aboutsummaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
authorBiwen Li <biwen.li@nxp.com>2018-12-12 09:56:18 +0800
committerHauke Mehrtens <hauke@hauke-m.de>2018-12-18 20:17:23 +0100
commit68904cb8fda3692c19fb39d3f99633c9d12efed7 (patch)
tree56b86c6463e7a554acc4c1e55d6260dc81f28268
parent328530c6e7569d7be24e3524483f4453910003e9 (diff)
downloadupstream-68904cb8fda3692c19fb39d3f99633c9d12efed7.tar.gz
upstream-68904cb8fda3692c19fb39d3f99633c9d12efed7.tar.bz2
upstream-68904cb8fda3692c19fb39d3f99633c9d12efed7.zip
layerscape: drop kernel 4.9 support
This patch is to drop kernel 4.9 support. Signed-off-by: Biwen Li <biwen.li@nxp.com> Signed-off-by: Yangbo Lu <yangbo.lu@nxp.com>
-rw-r--r--target/linux/layerscape/armv7/config-4.9952
-rw-r--r--target/linux/layerscape/armv8_32b/config-4.91211
-rw-r--r--target/linux/layerscape/armv8_64b/config-4.91242
-rw-r--r--target/linux/layerscape/patches-4.9/201-config-support-layerscape.patch374
-rw-r--r--target/linux/layerscape/patches-4.9/202-core-linux-support-layerscape.patch1458
-rw-r--r--target/linux/layerscape/patches-4.9/301-arch-support-layerscape.patch517
-rw-r--r--target/linux/layerscape/patches-4.9/302-dts-support-layercape.patch10549
-rw-r--r--target/linux/layerscape/patches-4.9/303-dts-layerscape-add-traverse-ls1043.patch95
-rw-r--r--target/linux/layerscape/patches-4.9/401-mtd-spi-nor-support-layerscape.patch993
-rw-r--r--target/linux/layerscape/patches-4.9/402-mtd-support-layerscape.patch392
-rw-r--r--target/linux/layerscape/patches-4.9/701-sdk_dpaa-support-layerscape.patch154756
-rw-r--r--target/linux/layerscape/patches-4.9/702-pci-support-layerscape.patch2098
-rw-r--r--target/linux/layerscape/patches-4.9/703-phy-support-layerscape.patch1763
-rw-r--r--target/linux/layerscape/patches-4.9/704-fsl-mc-layerscape-support.patch20043
-rw-r--r--target/linux/layerscape/patches-4.9/705-dpaa2-support-layerscape.patch23038
-rw-r--r--target/linux/layerscape/patches-4.9/706-fsl_ppfe-support-layercape.patch10654
-rw-r--r--target/linux/layerscape/patches-4.9/801-ata-support-layerscape.patch144
-rw-r--r--target/linux/layerscape/patches-4.9/802-clk-support-layerscape.patch323
-rw-r--r--target/linux/layerscape/patches-4.9/803-cpufreq-support-layerscape.patch1030
-rw-r--r--target/linux/layerscape/patches-4.9/804-crypto-support-layerscape.patch29223
-rw-r--r--target/linux/layerscape/patches-4.9/805-dma-support-layerscape.patch4372
-rw-r--r--target/linux/layerscape/patches-4.9/806-flextimer-support-layerscape.patch404
-rw-r--r--target/linux/layerscape/patches-4.9/807-gpu-support-layerscape.patch58
-rw-r--r--target/linux/layerscape/patches-4.9/808-guts-support-layerscape.patch421
-rw-r--r--target/linux/layerscape/patches-4.9/809-i2c-support-layerscape.patch360
-rw-r--r--target/linux/layerscape/patches-4.9/810-iommu-support-layerscape.patch1671
-rw-r--r--target/linux/layerscape/patches-4.9/811-irqchip-support-layerscape.patch182
-rw-r--r--target/linux/layerscape/patches-4.9/812-mmc-layerscape-support.patch594
-rw-r--r--target/linux/layerscape/patches-4.9/813-qe-support-layerscape.patch1976
-rw-r--r--target/linux/layerscape/patches-4.9/814-rtc-support-layerscape.patch682
-rw-r--r--target/linux/layerscape/patches-4.9/815-spi-support-layerscape.patch427
-rw-r--r--target/linux/layerscape/patches-4.9/816-tty-serial-support-layerscape.patch162
-rw-r--r--target/linux/layerscape/patches-4.9/817-usb-support-layerscape.patch1790
-rw-r--r--target/linux/layerscape/patches-4.9/818-vfio-support-layerscape.patch1165
-rw-r--r--target/linux/layerscape/patches-4.9/819-flexcan-support-layerscape.patch542
-rw-r--r--target/linux/layerscape/patches-4.9/820-kvm-support-layerscape.patch239
-rw-r--r--target/linux/layerscape/patches-4.9/821-add-esdhc-vsel-to-ls1043.patch10
-rw-r--r--target/linux/layerscape/patches-4.9/822-rgmii-fixed-link.patch62
38 files changed, 0 insertions, 275972 deletions
diff --git a/target/linux/layerscape/armv7/config-4.9 b/target/linux/layerscape/armv7/config-4.9
deleted file mode 100644
index 7c400527b3..0000000000
--- a/target/linux/layerscape/armv7/config-4.9
+++ /dev/null
@@ -1,952 +0,0 @@
-CONFIG_ABX500_CORE=y
-CONFIG_AD525X_DPOT=y
-CONFIG_AD525X_DPOT_I2C=y
-# CONFIG_AD525X_DPOT_SPI is not set
-CONFIG_AHCI_IMX=y
-CONFIG_AHCI_QORIQ=y
-CONFIG_AK8975=y
-CONFIG_ALIGNMENT_TRAP=y
-CONFIG_APDS9802ALS=y
-CONFIG_AQUANTIA_PHY=y
-# CONFIG_ARCH_AXXIA is not set
-CONFIG_ARCH_CLOCKSOURCE_DATA=y
-CONFIG_ARCH_DMA_ADDR_T_64BIT=y
-CONFIG_ARCH_HAS_ELF_RANDOMIZE=y
-CONFIG_ARCH_HAS_GCOV_PROFILE_ALL=y
-CONFIG_ARCH_HAS_RESET_CONTROLLER=y
-CONFIG_ARCH_HAS_SG_CHAIN=y
-CONFIG_ARCH_HAS_TICK_BROADCAST=y
-CONFIG_ARCH_HAVE_CUSTOM_GPIO_H=y
-CONFIG_ARCH_HIBERNATION_POSSIBLE=y
-CONFIG_ARCH_MIGHT_HAVE_PC_PARPORT=y
-CONFIG_ARCH_MULTIPLATFORM=y
-# CONFIG_ARCH_MULTI_CPU_AUTO is not set
-CONFIG_ARCH_MULTI_V6_V7=y
-CONFIG_ARCH_MULTI_V7=y
-CONFIG_ARCH_MXC=y
-CONFIG_ARCH_NR_GPIO=0
-CONFIG_ARCH_PHYS_ADDR_T_64BIT=y
-# CONFIG_ARCH_SELECT_MEMORY_MODEL is not set
-# CONFIG_ARCH_SPARSEMEM_DEFAULT is not set
-CONFIG_ARCH_SUPPORTS_ATOMIC_RMW=y
-CONFIG_ARCH_SUPPORTS_BIG_ENDIAN=y
-CONFIG_ARCH_SUPPORTS_UPROBES=y
-CONFIG_ARCH_SUSPEND_POSSIBLE=y
-CONFIG_ARCH_USE_BUILTIN_BSWAP=y
-CONFIG_ARCH_USE_CMPXCHG_LOCKREF=y
-CONFIG_ARCH_WANT_GENERAL_HUGETLB=y
-CONFIG_ARCH_WANT_IPC_PARSE_VERSION=y
-CONFIG_ARM=y
-CONFIG_ARM_APPENDED_DTB=y
-CONFIG_ARM_ARCH_TIMER=y
-CONFIG_ARM_ARCH_TIMER_EVTSTREAM=y
-CONFIG_ARM_ATAG_DTB_COMPAT=y
-# CONFIG_ARM_ATAG_DTB_COMPAT_CMDLINE_EXTEND is not set
-CONFIG_ARM_ATAG_DTB_COMPAT_CMDLINE_FROM_BOOTLOADER=y
-CONFIG_ARM_CPUIDLE=y
-CONFIG_ARM_CPU_SUSPEND=y
-CONFIG_ARM_CRYPTO=y
-CONFIG_ARM_ERRATA_430973=y
-CONFIG_ARM_ERRATA_643719=y
-CONFIG_ARM_ERRATA_720789=y
-CONFIG_ARM_ERRATA_754322=y
-CONFIG_ARM_ERRATA_754327=y
-CONFIG_ARM_ERRATA_764369=y
-CONFIG_ARM_ERRATA_775420=y
-CONFIG_ARM_ERRATA_798181=y
-CONFIG_ARM_GIC=y
-CONFIG_ARM_HAS_SG_CHAIN=y
-CONFIG_ARM_HEAVY_MB=y
-# CONFIG_ARM_HIGHBANK_CPUIDLE is not set
-CONFIG_ARM_IMX6Q_CPUFREQ=y
-CONFIG_ARM_L1_CACHE_SHIFT=6
-CONFIG_ARM_L1_CACHE_SHIFT_6=y
-CONFIG_ARM_LPAE=y
-CONFIG_ARM_PATCH_IDIV=y
-CONFIG_ARM_PATCH_PHYS_VIRT=y
-CONFIG_ARM_PMU=y
-CONFIG_ARM_PSCI=y
-CONFIG_ARM_PSCI_FW=y
-# CONFIG_ARM_SMMU is not set
-CONFIG_ARM_THUMB=y
-CONFIG_ARM_THUMBEE=y
-CONFIG_ARM_TIMER_SP804=y
-CONFIG_ARM_UNWIND=y
-CONFIG_ARM_VIRT_EXT=y
-CONFIG_ASSOCIATIVE_ARRAY=y
-CONFIG_AT803X_PHY=y
-CONFIG_ATA=y
-CONFIG_ATAGS=y
-CONFIG_ATA_VERBOSE_ERROR=y
-CONFIG_AUTOFS4_FS=y
-CONFIG_AUTO_ZRELADDR=y
-# CONFIG_AXP20X_POWER is not set
-CONFIG_BATTERY_ACT8945A=y
-CONFIG_BATTERY_SBS=y
-CONFIG_BCM_NET_PHYLIB=y
-# CONFIG_BLK_CGROUP is not set
-CONFIG_BLK_CMDLINE_PARSER=y
-CONFIG_BLK_DEV_BSG=y
-CONFIG_BLK_DEV_LOOP=y
-CONFIG_BLK_DEV_RAM=y
-CONFIG_BLK_DEV_RAM_COUNT=16
-CONFIG_BLK_DEV_RAM_SIZE=262144
-CONFIG_BLK_DEV_SD=y
-CONFIG_BLK_DEV_SR=y
-# CONFIG_BLK_DEV_SR_VENDOR is not set
-CONFIG_BLK_MQ_PCI=y
-# CONFIG_BOOTPARAM_HUNG_TASK_PANIC is not set
-CONFIG_BOOTPARAM_HUNG_TASK_PANIC_VALUE=0
-# CONFIG_BOOTPARAM_SOFTLOCKUP_PANIC is not set
-CONFIG_BOOTPARAM_SOFTLOCKUP_PANIC_VALUE=0
-CONFIG_BOUNCE=y
-# CONFIG_BPF_SYSCALL is not set
-CONFIG_BRCMSTB_GISB_ARB=y
-CONFIG_BROADCOM_PHY=y
-CONFIG_BUILD_BIN2C=y
-CONFIG_CACHE_L2X0=y
-# CONFIG_CACHE_L2X0_PMU is not set
-CONFIG_CC_OPTIMIZE_FOR_PERFORMANCE=y
-# CONFIG_CC_OPTIMIZE_FOR_SIZE is not set
-# CONFIG_CFS_BANDWIDTH is not set
-CONFIG_CGROUPS=y
-CONFIG_CGROUP_CPUACCT=y
-CONFIG_CGROUP_DEVICE=y
-CONFIG_CGROUP_FREEZER=y
-# CONFIG_CGROUP_PERF is not set
-CONFIG_CGROUP_PIDS=y
-CONFIG_CGROUP_SCHED=y
-# CONFIG_CHARGER_MAX14577 is not set
-CONFIG_CHARGER_TPS65090=y
-# CONFIG_CHARGER_TPS65217 is not set
-CONFIG_CHECKPOINT_RESTORE=y
-CONFIG_CHROME_PLATFORMS=y
-CONFIG_CHR_DEV_SG=y
-CONFIG_CLKDEV_LOOKUP=y
-CONFIG_CLKSRC_IMX_GPT=y
-CONFIG_CLKSRC_MMIO=y
-CONFIG_CLKSRC_OF=y
-CONFIG_CLKSRC_PROBE=y
-CONFIG_CLKSRC_VERSATILE=y
-CONFIG_CLK_QORIQ=y
-CONFIG_CLONE_BACKWARDS=y
-CONFIG_CMA=y
-CONFIG_CMA_ALIGNMENT=8
-CONFIG_CMA_AREAS=7
-# CONFIG_CMA_DEBUG is not set
-# CONFIG_CMA_DEBUGFS is not set
-CONFIG_CMA_SIZE_MBYTES=64
-# CONFIG_CMA_SIZE_SEL_MAX is not set
-CONFIG_CMA_SIZE_SEL_MBYTES=y
-# CONFIG_CMA_SIZE_SEL_MIN is not set
-# CONFIG_CMA_SIZE_SEL_PERCENTAGE is not set
-CONFIG_CMDLINE_PARTITION=y
-CONFIG_COMMON_CLK=y
-CONFIG_COMMON_CLK_MAX77686=y
-# CONFIG_COMMON_CLK_PALMAS is not set
-# CONFIG_COMMON_CLK_RK808 is not set
-# CONFIG_COMMON_CLK_S2MPS11 is not set
-CONFIG_COMPAT_BRK=y
-CONFIG_CONFIGFS_FS=y
-CONFIG_CONSOLE_TRANSLATIONS=y
-CONFIG_COREDUMP=y
-CONFIG_CORE_DUMP_DEFAULT_ELF_HEADERS=y
-# CONFIG_CORTINA_PHY is not set
-CONFIG_CPUFREQ_DT=y
-CONFIG_CPUFREQ_DT_PLATDEV=y
-CONFIG_CPUSETS=y
-CONFIG_CPU_32v6K=y
-CONFIG_CPU_32v7=y
-CONFIG_CPU_ABRT_EV7=y
-# CONFIG_CPU_BIG_ENDIAN is not set
-# CONFIG_CPU_BPREDICT_DISABLE is not set
-CONFIG_CPU_CACHE_V7=y
-CONFIG_CPU_CACHE_VIPT=y
-CONFIG_CPU_COPY_V6=y
-CONFIG_CPU_CP15=y
-CONFIG_CPU_CP15_MMU=y
-CONFIG_CPU_FREQ=y
-CONFIG_CPU_FREQ_DEFAULT_GOV_ONDEMAND=y
-# CONFIG_CPU_FREQ_DEFAULT_GOV_PERFORMANCE is not set
-CONFIG_CPU_FREQ_GOV_ATTR_SET=y
-CONFIG_CPU_FREQ_GOV_COMMON=y
-# CONFIG_CPU_FREQ_GOV_CONSERVATIVE is not set
-CONFIG_CPU_FREQ_GOV_ONDEMAND=y
-CONFIG_CPU_FREQ_GOV_PERFORMANCE=y
-# CONFIG_CPU_FREQ_GOV_POWERSAVE is not set
-CONFIG_CPU_FREQ_GOV_SCHEDUTIL=y
-# CONFIG_CPU_FREQ_GOV_USERSPACE is not set
-CONFIG_CPU_FREQ_STAT=y
-CONFIG_CPU_HAS_ASID=y
-# CONFIG_CPU_HOTPLUG_STATE_CONTROL is not set
-# CONFIG_CPU_ICACHE_DISABLE is not set
-CONFIG_CPU_IDLE=y
-# CONFIG_CPU_IDLE_GOV_LADDER is not set
-CONFIG_CPU_IDLE_GOV_MENU=y
-CONFIG_CPU_PABRT_V7=y
-CONFIG_CPU_PM=y
-CONFIG_CPU_RMAP=y
-CONFIG_CPU_THERMAL=y
-CONFIG_CPU_TLB_V7=y
-CONFIG_CPU_V7=y
-CONFIG_CRC16=y
-# CONFIG_CRC32_SARWATE is not set
-CONFIG_CRC32_SLICEBY8=y
-CONFIG_CROSS_MEMORY_ATTACH=y
-CONFIG_CRYPTO_AEAD=y
-CONFIG_CRYPTO_AEAD2=y
-# CONFIG_CRYPTO_AES_ARM_CE is not set
-CONFIG_CRYPTO_CRC32C=y
-CONFIG_CRYPTO_DEFLATE=y
-# CONFIG_CRYPTO_GHASH_ARM_CE is not set
-CONFIG_CRYPTO_HASH=y
-CONFIG_CRYPTO_HASH2=y
-CONFIG_CRYPTO_LZO=y
-CONFIG_CRYPTO_MANAGER=y
-CONFIG_CRYPTO_MANAGER2=y
-CONFIG_CRYPTO_NULL2=y
-# CONFIG_CRYPTO_PCRYPT is not set
-CONFIG_CRYPTO_RNG2=y
-# CONFIG_CRYPTO_SHA1_ARM_CE is not set
-# CONFIG_CRYPTO_SHA1_ARM_NEON is not set
-# CONFIG_CRYPTO_SHA256_ARM is not set
-# CONFIG_CRYPTO_SHA2_ARM_CE is not set
-# CONFIG_CRYPTO_SHA512_ARM is not set
-# CONFIG_CRYPTO_TLS is not set
-CONFIG_CRYPTO_WORKQUEUE=y
-CONFIG_DCACHE_WORD_ACCESS=y
-CONFIG_DEBUG_ALIGN_RODATA=y
-CONFIG_DEBUG_BUGVERBOSE=y
-CONFIG_DEBUG_IMX_UART_PORT=1
-CONFIG_DEBUG_LL_INCLUDE="mach/debug-macro.S"
-CONFIG_DEBUG_RODATA=y
-# CONFIG_DEBUG_UART_8250 is not set
-# CONFIG_DEBUG_USER is not set
-CONFIG_DECOMPRESS_BZIP2=y
-CONFIG_DECOMPRESS_GZIP=y
-CONFIG_DECOMPRESS_LZMA=y
-CONFIG_DECOMPRESS_LZO=y
-CONFIG_DECOMPRESS_XZ=y
-CONFIG_DEFAULT_CFQ=y
-# CONFIG_DEFAULT_DEADLINE is not set
-CONFIG_DEFAULT_HUNG_TASK_TIMEOUT=120
-CONFIG_DEFAULT_IOSCHED="cfq"
-CONFIG_DETECT_HUNG_TASK=y
-# CONFIG_DEVFREQ_GOV_PASSIVE is not set
-# CONFIG_DEVFREQ_GOV_PERFORMANCE is not set
-# CONFIG_DEVFREQ_GOV_POWERSAVE is not set
-# CONFIG_DEVFREQ_GOV_SIMPLE_ONDEMAND is not set
-# CONFIG_DEVFREQ_GOV_USERSPACE is not set
-# CONFIG_DEVFREQ_THERMAL is not set
-CONFIG_DEVKMEM=y
-CONFIG_DEVMEM=y
-CONFIG_DEVTMPFS=y
-CONFIG_DEVTMPFS_MOUNT=y
-CONFIG_DMADEVICES=y
-CONFIG_DMA_CMA=y
-CONFIG_DMA_ENGINE=y
-CONFIG_DMA_OF=y
-CONFIG_DMA_SHARED_BUFFER=y
-CONFIG_DMA_VIRTUAL_CHANNELS=y
-CONFIG_DNOTIFY=y
-CONFIG_DNS_RESOLVER=y
-CONFIG_DTC=y
-CONFIG_DT_IDLE_STATES=y
-CONFIG_DUMMY_CONSOLE=y
-CONFIG_DW_DMAC=y
-CONFIG_DW_DMAC_CORE=y
-CONFIG_DW_WATCHDOG=y
-CONFIG_EDAC=y
-CONFIG_EDAC_ATOMIC_SCRUB=y
-# CONFIG_EDAC_DEBUG is not set
-CONFIG_EDAC_LEGACY_SYSFS=y
-# CONFIG_EDAC_MM_EDAC is not set
-CONFIG_EDAC_SUPPORT=y
-CONFIG_EEPROM_93CX6=y
-CONFIG_EEPROM_AT24=y
-CONFIG_EFI=y
-# CONFIG_EFIVAR_FS is not set
-CONFIG_EFI_ARMSTUB=y
-# CONFIG_EFI_CAPSULE_LOADER is not set
-CONFIG_EFI_ESRT=y
-CONFIG_EFI_PARAMS_FROM_FDT=y
-CONFIG_EFI_RUNTIME_WRAPPERS=y
-CONFIG_EFI_STUB=y
-# CONFIG_EFI_TEST is not set
-# CONFIG_EFI_VARS is not set
-CONFIG_ELF_CORE=y
-# CONFIG_ENABLE_DEFAULT_TRACERS is not set
-CONFIG_ENABLE_MUST_CHECK=y
-CONFIG_EXPORTFS=y
-CONFIG_EXT4_FS=y
-CONFIG_EXT4_FS_POSIX_ACL=y
-CONFIG_EXT4_FS_SECURITY=y
-CONFIG_EXTCON=y
-# CONFIG_EXTCON_MAX14577 is not set
-# CONFIG_EXTCON_MAX8997 is not set
-# CONFIG_EXTCON_PALMAS is not set
-CONFIG_FAIR_GROUP_SCHED=y
-CONFIG_FAT_FS=y
-# CONFIG_FEC is not set
-CONFIG_FHANDLE=y
-CONFIG_FIRMWARE_IN_KERNEL=y
-CONFIG_FIXED_PHY=y
-CONFIG_FIX_EARLYCON_MEM=y
-CONFIG_FORCE_MAX_ZONEORDER=12
-CONFIG_FREEZER=y
-CONFIG_FSL_EDMA=y
-CONFIG_FSL_GUTS=y
-CONFIG_FSL_IFC=y
-# CONFIG_FSL_PPFE is not set
-CONFIG_FSL_PQ_MDIO=y
-# CONFIG_FSL_QDMA is not set
-# CONFIG_FSL_SDK_DPA is not set
-CONFIG_FSL_XGMAC_MDIO=y
-CONFIG_FS_MBCACHE=y
-CONFIG_FS_POSIX_ACL=y
-CONFIG_FTM_ALARM=y
-CONFIG_FTRACE=y
-# CONFIG_FTRACE_SYSCALLS is not set
-CONFIG_FUSE_FS=y
-# CONFIG_FW_LOADER_USER_HELPER_FALLBACK is not set
-CONFIG_GENERIC_ALLOCATOR=y
-CONFIG_GENERIC_BUG=y
-CONFIG_GENERIC_CLOCKEVENTS=y
-CONFIG_GENERIC_CLOCKEVENTS_BROADCAST=y
-CONFIG_GENERIC_EARLY_IOREMAP=y
-CONFIG_GENERIC_IDLE_POLL_SETUP=y
-CONFIG_GENERIC_IO=y
-CONFIG_GENERIC_IRQ_CHIP=y
-# CONFIG_GENERIC_IRQ_DEBUGFS is not set
-CONFIG_GENERIC_IRQ_SHOW=y
-CONFIG_GENERIC_IRQ_SHOW_LEVEL=y
-CONFIG_GENERIC_MSI_IRQ=y
-CONFIG_GENERIC_MSI_IRQ_DOMAIN=y
-CONFIG_GENERIC_PCI_IOMAP=y
-CONFIG_GENERIC_PHY=y
-CONFIG_GENERIC_PINCONF=y
-CONFIG_GENERIC_SCHED_CLOCK=y
-CONFIG_GENERIC_SMP_IDLE_THREAD=y
-CONFIG_GENERIC_STRNCPY_FROM_USER=y
-CONFIG_GENERIC_STRNLEN_USER=y
-CONFIG_GENERIC_TIME_VSYSCALL=y
-CONFIG_GIANFAR=y
-CONFIG_GLOB=y
-CONFIG_GPIOLIB=y
-CONFIG_GPIOLIB_IRQCHIP=y
-# CONFIG_GPIO_AXP209 is not set
-CONFIG_GPIO_DWAPB=y
-CONFIG_GPIO_EM=y
-CONFIG_GPIO_GENERIC=y
-CONFIG_GPIO_GENERIC_PLATFORM=y
-CONFIG_GPIO_MXC=y
-CONFIG_GPIO_PALMAS=y
-CONFIG_GPIO_PCA953X=y
-CONFIG_GPIO_PCA953X_IRQ=y
-CONFIG_GPIO_PCF857X=y
-# CONFIG_GPIO_STMPE is not set
-CONFIG_GPIO_SYSCON=y
-# CONFIG_GPIO_TPS65218 is not set
-CONFIG_GPIO_TPS6586X=y
-CONFIG_GPIO_TPS65910=y
-CONFIG_GPIO_TWL4030=y
-CONFIG_GPIO_XILINX=y
-CONFIG_HANDLE_DOMAIN_IRQ=y
-CONFIG_HARDEN_BRANCH_PREDICTOR=y
-CONFIG_HARDIRQS_SW_RESEND=y
-CONFIG_HAS_DMA=y
-CONFIG_HAS_IOMEM=y
-CONFIG_HAS_IOPORT_MAP=y
-# CONFIG_HAVE_64BIT_ALIGNED_ACCESS is not set
-CONFIG_HAVE_ARCH_AUDITSYSCALL=y
-CONFIG_HAVE_ARCH_BITREVERSE=y
-CONFIG_HAVE_ARCH_JUMP_LABEL=y
-CONFIG_HAVE_ARCH_KGDB=y
-CONFIG_HAVE_ARCH_PFN_VALID=y
-CONFIG_HAVE_ARCH_SECCOMP_FILTER=y
-CONFIG_HAVE_ARCH_TRACEHOOK=y
-CONFIG_HAVE_ARCH_TRANSPARENT_HUGEPAGE=y
-CONFIG_HAVE_ARM_ARCH_TIMER=y
-CONFIG_HAVE_ARM_SMCCC=y
-# CONFIG_HAVE_BOOTMEM_INFO_NODE is not set
-CONFIG_HAVE_CBPF_JIT=y
-CONFIG_HAVE_CC_STACKPROTECTOR=y
-CONFIG_HAVE_CLK=y
-CONFIG_HAVE_CLK_PREPARE=y
-CONFIG_HAVE_CONTEXT_TRACKING=y
-CONFIG_HAVE_C_RECORDMCOUNT=y
-CONFIG_HAVE_DEBUG_KMEMLEAK=y
-CONFIG_HAVE_DMA_API_DEBUG=y
-CONFIG_HAVE_DMA_CONTIGUOUS=y
-CONFIG_HAVE_DYNAMIC_FTRACE=y
-CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS=y
-CONFIG_HAVE_FTRACE_MCOUNT_RECORD=y
-CONFIG_HAVE_FUNCTION_GRAPH_TRACER=y
-CONFIG_HAVE_FUNCTION_TRACER=y
-CONFIG_HAVE_GENERIC_DMA_COHERENT=y
-CONFIG_HAVE_GENERIC_RCU_GUP=y
-CONFIG_HAVE_HW_BREAKPOINT=y
-CONFIG_HAVE_IDE=y
-CONFIG_HAVE_IMX_SRC=y
-CONFIG_HAVE_IRQ_TIME_ACCOUNTING=y
-CONFIG_HAVE_KVM_ARCH_TLB_FLUSH_ALL=y
-CONFIG_HAVE_KVM_CPU_RELAX_INTERCEPT=y
-CONFIG_HAVE_KVM_EVENTFD=y
-CONFIG_HAVE_KVM_IRQCHIP=y
-CONFIG_HAVE_KVM_IRQFD=y
-CONFIG_HAVE_KVM_IRQ_ROUTING=y
-CONFIG_HAVE_MEMBLOCK=y
-CONFIG_HAVE_MOD_ARCH_SPECIFIC=y
-CONFIG_HAVE_NET_DSA=y
-CONFIG_HAVE_OPROFILE=y
-CONFIG_HAVE_OPTPROBES=y
-CONFIG_HAVE_PERF_EVENTS=y
-CONFIG_HAVE_PERF_REGS=y
-CONFIG_HAVE_PERF_USER_STACK_DUMP=y
-CONFIG_HAVE_PROC_CPU=y
-CONFIG_HAVE_RCU_TABLE_FREE=y
-CONFIG_HAVE_REGS_AND_STACK_ACCESS_API=y
-CONFIG_HAVE_SMP=y
-CONFIG_HAVE_SYSCALL_TRACEPOINTS=y
-CONFIG_HAVE_UID16=y
-CONFIG_HAVE_VIRT_CPU_ACCOUNTING_GEN=y
-CONFIG_HID=y
-CONFIG_HID_GENERIC=y
-CONFIG_HIGHMEM=y
-CONFIG_HIGHPTE=y
-CONFIG_HOTPLUG_CPU=y
-# CONFIG_HUGETLBFS is not set
-CONFIG_HVC_DRIVER=y
-CONFIG_HWMON=y
-CONFIG_HW_CONSOLE=y
-CONFIG_HW_RANDOM=y
-CONFIG_HZ_FIXED=0
-CONFIG_I2C=y
-CONFIG_I2C_BOARDINFO=y
-CONFIG_I2C_CHARDEV=y
-CONFIG_I2C_COMPAT=y
-CONFIG_I2C_DEMUX_PINCTRL=y
-CONFIG_I2C_DESIGNWARE_CORE=y
-CONFIG_I2C_DESIGNWARE_PLATFORM=y
-CONFIG_I2C_HELPER_AUTO=y
-CONFIG_I2C_IMX=y
-CONFIG_I2C_MUX=y
-CONFIG_I2C_MUX_PCA954x=y
-CONFIG_I2C_MUX_PINCTRL=y
-CONFIG_I2C_RK3X=y
-CONFIG_I2C_SLAVE=y
-CONFIG_I2C_SLAVE_EEPROM=y
-CONFIG_I2C_XILINX=y
-CONFIG_ICPLUS_PHY=y
-CONFIG_ICS932S401=y
-CONFIG_IIO=y
-CONFIG_IIO_BUFFER=y
-CONFIG_IIO_CONFIGFS=y
-CONFIG_IIO_HRTIMER_TRIGGER=y
-CONFIG_IIO_KFIFO_BUF=y
-CONFIG_IIO_SW_TRIGGER=y
-# CONFIG_IIO_TIGHTLOOP_TRIGGER is not set
-CONFIG_IIO_TRIGGER=y
-CONFIG_IIO_TRIGGERED_BUFFER=y
-CONFIG_IKCONFIG=y
-CONFIG_IKCONFIG_PROC=y
-CONFIG_IMX2_WDT=y
-CONFIG_IMX_DMA=y
-CONFIG_IMX_SDMA=y
-# CONFIG_IMX_WEIM is not set
-CONFIG_INITRAMFS_SOURCE=""
-CONFIG_INPUT=y
-# CONFIG_INPUT_MISC is not set
-CONFIG_IOMMU_HELPER=y
-# CONFIG_IOMMU_IO_PGTABLE_ARMV7S is not set
-# CONFIG_IOMMU_IO_PGTABLE_LPAE is not set
-CONFIG_IOMMU_SUPPORT=y
-CONFIG_IOSCHED_CFQ=y
-CONFIG_IPC_NS=y
-CONFIG_IRQCHIP=y
-CONFIG_IRQ_DOMAIN=y
-CONFIG_IRQ_DOMAIN_DEBUG=y
-CONFIG_IRQ_DOMAIN_HIERARCHY=y
-CONFIG_IRQ_FORCED_THREADING=y
-CONFIG_IRQ_WORK=y
-# CONFIG_ISDN is not set
-CONFIG_ISL29003=y
-CONFIG_JBD2=y
-CONFIG_KALLSYMS=y
-CONFIG_KERNEL_GZIP=y
-# CONFIG_KERNEL_XZ is not set
-CONFIG_KEXEC=y
-CONFIG_KEXEC_CORE=y
-CONFIG_KEYS=y
-CONFIG_KVM=y
-CONFIG_KVM_ARM_HOST=y
-CONFIG_KVM_GENERIC_DIRTYLOG_READ_PROTECT=y
-CONFIG_KVM_MMIO=y
-CONFIG_KVM_VFIO=y
-CONFIG_LEDS_GPIO=y
-# CONFIG_LEDS_MAX8997 is not set
-CONFIG_LEDS_PWM=y
-CONFIG_LEDS_TRIGGER_BACKLIGHT=y
-CONFIG_LEDS_TRIGGER_CAMERA=y
-CONFIG_LEDS_TRIGGER_CPU=y
-CONFIG_LEDS_TRIGGER_GPIO=y
-CONFIG_LEDS_TRIGGER_HEARTBEAT=y
-CONFIG_LEDS_TRIGGER_ONESHOT=y
-CONFIG_LEDS_TRIGGER_TRANSIENT=y
-CONFIG_LEGACY_PTYS=y
-CONFIG_LEGACY_PTY_COUNT=256
-CONFIG_LIBFDT=y
-CONFIG_LOCALVERSION_AUTO=y
-CONFIG_LOCKUP_DETECTOR=y
-CONFIG_LOCK_SPIN_ON_OWNER=y
-CONFIG_LS_SCFG_MSI=y
-CONFIG_LS_SOC_DRIVERS=y
-CONFIG_LZO_COMPRESS=y
-CONFIG_LZO_DECOMPRESS=y
-CONFIG_MACVLAN=y
-CONFIG_MACVTAP=y
-CONFIG_MAGIC_SYSRQ=y
-CONFIG_MAILBOX=y
-# CONFIG_MAILBOX_TEST is not set
-CONFIG_MANDATORY_FILE_LOCKING=y
-CONFIG_MARVELL_PHY=y
-CONFIG_MCPM=y
-CONFIG_MDIO_BITBANG=y
-CONFIG_MDIO_BOARDINFO=y
-# CONFIG_MDIO_FSL_BACKPLANE is not set
-# CONFIG_MDIO_GPIO is not set
-CONFIG_MEMCG=y
-# CONFIG_MEMCG_SWAP is not set
-CONFIG_MEMORY=y
-CONFIG_MEMORY_ISOLATION=y
-CONFIG_MFD_ACT8945A=y
-CONFIG_MFD_AS3711=y
-CONFIG_MFD_AS3722=y
-CONFIG_MFD_ATMEL_FLEXCOM=y
-CONFIG_MFD_AXP20X=y
-CONFIG_MFD_AXP20X_I2C=y
-CONFIG_MFD_BCM590XX=y
-CONFIG_MFD_CORE=y
-CONFIG_MFD_MAX14577=y
-CONFIG_MFD_MAX77686=y
-CONFIG_MFD_MAX8907=y
-CONFIG_MFD_MAX8997=y
-CONFIG_MFD_MAX8998=y
-CONFIG_MFD_PALMAS=y
-CONFIG_MFD_RK808=y
-CONFIG_MFD_SEC_CORE=y
-CONFIG_MFD_STMPE=y
-CONFIG_MFD_SYSCON=y
-CONFIG_MFD_TPS65090=y
-CONFIG_MFD_TPS65217=y
-CONFIG_MFD_TPS65218=y
-CONFIG_MFD_TPS6586X=y
-CONFIG_MFD_TPS65910=y
-# CONFIG_MFD_TWL4030_AUDIO is not set
-CONFIG_MFD_VEXPRESS_SYSREG=y
-CONFIG_MICREL_PHY=y
-CONFIG_MIGHT_HAVE_CACHE_L2X0=y
-CONFIG_MIGHT_HAVE_PCI=y
-CONFIG_MIGRATION=y
-CONFIG_MMC=y
-CONFIG_MMC_BLOCK=y
-CONFIG_MMC_BLOCK_MINORS=16
-# CONFIG_MMC_MXC is not set
-CONFIG_MMC_SDHCI=y
-# CONFIG_MMC_SDHCI_ESDHC_IMX is not set
-CONFIG_MMC_SDHCI_IO_ACCESSORS=y
-CONFIG_MMC_SDHCI_OF_ESDHC=y
-# CONFIG_MMC_SDHCI_PCI is not set
-CONFIG_MMC_SDHCI_PLTFM=y
-# CONFIG_MMC_TIFM_SD is not set
-CONFIG_MMU_NOTIFIER=y
-CONFIG_MODULES_TREE_LOOKUP=y
-CONFIG_MODULES_USE_ELF_REL=y
-CONFIG_MSDOS_FS=y
-CONFIG_MTD_CFI_ADV_OPTIONS=y
-CONFIG_MTD_CFI_GEOMETRY=y
-CONFIG_MTD_CFI_STAA=y
-CONFIG_MTD_CMDLINE_PARTS=y
-# CONFIG_MTD_COMPLEX_MAPPINGS is not set
-CONFIG_MTD_DATAFLASH=y
-# CONFIG_MTD_DATAFLASH_OTP is not set
-# CONFIG_MTD_DATAFLASH_WRITE_VERIFY is not set
-CONFIG_MTD_M25P80=y
-CONFIG_MTD_NAND=y
-CONFIG_MTD_NAND_BRCMNAND=y
-CONFIG_MTD_NAND_DENALI=y
-CONFIG_MTD_NAND_DENALI_DT=y
-CONFIG_MTD_NAND_ECC=y
-CONFIG_MTD_NAND_FSL_IFC=y
-CONFIG_MTD_SPI_NOR=y
-CONFIG_MTD_SPI_NOR_USE_4K_SECTORS=y
-CONFIG_MTD_SST25L=y
-CONFIG_MTD_UBI=y
-CONFIG_MTD_UBI_BEB_LIMIT=20
-# CONFIG_MTD_UBI_BLOCK is not set
-# CONFIG_MTD_UBI_FASTMAP is not set
-# CONFIG_MTD_UBI_GLUEBI is not set
-CONFIG_MTD_UBI_WL_THRESHOLD=4096
-CONFIG_MULTI_IRQ_HANDLER=y
-CONFIG_MUTEX_SPIN_ON_OWNER=y
-CONFIG_MX3_IPU=y
-CONFIG_MX3_IPU_IRQS=4
-CONFIG_NAMESPACES=y
-CONFIG_NATIONAL_PHY=y
-CONFIG_NEED_DMA_MAP_STATE=y
-CONFIG_NEON=y
-# CONFIG_NET_CLS_CGROUP is not set
-CONFIG_NET_FLOW_LIMIT=y
-CONFIG_NET_NS=y
-CONFIG_NET_PTP_CLASSIFY=y
-CONFIG_NET_SWITCHDEV=y
-CONFIG_NLS=y
-CONFIG_NLS_CODEPAGE_437=y
-CONFIG_NLS_ISO8859_1=y
-CONFIG_NLS_UTF8=y
-CONFIG_NO_BOOTMEM=y
-CONFIG_NO_HZ=y
-CONFIG_NO_HZ_COMMON=y
-CONFIG_NO_HZ_IDLE=y
-CONFIG_NR_CPUS=16
-CONFIG_NTFS_FS=y
-CONFIG_NVMEM=y
-CONFIG_OF=y
-CONFIG_OF_ADDRESS=y
-CONFIG_OF_ADDRESS_PCI=y
-CONFIG_OF_DYNAMIC=y
-CONFIG_OF_EARLY_FLATTREE=y
-CONFIG_OF_FLATTREE=y
-CONFIG_OF_GPIO=y
-CONFIG_OF_IRQ=y
-CONFIG_OF_MDIO=y
-CONFIG_OF_NET=y
-CONFIG_OF_PCI=y
-CONFIG_OF_PCI_IRQ=y
-CONFIG_OF_RESERVED_MEM=y
-CONFIG_OLD_SIGACTION=y
-CONFIG_OLD_SIGSUSPEND3=y
-CONFIG_OUTER_CACHE=y
-CONFIG_OUTER_CACHE_SYNC=y
-CONFIG_PACKET_DIAG=y
-CONFIG_PAGE_COUNTER=y
-CONFIG_PAGE_OFFSET=0xC0000000
-# CONFIG_PANIC_ON_OOPS is not set
-CONFIG_PANIC_ON_OOPS_VALUE=0
-CONFIG_PANIC_TIMEOUT=0
-CONFIG_PCI=y
-CONFIG_PCIEAER=y
-CONFIG_PCIEASPM=y
-# CONFIG_PCIEASPM_DEBUG is not set
-CONFIG_PCIEASPM_DEFAULT=y
-# CONFIG_PCIEASPM_PERFORMANCE is not set
-# CONFIG_PCIEASPM_POWERSAVE is not set
-CONFIG_PCIEPORTBUS=y
-CONFIG_PCIE_DW=y
-CONFIG_PCIE_PME=y
-CONFIG_PCI_BUS_ADDR_T_64BIT=y
-CONFIG_PCI_DOMAINS=y
-CONFIG_PCI_DOMAINS_GENERIC=y
-CONFIG_PCI_ECAM=y
-CONFIG_PCI_HOST_COMMON=y
-CONFIG_PCI_HOST_GENERIC=y
-CONFIG_PCI_LAYERSCAPE=y
-CONFIG_PCI_MSI=y
-CONFIG_PCI_MSI_IRQ_DOMAIN=y
-CONFIG_PERF_EVENTS=y
-CONFIG_PERF_USE_VMALLOC=y
-CONFIG_PGTABLE_LEVELS=3
-CONFIG_PHYLIB=y
-CONFIG_PHYS_ADDR_T_64BIT=y
-CONFIG_PID_NS=y
-CONFIG_PINCTRL=y
-CONFIG_PINCTRL_AS3722=y
-CONFIG_PINCTRL_PALMAS=y
-CONFIG_PL310_ERRATA_588369=y
-CONFIG_PL310_ERRATA_727915=y
-CONFIG_PL310_ERRATA_753970=y
-CONFIG_PL310_ERRATA_769419=y
-CONFIG_PM=y
-CONFIG_PM_CLK=y
-# CONFIG_PM_DEBUG is not set
-CONFIG_PM_DEVFREQ=y
-# CONFIG_PM_DEVFREQ_EVENT is not set
-CONFIG_PM_OPP=y
-CONFIG_PM_SLEEP=y
-CONFIG_PM_SLEEP_SMP=y
-CONFIG_POSIX_MQUEUE=y
-CONFIG_POSIX_MQUEUE_SYSCTL=y
-CONFIG_POWER_RESET=y
-CONFIG_POWER_RESET_AS3722=y
-CONFIG_POWER_RESET_BRCMKONA=y
-CONFIG_POWER_RESET_BRCMSTB=y
-CONFIG_POWER_RESET_GPIO=y
-CONFIG_POWER_RESET_GPIO_RESTART=y
-CONFIG_POWER_RESET_SYSCON=y
-CONFIG_POWER_RESET_SYSCON_POWEROFF=y
-CONFIG_POWER_RESET_VEXPRESS=y
-CONFIG_POWER_SUPPLY=y
-CONFIG_PPS=y
-CONFIG_PREEMPT_NOTIFIERS=y
-CONFIG_PRINTK_TIME=y
-# CONFIG_PROBE_EVENTS is not set
-CONFIG_PROC_CHILDREN=y
-CONFIG_PROC_PAGE_MONITOR=y
-CONFIG_PROC_PID_CPUSET=y
-CONFIG_PSTORE=y
-CONFIG_PSTORE_CONSOLE=y
-# CONFIG_PSTORE_LZ4_COMPRESS is not set
-# CONFIG_PSTORE_LZO_COMPRESS is not set
-CONFIG_PSTORE_PMSG=y
-CONFIG_PSTORE_RAM=y
-CONFIG_PSTORE_ZLIB_COMPRESS=y
-CONFIG_PTP_1588_CLOCK=y
-CONFIG_PTP_1588_CLOCK_GIANFAR=y
-CONFIG_PWM=y
-# CONFIG_PWM_IMX is not set
-# CONFIG_PWM_STMPE is not set
-CONFIG_PWM_SYSFS=y
-# CONFIG_PWM_TWL is not set
-# CONFIG_PWM_TWL_LED is not set
-CONFIG_QORIQ_CPUFREQ=y
-# CONFIG_QUICC_ENGINE is not set
-CONFIG_RAS=y
-CONFIG_RATIONAL=y
-CONFIG_RCU_CPU_STALL_TIMEOUT=21
-# CONFIG_RCU_EXPERT is not set
-CONFIG_RCU_STALL_COMMON=y
-CONFIG_RD_BZIP2=y
-CONFIG_RD_GZIP=y
-CONFIG_RD_LZMA=y
-CONFIG_RD_LZO=y
-CONFIG_RD_XZ=y
-CONFIG_REALTEK_PHY=y
-CONFIG_REED_SOLOMON=y
-CONFIG_REED_SOLOMON_DEC8=y
-CONFIG_REED_SOLOMON_ENC8=y
-CONFIG_REGMAP=y
-CONFIG_REGMAP_I2C=y
-CONFIG_REGMAP_IRQ=y
-CONFIG_REGMAP_MMIO=y
-CONFIG_REGMAP_SPI=y
-CONFIG_REGULATOR=y
-CONFIG_REGULATOR_ACT8865=y
-CONFIG_REGULATOR_ACT8945A=y
-CONFIG_REGULATOR_ANATOP=y
-CONFIG_REGULATOR_AS3711=y
-CONFIG_REGULATOR_AS3722=y
-CONFIG_REGULATOR_AXP20X=y
-CONFIG_REGULATOR_BCM590XX=y
-CONFIG_REGULATOR_DA9210=y
-CONFIG_REGULATOR_FAN53555=y
-CONFIG_REGULATOR_FIXED_VOLTAGE=y
-CONFIG_REGULATOR_GPIO=y
-CONFIG_REGULATOR_LP872X=y
-# CONFIG_REGULATOR_MAX14577 is not set
-CONFIG_REGULATOR_MAX77686=y
-# CONFIG_REGULATOR_MAX77802 is not set
-CONFIG_REGULATOR_MAX8907=y
-CONFIG_REGULATOR_MAX8973=y
-# CONFIG_REGULATOR_MAX8997 is not set
-# CONFIG_REGULATOR_MAX8998 is not set
-CONFIG_REGULATOR_PALMAS=y
-CONFIG_REGULATOR_PWM=y
-# CONFIG_REGULATOR_QCOM_SPMI is not set
-CONFIG_REGULATOR_RK808=y
-# CONFIG_REGULATOR_S2MPA01 is not set
-CONFIG_REGULATOR_S2MPS11=y
-CONFIG_REGULATOR_S5M8767=y
-CONFIG_REGULATOR_TPS51632=y
-CONFIG_REGULATOR_TPS62360=y
-CONFIG_REGULATOR_TPS65090=y
-CONFIG_REGULATOR_TPS65217=y
-CONFIG_REGULATOR_TPS65218=y
-CONFIG_REGULATOR_TPS6586X=y
-CONFIG_REGULATOR_TPS65910=y
-CONFIG_REGULATOR_TWL4030=y
-CONFIG_REGULATOR_VEXPRESS=y
-CONFIG_RESET_CONTROLLER=y
-CONFIG_RFS_ACCEL=y
-CONFIG_RPS=y
-CONFIG_RTC_CLASS=y
-CONFIG_RTC_DRV_AS3722=y
-# CONFIG_RTC_DRV_CMOS is not set
-CONFIG_RTC_DRV_DS1307=y
-CONFIG_RTC_DRV_DS1307_HWMON=y
-CONFIG_RTC_DRV_DS3232=y
-# CONFIG_RTC_DRV_EFI is not set
-CONFIG_RTC_DRV_EM3027=y
-# CONFIG_RTC_DRV_IMXDI is not set
-CONFIG_RTC_DRV_MAX77686=y
-CONFIG_RTC_DRV_MAX8907=y
-# CONFIG_RTC_DRV_MAX8997 is not set
-# CONFIG_RTC_DRV_MAX8998 is not set
-# CONFIG_RTC_DRV_MXC is not set
-CONFIG_RTC_DRV_PALMAS=y
-CONFIG_RTC_DRV_PCF2127=y
-CONFIG_RTC_DRV_PCF85263=y
-# CONFIG_RTC_DRV_RK808 is not set
-# CONFIG_RTC_DRV_S5M is not set
-CONFIG_RTC_DRV_TPS6586X=y
-CONFIG_RTC_DRV_TPS65910=y
-CONFIG_RTC_DRV_TWL4030=y
-CONFIG_RTC_I2C_AND_SPI=y
-# CONFIG_RT_GROUP_SCHED is not set
-CONFIG_RWSEM_SPIN_ON_OWNER=y
-CONFIG_RWSEM_XCHGADD_ALGORITHM=y
-CONFIG_SATA_AHCI=y
-CONFIG_SATA_AHCI_PLATFORM=y
-CONFIG_SATA_PMP=y
-CONFIG_SATA_SIL24=y
-CONFIG_SCHED_DEBUG=y
-# CONFIG_SCHED_INFO is not set
-CONFIG_SCSI=y
-CONFIG_SECCOMP=y
-CONFIG_SECCOMP_FILTER=y
-# CONFIG_SECURITY_DMESG_RESTRICT is not set
-CONFIG_SENSORS_IIO_HWMON=y
-CONFIG_SENSORS_ISL29018=y
-CONFIG_SENSORS_ISL29028=y
-CONFIG_SENSORS_LM90=y
-CONFIG_SENSORS_LM95245=y
-CONFIG_SERIAL_8250_DEPRECATED_OPTIONS=y
-CONFIG_SERIAL_8250_DW=y
-CONFIG_SERIAL_8250_EM=y
-CONFIG_SERIAL_8250_FSL=y
-CONFIG_SERIAL_8250_NR_UARTS=4
-CONFIG_SERIAL_8250_PCI=y
-CONFIG_SERIAL_8250_RUNTIME_UARTS=4
-CONFIG_SERIAL_BCM63XX=y
-CONFIG_SERIAL_BCM63XX_CONSOLE=y
-CONFIG_SERIAL_CONEXANT_DIGICOLOR=y
-CONFIG_SERIAL_CONEXANT_DIGICOLOR_CONSOLE=y
-CONFIG_SERIAL_FSL_LPUART=y
-CONFIG_SERIAL_FSL_LPUART_CONSOLE=y
-CONFIG_SERIAL_IMX=y
-CONFIG_SERIAL_IMX_CONSOLE=y
-CONFIG_SERIAL_MCTRL_GPIO=y
-CONFIG_SERIAL_OF_PLATFORM=y
-CONFIG_SERIAL_ST_ASC=y
-CONFIG_SERIAL_ST_ASC_CONSOLE=y
-CONFIG_SERIAL_XILINX_PS_UART=y
-CONFIG_SERIAL_XILINX_PS_UART_CONSOLE=y
-CONFIG_SG_POOL=y
-CONFIG_SLUB_DEBUG=y
-CONFIG_SMP=y
-CONFIG_SMP_ON_UP=y
-CONFIG_SMSC_PHY=y
-CONFIG_SOCK_DIAG=y
-CONFIG_SOC_BRCMSTB=y
-CONFIG_SOC_BUS=y
-# CONFIG_SOC_IMX50 is not set
-# CONFIG_SOC_IMX51 is not set
-# CONFIG_SOC_IMX53 is not set
-# CONFIG_SOC_IMX6Q is not set
-# CONFIG_SOC_IMX6SL is not set
-# CONFIG_SOC_IMX6SX is not set
-# CONFIG_SOC_IMX6UL is not set
-# CONFIG_SOC_IMX7D is not set
-CONFIG_SOC_LS1021A=y
-# CONFIG_SOC_VF610 is not set
-CONFIG_SPARSE_IRQ=y
-CONFIG_SPI=y
-CONFIG_SPI_BITBANG=y
-CONFIG_SPI_CADENCE=y
-# CONFIG_SPI_FSL_QUADSPI is not set
-# CONFIG_SPI_IMX is not set
-CONFIG_SPI_MASTER=y
-CONFIG_SPI_SPIDEV=y
-CONFIG_SPI_XILINX=y
-CONFIG_SPMI=y
-# CONFIG_SQUASHFS_DECOMP_MULTI_PERCPU is not set
-CONFIG_SQUASHFS_DECOMP_SINGLE=y
-CONFIG_SQUASHFS_FILE_CACHE=y
-# CONFIG_SQUASHFS_FILE_DIRECT is not set
-CONFIG_SQUASHFS_LZO=y
-CONFIG_SQUASHFS_ZLIB=y
-CONFIG_SRAM=y
-CONFIG_SRCU=y
-CONFIG_STAGING_BOARD=y
-CONFIG_STMPE_I2C=y
-# CONFIG_STMPE_SPI is not set
-# CONFIG_STRIP_ASM_SYMS is not set
-CONFIG_SUSPEND=y
-CONFIG_SUSPEND_FREEZER=y
-CONFIG_SWIOTLB=y
-CONFIG_SWPHY=y
-CONFIG_SWP_EMULATE=y
-# CONFIG_SW_SYNC is not set
-CONFIG_SYNC_FILE=y
-# CONFIG_SYN_COOKIES is not set
-CONFIG_SYSFS_SYSCALL=y
-CONFIG_SYS_SUPPORTS_APM_EMULATION=y
-CONFIG_SYS_SUPPORTS_HUGETLBFS=y
-# CONFIG_TCP_CONG_ADVANCED is not set
-CONFIG_THERMAL=y
-CONFIG_THERMAL_DEFAULT_GOV_STEP_WISE=y
-CONFIG_THERMAL_GOV_STEP_WISE=y
-CONFIG_THERMAL_HWMON=y
-CONFIG_THERMAL_OF=y
-# CONFIG_THUMB2_KERNEL is not set
-CONFIG_TICK_CPU_ACCOUNTING=y
-CONFIG_TMPFS_POSIX_ACL=y
-CONFIG_TRACING_EVENTS_GPIO=y
-CONFIG_TREE_RCU=y
-CONFIG_TUN=y
-CONFIG_TWL4030_CORE=y
-CONFIG_TWL4030_POWER=y
-# CONFIG_TWL4030_WATCHDOG is not set
-CONFIG_UBIFS_FS=y
-# CONFIG_UBIFS_FS_ADVANCED_COMPR is not set
-CONFIG_UBIFS_FS_LZO=y
-CONFIG_UBIFS_FS_ZLIB=y
-CONFIG_UCS2_STRING=y
-CONFIG_UEVENT_HELPER_PATH=""
-CONFIG_UNCOMPRESS_INCLUDE="debug/uncompress.h"
-CONFIG_UNIX_DIAG=y
-CONFIG_USB_SUPPORT=y
-CONFIG_USER_NS=y
-CONFIG_USE_OF=y
-CONFIG_UTS_NS=y
-CONFIG_VDSO=y
-CONFIG_VECTORS_BASE=0xffff0000
-CONFIG_VETH=y
-CONFIG_VEXPRESS_CONFIG=y
-CONFIG_VEXPRESS_SYSCFG=y
-CONFIG_VFAT_FS=y
-CONFIG_VFP=y
-CONFIG_VFPv3=y
-CONFIG_VHOST=y
-CONFIG_VHOST_NET=y
-CONFIG_VIRTIO=y
-CONFIG_VIRTIO_BLK=y
-CONFIG_VIRTIO_CONSOLE=y
-CONFIG_VIRTIO_MMIO=y
-# CONFIG_VIRTIO_MMIO_CMDLINE_DEVICES is not set
-CONFIG_VIRTIO_NET=y
-CONFIG_VIRTIO_PCI=y
-CONFIG_VIRTIO_PCI_LEGACY=y
-CONFIG_VIRTUALIZATION=y
-CONFIG_VITESSE_PHY=y
-CONFIG_VM_EVENT_COUNTERS=y
-CONFIG_VT=y
-CONFIG_VT_CONSOLE=y
-CONFIG_VT_CONSOLE_SLEEP=y
-CONFIG_VT_HW_CONSOLE_BINDING=y
-CONFIG_WATCHDOG_CORE=y
-# CONFIG_WQ_POWER_EFFICIENT_DEFAULT is not set
-CONFIG_XFRM_ALGO=y
-CONFIG_XFRM_USER=y
-CONFIG_XILINX_WATCHDOG=y
-CONFIG_XPS=y
-CONFIG_XZ_DEC_ARM=y
-CONFIG_XZ_DEC_ARMTHUMB=y
-CONFIG_XZ_DEC_BCJ=y
-CONFIG_XZ_DEC_IA64=y
-CONFIG_XZ_DEC_POWERPC=y
-CONFIG_XZ_DEC_SPARC=y
-CONFIG_XZ_DEC_X86=y
-CONFIG_ZBOOT_ROM_BSS=0
-CONFIG_ZBOOT_ROM_TEXT=0
-CONFIG_ZLIB_DEFLATE=y
-CONFIG_ZLIB_INFLATE=y
diff --git a/target/linux/layerscape/armv8_32b/config-4.9 b/target/linux/layerscape/armv8_32b/config-4.9
deleted file mode 100644
index 933171aff1..0000000000
--- a/target/linux/layerscape/armv8_32b/config-4.9
+++ /dev/null
@@ -1,1211 +0,0 @@
-CONFIG_ABX500_CORE=y
-CONFIG_AD525X_DPOT=y
-CONFIG_AD525X_DPOT_I2C=y
-# CONFIG_AD525X_DPOT_SPI is not set
-CONFIG_AHCI_IMX=y
-CONFIG_AHCI_QORIQ=y
-CONFIG_AK8975=y
-CONFIG_ALIGNMENT_TRAP=y
-CONFIG_APDS9802ALS=y
-CONFIG_AQUANTIA_PHY=y
-# CONFIG_ARCH_AXXIA is not set
-CONFIG_ARCH_CLOCKSOURCE_DATA=y
-CONFIG_ARCH_DMA_ADDR_T_64BIT=y
-CONFIG_ARCH_HAS_ELF_RANDOMIZE=y
-CONFIG_ARCH_HAS_GCOV_PROFILE_ALL=y
-CONFIG_ARCH_HAS_RESET_CONTROLLER=y
-CONFIG_ARCH_HAS_SG_CHAIN=y
-CONFIG_ARCH_HAS_TICK_BROADCAST=y
-CONFIG_ARCH_HAVE_CUSTOM_GPIO_H=y
-CONFIG_ARCH_HIBERNATION_POSSIBLE=y
-CONFIG_ARCH_MIGHT_HAVE_PC_PARPORT=y
-CONFIG_ARCH_MULTIPLATFORM=y
-# CONFIG_ARCH_MULTI_CPU_AUTO is not set
-CONFIG_ARCH_MULTI_V6_V7=y
-CONFIG_ARCH_MULTI_V7=y
-CONFIG_ARCH_MXC=y
-CONFIG_ARCH_NR_GPIO=0
-CONFIG_ARCH_PHYS_ADDR_T_64BIT=y
-# CONFIG_ARCH_SELECT_MEMORY_MODEL is not set
-# CONFIG_ARCH_SPARSEMEM_DEFAULT is not set
-CONFIG_ARCH_SUPPORTS_ATOMIC_RMW=y
-CONFIG_ARCH_SUPPORTS_BIG_ENDIAN=y
-CONFIG_ARCH_SUPPORTS_UPROBES=y
-CONFIG_ARCH_SUSPEND_POSSIBLE=y
-CONFIG_ARCH_USE_BUILTIN_BSWAP=y
-CONFIG_ARCH_USE_CMPXCHG_LOCKREF=y
-CONFIG_ARCH_VIRT=y
-CONFIG_ARCH_WANT_GENERAL_HUGETLB=y
-CONFIG_ARCH_WANT_IPC_PARSE_VERSION=y
-CONFIG_ARM=y
-CONFIG_ARM_AMBA=y
-CONFIG_ARM_APPENDED_DTB=y
-CONFIG_ARM_ARCH_TIMER=y
-CONFIG_ARM_ARCH_TIMER_EVTSTREAM=y
-CONFIG_ARM_ATAG_DTB_COMPAT=y
-# CONFIG_ARM_ATAG_DTB_COMPAT_CMDLINE_EXTEND is not set
-CONFIG_ARM_ATAG_DTB_COMPAT_CMDLINE_FROM_BOOTLOADER=y
-CONFIG_ARM_CPUIDLE=y
-CONFIG_ARM_CPU_SUSPEND=y
-CONFIG_ARM_CRYPTO=y
-CONFIG_ARM_ERRATA_430973=y
-CONFIG_ARM_ERRATA_643719=y
-CONFIG_ARM_ERRATA_720789=y
-CONFIG_ARM_ERRATA_754322=y
-CONFIG_ARM_ERRATA_754327=y
-CONFIG_ARM_ERRATA_764369=y
-CONFIG_ARM_ERRATA_775420=y
-CONFIG_ARM_ERRATA_798181=y
-CONFIG_ARM_GIC=y
-CONFIG_ARM_GIC_V2M=y
-CONFIG_ARM_GIC_V3=y
-CONFIG_ARM_HAS_SG_CHAIN=y
-CONFIG_ARM_HEAVY_MB=y
-# CONFIG_ARM_HIGHBANK_CPUIDLE is not set
-CONFIG_ARM_IMX6Q_CPUFREQ=y
-CONFIG_ARM_L1_CACHE_SHIFT=6
-CONFIG_ARM_L1_CACHE_SHIFT_6=y
-CONFIG_ARM_LPAE=y
-CONFIG_ARM_PATCH_IDIV=y
-CONFIG_ARM_PATCH_PHYS_VIRT=y
-# CONFIG_ARM_PL172_MPMC is not set
-CONFIG_ARM_PMU=y
-CONFIG_ARM_PSCI=y
-CONFIG_ARM_PSCI_FW=y
-# CONFIG_ARM_SMMU is not set
-CONFIG_ARM_SP805_WATCHDOG=y
-CONFIG_ARM_THUMB=y
-CONFIG_ARM_THUMBEE=y
-CONFIG_ARM_TIMER_SP804=y
-CONFIG_ARM_UNWIND=y
-CONFIG_ARM_VIRT_EXT=y
-CONFIG_ASSOCIATIVE_ARRAY=y
-CONFIG_AT803X_PHY=y
-CONFIG_ATA=y
-CONFIG_ATAGS=y
-CONFIG_ATA_VERBOSE_ERROR=y
-CONFIG_AUTOFS4_FS=y
-CONFIG_AUTO_ZRELADDR=y
-# CONFIG_AXP20X_POWER is not set
-CONFIG_BACKLIGHT_AS3711=y
-CONFIG_BACKLIGHT_CLASS_DEVICE=y
-CONFIG_BACKLIGHT_GENERIC=y
-CONFIG_BACKLIGHT_LCD_SUPPORT=y
-CONFIG_BACKLIGHT_PWM=y
-# CONFIG_BACKLIGHT_TPS65217 is not set
-CONFIG_BATTERY_ACT8945A=y
-CONFIG_BATTERY_SBS=y
-CONFIG_BCMA=y
-CONFIG_BCMA_BLOCKIO=y
-# CONFIG_BCMA_DEBUG is not set
-CONFIG_BCMA_DRIVER_GMAC_CMN=y
-CONFIG_BCMA_DRIVER_GPIO=y
-CONFIG_BCMA_DRIVER_PCI=y
-CONFIG_BCMA_HOST_PCI=y
-CONFIG_BCMA_HOST_PCI_POSSIBLE=y
-CONFIG_BCMA_HOST_SOC=y
-CONFIG_BCMA_SFLASH=y
-CONFIG_BCM_NET_PHYLIB=y
-# CONFIG_BLK_CGROUP is not set
-CONFIG_BLK_CMDLINE_PARSER=y
-CONFIG_BLK_DEV_BSG=y
-CONFIG_BLK_DEV_LOOP=y
-CONFIG_BLK_DEV_NVME=y
-# CONFIG_BLK_DEV_NVME_SCSI is not set
-CONFIG_BLK_DEV_RAM=y
-CONFIG_BLK_DEV_RAM_COUNT=16
-CONFIG_BLK_DEV_RAM_SIZE=262144
-CONFIG_BLK_DEV_SD=y
-CONFIG_BLK_DEV_SR=y
-# CONFIG_BLK_DEV_SR_VENDOR is not set
-CONFIG_BLK_MQ_PCI=y
-# CONFIG_BOOTPARAM_HUNG_TASK_PANIC is not set
-CONFIG_BOOTPARAM_HUNG_TASK_PANIC_VALUE=0
-# CONFIG_BOOTPARAM_SOFTLOCKUP_PANIC is not set
-CONFIG_BOOTPARAM_SOFTLOCKUP_PANIC_VALUE=0
-CONFIG_BOUNCE=y
-# CONFIG_BPF_SYSCALL is not set
-CONFIG_BRCMSTB_GISB_ARB=y
-CONFIG_BROADCOM_PHY=y
-CONFIG_BUILD_BIN2C=y
-CONFIG_CACHE_L2X0=y
-# CONFIG_CACHE_L2X0_PMU is not set
-CONFIG_CAN=y
-CONFIG_CAN_BCM=y
-CONFIG_CAN_CALC_BITTIMING=y
-# CONFIG_CAN_CC770 is not set
-# CONFIG_CAN_C_CAN is not set
-CONFIG_CAN_DEV=y
-CONFIG_CAN_FLEXCAN=y
-# CONFIG_CAN_GRCAN is not set
-CONFIG_CAN_GW=y
-# CONFIG_CAN_LEDS is not set
-CONFIG_CAN_MCP251X=y
-CONFIG_CAN_RAW=y
-# CONFIG_CAN_SJA1000 is not set
-# CONFIG_CAN_SOFTING is not set
-# CONFIG_CAN_TI_HECC is not set
-CONFIG_CC_OPTIMIZE_FOR_PERFORMANCE=y
-# CONFIG_CC_OPTIMIZE_FOR_SIZE is not set
-# CONFIG_CFS_BANDWIDTH is not set
-CONFIG_CGROUPS=y
-CONFIG_CGROUP_CPUACCT=y
-CONFIG_CGROUP_DEVICE=y
-CONFIG_CGROUP_FREEZER=y
-# CONFIG_CGROUP_PERF is not set
-CONFIG_CGROUP_PIDS=y
-CONFIG_CGROUP_SCHED=y
-# CONFIG_CHARGER_MAX14577 is not set
-CONFIG_CHARGER_TPS65090=y
-# CONFIG_CHARGER_TPS65217 is not set
-CONFIG_CHECKPOINT_RESTORE=y
-CONFIG_CHROME_PLATFORMS=y
-CONFIG_CHR_DEV_SG=y
-CONFIG_CLKDEV_LOOKUP=y
-CONFIG_CLKSRC_IMX_GPT=y
-CONFIG_CLKSRC_MMIO=y
-CONFIG_CLKSRC_OF=y
-CONFIG_CLKSRC_PROBE=y
-CONFIG_CLKSRC_VERSATILE=y
-CONFIG_CLK_QORIQ=y
-CONFIG_CLONE_BACKWARDS=y
-CONFIG_CMA=y
-CONFIG_CMA_ALIGNMENT=8
-CONFIG_CMA_AREAS=7
-# CONFIG_CMA_DEBUG is not set
-# CONFIG_CMA_DEBUGFS is not set
-CONFIG_CMA_SIZE_MBYTES=64
-# CONFIG_CMA_SIZE_SEL_MAX is not set
-CONFIG_CMA_SIZE_SEL_MBYTES=y
-# CONFIG_CMA_SIZE_SEL_MIN is not set
-# CONFIG_CMA_SIZE_SEL_PERCENTAGE is not set
-CONFIG_CMDLINE_PARTITION=y
-CONFIG_COMMON_CLK=y
-CONFIG_COMMON_CLK_MAX77686=y
-# CONFIG_COMMON_CLK_PALMAS is not set
-# CONFIG_COMMON_CLK_RK808 is not set
-# CONFIG_COMMON_CLK_S2MPS11 is not set
-CONFIG_COMPAT_BRK=y
-CONFIG_CONFIGFS_FS=y
-CONFIG_CONSOLE_TRANSLATIONS=y
-CONFIG_COREDUMP=y
-CONFIG_CORE_DUMP_DEFAULT_ELF_HEADERS=y
-# CONFIG_CORTINA_PHY is not set
-CONFIG_CPUFREQ_DT=y
-CONFIG_CPUFREQ_DT_PLATDEV=y
-CONFIG_CPUSETS=y
-CONFIG_CPU_32v6K=y
-CONFIG_CPU_32v7=y
-CONFIG_CPU_ABRT_EV7=y
-# CONFIG_CPU_BIG_ENDIAN is not set
-# CONFIG_CPU_BPREDICT_DISABLE is not set
-CONFIG_CPU_CACHE_V7=y
-CONFIG_CPU_CACHE_VIPT=y
-CONFIG_CPU_COPY_V6=y
-CONFIG_CPU_CP15=y
-CONFIG_CPU_CP15_MMU=y
-CONFIG_CPU_FREQ=y
-CONFIG_CPU_FREQ_DEFAULT_GOV_ONDEMAND=y
-# CONFIG_CPU_FREQ_DEFAULT_GOV_PERFORMANCE is not set
-CONFIG_CPU_FREQ_GOV_ATTR_SET=y
-CONFIG_CPU_FREQ_GOV_COMMON=y
-# CONFIG_CPU_FREQ_GOV_CONSERVATIVE is not set
-CONFIG_CPU_FREQ_GOV_ONDEMAND=y
-CONFIG_CPU_FREQ_GOV_PERFORMANCE=y
-# CONFIG_CPU_FREQ_GOV_POWERSAVE is not set
-CONFIG_CPU_FREQ_GOV_SCHEDUTIL=y
-# CONFIG_CPU_FREQ_GOV_USERSPACE is not set
-CONFIG_CPU_FREQ_STAT=y
-CONFIG_CPU_HAS_ASID=y
-# CONFIG_CPU_HOTPLUG_STATE_CONTROL is not set
-# CONFIG_CPU_ICACHE_DISABLE is not set
-CONFIG_CPU_IDLE=y
-# CONFIG_CPU_IDLE_GOV_LADDER is not set
-CONFIG_CPU_IDLE_GOV_MENU=y
-CONFIG_CPU_PABRT_V7=y
-CONFIG_CPU_PM=y
-CONFIG_CPU_RMAP=y
-CONFIG_CPU_THERMAL=y
-CONFIG_CPU_TLB_V7=y
-CONFIG_CPU_V7=y
-CONFIG_CRC16=y
-# CONFIG_CRC32_SARWATE is not set
-CONFIG_CRC32_SLICEBY8=y
-CONFIG_CROSS_MEMORY_ATTACH=y
-CONFIG_CRYPTO_AEAD=y
-CONFIG_CRYPTO_AEAD2=y
-# CONFIG_CRYPTO_AES_ARM_CE is not set
-CONFIG_CRYPTO_AUTHENC=y
-CONFIG_CRYPTO_CBC=y
-CONFIG_CRYPTO_CRC32C=y
-CONFIG_CRYPTO_DEFLATE=y
-CONFIG_CRYPTO_DES=y
-# CONFIG_CRYPTO_DEV_FSL_CAAM_AHASH_API_DESC is not set
-# CONFIG_CRYPTO_DEV_FSL_CAAM_CRYPTO_API_DESC is not set
-# CONFIG_CRYPTO_DEV_MXC_SCC is not set
-CONFIG_CRYPTO_DRBG=y
-CONFIG_CRYPTO_DRBG_HMAC=y
-CONFIG_CRYPTO_DRBG_MENU=y
-CONFIG_CRYPTO_ECHAINIV=y
-# CONFIG_CRYPTO_GHASH_ARM_CE is not set
-CONFIG_CRYPTO_HASH=y
-CONFIG_CRYPTO_HASH2=y
-CONFIG_CRYPTO_HMAC=y
-CONFIG_CRYPTO_HW=y
-CONFIG_CRYPTO_JITTERENTROPY=y
-CONFIG_CRYPTO_LZO=y
-CONFIG_CRYPTO_MANAGER=y
-CONFIG_CRYPTO_MANAGER2=y
-CONFIG_CRYPTO_MD5=y
-CONFIG_CRYPTO_NULL=y
-CONFIG_CRYPTO_NULL2=y
-# CONFIG_CRYPTO_PCRYPT is not set
-CONFIG_CRYPTO_RNG=y
-CONFIG_CRYPTO_RNG2=y
-CONFIG_CRYPTO_RNG_DEFAULT=y
-CONFIG_CRYPTO_SHA1=y
-# CONFIG_CRYPTO_SHA1_ARM_CE is not set
-# CONFIG_CRYPTO_SHA1_ARM_NEON is not set
-CONFIG_CRYPTO_SHA256=y
-# CONFIG_CRYPTO_SHA256_ARM is not set
-# CONFIG_CRYPTO_SHA2_ARM_CE is not set
-# CONFIG_CRYPTO_SHA512_ARM is not set
-# CONFIG_CRYPTO_TLS is not set
-CONFIG_CRYPTO_WORKQUEUE=y
-CONFIG_DCACHE_WORD_ACCESS=y
-CONFIG_DEBUG_ALIGN_RODATA=y
-CONFIG_DEBUG_BUGVERBOSE=y
-CONFIG_DEBUG_IMX_UART_PORT=1
-CONFIG_DEBUG_LL_INCLUDE="mach/debug-macro.S"
-CONFIG_DEBUG_RODATA=y
-# CONFIG_DEBUG_UART_8250 is not set
-# CONFIG_DEBUG_USER is not set
-CONFIG_DECOMPRESS_BZIP2=y
-CONFIG_DECOMPRESS_GZIP=y
-CONFIG_DECOMPRESS_LZMA=y
-CONFIG_DECOMPRESS_LZO=y
-CONFIG_DECOMPRESS_XZ=y
-CONFIG_DEFAULT_CFQ=y
-# CONFIG_DEFAULT_DEADLINE is not set
-CONFIG_DEFAULT_HUNG_TASK_TIMEOUT=120
-CONFIG_DEFAULT_IOSCHED="cfq"
-CONFIG_DETECT_HUNG_TASK=y
-# CONFIG_DEVFREQ_GOV_PASSIVE is not set
-# CONFIG_DEVFREQ_GOV_PERFORMANCE is not set
-# CONFIG_DEVFREQ_GOV_POWERSAVE is not set
-# CONFIG_DEVFREQ_GOV_SIMPLE_ONDEMAND is not set
-# CONFIG_DEVFREQ_GOV_USERSPACE is not set
-# CONFIG_DEVFREQ_THERMAL is not set
-CONFIG_DEVTMPFS=y
-CONFIG_DEVTMPFS_MOUNT=y
-CONFIG_DMADEVICES=y
-CONFIG_DMA_CMA=y
-CONFIG_DMA_ENGINE=y
-CONFIG_DMA_OF=y
-CONFIG_DMA_SHARED_BUFFER=y
-CONFIG_DMA_VIRTUAL_CHANNELS=y
-CONFIG_DNOTIFY=y
-CONFIG_DNS_RESOLVER=y
-CONFIG_DRM=y
-CONFIG_DRM_BRIDGE=y
-CONFIG_DRM_FBDEV_EMULATION=y
-CONFIG_DRM_FSL_DCU=y
-CONFIG_DRM_GEM_CMA_HELPER=y
-CONFIG_DRM_KMS_CMA_HELPER=y
-CONFIG_DRM_KMS_FB_HELPER=y
-CONFIG_DRM_KMS_HELPER=y
-CONFIG_DRM_PANEL=y
-CONFIG_DRM_PANEL_SIMPLE=y
-CONFIG_DRM_SII902X=y
-# CONFIG_DRM_VIRTIO_GPU is not set
-CONFIG_DST_CACHE=y
-CONFIG_DTC=y
-CONFIG_DT_IDLE_STATES=y
-CONFIG_DUMMY_CONSOLE=y
-CONFIG_DWMAC_GENERIC=y
-CONFIG_DW_DMAC=y
-CONFIG_DW_DMAC_CORE=y
-CONFIG_DW_WATCHDOG=y
-CONFIG_E1000E=y
-CONFIG_EDAC=y
-CONFIG_EDAC_ATOMIC_SCRUB=y
-# CONFIG_EDAC_DEBUG is not set
-CONFIG_EDAC_LEGACY_SYSFS=y
-# CONFIG_EDAC_MM_EDAC is not set
-CONFIG_EDAC_SUPPORT=y
-CONFIG_EEPROM_93CX6=y
-CONFIG_EEPROM_AT24=y
-CONFIG_EFI=y
-# CONFIG_EFIVAR_FS is not set
-CONFIG_EFI_ARMSTUB=y
-# CONFIG_EFI_CAPSULE_LOADER is not set
-CONFIG_EFI_ESRT=y
-CONFIG_EFI_PARAMS_FROM_FDT=y
-CONFIG_EFI_RUNTIME_WRAPPERS=y
-CONFIG_EFI_STUB=y
-# CONFIG_EFI_TEST is not set
-# CONFIG_EFI_VARS is not set
-CONFIG_ELF_CORE=y
-# CONFIG_ENABLE_DEFAULT_TRACERS is not set
-CONFIG_ENABLE_MUST_CHECK=y
-CONFIG_EXPORTFS=y
-CONFIG_EXT4_FS=y
-CONFIG_EXT4_FS_POSIX_ACL=y
-CONFIG_EXT4_FS_SECURITY=y
-CONFIG_EXTCON=y
-# CONFIG_EXTCON_MAX14577 is not set
-# CONFIG_EXTCON_MAX8997 is not set
-# CONFIG_EXTCON_PALMAS is not set
-CONFIG_FAIR_GROUP_SCHED=y
-CONFIG_FAT_FS=y
-CONFIG_FB=y
-CONFIG_FB_ARMCLCD=y
-CONFIG_FB_CFB_COPYAREA=y
-CONFIG_FB_CFB_FILLRECT=y
-CONFIG_FB_CFB_IMAGEBLIT=y
-CONFIG_FB_CMDLINE=y
-CONFIG_FB_DEFERRED_IO=y
-CONFIG_FB_EFI=y
-CONFIG_FB_MODE_HELPERS=y
-CONFIG_FB_MX3=y
-CONFIG_FB_SIMPLE=y
-CONFIG_FB_SYS_COPYAREA=y
-CONFIG_FB_SYS_FILLRECT=y
-CONFIG_FB_SYS_FOPS=y
-CONFIG_FB_SYS_IMAGEBLIT=y
-CONFIG_FEC=y
-CONFIG_FHANDLE=y
-CONFIG_FIRMWARE_IN_KERNEL=y
-CONFIG_FIXED_PHY=y
-CONFIG_FIX_EARLYCON_MEM=y
-CONFIG_FMAN_ARM=y
-# CONFIG_FMAN_MIB_CNT_OVF_IRQ_EN is not set
-# CONFIG_FMAN_P1023 is not set
-# CONFIG_FMAN_P3040_P4080_P5020 is not set
-# CONFIG_FMAN_PFC is not set
-# CONFIG_FMAN_V3H is not set
-# CONFIG_FMAN_V3L is not set
-# CONFIG_FONTS is not set
-CONFIG_FONT_8x16=y
-CONFIG_FONT_8x8=y
-CONFIG_FONT_SUPPORT=y
-CONFIG_FORCE_MAX_ZONEORDER=12
-CONFIG_FRAMEBUFFER_CONSOLE=y
-CONFIG_FRAMEBUFFER_CONSOLE_DETECT_PRIMARY=y
-CONFIG_FRAMEBUFFER_CONSOLE_ROTATION=y
-CONFIG_FREEZER=y
-CONFIG_FSL_BMAN_CONFIG=y
-CONFIG_FSL_BMAN_DEBUGFS=y
-# CONFIG_FSL_BMAN_TEST is not set
-# CONFIG_FSL_DPAA_1588 is not set
-CONFIG_FSL_DPAA_ADVANCED_DRIVERS=y
-# CONFIG_FSL_DPAA_CEETM is not set
-CONFIG_FSL_DPAA_CS_THRESHOLD_10G=0x10000000
-CONFIG_FSL_DPAA_CS_THRESHOLD_1G=0x06000000
-# CONFIG_FSL_DPAA_DBG_LOOP is not set
-# CONFIG_FSL_DPAA_ETH_DEBUG is not set
-CONFIG_FSL_DPAA_ETH_DEBUGFS=y
-# CONFIG_FSL_DPAA_ETH_JUMBO_FRAME is not set
-CONFIG_FSL_DPAA_ETH_MAX_BUF_COUNT=128
-CONFIG_FSL_DPAA_ETH_REFILL_THRESHOLD=80
-CONFIG_FSL_DPAA_ETH_USE_NDO_SELECT_QUEUE=y
-# CONFIG_FSL_DPAA_HOOKS is not set
-CONFIG_FSL_DPAA_INGRESS_CS_THRESHOLD=0x10000000
-CONFIG_FSL_DPAA_OFFLINE_PORTS=y
-# CONFIG_FSL_DPAA_TS is not set
-CONFIG_FSL_DPA_CAN_WAIT=y
-CONFIG_FSL_DPA_CAN_WAIT_SYNC=y
-# CONFIG_FSL_DPA_CHECKING is not set
-CONFIG_FSL_DPA_PIRQ_FAST=y
-CONFIG_FSL_DPA_PIRQ_SLOW=y
-CONFIG_FSL_DPA_PORTAL_SHARE=y
-CONFIG_FSL_EDMA=y
-CONFIG_FSL_FM_MAX_FRAME_SIZE=1522
-CONFIG_FSL_FM_RX_EXTRA_HEADROOM=64
-CONFIG_FSL_GUTS=y
-CONFIG_FSL_IFC=y
-CONFIG_FSL_PPFE=y
-CONFIG_FSL_PPFE_UTIL_DISABLED=y
-CONFIG_FSL_PQ_MDIO=y
-# CONFIG_FSL_QDMA is not set
-CONFIG_FSL_QMAN_CI_SCHED_CFG_BMAN_W=2
-CONFIG_FSL_QMAN_CI_SCHED_CFG_RW_W=2
-CONFIG_FSL_QMAN_CI_SCHED_CFG_SRCCIV=4
-CONFIG_FSL_QMAN_CI_SCHED_CFG_SRQ_W=3
-CONFIG_FSL_QMAN_CONFIG=y
-CONFIG_FSL_QMAN_DEBUGFS=y
-CONFIG_FSL_QMAN_FQD_SZ=10
-# CONFIG_FSL_QMAN_FQ_LOOKUP is not set
-CONFIG_FSL_QMAN_INIT_TIMEOUT=10
-CONFIG_FSL_QMAN_PFDR_SZ=13
-CONFIG_FSL_QMAN_PIRQ_DQRR_ITHRESH=12
-CONFIG_FSL_QMAN_PIRQ_IPERIOD=100
-CONFIG_FSL_QMAN_PIRQ_MR_ITHRESH=4
-CONFIG_FSL_QMAN_POLL_LIMIT=32
-# CONFIG_FSL_QMAN_TEST is not set
-CONFIG_FSL_SDK_BMAN=y
-CONFIG_FSL_SDK_DPA=y
-CONFIG_FSL_SDK_DPAA_ETH=y
-CONFIG_FSL_SDK_FMAN=y
-# CONFIG_FSL_SDK_FMAN_TEST is not set
-CONFIG_FSL_SDK_QMAN=y
-CONFIG_FSL_USDPAA=y
-CONFIG_FSL_XGMAC_MDIO=y
-CONFIG_FS_MBCACHE=y
-CONFIG_FS_POSIX_ACL=y
-CONFIG_FTRACE=y
-# CONFIG_FTRACE_SYSCALLS is not set
-CONFIG_FUSE_FS=y
-# CONFIG_FW_LOADER_USER_HELPER_FALLBACK is not set
-CONFIG_GENERIC_ALLOCATOR=y
-CONFIG_GENERIC_BUG=y
-CONFIG_GENERIC_CLOCKEVENTS=y
-CONFIG_GENERIC_CLOCKEVENTS_BROADCAST=y
-CONFIG_GENERIC_EARLY_IOREMAP=y
-CONFIG_GENERIC_IDLE_POLL_SETUP=y
-CONFIG_GENERIC_IO=y
-CONFIG_GENERIC_IRQ_CHIP=y
-# CONFIG_GENERIC_IRQ_DEBUGFS is not set
-CONFIG_GENERIC_IRQ_SHOW=y
-CONFIG_GENERIC_IRQ_SHOW_LEVEL=y
-CONFIG_GENERIC_MSI_IRQ=y
-CONFIG_GENERIC_MSI_IRQ_DOMAIN=y
-CONFIG_GENERIC_PCI_IOMAP=y
-CONFIG_GENERIC_PHY=y
-CONFIG_GENERIC_PINCONF=y
-CONFIG_GENERIC_SCHED_CLOCK=y
-CONFIG_GENERIC_SMP_IDLE_THREAD=y
-CONFIG_GENERIC_STRNCPY_FROM_USER=y
-CONFIG_GENERIC_STRNLEN_USER=y
-CONFIG_GENERIC_TIME_VSYSCALL=y
-CONFIG_GIANFAR=y
-CONFIG_GLOB=y
-CONFIG_GPIOLIB=y
-CONFIG_GPIOLIB_IRQCHIP=y
-# CONFIG_GPIO_AXP209 is not set
-CONFIG_GPIO_DWAPB=y
-CONFIG_GPIO_EM=y
-CONFIG_GPIO_GENERIC=y
-CONFIG_GPIO_GENERIC_PLATFORM=y
-CONFIG_GPIO_MXC=y
-CONFIG_GPIO_PALMAS=y
-CONFIG_GPIO_PCA953X=y
-CONFIG_GPIO_PCA953X_IRQ=y
-CONFIG_GPIO_PCF857X=y
-CONFIG_GPIO_PL061=y
-# CONFIG_GPIO_STMPE is not set
-CONFIG_GPIO_SYSCON=y
-# CONFIG_GPIO_TPS65218 is not set
-CONFIG_GPIO_TPS6586X=y
-CONFIG_GPIO_TPS65910=y
-CONFIG_GPIO_TWL4030=y
-CONFIG_GPIO_XILINX=y
-CONFIG_HANDLE_DOMAIN_IRQ=y
-CONFIG_HARDEN_BRANCH_PREDICTOR=y
-CONFIG_HARDIRQS_SW_RESEND=y
-CONFIG_HAS_DMA=y
-CONFIG_HAS_IOMEM=y
-CONFIG_HAS_IOPORT_MAP=y
-# CONFIG_HAVE_64BIT_ALIGNED_ACCESS is not set
-CONFIG_HAVE_ARCH_AUDITSYSCALL=y
-CONFIG_HAVE_ARCH_BITREVERSE=y
-CONFIG_HAVE_ARCH_JUMP_LABEL=y
-CONFIG_HAVE_ARCH_KGDB=y
-CONFIG_HAVE_ARCH_PFN_VALID=y
-CONFIG_HAVE_ARCH_SECCOMP_FILTER=y
-CONFIG_HAVE_ARCH_TRACEHOOK=y
-CONFIG_HAVE_ARCH_TRANSPARENT_HUGEPAGE=y
-CONFIG_HAVE_ARM_ARCH_TIMER=y
-CONFIG_HAVE_ARM_SMCCC=y
-# CONFIG_HAVE_BOOTMEM_INFO_NODE is not set
-CONFIG_HAVE_CBPF_JIT=y
-CONFIG_HAVE_CC_STACKPROTECTOR=y
-CONFIG_HAVE_CLK=y
-CONFIG_HAVE_CLK_PREPARE=y
-CONFIG_HAVE_CONTEXT_TRACKING=y
-CONFIG_HAVE_C_RECORDMCOUNT=y
-CONFIG_HAVE_DEBUG_KMEMLEAK=y
-CONFIG_HAVE_DMA_API_DEBUG=y
-CONFIG_HAVE_DMA_CONTIGUOUS=y
-CONFIG_HAVE_DYNAMIC_FTRACE=y
-CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS=y
-CONFIG_HAVE_FTRACE_MCOUNT_RECORD=y
-CONFIG_HAVE_FUNCTION_GRAPH_TRACER=y
-CONFIG_HAVE_FUNCTION_TRACER=y
-CONFIG_HAVE_GENERIC_DMA_COHERENT=y
-CONFIG_HAVE_GENERIC_RCU_GUP=y
-CONFIG_HAVE_HW_BREAKPOINT=y
-CONFIG_HAVE_IDE=y
-CONFIG_HAVE_IMX_SRC=y
-CONFIG_HAVE_IRQ_TIME_ACCOUNTING=y
-CONFIG_HAVE_KVM_ARCH_TLB_FLUSH_ALL=y
-CONFIG_HAVE_KVM_CPU_RELAX_INTERCEPT=y
-CONFIG_HAVE_KVM_EVENTFD=y
-CONFIG_HAVE_KVM_IRQCHIP=y
-CONFIG_HAVE_KVM_IRQFD=y
-CONFIG_HAVE_KVM_IRQ_ROUTING=y
-CONFIG_HAVE_MEMBLOCK=y
-CONFIG_HAVE_MOD_ARCH_SPECIFIC=y
-CONFIG_HAVE_NET_DSA=y
-CONFIG_HAVE_OPROFILE=y
-CONFIG_HAVE_OPTPROBES=y
-CONFIG_HAVE_PERF_EVENTS=y
-CONFIG_HAVE_PERF_REGS=y
-CONFIG_HAVE_PERF_USER_STACK_DUMP=y
-CONFIG_HAVE_PROC_CPU=y
-CONFIG_HAVE_RCU_TABLE_FREE=y
-CONFIG_HAVE_REGS_AND_STACK_ACCESS_API=y
-CONFIG_HAVE_SMP=y
-CONFIG_HAVE_SYSCALL_TRACEPOINTS=y
-CONFIG_HAVE_UID16=y
-CONFIG_HAVE_VIRT_CPU_ACCOUNTING_GEN=y
-CONFIG_HDMI=y
-CONFIG_HID=y
-CONFIG_HID_GENERIC=y
-CONFIG_HIGHMEM=y
-CONFIG_HIGHPTE=y
-CONFIG_HIX5HD2_GMAC=y
-CONFIG_HOTPLUG_CPU=y
-# CONFIG_HUGETLBFS is not set
-CONFIG_HVC_DRIVER=y
-CONFIG_HWMON=y
-CONFIG_HW_CONSOLE=y
-CONFIG_HW_RANDOM=y
-CONFIG_HZ_FIXED=0
-CONFIG_I2C=y
-CONFIG_I2C_ALGOBIT=y
-CONFIG_I2C_BOARDINFO=y
-CONFIG_I2C_CHARDEV=y
-CONFIG_I2C_COMPAT=y
-CONFIG_I2C_DEMUX_PINCTRL=y
-CONFIG_I2C_DESIGNWARE_CORE=y
-CONFIG_I2C_DESIGNWARE_PLATFORM=y
-CONFIG_I2C_HELPER_AUTO=y
-CONFIG_I2C_IMX=y
-CONFIG_I2C_MUX=y
-CONFIG_I2C_MUX_PCA954x=y
-CONFIG_I2C_MUX_PINCTRL=y
-CONFIG_I2C_NOMADIK=y
-CONFIG_I2C_RK3X=y
-CONFIG_I2C_SLAVE=y
-CONFIG_I2C_SLAVE_EEPROM=y
-CONFIG_I2C_XILINX=y
-CONFIG_ICPLUS_PHY=y
-CONFIG_ICS932S401=y
-CONFIG_IGB=y
-CONFIG_IGB_HWMON=y
-CONFIG_IIO=y
-CONFIG_IIO_BUFFER=y
-CONFIG_IIO_CONFIGFS=y
-CONFIG_IIO_HRTIMER_TRIGGER=y
-CONFIG_IIO_KFIFO_BUF=y
-CONFIG_IIO_SW_TRIGGER=y
-# CONFIG_IIO_TIGHTLOOP_TRIGGER is not set
-CONFIG_IIO_TRIGGER=y
-CONFIG_IIO_TRIGGERED_BUFFER=y
-CONFIG_IKCONFIG=y
-CONFIG_IKCONFIG_PROC=y
-CONFIG_IMX2_WDT=y
-CONFIG_IMX_DMA=y
-CONFIG_IMX_SDMA=y
-# CONFIG_IMX_WEIM is not set
-CONFIG_INET6_XFRM_MODE_BEET=y
-CONFIG_INET6_XFRM_MODE_TRANSPORT=y
-CONFIG_INET6_XFRM_MODE_TUNNEL=y
-CONFIG_INET_DIAG=y
-# CONFIG_INET_DIAG_DESTROY is not set
-CONFIG_INET_ESP=y
-CONFIG_INET_TCP_DIAG=y
-CONFIG_INET_TUNNEL=y
-CONFIG_INET_XFRM_MODE_BEET=y
-CONFIG_INET_XFRM_MODE_TRANSPORT=y
-CONFIG_INET_XFRM_MODE_TUNNEL=y
-CONFIG_INITRAMFS_SOURCE=""
-CONFIG_INPUT=y
-# CONFIG_INPUT_AXP20X_PEK is not set
-CONFIG_INPUT_EVDEV=y
-CONFIG_INPUT_JOYDEV=y
-CONFIG_INPUT_KEYBOARD=y
-CONFIG_INPUT_LEDS=y
-CONFIG_INPUT_MATRIXKMAP=y
-# CONFIG_INPUT_MAX8997_HAPTIC is not set
-CONFIG_INPUT_MOUSE=y
-CONFIG_INPUT_MOUSEDEV=y
-CONFIG_INPUT_MOUSEDEV_PSAUX=y
-CONFIG_INPUT_MOUSEDEV_SCREEN_X=1024
-CONFIG_INPUT_MOUSEDEV_SCREEN_Y=768
-CONFIG_INPUT_TOUCHSCREEN=y
-CONFIG_IOMMU_HELPER=y
-# CONFIG_IOMMU_IO_PGTABLE_ARMV7S is not set
-# CONFIG_IOMMU_IO_PGTABLE_LPAE is not set
-CONFIG_IOMMU_SUPPORT=y
-CONFIG_IOSCHED_CFQ=y
-CONFIG_IPC_NS=y
-CONFIG_IPV6=y
-CONFIG_IPV6_MULTIPLE_TABLES=y
-CONFIG_IPV6_OPTIMISTIC_DAD=y
-CONFIG_IPV6_ROUTER_PREF=y
-CONFIG_IPV6_SIT=y
-# CONFIG_IPV6_SUBTREES is not set
-# CONFIG_IP_ADVANCED_ROUTER is not set
-CONFIG_IP_PNP=y
-CONFIG_IP_PNP_BOOTP=y
-CONFIG_IP_PNP_DHCP=y
-CONFIG_IP_PNP_RARP=y
-CONFIG_IRQCHIP=y
-CONFIG_IRQ_DOMAIN=y
-CONFIG_IRQ_DOMAIN_DEBUG=y
-CONFIG_IRQ_DOMAIN_HIERARCHY=y
-CONFIG_IRQ_FORCED_THREADING=y
-CONFIG_IRQ_WORK=y
-# CONFIG_ISDN is not set
-CONFIG_ISL29003=y
-CONFIG_JBD2=y
-CONFIG_KALLSYMS=y
-CONFIG_KERNEL_GZIP=y
-# CONFIG_KERNEL_XZ is not set
-CONFIG_KEXEC=y
-CONFIG_KEXEC_CORE=y
-CONFIG_KEYBOARD_ATKBD=y
-CONFIG_KEYBOARD_BCM=y
-CONFIG_KEYBOARD_GPIO=y
-# CONFIG_KEYBOARD_IMX is not set
-# CONFIG_KEYBOARD_STMPE is not set
-CONFIG_KEYS=y
-CONFIG_KS8851=y
-CONFIG_KVM=y
-CONFIG_KVM_ARM_HOST=y
-CONFIG_KVM_GENERIC_DIRTYLOG_READ_PROTECT=y
-CONFIG_KVM_MMIO=y
-CONFIG_KVM_VFIO=y
-# CONFIG_LCD_CLASS_DEVICE is not set
-CONFIG_LEDS_GPIO=y
-# CONFIG_LEDS_MAX8997 is not set
-CONFIG_LEDS_PWM=y
-CONFIG_LEDS_TRIGGER_BACKLIGHT=y
-CONFIG_LEDS_TRIGGER_CAMERA=y
-CONFIG_LEDS_TRIGGER_CPU=y
-CONFIG_LEDS_TRIGGER_GPIO=y
-CONFIG_LEDS_TRIGGER_HEARTBEAT=y
-CONFIG_LEDS_TRIGGER_ONESHOT=y
-CONFIG_LEDS_TRIGGER_TRANSIENT=y
-CONFIG_LEGACY_PTYS=y
-CONFIG_LEGACY_PTY_COUNT=256
-CONFIG_LIBFDT=y
-CONFIG_LOCALVERSION_AUTO=y
-CONFIG_LOCKUP_DETECTOR=y
-CONFIG_LOCK_SPIN_ON_OWNER=y
-CONFIG_LOGO=y
-CONFIG_LOGO_LINUX_CLUT224=y
-CONFIG_LOGO_LINUX_MONO=y
-CONFIG_LOGO_LINUX_VGA16=y
-CONFIG_LS_SCFG_MSI=y
-# CONFIG_LS_SOC_DRIVERS is not set
-CONFIG_LZO_COMPRESS=y
-CONFIG_LZO_DECOMPRESS=y
-CONFIG_MACB=y
-CONFIG_MACVLAN=y
-CONFIG_MACVTAP=y
-CONFIG_MAGIC_SYSRQ=y
-CONFIG_MAILBOX=y
-# CONFIG_MAILBOX_TEST is not set
-CONFIG_MANDATORY_FILE_LOCKING=y
-CONFIG_MARVELL_PHY=y
-CONFIG_MCPM=y
-CONFIG_MDIO_BITBANG=y
-CONFIG_MDIO_BOARDINFO=y
-CONFIG_MDIO_BUS_MUX=y
-CONFIG_MDIO_BUS_MUX_MMIOREG=y
-# CONFIG_MDIO_FSL_BACKPLANE is not set
-# CONFIG_MDIO_GPIO is not set
-CONFIG_MEMCG=y
-# CONFIG_MEMCG_SWAP is not set
-CONFIG_MEMORY=y
-CONFIG_MEMORY_ISOLATION=y
-CONFIG_MFD_ACT8945A=y
-CONFIG_MFD_AS3711=y
-CONFIG_MFD_AS3722=y
-CONFIG_MFD_ATMEL_FLEXCOM=y
-CONFIG_MFD_AXP20X=y
-CONFIG_MFD_AXP20X_I2C=y
-CONFIG_MFD_BCM590XX=y
-CONFIG_MFD_CORE=y
-CONFIG_MFD_MAX14577=y
-CONFIG_MFD_MAX77686=y
-CONFIG_MFD_MAX8907=y
-CONFIG_MFD_MAX8997=y
-CONFIG_MFD_MAX8998=y
-CONFIG_MFD_PALMAS=y
-CONFIG_MFD_RK808=y
-CONFIG_MFD_SEC_CORE=y
-CONFIG_MFD_STMPE=y
-CONFIG_MFD_SYSCON=y
-CONFIG_MFD_TPS65090=y
-CONFIG_MFD_TPS65217=y
-CONFIG_MFD_TPS65218=y
-CONFIG_MFD_TPS6586X=y
-CONFIG_MFD_TPS65910=y
-# CONFIG_MFD_TWL4030_AUDIO is not set
-CONFIG_MFD_VEXPRESS_SYSREG=y
-CONFIG_MICREL_PHY=y
-CONFIG_MIGHT_HAVE_CACHE_L2X0=y
-CONFIG_MIGHT_HAVE_PCI=y
-CONFIG_MIGRATION=y
-CONFIG_MMC=y
-CONFIG_MMC_ARMMMCI=y
-CONFIG_MMC_BLOCK=y
-CONFIG_MMC_BLOCK_MINORS=16
-CONFIG_MMC_DW=y
-CONFIG_MMC_DW_EXYNOS=y
-# CONFIG_MMC_DW_K3 is not set
-# CONFIG_MMC_DW_PCI is not set
-CONFIG_MMC_DW_PLTFM=y
-# CONFIG_MMC_MXC is not set
-CONFIG_MMC_SDHCI=y
-CONFIG_MMC_SDHCI_ESDHC_IMX=y
-CONFIG_MMC_SDHCI_IO_ACCESSORS=y
-CONFIG_MMC_SDHCI_OF_ARASAN=y
-CONFIG_MMC_SDHCI_OF_AT91=y
-CONFIG_MMC_SDHCI_OF_ESDHC=y
-# CONFIG_MMC_SDHCI_PCI is not set
-CONFIG_MMC_SDHCI_PLTFM=y
-# CONFIG_MMC_TIFM_SD is not set
-CONFIG_MMU_NOTIFIER=y
-CONFIG_MODULES_TREE_LOOKUP=y
-CONFIG_MODULES_USE_ELF_REL=y
-# CONFIG_MOUSE_BCM5974 is not set
-# CONFIG_MOUSE_CYAPA is not set
-CONFIG_MOUSE_ELAN_I2C=y
-CONFIG_MOUSE_ELAN_I2C_I2C=y
-# CONFIG_MOUSE_ELAN_I2C_SMBUS is not set
-CONFIG_MOUSE_PS2=y
-CONFIG_MOUSE_PS2_ALPS=y
-CONFIG_MOUSE_PS2_BYD=y
-CONFIG_MOUSE_PS2_CYPRESS=y
-CONFIG_MOUSE_PS2_ELANTECH=y
-CONFIG_MOUSE_PS2_FOCALTECH=y
-CONFIG_MOUSE_PS2_LOGIPS2PP=y
-CONFIG_MOUSE_PS2_SYNAPTICS=y
-# CONFIG_MOUSE_PS2_TOUCHKIT is not set
-CONFIG_MOUSE_PS2_TRACKPOINT=y
-# CONFIG_MOUSE_SERIAL is not set
-# CONFIG_MOUSE_VSXXXAA is not set
-CONFIG_MSDOS_FS=y
-# CONFIG_MTD_BCM47XXSFLASH is not set
-CONFIG_MTD_CFI_ADV_OPTIONS=y
-CONFIG_MTD_CFI_GEOMETRY=y
-CONFIG_MTD_CFI_STAA=y
-CONFIG_MTD_CMDLINE_PARTS=y
-# CONFIG_MTD_COMPLEX_MAPPINGS is not set
-CONFIG_MTD_DATAFLASH=y
-# CONFIG_MTD_DATAFLASH_OTP is not set
-# CONFIG_MTD_DATAFLASH_WRITE_VERIFY is not set
-CONFIG_MTD_M25P80=y
-CONFIG_MTD_NAND=y
-CONFIG_MTD_NAND_BRCMNAND=y
-CONFIG_MTD_NAND_DENALI=y
-CONFIG_MTD_NAND_DENALI_DT=y
-CONFIG_MTD_NAND_ECC=y
-CONFIG_MTD_NAND_FSL_IFC=y
-CONFIG_MTD_SPI_NOR=y
-CONFIG_MTD_SPI_NOR_USE_4K_SECTORS=y
-CONFIG_MTD_SPI_NOR_USE_4K_SECTORS_LIMIT=16384
-CONFIG_MTD_SST25L=y
-CONFIG_MTD_UBI=y
-CONFIG_MTD_UBI_BEB_LIMIT=20
-# CONFIG_MTD_UBI_BLOCK is not set
-# CONFIG_MTD_UBI_FASTMAP is not set
-# CONFIG_MTD_UBI_GLUEBI is not set
-CONFIG_MTD_UBI_WL_THRESHOLD=4096
-CONFIG_MULTI_IRQ_HANDLER=y
-CONFIG_MUTEX_SPIN_ON_OWNER=y
-CONFIG_MVMDIO=y
-CONFIG_MX3_IPU=y
-CONFIG_MX3_IPU_IRQS=4
-CONFIG_NAMESPACES=y
-CONFIG_NATIONAL_PHY=y
-CONFIG_NEED_DMA_MAP_STATE=y
-CONFIG_NEON=y
-# CONFIG_NET_CLS_CGROUP is not set
-CONFIG_NET_FLOW_LIMIT=y
-CONFIG_NET_IP_TUNNEL=y
-CONFIG_NET_NS=y
-CONFIG_NET_PTP_CLASSIFY=y
-CONFIG_NLS=y
-CONFIG_NLS_CODEPAGE_437=y
-CONFIG_NLS_ISO8859_1=y
-CONFIG_NLS_UTF8=y
-CONFIG_NO_BOOTMEM=y
-CONFIG_NO_HZ=y
-CONFIG_NO_HZ_COMMON=y
-CONFIG_NO_HZ_IDLE=y
-CONFIG_NR_CPUS=16
-CONFIG_NTFS_FS=y
-CONFIG_NVMEM=y
-CONFIG_NVME_CORE=y
-CONFIG_OF=y
-CONFIG_OF_ADDRESS=y
-CONFIG_OF_ADDRESS_PCI=y
-CONFIG_OF_DYNAMIC=y
-CONFIG_OF_EARLY_FLATTREE=y
-CONFIG_OF_FLATTREE=y
-CONFIG_OF_GPIO=y
-CONFIG_OF_IRQ=y
-CONFIG_OF_MDIO=y
-CONFIG_OF_NET=y
-CONFIG_OF_PCI=y
-CONFIG_OF_PCI_IRQ=y
-CONFIG_OF_RESERVED_MEM=y
-CONFIG_OLD_SIGACTION=y
-CONFIG_OLD_SIGSUSPEND3=y
-CONFIG_OUTER_CACHE=y
-CONFIG_OUTER_CACHE_SYNC=y
-CONFIG_PACKET_DIAG=y
-CONFIG_PAGE_COUNTER=y
-CONFIG_PAGE_OFFSET=0xC0000000
-# CONFIG_PANIC_ON_OOPS is not set
-CONFIG_PANIC_ON_OOPS_VALUE=0
-CONFIG_PANIC_TIMEOUT=0
-CONFIG_PARTITION_PERCPU=y
-CONFIG_PCI=y
-CONFIG_PCIEAER=y
-CONFIG_PCIEASPM=y
-# CONFIG_PCIEASPM_DEBUG is not set
-CONFIG_PCIEASPM_DEFAULT=y
-# CONFIG_PCIEASPM_PERFORMANCE is not set
-# CONFIG_PCIEASPM_POWERSAVE is not set
-CONFIG_PCIEPORTBUS=y
-CONFIG_PCIE_DW=y
-CONFIG_PCIE_PME=y
-CONFIG_PCI_BUS_ADDR_T_64BIT=y
-CONFIG_PCI_DOMAINS=y
-CONFIG_PCI_DOMAINS_GENERIC=y
-CONFIG_PCI_ECAM=y
-CONFIG_PCI_HOST_COMMON=y
-CONFIG_PCI_HOST_GENERIC=y
-CONFIG_PCI_LAYERSCAPE=y
-CONFIG_PCI_MSI=y
-CONFIG_PCI_MSI_IRQ_DOMAIN=y
-CONFIG_PERF_EVENTS=y
-CONFIG_PERF_USE_VMALLOC=y
-CONFIG_PGTABLE_LEVELS=3
-CONFIG_PHYLIB=y
-CONFIG_PHYS_ADDR_T_64BIT=y
-CONFIG_PID_NS=y
-CONFIG_PINCTRL=y
-CONFIG_PINCTRL_AS3722=y
-CONFIG_PINCTRL_PALMAS=y
-CONFIG_PL310_ERRATA_588369=y
-CONFIG_PL310_ERRATA_727915=y
-CONFIG_PL310_ERRATA_753970=y
-CONFIG_PL310_ERRATA_769419=y
-CONFIG_PL320_MBOX=y
-CONFIG_PL330_DMA=y
-# CONFIG_PLAT_VERSATILE_CLCD is not set
-CONFIG_PM=y
-CONFIG_PM_CLK=y
-# CONFIG_PM_DEBUG is not set
-CONFIG_PM_DEVFREQ=y
-# CONFIG_PM_DEVFREQ_EVENT is not set
-CONFIG_PM_OPP=y
-CONFIG_PM_SLEEP=y
-CONFIG_PM_SLEEP_SMP=y
-CONFIG_POSIX_MQUEUE=y
-CONFIG_POSIX_MQUEUE_SYSCTL=y
-CONFIG_POWER_AVS=y
-CONFIG_POWER_RESET=y
-CONFIG_POWER_RESET_AS3722=y
-CONFIG_POWER_RESET_BRCMKONA=y
-CONFIG_POWER_RESET_BRCMSTB=y
-CONFIG_POWER_RESET_GPIO=y
-CONFIG_POWER_RESET_GPIO_RESTART=y
-CONFIG_POWER_RESET_SYSCON=y
-CONFIG_POWER_RESET_SYSCON_POWEROFF=y
-CONFIG_POWER_RESET_VEXPRESS=y
-CONFIG_POWER_SUPPLY=y
-CONFIG_PPS=y
-CONFIG_PREEMPT_NOTIFIERS=y
-CONFIG_PRINTK_TIME=y
-# CONFIG_PROBE_EVENTS is not set
-CONFIG_PROC_CHILDREN=y
-CONFIG_PROC_PID_CPUSET=y
-CONFIG_PSTORE=y
-CONFIG_PSTORE_CONSOLE=y
-# CONFIG_PSTORE_LZ4_COMPRESS is not set
-# CONFIG_PSTORE_LZO_COMPRESS is not set
-CONFIG_PSTORE_PMSG=y
-CONFIG_PSTORE_RAM=y
-CONFIG_PSTORE_ZLIB_COMPRESS=y
-CONFIG_PTP_1588_CLOCK=y
-# CONFIG_PTP_1588_CLOCK_DPAA is not set
-CONFIG_PTP_1588_CLOCK_GIANFAR=y
-CONFIG_PWM=y
-# CONFIG_PWM_IMX is not set
-# CONFIG_PWM_STMPE is not set
-CONFIG_PWM_SYSFS=y
-# CONFIG_PWM_TWL is not set
-# CONFIG_PWM_TWL_LED is not set
-CONFIG_QMAN_CEETM_UPDATE_PERIOD=1000
-CONFIG_QORIQ_CPUFREQ=y
-# CONFIG_QUICC_ENGINE is not set
-CONFIG_R8169=y
-CONFIG_RAS=y
-CONFIG_RATIONAL=y
-CONFIG_RCU_CPU_STALL_TIMEOUT=21
-# CONFIG_RCU_EXPERT is not set
-CONFIG_RCU_STALL_COMMON=y
-CONFIG_RD_BZIP2=y
-CONFIG_RD_GZIP=y
-CONFIG_RD_LZMA=y
-CONFIG_RD_LZO=y
-CONFIG_RD_XZ=y
-CONFIG_REALTEK_PHY=y
-CONFIG_REED_SOLOMON=y
-CONFIG_REED_SOLOMON_DEC8=y
-CONFIG_REED_SOLOMON_ENC8=y
-CONFIG_REGMAP=y
-CONFIG_REGMAP_I2C=y
-CONFIG_REGMAP_IRQ=y
-CONFIG_REGMAP_MMIO=y
-CONFIG_REGMAP_SPI=y
-CONFIG_REGULATOR=y
-CONFIG_REGULATOR_ACT8865=y
-CONFIG_REGULATOR_ACT8945A=y
-CONFIG_REGULATOR_ANATOP=y
-CONFIG_REGULATOR_AS3711=y
-CONFIG_REGULATOR_AS3722=y
-CONFIG_REGULATOR_AXP20X=y
-CONFIG_REGULATOR_BCM590XX=y
-CONFIG_REGULATOR_DA9210=y
-CONFIG_REGULATOR_FAN53555=y
-CONFIG_REGULATOR_FIXED_VOLTAGE=y
-CONFIG_REGULATOR_GPIO=y
-CONFIG_REGULATOR_LP872X=y
-# CONFIG_REGULATOR_MAX14577 is not set
-CONFIG_REGULATOR_MAX77686=y
-# CONFIG_REGULATOR_MAX77802 is not set
-CONFIG_REGULATOR_MAX8907=y
-CONFIG_REGULATOR_MAX8973=y
-# CONFIG_REGULATOR_MAX8997 is not set
-# CONFIG_REGULATOR_MAX8998 is not set
-CONFIG_REGULATOR_PALMAS=y
-CONFIG_REGULATOR_PWM=y
-# CONFIG_REGULATOR_QCOM_SPMI is not set
-CONFIG_REGULATOR_RK808=y
-# CONFIG_REGULATOR_S2MPA01 is not set
-CONFIG_REGULATOR_S2MPS11=y
-CONFIG_REGULATOR_S5M8767=y
-CONFIG_REGULATOR_TPS51632=y
-CONFIG_REGULATOR_TPS62360=y
-CONFIG_REGULATOR_TPS65090=y
-CONFIG_REGULATOR_TPS65217=y
-CONFIG_REGULATOR_TPS65218=y
-CONFIG_REGULATOR_TPS6586X=y
-CONFIG_REGULATOR_TPS65910=y
-CONFIG_REGULATOR_TWL4030=y
-CONFIG_REGULATOR_VEXPRESS=y
-CONFIG_RESET_CONTROLLER=y
-CONFIG_RFS_ACCEL=y
-CONFIG_RPS=y
-CONFIG_RTC_CLASS=y
-CONFIG_RTC_DRV_AS3722=y
-# CONFIG_RTC_DRV_CMOS is not set
-CONFIG_RTC_DRV_DS1307=y
-CONFIG_RTC_DRV_DS1307_HWMON=y
-CONFIG_RTC_DRV_DS3232=y
-# CONFIG_RTC_DRV_EFI is not set
-CONFIG_RTC_DRV_EM3027=y
-# CONFIG_RTC_DRV_IMXDI is not set
-CONFIG_RTC_DRV_MAX77686=y
-CONFIG_RTC_DRV_MAX8907=y
-# CONFIG_RTC_DRV_MAX8997 is not set
-# CONFIG_RTC_DRV_MAX8998 is not set
-# CONFIG_RTC_DRV_MXC is not set
-CONFIG_RTC_DRV_PALMAS=y
-CONFIG_RTC_DRV_PCF2127=y
-CONFIG_RTC_DRV_PCF85263=y
-CONFIG_RTC_DRV_PL031=y
-# CONFIG_RTC_DRV_RK808 is not set
-# CONFIG_RTC_DRV_S5M is not set
-CONFIG_RTC_DRV_TPS6586X=y
-CONFIG_RTC_DRV_TPS65910=y
-CONFIG_RTC_DRV_TWL4030=y
-CONFIG_RTC_I2C_AND_SPI=y
-# CONFIG_RT_GROUP_SCHED is not set
-CONFIG_RWSEM_SPIN_ON_OWNER=y
-CONFIG_RWSEM_XCHGADD_ALGORITHM=y
-CONFIG_SATA_AHCI=y
-CONFIG_SATA_AHCI_PLATFORM=y
-CONFIG_SATA_MV=y
-CONFIG_SATA_PMP=y
-CONFIG_SATA_SIL24=y
-CONFIG_SCHED_DEBUG=y
-# CONFIG_SCHED_INFO is not set
-CONFIG_SCSI=y
-CONFIG_SECCOMP=y
-CONFIG_SECCOMP_FILTER=y
-CONFIG_SENSORS_IIO_HWMON=y
-CONFIG_SENSORS_ISL29018=y
-CONFIG_SENSORS_ISL29028=y
-CONFIG_SENSORS_LM90=y
-CONFIG_SENSORS_LM95245=y
-CONFIG_SERIAL_8250_DEPRECATED_OPTIONS=y
-CONFIG_SERIAL_8250_DW=y
-CONFIG_SERIAL_8250_EM=y
-CONFIG_SERIAL_8250_FSL=y
-CONFIG_SERIAL_8250_NR_UARTS=4
-CONFIG_SERIAL_8250_PCI=y
-CONFIG_SERIAL_8250_RUNTIME_UARTS=4
-CONFIG_SERIAL_AMBA_PL011=y
-CONFIG_SERIAL_AMBA_PL011_CONSOLE=y
-CONFIG_SERIAL_BCM63XX=y
-CONFIG_SERIAL_BCM63XX_CONSOLE=y
-CONFIG_SERIAL_CONEXANT_DIGICOLOR=y
-CONFIG_SERIAL_CONEXANT_DIGICOLOR_CONSOLE=y
-CONFIG_SERIAL_FSL_LPUART=y
-CONFIG_SERIAL_FSL_LPUART_CONSOLE=y
-CONFIG_SERIAL_IMX=y
-CONFIG_SERIAL_IMX_CONSOLE=y
-CONFIG_SERIAL_MCTRL_GPIO=y
-CONFIG_SERIAL_OF_PLATFORM=y
-CONFIG_SERIAL_ST_ASC=y
-CONFIG_SERIAL_ST_ASC_CONSOLE=y
-CONFIG_SERIAL_XILINX_PS_UART=y
-CONFIG_SERIAL_XILINX_PS_UART_CONSOLE=y
-CONFIG_SERIO=y
-CONFIG_SERIO_AMBAKMI=y
-CONFIG_SERIO_LIBPS2=y
-CONFIG_SERIO_SERPORT=y
-CONFIG_SG_POOL=y
-CONFIG_SLUB_DEBUG=y
-CONFIG_SMP=y
-CONFIG_SMP_ON_UP=y
-CONFIG_SMSC911X=y
-# CONFIG_SMSC911X_ARCH_HOOKS is not set
-CONFIG_SMSC_PHY=y
-CONFIG_SOCK_DIAG=y
-CONFIG_SOC_BRCMSTB=y
-CONFIG_SOC_BUS=y
-# CONFIG_SOC_IMX50 is not set
-# CONFIG_SOC_IMX51 is not set
-# CONFIG_SOC_IMX53 is not set
-# CONFIG_SOC_IMX6Q is not set
-# CONFIG_SOC_IMX6SL is not set
-# CONFIG_SOC_IMX6SX is not set
-# CONFIG_SOC_IMX6UL is not set
-# CONFIG_SOC_IMX7D is not set
-CONFIG_SOC_LS1021A=y
-# CONFIG_SOC_VF610 is not set
-CONFIG_SPARSE_IRQ=y
-CONFIG_SPI=y
-CONFIG_SPI_BITBANG=y
-CONFIG_SPI_CADENCE=y
-CONFIG_SPI_FSL_QUADSPI=y
-# CONFIG_SPI_IMX is not set
-CONFIG_SPI_MASTER=y
-CONFIG_SPI_PL022=y
-CONFIG_SPI_SPIDEV=y
-CONFIG_SPI_XILINX=y
-CONFIG_SPMI=y
-# CONFIG_SQUASHFS_DECOMP_MULTI_PERCPU is not set
-CONFIG_SQUASHFS_DECOMP_SINGLE=y
-CONFIG_SQUASHFS_FILE_CACHE=y
-# CONFIG_SQUASHFS_FILE_DIRECT is not set
-CONFIG_SQUASHFS_LZO=y
-CONFIG_SQUASHFS_ZLIB=y
-CONFIG_SRAM=y
-CONFIG_SRCU=y
-CONFIG_STAGING_BOARD=y
-CONFIG_STMMAC_ETH=y
-CONFIG_STMMAC_PLATFORM=y
-CONFIG_STMPE_I2C=y
-# CONFIG_STMPE_SPI is not set
-# CONFIG_STRIP_ASM_SYMS is not set
-CONFIG_SUSPEND=y
-CONFIG_SUSPEND_FREEZER=y
-CONFIG_SWIOTLB=y
-CONFIG_SWPHY=y
-CONFIG_SWP_EMULATE=y
-# CONFIG_SW_SYNC is not set
-CONFIG_SYNC_FILE=y
-# CONFIG_SYN_COOKIES is not set
-CONFIG_SYSFS_SYSCALL=y
-CONFIG_SYS_SUPPORTS_APM_EMULATION=y
-CONFIG_SYS_SUPPORTS_HUGETLBFS=y
-# CONFIG_TCP_CONG_ADVANCED is not set
-CONFIG_THERMAL=y
-CONFIG_THERMAL_DEFAULT_GOV_STEP_WISE=y
-CONFIG_THERMAL_GOV_STEP_WISE=y
-CONFIG_THERMAL_HWMON=y
-CONFIG_THERMAL_OF=y
-# CONFIG_THUMB2_KERNEL is not set
-CONFIG_TICK_CPU_ACCOUNTING=y
-CONFIG_TI_CPSW_ALE=y
-CONFIG_TMPFS_POSIX_ACL=y
-CONFIG_TOUCHSCREEN_PROPERTIES=y
-CONFIG_TOUCHSCREEN_STMPE=y
-CONFIG_TRACING_EVENTS_GPIO=y
-CONFIG_TREE_RCU=y
-CONFIG_TUN=y
-CONFIG_TWL4030_CORE=y
-CONFIG_TWL4030_POWER=y
-# CONFIG_TWL4030_WATCHDOG is not set
-CONFIG_UBIFS_FS=y
-# CONFIG_UBIFS_FS_ADVANCED_COMPR is not set
-CONFIG_UBIFS_FS_LZO=y
-CONFIG_UBIFS_FS_ZLIB=y
-CONFIG_UCS2_STRING=y
-CONFIG_UEVENT_HELPER_PATH=""
-CONFIG_UNCOMPRESS_INCLUDE="debug/uncompress.h"
-CONFIG_UNIX_DIAG=y
-# CONFIG_USB_IMX21_HCD is not set
-CONFIG_USB_SUPPORT=y
-# CONFIG_USERIO is not set
-CONFIG_USER_NS=y
-CONFIG_USE_OF=y
-CONFIG_UTS_NS=y
-CONFIG_VDSO=y
-CONFIG_VECTORS_BASE=0xffff0000
-CONFIG_VETH=y
-CONFIG_VEXPRESS_CONFIG=y
-CONFIG_VEXPRESS_SYSCFG=y
-CONFIG_VFAT_FS=y
-CONFIG_VFP=y
-CONFIG_VFPv3=y
-CONFIG_VGA_ARB=y
-CONFIG_VGA_ARB_MAX_GPUS=16
-CONFIG_VHOST=y
-CONFIG_VHOST_NET=y
-CONFIG_VIDEOMODE_HELPERS=y
-CONFIG_VIRTIO=y
-CONFIG_VIRTIO_BLK=y
-CONFIG_VIRTIO_CONSOLE=y
-CONFIG_VIRTIO_MMIO=y
-# CONFIG_VIRTIO_MMIO_CMDLINE_DEVICES is not set
-CONFIG_VIRTIO_NET=y
-CONFIG_VIRTIO_PCI=y
-CONFIG_VIRTIO_PCI_LEGACY=y
-CONFIG_VIRTUALIZATION=y
-CONFIG_VITESSE_PHY=y
-CONFIG_VM_EVENT_COUNTERS=y
-CONFIG_VT=y
-CONFIG_VT_CONSOLE=y
-CONFIG_VT_CONSOLE_SLEEP=y
-CONFIG_VT_HW_CONSOLE_BINDING=y
-CONFIG_WATCHDOG_CORE=y
-# CONFIG_WQ_POWER_EFFICIENT_DEFAULT is not set
-CONFIG_XFRM_ALGO=y
-CONFIG_XFRM_USER=y
-CONFIG_XILINX_WATCHDOG=y
-CONFIG_XPS=y
-CONFIG_XZ_DEC_ARM=y
-CONFIG_XZ_DEC_ARMTHUMB=y
-CONFIG_XZ_DEC_BCJ=y
-CONFIG_XZ_DEC_IA64=y
-CONFIG_XZ_DEC_POWERPC=y
-CONFIG_XZ_DEC_SPARC=y
-CONFIG_XZ_DEC_X86=y
-CONFIG_ZBOOT_ROM_BSS=0
-CONFIG_ZBOOT_ROM_TEXT=0
-CONFIG_ZLIB_DEFLATE=y
-CONFIG_ZLIB_INFLATE=y
diff --git a/target/linux/layerscape/armv8_64b/config-4.9 b/target/linux/layerscape/armv8_64b/config-4.9
deleted file mode 100644
index 34e6a01ec2..0000000000
--- a/target/linux/layerscape/armv8_64b/config-4.9
+++ /dev/null
@@ -1,1242 +0,0 @@
-CONFIG_64BIT=y
-CONFIG_ACPI=y
-CONFIG_ACPI_BUTTON=y
-CONFIG_ACPI_CCA_REQUIRED=y
-CONFIG_ACPI_CONTAINER=y
-# CONFIG_ACPI_CPPC_CPUFREQ is not set
-# CONFIG_ACPI_CUSTOM_DSDT is not set
-# CONFIG_ACPI_DEBUG is not set
-# CONFIG_ACPI_DEBUGGER is not set
-# CONFIG_ACPI_DOCK is not set
-# CONFIG_ACPI_EC_DEBUGFS is not set
-CONFIG_ACPI_FAN=y
-CONFIG_ACPI_GENERIC_GSI=y
-CONFIG_ACPI_HOTPLUG_CPU=y
-CONFIG_ACPI_I2C_OPREGION=y
-CONFIG_ACPI_IORT=y
-CONFIG_ACPI_MCFG=y
-CONFIG_ACPI_NUMA=y
-# CONFIG_ACPI_PCI_SLOT is not set
-CONFIG_ACPI_PROCESSOR=y
-CONFIG_ACPI_PROCESSOR_IDLE=y
-CONFIG_ACPI_REDUCED_HARDWARE_ONLY=y
-CONFIG_ACPI_SPCR_TABLE=y
-CONFIG_ACPI_TABLE_UPGRADE=y
-CONFIG_ACPI_THERMAL=y
-CONFIG_AHCI_CEVA=y
-CONFIG_AHCI_QORIQ=y
-CONFIG_AHCI_XGENE=y
-CONFIG_AMD_XGBE=y
-CONFIG_AQUANTIA_PHY=y
-CONFIG_ARCH_CLOCKSOURCE_DATA=y
-CONFIG_ARCH_DMA_ADDR_T_64BIT=y
-CONFIG_ARCH_HAS_ACPI_TABLE_UPGRADE=y
-CONFIG_ARCH_HAS_CACHE_LINE_SIZE=y
-CONFIG_ARCH_HAS_ELF_RANDOMIZE=y
-CONFIG_ARCH_HAS_GCOV_PROFILE_ALL=y
-CONFIG_ARCH_HAS_GIGANTIC_PAGE=y
-CONFIG_ARCH_HAS_HOLES_MEMORYMODEL=y
-CONFIG_ARCH_HAS_KCOV=y
-CONFIG_ARCH_HAS_SG_CHAIN=y
-CONFIG_ARCH_HAS_TICK_BROADCAST=y
-CONFIG_ARCH_HIBERNATION_HEADER=y
-CONFIG_ARCH_HIBERNATION_POSSIBLE=y
-CONFIG_ARCH_LAYERSCAPE=y
-CONFIG_ARCH_MMAP_RND_BITS=18
-CONFIG_ARCH_MMAP_RND_BITS_MAX=33
-CONFIG_ARCH_MMAP_RND_BITS_MIN=18
-CONFIG_ARCH_MMAP_RND_COMPAT_BITS=11
-CONFIG_ARCH_MMAP_RND_COMPAT_BITS_MIN=11
-CONFIG_ARCH_PHYS_ADDR_T_64BIT=y
-CONFIG_ARCH_SELECT_MEMORY_MODEL=y
-CONFIG_ARCH_SPARSEMEM_DEFAULT=y
-CONFIG_ARCH_SPARSEMEM_ENABLE=y
-CONFIG_ARCH_SUPPORTS_ATOMIC_RMW=y
-CONFIG_ARCH_SUPPORTS_DEBUG_PAGEALLOC=y
-CONFIG_ARCH_SUPPORTS_NUMA_BALANCING=y
-CONFIG_ARCH_SUSPEND_POSSIBLE=y
-CONFIG_ARCH_USE_CMPXCHG_LOCKREF=y
-CONFIG_ARCH_WANT_COMPAT_IPC_PARSE_VERSION=y
-CONFIG_ARCH_WANT_FRAME_POINTERS=y
-CONFIG_ARCH_WANT_HUGE_PMD_SHARE=y
-CONFIG_ARM64=y
-# CONFIG_ARM64_16K_PAGES is not set
-CONFIG_ARM64_4K_PAGES=y
-# CONFIG_ARM64_64K_PAGES is not set
-# CONFIG_ARM64_ACPI_PARKING_PROTOCOL is not set
-CONFIG_ARM64_CONT_SHIFT=4
-CONFIG_ARM64_CRYPTO=y
-CONFIG_ARM64_ERRATUM_819472=y
-CONFIG_ARM64_ERRATUM_824069=y
-CONFIG_ARM64_ERRATUM_826319=y
-CONFIG_ARM64_ERRATUM_827319=y
-CONFIG_ARM64_ERRATUM_832075=y
-CONFIG_ARM64_ERRATUM_834220=y
-CONFIG_ARM64_ERRATUM_843419=y
-CONFIG_ARM64_ERRATUM_845719=y
-CONFIG_ARM64_HW_AFDBM=y
-# CONFIG_ARM64_LSE_ATOMICS is not set
-CONFIG_ARM64_MODULE_CMODEL_LARGE=y
-CONFIG_ARM64_PAGE_SHIFT=12
-CONFIG_ARM64_PAN=y
-# CONFIG_ARM64_PTDUMP is not set
-# CONFIG_ARM64_RANDOMIZE_TEXT_OFFSET is not set
-CONFIG_ARM64_SSBD=y
-CONFIG_ARM64_UAO=y
-CONFIG_ARM64_VA_BITS=48
-# CONFIG_ARM64_VA_BITS_39 is not set
-CONFIG_ARM64_VA_BITS_48=y
-CONFIG_ARM64_VHE=y
-# CONFIG_ARMV8_DEPRECATED is not set
-CONFIG_ARM_AMBA=y
-CONFIG_ARM_ARCH_TIMER=y
-CONFIG_ARM_ARCH_TIMER_EVTSTREAM=y
-CONFIG_ARM_BIG_LITTLE_CPUFREQ=y
-CONFIG_ARM_CPUIDLE=y
-# CONFIG_ARM_DT_BL_CPUFREQ is not set
-CONFIG_ARM_GIC=y
-CONFIG_ARM_GIC_V2M=y
-CONFIG_ARM_GIC_V3=y
-CONFIG_ARM_GIC_V3_ITS=y
-CONFIG_ARM_GIC_V3_ITS_FSL_MC=y
-# CONFIG_ARM_PL172_MPMC is not set
-CONFIG_ARM_PMU=y
-CONFIG_ARM_PSCI_FW=y
-CONFIG_ARM_SMMU=y
-CONFIG_ARM_SMMU_V3=y
-CONFIG_ARM_SP805_WATCHDOG=y
-CONFIG_ARM_TIMER_SP804=y
-CONFIG_ASN1=y
-CONFIG_ASSOCIATIVE_ARRAY=y
-CONFIG_ASYNC_CORE=y
-CONFIG_ASYNC_TX_ENABLE_CHANNEL_SWITCH=y
-CONFIG_ATA=y
-CONFIG_ATA_ACPI=y
-CONFIG_ATA_VERBOSE_ERROR=y
-CONFIG_AUDIT=y
-CONFIG_AUDITSYSCALL=y
-CONFIG_AUDIT_ARCH_COMPAT_GENERIC=y
-CONFIG_AUDIT_COMPAT_GENERIC=y
-CONFIG_AUDIT_GENERIC=y
-CONFIG_AUDIT_TREE=y
-CONFIG_AUDIT_WATCH=y
-CONFIG_AUTOFS4_FS=y
-CONFIG_BACKLIGHT_CLASS_DEVICE=y
-CONFIG_BACKLIGHT_LCD_SUPPORT=y
-# CONFIG_BACKLIGHT_PWM is not set
-CONFIG_BALLOON_COMPACTION=y
-CONFIG_BATTERY_BQ27XXX=y
-CONFIG_BATTERY_BQ27XXX_I2C=y
-CONFIG_BLK_CGROUP=y
-CONFIG_BLK_DEV_BSG=y
-CONFIG_BLK_DEV_INTEGRITY=y
-CONFIG_BLK_DEV_LOOP=y
-CONFIG_BLK_DEV_NVME=y
-# CONFIG_BLK_DEV_NVME_SCSI is not set
-CONFIG_BLK_DEV_RAM=y
-CONFIG_BLK_DEV_RAM_COUNT=16
-CONFIG_BLK_DEV_RAM_SIZE=262144
-CONFIG_BLK_DEV_SD=y
-CONFIG_BLK_DEV_THROTTLING=y
-CONFIG_BLK_MQ_PCI=y
-CONFIG_BLOCK_COMPAT=y
-# CONFIG_BOOTPARAM_HUNG_TASK_PANIC is not set
-CONFIG_BOOTPARAM_HUNG_TASK_PANIC_VALUE=0
-# CONFIG_BOOTPARAM_SOFTLOCKUP_PANIC is not set
-CONFIG_BOOTPARAM_SOFTLOCKUP_PANIC_VALUE=0
-CONFIG_BOUNCE=y
-CONFIG_BPF_JIT=y
-# CONFIG_BPF_SYSCALL is not set
-CONFIG_BRIDGE_VLAN_FILTERING=y
-CONFIG_BSD_PROCESS_ACCT=y
-CONFIG_BSD_PROCESS_ACCT_V3=y
-CONFIG_BTRFS_FS=y
-# CONFIG_BTRFS_FS_CHECK_INTEGRITY is not set
-CONFIG_BTRFS_FS_POSIX_ACL=y
-CONFIG_BUILD_BIN2C=y
-CONFIG_CAVIUM_ERRATUM_22375=y
-CONFIG_CAVIUM_ERRATUM_23144=y
-CONFIG_CAVIUM_ERRATUM_23154=y
-CONFIG_CAVIUM_ERRATUM_27456=y
-CONFIG_CC_OPTIMIZE_FOR_PERFORMANCE=y
-# CONFIG_CC_OPTIMIZE_FOR_SIZE is not set
-CONFIG_CEPH_LIB=y
-# CONFIG_CEPH_LIB_PRETTYDEBUG is not set
-# CONFIG_CEPH_LIB_USE_DNS_RESOLVER is not set
-CONFIG_CFQ_GROUP_IOSCHED=y
-# CONFIG_CFS_BANDWIDTH is not set
-CONFIG_CGROUPS=y
-CONFIG_CGROUP_CPUACCT=y
-CONFIG_CGROUP_DEVICE=y
-CONFIG_CGROUP_FREEZER=y
-CONFIG_CGROUP_HUGETLB=y
-CONFIG_CGROUP_PERF=y
-CONFIG_CGROUP_PIDS=y
-CONFIG_CGROUP_SCHED=y
-CONFIG_CGROUP_WRITEBACK=y
-CONFIG_CHECKPOINT_RESTORE=y
-CONFIG_CHROME_PLATFORMS=y
-CONFIG_CLKDEV_LOOKUP=y
-CONFIG_CLKSRC_ACPI=y
-CONFIG_CLKSRC_MMIO=y
-CONFIG_CLKSRC_OF=y
-CONFIG_CLKSRC_PROBE=y
-CONFIG_CLKSRC_VERSATILE=y
-CONFIG_CLK_QORIQ=y
-CONFIG_CLK_SP810=y
-CONFIG_CLK_VEXPRESS_OSC=y
-CONFIG_CLONE_BACKWARDS=y
-CONFIG_CLZ_TAB=y
-CONFIG_CMA=y
-CONFIG_CMA_ALIGNMENT=8
-CONFIG_CMA_AREAS=7
-# CONFIG_CMA_DEBUG is not set
-# CONFIG_CMA_DEBUGFS is not set
-CONFIG_CMA_SIZE_MBYTES=16
-# CONFIG_CMA_SIZE_SEL_MAX is not set
-CONFIG_CMA_SIZE_SEL_MBYTES=y
-# CONFIG_CMA_SIZE_SEL_MIN is not set
-# CONFIG_CMA_SIZE_SEL_PERCENTAGE is not set
-CONFIG_COMMON_CLK=y
-CONFIG_COMMON_CLK_CS2000_CP=y
-# CONFIG_COMMON_CLK_MAX77686 is not set
-CONFIG_COMMON_CLK_PWM=y
-CONFIG_COMMON_CLK_RK808=y
-CONFIG_COMMON_CLK_S2MPS11=y
-CONFIG_COMMON_CLK_VERSATILE=y
-CONFIG_COMMON_CLK_XGENE=y
-CONFIG_COMPAT=y
-CONFIG_COMPAT_BINFMT_ELF=y
-CONFIG_COMPAT_NETLINK_MESSAGES=y
-CONFIG_COMPAT_OLD_SIGACTION=y
-CONFIG_CONFIGFS_FS=y
-CONFIG_CONSOLE_TRANSLATIONS=y
-CONFIG_COREDUMP=y
-# CONFIG_CORTINA_PHY is not set
-CONFIG_CPUFREQ_DT=y
-CONFIG_CPUFREQ_DT_PLATDEV=y
-CONFIG_CPUSETS=y
-# CONFIG_CPU_BIG_ENDIAN is not set
-CONFIG_CPU_FREQ=y
-CONFIG_CPU_FREQ_DEFAULT_GOV_PERFORMANCE=y
-CONFIG_CPU_FREQ_GOV_ATTR_SET=y
-CONFIG_CPU_FREQ_GOV_COMMON=y
-CONFIG_CPU_FREQ_GOV_CONSERVATIVE=y
-CONFIG_CPU_FREQ_GOV_ONDEMAND=y
-CONFIG_CPU_FREQ_GOV_PERFORMANCE=y
-CONFIG_CPU_FREQ_GOV_POWERSAVE=y
-CONFIG_CPU_FREQ_GOV_USERSPACE=y
-CONFIG_CPU_FREQ_STAT=y
-# CONFIG_CPU_HOTPLUG_STATE_CONTROL is not set
-CONFIG_CPU_IDLE=y
-# CONFIG_CPU_IDLE_GOV_LADDER is not set
-CONFIG_CPU_IDLE_GOV_MENU=y
-CONFIG_CPU_PM=y
-CONFIG_CPU_RMAP=y
-CONFIG_CPU_THERMAL=y
-CONFIG_CRC16=y
-# CONFIG_CRC32_SARWATE is not set
-CONFIG_CRC32_SLICEBY8=y
-CONFIG_CRC7=y
-CONFIG_CRC_ITU_T=y
-CONFIG_CRC_T10DIF=y
-CONFIG_CROSS_MEMORY_ATTACH=y
-# CONFIG_CROS_EC_CHARDEV is not set
-CONFIG_CROS_EC_PROTO=y
-# CONFIG_CROS_KBD_LED_BACKLIGHT is not set
-CONFIG_CRYPTO_ABLK_HELPER=y
-CONFIG_CRYPTO_AEAD=y
-CONFIG_CRYPTO_AEAD2=y
-CONFIG_CRYPTO_AES_ARM64_CE=y
-CONFIG_CRYPTO_AES_ARM64_CE_BLK=y
-CONFIG_CRYPTO_AES_ARM64_CE_CCM=y
-# CONFIG_CRYPTO_AES_ARM64_NEON_BLK is not set
-CONFIG_CRYPTO_AKCIPHER=y
-CONFIG_CRYPTO_AKCIPHER2=y
-CONFIG_CRYPTO_ANSI_CPRNG=y
-CONFIG_CRYPTO_AUTHENC=y
-CONFIG_CRYPTO_CBC=y
-CONFIG_CRYPTO_CRC32C=y
-# CONFIG_CRYPTO_CRC32_ARM64 is not set
-CONFIG_CRYPTO_CRCT10DIF=y
-CONFIG_CRYPTO_CRYPTD=y
-CONFIG_CRYPTO_DEFLATE=y
-CONFIG_CRYPTO_DES=y
-CONFIG_CRYPTO_DEV_FSL_CAAM=y
-CONFIG_CRYPTO_DEV_FSL_CAAM_AHASH_API=y
-CONFIG_CRYPTO_DEV_FSL_CAAM_AHASH_API_DESC=y
-CONFIG_CRYPTO_DEV_FSL_CAAM_COMMON=y
-CONFIG_CRYPTO_DEV_FSL_CAAM_CRYPTO_API=y
-CONFIG_CRYPTO_DEV_FSL_CAAM_CRYPTO_API_DESC=y
-CONFIG_CRYPTO_DEV_FSL_CAAM_CRYPTO_API_QI=y
-# CONFIG_CRYPTO_DEV_FSL_CAAM_DEBUG is not set
-CONFIG_CRYPTO_DEV_FSL_CAAM_DMA=y
-# CONFIG_CRYPTO_DEV_FSL_CAAM_INTC is not set
-CONFIG_CRYPTO_DEV_FSL_CAAM_JR=y
-CONFIG_CRYPTO_DEV_FSL_CAAM_PKC_API=y
-CONFIG_CRYPTO_DEV_FSL_CAAM_RINGSIZE=9
-CONFIG_CRYPTO_DEV_FSL_CAAM_RNG_API=y
-CONFIG_CRYPTO_DEV_FSL_DPAA2_CAAM=y
-CONFIG_CRYPTO_DRBG=y
-CONFIG_CRYPTO_DRBG_HMAC=y
-CONFIG_CRYPTO_DRBG_MENU=y
-CONFIG_CRYPTO_ECHAINIV=y
-CONFIG_CRYPTO_GHASH_ARM64_CE=y
-CONFIG_CRYPTO_HASH=y
-CONFIG_CRYPTO_HASH2=y
-CONFIG_CRYPTO_HMAC=y
-CONFIG_CRYPTO_HW=y
-CONFIG_CRYPTO_JITTERENTROPY=y
-CONFIG_CRYPTO_LZO=y
-CONFIG_CRYPTO_MANAGER=y
-CONFIG_CRYPTO_MANAGER2=y
-CONFIG_CRYPTO_MD5=y
-CONFIG_CRYPTO_NULL=y
-CONFIG_CRYPTO_NULL2=y
-# CONFIG_CRYPTO_PCRYPT is not set
-CONFIG_CRYPTO_RNG=y
-CONFIG_CRYPTO_RNG2=y
-CONFIG_CRYPTO_RNG_DEFAULT=y
-CONFIG_CRYPTO_RSA=y
-CONFIG_CRYPTO_SHA1=y
-CONFIG_CRYPTO_SHA1_ARM64_CE=y
-CONFIG_CRYPTO_SHA256=y
-CONFIG_CRYPTO_SHA2_ARM64_CE=y
-CONFIG_CRYPTO_SHA512=y
-# CONFIG_CRYPTO_TLS is not set
-CONFIG_CRYPTO_WORKQUEUE=y
-CONFIG_DCACHE_WORD_ACCESS=y
-# CONFIG_DEBUG_ALIGN_RODATA is not set
-# CONFIG_DEBUG_BLK_CGROUP is not set
-CONFIG_DEBUG_BUGVERBOSE=y
-CONFIG_DEBUG_INFO=y
-# CONFIG_DEBUG_INFO_REDUCED is not set
-CONFIG_DEBUG_MEMORY_INIT=y
-CONFIG_DEBUG_RODATA=y
-CONFIG_DEBUG_SET_MODULE_RONX=y
-CONFIG_DECOMPRESS_BZIP2=y
-CONFIG_DECOMPRESS_GZIP=y
-CONFIG_DECOMPRESS_LZMA=y
-CONFIG_DECOMPRESS_LZO=y
-CONFIG_DECOMPRESS_XZ=y
-CONFIG_DEFAULT_CFQ=y
-CONFIG_DEFAULT_HUNG_TASK_TIMEOUT=120
-CONFIG_DEFAULT_IOSCHED="cfq"
-CONFIG_DETECT_HUNG_TASK=y
-CONFIG_DEVTMPFS=y
-CONFIG_DEVTMPFS_MOUNT=y
-# CONFIG_DEV_DAX is not set
-CONFIG_DMADEVICES=y
-CONFIG_DMATEST=y
-CONFIG_DMA_ACPI=y
-CONFIG_DMA_CMA=y
-CONFIG_DMA_ENGINE=y
-CONFIG_DMA_ENGINE_RAID=y
-CONFIG_DMA_OF=y
-CONFIG_DMA_SHARED_BUFFER=y
-CONFIG_DMA_VIRTUAL_CHANNELS=y
-CONFIG_DMI=y
-CONFIG_DMIID=y
-# CONFIG_DMI_SYSFS is not set
-CONFIG_DNOTIFY=y
-CONFIG_DNS_RESOLVER=y
-CONFIG_DST_CACHE=y
-CONFIG_DTC=y
-CONFIG_DT_IDLE_STATES=y
-CONFIG_DUMMY_CONSOLE=y
-CONFIG_E1000=y
-CONFIG_E1000E=y
-CONFIG_EDAC_SUPPORT=y
-CONFIG_EEPROM_AT24=y
-CONFIG_EFI=y
-CONFIG_EFIVAR_FS=y
-CONFIG_EFI_ARMSTUB=y
-# CONFIG_EFI_CAPSULE_LOADER is not set
-CONFIG_EFI_ESRT=y
-CONFIG_EFI_PARAMS_FROM_FDT=y
-CONFIG_EFI_RUNTIME_WRAPPERS=y
-CONFIG_EFI_STUB=y
-# CONFIG_EFI_TEST is not set
-# CONFIG_EFI_VARS is not set
-CONFIG_ELF_CORE=y
-# CONFIG_EMBEDDED is not set
-CONFIG_ENABLE_MUST_CHECK=y
-# CONFIG_EVM is not set
-CONFIG_EXPORTFS=y
-CONFIG_EXT2_FS=y
-CONFIG_EXT2_FS_POSIX_ACL=y
-# CONFIG_EXT2_FS_SECURITY is not set
-CONFIG_EXT2_FS_XATTR=y
-CONFIG_EXT3_FS=y
-CONFIG_EXT3_FS_POSIX_ACL=y
-# CONFIG_EXT3_FS_SECURITY is not set
-CONFIG_EXT4_FS=y
-CONFIG_EXT4_FS_POSIX_ACL=y
-CONFIG_EXT4_FS_SECURITY=y
-CONFIG_EXTCON=y
-CONFIG_EXTCON_USB_GPIO=y
-CONFIG_FAIR_GROUP_SCHED=y
-CONFIG_FANOTIFY=y
-CONFIG_FANOTIFY_ACCESS_PERMISSIONS=y
-CONFIG_FAT_FS=y
-CONFIG_FB=y
-CONFIG_FB_ARMCLCD=y
-CONFIG_FB_CFB_COPYAREA=y
-CONFIG_FB_CFB_FILLRECT=y
-CONFIG_FB_CFB_IMAGEBLIT=y
-CONFIG_FB_CMDLINE=y
-CONFIG_FB_DEFERRED_IO=y
-# CONFIG_FB_EFI is not set
-CONFIG_FB_MODE_HELPERS=y
-CONFIG_FB_SYS_COPYAREA=y
-CONFIG_FB_SYS_FILLRECT=y
-CONFIG_FB_SYS_FOPS=y
-CONFIG_FB_SYS_IMAGEBLIT=y
-CONFIG_FHANDLE=y
-CONFIG_FIRMWARE_IN_KERNEL=y
-CONFIG_FIXED_PHY=y
-CONFIG_FIX_EARLYCON_MEM=y
-CONFIG_FMAN_ARM=y
-# CONFIG_FMAN_MIB_CNT_OVF_IRQ_EN is not set
-# CONFIG_FMAN_P1023 is not set
-# CONFIG_FMAN_P3040_P4080_P5020 is not set
-# CONFIG_FMAN_PFC is not set
-# CONFIG_FMAN_V3H is not set
-# CONFIG_FMAN_V3L is not set
-# CONFIG_FONTS is not set
-CONFIG_FONT_8x16=y
-CONFIG_FONT_8x8=y
-CONFIG_FONT_SUPPORT=y
-CONFIG_FRAMEBUFFER_CONSOLE=y
-CONFIG_FRAMEBUFFER_CONSOLE_DETECT_PRIMARY=y
-# CONFIG_FRAMEBUFFER_CONSOLE_ROTATION is not set
-CONFIG_FRAME_POINTER=y
-CONFIG_FRAME_WARN=2048
-CONFIG_FREEZER=y
-CONFIG_FSL_BMAN_CONFIG=y
-CONFIG_FSL_BMAN_DEBUGFS=y
-# CONFIG_FSL_BMAN_TEST is not set
-CONFIG_FSL_DPAA2=y
-CONFIG_FSL_DPAA2_ETH=y
-CONFIG_FSL_DPAA2_ETHSW=y
-# CONFIG_FSL_DPAA2_ETH_CEETM is not set
-# CONFIG_FSL_DPAA2_ETH_DEBUGFS is not set
-# CONFIG_FSL_DPAA2_ETH_USE_ERR_QUEUE is not set
-CONFIG_FSL_DPAA2_EVB=y
-CONFIG_FSL_DPAA2_MAC=y
-# CONFIG_FSL_DPAA2_MAC_NETDEVS is not set
-# CONFIG_FSL_DPAA2_QDMA is not set
-# CONFIG_FSL_DPAA_1588 is not set
-CONFIG_FSL_DPAA_ADVANCED_DRIVERS=y
-# CONFIG_FSL_DPAA_CEETM is not set
-CONFIG_FSL_DPAA_CS_THRESHOLD_10G=0x10000000
-CONFIG_FSL_DPAA_CS_THRESHOLD_1G=0x06000000
-# CONFIG_FSL_DPAA_DBG_LOOP is not set
-# CONFIG_FSL_DPAA_ETH_DEBUG is not set
-CONFIG_FSL_DPAA_ETH_DEBUGFS=y
-# CONFIG_FSL_DPAA_ETH_JUMBO_FRAME is not set
-CONFIG_FSL_DPAA_ETH_MAX_BUF_COUNT=128
-CONFIG_FSL_DPAA_ETH_REFILL_THRESHOLD=80
-CONFIG_FSL_DPAA_ETH_USE_NDO_SELECT_QUEUE=y
-# CONFIG_FSL_DPAA_HOOKS is not set
-CONFIG_FSL_DPAA_INGRESS_CS_THRESHOLD=0x10000000
-CONFIG_FSL_DPAA_OFFLINE_PORTS=y
-# CONFIG_FSL_DPAA_TS is not set
-CONFIG_FSL_DPA_CAN_WAIT=y
-CONFIG_FSL_DPA_CAN_WAIT_SYNC=y
-# CONFIG_FSL_DPA_CHECKING is not set
-CONFIG_FSL_DPA_PIRQ_FAST=y
-CONFIG_FSL_DPA_PIRQ_SLOW=y
-CONFIG_FSL_DPA_PORTAL_SHARE=y
-CONFIG_FSL_EDMA=y
-CONFIG_FSL_ERRATUM_A008585=y
-CONFIG_FSL_FM_MAX_FRAME_SIZE=1522
-CONFIG_FSL_FM_RX_EXTRA_HEADROOM=64
-CONFIG_FSL_GUTS=y
-CONFIG_FSL_IFC=y
-CONFIG_FSL_LS2_CONSOLE=y
-CONFIG_FSL_MC_BUS=y
-CONFIG_FSL_MC_DPIO=y
-CONFIG_FSL_MC_RESTOOL=y
-CONFIG_FSL_PPFE=y
-CONFIG_FSL_PPFE_UTIL_DISABLED=y
-# CONFIG_FSL_QBMAN_DEBUG is not set
-# CONFIG_FSL_QDMA is not set
-CONFIG_FSL_QMAN_CI_SCHED_CFG_BMAN_W=2
-CONFIG_FSL_QMAN_CI_SCHED_CFG_RW_W=2
-CONFIG_FSL_QMAN_CI_SCHED_CFG_SRCCIV=4
-CONFIG_FSL_QMAN_CI_SCHED_CFG_SRQ_W=3
-CONFIG_FSL_QMAN_CONFIG=y
-CONFIG_FSL_QMAN_DEBUGFS=y
-CONFIG_FSL_QMAN_FQD_SZ=10
-CONFIG_FSL_QMAN_FQ_LOOKUP=y
-CONFIG_FSL_QMAN_INIT_TIMEOUT=10
-CONFIG_FSL_QMAN_PFDR_SZ=13
-CONFIG_FSL_QMAN_PIRQ_DQRR_ITHRESH=12
-CONFIG_FSL_QMAN_PIRQ_IPERIOD=100
-CONFIG_FSL_QMAN_PIRQ_MR_ITHRESH=4
-CONFIG_FSL_QMAN_POLL_LIMIT=32
-# CONFIG_FSL_QMAN_TEST is not set
-CONFIG_FSL_SDK_BMAN=y
-CONFIG_FSL_SDK_DPA=y
-CONFIG_FSL_SDK_DPAA_ETH=y
-CONFIG_FSL_SDK_FMAN=y
-# CONFIG_FSL_SDK_FMAN_TEST is not set
-CONFIG_FSL_SDK_QMAN=y
-CONFIG_FSL_USDPAA=y
-CONFIG_FSL_XGMAC_MDIO=y
-CONFIG_FS_IOMAP=y
-CONFIG_FS_MBCACHE=y
-CONFIG_FS_POSIX_ACL=y
-CONFIG_FTM_ALARM=y
-CONFIG_FUSE_FS=y
-# CONFIG_FW_LOADER_USER_HELPER_FALLBACK is not set
-CONFIG_GARP=y
-CONFIG_GENERIC_ALLOCATOR=y
-CONFIG_GENERIC_BUG=y
-CONFIG_GENERIC_BUG_RELATIVE_POINTERS=y
-CONFIG_GENERIC_CLOCKEVENTS=y
-CONFIG_GENERIC_CLOCKEVENTS_BROADCAST=y
-CONFIG_GENERIC_CPU_AUTOPROBE=y
-CONFIG_GENERIC_CSUM=y
-CONFIG_GENERIC_EARLY_IOREMAP=y
-CONFIG_GENERIC_IDLE_POLL_SETUP=y
-CONFIG_GENERIC_IO=y
-CONFIG_GENERIC_IRQ_CHIP=y
-# CONFIG_GENERIC_IRQ_DEBUGFS is not set
-CONFIG_GENERIC_IRQ_MIGRATION=y
-CONFIG_GENERIC_IRQ_SHOW=y
-CONFIG_GENERIC_IRQ_SHOW_LEVEL=y
-CONFIG_GENERIC_MSI_IRQ=y
-CONFIG_GENERIC_MSI_IRQ_DOMAIN=y
-CONFIG_GENERIC_PCI_IOMAP=y
-CONFIG_GENERIC_PHY=y
-CONFIG_GENERIC_SCHED_CLOCK=y
-CONFIG_GENERIC_SMP_IDLE_THREAD=y
-CONFIG_GENERIC_STRNCPY_FROM_USER=y
-CONFIG_GENERIC_STRNLEN_USER=y
-CONFIG_GENERIC_TIME_VSYSCALL=y
-# CONFIG_GIANFAR is not set
-CONFIG_GLOB=y
-CONFIG_GPIOLIB=y
-CONFIG_GPIOLIB_IRQCHIP=y
-CONFIG_GPIO_ACPI=y
-CONFIG_GPIO_DWAPB=y
-CONFIG_GPIO_GENERIC=y
-CONFIG_GPIO_GENERIC_PLATFORM=y
-CONFIG_GPIO_MAX77620=y
-CONFIG_GPIO_MPC8XXX=y
-CONFIG_GPIO_PCA953X=y
-CONFIG_GPIO_PCA953X_IRQ=y
-CONFIG_GPIO_PL061=y
-CONFIG_GPIO_SYSFS=y
-CONFIG_GPIO_XGENE=y
-CONFIG_HANDLE_DOMAIN_IRQ=y
-CONFIG_HARDEN_BRANCH_PREDICTOR=y
-CONFIG_HARDIRQS_SW_RESEND=y
-CONFIG_HAS_DMA=y
-CONFIG_HAS_IOMEM=y
-CONFIG_HAS_IOPORT_MAP=y
-# CONFIG_HAVE_64BIT_ALIGNED_ACCESS is not set
-CONFIG_HAVE_ARCH_AUDITSYSCALL=y
-CONFIG_HAVE_ARCH_BITREVERSE=y
-CONFIG_HAVE_ARCH_HUGE_VMAP=y
-CONFIG_HAVE_ARCH_JUMP_LABEL=y
-CONFIG_HAVE_ARCH_KASAN=y
-CONFIG_HAVE_ARCH_KGDB=y
-CONFIG_HAVE_ARCH_MMAP_RND_COMPAT_BITS=y
-CONFIG_HAVE_ARCH_PFN_VALID=y
-CONFIG_HAVE_ARCH_SECCOMP_FILTER=y
-CONFIG_HAVE_ARCH_TRACEHOOK=y
-CONFIG_HAVE_ARCH_TRANSPARENT_HUGEPAGE=y
-CONFIG_HAVE_ARM_SMCCC=y
-# CONFIG_HAVE_BOOTMEM_INFO_NODE is not set
-CONFIG_HAVE_CC_STACKPROTECTOR=y
-CONFIG_HAVE_CLK=y
-CONFIG_HAVE_CLK_PREPARE=y
-CONFIG_HAVE_CMPXCHG_DOUBLE=y
-CONFIG_HAVE_CMPXCHG_LOCAL=y
-CONFIG_HAVE_CONTEXT_TRACKING=y
-CONFIG_HAVE_C_RECORDMCOUNT=y
-CONFIG_HAVE_DEBUG_BUGVERBOSE=y
-CONFIG_HAVE_DEBUG_KMEMLEAK=y
-CONFIG_HAVE_DMA_API_DEBUG=y
-CONFIG_HAVE_DMA_CONTIGUOUS=y
-CONFIG_HAVE_DYNAMIC_FTRACE=y
-CONFIG_HAVE_EBPF_JIT=y
-CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS=y
-CONFIG_HAVE_FTRACE_MCOUNT_RECORD=y
-CONFIG_HAVE_FUNCTION_GRAPH_TRACER=y
-CONFIG_HAVE_FUNCTION_TRACER=y
-CONFIG_HAVE_GENERIC_DMA_COHERENT=y
-CONFIG_HAVE_GENERIC_RCU_GUP=y
-CONFIG_HAVE_HW_BREAKPOINT=y
-CONFIG_HAVE_IRQ_TIME_ACCOUNTING=y
-CONFIG_HAVE_KVM_ARCH_TLB_FLUSH_ALL=y
-CONFIG_HAVE_KVM_CPU_RELAX_INTERCEPT=y
-CONFIG_HAVE_KVM_EVENTFD=y
-CONFIG_HAVE_KVM_IRQCHIP=y
-CONFIG_HAVE_KVM_IRQFD=y
-CONFIG_HAVE_KVM_IRQ_ROUTING=y
-CONFIG_HAVE_KVM_MSI=y
-CONFIG_HAVE_MEMBLOCK=y
-CONFIG_HAVE_MEMBLOCK_NODE_MAP=y
-CONFIG_HAVE_MEMORY_PRESENT=y
-CONFIG_HAVE_NET_DSA=y
-CONFIG_HAVE_PATA_PLATFORM=y
-CONFIG_HAVE_PERF_EVENTS=y
-CONFIG_HAVE_PERF_REGS=y
-CONFIG_HAVE_PERF_USER_STACK_DUMP=y
-CONFIG_HAVE_RCU_TABLE_FREE=y
-CONFIG_HAVE_REGS_AND_STACK_ACCESS_API=y
-CONFIG_HAVE_SETUP_PER_CPU_AREA=y
-CONFIG_HAVE_SYSCALL_TRACEPOINTS=y
-CONFIG_HAVE_UID16=y
-CONFIG_HAVE_VIRT_CPU_ACCOUNTING_GEN=y
-CONFIG_HIBERNATE_CALLBACKS=y
-CONFIG_HIBERNATION=y
-CONFIG_HID=y
-CONFIG_HID_A4TECH=y
-CONFIG_HID_APPLE=y
-CONFIG_HID_BELKIN=y
-CONFIG_HID_CHERRY=y
-CONFIG_HID_CHICONY=y
-CONFIG_HID_CYPRESS=y
-CONFIG_HID_EZKEY=y
-CONFIG_HID_GENERIC=y
-CONFIG_HID_KENSINGTON=y
-CONFIG_HID_LOGITECH=y
-CONFIG_HID_MICROSOFT=y
-CONFIG_HID_MONTEREY=y
-CONFIG_HNS=y
-CONFIG_HNS_DSAF=y
-CONFIG_HNS_ENET=y
-CONFIG_HNS_MDIO=y
-CONFIG_HOTPLUG_CPU=y
-# CONFIG_HPET is not set
-CONFIG_HUGETLBFS=y
-CONFIG_HUGETLB_PAGE=y
-CONFIG_HVC_DRIVER=y
-CONFIG_HVC_IRQ=y
-CONFIG_HVC_XEN=y
-CONFIG_HVC_XEN_FRONTEND=y
-CONFIG_HWMON=y
-CONFIG_HW_CONSOLE=y
-CONFIG_HW_RANDOM=y
-# CONFIG_HW_RANDOM_CAVIUM is not set
-CONFIG_HZ=250
-# CONFIG_HZ_100 is not set
-CONFIG_HZ_250=y
-CONFIG_I2C=y
-CONFIG_I2C_ALGOBIT=y
-CONFIG_I2C_BOARDINFO=y
-CONFIG_I2C_CHARDEV=y
-CONFIG_I2C_COMPAT=y
-CONFIG_I2C_CROS_EC_TUNNEL=y
-CONFIG_I2C_DESIGNWARE_CORE=y
-CONFIG_I2C_DESIGNWARE_PLATFORM=y
-CONFIG_I2C_HELPER_AUTO=y
-CONFIG_I2C_IMX=y
-CONFIG_I2C_MUX=y
-CONFIG_I2C_MUX_PCA954x=y
-CONFIG_I2C_RK3X=y
-CONFIG_I2C_SLAVE=y
-# CONFIG_I2C_SLAVE_EEPROM is not set
-CONFIG_IGB=y
-CONFIG_IGBVF=y
-CONFIG_IGB_HWMON=y
-CONFIG_IIO=y
-# CONFIG_IIO_BUFFER is not set
-# CONFIG_IIO_TRIGGER is not set
-CONFIG_IKCONFIG=y
-CONFIG_IKCONFIG_PROC=y
-CONFIG_ILLEGAL_POINTER_VALUE=0xdead000000000000
-# CONFIG_IMA is not set
-CONFIG_IMX2_WDT=y
-CONFIG_INET_DIAG=y
-# CONFIG_INET_DIAG_DESTROY is not set
-CONFIG_INET_ESP=y
-CONFIG_INET_TCP_DIAG=y
-CONFIG_INET_TUNNEL=y
-CONFIG_INET_XFRM_MODE_BEET=y
-CONFIG_INET_XFRM_MODE_TRANSPORT=y
-CONFIG_INET_XFRM_MODE_TUNNEL=y
-CONFIG_INITRAMFS_SOURCE=""
-CONFIG_INPUT=y
-CONFIG_INPUT_EVDEV=y
-CONFIG_INPUT_KEYBOARD=y
-CONFIG_INPUT_LEDS=y
-CONFIG_INPUT_MOUSE=y
-CONFIG_INPUT_MOUSEDEV=y
-CONFIG_INPUT_MOUSEDEV_PSAUX=y
-CONFIG_INPUT_MOUSEDEV_SCREEN_X=1024
-CONFIG_INPUT_MOUSEDEV_SCREEN_Y=768
-CONFIG_INPUT_XEN_KBDDEV_FRONTEND=y
-CONFIG_INTEGRITY=y
-CONFIG_INTEGRITY_AUDIT=y
-# CONFIG_INTEGRITY_SIGNATURE is not set
-CONFIG_IOMMU_API=y
-CONFIG_IOMMU_DMA=y
-CONFIG_IOMMU_HELPER=y
-CONFIG_IOMMU_IOVA=y
-CONFIG_IOMMU_IO_PGTABLE=y
-# CONFIG_IOMMU_IO_PGTABLE_ARMV7S is not set
-CONFIG_IOMMU_IO_PGTABLE_LPAE=y
-# CONFIG_IOMMU_IO_PGTABLE_LPAE_SELFTEST is not set
-CONFIG_IOMMU_SUPPORT=y
-CONFIG_IOSCHED_CFQ=y
-# CONFIG_IOSCHED_DEADLINE is not set
-CONFIG_IPC_NS=y
-CONFIG_IPV6=y
-CONFIG_IPV6_SIT=y
-# CONFIG_IP_ADVANCED_ROUTER is not set
-CONFIG_IP_PNP=y
-CONFIG_IP_PNP_BOOTP=y
-CONFIG_IP_PNP_DHCP=y
-# CONFIG_IP_PNP_RARP is not set
-CONFIG_IRQCHIP=y
-CONFIG_IRQ_BYPASS_MANAGER=y
-CONFIG_IRQ_DOMAIN=y
-CONFIG_IRQ_DOMAIN_HIERARCHY=y
-CONFIG_IRQ_FORCED_THREADING=y
-CONFIG_IRQ_WORK=y
-# CONFIG_ISDN is not set
-CONFIG_JBD2=y
-CONFIG_JUMP_LABEL=y
-CONFIG_KALLSYMS=y
-CONFIG_KALLSYMS_ALL=y
-CONFIG_KEXEC=y
-CONFIG_KEXEC_CORE=y
-CONFIG_KEYBOARD_ATKBD=y
-# CONFIG_KEYBOARD_CROS_EC is not set
-CONFIG_KEYBOARD_GPIO=y
-CONFIG_KEYS=y
-CONFIG_KEYS_COMPAT=y
-CONFIG_KSM=y
-CONFIG_KVM=y
-CONFIG_KVM_ARM_HOST=y
-CONFIG_KVM_ARM_PMU=y
-CONFIG_KVM_ARM_VGIC_V3_ITS=y
-CONFIG_KVM_COMPAT=y
-CONFIG_KVM_GENERIC_DIRTYLOG_READ_PROTECT=y
-CONFIG_KVM_MMIO=y
-CONFIG_KVM_VFIO=y
-# CONFIG_LCD_CLASS_DEVICE is not set
-CONFIG_LEDS_GPIO=y
-CONFIG_LEDS_PWM=y
-CONFIG_LEDS_SYSCON=y
-CONFIG_LEDS_TRIGGER_CPU=y
-CONFIG_LEDS_TRIGGER_HEARTBEAT=y
-# CONFIG_LEDS_TRIGGER_TIMER is not set
-CONFIG_LEGACY_PTYS=y
-CONFIG_LEGACY_PTY_COUNT=16
-CONFIG_LIBCRC32C=y
-CONFIG_LIBFDT=y
-CONFIG_LOCALVERSION_AUTO=y
-CONFIG_LOCKUP_DETECTOR=y
-CONFIG_LOCK_SPIN_ON_OWNER=y
-CONFIG_LOGO=y
-CONFIG_LOGO_LINUX_CLUT224=y
-# CONFIG_LOGO_LINUX_MONO is not set
-# CONFIG_LOGO_LINUX_VGA16 is not set
-CONFIG_LS_SCFG_MSI=y
-CONFIG_LS_SOC_DRIVERS=y
-CONFIG_LZO_COMPRESS=y
-CONFIG_LZO_DECOMPRESS=y
-CONFIG_MACB=y
-CONFIG_MACVLAN=y
-CONFIG_MACVTAP=y
-CONFIG_MAGIC_SYSRQ=y
-CONFIG_MANDATORY_FILE_LOCKING=y
-# CONFIG_MAX77620_THERMAL is not set
-# CONFIG_MAX77620_WATCHDOG is not set
-CONFIG_MDIO_BITBANG=y
-CONFIG_MDIO_BOARDINFO=y
-CONFIG_MDIO_BUS_MUX=y
-CONFIG_MDIO_BUS_MUX_MMIOREG=y
-# CONFIG_MDIO_FSL_BACKPLANE is not set
-# CONFIG_MDIO_GPIO is not set
-CONFIG_MEMCG=y
-CONFIG_MEMCG_SWAP=y
-CONFIG_MEMCG_SWAP_ENABLED=y
-CONFIG_MEMORY=y
-CONFIG_MEMORY_BALLOON=y
-CONFIG_MEMORY_ISOLATION=y
-CONFIG_MEMTEST=y
-CONFIG_MFD_CORE=y
-CONFIG_MFD_CROS_EC=y
-CONFIG_MFD_CROS_EC_I2C=y
-# CONFIG_MFD_CROS_EC_SPI is not set
-CONFIG_MFD_MAX77620=y
-CONFIG_MFD_RK808=y
-CONFIG_MFD_SEC_CORE=y
-CONFIG_MFD_SYSCON=y
-CONFIG_MFD_VEXPRESS_SYSREG=y
-CONFIG_MICREL_PHY=y
-CONFIG_MIGRATION=y
-CONFIG_MMC=y
-CONFIG_MMC_ARMMMCI=y
-CONFIG_MMC_BLOCK=y
-CONFIG_MMC_BLOCK_MINORS=32
-CONFIG_MMC_DW=y
-CONFIG_MMC_DW_EXYNOS=y
-CONFIG_MMC_DW_K3=y
-# CONFIG_MMC_DW_PCI is not set
-CONFIG_MMC_DW_PLTFM=y
-CONFIG_MMC_SDHCI=y
-CONFIG_MMC_SDHCI_ACPI=y
-CONFIG_MMC_SDHCI_IO_ACCESSORS=y
-CONFIG_MMC_SDHCI_OF_ARASAN=y
-CONFIG_MMC_SDHCI_OF_ESDHC=y
-# CONFIG_MMC_SDHCI_PCI is not set
-CONFIG_MMC_SDHCI_PLTFM=y
-CONFIG_MMC_SPI=y
-# CONFIG_MMC_TIFM_SD is not set
-CONFIG_MMU_NOTIFIER=y
-CONFIG_MODULES_TREE_LOOKUP=y
-CONFIG_MODULES_USE_ELF_RELA=y
-CONFIG_MODULE_FORCE_LOAD=y
-CONFIG_MODVERSIONS=y
-# CONFIG_MOUSE_BCM5974 is not set
-# CONFIG_MOUSE_CYAPA is not set
-CONFIG_MOUSE_PS2=y
-CONFIG_MOUSE_PS2_ALPS=y
-CONFIG_MOUSE_PS2_BYD=y
-CONFIG_MOUSE_PS2_CYPRESS=y
-# CONFIG_MOUSE_PS2_ELANTECH is not set
-CONFIG_MOUSE_PS2_FOCALTECH=y
-CONFIG_MOUSE_PS2_LOGIPS2PP=y
-CONFIG_MOUSE_PS2_SYNAPTICS=y
-# CONFIG_MOUSE_PS2_TOUCHKIT is not set
-CONFIG_MOUSE_PS2_TRACKPOINT=y
-# CONFIG_MOUSE_SERIAL is not set
-# CONFIG_MOUSE_VSXXXAA is not set
-CONFIG_MPILIB=y
-CONFIG_MRP=y
-CONFIG_MTD_CFI_ADV_OPTIONS=y
-# CONFIG_MTD_CFI_GEOMETRY is not set
-CONFIG_MTD_CFI_STAA=y
-CONFIG_MTD_CMDLINE_PARTS=y
-# CONFIG_MTD_COMPLEX_MAPPINGS is not set
-CONFIG_MTD_DATAFLASH=y
-# CONFIG_MTD_DATAFLASH_OTP is not set
-# CONFIG_MTD_DATAFLASH_WRITE_VERIFY is not set
-CONFIG_MTD_M25P80=y
-CONFIG_MTD_NAND=y
-CONFIG_MTD_NAND_DENALI=y
-CONFIG_MTD_NAND_DENALI_DT=y
-CONFIG_MTD_NAND_ECC=y
-CONFIG_MTD_NAND_FSL_IFC=y
-CONFIG_MTD_SPI_NOR=y
-CONFIG_MTD_SPI_NOR_USE_4K_SECTORS=y
-CONFIG_MTD_SPI_NOR_USE_4K_SECTORS_LIMIT=16384
-CONFIG_MTD_SST25L=y
-CONFIG_MTD_UBI=y
-CONFIG_MTD_UBI_BEB_LIMIT=20
-# CONFIG_MTD_UBI_BLOCK is not set
-# CONFIG_MTD_UBI_FASTMAP is not set
-# CONFIG_MTD_UBI_GLUEBI is not set
-CONFIG_MTD_UBI_WL_THRESHOLD=4096
-CONFIG_MUTEX_SPIN_ON_OWNER=y
-CONFIG_MV_XOR_V2=y
-CONFIG_NAMESPACES=y
-CONFIG_NEED_DMA_MAP_STATE=y
-CONFIG_NEED_MULTIPLE_NODES=y
-CONFIG_NEED_PER_CPU_EMBED_FIRST_CHUNK=y
-CONFIG_NEED_SG_DMA_LENGTH=y
-# CONFIG_NETLABEL is not set
-# CONFIG_NET_CLS_CGROUP is not set
-CONFIG_NET_FLOW_LIMIT=y
-CONFIG_NET_IP_TUNNEL=y
-CONFIG_NET_NS=y
-CONFIG_NET_PTP_CLASSIFY=y
-CONFIG_NET_SWITCHDEV=y
-CONFIG_NLS=y
-CONFIG_NLS_CODEPAGE_437=y
-CONFIG_NLS_ISO8859_1=y
-CONFIG_NODES_SHIFT=2
-CONFIG_NO_BOOTMEM=y
-CONFIG_NO_HZ_COMMON=y
-CONFIG_NO_HZ_IDLE=y
-CONFIG_NR_CPUS=64
-CONFIG_NUMA=y
-CONFIG_NUMA_BALANCING=y
-CONFIG_NUMA_BALANCING_DEFAULT_ENABLED=y
-CONFIG_NVMEM=y
-CONFIG_NVME_CORE=y
-CONFIG_OF=y
-CONFIG_OF_ADDRESS=y
-CONFIG_OF_ADDRESS_PCI=y
-CONFIG_OF_EARLY_FLATTREE=y
-CONFIG_OF_FLATTREE=y
-CONFIG_OF_GPIO=y
-CONFIG_OF_IOMMU=y
-CONFIG_OF_IRQ=y
-CONFIG_OF_MDIO=y
-CONFIG_OF_NET=y
-CONFIG_OF_NUMA=y
-CONFIG_OF_PCI=y
-CONFIG_OF_PCI_IRQ=y
-CONFIG_OF_RESERVED_MEM=y
-CONFIG_OLD_SIGSUSPEND3=y
-CONFIG_PACKET_DIAG=y
-CONFIG_PAGE_COUNTER=y
-# CONFIG_PANIC_ON_OOPS is not set
-CONFIG_PANIC_ON_OOPS_VALUE=0
-CONFIG_PANIC_TIMEOUT=0
-CONFIG_PARAVIRT=y
-# CONFIG_PARTITION_ADVANCED is not set
-CONFIG_PARTITION_PERCPU=y
-# CONFIG_PATA_ACPI is not set
-CONFIG_PATA_OF_PLATFORM=y
-CONFIG_PATA_PLATFORM=y
-CONFIG_PCI=y
-CONFIG_PCIEAER=y
-CONFIG_PCIEASPM=y
-# CONFIG_PCIEASPM_DEBUG is not set
-CONFIG_PCIEASPM_DEFAULT=y
-# CONFIG_PCIEASPM_PERFORMANCE is not set
-# CONFIG_PCIEASPM_POWERSAVE is not set
-CONFIG_PCIEPORTBUS=y
-CONFIG_PCIE_DW=y
-CONFIG_PCIE_PME=y
-CONFIG_PCI_ATS=y
-CONFIG_PCI_BUS_ADDR_T_64BIT=y
-CONFIG_PCI_DOMAINS=y
-CONFIG_PCI_DOMAINS_GENERIC=y
-CONFIG_PCI_ECAM=y
-CONFIG_PCI_HISI=y
-CONFIG_PCI_HOST_COMMON=y
-CONFIG_PCI_HOST_GENERIC=y
-CONFIG_PCI_IOV=y
-CONFIG_PCI_LABEL=y
-CONFIG_PCI_LAYERSCAPE=y
-CONFIG_PCI_MSI=y
-CONFIG_PCI_MSI_IRQ_DOMAIN=y
-CONFIG_PERF_EVENTS=y
-CONFIG_PGTABLE_LEVELS=4
-CONFIG_PHYLIB=y
-CONFIG_PHYS_ADDR_T_64BIT=y
-CONFIG_PHY_XGENE=y
-CONFIG_PID_IN_CONTEXTIDR=y
-CONFIG_PID_NS=y
-CONFIG_PL330_DMA=y
-CONFIG_PM=y
-# CONFIG_PMIC_OPREGION is not set
-CONFIG_PM_CLK=y
-# CONFIG_PM_DEBUG is not set
-CONFIG_PM_OPP=y
-CONFIG_PM_SLEEP=y
-CONFIG_PM_SLEEP_SMP=y
-CONFIG_PM_STD_PARTITION=""
-CONFIG_PNP=y
-CONFIG_PNPACPI=y
-CONFIG_PNP_DEBUG_MESSAGES=y
-CONFIG_POSIX_MQUEUE=y
-CONFIG_POSIX_MQUEUE_SYSCTL=y
-CONFIG_POWER_RESET=y
-CONFIG_POWER_RESET_SYSCON=y
-CONFIG_POWER_RESET_VEXPRESS=y
-CONFIG_POWER_RESET_XGENE=y
-CONFIG_POWER_SUPPLY=y
-CONFIG_PPS=y
-CONFIG_PREEMPT=y
-CONFIG_PREEMPT_COUNT=y
-# CONFIG_PREEMPT_NONE is not set
-CONFIG_PREEMPT_NOTIFIERS=y
-CONFIG_PREEMPT_RCU=y
-CONFIG_PRINTK_TIME=y
-CONFIG_PRINT_QUOTA_WARNING=y
-CONFIG_PROC_CHILDREN=y
-CONFIG_PROC_PID_CPUSET=y
-CONFIG_PROFILING=y
-CONFIG_PTP_1588_CLOCK=y
-# CONFIG_PTP_1588_CLOCK_DPAA is not set
-CONFIG_PTP_1588_CLOCK_DPAA2=y
-CONFIG_PWM=y
-# CONFIG_PWM_CROS_EC is not set
-CONFIG_PWM_SYSFS=y
-CONFIG_QCOM_HIDMA=y
-CONFIG_QCOM_HIDMA_MGMT=y
-CONFIG_QCOM_QDF2400_ERRATUM_0065=y
-# CONFIG_QFMT_V1 is not set
-# CONFIG_QFMT_V2 is not set
-CONFIG_QMAN_CEETM_UPDATE_PERIOD=1000
-CONFIG_QORIQ_CPUFREQ=y
-CONFIG_QORIQ_THERMAL=y
-# CONFIG_QUICC_ENGINE is not set
-CONFIG_QUOTA=y
-CONFIG_QUOTACTL=y
-# CONFIG_QUOTA_NETLINK_INTERFACE is not set
-CONFIG_RADIX_TREE_MULTIORDER=y
-CONFIG_RAID6_PQ=y
-# CONFIG_RANDOMIZE_BASE is not set
-CONFIG_RAS=y
-CONFIG_RATIONAL=y
-CONFIG_RCU_CPU_STALL_TIMEOUT=21
-# CONFIG_RCU_EXPERT is not set
-CONFIG_RCU_STALL_COMMON=y
-CONFIG_RD_BZIP2=y
-CONFIG_RD_GZIP=y
-CONFIG_RD_LZMA=y
-CONFIG_RD_LZO=y
-CONFIG_RD_XZ=y
-CONFIG_REALTEK_PHY=y
-CONFIG_REGMAP=y
-CONFIG_REGMAP_I2C=y
-CONFIG_REGMAP_IRQ=y
-CONFIG_REGMAP_MMIO=y
-CONFIG_REGMAP_SPI=y
-CONFIG_REGULATOR=y
-CONFIG_REGULATOR_FIXED_VOLTAGE=y
-CONFIG_REGULATOR_GPIO=y
-CONFIG_REGULATOR_MAX77620=y
-CONFIG_REGULATOR_PWM=y
-CONFIG_REGULATOR_QCOM_SPMI=y
-CONFIG_REGULATOR_RK808=y
-# CONFIG_REGULATOR_S2MPA01 is not set
-CONFIG_REGULATOR_S2MPS11=y
-# CONFIG_REGULATOR_S5M8767 is not set
-# CONFIG_REGULATOR_VEXPRESS is not set
-CONFIG_RESET_CONTROLLER=y
-CONFIG_RFS_ACCEL=y
-CONFIG_RPS=y
-CONFIG_RTC_CLASS=y
-CONFIG_RTC_DRV_DS1307=y
-CONFIG_RTC_DRV_DS1307_HWMON=y
-CONFIG_RTC_DRV_DS3232=y
-CONFIG_RTC_DRV_EFI=y
-CONFIG_RTC_DRV_MAX77686=y
-CONFIG_RTC_DRV_PCF2127=y
-CONFIG_RTC_DRV_PCF85263=y
-CONFIG_RTC_DRV_PL031=y
-# CONFIG_RTC_DRV_RK808 is not set
-CONFIG_RTC_DRV_S5M=y
-CONFIG_RTC_I2C_AND_SPI=y
-# CONFIG_RT_GROUP_SCHED is not set
-CONFIG_RWSEM_SPIN_ON_OWNER=y
-CONFIG_RWSEM_XCHGADD_ALGORITHM=y
-CONFIG_SATA_AHCI=y
-CONFIG_SATA_AHCI_PLATFORM=y
-CONFIG_SATA_PMP=y
-CONFIG_SATA_SIL24=y
-# CONFIG_SATA_ZPODD is not set
-CONFIG_SCHED_AUTOGROUP=y
-CONFIG_SCHED_INFO=y
-CONFIG_SCHED_MC=y
-CONFIG_SCSI=y
-CONFIG_SCSI_HISI_SAS=y
-# CONFIG_SCSI_PROC_FS is not set
-CONFIG_SCSI_SAS_ATA=y
-CONFIG_SCSI_SAS_ATTRS=y
-CONFIG_SCSI_SAS_HOST_SMP=y
-CONFIG_SCSI_SAS_LIBSAS=y
-CONFIG_SECCOMP=y
-CONFIG_SECCOMP_FILTER=y
-CONFIG_SECURITY=y
-# CONFIG_SECURITY_APPARMOR is not set
-# CONFIG_SECURITY_LOADPIN is not set
-# CONFIG_SECURITY_NETWORK is not set
-# CONFIG_SECURITY_PATH is not set
-# CONFIG_SECURITY_SMACK is not set
-# CONFIG_SECURITY_TOMOYO is not set
-# CONFIG_SECURITY_YAMA is not set
-CONFIG_SERIAL_8250_DEPRECATED_OPTIONS=y
-CONFIG_SERIAL_8250_DW=y
-CONFIG_SERIAL_8250_EXTENDED=y
-CONFIG_SERIAL_8250_FSL=y
-CONFIG_SERIAL_8250_NR_UARTS=4
-CONFIG_SERIAL_8250_PCI=y
-CONFIG_SERIAL_8250_PNP=y
-CONFIG_SERIAL_8250_RUNTIME_UARTS=4
-CONFIG_SERIAL_8250_SHARE_IRQ=y
-CONFIG_SERIAL_AMBA_PL011=y
-CONFIG_SERIAL_AMBA_PL011_CONSOLE=y
-CONFIG_SERIAL_FSL_LPUART=y
-CONFIG_SERIAL_FSL_LPUART_CONSOLE=y
-CONFIG_SERIAL_OF_PLATFORM=y
-CONFIG_SERIAL_XILINX_PS_UART=y
-CONFIG_SERIAL_XILINX_PS_UART_CONSOLE=y
-CONFIG_SERIO=y
-CONFIG_SERIO_AMBAKMI=y
-CONFIG_SERIO_LIBPS2=y
-CONFIG_SG_POOL=y
-CONFIG_SKY2=y
-CONFIG_SLAB=y
-# CONFIG_SLUB is not set
-CONFIG_SMC91X=y
-CONFIG_SMP=y
-CONFIG_SMSC911X=y
-# CONFIG_SMSC911X_ARCH_HOOKS is not set
-CONFIG_SND=y
-# CONFIG_SND_COMPRESS_OFFLOAD is not set
-CONFIG_SND_DMAENGINE_PCM=y
-CONFIG_SND_JACK=y
-CONFIG_SND_JACK_INPUT_DEV=y
-CONFIG_SND_PCM=y
-CONFIG_SND_PCM_TIMER=y
-CONFIG_SND_SIMPLE_CARD=y
-CONFIG_SND_SIMPLE_CARD_UTILS=y
-CONFIG_SND_SOC=y
-CONFIG_SND_SOC_AK4613=y
-CONFIG_SND_SOC_FSL_SAI=y
-CONFIG_SND_SOC_GENERIC_DMAENGINE_PCM=y
-CONFIG_SND_SOC_I2C_AND_SPI=y
-CONFIG_SND_SOC_SGTL5000=y
-CONFIG_SND_SPI=y
-CONFIG_SND_SUPPORT_OLD_API=y
-CONFIG_SND_TIMER=y
-CONFIG_SOCK_DIAG=y
-CONFIG_SOC_BUS=y
-CONFIG_SOUND=y
-CONFIG_SPARSEMEM=y
-CONFIG_SPARSEMEM_EXTREME=y
-CONFIG_SPARSEMEM_MANUAL=y
-CONFIG_SPARSEMEM_VMEMMAP=y
-CONFIG_SPARSEMEM_VMEMMAP_ENABLE=y
-CONFIG_SPARSE_IRQ=y
-CONFIG_SPI=y
-CONFIG_SPI_FSL_DSPI=y
-CONFIG_SPI_FSL_QUADSPI=y
-CONFIG_SPI_MASTER=y
-CONFIG_SPI_PL022=y
-CONFIG_SPMI=y
-# CONFIG_SQUASHFS_DECOMP_MULTI_PERCPU is not set
-CONFIG_SQUASHFS_DECOMP_SINGLE=y
-CONFIG_SQUASHFS_FILE_CACHE=y
-# CONFIG_SQUASHFS_FILE_DIRECT is not set
-# CONFIG_SQUASHFS_XZ is not set
-CONFIG_SQUASHFS_ZLIB=y
-CONFIG_SRAM=y
-CONFIG_SRCU=y
-# CONFIG_STRIP_ASM_SYMS is not set
-CONFIG_SUSPEND=y
-CONFIG_SUSPEND_FREEZER=y
-CONFIG_SWIOTLB=y
-CONFIG_SWIOTLB_XEN=y
-CONFIG_SWPHY=y
-# CONFIG_SW_SYNC is not set
-CONFIG_SYNC_FILE=y
-# CONFIG_SYN_COOKIES is not set
-CONFIG_SYSCTL_EXCEPTION_TRACE=y
-CONFIG_SYSFS_SYSCALL=y
-CONFIG_SYSVIPC_COMPAT=y
-CONFIG_SYS_HYPERVISOR=y
-CONFIG_SYS_SUPPORTS_HUGETLBFS=y
-CONFIG_TASKSTATS=y
-CONFIG_TASK_DELAY_ACCT=y
-CONFIG_TASK_IO_ACCOUNTING=y
-CONFIG_TASK_XACCT=y
-# CONFIG_TCP_CONG_ADVANCED is not set
-CONFIG_THERMAL=y
-CONFIG_THERMAL_DEFAULT_GOV_STEP_WISE=y
-CONFIG_THERMAL_EMULATION=y
-CONFIG_THERMAL_GOV_POWER_ALLOCATOR=y
-CONFIG_THERMAL_GOV_STEP_WISE=y
-CONFIG_THERMAL_HWMON=y
-CONFIG_THERMAL_OF=y
-CONFIG_TICK_CPU_ACCOUNTING=y
-CONFIG_TMPFS_POSIX_ACL=y
-CONFIG_TRANSPARENT_HUGEPAGE=y
-CONFIG_TRANSPARENT_HUGEPAGE_ALWAYS=y
-# CONFIG_TRANSPARENT_HUGEPAGE_MADVISE is not set
-CONFIG_TRANSPARENT_HUGE_PAGECACHE=y
-CONFIG_TUN=y
-CONFIG_UBIFS_FS=y
-# CONFIG_UBIFS_FS_ADVANCED_COMPR is not set
-CONFIG_UBIFS_FS_LZO=y
-CONFIG_UBIFS_FS_ZLIB=y
-CONFIG_UCS2_STRING=y
-CONFIG_UIO=y
-CONFIG_UIO_AEC=y
-CONFIG_UIO_CIF=y
-CONFIG_UIO_DMEM_GENIRQ=y
-CONFIG_UIO_MF624=y
-CONFIG_UIO_NETX=y
-CONFIG_UIO_PCI_GENERIC=y
-CONFIG_UIO_PDRV_GENIRQ=y
-# CONFIG_UIO_PRUSS is not set
-CONFIG_UIO_SERCOS3=y
-CONFIG_UNINLINE_SPIN_UNLOCK=y
-CONFIG_UNIX_DIAG=y
-CONFIG_UNMAP_KERNEL_AT_EL0=y
-CONFIG_USB_SUPPORT=y
-# CONFIG_USERIO is not set
-CONFIG_USER_NS=y
-CONFIG_USE_PERCPU_NUMA_NODE_ID=y
-CONFIG_UTS_NS=y
-CONFIG_VEXPRESS_CONFIG=y
-CONFIG_VEXPRESS_SYSCFG=y
-CONFIG_VFAT_FS=y
-CONFIG_VFIO=y
-CONFIG_VFIO_FSL_MC=y
-CONFIG_VFIO_IOMMU_TYPE1=y
-# CONFIG_VFIO_NOIOMMU is not set
-CONFIG_VFIO_PCI=y
-CONFIG_VFIO_PCI_INTX=y
-CONFIG_VFIO_PCI_MMAP=y
-# CONFIG_VFIO_PLATFORM is not set
-CONFIG_VFIO_VIRQFD=y
-CONFIG_VGA_ARB=y
-CONFIG_VGA_ARB_MAX_GPUS=16
-CONFIG_VHOST=y
-CONFIG_VHOST_NET=y
-CONFIG_VIDEOMODE_HELPERS=y
-CONFIG_VIRTIO=y
-CONFIG_VIRTIO_BALLOON=y
-CONFIG_VIRTIO_BLK=y
-CONFIG_VIRTIO_CONSOLE=y
-CONFIG_VIRTIO_MMIO=y
-# CONFIG_VIRTIO_MMIO_CMDLINE_DEVICES is not set
-CONFIG_VIRTIO_NET=y
-CONFIG_VIRTIO_PCI=y
-CONFIG_VIRTIO_PCI_LEGACY=y
-CONFIG_VIRTUALIZATION=y
-CONFIG_VITESSE_PHY=y
-CONFIG_VLAN_8021Q_GVRP=y
-CONFIG_VLAN_8021Q_MVRP=y
-CONFIG_VM_EVENT_COUNTERS=y
-CONFIG_VT=y
-CONFIG_VT_CONSOLE=y
-CONFIG_VT_CONSOLE_SLEEP=y
-CONFIG_VT_HW_CONSOLE_BINDING=y
-CONFIG_WATCHDOG_CORE=y
-# CONFIG_WQ_POWER_EFFICIENT_DEFAULT is not set
-CONFIG_XEN=y
-CONFIG_XENFS=y
-CONFIG_XEN_AUTO_XLATE=y
-CONFIG_XEN_BACKEND=y
-CONFIG_XEN_BALLOON=y
-# CONFIG_XEN_BLKDEV_BACKEND is not set
-CONFIG_XEN_BLKDEV_FRONTEND=y
-CONFIG_XEN_COMPAT_XENFS=y
-CONFIG_XEN_DEV_EVTCHN=y
-CONFIG_XEN_DOM0=y
-CONFIG_XEN_EFI=y
-CONFIG_XEN_FBDEV_FRONTEND=y
-CONFIG_XEN_GNTDEV=y
-CONFIG_XEN_GRANT_DEV_ALLOC=y
-# CONFIG_XEN_NETDEV_BACKEND is not set
-CONFIG_XEN_NETDEV_FRONTEND=y
-CONFIG_XEN_PRIVCMD=y
-CONFIG_XEN_SCRUB_PAGES=y
-# CONFIG_XEN_SCSI_FRONTEND is not set
-CONFIG_XEN_SYS_HYPERVISOR=y
-# CONFIG_XEN_WDT is not set
-CONFIG_XEN_XENBUS_FRONTEND=y
-CONFIG_XFRM_ALGO=y
-CONFIG_XFRM_USER=y
-CONFIG_XFS_FS=y
-CONFIG_XFS_POSIX_ACL=y
-CONFIG_XFS_RT=y
-CONFIG_XOR_BLOCKS=y
-CONFIG_XPS=y
-CONFIG_XZ_DEC_ARM=y
-CONFIG_XZ_DEC_ARMTHUMB=y
-CONFIG_XZ_DEC_BCJ=y
-CONFIG_XZ_DEC_IA64=y
-CONFIG_XZ_DEC_POWERPC=y
-CONFIG_XZ_DEC_SPARC=y
-CONFIG_XZ_DEC_X86=y
-CONFIG_ZLIB_DEFLATE=y
-CONFIG_ZLIB_INFLATE=y
diff --git a/target/linux/layerscape/patches-4.9/201-config-support-layerscape.patch b/target/linux/layerscape/patches-4.9/201-config-support-layerscape.patch
deleted file mode 100644
index ea3cc702d9..0000000000
--- a/target/linux/layerscape/patches-4.9/201-config-support-layerscape.patch
+++ /dev/null
@@ -1,374 +0,0 @@
-From 0774b97305507af18f8c43efb69aa00e6c57ae90 Mon Sep 17 00:00:00 2001
-From: Yangbo Lu <yangbo.lu@nxp.com>
-Date: Fri, 6 Jul 2018 15:31:14 +0800
-Subject: [PATCH] config: support layerscape
-MIME-Version: 1.0
-Content-Type: text/plain; charset=UTF-8
-Content-Transfer-Encoding: 8bit
-
-This is an integrated patch for layerscape config/makefile support.
-
-Signed-off-by: Yuantian Tang <andy.tang@nxp.com>
-Signed-off-by: Zhang Ying-22455 <ying.zhang22455@nxp.com>
-Signed-off-by: Zhao Qiang <qiang.zhao@nxp.com>
-Signed-off-by: Bogdan Purcareata <bogdan.purcareata@nxp.com>
-Signed-off-by: Zhao Qiang <qiang.zhao@nxp.com>
-Signed-off-by: Horia Geantă <horia.geanta@nxp.com>
-Signed-off-by: Yangbo Lu <yangbo.lu@nxp.com>
----
- arch/arm/mach-imx/Kconfig | 1 +
- drivers/base/Kconfig | 1 +
- drivers/crypto/Makefile | 2 +-
- drivers/net/ethernet/freescale/Kconfig | 4 +-
- drivers/net/ethernet/freescale/Makefile | 2 +
- drivers/ptp/Kconfig | 29 +++++++++++
- drivers/rtc/Kconfig | 8 +++
- drivers/rtc/Makefile | 1 +
- drivers/soc/Kconfig | 3 +-
- drivers/soc/fsl/Kconfig | 30 ++++++++++++
- drivers/soc/fsl/Kconfig.arm | 16 ++++++
- drivers/soc/fsl/Makefile | 5 ++
- drivers/soc/fsl/layerscape/Kconfig | 10 ++++
- drivers/soc/fsl/layerscape/Makefile | 1 +
- drivers/staging/Kconfig | 6 +++
- drivers/staging/Makefile | 3 ++
- drivers/staging/fsl-dpaa2/Kconfig | 65 +++++++++++++++++++++++++
- drivers/staging/fsl-dpaa2/Makefile | 9 ++++
- 18 files changed, 192 insertions(+), 4 deletions(-)
- create mode 100644 drivers/soc/fsl/Kconfig
- create mode 100644 drivers/soc/fsl/Kconfig.arm
- create mode 100644 drivers/soc/fsl/layerscape/Kconfig
- create mode 100644 drivers/soc/fsl/layerscape/Makefile
- create mode 100644 drivers/staging/fsl-dpaa2/Kconfig
- create mode 100644 drivers/staging/fsl-dpaa2/Makefile
-
---- a/arch/arm/mach-imx/Kconfig
-+++ b/arch/arm/mach-imx/Kconfig
-@@ -1,6 +1,7 @@
- menuconfig ARCH_MXC
- bool "Freescale i.MX family"
- depends on ARCH_MULTI_V4_V5 || ARCH_MULTI_V6_V7 || ARM_SINGLE_ARMV7M
-+ select ARCH_DMA_ADDR_T_64BIT if ARM_LPAE
- select ARCH_SUPPORTS_BIG_ENDIAN
- select CLKSRC_IMX_GPT
- select GENERIC_IRQ_CHIP
---- a/drivers/base/Kconfig
-+++ b/drivers/base/Kconfig
-@@ -240,6 +240,7 @@ config GENERIC_CPU_VULNERABILITIES
-
- config SOC_BUS
- bool
-+ select GLOB
-
- source "drivers/base/regmap/Kconfig"
-
---- a/drivers/crypto/Makefile
-+++ b/drivers/crypto/Makefile
-@@ -3,7 +3,7 @@ obj-$(CONFIG_CRYPTO_DEV_ATMEL_SHA) += at
- obj-$(CONFIG_CRYPTO_DEV_ATMEL_TDES) += atmel-tdes.o
- obj-$(CONFIG_CRYPTO_DEV_BFIN_CRC) += bfin_crc.o
- obj-$(CONFIG_CRYPTO_DEV_CCP) += ccp/
--obj-$(CONFIG_CRYPTO_DEV_FSL_CAAM) += caam/
-+obj-$(CONFIG_CRYPTO_DEV_FSL_CAAM_COMMON) += caam/
- obj-$(CONFIG_CRYPTO_DEV_GEODE) += geode-aes.o
- obj-$(CONFIG_CRYPTO_DEV_HIFN_795X) += hifn_795x.o
- obj-$(CONFIG_CRYPTO_DEV_IMGTEC_HASH) += img-hash.o
---- a/drivers/net/ethernet/freescale/Kconfig
-+++ b/drivers/net/ethernet/freescale/Kconfig
-@@ -5,7 +5,7 @@
- config NET_VENDOR_FREESCALE
- bool "Freescale devices"
- default y
-- depends on FSL_SOC || QUICC_ENGINE || CPM1 || CPM2 || PPC_MPC512x || \
-+ depends on FSL_SOC || (QUICC_ENGINE && PPC32) || CPM1 || CPM2 || PPC_MPC512x || \
- M523x || M527x || M5272 || M528x || M520x || M532x || \
- ARCH_MXC || ARCH_MXS || (PPC_MPC52xx && PPC_BESTCOMM) || \
- ARCH_LAYERSCAPE
-@@ -93,4 +93,6 @@ config GIANFAR
- and MPC86xx family of chips, the eTSEC on LS1021A and the FEC
- on the 8540.
-
-+source "drivers/net/ethernet/freescale/sdk_fman/Kconfig"
-+source "drivers/net/ethernet/freescale/sdk_dpaa/Kconfig"
- endif # NET_VENDOR_FREESCALE
---- a/drivers/net/ethernet/freescale/Makefile
-+++ b/drivers/net/ethernet/freescale/Makefile
-@@ -21,4 +21,6 @@ gianfar_driver-objs := gianfar.o \
- obj-$(CONFIG_UCC_GETH) += ucc_geth_driver.o
- ucc_geth_driver-objs := ucc_geth.o ucc_geth_ethtool.o
-
-+obj-$(if $(CONFIG_FSL_SDK_FMAN),y) += sdk_fman/
-+obj-$(if $(CONFIG_FSL_SDK_DPAA_ETH),y) += sdk_dpaa/
- obj-$(CONFIG_FSL_FMAN) += fman/
---- a/drivers/ptp/Kconfig
-+++ b/drivers/ptp/Kconfig
-@@ -39,6 +39,35 @@ config PTP_1588_CLOCK_GIANFAR
- To compile this driver as a module, choose M here: the module
- will be called gianfar_ptp.
-
-+config PTP_1588_CLOCK_DPAA
-+ tristate "Freescale DPAA as PTP clock"
-+ depends on FSL_SDK_DPAA_ETH
-+ select PTP_1588_CLOCK
-+ select FSL_DPAA_TS
-+ default n
-+ help
-+ This driver adds support for using the DPAA 1588 timer module
-+ as a PTP clock. This clock is only useful if your PTP programs are
-+ getting hardware time stamps on the PTP Ethernet packets
-+ using the SO_TIMESTAMPING API.
-+
-+ To compile this driver as a module, choose M here: the module
-+ will be called dpaa_ptp.
-+
-+config PTP_1588_CLOCK_DPAA2
-+ tristate "Freescale DPAA2 as PTP clock"
-+ depends on FSL_DPAA2_ETH
-+ select PTP_1588_CLOCK
-+ default y
-+ help
-+ This driver adds support for using the DPAA2 1588 timer module
-+ as a PTP clock. This clock is only useful if your PTP programs are
-+ getting hardware time stamps on the PTP Ethernet packets
-+ using the SO_TIMESTAMPING API.
-+
-+ To compile this driver as a module, choose M here: the module
-+ will be called dpaa2-rtc.
-+
- config PTP_1588_CLOCK_IXP46X
- tristate "Intel IXP46x as PTP clock"
- depends on IXP4XX_ETH
---- a/drivers/rtc/Kconfig
-+++ b/drivers/rtc/Kconfig
-@@ -414,6 +414,14 @@ config RTC_DRV_PCF85063
- This driver can also be built as a module. If so, the module
- will be called rtc-pcf85063.
-
-+config RTC_DRV_PCF85263
-+ tristate "NXP PCF85263"
-+ help
-+ If you say yes here you get support for the PCF85263 RTC chip
-+
-+ This driver can also be built as a module. If so, the module
-+ will be called rtc-pcf85263.
-+
- config RTC_DRV_PCF8563
- tristate "Philips PCF8563/Epson RTC8564"
- help
---- a/drivers/rtc/Makefile
-+++ b/drivers/rtc/Makefile
-@@ -111,6 +111,7 @@ obj-$(CONFIG_RTC_DRV_PCF2127) += rtc-pcf
- obj-$(CONFIG_RTC_DRV_PCF50633) += rtc-pcf50633.o
- obj-$(CONFIG_RTC_DRV_PCF85063) += rtc-pcf85063.o
- obj-$(CONFIG_RTC_DRV_PCF8523) += rtc-pcf8523.o
-+obj-$(CONFIG_RTC_DRV_PCF85263) += rtc-pcf85263.o
- obj-$(CONFIG_RTC_DRV_PCF8563) += rtc-pcf8563.o
- obj-$(CONFIG_RTC_DRV_PCF8583) += rtc-pcf8583.o
- obj-$(CONFIG_RTC_DRV_PIC32) += rtc-pic32.o
---- a/drivers/soc/Kconfig
-+++ b/drivers/soc/Kconfig
-@@ -1,8 +1,7 @@
- menu "SOC (System On Chip) specific Drivers"
-
- source "drivers/soc/bcm/Kconfig"
--source "drivers/soc/fsl/qbman/Kconfig"
--source "drivers/soc/fsl/qe/Kconfig"
-+source "drivers/soc/fsl/Kconfig"
- source "drivers/soc/mediatek/Kconfig"
- source "drivers/soc/qcom/Kconfig"
- source "drivers/soc/rockchip/Kconfig"
---- /dev/null
-+++ b/drivers/soc/fsl/Kconfig
-@@ -0,0 +1,30 @@
-+#
-+# Freescale SOC drivers
-+#
-+
-+source "drivers/soc/fsl/qbman/Kconfig"
-+source "drivers/soc/fsl/qe/Kconfig"
-+source "drivers/soc/fsl/ls2-console/Kconfig"
-+
-+config FSL_GUTS
-+ bool
-+ select SOC_BUS
-+ help
-+ The global utilities block controls power management, I/O device
-+ enabling, power-onreset(POR) configuration monitoring, alternate
-+ function selection for multiplexed signals,and clock control.
-+ This driver is to manage and access global utilities block.
-+ Initially only reading SVR and registering soc device are supported.
-+ Other guts accesses, such as reading RCW, should eventually be moved
-+ into this driver as well.
-+
-+config FSL_SLEEP_FSM
-+ bool
-+ help
-+ This driver configures a hardware FSM (Finite State Machine) for deep sleep.
-+ The FSM is used to finish clean-ups at the last stage of system entering deep
-+ sleep, and also wakes up system when a wake up event happens.
-+
-+if ARM || ARM64
-+source "drivers/soc/fsl/Kconfig.arm"
-+endif
---- /dev/null
-+++ b/drivers/soc/fsl/Kconfig.arm
-@@ -0,0 +1,16 @@
-+#
-+# Freescale ARM SOC Drivers
-+#
-+
-+config LS_SOC_DRIVERS
-+ bool "Layerscape Soc Drivers"
-+ depends on ARCH_LAYERSCAPE || SOC_LS1021A
-+ default n
-+ help
-+ Say y here to enable Freescale Layerscape Soc Device Drivers support.
-+ The Soc Drivers provides the device driver that is a specific block
-+ or feature on Layerscape platform.
-+
-+if LS_SOC_DRIVERS
-+ source "drivers/soc/fsl/layerscape/Kconfig"
-+endif
---- a/drivers/soc/fsl/Makefile
-+++ b/drivers/soc/fsl/Makefile
-@@ -5,3 +5,8 @@
- obj-$(CONFIG_FSL_DPAA) += qbman/
- obj-$(CONFIG_QUICC_ENGINE) += qe/
- obj-$(CONFIG_CPM) += qe/
-+obj-$(CONFIG_FSL_GUTS) += guts.o
-+obj-$(CONFIG_FSL_LS2_CONSOLE) += ls2-console/
-+obj-$(CONFIG_SUSPEND) += rcpm.o
-+obj-$(CONFIG_LS_SOC_DRIVERS) += layerscape/
-+obj-$(CONFIG_FSL_SLEEP_FSM) += sleep_fsm.o
---- /dev/null
-+++ b/drivers/soc/fsl/layerscape/Kconfig
-@@ -0,0 +1,10 @@
-+#
-+# Layerscape Soc drivers
-+#
-+config FTM_ALARM
-+ bool "FTM alarm driver"
-+ default n
-+ help
-+ Say y here to enable FTM alarm support. The FTM alarm provides
-+ alarm functions for wakeup system from deep sleep. There is only
-+ one FTM can be used in ALARM(FTM 0).
---- /dev/null
-+++ b/drivers/soc/fsl/layerscape/Makefile
-@@ -0,0 +1 @@
-+obj-$(CONFIG_FTM_ALARM) += ftm_alarm.o
---- a/drivers/staging/Kconfig
-+++ b/drivers/staging/Kconfig
-@@ -94,6 +94,8 @@ source "drivers/staging/fbtft/Kconfig"
-
- source "drivers/staging/fsl-mc/Kconfig"
-
-+source "drivers/staging/fsl-dpaa2/Kconfig"
-+
- source "drivers/staging/wilc1000/Kconfig"
-
- source "drivers/staging/most/Kconfig"
-@@ -106,4 +108,8 @@ source "drivers/staging/greybus/Kconfig"
-
- source "drivers/staging/vc04_services/Kconfig"
-
-+source "drivers/staging/fsl_qbman/Kconfig"
-+
-+source "drivers/staging/fsl_ppfe/Kconfig"
-+
- endif # STAGING
---- a/drivers/staging/Makefile
-+++ b/drivers/staging/Makefile
-@@ -36,9 +36,12 @@ obj-$(CONFIG_UNISYSSPAR) += unisys/
- obj-$(CONFIG_COMMON_CLK_XLNX_CLKWZRD) += clocking-wizard/
- obj-$(CONFIG_FB_TFT) += fbtft/
- obj-$(CONFIG_FSL_MC_BUS) += fsl-mc/
-+obj-$(CONFIG_FSL_DPAA2) += fsl-dpaa2/
- obj-$(CONFIG_WILC1000) += wilc1000/
- obj-$(CONFIG_MOST) += most/
- obj-$(CONFIG_ISDN_I4L) += i4l/
- obj-$(CONFIG_KS7010) += ks7010/
- obj-$(CONFIG_GREYBUS) += greybus/
- obj-$(CONFIG_BCM2708_VCHIQ) += vc04_services/
-+obj-$(CONFIG_FSL_SDK_DPA) += fsl_qbman/
-+obj-$(CONFIG_FSL_PPFE) += fsl_ppfe/
---- /dev/null
-+++ b/drivers/staging/fsl-dpaa2/Kconfig
-@@ -0,0 +1,65 @@
-+#
-+# Freescale DataPath Acceleration Architecture Gen2 (DPAA2) drivers
-+#
-+
-+config FSL_DPAA2
-+ bool "Freescale DPAA2 devices"
-+ depends on FSL_MC_BUS
-+ ---help---
-+ Build drivers for Freescale DataPath Acceleration
-+ Architecture (DPAA2) family of SoCs.
-+
-+config FSL_DPAA2_ETH
-+ tristate "Freescale DPAA2 Ethernet"
-+ depends on FSL_DPAA2 && FSL_MC_DPIO
-+ ---help---
-+ Ethernet driver for Freescale DPAA2 SoCs, using the
-+ Freescale MC bus driver
-+
-+if FSL_DPAA2_ETH
-+config FSL_DPAA2_ETH_USE_ERR_QUEUE
-+ bool "Enable Rx error queue"
-+ default n
-+ ---help---
-+ Allow Rx error frames to be enqueued on an error queue
-+ and processed by the driver (by default they are dropped
-+ in hardware).
-+ This may impact performance, recommended for debugging
-+ purposes only.
-+
-+# QBMAN_DEBUG requires some additional DPIO APIs
-+config FSL_DPAA2_ETH_DEBUGFS
-+ depends on DEBUG_FS
-+ bool "Enable debugfs support"
-+ default n
-+ ---help---
-+ Enable advanced statistics through debugfs interface.
-+
-+config FSL_DPAA2_ETH_DCB
-+ bool "Data Center Bridging (DCB) Support"
-+ default n
-+ depends on DCB
-+ ---help---
-+ Say Y here if you want to use Data Center Bridging (DCB) features
-+ (PFC) in the driver.
-+
-+ If unsure, say N.
-+
-+config FSL_DPAA2_ETH_CEETM
-+ depends on NET_SCHED
-+ bool "DPAA2 Ethernet CEETM QoS"
-+ default n
-+ ---help---
-+ Enable QoS offloading support through the CEETM hardware block.
-+endif
-+
-+source "drivers/staging/fsl-dpaa2/mac/Kconfig"
-+source "drivers/staging/fsl-dpaa2/evb/Kconfig"
-+
-+config FSL_DPAA2_ETHSW
-+ tristate "Freescale DPAA2 Ethernet Switch"
-+ depends on FSL_DPAA2
-+ depends on NET_SWITCHDEV
-+ ---help---
-+ Driver for Freescale DPAA2 Ethernet Switch. Select
-+ BRIDGE to have support for bridge tools.
---- /dev/null
-+++ b/drivers/staging/fsl-dpaa2/Makefile
-@@ -0,0 +1,9 @@
-+#
-+# Freescale DataPath Acceleration Architecture Gen2 (DPAA2) drivers
-+#
-+
-+obj-$(CONFIG_FSL_DPAA2_ETH) += ethernet/
-+obj-$(CONFIG_FSL_DPAA2_MAC) += mac/
-+obj-$(CONFIG_FSL_DPAA2_EVB) += evb/
-+obj-$(CONFIG_PTP_1588_CLOCK_DPAA2) += rtc/
-+obj-$(CONFIG_FSL_DPAA2_ETHSW) += ethsw/
diff --git a/target/linux/layerscape/patches-4.9/202-core-linux-support-layerscape.patch b/target/linux/layerscape/patches-4.9/202-core-linux-support-layerscape.patch
deleted file mode 100644
index 1e33604f6a..0000000000
--- a/target/linux/layerscape/patches-4.9/202-core-linux-support-layerscape.patch
+++ /dev/null
@@ -1,1458 +0,0 @@
-From f339945a8e81fff22df95284e142b79c37fd2333 Mon Sep 17 00:00:00 2001
-From: Yangbo Lu <yangbo.lu@nxp.com>
-Date: Thu, 5 Jul 2018 16:07:09 +0800
-Subject: [PATCH 02/32] core-linux: support layerscape
-
-This is an integrated patch for layerscape core-linux support.
-
-Signed-off-by: Madalin Bucur <madalin.bucur@freescale.com>
-Signed-off-by: Zhao Qiang <qiang.zhao@nxp.com>
-Signed-off-by: Camelia Groza <camelia.groza@nxp.com>
-Signed-off-by: Madalin Bucur <madalin.bucur@nxp.com>
-Signed-off-by: Zhang Ying-22455 <ying.zhang22455@nxp.com>
-Signed-off-by: Ramneek Mehresh <ramneek.mehresh@freescale.com>
-Signed-off-by: Jarod Wilson <jarod@redhat.com>
-Signed-off-by: Nikhil Badola <nikhil.badola@freescale.com>
-Signed-off-by: stephen hemminger <stephen@networkplumber.org>
-Signed-off-by: Arnd Bergmann <arnd@arndb.de>
-Signed-off-by: Yangbo Lu <yangbo.lu@nxp.com>
----
- drivers/base/devres.c | 66 ++++++
- drivers/base/soc.c | 70 ++++++
- .../net/ethernet/mellanox/mlxsw/spectrum.c | 2 +-
- .../mellanox/mlxsw/spectrum_switchdev.c | 2 +-
- drivers/net/ethernet/rocker/rocker_ofdpa.c | 4 +-
- include/linux/device.h | 19 ++
- include/linux/dma-mapping.h | 5 +
- include/linux/fsl/svr.h | 97 ++++++++
- include/linux/fsl_devices.h | 3 +
- include/linux/irqdesc.h | 4 +
- include/linux/irqdomain.h | 13 +-
- include/linux/netdev_features.h | 2 +
- include/linux/netdevice.h | 10 +-
- include/linux/skbuff.h | 2 +
- include/linux/sys_soc.h | 3 +
- include/net/switchdev.h | 8 +-
- include/uapi/linux/if_ether.h | 1 +
- kernel/irq/Kconfig | 11 +
- kernel/irq/Makefile | 1 +
- kernel/irq/debugfs.c | 215 ++++++++++++++++++
- kernel/irq/internals.h | 22 ++
- kernel/irq/irqdesc.c | 1 +
- kernel/irq/irqdomain.c | 171 ++++++++++----
- kernel/irq/manage.c | 1 +
- kernel/irq/msi.c | 2 +-
- net/bridge/br.c | 4 +-
- net/bridge/br_fdb.c | 2 +
- net/bridge/br_private.h | 7 +
- net/bridge/br_switchdev.c | 33 +++
- net/core/dev.c | 30 ++-
- net/core/net-sysfs.c | 20 +-
- net/core/rtnetlink.c | 4 +-
- net/core/skbuff.c | 29 ++-
- net/sched/sch_generic.c | 7 +
- 34 files changed, 809 insertions(+), 62 deletions(-)
- create mode 100644 include/linux/fsl/svr.h
- create mode 100644 kernel/irq/debugfs.c
-
---- a/drivers/base/devres.c
-+++ b/drivers/base/devres.c
-@@ -10,6 +10,7 @@
- #include <linux/device.h>
- #include <linux/module.h>
- #include <linux/slab.h>
-+#include <linux/percpu.h>
-
- #include "base.h"
-
-@@ -985,3 +986,68 @@ void devm_free_pages(struct device *dev,
- &devres));
- }
- EXPORT_SYMBOL_GPL(devm_free_pages);
-+
-+static void devm_percpu_release(struct device *dev, void *pdata)
-+{
-+ void __percpu *p;
-+
-+ p = *(void __percpu **)pdata;
-+ free_percpu(p);
-+}
-+
-+static int devm_percpu_match(struct device *dev, void *data, void *p)
-+{
-+ struct devres *devr = container_of(data, struct devres, data);
-+
-+ return *(void **)devr->data == p;
-+}
-+
-+/**
-+ * __devm_alloc_percpu - Resource-managed alloc_percpu
-+ * @dev: Device to allocate per-cpu memory for
-+ * @size: Size of per-cpu memory to allocate
-+ * @align: Alignment of per-cpu memory to allocate
-+ *
-+ * Managed alloc_percpu. Per-cpu memory allocated with this function is
-+ * automatically freed on driver detach.
-+ *
-+ * RETURNS:
-+ * Pointer to allocated memory on success, NULL on failure.
-+ */
-+void __percpu *__devm_alloc_percpu(struct device *dev, size_t size,
-+ size_t align)
-+{
-+ void *p;
-+ void __percpu *pcpu;
-+
-+ pcpu = __alloc_percpu(size, align);
-+ if (!pcpu)
-+ return NULL;
-+
-+ p = devres_alloc(devm_percpu_release, sizeof(void *), GFP_KERNEL);
-+ if (!p) {
-+ free_percpu(pcpu);
-+ return NULL;
-+ }
-+
-+ *(void __percpu **)p = pcpu;
-+
-+ devres_add(dev, p);
-+
-+ return pcpu;
-+}
-+EXPORT_SYMBOL_GPL(__devm_alloc_percpu);
-+
-+/**
-+ * devm_free_percpu - Resource-managed free_percpu
-+ * @dev: Device this memory belongs to
-+ * @pdata: Per-cpu memory to free
-+ *
-+ * Free memory allocated with devm_alloc_percpu().
-+ */
-+void devm_free_percpu(struct device *dev, void __percpu *pdata)
-+{
-+ WARN_ON(devres_destroy(dev, devm_percpu_release, devm_percpu_match,
-+ (void *)pdata));
-+}
-+EXPORT_SYMBOL_GPL(devm_free_percpu);
---- a/drivers/base/soc.c
-+++ b/drivers/base/soc.c
-@@ -13,6 +13,7 @@
- #include <linux/spinlock.h>
- #include <linux/sys_soc.h>
- #include <linux/err.h>
-+#include <linux/glob.h>
-
- static DEFINE_IDA(soc_ida);
-
-@@ -159,3 +160,72 @@ static int __init soc_bus_register(void)
- return bus_register(&soc_bus_type);
- }
- core_initcall(soc_bus_register);
-+
-+static int soc_device_match_one(struct device *dev, void *arg)
-+{
-+ struct soc_device *soc_dev = container_of(dev, struct soc_device, dev);
-+ const struct soc_device_attribute *match = arg;
-+
-+ if (match->machine &&
-+ (!soc_dev->attr->machine ||
-+ !glob_match(match->machine, soc_dev->attr->machine)))
-+ return 0;
-+
-+ if (match->family &&
-+ (!soc_dev->attr->family ||
-+ !glob_match(match->family, soc_dev->attr->family)))
-+ return 0;
-+
-+ if (match->revision &&
-+ (!soc_dev->attr->revision ||
-+ !glob_match(match->revision, soc_dev->attr->revision)))
-+ return 0;
-+
-+ if (match->soc_id &&
-+ (!soc_dev->attr->soc_id ||
-+ !glob_match(match->soc_id, soc_dev->attr->soc_id)))
-+ return 0;
-+
-+ return 1;
-+}
-+
-+/*
-+ * soc_device_match - identify the SoC in the machine
-+ * @matches: zero-terminated array of possible matches
-+ *
-+ * returns the first matching entry of the argument array, or NULL
-+ * if none of them match.
-+ *
-+ * This function is meant as a helper in place of of_match_node()
-+ * in cases where either no device tree is available or the information
-+ * in a device node is insufficient to identify a particular variant
-+ * by its compatible strings or other properties. For new devices,
-+ * the DT binding should always provide unique compatible strings
-+ * that allow the use of of_match_node() instead.
-+ *
-+ * The calling function can use the .data entry of the
-+ * soc_device_attribute to pass a structure or function pointer for
-+ * each entry.
-+ */
-+const struct soc_device_attribute *soc_device_match(
-+ const struct soc_device_attribute *matches)
-+{
-+ int ret = 0;
-+
-+ if (!matches)
-+ return NULL;
-+
-+ while (!ret) {
-+ if (!(matches->machine || matches->family ||
-+ matches->revision || matches->soc_id))
-+ break;
-+ ret = bus_for_each_dev(&soc_bus_type, NULL, (void *)matches,
-+ soc_device_match_one);
-+ if (!ret)
-+ matches++;
-+ else
-+ return matches;
-+ }
-+ return NULL;
-+}
-+EXPORT_SYMBOL_GPL(soc_device_match);
---- a/drivers/net/ethernet/mellanox/mlxsw/spectrum.c
-+++ b/drivers/net/ethernet/mellanox/mlxsw/spectrum.c
-@@ -859,7 +859,7 @@ mlxsw_sp_port_get_sw_stats64(const struc
- return 0;
- }
-
--static bool mlxsw_sp_port_has_offload_stats(int attr_id)
-+static bool mlxsw_sp_port_has_offload_stats(const struct net_device *dev, int attr_id)
- {
- switch (attr_id) {
- case IFLA_OFFLOAD_XSTATS_CPU_HIT:
---- a/drivers/net/ethernet/mellanox/mlxsw/spectrum_switchdev.c
-+++ b/drivers/net/ethernet/mellanox/mlxsw/spectrum_switchdev.c
-@@ -1405,7 +1405,7 @@ static void mlxsw_sp_fdb_call_notifiers(
- if (learning_sync) {
- info.addr = mac;
- info.vid = vid;
-- notifier_type = adding ? SWITCHDEV_FDB_ADD : SWITCHDEV_FDB_DEL;
-+ notifier_type = adding ? SWITCHDEV_FDB_ADD_TO_BRIDGE : SWITCHDEV_FDB_DEL_TO_BRIDGE;
- call_switchdev_notifiers(notifier_type, dev, &info.info);
- }
- }
---- a/drivers/net/ethernet/rocker/rocker_ofdpa.c
-+++ b/drivers/net/ethernet/rocker/rocker_ofdpa.c
-@@ -1939,10 +1939,10 @@ static void ofdpa_port_fdb_learn_work(st
-
- rtnl_lock();
- if (learned && removing)
-- call_switchdev_notifiers(SWITCHDEV_FDB_DEL,
-+ call_switchdev_notifiers(SWITCHDEV_FDB_DEL_TO_BRIDGE,
- lw->ofdpa_port->dev, &info.info);
- else if (learned && !removing)
-- call_switchdev_notifiers(SWITCHDEV_FDB_ADD,
-+ call_switchdev_notifiers(SWITCHDEV_FDB_ADD_TO_BRIDGE,
- lw->ofdpa_port->dev, &info.info);
- rtnl_unlock();
-
---- a/include/linux/device.h
-+++ b/include/linux/device.h
-@@ -688,6 +688,25 @@ void __iomem *devm_ioremap_resource(stru
- int devm_add_action(struct device *dev, void (*action)(void *), void *data);
- void devm_remove_action(struct device *dev, void (*action)(void *), void *data);
-
-+/**
-+ * devm_alloc_percpu - Resource-managed alloc_percpu
-+ * @dev: Device to allocate per-cpu memory for
-+ * @type: Type to allocate per-cpu memory for
-+ *
-+ * Managed alloc_percpu. Per-cpu memory allocated with this function is
-+ * automatically freed on driver detach.
-+ *
-+ * RETURNS:
-+ * Pointer to allocated memory on success, NULL on failure.
-+ */
-+#define devm_alloc_percpu(dev, type) \
-+ ((typeof(type) __percpu *)__devm_alloc_percpu((dev), sizeof(type), \
-+ __alignof__(type)))
-+
-+void __percpu *__devm_alloc_percpu(struct device *dev, size_t size,
-+ size_t align);
-+void devm_free_percpu(struct device *dev, void __percpu *pdata);
-+
- static inline int devm_add_action_or_reset(struct device *dev,
- void (*action)(void *), void *data)
- {
---- a/include/linux/dma-mapping.h
-+++ b/include/linux/dma-mapping.h
-@@ -164,6 +164,11 @@ int dma_mmap_from_coherent(struct device
-
- #ifdef CONFIG_HAS_DMA
- #include <asm/dma-mapping.h>
-+static inline void set_dma_ops(struct device *dev,
-+ struct dma_map_ops *dma_ops)
-+{
-+ dev->archdata.dma_ops = dma_ops;
-+}
- #else
- /*
- * Define the dma api to allow compilation but not linking of
---- /dev/null
-+++ b/include/linux/fsl/svr.h
-@@ -0,0 +1,97 @@
-+/*
-+ * MPC85xx cpu type detection
-+ *
-+ * Copyright 2011-2012 Freescale Semiconductor, Inc.
-+ *
-+ * This is free software; you can redistribute it and/or modify
-+ * it under the terms of the GNU General Public License as published by
-+ * the Free Software Foundation; either version 2 of the License, or
-+ * (at your option) any later version.
-+ */
-+
-+#ifndef FSL_SVR_H
-+#define FSL_SVR_H
-+
-+#define SVR_REV(svr) ((svr) & 0xFF) /* SOC design resision */
-+#define SVR_MAJ(svr) (((svr) >> 4) & 0xF) /* Major revision field*/
-+#define SVR_MIN(svr) (((svr) >> 0) & 0xF) /* Minor revision field*/
-+
-+/* Some parts define SVR[0:23] as the SOC version */
-+#define SVR_SOC_VER(svr) (((svr) >> 8) & 0xFFF7FF) /* SOC Version fields */
-+
-+#define SVR_8533 0x803400
-+#define SVR_8535 0x803701
-+#define SVR_8536 0x803700
-+#define SVR_8540 0x803000
-+#define SVR_8541 0x807200
-+#define SVR_8543 0x803200
-+#define SVR_8544 0x803401
-+#define SVR_8545 0x803102
-+#define SVR_8547 0x803101
-+#define SVR_8548 0x803100
-+#define SVR_8555 0x807100
-+#define SVR_8560 0x807000
-+#define SVR_8567 0x807501
-+#define SVR_8568 0x807500
-+#define SVR_8569 0x808000
-+#define SVR_8572 0x80E000
-+#define SVR_P1010 0x80F100
-+#define SVR_P1011 0x80E500
-+#define SVR_P1012 0x80E501
-+#define SVR_P1013 0x80E700
-+#define SVR_P1014 0x80F101
-+#define SVR_P1017 0x80F700
-+#define SVR_P1020 0x80E400
-+#define SVR_P1021 0x80E401
-+#define SVR_P1022 0x80E600
-+#define SVR_P1023 0x80F600
-+#define SVR_P1024 0x80E402
-+#define SVR_P1025 0x80E403
-+#define SVR_P2010 0x80E300
-+#define SVR_P2020 0x80E200
-+#define SVR_P2040 0x821000
-+#define SVR_P2041 0x821001
-+#define SVR_P3041 0x821103
-+#define SVR_P4040 0x820100
-+#define SVR_P4080 0x820000
-+#define SVR_P5010 0x822100
-+#define SVR_P5020 0x822000
-+#define SVR_P5021 0X820500
-+#define SVR_P5040 0x820400
-+#define SVR_T4240 0x824000
-+#define SVR_T4120 0x824001
-+#define SVR_T4160 0x824100
-+#define SVR_T4080 0x824102
-+#define SVR_C291 0x850000
-+#define SVR_C292 0x850020
-+#define SVR_C293 0x850030
-+#define SVR_B4860 0X868000
-+#define SVR_G4860 0x868001
-+#define SVR_G4060 0x868003
-+#define SVR_B4440 0x868100
-+#define SVR_G4440 0x868101
-+#define SVR_B4420 0x868102
-+#define SVR_B4220 0x868103
-+#define SVR_T1040 0x852000
-+#define SVR_T1041 0x852001
-+#define SVR_T1042 0x852002
-+#define SVR_T1020 0x852100
-+#define SVR_T1021 0x852101
-+#define SVR_T1022 0x852102
-+#define SVR_T1023 0x854100
-+#define SVR_T1024 0x854000
-+#define SVR_T2080 0x853000
-+#define SVR_T2081 0x853100
-+
-+#define SVR_8610 0x80A000
-+#define SVR_8641 0x809000
-+#define SVR_8641D 0x809001
-+
-+#define SVR_9130 0x860001
-+#define SVR_9131 0x860000
-+#define SVR_9132 0x861000
-+#define SVR_9232 0x861400
-+
-+#define SVR_Unknown 0xFFFFFF
-+
-+#endif
---- a/include/linux/fsl_devices.h
-+++ b/include/linux/fsl_devices.h
-@@ -99,7 +99,10 @@ struct fsl_usb2_platform_data {
- unsigned suspended:1;
- unsigned already_suspended:1;
- unsigned has_fsl_erratum_a007792:1;
-+ unsigned has_fsl_erratum_14:1;
- unsigned has_fsl_erratum_a005275:1;
-+ unsigned has_fsl_erratum_a006918:1;
-+ unsigned has_fsl_erratum_a005697:1;
- unsigned check_phy_clk_valid:1;
-
- /* register save area for suspend/resume */
---- a/include/linux/irqdesc.h
-+++ b/include/linux/irqdesc.h
-@@ -46,6 +46,7 @@ struct pt_regs;
- * @rcu: rcu head for delayed free
- * @kobj: kobject used to represent this struct in sysfs
- * @dir: /proc/irq/ procfs entry
-+ * @debugfs_file: dentry for the debugfs file
- * @name: flow handler name for /proc/interrupts output
- */
- struct irq_desc {
-@@ -88,6 +89,9 @@ struct irq_desc {
- #ifdef CONFIG_PROC_FS
- struct proc_dir_entry *dir;
- #endif
-+#ifdef CONFIG_GENERIC_IRQ_DEBUGFS
-+ struct dentry *debugfs_file;
-+#endif
- #ifdef CONFIG_SPARSE_IRQ
- struct rcu_head rcu;
- struct kobject kobj;
---- a/include/linux/irqdomain.h
-+++ b/include/linux/irqdomain.h
-@@ -138,6 +138,7 @@ struct irq_domain_chip_generic;
- * setting up one or more generic chips for interrupt controllers
- * drivers using the generic chip library which uses this pointer.
- * @parent: Pointer to parent irq_domain to support hierarchy irq_domains
-+ * @debugfs_file: dentry for the domain debugfs file
- *
- * Revmap data, used internally by irq_domain
- * @revmap_direct_max_irq: The largest hwirq that can be set for controllers that
-@@ -160,6 +161,9 @@ struct irq_domain {
- #ifdef CONFIG_IRQ_DOMAIN_HIERARCHY
- struct irq_domain *parent;
- #endif
-+#ifdef CONFIG_GENERIC_IRQ_DEBUGFS
-+ struct dentry *debugfs_file;
-+#endif
-
- /* reverse map data. The linear map gets appended to the irq_domain */
- irq_hw_number_t hwirq_max;
-@@ -174,8 +178,8 @@ enum {
- /* Irq domain is hierarchical */
- IRQ_DOMAIN_FLAG_HIERARCHY = (1 << 0),
-
-- /* Core calls alloc/free recursive through the domain hierarchy. */
-- IRQ_DOMAIN_FLAG_AUTO_RECURSIVE = (1 << 1),
-+ /* Irq domain name was allocated in __irq_domain_add() */
-+ IRQ_DOMAIN_NAME_ALLOCATED = (1 << 6),
-
- /* Irq domain is an IPI domain with virq per cpu */
- IRQ_DOMAIN_FLAG_IPI_PER_CPU = (1 << 2),
-@@ -231,6 +235,9 @@ static inline bool is_fwnode_irqchip(str
- return fwnode && fwnode->type == FWNODE_IRQCHIP;
- }
-
-+extern void irq_domain_update_bus_token(struct irq_domain *domain,
-+ enum irq_domain_bus_token bus_token);
-+
- static inline
- struct irq_domain *irq_find_matching_fwnode(struct fwnode_handle *fwnode,
- enum irq_domain_bus_token bus_token)
-@@ -403,7 +410,7 @@ static inline int irq_domain_alloc_irqs(
- NULL);
- }
-
--extern int irq_domain_alloc_irqs_recursive(struct irq_domain *domain,
-+extern int irq_domain_alloc_irqs_hierarchy(struct irq_domain *domain,
- unsigned int irq_base,
- unsigned int nr_irqs, void *arg);
- extern int irq_domain_set_hwirq_and_chip(struct irq_domain *domain,
---- a/include/linux/netdev_features.h
-+++ b/include/linux/netdev_features.h
-@@ -74,6 +74,7 @@ enum {
- NETIF_F_BUSY_POLL_BIT, /* Busy poll */
-
- NETIF_F_HW_TC_BIT, /* Offload TC infrastructure */
-+ NETIF_F_HW_ACCEL_MQ_BIT, /* Hardware-accelerated multiqueue */
-
- /*
- * Add your fresh new feature above and remember to update
-@@ -136,6 +137,7 @@ enum {
- #define NETIF_F_HW_L2FW_DOFFLOAD __NETIF_F(HW_L2FW_DOFFLOAD)
- #define NETIF_F_BUSY_POLL __NETIF_F(BUSY_POLL)
- #define NETIF_F_HW_TC __NETIF_F(HW_TC)
-+#define NETIF_F_HW_ACCEL_MQ __NETIF_F(HW_ACCEL_MQ)
-
- #define for_each_netdev_feature(mask_addr, bit) \
- for_each_set_bit(bit, (unsigned long *)mask_addr, NETDEV_FEATURE_COUNT)
---- a/include/linux/netdevice.h
-+++ b/include/linux/netdevice.h
-@@ -930,7 +930,7 @@ struct netdev_xdp {
- * 3. Update dev->stats asynchronously and atomically, and define
- * neither operation.
- *
-- * bool (*ndo_has_offload_stats)(int attr_id)
-+ * bool (*ndo_has_offload_stats)(const struct net_device *dev, int attr_id)
- * Return true if this device supports offload stats of this attr_id.
- *
- * int (*ndo_get_offload_stats)(int attr_id, const struct net_device *dev,
-@@ -1167,7 +1167,7 @@ struct net_device_ops {
-
- struct rtnl_link_stats64* (*ndo_get_stats64)(struct net_device *dev,
- struct rtnl_link_stats64 *storage);
-- bool (*ndo_has_offload_stats)(int attr_id);
-+ bool (*ndo_has_offload_stats)(const struct net_device *dev, int attr_id);
- int (*ndo_get_offload_stats)(int attr_id,
- const struct net_device *dev,
- void *attr_data);
-@@ -1509,6 +1509,8 @@ enum netdev_priv_flags {
- * @if_port: Selectable AUI, TP, ...
- * @dma: DMA channel
- * @mtu: Interface MTU value
-+ * @min_mtu: Interface Minimum MTU value
-+ * @max_mtu: Interface Maximum MTU value
- * @type: Interface hardware type
- * @hard_header_len: Maximum hardware header length.
- * @min_header_len: Minimum hardware header length
-@@ -1735,6 +1737,8 @@ struct net_device {
- unsigned char dma;
-
- unsigned int mtu;
-+ unsigned int min_mtu;
-+ unsigned int max_mtu;
- unsigned short type;
- unsigned short hard_header_len;
- unsigned short min_header_len;
-@@ -1938,6 +1942,8 @@ int netdev_set_prio_tc_map(struct net_de
- return 0;
- }
-
-+int netdev_txq_to_tc(struct net_device *dev, unsigned int txq);
-+
- static inline
- void netdev_reset_tc(struct net_device *dev)
- {
---- a/include/linux/skbuff.h
-+++ b/include/linux/skbuff.h
-@@ -908,6 +908,7 @@ void kfree_skb(struct sk_buff *skb);
- void kfree_skb_list(struct sk_buff *segs);
- void skb_tx_error(struct sk_buff *skb);
- void consume_skb(struct sk_buff *skb);
-+void skb_recycle(struct sk_buff *skb);
- void __kfree_skb(struct sk_buff *skb);
- extern struct kmem_cache *skbuff_head_cache;
-
-@@ -3081,6 +3082,7 @@ static inline void skb_free_datagram_loc
- }
- int skb_kill_datagram(struct sock *sk, struct sk_buff *skb, unsigned int flags);
- int skb_copy_bits(const struct sk_buff *skb, int offset, void *to, int len);
-+void copy_skb_header(struct sk_buff *new, const struct sk_buff *old);
- int skb_store_bits(struct sk_buff *skb, int offset, const void *from, int len);
- __wsum skb_copy_and_csum_bits(const struct sk_buff *skb, int offset, u8 *to,
- int len, __wsum csum);
---- a/include/linux/sys_soc.h
-+++ b/include/linux/sys_soc.h
-@@ -13,6 +13,7 @@ struct soc_device_attribute {
- const char *family;
- const char *revision;
- const char *soc_id;
-+ const void *data;
- };
-
- /**
-@@ -34,4 +35,6 @@ void soc_device_unregister(struct soc_de
- */
- struct device *soc_device_to_device(struct soc_device *soc);
-
-+const struct soc_device_attribute *soc_device_match(
-+ const struct soc_device_attribute *matches);
- #endif /* __SOC_BUS_H */
---- a/include/net/switchdev.h
-+++ b/include/net/switchdev.h
-@@ -46,6 +46,7 @@ enum switchdev_attr_id {
- SWITCHDEV_ATTR_ID_PORT_PARENT_ID,
- SWITCHDEV_ATTR_ID_PORT_STP_STATE,
- SWITCHDEV_ATTR_ID_PORT_BRIDGE_FLAGS,
-+ SWITCHDEV_ATTR_ID_PORT_BRIDGE_FLAGS_SUPPORT,
- SWITCHDEV_ATTR_ID_BRIDGE_AGEING_TIME,
- SWITCHDEV_ATTR_ID_BRIDGE_VLAN_FILTERING,
- };
-@@ -60,6 +61,7 @@ struct switchdev_attr {
- struct netdev_phys_item_id ppid; /* PORT_PARENT_ID */
- u8 stp_state; /* PORT_STP_STATE */
- unsigned long brport_flags; /* PORT_BRIDGE_FLAGS */
-+ unsigned long brport_flags_support; /* PORT_BRIDGE_FLAGS_SUPPORT */
- clock_t ageing_time; /* BRIDGE_AGEING_TIME */
- bool vlan_filtering; /* BRIDGE_VLAN_FILTERING */
- } u;
-@@ -149,8 +151,10 @@ struct switchdev_ops {
- };
-
- enum switchdev_notifier_type {
-- SWITCHDEV_FDB_ADD = 1,
-- SWITCHDEV_FDB_DEL,
-+ SWITCHDEV_FDB_ADD_TO_BRIDGE = 1,
-+ SWITCHDEV_FDB_DEL_TO_BRIDGE,
-+ SWITCHDEV_FDB_ADD_TO_DEVICE,
-+ SWITCHDEV_FDB_DEL_TO_DEVICE,
- };
-
- struct switchdev_notifier_info {
---- a/include/uapi/linux/if_ether.h
-+++ b/include/uapi/linux/if_ether.h
-@@ -36,6 +36,7 @@
- #define ETH_DATA_LEN 1500 /* Max. octets in payload */
- #define ETH_FRAME_LEN 1514 /* Max. octets in frame sans FCS */
- #define ETH_FCS_LEN 4 /* Octets in the FCS */
-+#define ETH_MIN_MTU 68 /* Min IPv4 MTU per RFC791 */
-
- /*
- * These are the defined Ethernet Protocol ID's.
---- a/kernel/irq/Kconfig
-+++ b/kernel/irq/Kconfig
-@@ -108,4 +108,15 @@ config SPARSE_IRQ
-
- If you don't know what to do here, say N.
-
-+config GENERIC_IRQ_DEBUGFS
-+ bool "Expose irq internals in debugfs"
-+ depends on DEBUG_FS
-+ default n
-+ ---help---
-+
-+ Exposes internal state information through debugfs. Mostly for
-+ developers and debugging of hard to diagnose interrupt problems.
-+
-+ If you don't know what to do here, say N.
-+
- endmenu
---- a/kernel/irq/Makefile
-+++ b/kernel/irq/Makefile
-@@ -10,3 +10,4 @@ obj-$(CONFIG_PM_SLEEP) += pm.o
- obj-$(CONFIG_GENERIC_MSI_IRQ) += msi.o
- obj-$(CONFIG_GENERIC_IRQ_IPI) += ipi.o
- obj-$(CONFIG_SMP) += affinity.o
-+obj-$(CONFIG_GENERIC_IRQ_DEBUGFS) += debugfs.o
---- /dev/null
-+++ b/kernel/irq/debugfs.c
-@@ -0,0 +1,215 @@
-+/*
-+ * Copyright 2017 Thomas Gleixner <tglx@linutronix.de>
-+ *
-+ * This file is licensed under the GPL V2.
-+ */
-+#include <linux/debugfs.h>
-+#include <linux/irqdomain.h>
-+#include <linux/irq.h>
-+
-+#include "internals.h"
-+
-+static struct dentry *irq_dir;
-+
-+struct irq_bit_descr {
-+ unsigned int mask;
-+ char *name;
-+};
-+#define BIT_MASK_DESCR(m) { .mask = m, .name = #m }
-+
-+static void irq_debug_show_bits(struct seq_file *m, int ind, unsigned int state,
-+ const struct irq_bit_descr *sd, int size)
-+{
-+ int i;
-+
-+ for (i = 0; i < size; i++, sd++) {
-+ if (state & sd->mask)
-+ seq_printf(m, "%*s%s\n", ind + 12, "", sd->name);
-+ }
-+}
-+
-+#ifdef CONFIG_SMP
-+static void irq_debug_show_masks(struct seq_file *m, struct irq_desc *desc)
-+{
-+ struct irq_data *data = irq_desc_get_irq_data(desc);
-+ struct cpumask *msk;
-+
-+ msk = irq_data_get_affinity_mask(data);
-+ seq_printf(m, "affinity: %*pbl\n", cpumask_pr_args(msk));
-+#ifdef CONFIG_GENERIC_PENDING_IRQ
-+ msk = desc->pending_mask;
-+ seq_printf(m, "pending: %*pbl\n", cpumask_pr_args(msk));
-+#endif
-+}
-+#else
-+static void irq_debug_show_masks(struct seq_file *m, struct irq_desc *desc) { }
-+#endif
-+
-+static const struct irq_bit_descr irqchip_flags[] = {
-+ BIT_MASK_DESCR(IRQCHIP_SET_TYPE_MASKED),
-+ BIT_MASK_DESCR(IRQCHIP_EOI_IF_HANDLED),
-+ BIT_MASK_DESCR(IRQCHIP_MASK_ON_SUSPEND),
-+ BIT_MASK_DESCR(IRQCHIP_ONOFFLINE_ENABLED),
-+ BIT_MASK_DESCR(IRQCHIP_SKIP_SET_WAKE),
-+ BIT_MASK_DESCR(IRQCHIP_ONESHOT_SAFE),
-+ BIT_MASK_DESCR(IRQCHIP_EOI_THREADED),
-+};
-+
-+static void
-+irq_debug_show_chip(struct seq_file *m, struct irq_data *data, int ind)
-+{
-+ struct irq_chip *chip = data->chip;
-+
-+ if (!chip) {
-+ seq_printf(m, "chip: None\n");
-+ return;
-+ }
-+ seq_printf(m, "%*schip: %s\n", ind, "", chip->name);
-+ seq_printf(m, "%*sflags: 0x%lx\n", ind + 1, "", chip->flags);
-+ irq_debug_show_bits(m, ind, chip->flags, irqchip_flags,
-+ ARRAY_SIZE(irqchip_flags));
-+}
-+
-+static void
-+irq_debug_show_data(struct seq_file *m, struct irq_data *data, int ind)
-+{
-+ seq_printf(m, "%*sdomain: %s\n", ind, "",
-+ data->domain ? data->domain->name : "");
-+ seq_printf(m, "%*shwirq: 0x%lx\n", ind + 1, "", data->hwirq);
-+ irq_debug_show_chip(m, data, ind + 1);
-+#ifdef CONFIG_IRQ_DOMAIN_HIERARCHY
-+ if (!data->parent_data)
-+ return;
-+ seq_printf(m, "%*sparent:\n", ind + 1, "");
-+ irq_debug_show_data(m, data->parent_data, ind + 4);
-+#endif
-+}
-+
-+static const struct irq_bit_descr irqdata_states[] = {
-+ BIT_MASK_DESCR(IRQ_TYPE_EDGE_RISING),
-+ BIT_MASK_DESCR(IRQ_TYPE_EDGE_FALLING),
-+ BIT_MASK_DESCR(IRQ_TYPE_LEVEL_HIGH),
-+ BIT_MASK_DESCR(IRQ_TYPE_LEVEL_LOW),
-+ BIT_MASK_DESCR(IRQD_LEVEL),
-+
-+ BIT_MASK_DESCR(IRQD_ACTIVATED),
-+ BIT_MASK_DESCR(IRQD_IRQ_STARTED),
-+ BIT_MASK_DESCR(IRQD_IRQ_DISABLED),
-+ BIT_MASK_DESCR(IRQD_IRQ_MASKED),
-+ BIT_MASK_DESCR(IRQD_IRQ_INPROGRESS),
-+
-+ BIT_MASK_DESCR(IRQD_PER_CPU),
-+ BIT_MASK_DESCR(IRQD_NO_BALANCING),
-+
-+ BIT_MASK_DESCR(IRQD_MOVE_PCNTXT),
-+ BIT_MASK_DESCR(IRQD_AFFINITY_SET),
-+ BIT_MASK_DESCR(IRQD_SETAFFINITY_PENDING),
-+ BIT_MASK_DESCR(IRQD_AFFINITY_MANAGED),
-+ BIT_MASK_DESCR(IRQD_MANAGED_SHUTDOWN),
-+
-+ BIT_MASK_DESCR(IRQD_FORWARDED_TO_VCPU),
-+
-+ BIT_MASK_DESCR(IRQD_WAKEUP_STATE),
-+ BIT_MASK_DESCR(IRQD_WAKEUP_ARMED),
-+};
-+
-+static const struct irq_bit_descr irqdesc_states[] = {
-+ BIT_MASK_DESCR(_IRQ_NOPROBE),
-+ BIT_MASK_DESCR(_IRQ_NOREQUEST),
-+ BIT_MASK_DESCR(_IRQ_NOTHREAD),
-+ BIT_MASK_DESCR(_IRQ_NOAUTOEN),
-+ BIT_MASK_DESCR(_IRQ_NESTED_THREAD),
-+ BIT_MASK_DESCR(_IRQ_PER_CPU_DEVID),
-+ BIT_MASK_DESCR(_IRQ_IS_POLLED),
-+ BIT_MASK_DESCR(_IRQ_DISABLE_UNLAZY),
-+};
-+
-+static const struct irq_bit_descr irqdesc_istates[] = {
-+ BIT_MASK_DESCR(IRQS_AUTODETECT),
-+ BIT_MASK_DESCR(IRQS_SPURIOUS_DISABLED),
-+ BIT_MASK_DESCR(IRQS_POLL_INPROGRESS),
-+ BIT_MASK_DESCR(IRQS_ONESHOT),
-+ BIT_MASK_DESCR(IRQS_REPLAY),
-+ BIT_MASK_DESCR(IRQS_WAITING),
-+ BIT_MASK_DESCR(IRQS_PENDING),
-+ BIT_MASK_DESCR(IRQS_SUSPENDED),
-+};
-+
-+
-+static int irq_debug_show(struct seq_file *m, void *p)
-+{
-+ struct irq_desc *desc = m->private;
-+ struct irq_data *data;
-+
-+ raw_spin_lock_irq(&desc->lock);
-+ data = irq_desc_get_irq_data(desc);
-+ seq_printf(m, "handler: %pf\n", desc->handle_irq);
-+ seq_printf(m, "status: 0x%08x\n", desc->status_use_accessors);
-+ irq_debug_show_bits(m, 0, desc->status_use_accessors, irqdesc_states,
-+ ARRAY_SIZE(irqdesc_states));
-+ seq_printf(m, "istate: 0x%08x\n", desc->istate);
-+ irq_debug_show_bits(m, 0, desc->istate, irqdesc_istates,
-+ ARRAY_SIZE(irqdesc_istates));
-+ seq_printf(m, "ddepth: %u\n", desc->depth);
-+ seq_printf(m, "wdepth: %u\n", desc->wake_depth);
-+ seq_printf(m, "dstate: 0x%08x\n", irqd_get(data));
-+ irq_debug_show_bits(m, 0, irqd_get(data), irqdata_states,
-+ ARRAY_SIZE(irqdata_states));
-+ seq_printf(m, "node: %d\n", irq_data_get_node(data));
-+ irq_debug_show_masks(m, desc);
-+ irq_debug_show_data(m, data, 0);
-+ raw_spin_unlock_irq(&desc->lock);
-+ return 0;
-+}
-+
-+static int irq_debug_open(struct inode *inode, struct file *file)
-+{
-+ return single_open(file, irq_debug_show, inode->i_private);
-+}
-+
-+static const struct file_operations dfs_irq_ops = {
-+ .open = irq_debug_open,
-+ .read = seq_read,
-+ .llseek = seq_lseek,
-+ .release = single_release,
-+};
-+
-+void irq_add_debugfs_entry(unsigned int irq, struct irq_desc *desc)
-+{
-+ char name [10];
-+
-+ if (!irq_dir || !desc || desc->debugfs_file)
-+ return;
-+
-+ sprintf(name, "%d", irq);
-+ desc->debugfs_file = debugfs_create_file(name, 0444, irq_dir, desc,
-+ &dfs_irq_ops);
-+}
-+
-+void irq_remove_debugfs_entry(struct irq_desc *desc)
-+{
-+ if (desc->debugfs_file)
-+ debugfs_remove(desc->debugfs_file);
-+}
-+
-+static int __init irq_debugfs_init(void)
-+{
-+ struct dentry *root_dir;
-+ int irq;
-+
-+ root_dir = debugfs_create_dir("irq", NULL);
-+ if (!root_dir)
-+ return -ENOMEM;
-+
-+ irq_domain_debugfs_init(root_dir);
-+
-+ irq_dir = debugfs_create_dir("irqs", root_dir);
-+
-+ irq_lock_sparse();
-+ for_each_active_irq(irq)
-+ irq_add_debugfs_entry(irq, irq_to_desc(irq));
-+ irq_unlock_sparse();
-+
-+ return 0;
-+}
-+__initcall(irq_debugfs_init);
---- a/kernel/irq/internals.h
-+++ b/kernel/irq/internals.h
-@@ -169,6 +169,11 @@ irq_put_desc_unlock(struct irq_desc *des
-
- #define __irqd_to_state(d) ACCESS_PRIVATE((d)->common, state_use_accessors)
-
-+static inline unsigned int irqd_get(struct irq_data *d)
-+{
-+ return __irqd_to_state(d);
-+}
-+
- /*
- * Manipulation functions for irq_data.state
- */
-@@ -226,3 +231,20 @@ irq_pm_install_action(struct irq_desc *d
- static inline void
- irq_pm_remove_action(struct irq_desc *desc, struct irqaction *action) { }
- #endif
-+
-+#ifdef CONFIG_GENERIC_IRQ_DEBUGFS
-+void irq_add_debugfs_entry(unsigned int irq, struct irq_desc *desc);
-+void irq_remove_debugfs_entry(struct irq_desc *desc);
-+# ifdef CONFIG_IRQ_DOMAIN
-+void irq_domain_debugfs_init(struct dentry *root);
-+# else
-+static inline void irq_domain_debugfs_init(struct dentry *root);
-+# endif
-+#else /* CONFIG_GENERIC_IRQ_DEBUGFS */
-+static inline void irq_add_debugfs_entry(unsigned int irq, struct irq_desc *d)
-+{
-+}
-+static inline void irq_remove_debugfs_entry(struct irq_desc *d)
-+{
-+}
-+#endif /* CONFIG_GENERIC_IRQ_DEBUGFS */
---- a/kernel/irq/irqdesc.c
-+++ b/kernel/irq/irqdesc.c
-@@ -394,6 +394,7 @@ static void free_desc(unsigned int irq)
- {
- struct irq_desc *desc = irq_to_desc(irq);
-
-+ irq_remove_debugfs_entry(desc);
- unregister_irq_proc(irq, desc);
-
- /*
---- a/kernel/irq/irqdomain.c
-+++ b/kernel/irq/irqdomain.c
-@@ -31,6 +31,14 @@ struct irqchip_fwid {
- void *data;
- };
-
-+#ifdef CONFIG_GENERIC_IRQ_DEBUGFS
-+static void debugfs_add_domain_dir(struct irq_domain *d);
-+static void debugfs_remove_domain_dir(struct irq_domain *d);
-+#else
-+static inline void debugfs_add_domain_dir(struct irq_domain *d) { }
-+static inline void debugfs_remove_domain_dir(struct irq_domain *d) { }
-+#endif
-+
- /**
- * irq_domain_alloc_fwnode - Allocate a fwnode_handle suitable for
- * identifying an irq domain
-@@ -117,6 +125,7 @@ struct irq_domain *__irq_domain_add(stru
- irq_domain_check_hierarchy(domain);
-
- mutex_lock(&irq_domain_mutex);
-+ debugfs_add_domain_dir(domain);
- list_add(&domain->link, &irq_domain_list);
- mutex_unlock(&irq_domain_mutex);
-
-@@ -136,6 +145,7 @@ EXPORT_SYMBOL_GPL(__irq_domain_add);
- void irq_domain_remove(struct irq_domain *domain)
- {
- mutex_lock(&irq_domain_mutex);
-+ debugfs_remove_domain_dir(domain);
-
- WARN_ON(!radix_tree_empty(&domain->revmap_tree));
-
-@@ -156,6 +166,37 @@ void irq_domain_remove(struct irq_domain
- }
- EXPORT_SYMBOL_GPL(irq_domain_remove);
-
-+void irq_domain_update_bus_token(struct irq_domain *domain,
-+ enum irq_domain_bus_token bus_token)
-+{
-+ char *name;
-+
-+ if (domain->bus_token == bus_token)
-+ return;
-+
-+ mutex_lock(&irq_domain_mutex);
-+
-+ domain->bus_token = bus_token;
-+
-+ name = kasprintf(GFP_KERNEL, "%s-%d", domain->name, bus_token);
-+ if (!name) {
-+ mutex_unlock(&irq_domain_mutex);
-+ return;
-+ }
-+
-+ debugfs_remove_domain_dir(domain);
-+
-+ if (domain->flags & IRQ_DOMAIN_NAME_ALLOCATED)
-+ kfree(domain->name);
-+ else
-+ domain->flags |= IRQ_DOMAIN_NAME_ALLOCATED;
-+
-+ domain->name = name;
-+ debugfs_add_domain_dir(domain);
-+
-+ mutex_unlock(&irq_domain_mutex);
-+}
-+
- /**
- * irq_domain_add_simple() - Register an irq_domain and optionally map a range of irqs
- * @of_node: pointer to interrupt controller's device tree node.
-@@ -1164,43 +1205,18 @@ void irq_domain_free_irqs_top(struct irq
- irq_domain_free_irqs_common(domain, virq, nr_irqs);
- }
-
--static bool irq_domain_is_auto_recursive(struct irq_domain *domain)
--{
-- return domain->flags & IRQ_DOMAIN_FLAG_AUTO_RECURSIVE;
--}
--
--static void irq_domain_free_irqs_recursive(struct irq_domain *domain,
-+static void irq_domain_free_irqs_hierarchy(struct irq_domain *domain,
- unsigned int irq_base,
- unsigned int nr_irqs)
- {
- domain->ops->free(domain, irq_base, nr_irqs);
-- if (irq_domain_is_auto_recursive(domain)) {
-- BUG_ON(!domain->parent);
-- irq_domain_free_irqs_recursive(domain->parent, irq_base,
-- nr_irqs);
-- }
- }
-
--int irq_domain_alloc_irqs_recursive(struct irq_domain *domain,
-+int irq_domain_alloc_irqs_hierarchy(struct irq_domain *domain,
- unsigned int irq_base,
- unsigned int nr_irqs, void *arg)
- {
-- int ret = 0;
-- struct irq_domain *parent = domain->parent;
-- bool recursive = irq_domain_is_auto_recursive(domain);
--
-- BUG_ON(recursive && !parent);
-- if (recursive)
-- ret = irq_domain_alloc_irqs_recursive(parent, irq_base,
-- nr_irqs, arg);
-- if (ret < 0)
-- return ret;
--
-- ret = domain->ops->alloc(domain, irq_base, nr_irqs, arg);
-- if (ret < 0 && recursive)
-- irq_domain_free_irqs_recursive(parent, irq_base, nr_irqs);
--
-- return ret;
-+ return domain->ops->alloc(domain, irq_base, nr_irqs, arg);
- }
-
- /**
-@@ -1261,7 +1277,7 @@ int __irq_domain_alloc_irqs(struct irq_d
- }
-
- mutex_lock(&irq_domain_mutex);
-- ret = irq_domain_alloc_irqs_recursive(domain, virq, nr_irqs, arg);
-+ ret = irq_domain_alloc_irqs_hierarchy(domain, virq, nr_irqs, arg);
- if (ret < 0) {
- mutex_unlock(&irq_domain_mutex);
- goto out_free_irq_data;
-@@ -1296,7 +1312,7 @@ void irq_domain_free_irqs(unsigned int v
- mutex_lock(&irq_domain_mutex);
- for (i = 0; i < nr_irqs; i++)
- irq_domain_remove_irq(virq + i);
-- irq_domain_free_irqs_recursive(data->domain, virq, nr_irqs);
-+ irq_domain_free_irqs_hierarchy(data->domain, virq, nr_irqs);
- mutex_unlock(&irq_domain_mutex);
-
- irq_domain_free_irq_data(virq, nr_irqs);
-@@ -1316,15 +1332,11 @@ int irq_domain_alloc_irqs_parent(struct
- unsigned int irq_base, unsigned int nr_irqs,
- void *arg)
- {
-- /* irq_domain_alloc_irqs_recursive() has called parent's alloc() */
-- if (irq_domain_is_auto_recursive(domain))
-- return 0;
-+ if (!domain->parent)
-+ return -ENOSYS;
-
-- domain = domain->parent;
-- if (domain)
-- return irq_domain_alloc_irqs_recursive(domain, irq_base,
-- nr_irqs, arg);
-- return -ENOSYS;
-+ return irq_domain_alloc_irqs_hierarchy(domain->parent, irq_base,
-+ nr_irqs, arg);
- }
- EXPORT_SYMBOL_GPL(irq_domain_alloc_irqs_parent);
-
-@@ -1339,10 +1351,10 @@ EXPORT_SYMBOL_GPL(irq_domain_alloc_irqs_
- void irq_domain_free_irqs_parent(struct irq_domain *domain,
- unsigned int irq_base, unsigned int nr_irqs)
- {
-- /* irq_domain_free_irqs_recursive() will call parent's free */
-- if (!irq_domain_is_auto_recursive(domain) && domain->parent)
-- irq_domain_free_irqs_recursive(domain->parent, irq_base,
-- nr_irqs);
-+ if (!domain->parent)
-+ return;
-+
-+ irq_domain_free_irqs_hierarchy(domain->parent, irq_base, nr_irqs);
- }
- EXPORT_SYMBOL_GPL(irq_domain_free_irqs_parent);
-
-@@ -1448,3 +1460,78 @@ static void irq_domain_check_hierarchy(s
- {
- }
- #endif /* CONFIG_IRQ_DOMAIN_HIERARCHY */
-+
-+#ifdef CONFIG_GENERIC_IRQ_DEBUGFS
-+static struct dentry *domain_dir;
-+
-+static void
-+irq_domain_debug_show_one(struct seq_file *m, struct irq_domain *d, int ind)
-+{
-+ seq_printf(m, "%*sname: %s\n", ind, "", d->name);
-+ seq_printf(m, "%*ssize: %u\n", ind + 1, "",
-+ d->revmap_size + d->revmap_direct_max_irq);
-+ seq_printf(m, "%*smapped: %u\n", ind + 1, "", d->mapcount);
-+ seq_printf(m, "%*sflags: 0x%08x\n", ind +1 , "", d->flags);
-+#ifdef CONFIG_IRQ_DOMAIN_HIERARCHY
-+ if (!d->parent)
-+ return;
-+ seq_printf(m, "%*sparent: %s\n", ind + 1, "", d->parent->name);
-+ irq_domain_debug_show_one(m, d->parent, ind + 4);
-+#endif
-+}
-+
-+static int irq_domain_debug_show(struct seq_file *m, void *p)
-+{
-+ struct irq_domain *d = m->private;
-+
-+ /* Default domain? Might be NULL */
-+ if (!d) {
-+ if (!irq_default_domain)
-+ return 0;
-+ d = irq_default_domain;
-+ }
-+ irq_domain_debug_show_one(m, d, 0);
-+ return 0;
-+}
-+
-+static int irq_domain_debug_open(struct inode *inode, struct file *file)
-+{
-+ return single_open(file, irq_domain_debug_show, inode->i_private);
-+}
-+
-+static const struct file_operations dfs_domain_ops = {
-+ .open = irq_domain_debug_open,
-+ .read = seq_read,
-+ .llseek = seq_lseek,
-+ .release = single_release,
-+};
-+
-+static void debugfs_add_domain_dir(struct irq_domain *d)
-+{
-+ if (!d->name || !domain_dir || d->debugfs_file)
-+ return;
-+ d->debugfs_file = debugfs_create_file(d->name, 0444, domain_dir, d,
-+ &dfs_domain_ops);
-+}
-+
-+static void debugfs_remove_domain_dir(struct irq_domain *d)
-+{
-+ if (d->debugfs_file)
-+ debugfs_remove(d->debugfs_file);
-+}
-+
-+void __init irq_domain_debugfs_init(struct dentry *root)
-+{
-+ struct irq_domain *d;
-+
-+ domain_dir = debugfs_create_dir("domains", root);
-+ if (!domain_dir)
-+ return;
-+
-+ debugfs_create_file("default", 0444, domain_dir, NULL, &dfs_domain_ops);
-+ mutex_lock(&irq_domain_mutex);
-+ list_for_each_entry(d, &irq_domain_list, link)
-+ debugfs_add_domain_dir(d);
-+ mutex_unlock(&irq_domain_mutex);
-+}
-+#endif
---- a/kernel/irq/manage.c
-+++ b/kernel/irq/manage.c
-@@ -1391,6 +1391,7 @@ __setup_irq(unsigned int irq, struct irq
- wake_up_process(new->secondary->thread);
-
- register_irq_proc(irq, desc);
-+ irq_add_debugfs_entry(irq, desc);
- new->dir = NULL;
- register_handler_proc(irq, new);
- free_cpumask_var(mask);
---- a/kernel/irq/msi.c
-+++ b/kernel/irq/msi.c
-@@ -310,7 +310,7 @@ int msi_domain_populate_irqs(struct irq_
-
- ops->set_desc(arg, desc);
- /* Assumes the domain mutex is held! */
-- ret = irq_domain_alloc_irqs_recursive(domain, virq, 1, arg);
-+ ret = irq_domain_alloc_irqs_hierarchy(domain, virq, 1, arg);
- if (ret)
- break;
-
---- a/net/bridge/br.c
-+++ b/net/bridge/br.c
-@@ -138,14 +138,14 @@ static int br_switchdev_event(struct not
- br = p->br;
-
- switch (event) {
-- case SWITCHDEV_FDB_ADD:
-+ case SWITCHDEV_FDB_ADD_TO_BRIDGE:
- fdb_info = ptr;
- err = br_fdb_external_learn_add(br, p, fdb_info->addr,
- fdb_info->vid);
- if (err)
- err = notifier_from_errno(err);
- break;
-- case SWITCHDEV_FDB_DEL:
-+ case SWITCHDEV_FDB_DEL_TO_BRIDGE:
- fdb_info = ptr;
- err = br_fdb_external_learn_del(br, p, fdb_info->addr,
- fdb_info->vid);
---- a/net/bridge/br_fdb.c
-+++ b/net/bridge/br_fdb.c
-@@ -688,6 +688,8 @@ static void fdb_notify(struct net_bridge
- struct sk_buff *skb;
- int err = -ENOBUFS;
-
-+ br_switchdev_fdb_notify(fdb, type);
-+
- skb = nlmsg_new(fdb_nlmsg_size(), GFP_ATOMIC);
- if (skb == NULL)
- goto errout;
---- a/net/bridge/br_private.h
-+++ b/net/bridge/br_private.h
-@@ -1060,6 +1060,8 @@ void nbp_switchdev_frame_mark(const stru
- struct sk_buff *skb);
- bool nbp_switchdev_allowed_egress(const struct net_bridge_port *p,
- const struct sk_buff *skb);
-+void br_switchdev_fdb_notify(const struct net_bridge_fdb_entry *fdb,
-+ int type);
- #else
- static inline int nbp_switchdev_mark_set(struct net_bridge_port *p)
- {
-@@ -1076,6 +1078,11 @@ static inline bool nbp_switchdev_allowed
- {
- return true;
- }
-+
-+static inline void
-+br_switchdev_fdb_notify(const struct net_bridge_fdb_entry *fdb, int type)
-+{
-+}
- #endif /* CONFIG_NET_SWITCHDEV */
-
- #endif
---- a/net/bridge/br_switchdev.c
-+++ b/net/bridge/br_switchdev.c
-@@ -55,3 +55,36 @@ bool nbp_switchdev_allowed_egress(const
- return !skb->offload_fwd_mark ||
- BR_INPUT_SKB_CB(skb)->offload_fwd_mark != p->offload_fwd_mark;
- }
-+
-+static void
-+br_switchdev_fdb_call_notifiers(bool adding, const unsigned char *mac,
-+ u16 vid, struct net_device *dev)
-+{
-+ struct switchdev_notifier_fdb_info info;
-+ unsigned long notifier_type;
-+
-+ info.addr = mac;
-+ info.vid = vid;
-+ notifier_type = adding ? SWITCHDEV_FDB_ADD_TO_DEVICE : SWITCHDEV_FDB_DEL_TO_DEVICE;
-+ call_switchdev_notifiers(notifier_type, dev, &info.info);
-+}
-+
-+void
-+br_switchdev_fdb_notify(const struct net_bridge_fdb_entry *fdb, int type)
-+{
-+ if (!fdb->added_by_user)
-+ return;
-+
-+ switch (type) {
-+ case RTM_DELNEIGH:
-+ br_switchdev_fdb_call_notifiers(false, fdb->addr.addr,
-+ fdb->vlan_id,
-+ fdb->dst->dev);
-+ break;
-+ case RTM_NEWNEIGH:
-+ br_switchdev_fdb_call_notifiers(true, fdb->addr.addr,
-+ fdb->vlan_id,
-+ fdb->dst->dev);
-+ break;
-+ }
-+}
---- a/net/core/dev.c
-+++ b/net/core/dev.c
-@@ -1990,6 +1990,23 @@ static void netif_setup_tc(struct net_de
- }
- }
-
-+int netdev_txq_to_tc(struct net_device *dev, unsigned int txq)
-+{
-+ if (dev->num_tc) {
-+ struct netdev_tc_txq *tc = &dev->tc_to_txq[0];
-+ int i;
-+
-+ for (i = 0; i < TC_MAX_QUEUE; i++, tc++) {
-+ if ((txq - tc->offset) < tc->count)
-+ return i;
-+ }
-+
-+ return -1;
-+ }
-+
-+ return 0;
-+}
-+
- #ifdef CONFIG_XPS
- static DEFINE_MUTEX(xps_map_mutex);
- #define xmap_dereference(P) \
-@@ -6656,9 +6673,18 @@ int dev_set_mtu(struct net_device *dev,
- if (new_mtu == dev->mtu)
- return 0;
-
-- /* MTU must be positive. */
-- if (new_mtu < 0)
-+ /* MTU must be positive, and in range */
-+ if (new_mtu < 0 || new_mtu < dev->min_mtu) {
-+ net_err_ratelimited("%s: Invalid MTU %d requested, hw min %d\n",
-+ dev->name, new_mtu, dev->min_mtu);
- return -EINVAL;
-+ }
-+
-+ if (dev->max_mtu > 0 && new_mtu > dev->max_mtu) {
-+ net_err_ratelimited("%s: Invalid MTU %d requested, hw max %d\n",
-+ dev->name, new_mtu, dev->min_mtu);
-+ return -EINVAL;
-+ }
-
- if (!netif_device_present(dev))
- return -ENODEV;
---- a/net/core/net-sysfs.c
-+++ b/net/core/net-sysfs.c
-@@ -1021,7 +1021,6 @@ static ssize_t show_trans_timeout(struct
- return sprintf(buf, "%lu", trans_timeout);
- }
-
--#ifdef CONFIG_XPS
- static unsigned int get_netdev_queue_index(struct netdev_queue *queue)
- {
- struct net_device *dev = queue->dev;
-@@ -1033,6 +1032,21 @@ static unsigned int get_netdev_queue_ind
- return i;
- }
-
-+static ssize_t show_traffic_class(struct netdev_queue *queue,
-+ struct netdev_queue_attribute *attribute,
-+ char *buf)
-+{
-+ struct net_device *dev = queue->dev;
-+ int index = get_netdev_queue_index(queue);
-+ int tc = netdev_txq_to_tc(dev, index);
-+
-+ if (tc < 0)
-+ return -EINVAL;
-+
-+ return sprintf(buf, "%u\n", tc);
-+}
-+
-+#ifdef CONFIG_XPS
- static ssize_t show_tx_maxrate(struct netdev_queue *queue,
- struct netdev_queue_attribute *attribute,
- char *buf)
-@@ -1075,6 +1089,9 @@ static struct netdev_queue_attribute que
- static struct netdev_queue_attribute queue_trans_timeout =
- __ATTR(tx_timeout, S_IRUGO, show_trans_timeout, NULL);
-
-+static struct netdev_queue_attribute queue_traffic_class =
-+ __ATTR(traffic_class, S_IRUGO, show_traffic_class, NULL);
-+
- #ifdef CONFIG_BQL
- /*
- * Byte queue limits sysfs structures and functions.
-@@ -1260,6 +1277,7 @@ static struct netdev_queue_attribute xps
-
- static struct attribute *netdev_queue_default_attrs[] = {
- &queue_trans_timeout.attr,
-+ &queue_traffic_class.attr,
- #ifdef CONFIG_XPS
- &xps_cpus_attribute.attr,
- &queue_tx_maxrate.attr,
---- a/net/core/rtnetlink.c
-+++ b/net/core/rtnetlink.c
-@@ -3709,7 +3709,7 @@ static int rtnl_get_offload_stats(struct
- if (!size)
- continue;
-
-- if (!dev->netdev_ops->ndo_has_offload_stats(attr_id))
-+ if (!dev->netdev_ops->ndo_has_offload_stats(dev, attr_id))
- continue;
-
- attr = nla_reserve_64bit(skb, attr_id, size,
-@@ -3750,7 +3750,7 @@ static int rtnl_get_offload_stats_size(c
-
- for (attr_id = IFLA_OFFLOAD_XSTATS_FIRST;
- attr_id <= IFLA_OFFLOAD_XSTATS_MAX; attr_id++) {
-- if (!dev->netdev_ops->ndo_has_offload_stats(attr_id))
-+ if (!dev->netdev_ops->ndo_has_offload_stats(dev, attr_id))
- continue;
- size = rtnl_get_offload_stats_attr_size(attr_id);
- nla_size += nla_total_size_64bit(size);
---- a/net/core/skbuff.c
-+++ b/net/core/skbuff.c
-@@ -842,6 +842,32 @@ void napi_consume_skb(struct sk_buff *sk
- }
- EXPORT_SYMBOL(napi_consume_skb);
-
-+/**
-+ * skb_recycle - clean up an skb for reuse
-+ * @skb: buffer
-+ *
-+ * Recycles the skb to be reused as a receive buffer. This
-+ * function does any necessary reference count dropping, and
-+ * cleans up the skbuff as if it just came from __alloc_skb().
-+ */
-+void skb_recycle(struct sk_buff *skb)
-+{
-+ struct skb_shared_info *shinfo;
-+ u8 head_frag = skb->head_frag;
-+
-+ skb_release_head_state(skb);
-+
-+ shinfo = skb_shinfo(skb);
-+ memset(shinfo, 0, offsetof(struct skb_shared_info, dataref));
-+ atomic_set(&shinfo->dataref, 1);
-+
-+ memset(skb, 0, offsetof(struct sk_buff, tail));
-+ skb->data = skb->head + NET_SKB_PAD;
-+ skb->head_frag = head_frag;
-+ skb_reset_tail_pointer(skb);
-+}
-+EXPORT_SYMBOL(skb_recycle);
-+
- /* Make sure a field is enclosed inside headers_start/headers_end section */
- #define CHECK_SKB_FIELD(field) \
- BUILD_BUG_ON(offsetof(struct sk_buff, field) < \
-@@ -1075,7 +1101,7 @@ static void skb_headers_offset_update(st
- skb->inner_mac_header += off;
- }
-
--static void copy_skb_header(struct sk_buff *new, const struct sk_buff *old)
-+void copy_skb_header(struct sk_buff *new, const struct sk_buff *old)
- {
- __copy_skb_header(new, old);
-
-@@ -1083,6 +1109,7 @@ static void copy_skb_header(struct sk_bu
- skb_shinfo(new)->gso_segs = skb_shinfo(old)->gso_segs;
- skb_shinfo(new)->gso_type = skb_shinfo(old)->gso_type;
- }
-+EXPORT_SYMBOL(copy_skb_header);
-
- static inline int skb_alloc_rx_flag(const struct sk_buff *skb)
- {
---- a/net/sched/sch_generic.c
-+++ b/net/sched/sch_generic.c
-@@ -309,6 +309,13 @@ static void dev_watchdog(unsigned long a
- txq->trans_timeout++;
- break;
- }
-+
-+ /* Devices with HW_ACCEL_MQ have multiple txqs
-+ * but update only the first one's transmission
-+ * timestamp so avoid checking the rest.
-+ */
-+ if (dev->features & NETIF_F_HW_ACCEL_MQ)
-+ break;
- }
-
- if (some_queue_timedout) {
diff --git a/target/linux/layerscape/patches-4.9/301-arch-support-layerscape.patch b/target/linux/layerscape/patches-4.9/301-arch-support-layerscape.patch
deleted file mode 100644
index 2985cff0b5..0000000000
--- a/target/linux/layerscape/patches-4.9/301-arch-support-layerscape.patch
+++ /dev/null
@@ -1,517 +0,0 @@
-From 2f2a0ab9e4b3186be981f7151a4f4f794d4b6caa Mon Sep 17 00:00:00 2001
-From: Yangbo Lu <yangbo.lu@nxp.com>
-Date: Thu, 5 Jul 2018 16:18:37 +0800
-Subject: [PATCH 03/32] arch: support layerscape
-MIME-Version: 1.0
-Content-Type: text/plain; charset=UTF-8
-Content-Transfer-Encoding: 8bit
-
-This is an integrated patch for layerscape arch support.
-
-Signed-off-by: Madalin Bucur <madalin.bucur@nxp.com>
-Signed-off-by: Nipun Gupta <nipun.gupta@nxp.com>
-Signed-off-by: Zhao Qiang <B45475@freescale.com>
-Signed-off-by: Camelia Groza <camelia.groza@nxp.com>
-Signed-off-by: Haiying Wang <Haiying.wang@freescale.com>
-Signed-off-by: Pan Jiafei <Jiafei.Pan@nxp.com>
-Signed-off-by: Po Liu <po.liu@nxp.com>
-Signed-off-by: Bharat Bhushan <Bharat.Bhushan@nxp.com>
-Signed-off-by: Jianhua Xie <jianhua.xie@nxp.com>
-Signed-off-by: Horia Geantă <horia.geanta@nxp.com>
-Signed-off-by: Yangbo Lu <yangbo.lu@nxp.com>
----
- arch/arm/include/asm/delay.h | 16 +++++++
- arch/arm/include/asm/dma-mapping.h | 6 ---
- arch/arm/include/asm/io.h | 31 +++++++++++++
- arch/arm/include/asm/mach/map.h | 4 +-
- arch/arm/include/asm/pgtable.h | 7 +++
- arch/arm/kernel/bios32.c | 43 ++++++++++++++++++
- arch/arm/mm/dma-mapping.c | 1 +
- arch/arm/mm/ioremap.c | 7 +++
- arch/arm/mm/mmu.c | 9 ++++
- arch/arm64/include/asm/cache.h | 2 +-
- arch/arm64/include/asm/io.h | 30 +++++++++++++
- arch/arm64/include/asm/pci.h | 4 ++
- arch/arm64/include/asm/pgtable-prot.h | 2 +
- arch/arm64/include/asm/pgtable.h | 5 +++
- arch/arm64/kernel/pci.c | 62 ++++++++++++++++++++++++++
- arch/arm64/mm/dma-mapping.c | 6 +++
- arch/powerpc/include/asm/dma-mapping.h | 5 ---
- arch/tile/include/asm/dma-mapping.h | 5 ---
- 18 files changed, 226 insertions(+), 19 deletions(-)
-
---- a/arch/arm/include/asm/delay.h
-+++ b/arch/arm/include/asm/delay.h
-@@ -57,6 +57,22 @@ extern void __bad_udelay(void);
- __const_udelay((n) * UDELAY_MULT)) : \
- __udelay(n))
-
-+#define spin_event_timeout(condition, timeout, delay) \
-+({ \
-+ typeof(condition) __ret; \
-+ int i = 0; \
-+ while (!(__ret = (condition)) && (i++ < timeout)) { \
-+ if (delay) \
-+ udelay(delay); \
-+ else \
-+ cpu_relax(); \
-+ udelay(1); \
-+ } \
-+ if (!__ret) \
-+ __ret = (condition); \
-+ __ret; \
-+})
-+
- /* Loop-based definitions for assembly code. */
- extern void __loop_delay(unsigned long loops);
- extern void __loop_udelay(unsigned long usecs);
---- a/arch/arm/include/asm/dma-mapping.h
-+++ b/arch/arm/include/asm/dma-mapping.h
-@@ -31,12 +31,6 @@ static inline struct dma_map_ops *get_dm
- return __generic_dma_ops(dev);
- }
-
--static inline void set_dma_ops(struct device *dev, struct dma_map_ops *ops)
--{
-- BUG_ON(!dev);
-- dev->archdata.dma_ops = ops;
--}
--
- #define HAVE_ARCH_DMA_SUPPORTED 1
- extern int dma_supported(struct device *dev, u64 mask);
-
---- a/arch/arm/include/asm/io.h
-+++ b/arch/arm/include/asm/io.h
-@@ -129,6 +129,7 @@ static inline u32 __raw_readl(const vola
- #define MT_DEVICE_NONSHARED 1
- #define MT_DEVICE_CACHED 2
- #define MT_DEVICE_WC 3
-+#define MT_MEMORY_RW_NS 4
- /*
- * types 4 onwards can be found in asm/mach/map.h and are undefined
- * for ioremap
-@@ -220,6 +221,34 @@ extern int pci_ioremap_io(unsigned int o
- #endif
- #endif
-
-+/* access ports */
-+#define setbits32(_addr, _v) iowrite32be(ioread32be(_addr) | (_v), (_addr))
-+#define clrbits32(_addr, _v) iowrite32be(ioread32be(_addr) & ~(_v), (_addr))
-+
-+#define setbits16(_addr, _v) iowrite16be(ioread16be(_addr) | (_v), (_addr))
-+#define clrbits16(_addr, _v) iowrite16be(ioread16be(_addr) & ~(_v), (_addr))
-+
-+#define setbits8(_addr, _v) iowrite8(ioread8(_addr) | (_v), (_addr))
-+#define clrbits8(_addr, _v) iowrite8(ioread8(_addr) & ~(_v), (_addr))
-+
-+/* Clear and set bits in one shot. These macros can be used to clear and
-+ * set multiple bits in a register using a single read-modify-write. These
-+ * macros can also be used to set a multiple-bit bit pattern using a mask,
-+ * by specifying the mask in the 'clear' parameter and the new bit pattern
-+ * in the 'set' parameter.
-+ */
-+
-+#define clrsetbits_be32(addr, clear, set) \
-+ iowrite32be((ioread32be(addr) & ~(clear)) | (set), (addr))
-+#define clrsetbits_le32(addr, clear, set) \
-+ iowrite32le((ioread32le(addr) & ~(clear)) | (set), (addr))
-+#define clrsetbits_be16(addr, clear, set) \
-+ iowrite16be((ioread16be(addr) & ~(clear)) | (set), (addr))
-+#define clrsetbits_le16(addr, clear, set) \
-+ iowrite16le((ioread16le(addr) & ~(clear)) | (set), (addr))
-+#define clrsetbits_8(addr, clear, set) \
-+ iowrite8((ioread8(addr) & ~(clear)) | (set), (addr))
-+
- /*
- * IO port access primitives
- * -------------------------
-@@ -408,6 +437,8 @@ void __iomem *ioremap_wc(resource_size_t
- #define ioremap_wc ioremap_wc
- #define ioremap_wt ioremap_wc
-
-+void __iomem *ioremap_cache_ns(resource_size_t res_cookie, size_t size);
-+
- void iounmap(volatile void __iomem *iomem_cookie);
- #define iounmap iounmap
-
---- a/arch/arm/include/asm/mach/map.h
-+++ b/arch/arm/include/asm/mach/map.h
-@@ -21,9 +21,9 @@ struct map_desc {
- unsigned int type;
- };
-
--/* types 0-3 are defined in asm/io.h */
-+/* types 0-4 are defined in asm/io.h */
- enum {
-- MT_UNCACHED = 4,
-+ MT_UNCACHED = 5,
- MT_CACHECLEAN,
- MT_MINICLEAN,
- MT_LOW_VECTORS,
---- a/arch/arm/include/asm/pgtable.h
-+++ b/arch/arm/include/asm/pgtable.h
-@@ -118,6 +118,13 @@ extern pgprot_t pgprot_s2_device;
- #define pgprot_noncached(prot) \
- __pgprot_modify(prot, L_PTE_MT_MASK, L_PTE_MT_UNCACHED)
-
-+#define pgprot_cached(prot) \
-+ __pgprot_modify(prot, L_PTE_MT_MASK, L_PTE_MT_DEV_CACHED)
-+
-+#define pgprot_cached_ns(prot) \
-+ __pgprot_modify(prot, L_PTE_MT_MASK, L_PTE_MT_DEV_CACHED | \
-+ L_PTE_MT_DEV_NONSHARED)
-+
- #define pgprot_writecombine(prot) \
- __pgprot_modify(prot, L_PTE_MT_MASK, L_PTE_MT_BUFFERABLE)
-
---- a/arch/arm/kernel/bios32.c
-+++ b/arch/arm/kernel/bios32.c
-@@ -11,6 +11,8 @@
- #include <linux/slab.h>
- #include <linux/init.h>
- #include <linux/io.h>
-+#include <linux/of_irq.h>
-+#include <linux/pcieport_if.h>
-
- #include <asm/mach-types.h>
- #include <asm/mach/map.h>
-@@ -64,6 +66,47 @@ void pcibios_report_status(u_int status_
- }
-
- /*
-+ * Check device tree if the service interrupts are there
-+ */
-+int pcibios_check_service_irqs(struct pci_dev *dev, int *irqs, int mask)
-+{
-+ int ret, count = 0;
-+ struct device_node *np = NULL;
-+
-+ if (dev->bus->dev.of_node)
-+ np = dev->bus->dev.of_node;
-+
-+ if (np == NULL)
-+ return 0;
-+
-+ if (!IS_ENABLED(CONFIG_OF_IRQ))
-+ return 0;
-+
-+ /* If root port doesn't support MSI/MSI-X/INTx in RC mode,
-+ * request irq for aer
-+ */
-+ if (mask & PCIE_PORT_SERVICE_AER) {
-+ ret = of_irq_get_byname(np, "aer");
-+ if (ret > 0) {
-+ irqs[PCIE_PORT_SERVICE_AER_SHIFT] = ret;
-+ count++;
-+ }
-+ }
-+
-+ if (mask & PCIE_PORT_SERVICE_PME) {
-+ ret = of_irq_get_byname(np, "pme");
-+ if (ret > 0) {
-+ irqs[PCIE_PORT_SERVICE_PME_SHIFT] = ret;
-+ count++;
-+ }
-+ }
-+
-+ /* TODO: add more service interrupts if there it is in the device tree*/
-+
-+ return count;
-+}
-+
-+/*
- * We don't use this to fix the device, but initialisation of it.
- * It's not the correct use for this, but it works.
- * Note that the arbiter/ISA bridge appears to be buggy, specifically in
---- a/arch/arm/mm/dma-mapping.c
-+++ b/arch/arm/mm/dma-mapping.c
-@@ -2410,6 +2410,7 @@ void arch_setup_dma_ops(struct device *d
-
- set_dma_ops(dev, dma_ops);
- }
-+EXPORT_SYMBOL(arch_setup_dma_ops);
-
- void arch_teardown_dma_ops(struct device *dev)
- {
---- a/arch/arm/mm/ioremap.c
-+++ b/arch/arm/mm/ioremap.c
-@@ -398,6 +398,13 @@ void __iomem *ioremap_wc(resource_size_t
- }
- EXPORT_SYMBOL(ioremap_wc);
-
-+void __iomem *ioremap_cache_ns(resource_size_t res_cookie, size_t size)
-+{
-+ return arch_ioremap_caller(res_cookie, size, MT_MEMORY_RW_NS,
-+ __builtin_return_address(0));
-+}
-+EXPORT_SYMBOL(ioremap_cache_ns);
-+
- /*
- * Remap an arbitrary physical address space into the kernel virtual
- * address space as memory. Needed when the kernel wants to execute
---- a/arch/arm/mm/mmu.c
-+++ b/arch/arm/mm/mmu.c
-@@ -313,6 +313,13 @@ static struct mem_type mem_types[] __ro_
- .prot_sect = PMD_TYPE_SECT | PMD_SECT_AP_WRITE,
- .domain = DOMAIN_KERNEL,
- },
-+ [MT_MEMORY_RW_NS] = {
-+ .prot_pte = L_PTE_PRESENT | L_PTE_YOUNG | L_PTE_DIRTY |
-+ L_PTE_XN,
-+ .prot_l1 = PMD_TYPE_TABLE,
-+ .prot_sect = PMD_TYPE_SECT | PMD_SECT_AP_WRITE | PMD_SECT_XN,
-+ .domain = DOMAIN_KERNEL,
-+ },
- [MT_ROM] = {
- .prot_sect = PMD_TYPE_SECT,
- .domain = DOMAIN_KERNEL,
-@@ -644,6 +651,7 @@ static void __init build_mem_type_table(
- }
- kern_pgprot |= PTE_EXT_AF;
- vecs_pgprot |= PTE_EXT_AF;
-+ mem_types[MT_MEMORY_RW_NS].prot_pte |= PTE_EXT_AF | cp->pte;
-
- /*
- * Set PXN for user mappings
-@@ -672,6 +680,7 @@ static void __init build_mem_type_table(
- mem_types[MT_MEMORY_RWX].prot_pte |= kern_pgprot;
- mem_types[MT_MEMORY_RW].prot_sect |= ecc_mask | cp->pmd;
- mem_types[MT_MEMORY_RW].prot_pte |= kern_pgprot;
-+ mem_types[MT_MEMORY_RW_NS].prot_sect |= ecc_mask | cp->pmd;
- mem_types[MT_MEMORY_DMA_READY].prot_pte |= kern_pgprot;
- mem_types[MT_MEMORY_RWX_NONCACHED].prot_sect |= ecc_mask;
- mem_types[MT_ROM].prot_sect |= cp->pmd;
---- a/arch/arm64/include/asm/cache.h
-+++ b/arch/arm64/include/asm/cache.h
-@@ -18,7 +18,7 @@
-
- #include <asm/cachetype.h>
-
--#define L1_CACHE_SHIFT 7
-+#define L1_CACHE_SHIFT 6
- #define L1_CACHE_BYTES (1 << L1_CACHE_SHIFT)
-
- /*
---- a/arch/arm64/include/asm/io.h
-+++ b/arch/arm64/include/asm/io.h
-@@ -171,6 +171,8 @@ extern void __iomem *ioremap_cache(phys_
- #define ioremap_nocache(addr, size) __ioremap((addr), (size), __pgprot(PROT_DEVICE_nGnRE))
- #define ioremap_wc(addr, size) __ioremap((addr), (size), __pgprot(PROT_NORMAL_NC))
- #define ioremap_wt(addr, size) __ioremap((addr), (size), __pgprot(PROT_DEVICE_nGnRE))
-+#define ioremap_cache_ns(addr, size) __ioremap((addr), (size), \
-+ __pgprot(PROT_NORMAL_NS))
- #define iounmap __iounmap
-
- /*
-@@ -184,6 +186,34 @@ extern void __iomem *ioremap_cache(phys_
- #define iowrite32be(v,p) ({ __iowmb(); __raw_writel((__force __u32)cpu_to_be32(v), p); })
- #define iowrite64be(v,p) ({ __iowmb(); __raw_writeq((__force __u64)cpu_to_be64(v), p); })
-
-+/* access ports */
-+#define setbits32(_addr, _v) iowrite32be(ioread32be(_addr) | (_v), (_addr))
-+#define clrbits32(_addr, _v) iowrite32be(ioread32be(_addr) & ~(_v), (_addr))
-+
-+#define setbits16(_addr, _v) iowrite16be(ioread16be(_addr) | (_v), (_addr))
-+#define clrbits16(_addr, _v) iowrite16be(ioread16be(_addr) & ~(_v), (_addr))
-+
-+#define setbits8(_addr, _v) iowrite8(ioread8(_addr) | (_v), (_addr))
-+#define clrbits8(_addr, _v) iowrite8(ioread8(_addr) & ~(_v), (_addr))
-+
-+/* Clear and set bits in one shot. These macros can be used to clear and
-+ * set multiple bits in a register using a single read-modify-write. These
-+ * macros can also be used to set a multiple-bit bit pattern using a mask,
-+ * by specifying the mask in the 'clear' parameter and the new bit pattern
-+ * in the 'set' parameter.
-+ */
-+
-+#define clrsetbits_be32(addr, clear, set) \
-+ iowrite32be((ioread32be(addr) & ~(clear)) | (set), (addr))
-+#define clrsetbits_le32(addr, clear, set) \
-+ iowrite32le((ioread32le(addr) & ~(clear)) | (set), (addr))
-+#define clrsetbits_be16(addr, clear, set) \
-+ iowrite16be((ioread16be(addr) & ~(clear)) | (set), (addr))
-+#define clrsetbits_le16(addr, clear, set) \
-+ iowrite16le((ioread16le(addr) & ~(clear)) | (set), (addr))
-+#define clrsetbits_8(addr, clear, set) \
-+ iowrite8((ioread8(addr) & ~(clear)) | (set), (addr))
-+
- #include <asm-generic/io.h>
-
- /*
---- a/arch/arm64/include/asm/pci.h
-+++ b/arch/arm64/include/asm/pci.h
-@@ -31,6 +31,10 @@ static inline int pci_get_legacy_ide_irq
- return -ENODEV;
- }
-
-+#define HAVE_PCI_MMAP
-+extern int pci_mmap_page_range(struct pci_dev *dev, struct vm_area_struct *vma,
-+ enum pci_mmap_state mmap_state,
-+ int write_combine);
- static inline int pci_proc_domain(struct pci_bus *bus)
- {
- return 1;
---- a/arch/arm64/include/asm/pgtable-prot.h
-+++ b/arch/arm64/include/asm/pgtable-prot.h
-@@ -48,6 +48,7 @@
- #define PROT_NORMAL_NC (PROT_DEFAULT | PTE_PXN | PTE_UXN | PTE_DIRTY | PTE_WRITE | PTE_ATTRINDX(MT_NORMAL_NC))
- #define PROT_NORMAL_WT (PROT_DEFAULT | PTE_PXN | PTE_UXN | PTE_DIRTY | PTE_WRITE | PTE_ATTRINDX(MT_NORMAL_WT))
- #define PROT_NORMAL (PROT_DEFAULT | PTE_PXN | PTE_UXN | PTE_DIRTY | PTE_WRITE | PTE_ATTRINDX(MT_NORMAL))
-+#define PROT_NORMAL_NS (PTE_TYPE_PAGE | PTE_AF | PTE_PXN | PTE_UXN | PTE_DIRTY | PTE_WRITE | PTE_ATTRINDX(MT_NORMAL))
-
- #define PROT_SECT_DEVICE_nGnRE (PROT_SECT_DEFAULT | PMD_SECT_PXN | PMD_SECT_UXN | PMD_ATTRINDX(MT_DEVICE_nGnRE))
- #define PROT_SECT_NORMAL (PROT_SECT_DEFAULT | PMD_SECT_PXN | PMD_SECT_UXN | PMD_ATTRINDX(MT_NORMAL))
-@@ -68,6 +69,7 @@
- #define PAGE_HYP_DEVICE __pgprot(PROT_DEVICE_nGnRE | PTE_HYP)
-
- #define PAGE_S2 __pgprot(_PROT_DEFAULT | PTE_S2_MEMATTR(MT_S2_NORMAL) | PTE_S2_RDONLY)
-+#define PAGE_S2_NS __pgprot(PTE_S2_MEMATTR(MT_S2_NORMAL) | PTE_S2_RDWR | PTE_TYPE_PAGE | PTE_AF)
- #define PAGE_S2_DEVICE __pgprot(_PROT_DEFAULT | PTE_S2_MEMATTR(MT_S2_DEVICE_nGnRE) | PTE_S2_RDONLY | PTE_UXN)
-
- #define PAGE_NONE __pgprot(((_PAGE_DEFAULT) & ~PTE_VALID) | PTE_PROT_NONE | PTE_NG | PTE_PXN | PTE_UXN)
---- a/arch/arm64/include/asm/pgtable.h
-+++ b/arch/arm64/include/asm/pgtable.h
-@@ -370,6 +370,11 @@ static inline int pmd_protnone(pmd_t pmd
- __pgprot_modify(prot, PTE_ATTRINDX_MASK, PTE_ATTRINDX(MT_DEVICE_nGnRnE) | PTE_PXN | PTE_UXN)
- #define pgprot_writecombine(prot) \
- __pgprot_modify(prot, PTE_ATTRINDX_MASK, PTE_ATTRINDX(MT_NORMAL_NC) | PTE_PXN | PTE_UXN)
-+#define pgprot_cached(prot) \
-+ __pgprot_modify(prot, PTE_ATTRINDX_MASK, PTE_ATTRINDX(MT_NORMAL) | \
-+ PTE_PXN | PTE_UXN)
-+#define pgprot_cached_ns(prot) \
-+ __pgprot(pgprot_val(pgprot_cached(prot)) ^ PTE_SHARED)
- #define pgprot_device(prot) \
- __pgprot_modify(prot, PTE_ATTRINDX_MASK, PTE_ATTRINDX(MT_DEVICE_nGnRE) | PTE_PXN | PTE_UXN)
- #define __HAVE_PHYS_MEM_ACCESS_PROT
---- a/arch/arm64/kernel/pci.c
-+++ b/arch/arm64/kernel/pci.c
-@@ -17,6 +17,8 @@
- #include <linux/mm.h>
- #include <linux/of_pci.h>
- #include <linux/of_platform.h>
-+#include <linux/of_irq.h>
-+#include <linux/pcieport_if.h>
- #include <linux/pci.h>
- #include <linux/pci-acpi.h>
- #include <linux/pci-ecam.h>
-@@ -53,6 +55,66 @@ int pcibios_alloc_irq(struct pci_dev *de
-
- return 0;
- }
-+
-+/*
-+ * Check device tree if the service interrupts are there
-+ */
-+int pcibios_check_service_irqs(struct pci_dev *dev, int *irqs, int mask)
-+{
-+ int ret, count = 0;
-+ struct device_node *np = NULL;
-+
-+ if (dev->bus->dev.of_node)
-+ np = dev->bus->dev.of_node;
-+
-+ if (np == NULL)
-+ return 0;
-+
-+ if (!IS_ENABLED(CONFIG_OF_IRQ))
-+ return 0;
-+
-+ /* If root port doesn't support MSI/MSI-X/INTx in RC mode,
-+ * request irq for aer
-+ */
-+ if (mask & PCIE_PORT_SERVICE_AER) {
-+ ret = of_irq_get_byname(np, "aer");
-+ if (ret > 0) {
-+ irqs[PCIE_PORT_SERVICE_AER_SHIFT] = ret;
-+ count++;
-+ }
-+ }
-+
-+ if (mask & PCIE_PORT_SERVICE_PME) {
-+ ret = of_irq_get_byname(np, "pme");
-+ if (ret > 0) {
-+ irqs[PCIE_PORT_SERVICE_PME_SHIFT] = ret;
-+ count++;
-+ }
-+ }
-+
-+ /* TODO: add more service interrupts if there it is in the device tree*/
-+
-+ return count;
-+}
-+
-+int pci_mmap_page_range(struct pci_dev *dev, struct vm_area_struct *vma,
-+ enum pci_mmap_state mmap_state, int write_combine)
-+{
-+ if (mmap_state == pci_mmap_io)
-+ return -EINVAL;
-+
-+ /*
-+ * Mark this as IO
-+ */
-+ vma->vm_page_prot = pgprot_noncached(vma->vm_page_prot);
-+
-+ if (remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff,
-+ vma->vm_end - vma->vm_start,
-+ vma->vm_page_prot))
-+ return -EAGAIN;
-+
-+ return 0;
-+}
-
- /*
- * raw_pci_read/write - Platform-specific PCI config space access.
---- a/arch/arm64/mm/dma-mapping.c
-+++ b/arch/arm64/mm/dma-mapping.c
-@@ -30,6 +30,7 @@
- #include <linux/swiotlb.h>
-
- #include <asm/cacheflush.h>
-+#include <linux/fsl/mc.h>
-
- static int swiotlb __ro_after_init;
-
-@@ -925,6 +926,10 @@ static int __init __iommu_dma_init(void)
- if (!ret)
- ret = register_iommu_dma_ops_notifier(&pci_bus_type);
- #endif
-+#ifdef CONFIG_FSL_MC_BUS
-+ if (!ret)
-+ ret = register_iommu_dma_ops_notifier(&fsl_mc_bus_type);
-+#endif
- return ret;
- }
- arch_initcall(__iommu_dma_init);
-@@ -978,3 +983,4 @@ void arch_setup_dma_ops(struct device *d
- dev->archdata.dma_coherent = coherent;
- __iommu_setup_dma_ops(dev, dma_base, size, iommu);
- }
-+EXPORT_SYMBOL(arch_setup_dma_ops);
---- a/arch/powerpc/include/asm/dma-mapping.h
-+++ b/arch/powerpc/include/asm/dma-mapping.h
-@@ -91,11 +91,6 @@ static inline struct dma_map_ops *get_dm
- return dev->archdata.dma_ops;
- }
-
--static inline void set_dma_ops(struct device *dev, struct dma_map_ops *ops)
--{
-- dev->archdata.dma_ops = ops;
--}
--
- /*
- * get_dma_offset()
- *
---- a/arch/tile/include/asm/dma-mapping.h
-+++ b/arch/tile/include/asm/dma-mapping.h
-@@ -59,11 +59,6 @@ static inline phys_addr_t dma_to_phys(st
-
- static inline void dma_mark_clean(void *addr, size_t size) {}
-
--static inline void set_dma_ops(struct device *dev, struct dma_map_ops *ops)
--{
-- dev->archdata.dma_ops = ops;
--}
--
- static inline bool dma_capable(struct device *dev, dma_addr_t addr, size_t size)
- {
- if (!dev->dma_mask)
diff --git a/target/linux/layerscape/patches-4.9/302-dts-support-layercape.patch b/target/linux/layerscape/patches-4.9/302-dts-support-layercape.patch
deleted file mode 100644
index 36854f1937..0000000000
--- a/target/linux/layerscape/patches-4.9/302-dts-support-layercape.patch
+++ /dev/null
@@ -1,10549 +0,0 @@
-From 2ba4c76bc645b7b4ff04364f294f3022d369108a Mon Sep 17 00:00:00 2001
-From: Yangbo Lu <yangbo.lu@nxp.com>
-Date: Thu, 5 Jul 2018 16:20:56 +0800
-Subject: [PATCH 04/32] dts: support layercape
-
-This is an integrated patch for layerscape dts support.
-
-Signed-off-by: Amrita Kumari <amrita.kumari@nxp.com>
-Signed-off-by: Alison Wang <b18965@freescale.com>
-Signed-off-by: Li Yang <leoyang.li@nxp.com>
-Signed-off-by: Ashish Kumar <Ashish.Kumar@nxp.com>
-Signed-off-by: Zhao Qiang <qiang.zhao@nxp.com>
-Signed-off-by: Rajesh Bhagat <rajesh.bhagat@freescale.com>
-Signed-off-by: Zhang Ying-22455 <ying.zhang22455@nxp.com>
-Signed-off-by: Madalin Bucur <madalin.bucur@nxp.com>
-Signed-off-by: Minghuan Lian <Minghuan.Lian@nxp.com>
-Signed-off-by: Suresh Gupta <suresh.gupta@nxp.com>
-Signed-off-by: Chenhui Zhao <chenhui.zhao@nxp.com>
-Signed-off-by: Priyanka Jain <priyanka.jain@nxp.com>
-Signed-off-by: Hou Zhiqiang <Zhiqiang.Hou@nxp.com>
-Signed-off-by: Changming Huang <jerry.huang@nxp.com>
-Signed-off-by: Bharat Bhushan <Bharat.Bhushan@nxp.com>
-Signed-off-by: Meng Yi <meng.yi@nxp.com>
-Signed-off-by: Shaohui Xie <Shaohui.Xie@nxp.com>
-Signed-off-by: Marc Zyngier <marc.zyngier@arm.com>
-Signed-off-by: Prabhakar Kushwaha <prabhakar.kushwaha@nxp.com>
-Signed-off-by: Ran Wang <ran.wang_1@nxp.com>
-Signed-off-by: Yangbo Lu <yangbo.lu@nxp.com>
----
- arch/arm/boot/dts/alpine.dtsi | 2 +-
- arch/arm/boot/dts/axm55xx.dtsi | 2 +-
- arch/arm/boot/dts/ecx-2000.dts | 2 +-
- arch/arm/boot/dts/imx6ul.dtsi | 4 +-
- arch/arm/boot/dts/keystone.dtsi | 4 +-
- arch/arm/boot/dts/ls1021a-qds.dts | 26 +
- arch/arm/boot/dts/ls1021a-twr.dts | 25 +
- arch/arm/boot/dts/ls1021a.dtsi | 284 ++++--
- arch/arm/boot/dts/mt6580.dtsi | 2 +-
- arch/arm/boot/dts/mt6589.dtsi | 2 +-
- arch/arm/boot/dts/mt8127.dtsi | 2 +-
- arch/arm/boot/dts/mt8135.dtsi | 2 +-
- arch/arm/boot/dts/rk3288.dtsi | 2 +-
- arch/arm/boot/dts/sun6i-a31.dtsi | 2 +-
- arch/arm/boot/dts/sun7i-a20.dtsi | 4 +-
- arch/arm/boot/dts/sun8i-a23-a33.dtsi | 2 +-
- arch/arm/boot/dts/sun9i-a80.dtsi | 2 +-
- arch/arm64/boot/dts/freescale/Makefile | 18 +
- .../boot/dts/freescale/fsl-ls1012a-2g5rdb.dts | 123 +++
- .../boot/dts/freescale/fsl-ls1012a-frdm.dts | 141 +++
- .../boot/dts/freescale/fsl-ls1012a-frwy.dts | 177 ++++
- .../boot/dts/freescale/fsl-ls1012a-qds.dts | 166 ++++
- .../boot/dts/freescale/fsl-ls1012a-rdb.dts | 102 ++
- .../arm64/boot/dts/freescale/fsl-ls1012a.dtsi | 567 +++++++++++
- .../boot/dts/freescale/fsl-ls1043-post.dtsi | 44 +
- .../dts/freescale/fsl-ls1043a-qds-sdk.dts | 71 ++
- .../boot/dts/freescale/fsl-ls1043a-qds.dts | 210 ++++-
- .../dts/freescale/fsl-ls1043a-rdb-sdk.dts | 71 ++
- .../dts/freescale/fsl-ls1043a-rdb-usdpaa.dts | 117 +++
- .../boot/dts/freescale/fsl-ls1043a-rdb.dts | 152 ++-
- .../arm64/boot/dts/freescale/fsl-ls1043a.dtsi | 357 +++++--
- .../boot/dts/freescale/fsl-ls1046-post.dtsi | 48 +
- .../dts/freescale/fsl-ls1046a-qds-sdk.dts | 112 +++
- .../boot/dts/freescale/fsl-ls1046a-qds.dts | 326 +++++++
- .../dts/freescale/fsl-ls1046a-rdb-sdk.dts | 85 ++
- .../dts/freescale/fsl-ls1046a-rdb-usdpaa.dts | 110 +++
- .../boot/dts/freescale/fsl-ls1046a-rdb.dts | 181 ++++
- .../arm64/boot/dts/freescale/fsl-ls1046a.dtsi | 771 +++++++++++++++
- .../boot/dts/freescale/fsl-ls1088a-qds.dts | 137 +++
- .../boot/dts/freescale/fsl-ls1088a-rdb.dts | 200 ++++
- .../arm64/boot/dts/freescale/fsl-ls1088a.dtsi | 800 ++++++++++++++++
- .../boot/dts/freescale/fsl-ls2080a-qds.dts | 229 +----
- .../boot/dts/freescale/fsl-ls2080a-rdb.dts | 207 ++--
- .../boot/dts/freescale/fsl-ls2080a-simu.dts | 47 +-
- .../arm64/boot/dts/freescale/fsl-ls2080a.dtsi | 803 ++--------------
- .../boot/dts/freescale/fsl-ls2081a-rdb.dts | 161 ++++
- .../boot/dts/freescale/fsl-ls2088a-qds.dts | 126 +++
- .../boot/dts/freescale/fsl-ls2088a-rdb.dts | 104 ++
- .../arm64/boot/dts/freescale/fsl-ls2088a.dtsi | 159 ++++
- .../boot/dts/freescale/fsl-ls208xa-qds.dtsi | 162 ++++
- .../boot/dts/freescale/fsl-ls208xa-rdb.dtsi | 136 +++
- .../arm64/boot/dts/freescale/fsl-ls208xa.dtsi | 885 ++++++++++++++++++
- .../dts/freescale/qoriq-bman-portals-sdk.dtsi | 55 ++
- .../dts/freescale/qoriq-bman-portals.dtsi | 77 ++
- .../boot/dts/freescale/qoriq-dpaa-eth.dtsi | 73 ++
- .../dts/freescale/qoriq-fman3-0-10g-0.dtsi | 43 +
- .../dts/freescale/qoriq-fman3-0-10g-1.dtsi | 43 +
- .../dts/freescale/qoriq-fman3-0-1g-0.dtsi | 42 +
- .../dts/freescale/qoriq-fman3-0-1g-1.dtsi | 42 +
- .../dts/freescale/qoriq-fman3-0-1g-2.dtsi | 42 +
- .../dts/freescale/qoriq-fman3-0-1g-3.dtsi | 42 +
- .../dts/freescale/qoriq-fman3-0-1g-4.dtsi | 42 +
- .../dts/freescale/qoriq-fman3-0-1g-5.dtsi | 42 +
- .../boot/dts/freescale/qoriq-fman3-0-6oh.dtsi | 47 +
- .../boot/dts/freescale/qoriq-fman3-0.dtsi | 130 +++
- .../dts/freescale/qoriq-qman-portals-sdk.dtsi | 38 +
- .../dts/freescale/qoriq-qman-portals.dtsi | 87 ++
- .../boot/dts/fsl/qoriq-bman1-portals.dtsi | 10 +
- .../boot/dts/fsl/qoriq-fman3-0-10g-0.dtsi | 4 +-
- .../boot/dts/fsl/qoriq-fman3-0-10g-1.dtsi | 4 +-
- 70 files changed, 8051 insertions(+), 1286 deletions(-)
- create mode 100644 arch/arm64/boot/dts/freescale/fsl-ls1012a-2g5rdb.dts
- create mode 100644 arch/arm64/boot/dts/freescale/fsl-ls1012a-frdm.dts
- create mode 100644 arch/arm64/boot/dts/freescale/fsl-ls1012a-frwy.dts
- create mode 100644 arch/arm64/boot/dts/freescale/fsl-ls1012a-qds.dts
- create mode 100644 arch/arm64/boot/dts/freescale/fsl-ls1012a-rdb.dts
- create mode 100644 arch/arm64/boot/dts/freescale/fsl-ls1012a.dtsi
- create mode 100644 arch/arm64/boot/dts/freescale/fsl-ls1043-post.dtsi
- create mode 100644 arch/arm64/boot/dts/freescale/fsl-ls1043a-qds-sdk.dts
- create mode 100644 arch/arm64/boot/dts/freescale/fsl-ls1043a-rdb-sdk.dts
- create mode 100644 arch/arm64/boot/dts/freescale/fsl-ls1043a-rdb-usdpaa.dts
- create mode 100644 arch/arm64/boot/dts/freescale/fsl-ls1046-post.dtsi
- create mode 100644 arch/arm64/boot/dts/freescale/fsl-ls1046a-qds-sdk.dts
- create mode 100644 arch/arm64/boot/dts/freescale/fsl-ls1046a-qds.dts
- create mode 100644 arch/arm64/boot/dts/freescale/fsl-ls1046a-rdb-sdk.dts
- create mode 100644 arch/arm64/boot/dts/freescale/fsl-ls1046a-rdb-usdpaa.dts
- create mode 100644 arch/arm64/boot/dts/freescale/fsl-ls1046a-rdb.dts
- create mode 100644 arch/arm64/boot/dts/freescale/fsl-ls1046a.dtsi
- create mode 100644 arch/arm64/boot/dts/freescale/fsl-ls1088a-qds.dts
- create mode 100644 arch/arm64/boot/dts/freescale/fsl-ls1088a-rdb.dts
- create mode 100644 arch/arm64/boot/dts/freescale/fsl-ls1088a.dtsi
- create mode 100644 arch/arm64/boot/dts/freescale/fsl-ls2081a-rdb.dts
- create mode 100644 arch/arm64/boot/dts/freescale/fsl-ls2088a-qds.dts
- create mode 100644 arch/arm64/boot/dts/freescale/fsl-ls2088a-rdb.dts
- create mode 100644 arch/arm64/boot/dts/freescale/fsl-ls2088a.dtsi
- create mode 100644 arch/arm64/boot/dts/freescale/fsl-ls208xa-qds.dtsi
- create mode 100644 arch/arm64/boot/dts/freescale/fsl-ls208xa-rdb.dtsi
- create mode 100644 arch/arm64/boot/dts/freescale/fsl-ls208xa.dtsi
- create mode 100644 arch/arm64/boot/dts/freescale/qoriq-bman-portals-sdk.dtsi
- create mode 100644 arch/arm64/boot/dts/freescale/qoriq-bman-portals.dtsi
- create mode 100644 arch/arm64/boot/dts/freescale/qoriq-dpaa-eth.dtsi
- create mode 100644 arch/arm64/boot/dts/freescale/qoriq-fman3-0-10g-0.dtsi
- create mode 100644 arch/arm64/boot/dts/freescale/qoriq-fman3-0-10g-1.dtsi
- create mode 100644 arch/arm64/boot/dts/freescale/qoriq-fman3-0-1g-0.dtsi
- create mode 100644 arch/arm64/boot/dts/freescale/qoriq-fman3-0-1g-1.dtsi
- create mode 100644 arch/arm64/boot/dts/freescale/qoriq-fman3-0-1g-2.dtsi
- create mode 100644 arch/arm64/boot/dts/freescale/qoriq-fman3-0-1g-3.dtsi
- create mode 100644 arch/arm64/boot/dts/freescale/qoriq-fman3-0-1g-4.dtsi
- create mode 100644 arch/arm64/boot/dts/freescale/qoriq-fman3-0-1g-5.dtsi
- create mode 100644 arch/arm64/boot/dts/freescale/qoriq-fman3-0-6oh.dtsi
- create mode 100644 arch/arm64/boot/dts/freescale/qoriq-fman3-0.dtsi
- create mode 100644 arch/arm64/boot/dts/freescale/qoriq-qman-portals-sdk.dtsi
- create mode 100644 arch/arm64/boot/dts/freescale/qoriq-qman-portals.dtsi
-
---- a/arch/arm/boot/dts/alpine.dtsi
-+++ b/arch/arm/boot/dts/alpine.dtsi
-@@ -93,7 +93,7 @@
- interrupt-controller;
- reg = <0x0 0xfb001000 0x0 0x1000>,
- <0x0 0xfb002000 0x0 0x2000>,
-- <0x0 0xfb004000 0x0 0x1000>,
-+ <0x0 0xfb004000 0x0 0x2000>,
- <0x0 0xfb006000 0x0 0x2000>;
- interrupts =
- <GIC_PPI 9 (GIC_CPU_MASK_SIMPLE(4) | IRQ_TYPE_LEVEL_HIGH)>;
---- a/arch/arm/boot/dts/axm55xx.dtsi
-+++ b/arch/arm/boot/dts/axm55xx.dtsi
-@@ -62,7 +62,7 @@
- #address-cells = <0>;
- interrupt-controller;
- reg = <0x20 0x01001000 0 0x1000>,
-- <0x20 0x01002000 0 0x1000>,
-+ <0x20 0x01002000 0 0x2000>,
- <0x20 0x01004000 0 0x2000>,
- <0x20 0x01006000 0 0x2000>;
- interrupts = <GIC_PPI 9 (GIC_CPU_MASK_SIMPLE(4) |
---- a/arch/arm/boot/dts/ecx-2000.dts
-+++ b/arch/arm/boot/dts/ecx-2000.dts
-@@ -99,7 +99,7 @@
- interrupt-controller;
- interrupts = <1 9 0xf04>;
- reg = <0xfff11000 0x1000>,
-- <0xfff12000 0x1000>,
-+ <0xfff12000 0x2000>,
- <0xfff14000 0x2000>,
- <0xfff16000 0x2000>;
- };
---- a/arch/arm/boot/dts/imx6ul.dtsi
-+++ b/arch/arm/boot/dts/imx6ul.dtsi
-@@ -89,11 +89,11 @@
- };
-
- intc: interrupt-controller@00a01000 {
-- compatible = "arm,cortex-a7-gic";
-+ compatible = "arm,gic-400", "arm,cortex-a7-gic";
- #interrupt-cells = <3>;
- interrupt-controller;
- reg = <0x00a01000 0x1000>,
-- <0x00a02000 0x1000>,
-+ <0x00a02000 0x2000>,
- <0x00a04000 0x2000>,
- <0x00a06000 0x2000>;
- };
---- a/arch/arm/boot/dts/keystone.dtsi
-+++ b/arch/arm/boot/dts/keystone.dtsi
-@@ -30,12 +30,12 @@
- };
-
- gic: interrupt-controller {
-- compatible = "arm,cortex-a15-gic";
-+ compatible = "arm,gic-400", "arm,cortex-a15-gic";
- #interrupt-cells = <3>;
- interrupt-controller;
- reg = <0x0 0x02561000 0x0 0x1000>,
- <0x0 0x02562000 0x0 0x2000>,
-- <0x0 0x02564000 0x0 0x1000>,
-+ <0x0 0x02564000 0x0 0x2000>,
- <0x0 0x02566000 0x0 0x2000>;
- interrupts = <GIC_PPI 9 (GIC_CPU_MASK_SIMPLE(4) |
- IRQ_TYPE_LEVEL_HIGH)>;
---- a/arch/arm/boot/dts/ls1021a-qds.dts
-+++ b/arch/arm/boot/dts/ls1021a-qds.dts
-@@ -124,6 +124,19 @@
- };
- };
-
-+&qspi {
-+ num-cs = <2>;
-+ status = "okay";
-+
-+ qflash0: s25fl128s@0 {
-+ compatible = "spansion,m25p80";
-+ #address-cells = <1>;
-+ #size-cells = <1>;
-+ spi-max-frequency = <20000000>;
-+ reg = <0>;
-+ };
-+};
-+
- &enet0 {
- tbi-handle = <&tbi0>;
- phy-handle = <&sgmii_phy1c>;
-@@ -239,6 +252,11 @@
- device-width = <1>;
- };
-
-+ nand@2,0 {
-+ compatible = "fsl,ifc-nand";
-+ reg = <0x2 0x0 0x10000>;
-+ };
-+
- fpga: board-control@3,0 {
- #address-cells = <1>;
- #size-cells = <1>;
-@@ -331,3 +349,11 @@
- &uart1 {
- status = "okay";
- };
-+
-+&can0 {
-+ status = "okay";
-+};
-+
-+&can1 {
-+ status = "okay";
-+};
---- a/arch/arm/boot/dts/ls1021a-twr.dts
-+++ b/arch/arm/boot/dts/ls1021a-twr.dts
-@@ -142,6 +142,19 @@
- };
- };
-
-+&qspi {
-+ num-cs = <2>;
-+ status = "okay";
-+
-+ qflash0: n25q128a13@0 {
-+ compatible = "n25q128a13", "jedec,spi-nor";
-+ #address-cells = <1>;
-+ #size-cells = <1>;
-+ spi-max-frequency = <20000000>;
-+ reg = <0>;
-+ };
-+};
-+
- &enet0 {
- tbi-handle = <&tbi1>;
- phy-handle = <&sgmii_phy2>;
-@@ -228,6 +241,10 @@
- };
- };
-
-+&esdhc {
-+ status = "okay";
-+};
-+
- &sai1 {
- status = "okay";
- };
-@@ -243,3 +260,11 @@
- &uart1 {
- status = "okay";
- };
-+
-+&can0 {
-+ status = "okay";
-+};
-+
-+&can1 {
-+ status = "okay";
-+};
---- a/arch/arm/boot/dts/ls1021a.dtsi
-+++ b/arch/arm/boot/dts/ls1021a.dtsi
-@@ -47,6 +47,7 @@
-
- #include "skeleton64.dtsi"
- #include <dt-bindings/interrupt-controller/arm-gic.h>
-+#include <dt-bindings/thermal/thermal.h>
-
- / {
- compatible = "fsl,ls1021a";
-@@ -70,21 +71,29 @@
- #address-cells = <1>;
- #size-cells = <0>;
-
-- cpu@f00 {
-+ cpu0: @f00 {
- compatible = "arm,cortex-a7";
- device_type = "cpu";
- reg = <0xf00>;
-- clocks = <&cluster1_clk>;
-+ clocks = <&clockgen 1 0>;
-+ #cooling-cells = <2>;
- };
-
-- cpu@f01 {
-+ cpu1: @f01 {
- compatible = "arm,cortex-a7";
- device_type = "cpu";
- reg = <0xf01>;
-- clocks = <&cluster1_clk>;
-+ clocks = <&clockgen 1 0>;
- };
- };
-
-+ sysclk: sysclk {
-+ compatible = "fixed-clock";
-+ #clock-cells = <0>;
-+ clock-frequency = <100000000>;
-+ clock-output-names = "sysclk";
-+ };
-+
- timer {
- compatible = "arm,armv7-timer";
- interrupts = <GIC_PPI 13 (GIC_CPU_MASK_SIMPLE(2) | IRQ_TYPE_LEVEL_LOW)>,
-@@ -108,11 +117,11 @@
- ranges;
-
- gic: interrupt-controller@1400000 {
-- compatible = "arm,cortex-a7-gic";
-+ compatible = "arm,gic-400", "arm,cortex-a7-gic";
- #interrupt-cells = <3>;
- interrupt-controller;
- reg = <0x0 0x1401000 0x0 0x1000>,
-- <0x0 0x1402000 0x0 0x1000>,
-+ <0x0 0x1402000 0x0 0x2000>,
- <0x0 0x1404000 0x0 0x2000>,
- <0x0 0x1406000 0x0 0x2000>;
- interrupts = <GIC_PPI 9 (GIC_CPU_MASK_SIMPLE(2) | IRQ_TYPE_LEVEL_HIGH)>;
-@@ -120,14 +129,14 @@
- };
-
- msi1: msi-controller@1570e00 {
-- compatible = "fsl,1s1021a-msi";
-+ compatible = "fsl,ls1021a-msi";
- reg = <0x0 0x1570e00 0x0 0x8>;
- msi-controller;
- interrupts = <GIC_SPI 179 IRQ_TYPE_LEVEL_HIGH>;
- };
-
- msi2: msi-controller@1570e08 {
-- compatible = "fsl,1s1021a-msi";
-+ compatible = "fsl,ls1021a-msi";
- reg = <0x0 0x1570e08 0x0 0x8>;
- msi-controller;
- interrupts = <GIC_SPI 180 IRQ_TYPE_LEVEL_HIGH>;
-@@ -137,11 +146,12 @@
- compatible = "fsl,ifc", "simple-bus";
- reg = <0x0 0x1530000 0x0 0x10000>;
- interrupts = <GIC_SPI 75 IRQ_TYPE_LEVEL_HIGH>;
-+ big-endian;
- };
-
- dcfg: dcfg@1ee0000 {
- compatible = "fsl,ls1021a-dcfg", "syscon";
-- reg = <0x0 0x1ee0000 0x0 0x10000>;
-+ reg = <0x0 0x1ee0000 0x0 0x1000>;
- big-endian;
- };
-
-@@ -163,7 +173,7 @@
- <0x0 0x20220520 0x0 0x4>;
- reg-names = "ahci", "sata-ecc";
- interrupts = <GIC_SPI 101 IRQ_TYPE_LEVEL_HIGH>;
-- clocks = <&platform_clk 1>;
-+ clocks = <&clockgen 4 1>;
- dma-coherent;
- status = "disabled";
- };
-@@ -214,43 +224,89 @@
- };
-
- clockgen: clocking@1ee1000 {
-- #address-cells = <1>;
-- #size-cells = <1>;
-- ranges = <0x0 0x0 0x1ee1000 0x10000>;
-+ compatible = "fsl,ls1021a-clockgen";
-+ reg = <0x0 0x1ee1000 0x0 0x1000>;
-+ #clock-cells = <2>;
-+ clocks = <&sysclk>;
-+ };
-
-- sysclk: sysclk {
-- compatible = "fixed-clock";
-- #clock-cells = <0>;
-- clock-output-names = "sysclk";
-- };
--
-- cga_pll1: pll@800 {
-- compatible = "fsl,qoriq-core-pll-2.0";
-- #clock-cells = <1>;
-- reg = <0x800 0x10>;
-- clocks = <&sysclk>;
-- clock-output-names = "cga-pll1", "cga-pll1-div2",
-- "cga-pll1-div4";
-- };
--
-- platform_clk: pll@c00 {
-- compatible = "fsl,qoriq-core-pll-2.0";
-- #clock-cells = <1>;
-- reg = <0xc00 0x10>;
-- clocks = <&sysclk>;
-- clock-output-names = "platform-clk", "platform-clk-div2";
-- };
--
-- cluster1_clk: clk0c0@0 {
-- compatible = "fsl,qoriq-core-mux-2.0";
-- #clock-cells = <0>;
-- reg = <0x0 0x10>;
-- clock-names = "pll1cga", "pll1cga-div2", "pll1cga-div4";
-- clocks = <&cga_pll1 0>, <&cga_pll1 1>, <&cga_pll1 2>;
-- clock-output-names = "cluster1-clk";
-+ tmu: tmu@1f00000 {
-+ compatible = "fsl,qoriq-tmu";
-+ reg = <0x0 0x1f00000 0x0 0x10000>;
-+ interrupts = <GIC_SPI 61 IRQ_TYPE_LEVEL_HIGH>;
-+ fsl,tmu-range = <0xb0000 0xa0026 0x80048 0x30061>;
-+ fsl,tmu-calibration = <0x00000000 0x0000000f
-+ 0x00000001 0x00000017
-+ 0x00000002 0x0000001e
-+ 0x00000003 0x00000026
-+ 0x00000004 0x0000002e
-+ 0x00000005 0x00000035
-+ 0x00000006 0x0000003d
-+ 0x00000007 0x00000044
-+ 0x00000008 0x0000004c
-+ 0x00000009 0x00000053
-+ 0x0000000a 0x0000005b
-+ 0x0000000b 0x00000064
-+
-+ 0x00010000 0x00000011
-+ 0x00010001 0x0000001c
-+ 0x00010002 0x00000024
-+ 0x00010003 0x0000002b
-+ 0x00010004 0x00000034
-+ 0x00010005 0x00000039
-+ 0x00010006 0x00000042
-+ 0x00010007 0x0000004c
-+ 0x00010008 0x00000051
-+ 0x00010009 0x0000005a
-+ 0x0001000a 0x00000063
-+
-+ 0x00020000 0x00000013
-+ 0x00020001 0x00000019
-+ 0x00020002 0x00000024
-+ 0x00020003 0x0000002c
-+ 0x00020004 0x00000035
-+ 0x00020005 0x0000003d
-+ 0x00020006 0x00000046
-+ 0x00020007 0x00000050
-+ 0x00020008 0x00000059
-+
-+ 0x00030000 0x00000002
-+ 0x00030001 0x0000000d
-+ 0x00030002 0x00000019
-+ 0x00030003 0x00000024>;
-+ #thermal-sensor-cells = <1>;
-+ };
-+
-+ thermal-zones {
-+ cpu_thermal: cpu-thermal {
-+ polling-delay-passive = <1000>;
-+ polling-delay = <5000>;
-+
-+ thermal-sensors = <&tmu 0>;
-+
-+ trips {
-+ cpu_alert: cpu-alert {
-+ temperature = <85000>;
-+ hysteresis = <2000>;
-+ type = "passive";
-+ };
-+ cpu_crit: cpu-crit {
-+ temperature = <95000>;
-+ hysteresis = <2000>;
-+ type = "critical";
-+ };
-+ };
-+
-+ cooling-maps {
-+ map0 {
-+ trip = <&cpu_alert>;
-+ cooling-device =
-+ <&cpu0 THERMAL_NO_LIMIT
-+ THERMAL_NO_LIMIT>;
-+ };
-+ };
- };
- };
--
- dspi0: dspi@2100000 {
- compatible = "fsl,ls1021a-v1.0-dspi";
- #address-cells = <1>;
-@@ -258,7 +314,7 @@
- reg = <0x0 0x2100000 0x0 0x10000>;
- interrupts = <GIC_SPI 96 IRQ_TYPE_LEVEL_HIGH>;
- clock-names = "dspi";
-- clocks = <&platform_clk 1>;
-+ clocks = <&clockgen 4 1>;
- spi-num-chipselects = <6>;
- big-endian;
- status = "disabled";
-@@ -271,31 +327,48 @@
- reg = <0x0 0x2110000 0x0 0x10000>;
- interrupts = <GIC_SPI 97 IRQ_TYPE_LEVEL_HIGH>;
- clock-names = "dspi";
-- clocks = <&platform_clk 1>;
-+ clocks = <&clockgen 4 1>;
- spi-num-chipselects = <6>;
- big-endian;
- status = "disabled";
- };
-
-+ qspi: quadspi@1550000 {
-+ compatible = "fsl,ls1021a-qspi";
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ reg = <0x0 0x1550000 0x0 0x10000>,
-+ <0x0 0x40000000 0x0 0x4000000>;
-+ reg-names = "QuadSPI", "QuadSPI-memory";
-+ interrupts = <GIC_SPI 131 IRQ_TYPE_LEVEL_HIGH>;
-+ clock-names = "qspi_en", "qspi";
-+ clocks = <&clockgen 4 1>, <&clockgen 4 1>;
-+ big-endian;
-+ amba-base = <0x40000000>;
-+ status = "disabled";
-+ };
-+
- i2c0: i2c@2180000 {
-- compatible = "fsl,vf610-i2c";
-+ compatible = "fsl,vf610-i2c", "fsl,ls1021a-vf610-i2c";
- #address-cells = <1>;
- #size-cells = <0>;
- reg = <0x0 0x2180000 0x0 0x10000>;
- interrupts = <GIC_SPI 88 IRQ_TYPE_LEVEL_HIGH>;
- clock-names = "i2c";
-- clocks = <&platform_clk 1>;
-+ clocks = <&clockgen 4 1>;
-+ fsl-scl-gpio = <&gpio3 23 0>;
- status = "disabled";
- };
-
- i2c1: i2c@2190000 {
-- compatible = "fsl,vf610-i2c";
-+ compatible = "fsl,vf610-i2c", "fsl,ls1021a-vf610-i2c";
- #address-cells = <1>;
- #size-cells = <0>;
- reg = <0x0 0x2190000 0x0 0x10000>;
- interrupts = <GIC_SPI 89 IRQ_TYPE_LEVEL_HIGH>;
- clock-names = "i2c";
-- clocks = <&platform_clk 1>;
-+ clocks = <&clockgen 4 1>;
-+ fsl-scl-gpio = <&gpio3 23 0>;
- status = "disabled";
- };
-
-@@ -306,7 +379,7 @@
- reg = <0x0 0x21a0000 0x0 0x10000>;
- interrupts = <GIC_SPI 90 IRQ_TYPE_LEVEL_HIGH>;
- clock-names = "i2c";
-- clocks = <&platform_clk 1>;
-+ clocks = <&clockgen 4 1>;
- status = "disabled";
- };
-
-@@ -399,7 +472,7 @@
- compatible = "fsl,ls1021a-lpuart";
- reg = <0x0 0x2960000 0x0 0x1000>;
- interrupts = <GIC_SPI 81 IRQ_TYPE_LEVEL_HIGH>;
-- clocks = <&platform_clk 1>;
-+ clocks = <&clockgen 4 1>;
- clock-names = "ipg";
- status = "disabled";
- };
-@@ -408,7 +481,7 @@
- compatible = "fsl,ls1021a-lpuart";
- reg = <0x0 0x2970000 0x0 0x1000>;
- interrupts = <GIC_SPI 82 IRQ_TYPE_LEVEL_HIGH>;
-- clocks = <&platform_clk 1>;
-+ clocks = <&clockgen 4 1>;
- clock-names = "ipg";
- status = "disabled";
- };
-@@ -417,7 +490,7 @@
- compatible = "fsl,ls1021a-lpuart";
- reg = <0x0 0x2980000 0x0 0x1000>;
- interrupts = <GIC_SPI 83 IRQ_TYPE_LEVEL_HIGH>;
-- clocks = <&platform_clk 1>;
-+ clocks = <&clockgen 4 1>;
- clock-names = "ipg";
- status = "disabled";
- };
-@@ -426,7 +499,7 @@
- compatible = "fsl,ls1021a-lpuart";
- reg = <0x0 0x2990000 0x0 0x1000>;
- interrupts = <GIC_SPI 84 IRQ_TYPE_LEVEL_HIGH>;
-- clocks = <&platform_clk 1>;
-+ clocks = <&clockgen 4 1>;
- clock-names = "ipg";
- status = "disabled";
- };
-@@ -435,16 +508,26 @@
- compatible = "fsl,ls1021a-lpuart";
- reg = <0x0 0x29a0000 0x0 0x1000>;
- interrupts = <GIC_SPI 85 IRQ_TYPE_LEVEL_HIGH>;
-- clocks = <&platform_clk 1>;
-+ clocks = <&clockgen 4 1>;
- clock-names = "ipg";
- status = "disabled";
- };
-
-+ ftm0: ftm0@29d0000 {
-+ compatible = "fsl,ls1021a-ftm";
-+ reg = <0x0 0x29d0000 0x0 0x10000>,
-+ <0x0 0x1ee2140 0x0 0x4>;
-+ reg-names = "ftm", "FlexTimer1";
-+ interrupts = <GIC_SPI 118 IRQ_TYPE_LEVEL_HIGH>;
-+ big-endian;
-+ status = "okay";
-+ };
-+
- wdog0: watchdog@2ad0000 {
- compatible = "fsl,imx21-wdt";
- reg = <0x0 0x2ad0000 0x0 0x10000>;
- interrupts = <GIC_SPI 115 IRQ_TYPE_LEVEL_HIGH>;
-- clocks = <&platform_clk 1>;
-+ clocks = <&clockgen 4 1>;
- clock-names = "wdog-en";
- big-endian;
- };
-@@ -454,8 +537,8 @@
- compatible = "fsl,vf610-sai";
- reg = <0x0 0x2b50000 0x0 0x10000>;
- interrupts = <GIC_SPI 132 IRQ_TYPE_LEVEL_HIGH>;
-- clocks = <&platform_clk 1>, <&platform_clk 1>,
-- <&platform_clk 1>, <&platform_clk 1>;
-+ clocks = <&clockgen 4 1>, <&clockgen 4 1>,
-+ <&clockgen 4 1>, <&clockgen 4 1>;
- clock-names = "bus", "mclk1", "mclk2", "mclk3";
- dma-names = "tx", "rx";
- dmas = <&edma0 1 47>,
-@@ -468,8 +551,8 @@
- compatible = "fsl,vf610-sai";
- reg = <0x0 0x2b60000 0x0 0x10000>;
- interrupts = <GIC_SPI 133 IRQ_TYPE_LEVEL_HIGH>;
-- clocks = <&platform_clk 1>, <&platform_clk 1>,
-- <&platform_clk 1>, <&platform_clk 1>;
-+ clocks = <&clockgen 4 1>, <&clockgen 4 1>,
-+ <&clockgen 4 1>, <&clockgen 4 1>;
- clock-names = "bus", "mclk1", "mclk2", "mclk3";
- dma-names = "tx", "rx";
- dmas = <&edma0 1 45>,
-@@ -489,16 +572,31 @@
- dma-channels = <32>;
- big-endian;
- clock-names = "dmamux0", "dmamux1";
-- clocks = <&platform_clk 1>,
-- <&platform_clk 1>;
-+ clocks = <&clockgen 4 1>,
-+ <&clockgen 4 1>;
-+ };
-+
-+ qdma: qdma@8390000 {
-+ compatible = "fsl,ls1021a-qdma";
-+ reg = <0x0 0x8398000 0x0 0x1000>, /* Controller regs */
-+ <0x0 0x8399000 0x0 0x1000>, /* Status regs */
-+ <0x0 0x839a000 0x0 0x2000>; /* Block regs */
-+ interrupts = <GIC_SPI 185 IRQ_TYPE_LEVEL_HIGH>,
-+ <GIC_SPI 76 IRQ_TYPE_LEVEL_HIGH>;
-+ interrupt-names = "qdma-error", "qdma-queue";
-+ channels = <8>;
-+ queues = <2>;
-+ status-sizes = <64>;
-+ queue-sizes = <64 64>;
-+ big-endian;
- };
-
- dcu: dcu@2ce0000 {
- compatible = "fsl,ls1021a-dcu";
- reg = <0x0 0x2ce0000 0x0 0x10000>;
- interrupts = <GIC_SPI 172 IRQ_TYPE_LEVEL_HIGH>;
-- clocks = <&platform_clk 0>,
-- <&platform_clk 0>;
-+ clocks = <&clockgen 4 0>,
-+ <&clockgen 4 0>;
- clock-names = "dcu", "pix";
- big-endian;
- status = "disabled";
-@@ -626,7 +724,11 @@
- interrupts = <GIC_SPI 93 IRQ_TYPE_LEVEL_HIGH>;
- dr_mode = "host";
- snps,quirk-frame-length-adjustment = <0x20>;
-+ configure-gfladj;
-+ dma-coherent;
- snps,dis_rxdet_inp3_quirk;
-+ usb3-lpm-capable;
-+ snps,dis-u1u2-when-u3-quirk;
- };
-
- pcie@3400000 {
-@@ -634,7 +736,9 @@
- reg = <0x00 0x03400000 0x0 0x00010000 /* controller registers */
- 0x40 0x00000000 0x0 0x00002000>; /* configuration space */
- reg-names = "regs", "config";
-- interrupts = <GIC_SPI 177 IRQ_TYPE_LEVEL_HIGH>; /* controller interrupt */
-+ interrupts = <GIC_SPI 181 IRQ_TYPE_LEVEL_HIGH>,
-+ <GIC_SPI 183 IRQ_TYPE_LEVEL_HIGH>; /* aer interrupt */
-+ interrupt-names = "pme", "aer";
- fsl,pcie-scfg = <&scfg 0>;
- #address-cells = <3>;
- #size-cells = <2>;
-@@ -643,7 +747,7 @@
- bus-range = <0x0 0xff>;
- ranges = <0x81000000 0x0 0x00000000 0x40 0x00010000 0x0 0x00010000 /* downstream I/O */
- 0x82000000 0x0 0x40000000 0x40 0x40000000 0x0 0x40000000>; /* non-prefetchable memory */
-- msi-parent = <&msi1>;
-+ msi-parent = <&msi1>, <&msi2>;
- #interrupt-cells = <1>;
- interrupt-map-mask = <0 0 0 7>;
- interrupt-map = <0000 0 0 1 &gic GIC_SPI 91 IRQ_TYPE_LEVEL_HIGH>,
-@@ -657,7 +761,9 @@
- reg = <0x00 0x03500000 0x0 0x00010000 /* controller registers */
- 0x48 0x00000000 0x0 0x00002000>; /* configuration space */
- reg-names = "regs", "config";
-- interrupts = <GIC_SPI 178 IRQ_TYPE_LEVEL_HIGH>;
-+ interrupts = <GIC_SPI 182 IRQ_TYPE_LEVEL_HIGH>,
-+ <GIC_SPI 184 IRQ_TYPE_LEVEL_HIGH>; /* aer interrupt */
-+ interrupt-names = "pme", "aer";
- fsl,pcie-scfg = <&scfg 1>;
- #address-cells = <3>;
- #size-cells = <2>;
-@@ -666,7 +772,7 @@
- bus-range = <0x0 0xff>;
- ranges = <0x81000000 0x0 0x00000000 0x48 0x00010000 0x0 0x00010000 /* downstream I/O */
- 0x82000000 0x0 0x40000000 0x48 0x40000000 0x0 0x40000000>; /* non-prefetchable memory */
-- msi-parent = <&msi2>;
-+ msi-parent = <&msi1>, <&msi2>;
- #interrupt-cells = <1>;
- interrupt-map-mask = <0 0 0 7>;
- interrupt-map = <0000 0 0 1 &gic GIC_SPI 92 IRQ_TYPE_LEVEL_HIGH>,
-@@ -674,5 +780,45 @@
- <0000 0 0 3 &gic GIC_SPI 191 IRQ_TYPE_LEVEL_HIGH>,
- <0000 0 0 4 &gic GIC_SPI 193 IRQ_TYPE_LEVEL_HIGH>;
- };
-+
-+ can0: can@2a70000 {
-+ compatible = "fsl,ls1021ar2-flexcan";
-+ reg = <0x0 0x2a70000 0x0 0x1000>;
-+ interrupts = <GIC_SPI 126 IRQ_TYPE_LEVEL_HIGH>;
-+ clocks = <&clockgen 4 1>, <&clockgen 4 1>;
-+ clock-names = "ipg", "per";
-+ big-endian;
-+ status = "disabled";
-+ };
-+
-+ can1: can@2a80000 {
-+ compatible = "fsl,ls1021ar2-flexcan";
-+ reg = <0x0 0x2a80000 0x0 0x1000>;
-+ interrupts = <GIC_SPI 127 IRQ_TYPE_LEVEL_HIGH>;
-+ clocks = <&clockgen 4 1>, <&clockgen 4 1>;
-+ clock-names = "ipg", "per";
-+ big-endian;
-+ status = "disabled";
-+ };
-+
-+ can2: can@2a90000 {
-+ compatible = "fsl,ls1021ar2-flexcan";
-+ reg = <0x0 0x2a90000 0x0 0x1000>;
-+ interrupts = <GIC_SPI 128 IRQ_TYPE_LEVEL_HIGH>;
-+ clocks = <&clockgen 4 1>, <&clockgen 4 1>;
-+ clock-names = "ipg", "per";
-+ big-endian;
-+ status = "disabled";
-+ };
-+
-+ can3: can@2aa0000 {
-+ compatible = "fsl,ls1021ar2-flexcan";
-+ reg = <0x0 0x2aa0000 0x0 0x1000>;
-+ interrupts = <GIC_SPI 129 IRQ_TYPE_LEVEL_HIGH>;
-+ clocks = <&clockgen 4 1>, <&clockgen 4 1>;
-+ clock-names = "ipg", "per";
-+ big-endian;
-+ status = "disabled";
-+ };
- };
- };
---- a/arch/arm/boot/dts/mt6580.dtsi
-+++ b/arch/arm/boot/dts/mt6580.dtsi
-@@ -91,7 +91,7 @@
- #interrupt-cells = <3>;
- interrupt-parent = <&gic>;
- reg = <0x10211000 0x1000>,
-- <0x10212000 0x1000>,
-+ <0x10212000 0x2000>,
- <0x10214000 0x2000>,
- <0x10216000 0x2000>;
- };
---- a/arch/arm/boot/dts/mt6589.dtsi
-+++ b/arch/arm/boot/dts/mt6589.dtsi
-@@ -102,7 +102,7 @@
- #interrupt-cells = <3>;
- interrupt-parent = <&gic>;
- reg = <0x10211000 0x1000>,
-- <0x10212000 0x1000>,
-+ <0x10212000 0x2000>,
- <0x10214000 0x2000>,
- <0x10216000 0x2000>;
- };
---- a/arch/arm/boot/dts/mt8127.dtsi
-+++ b/arch/arm/boot/dts/mt8127.dtsi
-@@ -129,7 +129,7 @@
- #interrupt-cells = <3>;
- interrupt-parent = <&gic>;
- reg = <0 0x10211000 0 0x1000>,
-- <0 0x10212000 0 0x1000>,
-+ <0 0x10212000 0 0x2000>,
- <0 0x10214000 0 0x2000>,
- <0 0x10216000 0 0x2000>;
- };
---- a/arch/arm/boot/dts/mt8135.dtsi
-+++ b/arch/arm/boot/dts/mt8135.dtsi
-@@ -221,7 +221,7 @@
- #interrupt-cells = <3>;
- interrupt-parent = <&gic>;
- reg = <0 0x10211000 0 0x1000>,
-- <0 0x10212000 0 0x1000>,
-+ <0 0x10212000 0 0x2000>,
- <0 0x10214000 0 0x2000>,
- <0 0x10216000 0 0x2000>;
- };
---- a/arch/arm/boot/dts/rk3288.dtsi
-+++ b/arch/arm/boot/dts/rk3288.dtsi
-@@ -1109,7 +1109,7 @@
- #address-cells = <0>;
-
- reg = <0xffc01000 0x1000>,
-- <0xffc02000 0x1000>,
-+ <0xffc02000 0x2000>,
- <0xffc04000 0x2000>,
- <0xffc06000 0x2000>;
- interrupts = <GIC_PPI 9 0xf04>;
---- a/arch/arm/boot/dts/sun6i-a31.dtsi
-+++ b/arch/arm/boot/dts/sun6i-a31.dtsi
-@@ -791,7 +791,7 @@
- gic: interrupt-controller@01c81000 {
- compatible = "arm,cortex-a7-gic", "arm,cortex-a15-gic";
- reg = <0x01c81000 0x1000>,
-- <0x01c82000 0x1000>,
-+ <0x01c82000 0x2000>,
- <0x01c84000 0x2000>,
- <0x01c86000 0x2000>;
- interrupt-controller;
---- a/arch/arm/boot/dts/sun7i-a20.dtsi
-+++ b/arch/arm/boot/dts/sun7i-a20.dtsi
-@@ -1685,9 +1685,9 @@
- };
-
- gic: interrupt-controller@01c81000 {
-- compatible = "arm,cortex-a7-gic", "arm,cortex-a15-gic";
-+ compatible = "arm,gic-400", "arm,cortex-a7-gic", "arm,cortex-a15-gic";
- reg = <0x01c81000 0x1000>,
-- <0x01c82000 0x1000>,
-+ <0x01c82000 0x2000>,
- <0x01c84000 0x2000>,
- <0x01c86000 0x2000>;
- interrupt-controller;
---- a/arch/arm/boot/dts/sun8i-a23-a33.dtsi
-+++ b/arch/arm/boot/dts/sun8i-a23-a33.dtsi
-@@ -488,7 +488,7 @@
- gic: interrupt-controller@01c81000 {
- compatible = "arm,cortex-a7-gic", "arm,cortex-a15-gic";
- reg = <0x01c81000 0x1000>,
-- <0x01c82000 0x1000>,
-+ <0x01c82000 0x2000>,
- <0x01c84000 0x2000>,
- <0x01c86000 0x2000>;
- interrupt-controller;
---- a/arch/arm/boot/dts/sun9i-a80.dtsi
-+++ b/arch/arm/boot/dts/sun9i-a80.dtsi
-@@ -613,7 +613,7 @@
- gic: interrupt-controller@01c41000 {
- compatible = "arm,cortex-a7-gic", "arm,cortex-a15-gic";
- reg = <0x01c41000 0x1000>,
-- <0x01c42000 0x1000>,
-+ <0x01c42000 0x2000>,
- <0x01c44000 0x2000>,
- <0x01c46000 0x2000>;
- interrupt-controller;
---- a/arch/arm64/boot/dts/freescale/Makefile
-+++ b/arch/arm64/boot/dts/freescale/Makefile
-@@ -1,8 +1,26 @@
-+dtb-$(CONFIG_ARCH_LAYERSCAPE) += fsl-ls1012a-frdm.dtb
-+dtb-$(CONFIG_ARCH_LAYERSCAPE) += fsl-ls1012a-frwy.dtb
-+dtb-$(CONFIG_ARCH_LAYERSCAPE) += fsl-ls1012a-qds.dtb
-+dtb-$(CONFIG_ARCH_LAYERSCAPE) += fsl-ls1012a-rdb.dtb
-+dtb-$(CONFIG_ARCH_LAYERSCAPE) += fsl-ls1012a-2g5rdb.dtb
- dtb-$(CONFIG_ARCH_LAYERSCAPE) += fsl-ls1043a-qds.dtb
-+dtb-$(CONFIG_ARCH_LAYERSCAPE) += fsl-ls1043a-qds-sdk.dtb
- dtb-$(CONFIG_ARCH_LAYERSCAPE) += fsl-ls1043a-rdb.dtb
-+dtb-$(CONFIG_ARCH_LAYERSCAPE) += fsl-ls1043a-rdb-sdk.dtb
-+dtb-$(CONFIG_ARCH_LAYERSCAPE) += fsl-ls1043a-rdb-usdpaa.dtb
-+dtb-$(CONFIG_ARCH_LAYERSCAPE) += fsl-ls1046a-qds.dtb
-+dtb-$(CONFIG_ARCH_LAYERSCAPE) += fsl-ls1046a-qds-sdk.dtb
-+dtb-$(CONFIG_ARCH_LAYERSCAPE) += fsl-ls1046a-rdb.dtb
-+dtb-$(CONFIG_ARCH_LAYERSCAPE) += fsl-ls1046a-rdb-sdk.dtb
-+dtb-$(CONFIG_ARCH_LAYERSCAPE) += fsl-ls1046a-rdb-usdpaa.dtb
-+dtb-$(CONFIG_ARCH_LAYERSCAPE) += fsl-ls1088a-qds.dtb
-+dtb-$(CONFIG_ARCH_LAYERSCAPE) += fsl-ls1088a-rdb.dtb
- dtb-$(CONFIG_ARCH_LAYERSCAPE) += fsl-ls2080a-qds.dtb
- dtb-$(CONFIG_ARCH_LAYERSCAPE) += fsl-ls2080a-rdb.dtb
-+dtb-$(CONFIG_ARCH_LAYERSCAPE) += fsl-ls2081a-rdb.dtb
- dtb-$(CONFIG_ARCH_LAYERSCAPE) += fsl-ls2080a-simu.dtb
-+dtb-$(CONFIG_ARCH_LAYERSCAPE) += fsl-ls2088a-qds.dtb
-+dtb-$(CONFIG_ARCH_LAYERSCAPE) += fsl-ls2088a-rdb.dtb
-
- always := $(dtb-y)
- subdir-y := $(dts-dirs)
---- /dev/null
-+++ b/arch/arm64/boot/dts/freescale/fsl-ls1012a-2g5rdb.dts
-@@ -0,0 +1,123 @@
-+/*
-+ * Device Tree file for NXP LS1012A 2G5RDB Board.
-+ *
-+ * Copyright 2017 NXP
-+ *
-+ * Bhaskar Upadhaya <bhaskar.upadhaya@nxp.com>
-+ *
-+ * This file is dual-licensed: you can use it either under the terms
-+ * of the GPLv2 or the X11 license, at your option. Note that this dual
-+ * licensing only applies to this file, and not this project as a
-+ * whole.
-+ *
-+ * a) This library is free software; you can redistribute it and/or
-+ * modify it under the terms of the GNU General Public License as
-+ * published by the Free Software Foundation; either version 2 of the
-+ * License, or (at your option) any later version.
-+ *
-+ * This library is distributed in the hope that it will be useful,
-+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
-+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
-+ * GNU General Public License for more details.
-+ *
-+ * Or, alternatively,
-+ *
-+ * b) Permission is hereby granted, free of charge, to any person
-+ * obtaining a copy of this software and associated documentation
-+ * files (the "Software"), to deal in the Software without
-+ * restriction, including without limitation the rights to use,
-+ * copy, modify, merge, publish, distribute, sublicense, and/or
-+ * sell copies of the Software, and to permit persons to whom the
-+ * Software is furnished to do so, subject to the following
-+ * conditions:
-+ *
-+ * The above copyright notice and this permission notice shall be
-+ * included in all copies or substantial portions of the Software.
-+ *
-+ * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
-+ * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES
-+ * OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
-+ * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT
-+ * HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY,
-+ * WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
-+ * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
-+ * OTHER DEALINGS IN THE SOFTWARE.
-+ */
-+/dts-v1/;
-+
-+#include "fsl-ls1012a.dtsi"
-+
-+/ {
-+ model = "LS1012A 2G5RDB Board";
-+ compatible = "fsl,ls1012a-rdb", "fsl,ls1012a";
-+
-+ aliases {
-+ ethernet0 = &pfe_mac0;
-+ ethernet1 = &pfe_mac1;
-+ };
-+};
-+
-+&duart0 {
-+ status = "okay";
-+};
-+
-+&i2c0 {
-+ status = "okay";
-+};
-+
-+&qspi {
-+ num-cs = <2>;
-+ bus-num = <0>;
-+ status = "okay";
-+
-+ qflash0: s25fs512s@0 {
-+ compatible = "spansion,m25p80";
-+ #address-cells = <1>;
-+ #size-cells = <1>;
-+ spi-max-frequency = <20000000>;
-+ m25p,fast-read;
-+ reg = <0>;
-+ };
-+};
-+
-+&sata {
-+ status = "okay";
-+};
-+
-+&pfe {
-+ status = "okay";
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+
-+ ethernet@0 {
-+ compatible = "fsl,pfe-gemac-port";
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ reg = <0x0>; /* GEM_ID */
-+ fsl,gemac-bus-id = <0x0>; /* BUS_ID */
-+ fsl,gemac-phy-id = <0x1>; /* PHY_ID */
-+ fsl,mdio-mux-val = <0x0>;
-+ phy-mode = "sgmii-2500";
-+ fsl,pfe-phy-if-flags = <0x0>;
-+
-+ mdio@0 {
-+ reg = <0x1>; /* enabled/disabled */
-+ };
-+ };
-+
-+ ethernet@1 {
-+ compatible = "fsl,pfe-gemac-port";
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ reg = <0x1>; /* GEM_ID */
-+ fsl,gemac-bus-id = < 0x0>; /* BUS_ID */
-+ fsl,gemac-phy-id = < 0x2>; /* PHY_ID */
-+ fsl,mdio-mux-val = <0x0>;
-+ phy-mode = "sgmii-2500";
-+ fsl,pfe-phy-if-flags = <0x0>;
-+
-+ mdio@0 {
-+ reg = <0x0>; /* enabled/disabled */
-+ };
-+ };
-+};
---- /dev/null
-+++ b/arch/arm64/boot/dts/freescale/fsl-ls1012a-frdm.dts
-@@ -0,0 +1,141 @@
-+// SPDX-License-Identifier: (GPL-2.0+ OR MIT)
-+/*
-+ * Device Tree file for Freescale LS1012A Freedom Board.
-+ *
-+ * Copyright 2016 Freescale Semiconductor, Inc.
-+ *
-+ */
-+/dts-v1/;
-+
-+#include "fsl-ls1012a.dtsi"
-+
-+/ {
-+ model = "LS1012A Freedom Board";
-+ compatible = "fsl,ls1012a-frdm", "fsl,ls1012a";
-+
-+ aliases {
-+ ethernet0 = &pfe_mac0;
-+ ethernet1 = &pfe_mac1;
-+ };
-+
-+ sys_mclk: clock-mclk {
-+ compatible = "fixed-clock";
-+ #clock-cells = <0>;
-+ clock-frequency = <25000000>;
-+ };
-+
-+ reg_1p8v: regulator-1p8v {
-+ compatible = "regulator-fixed";
-+ regulator-name = "1P8V";
-+ regulator-min-microvolt = <1800000>;
-+ regulator-max-microvolt = <1800000>;
-+ regulator-always-on;
-+ };
-+
-+ sound {
-+ compatible = "simple-audio-card";
-+ simple-audio-card,format = "i2s";
-+ simple-audio-card,widgets =
-+ "Microphone", "Microphone Jack",
-+ "Headphone", "Headphone Jack",
-+ "Speaker", "Speaker Ext",
-+ "Line", "Line In Jack";
-+ simple-audio-card,routing =
-+ "MIC_IN", "Microphone Jack",
-+ "Microphone Jack", "Mic Bias",
-+ "LINE_IN", "Line In Jack",
-+ "Headphone Jack", "HP_OUT",
-+ "Speaker Ext", "LINE_OUT";
-+
-+ simple-audio-card,cpu {
-+ sound-dai = <&sai2>;
-+ frame-master;
-+ bitclock-master;
-+ };
-+
-+ simple-audio-card,codec {
-+ sound-dai = <&codec>;
-+ frame-master;
-+ bitclock-master;
-+ system-clock-frequency = <25000000>;
-+ };
-+ };
-+};
-+
-+&duart0 {
-+ status = "okay";
-+};
-+
-+&i2c0 {
-+ status = "okay";
-+
-+ codec: sgtl5000@a {
-+ #sound-dai-cells = <0>;
-+ compatible = "fsl,sgtl5000";
-+ reg = <0xa>;
-+ VDDA-supply = <&reg_1p8v>;
-+ VDDIO-supply = <&reg_1p8v>;
-+ clocks = <&sys_mclk>;
-+ };
-+};
-+
-+&qspi {
-+ num-cs = <1>;
-+ bus-num = <0>;
-+ status = "okay";
-+
-+ qflash0: s25fs512s@0 {
-+ compatible = "spansion,m25p80";
-+ #address-cells = <1>;
-+ #size-cells = <1>;
-+ m25p,fast-read;
-+ spi-max-frequency = <20000000>;
-+ reg = <0>;
-+ };
-+};
-+
-+&pfe {
-+ status = "okay";
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+
-+ ethernet@0 {
-+ compatible = "fsl,pfe-gemac-port";
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ reg = <0x0>; /* GEM_ID */
-+ fsl,gemac-bus-id = <0x0>; /* BUS_ID */
-+ fsl,gemac-phy-id = <0x2>; /* PHY_ID */
-+ fsl,mdio-mux-val = <0x0>;
-+ phy-mode = "sgmii";
-+ fsl,pfe-phy-if-flags = <0x0>;
-+
-+ mdio@0 {
-+ reg = <0x1>; /* enabled/disabled */
-+ };
-+ };
-+
-+ ethernet@1 {
-+ compatible = "fsl,pfe-gemac-port";
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ reg = <0x1>; /* GEM_ID */
-+ fsl,gemac-bus-id = <0x1>; /* BUS_ID */
-+ fsl,gemac-phy-id = <0x1>; /* PHY_ID */
-+ fsl,mdio-mux-val = <0x0>;
-+ phy-mode = "sgmii";
-+ fsl,pfe-phy-if-flags = <0x0>;
-+
-+ mdio@0 {
-+ reg = <0x0>; /* enabled/disabled */
-+ };
-+ };
-+};
-+
-+&sai2 {
-+ status = "okay";
-+};
-+
-+&sata {
-+ status = "okay";
-+};
---- /dev/null
-+++ b/arch/arm64/boot/dts/freescale/fsl-ls1012a-frwy.dts
-@@ -0,0 +1,177 @@
-+/*
-+ * Device Tree file for NXP LS1012A FRWY Board.
-+ *
-+ * Copyright 2018 NXP
-+ *
-+ * This file is dual-licensed: you can use it either under the terms
-+ * of the GPLv2 or the X11 license, at your option. Note that this dual
-+ * licensing only applies to this file, and not this project as a
-+ * whole.
-+ *
-+ * a) This library is free software; you can redistribute it and/or
-+ * modify it under the terms of the GNU General Public License as
-+ * published by the Free Software Foundation; either version 2 of the
-+ * License, or (at your option) any later version.
-+ *
-+ * This library is distributed in the hope that it will be useful,
-+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
-+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
-+ * GNU General Public License for more details.
-+ *
-+ * Or, alternatively,
-+ *
-+ * b) Permission is hereby granted, free of charge, to any person
-+ * obtaining a copy of this software and associated documentation
-+ * files (the "Software"), to deal in the Software without
-+ * restriction, including without limitation the rights to use,
-+ * copy, modify, merge, publish, distribute, sublicense, and/or
-+ * sell copies of the Software, and to permit persons to whom the
-+ * Software is furnished to do so, subject to the following
-+ * conditions:
-+ *
-+ * The above copyright notice and this permission notice shall be
-+ * included in all copies or substantial portions of the Software.
-+ *
-+ * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
-+ * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES
-+ * OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
-+ * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT
-+ * HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY,
-+ * WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
-+ * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
-+ * OTHER DEALINGS IN THE SOFTWARE.
-+ */
-+/dts-v1/;
-+
-+#include "fsl-ls1012a.dtsi"
-+
-+/ {
-+ model = "LS1012A FRWY Board";
-+ compatible = "fsl,ls1012a-frwy", "fsl,ls1012a";
-+
-+ aliases {
-+ ethernet0 = &pfe_mac0;
-+ ethernet1 = &pfe_mac1;
-+ };
-+
-+ sys_mclk: clock-mclk {
-+ compatible = "fixed-clock";
-+ #clock-cells = <0>;
-+ clock-frequency = <25000000>;
-+ };
-+
-+ reg_1p8v: regulator-1p8v {
-+ compatible = "regulator-fixed";
-+ regulator-name = "1P8V";
-+ regulator-min-microvolt = <1800000>;
-+ regulator-max-microvolt = <1800000>;
-+ regulator-always-on;
-+ };
-+
-+ sound {
-+ compatible = "simple-audio-card";
-+ simple-audio-card,format = "i2s";
-+ simple-audio-card,widgets =
-+ "Microphone", "Microphone Jack",
-+ "Headphone", "Headphone Jack",
-+ "Speaker", "Speaker Ext",
-+ "Line", "Line In Jack";
-+ simple-audio-card,routing =
-+ "MIC_IN", "Microphone Jack",
-+ "Microphone Jack", "Mic Bias",
-+ "LINE_IN", "Line In Jack",
-+ "Headphone Jack", "HP_OUT",
-+ "Speaker Ext", "LINE_OUT";
-+
-+ simple-audio-card,cpu {
-+ sound-dai = <&sai2>;
-+ frame-master;
-+ bitclock-master;
-+ };
-+
-+ simple-audio-card,codec {
-+ sound-dai = <&codec>;
-+ frame-master;
-+ bitclock-master;
-+ system-clock-frequency = <25000000>;
-+ };
-+ };
-+};
-+
-+&pcie {
-+ status = "okay";
-+};
-+
-+&duart0 {
-+ status = "okay";
-+};
-+
-+&i2c0 {
-+ status = "okay";
-+
-+ codec: sgtl5000@a {
-+ compatible = "fsl,sgtl5000";
-+ #sound-dai-cells = <0>;
-+ reg = <0xa>;
-+ VDDA-supply = <&reg_1p8v>;
-+ VDDIO-supply = <&reg_1p8v>;
-+ clocks = <&sys_mclk>;
-+ };
-+};
-+
-+&qspi {
-+ num-cs = <1>;
-+ bus-num = <0>;
-+ status = "okay";
-+
-+ qflash0: w25q16dw@0 {
-+ compatible = "spansion,m25p80";
-+ #address-cells = <1>;
-+ #size-cells = <1>;
-+ m25p,fast-read;
-+ spi-max-frequency = <20000000>;
-+ reg = <0>;
-+ };
-+};
-+
-+&pfe {
-+ status = "okay";
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+
-+ ethernet@0 {
-+ compatible = "fsl,pfe-gemac-port";
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ reg = <0x0>; /* GEM_ID */
-+ fsl,gemac-bus-id = <0x0>; /* BUS_ID */
-+ fsl,gemac-phy-id = <0x2>; /* PHY_ID */
-+ fsl,mdio-mux-val = <0x0>;
-+ phy-mode = "sgmii";
-+ fsl,pfe-phy-if-flags = <0x0>;
-+
-+ mdio@0 {
-+ reg = <0x1>; /* enabled/disabled */
-+ };
-+ };
-+
-+ ethernet@1 {
-+ compatible = "fsl,pfe-gemac-port";
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ reg = <0x1>; /* GEM_ID */
-+ fsl,gemac-bus-id = <0x1>; /* BUS_ID */
-+ fsl,gemac-phy-id = <0x1>; /* PHY_ID */
-+ fsl,mdio-mux-val = <0x0>;
-+ phy-mode = "sgmii";
-+ fsl,pfe-phy-if-flags = <0x0>;
-+
-+ mdio@0 {
-+ reg = <0x0>; /* enabled/disabled */
-+ };
-+ };
-+};
-+
-+&sai2 {
-+ status = "okay";
-+};
---- /dev/null
-+++ b/arch/arm64/boot/dts/freescale/fsl-ls1012a-qds.dts
-@@ -0,0 +1,166 @@
-+// SPDX-License-Identifier: (GPL-2.0+ OR MIT)
-+/*
-+ * Device Tree file for Freescale LS1012A QDS Board.
-+ *
-+ * Copyright 2016 Freescale Semiconductor, Inc.
-+ *
-+ */
-+/dts-v1/;
-+
-+#include "fsl-ls1012a.dtsi"
-+
-+/ {
-+ model = "LS1012A QDS Board";
-+ compatible = "fsl,ls1012a-qds", "fsl,ls1012a";
-+
-+ aliases {
-+ ethernet0 = &pfe_mac0;
-+ ethernet1 = &pfe_mac1;
-+ };
-+
-+ sys_mclk: clock-mclk {
-+ compatible = "fixed-clock";
-+ #clock-cells = <0>;
-+ clock-frequency = <24576000>;
-+ };
-+
-+ reg_3p3v: regulator-3p3v {
-+ compatible = "regulator-fixed";
-+ regulator-name = "3P3V";
-+ regulator-min-microvolt = <3300000>;
-+ regulator-max-microvolt = <3300000>;
-+ regulator-always-on;
-+ };
-+
-+ sound {
-+ compatible = "simple-audio-card";
-+ simple-audio-card,format = "i2s";
-+ simple-audio-card,widgets =
-+ "Microphone", "Microphone Jack",
-+ "Headphone", "Headphone Jack",
-+ "Speaker", "Speaker Ext",
-+ "Line", "Line In Jack";
-+ simple-audio-card,routing =
-+ "MIC_IN", "Microphone Jack",
-+ "Microphone Jack", "Mic Bias",
-+ "LINE_IN", "Line In Jack",
-+ "Headphone Jack", "HP_OUT",
-+ "Speaker Ext", "LINE_OUT";
-+
-+ simple-audio-card,cpu {
-+ sound-dai = <&sai2>;
-+ frame-master;
-+ bitclock-master;
-+ };
-+
-+ simple-audio-card,codec {
-+ sound-dai = <&codec>;
-+ frame-master;
-+ bitclock-master;
-+ system-clock-frequency = <24576000>;
-+ };
-+ };
-+};
-+
-+&pcie {
-+ status = "okay";
-+};
-+
-+&duart0 {
-+ status = "okay";
-+};
-+
-+&i2c0 {
-+ status = "okay";
-+
-+ pca9547@77 {
-+ compatible = "nxp,pca9547";
-+ reg = <0x77>;
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+
-+ i2c@4 {
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ reg = <0x4>;
-+
-+ codec: sgtl5000@a {
-+ #sound-dai-cells = <0>;
-+ compatible = "fsl,sgtl5000";
-+ reg = <0xa>;
-+ VDDA-supply = <&reg_3p3v>;
-+ VDDIO-supply = <&reg_3p3v>;
-+ clocks = <&sys_mclk>;
-+ };
-+ };
-+ };
-+};
-+
-+&qspi {
-+ num-cs = <2>;
-+ bus-num = <0>;
-+ status = "okay";
-+
-+ qflash0: s25fs512s@0 {
-+ compatible = "spansion,m25p80";
-+ #address-cells = <1>;
-+ #size-cells = <1>;
-+ spi-max-frequency = <20000000>;
-+ m25p,fast-read;
-+ reg = <0>;
-+ };
-+};
-+
-+&pfe {
-+ status = "okay";
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+
-+ ethernet@0 {
-+ compatible = "fsl,pfe-gemac-port";
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ reg = <0x0>; /* GEM_ID */
-+ fsl,gemac-bus-id = <0x0>; /* BUS_ID */
-+ fsl,gemac-phy-id = <0x1>; /* PHY_ID */
-+ fsl,mdio-mux-val = <0x2>;
-+ phy-mode = "sgmii-2500";
-+ fsl,pfe-phy-if-flags = <0x0>;
-+
-+ mdio@0 {
-+ reg = <0x1>; /* enabled/disabled */
-+ };
-+ };
-+
-+ ethernet@1 {
-+ compatible = "fsl,pfe-gemac-port";
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ reg = <0x1>; /* GEM_ID */
-+ fsl,gemac-bus-id = <0x1>; /* BUS_ID */
-+ fsl,gemac-phy-id = <0x2>; /* PHY_ID */
-+ fsl,mdio-mux-val = <0x3>;
-+ phy-mode = "sgmii-2500";
-+ fsl,pfe-phy-if-flags = <0x0>;
-+
-+ mdio@0 {
-+ reg = <0x0>; /* enabled/disabled */
-+ };
-+ };
-+};
-+
-+&sai2 {
-+ status = "okay";
-+};
-+
-+&sata {
-+ status = "okay";
-+};
-+
-+&esdhc0 {
-+ status = "okay";
-+};
-+
-+&esdhc1 {
-+ status = "okay";
-+};
---- /dev/null
-+++ b/arch/arm64/boot/dts/freescale/fsl-ls1012a-rdb.dts
-@@ -0,0 +1,102 @@
-+// SPDX-License-Identifier: (GPL-2.0+ OR MIT)
-+/*
-+ * Device Tree file for Freescale LS1012A RDB Board.
-+ *
-+ * Copyright 2016 Freescale Semiconductor, Inc.
-+ *
-+ */
-+/dts-v1/;
-+
-+#include "fsl-ls1012a.dtsi"
-+
-+/ {
-+ model = "LS1012A RDB Board";
-+ compatible = "fsl,ls1012a-rdb", "fsl,ls1012a";
-+
-+ aliases {
-+ ethernet0 = &pfe_mac0;
-+ ethernet1 = &pfe_mac1;
-+ };
-+};
-+
-+&pcie {
-+ status = "okay";
-+};
-+
-+&duart0 {
-+ status = "okay";
-+};
-+
-+&i2c0 {
-+ status = "okay";
-+};
-+
-+&qspi {
-+ num-cs = <2>;
-+ bus-num = <0>;
-+ status = "okay";
-+
-+ qflash0: s25fs512s@0 {
-+ compatible = "spansion,m25p80";
-+ #address-cells = <1>;
-+ #size-cells = <1>;
-+ spi-max-frequency = <20000000>;
-+ m25p,fast-read;
-+ reg = <0>;
-+ };
-+};
-+
-+&sata {
-+ status = "okay";
-+};
-+
-+&esdhc0 {
-+ sd-uhs-sdr104;
-+ sd-uhs-sdr50;
-+ sd-uhs-sdr25;
-+ sd-uhs-sdr12;
-+ status = "okay";
-+};
-+
-+&esdhc1 {
-+ mmc-hs200-1_8v;
-+ status = "okay";
-+};
-+
-+&pfe {
-+ status = "okay";
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+
-+ ethernet@0 {
-+ compatible = "fsl,pfe-gemac-port";
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ reg = <0x0>; /* GEM_ID */
-+ fsl,gemac-bus-id = <0x0>; /* BUS_ID */
-+ fsl,gemac-phy-id = <0x2>; /* PHY_ID */
-+ fsl,mdio-mux-val = <0x0>;
-+ phy-mode = "sgmii";
-+ fsl,pfe-phy-if-flags = <0x0>;
-+
-+ mdio@0 {
-+ reg = <0x1>; /* enabled/disabled */
-+ };
-+ };
-+
-+ ethernet@1 {
-+ compatible = "fsl,pfe-gemac-port";
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ reg = <0x1>; /* GEM_ID */
-+ fsl,gemac-bus-id = < 0x1 >; /* BUS_ID */
-+ fsl,gemac-phy-id = < 0x1 >; /* PHY_ID */
-+ fsl,mdio-mux-val = <0x0>;
-+ phy-mode = "rgmii-txid";
-+ fsl,pfe-phy-if-flags = <0x0>;
-+
-+ mdio@0 {
-+ reg = <0x0>; /* enabled/disabled */
-+ };
-+ };
-+};
---- /dev/null
-+++ b/arch/arm64/boot/dts/freescale/fsl-ls1012a.dtsi
-@@ -0,0 +1,567 @@
-+// SPDX-License-Identifier: (GPL-2.0+ OR MIT)
-+/*
-+ * Device Tree Include file for Freescale Layerscape-1012A family SoC.
-+ *
-+ * Copyright 2016 Freescale Semiconductor, Inc.
-+ *
-+ */
-+
-+#include <dt-bindings/interrupt-controller/arm-gic.h>
-+#include <dt-bindings/thermal/thermal.h>
-+
-+/ {
-+ compatible = "fsl,ls1012a";
-+ interrupt-parent = <&gic>;
-+ #address-cells = <2>;
-+ #size-cells = <2>;
-+
-+ aliases {
-+ crypto = &crypto;
-+ rtic_a = &rtic_a;
-+ rtic_b = &rtic_b;
-+ rtic_c = &rtic_c;
-+ rtic_d = &rtic_d;
-+ sec_mon = &sec_mon;
-+ };
-+
-+ cpus {
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+
-+ cpu0: cpu@0 {
-+ device_type = "cpu";
-+ compatible = "arm,cortex-a53";
-+ reg = <0x0>;
-+ clocks = <&clockgen 1 0>;
-+ #cooling-cells = <2>;
-+ cpu-idle-states = <&CPU_PH20>;
-+ };
-+ };
-+
-+ idle-states {
-+ /*
-+ * PSCI node is not added default, U-boot will add missing
-+ * parts if it determines to use PSCI.
-+ */
-+ entry-method = "arm,psci";
-+
-+ CPU_PH20: cpu-ph20 {
-+ compatible = "arm,idle-state";
-+ idle-state-name = "PH20";
-+ arm,psci-suspend-param = <0x0>;
-+ entry-latency-us = <1000>;
-+ exit-latency-us = <1000>;
-+ min-residency-us = <3000>;
-+ };
-+ };
-+
-+ sysclk: sysclk {
-+ compatible = "fixed-clock";
-+ #clock-cells = <0>;
-+ clock-frequency = <125000000>;
-+ clock-output-names = "sysclk";
-+ };
-+
-+ coreclk: coreclk {
-+ compatible = "fixed-clock";
-+ #clock-cells = <0>;
-+ clock-frequency = <100000000>;
-+ clock-output-names = "coreclk";
-+ };
-+
-+ timer {
-+ compatible = "arm,armv8-timer";
-+ interrupts = <1 13 IRQ_TYPE_LEVEL_LOW>,/* Physical Secure PPI */
-+ <1 14 IRQ_TYPE_LEVEL_LOW>,/* Physical Non-Secure PPI */
-+ <1 11 IRQ_TYPE_LEVEL_LOW>,/* Virtual PPI */
-+ <1 10 IRQ_TYPE_LEVEL_LOW>;/* Hypervisor PPI */
-+ };
-+
-+ pmu {
-+ compatible = "arm,armv8-pmuv3";
-+ interrupts = <0 106 IRQ_TYPE_LEVEL_HIGH>;
-+ };
-+
-+ gic: interrupt-controller@1400000 {
-+ compatible = "arm,gic-400";
-+ #interrupt-cells = <3>;
-+ interrupt-controller;
-+ reg = <0x0 0x1401000 0 0x1000>, /* GICD */
-+ <0x0 0x1402000 0 0x2000>, /* GICC */
-+ <0x0 0x1404000 0 0x2000>, /* GICH */
-+ <0x0 0x1406000 0 0x2000>; /* GICV */
-+ interrupts = <1 9 IRQ_TYPE_LEVEL_LOW>;
-+ };
-+
-+ reboot {
-+ compatible = "syscon-reboot";
-+ regmap = <&dcfg>;
-+ offset = <0xb0>;
-+ mask = <0x02>;
-+ };
-+
-+ soc {
-+ compatible = "simple-bus";
-+ #address-cells = <2>;
-+ #size-cells = <2>;
-+ ranges;
-+
-+ scfg: scfg@1570000 {
-+ compatible = "fsl,ls1012a-scfg", "syscon";
-+ reg = <0x0 0x1570000 0x0 0x10000>;
-+ big-endian;
-+ };
-+
-+ crypto: crypto@1700000 {
-+ compatible = "fsl,sec-v5.4", "fsl,sec-v5.0",
-+ "fsl,sec-v4.0";
-+ fsl,sec-era = <8>;
-+ #address-cells = <1>;
-+ #size-cells = <1>;
-+ ranges = <0x0 0x00 0x1700000 0x100000>;
-+ reg = <0x00 0x1700000 0x0 0x100000>;
-+ interrupts = <GIC_SPI 75 IRQ_TYPE_LEVEL_HIGH>;
-+
-+ sec_jr0: jr@10000 {
-+ compatible = "fsl,sec-v5.4-job-ring",
-+ "fsl,sec-v5.0-job-ring",
-+ "fsl,sec-v4.0-job-ring";
-+ reg = <0x10000 0x10000>;
-+ interrupts = <GIC_SPI 71 IRQ_TYPE_LEVEL_HIGH>;
-+ };
-+
-+ sec_jr1: jr@20000 {
-+ compatible = "fsl,sec-v5.4-job-ring",
-+ "fsl,sec-v5.0-job-ring",
-+ "fsl,sec-v4.0-job-ring";
-+ reg = <0x20000 0x10000>;
-+ interrupts = <GIC_SPI 72 IRQ_TYPE_LEVEL_HIGH>;
-+ };
-+
-+ sec_jr2: jr@30000 {
-+ compatible = "fsl,sec-v5.4-job-ring",
-+ "fsl,sec-v5.0-job-ring",
-+ "fsl,sec-v4.0-job-ring";
-+ reg = <0x30000 0x10000>;
-+ interrupts = <GIC_SPI 73 IRQ_TYPE_LEVEL_HIGH>;
-+ };
-+
-+ sec_jr3: jr@40000 {
-+ compatible = "fsl,sec-v5.4-job-ring",
-+ "fsl,sec-v5.0-job-ring",
-+ "fsl,sec-v4.0-job-ring";
-+ reg = <0x40000 0x10000>;
-+ interrupts = <GIC_SPI 74 IRQ_TYPE_LEVEL_HIGH>;
-+ };
-+
-+ caam-dma {
-+ compatible = "fsl,sec-v5.4-dma",
-+ "fsl,sec-v5.0-dma",
-+ "fsl,sec-v4.0-dma";
-+ };
-+
-+ rtic@60000 {
-+ compatible = "fsl,sec-v5.4-rtic",
-+ "fsl,sec-v5.0-rtic",
-+ "fsl,sec-v4.0-rtic";
-+ #address-cells = <1>;
-+ #size-cells = <1>;
-+ reg = <0x60000 0x100 0x60e00 0x18>;
-+ ranges = <0x0 0x60100 0x500>;
-+
-+ rtic_a: rtic-a@0 {
-+ compatible = "fsl,sec-v5.4-rtic-memory",
-+ "fsl,sec-v5.0-rtic-memory",
-+ "fsl,sec-v4.0-rtic-memory";
-+ reg = <0x00 0x20 0x100 0x100>;
-+ };
-+
-+ rtic_b: rtic-b@20 {
-+ compatible = "fsl,sec-v5.4-rtic-memory",
-+ "fsl,sec-v5.0-rtic-memory",
-+ "fsl,sec-v4.0-rtic-memory";
-+ reg = <0x20 0x20 0x200 0x100>;
-+ };
-+
-+ rtic_c: rtic-c@40 {
-+ compatible = "fsl,sec-v5.4-rtic-memory",
-+ "fsl,sec-v5.0-rtic-memory",
-+ "fsl,sec-v4.0-rtic-memory";
-+ reg = <0x40 0x20 0x300 0x100>;
-+ };
-+
-+ rtic_d: rtic-d@60 {
-+ compatible = "fsl,sec-v5.4-rtic-memory",
-+ "fsl,sec-v5.0-rtic-memory",
-+ "fsl,sec-v4.0-rtic-memory";
-+ reg = <0x60 0x20 0x400 0x100>;
-+ };
-+ };
-+ };
-+
-+ sec_mon: sec_mon@1e90000 {
-+ compatible = "fsl,sec-v5.4-mon", "fsl,sec-v5.0-mon",
-+ "fsl,sec-v4.0-mon";
-+ reg = <0x0 0x1e90000 0x0 0x10000>;
-+ interrupts = <GIC_SPI 78 IRQ_TYPE_LEVEL_HIGH>,
-+ <GIC_SPI 79 IRQ_TYPE_LEVEL_HIGH>;
-+ };
-+
-+ dcfg: dcfg@1ee0000 {
-+ compatible = "fsl,ls1012a-dcfg",
-+ "syscon";
-+ reg = <0x0 0x1ee0000 0x0 0x10000>;
-+ big-endian;
-+ };
-+
-+ clockgen: clocking@1ee1000 {
-+ compatible = "fsl,ls1012a-clockgen";
-+ reg = <0x0 0x1ee1000 0x0 0x1000>;
-+ #clock-cells = <2>;
-+ clocks = <&sysclk &coreclk>;
-+ clock-names = "sysclk", "coreclk";
-+ };
-+
-+ tmu: tmu@1f00000 {
-+ compatible = "fsl,qoriq-tmu";
-+ reg = <0x0 0x1f00000 0x0 0x10000>;
-+ interrupts = <0 33 0x4>;
-+ fsl,tmu-range = <0xb0000 0x9002a 0x6004c 0x30062>;
-+ fsl,tmu-calibration = <0x00000000 0x00000026
-+ 0x00000001 0x0000002d
-+ 0x00000002 0x00000032
-+ 0x00000003 0x00000039
-+ 0x00000004 0x0000003f
-+ 0x00000005 0x00000046
-+ 0x00000006 0x0000004d
-+ 0x00000007 0x00000054
-+ 0x00000008 0x0000005a
-+ 0x00000009 0x00000061
-+ 0x0000000a 0x0000006a
-+ 0x0000000b 0x00000071
-+
-+ 0x00010000 0x00000025
-+ 0x00010001 0x0000002c
-+ 0x00010002 0x00000035
-+ 0x00010003 0x0000003d
-+ 0x00010004 0x00000045
-+ 0x00010005 0x0000004e
-+ 0x00010006 0x00000057
-+ 0x00010007 0x00000061
-+ 0x00010008 0x0000006b
-+ 0x00010009 0x00000076
-+
-+ 0x00020000 0x00000029
-+ 0x00020001 0x00000033
-+ 0x00020002 0x0000003d
-+ 0x00020003 0x00000049
-+ 0x00020004 0x00000056
-+ 0x00020005 0x00000061
-+ 0x00020006 0x0000006d
-+
-+ 0x00030000 0x00000021
-+ 0x00030001 0x0000002a
-+ 0x00030002 0x0000003c
-+ 0x00030003 0x0000004e>;
-+ big-endian;
-+ #thermal-sensor-cells = <1>;
-+ };
-+
-+ thermal-zones {
-+ cpu_thermal: cpu-thermal {
-+ polling-delay-passive = <1000>;
-+ polling-delay = <5000>;
-+ thermal-sensors = <&tmu 0>;
-+
-+ trips {
-+ cpu_alert: cpu-alert {
-+ temperature = <85000>;
-+ hysteresis = <2000>;
-+ type = "passive";
-+ };
-+
-+ cpu_crit: cpu-crit {
-+ temperature = <95000>;
-+ hysteresis = <2000>;
-+ type = "critical";
-+ };
-+ };
-+
-+ cooling-maps {
-+ map0 {
-+ trip = <&cpu_alert>;
-+ cooling-device =
-+ <&cpu0 THERMAL_NO_LIMIT
-+ THERMAL_NO_LIMIT>;
-+ };
-+ };
-+ };
-+ };
-+
-+ esdhc0: esdhc@1560000 {
-+ compatible = "fsl,ls1012a-esdhc", "fsl,esdhc";
-+ reg = <0x0 0x1560000 0x0 0x10000>;
-+ interrupts = <0 62 0x4>;
-+ clocks = <&clockgen 4 0>;
-+ voltage-ranges = <1800 1800 3300 3300>;
-+ sdhci,auto-cmd12;
-+ big-endian;
-+ bus-width = <4>;
-+ status = "disabled";
-+ };
-+
-+ esdhc1: esdhc@1580000 {
-+ compatible = "fsl,ls1012a-esdhc", "fsl,esdhc";
-+ reg = <0x0 0x1580000 0x0 0x10000>;
-+ interrupts = <0 65 0x4>;
-+ clocks = <&clockgen 4 0>;
-+ voltage-ranges = <1800 1800 3300 3300>;
-+ sdhci,auto-cmd12;
-+ big-endian;
-+ broken-cd;
-+ bus-width = <4>;
-+ status = "disabled";
-+ };
-+
-+ rcpm: rcpm@1ee2000 {
-+ compatible = "fsl,ls1012a-rcpm", "fsl,qoriq-rcpm-2.1";
-+ reg = <0x0 0x1ee2000 0x0 0x1000>;
-+ fsl,#rcpm-wakeup-cells = <1>;
-+ };
-+
-+ ftm0: ftm0@29d0000 {
-+ compatible = "fsl,ls1012a-ftm";
-+ reg = <0x0 0x29d0000 0x0 0x10000>,
-+ <0x0 0x1ee2140 0x0 0x4>;
-+ reg-names = "ftm", "FlexTimer1";
-+ interrupts = <0 86 0x4>;
-+ big-endian;
-+ };
-+
-+ i2c0: i2c@2180000 {
-+ compatible = "fsl,vf610-i2c", "fsl,ls1012a-vf610-i2c";
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ reg = <0x0 0x2180000 0x0 0x10000>;
-+ interrupts = <0 56 IRQ_TYPE_LEVEL_HIGH>;
-+ clocks = <&clockgen 4 3>;
-+ fsl-scl-gpio = <&gpio0 13 0>;
-+ status = "disabled";
-+ };
-+
-+ i2c1: i2c@2190000 {
-+ compatible = "fsl,vf610-i2c";
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ reg = <0x0 0x2190000 0x0 0x10000>;
-+ interrupts = <0 57 IRQ_TYPE_LEVEL_HIGH>;
-+ clocks = <&clockgen 4 3>;
-+ status = "disabled";
-+ };
-+
-+ duart0: serial@21c0500 {
-+ compatible = "fsl,ns16550", "ns16550a";
-+ reg = <0x00 0x21c0500 0x0 0x100>;
-+ interrupts = <0 54 IRQ_TYPE_LEVEL_HIGH>;
-+ clocks = <&clockgen 4 0>;
-+ status = "disabled";
-+ };
-+
-+ duart1: serial@21c0600 {
-+ compatible = "fsl,ns16550", "ns16550a";
-+ reg = <0x00 0x21c0600 0x0 0x100>;
-+ interrupts = <0 54 IRQ_TYPE_LEVEL_HIGH>;
-+ clocks = <&clockgen 4 0>;
-+ status = "disabled";
-+ };
-+
-+ gpio0: gpio@2300000 {
-+ compatible = "fsl,qoriq-gpio";
-+ reg = <0x0 0x2300000 0x0 0x10000>;
-+ interrupts = <0 66 IRQ_TYPE_LEVEL_HIGH>;
-+ gpio-controller;
-+ #gpio-cells = <2>;
-+ interrupt-controller;
-+ #interrupt-cells = <2>;
-+ };
-+
-+ gpio1: gpio@2310000 {
-+ compatible = "fsl,qoriq-gpio";
-+ reg = <0x0 0x2310000 0x0 0x10000>;
-+ interrupts = <0 67 IRQ_TYPE_LEVEL_HIGH>;
-+ gpio-controller;
-+ #gpio-cells = <2>;
-+ interrupt-controller;
-+ #interrupt-cells = <2>;
-+ };
-+
-+ qspi: quadspi@1550000 {
-+ compatible = "fsl,ls1012a-qspi", "fsl,ls1021a-qspi";
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ reg = <0x0 0x1550000 0x0 0x10000>,
-+ <0x0 0x40000000 0x0 0x10000000>;
-+ reg-names = "QuadSPI", "QuadSPI-memory";
-+ interrupts = <0 99 IRQ_TYPE_LEVEL_HIGH>;
-+ clock-names = "qspi_en", "qspi";
-+ clocks = <&clockgen 4 0>, <&clockgen 4 0>;
-+ big-endian;
-+ fsl,qspi-has-second-chip;
-+ status = "disabled";
-+ };
-+
-+ wdog0: wdog@2ad0000 {
-+ compatible = "fsl,ls1012a-wdt",
-+ "fsl,imx21-wdt";
-+ reg = <0x0 0x2ad0000 0x0 0x10000>;
-+ interrupts = <0 83 IRQ_TYPE_LEVEL_HIGH>;
-+ clocks = <&clockgen 4 0>;
-+ big-endian;
-+ };
-+
-+ sai1: sai@2b50000 {
-+ #sound-dai-cells = <0>;
-+ compatible = "fsl,vf610-sai";
-+ reg = <0x0 0x2b50000 0x0 0x10000>;
-+ interrupts = <0 148 IRQ_TYPE_LEVEL_HIGH>;
-+ clocks = <&clockgen 4 3>, <&clockgen 4 3>,
-+ <&clockgen 4 3>, <&clockgen 4 3>;
-+ clock-names = "bus", "mclk1", "mclk2", "mclk3";
-+ dma-names = "tx", "rx";
-+ dmas = <&edma0 1 47>,
-+ <&edma0 1 46>;
-+ status = "disabled";
-+ };
-+
-+ sai2: sai@2b60000 {
-+ #sound-dai-cells = <0>;
-+ compatible = "fsl,vf610-sai";
-+ reg = <0x0 0x2b60000 0x0 0x10000>;
-+ interrupts = <0 149 IRQ_TYPE_LEVEL_HIGH>;
-+ clocks = <&clockgen 4 3>, <&clockgen 4 3>,
-+ <&clockgen 4 3>, <&clockgen 4 3>;
-+ clock-names = "bus", "mclk1", "mclk2", "mclk3";
-+ dma-names = "tx", "rx";
-+ dmas = <&edma0 1 45>,
-+ <&edma0 1 44>;
-+ status = "disabled";
-+ };
-+
-+ edma0: edma@2c00000 {
-+ #dma-cells = <2>;
-+ compatible = "fsl,vf610-edma";
-+ reg = <0x0 0x2c00000 0x0 0x10000>,
-+ <0x0 0x2c10000 0x0 0x10000>,
-+ <0x0 0x2c20000 0x0 0x10000>;
-+ interrupts = <0 103 IRQ_TYPE_LEVEL_HIGH>,
-+ <0 103 IRQ_TYPE_LEVEL_HIGH>;
-+ interrupt-names = "edma-tx", "edma-err";
-+ dma-channels = <32>;
-+ big-endian;
-+ clock-names = "dmamux0", "dmamux1";
-+ clocks = <&clockgen 4 3>,
-+ <&clockgen 4 3>;
-+ };
-+
-+ usb0: usb3@2f00000 {
-+ compatible = "snps,dwc3";
-+ reg = <0x0 0x2f00000 0x0 0x10000>;
-+ interrupts = <0 60 0x4>;
-+ dr_mode = "host";
-+ snps,quirk-frame-length-adjustment = <0x20>;
-+ snps,dis_rxdet_inp3_quirk;
-+ };
-+
-+ usb1: usb2@8600000 {
-+ compatible = "fsl-usb2-dr-v2.5", "fsl-usb2-dr";
-+ reg = <0x0 0x8600000 0x0 0x1000>;
-+ interrupts = <0 139 0x4>;
-+ dr_mode = "host";
-+ phy_type = "ulpi";
-+ };
-+
-+ sata: sata@3200000 {
-+ compatible = "fsl,ls1012a-ahci", "fsl,ls1043a-ahci";
-+ reg = <0x0 0x3200000 0x0 0x10000>,
-+ <0x0 0x20140520 0x0 0x4>;
-+ reg-names = "ahci", "sata-ecc";
-+ interrupts = <0 69 IRQ_TYPE_LEVEL_HIGH>;
-+ clocks = <&clockgen 4 0>;
-+ dma-coherent;
-+ status = "disabled";
-+ };
-+
-+ msi: msi-controller1@1572000 {
-+ compatible = "fsl,ls1012a-msi";
-+ reg = <0x0 0x1572000 0x0 0x8>;
-+ msi-controller;
-+ interrupts = <0 126 IRQ_TYPE_LEVEL_HIGH>;
-+ };
-+
-+ pcie: pcie@3400000 {
-+ compatible = "fsl,ls1012a-pcie", "snps,dw-pcie";
-+ reg = <0x00 0x03400000 0x0 0x00100000 /* controller registers */
-+ 0x40 0x00000000 0x0 0x00002000>; /* configuration space */
-+ reg-names = "regs", "config";
-+ interrupts = <0 118 0x4>, /* AER interrupt */
-+ <0 117 0x4>; /* PME interrupt */
-+ interrupt-names = "aer", "pme";
-+ #address-cells = <3>;
-+ #size-cells = <2>;
-+ device_type = "pci";
-+ num-lanes = <4>;
-+ bus-range = <0x0 0xff>;
-+ ranges = <0x81000000 0x0 0x00000000 0x40 0x00010000 0x0 0x00010000 /* downstream I/O */
-+ 0x82000000 0x0 0x40000000 0x40 0x40000000 0x0 0x40000000>; /* non-prefetchable memory */
-+ msi-parent = <&msi>;
-+ #interrupt-cells = <1>;
-+ interrupt-map-mask = <0 0 0 7>;
-+ interrupt-map = <0000 0 0 1 &gic 0 110 IRQ_TYPE_LEVEL_HIGH>,
-+ <0000 0 0 2 &gic 0 111 IRQ_TYPE_LEVEL_HIGH>,
-+ <0000 0 0 3 &gic 0 112 IRQ_TYPE_LEVEL_HIGH>,
-+ <0000 0 0 4 &gic 0 113 IRQ_TYPE_LEVEL_HIGH>;
-+ status = "disabled";
-+ };
-+ };
-+
-+ reserved-memory {
-+ #address-cells = <2>;
-+ #size-cells = <2>;
-+ ranges;
-+
-+ pfe_reserved: packetbuffer@83400000 {
-+ reg = <0 0x83400000 0 0xc00000>;
-+ };
-+ };
-+
-+ pfe: pfe@04000000 {
-+ compatible = "fsl,pfe";
-+ reg = <0x0 0x04000000 0x0 0xc00000>, /* AXI 16M */
-+ <0x0 0x83400000 0x0 0xc00000>; /* PFE DDR 12M */
-+ reg-names = "pfe", "pfe-ddr";
-+ fsl,pfe-num-interfaces = <0x2>;
-+ interrupts = <0 172 0x4>, /* HIF interrupt */
-+ <0 173 0x4>, /*HIF_NOCPY interrupt */
-+ <0 174 0x4>; /* WoL interrupt */
-+ interrupt-names = "pfe_hif", "pfe_hif_nocpy", "pfe_wol";
-+ memory-region = <&pfe_reserved>;
-+ fsl,pfe-scfg = <&scfg 0>;
-+ fsl,rcpm-wakeup = <&rcpm 0xf0000020>;
-+ clocks = <&clockgen 4 0>;
-+ clock-names = "pfe";
-+
-+ status = "okay";
-+ pfe_mac0: ethernet@0 {
-+ };
-+
-+ pfe_mac1: ethernet@1 {
-+ };
-+ };
-+
-+ firmware {
-+ optee {
-+ compatible = "linaro,optee-tz";
-+ method = "smc";
-+ };
-+ };
-+};
---- /dev/null
-+++ b/arch/arm64/boot/dts/freescale/fsl-ls1043-post.dtsi
-@@ -0,0 +1,44 @@
-+// SPDX-License-Identifier: (GPL-2.0+ OR BSD-3-Clause)
-+/*
-+ * QorIQ FMan v3 device tree nodes for ls1043
-+ *
-+ * Copyright 2015-2016 Freescale Semiconductor Inc.
-+ */
-+
-+&soc {
-+
-+/* include used FMan blocks */
-+#include "qoriq-fman3-0.dtsi"
-+#include "qoriq-fman3-0-1g-0.dtsi"
-+#include "qoriq-fman3-0-1g-1.dtsi"
-+#include "qoriq-fman3-0-1g-2.dtsi"
-+#include "qoriq-fman3-0-1g-3.dtsi"
-+#include "qoriq-fman3-0-1g-4.dtsi"
-+#include "qoriq-fman3-0-1g-5.dtsi"
-+#include "qoriq-fman3-0-10g-0.dtsi"
-+
-+};
-+
-+&fman0 {
-+ /* these aliases provide the FMan ports mapping */
-+ enet0: ethernet@e0000 {
-+ };
-+
-+ enet1: ethernet@e2000 {
-+ };
-+
-+ enet2: ethernet@e4000 {
-+ };
-+
-+ enet3: ethernet@e6000 {
-+ };
-+
-+ enet4: ethernet@e8000 {
-+ };
-+
-+ enet5: ethernet@ea000 {
-+ };
-+
-+ enet6: ethernet@f0000 {
-+ };
-+};
---- /dev/null
-+++ b/arch/arm64/boot/dts/freescale/fsl-ls1043a-qds-sdk.dts
-@@ -0,0 +1,71 @@
-+/*
-+ * Device Tree Include file for Freescale Layerscape-1043A family SoC.
-+ *
-+ * Copyright 2014-2015 Freescale Semiconductor, Inc.
-+ *
-+ * Mingkai Hu <Mingkai.hu@freescale.com>
-+ *
-+ * This file is dual-licensed: you can use it either under the terms
-+ * of the GPLv2 or the X11 license, at your option. Note that this dual
-+ * licensing only applies to this file, and not this project as a
-+ * whole.
-+ *
-+ * a) This library is free software; you can redistribute it and/or
-+ * modify it under the terms of the GNU General Public License as
-+ * published by the Free Software Foundation; either version 2 of the
-+ * License, or (at your option) any later version.
-+ *
-+ * This library is distributed in the hope that it will be useful,
-+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
-+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
-+ * GNU General Public License for more details.
-+ *
-+ * Or, alternatively,
-+ *
-+ * b) Permission is hereby granted, free of charge, to any person
-+ * obtaining a copy of this software and associated documentation
-+ * files (the "Software"), to deal in the Software without
-+ * restriction, including without limitation the rights to use,
-+ * copy, modify, merge, publish, distribute, sublicense, and/or
-+ * sell copies of the Software, and to permit persons to whom the
-+ * Software is furnished to do so, subject to the following
-+ * conditions:
-+ *
-+ * The above copyright notice and this permission notice shall be
-+ * included in all copies or substantial portions of the Software.
-+ *
-+ * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
-+ * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES
-+ * OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
-+ * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT
-+ * HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY,
-+ * WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
-+ * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
-+ * OTHER DEALINGS IN THE SOFTWARE.
-+ */
-+
-+#include "fsl-ls1043a-qds.dts"
-+#include "qoriq-qman-portals-sdk.dtsi"
-+#include "qoriq-bman-portals-sdk.dtsi"
-+
-+&bman_fbpr {
-+ compatible = "fsl,bman-fbpr";
-+ alloc-ranges = <0 0 0x10000 0>;
-+};
-+&qman_fqd {
-+ compatible = "fsl,qman-fqd";
-+ alloc-ranges = <0 0 0x10000 0>;
-+};
-+&qman_pfdr {
-+ compatible = "fsl,qman-pfdr";
-+ alloc-ranges = <0 0 0x10000 0>;
-+};
-+
-+&soc {
-+#include "qoriq-dpaa-eth.dtsi"
-+#include "qoriq-fman3-0-6oh.dtsi"
-+};
-+
-+&fman0 {
-+ compatible = "fsl,fman", "simple-bus";
-+};
---- a/arch/arm64/boot/dts/freescale/fsl-ls1043a-qds.dts
-+++ b/arch/arm64/boot/dts/freescale/fsl-ls1043a-qds.dts
-@@ -1,51 +1,14 @@
-+// SPDX-License-Identifier: (GPL-2.0+ OR MIT)
- /*
- * Device Tree Include file for Freescale Layerscape-1043A family SoC.
- *
-- * Copyright 2014-2015, Freescale Semiconductor
-+ * Copyright 2014-2015 Freescale Semiconductor, Inc.
- *
- * Mingkai Hu <Mingkai.hu@freescale.com>
-- *
-- * This file is dual-licensed: you can use it either under the terms
-- * of the GPLv2 or the X11 license, at your option. Note that this dual
-- * licensing only applies to this file, and not this project as a
-- * whole.
-- *
-- * a) This library is free software; you can redistribute it and/or
-- * modify it under the terms of the GNU General Public License as
-- * published by the Free Software Foundation; either version 2 of the
-- * License, or (at your option) any later version.
-- *
-- * This library is distributed in the hope that it will be useful,
-- * but WITHOUT ANY WARRANTY; without even the implied warranty of
-- * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
-- * GNU General Public License for more details.
-- *
-- * Or, alternatively,
-- *
-- * b) Permission is hereby granted, free of charge, to any person
-- * obtaining a copy of this software and associated documentation
-- * files (the "Software"), to deal in the Software without
-- * restriction, including without limitation the rights to use,
-- * copy, modify, merge, publish, distribute, sublicense, and/or
-- * sell copies of the Software, and to permit persons to whom the
-- * Software is furnished to do so, subject to the following
-- * conditions:
-- *
-- * The above copyright notice and this permission notice shall be
-- * included in all copies or substantial portions of the Software.
-- *
-- * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
-- * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES
-- * OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
-- * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT
-- * HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY,
-- * WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
-- * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
-- * OTHER DEALINGS IN THE SOFTWARE.
- */
-
- /dts-v1/;
--/include/ "fsl-ls1043a.dtsi"
-+#include "fsl-ls1043a.dtsi"
-
- / {
- model = "LS1043A QDS Board";
-@@ -60,6 +23,22 @@
- serial1 = &duart1;
- serial2 = &duart2;
- serial3 = &duart3;
-+ sgmii_riser_s1_p1 = &sgmii_phy_s1_p1;
-+ sgmii_riser_s2_p1 = &sgmii_phy_s2_p1;
-+ sgmii_riser_s3_p1 = &sgmii_phy_s3_p1;
-+ sgmii_riser_s4_p1 = &sgmii_phy_s4_p1;
-+ qsgmii_s1_p1 = &qsgmii_phy_s1_p1;
-+ qsgmii_s1_p2 = &qsgmii_phy_s1_p2;
-+ qsgmii_s1_p3 = &qsgmii_phy_s1_p3;
-+ qsgmii_s1_p4 = &qsgmii_phy_s1_p4;
-+ qsgmii_s2_p1 = &qsgmii_phy_s2_p1;
-+ qsgmii_s2_p2 = &qsgmii_phy_s2_p2;
-+ qsgmii_s2_p3 = &qsgmii_phy_s2_p3;
-+ qsgmii_s2_p4 = &qsgmii_phy_s2_p4;
-+ emi1_slot1 = &ls1043mdio_s1;
-+ emi1_slot2 = &ls1043mdio_s2;
-+ emi1_slot3 = &ls1043mdio_s3;
-+ emi1_slot4 = &ls1043mdio_s4;
- };
-
- chosen {
-@@ -97,8 +76,11 @@
- };
-
- fpga: board-control@2,0 {
-- compatible = "fsl,ls1043aqds-fpga", "fsl,fpga-qixis";
-+ #address-cells = <1>;
-+ #size-cells = <1>;
-+ compatible = "fsl,ls1043aqds-fpga", "fsl,fpga-qixis", "simple-bus";
- reg = <0x2 0x0 0x0000100>;
-+ ranges = <0 2 0 0x100>;
- };
- };
-
-@@ -181,3 +163,149 @@
- reg = <0>;
- };
- };
-+
-+#include "fsl-ls1043-post.dtsi"
-+
-+&fman0 {
-+ ethernet@e0000 {
-+ phy-handle = <&qsgmii_phy_s2_p1>;
-+ phy-connection-type = "sgmii";
-+ };
-+
-+ ethernet@e2000 {
-+ phy-handle = <&qsgmii_phy_s2_p2>;
-+ phy-connection-type = "sgmii";
-+ };
-+
-+ ethernet@e4000 {
-+ phy-handle = <&rgmii_phy1>;
-+ phy-connection-type = "rgmii";
-+ };
-+
-+ ethernet@e6000 {
-+ phy-handle = <&rgmii_phy2>;
-+ phy-connection-type = "rgmii";
-+ };
-+
-+ ethernet@e8000 {
-+ phy-handle = <&qsgmii_phy_s2_p3>;
-+ phy-connection-type = "sgmii";
-+ };
-+
-+ ethernet@ea000 {
-+ phy-handle = <&qsgmii_phy_s2_p4>;
-+ phy-connection-type = "sgmii";
-+ };
-+
-+ ethernet@f0000 { /* DTSEC9/10GEC1 */
-+ fixed-link = <1 1 10000 0 0>;
-+ phy-connection-type = "xgmii";
-+ };
-+};
-+
-+&fpga {
-+ mdio-mux-emi1 {
-+ compatible = "mdio-mux-mmioreg", "mdio-mux";
-+ mdio-parent-bus = <&mdio0>;
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ reg = <0x54 1>; /* BRDCFG4 */
-+ mux-mask = <0xe0>; /* EMI1 */
-+
-+ /* On-board RGMII1 PHY */
-+ ls1043mdio0: mdio@0 {
-+ reg = <0>;
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+
-+ rgmii_phy1: ethernet-phy@1 { /* MAC3 */
-+ reg = <0x1>;
-+ };
-+ };
-+
-+ /* On-board RGMII2 PHY */
-+ ls1043mdio1: mdio@1 {
-+ reg = <0x20>;
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+
-+ rgmii_phy2: ethernet-phy@2 { /* MAC4 */
-+ reg = <0x2>;
-+ };
-+ };
-+
-+ /* Slot 1 */
-+ ls1043mdio_s1: mdio@2 {
-+ reg = <0x40>;
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ status = "disabled";
-+
-+ qsgmii_phy_s1_p1: ethernet-phy@4 {
-+ reg = <0x4>;
-+ };
-+ qsgmii_phy_s1_p2: ethernet-phy@5 {
-+ reg = <0x5>;
-+ };
-+ qsgmii_phy_s1_p3: ethernet-phy@6 {
-+ reg = <0x6>;
-+ };
-+ qsgmii_phy_s1_p4: ethernet-phy@7 {
-+ reg = <0x7>;
-+ };
-+
-+ sgmii_phy_s1_p1: ethernet-phy@1c {
-+ reg = <0x1c>;
-+ };
-+ };
-+
-+ /* Slot 2 */
-+ ls1043mdio_s2: mdio@3 {
-+ reg = <0x60>;
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ status = "disabled";
-+
-+ qsgmii_phy_s2_p1: ethernet-phy@8 {
-+ reg = <0x8>;
-+ };
-+ qsgmii_phy_s2_p2: ethernet-phy@9 {
-+ reg = <0x9>;
-+ };
-+ qsgmii_phy_s2_p3: ethernet-phy@a {
-+ reg = <0xa>;
-+ };
-+ qsgmii_phy_s2_p4: ethernet-phy@b {
-+ reg = <0xb>;
-+ };
-+
-+ sgmii_phy_s2_p1: ethernet-phy@1c {
-+ reg = <0x1c>;
-+ };
-+ };
-+
-+ /* Slot 3 */
-+ ls1043mdio_s3: mdio@4 {
-+ reg = <0x80>;
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ status = "disabled";
-+
-+ sgmii_phy_s3_p1: ethernet-phy@1c {
-+ reg = <0x1c>;
-+ };
-+ };
-+
-+ /* Slot 4 */
-+ ls1043mdio_s4: mdio@5 {
-+ reg = <0xa0>;
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ status = "disabled";
-+
-+ sgmii_phy_s4_p1: ethernet-phy@1c {
-+ reg = <0x1c>;
-+ };
-+ };
-+ };
-+};
---- /dev/null
-+++ b/arch/arm64/boot/dts/freescale/fsl-ls1043a-rdb-sdk.dts
-@@ -0,0 +1,71 @@
-+/*
-+ * Device Tree Include file for Freescale Layerscape-1043A family SoC.
-+ *
-+ * Copyright 2014-2015 Freescale Semiconductor, Inc.
-+ *
-+ * Mingkai Hu <Mingkai.hu@freescale.com>
-+ *
-+ * This file is dual-licensed: you can use it either under the terms
-+ * of the GPLv2 or the X11 license, at your option. Note that this dual
-+ * licensing only applies to this file, and not this project as a
-+ * whole.
-+ *
-+ * a) This library is free software; you can redistribute it and/or
-+ * modify it under the terms of the GNU General Public License as
-+ * published by the Free Software Foundation; either version 2 of the
-+ * License, or (at your option) any later version.
-+ *
-+ * This library is distributed in the hope that it will be useful,
-+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
-+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
-+ * GNU General Public License for more details.
-+ *
-+ * Or, alternatively,
-+ *
-+ * b) Permission is hereby granted, free of charge, to any person
-+ * obtaining a copy of this software and associated documentation
-+ * files (the "Software"), to deal in the Software without
-+ * restriction, including without limitation the rights to use,
-+ * copy, modify, merge, publish, distribute, sublicense, and/or
-+ * sell copies of the Software, and to permit persons to whom the
-+ * Software is furnished to do so, subject to the following
-+ * conditions:
-+ *
-+ * The above copyright notice and this permission notice shall be
-+ * included in all copies or substantial portions of the Software.
-+ *
-+ * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
-+ * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES
-+ * OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
-+ * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT
-+ * HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY,
-+ * WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
-+ * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
-+ * OTHER DEALINGS IN THE SOFTWARE.
-+ */
-+
-+#include "fsl-ls1043a-rdb.dts"
-+#include "qoriq-qman-portals-sdk.dtsi"
-+#include "qoriq-bman-portals-sdk.dtsi"
-+
-+&bman_fbpr {
-+ compatible = "fsl,bman-fbpr";
-+ alloc-ranges = <0 0 0x10000 0>;
-+};
-+&qman_fqd {
-+ compatible = "fsl,qman-fqd";
-+ alloc-ranges = <0 0 0x10000 0>;
-+};
-+&qman_pfdr {
-+ compatible = "fsl,qman-pfdr";
-+ alloc-ranges = <0 0 0x10000 0>;
-+};
-+
-+&soc {
-+#include "qoriq-dpaa-eth.dtsi"
-+#include "qoriq-fman3-0-6oh.dtsi"
-+};
-+
-+&fman0 {
-+ compatible = "fsl,fman", "simple-bus";
-+};
---- /dev/null
-+++ b/arch/arm64/boot/dts/freescale/fsl-ls1043a-rdb-usdpaa.dts
-@@ -0,0 +1,117 @@
-+/*
-+ * Device Tree Include file for Freescale Layerscape-1043A family SoC.
-+ *
-+ * Copyright 2014-2015 Freescale Semiconductor, Inc.
-+ *
-+ * This file is licensed under the terms of the GNU General Public
-+ * License version 2. This program is licensed "as is" without any
-+ * warranty of any kind, whether express or implied.
-+ */
-+
-+#include "fsl-ls1043a-rdb-sdk.dts"
-+
-+&soc {
-+ bp7: buffer-pool@7 {
-+ compatible = "fsl,p4080-bpool", "fsl,bpool";
-+ fsl,bpid = <7>;
-+ fsl,bpool-ethernet-cfg = <0 0 0 192 0 0xdeadbeef>;
-+ fsl,bpool-thresholds = <0x400 0xc00 0x0 0x0>;
-+ };
-+
-+ bp8: buffer-pool@8 {
-+ compatible = "fsl,p4080-bpool", "fsl,bpool";
-+ fsl,bpid = <8>;
-+ fsl,bpool-ethernet-cfg = <0 0 0 576 0 0xabbaf00d>;
-+ fsl,bpool-thresholds = <0x100 0x300 0x0 0x0>;
-+ };
-+
-+ bp9: buffer-pool@9 {
-+ compatible = "fsl,p4080-bpool", "fsl,bpool";
-+ fsl,bpid = <9>;
-+ fsl,bpool-ethernet-cfg = <0 0 0 2048 0 0xfeedabba>;
-+ fsl,bpool-thresholds = <0x100 0x300 0x0 0x0>;
-+ };
-+
-+ fsl,dpaa {
-+ compatible = "fsl,ls1043a", "fsl,dpaa", "simple-bus";
-+
-+ ethernet@0 {
-+ compatible = "fsl,dpa-ethernet-init";
-+ fsl,bman-buffer-pools = <&bp7 &bp8 &bp9>;
-+ fsl,qman-frame-queues-rx = <0x50 1 0x51 1>;
-+ fsl,qman-frame-queues-tx = <0x70 1 0x71 1>;
-+ };
-+
-+ ethernet@1 {
-+ compatible = "fsl,dpa-ethernet-init";
-+ fsl,bman-buffer-pools = <&bp7 &bp8 &bp9>;
-+ fsl,qman-frame-queues-rx = <0x52 1 0x53 1>;
-+ fsl,qman-frame-queues-tx = <0x72 1 0x73 1>;
-+ };
-+
-+ ethernet@2 {
-+ compatible = "fsl,dpa-ethernet-init";
-+ fsl,bman-buffer-pools = <&bp7 &bp8 &bp9>;
-+ fsl,qman-frame-queues-rx = <0x54 1 0x55 1>;
-+ fsl,qman-frame-queues-tx = <0x74 1 0x75 1>;
-+ };
-+
-+ ethernet@3 {
-+ compatible = "fsl,dpa-ethernet-init";
-+ fsl,bman-buffer-pools = <&bp7 &bp8 &bp9>;
-+ fsl,qman-frame-queues-rx = <0x56 1 0x57 1>;
-+ fsl,qman-frame-queues-tx = <0x76 1 0x77 1>;
-+ };
-+
-+ ethernet@4 {
-+ compatible = "fsl,dpa-ethernet-init";
-+ fsl,bman-buffer-pools = <&bp7 &bp8 &bp9>;
-+ fsl,qman-frame-queues-rx = <0x58 1 0x59 1>;
-+ fsl,qman-frame-queues-tx = <0x78 1 0x79 1>;
-+ };
-+
-+ ethernet@5 {
-+ compatible = "fsl,dpa-ethernet-init";
-+ fsl,bman-buffer-pools = <&bp7 &bp8 &bp9>;
-+ fsl,qman-frame-queues-rx = <0x60 1 0x61 1>;
-+ fsl,qman-frame-queues-tx = <0x80 1 0x81 1>;
-+ };
-+
-+ ethernet@8 {
-+ compatible = "fsl,dpa-ethernet-init";
-+ fsl,bman-buffer-pools = <&bp7 &bp8 &bp9>;
-+ fsl,qman-frame-queues-rx = <0x5c 1 0x5d 1>;
-+ fsl,qman-frame-queues-tx = <0x7c 1 0x7d 1>;
-+
-+ };
-+ dpa-fman0-oh@2 {
-+ compatible = "fsl,dpa-oh";
-+ /* Define frame queues for the OH port*/
-+ /* <OH Rx error, OH Rx default> */
-+ fsl,qman-frame-queues-oh = <0x5a 1 0x5b 1>;
-+ fsl,fman-oh-port = <&fman0_oh2>;
-+ };
-+ };
-+};
-+/ {
-+ reserved-memory {
-+ #address-cells = <2>;
-+ #size-cells = <2>;
-+ ranges;
-+
-+ usdpaa_mem: usdpaa_mem {
-+ compatible = "fsl,usdpaa-mem";
-+ alloc-ranges = <0 0 0x10000 0>;
-+ size = <0 0x10000000>;
-+ alignment = <0 0x10000000>;
-+ };
-+ };
-+};
-+
-+&fman0 {
-+ fman0_oh2: port@83000 {
-+ cell-index = <1>;
-+ compatible = "fsl,fman-port-oh";
-+ reg = <0x83000 0x1000>;
-+ };
-+};
---- a/arch/arm64/boot/dts/freescale/fsl-ls1043a-rdb.dts
-+++ b/arch/arm64/boot/dts/freescale/fsl-ls1043a-rdb.dts
-@@ -1,51 +1,14 @@
-+// SPDX-License-Identifier: (GPL-2.0+ OR MIT)
- /*
- * Device Tree Include file for Freescale Layerscape-1043A family SoC.
- *
-- * Copyright 2014-2015, Freescale Semiconductor
-+ * Copyright 2014-2015 Freescale Semiconductor, Inc.
- *
- * Mingkai Hu <Mingkai.hu@freescale.com>
-- *
-- * This file is dual-licensed: you can use it either under the terms
-- * of the GPLv2 or the X11 license, at your option. Note that this dual
-- * licensing only applies to this file, and not this project as a
-- * whole.
-- *
-- * a) This library is free software; you can redistribute it and/or
-- * modify it under the terms of the GNU General Public License as
-- * published by the Free Software Foundation; either version 2 of the
-- * License, or (at your option) any later version.
-- *
-- * This library is distributed in the hope that it will be useful,
-- * but WITHOUT ANY WARRANTY; without even the implied warranty of
-- * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
-- * GNU General Public License for more details.
-- *
-- * Or, alternatively,
-- *
-- * b) Permission is hereby granted, free of charge, to any person
-- * obtaining a copy of this software and associated documentation
-- * files (the "Software"), to deal in the Software without
-- * restriction, including without limitation the rights to use,
-- * copy, modify, merge, publish, distribute, sublicense, and/or
-- * sell copies of the Software, and to permit persons to whom the
-- * Software is furnished to do so, subject to the following
-- * conditions:
-- *
-- * The above copyright notice and this permission notice shall be
-- * included in all copies or substantial portions of the Software.
-- *
-- * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
-- * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES
-- * OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
-- * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT
-- * HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY,
-- * WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
-- * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
-- * OTHER DEALINGS IN THE SOFTWARE.
- */
-
- /dts-v1/;
--/include/ "fsl-ls1043a.dtsi"
-+#include "fsl-ls1043a.dtsi"
-
- / {
- model = "LS1043A RDB Board";
-@@ -86,6 +49,10 @@
- compatible = "pericom,pt7c4338";
- reg = <0x68>;
- };
-+ rtc@51 {
-+ compatible = "nxp,pcf85263";
-+ reg = <0x51>;
-+ };
- };
-
- &ifc {
-@@ -130,6 +97,38 @@
- reg = <0>;
- spi-max-frequency = <1000000>; /* input clock */
- };
-+
-+ slic@2 {
-+ compatible = "maxim,ds26522";
-+ reg = <2>;
-+ spi-max-frequency = <2000000>;
-+ fsl,spi-cs-sck-delay = <100>;
-+ fsl,spi-sck-cs-delay = <50>;
-+ };
-+
-+ slic@3 {
-+ compatible = "maxim,ds26522";
-+ reg = <3>;
-+ spi-max-frequency = <2000000>;
-+ fsl,spi-cs-sck-delay = <100>;
-+ fsl,spi-sck-cs-delay = <50>;
-+ };
-+};
-+
-+&uqe {
-+ ucc_hdlc: ucc@2000 {
-+ compatible = "fsl,ucc-hdlc";
-+ rx-clock-name = "clk8";
-+ tx-clock-name = "clk9";
-+ fsl,rx-sync-clock = "rsync_pin";
-+ fsl,tx-sync-clock = "tsync_pin";
-+ fsl,tx-timeslot-mask = <0xfffffffe>;
-+ fsl,rx-timeslot-mask = <0xfffffffe>;
-+ fsl,tdm-framer-type = "e1";
-+ fsl,tdm-id = <0>;
-+ fsl,siram-entry-id = <0>;
-+ fsl,tdm-interface;
-+ };
- };
-
- &duart0 {
-@@ -139,3 +138,76 @@
- &duart1 {
- status = "okay";
- };
-+
-+#include "fsl-ls1043-post.dtsi"
-+
-+&fman0 {
-+ ethernet@e0000 {
-+ phy-handle = <&qsgmii_phy1>;
-+ phy-connection-type = "qsgmii";
-+ };
-+
-+ ethernet@e2000 {
-+ phy-handle = <&qsgmii_phy2>;
-+ phy-connection-type = "qsgmii";
-+ };
-+
-+ ethernet@e4000 {
-+ phy-handle = <&rgmii_phy1>;
-+ phy-connection-type = "rgmii-txid";
-+ };
-+
-+ ethernet@e6000 {
-+ phy-handle = <&rgmii_phy2>;
-+ phy-connection-type = "rgmii-txid";
-+ };
-+
-+ ethernet@e8000 {
-+ phy-handle = <&qsgmii_phy3>;
-+ phy-connection-type = "qsgmii";
-+ };
-+
-+ ethernet@ea000 {
-+ phy-handle = <&qsgmii_phy4>;
-+ phy-connection-type = "qsgmii";
-+ };
-+
-+ ethernet@f0000 { /* 10GEC1 */
-+ phy-handle = <&aqr105_phy>;
-+ phy-connection-type = "xgmii";
-+ };
-+
-+ mdio@fc000 {
-+ rgmii_phy1: ethernet-phy@1 {
-+ reg = <0x1>;
-+ };
-+
-+ rgmii_phy2: ethernet-phy@2 {
-+ reg = <0x2>;
-+ };
-+
-+ qsgmii_phy1: ethernet-phy@4 {
-+ reg = <0x4>;
-+ };
-+
-+ qsgmii_phy2: ethernet-phy@5 {
-+ reg = <0x5>;
-+ };
-+
-+ qsgmii_phy3: ethernet-phy@6 {
-+ reg = <0x6>;
-+ };
-+
-+ qsgmii_phy4: ethernet-phy@7 {
-+ reg = <0x7>;
-+ };
-+ };
-+
-+ mdio@fd000 {
-+ aqr105_phy: ethernet-phy@1 {
-+ compatible = "ethernet-phy-ieee802.3-c45";
-+ interrupts = <0 132 4>;
-+ reg = <0x1>;
-+ };
-+ };
-+};
---- a/arch/arm64/boot/dts/freescale/fsl-ls1043a.dtsi
-+++ b/arch/arm64/boot/dts/freescale/fsl-ls1043a.dtsi
-@@ -1,55 +1,32 @@
-+// SPDX-License-Identifier: (GPL-2.0+ OR MIT)
- /*
- * Device Tree Include file for Freescale Layerscape-1043A family SoC.
- *
-- * Copyright 2014-2015, Freescale Semiconductor
-+ * Copyright 2014-2015 Freescale Semiconductor, Inc.
- *
- * Mingkai Hu <Mingkai.hu@freescale.com>
-- *
-- * This file is dual-licensed: you can use it either under the terms
-- * of the GPLv2 or the X11 license, at your option. Note that this dual
-- * licensing only applies to this file, and not this project as a
-- * whole.
-- *
-- * a) This library is free software; you can redistribute it and/or
-- * modify it under the terms of the GNU General Public License as
-- * published by the Free Software Foundation; either version 2 of the
-- * License, or (at your option) any later version.
-- *
-- * This library is distributed in the hope that it will be useful,
-- * but WITHOUT ANY WARRANTY; without even the implied warranty of
-- * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
-- * GNU General Public License for more details.
-- *
-- * Or, alternatively,
-- *
-- * b) Permission is hereby granted, free of charge, to any person
-- * obtaining a copy of this software and associated documentation
-- * files (the "Software"), to deal in the Software without
-- * restriction, including without limitation the rights to use,
-- * copy, modify, merge, publish, distribute, sublicense, and/or
-- * sell copies of the Software, and to permit persons to whom the
-- * Software is furnished to do so, subject to the following
-- * conditions:
-- *
-- * The above copyright notice and this permission notice shall be
-- * included in all copies or substantial portions of the Software.
-- *
-- * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
-- * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES
-- * OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
-- * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT
-- * HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY,
-- * WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
-- * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
-- * OTHER DEALINGS IN THE SOFTWARE.
- */
-
-+#include <dt-bindings/thermal/thermal.h>
-+#include <dt-bindings/interrupt-controller/arm-gic.h>
-+
- / {
- compatible = "fsl,ls1043a";
- interrupt-parent = <&gic>;
- #address-cells = <2>;
- #size-cells = <2>;
-
-+ aliases {
-+ fman0 = &fman0;
-+ ethernet0 = &enet0;
-+ ethernet1 = &enet1;
-+ ethernet2 = &enet2;
-+ ethernet3 = &enet3;
-+ ethernet4 = &enet4;
-+ ethernet5 = &enet5;
-+ ethernet6 = &enet6;
-+ };
-+
- cpus {
- #address-cells = <1>;
- #size-cells = <0>;
-@@ -66,6 +43,8 @@
- reg = <0x0>;
- clocks = <&clockgen 1 0>;
- next-level-cache = <&l2>;
-+ #cooling-cells = <2>;
-+ cpu-idle-states = <&CPU_PH20>;
- };
-
- cpu1: cpu@1 {
-@@ -74,6 +53,7 @@
- reg = <0x1>;
- clocks = <&clockgen 1 0>;
- next-level-cache = <&l2>;
-+ cpu-idle-states = <&CPU_PH20>;
- };
-
- cpu2: cpu@2 {
-@@ -82,6 +62,7 @@
- reg = <0x2>;
- clocks = <&clockgen 1 0>;
- next-level-cache = <&l2>;
-+ cpu-idle-states = <&CPU_PH20>;
- };
-
- cpu3: cpu@3 {
-@@ -90,6 +71,7 @@
- reg = <0x3>;
- clocks = <&clockgen 1 0>;
- next-level-cache = <&l2>;
-+ cpu-idle-states = <&CPU_PH20>;
- };
-
- l2: l2-cache {
-@@ -97,12 +79,56 @@
- };
- };
-
-+ idle-states {
-+ /*
-+ * PSCI node is not added default, U-boot will add missing
-+ * parts if it determines to use PSCI.
-+ */
-+ entry-method = "arm,psci";
-+
-+ CPU_PH20: cpu-ph20 {
-+ compatible = "arm,idle-state";
-+ idle-state-name = "PH20";
-+ arm,psci-suspend-param = <0x0>;
-+ entry-latency-us = <1000>;
-+ exit-latency-us = <1000>;
-+ min-residency-us = <3000>;
-+ };
-+ };
-+
- memory@80000000 {
- device_type = "memory";
- reg = <0x0 0x80000000 0 0x80000000>;
- /* DRAM space 1, size: 2GiB DRAM */
- };
-
-+ reserved-memory {
-+ #address-cells = <2>;
-+ #size-cells = <2>;
-+ ranges;
-+
-+ bman_fbpr: bman-fbpr {
-+ compatible = "shared-dma-pool";
-+ size = <0 0x1000000>;
-+ alignment = <0 0x1000000>;
-+ no-map;
-+ };
-+
-+ qman_fqd: qman-fqd {
-+ compatible = "shared-dma-pool";
-+ size = <0 0x400000>;
-+ alignment = <0 0x400000>;
-+ no-map;
-+ };
-+
-+ qman_pfdr: qman-pfdr {
-+ compatible = "shared-dma-pool";
-+ size = <0 0x2000000>;
-+ alignment = <0 0x2000000>;
-+ no-map;
-+ };
-+ };
-+
- sysclk: sysclk {
- compatible = "fixed-clock";
- #clock-cells = <0>;
-@@ -149,7 +175,7 @@
- interrupts = <1 9 0xf08>;
- };
-
-- soc {
-+ soc: soc {
- compatible = "simple-bus";
- #address-cells = <2>;
- #size-cells = <2>;
-@@ -213,13 +239,14 @@
-
- dcfg: dcfg@1ee0000 {
- compatible = "fsl,ls1043a-dcfg", "syscon";
-- reg = <0x0 0x1ee0000 0x0 0x10000>;
-+ reg = <0x0 0x1ee0000 0x0 0x1000>;
- big-endian;
- };
-
- ifc: ifc@1530000 {
- compatible = "fsl,ifc", "simple-bus";
- reg = <0x0 0x1530000 0x0 0x10000>;
-+ big-endian;
- interrupts = <0 43 0x4>;
- };
-
-@@ -255,6 +282,103 @@
- big-endian;
- };
-
-+ tmu: tmu@1f00000 {
-+ compatible = "fsl,qoriq-tmu";
-+ reg = <0x0 0x1f00000 0x0 0x10000>;
-+ interrupts = <0 33 0x4>;
-+ fsl,tmu-range = <0xb0000 0x9002a 0x6004c 0x30062>;
-+ fsl,tmu-calibration = <0x00000000 0x00000026
-+ 0x00000001 0x0000002d
-+ 0x00000002 0x00000032
-+ 0x00000003 0x00000039
-+ 0x00000004 0x0000003f
-+ 0x00000005 0x00000046
-+ 0x00000006 0x0000004d
-+ 0x00000007 0x00000054
-+ 0x00000008 0x0000005a
-+ 0x00000009 0x00000061
-+ 0x0000000a 0x0000006a
-+ 0x0000000b 0x00000071
-+
-+ 0x00010000 0x00000025
-+ 0x00010001 0x0000002c
-+ 0x00010002 0x00000035
-+ 0x00010003 0x0000003d
-+ 0x00010004 0x00000045
-+ 0x00010005 0x0000004e
-+ 0x00010006 0x00000057
-+ 0x00010007 0x00000061
-+ 0x00010008 0x0000006b
-+ 0x00010009 0x00000076
-+
-+ 0x00020000 0x00000029
-+ 0x00020001 0x00000033
-+ 0x00020002 0x0000003d
-+ 0x00020003 0x00000049
-+ 0x00020004 0x00000056
-+ 0x00020005 0x00000061
-+ 0x00020006 0x0000006d
-+
-+ 0x00030000 0x00000021
-+ 0x00030001 0x0000002a
-+ 0x00030002 0x0000003c
-+ 0x00030003 0x0000004e>;
-+ #thermal-sensor-cells = <1>;
-+ };
-+
-+ thermal-zones {
-+ cpu_thermal: cpu-thermal {
-+ polling-delay-passive = <1000>;
-+ polling-delay = <5000>;
-+
-+ thermal-sensors = <&tmu 3>;
-+
-+ trips {
-+ cpu_alert: cpu-alert {
-+ temperature = <85000>;
-+ hysteresis = <2000>;
-+ type = "passive";
-+ };
-+ cpu_crit: cpu-crit {
-+ temperature = <95000>;
-+ hysteresis = <2000>;
-+ type = "critical";
-+ };
-+ };
-+
-+ cooling-maps {
-+ map0 {
-+ trip = <&cpu_alert>;
-+ cooling-device =
-+ <&cpu0 THERMAL_NO_LIMIT
-+ THERMAL_NO_LIMIT>;
-+ };
-+ };
-+ };
-+ };
-+
-+ qman: qman@1880000 {
-+ compatible = "fsl,qman";
-+ reg = <0x00 0x1880000 0x0 0x10000>;
-+ interrupts = <0 45 0x4>;
-+ memory-region = <&qman_fqd &qman_pfdr>;
-+ };
-+
-+ bman: bman@1890000 {
-+ compatible = "fsl,bman";
-+ reg = <0x00 0x1890000 0x0 0x10000>;
-+ interrupts = <0 45 0x4>;
-+ memory-region = <&bman_fbpr>;
-+ };
-+
-+ bportals: bman-portals@508000000 {
-+ ranges = <0x0 0x5 0x08000000 0x8000000>;
-+ };
-+
-+ qportals: qman-portals@500000000 {
-+ ranges = <0x0 0x5 0x00000000 0x8000000>;
-+ };
-+
- dspi0: dspi@2100000 {
- compatible = "fsl,ls1043a-dspi", "fsl,ls1021a-v1.0-dspi";
- #address-cells = <1>;
-@@ -282,7 +406,7 @@
- };
-
- i2c0: i2c@2180000 {
-- compatible = "fsl,vf610-i2c";
-+ compatible = "fsl,vf610-i2c", "fsl,ls1043a-vf610-i2c";
- #address-cells = <1>;
- #size-cells = <0>;
- reg = <0x0 0x2180000 0x0 0x10000>;
-@@ -292,6 +416,7 @@
- dmas = <&edma0 1 39>,
- <&edma0 1 38>;
- dma-names = "tx", "rx";
-+ fsl-scl-gpio = <&gpio4 12 0>;
- status = "disabled";
- };
-
-@@ -396,6 +521,72 @@
- #interrupt-cells = <2>;
- };
-
-+ uqe: uqe@2400000 {
-+ #address-cells = <1>;
-+ #size-cells = <1>;
-+ device_type = "qe";
-+ compatible = "fsl,qe", "simple-bus";
-+ ranges = <0x0 0x0 0x2400000 0x40000>;
-+ reg = <0x0 0x2400000 0x0 0x480>;
-+ brg-frequency = <100000000>;
-+ bus-frequency = <200000000>;
-+
-+ fsl,qe-num-riscs = <1>;
-+ fsl,qe-num-snums = <28>;
-+
-+ qeic: qeic@80 {
-+ compatible = "fsl,qe-ic";
-+ reg = <0x80 0x80>;
-+ #address-cells = <0>;
-+ interrupt-controller;
-+ #interrupt-cells = <1>;
-+ interrupts = <0 77 0x04 0 77 0x04>;
-+ };
-+
-+ si1: si@700 {
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ compatible = "fsl,ls1043-qe-si",
-+ "fsl,t1040-qe-si";
-+ reg = <0x700 0x80>;
-+ };
-+
-+ siram1: siram@1000 {
-+ #address-cells = <1>;
-+ #size-cells = <1>;
-+ compatible = "fsl,ls1043-qe-siram",
-+ "fsl,t1040-qe-siram";
-+ reg = <0x1000 0x800>;
-+ };
-+
-+ ucc@2000 {
-+ cell-index = <1>;
-+ reg = <0x2000 0x200>;
-+ interrupts = <32>;
-+ interrupt-parent = <&qeic>;
-+ };
-+
-+ ucc@2200 {
-+ cell-index = <3>;
-+ reg = <0x2200 0x200>;
-+ interrupts = <34>;
-+ interrupt-parent = <&qeic>;
-+ };
-+
-+ muram@10000 {
-+ #address-cells = <1>;
-+ #size-cells = <1>;
-+ compatible = "fsl,qe-muram", "fsl,cpm-muram";
-+ ranges = <0x0 0x10000 0x6000>;
-+
-+ data-only@0 {
-+ compatible = "fsl,qe-muram-data",
-+ "fsl,cpm-muram-data";
-+ reg = <0x0 0x6000>;
-+ };
-+ };
-+ };
-+
- lpuart0: serial@2950000 {
- compatible = "fsl,ls1021a-lpuart";
- reg = <0x0 0x2950000 0x0 0x1000>;
-@@ -450,6 +641,16 @@
- status = "disabled";
- };
-
-+ ftm0: ftm0@29d0000 {
-+ compatible = "fsl,ls1043a-ftm";
-+ reg = <0x0 0x29d0000 0x0 0x10000>,
-+ <0x0 0x1ee2140 0x0 0x4>;
-+ reg-names = "ftm", "FlexTimer1";
-+ interrupts = <0 86 0x4>;
-+ big-endian;
-+ status = "okay";
-+ };
-+
- wdog0: wdog@2ad0000 {
- compatible = "fsl,ls1043a-wdt", "fsl,imx21-wdt";
- reg = <0x0 0x2ad0000 0x0 0x10000>;
-@@ -482,6 +683,10 @@
- dr_mode = "host";
- snps,quirk-frame-length-adjustment = <0x20>;
- snps,dis_rxdet_inp3_quirk;
-+ snps,incr-burst-type-adjustment = <1>, <4>, <8>, <16>;
-+ snps,dma-snooping;
-+ usb3-lpm-capable;
-+ snps,dis-u1u2-when-u3-quirk;
- };
-
- usb1: usb3@3000000 {
-@@ -491,6 +696,11 @@
- dr_mode = "host";
- snps,quirk-frame-length-adjustment = <0x20>;
- snps,dis_rxdet_inp3_quirk;
-+ snps,incr-burst-type-adjustment = <1>, <4>, <8>, <16>;
-+ snps,dma-snooping;
-+ configure-gfladj;
-+ usb3-lpm-capable;
-+ snps,dis-u1u2-when-u3-quirk;
- };
-
- usb2: usb3@3100000 {
-@@ -500,32 +710,54 @@
- dr_mode = "host";
- snps,quirk-frame-length-adjustment = <0x20>;
- snps,dis_rxdet_inp3_quirk;
-+ snps,incr-burst-type-adjustment = <1>, <4>, <8>, <16>;
-+ snps,dma-snooping;
-+ configure-gfladj;
-+ usb3-lpm-capable;
-+ snps,dis-u1u2-when-u3-quirk;
- };
-
- sata: sata@3200000 {
- compatible = "fsl,ls1043a-ahci";
-- reg = <0x0 0x3200000 0x0 0x10000>;
-+ reg = <0x0 0x3200000 0x0 0x10000>,
-+ <0x0 0x20140520 0x0 0x4>;
-+ reg-names = "ahci", "sata-ecc";
- interrupts = <0 69 0x4>;
- clocks = <&clockgen 4 0>;
- dma-coherent;
- };
-
-+ qdma: qdma@8380000 {
-+ compatible = "fsl,ls1021a-qdma", "fsl,ls1043a-qdma";
-+ reg = <0x0 0x8380000 0x0 0x1000>, /* Controller regs */
-+ <0x0 0x8390000 0x0 0x10000>, /* Status regs */
-+ <0x0 0x83a0000 0x0 0x40000>; /* Block regs */
-+ interrupts = <0 152 0x4>,
-+ <0 39 0x4>;
-+ interrupt-names = "qdma-error", "qdma-queue";
-+ channels = <8>;
-+ queues = <2>;
-+ status-sizes = <64>;
-+ queue-sizes = <64 64>;
-+ big-endian;
-+ };
-+
- msi1: msi-controller1@1571000 {
-- compatible = "fsl,1s1043a-msi";
-+ compatible = "fsl,ls1043a-msi";
- reg = <0x0 0x1571000 0x0 0x8>;
- msi-controller;
- interrupts = <0 116 0x4>;
- };
-
- msi2: msi-controller2@1572000 {
-- compatible = "fsl,1s1043a-msi";
-+ compatible = "fsl,ls1043a-msi";
- reg = <0x0 0x1572000 0x0 0x8>;
- msi-controller;
- interrupts = <0 126 0x4>;
- };
-
- msi3: msi-controller3@1573000 {
-- compatible = "fsl,1s1043a-msi";
-+ compatible = "fsl,ls1043a-msi";
- reg = <0x0 0x1573000 0x0 0x8>;
- msi-controller;
- interrupts = <0 160 0x4>;
-@@ -536,9 +768,9 @@
- reg = <0x00 0x03400000 0x0 0x00100000 /* controller registers */
- 0x40 0x00000000 0x0 0x00002000>; /* configuration space */
- reg-names = "regs", "config";
-- interrupts = <0 118 0x4>, /* controller interrupt */
-- <0 117 0x4>; /* PME interrupt */
-- interrupt-names = "intr", "pme";
-+ interrupts = <0 117 0x4>, /* PME interrupt */
-+ <0 118 0x4>; /* aer interrupt */
-+ interrupt-names = "pme", "aer";
- #address-cells = <3>;
- #size-cells = <2>;
- device_type = "pci";
-@@ -547,7 +779,7 @@
- bus-range = <0x0 0xff>;
- ranges = <0x81000000 0x0 0x00000000 0x40 0x00010000 0x0 0x00010000 /* downstream I/O */
- 0x82000000 0x0 0x40000000 0x40 0x40000000 0x0 0x40000000>; /* non-prefetchable memory */
-- msi-parent = <&msi1>;
-+ msi-parent = <&msi1>, <&msi2>, <&msi3>;
- #interrupt-cells = <1>;
- interrupt-map-mask = <0 0 0 7>;
- interrupt-map = <0000 0 0 1 &gic 0 110 0x4>,
-@@ -561,9 +793,9 @@
- reg = <0x00 0x03500000 0x0 0x00100000 /* controller registers */
- 0x48 0x00000000 0x0 0x00002000>; /* configuration space */
- reg-names = "regs", "config";
-- interrupts = <0 128 0x4>,
-- <0 127 0x4>;
-- interrupt-names = "intr", "pme";
-+ interrupts = <0 127 0x4>,
-+ <0 128 0x4>;
-+ interrupt-names = "pme", "aer";
- #address-cells = <3>;
- #size-cells = <2>;
- device_type = "pci";
-@@ -572,7 +804,7 @@
- bus-range = <0x0 0xff>;
- ranges = <0x81000000 0x0 0x00000000 0x48 0x00010000 0x0 0x00010000 /* downstream I/O */
- 0x82000000 0x0 0x40000000 0x48 0x40000000 0x0 0x40000000>; /* non-prefetchable memory */
-- msi-parent = <&msi2>;
-+ msi-parent = <&msi1>, <&msi2>, <&msi3>;
- #interrupt-cells = <1>;
- interrupt-map-mask = <0 0 0 7>;
- interrupt-map = <0000 0 0 1 &gic 0 120 0x4>,
-@@ -586,9 +818,9 @@
- reg = <0x00 0x03600000 0x0 0x00100000 /* controller registers */
- 0x50 0x00000000 0x0 0x00002000>; /* configuration space */
- reg-names = "regs", "config";
-- interrupts = <0 162 0x4>,
-- <0 161 0x4>;
-- interrupt-names = "intr", "pme";
-+ interrupts = <0 161 0x4>,
-+ <0 162 0x4>;
-+ interrupt-names = "pme", "aer";
- #address-cells = <3>;
- #size-cells = <2>;
- device_type = "pci";
-@@ -597,7 +829,7 @@
- bus-range = <0x0 0xff>;
- ranges = <0x81000000 0x0 0x00000000 0x50 0x00010000 0x0 0x00010000 /* downstream I/O */
- 0x82000000 0x0 0x40000000 0x50 0x40000000 0x0 0x40000000>; /* non-prefetchable memory */
-- msi-parent = <&msi3>;
-+ msi-parent = <&msi1>, <&msi2>, <&msi3>;
- #interrupt-cells = <1>;
- interrupt-map-mask = <0 0 0 7>;
- interrupt-map = <0000 0 0 1 &gic 0 154 0x4>,
-@@ -607,4 +839,13 @@
- };
- };
-
-+ firmware {
-+ optee {
-+ compatible = "linaro,optee-tz";
-+ method = "smc";
-+ };
-+ };
- };
-+
-+#include "qoriq-qman-portals.dtsi"
-+#include "qoriq-bman-portals.dtsi"
---- /dev/null
-+++ b/arch/arm64/boot/dts/freescale/fsl-ls1046-post.dtsi
-@@ -0,0 +1,48 @@
-+// SPDX-License-Identifier: (GPL-2.0+ OR BSD-3-Clause)
-+/*
-+ * QorIQ FMan v3 device tree nodes for ls1046
-+ *
-+ * Copyright 2015-2016 Freescale Semiconductor Inc.
-+ *
-+ */
-+
-+&soc {
-+
-+/* include used FMan blocks */
-+#include "qoriq-fman3-0.dtsi"
-+#include "qoriq-fman3-0-1g-0.dtsi"
-+#include "qoriq-fman3-0-1g-1.dtsi"
-+#include "qoriq-fman3-0-1g-2.dtsi"
-+#include "qoriq-fman3-0-1g-3.dtsi"
-+#include "qoriq-fman3-0-1g-4.dtsi"
-+#include "qoriq-fman3-0-1g-5.dtsi"
-+#include "qoriq-fman3-0-10g-0.dtsi"
-+#include "qoriq-fman3-0-10g-1.dtsi"
-+};
-+
-+&fman0 {
-+ /* these aliases provide the FMan ports mapping */
-+ enet0: ethernet@e0000 {
-+ };
-+
-+ enet1: ethernet@e2000 {
-+ };
-+
-+ enet2: ethernet@e4000 {
-+ };
-+
-+ enet3: ethernet@e6000 {
-+ };
-+
-+ enet4: ethernet@e8000 {
-+ };
-+
-+ enet5: ethernet@ea000 {
-+ };
-+
-+ enet6: ethernet@f0000 {
-+ };
-+
-+ enet7: ethernet@f2000 {
-+ };
-+};
---- /dev/null
-+++ b/arch/arm64/boot/dts/freescale/fsl-ls1046a-qds-sdk.dts
-@@ -0,0 +1,112 @@
-+/*
-+ * Device Tree Include file for Freescale Layerscape-1046A family SoC.
-+ *
-+ * Copyright 2014-2015 Freescale Semiconductor, Inc.
-+ *
-+ * Mingkai Hu <Mingkai.hu@freescale.com>
-+ *
-+ * This file is dual-licensed: you can use it either under the terms
-+ * of the GPLv2 or the X11 license, at your option. Note that this dual
-+ * licensing only applies to this file, and not this project as a
-+ * whole.
-+ *
-+ * a) This library is free software; you can redistribute it and/or
-+ * modify it under the terms of the GNU General Public License as
-+ * published by the Free Software Foundation; either version 2 of the
-+ * License, or (at your option) any later version.
-+ *
-+ * This library is distributed in the hope that it will be useful,
-+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
-+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
-+ * GNU General Public License for more details.
-+ *
-+ * Or, alternatively,
-+ *
-+ * b) Permission is hereby granted, free of charge, to any person
-+ * obtaining a copy of this software and associated documentation
-+ * files (the "Software"), to deal in the Software without
-+ * restriction, including without limitation the rights to use,
-+ * copy, modify, merge, publish, distribute, sublicense, and/or
-+ * sell copies of the Software, and to permit persons to whom the
-+ * Software is furnished to do so, subject to the following
-+ * conditions:
-+ *
-+ * The above copyright notice and this permission notice shall be
-+ * included in all copies or substantial portions of the Software.
-+ *
-+ * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
-+ * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES
-+ * OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
-+ * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT
-+ * HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY,
-+ * WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
-+ * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
-+ * OTHER DEALINGS IN THE SOFTWARE.
-+ */
-+
-+#include "fsl-ls1046a-qds.dts"
-+#include "qoriq-qman-portals-sdk.dtsi"
-+#include "qoriq-bman-portals-sdk.dtsi"
-+
-+&bman_fbpr {
-+ compatible = "fsl,bman-fbpr";
-+ alloc-ranges = <0 0 0x10000 0>;
-+};
-+&qman_fqd {
-+ compatible = "fsl,qman-fqd";
-+ alloc-ranges = <0 0 0x10000 0>;
-+};
-+&qman_pfdr {
-+ compatible = "fsl,qman-pfdr";
-+ alloc-ranges = <0 0 0x10000 0>;
-+};
-+
-+&soc {
-+#include "qoriq-dpaa-eth.dtsi"
-+#include "qoriq-fman3-0-6oh.dtsi"
-+};
-+
-+&fsldpaa {
-+ ethernet@9 {
-+ compatible = "fsl,dpa-ethernet";
-+ fsl,fman-mac = <&enet7>;
-+ dma-coherent;
-+ };
-+};
-+
-+&fman0 {
-+ compatible = "fsl,fman", "simple-bus";
-+};
-+
-+&dspi {
-+ bus-num = <0>;
-+ status = "okay";
-+
-+ flash@0 {
-+ #address-cells = <1>;
-+ #size-cells = <1>;
-+ compatible = "n25q128a11", "jedec,spi-nor";
-+ reg = <0>;
-+ spi-max-frequency = <10000000>;
-+ };
-+
-+ flash@1 {
-+ #address-cells = <1>;
-+ #size-cells = <1>;
-+ compatible = "sst25wf040b", "jedec,spi-nor";
-+ spi-cpol;
-+ spi-cpha;
-+ reg = <1>;
-+ spi-max-frequency = <10000000>;
-+ };
-+
-+ flash@2 {
-+ #address-cells = <1>;
-+ #size-cells = <1>;
-+ compatible = "en25s64", "jedec,spi-nor";
-+ spi-cpol;
-+ spi-cpha;
-+ reg = <2>;
-+ spi-max-frequency = <10000000>;
-+ };
-+};
---- /dev/null
-+++ b/arch/arm64/boot/dts/freescale/fsl-ls1046a-qds.dts
-@@ -0,0 +1,326 @@
-+// SPDX-License-Identifier: (GPL-2.0+ OR MIT)
-+/*
-+ * Device Tree Include file for Freescale Layerscape-1046A family SoC.
-+ *
-+ * Copyright 2016 Freescale Semiconductor, Inc.
-+ *
-+ * Shaohui Xie <Shaohui.Xie@nxp.com>
-+ */
-+
-+/dts-v1/;
-+
-+#include "fsl-ls1046a.dtsi"
-+
-+/ {
-+ model = "LS1046A QDS Board";
-+ compatible = "fsl,ls1046a-qds", "fsl,ls1046a";
-+
-+ aliases {
-+ gpio0 = &gpio0;
-+ gpio1 = &gpio1;
-+ gpio2 = &gpio2;
-+ gpio3 = &gpio3;
-+ serial0 = &duart0;
-+ serial1 = &duart1;
-+ serial2 = &duart2;
-+ serial3 = &duart3;
-+
-+ emi1_slot1 = &ls1046mdio_s1;
-+ emi1_slot2 = &ls1046mdio_s2;
-+ emi1_slot4 = &ls1046mdio_s4;
-+
-+ sgmii_s1_p1 = &sgmii_phy_s1_p1;
-+ sgmii_s1_p2 = &sgmii_phy_s1_p2;
-+ sgmii_s1_p3 = &sgmii_phy_s1_p3;
-+ sgmii_s1_p4 = &sgmii_phy_s1_p4;
-+ sgmii_s4_p1 = &sgmii_phy_s4_p1;
-+ qsgmii_s2_p1 = &qsgmii_phy_s2_p1;
-+ qsgmii_s2_p2 = &qsgmii_phy_s2_p2;
-+ qsgmii_s2_p3 = &qsgmii_phy_s2_p3;
-+ qsgmii_s2_p4 = &qsgmii_phy_s2_p4;
-+ };
-+
-+ chosen {
-+ stdout-path = "serial0:115200n8";
-+ };
-+};
-+
-+&dspi {
-+ bus-num = <0>;
-+ status = "okay";
-+
-+ flash@0 {
-+ #address-cells = <1>;
-+ #size-cells = <1>;
-+ compatible = "n25q128a11", "jedec,spi-nor";
-+ reg = <0>;
-+ spi-max-frequency = <10000000>;
-+ };
-+
-+ flash@1 {
-+ #address-cells = <1>;
-+ #size-cells = <1>;
-+ compatible = "sst25wf040b", "jedec,spi-nor";
-+ spi-cpol;
-+ spi-cpha;
-+ reg = <1>;
-+ spi-max-frequency = <10000000>;
-+ };
-+
-+ flash@2 {
-+ #address-cells = <1>;
-+ #size-cells = <1>;
-+ compatible = "en25s64", "jedec,spi-nor";
-+ spi-cpol;
-+ spi-cpha;
-+ reg = <2>;
-+ spi-max-frequency = <10000000>;
-+ };
-+};
-+
-+&duart0 {
-+ status = "okay";
-+};
-+
-+&duart1 {
-+ status = "okay";
-+};
-+
-+&i2c0 {
-+ status = "okay";
-+
-+ pca9547@77 {
-+ compatible = "nxp,pca9547";
-+ reg = <0x77>;
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+
-+ i2c@2 {
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ reg = <0x2>;
-+
-+ ina220@40 {
-+ compatible = "ti,ina220";
-+ reg = <0x40>;
-+ shunt-resistor = <1000>;
-+ };
-+
-+ ina220@41 {
-+ compatible = "ti,ina220";
-+ reg = <0x41>;
-+ shunt-resistor = <1000>;
-+ };
-+ };
-+
-+ i2c@3 {
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ reg = <0x3>;
-+
-+ rtc@51 {
-+ compatible = "nxp,pcf2129";
-+ reg = <0x51>;
-+ /* IRQ10_B */
-+ interrupts = <0 150 0x4>;
-+ };
-+
-+ eeprom@56 {
-+ compatible = "atmel,24c512";
-+ reg = <0x56>;
-+ };
-+
-+ eeprom@57 {
-+ compatible = "atmel,24c512";
-+ reg = <0x57>;
-+ };
-+
-+ temp-sensor@4c {
-+ compatible = "adi,adt7461a";
-+ reg = <0x4c>;
-+ };
-+ };
-+ };
-+};
-+
-+&ifc {
-+ #address-cells = <2>;
-+ #size-cells = <1>;
-+ /* NOR, NAND Flashes and FPGA on board */
-+ ranges = <0x0 0x0 0x0 0x60000000 0x08000000
-+ 0x1 0x0 0x0 0x7e800000 0x00010000
-+ 0x2 0x0 0x0 0x7fb00000 0x00000100>;
-+ status = "okay";
-+
-+ nor@0,0 {
-+ compatible = "cfi-flash";
-+ reg = <0x0 0x0 0x8000000>;
-+ bank-width = <2>;
-+ device-width = <1>;
-+ };
-+
-+ nand@1,0 {
-+ compatible = "fsl,ifc-nand";
-+ reg = <0x1 0x0 0x10000>;
-+ };
-+
-+ fpga: board-control@2,0 {
-+ compatible = "fsl,ls1046aqds-fpga", "fsl,fpga-qixis", "simple-bus";
-+ reg = <0x2 0x0 0x0000100>;
-+ ranges = <0 2 0 0x100>;
-+ };
-+};
-+
-+&lpuart0 {
-+ status = "okay";
-+};
-+
-+&qspi {
-+ num-cs = <2>;
-+ bus-num = <0>;
-+ status = "okay";
-+
-+ qflash0: s25fl128s@0 {
-+ compatible = "spansion,m25p80";
-+ #address-cells = <1>;
-+ #size-cells = <1>;
-+ spi-max-frequency = <20000000>;
-+ reg = <0>;
-+ };
-+};
-+
-+#include "fsl-ls1046-post.dtsi"
-+
-+&fman0 {
-+ ethernet@e0000 {
-+ phy-handle = <&qsgmii_phy_s2_p1>;
-+ phy-connection-type = "sgmii";
-+ };
-+
-+ ethernet@e2000 {
-+ phy-handle = <&sgmii_phy_s4_p1>;
-+ phy-connection-type = "sgmii";
-+ };
-+
-+ ethernet@e4000 {
-+ phy-handle = <&rgmii_phy1>;
-+ phy-connection-type = "rgmii";
-+ };
-+
-+ ethernet@e6000 {
-+ phy-handle = <&rgmii_phy2>;
-+ phy-connection-type = "rgmii";
-+ };
-+
-+ ethernet@e8000 {
-+ phy-handle = <&sgmii_phy_s1_p3>;
-+ phy-connection-type = "sgmii";
-+ };
-+
-+ ethernet@ea000 {
-+ phy-handle = <&sgmii_phy_s1_p4>;
-+ phy-connection-type = "sgmii";
-+ };
-+
-+ ethernet@f0000 { /* DTSEC9/10GEC1 */
-+ phy-handle = <&sgmii_phy_s1_p1>;
-+ phy-connection-type = "xgmii";
-+ };
-+
-+ ethernet@f2000 { /* DTSEC10/10GEC2 */
-+ phy-handle = <&sgmii_phy_s1_p2>;
-+ phy-connection-type = "xgmii";
-+ };
-+};
-+
-+&fpga {
-+ #address-cells = <1>;
-+ #size-cells = <1>;
-+ mdio-mux-emi1 {
-+ compatible = "mdio-mux-mmioreg", "mdio-mux";
-+ mdio-parent-bus = <&mdio0>;
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ reg = <0x54 1>; /* BRDCFG4 */
-+ mux-mask = <0xe0>; /* EMI1 */
-+
-+ /* On-board RGMII1 PHY */
-+ ls1046mdio0: mdio@0 {
-+ reg = <0>;
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+
-+ rgmii_phy1: ethernet-phy@1 { /* MAC3 */
-+ reg = <0x1>;
-+ };
-+ };
-+
-+ /* On-board RGMII2 PHY */
-+ ls1046mdio1: mdio@1 {
-+ reg = <0x20>;
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+
-+ rgmii_phy2: ethernet-phy@2 { /* MAC4 */
-+ reg = <0x2>;
-+ };
-+ };
-+
-+ /* Slot 1 */
-+ ls1046mdio_s1: mdio@2 {
-+ reg = <0x40>;
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ status = "disabled";
-+
-+ sgmii_phy_s1_p1: ethernet-phy@1c {
-+ reg = <0x1c>;
-+ };
-+
-+ sgmii_phy_s1_p2: ethernet-phy@1d {
-+ reg = <0x1d>;
-+ };
-+
-+ sgmii_phy_s1_p3: ethernet-phy@1e {
-+ reg = <0x1e>;
-+ };
-+
-+ sgmii_phy_s1_p4: ethernet-phy@1f {
-+ reg = <0x1f>;
-+ };
-+ };
-+
-+ /* Slot 2 */
-+ ls1046mdio_s2: mdio@3 {
-+ reg = <0x60>;
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ status = "disabled";
-+
-+ qsgmii_phy_s2_p1: ethernet-phy@8 {
-+ reg = <0x8>;
-+ };
-+ qsgmii_phy_s2_p2: ethernet-phy@9 {
-+ reg = <0x9>;
-+ };
-+ qsgmii_phy_s2_p3: ethernet-phy@a {
-+ reg = <0xa>;
-+ };
-+ qsgmii_phy_s2_p4: ethernet-phy@b {
-+ reg = <0xb>;
-+ };
-+ };
-+
-+ /* Slot 4 */
-+ ls1046mdio_s4: mdio@5 {
-+ reg = <0x80>;
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ status = "disabled";
-+
-+ sgmii_phy_s4_p1: ethernet-phy@1c {
-+ reg = <0x1c>;
-+ };
-+ };
-+ };
-+};
---- /dev/null
-+++ b/arch/arm64/boot/dts/freescale/fsl-ls1046a-rdb-sdk.dts
-@@ -0,0 +1,85 @@
-+/*
-+ * Device Tree Include file for Freescale Layerscape-1046A family SoC.
-+ *
-+ * Copyright 2014-2015 Freescale Semiconductor, Inc.
-+ *
-+ * Mingkai Hu <Mingkai.hu@freescale.com>
-+ *
-+ * This file is dual-licensed: you can use it either under the terms
-+ * of the GPLv2 or the X11 license, at your option. Note that this dual
-+ * licensing only applies to this file, and not this project as a
-+ * whole.
-+ *
-+ * a) This library is free software; you can redistribute it and/or
-+ * modify it under the terms of the GNU General Public License as
-+ * published by the Free Software Foundation; either version 2 of the
-+ * License, or (at your option) any later version.
-+ *
-+ * This library is distributed in the hope that it will be useful,
-+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
-+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
-+ * GNU General Public License for more details.
-+ *
-+ * Or, alternatively,
-+ *
-+ * b) Permission is hereby granted, free of charge, to any person
-+ * obtaining a copy of this software and associated documentation
-+ * files (the "Software"), to deal in the Software without
-+ * restriction, including without limitation the rights to use,
-+ * copy, modify, merge, publish, distribute, sublicense, and/or
-+ * sell copies of the Software, and to permit persons to whom the
-+ * Software is furnished to do so, subject to the following
-+ * conditions:
-+ *
-+ * The above copyright notice and this permission notice shall be
-+ * included in all copies or substantial portions of the Software.
-+ *
-+ * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
-+ * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES
-+ * OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
-+ * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT
-+ * HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY,
-+ * WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
-+ * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
-+ * OTHER DEALINGS IN THE SOFTWARE.
-+ */
-+
-+#include "fsl-ls1046a-rdb.dts"
-+#include "qoriq-qman-portals-sdk.dtsi"
-+#include "qoriq-bman-portals-sdk.dtsi"
-+
-+&bman_fbpr {
-+ compatible = "fsl,bman-fbpr";
-+ alloc-ranges = <0 0 0x10000 0>;
-+};
-+&qman_fqd {
-+ compatible = "fsl,qman-fqd";
-+ alloc-ranges = <0 0 0x10000 0>;
-+};
-+&qman_pfdr {
-+ compatible = "fsl,qman-pfdr";
-+ alloc-ranges = <0 0 0x10000 0>;
-+};
-+
-+&soc {
-+#include "qoriq-dpaa-eth.dtsi"
-+#include "qoriq-fman3-0-6oh.dtsi"
-+};
-+
-+&fsldpaa {
-+ ethernet@0 {
-+ status = "disabled";
-+ };
-+ ethernet@1 {
-+ status = "disabled";
-+ };
-+ ethernet@9 {
-+ compatible = "fsl,dpa-ethernet";
-+ fsl,fman-mac = <&enet7>;
-+ dma-coherent;
-+ };
-+};
-+
-+&fman0 {
-+ compatible = "fsl,fman", "simple-bus";
-+};
---- /dev/null
-+++ b/arch/arm64/boot/dts/freescale/fsl-ls1046a-rdb-usdpaa.dts
-@@ -0,0 +1,110 @@
-+/*
-+ * Device Tree Include file for Freescale Layerscape-1046A family SoC.
-+ *
-+ * Copyright 2016 Freescale Semiconductor, Inc.
-+ *
-+ * This file is licensed under the terms of the GNU General Public
-+ * License version 2. This program is licensed "as is" without any
-+ * warranty of any kind, whether express or implied.
-+ */
-+
-+#include "fsl-ls1046a-rdb-sdk.dts"
-+
-+&soc {
-+ bp7: buffer-pool@7 {
-+ compatible = "fsl,ls1046a-bpool", "fsl,bpool";
-+ fsl,bpid = <7>;
-+ fsl,bpool-ethernet-cfg = <0 0 0 192 0 0xdeadbeef>;
-+ fsl,bpool-thresholds = <0x400 0xc00 0x0 0x0>;
-+ };
-+
-+ bp8: buffer-pool@8 {
-+ compatible = "fsl,ls1046a-bpool", "fsl,bpool";
-+ fsl,bpid = <8>;
-+ fsl,bpool-ethernet-cfg = <0 0 0 576 0 0xabbaf00d>;
-+ fsl,bpool-thresholds = <0x100 0x300 0x0 0x0>;
-+ };
-+
-+ bp9: buffer-pool@9 {
-+ compatible = "fsl,ls1046a-bpool", "fsl,bpool";
-+ fsl,bpid = <9>;
-+ fsl,bpool-ethernet-cfg = <0 0 0 2048 0 0xfeedabba>;
-+ fsl,bpool-thresholds = <0x100 0x300 0x0 0x0>;
-+ };
-+
-+ fsl,dpaa {
-+ compatible = "fsl,ls1046a", "fsl,dpaa", "simple-bus";
-+
-+ ethernet@2 {
-+ compatible = "fsl,dpa-ethernet-init";
-+ fsl,bman-buffer-pools = <&bp7 &bp8 &bp9>;
-+ fsl,qman-frame-queues-rx = <0x54 1 0x55 1>;
-+ fsl,qman-frame-queues-tx = <0x74 1 0x75 1>;
-+ };
-+
-+ ethernet@3 {
-+ compatible = "fsl,dpa-ethernet-init";
-+ fsl,bman-buffer-pools = <&bp7 &bp8 &bp9>;
-+ fsl,qman-frame-queues-rx = <0x56 1 0x57 1>;
-+ fsl,qman-frame-queues-tx = <0x76 1 0x77 1>;
-+ };
-+
-+ ethernet@4 {
-+ compatible = "fsl,dpa-ethernet-init";
-+ fsl,bman-buffer-pools = <&bp7 &bp8 &bp9>;
-+ fsl,qman-frame-queues-rx = <0x58 1 0x59 1>;
-+ fsl,qman-frame-queues-tx = <0x78 1 0x79 1>;
-+ };
-+
-+ ethernet@5 {
-+ compatible = "fsl,dpa-ethernet-init";
-+ fsl,bman-buffer-pools = <&bp7 &bp8 &bp9>;
-+ fsl,qman-frame-queues-rx = <0x5a 1 0x5b 1>;
-+ fsl,qman-frame-queues-tx = <0x7a 1 0x7b 1>;
-+ };
-+
-+ ethernet@8 {
-+ compatible = "fsl,dpa-ethernet-init";
-+ fsl,bman-buffer-pools = <&bp7 &bp8 &bp9>;
-+ fsl,qman-frame-queues-rx = <0x5c 1 0x5d 1>;
-+ fsl,qman-frame-queues-tx = <0x7c 1 0x7d 1>;
-+ };
-+
-+ ethernet@9 {
-+ compatible = "fsl,dpa-ethernet-init";
-+ fsl,bman-buffer-pools = <&bp7 &bp8 &bp9>;
-+ fsl,qman-frame-queues-rx = <0x5e 1 0x5f 1>;
-+ fsl,qman-frame-queues-tx = <0x7e 1 0x7f 1>;
-+ };
-+
-+ dpa-fman0-oh@2 {
-+ compatible = "fsl,dpa-oh";
-+ /* Define frame queues for the OH port*/
-+ /* <OH Rx error, OH Rx default> */
-+ fsl,qman-frame-queues-oh = <0x60 1 0x61 1>;
-+ fsl,fman-oh-port = <&fman0_oh2>;
-+ };
-+ };
-+};
-+/ {
-+ reserved-memory {
-+ #address-cells = <2>;
-+ #size-cells = <2>;
-+ ranges;
-+
-+ usdpaa_mem: usdpaa_mem {
-+ compatible = "fsl,usdpaa-mem";
-+ alloc-ranges = <0 0 0x10000 0>;
-+ size = <0 0x10000000>;
-+ alignment = <0 0x10000000>;
-+ };
-+ };
-+};
-+
-+&fman0 {
-+ fman0_oh2: port@83000 {
-+ cell-index = <1>;
-+ compatible = "fsl,fman-port-oh";
-+ reg = <0x83000 0x1000>;
-+ };
-+};
---- /dev/null
-+++ b/arch/arm64/boot/dts/freescale/fsl-ls1046a-rdb.dts
-@@ -0,0 +1,181 @@
-+// SPDX-License-Identifier: (GPL-2.0+ OR MIT)
-+/*
-+ * Device Tree Include file for Freescale Layerscape-1046A family SoC.
-+ *
-+ * Copyright 2016 Freescale Semiconductor, Inc.
-+ *
-+ * Mingkai Hu <mingkai.hu@nxp.com>
-+ */
-+
-+/dts-v1/;
-+
-+#include "fsl-ls1046a.dtsi"
-+
-+/ {
-+ model = "LS1046A RDB Board";
-+ compatible = "fsl,ls1046a-rdb", "fsl,ls1046a";
-+
-+ aliases {
-+ serial0 = &duart0;
-+ serial1 = &duart1;
-+ serial2 = &duart2;
-+ serial3 = &duart3;
-+ };
-+
-+ chosen {
-+ stdout-path = "serial0:115200n8";
-+ };
-+};
-+
-+&esdhc {
-+ mmc-hs200-1_8v;
-+ sd-uhs-sdr104;
-+ sd-uhs-sdr50;
-+ sd-uhs-sdr25;
-+ sd-uhs-sdr12;
-+};
-+
-+&duart0 {
-+ status = "okay";
-+};
-+
-+&duart1 {
-+ status = "okay";
-+};
-+
-+&i2c0 {
-+ status = "okay";
-+
-+ ina220@40 {
-+ compatible = "ti,ina220";
-+ reg = <0x40>;
-+ shunt-resistor = <1000>;
-+ };
-+
-+ temp-sensor@4c {
-+ compatible = "adi,adt7461";
-+ reg = <0x4c>;
-+ };
-+
-+ eeprom@56 {
-+ compatible = "atmel,24c512";
-+ reg = <0x52>;
-+ };
-+
-+ eeprom@57 {
-+ compatible = "atmel,24c512";
-+ reg = <0x53>;
-+ };
-+};
-+
-+&i2c3 {
-+ status = "okay";
-+
-+ rtc@51 {
-+ compatible = "nxp,pcf2129";
-+ reg = <0x51>;
-+ };
-+};
-+
-+&ifc {
-+ #address-cells = <2>;
-+ #size-cells = <1>;
-+ /* NAND Flashe and CPLD on board */
-+ ranges = <0x0 0x0 0x0 0x7e800000 0x00010000
-+ 0x2 0x0 0x0 0x7fb00000 0x00000100>;
-+ status = "okay";
-+
-+ nand@0,0 {
-+ compatible = "fsl,ifc-nand";
-+ #address-cells = <1>;
-+ #size-cells = <1>;
-+ reg = <0x0 0x0 0x10000>;
-+ };
-+
-+ cpld: board-control@2,0 {
-+ compatible = "fsl,ls1046ardb-cpld";
-+ reg = <0x2 0x0 0x0000100>;
-+ };
-+};
-+
-+&qspi {
-+ num-cs = <2>;
-+ bus-num = <0>;
-+ status = "okay";
-+
-+ qflash0: s25fs512s@0 {
-+ compatible = "spansion,m25p80";
-+ #address-cells = <1>;
-+ #size-cells = <1>;
-+ spi-max-frequency = <20000000>;
-+ reg = <0>;
-+ };
-+
-+ qflash1: s25fs512s@1 {
-+ compatible = "spansion,m25p80";
-+ #address-cells = <1>;
-+ #size-cells = <1>;
-+ spi-max-frequency = <20000000>;
-+ reg = <1>;
-+ };
-+};
-+
-+#include "fsl-ls1046-post.dtsi"
-+
-+&fman0 {
-+ ethernet@e4000 {
-+ phy-handle = <&rgmii_phy1>;
-+ phy-connection-type = "rgmii";
-+ };
-+
-+ ethernet@e6000 {
-+ phy-handle = <&rgmii_phy2>;
-+ phy-connection-type = "rgmii";
-+ };
-+
-+ ethernet@e8000 {
-+ phy-handle = <&sgmii_phy1>;
-+ phy-connection-type = "sgmii";
-+ };
-+
-+ ethernet@ea000 {
-+ phy-handle = <&sgmii_phy2>;
-+ phy-connection-type = "sgmii";
-+ };
-+
-+ ethernet@f0000 { /* 10GEC1 */
-+ phy-handle = <&aqr106_phy>;
-+ phy-connection-type = "xgmii";
-+ };
-+
-+ ethernet@f2000 { /* 10GEC2 */
-+ fixed-link = <0 1 1000 0 0>;
-+ phy-connection-type = "xgmii";
-+ };
-+
-+ mdio@fc000 {
-+ rgmii_phy1: ethernet-phy@1 {
-+ reg = <0x1>;
-+ };
-+
-+ rgmii_phy2: ethernet-phy@2 {
-+ reg = <0x2>;
-+ };
-+
-+ sgmii_phy1: ethernet-phy@3 {
-+ reg = <0x3>;
-+ };
-+
-+ sgmii_phy2: ethernet-phy@4 {
-+ reg = <0x4>;
-+ };
-+ };
-+
-+ mdio@fd000 {
-+ aqr106_phy: ethernet-phy@0 {
-+ compatible = "ethernet-phy-ieee802.3-c45";
-+ interrupts = <0 131 4>;
-+ reg = <0x0>;
-+ };
-+ };
-+};
---- /dev/null
-+++ b/arch/arm64/boot/dts/freescale/fsl-ls1046a.dtsi
-@@ -0,0 +1,771 @@
-+// SPDX-License-Identifier: (GPL-2.0+ OR MIT)
-+/*
-+ * Device Tree Include file for Freescale Layerscape-1046A family SoC.
-+ *
-+ * Copyright 2016 Freescale Semiconductor, Inc.
-+ *
-+ * Mingkai Hu <mingkai.hu@nxp.com>
-+ */
-+
-+#include <dt-bindings/interrupt-controller/arm-gic.h>
-+#include <dt-bindings/thermal/thermal.h>
-+
-+/ {
-+ compatible = "fsl,ls1046a";
-+ interrupt-parent = <&gic>;
-+ #address-cells = <2>;
-+ #size-cells = <2>;
-+
-+ aliases {
-+ crypto = &crypto;
-+ fman0 = &fman0;
-+ ethernet0 = &enet0;
-+ ethernet1 = &enet1;
-+ ethernet2 = &enet2;
-+ ethernet3 = &enet3;
-+ ethernet4 = &enet4;
-+ ethernet5 = &enet5;
-+ ethernet6 = &enet6;
-+ ethernet7 = &enet7;
-+ };
-+
-+ cpus {
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+
-+ cpu0: cpu@0 {
-+ device_type = "cpu";
-+ compatible = "arm,cortex-a72";
-+ reg = <0x0>;
-+ clocks = <&clockgen 1 0>;
-+ next-level-cache = <&l2>;
-+ cpu-idle-states = <&CPU_PH20>;
-+ #cooling-cells = <2>;
-+ };
-+
-+ cpu1: cpu@1 {
-+ device_type = "cpu";
-+ compatible = "arm,cortex-a72";
-+ reg = <0x1>;
-+ clocks = <&clockgen 1 0>;
-+ next-level-cache = <&l2>;
-+ cpu-idle-states = <&CPU_PH20>;
-+ };
-+
-+ cpu2: cpu@2 {
-+ device_type = "cpu";
-+ compatible = "arm,cortex-a72";
-+ reg = <0x2>;
-+ clocks = <&clockgen 1 0>;
-+ next-level-cache = <&l2>;
-+ cpu-idle-states = <&CPU_PH20>;
-+ };
-+
-+ cpu3: cpu@3 {
-+ device_type = "cpu";
-+ compatible = "arm,cortex-a72";
-+ reg = <0x3>;
-+ clocks = <&clockgen 1 0>;
-+ next-level-cache = <&l2>;
-+ cpu-idle-states = <&CPU_PH20>;
-+ };
-+
-+ l2: l2-cache {
-+ compatible = "cache";
-+ };
-+ };
-+
-+ idle-states {
-+ /*
-+ * PSCI node is not added default, U-boot will add missing
-+ * parts if it determines to use PSCI.
-+ */
-+ entry-method = "arm,psci";
-+
-+ CPU_PH20: cpu-ph20 {
-+ compatible = "arm,idle-state";
-+ idle-state-name = "PH20";
-+ arm,psci-suspend-param = <0x0>;
-+ entry-latency-us = <1000>;
-+ exit-latency-us = <1000>;
-+ min-residency-us = <3000>;
-+ };
-+ };
-+
-+ memory@80000000 {
-+ device_type = "memory";
-+ };
-+
-+ sysclk: sysclk {
-+ compatible = "fixed-clock";
-+ #clock-cells = <0>;
-+ clock-frequency = <100000000>;
-+ clock-output-names = "sysclk";
-+ };
-+
-+ reboot {
-+ compatible ="syscon-reboot";
-+ regmap = <&dcfg>;
-+ offset = <0xb0>;
-+ mask = <0x02>;
-+ };
-+
-+ timer {
-+ compatible = "arm,armv8-timer";
-+ interrupts = <GIC_PPI 13 (GIC_CPU_MASK_RAW(0xf) |
-+ IRQ_TYPE_LEVEL_LOW)>,
-+ <GIC_PPI 14 (GIC_CPU_MASK_RAW(0xf) |
-+ IRQ_TYPE_LEVEL_LOW)>,
-+ <GIC_PPI 11 (GIC_CPU_MASK_RAW(0xf) |
-+ IRQ_TYPE_LEVEL_LOW)>,
-+ <GIC_PPI 10 (GIC_CPU_MASK_RAW(0xf) |
-+ IRQ_TYPE_LEVEL_LOW)>;
-+ };
-+
-+ pmu {
-+ compatible = "arm,cortex-a72-pmu";
-+ interrupts = <GIC_SPI 106 IRQ_TYPE_LEVEL_HIGH>,
-+ <GIC_SPI 107 IRQ_TYPE_LEVEL_HIGH>,
-+ <GIC_SPI 95 IRQ_TYPE_LEVEL_HIGH>,
-+ <GIC_SPI 97 IRQ_TYPE_LEVEL_HIGH>;
-+ interrupt-affinity = <&cpu0>,
-+ <&cpu1>,
-+ <&cpu2>,
-+ <&cpu3>;
-+ };
-+
-+ gic: interrupt-controller@1400000 {
-+ compatible = "arm,gic-400";
-+ #interrupt-cells = <3>;
-+ interrupt-controller;
-+ reg = <0x0 0x1410000 0 0x10000>, /* GICD */
-+ <0x0 0x1420000 0 0x20000>, /* GICC */
-+ <0x0 0x1440000 0 0x20000>, /* GICH */
-+ <0x0 0x1460000 0 0x20000>; /* GICV */
-+ interrupts = <GIC_PPI 9 (GIC_CPU_MASK_RAW(0xf) |
-+ IRQ_TYPE_LEVEL_LOW)>;
-+ };
-+
-+ soc: soc {
-+ compatible = "simple-bus";
-+ #address-cells = <2>;
-+ #size-cells = <2>;
-+ ranges;
-+
-+ ddr: memory-controller@1080000 {
-+ compatible = "fsl,qoriq-memory-controller";
-+ reg = <0x0 0x1080000 0x0 0x1000>;
-+ interrupts = <GIC_SPI 144 IRQ_TYPE_LEVEL_HIGH>;
-+ big-endian;
-+ };
-+
-+ ifc: ifc@1530000 {
-+ compatible = "fsl,ifc", "simple-bus";
-+ reg = <0x0 0x1530000 0x0 0x10000>;
-+ big-endian;
-+ interrupts = <GIC_SPI 43 IRQ_TYPE_LEVEL_HIGH>;
-+ };
-+
-+ qspi: quadspi@1550000 {
-+ compatible = "fsl,ls1021a-qspi";
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ reg = <0x0 0x1550000 0x0 0x10000>,
-+ <0x0 0x40000000 0x0 0x10000000>;
-+ reg-names = "QuadSPI", "QuadSPI-memory";
-+ interrupts = <GIC_SPI 99 IRQ_TYPE_LEVEL_HIGH>;
-+ clock-names = "qspi_en", "qspi";
-+ clocks = <&clockgen 4 1>, <&clockgen 4 1>;
-+ big-endian;
-+ fsl,qspi-has-second-chip;
-+ status = "disabled";
-+ };
-+
-+ esdhc: esdhc@1560000 {
-+ compatible = "fsl,ls1046a-esdhc", "fsl,esdhc";
-+ reg = <0x0 0x1560000 0x0 0x10000>;
-+ interrupts = <GIC_SPI 62 IRQ_TYPE_LEVEL_HIGH>;
-+ clocks = <&clockgen 2 1>;
-+ voltage-ranges = <1800 1800 3300 3300>;
-+ sdhci,auto-cmd12;
-+ big-endian;
-+ bus-width = <4>;
-+ };
-+
-+ scfg: scfg@1570000 {
-+ compatible = "fsl,ls1046a-scfg", "syscon";
-+ reg = <0x0 0x1570000 0x0 0x10000>;
-+ big-endian;
-+ };
-+
-+ crypto: crypto@1700000 {
-+ compatible = "fsl,sec-v5.4", "fsl,sec-v5.0",
-+ "fsl,sec-v4.0";
-+ fsl,sec-era = <8>;
-+ #address-cells = <1>;
-+ #size-cells = <1>;
-+ ranges = <0x0 0x00 0x1700000 0x100000>;
-+ reg = <0x00 0x1700000 0x0 0x100000>;
-+ interrupts = <GIC_SPI 75 IRQ_TYPE_LEVEL_HIGH>;
-+
-+ sec_jr0: jr@10000 {
-+ compatible = "fsl,sec-v5.4-job-ring",
-+ "fsl,sec-v5.0-job-ring",
-+ "fsl,sec-v4.0-job-ring";
-+ reg = <0x10000 0x10000>;
-+ interrupts = <GIC_SPI 71 IRQ_TYPE_LEVEL_HIGH>;
-+ };
-+
-+ sec_jr1: jr@20000 {
-+ compatible = "fsl,sec-v5.4-job-ring",
-+ "fsl,sec-v5.0-job-ring",
-+ "fsl,sec-v4.0-job-ring";
-+ reg = <0x20000 0x10000>;
-+ interrupts = <GIC_SPI 72 IRQ_TYPE_LEVEL_HIGH>;
-+ };
-+
-+ sec_jr2: jr@30000 {
-+ compatible = "fsl,sec-v5.4-job-ring",
-+ "fsl,sec-v5.0-job-ring",
-+ "fsl,sec-v4.0-job-ring";
-+ reg = <0x30000 0x10000>;
-+ interrupts = <GIC_SPI 73 IRQ_TYPE_LEVEL_HIGH>;
-+ };
-+
-+ sec_jr3: jr@40000 {
-+ compatible = "fsl,sec-v5.4-job-ring",
-+ "fsl,sec-v5.0-job-ring",
-+ "fsl,sec-v4.0-job-ring";
-+ reg = <0x40000 0x10000>;
-+ interrupts = <GIC_SPI 74 IRQ_TYPE_LEVEL_HIGH>;
-+ };
-+ };
-+
-+ qman: qman@1880000 {
-+ compatible = "fsl,qman";
-+ reg = <0x00 0x1880000 0x0 0x10000>;
-+ interrupts = <0 45 0x4>;
-+ memory-region = <&qman_fqd &qman_pfdr>;
-+
-+ };
-+
-+ bman: bman@1890000 {
-+ compatible = "fsl,bman";
-+ reg = <0x00 0x1890000 0x0 0x10000>;
-+ interrupts = <0 45 0x4>;
-+ memory-region = <&bman_fbpr>;
-+
-+ };
-+
-+ qportals: qman-portals@500000000 {
-+ ranges = <0x0 0x5 0x00000000 0x8000000>;
-+ };
-+
-+ bportals: bman-portals@508000000 {
-+ ranges = <0x0 0x5 0x08000000 0x8000000>;
-+ };
-+
-+ dcfg: dcfg@1ee0000 {
-+ compatible = "fsl,ls1046a-dcfg", "syscon";
-+ reg = <0x0 0x1ee0000 0x0 0x1000>;
-+ big-endian;
-+ };
-+
-+ clockgen: clocking@1ee1000 {
-+ compatible = "fsl,ls1046a-clockgen";
-+ reg = <0x0 0x1ee1000 0x0 0x1000>;
-+ #clock-cells = <2>;
-+ clocks = <&sysclk>;
-+ };
-+
-+ tmu: tmu@1f00000 {
-+ compatible = "fsl,qoriq-tmu";
-+ reg = <0x0 0x1f00000 0x0 0x10000>;
-+ interrupts = <0 33 0x4>;
-+ fsl,tmu-range = <0xb0000 0x9002a 0x6004c 0x30062>;
-+ fsl,tmu-calibration =
-+ /* Calibration data group 1 */
-+ <0x00000000 0x00000026
-+ 0x00000001 0x0000002d
-+ 0x00000002 0x00000032
-+ 0x00000003 0x00000039
-+ 0x00000004 0x0000003f
-+ 0x00000005 0x00000046
-+ 0x00000006 0x0000004d
-+ 0x00000007 0x00000054
-+ 0x00000008 0x0000005a
-+ 0x00000009 0x00000061
-+ 0x0000000a 0x0000006a
-+ 0x0000000b 0x00000071
-+ /* Calibration data group 2 */
-+ 0x00010000 0x00000025
-+ 0x00010001 0x0000002c
-+ 0x00010002 0x00000035
-+ 0x00010003 0x0000003d
-+ 0x00010004 0x00000045
-+ 0x00010005 0x0000004e
-+ 0x00010006 0x00000057
-+ 0x00010007 0x00000061
-+ 0x00010008 0x0000006b
-+ 0x00010009 0x00000076
-+ /* Calibration data group 3 */
-+ 0x00020000 0x00000029
-+ 0x00020001 0x00000033
-+ 0x00020002 0x0000003d
-+ 0x00020003 0x00000049
-+ 0x00020004 0x00000056
-+ 0x00020005 0x00000061
-+ 0x00020006 0x0000006d
-+ /* Calibration data group 4 */
-+ 0x00030000 0x00000021
-+ 0x00030001 0x0000002a
-+ 0x00030002 0x0000003c
-+ 0x00030003 0x0000004e>;
-+ big-endian;
-+ #thermal-sensor-cells = <1>;
-+ };
-+
-+ thermal-zones {
-+ cpu_thermal: cpu-thermal {
-+ polling-delay-passive = <1000>;
-+ polling-delay = <5000>;
-+ thermal-sensors = <&tmu 3>;
-+
-+ trips {
-+ cpu_alert: cpu-alert {
-+ temperature = <85000>;
-+ hysteresis = <2000>;
-+ type = "passive";
-+ };
-+
-+ cpu_crit: cpu-crit {
-+ temperature = <95000>;
-+ hysteresis = <2000>;
-+ type = "critical";
-+ };
-+ };
-+
-+ cooling-maps {
-+ map0 {
-+ trip = <&cpu_alert>;
-+ cooling-device =
-+ <&cpu0 THERMAL_NO_LIMIT
-+ THERMAL_NO_LIMIT>;
-+ };
-+ };
-+ };
-+ };
-+
-+ dspi: dspi@2100000 {
-+ compatible = "fsl,ls1021a-v1.0-dspi";
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ reg = <0x0 0x2100000 0x0 0x10000>;
-+ interrupts = <GIC_SPI 64 IRQ_TYPE_LEVEL_HIGH>;
-+ clock-names = "dspi";
-+ clocks = <&clockgen 4 1>;
-+ spi-num-chipselects = <5>;
-+ big-endian;
-+ status = "disabled";
-+ };
-+
-+ i2c0: i2c@2180000 {
-+ compatible = "fsl,vf610-i2c", "fsl,ls1046a-vf610-i2c";
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ reg = <0x0 0x2180000 0x0 0x10000>;
-+ interrupts = <GIC_SPI 56 IRQ_TYPE_LEVEL_HIGH>;
-+ clocks = <&clockgen 4 1>;
-+ dmas = <&edma0 1 39>,
-+ <&edma0 1 38>;
-+ dma-names = "tx", "rx";
-+ fsl-scl-gpio = <&gpio3 12 0>;
-+ status = "disabled";
-+ };
-+
-+ i2c1: i2c@2190000 {
-+ compatible = "fsl,vf610-i2c";
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ reg = <0x0 0x2190000 0x0 0x10000>;
-+ interrupts = <GIC_SPI 57 IRQ_TYPE_LEVEL_HIGH>;
-+ clocks = <&clockgen 4 1>;
-+ status = "disabled";
-+ };
-+
-+ i2c2: i2c@21a0000 {
-+ compatible = "fsl,vf610-i2c";
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ reg = <0x0 0x21a0000 0x0 0x10000>;
-+ interrupts = <GIC_SPI 58 IRQ_TYPE_LEVEL_HIGH>;
-+ clocks = <&clockgen 4 1>;
-+ status = "disabled";
-+ };
-+
-+ i2c3: i2c@21b0000 {
-+ compatible = "fsl,vf610-i2c", "fsl,ls1046a-vf610-i2c";
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ reg = <0x0 0x21b0000 0x0 0x10000>;
-+ interrupts = <GIC_SPI 59 IRQ_TYPE_LEVEL_HIGH>;
-+ clocks = <&clockgen 4 1>;
-+ fsl-scl-gpio = <&gpio3 12 0>;
-+ status = "disabled";
-+ };
-+
-+ duart0: serial@21c0500 {
-+ compatible = "fsl,ns16550", "ns16550a";
-+ reg = <0x00 0x21c0500 0x0 0x100>;
-+ interrupts = <GIC_SPI 54 IRQ_TYPE_LEVEL_HIGH>;
-+ clocks = <&clockgen 4 1>;
-+ };
-+
-+ duart1: serial@21c0600 {
-+ compatible = "fsl,ns16550", "ns16550a";
-+ reg = <0x00 0x21c0600 0x0 0x100>;
-+ interrupts = <GIC_SPI 54 IRQ_TYPE_LEVEL_HIGH>;
-+ clocks = <&clockgen 4 1>;
-+ };
-+
-+ duart2: serial@21d0500 {
-+ compatible = "fsl,ns16550", "ns16550a";
-+ reg = <0x0 0x21d0500 0x0 0x100>;
-+ interrupts = <GIC_SPI 55 IRQ_TYPE_LEVEL_HIGH>;
-+ clocks = <&clockgen 4 1>;
-+ };
-+
-+ duart3: serial@21d0600 {
-+ compatible = "fsl,ns16550", "ns16550a";
-+ reg = <0x0 0x21d0600 0x0 0x100>;
-+ interrupts = <GIC_SPI 55 IRQ_TYPE_LEVEL_HIGH>;
-+ clocks = <&clockgen 4 1>;
-+ };
-+
-+ gpio0: gpio@2300000 {
-+ compatible = "fsl,qoriq-gpio";
-+ reg = <0x0 0x2300000 0x0 0x10000>;
-+ interrupts = <GIC_SPI 66 IRQ_TYPE_LEVEL_HIGH>;
-+ gpio-controller;
-+ #gpio-cells = <2>;
-+ interrupt-controller;
-+ #interrupt-cells = <2>;
-+ };
-+
-+ gpio1: gpio@2310000 {
-+ compatible = "fsl,qoriq-gpio";
-+ reg = <0x0 0x2310000 0x0 0x10000>;
-+ interrupts = <GIC_SPI 67 IRQ_TYPE_LEVEL_HIGH>;
-+ gpio-controller;
-+ #gpio-cells = <2>;
-+ interrupt-controller;
-+ #interrupt-cells = <2>;
-+ };
-+
-+ gpio2: gpio@2320000 {
-+ compatible = "fsl,qoriq-gpio";
-+ reg = <0x0 0x2320000 0x0 0x10000>;
-+ interrupts = <GIC_SPI 68 IRQ_TYPE_LEVEL_HIGH>;
-+ gpio-controller;
-+ #gpio-cells = <2>;
-+ interrupt-controller;
-+ #interrupt-cells = <2>;
-+ };
-+
-+ gpio3: gpio@2330000 {
-+ compatible = "fsl,qoriq-gpio";
-+ reg = <0x0 0x2330000 0x0 0x10000>;
-+ interrupts = <GIC_SPI 134 IRQ_TYPE_LEVEL_HIGH>;
-+ gpio-controller;
-+ #gpio-cells = <2>;
-+ interrupt-controller;
-+ #interrupt-cells = <2>;
-+ };
-+
-+ lpuart0: serial@2950000 {
-+ compatible = "fsl,ls1021a-lpuart";
-+ reg = <0x0 0x2950000 0x0 0x1000>;
-+ interrupts = <GIC_SPI 48 IRQ_TYPE_LEVEL_HIGH>;
-+ clocks = <&clockgen 4 0>;
-+ clock-names = "ipg";
-+ status = "disabled";
-+ };
-+
-+ lpuart1: serial@2960000 {
-+ compatible = "fsl,ls1021a-lpuart";
-+ reg = <0x0 0x2960000 0x0 0x1000>;
-+ interrupts = <GIC_SPI 49 IRQ_TYPE_LEVEL_HIGH>;
-+ clocks = <&clockgen 4 1>;
-+ clock-names = "ipg";
-+ status = "disabled";
-+ };
-+
-+ lpuart2: serial@2970000 {
-+ compatible = "fsl,ls1021a-lpuart";
-+ reg = <0x0 0x2970000 0x0 0x1000>;
-+ interrupts = <GIC_SPI 50 IRQ_TYPE_LEVEL_HIGH>;
-+ clocks = <&clockgen 4 1>;
-+ clock-names = "ipg";
-+ status = "disabled";
-+ };
-+
-+ lpuart3: serial@2980000 {
-+ compatible = "fsl,ls1021a-lpuart";
-+ reg = <0x0 0x2980000 0x0 0x1000>;
-+ interrupts = <GIC_SPI 51 IRQ_TYPE_LEVEL_HIGH>;
-+ clocks = <&clockgen 4 1>;
-+ clock-names = "ipg";
-+ status = "disabled";
-+ };
-+
-+ lpuart4: serial@2990000 {
-+ compatible = "fsl,ls1021a-lpuart";
-+ reg = <0x0 0x2990000 0x0 0x1000>;
-+ interrupts = <GIC_SPI 52 IRQ_TYPE_LEVEL_HIGH>;
-+ clocks = <&clockgen 4 1>;
-+ clock-names = "ipg";
-+ status = "disabled";
-+ };
-+
-+ lpuart5: serial@29a0000 {
-+ compatible = "fsl,ls1021a-lpuart";
-+ reg = <0x0 0x29a0000 0x0 0x1000>;
-+ interrupts = <GIC_SPI 53 IRQ_TYPE_LEVEL_HIGH>;
-+ clocks = <&clockgen 4 1>;
-+ clock-names = "ipg";
-+ status = "disabled";
-+ };
-+
-+ ftm0: ftm0@29d0000 {
-+ compatible = "fsl,ls1046a-ftm";
-+ reg = <0x0 0x29d0000 0x0 0x10000>,
-+ <0x0 0x1ee2140 0x0 0x4>;
-+ reg-names = "ftm", "FlexTimer1";
-+ interrupts = <GIC_SPI 86 IRQ_TYPE_LEVEL_HIGH>;
-+ big-endian;
-+ };
-+
-+ wdog0: watchdog@2ad0000 {
-+ compatible = "fsl,imx21-wdt";
-+ reg = <0x0 0x2ad0000 0x0 0x10000>;
-+ interrupts = <GIC_SPI 83 IRQ_TYPE_LEVEL_HIGH>;
-+ clocks = <&clockgen 4 1>;
-+ big-endian;
-+ };
-+
-+ edma0: edma@2c00000 {
-+ #dma-cells = <2>;
-+ compatible = "fsl,vf610-edma";
-+ reg = <0x0 0x2c00000 0x0 0x10000>,
-+ <0x0 0x2c10000 0x0 0x10000>,
-+ <0x0 0x2c20000 0x0 0x10000>;
-+ interrupts = <GIC_SPI 103 IRQ_TYPE_LEVEL_HIGH>,
-+ <GIC_SPI 103 IRQ_TYPE_LEVEL_HIGH>;
-+ interrupt-names = "edma-tx", "edma-err";
-+ dma-channels = <32>;
-+ big-endian;
-+ clock-names = "dmamux0", "dmamux1";
-+ clocks = <&clockgen 4 1>,
-+ <&clockgen 4 1>;
-+ };
-+
-+ usb0: usb@2f00000 {
-+ compatible = "snps,dwc3";
-+ reg = <0x0 0x2f00000 0x0 0x10000>;
-+ interrupts = <GIC_SPI 60 IRQ_TYPE_LEVEL_HIGH>;
-+ dr_mode = "host";
-+ snps,quirk-frame-length-adjustment = <0x20>;
-+ snps,dis_rxdet_inp3_quirk;
-+ usb3-lpm-capable;
-+ snps,dis-u1u2-when-u3-quirk;
-+ };
-+
-+ usb1: usb@3000000 {
-+ compatible = "snps,dwc3";
-+ reg = <0x0 0x3000000 0x0 0x10000>;
-+ interrupts = <GIC_SPI 61 IRQ_TYPE_LEVEL_HIGH>;
-+ dr_mode = "host";
-+ snps,quirk-frame-length-adjustment = <0x20>;
-+ snps,dis_rxdet_inp3_quirk;
-+ usb3-lpm-capable;
-+ snps,dis-u1u2-when-u3-quirk;
-+ };
-+
-+ usb2: usb@3100000 {
-+ compatible = "snps,dwc3";
-+ reg = <0x0 0x3100000 0x0 0x10000>;
-+ interrupts = <GIC_SPI 63 IRQ_TYPE_LEVEL_HIGH>;
-+ dr_mode = "host";
-+ snps,quirk-frame-length-adjustment = <0x20>;
-+ snps,dis_rxdet_inp3_quirk;
-+ usb3-lpm-capable;
-+ snps,dis-u1u2-when-u3-quirk;
-+ };
-+
-+ sata: sata@3200000 {
-+ compatible = "fsl,ls1046a-ahci";
-+ reg = <0x0 0x3200000 0x0 0x10000>,
-+ <0x0 0x20140520 0x0 0x4>;
-+ reg-names = "ahci", "sata-ecc";
-+ interrupts = <GIC_SPI 69 IRQ_TYPE_LEVEL_HIGH>;
-+ clocks = <&clockgen 4 1>;
-+ dma-coherent;
-+ };
-+
-+ qdma: qdma@8380000 {
-+ compatible = "fsl,ls1046a-qdma", "fsl,ls1021a-qdma";
-+ reg = <0x0 0x8380000 0x0 0x1000>, /* Controller regs */
-+ <0x0 0x8390000 0x0 0x10000>, /* Status regs */
-+ <0x0 0x83a0000 0x0 0x40000>; /* Block regs */
-+ interrupts = <0 153 0x4>,
-+ <0 39 0x4>;
-+ interrupt-names = "qdma-error", "qdma-queue";
-+ channels = <8>;
-+ queues = <2>;
-+ status-sizes = <64>;
-+ queue-sizes = <64 64>;
-+ big-endian;
-+ };
-+
-+ msi1: msi-controller@1580000 {
-+ compatible = "fsl,ls1046a-msi";
-+ msi-controller;
-+ reg = <0x0 0x1580000 0x0 0x10000>;
-+ interrupts = <GIC_SPI 116 IRQ_TYPE_LEVEL_HIGH>,
-+ <GIC_SPI 111 IRQ_TYPE_LEVEL_HIGH>,
-+ <GIC_SPI 112 IRQ_TYPE_LEVEL_HIGH>,
-+ <GIC_SPI 113 IRQ_TYPE_LEVEL_HIGH>;
-+ };
-+
-+ msi2: msi-controller@1590000 {
-+ compatible = "fsl,ls1046a-msi";
-+ msi-controller;
-+ reg = <0x0 0x1590000 0x0 0x10000>;
-+ interrupts = <GIC_SPI 126 IRQ_TYPE_LEVEL_HIGH>,
-+ <GIC_SPI 121 IRQ_TYPE_LEVEL_HIGH>,
-+ <GIC_SPI 122 IRQ_TYPE_LEVEL_HIGH>,
-+ <GIC_SPI 123 IRQ_TYPE_LEVEL_HIGH>;
-+ };
-+
-+ msi3: msi-controller@15a0000 {
-+ compatible = "fsl,ls1046a-msi";
-+ msi-controller;
-+ reg = <0x0 0x15a0000 0x0 0x10000>;
-+ interrupts = <GIC_SPI 160 IRQ_TYPE_LEVEL_HIGH>,
-+ <GIC_SPI 155 IRQ_TYPE_LEVEL_HIGH>,
-+ <GIC_SPI 156 IRQ_TYPE_LEVEL_HIGH>,
-+ <GIC_SPI 157 IRQ_TYPE_LEVEL_HIGH>;
-+ };
-+
-+ pcie@3400000 {
-+ compatible = "fsl,ls1046a-pcie", "snps,dw-pcie";
-+ reg = <0x00 0x03400000 0x0 0x00100000 /* controller registers */
-+ 0x40 0x00000000 0x0 0x00002000>; /* configuration space */
-+ reg-names = "regs", "config";
-+ interrupts = <GIC_SPI 117 IRQ_TYPE_LEVEL_HIGH>, /* PME interrupt */
-+ <GIC_SPI 118 IRQ_TYPE_LEVEL_HIGH>; /* aer interrupt */
-+ interrupt-names = "pme", "aer";
-+ #address-cells = <3>;
-+ #size-cells = <2>;
-+ device_type = "pci";
-+ dma-coherent;
-+ num-lanes = <4>;
-+ bus-range = <0x0 0xff>;
-+ ranges = <0x81000000 0x0 0x00000000 0x40 0x00010000 0x0 0x00010000 /* downstream I/O */
-+ 0x82000000 0x0 0x40000000 0x40 0x40000000 0x0 0x40000000>; /* non-prefetchable memory */
-+ msi-parent = <&msi1>, <&msi2>, <&msi3>;
-+ #interrupt-cells = <1>;
-+ interrupt-map-mask = <0 0 0 7>;
-+ interrupt-map = <0000 0 0 1 &gic GIC_SPI 110 IRQ_TYPE_LEVEL_HIGH>,
-+ <0000 0 0 2 &gic GIC_SPI 110 IRQ_TYPE_LEVEL_HIGH>,
-+ <0000 0 0 3 &gic GIC_SPI 110 IRQ_TYPE_LEVEL_HIGH>,
-+ <0000 0 0 4 &gic GIC_SPI 110 IRQ_TYPE_LEVEL_HIGH>;
-+ };
-+
-+ pcie@3500000 {
-+ compatible = "fsl,ls1046a-pcie", "snps,dw-pcie";
-+ reg = <0x00 0x03500000 0x0 0x00100000 /* controller registers */
-+ 0x48 0x00000000 0x0 0x00002000>; /* configuration space */
-+ reg-names = "regs", "config";
-+ interrupts = <GIC_SPI 127 IRQ_TYPE_LEVEL_HIGH>,
-+ <GIC_SPI 128 IRQ_TYPE_LEVEL_HIGH>;
-+ interrupt-names = "pme", "aer";
-+ #address-cells = <3>;
-+ #size-cells = <2>;
-+ device_type = "pci";
-+ dma-coherent;
-+ num-lanes = <2>;
-+ bus-range = <0x0 0xff>;
-+ ranges = <0x81000000 0x0 0x00000000 0x48 0x00010000 0x0 0x00010000 /* downstream I/O */
-+ 0x82000000 0x0 0x40000000 0x48 0x40000000 0x0 0x40000000>; /* non-prefetchable memory */
-+ msi-parent = <&msi1>, <&msi2>, <&msi3>;
-+ #interrupt-cells = <1>;
-+ interrupt-map-mask = <0 0 0 7>;
-+ interrupt-map = <0000 0 0 1 &gic GIC_SPI 120 IRQ_TYPE_LEVEL_HIGH>,
-+ <0000 0 0 2 &gic GIC_SPI 120 IRQ_TYPE_LEVEL_HIGH>,
-+ <0000 0 0 3 &gic GIC_SPI 120 IRQ_TYPE_LEVEL_HIGH>,
-+ <0000 0 0 4 &gic GIC_SPI 120 IRQ_TYPE_LEVEL_HIGH>;
-+ };
-+
-+ pcie@3600000 {
-+ compatible = "fsl,ls1046a-pcie", "snps,dw-pcie";
-+ reg = <0x00 0x03600000 0x0 0x00100000 /* controller registers */
-+ 0x50 0x00000000 0x0 0x00002000>; /* configuration space */
-+ reg-names = "regs", "config";
-+ interrupts = <GIC_SPI 161 IRQ_TYPE_LEVEL_HIGH>,
-+ <GIC_SPI 162 IRQ_TYPE_LEVEL_HIGH>;
-+ interrupt-names = "pme", "aer";
-+ #address-cells = <3>;
-+ #size-cells = <2>;
-+ device_type = "pci";
-+ dma-coherent;
-+ num-lanes = <2>;
-+ bus-range = <0x0 0xff>;
-+ ranges = <0x81000000 0x0 0x00000000 0x50 0x00010000 0x0 0x00010000 /* downstream I/O */
-+ 0x82000000 0x0 0x40000000 0x50 0x40000000 0x0 0x40000000>; /* non-prefetchable memory */
-+ msi-parent = <&msi1>, <&msi2>, <&msi3>;
-+ #interrupt-cells = <1>;
-+ interrupt-map-mask = <0 0 0 7>;
-+ interrupt-map = <0000 0 0 1 &gic GIC_SPI 154 IRQ_TYPE_LEVEL_HIGH>,
-+ <0000 0 0 2 &gic GIC_SPI 154 IRQ_TYPE_LEVEL_HIGH>,
-+ <0000 0 0 3 &gic GIC_SPI 154 IRQ_TYPE_LEVEL_HIGH>,
-+ <0000 0 0 4 &gic GIC_SPI 154 IRQ_TYPE_LEVEL_HIGH>;
-+ };
-+
-+ };
-+
-+ reserved-memory {
-+ #address-cells = <2>;
-+ #size-cells = <2>;
-+ ranges;
-+
-+ bman_fbpr: bman-fbpr {
-+ compatible = "shared-dma-pool";
-+ size = <0 0x1000000>;
-+ alignment = <0 0x1000000>;
-+ no-map;
-+ };
-+ qman_fqd: qman-fqd {
-+ compatible = "shared-dma-pool";
-+ size = <0 0x800000>;
-+ alignment = <0 0x800000>;
-+ no-map;
-+ };
-+ qman_pfdr: qman-pfdr {
-+ compatible = "shared-dma-pool";
-+ size = <0 0x2000000>;
-+ alignment = <0 0x2000000>;
-+ no-map;
-+ };
-+ };
-+
-+ firmware {
-+ optee {
-+ compatible = "linaro,optee-tz";
-+ method = "smc";
-+ };
-+ };
-+};
-+
-+#include "qoriq-qman-portals.dtsi"
-+#include "qoriq-bman-portals.dtsi"
---- /dev/null
-+++ b/arch/arm64/boot/dts/freescale/fsl-ls1088a-qds.dts
-@@ -0,0 +1,137 @@
-+// SPDX-License-Identifier: (GPL-2.0+ OR MIT)
-+/*
-+ * Device Tree file for NXP LS1088A QDS Board.
-+ *
-+ * Copyright 2017 NXP
-+ *
-+ * Harninder Rai <harninder.rai@nxp.com>
-+ *
-+ */
-+
-+/dts-v1/;
-+
-+#include "fsl-ls1088a.dtsi"
-+
-+/ {
-+ model = "LS1088A QDS Board";
-+ compatible = "fsl,ls1088a-qds", "fsl,ls1088a";
-+};
-+
-+&i2c0 {
-+ status = "okay";
-+
-+ i2c-switch@77 {
-+ compatible = "nxp,pca9547";
-+ reg = <0x77>;
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+
-+ i2c@2 {
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ reg = <0x2>;
-+
-+ ina220@40 {
-+ compatible = "ti,ina220";
-+ reg = <0x40>;
-+ shunt-resistor = <1000>;
-+ };
-+
-+ ina220@41 {
-+ compatible = "ti,ina220";
-+ reg = <0x41>;
-+ shunt-resistor = <1000>;
-+ };
-+ };
-+
-+ i2c@3 {
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ reg = <0x3>;
-+
-+ temp-sensor@4c {
-+ compatible = "adi,adt7461a";
-+ reg = <0x4c>;
-+ };
-+
-+ rtc@51 {
-+ compatible = "nxp,pcf2129";
-+ reg = <0x51>;
-+ /* IRQ10_B */
-+ interrupts = <0 150 IRQ_TYPE_LEVEL_HIGH>;
-+ };
-+
-+ eeprom@56 {
-+ compatible = "atmel,24c512";
-+ reg = <0x56>;
-+ };
-+
-+ eeprom@57 {
-+ compatible = "atmel,24c512";
-+ reg = <0x57>;
-+ };
-+ };
-+ };
-+};
-+
-+&qspi {
-+ status = "okay";
-+ qflash0: s25fs512s@0 {
-+ compatible = "spansion,m25p80";
-+ #address-cells = <1>;
-+ #size-cells = <1>;
-+ spi-max-frequency = <20000000>;
-+ m25p,fast-read;
-+ reg = <0>;
-+ };
-+
-+ qflash1: s25fs512s@1 {
-+ compatible = "spansion,m25p80";
-+ #address-cells = <1>;
-+ #size-cells = <1>;
-+ spi-max-frequency = <20000000>;
-+ m25p,fast-read;
-+ reg = <1>;
-+ };
-+};
-+
-+&ifc {
-+ status = "okay";
-+
-+ ranges = <0 0 0x5 0x80000000 0x08000000
-+ 2 0 0x5 0x30000000 0x00010000
-+ 3 0 0x5 0x20000000 0x00010000>;
-+
-+ nor@0,0 {
-+ compatible = "cfi-flash";
-+ reg = <0x0 0x0 0x8000000>;
-+ bank-width = <2>;
-+ device-width = <1>;
-+ };
-+
-+ nand@2,0 {
-+ compatible = "fsl,ifc-nand";
-+ reg = <0x2 0x0 0x10000>;
-+ };
-+
-+ fpga: board-control@3,0 {
-+ compatible = "fsl,ls1088aqds-fpga", "fsl,fpga-qixis";
-+ reg = <0x3 0x0 0x0000100>;
-+ };
-+};
-+
-+&duart0 {
-+ status = "okay";
-+};
-+
-+&duart1 {
-+ status = "okay";
-+};
-+
-+&esdhc {
-+ status = "okay";
-+};
-+
-+&sata {
-+ status = "okay";
-+};
---- /dev/null
-+++ b/arch/arm64/boot/dts/freescale/fsl-ls1088a-rdb.dts
-@@ -0,0 +1,200 @@
-+// SPDX-License-Identifier: (GPL-2.0+ OR MIT)
-+/*
-+ * Device Tree file for NXP LS1088A RDB Board.
-+ *
-+ * Copyright 2017 NXP
-+ *
-+ * Harninder Rai <harninder.rai@nxp.com>
-+ *
-+ */
-+
-+/dts-v1/;
-+
-+#include "fsl-ls1088a.dtsi"
-+
-+/ {
-+ model = "L1088A RDB Board";
-+ compatible = "fsl,ls1088a-rdb", "fsl,ls1088a";
-+};
-+
-+&i2c0 {
-+ status = "okay";
-+
-+ i2c-switch@77 {
-+ compatible = "nxp,pca9547";
-+ reg = <0x77>;
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+
-+ i2c@2 {
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ reg = <0x2>;
-+
-+ ina220@40 {
-+ compatible = "ti,ina220";
-+ reg = <0x40>;
-+ shunt-resistor = <1000>;
-+ };
-+ };
-+
-+ i2c@3 {
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ reg = <0x3>;
-+
-+ temp-sensor@4c {
-+ compatible = "adi,adt7461a";
-+ reg = <0x4c>;
-+ };
-+
-+ rtc@51 {
-+ compatible = "nxp,pcf2129";
-+ reg = <0x51>;
-+ /* IRQ10_B */
-+ interrupts = <0 150 IRQ_TYPE_LEVEL_HIGH>;
-+ };
-+ };
-+ };
-+};
-+
-+&qspi {
-+ status = "okay";
-+ qflash0: s25fs512s@0 {
-+ compatible = "spansion,m25p80";
-+ #address-cells = <1>;
-+ #size-cells = <1>;
-+ m25p,fast-read;
-+ spi-max-frequency = <20000000>;
-+ reg = <0>;
-+ };
-+
-+ qflash1: s25fs512s@1 {
-+ compatible = "spansion,m25p80";
-+ #address-cells = <1>;
-+ #size-cells = <1>;
-+ m25p,fast-read;
-+ spi-max-frequency = <20000000>;
-+ reg = <1>;
-+ };
-+};
-+
-+&ifc {
-+ status = "okay";
-+
-+ ranges = <0 0 0x5 0x30000000 0x00010000
-+ 2 0 0x5 0x20000000 0x00010000>;
-+
-+ nand@0,0 {
-+ compatible = "fsl,ifc-nand";
-+ reg = <0x0 0x0 0x10000>;
-+ };
-+
-+ fpga: board-control@2,0 {
-+ compatible = "fsl,ls1088ardb-fpga", "fsl,fpga-qixis";
-+ reg = <0x2 0x0 0x0000100>;
-+ };
-+};
-+
-+&duart0 {
-+ status = "okay";
-+};
-+
-+&duart1 {
-+ status = "okay";
-+};
-+
-+&usb0 {
-+ status = "okay";
-+};
-+
-+&usb1 {
-+ status = "okay";
-+};
-+
-+&esdhc {
-+ status = "okay";
-+};
-+
-+&sata {
-+ status = "okay";
-+};
-+
-+&emdio1 {
-+ /* Freescale F104 PHY1 */
-+ mdio1_phy1: emdio1_phy@1 {
-+ reg = <0x1c>;
-+ phy-connection-type = "qsgmii";
-+ };
-+ mdio1_phy2: emdio1_phy@2 {
-+ reg = <0x1d>;
-+ phy-connection-type = "qsgmii";
-+ };
-+ mdio1_phy3: emdio1_phy@3 {
-+ reg = <0x1e>;
-+ phy-connection-type = "qsgmii";
-+ };
-+ mdio1_phy4: emdio1_phy@4 {
-+ reg = <0x1f>;
-+ phy-connection-type = "qsgmii";
-+ };
-+ /* F104 PHY2 */
-+ mdio1_phy5: emdio1_phy@5 {
-+ reg = <0x0c>;
-+ phy-connection-type = "qsgmii";
-+ };
-+ mdio1_phy6: emdio1_phy@6 {
-+ reg = <0x0d>;
-+ phy-connection-type = "qsgmii";
-+ };
-+ mdio1_phy7: emdio1_phy@7 {
-+ reg = <0x0e>;
-+ phy-connection-type = "qsgmii";
-+ };
-+ mdio1_phy8: emdio1_phy@8 {
-+ reg = <0x0f>;
-+ phy-connection-type = "qsgmii";
-+ };
-+};
-+
-+&emdio2 {
-+ /* Aquantia AQR105 10G PHY */
-+ mdio2_phy1: emdio2_phy@1 {
-+ compatible = "ethernet-phy-ieee802.3-c45";
-+ interrupts = <0 2 0x4>;
-+ reg = <0x0>;
-+ phy-connection-type = "xfi";
-+ };
-+};
-+
-+/* DPMAC connections to external PHYs
-+ * based on LS1088A RM RevC - $24.1.2 SerDes Options
-+ */
-+/* DPMAC1 is 10G SFP+, fixed link */
-+&dpmac2 {
-+ phy-handle = <&mdio2_phy1>;
-+};
-+&dpmac3 {
-+ phy-handle = <&mdio1_phy5>;
-+};
-+&dpmac4 {
-+ phy-handle = <&mdio1_phy6>;
-+};
-+&dpmac5 {
-+ phy-handle = <&mdio1_phy7>;
-+};
-+&dpmac6 {
-+ phy-handle = <&mdio1_phy8>;
-+};
-+&dpmac7 {
-+ phy-handle = <&mdio1_phy1>;
-+};
-+&dpmac8 {
-+ phy-handle = <&mdio1_phy2>;
-+};
-+&dpmac9 {
-+ phy-handle = <&mdio1_phy3>;
-+};
-+&dpmac10 {
-+ phy-handle = <&mdio1_phy4>;
-+};
---- /dev/null
-+++ b/arch/arm64/boot/dts/freescale/fsl-ls1088a.dtsi
-@@ -0,0 +1,800 @@
-+// SPDX-License-Identifier: (GPL-2.0+ OR MIT)
-+/*
-+ * Device Tree Include file for NXP Layerscape-1088A family SoC.
-+ *
-+ * Copyright 2017 NXP
-+ *
-+ * Harninder Rai <harninder.rai@nxp.com>
-+ *
-+ */
-+#include <dt-bindings/interrupt-controller/arm-gic.h>
-+#include <dt-bindings/thermal/thermal.h>
-+
-+/ {
-+ compatible = "fsl,ls1088a";
-+ interrupt-parent = <&gic>;
-+ #address-cells = <2>;
-+ #size-cells = <2>;
-+
-+ aliases {
-+ crypto = &crypto;
-+ };
-+
-+ cpus {
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+
-+ /* We have 2 clusters having 4 Cortex-A53 cores each */
-+ cpu0: cpu@0 {
-+ device_type = "cpu";
-+ compatible = "arm,cortex-a53";
-+ reg = <0x0>;
-+ clocks = <&clockgen 1 0>;
-+ #cooling-cells = <2>;
-+ cpu-idle-states = <&CPU_PH20>;
-+ };
-+
-+ cpu1: cpu@1 {
-+ device_type = "cpu";
-+ compatible = "arm,cortex-a53";
-+ reg = <0x1>;
-+ clocks = <&clockgen 1 0>;
-+ cpu-idle-states = <&CPU_PH20>;
-+ };
-+
-+ cpu2: cpu@2 {
-+ device_type = "cpu";
-+ compatible = "arm,cortex-a53";
-+ reg = <0x2>;
-+ clocks = <&clockgen 1 0>;
-+ cpu-idle-states = <&CPU_PH20>;
-+ };
-+
-+ cpu3: cpu@3 {
-+ device_type = "cpu";
-+ compatible = "arm,cortex-a53";
-+ reg = <0x3>;
-+ clocks = <&clockgen 1 0>;
-+ cpu-idle-states = <&CPU_PH20>;
-+ };
-+
-+ cpu4: cpu@100 {
-+ device_type = "cpu";
-+ compatible = "arm,cortex-a53";
-+ reg = <0x100>;
-+ clocks = <&clockgen 1 1>;
-+ #cooling-cells = <2>;
-+ cpu-idle-states = <&CPU_PH20>;
-+ };
-+
-+ cpu5: cpu@101 {
-+ device_type = "cpu";
-+ compatible = "arm,cortex-a53";
-+ reg = <0x101>;
-+ clocks = <&clockgen 1 1>;
-+ cpu-idle-states = <&CPU_PH20>;
-+ };
-+
-+ cpu6: cpu@102 {
-+ device_type = "cpu";
-+ compatible = "arm,cortex-a53";
-+ reg = <0x102>;
-+ clocks = <&clockgen 1 1>;
-+ cpu-idle-states = <&CPU_PH20>;
-+ };
-+
-+ cpu7: cpu@103 {
-+ device_type = "cpu";
-+ compatible = "arm,cortex-a53";
-+ reg = <0x103>;
-+ clocks = <&clockgen 1 1>;
-+ cpu-idle-states = <&CPU_PH20>;
-+ };
-+ };
-+
-+ idle-states {
-+ /*
-+ * PSCI node is not added default, U-boot will add missing
-+ * parts if it determines to use PSCI.
-+ */
-+ entry-method = "arm,psci";
-+
-+ CPU_PH20: cpu-ph20 {
-+ compatible = "arm,idle-state";
-+ idle-state-name = "PH20";
-+ arm,psci-suspend-param = <0x0>;
-+ entry-latency-us = <1000>;
-+ exit-latency-us = <1000>;
-+ min-residency-us = <3000>;
-+ };
-+ };
-+
-+ gic: interrupt-controller@6000000 {
-+ compatible = "arm,gic-v3";
-+ #interrupt-cells = <3>;
-+ #address-cells = <2>;
-+ #size-cells = <2>;
-+ ranges;
-+ interrupt-controller;
-+ reg = <0x0 0x06000000 0 0x10000>, /* GIC Dist */
-+ <0x0 0x06100000 0 0x100000>, /* GICR(RD_base+SGI_base)*/
-+ <0x0 0x0c0c0000 0 0x2000>, /* GICC */
-+ <0x0 0x0c0d0000 0 0x1000>, /* GICH */
-+ <0x0 0x0c0e0000 0 0x20000>; /* GICV */
-+ interrupts = <1 9 IRQ_TYPE_LEVEL_HIGH>;
-+
-+ its: gic-its@6020000 {
-+ compatible = "arm,gic-v3-its";
-+ msi-controller;
-+ reg = <0x0 0x6020000 0 0x20000>;
-+ };
-+ };
-+
-+ timer {
-+ compatible = "arm,armv8-timer";
-+ interrupts = <1 13 IRQ_TYPE_LEVEL_LOW>,/* Physical Secure PPI */
-+ <1 14 IRQ_TYPE_LEVEL_LOW>,/* Physical Non-Secure PPI */
-+ <1 11 IRQ_TYPE_LEVEL_LOW>,/* Virtual PPI */
-+ <1 10 IRQ_TYPE_LEVEL_LOW>;/* Hypervisor PPI */
-+ };
-+
-+ sysclk: sysclk {
-+ compatible = "fixed-clock";
-+ #clock-cells = <0>;
-+ clock-frequency = <100000000>;
-+ clock-output-names = "sysclk";
-+ };
-+
-+ dcfg: dcfg@1e00000 {
-+ compatible = "fsl,ls1088a-dcfg", "syscon";
-+ reg = <0x0 0x1e00000 0x0 0x10000>;
-+ little-endian;
-+ };
-+
-+ rstcr: syscon@1e60000 {
-+ compatible = "fsl,ls1088a-rstcr", "syscon";
-+ reg = <0x0 0x1e60000 0x0 0x4>;
-+ };
-+
-+ reboot {
-+ compatible = "syscon-reboot";
-+ regmap = <&rstcr>;
-+ offset = <0x0>;
-+ mask = <0x02>;
-+ };
-+
-+
-+ soc {
-+ compatible = "simple-bus";
-+ #address-cells = <2>;
-+ #size-cells = <2>;
-+ ranges;
-+ dma-ranges = <0x0 0x0 0x0 0x0 0x10000 0x00000000>;
-+
-+ clockgen: clocking@1300000 {
-+ compatible = "fsl,ls1088a-clockgen";
-+ reg = <0 0x1300000 0 0xa0000>;
-+ #clock-cells = <2>;
-+ clocks = <&sysclk>;
-+ };
-+
-+ tmu: tmu@1f80000 {
-+ compatible = "fsl,qoriq-tmu";
-+ reg = <0x0 0x1f80000 0x0 0x10000>;
-+ interrupts = <0 23 0x4>;
-+ fsl,tmu-range = <0xb0000 0x9002a 0x6004c 0x30062>;
-+ fsl,tmu-calibration =
-+ /* Calibration data group 1 */
-+ <0x00000000 0x00000026
-+ 0x00000001 0x0000002d
-+ 0x00000002 0x00000032
-+ 0x00000003 0x00000039
-+ 0x00000004 0x0000003f
-+ 0x00000005 0x00000046
-+ 0x00000006 0x0000004d
-+ 0x00000007 0x00000054
-+ 0x00000008 0x0000005a
-+ 0x00000009 0x00000061
-+ 0x0000000a 0x0000006a
-+ 0x0000000b 0x00000071
-+ /* Calibration data group 2 */
-+ 0x00010000 0x00000025
-+ 0x00010001 0x0000002c
-+ 0x00010002 0x00000035
-+ 0x00010003 0x0000003d
-+ 0x00010004 0x00000045
-+ 0x00010005 0x0000004e
-+ 0x00010006 0x00000057
-+ 0x00010007 0x00000061
-+ 0x00010008 0x0000006b
-+ 0x00010009 0x00000076
-+ /* Calibration data group 3 */
-+ 0x00020000 0x00000029
-+ 0x00020001 0x00000033
-+ 0x00020002 0x0000003d
-+ 0x00020003 0x00000049
-+ 0x00020004 0x00000056
-+ 0x00020005 0x00000061
-+ 0x00020006 0x0000006d
-+ /* Calibration data group 4 */
-+ 0x00030000 0x00000021
-+ 0x00030001 0x0000002a
-+ 0x00030002 0x0000003c
-+ 0x00030003 0x0000004e>;
-+ little-endian;
-+ #thermal-sensor-cells = <1>;
-+ };
-+
-+ thermal-zones {
-+ cpu_thermal: cpu-thermal {
-+ polling-delay-passive = <1000>;
-+ polling-delay = <5000>;
-+ thermal-sensors = <&tmu 0>;
-+
-+ trips {
-+ cpu_alert: cpu-alert {
-+ temperature = <85000>;
-+ hysteresis = <2000>;
-+ type = "passive";
-+ };
-+
-+ cpu_crit: cpu-crit {
-+ temperature = <95000>;
-+ hysteresis = <2000>;
-+ type = "critical";
-+ };
-+ };
-+
-+ cooling-maps {
-+ map0 {
-+ trip = <&cpu_alert>;
-+ cooling-device =
-+ <&cpu0 THERMAL_NO_LIMIT
-+ THERMAL_NO_LIMIT>;
-+ };
-+ map1 {
-+ trip = <&cpu_alert>;
-+ cooling-device =
-+ <&cpu4 THERMAL_NO_LIMIT
-+ THERMAL_NO_LIMIT>;
-+ };
-+ };
-+ };
-+ };
-+
-+ duart0: serial@21c0500 {
-+ compatible = "fsl,ns16550", "ns16550a";
-+ reg = <0x0 0x21c0500 0x0 0x100>;
-+ clocks = <&clockgen 4 3>;
-+ interrupts = <0 32 IRQ_TYPE_LEVEL_HIGH>;
-+ status = "disabled";
-+ };
-+
-+ duart1: serial@21c0600 {
-+ compatible = "fsl,ns16550", "ns16550a";
-+ reg = <0x0 0x21c0600 0x0 0x100>;
-+ clocks = <&clockgen 4 3>;
-+ interrupts = <0 32 IRQ_TYPE_LEVEL_HIGH>;
-+ status = "disabled";
-+ };
-+
-+ cluster1_core0_watchdog: wdt@c000000 {
-+ compatible = "arm,sp805-wdt", "arm,primecell";
-+ reg = <0x0 0xc000000 0x0 0x1000>;
-+ clocks = <&clockgen 4 3>, <&clockgen 4 3>;
-+ clock-names = "apb_pclk", "wdog_clk";
-+ };
-+
-+ cluster1_core1_watchdog: wdt@c010000 {
-+ compatible = "arm,sp805-wdt", "arm,primecell";
-+ reg = <0x0 0xc010000 0x0 0x1000>;
-+ clocks = <&clockgen 4 3>, <&clockgen 4 3>;
-+ clock-names = "apb_pclk", "wdog_clk";
-+ };
-+
-+ cluster1_core2_watchdog: wdt@c020000 {
-+ compatible = "arm,sp805-wdt", "arm,primecell";
-+ reg = <0x0 0xc020000 0x0 0x1000>;
-+ clocks = <&clockgen 4 3>, <&clockgen 4 3>;
-+ clock-names = "apb_pclk", "wdog_clk";
-+ };
-+
-+ cluster1_core3_watchdog: wdt@c030000 {
-+ compatible = "arm,sp805-wdt", "arm,primecell";
-+ reg = <0x0 0xc030000 0x0 0x1000>;
-+ clocks = <&clockgen 4 3>, <&clockgen 4 3>;
-+ clock-names = "apb_pclk", "wdog_clk";
-+ };
-+
-+ cluster2_core0_watchdog: wdt@c100000 {
-+ compatible = "arm,sp805-wdt", "arm,primecell";
-+ reg = <0x0 0xc100000 0x0 0x1000>;
-+ clocks = <&clockgen 4 3>, <&clockgen 4 3>;
-+ clock-names = "apb_pclk", "wdog_clk";
-+ };
-+
-+ cluster2_core1_watchdog: wdt@c110000 {
-+ compatible = "arm,sp805-wdt", "arm,primecell";
-+ reg = <0x0 0xc110000 0x0 0x1000>;
-+ clocks = <&clockgen 4 3>, <&clockgen 4 3>;
-+ clock-names = "apb_pclk", "wdog_clk";
-+ };
-+
-+ cluster2_core2_watchdog: wdt@c120000 {
-+ compatible = "arm,sp805-wdt", "arm,primecell";
-+ reg = <0x0 0xc120000 0x0 0x1000>;
-+ clocks = <&clockgen 4 3>, <&clockgen 4 3>;
-+ clock-names = "apb_pclk", "wdog_clk";
-+ };
-+
-+ cluster2_core3_watchdog: wdt@c130000 {
-+ compatible = "arm,sp805-wdt", "arm,primecell";
-+ reg = <0x0 0xc130000 0x0 0x1000>;
-+ clocks = <&clockgen 4 3>, <&clockgen 4 3>;
-+ clock-names = "apb_pclk", "wdog_clk";
-+ };
-+
-+ gpio0: gpio@2300000 {
-+ compatible = "fsl,qoriq-gpio";
-+ reg = <0x0 0x2300000 0x0 0x10000>;
-+ interrupts = <0 36 IRQ_TYPE_LEVEL_HIGH>;
-+ gpio-controller;
-+ #gpio-cells = <2>;
-+ interrupt-controller;
-+ #interrupt-cells = <2>;
-+ };
-+
-+ gpio1: gpio@2310000 {
-+ compatible = "fsl,qoriq-gpio";
-+ reg = <0x0 0x2310000 0x0 0x10000>;
-+ interrupts = <0 36 IRQ_TYPE_LEVEL_HIGH>;
-+ gpio-controller;
-+ #gpio-cells = <2>;
-+ interrupt-controller;
-+ #interrupt-cells = <2>;
-+ };
-+
-+ gpio2: gpio@2320000 {
-+ compatible = "fsl,qoriq-gpio";
-+ reg = <0x0 0x2320000 0x0 0x10000>;
-+ interrupts = <0 37 IRQ_TYPE_LEVEL_HIGH>;
-+ gpio-controller;
-+ #gpio-cells = <2>;
-+ interrupt-controller;
-+ #interrupt-cells = <2>;
-+ };
-+
-+ gpio3: gpio@2330000 {
-+ compatible = "fsl,qoriq-gpio";
-+ reg = <0x0 0x2330000 0x0 0x10000>;
-+ interrupts = <0 37 IRQ_TYPE_LEVEL_HIGH>;
-+ gpio-controller;
-+ #gpio-cells = <2>;
-+ interrupt-controller;
-+ #interrupt-cells = <2>;
-+ };
-+
-+ /* TODO: WRIOP (CCSR?) */
-+ emdio1: mdio@0x8B96000 { /* WRIOP0: 0x8B8_0000,
-+ * E-MDIO1: 0x1_6000
-+ */
-+ compatible = "fsl,fman-memac-mdio";
-+ reg = <0x0 0x8B96000 0x0 0x1000>;
-+ device_type = "mdio";
-+ little-endian; /* force the driver in LE mode */
-+
-+ /* Not necessary on the QDS, but needed on the RDB */
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ };
-+
-+ emdio2: mdio@0x8B97000 { /* WRIOP0: 0x8B8_0000,
-+ * E-MDIO2: 0x1_7000
-+ */
-+ compatible = "fsl,fman-memac-mdio";
-+ reg = <0x0 0x8B97000 0x0 0x1000>;
-+ device_type = "mdio";
-+ little-endian; /* force the driver in LE mode */
-+
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ };
-+
-+ ifc: ifc@2240000 {
-+ compatible = "fsl,ifc", "simple-bus";
-+ reg = <0x0 0x2240000 0x0 0x20000>;
-+ interrupts = <0 21 IRQ_TYPE_LEVEL_HIGH>;
-+ little-endian;
-+ #address-cells = <2>;
-+ #size-cells = <1>;
-+
-+ };
-+
-+ ftm0: ftm0@2800000 {
-+ compatible = "fsl,ls1088a-ftm";
-+ reg = <0x0 0x2800000 0x0 0x10000>,
-+ <0x0 0x1e34050 0x0 0x4>;
-+ interrupts = <0 44 4>;
-+ reg-names = "ftm", "FlexTimer1";
-+ };
-+
-+ i2c0: i2c@2000000 {
-+ compatible = "fsl,vf610-i2c", "fsl,ls1088a-vf610-i2c";
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ reg = <0x0 0x2000000 0x0 0x10000>;
-+ interrupts = <0 34 IRQ_TYPE_LEVEL_HIGH>;
-+ clocks = <&clockgen 4 7>;
-+ fsl-scl-gpio = <&gpio3 30 0>;
-+ status = "disabled";
-+ };
-+
-+ i2c1: i2c@2010000 {
-+ compatible = "fsl,vf610-i2c";
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ reg = <0x0 0x2010000 0x0 0x10000>;
-+ interrupts = <0 34 IRQ_TYPE_LEVEL_HIGH>;
-+ clocks = <&clockgen 4 7>;
-+ status = "disabled";
-+ };
-+
-+ i2c2: i2c@2020000 {
-+ compatible = "fsl,vf610-i2c";
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ reg = <0x0 0x2020000 0x0 0x10000>;
-+ interrupts = <0 35 IRQ_TYPE_LEVEL_HIGH>;
-+ clocks = <&clockgen 4 7>;
-+ status = "disabled";
-+ };
-+
-+ i2c3: i2c@2030000 {
-+ compatible = "fsl,vf610-i2c";
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ reg = <0x0 0x2030000 0x0 0x10000>;
-+ interrupts = <0 35 IRQ_TYPE_LEVEL_HIGH>;
-+ clocks = <&clockgen 4 7>;
-+ status = "disabled";
-+ };
-+
-+ qspi: quadspi@20c0000 {
-+ compatible = "fsl,ls2080a-qspi", "fsl,ls1088a-qspi";
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ reg = <0x0 0x20c0000 0x0 0x10000>,
-+ <0x0 0x20000000 0x0 0x10000000>;
-+ reg-names = "QuadSPI", "QuadSPI-memory";
-+ interrupts = <0 25 0x4>; /* Level high type */
-+ clocks = <&clockgen 4 3>, <&clockgen 4 3>;
-+ clock-names = "qspi_en", "qspi";
-+ fsl,qspi-has-second-chip;
-+ };
-+
-+ esdhc: esdhc@2140000 {
-+ compatible = "fsl,ls1088a-esdhc", "fsl,esdhc";
-+ reg = <0x0 0x2140000 0x0 0x10000>;
-+ interrupts = <0 28 0x4>; /* Level high type */
-+ clock-frequency = <0>;
-+ voltage-ranges = <1800 1800 3300 3300>;
-+ sdhci,auto-cmd12;
-+ little-endian;
-+ bus-width = <4>;
-+ status = "disabled";
-+ };
-+
-+ usb0: usb3@3100000 {
-+ compatible = "snps,dwc3";
-+ reg = <0x0 0x3100000 0x0 0x10000>;
-+ interrupts = <0 80 0x4>; /* Level high type */
-+ dr_mode = "host";
-+ configure-gfladj;
-+ snps,dis_rxdet_inp3_quirk;
-+ };
-+
-+ usb1: usb3@3110000 {
-+ compatible = "snps,dwc3";
-+ reg = <0x0 0x3110000 0x0 0x10000>;
-+ interrupts = <0 81 0x4>; /* Level high type */
-+ dr_mode = "host";
-+ configure-gfladj;
-+ snps,dis_rxdet_inp3_quirk;
-+ };
-+
-+ sata: sata@3200000 {
-+ compatible = "fsl,ls1088a-ahci";
-+ reg = <0x0 0x3200000 0x0 0x10000>,
-+ <0x7 0x100520 0x0 0x4>;
-+ reg-names = "ahci", "sata-ecc";
-+ interrupts = <0 133 IRQ_TYPE_LEVEL_HIGH>;
-+ clocks = <&clockgen 4 3>;
-+ dma-coherent;
-+ status = "disabled";
-+ };
-+
-+ pcie@3400000 {
-+ compatible = "fsl,ls2088a-pcie", "fsl,ls1088a-pcie",
-+ "snps,dw-pcie";
-+ reg = <0x00 0x03400000 0x0 0x00100000 /* controller registers */
-+ 0x20 0x00000000 0x0 0x00002000>; /* configuration space */
-+ reg-names = "regs", "config";
-+ interrupts = <0 108 IRQ_TYPE_LEVEL_HIGH>; /* aer interrupt */
-+ interrupt-names = "aer";
-+ #address-cells = <3>;
-+ #size-cells = <2>;
-+ device_type = "pci";
-+ dma-coherent;
-+ num-lanes = <4>;
-+ bus-range = <0x0 0xff>;
-+ ranges = <0x81000000 0x0 0x00000000 0x20 0x00010000 0x0 0x00010000 /* downstream I/O */
-+ 0x82000000 0x0 0x40000000 0x20 0x40000000 0x0 0x40000000>; /* non-prefetchable memory */
-+ msi-parent = <&its>;
-+ iommu-map = <0 &smmu 0 1>; /* This is fixed-up by u-boot */
-+ #interrupt-cells = <1>;
-+ interrupt-map-mask = <0 0 0 7>;
-+ interrupt-map = <0000 0 0 1 &gic 0 0 0 109 IRQ_TYPE_LEVEL_HIGH>,
-+ <0000 0 0 2 &gic 0 0 0 110 IRQ_TYPE_LEVEL_HIGH>,
-+ <0000 0 0 3 &gic 0 0 0 111 IRQ_TYPE_LEVEL_HIGH>,
-+ <0000 0 0 4 &gic 0 0 0 112 IRQ_TYPE_LEVEL_HIGH>;
-+ };
-+
-+ pcie@3500000 {
-+ compatible = "fsl,ls2088a-pcie", "fsl,ls1088a-pcie",
-+ "snps,dw-pcie";
-+ reg = <0x00 0x03500000 0x0 0x00100000 /* controller registers */
-+ 0x28 0x00000000 0x0 0x00002000>; /* configuration space */
-+ reg-names = "regs", "config";
-+ interrupts = <0 113 IRQ_TYPE_LEVEL_HIGH>; /* aer interrupt */
-+ interrupt-names = "aer";
-+ #address-cells = <3>;
-+ #size-cells = <2>;
-+ device_type = "pci";
-+ dma-coherent;
-+ num-lanes = <4>;
-+ bus-range = <0x0 0xff>;
-+ ranges = <0x81000000 0x0 0x00000000 0x28 0x00010000 0x0 0x00010000 /* downstream I/O */
-+ 0x82000000 0x0 0x40000000 0x28 0x40000000 0x0 0x40000000>; /* non-prefetchable memory */
-+ msi-parent = <&its>;
-+ iommu-map = <0 &smmu 0 1>; /* This is fixed-up by u-boot */
-+ #interrupt-cells = <1>;
-+ interrupt-map-mask = <0 0 0 7>;
-+ interrupt-map = <0000 0 0 1 &gic 0 0 0 114 IRQ_TYPE_LEVEL_HIGH>,
-+ <0000 0 0 2 &gic 0 0 0 115 IRQ_TYPE_LEVEL_HIGH>,
-+ <0000 0 0 3 &gic 0 0 0 116 IRQ_TYPE_LEVEL_HIGH>,
-+ <0000 0 0 4 &gic 0 0 0 117 IRQ_TYPE_LEVEL_HIGH>;
-+ };
-+
-+ pcie@3600000 {
-+ compatible = "fsl,ls2088a-pcie", "fsl,ls1088a-pcie",
-+ "snps,dw-pcie";
-+ reg = <0x00 0x03600000 0x0 0x00100000 /* controller registers */
-+ 0x30 0x00000000 0x0 0x00002000>; /* configuration space */
-+ reg-names = "regs", "config";
-+ interrupts = <0 118 IRQ_TYPE_LEVEL_HIGH>; /* aer interrupt */
-+ interrupt-names = "aer";
-+ #address-cells = <3>;
-+ #size-cells = <2>;
-+ device_type = "pci";
-+ dma-coherent;
-+ num-lanes = <8>;
-+ bus-range = <0x0 0xff>;
-+ ranges = <0x81000000 0x0 0x00000000 0x30 0x00010000 0x0 0x00010000 /* downstream I/O */
-+ 0x82000000 0x0 0x40000000 0x30 0x40000000 0x0 0x40000000>; /* non-prefetchable memory */
-+ msi-parent = <&its>;
-+ iommu-map = <0 &smmu 0 1>; /* This is fixed-up by u-boot */
-+ #interrupt-cells = <1>;
-+ interrupt-map-mask = <0 0 0 7>;
-+ interrupt-map = <0000 0 0 1 &gic 0 0 0 119 IRQ_TYPE_LEVEL_HIGH>,
-+ <0000 0 0 2 &gic 0 0 0 120 IRQ_TYPE_LEVEL_HIGH>,
-+ <0000 0 0 3 &gic 0 0 0 121 IRQ_TYPE_LEVEL_HIGH>,
-+ <0000 0 0 4 &gic 0 0 0 122 IRQ_TYPE_LEVEL_HIGH>;
-+ };
-+
-+ fsl_mc: fsl-mc@80c000000 {
-+ compatible = "fsl,qoriq-mc";
-+ reg = <0x00000008 0x0c000000 0 0x40>, /* MC portal base */
-+ <0x00000000 0x08340000 0 0x40000>; /* MC control reg */
-+ msi-parent = <&its>;
-+ iommu-map = <0 &smmu 0 0>; /* This is fixed-up by u-boot */
-+ dma-coherent;
-+ #address-cells = <3>;
-+ #size-cells = <1>;
-+
-+ /*
-+ * Region type 0x0 - MC portals
-+ * Region type 0x1 - QBMAN portals
-+ */
-+ ranges = <0x0 0x0 0x0 0x8 0x0c000000 0x4000000
-+ 0x1 0x0 0x0 0x8 0x18000000 0x8000000>;
-+
-+ dpmacs {
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+
-+ dpmac1: dpmac@1 {
-+ compatible = "fsl,qoriq-mc-dpmac";
-+ reg = <1>;
-+ };
-+
-+ dpmac2: dpmac@2 {
-+ compatible = "fsl,qoriq-mc-dpmac";
-+ reg = <2>;
-+ };
-+
-+ dpmac3: dpmac@3 {
-+ compatible = "fsl,qoriq-mc-dpmac";
-+ reg = <3>;
-+ };
-+
-+ dpmac4: dpmac@4 {
-+ compatible = "fsl,qoriq-mc-dpmac";
-+ reg = <4>;
-+ };
-+
-+ dpmac5: dpmac@5 {
-+ compatible = "fsl,qoriq-mc-dpmac";
-+ reg = <5>;
-+ };
-+
-+ dpmac6: dpmac@6 {
-+ compatible = "fsl,qoriq-mc-dpmac";
-+ reg = <6>;
-+ };
-+
-+ dpmac7: dpmac@7 {
-+ compatible = "fsl,qoriq-mc-dpmac";
-+ reg = <7>;
-+ };
-+
-+ dpmac8: dpmac@8 {
-+ compatible = "fsl,qoriq-mc-dpmac";
-+ reg = <8>;
-+ };
-+
-+ dpmac9: dpmac@9 {
-+ compatible = "fsl,qoriq-mc-dpmac";
-+ reg = <9>;
-+ };
-+
-+ dpmac10: dpmac@a {
-+ compatible = "fsl,qoriq-mc-dpmac";
-+ reg = <0xa>;
-+ };
-+ };
-+ };
-+
-+ smmu: iommu@5000000 {
-+ compatible = "arm,mmu-500";
-+ reg = <0 0x5000000 0 0x800000>;
-+ #global-interrupts = <12>;
-+ #iommu-cells = <1>;
-+ stream-match-mask = <0x7C00>;
-+ interrupts = <0 13 4>, /* global secure fault */
-+ <0 14 4>, /* combined secure interrupt */
-+ <0 15 4>, /* global non-secure fault */
-+ <0 16 4>, /* combined non-secure interrupt */
-+ /* performance counter interrupts 0-7 */
-+ <0 211 4>,
-+ <0 212 4>,
-+ <0 213 4>,
-+ <0 214 4>,
-+ <0 215 4>,
-+ <0 216 4>,
-+ <0 217 4>,
-+ <0 218 4>,
-+ /* per context interrupt, 64 interrupts */
-+ <0 146 4>,
-+ <0 147 4>,
-+ <0 148 4>,
-+ <0 149 4>,
-+ <0 150 4>,
-+ <0 151 4>,
-+ <0 152 4>,
-+ <0 153 4>,
-+ <0 154 4>,
-+ <0 155 4>,
-+ <0 156 4>,
-+ <0 157 4>,
-+ <0 158 4>,
-+ <0 159 4>,
-+ <0 160 4>,
-+ <0 161 4>,
-+ <0 162 4>,
-+ <0 163 4>,
-+ <0 164 4>,
-+ <0 165 4>,
-+ <0 166 4>,
-+ <0 167 4>,
-+ <0 168 4>,
-+ <0 169 4>,
-+ <0 170 4>,
-+ <0 171 4>,
-+ <0 172 4>,
-+ <0 173 4>,
-+ <0 174 4>,
-+ <0 175 4>,
-+ <0 176 4>,
-+ <0 177 4>,
-+ <0 178 4>,
-+ <0 179 4>,
-+ <0 180 4>,
-+ <0 181 4>,
-+ <0 182 4>,
-+ <0 183 4>,
-+ <0 184 4>,
-+ <0 185 4>,
-+ <0 186 4>,
-+ <0 187 4>,
-+ <0 188 4>,
-+ <0 189 4>,
-+ <0 190 4>,
-+ <0 191 4>,
-+ <0 192 4>,
-+ <0 193 4>,
-+ <0 194 4>,
-+ <0 195 4>,
-+ <0 196 4>,
-+ <0 197 4>,
-+ <0 198 4>,
-+ <0 199 4>,
-+ <0 200 4>,
-+ <0 201 4>,
-+ <0 202 4>,
-+ <0 203 4>,
-+ <0 204 4>,
-+ <0 205 4>,
-+ <0 206 4>,
-+ <0 207 4>,
-+ <0 208 4>,
-+ <0 209 4>;
-+ };
-+
-+ crypto: crypto@8000000 {
-+ compatible = "fsl,sec-v5.0", "fsl,sec-v4.0";
-+ fsl,sec-era = <8>;
-+ #address-cells = <1>;
-+ #size-cells = <1>;
-+ ranges = <0x0 0x00 0x8000000 0x100000>;
-+ reg = <0x00 0x8000000 0x0 0x100000>;
-+ interrupts = <GIC_SPI 139 IRQ_TYPE_LEVEL_HIGH>;
-+ dma-coherent;
-+
-+ sec_jr0: jr@10000 {
-+ compatible = "fsl,sec-v5.0-job-ring",
-+ "fsl,sec-v4.0-job-ring";
-+ reg = <0x10000 0x10000>;
-+ interrupts = <GIC_SPI 140 IRQ_TYPE_LEVEL_HIGH>;
-+ };
-+
-+ sec_jr1: jr@20000 {
-+ compatible = "fsl,sec-v5.0-job-ring",
-+ "fsl,sec-v4.0-job-ring";
-+ reg = <0x20000 0x10000>;
-+ interrupts = <GIC_SPI 141 IRQ_TYPE_LEVEL_HIGH>;
-+ };
-+
-+ sec_jr2: jr@30000 {
-+ compatible = "fsl,sec-v5.0-job-ring",
-+ "fsl,sec-v4.0-job-ring";
-+ reg = <0x30000 0x10000>;
-+ interrupts = <GIC_SPI 142 IRQ_TYPE_LEVEL_HIGH>;
-+ };
-+
-+ sec_jr3: jr@40000 {
-+ compatible = "fsl,sec-v5.0-job-ring",
-+ "fsl,sec-v4.0-job-ring";
-+ reg = <0x40000 0x10000>;
-+ interrupts = <GIC_SPI 143 IRQ_TYPE_LEVEL_HIGH>;
-+ };
-+ };
-+ };
-+
-+ firmware {
-+ optee {
-+ compatible = "linaro,optee-tz";
-+ method = "smc";
-+ };
-+ };
-+
-+};
---- a/arch/arm64/boot/dts/freescale/fsl-ls2080a-qds.dts
-+++ b/arch/arm64/boot/dts/freescale/fsl-ls2080a-qds.dts
-@@ -1,214 +1,87 @@
-+// SPDX-License-Identifier: (GPL-2.0+ OR MIT)
- /*
- * Device Tree file for Freescale LS2080a QDS Board.
- *
-- * Copyright (C) 2015, Freescale Semiconductor
-+ * Copyright 2015-2016 Freescale Semiconductor, Inc.
-+ * Copyright 2017 NXP
- *
-+ * Abhimanyu Saini <abhimanyu.saini@nxp.com>
- * Bhupesh Sharma <bhupesh.sharma@freescale.com>
- *
-- * This file is dual-licensed: you can use it either under the terms
-- * of the GPLv2 or the X11 license, at your option. Note that this dual
-- * licensing only applies to this file, and not this project as a
-- * whole.
-- *
-- * a) This library is free software; you can redistribute it and/or
-- * modify it under the terms of the GNU General Public License as
-- * published by the Free Software Foundation; either version 2 of the
-- * License, or (at your option) any later version.
-- *
-- * This library is distributed in the hope that it will be useful,
-- * but WITHOUT ANY WARRANTY; without even the implied warranty of
-- * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
-- * GNU General Public License for more details.
-- *
-- * Or, alternatively,
-- *
-- * b) Permission is hereby granted, free of charge, to any person
-- * obtaining a copy of this software and associated documentation
-- * files (the "Software"), to deal in the Software without
-- * restriction, including without limitation the rights to use,
-- * copy, modify, merge, publish, distribute, sublicense, and/or
-- * sell copies of the Software, and to permit persons to whom the
-- * Software is furnished to do so, subject to the following
-- * conditions:
-- *
-- * The above copyright notice and this permission notice shall be
-- * included in all copies or substantial portions of the Software.
-- *
-- * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
-- * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES
-- * OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
-- * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT
-- * HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY,
-- * WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
-- * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
-- * OTHER DEALINGS IN THE SOFTWARE.
- */
-
- /dts-v1/;
-
--/include/ "fsl-ls2080a.dtsi"
-+#include "fsl-ls2080a.dtsi"
-+#include "fsl-ls208xa-qds.dtsi"
-
- / {
- model = "Freescale Layerscape 2080a QDS Board";
- compatible = "fsl,ls2080a-qds", "fsl,ls2080a";
-
-- aliases {
-- serial0 = &serial0;
-- serial1 = &serial1;
-- };
--
- chosen {
- stdout-path = "serial0:115200n8";
- };
- };
-
--&esdhc {
-- status = "okay";
--};
--
- &ifc {
-- status = "okay";
-- #address-cells = <2>;
-- #size-cells = <1>;
-- ranges = <0x0 0x0 0x5 0x80000000 0x08000000
-- 0x2 0x0 0x5 0x30000000 0x00010000
-- 0x3 0x0 0x5 0x20000000 0x00010000>;
--
-- nor@0,0 {
-+ boardctrl: board-control@3,0 {
- #address-cells = <1>;
- #size-cells = <1>;
-- compatible = "cfi-flash";
-- reg = <0x0 0x0 0x8000000>;
-- bank-width = <2>;
-- device-width = <1>;
-- };
--
-- nand@2,0 {
-- compatible = "fsl,ifc-nand";
-- reg = <0x2 0x0 0x10000>;
-- };
-+ compatible = "fsl,tetra-fpga", "fsl,fpga-qixis", "simple-bus";
-+ reg = <3 0 0x300>; /* TODO check address */
-+ ranges = <0 3 0 0x300>;
-
-- cpld@3,0 {
-- reg = <0x3 0x0 0x10000>;
-- compatible = "fsl,ls2080aqds-fpga", "fsl,fpga-qixis";
-- };
--};
--
--&i2c0 {
-- status = "okay";
-- pca9547@77 {
-- compatible = "nxp,pca9547";
-- reg = <0x77>;
-- #address-cells = <1>;
-- #size-cells = <0>;
-- i2c@0 {
-- #address-cells = <1>;
-- #size-cells = <0>;
-- reg = <0x00>;
-- rtc@68 {
-- compatible = "dallas,ds3232";
-- reg = <0x68>;
-- };
-- };
--
-- i2c@2 {
-- #address-cells = <1>;
-- #size-cells = <0>;
-- reg = <0x02>;
--
-- ina220@40 {
-- compatible = "ti,ina220";
-- reg = <0x40>;
-- shunt-resistor = <500>;
-- };
--
-- ina220@41 {
-- compatible = "ti,ina220";
-- reg = <0x41>;
-- shunt-resistor = <1000>;
-- };
-- };
-+ mdio_mux_emi1 {
-+ compatible = "mdio-mux-mmioreg", "mdio-mux";
-+ mdio-parent-bus = <&emdio1>;
-+ reg = <0x54 1>; /* BRDCFG4 */
-+ mux-mask = <0xe0>; /* EMI1_MDIO */
-
-- i2c@3 {
-- #address-cells = <1>;
-+ #address-cells=<1>;
- #size-cells = <0>;
-- reg = <0x3>;
-
-- adt7481@4c {
-- compatible = "adi,adt7461";
-- reg = <0x4c>;
-+ /* Child MDIO buses, one for each riser card:
-+ * reg = 0x0, 0x20, 0x40, 0x60, 0x80, 0xa0.
-+ * VSC8234 PHYs on the riser cards.
-+ */
-+
-+ mdio_mux3: mdio@60 {
-+ reg = <0x60>;
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+
-+ mdio0_phy12: mdio_phy0@1c {
-+ reg = <0x1c>;
-+ phy-connection-type = "sgmii";
-+ };
-+ mdio0_phy13: mdio_phy1@1d {
-+ reg = <0x1d>;
-+ phy-connection-type = "sgmii";
-+ };
-+ mdio0_phy14: mdio_phy2@1e {
-+ reg = <0x1e>;
-+ phy-connection-type = "sgmii";
-+ };
-+ mdio0_phy15: mdio_phy3@1f {
-+ reg = <0x1f>;
-+ phy-connection-type = "sgmii";
-+ };
- };
- };
- };
- };
-
--&i2c1 {
-- status = "disabled";
-+/* Update DPMAC connections to external PHYs, under SerDes 0x2a_0x49. */
-+&dpmac9 {
-+ phy-handle = <&mdio0_phy12>;
- };
--
--&i2c2 {
-- status = "disabled";
--};
--
--&i2c3 {
-- status = "disabled";
--};
--
--&dspi {
-- status = "okay";
-- dflash0: n25q128a {
-- #address-cells = <1>;
-- #size-cells = <1>;
-- compatible = "st,m25p80";
-- spi-max-frequency = <3000000>;
-- reg = <0>;
-- };
-- dflash1: sst25wf040b {
-- #address-cells = <1>;
-- #size-cells = <1>;
-- compatible = "st,m25p80";
-- spi-max-frequency = <3000000>;
-- reg = <1>;
-- };
-- dflash2: en25s64 {
-- #address-cells = <1>;
-- #size-cells = <1>;
-- compatible = "st,m25p80";
-- spi-max-frequency = <3000000>;
-- reg = <2>;
-- };
-+&dpmac10 {
-+ phy-handle = <&mdio0_phy13>;
- };
--
--&qspi {
-- status = "okay";
-- flash0: s25fl256s1@0 {
-- #address-cells = <1>;
-- #size-cells = <1>;
-- compatible = "st,m25p80";
-- spi-max-frequency = <20000000>;
-- reg = <0>;
-- };
-- flash2: s25fl256s1@2 {
-- #address-cells = <1>;
-- #size-cells = <1>;
-- compatible = "st,m25p80";
-- spi-max-frequency = <20000000>;
-- reg = <0>;
-- };
-+&dpmac11 {
-+ phy-handle = <&mdio0_phy14>;
- };
--
--&sata0 {
-- status = "okay";
--};
--
--&sata1 {
-- status = "okay";
--};
--
--&usb0 {
-- status = "okay";
--};
--
--&usb1 {
-- status = "okay";
-+&dpmac12 {
-+ phy-handle = <&mdio0_phy15>;
- };
---- a/arch/arm64/boot/dts/freescale/fsl-ls2080a-rdb.dts
-+++ b/arch/arm64/boot/dts/freescale/fsl-ls2080a-rdb.dts
-@@ -1,170 +1,105 @@
-+// SPDX-License-Identifier: (GPL-2.0+ OR MIT)
- /*
- * Device Tree file for Freescale LS2080a RDB Board.
- *
-- * Copyright (C) 2015, Freescale Semiconductor
-+ * Copyright 2016 Freescale Semiconductor, Inc.
-+ * Copyright 2017 NXP
- *
-+ * Abhimanyu Saini <abhimanyu.saini@nxp.com>
- * Bhupesh Sharma <bhupesh.sharma@freescale.com>
- *
-- * This file is dual-licensed: you can use it either under the terms
-- * of the GPLv2 or the X11 license, at your option. Note that this dual
-- * licensing only applies to this file, and not this project as a
-- * whole.
-- *
-- * a) This library is free software; you can redistribute it and/or
-- * modify it under the terms of the GNU General Public License as
-- * published by the Free Software Foundation; either version 2 of the
-- * License, or (at your option) any later version.
-- *
-- * This library is distributed in the hope that it will be useful,
-- * but WITHOUT ANY WARRANTY; without even the implied warranty of
-- * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
-- * GNU General Public License for more details.
-- *
-- * Or, alternatively,
-- *
-- * b) Permission is hereby granted, free of charge, to any person
-- * obtaining a copy of this software and associated documentation
-- * files (the "Software"), to deal in the Software without
-- * restriction, including without limitation the rights to use,
-- * copy, modify, merge, publish, distribute, sublicense, and/or
-- * sell copies of the Software, and to permit persons to whom the
-- * Software is furnished to do so, subject to the following
-- * conditions:
-- *
-- * The above copyright notice and this permission notice shall be
-- * included in all copies or substantial portions of the Software.
-- *
-- * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
-- * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES
-- * OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
-- * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT
-- * HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY,
-- * WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
-- * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
-- * OTHER DEALINGS IN THE SOFTWARE.
- */
-
- /dts-v1/;
-
--/include/ "fsl-ls2080a.dtsi"
-+#include "fsl-ls2080a.dtsi"
-+#include "fsl-ls208xa-rdb.dtsi"
-
- / {
- model = "Freescale Layerscape 2080a RDB Board";
- compatible = "fsl,ls2080a-rdb", "fsl,ls2080a";
-
-- aliases {
-- serial0 = &serial0;
-- serial1 = &serial1;
-- };
--
- chosen {
- stdout-path = "serial1:115200n8";
- };
- };
-
--&esdhc {
-- status = "okay";
--};
--
--&ifc {
-- status = "okay";
-- #address-cells = <2>;
-- #size-cells = <1>;
-- ranges = <0x0 0x0 0x5 0x80000000 0x08000000
-- 0x2 0x0 0x5 0x30000000 0x00010000
-- 0x3 0x0 0x5 0x20000000 0x00010000>;
--
-- nor@0,0 {
-- #address-cells = <1>;
-- #size-cells = <1>;
-- compatible = "cfi-flash";
-- reg = <0x0 0x0 0x8000000>;
-- bank-width = <2>;
-- device-width = <1>;
-- };
--
-- nand@2,0 {
-- compatible = "fsl,ifc-nand";
-- reg = <0x2 0x0 0x10000>;
-- };
--
-- cpld@3,0 {
-- reg = <0x3 0x0 0x10000>;
-- compatible = "fsl,ls2080aqds-fpga", "fsl,fpga-qixis";
-- };
--
--};
--
--&i2c0 {
-- status = "okay";
-- pca9547@75 {
-- compatible = "nxp,pca9547";
-- reg = <0x75>;
-- #address-cells = <1>;
-- #size-cells = <0>;
-- status = "disabled";
-- i2c@1 {
-- #address-cells = <1>;
-- #size-cells = <0>;
-- reg = <0x01>;
-- rtc@68 {
-- compatible = "dallas,ds3232";
-- reg = <0x68>;
-- };
-- };
--
-- i2c@3 {
-- #address-cells = <1>;
-- #size-cells = <0>;
-- reg = <0x3>;
--
-- adt7481@4c {
-- compatible = "adi,adt7461";
-- reg = <0x4c>;
-- };
-- };
-- };
--};
--
--&i2c1 {
-- status = "disabled";
--};
--
--&i2c2 {
-- status = "disabled";
--};
--
--&i2c3 {
-+&emdio1 {
- status = "disabled";
-+ /* CS4340 PHYs */
-+ mdio1_phy1: emdio1_phy@1 {
-+ reg = <0x10>;
-+ phy-connection-type = "xfi";
-+ };
-+ mdio1_phy2: emdio1_phy@2 {
-+ reg = <0x11>;
-+ phy-connection-type = "xfi";
-+ };
-+ mdio1_phy3: emdio1_phy@3 {
-+ reg = <0x12>;
-+ phy-connection-type = "xfi";
-+ };
-+ mdio1_phy4: emdio1_phy@4 {
-+ reg = <0x13>;
-+ phy-connection-type = "xfi";
-+ };
- };
-
--&dspi {
-- status = "okay";
-- dflash0: n25q512a {
-- #address-cells = <1>;
-- #size-cells = <1>;
-- compatible = "st,m25p80";
-- spi-max-frequency = <3000000>;
-- reg = <0>;
-+&emdio2 {
-+ /* AQR405 PHYs */
-+ mdio2_phy1: emdio2_phy@1 {
-+ compatible = "ethernet-phy-ieee802.3-c45";
-+ interrupts = <0 1 0x4>; /* Level high type */
-+ reg = <0x0>;
-+ phy-connection-type = "xfi";
-+ };
-+ mdio2_phy2: emdio2_phy@2 {
-+ compatible = "ethernet-phy-ieee802.3-c45";
-+ interrupts = <0 2 0x4>; /* Level high type */
-+ reg = <0x1>;
-+ phy-connection-type = "xfi";
-+ };
-+ mdio2_phy3: emdio2_phy@3 {
-+ compatible = "ethernet-phy-ieee802.3-c45";
-+ interrupts = <0 4 0x4>; /* Level high type */
-+ reg = <0x2>;
-+ phy-connection-type = "xfi";
-+ };
-+ mdio2_phy4: emdio2_phy@4 {
-+ compatible = "ethernet-phy-ieee802.3-c45";
-+ interrupts = <0 5 0x4>; /* Level high type */
-+ reg = <0x3>;
-+ phy-connection-type = "xfi";
- };
- };
-
--&qspi {
-- status = "disabled";
--};
-+/* Update DPMAC connections to external PHYs, under the assumption of
-+ * SerDes 0x2a_0x41. This is currently the only SerDes supported on the board.
-+ */
-+/* Leave Cortina nodes commented out until driver is integrated
-+ *&dpmac1 {
-+ * phy-handle = <&mdio1_phy1>;
-+ *};
-+ *&dpmac2 {
-+ * phy-handle = <&mdio1_phy2>;
-+ *};
-+ *&dpmac3 {
-+ * phy-handle = <&mdio1_phy3>;
-+ *};
-+ *&dpmac4 {
-+ * phy-handle = <&mdio1_phy4>;
-+ *};
-+ */
-
--&sata0 {
-- status = "okay";
-+&dpmac5 {
-+ phy-handle = <&mdio2_phy1>;
- };
--
--&sata1 {
-- status = "okay";
-+&dpmac6 {
-+ phy-handle = <&mdio2_phy2>;
- };
--
--&usb0 {
-- status = "okay";
-+&dpmac7 {
-+ phy-handle = <&mdio2_phy3>;
- };
--
--&usb1 {
-- status = "okay";
-+&dpmac8 {
-+ phy-handle = <&mdio2_phy4>;
- };
---- a/arch/arm64/boot/dts/freescale/fsl-ls2080a-simu.dts
-+++ b/arch/arm64/boot/dts/freescale/fsl-ls2080a-simu.dts
-@@ -1,62 +1,21 @@
-+// SPDX-License-Identifier: (GPL-2.0+ OR MIT)
- /*
- * Device Tree file for Freescale LS2080a software Simulator model
- *
-- * Copyright (C) 2014-2015, Freescale Semiconductor
-+ * Copyright 2014-2015 Freescale Semiconductor, Inc.
- *
- * Bhupesh Sharma <bhupesh.sharma@freescale.com>
- *
-- * This file is dual-licensed: you can use it either under the terms
-- * of the GPL or the X11 license, at your option. Note that this dual
-- * licensing only applies to this file, and not this project as a
-- * whole.
-- *
-- * a) This library is free software; you can redistribute it and/or
-- * modify it under the terms of the GNU General Public License as
-- * published by the Free Software Foundation; either version 2 of the
-- * License, or (at your option) any later version.
-- *
-- * This library is distributed in the hope that it will be useful,
-- * but WITHOUT ANY WARRANTY; without even the implied warranty of
-- * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
-- * GNU General Public License for more details.
-- *
-- * Or, alternatively,
-- *
-- * b) Permission is hereby granted, free of charge, to any person
-- * obtaining a copy of this software and associated documentation
-- * files (the "Software"), to deal in the Software without
-- * restriction, including without limitation the rights to use,
-- * copy, modify, merge, publish, distribute, sublicense, and/or
-- * sell copies of the Software, and to permit persons to whom the
-- * Software is furnished to do so, subject to the following
-- * conditions:
-- *
-- * The above copyright notice and this permission notice shall be
-- * included in all copies or substantial portions of the Software.
-- *
-- * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
-- * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES
-- * OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
-- * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT
-- * HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY,
-- * WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
-- * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
-- * OTHER DEALINGS IN THE SOFTWARE.
- */
-
- /dts-v1/;
-
--/include/ "fsl-ls2080a.dtsi"
-+#include "fsl-ls2080a.dtsi"
-
- / {
- model = "Freescale Layerscape 2080a software Simulator model";
- compatible = "fsl,ls2080a-simu", "fsl,ls2080a";
-
-- aliases {
-- serial0 = &serial0;
-- serial1 = &serial1;
-- };
--
- ethernet@2210000 {
- compatible = "smsc,lan91c111";
- reg = <0x0 0x2210000 0x0 0x100>;
---- a/arch/arm64/boot/dts/freescale/fsl-ls2080a.dtsi
-+++ b/arch/arm64/boot/dts/freescale/fsl-ls2080a.dtsi
-@@ -1,739 +1,144 @@
-+// SPDX-License-Identifier: (GPL-2.0+ OR MIT)
- /*
- * Device Tree Include file for Freescale Layerscape-2080A family SoC.
- *
-- * Copyright (C) 2014-2015, Freescale Semiconductor
-+ * Copyright 2014-2016 Freescale Semiconductor, Inc.
- *
-+ * Abhimanyu Saini <abhimanyu.saini@nxp.com>
- * Bhupesh Sharma <bhupesh.sharma@freescale.com>
- *
-- * This file is dual-licensed: you can use it either under the terms
-- * of the GPLv2 or the X11 license, at your option. Note that this dual
-- * licensing only applies to this file, and not this project as a
-- * whole.
-- *
-- * a) This library is free software; you can redistribute it and/or
-- * modify it under the terms of the GNU General Public License as
-- * published by the Free Software Foundation; either version 2 of the
-- * License, or (at your option) any later version.
-- *
-- * This library is distributed in the hope that it will be useful,
-- * but WITHOUT ANY WARRANTY; without even the implied warranty of
-- * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
-- * GNU General Public License for more details.
-- *
-- * Or, alternatively,
-- *
-- * b) Permission is hereby granted, free of charge, to any person
-- * obtaining a copy of this software and associated documentation
-- * files (the "Software"), to deal in the Software without
-- * restriction, including without limitation the rights to use,
-- * copy, modify, merge, publish, distribute, sublicense, and/or
-- * sell copies of the Software, and to permit persons to whom the
-- * Software is furnished to do so, subject to the following
-- * conditions:
-- *
-- * The above copyright notice and this permission notice shall be
-- * included in all copies or substantial portions of the Software.
-- *
-- * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
-- * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES
-- * OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
-- * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT
-- * HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY,
-- * WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
-- * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
-- * OTHER DEALINGS IN THE SOFTWARE.
- */
-
--/ {
-- compatible = "fsl,ls2080a";
-- interrupt-parent = <&gic>;
-- #address-cells = <2>;
-- #size-cells = <2>;
--
-- cpus {
-- #address-cells = <1>;
-- #size-cells = <0>;
--
-- /*
-- * We expect the enable-method for cpu's to be "psci", but this
-- * is dependent on the SoC FW, which will fill this in.
-- *
-- * Currently supported enable-method is psci v0.2
-- */
--
-- /* We have 4 clusters having 2 Cortex-A57 cores each */
-- cpu@0 {
-- device_type = "cpu";
-- compatible = "arm,cortex-a57";
-- reg = <0x0>;
-- clocks = <&clockgen 1 0>;
-- next-level-cache = <&cluster0_l2>;
-- };
--
-- cpu@1 {
-- device_type = "cpu";
-- compatible = "arm,cortex-a57";
-- reg = <0x1>;
-- clocks = <&clockgen 1 0>;
-- next-level-cache = <&cluster0_l2>;
-- };
--
-- cpu@100 {
-- device_type = "cpu";
-- compatible = "arm,cortex-a57";
-- reg = <0x100>;
-- clocks = <&clockgen 1 1>;
-- next-level-cache = <&cluster1_l2>;
-- };
--
-- cpu@101 {
-- device_type = "cpu";
-- compatible = "arm,cortex-a57";
-- reg = <0x101>;
-- clocks = <&clockgen 1 1>;
-- next-level-cache = <&cluster1_l2>;
-- };
--
-- cpu@200 {
-- device_type = "cpu";
-- compatible = "arm,cortex-a57";
-- reg = <0x200>;
-- clocks = <&clockgen 1 2>;
-- next-level-cache = <&cluster2_l2>;
-- };
--
-- cpu@201 {
-- device_type = "cpu";
-- compatible = "arm,cortex-a57";
-- reg = <0x201>;
-- clocks = <&clockgen 1 2>;
-- next-level-cache = <&cluster2_l2>;
-- };
--
-- cpu@300 {
-- device_type = "cpu";
-- compatible = "arm,cortex-a57";
-- reg = <0x300>;
-- clocks = <&clockgen 1 3>;
-- next-level-cache = <&cluster3_l2>;
-- };
--
-- cpu@301 {
-- device_type = "cpu";
-- compatible = "arm,cortex-a57";
-- reg = <0x301>;
-- clocks = <&clockgen 1 3>;
-- next-level-cache = <&cluster3_l2>;
-- };
--
-- cluster0_l2: l2-cache0 {
-- compatible = "cache";
-- };
--
-- cluster1_l2: l2-cache1 {
-- compatible = "cache";
-- };
--
-- cluster2_l2: l2-cache2 {
-- compatible = "cache";
-- };
--
-- cluster3_l2: l2-cache3 {
-- compatible = "cache";
-- };
-- };
--
-- memory@80000000 {
-- device_type = "memory";
-- reg = <0x00000000 0x80000000 0 0x80000000>;
-- /* DRAM space - 1, size : 2 GB DRAM */
-- };
--
-- sysclk: sysclk {
-- compatible = "fixed-clock";
-- #clock-cells = <0>;
-- clock-frequency = <100000000>;
-- clock-output-names = "sysclk";
-- };
--
-- gic: interrupt-controller@6000000 {
-- compatible = "arm,gic-v3";
-- reg = <0x0 0x06000000 0 0x10000>, /* GIC Dist */
-- <0x0 0x06100000 0 0x100000>, /* GICR (RD_base + SGI_base) */
-- <0x0 0x0c0c0000 0 0x2000>, /* GICC */
-- <0x0 0x0c0d0000 0 0x1000>, /* GICH */
-- <0x0 0x0c0e0000 0 0x20000>; /* GICV */
-- #interrupt-cells = <3>;
-- #address-cells = <2>;
-- #size-cells = <2>;
-- ranges;
-- interrupt-controller;
-- interrupts = <1 9 0x4>;
--
-- its: gic-its@6020000 {
-- compatible = "arm,gic-v3-its";
-- msi-controller;
-- reg = <0x0 0x6020000 0 0x20000>;
-- };
-- };
--
-- rstcr: syscon@1e60000 {
-- compatible = "fsl,ls2080a-rstcr", "syscon";
-- reg = <0x0 0x1e60000 0x0 0x4>;
-- };
--
-- reboot {
-- compatible ="syscon-reboot";
-- regmap = <&rstcr>;
-- offset = <0x0>;
-- mask = <0x2>;
-- };
--
-- timer {
-- compatible = "arm,armv8-timer";
-- interrupts = <1 13 4>, /* Physical Secure PPI, active-low */
-- <1 14 4>, /* Physical Non-Secure PPI, active-low */
-- <1 11 4>, /* Virtual PPI, active-low */
-- <1 10 4>; /* Hypervisor PPI, active-low */
-- fsl,erratum-a008585;
-- };
--
-- pmu {
-- compatible = "arm,armv8-pmuv3";
-- interrupts = <1 7 0x8>; /* PMU PPI, Level low type */
-- };
--
-- soc {
-- compatible = "simple-bus";
-- #address-cells = <2>;
-- #size-cells = <2>;
-- ranges;
--
-- clockgen: clocking@1300000 {
-- compatible = "fsl,ls2080a-clockgen";
-- reg = <0 0x1300000 0 0xa0000>;
-- #clock-cells = <2>;
-- clocks = <&sysclk>;
-- };
--
-- serial0: serial@21c0500 {
-- compatible = "fsl,ns16550", "ns16550a";
-- reg = <0x0 0x21c0500 0x0 0x100>;
-- clocks = <&clockgen 4 3>;
-- interrupts = <0 32 0x4>; /* Level high type */
-- };
--
-- serial1: serial@21c0600 {
-- compatible = "fsl,ns16550", "ns16550a";
-- reg = <0x0 0x21c0600 0x0 0x100>;
-- clocks = <&clockgen 4 3>;
-- interrupts = <0 32 0x4>; /* Level high type */
-- };
--
-- cluster1_core0_watchdog: wdt@c000000 {
-- compatible = "arm,sp805-wdt", "arm,primecell";
-- reg = <0x0 0xc000000 0x0 0x1000>;
-- clocks = <&clockgen 4 3>, <&clockgen 4 3>;
-- clock-names = "apb_pclk", "wdog_clk";
-- };
--
-- cluster1_core1_watchdog: wdt@c010000 {
-- compatible = "arm,sp805-wdt", "arm,primecell";
-- reg = <0x0 0xc010000 0x0 0x1000>;
-- clocks = <&clockgen 4 3>, <&clockgen 4 3>;
-- clock-names = "apb_pclk", "wdog_clk";
-- };
--
-- cluster2_core0_watchdog: wdt@c100000 {
-- compatible = "arm,sp805-wdt", "arm,primecell";
-- reg = <0x0 0xc100000 0x0 0x1000>;
-- clocks = <&clockgen 4 3>, <&clockgen 4 3>;
-- clock-names = "apb_pclk", "wdog_clk";
-- };
--
-- cluster2_core1_watchdog: wdt@c110000 {
-- compatible = "arm,sp805-wdt", "arm,primecell";
-- reg = <0x0 0xc110000 0x0 0x1000>;
-- clocks = <&clockgen 4 3>, <&clockgen 4 3>;
-- clock-names = "apb_pclk", "wdog_clk";
-- };
--
-- cluster3_core0_watchdog: wdt@c200000 {
-- compatible = "arm,sp805-wdt", "arm,primecell";
-- reg = <0x0 0xc200000 0x0 0x1000>;
-- clocks = <&clockgen 4 3>, <&clockgen 4 3>;
-- clock-names = "apb_pclk", "wdog_clk";
-- };
--
-- cluster3_core1_watchdog: wdt@c210000 {
-- compatible = "arm,sp805-wdt", "arm,primecell";
-- reg = <0x0 0xc210000 0x0 0x1000>;
-- clocks = <&clockgen 4 3>, <&clockgen 4 3>;
-- clock-names = "apb_pclk", "wdog_clk";
-- };
--
-- cluster4_core0_watchdog: wdt@c300000 {
-- compatible = "arm,sp805-wdt", "arm,primecell";
-- reg = <0x0 0xc300000 0x0 0x1000>;
-- clocks = <&clockgen 4 3>, <&clockgen 4 3>;
-- clock-names = "apb_pclk", "wdog_clk";
-- };
--
-- cluster4_core1_watchdog: wdt@c310000 {
-- compatible = "arm,sp805-wdt", "arm,primecell";
-- reg = <0x0 0xc310000 0x0 0x1000>;
-- clocks = <&clockgen 4 3>, <&clockgen 4 3>;
-- clock-names = "apb_pclk", "wdog_clk";
-- };
--
-- fsl_mc: fsl-mc@80c000000 {
-- compatible = "fsl,qoriq-mc";
-- reg = <0x00000008 0x0c000000 0 0x40>, /* MC portal base */
-- <0x00000000 0x08340000 0 0x40000>; /* MC control reg */
-- msi-parent = <&its>;
-- #address-cells = <3>;
-- #size-cells = <1>;
--
-- /*
-- * Region type 0x0 - MC portals
-- * Region type 0x1 - QBMAN portals
-- */
-- ranges = <0x0 0x0 0x0 0x8 0x0c000000 0x4000000
-- 0x1 0x0 0x0 0x8 0x18000000 0x8000000>;
--
-- /*
-- * Define the maximum number of MACs present on the SoC.
-- */
-- dpmacs {
-- #address-cells = <1>;
-- #size-cells = <0>;
--
-- dpmac1: dpmac@1 {
-- compatible = "fsl,qoriq-mc-dpmac";
-- reg = <0x1>;
-- };
--
-- dpmac2: dpmac@2 {
-- compatible = "fsl,qoriq-mc-dpmac";
-- reg = <0x2>;
-- };
--
-- dpmac3: dpmac@3 {
-- compatible = "fsl,qoriq-mc-dpmac";
-- reg = <0x3>;
-- };
--
-- dpmac4: dpmac@4 {
-- compatible = "fsl,qoriq-mc-dpmac";
-- reg = <0x4>;
-- };
--
-- dpmac5: dpmac@5 {
-- compatible = "fsl,qoriq-mc-dpmac";
-- reg = <0x5>;
-- };
--
-- dpmac6: dpmac@6 {
-- compatible = "fsl,qoriq-mc-dpmac";
-- reg = <0x6>;
-- };
--
-- dpmac7: dpmac@7 {
-- compatible = "fsl,qoriq-mc-dpmac";
-- reg = <0x7>;
-- };
--
-- dpmac8: dpmac@8 {
-- compatible = "fsl,qoriq-mc-dpmac";
-- reg = <0x8>;
-- };
--
-- dpmac9: dpmac@9 {
-- compatible = "fsl,qoriq-mc-dpmac";
-- reg = <0x9>;
-- };
--
-- dpmac10: dpmac@a {
-- compatible = "fsl,qoriq-mc-dpmac";
-- reg = <0xa>;
-- };
--
-- dpmac11: dpmac@b {
-- compatible = "fsl,qoriq-mc-dpmac";
-- reg = <0xb>;
-- };
--
-- dpmac12: dpmac@c {
-- compatible = "fsl,qoriq-mc-dpmac";
-- reg = <0xc>;
-- };
--
-- dpmac13: dpmac@d {
-- compatible = "fsl,qoriq-mc-dpmac";
-- reg = <0xd>;
-- };
--
-- dpmac14: dpmac@e {
-- compatible = "fsl,qoriq-mc-dpmac";
-- reg = <0xe>;
-- };
--
-- dpmac15: dpmac@f {
-- compatible = "fsl,qoriq-mc-dpmac";
-- reg = <0xf>;
-- };
--
-- dpmac16: dpmac@10 {
-- compatible = "fsl,qoriq-mc-dpmac";
-- reg = <0x10>;
-- };
-- };
-- };
--
-- smmu: iommu@5000000 {
-- compatible = "arm,mmu-500";
-- reg = <0 0x5000000 0 0x800000>;
-- #global-interrupts = <12>;
-- interrupts = <0 13 4>, /* global secure fault */
-- <0 14 4>, /* combined secure interrupt */
-- <0 15 4>, /* global non-secure fault */
-- <0 16 4>, /* combined non-secure interrupt */
-- /* performance counter interrupts 0-7 */
-- <0 211 4>, <0 212 4>,
-- <0 213 4>, <0 214 4>,
-- <0 215 4>, <0 216 4>,
-- <0 217 4>, <0 218 4>,
-- /* per context interrupt, 64 interrupts */
-- <0 146 4>, <0 147 4>,
-- <0 148 4>, <0 149 4>,
-- <0 150 4>, <0 151 4>,
-- <0 152 4>, <0 153 4>,
-- <0 154 4>, <0 155 4>,
-- <0 156 4>, <0 157 4>,
-- <0 158 4>, <0 159 4>,
-- <0 160 4>, <0 161 4>,
-- <0 162 4>, <0 163 4>,
-- <0 164 4>, <0 165 4>,
-- <0 166 4>, <0 167 4>,
-- <0 168 4>, <0 169 4>,
-- <0 170 4>, <0 171 4>,
-- <0 172 4>, <0 173 4>,
-- <0 174 4>, <0 175 4>,
-- <0 176 4>, <0 177 4>,
-- <0 178 4>, <0 179 4>,
-- <0 180 4>, <0 181 4>,
-- <0 182 4>, <0 183 4>,
-- <0 184 4>, <0 185 4>,
-- <0 186 4>, <0 187 4>,
-- <0 188 4>, <0 189 4>,
-- <0 190 4>, <0 191 4>,
-- <0 192 4>, <0 193 4>,
-- <0 194 4>, <0 195 4>,
-- <0 196 4>, <0 197 4>,
-- <0 198 4>, <0 199 4>,
-- <0 200 4>, <0 201 4>,
-- <0 202 4>, <0 203 4>,
-- <0 204 4>, <0 205 4>,
-- <0 206 4>, <0 207 4>,
-- <0 208 4>, <0 209 4>;
-- mmu-masters = <&fsl_mc 0x300 0>;
-- };
--
-- dspi: dspi@2100000 {
-- status = "disabled";
-- compatible = "fsl,ls2080a-dspi", "fsl,ls2085a-dspi";
-- #address-cells = <1>;
-- #size-cells = <0>;
-- reg = <0x0 0x2100000 0x0 0x10000>;
-- interrupts = <0 26 0x4>; /* Level high type */
-- clocks = <&clockgen 4 3>;
-- clock-names = "dspi";
-- spi-num-chipselects = <5>;
-- bus-num = <0>;
-- };
--
-- esdhc: esdhc@2140000 {
-- status = "disabled";
-- compatible = "fsl,ls2080a-esdhc", "fsl,esdhc";
-- reg = <0x0 0x2140000 0x0 0x10000>;
-- interrupts = <0 28 0x4>; /* Level high type */
-- clock-frequency = <0>; /* Updated by bootloader */
-- voltage-ranges = <1800 1800 3300 3300>;
-- sdhci,auto-cmd12;
-- little-endian;
-- bus-width = <4>;
-- };
--
-- gpio0: gpio@2300000 {
-- compatible = "fsl,ls2080a-gpio", "fsl,qoriq-gpio";
-- reg = <0x0 0x2300000 0x0 0x10000>;
-- interrupts = <0 36 0x4>; /* Level high type */
-- gpio-controller;
-- little-endian;
-- #gpio-cells = <2>;
-- interrupt-controller;
-- #interrupt-cells = <2>;
-- };
--
-- gpio1: gpio@2310000 {
-- compatible = "fsl,ls2080a-gpio", "fsl,qoriq-gpio";
-- reg = <0x0 0x2310000 0x0 0x10000>;
-- interrupts = <0 36 0x4>; /* Level high type */
-- gpio-controller;
-- little-endian;
-- #gpio-cells = <2>;
-- interrupt-controller;
-- #interrupt-cells = <2>;
-- };
--
-- gpio2: gpio@2320000 {
-- compatible = "fsl,ls2080a-gpio", "fsl,qoriq-gpio";
-- reg = <0x0 0x2320000 0x0 0x10000>;
-- interrupts = <0 37 0x4>; /* Level high type */
-- gpio-controller;
-- little-endian;
-- #gpio-cells = <2>;
-- interrupt-controller;
-- #interrupt-cells = <2>;
-- };
--
-- gpio3: gpio@2330000 {
-- compatible = "fsl,ls2080a-gpio", "fsl,qoriq-gpio";
-- reg = <0x0 0x2330000 0x0 0x10000>;
-- interrupts = <0 37 0x4>; /* Level high type */
-- gpio-controller;
-- little-endian;
-- #gpio-cells = <2>;
-- interrupt-controller;
-- #interrupt-cells = <2>;
-- };
--
-- i2c0: i2c@2000000 {
-- status = "disabled";
-- compatible = "fsl,vf610-i2c";
-- #address-cells = <1>;
-- #size-cells = <0>;
-- reg = <0x0 0x2000000 0x0 0x10000>;
-- interrupts = <0 34 0x4>; /* Level high type */
-- clock-names = "i2c";
-- clocks = <&clockgen 4 3>;
-- };
--
-- i2c1: i2c@2010000 {
-- status = "disabled";
-- compatible = "fsl,vf610-i2c";
-- #address-cells = <1>;
-- #size-cells = <0>;
-- reg = <0x0 0x2010000 0x0 0x10000>;
-- interrupts = <0 34 0x4>; /* Level high type */
-- clock-names = "i2c";
-- clocks = <&clockgen 4 3>;
-- };
--
-- i2c2: i2c@2020000 {
-- status = "disabled";
-- compatible = "fsl,vf610-i2c";
-- #address-cells = <1>;
-- #size-cells = <0>;
-- reg = <0x0 0x2020000 0x0 0x10000>;
-- interrupts = <0 35 0x4>; /* Level high type */
-- clock-names = "i2c";
-- clocks = <&clockgen 4 3>;
-- };
--
-- i2c3: i2c@2030000 {
-- status = "disabled";
-- compatible = "fsl,vf610-i2c";
-- #address-cells = <1>;
-- #size-cells = <0>;
-- reg = <0x0 0x2030000 0x0 0x10000>;
-- interrupts = <0 35 0x4>; /* Level high type */
-- clock-names = "i2c";
-- clocks = <&clockgen 4 3>;
-- };
--
-- ifc: ifc@2240000 {
-- compatible = "fsl,ifc", "simple-bus";
-- reg = <0x0 0x2240000 0x0 0x20000>;
-- interrupts = <0 21 0x4>; /* Level high type */
-- little-endian;
-- #address-cells = <2>;
-- #size-cells = <1>;
--
-- ranges = <0 0 0x5 0x80000000 0x08000000
-- 2 0 0x5 0x30000000 0x00010000
-- 3 0 0x5 0x20000000 0x00010000>;
-- };
--
-- qspi: quadspi@20c0000 {
-- status = "disabled";
-- compatible = "fsl,ls2080a-qspi", "fsl,ls1021a-qspi";
-- #address-cells = <1>;
-- #size-cells = <0>;
-- reg = <0x0 0x20c0000 0x0 0x10000>,
-- <0x0 0x20000000 0x0 0x10000000>;
-- reg-names = "QuadSPI", "QuadSPI-memory";
-- interrupts = <0 25 0x4>; /* Level high type */
-- clocks = <&clockgen 4 3>, <&clockgen 4 3>;
-- clock-names = "qspi_en", "qspi";
-- };
--
-- pcie@3400000 {
-- compatible = "fsl,ls2080a-pcie", "fsl,ls2085a-pcie",
-- "snps,dw-pcie";
-- reg = <0x00 0x03400000 0x0 0x00100000 /* controller registers */
-- 0x10 0x00000000 0x0 0x00002000>; /* configuration space */
-- reg-names = "regs", "config";
-- interrupts = <0 108 0x4>; /* Level high type */
-- interrupt-names = "intr";
-- #address-cells = <3>;
-- #size-cells = <2>;
-- device_type = "pci";
-- dma-coherent;
-- num-lanes = <4>;
-- bus-range = <0x0 0xff>;
-- ranges = <0x81000000 0x0 0x00000000 0x10 0x00010000 0x0 0x00010000 /* downstream I/O */
-- 0x82000000 0x0 0x40000000 0x10 0x40000000 0x0 0x40000000>; /* non-prefetchable memory */
-- msi-parent = <&its>;
-- #interrupt-cells = <1>;
-- interrupt-map-mask = <0 0 0 7>;
-- interrupt-map = <0000 0 0 1 &gic 0 0 0 109 4>,
-- <0000 0 0 2 &gic 0 0 0 110 4>,
-- <0000 0 0 3 &gic 0 0 0 111 4>,
-- <0000 0 0 4 &gic 0 0 0 112 4>;
-- };
--
-- pcie@3500000 {
-- compatible = "fsl,ls2080a-pcie", "fsl,ls2085a-pcie",
-- "snps,dw-pcie";
-- reg = <0x00 0x03500000 0x0 0x00100000 /* controller registers */
-- 0x12 0x00000000 0x0 0x00002000>; /* configuration space */
-- reg-names = "regs", "config";
-- interrupts = <0 113 0x4>; /* Level high type */
-- interrupt-names = "intr";
-- #address-cells = <3>;
-- #size-cells = <2>;
-- device_type = "pci";
-- dma-coherent;
-- num-lanes = <4>;
-- bus-range = <0x0 0xff>;
-- ranges = <0x81000000 0x0 0x00000000 0x12 0x00010000 0x0 0x00010000 /* downstream I/O */
-- 0x82000000 0x0 0x40000000 0x12 0x40000000 0x0 0x40000000>; /* non-prefetchable memory */
-- msi-parent = <&its>;
-- #interrupt-cells = <1>;
-- interrupt-map-mask = <0 0 0 7>;
-- interrupt-map = <0000 0 0 1 &gic 0 0 0 114 4>,
-- <0000 0 0 2 &gic 0 0 0 115 4>,
-- <0000 0 0 3 &gic 0 0 0 116 4>,
-- <0000 0 0 4 &gic 0 0 0 117 4>;
-- };
--
-- pcie@3600000 {
-- compatible = "fsl,ls2080a-pcie", "fsl,ls2085a-pcie",
-- "snps,dw-pcie";
-- reg = <0x00 0x03600000 0x0 0x00100000 /* controller registers */
-- 0x14 0x00000000 0x0 0x00002000>; /* configuration space */
-- reg-names = "regs", "config";
-- interrupts = <0 118 0x4>; /* Level high type */
-- interrupt-names = "intr";
-- #address-cells = <3>;
-- #size-cells = <2>;
-- device_type = "pci";
-- dma-coherent;
-- num-lanes = <8>;
-- bus-range = <0x0 0xff>;
-- ranges = <0x81000000 0x0 0x00000000 0x14 0x00010000 0x0 0x00010000 /* downstream I/O */
-- 0x82000000 0x0 0x40000000 0x14 0x40000000 0x0 0x40000000>; /* non-prefetchable memory */
-- msi-parent = <&its>;
-- #interrupt-cells = <1>;
-- interrupt-map-mask = <0 0 0 7>;
-- interrupt-map = <0000 0 0 1 &gic 0 0 0 119 4>,
-- <0000 0 0 2 &gic 0 0 0 120 4>,
-- <0000 0 0 3 &gic 0 0 0 121 4>,
-- <0000 0 0 4 &gic 0 0 0 122 4>;
-- };
--
-- pcie@3700000 {
-- compatible = "fsl,ls2080a-pcie", "fsl,ls2085a-pcie",
-- "snps,dw-pcie";
-- reg = <0x00 0x03700000 0x0 0x00100000 /* controller registers */
-- 0x16 0x00000000 0x0 0x00002000>; /* configuration space */
-- reg-names = "regs", "config";
-- interrupts = <0 123 0x4>; /* Level high type */
-- interrupt-names = "intr";
-- #address-cells = <3>;
-- #size-cells = <2>;
-- device_type = "pci";
-- dma-coherent;
-- num-lanes = <4>;
-- bus-range = <0x0 0xff>;
-- ranges = <0x81000000 0x0 0x00000000 0x16 0x00010000 0x0 0x00010000 /* downstream I/O */
-- 0x82000000 0x0 0x40000000 0x16 0x40000000 0x0 0x40000000>; /* non-prefetchable memory */
-- msi-parent = <&its>;
-- #interrupt-cells = <1>;
-- interrupt-map-mask = <0 0 0 7>;
-- interrupt-map = <0000 0 0 1 &gic 0 0 0 124 4>,
-- <0000 0 0 2 &gic 0 0 0 125 4>,
-- <0000 0 0 3 &gic 0 0 0 126 4>,
-- <0000 0 0 4 &gic 0 0 0 127 4>;
-- };
--
-- sata0: sata@3200000 {
-- status = "disabled";
-- compatible = "fsl,ls2080a-ahci";
-- reg = <0x0 0x3200000 0x0 0x10000>;
-- interrupts = <0 133 0x4>; /* Level high type */
-- clocks = <&clockgen 4 3>;
-- dma-coherent;
-- };
--
-- sata1: sata@3210000 {
-- status = "disabled";
-- compatible = "fsl,ls2080a-ahci";
-- reg = <0x0 0x3210000 0x0 0x10000>;
-- interrupts = <0 136 0x4>; /* Level high type */
-- clocks = <&clockgen 4 3>;
-- dma-coherent;
-- };
--
-- usb0: usb3@3100000 {
-- status = "disabled";
-- compatible = "snps,dwc3";
-- reg = <0x0 0x3100000 0x0 0x10000>;
-- interrupts = <0 80 0x4>; /* Level high type */
-- dr_mode = "host";
-- snps,quirk-frame-length-adjustment = <0x20>;
-- snps,dis_rxdet_inp3_quirk;
-- };
--
-- usb1: usb3@3110000 {
-- status = "disabled";
-- compatible = "snps,dwc3";
-- reg = <0x0 0x3110000 0x0 0x10000>;
-- interrupts = <0 81 0x4>; /* Level high type */
-- dr_mode = "host";
-- snps,quirk-frame-length-adjustment = <0x20>;
-- snps,dis_rxdet_inp3_quirk;
-- };
--
-- ccn@4000000 {
-- compatible = "arm,ccn-504";
-- reg = <0x0 0x04000000 0x0 0x01000000>;
-- interrupts = <0 12 4>;
-- };
-- };
--
-- ddr1: memory-controller@1080000 {
-- compatible = "fsl,qoriq-memory-controller";
-- reg = <0x0 0x1080000 0x0 0x1000>;
-- interrupts = <0 17 0x4>;
-- little-endian;
-- };
--
-- ddr2: memory-controller@1090000 {
-- compatible = "fsl,qoriq-memory-controller";
-- reg = <0x0 0x1090000 0x0 0x1000>;
-- interrupts = <0 18 0x4>;
-- little-endian;
-+#include "fsl-ls208xa.dtsi"
-+
-+&cpu {
-+ cpu0: cpu@0 {
-+ device_type = "cpu";
-+ compatible = "arm,cortex-a57";
-+ reg = <0x0>;
-+ clocks = <&clockgen 1 0>;
-+ next-level-cache = <&cluster0_l2>;
-+ #cooling-cells = <2>;
-+ };
-+
-+ cpu1: cpu@1 {
-+ device_type = "cpu";
-+ compatible = "arm,cortex-a57";
-+ reg = <0x1>;
-+ clocks = <&clockgen 1 0>;
-+ next-level-cache = <&cluster0_l2>;
-+ };
-+
-+ cpu2: cpu@100 {
-+ device_type = "cpu";
-+ compatible = "arm,cortex-a57";
-+ reg = <0x100>;
-+ clocks = <&clockgen 1 1>;
-+ next-level-cache = <&cluster1_l2>;
-+ #cooling-cells = <2>;
-+ };
-+
-+ cpu3: cpu@101 {
-+ device_type = "cpu";
-+ compatible = "arm,cortex-a57";
-+ reg = <0x101>;
-+ clocks = <&clockgen 1 1>;
-+ next-level-cache = <&cluster1_l2>;
-+ };
-+
-+ cpu4: cpu@200 {
-+ device_type = "cpu";
-+ compatible = "arm,cortex-a57";
-+ reg = <0x200>;
-+ clocks = <&clockgen 1 2>;
-+ next-level-cache = <&cluster2_l2>;
-+ #cooling-cells = <2>;
-+ };
-+
-+ cpu5: cpu@201 {
-+ device_type = "cpu";
-+ compatible = "arm,cortex-a57";
-+ reg = <0x201>;
-+ clocks = <&clockgen 1 2>;
-+ next-level-cache = <&cluster2_l2>;
-+ };
-+
-+ cpu6: cpu@300 {
-+ device_type = "cpu";
-+ compatible = "arm,cortex-a57";
-+ reg = <0x300>;
-+ clocks = <&clockgen 1 3>;
-+ next-level-cache = <&cluster3_l2>;
-+ #cooling-cells = <2>;
-+ };
-+
-+ cpu7: cpu@301 {
-+ device_type = "cpu";
-+ compatible = "arm,cortex-a57";
-+ reg = <0x301>;
-+ clocks = <&clockgen 1 3>;
-+ next-level-cache = <&cluster3_l2>;
- };
-+
-+ cluster0_l2: l2-cache0 {
-+ compatible = "cache";
-+ };
-+
-+ cluster1_l2: l2-cache1 {
-+ compatible = "cache";
-+ };
-+
-+ cluster2_l2: l2-cache2 {
-+ compatible = "cache";
-+ };
-+
-+ cluster3_l2: l2-cache3 {
-+ compatible = "cache";
-+ };
-+};
-+
-+&usb0 {
-+ snps,incr-burst-type-adjustment = <1>, <4>, <8>, <16>;
-+ snps,dma-snooping;
-+};
-+
-+&usb1 {
-+ snps,incr-burst-type-adjustment = <1>, <4>, <8>, <16>;
-+ snps,dma-snooping;
-+};
-+
-+&timer {
-+ fsl,erratum-a008585;
-+};
-+
-+&pcie1 {
-+ reg = <0x00 0x03400000 0x0 0x00100000 /* controller registers */
-+ 0x10 0x00000000 0x0 0x00002000>; /* configuration space */
-+
-+ ranges = <0x81000000 0x0 0x00000000 0x10 0x00010000 0x0 0x00010000 /* downstream I/O */
-+ 0x82000000 0x0 0x40000000 0x10 0x40000000 0x0 0x40000000>; /* non-prefetchable memory */
-+};
-+
-+&pcie2 {
-+ reg = <0x00 0x03500000 0x0 0x00100000 /* controller registers */
-+ 0x12 0x00000000 0x0 0x00002000>; /* configuration space */
-+
-+ ranges = <0x81000000 0x0 0x00000000 0x12 0x00010000 0x0 0x00010000 /* downstream I/O */
-+ 0x82000000 0x0 0x40000000 0x12 0x40000000 0x0 0x40000000>; /* non-prefetchable memory */
-+};
-+
-+&pcie3 {
-+ reg = <0x00 0x03600000 0x0 0x00100000 /* controller registers */
-+ 0x14 0x00000000 0x0 0x00002000>; /* configuration space */
-+
-+ ranges = <0x81000000 0x0 0x00000000 0x14 0x00010000 0x0 0x00010000 /* downstream I/O */
-+ 0x82000000 0x0 0x40000000 0x14 0x40000000 0x0 0x40000000>; /* non-prefetchable memory */
-+};
-+
-+&pcie4 {
-+ reg = <0x00 0x03700000 0x0 0x00100000 /* controller registers */
-+ 0x16 0x00000000 0x0 0x00002000>; /* configuration space */
-+
-+ ranges = <0x81000000 0x0 0x00000000 0x16 0x00010000 0x0 0x00010000 /* downstream I/O */
-+ 0x82000000 0x0 0x40000000 0x16 0x40000000 0x0 0x40000000>; /* non-prefetchable memory */
- };
---- /dev/null
-+++ b/arch/arm64/boot/dts/freescale/fsl-ls2081a-rdb.dts
-@@ -0,0 +1,161 @@
-+/*
-+ * Device Tree file for NXP LS2081A RDB Board.
-+ *
-+ * Copyright 2017 NXP
-+ *
-+ * Priyanka Jain <priyanka.jain@nxp.com>
-+ *
-+ * This file is dual-licensed: you can use it either under the terms
-+ * of the GPLv2 or the X11 license, at your option. Note that this dual
-+ * licensing only applies to this file, and not this project as a
-+ * whole.
-+ *
-+ * a) This library is free software; you can redistribute it and/or
-+ * modify it under the terms of the GNU General Public License as
-+ * published by the Free Software Foundation; either version 2 of the
-+ * License, or (at your option) any later version.
-+ *
-+ * This library is distributed in the hope that it will be useful,
-+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
-+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
-+ * GNU General Public License for more details.
-+ *
-+ * Or, alternatively,
-+ *
-+ * b) Permission is hereby granted, free of charge, to any person
-+ * obtaining a copy of this software and associated documentation
-+ * files (the "Software"), to deal in the Software without
-+ * restriction, including without limitation the rights to use,
-+ * copy, modify, merge, publish, distribute, sublicense, and/or
-+ * sell copies of the Software, and to permit persons to whom the
-+ * Software is furnished to do so, subject to the following
-+ * conditions:
-+ *
-+ * The above copyright notice and this permission notice shall be
-+ * included in all copies or substantial portions of the Software.
-+ *
-+ * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
-+ * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES
-+ * OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
-+ * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT
-+ * HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY,
-+ * WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
-+ * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
-+ * OTHER DEALINGS IN THE SOFTWARE.
-+ */
-+
-+/dts-v1/;
-+
-+#include "fsl-ls2088a.dtsi"
-+
-+/ {
-+ model = "NXP Layerscape 2081A RDB Board";
-+ compatible = "fsl,ls2081a-rdb", "fsl,ls2081a";
-+
-+ aliases {
-+ serial0 = &serial0;
-+ serial1 = &serial1;
-+ };
-+
-+ chosen {
-+ stdout-path = "serial1:115200n8";
-+ };
-+};
-+
-+&esdhc {
-+ status = "okay";
-+};
-+
-+&ifc {
-+ status = "disabled";
-+};
-+
-+&i2c0 {
-+ status = "okay";
-+ pca9547@75 {
-+ compatible = "nxp,pca9547";
-+ reg = <0x75>;
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ i2c@1 {
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ reg = <0x01>;
-+ rtc@51 {
-+ compatible = "nxp,pcf2129";
-+ reg = <0x51>;
-+ };
-+ };
-+
-+ i2c@2 {
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ reg = <0x02>;
-+
-+ ina220@40 {
-+ compatible = "ti,ina220";
-+ reg = <0x40>;
-+ shunt-resistor = <500>;
-+ };
-+ };
-+
-+ i2c@3 {
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ reg = <0x3>;
-+
-+ adt7481@4c {
-+ compatible = "adi,adt7461";
-+ reg = <0x4c>;
-+ };
-+ };
-+ };
-+};
-+
-+&dspi {
-+ status = "okay";
-+ dflash0: n25q512a {
-+ #address-cells = <1>;
-+ #size-cells = <1>;
-+ compatible = "st,m25p80";
-+ spi-max-frequency = <3000000>;
-+ reg = <0>;
-+ };
-+};
-+
-+&qspi {
-+ status = "okay";
-+ fsl,qspi-has-second-chip;
-+ flash0: s25fs512s@0 {
-+ #address-cells = <1>;
-+ #size-cells = <1>;
-+ compatible = "spansion,m25p80";
-+ m25p,fast-read;
-+ spi-max-frequency = <20000000>;
-+ reg = <0>;
-+ };
-+ flash1: s25fs512s@1 {
-+ #address-cells = <1>;
-+ #size-cells = <1>;
-+ compatible = "spansion,m25p80";
-+ m25p,fast-read;
-+ spi-max-frequency = <20000000>;
-+ reg = <1>;
-+ };
-+};
-+
-+&sata0 {
-+ status = "okay";
-+};
-+
-+&sata1 {
-+ status = "okay";
-+};
-+
-+&usb0 {
-+ status = "okay";
-+};
-+
-+&usb1 {
-+ status = "okay";
-+};
---- /dev/null
-+++ b/arch/arm64/boot/dts/freescale/fsl-ls2088a-qds.dts
-@@ -0,0 +1,126 @@
-+// SPDX-License-Identifier: (GPL-2.0+ OR MIT)
-+/*
-+ * Device Tree file for Freescale LS2088A QDS Board.
-+ *
-+ * Copyright 2016 Freescale Semiconductor, Inc.
-+ * Copyright 2017 NXP
-+ *
-+ * Abhimanyu Saini <abhimanyu.saini@nxp.com>
-+ *
-+ */
-+
-+/dts-v1/;
-+
-+#include "fsl-ls2088a.dtsi"
-+#include "fsl-ls208xa-qds.dtsi"
-+
-+/ {
-+ model = "Freescale Layerscape 2088A QDS Board";
-+ compatible = "fsl,ls2088a-qds", "fsl,ls2088a";
-+
-+ chosen {
-+ stdout-path = "serial0:115200n8";
-+ };
-+};
-+
-+&ifc {
-+ boardctrl: board-control@3,0 {
-+ #address-cells = <1>;
-+ #size-cells = <1>;
-+ compatible = "fsl,tetra-fpga", "fsl,fpga-qixis", "simple-bus";
-+ reg = <3 0 0x300>; /* TODO check address */
-+ ranges = <0 3 0 0x300>;
-+
-+ mdio_mux_emi1 {
-+ compatible = "mdio-mux-mmioreg", "mdio-mux";
-+ mdio-parent-bus = <&emdio1>;
-+ reg = <0x54 1>; /* BRDCFG4 */
-+ mux-mask = <0xe0>; /* EMI1_MDIO */
-+
-+ #address-cells=<1>;
-+ #size-cells = <0>;
-+
-+ /* Child MDIO buses, one for each riser card:
-+ * reg = 0x0, 0x20, 0x40, 0x60, 0x80, 0xa0.
-+ * VSC8234 PHYs on the riser cards.
-+ */
-+
-+ mdio_mux3: mdio@60 {
-+ reg = <0x60>;
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+
-+ mdio0_phy12: mdio_phy0@1c {
-+ reg = <0x1c>;
-+ phy-connection-type = "sgmii";
-+ };
-+ mdio0_phy13: mdio_phy1@1d {
-+ reg = <0x1d>;
-+ phy-connection-type = "sgmii";
-+ };
-+ mdio0_phy14: mdio_phy2@1e {
-+ reg = <0x1e>;
-+ phy-connection-type = "sgmii";
-+ };
-+ mdio0_phy15: mdio_phy3@1f {
-+ reg = <0x1f>;
-+ phy-connection-type = "sgmii";
-+ };
-+ };
-+ };
-+ };
-+};
-+
-+&pcs_mdio1 {
-+ pcs_phy1: ethernet-phy@0 {
-+ backplane-mode = "10gbase-kr";
-+ compatible = "ethernet-phy-ieee802.3-c45";
-+ reg = <0x0>;
-+ fsl,lane-handle = <&serdes1>;
-+ fsl,lane-reg = <0x9C0 0x40>;/* lane H */
-+ };
-+};
-+
-+&pcs_mdio2 {
-+ pcs_phy2: ethernet-phy@0 {
-+ backplane-mode = "10gbase-kr";
-+ compatible = "ethernet-phy-ieee802.3-c45";
-+ reg = <0x0>;
-+ fsl,lane-handle = <&serdes1>;
-+ fsl,lane-reg = <0x980 0x40>;/* lane G */
-+ };
-+};
-+
-+&pcs_mdio3 {
-+ pcs_phy3: ethernet-phy@0 {
-+ backplane-mode = "10gbase-kr";
-+ compatible = "ethernet-phy-ieee802.3-c45";
-+ reg = <0x0>;
-+ fsl,lane-handle = <&serdes1>;
-+ fsl,lane-reg = <0x940 0x40>;/* lane F */
-+ };
-+};
-+
-+&pcs_mdio4 {
-+ pcs_phy4: ethernet-phy@0 {
-+ backplane-mode = "10gbase-kr";
-+ compatible = "ethernet-phy-ieee802.3-c45";
-+ reg = <0x0>;
-+ fsl,lane-handle = <&serdes1>;
-+ fsl,lane-reg = <0x900 0x40>;/* lane E */
-+ };
-+};
-+
-+/* Update DPMAC connections to external PHYs, under SerDes 0x2a_0x49. */
-+&dpmac9 {
-+ phy-handle = <&mdio0_phy12>;
-+};
-+&dpmac10 {
-+ phy-handle = <&mdio0_phy13>;
-+};
-+&dpmac11 {
-+ phy-handle = <&mdio0_phy14>;
-+};
-+&dpmac12 {
-+ phy-handle = <&mdio0_phy15>;
-+};
---- /dev/null
-+++ b/arch/arm64/boot/dts/freescale/fsl-ls2088a-rdb.dts
-@@ -0,0 +1,104 @@
-+// SPDX-License-Identifier: (GPL-2.0+ OR MIT)
-+/*
-+ * Device Tree file for Freescale LS2088A RDB Board.
-+ *
-+ * Copyright 2016 Freescale Semiconductor, Inc.
-+ * Copyright 2017 NXP
-+ *
-+ * Abhimanyu Saini <abhimanyu.saini@nxp.com>
-+ *
-+ */
-+
-+/dts-v1/;
-+
-+#include "fsl-ls2088a.dtsi"
-+#include "fsl-ls208xa-rdb.dtsi"
-+
-+/ {
-+ model = "Freescale Layerscape 2088A RDB Board";
-+ compatible = "fsl,ls2088a-rdb", "fsl,ls2088a";
-+
-+ chosen {
-+ stdout-path = "serial1:115200n8";
-+ };
-+};
-+
-+&emdio1 {
-+ status = "disabled";
-+ /* CS4340 PHYs */
-+ mdio1_phy1: emdio1_phy@1 {
-+ reg = <0x10>;
-+ phy-connection-type = "xfi";
-+ };
-+ mdio1_phy2: emdio1_phy@2 {
-+ reg = <0x11>;
-+ phy-connection-type = "xfi";
-+ };
-+ mdio1_phy3: emdio1_phy@3 {
-+ reg = <0x12>;
-+ phy-connection-type = "xfi";
-+ };
-+ mdio1_phy4: emdio1_phy@4 {
-+ reg = <0x13>;
-+ phy-connection-type = "xfi";
-+ };
-+};
-+
-+&emdio2 {
-+ /* AQR405 PHYs */
-+ mdio2_phy1: emdio2_phy@1 {
-+ compatible = "ethernet-phy-ieee802.3-c45";
-+ interrupts = <0 1 0x4>; /* Level high type */
-+ reg = <0x0>;
-+ phy-connection-type = "xfi";
-+ };
-+ mdio2_phy2: emdio2_phy@2 {
-+ compatible = "ethernet-phy-ieee802.3-c45";
-+ interrupts = <0 2 0x4>; /* Level high type */
-+ reg = <0x1>;
-+ phy-connection-type = "xfi";
-+ };
-+ mdio2_phy3: emdio2_phy@3 {
-+ compatible = "ethernet-phy-ieee802.3-c45";
-+ interrupts = <0 4 0x4>; /* Level high type */
-+ reg = <0x2>;
-+ phy-connection-type = "xfi";
-+ };
-+ mdio2_phy4: emdio2_phy@4 {
-+ compatible = "ethernet-phy-ieee802.3-c45";
-+ interrupts = <0 5 0x4>; /* Level high type */
-+ reg = <0x3>;
-+ phy-connection-type = "xfi";
-+ };
-+};
-+
-+/* Update DPMAC connections to external PHYs, under the assumption of
-+ * SerDes 0x2a_0x41. This is currently the only SerDes supported on the board.
-+ */
-+/* Leave Cortina PHYs commented out until proper driver is integrated
-+ *&dpmac1 {
-+ * phy-handle = <&mdio1_phy1>;
-+ *};
-+ *&dpmac2 {
-+ * phy-handle = <&mdio1_phy2>;
-+ *};
-+ *&dpmac3 {
-+ * phy-handle = <&mdio1_phy3>;
-+ *};
-+ *&dpmac4 {
-+ * phy-handle = <&mdio1_phy4>;
-+ *};
-+ */
-+
-+&dpmac5 {
-+ phy-handle = <&mdio2_phy1>;
-+};
-+&dpmac6 {
-+ phy-handle = <&mdio2_phy2>;
-+};
-+&dpmac7 {
-+ phy-handle = <&mdio2_phy3>;
-+};
-+&dpmac8 {
-+ phy-handle = <&mdio2_phy4>;
-+};
---- /dev/null
-+++ b/arch/arm64/boot/dts/freescale/fsl-ls2088a.dtsi
-@@ -0,0 +1,159 @@
-+// SPDX-License-Identifier: (GPL-2.0+ OR MIT)
-+/*
-+ * Device Tree Include file for Freescale Layerscape-2088A family SoC.
-+ *
-+ * Copyright 2016 Freescale Semiconductor, Inc.
-+ * Copyright 2017 NXP
-+ *
-+ * Abhimanyu Saini <abhimanyu.saini@nxp.com>
-+ *
-+ */
-+
-+#include "fsl-ls208xa.dtsi"
-+
-+&cpu {
-+ cpu0: cpu@0 {
-+ device_type = "cpu";
-+ compatible = "arm,cortex-a72";
-+ reg = <0x0>;
-+ clocks = <&clockgen 1 0>;
-+ next-level-cache = <&cluster0_l2>;
-+ #cooling-cells = <2>;
-+ cpu-idle-states = <&CPU_PH20>;
-+ };
-+
-+ cpu1: cpu@1 {
-+ device_type = "cpu";
-+ compatible = "arm,cortex-a72";
-+ reg = <0x1>;
-+ clocks = <&clockgen 1 0>;
-+ next-level-cache = <&cluster0_l2>;
-+ cpu-idle-states = <&CPU_PH20>;
-+ };
-+
-+ cpu2: cpu@100 {
-+ device_type = "cpu";
-+ compatible = "arm,cortex-a72";
-+ reg = <0x100>;
-+ clocks = <&clockgen 1 1>;
-+ next-level-cache = <&cluster1_l2>;
-+ #cooling-cells = <2>;
-+ cpu-idle-states = <&CPU_PH20>;
-+ };
-+
-+ cpu3: cpu@101 {
-+ device_type = "cpu";
-+ compatible = "arm,cortex-a72";
-+ reg = <0x101>;
-+ clocks = <&clockgen 1 1>;
-+ next-level-cache = <&cluster1_l2>;
-+ cpu-idle-states = <&CPU_PH20>;
-+ };
-+
-+ cpu4: cpu@200 {
-+ device_type = "cpu";
-+ compatible = "arm,cortex-a72";
-+ reg = <0x200>;
-+ clocks = <&clockgen 1 2>;
-+ next-level-cache = <&cluster2_l2>;
-+ #cooling-cells = <2>;
-+ cpu-idle-states = <&CPU_PH20>;
-+ };
-+
-+ cpu5: cpu@201 {
-+ device_type = "cpu";
-+ compatible = "arm,cortex-a72";
-+ reg = <0x201>;
-+ clocks = <&clockgen 1 2>;
-+ next-level-cache = <&cluster2_l2>;
-+ cpu-idle-states = <&CPU_PH20>;
-+ };
-+
-+ cpu6: cpu@300 {
-+ device_type = "cpu";
-+ compatible = "arm,cortex-a72";
-+ reg = <0x300>;
-+ clocks = <&clockgen 1 3>;
-+ next-level-cache = <&cluster3_l2>;
-+ #cooling-cells = <2>;
-+ cpu-idle-states = <&CPU_PH20>;
-+ };
-+
-+ cpu7: cpu@301 {
-+ device_type = "cpu";
-+ compatible = "arm,cortex-a72";
-+ reg = <0x301>;
-+ clocks = <&clockgen 1 3>;
-+ next-level-cache = <&cluster3_l2>;
-+ cpu-idle-states = <&CPU_PH20>;
-+ };
-+
-+ idle-states {
-+ /*
-+ * PSCI node is not added default, U-boot will add missing
-+ * parts if it determines to use PSCI.
-+ */
-+ entry-method = "arm,psci";
-+
-+ CPU_PH20: cpu-ph20 {
-+ compatible = "arm,idle-state";
-+ idle-state-name = "PH20";
-+ arm,psci-suspend-param = <0x0>;
-+ entry-latency-us = <1000>;
-+ exit-latency-us = <1000>;
-+ min-residency-us = <3000>;
-+ };
-+ };
-+
-+ cluster0_l2: l2-cache0 {
-+ compatible = "cache";
-+ };
-+
-+ cluster1_l2: l2-cache1 {
-+ compatible = "cache";
-+ };
-+
-+ cluster2_l2: l2-cache2 {
-+ compatible = "cache";
-+ };
-+
-+ cluster3_l2: l2-cache3 {
-+ compatible = "cache";
-+ };
-+};
-+
-+&pcie1 {
-+ compatible = "fsl,ls2088a-pcie", "snps,dw-pcie";
-+ reg = <0x00 0x03400000 0x0 0x00100000 /* controller registers */
-+ 0x20 0x00000000 0x0 0x00002000>; /* configuration space */
-+
-+ ranges = <0x81000000 0x0 0x00000000 0x20 0x00010000 0x0 0x00010000
-+ 0x82000000 0x0 0x40000000 0x20 0x40000000 0x0 0x40000000>;
-+};
-+
-+&pcie2 {
-+ compatible = "fsl,ls2088a-pcie", "snps,dw-pcie";
-+ reg = <0x00 0x03500000 0x0 0x00100000 /* controller registers */
-+ 0x28 0x00000000 0x0 0x00002000>; /* configuration space */
-+
-+ ranges = <0x81000000 0x0 0x00000000 0x28 0x00010000 0x0 0x00010000
-+ 0x82000000 0x0 0x40000000 0x28 0x40000000 0x0 0x40000000>;
-+};
-+
-+&pcie3 {
-+ compatible = "fsl,ls2088a-pcie", "snps,dw-pcie";
-+ reg = <0x00 0x03600000 0x0 0x00100000 /* controller registers */
-+ 0x30 0x00000000 0x0 0x00002000>; /* configuration space */
-+
-+ ranges = <0x81000000 0x0 0x00000000 0x30 0x00010000 0x0 0x00010000
-+ 0x82000000 0x0 0x40000000 0x30 0x40000000 0x0 0x40000000>;
-+};
-+
-+&pcie4 {
-+ compatible = "fsl,ls2088a-pcie", "snps,dw-pcie";
-+ reg = <0x00 0x03700000 0x0 0x00100000 /* controller registers */
-+ 0x38 0x00000000 0x0 0x00002000>; /* configuration space */
-+
-+ ranges = <0x81000000 0x0 0x00000000 0x38 0x00010000 0x0 0x00010000
-+ 0x82000000 0x0 0x40000000 0x38 0x40000000 0x0 0x40000000>;
-+};
---- /dev/null
-+++ b/arch/arm64/boot/dts/freescale/fsl-ls208xa-qds.dtsi
-@@ -0,0 +1,162 @@
-+// SPDX-License-Identifier: (GPL-2.0+ OR MIT)
-+/*
-+ * Device Tree file for Freescale LS2080A QDS Board.
-+ *
-+ * Copyright 2016 Freescale Semiconductor, Inc.
-+ * Copyright 2017 NXP
-+ *
-+ * Abhimanyu Saini <abhimanyu.saini@nxp.com>
-+ *
-+ */
-+
-+&esdhc {
-+ mmc-hs200-1_8v;
-+ status = "okay";
-+};
-+
-+&ifc {
-+ status = "okay";
-+ #address-cells = <2>;
-+ #size-cells = <1>;
-+ ranges = <0x0 0x0 0x5 0x80000000 0x08000000
-+ 0x2 0x0 0x5 0x30000000 0x00010000
-+ 0x3 0x0 0x5 0x20000000 0x00010000>;
-+
-+ nor@0,0 {
-+ #address-cells = <1>;
-+ #size-cells = <1>;
-+ compatible = "cfi-flash";
-+ reg = <0x0 0x0 0x8000000>;
-+ bank-width = <2>;
-+ device-width = <1>;
-+ };
-+
-+ nand@2,0 {
-+ compatible = "fsl,ifc-nand";
-+ reg = <0x2 0x0 0x10000>;
-+ };
-+
-+ cpld@3,0 {
-+ reg = <0x3 0x0 0x10000>;
-+ compatible = "fsl,ls2080aqds-fpga", "fsl,fpga-qixis";
-+ };
-+};
-+
-+&i2c0 {
-+ status = "okay";
-+ pca9547@77 {
-+ compatible = "nxp,pca9547";
-+ reg = <0x77>;
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ i2c@0 {
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ reg = <0x00>;
-+ rtc@68 {
-+ compatible = "dallas,ds3232";
-+ reg = <0x68>;
-+ };
-+ };
-+
-+ i2c@2 {
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ reg = <0x02>;
-+
-+ ina220@40 {
-+ compatible = "ti,ina220";
-+ reg = <0x40>;
-+ shunt-resistor = <500>;
-+ };
-+
-+ ina220@41 {
-+ compatible = "ti,ina220";
-+ reg = <0x41>;
-+ shunt-resistor = <1000>;
-+ };
-+ };
-+
-+ i2c@3 {
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ reg = <0x3>;
-+
-+ adt7481@4c {
-+ compatible = "adi,adt7461";
-+ reg = <0x4c>;
-+ };
-+ };
-+ };
-+};
-+
-+&i2c1 {
-+ status = "disabled";
-+};
-+
-+&i2c2 {
-+ status = "disabled";
-+};
-+
-+&i2c3 {
-+ status = "disabled";
-+};
-+
-+&dspi {
-+ status = "okay";
-+ dflash0: n25q128a {
-+ #address-cells = <1>;
-+ #size-cells = <1>;
-+ compatible = "st,m25p80";
-+ spi-max-frequency = <3000000>;
-+ reg = <0>;
-+ };
-+ dflash1: sst25wf040b {
-+ #address-cells = <1>;
-+ #size-cells = <1>;
-+ compatible = "st,m25p80";
-+ spi-max-frequency = <3000000>;
-+ reg = <1>;
-+ };
-+ dflash2: en25s64 {
-+ #address-cells = <1>;
-+ #size-cells = <1>;
-+ compatible = "st,m25p80";
-+ spi-max-frequency = <3000000>;
-+ reg = <2>;
-+ };
-+};
-+
-+&qspi {
-+ status = "okay";
-+ flash0: s25fl256s1@0 {
-+ #address-cells = <1>;
-+ #size-cells = <1>;
-+ compatible = "st,m25p80";
-+ spi-max-frequency = <20000000>;
-+ reg = <0>;
-+ };
-+ flash2: s25fl256s1@2 {
-+ #address-cells = <1>;
-+ #size-cells = <1>;
-+ compatible = "st,m25p80";
-+ spi-max-frequency = <20000000>;
-+ reg = <0>;
-+ };
-+};
-+
-+&sata0 {
-+ status = "okay";
-+};
-+
-+&sata1 {
-+ status = "okay";
-+};
-+
-+&usb0 {
-+ status = "okay";
-+};
-+
-+&usb1 {
-+ status = "okay";
-+};
---- /dev/null
-+++ b/arch/arm64/boot/dts/freescale/fsl-ls208xa-rdb.dtsi
-@@ -0,0 +1,136 @@
-+// SPDX-License-Identifier: (GPL-2.0+ OR MIT)
-+/*
-+ * Device Tree file for Freescale LS2080A RDB Board.
-+ *
-+ * Copyright 2016 Freescale Semiconductor, Inc.
-+ * Copyright 2017 NXP
-+ *
-+ * Abhimanyu Saini <abhimanyu.saini@nxp.com>
-+ *
-+ */
-+
-+&esdhc {
-+ status = "okay";
-+};
-+
-+&ifc {
-+ status = "okay";
-+ #address-cells = <2>;
-+ #size-cells = <1>;
-+ ranges = <0x0 0x0 0x5 0x80000000 0x08000000
-+ 0x2 0x0 0x5 0x30000000 0x00010000
-+ 0x3 0x0 0x5 0x20000000 0x00010000>;
-+
-+ nor@0,0 {
-+ #address-cells = <1>;
-+ #size-cells = <1>;
-+ compatible = "cfi-flash";
-+ reg = <0x0 0x0 0x8000000>;
-+ bank-width = <2>;
-+ device-width = <1>;
-+ };
-+
-+ nand@2,0 {
-+ compatible = "fsl,ifc-nand";
-+ reg = <0x2 0x0 0x10000>;
-+ };
-+
-+ cpld@3,0 {
-+ reg = <0x3 0x0 0x10000>;
-+ compatible = "fsl,ls2080aqds-fpga", "fsl,fpga-qixis";
-+ };
-+
-+};
-+
-+&i2c0 {
-+ status = "okay";
-+ pca9547@75 {
-+ compatible = "nxp,pca9547";
-+ reg = <0x75>;
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ i2c-mux-never-disable;
-+ i2c@1 {
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ reg = <0x01>;
-+ rtc@68 {
-+ compatible = "dallas,ds3232";
-+ reg = <0x68>;
-+ };
-+ };
-+
-+ i2c@2 {
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ reg = <0x02>;
-+ ina220@40 {
-+ compatible = "ti,ina220";
-+ reg = <0x40>;
-+ shunt-resistor = <500>;
-+ };
-+ };
-+
-+ i2c@3 {
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ reg = <0x3>;
-+
-+ adt7481@4c {
-+ compatible = "adi,adt7461";
-+ reg = <0x4c>;
-+ };
-+ };
-+ };
-+};
-+
-+&i2c1 {
-+ status = "disabled";
-+};
-+
-+&i2c2 {
-+ status = "disabled";
-+};
-+
-+&i2c3 {
-+ status = "disabled";
-+};
-+
-+&dspi {
-+ status = "okay";
-+ dflash0: n25q512a {
-+ #address-cells = <1>;
-+ #size-cells = <1>;
-+ compatible = "st,m25p80";
-+ spi-max-frequency = <3000000>;
-+ reg = <0>;
-+ };
-+};
-+
-+&qspi {
-+ status = "okay";
-+ flash0: s25fs512s@0 {
-+ #address-cells = <1>;
-+ #size-cells = <1>;
-+ compatible = "spansion,m25p80";
-+ m25p,fast-read;
-+ spi-max-frequency = <20000000>;
-+ reg = <0>;
-+ };
-+};
-+
-+&sata0 {
-+ status = "okay";
-+};
-+
-+&sata1 {
-+ status = "okay";
-+};
-+
-+&usb0 {
-+ status = "okay";
-+};
-+
-+&usb1 {
-+ status = "okay";
-+};
---- /dev/null
-+++ b/arch/arm64/boot/dts/freescale/fsl-ls208xa.dtsi
-@@ -0,0 +1,885 @@
-+// SPDX-License-Identifier: (GPL-2.0+ OR MIT)
-+/*
-+ * Device Tree Include file for Freescale Layerscape-2080A family SoC.
-+ *
-+ * Copyright 2016 Freescale Semiconductor, Inc.
-+ * Copyright 2017 NXP
-+ *
-+ * Abhimanyu Saini <abhimanyu.saini@nxp.com>
-+ *
-+ */
-+
-+#include <dt-bindings/thermal/thermal.h>
-+#include <dt-bindings/interrupt-controller/arm-gic.h>
-+
-+/ {
-+ compatible = "fsl,ls2080a";
-+ interrupt-parent = <&gic>;
-+ #address-cells = <2>;
-+ #size-cells = <2>;
-+
-+ aliases {
-+ crypto = &crypto;
-+ serial0 = &serial0;
-+ serial1 = &serial1;
-+ };
-+
-+ cpu: cpus {
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ };
-+
-+ memory@80000000 {
-+ device_type = "memory";
-+ reg = <0x00000000 0x80000000 0 0x80000000>;
-+ /* DRAM space - 1, size : 2 GB DRAM */
-+ };
-+
-+ sysclk: sysclk {
-+ compatible = "fixed-clock";
-+ #clock-cells = <0>;
-+ clock-frequency = <100000000>;
-+ clock-output-names = "sysclk";
-+ };
-+
-+ gic: interrupt-controller@6000000 {
-+ compatible = "arm,gic-v3";
-+ reg = <0x0 0x06000000 0 0x10000>, /* GIC Dist */
-+ <0x0 0x06100000 0 0x100000>, /* GICR (RD_base + SGI_base) */
-+ <0x0 0x0c0c0000 0 0x2000>, /* GICC */
-+ <0x0 0x0c0d0000 0 0x1000>, /* GICH */
-+ <0x0 0x0c0e0000 0 0x20000>; /* GICV */
-+ #interrupt-cells = <3>;
-+ #address-cells = <2>;
-+ #size-cells = <2>;
-+ ranges;
-+ interrupt-controller;
-+ interrupts = <1 9 0x4>;
-+
-+ its: gic-its@6020000 {
-+ compatible = "arm,gic-v3-its";
-+ msi-controller;
-+ reg = <0x0 0x6020000 0 0x20000>;
-+ };
-+ };
-+
-+ rstcr: syscon@1e60000 {
-+ compatible = "fsl,ls2080a-rstcr", "syscon";
-+ reg = <0x0 0x1e60000 0x0 0x4>;
-+ };
-+
-+ reboot {
-+ compatible ="syscon-reboot";
-+ regmap = <&rstcr>;
-+ offset = <0x0>;
-+ mask = <0x2>;
-+ };
-+
-+ timer: timer {
-+ compatible = "arm,armv8-timer";
-+ interrupts = <1 13 4>, /* Physical Secure PPI, active-low */
-+ <1 14 4>, /* Physical Non-Secure PPI, active-low */
-+ <1 11 4>, /* Virtual PPI, active-low */
-+ <1 10 4>; /* Hypervisor PPI, active-low */
-+ };
-+
-+ pmu {
-+ compatible = "arm,armv8-pmuv3";
-+ interrupts = <1 7 0x8>; /* PMU PPI, Level low type */
-+ };
-+
-+ soc {
-+ compatible = "simple-bus";
-+ #address-cells = <2>;
-+ #size-cells = <2>;
-+ ranges;
-+ dma-ranges = <0x0 0x0 0x0 0x0 0x10000 0x00000000>;
-+
-+ clockgen: clocking@1300000 {
-+ compatible = "fsl,ls2080a-clockgen";
-+ reg = <0 0x1300000 0 0xa0000>;
-+ #clock-cells = <2>;
-+ clocks = <&sysclk>;
-+ };
-+
-+ dcfg: dcfg@1e00000 {
-+ compatible = "fsl,ls2080a-dcfg", "syscon";
-+ reg = <0x0 0x1e00000 0x0 0x10000>;
-+ little-endian;
-+ };
-+
-+ tmu: tmu@1f80000 {
-+ compatible = "fsl,qoriq-tmu";
-+ reg = <0x0 0x1f80000 0x0 0x10000>;
-+ interrupts = <0 23 0x4>;
-+ fsl,tmu-range = <0xb0000 0x9002a 0x6004c 0x30062>;
-+ fsl,tmu-calibration = <0x00000000 0x00000026
-+ 0x00000001 0x0000002d
-+ 0x00000002 0x00000032
-+ 0x00000003 0x00000039
-+ 0x00000004 0x0000003f
-+ 0x00000005 0x00000046
-+ 0x00000006 0x0000004d
-+ 0x00000007 0x00000054
-+ 0x00000008 0x0000005a
-+ 0x00000009 0x00000061
-+ 0x0000000a 0x0000006a
-+ 0x0000000b 0x00000071
-+
-+ 0x00010000 0x00000025
-+ 0x00010001 0x0000002c
-+ 0x00010002 0x00000035
-+ 0x00010003 0x0000003d
-+ 0x00010004 0x00000045
-+ 0x00010005 0x0000004e
-+ 0x00010006 0x00000057
-+ 0x00010007 0x00000061
-+ 0x00010008 0x0000006b
-+ 0x00010009 0x00000076
-+
-+ 0x00020000 0x00000029
-+ 0x00020001 0x00000033
-+ 0x00020002 0x0000003d
-+ 0x00020003 0x00000049
-+ 0x00020004 0x00000056
-+ 0x00020005 0x00000061
-+ 0x00020006 0x0000006d
-+
-+ 0x00030000 0x00000021
-+ 0x00030001 0x0000002a
-+ 0x00030002 0x0000003c
-+ 0x00030003 0x0000004e>;
-+ little-endian;
-+ #thermal-sensor-cells = <1>;
-+ };
-+
-+ thermal-zones {
-+ cpu_thermal: cpu-thermal {
-+ polling-delay-passive = <1000>;
-+ polling-delay = <5000>;
-+
-+ thermal-sensors = <&tmu 4>;
-+
-+ trips {
-+ cpu_alert: cpu-alert {
-+ temperature = <75000>;
-+ hysteresis = <2000>;
-+ type = "passive";
-+ };
-+ cpu_crit: cpu-crit {
-+ temperature = <85000>;
-+ hysteresis = <2000>;
-+ type = "critical";
-+ };
-+ };
-+
-+ cooling-maps {
-+ map0 {
-+ trip = <&cpu_alert>;
-+ cooling-device =
-+ <&cpu0 THERMAL_NO_LIMIT
-+ THERMAL_NO_LIMIT>;
-+ };
-+ map1 {
-+ trip = <&cpu_alert>;
-+ cooling-device =
-+ <&cpu2 THERMAL_NO_LIMIT
-+ THERMAL_NO_LIMIT>;
-+ };
-+ map2 {
-+ trip = <&cpu_alert>;
-+ cooling-device =
-+ <&cpu4 THERMAL_NO_LIMIT
-+ THERMAL_NO_LIMIT>;
-+ };
-+ map3 {
-+ trip = <&cpu_alert>;
-+ cooling-device =
-+ <&cpu6 THERMAL_NO_LIMIT
-+ THERMAL_NO_LIMIT>;
-+ };
-+ };
-+ };
-+ };
-+
-+ serial0: serial@21c0500 {
-+ compatible = "fsl,ns16550", "ns16550a";
-+ reg = <0x0 0x21c0500 0x0 0x100>;
-+ clocks = <&clockgen 4 3>;
-+ interrupts = <0 32 0x4>; /* Level high type */
-+ };
-+
-+ serial1: serial@21c0600 {
-+ compatible = "fsl,ns16550", "ns16550a";
-+ reg = <0x0 0x21c0600 0x0 0x100>;
-+ clocks = <&clockgen 4 3>;
-+ interrupts = <0 32 0x4>; /* Level high type */
-+ };
-+
-+ cluster1_core0_watchdog: wdt@c000000 {
-+ compatible = "arm,sp805-wdt", "arm,primecell";
-+ reg = <0x0 0xc000000 0x0 0x1000>;
-+ clocks = <&clockgen 4 3>, <&clockgen 4 3>;
-+ clock-names = "apb_pclk", "wdog_clk";
-+ };
-+
-+ cluster1_core1_watchdog: wdt@c010000 {
-+ compatible = "arm,sp805-wdt", "arm,primecell";
-+ reg = <0x0 0xc010000 0x0 0x1000>;
-+ clocks = <&clockgen 4 3>, <&clockgen 4 3>;
-+ clock-names = "apb_pclk", "wdog_clk";
-+ };
-+
-+ cluster2_core0_watchdog: wdt@c100000 {
-+ compatible = "arm,sp805-wdt", "arm,primecell";
-+ reg = <0x0 0xc100000 0x0 0x1000>;
-+ clocks = <&clockgen 4 3>, <&clockgen 4 3>;
-+ clock-names = "apb_pclk", "wdog_clk";
-+ };
-+
-+ cluster2_core1_watchdog: wdt@c110000 {
-+ compatible = "arm,sp805-wdt", "arm,primecell";
-+ reg = <0x0 0xc110000 0x0 0x1000>;
-+ clocks = <&clockgen 4 3>, <&clockgen 4 3>;
-+ clock-names = "apb_pclk", "wdog_clk";
-+ };
-+
-+ cluster3_core0_watchdog: wdt@c200000 {
-+ compatible = "arm,sp805-wdt", "arm,primecell";
-+ reg = <0x0 0xc200000 0x0 0x1000>;
-+ clocks = <&clockgen 4 3>, <&clockgen 4 3>;
-+ clock-names = "apb_pclk", "wdog_clk";
-+ };
-+
-+ cluster3_core1_watchdog: wdt@c210000 {
-+ compatible = "arm,sp805-wdt", "arm,primecell";
-+ reg = <0x0 0xc210000 0x0 0x1000>;
-+ clocks = <&clockgen 4 3>, <&clockgen 4 3>;
-+ clock-names = "apb_pclk", "wdog_clk";
-+ };
-+
-+ cluster4_core0_watchdog: wdt@c300000 {
-+ compatible = "arm,sp805-wdt", "arm,primecell";
-+ reg = <0x0 0xc300000 0x0 0x1000>;
-+ clocks = <&clockgen 4 3>, <&clockgen 4 3>;
-+ clock-names = "apb_pclk", "wdog_clk";
-+ };
-+
-+ cluster4_core1_watchdog: wdt@c310000 {
-+ compatible = "arm,sp805-wdt", "arm,primecell";
-+ reg = <0x0 0xc310000 0x0 0x1000>;
-+ clocks = <&clockgen 4 3>, <&clockgen 4 3>;
-+ clock-names = "apb_pclk", "wdog_clk";
-+ };
-+
-+ crypto: crypto@8000000 {
-+ compatible = "fsl,sec-v5.0", "fsl,sec-v4.0";
-+ fsl,sec-era = <8>;
-+ #address-cells = <1>;
-+ #size-cells = <1>;
-+ ranges = <0x0 0x00 0x8000000 0x100000>;
-+ reg = <0x00 0x8000000 0x0 0x100000>;
-+ interrupts = <GIC_SPI 139 IRQ_TYPE_LEVEL_HIGH>;
-+ dma-coherent;
-+
-+ sec_jr0: jr@10000 {
-+ compatible = "fsl,sec-v5.0-job-ring",
-+ "fsl,sec-v4.0-job-ring";
-+ reg = <0x10000 0x10000>;
-+ interrupts = <GIC_SPI 140 IRQ_TYPE_LEVEL_HIGH>;
-+ };
-+
-+ sec_jr1: jr@20000 {
-+ compatible = "fsl,sec-v5.0-job-ring",
-+ "fsl,sec-v4.0-job-ring";
-+ reg = <0x20000 0x10000>;
-+ interrupts = <GIC_SPI 141 IRQ_TYPE_LEVEL_HIGH>;
-+ };
-+
-+ sec_jr2: jr@30000 {
-+ compatible = "fsl,sec-v5.0-job-ring",
-+ "fsl,sec-v4.0-job-ring";
-+ reg = <0x30000 0x10000>;
-+ interrupts = <GIC_SPI 142 IRQ_TYPE_LEVEL_HIGH>;
-+ };
-+
-+ sec_jr3: jr@40000 {
-+ compatible = "fsl,sec-v5.0-job-ring",
-+ "fsl,sec-v4.0-job-ring";
-+ reg = <0x40000 0x10000>;
-+ interrupts = <GIC_SPI 143 IRQ_TYPE_LEVEL_HIGH>;
-+ };
-+ };
-+
-+ fsl_mc: fsl-mc@80c000000 {
-+ compatible = "fsl,qoriq-mc";
-+ reg = <0x00000008 0x0c000000 0 0x40>, /* MC portal base */
-+ <0x00000000 0x08340000 0 0x40000>; /* MC control reg */
-+ msi-parent = <&its>;
-+ iommu-map = <0 &smmu 0 0>; /* This is fixed-up by u-boot */
-+ dma-coherent;
-+ #address-cells = <3>;
-+ #size-cells = <1>;
-+
-+ /*
-+ * Region type 0x0 - MC portals
-+ * Region type 0x1 - QBMAN portals
-+ */
-+ ranges = <0x0 0x0 0x0 0x8 0x0c000000 0x4000000
-+ 0x1 0x0 0x0 0x8 0x18000000 0x8000000>;
-+
-+ /*
-+ * Define the maximum number of MACs present on the SoC.
-+ */
-+ dpmacs {
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+
-+ dpmac1: dpmac@1 {
-+ compatible = "fsl,qoriq-mc-dpmac";
-+ reg = <0x1>;
-+ };
-+
-+ dpmac2: dpmac@2 {
-+ compatible = "fsl,qoriq-mc-dpmac";
-+ reg = <0x2>;
-+ };
-+
-+ dpmac3: dpmac@3 {
-+ compatible = "fsl,qoriq-mc-dpmac";
-+ reg = <0x3>;
-+ };
-+
-+ dpmac4: dpmac@4 {
-+ compatible = "fsl,qoriq-mc-dpmac";
-+ reg = <0x4>;
-+ };
-+
-+ dpmac5: dpmac@5 {
-+ compatible = "fsl,qoriq-mc-dpmac";
-+ reg = <0x5>;
-+ };
-+
-+ dpmac6: dpmac@6 {
-+ compatible = "fsl,qoriq-mc-dpmac";
-+ reg = <0x6>;
-+ };
-+
-+ dpmac7: dpmac@7 {
-+ compatible = "fsl,qoriq-mc-dpmac";
-+ reg = <0x7>;
-+ };
-+
-+ dpmac8: dpmac@8 {
-+ compatible = "fsl,qoriq-mc-dpmac";
-+ reg = <0x8>;
-+ };
-+
-+ dpmac9: dpmac@9 {
-+ compatible = "fsl,qoriq-mc-dpmac";
-+ reg = <0x9>;
-+ };
-+
-+ dpmac10: dpmac@a {
-+ compatible = "fsl,qoriq-mc-dpmac";
-+ reg = <0xa>;
-+ };
-+
-+ dpmac11: dpmac@b {
-+ compatible = "fsl,qoriq-mc-dpmac";
-+ reg = <0xb>;
-+ };
-+
-+ dpmac12: dpmac@c {
-+ compatible = "fsl,qoriq-mc-dpmac";
-+ reg = <0xc>;
-+ };
-+
-+ dpmac13: dpmac@d {
-+ compatible = "fsl,qoriq-mc-dpmac";
-+ reg = <0xd>;
-+ };
-+
-+ dpmac14: dpmac@e {
-+ compatible = "fsl,qoriq-mc-dpmac";
-+ reg = <0xe>;
-+ };
-+
-+ dpmac15: dpmac@f {
-+ compatible = "fsl,qoriq-mc-dpmac";
-+ reg = <0xf>;
-+ };
-+
-+ dpmac16: dpmac@10 {
-+ compatible = "fsl,qoriq-mc-dpmac";
-+ reg = <0x10>;
-+ };
-+ };
-+ };
-+
-+ smmu: iommu@5000000 {
-+ compatible = "arm,mmu-500";
-+ reg = <0 0x5000000 0 0x800000>;
-+ #global-interrupts = <12>;
-+ #iommu-cells = <1>;
-+ stream-match-mask = <0x7C00>;
-+ interrupts = <0 13 4>, /* global secure fault */
-+ <0 14 4>, /* combined secure interrupt */
-+ <0 15 4>, /* global non-secure fault */
-+ <0 16 4>, /* combined non-secure interrupt */
-+ /* performance counter interrupts 0-7 */
-+ <0 211 4>, <0 212 4>,
-+ <0 213 4>, <0 214 4>,
-+ <0 215 4>, <0 216 4>,
-+ <0 217 4>, <0 218 4>,
-+ /* per context interrupt, 64 interrupts */
-+ <0 146 4>, <0 147 4>,
-+ <0 148 4>, <0 149 4>,
-+ <0 150 4>, <0 151 4>,
-+ <0 152 4>, <0 153 4>,
-+ <0 154 4>, <0 155 4>,
-+ <0 156 4>, <0 157 4>,
-+ <0 158 4>, <0 159 4>,
-+ <0 160 4>, <0 161 4>,
-+ <0 162 4>, <0 163 4>,
-+ <0 164 4>, <0 165 4>,
-+ <0 166 4>, <0 167 4>,
-+ <0 168 4>, <0 169 4>,
-+ <0 170 4>, <0 171 4>,
-+ <0 172 4>, <0 173 4>,
-+ <0 174 4>, <0 175 4>,
-+ <0 176 4>, <0 177 4>,
-+ <0 178 4>, <0 179 4>,
-+ <0 180 4>, <0 181 4>,
-+ <0 182 4>, <0 183 4>,
-+ <0 184 4>, <0 185 4>,
-+ <0 186 4>, <0 187 4>,
-+ <0 188 4>, <0 189 4>,
-+ <0 190 4>, <0 191 4>,
-+ <0 192 4>, <0 193 4>,
-+ <0 194 4>, <0 195 4>,
-+ <0 196 4>, <0 197 4>,
-+ <0 198 4>, <0 199 4>,
-+ <0 200 4>, <0 201 4>,
-+ <0 202 4>, <0 203 4>,
-+ <0 204 4>, <0 205 4>,
-+ <0 206 4>, <0 207 4>,
-+ <0 208 4>, <0 209 4>;
-+ };
-+
-+ dspi: dspi@2100000 {
-+ status = "disabled";
-+ compatible = "fsl,ls2080a-dspi", "fsl,ls2085a-dspi";
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ reg = <0x0 0x2100000 0x0 0x10000>;
-+ interrupts = <0 26 0x4>; /* Level high type */
-+ clocks = <&clockgen 4 3>;
-+ clock-names = "dspi";
-+ spi-num-chipselects = <5>;
-+ bus-num = <0>;
-+ };
-+
-+ esdhc: esdhc@2140000 {
-+ status = "disabled";
-+ compatible = "fsl,ls2080a-esdhc", "fsl,esdhc";
-+ reg = <0x0 0x2140000 0x0 0x10000>;
-+ interrupts = <0 28 0x4>; /* Level high type */
-+ clocks = <&clockgen 4 1>;
-+ voltage-ranges = <1800 1800 3300 3300>;
-+ sdhci,auto-cmd12;
-+ little-endian;
-+ bus-width = <4>;
-+ };
-+
-+ gpio0: gpio@2300000 {
-+ compatible = "fsl,ls2080a-gpio", "fsl,qoriq-gpio";
-+ reg = <0x0 0x2300000 0x0 0x10000>;
-+ interrupts = <0 36 0x4>; /* Level high type */
-+ gpio-controller;
-+ little-endian;
-+ #gpio-cells = <2>;
-+ interrupt-controller;
-+ #interrupt-cells = <2>;
-+ };
-+
-+ gpio1: gpio@2310000 {
-+ compatible = "fsl,ls2080a-gpio", "fsl,qoriq-gpio";
-+ reg = <0x0 0x2310000 0x0 0x10000>;
-+ interrupts = <0 36 0x4>; /* Level high type */
-+ gpio-controller;
-+ little-endian;
-+ #gpio-cells = <2>;
-+ interrupt-controller;
-+ #interrupt-cells = <2>;
-+ };
-+
-+ gpio2: gpio@2320000 {
-+ compatible = "fsl,ls2080a-gpio", "fsl,qoriq-gpio";
-+ reg = <0x0 0x2320000 0x0 0x10000>;
-+ interrupts = <0 37 0x4>; /* Level high type */
-+ gpio-controller;
-+ little-endian;
-+ #gpio-cells = <2>;
-+ interrupt-controller;
-+ #interrupt-cells = <2>;
-+ };
-+
-+ gpio3: gpio@2330000 {
-+ compatible = "fsl,ls2080a-gpio", "fsl,qoriq-gpio";
-+ reg = <0x0 0x2330000 0x0 0x10000>;
-+ interrupts = <0 37 0x4>; /* Level high type */
-+ gpio-controller;
-+ little-endian;
-+ #gpio-cells = <2>;
-+ interrupt-controller;
-+ #interrupt-cells = <2>;
-+ };
-+
-+ /* TODO: WRIOP (CCSR?) */
-+ emdio1: mdio@0x8B96000 { /* WRIOP0: 0x8B8_0000,
-+ * E-MDIO1: 0x1_6000
-+ */
-+ compatible = "fsl,fman-memac-mdio";
-+ reg = <0x0 0x8B96000 0x0 0x1000>;
-+ device_type = "mdio"; /* TODO: is this necessary? */
-+ little-endian; /* force the driver in LE mode */
-+
-+ /* Not necessary on the QDS, but needed on the RDB */
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ };
-+
-+ emdio2: mdio@0x8B97000 { /* WRIOP0: 0x8B8_0000,
-+ * E-MDIO2: 0x1_7000
-+ */
-+ compatible = "fsl,fman-memac-mdio";
-+ reg = <0x0 0x8B97000 0x0 0x1000>;
-+ device_type = "mdio"; /* TODO: is this necessary? */
-+ little-endian; /* force the driver in LE mode */
-+
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ };
-+
-+ pcs_mdio1: mdio@0x8c07000 {
-+ compatible = "fsl,fman-memac-mdio";
-+ reg = <0x0 0x8c07000 0x0 0x1000>;
-+ device_type = "mdio";
-+ little-endian;
-+
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ };
-+
-+ pcs_mdio2: mdio@0x8c0b000 {
-+ compatible = "fsl,fman-memac-mdio";
-+ reg = <0x0 0x8c0b000 0x0 0x1000>;
-+ device_type = "mdio";
-+ little-endian;
-+
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ };
-+
-+ pcs_mdio3: mdio@0x8c0f000 {
-+ compatible = "fsl,fman-memac-mdio";
-+ reg = <0x0 0x8c0f000 0x0 0x1000>;
-+ device_type = "mdio";
-+ little-endian;
-+
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ };
-+
-+ pcs_mdio4: mdio@0x8c13000 {
-+ compatible = "fsl,fman-memac-mdio";
-+ reg = <0x0 0x8c13000 0x0 0x1000>;
-+ device_type = "mdio";
-+ little-endian;
-+
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ };
-+
-+ pcs_mdio5: mdio@0x8c17000 {
-+ status = "disabled";
-+ compatible = "fsl,fman-memac-mdio";
-+ reg = <0x0 0x8c17000 0x0 0x1000>;
-+ device_type = "mdio";
-+ little-endian;
-+
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ };
-+
-+ pcs_mdio6: mdio@0x8c1b000 {
-+ status = "disabled";
-+ compatible = "fsl,fman-memac-mdio";
-+ reg = <0x0 0x8c1b000 0x0 0x1000>;
-+ device_type = "mdio";
-+ little-endian;
-+
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ };
-+
-+ pcs_mdio7: mdio@0x8c1f000 {
-+ status = "disabled";
-+ compatible = "fsl,fman-memac-mdio";
-+ reg = <0x0 0x8c1f000 0x0 0x1000>;
-+ device_type = "mdio";
-+ little-endian;
-+
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ };
-+
-+ pcs_mdio8: mdio@0x8c23000 {
-+ status = "disabled";
-+ compatible = "fsl,fman-memac-mdio";
-+ reg = <0x0 0x8c23000 0x0 0x1000>;
-+ device_type = "mdio";
-+ little-endian;
-+
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ };
-+
-+ i2c0: i2c@2000000 {
-+ status = "disabled";
-+ compatible = "fsl,vf610-i2c", "fsl,ls208xa-vf610-i2c";
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ reg = <0x0 0x2000000 0x0 0x10000>;
-+ interrupts = <0 34 0x4>; /* Level high type */
-+ clock-names = "i2c";
-+ clocks = <&clockgen 4 1>;
-+ fsl-scl-gpio = <&gpio3 10 0>;
-+ };
-+
-+ i2c1: i2c@2010000 {
-+ status = "disabled";
-+ compatible = "fsl,vf610-i2c";
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ reg = <0x0 0x2010000 0x0 0x10000>;
-+ interrupts = <0 34 0x4>; /* Level high type */
-+ clock-names = "i2c";
-+ clocks = <&clockgen 4 1>;
-+ };
-+
-+ i2c2: i2c@2020000 {
-+ status = "disabled";
-+ compatible = "fsl,vf610-i2c";
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ reg = <0x0 0x2020000 0x0 0x10000>;
-+ interrupts = <0 35 0x4>; /* Level high type */
-+ clock-names = "i2c";
-+ clocks = <&clockgen 4 1>;
-+ };
-+
-+ i2c3: i2c@2030000 {
-+ status = "disabled";
-+ compatible = "fsl,vf610-i2c";
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ reg = <0x0 0x2030000 0x0 0x10000>;
-+ interrupts = <0 35 0x4>; /* Level high type */
-+ clock-names = "i2c";
-+ clocks = <&clockgen 4 1>;
-+ };
-+
-+ ifc: ifc@2240000 {
-+ compatible = "fsl,ifc", "simple-bus";
-+ reg = <0x0 0x2240000 0x0 0x20000>;
-+ interrupts = <0 21 0x4>; /* Level high type */
-+ little-endian;
-+ #address-cells = <2>;
-+ #size-cells = <1>;
-+
-+ ranges = <0 0 0x5 0x80000000 0x08000000
-+ 2 0 0x5 0x30000000 0x00010000
-+ 3 0 0x5 0x20000000 0x00010000>;
-+ };
-+
-+ qspi: quadspi@20c0000 {
-+ status = "disabled";
-+ compatible = "fsl,ls2080a-qspi", "fsl,ls1021a-qspi";
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ reg = <0x0 0x20c0000 0x0 0x10000>,
-+ <0x0 0x20000000 0x0 0x10000000>;
-+ reg-names = "QuadSPI", "QuadSPI-memory";
-+ interrupts = <0 25 0x4>; /* Level high type */
-+ clocks = <&clockgen 4 3>, <&clockgen 4 3>;
-+ clock-names = "qspi_en", "qspi";
-+ };
-+
-+ pcie1: pcie@3400000 {
-+ compatible = "fsl,ls2080a-pcie", "fsl,ls2085a-pcie",
-+ "snps,dw-pcie";
-+ reg-names = "regs", "config";
-+ interrupts = <0 108 0x4>; /* aer interrupt */
-+ interrupt-names = "aer";
-+ #address-cells = <3>;
-+ #size-cells = <2>;
-+ device_type = "pci";
-+ dma-coherent;
-+ num-lanes = <4>;
-+ bus-range = <0x0 0xff>;
-+ msi-parent = <&its>;
-+ iommu-map = <0 &smmu 0 1>; /* This is fixed-up by u-boot */
-+ #interrupt-cells = <1>;
-+ interrupt-map-mask = <0 0 0 7>;
-+ interrupt-map = <0000 0 0 1 &gic 0 0 0 109 4>,
-+ <0000 0 0 2 &gic 0 0 0 110 4>,
-+ <0000 0 0 3 &gic 0 0 0 111 4>,
-+ <0000 0 0 4 &gic 0 0 0 112 4>;
-+ };
-+
-+ pcie2: pcie@3500000 {
-+ compatible = "fsl,ls2080a-pcie", "fsl,ls2085a-pcie",
-+ "snps,dw-pcie";
-+ reg-names = "regs", "config";
-+ interrupts = <0 113 0x4>; /* aer interrupt */
-+ interrupt-names = "aer";
-+ #address-cells = <3>;
-+ #size-cells = <2>;
-+ device_type = "pci";
-+ dma-coherent;
-+ num-lanes = <4>;
-+ bus-range = <0x0 0xff>;
-+ msi-parent = <&its>;
-+ iommu-map = <0 &smmu 0 1>; /* This is fixed-up by u-boot */
-+ #interrupt-cells = <1>;
-+ interrupt-map-mask = <0 0 0 7>;
-+ interrupt-map = <0000 0 0 1 &gic 0 0 0 114 4>,
-+ <0000 0 0 2 &gic 0 0 0 115 4>,
-+ <0000 0 0 3 &gic 0 0 0 116 4>,
-+ <0000 0 0 4 &gic 0 0 0 117 4>;
-+ };
-+
-+ pcie3: pcie@3600000 {
-+ compatible = "fsl,ls2080a-pcie", "fsl,ls2085a-pcie",
-+ "snps,dw-pcie";
-+ reg-names = "regs", "config";
-+ interrupts = <0 118 0x4>; /* aer interrupt */
-+ interrupt-names = "aer";
-+ #address-cells = <3>;
-+ #size-cells = <2>;
-+ device_type = "pci";
-+ dma-coherent;
-+ num-lanes = <8>;
-+ bus-range = <0x0 0xff>;
-+ msi-parent = <&its>;
-+ iommu-map = <0 &smmu 0 1>; /* This is fixed-up by u-boot */
-+ #interrupt-cells = <1>;
-+ interrupt-map-mask = <0 0 0 7>;
-+ interrupt-map = <0000 0 0 1 &gic 0 0 0 119 4>,
-+ <0000 0 0 2 &gic 0 0 0 120 4>,
-+ <0000 0 0 3 &gic 0 0 0 121 4>,
-+ <0000 0 0 4 &gic 0 0 0 122 4>;
-+ };
-+
-+ pcie4: pcie@3700000 {
-+ compatible = "fsl,ls2080a-pcie", "fsl,ls2085a-pcie",
-+ "snps,dw-pcie";
-+ reg-names = "regs", "config";
-+ interrupts = <0 123 0x4>; /* aer interrupt */
-+ interrupt-names = "aer";
-+ #address-cells = <3>;
-+ #size-cells = <2>;
-+ device_type = "pci";
-+ dma-coherent;
-+ num-lanes = <4>;
-+ bus-range = <0x0 0xff>;
-+ msi-parent = <&its>;
-+ iommu-map = <0 &smmu 0 1>; /* This is fixed-up by u-boot */
-+ #interrupt-cells = <1>;
-+ interrupt-map-mask = <0 0 0 7>;
-+ interrupt-map = <0000 0 0 1 &gic 0 0 0 124 4>,
-+ <0000 0 0 2 &gic 0 0 0 125 4>,
-+ <0000 0 0 3 &gic 0 0 0 126 4>,
-+ <0000 0 0 4 &gic 0 0 0 127 4>;
-+ };
-+
-+ sata0: sata@3200000 {
-+ status = "disabled";
-+ compatible = "fsl,ls2080a-ahci";
-+ reg = <0x0 0x3200000 0x0 0x10000>;
-+ interrupts = <0 133 0x4>; /* Level high type */
-+ clocks = <&clockgen 4 3>;
-+ dma-coherent;
-+ };
-+
-+ sata1: sata@3210000 {
-+ status = "disabled";
-+ compatible = "fsl,ls2080a-ahci";
-+ reg = <0x0 0x3210000 0x0 0x10000>;
-+ interrupts = <0 136 0x4>; /* Level high type */
-+ clocks = <&clockgen 4 3>;
-+ dma-coherent;
-+ };
-+
-+ usb0: usb3@3100000 {
-+ status = "disabled";
-+ compatible = "snps,dwc3";
-+ reg = <0x0 0x3100000 0x0 0x10000>;
-+ interrupts = <0 80 0x4>; /* Level high type */
-+ dr_mode = "host";
-+ snps,quirk-frame-length-adjustment = <0x20>;
-+ snps,dis_rxdet_inp3_quirk;
-+ };
-+
-+ usb1: usb3@3110000 {
-+ status = "disabled";
-+ compatible = "snps,dwc3";
-+ reg = <0x0 0x3110000 0x0 0x10000>;
-+ interrupts = <0 81 0x4>; /* Level high type */
-+ dr_mode = "host";
-+ snps,quirk-frame-length-adjustment = <0x20>;
-+ snps,dis_rxdet_inp3_quirk;
-+ };
-+
-+ serdes1: serdes@1ea0000 {
-+ reg = <0x0 0x1ea0000 0 0x00002000>;
-+ };
-+
-+ ccn@4000000 {
-+ compatible = "arm,ccn-504";
-+ reg = <0x0 0x04000000 0x0 0x01000000>;
-+ interrupts = <0 12 4>;
-+ };
-+
-+ ftm0: ftm0@2800000 {
-+ compatible = "fsl,ls208xa-ftm";
-+ reg = <0x0 0x2800000 0x0 0x10000>,
-+ <0x0 0x1e34050 0x0 0x4>;
-+ interrupts = <0 44 4>;
-+ reg-names = "ftm", "FlexTimer1";
-+ };
-+ };
-+
-+ ddr1: memory-controller@1080000 {
-+ compatible = "fsl,qoriq-memory-controller";
-+ reg = <0x0 0x1080000 0x0 0x1000>;
-+ interrupts = <0 17 0x4>;
-+ little-endian;
-+ };
-+
-+ ddr2: memory-controller@1090000 {
-+ compatible = "fsl,qoriq-memory-controller";
-+ reg = <0x0 0x1090000 0x0 0x1000>;
-+ interrupts = <0 18 0x4>;
-+ little-endian;
-+ };
-+
-+ firmware {
-+ optee {
-+ compatible = "linaro,optee-tz";
-+ method = "smc";
-+ };
-+ };
-+};
---- /dev/null
-+++ b/arch/arm64/boot/dts/freescale/qoriq-bman-portals-sdk.dtsi
-@@ -0,0 +1,55 @@
-+/*
-+ * QorIQ BMan SDK Portals device tree nodes
-+ *
-+ * Copyright 2011-2016 Freescale Semiconductor Inc.
-+ * Copyright 2017 NXP
-+ *
-+ * SPDX-License-Identifier: (GPL-2.0+ OR BSD-3-Clause)
-+ */
-+
-+&bportals {
-+ bman-portal@0 {
-+ cell-index = <0>;
-+ };
-+
-+ bman-portal@10000 {
-+ cell-index = <1>;
-+ };
-+
-+ bman-portal@20000 {
-+ cell-index = <2>;
-+ };
-+
-+ bman-portal@30000 {
-+ cell-index = <3>;
-+ };
-+
-+ bman-portal@40000 {
-+ cell-index = <4>;
-+ };
-+
-+ bman-portal@50000 {
-+ cell-index = <5>;
-+ };
-+
-+ bman-portal@60000 {
-+ cell-index = <6>;
-+ };
-+
-+ bman-portal@70000 {
-+ cell-index = <7>;
-+ };
-+
-+ bman-portal@80000 {
-+ cell-index = <8>;
-+ };
-+
-+ bman-portal@90000 {
-+ cell-index = <9>;
-+ };
-+
-+ bman-bpids@0 {
-+ compatible = "fsl,bpid-range";
-+ fsl,bpid-range = <32 32>;
-+ };
-+};
---- /dev/null
-+++ b/arch/arm64/boot/dts/freescale/qoriq-bman-portals.dtsi
-@@ -0,0 +1,77 @@
-+// SPDX-License-Identifier: (GPL-2.0+ OR BSD-3-Clause)
-+/*
-+ * QorIQ BMan Portals device tree
-+ *
-+ * Copyright 2011-2016 Freescale Semiconductor Inc.
-+ *
-+ */
-+
-+&bportals {
-+ #address-cells = <1>;
-+ #size-cells = <1>;
-+ compatible = "simple-bus";
-+
-+ bman-portal@0 {
-+ /*
-+ * bootloader fix-ups are expected to provide the
-+ * "fsl,bman-portal-<hardware revision>" compatible
-+ */
-+ compatible = "fsl,bman-portal";
-+ reg = <0x0 0x4000>, <0x4000000 0x4000>;
-+ interrupts = <GIC_SPI 173 IRQ_TYPE_LEVEL_HIGH>;
-+ };
-+
-+ bman-portal@10000 {
-+ compatible = "fsl,bman-portal";
-+ reg = <0x10000 0x4000>, <0x4010000 0x4000>;
-+ interrupts = <GIC_SPI 175 IRQ_TYPE_LEVEL_HIGH>;
-+ };
-+
-+ bman-portal@20000 {
-+ compatible = "fsl,bman-portal";
-+ reg = <0x20000 0x4000>, <0x4020000 0x4000>;
-+ interrupts = <GIC_SPI 177 IRQ_TYPE_LEVEL_HIGH>;
-+ };
-+
-+ bman-portal@30000 {
-+ compatible = "fsl,bman-portal";
-+ reg = <0x30000 0x4000>, <0x4030000 0x4000>;
-+ interrupts = <GIC_SPI 179 IRQ_TYPE_LEVEL_HIGH>;
-+ };
-+
-+ bman-portal@40000 {
-+ compatible = "fsl,bman-portal";
-+ reg = <0x40000 0x4000>, <0x4040000 0x4000>;
-+ interrupts = <GIC_SPI 181 IRQ_TYPE_LEVEL_HIGH>;
-+ };
-+
-+ bman-portal@50000 {
-+ compatible = "fsl,bman-portal";
-+ reg = <0x50000 0x4000>, <0x4050000 0x4000>;
-+ interrupts = <GIC_SPI 183 IRQ_TYPE_LEVEL_HIGH>;
-+ };
-+
-+ bman-portal@60000 {
-+ compatible = "fsl,bman-portal";
-+ reg = <0x60000 0x4000>, <0x4060000 0x4000>;
-+ interrupts = <GIC_SPI 185 IRQ_TYPE_LEVEL_HIGH>;
-+ };
-+
-+ bman-portal@70000 {
-+ compatible = "fsl,bman-portal";
-+ reg = <0x70000 0x4000>, <0x4070000 0x4000>;
-+ interrupts = <GIC_SPI 187 IRQ_TYPE_LEVEL_HIGH>;
-+ };
-+
-+ bman-portal@80000 {
-+ compatible = "fsl,bman-portal";
-+ reg = <0x80000 0x4000>, <0x4080000 0x4000>;
-+ interrupts = <GIC_SPI 189 IRQ_TYPE_LEVEL_HIGH>;
-+ };
-+
-+ bman-portal@90000 {
-+ compatible = "fsl,bman-portal";
-+ reg = <0x90000 0x4000>, <0x4090000 0x4000>;
-+ interrupts = <GIC_SPI 191 IRQ_TYPE_LEVEL_HIGH>;
-+ };
-+};
---- /dev/null
-+++ b/arch/arm64/boot/dts/freescale/qoriq-dpaa-eth.dtsi
-@@ -0,0 +1,73 @@
-+/*
-+ * QorIQ FMan v3 10g port #1 device tree stub [ controller @ offset 0x400000 ]
-+ *
-+ * Copyright 2012 - 2015 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+fsldpaa: fsl,dpaa {
-+ compatible = "fsl,ls1043a-dpaa", "simple-bus", "fsl,dpaa";
-+ ethernet@0 {
-+ compatible = "fsl,dpa-ethernet";
-+ fsl,fman-mac = <&enet0>;
-+ dma-coherent;
-+ };
-+ ethernet@1 {
-+ compatible = "fsl,dpa-ethernet";
-+ fsl,fman-mac = <&enet1>;
-+ dma-coherent;
-+ };
-+ ethernet@2 {
-+ compatible = "fsl,dpa-ethernet";
-+ fsl,fman-mac = <&enet2>;
-+ dma-coherent;
-+ };
-+ ethernet@3 {
-+ compatible = "fsl,dpa-ethernet";
-+ fsl,fman-mac = <&enet3>;
-+ dma-coherent;
-+ };
-+ ethernet@4 {
-+ compatible = "fsl,dpa-ethernet";
-+ fsl,fman-mac = <&enet4>;
-+ dma-coherent;
-+ };
-+ ethernet@5 {
-+ compatible = "fsl,dpa-ethernet";
-+ fsl,fman-mac = <&enet5>;
-+ dma-coherent;
-+ };
-+ ethernet@8 {
-+ compatible = "fsl,dpa-ethernet";
-+ fsl,fman-mac = <&enet6>;
-+ dma-coherent;
-+ };
-+};
-+
---- /dev/null
-+++ b/arch/arm64/boot/dts/freescale/qoriq-fman3-0-10g-0.dtsi
-@@ -0,0 +1,43 @@
-+// SPDX-License-Identifier: (GPL-2.0+ OR BSD-3-Clause)
-+/*
-+ * QorIQ FMan v3 10g port #0 device tree
-+ *
-+ * Copyright 2012-2015 Freescale Semiconductor Inc.
-+ *
-+ */
-+
-+fman@1a00000 {
-+ fman0_rx_0x10: port@90000 {
-+ cell-index = <0x10>;
-+ compatible = "fsl,fman-v3-port-rx", "fsl,fman-port-10g-rx";
-+ reg = <0x90000 0x1000>;
-+ fsl,fman-10g-port;
-+ };
-+
-+ fman0_tx_0x30: port@b0000 {
-+ cell-index = <0x30>;
-+ compatible = "fsl,fman-v3-port-tx", "fsl,fman-port-10g-tx";
-+ reg = <0xb0000 0x1000>;
-+ fsl,fman-10g-port;
-+ fsl,qman-channel-id = <0x800>;
-+ };
-+
-+ ethernet@f0000 {
-+ cell-index = <0x8>;
-+ compatible = "fsl,fman-memac";
-+ reg = <0xf0000 0x1000>;
-+ fsl,fman-ports = <&fman0_rx_0x10 &fman0_tx_0x30>;
-+ pcsphy-handle = <&pcsphy6>;
-+ };
-+
-+ mdio@f1000 {
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ compatible = "fsl,fman-memac-mdio", "fsl,fman-xmdio";
-+ reg = <0xf1000 0x1000>;
-+
-+ pcsphy6: ethernet-phy@0 {
-+ reg = <0x0>;
-+ };
-+ };
-+};
---- /dev/null
-+++ b/arch/arm64/boot/dts/freescale/qoriq-fman3-0-10g-1.dtsi
-@@ -0,0 +1,43 @@
-+// SPDX-License-Identifier: (GPL-2.0+ OR BSD-3-Clause)
-+/*
-+ * QorIQ FMan v3 10g port #1 device tree
-+ *
-+ * Copyright 2012-2015 Freescale Semiconductor Inc.
-+ *
-+ */
-+
-+fman@1a00000 {
-+ fman0_rx_0x11: port@91000 {
-+ cell-index = <0x11>;
-+ compatible = "fsl,fman-v3-port-rx", "fsl,fman-port-10g-rx";
-+ reg = <0x91000 0x1000>;
-+ fsl,fman-10g-port;
-+ };
-+
-+ fman0_tx_0x31: port@b1000 {
-+ cell-index = <0x31>;
-+ compatible = "fsl,fman-v3-port-tx", "fsl,fman-port-10g-tx";
-+ reg = <0xb1000 0x1000>;
-+ fsl,fman-10g-port;
-+ fsl,qman-channel-id = <0x801>;
-+ };
-+
-+ ethernet@f2000 {
-+ cell-index = <0x9>;
-+ compatible = "fsl,fman-memac";
-+ reg = <0xf2000 0x1000>;
-+ fsl,fman-ports = <&fman0_rx_0x11 &fman0_tx_0x31>;
-+ pcsphy-handle = <&pcsphy7>;
-+ };
-+
-+ mdio@f3000 {
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ compatible = "fsl,fman-memac-mdio", "fsl,fman-xmdio";
-+ reg = <0xf3000 0x1000>;
-+
-+ pcsphy7: ethernet-phy@0 {
-+ reg = <0x0>;
-+ };
-+ };
-+};
---- /dev/null
-+++ b/arch/arm64/boot/dts/freescale/qoriq-fman3-0-1g-0.dtsi
-@@ -0,0 +1,42 @@
-+// SPDX-License-Identifier: (GPL-2.0+ OR BSD-3-Clause)
-+/*
-+ * QorIQ FMan v3 1g port #0 device tree
-+ *
-+ * Copyright 2012-2015 Freescale Semiconductor Inc.
-+ *
-+ */
-+
-+fman@1a00000 {
-+ fman0_rx_0x08: port@88000 {
-+ cell-index = <0x8>;
-+ compatible = "fsl,fman-v3-port-rx", "fsl,fman-port-1g-rx";
-+ reg = <0x88000 0x1000>;
-+ };
-+
-+ fman0_tx_0x28: port@a8000 {
-+ cell-index = <0x28>;
-+ compatible = "fsl,fman-v3-port-tx", "fsl,fman-port-1g-tx";
-+ reg = <0xa8000 0x1000>;
-+ fsl,qman-channel-id = <0x802>;
-+ };
-+
-+ ethernet@e0000 {
-+ cell-index = <0>;
-+ compatible = "fsl,fman-memac";
-+ reg = <0xe0000 0x1000>;
-+ fsl,fman-ports = <&fman0_rx_0x08 &fman0_tx_0x28>;
-+ ptp-timer = <&ptp_timer0>;
-+ pcsphy-handle = <&pcsphy0>;
-+ };
-+
-+ mdio@e1000 {
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ compatible = "fsl,fman-memac-mdio", "fsl,fman-xmdio";
-+ reg = <0xe1000 0x1000>;
-+
-+ pcsphy0: ethernet-phy@0 {
-+ reg = <0x0>;
-+ };
-+ };
-+};
---- /dev/null
-+++ b/arch/arm64/boot/dts/freescale/qoriq-fman3-0-1g-1.dtsi
-@@ -0,0 +1,42 @@
-+// SPDX-License-Identifier: (GPL-2.0+ OR BSD-3-Clause)
-+/*
-+ * QorIQ FMan v3 1g port #1 device tree
-+ *
-+ * Copyright 2012-2015 Freescale Semiconductor Inc.
-+ *
-+ */
-+
-+fman@1a00000 {
-+ fman0_rx_0x09: port@89000 {
-+ cell-index = <0x9>;
-+ compatible = "fsl,fman-v3-port-rx", "fsl,fman-port-1g-rx";
-+ reg = <0x89000 0x1000>;
-+ };
-+
-+ fman0_tx_0x29: port@a9000 {
-+ cell-index = <0x29>;
-+ compatible = "fsl,fman-v3-port-tx", "fsl,fman-port-1g-tx";
-+ reg = <0xa9000 0x1000>;
-+ fsl,qman-channel-id = <0x803>;
-+ };
-+
-+ ethernet@e2000 {
-+ cell-index = <1>;
-+ compatible = "fsl,fman-memac";
-+ reg = <0xe2000 0x1000>;
-+ fsl,fman-ports = <&fman0_rx_0x09 &fman0_tx_0x29>;
-+ ptp-timer = <&ptp_timer0>;
-+ pcsphy-handle = <&pcsphy1>;
-+ };
-+
-+ mdio@e3000 {
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ compatible = "fsl,fman-memac-mdio", "fsl,fman-xmdio";
-+ reg = <0xe3000 0x1000>;
-+
-+ pcsphy1: ethernet-phy@0 {
-+ reg = <0x0>;
-+ };
-+ };
-+};
---- /dev/null
-+++ b/arch/arm64/boot/dts/freescale/qoriq-fman3-0-1g-2.dtsi
-@@ -0,0 +1,42 @@
-+// SPDX-License-Identifier: (GPL-2.0+ OR BSD-3-Clause)
-+/*
-+ * QorIQ FMan v3 1g port #2 device tree
-+ *
-+ * Copyright 2012-2015 Freescale Semiconductor Inc.
-+ *
-+ */
-+
-+fman@1a00000 {
-+ fman0_rx_0x0a: port@8a000 {
-+ cell-index = <0xa>;
-+ compatible = "fsl,fman-v3-port-rx", "fsl,fman-port-1g-rx";
-+ reg = <0x8a000 0x1000>;
-+ };
-+
-+ fman0_tx_0x2a: port@aa000 {
-+ cell-index = <0x2a>;
-+ compatible = "fsl,fman-v3-port-tx", "fsl,fman-port-1g-tx";
-+ reg = <0xaa000 0x1000>;
-+ fsl,qman-channel-id = <0x804>;
-+ };
-+
-+ ethernet@e4000 {
-+ cell-index = <2>;
-+ compatible = "fsl,fman-memac";
-+ reg = <0xe4000 0x1000>;
-+ fsl,fman-ports = <&fman0_rx_0x0a &fman0_tx_0x2a>;
-+ ptp-timer = <&ptp_timer0>;
-+ pcsphy-handle = <&pcsphy2>;
-+ };
-+
-+ mdio@e5000 {
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ compatible = "fsl,fman-memac-mdio", "fsl,fman-xmdio";
-+ reg = <0xe5000 0x1000>;
-+
-+ pcsphy2: ethernet-phy@0 {
-+ reg = <0x0>;
-+ };
-+ };
-+};
---- /dev/null
-+++ b/arch/arm64/boot/dts/freescale/qoriq-fman3-0-1g-3.dtsi
-@@ -0,0 +1,42 @@
-+// SPDX-License-Identifier: (GPL-2.0+ OR BSD-3-Clause)
-+/*
-+ * QorIQ FMan v3 1g port #3 device tree
-+ *
-+ * Copyright 2012-2015 Freescale Semiconductor Inc.
-+ *
-+ */
-+
-+fman@1a00000 {
-+ fman0_rx_0x0b: port@8b000 {
-+ cell-index = <0xb>;
-+ compatible = "fsl,fman-v3-port-rx", "fsl,fman-port-1g-rx";
-+ reg = <0x8b000 0x1000>;
-+ };
-+
-+ fman0_tx_0x2b: port@ab000 {
-+ cell-index = <0x2b>;
-+ compatible = "fsl,fman-v3-port-tx", "fsl,fman-port-1g-tx";
-+ reg = <0xab000 0x1000>;
-+ fsl,qman-channel-id = <0x805>;
-+ };
-+
-+ ethernet@e6000 {
-+ cell-index = <3>;
-+ compatible = "fsl,fman-memac";
-+ reg = <0xe6000 0x1000>;
-+ fsl,fman-ports = <&fman0_rx_0x0b &fman0_tx_0x2b>;
-+ ptp-timer = <&ptp_timer0>;
-+ pcsphy-handle = <&pcsphy3>;
-+ };
-+
-+ mdio@e7000 {
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ compatible = "fsl,fman-memac-mdio", "fsl,fman-xmdio";
-+ reg = <0xe7000 0x1000>;
-+
-+ pcsphy3: ethernet-phy@0 {
-+ reg = <0x0>;
-+ };
-+ };
-+};
---- /dev/null
-+++ b/arch/arm64/boot/dts/freescale/qoriq-fman3-0-1g-4.dtsi
-@@ -0,0 +1,42 @@
-+// SPDX-License-Identifier: (GPL-2.0+ OR BSD-3-Clause)
-+/*
-+ * QorIQ FMan v3 1g port #4 device tree
-+ *
-+ * Copyright 2012-2015 Freescale Semiconductor Inc.
-+ *
-+ */
-+
-+fman@1a00000 {
-+ fman0_rx_0x0c: port@8c000 {
-+ cell-index = <0xc>;
-+ compatible = "fsl,fman-v3-port-rx", "fsl,fman-port-1g-rx";
-+ reg = <0x8c000 0x1000>;
-+ };
-+
-+ fman0_tx_0x2c: port@ac000 {
-+ cell-index = <0x2c>;
-+ compatible = "fsl,fman-v3-port-tx", "fsl,fman-port-1g-tx";
-+ reg = <0xac000 0x1000>;
-+ fsl,qman-channel-id = <0x806>;
-+ };
-+
-+ ethernet@e8000 {
-+ cell-index = <4>;
-+ compatible = "fsl,fman-memac";
-+ reg = <0xe8000 0x1000>;
-+ fsl,fman-ports = <&fman0_rx_0x0c &fman0_tx_0x2c>;
-+ ptp-timer = <&ptp_timer0>;
-+ pcsphy-handle = <&pcsphy4>;
-+ };
-+
-+ mdio@e9000 {
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ compatible = "fsl,fman-memac-mdio", "fsl,fman-xmdio";
-+ reg = <0xe9000 0x1000>;
-+
-+ pcsphy4: ethernet-phy@0 {
-+ reg = <0x0>;
-+ };
-+ };
-+};
---- /dev/null
-+++ b/arch/arm64/boot/dts/freescale/qoriq-fman3-0-1g-5.dtsi
-@@ -0,0 +1,42 @@
-+// SPDX-License-Identifier: (GPL-2.0+ OR BSD-3-Clause)
-+/*
-+ * QorIQ FMan v3 1g port #5 device tree
-+ *
-+ * Copyright 2012-2015 Freescale Semiconductor Inc.
-+ *
-+ */
-+
-+fman@1a00000 {
-+ fman0_rx_0x0d: port@8d000 {
-+ cell-index = <0xd>;
-+ compatible = "fsl,fman-v3-port-rx", "fsl,fman-port-1g-rx";
-+ reg = <0x8d000 0x1000>;
-+ };
-+
-+ fman0_tx_0x2d: port@ad000 {
-+ cell-index = <0x2d>;
-+ compatible = "fsl,fman-v3-port-tx", "fsl,fman-port-1g-tx";
-+ reg = <0xad000 0x1000>;
-+ fsl,qman-channel-id = <0x807>;
-+ };
-+
-+ ethernet@ea000 {
-+ cell-index = <5>;
-+ compatible = "fsl,fman-memac";
-+ reg = <0xea000 0x1000>;
-+ fsl,fman-ports = <&fman0_rx_0x0d &fman0_tx_0x2d>;
-+ ptp-timer = <&ptp_timer0>;
-+ pcsphy-handle = <&pcsphy5>;
-+ };
-+
-+ mdio@eb000 {
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ compatible = "fsl,fman-memac-mdio", "fsl,fman-xmdio";
-+ reg = <0xeb000 0x1000>;
-+
-+ pcsphy5: ethernet-phy@0 {
-+ reg = <0x0>;
-+ };
-+ };
-+};
---- /dev/null
-+++ b/arch/arm64/boot/dts/freescale/qoriq-fman3-0-6oh.dtsi
-@@ -0,0 +1,47 @@
-+/*
-+ * QorIQ FMan v3 OH ports device tree
-+ *
-+ * Copyright 2012-2015 Freescale Semiconductor Inc.
-+ *
-+ * SPDX-License-Identifier: (GPL-2.0+ OR BSD-3-Clause)
-+ */
-+
-+fman@1a00000 {
-+
-+ fman0_oh1: port@82000 {
-+ cell-index = <0>;
-+ compatible = "fsl,fman-port-oh";
-+ reg = <0x82000 0x1000>;
-+ };
-+
-+ fman0_oh2: port@83000 {
-+ cell-index = <1>;
-+ compatible = "fsl,fman-port-oh";
-+ reg = <0x83000 0x1000>;
-+ };
-+
-+ fman0_oh3: port@84000 {
-+ cell-index = <2>;
-+ compatible = "fsl,fman-port-oh";
-+ reg = <0x84000 0x1000>;
-+ };
-+
-+ fman0_oh4: port@85000 {
-+ cell-index = <3>;
-+ compatible = "fsl,fman-port-oh";
-+ reg = <0x85000 0x1000>;
-+ };
-+
-+ fman0_oh5: port@86000 {
-+ cell-index = <4>;
-+ compatible = "fsl,fman-port-oh";
-+ reg = <0x86000 0x1000>;
-+ };
-+
-+ fman0_oh6: port@87000 {
-+ cell-index = <5>;
-+ compatible = "fsl,fman-port-oh";
-+ reg = <0x87000 0x1000>;
-+ };
-+
-+};
---- /dev/null
-+++ b/arch/arm64/boot/dts/freescale/qoriq-fman3-0.dtsi
-@@ -0,0 +1,130 @@
-+// SPDX-License-Identifier: (GPL-2.0+ OR BSD-3-Clause)
-+/*
-+ * QorIQ FMan v3 device tree
-+ *
-+ * Copyright 2012-2015 Freescale Semiconductor Inc.
-+ *
-+ */
-+
-+fman0: fman@1a00000 {
-+ #address-cells = <1>;
-+ #size-cells = <1>;
-+ cell-index = <0>;
-+ compatible = "fsl,fman";
-+ ranges = <0x0 0x00 0x1a00000 0x100000>;
-+ reg = <0x0 0x1a00000 0x0 0x100000>;
-+ interrupts = <0 44 0x4>, <0 45 0x4>;
-+ clocks = <&clockgen 3 0>;
-+ clock-names = "fmanclk";
-+ fsl,qman-channel-range = <0x800 0x10>;
-+
-+ cc {
-+ compatible = "fsl,fman-cc";
-+ };
-+
-+ muram@0 {
-+ compatible = "fsl,fman-muram";
-+ reg = <0x0 0x60000>;
-+ };
-+
-+ bmi@80000 {
-+ compatible = "fsl,fman-bmi";
-+ reg = <0x80000 0x400>;
-+ };
-+
-+ qmi@80400 {
-+ compatible = "fsl,fman-qmi";
-+ reg = <0x80400 0x400>;
-+ };
-+
-+ fman0_oh_0x2: port@82000 {
-+ cell-index = <0x2>;
-+ compatible = "fsl,fman-v3-port-oh";
-+ reg = <0x82000 0x1000>;
-+ fsl,qman-channel-id = <0x809>;
-+ };
-+
-+ fman0_oh_0x3: port@83000 {
-+ cell-index = <0x3>;
-+ compatible = "fsl,fman-v3-port-oh";
-+ reg = <0x83000 0x1000>;
-+ fsl,qman-channel-id = <0x80a>;
-+ };
-+
-+ fman0_oh_0x4: port@84000 {
-+ cell-index = <0x4>;
-+ compatible = "fsl,fman-v3-port-oh";
-+ reg = <0x84000 0x1000>;
-+ fsl,qman-channel-id = <0x80b>;
-+ };
-+
-+ fman0_oh_0x5: port@85000 {
-+ cell-index = <0x5>;
-+ compatible = "fsl,fman-v3-port-oh";
-+ reg = <0x85000 0x1000>;
-+ fsl,qman-channel-id = <0x80c>;
-+ };
-+
-+ fman0_oh_0x6: port@86000 {
-+ cell-index = <0x6>;
-+ compatible = "fsl,fman-v3-port-oh";
-+ reg = <0x86000 0x1000>;
-+ fsl,qman-channel-id = <0x80d>;
-+ };
-+
-+ fman0_oh_0x7: port@87000 {
-+ cell-index = <0x7>;
-+ compatible = "fsl,fman-v3-port-oh";
-+ reg = <0x87000 0x1000>;
-+ fsl,qman-channel-id = <0x80e>;
-+ };
-+
-+ policer@c0000 {
-+ compatible = "fsl,fman-policer";
-+ reg = <0xc0000 0x1000>;
-+ };
-+
-+ keygen@c1000 {
-+ compatible = "fsl,fman-keygen";
-+ reg = <0xc1000 0x1000>;
-+ };
-+
-+ dma@c2000 {
-+ compatible = "fsl,fman-dma";
-+ reg = <0xc2000 0x1000>;
-+ };
-+
-+ fpm@c3000 {
-+ compatible = "fsl,fman-fpm";
-+ reg = <0xc3000 0x1000>;
-+ };
-+
-+ parser@c7000 {
-+ compatible = "fsl,fman-parser";
-+ reg = <0xc7000 0x1000>;
-+ };
-+
-+ vsps@dc000 {
-+ compatible = "fsl,fman-vsps";
-+ reg = <0xdc000 0x1000>;
-+ };
-+
-+ mdio0: mdio@fc000 {
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ compatible = "fsl,fman-memac-mdio", "fsl,fman-xmdio";
-+ reg = <0xfc000 0x1000>;
-+ };
-+
-+ xmdio0: mdio@fd000 {
-+ #address-cells = <1>;
-+ #size-cells = <0>;
-+ compatible = "fsl,fman-memac-mdio", "fsl,fman-xmdio";
-+ reg = <0xfd000 0x1000>;
-+ };
-+
-+ ptp_timer0: ptp-timer@fe000 {
-+ compatible = "fsl,fman-ptp-timer", "fsl,fman-rtc";
-+ reg = <0xfe000 0x1000>;
-+ };
-+};
---- /dev/null
-+++ b/arch/arm64/boot/dts/freescale/qoriq-qman-portals-sdk.dtsi
-@@ -0,0 +1,38 @@
-+/*
-+ * QorIQ QMan SDK Portals device tree nodes
-+ *
-+ * Copyright 2011-2016 Freescale Semiconductor Inc.
-+ * Copyright 2017 NXP
-+ *
-+ * SPDX-License-Identifier: (GPL-2.0+ OR BSD-3-Clause)
-+ */
-+
-+&qportals {
-+ qman-fqids@0 {
-+ compatible = "fsl,fqid-range";
-+ fsl,fqid-range = <256 256>;
-+ };
-+
-+ qman-fqids@1 {
-+ compatible = "fsl,fqid-range";
-+ fsl,fqid-range = <32768 32768>;
-+ };
-+
-+ qman-pools@0 {
-+ compatible = "fsl,pool-channel-range";
-+ fsl,pool-channel-range = <0x401 0xf>;
-+ };
-+
-+ qman-cgrids@0 {
-+ compatible = "fsl,cgrid-range";
-+ fsl,cgrid-range = <0 256>;
-+ };
-+
-+ qman-ceetm@0 {
-+ compatible = "fsl,qman-ceetm";
-+ fsl,ceetm-lfqid-range = <0xf00000 0x1000>;
-+ fsl,ceetm-sp-range = <0 16>;
-+ fsl,ceetm-lni-range = <0 8>;
-+ fsl,ceetm-channel-range = <0 32>;
-+ };
-+};
---- /dev/null
-+++ b/arch/arm64/boot/dts/freescale/qoriq-qman-portals.dtsi
-@@ -0,0 +1,87 @@
-+// SPDX-License-Identifier: (GPL-2.0+ OR BSD-3-Clause)
-+/*
-+ * QorIQ QMan Portals device tree
-+ *
-+ * Copyright 2011-2016 Freescale Semiconductor Inc.
-+ *
-+ */
-+
-+&qportals {
-+ #address-cells = <1>;
-+ #size-cells = <1>;
-+ compatible = "simple-bus";
-+
-+ qportal0: qman-portal@0 {
-+ /*
-+ * bootloader fix-ups are expected to provide the
-+ * "fsl,bman-portal-<hardware revision>" compatible
-+ */
-+ compatible = "fsl,qman-portal";
-+ reg = <0x0 0x4000>, <0x4000000 0x4000>;
-+ interrupts = <GIC_SPI 172 IRQ_TYPE_LEVEL_HIGH>;
-+ cell-index = <0>;
-+ };
-+
-+ qportal1: qman-portal@10000 {
-+ compatible = "fsl,qman-portal";
-+ reg = <0x10000 0x4000>, <0x4010000 0x4000>;
-+ interrupts = <GIC_SPI 174 IRQ_TYPE_LEVEL_HIGH>;
-+ cell-index = <1>;
-+ };
-+
-+ qportal2: qman-portal@20000 {
-+ compatible = "fsl,qman-portal";
-+ reg = <0x20000 0x4000>, <0x4020000 0x4000>;
-+ interrupts = <GIC_SPI 176 IRQ_TYPE_LEVEL_HIGH>;
-+ cell-index = <2>;
-+ };
-+
-+ qportal3: qman-portal@30000 {
-+ compatible = "fsl,qman-portal";
-+ reg = <0x30000 0x4000>, <0x4030000 0x4000>;
-+ interrupts = <GIC_SPI 178 IRQ_TYPE_LEVEL_HIGH>;
-+ cell-index = <3>;
-+ };
-+
-+ qportal4: qman-portal@40000 {
-+ compatible = "fsl,qman-portal";
-+ reg = <0x40000 0x4000>, <0x4040000 0x4000>;
-+ interrupts = <GIC_SPI 180 IRQ_TYPE_LEVEL_HIGH>;
-+ cell-index = <4>;
-+ };
-+
-+ qportal5: qman-portal@50000 {
-+ compatible = "fsl,qman-portal";
-+ reg = <0x50000 0x4000>, <0x4050000 0x4000>;
-+ interrupts = <GIC_SPI 182 IRQ_TYPE_LEVEL_HIGH>;
-+ cell-index = <5>;
-+ };
-+
-+ qportal6: qman-portal@60000 {
-+ compatible = "fsl,qman-portal";
-+ reg = <0x60000 0x4000>, <0x4060000 0x4000>;
-+ interrupts = <GIC_SPI 184 IRQ_TYPE_LEVEL_HIGH>;
-+ cell-index = <6>;
-+ };
-+
-+ qportal7: qman-portal@70000 {
-+ compatible = "fsl,qman-portal";
-+ reg = <0x70000 0x4000>, <0x4070000 0x4000>;
-+ interrupts = <GIC_SPI 186 IRQ_TYPE_LEVEL_HIGH>;
-+ cell-index = <7>;
-+ };
-+
-+ qportal8: qman-portal@80000 {
-+ compatible = "fsl,qman-portal";
-+ reg = <0x80000 0x4000>, <0x4080000 0x4000>;
-+ interrupts = <GIC_SPI 188 IRQ_TYPE_LEVEL_HIGH>;
-+ cell-index = <8>;
-+ };
-+
-+ qportal9: qman-portal@90000 {
-+ compatible = "fsl,qman-portal";
-+ reg = <0x90000 0x4000>, <0x4090000 0x4000>;
-+ interrupts = <GIC_SPI 190 IRQ_TYPE_LEVEL_HIGH>;
-+ cell-index = <9>;
-+ };
-+};
---- a/arch/powerpc/boot/dts/fsl/qoriq-bman1-portals.dtsi
-+++ b/arch/powerpc/boot/dts/fsl/qoriq-bman1-portals.dtsi
-@@ -38,51 +38,61 @@
- compatible = "simple-bus";
-
- bman-portal@0 {
-+ cell-index = <0>;
- compatible = "fsl,bman-portal";
- reg = <0x0 0x4000>, <0x100000 0x1000>;
- interrupts = <105 2 0 0>;
- };
- bman-portal@4000 {
-+ cell-index = <1>;
- compatible = "fsl,bman-portal";
- reg = <0x4000 0x4000>, <0x101000 0x1000>;
- interrupts = <107 2 0 0>;
- };
- bman-portal@8000 {
-+ cell-index = <2>;
- compatible = "fsl,bman-portal";
- reg = <0x8000 0x4000>, <0x102000 0x1000>;
- interrupts = <109 2 0 0>;
- };
- bman-portal@c000 {
-+ cell-index = <3>;
- compatible = "fsl,bman-portal";
- reg = <0xc000 0x4000>, <0x103000 0x1000>;
- interrupts = <111 2 0 0>;
- };
- bman-portal@10000 {
-+ cell-index = <4>;
- compatible = "fsl,bman-portal";
- reg = <0x10000 0x4000>, <0x104000 0x1000>;
- interrupts = <113 2 0 0>;
- };
- bman-portal@14000 {
-+ cell-index = <5>;
- compatible = "fsl,bman-portal";
- reg = <0x14000 0x4000>, <0x105000 0x1000>;
- interrupts = <115 2 0 0>;
- };
- bman-portal@18000 {
-+ cell-index = <6>;
- compatible = "fsl,bman-portal";
- reg = <0x18000 0x4000>, <0x106000 0x1000>;
- interrupts = <117 2 0 0>;
- };
- bman-portal@1c000 {
-+ cell-index = <7>;
- compatible = "fsl,bman-portal";
- reg = <0x1c000 0x4000>, <0x107000 0x1000>;
- interrupts = <119 2 0 0>;
- };
- bman-portal@20000 {
-+ cell-index = <8>;
- compatible = "fsl,bman-portal";
- reg = <0x20000 0x4000>, <0x108000 0x1000>;
- interrupts = <121 2 0 0>;
- };
- bman-portal@24000 {
-+ cell-index = <9>;
- compatible = "fsl,bman-portal";
- reg = <0x24000 0x4000>, <0x109000 0x1000>;
- interrupts = <123 2 0 0>;
---- a/arch/powerpc/boot/dts/fsl/qoriq-fman3-0-10g-0.dtsi
-+++ b/arch/powerpc/boot/dts/fsl/qoriq-fman3-0-10g-0.dtsi
-@@ -35,14 +35,14 @@
- fman@400000 {
- fman0_rx_0x10: port@90000 {
- cell-index = <0x10>;
-- compatible = "fsl,fman-v3-port-rx";
-+ compatible = "fsl,fman-v3-port-rx", "fsl,fman-port-10g-rx";
- reg = <0x90000 0x1000>;
- fsl,fman-10g-port;
- };
-
- fman0_tx_0x30: port@b0000 {
- cell-index = <0x30>;
-- compatible = "fsl,fman-v3-port-tx";
-+ compatible = "fsl,fman-v3-port-tx", "fsl,fman-port-10g-tx";
- reg = <0xb0000 0x1000>;
- fsl,fman-10g-port;
- };
---- a/arch/powerpc/boot/dts/fsl/qoriq-fman3-0-10g-1.dtsi
-+++ b/arch/powerpc/boot/dts/fsl/qoriq-fman3-0-10g-1.dtsi
-@@ -35,14 +35,14 @@
- fman@400000 {
- fman0_rx_0x11: port@91000 {
- cell-index = <0x11>;
-- compatible = "fsl,fman-v3-port-rx";
-+ compatible = "fsl,fman-v3-port-rx", "fsl,fman-port-10g-rx";
- reg = <0x91000 0x1000>;
- fsl,fman-10g-port;
- };
-
- fman0_tx_0x31: port@b1000 {
- cell-index = <0x31>;
-- compatible = "fsl,fman-v3-port-tx";
-+ compatible = "fsl,fman-v3-port-tx", "fsl,fman-port-10g-tx";
- reg = <0xb1000 0x1000>;
- fsl,fman-10g-port;
- };
diff --git a/target/linux/layerscape/patches-4.9/303-dts-layerscape-add-traverse-ls1043.patch b/target/linux/layerscape/patches-4.9/303-dts-layerscape-add-traverse-ls1043.patch
deleted file mode 100644
index 6232e56782..0000000000
--- a/target/linux/layerscape/patches-4.9/303-dts-layerscape-add-traverse-ls1043.patch
+++ /dev/null
@@ -1,95 +0,0 @@
-From c0612164b379ebc8964da6bc6f6ced9736dce488 Mon Sep 17 00:00:00 2001
-From: Mathew McBride <matt@traverse.com.au>
-Date: Tue, 17 Apr 2018 10:01:03 +1000
-Subject: [PATCH] add DTS for Traverse LS1043 Boards
-
-Signed-off-by: Mathew McBride <matt@traverse.com.au>
----
- arch/arm64/boot/dts/freescale/Makefile | 5 ++++-
- 1 file changed, 4 insertions(+), 1 deletion(-)
- create mode 100644 arch/arm64/boot/dts/freescale/traverse-ls1043v.dts
-
---- a/arch/arm64/boot/dts/freescale/Makefile
-+++ b/arch/arm64/boot/dts/freescale/Makefile
-@@ -21,7 +21,10 @@ dtb-$(CONFIG_ARCH_LAYERSCAPE) += fsl-ls2
- dtb-$(CONFIG_ARCH_LAYERSCAPE) += fsl-ls2080a-simu.dtb
- dtb-$(CONFIG_ARCH_LAYERSCAPE) += fsl-ls2088a-qds.dtb
- dtb-$(CONFIG_ARCH_LAYERSCAPE) += fsl-ls2088a-rdb.dtb
--
-+
-+dtb-$(CONFIG_ARCH_LAYERSCAPE) += traverse-ls1043v.dtb
-+dtb-$(CONFIG_ARCH_LAYERSCAPE) += traverse-ls1043s.dtb
-+
- always := $(dtb-y)
- subdir-y := $(dts-dirs)
- clean-files := *.dtb
---- a/arch/arm64/boot/dts/freescale/traverse-ls1043s.dts
-+++ b/arch/arm64/boot/dts/freescale/traverse-ls1043s.dts
-@@ -330,3 +330,32 @@
- &sata {
- status = "disabled";
- };
-+
-+/* Additions for Layerscape SDK (4.4/4.9) Kernel only
-+ * These kernels need additional setup for FMan/QMan DMA shared memory
-+ */
-+
-+#include "qoriq-qman-portals-sdk.dtsi"
-+#include "qoriq-bman-portals-sdk.dtsi"
-+
-+&bman_fbpr {
-+ compatible = "fsl,bman-fbpr";
-+ alloc-ranges = <0 0 0x10000 0>;
-+};
-+&qman_fqd {
-+ compatible = "fsl,qman-fqd";
-+ alloc-ranges = <0 0 0x10000 0>;
-+};
-+&qman_pfdr {
-+ compatible = "fsl,qman-pfdr";
-+ alloc-ranges = <0 0 0x10000 0>;
-+};
-+
-+&soc {
-+#include "qoriq-dpaa-eth.dtsi"
-+#include "qoriq-fman3-0-6oh.dtsi"
-+};
-+
-+&fman0 {
-+ compatible = "fsl,fman", "simple-bus";
-+};
---- a/arch/arm64/boot/dts/freescale/traverse-ls1043v.dts
-+++ b/arch/arm64/boot/dts/freescale/traverse-ls1043v.dts
-@@ -251,3 +251,32 @@
- &sata {
- status = "disabled";
- };
-+
-+/* Additions for Layerscape SDK (4.4/4.9) Kernel only
-+ * These kernels need additional setup for FMan/QMan DMA shared memory
-+ */
-+
-+#include "qoriq-qman-portals-sdk.dtsi"
-+#include "qoriq-bman-portals-sdk.dtsi"
-+
-+&bman_fbpr {
-+ compatible = "fsl,bman-fbpr";
-+ alloc-ranges = <0 0 0x10000 0>;
-+};
-+&qman_fqd {
-+ compatible = "fsl,qman-fqd";
-+ alloc-ranges = <0 0 0x10000 0>;
-+};
-+&qman_pfdr {
-+ compatible = "fsl,qman-pfdr";
-+ alloc-ranges = <0 0 0x10000 0>;
-+};
-+
-+&soc {
-+#include "qoriq-dpaa-eth.dtsi"
-+#include "qoriq-fman3-0-6oh.dtsi"
-+};
-+
-+&fman0 {
-+ compatible = "fsl,fman", "simple-bus";
-+};
diff --git a/target/linux/layerscape/patches-4.9/401-mtd-spi-nor-support-layerscape.patch b/target/linux/layerscape/patches-4.9/401-mtd-spi-nor-support-layerscape.patch
deleted file mode 100644
index 89139d93b8..0000000000
--- a/target/linux/layerscape/patches-4.9/401-mtd-spi-nor-support-layerscape.patch
+++ /dev/null
@@ -1,993 +0,0 @@
-From c03c545e064a81515fe109ddcc4ecb3895528e58 Mon Sep 17 00:00:00 2001
-From: Yangbo Lu <yangbo.lu@nxp.com>
-Date: Fri, 6 Jul 2018 15:32:05 +0800
-Subject: [PATCH] mtd: spi-nor: support layerscape
-
-This is an integrated patch for layerscape qspi support.
-
-Signed-off-by: Suresh Gupta <suresh.gupta@nxp.com>
-Signed-off-by: Yunhui Cui <B56489@freescale.com>
-Signed-off-by: mar.krzeminski <mar.krzeminski@gmail.com>
-Signed-off-by: Alison Wang <b18965@freescale.com>
-Signed-off-by: Nobuhiro Iwamatsu <nobuhiro.iwamatsu.kw@hitachi.com>
-Signed-off-by: LABBE Corentin <clabbe.montjoie@gmail.com>
-Signed-off-by: Yuan Yao <yao.yuan@nxp.com>
-Signed-off-by: Alexander Kurz <akurz@blala.de>
-Signed-off-by: L. D. Pinney <ldpinney@gmail.com>
-Signed-off-by: Ash Benz <ash.benz@bk.ru>
-Signed-off-by: Yangbo Lu <yangbo.lu@nxp.com>
----
- drivers/mtd/devices/m25p80.c | 3 +-
- drivers/mtd/mtdchar.c | 2 +-
- drivers/mtd/spi-nor/fsl-quadspi.c | 327 ++++++++++++++++++++++++------
- drivers/mtd/spi-nor/spi-nor.c | 141 ++++++++++++-
- include/linux/mtd/spi-nor.h | 14 +-
- 5 files changed, 416 insertions(+), 71 deletions(-)
-
---- a/drivers/mtd/mtdchar.c
-+++ b/drivers/mtd/mtdchar.c
-@@ -455,7 +455,7 @@ static int mtdchar_readoob(struct file *
- * data. For our userspace tools it is important to dump areas
- * with ECC errors!
- * For kernel internal usage it also might return -EUCLEAN
-- * to signal the caller that a bitflip has occured and has
-+ * to signal the caller that a bitflip has occurred and has
- * been corrected by the ECC algorithm.
- *
- * Note: currently the standard NAND function, nand_read_oob_std,
---- a/drivers/mtd/spi-nor/fsl-quadspi.c
-+++ b/drivers/mtd/spi-nor/fsl-quadspi.c
-@@ -41,6 +41,8 @@
- #define QUADSPI_QUIRK_TKT253890 (1 << 2)
- /* Controller cannot wake up from wait mode, TKT245618 */
- #define QUADSPI_QUIRK_TKT245618 (1 << 3)
-+/* QSPI_AMBA_BASE is internally added by SOC design */
-+#define QUADSPI_AMBA_BASE_INTERNAL (0x10000)
-
- /* The registers */
- #define QUADSPI_MCR 0x00
-@@ -193,7 +195,7 @@
- #define QUADSPI_LUT_NUM 64
-
- /* SEQID -- we can have 16 seqids at most. */
--#define SEQID_QUAD_READ 0
-+#define SEQID_READ 0
- #define SEQID_WREN 1
- #define SEQID_WRDI 2
- #define SEQID_RDSR 3
-@@ -205,15 +207,22 @@
- #define SEQID_RDCR 9
- #define SEQID_EN4B 10
- #define SEQID_BRWR 11
-+#define SEQID_RDAR_OR_RD_EVCR 12
-+#define SEQID_WRAR 13
-+#define SEQID_WD_EVCR 14
-
- #define QUADSPI_MIN_IOMAP SZ_4M
-
-+#define FLASH_VENDOR_SPANSION_FS "s25fs"
-+#define SPANSION_S25FS_FAMILY (1 << 1)
-+
- enum fsl_qspi_devtype {
- FSL_QUADSPI_VYBRID,
- FSL_QUADSPI_IMX6SX,
- FSL_QUADSPI_IMX7D,
- FSL_QUADSPI_IMX6UL,
- FSL_QUADSPI_LS1021A,
-+ FSL_QUADSPI_LS2080A,
- };
-
- struct fsl_qspi_devtype_data {
-@@ -224,7 +233,7 @@ struct fsl_qspi_devtype_data {
- int driver_data;
- };
-
--static struct fsl_qspi_devtype_data vybrid_data = {
-+static const struct fsl_qspi_devtype_data vybrid_data = {
- .devtype = FSL_QUADSPI_VYBRID,
- .rxfifo = 128,
- .txfifo = 64,
-@@ -232,7 +241,7 @@ static struct fsl_qspi_devtype_data vybr
- .driver_data = QUADSPI_QUIRK_SWAP_ENDIAN,
- };
-
--static struct fsl_qspi_devtype_data imx6sx_data = {
-+static const struct fsl_qspi_devtype_data imx6sx_data = {
- .devtype = FSL_QUADSPI_IMX6SX,
- .rxfifo = 128,
- .txfifo = 512,
-@@ -241,7 +250,7 @@ static struct fsl_qspi_devtype_data imx6
- | QUADSPI_QUIRK_TKT245618,
- };
-
--static struct fsl_qspi_devtype_data imx7d_data = {
-+static const struct fsl_qspi_devtype_data imx7d_data = {
- .devtype = FSL_QUADSPI_IMX7D,
- .rxfifo = 512,
- .txfifo = 512,
-@@ -250,7 +259,7 @@ static struct fsl_qspi_devtype_data imx7
- | QUADSPI_QUIRK_4X_INT_CLK,
- };
-
--static struct fsl_qspi_devtype_data imx6ul_data = {
-+static const struct fsl_qspi_devtype_data imx6ul_data = {
- .devtype = FSL_QUADSPI_IMX6UL,
- .rxfifo = 128,
- .txfifo = 512,
-@@ -267,6 +276,14 @@ static struct fsl_qspi_devtype_data ls10
- .driver_data = 0,
- };
-
-+static struct fsl_qspi_devtype_data ls2080a_data = {
-+ .devtype = FSL_QUADSPI_LS2080A,
-+ .rxfifo = 128,
-+ .txfifo = 64,
-+ .ahb_buf_size = 1024,
-+ .driver_data = QUADSPI_AMBA_BASE_INTERNAL | QUADSPI_QUIRK_TKT253890,
-+};
-+
- #define FSL_QSPI_MAX_CHIP 4
- struct fsl_qspi {
- struct spi_nor nor[FSL_QSPI_MAX_CHIP];
-@@ -282,6 +299,7 @@ struct fsl_qspi {
- u32 nor_size;
- u32 nor_num;
- u32 clk_rate;
-+ u32 ddr_smp;
- unsigned int chip_base_addr; /* We may support two chips. */
- bool has_second_chip;
- bool big_endian;
-@@ -309,6 +327,23 @@ static inline int needs_wakeup_wait_mode
- return q->devtype_data->driver_data & QUADSPI_QUIRK_TKT245618;
- }
-
-+static inline int has_added_amba_base_internal(struct fsl_qspi *q)
-+{
-+ return q->devtype_data->driver_data & QUADSPI_AMBA_BASE_INTERNAL;
-+}
-+
-+static u32 fsl_get_nor_vendor(struct spi_nor *nor)
-+{
-+ u32 vendor_id;
-+
-+ if (nor->vendor) {
-+ if (memcmp(nor->vendor, FLASH_VENDOR_SPANSION_FS,
-+ sizeof(FLASH_VENDOR_SPANSION_FS) - 1))
-+ vendor_id = SPANSION_S25FS_FAMILY;
-+ }
-+ return vendor_id;
-+}
-+
- /*
- * R/W functions for big- or little-endian registers:
- * The qSPI controller's endian is independent of the CPU core's endian.
-@@ -331,6 +366,31 @@ static u32 qspi_readl(struct fsl_qspi *q
- return ioread32(addr);
- }
-
-+static inline u32 *u8tou32(u32 *dest, const u8 *src, size_t n)
-+{
-+ size_t i;
-+ *dest = 0;
-+
-+ n = n > 4 ? 4 : n;
-+ for (i = 0; i < n; i++)
-+ *dest |= *src++ << i * 8;
-+
-+ return dest;
-+
-+}
-+
-+static inline u8 *u32tou8(u8 *dest, const u32 *src, size_t n)
-+{
-+ size_t i;
-+ u8 *xdest = dest;
-+
-+ n = n > 4 ? 4 : n;
-+ for (i = 0; i < n; i++)
-+ *xdest++ = *src >> i * 8;
-+
-+ return dest;
-+}
-+
- /*
- * An IC bug makes us to re-arrange the 32-bit data.
- * The following chips, such as IMX6SLX, have fixed this bug.
-@@ -373,8 +433,15 @@ static void fsl_qspi_init_lut(struct fsl
- void __iomem *base = q->iobase;
- int rxfifo = q->devtype_data->rxfifo;
- u32 lut_base;
-- u8 cmd, addrlen, dummy;
- int i;
-+ u32 vendor;
-+
-+ struct spi_nor *nor = &q->nor[0];
-+ u8 addrlen = (nor->addr_width == 3) ? ADDR24BIT : ADDR32BIT;
-+ u8 read_op = nor->read_opcode;
-+ u8 read_dm = nor->read_dummy;
-+
-+ vendor = fsl_get_nor_vendor(nor);
-
- fsl_qspi_unlock_lut(q);
-
-@@ -382,24 +449,50 @@ static void fsl_qspi_init_lut(struct fsl
- for (i = 0; i < QUADSPI_LUT_NUM; i++)
- qspi_writel(q, 0, base + QUADSPI_LUT_BASE + i * 4);
-
-- /* Quad Read */
-- lut_base = SEQID_QUAD_READ * 4;
-+ /* Read */
-+ lut_base = SEQID_READ * 4;
-
-- if (q->nor_size <= SZ_16M) {
-- cmd = SPINOR_OP_READ_1_1_4;
-- addrlen = ADDR24BIT;
-- dummy = 8;
-- } else {
-- /* use the 4-byte address */
-- cmd = SPINOR_OP_READ_1_1_4;
-- addrlen = ADDR32BIT;
-- dummy = 8;
-- }
--
-- qspi_writel(q, LUT0(CMD, PAD1, cmd) | LUT1(ADDR, PAD1, addrlen),
-+ if (nor->flash_read == SPI_NOR_FAST) {
-+ qspi_writel(q, LUT0(CMD, PAD1, read_op) |
-+ LUT1(ADDR, PAD1, addrlen),
-+ base + QUADSPI_LUT(lut_base));
-+ qspi_writel(q, LUT0(DUMMY, PAD1, read_dm) |
-+ LUT1(FSL_READ, PAD1, rxfifo),
-+ base + QUADSPI_LUT(lut_base + 1));
-+ } else if (nor->flash_read == SPI_NOR_QUAD) {
-+ if (q->nor_size == 0x4000000) {
-+ read_op = 0xEC;
-+ qspi_writel(q,
-+ LUT0(CMD, PAD1, read_op) | LUT1(ADDR, PAD4, addrlen),
- base + QUADSPI_LUT(lut_base));
-- qspi_writel(q, LUT0(DUMMY, PAD1, dummy) | LUT1(FSL_READ, PAD4, rxfifo),
-+ qspi_writel(q,
-+ LUT0(MODE, PAD4, 0xff) | LUT1(DUMMY, PAD4, read_dm),
- base + QUADSPI_LUT(lut_base + 1));
-+ qspi_writel(q,
-+ LUT0(FSL_READ, PAD4, rxfifo),
-+ base + QUADSPI_LUT(lut_base + 2));
-+ } else {
-+ qspi_writel(q, LUT0(CMD, PAD1, read_op) |
-+ LUT1(ADDR, PAD1, addrlen),
-+ base + QUADSPI_LUT(lut_base));
-+ qspi_writel(q, LUT0(DUMMY, PAD1, read_dm) |
-+ LUT1(FSL_READ, PAD4, rxfifo),
-+ base + QUADSPI_LUT(lut_base + 1));
-+ }
-+ } else if (nor->flash_read == SPI_NOR_DDR_QUAD) {
-+ /* read mode : 1-4-4, such as Spansion s25fl128s. */
-+ qspi_writel(q, LUT0(CMD, PAD1, read_op)
-+ | LUT1(ADDR_DDR, PAD4, addrlen),
-+ base + QUADSPI_LUT(lut_base));
-+
-+ qspi_writel(q, LUT0(MODE_DDR, PAD4, 0xff)
-+ | LUT1(DUMMY, PAD1, read_dm),
-+ base + QUADSPI_LUT(lut_base + 1));
-+
-+ qspi_writel(q, LUT0(FSL_READ_DDR, PAD4, rxfifo)
-+ | LUT1(JMP_ON_CS, PAD1, 0),
-+ base + QUADSPI_LUT(lut_base + 2));
-+ }
-
- /* Write enable */
- lut_base = SEQID_WREN * 4;
-@@ -409,16 +502,8 @@ static void fsl_qspi_init_lut(struct fsl
- /* Page Program */
- lut_base = SEQID_PP * 4;
-
-- if (q->nor_size <= SZ_16M) {
-- cmd = SPINOR_OP_PP;
-- addrlen = ADDR24BIT;
-- } else {
-- /* use the 4-byte address */
-- cmd = SPINOR_OP_PP;
-- addrlen = ADDR32BIT;
-- }
--
-- qspi_writel(q, LUT0(CMD, PAD1, cmd) | LUT1(ADDR, PAD1, addrlen),
-+ qspi_writel(q, LUT0(CMD, PAD1, nor->program_opcode) |
-+ LUT1(ADDR, PAD1, addrlen),
- base + QUADSPI_LUT(lut_base));
- qspi_writel(q, LUT0(FSL_WRITE, PAD1, 0),
- base + QUADSPI_LUT(lut_base + 1));
-@@ -432,10 +517,8 @@ static void fsl_qspi_init_lut(struct fsl
- /* Erase a sector */
- lut_base = SEQID_SE * 4;
-
-- cmd = q->nor[0].erase_opcode;
-- addrlen = q->nor_size <= SZ_16M ? ADDR24BIT : ADDR32BIT;
--
-- qspi_writel(q, LUT0(CMD, PAD1, cmd) | LUT1(ADDR, PAD1, addrlen),
-+ qspi_writel(q, LUT0(CMD, PAD1, nor->erase_opcode) |
-+ LUT1(ADDR, PAD1, addrlen),
- base + QUADSPI_LUT(lut_base));
-
- /* Erase the whole chip */
-@@ -476,6 +559,44 @@ static void fsl_qspi_init_lut(struct fsl
- qspi_writel(q, LUT0(CMD, PAD1, SPINOR_OP_BRWR),
- base + QUADSPI_LUT(lut_base));
-
-+
-+ /*
-+ * Flash Micron and Spansion command confilict
-+ * use the same value 0x65. But it indicates different meaning.
-+ */
-+ lut_base = SEQID_RDAR_OR_RD_EVCR * 4;
-+
-+ if (vendor == SPANSION_S25FS_FAMILY) {
-+ /*
-+ * Read any device register.
-+ * Used for Spansion S25FS-S family flash only.
-+ */
-+ qspi_writel(q, LUT0(CMD, PAD1, SPINOR_OP_SPANSION_RDAR) |
-+ LUT1(ADDR, PAD1, ADDR24BIT),
-+ base + QUADSPI_LUT(lut_base));
-+ qspi_writel(q, LUT0(DUMMY, PAD1, 8) | LUT1(FSL_READ, PAD1, 1),
-+ base + QUADSPI_LUT(lut_base + 1));
-+ } else {
-+ qspi_writel(q, LUT0(CMD, PAD1, SPINOR_OP_RD_EVCR),
-+ base + QUADSPI_LUT(lut_base));
-+ }
-+
-+ /*
-+ * Write any device register.
-+ * Used for Spansion S25FS-S family flash only.
-+ */
-+ lut_base = SEQID_WRAR * 4;
-+ qspi_writel(q, LUT0(CMD, PAD1, SPINOR_OP_SPANSION_WRAR) |
-+ LUT1(ADDR, PAD1, ADDR24BIT),
-+ base + QUADSPI_LUT(lut_base));
-+ qspi_writel(q, LUT0(FSL_WRITE, PAD1, 1),
-+ base + QUADSPI_LUT(lut_base + 1));
-+
-+ /* Write EVCR register */
-+ lut_base = SEQID_WD_EVCR * 4;
-+ qspi_writel(q, LUT0(CMD, PAD1, SPINOR_OP_WD_EVCR),
-+ base + QUADSPI_LUT(lut_base));
-+
- fsl_qspi_lock_lut(q);
- }
-
-@@ -483,8 +604,24 @@ static void fsl_qspi_init_lut(struct fsl
- static int fsl_qspi_get_seqid(struct fsl_qspi *q, u8 cmd)
- {
- switch (cmd) {
-+ case SPINOR_OP_READ_1_4_4_D:
-+ case SPINOR_OP_READ4_1_4_4_D:
-+ case SPINOR_OP_READ4_1_1_4:
- case SPINOR_OP_READ_1_1_4:
-- return SEQID_QUAD_READ;
-+ case SPINOR_OP_READ_FAST:
-+ case SPINOR_OP_READ4_FAST:
-+ return SEQID_READ;
-+ /*
-+ * Spansion & Micron use the same command value 0x65
-+ * Spansion: SPINOR_OP_SPANSION_RDAR, read any register.
-+ * Micron: SPINOR_OP_RD_EVCR,
-+ * read enhanced volatile configuration register.
-+ * case SPINOR_OP_RD_EVCR:
-+ */
-+ case SPINOR_OP_SPANSION_RDAR:
-+ return SEQID_RDAR_OR_RD_EVCR;
-+ case SPINOR_OP_SPANSION_WRAR:
-+ return SEQID_WRAR;
- case SPINOR_OP_WREN:
- return SEQID_WREN;
- case SPINOR_OP_WRDI:
-@@ -496,6 +633,7 @@ static int fsl_qspi_get_seqid(struct fsl
- case SPINOR_OP_CHIP_ERASE:
- return SEQID_CHIP_ERASE;
- case SPINOR_OP_PP:
-+ case SPINOR_OP_PP_4B:
- return SEQID_PP;
- case SPINOR_OP_RDID:
- return SEQID_RDID;
-@@ -507,6 +645,8 @@ static int fsl_qspi_get_seqid(struct fsl
- return SEQID_EN4B;
- case SPINOR_OP_BRWR:
- return SEQID_BRWR;
-+ case SPINOR_OP_WD_EVCR:
-+ return SEQID_WD_EVCR;
- default:
- if (cmd == q->nor[0].erase_opcode)
- return SEQID_SE;
-@@ -531,8 +671,11 @@ fsl_qspi_runcmd(struct fsl_qspi *q, u8 c
- /* save the reg */
- reg = qspi_readl(q, base + QUADSPI_MCR);
-
-- qspi_writel(q, q->memmap_phy + q->chip_base_addr + addr,
-- base + QUADSPI_SFAR);
-+ if (has_added_amba_base_internal(q))
-+ qspi_writel(q, q->chip_base_addr + addr, base + QUADSPI_SFAR);
-+ else
-+ qspi_writel(q, q->memmap_phy + q->chip_base_addr + addr,
-+ base + QUADSPI_SFAR);
- qspi_writel(q, QUADSPI_RBCT_WMRK_MASK | QUADSPI_RBCT_RXBRD_USEIPS,
- base + QUADSPI_RBCT);
- qspi_writel(q, reg | QUADSPI_MCR_CLR_RXF_MASK, base + QUADSPI_MCR);
-@@ -582,10 +725,10 @@ static void fsl_qspi_read_data(struct fs
- q->chip_base_addr, tmp);
-
- if (len >= 4) {
-- *((u32 *)rxbuf) = tmp;
-+ u32tou8(rxbuf, &tmp, 4);
- rxbuf += 4;
- } else {
-- memcpy(rxbuf, &tmp, len);
-+ u32tou8(rxbuf, &tmp, len);
- break;
- }
-
-@@ -619,11 +762,12 @@ static inline void fsl_qspi_invalid(stru
- }
-
- static ssize_t fsl_qspi_nor_write(struct fsl_qspi *q, struct spi_nor *nor,
-- u8 opcode, unsigned int to, u32 *txbuf,
-+ u8 opcode, unsigned int to, u8 *txbuf,
- unsigned count)
- {
- int ret, i, j;
- u32 tmp;
-+ u8 byts;
-
- dev_dbg(q->dev, "to 0x%.8x:0x%.8x, len : %d\n",
- q->chip_base_addr, to, count);
-@@ -633,10 +777,13 @@ static ssize_t fsl_qspi_nor_write(struct
- qspi_writel(q, tmp | QUADSPI_MCR_CLR_TXF_MASK, q->iobase + QUADSPI_MCR);
-
- /* fill the TX data to the FIFO */
-+ byts = count;
- for (j = 0, i = ((count + 3) / 4); j < i; j++) {
-- tmp = fsl_qspi_endian_xchg(q, *txbuf);
-+ u8tou32(&tmp, txbuf, byts);
-+ tmp = fsl_qspi_endian_xchg(q, tmp);
- qspi_writel(q, tmp, q->iobase + QUADSPI_TBDR);
-- txbuf++;
-+ txbuf += 4;
-+ byts -= 4;
- }
-
- /* fill the TXFIFO upto 16 bytes for i.MX7d */
-@@ -657,11 +804,43 @@ static void fsl_qspi_set_map_addr(struct
- {
- int nor_size = q->nor_size;
- void __iomem *base = q->iobase;
-+ u32 mem_base;
-+
-+ if (has_added_amba_base_internal(q))
-+ mem_base = 0x0;
-+ else
-+ mem_base = q->memmap_phy;
-+
-+ qspi_writel(q, nor_size + mem_base, base + QUADSPI_SFA1AD);
-+ qspi_writel(q, nor_size * 2 + mem_base, base + QUADSPI_SFA2AD);
-+ qspi_writel(q, nor_size * 3 + mem_base, base + QUADSPI_SFB1AD);
-+ qspi_writel(q, nor_size * 4 + mem_base, base + QUADSPI_SFB2AD);
-+}
-+
-+/*
-+ * enable controller ddr quad mode to support different
-+ * vender flashes ddr quad mode.
-+ */
-+static void set_ddr_quad_mode(struct fsl_qspi *q)
-+{
-+ u32 reg, reg2;
-+
-+ reg = qspi_readl(q, q->iobase + QUADSPI_MCR);
-+
-+ /* Firstly, disable the module */
-+ qspi_writel(q, reg | QUADSPI_MCR_MDIS_MASK, q->iobase + QUADSPI_MCR);
-+
-+ /* Set the Sampling Register for DDR */
-+ reg2 = qspi_readl(q, q->iobase + QUADSPI_SMPR);
-+ reg2 &= ~QUADSPI_SMPR_DDRSMP_MASK;
-+ reg2 |= (((q->ddr_smp) << QUADSPI_SMPR_DDRSMP_SHIFT) &
-+ QUADSPI_SMPR_DDRSMP_MASK);
-+ qspi_writel(q, reg2, q->iobase + QUADSPI_SMPR);
-+
-+ /* Enable the module again (enable the DDR too) */
-+ reg |= QUADSPI_MCR_DDR_EN_MASK;
-+ qspi_writel(q, reg, q->iobase + QUADSPI_MCR);
-
-- qspi_writel(q, nor_size + q->memmap_phy, base + QUADSPI_SFA1AD);
-- qspi_writel(q, nor_size * 2 + q->memmap_phy, base + QUADSPI_SFA2AD);
-- qspi_writel(q, nor_size * 3 + q->memmap_phy, base + QUADSPI_SFB1AD);
-- qspi_writel(q, nor_size * 4 + q->memmap_phy, base + QUADSPI_SFB2AD);
- }
-
- /*
-@@ -704,6 +883,11 @@ static void fsl_qspi_init_abh_read(struc
- seqid = fsl_qspi_get_seqid(q, q->nor[0].read_opcode);
- qspi_writel(q, seqid << QUADSPI_BFGENCR_SEQID_SHIFT,
- q->iobase + QUADSPI_BFGENCR);
-+
-+ /* enable the DDR quad read */
-+ if (q->nor->flash_read == SPI_NOR_DDR_QUAD)
-+ set_ddr_quad_mode(q);
-+
- }
-
- /* This function was used to prepare and enable QSPI clock */
-@@ -822,6 +1006,7 @@ static const struct of_device_id fsl_qsp
- { .compatible = "fsl,imx7d-qspi", .data = (void *)&imx7d_data, },
- { .compatible = "fsl,imx6ul-qspi", .data = (void *)&imx6ul_data, },
- { .compatible = "fsl,ls1021a-qspi", .data = (void *)&ls1021a_data, },
-+ { .compatible = "fsl,ls2080a-qspi", .data = (void *)&ls2080a_data, },
- { /* sentinel */ }
- };
- MODULE_DEVICE_TABLE(of, fsl_qspi_dt_ids);
-@@ -835,8 +1020,12 @@ static int fsl_qspi_read_reg(struct spi_
- {
- int ret;
- struct fsl_qspi *q = nor->priv;
-+ u32 to = 0;
-
-- ret = fsl_qspi_runcmd(q, opcode, 0, len);
-+ if (opcode == SPINOR_OP_SPANSION_RDAR)
-+ u8tou32(&to, nor->cmd_buf, 4);
-+
-+ ret = fsl_qspi_runcmd(q, opcode, to, len);
- if (ret)
- return ret;
-
-@@ -848,9 +1037,13 @@ static int fsl_qspi_write_reg(struct spi
- {
- struct fsl_qspi *q = nor->priv;
- int ret;
-+ u32 to = 0;
-+
-+ if (opcode == SPINOR_OP_SPANSION_WRAR)
-+ u8tou32(&to, nor->cmd_buf, 4);
-
- if (!buf) {
-- ret = fsl_qspi_runcmd(q, opcode, 0, 1);
-+ ret = fsl_qspi_runcmd(q, opcode, to, 1);
- if (ret)
- return ret;
-
-@@ -859,7 +1052,7 @@ static int fsl_qspi_write_reg(struct spi
-
- } else if (len > 0) {
- ret = fsl_qspi_nor_write(q, nor, opcode, 0,
-- (u32 *)buf, len);
-+ buf, len);
- if (ret > 0)
- return 0;
- } else {
-@@ -875,7 +1068,7 @@ static ssize_t fsl_qspi_write(struct spi
- {
- struct fsl_qspi *q = nor->priv;
- ssize_t ret = fsl_qspi_nor_write(q, nor, nor->program_opcode, to,
-- (u32 *)buf, len);
-+ (u8 *)buf, len);
-
- /* invalid the data in the AHB buffer. */
- fsl_qspi_invalid(q);
-@@ -922,7 +1115,7 @@ static ssize_t fsl_qspi_read(struct spi_
- len);
-
- /* Read out the data directly from the AHB buffer.*/
-- memcpy(buf, q->ahb_addr + q->chip_base_addr + from - q->memmap_offs,
-+ memcpy_toio(buf, q->ahb_addr + q->chip_base_addr + from - q->memmap_offs,
- len);
-
- return len;
-@@ -980,6 +1173,8 @@ static int fsl_qspi_probe(struct platfor
- struct spi_nor *nor;
- struct mtd_info *mtd;
- int ret, i = 0;
-+ int find_node;
-+ enum read_mode mode = SPI_NOR_QUAD;
-
- q = devm_kzalloc(dev, sizeof(*q), GFP_KERNEL);
- if (!q)
-@@ -1027,6 +1222,12 @@ static int fsl_qspi_probe(struct platfor
- goto clk_failed;
- }
-
-+ /* find ddrsmp value */
-+ ret = of_property_read_u32(dev->of_node, "fsl,ddr-sampling-point",
-+ &q->ddr_smp);
-+ if (ret)
-+ q->ddr_smp = 0;
-+
- /* find the irq */
- ret = platform_get_irq(pdev, 0);
- if (ret < 0) {
-@@ -1050,6 +1251,7 @@ static int fsl_qspi_probe(struct platfor
-
- mutex_init(&q->lock);
-
-+ find_node = 0;
- /* iterate the subnodes. */
- for_each_available_child_of_node(dev->of_node, np) {
- /* skip the holes */
-@@ -1076,18 +1278,25 @@ static int fsl_qspi_probe(struct platfor
- ret = of_property_read_u32(np, "spi-max-frequency",
- &q->clk_rate);
- if (ret < 0)
-- goto mutex_failed;
-+ continue;
-
- /* set the chip address for READID */
- fsl_qspi_set_base_addr(q, nor);
-
-- ret = spi_nor_scan(nor, NULL, SPI_NOR_QUAD);
-+ ret = of_property_read_bool(np, "m25p,fast-read");
-+ mode = (ret) ? SPI_NOR_FAST : SPI_NOR_QUAD;
-+ /* Can we enable the DDR Quad Read? */
-+ ret = of_property_read_bool(np, "ddr-quad-read");
- if (ret)
-- goto mutex_failed;
-+ mode = SPI_NOR_DDR_QUAD;
-+
-+ ret = spi_nor_scan(nor, NULL, mode);
-+ if (ret)
-+ continue;
-
- ret = mtd_device_register(mtd, NULL, 0);
- if (ret)
-- goto mutex_failed;
-+ continue;
-
- /* Set the correct NOR size now. */
- if (q->nor_size == 0) {
-@@ -1110,8 +1319,12 @@ static int fsl_qspi_probe(struct platfor
- nor->page_size = q->devtype_data->txfifo;
-
- i++;
-+ find_node++;
- }
-
-+ if (find_node == 0)
-+ goto mutex_failed;
-+
- /* finish the rest init. */
- ret = fsl_qspi_nor_setup_last(q);
- if (ret)
---- a/drivers/mtd/spi-nor/spi-nor.c
-+++ b/drivers/mtd/spi-nor/spi-nor.c
-@@ -40,6 +40,13 @@
- #define SPI_NOR_MAX_ID_LEN 6
- #define SPI_NOR_MAX_ADDR_WIDTH 4
-
-+#define SPI_NOR_MICRON_WRITE_ENABLE 0x7f
-+/* Added for S25FS-S family flash */
-+#define SPINOR_CONFIG_REG3_OFFSET 0x800004
-+#define CR3V_4KB_ERASE_UNABLE 0x8
-+#define SPINOR_S25FS_FAMILY_ID 0x81
-+
-+
- struct flash_info {
- char *name;
-
-@@ -68,7 +75,8 @@ struct flash_info {
- #define SECT_4K_PMC BIT(4) /* SPINOR_OP_BE_4K_PMC works uniformly */
- #define SPI_NOR_DUAL_READ BIT(5) /* Flash supports Dual Read */
- #define SPI_NOR_QUAD_READ BIT(6) /* Flash supports Quad Read */
--#define USE_FSR BIT(7) /* use flag status register */
-+#define USE_FSR BIT(13) /* use flag status register */
-+#define SPI_NOR_DDR_QUAD_READ BIT(7) /* Flash supports DDR Quad Read */
- #define SPI_NOR_HAS_LOCK BIT(8) /* Flash supports lock/unlock via SR */
- #define SPI_NOR_HAS_TB BIT(9) /*
- * Flash SR has Top/Bottom (TB) protect
-@@ -85,9 +93,11 @@ struct flash_info {
- * Use dedicated 4byte address op codes
- * to support memory size above 128Mib.
- */
-+#define NO_CHIP_ERASE BIT(12) /* Chip does not support chip erase */
- };
-
- #define JEDEC_MFR(info) ((info)->id[0])
-+#define EXT_ID(info) ((info)->id[5])
-
- static const struct flash_info *spi_nor_match_id(const char *name);
-
-@@ -132,7 +142,7 @@ static int read_fsr(struct spi_nor *nor)
- /*
- * Read configuration register, returning its value in the
- * location. Return the configuration register value.
-- * Returns negative if error occured.
-+ * Returns negative if error occurred.
- */
- static int read_cr(struct spi_nor *nor)
- {
-@@ -160,6 +170,8 @@ static inline int spi_nor_read_dummy_cyc
- case SPI_NOR_DUAL:
- case SPI_NOR_QUAD:
- return 8;
-+ case SPI_NOR_DDR_QUAD:
-+ return 6;
- case SPI_NOR_NORMAL:
- return 0;
- }
-@@ -962,6 +974,8 @@ static const struct flash_info spi_nor_i
-
- /* ESMT */
- { "f25l32pa", INFO(0x8c2016, 0, 64 * 1024, 64, SECT_4K | SPI_NOR_HAS_LOCK) },
-+ { "f25l32qa", INFO(0x8c4116, 0, 64 * 1024, 64, SECT_4K | SPI_NOR_HAS_LOCK) },
-+ { "f25l64qa", INFO(0x8c4117, 0, 64 * 1024, 128, SECT_4K | SPI_NOR_HAS_LOCK) },
-
- /* Everspin */
- { "mr25h256", CAT25_INFO( 32 * 1024, 1, 256, 2, SPI_NOR_NO_ERASE | SPI_NOR_NO_FR) },
-@@ -1021,12 +1035,15 @@ static const struct flash_info spi_nor_i
- { "mx25l3205d", INFO(0xc22016, 0, 64 * 1024, 64, SECT_4K) },
- { "mx25l3255e", INFO(0xc29e16, 0, 64 * 1024, 64, SECT_4K) },
- { "mx25l6405d", INFO(0xc22017, 0, 64 * 1024, 128, SECT_4K) },
-+ { "mx25u2033e", INFO(0xc22532, 0, 64 * 1024, 4, SECT_4K) },
-+ { "mx25u4035", INFO(0xc22533, 0, 64 * 1024, 8, SECT_4K) },
-+ { "mx25u8035", INFO(0xc22534, 0, 64 * 1024, 16, SECT_4K) },
- { "mx25u3235f", INFO(0xc22536, 0, 64 * 1024, 64, 0) },
- { "mx25u6435f", INFO(0xc22537, 0, 64 * 1024, 128, SECT_4K) },
- { "mx25l12805d", INFO(0xc22018, 0, 64 * 1024, 256, 0) },
- { "mx25l12855e", INFO(0xc22618, 0, 64 * 1024, 256, 0) },
- { "mx25l25635e", INFO(0xc22019, 0, 64 * 1024, 512, 0) },
-- { "mx25u25635f", INFO(0xc22539, 0, 64 * 1024, 512, SECT_4K | SPI_NOR_4B_OPCODES) },
-+ { "mx25u25635f", INFO(0xc22539, 0, 64 * 1024, 512, SECT_4K) },
- { "mx25l25655e", INFO(0xc22619, 0, 64 * 1024, 512, 0) },
- { "mx66l51235l", INFO(0xc2201a, 0, 64 * 1024, 1024, SPI_NOR_QUAD_READ) },
- { "mx66l1g55g", INFO(0xc2261b, 0, 64 * 1024, 2048, SPI_NOR_QUAD_READ) },
-@@ -1040,10 +1057,11 @@ static const struct flash_info spi_nor_i
- { "n25q128a11", INFO(0x20bb18, 0, 64 * 1024, 256, SECT_4K | SPI_NOR_QUAD_READ) },
- { "n25q128a13", INFO(0x20ba18, 0, 64 * 1024, 256, SECT_4K | SPI_NOR_QUAD_READ) },
- { "n25q256a", INFO(0x20ba19, 0, 64 * 1024, 512, SECT_4K | SPI_NOR_QUAD_READ) },
-+ { "n25q256ax1", INFO(0x20bb19, 0, 64 * 1024, 512, SECT_4K | SPI_NOR_QUAD_READ) },
- { "n25q512a", INFO(0x20bb20, 0, 64 * 1024, 1024, SECT_4K | USE_FSR | SPI_NOR_QUAD_READ) },
- { "n25q512ax3", INFO(0x20ba20, 0, 64 * 1024, 1024, SECT_4K | USE_FSR | SPI_NOR_QUAD_READ) },
-- { "n25q00", INFO(0x20ba21, 0, 64 * 1024, 2048, SECT_4K | USE_FSR | SPI_NOR_QUAD_READ) },
-- { "n25q00a", INFO(0x20bb21, 0, 64 * 1024, 2048, SECT_4K | USE_FSR | SPI_NOR_QUAD_READ) },
-+ { "n25q00", INFO(0x20ba21, 0, 64 * 1024, 2048, SECT_4K | USE_FSR | SPI_NOR_QUAD_READ | NO_CHIP_ERASE) },
-+ { "n25q00a", INFO(0x20bb21, 0, 64 * 1024, 2048, SECT_4K | USE_FSR | SPI_NOR_QUAD_READ | NO_CHIP_ERASE) },
-
- /* PMC */
- { "pm25lv512", INFO(0, 0, 32 * 1024, 2, SECT_4K_PMC) },
-@@ -1061,8 +1079,11 @@ static const struct flash_info spi_nor_i
- { "s70fl01gs", INFO(0x010221, 0x4d00, 256 * 1024, 256, 0) },
- { "s25sl12800", INFO(0x012018, 0x0300, 256 * 1024, 64, 0) },
- { "s25sl12801", INFO(0x012018, 0x0301, 64 * 1024, 256, 0) },
-- { "s25fl128s", INFO6(0x012018, 0x4d0180, 64 * 1024, 256, SPI_NOR_DUAL_READ | SPI_NOR_QUAD_READ) },
-+ { "s25fs256s1", INFO6(0x010219, 0x4d0181, 64 * 1024, 512, 0)},
-+ { "s25fl128s", INFO6(0x012018, 0x4d0180, 64 * 1024, 256, SPI_NOR_QUAD_READ
-+ | SPI_NOR_DDR_QUAD_READ) },
- { "s25fl129p0", INFO(0x012018, 0x4d00, 256 * 1024, 64, SPI_NOR_DUAL_READ | SPI_NOR_QUAD_READ) },
-+ { "s25fs512s", INFO6(0x010220, 0x4d0081, 256 * 1024, 256, SPI_NOR_DUAL_READ | SPI_NOR_QUAD_READ)},
- { "s25fl129p1", INFO(0x012018, 0x4d01, 64 * 1024, 256, SPI_NOR_DUAL_READ | SPI_NOR_QUAD_READ) },
- { "s25sl004a", INFO(0x010212, 0, 64 * 1024, 8, 0) },
- { "s25sl008a", INFO(0x010213, 0, 64 * 1024, 16, 0) },
-@@ -1136,7 +1157,15 @@ static const struct flash_info spi_nor_i
- { "w25x40", INFO(0xef3013, 0, 64 * 1024, 8, SECT_4K) },
- { "w25x80", INFO(0xef3014, 0, 64 * 1024, 16, SECT_4K) },
- { "w25x16", INFO(0xef3015, 0, 64 * 1024, 32, SECT_4K) },
-+ {
-+ "w25q16dw", INFO(0xef6015, 0, 64 * 1024, 32,
-+ SECT_4K | SPI_NOR_DUAL_READ | SPI_NOR_QUAD_READ |
-+ SPI_NOR_HAS_LOCK | SPI_NOR_HAS_TB)
-+ },
- { "w25x32", INFO(0xef3016, 0, 64 * 1024, 64, SECT_4K) },
-+ { "w25q20cl", INFO(0xef4012, 0, 64 * 1024, 4, SECT_4K) },
-+ { "w25q20bw", INFO(0xef5012, 0, 64 * 1024, 4, SECT_4K) },
-+ { "w25q20ew", INFO(0xef6012, 0, 64 * 1024, 4, SECT_4K) },
- { "w25q32", INFO(0xef4016, 0, 64 * 1024, 64, SECT_4K) },
- {
- "w25q32dw", INFO(0xef6016, 0, 64 * 1024, 64,
-@@ -1208,6 +1237,53 @@ static const struct flash_info *spi_nor_
- id[0], id[1], id[2]);
- return ERR_PTR(-ENODEV);
- }
-+/*
-+ * The S25FS-S family physical sectors may be configured as a
-+ * hybrid combination of eight 4-kB parameter sectors
-+ * at the top or bottom of the address space with all
-+ * but one of the remaining sectors being uniform size.
-+ * The Parameter Sector Erase commands (20h or 21h) must
-+ * be used to erase the 4-kB parameter sectors individually.
-+ * The Sector (uniform sector) Erase commands (D8h or DCh)
-+ * must be used to erase any of the remaining
-+ * sectors, including the portion of highest or lowest address
-+ * sector that is not overlaid by the parameter sectors.
-+ * The uniform sector erase command has no effect on parameter sectors.
-+ */
-+static int spansion_s25fs_disable_4kb_erase(struct spi_nor *nor)
-+{
-+ struct fsl_qspi *q;
-+ u32 cr3v_addr = SPINOR_CONFIG_REG3_OFFSET;
-+ u8 cr3v = 0x0;
-+ int ret = 0x0;
-+
-+ q = nor->priv;
-+
-+ nor->cmd_buf[2] = cr3v_addr >> 16;
-+ nor->cmd_buf[1] = cr3v_addr >> 8;
-+ nor->cmd_buf[0] = cr3v_addr >> 0;
-+
-+ ret = nor->read_reg(nor, SPINOR_OP_SPANSION_RDAR, &cr3v, 1);
-+ if (ret)
-+ return ret;
-+ if (cr3v & CR3V_4KB_ERASE_UNABLE)
-+ return 0;
-+ ret = nor->write_reg(nor, SPINOR_OP_WREN, NULL, 0);
-+ if (ret)
-+ return ret;
-+ cr3v = CR3V_4KB_ERASE_UNABLE;
-+ nor->program_opcode = SPINOR_OP_SPANSION_WRAR;
-+ nor->write(nor, cr3v_addr, 1, &cr3v);
-+
-+ ret = nor->read_reg(nor, SPINOR_OP_SPANSION_RDAR, &cr3v, 1);
-+ if (ret)
-+ return ret;
-+ if (!(cr3v & CR3V_4KB_ERASE_UNABLE))
-+ return -EPERM;
-+
-+ return 0;
-+}
-+
-
- static int spi_nor_read(struct mtd_info *mtd, loff_t from, size_t len,
- size_t *retlen, u_char *buf)
-@@ -1427,7 +1503,7 @@ static int macronix_quad_enable(struct s
- * Write status Register and configuration register with 2 bytes
- * The first byte will be written to the status register, while the
- * second byte will be written to the configuration register.
-- * Return negative if error occured.
-+ * Return negative if error occurred.
- */
- static int write_sr_cr(struct spi_nor *nor, u16 val)
- {
-@@ -1475,6 +1551,24 @@ static int spansion_quad_enable(struct s
- return 0;
- }
-
-+static int set_ddr_quad_mode(struct spi_nor *nor, const struct flash_info *info)
-+{
-+ int status;
-+
-+ switch (JEDEC_MFR(info)) {
-+ case SNOR_MFR_SPANSION:
-+ status = spansion_quad_enable(nor);
-+ if (status) {
-+ dev_err(nor->dev, "Spansion DDR quad-read not enabled\n");
-+ return status;
-+ }
-+ return status;
-+ default:
-+ return -EINVAL;
-+ }
-+}
-+
-+
- static int set_quad_mode(struct spi_nor *nor, const struct flash_info *info)
- {
- int status;
-@@ -1621,9 +1715,25 @@ int spi_nor_scan(struct spi_nor *nor, co
- write_sr(nor, 0);
- spi_nor_wait_till_ready(nor);
- }
-+ if (JEDEC_MFR(info) == SNOR_MFR_MICRON) {
-+ ret = read_sr(nor);
-+ ret &= SPI_NOR_MICRON_WRITE_ENABLE;
-+
-+ write_enable(nor);
-+ write_sr(nor, ret);
-+ }
-+
-+ if (EXT_ID(info) == SPINOR_S25FS_FAMILY_ID) {
-+ ret = spansion_s25fs_disable_4kb_erase(nor);
-+ if (ret)
-+ return ret;
-+ }
-+
-
- if (!mtd->name)
- mtd->name = dev_name(dev);
-+ if (info->name)
-+ nor->vendor = info->name;
- mtd->priv = nor;
- mtd->type = MTD_NORFLASH;
- mtd->writesize = 1;
-@@ -1657,6 +1767,8 @@ int spi_nor_scan(struct spi_nor *nor, co
- nor->flags |= SNOR_F_USE_FSR;
- if (info->flags & SPI_NOR_HAS_TB)
- nor->flags |= SNOR_F_HAS_SR_TB;
-+ if (info->flags & NO_CHIP_ERASE)
-+ nor->flags |= SNOR_F_NO_OP_CHIP_ERASE;
-
- #ifdef CONFIG_MTD_SPI_NOR_USE_4K_SECTORS
- /* prefer "small sector" erase if possible */
-@@ -1696,9 +1808,15 @@ int spi_nor_scan(struct spi_nor *nor, co
- /* Some devices cannot do fast-read, no matter what DT tells us */
- if (info->flags & SPI_NOR_NO_FR)
- nor->flash_read = SPI_NOR_NORMAL;
--
-- /* Quad/Dual-read mode takes precedence over fast/normal */
-- if (mode == SPI_NOR_QUAD && info->flags & SPI_NOR_QUAD_READ) {
-+ /* DDR Quad/Quad/Dual-read mode takes precedence over fast/normal */
-+ if (mode == SPI_NOR_DDR_QUAD && info->flags & SPI_NOR_DDR_QUAD_READ) {
-+ ret = set_ddr_quad_mode(nor, info);
-+ if (ret) {
-+ dev_err(dev, "DDR quad mode not supported\n");
-+ return ret;
-+ }
-+ nor->flash_read = SPI_NOR_DDR_QUAD;
-+ } else if (mode == SPI_NOR_QUAD && info->flags & SPI_NOR_QUAD_READ) {
- ret = set_quad_mode(nor, info);
- if (ret) {
- dev_err(dev, "quad mode not supported\n");
-@@ -1711,6 +1829,9 @@ int spi_nor_scan(struct spi_nor *nor, co
-
- /* Default commands */
- switch (nor->flash_read) {
-+ case SPI_NOR_DDR_QUAD:
-+ nor->read_opcode = SPINOR_OP_READ4_1_4_4_D;
-+ break;
- case SPI_NOR_QUAD:
- nor->read_opcode = SPINOR_OP_READ_1_1_4;
- break;
---- a/include/linux/mtd/spi-nor.h
-+++ b/include/linux/mtd/spi-nor.h
-@@ -31,10 +31,10 @@
-
- /*
- * Note on opcode nomenclature: some opcodes have a format like
-- * SPINOR_OP_FUNCTION{4,}_x_y_z. The numbers x, y, and z stand for the number
-+ * SPINOR_OP_FUNCTION{4,}_x_y_z{_D}. The numbers x, y,and z stand for the number
- * of I/O lines used for the opcode, address, and data (respectively). The
- * FUNCTION has an optional suffix of '4', to represent an opcode which
-- * requires a 4-byte (32-bit) address.
-+ * requires a 4-byte (32-bit) address. The suffix of 'D' stands for the
- */
-
- /* Flash opcodes. */
-@@ -46,7 +46,9 @@
- #define SPINOR_OP_READ_1_1_2 0x3b /* Read data bytes (Dual Output SPI) */
- #define SPINOR_OP_READ_1_2_2 0xbb /* Read data bytes (Dual I/O SPI) */
- #define SPINOR_OP_READ_1_1_4 0x6b /* Read data bytes (Quad Output SPI) */
-+#define SPINOR_OP_READ_1_4_4_D 0xed /* Read data bytes (DDR Quad SPI) */
- #define SPINOR_OP_READ_1_4_4 0xeb /* Read data bytes (Quad I/O SPI) */
-+#define SPINOR_OP_READ4_1_4_4_D 0xee /* Read data bytes (DDR Quad SPI) */
- #define SPINOR_OP_PP 0x02 /* Page program (up to 256 bytes) */
- #define SPINOR_OP_PP_1_1_4 0x32 /* Quad page program */
- #define SPINOR_OP_PP_1_4_4 0x38 /* Quad page program */
-@@ -62,9 +64,11 @@
- /* 4-byte address opcodes - used on Spansion and some Macronix flashes. */
- #define SPINOR_OP_READ_4B 0x13 /* Read data bytes (low frequency) */
- #define SPINOR_OP_READ_FAST_4B 0x0c /* Read data bytes (high frequency) */
-+#define SPINOR_OP_READ4_FAST 0x0c /* Read data bytes (high frequency) */
- #define SPINOR_OP_READ_1_1_2_4B 0x3c /* Read data bytes (Dual Output SPI) */
- #define SPINOR_OP_READ_1_2_2_4B 0xbc /* Read data bytes (Dual I/O SPI) */
- #define SPINOR_OP_READ_1_1_4_4B 0x6c /* Read data bytes (Quad Output SPI) */
-+#define SPINOR_OP_READ4_1_1_4 0x6c /* Read data bytes (Quad SPI) */
- #define SPINOR_OP_READ_1_4_4_4B 0xec /* Read data bytes (Quad I/O SPI) */
- #define SPINOR_OP_PP_4B 0x12 /* Page program (up to 256 bytes) */
- #define SPINOR_OP_PP_1_1_4_4B 0x34 /* Quad page program */
-@@ -94,6 +98,10 @@
- /* Used for Spansion flashes only. */
- #define SPINOR_OP_BRWR 0x17 /* Bank register write */
-
-+/* Used for Spansion S25FS-S family flash only. */
-+#define SPINOR_OP_SPANSION_RDAR 0x65 /* Read any device register */
-+#define SPINOR_OP_SPANSION_WRAR 0x71 /* Write any device register */
-+
- /* Used for Micron flashes only. */
- #define SPINOR_OP_RD_EVCR 0x65 /* Read EVCR register */
- #define SPINOR_OP_WD_EVCR 0x61 /* Write EVCR register */
-@@ -124,6 +132,7 @@ enum read_mode {
- SPI_NOR_FAST,
- SPI_NOR_DUAL,
- SPI_NOR_QUAD,
-+ SPI_NOR_DDR_QUAD,
- };
-
- #define SPI_NOR_MAX_CMD_SIZE 8
-@@ -189,6 +198,7 @@ struct spi_nor {
- bool sst_write_second;
- u32 flags;
- u8 cmd_buf[SPI_NOR_MAX_CMD_SIZE];
-+ char *vendor;
-
- int (*prepare)(struct spi_nor *nor, enum spi_nor_ops ops);
- void (*unprepare)(struct spi_nor *nor, enum spi_nor_ops ops);
diff --git a/target/linux/layerscape/patches-4.9/402-mtd-support-layerscape.patch b/target/linux/layerscape/patches-4.9/402-mtd-support-layerscape.patch
deleted file mode 100644
index 9f44e9e723..0000000000
--- a/target/linux/layerscape/patches-4.9/402-mtd-support-layerscape.patch
+++ /dev/null
@@ -1,392 +0,0 @@
-From d9d0181f74146507026c31cccd52dda27ec3d966 Mon Sep 17 00:00:00 2001
-From: Yangbo Lu <yangbo.lu@nxp.com>
-Date: Wed, 17 Jan 2018 14:57:31 +0800
-Subject: [PATCH 06/30] mtd: support layerscape
-
-This is an integrated patch for layerscape ifc-nor-nand support.
-
-Signed-off-by: Alison Wang <b18965@freescale.com>
-Signed-off-by: Prabhakar Kushwaha <prabhakar.kushwaha@nxp.com>
-Signed-off-by: Yangbo Lu <yangbo.lu@nxp.com>
----
- drivers/memory/Kconfig | 2 +-
- drivers/memory/fsl_ifc.c | 263 ++++++++++++++++++++++++++++++++++++++++
- drivers/mtd/maps/physmap_of.c | 4 +
- drivers/mtd/nand/Kconfig | 2 +-
- drivers/mtd/nand/fsl_ifc_nand.c | 5 +-
- include/linux/fsl_ifc.h | 7 ++
- 6 files changed, 280 insertions(+), 3 deletions(-)
-
---- a/drivers/memory/Kconfig
-+++ b/drivers/memory/Kconfig
-@@ -115,7 +115,7 @@ config FSL_CORENET_CF
-
- config FSL_IFC
- bool
-- depends on FSL_SOC || ARCH_LAYERSCAPE
-+ depends on FSL_SOC || ARCH_LAYERSCAPE || SOC_LS1021A
-
- config JZ4780_NEMC
- bool "Ingenic JZ4780 SoC NEMC driver"
---- a/drivers/memory/fsl_ifc.c
-+++ b/drivers/memory/fsl_ifc.c
-@@ -24,6 +24,7 @@
- #include <linux/compiler.h>
- #include <linux/sched.h>
- #include <linux/spinlock.h>
-+#include <linux/delay.h>
- #include <linux/types.h>
- #include <linux/slab.h>
- #include <linux/io.h>
-@@ -37,6 +38,8 @@
-
- struct fsl_ifc_ctrl *fsl_ifc_ctrl_dev;
- EXPORT_SYMBOL(fsl_ifc_ctrl_dev);
-+#define FSL_IFC_V1_3_0 0x01030000
-+#define IFC_TIMEOUT_MSECS 1000 /* 1000ms */
-
- /*
- * convert_ifc_address - convert the base address
-@@ -311,6 +314,261 @@ err:
- return ret;
- }
-
-+#ifdef CONFIG_PM_SLEEP
-+/* save ifc registers */
-+static int fsl_ifc_suspend(struct device *dev)
-+{
-+ struct fsl_ifc_ctrl *ctrl = dev_get_drvdata(dev);
-+ struct fsl_ifc_global __iomem *fcm = ctrl->gregs;
-+ struct fsl_ifc_runtime __iomem *runtime = ctrl->rregs;
-+ __be32 nand_evter_intr_en, cm_evter_intr_en, nor_evter_intr_en,
-+ gpcm_evter_intr_en;
-+ uint32_t ifc_bank, i;
-+
-+ ctrl->saved_gregs = kzalloc(sizeof(struct fsl_ifc_global), GFP_KERNEL);
-+ if (!ctrl->saved_gregs)
-+ return -ENOMEM;
-+ ctrl->saved_rregs = kzalloc(sizeof(struct fsl_ifc_runtime), GFP_KERNEL);
-+ if (!ctrl->saved_rregs)
-+ return -ENOMEM;
-+
-+ cm_evter_intr_en = ifc_in32(&fcm->cm_evter_intr_en);
-+ nand_evter_intr_en = ifc_in32(&runtime->ifc_nand.nand_evter_intr_en);
-+ nor_evter_intr_en = ifc_in32(&runtime->ifc_nor.nor_evter_intr_en);
-+ gpcm_evter_intr_en = ifc_in32(&runtime->ifc_gpcm.gpcm_evter_intr_en);
-+
-+/* IFC interrupts disabled */
-+
-+ ifc_out32(0x0, &fcm->cm_evter_intr_en);
-+ ifc_out32(0x0, &runtime->ifc_nand.nand_evter_intr_en);
-+ ifc_out32(0x0, &runtime->ifc_nor.nor_evter_intr_en);
-+ ifc_out32(0x0, &runtime->ifc_gpcm.gpcm_evter_intr_en);
-+
-+ if (ctrl->saved_gregs) {
-+ for (ifc_bank = 0; ifc_bank < FSL_IFC_BANK_COUNT; ifc_bank++) {
-+ ctrl->saved_gregs->cspr_cs[ifc_bank].cspr_ext =
-+ ifc_in32(&fcm->cspr_cs[ifc_bank].cspr_ext);
-+ ctrl->saved_gregs->cspr_cs[ifc_bank].cspr =
-+ ifc_in32(&fcm->cspr_cs[ifc_bank].cspr);
-+ ctrl->saved_gregs->amask_cs[ifc_bank].amask =
-+ ifc_in32(&fcm->amask_cs[ifc_bank].amask);
-+ ctrl->saved_gregs->csor_cs[ifc_bank].csor_ext =
-+ ifc_in32(&fcm->csor_cs[ifc_bank].csor_ext);
-+ ctrl->saved_gregs->csor_cs[ifc_bank].csor =
-+ ifc_in32(&fcm->csor_cs[ifc_bank].csor);
-+ for (i = 0; i < 4; i++) {
-+ ctrl->saved_gregs->ftim_cs[ifc_bank].ftim[i] =
-+ ifc_in32(
-+ &fcm->ftim_cs[ifc_bank].ftim[i]);
-+ }
-+ }
-+
-+ ctrl->saved_gregs->rb_map = ifc_in32(&fcm->rb_map);
-+ ctrl->saved_gregs->wb_map = ifc_in32(&fcm->wb_map);
-+ ctrl->saved_gregs->ifc_gcr = ifc_in32(&fcm->ifc_gcr);
-+ ctrl->saved_gregs->ddr_ccr_low = ifc_in32(&fcm->ddr_ccr_low);
-+ ctrl->saved_gregs->cm_evter_en = ifc_in32(&fcm->cm_evter_en);
-+ }
-+
-+ if (ctrl->saved_rregs) {
-+ /* IFC controller NAND machine registers */
-+ ctrl->saved_rregs->ifc_nand.ncfgr =
-+ ifc_in32(&runtime->ifc_nand.ncfgr);
-+ ctrl->saved_rregs->ifc_nand.nand_fcr0 =
-+ ifc_in32(&runtime->ifc_nand.nand_fcr0);
-+ ctrl->saved_rregs->ifc_nand.nand_fcr1 =
-+ ifc_in32(&runtime->ifc_nand.nand_fcr1);
-+ ctrl->saved_rregs->ifc_nand.row0 =
-+ ifc_in32(&runtime->ifc_nand.row0);
-+ ctrl->saved_rregs->ifc_nand.row1 =
-+ ifc_in32(&runtime->ifc_nand.row1);
-+ ctrl->saved_rregs->ifc_nand.col0 =
-+ ifc_in32(&runtime->ifc_nand.col0);
-+ ctrl->saved_rregs->ifc_nand.col1 =
-+ ifc_in32(&runtime->ifc_nand.col1);
-+ ctrl->saved_rregs->ifc_nand.row2 =
-+ ifc_in32(&runtime->ifc_nand.row2);
-+ ctrl->saved_rregs->ifc_nand.col2 =
-+ ifc_in32(&runtime->ifc_nand.col2);
-+ ctrl->saved_rregs->ifc_nand.row3 =
-+ ifc_in32(&runtime->ifc_nand.row3);
-+ ctrl->saved_rregs->ifc_nand.col3 =
-+ ifc_in32(&runtime->ifc_nand.col3);
-+
-+ ctrl->saved_rregs->ifc_nand.nand_fbcr =
-+ ifc_in32(&runtime->ifc_nand.nand_fbcr);
-+ ctrl->saved_rregs->ifc_nand.nand_fir0 =
-+ ifc_in32(&runtime->ifc_nand.nand_fir0);
-+ ctrl->saved_rregs->ifc_nand.nand_fir1 =
-+ ifc_in32(&runtime->ifc_nand.nand_fir1);
-+ ctrl->saved_rregs->ifc_nand.nand_fir2 =
-+ ifc_in32(&runtime->ifc_nand.nand_fir2);
-+ ctrl->saved_rregs->ifc_nand.nand_csel =
-+ ifc_in32(&runtime->ifc_nand.nand_csel);
-+ ctrl->saved_rregs->ifc_nand.nandseq_strt =
-+ ifc_in32(
-+ &runtime->ifc_nand.nandseq_strt);
-+ ctrl->saved_rregs->ifc_nand.nand_evter_en =
-+ ifc_in32(
-+ &runtime->ifc_nand.nand_evter_en);
-+ ctrl->saved_rregs->ifc_nand.nanndcr =
-+ ifc_in32(&runtime->ifc_nand.nanndcr);
-+ ctrl->saved_rregs->ifc_nand.nand_dll_lowcfg0 =
-+ ifc_in32(
-+ &runtime->ifc_nand.nand_dll_lowcfg0);
-+ ctrl->saved_rregs->ifc_nand.nand_dll_lowcfg1 =
-+ ifc_in32(
-+ &runtime->ifc_nand.nand_dll_lowcfg1);
-+
-+ /* IFC controller NOR machine registers */
-+ ctrl->saved_rregs->ifc_nor.nor_evter_en =
-+ ifc_in32(
-+ &runtime->ifc_nor.nor_evter_en);
-+ ctrl->saved_rregs->ifc_nor.norcr =
-+ ifc_in32(&runtime->ifc_nor.norcr);
-+
-+ /* IFC controller GPCM Machine registers */
-+ ctrl->saved_rregs->ifc_gpcm.gpcm_evter_en =
-+ ifc_in32(
-+ &runtime->ifc_gpcm.gpcm_evter_en);
-+ }
-+
-+/* save the interrupt values */
-+ ctrl->saved_gregs->cm_evter_intr_en = cm_evter_intr_en;
-+ ctrl->saved_rregs->ifc_nand.nand_evter_intr_en = nand_evter_intr_en;
-+ ctrl->saved_rregs->ifc_nor.nor_evter_intr_en = nor_evter_intr_en;
-+ ctrl->saved_rregs->ifc_gpcm.gpcm_evter_intr_en = gpcm_evter_intr_en;
-+
-+ return 0;
-+}
-+
-+/* restore ifc registers */
-+static int fsl_ifc_resume(struct device *dev)
-+{
-+ struct fsl_ifc_ctrl *ctrl = dev_get_drvdata(dev);
-+ struct fsl_ifc_global __iomem *fcm = ctrl->gregs;
-+ struct fsl_ifc_runtime __iomem *runtime = ctrl->rregs;
-+ struct fsl_ifc_global *savd_gregs = ctrl->saved_gregs;
-+ struct fsl_ifc_runtime *savd_rregs = ctrl->saved_rregs;
-+ uint32_t ver = 0, ncfgr, timeout, ifc_bank, i;
-+
-+/*
-+ * IFC interrupts disabled
-+ */
-+ ifc_out32(0x0, &fcm->cm_evter_intr_en);
-+ ifc_out32(0x0, &runtime->ifc_nand.nand_evter_intr_en);
-+ ifc_out32(0x0, &runtime->ifc_nor.nor_evter_intr_en);
-+ ifc_out32(0x0, &runtime->ifc_gpcm.gpcm_evter_intr_en);
-+
-+
-+ if (ctrl->saved_gregs) {
-+ for (ifc_bank = 0; ifc_bank < FSL_IFC_BANK_COUNT; ifc_bank++) {
-+ ifc_out32(savd_gregs->cspr_cs[ifc_bank].cspr_ext,
-+ &fcm->cspr_cs[ifc_bank].cspr_ext);
-+ ifc_out32(savd_gregs->cspr_cs[ifc_bank].cspr,
-+ &fcm->cspr_cs[ifc_bank].cspr);
-+ ifc_out32(savd_gregs->amask_cs[ifc_bank].amask,
-+ &fcm->amask_cs[ifc_bank].amask);
-+ ifc_out32(savd_gregs->csor_cs[ifc_bank].csor_ext,
-+ &fcm->csor_cs[ifc_bank].csor_ext);
-+ ifc_out32(savd_gregs->csor_cs[ifc_bank].csor,
-+ &fcm->csor_cs[ifc_bank].csor);
-+ for (i = 0; i < 4; i++) {
-+ ifc_out32(savd_gregs->ftim_cs[ifc_bank].ftim[i],
-+ &fcm->ftim_cs[ifc_bank].ftim[i]);
-+ }
-+ }
-+ ifc_out32(savd_gregs->rb_map, &fcm->rb_map);
-+ ifc_out32(savd_gregs->wb_map, &fcm->wb_map);
-+ ifc_out32(savd_gregs->ifc_gcr, &fcm->ifc_gcr);
-+ ifc_out32(savd_gregs->ddr_ccr_low, &fcm->ddr_ccr_low);
-+ ifc_out32(savd_gregs->cm_evter_en, &fcm->cm_evter_en);
-+ }
-+
-+ if (ctrl->saved_rregs) {
-+ /* IFC controller NAND machine registers */
-+ ifc_out32(savd_rregs->ifc_nand.ncfgr,
-+ &runtime->ifc_nand.ncfgr);
-+ ifc_out32(savd_rregs->ifc_nand.nand_fcr0,
-+ &runtime->ifc_nand.nand_fcr0);
-+ ifc_out32(savd_rregs->ifc_nand.nand_fcr1,
-+ &runtime->ifc_nand.nand_fcr1);
-+ ifc_out32(savd_rregs->ifc_nand.row0, &runtime->ifc_nand.row0);
-+ ifc_out32(savd_rregs->ifc_nand.row1, &runtime->ifc_nand.row1);
-+ ifc_out32(savd_rregs->ifc_nand.col0, &runtime->ifc_nand.col0);
-+ ifc_out32(savd_rregs->ifc_nand.col1, &runtime->ifc_nand.col1);
-+ ifc_out32(savd_rregs->ifc_nand.row2, &runtime->ifc_nand.row2);
-+ ifc_out32(savd_rregs->ifc_nand.col2, &runtime->ifc_nand.col2);
-+ ifc_out32(savd_rregs->ifc_nand.row3, &runtime->ifc_nand.row3);
-+ ifc_out32(savd_rregs->ifc_nand.col3, &runtime->ifc_nand.col3);
-+ ifc_out32(savd_rregs->ifc_nand.nand_fbcr,
-+ &runtime->ifc_nand.nand_fbcr);
-+ ifc_out32(savd_rregs->ifc_nand.nand_fir0,
-+ &runtime->ifc_nand.nand_fir0);
-+ ifc_out32(savd_rregs->ifc_nand.nand_fir1,
-+ &runtime->ifc_nand.nand_fir1);
-+ ifc_out32(savd_rregs->ifc_nand.nand_fir2,
-+ &runtime->ifc_nand.nand_fir2);
-+ ifc_out32(savd_rregs->ifc_nand.nand_csel,
-+ &runtime->ifc_nand.nand_csel);
-+ ifc_out32(savd_rregs->ifc_nand.nandseq_strt,
-+ &runtime->ifc_nand.nandseq_strt);
-+ ifc_out32(savd_rregs->ifc_nand.nand_evter_en,
-+ &runtime->ifc_nand.nand_evter_en);
-+ ifc_out32(savd_rregs->ifc_nand.nanndcr,
-+ &runtime->ifc_nand.nanndcr);
-+ ifc_out32(savd_rregs->ifc_nand.nand_dll_lowcfg0,
-+ &runtime->ifc_nand.nand_dll_lowcfg0);
-+ ifc_out32(savd_rregs->ifc_nand.nand_dll_lowcfg1,
-+ &runtime->ifc_nand.nand_dll_lowcfg1);
-+
-+ /* IFC controller NOR machine registers */
-+ ifc_out32(savd_rregs->ifc_nor.nor_evter_en,
-+ &runtime->ifc_nor.nor_evter_en);
-+ ifc_out32(savd_rregs->ifc_nor.norcr, &runtime->ifc_nor.norcr);
-+
-+ /* IFC controller GPCM Machine registers */
-+ ifc_out32(savd_rregs->ifc_gpcm.gpcm_evter_en,
-+ &runtime->ifc_gpcm.gpcm_evter_en);
-+
-+ /* IFC interrupts enabled */
-+ ifc_out32(ctrl->saved_gregs->cm_evter_intr_en,
-+ &fcm->cm_evter_intr_en);
-+ ifc_out32(ctrl->saved_rregs->ifc_nand.nand_evter_intr_en,
-+ &runtime->ifc_nand.nand_evter_intr_en);
-+ ifc_out32(ctrl->saved_rregs->ifc_nor.nor_evter_intr_en,
-+ &runtime->ifc_nor.nor_evter_intr_en);
-+ ifc_out32(ctrl->saved_rregs->ifc_gpcm.gpcm_evter_intr_en,
-+ &runtime->ifc_gpcm.gpcm_evter_intr_en);
-+
-+ kfree(ctrl->saved_gregs);
-+ kfree(ctrl->saved_rregs);
-+ ctrl->saved_gregs = NULL;
-+ ctrl->saved_rregs = NULL;
-+ }
-+
-+ ver = ifc_in32(&fcm->ifc_rev);
-+ ncfgr = ifc_in32(&runtime->ifc_nand.ncfgr);
-+ if (ver >= FSL_IFC_V1_3_0) {
-+
-+ ifc_out32(ncfgr | IFC_NAND_SRAM_INIT_EN,
-+ &runtime->ifc_nand.ncfgr);
-+ /* wait for SRAM_INIT bit to be clear or timeout */
-+ timeout = 10;
-+ while ((ifc_in32(&runtime->ifc_nand.ncfgr) &
-+ IFC_NAND_SRAM_INIT_EN) && timeout) {
-+ mdelay(IFC_TIMEOUT_MSECS);
-+ timeout--;
-+ }
-+
-+ if (!timeout)
-+ dev_err(ctrl->dev, "Timeout waiting for IFC SRAM INIT");
-+ }
-+
-+ return 0;
-+}
-+#endif /* CONFIG_PM_SLEEP */
-+
- static const struct of_device_id fsl_ifc_match[] = {
- {
- .compatible = "fsl,ifc",
-@@ -318,10 +576,15 @@ static const struct of_device_id fsl_ifc
- {},
- };
-
-+static const struct dev_pm_ops ifc_pm_ops = {
-+ SET_SYSTEM_SLEEP_PM_OPS(fsl_ifc_suspend, fsl_ifc_resume)
-+};
-+
- static struct platform_driver fsl_ifc_ctrl_driver = {
- .driver = {
- .name = "fsl-ifc",
- .of_match_table = fsl_ifc_match,
-+ .pm = &ifc_pm_ops,
- },
- .probe = fsl_ifc_ctrl_probe,
- .remove = fsl_ifc_ctrl_remove,
---- a/drivers/mtd/maps/physmap_of.c
-+++ b/drivers/mtd/maps/physmap_of.c
-@@ -20,6 +20,7 @@
- #include <linux/mtd/map.h>
- #include <linux/mtd/partitions.h>
- #include <linux/mtd/concat.h>
-+#include <linux/mtd/cfi_endian.h>
- #include <linux/of.h>
- #include <linux/of_address.h>
- #include <linux/of_platform.h>
-@@ -209,6 +210,9 @@ static int of_flash_probe(struct platfor
- return err;
- }
-
-+ if (of_property_read_bool(dp->parent, "big-endian"))
-+ info->list[i].map.swap = CFI_BIG_ENDIAN;
-+
- err = -ENOMEM;
- info->list[i].map.virt = ioremap(info->list[i].map.phys,
- info->list[i].map.size);
---- a/drivers/mtd/nand/Kconfig
-+++ b/drivers/mtd/nand/Kconfig
-@@ -438,7 +438,7 @@ config MTD_NAND_FSL_ELBC
-
- config MTD_NAND_FSL_IFC
- tristate "NAND support for Freescale IFC controller"
-- depends on FSL_SOC || ARCH_LAYERSCAPE
-+ depends on FSL_SOC || ARCH_LAYERSCAPE || SOC_LS1021A
- select FSL_IFC
- select MEMORY
- help
---- a/drivers/mtd/nand/fsl_ifc_nand.c
-+++ b/drivers/mtd/nand/fsl_ifc_nand.c
-@@ -901,7 +901,7 @@ static int fsl_ifc_chip_init(struct fsl_
- chip->ecc.algo = NAND_ECC_HAMMING;
- }
-
-- if (ctrl->version == FSL_IFC_VERSION_1_1_0)
-+ if (ctrl->version >= FSL_IFC_VERSION_1_1_0)
- fsl_ifc_sram_init(priv);
-
- /*
---- a/include/linux/fsl_ifc.h
-+++ b/include/linux/fsl_ifc.h
-@@ -274,6 +274,8 @@
- */
- /* Auto Boot Mode */
- #define IFC_NAND_NCFGR_BOOT 0x80000000
-+/* SRAM INIT EN */
-+#define IFC_NAND_SRAM_INIT_EN 0x20000000
- /* Addressing Mode-ROW0+n/COL0 */
- #define IFC_NAND_NCFGR_ADDR_MODE_RC0 0x00000000
- /* Addressing Mode-ROW0+n/COL0+n */
-@@ -857,6 +859,11 @@ struct fsl_ifc_ctrl {
- u32 nand_stat;
- wait_queue_head_t nand_wait;
- bool little_endian;
-+#ifdef CONFIG_PM_SLEEP
-+ /*save regs when system goes to deep sleep*/
-+ struct fsl_ifc_global *saved_gregs;
-+ struct fsl_ifc_runtime *saved_rregs;
-+#endif
- };
-
- extern struct fsl_ifc_ctrl *fsl_ifc_ctrl_dev;
diff --git a/target/linux/layerscape/patches-4.9/701-sdk_dpaa-support-layerscape.patch b/target/linux/layerscape/patches-4.9/701-sdk_dpaa-support-layerscape.patch
deleted file mode 100644
index 0ea829aef2..0000000000
--- a/target/linux/layerscape/patches-4.9/701-sdk_dpaa-support-layerscape.patch
+++ /dev/null
@@ -1,154756 +0,0 @@
-From 6cc4cbfd0456c752f9f59d7d07fbb4b514dc6909 Mon Sep 17 00:00:00 2001
-From: Yangbo Lu <yangbo.lu@nxp.com>
-Date: Thu, 5 Jul 2018 16:25:00 +0800
-Subject: [PATCH 07/32] sdk_dpaa: support layerscape
-
-This is an integrated patch for layerscape dpaa1-sdk support.
-
-Signed-off-by: Camelia Groza <camelia.groza@nxp.com>
-Signed-off-by: Zhao Qiang <qiang.zhao@nxp.com>
-Signed-off-by: Zhang Ying-22455 <ying.zhang22455@nxp.com>
-Signed-off-by: Madalin Bucur <madalin.bucur@nxp.com>
-Signed-off-by: Mathew McBride <matt@traverse.com.au>
-Signed-off-by: Yangbo Lu <yangbo.lu@nxp.com>
----
- .../net/ethernet/freescale/sdk_dpaa/Kconfig | 196 +
- .../net/ethernet/freescale/sdk_dpaa/Makefile | 46 +
- .../ethernet/freescale/sdk_dpaa/dpaa_1588.c | 580 ++
- .../ethernet/freescale/sdk_dpaa/dpaa_1588.h | 138 +
- .../freescale/sdk_dpaa/dpaa_debugfs.c | 180 +
- .../freescale/sdk_dpaa/dpaa_debugfs.h | 43 +
- .../ethernet/freescale/sdk_dpaa/dpaa_eth.c | 1224 +++
- .../ethernet/freescale/sdk_dpaa/dpaa_eth.h | 687 ++
- .../freescale/sdk_dpaa/dpaa_eth_base.c | 205 +
- .../freescale/sdk_dpaa/dpaa_eth_base.h | 49 +
- .../freescale/sdk_dpaa/dpaa_eth_ceetm.c | 2115 +++++
- .../freescale/sdk_dpaa/dpaa_eth_ceetm.h | 240 +
- .../freescale/sdk_dpaa/dpaa_eth_common.c | 1802 ++++
- .../freescale/sdk_dpaa/dpaa_eth_common.h | 225 +
- .../freescale/sdk_dpaa/dpaa_eth_proxy.c | 381 +
- .../ethernet/freescale/sdk_dpaa/dpaa_eth_sg.c | 1193 +++
- .../freescale/sdk_dpaa/dpaa_eth_sysfs.c | 278 +
- .../freescale/sdk_dpaa/dpaa_eth_trace.h | 144 +
- .../freescale/sdk_dpaa/dpaa_ethtool.c | 544 ++
- .../ethernet/freescale/sdk_dpaa/dpaa_ptp.c | 291 +
- .../net/ethernet/freescale/sdk_dpaa/mac-api.c | 931 ++
- drivers/net/ethernet/freescale/sdk_dpaa/mac.c | 489 ++
- drivers/net/ethernet/freescale/sdk_dpaa/mac.h | 135 +
- .../freescale/sdk_dpaa/offline_port.c | 848 ++
- .../freescale/sdk_dpaa/offline_port.h | 59 +
- .../net/ethernet/freescale/sdk_fman/Kconfig | 153 +
- .../net/ethernet/freescale/sdk_fman/Makefile | 11 +
- .../sdk_fman/Peripherals/FM/HC/Makefile | 15 +
- .../freescale/sdk_fman/Peripherals/FM/HC/hc.c | 1232 +++
- .../sdk_fman/Peripherals/FM/MAC/Makefile | 28 +
- .../sdk_fman/Peripherals/FM/MAC/dtsec.c | 1504 ++++
- .../sdk_fman/Peripherals/FM/MAC/dtsec.h | 228 +
- .../Peripherals/FM/MAC/dtsec_mii_acc.c | 97 +
- .../Peripherals/FM/MAC/dtsec_mii_acc.h | 42 +
- .../sdk_fman/Peripherals/FM/MAC/fm_mac.c | 674 ++
- .../sdk_fman/Peripherals/FM/MAC/fm_mac.h | 226 +
- .../sdk_fman/Peripherals/FM/MAC/fman_crc32.c | 119 +
- .../sdk_fman/Peripherals/FM/MAC/fman_crc32.h | 43 +
- .../sdk_fman/Peripherals/FM/MAC/fman_dtsec.c | 845 ++
- .../Peripherals/FM/MAC/fman_dtsec_mii_acc.c | 163 +
- .../sdk_fman/Peripherals/FM/MAC/fman_memac.c | 532 ++
- .../Peripherals/FM/MAC/fman_memac_mii_acc.c | 213 +
- .../sdk_fman/Peripherals/FM/MAC/fman_tgec.c | 367 +
- .../sdk_fman/Peripherals/FM/MAC/memac.c | 1153 +++
- .../sdk_fman/Peripherals/FM/MAC/memac.h | 110 +
- .../Peripherals/FM/MAC/memac_mii_acc.c | 78 +
- .../Peripherals/FM/MAC/memac_mii_acc.h | 73 +
- .../sdk_fman/Peripherals/FM/MAC/tgec.c | 1017 +++
- .../sdk_fman/Peripherals/FM/MAC/tgec.h | 151 +
- .../Peripherals/FM/MAC/tgec_mii_acc.c | 139 +
- .../Peripherals/FM/MAC/tgec_mii_acc.h | 80 +
- .../sdk_fman/Peripherals/FM/MACSEC/Makefile | 15 +
- .../Peripherals/FM/MACSEC/fm_macsec.c | 237 +
- .../Peripherals/FM/MACSEC/fm_macsec.h | 203 +
- .../Peripherals/FM/MACSEC/fm_macsec_guest.c | 59 +
- .../Peripherals/FM/MACSEC/fm_macsec_master.c | 1031 +++
- .../Peripherals/FM/MACSEC/fm_macsec_master.h | 479 ++
- .../Peripherals/FM/MACSEC/fm_macsec_secy.c | 883 ++
- .../Peripherals/FM/MACSEC/fm_macsec_secy.h | 144 +
- .../sdk_fman/Peripherals/FM/Makefile | 23 +
- .../sdk_fman/Peripherals/FM/Pcd/Makefile | 26 +
- .../sdk_fman/Peripherals/FM/Pcd/crc64.h | 360 +
- .../sdk_fman/Peripherals/FM/Pcd/fm_cc.c | 7582 +++++++++++++++++
- .../sdk_fman/Peripherals/FM/Pcd/fm_cc.h | 399 +
- .../sdk_fman/Peripherals/FM/Pcd/fm_kg.c | 3242 +++++++
- .../sdk_fman/Peripherals/FM/Pcd/fm_kg.h | 206 +
- .../sdk_fman/Peripherals/FM/Pcd/fm_manip.c | 5571 ++++++++++++
- .../sdk_fman/Peripherals/FM/Pcd/fm_manip.h | 555 ++
- .../sdk_fman/Peripherals/FM/Pcd/fm_pcd.c | 2095 +++++
- .../sdk_fman/Peripherals/FM/Pcd/fm_pcd.h | 543 ++
- .../sdk_fman/Peripherals/FM/Pcd/fm_pcd_ipc.h | 280 +
- .../sdk_fman/Peripherals/FM/Pcd/fm_plcr.c | 1847 ++++
- .../sdk_fman/Peripherals/FM/Pcd/fm_plcr.h | 165 +
- .../sdk_fman/Peripherals/FM/Pcd/fm_prs.c | 423 +
- .../sdk_fman/Peripherals/FM/Pcd/fm_prs.h | 316 +
- .../sdk_fman/Peripherals/FM/Pcd/fm_replic.c | 984 +++
- .../sdk_fman/Peripherals/FM/Pcd/fm_replic.h | 101 +
- .../sdk_fman/Peripherals/FM/Pcd/fman_kg.c | 888 ++
- .../sdk_fman/Peripherals/FM/Pcd/fman_prs.c | 129 +
- .../sdk_fman/Peripherals/FM/Port/Makefile | 15 +
- .../sdk_fman/Peripherals/FM/Port/fm_port.c | 6436 ++++++++++++++
- .../sdk_fman/Peripherals/FM/Port/fm_port.h | 999 +++
- .../Peripherals/FM/Port/fm_port_dsar.h | 494 ++
- .../sdk_fman/Peripherals/FM/Port/fm_port_im.c | 753 ++
- .../sdk_fman/Peripherals/FM/Port/fman_port.c | 1568 ++++
- .../sdk_fman/Peripherals/FM/Rtc/Makefile | 15 +
- .../sdk_fman/Peripherals/FM/Rtc/fm_rtc.c | 692 ++
- .../sdk_fman/Peripherals/FM/Rtc/fm_rtc.h | 96 +
- .../sdk_fman/Peripherals/FM/Rtc/fman_rtc.c | 334 +
- .../sdk_fman/Peripherals/FM/SP/Makefile | 15 +
- .../sdk_fman/Peripherals/FM/SP/fm_sp.c | 757 ++
- .../sdk_fman/Peripherals/FM/SP/fm_sp.h | 85 +
- .../sdk_fman/Peripherals/FM/SP/fman_sp.c | 197 +
- .../freescale/sdk_fman/Peripherals/FM/fm.c | 5216 ++++++++++++
- .../freescale/sdk_fman/Peripherals/FM/fm.h | 648 ++
- .../sdk_fman/Peripherals/FM/fm_ipc.h | 465 +
- .../sdk_fman/Peripherals/FM/fm_muram.c | 174 +
- .../freescale/sdk_fman/Peripherals/FM/fman.c | 1398 +++
- .../sdk_fman/Peripherals/FM/inc/fm_common.h | 1214 +++
- .../sdk_fman/Peripherals/FM/inc/fm_hc.h | 93 +
- .../Peripherals/FM/inc/fm_sp_common.h | 117 +
- .../ethernet/freescale/sdk_fman/etc/Makefile | 12 +
- .../ethernet/freescale/sdk_fman/etc/error.c | 95 +
- .../ethernet/freescale/sdk_fman/etc/list.c | 71 +
- .../ethernet/freescale/sdk_fman/etc/memcpy.c | 620 ++
- .../net/ethernet/freescale/sdk_fman/etc/mm.c | 1155 +++
- .../net/ethernet/freescale/sdk_fman/etc/mm.h | 105 +
- .../ethernet/freescale/sdk_fman/etc/sprint.c | 81 +
- .../freescale/sdk_fman/fmanv3h_dflags.h | 57 +
- .../freescale/sdk_fman/fmanv3l_dflags.h | 56 +
- .../inc/Peripherals/crc_mac_addr_ext.h | 364 +
- .../sdk_fman/inc/Peripherals/dpaa_ext.h | 210 +
- .../sdk_fman/inc/Peripherals/fm_ext.h | 1731 ++++
- .../sdk_fman/inc/Peripherals/fm_mac_ext.h | 887 ++
- .../sdk_fman/inc/Peripherals/fm_macsec_ext.h | 1271 +++
- .../sdk_fman/inc/Peripherals/fm_muram_ext.h | 170 +
- .../sdk_fman/inc/Peripherals/fm_pcd_ext.h | 3974 +++++++++
- .../sdk_fman/inc/Peripherals/fm_port_ext.h | 2608 ++++++
- .../sdk_fman/inc/Peripherals/fm_rtc_ext.h | 619 ++
- .../sdk_fman/inc/Peripherals/fm_vsp_ext.h | 411 +
- .../sdk_fman/inc/Peripherals/mii_acc_ext.h | 76 +
- .../freescale/sdk_fman/inc/core_ext.h | 90 +
- .../freescale/sdk_fman/inc/cores/arm_ext.h | 55 +
- .../freescale/sdk_fman/inc/cores/e500v2_ext.h | 476 ++
- .../freescale/sdk_fman/inc/cores/ppc_ext.h | 141 +
- .../freescale/sdk_fman/inc/ddr_std_ext.h | 77 +
- .../freescale/sdk_fman/inc/debug_ext.h | 233 +
- .../freescale/sdk_fman/inc/endian_ext.h | 447 +
- .../freescale/sdk_fman/inc/enet_ext.h | 205 +
- .../freescale/sdk_fman/inc/error_ext.h | 529 ++
- .../freescale/sdk_fman/inc/etc/list_ext.h | 358 +
- .../freescale/sdk_fman/inc/etc/mem_ext.h | 318 +
- .../freescale/sdk_fman/inc/etc/memcpy_ext.h | 208 +
- .../freescale/sdk_fman/inc/etc/mm_ext.h | 310 +
- .../freescale/sdk_fman/inc/etc/sprint_ext.h | 118 +
- .../inc/flib/common/arch/ppc_access.h | 37 +
- .../sdk_fman/inc/flib/common/general.h | 52 +
- .../freescale/sdk_fman/inc/flib/fman_common.h | 78 +
- .../freescale/sdk_fman/inc/flib/fsl_enet.h | 273 +
- .../freescale/sdk_fman/inc/flib/fsl_fman.h | 825 ++
- .../sdk_fman/inc/flib/fsl_fman_dtsec.h | 1096 +++
- .../inc/flib/fsl_fman_dtsec_mii_acc.h | 107 +
- .../freescale/sdk_fman/inc/flib/fsl_fman_kg.h | 514 ++
- .../sdk_fman/inc/flib/fsl_fman_memac.h | 434 +
- .../inc/flib/fsl_fman_memac_mii_acc.h | 78 +
- .../sdk_fman/inc/flib/fsl_fman_port.h | 593 ++
- .../sdk_fman/inc/flib/fsl_fman_prs.h | 102 +
- .../sdk_fman/inc/flib/fsl_fman_rtc.h | 449 +
- .../freescale/sdk_fman/inc/flib/fsl_fman_sp.h | 138 +
- .../sdk_fman/inc/flib/fsl_fman_tgec.h | 479 ++
- .../FMANV3H/dpaa_integration_ext.h | 291 +
- .../inc/integrations/FMANV3H/part_ext.h | 71 +
- .../FMANV3H/part_integration_ext.h | 304 +
- .../FMANV3L/dpaa_integration_ext.h | 293 +
- .../inc/integrations/FMANV3L/part_ext.h | 59 +
- .../FMANV3L/part_integration_ext.h | 304 +
- .../LS1043/dpaa_integration_ext.h | 291 +
- .../inc/integrations/LS1043/part_ext.h | 64 +
- .../LS1043/part_integration_ext.h | 185 +
- .../integrations/P1023/dpaa_integration_ext.h | 213 +
- .../inc/integrations/P1023/part_ext.h | 82 +
- .../integrations/P1023/part_integration_ext.h | 635 ++
- .../P3040_P4080_P5020/dpaa_integration_ext.h | 276 +
- .../integrations/P3040_P4080_P5020/part_ext.h | 83 +
- .../P3040_P4080_P5020/part_integration_ext.h | 336 +
- .../freescale/sdk_fman/inc/math_ext.h | 100 +
- .../freescale/sdk_fman/inc/ncsw_ext.h | 435 +
- .../ethernet/freescale/sdk_fman/inc/net_ext.h | 430 +
- .../ethernet/freescale/sdk_fman/inc/std_ext.h | 48 +
- .../freescale/sdk_fman/inc/stdarg_ext.h | 49 +
- .../freescale/sdk_fman/inc/stdlib_ext.h | 162 +
- .../freescale/sdk_fman/inc/string_ext.h | 56 +
- .../freescale/sdk_fman/inc/types_ext.h | 62 +
- .../freescale/sdk_fman/inc/xx_common.h | 56 +
- .../ethernet/freescale/sdk_fman/inc/xx_ext.h | 791 ++
- .../freescale/sdk_fman/ls1043_dflags.h | 56 +
- .../freescale/sdk_fman/ncsw_config.mk | 53 +
- .../freescale/sdk_fman/p1023_dflags.h | 65 +
- .../sdk_fman/p3040_4080_5020_dflags.h | 62 +
- .../ethernet/freescale/sdk_fman/src/Makefile | 11 +
- .../sdk_fman/src/inc/system/sys_ext.h | 118 +
- .../sdk_fman/src/inc/system/sys_io_ext.h | 46 +
- .../freescale/sdk_fman/src/inc/types_linux.h | 208 +
- .../sdk_fman/src/inc/wrapper/fsl_fman_test.h | 84 +
- .../sdk_fman/src/inc/wrapper/lnxwrp_exp_sym.h | 130 +
- .../sdk_fman/src/inc/wrapper/lnxwrp_fm_ext.h | 163 +
- .../src/inc/wrapper/lnxwrp_fsl_fman.h | 921 ++
- .../freescale/sdk_fman/src/inc/xx/xx.h | 50 +
- .../freescale/sdk_fman/src/system/Makefile | 10 +
- .../freescale/sdk_fman/src/system/sys_io.c | 171 +
- .../freescale/sdk_fman/src/wrapper/Makefile | 19 +
- .../sdk_fman/src/wrapper/fman_test.c | 1665 ++++
- .../sdk_fman/src/wrapper/lnxwrp_fm.c | 2908 +++++++
- .../sdk_fman/src/wrapper/lnxwrp_fm.h | 294 +
- .../sdk_fman/src/wrapper/lnxwrp_fm_port.c | 1512 ++++
- .../sdk_fman/src/wrapper/lnxwrp_ioctls_fm.c | 4854 +++++++++++
- .../src/wrapper/lnxwrp_ioctls_fm_compat.c | 1297 +++
- .../src/wrapper/lnxwrp_ioctls_fm_compat.h | 755 ++
- .../sdk_fman/src/wrapper/lnxwrp_resources.h | 121 +
- .../src/wrapper/lnxwrp_resources_ut.c | 191 +
- .../src/wrapper/lnxwrp_resources_ut.h | 144 +
- .../src/wrapper/lnxwrp_resources_ut.make | 28 +
- .../sdk_fman/src/wrapper/lnxwrp_sysfs.c | 60 +
- .../sdk_fman/src/wrapper/lnxwrp_sysfs.h | 60 +
- .../sdk_fman/src/wrapper/lnxwrp_sysfs_fm.c | 1855 ++++
- .../sdk_fman/src/wrapper/lnxwrp_sysfs_fm.h | 136 +
- .../src/wrapper/lnxwrp_sysfs_fm_port.c | 1268 +++
- .../src/wrapper/lnxwrp_sysfs_fm_port.h | 56 +
- .../freescale/sdk_fman/src/xx/Makefile | 18 +
- .../sdk_fman/src/xx/module_strings.c | 46 +
- .../freescale/sdk_fman/src/xx/xx_arm_linux.c | 905 ++
- .../freescale/sdk_fman/src/xx/xx_linux.c | 918 ++
- drivers/staging/fsl_qbman/Kconfig | 228 +
- drivers/staging/fsl_qbman/Makefile | 28 +
- drivers/staging/fsl_qbman/bman_config.c | 720 ++
- drivers/staging/fsl_qbman/bman_debugfs.c | 119 +
- drivers/staging/fsl_qbman/bman_driver.c | 575 ++
- drivers/staging/fsl_qbman/bman_high.c | 1145 +++
- drivers/staging/fsl_qbman/bman_low.h | 565 ++
- drivers/staging/fsl_qbman/bman_private.h | 166 +
- drivers/staging/fsl_qbman/bman_test.c | 56 +
- drivers/staging/fsl_qbman/bman_test.h | 44 +
- drivers/staging/fsl_qbman/bman_test_high.c | 183 +
- drivers/staging/fsl_qbman/bman_test_thresh.c | 196 +
- drivers/staging/fsl_qbman/dpa_alloc.c | 706 ++
- drivers/staging/fsl_qbman/dpa_sys.h | 259 +
- drivers/staging/fsl_qbman/dpa_sys_arm.h | 95 +
- drivers/staging/fsl_qbman/dpa_sys_arm64.h | 102 +
- drivers/staging/fsl_qbman/dpa_sys_ppc32.h | 70 +
- drivers/staging/fsl_qbman/dpa_sys_ppc64.h | 79 +
- drivers/staging/fsl_qbman/fsl_usdpaa.c | 2007 +++++
- drivers/staging/fsl_qbman/fsl_usdpaa_irq.c | 289 +
- drivers/staging/fsl_qbman/qbman_driver.c | 88 +
- drivers/staging/fsl_qbman/qman_config.c | 1224 +++
- drivers/staging/fsl_qbman/qman_debugfs.c | 1594 ++++
- drivers/staging/fsl_qbman/qman_driver.c | 977 +++
- drivers/staging/fsl_qbman/qman_high.c | 5652 ++++++++++++
- drivers/staging/fsl_qbman/qman_low.h | 1442 ++++
- drivers/staging/fsl_qbman/qman_private.h | 398 +
- drivers/staging/fsl_qbman/qman_test.c | 57 +
- drivers/staging/fsl_qbman/qman_test.h | 45 +
- drivers/staging/fsl_qbman/qman_test_high.c | 216 +
- .../staging/fsl_qbman/qman_test_hotpotato.c | 502 ++
- drivers/staging/fsl_qbman/qman_utility.c | 129 +
- include/linux/fsl_bman.h | 532 ++
- include/linux/fsl_qman.h | 3900 +++++++++
- include/linux/fsl_usdpaa.h | 372 +
- include/uapi/linux/fmd/Kbuild | 5 +
- include/uapi/linux/fmd/Peripherals/Kbuild | 4 +
- .../uapi/linux/fmd/Peripherals/fm_ioctls.h | 628 ++
- .../linux/fmd/Peripherals/fm_pcd_ioctls.h | 3084 +++++++
- .../linux/fmd/Peripherals/fm_port_ioctls.h | 973 +++
- .../linux/fmd/Peripherals/fm_test_ioctls.h | 208 +
- include/uapi/linux/fmd/integrations/Kbuild | 1 +
- .../fmd/integrations/integration_ioctls.h | 56 +
- include/uapi/linux/fmd/ioctls.h | 96 +
- include/uapi/linux/fmd/net_ioctls.h | 430 +
- 257 files changed, 153455 insertions(+)
- create mode 100644 drivers/net/ethernet/freescale/sdk_dpaa/Kconfig
- create mode 100644 drivers/net/ethernet/freescale/sdk_dpaa/Makefile
- create mode 100644 drivers/net/ethernet/freescale/sdk_dpaa/dpaa_1588.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_dpaa/dpaa_1588.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_dpaa/dpaa_debugfs.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_dpaa/dpaa_debugfs.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_dpaa/dpaa_eth.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_dpaa/dpaa_eth.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_dpaa/dpaa_eth_base.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_dpaa/dpaa_eth_base.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_dpaa/dpaa_eth_ceetm.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_dpaa/dpaa_eth_ceetm.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_dpaa/dpaa_eth_common.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_dpaa/dpaa_eth_common.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_dpaa/dpaa_eth_proxy.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_dpaa/dpaa_eth_sg.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_dpaa/dpaa_eth_sysfs.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_dpaa/dpaa_eth_trace.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_dpaa/dpaa_ethtool.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_dpaa/dpaa_ptp.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_dpaa/mac-api.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_dpaa/mac.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_dpaa/mac.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_dpaa/offline_port.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_dpaa/offline_port.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Kconfig
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Makefile
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/HC/Makefile
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/HC/hc.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MAC/Makefile
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MAC/dtsec.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MAC/dtsec.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MAC/dtsec_mii_acc.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MAC/dtsec_mii_acc.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MAC/fm_mac.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MAC/fm_mac.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MAC/fman_crc32.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MAC/fman_crc32.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MAC/fman_dtsec.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MAC/fman_dtsec_mii_acc.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MAC/fman_memac.c
- create mode 100755 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MAC/fman_memac_mii_acc.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MAC/fman_tgec.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MAC/memac.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MAC/memac.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MAC/memac_mii_acc.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MAC/memac_mii_acc.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MAC/tgec.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MAC/tgec.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MAC/tgec_mii_acc.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MAC/tgec_mii_acc.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MACSEC/Makefile
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MACSEC/fm_macsec.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MACSEC/fm_macsec.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MACSEC/fm_macsec_guest.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MACSEC/fm_macsec_master.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MACSEC/fm_macsec_master.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MACSEC/fm_macsec_secy.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MACSEC/fm_macsec_secy.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/Makefile
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/Pcd/Makefile
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/Pcd/crc64.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/Pcd/fm_cc.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/Pcd/fm_cc.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/Pcd/fm_kg.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/Pcd/fm_kg.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/Pcd/fm_manip.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/Pcd/fm_manip.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/Pcd/fm_pcd.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/Pcd/fm_pcd.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/Pcd/fm_pcd_ipc.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/Pcd/fm_plcr.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/Pcd/fm_plcr.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/Pcd/fm_prs.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/Pcd/fm_prs.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/Pcd/fm_replic.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/Pcd/fm_replic.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/Pcd/fman_kg.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/Pcd/fman_prs.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/Port/Makefile
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/Port/fm_port.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/Port/fm_port.h
- create mode 100755 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/Port/fm_port_dsar.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/Port/fm_port_im.c
- create mode 100755 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/Port/fman_port.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/Rtc/Makefile
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/Rtc/fm_rtc.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/Rtc/fm_rtc.h
- create mode 100755 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/Rtc/fman_rtc.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/SP/Makefile
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/SP/fm_sp.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/SP/fm_sp.h
- create mode 100755 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/SP/fman_sp.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/fm.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/fm.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/fm_ipc.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/fm_muram.c
- create mode 100755 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/fman.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/inc/fm_common.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/inc/fm_hc.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/inc/fm_sp_common.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/etc/Makefile
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/etc/error.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/etc/list.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/etc/memcpy.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/etc/mm.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/etc/mm.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/etc/sprint.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/fmanv3h_dflags.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/fmanv3l_dflags.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/inc/Peripherals/crc_mac_addr_ext.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/inc/Peripherals/dpaa_ext.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/inc/Peripherals/fm_ext.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/inc/Peripherals/fm_mac_ext.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/inc/Peripherals/fm_macsec_ext.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/inc/Peripherals/fm_muram_ext.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/inc/Peripherals/fm_pcd_ext.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/inc/Peripherals/fm_port_ext.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/inc/Peripherals/fm_rtc_ext.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/inc/Peripherals/fm_vsp_ext.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/inc/Peripherals/mii_acc_ext.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/inc/core_ext.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/inc/cores/arm_ext.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/inc/cores/e500v2_ext.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/inc/cores/ppc_ext.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/inc/ddr_std_ext.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/inc/debug_ext.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/inc/endian_ext.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/inc/enet_ext.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/inc/error_ext.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/inc/etc/list_ext.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/inc/etc/mem_ext.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/inc/etc/memcpy_ext.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/inc/etc/mm_ext.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/inc/etc/sprint_ext.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/inc/flib/common/arch/ppc_access.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/inc/flib/common/general.h
- create mode 100755 drivers/net/ethernet/freescale/sdk_fman/inc/flib/fman_common.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/inc/flib/fsl_enet.h
- create mode 100755 drivers/net/ethernet/freescale/sdk_fman/inc/flib/fsl_fman.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/inc/flib/fsl_fman_dtsec.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/inc/flib/fsl_fman_dtsec_mii_acc.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/inc/flib/fsl_fman_kg.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/inc/flib/fsl_fman_memac.h
- create mode 100755 drivers/net/ethernet/freescale/sdk_fman/inc/flib/fsl_fman_memac_mii_acc.h
- create mode 100755 drivers/net/ethernet/freescale/sdk_fman/inc/flib/fsl_fman_port.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/inc/flib/fsl_fman_prs.h
- create mode 100755 drivers/net/ethernet/freescale/sdk_fman/inc/flib/fsl_fman_rtc.h
- create mode 100755 drivers/net/ethernet/freescale/sdk_fman/inc/flib/fsl_fman_sp.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/inc/flib/fsl_fman_tgec.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/inc/integrations/FMANV3H/dpaa_integration_ext.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/inc/integrations/FMANV3H/part_ext.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/inc/integrations/FMANV3H/part_integration_ext.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/inc/integrations/FMANV3L/dpaa_integration_ext.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/inc/integrations/FMANV3L/part_ext.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/inc/integrations/FMANV3L/part_integration_ext.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/inc/integrations/LS1043/dpaa_integration_ext.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/inc/integrations/LS1043/part_ext.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/inc/integrations/LS1043/part_integration_ext.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/inc/integrations/P1023/dpaa_integration_ext.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/inc/integrations/P1023/part_ext.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/inc/integrations/P1023/part_integration_ext.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/inc/integrations/P3040_P4080_P5020/dpaa_integration_ext.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/inc/integrations/P3040_P4080_P5020/part_ext.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/inc/integrations/P3040_P4080_P5020/part_integration_ext.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/inc/math_ext.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/inc/ncsw_ext.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/inc/net_ext.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/inc/std_ext.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/inc/stdarg_ext.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/inc/stdlib_ext.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/inc/string_ext.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/inc/types_ext.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/inc/xx_common.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/inc/xx_ext.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/ls1043_dflags.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/ncsw_config.mk
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/p1023_dflags.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/p3040_4080_5020_dflags.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/src/Makefile
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/src/inc/system/sys_ext.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/src/inc/system/sys_io_ext.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/src/inc/types_linux.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/src/inc/wrapper/fsl_fman_test.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/src/inc/wrapper/lnxwrp_exp_sym.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/src/inc/wrapper/lnxwrp_fm_ext.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/src/inc/wrapper/lnxwrp_fsl_fman.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/src/inc/xx/xx.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/src/system/Makefile
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/src/system/sys_io.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/src/wrapper/Makefile
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/src/wrapper/fman_test.c
- create mode 100755 drivers/net/ethernet/freescale/sdk_fman/src/wrapper/lnxwrp_fm.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/src/wrapper/lnxwrp_fm.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/src/wrapper/lnxwrp_fm_port.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/src/wrapper/lnxwrp_ioctls_fm.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/src/wrapper/lnxwrp_ioctls_fm_compat.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/src/wrapper/lnxwrp_ioctls_fm_compat.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/src/wrapper/lnxwrp_resources.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/src/wrapper/lnxwrp_resources_ut.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/src/wrapper/lnxwrp_resources_ut.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/src/wrapper/lnxwrp_resources_ut.make
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/src/wrapper/lnxwrp_sysfs.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/src/wrapper/lnxwrp_sysfs.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/src/wrapper/lnxwrp_sysfs_fm.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/src/wrapper/lnxwrp_sysfs_fm.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/src/wrapper/lnxwrp_sysfs_fm_port.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/src/wrapper/lnxwrp_sysfs_fm_port.h
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/src/xx/Makefile
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/src/xx/module_strings.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/src/xx/xx_arm_linux.c
- create mode 100644 drivers/net/ethernet/freescale/sdk_fman/src/xx/xx_linux.c
- create mode 100644 drivers/staging/fsl_qbman/Kconfig
- create mode 100644 drivers/staging/fsl_qbman/Makefile
- create mode 100644 drivers/staging/fsl_qbman/bman_config.c
- create mode 100644 drivers/staging/fsl_qbman/bman_debugfs.c
- create mode 100644 drivers/staging/fsl_qbman/bman_driver.c
- create mode 100644 drivers/staging/fsl_qbman/bman_high.c
- create mode 100644 drivers/staging/fsl_qbman/bman_low.h
- create mode 100644 drivers/staging/fsl_qbman/bman_private.h
- create mode 100644 drivers/staging/fsl_qbman/bman_test.c
- create mode 100644 drivers/staging/fsl_qbman/bman_test.h
- create mode 100644 drivers/staging/fsl_qbman/bman_test_high.c
- create mode 100644 drivers/staging/fsl_qbman/bman_test_thresh.c
- create mode 100644 drivers/staging/fsl_qbman/dpa_alloc.c
- create mode 100644 drivers/staging/fsl_qbman/dpa_sys.h
- create mode 100644 drivers/staging/fsl_qbman/dpa_sys_arm.h
- create mode 100644 drivers/staging/fsl_qbman/dpa_sys_arm64.h
- create mode 100644 drivers/staging/fsl_qbman/dpa_sys_ppc32.h
- create mode 100644 drivers/staging/fsl_qbman/dpa_sys_ppc64.h
- create mode 100644 drivers/staging/fsl_qbman/fsl_usdpaa.c
- create mode 100644 drivers/staging/fsl_qbman/fsl_usdpaa_irq.c
- create mode 100644 drivers/staging/fsl_qbman/qbman_driver.c
- create mode 100644 drivers/staging/fsl_qbman/qman_config.c
- create mode 100644 drivers/staging/fsl_qbman/qman_debugfs.c
- create mode 100644 drivers/staging/fsl_qbman/qman_driver.c
- create mode 100644 drivers/staging/fsl_qbman/qman_high.c
- create mode 100644 drivers/staging/fsl_qbman/qman_low.h
- create mode 100644 drivers/staging/fsl_qbman/qman_private.h
- create mode 100644 drivers/staging/fsl_qbman/qman_test.c
- create mode 100644 drivers/staging/fsl_qbman/qman_test.h
- create mode 100644 drivers/staging/fsl_qbman/qman_test_high.c
- create mode 100644 drivers/staging/fsl_qbman/qman_test_hotpotato.c
- create mode 100644 drivers/staging/fsl_qbman/qman_utility.c
- create mode 100644 include/linux/fsl_bman.h
- create mode 100644 include/linux/fsl_qman.h
- create mode 100644 include/linux/fsl_usdpaa.h
- create mode 100644 include/uapi/linux/fmd/Kbuild
- create mode 100644 include/uapi/linux/fmd/Peripherals/Kbuild
- create mode 100644 include/uapi/linux/fmd/Peripherals/fm_ioctls.h
- create mode 100644 include/uapi/linux/fmd/Peripherals/fm_pcd_ioctls.h
- create mode 100644 include/uapi/linux/fmd/Peripherals/fm_port_ioctls.h
- create mode 100644 include/uapi/linux/fmd/Peripherals/fm_test_ioctls.h
- create mode 100644 include/uapi/linux/fmd/integrations/Kbuild
- create mode 100644 include/uapi/linux/fmd/integrations/integration_ioctls.h
- create mode 100644 include/uapi/linux/fmd/ioctls.h
- create mode 100644 include/uapi/linux/fmd/net_ioctls.h
-
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_dpaa/Kconfig
-@@ -0,0 +1,196 @@
-+menuconfig FSL_SDK_DPAA_ETH
-+ tristate "DPAA Ethernet"
-+ depends on (FSL_SOC || ARM64 || ARM) && FSL_SDK_BMAN && FSL_SDK_QMAN && FSL_SDK_FMAN && !FSL_DPAA_ETH
-+ select PHYLIB
-+ help
-+ Data Path Acceleration Architecture Ethernet driver,
-+ supporting the Freescale QorIQ chips.
-+ Depends on Freescale Buffer Manager and Queue Manager
-+ driver and Frame Manager Driver.
-+
-+if FSL_SDK_DPAA_ETH
-+
-+config FSL_DPAA_HOOKS
-+ bool "DPAA Ethernet driver hooks"
-+
-+config FSL_DPAA_CEETM
-+ bool "DPAA CEETM QoS"
-+ depends on NET_SCHED
-+ default n
-+ help
-+ Enable QoS offloading support through the CEETM hardware block.
-+
-+config FSL_DPAA_CEETM_CCS_THRESHOLD_1G
-+ hex "CEETM egress congestion threshold on 1G ports"
-+ depends on FSL_DPAA_CEETM
-+ range 0x1000 0x10000000
-+ default "0x000a0000"
-+ help
-+ The size in bytes of the CEETM egress Class Congestion State threshold on 1G ports.
-+ The threshold needs to be configured keeping in mind the following factors:
-+ - A threshold too large will buffer frames for a long time in the TX queues,
-+ when a small shaping rate is configured. This will cause buffer pool depletion
-+ or out of memory errors. This in turn will cause frame loss on RX;
-+ - A threshold too small will cause unnecessary frame loss by entering
-+ congestion too often.
-+
-+config FSL_DPAA_CEETM_CCS_THRESHOLD_10G
-+ hex "CEETM egress congestion threshold on 10G ports"
-+ depends on FSL_DPAA_CEETM
-+ range 0x1000 0x20000000
-+ default "0x00640000"
-+ help
-+ The size in bytes of the CEETM egress Class Congestion State threshold on 10G ports.
-+ See FSL_DPAA_CEETM_CCS_THRESHOLD_1G for details.
-+
-+config FSL_DPAA_OFFLINE_PORTS
-+ bool "Offline Ports support"
-+ depends on FSL_SDK_DPAA_ETH
-+ default y
-+ help
-+ The Offline Parsing / Host Command ports (short: OH ports, of Offline ports) provide
-+ most of the functionality of the regular, online ports, except they receive their
-+ frames from a core or an accelerator on the SoC, via QMan frame queues,
-+ rather than directly from the network.
-+ Offline ports are configured via PCD (Parse-Classify-Distribute) schemes, just like
-+ any online FMan port. They deliver the processed frames to frame queues, according
-+ to the applied PCD configurations.
-+
-+ Choosing this feature will not impact the functionality and/or performance of the system,
-+ so it is safe to have it.
-+
-+config FSL_DPAA_ADVANCED_DRIVERS
-+ bool "Advanced DPAA Ethernet drivers"
-+ depends on FSL_SDK_DPAA_ETH
-+ default y
-+ help
-+ Besides the standard DPAA Ethernet driver the DPAA Proxy initialization driver
-+ is needed to support advanced scenarios. Select this to also build the advanced
-+ drivers.
-+
-+config FSL_DPAA_ETH_JUMBO_FRAME
-+ bool "Optimize for jumbo frames"
-+ default n
-+ help
-+ Optimize the DPAA Ethernet driver throughput for large frames
-+ termination traffic (e.g. 4K and above).
-+ NOTE: This option can only be used if FSL_FM_MAX_FRAME_SIZE
-+ is set to 9600 bytes.
-+ Using this option in combination with small frames increases
-+ significantly the driver's memory footprint and may even deplete
-+ the system memory. Also, the skb truesize is altered and messages
-+ from the stack that warn against this are bypassed.
-+ This option is not available on LS1043.
-+
-+config FSL_DPAA_TS
-+ bool "Linux compliant timestamping"
-+ depends on FSL_SDK_DPAA_ETH
-+ default n
-+ help
-+ Enable Linux API compliant timestamping support.
-+
-+config FSL_DPAA_1588
-+ bool "IEEE 1588-compliant timestamping"
-+ depends on FSL_SDK_DPAA_ETH
-+ select FSL_DPAA_TS
-+ default n
-+ help
-+ Enable IEEE1588 support code.
-+
-+config FSL_DPAA_ETH_USE_NDO_SELECT_QUEUE
-+ bool "Use driver's Tx queue selection mechanism"
-+ default y
-+ depends on FSL_SDK_DPAA_ETH
-+ help
-+ The DPAA-Ethernet driver defines a ndo_select_queue() callback for optimal selection
-+ of the egress FQ. That will override the XPS support for this netdevice.
-+ If for whatever reason you want to be in control of the egress FQ-to-CPU selection and mapping,
-+ or simply don't want to use the driver's ndo_select_queue() callback, then unselect this
-+ and use the standard XPS support instead.
-+
-+config FSL_DPAA_ETH_MAX_BUF_COUNT
-+ int "Maximum nuber of buffers in private bpool"
-+ depends on FSL_SDK_DPAA_ETH
-+ range 64 2048
-+ default "128"
-+ help
-+ The maximum number of buffers to be by default allocated in the DPAA-Ethernet private port's
-+ buffer pool. One needn't normally modify this, as it has probably been tuned for performance
-+ already. This cannot be lower than DPAA_ETH_REFILL_THRESHOLD.
-+
-+config FSL_DPAA_ETH_REFILL_THRESHOLD
-+ int "Private bpool refill threshold"
-+ depends on FSL_SDK_DPAA_ETH
-+ range 32 FSL_DPAA_ETH_MAX_BUF_COUNT
-+ default "80"
-+ help
-+ The DPAA-Ethernet driver will start replenishing buffer pools whose count
-+ falls below this threshold. This must be related to DPAA_ETH_MAX_BUF_COUNT. One needn't normally
-+ modify this value unless one has very specific performance reasons.
-+
-+config FSL_DPAA_CS_THRESHOLD_1G
-+ hex "Egress congestion threshold on 1G ports"
-+ depends on FSL_SDK_DPAA_ETH
-+ range 0x1000 0x10000000
-+ default "0x06000000"
-+ help
-+ The size in bytes of the egress Congestion State notification threshold on 1G ports.
-+ The 1G dTSECs can quite easily be flooded by cores doing Tx in a tight loop
-+ (e.g. by sending UDP datagrams at "while(1) speed"),
-+ and the larger the frame size, the more acute the problem.
-+ So we have to find a balance between these factors:
-+ - avoiding the device staying congested for a prolonged time (risking
-+ the netdev watchdog to fire - see also the tx_timeout module param);
-+ - affecting performance of protocols such as TCP, which otherwise
-+ behave well under the congestion notification mechanism;
-+ - preventing the Tx cores from tightly-looping (as if the congestion
-+ threshold was too low to be effective);
-+ - running out of memory if the CS threshold is set too high.
-+
-+config FSL_DPAA_CS_THRESHOLD_10G
-+ hex "Egress congestion threshold on 10G ports"
-+ depends on FSL_SDK_DPAA_ETH
-+ range 0x1000 0x20000000
-+ default "0x10000000"
-+ help
-+ The size in bytes of the egress Congestion State notification threshold on 10G ports.
-+
-+config FSL_DPAA_INGRESS_CS_THRESHOLD
-+ hex "Ingress congestion threshold on FMan ports"
-+ depends on FSL_SDK_DPAA_ETH
-+ default "0x10000000"
-+ help
-+ The size in bytes of the ingress tail-drop threshold on FMan ports.
-+ Traffic piling up above this value will be rejected by QMan and discarded by FMan.
-+
-+config FSL_DPAA_ETH_DEBUGFS
-+ bool "DPAA Ethernet debugfs interface"
-+ depends on DEBUG_FS && FSL_SDK_DPAA_ETH
-+ default y
-+ help
-+ This option compiles debugfs code for the DPAA Ethernet driver.
-+
-+config FSL_DPAA_ETH_DEBUG
-+ bool "DPAA Ethernet Debug Support"
-+ depends on FSL_SDK_DPAA_ETH
-+ default n
-+ help
-+ This option compiles debug code for the DPAA Ethernet driver.
-+
-+config FSL_DPAA_DBG_LOOP
-+ bool "DPAA Ethernet Debug loopback"
-+ depends on FSL_DPAA_ETH_DEBUGFS && FSL_DPAA_ETH_USE_NDO_SELECT_QUEUE
-+ default n
-+ help
-+ This option allows to divert all received traffic on a certain interface A towards a
-+ selected interface B. This option is used to benchmark the HW + Ethernet driver in
-+ isolation from the Linux networking stack. The loops are controlled by debugfs entries,
-+ one for each interface. By default all loops are disabled (target value is -1). I.e. to
-+ change the loop setting for interface 4 and divert all received traffic to interface 5
-+ write Tx interface number in the receive interface debugfs file:
-+ # cat /sys/kernel/debug/powerpc/fsl_dpa/eth4_loop
-+ 4->-1
-+ # echo 5 > /sys/kernel/debug/powerpc/fsl_dpa/eth4_loop
-+ # cat /sys/kernel/debug/powerpc/fsl_dpa/eth4_loop
-+ 4->5
-+endif # FSL_SDK_DPAA_ETH
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_dpaa/Makefile
-@@ -0,0 +1,46 @@
-+#
-+# Makefile for the Freescale Ethernet controllers
-+#
-+ccflags-y += -DVERSION=\"\"
-+#
-+# Include netcomm SW specific definitions
-+include $(srctree)/drivers/net/ethernet/freescale/sdk_fman/ncsw_config.mk
-+
-+ccflags-y += -I$(NET_DPA)
-+
-+obj-$(CONFIG_FSL_SDK_DPAA_ETH) += fsl_mac.o fsl_dpa.o
-+obj-$(CONFIG_PTP_1588_CLOCK_DPAA) += dpaa_ptp.o
-+
-+fsl_dpa-objs += dpaa_ethtool.o dpaa_eth_sysfs.o dpaa_eth.o dpaa_eth_sg.o dpaa_eth_common.o
-+ifeq ($(CONFIG_FSL_DPAA_DBG_LOOP),y)
-+fsl_dpa-objs += dpaa_debugfs.o
-+endif
-+ifeq ($(CONFIG_FSL_DPAA_1588),y)
-+fsl_dpa-objs += dpaa_1588.o
-+endif
-+ifeq ($(CONFIG_FSL_DPAA_CEETM),y)
-+ccflags-y += -Idrivers/net/ethernet/freescale/sdk_fman/src/wrapper
-+fsl_dpa-objs += dpaa_eth_ceetm.o
-+endif
-+
-+fsl_mac-objs += mac.o mac-api.o
-+
-+# Advanced drivers
-+ifeq ($(CONFIG_FSL_DPAA_ADVANCED_DRIVERS),y)
-+obj-$(CONFIG_FSL_SDK_DPAA_ETH) += fsl_advanced.o
-+obj-$(CONFIG_FSL_SDK_DPAA_ETH) += fsl_proxy.o
-+
-+fsl_advanced-objs += dpaa_eth_base.o
-+# suport for multiple drivers per kernel module comes in kernel 3.14
-+# so we are forced to generate several modules for the advanced drivers
-+fsl_proxy-objs += dpaa_eth_proxy.o
-+
-+ifeq ($(CONFIG_FSL_DPAA_OFFLINE_PORTS),y)
-+obj-$(CONFIG_FSL_SDK_DPAA_ETH) += fsl_oh.o
-+
-+fsl_oh-objs += offline_port.o
-+endif
-+endif
-+
-+# Needed by the tracing framework
-+CFLAGS_dpaa_eth.o := -I$(src)
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_dpaa/dpaa_1588.c
-@@ -0,0 +1,580 @@
-+/* Copyright (C) 2011 Freescale Semiconductor, Inc.
-+ * Copyright (C) 2009 IXXAT Automation, GmbH
-+ *
-+ * DPAA Ethernet Driver -- IEEE 1588 interface functionality
-+ *
-+ * This program is free software; you can redistribute it and/or modify
-+ * it under the terms of the GNU General Public License as published by
-+ * the Free Software Foundation; either version 2 of the License, or
-+ * (at your option) any later version.
-+ *
-+ * This program is distributed in the hope that it will be useful,
-+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
-+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
-+ * GNU General Public License for more details.
-+ *
-+ * You should have received a copy of the GNU General Public License along
-+ * with this program; if not, write to the Free Software Foundation, Inc.,
-+ * 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA.
-+ *
-+ */
-+#include <linux/io.h>
-+#include <linux/device.h>
-+#include <linux/fs.h>
-+#include <linux/vmalloc.h>
-+#include <linux/spinlock.h>
-+#include <linux/ip.h>
-+#include <linux/ipv6.h>
-+#include <linux/udp.h>
-+#include <asm/div64.h>
-+#include "dpaa_eth.h"
-+#include "dpaa_eth_common.h"
-+#include "dpaa_1588.h"
-+#include "mac.h"
-+
-+static int dpa_ptp_init_circ(struct dpa_ptp_circ_buf *ptp_buf, u32 size)
-+{
-+ struct circ_buf *circ_buf = &ptp_buf->circ_buf;
-+
-+ circ_buf->buf = vmalloc(sizeof(struct dpa_ptp_data) * size);
-+ if (!circ_buf->buf)
-+ return 1;
-+
-+ circ_buf->head = 0;
-+ circ_buf->tail = 0;
-+ ptp_buf->size = size;
-+ spin_lock_init(&ptp_buf->ptp_lock);
-+
-+ return 0;
-+}
-+
-+static void dpa_ptp_reset_circ(struct dpa_ptp_circ_buf *ptp_buf, u32 size)
-+{
-+ struct circ_buf *circ_buf = &ptp_buf->circ_buf;
-+
-+ circ_buf->head = 0;
-+ circ_buf->tail = 0;
-+ ptp_buf->size = size;
-+}
-+
-+static int dpa_ptp_insert(struct dpa_ptp_circ_buf *ptp_buf,
-+ struct dpa_ptp_data *data)
-+{
-+ struct circ_buf *circ_buf = &ptp_buf->circ_buf;
-+ int size = ptp_buf->size;
-+ struct dpa_ptp_data *tmp;
-+ unsigned long flags;
-+ int head, tail;
-+
-+ spin_lock_irqsave(&ptp_buf->ptp_lock, flags);
-+
-+ head = circ_buf->head;
-+ tail = circ_buf->tail;
-+
-+ if (CIRC_SPACE(head, tail, size) <= 0)
-+ circ_buf->tail = (tail + 1) & (size - 1);
-+
-+ tmp = (struct dpa_ptp_data *)(circ_buf->buf) + head;
-+ memcpy(tmp, data, sizeof(struct dpa_ptp_data));
-+
-+ circ_buf->head = (head + 1) & (size - 1);
-+
-+ spin_unlock_irqrestore(&ptp_buf->ptp_lock, flags);
-+
-+ return 0;
-+}
-+
-+static int dpa_ptp_is_ident_match(struct dpa_ptp_ident *dst,
-+ struct dpa_ptp_ident *src)
-+{
-+ int ret;
-+
-+ if ((dst->version != src->version) || (dst->msg_type != src->msg_type))
-+ return 0;
-+
-+ if ((dst->netw_prot == src->netw_prot)
-+ || src->netw_prot == DPA_PTP_PROT_DONTCARE) {
-+ if (dst->seq_id != src->seq_id)
-+ return 0;
-+
-+ ret = memcmp(dst->snd_port_id, src->snd_port_id,
-+ DPA_PTP_SOURCE_PORT_LENGTH);
-+ if (ret)
-+ return 0;
-+ else
-+ return 1;
-+ }
-+
-+ return 0;
-+}
-+
-+static int dpa_ptp_find_and_remove(struct dpa_ptp_circ_buf *ptp_buf,
-+ struct dpa_ptp_ident *ident,
-+ struct dpa_ptp_time *ts)
-+{
-+ struct circ_buf *circ_buf = &ptp_buf->circ_buf;
-+ int size = ptp_buf->size;
-+ int head, tail, idx;
-+ unsigned long flags;
-+ struct dpa_ptp_data *tmp, *tmp2;
-+ struct dpa_ptp_ident *tmp_ident;
-+
-+ spin_lock_irqsave(&ptp_buf->ptp_lock, flags);
-+
-+ head = circ_buf->head;
-+ tail = idx = circ_buf->tail;
-+
-+ if (CIRC_CNT(head, tail, size) == 0) {
-+ spin_unlock_irqrestore(&ptp_buf->ptp_lock, flags);
-+ return 1;
-+ }
-+
-+ while (idx != head) {
-+ tmp = (struct dpa_ptp_data *)(circ_buf->buf) + idx;
-+ tmp_ident = &tmp->ident;
-+ if (dpa_ptp_is_ident_match(tmp_ident, ident))
-+ break;
-+ idx = (idx + 1) & (size - 1);
-+ }
-+
-+ if (idx == head) {
-+ spin_unlock_irqrestore(&ptp_buf->ptp_lock, flags);
-+ return 1;
-+ }
-+
-+ ts->sec = tmp->ts.sec;
-+ ts->nsec = tmp->ts.nsec;
-+
-+ if (idx != tail) {
-+ if (CIRC_CNT(idx, tail, size) > TS_ACCUMULATION_THRESHOLD) {
-+ tail = circ_buf->tail =
-+ (idx - TS_ACCUMULATION_THRESHOLD) & (size - 1);
-+ }
-+
-+ while (CIRC_CNT(idx, tail, size) > 0) {
-+ tmp = (struct dpa_ptp_data *)(circ_buf->buf) + idx;
-+ idx = (idx - 1) & (size - 1);
-+ tmp2 = (struct dpa_ptp_data *)(circ_buf->buf) + idx;
-+ *tmp = *tmp2;
-+ }
-+ }
-+ circ_buf->tail = (tail + 1) & (size - 1);
-+
-+ spin_unlock_irqrestore(&ptp_buf->ptp_lock, flags);
-+
-+ return 0;
-+}
-+
-+/* Parse the PTP packets
-+ *
-+ * The PTP header can be found in an IPv4 packet, IPv6 patcket or in
-+ * an IEEE802.3 ethernet frame. This function returns the position of
-+ * the PTP packet or NULL if no PTP found
-+ */
-+static u8 *dpa_ptp_parse_packet(struct sk_buff *skb, u16 *eth_type)
-+{
-+ u8 *pos = skb->data + ETH_ALEN + ETH_ALEN;
-+ u8 *ptp_loc = NULL;
-+ u8 msg_type;
-+ u32 access_len = ETH_ALEN + ETH_ALEN + DPA_ETYPE_LEN;
-+ struct iphdr *iph;
-+ struct udphdr *udph;
-+ struct ipv6hdr *ipv6h;
-+
-+ /* when we can receive S/G frames we need to check the data we want to
-+ * access is in the linear skb buffer
-+ */
-+ if (!pskb_may_pull(skb, access_len))
-+ return NULL;
-+
-+ *eth_type = *((u16 *)pos);
-+
-+ /* Check if inner tag is here */
-+ if (*eth_type == ETH_P_8021Q) {
-+ access_len += DPA_VLAN_TAG_LEN;
-+
-+ if (!pskb_may_pull(skb, access_len))
-+ return NULL;
-+
-+ pos += DPA_VLAN_TAG_LEN;
-+ *eth_type = *((u16 *)pos);
-+ }
-+
-+ pos += DPA_ETYPE_LEN;
-+
-+ switch (*eth_type) {
-+ /* Transport of PTP over Ethernet */
-+ case ETH_P_1588:
-+ ptp_loc = pos;
-+
-+ if (!pskb_may_pull(skb, access_len + PTP_OFFS_MSG_TYPE + 1))
-+ return NULL;
-+
-+ msg_type = *((u8 *)(ptp_loc + PTP_OFFS_MSG_TYPE)) & 0xf;
-+ if ((msg_type == PTP_MSGTYPE_SYNC)
-+ || (msg_type == PTP_MSGTYPE_DELREQ)
-+ || (msg_type == PTP_MSGTYPE_PDELREQ)
-+ || (msg_type == PTP_MSGTYPE_PDELRESP))
-+ return ptp_loc;
-+ break;
-+ /* Transport of PTP over IPv4 */
-+ case ETH_P_IP:
-+ iph = (struct iphdr *)pos;
-+ access_len += sizeof(struct iphdr);
-+
-+ if (!pskb_may_pull(skb, access_len))
-+ return NULL;
-+
-+ if (ntohs(iph->protocol) != IPPROTO_UDP)
-+ return NULL;
-+
-+ access_len += iph->ihl * 4 - sizeof(struct iphdr) +
-+ sizeof(struct udphdr);
-+
-+ if (!pskb_may_pull(skb, access_len))
-+ return NULL;
-+
-+ pos += iph->ihl * 4;
-+ udph = (struct udphdr *)pos;
-+ if (ntohs(udph->dest) != 319)
-+ return NULL;
-+ ptp_loc = pos + sizeof(struct udphdr);
-+ break;
-+ /* Transport of PTP over IPv6 */
-+ case ETH_P_IPV6:
-+ ipv6h = (struct ipv6hdr *)pos;
-+
-+ access_len += sizeof(struct ipv6hdr) + sizeof(struct udphdr);
-+
-+ if (ntohs(ipv6h->nexthdr) != IPPROTO_UDP)
-+ return NULL;
-+
-+ pos += sizeof(struct ipv6hdr);
-+ udph = (struct udphdr *)pos;
-+ if (ntohs(udph->dest) != 319)
-+ return NULL;
-+ ptp_loc = pos + sizeof(struct udphdr);
-+ break;
-+ default:
-+ break;
-+ }
-+
-+ return ptp_loc;
-+}
-+
-+static int dpa_ptp_store_stamp(const struct dpa_priv_s *priv,
-+ struct sk_buff *skb, void *data, enum port_type rx_tx,
-+ struct dpa_ptp_data *ptp_data)
-+{
-+ u64 nsec;
-+ u32 mod;
-+ u8 *ptp_loc;
-+ u16 eth_type;
-+
-+ ptp_loc = dpa_ptp_parse_packet(skb, &eth_type);
-+ if (!ptp_loc)
-+ return -EINVAL;
-+
-+ switch (eth_type) {
-+ case ETH_P_IP:
-+ ptp_data->ident.netw_prot = DPA_PTP_PROT_IPV4;
-+ break;
-+ case ETH_P_IPV6:
-+ ptp_data->ident.netw_prot = DPA_PTP_PROT_IPV6;
-+ break;
-+ case ETH_P_1588:
-+ ptp_data->ident.netw_prot = DPA_PTP_PROT_802_3;
-+ break;
-+ default:
-+ return -EINVAL;
-+ }
-+
-+ if (!pskb_may_pull(skb, ptp_loc - skb->data + PTP_OFFS_SEQ_ID + 2))
-+ return -EINVAL;
-+
-+ ptp_data->ident.version = *(ptp_loc + PTP_OFFS_VER_PTP) & 0xf;
-+ ptp_data->ident.msg_type = *(ptp_loc + PTP_OFFS_MSG_TYPE) & 0xf;
-+ ptp_data->ident.seq_id = *((u16 *)(ptp_loc + PTP_OFFS_SEQ_ID));
-+ memcpy(ptp_data->ident.snd_port_id, ptp_loc + PTP_OFFS_SRCPRTID,
-+ DPA_PTP_SOURCE_PORT_LENGTH);
-+
-+ nsec = dpa_get_timestamp_ns(priv, rx_tx, data);
-+ mod = do_div(nsec, NANOSEC_PER_SECOND);
-+ ptp_data->ts.sec = nsec;
-+ ptp_data->ts.nsec = mod;
-+
-+ return 0;
-+}
-+
-+void dpa_ptp_store_txstamp(const struct dpa_priv_s *priv,
-+ struct sk_buff *skb, void *data)
-+{
-+ struct dpa_ptp_tsu *tsu = priv->tsu;
-+ struct dpa_ptp_data ptp_tx_data;
-+
-+ if (dpa_ptp_store_stamp(priv, skb, data, TX, &ptp_tx_data))
-+ return;
-+
-+ dpa_ptp_insert(&tsu->tx_timestamps, &ptp_tx_data);
-+}
-+
-+void dpa_ptp_store_rxstamp(const struct dpa_priv_s *priv,
-+ struct sk_buff *skb, void *data)
-+{
-+ struct dpa_ptp_tsu *tsu = priv->tsu;
-+ struct dpa_ptp_data ptp_rx_data;
-+
-+ if (dpa_ptp_store_stamp(priv, skb, data, RX, &ptp_rx_data))
-+ return;
-+
-+ dpa_ptp_insert(&tsu->rx_timestamps, &ptp_rx_data);
-+}
-+
-+static uint8_t dpa_get_tx_timestamp(struct dpa_ptp_tsu *ptp_tsu,
-+ struct dpa_ptp_ident *ident,
-+ struct dpa_ptp_time *ts)
-+{
-+ struct dpa_ptp_tsu *tsu = ptp_tsu;
-+ struct dpa_ptp_time tmp;
-+ int flag;
-+
-+ flag = dpa_ptp_find_and_remove(&tsu->tx_timestamps, ident, &tmp);
-+ if (!flag) {
-+ ts->sec = tmp.sec;
-+ ts->nsec = tmp.nsec;
-+ return 0;
-+ }
-+
-+ return -1;
-+}
-+
-+static uint8_t dpa_get_rx_timestamp(struct dpa_ptp_tsu *ptp_tsu,
-+ struct dpa_ptp_ident *ident,
-+ struct dpa_ptp_time *ts)
-+{
-+ struct dpa_ptp_tsu *tsu = ptp_tsu;
-+ struct dpa_ptp_time tmp;
-+ int flag;
-+
-+ flag = dpa_ptp_find_and_remove(&tsu->rx_timestamps, ident, &tmp);
-+ if (!flag) {
-+ ts->sec = tmp.sec;
-+ ts->nsec = tmp.nsec;
-+ return 0;
-+ }
-+
-+ return -1;
-+}
-+
-+static void dpa_set_fiper_alarm(struct dpa_ptp_tsu *tsu,
-+ struct dpa_ptp_time *cnt_time)
-+{
-+ struct mac_device *mac_dev = tsu->dpa_priv->mac_dev;
-+ u64 tmp, fiper;
-+
-+ if (mac_dev->fm_rtc_disable)
-+ mac_dev->fm_rtc_disable(get_fm_handle(tsu->dpa_priv->net_dev));
-+
-+ /* TMR_FIPER1 will pulse every second after ALARM1 expired */
-+ tmp = (u64)cnt_time->sec * NANOSEC_PER_SECOND + (u64)cnt_time->nsec;
-+ fiper = NANOSEC_PER_SECOND - DPA_PTP_NOMINAL_FREQ_PERIOD_NS;
-+ if (mac_dev->fm_rtc_set_alarm)
-+ mac_dev->fm_rtc_set_alarm(get_fm_handle(tsu->dpa_priv->net_dev),
-+ 0, tmp);
-+ if (mac_dev->fm_rtc_set_fiper)
-+ mac_dev->fm_rtc_set_fiper(get_fm_handle(tsu->dpa_priv->net_dev),
-+ 0, fiper);
-+
-+ if (mac_dev->fm_rtc_enable)
-+ mac_dev->fm_rtc_enable(get_fm_handle(tsu->dpa_priv->net_dev));
-+}
-+
-+static void dpa_get_curr_cnt(struct dpa_ptp_tsu *tsu,
-+ struct dpa_ptp_time *curr_time)
-+{
-+ struct mac_device *mac_dev = tsu->dpa_priv->mac_dev;
-+ u64 tmp;
-+ u32 mod;
-+
-+ if (mac_dev->fm_rtc_get_cnt)
-+ mac_dev->fm_rtc_get_cnt(get_fm_handle(tsu->dpa_priv->net_dev),
-+ &tmp);
-+
-+ mod = do_div(tmp, NANOSEC_PER_SECOND);
-+ curr_time->sec = (u32)tmp;
-+ curr_time->nsec = mod;
-+}
-+
-+static void dpa_set_1588cnt(struct dpa_ptp_tsu *tsu,
-+ struct dpa_ptp_time *cnt_time)
-+{
-+ struct mac_device *mac_dev = tsu->dpa_priv->mac_dev;
-+ u64 tmp;
-+
-+ tmp = (u64)cnt_time->sec * NANOSEC_PER_SECOND + (u64)cnt_time->nsec;
-+
-+ if (mac_dev->fm_rtc_set_cnt)
-+ mac_dev->fm_rtc_set_cnt(get_fm_handle(tsu->dpa_priv->net_dev),
-+ tmp);
-+
-+ /* Restart fiper two seconds later */
-+ cnt_time->sec += 2;
-+ cnt_time->nsec = 0;
-+ dpa_set_fiper_alarm(tsu, cnt_time);
-+}
-+
-+static void dpa_get_drift(struct dpa_ptp_tsu *tsu, u32 *addend)
-+{
-+ struct mac_device *mac_dev = tsu->dpa_priv->mac_dev;
-+ u32 drift;
-+
-+ if (mac_dev->fm_rtc_get_drift)
-+ mac_dev->fm_rtc_get_drift(get_fm_handle(tsu->dpa_priv->net_dev),
-+ &drift);
-+
-+ *addend = drift;
-+}
-+
-+static void dpa_set_drift(struct dpa_ptp_tsu *tsu, u32 addend)
-+{
-+ struct mac_device *mac_dev = tsu->dpa_priv->mac_dev;
-+
-+ if (mac_dev->fm_rtc_set_drift)
-+ mac_dev->fm_rtc_set_drift(get_fm_handle(tsu->dpa_priv->net_dev),
-+ addend);
-+}
-+
-+static void dpa_flush_timestamp(struct dpa_ptp_tsu *tsu)
-+{
-+ dpa_ptp_reset_circ(&tsu->rx_timestamps, DEFAULT_PTP_RX_BUF_SZ);
-+ dpa_ptp_reset_circ(&tsu->tx_timestamps, DEFAULT_PTP_TX_BUF_SZ);
-+}
-+
-+int dpa_ioctl_1588(struct net_device *dev, struct ifreq *ifr, int cmd)
-+{
-+ struct dpa_priv_s *priv = netdev_priv(dev);
-+ struct dpa_ptp_tsu *tsu = priv->tsu;
-+ struct mac_device *mac_dev = priv->mac_dev;
-+ struct dpa_ptp_data ptp_data;
-+ struct dpa_ptp_data *ptp_data_user;
-+ struct dpa_ptp_time act_time;
-+ u32 addend;
-+ int retval = 0;
-+
-+ if (!tsu || !tsu->valid)
-+ return -ENODEV;
-+
-+ switch (cmd) {
-+ case PTP_ENBL_TXTS_IOCTL:
-+ tsu->hwts_tx_en_ioctl = 1;
-+ if (mac_dev->fm_rtc_enable)
-+ mac_dev->fm_rtc_enable(get_fm_handle(dev));
-+ if (mac_dev->ptp_enable)
-+ mac_dev->ptp_enable(mac_dev->get_mac_handle(mac_dev));
-+ break;
-+ case PTP_DSBL_TXTS_IOCTL:
-+ tsu->hwts_tx_en_ioctl = 0;
-+ if (mac_dev->fm_rtc_disable)
-+ mac_dev->fm_rtc_disable(get_fm_handle(dev));
-+ if (mac_dev->ptp_disable)
-+ mac_dev->ptp_disable(mac_dev->get_mac_handle(mac_dev));
-+ break;
-+ case PTP_ENBL_RXTS_IOCTL:
-+ tsu->hwts_rx_en_ioctl = 1;
-+ break;
-+ case PTP_DSBL_RXTS_IOCTL:
-+ tsu->hwts_rx_en_ioctl = 0;
-+ break;
-+ case PTP_GET_RX_TIMESTAMP:
-+ ptp_data_user = (struct dpa_ptp_data *)ifr->ifr_data;
-+ if (copy_from_user(&ptp_data.ident,
-+ &ptp_data_user->ident, sizeof(ptp_data.ident)))
-+ return -EINVAL;
-+
-+ if (dpa_get_rx_timestamp(tsu, &ptp_data.ident, &ptp_data.ts))
-+ return -EAGAIN;
-+
-+ if (copy_to_user((void __user *)&ptp_data_user->ts,
-+ &ptp_data.ts, sizeof(ptp_data.ts)))
-+ return -EFAULT;
-+ break;
-+ case PTP_GET_TX_TIMESTAMP:
-+ ptp_data_user = (struct dpa_ptp_data *)ifr->ifr_data;
-+ if (copy_from_user(&ptp_data.ident,
-+ &ptp_data_user->ident, sizeof(ptp_data.ident)))
-+ return -EINVAL;
-+
-+ if (dpa_get_tx_timestamp(tsu, &ptp_data.ident, &ptp_data.ts))
-+ return -EAGAIN;
-+
-+ if (copy_to_user((void __user *)&ptp_data_user->ts,
-+ &ptp_data.ts, sizeof(ptp_data.ts)))
-+ return -EFAULT;
-+ break;
-+ case PTP_GET_TIME:
-+ dpa_get_curr_cnt(tsu, &act_time);
-+ if (copy_to_user(ifr->ifr_data, &act_time, sizeof(act_time)))
-+ return -EFAULT;
-+ break;
-+ case PTP_SET_TIME:
-+ if (copy_from_user(&act_time, ifr->ifr_data, sizeof(act_time)))
-+ return -EINVAL;
-+ dpa_set_1588cnt(tsu, &act_time);
-+ break;
-+ case PTP_GET_ADJ:
-+ dpa_get_drift(tsu, &addend);
-+ if (copy_to_user(ifr->ifr_data, &addend, sizeof(addend)))
-+ return -EFAULT;
-+ break;
-+ case PTP_SET_ADJ:
-+ if (copy_from_user(&addend, ifr->ifr_data, sizeof(addend)))
-+ return -EINVAL;
-+ dpa_set_drift(tsu, addend);
-+ break;
-+ case PTP_SET_FIPER_ALARM:
-+ if (copy_from_user(&act_time, ifr->ifr_data, sizeof(act_time)))
-+ return -EINVAL;
-+ dpa_set_fiper_alarm(tsu, &act_time);
-+ break;
-+ case PTP_CLEANUP_TS:
-+ dpa_flush_timestamp(tsu);
-+ break;
-+ default:
-+ return -EINVAL;
-+ }
-+
-+ return retval;
-+}
-+
-+int dpa_ptp_init(struct dpa_priv_s *priv)
-+{
-+ struct dpa_ptp_tsu *tsu;
-+
-+ /* Allocate memory for PTP structure */
-+ tsu = kzalloc(sizeof(struct dpa_ptp_tsu), GFP_KERNEL);
-+ if (!tsu)
-+ return -ENOMEM;
-+
-+ tsu->valid = TRUE;
-+ tsu->dpa_priv = priv;
-+
-+ dpa_ptp_init_circ(&tsu->rx_timestamps, DEFAULT_PTP_RX_BUF_SZ);
-+ dpa_ptp_init_circ(&tsu->tx_timestamps, DEFAULT_PTP_TX_BUF_SZ);
-+
-+ priv->tsu = tsu;
-+
-+ return 0;
-+}
-+EXPORT_SYMBOL(dpa_ptp_init);
-+
-+void dpa_ptp_cleanup(struct dpa_priv_s *priv)
-+{
-+ struct dpa_ptp_tsu *tsu = priv->tsu;
-+
-+ tsu->valid = FALSE;
-+ vfree(tsu->rx_timestamps.circ_buf.buf);
-+ vfree(tsu->tx_timestamps.circ_buf.buf);
-+
-+ kfree(tsu);
-+}
-+EXPORT_SYMBOL(dpa_ptp_cleanup);
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_dpaa/dpaa_1588.h
-@@ -0,0 +1,138 @@
-+/* Copyright (C) 2011 Freescale Semiconductor, Inc.
-+ *
-+ * This program is free software; you can redistribute it and/or modify
-+ * it under the terms of the GNU General Public License as published by
-+ * the Free Software Foundation; either version 2 of the License, or
-+ * (at your option) any later version.
-+ *
-+ * This program is distributed in the hope that it will be useful,
-+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
-+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
-+ * GNU General Public License for more details.
-+ *
-+ * You should have received a copy of the GNU General Public License along
-+ * with this program; if not, write to the Free Software Foundation, Inc.,
-+ * 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA.
-+ *
-+ */
-+#ifndef __DPAA_1588_H__
-+#define __DPAA_1588_H__
-+
-+#include <linux/netdevice.h>
-+#include <linux/etherdevice.h>
-+#include <linux/circ_buf.h>
-+#include <linux/fsl_qman.h>
-+
-+#define DEFAULT_PTP_RX_BUF_SZ 256
-+#define DEFAULT_PTP_TX_BUF_SZ 256
-+
-+/* 1588 private ioctl calls */
-+#define PTP_ENBL_TXTS_IOCTL SIOCDEVPRIVATE
-+#define PTP_DSBL_TXTS_IOCTL (SIOCDEVPRIVATE + 1)
-+#define PTP_ENBL_RXTS_IOCTL (SIOCDEVPRIVATE + 2)
-+#define PTP_DSBL_RXTS_IOCTL (SIOCDEVPRIVATE + 3)
-+#define PTP_GET_TX_TIMESTAMP (SIOCDEVPRIVATE + 4)
-+#define PTP_GET_RX_TIMESTAMP (SIOCDEVPRIVATE + 5)
-+#define PTP_SET_TIME (SIOCDEVPRIVATE + 6)
-+#define PTP_GET_TIME (SIOCDEVPRIVATE + 7)
-+#define PTP_SET_FIPER_ALARM (SIOCDEVPRIVATE + 8)
-+#define PTP_SET_ADJ (SIOCDEVPRIVATE + 9)
-+#define PTP_GET_ADJ (SIOCDEVPRIVATE + 10)
-+#define PTP_CLEANUP_TS (SIOCDEVPRIVATE + 11)
-+
-+/* PTP V2 message type */
-+enum {
-+ PTP_MSGTYPE_SYNC = 0x0,
-+ PTP_MSGTYPE_DELREQ = 0x1,
-+ PTP_MSGTYPE_PDELREQ = 0x2,
-+ PTP_MSGTYPE_PDELRESP = 0x3,
-+ PTP_MSGTYPE_FLWUP = 0x8,
-+ PTP_MSGTYPE_DELRESP = 0x9,
-+ PTP_MSGTYPE_PDELRES_FLWUP = 0xA,
-+ PTP_MSGTYPE_ANNOUNCE = 0xB,
-+ PTP_MSGTYPE_SGNLNG = 0xC,
-+ PTP_MSGTYPE_MNGMNT = 0xD,
-+};
-+
-+/* Byte offset of data in the PTP V2 headers */
-+#define PTP_OFFS_MSG_TYPE 0
-+#define PTP_OFFS_VER_PTP 1
-+#define PTP_OFFS_MSG_LEN 2
-+#define PTP_OFFS_DOM_NMB 4
-+#define PTP_OFFS_FLAGS 6
-+#define PTP_OFFS_CORFIELD 8
-+#define PTP_OFFS_SRCPRTID 20
-+#define PTP_OFFS_SEQ_ID 30
-+#define PTP_OFFS_CTRL 32
-+#define PTP_OFFS_LOGMEAN 33
-+
-+#define PTP_IP_OFFS 14
-+#define PTP_UDP_OFFS 34
-+#define PTP_HEADER_OFFS 42
-+#define PTP_MSG_TYPE_OFFS (PTP_HEADER_OFFS + PTP_OFFS_MSG_TYPE)
-+#define PTP_SPORT_ID_OFFS (PTP_HEADER_OFFS + PTP_OFFS_SRCPRTID)
-+#define PTP_SEQ_ID_OFFS (PTP_HEADER_OFFS + PTP_OFFS_SEQ_ID)
-+#define PTP_CTRL_OFFS (PTP_HEADER_OFFS + PTP_OFFS_CTRL)
-+
-+/* 1588-2008 network protocol enumeration values */
-+#define DPA_PTP_PROT_IPV4 1
-+#define DPA_PTP_PROT_IPV6 2
-+#define DPA_PTP_PROT_802_3 3
-+#define DPA_PTP_PROT_DONTCARE 0xFFFF
-+
-+#define DPA_PTP_SOURCE_PORT_LENGTH 10
-+#define DPA_PTP_HEADER_SZE 34
-+#define DPA_ETYPE_LEN 2
-+#define DPA_VLAN_TAG_LEN 4
-+#define NANOSEC_PER_SECOND 1000000000
-+
-+/* The threshold between the current found one and the oldest one */
-+#define TS_ACCUMULATION_THRESHOLD 50
-+
-+/* Struct needed to identify a timestamp */
-+struct dpa_ptp_ident {
-+ u8 version;
-+ u8 msg_type;
-+ u16 netw_prot;
-+ u16 seq_id;
-+ u8 snd_port_id[DPA_PTP_SOURCE_PORT_LENGTH];
-+};
-+
-+/* Timestamp format in 1588-2008 */
-+struct dpa_ptp_time {
-+ u64 sec; /* just 48 bit used */
-+ u32 nsec;
-+};
-+
-+/* needed for timestamp data over ioctl */
-+struct dpa_ptp_data {
-+ struct dpa_ptp_ident ident;
-+ struct dpa_ptp_time ts;
-+};
-+
-+struct dpa_ptp_circ_buf {
-+ struct circ_buf circ_buf;
-+ u32 size;
-+ spinlock_t ptp_lock;
-+};
-+
-+/* PTP TSU control structure */
-+struct dpa_ptp_tsu {
-+ struct dpa_priv_s *dpa_priv;
-+ bool valid;
-+ struct dpa_ptp_circ_buf rx_timestamps;
-+ struct dpa_ptp_circ_buf tx_timestamps;
-+
-+ /* HW timestamping over ioctl enabled flag */
-+ int hwts_tx_en_ioctl;
-+ int hwts_rx_en_ioctl;
-+};
-+
-+extern int dpa_ptp_init(struct dpa_priv_s *priv);
-+extern void dpa_ptp_cleanup(struct dpa_priv_s *priv);
-+extern void dpa_ptp_store_txstamp(const struct dpa_priv_s *priv,
-+ struct sk_buff *skb, void *data);
-+extern void dpa_ptp_store_rxstamp(const struct dpa_priv_s *priv,
-+ struct sk_buff *skb, void *data);
-+extern int dpa_ioctl_1588(struct net_device *dev, struct ifreq *ifr, int cmd);
-+#endif
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_dpaa/dpaa_debugfs.c
-@@ -0,0 +1,180 @@
-+/* Copyright 2008-2013 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#include <linux/module.h>
-+#include <linux/fsl_qman.h> /* struct qm_mcr_querycgr */
-+#include <linux/debugfs.h>
-+#include "dpaa_debugfs.h"
-+#include "dpaa_eth.h" /* struct dpa_priv_s, dpa_percpu_priv_s, dpa_bp */
-+
-+#define DPA_DEBUGFS_DESCRIPTION "FSL DPAA Ethernet debugfs entries"
-+#define DPA_ETH_DEBUGFS_ROOT "fsl_dpa"
-+
-+static struct dentry *dpa_debugfs_root;
-+
-+static int __cold dpa_debugfs_loop_open(struct inode *inode, struct file *file);
-+static ssize_t dpa_loop_write(struct file *f,
-+ const char __user *buf, size_t count, loff_t *off);
-+
-+static const struct file_operations dpa_debugfs_lp_fops = {
-+ .open = dpa_debugfs_loop_open,
-+ .write = dpa_loop_write,
-+ .read = seq_read,
-+ .llseek = seq_lseek,
-+ .release = single_release,
-+};
-+
-+static int dpa_debugfs_loop_show(struct seq_file *file, void *offset)
-+{
-+ struct dpa_priv_s *priv;
-+
-+ BUG_ON(offset == NULL);
-+
-+ priv = netdev_priv((struct net_device *)file->private);
-+ seq_printf(file, "%d->%d\n", priv->loop_id, priv->loop_to);
-+
-+ return 0;
-+}
-+
-+static int user_input_convert(const char __user *user_buf, size_t count,
-+ long *val)
-+{
-+ char buf[12];
-+
-+ if (count > sizeof(buf) - 1)
-+ return -EINVAL;
-+ if (copy_from_user(buf, user_buf, count))
-+ return -EFAULT;
-+ buf[count] = '\0';
-+ if (kstrtol(buf, 0, val))
-+ return -EINVAL;
-+ return 0;
-+}
-+
-+static ssize_t dpa_loop_write(struct file *f,
-+ const char __user *buf, size_t count, loff_t *off)
-+{
-+ struct dpa_priv_s *priv;
-+ struct net_device *netdev;
-+ struct seq_file *sf;
-+ int ret;
-+ long val;
-+
-+ ret = user_input_convert(buf, count, &val);
-+ if (ret)
-+ return ret;
-+
-+ sf = (struct seq_file *)f->private_data;
-+ netdev = (struct net_device *)sf->private;
-+ priv = netdev_priv(netdev);
-+
-+ priv->loop_to = ((val < 0) || (val > 20)) ? -1 : val;
-+
-+ return count;
-+}
-+
-+static int __cold dpa_debugfs_loop_open(struct inode *inode, struct file *file)
-+{
-+ int _errno;
-+ const struct net_device *net_dev;
-+
-+ _errno = single_open(file, dpa_debugfs_loop_show, inode->i_private);
-+ if (unlikely(_errno < 0)) {
-+ net_dev = (struct net_device *)inode->i_private;
-+
-+ if (netif_msg_drv((struct dpa_priv_s *)netdev_priv(net_dev)))
-+ netdev_err(net_dev, "single_open() = %d\n",
-+ _errno);
-+ }
-+
-+ return _errno;
-+}
-+
-+
-+int dpa_netdev_debugfs_create(struct net_device *net_dev)
-+{
-+ struct dpa_priv_s *priv = netdev_priv(net_dev);
-+ static int cnt;
-+ char loop_file_name[100];
-+
-+ if (unlikely(dpa_debugfs_root == NULL)) {
-+ pr_err(KBUILD_MODNAME ": %s:%hu:%s(): \t%s\n",
-+ KBUILD_BASENAME".c", __LINE__, __func__,
-+ "root debugfs missing, possible module ordering issue");
-+ return -ENOMEM;
-+ }
-+
-+ sprintf(loop_file_name, "eth%d_loop", ++cnt);
-+ priv->debugfs_loop_file = debugfs_create_file(loop_file_name,
-+ S_IRUGO,
-+ dpa_debugfs_root,
-+ net_dev,
-+ &dpa_debugfs_lp_fops);
-+ if (unlikely(priv->debugfs_loop_file == NULL)) {
-+ netdev_err(net_dev, "debugfs_create_file(%s/%s)",
-+ dpa_debugfs_root->d_iname,
-+ loop_file_name);
-+
-+ return -ENOMEM;
-+ }
-+ return 0;
-+}
-+
-+void dpa_netdev_debugfs_remove(struct net_device *net_dev)
-+{
-+ struct dpa_priv_s *priv = netdev_priv(net_dev);
-+
-+ debugfs_remove(priv->debugfs_loop_file);
-+}
-+
-+int __init dpa_debugfs_module_init(void)
-+{
-+ int _errno = 0;
-+
-+ pr_info(KBUILD_MODNAME ": " DPA_DEBUGFS_DESCRIPTION "\n");
-+
-+ dpa_debugfs_root = debugfs_create_dir(DPA_ETH_DEBUGFS_ROOT, NULL);
-+
-+ if (unlikely(dpa_debugfs_root == NULL)) {
-+ _errno = -ENOMEM;
-+ pr_err(KBUILD_MODNAME ": %s:%hu:%s():\n",
-+ KBUILD_BASENAME".c", __LINE__, __func__);
-+ pr_err("\tdebugfs_create_dir(%s/"KBUILD_MODNAME") = %d\n",
-+ DPA_ETH_DEBUGFS_ROOT, _errno);
-+ }
-+
-+ return _errno;
-+}
-+
-+void __exit dpa_debugfs_module_exit(void)
-+{
-+ debugfs_remove(dpa_debugfs_root);
-+}
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_dpaa/dpaa_debugfs.h
-@@ -0,0 +1,43 @@
-+/* Copyright 2008-2013 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#ifndef DPAA_DEBUGFS_H_
-+#define DPAA_DEBUGFS_H_
-+
-+#include <linux/netdevice.h>
-+#include <linux/dcache.h> /* struct dentry needed in dpaa_eth.h */
-+
-+int dpa_netdev_debugfs_create(struct net_device *net_dev);
-+void dpa_netdev_debugfs_remove(struct net_device *net_dev);
-+int __init dpa_debugfs_module_init(void);
-+void __exit dpa_debugfs_module_exit(void);
-+
-+#endif /* DPAA_DEBUGFS_H_ */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_dpaa/dpaa_eth.c
-@@ -0,0 +1,1224 @@
-+/* Copyright 2008-2013 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#ifdef CONFIG_FSL_DPAA_ETH_DEBUG
-+#define pr_fmt(fmt) \
-+ KBUILD_MODNAME ": %s:%hu:%s() " fmt, \
-+ KBUILD_BASENAME".c", __LINE__, __func__
-+#else
-+#define pr_fmt(fmt) \
-+ KBUILD_MODNAME ": " fmt
-+#endif
-+
-+#include <linux/init.h>
-+#include <linux/module.h>
-+#include <linux/of_mdio.h>
-+#include <linux/of_net.h>
-+#include <linux/kthread.h>
-+#include <linux/io.h>
-+#include <linux/if_arp.h> /* arp_hdr_len() */
-+#include <linux/if_vlan.h> /* VLAN_HLEN */
-+#include <linux/icmp.h> /* struct icmphdr */
-+#include <linux/ip.h> /* struct iphdr */
-+#include <linux/ipv6.h> /* struct ipv6hdr */
-+#include <linux/udp.h> /* struct udphdr */
-+#include <linux/tcp.h> /* struct tcphdr */
-+#include <linux/net.h> /* net_ratelimit() */
-+#include <linux/if_ether.h> /* ETH_P_IP and ETH_P_IPV6 */
-+#include <linux/highmem.h>
-+#include <linux/percpu.h>
-+#include <linux/dma-mapping.h>
-+#include <linux/fsl_bman.h>
-+#ifdef CONFIG_SOC_BUS
-+#include <linux/sys_soc.h> /* soc_device_match */
-+#endif
-+
-+#include "fsl_fman.h"
-+#include "fm_ext.h"
-+#include "fm_port_ext.h"
-+
-+#include "mac.h"
-+#include "dpaa_eth.h"
-+#include "dpaa_eth_common.h"
-+#ifdef CONFIG_FSL_DPAA_DBG_LOOP
-+#include "dpaa_debugfs.h"
-+#endif /* CONFIG_FSL_DPAA_DBG_LOOP */
-+
-+/* CREATE_TRACE_POINTS only needs to be defined once. Other dpa files
-+ * using trace events only need to #include <trace/events/sched.h>
-+ */
-+#define CREATE_TRACE_POINTS
-+#include "dpaa_eth_trace.h"
-+
-+#define DPA_NAPI_WEIGHT 64
-+
-+/* Valid checksum indication */
-+#define DPA_CSUM_VALID 0xFFFF
-+
-+#define DPA_DESCRIPTION "FSL DPAA Ethernet driver"
-+
-+MODULE_LICENSE("Dual BSD/GPL");
-+
-+MODULE_AUTHOR("Andy Fleming <afleming@freescale.com>");
-+
-+MODULE_DESCRIPTION(DPA_DESCRIPTION);
-+
-+static uint8_t debug = -1;
-+module_param(debug, byte, S_IRUGO);
-+MODULE_PARM_DESC(debug, "Module/Driver verbosity level");
-+
-+/* This has to work in tandem with the DPA_CS_THRESHOLD_xxx values. */
-+static uint16_t tx_timeout = 1000;
-+module_param(tx_timeout, ushort, S_IRUGO);
-+MODULE_PARM_DESC(tx_timeout, "The Tx timeout in ms");
-+
-+static const char rtx[][3] = {
-+ [RX] = "RX",
-+ [TX] = "TX"
-+};
-+
-+#ifndef CONFIG_PPC
-+bool dpaa_errata_a010022;
-+EXPORT_SYMBOL(dpaa_errata_a010022);
-+#endif
-+
-+/* BM */
-+
-+#define DPAA_ETH_MAX_PAD (L1_CACHE_BYTES * 8)
-+
-+static uint8_t dpa_priv_common_bpid;
-+
-+#ifdef CONFIG_FSL_DPAA_DBG_LOOP
-+struct net_device *dpa_loop_netdevs[20];
-+#endif
-+
-+#ifdef CONFIG_PM
-+
-+static int dpaa_suspend(struct device *dev)
-+{
-+ struct net_device *net_dev;
-+ struct dpa_priv_s *priv;
-+ struct mac_device *mac_dev;
-+ int err = 0;
-+
-+ net_dev = dev_get_drvdata(dev);
-+
-+ if (net_dev->flags & IFF_UP) {
-+ priv = netdev_priv(net_dev);
-+ mac_dev = priv->mac_dev;
-+
-+ if (priv->wol & DPAA_WOL_MAGIC) {
-+ err = priv->mac_dev->set_wol(mac_dev->port_dev[RX],
-+ priv->mac_dev->get_mac_handle(mac_dev), true);
-+ if (err) {
-+ netdev_err(net_dev, "set_wol() = %d\n", err);
-+ goto set_wol_failed;
-+ }
-+ }
-+
-+ err = fm_port_suspend(mac_dev->port_dev[RX]);
-+ if (err) {
-+ netdev_err(net_dev, "fm_port_suspend(RX) = %d\n", err);
-+ goto rx_port_suspend_failed;
-+ }
-+
-+ err = fm_port_suspend(mac_dev->port_dev[TX]);
-+ if (err) {
-+ netdev_err(net_dev, "fm_port_suspend(TX) = %d\n", err);
-+ goto tx_port_suspend_failed;
-+ }
-+ }
-+
-+ return 0;
-+
-+tx_port_suspend_failed:
-+ fm_port_resume(mac_dev->port_dev[RX]);
-+rx_port_suspend_failed:
-+ if (priv->wol & DPAA_WOL_MAGIC) {
-+ priv->mac_dev->set_wol(mac_dev->port_dev[RX],
-+ priv->mac_dev->get_mac_handle(mac_dev), false);
-+ }
-+set_wol_failed:
-+ return err;
-+}
-+
-+static int dpaa_resume(struct device *dev)
-+{
-+ struct net_device *net_dev;
-+ struct dpa_priv_s *priv;
-+ struct mac_device *mac_dev;
-+ int err = 0;
-+
-+ net_dev = dev_get_drvdata(dev);
-+
-+ if (net_dev->flags & IFF_UP) {
-+ priv = netdev_priv(net_dev);
-+ mac_dev = priv->mac_dev;
-+
-+ err = fm_mac_resume(mac_dev->get_mac_handle(mac_dev));
-+ if (err) {
-+ netdev_err(net_dev, "fm_mac_resume = %d\n", err);
-+ goto resume_failed;
-+ }
-+
-+ err = fm_port_resume(mac_dev->port_dev[TX]);
-+ if (err) {
-+ netdev_err(net_dev, "fm_port_resume(TX) = %d\n", err);
-+ goto resume_failed;
-+ }
-+
-+ err = fm_port_resume(mac_dev->port_dev[RX]);
-+ if (err) {
-+ netdev_err(net_dev, "fm_port_resume(RX) = %d\n", err);
-+ goto resume_failed;
-+ }
-+
-+ if (priv->wol & DPAA_WOL_MAGIC) {
-+ err = priv->mac_dev->set_wol(mac_dev->port_dev[RX],
-+ priv->mac_dev->get_mac_handle(mac_dev), false);
-+ if (err) {
-+ netdev_err(net_dev, "set_wol() = %d\n", err);
-+ goto resume_failed;
-+ }
-+ }
-+ }
-+
-+ return 0;
-+
-+resume_failed:
-+ return err;
-+}
-+
-+static const struct dev_pm_ops dpaa_pm_ops = {
-+ .suspend = dpaa_suspend,
-+ .resume = dpaa_resume,
-+};
-+
-+#define DPAA_PM_OPS (&dpaa_pm_ops)
-+
-+#else /* CONFIG_PM */
-+
-+#define DPAA_PM_OPS NULL
-+
-+#endif /* CONFIG_PM */
-+
-+/* Checks whether the checksum field in Parse Results array is valid
-+ * (equals 0xFFFF) and increments the .cse counter otherwise
-+ */
-+static inline void
-+dpa_csum_validation(const struct dpa_priv_s *priv,
-+ struct dpa_percpu_priv_s *percpu_priv,
-+ const struct qm_fd *fd)
-+{
-+ dma_addr_t addr = qm_fd_addr(fd);
-+ struct dpa_bp *dpa_bp = priv->dpa_bp;
-+ void *frm = phys_to_virt(addr);
-+ fm_prs_result_t *parse_result;
-+
-+ if (unlikely(!frm))
-+ return;
-+
-+ dma_sync_single_for_cpu(dpa_bp->dev, addr, DPA_RX_PRIV_DATA_SIZE +
-+ DPA_PARSE_RESULTS_SIZE, DMA_BIDIRECTIONAL);
-+
-+ parse_result = (fm_prs_result_t *)(frm + DPA_RX_PRIV_DATA_SIZE);
-+
-+ if (parse_result->cksum != DPA_CSUM_VALID)
-+ percpu_priv->rx_errors.cse++;
-+}
-+
-+static void _dpa_rx_error(struct net_device *net_dev,
-+ const struct dpa_priv_s *priv,
-+ struct dpa_percpu_priv_s *percpu_priv,
-+ const struct qm_fd *fd,
-+ u32 fqid)
-+{
-+ /* limit common, possibly innocuous Rx FIFO Overflow errors'
-+ * interference with zero-loss convergence benchmark results.
-+ */
-+ if (likely(fd->status & FM_FD_STAT_ERR_PHYSICAL))
-+ pr_warn_once("fsl-dpa: non-zero error counters in fman statistics (sysfs)\n");
-+ else
-+ if (netif_msg_hw(priv) && net_ratelimit())
-+ netdev_dbg(net_dev, "Err FD status = 0x%08x\n",
-+ fd->status & FM_FD_STAT_RX_ERRORS);
-+#ifdef CONFIG_FSL_DPAA_HOOKS
-+ if (dpaa_eth_hooks.rx_error &&
-+ dpaa_eth_hooks.rx_error(net_dev, fd, fqid) == DPAA_ETH_STOLEN)
-+ /* it's up to the hook to perform resource cleanup */
-+ return;
-+#endif
-+ percpu_priv->stats.rx_errors++;
-+
-+ if (fd->status & FM_PORT_FRM_ERR_DMA)
-+ percpu_priv->rx_errors.dme++;
-+ if (fd->status & FM_PORT_FRM_ERR_PHYSICAL)
-+ percpu_priv->rx_errors.fpe++;
-+ if (fd->status & FM_PORT_FRM_ERR_SIZE)
-+ percpu_priv->rx_errors.fse++;
-+ if (fd->status & FM_PORT_FRM_ERR_PRS_HDR_ERR)
-+ percpu_priv->rx_errors.phe++;
-+ if (fd->status & FM_FD_STAT_L4CV)
-+ dpa_csum_validation(priv, percpu_priv, fd);
-+
-+ dpa_fd_release(net_dev, fd);
-+}
-+
-+static void _dpa_tx_error(struct net_device *net_dev,
-+ const struct dpa_priv_s *priv,
-+ struct dpa_percpu_priv_s *percpu_priv,
-+ const struct qm_fd *fd,
-+ u32 fqid)
-+{
-+ struct sk_buff *skb;
-+
-+ if (netif_msg_hw(priv) && net_ratelimit())
-+ netdev_warn(net_dev, "FD status = 0x%08x\n",
-+ fd->status & FM_FD_STAT_TX_ERRORS);
-+#ifdef CONFIG_FSL_DPAA_HOOKS
-+ if (dpaa_eth_hooks.tx_error &&
-+ dpaa_eth_hooks.tx_error(net_dev, fd, fqid) == DPAA_ETH_STOLEN)
-+ /* now the hook must ensure proper cleanup */
-+ return;
-+#endif
-+ percpu_priv->stats.tx_errors++;
-+
-+ /* If we intended the buffers from this frame to go into the bpools
-+ * when the FMan transmit was done, we need to put it in manually.
-+ */
-+ if (fd->bpid != 0xff) {
-+ dpa_fd_release(net_dev, fd);
-+ return;
-+ }
-+
-+ skb = _dpa_cleanup_tx_fd(priv, fd);
-+ dev_kfree_skb(skb);
-+}
-+
-+/* Helper function to factor out frame validation logic on all Rx paths. Its
-+ * purpose is to extract from the Parse Results structure information about
-+ * the integrity of the frame, its checksum, the length of the parsed headers
-+ * and whether the frame is suitable for GRO.
-+ *
-+ * Assumes no parser errors, since any error frame is dropped before this
-+ * function is called.
-+ *
-+ * @skb will have its ip_summed field overwritten;
-+ * @use_gro will only be written with 0, if the frame is definitely not
-+ * GRO-able; otherwise, it will be left unchanged;
-+ * @hdr_size will be written with a safe value, at least the size of the
-+ * headers' length.
-+ */
-+void __hot _dpa_process_parse_results(const fm_prs_result_t *parse_results,
-+ const struct qm_fd *fd,
-+ struct sk_buff *skb, int *use_gro)
-+{
-+ if (fd->status & FM_FD_STAT_L4CV) {
-+ /* The parser has run and performed L4 checksum validation.
-+ * We know there were no parser errors (and implicitly no
-+ * L4 csum error), otherwise we wouldn't be here.
-+ */
-+ skb->ip_summed = CHECKSUM_UNNECESSARY;
-+
-+ /* Don't go through GRO for certain types of traffic that
-+ * we know are not GRO-able, such as dgram-based protocols.
-+ * In the worst-case scenarios, such as small-pkt terminating
-+ * UDP, the extra GRO processing would be overkill.
-+ *
-+ * The only protocol the Parser supports that is also GRO-able
-+ * is currently TCP.
-+ */
-+ if (!fm_l4_frame_is_tcp(parse_results))
-+ *use_gro = 0;
-+
-+ return;
-+ }
-+
-+ /* We're here because either the parser didn't run or the L4 checksum
-+ * was not verified. This may include the case of a UDP frame with
-+ * checksum zero or an L4 proto other than TCP/UDP
-+ */
-+ skb->ip_summed = CHECKSUM_NONE;
-+
-+ /* Bypass GRO for unknown traffic or if no PCDs are applied */
-+ *use_gro = 0;
-+}
-+
-+int dpaa_eth_poll(struct napi_struct *napi, int budget)
-+{
-+ struct dpa_napi_portal *np =
-+ container_of(napi, struct dpa_napi_portal, napi);
-+
-+ int cleaned = qman_p_poll_dqrr(np->p, budget);
-+
-+ if (cleaned < budget) {
-+ int tmp;
-+ napi_complete(napi);
-+ tmp = qman_p_irqsource_add(np->p, QM_PIRQ_DQRI);
-+ DPA_BUG_ON(tmp);
-+ }
-+
-+ return cleaned;
-+}
-+EXPORT_SYMBOL(dpaa_eth_poll);
-+
-+static void __hot _dpa_tx_conf(struct net_device *net_dev,
-+ const struct dpa_priv_s *priv,
-+ struct dpa_percpu_priv_s *percpu_priv,
-+ const struct qm_fd *fd,
-+ u32 fqid)
-+{
-+ struct sk_buff *skb;
-+
-+ /* do we need the timestamp for the error frames? */
-+
-+ if (unlikely(fd->status & FM_FD_STAT_TX_ERRORS) != 0) {
-+ if (netif_msg_hw(priv) && net_ratelimit())
-+ netdev_warn(net_dev, "FD status = 0x%08x\n",
-+ fd->status & FM_FD_STAT_TX_ERRORS);
-+
-+ percpu_priv->stats.tx_errors++;
-+ }
-+
-+ /* hopefully we need not get the timestamp before the hook */
-+#ifdef CONFIG_FSL_DPAA_HOOKS
-+ if (dpaa_eth_hooks.tx_confirm && dpaa_eth_hooks.tx_confirm(net_dev,
-+ fd, fqid) == DPAA_ETH_STOLEN)
-+ /* it's the hook that must now perform cleanup */
-+ return;
-+#endif
-+ /* This might not perfectly reflect the reality, if the core dequeuing
-+ * the Tx confirmation is different from the one that did the enqueue,
-+ * but at least it'll show up in the total count.
-+ */
-+ percpu_priv->tx_confirm++;
-+
-+ skb = _dpa_cleanup_tx_fd(priv, fd);
-+
-+ dev_kfree_skb(skb);
-+}
-+
-+enum qman_cb_dqrr_result
-+priv_rx_error_dqrr(struct qman_portal *portal,
-+ struct qman_fq *fq,
-+ const struct qm_dqrr_entry *dq)
-+{
-+ struct net_device *net_dev;
-+ struct dpa_priv_s *priv;
-+ struct dpa_percpu_priv_s *percpu_priv;
-+ int *count_ptr;
-+
-+ net_dev = ((struct dpa_fq *)fq)->net_dev;
-+ priv = netdev_priv(net_dev);
-+
-+ percpu_priv = raw_cpu_ptr(priv->percpu_priv);
-+ count_ptr = raw_cpu_ptr(priv->dpa_bp->percpu_count);
-+
-+ if (dpaa_eth_napi_schedule(percpu_priv, portal))
-+ return qman_cb_dqrr_stop;
-+
-+ if (unlikely(dpaa_eth_refill_bpools(priv->dpa_bp, count_ptr)))
-+ /* Unable to refill the buffer pool due to insufficient
-+ * system memory. Just release the frame back into the pool,
-+ * otherwise we'll soon end up with an empty buffer pool.
-+ */
-+ dpa_fd_release(net_dev, &dq->fd);
-+ else
-+ _dpa_rx_error(net_dev, priv, percpu_priv, &dq->fd, fq->fqid);
-+
-+ return qman_cb_dqrr_consume;
-+}
-+
-+
-+enum qman_cb_dqrr_result __hot
-+priv_rx_default_dqrr(struct qman_portal *portal,
-+ struct qman_fq *fq,
-+ const struct qm_dqrr_entry *dq)
-+{
-+ struct net_device *net_dev;
-+ struct dpa_priv_s *priv;
-+ struct dpa_percpu_priv_s *percpu_priv;
-+ int *count_ptr;
-+ struct dpa_bp *dpa_bp;
-+
-+ net_dev = ((struct dpa_fq *)fq)->net_dev;
-+ priv = netdev_priv(net_dev);
-+ dpa_bp = priv->dpa_bp;
-+
-+ /* Trace the Rx fd */
-+ trace_dpa_rx_fd(net_dev, fq, &dq->fd);
-+
-+ /* IRQ handler, non-migratable; safe to use raw_cpu_ptr here */
-+ percpu_priv = raw_cpu_ptr(priv->percpu_priv);
-+ count_ptr = raw_cpu_ptr(dpa_bp->percpu_count);
-+
-+ if (unlikely(dpaa_eth_napi_schedule(percpu_priv, portal)))
-+ return qman_cb_dqrr_stop;
-+
-+ /* Vale of plenty: make sure we didn't run out of buffers */
-+
-+ if (unlikely(dpaa_eth_refill_bpools(dpa_bp, count_ptr)))
-+ /* Unable to refill the buffer pool due to insufficient
-+ * system memory. Just release the frame back into the pool,
-+ * otherwise we'll soon end up with an empty buffer pool.
-+ */
-+ dpa_fd_release(net_dev, &dq->fd);
-+ else
-+ _dpa_rx(net_dev, portal, priv, percpu_priv, &dq->fd, fq->fqid,
-+ count_ptr);
-+
-+ return qman_cb_dqrr_consume;
-+}
-+
-+enum qman_cb_dqrr_result
-+priv_tx_conf_error_dqrr(struct qman_portal *portal,
-+ struct qman_fq *fq,
-+ const struct qm_dqrr_entry *dq)
-+{
-+ struct net_device *net_dev;
-+ struct dpa_priv_s *priv;
-+ struct dpa_percpu_priv_s *percpu_priv;
-+
-+ net_dev = ((struct dpa_fq *)fq)->net_dev;
-+ priv = netdev_priv(net_dev);
-+
-+ percpu_priv = raw_cpu_ptr(priv->percpu_priv);
-+
-+ if (dpaa_eth_napi_schedule(percpu_priv, portal))
-+ return qman_cb_dqrr_stop;
-+
-+ _dpa_tx_error(net_dev, priv, percpu_priv, &dq->fd, fq->fqid);
-+
-+ return qman_cb_dqrr_consume;
-+}
-+
-+enum qman_cb_dqrr_result __hot
-+priv_tx_conf_default_dqrr(struct qman_portal *portal,
-+ struct qman_fq *fq,
-+ const struct qm_dqrr_entry *dq)
-+{
-+ struct net_device *net_dev;
-+ struct dpa_priv_s *priv;
-+ struct dpa_percpu_priv_s *percpu_priv;
-+
-+ net_dev = ((struct dpa_fq *)fq)->net_dev;
-+ priv = netdev_priv(net_dev);
-+
-+ /* Trace the fd */
-+ trace_dpa_tx_conf_fd(net_dev, fq, &dq->fd);
-+
-+ /* Non-migratable context, safe to use raw_cpu_ptr */
-+ percpu_priv = raw_cpu_ptr(priv->percpu_priv);
-+
-+ if (dpaa_eth_napi_schedule(percpu_priv, portal))
-+ return qman_cb_dqrr_stop;
-+
-+ _dpa_tx_conf(net_dev, priv, percpu_priv, &dq->fd, fq->fqid);
-+
-+ return qman_cb_dqrr_consume;
-+}
-+
-+void priv_ern(struct qman_portal *portal,
-+ struct qman_fq *fq,
-+ const struct qm_mr_entry *msg)
-+{
-+ struct net_device *net_dev;
-+ const struct dpa_priv_s *priv;
-+ struct sk_buff *skb;
-+ struct dpa_percpu_priv_s *percpu_priv;
-+ struct qm_fd fd = msg->ern.fd;
-+
-+ net_dev = ((struct dpa_fq *)fq)->net_dev;
-+ priv = netdev_priv(net_dev);
-+ /* Non-migratable context, safe to use raw_cpu_ptr */
-+ percpu_priv = raw_cpu_ptr(priv->percpu_priv);
-+
-+ percpu_priv->stats.tx_dropped++;
-+ percpu_priv->stats.tx_fifo_errors++;
-+ count_ern(percpu_priv, msg);
-+
-+ /* If we intended this buffer to go into the pool
-+ * when the FM was done, we need to put it in
-+ * manually.
-+ */
-+ if (msg->ern.fd.bpid != 0xff) {
-+ dpa_fd_release(net_dev, &fd);
-+ return;
-+ }
-+
-+ skb = _dpa_cleanup_tx_fd(priv, &fd);
-+ dev_kfree_skb_any(skb);
-+}
-+
-+const struct dpa_fq_cbs_t private_fq_cbs = {
-+ .rx_defq = { .cb = { .dqrr = priv_rx_default_dqrr } },
-+ .tx_defq = { .cb = { .dqrr = priv_tx_conf_default_dqrr } },
-+ .rx_errq = { .cb = { .dqrr = priv_rx_error_dqrr } },
-+ .tx_errq = { .cb = { .dqrr = priv_tx_conf_error_dqrr } },
-+ .egress_ern = { .cb = { .ern = priv_ern } }
-+};
-+EXPORT_SYMBOL(private_fq_cbs);
-+
-+static void dpaa_eth_napi_enable(struct dpa_priv_s *priv)
-+{
-+ struct dpa_percpu_priv_s *percpu_priv;
-+ int i, j;
-+
-+ for_each_possible_cpu(i) {
-+ percpu_priv = per_cpu_ptr(priv->percpu_priv, i);
-+
-+ for (j = 0; j < qman_portal_max; j++)
-+ napi_enable(&percpu_priv->np[j].napi);
-+ }
-+}
-+
-+static void dpaa_eth_napi_disable(struct dpa_priv_s *priv)
-+{
-+ struct dpa_percpu_priv_s *percpu_priv;
-+ int i, j;
-+
-+ for_each_possible_cpu(i) {
-+ percpu_priv = per_cpu_ptr(priv->percpu_priv, i);
-+
-+ for (j = 0; j < qman_portal_max; j++)
-+ napi_disable(&percpu_priv->np[j].napi);
-+ }
-+}
-+
-+static int __cold dpa_eth_priv_start(struct net_device *net_dev)
-+{
-+ int err;
-+ struct dpa_priv_s *priv;
-+
-+ priv = netdev_priv(net_dev);
-+
-+ dpaa_eth_napi_enable(priv);
-+
-+ err = dpa_start(net_dev);
-+ if (err < 0)
-+ dpaa_eth_napi_disable(priv);
-+
-+ return err;
-+}
-+
-+
-+
-+static int __cold dpa_eth_priv_stop(struct net_device *net_dev)
-+{
-+ int _errno;
-+ struct dpa_priv_s *priv;
-+
-+ _errno = dpa_stop(net_dev);
-+ /* Allow NAPI to consume any frame still in the Rx/TxConfirm
-+ * ingress queues. This is to avoid a race between the current
-+ * context and ksoftirqd which could leave NAPI disabled while
-+ * in fact there's still Rx traffic to be processed.
-+ */
-+ usleep_range(5000, 10000);
-+
-+ priv = netdev_priv(net_dev);
-+ dpaa_eth_napi_disable(priv);
-+
-+ return _errno;
-+}
-+
-+#ifdef CONFIG_NET_POLL_CONTROLLER
-+static void dpaa_eth_poll_controller(struct net_device *net_dev)
-+{
-+ struct dpa_priv_s *priv = netdev_priv(net_dev);
-+ struct dpa_percpu_priv_s *percpu_priv =
-+ raw_cpu_ptr(priv->percpu_priv);
-+ struct qman_portal *p;
-+ const struct qman_portal_config *pc;
-+ struct dpa_napi_portal *np;
-+
-+ p = (struct qman_portal *)qman_get_affine_portal(smp_processor_id());
-+ pc = qman_p_get_portal_config(p);
-+ np = &percpu_priv->np[pc->index];
-+
-+ qman_p_irqsource_remove(np->p, QM_PIRQ_DQRI);
-+ qman_p_poll_dqrr(np->p, np->napi.weight);
-+ qman_p_irqsource_add(np->p, QM_PIRQ_DQRI);
-+}
-+#endif
-+
-+static const struct net_device_ops dpa_private_ops = {
-+ .ndo_open = dpa_eth_priv_start,
-+ .ndo_start_xmit = dpa_tx,
-+ .ndo_stop = dpa_eth_priv_stop,
-+ .ndo_tx_timeout = dpa_timeout,
-+ .ndo_get_stats64 = dpa_get_stats64,
-+ .ndo_set_mac_address = dpa_set_mac_address,
-+ .ndo_validate_addr = eth_validate_addr,
-+#ifdef CONFIG_FSL_DPAA_ETH_USE_NDO_SELECT_QUEUE
-+ .ndo_select_queue = dpa_select_queue,
-+#endif
-+ .ndo_change_mtu = dpa_change_mtu,
-+ .ndo_set_rx_mode = dpa_set_rx_mode,
-+ .ndo_init = dpa_ndo_init,
-+ .ndo_set_features = dpa_set_features,
-+ .ndo_fix_features = dpa_fix_features,
-+ .ndo_do_ioctl = dpa_ioctl,
-+#ifdef CONFIG_NET_POLL_CONTROLLER
-+ .ndo_poll_controller = dpaa_eth_poll_controller,
-+#endif
-+};
-+
-+static int dpa_private_napi_add(struct net_device *net_dev)
-+{
-+ struct dpa_priv_s *priv = netdev_priv(net_dev);
-+ struct dpa_percpu_priv_s *percpu_priv;
-+ int i, cpu;
-+
-+ for_each_possible_cpu(cpu) {
-+ percpu_priv = per_cpu_ptr(priv->percpu_priv, cpu);
-+
-+ percpu_priv->np = devm_kzalloc(net_dev->dev.parent,
-+ qman_portal_max * sizeof(struct dpa_napi_portal),
-+ GFP_KERNEL);
-+
-+ if (unlikely(percpu_priv->np == NULL)) {
-+ dev_err(net_dev->dev.parent, "devm_kzalloc() failed\n");
-+ return -ENOMEM;
-+ }
-+
-+ for (i = 0; i < qman_portal_max; i++)
-+ netif_napi_add(net_dev, &percpu_priv->np[i].napi,
-+ dpaa_eth_poll, DPA_NAPI_WEIGHT);
-+ }
-+
-+ return 0;
-+}
-+
-+void dpa_private_napi_del(struct net_device *net_dev)
-+{
-+ struct dpa_priv_s *priv = netdev_priv(net_dev);
-+ struct dpa_percpu_priv_s *percpu_priv;
-+ int i, cpu;
-+
-+ for_each_possible_cpu(cpu) {
-+ percpu_priv = per_cpu_ptr(priv->percpu_priv, cpu);
-+
-+ if (percpu_priv->np) {
-+ for (i = 0; i < qman_portal_max; i++)
-+ netif_napi_del(&percpu_priv->np[i].napi);
-+
-+ devm_kfree(net_dev->dev.parent, percpu_priv->np);
-+ }
-+ }
-+}
-+EXPORT_SYMBOL(dpa_private_napi_del);
-+
-+static int dpa_private_netdev_init(struct net_device *net_dev)
-+{
-+ int i;
-+ struct dpa_priv_s *priv = netdev_priv(net_dev);
-+ struct dpa_percpu_priv_s *percpu_priv;
-+ const uint8_t *mac_addr;
-+
-+ /* Although we access another CPU's private data here
-+ * we do it at initialization so it is safe
-+ */
-+ for_each_possible_cpu(i) {
-+ percpu_priv = per_cpu_ptr(priv->percpu_priv, i);
-+ percpu_priv->net_dev = net_dev;
-+ }
-+
-+ net_dev->netdev_ops = &dpa_private_ops;
-+ mac_addr = priv->mac_dev->addr;
-+
-+ net_dev->mem_start = priv->mac_dev->res->start;
-+ net_dev->mem_end = priv->mac_dev->res->end;
-+
-+ net_dev->hw_features |= (NETIF_F_IP_CSUM | NETIF_F_IPV6_CSUM |
-+ NETIF_F_LLTX);
-+
-+ /* Advertise S/G and HIGHDMA support for private interfaces */
-+ net_dev->hw_features |= NETIF_F_SG | NETIF_F_HIGHDMA;
-+ /* Recent kernels enable GSO automatically, if
-+ * we declare NETIF_F_SG. For conformity, we'll
-+ * still declare GSO explicitly.
-+ */
-+ net_dev->features |= NETIF_F_GSO;
-+
-+ /* Advertise GRO support */
-+ net_dev->features |= NETIF_F_GRO;
-+
-+ /* Advertise NETIF_F_HW_ACCEL_MQ to avoid Tx timeout warnings */
-+ net_dev->features |= NETIF_F_HW_ACCEL_MQ;
-+
-+#ifndef CONFIG_PPC
-+ /* Due to the A010022 FMan errata, we can not use contig frames larger
-+ * than 4K, nor S/G frames. We need to stop advertising S/G and GSO
-+ * support.
-+ */
-+ if (unlikely(dpaa_errata_a010022)) {
-+ net_dev->hw_features &= ~NETIF_F_SG;
-+ net_dev->features &= ~NETIF_F_GSO;
-+ }
-+#endif
-+
-+ return dpa_netdev_init(net_dev, mac_addr, tx_timeout);
-+}
-+
-+static struct dpa_bp * __cold
-+dpa_priv_bp_probe(struct device *dev)
-+{
-+ struct dpa_bp *dpa_bp;
-+
-+ dpa_bp = devm_kzalloc(dev, sizeof(*dpa_bp), GFP_KERNEL);
-+ if (unlikely(dpa_bp == NULL)) {
-+ dev_err(dev, "devm_kzalloc() failed\n");
-+ return ERR_PTR(-ENOMEM);
-+ }
-+
-+ dpa_bp->percpu_count = devm_alloc_percpu(dev, *dpa_bp->percpu_count);
-+ dpa_bp->target_count = CONFIG_FSL_DPAA_ETH_MAX_BUF_COUNT;
-+
-+ dpa_bp->seed_cb = dpa_bp_priv_seed;
-+ dpa_bp->free_buf_cb = _dpa_bp_free_pf;
-+
-+ return dpa_bp;
-+}
-+
-+/* Place all ingress FQs (Rx Default, Rx Error, PCD FQs) in a dedicated CGR.
-+ * We won't be sending congestion notifications to FMan; for now, we just use
-+ * this CGR to generate enqueue rejections to FMan in order to drop the frames
-+ * before they reach our ingress queues and eat up memory.
-+ */
-+static int dpaa_eth_priv_ingress_cgr_init(struct dpa_priv_s *priv)
-+{
-+ struct qm_mcc_initcgr initcgr;
-+ u32 cs_th;
-+ int err;
-+
-+ err = qman_alloc_cgrid(&priv->ingress_cgr.cgrid);
-+ if (err < 0) {
-+ pr_err("Error %d allocating CGR ID\n", err);
-+ goto out_error;
-+ }
-+
-+ /* Enable CS TD, but disable Congestion State Change Notifications. */
-+ initcgr.we_mask = QM_CGR_WE_CS_THRES;
-+ initcgr.cgr.cscn_en = QM_CGR_EN;
-+ cs_th = CONFIG_FSL_DPAA_INGRESS_CS_THRESHOLD;
-+ qm_cgr_cs_thres_set64(&initcgr.cgr.cs_thres, cs_th, 1);
-+
-+ initcgr.we_mask |= QM_CGR_WE_CSTD_EN;
-+ initcgr.cgr.cstd_en = QM_CGR_EN;
-+
-+ /* This is actually a hack, because this CGR will be associated with
-+ * our affine SWP. However, we'll place our ingress FQs in it.
-+ */
-+ err = qman_create_cgr(&priv->ingress_cgr, QMAN_CGR_FLAG_USE_INIT,
-+ &initcgr);
-+ if (err < 0) {
-+ pr_err("Error %d creating ingress CGR with ID %d\n", err,
-+ priv->ingress_cgr.cgrid);
-+ qman_release_cgrid(priv->ingress_cgr.cgrid);
-+ goto out_error;
-+ }
-+ pr_debug("Created ingress CGR %d for netdev with hwaddr %pM\n",
-+ priv->ingress_cgr.cgrid, priv->mac_dev->addr);
-+
-+ /* struct qman_cgr allows special cgrid values (i.e. outside the 0..255
-+ * range), but we have no common initialization path between the
-+ * different variants of the DPAA Eth driver, so we do it here rather
-+ * than modifying every other variant than "private Eth".
-+ */
-+ priv->use_ingress_cgr = true;
-+
-+out_error:
-+ return err;
-+}
-+
-+static int dpa_priv_bp_create(struct net_device *net_dev, struct dpa_bp *dpa_bp,
-+ size_t count)
-+{
-+ struct dpa_priv_s *priv = netdev_priv(net_dev);
-+ int i;
-+
-+ if (netif_msg_probe(priv))
-+ dev_dbg(net_dev->dev.parent,
-+ "Using private BM buffer pools\n");
-+
-+ priv->bp_count = count;
-+
-+ for (i = 0; i < count; i++) {
-+ int err;
-+ err = dpa_bp_alloc(&dpa_bp[i], net_dev->dev.parent);
-+ if (err < 0) {
-+ dpa_bp_free(priv);
-+ priv->dpa_bp = NULL;
-+ return err;
-+ }
-+
-+ priv->dpa_bp = &dpa_bp[i];
-+ }
-+
-+ dpa_priv_common_bpid = priv->dpa_bp->bpid;
-+ return 0;
-+}
-+
-+static const struct of_device_id dpa_match[];
-+
-+#ifdef CONFIG_FSL_DPAA_DBG_LOOP
-+static int dpa_new_loop_id(void)
-+{
-+ static int if_id;
-+
-+ return if_id++;
-+}
-+#endif
-+
-+static int
-+dpaa_eth_priv_probe(struct platform_device *_of_dev)
-+{
-+ int err = 0, i, channel;
-+ struct device *dev;
-+ struct device_node *dpa_node;
-+ struct dpa_bp *dpa_bp;
-+ size_t count = 1;
-+ struct net_device *net_dev = NULL;
-+ struct dpa_priv_s *priv = NULL;
-+ struct dpa_percpu_priv_s *percpu_priv;
-+ struct fm_port_fqs port_fqs;
-+ struct dpa_buffer_layout_s *buf_layout = NULL;
-+ struct mac_device *mac_dev;
-+
-+ dev = &_of_dev->dev;
-+
-+ dpa_node = dev->of_node;
-+
-+ if (!of_device_is_available(dpa_node))
-+ return -ENODEV;
-+
-+ /* Get the buffer pools assigned to this interface;
-+ * run only once the default pool probing code
-+ */
-+ dpa_bp = (dpa_bpid2pool(dpa_priv_common_bpid)) ? :
-+ dpa_priv_bp_probe(dev);
-+ if (IS_ERR(dpa_bp))
-+ return PTR_ERR(dpa_bp);
-+
-+ /* Allocate this early, so we can store relevant information in
-+ * the private area (needed by 1588 code in dpa_mac_probe)
-+ */
-+ net_dev = alloc_etherdev_mq(sizeof(*priv), DPAA_ETH_TX_QUEUES);
-+ if (!net_dev) {
-+ dev_err(dev, "alloc_etherdev_mq() failed\n");
-+ goto alloc_etherdev_mq_failed;
-+ }
-+
-+ /* Do this here, so we can be verbose early */
-+ SET_NETDEV_DEV(net_dev, dev);
-+ dev_set_drvdata(dev, net_dev);
-+
-+ priv = netdev_priv(net_dev);
-+ priv->net_dev = net_dev;
-+ strcpy(priv->if_type, "private");
-+
-+ priv->msg_enable = netif_msg_init(debug, -1);
-+
-+#ifdef CONFIG_FSL_DPAA_DBG_LOOP
-+ priv->loop_id = dpa_new_loop_id();
-+ priv->loop_to = -1; /* disabled by default */
-+ dpa_loop_netdevs[priv->loop_id] = net_dev;
-+#endif
-+
-+ mac_dev = dpa_mac_probe(_of_dev);
-+ if (IS_ERR(mac_dev) || !mac_dev) {
-+ err = PTR_ERR(mac_dev);
-+ goto mac_probe_failed;
-+ }
-+
-+ /* We have physical ports, so we need to establish
-+ * the buffer layout.
-+ */
-+ buf_layout = devm_kzalloc(dev, 2 * sizeof(*buf_layout),
-+ GFP_KERNEL);
-+ if (!buf_layout) {
-+ dev_err(dev, "devm_kzalloc() failed\n");
-+ goto alloc_failed;
-+ }
-+ dpa_set_buffers_layout(mac_dev, buf_layout);
-+
-+ /* For private ports, need to compute the size of the default
-+ * buffer pool, based on FMan port buffer layout;also update
-+ * the maximum buffer size for private ports if necessary
-+ */
-+ dpa_bp->size = dpa_bp_size(&buf_layout[RX]);
-+
-+#ifdef CONFIG_FSL_DPAA_ETH_JUMBO_FRAME
-+ /* We only want to use jumbo frame optimization if we actually have
-+ * L2 MAX FRM set for jumbo frames as well.
-+ */
-+#ifndef CONFIG_PPC
-+ if (likely(!dpaa_errata_a010022))
-+#endif
-+ if(fm_get_max_frm() < 9600)
-+ dev_warn(dev,
-+ "Invalid configuration: if jumbo frames support is on, FSL_FM_MAX_FRAME_SIZE should be set to 9600\n");
-+#endif
-+
-+ INIT_LIST_HEAD(&priv->dpa_fq_list);
-+
-+ memset(&port_fqs, 0, sizeof(port_fqs));
-+
-+ err = dpa_fq_probe_mac(dev, &priv->dpa_fq_list, &port_fqs, true, RX);
-+ if (!err)
-+ err = dpa_fq_probe_mac(dev, &priv->dpa_fq_list,
-+ &port_fqs, true, TX);
-+
-+ if (err < 0)
-+ goto fq_probe_failed;
-+
-+ /* bp init */
-+
-+ err = dpa_priv_bp_create(net_dev, dpa_bp, count);
-+
-+ if (err < 0)
-+ goto bp_create_failed;
-+
-+ priv->mac_dev = mac_dev;
-+
-+ channel = dpa_get_channel();
-+
-+ if (channel < 0) {
-+ err = channel;
-+ goto get_channel_failed;
-+ }
-+
-+ priv->channel = (uint16_t)channel;
-+ dpaa_eth_add_channel(priv->channel);
-+
-+ dpa_fq_setup(priv, &private_fq_cbs, priv->mac_dev->port_dev[TX]);
-+
-+ /* Create a congestion group for this netdev, with
-+ * dynamically-allocated CGR ID.
-+ * Must be executed after probing the MAC, but before
-+ * assigning the egress FQs to the CGRs.
-+ */
-+ err = dpaa_eth_cgr_init(priv);
-+ if (err < 0) {
-+ dev_err(dev, "Error initializing CGR\n");
-+ goto tx_cgr_init_failed;
-+ }
-+ err = dpaa_eth_priv_ingress_cgr_init(priv);
-+ if (err < 0) {
-+ dev_err(dev, "Error initializing ingress CGR\n");
-+ goto rx_cgr_init_failed;
-+ }
-+
-+ /* Add the FQs to the interface, and make them active */
-+ err = dpa_fqs_init(dev, &priv->dpa_fq_list, false);
-+ if (err < 0)
-+ goto fq_alloc_failed;
-+
-+ priv->buf_layout = buf_layout;
-+ priv->tx_headroom = dpa_get_headroom(&priv->buf_layout[TX]);
-+ priv->rx_headroom = dpa_get_headroom(&priv->buf_layout[RX]);
-+
-+ /* All real interfaces need their ports initialized */
-+ dpaa_eth_init_ports(mac_dev, dpa_bp, count, &port_fqs,
-+ buf_layout, dev);
-+
-+#ifdef CONFIG_FMAN_PFC
-+ for (i = 0; i < CONFIG_FMAN_PFC_COS_COUNT; i++) {
-+ err = fm_port_set_pfc_priorities_mapping_to_qman_wq(
-+ mac_dev->port_dev[TX], i, i);
-+ if (unlikely(err != 0)) {
-+ dev_err(dev, "Error maping PFC %u to WQ %u\n", i, i);
-+ goto pfc_mapping_failed;
-+ }
-+ }
-+#endif
-+
-+ priv->percpu_priv = devm_alloc_percpu(dev, *priv->percpu_priv);
-+
-+ if (priv->percpu_priv == NULL) {
-+ dev_err(dev, "devm_alloc_percpu() failed\n");
-+ err = -ENOMEM;
-+ goto alloc_percpu_failed;
-+ }
-+ for_each_possible_cpu(i) {
-+ percpu_priv = per_cpu_ptr(priv->percpu_priv, i);
-+ memset(percpu_priv, 0, sizeof(*percpu_priv));
-+ }
-+
-+ /* Initialize NAPI */
-+ err = dpa_private_napi_add(net_dev);
-+
-+ if (err < 0)
-+ goto napi_add_failed;
-+
-+ err = dpa_private_netdev_init(net_dev);
-+
-+ if (err < 0)
-+ goto netdev_init_failed;
-+
-+ dpaa_eth_sysfs_init(&net_dev->dev);
-+
-+#ifdef CONFIG_PM
-+ device_set_wakeup_capable(dev, true);
-+#endif
-+
-+ pr_info("fsl_dpa: Probed interface %s\n", net_dev->name);
-+
-+ return 0;
-+
-+netdev_init_failed:
-+napi_add_failed:
-+ dpa_private_napi_del(net_dev);
-+alloc_percpu_failed:
-+#ifdef CONFIG_FMAN_PFC
-+pfc_mapping_failed:
-+#endif
-+ dpa_fq_free(dev, &priv->dpa_fq_list);
-+fq_alloc_failed:
-+ qman_delete_cgr_safe(&priv->ingress_cgr);
-+ qman_release_cgrid(priv->ingress_cgr.cgrid);
-+rx_cgr_init_failed:
-+ qman_delete_cgr_safe(&priv->cgr_data.cgr);
-+ qman_release_cgrid(priv->cgr_data.cgr.cgrid);
-+tx_cgr_init_failed:
-+get_channel_failed:
-+ dpa_bp_free(priv);
-+bp_create_failed:
-+fq_probe_failed:
-+alloc_failed:
-+mac_probe_failed:
-+ dev_set_drvdata(dev, NULL);
-+ free_netdev(net_dev);
-+alloc_etherdev_mq_failed:
-+ if (atomic_read(&dpa_bp->refs) == 0)
-+ devm_kfree(dev, dpa_bp);
-+
-+ return err;
-+}
-+
-+static const struct of_device_id dpa_match[] = {
-+ {
-+ .compatible = "fsl,dpa-ethernet"
-+ },
-+ {}
-+};
-+MODULE_DEVICE_TABLE(of, dpa_match);
-+
-+static struct platform_driver dpa_driver = {
-+ .driver = {
-+ .name = KBUILD_MODNAME,
-+ .of_match_table = dpa_match,
-+ .owner = THIS_MODULE,
-+ .pm = DPAA_PM_OPS,
-+ },
-+ .probe = dpaa_eth_priv_probe,
-+ .remove = dpa_remove
-+};
-+
-+#ifndef CONFIG_PPC
-+static bool __init __cold soc_has_errata_a010022(void)
-+{
-+#ifdef CONFIG_SOC_BUS
-+ const struct soc_device_attribute soc_msi_matches[] = {
-+ { .family = "QorIQ LS1043A",
-+ .data = NULL },
-+ { },
-+ };
-+
-+ if (soc_device_match(soc_msi_matches))
-+ return true;
-+
-+ return false;
-+#else
-+ return true; /* cannot identify SoC */
-+#endif
-+}
-+#endif
-+
-+static int __init __cold dpa_load(void)
-+{
-+ int _errno;
-+
-+ pr_info(DPA_DESCRIPTION "\n");
-+
-+#ifdef CONFIG_FSL_DPAA_DBG_LOOP
-+ dpa_debugfs_module_init();
-+#endif /* CONFIG_FSL_DPAA_DBG_LOOP */
-+
-+ /* initialise dpaa_eth mirror values */
-+ dpa_rx_extra_headroom = fm_get_rx_extra_headroom();
-+ dpa_max_frm = fm_get_max_frm();
-+ dpa_num_cpus = num_possible_cpus();
-+
-+#ifndef CONFIG_PPC
-+ /* Detect if the current SoC requires the 4K alignment workaround */
-+ dpaa_errata_a010022 = soc_has_errata_a010022();
-+#endif
-+
-+#ifdef CONFIG_FSL_DPAA_DBG_LOOP
-+ memset(dpa_loop_netdevs, 0, sizeof(dpa_loop_netdevs));
-+#endif
-+
-+ _errno = platform_driver_register(&dpa_driver);
-+ if (unlikely(_errno < 0)) {
-+ pr_err(KBUILD_MODNAME
-+ ": %s:%hu:%s(): platform_driver_register() = %d\n",
-+ KBUILD_BASENAME".c", __LINE__, __func__, _errno);
-+ }
-+
-+ pr_debug(KBUILD_MODNAME ": %s:%s() ->\n",
-+ KBUILD_BASENAME".c", __func__);
-+
-+ return _errno;
-+}
-+module_init(dpa_load);
-+
-+static void __exit __cold dpa_unload(void)
-+{
-+ pr_debug(KBUILD_MODNAME ": -> %s:%s()\n",
-+ KBUILD_BASENAME".c", __func__);
-+
-+ platform_driver_unregister(&dpa_driver);
-+
-+#ifdef CONFIG_FSL_DPAA_DBG_LOOP
-+ dpa_debugfs_module_exit();
-+#endif /* CONFIG_FSL_DPAA_DBG_LOOP */
-+
-+ /* Only one channel is used and needs to be relased after all
-+ * interfaces are removed
-+ */
-+ dpa_release_channel();
-+
-+ pr_debug(KBUILD_MODNAME ": %s:%s() ->\n",
-+ KBUILD_BASENAME".c", __func__);
-+}
-+module_exit(dpa_unload);
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_dpaa/dpaa_eth.h
-@@ -0,0 +1,687 @@
-+/* Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#ifndef __DPA_H
-+#define __DPA_H
-+
-+#include <linux/netdevice.h>
-+#include <linux/fsl_qman.h> /* struct qman_fq */
-+
-+#include "fm_ext.h"
-+#include "dpaa_eth_trace.h"
-+
-+extern int dpa_rx_extra_headroom;
-+extern int dpa_max_frm;
-+extern int dpa_num_cpus;
-+
-+#define dpa_get_rx_extra_headroom() dpa_rx_extra_headroom
-+#define dpa_get_max_frm() dpa_max_frm
-+
-+#define dpa_get_max_mtu() \
-+ (dpa_get_max_frm() - (VLAN_ETH_HLEN + ETH_FCS_LEN))
-+
-+#define __hot
-+
-+/* Simple enum of FQ types - used for array indexing */
-+enum port_type {RX, TX};
-+
-+/* TODO: This structure should be renamed & moved to the FMD wrapper */
-+struct dpa_buffer_layout_s {
-+ uint16_t priv_data_size;
-+ bool parse_results;
-+ bool time_stamp;
-+ bool hash_results;
-+ uint8_t manip_extra_space;
-+ uint16_t data_align;
-+};
-+
-+#ifdef CONFIG_FSL_DPAA_ETH_DEBUG
-+#define DPA_BUG_ON(cond) BUG_ON(cond)
-+#else
-+#define DPA_BUG_ON(cond)
-+#endif
-+
-+#define DPA_TX_PRIV_DATA_SIZE 16
-+#define DPA_PARSE_RESULTS_SIZE sizeof(fm_prs_result_t)
-+#define DPA_TIME_STAMP_SIZE 8
-+#define DPA_HASH_RESULTS_SIZE 8
-+#define DPA_RX_PRIV_DATA_SIZE (DPA_TX_PRIV_DATA_SIZE + \
-+ dpa_get_rx_extra_headroom())
-+
-+#define FM_FD_STAT_RX_ERRORS \
-+ (FM_PORT_FRM_ERR_DMA | FM_PORT_FRM_ERR_PHYSICAL | \
-+ FM_PORT_FRM_ERR_SIZE | FM_PORT_FRM_ERR_CLS_DISCARD | \
-+ FM_PORT_FRM_ERR_EXTRACTION | FM_PORT_FRM_ERR_NO_SCHEME | \
-+ FM_PORT_FRM_ERR_ILL_PLCR | FM_PORT_FRM_ERR_PRS_TIMEOUT | \
-+ FM_PORT_FRM_ERR_PRS_ILL_INSTRUCT | FM_PORT_FRM_ERR_PRS_HDR_ERR)
-+
-+#define FM_FD_STAT_TX_ERRORS \
-+ (FM_PORT_FRM_ERR_UNSUPPORTED_FORMAT | \
-+ FM_PORT_FRM_ERR_LENGTH | FM_PORT_FRM_ERR_DMA)
-+
-+#ifndef CONFIG_FSL_DPAA_ETH_JUMBO_FRAME
-+/* The raw buffer size must be cacheline aligned.
-+ * Normally we use 2K buffers.
-+ */
-+#define DPA_BP_RAW_SIZE 2048
-+#else
-+/* For jumbo frame optimizations, use buffers large enough to accommodate
-+ * 9.6K frames, FD maximum offset, skb sh_info overhead and some extra
-+ * space to account for further alignments.
-+ */
-+#define DPA_MAX_FRM_SIZE 9600
-+#ifdef CONFIG_PPC
-+#define DPA_BP_RAW_SIZE \
-+ ((DPA_MAX_FRM_SIZE + DPA_MAX_FD_OFFSET + \
-+ sizeof(struct skb_shared_info) + 128) & ~(SMP_CACHE_BYTES - 1))
-+#else /* CONFIG_PPC */
-+#define DPA_BP_RAW_SIZE ((unlikely(dpaa_errata_a010022)) ? 2048 : \
-+ ((DPA_MAX_FRM_SIZE + DPA_MAX_FD_OFFSET + \
-+ sizeof(struct skb_shared_info) + 128) & ~(SMP_CACHE_BYTES - 1)))
-+#endif /* CONFIG_PPC */
-+#endif /* CONFIG_FSL_DPAA_ETH_JUMBO_FRAME */
-+
-+/* This is what FMan is ever allowed to use.
-+ * FMan-DMA requires 16-byte alignment for Rx buffers, but SKB_DATA_ALIGN is
-+ * even stronger (SMP_CACHE_BYTES-aligned), so we just get away with that,
-+ * via SKB_WITH_OVERHEAD(). We can't rely on netdev_alloc_frag() giving us
-+ * half-page-aligned buffers (can we?), so we reserve some more space
-+ * for start-of-buffer alignment.
-+ */
-+#define dpa_bp_size(buffer_layout) (SKB_WITH_OVERHEAD(DPA_BP_RAW_SIZE) - \
-+ SMP_CACHE_BYTES)
-+/* We must ensure that skb_shinfo is always cacheline-aligned. */
-+#define DPA_SKB_SIZE(size) ((size) & ~(SMP_CACHE_BYTES - 1))
-+
-+/* Maximum size of a buffer for which recycling is allowed.
-+ * We need an upper limit such that forwarded skbs that get reallocated on Tx
-+ * aren't allowed to grow unboundedly. On the other hand, we need to make sure
-+ * that skbs allocated by us will not fail to be recycled due to their size.
-+ *
-+ * For a requested size, the kernel allocator provides the next power of two
-+ * sized block, which the stack will use as is, regardless of the actual size
-+ * it required; since we must accommodate at most 9.6K buffers (L2 maximum
-+ * supported frame size), set the recycling upper limit to 16K.
-+ */
-+#define DPA_RECYCLE_MAX_SIZE 16384
-+
-+#if defined(CONFIG_FSL_SDK_FMAN_TEST)
-+/*TODO: temporary for fman pcd testing */
-+#define FMAN_PCD_TESTS_MAX_NUM_RANGES 20
-+#endif
-+
-+#define DPAA_ETH_FQ_DELTA 0x10000
-+
-+#define DPAA_ETH_PCD_FQ_BASE(device_addr) \
-+ (((device_addr) & 0x1fffff) >> 6)
-+
-+#define DPAA_ETH_PCD_FQ_HI_PRIO_BASE(device_addr) \
-+ (DPAA_ETH_FQ_DELTA + DPAA_ETH_PCD_FQ_BASE(device_addr))
-+
-+/* Largest value that the FQD's OAL field can hold.
-+ * This is DPAA-1.x specific.
-+ * TODO: This rather belongs in fsl_qman.h
-+ */
-+#define FSL_QMAN_MAX_OAL 127
-+
-+/* Maximum offset value for a contig or sg FD (represented on 9 bits) */
-+#define DPA_MAX_FD_OFFSET ((1 << 9) - 1)
-+
-+/* Default alignment for start of data in an Rx FD */
-+#define DPA_FD_DATA_ALIGNMENT 16
-+
-+/* Values for the L3R field of the FM Parse Results
-+ */
-+/* L3 Type field: First IP Present IPv4 */
-+#define FM_L3_PARSE_RESULT_IPV4 0x8000
-+/* L3 Type field: First IP Present IPv6 */
-+#define FM_L3_PARSE_RESULT_IPV6 0x4000
-+
-+/* Values for the L4R field of the FM Parse Results
-+ * See $8.8.4.7.20 - L4 HXS - L4 Results from DPAA-Rev2 Reference Manual.
-+ */
-+/* L4 Type field: UDP */
-+#define FM_L4_PARSE_RESULT_UDP 0x40
-+/* L4 Type field: TCP */
-+#define FM_L4_PARSE_RESULT_TCP 0x20
-+/* FD status field indicating whether the FM Parser has attempted to validate
-+ * the L4 csum of the frame.
-+ * Note that having this bit set doesn't necessarily imply that the checksum
-+ * is valid. One would have to check the parse results to find that out.
-+ */
-+#define FM_FD_STAT_L4CV 0x00000004
-+
-+
-+#define FM_FD_STAT_ERR_PHYSICAL FM_PORT_FRM_ERR_PHYSICAL
-+
-+/* Check if the parsed frame was found to be a TCP segment.
-+ *
-+ * @parse_result_ptr must be of type (fm_prs_result_t *).
-+ */
-+#define fm_l4_frame_is_tcp(parse_result_ptr) \
-+ ((parse_result_ptr)->l4r & FM_L4_PARSE_RESULT_TCP)
-+
-+/* number of Tx queues to FMan */
-+#ifdef CONFIG_FMAN_PFC
-+#define DPAA_ETH_TX_QUEUES (NR_CPUS * CONFIG_FMAN_PFC_COS_COUNT)
-+#else
-+#define DPAA_ETH_TX_QUEUES NR_CPUS
-+#endif
-+
-+#define DPAA_ETH_RX_QUEUES 128
-+
-+/* Convenience macros for storing/retrieving the skb back-pointers. They must
-+ * accommodate both recycling and confirmation paths - i.e. cases when the buf
-+ * was allocated by ourselves, respectively by the stack. In the former case,
-+ * we could store the skb at negative offset; in the latter case, we can't,
-+ * so we'll use 0 as offset.
-+ *
-+ * NB: @off is an offset from a (struct sk_buff **) pointer!
-+ */
-+#define DPA_WRITE_SKB_PTR(skb, skbh, addr, off) \
-+{ \
-+ skbh = (struct sk_buff **)addr; \
-+ *(skbh + (off)) = skb; \
-+}
-+#define DPA_READ_SKB_PTR(skb, skbh, addr, off) \
-+{ \
-+ skbh = (struct sk_buff **)addr; \
-+ skb = *(skbh + (off)); \
-+}
-+
-+#ifdef CONFIG_PM
-+/* Magic Packet wakeup */
-+#define DPAA_WOL_MAGIC 0x00000001
-+#endif
-+
-+#if defined(CONFIG_FSL_SDK_FMAN_TEST)
-+struct pcd_range {
-+ uint32_t base;
-+ uint32_t count;
-+};
-+#endif
-+
-+/* More detailed FQ types - used for fine-grained WQ assignments */
-+enum dpa_fq_type {
-+ FQ_TYPE_RX_DEFAULT = 1, /* Rx Default FQs */
-+ FQ_TYPE_RX_ERROR, /* Rx Error FQs */
-+ FQ_TYPE_RX_PCD, /* User-defined PCDs */
-+ FQ_TYPE_TX, /* "Real" Tx FQs */
-+ FQ_TYPE_TX_CONFIRM, /* Tx default Conf FQ (actually an Rx FQ) */
-+ FQ_TYPE_TX_CONF_MQ, /* Tx conf FQs (one for each Tx FQ) */
-+ FQ_TYPE_TX_ERROR, /* Tx Error FQs (these are actually Rx FQs) */
-+ FQ_TYPE_RX_PCD_HI_PRIO, /* User-defined high-priority PCDs */
-+};
-+
-+struct dpa_fq {
-+ struct qman_fq fq_base;
-+ struct list_head list;
-+ struct net_device *net_dev;
-+ bool init;
-+ uint32_t fqid;
-+ uint32_t flags;
-+ uint16_t channel;
-+ uint8_t wq;
-+ enum dpa_fq_type fq_type;
-+};
-+
-+struct dpa_fq_cbs_t {
-+ struct qman_fq rx_defq;
-+ struct qman_fq tx_defq;
-+ struct qman_fq rx_errq;
-+ struct qman_fq tx_errq;
-+ struct qman_fq egress_ern;
-+};
-+
-+struct fqid_cell {
-+ uint32_t start;
-+ uint32_t count;
-+};
-+
-+struct dpa_bp {
-+ struct bman_pool *pool;
-+ uint8_t bpid;
-+ struct device *dev;
-+ union {
-+ /* The buffer pools used for the private ports are initialized
-+ * with target_count buffers for each CPU; at runtime the
-+ * number of buffers per CPU is constantly brought back to this
-+ * level
-+ */
-+ int target_count;
-+ /* The configured value for the number of buffers in the pool,
-+ * used for shared port buffer pools
-+ */
-+ int config_count;
-+ };
-+ size_t size;
-+ bool seed_pool;
-+ /* physical address of the contiguous memory used by the pool to store
-+ * the buffers
-+ */
-+ dma_addr_t paddr;
-+ /* virtual address of the contiguous memory used by the pool to store
-+ * the buffers
-+ */
-+ void __iomem *vaddr;
-+ /* current number of buffers in the bpool alloted to this CPU */
-+ int __percpu *percpu_count;
-+ atomic_t refs;
-+ /* some bpools need to be seeded before use by this cb */
-+ int (*seed_cb)(struct dpa_bp *);
-+ /* some bpools need to be emptied before freeing; this cb is used
-+ * for freeing of individual buffers taken from the pool
-+ */
-+ void (*free_buf_cb)(void *addr);
-+};
-+
-+struct dpa_rx_errors {
-+ u64 dme; /* DMA Error */
-+ u64 fpe; /* Frame Physical Error */
-+ u64 fse; /* Frame Size Error */
-+ u64 phe; /* Header Error */
-+ u64 cse; /* Checksum Validation Error */
-+};
-+
-+/* Counters for QMan ERN frames - one counter per rejection code */
-+struct dpa_ern_cnt {
-+ u64 cg_tdrop; /* Congestion group taildrop */
-+ u64 wred; /* WRED congestion */
-+ u64 err_cond; /* Error condition */
-+ u64 early_window; /* Order restoration, frame too early */
-+ u64 late_window; /* Order restoration, frame too late */
-+ u64 fq_tdrop; /* FQ taildrop */
-+ u64 fq_retired; /* FQ is retired */
-+ u64 orp_zero; /* ORP disabled */
-+};
-+
-+struct dpa_napi_portal {
-+ struct napi_struct napi;
-+ struct qman_portal *p;
-+};
-+
-+struct dpa_percpu_priv_s {
-+ struct net_device *net_dev;
-+ struct dpa_napi_portal *np;
-+ u64 in_interrupt;
-+ u64 tx_returned;
-+ u64 tx_confirm;
-+ /* fragmented (non-linear) skbuffs received from the stack */
-+ u64 tx_frag_skbuffs;
-+ /* number of S/G frames received */
-+ u64 rx_sg;
-+
-+ struct rtnl_link_stats64 stats;
-+ struct dpa_rx_errors rx_errors;
-+ struct dpa_ern_cnt ern_cnt;
-+};
-+
-+struct dpa_priv_s {
-+ struct dpa_percpu_priv_s __percpu *percpu_priv;
-+ struct dpa_bp *dpa_bp;
-+ /* Store here the needed Tx headroom for convenience and speed
-+ * (even though it can be computed based on the fields of buf_layout)
-+ */
-+ uint16_t tx_headroom;
-+ struct net_device *net_dev;
-+ struct mac_device *mac_dev;
-+ struct qman_fq *egress_fqs[DPAA_ETH_TX_QUEUES];
-+ struct qman_fq *conf_fqs[DPAA_ETH_TX_QUEUES];
-+
-+ size_t bp_count;
-+
-+ uint16_t channel; /* "fsl,qman-channel-id" */
-+ struct list_head dpa_fq_list;
-+
-+#ifdef CONFIG_FSL_DPAA_DBG_LOOP
-+ struct dentry *debugfs_loop_file;
-+#endif
-+
-+ uint32_t msg_enable; /* net_device message level */
-+#ifdef CONFIG_FSL_DPAA_1588
-+ struct dpa_ptp_tsu *tsu;
-+#endif
-+
-+#if defined(CONFIG_FSL_SDK_FMAN_TEST)
-+/* TODO: this is temporary until pcd support is implemented in dpaa */
-+ int priv_pcd_num_ranges;
-+ struct pcd_range priv_pcd_ranges[FMAN_PCD_TESTS_MAX_NUM_RANGES];
-+#endif
-+
-+ struct {
-+ /**
-+ * All egress queues to a given net device belong to one
-+ * (and the same) congestion group.
-+ */
-+ struct qman_cgr cgr;
-+ /* If congested, when it began. Used for performance stats. */
-+ u32 congestion_start_jiffies;
-+ /* Number of jiffies the Tx port was congested. */
-+ u32 congested_jiffies;
-+ /**
-+ * Counter for the number of times the CGR
-+ * entered congestion state
-+ */
-+ u32 cgr_congested_count;
-+ } cgr_data;
-+ /* Use a per-port CGR for ingress traffic. */
-+ bool use_ingress_cgr;
-+ struct qman_cgr ingress_cgr;
-+
-+#ifdef CONFIG_FSL_DPAA_TS
-+ bool ts_tx_en; /* Tx timestamping enabled */
-+ bool ts_rx_en; /* Rx timestamping enabled */
-+#endif /* CONFIG_FSL_DPAA_TS */
-+
-+ struct dpa_buffer_layout_s *buf_layout;
-+ uint16_t rx_headroom;
-+ char if_type[30];
-+
-+ void *peer;
-+#ifdef CONFIG_PM
-+ u32 wol;
-+#endif
-+#ifdef CONFIG_FSL_DPAA_DBG_LOOP
-+ int loop_id;
-+ int loop_to;
-+#endif
-+#ifdef CONFIG_FSL_DPAA_CEETM
-+ bool ceetm_en; /* CEETM QoS enabled */
-+#endif
-+};
-+
-+struct fm_port_fqs {
-+ struct dpa_fq *tx_defq;
-+ struct dpa_fq *tx_errq;
-+ struct dpa_fq *rx_defq;
-+ struct dpa_fq *rx_errq;
-+};
-+
-+
-+#ifdef CONFIG_FSL_DPAA_DBG_LOOP
-+extern struct net_device *dpa_loop_netdevs[20];
-+#endif
-+
-+/* functions with different implementation for SG and non-SG: */
-+int dpa_bp_priv_seed(struct dpa_bp *dpa_bp);
-+int dpaa_eth_refill_bpools(struct dpa_bp *dpa_bp, int *count_ptr);
-+void __hot _dpa_rx(struct net_device *net_dev,
-+ struct qman_portal *portal,
-+ const struct dpa_priv_s *priv,
-+ struct dpa_percpu_priv_s *percpu_priv,
-+ const struct qm_fd *fd,
-+ u32 fqid,
-+ int *count_ptr);
-+int __hot dpa_tx(struct sk_buff *skb, struct net_device *net_dev);
-+int __hot dpa_tx_extended(struct sk_buff *skb, struct net_device *net_dev,
-+ struct qman_fq *egress_fq, struct qman_fq *conf_fq);
-+struct sk_buff *_dpa_cleanup_tx_fd(const struct dpa_priv_s *priv,
-+ const struct qm_fd *fd);
-+void __hot _dpa_process_parse_results(const fm_prs_result_t *parse_results,
-+ const struct qm_fd *fd,
-+ struct sk_buff *skb,
-+ int *use_gro);
-+#ifndef CONFIG_FSL_DPAA_TS
-+bool dpa_skb_is_recyclable(struct sk_buff *skb);
-+bool dpa_buf_is_recyclable(struct sk_buff *skb,
-+ uint32_t min_size,
-+ uint16_t min_offset,
-+ unsigned char **new_buf_start);
-+#endif
-+int __hot skb_to_contig_fd(struct dpa_priv_s *priv,
-+ struct sk_buff *skb, struct qm_fd *fd,
-+ int *count_ptr, int *offset);
-+int __hot skb_to_sg_fd(struct dpa_priv_s *priv,
-+ struct sk_buff *skb, struct qm_fd *fd);
-+int __cold __attribute__((nonnull))
-+ _dpa_fq_free(struct device *dev, struct qman_fq *fq);
-+
-+/* Turn on HW checksum computation for this outgoing frame.
-+ * If the current protocol is not something we support in this regard
-+ * (or if the stack has already computed the SW checksum), we do nothing.
-+ *
-+ * Returns 0 if all goes well (or HW csum doesn't apply), and a negative value
-+ * otherwise.
-+ *
-+ * Note that this function may modify the fd->cmd field and the skb data buffer
-+ * (the Parse Results area).
-+ */
-+int dpa_enable_tx_csum(struct dpa_priv_s *priv,
-+ struct sk_buff *skb, struct qm_fd *fd, char *parse_results);
-+
-+static inline int dpaa_eth_napi_schedule(struct dpa_percpu_priv_s *percpu_priv,
-+ struct qman_portal *portal)
-+{
-+ /* In case of threaded ISR for RT enable kernel,
-+ * in_irq() does not return appropriate value, so use
-+ * in_serving_softirq to distinguish softirq or irq context.
-+ */
-+ if (unlikely(in_irq() || !in_serving_softirq())) {
-+ /* Disable QMan IRQ and invoke NAPI */
-+ int ret = qman_p_irqsource_remove(portal, QM_PIRQ_DQRI);
-+ if (likely(!ret)) {
-+ const struct qman_portal_config *pc =
-+ qman_p_get_portal_config(portal);
-+ struct dpa_napi_portal *np =
-+ &percpu_priv->np[pc->index];
-+
-+ np->p = portal;
-+ napi_schedule(&np->napi);
-+ percpu_priv->in_interrupt++;
-+ return 1;
-+ }
-+ }
-+ return 0;
-+}
-+
-+static inline ssize_t __const __must_check __attribute__((nonnull))
-+dpa_fd_length(const struct qm_fd *fd)
-+{
-+ return fd->length20;
-+}
-+
-+static inline ssize_t __const __must_check __attribute__((nonnull))
-+dpa_fd_offset(const struct qm_fd *fd)
-+{
-+ return fd->offset;
-+}
-+
-+static inline uint16_t dpa_get_headroom(struct dpa_buffer_layout_s *bl)
-+{
-+ uint16_t headroom;
-+ /* The frame headroom must accommodate:
-+ * - the driver private data area
-+ * - parse results, hash results, timestamp if selected
-+ * - manip extra space
-+ * If either hash results or time stamp are selected, both will
-+ * be copied to/from the frame headroom, as TS is located between PR and
-+ * HR in the IC and IC copy size has a granularity of 16bytes
-+ * (see description of FMBM_RICP and FMBM_TICP registers in DPAARM)
-+ *
-+ * Also make sure the headroom is a multiple of data_align bytes
-+ */
-+ headroom = (uint16_t)(bl->priv_data_size +
-+ (bl->parse_results ? DPA_PARSE_RESULTS_SIZE : 0) +
-+ (bl->hash_results || bl->time_stamp ?
-+ DPA_TIME_STAMP_SIZE + DPA_HASH_RESULTS_SIZE : 0) +
-+ bl->manip_extra_space);
-+
-+ return bl->data_align ? ALIGN(headroom, bl->data_align) : headroom;
-+}
-+
-+int fm_mac_dump_regs(struct mac_device *h_dev, char *buf, int n);
-+int fm_mac_dump_rx_stats(struct mac_device *h_dev, char *buf, int n);
-+int fm_mac_dump_tx_stats(struct mac_device *h_dev, char *buf, int n);
-+
-+void dpaa_eth_sysfs_remove(struct device *dev);
-+void dpaa_eth_sysfs_init(struct device *dev);
-+int dpaa_eth_poll(struct napi_struct *napi, int budget);
-+
-+void dpa_private_napi_del(struct net_device *net_dev);
-+
-+/* Equivalent to a memset(0), but works faster */
-+static inline void clear_fd(struct qm_fd *fd)
-+{
-+ fd->opaque_addr = 0;
-+ fd->opaque = 0;
-+ fd->cmd = 0;
-+}
-+
-+static inline int _dpa_tx_fq_to_id(const struct dpa_priv_s *priv,
-+ struct qman_fq *tx_fq)
-+{
-+ int i;
-+
-+ for (i = 0; i < DPAA_ETH_TX_QUEUES; i++)
-+ if (priv->egress_fqs[i] == tx_fq)
-+ return i;
-+
-+ return -EINVAL;
-+}
-+
-+static inline int __hot dpa_xmit(struct dpa_priv_s *priv,
-+ struct rtnl_link_stats64 *percpu_stats,
-+ struct qm_fd *fd, struct qman_fq *egress_fq,
-+ struct qman_fq *conf_fq)
-+{
-+ int err, i;
-+
-+ if (fd->bpid == 0xff)
-+ fd->cmd |= qman_fq_fqid(conf_fq);
-+
-+ /* Trace this Tx fd */
-+ trace_dpa_tx_fd(priv->net_dev, egress_fq, fd);
-+
-+ for (i = 0; i < 100000; i++) {
-+ err = qman_enqueue(egress_fq, fd, 0);
-+ if (err != -EBUSY)
-+ break;
-+ }
-+
-+ if (unlikely(err < 0)) {
-+ /* TODO differentiate b/w -EBUSY (EQCR full) and other codes? */
-+ percpu_stats->tx_errors++;
-+ percpu_stats->tx_fifo_errors++;
-+ return err;
-+ }
-+
-+ percpu_stats->tx_packets++;
-+ percpu_stats->tx_bytes += dpa_fd_length(fd);
-+
-+ return 0;
-+}
-+
-+/* Use multiple WQs for FQ assignment:
-+ * - Tx Confirmation queues go to WQ1.
-+ * - Rx Default, Tx and PCD queues go to WQ3 (no differentiation between
-+ * Rx and Tx traffic, or between Rx Default and Rx PCD frames).
-+ * - Rx Error and Tx Error queues go to WQ2 (giving them a better chance
-+ * to be scheduled, in case there are many more FQs in WQ3).
-+ * This ensures that Tx-confirmed buffers are timely released. In particular,
-+ * it avoids congestion on the Tx Confirm FQs, which can pile up PFDRs if they
-+ * are greatly outnumbered by other FQs in the system (usually PCDs), while
-+ * dequeue scheduling is round-robin.
-+ */
-+static inline void _dpa_assign_wq(struct dpa_fq *fq)
-+{
-+ switch (fq->fq_type) {
-+ case FQ_TYPE_TX_CONFIRM:
-+ case FQ_TYPE_TX_CONF_MQ:
-+ fq->wq = 1;
-+ break;
-+ case FQ_TYPE_RX_DEFAULT:
-+ case FQ_TYPE_TX:
-+ fq->wq = 3;
-+ break;
-+ case FQ_TYPE_RX_ERROR:
-+ case FQ_TYPE_TX_ERROR:
-+ case FQ_TYPE_RX_PCD_HI_PRIO:
-+ fq->wq = 2;
-+ break;
-+ case FQ_TYPE_RX_PCD:
-+ fq->wq = 5;
-+ break;
-+ default:
-+ WARN(1, "Invalid FQ type %d for FQID %d!\n",
-+ fq->fq_type, fq->fqid);
-+ }
-+}
-+
-+#ifdef CONFIG_FSL_DPAA_ETH_USE_NDO_SELECT_QUEUE
-+/* Use in lieu of skb_get_queue_mapping() */
-+#ifdef CONFIG_FMAN_PFC
-+#define dpa_get_queue_mapping(skb) \
-+ (((skb)->priority < CONFIG_FMAN_PFC_COS_COUNT) ? \
-+ ((skb)->priority * dpa_num_cpus + smp_processor_id()) : \
-+ ((CONFIG_FMAN_PFC_COS_COUNT - 1) * \
-+ dpa_num_cpus + smp_processor_id()));
-+
-+#else
-+#define dpa_get_queue_mapping(skb) \
-+ raw_smp_processor_id()
-+#endif
-+#else
-+/* Use the queue selected by XPS */
-+#define dpa_get_queue_mapping(skb) \
-+ skb_get_queue_mapping(skb)
-+#endif
-+
-+#ifdef CONFIG_PTP_1588_CLOCK_DPAA
-+struct ptp_priv_s {
-+ struct device_node *node;
-+ struct platform_device *of_dev;
-+ struct ptp_clock *clock;
-+ struct mac_device *mac_dev;
-+};
-+extern struct ptp_priv_s ptp_priv;
-+#endif
-+
-+static inline void _dpa_bp_free_pf(void *addr)
-+{
-+ put_page(virt_to_head_page(addr));
-+}
-+
-+/* LS1043A SoC has a HW issue regarding FMan DMA transactions; The issue
-+ * manifests itself at high traffic rates when frames cross 4K memory
-+ * boundaries or when they are not aligned to 16 bytes; For the moment, we
-+ * use a SW workaround to avoid frames larger than 4K or that exceed 4K
-+ * alignments and to realign the frames to 16 bytes.
-+ */
-+
-+#ifndef CONFIG_PPC
-+extern bool dpaa_errata_a010022; /* SoC affected by A010022 errata */
-+#define NONREC_MARK 0x01
-+#define HAS_DMA_ISSUE(start, size) \
-+ (((uintptr_t)(start) + (size)) > \
-+ (((uintptr_t)(start) + 0x1000) & ~0xFFF))
-+#endif /* !CONFIG_PPC */
-+
-+#endif /* __DPA_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_dpaa/dpaa_eth_base.c
-@@ -0,0 +1,205 @@
-+/* Copyright 2008-2013 Freescale Semiconductor, Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#ifdef CONFIG_FSL_DPAA_ETH_DEBUG
-+#define pr_fmt(fmt) \
-+ KBUILD_MODNAME ": %s:%hu:%s() " fmt, \
-+ KBUILD_BASENAME".c", __LINE__, __func__
-+#else
-+#define pr_fmt(fmt) \
-+ KBUILD_MODNAME ": " fmt
-+#endif
-+
-+#include <linux/init.h>
-+#include <linux/module.h>
-+#include <linux/io.h>
-+#include <linux/of_platform.h>
-+#include <linux/of_net.h>
-+#include <linux/etherdevice.h>
-+#include <linux/kthread.h>
-+#include <linux/percpu.h>
-+#include <linux/highmem.h>
-+#include <linux/sort.h>
-+#include <linux/fsl_qman.h>
-+#include "dpaa_eth.h"
-+#include "dpaa_eth_common.h"
-+#include "dpaa_eth_base.h"
-+
-+#define DPA_DESCRIPTION "FSL DPAA Advanced drivers:"
-+
-+MODULE_LICENSE("Dual BSD/GPL");
-+
-+uint8_t advanced_debug = -1;
-+module_param(advanced_debug, byte, S_IRUGO);
-+MODULE_PARM_DESC(advanced_debug, "Module/Driver verbosity level");
-+EXPORT_SYMBOL(advanced_debug);
-+
-+static int dpa_bp_cmp(const void *dpa_bp0, const void *dpa_bp1)
-+{
-+ return ((struct dpa_bp *)dpa_bp0)->size -
-+ ((struct dpa_bp *)dpa_bp1)->size;
-+}
-+
-+struct dpa_bp * __cold __must_check /* __attribute__((nonnull)) */
-+dpa_bp_probe(struct platform_device *_of_dev, size_t *count)
-+{
-+ int i, lenp, na, ns, err;
-+ struct device *dev;
-+ struct device_node *dev_node;
-+ const __be32 *bpool_cfg;
-+ struct dpa_bp *dpa_bp;
-+ u32 bpid;
-+
-+ dev = &_of_dev->dev;
-+
-+ *count = of_count_phandle_with_args(dev->of_node,
-+ "fsl,bman-buffer-pools", NULL);
-+ if (*count < 1) {
-+ dev_err(dev, "missing fsl,bman-buffer-pools device tree entry\n");
-+ return ERR_PTR(-EINVAL);
-+ }
-+
-+ dpa_bp = devm_kzalloc(dev, *count * sizeof(*dpa_bp), GFP_KERNEL);
-+ if (dpa_bp == NULL) {
-+ dev_err(dev, "devm_kzalloc() failed\n");
-+ return ERR_PTR(-ENOMEM);
-+ }
-+
-+ dev_node = of_find_node_by_path("/");
-+ if (unlikely(dev_node == NULL)) {
-+ dev_err(dev, "of_find_node_by_path(/) failed\n");
-+ return ERR_PTR(-EINVAL);
-+ }
-+
-+ na = of_n_addr_cells(dev_node);
-+ ns = of_n_size_cells(dev_node);
-+
-+ for (i = 0; i < *count; i++) {
-+ of_node_put(dev_node);
-+
-+ dev_node = of_parse_phandle(dev->of_node,
-+ "fsl,bman-buffer-pools", i);
-+ if (dev_node == NULL) {
-+ dev_err(dev, "of_find_node_by_phandle() failed\n");
-+ return ERR_PTR(-EFAULT);
-+ }
-+
-+ if (unlikely(!of_device_is_compatible(dev_node, "fsl,bpool"))) {
-+ dev_err(dev,
-+ "!of_device_is_compatible(%s, fsl,bpool)\n",
-+ dev_node->full_name);
-+ dpa_bp = ERR_PTR(-EINVAL);
-+ goto _return_of_node_put;
-+ }
-+
-+ err = of_property_read_u32(dev_node, "fsl,bpid", &bpid);
-+ if (err) {
-+ dev_err(dev, "Cannot find buffer pool ID in the device tree\n");
-+ dpa_bp = ERR_PTR(-EINVAL);
-+ goto _return_of_node_put;
-+ }
-+ dpa_bp[i].bpid = (uint8_t)bpid;
-+
-+ bpool_cfg = of_get_property(dev_node, "fsl,bpool-ethernet-cfg",
-+ &lenp);
-+ if (bpool_cfg && (lenp == (2 * ns + na) * sizeof(*bpool_cfg))) {
-+ const uint32_t *seed_pool;
-+
-+ dpa_bp[i].config_count =
-+ (int)of_read_number(bpool_cfg, ns);
-+ dpa_bp[i].size =
-+ (size_t)of_read_number(bpool_cfg + ns, ns);
-+ dpa_bp[i].paddr =
-+ of_read_number(bpool_cfg + 2 * ns, na);
-+
-+ seed_pool = of_get_property(dev_node,
-+ "fsl,bpool-ethernet-seeds", &lenp);
-+ dpa_bp[i].seed_pool = !!seed_pool;
-+
-+ } else {
-+ dev_err(dev,
-+ "Missing/invalid fsl,bpool-ethernet-cfg device tree entry for node %s\n",
-+ dev_node->full_name);
-+ dpa_bp = ERR_PTR(-EINVAL);
-+ goto _return_of_node_put;
-+ }
-+ }
-+
-+ sort(dpa_bp, *count, sizeof(*dpa_bp), dpa_bp_cmp, NULL);
-+
-+ return dpa_bp;
-+
-+_return_of_node_put:
-+ if (dev_node)
-+ of_node_put(dev_node);
-+
-+ return dpa_bp;
-+}
-+EXPORT_SYMBOL(dpa_bp_probe);
-+
-+int dpa_bp_create(struct net_device *net_dev, struct dpa_bp *dpa_bp,
-+ size_t count)
-+{
-+ struct dpa_priv_s *priv = netdev_priv(net_dev);
-+ int i;
-+
-+ priv->dpa_bp = dpa_bp;
-+ priv->bp_count = count;
-+
-+ for (i = 0; i < count; i++) {
-+ int err;
-+ err = dpa_bp_alloc(&dpa_bp[i], net_dev->dev.parent);
-+ if (err < 0) {
-+ dpa_bp_free(priv);
-+ priv->dpa_bp = NULL;
-+ return err;
-+ }
-+ }
-+
-+ return 0;
-+}
-+EXPORT_SYMBOL(dpa_bp_create);
-+
-+static int __init __cold dpa_advanced_load(void)
-+{
-+ pr_info(DPA_DESCRIPTION "\n");
-+
-+ return 0;
-+}
-+module_init(dpa_advanced_load);
-+
-+static void __exit __cold dpa_advanced_unload(void)
-+{
-+ pr_debug(KBUILD_MODNAME ": -> %s:%s()\n",
-+ KBUILD_BASENAME".c", __func__);
-+
-+}
-+module_exit(dpa_advanced_unload);
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_dpaa/dpaa_eth_base.h
-@@ -0,0 +1,49 @@
-+/* Copyright 2008-2013 Freescale Semiconductor, Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#ifndef __DPAA_ETH_BASE_H
-+#define __DPAA_ETH_BASE_H
-+
-+#include <linux/etherdevice.h> /* struct net_device */
-+#include <linux/fsl_bman.h> /* struct bm_buffer */
-+#include <linux/of_platform.h> /* struct platform_device */
-+#include <linux/net_tstamp.h> /* struct hwtstamp_config */
-+
-+extern uint8_t advanced_debug;
-+extern const struct dpa_fq_cbs_t shared_fq_cbs;
-+extern int __hot dpa_shared_tx(struct sk_buff *skb, struct net_device *net_dev);
-+
-+struct dpa_bp * __cold __must_check /* __attribute__((nonnull)) */
-+dpa_bp_probe(struct platform_device *_of_dev, size_t *count);
-+int dpa_bp_create(struct net_device *net_dev, struct dpa_bp *dpa_bp,
-+ size_t count);
-+
-+#endif /* __DPAA_ETH_BASE_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_dpaa/dpaa_eth_ceetm.c
-@@ -0,0 +1,2115 @@
-+/* Copyright 2008-2016 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#include <linux/init.h>
-+#include "dpaa_eth_ceetm.h"
-+
-+#define DPA_CEETM_DESCRIPTION "FSL DPAA CEETM qdisc"
-+
-+const struct nla_policy ceetm_policy[TCA_CEETM_MAX + 1] = {
-+ [TCA_CEETM_COPT] = { .len = sizeof(struct tc_ceetm_copt) },
-+ [TCA_CEETM_QOPS] = { .len = sizeof(struct tc_ceetm_qopt) },
-+};
-+
-+struct Qdisc_ops ceetm_qdisc_ops;
-+
-+/* Obtain the DCP and the SP ids from the FMan port */
-+static void get_dcp_and_sp(struct net_device *dev, enum qm_dc_portal *dcp_id,
-+ unsigned int *sp_id)
-+{
-+ uint32_t channel;
-+ t_LnxWrpFmPortDev *port_dev;
-+ struct dpa_priv_s *dpa_priv = netdev_priv(dev);
-+ struct mac_device *mac_dev = dpa_priv->mac_dev;
-+
-+ port_dev = (t_LnxWrpFmPortDev *)mac_dev->port_dev[TX];
-+ channel = port_dev->txCh;
-+
-+ *sp_id = channel & CHANNEL_SP_MASK;
-+ pr_debug(KBUILD_BASENAME " : FM sub-portal ID %d\n", *sp_id);
-+
-+ if (channel < DCP0_MAX_CHANNEL) {
-+ *dcp_id = qm_dc_portal_fman0;
-+ pr_debug(KBUILD_BASENAME " : DCP ID 0\n");
-+ } else {
-+ *dcp_id = qm_dc_portal_fman1;
-+ pr_debug(KBUILD_BASENAME " : DCP ID 1\n");
-+ }
-+}
-+
-+/* Wait for the DPAA Eth driver WQ TX FQs to empty */
-+static void dpaa_drain_fqs(struct net_device *dev)
-+{
-+ const struct dpa_priv_s *priv = netdev_priv(dev);
-+ struct qm_mcr_queryfq_np np;
-+ struct qman_fq *fq;
-+ int ret, i;
-+
-+ for (i = 0; i < DPAA_ETH_TX_QUEUES; i ++) {
-+ fq = priv->egress_fqs[i];
-+ while (true) {
-+ ret = qman_query_fq_np(fq, &np);
-+ if (unlikely(ret)) {
-+ pr_err(KBUILD_BASENAME
-+ " : %s : unable to query FQ %x: %d\n",
-+ __func__, fq->fqid, ret);
-+ break;
-+ }
-+
-+ if (np.frm_cnt == 0)
-+ break;
-+ }
-+ }
-+}
-+
-+/* Wait for the DPAA CEETM TX CQs to empty */
-+static void ceetm_drain_class(struct ceetm_class *cl)
-+{
-+ struct qm_mcr_ceetm_cq_query cq_query;
-+ struct qm_ceetm_cq *cq;
-+ unsigned int idx;
-+ int ret;
-+
-+ if (!cl)
-+ return;
-+
-+ switch (cl->type) {
-+ case CEETM_ROOT:
-+ /* The ROOT classes aren't directly linked to CEETM CQs */
-+ return;
-+ case CEETM_PRIO:
-+ cq = (struct qm_ceetm_cq*)cl->prio.cq;
-+ break;
-+ case CEETM_WBFS:
-+ cq = (struct qm_ceetm_cq*)cl->wbfs.cq;
-+ break;
-+ }
-+
-+ if (!cq || !cl->ch)
-+ return;
-+
-+ /* Build the query CQID by merging the channel and the CQ IDs */
-+ idx = (cq->parent->idx << 4) | cq->idx;
-+
-+ while (true) {
-+ ret = qman_ceetm_query_cq(idx,
-+ cl->ch->dcp_idx,
-+ &cq_query);
-+ if (unlikely(ret)) {
-+ pr_err(KBUILD_BASENAME
-+ " : %s : unable to query CQ %x: %d\n",
-+ __func__, idx, ret);
-+ break;
-+ }
-+
-+ if (cq_query.frm_cnt == 0)
-+ break;
-+ }
-+}
-+
-+/* Enqueue Rejection Notification callback */
-+static void ceetm_ern(struct qman_portal *portal, struct qman_fq *fq,
-+ const struct qm_mr_entry *msg)
-+{
-+ struct dpa_percpu_priv_s *dpa_percpu_priv;
-+ struct ceetm_class_stats *cstats = NULL;
-+ const struct dpa_priv_s *dpa_priv;
-+ struct qm_fd fd = msg->ern.fd;
-+ struct net_device *net_dev;
-+ struct ceetm_fq *ceetm_fq;
-+ struct ceetm_class *cls;
-+ struct sk_buff *skb;
-+
-+ ceetm_fq = container_of(fq, struct ceetm_fq, fq);
-+ net_dev = ceetm_fq->net_dev;
-+ dpa_priv = netdev_priv(net_dev);
-+ dpa_percpu_priv = raw_cpu_ptr(dpa_priv->percpu_priv);
-+
-+ /* Increment DPA counters */
-+ dpa_percpu_priv->stats.tx_dropped++;
-+ dpa_percpu_priv->stats.tx_fifo_errors++;
-+ count_ern(dpa_percpu_priv, msg);
-+
-+ /* Increment CEETM counters */
-+ cls = ceetm_fq->ceetm_cls;
-+ switch (cls->type) {
-+ case CEETM_PRIO:
-+ cstats = this_cpu_ptr(cls->prio.cstats);
-+ break;
-+ case CEETM_WBFS:
-+ cstats = this_cpu_ptr(cls->wbfs.cstats);
-+ break;
-+ }
-+
-+ if (cstats)
-+ cstats->ern_drop_count++;
-+
-+ /* Release the buffers that were supposed to be recycled. */
-+ if (fd.bpid != 0xff) {
-+ dpa_fd_release(net_dev, &fd);
-+ return;
-+ }
-+
-+ /* Release the frames that were supposed to return on the
-+ * confirmation path.
-+ */
-+ skb = _dpa_cleanup_tx_fd(dpa_priv, &fd);
-+ dev_kfree_skb_any(skb);
-+}
-+
-+/* Congestion State Change Notification callback */
-+static void ceetm_cscn(struct qm_ceetm_ccg *ccg, void *cb_ctx, int congested)
-+{
-+ struct ceetm_fq *ceetm_fq = (struct ceetm_fq *)cb_ctx;
-+ struct dpa_priv_s *dpa_priv = netdev_priv(ceetm_fq->net_dev);
-+ struct ceetm_class *cls = ceetm_fq->ceetm_cls;
-+ struct ceetm_class_stats *cstats = NULL;
-+
-+ switch (cls->type) {
-+ case CEETM_PRIO:
-+ cstats = this_cpu_ptr(cls->prio.cstats);
-+ break;
-+ case CEETM_WBFS:
-+ cstats = this_cpu_ptr(cls->wbfs.cstats);
-+ break;
-+ }
-+
-+ ceetm_fq->congested = congested;
-+
-+ if (congested) {
-+ dpa_priv->cgr_data.congestion_start_jiffies = jiffies;
-+ dpa_priv->cgr_data.cgr_congested_count++;
-+ if (cstats)
-+ cstats->congested_count++;
-+ } else {
-+ dpa_priv->cgr_data.congested_jiffies +=
-+ (jiffies - dpa_priv->cgr_data.congestion_start_jiffies);
-+ }
-+}
-+
-+/* Allocate a ceetm fq */
-+static int ceetm_alloc_fq(struct ceetm_fq **fq, struct net_device *dev,
-+ struct ceetm_class *cls)
-+{
-+ *fq = kzalloc(sizeof(**fq), GFP_KERNEL);
-+ if (!*fq)
-+ return -ENOMEM;
-+
-+ (*fq)->net_dev = dev;
-+ (*fq)->ceetm_cls = cls;
-+ (*fq)->congested = 0;
-+ return 0;
-+}
-+
-+/* Configure a ceetm Class Congestion Group */
-+static int ceetm_config_ccg(struct qm_ceetm_ccg **ccg,
-+ struct qm_ceetm_channel *channel, unsigned int id,
-+ struct ceetm_fq *fq, struct dpa_priv_s *dpa_priv)
-+{
-+ int err;
-+ u32 cs_th;
-+ u16 ccg_mask;
-+ struct qm_ceetm_ccg_params ccg_params;
-+
-+ err = qman_ceetm_ccg_claim(ccg, channel, id, ceetm_cscn, fq);
-+ if (err)
-+ return err;
-+
-+ /* Configure the count mode (frames/bytes), enable congestion state
-+ * notifications, configure the congestion entry and exit thresholds,
-+ * enable tail-drop, configure the tail-drop mode, and set the
-+ * overhead accounting limit
-+ */
-+ ccg_mask = QM_CCGR_WE_MODE |
-+ QM_CCGR_WE_CSCN_EN |
-+ QM_CCGR_WE_CS_THRES_IN | QM_CCGR_WE_CS_THRES_OUT |
-+ QM_CCGR_WE_TD_EN | QM_CCGR_WE_TD_MODE |
-+ QM_CCGR_WE_OAL;
-+
-+ ccg_params.mode = 0; /* count bytes */
-+ ccg_params.cscn_en = 1; /* generate notifications */
-+ ccg_params.td_en = 1; /* enable tail-drop */
-+ ccg_params.td_mode = 0; /* tail-drop on congestion state */
-+ ccg_params.oal = (signed char)(min(sizeof(struct sk_buff) +
-+ dpa_priv->tx_headroom, (size_t)FSL_QMAN_MAX_OAL));
-+
-+ /* Set the congestion state thresholds according to the link speed */
-+ if (dpa_priv->mac_dev->if_support & SUPPORTED_10000baseT_Full)
-+ cs_th = CONFIG_FSL_DPAA_CEETM_CCS_THRESHOLD_10G;
-+ else
-+ cs_th = CONFIG_FSL_DPAA_CEETM_CCS_THRESHOLD_1G;
-+
-+ qm_cgr_cs_thres_set64(&ccg_params.cs_thres_in, cs_th, 1);
-+ qm_cgr_cs_thres_set64(&ccg_params.cs_thres_out,
-+ cs_th * CEETM_CCGR_RATIO, 1);
-+
-+ err = qman_ceetm_ccg_set(*ccg, ccg_mask, &ccg_params);
-+ if (err)
-+ return err;
-+
-+ return 0;
-+}
-+
-+/* Configure a ceetm Logical Frame Queue */
-+static int ceetm_config_lfq(struct qm_ceetm_cq *cq, struct ceetm_fq *fq,
-+ struct qm_ceetm_lfq **lfq)
-+{
-+ int err;
-+ u64 context_a;
-+ u32 context_b;
-+
-+ err = qman_ceetm_lfq_claim(lfq, cq);
-+ if (err)
-+ return err;
-+
-+ /* Get the former contexts in order to preserve context B */
-+ err = qman_ceetm_lfq_get_context(*lfq, &context_a, &context_b);
-+ if (err)
-+ return err;
-+
-+ context_a = CEETM_CONTEXT_A;
-+ err = qman_ceetm_lfq_set_context(*lfq, context_a, context_b);
-+ if (err)
-+ return err;
-+
-+ (*lfq)->ern = ceetm_ern;
-+
-+ err = qman_ceetm_create_fq(*lfq, &fq->fq);
-+ if (err)
-+ return err;
-+
-+ return 0;
-+}
-+
-+/* Configure a prio ceetm class */
-+static int ceetm_config_prio_cls(struct ceetm_class *cls,
-+ struct net_device *dev,
-+ unsigned int id)
-+{
-+ int err;
-+ struct dpa_priv_s *dpa_priv = netdev_priv(dev);
-+
-+ err = ceetm_alloc_fq(&cls->prio.fq, dev, cls);
-+ if (err)
-+ return err;
-+
-+ /* Claim and configure the CCG */
-+ err = ceetm_config_ccg(&cls->prio.ccg, cls->ch, id, cls->prio.fq,
-+ dpa_priv);
-+ if (err)
-+ return err;
-+
-+ /* Claim and configure the CQ */
-+ err = qman_ceetm_cq_claim(&cls->prio.cq, cls->ch, id, cls->prio.ccg);
-+ if (err)
-+ return err;
-+
-+ if (cls->shaped) {
-+ err = qman_ceetm_channel_set_cq_cr_eligibility(cls->ch, id, 1);
-+ if (err)
-+ return err;
-+
-+ err = qman_ceetm_channel_set_cq_er_eligibility(cls->ch, id, 1);
-+ if (err)
-+ return err;
-+ }
-+
-+ /* Claim and configure a LFQ */
-+ err = ceetm_config_lfq(cls->prio.cq, cls->prio.fq, &cls->prio.lfq);
-+ if (err)
-+ return err;
-+
-+ return 0;
-+}
-+
-+/* Configure a wbfs ceetm class */
-+static int ceetm_config_wbfs_cls(struct ceetm_class *cls,
-+ struct net_device *dev,
-+ unsigned int id, int type)
-+{
-+ int err;
-+ struct dpa_priv_s *dpa_priv = netdev_priv(dev);
-+
-+ err = ceetm_alloc_fq(&cls->wbfs.fq, dev, cls);
-+ if (err)
-+ return err;
-+
-+ /* Claim and configure the CCG */
-+ err = ceetm_config_ccg(&cls->wbfs.ccg, cls->ch, id, cls->wbfs.fq,
-+ dpa_priv);
-+ if (err)
-+ return err;
-+
-+ /* Claim and configure the CQ */
-+ if (type == WBFS_GRP_B)
-+ err = qman_ceetm_cq_claim_B(&cls->wbfs.cq, cls->ch, id,
-+ cls->wbfs.ccg);
-+ else
-+ err = qman_ceetm_cq_claim_A(&cls->wbfs.cq, cls->ch, id,
-+ cls->wbfs.ccg);
-+ if (err)
-+ return err;
-+
-+ /* Configure the CQ weight: real number multiplied by 100 to get rid
-+ * of the fraction
-+ */
-+ err = qman_ceetm_set_queue_weight_in_ratio(cls->wbfs.cq,
-+ cls->wbfs.weight * 100);
-+ if (err)
-+ return err;
-+
-+ /* Claim and configure a LFQ */
-+ err = ceetm_config_lfq(cls->wbfs.cq, cls->wbfs.fq, &cls->wbfs.lfq);
-+ if (err)
-+ return err;
-+
-+ return 0;
-+}
-+
-+/* Find class in qdisc hash table using given handle */
-+static inline struct ceetm_class *ceetm_find(u32 handle, struct Qdisc *sch)
-+{
-+ struct ceetm_qdisc *priv = qdisc_priv(sch);
-+ struct Qdisc_class_common *clc;
-+
-+ pr_debug(KBUILD_BASENAME " : %s : find class %X in qdisc %X\n",
-+ __func__, handle, sch->handle);
-+
-+ clc = qdisc_class_find(&priv->clhash, handle);
-+ return clc ? container_of(clc, struct ceetm_class, common) : NULL;
-+}
-+
-+/* Insert a class in the qdisc's class hash */
-+static void ceetm_link_class(struct Qdisc *sch,
-+ struct Qdisc_class_hash *clhash,
-+ struct Qdisc_class_common *common)
-+{
-+ sch_tree_lock(sch);
-+ qdisc_class_hash_insert(clhash, common);
-+ sch_tree_unlock(sch);
-+ qdisc_class_hash_grow(sch, clhash);
-+}
-+
-+/* Destroy a ceetm class */
-+static void ceetm_cls_destroy(struct Qdisc *sch, struct ceetm_class *cl)
-+{
-+ struct net_device *dev = qdisc_dev(sch);
-+
-+ if (!cl)
-+ return;
-+
-+ pr_debug(KBUILD_BASENAME " : %s : destroy class %X from under %X\n",
-+ __func__, cl->common.classid, sch->handle);
-+
-+ switch (cl->type) {
-+ case CEETM_ROOT:
-+ if (cl->root.child) {
-+ qdisc_destroy(cl->root.child);
-+ cl->root.child = NULL;
-+ }
-+
-+ if (cl->ch && qman_ceetm_channel_release(cl->ch))
-+ pr_err(KBUILD_BASENAME
-+ " : %s : error releasing the channel %d\n",
-+ __func__, cl->ch->idx);
-+
-+ break;
-+
-+ case CEETM_PRIO:
-+ if (cl->prio.child) {
-+ qdisc_destroy(cl->prio.child);
-+ cl->prio.child = NULL;
-+ }
-+
-+ /* We must make sure the CQ is empty before releasing it.
-+ * Pause all transmissions while we wait for it to drain.
-+ */
-+ netif_tx_stop_all_queues(dev);
-+ ceetm_drain_class(cl);
-+
-+ if (cl->prio.lfq && qman_ceetm_lfq_release(cl->prio.lfq))
-+ pr_err(KBUILD_BASENAME
-+ " : %s : error releasing the LFQ %d\n",
-+ __func__, cl->prio.lfq->idx);
-+
-+ if (cl->prio.cq && qman_ceetm_cq_release(cl->prio.cq))
-+ pr_err(KBUILD_BASENAME
-+ " : %s : error releasing the CQ %d\n",
-+ __func__, cl->prio.cq->idx);
-+
-+ if (cl->prio.ccg && qman_ceetm_ccg_release(cl->prio.ccg))
-+ pr_err(KBUILD_BASENAME
-+ " : %s : error releasing the CCG %d\n",
-+ __func__, cl->prio.ccg->idx);
-+
-+ kfree(cl->prio.fq);
-+
-+ if (cl->prio.cstats)
-+ free_percpu(cl->prio.cstats);
-+
-+ netif_tx_wake_all_queues(dev);
-+ break;
-+
-+ case CEETM_WBFS:
-+ /* We must make sure the CQ is empty before releasing it.
-+ * Pause all transmissions while we wait for it to drain.
-+ */
-+ netif_tx_stop_all_queues(dev);
-+ ceetm_drain_class(cl);
-+
-+ if (cl->wbfs.lfq && qman_ceetm_lfq_release(cl->wbfs.lfq))
-+ pr_err(KBUILD_BASENAME
-+ " : %s : error releasing the LFQ %d\n",
-+ __func__, cl->wbfs.lfq->idx);
-+
-+ if (cl->wbfs.cq && qman_ceetm_cq_release(cl->wbfs.cq))
-+ pr_err(KBUILD_BASENAME
-+ " : %s : error releasing the CQ %d\n",
-+ __func__, cl->wbfs.cq->idx);
-+
-+ if (cl->wbfs.ccg && qman_ceetm_ccg_release(cl->wbfs.ccg))
-+ pr_err(KBUILD_BASENAME
-+ " : %s : error releasing the CCG %d\n",
-+ __func__, cl->wbfs.ccg->idx);
-+
-+ kfree(cl->wbfs.fq);
-+
-+ if (cl->wbfs.cstats)
-+ free_percpu(cl->wbfs.cstats);
-+
-+ netif_tx_wake_all_queues(dev);
-+ }
-+
-+ tcf_destroy_chain(&cl->filter_list);
-+ kfree(cl);
-+}
-+
-+/* Destroy a ceetm qdisc */
-+static void ceetm_destroy(struct Qdisc *sch)
-+{
-+ unsigned int ntx, i;
-+ struct hlist_node *next;
-+ struct ceetm_class *cl;
-+ struct ceetm_qdisc *priv = qdisc_priv(sch);
-+ struct net_device *dev = qdisc_dev(sch);
-+
-+ pr_debug(KBUILD_BASENAME " : %s : destroy qdisc %X\n",
-+ __func__, sch->handle);
-+
-+ /* All filters need to be removed before destroying the classes */
-+ tcf_destroy_chain(&priv->filter_list);
-+
-+ for (i = 0; i < priv->clhash.hashsize; i++) {
-+ hlist_for_each_entry(cl, &priv->clhash.hash[i], common.hnode)
-+ tcf_destroy_chain(&cl->filter_list);
-+ }
-+
-+ for (i = 0; i < priv->clhash.hashsize; i++) {
-+ hlist_for_each_entry_safe(cl, next, &priv->clhash.hash[i],
-+ common.hnode)
-+ ceetm_cls_destroy(sch, cl);
-+ }
-+
-+ qdisc_class_hash_destroy(&priv->clhash);
-+
-+ switch (priv->type) {
-+ case CEETM_ROOT:
-+ dpa_disable_ceetm(dev);
-+
-+ if (priv->root.lni && qman_ceetm_lni_release(priv->root.lni))
-+ pr_err(KBUILD_BASENAME
-+ " : %s : error releasing the LNI %d\n",
-+ __func__, priv->root.lni->idx);
-+
-+ if (priv->root.sp && qman_ceetm_sp_release(priv->root.sp))
-+ pr_err(KBUILD_BASENAME
-+ " : %s : error releasing the SP %d\n",
-+ __func__, priv->root.sp->idx);
-+
-+ if (priv->root.qstats)
-+ free_percpu(priv->root.qstats);
-+
-+ if (!priv->root.qdiscs)
-+ break;
-+
-+ /* Destroy the pfifo qdiscs in case they haven't been attached
-+ * to the netdev queues yet.
-+ */
-+ for (ntx = 0; ntx < dev->num_tx_queues; ntx++)
-+ if (priv->root.qdiscs[ntx])
-+ qdisc_destroy(priv->root.qdiscs[ntx]);
-+
-+ kfree(priv->root.qdiscs);
-+ break;
-+
-+ case CEETM_PRIO:
-+ if (priv->prio.parent)
-+ priv->prio.parent->root.child = NULL;
-+ break;
-+
-+ case CEETM_WBFS:
-+ /* Reset the WBFS groups and priorities */
-+ if (priv->wbfs.ch)
-+ qman_ceetm_channel_set_group(priv->wbfs.ch, 1, 0, 0);
-+
-+ if (priv->wbfs.parent)
-+ priv->wbfs.parent->prio.child = NULL;
-+ break;
-+ }
-+}
-+
-+static int ceetm_dump(struct Qdisc *sch, struct sk_buff *skb)
-+{
-+ struct Qdisc *qdisc;
-+ unsigned int ntx, i;
-+ struct nlattr *nest;
-+ struct tc_ceetm_qopt qopt;
-+ struct ceetm_qdisc_stats *qstats;
-+ struct net_device *dev = qdisc_dev(sch);
-+ struct ceetm_qdisc *priv = qdisc_priv(sch);
-+
-+ pr_debug(KBUILD_BASENAME " : %s : qdisc %X\n", __func__, sch->handle);
-+
-+ sch_tree_lock(sch);
-+ memset(&qopt, 0, sizeof(qopt));
-+ qopt.type = priv->type;
-+ qopt.shaped = priv->shaped;
-+
-+ switch (priv->type) {
-+ case CEETM_ROOT:
-+ /* Gather statistics from the underlying pfifo qdiscs */
-+ sch->q.qlen = 0;
-+ memset(&sch->bstats, 0, sizeof(sch->bstats));
-+ memset(&sch->qstats, 0, sizeof(sch->qstats));
-+
-+ for (ntx = 0; ntx < dev->num_tx_queues; ntx++) {
-+ qdisc = netdev_get_tx_queue(dev, ntx)->qdisc_sleeping;
-+ sch->q.qlen += qdisc->q.qlen;
-+ sch->bstats.bytes += qdisc->bstats.bytes;
-+ sch->bstats.packets += qdisc->bstats.packets;
-+ sch->qstats.qlen += qdisc->qstats.qlen;
-+ sch->qstats.backlog += qdisc->qstats.backlog;
-+ sch->qstats.drops += qdisc->qstats.drops;
-+ sch->qstats.requeues += qdisc->qstats.requeues;
-+ sch->qstats.overlimits += qdisc->qstats.overlimits;
-+ }
-+
-+ for_each_online_cpu(i) {
-+ qstats = per_cpu_ptr(priv->root.qstats, i);
-+ sch->qstats.drops += qstats->drops;
-+ }
-+
-+ qopt.rate = priv->root.rate;
-+ qopt.ceil = priv->root.ceil;
-+ qopt.overhead = priv->root.overhead;
-+ break;
-+
-+ case CEETM_PRIO:
-+ qopt.qcount = priv->prio.qcount;
-+ break;
-+
-+ case CEETM_WBFS:
-+ qopt.qcount = priv->wbfs.qcount;
-+ qopt.cr = priv->wbfs.cr;
-+ qopt.er = priv->wbfs.er;
-+ break;
-+
-+ default:
-+ pr_err(KBUILD_BASENAME " : %s : invalid qdisc\n", __func__);
-+ sch_tree_unlock(sch);
-+ return -EINVAL;
-+ }
-+
-+ nest = nla_nest_start(skb, TCA_OPTIONS);
-+ if (!nest)
-+ goto nla_put_failure;
-+ if (nla_put(skb, TCA_CEETM_QOPS, sizeof(qopt), &qopt))
-+ goto nla_put_failure;
-+ nla_nest_end(skb, nest);
-+
-+ sch_tree_unlock(sch);
-+ return skb->len;
-+
-+nla_put_failure:
-+ sch_tree_unlock(sch);
-+ nla_nest_cancel(skb, nest);
-+ return -EMSGSIZE;
-+}
-+
-+/* Configure a root ceetm qdisc */
-+static int ceetm_init_root(struct Qdisc *sch, struct ceetm_qdisc *priv,
-+ struct tc_ceetm_qopt *qopt)
-+{
-+ struct netdev_queue *dev_queue;
-+ struct Qdisc *qdisc;
-+ enum qm_dc_portal dcp_id;
-+ unsigned int i, sp_id, parent_id;
-+ int err;
-+ u64 bps;
-+ struct qm_ceetm_sp *sp;
-+ struct qm_ceetm_lni *lni;
-+ struct net_device *dev = qdisc_dev(sch);
-+ struct dpa_priv_s *dpa_priv = netdev_priv(dev);
-+ struct mac_device *mac_dev = dpa_priv->mac_dev;
-+
-+ pr_debug(KBUILD_BASENAME " : %s : qdisc %X\n", __func__, sch->handle);
-+
-+ /* Validate inputs */
-+ if (sch->parent != TC_H_ROOT) {
-+ pr_err("CEETM: a root ceetm qdisc can not be attached to a class\n");
-+ tcf_destroy_chain(&priv->filter_list);
-+ qdisc_class_hash_destroy(&priv->clhash);
-+ return -EINVAL;
-+ }
-+
-+ if (!mac_dev) {
-+ pr_err("CEETM: the interface is lacking a mac\n");
-+ err = -EINVAL;
-+ goto err_init_root;
-+ }
-+
-+ /* Pre-allocate underlying pfifo qdiscs.
-+ *
-+ * We want to offload shaping and scheduling decisions to the hardware.
-+ * The pfifo qdiscs will be attached to the netdev queues and will
-+ * guide the traffic from the IP stack down to the driver with minimum
-+ * interference.
-+ *
-+ * The CEETM qdiscs and classes will be crossed when the traffic
-+ * reaches the driver.
-+ */
-+ priv->root.qdiscs = kcalloc(dev->num_tx_queues,
-+ sizeof(priv->root.qdiscs[0]),
-+ GFP_KERNEL);
-+ if (!priv->root.qdiscs) {
-+ err = -ENOMEM;
-+ goto err_init_root;
-+ }
-+
-+ for (i = 0; i < dev->num_tx_queues; i++) {
-+ dev_queue = netdev_get_tx_queue(dev, i);
-+ parent_id = TC_H_MAKE(TC_H_MAJ(sch->handle),
-+ TC_H_MIN(i + PFIFO_MIN_OFFSET));
-+
-+ qdisc = qdisc_create_dflt(dev_queue, &pfifo_qdisc_ops,
-+ parent_id);
-+ if (!qdisc) {
-+ err = -ENOMEM;
-+ goto err_init_root;
-+ }
-+
-+ priv->root.qdiscs[i] = qdisc;
-+ qdisc->flags |= TCQ_F_ONETXQUEUE;
-+ }
-+
-+ sch->flags |= TCQ_F_MQROOT;
-+
-+ priv->root.qstats = alloc_percpu(struct ceetm_qdisc_stats);
-+ if (!priv->root.qstats) {
-+ pr_err(KBUILD_BASENAME " : %s : alloc_percpu() failed\n",
-+ __func__);
-+ err = -ENOMEM;
-+ goto err_init_root;
-+ }
-+
-+ priv->shaped = qopt->shaped;
-+ priv->root.rate = qopt->rate;
-+ priv->root.ceil = qopt->ceil;
-+ priv->root.overhead = qopt->overhead;
-+
-+ /* Claim the SP */
-+ get_dcp_and_sp(dev, &dcp_id, &sp_id);
-+ err = qman_ceetm_sp_claim(&sp, dcp_id, sp_id);
-+ if (err) {
-+ pr_err(KBUILD_BASENAME " : %s : failed to claim the SP\n",
-+ __func__);
-+ goto err_init_root;
-+ }
-+
-+ priv->root.sp = sp;
-+
-+ /* Claim the LNI - will use the same id as the SP id since SPs 0-7
-+ * are connected to the TX FMan ports
-+ */
-+ err = qman_ceetm_lni_claim(&lni, dcp_id, sp_id);
-+ if (err) {
-+ pr_err(KBUILD_BASENAME " : %s : failed to claim the LNI\n",
-+ __func__);
-+ goto err_init_root;
-+ }
-+
-+ priv->root.lni = lni;
-+
-+ err = qman_ceetm_sp_set_lni(sp, lni);
-+ if (err) {
-+ pr_err(KBUILD_BASENAME " : %s : failed to link the SP and LNI\n",
-+ __func__);
-+ goto err_init_root;
-+ }
-+
-+ lni->sp = sp;
-+
-+ /* Configure the LNI shaper */
-+ if (priv->shaped) {
-+ err = qman_ceetm_lni_enable_shaper(lni, 1, priv->root.overhead);
-+ if (err) {
-+ pr_err(KBUILD_BASENAME " : %s : failed to configure the LNI shaper\n",
-+ __func__);
-+ goto err_init_root;
-+ }
-+
-+ bps = priv->root.rate << 3; /* Bps -> bps */
-+ err = qman_ceetm_lni_set_commit_rate_bps(lni, bps, dev->mtu);
-+ if (err) {
-+ pr_err(KBUILD_BASENAME " : %s : failed to configure the LNI shaper\n",
-+ __func__);
-+ goto err_init_root;
-+ }
-+
-+ bps = priv->root.ceil << 3; /* Bps -> bps */
-+ err = qman_ceetm_lni_set_excess_rate_bps(lni, bps, dev->mtu);
-+ if (err) {
-+ pr_err(KBUILD_BASENAME " : %s : failed to configure the LNI shaper\n",
-+ __func__);
-+ goto err_init_root;
-+ }
-+ }
-+
-+ /* TODO default configuration */
-+
-+ dpa_enable_ceetm(dev);
-+ return 0;
-+
-+err_init_root:
-+ ceetm_destroy(sch);
-+ return err;
-+}
-+
-+/* Configure a prio ceetm qdisc */
-+static int ceetm_init_prio(struct Qdisc *sch, struct ceetm_qdisc *priv,
-+ struct tc_ceetm_qopt *qopt)
-+{
-+ int err;
-+ unsigned int i;
-+ struct ceetm_class *parent_cl, *child_cl;
-+ struct Qdisc *parent_qdisc;
-+ struct net_device *dev = qdisc_dev(sch);
-+
-+ pr_debug(KBUILD_BASENAME " : %s : qdisc %X\n", __func__, sch->handle);
-+
-+ if (sch->parent == TC_H_ROOT) {
-+ pr_err("CEETM: a prio ceetm qdisc can not be root\n");
-+ err = -EINVAL;
-+ goto err_init_prio;
-+ }
-+
-+ parent_qdisc = qdisc_lookup(dev, TC_H_MAJ(sch->parent));
-+ if (strcmp(parent_qdisc->ops->id, ceetm_qdisc_ops.id)) {
-+ pr_err("CEETM: a ceetm qdisc can not be attached to other qdisc/class types\n");
-+ err = -EINVAL;
-+ goto err_init_prio;
-+ }
-+
-+ /* Obtain the parent root ceetm_class */
-+ parent_cl = ceetm_find(sch->parent, parent_qdisc);
-+
-+ if (!parent_cl || parent_cl->type != CEETM_ROOT) {
-+ pr_err("CEETM: a prio ceetm qdiscs can be added only under a root ceetm class\n");
-+ err = -EINVAL;
-+ goto err_init_prio;
-+ }
-+
-+ priv->prio.parent = parent_cl;
-+ parent_cl->root.child = sch;
-+
-+ priv->shaped = parent_cl->shaped;
-+ priv->prio.qcount = qopt->qcount;
-+ priv->prio.ch = parent_cl->ch;
-+
-+ /* Create and configure qcount child classes */
-+ for (i = 0; i < priv->prio.qcount; i++) {
-+ child_cl = kzalloc(sizeof(*child_cl), GFP_KERNEL);
-+ if (!child_cl) {
-+ pr_err(KBUILD_BASENAME " : %s : kzalloc() failed\n",
-+ __func__);
-+ err = -ENOMEM;
-+ goto err_init_prio;
-+ }
-+
-+ child_cl->prio.cstats = alloc_percpu(struct ceetm_class_stats);
-+ if (!child_cl->prio.cstats) {
-+ pr_err(KBUILD_BASENAME " : %s : alloc_percpu() failed\n",
-+ __func__);
-+ err = -ENOMEM;
-+ goto err_init_prio_cls;
-+ }
-+
-+ child_cl->common.classid = TC_H_MAKE(sch->handle, (i + 1));
-+ child_cl->refcnt = 1;
-+ child_cl->parent = sch;
-+ child_cl->type = CEETM_PRIO;
-+ child_cl->shaped = priv->shaped;
-+ child_cl->prio.child = NULL;
-+ child_cl->ch = priv->prio.ch;
-+
-+ /* All shaped CQs have CR and ER enabled by default */
-+ child_cl->prio.cr = child_cl->shaped;
-+ child_cl->prio.er = child_cl->shaped;
-+ child_cl->prio.fq = NULL;
-+ child_cl->prio.cq = NULL;
-+
-+ /* Configure the corresponding hardware CQ */
-+ err = ceetm_config_prio_cls(child_cl, dev, i);
-+ if (err) {
-+ pr_err(KBUILD_BASENAME " : %s : failed to configure the ceetm prio class %X\n",
-+ __func__, child_cl->common.classid);
-+ goto err_init_prio_cls;
-+ }
-+
-+ /* Add class handle in Qdisc */
-+ ceetm_link_class(sch, &priv->clhash, &child_cl->common);
-+ pr_debug(KBUILD_BASENAME " : %s : added ceetm prio class %X associated with CQ %d and CCG %d\n",
-+ __func__, child_cl->common.classid,
-+ child_cl->prio.cq->idx, child_cl->prio.ccg->idx);
-+ }
-+
-+ return 0;
-+
-+err_init_prio_cls:
-+ ceetm_cls_destroy(sch, child_cl);
-+err_init_prio:
-+ ceetm_destroy(sch);
-+ return err;
-+}
-+
-+/* Configure a wbfs ceetm qdisc */
-+static int ceetm_init_wbfs(struct Qdisc *sch, struct ceetm_qdisc *priv,
-+ struct tc_ceetm_qopt *qopt)
-+{
-+ int err, group_b, small_group;
-+ unsigned int i, id, prio_a, prio_b;
-+ struct ceetm_class *parent_cl, *child_cl, *root_cl;
-+ struct Qdisc *parent_qdisc;
-+ struct ceetm_qdisc *parent_priv;
-+ struct net_device *dev = qdisc_dev(sch);
-+
-+ pr_debug(KBUILD_BASENAME " : %s : qdisc %X\n", __func__, sch->handle);
-+
-+ /* Validate inputs */
-+ if (sch->parent == TC_H_ROOT) {
-+ pr_err("CEETM: a wbfs ceetm qdiscs can not be root\n");
-+ err = -EINVAL;
-+ goto err_init_wbfs;
-+ }
-+
-+ /* Obtain the parent prio ceetm qdisc */
-+ parent_qdisc = qdisc_lookup(dev, TC_H_MAJ(sch->parent));
-+ if (strcmp(parent_qdisc->ops->id, ceetm_qdisc_ops.id)) {
-+ pr_err("CEETM: a ceetm qdisc can not be attached to other qdisc/class types\n");
-+ err = -EINVAL;
-+ goto err_init_wbfs;
-+ }
-+
-+ /* Obtain the parent prio ceetm class */
-+ parent_cl = ceetm_find(sch->parent, parent_qdisc);
-+ parent_priv = qdisc_priv(parent_qdisc);
-+
-+ if (!parent_cl || parent_cl->type != CEETM_PRIO) {
-+ pr_err("CEETM: a wbfs ceetm qdiscs can be added only under a prio ceetm class\n");
-+ err = -EINVAL;
-+ goto err_init_wbfs;
-+ }
-+
-+ if (!qopt->qcount || !qopt->qweight[0]) {
-+ pr_err("CEETM: qcount and qweight are mandatory for a wbfs ceetm qdisc\n");
-+ err = -EINVAL;
-+ goto err_init_wbfs;
-+ }
-+
-+ priv->shaped = parent_cl->shaped;
-+
-+ if (!priv->shaped && (qopt->cr || qopt->er)) {
-+ pr_err("CEETM: CR/ER can be enabled only for shaped wbfs ceetm qdiscs\n");
-+ err = -EINVAL;
-+ goto err_init_wbfs;
-+ }
-+
-+ if (priv->shaped && !(qopt->cr || qopt->er)) {
-+ pr_err("CEETM: either CR or ER must be enabled for shaped wbfs ceetm qdiscs\n");
-+ err = -EINVAL;
-+ goto err_init_wbfs;
-+ }
-+
-+ /* Obtain the parent root ceetm class */
-+ root_cl = parent_priv->prio.parent;
-+ if ((root_cl->root.wbfs_grp_a && root_cl->root.wbfs_grp_b) ||
-+ root_cl->root.wbfs_grp_large) {
-+ pr_err("CEETM: no more wbfs classes are available\n");
-+ err = -EINVAL;
-+ goto err_init_wbfs;
-+ }
-+
-+ if ((root_cl->root.wbfs_grp_a || root_cl->root.wbfs_grp_b) &&
-+ qopt->qcount == CEETM_MAX_WBFS_QCOUNT) {
-+ pr_err("CEETM: only %d wbfs classes are available\n",
-+ CEETM_MIN_WBFS_QCOUNT);
-+ err = -EINVAL;
-+ goto err_init_wbfs;
-+ }
-+
-+ priv->wbfs.parent = parent_cl;
-+ parent_cl->prio.child = sch;
-+
-+ priv->wbfs.qcount = qopt->qcount;
-+ priv->wbfs.cr = qopt->cr;
-+ priv->wbfs.er = qopt->er;
-+ priv->wbfs.ch = parent_cl->ch;
-+
-+ /* Configure the hardware wbfs channel groups */
-+ if (priv->wbfs.qcount == CEETM_MAX_WBFS_QCOUNT) {
-+ /* Configure the large group A */
-+ priv->wbfs.group_type = WBFS_GRP_LARGE;
-+ small_group = false;
-+ group_b = false;
-+ prio_a = TC_H_MIN(parent_cl->common.classid) - 1;
-+ prio_b = prio_a;
-+
-+ } else if (root_cl->root.wbfs_grp_a) {
-+ /* Configure the group B */
-+ priv->wbfs.group_type = WBFS_GRP_B;
-+
-+ err = qman_ceetm_channel_get_group(priv->wbfs.ch, &small_group,
-+ &prio_a, &prio_b);
-+ if (err) {
-+ pr_err(KBUILD_BASENAME " : %s : failed to get group details\n",
-+ __func__);
-+ goto err_init_wbfs;
-+ }
-+
-+ small_group = true;
-+ group_b = true;
-+ prio_b = TC_H_MIN(parent_cl->common.classid) - 1;
-+ /* If group A isn't configured, configure it as group B */
-+ prio_a = prio_a ? : prio_b;
-+
-+ } else {
-+ /* Configure the small group A */
-+ priv->wbfs.group_type = WBFS_GRP_A;
-+
-+ err = qman_ceetm_channel_get_group(priv->wbfs.ch, &small_group,
-+ &prio_a, &prio_b);
-+ if (err) {
-+ pr_err(KBUILD_BASENAME " : %s : failed to get group details\n",
-+ __func__);
-+ goto err_init_wbfs;
-+ }
-+
-+ small_group = true;
-+ group_b = false;
-+ prio_a = TC_H_MIN(parent_cl->common.classid) - 1;
-+ /* If group B isn't configured, configure it as group A */
-+ prio_b = prio_b ? : prio_a;
-+ }
-+
-+ err = qman_ceetm_channel_set_group(priv->wbfs.ch, small_group, prio_a,
-+ prio_b);
-+ if (err)
-+ goto err_init_wbfs;
-+
-+ if (priv->shaped) {
-+ err = qman_ceetm_channel_set_group_cr_eligibility(priv->wbfs.ch,
-+ group_b,
-+ priv->wbfs.cr);
-+ if (err) {
-+ pr_err(KBUILD_BASENAME " : %s : failed to set group CR eligibility\n",
-+ __func__);
-+ goto err_init_wbfs;
-+ }
-+
-+ err = qman_ceetm_channel_set_group_er_eligibility(priv->wbfs.ch,
-+ group_b,
-+ priv->wbfs.er);
-+ if (err) {
-+ pr_err(KBUILD_BASENAME " : %s : failed to set group ER eligibility\n",
-+ __func__);
-+ goto err_init_wbfs;
-+ }
-+ }
-+
-+ /* Create qcount child classes */
-+ for (i = 0; i < priv->wbfs.qcount; i++) {
-+ child_cl = kzalloc(sizeof(*child_cl), GFP_KERNEL);
-+ if (!child_cl) {
-+ pr_err(KBUILD_BASENAME " : %s : kzalloc() failed\n",
-+ __func__);
-+ err = -ENOMEM;
-+ goto err_init_wbfs;
-+ }
-+
-+ child_cl->wbfs.cstats = alloc_percpu(struct ceetm_class_stats);
-+ if (!child_cl->wbfs.cstats) {
-+ pr_err(KBUILD_BASENAME " : %s : alloc_percpu() failed\n",
-+ __func__);
-+ err = -ENOMEM;
-+ goto err_init_wbfs_cls;
-+ }
-+
-+ child_cl->common.classid = TC_H_MAKE(sch->handle, (i + 1));
-+ child_cl->refcnt = 1;
-+ child_cl->parent = sch;
-+ child_cl->type = CEETM_WBFS;
-+ child_cl->shaped = priv->shaped;
-+ child_cl->wbfs.fq = NULL;
-+ child_cl->wbfs.cq = NULL;
-+ child_cl->wbfs.weight = qopt->qweight[i];
-+ child_cl->ch = priv->wbfs.ch;
-+
-+ if (priv->wbfs.group_type == WBFS_GRP_B)
-+ id = WBFS_GRP_B_OFFSET + i;
-+ else
-+ id = WBFS_GRP_A_OFFSET + i;
-+
-+ err = ceetm_config_wbfs_cls(child_cl, dev, id,
-+ priv->wbfs.group_type);
-+ if (err) {
-+ pr_err(KBUILD_BASENAME " : %s : failed to configure the ceetm wbfs class %X\n",
-+ __func__, child_cl->common.classid);
-+ goto err_init_wbfs_cls;
-+ }
-+
-+ /* Add class handle in Qdisc */
-+ ceetm_link_class(sch, &priv->clhash, &child_cl->common);
-+ pr_debug(KBUILD_BASENAME " : %s : added ceetm wbfs class %X associated with CQ %d and CCG %d\n",
-+ __func__, child_cl->common.classid,
-+ child_cl->wbfs.cq->idx, child_cl->wbfs.ccg->idx);
-+ }
-+
-+ /* Signal the root class that a group has been configured */
-+ switch (priv->wbfs.group_type) {
-+ case WBFS_GRP_LARGE:
-+ root_cl->root.wbfs_grp_large = true;
-+ break;
-+ case WBFS_GRP_A:
-+ root_cl->root.wbfs_grp_a = true;
-+ break;
-+ case WBFS_GRP_B:
-+ root_cl->root.wbfs_grp_b = true;
-+ break;
-+ }
-+
-+ return 0;
-+
-+err_init_wbfs_cls:
-+ ceetm_cls_destroy(sch, child_cl);
-+err_init_wbfs:
-+ ceetm_destroy(sch);
-+ return err;
-+}
-+
-+/* Configure a generic ceetm qdisc */
-+static int ceetm_init(struct Qdisc *sch, struct nlattr *opt)
-+{
-+ struct tc_ceetm_qopt *qopt;
-+ struct nlattr *tb[TCA_CEETM_QOPS + 1];
-+ int ret;
-+ struct ceetm_qdisc *priv = qdisc_priv(sch);
-+ struct net_device *dev = qdisc_dev(sch);
-+
-+ pr_debug(KBUILD_BASENAME " : %s : qdisc %X\n", __func__, sch->handle);
-+
-+ if (!netif_is_multiqueue(dev))
-+ return -EOPNOTSUPP;
-+
-+ if (!opt) {
-+ pr_err(KBUILD_BASENAME " : %s : tc error\n", __func__);
-+ return -EINVAL;
-+ }
-+
-+ ret = nla_parse_nested(tb, TCA_CEETM_QOPS, opt, ceetm_policy);
-+ if (ret < 0) {
-+ pr_err(KBUILD_BASENAME " : %s : tc error\n", __func__);
-+ return ret;
-+ }
-+
-+ if (!tb[TCA_CEETM_QOPS]) {
-+ pr_err(KBUILD_BASENAME " : %s : tc error\n", __func__);
-+ return -EINVAL;
-+ }
-+
-+ if (TC_H_MIN(sch->handle)) {
-+ pr_err("CEETM: a qdisc should not have a minor\n");
-+ return -EINVAL;
-+ }
-+
-+ qopt = nla_data(tb[TCA_CEETM_QOPS]);
-+
-+ /* Initialize the class hash list. Each qdisc has its own class hash */
-+ ret = qdisc_class_hash_init(&priv->clhash);
-+ if (ret < 0) {
-+ pr_err(KBUILD_BASENAME " : %s : qdisc_class_hash_init failed\n",
-+ __func__);
-+ return ret;
-+ }
-+
-+ priv->type = qopt->type;
-+
-+ switch (priv->type) {
-+ case CEETM_ROOT:
-+ netif_tx_stop_all_queues(dev);
-+ dpaa_drain_fqs(dev);
-+ ret = ceetm_init_root(sch, priv, qopt);
-+ netif_tx_wake_all_queues(dev);
-+ break;
-+ case CEETM_PRIO:
-+ ret = ceetm_init_prio(sch, priv, qopt);
-+ break;
-+ case CEETM_WBFS:
-+ ret = ceetm_init_wbfs(sch, priv, qopt);
-+ break;
-+ default:
-+ pr_err(KBUILD_BASENAME " : %s : invalid qdisc\n", __func__);
-+ ceetm_destroy(sch);
-+ ret = -EINVAL;
-+ }
-+
-+ return ret;
-+}
-+
-+/* Edit a root ceetm qdisc */
-+static int ceetm_change_root(struct Qdisc *sch, struct ceetm_qdisc *priv,
-+ struct net_device *dev,
-+ struct tc_ceetm_qopt *qopt)
-+{
-+ int err = 0;
-+ u64 bps;
-+
-+ if (priv->shaped != (bool)qopt->shaped) {
-+ pr_err("CEETM: qdisc %X is %s\n", sch->handle,
-+ priv->shaped ? "shaped" : "unshaped");
-+ return -EINVAL;
-+ }
-+
-+ /* Nothing to modify for unshaped qdiscs */
-+ if (!priv->shaped)
-+ return 0;
-+
-+ /* Configure the LNI shaper */
-+ if (priv->root.overhead != qopt->overhead) {
-+ err = qman_ceetm_lni_enable_shaper(priv->root.lni, 1,
-+ qopt->overhead);
-+ if (err)
-+ goto change_err;
-+ priv->root.overhead = qopt->overhead;
-+ }
-+
-+ if (priv->root.rate != qopt->rate) {
-+ bps = qopt->rate << 3; /* Bps -> bps */
-+ err = qman_ceetm_lni_set_commit_rate_bps(priv->root.lni, bps,
-+ dev->mtu);
-+ if (err)
-+ goto change_err;
-+ priv->root.rate = qopt->rate;
-+ }
-+
-+ if (priv->root.ceil != qopt->ceil) {
-+ bps = qopt->ceil << 3; /* Bps -> bps */
-+ err = qman_ceetm_lni_set_excess_rate_bps(priv->root.lni, bps,
-+ dev->mtu);
-+ if (err)
-+ goto change_err;
-+ priv->root.ceil = qopt->ceil;
-+ }
-+
-+ return 0;
-+
-+change_err:
-+ pr_err(KBUILD_BASENAME " : %s : failed to configure the root ceetm qdisc %X\n",
-+ __func__, sch->handle);
-+ return err;
-+}
-+
-+/* Edit a wbfs ceetm qdisc */
-+static int ceetm_change_wbfs(struct Qdisc *sch, struct ceetm_qdisc *priv,
-+ struct tc_ceetm_qopt *qopt)
-+{
-+ int err;
-+ bool group_b;
-+
-+ if (qopt->qcount) {
-+ pr_err("CEETM: the qcount can not be modified\n");
-+ return -EINVAL;
-+ }
-+
-+ if (qopt->qweight[0]) {
-+ pr_err("CEETM: the qweight can be modified through the wbfs classes\n");
-+ return -EINVAL;
-+ }
-+
-+ if (!priv->shaped && (qopt->cr || qopt->er)) {
-+ pr_err("CEETM: CR/ER can be enabled only for shaped wbfs ceetm qdiscs\n");
-+ return -EINVAL;
-+ }
-+
-+ if (priv->shaped && !(qopt->cr || qopt->er)) {
-+ pr_err("CEETM: either CR or ER must be enabled for shaped wbfs ceetm qdiscs\n");
-+ return -EINVAL;
-+ }
-+
-+ /* Nothing to modify for unshaped qdiscs */
-+ if (!priv->shaped)
-+ return 0;
-+
-+ group_b = priv->wbfs.group_type == WBFS_GRP_B;
-+
-+ if (qopt->cr != priv->wbfs.cr) {
-+ err = qman_ceetm_channel_set_group_cr_eligibility(priv->wbfs.ch,
-+ group_b,
-+ qopt->cr);
-+ if (err)
-+ goto change_err;
-+ priv->wbfs.cr = qopt->cr;
-+ }
-+
-+ if (qopt->er != priv->wbfs.er) {
-+ err = qman_ceetm_channel_set_group_er_eligibility(priv->wbfs.ch,
-+ group_b,
-+ qopt->er);
-+ if (err)
-+ goto change_err;
-+ priv->wbfs.er = qopt->er;
-+ }
-+
-+ return 0;
-+
-+change_err:
-+ pr_err(KBUILD_BASENAME " : %s : failed to configure the wbfs ceetm qdisc %X\n",
-+ __func__, sch->handle);
-+ return err;
-+}
-+
-+/* Edit a ceetm qdisc */
-+static int ceetm_change(struct Qdisc *sch, struct nlattr *opt)
-+{
-+ struct tc_ceetm_qopt *qopt;
-+ struct nlattr *tb[TCA_CEETM_QOPS + 1];
-+ int ret;
-+ struct ceetm_qdisc *priv = qdisc_priv(sch);
-+ struct net_device *dev = qdisc_dev(sch);
-+
-+ pr_debug(KBUILD_BASENAME " : %s : qdisc %X\n", __func__, sch->handle);
-+
-+ ret = nla_parse_nested(tb, TCA_CEETM_QOPS, opt, ceetm_policy);
-+ if (ret < 0) {
-+ pr_err(KBUILD_BASENAME " : %s : tc error\n", __func__);
-+ return ret;
-+ }
-+
-+ if (!tb[TCA_CEETM_QOPS]) {
-+ pr_err(KBUILD_BASENAME " : %s : tc error\n", __func__);
-+ return -EINVAL;
-+ }
-+
-+ if (TC_H_MIN(sch->handle)) {
-+ pr_err("CEETM: a qdisc should not have a minor\n");
-+ return -EINVAL;
-+ }
-+
-+ qopt = nla_data(tb[TCA_CEETM_QOPS]);
-+
-+ if (priv->type != qopt->type) {
-+ pr_err("CEETM: qdisc %X is not of the provided type\n",
-+ sch->handle);
-+ return -EINVAL;
-+ }
-+
-+ switch (priv->type) {
-+ case CEETM_ROOT:
-+ ret = ceetm_change_root(sch, priv, dev, qopt);
-+ break;
-+ case CEETM_PRIO:
-+ pr_err("CEETM: prio qdiscs can not be modified\n");
-+ ret = -EINVAL;
-+ break;
-+ case CEETM_WBFS:
-+ ret = ceetm_change_wbfs(sch, priv, qopt);
-+ break;
-+ default:
-+ pr_err(KBUILD_BASENAME " : %s : invalid qdisc\n", __func__);
-+ ret = -EINVAL;
-+ }
-+
-+ return ret;
-+}
-+
-+/* Graft the underlying pfifo qdiscs to the netdev queues.
-+ * It's safe to remove our references at this point, since the kernel will
-+ * destroy the qdiscs on its own and no cleanup from our part is required.
-+ */
-+static void ceetm_attach(struct Qdisc *sch)
-+{
-+ struct net_device *dev = qdisc_dev(sch);
-+ struct ceetm_qdisc *priv = qdisc_priv(sch);
-+ struct Qdisc *qdisc, *old_qdisc;
-+ unsigned int i;
-+
-+ pr_debug(KBUILD_BASENAME " : %s : qdisc %X\n", __func__, sch->handle);
-+
-+ for (i = 0; i < dev->num_tx_queues; i++) {
-+ qdisc = priv->root.qdiscs[i];
-+ old_qdisc = dev_graft_qdisc(qdisc->dev_queue, qdisc);
-+ if (old_qdisc)
-+ qdisc_destroy(old_qdisc);
-+ }
-+
-+ kfree(priv->root.qdiscs);
-+ priv->root.qdiscs = NULL;
-+}
-+
-+static unsigned long ceetm_cls_get(struct Qdisc *sch, u32 classid)
-+{
-+ struct ceetm_class *cl;
-+
-+ pr_debug(KBUILD_BASENAME " : %s : classid %X from qdisc %X\n",
-+ __func__, classid, sch->handle);
-+ cl = ceetm_find(classid, sch);
-+
-+ if (cl)
-+ cl->refcnt++; /* Will decrement in put() */
-+ return (unsigned long)cl;
-+}
-+
-+static void ceetm_cls_put(struct Qdisc *sch, unsigned long arg)
-+{
-+ struct ceetm_class *cl = (struct ceetm_class *)arg;
-+
-+ pr_debug(KBUILD_BASENAME " : %s : classid %X from qdisc %X\n",
-+ __func__, cl->common.classid, sch->handle);
-+ cl->refcnt--;
-+
-+ if (cl->refcnt == 0)
-+ ceetm_cls_destroy(sch, cl);
-+}
-+
-+static int ceetm_cls_change_root(struct ceetm_class *cl,
-+ struct tc_ceetm_copt *copt,
-+ struct net_device *dev)
-+{
-+ int err;
-+ u64 bps;
-+
-+ if ((bool)copt->shaped != cl->shaped) {
-+ pr_err("CEETM: class %X is %s\n", cl->common.classid,
-+ cl->shaped ? "shaped" : "unshaped");
-+ return -EINVAL;
-+ }
-+
-+ if (cl->shaped && cl->root.rate != copt->rate) {
-+ bps = copt->rate << 3; /* Bps -> bps */
-+ err = qman_ceetm_channel_set_commit_rate_bps(cl->ch, bps,
-+ dev->mtu);
-+ if (err)
-+ goto change_cls_err;
-+ cl->root.rate = copt->rate;
-+ }
-+
-+ if (cl->shaped && cl->root.ceil != copt->ceil) {
-+ bps = copt->ceil << 3; /* Bps -> bps */
-+ err = qman_ceetm_channel_set_excess_rate_bps(cl->ch, bps,
-+ dev->mtu);
-+ if (err)
-+ goto change_cls_err;
-+ cl->root.ceil = copt->ceil;
-+ }
-+
-+ if (!cl->shaped && cl->root.tbl != copt->tbl) {
-+ err = qman_ceetm_channel_set_weight(cl->ch, copt->tbl);
-+ if (err)
-+ goto change_cls_err;
-+ cl->root.tbl = copt->tbl;
-+ }
-+
-+ return 0;
-+
-+change_cls_err:
-+ pr_err(KBUILD_BASENAME " : %s : failed to configure the ceetm root class %X\n",
-+ __func__, cl->common.classid);
-+ return err;
-+}
-+
-+static int ceetm_cls_change_prio(struct ceetm_class *cl,
-+ struct tc_ceetm_copt *copt)
-+{
-+ int err;
-+
-+ if (!cl->shaped && (copt->cr || copt->er)) {
-+ pr_err("CEETM: only shaped classes can have CR and ER enabled\n");
-+ return -EINVAL;
-+ }
-+
-+ if (cl->prio.cr != (bool)copt->cr) {
-+ err = qman_ceetm_channel_set_cq_cr_eligibility(
-+ cl->prio.cq->parent,
-+ cl->prio.cq->idx,
-+ copt->cr);
-+ if (err)
-+ goto change_cls_err;
-+ cl->prio.cr = copt->cr;
-+ }
-+
-+ if (cl->prio.er != (bool)copt->er) {
-+ err = qman_ceetm_channel_set_cq_er_eligibility(
-+ cl->prio.cq->parent,
-+ cl->prio.cq->idx,
-+ copt->er);
-+ if (err)
-+ goto change_cls_err;
-+ cl->prio.er = copt->er;
-+ }
-+
-+ return 0;
-+
-+change_cls_err:
-+ pr_err(KBUILD_BASENAME " : %s : failed to configure the ceetm prio class %X\n",
-+ __func__, cl->common.classid);
-+ return err;
-+}
-+
-+static int ceetm_cls_change_wbfs(struct ceetm_class *cl,
-+ struct tc_ceetm_copt *copt)
-+{
-+ int err;
-+
-+ if (copt->weight != cl->wbfs.weight) {
-+ /* Configure the CQ weight: real number multiplied by 100 to
-+ * get rid of the fraction
-+ */
-+ err = qman_ceetm_set_queue_weight_in_ratio(cl->wbfs.cq,
-+ copt->weight * 100);
-+
-+ if (err) {
-+ pr_err(KBUILD_BASENAME " : %s : failed to configure the ceetm wbfs class %X\n",
-+ __func__, cl->common.classid);
-+ return err;
-+ }
-+
-+ cl->wbfs.weight = copt->weight;
-+ }
-+
-+ return 0;
-+}
-+
-+/* Add a ceetm root class or configure a ceetm root/prio/wbfs class */
-+static int ceetm_cls_change(struct Qdisc *sch, u32 classid, u32 parentid,
-+ struct nlattr **tca, unsigned long *arg)
-+{
-+ int err;
-+ u64 bps;
-+ struct ceetm_qdisc *priv;
-+ struct ceetm_class *cl = (struct ceetm_class *)*arg;
-+ struct nlattr *opt = tca[TCA_OPTIONS];
-+ struct nlattr *tb[__TCA_CEETM_MAX];
-+ struct tc_ceetm_copt *copt;
-+ struct qm_ceetm_channel *channel;
-+ struct net_device *dev = qdisc_dev(sch);
-+
-+ pr_debug(KBUILD_BASENAME " : %s : classid %X under qdisc %X\n",
-+ __func__, classid, sch->handle);
-+
-+ if (strcmp(sch->ops->id, ceetm_qdisc_ops.id)) {
-+ pr_err("CEETM: a ceetm class can not be attached to other qdisc/class types\n");
-+ return -EINVAL;
-+ }
-+
-+ priv = qdisc_priv(sch);
-+
-+ if (!opt) {
-+ pr_err(KBUILD_BASENAME " : %s : tc error\n", __func__);
-+ return -EINVAL;
-+ }
-+
-+ if (!cl && sch->handle != parentid) {
-+ pr_err("CEETM: classes can be attached to the root ceetm qdisc only\n");
-+ return -EINVAL;
-+ }
-+
-+ if (!cl && priv->type != CEETM_ROOT) {
-+ pr_err("CEETM: only root ceetm classes can be attached to the root ceetm qdisc\n");
-+ return -EINVAL;
-+ }
-+
-+ err = nla_parse_nested(tb, TCA_CEETM_COPT, opt, ceetm_policy);
-+ if (err < 0) {
-+ pr_err(KBUILD_BASENAME " : %s : tc error\n", __func__);
-+ return -EINVAL;
-+ }
-+
-+ if (!tb[TCA_CEETM_COPT]) {
-+ pr_err(KBUILD_BASENAME " : %s : tc error\n", __func__);
-+ return -EINVAL;
-+ }
-+
-+ if (TC_H_MIN(classid) >= PFIFO_MIN_OFFSET) {
-+ pr_err("CEETM: only minors 0x01 to 0x20 can be used for ceetm root classes\n");
-+ return -EINVAL;
-+ }
-+
-+ copt = nla_data(tb[TCA_CEETM_COPT]);
-+
-+ /* Configure an existing ceetm class */
-+ if (cl) {
-+ if (copt->type != cl->type) {
-+ pr_err("CEETM: class %X is not of the provided type\n",
-+ cl->common.classid);
-+ return -EINVAL;
-+ }
-+
-+ switch (copt->type) {
-+ case CEETM_ROOT:
-+ return ceetm_cls_change_root(cl, copt, dev);
-+
-+ case CEETM_PRIO:
-+ return ceetm_cls_change_prio(cl, copt);
-+
-+ case CEETM_WBFS:
-+ return ceetm_cls_change_wbfs(cl, copt);
-+
-+ default:
-+ pr_err(KBUILD_BASENAME " : %s : invalid class\n",
-+ __func__);
-+ return -EINVAL;
-+ }
-+ }
-+
-+ /* Add a new root ceetm class */
-+ if (copt->type != CEETM_ROOT) {
-+ pr_err("CEETM: only root ceetm classes can be attached to the root ceetm qdisc\n");
-+ return -EINVAL;
-+ }
-+
-+ if (copt->shaped && !priv->shaped) {
-+ pr_err("CEETM: can not add a shaped ceetm root class under an unshaped ceetm root qdisc\n");
-+ return -EINVAL;
-+ }
-+
-+ cl = kzalloc(sizeof(*cl), GFP_KERNEL);
-+ if (!cl)
-+ return -ENOMEM;
-+
-+ cl->type = copt->type;
-+ cl->shaped = copt->shaped;
-+ cl->root.rate = copt->rate;
-+ cl->root.ceil = copt->ceil;
-+ cl->root.tbl = copt->tbl;
-+
-+ cl->common.classid = classid;
-+ cl->refcnt = 1;
-+ cl->parent = sch;
-+ cl->root.child = NULL;
-+ cl->root.wbfs_grp_a = false;
-+ cl->root.wbfs_grp_b = false;
-+ cl->root.wbfs_grp_large = false;
-+
-+ /* Claim a CEETM channel */
-+ err = qman_ceetm_channel_claim(&channel, priv->root.lni);
-+ if (err) {
-+ pr_err(KBUILD_BASENAME " : %s : failed to claim a channel\n",
-+ __func__);
-+ goto claim_err;
-+ }
-+
-+ cl->ch = channel;
-+
-+ if (cl->shaped) {
-+ /* Configure the channel shaper */
-+ err = qman_ceetm_channel_enable_shaper(channel, 1);
-+ if (err)
-+ goto channel_err;
-+
-+ bps = cl->root.rate << 3; /* Bps -> bps */
-+ err = qman_ceetm_channel_set_commit_rate_bps(channel, bps,
-+ dev->mtu);
-+ if (err)
-+ goto channel_err;
-+
-+ bps = cl->root.ceil << 3; /* Bps -> bps */
-+ err = qman_ceetm_channel_set_excess_rate_bps(channel, bps,
-+ dev->mtu);
-+ if (err)
-+ goto channel_err;
-+
-+ } else {
-+ /* Configure the uFQ algorithm */
-+ err = qman_ceetm_channel_set_weight(channel, cl->root.tbl);
-+ if (err)
-+ goto channel_err;
-+ }
-+
-+ /* Add class handle in Qdisc */
-+ ceetm_link_class(sch, &priv->clhash, &cl->common);
-+
-+ pr_debug(KBUILD_BASENAME " : %s : configured class %X associated with channel %d\n",
-+ __func__, classid, channel->idx);
-+ *arg = (unsigned long)cl;
-+ return 0;
-+
-+channel_err:
-+ pr_err(KBUILD_BASENAME " : %s : failed to configure the channel %d\n",
-+ __func__, channel->idx);
-+ if (qman_ceetm_channel_release(channel))
-+ pr_err(KBUILD_BASENAME " : %s : failed to release the channel %d\n",
-+ __func__, channel->idx);
-+claim_err:
-+ kfree(cl);
-+ return err;
-+}
-+
-+static void ceetm_cls_walk(struct Qdisc *sch, struct qdisc_walker *arg)
-+{
-+ struct ceetm_qdisc *priv = qdisc_priv(sch);
-+ struct ceetm_class *cl;
-+ unsigned int i;
-+
-+ pr_debug(KBUILD_BASENAME " : %s : qdisc %X\n", __func__, sch->handle);
-+
-+ if (arg->stop)
-+ return;
-+
-+ for (i = 0; i < priv->clhash.hashsize; i++) {
-+ hlist_for_each_entry(cl, &priv->clhash.hash[i], common.hnode) {
-+ if (arg->count < arg->skip) {
-+ arg->count++;
-+ continue;
-+ }
-+ if (arg->fn(sch, (unsigned long)cl, arg) < 0) {
-+ arg->stop = 1;
-+ return;
-+ }
-+ arg->count++;
-+ }
-+ }
-+}
-+
-+static int ceetm_cls_dump(struct Qdisc *sch, unsigned long arg,
-+ struct sk_buff *skb, struct tcmsg *tcm)
-+{
-+ struct ceetm_class *cl = (struct ceetm_class *)arg;
-+ struct nlattr *nest;
-+ struct tc_ceetm_copt copt;
-+
-+ pr_debug(KBUILD_BASENAME " : %s : class %X under qdisc %X\n",
-+ __func__, cl->common.classid, sch->handle);
-+
-+ sch_tree_lock(sch);
-+
-+ tcm->tcm_parent = ((struct Qdisc *)cl->parent)->handle;
-+ tcm->tcm_handle = cl->common.classid;
-+
-+ memset(&copt, 0, sizeof(copt));
-+
-+ copt.shaped = cl->shaped;
-+ copt.type = cl->type;
-+
-+ switch (cl->type) {
-+ case CEETM_ROOT:
-+ if (cl->root.child)
-+ tcm->tcm_info = cl->root.child->handle;
-+
-+ copt.rate = cl->root.rate;
-+ copt.ceil = cl->root.ceil;
-+ copt.tbl = cl->root.tbl;
-+ break;
-+
-+ case CEETM_PRIO:
-+ if (cl->prio.child)
-+ tcm->tcm_info = cl->prio.child->handle;
-+
-+ copt.cr = cl->prio.cr;
-+ copt.er = cl->prio.er;
-+ break;
-+
-+ case CEETM_WBFS:
-+ copt.weight = cl->wbfs.weight;
-+ break;
-+ }
-+
-+ nest = nla_nest_start(skb, TCA_OPTIONS);
-+ if (!nest)
-+ goto nla_put_failure;
-+ if (nla_put(skb, TCA_CEETM_COPT, sizeof(copt), &copt))
-+ goto nla_put_failure;
-+ nla_nest_end(skb, nest);
-+ sch_tree_unlock(sch);
-+ return skb->len;
-+
-+nla_put_failure:
-+ sch_tree_unlock(sch);
-+ nla_nest_cancel(skb, nest);
-+ return -EMSGSIZE;
-+}
-+
-+static int ceetm_cls_delete(struct Qdisc *sch, unsigned long arg)
-+{
-+ struct ceetm_qdisc *priv = qdisc_priv(sch);
-+ struct ceetm_class *cl = (struct ceetm_class *)arg;
-+
-+ pr_debug(KBUILD_BASENAME " : %s : class %X under qdisc %X\n",
-+ __func__, cl->common.classid, sch->handle);
-+
-+ sch_tree_lock(sch);
-+ qdisc_class_hash_remove(&priv->clhash, &cl->common);
-+ cl->refcnt--;
-+
-+ /* The refcnt should be at least 1 since we have incremented it in
-+ * get(). Will decrement again in put() where we will call destroy()
-+ * to actually free the memory if it reaches 0.
-+ */
-+ WARN_ON(cl->refcnt == 0);
-+
-+ sch_tree_unlock(sch);
-+ return 0;
-+}
-+
-+/* Get the class' child qdisc, if any */
-+static struct Qdisc *ceetm_cls_leaf(struct Qdisc *sch, unsigned long arg)
-+{
-+ struct ceetm_class *cl = (struct ceetm_class *)arg;
-+
-+ pr_debug(KBUILD_BASENAME " : %s : class %X under qdisc %X\n",
-+ __func__, cl->common.classid, sch->handle);
-+
-+ switch (cl->type) {
-+ case CEETM_ROOT:
-+ return cl->root.child;
-+
-+ case CEETM_PRIO:
-+ return cl->prio.child;
-+ }
-+
-+ return NULL;
-+}
-+
-+static int ceetm_cls_graft(struct Qdisc *sch, unsigned long arg,
-+ struct Qdisc *new, struct Qdisc **old)
-+{
-+ if (new && strcmp(new->ops->id, ceetm_qdisc_ops.id)) {
-+ pr_err("CEETM: only ceetm qdiscs can be attached to ceetm classes\n");
-+ return -EOPNOTSUPP;
-+ }
-+
-+ return 0;
-+}
-+
-+static int ceetm_cls_dump_stats(struct Qdisc *sch, unsigned long arg,
-+ struct gnet_dump *d)
-+{
-+ unsigned int i;
-+ struct ceetm_class *cl = (struct ceetm_class *)arg;
-+ struct gnet_stats_basic_packed tmp_bstats;
-+ struct ceetm_class_stats *cstats = NULL;
-+ struct qm_ceetm_cq *cq = NULL;
-+ struct tc_ceetm_xstats xstats;
-+
-+ memset(&xstats, 0, sizeof(xstats));
-+ memset(&tmp_bstats, 0, sizeof(tmp_bstats));
-+
-+ switch (cl->type) {
-+ case CEETM_ROOT:
-+ return 0;
-+ case CEETM_PRIO:
-+ cq = cl->prio.cq;
-+ break;
-+ case CEETM_WBFS:
-+ cq = cl->wbfs.cq;
-+ break;
-+ }
-+
-+ for_each_online_cpu(i) {
-+ switch (cl->type) {
-+ case CEETM_PRIO:
-+ cstats = per_cpu_ptr(cl->prio.cstats, i);
-+ break;
-+ case CEETM_WBFS:
-+ cstats = per_cpu_ptr(cl->wbfs.cstats, i);
-+ break;
-+ }
-+
-+ if (cstats) {
-+ xstats.ern_drop_count += cstats->ern_drop_count;
-+ xstats.congested_count += cstats->congested_count;
-+ tmp_bstats.bytes += cstats->bstats.bytes;
-+ tmp_bstats.packets += cstats->bstats.packets;
-+ }
-+ }
-+
-+ if (gnet_stats_copy_basic(qdisc_root_sleeping_running(sch),
-+ d, NULL, &tmp_bstats) < 0)
-+ return -1;
-+
-+ if (cq && qman_ceetm_cq_get_dequeue_statistics(cq, 0,
-+ &xstats.frame_count,
-+ &xstats.byte_count))
-+ return -1;
-+
-+ return gnet_stats_copy_app(d, &xstats, sizeof(xstats));
-+}
-+
-+static struct tcf_proto **ceetm_tcf_chain(struct Qdisc *sch, unsigned long arg)
-+{
-+ struct ceetm_qdisc *priv = qdisc_priv(sch);
-+ struct ceetm_class *cl = (struct ceetm_class *)arg;
-+ struct tcf_proto **fl = cl ? &cl->filter_list : &priv->filter_list;
-+
-+ pr_debug(KBUILD_BASENAME " : %s : class %X under qdisc %X\n", __func__,
-+ cl ? cl->common.classid : 0, sch->handle);
-+ return fl;
-+}
-+
-+static unsigned long ceetm_tcf_bind(struct Qdisc *sch, unsigned long parent,
-+ u32 classid)
-+{
-+ struct ceetm_class *cl = ceetm_find(classid, sch);
-+
-+ pr_debug(KBUILD_BASENAME " : %s : class %X under qdisc %X\n", __func__,
-+ cl ? cl->common.classid : 0, sch->handle);
-+ return (unsigned long)cl;
-+}
-+
-+static void ceetm_tcf_unbind(struct Qdisc *sch, unsigned long arg)
-+{
-+ struct ceetm_class *cl = (struct ceetm_class *)arg;
-+
-+ pr_debug(KBUILD_BASENAME " : %s : class %X under qdisc %X\n", __func__,
-+ cl ? cl->common.classid : 0, sch->handle);
-+}
-+
-+const struct Qdisc_class_ops ceetm_cls_ops = {
-+ .graft = ceetm_cls_graft,
-+ .leaf = ceetm_cls_leaf,
-+ .get = ceetm_cls_get,
-+ .put = ceetm_cls_put,
-+ .change = ceetm_cls_change,
-+ .delete = ceetm_cls_delete,
-+ .walk = ceetm_cls_walk,
-+ .tcf_chain = ceetm_tcf_chain,
-+ .bind_tcf = ceetm_tcf_bind,
-+ .unbind_tcf = ceetm_tcf_unbind,
-+ .dump = ceetm_cls_dump,
-+ .dump_stats = ceetm_cls_dump_stats,
-+};
-+
-+struct Qdisc_ops ceetm_qdisc_ops __read_mostly = {
-+ .id = "ceetm",
-+ .priv_size = sizeof(struct ceetm_qdisc),
-+ .cl_ops = &ceetm_cls_ops,
-+ .init = ceetm_init,
-+ .destroy = ceetm_destroy,
-+ .change = ceetm_change,
-+ .dump = ceetm_dump,
-+ .attach = ceetm_attach,
-+ .owner = THIS_MODULE,
-+};
-+
-+/* Run the filters and classifiers attached to the qdisc on the provided skb */
-+static struct ceetm_class *ceetm_classify(struct sk_buff *skb,
-+ struct Qdisc *sch, int *qerr,
-+ bool *act_drop)
-+{
-+ struct ceetm_qdisc *priv = qdisc_priv(sch);
-+ struct ceetm_class *cl = NULL, *wbfs_cl;
-+ struct tcf_result res;
-+ struct tcf_proto *tcf;
-+ int result;
-+
-+ *qerr = NET_XMIT_SUCCESS | __NET_XMIT_BYPASS;
-+ tcf = priv->filter_list;
-+ while (tcf && (result = tc_classify(skb, tcf, &res, false)) >= 0) {
-+#ifdef CONFIG_NET_CLS_ACT
-+ switch (result) {
-+ case TC_ACT_QUEUED:
-+ case TC_ACT_STOLEN:
-+ *qerr = NET_XMIT_SUCCESS | __NET_XMIT_STOLEN;
-+ case TC_ACT_SHOT:
-+ /* No valid class found due to action */
-+ *act_drop = true;
-+ return NULL;
-+ }
-+#endif
-+ cl = (void *)res.class;
-+ if (!cl) {
-+ if (res.classid == sch->handle) {
-+ /* The filter leads to the qdisc */
-+ /* TODO default qdisc */
-+ return NULL;
-+ }
-+
-+ cl = ceetm_find(res.classid, sch);
-+ if (!cl)
-+ /* The filter leads to an invalid class */
-+ break;
-+ }
-+
-+ /* The class might have its own filters attached */
-+ tcf = cl->filter_list;
-+ }
-+
-+ if (!cl) {
-+ /* No valid class found */
-+ /* TODO default qdisc */
-+ return NULL;
-+ }
-+
-+ switch (cl->type) {
-+ case CEETM_ROOT:
-+ if (cl->root.child) {
-+ /* Run the prio qdisc classifiers */
-+ return ceetm_classify(skb, cl->root.child, qerr,
-+ act_drop);
-+ } else {
-+ /* The root class does not have a child prio qdisc */
-+ /* TODO default qdisc */
-+ return NULL;
-+ }
-+ case CEETM_PRIO:
-+ if (cl->prio.child) {
-+ /* If filters lead to a wbfs class, return it.
-+ * Otherwise, return the prio class
-+ */
-+ wbfs_cl = ceetm_classify(skb, cl->prio.child, qerr,
-+ act_drop);
-+ /* A NULL result might indicate either an erroneous
-+ * filter, or no filters at all. We will assume the
-+ * latter
-+ */
-+ return wbfs_cl ? : cl;
-+ }
-+ }
-+
-+ /* For wbfs and childless prio classes, return the class directly */
-+ return cl;
-+}
-+
-+int __hot ceetm_tx(struct sk_buff *skb, struct net_device *net_dev)
-+{
-+ const int queue_mapping = dpa_get_queue_mapping(skb);
-+ struct Qdisc *sch = net_dev->qdisc;
-+ struct ceetm_class_stats *cstats;
-+ struct ceetm_qdisc_stats *qstats;
-+ struct dpa_priv_s *priv_dpa;
-+ struct ceetm_fq *ceetm_fq;
-+ struct ceetm_qdisc *priv;
-+ struct qman_fq *conf_fq;
-+ struct ceetm_class *cl;
-+ spinlock_t *root_lock;
-+ bool act_drop = false;
-+ int ret;
-+
-+ root_lock = qdisc_lock(sch);
-+ priv = qdisc_priv(sch);
-+ qstats = this_cpu_ptr(priv->root.qstats);
-+
-+ spin_lock(root_lock);
-+ cl = ceetm_classify(skb, sch, &ret, &act_drop);
-+ spin_unlock(root_lock);
-+
-+#ifdef CONFIG_NET_CLS_ACT
-+ if (act_drop) {
-+ if (ret & __NET_XMIT_BYPASS)
-+ qstats->drops++;
-+ goto drop;
-+ }
-+#endif
-+ /* TODO default class */
-+ if (unlikely(!cl)) {
-+ qstats->drops++;
-+ goto drop;
-+ }
-+
-+ priv_dpa = netdev_priv(net_dev);
-+ conf_fq = priv_dpa->conf_fqs[queue_mapping];
-+
-+ /* Choose the proper tx fq and update the basic stats (bytes and
-+ * packets sent by the class)
-+ */
-+ switch (cl->type) {
-+ case CEETM_PRIO:
-+ ceetm_fq = cl->prio.fq;
-+ cstats = this_cpu_ptr(cl->prio.cstats);
-+ break;
-+ case CEETM_WBFS:
-+ ceetm_fq = cl->wbfs.fq;
-+ cstats = this_cpu_ptr(cl->wbfs.cstats);
-+ break;
-+ default:
-+ qstats->drops++;
-+ goto drop;
-+ }
-+
-+ /* If the FQ is congested, avoid enqueuing the frame and dropping it
-+ * when it returns on the ERN path. Drop it here directly instead.
-+ */
-+ if (unlikely(ceetm_fq->congested)) {
-+ qstats->drops++;
-+ goto drop;
-+ }
-+
-+ bstats_update(&cstats->bstats, skb);
-+ return dpa_tx_extended(skb, net_dev, &ceetm_fq->fq, conf_fq);
-+
-+drop:
-+ dev_kfree_skb_any(skb);
-+ return NET_XMIT_SUCCESS;
-+}
-+
-+static int __init ceetm_register(void)
-+{
-+ int _errno = 0;
-+
-+ pr_info(KBUILD_MODNAME ": " DPA_CEETM_DESCRIPTION "\n");
-+
-+ _errno = register_qdisc(&ceetm_qdisc_ops);
-+ if (unlikely(_errno))
-+ pr_err(KBUILD_MODNAME
-+ ": %s:%hu:%s(): register_qdisc() = %d\n",
-+ KBUILD_BASENAME ".c", __LINE__, __func__, _errno);
-+
-+ return _errno;
-+}
-+
-+static void __exit ceetm_unregister(void)
-+{
-+ pr_debug(KBUILD_MODNAME ": %s:%s() ->\n",
-+ KBUILD_BASENAME ".c", __func__);
-+
-+ unregister_qdisc(&ceetm_qdisc_ops);
-+}
-+
-+module_init(ceetm_register);
-+module_exit(ceetm_unregister);
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_dpaa/dpaa_eth_ceetm.h
-@@ -0,0 +1,240 @@
-+/* Copyright 2008-2016 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#ifndef __DPAA_ETH_CEETM_H
-+#define __DPAA_ETH_CEETM_H
-+
-+#include <net/pkt_sched.h>
-+#include <net/pkt_cls.h>
-+#include <net/netlink.h>
-+#include <lnxwrp_fm.h>
-+
-+#include "mac.h"
-+#include "dpaa_eth_common.h"
-+
-+/* Mask to determine the sub-portal id from a channel number */
-+#define CHANNEL_SP_MASK 0x1f
-+/* The number of the last channel that services DCP0, connected to FMan 0.
-+ * Value validated for B4 and T series platforms.
-+ */
-+#define DCP0_MAX_CHANNEL 0x80f
-+/* A2V=1 - field A2 is valid
-+ * A0V=1 - field A0 is valid - enables frame confirmation
-+ * OVOM=1 - override operation mode bits with values from A2
-+ * EBD=1 - external buffers are deallocated at the end of the FMan flow
-+ * NL=0 - the BMI releases all the internal buffers
-+ */
-+#define CEETM_CONTEXT_A 0x1a00000080000000
-+/* The ratio between the superior and inferior congestion state thresholds. The
-+ * lower threshold is set to 7/8 of the superior one (as the default for WQ
-+ * scheduling).
-+ */
-+#define CEETM_CCGR_RATIO 0.875
-+/* For functional purposes, there are num_tx_queues pfifo qdiscs through which
-+ * frames reach the driver. Their handles start from 1:21. Handles 1:1 to 1:20
-+ * are reserved for the maximum 32 CEETM channels (majors and minors are in
-+ * hex).
-+ */
-+#define PFIFO_MIN_OFFSET 0x21
-+
-+/* A maximum of 8 CQs can be linked to a CQ channel or to a WBFS scheduler. */
-+#define CEETM_MAX_PRIO_QCOUNT 8
-+#define CEETM_MAX_WBFS_QCOUNT 8
-+#define CEETM_MIN_WBFS_QCOUNT 4
-+
-+/* The id offsets of the CQs belonging to WBFS groups (ids 8-11/15 for group A
-+ * and/or 12-15 for group B).
-+ */
-+#define WBFS_GRP_A_OFFSET 8
-+#define WBFS_GRP_B_OFFSET 12
-+
-+#define WBFS_GRP_A 1
-+#define WBFS_GRP_B 2
-+#define WBFS_GRP_LARGE 3
-+
-+enum {
-+ TCA_CEETM_UNSPEC,
-+ TCA_CEETM_COPT,
-+ TCA_CEETM_QOPS,
-+ __TCA_CEETM_MAX,
-+};
-+
-+/* CEETM configuration types */
-+enum {
-+ CEETM_ROOT = 1,
-+ CEETM_PRIO,
-+ CEETM_WBFS
-+};
-+
-+#define TCA_CEETM_MAX (__TCA_CEETM_MAX - 1)
-+extern const struct nla_policy ceetm_policy[TCA_CEETM_MAX + 1];
-+
-+struct ceetm_class;
-+struct ceetm_qdisc_stats;
-+struct ceetm_class_stats;
-+
-+struct ceetm_fq {
-+ struct qman_fq fq;
-+ struct net_device *net_dev;
-+ struct ceetm_class *ceetm_cls;
-+ int congested; /* Congestion status */
-+};
-+
-+struct root_q {
-+ struct Qdisc **qdiscs;
-+ __u16 overhead;
-+ __u32 rate;
-+ __u32 ceil;
-+ struct qm_ceetm_sp *sp;
-+ struct qm_ceetm_lni *lni;
-+ struct ceetm_qdisc_stats __percpu *qstats;
-+};
-+
-+struct prio_q {
-+ __u16 qcount;
-+ struct ceetm_class *parent;
-+ struct qm_ceetm_channel *ch;
-+};
-+
-+struct wbfs_q {
-+ __u16 qcount;
-+ int group_type;
-+ struct ceetm_class *parent;
-+ struct qm_ceetm_channel *ch;
-+ __u16 cr;
-+ __u16 er;
-+};
-+
-+struct ceetm_qdisc {
-+ int type; /* LNI/CHNL/WBFS */
-+ bool shaped;
-+ union {
-+ struct root_q root;
-+ struct prio_q prio;
-+ struct wbfs_q wbfs;
-+ };
-+ struct Qdisc_class_hash clhash;
-+ struct tcf_proto *filter_list; /* qdisc attached filters */
-+};
-+
-+/* CEETM Qdisc configuration parameters */
-+struct tc_ceetm_qopt {
-+ __u32 type;
-+ __u16 shaped;
-+ __u16 qcount;
-+ __u16 overhead;
-+ __u32 rate;
-+ __u32 ceil;
-+ __u16 cr;
-+ __u16 er;
-+ __u8 qweight[CEETM_MAX_WBFS_QCOUNT];
-+};
-+
-+struct root_c {
-+ unsigned int rate;
-+ unsigned int ceil;
-+ unsigned int tbl;
-+ bool wbfs_grp_a;
-+ bool wbfs_grp_b;
-+ bool wbfs_grp_large;
-+ struct Qdisc *child;
-+};
-+
-+struct prio_c {
-+ bool cr;
-+ bool er;
-+ struct ceetm_fq *fq; /* Hardware FQ instance Handle */
-+ struct qm_ceetm_lfq *lfq;
-+ struct qm_ceetm_cq *cq; /* Hardware Class Queue instance Handle */
-+ struct qm_ceetm_ccg *ccg;
-+ /* only one wbfs can be linked to one priority CQ */
-+ struct Qdisc *child;
-+ struct ceetm_class_stats __percpu *cstats;
-+};
-+
-+struct wbfs_c {
-+ __u8 weight; /* The weight of the class between 1 and 248 */
-+ struct ceetm_fq *fq; /* Hardware FQ instance Handle */
-+ struct qm_ceetm_lfq *lfq;
-+ struct qm_ceetm_cq *cq; /* Hardware Class Queue instance Handle */
-+ struct qm_ceetm_ccg *ccg;
-+ struct ceetm_class_stats __percpu *cstats;
-+};
-+
-+struct ceetm_class {
-+ struct Qdisc_class_common common;
-+ int refcnt; /* usage count of this class */
-+ struct tcf_proto *filter_list; /* class attached filters */
-+ struct Qdisc *parent;
-+ struct qm_ceetm_channel *ch;
-+ bool shaped;
-+ int type; /* ROOT/PRIO/WBFS */
-+ union {
-+ struct root_c root;
-+ struct prio_c prio;
-+ struct wbfs_c wbfs;
-+ };
-+};
-+
-+/* CEETM Class configuration parameters */
-+struct tc_ceetm_copt {
-+ __u32 type;
-+ __u16 shaped;
-+ __u32 rate;
-+ __u32 ceil;
-+ __u16 tbl;
-+ __u16 cr;
-+ __u16 er;
-+ __u8 weight;
-+};
-+
-+/* CEETM stats */
-+struct ceetm_qdisc_stats {
-+ __u32 drops;
-+};
-+
-+struct ceetm_class_stats {
-+ /* Software counters */
-+ struct gnet_stats_basic_packed bstats;
-+ __u32 ern_drop_count;
-+ __u32 congested_count;
-+};
-+
-+struct tc_ceetm_xstats {
-+ __u32 ern_drop_count;
-+ __u32 congested_count;
-+ /* Hardware counters */
-+ __u64 frame_count;
-+ __u64 byte_count;
-+};
-+
-+int __hot ceetm_tx(struct sk_buff *skb, struct net_device *net_dev);
-+#endif
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_dpaa/dpaa_eth_common.c
-@@ -0,0 +1,1802 @@
-+/* Copyright 2008-2013 Freescale Semiconductor, Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#include <linux/init.h>
-+#include <linux/module.h>
-+#include <linux/of_platform.h>
-+#include <linux/of_net.h>
-+#include <linux/etherdevice.h>
-+#include <linux/kthread.h>
-+#include <linux/percpu.h>
-+#include <linux/highmem.h>
-+#include <linux/sort.h>
-+#include <linux/fsl_qman.h>
-+#include <linux/ip.h>
-+#include <linux/ipv6.h>
-+#include <linux/if_vlan.h> /* vlan_eth_hdr */
-+#include "dpaa_eth.h"
-+#include "dpaa_eth_common.h"
-+#ifdef CONFIG_FSL_DPAA_1588
-+#include "dpaa_1588.h"
-+#endif
-+#ifdef CONFIG_FSL_DPAA_DBG_LOOP
-+#include "dpaa_debugfs.h"
-+#endif /* CONFIG_FSL_DPAA_DBG_LOOP */
-+#include "mac.h"
-+
-+/* Size in bytes of the FQ taildrop threshold */
-+#define DPA_FQ_TD 0x200000
-+
-+#ifdef CONFIG_PTP_1588_CLOCK_DPAA
-+struct ptp_priv_s ptp_priv;
-+#endif
-+
-+static struct dpa_bp *dpa_bp_array[64];
-+
-+int dpa_max_frm;
-+EXPORT_SYMBOL(dpa_max_frm);
-+
-+int dpa_rx_extra_headroom;
-+EXPORT_SYMBOL(dpa_rx_extra_headroom);
-+
-+int dpa_num_cpus = NR_CPUS;
-+
-+static const struct fqid_cell tx_confirm_fqids[] = {
-+ {0, DPAA_ETH_TX_QUEUES}
-+};
-+
-+static struct fqid_cell default_fqids[][3] = {
-+ [RX] = { {0, 1}, {0, 1}, {0, DPAA_ETH_RX_QUEUES} },
-+ [TX] = { {0, 1}, {0, 1}, {0, DPAA_ETH_TX_QUEUES} }
-+};
-+
-+static const char fsl_qman_frame_queues[][25] = {
-+ [RX] = "fsl,qman-frame-queues-rx",
-+ [TX] = "fsl,qman-frame-queues-tx"
-+};
-+#ifdef CONFIG_FSL_DPAA_HOOKS
-+/* A set of callbacks for hooking into the fastpath at different points. */
-+struct dpaa_eth_hooks_s dpaa_eth_hooks;
-+EXPORT_SYMBOL(dpaa_eth_hooks);
-+/* This function should only be called on the probe paths, since it makes no
-+ * effort to guarantee consistency of the destination hooks structure.
-+ */
-+void fsl_dpaa_eth_set_hooks(struct dpaa_eth_hooks_s *hooks)
-+{
-+ if (hooks)
-+ dpaa_eth_hooks = *hooks;
-+ else
-+ pr_err("NULL pointer to hooks!\n");
-+}
-+EXPORT_SYMBOL(fsl_dpaa_eth_set_hooks);
-+#endif
-+
-+int dpa_netdev_init(struct net_device *net_dev,
-+ const uint8_t *mac_addr,
-+ uint16_t tx_timeout)
-+{
-+ int err;
-+ struct dpa_priv_s *priv = netdev_priv(net_dev);
-+ struct device *dev = net_dev->dev.parent;
-+
-+ net_dev->priv_flags |= IFF_LIVE_ADDR_CHANGE;
-+
-+ net_dev->features |= net_dev->hw_features;
-+ net_dev->vlan_features = net_dev->features;
-+
-+ memcpy(net_dev->perm_addr, mac_addr, net_dev->addr_len);
-+ memcpy(net_dev->dev_addr, mac_addr, net_dev->addr_len);
-+
-+ net_dev->ethtool_ops = &dpa_ethtool_ops;
-+
-+ net_dev->needed_headroom = priv->tx_headroom;
-+ net_dev->watchdog_timeo = msecs_to_jiffies(tx_timeout);
-+
-+ err = register_netdev(net_dev);
-+ if (err < 0) {
-+ dev_err(dev, "register_netdev() = %d\n", err);
-+ return err;
-+ }
-+
-+#ifdef CONFIG_FSL_DPAA_DBG_LOOP
-+ /* create debugfs entry for this net_device */
-+ err = dpa_netdev_debugfs_create(net_dev);
-+ if (err) {
-+ unregister_netdev(net_dev);
-+ return err;
-+ }
-+#endif /* CONFIG_FSL_DPAA_DBG_LOOP */
-+
-+ return 0;
-+}
-+EXPORT_SYMBOL(dpa_netdev_init);
-+
-+int __cold dpa_start(struct net_device *net_dev)
-+{
-+ int err, i;
-+ struct dpa_priv_s *priv;
-+ struct mac_device *mac_dev;
-+
-+ priv = netdev_priv(net_dev);
-+ mac_dev = priv->mac_dev;
-+
-+ err = mac_dev->init_phy(net_dev, priv->mac_dev);
-+ if (err < 0) {
-+ if (netif_msg_ifup(priv))
-+ netdev_err(net_dev, "init_phy() = %d\n", err);
-+ return err;
-+ }
-+
-+ for_each_port_device(i, mac_dev->port_dev) {
-+ err = fm_port_enable(mac_dev->port_dev[i]);
-+ if (err)
-+ goto mac_start_failed;
-+ }
-+
-+ err = priv->mac_dev->start(mac_dev);
-+ if (err < 0) {
-+ if (netif_msg_ifup(priv))
-+ netdev_err(net_dev, "mac_dev->start() = %d\n", err);
-+ goto mac_start_failed;
-+ }
-+
-+ netif_tx_start_all_queues(net_dev);
-+
-+ return 0;
-+
-+mac_start_failed:
-+ for_each_port_device(i, mac_dev->port_dev)
-+ fm_port_disable(mac_dev->port_dev[i]);
-+
-+ return err;
-+}
-+EXPORT_SYMBOL(dpa_start);
-+
-+int __cold dpa_stop(struct net_device *net_dev)
-+{
-+ int _errno, i, err;
-+ struct dpa_priv_s *priv;
-+ struct mac_device *mac_dev;
-+
-+ priv = netdev_priv(net_dev);
-+ mac_dev = priv->mac_dev;
-+
-+ netif_tx_stop_all_queues(net_dev);
-+ /* Allow the Fman (Tx) port to process in-flight frames before we
-+ * try switching it off.
-+ */
-+ usleep_range(5000, 10000);
-+
-+ _errno = mac_dev->stop(mac_dev);
-+ if (unlikely(_errno < 0))
-+ if (netif_msg_ifdown(priv))
-+ netdev_err(net_dev, "mac_dev->stop() = %d\n",
-+ _errno);
-+
-+ for_each_port_device(i, mac_dev->port_dev) {
-+ err = fm_port_disable(mac_dev->port_dev[i]);
-+ _errno = err ? err : _errno;
-+ }
-+
-+ if (mac_dev->phy_dev)
-+ phy_disconnect(mac_dev->phy_dev);
-+ mac_dev->phy_dev = NULL;
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(dpa_stop);
-+
-+void __cold dpa_timeout(struct net_device *net_dev)
-+{
-+ const struct dpa_priv_s *priv;
-+ struct dpa_percpu_priv_s *percpu_priv;
-+
-+ priv = netdev_priv(net_dev);
-+ percpu_priv = raw_cpu_ptr(priv->percpu_priv);
-+
-+ if (netif_msg_timer(priv))
-+ netdev_crit(net_dev, "Transmit timeout!\n");
-+
-+ percpu_priv->stats.tx_errors++;
-+}
-+EXPORT_SYMBOL(dpa_timeout);
-+
-+/* net_device */
-+
-+/**
-+ * @param net_dev the device for which statistics are calculated
-+ * @param stats the function fills this structure with the device's statistics
-+ * @return the address of the structure containing the statistics
-+ *
-+ * Calculates the statistics for the given device by adding the statistics
-+ * collected by each CPU.
-+ */
-+struct rtnl_link_stats64 __cold
-+*dpa_get_stats64(struct net_device *net_dev,
-+ struct rtnl_link_stats64 *stats)
-+{
-+ struct dpa_priv_s *priv = netdev_priv(net_dev);
-+ u64 *cpustats;
-+ u64 *netstats = (u64 *)stats;
-+ int i, j;
-+ struct dpa_percpu_priv_s *percpu_priv;
-+ int numstats = sizeof(struct rtnl_link_stats64) / sizeof(u64);
-+
-+ for_each_possible_cpu(i) {
-+ percpu_priv = per_cpu_ptr(priv->percpu_priv, i);
-+
-+ cpustats = (u64 *)&percpu_priv->stats;
-+
-+ for (j = 0; j < numstats; j++)
-+ netstats[j] += cpustats[j];
-+ }
-+ return stats;
-+}
-+EXPORT_SYMBOL(dpa_get_stats64);
-+
-+int dpa_change_mtu(struct net_device *net_dev, int new_mtu)
-+{
-+ int max_mtu = dpa_get_max_mtu();
-+
-+#ifndef CONFIG_PPC
-+ /* Due to the A010022 FMan errata, we can not use contig frames larger
-+ * than 4K, nor S/G frames. We need to prevent the user from setting a
-+ * large MTU.
-+ */
-+ if (unlikely(dpaa_errata_a010022))
-+ max_mtu = DPA_BP_RAW_SIZE;
-+#endif
-+
-+ /* Make sure we don't exceed the Ethernet controller's MAXFRM */
-+ if (new_mtu < 68 || new_mtu > max_mtu) {
-+ netdev_err(net_dev, "Invalid L3 mtu %d (must be between %d and %d).\n",
-+ new_mtu, 68, max_mtu);
-+ return -EINVAL;
-+ }
-+ net_dev->mtu = new_mtu;
-+
-+ return 0;
-+}
-+EXPORT_SYMBOL(dpa_change_mtu);
-+
-+/* .ndo_init callback */
-+int dpa_ndo_init(struct net_device *net_dev)
-+{
-+ /* If fsl_fm_max_frm is set to a higher value than the all-common 1500,
-+ * we choose conservatively and let the user explicitly set a higher
-+ * MTU via ifconfig. Otherwise, the user may end up with different MTUs
-+ * in the same LAN.
-+ * If on the other hand fsl_fm_max_frm has been chosen below 1500,
-+ * start with the maximum allowed.
-+ */
-+ int init_mtu = min(dpa_get_max_mtu(), ETH_DATA_LEN);
-+
-+ pr_debug("Setting initial MTU on net device: %d\n", init_mtu);
-+ net_dev->mtu = init_mtu;
-+
-+ return 0;
-+}
-+EXPORT_SYMBOL(dpa_ndo_init);
-+
-+int dpa_set_features(struct net_device *dev, netdev_features_t features)
-+{
-+ /* Not much to do here for now */
-+ dev->features = features;
-+ return 0;
-+}
-+EXPORT_SYMBOL(dpa_set_features);
-+
-+netdev_features_t dpa_fix_features(struct net_device *dev,
-+ netdev_features_t features)
-+{
-+ netdev_features_t unsupported_features = 0;
-+
-+ /* In theory we should never be requested to enable features that
-+ * we didn't set in netdev->features and netdev->hw_features at probe
-+ * time, but double check just to be on the safe side.
-+ * We don't support enabling Rx csum through ethtool yet
-+ */
-+ unsupported_features |= NETIF_F_RXCSUM;
-+
-+ features &= ~unsupported_features;
-+
-+ return features;
-+}
-+EXPORT_SYMBOL(dpa_fix_features);
-+
-+#ifdef CONFIG_FSL_DPAA_TS
-+u64 dpa_get_timestamp_ns(const struct dpa_priv_s *priv, enum port_type rx_tx,
-+ const void *data)
-+{
-+ u64 *ts, ns;
-+
-+ ts = fm_port_get_buffer_time_stamp(priv->mac_dev->port_dev[rx_tx],
-+ data);
-+
-+ if (!ts || *ts == 0)
-+ return 0;
-+
-+ be64_to_cpus(ts);
-+
-+ /* multiple DPA_PTP_NOMINAL_FREQ_PERIOD_NS for case of non power of 2 */
-+ ns = *ts << DPA_PTP_NOMINAL_FREQ_PERIOD_SHIFT;
-+
-+ return ns;
-+}
-+
-+int dpa_get_ts(const struct dpa_priv_s *priv, enum port_type rx_tx,
-+ struct skb_shared_hwtstamps *shhwtstamps, const void *data)
-+{
-+ u64 ns;
-+
-+ ns = dpa_get_timestamp_ns(priv, rx_tx, data);
-+
-+ if (ns == 0)
-+ return -EINVAL;
-+
-+ memset(shhwtstamps, 0, sizeof(*shhwtstamps));
-+ shhwtstamps->hwtstamp = ns_to_ktime(ns);
-+
-+ return 0;
-+}
-+
-+static void dpa_ts_tx_enable(struct net_device *dev)
-+{
-+ struct dpa_priv_s *priv = netdev_priv(dev);
-+ struct mac_device *mac_dev = priv->mac_dev;
-+
-+ if (mac_dev->fm_rtc_enable)
-+ mac_dev->fm_rtc_enable(get_fm_handle(dev));
-+ if (mac_dev->ptp_enable)
-+ mac_dev->ptp_enable(mac_dev->get_mac_handle(mac_dev));
-+
-+ priv->ts_tx_en = true;
-+}
-+
-+static void dpa_ts_tx_disable(struct net_device *dev)
-+{
-+ struct dpa_priv_s *priv = netdev_priv(dev);
-+
-+#if 0
-+/* the RTC might be needed by the Rx Ts, cannot disable here
-+ * no separate ptp_disable API for Rx/Tx, cannot disable here
-+ */
-+ struct mac_device *mac_dev = priv->mac_dev;
-+
-+ if (mac_dev->fm_rtc_disable)
-+ mac_dev->fm_rtc_disable(get_fm_handle(dev));
-+
-+ if (mac_dev->ptp_disable)
-+ mac_dev->ptp_disable(mac_dev->get_mac_handle(mac_dev));
-+#endif
-+
-+ priv->ts_tx_en = false;
-+}
-+
-+static void dpa_ts_rx_enable(struct net_device *dev)
-+{
-+ struct dpa_priv_s *priv = netdev_priv(dev);
-+ struct mac_device *mac_dev = priv->mac_dev;
-+
-+ if (mac_dev->fm_rtc_enable)
-+ mac_dev->fm_rtc_enable(get_fm_handle(dev));
-+ if (mac_dev->ptp_enable)
-+ mac_dev->ptp_enable(mac_dev->get_mac_handle(mac_dev));
-+
-+ priv->ts_rx_en = true;
-+}
-+
-+static void dpa_ts_rx_disable(struct net_device *dev)
-+{
-+ struct dpa_priv_s *priv = netdev_priv(dev);
-+
-+#if 0
-+/* the RTC might be needed by the Tx Ts, cannot disable here
-+ * no separate ptp_disable API for Rx/Tx, cannot disable here
-+ */
-+ struct mac_device *mac_dev = priv->mac_dev;
-+
-+ if (mac_dev->fm_rtc_disable)
-+ mac_dev->fm_rtc_disable(get_fm_handle(dev));
-+
-+ if (mac_dev->ptp_disable)
-+ mac_dev->ptp_disable(mac_dev->get_mac_handle(mac_dev));
-+#endif
-+
-+ priv->ts_rx_en = false;
-+}
-+
-+static int dpa_ts_ioctl(struct net_device *dev, struct ifreq *rq, int cmd)
-+{
-+ struct hwtstamp_config config;
-+
-+ if (copy_from_user(&config, rq->ifr_data, sizeof(config)))
-+ return -EFAULT;
-+
-+ switch (config.tx_type) {
-+ case HWTSTAMP_TX_OFF:
-+ dpa_ts_tx_disable(dev);
-+ break;
-+ case HWTSTAMP_TX_ON:
-+ dpa_ts_tx_enable(dev);
-+ break;
-+ default:
-+ return -ERANGE;
-+ }
-+
-+ if (config.rx_filter == HWTSTAMP_FILTER_NONE)
-+ dpa_ts_rx_disable(dev);
-+ else {
-+ dpa_ts_rx_enable(dev);
-+ /* TS is set for all frame types, not only those requested */
-+ config.rx_filter = HWTSTAMP_FILTER_ALL;
-+ }
-+
-+ return copy_to_user(rq->ifr_data, &config, sizeof(config)) ?
-+ -EFAULT : 0;
-+}
-+#endif /* CONFIG_FSL_DPAA_TS */
-+
-+int dpa_ioctl(struct net_device *dev, struct ifreq *rq, int cmd)
-+{
-+#ifdef CONFIG_FSL_DPAA_1588
-+ struct dpa_priv_s *priv = netdev_priv(dev);
-+#endif
-+ int ret = -EINVAL;
-+
-+ if (!netif_running(dev))
-+ return -EINVAL;
-+
-+ if (cmd == SIOCGMIIREG) {
-+ if (!dev->phydev)
-+ ret = -EINVAL;
-+ else
-+ ret = phy_mii_ioctl(dev->phydev, rq, cmd);
-+ }
-+
-+#ifdef CONFIG_FSL_DPAA_TS
-+ if (cmd == SIOCSHWTSTAMP)
-+ return dpa_ts_ioctl(dev, rq, cmd);
-+#endif /* CONFIG_FSL_DPAA_TS */
-+
-+#ifdef CONFIG_FSL_DPAA_1588
-+ if ((cmd >= PTP_ENBL_TXTS_IOCTL) && (cmd <= PTP_CLEANUP_TS)) {
-+ if (priv->tsu && priv->tsu->valid)
-+ ret = dpa_ioctl_1588(dev, rq, cmd);
-+ else
-+ ret = -ENODEV;
-+ }
-+#endif
-+
-+ return ret;
-+}
-+EXPORT_SYMBOL(dpa_ioctl);
-+
-+int __cold dpa_remove(struct platform_device *of_dev)
-+{
-+ int err;
-+ struct device *dev;
-+ struct net_device *net_dev;
-+ struct dpa_priv_s *priv;
-+
-+ dev = &of_dev->dev;
-+ net_dev = dev_get_drvdata(dev);
-+
-+ priv = netdev_priv(net_dev);
-+
-+ dpaa_eth_sysfs_remove(dev);
-+
-+ dev_set_drvdata(dev, NULL);
-+ unregister_netdev(net_dev);
-+
-+ err = dpa_fq_free(dev, &priv->dpa_fq_list);
-+
-+ qman_delete_cgr_safe(&priv->ingress_cgr);
-+ qman_release_cgrid(priv->ingress_cgr.cgrid);
-+ qman_delete_cgr_safe(&priv->cgr_data.cgr);
-+ qman_release_cgrid(priv->cgr_data.cgr.cgrid);
-+
-+ dpa_private_napi_del(net_dev);
-+
-+ dpa_bp_free(priv);
-+
-+ if (priv->buf_layout)
-+ devm_kfree(dev, priv->buf_layout);
-+
-+#ifdef CONFIG_FSL_DPAA_DBG_LOOP
-+ /* remove debugfs entry for this net_device */
-+ dpa_netdev_debugfs_remove(net_dev);
-+#endif /* CONFIG_FSL_DPAA_DBG_LOOP */
-+
-+#ifdef CONFIG_FSL_DPAA_1588
-+ if (priv->tsu && priv->tsu->valid)
-+ dpa_ptp_cleanup(priv);
-+#endif
-+
-+ free_netdev(net_dev);
-+
-+ return err;
-+}
-+EXPORT_SYMBOL(dpa_remove);
-+
-+struct mac_device * __cold __must_check
-+__attribute__((nonnull))
-+dpa_mac_probe(struct platform_device *_of_dev)
-+{
-+ struct device *dpa_dev, *dev;
-+ struct device_node *mac_node;
-+ struct platform_device *of_dev;
-+ struct mac_device *mac_dev;
-+#ifdef CONFIG_FSL_DPAA_1588
-+ int lenp;
-+ const phandle *phandle_prop;
-+ struct net_device *net_dev = NULL;
-+ struct dpa_priv_s *priv = NULL;
-+ struct device_node *timer_node;
-+#endif
-+ dpa_dev = &_of_dev->dev;
-+
-+ mac_node = of_parse_phandle(_of_dev->dev.of_node, "fsl,fman-mac", 0);
-+ if (unlikely(mac_node == NULL)) {
-+ dev_err(dpa_dev, "Cannot find MAC device device tree node\n");
-+ return ERR_PTR(-EFAULT);
-+ }
-+
-+ of_dev = of_find_device_by_node(mac_node);
-+ if (unlikely(of_dev == NULL)) {
-+ dev_err(dpa_dev, "of_find_device_by_node(%s) failed\n",
-+ mac_node->full_name);
-+ of_node_put(mac_node);
-+ return ERR_PTR(-EINVAL);
-+ }
-+ of_node_put(mac_node);
-+
-+ dev = &of_dev->dev;
-+
-+ mac_dev = dev_get_drvdata(dev);
-+ if (unlikely(mac_dev == NULL)) {
-+ dev_err(dpa_dev, "dev_get_drvdata(%s) failed\n",
-+ dev_name(dev));
-+ return ERR_PTR(-EINVAL);
-+ }
-+
-+#ifdef CONFIG_FSL_DPAA_1588
-+ phandle_prop = of_get_property(mac_node, "ptp-timer", &lenp);
-+ if (phandle_prop && ((mac_dev->phy_if != PHY_INTERFACE_MODE_SGMII) ||
-+ ((mac_dev->phy_if == PHY_INTERFACE_MODE_SGMII) &&
-+ (mac_dev->speed == SPEED_1000)))) {
-+ timer_node = of_find_node_by_phandle(*phandle_prop);
-+ if (timer_node)
-+ net_dev = dev_get_drvdata(dpa_dev);
-+ if (timer_node && net_dev) {
-+ priv = netdev_priv(net_dev);
-+ if (!dpa_ptp_init(priv))
-+ dev_info(dev, "%s: ptp 1588 is initialized.\n",
-+ mac_node->full_name);
-+ }
-+ }
-+#endif
-+
-+#ifdef CONFIG_PTP_1588_CLOCK_DPAA
-+ if ((mac_dev->phy_if != PHY_INTERFACE_MODE_SGMII) ||
-+ ((mac_dev->phy_if == PHY_INTERFACE_MODE_SGMII) &&
-+ (mac_dev->speed == SPEED_1000))) {
-+ ptp_priv.node = of_parse_phandle(mac_node, "ptp-timer", 0);
-+ if (ptp_priv.node) {
-+ ptp_priv.of_dev = of_find_device_by_node(ptp_priv.node);
-+ if (unlikely(ptp_priv.of_dev == NULL)) {
-+ dev_err(dpa_dev,
-+ "Cannot find device represented by timer_node\n");
-+ of_node_put(ptp_priv.node);
-+ return ERR_PTR(-EINVAL);
-+ }
-+ ptp_priv.mac_dev = mac_dev;
-+ }
-+ }
-+#endif
-+ return mac_dev;
-+}
-+EXPORT_SYMBOL(dpa_mac_probe);
-+
-+int dpa_set_mac_address(struct net_device *net_dev, void *addr)
-+{
-+ const struct dpa_priv_s *priv;
-+ int _errno;
-+ struct mac_device *mac_dev;
-+
-+ priv = netdev_priv(net_dev);
-+
-+ _errno = eth_mac_addr(net_dev, addr);
-+ if (_errno < 0) {
-+ if (netif_msg_drv(priv))
-+ netdev_err(net_dev,
-+ "eth_mac_addr() = %d\n",
-+ _errno);
-+ return _errno;
-+ }
-+
-+ mac_dev = priv->mac_dev;
-+
-+ _errno = mac_dev->change_addr(mac_dev->get_mac_handle(mac_dev),
-+ net_dev->dev_addr);
-+ if (_errno < 0) {
-+ if (netif_msg_drv(priv))
-+ netdev_err(net_dev,
-+ "mac_dev->change_addr() = %d\n",
-+ _errno);
-+ return _errno;
-+ }
-+
-+ return 0;
-+}
-+EXPORT_SYMBOL(dpa_set_mac_address);
-+
-+void dpa_set_rx_mode(struct net_device *net_dev)
-+{
-+ int _errno;
-+ const struct dpa_priv_s *priv;
-+
-+ priv = netdev_priv(net_dev);
-+
-+ if (!!(net_dev->flags & IFF_PROMISC) != priv->mac_dev->promisc) {
-+ priv->mac_dev->promisc = !priv->mac_dev->promisc;
-+ _errno = priv->mac_dev->set_promisc(
-+ priv->mac_dev->get_mac_handle(priv->mac_dev),
-+ priv->mac_dev->promisc);
-+ if (unlikely(_errno < 0) && netif_msg_drv(priv))
-+ netdev_err(net_dev,
-+ "mac_dev->set_promisc() = %d\n",
-+ _errno);
-+ }
-+
-+ _errno = priv->mac_dev->set_multi(net_dev, priv->mac_dev);
-+ if (unlikely(_errno < 0) && netif_msg_drv(priv))
-+ netdev_err(net_dev, "mac_dev->set_multi() = %d\n", _errno);
-+}
-+EXPORT_SYMBOL(dpa_set_rx_mode);
-+
-+void dpa_set_buffers_layout(struct mac_device *mac_dev,
-+ struct dpa_buffer_layout_s *layout)
-+{
-+ struct fm_port_params params;
-+
-+ /* Rx */
-+ layout[RX].priv_data_size = (uint16_t)DPA_RX_PRIV_DATA_SIZE;
-+ layout[RX].parse_results = true;
-+ layout[RX].hash_results = true;
-+#ifdef CONFIG_FSL_DPAA_TS
-+ layout[RX].time_stamp = true;
-+#endif
-+ fm_port_get_buff_layout_ext_params(mac_dev->port_dev[RX], &params);
-+ layout[RX].manip_extra_space = params.manip_extra_space;
-+ /* a value of zero for data alignment means "don't care", so align to
-+ * a non-zero value to prevent FMD from using its own default
-+ */
-+ layout[RX].data_align = params.data_align ? : DPA_FD_DATA_ALIGNMENT;
-+
-+ /* Tx */
-+ layout[TX].priv_data_size = DPA_TX_PRIV_DATA_SIZE;
-+ layout[TX].parse_results = true;
-+ layout[TX].hash_results = true;
-+#ifdef CONFIG_FSL_DPAA_TS
-+ layout[TX].time_stamp = true;
-+#endif
-+ fm_port_get_buff_layout_ext_params(mac_dev->port_dev[TX], &params);
-+ layout[TX].manip_extra_space = params.manip_extra_space;
-+ layout[TX].data_align = params.data_align ? : DPA_FD_DATA_ALIGNMENT;
-+}
-+EXPORT_SYMBOL(dpa_set_buffers_layout);
-+
-+int __attribute__((nonnull))
-+dpa_bp_alloc(struct dpa_bp *dpa_bp, struct device *dev)
-+{
-+ int err;
-+ struct bman_pool_params bp_params;
-+
-+ if (dpa_bp->size == 0 || dpa_bp->config_count == 0) {
-+ pr_err("Buffer pool is not properly initialized! Missing size or initial number of buffers");
-+ return -EINVAL;
-+ }
-+
-+ memset(&bp_params, 0, sizeof(struct bman_pool_params));
-+#ifdef CONFIG_FMAN_PFC
-+ bp_params.flags = BMAN_POOL_FLAG_THRESH;
-+ bp_params.thresholds[0] = bp_params.thresholds[2] =
-+ CONFIG_FSL_DPAA_ETH_REFILL_THRESHOLD;
-+ bp_params.thresholds[1] = bp_params.thresholds[3] =
-+ CONFIG_FSL_DPAA_ETH_MAX_BUF_COUNT;
-+#endif
-+
-+ /* If the pool is already specified, we only create one per bpid */
-+ if (dpa_bpid2pool_use(dpa_bp->bpid))
-+ return 0;
-+
-+ if (dpa_bp->bpid == 0)
-+ bp_params.flags |= BMAN_POOL_FLAG_DYNAMIC_BPID;
-+ else
-+ bp_params.bpid = dpa_bp->bpid;
-+
-+ dpa_bp->pool = bman_new_pool(&bp_params);
-+ if (unlikely(dpa_bp->pool == NULL)) {
-+ pr_err("bman_new_pool() failed\n");
-+ return -ENODEV;
-+ }
-+
-+ dpa_bp->bpid = (uint8_t)bman_get_params(dpa_bp->pool)->bpid;
-+
-+ err = dma_coerce_mask_and_coherent(dev, DMA_BIT_MASK(40));
-+ if (err) {
-+ pr_err("dma_coerce_mask_and_coherent() failed\n");
-+ goto bman_free_pool;
-+ }
-+
-+ dpa_bp->dev = dev;
-+
-+ if (dpa_bp->seed_cb) {
-+ err = dpa_bp->seed_cb(dpa_bp);
-+ if (err)
-+ goto bman_free_pool;
-+ }
-+
-+ dpa_bpid2pool_map(dpa_bp->bpid, dpa_bp);
-+
-+ return 0;
-+
-+bman_free_pool:
-+ bman_free_pool(dpa_bp->pool);
-+
-+ return err;
-+}
-+EXPORT_SYMBOL(dpa_bp_alloc);
-+
-+void dpa_bp_drain(struct dpa_bp *bp)
-+{
-+ int ret, num = 8;
-+
-+ do {
-+ struct bm_buffer bmb[8];
-+ int i;
-+
-+ ret = bman_acquire(bp->pool, bmb, num, 0);
-+ if (ret < 0) {
-+ if (num == 8) {
-+ /* we have less than 8 buffers left;
-+ * drain them one by one
-+ */
-+ num = 1;
-+ ret = 1;
-+ continue;
-+ } else {
-+ /* Pool is fully drained */
-+ break;
-+ }
-+ }
-+
-+ for (i = 0; i < num; i++) {
-+ dma_addr_t addr = bm_buf_addr(&bmb[i]);
-+
-+ dma_unmap_single(bp->dev, addr, bp->size,
-+ DMA_BIDIRECTIONAL);
-+
-+ bp->free_buf_cb(phys_to_virt(addr));
-+ }
-+ } while (ret > 0);
-+}
-+EXPORT_SYMBOL(dpa_bp_drain);
-+
-+static void __cold __attribute__((nonnull))
-+_dpa_bp_free(struct dpa_bp *dpa_bp)
-+{
-+ struct dpa_bp *bp = dpa_bpid2pool(dpa_bp->bpid);
-+
-+ /* the mapping between bpid and dpa_bp is done very late in the
-+ * allocation procedure; if something failed before the mapping, the bp
-+ * was not configured, therefore we don't need the below instructions
-+ */
-+ if (!bp)
-+ return;
-+
-+ if (!atomic_dec_and_test(&bp->refs))
-+ return;
-+
-+ if (bp->free_buf_cb)
-+ dpa_bp_drain(bp);
-+
-+ dpa_bp_array[bp->bpid] = NULL;
-+ bman_free_pool(bp->pool);
-+}
-+
-+void __cold __attribute__((nonnull))
-+dpa_bp_free(struct dpa_priv_s *priv)
-+{
-+ int i;
-+
-+ if (priv->dpa_bp)
-+ for (i = 0; i < priv->bp_count; i++)
-+ _dpa_bp_free(&priv->dpa_bp[i]);
-+}
-+EXPORT_SYMBOL(dpa_bp_free);
-+
-+struct dpa_bp *dpa_bpid2pool(int bpid)
-+{
-+ return dpa_bp_array[bpid];
-+}
-+EXPORT_SYMBOL(dpa_bpid2pool);
-+
-+void dpa_bpid2pool_map(int bpid, struct dpa_bp *dpa_bp)
-+{
-+ dpa_bp_array[bpid] = dpa_bp;
-+ atomic_set(&dpa_bp->refs, 1);
-+}
-+
-+bool dpa_bpid2pool_use(int bpid)
-+{
-+ if (dpa_bpid2pool(bpid)) {
-+ atomic_inc(&dpa_bp_array[bpid]->refs);
-+ return true;
-+ }
-+
-+ return false;
-+}
-+
-+#ifdef CONFIG_FSL_DPAA_ETH_USE_NDO_SELECT_QUEUE
-+u16 dpa_select_queue(struct net_device *net_dev, struct sk_buff *skb,
-+ void *accel_priv, select_queue_fallback_t fallback)
-+{
-+ return dpa_get_queue_mapping(skb);
-+}
-+EXPORT_SYMBOL(dpa_select_queue);
-+#endif
-+
-+struct dpa_fq *dpa_fq_alloc(struct device *dev,
-+ u32 fq_start,
-+ u32 fq_count,
-+ struct list_head *list,
-+ enum dpa_fq_type fq_type)
-+{
-+ int i;
-+ struct dpa_fq *dpa_fq;
-+
-+ dpa_fq = devm_kzalloc(dev, sizeof(*dpa_fq) * fq_count, GFP_KERNEL);
-+ if (dpa_fq == NULL)
-+ return NULL;
-+
-+ for (i = 0; i < fq_count; i++) {
-+ dpa_fq[i].fq_type = fq_type;
-+ if (fq_type == FQ_TYPE_RX_PCD_HI_PRIO)
-+ dpa_fq[i].fqid = fq_start ?
-+ DPAA_ETH_FQ_DELTA + fq_start + i : 0;
-+ else
-+ dpa_fq[i].fqid = fq_start ? fq_start + i : 0;
-+
-+ list_add_tail(&dpa_fq[i].list, list);
-+ }
-+
-+#ifdef CONFIG_FMAN_PFC
-+ if (fq_type == FQ_TYPE_TX)
-+ for (i = 0; i < fq_count; i++)
-+ dpa_fq[i].wq = i / dpa_num_cpus;
-+ else
-+#endif
-+ for (i = 0; i < fq_count; i++)
-+ _dpa_assign_wq(dpa_fq + i);
-+
-+ return dpa_fq;
-+}
-+EXPORT_SYMBOL(dpa_fq_alloc);
-+
-+/* Probing of FQs for MACful ports */
-+int dpa_fq_probe_mac(struct device *dev, struct list_head *list,
-+ struct fm_port_fqs *port_fqs,
-+ bool alloc_tx_conf_fqs,
-+ enum port_type ptype)
-+{
-+ struct fqid_cell *fqids = NULL;
-+ const void *fqids_off = NULL;
-+ struct dpa_fq *dpa_fq = NULL;
-+ struct device_node *np = dev->of_node;
-+ int num_ranges;
-+ int i, lenp;
-+
-+ if (ptype == TX && alloc_tx_conf_fqs) {
-+ if (!dpa_fq_alloc(dev, tx_confirm_fqids->start,
-+ tx_confirm_fqids->count, list,
-+ FQ_TYPE_TX_CONF_MQ))
-+ goto fq_alloc_failed;
-+ }
-+
-+ fqids_off = of_get_property(np, fsl_qman_frame_queues[ptype], &lenp);
-+ if (fqids_off == NULL) {
-+ /* No dts definition, so use the defaults. */
-+ fqids = default_fqids[ptype];
-+ num_ranges = 3;
-+ } else {
-+ num_ranges = lenp / sizeof(*fqids);
-+
-+ fqids = devm_kzalloc(dev, sizeof(*fqids) * num_ranges,
-+ GFP_KERNEL);
-+ if (fqids == NULL)
-+ goto fqids_alloc_failed;
-+
-+ /* convert to CPU endianess */
-+ for (i = 0; i < num_ranges; i++) {
-+ fqids[i].start = be32_to_cpup(fqids_off +
-+ i * sizeof(*fqids));
-+ fqids[i].count = be32_to_cpup(fqids_off +
-+ i * sizeof(*fqids) + sizeof(__be32));
-+ }
-+ }
-+
-+ for (i = 0; i < num_ranges; i++) {
-+ switch (i) {
-+ case 0:
-+ /* The first queue is the error queue */
-+ if (fqids[i].count != 1)
-+ goto invalid_error_queue;
-+
-+ dpa_fq = dpa_fq_alloc(dev, fqids[i].start,
-+ fqids[i].count, list,
-+ ptype == RX ?
-+ FQ_TYPE_RX_ERROR :
-+ FQ_TYPE_TX_ERROR);
-+ if (dpa_fq == NULL)
-+ goto fq_alloc_failed;
-+
-+ if (ptype == RX)
-+ port_fqs->rx_errq = &dpa_fq[0];
-+ else
-+ port_fqs->tx_errq = &dpa_fq[0];
-+ break;
-+ case 1:
-+ /* the second queue is the default queue */
-+ if (fqids[i].count != 1)
-+ goto invalid_default_queue;
-+
-+ dpa_fq = dpa_fq_alloc(dev, fqids[i].start,
-+ fqids[i].count, list,
-+ ptype == RX ?
-+ FQ_TYPE_RX_DEFAULT :
-+ FQ_TYPE_TX_CONFIRM);
-+ if (dpa_fq == NULL)
-+ goto fq_alloc_failed;
-+
-+ if (ptype == RX)
-+ port_fqs->rx_defq = &dpa_fq[0];
-+ else
-+ port_fqs->tx_defq = &dpa_fq[0];
-+ break;
-+ default:
-+ /* all subsequent queues are either RX* PCD or Tx */
-+ if (ptype == RX) {
-+ if (!dpa_fq_alloc(dev, fqids[i].start,
-+ fqids[i].count, list,
-+ FQ_TYPE_RX_PCD) ||
-+ !dpa_fq_alloc(dev, fqids[i].start,
-+ fqids[i].count, list,
-+ FQ_TYPE_RX_PCD_HI_PRIO))
-+ goto fq_alloc_failed;
-+ } else {
-+ if (!dpa_fq_alloc(dev, fqids[i].start,
-+ fqids[i].count, list,
-+ FQ_TYPE_TX))
-+ goto fq_alloc_failed;
-+ }
-+ break;
-+ }
-+ }
-+
-+ return 0;
-+
-+fq_alloc_failed:
-+fqids_alloc_failed:
-+ dev_err(dev, "Cannot allocate memory for frame queues\n");
-+ return -ENOMEM;
-+
-+invalid_default_queue:
-+invalid_error_queue:
-+ dev_err(dev, "Too many default or error queues\n");
-+ return -EINVAL;
-+}
-+EXPORT_SYMBOL(dpa_fq_probe_mac);
-+
-+static u32 rx_pool_channel;
-+static DEFINE_SPINLOCK(rx_pool_channel_init);
-+
-+int dpa_get_channel(void)
-+{
-+ spin_lock(&rx_pool_channel_init);
-+ if (!rx_pool_channel) {
-+ u32 pool;
-+ int ret = qman_alloc_pool(&pool);
-+ if (!ret)
-+ rx_pool_channel = pool;
-+ }
-+ spin_unlock(&rx_pool_channel_init);
-+ if (!rx_pool_channel)
-+ return -ENOMEM;
-+ return rx_pool_channel;
-+}
-+EXPORT_SYMBOL(dpa_get_channel);
-+
-+void dpa_release_channel(void)
-+{
-+ qman_release_pool(rx_pool_channel);
-+}
-+EXPORT_SYMBOL(dpa_release_channel);
-+
-+void dpaa_eth_add_channel(u16 channel)
-+{
-+ const cpumask_t *cpus = qman_affine_cpus();
-+ u32 pool = QM_SDQCR_CHANNELS_POOL_CONV(channel);
-+ int cpu;
-+ struct qman_portal *portal;
-+
-+ for_each_cpu(cpu, cpus) {
-+ portal = (struct qman_portal *)qman_get_affine_portal(cpu);
-+ qman_p_static_dequeue_add(portal, pool);
-+ }
-+}
-+EXPORT_SYMBOL(dpaa_eth_add_channel);
-+
-+/**
-+ * Congestion group state change notification callback.
-+ * Stops the device's egress queues while they are congested and
-+ * wakes them upon exiting congested state.
-+ * Also updates some CGR-related stats.
-+ */
-+static void dpaa_eth_cgscn(struct qman_portal *qm, struct qman_cgr *cgr,
-+
-+ int congested)
-+{
-+ struct dpa_priv_s *priv = (struct dpa_priv_s *)container_of(cgr,
-+ struct dpa_priv_s, cgr_data.cgr);
-+
-+ if (congested) {
-+ priv->cgr_data.congestion_start_jiffies = jiffies;
-+ netif_tx_stop_all_queues(priv->net_dev);
-+ priv->cgr_data.cgr_congested_count++;
-+ } else {
-+ priv->cgr_data.congested_jiffies +=
-+ (jiffies - priv->cgr_data.congestion_start_jiffies);
-+ netif_tx_wake_all_queues(priv->net_dev);
-+ }
-+}
-+
-+int dpaa_eth_cgr_init(struct dpa_priv_s *priv)
-+{
-+ struct qm_mcc_initcgr initcgr;
-+ u32 cs_th;
-+ int err;
-+
-+ err = qman_alloc_cgrid(&priv->cgr_data.cgr.cgrid);
-+ if (err < 0) {
-+ pr_err("Error %d allocating CGR ID\n", err);
-+ goto out_error;
-+ }
-+ priv->cgr_data.cgr.cb = dpaa_eth_cgscn;
-+
-+ /* Enable Congestion State Change Notifications and CS taildrop */
-+ initcgr.we_mask = QM_CGR_WE_CSCN_EN | QM_CGR_WE_CS_THRES;
-+ initcgr.cgr.cscn_en = QM_CGR_EN;
-+
-+ /* Set different thresholds based on the MAC speed.
-+ * TODO: this may turn suboptimal if the MAC is reconfigured at a speed
-+ * lower than its max, e.g. if a dTSEC later negotiates a 100Mbps link.
-+ * In such cases, we ought to reconfigure the threshold, too.
-+ */
-+ if (priv->mac_dev->if_support & SUPPORTED_10000baseT_Full)
-+ cs_th = CONFIG_FSL_DPAA_CS_THRESHOLD_10G;
-+ else
-+ cs_th = CONFIG_FSL_DPAA_CS_THRESHOLD_1G;
-+ qm_cgr_cs_thres_set64(&initcgr.cgr.cs_thres, cs_th, 1);
-+
-+ initcgr.we_mask |= QM_CGR_WE_CSTD_EN;
-+ initcgr.cgr.cstd_en = QM_CGR_EN;
-+
-+ err = qman_create_cgr(&priv->cgr_data.cgr, QMAN_CGR_FLAG_USE_INIT,
-+ &initcgr);
-+ if (err < 0) {
-+ pr_err("Error %d creating CGR with ID %d\n", err,
-+ priv->cgr_data.cgr.cgrid);
-+ qman_release_cgrid(priv->cgr_data.cgr.cgrid);
-+ goto out_error;
-+ }
-+ pr_debug("Created CGR %d for netdev with hwaddr %pM on QMan channel %d\n",
-+ priv->cgr_data.cgr.cgrid, priv->mac_dev->addr,
-+ priv->cgr_data.cgr.chan);
-+
-+out_error:
-+ return err;
-+}
-+EXPORT_SYMBOL(dpaa_eth_cgr_init);
-+
-+static inline void dpa_setup_ingress(const struct dpa_priv_s *priv,
-+ struct dpa_fq *fq,
-+ const struct qman_fq *template)
-+{
-+ fq->fq_base = *template;
-+ fq->net_dev = priv->net_dev;
-+
-+ fq->flags = QMAN_FQ_FLAG_NO_ENQUEUE;
-+ fq->channel = priv->channel;
-+}
-+
-+static inline void dpa_setup_egress(const struct dpa_priv_s *priv,
-+ struct dpa_fq *fq,
-+ struct fm_port *port,
-+ const struct qman_fq *template)
-+{
-+ fq->fq_base = *template;
-+ fq->net_dev = priv->net_dev;
-+
-+ if (port) {
-+ fq->flags = QMAN_FQ_FLAG_TO_DCPORTAL;
-+ fq->channel = (uint16_t)fm_get_tx_port_channel(port);
-+ } else {
-+ fq->flags = QMAN_FQ_FLAG_NO_MODIFY;
-+ }
-+}
-+
-+void dpa_fq_setup(struct dpa_priv_s *priv, const struct dpa_fq_cbs_t *fq_cbs,
-+ struct fm_port *tx_port)
-+{
-+ struct dpa_fq *fq;
-+ uint16_t portals[NR_CPUS];
-+ int cpu, portal_cnt = 0, num_portals = 0;
-+ uint32_t pcd_fqid, pcd_fqid_hi_prio;
-+ const cpumask_t *affine_cpus = qman_affine_cpus();
-+ int egress_cnt = 0, conf_cnt = 0;
-+
-+ /* Prepare for PCD FQs init */
-+ for_each_cpu(cpu, affine_cpus)
-+ portals[num_portals++] = qman_affine_channel(cpu);
-+ if (num_portals == 0)
-+ dev_err(priv->net_dev->dev.parent,
-+ "No Qman software (affine) channels found");
-+
-+ pcd_fqid = (priv->mac_dev) ?
-+ DPAA_ETH_PCD_FQ_BASE(priv->mac_dev->res->start) : 0;
-+ pcd_fqid_hi_prio = (priv->mac_dev) ?
-+ DPAA_ETH_PCD_FQ_HI_PRIO_BASE(priv->mac_dev->res->start) : 0;
-+
-+ /* Initialize each FQ in the list */
-+ list_for_each_entry(fq, &priv->dpa_fq_list, list) {
-+ switch (fq->fq_type) {
-+ case FQ_TYPE_RX_DEFAULT:
-+ BUG_ON(!priv->mac_dev);
-+ dpa_setup_ingress(priv, fq, &fq_cbs->rx_defq);
-+ break;
-+ case FQ_TYPE_RX_ERROR:
-+ BUG_ON(!priv->mac_dev);
-+ dpa_setup_ingress(priv, fq, &fq_cbs->rx_errq);
-+ break;
-+ case FQ_TYPE_RX_PCD:
-+ /* For MACless we can't have dynamic Rx queues */
-+ BUG_ON(!priv->mac_dev && !fq->fqid);
-+ dpa_setup_ingress(priv, fq, &fq_cbs->rx_defq);
-+ if (!fq->fqid)
-+ fq->fqid = pcd_fqid++;
-+ fq->channel = portals[portal_cnt];
-+ portal_cnt = (portal_cnt + 1) % num_portals;
-+ break;
-+ case FQ_TYPE_RX_PCD_HI_PRIO:
-+ /* For MACless we can't have dynamic Hi Pri Rx queues */
-+ BUG_ON(!priv->mac_dev && !fq->fqid);
-+ dpa_setup_ingress(priv, fq, &fq_cbs->rx_defq);
-+ if (!fq->fqid)
-+ fq->fqid = pcd_fqid_hi_prio++;
-+ fq->channel = portals[portal_cnt];
-+ portal_cnt = (portal_cnt + 1) % num_portals;
-+ break;
-+ case FQ_TYPE_TX:
-+ dpa_setup_egress(priv, fq, tx_port,
-+ &fq_cbs->egress_ern);
-+ /* If we have more Tx queues than the number of cores,
-+ * just ignore the extra ones.
-+ */
-+ if (egress_cnt < DPAA_ETH_TX_QUEUES)
-+ priv->egress_fqs[egress_cnt++] = &fq->fq_base;
-+ break;
-+ case FQ_TYPE_TX_CONFIRM:
-+ BUG_ON(!priv->mac_dev);
-+ dpa_setup_ingress(priv, fq, &fq_cbs->tx_defq);
-+ break;
-+ case FQ_TYPE_TX_CONF_MQ:
-+ BUG_ON(!priv->mac_dev);
-+ dpa_setup_ingress(priv, fq, &fq_cbs->tx_defq);
-+ priv->conf_fqs[conf_cnt++] = &fq->fq_base;
-+ break;
-+ case FQ_TYPE_TX_ERROR:
-+ BUG_ON(!priv->mac_dev);
-+ dpa_setup_ingress(priv, fq, &fq_cbs->tx_errq);
-+ break;
-+ default:
-+ dev_warn(priv->net_dev->dev.parent,
-+ "Unknown FQ type detected!\n");
-+ break;
-+ }
-+ }
-+
-+ /* The number of Tx queues may be smaller than the number of cores, if
-+ * the Tx queue range is specified in the device tree instead of being
-+ * dynamically allocated.
-+ * Make sure all CPUs receive a corresponding Tx queue.
-+ */
-+ while (egress_cnt < DPAA_ETH_TX_QUEUES) {
-+ list_for_each_entry(fq, &priv->dpa_fq_list, list) {
-+ if (fq->fq_type != FQ_TYPE_TX)
-+ continue;
-+ priv->egress_fqs[egress_cnt++] = &fq->fq_base;
-+ if (egress_cnt == DPAA_ETH_TX_QUEUES)
-+ break;
-+ }
-+ }
-+}
-+EXPORT_SYMBOL(dpa_fq_setup);
-+
-+int dpa_fq_init(struct dpa_fq *dpa_fq, bool td_enable)
-+{
-+ int _errno;
-+ const struct dpa_priv_s *priv;
-+ struct device *dev;
-+ struct qman_fq *fq;
-+ struct qm_mcc_initfq initfq;
-+ struct qman_fq *confq;
-+ int queue_id;
-+
-+ priv = netdev_priv(dpa_fq->net_dev);
-+ dev = dpa_fq->net_dev->dev.parent;
-+
-+ if (dpa_fq->fqid == 0)
-+ dpa_fq->flags |= QMAN_FQ_FLAG_DYNAMIC_FQID;
-+
-+ dpa_fq->init = !(dpa_fq->flags & QMAN_FQ_FLAG_NO_MODIFY);
-+
-+ _errno = qman_create_fq(dpa_fq->fqid, dpa_fq->flags, &dpa_fq->fq_base);
-+ if (_errno) {
-+ dev_err(dev, "qman_create_fq() failed\n");
-+ return _errno;
-+ }
-+ fq = &dpa_fq->fq_base;
-+
-+ if (dpa_fq->init) {
-+ memset(&initfq, 0, sizeof(initfq));
-+
-+ initfq.we_mask = QM_INITFQ_WE_FQCTRL;
-+ /* FIXME: why would we want to keep an empty FQ in cache? */
-+ initfq.fqd.fq_ctrl = QM_FQCTRL_PREFERINCACHE;
-+
-+ /* Try to reduce the number of portal interrupts for
-+ * Tx Confirmation FQs.
-+ */
-+ if (dpa_fq->fq_type == FQ_TYPE_TX_CONFIRM)
-+ initfq.fqd.fq_ctrl |= QM_FQCTRL_HOLDACTIVE;
-+
-+ /* FQ placement */
-+ initfq.we_mask |= QM_INITFQ_WE_DESTWQ;
-+
-+ initfq.fqd.dest.channel = dpa_fq->channel;
-+ initfq.fqd.dest.wq = dpa_fq->wq;
-+
-+ /* Put all egress queues in a congestion group of their own.
-+ * Sensu stricto, the Tx confirmation queues are Rx FQs,
-+ * rather than Tx - but they nonetheless account for the
-+ * memory footprint on behalf of egress traffic. We therefore
-+ * place them in the netdev's CGR, along with the Tx FQs.
-+ */
-+ if (dpa_fq->fq_type == FQ_TYPE_TX ||
-+ dpa_fq->fq_type == FQ_TYPE_TX_CONFIRM ||
-+ dpa_fq->fq_type == FQ_TYPE_TX_CONF_MQ) {
-+ initfq.we_mask |= QM_INITFQ_WE_CGID;
-+ initfq.fqd.fq_ctrl |= QM_FQCTRL_CGE;
-+ initfq.fqd.cgid = (uint8_t)priv->cgr_data.cgr.cgrid;
-+ /* Set a fixed overhead accounting, in an attempt to
-+ * reduce the impact of fixed-size skb shells and the
-+ * driver's needed headroom on system memory. This is
-+ * especially the case when the egress traffic is
-+ * composed of small datagrams.
-+ * Unfortunately, QMan's OAL value is capped to an
-+ * insufficient value, but even that is better than
-+ * no overhead accounting at all.
-+ */
-+ initfq.we_mask |= QM_INITFQ_WE_OAC;
-+ initfq.fqd.oac_init.oac = QM_OAC_CG;
-+ initfq.fqd.oac_init.oal =
-+ (signed char)(min(sizeof(struct sk_buff) +
-+ priv->tx_headroom, (size_t)FSL_QMAN_MAX_OAL));
-+ }
-+
-+ if (td_enable) {
-+ initfq.we_mask |= QM_INITFQ_WE_TDTHRESH;
-+ qm_fqd_taildrop_set(&initfq.fqd.td,
-+ DPA_FQ_TD, 1);
-+ initfq.fqd.fq_ctrl = QM_FQCTRL_TDE;
-+ }
-+
-+ /* Configure the Tx confirmation queue, now that we know
-+ * which Tx queue it pairs with.
-+ */
-+ if (dpa_fq->fq_type == FQ_TYPE_TX) {
-+ queue_id = _dpa_tx_fq_to_id(priv, &dpa_fq->fq_base);
-+ if (queue_id >= 0) {
-+ confq = priv->conf_fqs[queue_id];
-+ if (confq) {
-+ initfq.we_mask |= QM_INITFQ_WE_CONTEXTA;
-+ /* ContextA: OVOM=1 (use contextA2 bits instead of ICAD)
-+ * A2V=1 (contextA A2 field is valid)
-+ * A0V=1 (contextA A0 field is valid)
-+ * B0V=1 (contextB field is valid)
-+ * ContextA A2: EBD=1 (deallocate buffers inside FMan)
-+ * ContextB B0(ASPID): 0 (absolute Virtual Storage ID)
-+ */
-+ initfq.fqd.context_a.hi = 0x1e000000;
-+ initfq.fqd.context_a.lo = 0x80000000;
-+ }
-+ }
-+ }
-+
-+ /* Put all *private* ingress queues in our "ingress CGR". */
-+ if (priv->use_ingress_cgr &&
-+ (dpa_fq->fq_type == FQ_TYPE_RX_DEFAULT ||
-+ dpa_fq->fq_type == FQ_TYPE_RX_ERROR ||
-+ dpa_fq->fq_type == FQ_TYPE_RX_PCD ||
-+ dpa_fq->fq_type == FQ_TYPE_RX_PCD_HI_PRIO)) {
-+ initfq.we_mask |= QM_INITFQ_WE_CGID;
-+ initfq.fqd.fq_ctrl |= QM_FQCTRL_CGE;
-+ initfq.fqd.cgid = (uint8_t)priv->ingress_cgr.cgrid;
-+ /* Set a fixed overhead accounting, just like for the
-+ * egress CGR.
-+ */
-+ initfq.we_mask |= QM_INITFQ_WE_OAC;
-+ initfq.fqd.oac_init.oac = QM_OAC_CG;
-+ initfq.fqd.oac_init.oal =
-+ (signed char)(min(sizeof(struct sk_buff) +
-+ priv->tx_headroom, (size_t)FSL_QMAN_MAX_OAL));
-+ }
-+
-+ /* Initialization common to all ingress queues */
-+ if (dpa_fq->flags & QMAN_FQ_FLAG_NO_ENQUEUE) {
-+ initfq.we_mask |= QM_INITFQ_WE_CONTEXTA;
-+ initfq.fqd.fq_ctrl |=
-+ QM_FQCTRL_CTXASTASHING | QM_FQCTRL_AVOIDBLOCK;
-+ initfq.fqd.context_a.stashing.exclusive =
-+ QM_STASHING_EXCL_DATA | QM_STASHING_EXCL_CTX |
-+ QM_STASHING_EXCL_ANNOTATION;
-+ initfq.fqd.context_a.stashing.data_cl = 2;
-+ initfq.fqd.context_a.stashing.annotation_cl = 1;
-+ initfq.fqd.context_a.stashing.context_cl =
-+ DIV_ROUND_UP(sizeof(struct qman_fq), 64);
-+ }
-+
-+ _errno = qman_init_fq(fq, QMAN_INITFQ_FLAG_SCHED, &initfq);
-+ if (_errno < 0) {
-+ if (DPA_RX_PCD_HI_PRIO_FQ_INIT_FAIL(dpa_fq, _errno)) {
-+ dpa_fq->init = 0;
-+ } else {
-+ dev_err(dev, "qman_init_fq(%u) = %d\n",
-+ qman_fq_fqid(fq), _errno);
-+ qman_destroy_fq(fq, 0);
-+ }
-+ return _errno;
-+ }
-+ }
-+
-+ dpa_fq->fqid = qman_fq_fqid(fq);
-+
-+ return 0;
-+}
-+EXPORT_SYMBOL(dpa_fq_init);
-+
-+int __cold __attribute__((nonnull))
-+_dpa_fq_free(struct device *dev, struct qman_fq *fq)
-+{
-+ int _errno, __errno;
-+ struct dpa_fq *dpa_fq;
-+ const struct dpa_priv_s *priv;
-+
-+ _errno = 0;
-+
-+ dpa_fq = container_of(fq, struct dpa_fq, fq_base);
-+ priv = netdev_priv(dpa_fq->net_dev);
-+
-+ if (dpa_fq->init) {
-+ _errno = qman_retire_fq(fq, NULL);
-+ if (unlikely(_errno < 0) && netif_msg_drv(priv))
-+ dev_err(dev, "qman_retire_fq(%u) = %d\n",
-+ qman_fq_fqid(fq), _errno);
-+
-+ __errno = qman_oos_fq(fq);
-+ if (unlikely(__errno < 0) && netif_msg_drv(priv)) {
-+ dev_err(dev, "qman_oos_fq(%u) = %d\n",
-+ qman_fq_fqid(fq), __errno);
-+ if (_errno >= 0)
-+ _errno = __errno;
-+ }
-+ }
-+
-+ qman_destroy_fq(fq, 0);
-+ list_del(&dpa_fq->list);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(_dpa_fq_free);
-+
-+int __cold __attribute__((nonnull))
-+dpa_fq_free(struct device *dev, struct list_head *list)
-+{
-+ int _errno, __errno;
-+ struct dpa_fq *dpa_fq, *tmp;
-+
-+ _errno = 0;
-+ list_for_each_entry_safe(dpa_fq, tmp, list, list) {
-+ __errno = _dpa_fq_free(dev, (struct qman_fq *)dpa_fq);
-+ if (unlikely(__errno < 0) && _errno >= 0)
-+ _errno = __errno;
-+ }
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(dpa_fq_free);
-+
-+int dpa_fqs_init(struct device *dev, struct list_head *list, bool td_enable)
-+{
-+ int _errno, __errno;
-+ struct dpa_fq *dpa_fq, *tmp;
-+ static bool print_msg __read_mostly;
-+
-+ _errno = 0;
-+ print_msg = true;
-+ list_for_each_entry_safe(dpa_fq, tmp, list, list) {
-+ __errno = dpa_fq_init(dpa_fq, td_enable);
-+ if (unlikely(__errno < 0) && _errno >= 0) {
-+ if (DPA_RX_PCD_HI_PRIO_FQ_INIT_FAIL(dpa_fq, __errno)) {
-+ if (print_msg) {
-+ dev_warn(dev,
-+ "Skip RX PCD High Priority FQs initialization\n");
-+ print_msg = false;
-+ }
-+ if (_dpa_fq_free(dev, (struct qman_fq *)dpa_fq))
-+ dev_warn(dev,
-+ "Error freeing frame queues\n");
-+ } else {
-+ _errno = __errno;
-+ break;
-+ }
-+ }
-+ }
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(dpa_fqs_init);
-+static void
-+dpaa_eth_init_tx_port(struct fm_port *port, struct dpa_fq *errq,
-+ struct dpa_fq *defq, struct dpa_buffer_layout_s *buf_layout)
-+{
-+ struct fm_port_params tx_port_param;
-+ bool frag_enabled = false;
-+
-+ memset(&tx_port_param, 0, sizeof(tx_port_param));
-+ dpaa_eth_init_port(tx, port, tx_port_param, errq->fqid, defq->fqid,
-+ buf_layout, frag_enabled);
-+}
-+
-+static void
-+dpaa_eth_init_rx_port(struct fm_port *port, struct dpa_bp *bp, size_t count,
-+ struct dpa_fq *errq, struct dpa_fq *defq,
-+ struct dpa_buffer_layout_s *buf_layout)
-+{
-+ struct fm_port_params rx_port_param;
-+ int i;
-+ bool frag_enabled = false;
-+
-+ memset(&rx_port_param, 0, sizeof(rx_port_param));
-+ count = min(ARRAY_SIZE(rx_port_param.pool_param), count);
-+ rx_port_param.num_pools = (uint8_t)count;
-+ for (i = 0; i < count; i++) {
-+ if (i >= rx_port_param.num_pools)
-+ break;
-+ rx_port_param.pool_param[i].id = bp[i].bpid;
-+ rx_port_param.pool_param[i].size = (uint16_t)bp[i].size;
-+ }
-+
-+ dpaa_eth_init_port(rx, port, rx_port_param, errq->fqid, defq->fqid,
-+ buf_layout, frag_enabled);
-+}
-+
-+#if defined(CONFIG_FSL_SDK_FMAN_TEST)
-+/* Defined as weak, to be implemented by fman pcd tester. */
-+int dpa_alloc_pcd_fqids(struct device *, uint32_t, uint8_t, uint32_t *)
-+__attribute__((weak));
-+
-+int dpa_free_pcd_fqids(struct device *, uint32_t) __attribute__((weak));
-+#else
-+int dpa_alloc_pcd_fqids(struct device *, uint32_t, uint8_t, uint32_t *);
-+
-+int dpa_free_pcd_fqids(struct device *, uint32_t);
-+
-+#endif /* CONFIG_FSL_SDK_FMAN_TEST */
-+
-+
-+int dpa_alloc_pcd_fqids(struct device *dev, uint32_t num,
-+ uint8_t alignment, uint32_t *base_fqid)
-+{
-+ dev_crit(dev, "callback not implemented!\n");
-+
-+ return 0;
-+}
-+
-+int dpa_free_pcd_fqids(struct device *dev, uint32_t base_fqid)
-+{
-+
-+ dev_crit(dev, "callback not implemented!\n");
-+
-+ return 0;
-+}
-+
-+void dpaa_eth_init_ports(struct mac_device *mac_dev,
-+ struct dpa_bp *bp, size_t count,
-+ struct fm_port_fqs *port_fqs,
-+ struct dpa_buffer_layout_s *buf_layout,
-+ struct device *dev)
-+{
-+ struct fm_port_pcd_param rx_port_pcd_param;
-+ struct fm_port *rxport = mac_dev->port_dev[RX];
-+ struct fm_port *txport = mac_dev->port_dev[TX];
-+
-+ dpaa_eth_init_tx_port(txport, port_fqs->tx_errq,
-+ port_fqs->tx_defq, &buf_layout[TX]);
-+ dpaa_eth_init_rx_port(rxport, bp, count, port_fqs->rx_errq,
-+ port_fqs->rx_defq, &buf_layout[RX]);
-+
-+ rx_port_pcd_param.cba = dpa_alloc_pcd_fqids;
-+ rx_port_pcd_param.cbf = dpa_free_pcd_fqids;
-+ rx_port_pcd_param.dev = dev;
-+ fm_port_pcd_bind(rxport, &rx_port_pcd_param);
-+}
-+EXPORT_SYMBOL(dpaa_eth_init_ports);
-+
-+void dpa_release_sgt(struct qm_sg_entry *sgt)
-+{
-+ struct dpa_bp *dpa_bp;
-+ struct bm_buffer bmb[DPA_BUFF_RELEASE_MAX];
-+ uint8_t i = 0, j;
-+
-+ memset(bmb, 0, DPA_BUFF_RELEASE_MAX * sizeof(struct bm_buffer));
-+
-+ do {
-+ dpa_bp = dpa_bpid2pool(qm_sg_entry_get_bpid(&sgt[i]));
-+ DPA_BUG_ON(!dpa_bp);
-+
-+ j = 0;
-+ do {
-+ DPA_BUG_ON(qm_sg_entry_get_ext(&sgt[i]));
-+ bm_buffer_set64(&bmb[j], qm_sg_addr(&sgt[i]));
-+
-+ j++; i++;
-+ } while (j < ARRAY_SIZE(bmb) &&
-+ !qm_sg_entry_get_final(&sgt[i-1]) &&
-+ qm_sg_entry_get_bpid(&sgt[i-1]) ==
-+ qm_sg_entry_get_bpid(&sgt[i]));
-+
-+ while (bman_release(dpa_bp->pool, bmb, j, 0))
-+ cpu_relax();
-+ } while (!qm_sg_entry_get_final(&sgt[i-1]));
-+}
-+EXPORT_SYMBOL(dpa_release_sgt);
-+
-+void __attribute__((nonnull))
-+dpa_fd_release(const struct net_device *net_dev, const struct qm_fd *fd)
-+{
-+ struct qm_sg_entry *sgt;
-+ struct dpa_bp *dpa_bp;
-+ struct bm_buffer bmb;
-+ dma_addr_t addr;
-+ void *vaddr;
-+
-+ bmb.opaque = 0;
-+ bm_buffer_set64(&bmb, qm_fd_addr(fd));
-+
-+ dpa_bp = dpa_bpid2pool(fd->bpid);
-+ DPA_BUG_ON(!dpa_bp);
-+
-+ if (fd->format == qm_fd_sg) {
-+ vaddr = phys_to_virt(qm_fd_addr(fd));
-+ sgt = vaddr + dpa_fd_offset(fd);
-+
-+ dma_unmap_single(dpa_bp->dev, qm_fd_addr(fd), dpa_bp->size,
-+ DMA_BIDIRECTIONAL);
-+
-+ dpa_release_sgt(sgt);
-+ addr = dma_map_single(dpa_bp->dev, vaddr, dpa_bp->size,
-+ DMA_BIDIRECTIONAL);
-+ if (unlikely(dma_mapping_error(dpa_bp->dev, addr))) {
-+ dev_err(dpa_bp->dev, "DMA mapping failed");
-+ return;
-+ }
-+ bm_buffer_set64(&bmb, addr);
-+ }
-+
-+ while (bman_release(dpa_bp->pool, &bmb, 1, 0))
-+ cpu_relax();
-+}
-+EXPORT_SYMBOL(dpa_fd_release);
-+
-+void count_ern(struct dpa_percpu_priv_s *percpu_priv,
-+ const struct qm_mr_entry *msg)
-+{
-+ switch (msg->ern.rc & QM_MR_RC_MASK) {
-+ case QM_MR_RC_CGR_TAILDROP:
-+ percpu_priv->ern_cnt.cg_tdrop++;
-+ break;
-+ case QM_MR_RC_WRED:
-+ percpu_priv->ern_cnt.wred++;
-+ break;
-+ case QM_MR_RC_ERROR:
-+ percpu_priv->ern_cnt.err_cond++;
-+ break;
-+ case QM_MR_RC_ORPWINDOW_EARLY:
-+ percpu_priv->ern_cnt.early_window++;
-+ break;
-+ case QM_MR_RC_ORPWINDOW_LATE:
-+ percpu_priv->ern_cnt.late_window++;
-+ break;
-+ case QM_MR_RC_FQ_TAILDROP:
-+ percpu_priv->ern_cnt.fq_tdrop++;
-+ break;
-+ case QM_MR_RC_ORPWINDOW_RETIRED:
-+ percpu_priv->ern_cnt.fq_retired++;
-+ break;
-+ case QM_MR_RC_ORP_ZERO:
-+ percpu_priv->ern_cnt.orp_zero++;
-+ break;
-+ }
-+}
-+EXPORT_SYMBOL(count_ern);
-+
-+/**
-+ * Turn on HW checksum computation for this outgoing frame.
-+ * If the current protocol is not something we support in this regard
-+ * (or if the stack has already computed the SW checksum), we do nothing.
-+ *
-+ * Returns 0 if all goes well (or HW csum doesn't apply), and a negative value
-+ * otherwise.
-+ *
-+ * Note that this function may modify the fd->cmd field and the skb data buffer
-+ * (the Parse Results area).
-+ */
-+int dpa_enable_tx_csum(struct dpa_priv_s *priv,
-+ struct sk_buff *skb, struct qm_fd *fd, char *parse_results)
-+{
-+ fm_prs_result_t *parse_result;
-+ struct iphdr *iph;
-+ struct ipv6hdr *ipv6h = NULL;
-+ u8 l4_proto;
-+ u16 ethertype = ntohs(skb->protocol);
-+ int retval = 0;
-+
-+ if (skb->ip_summed != CHECKSUM_PARTIAL)
-+ return 0;
-+
-+ /* Note: L3 csum seems to be already computed in sw, but we can't choose
-+ * L4 alone from the FM configuration anyway.
-+ */
-+
-+ /* Fill in some fields of the Parse Results array, so the FMan
-+ * can find them as if they came from the FMan Parser.
-+ */
-+ parse_result = (fm_prs_result_t *)parse_results;
-+
-+ /* If we're dealing with VLAN, get the real Ethernet type */
-+ if (ethertype == ETH_P_8021Q) {
-+ /* We can't always assume the MAC header is set correctly
-+ * by the stack, so reset to beginning of skb->data
-+ */
-+ skb_reset_mac_header(skb);
-+ ethertype = ntohs(vlan_eth_hdr(skb)->h_vlan_encapsulated_proto);
-+ }
-+
-+ /* Fill in the relevant L3 parse result fields
-+ * and read the L4 protocol type
-+ */
-+ switch (ethertype) {
-+ case ETH_P_IP:
-+ parse_result->l3r = cpu_to_be16(FM_L3_PARSE_RESULT_IPV4);
-+ iph = ip_hdr(skb);
-+ DPA_BUG_ON(iph == NULL);
-+ l4_proto = iph->protocol;
-+ break;
-+ case ETH_P_IPV6:
-+ parse_result->l3r = cpu_to_be16(FM_L3_PARSE_RESULT_IPV6);
-+ ipv6h = ipv6_hdr(skb);
-+ DPA_BUG_ON(ipv6h == NULL);
-+ l4_proto = ipv6h->nexthdr;
-+ break;
-+ default:
-+ /* We shouldn't even be here */
-+ if (netif_msg_tx_err(priv) && net_ratelimit())
-+ netdev_alert(priv->net_dev,
-+ "Can't compute HW csum for L3 proto 0x%x\n",
-+ ntohs(skb->protocol));
-+ retval = -EIO;
-+ goto return_error;
-+ }
-+
-+ /* Fill in the relevant L4 parse result fields */
-+ switch (l4_proto) {
-+ case IPPROTO_UDP:
-+ parse_result->l4r = FM_L4_PARSE_RESULT_UDP;
-+ break;
-+ case IPPROTO_TCP:
-+ parse_result->l4r = FM_L4_PARSE_RESULT_TCP;
-+ break;
-+ default:
-+ /* This can as well be a BUG() */
-+ if (netif_msg_tx_err(priv) && net_ratelimit())
-+ netdev_alert(priv->net_dev,
-+ "Can't compute HW csum for L4 proto 0x%x\n",
-+ l4_proto);
-+ retval = -EIO;
-+ goto return_error;
-+ }
-+
-+ /* At index 0 is IPOffset_1 as defined in the Parse Results */
-+ parse_result->ip_off[0] = (uint8_t)skb_network_offset(skb);
-+ parse_result->l4_off = (uint8_t)skb_transport_offset(skb);
-+
-+ /* Enable L3 (and L4, if TCP or UDP) HW checksum. */
-+ fd->cmd |= FM_FD_CMD_RPD | FM_FD_CMD_DTC;
-+
-+ /* On P1023 and similar platforms fd->cmd interpretation could
-+ * be disabled by setting CONTEXT_A bit ICMD; currently this bit
-+ * is not set so we do not need to check; in the future, if/when
-+ * using context_a we need to check this bit
-+ */
-+
-+return_error:
-+ return retval;
-+}
-+EXPORT_SYMBOL(dpa_enable_tx_csum);
-+
-+#ifdef CONFIG_FSL_DPAA_CEETM
-+void dpa_enable_ceetm(struct net_device *dev)
-+{
-+ struct dpa_priv_s *priv = netdev_priv(dev);
-+ priv->ceetm_en = true;
-+}
-+EXPORT_SYMBOL(dpa_enable_ceetm);
-+
-+void dpa_disable_ceetm(struct net_device *dev)
-+{
-+ struct dpa_priv_s *priv = netdev_priv(dev);
-+ priv->ceetm_en = false;
-+}
-+EXPORT_SYMBOL(dpa_disable_ceetm);
-+#endif
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_dpaa/dpaa_eth_common.h
-@@ -0,0 +1,225 @@
-+/* Copyright 2008-2013 Freescale Semiconductor, Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#ifndef __DPAA_ETH_COMMON_H
-+#define __DPAA_ETH_COMMON_H
-+
-+#include <linux/etherdevice.h> /* struct net_device */
-+#include <linux/fsl_bman.h> /* struct bm_buffer */
-+#include <linux/of_platform.h> /* struct platform_device */
-+#include <linux/net_tstamp.h> /* struct hwtstamp_config */
-+
-+#include "dpaa_eth.h"
-+#include "lnxwrp_fsl_fman.h"
-+
-+#define dpaa_eth_init_port(type, port, param, errq_id, defq_id, buf_layout,\
-+ frag_enabled) \
-+{ \
-+ param.errq = errq_id; \
-+ param.defq = defq_id; \
-+ param.priv_data_size = buf_layout->priv_data_size; \
-+ param.parse_results = buf_layout->parse_results; \
-+ param.hash_results = buf_layout->hash_results; \
-+ param.frag_enable = frag_enabled; \
-+ param.time_stamp = buf_layout->time_stamp; \
-+ param.manip_extra_space = buf_layout->manip_extra_space; \
-+ param.data_align = buf_layout->data_align; \
-+ fm_set_##type##_port_params(port, &param); \
-+}
-+
-+#define DPA_SGT_MAX_ENTRIES 16 /* maximum number of entries in SG Table */
-+
-+#define DPA_SGT_ENTRIES_THRESHOLD DPA_SGT_MAX_ENTRIES
-+
-+#define DPA_BUFF_RELEASE_MAX 8 /* maximum number of buffers released at once */
-+
-+#define DPA_RX_PCD_HI_PRIO_FQ_INIT_FAIL(dpa_fq, _errno) \
-+ (((dpa_fq)->fq_type == FQ_TYPE_RX_PCD_HI_PRIO) && \
-+ (_errno == -EIO))
-+/* return codes for the dpaa-eth hooks */
-+enum dpaa_eth_hook_result {
-+ /* fd/skb was retained by the hook.
-+ *
-+ * On the Rx path, this means the Ethernet driver will _not_
-+ * deliver the skb to the stack. Instead, the hook implementation
-+ * is expected to properly dispose of the skb.
-+ *
-+ * On the Tx path, the Ethernet driver's dpa_tx() function will
-+ * immediately return NETDEV_TX_OK. The hook implementation is expected
-+ * to free the skb. *DO*NOT* release it to BMan, or enqueue it to FMan,
-+ * unless you know exactly what you're doing!
-+ *
-+ * On the confirmation/error paths, the Ethernet driver will _not_
-+ * perform any fd cleanup, nor update the interface statistics.
-+ */
-+ DPAA_ETH_STOLEN,
-+ /* fd/skb was returned to the Ethernet driver for regular processing.
-+ * The hook is not allowed to, for instance, reallocate the skb (as if
-+ * by linearizing, copying, cloning or reallocating the headroom).
-+ */
-+ DPAA_ETH_CONTINUE
-+};
-+
-+typedef enum dpaa_eth_hook_result (*dpaa_eth_ingress_hook_t)(
-+ struct sk_buff *skb, struct net_device *net_dev, u32 fqid);
-+typedef enum dpaa_eth_hook_result (*dpaa_eth_egress_hook_t)(
-+ struct sk_buff *skb, struct net_device *net_dev);
-+typedef enum dpaa_eth_hook_result (*dpaa_eth_confirm_hook_t)(
-+ struct net_device *net_dev, const struct qm_fd *fd, u32 fqid);
-+
-+/* used in napi related functions */
-+extern u16 qman_portal_max;
-+
-+/* from dpa_ethtool.c */
-+extern const struct ethtool_ops dpa_ethtool_ops;
-+
-+#ifdef CONFIG_FSL_DPAA_HOOKS
-+/* Various hooks used for unit-testing and/or fastpath optimizations.
-+ * Currently only one set of such hooks is supported.
-+ */
-+struct dpaa_eth_hooks_s {
-+ /* Invoked on the Tx private path, immediately after receiving the skb
-+ * from the stack.
-+ */
-+ dpaa_eth_egress_hook_t tx;
-+
-+ /* Invoked on the Rx private path, right before passing the skb
-+ * up the stack. At that point, the packet's protocol id has already
-+ * been set. The skb's data pointer is now at the L3 header, and
-+ * skb->mac_header points to the L2 header. skb->len has been adjusted
-+ * to be the length of L3+payload (i.e., the length of the
-+ * original frame minus the L2 header len).
-+ * For more details on what the skb looks like, see eth_type_trans().
-+ */
-+ dpaa_eth_ingress_hook_t rx_default;
-+
-+ /* Driver hook for the Rx error private path. */
-+ dpaa_eth_confirm_hook_t rx_error;
-+ /* Driver hook for the Tx confirmation private path. */
-+ dpaa_eth_confirm_hook_t tx_confirm;
-+ /* Driver hook for the Tx error private path. */
-+ dpaa_eth_confirm_hook_t tx_error;
-+};
-+
-+void fsl_dpaa_eth_set_hooks(struct dpaa_eth_hooks_s *hooks);
-+
-+extern struct dpaa_eth_hooks_s dpaa_eth_hooks;
-+#endif
-+
-+int dpa_netdev_init(struct net_device *net_dev,
-+ const uint8_t *mac_addr,
-+ uint16_t tx_timeout);
-+int __cold dpa_start(struct net_device *net_dev);
-+int __cold dpa_stop(struct net_device *net_dev);
-+void __cold dpa_timeout(struct net_device *net_dev);
-+struct rtnl_link_stats64 __cold
-+*dpa_get_stats64(struct net_device *net_dev,
-+ struct rtnl_link_stats64 *stats);
-+int dpa_change_mtu(struct net_device *net_dev, int new_mtu);
-+int dpa_ndo_init(struct net_device *net_dev);
-+int dpa_set_features(struct net_device *dev, netdev_features_t features);
-+netdev_features_t dpa_fix_features(struct net_device *dev,
-+ netdev_features_t features);
-+#ifdef CONFIG_FSL_DPAA_TS
-+u64 dpa_get_timestamp_ns(const struct dpa_priv_s *priv,
-+ enum port_type rx_tx, const void *data);
-+/* Updates the skb shared hw timestamp from the hardware timestamp */
-+int dpa_get_ts(const struct dpa_priv_s *priv, enum port_type rx_tx,
-+ struct skb_shared_hwtstamps *shhwtstamps, const void *data);
-+#endif /* CONFIG_FSL_DPAA_TS */
-+int dpa_ioctl(struct net_device *dev, struct ifreq *rq, int cmd);
-+int __cold dpa_remove(struct platform_device *of_dev);
-+struct mac_device * __cold __must_check
-+__attribute__((nonnull)) dpa_mac_probe(struct platform_device *_of_dev);
-+int dpa_set_mac_address(struct net_device *net_dev, void *addr);
-+void dpa_set_rx_mode(struct net_device *net_dev);
-+void dpa_set_buffers_layout(struct mac_device *mac_dev,
-+ struct dpa_buffer_layout_s *layout);
-+int __attribute__((nonnull))
-+dpa_bp_alloc(struct dpa_bp *dpa_bp, struct device *dev);
-+void __cold __attribute__((nonnull))
-+dpa_bp_free(struct dpa_priv_s *priv);
-+struct dpa_bp *dpa_bpid2pool(int bpid);
-+void dpa_bpid2pool_map(int bpid, struct dpa_bp *dpa_bp);
-+bool dpa_bpid2pool_use(int bpid);
-+void dpa_bp_drain(struct dpa_bp *bp);
-+#ifdef CONFIG_FSL_DPAA_ETH_USE_NDO_SELECT_QUEUE
-+u16 dpa_select_queue(struct net_device *net_dev, struct sk_buff *skb,
-+ void *accel_priv, select_queue_fallback_t fallback);
-+#endif
-+struct dpa_fq *dpa_fq_alloc(struct device *dev,
-+ u32 fq_start,
-+ u32 fq_count,
-+ struct list_head *list,
-+ enum dpa_fq_type fq_type);
-+int dpa_fq_probe_mac(struct device *dev, struct list_head *list,
-+ struct fm_port_fqs *port_fqs,
-+ bool tx_conf_fqs_per_core,
-+ enum port_type ptype);
-+int dpa_get_channel(void);
-+void dpa_release_channel(void);
-+void dpaa_eth_add_channel(u16 channel);
-+int dpaa_eth_cgr_init(struct dpa_priv_s *priv);
-+void dpa_fq_setup(struct dpa_priv_s *priv, const struct dpa_fq_cbs_t *fq_cbs,
-+ struct fm_port *tx_port);
-+int dpa_fq_init(struct dpa_fq *dpa_fq, bool td_enable);
-+int dpa_fqs_init(struct device *dev, struct list_head *list, bool td_enable);
-+int __cold __attribute__((nonnull))
-+dpa_fq_free(struct device *dev, struct list_head *list);
-+void dpaa_eth_init_ports(struct mac_device *mac_dev,
-+ struct dpa_bp *bp, size_t count,
-+ struct fm_port_fqs *port_fqs,
-+ struct dpa_buffer_layout_s *buf_layout,
-+ struct device *dev);
-+void dpa_release_sgt(struct qm_sg_entry *sgt);
-+void __attribute__((nonnull))
-+dpa_fd_release(const struct net_device *net_dev, const struct qm_fd *fd);
-+void count_ern(struct dpa_percpu_priv_s *percpu_priv,
-+ const struct qm_mr_entry *msg);
-+int dpa_enable_tx_csum(struct dpa_priv_s *priv,
-+ struct sk_buff *skb, struct qm_fd *fd, char *parse_results);
-+#ifdef CONFIG_FSL_DPAA_CEETM
-+void dpa_enable_ceetm(struct net_device *dev);
-+void dpa_disable_ceetm(struct net_device *dev);
-+#endif
-+struct proxy_device {
-+ struct mac_device *mac_dev;
-+};
-+
-+/* mac device control functions exposed by proxy interface*/
-+int dpa_proxy_start(struct net_device *net_dev);
-+int dpa_proxy_stop(struct proxy_device *proxy_dev, struct net_device *net_dev);
-+int dpa_proxy_set_mac_address(struct proxy_device *proxy_dev,
-+ struct net_device *net_dev);
-+int dpa_proxy_set_rx_mode(struct proxy_device *proxy_dev,
-+ struct net_device *net_dev);
-+
-+#endif /* __DPAA_ETH_COMMON_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_dpaa/dpaa_eth_proxy.c
-@@ -0,0 +1,381 @@
-+/* Copyright 2008-2013 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#ifdef CONFIG_FSL_DPAA_ETH_DEBUG
-+#define pr_fmt(fmt) \
-+ KBUILD_MODNAME ": %s:%hu:%s() " fmt, \
-+ KBUILD_BASENAME".c", __LINE__, __func__
-+#else
-+#define pr_fmt(fmt) \
-+ KBUILD_MODNAME ": " fmt
-+#endif
-+
-+#include <linux/init.h>
-+#include <linux/module.h>
-+#include <linux/of_platform.h>
-+#include "dpaa_eth.h"
-+#include "dpaa_eth_common.h"
-+#include "dpaa_eth_base.h"
-+#include "lnxwrp_fsl_fman.h" /* fm_get_rx_extra_headroom(), fm_get_max_frm() */
-+#include "mac.h"
-+
-+#define DPA_DESCRIPTION "FSL DPAA Proxy initialization driver"
-+
-+MODULE_LICENSE("Dual BSD/GPL");
-+
-+MODULE_DESCRIPTION(DPA_DESCRIPTION);
-+
-+static int __cold dpa_eth_proxy_remove(struct platform_device *of_dev);
-+#ifdef CONFIG_PM
-+
-+static int proxy_suspend(struct device *dev)
-+{
-+ struct proxy_device *proxy_dev = dev_get_drvdata(dev);
-+ struct mac_device *mac_dev = proxy_dev->mac_dev;
-+ int err = 0;
-+
-+ err = fm_port_suspend(mac_dev->port_dev[RX]);
-+ if (err)
-+ goto port_suspend_failed;
-+
-+ err = fm_port_suspend(mac_dev->port_dev[TX]);
-+ if (err)
-+ err = fm_port_resume(mac_dev->port_dev[RX]);
-+
-+port_suspend_failed:
-+ return err;
-+}
-+
-+static int proxy_resume(struct device *dev)
-+{
-+ struct proxy_device *proxy_dev = dev_get_drvdata(dev);
-+ struct mac_device *mac_dev = proxy_dev->mac_dev;
-+ int err = 0;
-+
-+ err = fm_port_resume(mac_dev->port_dev[TX]);
-+ if (err)
-+ goto port_resume_failed;
-+
-+ err = fm_port_resume(mac_dev->port_dev[RX]);
-+ if (err)
-+ err = fm_port_suspend(mac_dev->port_dev[TX]);
-+
-+port_resume_failed:
-+ return err;
-+}
-+
-+static const struct dev_pm_ops proxy_pm_ops = {
-+ .suspend = proxy_suspend,
-+ .resume = proxy_resume,
-+};
-+
-+#define PROXY_PM_OPS (&proxy_pm_ops)
-+
-+#else /* CONFIG_PM */
-+
-+#define PROXY_PM_OPS NULL
-+
-+#endif /* CONFIG_PM */
-+
-+static int dpaa_eth_proxy_probe(struct platform_device *_of_dev)
-+{
-+ int err = 0, i;
-+ struct device *dev;
-+ struct device_node *dpa_node;
-+ struct dpa_bp *dpa_bp;
-+ struct list_head proxy_fq_list;
-+ size_t count;
-+ struct fm_port_fqs port_fqs;
-+ struct dpa_buffer_layout_s *buf_layout = NULL;
-+ struct mac_device *mac_dev;
-+ struct proxy_device *proxy_dev;
-+
-+ dev = &_of_dev->dev;
-+
-+ dpa_node = dev->of_node;
-+
-+ if (!of_device_is_available(dpa_node))
-+ return -ENODEV;
-+
-+ /* Get the buffer pools assigned to this interface */
-+ dpa_bp = dpa_bp_probe(_of_dev, &count);
-+ if (IS_ERR(dpa_bp))
-+ return PTR_ERR(dpa_bp);
-+
-+ mac_dev = dpa_mac_probe(_of_dev);
-+ if (IS_ERR(mac_dev))
-+ return PTR_ERR(mac_dev);
-+
-+ proxy_dev = devm_kzalloc(dev, sizeof(*proxy_dev), GFP_KERNEL);
-+ if (!proxy_dev) {
-+ dev_err(dev, "devm_kzalloc() failed\n");
-+ return -ENOMEM;
-+ }
-+
-+ proxy_dev->mac_dev = mac_dev;
-+ dev_set_drvdata(dev, proxy_dev);
-+
-+ /* We have physical ports, so we need to establish
-+ * the buffer layout.
-+ */
-+ buf_layout = devm_kzalloc(dev, 2 * sizeof(*buf_layout),
-+ GFP_KERNEL);
-+ if (!buf_layout) {
-+ dev_err(dev, "devm_kzalloc() failed\n");
-+ return -ENOMEM;
-+ }
-+ dpa_set_buffers_layout(mac_dev, buf_layout);
-+
-+ INIT_LIST_HEAD(&proxy_fq_list);
-+
-+ memset(&port_fqs, 0, sizeof(port_fqs));
-+
-+ err = dpa_fq_probe_mac(dev, &proxy_fq_list, &port_fqs, true, RX);
-+ if (!err)
-+ err = dpa_fq_probe_mac(dev, &proxy_fq_list, &port_fqs, true,
-+ TX);
-+ if (err < 0) {
-+ devm_kfree(dev, buf_layout);
-+ return err;
-+ }
-+
-+ /* Proxy initializer - Just configures the MAC on behalf of
-+ * another partition.
-+ */
-+ dpaa_eth_init_ports(mac_dev, dpa_bp, count, &port_fqs,
-+ buf_layout, dev);
-+
-+ /* Proxy interfaces need to be started, and the allocated
-+ * memory freed
-+ */
-+ devm_kfree(dev, buf_layout);
-+ devm_kfree(dev, dpa_bp);
-+
-+ /* Free FQ structures */
-+ devm_kfree(dev, port_fqs.rx_defq);
-+ devm_kfree(dev, port_fqs.rx_errq);
-+ devm_kfree(dev, port_fqs.tx_defq);
-+ devm_kfree(dev, port_fqs.tx_errq);
-+
-+ for_each_port_device(i, mac_dev->port_dev) {
-+ err = fm_port_enable(mac_dev->port_dev[i]);
-+ if (err)
-+ goto port_enable_fail;
-+ }
-+
-+ dev_info(dev, "probed MAC device with MAC address: %02hx:%02hx:%02hx:%02hx:%02hx:%02hx\n",
-+ mac_dev->addr[0], mac_dev->addr[1], mac_dev->addr[2],
-+ mac_dev->addr[3], mac_dev->addr[4], mac_dev->addr[5]);
-+
-+ return 0; /* Proxy interface initialization ended */
-+
-+port_enable_fail:
-+ for_each_port_device(i, mac_dev->port_dev)
-+ fm_port_disable(mac_dev->port_dev[i]);
-+ dpa_eth_proxy_remove(_of_dev);
-+
-+ return err;
-+}
-+
-+int dpa_proxy_set_mac_address(struct proxy_device *proxy_dev,
-+ struct net_device *net_dev)
-+{
-+ struct mac_device *mac_dev;
-+ int _errno;
-+
-+ mac_dev = proxy_dev->mac_dev;
-+
-+ _errno = mac_dev->change_addr(mac_dev->get_mac_handle(mac_dev),
-+ net_dev->dev_addr);
-+ if (_errno < 0)
-+ return _errno;
-+
-+ return 0;
-+}
-+EXPORT_SYMBOL(dpa_proxy_set_mac_address);
-+
-+int dpa_proxy_set_rx_mode(struct proxy_device *proxy_dev,
-+ struct net_device *net_dev)
-+{
-+ struct mac_device *mac_dev = proxy_dev->mac_dev;
-+ int _errno;
-+
-+ if (!!(net_dev->flags & IFF_PROMISC) != mac_dev->promisc) {
-+ mac_dev->promisc = !mac_dev->promisc;
-+ _errno = mac_dev->set_promisc(mac_dev->get_mac_handle(mac_dev),
-+ mac_dev->promisc);
-+ if (unlikely(_errno < 0))
-+ netdev_err(net_dev, "mac_dev->set_promisc() = %d\n",
-+ _errno);
-+ }
-+
-+ _errno = mac_dev->set_multi(net_dev, mac_dev);
-+ if (unlikely(_errno < 0))
-+ return _errno;
-+
-+ return 0;
-+}
-+EXPORT_SYMBOL(dpa_proxy_set_rx_mode);
-+
-+int dpa_proxy_start(struct net_device *net_dev)
-+{
-+ struct mac_device *mac_dev;
-+ const struct dpa_priv_s *priv;
-+ struct proxy_device *proxy_dev;
-+ int _errno;
-+ int i;
-+
-+ priv = netdev_priv(net_dev);
-+ proxy_dev = (struct proxy_device *)priv->peer;
-+ mac_dev = proxy_dev->mac_dev;
-+
-+ _errno = mac_dev->init_phy(net_dev, mac_dev);
-+ if (_errno < 0) {
-+ if (netif_msg_drv(priv))
-+ netdev_err(net_dev, "init_phy() = %d\n",
-+ _errno);
-+ return _errno;
-+ }
-+
-+ for_each_port_device(i, mac_dev->port_dev) {
-+ _errno = fm_port_enable(mac_dev->port_dev[i]);
-+ if (_errno)
-+ goto port_enable_fail;
-+ }
-+
-+ _errno = mac_dev->start(mac_dev);
-+ if (_errno < 0) {
-+ if (netif_msg_drv(priv))
-+ netdev_err(net_dev, "mac_dev->start() = %d\n",
-+ _errno);
-+ goto port_enable_fail;
-+ }
-+
-+ return _errno;
-+
-+port_enable_fail:
-+ for_each_port_device(i, mac_dev->port_dev)
-+ fm_port_disable(mac_dev->port_dev[i]);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(dpa_proxy_start);
-+
-+int dpa_proxy_stop(struct proxy_device *proxy_dev, struct net_device *net_dev)
-+{
-+ struct mac_device *mac_dev = proxy_dev->mac_dev;
-+ const struct dpa_priv_s *priv = netdev_priv(net_dev);
-+ int _errno, i, err;
-+
-+ _errno = mac_dev->stop(mac_dev);
-+ if (_errno < 0) {
-+ if (netif_msg_drv(priv))
-+ netdev_err(net_dev, "mac_dev->stop() = %d\n",
-+ _errno);
-+ return _errno;
-+ }
-+
-+ for_each_port_device(i, mac_dev->port_dev) {
-+ err = fm_port_disable(mac_dev->port_dev[i]);
-+ _errno = err ? err : _errno;
-+ }
-+
-+ if (mac_dev->phy_dev)
-+ phy_disconnect(mac_dev->phy_dev);
-+ mac_dev->phy_dev = NULL;
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(dpa_proxy_stop);
-+
-+static int __cold dpa_eth_proxy_remove(struct platform_device *of_dev)
-+{
-+ struct device *dev = &of_dev->dev;
-+ struct proxy_device *proxy_dev = dev_get_drvdata(dev);
-+
-+ kfree(proxy_dev);
-+
-+ dev_set_drvdata(dev, NULL);
-+
-+ return 0;
-+}
-+
-+static const struct of_device_id dpa_proxy_match[] = {
-+ {
-+ .compatible = "fsl,dpa-ethernet-init"
-+ },
-+ {}
-+};
-+MODULE_DEVICE_TABLE(of, dpa_proxy_match);
-+
-+static struct platform_driver dpa_proxy_driver = {
-+ .driver = {
-+ .name = KBUILD_MODNAME "-proxy",
-+ .of_match_table = dpa_proxy_match,
-+ .owner = THIS_MODULE,
-+ .pm = PROXY_PM_OPS,
-+ },
-+ .probe = dpaa_eth_proxy_probe,
-+ .remove = dpa_eth_proxy_remove
-+};
-+
-+static int __init __cold dpa_proxy_load(void)
-+{
-+ int _errno;
-+
-+ pr_info(DPA_DESCRIPTION "\n");
-+
-+ /* Initialize dpaa_eth mirror values */
-+ dpa_rx_extra_headroom = fm_get_rx_extra_headroom();
-+ dpa_max_frm = fm_get_max_frm();
-+
-+ _errno = platform_driver_register(&dpa_proxy_driver);
-+ if (unlikely(_errno < 0)) {
-+ pr_err(KBUILD_MODNAME
-+ ": %s:%hu:%s(): platform_driver_register() = %d\n",
-+ KBUILD_BASENAME".c", __LINE__, __func__, _errno);
-+ }
-+
-+ pr_debug(KBUILD_MODNAME ": %s:%s() ->\n",
-+ KBUILD_BASENAME".c", __func__);
-+
-+ return _errno;
-+}
-+module_init(dpa_proxy_load);
-+
-+static void __exit __cold dpa_proxy_unload(void)
-+{
-+ platform_driver_unregister(&dpa_proxy_driver);
-+
-+ pr_debug(KBUILD_MODNAME ": %s:%s() ->\n",
-+ KBUILD_BASENAME".c", __func__);
-+}
-+module_exit(dpa_proxy_unload);
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_dpaa/dpaa_eth_sg.c
-@@ -0,0 +1,1193 @@
-+/* Copyright 2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#ifdef CONFIG_FSL_DPAA_ETH_DEBUG
-+#define pr_fmt(fmt) \
-+ KBUILD_MODNAME ": %s:%hu:%s() " fmt, \
-+ KBUILD_BASENAME".c", __LINE__, __func__
-+#else
-+#define pr_fmt(fmt) \
-+ KBUILD_MODNAME ": " fmt
-+#endif
-+
-+#include <linux/init.h>
-+#include <linux/skbuff.h>
-+#include <linux/highmem.h>
-+#include <linux/fsl_bman.h>
-+#include <net/sock.h>
-+
-+#include "dpaa_eth.h"
-+#include "dpaa_eth_common.h"
-+#ifdef CONFIG_FSL_DPAA_1588
-+#include "dpaa_1588.h"
-+#endif
-+#ifdef CONFIG_FSL_DPAA_CEETM
-+#include "dpaa_eth_ceetm.h"
-+#endif
-+
-+/* DMA map and add a page frag back into the bpool.
-+ * @vaddr fragment must have been allocated with netdev_alloc_frag(),
-+ * specifically for fitting into @dpa_bp.
-+ */
-+static void dpa_bp_recycle_frag(struct dpa_bp *dpa_bp, unsigned long vaddr,
-+ int *count_ptr)
-+{
-+ struct bm_buffer bmb;
-+ dma_addr_t addr;
-+
-+ bmb.opaque = 0;
-+
-+ addr = dma_map_single(dpa_bp->dev, (void *)vaddr, dpa_bp->size,
-+ DMA_BIDIRECTIONAL);
-+ if (unlikely(dma_mapping_error(dpa_bp->dev, addr))) {
-+ dev_err(dpa_bp->dev, "DMA mapping failed");
-+ return;
-+ }
-+
-+ bm_buffer_set64(&bmb, addr);
-+
-+ while (bman_release(dpa_bp->pool, &bmb, 1, 0))
-+ cpu_relax();
-+
-+ (*count_ptr)++;
-+}
-+
-+static int _dpa_bp_add_8_bufs(const struct dpa_bp *dpa_bp)
-+{
-+ struct bm_buffer bmb[8];
-+ void *new_buf;
-+ dma_addr_t addr;
-+ uint8_t i;
-+ struct device *dev = dpa_bp->dev;
-+ struct sk_buff *skb, **skbh;
-+
-+ memset(bmb, 0, sizeof(struct bm_buffer) * 8);
-+
-+ for (i = 0; i < 8; i++) {
-+ /* We'll prepend the skb back-pointer; can't use the DPA
-+ * priv space, because FMan will overwrite it (from offset 0)
-+ * if it ends up being the second, third, etc. fragment
-+ * in a S/G frame.
-+ *
-+ * We only need enough space to store a pointer, but allocate
-+ * an entire cacheline for performance reasons.
-+ */
-+#ifndef CONFIG_PPC
-+ if (unlikely(dpaa_errata_a010022)) {
-+ struct page *new_page = alloc_page(GFP_ATOMIC);
-+ if (unlikely(!new_page))
-+ goto netdev_alloc_failed;
-+ new_buf = page_address(new_page);
-+ }
-+ else
-+#endif
-+ new_buf = netdev_alloc_frag(SMP_CACHE_BYTES + DPA_BP_RAW_SIZE);
-+
-+ if (unlikely(!new_buf))
-+ goto netdev_alloc_failed;
-+ new_buf = PTR_ALIGN(new_buf + SMP_CACHE_BYTES, SMP_CACHE_BYTES);
-+
-+ skb = build_skb(new_buf, DPA_SKB_SIZE(dpa_bp->size) +
-+ SKB_DATA_ALIGN(sizeof(struct skb_shared_info)));
-+ if (unlikely(!skb)) {
-+ put_page(virt_to_head_page(new_buf));
-+ goto build_skb_failed;
-+ }
-+
-+ /* Store the skb back-pointer before the start of the buffer.
-+ * Otherwise it will be overwritten by the FMan.
-+ */
-+ DPA_WRITE_SKB_PTR(skb, skbh, new_buf, -1);
-+
-+ addr = dma_map_single(dev, new_buf,
-+ dpa_bp->size, DMA_BIDIRECTIONAL);
-+ if (unlikely(dma_mapping_error(dev, addr)))
-+ goto dma_map_failed;
-+
-+ bm_buffer_set64(&bmb[i], addr);
-+ }
-+
-+release_bufs:
-+ /* Release the buffers. In case bman is busy, keep trying
-+ * until successful. bman_release() is guaranteed to succeed
-+ * in a reasonable amount of time
-+ */
-+ while (unlikely(bman_release(dpa_bp->pool, bmb, i, 0)))
-+ cpu_relax();
-+ return i;
-+
-+dma_map_failed:
-+ kfree_skb(skb);
-+
-+build_skb_failed:
-+netdev_alloc_failed:
-+ net_err_ratelimited("dpa_bp_add_8_bufs() failed\n");
-+ WARN_ONCE(1, "Memory allocation failure on Rx\n");
-+
-+ bm_buffer_set64(&bmb[i], 0);
-+ /* Avoid releasing a completely null buffer; bman_release() requires
-+ * at least one buffer.
-+ */
-+ if (likely(i))
-+ goto release_bufs;
-+
-+ return 0;
-+}
-+
-+/* Cold path wrapper over _dpa_bp_add_8_bufs(). */
-+static void dpa_bp_add_8_bufs(const struct dpa_bp *dpa_bp, int cpu)
-+{
-+ int *count_ptr = per_cpu_ptr(dpa_bp->percpu_count, cpu);
-+ *count_ptr += _dpa_bp_add_8_bufs(dpa_bp);
-+}
-+
-+int dpa_bp_priv_seed(struct dpa_bp *dpa_bp)
-+{
-+ int i;
-+
-+ /* Give each CPU an allotment of "config_count" buffers */
-+ for_each_possible_cpu(i) {
-+ int j;
-+
-+ /* Although we access another CPU's counters here
-+ * we do it at boot time so it is safe
-+ */
-+ for (j = 0; j < dpa_bp->config_count; j += 8)
-+ dpa_bp_add_8_bufs(dpa_bp, i);
-+ }
-+ return 0;
-+}
-+EXPORT_SYMBOL(dpa_bp_priv_seed);
-+
-+/* Add buffers/(pages) for Rx processing whenever bpool count falls below
-+ * REFILL_THRESHOLD.
-+ */
-+int dpaa_eth_refill_bpools(struct dpa_bp *dpa_bp, int *countptr)
-+{
-+ int count = *countptr;
-+ int new_bufs;
-+
-+ if (unlikely(count < CONFIG_FSL_DPAA_ETH_REFILL_THRESHOLD)) {
-+ do {
-+ new_bufs = _dpa_bp_add_8_bufs(dpa_bp);
-+ if (unlikely(!new_bufs)) {
-+ /* Avoid looping forever if we've temporarily
-+ * run out of memory. We'll try again at the
-+ * next NAPI cycle.
-+ */
-+ break;
-+ }
-+ count += new_bufs;
-+ } while (count < CONFIG_FSL_DPAA_ETH_MAX_BUF_COUNT);
-+
-+ *countptr = count;
-+ if (unlikely(count < CONFIG_FSL_DPAA_ETH_MAX_BUF_COUNT))
-+ return -ENOMEM;
-+ }
-+
-+ return 0;
-+}
-+EXPORT_SYMBOL(dpaa_eth_refill_bpools);
-+
-+/* Cleanup function for outgoing frame descriptors that were built on Tx path,
-+ * either contiguous frames or scatter/gather ones.
-+ * Skb freeing is not handled here.
-+ *
-+ * This function may be called on error paths in the Tx function, so guard
-+ * against cases when not all fd relevant fields were filled in.
-+ *
-+ * Return the skb backpointer, since for S/G frames the buffer containing it
-+ * gets freed here.
-+ */
-+struct sk_buff *_dpa_cleanup_tx_fd(const struct dpa_priv_s *priv,
-+ const struct qm_fd *fd)
-+{
-+ const struct qm_sg_entry *sgt;
-+ int i;
-+ struct dpa_bp *dpa_bp = priv->dpa_bp;
-+ dma_addr_t addr = qm_fd_addr(fd);
-+ dma_addr_t sg_addr;
-+ struct sk_buff **skbh;
-+ struct sk_buff *skb = NULL;
-+ const enum dma_data_direction dma_dir = DMA_TO_DEVICE;
-+ int nr_frags;
-+ int sg_len;
-+
-+ /* retrieve skb back pointer */
-+ DPA_READ_SKB_PTR(skb, skbh, phys_to_virt(addr), 0);
-+
-+ if (unlikely(fd->format == qm_fd_sg)) {
-+ nr_frags = skb_shinfo(skb)->nr_frags;
-+ dma_unmap_single(dpa_bp->dev, addr, dpa_fd_offset(fd) +
-+ sizeof(struct qm_sg_entry) * (1 + nr_frags),
-+ dma_dir);
-+
-+ /* The sgt buffer has been allocated with netdev_alloc_frag(),
-+ * it's from lowmem.
-+ */
-+ sgt = phys_to_virt(addr + dpa_fd_offset(fd));
-+#ifdef CONFIG_FSL_DPAA_1588
-+ if (priv->tsu && priv->tsu->valid &&
-+ priv->tsu->hwts_tx_en_ioctl)
-+ dpa_ptp_store_txstamp(priv, skb, (void *)skbh);
-+#endif
-+#ifdef CONFIG_FSL_DPAA_TS
-+ if (unlikely(priv->ts_tx_en &&
-+ skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP)) {
-+ struct skb_shared_hwtstamps shhwtstamps;
-+
-+ dpa_get_ts(priv, TX, &shhwtstamps, (void *)skbh);
-+ skb_tstamp_tx(skb, &shhwtstamps);
-+ }
-+#endif /* CONFIG_FSL_DPAA_TS */
-+
-+ /* sgt[0] is from lowmem, was dma_map_single()-ed */
-+ sg_addr = qm_sg_addr(&sgt[0]);
-+ sg_len = qm_sg_entry_get_len(&sgt[0]);
-+ dma_unmap_single(dpa_bp->dev, sg_addr, sg_len, dma_dir);
-+
-+ /* remaining pages were mapped with dma_map_page() */
-+ for (i = 1; i <= nr_frags; i++) {
-+ DPA_BUG_ON(qm_sg_entry_get_ext(&sgt[i]));
-+ sg_addr = qm_sg_addr(&sgt[i]);
-+ sg_len = qm_sg_entry_get_len(&sgt[i]);
-+ dma_unmap_page(dpa_bp->dev, sg_addr, sg_len, dma_dir);
-+ }
-+
-+ /* Free the page frag that we allocated on Tx */
-+ put_page(virt_to_head_page(sgt));
-+ } else {
-+ dma_unmap_single(dpa_bp->dev, addr,
-+ skb_tail_pointer(skb) - (u8 *)skbh, dma_dir);
-+#ifdef CONFIG_FSL_DPAA_TS
-+ /* get the timestamp for non-SG frames */
-+#ifdef CONFIG_FSL_DPAA_1588
-+ if (priv->tsu && priv->tsu->valid &&
-+ priv->tsu->hwts_tx_en_ioctl)
-+ dpa_ptp_store_txstamp(priv, skb, (void *)skbh);
-+#endif
-+ if (unlikely(priv->ts_tx_en &&
-+ skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP)) {
-+ struct skb_shared_hwtstamps shhwtstamps;
-+
-+ dpa_get_ts(priv, TX, &shhwtstamps, (void *)skbh);
-+ skb_tstamp_tx(skb, &shhwtstamps);
-+ }
-+#endif
-+ }
-+
-+ return skb;
-+}
-+EXPORT_SYMBOL(_dpa_cleanup_tx_fd);
-+
-+#ifndef CONFIG_FSL_DPAA_TS
-+bool dpa_skb_is_recyclable(struct sk_buff *skb)
-+{
-+#ifndef CONFIG_PPC
-+ /* Do no recycle skbs realigned by the errata workaround */
-+ if (unlikely(dpaa_errata_a010022) && skb->mark == NONREC_MARK)
-+ return false;
-+#endif
-+
-+ /* No recycling possible if skb buffer is kmalloc'ed */
-+ if (skb->head_frag == 0)
-+ return false;
-+
-+ /* or if it's an userspace buffer */
-+ if (skb_shinfo(skb)->tx_flags & SKBTX_DEV_ZEROCOPY)
-+ return false;
-+
-+ /* or if it's cloned or shared */
-+ if (skb_shared(skb) || skb_cloned(skb) ||
-+ skb->fclone != SKB_FCLONE_UNAVAILABLE)
-+ return false;
-+
-+ return true;
-+}
-+EXPORT_SYMBOL(dpa_skb_is_recyclable);
-+
-+bool dpa_buf_is_recyclable(struct sk_buff *skb,
-+ uint32_t min_size,
-+ uint16_t min_offset,
-+ unsigned char **new_buf_start)
-+{
-+ unsigned char *new;
-+
-+ /* In order to recycle a buffer, the following conditions must be met:
-+ * - buffer size no less than the buffer pool size
-+ * - buffer size no higher than an upper limit (to avoid moving too much
-+ * system memory to the buffer pools)
-+ * - buffer address aligned to cacheline bytes
-+ * - offset of data from start of buffer no lower than a minimum value
-+ * - offset of data from start of buffer no higher than a maximum value
-+ * - the skb back-pointer is stored safely
-+ */
-+
-+ /* guarantee both the minimum size and the minimum data offset */
-+ new = min(skb_end_pointer(skb) - min_size, skb->data - min_offset);
-+
-+ /* left align to the nearest cacheline */
-+ new = (unsigned char *)((unsigned long)new & ~(SMP_CACHE_BYTES - 1));
-+
-+ /* Make sure there is enough space to store the skb back-pointer in
-+ * the headroom, right before the start of the buffer.
-+ *
-+ * Guarantee that both maximum size and maximum data offsets aren't
-+ * crossed.
-+ */
-+ if (likely(new >= (skb->head + sizeof(void *)) &&
-+ new >= (skb->data - DPA_MAX_FD_OFFSET) &&
-+ skb_end_pointer(skb) - new <= DPA_RECYCLE_MAX_SIZE)) {
-+ *new_buf_start = new;
-+ return true;
-+ }
-+
-+ return false;
-+}
-+EXPORT_SYMBOL(dpa_buf_is_recyclable);
-+#endif
-+
-+/* Build a linear skb around the received buffer.
-+ * We are guaranteed there is enough room at the end of the data buffer to
-+ * accommodate the shared info area of the skb.
-+ */
-+static struct sk_buff *__hot contig_fd_to_skb(const struct dpa_priv_s *priv,
-+ const struct qm_fd *fd, int *use_gro)
-+{
-+ dma_addr_t addr = qm_fd_addr(fd);
-+ ssize_t fd_off = dpa_fd_offset(fd);
-+ void *vaddr;
-+ const fm_prs_result_t *parse_results;
-+ struct sk_buff *skb = NULL, **skbh;
-+
-+ vaddr = phys_to_virt(addr);
-+ DPA_BUG_ON(!IS_ALIGNED((unsigned long)vaddr, SMP_CACHE_BYTES));
-+
-+ /* Retrieve the skb and adjust data and tail pointers, to make sure
-+ * forwarded skbs will have enough space on Tx if extra headers
-+ * are added.
-+ */
-+ DPA_READ_SKB_PTR(skb, skbh, vaddr, -1);
-+
-+#ifdef CONFIG_FSL_DPAA_ETH_JUMBO_FRAME
-+ /* When using jumbo Rx buffers, we risk having frames dropped due to
-+ * the socket backlog reaching its maximum allowed size.
-+ * Use the frame length for the skb truesize instead of the buffer
-+ * size, as this is the size of the data that actually gets copied to
-+ * userspace.
-+ * The stack may increase the payload. In this case, it will want to
-+ * warn us that the frame length is larger than the truesize. We
-+ * bypass the warning.
-+ */
-+#ifndef CONFIG_PPC
-+ /* We do not support Jumbo frames on LS1043 and thus we edit
-+ * the skb truesize only when the 4k errata is not present.
-+ */
-+ if (likely(!dpaa_errata_a010022))
-+#endif
-+ skb->truesize = SKB_TRUESIZE(dpa_fd_length(fd));
-+#endif
-+
-+ DPA_BUG_ON(fd_off != priv->rx_headroom);
-+ skb_reserve(skb, fd_off);
-+ skb_put(skb, dpa_fd_length(fd));
-+
-+ /* Peek at the parse results for csum validation */
-+ parse_results = (const fm_prs_result_t *)(vaddr +
-+ DPA_RX_PRIV_DATA_SIZE);
-+ _dpa_process_parse_results(parse_results, fd, skb, use_gro);
-+
-+#ifdef CONFIG_FSL_DPAA_1588
-+ if (priv->tsu && priv->tsu->valid && priv->tsu->hwts_rx_en_ioctl)
-+ dpa_ptp_store_rxstamp(priv, skb, vaddr);
-+#endif
-+#ifdef CONFIG_FSL_DPAA_TS
-+ if (priv->ts_rx_en)
-+ dpa_get_ts(priv, RX, skb_hwtstamps(skb), vaddr);
-+#endif /* CONFIG_FSL_DPAA_TS */
-+
-+ return skb;
-+}
-+
-+
-+/* Build an skb with the data of the first S/G entry in the linear portion and
-+ * the rest of the frame as skb fragments.
-+ *
-+ * The page fragment holding the S/G Table is recycled here.
-+ */
-+static struct sk_buff *__hot sg_fd_to_skb(const struct dpa_priv_s *priv,
-+ const struct qm_fd *fd, int *use_gro,
-+ int *count_ptr)
-+{
-+ const struct qm_sg_entry *sgt;
-+ dma_addr_t addr = qm_fd_addr(fd);
-+ ssize_t fd_off = dpa_fd_offset(fd);
-+ dma_addr_t sg_addr;
-+ void *vaddr, *sg_vaddr;
-+ struct dpa_bp *dpa_bp;
-+ struct page *page, *head_page;
-+ int frag_offset, frag_len;
-+ int page_offset;
-+ int i;
-+ const fm_prs_result_t *parse_results;
-+ struct sk_buff *skb = NULL, *skb_tmp, **skbh;
-+
-+ vaddr = phys_to_virt(addr);
-+ DPA_BUG_ON(!IS_ALIGNED((unsigned long)vaddr, SMP_CACHE_BYTES));
-+
-+ dpa_bp = priv->dpa_bp;
-+ /* Iterate through the SGT entries and add data buffers to the skb */
-+ sgt = vaddr + fd_off;
-+ for (i = 0; i < DPA_SGT_MAX_ENTRIES; i++) {
-+ /* Extension bit is not supported */
-+ DPA_BUG_ON(qm_sg_entry_get_ext(&sgt[i]));
-+
-+ /* We use a single global Rx pool */
-+ DPA_BUG_ON(dpa_bp !=
-+ dpa_bpid2pool(qm_sg_entry_get_bpid(&sgt[i])));
-+
-+ sg_addr = qm_sg_addr(&sgt[i]);
-+ sg_vaddr = phys_to_virt(sg_addr);
-+ DPA_BUG_ON(!IS_ALIGNED((unsigned long)sg_vaddr,
-+ SMP_CACHE_BYTES));
-+
-+ dma_unmap_single(dpa_bp->dev, sg_addr, dpa_bp->size,
-+ DMA_BIDIRECTIONAL);
-+ if (i == 0) {
-+ DPA_READ_SKB_PTR(skb, skbh, sg_vaddr, -1);
-+ DPA_BUG_ON(skb->head != sg_vaddr);
-+#ifdef CONFIG_FSL_DPAA_1588
-+ if (priv->tsu && priv->tsu->valid &&
-+ priv->tsu->hwts_rx_en_ioctl)
-+ dpa_ptp_store_rxstamp(priv, skb, vaddr);
-+#endif
-+#ifdef CONFIG_FSL_DPAA_TS
-+ if (priv->ts_rx_en)
-+ dpa_get_ts(priv, RX, skb_hwtstamps(skb), vaddr);
-+#endif /* CONFIG_FSL_DPAA_TS */
-+
-+ /* In the case of a SG frame, FMan stores the Internal
-+ * Context in the buffer containing the sgt.
-+ * Inspect the parse results before anything else.
-+ */
-+ parse_results = (const fm_prs_result_t *)(vaddr +
-+ DPA_RX_PRIV_DATA_SIZE);
-+ _dpa_process_parse_results(parse_results, fd, skb,
-+ use_gro);
-+
-+ /* Make sure forwarded skbs will have enough space
-+ * on Tx, if extra headers are added.
-+ */
-+ DPA_BUG_ON(fd_off != priv->rx_headroom);
-+ skb_reserve(skb, fd_off);
-+ skb_put(skb, qm_sg_entry_get_len(&sgt[i]));
-+ } else {
-+ /* Not the first S/G entry; all data from buffer will
-+ * be added in an skb fragment; fragment index is offset
-+ * by one since first S/G entry was incorporated in the
-+ * linear part of the skb.
-+ *
-+ * Caution: 'page' may be a tail page.
-+ */
-+ DPA_READ_SKB_PTR(skb_tmp, skbh, sg_vaddr, -1);
-+ page = virt_to_page(sg_vaddr);
-+ head_page = virt_to_head_page(sg_vaddr);
-+
-+ /* Free (only) the skbuff shell because its data buffer
-+ * is already a frag in the main skb.
-+ */
-+ get_page(head_page);
-+ dev_kfree_skb(skb_tmp);
-+
-+ /* Compute offset in (possibly tail) page */
-+ page_offset = ((unsigned long)sg_vaddr &
-+ (PAGE_SIZE - 1)) +
-+ (page_address(page) - page_address(head_page));
-+ /* page_offset only refers to the beginning of sgt[i];
-+ * but the buffer itself may have an internal offset.
-+ */
-+ frag_offset = qm_sg_entry_get_offset(&sgt[i]) +
-+ page_offset;
-+ frag_len = qm_sg_entry_get_len(&sgt[i]);
-+ /* skb_add_rx_frag() does no checking on the page; if
-+ * we pass it a tail page, we'll end up with
-+ * bad page accounting and eventually with segafults.
-+ */
-+ skb_add_rx_frag(skb, i - 1, head_page, frag_offset,
-+ frag_len, dpa_bp->size);
-+ }
-+ /* Update the pool count for the current {cpu x bpool} */
-+ (*count_ptr)--;
-+
-+ if (qm_sg_entry_get_final(&sgt[i]))
-+ break;
-+ }
-+ WARN_ONCE(i == DPA_SGT_MAX_ENTRIES, "No final bit on SGT\n");
-+
-+ /* recycle the SGT fragment */
-+ DPA_BUG_ON(dpa_bp != dpa_bpid2pool(fd->bpid));
-+ dpa_bp_recycle_frag(dpa_bp, (unsigned long)vaddr, count_ptr);
-+ return skb;
-+}
-+
-+#ifdef CONFIG_FSL_DPAA_DBG_LOOP
-+static inline int dpa_skb_loop(const struct dpa_priv_s *priv,
-+ struct sk_buff *skb)
-+{
-+ if (unlikely(priv->loop_to < 0))
-+ return 0; /* loop disabled by default */
-+
-+ skb_push(skb, ETH_HLEN); /* compensate for eth_type_trans */
-+ dpa_tx(skb, dpa_loop_netdevs[priv->loop_to]);
-+
-+ return 1; /* Frame Tx on the selected interface */
-+}
-+#endif
-+
-+void __hot _dpa_rx(struct net_device *net_dev,
-+ struct qman_portal *portal,
-+ const struct dpa_priv_s *priv,
-+ struct dpa_percpu_priv_s *percpu_priv,
-+ const struct qm_fd *fd,
-+ u32 fqid,
-+ int *count_ptr)
-+{
-+ struct dpa_bp *dpa_bp;
-+ struct sk_buff *skb;
-+ dma_addr_t addr = qm_fd_addr(fd);
-+ u32 fd_status = fd->status;
-+ unsigned int skb_len;
-+ struct rtnl_link_stats64 *percpu_stats = &percpu_priv->stats;
-+ int use_gro = net_dev->features & NETIF_F_GRO;
-+
-+ if (unlikely(fd_status & FM_FD_STAT_RX_ERRORS) != 0) {
-+ if (netif_msg_hw(priv) && net_ratelimit())
-+ netdev_warn(net_dev, "FD status = 0x%08x\n",
-+ fd_status & FM_FD_STAT_RX_ERRORS);
-+
-+ percpu_stats->rx_errors++;
-+ goto _release_frame;
-+ }
-+
-+ dpa_bp = priv->dpa_bp;
-+ DPA_BUG_ON(dpa_bp != dpa_bpid2pool(fd->bpid));
-+
-+ /* prefetch the first 64 bytes of the frame or the SGT start */
-+ dma_unmap_single(dpa_bp->dev, addr, dpa_bp->size, DMA_BIDIRECTIONAL);
-+ prefetch(phys_to_virt(addr) + dpa_fd_offset(fd));
-+
-+ /* The only FD types that we may receive are contig and S/G */
-+ DPA_BUG_ON((fd->format != qm_fd_contig) && (fd->format != qm_fd_sg));
-+
-+ if (likely(fd->format == qm_fd_contig)) {
-+#ifdef CONFIG_FSL_DPAA_HOOKS
-+ /* Execute the Rx processing hook, if it exists. */
-+ if (dpaa_eth_hooks.rx_default &&
-+ dpaa_eth_hooks.rx_default((void *)fd, net_dev,
-+ fqid) == DPAA_ETH_STOLEN) {
-+ /* won't count the rx bytes in */
-+ return;
-+ }
-+#endif
-+ skb = contig_fd_to_skb(priv, fd, &use_gro);
-+ } else {
-+ skb = sg_fd_to_skb(priv, fd, &use_gro, count_ptr);
-+ percpu_priv->rx_sg++;
-+ }
-+
-+ /* Account for either the contig buffer or the SGT buffer (depending on
-+ * which case we were in) having been removed from the pool.
-+ */
-+ (*count_ptr)--;
-+ skb->protocol = eth_type_trans(skb, net_dev);
-+
-+ skb_len = skb->len;
-+
-+#ifdef CONFIG_FSL_DPAA_DBG_LOOP
-+ if (dpa_skb_loop(priv, skb)) {
-+ percpu_stats->rx_packets++;
-+ percpu_stats->rx_bytes += skb_len;
-+ return;
-+ }
-+#endif
-+
-+ if (use_gro) {
-+ gro_result_t gro_result;
-+ const struct qman_portal_config *pc =
-+ qman_p_get_portal_config(portal);
-+ struct dpa_napi_portal *np = &percpu_priv->np[pc->index];
-+
-+ np->p = portal;
-+ gro_result = napi_gro_receive(&np->napi, skb);
-+ /* If frame is dropped by the stack, rx_dropped counter is
-+ * incremented automatically, so no need for us to update it
-+ */
-+ if (unlikely(gro_result == GRO_DROP))
-+ goto packet_dropped;
-+ } else if (unlikely(netif_receive_skb(skb) == NET_RX_DROP))
-+ goto packet_dropped;
-+
-+ percpu_stats->rx_packets++;
-+ percpu_stats->rx_bytes += skb_len;
-+
-+packet_dropped:
-+ return;
-+
-+_release_frame:
-+ dpa_fd_release(net_dev, fd);
-+}
-+
-+int __hot skb_to_contig_fd(struct dpa_priv_s *priv,
-+ struct sk_buff *skb, struct qm_fd *fd,
-+ int *count_ptr, int *offset)
-+{
-+ struct sk_buff **skbh;
-+ dma_addr_t addr;
-+ struct dpa_bp *dpa_bp = priv->dpa_bp;
-+ struct net_device *net_dev = priv->net_dev;
-+ int err;
-+ enum dma_data_direction dma_dir;
-+ unsigned char *buffer_start;
-+ int dma_map_size;
-+
-+#ifndef CONFIG_FSL_DPAA_TS
-+ /* Check recycling conditions; only if timestamp support is not
-+ * enabled, otherwise we need the fd back on tx confirmation
-+ */
-+
-+ /* We can recycle the buffer if:
-+ * - the pool is not full
-+ * - the buffer meets the skb recycling conditions
-+ * - the buffer meets our own (size, offset, align) conditions
-+ */
-+ if (likely((*count_ptr < dpa_bp->target_count) &&
-+ dpa_skb_is_recyclable(skb) &&
-+ dpa_buf_is_recyclable(skb, dpa_bp->size,
-+ priv->tx_headroom, &buffer_start))) {
-+ /* Buffer is recyclable; use the new start address
-+ * and set fd parameters and DMA mapping direction
-+ */
-+ fd->bpid = dpa_bp->bpid;
-+ DPA_BUG_ON(skb->data - buffer_start > DPA_MAX_FD_OFFSET);
-+ fd->offset = (uint16_t)(skb->data - buffer_start);
-+ dma_dir = DMA_BIDIRECTIONAL;
-+ dma_map_size = dpa_bp->size;
-+
-+ /* Store the skb back-pointer before the start of the buffer.
-+ * Otherwise it will be overwritten by the FMan.
-+ */
-+ DPA_WRITE_SKB_PTR(skb, skbh, buffer_start, -1);
-+ *offset = skb_headroom(skb) - fd->offset;
-+ } else
-+#endif
-+ {
-+ /* Not recyclable.
-+ * We are guaranteed to have at least tx_headroom bytes
-+ * available, so just use that for offset.
-+ */
-+ fd->bpid = 0xff;
-+ buffer_start = skb->data - priv->tx_headroom;
-+ fd->offset = priv->tx_headroom;
-+ dma_dir = DMA_TO_DEVICE;
-+ dma_map_size = skb_tail_pointer(skb) - buffer_start;
-+
-+ /* The buffer will be Tx-confirmed, but the TxConf cb must
-+ * necessarily look at our Tx private data to retrieve the
-+ * skbuff. Store the back-pointer inside the buffer.
-+ */
-+ DPA_WRITE_SKB_PTR(skb, skbh, buffer_start, 0);
-+ }
-+
-+ /* Enable L3/L4 hardware checksum computation.
-+ *
-+ * We must do this before dma_map_single(DMA_TO_DEVICE), because we may
-+ * need to write into the skb.
-+ */
-+ err = dpa_enable_tx_csum(priv, skb, fd,
-+ ((char *)skbh) + DPA_TX_PRIV_DATA_SIZE);
-+ if (unlikely(err < 0)) {
-+ if (netif_msg_tx_err(priv) && net_ratelimit())
-+ netdev_err(net_dev, "HW csum error: %d\n", err);
-+ return err;
-+ }
-+
-+ /* Fill in the rest of the FD fields */
-+ fd->format = qm_fd_contig;
-+ fd->length20 = skb->len;
-+ fd->cmd |= FM_FD_CMD_FCO;
-+
-+ /* Map the entire buffer size that may be seen by FMan, but no more */
-+ addr = dma_map_single(dpa_bp->dev, skbh, dma_map_size, dma_dir);
-+ if (unlikely(dma_mapping_error(dpa_bp->dev, addr))) {
-+ if (netif_msg_tx_err(priv) && net_ratelimit())
-+ netdev_err(net_dev, "dma_map_single() failed\n");
-+ return -EINVAL;
-+ }
-+ qm_fd_addr_set64(fd, addr);
-+
-+ return 0;
-+}
-+EXPORT_SYMBOL(skb_to_contig_fd);
-+
-+#ifndef CONFIG_PPC
-+/* Verify the conditions that trigger the A010022 errata: data unaligned to
-+ * 16 bytes and 4K memory address crossings.
-+ */
-+static bool a010022_check_skb(struct sk_buff *skb, struct dpa_priv_s *priv)
-+{
-+ int nr_frags, i = 0;
-+ skb_frag_t *frag;
-+
-+ /* Check if the headroom is aligned */
-+ if (((uintptr_t)skb->data - priv->tx_headroom) %
-+ priv->buf_layout[TX].data_align != 0)
-+ return true;
-+
-+ /* Check if the headroom crosses a boundary */
-+ if (HAS_DMA_ISSUE(skb->head, skb_headroom(skb)))
-+ return true;
-+
-+ /* Check if the non-paged data crosses a boundary */
-+ if (HAS_DMA_ISSUE(skb->data, skb_headlen(skb)))
-+ return true;
-+
-+ /* Check if the entire linear skb crosses a boundary */
-+ if (HAS_DMA_ISSUE(skb->head, skb_end_offset(skb)))
-+ return true;
-+
-+ nr_frags = skb_shinfo(skb)->nr_frags;
-+
-+ while (i < nr_frags) {
-+ frag = &skb_shinfo(skb)->frags[i];
-+
-+ /* Check if a paged fragment crosses a boundary from its
-+ * offset to its end.
-+ */
-+ if (HAS_DMA_ISSUE(frag->page_offset, frag->size))
-+ return true;
-+
-+ i++;
-+ }
-+
-+ return false;
-+}
-+
-+/* Realign the skb by copying its contents at the start of a newly allocated
-+ * page. Build a new skb around the new buffer and release the old one.
-+ * A performance drop should be expected.
-+ */
-+static struct sk_buff *a010022_realign_skb(struct sk_buff *skb,
-+ struct dpa_priv_s *priv)
-+{
-+ int trans_offset = skb_transport_offset(skb);
-+ int net_offset = skb_network_offset(skb);
-+ struct sk_buff *nskb = NULL;
-+ int nsize, headroom;
-+ struct page *npage;
-+ void *npage_addr;
-+
-+ /* Guarantee the minimum required headroom */
-+ if (skb_headroom(skb) >= priv->tx_headroom)
-+ headroom = skb_headroom(skb);
-+ else
-+ headroom = priv->tx_headroom;
-+
-+ npage = alloc_page(GFP_ATOMIC);
-+ if (unlikely(!npage)) {
-+ WARN_ONCE(1, "Memory allocation failure\n");
-+ return NULL;
-+ }
-+ npage_addr = page_address(npage);
-+
-+ /* For the new skb we only need the old one's data (both non-paged and
-+ * paged) and a headroom large enough to fit our private info. We can
-+ * skip the old tailroom.
-+ *
-+ * Make sure the new linearized buffer will not exceed a page's size.
-+ */
-+ nsize = headroom + skb->len +
-+ SKB_DATA_ALIGN(sizeof(struct skb_shared_info));
-+ if (unlikely(nsize > 4096))
-+ goto err;
-+
-+ nskb = build_skb(npage_addr, nsize);
-+ if (unlikely(!nskb))
-+ goto err;
-+
-+ /* Reserve only the needed headroom in order to guarantee the data's
-+ * alignment.
-+ * Code borrowed and adapted from skb_copy().
-+ */
-+ skb_reserve(nskb, priv->tx_headroom);
-+ skb_put(nskb, skb->len);
-+ if (skb_copy_bits(skb, 0, nskb->data, skb->len)) {
-+ WARN_ONCE(1, "skb parsing failure\n");
-+ goto err;
-+ }
-+ copy_skb_header(nskb, skb);
-+
-+#ifdef CONFIG_FSL_DPAA_TS
-+ /* Copy relevant timestamp info from the old skb to the new */
-+ if (priv->ts_tx_en) {
-+ skb_shinfo(nskb)->tx_flags = skb_shinfo(skb)->tx_flags;
-+ skb_shinfo(nskb)->hwtstamps = skb_shinfo(skb)->hwtstamps;
-+ skb_shinfo(nskb)->tskey = skb_shinfo(skb)->tskey;
-+ if (skb->sk)
-+ skb_set_owner_w(nskb, skb->sk);
-+ }
-+#endif
-+ /* We move the headroom when we align it so we have to reset the
-+ * network and transport header offsets relative to the new data
-+ * pointer. The checksum offload relies on these offsets.
-+ */
-+ skb_set_network_header(nskb, net_offset);
-+ skb_set_transport_header(nskb, trans_offset);
-+
-+ /* We don't want the buffer to be recycled so we mark it accordingly */
-+ nskb->mark = NONREC_MARK;
-+
-+ dev_kfree_skb(skb);
-+ return nskb;
-+
-+err:
-+ if (nskb)
-+ dev_kfree_skb(nskb);
-+ put_page(npage);
-+ return NULL;
-+}
-+#endif
-+
-+int __hot skb_to_sg_fd(struct dpa_priv_s *priv,
-+ struct sk_buff *skb, struct qm_fd *fd)
-+{
-+ struct dpa_bp *dpa_bp = priv->dpa_bp;
-+ dma_addr_t addr;
-+ dma_addr_t sg_addr;
-+ struct sk_buff **skbh;
-+ struct net_device *net_dev = priv->net_dev;
-+ int sg_len, sgt_size;
-+ int err;
-+
-+ struct qm_sg_entry *sgt;
-+ void *sgt_buf;
-+ skb_frag_t *frag;
-+ int i = 0, j = 0;
-+ int nr_frags;
-+ const enum dma_data_direction dma_dir = DMA_TO_DEVICE;
-+
-+ nr_frags = skb_shinfo(skb)->nr_frags;
-+ fd->format = qm_fd_sg;
-+
-+ sgt_size = sizeof(struct qm_sg_entry) * (1 + nr_frags);
-+
-+ /* Get a page frag to store the SGTable, or a full page if the errata
-+ * is in place and we need to avoid crossing a 4k boundary.
-+ */
-+#ifndef CONFIG_PPC
-+ if (unlikely(dpaa_errata_a010022))
-+ sgt_buf = page_address(alloc_page(GFP_ATOMIC));
-+ else
-+#endif
-+ sgt_buf = netdev_alloc_frag(priv->tx_headroom + sgt_size);
-+ if (unlikely(!sgt_buf)) {
-+ dev_err(dpa_bp->dev, "netdev_alloc_frag() failed\n");
-+ return -ENOMEM;
-+ }
-+
-+ /* it seems that the memory allocator does not zero the allocated mem */
-+ memset(sgt_buf, 0, priv->tx_headroom + sgt_size);
-+
-+ /* Enable L3/L4 hardware checksum computation.
-+ *
-+ * We must do this before dma_map_single(DMA_TO_DEVICE), because we may
-+ * need to write into the skb.
-+ */
-+ err = dpa_enable_tx_csum(priv, skb, fd,
-+ sgt_buf + DPA_TX_PRIV_DATA_SIZE);
-+ if (unlikely(err < 0)) {
-+ if (netif_msg_tx_err(priv) && net_ratelimit())
-+ netdev_err(net_dev, "HW csum error: %d\n", err);
-+ goto csum_failed;
-+ }
-+
-+ /* Assign the data from skb->data to the first SG list entry */
-+ sgt = (struct qm_sg_entry *)(sgt_buf + priv->tx_headroom);
-+ sg_len = skb_headlen(skb);
-+ qm_sg_entry_set_bpid(&sgt[0], 0xff);
-+ qm_sg_entry_set_offset(&sgt[0], 0);
-+ qm_sg_entry_set_len(&sgt[0], sg_len);
-+ qm_sg_entry_set_ext(&sgt[0], 0);
-+ qm_sg_entry_set_final(&sgt[0], 0);
-+
-+ addr = dma_map_single(dpa_bp->dev, skb->data, sg_len, dma_dir);
-+ if (unlikely(dma_mapping_error(dpa_bp->dev, addr))) {
-+ dev_err(dpa_bp->dev, "DMA mapping failed");
-+ err = -EINVAL;
-+ goto sg0_map_failed;
-+ }
-+
-+ qm_sg_entry_set64(&sgt[0], addr);
-+
-+ /* populate the rest of SGT entries */
-+ for (i = 1; i <= nr_frags; i++) {
-+ frag = &skb_shinfo(skb)->frags[i - 1];
-+ qm_sg_entry_set_bpid(&sgt[i], 0xff);
-+ qm_sg_entry_set_offset(&sgt[i], 0);
-+ qm_sg_entry_set_len(&sgt[i], frag->size);
-+ qm_sg_entry_set_ext(&sgt[i], 0);
-+
-+ if (i == nr_frags)
-+ qm_sg_entry_set_final(&sgt[i], 1);
-+ else
-+ qm_sg_entry_set_final(&sgt[i], 0);
-+
-+ DPA_BUG_ON(!skb_frag_page(frag));
-+ addr = skb_frag_dma_map(dpa_bp->dev, frag, 0, frag->size,
-+ dma_dir);
-+ if (unlikely(dma_mapping_error(dpa_bp->dev, addr))) {
-+ dev_err(dpa_bp->dev, "DMA mapping failed");
-+ err = -EINVAL;
-+ goto sg_map_failed;
-+ }
-+
-+ /* keep the offset in the address */
-+ qm_sg_entry_set64(&sgt[i], addr);
-+ }
-+
-+ fd->length20 = skb->len;
-+ fd->offset = priv->tx_headroom;
-+
-+ /* DMA map the SGT page
-+ *
-+ * It's safe to store the skb back-pointer inside the buffer since
-+ * S/G frames are non-recyclable.
-+ */
-+ DPA_WRITE_SKB_PTR(skb, skbh, sgt_buf, 0);
-+ addr = dma_map_single(dpa_bp->dev, sgt_buf,
-+ priv->tx_headroom + sgt_size,
-+ dma_dir);
-+
-+ if (unlikely(dma_mapping_error(dpa_bp->dev, addr))) {
-+ dev_err(dpa_bp->dev, "DMA mapping failed");
-+ err = -EINVAL;
-+ goto sgt_map_failed;
-+ }
-+
-+ qm_fd_addr_set64(fd, addr);
-+ fd->bpid = 0xff;
-+ fd->cmd |= FM_FD_CMD_FCO;
-+
-+ return 0;
-+
-+sgt_map_failed:
-+sg_map_failed:
-+ for (j = 0; j < i; j++) {
-+ sg_addr = qm_sg_addr(&sgt[j]);
-+ dma_unmap_page(dpa_bp->dev, sg_addr,
-+ qm_sg_entry_get_len(&sgt[j]), dma_dir);
-+ }
-+sg0_map_failed:
-+csum_failed:
-+ put_page(virt_to_head_page(sgt_buf));
-+
-+ return err;
-+}
-+EXPORT_SYMBOL(skb_to_sg_fd);
-+
-+int __hot dpa_tx(struct sk_buff *skb, struct net_device *net_dev)
-+{
-+ struct dpa_priv_s *priv;
-+ const int queue_mapping = dpa_get_queue_mapping(skb);
-+ struct qman_fq *egress_fq, *conf_fq;
-+
-+#ifdef CONFIG_FSL_DPAA_HOOKS
-+ /* If there is a Tx hook, run it. */
-+ if (dpaa_eth_hooks.tx &&
-+ dpaa_eth_hooks.tx(skb, net_dev) == DPAA_ETH_STOLEN)
-+ /* won't update any Tx stats */
-+ return NETDEV_TX_OK;
-+#endif
-+
-+ priv = netdev_priv(net_dev);
-+
-+#ifdef CONFIG_FSL_DPAA_CEETM
-+ if (priv->ceetm_en)
-+ return ceetm_tx(skb, net_dev);
-+#endif
-+
-+ egress_fq = priv->egress_fqs[queue_mapping];
-+ conf_fq = priv->conf_fqs[queue_mapping];
-+
-+ return dpa_tx_extended(skb, net_dev, egress_fq, conf_fq);
-+}
-+
-+int __hot dpa_tx_extended(struct sk_buff *skb, struct net_device *net_dev,
-+ struct qman_fq *egress_fq, struct qman_fq *conf_fq)
-+{
-+ struct dpa_priv_s *priv;
-+ struct qm_fd fd;
-+ struct dpa_percpu_priv_s *percpu_priv;
-+ struct rtnl_link_stats64 *percpu_stats;
-+ int err = 0;
-+ bool nonlinear;
-+ int *countptr, offset = 0;
-+
-+ priv = netdev_priv(net_dev);
-+ /* Non-migratable context, safe to use raw_cpu_ptr */
-+ percpu_priv = raw_cpu_ptr(priv->percpu_priv);
-+ percpu_stats = &percpu_priv->stats;
-+ countptr = raw_cpu_ptr(priv->dpa_bp->percpu_count);
-+
-+ clear_fd(&fd);
-+
-+#ifndef CONFIG_PPC
-+ if (unlikely(dpaa_errata_a010022) && a010022_check_skb(skb, priv)) {
-+ skb = a010022_realign_skb(skb, priv);
-+ if (!skb)
-+ goto skb_to_fd_failed;
-+ }
-+#endif
-+
-+ nonlinear = skb_is_nonlinear(skb);
-+
-+#ifdef CONFIG_FSL_DPAA_1588
-+ if (priv->tsu && priv->tsu->valid && priv->tsu->hwts_tx_en_ioctl)
-+ fd.cmd |= FM_FD_CMD_UPD;
-+#endif
-+#ifdef CONFIG_FSL_DPAA_TS
-+ if (unlikely(priv->ts_tx_en &&
-+ skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP))
-+ fd.cmd |= FM_FD_CMD_UPD;
-+ skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS;
-+#endif /* CONFIG_FSL_DPAA_TS */
-+
-+ /* MAX_SKB_FRAGS is larger than our DPA_SGT_MAX_ENTRIES; make sure
-+ * we don't feed FMan with more fragments than it supports.
-+ * Btw, we're using the first sgt entry to store the linear part of
-+ * the skb, so we're one extra frag short.
-+ */
-+ if (nonlinear &&
-+ likely(skb_shinfo(skb)->nr_frags < DPA_SGT_MAX_ENTRIES)) {
-+ /* Just create a S/G fd based on the skb */
-+ err = skb_to_sg_fd(priv, skb, &fd);
-+ percpu_priv->tx_frag_skbuffs++;
-+ } else {
-+ /* Make sure we have enough headroom to accommodate private
-+ * data, parse results, etc. Normally this shouldn't happen if
-+ * we're here via the standard kernel stack.
-+ */
-+ if (unlikely(skb_headroom(skb) < priv->tx_headroom)) {
-+ struct sk_buff *skb_new;
-+
-+ skb_new = skb_realloc_headroom(skb, priv->tx_headroom);
-+ if (unlikely(!skb_new)) {
-+ dev_kfree_skb(skb);
-+ percpu_stats->tx_errors++;
-+ return NETDEV_TX_OK;
-+ }
-+
-+ /* propagate the skb ownership information */
-+ if (skb->sk)
-+ skb_set_owner_w(skb_new, skb->sk);
-+
-+ dev_kfree_skb(skb);
-+ skb = skb_new;
-+ }
-+
-+ /* We're going to store the skb backpointer at the beginning
-+ * of the data buffer, so we need a privately owned skb
-+ */
-+
-+ /* Code borrowed from skb_unshare(). */
-+ if (skb_cloned(skb)) {
-+ struct sk_buff *nskb = skb_copy(skb, GFP_ATOMIC);
-+ kfree_skb(skb);
-+ skb = nskb;
-+#ifndef CONFIG_PPC
-+ if (unlikely(dpaa_errata_a010022) &&
-+ a010022_check_skb(skb, priv)) {
-+ skb = a010022_realign_skb(skb, priv);
-+ if (!skb)
-+ goto skb_to_fd_failed;
-+ }
-+#endif
-+ /* skb_copy() has now linearized the skbuff. */
-+ } else if (unlikely(nonlinear)) {
-+ /* We are here because the egress skb contains
-+ * more fragments than we support. In this case,
-+ * we have no choice but to linearize it ourselves.
-+ */
-+ err = __skb_linearize(skb);
-+ }
-+ if (unlikely(!skb || err < 0))
-+ /* Common out-of-memory error path */
-+ goto enomem;
-+
-+ err = skb_to_contig_fd(priv, skb, &fd, countptr, &offset);
-+ }
-+ if (unlikely(err < 0))
-+ goto skb_to_fd_failed;
-+
-+ if (fd.bpid != 0xff) {
-+ skb_recycle(skb);
-+ /* skb_recycle() reserves NET_SKB_PAD as skb headroom,
-+ * but we need the skb to look as if returned by build_skb().
-+ * We need to manually adjust the tailptr as well.
-+ */
-+ skb->data = skb->head + offset;
-+ skb_reset_tail_pointer(skb);
-+
-+ (*countptr)++;
-+ percpu_priv->tx_returned++;
-+ }
-+
-+ if (unlikely(dpa_xmit(priv, percpu_stats, &fd, egress_fq, conf_fq) < 0))
-+ goto xmit_failed;
-+
-+ netif_trans_update(net_dev);
-+ return NETDEV_TX_OK;
-+
-+xmit_failed:
-+ if (fd.bpid != 0xff) {
-+ (*countptr)--;
-+ percpu_priv->tx_returned--;
-+ dpa_fd_release(net_dev, &fd);
-+ percpu_stats->tx_errors++;
-+ return NETDEV_TX_OK;
-+ }
-+ _dpa_cleanup_tx_fd(priv, &fd);
-+skb_to_fd_failed:
-+enomem:
-+ percpu_stats->tx_errors++;
-+ dev_kfree_skb(skb);
-+ return NETDEV_TX_OK;
-+}
-+EXPORT_SYMBOL(dpa_tx_extended);
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_dpaa/dpaa_eth_sysfs.c
-@@ -0,0 +1,278 @@
-+/* Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#include <linux/init.h>
-+#include <linux/module.h>
-+#include <linux/kthread.h>
-+#include <linux/io.h>
-+#include <linux/of_net.h>
-+#include "dpaa_eth.h"
-+#include "mac.h" /* struct mac_device */
-+#ifdef CONFIG_FSL_DPAA_1588
-+#include "dpaa_1588.h"
-+#endif
-+
-+static ssize_t dpaa_eth_show_addr(struct device *dev,
-+ struct device_attribute *attr, char *buf)
-+{
-+ struct dpa_priv_s *priv = netdev_priv(to_net_dev(dev));
-+ struct mac_device *mac_dev = priv->mac_dev;
-+
-+ if (mac_dev)
-+ return sprintf(buf, "%llx",
-+ (unsigned long long)mac_dev->res->start);
-+ else
-+ return sprintf(buf, "none");
-+}
-+
-+static ssize_t dpaa_eth_show_type(struct device *dev,
-+ struct device_attribute *attr, char *buf)
-+{
-+ struct dpa_priv_s *priv = netdev_priv(to_net_dev(dev));
-+ ssize_t res = 0;
-+
-+ if (priv)
-+ res = sprintf(buf, "%s", priv->if_type);
-+
-+ return res;
-+}
-+
-+static ssize_t dpaa_eth_show_fqids(struct device *dev,
-+ struct device_attribute *attr, char *buf)
-+{
-+ struct dpa_priv_s *priv = netdev_priv(to_net_dev(dev));
-+ ssize_t bytes = 0;
-+ int i = 0;
-+ char *str;
-+ struct dpa_fq *fq;
-+ struct dpa_fq *tmp;
-+ struct dpa_fq *prev = NULL;
-+ u32 first_fqid = 0;
-+ u32 last_fqid = 0;
-+ char *prevstr = NULL;
-+
-+ list_for_each_entry_safe(fq, tmp, &priv->dpa_fq_list, list) {
-+ switch (fq->fq_type) {
-+ case FQ_TYPE_RX_DEFAULT:
-+ str = "Rx default";
-+ break;
-+ case FQ_TYPE_RX_ERROR:
-+ str = "Rx error";
-+ break;
-+ case FQ_TYPE_RX_PCD:
-+ str = "Rx PCD";
-+ break;
-+ case FQ_TYPE_TX_CONFIRM:
-+ str = "Tx default confirmation";
-+ break;
-+ case FQ_TYPE_TX_CONF_MQ:
-+ str = "Tx confirmation (mq)";
-+ break;
-+ case FQ_TYPE_TX_ERROR:
-+ str = "Tx error";
-+ break;
-+ case FQ_TYPE_TX:
-+ str = "Tx";
-+ break;
-+ case FQ_TYPE_RX_PCD_HI_PRIO:
-+ str ="Rx PCD High Priority";
-+ break;
-+ default:
-+ str = "Unknown";
-+ }
-+
-+ if (prev && (abs(fq->fqid - prev->fqid) != 1 ||
-+ str != prevstr)) {
-+ if (last_fqid == first_fqid)
-+ bytes += sprintf(buf + bytes,
-+ "%s: %d\n", prevstr, prev->fqid);
-+ else
-+ bytes += sprintf(buf + bytes,
-+ "%s: %d - %d\n", prevstr,
-+ first_fqid, last_fqid);
-+ }
-+
-+ if (prev && abs(fq->fqid - prev->fqid) == 1 && str == prevstr)
-+ last_fqid = fq->fqid;
-+ else
-+ first_fqid = last_fqid = fq->fqid;
-+
-+ prev = fq;
-+ prevstr = str;
-+ i++;
-+ }
-+
-+ if (prev) {
-+ if (last_fqid == first_fqid)
-+ bytes += sprintf(buf + bytes, "%s: %d\n", prevstr,
-+ prev->fqid);
-+ else
-+ bytes += sprintf(buf + bytes, "%s: %d - %d\n", prevstr,
-+ first_fqid, last_fqid);
-+ }
-+
-+ return bytes;
-+}
-+
-+static ssize_t dpaa_eth_show_bpids(struct device *dev,
-+ struct device_attribute *attr, char *buf)
-+{
-+ ssize_t bytes = 0;
-+ struct dpa_priv_s *priv = netdev_priv(to_net_dev(dev));
-+ struct dpa_bp *dpa_bp = priv->dpa_bp;
-+ int i = 0;
-+
-+ for (i = 0; i < priv->bp_count; i++)
-+ bytes += snprintf(buf + bytes, PAGE_SIZE, "%u\n",
-+ dpa_bp[i].bpid);
-+
-+ return bytes;
-+}
-+
-+static ssize_t dpaa_eth_show_mac_regs(struct device *dev,
-+ struct device_attribute *attr, char *buf)
-+{
-+ struct dpa_priv_s *priv = netdev_priv(to_net_dev(dev));
-+ struct mac_device *mac_dev = priv->mac_dev;
-+ int n = 0;
-+
-+ if (mac_dev)
-+ n = fm_mac_dump_regs(mac_dev, buf, n);
-+ else
-+ return sprintf(buf, "no mac registers\n");
-+
-+ return n;
-+}
-+
-+static ssize_t dpaa_eth_show_mac_rx_stats(struct device *dev,
-+ struct device_attribute *attr, char *buf)
-+{
-+ struct dpa_priv_s *priv = netdev_priv(to_net_dev(dev));
-+ struct mac_device *mac_dev = priv->mac_dev;
-+ int n = 0;
-+
-+ if (mac_dev)
-+ n = fm_mac_dump_rx_stats(mac_dev, buf, n);
-+ else
-+ return sprintf(buf, "no mac rx stats\n");
-+
-+ return n;
-+}
-+
-+static ssize_t dpaa_eth_show_mac_tx_stats(struct device *dev,
-+ struct device_attribute *attr, char *buf)
-+{
-+ struct dpa_priv_s *priv = netdev_priv(to_net_dev(dev));
-+ struct mac_device *mac_dev = priv->mac_dev;
-+ int n = 0;
-+
-+ if (mac_dev)
-+ n = fm_mac_dump_tx_stats(mac_dev, buf, n);
-+ else
-+ return sprintf(buf, "no mac tx stats\n");
-+
-+ return n;
-+}
-+
-+#ifdef CONFIG_FSL_DPAA_1588
-+static ssize_t dpaa_eth_show_ptp_1588(struct device *dev,
-+ struct device_attribute *attr, char *buf)
-+{
-+ struct dpa_priv_s *priv = netdev_priv(to_net_dev(dev));
-+
-+ if (priv->tsu && priv->tsu->valid)
-+ return sprintf(buf, "1\n");
-+ else
-+ return sprintf(buf, "0\n");
-+}
-+
-+static ssize_t dpaa_eth_set_ptp_1588(struct device *dev,
-+ struct device_attribute *attr,
-+ const char *buf, size_t count)
-+{
-+ struct dpa_priv_s *priv = netdev_priv(to_net_dev(dev));
-+ unsigned int num;
-+ unsigned long flags;
-+
-+ if (kstrtouint(buf, 0, &num) < 0)
-+ return -EINVAL;
-+
-+ local_irq_save(flags);
-+
-+ if (num) {
-+ if (priv->tsu)
-+ priv->tsu->valid = TRUE;
-+ } else {
-+ if (priv->tsu)
-+ priv->tsu->valid = FALSE;
-+ }
-+
-+ local_irq_restore(flags);
-+
-+ return count;
-+}
-+#endif
-+
-+static struct device_attribute dpaa_eth_attrs[] = {
-+ __ATTR(device_addr, S_IRUGO, dpaa_eth_show_addr, NULL),
-+ __ATTR(device_type, S_IRUGO, dpaa_eth_show_type, NULL),
-+ __ATTR(fqids, S_IRUGO, dpaa_eth_show_fqids, NULL),
-+ __ATTR(bpids, S_IRUGO, dpaa_eth_show_bpids, NULL),
-+ __ATTR(mac_regs, S_IRUGO, dpaa_eth_show_mac_regs, NULL),
-+ __ATTR(mac_rx_stats, S_IRUGO, dpaa_eth_show_mac_rx_stats, NULL),
-+ __ATTR(mac_tx_stats, S_IRUGO, dpaa_eth_show_mac_tx_stats, NULL),
-+#ifdef CONFIG_FSL_DPAA_1588
-+ __ATTR(ptp_1588, S_IRUGO | S_IWUSR, dpaa_eth_show_ptp_1588,
-+ dpaa_eth_set_ptp_1588),
-+#endif
-+};
-+
-+void dpaa_eth_sysfs_init(struct device *dev)
-+{
-+ int i;
-+
-+ for (i = 0; i < ARRAY_SIZE(dpaa_eth_attrs); i++)
-+ if (device_create_file(dev, &dpaa_eth_attrs[i])) {
-+ dev_err(dev, "Error creating sysfs file\n");
-+ while (i > 0)
-+ device_remove_file(dev, &dpaa_eth_attrs[--i]);
-+ return;
-+ }
-+}
-+EXPORT_SYMBOL(dpaa_eth_sysfs_init);
-+
-+void dpaa_eth_sysfs_remove(struct device *dev)
-+{
-+ int i;
-+
-+ for (i = 0; i < ARRAY_SIZE(dpaa_eth_attrs); i++)
-+ device_remove_file(dev, &dpaa_eth_attrs[i]);
-+}
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_dpaa/dpaa_eth_trace.h
-@@ -0,0 +1,144 @@
-+/* Copyright 2013 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#undef TRACE_SYSTEM
-+#define TRACE_SYSTEM dpaa_eth
-+
-+#if !defined(_DPAA_ETH_TRACE_H) || defined(TRACE_HEADER_MULTI_READ)
-+#define _DPAA_ETH_TRACE_H
-+
-+#include <linux/skbuff.h>
-+#include <linux/netdevice.h>
-+#include "dpaa_eth.h"
-+#include <linux/tracepoint.h>
-+
-+#define fd_format_name(format) { qm_fd_##format, #format }
-+#define fd_format_list \
-+ fd_format_name(contig), \
-+ fd_format_name(sg)
-+#define TR_FMT "[%s] fqid=%d, fd: addr=0x%llx, format=%s, off=%u, len=%u," \
-+ " status=0x%08x"
-+
-+/* This is used to declare a class of events.
-+ * individual events of this type will be defined below.
-+ */
-+
-+/* Store details about a frame descriptor and the FQ on which it was
-+ * transmitted/received.
-+ */
-+DECLARE_EVENT_CLASS(dpaa_eth_fd,
-+ /* Trace function prototype */
-+ TP_PROTO(struct net_device *netdev,
-+ struct qman_fq *fq,
-+ const struct qm_fd *fd),
-+
-+ /* Repeat argument list here */
-+ TP_ARGS(netdev, fq, fd),
-+
-+ /* A structure containing the relevant information we want to record.
-+ * Declare name and type for each normal element, name, type and size
-+ * for arrays. Use __string for variable length strings.
-+ */
-+ TP_STRUCT__entry(
-+ __field(u32, fqid)
-+ __field(u64, fd_addr)
-+ __field(u8, fd_format)
-+ __field(u16, fd_offset)
-+ __field(u32, fd_length)
-+ __field(u32, fd_status)
-+ __string(name, netdev->name)
-+ ),
-+
-+ /* The function that assigns values to the above declared fields */
-+ TP_fast_assign(
-+ __entry->fqid = fq->fqid;
-+ __entry->fd_addr = qm_fd_addr_get64(fd);
-+ __entry->fd_format = fd->format;
-+ __entry->fd_offset = dpa_fd_offset(fd);
-+ __entry->fd_length = dpa_fd_length(fd);
-+ __entry->fd_status = fd->status;
-+ __assign_str(name, netdev->name);
-+ ),
-+
-+ /* This is what gets printed when the trace event is triggered */
-+ /* TODO: print the status using __print_flags() */
-+ TP_printk(TR_FMT,
-+ __get_str(name), __entry->fqid, __entry->fd_addr,
-+ __print_symbolic(__entry->fd_format, fd_format_list),
-+ __entry->fd_offset, __entry->fd_length, __entry->fd_status)
-+);
-+
-+/* Now declare events of the above type. Format is:
-+ * DEFINE_EVENT(class, name, proto, args), with proto and args same as for class
-+ */
-+
-+/* Tx (egress) fd */
-+DEFINE_EVENT(dpaa_eth_fd, dpa_tx_fd,
-+
-+ TP_PROTO(struct net_device *netdev,
-+ struct qman_fq *fq,
-+ const struct qm_fd *fd),
-+
-+ TP_ARGS(netdev, fq, fd)
-+);
-+
-+/* Rx fd */
-+DEFINE_EVENT(dpaa_eth_fd, dpa_rx_fd,
-+
-+ TP_PROTO(struct net_device *netdev,
-+ struct qman_fq *fq,
-+ const struct qm_fd *fd),
-+
-+ TP_ARGS(netdev, fq, fd)
-+);
-+
-+/* Tx confirmation fd */
-+DEFINE_EVENT(dpaa_eth_fd, dpa_tx_conf_fd,
-+
-+ TP_PROTO(struct net_device *netdev,
-+ struct qman_fq *fq,
-+ const struct qm_fd *fd),
-+
-+ TP_ARGS(netdev, fq, fd)
-+);
-+
-+/* If only one event of a certain type needs to be declared, use TRACE_EVENT().
-+ * The syntax is the same as for DECLARE_EVENT_CLASS().
-+ */
-+
-+#endif /* _DPAA_ETH_TRACE_H */
-+
-+/* This must be outside ifdef _DPAA_ETH_TRACE_H */
-+#undef TRACE_INCLUDE_PATH
-+#define TRACE_INCLUDE_PATH .
-+#undef TRACE_INCLUDE_FILE
-+#define TRACE_INCLUDE_FILE dpaa_eth_trace
-+#include <trace/define_trace.h>
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_dpaa/dpaa_ethtool.c
-@@ -0,0 +1,544 @@
-+/* Copyright 2008-2012 Freescale Semiconductor, Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#ifdef CONFIG_FSL_DPAA_ETH_DEBUG
-+#define pr_fmt(fmt) \
-+ KBUILD_MODNAME ": %s:%hu:%s() " fmt, \
-+ KBUILD_BASENAME".c", __LINE__, __func__
-+#else
-+#define pr_fmt(fmt) \
-+ KBUILD_MODNAME ": " fmt
-+#endif
-+
-+#include <linux/string.h>
-+
-+#include "dpaa_eth.h"
-+#include "mac.h" /* struct mac_device */
-+#include "dpaa_eth_common.h"
-+
-+static const char dpa_stats_percpu[][ETH_GSTRING_LEN] = {
-+ "interrupts",
-+ "rx packets",
-+ "tx packets",
-+ "tx recycled",
-+ "tx confirm",
-+ "tx S/G",
-+ "rx S/G",
-+ "tx error",
-+ "rx error",
-+ "bp count"
-+};
-+
-+static char dpa_stats_global[][ETH_GSTRING_LEN] = {
-+ /* dpa rx errors */
-+ "rx dma error",
-+ "rx frame physical error",
-+ "rx frame size error",
-+ "rx header error",
-+ "rx csum error",
-+
-+ /* demultiplexing errors */
-+ "qman cg_tdrop",
-+ "qman wred",
-+ "qman error cond",
-+ "qman early window",
-+ "qman late window",
-+ "qman fq tdrop",
-+ "qman fq retired",
-+ "qman orp disabled",
-+
-+ /* congestion related stats */
-+ "congestion time (ms)",
-+ "entered congestion",
-+ "congested (0/1)"
-+};
-+
-+#define DPA_STATS_PERCPU_LEN ARRAY_SIZE(dpa_stats_percpu)
-+#define DPA_STATS_GLOBAL_LEN ARRAY_SIZE(dpa_stats_global)
-+
-+static int __cold dpa_get_settings(struct net_device *net_dev,
-+ struct ethtool_cmd *et_cmd)
-+{
-+ int _errno;
-+ struct dpa_priv_s *priv;
-+
-+ priv = netdev_priv(net_dev);
-+
-+ if (priv->mac_dev == NULL) {
-+ netdev_info(net_dev, "This is a MAC-less interface\n");
-+ return -ENODEV;
-+ }
-+ if (unlikely(priv->mac_dev->phy_dev == NULL)) {
-+ netdev_dbg(net_dev, "phy device not initialized\n");
-+ return 0;
-+ }
-+
-+ _errno = phy_ethtool_gset(priv->mac_dev->phy_dev, et_cmd);
-+ if (unlikely(_errno < 0))
-+ netdev_err(net_dev, "phy_ethtool_gset() = %d\n", _errno);
-+
-+ return _errno;
-+}
-+
-+static int __cold dpa_set_settings(struct net_device *net_dev,
-+ struct ethtool_cmd *et_cmd)
-+{
-+ int _errno;
-+ struct dpa_priv_s *priv;
-+
-+ priv = netdev_priv(net_dev);
-+
-+ if (priv->mac_dev == NULL) {
-+ netdev_info(net_dev, "This is a MAC-less interface\n");
-+ return -ENODEV;
-+ }
-+ if (unlikely(priv->mac_dev->phy_dev == NULL)) {
-+ netdev_err(net_dev, "phy device not initialized\n");
-+ return -ENODEV;
-+ }
-+
-+ _errno = phy_ethtool_sset(priv->mac_dev->phy_dev, et_cmd);
-+ if (unlikely(_errno < 0))
-+ netdev_err(net_dev, "phy_ethtool_sset() = %d\n", _errno);
-+
-+ return _errno;
-+}
-+
-+static void __cold dpa_get_drvinfo(struct net_device *net_dev,
-+ struct ethtool_drvinfo *drvinfo)
-+{
-+ int _errno;
-+
-+ strncpy(drvinfo->driver, KBUILD_MODNAME,
-+ sizeof(drvinfo->driver) - 1)[sizeof(drvinfo->driver)-1] = 0;
-+ _errno = snprintf(drvinfo->fw_version, sizeof(drvinfo->fw_version),
-+ "%X", 0);
-+
-+ if (unlikely(_errno >= sizeof(drvinfo->fw_version))) {
-+ /* Truncated output */
-+ netdev_notice(net_dev, "snprintf() = %d\n", _errno);
-+ } else if (unlikely(_errno < 0)) {
-+ netdev_warn(net_dev, "snprintf() = %d\n", _errno);
-+ memset(drvinfo->fw_version, 0, sizeof(drvinfo->fw_version));
-+ }
-+ strncpy(drvinfo->bus_info, dev_name(net_dev->dev.parent->parent),
-+ sizeof(drvinfo->bus_info)-1)[sizeof(drvinfo->bus_info)-1] = 0;
-+}
-+
-+static uint32_t __cold dpa_get_msglevel(struct net_device *net_dev)
-+{
-+ return ((struct dpa_priv_s *)netdev_priv(net_dev))->msg_enable;
-+}
-+
-+static void __cold dpa_set_msglevel(struct net_device *net_dev,
-+ uint32_t msg_enable)
-+{
-+ ((struct dpa_priv_s *)netdev_priv(net_dev))->msg_enable = msg_enable;
-+}
-+
-+static int __cold dpa_nway_reset(struct net_device *net_dev)
-+{
-+ int _errno;
-+ struct dpa_priv_s *priv;
-+
-+ priv = netdev_priv(net_dev);
-+
-+ if (priv->mac_dev == NULL) {
-+ netdev_info(net_dev, "This is a MAC-less interface\n");
-+ return -ENODEV;
-+ }
-+ if (unlikely(priv->mac_dev->phy_dev == NULL)) {
-+ netdev_err(net_dev, "phy device not initialized\n");
-+ return -ENODEV;
-+ }
-+
-+ _errno = 0;
-+ if (priv->mac_dev->phy_dev->autoneg) {
-+ _errno = phy_start_aneg(priv->mac_dev->phy_dev);
-+ if (unlikely(_errno < 0))
-+ netdev_err(net_dev, "phy_start_aneg() = %d\n",
-+ _errno);
-+ }
-+
-+ return _errno;
-+}
-+
-+static void __cold dpa_get_pauseparam(struct net_device *net_dev,
-+ struct ethtool_pauseparam *epause)
-+{
-+ struct dpa_priv_s *priv;
-+ struct mac_device *mac_dev;
-+ struct phy_device *phy_dev;
-+
-+ priv = netdev_priv(net_dev);
-+ mac_dev = priv->mac_dev;
-+
-+ if (mac_dev == NULL) {
-+ netdev_info(net_dev, "This is a MAC-less interface\n");
-+ return;
-+ }
-+
-+ phy_dev = mac_dev->phy_dev;
-+ if (unlikely(phy_dev == NULL)) {
-+ netdev_err(net_dev, "phy device not initialized\n");
-+ return;
-+ }
-+
-+ epause->autoneg = mac_dev->autoneg_pause;
-+ epause->rx_pause = mac_dev->rx_pause_active;
-+ epause->tx_pause = mac_dev->tx_pause_active;
-+}
-+
-+static int __cold dpa_set_pauseparam(struct net_device *net_dev,
-+ struct ethtool_pauseparam *epause)
-+{
-+ struct dpa_priv_s *priv;
-+ struct mac_device *mac_dev;
-+ struct phy_device *phy_dev;
-+ int _errno;
-+ u32 newadv, oldadv;
-+ bool rx_pause, tx_pause;
-+
-+ priv = netdev_priv(net_dev);
-+ mac_dev = priv->mac_dev;
-+
-+ if (mac_dev == NULL) {
-+ netdev_info(net_dev, "This is a MAC-less interface\n");
-+ return -ENODEV;
-+ }
-+
-+ phy_dev = mac_dev->phy_dev;
-+ if (unlikely(phy_dev == NULL)) {
-+ netdev_err(net_dev, "phy device not initialized\n");
-+ return -ENODEV;
-+ }
-+
-+ if (!(phy_dev->supported & SUPPORTED_Pause) ||
-+ (!(phy_dev->supported & SUPPORTED_Asym_Pause) &&
-+ (epause->rx_pause != epause->tx_pause)))
-+ return -EINVAL;
-+
-+ /* The MAC should know how to handle PAUSE frame autonegotiation before
-+ * adjust_link is triggered by a forced renegotiation of sym/asym PAUSE
-+ * settings.
-+ */
-+ mac_dev->autoneg_pause = !!epause->autoneg;
-+ mac_dev->rx_pause_req = !!epause->rx_pause;
-+ mac_dev->tx_pause_req = !!epause->tx_pause;
-+
-+ /* Determine the sym/asym advertised PAUSE capabilities from the desired
-+ * rx/tx pause settings.
-+ */
-+ newadv = 0;
-+ if (epause->rx_pause)
-+ newadv = ADVERTISED_Pause | ADVERTISED_Asym_Pause;
-+ if (epause->tx_pause)
-+ newadv |= ADVERTISED_Asym_Pause;
-+
-+ oldadv = phy_dev->advertising &
-+ (ADVERTISED_Pause | ADVERTISED_Asym_Pause);
-+
-+ /* If there are differences between the old and the new advertised
-+ * values, restart PHY autonegotiation and advertise the new values.
-+ */
-+ if (oldadv != newadv) {
-+ phy_dev->advertising &= ~(ADVERTISED_Pause
-+ | ADVERTISED_Asym_Pause);
-+ phy_dev->advertising |= newadv;
-+ if (phy_dev->autoneg) {
-+ _errno = phy_start_aneg(phy_dev);
-+ if (unlikely(_errno < 0))
-+ netdev_err(net_dev, "phy_start_aneg() = %d\n",
-+ _errno);
-+ }
-+ }
-+
-+ get_pause_cfg(mac_dev, &rx_pause, &tx_pause);
-+ _errno = set_mac_active_pause(mac_dev, rx_pause, tx_pause);
-+ if (unlikely(_errno < 0))
-+ netdev_err(net_dev, "set_mac_active_pause() = %d\n", _errno);
-+
-+ return _errno;
-+}
-+
-+#ifdef CONFIG_PM
-+static void dpa_get_wol(struct net_device *net_dev, struct ethtool_wolinfo *wol)
-+{
-+ struct dpa_priv_s *priv = netdev_priv(net_dev);
-+
-+ wol->supported = 0;
-+ wol->wolopts = 0;
-+
-+ if (!priv->wol || !device_can_wakeup(net_dev->dev.parent))
-+ return;
-+
-+ if (priv->wol & DPAA_WOL_MAGIC) {
-+ wol->supported = WAKE_MAGIC;
-+ wol->wolopts = WAKE_MAGIC;
-+ }
-+}
-+
-+static int dpa_set_wol(struct net_device *net_dev, struct ethtool_wolinfo *wol)
-+{
-+ struct dpa_priv_s *priv = netdev_priv(net_dev);
-+
-+ if (priv->mac_dev == NULL) {
-+ netdev_info(net_dev, "This is a MAC-less interface\n");
-+ return -ENODEV;
-+ }
-+
-+ if (unlikely(priv->mac_dev->phy_dev == NULL)) {
-+ netdev_dbg(net_dev, "phy device not initialized\n");
-+ return -ENODEV;
-+ }
-+
-+ if (!device_can_wakeup(net_dev->dev.parent) ||
-+ (wol->wolopts & ~WAKE_MAGIC))
-+ return -EOPNOTSUPP;
-+
-+ priv->wol = 0;
-+
-+ if (wol->wolopts & WAKE_MAGIC) {
-+ priv->wol = DPAA_WOL_MAGIC;
-+ device_set_wakeup_enable(net_dev->dev.parent, 1);
-+ } else {
-+ device_set_wakeup_enable(net_dev->dev.parent, 0);
-+ }
-+
-+ return 0;
-+}
-+#endif
-+
-+static int dpa_get_eee(struct net_device *net_dev, struct ethtool_eee *et_eee)
-+{
-+ struct dpa_priv_s *priv;
-+
-+ priv = netdev_priv(net_dev);
-+ if (priv->mac_dev == NULL) {
-+ netdev_info(net_dev, "This is a MAC-less interface\n");
-+ return -ENODEV;
-+ }
-+
-+ if (unlikely(priv->mac_dev->phy_dev == NULL)) {
-+ netdev_err(net_dev, "phy device not initialized\n");
-+ return -ENODEV;
-+ }
-+
-+ return phy_ethtool_get_eee(priv->mac_dev->phy_dev, et_eee);
-+}
-+
-+static int dpa_set_eee(struct net_device *net_dev, struct ethtool_eee *et_eee)
-+{
-+ struct dpa_priv_s *priv;
-+
-+ priv = netdev_priv(net_dev);
-+ if (priv->mac_dev == NULL) {
-+ netdev_info(net_dev, "This is a MAC-less interface\n");
-+ return -ENODEV;
-+ }
-+
-+ if (unlikely(priv->mac_dev->phy_dev == NULL)) {
-+ netdev_err(net_dev, "phy device not initialized\n");
-+ return -ENODEV;
-+ }
-+
-+ return phy_ethtool_set_eee(priv->mac_dev->phy_dev, et_eee);
-+}
-+
-+static int dpa_get_sset_count(struct net_device *net_dev, int type)
-+{
-+ unsigned int total_stats, num_stats;
-+
-+ num_stats = num_online_cpus() + 1;
-+ total_stats = num_stats * DPA_STATS_PERCPU_LEN + DPA_STATS_GLOBAL_LEN;
-+
-+ switch (type) {
-+ case ETH_SS_STATS:
-+ return total_stats;
-+ default:
-+ return -EOPNOTSUPP;
-+ }
-+}
-+
-+static void copy_stats(struct dpa_percpu_priv_s *percpu_priv, int num_cpus,
-+ int crr_cpu, u64 bp_count, u64 *data)
-+{
-+ int num_stat_values = num_cpus + 1;
-+ int crr_stat = 0;
-+
-+ /* update current CPU's stats and also add them to the total values */
-+ data[crr_stat * num_stat_values + crr_cpu] = percpu_priv->in_interrupt;
-+ data[crr_stat++ * num_stat_values + num_cpus] += percpu_priv->in_interrupt;
-+
-+ data[crr_stat * num_stat_values + crr_cpu] = percpu_priv->stats.rx_packets;
-+ data[crr_stat++ * num_stat_values + num_cpus] += percpu_priv->stats.rx_packets;
-+
-+ data[crr_stat * num_stat_values + crr_cpu] = percpu_priv->stats.tx_packets;
-+ data[crr_stat++ * num_stat_values + num_cpus] += percpu_priv->stats.tx_packets;
-+
-+ data[crr_stat * num_stat_values + crr_cpu] = percpu_priv->tx_returned;
-+ data[crr_stat++ * num_stat_values + num_cpus] += percpu_priv->tx_returned;
-+
-+ data[crr_stat * num_stat_values + crr_cpu] = percpu_priv->tx_confirm;
-+ data[crr_stat++ * num_stat_values + num_cpus] += percpu_priv->tx_confirm;
-+
-+ data[crr_stat * num_stat_values + crr_cpu] = percpu_priv->tx_frag_skbuffs;
-+ data[crr_stat++ * num_stat_values + num_cpus] += percpu_priv->tx_frag_skbuffs;
-+
-+ data[crr_stat * num_stat_values + crr_cpu] = percpu_priv->rx_sg;
-+ data[crr_stat++ * num_stat_values + num_cpus] += percpu_priv->rx_sg;
-+
-+ data[crr_stat * num_stat_values + crr_cpu] = percpu_priv->stats.tx_errors;
-+ data[crr_stat++ * num_stat_values + num_cpus] += percpu_priv->stats.tx_errors;
-+
-+ data[crr_stat * num_stat_values + crr_cpu] = percpu_priv->stats.rx_errors;
-+ data[crr_stat++ * num_stat_values + num_cpus] += percpu_priv->stats.rx_errors;
-+
-+ data[crr_stat * num_stat_values + crr_cpu] = bp_count;
-+ data[crr_stat++ * num_stat_values + num_cpus] += bp_count;
-+}
-+
-+static void dpa_get_ethtool_stats(struct net_device *net_dev,
-+ struct ethtool_stats *stats, u64 *data)
-+{
-+ u64 bp_count, cg_time, cg_num, cg_status;
-+ struct dpa_percpu_priv_s *percpu_priv;
-+ struct qm_mcr_querycgr query_cgr;
-+ struct dpa_rx_errors rx_errors;
-+ struct dpa_ern_cnt ern_cnt;
-+ struct dpa_priv_s *priv;
-+ unsigned int num_cpus, offset;
-+ struct dpa_bp *dpa_bp;
-+ int total_stats, i;
-+
-+ total_stats = dpa_get_sset_count(net_dev, ETH_SS_STATS);
-+ priv = netdev_priv(net_dev);
-+ dpa_bp = priv->dpa_bp;
-+ num_cpus = num_online_cpus();
-+ bp_count = 0;
-+
-+ memset(&rx_errors, 0, sizeof(struct dpa_rx_errors));
-+ memset(&ern_cnt, 0, sizeof(struct dpa_ern_cnt));
-+ memset(data, 0, total_stats * sizeof(u64));
-+
-+ for_each_online_cpu(i) {
-+ percpu_priv = per_cpu_ptr(priv->percpu_priv, i);
-+
-+ if (dpa_bp->percpu_count)
-+ bp_count = *(per_cpu_ptr(dpa_bp->percpu_count, i));
-+
-+ rx_errors.dme += percpu_priv->rx_errors.dme;
-+ rx_errors.fpe += percpu_priv->rx_errors.fpe;
-+ rx_errors.fse += percpu_priv->rx_errors.fse;
-+ rx_errors.phe += percpu_priv->rx_errors.phe;
-+ rx_errors.cse += percpu_priv->rx_errors.cse;
-+
-+ ern_cnt.cg_tdrop += percpu_priv->ern_cnt.cg_tdrop;
-+ ern_cnt.wred += percpu_priv->ern_cnt.wred;
-+ ern_cnt.err_cond += percpu_priv->ern_cnt.err_cond;
-+ ern_cnt.early_window += percpu_priv->ern_cnt.early_window;
-+ ern_cnt.late_window += percpu_priv->ern_cnt.late_window;
-+ ern_cnt.fq_tdrop += percpu_priv->ern_cnt.fq_tdrop;
-+ ern_cnt.fq_retired += percpu_priv->ern_cnt.fq_retired;
-+ ern_cnt.orp_zero += percpu_priv->ern_cnt.orp_zero;
-+
-+ copy_stats(percpu_priv, num_cpus, i, bp_count, data);
-+ }
-+
-+ offset = (num_cpus + 1) * DPA_STATS_PERCPU_LEN;
-+ memcpy(data + offset, &rx_errors, sizeof(struct dpa_rx_errors));
-+
-+ offset += sizeof(struct dpa_rx_errors) / sizeof(u64);
-+ memcpy(data + offset, &ern_cnt, sizeof(struct dpa_ern_cnt));
-+
-+ /* gather congestion related counters */
-+ cg_num = 0;
-+ cg_status = 0;
-+ cg_time = jiffies_to_msecs(priv->cgr_data.congested_jiffies);
-+ if (qman_query_cgr(&priv->cgr_data.cgr, &query_cgr) == 0) {
-+ cg_num = priv->cgr_data.cgr_congested_count;
-+ cg_status = query_cgr.cgr.cs;
-+
-+ /* reset congestion stats (like QMan API does */
-+ priv->cgr_data.congested_jiffies = 0;
-+ priv->cgr_data.cgr_congested_count = 0;
-+ }
-+
-+ offset += sizeof(struct dpa_ern_cnt) / sizeof(u64);
-+ data[offset++] = cg_time;
-+ data[offset++] = cg_num;
-+ data[offset++] = cg_status;
-+}
-+
-+static void dpa_get_strings(struct net_device *net_dev, u32 stringset, u8 *data)
-+{
-+ unsigned int i, j, num_cpus, size;
-+ char stat_string_cpu[ETH_GSTRING_LEN];
-+ u8 *strings;
-+
-+ strings = data;
-+ num_cpus = num_online_cpus();
-+ size = DPA_STATS_GLOBAL_LEN * ETH_GSTRING_LEN;
-+
-+ for (i = 0; i < DPA_STATS_PERCPU_LEN; i++) {
-+ for (j = 0; j < num_cpus; j++) {
-+ snprintf(stat_string_cpu, ETH_GSTRING_LEN, "%s [CPU %d]", dpa_stats_percpu[i], j);
-+ memcpy(strings, stat_string_cpu, ETH_GSTRING_LEN);
-+ strings += ETH_GSTRING_LEN;
-+ }
-+ snprintf(stat_string_cpu, ETH_GSTRING_LEN, "%s [TOTAL]", dpa_stats_percpu[i]);
-+ memcpy(strings, stat_string_cpu, ETH_GSTRING_LEN);
-+ strings += ETH_GSTRING_LEN;
-+ }
-+ memcpy(strings, dpa_stats_global, size);
-+}
-+
-+const struct ethtool_ops dpa_ethtool_ops = {
-+ .get_settings = dpa_get_settings,
-+ .set_settings = dpa_set_settings,
-+ .get_drvinfo = dpa_get_drvinfo,
-+ .get_msglevel = dpa_get_msglevel,
-+ .set_msglevel = dpa_set_msglevel,
-+ .nway_reset = dpa_nway_reset,
-+ .get_pauseparam = dpa_get_pauseparam,
-+ .set_pauseparam = dpa_set_pauseparam,
-+ .self_test = NULL, /* TODO invoke the cold-boot unit-test? */
-+ .get_link = ethtool_op_get_link,
-+ .get_eee = dpa_get_eee,
-+ .set_eee = dpa_set_eee,
-+ .get_sset_count = dpa_get_sset_count,
-+ .get_ethtool_stats = dpa_get_ethtool_stats,
-+ .get_strings = dpa_get_strings,
-+#ifdef CONFIG_PM
-+ .get_wol = dpa_get_wol,
-+ .set_wol = dpa_set_wol,
-+#endif
-+};
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_dpaa/dpaa_ptp.c
-@@ -0,0 +1,291 @@
-+/*
-+ * DPAA Ethernet Driver -- PTP 1588 clock using the dTSEC
-+ *
-+ * Author: Yangbo Lu <yangbo.lu@freescale.com>
-+ *
-+ * Copyright 2014 Freescale Semiconductor, Inc.
-+ *
-+ * This program is free software; you can redistribute it and/or modify it
-+ * under the terms of the GNU General Public License as published by the
-+ * Free Software Foundation; either version 2 of the License, or (at your
-+ * option) any later version.
-+*/
-+
-+#include <linux/device.h>
-+#include <linux/hrtimer.h>
-+#include <linux/init.h>
-+#include <linux/interrupt.h>
-+#include <linux/kernel.h>
-+#include <linux/module.h>
-+#include <linux/of.h>
-+#include <linux/of_platform.h>
-+#include <linux/timex.h>
-+#include <linux/io.h>
-+
-+#include <linux/ptp_clock_kernel.h>
-+
-+#include "dpaa_eth.h"
-+#include "mac.h"
-+
-+static struct mac_device *mac_dev;
-+static u32 freqCompensation;
-+
-+/* Bit definitions for the TMR_CTRL register */
-+#define ALM1P (1<<31) /* Alarm1 output polarity */
-+#define ALM2P (1<<30) /* Alarm2 output polarity */
-+#define FS (1<<28) /* FIPER start indication */
-+#define PP1L (1<<27) /* Fiper1 pulse loopback mode enabled. */
-+#define PP2L (1<<26) /* Fiper2 pulse loopback mode enabled. */
-+#define TCLK_PERIOD_SHIFT (16) /* 1588 timer reference clock period. */
-+#define TCLK_PERIOD_MASK (0x3ff)
-+#define RTPE (1<<15) /* Record Tx Timestamp to PAL Enable. */
-+#define FRD (1<<14) /* FIPER Realignment Disable */
-+#define ESFDP (1<<11) /* External Tx/Rx SFD Polarity. */
-+#define ESFDE (1<<10) /* External Tx/Rx SFD Enable. */
-+#define ETEP2 (1<<9) /* External trigger 2 edge polarity */
-+#define ETEP1 (1<<8) /* External trigger 1 edge polarity */
-+#define COPH (1<<7) /* Generated clock output phase. */
-+#define CIPH (1<<6) /* External oscillator input clock phase */
-+#define TMSR (1<<5) /* Timer soft reset. */
-+#define BYP (1<<3) /* Bypass drift compensated clock */
-+#define TE (1<<2) /* 1588 timer enable. */
-+#define CKSEL_SHIFT (0) /* 1588 Timer reference clock source */
-+#define CKSEL_MASK (0x3)
-+
-+/* Bit definitions for the TMR_TEVENT register */
-+#define ETS2 (1<<25) /* External trigger 2 timestamp sampled */
-+#define ETS1 (1<<24) /* External trigger 1 timestamp sampled */
-+#define ALM2 (1<<17) /* Current time = alarm time register 2 */
-+#define ALM1 (1<<16) /* Current time = alarm time register 1 */
-+#define PP1 (1<<7) /* periodic pulse generated on FIPER1 */
-+#define PP2 (1<<6) /* periodic pulse generated on FIPER2 */
-+#define PP3 (1<<5) /* periodic pulse generated on FIPER3 */
-+
-+/* Bit definitions for the TMR_TEMASK register */
-+#define ETS2EN (1<<25) /* External trigger 2 timestamp enable */
-+#define ETS1EN (1<<24) /* External trigger 1 timestamp enable */
-+#define ALM2EN (1<<17) /* Timer ALM2 event enable */
-+#define ALM1EN (1<<16) /* Timer ALM1 event enable */
-+#define PP1EN (1<<7) /* Periodic pulse event 1 enable */
-+#define PP2EN (1<<6) /* Periodic pulse event 2 enable */
-+
-+/* Bit definitions for the TMR_PEVENT register */
-+#define TXP2 (1<<9) /* PTP transmitted timestamp im TXTS2 */
-+#define TXP1 (1<<8) /* PTP transmitted timestamp in TXTS1 */
-+#define RXP (1<<0) /* PTP frame has been received */
-+
-+/* Bit definitions for the TMR_PEMASK register */
-+#define TXP2EN (1<<9) /* Transmit PTP packet event 2 enable */
-+#define TXP1EN (1<<8) /* Transmit PTP packet event 1 enable */
-+#define RXPEN (1<<0) /* Receive PTP packet event enable */
-+
-+/* Bit definitions for the TMR_STAT register */
-+#define STAT_VEC_SHIFT (0) /* Timer general purpose status vector */
-+#define STAT_VEC_MASK (0x3f)
-+
-+/* Bit definitions for the TMR_PRSC register */
-+#define PRSC_OCK_SHIFT (0) /* Output clock division/prescale factor. */
-+#define PRSC_OCK_MASK (0xffff)
-+
-+
-+#define N_EXT_TS 2
-+
-+static void set_alarm(void)
-+{
-+ u64 ns;
-+
-+ if (mac_dev->fm_rtc_get_cnt)
-+ mac_dev->fm_rtc_get_cnt(mac_dev->fm_dev, &ns);
-+ ns += 1500000000ULL;
-+ ns = div_u64(ns, 1000000000UL) * 1000000000ULL;
-+ ns -= DPA_PTP_NOMINAL_FREQ_PERIOD_NS;
-+ if (mac_dev->fm_rtc_set_alarm)
-+ mac_dev->fm_rtc_set_alarm(mac_dev->fm_dev, 0, ns);
-+}
-+
-+static void set_fipers(void)
-+{
-+ u64 fiper;
-+
-+ if (mac_dev->fm_rtc_disable)
-+ mac_dev->fm_rtc_disable(mac_dev->fm_dev);
-+
-+ set_alarm();
-+ fiper = 1000000000ULL - DPA_PTP_NOMINAL_FREQ_PERIOD_NS;
-+ if (mac_dev->fm_rtc_set_fiper)
-+ mac_dev->fm_rtc_set_fiper(mac_dev->fm_dev, 0, fiper);
-+
-+ if (mac_dev->fm_rtc_enable)
-+ mac_dev->fm_rtc_enable(mac_dev->fm_dev);
-+}
-+
-+/* PTP clock operations */
-+
-+static int ptp_dpa_adjfreq(struct ptp_clock_info *ptp, s32 ppb)
-+{
-+ u64 adj;
-+ u32 diff, tmr_add;
-+ int neg_adj = 0;
-+
-+ if (ppb < 0) {
-+ neg_adj = 1;
-+ ppb = -ppb;
-+ }
-+
-+ tmr_add = freqCompensation;
-+ adj = tmr_add;
-+ adj *= ppb;
-+ diff = div_u64(adj, 1000000000ULL);
-+
-+ tmr_add = neg_adj ? tmr_add - diff : tmr_add + diff;
-+
-+ if (mac_dev->fm_rtc_set_drift)
-+ mac_dev->fm_rtc_set_drift(mac_dev->fm_dev, tmr_add);
-+
-+ return 0;
-+}
-+
-+static int ptp_dpa_adjtime(struct ptp_clock_info *ptp, s64 delta)
-+{
-+ s64 now;
-+
-+ if (mac_dev->fm_rtc_get_cnt)
-+ mac_dev->fm_rtc_get_cnt(mac_dev->fm_dev, &now);
-+
-+ now += delta;
-+
-+ if (mac_dev->fm_rtc_set_cnt)
-+ mac_dev->fm_rtc_set_cnt(mac_dev->fm_dev, now);
-+ set_fipers();
-+
-+ return 0;
-+}
-+
-+static int ptp_dpa_gettime(struct ptp_clock_info *ptp, struct timespec64 *ts)
-+{
-+ u64 ns;
-+ u32 remainder;
-+
-+ if (mac_dev->fm_rtc_get_cnt)
-+ mac_dev->fm_rtc_get_cnt(mac_dev->fm_dev, &ns);
-+
-+ ts->tv_sec = div_u64_rem(ns, 1000000000, &remainder);
-+ ts->tv_nsec = remainder;
-+ return 0;
-+}
-+
-+static int ptp_dpa_settime(struct ptp_clock_info *ptp,
-+ const struct timespec64 *ts)
-+{
-+ u64 ns;
-+
-+ ns = ts->tv_sec * 1000000000ULL;
-+ ns += ts->tv_nsec;
-+
-+ if (mac_dev->fm_rtc_set_cnt)
-+ mac_dev->fm_rtc_set_cnt(mac_dev->fm_dev, ns);
-+ set_fipers();
-+ return 0;
-+}
-+
-+static int ptp_dpa_enable(struct ptp_clock_info *ptp,
-+ struct ptp_clock_request *rq, int on)
-+{
-+ u32 bit;
-+
-+ switch (rq->type) {
-+ case PTP_CLK_REQ_EXTTS:
-+ switch (rq->extts.index) {
-+ case 0:
-+ bit = ETS1EN;
-+ break;
-+ case 1:
-+ bit = ETS2EN;
-+ break;
-+ default:
-+ return -EINVAL;
-+ }
-+ if (on) {
-+ if (mac_dev->fm_rtc_enable_interrupt)
-+ mac_dev->fm_rtc_enable_interrupt(
-+ mac_dev->fm_dev, bit);
-+ } else {
-+ if (mac_dev->fm_rtc_disable_interrupt)
-+ mac_dev->fm_rtc_disable_interrupt(
-+ mac_dev->fm_dev, bit);
-+ }
-+ return 0;
-+
-+ case PTP_CLK_REQ_PPS:
-+ if (on) {
-+ if (mac_dev->fm_rtc_enable_interrupt)
-+ mac_dev->fm_rtc_enable_interrupt(
-+ mac_dev->fm_dev, PP1EN);
-+ } else {
-+ if (mac_dev->fm_rtc_disable_interrupt)
-+ mac_dev->fm_rtc_disable_interrupt(
-+ mac_dev->fm_dev, PP1EN);
-+ }
-+ return 0;
-+
-+ default:
-+ break;
-+ }
-+
-+ return -EOPNOTSUPP;
-+}
-+
-+static struct ptp_clock_info ptp_dpa_caps = {
-+ .owner = THIS_MODULE,
-+ .name = "dpaa clock",
-+ .max_adj = 512000,
-+ .n_alarm = 0,
-+ .n_ext_ts = N_EXT_TS,
-+ .n_per_out = 0,
-+ .pps = 1,
-+ .adjfreq = ptp_dpa_adjfreq,
-+ .adjtime = ptp_dpa_adjtime,
-+ .gettime64 = ptp_dpa_gettime,
-+ .settime64 = ptp_dpa_settime,
-+ .enable = ptp_dpa_enable,
-+};
-+
-+static int __init __cold dpa_ptp_load(void)
-+{
-+ struct device *ptp_dev;
-+ struct timespec64 now;
-+ struct ptp_clock *clock = ptp_priv.clock;
-+ int dpa_phc_index;
-+ int err;
-+
-+ if (!(ptp_priv.of_dev && ptp_priv.mac_dev))
-+ return -ENODEV;
-+
-+ ptp_dev = &ptp_priv.of_dev->dev;
-+ mac_dev = ptp_priv.mac_dev;
-+
-+ if (mac_dev->fm_rtc_get_drift)
-+ mac_dev->fm_rtc_get_drift(mac_dev->fm_dev, &freqCompensation);
-+
-+ getnstimeofday64(&now);
-+ ptp_dpa_settime(&ptp_dpa_caps, &now);
-+
-+ clock = ptp_clock_register(&ptp_dpa_caps, ptp_dev);
-+ if (IS_ERR(clock)) {
-+ err = PTR_ERR(clock);
-+ return err;
-+ }
-+ dpa_phc_index = ptp_clock_index(clock);
-+ return 0;
-+}
-+module_init(dpa_ptp_load);
-+
-+static void __exit __cold dpa_ptp_unload(void)
-+{
-+ struct ptp_clock *clock = ptp_priv.clock;
-+
-+ if (mac_dev->fm_rtc_disable_interrupt)
-+ mac_dev->fm_rtc_disable_interrupt(mac_dev->fm_dev, 0xffffffff);
-+ ptp_clock_unregister(clock);
-+}
-+module_exit(dpa_ptp_unload);
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_dpaa/mac-api.c
-@@ -0,0 +1,931 @@
-+/* Copyright 2008-2012 Freescale Semiconductor, Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#ifdef CONFIG_FSL_DPAA_ETH_DEBUG
-+#define pr_fmt(fmt) \
-+ KBUILD_MODNAME ": %s:%hu:%s() " fmt, \
-+ KBUILD_BASENAME".c", __LINE__, __func__
-+#else
-+#define pr_fmt(fmt) \
-+ KBUILD_MODNAME ": " fmt
-+#endif
-+
-+#include <linux/init.h>
-+#include <linux/module.h>
-+#include <linux/io.h>
-+#include <linux/of_platform.h>
-+#include <linux/of_mdio.h>
-+#include <linux/phy.h>
-+#include <linux/netdevice.h>
-+
-+#include "dpaa_eth.h"
-+#include "mac.h"
-+#include "lnxwrp_fsl_fman.h"
-+
-+#include "error_ext.h" /* GET_ERROR_TYPE, E_OK */
-+
-+#include "fsl_fman_dtsec.h"
-+#include "fsl_fman_tgec.h"
-+#include "fsl_fman_memac.h"
-+#include "../sdk_fman/src/wrapper/lnxwrp_sysfs_fm.h"
-+
-+#define MAC_DESCRIPTION "FSL FMan MAC API based driver"
-+
-+MODULE_LICENSE("Dual BSD/GPL");
-+
-+MODULE_AUTHOR("Emil Medve <Emilian.Medve@Freescale.com>");
-+
-+MODULE_DESCRIPTION(MAC_DESCRIPTION);
-+
-+struct mac_priv_s {
-+ struct fm_mac_dev *fm_mac;
-+};
-+
-+const char *mac_driver_description __initconst = MAC_DESCRIPTION;
-+const size_t mac_sizeof_priv[] = {
-+ [DTSEC] = sizeof(struct mac_priv_s),
-+ [XGMAC] = sizeof(struct mac_priv_s),
-+ [MEMAC] = sizeof(struct mac_priv_s)
-+};
-+
-+static const enet_mode_t _100[] = {
-+ [PHY_INTERFACE_MODE_MII] = e_ENET_MODE_MII_100,
-+ [PHY_INTERFACE_MODE_RMII] = e_ENET_MODE_RMII_100
-+};
-+
-+static const enet_mode_t _1000[] = {
-+ [PHY_INTERFACE_MODE_GMII] = e_ENET_MODE_GMII_1000,
-+ [PHY_INTERFACE_MODE_SGMII] = e_ENET_MODE_SGMII_1000,
-+ [PHY_INTERFACE_MODE_QSGMII] = e_ENET_MODE_QSGMII_1000,
-+ [PHY_INTERFACE_MODE_TBI] = e_ENET_MODE_TBI_1000,
-+ [PHY_INTERFACE_MODE_RGMII] = e_ENET_MODE_RGMII_1000,
-+ [PHY_INTERFACE_MODE_RGMII_ID] = e_ENET_MODE_RGMII_1000,
-+ [PHY_INTERFACE_MODE_RGMII_RXID] = e_ENET_MODE_RGMII_1000,
-+ [PHY_INTERFACE_MODE_RGMII_TXID] = e_ENET_MODE_RGMII_1000,
-+ [PHY_INTERFACE_MODE_RTBI] = e_ENET_MODE_RTBI_1000
-+};
-+
-+static enet_mode_t __cold __attribute__((nonnull))
-+macdev2enetinterface(const struct mac_device *mac_dev)
-+{
-+ switch (mac_dev->max_speed) {
-+ case SPEED_100:
-+ return _100[mac_dev->phy_if];
-+ case SPEED_1000:
-+ return _1000[mac_dev->phy_if];
-+ case SPEED_2500:
-+ return e_ENET_MODE_SGMII_2500;
-+ case SPEED_10000:
-+ return e_ENET_MODE_XGMII_10000;
-+ default:
-+ return e_ENET_MODE_MII_100;
-+ }
-+}
-+
-+static void mac_exception(handle_t _mac_dev, e_FmMacExceptions exception)
-+{
-+ struct mac_device *mac_dev;
-+
-+ mac_dev = (struct mac_device *)_mac_dev;
-+
-+ if (e_FM_MAC_EX_10G_RX_FIFO_OVFL == exception) {
-+ /* don't flag RX FIFO after the first */
-+ fm_mac_set_exception(mac_dev->get_mac_handle(mac_dev),
-+ e_FM_MAC_EX_10G_RX_FIFO_OVFL, false);
-+ dev_err(mac_dev->dev, "10G MAC got RX FIFO Error = %x\n",
-+ exception);
-+ }
-+
-+ dev_dbg(mac_dev->dev, "%s:%s() -> %d\n", KBUILD_BASENAME".c", __func__,
-+ exception);
-+}
-+
-+static int __cold init(struct mac_device *mac_dev)
-+{
-+ int _errno;
-+ struct mac_priv_s *priv;
-+ t_FmMacParams param;
-+ uint32_t version;
-+
-+ priv = macdev_priv(mac_dev);
-+
-+ param.baseAddr = (typeof(param.baseAddr))(uintptr_t)devm_ioremap(
-+ mac_dev->dev, mac_dev->res->start, 0x2000);
-+ param.enetMode = macdev2enetinterface(mac_dev);
-+ memcpy(&param.addr, mac_dev->addr, min(sizeof(param.addr),
-+ sizeof(mac_dev->addr)));
-+ param.macId = mac_dev->cell_index;
-+ param.h_Fm = (handle_t)mac_dev->fm;
-+ param.mdioIrq = NO_IRQ;
-+ param.f_Exception = mac_exception;
-+ param.f_Event = mac_exception;
-+ param.h_App = mac_dev;
-+
-+ priv->fm_mac = fm_mac_config(&param);
-+ if (unlikely(priv->fm_mac == NULL)) {
-+ _errno = -EINVAL;
-+ goto _return;
-+ }
-+
-+ fm_mac_set_handle(mac_dev->fm_dev, priv->fm_mac,
-+ (macdev2enetinterface(mac_dev) != e_ENET_MODE_XGMII_10000) ?
-+ param.macId : param.macId + FM_MAX_NUM_OF_1G_MACS);
-+
-+ _errno = fm_mac_config_max_frame_length(priv->fm_mac,
-+ fm_get_max_frm());
-+ if (unlikely(_errno < 0))
-+ goto _return_fm_mac_free;
-+
-+ if (macdev2enetinterface(mac_dev) != e_ENET_MODE_XGMII_10000) {
-+ /* 10G always works with pad and CRC */
-+ _errno = fm_mac_config_pad_and_crc(priv->fm_mac, true);
-+ if (unlikely(_errno < 0))
-+ goto _return_fm_mac_free;
-+
-+ _errno = fm_mac_config_half_duplex(priv->fm_mac,
-+ mac_dev->half_duplex);
-+ if (unlikely(_errno < 0))
-+ goto _return_fm_mac_free;
-+ } else {
-+ _errno = fm_mac_config_reset_on_init(priv->fm_mac, true);
-+ if (unlikely(_errno < 0))
-+ goto _return_fm_mac_free;
-+ }
-+
-+ _errno = fm_mac_init(priv->fm_mac);
-+ if (unlikely(_errno < 0))
-+ goto _return_fm_mac_free;
-+
-+#ifndef CONFIG_FMAN_MIB_CNT_OVF_IRQ_EN
-+ /* For 1G MAC, disable by default the MIB counters overflow interrupt */
-+ if (macdev2enetinterface(mac_dev) != e_ENET_MODE_XGMII_10000) {
-+ _errno = fm_mac_set_exception(mac_dev->get_mac_handle(mac_dev),
-+ e_FM_MAC_EX_1G_RX_MIB_CNT_OVFL, FALSE);
-+ if (unlikely(_errno < 0))
-+ goto _return_fm_mac_free;
-+ }
-+#endif /* !CONFIG_FMAN_MIB_CNT_OVF_IRQ_EN */
-+
-+ /* For 10G MAC, disable Tx ECC exception */
-+ if (macdev2enetinterface(mac_dev) == e_ENET_MODE_XGMII_10000) {
-+ _errno = fm_mac_set_exception(mac_dev->get_mac_handle(mac_dev),
-+ e_FM_MAC_EX_10G_1TX_ECC_ER, FALSE);
-+ if (unlikely(_errno < 0))
-+ goto _return_fm_mac_free;
-+ }
-+
-+ _errno = fm_mac_get_version(priv->fm_mac, &version);
-+ if (unlikely(_errno < 0))
-+ goto _return_fm_mac_free;
-+
-+ dev_info(mac_dev->dev, "FMan %s version: 0x%08x\n",
-+ ((macdev2enetinterface(mac_dev) != e_ENET_MODE_XGMII_10000) ?
-+ "dTSEC" : "XGEC"), version);
-+
-+ goto _return;
-+
-+
-+_return_fm_mac_free:
-+ fm_mac_free(mac_dev->get_mac_handle(mac_dev));
-+
-+_return:
-+ return _errno;
-+}
-+
-+static int __cold memac_init(struct mac_device *mac_dev)
-+{
-+ int _errno;
-+ struct mac_priv_s *priv;
-+ t_FmMacParams param;
-+
-+ priv = macdev_priv(mac_dev);
-+
-+ param.baseAddr = (typeof(param.baseAddr))(uintptr_t)devm_ioremap(
-+ mac_dev->dev, mac_dev->res->start, 0x2000);
-+ param.enetMode = macdev2enetinterface(mac_dev);
-+ memcpy(&param.addr, mac_dev->addr, sizeof(mac_dev->addr));
-+ param.macId = mac_dev->cell_index;
-+ param.h_Fm = (handle_t)mac_dev->fm;
-+ param.mdioIrq = NO_IRQ;
-+ param.f_Exception = mac_exception;
-+ param.f_Event = mac_exception;
-+ param.h_App = mac_dev;
-+
-+ priv->fm_mac = fm_mac_config(&param);
-+ if (unlikely(priv->fm_mac == NULL)) {
-+ _errno = -EINVAL;
-+ goto _return;
-+ }
-+
-+ fm_mac_set_handle(mac_dev->fm_dev, priv->fm_mac,
-+ (macdev2enetinterface(mac_dev) != e_ENET_MODE_XGMII_10000) ?
-+ param.macId : param.macId + FM_MAX_NUM_OF_1G_MACS);
-+
-+ _errno = fm_mac_config_max_frame_length(priv->fm_mac, fm_get_max_frm());
-+ if (unlikely(_errno < 0))
-+ goto _return_fm_mac_free;
-+
-+ _errno = fm_mac_config_reset_on_init(priv->fm_mac, true);
-+ if (unlikely(_errno < 0))
-+ goto _return_fm_mac_free;
-+
-+ _errno = fm_mac_init(priv->fm_mac);
-+ if (unlikely(_errno < 0))
-+ goto _return_fm_mac_free;
-+
-+ dev_info(mac_dev->dev, "FMan MEMAC\n");
-+
-+ goto _return;
-+
-+_return_fm_mac_free:
-+ fm_mac_free(priv->fm_mac);
-+
-+_return:
-+ return _errno;
-+}
-+
-+static int __cold start(struct mac_device *mac_dev)
-+{
-+ int _errno;
-+ struct phy_device *phy_dev = mac_dev->phy_dev;
-+
-+ _errno = fm_mac_enable(mac_dev->get_mac_handle(mac_dev));
-+
-+ if (!_errno && phy_dev)
-+ phy_start(phy_dev);
-+
-+ return _errno;
-+}
-+
-+static int __cold stop(struct mac_device *mac_dev)
-+{
-+ if (mac_dev->phy_dev)
-+ phy_stop(mac_dev->phy_dev);
-+
-+ return fm_mac_disable(mac_dev->get_mac_handle(mac_dev));
-+}
-+
-+static int __cold set_multi(struct net_device *net_dev,
-+ struct mac_device *mac_dev)
-+{
-+ struct mac_priv_s *mac_priv;
-+ struct mac_address *old_addr, *tmp;
-+ struct netdev_hw_addr *ha;
-+ int _errno;
-+
-+ mac_priv = macdev_priv(mac_dev);
-+
-+ /* Clear previous address list */
-+ list_for_each_entry_safe(old_addr, tmp, &mac_dev->mc_addr_list, list) {
-+ _errno = fm_mac_remove_hash_mac_addr(mac_priv->fm_mac,
-+ (t_EnetAddr *)old_addr->addr);
-+ if (_errno < 0)
-+ return _errno;
-+
-+ list_del(&old_addr->list);
-+ kfree(old_addr);
-+ }
-+
-+ /* Add all the addresses from the new list */
-+ netdev_for_each_mc_addr(ha, net_dev) {
-+ _errno = fm_mac_add_hash_mac_addr(mac_priv->fm_mac,
-+ (t_EnetAddr *)ha->addr);
-+ if (_errno < 0)
-+ return _errno;
-+
-+ tmp = kmalloc(sizeof(struct mac_address), GFP_ATOMIC);
-+ if (!tmp) {
-+ dev_err(mac_dev->dev, "Out of memory\n");
-+ return -ENOMEM;
-+ }
-+ memcpy(tmp->addr, ha->addr, ETH_ALEN);
-+ list_add(&tmp->list, &mac_dev->mc_addr_list);
-+ }
-+ return 0;
-+}
-+
-+/* Avoid redundant calls to FMD, if the MAC driver already contains the desired
-+ * active PAUSE settings. Otherwise, the new active settings should be reflected
-+ * in FMan.
-+ */
-+int set_mac_active_pause(struct mac_device *mac_dev, bool rx, bool tx)
-+{
-+ struct fm_mac_dev *fm_mac_dev = mac_dev->get_mac_handle(mac_dev);
-+ int _errno = 0;
-+
-+ if (unlikely(rx != mac_dev->rx_pause_active)) {
-+ _errno = fm_mac_set_rx_pause_frames(fm_mac_dev, rx);
-+ if (likely(_errno == 0))
-+ mac_dev->rx_pause_active = rx;
-+ }
-+
-+ if (unlikely(tx != mac_dev->tx_pause_active)) {
-+ _errno = fm_mac_set_tx_pause_frames(fm_mac_dev, tx);
-+ if (likely(_errno == 0))
-+ mac_dev->tx_pause_active = tx;
-+ }
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(set_mac_active_pause);
-+
-+/* Determine the MAC RX/TX PAUSE frames settings based on PHY
-+ * autonegotiation or values set by eththool.
-+ */
-+void get_pause_cfg(struct mac_device *mac_dev, bool *rx_pause, bool *tx_pause)
-+{
-+ struct phy_device *phy_dev = mac_dev->phy_dev;
-+ u16 lcl_adv, rmt_adv;
-+ u8 flowctrl;
-+
-+ *rx_pause = *tx_pause = false;
-+
-+ if (!phy_dev->duplex)
-+ return;
-+
-+ /* If PAUSE autonegotiation is disabled, the TX/RX PAUSE settings
-+ * are those set by ethtool.
-+ */
-+ if (!mac_dev->autoneg_pause) {
-+ *rx_pause = mac_dev->rx_pause_req;
-+ *tx_pause = mac_dev->tx_pause_req;
-+ return;
-+ }
-+
-+ /* Else if PAUSE autonegotiation is enabled, the TX/RX PAUSE
-+ * settings depend on the result of the link negotiation.
-+ */
-+
-+ /* get local capabilities */
-+ lcl_adv = 0;
-+ if (phy_dev->advertising & ADVERTISED_Pause)
-+ lcl_adv |= ADVERTISE_PAUSE_CAP;
-+ if (phy_dev->advertising & ADVERTISED_Asym_Pause)
-+ lcl_adv |= ADVERTISE_PAUSE_ASYM;
-+
-+ /* get link partner capabilities */
-+ rmt_adv = 0;
-+ if (phy_dev->pause)
-+ rmt_adv |= LPA_PAUSE_CAP;
-+ if (phy_dev->asym_pause)
-+ rmt_adv |= LPA_PAUSE_ASYM;
-+
-+ /* Calculate TX/RX settings based on local and peer advertised
-+ * symmetric/asymmetric PAUSE capabilities.
-+ */
-+ flowctrl = mii_resolve_flowctrl_fdx(lcl_adv, rmt_adv);
-+ if (flowctrl & FLOW_CTRL_RX)
-+ *rx_pause = true;
-+ if (flowctrl & FLOW_CTRL_TX)
-+ *tx_pause = true;
-+}
-+EXPORT_SYMBOL(get_pause_cfg);
-+
-+static void adjust_link_void(struct net_device *net_dev)
-+{
-+}
-+
-+static void adjust_link(struct net_device *net_dev)
-+{
-+ struct dpa_priv_s *priv = netdev_priv(net_dev);
-+ struct mac_device *mac_dev = priv->mac_dev;
-+ struct phy_device *phy_dev = mac_dev->phy_dev;
-+ struct fm_mac_dev *fm_mac_dev;
-+ bool rx_pause, tx_pause;
-+ int _errno;
-+
-+ fm_mac_dev = mac_dev->get_mac_handle(mac_dev);
-+ fm_mac_adjust_link(fm_mac_dev, phy_dev->link, phy_dev->speed,
-+ phy_dev->duplex);
-+
-+ get_pause_cfg(mac_dev, &rx_pause, &tx_pause);
-+ _errno = set_mac_active_pause(mac_dev, rx_pause, tx_pause);
-+ if (unlikely(_errno < 0))
-+ netdev_err(net_dev, "set_mac_active_pause() = %d\n", _errno);
-+}
-+
-+/* Initializes driver's PHY state, and attaches to the PHY.
-+ * Returns 0 on success.
-+ */
-+static int dtsec_init_phy(struct net_device *net_dev,
-+ struct mac_device *mac_dev)
-+{
-+ struct phy_device *phy_dev;
-+
-+ if (of_phy_is_fixed_link(mac_dev->phy_node))
-+ phy_dev = of_phy_attach(net_dev, mac_dev->phy_node,
-+ 0, mac_dev->phy_if);
-+ else
-+ phy_dev = of_phy_connect(net_dev, mac_dev->phy_node,
-+ &adjust_link, 0, mac_dev->phy_if);
-+ if (unlikely(phy_dev == NULL) || IS_ERR(phy_dev)) {
-+ netdev_err(net_dev, "Could not connect to PHY %s\n",
-+ mac_dev->phy_node ?
-+ mac_dev->phy_node->full_name :
-+ mac_dev->fixed_bus_id);
-+ return phy_dev == NULL ? -ENODEV : PTR_ERR(phy_dev);
-+ }
-+
-+ /* Remove any features not supported by the controller */
-+ phy_dev->supported &= mac_dev->if_support;
-+ /* Enable the symmetric and asymmetric PAUSE frame advertisements,
-+ * as most of the PHY drivers do not enable them by default.
-+ */
-+ phy_dev->supported |= (SUPPORTED_Pause | SUPPORTED_Asym_Pause);
-+ phy_dev->advertising = phy_dev->supported;
-+
-+ mac_dev->phy_dev = phy_dev;
-+
-+ return 0;
-+}
-+
-+static int xgmac_init_phy(struct net_device *net_dev,
-+ struct mac_device *mac_dev)
-+{
-+ struct phy_device *phy_dev;
-+
-+ if (of_phy_is_fixed_link(mac_dev->phy_node))
-+ phy_dev = of_phy_attach(net_dev, mac_dev->phy_node,
-+ 0, mac_dev->phy_if);
-+ else
-+ phy_dev = of_phy_connect(net_dev, mac_dev->phy_node,
-+ &adjust_link_void, 0, mac_dev->phy_if);
-+ if (unlikely(phy_dev == NULL) || IS_ERR(phy_dev)) {
-+ netdev_err(net_dev, "Could not attach to PHY %s\n",
-+ mac_dev->phy_node ?
-+ mac_dev->phy_node->full_name :
-+ mac_dev->fixed_bus_id);
-+ return phy_dev == NULL ? -ENODEV : PTR_ERR(phy_dev);
-+ }
-+
-+ phy_dev->supported &= mac_dev->if_support;
-+ /* Enable the symmetric and asymmetric PAUSE frame advertisements,
-+ * as most of the PHY drivers do not enable them by default.
-+ */
-+ phy_dev->supported |= (SUPPORTED_Pause | SUPPORTED_Asym_Pause);
-+ phy_dev->advertising = phy_dev->supported;
-+
-+ mac_dev->phy_dev = phy_dev;
-+
-+ return 0;
-+}
-+
-+static int memac_init_phy(struct net_device *net_dev,
-+ struct mac_device *mac_dev)
-+{
-+ struct phy_device *phy_dev;
-+ void (*adjust_link_handler)(struct net_device *);
-+
-+ if ((macdev2enetinterface(mac_dev) == e_ENET_MODE_XGMII_10000) ||
-+ (macdev2enetinterface(mac_dev) == e_ENET_MODE_SGMII_2500)) {
-+ /* Pass a void link state handler to the PHY state machine
-+ * for XGMII (10G) and SGMII 2.5G, as the hardware does not
-+ * permit dynamic link speed adjustments. */
-+ adjust_link_handler = adjust_link_void;
-+ } else if (macdev2enetinterface(mac_dev) & e_ENET_IF_RGMII) {
-+ /* Regular RGMII ports connected to a PHY, as well as
-+ * ports that are marked as "fixed-link" in the DTS,
-+ * will have the adjust_link callback. This calls
-+ * fman_memac_adjust_link in order to configure the
-+ * IF_MODE register, which is needed in both cases.
-+ */
-+ adjust_link_handler = adjust_link;
-+ } else if (of_phy_is_fixed_link(mac_dev->phy_node)) {
-+ /* Pass a void link state handler for fixed-link
-+ * interfaces that are not RGMII. Only RGMII has been
-+ * tested and confirmed to work with fixed-link. Other
-+ * MII interfaces may need further work.
-+ * TODO: Change this as needed.
-+ */
-+ adjust_link_handler = adjust_link_void;
-+ } else {
-+ /* MII, RMII, SMII, GMII, SGMII, BASEX ports,
-+ * that are NOT fixed-link.
-+ * TODO: May not be needed for interfaces that
-+ * pass through the SerDes block (*SGMII, XFI).
-+ */
-+ adjust_link_handler = adjust_link;
-+ }
-+ phy_dev = of_phy_connect(net_dev, mac_dev->phy_node,
-+ adjust_link_handler, 0,
-+ mac_dev->phy_if);
-+
-+ if (unlikely(phy_dev == NULL) || IS_ERR(phy_dev)) {
-+ netdev_err(net_dev, "Could not connect to PHY %s\n",
-+ mac_dev->phy_node ?
-+ mac_dev->phy_node->full_name :
-+ mac_dev->fixed_bus_id);
-+ return phy_dev == NULL ? -ENODEV : PTR_ERR(phy_dev);
-+ }
-+
-+ /* Remove any features not supported by the controller */
-+ phy_dev->supported &= mac_dev->if_support;
-+ /* Enable the symmetric and asymmetric PAUSE frame advertisements,
-+ * as most of the PHY drivers do not enable them by default.
-+ */
-+ phy_dev->supported |= (SUPPORTED_Pause | SUPPORTED_Asym_Pause);
-+ phy_dev->advertising = phy_dev->supported;
-+
-+ mac_dev->phy_dev = phy_dev;
-+
-+ return 0;
-+}
-+
-+static int __cold uninit(struct fm_mac_dev *fm_mac_dev)
-+{
-+ int _errno, __errno;
-+
-+ _errno = fm_mac_disable(fm_mac_dev);
-+ __errno = fm_mac_free(fm_mac_dev);
-+
-+ if (unlikely(__errno < 0))
-+ _errno = __errno;
-+
-+ return _errno;
-+}
-+
-+static struct fm_mac_dev *get_mac_handle(struct mac_device *mac_dev)
-+{
-+ const struct mac_priv_s *priv;
-+ priv = macdev_priv(mac_dev);
-+ return priv->fm_mac;
-+}
-+
-+static int dtsec_dump_regs(struct mac_device *h_mac, char *buf, int nn)
-+{
-+ struct dtsec_regs *p_mm = (struct dtsec_regs *) h_mac->vaddr;
-+ int i = 0, n = nn;
-+
-+ FM_DMP_SUBTITLE(buf, n, "\n");
-+
-+ FM_DMP_TITLE(buf, n, p_mm, "FM MAC - DTSEC-%d", h_mac->cell_index);
-+
-+ FM_DMP_V32(buf, n, p_mm, tsec_id);
-+ FM_DMP_V32(buf, n, p_mm, tsec_id2);
-+ FM_DMP_V32(buf, n, p_mm, ievent);
-+ FM_DMP_V32(buf, n, p_mm, imask);
-+ FM_DMP_V32(buf, n, p_mm, ecntrl);
-+ FM_DMP_V32(buf, n, p_mm, ptv);
-+ FM_DMP_V32(buf, n, p_mm, tmr_ctrl);
-+ FM_DMP_V32(buf, n, p_mm, tmr_pevent);
-+ FM_DMP_V32(buf, n, p_mm, tmr_pemask);
-+ FM_DMP_V32(buf, n, p_mm, tctrl);
-+ FM_DMP_V32(buf, n, p_mm, rctrl);
-+ FM_DMP_V32(buf, n, p_mm, maccfg1);
-+ FM_DMP_V32(buf, n, p_mm, maccfg2);
-+ FM_DMP_V32(buf, n, p_mm, ipgifg);
-+ FM_DMP_V32(buf, n, p_mm, hafdup);
-+ FM_DMP_V32(buf, n, p_mm, maxfrm);
-+
-+ FM_DMP_V32(buf, n, p_mm, macstnaddr1);
-+ FM_DMP_V32(buf, n, p_mm, macstnaddr2);
-+
-+ for (i = 0; i < 7; ++i) {
-+ FM_DMP_V32(buf, n, p_mm, macaddr[i].exact_match1);
-+ FM_DMP_V32(buf, n, p_mm, macaddr[i].exact_match2);
-+ }
-+
-+ FM_DMP_V32(buf, n, p_mm, car1);
-+ FM_DMP_V32(buf, n, p_mm, car2);
-+
-+ return n;
-+}
-+
-+static int xgmac_dump_regs(struct mac_device *h_mac, char *buf, int nn)
-+{
-+ struct tgec_regs *p_mm = (struct tgec_regs *) h_mac->vaddr;
-+ int n = nn;
-+
-+ FM_DMP_SUBTITLE(buf, n, "\n");
-+ FM_DMP_TITLE(buf, n, p_mm, "FM MAC - TGEC -%d", h_mac->cell_index);
-+
-+ FM_DMP_V32(buf, n, p_mm, tgec_id);
-+ FM_DMP_V32(buf, n, p_mm, command_config);
-+ FM_DMP_V32(buf, n, p_mm, mac_addr_0);
-+ FM_DMP_V32(buf, n, p_mm, mac_addr_1);
-+ FM_DMP_V32(buf, n, p_mm, maxfrm);
-+ FM_DMP_V32(buf, n, p_mm, pause_quant);
-+ FM_DMP_V32(buf, n, p_mm, rx_fifo_sections);
-+ FM_DMP_V32(buf, n, p_mm, tx_fifo_sections);
-+ FM_DMP_V32(buf, n, p_mm, rx_fifo_almost_f_e);
-+ FM_DMP_V32(buf, n, p_mm, tx_fifo_almost_f_e);
-+ FM_DMP_V32(buf, n, p_mm, hashtable_ctrl);
-+ FM_DMP_V32(buf, n, p_mm, mdio_cfg_status);
-+ FM_DMP_V32(buf, n, p_mm, mdio_command);
-+ FM_DMP_V32(buf, n, p_mm, mdio_data);
-+ FM_DMP_V32(buf, n, p_mm, mdio_regaddr);
-+ FM_DMP_V32(buf, n, p_mm, status);
-+ FM_DMP_V32(buf, n, p_mm, tx_ipg_len);
-+ FM_DMP_V32(buf, n, p_mm, mac_addr_2);
-+ FM_DMP_V32(buf, n, p_mm, mac_addr_3);
-+ FM_DMP_V32(buf, n, p_mm, rx_fifo_ptr_rd);
-+ FM_DMP_V32(buf, n, p_mm, rx_fifo_ptr_wr);
-+ FM_DMP_V32(buf, n, p_mm, tx_fifo_ptr_rd);
-+ FM_DMP_V32(buf, n, p_mm, tx_fifo_ptr_wr);
-+ FM_DMP_V32(buf, n, p_mm, imask);
-+ FM_DMP_V32(buf, n, p_mm, ievent);
-+
-+ return n;
-+}
-+
-+static int memac_dump_regs(struct mac_device *h_mac, char *buf, int nn)
-+{
-+ struct memac_regs *p_mm = (struct memac_regs *) h_mac->vaddr;
-+ int i = 0, n = nn;
-+
-+ FM_DMP_SUBTITLE(buf, n, "\n");
-+ FM_DMP_TITLE(buf, n, p_mm, "FM MAC - MEMAC -%d", h_mac->cell_index);
-+
-+ FM_DMP_V32(buf, n, p_mm, command_config);
-+ FM_DMP_V32(buf, n, p_mm, mac_addr0.mac_addr_l);
-+ FM_DMP_V32(buf, n, p_mm, mac_addr0.mac_addr_u);
-+ FM_DMP_V32(buf, n, p_mm, maxfrm);
-+ FM_DMP_V32(buf, n, p_mm, hashtable_ctrl);
-+ FM_DMP_V32(buf, n, p_mm, ievent);
-+ FM_DMP_V32(buf, n, p_mm, tx_ipg_length);
-+ FM_DMP_V32(buf, n, p_mm, imask);
-+
-+ for (i = 0; i < 4; ++i)
-+ FM_DMP_V32(buf, n, p_mm, pause_quanta[i]);
-+
-+ for (i = 0; i < 4; ++i)
-+ FM_DMP_V32(buf, n, p_mm, pause_thresh[i]);
-+
-+ FM_DMP_V32(buf, n, p_mm, rx_pause_status);
-+
-+ for (i = 0; i < MEMAC_NUM_OF_PADDRS; ++i) {
-+ FM_DMP_V32(buf, n, p_mm, mac_addr[i].mac_addr_l);
-+ FM_DMP_V32(buf, n, p_mm, mac_addr[i].mac_addr_u);
-+ }
-+
-+ FM_DMP_V32(buf, n, p_mm, lpwake_timer);
-+ FM_DMP_V32(buf, n, p_mm, sleep_timer);
-+ FM_DMP_V32(buf, n, p_mm, statn_config);
-+ FM_DMP_V32(buf, n, p_mm, if_mode);
-+ FM_DMP_V32(buf, n, p_mm, if_status);
-+ FM_DMP_V32(buf, n, p_mm, hg_config);
-+ FM_DMP_V32(buf, n, p_mm, hg_pause_quanta);
-+ FM_DMP_V32(buf, n, p_mm, hg_pause_thresh);
-+ FM_DMP_V32(buf, n, p_mm, hgrx_pause_status);
-+ FM_DMP_V32(buf, n, p_mm, hg_fifos_status);
-+ FM_DMP_V32(buf, n, p_mm, rhm);
-+ FM_DMP_V32(buf, n, p_mm, thm);
-+
-+ return n;
-+}
-+
-+static int memac_dump_regs_rx(struct mac_device *h_mac, char *buf, int nn)
-+{
-+ struct memac_regs *p_mm = (struct memac_regs *) h_mac->vaddr;
-+ int n = nn;
-+
-+ FM_DMP_SUBTITLE(buf, n, "\n");
-+ FM_DMP_TITLE(buf, n, p_mm, "FM MAC - MEMAC -%d Rx stats", h_mac->cell_index);
-+
-+ /* Rx Statistics Counter */
-+ FM_DMP_V32(buf, n, p_mm, reoct_l);
-+ FM_DMP_V32(buf, n, p_mm, reoct_u);
-+ FM_DMP_V32(buf, n, p_mm, roct_l);
-+ FM_DMP_V32(buf, n, p_mm, roct_u);
-+ FM_DMP_V32(buf, n, p_mm, raln_l);
-+ FM_DMP_V32(buf, n, p_mm, raln_u);
-+ FM_DMP_V32(buf, n, p_mm, rxpf_l);
-+ FM_DMP_V32(buf, n, p_mm, rxpf_u);
-+ FM_DMP_V32(buf, n, p_mm, rfrm_l);
-+ FM_DMP_V32(buf, n, p_mm, rfrm_u);
-+ FM_DMP_V32(buf, n, p_mm, rfcs_l);
-+ FM_DMP_V32(buf, n, p_mm, rfcs_u);
-+ FM_DMP_V32(buf, n, p_mm, rvlan_l);
-+ FM_DMP_V32(buf, n, p_mm, rvlan_u);
-+ FM_DMP_V32(buf, n, p_mm, rerr_l);
-+ FM_DMP_V32(buf, n, p_mm, rerr_u);
-+ FM_DMP_V32(buf, n, p_mm, ruca_l);
-+ FM_DMP_V32(buf, n, p_mm, ruca_u);
-+ FM_DMP_V32(buf, n, p_mm, rmca_l);
-+ FM_DMP_V32(buf, n, p_mm, rmca_u);
-+ FM_DMP_V32(buf, n, p_mm, rbca_l);
-+ FM_DMP_V32(buf, n, p_mm, rbca_u);
-+ FM_DMP_V32(buf, n, p_mm, rdrp_l);
-+ FM_DMP_V32(buf, n, p_mm, rdrp_u);
-+ FM_DMP_V32(buf, n, p_mm, rpkt_l);
-+ FM_DMP_V32(buf, n, p_mm, rpkt_u);
-+ FM_DMP_V32(buf, n, p_mm, rund_l);
-+ FM_DMP_V32(buf, n, p_mm, rund_u);
-+ FM_DMP_V32(buf, n, p_mm, r64_l);
-+ FM_DMP_V32(buf, n, p_mm, r64_u);
-+ FM_DMP_V32(buf, n, p_mm, r127_l);
-+ FM_DMP_V32(buf, n, p_mm, r127_u);
-+ FM_DMP_V32(buf, n, p_mm, r255_l);
-+ FM_DMP_V32(buf, n, p_mm, r255_u);
-+ FM_DMP_V32(buf, n, p_mm, r511_l);
-+ FM_DMP_V32(buf, n, p_mm, r511_u);
-+ FM_DMP_V32(buf, n, p_mm, r1023_l);
-+ FM_DMP_V32(buf, n, p_mm, r1023_u);
-+ FM_DMP_V32(buf, n, p_mm, r1518_l);
-+ FM_DMP_V32(buf, n, p_mm, r1518_u);
-+ FM_DMP_V32(buf, n, p_mm, r1519x_l);
-+ FM_DMP_V32(buf, n, p_mm, r1519x_u);
-+ FM_DMP_V32(buf, n, p_mm, rovr_l);
-+ FM_DMP_V32(buf, n, p_mm, rovr_u);
-+ FM_DMP_V32(buf, n, p_mm, rjbr_l);
-+ FM_DMP_V32(buf, n, p_mm, rjbr_u);
-+ FM_DMP_V32(buf, n, p_mm, rfrg_l);
-+ FM_DMP_V32(buf, n, p_mm, rfrg_u);
-+ FM_DMP_V32(buf, n, p_mm, rcnp_l);
-+ FM_DMP_V32(buf, n, p_mm, rcnp_u);
-+ FM_DMP_V32(buf, n, p_mm, rdrntp_l);
-+ FM_DMP_V32(buf, n, p_mm, rdrntp_u);
-+
-+ return n;
-+}
-+
-+static int memac_dump_regs_tx(struct mac_device *h_mac, char *buf, int nn)
-+{
-+ struct memac_regs *p_mm = (struct memac_regs *) h_mac->vaddr;
-+ int n = nn;
-+
-+ FM_DMP_SUBTITLE(buf, n, "\n");
-+ FM_DMP_TITLE(buf, n, p_mm, "FM MAC - MEMAC -%d Tx stats", h_mac->cell_index);
-+
-+
-+ /* Tx Statistics Counter */
-+ FM_DMP_V32(buf, n, p_mm, teoct_l);
-+ FM_DMP_V32(buf, n, p_mm, teoct_u);
-+ FM_DMP_V32(buf, n, p_mm, toct_l);
-+ FM_DMP_V32(buf, n, p_mm, toct_u);
-+ FM_DMP_V32(buf, n, p_mm, txpf_l);
-+ FM_DMP_V32(buf, n, p_mm, txpf_u);
-+ FM_DMP_V32(buf, n, p_mm, tfrm_l);
-+ FM_DMP_V32(buf, n, p_mm, tfrm_u);
-+ FM_DMP_V32(buf, n, p_mm, tfcs_l);
-+ FM_DMP_V32(buf, n, p_mm, tfcs_u);
-+ FM_DMP_V32(buf, n, p_mm, tvlan_l);
-+ FM_DMP_V32(buf, n, p_mm, tvlan_u);
-+ FM_DMP_V32(buf, n, p_mm, terr_l);
-+ FM_DMP_V32(buf, n, p_mm, terr_u);
-+ FM_DMP_V32(buf, n, p_mm, tuca_l);
-+ FM_DMP_V32(buf, n, p_mm, tuca_u);
-+ FM_DMP_V32(buf, n, p_mm, tmca_l);
-+ FM_DMP_V32(buf, n, p_mm, tmca_u);
-+ FM_DMP_V32(buf, n, p_mm, tbca_l);
-+ FM_DMP_V32(buf, n, p_mm, tbca_u);
-+ FM_DMP_V32(buf, n, p_mm, tpkt_l);
-+ FM_DMP_V32(buf, n, p_mm, tpkt_u);
-+ FM_DMP_V32(buf, n, p_mm, tund_l);
-+ FM_DMP_V32(buf, n, p_mm, tund_u);
-+ FM_DMP_V32(buf, n, p_mm, t64_l);
-+ FM_DMP_V32(buf, n, p_mm, t64_u);
-+ FM_DMP_V32(buf, n, p_mm, t127_l);
-+ FM_DMP_V32(buf, n, p_mm, t127_u);
-+ FM_DMP_V32(buf, n, p_mm, t255_l);
-+ FM_DMP_V32(buf, n, p_mm, t255_u);
-+ FM_DMP_V32(buf, n, p_mm, t511_l);
-+ FM_DMP_V32(buf, n, p_mm, t511_u);
-+ FM_DMP_V32(buf, n, p_mm, t1023_l);
-+ FM_DMP_V32(buf, n, p_mm, t1023_u);
-+ FM_DMP_V32(buf, n, p_mm, t1518_l);
-+ FM_DMP_V32(buf, n, p_mm, t1518_u);
-+ FM_DMP_V32(buf, n, p_mm, t1519x_l);
-+ FM_DMP_V32(buf, n, p_mm, t1519x_u);
-+ FM_DMP_V32(buf, n, p_mm, tcnp_l);
-+ FM_DMP_V32(buf, n, p_mm, tcnp_u);
-+
-+ return n;
-+}
-+
-+int fm_mac_dump_regs(struct mac_device *h_mac, char *buf, int nn)
-+{
-+ int n = nn;
-+
-+ n = h_mac->dump_mac_regs(h_mac, buf, n);
-+
-+ return n;
-+}
-+EXPORT_SYMBOL(fm_mac_dump_regs);
-+
-+int fm_mac_dump_rx_stats(struct mac_device *h_mac, char *buf, int nn)
-+{
-+ int n = nn;
-+
-+ if(h_mac->dump_mac_rx_stats)
-+ n = h_mac->dump_mac_rx_stats(h_mac, buf, n);
-+
-+ return n;
-+}
-+EXPORT_SYMBOL(fm_mac_dump_rx_stats);
-+
-+int fm_mac_dump_tx_stats(struct mac_device *h_mac, char *buf, int nn)
-+{
-+ int n = nn;
-+
-+ if(h_mac->dump_mac_tx_stats)
-+ n = h_mac->dump_mac_tx_stats(h_mac, buf, n);
-+
-+ return n;
-+}
-+EXPORT_SYMBOL(fm_mac_dump_tx_stats);
-+
-+static void __cold setup_dtsec(struct mac_device *mac_dev)
-+{
-+ mac_dev->init_phy = dtsec_init_phy;
-+ mac_dev->init = init;
-+ mac_dev->start = start;
-+ mac_dev->stop = stop;
-+ mac_dev->set_promisc = fm_mac_set_promiscuous;
-+ mac_dev->change_addr = fm_mac_modify_mac_addr;
-+ mac_dev->set_multi = set_multi;
-+ mac_dev->uninit = uninit;
-+ mac_dev->ptp_enable = fm_mac_enable_1588_time_stamp;
-+ mac_dev->ptp_disable = fm_mac_disable_1588_time_stamp;
-+ mac_dev->get_mac_handle = get_mac_handle;
-+ mac_dev->set_tx_pause = fm_mac_set_tx_pause_frames;
-+ mac_dev->set_rx_pause = fm_mac_set_rx_pause_frames;
-+ mac_dev->fm_rtc_enable = fm_rtc_enable;
-+ mac_dev->fm_rtc_disable = fm_rtc_disable;
-+ mac_dev->fm_rtc_get_cnt = fm_rtc_get_cnt;
-+ mac_dev->fm_rtc_set_cnt = fm_rtc_set_cnt;
-+ mac_dev->fm_rtc_get_drift = fm_rtc_get_drift;
-+ mac_dev->fm_rtc_set_drift = fm_rtc_set_drift;
-+ mac_dev->fm_rtc_set_alarm = fm_rtc_set_alarm;
-+ mac_dev->fm_rtc_set_fiper = fm_rtc_set_fiper;
-+ mac_dev->set_wol = fm_mac_set_wol;
-+ mac_dev->dump_mac_regs = dtsec_dump_regs;
-+}
-+
-+static void __cold setup_xgmac(struct mac_device *mac_dev)
-+{
-+ mac_dev->init_phy = xgmac_init_phy;
-+ mac_dev->init = init;
-+ mac_dev->start = start;
-+ mac_dev->stop = stop;
-+ mac_dev->set_promisc = fm_mac_set_promiscuous;
-+ mac_dev->change_addr = fm_mac_modify_mac_addr;
-+ mac_dev->set_multi = set_multi;
-+ mac_dev->uninit = uninit;
-+ mac_dev->get_mac_handle = get_mac_handle;
-+ mac_dev->set_tx_pause = fm_mac_set_tx_pause_frames;
-+ mac_dev->set_rx_pause = fm_mac_set_rx_pause_frames;
-+ mac_dev->set_wol = fm_mac_set_wol;
-+ mac_dev->dump_mac_regs = xgmac_dump_regs;
-+}
-+
-+static void __cold setup_memac(struct mac_device *mac_dev)
-+{
-+ mac_dev->init_phy = memac_init_phy;
-+ mac_dev->init = memac_init;
-+ mac_dev->start = start;
-+ mac_dev->stop = stop;
-+ mac_dev->set_promisc = fm_mac_set_promiscuous;
-+ mac_dev->change_addr = fm_mac_modify_mac_addr;
-+ mac_dev->set_multi = set_multi;
-+ mac_dev->uninit = uninit;
-+ mac_dev->get_mac_handle = get_mac_handle;
-+ mac_dev->set_tx_pause = fm_mac_set_tx_pause_frames;
-+ mac_dev->set_rx_pause = fm_mac_set_rx_pause_frames;
-+ mac_dev->fm_rtc_enable = fm_rtc_enable;
-+ mac_dev->fm_rtc_disable = fm_rtc_disable;
-+ mac_dev->fm_rtc_get_cnt = fm_rtc_get_cnt;
-+ mac_dev->fm_rtc_set_cnt = fm_rtc_set_cnt;
-+ mac_dev->fm_rtc_get_drift = fm_rtc_get_drift;
-+ mac_dev->fm_rtc_set_drift = fm_rtc_set_drift;
-+ mac_dev->fm_rtc_set_alarm = fm_rtc_set_alarm;
-+ mac_dev->fm_rtc_set_fiper = fm_rtc_set_fiper;
-+ mac_dev->set_wol = fm_mac_set_wol;
-+ mac_dev->dump_mac_regs = memac_dump_regs;
-+ mac_dev->dump_mac_rx_stats = memac_dump_regs_rx;
-+ mac_dev->dump_mac_tx_stats = memac_dump_regs_tx;
-+}
-+
-+void (*const mac_setup[])(struct mac_device *mac_dev) = {
-+ [DTSEC] = setup_dtsec,
-+ [XGMAC] = setup_xgmac,
-+ [MEMAC] = setup_memac
-+};
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_dpaa/mac.c
-@@ -0,0 +1,489 @@
-+/* Copyright 2008-2012 Freescale Semiconductor, Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#ifdef CONFIG_FSL_DPAA_ETH_DEBUG
-+#define pr_fmt(fmt) \
-+ KBUILD_MODNAME ": %s:%hu:%s() " fmt, \
-+ KBUILD_BASENAME".c", __LINE__, __func__
-+#else
-+#define pr_fmt(fmt) \
-+ KBUILD_MODNAME ": " fmt
-+#endif
-+
-+#include <linux/init.h>
-+#include <linux/module.h>
-+#include <linux/of_address.h>
-+#include <linux/of_platform.h>
-+#include <linux/of_net.h>
-+#include <linux/of_mdio.h>
-+#include <linux/phy_fixed.h>
-+#include <linux/device.h>
-+#include <linux/phy.h>
-+#include <linux/io.h>
-+
-+#include "lnxwrp_fm_ext.h"
-+
-+#include "mac.h"
-+
-+#define DTSEC_SUPPORTED \
-+ (SUPPORTED_10baseT_Half \
-+ | SUPPORTED_10baseT_Full \
-+ | SUPPORTED_100baseT_Half \
-+ | SUPPORTED_100baseT_Full \
-+ | SUPPORTED_Autoneg \
-+ | SUPPORTED_Pause \
-+ | SUPPORTED_Asym_Pause \
-+ | SUPPORTED_MII)
-+
-+static const char phy_str[][11] = {
-+ [PHY_INTERFACE_MODE_MII] = "mii",
-+ [PHY_INTERFACE_MODE_GMII] = "gmii",
-+ [PHY_INTERFACE_MODE_SGMII] = "sgmii",
-+ [PHY_INTERFACE_MODE_QSGMII] = "qsgmii",
-+ [PHY_INTERFACE_MODE_TBI] = "tbi",
-+ [PHY_INTERFACE_MODE_RMII] = "rmii",
-+ [PHY_INTERFACE_MODE_RGMII] = "rgmii",
-+ [PHY_INTERFACE_MODE_RGMII_ID] = "rgmii-id",
-+ [PHY_INTERFACE_MODE_RGMII_RXID] = "rgmii-rxid",
-+ [PHY_INTERFACE_MODE_RGMII_TXID] = "rgmii-txid",
-+ [PHY_INTERFACE_MODE_RTBI] = "rtbi",
-+ [PHY_INTERFACE_MODE_XGMII] = "xgmii",
-+ [PHY_INTERFACE_MODE_2500SGMII] = "sgmii-2500",
-+};
-+
-+static phy_interface_t __pure __attribute__((nonnull)) str2phy(const char *str)
-+{
-+ int i;
-+
-+ for (i = 0; i < ARRAY_SIZE(phy_str); i++)
-+ if (strcmp(str, phy_str[i]) == 0)
-+ return (phy_interface_t)i;
-+
-+ return PHY_INTERFACE_MODE_MII;
-+}
-+
-+static const uint16_t phy2speed[] = {
-+ [PHY_INTERFACE_MODE_MII] = SPEED_100,
-+ [PHY_INTERFACE_MODE_GMII] = SPEED_1000,
-+ [PHY_INTERFACE_MODE_SGMII] = SPEED_1000,
-+ [PHY_INTERFACE_MODE_QSGMII] = SPEED_1000,
-+ [PHY_INTERFACE_MODE_TBI] = SPEED_1000,
-+ [PHY_INTERFACE_MODE_RMII] = SPEED_100,
-+ [PHY_INTERFACE_MODE_RGMII] = SPEED_1000,
-+ [PHY_INTERFACE_MODE_RGMII_ID] = SPEED_1000,
-+ [PHY_INTERFACE_MODE_RGMII_RXID] = SPEED_1000,
-+ [PHY_INTERFACE_MODE_RGMII_TXID] = SPEED_1000,
-+ [PHY_INTERFACE_MODE_RTBI] = SPEED_1000,
-+ [PHY_INTERFACE_MODE_XGMII] = SPEED_10000,
-+ [PHY_INTERFACE_MODE_2500SGMII] = SPEED_2500,
-+};
-+
-+static struct mac_device * __cold
-+alloc_macdev(struct device *dev, size_t sizeof_priv,
-+ void (*setup)(struct mac_device *mac_dev))
-+{
-+ struct mac_device *mac_dev;
-+
-+ mac_dev = devm_kzalloc(dev, sizeof(*mac_dev) + sizeof_priv, GFP_KERNEL);
-+ if (unlikely(mac_dev == NULL))
-+ mac_dev = ERR_PTR(-ENOMEM);
-+ else {
-+ mac_dev->dev = dev;
-+ dev_set_drvdata(dev, mac_dev);
-+ setup(mac_dev);
-+ }
-+
-+ return mac_dev;
-+}
-+
-+static int __cold free_macdev(struct mac_device *mac_dev)
-+{
-+ dev_set_drvdata(mac_dev->dev, NULL);
-+
-+ return mac_dev->uninit(mac_dev->get_mac_handle(mac_dev));
-+}
-+
-+static const struct of_device_id mac_match[] = {
-+ [DTSEC] = {
-+ .compatible = "fsl,fman-dtsec"
-+ },
-+ [XGMAC] = {
-+ .compatible = "fsl,fman-xgec"
-+ },
-+ [MEMAC] = {
-+ .compatible = "fsl,fman-memac"
-+ },
-+ {}
-+};
-+MODULE_DEVICE_TABLE(of, mac_match);
-+
-+static int __cold mac_probe(struct platform_device *_of_dev)
-+{
-+ int _errno, i;
-+ struct device *dev;
-+ struct device_node *mac_node, *dev_node;
-+ struct mac_device *mac_dev;
-+ struct platform_device *of_dev;
-+ struct resource res;
-+ const uint8_t *mac_addr;
-+ const char *char_prop;
-+ int nph;
-+ u32 cell_index;
-+ const struct of_device_id *match;
-+
-+ dev = &_of_dev->dev;
-+ mac_node = dev->of_node;
-+
-+ match = of_match_device(mac_match, dev);
-+ if (!match)
-+ return -EINVAL;
-+
-+ for (i = 0; i < ARRAY_SIZE(mac_match) - 1 && match != mac_match + i;
-+ i++)
-+ ;
-+ BUG_ON(i >= ARRAY_SIZE(mac_match) - 1);
-+
-+ mac_dev = alloc_macdev(dev, mac_sizeof_priv[i], mac_setup[i]);
-+ if (IS_ERR(mac_dev)) {
-+ _errno = PTR_ERR(mac_dev);
-+ dev_err(dev, "alloc_macdev() = %d\n", _errno);
-+ goto _return;
-+ }
-+
-+ INIT_LIST_HEAD(&mac_dev->mc_addr_list);
-+
-+ /* Get the FM node */
-+ dev_node = of_get_parent(mac_node);
-+ if (unlikely(dev_node == NULL)) {
-+ dev_err(dev, "of_get_parent(%s) failed\n",
-+ mac_node->full_name);
-+ _errno = -EINVAL;
-+ goto _return_dev_set_drvdata;
-+ }
-+
-+ of_dev = of_find_device_by_node(dev_node);
-+ if (unlikely(of_dev == NULL)) {
-+ dev_err(dev, "of_find_device_by_node(%s) failed\n",
-+ dev_node->full_name);
-+ _errno = -EINVAL;
-+ goto _return_of_node_put;
-+ }
-+
-+ mac_dev->fm_dev = fm_bind(&of_dev->dev);
-+ if (unlikely(mac_dev->fm_dev == NULL)) {
-+ dev_err(dev, "fm_bind(%s) failed\n", dev_node->full_name);
-+ _errno = -ENODEV;
-+ goto _return_of_node_put;
-+ }
-+
-+ mac_dev->fm = (void *)fm_get_handle(mac_dev->fm_dev);
-+ of_node_put(dev_node);
-+
-+ /* Get the address of the memory mapped registers */
-+ _errno = of_address_to_resource(mac_node, 0, &res);
-+ if (unlikely(_errno < 0)) {
-+ dev_err(dev, "of_address_to_resource(%s) = %d\n",
-+ mac_node->full_name, _errno);
-+ goto _return_dev_set_drvdata;
-+ }
-+
-+ mac_dev->res = __devm_request_region(
-+ dev,
-+ fm_get_mem_region(mac_dev->fm_dev),
-+ res.start, res.end + 1 - res.start, "mac");
-+ if (unlikely(mac_dev->res == NULL)) {
-+ dev_err(dev, "__devm_request_mem_region(mac) failed\n");
-+ _errno = -EBUSY;
-+ goto _return_dev_set_drvdata;
-+ }
-+
-+ mac_dev->vaddr = devm_ioremap(dev, mac_dev->res->start,
-+ mac_dev->res->end + 1
-+ - mac_dev->res->start);
-+ if (unlikely(mac_dev->vaddr == NULL)) {
-+ dev_err(dev, "devm_ioremap() failed\n");
-+ _errno = -EIO;
-+ goto _return_dev_set_drvdata;
-+ }
-+
-+#define TBIPA_OFFSET 0x1c
-+#define TBIPA_DEFAULT_ADDR 5 /* override if used as external PHY addr. */
-+ mac_dev->tbi_node = of_parse_phandle(mac_node, "tbi-handle", 0);
-+ if (mac_dev->tbi_node) {
-+ u32 tbiaddr = TBIPA_DEFAULT_ADDR;
-+ const __be32 *tbi_reg;
-+ void __iomem *addr;
-+
-+ tbi_reg = of_get_property(mac_dev->tbi_node, "reg", NULL);
-+ if (tbi_reg)
-+ tbiaddr = be32_to_cpup(tbi_reg);
-+ addr = mac_dev->vaddr + TBIPA_OFFSET;
-+ /* TODO: out_be32 does not exist on ARM */
-+ out_be32(addr, tbiaddr);
-+ }
-+
-+ if (!of_device_is_available(mac_node)) {
-+ devm_iounmap(dev, mac_dev->vaddr);
-+ __devm_release_region(dev, fm_get_mem_region(mac_dev->fm_dev),
-+ res.start, res.end + 1 - res.start);
-+ fm_unbind(mac_dev->fm_dev);
-+ devm_kfree(dev, mac_dev);
-+ dev_set_drvdata(dev, NULL);
-+ return -ENODEV;
-+ }
-+
-+ /* Get the cell-index */
-+ _errno = of_property_read_u32(mac_node, "cell-index", &cell_index);
-+ if (unlikely(_errno)) {
-+ dev_err(dev, "Cannot read cell-index of mac node %s from device tree\n",
-+ mac_node->full_name);
-+ goto _return_dev_set_drvdata;
-+ }
-+ mac_dev->cell_index = (uint8_t)cell_index;
-+ if (mac_dev->cell_index >= 8)
-+ mac_dev->cell_index -= 8;
-+
-+ /* Get the MAC address */
-+ mac_addr = of_get_mac_address(mac_node);
-+ if (unlikely(mac_addr == NULL)) {
-+ dev_err(dev, "of_get_mac_address(%s) failed\n",
-+ mac_node->full_name);
-+ _errno = -EINVAL;
-+ goto _return_dev_set_drvdata;
-+ }
-+ memcpy(mac_dev->addr, mac_addr, sizeof(mac_dev->addr));
-+
-+ /* Verify the number of port handles */
-+ nph = of_count_phandle_with_args(mac_node, "fsl,fman-ports", NULL);
-+ if (unlikely(nph < 0)) {
-+ dev_err(dev, "Cannot read port handles of mac node %s from device tree\n",
-+ mac_node->full_name);
-+ _errno = nph;
-+ goto _return_dev_set_drvdata;
-+ }
-+
-+ if (nph != ARRAY_SIZE(mac_dev->port_dev)) {
-+ dev_err(dev, "Not supported number of port handles of mac node %s from device tree\n",
-+ mac_node->full_name);
-+ _errno = -EINVAL;
-+ goto _return_dev_set_drvdata;
-+ }
-+
-+ for_each_port_device(i, mac_dev->port_dev) {
-+ dev_node = of_parse_phandle(mac_node, "fsl,fman-ports", i);
-+ if (unlikely(dev_node == NULL)) {
-+ dev_err(dev, "Cannot find port node referenced by mac node %s from device tree\n",
-+ mac_node->full_name);
-+ _errno = -EINVAL;
-+ goto _return_of_node_put;
-+ }
-+
-+ of_dev = of_find_device_by_node(dev_node);
-+ if (unlikely(of_dev == NULL)) {
-+ dev_err(dev, "of_find_device_by_node(%s) failed\n",
-+ dev_node->full_name);
-+ _errno = -EINVAL;
-+ goto _return_of_node_put;
-+ }
-+
-+ mac_dev->port_dev[i] = fm_port_bind(&of_dev->dev);
-+ if (unlikely(mac_dev->port_dev[i] == NULL)) {
-+ dev_err(dev, "dev_get_drvdata(%s) failed\n",
-+ dev_node->full_name);
-+ _errno = -EINVAL;
-+ goto _return_of_node_put;
-+ }
-+ of_node_put(dev_node);
-+ }
-+
-+ /* Get the PHY connection type */
-+ _errno = of_property_read_string(mac_node, "phy-connection-type",
-+ &char_prop);
-+ if (unlikely(_errno)) {
-+ dev_warn(dev,
-+ "Cannot read PHY connection type of mac node %s from device tree. Defaulting to MII\n",
-+ mac_node->full_name);
-+ mac_dev->phy_if = PHY_INTERFACE_MODE_MII;
-+ } else
-+ mac_dev->phy_if = str2phy(char_prop);
-+
-+ mac_dev->link = false;
-+ mac_dev->half_duplex = false;
-+ mac_dev->speed = phy2speed[mac_dev->phy_if];
-+ mac_dev->max_speed = mac_dev->speed;
-+ mac_dev->if_support = DTSEC_SUPPORTED;
-+ /* We don't support half-duplex in SGMII mode */
-+ if (strstr(char_prop, "sgmii") || strstr(char_prop, "qsgmii") ||
-+ strstr(char_prop, "sgmii-2500"))
-+ mac_dev->if_support &= ~(SUPPORTED_10baseT_Half |
-+ SUPPORTED_100baseT_Half);
-+
-+ /* Gigabit support (no half-duplex) */
-+ if (mac_dev->max_speed == SPEED_1000 ||
-+ mac_dev->max_speed == SPEED_2500)
-+ mac_dev->if_support |= SUPPORTED_1000baseT_Full;
-+
-+ /* The 10G interface only supports one mode */
-+ if (strstr(char_prop, "xgmii"))
-+ mac_dev->if_support = SUPPORTED_10000baseT_Full;
-+
-+ /* Get the rest of the PHY information */
-+ mac_dev->phy_node = of_parse_phandle(mac_node, "phy-handle", 0);
-+ if (!mac_dev->phy_node) {
-+ struct phy_device *phy;
-+
-+ if (!of_phy_is_fixed_link(mac_node)) {
-+ dev_err(dev, "Wrong PHY information of mac node %s\n",
-+ mac_node->full_name);
-+ goto _return_dev_set_drvdata;
-+ }
-+
-+ _errno = of_phy_register_fixed_link(mac_node);
-+ if (_errno)
-+ goto _return_dev_set_drvdata;
-+
-+ mac_dev->fixed_link = devm_kzalloc(mac_dev->dev,
-+ sizeof(*mac_dev->fixed_link),
-+ GFP_KERNEL);
-+ if (!mac_dev->fixed_link)
-+ goto _return_dev_set_drvdata;
-+
-+ mac_dev->phy_node = of_node_get(mac_node);
-+ phy = of_phy_find_device(mac_dev->phy_node);
-+ if (!phy)
-+ goto _return_dev_set_drvdata;
-+
-+ mac_dev->fixed_link->link = phy->link;
-+ mac_dev->fixed_link->speed = phy->speed;
-+ mac_dev->fixed_link->duplex = phy->duplex;
-+ mac_dev->fixed_link->pause = phy->pause;
-+ mac_dev->fixed_link->asym_pause = phy->asym_pause;
-+ }
-+
-+ _errno = mac_dev->init(mac_dev);
-+ if (unlikely(_errno < 0)) {
-+ dev_err(dev, "mac_dev->init() = %d\n", _errno);
-+ goto _return_dev_set_drvdata;
-+ }
-+
-+ /* pause frame autonegotiation enabled*/
-+ mac_dev->autoneg_pause = true;
-+
-+ /* by intializing the values to false, force FMD to enable PAUSE frames
-+ * on RX and TX
-+ */
-+ mac_dev->rx_pause_req = mac_dev->tx_pause_req = true;
-+ mac_dev->rx_pause_active = mac_dev->tx_pause_active = false;
-+ _errno = set_mac_active_pause(mac_dev, true, true);
-+ if (unlikely(_errno < 0))
-+ dev_err(dev, "set_mac_active_pause() = %d\n", _errno);
-+
-+ dev_info(dev,
-+ "FMan MAC address: %02hx:%02hx:%02hx:%02hx:%02hx:%02hx\n",
-+ mac_dev->addr[0], mac_dev->addr[1], mac_dev->addr[2],
-+ mac_dev->addr[3], mac_dev->addr[4], mac_dev->addr[5]);
-+
-+ goto _return;
-+
-+_return_of_node_put:
-+ of_node_put(dev_node);
-+_return_dev_set_drvdata:
-+ dev_set_drvdata(dev, NULL);
-+_return:
-+ return _errno;
-+}
-+
-+static int __cold mac_remove(struct platform_device *of_dev)
-+{
-+ int i, _errno;
-+ struct device *dev;
-+ struct mac_device *mac_dev;
-+
-+ dev = &of_dev->dev;
-+ mac_dev = (struct mac_device *)dev_get_drvdata(dev);
-+
-+ for_each_port_device(i, mac_dev->port_dev)
-+ fm_port_unbind(mac_dev->port_dev[i]);
-+
-+ fm_unbind(mac_dev->fm_dev);
-+
-+ _errno = free_macdev(mac_dev);
-+
-+ return _errno;
-+}
-+
-+static struct platform_driver mac_driver = {
-+ .driver = {
-+ .name = KBUILD_MODNAME,
-+ .of_match_table = mac_match,
-+ .owner = THIS_MODULE,
-+ },
-+ .probe = mac_probe,
-+ .remove = mac_remove
-+};
-+
-+static int __init __cold mac_load(void)
-+{
-+ int _errno;
-+
-+ pr_debug(KBUILD_MODNAME ": -> %s:%s()\n",
-+ KBUILD_BASENAME".c", __func__);
-+
-+ pr_info(KBUILD_MODNAME ": %s\n", mac_driver_description);
-+
-+ _errno = platform_driver_register(&mac_driver);
-+ if (unlikely(_errno < 0)) {
-+ pr_err(KBUILD_MODNAME ": %s:%hu:%s(): platform_driver_register() = %d\n",
-+ KBUILD_BASENAME".c", __LINE__, __func__, _errno);
-+ goto _return;
-+ }
-+
-+ goto _return;
-+
-+_return:
-+ pr_debug(KBUILD_MODNAME ": %s:%s() ->\n",
-+ KBUILD_BASENAME".c", __func__);
-+
-+ return _errno;
-+}
-+module_init(mac_load);
-+
-+static void __exit __cold mac_unload(void)
-+{
-+ pr_debug(KBUILD_MODNAME ": -> %s:%s()\n",
-+ KBUILD_BASENAME".c", __func__);
-+
-+ platform_driver_unregister(&mac_driver);
-+
-+ pr_debug(KBUILD_MODNAME ": %s:%s() ->\n",
-+ KBUILD_BASENAME".c", __func__);
-+}
-+module_exit(mac_unload);
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_dpaa/mac.h
-@@ -0,0 +1,135 @@
-+/* Copyright 2008-2011 Freescale Semiconductor, Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#ifndef __MAC_H
-+#define __MAC_H
-+
-+#include <linux/device.h> /* struct device, BUS_ID_SIZE */
-+#include <linux/if_ether.h> /* ETH_ALEN */
-+#include <linux/phy.h> /* phy_interface_t, struct phy_device */
-+#include <linux/list.h>
-+
-+#include "lnxwrp_fsl_fman.h" /* struct port_device */
-+
-+enum {DTSEC, XGMAC, MEMAC};
-+
-+struct mac_device {
-+ struct device *dev;
-+ void *priv;
-+ uint8_t cell_index;
-+ struct resource *res;
-+ void __iomem *vaddr;
-+ uint8_t addr[ETH_ALEN];
-+ bool promisc;
-+
-+ struct fm *fm_dev;
-+ struct fm_port *port_dev[2];
-+
-+ phy_interface_t phy_if;
-+ u32 if_support;
-+ bool link;
-+ bool half_duplex;
-+ uint16_t speed;
-+ uint16_t max_speed;
-+ struct device_node *phy_node;
-+ char fixed_bus_id[MII_BUS_ID_SIZE + 3];
-+ struct device_node *tbi_node;
-+ struct phy_device *phy_dev;
-+ void *fm;
-+ /* List of multicast addresses */
-+ struct list_head mc_addr_list;
-+ struct fixed_phy_status *fixed_link;
-+
-+ bool autoneg_pause;
-+ bool rx_pause_req;
-+ bool tx_pause_req;
-+ bool rx_pause_active;
-+ bool tx_pause_active;
-+
-+ struct fm_mac_dev *(*get_mac_handle)(struct mac_device *mac_dev);
-+ int (*init_phy)(struct net_device *net_dev, struct mac_device *mac_dev);
-+ int (*init)(struct mac_device *mac_dev);
-+ int (*start)(struct mac_device *mac_dev);
-+ int (*stop)(struct mac_device *mac_dev);
-+ int (*set_promisc)(struct fm_mac_dev *fm_mac_dev, bool enable);
-+ int (*change_addr)(struct fm_mac_dev *fm_mac_dev, uint8_t *addr);
-+ int (*set_multi)(struct net_device *net_dev,
-+ struct mac_device *mac_dev);
-+ int (*uninit)(struct fm_mac_dev *fm_mac_dev);
-+ int (*ptp_enable)(struct fm_mac_dev *fm_mac_dev);
-+ int (*ptp_disable)(struct fm_mac_dev *fm_mac_dev);
-+ int (*set_rx_pause)(struct fm_mac_dev *fm_mac_dev, bool en);
-+ int (*set_tx_pause)(struct fm_mac_dev *fm_mac_dev, bool en);
-+ int (*fm_rtc_enable)(struct fm *fm_dev);
-+ int (*fm_rtc_disable)(struct fm *fm_dev);
-+ int (*fm_rtc_get_cnt)(struct fm *fm_dev, uint64_t *ts);
-+ int (*fm_rtc_set_cnt)(struct fm *fm_dev, uint64_t ts);
-+ int (*fm_rtc_get_drift)(struct fm *fm_dev, uint32_t *drift);
-+ int (*fm_rtc_set_drift)(struct fm *fm_dev, uint32_t drift);
-+ int (*fm_rtc_set_alarm)(struct fm *fm_dev, uint32_t id, uint64_t time);
-+ int (*fm_rtc_set_fiper)(struct fm *fm_dev, uint32_t id,
-+ uint64_t fiper);
-+#ifdef CONFIG_PTP_1588_CLOCK_DPAA
-+ int (*fm_rtc_enable_interrupt)(struct fm *fm_dev, uint32_t events);
-+ int (*fm_rtc_disable_interrupt)(struct fm *fm_dev, uint32_t events);
-+#endif
-+ int (*set_wol)(struct fm_port *port, struct fm_mac_dev *fm_mac_dev,
-+ bool en);
-+ int (*dump_mac_regs)(struct mac_device *h_mac, char *buf, int nn);
-+ int (*dump_mac_rx_stats)(struct mac_device *h_mac, char *buf, int nn);
-+ int (*dump_mac_tx_stats)(struct mac_device *h_mac, char *buf, int nn);
-+};
-+
-+struct mac_address {
-+ uint8_t addr[ETH_ALEN];
-+ struct list_head list;
-+};
-+
-+#define get_fm_handle(net_dev) \
-+ (((struct dpa_priv_s *)netdev_priv(net_dev))->mac_dev->fm_dev)
-+
-+#define for_each_port_device(i, port_dev) \
-+ for (i = 0; i < ARRAY_SIZE(port_dev); i++)
-+
-+static inline __attribute((nonnull)) void *macdev_priv(
-+ const struct mac_device *mac_dev)
-+{
-+ return (void *)mac_dev + sizeof(*mac_dev);
-+}
-+
-+extern const char *mac_driver_description;
-+extern const size_t mac_sizeof_priv[];
-+extern void (*const mac_setup[])(struct mac_device *mac_dev);
-+
-+int set_mac_active_pause(struct mac_device *mac_dev, bool rx, bool tx);
-+void get_pause_cfg(struct mac_device *mac_dev, bool *rx_pause, bool *tx_pause);
-+
-+#endif /* __MAC_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_dpaa/offline_port.c
-@@ -0,0 +1,848 @@
-+/* Copyright 2011-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+/* Offline Parsing / Host Command port driver for FSL QorIQ FMan.
-+ * Validates device-tree configuration and sets up the offline ports.
-+ */
-+
-+#ifdef CONFIG_FSL_DPAA_ETH_DEBUG
-+#define pr_fmt(fmt) \
-+ KBUILD_MODNAME ": %s:%hu:%s() " fmt, \
-+ KBUILD_BASENAME".c", __LINE__, __func__
-+#else
-+#define pr_fmt(fmt) \
-+ KBUILD_MODNAME ": " fmt
-+#endif
-+
-+
-+#include <linux/init.h>
-+#include <linux/module.h>
-+#include <linux/of_platform.h>
-+#include <linux/fsl_qman.h>
-+
-+#include "offline_port.h"
-+#include "dpaa_eth.h"
-+#include "dpaa_eth_common.h"
-+
-+#define OH_MOD_DESCRIPTION "FSL FMan Offline Parsing port driver"
-+/* Manip extra space and data alignment for fragmentation */
-+#define FRAG_MANIP_SPACE 128
-+#define FRAG_DATA_ALIGN 64
-+
-+
-+MODULE_LICENSE("Dual BSD/GPL");
-+MODULE_AUTHOR("Bogdan Hamciuc <bogdan.hamciuc@freescale.com>");
-+MODULE_DESCRIPTION(OH_MOD_DESCRIPTION);
-+
-+
-+static const struct of_device_id oh_port_match_table[] = {
-+ {
-+ .compatible = "fsl,dpa-oh"
-+ },
-+ {
-+ .compatible = "fsl,dpa-oh-shared"
-+ },
-+ {}
-+};
-+MODULE_DEVICE_TABLE(of, oh_port_match_table);
-+
-+#ifdef CONFIG_PM
-+
-+static int oh_suspend(struct device *dev)
-+{
-+ struct dpa_oh_config_s *oh_config;
-+
-+ oh_config = dev_get_drvdata(dev);
-+ return fm_port_suspend(oh_config->oh_port);
-+}
-+
-+static int oh_resume(struct device *dev)
-+{
-+ struct dpa_oh_config_s *oh_config;
-+
-+ oh_config = dev_get_drvdata(dev);
-+ return fm_port_resume(oh_config->oh_port);
-+}
-+
-+static const struct dev_pm_ops oh_pm_ops = {
-+ .suspend = oh_suspend,
-+ .resume = oh_resume,
-+};
-+
-+#define OH_PM_OPS (&oh_pm_ops)
-+
-+#else /* CONFIG_PM */
-+
-+#define OH_PM_OPS NULL
-+
-+#endif /* CONFIG_PM */
-+
-+/* Creates Frame Queues */
-+static uint32_t oh_fq_create(struct qman_fq *fq,
-+ uint32_t fq_id, uint16_t channel,
-+ uint16_t wq_id)
-+{
-+ struct qm_mcc_initfq fq_opts;
-+ uint32_t create_flags, init_flags;
-+ uint32_t ret = 0;
-+
-+ if (fq == NULL)
-+ return 1;
-+
-+ /* Set flags for FQ create */
-+ create_flags = QMAN_FQ_FLAG_LOCKED | QMAN_FQ_FLAG_TO_DCPORTAL;
-+
-+ /* Create frame queue */
-+ ret = qman_create_fq(fq_id, create_flags, fq);
-+ if (ret != 0)
-+ return 1;
-+
-+ /* Set flags for FQ init */
-+ init_flags = QMAN_INITFQ_FLAG_SCHED;
-+
-+ /* Set FQ init options. Specify destination WQ ID and channel */
-+ fq_opts.we_mask = QM_INITFQ_WE_DESTWQ;
-+ fq_opts.fqd.dest.wq = wq_id;
-+ fq_opts.fqd.dest.channel = channel;
-+
-+ /* Initialize frame queue */
-+ ret = qman_init_fq(fq, init_flags, &fq_opts);
-+ if (ret != 0) {
-+ qman_destroy_fq(fq, 0);
-+ return 1;
-+ }
-+
-+ return 0;
-+}
-+
-+static void dump_fq(struct device *dev, int fqid, uint16_t channel)
-+{
-+ if (channel) {
-+ /* display fqs with a valid (!= 0) destination channel */
-+ dev_info(dev, "FQ ID:%d Channel ID:%d\n", fqid, channel);
-+ }
-+}
-+
-+static void dump_fq_duple(struct device *dev, struct qman_fq *fqs,
-+ int fqs_count, uint16_t channel_id)
-+{
-+ int i;
-+ for (i = 0; i < fqs_count; i++)
-+ dump_fq(dev, (fqs + i)->fqid, channel_id);
-+}
-+
-+static void dump_oh_config(struct device *dev, struct dpa_oh_config_s *conf)
-+{
-+ struct list_head *fq_list;
-+ struct fq_duple *fqd;
-+ int i;
-+
-+ dev_info(dev, "Default egress frame queue: %d\n", conf->default_fqid);
-+ dev_info(dev, "Default error frame queue: %d\n", conf->error_fqid);
-+
-+ /* TX queues (old initialization) */
-+ dev_info(dev, "Initialized queues:");
-+ for (i = 0; i < conf->egress_cnt; i++)
-+ dump_fq_duple(dev, conf->egress_fqs, conf->egress_cnt,
-+ conf->channel);
-+
-+ /* initialized ingress queues */
-+ list_for_each(fq_list, &conf->fqs_ingress_list) {
-+ fqd = list_entry(fq_list, struct fq_duple, fq_list);
-+ dump_fq_duple(dev, fqd->fqs, fqd->fqs_count, fqd->channel_id);
-+ }
-+
-+ /* initialized egress queues */
-+ list_for_each(fq_list, &conf->fqs_egress_list) {
-+ fqd = list_entry(fq_list, struct fq_duple, fq_list);
-+ dump_fq_duple(dev, fqd->fqs, fqd->fqs_count, fqd->channel_id);
-+ }
-+}
-+
-+/* Destroys Frame Queues */
-+static void oh_fq_destroy(struct qman_fq *fq)
-+{
-+ int _errno = 0;
-+
-+ _errno = qman_retire_fq(fq, NULL);
-+ if (unlikely(_errno < 0))
-+ pr_err(KBUILD_MODNAME": %s:%hu:%s(): qman_retire_fq(%u)=%d\n",
-+ KBUILD_BASENAME".c", __LINE__, __func__,
-+ qman_fq_fqid(fq), _errno);
-+
-+ _errno = qman_oos_fq(fq);
-+ if (unlikely(_errno < 0)) {
-+ pr_err(KBUILD_MODNAME": %s:%hu:%s(): qman_oos_fq(%u)=%d\n",
-+ KBUILD_BASENAME".c", __LINE__, __func__,
-+ qman_fq_fqid(fq), _errno);
-+ }
-+
-+ qman_destroy_fq(fq, 0);
-+}
-+
-+/* Allocation code for the OH port's PCD frame queues */
-+static int __cold oh_alloc_pcd_fqids(struct device *dev,
-+ uint32_t num,
-+ uint8_t alignment,
-+ uint32_t *base_fqid)
-+{
-+ dev_crit(dev, "callback not implemented!\n");
-+ BUG();
-+
-+ return 0;
-+}
-+
-+static int __cold oh_free_pcd_fqids(struct device *dev, uint32_t base_fqid)
-+{
-+ dev_crit(dev, "callback not implemented!\n");
-+ BUG();
-+
-+ return 0;
-+}
-+
-+static void oh_set_buffer_layout(struct fm_port *port,
-+ struct dpa_buffer_layout_s *layout)
-+{
-+ struct fm_port_params params;
-+
-+ layout->priv_data_size = DPA_TX_PRIV_DATA_SIZE;
-+ layout->parse_results = true;
-+ layout->hash_results = true;
-+ layout->time_stamp = false;
-+
-+ fm_port_get_buff_layout_ext_params(port, &params);
-+ layout->manip_extra_space = params.manip_extra_space;
-+ layout->data_align = params.data_align;
-+}
-+
-+static int
-+oh_port_probe(struct platform_device *_of_dev)
-+{
-+ struct device *dpa_oh_dev;
-+ struct device_node *dpa_oh_node;
-+ int lenp, _errno = 0, fq_idx, duple_idx;
-+ int n_size, i, j, ret, duples_count;
-+ struct platform_device *oh_of_dev;
-+ struct device_node *oh_node, *bpool_node = NULL, *root_node;
-+ struct device *oh_dev;
-+ struct dpa_oh_config_s *oh_config = NULL;
-+ const __be32 *oh_all_queues;
-+ const __be32 *channel_ids;
-+ const __be32 *oh_tx_queues;
-+ uint32_t queues_count;
-+ uint32_t crt_fqid_base;
-+ uint32_t crt_fq_count;
-+ bool frag_enabled = false;
-+ struct fm_port_params oh_port_tx_params;
-+ struct fm_port_pcd_param oh_port_pcd_params;
-+ struct dpa_buffer_layout_s buf_layout;
-+
-+ /* True if the current partition owns the OH port. */
-+ bool init_oh_port;
-+
-+ const struct of_device_id *match;
-+ int crt_ext_pools_count;
-+ u32 ext_pool_size;
-+ u32 port_id;
-+ u32 channel_id;
-+
-+ int channel_ids_count;
-+ int channel_idx;
-+ struct fq_duple *fqd;
-+ struct list_head *fq_list, *fq_list_tmp;
-+
-+ const __be32 *bpool_cfg;
-+ uint32_t bpid;
-+
-+ memset(&oh_port_tx_params, 0, sizeof(oh_port_tx_params));
-+ dpa_oh_dev = &_of_dev->dev;
-+ dpa_oh_node = dpa_oh_dev->of_node;
-+ BUG_ON(dpa_oh_node == NULL);
-+
-+ match = of_match_device(oh_port_match_table, dpa_oh_dev);
-+ if (!match)
-+ return -EINVAL;
-+
-+ dev_dbg(dpa_oh_dev, "Probing OH port...\n");
-+
-+ /* Find the referenced OH node */
-+ oh_node = of_parse_phandle(dpa_oh_node, "fsl,fman-oh-port", 0);
-+ if (oh_node == NULL) {
-+ dev_err(dpa_oh_dev,
-+ "Can't find OH node referenced from node %s\n",
-+ dpa_oh_node->full_name);
-+ return -EINVAL;
-+ }
-+ dev_info(dpa_oh_dev, "Found OH node handle compatible with %s\n",
-+ match->compatible);
-+
-+ _errno = of_property_read_u32(oh_node, "cell-index", &port_id);
-+ if (_errno) {
-+ dev_err(dpa_oh_dev, "No port id found in node %s\n",
-+ dpa_oh_node->full_name);
-+ goto return_kfree;
-+ }
-+
-+ _errno = of_property_read_u32(oh_node, "fsl,qman-channel-id",
-+ &channel_id);
-+ if (_errno) {
-+ dev_err(dpa_oh_dev, "No channel id found in node %s\n",
-+ dpa_oh_node->full_name);
-+ goto return_kfree;
-+ }
-+
-+ oh_of_dev = of_find_device_by_node(oh_node);
-+ BUG_ON(oh_of_dev == NULL);
-+ oh_dev = &oh_of_dev->dev;
-+
-+ /* The OH port must be initialized exactly once.
-+ * The following scenarios are of interest:
-+ * - the node is Linux-private (will always initialize it);
-+ * - the node is shared between two Linux partitions
-+ * (only one of them will initialize it);
-+ * - the node is shared between a Linux and a LWE partition
-+ * (Linux will initialize it) - "fsl,dpa-oh-shared"
-+ */
-+
-+ /* Check if the current partition owns the OH port
-+ * and ought to initialize it. It may be the case that we leave this
-+ * to another (also Linux) partition.
-+ */
-+ init_oh_port = strcmp(match->compatible, "fsl,dpa-oh-shared");
-+
-+ /* If we aren't the "owner" of the OH node, we're done here. */
-+ if (!init_oh_port) {
-+ dev_dbg(dpa_oh_dev,
-+ "Not owning the shared OH port %s, will not initialize it.\n",
-+ oh_node->full_name);
-+ of_node_put(oh_node);
-+ return 0;
-+ }
-+
-+ /* Allocate OH dev private data */
-+ oh_config = devm_kzalloc(dpa_oh_dev, sizeof(*oh_config), GFP_KERNEL);
-+ if (oh_config == NULL) {
-+ dev_err(dpa_oh_dev,
-+ "Can't allocate private data for OH node %s referenced from node %s!\n",
-+ oh_node->full_name, dpa_oh_node->full_name);
-+ _errno = -ENOMEM;
-+ goto return_kfree;
-+ }
-+
-+ INIT_LIST_HEAD(&oh_config->fqs_ingress_list);
-+ INIT_LIST_HEAD(&oh_config->fqs_egress_list);
-+
-+ /* FQs that enter OH port */
-+ lenp = 0;
-+ oh_all_queues = of_get_property(dpa_oh_node,
-+ "fsl,qman-frame-queues-ingress", &lenp);
-+ if (lenp % (2 * sizeof(*oh_all_queues))) {
-+ dev_warn(dpa_oh_dev,
-+ "Wrong ingress queues format for OH node %s referenced from node %s!\n",
-+ oh_node->full_name, dpa_oh_node->full_name);
-+ /* just ignore the last unpaired value */
-+ }
-+
-+ duples_count = lenp / (2 * sizeof(*oh_all_queues));
-+ dev_err(dpa_oh_dev, "Allocating %d ingress frame queues duples\n",
-+ duples_count);
-+ for (duple_idx = 0; duple_idx < duples_count; duple_idx++) {
-+ crt_fqid_base = be32_to_cpu(oh_all_queues[2 * duple_idx]);
-+ crt_fq_count = be32_to_cpu(oh_all_queues[2 * duple_idx + 1]);
-+
-+ fqd = devm_kzalloc(dpa_oh_dev,
-+ sizeof(struct fq_duple), GFP_KERNEL);
-+ if (!fqd) {
-+ dev_err(dpa_oh_dev, "Can't allocate structures for ingress frame queues for OH node %s referenced from node %s!\n",
-+ oh_node->full_name,
-+ dpa_oh_node->full_name);
-+ _errno = -ENOMEM;
-+ goto return_kfree;
-+ }
-+
-+ fqd->fqs = devm_kzalloc(dpa_oh_dev,
-+ crt_fq_count * sizeof(struct qman_fq),
-+ GFP_KERNEL);
-+ if (!fqd->fqs) {
-+ dev_err(dpa_oh_dev, "Can't allocate structures for ingress frame queues for OH node %s referenced from node %s!\n",
-+ oh_node->full_name,
-+ dpa_oh_node->full_name);
-+ _errno = -ENOMEM;
-+ goto return_kfree;
-+ }
-+
-+ for (j = 0; j < crt_fq_count; j++)
-+ (fqd->fqs + j)->fqid = crt_fqid_base + j;
-+ fqd->fqs_count = crt_fq_count;
-+ fqd->channel_id = (uint16_t)channel_id;
-+ list_add(&fqd->fq_list, &oh_config->fqs_ingress_list);
-+ }
-+
-+ /* create the ingress queues */
-+ list_for_each(fq_list, &oh_config->fqs_ingress_list) {
-+ fqd = list_entry(fq_list, struct fq_duple, fq_list);
-+
-+ for (j = 0; j < fqd->fqs_count; j++) {
-+ ret = oh_fq_create(fqd->fqs + j,
-+ (fqd->fqs + j)->fqid,
-+ fqd->channel_id, 3);
-+ if (ret != 0) {
-+ dev_err(dpa_oh_dev, "Unable to create ingress frame queue %d for OH node %s referenced from node %s!\n",
-+ (fqd->fqs + j)->fqid,
-+ oh_node->full_name,
-+ dpa_oh_node->full_name);
-+ _errno = -EINVAL;
-+ goto return_kfree;
-+ }
-+ }
-+ }
-+
-+ /* FQs that exit OH port */
-+ lenp = 0;
-+ oh_all_queues = of_get_property(dpa_oh_node,
-+ "fsl,qman-frame-queues-egress", &lenp);
-+ if (lenp % (2 * sizeof(*oh_all_queues))) {
-+ dev_warn(dpa_oh_dev,
-+ "Wrong egress queues format for OH node %s referenced from node %s!\n",
-+ oh_node->full_name, dpa_oh_node->full_name);
-+ /* just ignore the last unpaired value */
-+ }
-+
-+ duples_count = lenp / (2 * sizeof(*oh_all_queues));
-+ dev_dbg(dpa_oh_dev, "Allocating %d egress frame queues duples\n",
-+ duples_count);
-+ for (duple_idx = 0; duple_idx < duples_count; duple_idx++) {
-+ crt_fqid_base = be32_to_cpu(oh_all_queues[2 * duple_idx]);
-+ crt_fq_count = be32_to_cpu(oh_all_queues[2 * duple_idx + 1]);
-+
-+ fqd = devm_kzalloc(dpa_oh_dev,
-+ sizeof(struct fq_duple), GFP_KERNEL);
-+ if (!fqd) {
-+ dev_err(dpa_oh_dev, "Can't allocate structures for egress frame queues for OH node %s referenced from node %s!\n",
-+ oh_node->full_name,
-+ dpa_oh_node->full_name);
-+ _errno = -ENOMEM;
-+ goto return_kfree;
-+ }
-+
-+ fqd->fqs = devm_kzalloc(dpa_oh_dev,
-+ crt_fq_count * sizeof(struct qman_fq),
-+ GFP_KERNEL);
-+ if (!fqd->fqs) {
-+ dev_err(dpa_oh_dev,
-+ "Can't allocate structures for egress frame queues for OH node %s referenced from node %s!\n",
-+ oh_node->full_name,
-+ dpa_oh_node->full_name);
-+ _errno = -ENOMEM;
-+ goto return_kfree;
-+ }
-+
-+ for (j = 0; j < crt_fq_count; j++)
-+ (fqd->fqs + j)->fqid = crt_fqid_base + j;
-+ fqd->fqs_count = crt_fq_count;
-+ /* channel ID is specified in another attribute */
-+ fqd->channel_id = 0;
-+ list_add_tail(&fqd->fq_list, &oh_config->fqs_egress_list);
-+
-+ /* allocate the queue */
-+
-+ }
-+
-+ /* channel_ids for FQs that exit OH port */
-+ lenp = 0;
-+ channel_ids = of_get_property(dpa_oh_node,
-+ "fsl,qman-channel-ids-egress", &lenp);
-+
-+ channel_ids_count = lenp / (sizeof(*channel_ids));
-+ if (channel_ids_count != duples_count) {
-+ dev_warn(dpa_oh_dev,
-+ "Not all egress queues have a channel id for OH node %s referenced from node %s!\n",
-+ oh_node->full_name, dpa_oh_node->full_name);
-+ /* just ignore the queues that do not have a Channel ID */
-+ }
-+
-+ channel_idx = 0;
-+ list_for_each(fq_list, &oh_config->fqs_egress_list) {
-+ if (channel_idx + 1 > channel_ids_count)
-+ break;
-+ fqd = list_entry(fq_list, struct fq_duple, fq_list);
-+ fqd->channel_id =
-+ (uint16_t)be32_to_cpu(channel_ids[channel_idx++]);
-+ }
-+
-+ /* create egress queues */
-+ list_for_each(fq_list, &oh_config->fqs_egress_list) {
-+ fqd = list_entry(fq_list, struct fq_duple, fq_list);
-+
-+ if (fqd->channel_id == 0) {
-+ /* missing channel id in dts */
-+ continue;
-+ }
-+
-+ for (j = 0; j < fqd->fqs_count; j++) {
-+ ret = oh_fq_create(fqd->fqs + j,
-+ (fqd->fqs + j)->fqid,
-+ fqd->channel_id, 3);
-+ if (ret != 0) {
-+ dev_err(dpa_oh_dev, "Unable to create egress frame queue %d for OH node %s referenced from node %s!\n",
-+ (fqd->fqs + j)->fqid,
-+ oh_node->full_name,
-+ dpa_oh_node->full_name);
-+ _errno = -EINVAL;
-+ goto return_kfree;
-+ }
-+ }
-+ }
-+
-+ /* Read FQ ids/nums for the DPA OH node */
-+ oh_all_queues = of_get_property(dpa_oh_node,
-+ "fsl,qman-frame-queues-oh", &lenp);
-+ if (oh_all_queues == NULL) {
-+ dev_err(dpa_oh_dev,
-+ "No frame queues have been defined for OH node %s referenced from node %s\n",
-+ oh_node->full_name, dpa_oh_node->full_name);
-+ _errno = -EINVAL;
-+ goto return_kfree;
-+ }
-+
-+ /* Check that the OH error and default FQs are there */
-+ BUG_ON(lenp % (2 * sizeof(*oh_all_queues)));
-+ queues_count = lenp / (2 * sizeof(*oh_all_queues));
-+ if (queues_count != 2) {
-+ dev_err(dpa_oh_dev,
-+ "Error and Default queues must be defined for OH node %s referenced from node %s\n",
-+ oh_node->full_name, dpa_oh_node->full_name);
-+ _errno = -EINVAL;
-+ goto return_kfree;
-+ }
-+
-+ /* Read the FQIDs defined for this OH port */
-+ dev_dbg(dpa_oh_dev, "Reading %d queues...\n", queues_count);
-+ fq_idx = 0;
-+
-+ /* Error FQID - must be present */
-+ crt_fqid_base = be32_to_cpu(oh_all_queues[fq_idx++]);
-+ crt_fq_count = be32_to_cpu(oh_all_queues[fq_idx++]);
-+ if (crt_fq_count != 1) {
-+ dev_err(dpa_oh_dev,
-+ "Only 1 Error FQ allowed in OH node %s referenced from node %s (read: %d FQIDs).\n",
-+ oh_node->full_name, dpa_oh_node->full_name,
-+ crt_fq_count);
-+ _errno = -EINVAL;
-+ goto return_kfree;
-+ }
-+ oh_config->error_fqid = crt_fqid_base;
-+ dev_dbg(dpa_oh_dev, "Read Error FQID 0x%x for OH port %s.\n",
-+ oh_config->error_fqid, oh_node->full_name);
-+
-+ /* Default FQID - must be present */
-+ crt_fqid_base = be32_to_cpu(oh_all_queues[fq_idx++]);
-+ crt_fq_count = be32_to_cpu(oh_all_queues[fq_idx++]);
-+ if (crt_fq_count != 1) {
-+ dev_err(dpa_oh_dev,
-+ "Only 1 Default FQ allowed in OH node %s referenced from %s (read: %d FQIDs).\n",
-+ oh_node->full_name, dpa_oh_node->full_name,
-+ crt_fq_count);
-+ _errno = -EINVAL;
-+ goto return_kfree;
-+ }
-+ oh_config->default_fqid = crt_fqid_base;
-+ dev_dbg(dpa_oh_dev, "Read Default FQID 0x%x for OH port %s.\n",
-+ oh_config->default_fqid, oh_node->full_name);
-+
-+ /* TX FQID - presence is optional */
-+ oh_tx_queues = of_get_property(dpa_oh_node, "fsl,qman-frame-queues-tx",
-+ &lenp);
-+ if (oh_tx_queues == NULL) {
-+ dev_dbg(dpa_oh_dev,
-+ "No tx queues have been defined for OH node %s referenced from node %s\n",
-+ oh_node->full_name, dpa_oh_node->full_name);
-+ goto config_port;
-+ }
-+
-+ /* Check that queues-tx has only a base and a count defined */
-+ BUG_ON(lenp % (2 * sizeof(*oh_tx_queues)));
-+ queues_count = lenp / (2 * sizeof(*oh_tx_queues));
-+ if (queues_count != 1) {
-+ dev_err(dpa_oh_dev,
-+ "TX queues must be defined in only one <base count> tuple for OH node %s referenced from node %s\n",
-+ oh_node->full_name, dpa_oh_node->full_name);
-+ _errno = -EINVAL;
-+ goto return_kfree;
-+ }
-+
-+ fq_idx = 0;
-+ crt_fqid_base = be32_to_cpu(oh_tx_queues[fq_idx++]);
-+ crt_fq_count = be32_to_cpu(oh_tx_queues[fq_idx++]);
-+ oh_config->egress_cnt = crt_fq_count;
-+
-+ /* Allocate TX queues */
-+ dev_dbg(dpa_oh_dev, "Allocating %d queues for TX...\n", crt_fq_count);
-+ oh_config->egress_fqs = devm_kzalloc(dpa_oh_dev,
-+ crt_fq_count * sizeof(struct qman_fq), GFP_KERNEL);
-+ if (oh_config->egress_fqs == NULL) {
-+ dev_err(dpa_oh_dev,
-+ "Can't allocate private data for TX queues for OH node %s referenced from node %s!\n",
-+ oh_node->full_name, dpa_oh_node->full_name);
-+ _errno = -ENOMEM;
-+ goto return_kfree;
-+ }
-+
-+ /* Create TX queues */
-+ for (i = 0; i < crt_fq_count; i++) {
-+ ret = oh_fq_create(oh_config->egress_fqs + i,
-+ crt_fqid_base + i, (uint16_t)channel_id, 3);
-+ if (ret != 0) {
-+ dev_err(dpa_oh_dev,
-+ "Unable to create TX frame queue %d for OH node %s referenced from node %s!\n",
-+ crt_fqid_base + i, oh_node->full_name,
-+ dpa_oh_node->full_name);
-+ _errno = -EINVAL;
-+ goto return_kfree;
-+ }
-+ }
-+
-+config_port:
-+ /* Get a handle to the fm_port so we can set
-+ * its configuration params
-+ */
-+ oh_config->oh_port = fm_port_bind(oh_dev);
-+ if (oh_config->oh_port == NULL) {
-+ dev_err(dpa_oh_dev, "NULL drvdata from fm port dev %s!\n",
-+ oh_node->full_name);
-+ _errno = -EINVAL;
-+ goto return_kfree;
-+ }
-+
-+ oh_set_buffer_layout(oh_config->oh_port, &buf_layout);
-+
-+ /* read the pool handlers */
-+ crt_ext_pools_count = of_count_phandle_with_args(dpa_oh_node,
-+ "fsl,bman-buffer-pools", NULL);
-+ if (crt_ext_pools_count <= 0) {
-+ dev_info(dpa_oh_dev,
-+ "OH port %s has no buffer pool. Fragmentation will not be enabled\n",
-+ oh_node->full_name);
-+ goto init_port;
-+ }
-+
-+ /* used for reading ext_pool_size*/
-+ root_node = of_find_node_by_path("/");
-+ if (root_node == NULL) {
-+ dev_err(dpa_oh_dev, "of_find_node_by_path(/) failed\n");
-+ _errno = -EINVAL;
-+ goto return_kfree;
-+ }
-+
-+ n_size = of_n_size_cells(root_node);
-+ of_node_put(root_node);
-+
-+ dev_dbg(dpa_oh_dev, "OH port number of pools = %d\n",
-+ crt_ext_pools_count);
-+
-+ oh_port_tx_params.num_pools = (uint8_t)crt_ext_pools_count;
-+
-+ for (i = 0; i < crt_ext_pools_count; i++) {
-+ bpool_node = of_parse_phandle(dpa_oh_node,
-+ "fsl,bman-buffer-pools", i);
-+ if (bpool_node == NULL) {
-+ dev_err(dpa_oh_dev, "Invalid Buffer pool node\n");
-+ _errno = -EINVAL;
-+ goto return_kfree;
-+ }
-+
-+ _errno = of_property_read_u32(bpool_node, "fsl,bpid", &bpid);
-+ if (_errno) {
-+ dev_err(dpa_oh_dev, "Invalid Buffer Pool ID\n");
-+ _errno = -EINVAL;
-+ goto return_kfree;
-+ }
-+
-+ oh_port_tx_params.pool_param[i].id = (uint8_t)bpid;
-+ dev_dbg(dpa_oh_dev, "OH port bpool id = %u\n", bpid);
-+
-+ bpool_cfg = of_get_property(bpool_node,
-+ "fsl,bpool-ethernet-cfg", &lenp);
-+ if (bpool_cfg == NULL) {
-+ dev_err(dpa_oh_dev, "Invalid Buffer pool config params\n");
-+ _errno = -EINVAL;
-+ goto return_kfree;
-+ }
-+
-+ ext_pool_size = of_read_number(bpool_cfg + n_size, n_size);
-+ oh_port_tx_params.pool_param[i].size = (uint16_t)ext_pool_size;
-+ dev_dbg(dpa_oh_dev, "OH port bpool size = %u\n",
-+ ext_pool_size);
-+ of_node_put(bpool_node);
-+
-+ }
-+
-+ if (buf_layout.data_align != FRAG_DATA_ALIGN ||
-+ buf_layout.manip_extra_space != FRAG_MANIP_SPACE)
-+ goto init_port;
-+
-+ frag_enabled = true;
-+ dev_info(dpa_oh_dev, "IP Fragmentation enabled for OH port %d",
-+ port_id);
-+
-+init_port:
-+ of_node_put(oh_node);
-+ /* Set Tx params */
-+ dpaa_eth_init_port(tx, oh_config->oh_port, oh_port_tx_params,
-+ oh_config->error_fqid, oh_config->default_fqid, (&buf_layout),
-+ frag_enabled);
-+ /* Set PCD params */
-+ oh_port_pcd_params.cba = oh_alloc_pcd_fqids;
-+ oh_port_pcd_params.cbf = oh_free_pcd_fqids;
-+ oh_port_pcd_params.dev = dpa_oh_dev;
-+ fm_port_pcd_bind(oh_config->oh_port, &oh_port_pcd_params);
-+
-+ dev_set_drvdata(dpa_oh_dev, oh_config);
-+
-+ /* Enable the OH port */
-+ _errno = fm_port_enable(oh_config->oh_port);
-+ if (_errno)
-+ goto return_kfree;
-+
-+ dev_info(dpa_oh_dev, "OH port %s enabled.\n", oh_node->full_name);
-+
-+ /* print of all referenced & created queues */
-+ dump_oh_config(dpa_oh_dev, oh_config);
-+
-+ return 0;
-+
-+return_kfree:
-+ if (bpool_node)
-+ of_node_put(bpool_node);
-+ if (oh_node)
-+ of_node_put(oh_node);
-+ if (oh_config && oh_config->egress_fqs)
-+ devm_kfree(dpa_oh_dev, oh_config->egress_fqs);
-+
-+ list_for_each_safe(fq_list, fq_list_tmp, &oh_config->fqs_ingress_list) {
-+ fqd = list_entry(fq_list, struct fq_duple, fq_list);
-+ list_del(fq_list);
-+ devm_kfree(dpa_oh_dev, fqd->fqs);
-+ devm_kfree(dpa_oh_dev, fqd);
-+ }
-+
-+ list_for_each_safe(fq_list, fq_list_tmp, &oh_config->fqs_egress_list) {
-+ fqd = list_entry(fq_list, struct fq_duple, fq_list);
-+ list_del(fq_list);
-+ devm_kfree(dpa_oh_dev, fqd->fqs);
-+ devm_kfree(dpa_oh_dev, fqd);
-+ }
-+
-+ devm_kfree(dpa_oh_dev, oh_config);
-+ return _errno;
-+}
-+
-+static int __cold oh_port_remove(struct platform_device *_of_dev)
-+{
-+ int _errno = 0, i;
-+ struct dpa_oh_config_s *oh_config;
-+
-+ pr_info("Removing OH port...\n");
-+
-+ oh_config = dev_get_drvdata(&_of_dev->dev);
-+ if (oh_config == NULL) {
-+ pr_err(KBUILD_MODNAME
-+ ": %s:%hu:%s(): No OH config in device private data!\n",
-+ KBUILD_BASENAME".c", __LINE__, __func__);
-+ _errno = -ENODEV;
-+ goto return_error;
-+ }
-+
-+ if (oh_config->egress_fqs)
-+ for (i = 0; i < oh_config->egress_cnt; i++)
-+ oh_fq_destroy(oh_config->egress_fqs + i);
-+
-+ if (oh_config->oh_port == NULL) {
-+ pr_err(KBUILD_MODNAME
-+ ": %s:%hu:%s(): No fm port in device private data!\n",
-+ KBUILD_BASENAME".c", __LINE__, __func__);
-+ _errno = -EINVAL;
-+ goto free_egress_fqs;
-+ }
-+
-+ _errno = fm_port_disable(oh_config->oh_port);
-+
-+free_egress_fqs:
-+ if (oh_config->egress_fqs)
-+ devm_kfree(&_of_dev->dev, oh_config->egress_fqs);
-+ devm_kfree(&_of_dev->dev, oh_config);
-+ dev_set_drvdata(&_of_dev->dev, NULL);
-+
-+return_error:
-+ return _errno;
-+}
-+
-+static struct platform_driver oh_port_driver = {
-+ .driver = {
-+ .name = KBUILD_MODNAME,
-+ .of_match_table = oh_port_match_table,
-+ .owner = THIS_MODULE,
-+ .pm = OH_PM_OPS,
-+ },
-+ .probe = oh_port_probe,
-+ .remove = oh_port_remove
-+};
-+
-+static int __init __cold oh_port_load(void)
-+{
-+ int _errno;
-+
-+ pr_info(OH_MOD_DESCRIPTION "\n");
-+
-+ _errno = platform_driver_register(&oh_port_driver);
-+ if (_errno < 0) {
-+ pr_err(KBUILD_MODNAME
-+ ": %s:%hu:%s(): platform_driver_register() = %d\n",
-+ KBUILD_BASENAME".c", __LINE__, __func__, _errno);
-+ }
-+
-+ pr_debug(KBUILD_MODNAME ": %s:%s() ->\n",
-+ KBUILD_BASENAME".c", __func__);
-+ return _errno;
-+}
-+module_init(oh_port_load);
-+
-+static void __exit __cold oh_port_unload(void)
-+{
-+ pr_debug(KBUILD_MODNAME ": -> %s:%s()\n",
-+ KBUILD_BASENAME".c", __func__);
-+
-+ platform_driver_unregister(&oh_port_driver);
-+
-+ pr_debug(KBUILD_MODNAME ": %s:%s() ->\n",
-+ KBUILD_BASENAME".c", __func__);
-+}
-+module_exit(oh_port_unload);
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_dpaa/offline_port.h
-@@ -0,0 +1,59 @@
-+/* Copyright 2011 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#ifndef __OFFLINE_PORT_H
-+#define __OFFLINE_PORT_H
-+
-+struct fm_port;
-+struct qman_fq;
-+
-+/* fqs are defined in duples (base_fq, fq_count) */
-+struct fq_duple {
-+ struct qman_fq *fqs;
-+ int fqs_count;
-+ uint16_t channel_id;
-+ struct list_head fq_list;
-+};
-+
-+/* OH port configuration */
-+struct dpa_oh_config_s {
-+ uint32_t error_fqid;
-+ uint32_t default_fqid;
-+ struct fm_port *oh_port;
-+ uint32_t egress_cnt;
-+ struct qman_fq *egress_fqs;
-+ uint16_t channel;
-+
-+ struct list_head fqs_ingress_list;
-+ struct list_head fqs_egress_list;
-+};
-+
-+#endif /* __OFFLINE_PORT_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Kconfig
-@@ -0,0 +1,153 @@
-+menu "Frame Manager support"
-+
-+menuconfig FSL_SDK_FMAN
-+ bool "Freescale Frame Manager (datapath) support - SDK driver"
-+ depends on (FSL_SOC || ARM64 || ARM) && FSL_SDK_BMAN && FSL_SDK_QMAN && !FSL_FMAN
-+ default y
-+ ---help---
-+ If unsure, say Y.
-+
-+if FSL_SDK_FMAN
-+
-+config FSL_SDK_FMAN_TEST
-+ bool "FMan test module"
-+ default n
-+ select FSL_DPAA_HOOKS
-+ ---help---
-+ This option compiles test code for FMan.
-+
-+menu "FMAN Processor support"
-+choice
-+ depends on FSL_SDK_FMAN
-+ prompt "Processor Type"
-+
-+config FMAN_ARM
-+ bool "LS1043"
-+ depends on ARM64 || ARM
-+ ---help---
-+ Choose "LS1043" for the ARM platforms:
-+ LS1043
-+
-+config FMAN_P3040_P4080_P5020
-+ bool "P3040 P4080 5020"
-+
-+config FMAN_P1023
-+ bool "P1023"
-+
-+config FMAN_V3H
-+ bool "FmanV3H"
-+ ---help---
-+ Choose "FmanV3H" for Fman rev3H:
-+ B4860, T4240, T4160, etc
-+
-+config FMAN_V3L
-+ bool "FmanV3L"
-+ ---help---
-+ Choose "FmanV3L" for Fman rev3L:
-+ T1040, T1042, T1020, T1022, T1023, T1024, etc
-+
-+endchoice
-+endmenu
-+
-+config FMAN_MIB_CNT_OVF_IRQ_EN
-+ bool "Enable the dTSEC MIB counters overflow interrupt"
-+ default n
-+ ---help---
-+ Enable the dTSEC MIB counters overflow interrupt to get
-+ accurate MIB counters values. Enabled it compensates
-+ for the counters overflow but reduces performance and
-+ triggers error messages in HV setups.
-+
-+config FSL_FM_MAX_FRAME_SIZE
-+ int "Maximum L2 frame size"
-+ depends on FSL_SDK_FMAN
-+ range 64 9600
-+ default "1522"
-+ help
-+ Configure this in relation to the maximum possible MTU of your
-+ network configuration. In particular, one would need to
-+ increase this value in order to use jumbo frames.
-+ FSL_FM_MAX_FRAME_SIZE must accommodate the Ethernet FCS (4 bytes)
-+ and one ETH+VLAN header (18 bytes), to a total of 22 bytes in
-+ excess of the desired L3 MTU.
-+
-+ Note that having too large a FSL_FM_MAX_FRAME_SIZE (much larger
-+ than the actual MTU) may lead to buffer exhaustion, especially
-+ in the case of badly fragmented datagrams on the Rx path.
-+ Conversely, having a FSL_FM_MAX_FRAME_SIZE smaller than the actual
-+ MTU will lead to frames being dropped.
-+
-+ This can be overridden by specifying "fsl_fm_max_frm" in
-+ the kernel bootargs:
-+ * in Hypervisor-based scenarios, by adding a "chosen" node
-+ with the "bootargs" property specifying
-+ "fsl_fm_max_frm=<YourValue>";
-+ * in non-Hypervisor-based scenarios, via u-boot's env, by
-+ modifying the "bootargs" env variable.
-+
-+config FSL_FM_RX_EXTRA_HEADROOM
-+ int "Add extra headroom at beginning of data buffers"
-+ depends on FSL_SDK_FMAN
-+ range 16 384
-+ default "64"
-+ help
-+ Configure this to tell the Frame Manager to reserve some extra
-+ space at the beginning of a data buffer on the receive path,
-+ before Internal Context fields are copied. This is in addition
-+ to the private data area already reserved for driver internal
-+ use. The provided value must be a multiple of 16.
-+
-+ This setting can be overridden by specifying
-+ "fsl_fm_rx_extra_headroom" in the kernel bootargs:
-+ * in Hypervisor-based scenarios, by adding a "chosen" node
-+ with the "bootargs" property specifying
-+ "fsl_fm_rx_extra_headroom=<YourValue>";
-+ * in non-Hypervisor-based scenarios, via u-boot's env, by
-+ modifying the "bootargs" env variable.
-+
-+config FMAN_PFC
-+ bool "FMan PFC support (EXPERIMENTAL)"
-+ depends on ( FMAN_V3H || FMAN_V3L || FMAN_ARM) && FSL_SDK_FMAN
-+ default n
-+ help
-+ This option enables PFC support on FMan v3 ports.
-+ Data Center Bridging defines Classes of Service that are
-+ flow-controlled using PFC pause frames.
-+
-+if FMAN_PFC
-+config FMAN_PFC_COS_COUNT
-+ int "Number of PFC Classes of Service"
-+ depends on FMAN_PFC && FSL_SDK_FMAN
-+ range 1 4
-+ default "3"
-+ help
-+ The number of Classes of Service controlled by PFC.
-+
-+config FMAN_PFC_QUANTA_0
-+ int "The pause quanta for PFC CoS 0"
-+ depends on FMAN_PFC && FSL_SDK_FMAN
-+ range 0 65535
-+ default "65535"
-+
-+config FMAN_PFC_QUANTA_1
-+ int "The pause quanta for PFC CoS 1"
-+ depends on FMAN_PFC && FSL_SDK_FMAN
-+ range 0 65535
-+ default "65535"
-+
-+config FMAN_PFC_QUANTA_2
-+ int "The pause quanta for PFC CoS 2"
-+ depends on FMAN_PFC && FSL_SDK_FMAN
-+ range 0 65535
-+ default "65535"
-+
-+config FMAN_PFC_QUANTA_3
-+ int "The pause quanta for PFC CoS 3"
-+ depends on FMAN_PFC && FSL_SDK_FMAN
-+ range 0 65535
-+ default "65535"
-+endif
-+
-+endif # FSL_SDK_FMAN
-+
-+endmenu
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Makefile
-@@ -0,0 +1,11 @@
-+#
-+# Makefile for the Freescale Ethernet controllers
-+#
-+ccflags-y += -DVERSION=\"\"
-+#
-+#Include netcomm SW specific definitions
-+include $(srctree)/drivers/net/ethernet/freescale/sdk_fman/ncsw_config.mk
-+#
-+obj-y += etc/
-+obj-y += Peripherals/FM/
-+obj-y += src/
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/HC/Makefile
-@@ -0,0 +1,15 @@
-+#
-+# Makefile for the Freescale Ethernet controllers
-+#
-+ccflags-y += -DVERSION=\"\"
-+#
-+#Include netcomm SW specific definitions
-+include $(srctree)/drivers/net/ethernet/freescale/sdk_fman/ncsw_config.mk
-+
-+NCSW_FM_INC = $(srctree)/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/inc
-+
-+ccflags-y += -I$(NCSW_FM_INC)
-+
-+obj-y += fsl-ncsw-Hc.o
-+
-+fsl-ncsw-Hc-objs := hc.o
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/HC/hc.c
-@@ -0,0 +1,1232 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+#include "std_ext.h"
-+#include "error_ext.h"
-+#include "sprint_ext.h"
-+#include "string_ext.h"
-+
-+#include "fm_common.h"
-+#include "fm_hc.h"
-+
-+
-+/**************************************************************************//**
-+ @Description defaults
-+*//***************************************************************************/
-+#define DEFAULT_dataMemId 0
-+
-+#define HC_HCOR_OPCODE_PLCR_PRFL 0x0
-+#define HC_HCOR_OPCODE_KG_SCM 0x1
-+#define HC_HCOR_OPCODE_SYNC 0x2
-+#define HC_HCOR_OPCODE_CC 0x3
-+#define HC_HCOR_OPCODE_CC_AGE_MASK 0x4
-+#define HC_HCOR_OPCODE_CC_CAPWAP_REASSM_TIMEOUT 0x5
-+#define HC_HCOR_OPCODE_CC_REASSM_TIMEOUT 0x10
-+#define HC_HCOR_OPCODE_CC_IP_FRAG_INITIALIZATION 0x11
-+#define HC_HCOR_OPCODE_CC_UPDATE_WITH_AGING 0x13
-+#define HC_HCOR_ACTION_REG_REASSM_TIMEOUT_ACTIVE_SHIFT 24
-+#define HC_HCOR_EXTRA_REG_REASSM_TIMEOUT_TSBS_SHIFT 24
-+#define HC_HCOR_EXTRA_REG_CC_AGING_ADD 0x80000000
-+#define HC_HCOR_EXTRA_REG_CC_AGING_REMOVE 0x40000000
-+#define HC_HCOR_EXTRA_REG_CC_AGING_CHANGE_MASK 0xC0000000
-+#define HC_HCOR_EXTRA_REG_CC_REMOVE_INDX_SHIFT 24
-+#define HC_HCOR_EXTRA_REG_CC_REMOVE_INDX_MASK 0x1F000000
-+#define HC_HCOR_ACTION_REG_REASSM_TIMEOUT_RES_SHIFT 16
-+#define HC_HCOR_ACTION_REG_REASSM_TIMEOUT_RES_MASK 0xF
-+#define HC_HCOR_ACTION_REG_IP_FRAG_SCRATCH_POOL_CMD_SHIFT 24
-+#define HC_HCOR_ACTION_REG_IP_FRAG_SCRATCH_POOL_BPID 16
-+
-+#define HC_HCOR_GBL 0x20000000
-+
-+#define HC_HCOR_KG_SCHEME_COUNTER 0x00000400
-+
-+#if (DPAA_VERSION == 10)
-+#define HC_HCOR_KG_SCHEME_REGS_MASK 0xFFFFF800
-+#else
-+#define HC_HCOR_KG_SCHEME_REGS_MASK 0xFFFFFE00
-+#endif /* (DPAA_VERSION == 10) */
-+
-+#define SIZE_OF_HC_FRAME_PORT_REGS (sizeof(t_HcFrame)-sizeof(struct fman_kg_scheme_regs)+sizeof(t_FmPcdKgPortRegs))
-+#define SIZE_OF_HC_FRAME_SCHEME_REGS sizeof(t_HcFrame)
-+#define SIZE_OF_HC_FRAME_PROFILES_REGS (sizeof(t_HcFrame)-sizeof(struct fman_kg_scheme_regs)+sizeof(t_FmPcdPlcrProfileRegs))
-+#define SIZE_OF_HC_FRAME_PROFILE_CNT (sizeof(t_HcFrame)-sizeof(t_FmPcdPlcrProfileRegs)+sizeof(uint32_t))
-+#define SIZE_OF_HC_FRAME_READ_OR_CC_DYNAMIC 16
-+
-+#define HC_CMD_POOL_SIZE (INTG_MAX_NUM_OF_CORES)
-+
-+#define BUILD_FD(len) \
-+do { \
-+ memset(&fmFd, 0, sizeof(t_DpaaFD)); \
-+ DPAA_FD_SET_ADDR(&fmFd, p_HcFrame); \
-+ DPAA_FD_SET_OFFSET(&fmFd, 0); \
-+ DPAA_FD_SET_LENGTH(&fmFd, len); \
-+} while (0)
-+
-+
-+#if defined(__MWERKS__) && !defined(__GNUC__)
-+#pragma pack(push,1)
-+#endif /* defined(__MWERKS__) && ... */
-+
-+typedef struct t_FmPcdKgPortRegs {
-+ volatile uint32_t spReg;
-+ volatile uint32_t cppReg;
-+} t_FmPcdKgPortRegs;
-+
-+typedef struct t_HcFrame {
-+ volatile uint32_t opcode;
-+ volatile uint32_t actionReg;
-+ volatile uint32_t extraReg;
-+ volatile uint32_t commandSequence;
-+ union {
-+ struct fman_kg_scheme_regs schemeRegs;
-+ struct fman_kg_scheme_regs schemeRegsWithoutCounter;
-+ t_FmPcdPlcrProfileRegs profileRegs;
-+ volatile uint32_t singleRegForWrite; /* for writing SP, CPP, profile counter */
-+ t_FmPcdKgPortRegs portRegsForRead;
-+ volatile uint32_t clsPlanEntries[CLS_PLAN_NUM_PER_GRP];
-+ t_FmPcdCcCapwapReassmTimeoutParams ccCapwapReassmTimeout;
-+ t_FmPcdCcReassmTimeoutParams ccReassmTimeout;
-+ } hcSpecificData;
-+} t_HcFrame;
-+
-+#if defined(__MWERKS__) && !defined(__GNUC__)
-+#pragma pack(pop)
-+#endif /* defined(__MWERKS__) && ... */
-+
-+
-+typedef struct t_FmHc {
-+ t_Handle h_FmPcd;
-+ t_Handle h_HcPortDev;
-+ t_FmPcdQmEnqueueCallback *f_QmEnqueue; /**< A callback for enqueuing frames to the QM */
-+ t_Handle h_QmArg; /**< A handle to the QM module */
-+ uint8_t dataMemId; /**< Memory partition ID for data buffers */
-+
-+ uint32_t seqNum[HC_CMD_POOL_SIZE]; /* FIFO of seqNum to use when
-+ taking buffer */
-+ uint32_t nextSeqNumLocation; /* seqNum location in seqNum[] for next buffer */
-+ volatile bool enqueued[HC_CMD_POOL_SIZE]; /* HC is active - frame is enqueued
-+ and not confirmed yet */
-+ t_HcFrame *p_Frm[HC_CMD_POOL_SIZE];
-+} t_FmHc;
-+
-+
-+static t_Error FillBufPool(t_FmHc *p_FmHc)
-+{
-+ uint32_t i;
-+
-+ ASSERT_COND(p_FmHc);
-+
-+ for (i = 0; i < HC_CMD_POOL_SIZE; i++)
-+ {
-+#ifdef FM_LOCKUP_ALIGNMENT_ERRATA_FMAN_SW004
-+ p_FmHc->p_Frm[i] = (t_HcFrame *)XX_MallocSmart((sizeof(t_HcFrame) + (16 - (sizeof(t_FmHc) % 16))),
-+ p_FmHc->dataMemId,
-+ 16);
-+#else
-+ p_FmHc->p_Frm[i] = (t_HcFrame *)XX_MallocSmart(sizeof(t_HcFrame),
-+ p_FmHc->dataMemId,
-+ 16);
-+#endif /* FM_LOCKUP_ALIGNMENT_ERRATA_FMAN_SW004 */
-+ if (!p_FmHc->p_Frm[i])
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, ("FM HC frames!"));
-+ }
-+
-+ /* Initialize FIFO of seqNum to use during GetBuf */
-+ for (i = 0; i < HC_CMD_POOL_SIZE; i++)
-+ {
-+ p_FmHc->seqNum[i] = i;
-+ }
-+ p_FmHc->nextSeqNumLocation = 0;
-+
-+ return E_OK;
-+}
-+
-+static __inline__ t_HcFrame * GetBuf(t_FmHc *p_FmHc, uint32_t *p_SeqNum)
-+{
-+ uint32_t intFlags;
-+
-+ ASSERT_COND(p_FmHc);
-+
-+ intFlags = FmPcdLock(p_FmHc->h_FmPcd);
-+
-+ if (p_FmHc->nextSeqNumLocation == HC_CMD_POOL_SIZE)
-+ {
-+ /* No more buffers */
-+ FmPcdUnlock(p_FmHc->h_FmPcd, intFlags);
-+ return NULL;
-+ }
-+
-+ *p_SeqNum = p_FmHc->seqNum[p_FmHc->nextSeqNumLocation];
-+ p_FmHc->nextSeqNumLocation++;
-+
-+ FmPcdUnlock(p_FmHc->h_FmPcd, intFlags);
-+ return p_FmHc->p_Frm[*p_SeqNum];
-+}
-+
-+static __inline__ void PutBuf(t_FmHc *p_FmHc, t_HcFrame *p_Buf, uint32_t seqNum)
-+{
-+ uint32_t intFlags;
-+
-+ UNUSED(p_Buf);
-+
-+ intFlags = FmPcdLock(p_FmHc->h_FmPcd);
-+ ASSERT_COND(p_FmHc->nextSeqNumLocation);
-+ p_FmHc->nextSeqNumLocation--;
-+ p_FmHc->seqNum[p_FmHc->nextSeqNumLocation] = seqNum;
-+ FmPcdUnlock(p_FmHc->h_FmPcd, intFlags);
-+}
-+
-+static __inline__ t_Error EnQFrm(t_FmHc *p_FmHc, t_DpaaFD *p_FmFd, uint32_t seqNum)
-+{
-+ t_Error err = E_OK;
-+ uint32_t intFlags;
-+ uint32_t timeout=100;
-+
-+ intFlags = FmPcdLock(p_FmHc->h_FmPcd);
-+ ASSERT_COND(!p_FmHc->enqueued[seqNum]);
-+ p_FmHc->enqueued[seqNum] = TRUE;
-+ FmPcdUnlock(p_FmHc->h_FmPcd, intFlags);
-+ DBG(TRACE, ("Send Hc, SeqNum %d, buff@0x%x, fd offset 0x%x",
-+ seqNum,
-+ DPAA_FD_GET_ADDR(p_FmFd),
-+ DPAA_FD_GET_OFFSET(p_FmFd)));
-+ err = p_FmHc->f_QmEnqueue(p_FmHc->h_QmArg, (void *)p_FmFd);
-+ if (err)
-+ RETURN_ERROR(MINOR, err, ("HC enqueue failed"));
-+
-+ while (p_FmHc->enqueued[seqNum] && --timeout)
-+ XX_UDelay(100);
-+
-+ if (!timeout)
-+ RETURN_ERROR(MINOR, E_TIMEOUT, ("HC Callback, timeout exceeded"));
-+
-+ return err;
-+}
-+
-+
-+t_Handle FmHcConfigAndInit(t_FmHcParams *p_FmHcParams)
-+{
-+ t_FmHc *p_FmHc;
-+ t_FmPortParams fmPortParam;
-+ t_Error err;
-+
-+ p_FmHc = (t_FmHc *)XX_Malloc(sizeof(t_FmHc));
-+ if (!p_FmHc)
-+ {
-+ REPORT_ERROR(MINOR, E_NO_MEMORY, ("HC obj"));
-+ return NULL;
-+ }
-+ memset(p_FmHc,0,sizeof(t_FmHc));
-+
-+ p_FmHc->h_FmPcd = p_FmHcParams->h_FmPcd;
-+ p_FmHc->f_QmEnqueue = p_FmHcParams->params.f_QmEnqueue;
-+ p_FmHc->h_QmArg = p_FmHcParams->params.h_QmArg;
-+ p_FmHc->dataMemId = DEFAULT_dataMemId;
-+
-+ err = FillBufPool(p_FmHc);
-+ if (err != E_OK)
-+ {
-+ REPORT_ERROR(MAJOR, err, NO_MSG);
-+ FmHcFree(p_FmHc);
-+ return NULL;
-+ }
-+
-+ if (!FmIsMaster(p_FmHcParams->h_Fm))
-+ return (t_Handle)p_FmHc;
-+
-+ memset(&fmPortParam, 0, sizeof(fmPortParam));
-+ fmPortParam.baseAddr = p_FmHcParams->params.portBaseAddr;
-+ fmPortParam.portType = e_FM_PORT_TYPE_OH_HOST_COMMAND;
-+ fmPortParam.portId = p_FmHcParams->params.portId;
-+ fmPortParam.liodnBase = p_FmHcParams->params.liodnBase;
-+ fmPortParam.h_Fm = p_FmHcParams->h_Fm;
-+
-+ fmPortParam.specificParams.nonRxParams.errFqid = p_FmHcParams->params.errFqid;
-+ fmPortParam.specificParams.nonRxParams.dfltFqid = p_FmHcParams->params.confFqid;
-+ fmPortParam.specificParams.nonRxParams.qmChannel = p_FmHcParams->params.qmChannel;
-+
-+ p_FmHc->h_HcPortDev = FM_PORT_Config(&fmPortParam);
-+ if (!p_FmHc->h_HcPortDev)
-+ {
-+ REPORT_ERROR(MAJOR, E_INVALID_HANDLE, ("FM HC port!"));
-+ XX_Free(p_FmHc);
-+ return NULL;
-+ }
-+
-+ err = FM_PORT_ConfigMaxFrameLength(p_FmHc->h_HcPortDev,
-+ (uint16_t)sizeof(t_HcFrame));
-+
-+ if (err != E_OK)
-+ {
-+ REPORT_ERROR(MAJOR, err, ("FM HC port init!"));
-+ FmHcFree(p_FmHc);
-+ return NULL;
-+ }
-+
-+ /* final init */
-+ err = FM_PORT_Init(p_FmHc->h_HcPortDev);
-+ if (err != E_OK)
-+ {
-+ REPORT_ERROR(MAJOR, err, ("FM HC port init!"));
-+ FmHcFree(p_FmHc);
-+ return NULL;
-+ }
-+
-+ err = FM_PORT_Enable(p_FmHc->h_HcPortDev);
-+ if (err != E_OK)
-+ {
-+ REPORT_ERROR(MAJOR, err, ("FM HC port enable!"));
-+ FmHcFree(p_FmHc);
-+ return NULL;
-+ }
-+
-+ return (t_Handle)p_FmHc;
-+}
-+
-+void FmHcFree(t_Handle h_FmHc)
-+{
-+ t_FmHc *p_FmHc = (t_FmHc*)h_FmHc;
-+ int i;
-+
-+ if (!p_FmHc)
-+ return;
-+
-+ for (i=0; i<HC_CMD_POOL_SIZE; i++)
-+ if (p_FmHc->p_Frm[i])
-+ XX_FreeSmart(p_FmHc->p_Frm[i]);
-+ else
-+ break;
-+
-+ if (p_FmHc->h_HcPortDev)
-+ FM_PORT_Free(p_FmHc->h_HcPortDev);
-+
-+ XX_Free(p_FmHc);
-+}
-+
-+/*****************************************************************************/
-+t_Error FmHcSetFramesDataMemory(t_Handle h_FmHc,
-+ uint8_t memId)
-+{
-+ t_FmHc *p_FmHc = (t_FmHc*)h_FmHc;
-+ int i;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmHc, E_INVALID_HANDLE);
-+
-+ p_FmHc->dataMemId = memId;
-+
-+ for (i=0; i<HC_CMD_POOL_SIZE; i++)
-+ if (p_FmHc->p_Frm[i])
-+ XX_FreeSmart(p_FmHc->p_Frm[i]);
-+
-+ return FillBufPool(p_FmHc);
-+}
-+
-+void FmHcTxConf(t_Handle h_FmHc, t_DpaaFD *p_Fd)
-+{
-+ t_FmHc *p_FmHc = (t_FmHc*)h_FmHc;
-+ t_HcFrame *p_HcFrame;
-+ uint32_t intFlags;
-+
-+ ASSERT_COND(p_FmHc);
-+
-+ intFlags = FmPcdLock(p_FmHc->h_FmPcd);
-+ p_HcFrame = (t_HcFrame *)PTR_MOVE(DPAA_FD_GET_ADDR(p_Fd), DPAA_FD_GET_OFFSET(p_Fd));
-+
-+ DBG(TRACE, ("Hc Conf, SeqNum %d, FD@0x%x, fd offset 0x%x",
-+ p_HcFrame->commandSequence, DPAA_FD_GET_ADDR(p_Fd), DPAA_FD_GET_OFFSET(p_Fd)));
-+
-+ if (!(p_FmHc->enqueued[p_HcFrame->commandSequence]))
-+ REPORT_ERROR(MINOR, E_INVALID_FRAME, ("Not an Host-Command frame received!"));
-+ else
-+ p_FmHc->enqueued[p_HcFrame->commandSequence] = FALSE;
-+ FmPcdUnlock(p_FmHc->h_FmPcd, intFlags);
-+}
-+
-+t_Error FmHcPcdKgSetScheme(t_Handle h_FmHc,
-+ t_Handle h_Scheme,
-+ struct fman_kg_scheme_regs *p_SchemeRegs,
-+ bool updateCounter)
-+{
-+ t_FmHc *p_FmHc = (t_FmHc*)h_FmHc;
-+ t_Error err = E_OK;
-+ t_HcFrame *p_HcFrame;
-+ t_DpaaFD fmFd;
-+ uint8_t physicalSchemeId;
-+ uint32_t seqNum;
-+
-+ p_HcFrame = GetBuf(p_FmHc, &seqNum);
-+ if (!p_HcFrame)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("HC Frame object"));
-+
-+ physicalSchemeId = FmPcdKgGetSchemeId(h_Scheme);
-+
-+ memset(p_HcFrame, 0, sizeof(t_HcFrame));
-+ p_HcFrame->opcode = (uint32_t)(HC_HCOR_GBL | HC_HCOR_OPCODE_KG_SCM);
-+ p_HcFrame->actionReg = FmPcdKgBuildWriteSchemeActionReg(physicalSchemeId, updateCounter);
-+ p_HcFrame->extraReg = HC_HCOR_KG_SCHEME_REGS_MASK;
-+ memcpy(&p_HcFrame->hcSpecificData.schemeRegs, p_SchemeRegs, sizeof(struct fman_kg_scheme_regs));
-+ if (!updateCounter)
-+ {
-+ p_HcFrame->hcSpecificData.schemeRegs.kgse_dv0 = p_SchemeRegs->kgse_dv0;
-+ p_HcFrame->hcSpecificData.schemeRegs.kgse_dv1 = p_SchemeRegs->kgse_dv1;
-+ p_HcFrame->hcSpecificData.schemeRegs.kgse_ccbs = p_SchemeRegs->kgse_ccbs;
-+ p_HcFrame->hcSpecificData.schemeRegs.kgse_mv = p_SchemeRegs->kgse_mv;
-+ }
-+ p_HcFrame->commandSequence = seqNum;
-+
-+ BUILD_FD(sizeof(t_HcFrame));
-+
-+ err = EnQFrm(p_FmHc, &fmFd, seqNum);
-+
-+ PutBuf(p_FmHc, p_HcFrame, seqNum);
-+
-+ if (err != E_OK)
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+
-+ return E_OK;
-+}
-+
-+t_Error FmHcPcdKgDeleteScheme(t_Handle h_FmHc, t_Handle h_Scheme)
-+{
-+ t_FmHc *p_FmHc = (t_FmHc*)h_FmHc;
-+ t_Error err = E_OK;
-+ t_HcFrame *p_HcFrame;
-+ t_DpaaFD fmFd;
-+ uint8_t physicalSchemeId = FmPcdKgGetSchemeId(h_Scheme);
-+ uint32_t seqNum;
-+
-+ p_HcFrame = GetBuf(p_FmHc, &seqNum);
-+ if (!p_HcFrame)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("HC Frame object"));
-+
-+ memset(p_HcFrame, 0, sizeof(t_HcFrame));
-+ p_HcFrame->opcode = (uint32_t)(HC_HCOR_GBL | HC_HCOR_OPCODE_KG_SCM);
-+ p_HcFrame->actionReg = FmPcdKgBuildWriteSchemeActionReg(physicalSchemeId, TRUE);
-+ p_HcFrame->extraReg = HC_HCOR_KG_SCHEME_REGS_MASK;
-+ memset(&p_HcFrame->hcSpecificData.schemeRegs, 0, sizeof(struct fman_kg_scheme_regs));
-+ p_HcFrame->commandSequence = seqNum;
-+
-+ BUILD_FD(sizeof(t_HcFrame));
-+
-+ err = EnQFrm(p_FmHc, &fmFd, seqNum);
-+
-+ PutBuf(p_FmHc, p_HcFrame, seqNum);
-+
-+ if (err != E_OK)
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+
-+ return E_OK;
-+}
-+
-+t_Error FmHcPcdKgCcGetSetParams(t_Handle h_FmHc, t_Handle h_Scheme, uint32_t requiredAction, uint32_t value)
-+{
-+ t_FmHc *p_FmHc = (t_FmHc*)h_FmHc;
-+ t_Error err = E_OK;
-+ t_HcFrame *p_HcFrame;
-+ t_DpaaFD fmFd;
-+ uint8_t relativeSchemeId;
-+ uint8_t physicalSchemeId = FmPcdKgGetSchemeId(h_Scheme);
-+ uint32_t tmpReg32 = 0;
-+ uint32_t seqNum;
-+
-+ /* Scheme is locked by calling routine */
-+ /* WARNING - this lock will not be efficient if other HC routine will attempt to change
-+ * "kgse_mode" or "kgse_om" without locking scheme !
-+ */
-+
-+ relativeSchemeId = FmPcdKgGetRelativeSchemeId(p_FmHc->h_FmPcd, physicalSchemeId);
-+ if ( relativeSchemeId == FM_PCD_KG_NUM_OF_SCHEMES)
-+ RETURN_ERROR(MAJOR, E_NOT_IN_RANGE, NO_MSG);
-+
-+ if (!FmPcdKgGetRequiredActionFlag(p_FmHc->h_FmPcd, relativeSchemeId) ||
-+ !(FmPcdKgGetRequiredAction(p_FmHc->h_FmPcd, relativeSchemeId) & requiredAction))
-+ {
-+ if ((requiredAction & UPDATE_NIA_ENQ_WITHOUT_DMA) &&
-+ (FmPcdKgGetNextEngine(p_FmHc->h_FmPcd, relativeSchemeId) == e_FM_PCD_PLCR))
-+ {
-+ if ((FmPcdKgIsDirectPlcr(p_FmHc->h_FmPcd, relativeSchemeId) == FALSE) ||
-+ (FmPcdKgIsDistrOnPlcrProfile(p_FmHc->h_FmPcd, relativeSchemeId) == TRUE))
-+ RETURN_ERROR(MAJOR, E_NOT_SUPPORTED, ("In this situation PP can not be with distribution and has to be shared"));
-+ err = FmPcdPlcrCcGetSetParams(p_FmHc->h_FmPcd, FmPcdKgGetRelativeProfileId(p_FmHc->h_FmPcd, relativeSchemeId), requiredAction);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+ else /* From here we deal with KG-Schemes only */
-+ {
-+ /* Pre change general code */
-+ p_HcFrame = GetBuf(p_FmHc, &seqNum);
-+ if (!p_HcFrame)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("HC Frame object"));
-+ memset(p_HcFrame, 0, sizeof(t_HcFrame));
-+ p_HcFrame->opcode = (uint32_t)(HC_HCOR_GBL | HC_HCOR_OPCODE_KG_SCM);
-+ p_HcFrame->actionReg = FmPcdKgBuildReadSchemeActionReg(physicalSchemeId);
-+ p_HcFrame->extraReg = HC_HCOR_KG_SCHEME_REGS_MASK;
-+ p_HcFrame->commandSequence = seqNum;
-+ BUILD_FD(SIZE_OF_HC_FRAME_READ_OR_CC_DYNAMIC);
-+ if ((err = EnQFrm(p_FmHc, &fmFd, seqNum)) != E_OK)
-+ {
-+ PutBuf(p_FmHc, p_HcFrame, seqNum);
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+ }
-+
-+ /* specific change */
-+ if ((requiredAction & UPDATE_NIA_ENQ_WITHOUT_DMA) &&
-+ ((FmPcdKgGetNextEngine(p_FmHc->h_FmPcd, relativeSchemeId) == e_FM_PCD_DONE) &&
-+ (FmPcdKgGetDoneAction(p_FmHc->h_FmPcd, relativeSchemeId) == e_FM_PCD_ENQ_FRAME)))
-+ {
-+ tmpReg32 = p_HcFrame->hcSpecificData.schemeRegs.kgse_mode;
-+ ASSERT_COND(tmpReg32 & (NIA_ENG_BMI | NIA_BMI_AC_ENQ_FRAME));
-+ p_HcFrame->hcSpecificData.schemeRegs.kgse_mode = tmpReg32 | NIA_BMI_AC_ENQ_FRAME_WITHOUT_DMA;
-+ }
-+
-+ if ((requiredAction & UPDATE_KG_NIA_CC_WA) &&
-+ (FmPcdKgGetNextEngine(p_FmHc->h_FmPcd, relativeSchemeId) == e_FM_PCD_CC))
-+ {
-+ tmpReg32 = p_HcFrame->hcSpecificData.schemeRegs.kgse_mode;
-+ ASSERT_COND(tmpReg32 & (NIA_ENG_FM_CTL | NIA_FM_CTL_AC_CC));
-+ tmpReg32 &= ~NIA_FM_CTL_AC_CC;
-+ p_HcFrame->hcSpecificData.schemeRegs.kgse_mode = tmpReg32 | NIA_FM_CTL_AC_PRE_CC;
-+ }
-+
-+ if (requiredAction & UPDATE_KG_OPT_MODE)
-+ p_HcFrame->hcSpecificData.schemeRegs.kgse_om = value;
-+
-+ if (requiredAction & UPDATE_KG_NIA)
-+ {
-+ tmpReg32 = p_HcFrame->hcSpecificData.schemeRegs.kgse_mode;
-+ tmpReg32 &= ~(NIA_ENG_MASK | NIA_AC_MASK);
-+ tmpReg32 |= value;
-+ p_HcFrame->hcSpecificData.schemeRegs.kgse_mode = tmpReg32;
-+ }
-+
-+ /* Post change general code */
-+ p_HcFrame->opcode = (uint32_t)(HC_HCOR_GBL | HC_HCOR_OPCODE_KG_SCM);
-+ p_HcFrame->actionReg = FmPcdKgBuildWriteSchemeActionReg(physicalSchemeId, FALSE);
-+ p_HcFrame->extraReg = HC_HCOR_KG_SCHEME_REGS_MASK;
-+
-+ BUILD_FD(sizeof(t_HcFrame));
-+ err = EnQFrm(p_FmHc, &fmFd, seqNum);
-+
-+ PutBuf(p_FmHc, p_HcFrame, seqNum);
-+
-+ if (err != E_OK)
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+ }
-+ }
-+
-+ return E_OK;
-+}
-+
-+uint32_t FmHcPcdKgGetSchemeCounter(t_Handle h_FmHc, t_Handle h_Scheme)
-+{
-+ t_FmHc *p_FmHc = (t_FmHc*)h_FmHc;
-+ t_Error err;
-+ t_HcFrame *p_HcFrame;
-+ t_DpaaFD fmFd;
-+ uint32_t retVal;
-+ uint8_t relativeSchemeId;
-+ uint8_t physicalSchemeId = FmPcdKgGetSchemeId(h_Scheme);
-+ uint32_t seqNum;
-+
-+ relativeSchemeId = FmPcdKgGetRelativeSchemeId(p_FmHc->h_FmPcd, physicalSchemeId);
-+ if ( relativeSchemeId == FM_PCD_KG_NUM_OF_SCHEMES)
-+ {
-+ REPORT_ERROR(MAJOR, E_NOT_IN_RANGE, NO_MSG);
-+ return 0;
-+ }
-+
-+ /* first read scheme and check that it is valid */
-+ p_HcFrame = GetBuf(p_FmHc, &seqNum);
-+ if (!p_HcFrame)
-+ {
-+ REPORT_ERROR(MINOR, E_NO_MEMORY, ("HC Frame object"));
-+ return 0;
-+ }
-+ memset(p_HcFrame, 0, sizeof(t_HcFrame));
-+ p_HcFrame->opcode = (uint32_t)(HC_HCOR_GBL | HC_HCOR_OPCODE_KG_SCM);
-+ p_HcFrame->actionReg = FmPcdKgBuildReadSchemeActionReg(physicalSchemeId);
-+ p_HcFrame->extraReg = HC_HCOR_KG_SCHEME_REGS_MASK;
-+ p_HcFrame->commandSequence = seqNum;
-+
-+ BUILD_FD(SIZE_OF_HC_FRAME_READ_OR_CC_DYNAMIC);
-+
-+ err = EnQFrm(p_FmHc, &fmFd, seqNum);
-+ if (err != E_OK)
-+ {
-+ PutBuf(p_FmHc, p_HcFrame, seqNum);
-+ REPORT_ERROR(MINOR, err, NO_MSG);
-+ return 0;
-+ }
-+
-+ if (!FmPcdKgHwSchemeIsValid(p_HcFrame->hcSpecificData.schemeRegs.kgse_mode))
-+ {
-+ PutBuf(p_FmHc, p_HcFrame, seqNum);
-+ REPORT_ERROR(MAJOR, E_ALREADY_EXISTS, ("Scheme is invalid"));
-+ return 0;
-+ }
-+
-+ retVal = p_HcFrame->hcSpecificData.schemeRegs.kgse_spc;
-+ PutBuf(p_FmHc, p_HcFrame, seqNum);
-+
-+ return retVal;
-+}
-+
-+t_Error FmHcPcdKgSetSchemeCounter(t_Handle h_FmHc, t_Handle h_Scheme, uint32_t value)
-+{
-+ t_FmHc *p_FmHc = (t_FmHc*)h_FmHc;
-+ t_Error err = E_OK;
-+ t_HcFrame *p_HcFrame;
-+ t_DpaaFD fmFd;
-+ uint8_t relativeSchemeId, physicalSchemeId;
-+ uint32_t seqNum;
-+
-+ physicalSchemeId = FmPcdKgGetSchemeId(h_Scheme);
-+ relativeSchemeId = FmPcdKgGetRelativeSchemeId(p_FmHc->h_FmPcd, physicalSchemeId);
-+ if ( relativeSchemeId == FM_PCD_KG_NUM_OF_SCHEMES)
-+ RETURN_ERROR(MAJOR, E_NOT_IN_RANGE, NO_MSG);
-+
-+ /* first read scheme and check that it is valid */
-+ p_HcFrame = GetBuf(p_FmHc, &seqNum);
-+ if (!p_HcFrame)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("HC Frame object"));
-+ memset(p_HcFrame, 0, sizeof(t_HcFrame));
-+ p_HcFrame->opcode = (uint32_t)(HC_HCOR_GBL | HC_HCOR_OPCODE_KG_SCM);
-+ p_HcFrame->actionReg = FmPcdKgBuildWriteSchemeActionReg(physicalSchemeId, TRUE);
-+ p_HcFrame->extraReg = HC_HCOR_KG_SCHEME_COUNTER;
-+ /* write counter */
-+ p_HcFrame->hcSpecificData.singleRegForWrite = value;
-+ p_HcFrame->commandSequence = seqNum;
-+
-+ BUILD_FD(sizeof(t_HcFrame));
-+
-+ err = EnQFrm(p_FmHc, &fmFd, seqNum);
-+
-+ PutBuf(p_FmHc, p_HcFrame, seqNum);
-+ return err;
-+}
-+
-+t_Error FmHcPcdKgSetClsPlan(t_Handle h_FmHc, t_FmPcdKgInterModuleClsPlanSet *p_Set)
-+{
-+ t_FmHc *p_FmHc = (t_FmHc*)h_FmHc;
-+ t_HcFrame *p_HcFrame;
-+ t_DpaaFD fmFd;
-+ uint8_t i, idx;
-+ uint32_t seqNum;
-+ t_Error err = E_OK;
-+
-+ ASSERT_COND(p_FmHc);
-+
-+ p_HcFrame = GetBuf(p_FmHc, &seqNum);
-+ if (!p_HcFrame)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("HC Frame object"));
-+
-+ for (i = p_Set->baseEntry; i < (p_Set->baseEntry+p_Set->numOfClsPlanEntries); i+=8)
-+ {
-+ memset(p_HcFrame, 0, sizeof(t_HcFrame));
-+ p_HcFrame->opcode = (uint32_t)(HC_HCOR_GBL | HC_HCOR_OPCODE_KG_SCM);
-+ p_HcFrame->actionReg = FmPcdKgBuildWriteClsPlanBlockActionReg((uint8_t)(i / CLS_PLAN_NUM_PER_GRP));
-+ p_HcFrame->extraReg = HC_HCOR_KG_SCHEME_REGS_MASK;
-+
-+ idx = (uint8_t)(i - p_Set->baseEntry);
-+ ASSERT_COND(idx < FM_PCD_MAX_NUM_OF_CLS_PLANS);
-+ memcpy(&p_HcFrame->hcSpecificData.clsPlanEntries, &p_Set->vectors[idx], CLS_PLAN_NUM_PER_GRP*sizeof(uint32_t));
-+ p_HcFrame->commandSequence = seqNum;
-+
-+ BUILD_FD(sizeof(t_HcFrame));
-+
-+ if ((err = EnQFrm(p_FmHc, &fmFd, seqNum)) != E_OK)
-+ {
-+ PutBuf(p_FmHc, p_HcFrame, seqNum);
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+ }
-+ }
-+
-+ PutBuf(p_FmHc, p_HcFrame, seqNum);
-+ return err;
-+}
-+
-+t_Error FmHcPcdKgDeleteClsPlan(t_Handle h_FmHc, uint8_t grpId)
-+{
-+ t_FmHc *p_FmHc = (t_FmHc*)h_FmHc;
-+ t_FmPcdKgInterModuleClsPlanSet *p_ClsPlanSet;
-+
-+ p_ClsPlanSet = (t_FmPcdKgInterModuleClsPlanSet *)XX_Malloc(sizeof(t_FmPcdKgInterModuleClsPlanSet));
-+ if (!p_ClsPlanSet)
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, ("Classification plan set"));
-+
-+ memset(p_ClsPlanSet, 0, sizeof(t_FmPcdKgInterModuleClsPlanSet));
-+
-+ p_ClsPlanSet->baseEntry = FmPcdKgGetClsPlanGrpBase(p_FmHc->h_FmPcd, grpId);
-+ p_ClsPlanSet->numOfClsPlanEntries = FmPcdKgGetClsPlanGrpSize(p_FmHc->h_FmPcd, grpId);
-+ ASSERT_COND(p_ClsPlanSet->numOfClsPlanEntries <= FM_PCD_MAX_NUM_OF_CLS_PLANS);
-+
-+ if (FmHcPcdKgSetClsPlan(p_FmHc, p_ClsPlanSet) != E_OK)
-+ {
-+ XX_Free(p_ClsPlanSet);
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, NO_MSG);
-+ }
-+
-+ XX_Free(p_ClsPlanSet);
-+ FmPcdKgDestroyClsPlanGrp(p_FmHc->h_FmPcd, grpId);
-+
-+ return E_OK;
-+}
-+
-+t_Error FmHcPcdCcCapwapTimeoutReassm(t_Handle h_FmHc, t_FmPcdCcCapwapReassmTimeoutParams *p_CcCapwapReassmTimeoutParams )
-+{
-+ t_FmHc *p_FmHc = (t_FmHc*)h_FmHc;
-+ t_HcFrame *p_HcFrame;
-+ t_DpaaFD fmFd;
-+ t_Error err;
-+ uint32_t seqNum;
-+
-+ SANITY_CHECK_RETURN_VALUE(h_FmHc, E_INVALID_HANDLE,0);
-+
-+ p_HcFrame = GetBuf(p_FmHc, &seqNum);
-+ if (!p_HcFrame)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("HC Frame object"));
-+
-+ memset(p_HcFrame, 0, sizeof(t_HcFrame));
-+ p_HcFrame->opcode = (uint32_t)(HC_HCOR_GBL | HC_HCOR_OPCODE_CC_CAPWAP_REASSM_TIMEOUT);
-+ memcpy(&p_HcFrame->hcSpecificData.ccCapwapReassmTimeout, p_CcCapwapReassmTimeoutParams, sizeof(t_FmPcdCcCapwapReassmTimeoutParams));
-+ p_HcFrame->commandSequence = seqNum;
-+ BUILD_FD(sizeof(t_HcFrame));
-+
-+ err = EnQFrm(p_FmHc, &fmFd, seqNum);
-+
-+ PutBuf(p_FmHc, p_HcFrame, seqNum);
-+ return err;
-+}
-+
-+t_Error FmHcPcdCcIpFragScratchPollCmd(t_Handle h_FmHc, bool fill, t_FmPcdCcFragScratchPoolCmdParams *p_FmPcdCcFragScratchPoolCmdParams)
-+{
-+ t_FmHc *p_FmHc = (t_FmHc*)h_FmHc;
-+ t_HcFrame *p_HcFrame;
-+ t_DpaaFD fmFd;
-+ t_Error err;
-+ uint32_t seqNum;
-+
-+ SANITY_CHECK_RETURN_VALUE(h_FmHc, E_INVALID_HANDLE,0);
-+
-+ p_HcFrame = GetBuf(p_FmHc, &seqNum);
-+ if (!p_HcFrame)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("HC Frame object"));
-+
-+ memset(p_HcFrame, 0, sizeof(t_HcFrame));
-+
-+ p_HcFrame->opcode = (uint32_t)(HC_HCOR_GBL | HC_HCOR_OPCODE_CC_IP_FRAG_INITIALIZATION);
-+ p_HcFrame->actionReg = (uint32_t)(((fill == TRUE) ? 0 : 1) << HC_HCOR_ACTION_REG_IP_FRAG_SCRATCH_POOL_CMD_SHIFT);
-+ p_HcFrame->actionReg |= p_FmPcdCcFragScratchPoolCmdParams->bufferPoolId << HC_HCOR_ACTION_REG_IP_FRAG_SCRATCH_POOL_BPID;
-+ if (fill == TRUE)
-+ {
-+ p_HcFrame->extraReg = p_FmPcdCcFragScratchPoolCmdParams->numOfBuffers;
-+ }
-+ p_HcFrame->commandSequence = seqNum;
-+
-+ BUILD_FD(sizeof(t_HcFrame));
-+ if ((err = EnQFrm(p_FmHc, &fmFd, seqNum)) != E_OK)
-+ {
-+ PutBuf(p_FmHc, p_HcFrame, seqNum);
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+ }
-+
-+ p_FmPcdCcFragScratchPoolCmdParams->numOfBuffers = p_HcFrame->extraReg;
-+
-+ PutBuf(p_FmHc, p_HcFrame, seqNum);
-+ return E_OK;
-+}
-+
-+t_Error FmHcPcdCcTimeoutReassm(t_Handle h_FmHc, t_FmPcdCcReassmTimeoutParams *p_CcReassmTimeoutParams, uint8_t *p_Result)
-+{
-+ t_FmHc *p_FmHc = (t_FmHc*)h_FmHc;
-+ t_HcFrame *p_HcFrame;
-+ t_DpaaFD fmFd;
-+ t_Error err;
-+ uint32_t seqNum;
-+
-+ SANITY_CHECK_RETURN_VALUE(h_FmHc, E_INVALID_HANDLE,0);
-+
-+ p_HcFrame = GetBuf(p_FmHc, &seqNum);
-+ if (!p_HcFrame)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("HC Frame object"));
-+
-+ memset(p_HcFrame, 0, sizeof(t_HcFrame));
-+ p_HcFrame->opcode = (uint32_t)(HC_HCOR_GBL | HC_HCOR_OPCODE_CC_REASSM_TIMEOUT);
-+ p_HcFrame->actionReg = (uint32_t)((p_CcReassmTimeoutParams->activate ? 0 : 1) << HC_HCOR_ACTION_REG_REASSM_TIMEOUT_ACTIVE_SHIFT);
-+ p_HcFrame->extraReg = (p_CcReassmTimeoutParams->tsbs << HC_HCOR_EXTRA_REG_REASSM_TIMEOUT_TSBS_SHIFT) | p_CcReassmTimeoutParams->iprcpt;
-+ p_HcFrame->commandSequence = seqNum;
-+
-+ BUILD_FD(sizeof(t_HcFrame));
-+ if ((err = EnQFrm(p_FmHc, &fmFd, seqNum)) != E_OK)
-+ {
-+ PutBuf(p_FmHc, p_HcFrame, seqNum);
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+ }
-+
-+ *p_Result = (uint8_t)
-+ ((p_HcFrame->actionReg >> HC_HCOR_ACTION_REG_REASSM_TIMEOUT_RES_SHIFT) & HC_HCOR_ACTION_REG_REASSM_TIMEOUT_RES_MASK);
-+
-+ PutBuf(p_FmHc, p_HcFrame, seqNum);
-+ return E_OK;
-+}
-+
-+t_Error FmHcPcdPlcrCcGetSetParams(t_Handle h_FmHc,uint16_t absoluteProfileId, uint32_t requiredAction)
-+{
-+ t_FmHc *p_FmHc = (t_FmHc*)h_FmHc;
-+ t_HcFrame *p_HcFrame;
-+ t_DpaaFD fmFd;
-+ t_Error err;
-+ uint32_t tmpReg32 = 0;
-+ uint32_t requiredActionTmp, requiredActionFlag;
-+ uint32_t seqNum;
-+
-+ SANITY_CHECK_RETURN_VALUE(h_FmHc, E_INVALID_HANDLE,0);
-+
-+ /* Profile is locked by calling routine */
-+ /* WARNING - this lock will not be efficient if other HC routine will attempt to change
-+ * "fmpl_pegnia" "fmpl_peynia" or "fmpl_pernia" without locking Profile !
-+ */
-+
-+ requiredActionTmp = FmPcdPlcrGetRequiredAction(p_FmHc->h_FmPcd, absoluteProfileId);
-+ requiredActionFlag = FmPcdPlcrGetRequiredActionFlag(p_FmHc->h_FmPcd, absoluteProfileId);
-+
-+ if (!requiredActionFlag || !(requiredActionTmp & requiredAction))
-+ {
-+ if (requiredAction & UPDATE_NIA_ENQ_WITHOUT_DMA)
-+ {
-+ p_HcFrame = GetBuf(p_FmHc, &seqNum);
-+ if (!p_HcFrame)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("HC Frame object"));
-+ /* first read scheme and check that it is valid */
-+ memset(p_HcFrame, 0, sizeof(t_HcFrame));
-+ p_HcFrame->opcode = (uint32_t)(HC_HCOR_GBL | HC_HCOR_OPCODE_PLCR_PRFL);
-+ p_HcFrame->actionReg = FmPcdPlcrBuildReadPlcrActionReg(absoluteProfileId);
-+ p_HcFrame->extraReg = 0x00008000;
-+ p_HcFrame->commandSequence = seqNum;
-+
-+ BUILD_FD(SIZE_OF_HC_FRAME_READ_OR_CC_DYNAMIC);
-+
-+ if ((err = EnQFrm(p_FmHc, &fmFd, seqNum)) != E_OK)
-+ {
-+ PutBuf(p_FmHc, p_HcFrame, seqNum);
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+ }
-+
-+ tmpReg32 = p_HcFrame->hcSpecificData.profileRegs.fmpl_pegnia;
-+ if (!(tmpReg32 & (NIA_ENG_BMI | NIA_BMI_AC_ENQ_FRAME)))
-+ {
-+ PutBuf(p_FmHc, p_HcFrame, seqNum);
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE,
-+ ("Next engine of this policer profile has to be assigned to FM_PCD_DONE"));
-+ }
-+
-+ tmpReg32 |= NIA_BMI_AC_ENQ_FRAME_WITHOUT_DMA;
-+
-+ p_HcFrame->opcode = (uint32_t)(HC_HCOR_GBL | HC_HCOR_OPCODE_PLCR_PRFL);
-+ p_HcFrame->actionReg = FmPcdPlcrBuildWritePlcrActionReg(absoluteProfileId);
-+ p_HcFrame->actionReg |= FmPcdPlcrBuildNiaProfileReg(TRUE, FALSE, FALSE);
-+ p_HcFrame->extraReg = 0x00008000;
-+ p_HcFrame->hcSpecificData.singleRegForWrite = tmpReg32;
-+
-+ BUILD_FD(SIZE_OF_HC_FRAME_PROFILE_CNT);
-+
-+ if ((err = EnQFrm(p_FmHc, &fmFd, seqNum)) != E_OK)
-+ {
-+ PutBuf(p_FmHc, p_HcFrame, seqNum);
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+ }
-+
-+ tmpReg32 = p_HcFrame->hcSpecificData.profileRegs.fmpl_peynia;
-+ if (!(tmpReg32 & (NIA_ENG_BMI | NIA_BMI_AC_ENQ_FRAME)))
-+ {
-+ PutBuf(p_FmHc, p_HcFrame, seqNum);
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("Next engine of this policer profile has to be assigned to FM_PCD_DONE"));
-+ }
-+
-+ tmpReg32 |= NIA_BMI_AC_ENQ_FRAME_WITHOUT_DMA;
-+
-+ p_HcFrame->opcode = (uint32_t)(HC_HCOR_GBL | HC_HCOR_OPCODE_PLCR_PRFL);
-+ p_HcFrame->actionReg = FmPcdPlcrBuildWritePlcrActionReg(absoluteProfileId);
-+ p_HcFrame->actionReg |= FmPcdPlcrBuildNiaProfileReg(FALSE, TRUE, FALSE);
-+ p_HcFrame->extraReg = 0x00008000;
-+ p_HcFrame->hcSpecificData.singleRegForWrite = tmpReg32;
-+
-+ BUILD_FD(SIZE_OF_HC_FRAME_PROFILE_CNT);
-+
-+ if ((err = EnQFrm(p_FmHc, &fmFd, seqNum)) != E_OK)
-+ {
-+ PutBuf(p_FmHc, p_HcFrame, seqNum);
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+ }
-+
-+ tmpReg32 = p_HcFrame->hcSpecificData.profileRegs.fmpl_pernia;
-+ if (!(tmpReg32 & (NIA_ENG_BMI | NIA_BMI_AC_ENQ_FRAME)))
-+ {
-+ PutBuf(p_FmHc, p_HcFrame, seqNum);
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("Next engine of this policer profile has to be assigned to FM_PCD_DONE"));
-+ }
-+
-+ tmpReg32 |= NIA_BMI_AC_ENQ_FRAME_WITHOUT_DMA;
-+
-+ p_HcFrame->opcode = (uint32_t)(HC_HCOR_GBL | HC_HCOR_OPCODE_PLCR_PRFL);
-+ p_HcFrame->actionReg = FmPcdPlcrBuildWritePlcrActionReg(absoluteProfileId);
-+ p_HcFrame->actionReg |= FmPcdPlcrBuildNiaProfileReg(FALSE, FALSE, TRUE);
-+ p_HcFrame->extraReg = 0x00008000;
-+ p_HcFrame->hcSpecificData.singleRegForWrite = tmpReg32;
-+
-+ BUILD_FD(SIZE_OF_HC_FRAME_PROFILE_CNT);
-+
-+ if ((err = EnQFrm(p_FmHc, &fmFd, seqNum)) != E_OK)
-+ {
-+ PutBuf(p_FmHc, p_HcFrame, seqNum);
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+ }
-+
-+ PutBuf(p_FmHc, p_HcFrame, seqNum);
-+ }
-+ }
-+
-+ return E_OK;
-+}
-+
-+t_Error FmHcPcdPlcrSetProfile(t_Handle h_FmHc, t_Handle h_Profile, t_FmPcdPlcrProfileRegs *p_PlcrRegs)
-+{
-+ t_FmHc *p_FmHc = (t_FmHc*)h_FmHc;
-+ t_Error err = E_OK;
-+ uint16_t profileIndx;
-+ t_HcFrame *p_HcFrame;
-+ t_DpaaFD fmFd;
-+ uint32_t seqNum;
-+
-+ p_HcFrame = GetBuf(p_FmHc, &seqNum);
-+ if (!p_HcFrame)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("HC Frame object"));
-+
-+ profileIndx = FmPcdPlcrProfileGetAbsoluteId(h_Profile);
-+
-+ memset(p_HcFrame, 0, sizeof(t_HcFrame));
-+ p_HcFrame->opcode = (uint32_t)(HC_HCOR_GBL | HC_HCOR_OPCODE_PLCR_PRFL);
-+ p_HcFrame->actionReg = FmPcdPlcrBuildWritePlcrActionRegs(profileIndx);
-+ p_HcFrame->extraReg = 0x00008000;
-+ memcpy(&p_HcFrame->hcSpecificData.profileRegs, p_PlcrRegs, sizeof(t_FmPcdPlcrProfileRegs));
-+ p_HcFrame->commandSequence = seqNum;
-+
-+ BUILD_FD(sizeof(t_HcFrame));
-+
-+ err = EnQFrm(p_FmHc, &fmFd, seqNum);
-+
-+ PutBuf(p_FmHc, p_HcFrame, seqNum);
-+
-+ if (err != E_OK)
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+
-+ return E_OK;
-+}
-+
-+t_Error FmHcPcdPlcrDeleteProfile(t_Handle h_FmHc, t_Handle h_Profile)
-+{
-+ t_FmHc *p_FmHc = (t_FmHc*)h_FmHc;
-+ uint16_t absoluteProfileId = FmPcdPlcrProfileGetAbsoluteId(h_Profile);
-+ t_Error err = E_OK;
-+ t_HcFrame *p_HcFrame;
-+ t_DpaaFD fmFd;
-+ uint32_t seqNum;
-+
-+ p_HcFrame = GetBuf(p_FmHc, &seqNum);
-+ if (!p_HcFrame)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("HC Frame object"));
-+ memset(p_HcFrame, 0, sizeof(t_HcFrame));
-+ p_HcFrame->opcode = (uint32_t)(HC_HCOR_GBL | HC_HCOR_OPCODE_PLCR_PRFL);
-+ p_HcFrame->actionReg = FmPcdPlcrBuildWritePlcrActionReg(absoluteProfileId);
-+ p_HcFrame->actionReg |= 0x00008000;
-+ p_HcFrame->extraReg = 0x00008000;
-+ memset(&p_HcFrame->hcSpecificData.profileRegs, 0, sizeof(t_FmPcdPlcrProfileRegs));
-+ p_HcFrame->commandSequence = seqNum;
-+
-+ BUILD_FD(sizeof(t_HcFrame));
-+
-+ err = EnQFrm(p_FmHc, &fmFd, seqNum);
-+
-+ PutBuf(p_FmHc, p_HcFrame, seqNum);
-+
-+ if (err != E_OK)
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+
-+ return E_OK;
-+}
-+
-+t_Error FmHcPcdPlcrSetProfileCounter(t_Handle h_FmHc, t_Handle h_Profile, e_FmPcdPlcrProfileCounters counter, uint32_t value)
-+{
-+
-+ t_FmHc *p_FmHc = (t_FmHc*)h_FmHc;
-+ uint16_t absoluteProfileId = FmPcdPlcrProfileGetAbsoluteId(h_Profile);
-+ t_Error err = E_OK;
-+ t_HcFrame *p_HcFrame;
-+ t_DpaaFD fmFd;
-+ uint32_t seqNum;
-+
-+ /* first read scheme and check that it is valid */
-+ p_HcFrame = GetBuf(p_FmHc, &seqNum);
-+ if (!p_HcFrame)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("HC Frame object"));
-+ memset(p_HcFrame, 0, sizeof(t_HcFrame));
-+ p_HcFrame->opcode = (uint32_t)(HC_HCOR_GBL | HC_HCOR_OPCODE_PLCR_PRFL);
-+ p_HcFrame->actionReg = FmPcdPlcrBuildWritePlcrActionReg(absoluteProfileId);
-+ p_HcFrame->actionReg |= FmPcdPlcrBuildCounterProfileReg(counter);
-+ p_HcFrame->extraReg = 0x00008000;
-+ p_HcFrame->hcSpecificData.singleRegForWrite = value;
-+ p_HcFrame->commandSequence = seqNum;
-+
-+ BUILD_FD(SIZE_OF_HC_FRAME_PROFILE_CNT);
-+
-+ err = EnQFrm(p_FmHc, &fmFd, seqNum);
-+
-+ PutBuf(p_FmHc, p_HcFrame, seqNum);
-+
-+ if (err != E_OK)
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+
-+ return E_OK;
-+}
-+
-+uint32_t FmHcPcdPlcrGetProfileCounter(t_Handle h_FmHc, t_Handle h_Profile, e_FmPcdPlcrProfileCounters counter)
-+{
-+ t_FmHc *p_FmHc = (t_FmHc*)h_FmHc;
-+ uint16_t absoluteProfileId = FmPcdPlcrProfileGetAbsoluteId(h_Profile);
-+ t_Error err;
-+ t_HcFrame *p_HcFrame;
-+ t_DpaaFD fmFd;
-+ uint32_t retVal = 0;
-+ uint32_t seqNum;
-+
-+ SANITY_CHECK_RETURN_VALUE(h_FmHc, E_INVALID_HANDLE,0);
-+
-+ /* first read scheme and check that it is valid */
-+ p_HcFrame = GetBuf(p_FmHc, &seqNum);
-+ if (!p_HcFrame)
-+ {
-+ REPORT_ERROR(MINOR, E_NO_MEMORY, ("HC Frame object"));
-+ return 0;
-+ }
-+ memset(p_HcFrame, 0, sizeof(t_HcFrame));
-+ p_HcFrame->opcode = (uint32_t)(HC_HCOR_GBL | HC_HCOR_OPCODE_PLCR_PRFL);
-+ p_HcFrame->actionReg = FmPcdPlcrBuildReadPlcrActionReg(absoluteProfileId);
-+ p_HcFrame->extraReg = 0x00008000;
-+ p_HcFrame->commandSequence = seqNum;
-+
-+ BUILD_FD(SIZE_OF_HC_FRAME_READ_OR_CC_DYNAMIC);
-+
-+ err = EnQFrm(p_FmHc, &fmFd, seqNum);
-+ if (err != E_OK)
-+ {
-+ PutBuf(p_FmHc, p_HcFrame, seqNum);
-+ REPORT_ERROR(MINOR, err, NO_MSG);
-+ return 0;
-+ }
-+
-+ switch (counter)
-+ {
-+ case e_FM_PCD_PLCR_PROFILE_GREEN_PACKET_TOTAL_COUNTER:
-+ retVal = p_HcFrame->hcSpecificData.profileRegs.fmpl_pegpc;
-+ break;
-+ case e_FM_PCD_PLCR_PROFILE_YELLOW_PACKET_TOTAL_COUNTER:
-+ retVal = p_HcFrame->hcSpecificData.profileRegs.fmpl_peypc;
-+ break;
-+ case e_FM_PCD_PLCR_PROFILE_RED_PACKET_TOTAL_COUNTER:
-+ retVal = p_HcFrame->hcSpecificData.profileRegs.fmpl_perpc;
-+ break;
-+ case e_FM_PCD_PLCR_PROFILE_RECOLOURED_YELLOW_PACKET_TOTAL_COUNTER:
-+ retVal = p_HcFrame->hcSpecificData.profileRegs.fmpl_perypc;
-+ break;
-+ case e_FM_PCD_PLCR_PROFILE_RECOLOURED_RED_PACKET_TOTAL_COUNTER:
-+ retVal = p_HcFrame->hcSpecificData.profileRegs.fmpl_perrpc;
-+ break;
-+ default:
-+ REPORT_ERROR(MAJOR, E_INVALID_SELECTION, NO_MSG);
-+ }
-+
-+ PutBuf(p_FmHc, p_HcFrame, seqNum);
-+ return retVal;
-+}
-+
-+t_Error FmHcKgWriteSp(t_Handle h_FmHc, uint8_t hardwarePortId, uint32_t spReg, bool add)
-+{
-+ t_FmHc *p_FmHc = (t_FmHc*)h_FmHc;
-+ t_HcFrame *p_HcFrame;
-+ t_DpaaFD fmFd;
-+ t_Error err = E_OK;
-+ uint32_t seqNum;
-+
-+ ASSERT_COND(p_FmHc);
-+
-+ p_HcFrame = GetBuf(p_FmHc, &seqNum);
-+ if (!p_HcFrame)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("HC Frame object"));
-+ memset(p_HcFrame, 0, sizeof(t_HcFrame));
-+ /* first read SP register */
-+ p_HcFrame->opcode = (uint32_t)(HC_HCOR_GBL | HC_HCOR_OPCODE_KG_SCM);
-+ p_HcFrame->actionReg = FmPcdKgBuildReadPortSchemeBindActionReg(hardwarePortId);
-+ p_HcFrame->extraReg = HC_HCOR_KG_SCHEME_REGS_MASK;
-+ p_HcFrame->commandSequence = seqNum;
-+
-+ BUILD_FD(SIZE_OF_HC_FRAME_PORT_REGS);
-+
-+ if ((err = EnQFrm(p_FmHc, &fmFd, seqNum)) != E_OK)
-+ {
-+ PutBuf(p_FmHc, p_HcFrame, seqNum);
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+ }
-+
-+ /* spReg is the first reg, so we can use it both for read and for write */
-+ if (add)
-+ p_HcFrame->hcSpecificData.portRegsForRead.spReg |= spReg;
-+ else
-+ p_HcFrame->hcSpecificData.portRegsForRead.spReg &= ~spReg;
-+
-+ p_HcFrame->actionReg = FmPcdKgBuildWritePortSchemeBindActionReg(hardwarePortId);
-+
-+ BUILD_FD(sizeof(t_HcFrame));
-+
-+ err = EnQFrm(p_FmHc, &fmFd, seqNum);
-+
-+ PutBuf(p_FmHc, p_HcFrame, seqNum);
-+
-+ if (err != E_OK)
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+
-+ return E_OK;
-+}
-+
-+t_Error FmHcKgWriteCpp(t_Handle h_FmHc, uint8_t hardwarePortId, uint32_t cppReg)
-+{
-+ t_FmHc *p_FmHc = (t_FmHc*)h_FmHc;
-+ t_HcFrame *p_HcFrame;
-+ t_DpaaFD fmFd;
-+ t_Error err = E_OK;
-+ uint32_t seqNum;
-+
-+ ASSERT_COND(p_FmHc);
-+
-+ p_HcFrame = GetBuf(p_FmHc, &seqNum);
-+ if (!p_HcFrame)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("HC Frame object"));
-+ memset(p_HcFrame, 0, sizeof(t_HcFrame));
-+ /* first read SP register */
-+ p_HcFrame->opcode = (uint32_t)(HC_HCOR_GBL | HC_HCOR_OPCODE_KG_SCM);
-+ p_HcFrame->actionReg = FmPcdKgBuildWritePortClsPlanBindActionReg(hardwarePortId);
-+ p_HcFrame->extraReg = HC_HCOR_KG_SCHEME_REGS_MASK;
-+ p_HcFrame->hcSpecificData.singleRegForWrite = cppReg;
-+ p_HcFrame->commandSequence = seqNum;
-+
-+ BUILD_FD(sizeof(t_HcFrame));
-+
-+ err = EnQFrm(p_FmHc, &fmFd, seqNum);
-+
-+ PutBuf(p_FmHc, p_HcFrame, seqNum);
-+
-+ if (err != E_OK)
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+
-+ return E_OK;
-+}
-+
-+t_Error FmHcPcdCcDoDynamicChange(t_Handle h_FmHc, uint32_t oldAdAddrOffset, uint32_t newAdAddrOffset)
-+{
-+ t_FmHc *p_FmHc = (t_FmHc*)h_FmHc;
-+ t_HcFrame *p_HcFrame;
-+ t_DpaaFD fmFd;
-+ t_Error err = E_OK;
-+ uint32_t seqNum;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmHc, E_INVALID_HANDLE);
-+
-+ p_HcFrame = GetBuf(p_FmHc, &seqNum);
-+ if (!p_HcFrame)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("HC Frame object"));
-+ memset(p_HcFrame, 0, sizeof(t_HcFrame));
-+
-+ p_HcFrame->opcode = (uint32_t)(HC_HCOR_GBL | HC_HCOR_OPCODE_CC);
-+ p_HcFrame->actionReg = newAdAddrOffset;
-+ p_HcFrame->actionReg |= 0xc0000000;
-+ p_HcFrame->extraReg = oldAdAddrOffset;
-+ p_HcFrame->commandSequence = seqNum;
-+
-+ BUILD_FD(SIZE_OF_HC_FRAME_READ_OR_CC_DYNAMIC);
-+
-+ err = EnQFrm(p_FmHc, &fmFd, seqNum);
-+
-+ PutBuf(p_FmHc, p_HcFrame, seqNum);
-+
-+ if (err != E_OK)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+
-+ return E_OK;
-+}
-+
-+t_Error FmHcPcdSync(t_Handle h_FmHc)
-+{
-+ t_FmHc *p_FmHc = (t_FmHc*)h_FmHc;
-+ t_HcFrame *p_HcFrame;
-+ t_DpaaFD fmFd;
-+ t_Error err = E_OK;
-+ uint32_t seqNum;
-+
-+ ASSERT_COND(p_FmHc);
-+
-+ p_HcFrame = GetBuf(p_FmHc, &seqNum);
-+ if (!p_HcFrame)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("HC Frame object"));
-+ memset(p_HcFrame, 0, sizeof(t_HcFrame));
-+ /* first read SP register */
-+ p_HcFrame->opcode = (uint32_t)(HC_HCOR_GBL | HC_HCOR_OPCODE_SYNC);
-+ p_HcFrame->actionReg = 0;
-+ p_HcFrame->extraReg = 0;
-+ p_HcFrame->commandSequence = seqNum;
-+
-+ BUILD_FD(sizeof(t_HcFrame));
-+
-+ err = EnQFrm(p_FmHc, &fmFd, seqNum);
-+
-+ PutBuf(p_FmHc, p_HcFrame, seqNum);
-+
-+ if (err != E_OK)
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+
-+ return E_OK;
-+}
-+
-+t_Handle FmHcGetPort(t_Handle h_FmHc)
-+{
-+ t_FmHc *p_FmHc = (t_FmHc*)h_FmHc;
-+ return p_FmHc->h_HcPortDev;
-+}
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MAC/Makefile
-@@ -0,0 +1,28 @@
-+#
-+# Makefile for the Freescale Ethernet controllers
-+#
-+ccflags-y += -DVERSION=\"\"
-+#
-+#Include netcomm SW specific definitions
-+include $(srctree)/drivers/net/ethernet/freescale/sdk_fman/ncsw_config.mk
-+
-+NCSW_FM_INC = $(srctree)/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/inc
-+
-+ccflags-y += -I$(NCSW_FM_INC)
-+
-+obj-y += fsl-ncsw-MAC.o
-+
-+fsl-ncsw-MAC-objs := dtsec.o dtsec_mii_acc.o fm_mac.o tgec.o tgec_mii_acc.o \
-+ fman_dtsec.o fman_dtsec_mii_acc.o fman_memac.o \
-+ fman_tgec.o fman_crc32.o
-+
-+ifeq ($(CONFIG_FMAN_V3H),y)
-+fsl-ncsw-MAC-objs += memac.o memac_mii_acc.o fman_memac_mii_acc.o
-+endif
-+ifeq ($(CONFIG_FMAN_V3L),y)
-+fsl-ncsw-MAC-objs += memac.o memac_mii_acc.o fman_memac_mii_acc.o
-+endif
-+ifeq ($(CONFIG_FMAN_ARM),y)
-+fsl-ncsw-MAC-objs += memac.o memac_mii_acc.o fman_memac_mii_acc.o
-+endif
-+
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MAC/dtsec.c
-@@ -0,0 +1,1504 @@
-+/*
-+ * Copyright 2008-2013 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+/******************************************************************************
-+ @File dtsec.c
-+
-+ @Description FMan dTSEC driver
-+*//***************************************************************************/
-+
-+#include "std_ext.h"
-+#include "error_ext.h"
-+#include "string_ext.h"
-+#include "xx_ext.h"
-+#include "endian_ext.h"
-+#include "debug_ext.h"
-+#include "crc_mac_addr_ext.h"
-+
-+#include "fm_common.h"
-+#include "dtsec.h"
-+#include "fsl_fman_dtsec.h"
-+#include "fsl_fman_dtsec_mii_acc.h"
-+
-+/*****************************************************************************/
-+/* Internal routines */
-+/*****************************************************************************/
-+
-+static t_Error CheckInitParameters(t_Dtsec *p_Dtsec)
-+{
-+ if (ENET_SPEED_FROM_MODE(p_Dtsec->enetMode) >= e_ENET_SPEED_10000)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("Ethernet 1G MAC driver only supports 1G or lower speeds"));
-+ if (p_Dtsec->macId >= FM_MAX_NUM_OF_1G_MACS)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("macId can not be greater than the number of 1G MACs"));
-+ if (p_Dtsec->addr == 0)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("Ethernet MAC Must have a valid MAC Address"));
-+ if ((ENET_SPEED_FROM_MODE(p_Dtsec->enetMode) >= e_ENET_SPEED_1000) &&
-+ p_Dtsec->p_DtsecDriverParam->halfdup_on)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("Ethernet MAC 1G can't work in half duplex"));
-+ if (p_Dtsec->p_DtsecDriverParam->halfdup_on && (p_Dtsec->p_DtsecDriverParam)->loopback)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("LoopBack is not supported in halfDuplex mode"));
-+#ifdef FM_RX_PREAM_4_ERRATA_DTSEC_A001
-+ if (p_Dtsec->fmMacControllerDriver.fmRevInfo.majorRev <= 6) /* fixed for rev3 */
-+ if (p_Dtsec->p_DtsecDriverParam->rx_preamble)
-+ RETURN_ERROR(MAJOR, E_NOT_SUPPORTED, ("preambleRxEn"));
-+#endif /* FM_RX_PREAM_4_ERRATA_DTSEC_A001 */
-+ if (((p_Dtsec->p_DtsecDriverParam)->tx_preamble || (p_Dtsec->p_DtsecDriverParam)->rx_preamble) &&( (p_Dtsec->p_DtsecDriverParam)->preamble_len != 0x7))
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("Preamble length should be 0x7 bytes"));
-+ if ((p_Dtsec->p_DtsecDriverParam)->halfdup_on &&
-+ (p_Dtsec->p_DtsecDriverParam->tx_time_stamp_en || p_Dtsec->p_DtsecDriverParam->rx_time_stamp_en))
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("dTSEC in half duplex mode has to be with 1588 timeStamping diable"));
-+ if ((p_Dtsec->p_DtsecDriverParam)->rx_flow && (p_Dtsec->p_DtsecDriverParam)->rx_ctrl_acc )
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("Receive control frame are not passed to the system memory so it can not be accept "));
-+ if ((p_Dtsec->p_DtsecDriverParam)->rx_prepend > MAX_PACKET_ALIGNMENT)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("packetAlignmentPadding can't be greater than %d ",MAX_PACKET_ALIGNMENT ));
-+ if (((p_Dtsec->p_DtsecDriverParam)->non_back_to_back_ipg1 > MAX_INTER_PACKET_GAP) ||
-+ ((p_Dtsec->p_DtsecDriverParam)->non_back_to_back_ipg2 > MAX_INTER_PACKET_GAP) ||
-+ ((p_Dtsec->p_DtsecDriverParam)->back_to_back_ipg > MAX_INTER_PACKET_GAP))
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("Inter packet gap can't be greater than %d ",MAX_INTER_PACKET_GAP ));
-+ if ((p_Dtsec->p_DtsecDriverParam)->halfdup_alt_backoff_val > MAX_INTER_PALTERNATE_BEB)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("alternateBackoffVal can't be greater than %d ",MAX_INTER_PALTERNATE_BEB ));
-+ if ((p_Dtsec->p_DtsecDriverParam)->halfdup_retransmit > MAX_RETRANSMISSION)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("maxRetransmission can't be greater than %d ",MAX_RETRANSMISSION ));
-+ if ((p_Dtsec->p_DtsecDriverParam)->halfdup_coll_window > MAX_COLLISION_WINDOW)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("collisionWindow can't be greater than %d ",MAX_COLLISION_WINDOW ));
-+
-+ /* If Auto negotiation process is disabled, need to */
-+ /* Set up the PHY using the MII Management Interface */
-+ if (p_Dtsec->p_DtsecDriverParam->tbipa > MAX_PHYS)
-+ RETURN_ERROR(MAJOR, E_NOT_IN_RANGE, ("PHY address (should be 0-%d)", MAX_PHYS));
-+ if (!p_Dtsec->f_Exception)
-+ RETURN_ERROR(MAJOR, E_INVALID_HANDLE, ("uninitialized f_Exception"));
-+ if (!p_Dtsec->f_Event)
-+ RETURN_ERROR(MAJOR, E_INVALID_HANDLE, ("uninitialized f_Event"));
-+
-+#ifdef FM_LEN_CHECK_ERRATA_FMAN_SW002
-+ if (p_Dtsec->p_DtsecDriverParam->rx_len_check)
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, ("LengthCheck!"));
-+#endif /* FM_LEN_CHECK_ERRATA_FMAN_SW002 */
-+
-+ return E_OK;
-+}
-+
-+/* ......................................................................... */
-+
-+static uint32_t GetMacAddrHashCode(uint64_t ethAddr)
-+{
-+ uint32_t crc;
-+
-+ /* CRC calculation */
-+ GET_MAC_ADDR_CRC(ethAddr, crc);
-+
-+ crc = GetMirror32(crc);
-+
-+ return crc;
-+}
-+
-+/* ......................................................................... */
-+
-+static void UpdateStatistics(t_Dtsec *p_Dtsec)
-+{
-+ uint32_t car1, car2;
-+
-+ fman_dtsec_get_clear_carry_regs(p_Dtsec->p_MemMap, &car1, &car2);
-+
-+ if (car1)
-+ {
-+ if (car1 & CAR1_TR64)
-+ p_Dtsec->internalStatistics.tr64 += VAL22BIT;
-+ if (car1 & CAR1_TR127)
-+ p_Dtsec->internalStatistics.tr127 += VAL22BIT;
-+ if (car1 & CAR1_TR255)
-+ p_Dtsec->internalStatistics.tr255 += VAL22BIT;
-+ if (car1 & CAR1_TR511)
-+ p_Dtsec->internalStatistics.tr511 += VAL22BIT;
-+ if (car1 & CAR1_TRK1)
-+ p_Dtsec->internalStatistics.tr1k += VAL22BIT;
-+ if (car1 & CAR1_TRMAX)
-+ p_Dtsec->internalStatistics.trmax += VAL22BIT;
-+ if (car1 & CAR1_TRMGV)
-+ p_Dtsec->internalStatistics.trmgv += VAL22BIT;
-+ if (car1 & CAR1_RBYT)
-+ p_Dtsec->internalStatistics.rbyt += (uint64_t)VAL32BIT;
-+ if (car1 & CAR1_RPKT)
-+ p_Dtsec->internalStatistics.rpkt += VAL22BIT;
-+ if (car1 & CAR1_RMCA)
-+ p_Dtsec->internalStatistics.rmca += VAL22BIT;
-+ if (car1 & CAR1_RBCA)
-+ p_Dtsec->internalStatistics.rbca += VAL22BIT;
-+ if (car1 & CAR1_RXPF)
-+ p_Dtsec->internalStatistics.rxpf += VAL16BIT;
-+ if (car1 & CAR1_RALN)
-+ p_Dtsec->internalStatistics.raln += VAL16BIT;
-+ if (car1 & CAR1_RFLR)
-+ p_Dtsec->internalStatistics.rflr += VAL16BIT;
-+ if (car1 & CAR1_RCDE)
-+ p_Dtsec->internalStatistics.rcde += VAL16BIT;
-+ if (car1 & CAR1_RCSE)
-+ p_Dtsec->internalStatistics.rcse += VAL16BIT;
-+ if (car1 & CAR1_RUND)
-+ p_Dtsec->internalStatistics.rund += VAL16BIT;
-+ if (car1 & CAR1_ROVR)
-+ p_Dtsec->internalStatistics.rovr += VAL16BIT;
-+ if (car1 & CAR1_RFRG)
-+ p_Dtsec->internalStatistics.rfrg += VAL16BIT;
-+ if (car1 & CAR1_RJBR)
-+ p_Dtsec->internalStatistics.rjbr += VAL16BIT;
-+ if (car1 & CAR1_RDRP)
-+ p_Dtsec->internalStatistics.rdrp += VAL16BIT;
-+ }
-+ if (car2)
-+ {
-+ if (car2 & CAR2_TFCS)
-+ p_Dtsec->internalStatistics.tfcs += VAL12BIT;
-+ if (car2 & CAR2_TBYT)
-+ p_Dtsec->internalStatistics.tbyt += (uint64_t)VAL32BIT;
-+ if (car2 & CAR2_TPKT)
-+ p_Dtsec->internalStatistics.tpkt += VAL22BIT;
-+ if (car2 & CAR2_TMCA)
-+ p_Dtsec->internalStatistics.tmca += VAL22BIT;
-+ if (car2 & CAR2_TBCA)
-+ p_Dtsec->internalStatistics.tbca += VAL22BIT;
-+ if (car2 & CAR2_TXPF)
-+ p_Dtsec->internalStatistics.txpf += VAL16BIT;
-+ if (car2 & CAR2_TDRP)
-+ p_Dtsec->internalStatistics.tdrp += VAL16BIT;
-+ }
-+}
-+
-+/* .............................................................................. */
-+
-+static uint16_t DtsecGetMaxFrameLength(t_Handle h_Dtsec)
-+{
-+ t_Dtsec *p_Dtsec = (t_Dtsec *)h_Dtsec;
-+
-+ SANITY_CHECK_RETURN_VALUE(p_Dtsec, E_INVALID_HANDLE, 0);
-+ SANITY_CHECK_RETURN_VALUE(!p_Dtsec->p_DtsecDriverParam, E_INVALID_STATE, 0);
-+
-+ return fman_dtsec_get_max_frame_len(p_Dtsec->p_MemMap);
-+}
-+
-+/* .............................................................................. */
-+
-+static void DtsecIsr(t_Handle h_Dtsec)
-+{
-+ t_Dtsec *p_Dtsec = (t_Dtsec *)h_Dtsec;
-+ uint32_t event;
-+ struct dtsec_regs *p_DtsecMemMap = p_Dtsec->p_MemMap;
-+
-+ /* do not handle MDIO events */
-+ event = fman_dtsec_get_event(p_DtsecMemMap, (uint32_t)(~(DTSEC_IMASK_MMRDEN | DTSEC_IMASK_MMWREN)));
-+
-+ event &= fman_dtsec_get_interrupt_mask(p_DtsecMemMap);
-+
-+ fman_dtsec_ack_event(p_DtsecMemMap, event);
-+
-+ if (event & DTSEC_IMASK_BREN)
-+ p_Dtsec->f_Exception(p_Dtsec->h_App, e_FM_MAC_EX_1G_BAB_RX);
-+ if (event & DTSEC_IMASK_RXCEN)
-+ p_Dtsec->f_Exception(p_Dtsec->h_App, e_FM_MAC_EX_1G_RX_CTL);
-+ if (event & DTSEC_IMASK_MSROEN)
-+ UpdateStatistics(p_Dtsec);
-+ if (event & DTSEC_IMASK_GTSCEN)
-+ p_Dtsec->f_Exception(p_Dtsec->h_App, e_FM_MAC_EX_1G_GRATEFUL_TX_STP_COMPLET);
-+ if (event & DTSEC_IMASK_BTEN)
-+ p_Dtsec->f_Exception(p_Dtsec->h_App, e_FM_MAC_EX_1G_BAB_TX);
-+ if (event & DTSEC_IMASK_TXCEN)
-+ p_Dtsec->f_Exception(p_Dtsec->h_App, e_FM_MAC_EX_1G_TX_CTL);
-+ if (event & DTSEC_IMASK_TXEEN)
-+ p_Dtsec->f_Exception(p_Dtsec->h_App, e_FM_MAC_EX_1G_TX_ERR);
-+ if (event & DTSEC_IMASK_LCEN)
-+ p_Dtsec->f_Exception(p_Dtsec->h_App, e_FM_MAC_EX_1G_LATE_COL);
-+ if (event & DTSEC_IMASK_CRLEN)
-+ p_Dtsec->f_Exception(p_Dtsec->h_App, e_FM_MAC_EX_1G_COL_RET_LMT);
-+ if (event & DTSEC_IMASK_XFUNEN)
-+ {
-+#ifdef FM_TX_LOCKUP_ERRATA_DTSEC6
-+ if (p_Dtsec->fmMacControllerDriver.fmRevInfo.majorRev == 2)
-+ {
-+ uint32_t tpkt1, tmpReg1, tpkt2, tmpReg2, i;
-+ /* a. Write 0x00E0_0C00 to DTSEC_ID */
-+ /* This is a read only regidter */
-+
-+ /* b. Read and save the value of TPKT */
-+ tpkt1 = GET_UINT32(p_DtsecMemMap->tpkt);
-+
-+ /* c. Read the register at dTSEC address offset 0x32C */
-+ tmpReg1 = GET_UINT32(*(uint32_t*)((uint8_t*)p_DtsecMemMap + 0x32c));
-+
-+ /* d. Compare bits [9:15] to bits [25:31] of the register at address offset 0x32C. */
-+ if ((tmpReg1 & 0x007F0000) != (tmpReg1 & 0x0000007F))
-+ {
-+ /* If they are not equal, save the value of this register and wait for at least
-+ * MAXFRM*16 ns */
-+ XX_UDelay((uint32_t)(MIN(DtsecGetMaxFrameLength(p_Dtsec)*16/1000, 1)));
-+ }
-+
-+ /* e. Read and save TPKT again and read the register at dTSEC address offset
-+ 0x32C again*/
-+ tpkt2 = GET_UINT32(p_DtsecMemMap->tpkt);
-+ tmpReg2 = GET_UINT32(*(uint32_t*)((uint8_t*)p_DtsecMemMap + 0x32c));
-+
-+ /* f. Compare the value of TPKT saved in step b to value read in step e. Also
-+ compare bits [9:15] of the register at offset 0x32C saved in step d to the value
-+ of bits [9:15] saved in step e. If the two registers values are unchanged, then
-+ the transmit portion of the dTSEC controller is locked up and the user should
-+ proceed to the recover sequence. */
-+ if ((tpkt1 == tpkt2) && ((tmpReg1 & 0x007F0000) == (tmpReg2 & 0x007F0000)))
-+ {
-+ /* recover sequence */
-+
-+ /* a.Write a 1 to RCTRL[GRS]*/
-+
-+ WRITE_UINT32(p_DtsecMemMap->rctrl, GET_UINT32(p_DtsecMemMap->rctrl) | RCTRL_GRS);
-+
-+ /* b.Wait until IEVENT[GRSC]=1, or at least 100 us has elapsed. */
-+ for (i = 0 ; i < 100 ; i++ )
-+ {
-+ if (GET_UINT32(p_DtsecMemMap->ievent) & DTSEC_IMASK_GRSCEN)
-+ break;
-+ XX_UDelay(1);
-+ }
-+ if (GET_UINT32(p_DtsecMemMap->ievent) & DTSEC_IMASK_GRSCEN)
-+ WRITE_UINT32(p_DtsecMemMap->ievent, DTSEC_IMASK_GRSCEN);
-+ else
-+ DBG(INFO,("Rx lockup due to dTSEC Tx lockup"));
-+
-+ /* c.Write a 1 to bit n of FM_RSTC (offset 0x0CC of FPM)*/
-+ FmResetMac(p_Dtsec->fmMacControllerDriver.h_Fm, e_FM_MAC_1G, p_Dtsec->fmMacControllerDriver.macId);
-+
-+ /* d.Wait 4 Tx clocks (32 ns) */
-+ XX_UDelay(1);
-+
-+ /* e.Write a 0 to bit n of FM_RSTC. */
-+ /* cleared by FMAN */
-+ }
-+ }
-+#endif /* FM_TX_LOCKUP_ERRATA_DTSEC6 */
-+
-+ p_Dtsec->f_Exception(p_Dtsec->h_App, e_FM_MAC_EX_1G_TX_FIFO_UNDRN);
-+ }
-+ if (event & DTSEC_IMASK_MAGEN)
-+ p_Dtsec->f_Exception(p_Dtsec->h_App, e_FM_MAC_EX_1G_MAG_PCKT);
-+ if (event & DTSEC_IMASK_GRSCEN)
-+ p_Dtsec->f_Exception(p_Dtsec->h_App, e_FM_MAC_EX_1G_GRATEFUL_RX_STP_COMPLET);
-+ if (event & DTSEC_IMASK_TDPEEN)
-+ p_Dtsec->f_Exception(p_Dtsec->h_App, e_FM_MAC_EX_1G_TX_DATA_ERR);
-+ if (event & DTSEC_IMASK_RDPEEN)
-+ p_Dtsec->f_Exception(p_Dtsec->h_App, e_FM_MAC_EX_1G_RX_DATA_ERR);
-+
-+ /* - masked interrupts */
-+ ASSERT_COND(!(event & DTSEC_IMASK_ABRTEN));
-+ ASSERT_COND(!(event & DTSEC_IMASK_IFERREN));
-+}
-+
-+static void DtsecMdioIsr(t_Handle h_Dtsec)
-+{
-+ t_Dtsec *p_Dtsec = (t_Dtsec *)h_Dtsec;
-+ uint32_t event;
-+ struct dtsec_regs *p_DtsecMemMap = p_Dtsec->p_MemMap;
-+
-+ event = GET_UINT32(p_DtsecMemMap->ievent);
-+ /* handle only MDIO events */
-+ event &= (DTSEC_IMASK_MMRDEN | DTSEC_IMASK_MMWREN);
-+ if (event)
-+ {
-+ event &= GET_UINT32(p_DtsecMemMap->imask);
-+
-+ WRITE_UINT32(p_DtsecMemMap->ievent, event);
-+
-+ if (event & DTSEC_IMASK_MMRDEN)
-+ p_Dtsec->f_Event(p_Dtsec->h_App, e_FM_MAC_EX_1G_MII_MNG_RD_COMPLET);
-+ if (event & DTSEC_IMASK_MMWREN)
-+ p_Dtsec->f_Event(p_Dtsec->h_App, e_FM_MAC_EX_1G_MII_MNG_WR_COMPLET);
-+ }
-+}
-+
-+static void Dtsec1588Isr(t_Handle h_Dtsec)
-+{
-+ t_Dtsec *p_Dtsec = (t_Dtsec *)h_Dtsec;
-+ uint32_t event;
-+ struct dtsec_regs *p_DtsecMemMap = p_Dtsec->p_MemMap;
-+
-+ if (p_Dtsec->ptpTsuEnabled)
-+ {
-+ event = fman_dtsec_check_and_clear_tmr_event(p_DtsecMemMap);
-+
-+ if (event)
-+ {
-+ ASSERT_COND(event & TMR_PEVENT_TSRE);
-+ p_Dtsec->f_Exception(p_Dtsec->h_App, e_FM_MAC_EX_1G_1588_TS_RX_ERR);
-+ }
-+ }
-+}
-+
-+/* ........................................................................... */
-+
-+static void FreeInitResources(t_Dtsec *p_Dtsec)
-+{
-+ if (p_Dtsec->mdioIrq != NO_IRQ)
-+ {
-+ XX_DisableIntr(p_Dtsec->mdioIrq);
-+ XX_FreeIntr(p_Dtsec->mdioIrq);
-+ }
-+ FmUnregisterIntr(p_Dtsec->fmMacControllerDriver.h_Fm, e_FM_MOD_1G_MAC, p_Dtsec->macId, e_FM_INTR_TYPE_ERR);
-+ FmUnregisterIntr(p_Dtsec->fmMacControllerDriver.h_Fm, e_FM_MOD_1G_MAC, p_Dtsec->macId, e_FM_INTR_TYPE_NORMAL);
-+
-+ /* release the driver's group hash table */
-+ FreeHashTable(p_Dtsec->p_MulticastAddrHash);
-+ p_Dtsec->p_MulticastAddrHash = NULL;
-+
-+ /* release the driver's individual hash table */
-+ FreeHashTable(p_Dtsec->p_UnicastAddrHash);
-+ p_Dtsec->p_UnicastAddrHash = NULL;
-+}
-+
-+/* ........................................................................... */
-+
-+static t_Error GracefulStop(t_Dtsec *p_Dtsec, e_CommMode mode)
-+{
-+ struct dtsec_regs *p_MemMap;
-+ int pollTimeout = 0;
-+
-+ ASSERT_COND(p_Dtsec);
-+
-+ p_MemMap = p_Dtsec->p_MemMap;
-+ ASSERT_COND(p_MemMap);
-+
-+ /* Assert the graceful transmit stop bit */
-+ if (mode & e_COMM_MODE_RX)
-+ {
-+ fman_dtsec_stop_rx(p_MemMap);
-+
-+#ifdef FM_GRS_ERRATA_DTSEC_A002
-+ if (p_Dtsec->fmMacControllerDriver.fmRevInfo.majorRev == 2)
-+ XX_UDelay(100);
-+#else /* FM_GRS_ERRATA_DTSEC_A002 */
-+#ifdef FM_GTS_AFTER_DROPPED_FRAME_ERRATA_DTSEC_A004839
-+ XX_UDelay(10);
-+#endif /* FM_GTS_AFTER_DROPPED_FRAME_ERRATA_DTSEC_A004839 */
-+#endif /* FM_GRS_ERRATA_DTSEC_A002 */
-+ }
-+
-+ if (mode & e_COMM_MODE_TX)
-+ {
-+#if defined(FM_GTS_ERRATA_DTSEC_A004)
-+ if (p_Dtsec->fmMacControllerDriver.fmRevInfo.majorRev == 2)
-+ DBG(INFO, ("GTS not supported due to DTSEC_A004 errata."));
-+#else /* not defined(FM_GTS_ERRATA_DTSEC_A004) */
-+
-+ fman_dtsec_stop_tx(p_MemMap);
-+
-+#if defined(FM_GTS_UNDERRUN_ERRATA_DTSEC_A0014) || defined(FM_GTS_AFTER_MAC_ABORTED_FRAME_ERRATA_DTSEC_A0012)
-+ XX_UDelay(10);
-+#endif /* FM_GTS_UNDERRUN_ERRATA_DTSEC_A0014 || FM_GTS_AFTER_MAC_ABORTED_FRAME_ERRATA_DTSEC_A0012 */
-+#endif /* defined(FM_GTS_ERRATA_DTSEC_A004) */
-+ }
-+
-+ /* Poll GRSC/GTSC bits in IEVENT register until both are set */
-+#if defined(FM_GRS_ERRATA_DTSEC_A002) || defined(FM_GTS_ERRATA_DTSEC_A004) || defined(FM_GTS_AFTER_MAC_ABORTED_FRAME_ERRATA_DTSEC_A0012) || defined(FM_GTS_UNDERRUN_ERRATA_DTSEC_A0014) || defined(FM_GTS_AFTER_DROPPED_FRAME_ERRATA_DTSEC_A004839)
-+ XX_UDelay(10);
-+#else
-+ while (fman_dtsec_get_event(p_MemMap, DTSEC_IMASK_GRSCEN | DTSEC_IMASK_GTSCEN) != (DTSEC_IMASK_GRSCEN | DTSEC_IMASK_GTSCEN))
-+ {
-+ if (pollTimeout == 100)
-+ break;
-+ XX_UDelay(1);
-+ pollTimeout++;
-+ }
-+#endif
-+
-+ return E_OK;
-+}
-+
-+/* .............................................................................. */
-+
-+static t_Error GracefulRestart(t_Dtsec *p_Dtsec, e_CommMode mode)
-+{
-+ struct dtsec_regs *p_MemMap;
-+
-+ ASSERT_COND(p_Dtsec);
-+ p_MemMap = p_Dtsec->p_MemMap;
-+ ASSERT_COND(p_MemMap);
-+
-+ /* clear the graceful receive stop bit */
-+ if (mode & e_COMM_MODE_TX)
-+ fman_dtsec_start_tx(p_MemMap);
-+
-+ if (mode & e_COMM_MODE_RX)
-+ fman_dtsec_start_rx(p_MemMap);
-+
-+ return E_OK;
-+}
-+
-+
-+/*****************************************************************************/
-+/* dTSEC Configs modification functions */
-+/*****************************************************************************/
-+
-+/* .............................................................................. */
-+
-+static t_Error DtsecConfigLoopback(t_Handle h_Dtsec, bool newVal)
-+{
-+
-+ t_Dtsec *p_Dtsec = (t_Dtsec *)h_Dtsec;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Dtsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Dtsec->p_DtsecDriverParam, E_INVALID_STATE);
-+
-+ p_Dtsec->p_DtsecDriverParam->loopback = newVal;
-+
-+ return E_OK;
-+}
-+
-+/* .............................................................................. */
-+
-+static t_Error DtsecConfigMaxFrameLength(t_Handle h_Dtsec, uint16_t newVal)
-+{
-+ t_Dtsec *p_Dtsec = (t_Dtsec *)h_Dtsec;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Dtsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Dtsec->p_DtsecDriverParam, E_INVALID_STATE);
-+
-+ p_Dtsec->p_DtsecDriverParam->maximum_frame = newVal;
-+
-+ return E_OK;
-+}
-+
-+/* .............................................................................. */
-+
-+static t_Error DtsecConfigPadAndCrc(t_Handle h_Dtsec, bool newVal)
-+{
-+ t_Dtsec *p_Dtsec = (t_Dtsec *)h_Dtsec;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Dtsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Dtsec->p_DtsecDriverParam, E_INVALID_STATE);
-+
-+ p_Dtsec->p_DtsecDriverParam->tx_pad_crc = newVal;
-+
-+ return E_OK;
-+}
-+
-+/* .............................................................................. */
-+
-+static t_Error DtsecConfigHalfDuplex(t_Handle h_Dtsec, bool newVal)
-+{
-+ t_Dtsec *p_Dtsec = (t_Dtsec *)h_Dtsec;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Dtsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Dtsec->p_DtsecDriverParam, E_INVALID_STATE);
-+
-+ p_Dtsec->p_DtsecDriverParam->halfdup_on = newVal;
-+
-+ return E_OK;
-+}
-+
-+/* .............................................................................. */
-+
-+static t_Error DtsecConfigTbiPhyAddr(t_Handle h_Dtsec, uint8_t newVal)
-+{
-+ t_Dtsec *p_Dtsec = (t_Dtsec *)h_Dtsec;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Dtsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Dtsec->p_DtsecDriverParam, E_INVALID_STATE);
-+
-+ p_Dtsec->p_DtsecDriverParam->tbi_phy_addr = newVal;
-+
-+ return E_OK;
-+}
-+
-+/* .............................................................................. */
-+
-+static t_Error DtsecConfigLengthCheck(t_Handle h_Dtsec, bool newVal)
-+{
-+ t_Dtsec *p_Dtsec = (t_Dtsec *)h_Dtsec;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Dtsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Dtsec->p_DtsecDriverParam, E_INVALID_STATE);
-+
-+ p_Dtsec->p_DtsecDriverParam->rx_len_check = newVal;
-+
-+ return E_OK;
-+}
-+
-+/* .............................................................................. */
-+
-+static t_Error DtsecConfigException(t_Handle h_Dtsec, e_FmMacExceptions exception, bool enable)
-+{
-+ t_Dtsec *p_Dtsec = (t_Dtsec *)h_Dtsec;
-+ uint32_t bitMask = 0;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Dtsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Dtsec->p_DtsecDriverParam, E_INVALID_STATE);
-+
-+ if (exception != e_FM_MAC_EX_1G_1588_TS_RX_ERR)
-+ {
-+ GET_EXCEPTION_FLAG(bitMask, exception);
-+ if (bitMask)
-+ {
-+ if (enable)
-+ p_Dtsec->exceptions |= bitMask;
-+ else
-+ p_Dtsec->exceptions &= ~bitMask;
-+ }
-+ else
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("Undefined exception"));
-+ }
-+ else
-+ {
-+ if (!p_Dtsec->ptpTsuEnabled)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("Exception valid for 1588 only"));
-+
-+ if (enable)
-+ p_Dtsec->enTsuErrExeption = TRUE;
-+ else
-+ p_Dtsec->enTsuErrExeption = FALSE;
-+ }
-+
-+ return E_OK;
-+}
-+
-+
-+/*****************************************************************************/
-+/* dTSEC Run Time API functions */
-+/*****************************************************************************/
-+
-+/* .............................................................................. */
-+
-+static t_Error DtsecEnable(t_Handle h_Dtsec, e_CommMode mode)
-+{
-+ t_Dtsec *p_Dtsec = (t_Dtsec *)h_Dtsec;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Dtsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Dtsec->p_DtsecDriverParam, E_INVALID_STATE);
-+
-+ fman_dtsec_enable(p_Dtsec->p_MemMap,
-+ (bool)!!(mode & e_COMM_MODE_RX),
-+ (bool)!!(mode & e_COMM_MODE_TX));
-+
-+ GracefulRestart(p_Dtsec, mode);
-+
-+ return E_OK;
-+}
-+
-+/* .............................................................................. */
-+
-+static t_Error DtsecDisable (t_Handle h_Dtsec, e_CommMode mode)
-+{
-+ t_Dtsec *p_Dtsec = (t_Dtsec *)h_Dtsec;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Dtsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Dtsec->p_DtsecDriverParam, E_INVALID_STATE);
-+
-+ GracefulStop(p_Dtsec, mode);
-+
-+ fman_dtsec_disable(p_Dtsec->p_MemMap,
-+ (bool)!!(mode & e_COMM_MODE_RX),
-+ (bool)!!(mode & e_COMM_MODE_TX));
-+
-+ return E_OK;
-+}
-+
-+/* .............................................................................. */
-+
-+static t_Error DtsecSetTxPauseFrames(t_Handle h_Dtsec,
-+ uint8_t priority,
-+ uint16_t pauseTime,
-+ uint16_t threshTime)
-+{
-+ t_Dtsec *p_Dtsec = (t_Dtsec *)h_Dtsec;
-+
-+ UNUSED(priority);UNUSED(threshTime);
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Dtsec, E_INVALID_STATE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Dtsec->p_DtsecDriverParam, E_INVALID_STATE);
-+
-+#ifdef FM_BAD_TX_TS_IN_B_2_B_ERRATA_DTSEC_A003
-+ if (p_Dtsec->fmMacControllerDriver.fmRevInfo.majorRev == 2)
-+ if (0 < pauseTime && pauseTime <= 320)
-+ RETURN_ERROR(MINOR, E_INVALID_VALUE,
-+ ("This pause-time value of %d is illegal due to errata dTSEC-A003!"
-+ " value should be greater than 320."));
-+#endif /* FM_BAD_TX_TS_IN_B_2_B_ERRATA_DTSEC_A003 */
-+
-+ GracefulStop(p_Dtsec, e_COMM_MODE_RX_AND_TX);
-+
-+ fman_dtsec_set_tx_pause_frames(p_Dtsec->p_MemMap, pauseTime);
-+
-+ GracefulRestart(p_Dtsec, e_COMM_MODE_RX_AND_TX);
-+
-+ return E_OK;
-+}
-+
-+/* .............................................................................. */
-+/* backward compatibility. will be removed in the future. */
-+static t_Error DtsecTxMacPause(t_Handle h_Dtsec, uint16_t pauseTime)
-+{
-+ return DtsecSetTxPauseFrames(h_Dtsec, 0, pauseTime, 0);
-+}
-+
-+/* .............................................................................. */
-+
-+static t_Error DtsecRxIgnoreMacPause(t_Handle h_Dtsec, bool en)
-+{
-+ t_Dtsec *p_Dtsec = (t_Dtsec *)h_Dtsec;
-+ bool accept_pause = !en;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Dtsec, E_INVALID_STATE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Dtsec->p_DtsecDriverParam, E_INVALID_STATE);
-+
-+ GracefulStop(p_Dtsec, e_COMM_MODE_RX_AND_TX);
-+
-+ fman_dtsec_handle_rx_pause(p_Dtsec->p_MemMap, accept_pause);
-+
-+ GracefulRestart(p_Dtsec, e_COMM_MODE_RX_AND_TX);
-+
-+ return E_OK;
-+}
-+
-+/* .............................................................................. */
-+
-+static t_Error DtsecEnable1588TimeStamp(t_Handle h_Dtsec)
-+{
-+ t_Dtsec *p_Dtsec = (t_Dtsec *)h_Dtsec;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Dtsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Dtsec->p_DtsecDriverParam, E_INVALID_STATE);
-+
-+ p_Dtsec->ptpTsuEnabled = TRUE;
-+ fman_dtsec_set_ts(p_Dtsec->p_MemMap, TRUE);
-+
-+ return E_OK;
-+}
-+
-+/* .............................................................................. */
-+
-+static t_Error DtsecDisable1588TimeStamp(t_Handle h_Dtsec)
-+{
-+ t_Dtsec *p_Dtsec = (t_Dtsec *)h_Dtsec;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Dtsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Dtsec->p_DtsecDriverParam, E_INVALID_STATE);
-+
-+ p_Dtsec->ptpTsuEnabled = FALSE;
-+ fman_dtsec_set_ts(p_Dtsec->p_MemMap, FALSE);
-+
-+ return E_OK;
-+}
-+
-+/* .............................................................................. */
-+
-+static t_Error DtsecGetStatistics(t_Handle h_Dtsec, t_FmMacStatistics *p_Statistics)
-+{
-+ t_Dtsec *p_Dtsec = (t_Dtsec *)h_Dtsec;
-+ struct dtsec_regs *p_DtsecMemMap;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Dtsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Dtsec->p_DtsecDriverParam, E_INVALID_STATE);
-+ SANITY_CHECK_RETURN_ERROR(p_Statistics, E_NULL_POINTER);
-+
-+ p_DtsecMemMap = p_Dtsec->p_MemMap;
-+
-+ if (p_Dtsec->statisticsLevel == e_FM_MAC_NONE_STATISTICS)
-+ RETURN_ERROR(MINOR, E_INVALID_STATE, ("Statistics disabled"));
-+
-+ memset(p_Statistics, 0xff, sizeof(t_FmMacStatistics));
-+
-+ if (p_Dtsec->statisticsLevel == e_FM_MAC_FULL_STATISTICS)
-+ {
-+ p_Statistics->eStatPkts64 = fman_dtsec_get_stat_counter(p_DtsecMemMap, E_DTSEC_STAT_TR64)
-+ + p_Dtsec->internalStatistics.tr64;
-+ p_Statistics->eStatPkts65to127 = fman_dtsec_get_stat_counter(p_DtsecMemMap, E_DTSEC_STAT_TR127)
-+ + p_Dtsec->internalStatistics.tr127;
-+ p_Statistics->eStatPkts128to255 = fman_dtsec_get_stat_counter(p_DtsecMemMap, E_DTSEC_STAT_TR255)
-+ + p_Dtsec->internalStatistics.tr255;
-+ p_Statistics->eStatPkts256to511 = fman_dtsec_get_stat_counter(p_DtsecMemMap, E_DTSEC_STAT_TR511)
-+ + p_Dtsec->internalStatistics.tr511;
-+ p_Statistics->eStatPkts512to1023 = fman_dtsec_get_stat_counter(p_DtsecMemMap, E_DTSEC_STAT_TR1K)
-+ + p_Dtsec->internalStatistics.tr1k;
-+ p_Statistics->eStatPkts1024to1518 = fman_dtsec_get_stat_counter(p_DtsecMemMap, E_DTSEC_STAT_TRMAX)
-+ + p_Dtsec->internalStatistics.trmax;
-+ p_Statistics->eStatPkts1519to1522 = fman_dtsec_get_stat_counter(p_DtsecMemMap, E_DTSEC_STAT_TRMGV)
-+ + p_Dtsec->internalStatistics.trmgv;
-+
-+ /* MIB II */
-+ p_Statistics->ifInOctets = fman_dtsec_get_stat_counter(p_DtsecMemMap, E_DTSEC_STAT_RBYT)
-+ + p_Dtsec->internalStatistics.rbyt;
-+ p_Statistics->ifInPkts = fman_dtsec_get_stat_counter(p_DtsecMemMap, E_DTSEC_STAT_RPKT)
-+ + p_Dtsec->internalStatistics.rpkt;
-+ p_Statistics->ifInUcastPkts = 0;
-+ p_Statistics->ifInMcastPkts = fman_dtsec_get_stat_counter(p_DtsecMemMap, E_DTSEC_STAT_RMCA)
-+ + p_Dtsec->internalStatistics.rmca;
-+ p_Statistics->ifInBcastPkts = fman_dtsec_get_stat_counter(p_DtsecMemMap, E_DTSEC_STAT_RBCA)
-+ + p_Dtsec->internalStatistics.rbca;
-+ p_Statistics->ifOutOctets = fman_dtsec_get_stat_counter(p_DtsecMemMap, E_DTSEC_STAT_TBYT)
-+ + p_Dtsec->internalStatistics.tbyt;
-+ p_Statistics->ifOutPkts = fman_dtsec_get_stat_counter(p_DtsecMemMap, E_DTSEC_STAT_TPKT)
-+ + p_Dtsec->internalStatistics.tpkt;
-+ p_Statistics->ifOutUcastPkts = 0;
-+ p_Statistics->ifOutMcastPkts = fman_dtsec_get_stat_counter(p_DtsecMemMap, E_DTSEC_STAT_TMCA)
-+ + p_Dtsec->internalStatistics.tmca;
-+ p_Statistics->ifOutBcastPkts = fman_dtsec_get_stat_counter(p_DtsecMemMap, E_DTSEC_STAT_TBCA)
-+ + p_Dtsec->internalStatistics.tbca;
-+ }
-+
-+ p_Statistics->eStatFragments = fman_dtsec_get_stat_counter(p_DtsecMemMap, E_DTSEC_STAT_RFRG)
-+ + p_Dtsec->internalStatistics.rfrg;
-+ p_Statistics->eStatJabbers = fman_dtsec_get_stat_counter(p_DtsecMemMap, E_DTSEC_STAT_RJBR)
-+ + p_Dtsec->internalStatistics.rjbr;
-+ p_Statistics->eStatsDropEvents = fman_dtsec_get_stat_counter(p_DtsecMemMap, E_DTSEC_STAT_RDRP)
-+ + p_Dtsec->internalStatistics.rdrp;
-+ p_Statistics->eStatCRCAlignErrors = fman_dtsec_get_stat_counter(p_DtsecMemMap, E_DTSEC_STAT_RALN)
-+ + p_Dtsec->internalStatistics.raln;
-+ p_Statistics->eStatUndersizePkts = fman_dtsec_get_stat_counter(p_DtsecMemMap, E_DTSEC_STAT_RUND)
-+ + p_Dtsec->internalStatistics.rund;
-+ p_Statistics->eStatOversizePkts = fman_dtsec_get_stat_counter(p_DtsecMemMap, E_DTSEC_STAT_ROVR)
-+ + p_Dtsec->internalStatistics.rovr;
-+ p_Statistics->reStatPause = fman_dtsec_get_stat_counter(p_DtsecMemMap, E_DTSEC_STAT_RXPF)
-+ + p_Dtsec->internalStatistics.rxpf;
-+ p_Statistics->teStatPause = fman_dtsec_get_stat_counter(p_DtsecMemMap, E_DTSEC_STAT_TXPF)
-+ + p_Dtsec->internalStatistics.txpf;
-+ p_Statistics->ifInDiscards = p_Statistics->eStatsDropEvents;
-+ p_Statistics->ifInErrors = p_Statistics->eStatsDropEvents + p_Statistics->eStatCRCAlignErrors
-+ + fman_dtsec_get_stat_counter(p_DtsecMemMap,E_DTSEC_STAT_RFLR) + p_Dtsec->internalStatistics.rflr
-+ + fman_dtsec_get_stat_counter(p_DtsecMemMap,E_DTSEC_STAT_RCDE) + p_Dtsec->internalStatistics.rcde
-+ + fman_dtsec_get_stat_counter(p_DtsecMemMap,E_DTSEC_STAT_RCSE) + p_Dtsec->internalStatistics.rcse;
-+
-+ p_Statistics->ifOutDiscards = fman_dtsec_get_stat_counter(p_DtsecMemMap, E_DTSEC_STAT_TDRP)
-+ + p_Dtsec->internalStatistics.tdrp;
-+ p_Statistics->ifOutErrors = p_Statistics->ifOutDiscards /**< Number of frames transmitted with error: */
-+ + fman_dtsec_get_stat_counter(p_DtsecMemMap,E_DTSEC_STAT_TFCS)
-+ + p_Dtsec->internalStatistics.tfcs;
-+
-+ return E_OK;
-+}
-+
-+/* .............................................................................. */
-+
-+static t_Error DtsecModifyMacAddress (t_Handle h_Dtsec, t_EnetAddr *p_EnetAddr)
-+{
-+ t_Dtsec *p_Dtsec = (t_Dtsec *)h_Dtsec;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Dtsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Dtsec->p_DtsecDriverParam, E_INVALID_STATE);
-+
-+ /* Initialize MAC Station Address registers (1 & 2) */
-+ /* Station address have to be swapped (big endian to little endian */
-+ p_Dtsec->addr = ENET_ADDR_TO_UINT64(*p_EnetAddr);
-+
-+ GracefulStop(p_Dtsec, e_COMM_MODE_RX_AND_TX);
-+
-+ fman_dtsec_set_mac_address(p_Dtsec->p_MemMap, (uint8_t *)(*p_EnetAddr));
-+
-+ GracefulRestart(p_Dtsec, e_COMM_MODE_RX_AND_TX);
-+
-+ return E_OK;
-+}
-+
-+/* .............................................................................. */
-+
-+static t_Error DtsecResetCounters (t_Handle h_Dtsec)
-+{
-+ t_Dtsec *p_Dtsec = (t_Dtsec *)h_Dtsec;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Dtsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Dtsec->p_DtsecDriverParam, E_INVALID_STATE);
-+
-+ /* clear HW counters */
-+ fman_dtsec_reset_stat(p_Dtsec->p_MemMap);
-+
-+ /* clear SW counters holding carries */
-+ memset(&p_Dtsec->internalStatistics, 0, sizeof(t_InternalStatistics));
-+
-+ return E_OK;
-+}
-+
-+/* .............................................................................. */
-+
-+static t_Error DtsecAddExactMatchMacAddress(t_Handle h_Dtsec, t_EnetAddr *p_EthAddr)
-+{
-+ t_Dtsec *p_Dtsec = (t_Dtsec *) h_Dtsec;
-+ uint64_t ethAddr;
-+ uint8_t paddrNum;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Dtsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Dtsec->p_DtsecDriverParam, E_INVALID_STATE);
-+
-+ ethAddr = ENET_ADDR_TO_UINT64(*p_EthAddr);
-+
-+ if (ethAddr & GROUP_ADDRESS)
-+ /* Multicast address has no effect in PADDR */
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("Multicast address"));
-+
-+ /* Make sure no PADDR contains this address */
-+ for (paddrNum = 0; paddrNum < DTSEC_NUM_OF_PADDRS; paddrNum++)
-+ if (p_Dtsec->indAddrRegUsed[paddrNum])
-+ if (p_Dtsec->paddr[paddrNum] == ethAddr)
-+ RETURN_ERROR(MAJOR, E_ALREADY_EXISTS, NO_MSG);
-+
-+ /* Find first unused PADDR */
-+ for (paddrNum = 0; paddrNum < DTSEC_NUM_OF_PADDRS; paddrNum++)
-+ if (!(p_Dtsec->indAddrRegUsed[paddrNum]))
-+ {
-+ /* mark this PADDR as used */
-+ p_Dtsec->indAddrRegUsed[paddrNum] = TRUE;
-+ /* store address */
-+ p_Dtsec->paddr[paddrNum] = ethAddr;
-+
-+ /* put in hardware */
-+ fman_dtsec_add_addr_in_paddr(p_Dtsec->p_MemMap, (uint64_t)PTR_TO_UINT(&ethAddr), paddrNum);
-+ p_Dtsec->numOfIndAddrInRegs++;
-+
-+ return E_OK;
-+ }
-+
-+ /* No free PADDR */
-+ RETURN_ERROR(MAJOR, E_FULL, NO_MSG);
-+}
-+
-+/* .............................................................................. */
-+
-+static t_Error DtsecDelExactMatchMacAddress(t_Handle h_Dtsec, t_EnetAddr *p_EthAddr)
-+{
-+ t_Dtsec *p_Dtsec = (t_Dtsec *) h_Dtsec;
-+ uint64_t ethAddr;
-+ uint8_t paddrNum;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Dtsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Dtsec->p_DtsecDriverParam, E_INVALID_STATE);
-+
-+ ethAddr = ENET_ADDR_TO_UINT64(*p_EthAddr);
-+
-+ /* Find used PADDR containing this address */
-+ for (paddrNum = 0; paddrNum < DTSEC_NUM_OF_PADDRS; paddrNum++)
-+ {
-+ if ((p_Dtsec->indAddrRegUsed[paddrNum]) &&
-+ (p_Dtsec->paddr[paddrNum] == ethAddr))
-+ {
-+ /* mark this PADDR as not used */
-+ p_Dtsec->indAddrRegUsed[paddrNum] = FALSE;
-+ /* clear in hardware */
-+ fman_dtsec_clear_addr_in_paddr(p_Dtsec->p_MemMap, paddrNum);
-+ p_Dtsec->numOfIndAddrInRegs--;
-+
-+ return E_OK;
-+ }
-+ }
-+
-+ RETURN_ERROR(MAJOR, E_NOT_FOUND, NO_MSG);
-+}
-+
-+/* .............................................................................. */
-+
-+static t_Error DtsecAddHashMacAddress(t_Handle h_Dtsec, t_EnetAddr *p_EthAddr)
-+{
-+ t_Dtsec *p_Dtsec = (t_Dtsec *)h_Dtsec;
-+ t_EthHashEntry *p_HashEntry;
-+ uint64_t ethAddr;
-+ int32_t bucket;
-+ uint32_t crc;
-+ bool mcast, ghtx;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Dtsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Dtsec->p_DtsecDriverParam, E_INVALID_STATE);
-+
-+ ethAddr = ENET_ADDR_TO_UINT64(*p_EthAddr);
-+
-+ ghtx = (bool)((fman_dtsec_get_rctrl(p_Dtsec->p_MemMap) & RCTRL_GHTX) ? TRUE : FALSE);
-+ mcast = (bool)((ethAddr & MAC_GROUP_ADDRESS) ? TRUE : FALSE);
-+
-+ if (ghtx && !mcast) /* Cannot handle unicast mac addr when GHTX is on */
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("Could not compute hash bucket"));
-+
-+ crc = GetMacAddrHashCode(ethAddr);
-+
-+ /* considering the 9 highest order bits in crc H[8:0]:
-+ * if ghtx = 0 H[8:6] (highest order 3 bits) identify the hash register
-+ * and H[5:1] (next 5 bits) identify the hash bit
-+ * if ghts = 1 H[8:5] (highest order 4 bits) identify the hash register
-+ * and H[4:0] (next 5 bits) identify the hash bit.
-+ *
-+ * In bucket index output the low 5 bits identify the hash register bit,
-+ * while the higher 4 bits identify the hash register
-+ */
-+
-+ if (ghtx)
-+ bucket = (int32_t)((crc >> 23) & 0x1ff);
-+ else {
-+ bucket = (int32_t)((crc >> 24) & 0xff);
-+ /* if !ghtx and mcast the bit must be set in gaddr instead of igaddr. */
-+ if (mcast)
-+ bucket += 0x100;
-+ }
-+
-+ fman_dtsec_set_bucket(p_Dtsec->p_MemMap, bucket, TRUE);
-+
-+ /* Create element to be added to the driver hash table */
-+ p_HashEntry = (t_EthHashEntry *)XX_Malloc(sizeof(t_EthHashEntry));
-+ p_HashEntry->addr = ethAddr;
-+ INIT_LIST(&p_HashEntry->node);
-+
-+ if (ethAddr & MAC_GROUP_ADDRESS)
-+ /* Group Address */
-+ LIST_AddToTail(&(p_HashEntry->node), &(p_Dtsec->p_MulticastAddrHash->p_Lsts[bucket]));
-+ else
-+ LIST_AddToTail(&(p_HashEntry->node), &(p_Dtsec->p_UnicastAddrHash->p_Lsts[bucket]));
-+
-+ return E_OK;
-+}
-+
-+/* .............................................................................. */
-+
-+static t_Error DtsecDelHashMacAddress(t_Handle h_Dtsec, t_EnetAddr *p_EthAddr)
-+{
-+ t_Dtsec *p_Dtsec = (t_Dtsec *)h_Dtsec;
-+ t_List *p_Pos;
-+ t_EthHashEntry *p_HashEntry = NULL;
-+ uint64_t ethAddr;
-+ int32_t bucket;
-+ uint32_t crc;
-+ bool mcast, ghtx;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Dtsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Dtsec->p_DtsecDriverParam, E_INVALID_STATE);
-+
-+ ethAddr = ENET_ADDR_TO_UINT64(*p_EthAddr);
-+
-+ ghtx = (bool)((fman_dtsec_get_rctrl(p_Dtsec->p_MemMap) & RCTRL_GHTX) ? TRUE : FALSE);
-+ mcast = (bool)((ethAddr & MAC_GROUP_ADDRESS) ? TRUE : FALSE);
-+
-+ if (ghtx && !mcast) /* Cannot handle unicast mac addr when GHTX is on */
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("Could not compute hash bucket"));
-+
-+ crc = GetMacAddrHashCode(ethAddr);
-+
-+ if (ghtx)
-+ bucket = (int32_t)((crc >> 23) & 0x1ff);
-+ else {
-+ bucket = (int32_t)((crc >> 24) & 0xff);
-+ /* if !ghtx and mcast the bit must be set in gaddr instead of igaddr. */
-+ if (mcast)
-+ bucket += 0x100;
-+ }
-+
-+ if (ethAddr & MAC_GROUP_ADDRESS)
-+ {
-+ /* Group Address */
-+ LIST_FOR_EACH(p_Pos, &(p_Dtsec->p_MulticastAddrHash->p_Lsts[bucket]))
-+ {
-+ p_HashEntry = ETH_HASH_ENTRY_OBJ(p_Pos);
-+ if (p_HashEntry->addr == ethAddr)
-+ {
-+ LIST_DelAndInit(&p_HashEntry->node);
-+ XX_Free(p_HashEntry);
-+ break;
-+ }
-+ }
-+ if (LIST_IsEmpty(&p_Dtsec->p_MulticastAddrHash->p_Lsts[bucket]))
-+ fman_dtsec_set_bucket(p_Dtsec->p_MemMap, bucket, FALSE);
-+ }
-+ else
-+ {
-+ /* Individual Address */
-+ LIST_FOR_EACH(p_Pos, &(p_Dtsec->p_UnicastAddrHash->p_Lsts[bucket]))
-+ {
-+ p_HashEntry = ETH_HASH_ENTRY_OBJ(p_Pos);
-+ if (p_HashEntry->addr == ethAddr)
-+ {
-+ LIST_DelAndInit(&p_HashEntry->node);
-+ XX_Free(p_HashEntry);
-+ break;
-+ }
-+ }
-+ if (LIST_IsEmpty(&p_Dtsec->p_UnicastAddrHash->p_Lsts[bucket]))
-+ fman_dtsec_set_bucket(p_Dtsec->p_MemMap, bucket, FALSE);
-+ }
-+
-+ /* address does not exist */
-+ ASSERT_COND(p_HashEntry != NULL);
-+
-+ return E_OK;
-+}
-+
-+/* .............................................................................. */
-+
-+static t_Error DtsecSetPromiscuous(t_Handle h_Dtsec, bool newVal)
-+{
-+ t_Dtsec *p_Dtsec = (t_Dtsec *)h_Dtsec;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Dtsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Dtsec->p_DtsecDriverParam, E_INVALID_STATE);
-+
-+ fman_dtsec_set_uc_promisc(p_Dtsec->p_MemMap, newVal);
-+ fman_dtsec_set_mc_promisc(p_Dtsec->p_MemMap, newVal);
-+
-+ return E_OK;
-+}
-+
-+/* .............................................................................. */
-+
-+static t_Error DtsecSetStatistics(t_Handle h_Dtsec, e_FmMacStatisticsLevel statisticsLevel)
-+{
-+ t_Dtsec *p_Dtsec = (t_Dtsec *)h_Dtsec;
-+ t_Error err;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Dtsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Dtsec->p_DtsecDriverParam, E_INVALID_STATE);
-+
-+ p_Dtsec->statisticsLevel = statisticsLevel;
-+
-+ err = (t_Error)fman_dtsec_set_stat_level(p_Dtsec->p_MemMap,
-+ (enum dtsec_stat_level)statisticsLevel);
-+ if (err != E_OK)
-+ return err;
-+
-+ switch (statisticsLevel)
-+ {
-+ case (e_FM_MAC_NONE_STATISTICS):
-+ p_Dtsec->exceptions &= ~DTSEC_IMASK_MSROEN;
-+ break;
-+ case (e_FM_MAC_PARTIAL_STATISTICS):
-+ p_Dtsec->exceptions |= DTSEC_IMASK_MSROEN;
-+ break;
-+ case (e_FM_MAC_FULL_STATISTICS):
-+ p_Dtsec->exceptions |= DTSEC_IMASK_MSROEN;
-+ break;
-+ default:
-+ RETURN_ERROR(MINOR, E_INVALID_SELECTION, NO_MSG);
-+ }
-+
-+ return E_OK;
-+}
-+
-+/* .............................................................................. */
-+
-+static t_Error DtsecSetWakeOnLan(t_Handle h_Dtsec, bool en)
-+{
-+ t_Dtsec *p_Dtsec = (t_Dtsec *)h_Dtsec;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Dtsec, E_INVALID_STATE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Dtsec->p_DtsecDriverParam, E_INVALID_STATE);
-+
-+ GracefulStop(p_Dtsec, e_COMM_MODE_RX_AND_TX);
-+
-+ fman_dtsec_set_wol(p_Dtsec->p_MemMap, en);
-+
-+ GracefulRestart(p_Dtsec, e_COMM_MODE_RX_AND_TX);
-+
-+ return E_OK;
-+}
-+
-+/* .............................................................................. */
-+
-+static t_Error DtsecAdjustLink(t_Handle h_Dtsec, e_EnetSpeed speed, bool fullDuplex)
-+{
-+ t_Dtsec *p_Dtsec = (t_Dtsec *)h_Dtsec;
-+ int err;
-+ enum enet_interface enet_interface;
-+ enum enet_speed enet_speed;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Dtsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Dtsec->p_DtsecDriverParam, E_INVALID_STATE);
-+
-+ p_Dtsec->enetMode = MAKE_ENET_MODE(ENET_INTERFACE_FROM_MODE(p_Dtsec->enetMode), speed);
-+ enet_interface = (enum enet_interface) ENET_INTERFACE_FROM_MODE(p_Dtsec->enetMode);
-+ enet_speed = (enum enet_speed) ENET_SPEED_FROM_MODE(p_Dtsec->enetMode);
-+ p_Dtsec->halfDuplex = !fullDuplex;
-+
-+ GracefulStop(p_Dtsec, e_COMM_MODE_RX_AND_TX);
-+
-+ err = fman_dtsec_adjust_link(p_Dtsec->p_MemMap, enet_interface, enet_speed, fullDuplex);
-+
-+ if (err == -EINVAL)
-+ RETURN_ERROR(MAJOR, E_CONFLICT, ("Ethernet interface does not support Half Duplex mode"));
-+
-+ GracefulRestart(p_Dtsec, e_COMM_MODE_RX_AND_TX);
-+
-+ return (t_Error)err;
-+}
-+
-+/* .............................................................................. */
-+
-+static t_Error DtsecRestartAutoneg(t_Handle h_Dtsec)
-+{
-+ t_Dtsec *p_Dtsec = (t_Dtsec *)h_Dtsec;
-+ uint16_t tmpReg16;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Dtsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Dtsec->p_DtsecDriverParam, E_INVALID_STATE);
-+
-+ DTSEC_MII_ReadPhyReg(p_Dtsec, p_Dtsec->tbi_phy_addr, 0, &tmpReg16);
-+
-+ tmpReg16 &= ~( PHY_CR_SPEED0 | PHY_CR_SPEED1 );
-+ tmpReg16 |= (PHY_CR_ANE | PHY_CR_RESET_AN | PHY_CR_FULLDUPLEX | PHY_CR_SPEED1);
-+
-+ DTSEC_MII_WritePhyReg(p_Dtsec, p_Dtsec->tbi_phy_addr, 0, tmpReg16);
-+
-+ return E_OK;
-+}
-+
-+/* .............................................................................. */
-+
-+static t_Error DtsecGetId(t_Handle h_Dtsec, uint32_t *macId)
-+{
-+ t_Dtsec *p_Dtsec = (t_Dtsec *)h_Dtsec;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Dtsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Dtsec->p_DtsecDriverParam, E_INVALID_STATE);
-+
-+ *macId = p_Dtsec->macId;
-+
-+ return E_OK;
-+}
-+
-+/* .............................................................................. */
-+
-+static t_Error DtsecGetVersion(t_Handle h_Dtsec, uint32_t *macVersion)
-+{
-+ t_Dtsec *p_Dtsec = (t_Dtsec *)h_Dtsec;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Dtsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Dtsec->p_DtsecDriverParam, E_INVALID_STATE);
-+
-+ *macVersion = fman_dtsec_get_revision(p_Dtsec->p_MemMap);
-+
-+ return E_OK;
-+}
-+
-+/* .............................................................................. */
-+
-+static t_Error DtsecSetException(t_Handle h_Dtsec, e_FmMacExceptions exception, bool enable)
-+{
-+ t_Dtsec *p_Dtsec = (t_Dtsec *)h_Dtsec;
-+ uint32_t bitMask = 0;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Dtsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Dtsec->p_DtsecDriverParam, E_INVALID_STATE);
-+
-+ if (exception != e_FM_MAC_EX_1G_1588_TS_RX_ERR)
-+ {
-+ GET_EXCEPTION_FLAG(bitMask, exception);
-+ if (bitMask)
-+ {
-+ if (enable)
-+ p_Dtsec->exceptions |= bitMask;
-+ else
-+ p_Dtsec->exceptions &= ~bitMask;
-+ }
-+ else
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("Undefined exception"));
-+
-+ if (enable)
-+ fman_dtsec_enable_interrupt(p_Dtsec->p_MemMap, bitMask);
-+ else
-+ fman_dtsec_disable_interrupt(p_Dtsec->p_MemMap, bitMask);
-+ }
-+ else
-+ {
-+ if (!p_Dtsec->ptpTsuEnabled)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("Exception valid for 1588 only"));
-+
-+ if (enable)
-+ {
-+ p_Dtsec->enTsuErrExeption = TRUE;
-+ fman_dtsec_enable_tmr_interrupt(p_Dtsec->p_MemMap);
-+ }
-+ else
-+ {
-+ p_Dtsec->enTsuErrExeption = FALSE;
-+ fman_dtsec_disable_tmr_interrupt(p_Dtsec->p_MemMap);
-+ }
-+ }
-+
-+ return E_OK;
-+}
-+
-+
-+/*****************************************************************************/
-+/* dTSEC Init & Free API */
-+/*****************************************************************************/
-+
-+/* .............................................................................. */
-+
-+static t_Error DtsecInit(t_Handle h_Dtsec)
-+{
-+ t_Dtsec *p_Dtsec = (t_Dtsec *)h_Dtsec;
-+ struct dtsec_cfg *p_DtsecDriverParam;
-+ t_Error err;
-+ uint16_t maxFrmLn;
-+ enum enet_interface enet_interface;
-+ enum enet_speed enet_speed;
-+ t_EnetAddr ethAddr;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Dtsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Dtsec->p_DtsecDriverParam, E_INVALID_STATE);
-+ SANITY_CHECK_RETURN_ERROR(p_Dtsec->fmMacControllerDriver.h_Fm, E_INVALID_HANDLE);
-+
-+ FM_GetRevision(p_Dtsec->fmMacControllerDriver.h_Fm, &p_Dtsec->fmMacControllerDriver.fmRevInfo);
-+ CHECK_INIT_PARAMETERS(p_Dtsec, CheckInitParameters);
-+
-+ p_DtsecDriverParam = p_Dtsec->p_DtsecDriverParam;
-+ p_Dtsec->halfDuplex = p_DtsecDriverParam->halfdup_on;
-+
-+ enet_interface = (enum enet_interface)ENET_INTERFACE_FROM_MODE(p_Dtsec->enetMode);
-+ enet_speed = (enum enet_speed)ENET_SPEED_FROM_MODE(p_Dtsec->enetMode);
-+ MAKE_ENET_ADDR_FROM_UINT64(p_Dtsec->addr, ethAddr);
-+
-+ err = (t_Error)fman_dtsec_init(p_Dtsec->p_MemMap,
-+ p_DtsecDriverParam,
-+ enet_interface,
-+ enet_speed,
-+ (uint8_t*)ethAddr,
-+ p_Dtsec->fmMacControllerDriver.fmRevInfo.majorRev,
-+ p_Dtsec->fmMacControllerDriver.fmRevInfo.minorRev,
-+ p_Dtsec->exceptions);
-+ if (err)
-+ {
-+ FreeInitResources(p_Dtsec);
-+ RETURN_ERROR(MAJOR, err, ("This DTSEC version does not support the required i/f mode"));
-+ }
-+
-+ if (ENET_INTERFACE_FROM_MODE(p_Dtsec->enetMode) == e_ENET_IF_SGMII)
-+ {
-+ uint16_t tmpReg16;
-+
-+ /* Configure the TBI PHY Control Register */
-+ tmpReg16 = PHY_TBICON_CLK_SEL | PHY_TBICON_SRESET;
-+ DTSEC_MII_WritePhyReg(p_Dtsec, (uint8_t)p_DtsecDriverParam->tbipa, 17, tmpReg16);
-+
-+ tmpReg16 = PHY_TBICON_CLK_SEL;
-+ DTSEC_MII_WritePhyReg(p_Dtsec, (uint8_t)p_DtsecDriverParam->tbipa, 17, tmpReg16);
-+
-+ tmpReg16 = (PHY_CR_PHY_RESET | PHY_CR_ANE | PHY_CR_FULLDUPLEX | PHY_CR_SPEED1);
-+ DTSEC_MII_WritePhyReg(p_Dtsec, (uint8_t)p_DtsecDriverParam->tbipa, 0, tmpReg16);
-+
-+ if (p_Dtsec->enetMode & ENET_IF_SGMII_BASEX)
-+ tmpReg16 = PHY_TBIANA_1000X;
-+ else
-+ tmpReg16 = PHY_TBIANA_SGMII;
-+ DTSEC_MII_WritePhyReg(p_Dtsec, (uint8_t)p_DtsecDriverParam->tbipa, 4, tmpReg16);
-+
-+ tmpReg16 = (PHY_CR_ANE | PHY_CR_RESET_AN | PHY_CR_FULLDUPLEX | PHY_CR_SPEED1);
-+
-+ DTSEC_MII_WritePhyReg(p_Dtsec, (uint8_t)p_DtsecDriverParam->tbipa, 0, tmpReg16);
-+ }
-+
-+ /* Max Frame Length */
-+ maxFrmLn = fman_dtsec_get_max_frame_len(p_Dtsec->p_MemMap);
-+ err = FmSetMacMaxFrame(p_Dtsec->fmMacControllerDriver.h_Fm, e_FM_MAC_1G,
-+ p_Dtsec->fmMacControllerDriver.macId, maxFrmLn);
-+ if (err)
-+ RETURN_ERROR(MINOR,err, NO_MSG);
-+
-+ p_Dtsec->p_MulticastAddrHash = AllocHashTable(EXTENDED_HASH_TABLE_SIZE);
-+ if (!p_Dtsec->p_MulticastAddrHash) {
-+ FreeInitResources(p_Dtsec);
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, ("MC hash table is FAILED"));
-+ }
-+
-+ p_Dtsec->p_UnicastAddrHash = AllocHashTable(HASH_TABLE_SIZE);
-+ if (!p_Dtsec->p_UnicastAddrHash)
-+ {
-+ FreeInitResources(p_Dtsec);
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, ("UC hash table is FAILED"));
-+ }
-+
-+ /* register err intr handler for dtsec to FPM (err)*/
-+ FmRegisterIntr(p_Dtsec->fmMacControllerDriver.h_Fm,
-+ e_FM_MOD_1G_MAC,
-+ p_Dtsec->macId,
-+ e_FM_INTR_TYPE_ERR,
-+ DtsecIsr,
-+ p_Dtsec);
-+ /* register 1588 intr handler for TMR to FPM (normal)*/
-+ FmRegisterIntr(p_Dtsec->fmMacControllerDriver.h_Fm,
-+ e_FM_MOD_1G_MAC,
-+ p_Dtsec->macId,
-+ e_FM_INTR_TYPE_NORMAL,
-+ Dtsec1588Isr,
-+ p_Dtsec);
-+ /* register normal intr handler for dtsec to main interrupt controller. */
-+ if (p_Dtsec->mdioIrq != NO_IRQ)
-+ {
-+ XX_SetIntr(p_Dtsec->mdioIrq, DtsecMdioIsr, p_Dtsec);
-+ XX_EnableIntr(p_Dtsec->mdioIrq);
-+ }
-+
-+ XX_Free(p_DtsecDriverParam);
-+ p_Dtsec->p_DtsecDriverParam = NULL;
-+
-+ err = DtsecSetStatistics(h_Dtsec, e_FM_MAC_FULL_STATISTICS);
-+ if (err)
-+ {
-+ FreeInitResources(p_Dtsec);
-+ RETURN_ERROR(MAJOR, err, ("Undefined statistics level"));
-+ }
-+
-+ return E_OK;
-+}
-+
-+/* ........................................................................... */
-+
-+static t_Error DtsecFree(t_Handle h_Dtsec)
-+{
-+ t_Dtsec *p_Dtsec = (t_Dtsec *)h_Dtsec;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Dtsec, E_INVALID_HANDLE);
-+
-+ if (p_Dtsec->p_DtsecDriverParam)
-+ {
-+ /* Called after config */
-+ XX_Free(p_Dtsec->p_DtsecDriverParam);
-+ p_Dtsec->p_DtsecDriverParam = NULL;
-+ }
-+ else
-+ /* Called after init */
-+ FreeInitResources(p_Dtsec);
-+
-+ XX_Free(p_Dtsec);
-+
-+ return E_OK;
-+}
-+
-+/* .............................................................................. */
-+
-+static void InitFmMacControllerDriver(t_FmMacControllerDriver *p_FmMacControllerDriver)
-+{
-+ p_FmMacControllerDriver->f_FM_MAC_Init = DtsecInit;
-+ p_FmMacControllerDriver->f_FM_MAC_Free = DtsecFree;
-+
-+ p_FmMacControllerDriver->f_FM_MAC_SetStatistics = DtsecSetStatistics;
-+ p_FmMacControllerDriver->f_FM_MAC_ConfigLoopback = DtsecConfigLoopback;
-+ p_FmMacControllerDriver->f_FM_MAC_ConfigMaxFrameLength = DtsecConfigMaxFrameLength;
-+
-+ p_FmMacControllerDriver->f_FM_MAC_ConfigWan = NULL; /* Not supported on dTSEC */
-+
-+ p_FmMacControllerDriver->f_FM_MAC_ConfigPadAndCrc = DtsecConfigPadAndCrc;
-+ p_FmMacControllerDriver->f_FM_MAC_ConfigHalfDuplex = DtsecConfigHalfDuplex;
-+ p_FmMacControllerDriver->f_FM_MAC_ConfigLengthCheck = DtsecConfigLengthCheck;
-+ p_FmMacControllerDriver->f_FM_MAC_ConfigTbiPhyAddr = DtsecConfigTbiPhyAddr;
-+ p_FmMacControllerDriver->f_FM_MAC_ConfigException = DtsecConfigException;
-+ p_FmMacControllerDriver->f_FM_MAC_ConfigResetOnInit = NULL;
-+
-+ p_FmMacControllerDriver->f_FM_MAC_Enable = DtsecEnable;
-+ p_FmMacControllerDriver->f_FM_MAC_Disable = DtsecDisable;
-+ p_FmMacControllerDriver->f_FM_MAC_Resume = NULL;
-+
-+ p_FmMacControllerDriver->f_FM_MAC_SetException = DtsecSetException;
-+
-+ p_FmMacControllerDriver->f_FM_MAC_SetPromiscuous = DtsecSetPromiscuous;
-+ p_FmMacControllerDriver->f_FM_MAC_AdjustLink = DtsecAdjustLink;
-+ p_FmMacControllerDriver->f_FM_MAC_SetWakeOnLan = DtsecSetWakeOnLan;
-+ p_FmMacControllerDriver->f_FM_MAC_RestartAutoneg = DtsecRestartAutoneg;
-+
-+ p_FmMacControllerDriver->f_FM_MAC_Enable1588TimeStamp = DtsecEnable1588TimeStamp;
-+ p_FmMacControllerDriver->f_FM_MAC_Disable1588TimeStamp = DtsecDisable1588TimeStamp;
-+
-+ p_FmMacControllerDriver->f_FM_MAC_SetTxAutoPauseFrames = DtsecTxMacPause;
-+ p_FmMacControllerDriver->f_FM_MAC_SetTxPauseFrames = DtsecSetTxPauseFrames;
-+ p_FmMacControllerDriver->f_FM_MAC_SetRxIgnorePauseFrames = DtsecRxIgnoreMacPause;
-+
-+ p_FmMacControllerDriver->f_FM_MAC_ResetCounters = DtsecResetCounters;
-+ p_FmMacControllerDriver->f_FM_MAC_GetStatistics = DtsecGetStatistics;
-+ p_FmMacControllerDriver->f_FM_MAC_GetFrameSizeCounters = NULL;
-+
-+ p_FmMacControllerDriver->f_FM_MAC_ModifyMacAddr = DtsecModifyMacAddress;
-+ p_FmMacControllerDriver->f_FM_MAC_AddHashMacAddr = DtsecAddHashMacAddress;
-+ p_FmMacControllerDriver->f_FM_MAC_RemoveHashMacAddr = DtsecDelHashMacAddress;
-+ p_FmMacControllerDriver->f_FM_MAC_AddExactMatchMacAddr = DtsecAddExactMatchMacAddress;
-+ p_FmMacControllerDriver->f_FM_MAC_RemovelExactMatchMacAddr = DtsecDelExactMatchMacAddress;
-+ p_FmMacControllerDriver->f_FM_MAC_GetId = DtsecGetId;
-+ p_FmMacControllerDriver->f_FM_MAC_GetVersion = DtsecGetVersion;
-+ p_FmMacControllerDriver->f_FM_MAC_GetMaxFrameLength = DtsecGetMaxFrameLength;
-+
-+ p_FmMacControllerDriver->f_FM_MAC_MII_WritePhyReg = DTSEC_MII_WritePhyReg;
-+ p_FmMacControllerDriver->f_FM_MAC_MII_ReadPhyReg = DTSEC_MII_ReadPhyReg;
-+
-+}
-+
-+
-+/*****************************************************************************/
-+/* dTSEC Config Main Entry */
-+/*****************************************************************************/
-+
-+/* .............................................................................. */
-+
-+t_Handle DTSEC_Config(t_FmMacParams *p_FmMacParam)
-+{
-+ t_Dtsec *p_Dtsec;
-+ struct dtsec_cfg *p_DtsecDriverParam;
-+ uintptr_t baseAddr;
-+
-+ SANITY_CHECK_RETURN_VALUE(p_FmMacParam, E_NULL_POINTER, NULL);
-+
-+ baseAddr = p_FmMacParam->baseAddr;
-+
-+ /* allocate memory for the UCC GETH data structure. */
-+ p_Dtsec = (t_Dtsec *)XX_Malloc(sizeof(t_Dtsec));
-+ if (!p_Dtsec)
-+ {
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("dTSEC driver structure"));
-+ return NULL;
-+ }
-+ memset(p_Dtsec, 0, sizeof(t_Dtsec));
-+ InitFmMacControllerDriver(&p_Dtsec->fmMacControllerDriver);
-+
-+ /* allocate memory for the dTSEC driver parameters data structure. */
-+ p_DtsecDriverParam = (struct dtsec_cfg *) XX_Malloc(sizeof(struct dtsec_cfg));
-+ if (!p_DtsecDriverParam)
-+ {
-+ XX_Free(p_Dtsec);
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("dTSEC driver parameters"));
-+ return NULL;
-+ }
-+ memset(p_DtsecDriverParam, 0, sizeof(struct dtsec_cfg));
-+
-+ /* Plant parameter structure pointer */
-+ p_Dtsec->p_DtsecDriverParam = p_DtsecDriverParam;
-+
-+ fman_dtsec_defconfig(p_DtsecDriverParam);
-+
-+ p_Dtsec->p_MemMap = (struct dtsec_regs *)UINT_TO_PTR(baseAddr);
-+ p_Dtsec->p_MiiMemMap = (struct dtsec_mii_reg *)UINT_TO_PTR(baseAddr + DTSEC_TO_MII_OFFSET);
-+ p_Dtsec->addr = ENET_ADDR_TO_UINT64(p_FmMacParam->addr);
-+ p_Dtsec->enetMode = p_FmMacParam->enetMode;
-+ p_Dtsec->macId = p_FmMacParam->macId;
-+ p_Dtsec->exceptions = DEFAULT_exceptions;
-+ p_Dtsec->mdioIrq = p_FmMacParam->mdioIrq;
-+ p_Dtsec->f_Exception = p_FmMacParam->f_Exception;
-+ p_Dtsec->f_Event = p_FmMacParam->f_Event;
-+ p_Dtsec->h_App = p_FmMacParam->h_App;
-+ p_Dtsec->ptpTsuEnabled = p_Dtsec->p_DtsecDriverParam->ptp_tsu_en;
-+ p_Dtsec->enTsuErrExeption = p_Dtsec->p_DtsecDriverParam->ptp_exception_en;
-+ p_Dtsec->tbi_phy_addr = p_Dtsec->p_DtsecDriverParam->tbi_phy_addr;
-+
-+ return p_Dtsec;
-+}
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MAC/dtsec.h
-@@ -0,0 +1,228 @@
-+/*
-+ * Copyright 2008-2013 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+/******************************************************************************
-+ @File dtsec.h
-+
-+ @Description FM dTSEC ...
-+*//***************************************************************************/
-+#ifndef __DTSEC_H
-+#define __DTSEC_H
-+
-+#include "std_ext.h"
-+#include "error_ext.h"
-+#include "list_ext.h"
-+#include "enet_ext.h"
-+
-+#include "dtsec_mii_acc.h"
-+#include "fm_mac.h"
-+
-+
-+#define DEFAULT_exceptions \
-+ ((uint32_t)(DTSEC_IMASK_BREN | \
-+ DTSEC_IMASK_RXCEN | \
-+ DTSEC_IMASK_BTEN | \
-+ DTSEC_IMASK_TXCEN | \
-+ DTSEC_IMASK_TXEEN | \
-+ DTSEC_IMASK_ABRTEN | \
-+ DTSEC_IMASK_LCEN | \
-+ DTSEC_IMASK_CRLEN | \
-+ DTSEC_IMASK_XFUNEN | \
-+ DTSEC_IMASK_IFERREN | \
-+ DTSEC_IMASK_MAGEN | \
-+ DTSEC_IMASK_TDPEEN | \
-+ DTSEC_IMASK_RDPEEN))
-+
-+#define GET_EXCEPTION_FLAG(bitMask, exception) switch (exception){ \
-+ case e_FM_MAC_EX_1G_BAB_RX: \
-+ bitMask = DTSEC_IMASK_BREN; break; \
-+ case e_FM_MAC_EX_1G_RX_CTL: \
-+ bitMask = DTSEC_IMASK_RXCEN; break; \
-+ case e_FM_MAC_EX_1G_GRATEFUL_TX_STP_COMPLET: \
-+ bitMask = DTSEC_IMASK_GTSCEN ; break; \
-+ case e_FM_MAC_EX_1G_BAB_TX: \
-+ bitMask = DTSEC_IMASK_BTEN ; break; \
-+ case e_FM_MAC_EX_1G_TX_CTL: \
-+ bitMask = DTSEC_IMASK_TXCEN ; break; \
-+ case e_FM_MAC_EX_1G_TX_ERR: \
-+ bitMask = DTSEC_IMASK_TXEEN ; break; \
-+ case e_FM_MAC_EX_1G_LATE_COL: \
-+ bitMask = DTSEC_IMASK_LCEN ; break; \
-+ case e_FM_MAC_EX_1G_COL_RET_LMT: \
-+ bitMask = DTSEC_IMASK_CRLEN ; break; \
-+ case e_FM_MAC_EX_1G_TX_FIFO_UNDRN: \
-+ bitMask = DTSEC_IMASK_XFUNEN ; break; \
-+ case e_FM_MAC_EX_1G_MAG_PCKT: \
-+ bitMask = DTSEC_IMASK_MAGEN ; break; \
-+ case e_FM_MAC_EX_1G_MII_MNG_RD_COMPLET: \
-+ bitMask = DTSEC_IMASK_MMRDEN; break; \
-+ case e_FM_MAC_EX_1G_MII_MNG_WR_COMPLET: \
-+ bitMask = DTSEC_IMASK_MMWREN ; break; \
-+ case e_FM_MAC_EX_1G_GRATEFUL_RX_STP_COMPLET: \
-+ bitMask = DTSEC_IMASK_GRSCEN; break; \
-+ case e_FM_MAC_EX_1G_TX_DATA_ERR: \
-+ bitMask = DTSEC_IMASK_TDPEEN; break; \
-+ case e_FM_MAC_EX_1G_RX_MIB_CNT_OVFL: \
-+ bitMask = DTSEC_IMASK_MSROEN ; break; \
-+ default: bitMask = 0;break;}
-+
-+
-+#define MAX_PACKET_ALIGNMENT 31
-+#define MAX_INTER_PACKET_GAP 0x7f
-+#define MAX_INTER_PALTERNATE_BEB 0x0f
-+#define MAX_RETRANSMISSION 0x0f
-+#define MAX_COLLISION_WINDOW 0x03ff
-+
-+
-+/********************* From mac ext ******************************************/
-+typedef uint32_t t_ErrorDisable;
-+
-+#define ERROR_DISABLE_TRANSMIT 0x00400000
-+#define ERROR_DISABLE_LATE_COLLISION 0x00040000
-+#define ERROR_DISABLE_COLLISION_RETRY_LIMIT 0x00020000
-+#define ERROR_DISABLE_TxFIFO_UNDERRUN 0x00010000
-+#define ERROR_DISABLE_TxABORT 0x00008000
-+#define ERROR_DISABLE_INTERFACE 0x00004000
-+#define ERROR_DISABLE_TxDATA_PARITY 0x00000002
-+#define ERROR_DISABLE_RxDATA_PARITY 0x00000001
-+
-+/*****************************************************************************/
-+#define DTSEC_NUM_OF_PADDRS 15 /* number of pattern match registers (entries) */
-+
-+#define GROUP_ADDRESS 0x0000010000000000LL /* Group address bit indication */
-+
-+#define HASH_TABLE_SIZE 256 /* Hash table size (= 32 bits * 8 regs) */
-+
-+#define HASH_TABLE_SIZE 256 /* Hash table size (32 bits * 8 regs) */
-+#define EXTENDED_HASH_TABLE_SIZE 512 /* Extended Hash table size (32 bits * 16 regs) */
-+
-+#define DTSEC_TO_MII_OFFSET 0x1000 /* number of pattern match registers (entries) */
-+
-+#define MAX_PHYS 32 /* maximum number of phys */
-+
-+#define VAL32BIT 0x100000000LL
-+#define VAL22BIT 0x00400000
-+#define VAL16BIT 0x00010000
-+#define VAL12BIT 0x00001000
-+
-+/* CAR1/2 bits */
-+#define CAR1_TR64 0x80000000
-+#define CAR1_TR127 0x40000000
-+#define CAR1_TR255 0x20000000
-+#define CAR1_TR511 0x10000000
-+#define CAR1_TRK1 0x08000000
-+#define CAR1_TRMAX 0x04000000
-+#define CAR1_TRMGV 0x02000000
-+
-+#define CAR1_RBYT 0x00010000
-+#define CAR1_RPKT 0x00008000
-+#define CAR1_RMCA 0x00002000
-+#define CAR1_RBCA 0x00001000
-+#define CAR1_RXPF 0x00000400
-+#define CAR1_RALN 0x00000100
-+#define CAR1_RFLR 0x00000080
-+#define CAR1_RCDE 0x00000040
-+#define CAR1_RCSE 0x00000020
-+#define CAR1_RUND 0x00000010
-+#define CAR1_ROVR 0x00000008
-+#define CAR1_RFRG 0x00000004
-+#define CAR1_RJBR 0x00000002
-+#define CAR1_RDRP 0x00000001
-+
-+#define CAR2_TFCS 0x00040000
-+#define CAR2_TBYT 0x00002000
-+#define CAR2_TPKT 0x00001000
-+#define CAR2_TMCA 0x00000800
-+#define CAR2_TBCA 0x00000400
-+#define CAR2_TXPF 0x00000200
-+#define CAR2_TDRP 0x00000001
-+
-+typedef struct t_InternalStatistics
-+{
-+ uint64_t tr64;
-+ uint64_t tr127;
-+ uint64_t tr255;
-+ uint64_t tr511;
-+ uint64_t tr1k;
-+ uint64_t trmax;
-+ uint64_t trmgv;
-+ uint64_t rfrg;
-+ uint64_t rjbr;
-+ uint64_t rdrp;
-+ uint64_t raln;
-+ uint64_t rund;
-+ uint64_t rovr;
-+ uint64_t rxpf;
-+ uint64_t txpf;
-+ uint64_t rbyt;
-+ uint64_t rpkt;
-+ uint64_t rmca;
-+ uint64_t rbca;
-+ uint64_t rflr;
-+ uint64_t rcde;
-+ uint64_t rcse;
-+ uint64_t tbyt;
-+ uint64_t tpkt;
-+ uint64_t tmca;
-+ uint64_t tbca;
-+ uint64_t tdrp;
-+ uint64_t tfcs;
-+} t_InternalStatistics;
-+
-+typedef struct {
-+ t_FmMacControllerDriver fmMacControllerDriver;
-+ t_Handle h_App; /**< Handle to the upper layer application */
-+ struct dtsec_regs *p_MemMap; /**< pointer to dTSEC memory mapped registers. */
-+ struct dtsec_mii_reg *p_MiiMemMap; /**< pointer to dTSEC MII memory mapped registers. */
-+ uint64_t addr; /**< MAC address of device; */
-+ e_EnetMode enetMode; /**< Ethernet physical interface */
-+ t_FmMacExceptionCallback *f_Exception;
-+ int mdioIrq;
-+ t_FmMacExceptionCallback *f_Event;
-+ bool indAddrRegUsed[DTSEC_NUM_OF_PADDRS]; /**< Whether a particular individual address recognition register is being used */
-+ uint64_t paddr[DTSEC_NUM_OF_PADDRS]; /**< MAC address for particular individual address recognition register */
-+ uint8_t numOfIndAddrInRegs; /**< Number of individual addresses in registers for this station. */
-+ bool halfDuplex;
-+ t_InternalStatistics internalStatistics;
-+ t_EthHash *p_MulticastAddrHash; /* pointer to driver's global address hash table */
-+ t_EthHash *p_UnicastAddrHash; /* pointer to driver's individual address hash table */
-+ uint8_t macId;
-+ uint8_t tbi_phy_addr;
-+ uint32_t exceptions;
-+ bool ptpTsuEnabled;
-+ bool enTsuErrExeption;
-+ e_FmMacStatisticsLevel statisticsLevel;
-+ struct dtsec_cfg *p_DtsecDriverParam;
-+} t_Dtsec;
-+
-+
-+#endif /* __DTSEC_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MAC/dtsec_mii_acc.c
-@@ -0,0 +1,97 @@
-+/*
-+ * Copyright 2008-2013 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/******************************************************************************
-+ @File dtsec_mii_acc.c
-+
-+ @Description FM dtsec MII register access MAC ...
-+*//***************************************************************************/
-+
-+#include "error_ext.h"
-+#include "std_ext.h"
-+#include "fm_mac.h"
-+#include "dtsec.h"
-+#include "fsl_fman_dtsec_mii_acc.h"
-+
-+
-+/*****************************************************************************/
-+t_Error DTSEC_MII_WritePhyReg(t_Handle h_Dtsec,
-+ uint8_t phyAddr,
-+ uint8_t reg,
-+ uint16_t data)
-+{
-+ t_Dtsec *p_Dtsec = (t_Dtsec *)h_Dtsec;
-+ struct dtsec_mii_reg *miiregs;
-+ uint16_t dtsec_freq;
-+ t_Error err;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Dtsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Dtsec->p_MiiMemMap, E_INVALID_HANDLE);
-+
-+ dtsec_freq = (uint16_t)(p_Dtsec->fmMacControllerDriver.clkFreq >> 1);
-+ miiregs = p_Dtsec->p_MiiMemMap;
-+
-+ err = (t_Error)fman_dtsec_mii_write_reg(miiregs, phyAddr, reg, data, dtsec_freq);
-+
-+ return err;
-+}
-+
-+/*****************************************************************************/
-+t_Error DTSEC_MII_ReadPhyReg(t_Handle h_Dtsec,
-+ uint8_t phyAddr,
-+ uint8_t reg,
-+ uint16_t *p_Data)
-+{
-+ t_Dtsec *p_Dtsec = (t_Dtsec *)h_Dtsec;
-+ struct dtsec_mii_reg *miiregs;
-+ uint16_t dtsec_freq;
-+ t_Error err;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Dtsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Dtsec->p_MiiMemMap, E_INVALID_HANDLE);
-+
-+ dtsec_freq = (uint16_t)(p_Dtsec->fmMacControllerDriver.clkFreq >> 1);
-+ miiregs = p_Dtsec->p_MiiMemMap;
-+
-+ err = fman_dtsec_mii_read_reg(miiregs, phyAddr, reg, p_Data, dtsec_freq);
-+
-+ if (*p_Data == 0xffff)
-+ RETURN_ERROR(MINOR, E_NO_DEVICE,
-+ ("Read wrong data (0xffff): phyAddr 0x%x, reg 0x%x",
-+ phyAddr, reg));
-+ if (err)
-+ RETURN_ERROR(MINOR, (t_Error)err, NO_MSG);
-+
-+ return E_OK;
-+}
-+
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MAC/dtsec_mii_acc.h
-@@ -0,0 +1,42 @@
-+/*
-+ * Copyright 2008-2013 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#ifndef __DTSEC_MII_ACC_H
-+#define __DTSEC_MII_ACC_H
-+
-+#include "std_ext.h"
-+
-+
-+t_Error DTSEC_MII_WritePhyReg(t_Handle h_Dtsec, uint8_t phyAddr, uint8_t reg, uint16_t data);
-+t_Error DTSEC_MII_ReadPhyReg(t_Handle h_Dtsec, uint8_t phyAddr, uint8_t reg, uint16_t *p_Data);
-+
-+#endif /* __DTSEC_MII_ACC_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MAC/fm_mac.c
-@@ -0,0 +1,674 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/******************************************************************************
-+ @File fm_mac.c
-+
-+ @Description FM MAC ...
-+*//***************************************************************************/
-+#include "std_ext.h"
-+#include "string_ext.h"
-+#include "sprint_ext.h"
-+#include "error_ext.h"
-+#include "fm_ext.h"
-+
-+#include "fm_common.h"
-+#include "fm_mac.h"
-+
-+
-+/* ......................................................................... */
-+
-+t_Handle FM_MAC_Config (t_FmMacParams *p_FmMacParam)
-+{
-+ t_FmMacControllerDriver *p_FmMacControllerDriver;
-+ uint16_t fmClkFreq;
-+
-+ SANITY_CHECK_RETURN_VALUE(p_FmMacParam, E_INVALID_HANDLE, NULL);
-+
-+ fmClkFreq = FmGetClockFreq(p_FmMacParam->h_Fm);
-+ if (fmClkFreq == 0)
-+ {
-+ REPORT_ERROR(MAJOR, E_INVALID_STATE, ("Can't get clock for MAC!"));
-+ return NULL;
-+ }
-+
-+#if (DPAA_VERSION == 10)
-+ if (ENET_SPEED_FROM_MODE(p_FmMacParam->enetMode) < e_ENET_SPEED_10000)
-+ p_FmMacControllerDriver = (t_FmMacControllerDriver *)DTSEC_Config(p_FmMacParam);
-+ else
-+#if FM_MAX_NUM_OF_10G_MACS > 0
-+ p_FmMacControllerDriver = (t_FmMacControllerDriver *)TGEC_Config(p_FmMacParam);
-+#else
-+ p_FmMacControllerDriver = NULL;
-+#endif /* FM_MAX_NUM_OF_10G_MACS > 0 */
-+#else
-+ p_FmMacControllerDriver = (t_FmMacControllerDriver *)MEMAC_Config(p_FmMacParam);
-+#endif /* (DPAA_VERSION == 10) */
-+
-+ if (!p_FmMacControllerDriver)
-+ return NULL;
-+
-+ p_FmMacControllerDriver->h_Fm = p_FmMacParam->h_Fm;
-+ p_FmMacControllerDriver->enetMode = p_FmMacParam->enetMode;
-+ p_FmMacControllerDriver->macId = p_FmMacParam->macId;
-+ p_FmMacControllerDriver->resetOnInit = DEFAULT_resetOnInit;
-+
-+ p_FmMacControllerDriver->clkFreq = fmClkFreq;
-+
-+ return (t_Handle)p_FmMacControllerDriver;
-+}
-+
-+/* ......................................................................... */
-+
-+t_Error FM_MAC_Init (t_Handle h_FmMac)
-+{
-+ t_FmMacControllerDriver *p_FmMacControllerDriver = (t_FmMacControllerDriver *)h_FmMac;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacControllerDriver, E_INVALID_HANDLE);
-+
-+ if (p_FmMacControllerDriver->resetOnInit &&
-+ !p_FmMacControllerDriver->f_FM_MAC_ConfigResetOnInit &&
-+ (FmResetMac(p_FmMacControllerDriver->h_Fm,
-+ ((ENET_INTERFACE_FROM_MODE(p_FmMacControllerDriver->enetMode) == e_ENET_IF_XGMII) ?
-+ e_FM_MAC_10G : e_FM_MAC_1G),
-+ p_FmMacControllerDriver->macId) != E_OK))
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("Can't reset MAC!"));
-+
-+ if (p_FmMacControllerDriver->f_FM_MAC_Init)
-+ return p_FmMacControllerDriver->f_FM_MAC_Init(h_FmMac);
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, NO_MSG);
-+}
-+
-+/* ......................................................................... */
-+
-+t_Error FM_MAC_Free (t_Handle h_FmMac)
-+{
-+ t_FmMacControllerDriver *p_FmMacControllerDriver = (t_FmMacControllerDriver *)h_FmMac;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacControllerDriver, E_INVALID_HANDLE);
-+
-+ if (p_FmMacControllerDriver->f_FM_MAC_Free)
-+ return p_FmMacControllerDriver->f_FM_MAC_Free(h_FmMac);
-+
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, NO_MSG);
-+}
-+
-+/* ......................................................................... */
-+
-+t_Error FM_MAC_ConfigResetOnInit (t_Handle h_FmMac, bool enable)
-+{
-+ t_FmMacControllerDriver *p_FmMacControllerDriver = (t_FmMacControllerDriver *)h_FmMac;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacControllerDriver, E_INVALID_HANDLE);
-+
-+ if (p_FmMacControllerDriver->f_FM_MAC_ConfigResetOnInit)
-+ return p_FmMacControllerDriver->f_FM_MAC_ConfigResetOnInit(h_FmMac, enable);
-+
-+ p_FmMacControllerDriver->resetOnInit = enable;
-+
-+ return E_OK;
-+}
-+
-+/* ......................................................................... */
-+
-+t_Error FM_MAC_ConfigLoopback (t_Handle h_FmMac, bool newVal)
-+{
-+ t_FmMacControllerDriver *p_FmMacControllerDriver = (t_FmMacControllerDriver *)h_FmMac;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacControllerDriver, E_INVALID_HANDLE);
-+
-+ if (p_FmMacControllerDriver->f_FM_MAC_ConfigLoopback)
-+ return p_FmMacControllerDriver->f_FM_MAC_ConfigLoopback(h_FmMac, newVal);
-+
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, NO_MSG);
-+}
-+
-+/* ......................................................................... */
-+
-+t_Error FM_MAC_ConfigMaxFrameLength (t_Handle h_FmMac, uint16_t newVal)
-+{
-+ t_FmMacControllerDriver *p_FmMacControllerDriver = (t_FmMacControllerDriver *)h_FmMac;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacControllerDriver, E_INVALID_HANDLE);
-+
-+ if (p_FmMacControllerDriver->f_FM_MAC_ConfigMaxFrameLength)
-+ return p_FmMacControllerDriver->f_FM_MAC_ConfigMaxFrameLength(h_FmMac, newVal);
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, NO_MSG);
-+}
-+
-+/* ......................................................................... */
-+
-+t_Error FM_MAC_ConfigWan (t_Handle h_FmMac, bool flag)
-+{
-+ t_FmMacControllerDriver *p_FmMacControllerDriver = (t_FmMacControllerDriver *)h_FmMac;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacControllerDriver, E_INVALID_HANDLE);
-+
-+ if (p_FmMacControllerDriver->f_FM_MAC_ConfigWan)
-+ return p_FmMacControllerDriver->f_FM_MAC_ConfigWan(h_FmMac, flag);
-+
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, NO_MSG);
-+}
-+
-+/* ......................................................................... */
-+
-+t_Error FM_MAC_ConfigPadAndCrc (t_Handle h_FmMac, bool newVal)
-+{
-+ t_FmMacControllerDriver *p_FmMacControllerDriver = (t_FmMacControllerDriver *)h_FmMac;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacControllerDriver, E_INVALID_HANDLE);
-+
-+ if (p_FmMacControllerDriver->f_FM_MAC_ConfigPadAndCrc)
-+ return p_FmMacControllerDriver->f_FM_MAC_ConfigPadAndCrc(h_FmMac, newVal);
-+
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, NO_MSG);
-+}
-+
-+/* ......................................................................... */
-+
-+t_Error FM_MAC_ConfigHalfDuplex (t_Handle h_FmMac, bool newVal)
-+{
-+ t_FmMacControllerDriver *p_FmMacControllerDriver = (t_FmMacControllerDriver *)h_FmMac;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacControllerDriver, E_INVALID_HANDLE);
-+
-+ if (p_FmMacControllerDriver->f_FM_MAC_ConfigHalfDuplex)
-+ return p_FmMacControllerDriver->f_FM_MAC_ConfigHalfDuplex(h_FmMac,newVal);
-+
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, NO_MSG);
-+}
-+
-+/* ......................................................................... */
-+
-+t_Error FM_MAC_ConfigTbiPhyAddr (t_Handle h_FmMac, uint8_t newVal)
-+{
-+ t_FmMacControllerDriver *p_FmMacControllerDriver = (t_FmMacControllerDriver *)h_FmMac;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacControllerDriver, E_INVALID_HANDLE);
-+
-+ if (p_FmMacControllerDriver->f_FM_MAC_ConfigTbiPhyAddr)
-+ return p_FmMacControllerDriver->f_FM_MAC_ConfigTbiPhyAddr(h_FmMac,newVal);
-+
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, NO_MSG);
-+}
-+
-+/* ......................................................................... */
-+
-+t_Error FM_MAC_ConfigLengthCheck (t_Handle h_FmMac, bool newVal)
-+{
-+ t_FmMacControllerDriver *p_FmMacControllerDriver = (t_FmMacControllerDriver *)h_FmMac;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacControllerDriver, E_INVALID_HANDLE);
-+
-+ if (p_FmMacControllerDriver->f_FM_MAC_ConfigLengthCheck)
-+ return p_FmMacControllerDriver->f_FM_MAC_ConfigLengthCheck(h_FmMac,newVal);
-+
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, NO_MSG);
-+}
-+
-+/* ......................................................................... */
-+
-+t_Error FM_MAC_ConfigException (t_Handle h_FmMac, e_FmMacExceptions ex, bool enable)
-+{
-+ t_FmMacControllerDriver *p_FmMacControllerDriver = (t_FmMacControllerDriver *)h_FmMac;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacControllerDriver, E_INVALID_HANDLE);
-+
-+ if (p_FmMacControllerDriver->f_FM_MAC_ConfigException)
-+ return p_FmMacControllerDriver->f_FM_MAC_ConfigException(h_FmMac, ex, enable);
-+
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, NO_MSG);
-+}
-+
-+#ifdef FM_TX_ECC_FRMS_ERRATA_10GMAC_A004
-+/* ......................................................................... */
-+
-+t_Error FM_MAC_ConfigSkipFman11Workaround (t_Handle h_FmMac)
-+{
-+ t_FmMacControllerDriver *p_FmMacControllerDriver = (t_FmMacControllerDriver *)h_FmMac;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacControllerDriver, E_INVALID_HANDLE);
-+
-+ if (p_FmMacControllerDriver->f_FM_MAC_ConfigSkipFman11Workaround)
-+ return p_FmMacControllerDriver->f_FM_MAC_ConfigSkipFman11Workaround(h_FmMac);
-+
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, NO_MSG);
-+}
-+#endif /* FM_TX_ECC_FRMS_ERRATA_10GMAC_A004 */
-+
-+
-+/*****************************************************************************/
-+/* Run Time Control */
-+/*****************************************************************************/
-+
-+/* ......................................................................... */
-+
-+t_Error FM_MAC_Enable (t_Handle h_FmMac, e_CommMode mode)
-+{
-+ t_FmMacControllerDriver *p_FmMacControllerDriver = (t_FmMacControllerDriver *)h_FmMac;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacControllerDriver, E_INVALID_HANDLE);
-+
-+ if (p_FmMacControllerDriver->f_FM_MAC_Enable)
-+ return p_FmMacControllerDriver->f_FM_MAC_Enable(h_FmMac, mode);
-+
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, NO_MSG);
-+}
-+
-+/* ......................................................................... */
-+
-+t_Error FM_MAC_Disable (t_Handle h_FmMac, e_CommMode mode)
-+{
-+ t_FmMacControllerDriver *p_FmMacControllerDriver = (t_FmMacControllerDriver *)h_FmMac;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacControllerDriver, E_INVALID_HANDLE);
-+
-+ if (p_FmMacControllerDriver->f_FM_MAC_Disable)
-+ return p_FmMacControllerDriver->f_FM_MAC_Disable(h_FmMac, mode);
-+
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, NO_MSG);
-+}
-+
-+t_Error FM_MAC_Resume (t_Handle h_FmMac)
-+{
-+ t_FmMacControllerDriver *p_FmMacControllerDriver = (t_FmMacControllerDriver *)h_FmMac;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacControllerDriver, E_INVALID_HANDLE);
-+
-+ if (p_FmMacControllerDriver->f_FM_MAC_Resume)
-+ return p_FmMacControllerDriver->f_FM_MAC_Resume(h_FmMac);
-+
-+ return E_OK;
-+}
-+
-+/* ......................................................................... */
-+
-+t_Error FM_MAC_Enable1588TimeStamp (t_Handle h_FmMac)
-+{
-+ t_FmMacControllerDriver *p_FmMacControllerDriver = (t_FmMacControllerDriver *)h_FmMac;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacControllerDriver, E_INVALID_HANDLE);
-+
-+ if (p_FmMacControllerDriver->f_FM_MAC_Enable1588TimeStamp)
-+ return p_FmMacControllerDriver->f_FM_MAC_Enable1588TimeStamp(h_FmMac);
-+
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, NO_MSG);
-+}
-+
-+/* ......................................................................... */
-+
-+t_Error FM_MAC_Disable1588TimeStamp (t_Handle h_FmMac)
-+{
-+ t_FmMacControllerDriver *p_FmMacControllerDriver = (t_FmMacControllerDriver *)h_FmMac;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacControllerDriver, E_INVALID_HANDLE);
-+
-+ if (p_FmMacControllerDriver->f_FM_MAC_Disable1588TimeStamp)
-+ return p_FmMacControllerDriver->f_FM_MAC_Disable1588TimeStamp(h_FmMac);
-+
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, NO_MSG);
-+}
-+
-+/* ......................................................................... */
-+
-+t_Error FM_MAC_SetTxAutoPauseFrames(t_Handle h_FmMac,
-+ uint16_t pauseTime)
-+{
-+ t_FmMacControllerDriver *p_FmMacControllerDriver = (t_FmMacControllerDriver *)h_FmMac;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacControllerDriver, E_INVALID_HANDLE);
-+
-+ if (p_FmMacControllerDriver->f_FM_MAC_SetTxAutoPauseFrames)
-+ return p_FmMacControllerDriver->f_FM_MAC_SetTxAutoPauseFrames(h_FmMac,
-+ pauseTime);
-+
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, NO_MSG);
-+}
-+
-+/* ......................................................................... */
-+
-+t_Error FM_MAC_SetTxPauseFrames(t_Handle h_FmMac,
-+ uint8_t priority,
-+ uint16_t pauseTime,
-+ uint16_t threshTime)
-+{
-+ t_FmMacControllerDriver *p_FmMacControllerDriver = (t_FmMacControllerDriver *)h_FmMac;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacControllerDriver, E_INVALID_HANDLE);
-+
-+ if (p_FmMacControllerDriver->f_FM_MAC_SetTxPauseFrames)
-+ return p_FmMacControllerDriver->f_FM_MAC_SetTxPauseFrames(h_FmMac,
-+ priority,
-+ pauseTime,
-+ threshTime);
-+
-+ RETURN_ERROR(MAJOR, E_NOT_SUPPORTED, NO_MSG);
-+}
-+
-+/* ......................................................................... */
-+
-+t_Error FM_MAC_SetRxIgnorePauseFrames (t_Handle h_FmMac, bool en)
-+{
-+ t_FmMacControllerDriver *p_FmMacControllerDriver = (t_FmMacControllerDriver *)h_FmMac;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacControllerDriver, E_INVALID_HANDLE);
-+
-+ if (p_FmMacControllerDriver->f_FM_MAC_SetRxIgnorePauseFrames)
-+ return p_FmMacControllerDriver->f_FM_MAC_SetRxIgnorePauseFrames(h_FmMac, en);
-+
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, NO_MSG);
-+}
-+
-+/* ......................................................................... */
-+
-+t_Error FM_MAC_SetWakeOnLan (t_Handle h_FmMac, bool en)
-+{
-+ t_FmMacControllerDriver *p_FmMacControllerDriver = (t_FmMacControllerDriver *)h_FmMac;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacControllerDriver, E_INVALID_HANDLE);
-+
-+ if (p_FmMacControllerDriver->f_FM_MAC_SetWakeOnLan)
-+ return p_FmMacControllerDriver->f_FM_MAC_SetWakeOnLan(h_FmMac, en);
-+
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, NO_MSG);
-+}
-+
-+/* ......................................................................... */
-+
-+t_Error FM_MAC_ResetCounters (t_Handle h_FmMac)
-+{
-+ t_FmMacControllerDriver *p_FmMacControllerDriver = (t_FmMacControllerDriver *)h_FmMac;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacControllerDriver, E_INVALID_HANDLE);
-+
-+ if (p_FmMacControllerDriver->f_FM_MAC_ResetCounters)
-+ return p_FmMacControllerDriver->f_FM_MAC_ResetCounters(h_FmMac);
-+
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, NO_MSG);
-+}
-+
-+/* ......................................................................... */
-+
-+t_Error FM_MAC_SetException(t_Handle h_FmMac, e_FmMacExceptions ex, bool enable)
-+{
-+ t_FmMacControllerDriver *p_FmMacControllerDriver = (t_FmMacControllerDriver *)h_FmMac;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacControllerDriver, E_INVALID_HANDLE);
-+
-+ if (p_FmMacControllerDriver->f_FM_MAC_SetException)
-+ return p_FmMacControllerDriver->f_FM_MAC_SetException(h_FmMac, ex, enable);
-+
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, NO_MSG);
-+}
-+
-+/* ......................................................................... */
-+
-+t_Error FM_MAC_SetStatistics (t_Handle h_FmMac, e_FmMacStatisticsLevel statisticsLevel)
-+{
-+ t_FmMacControllerDriver *p_FmMacControllerDriver = (t_FmMacControllerDriver *)h_FmMac;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacControllerDriver, E_INVALID_HANDLE);
-+
-+ if (p_FmMacControllerDriver->f_FM_MAC_SetStatistics)
-+ return p_FmMacControllerDriver->f_FM_MAC_SetStatistics(h_FmMac, statisticsLevel);
-+
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, NO_MSG);
-+}
-+
-+/* ......................................................................... */
-+
-+t_Error FM_MAC_GetStatistics (t_Handle h_FmMac, t_FmMacStatistics *p_Statistics)
-+{
-+ t_FmMacControllerDriver *p_FmMacControllerDriver = (t_FmMacControllerDriver *)h_FmMac;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacControllerDriver, E_INVALID_HANDLE);
-+
-+ if (p_FmMacControllerDriver->f_FM_MAC_GetStatistics)
-+ return p_FmMacControllerDriver->f_FM_MAC_GetStatistics(h_FmMac, p_Statistics);
-+
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, NO_MSG);
-+}
-+
-+/* ......................................................................... */
-+
-+t_Error FM_MAC_GetFrameSizeCounters(t_Handle h_FmMac, t_FmMacFrameSizeCounters *p_FrameSizeCounters, e_CommMode type)
-+{
-+ t_FmMacControllerDriver *p_FmMacControllerDriver = (t_FmMacControllerDriver *)h_FmMac;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacControllerDriver, E_INVALID_HANDLE);
-+
-+ memset(p_FrameSizeCounters, 0, sizeof(t_FmMacFrameSizeCounters));
-+
-+ if (p_FmMacControllerDriver->f_FM_MAC_GetFrameSizeCounters)
-+ return p_FmMacControllerDriver->f_FM_MAC_GetFrameSizeCounters(h_FmMac, p_FrameSizeCounters, type);
-+
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, NO_MSG);
-+}
-+
-+/* ......................................................................... */
-+
-+t_Error FM_MAC_ModifyMacAddr (t_Handle h_FmMac, t_EnetAddr *p_EnetAddr)
-+{
-+ t_FmMacControllerDriver *p_FmMacControllerDriver = (t_FmMacControllerDriver *)h_FmMac;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacControllerDriver, E_INVALID_HANDLE);
-+
-+ if (p_FmMacControllerDriver->f_FM_MAC_ModifyMacAddr)
-+ return p_FmMacControllerDriver->f_FM_MAC_ModifyMacAddr(h_FmMac, p_EnetAddr);
-+
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, NO_MSG);
-+}
-+
-+/* ......................................................................... */
-+
-+t_Error FM_MAC_AddHashMacAddr (t_Handle h_FmMac, t_EnetAddr *p_EnetAddr)
-+{
-+ t_FmMacControllerDriver *p_FmMacControllerDriver = (t_FmMacControllerDriver *)h_FmMac;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacControllerDriver, E_INVALID_HANDLE);
-+
-+ if (p_FmMacControllerDriver->f_FM_MAC_AddHashMacAddr)
-+ return p_FmMacControllerDriver->f_FM_MAC_AddHashMacAddr(h_FmMac, p_EnetAddr);
-+
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, NO_MSG);
-+}
-+
-+/* ......................................................................... */
-+
-+t_Error FM_MAC_RemoveHashMacAddr (t_Handle h_FmMac, t_EnetAddr *p_EnetAddr)
-+{
-+ t_FmMacControllerDriver *p_FmMacControllerDriver = (t_FmMacControllerDriver *)h_FmMac;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacControllerDriver, E_INVALID_HANDLE);
-+
-+ if (p_FmMacControllerDriver->f_FM_MAC_RemoveHashMacAddr)
-+ return p_FmMacControllerDriver->f_FM_MAC_RemoveHashMacAddr(h_FmMac, p_EnetAddr);
-+
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, NO_MSG);
-+}
-+
-+/* ......................................................................... */
-+
-+t_Error FM_MAC_AddExactMatchMacAddr (t_Handle h_FmMac, t_EnetAddr *p_EnetAddr)
-+{
-+ t_FmMacControllerDriver *p_FmMacControllerDriver = (t_FmMacControllerDriver *)h_FmMac;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacControllerDriver, E_INVALID_HANDLE);
-+
-+ if (p_FmMacControllerDriver->f_FM_MAC_AddExactMatchMacAddr)
-+ return p_FmMacControllerDriver->f_FM_MAC_AddExactMatchMacAddr(h_FmMac, p_EnetAddr);
-+
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, NO_MSG);
-+}
-+
-+/* ......................................................................... */
-+
-+t_Error FM_MAC_RemovelExactMatchMacAddr (t_Handle h_FmMac, t_EnetAddr *p_EnetAddr)
-+{
-+ t_FmMacControllerDriver *p_FmMacControllerDriver = (t_FmMacControllerDriver *)h_FmMac;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacControllerDriver, E_INVALID_HANDLE);
-+
-+ if (p_FmMacControllerDriver->f_FM_MAC_RemovelExactMatchMacAddr)
-+ return p_FmMacControllerDriver->f_FM_MAC_RemovelExactMatchMacAddr(h_FmMac, p_EnetAddr);
-+
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, NO_MSG);
-+}
-+
-+/* ......................................................................... */
-+
-+t_Error FM_MAC_GetVesrion (t_Handle h_FmMac, uint32_t *macVresion)
-+{
-+ t_FmMacControllerDriver *p_FmMacControllerDriver = (t_FmMacControllerDriver *)h_FmMac;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacControllerDriver, E_INVALID_HANDLE);
-+
-+ if (p_FmMacControllerDriver->f_FM_MAC_GetVersion)
-+ return p_FmMacControllerDriver->f_FM_MAC_GetVersion(h_FmMac, macVresion);
-+
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, NO_MSG);
-+
-+}
-+
-+/* ......................................................................... */
-+
-+t_Error FM_MAC_GetId (t_Handle h_FmMac, uint32_t *macId)
-+{
-+ t_FmMacControllerDriver *p_FmMacControllerDriver = (t_FmMacControllerDriver *)h_FmMac;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacControllerDriver, E_INVALID_HANDLE);
-+
-+ if (p_FmMacControllerDriver->f_FM_MAC_GetId)
-+ return p_FmMacControllerDriver->f_FM_MAC_GetId(h_FmMac, macId);
-+
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, NO_MSG);
-+}
-+
-+/* ......................................................................... */
-+
-+t_Error FM_MAC_SetPromiscuous (t_Handle h_FmMac, bool newVal)
-+{
-+ t_FmMacControllerDriver *p_FmMacControllerDriver = (t_FmMacControllerDriver *)h_FmMac;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacControllerDriver, E_INVALID_HANDLE);
-+
-+ if (p_FmMacControllerDriver->f_FM_MAC_SetPromiscuous)
-+ return p_FmMacControllerDriver->f_FM_MAC_SetPromiscuous(h_FmMac, newVal);
-+
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, NO_MSG);
-+}
-+
-+/* ......................................................................... */
-+
-+t_Error FM_MAC_AdjustLink(t_Handle h_FmMac, e_EnetSpeed speed, bool fullDuplex)
-+{
-+ t_FmMacControllerDriver *p_FmMacControllerDriver = (t_FmMacControllerDriver *)h_FmMac;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacControllerDriver, E_INVALID_HANDLE);
-+
-+ if (p_FmMacControllerDriver->f_FM_MAC_AdjustLink)
-+ return p_FmMacControllerDriver->f_FM_MAC_AdjustLink(h_FmMac, speed, fullDuplex);
-+
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, NO_MSG);
-+}
-+
-+/* ......................................................................... */
-+
-+t_Error FM_MAC_RestartAutoneg(t_Handle h_FmMac)
-+{
-+ t_FmMacControllerDriver *p_FmMacControllerDriver = (t_FmMacControllerDriver *)h_FmMac;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacControllerDriver, E_INVALID_HANDLE);
-+
-+ if (p_FmMacControllerDriver->f_FM_MAC_RestartAutoneg)
-+ return p_FmMacControllerDriver->f_FM_MAC_RestartAutoneg(h_FmMac);
-+
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, NO_MSG);
-+}
-+
-+/* ......................................................................... */
-+
-+t_Error FM_MAC_MII_WritePhyReg (t_Handle h_FmMac, uint8_t phyAddr, uint8_t reg, uint16_t data)
-+{
-+ t_FmMacControllerDriver *p_FmMacControllerDriver = (t_FmMacControllerDriver *)h_FmMac;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacControllerDriver, E_INVALID_HANDLE);
-+
-+ if (p_FmMacControllerDriver->f_FM_MAC_MII_WritePhyReg)
-+ return p_FmMacControllerDriver->f_FM_MAC_MII_WritePhyReg(h_FmMac, phyAddr, reg, data);
-+
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, NO_MSG);
-+}
-+
-+/* ......................................................................... */
-+
-+t_Error FM_MAC_MII_ReadPhyReg(t_Handle h_FmMac, uint8_t phyAddr, uint8_t reg, uint16_t *p_Data)
-+{
-+ t_FmMacControllerDriver *p_FmMacControllerDriver = (t_FmMacControllerDriver *)h_FmMac;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacControllerDriver, E_INVALID_HANDLE);
-+
-+ if (p_FmMacControllerDriver->f_FM_MAC_MII_ReadPhyReg)
-+ return p_FmMacControllerDriver->f_FM_MAC_MII_ReadPhyReg(h_FmMac, phyAddr, reg, p_Data);
-+
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, NO_MSG);
-+}
-+
-+/* ......................................................................... */
-+
-+uint16_t FM_MAC_GetMaxFrameLength(t_Handle h_FmMac)
-+{
-+ t_FmMacControllerDriver *p_FmMacControllerDriver = (t_FmMacControllerDriver *)h_FmMac;
-+
-+ SANITY_CHECK_RETURN_VALUE(p_FmMacControllerDriver, E_INVALID_HANDLE, 0);
-+
-+ if (p_FmMacControllerDriver->f_FM_MAC_GetMaxFrameLength)
-+ return p_FmMacControllerDriver->f_FM_MAC_GetMaxFrameLength(h_FmMac);
-+
-+ REPORT_ERROR(MINOR, E_NOT_SUPPORTED, NO_MSG);
-+ return 0;
-+}
-+
-+#if (defined(DEBUG_ERRORS) && (DEBUG_ERRORS > 0))
-+/*****************************************************************************/
-+t_Error FM_MAC_DumpRegs(t_Handle h_FmMac)
-+{
-+ t_FmMacControllerDriver *p_FmMacControllerDriver = (t_FmMacControllerDriver *)h_FmMac;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacControllerDriver, E_INVALID_HANDLE);
-+
-+ if (p_FmMacControllerDriver->f_FM_MAC_DumpRegs)
-+ return p_FmMacControllerDriver->f_FM_MAC_DumpRegs(h_FmMac);
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, NO_MSG);
-+}
-+#endif /* (defined(DEBUG_ERRORS) && ... */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MAC/fm_mac.h
-@@ -0,0 +1,226 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/******************************************************************************
-+ @File fm_mac.h
-+
-+ @Description FM MAC ...
-+*//***************************************************************************/
-+#ifndef __FM_MAC_H
-+#define __FM_MAC_H
-+
-+#include "std_ext.h"
-+#include "error_ext.h"
-+#include "list_ext.h"
-+#include "fm_mac_ext.h"
-+#include "fm_common.h"
-+
-+
-+#define __ERR_MODULE__ MODULE_FM_MAC
-+
-+/**************************************************************************//**
-+ @Description defaults
-+*//***************************************************************************/
-+
-+
-+#define DEFAULT_halfDuplex FALSE
-+#define DEFAULT_padAndCrcEnable TRUE
-+#define DEFAULT_resetOnInit FALSE
-+
-+
-+typedef struct {
-+ uint64_t addr; /* Ethernet Address */
-+ t_List node;
-+} t_EthHashEntry;
-+#define ETH_HASH_ENTRY_OBJ(ptr) LIST_OBJECT(ptr, t_EthHashEntry, node)
-+
-+typedef struct {
-+ uint16_t size;
-+ t_List *p_Lsts;
-+} t_EthHash;
-+
-+typedef struct {
-+ t_Error (*f_FM_MAC_Init) (t_Handle h_FmMac);
-+ t_Error (*f_FM_MAC_Free) (t_Handle h_FmMac);
-+
-+ t_Error (*f_FM_MAC_SetStatistics) (t_Handle h_FmMac, e_FmMacStatisticsLevel statisticsLevel);
-+ t_Error (*f_FM_MAC_ConfigLoopback) (t_Handle h_FmMac, bool newVal);
-+ t_Error (*f_FM_MAC_ConfigMaxFrameLength) (t_Handle h_FmMac, uint16_t newVal);
-+ t_Error (*f_FM_MAC_ConfigWan) (t_Handle h_FmMac, bool flag);
-+ t_Error (*f_FM_MAC_ConfigPadAndCrc) (t_Handle h_FmMac, bool newVal);
-+ t_Error (*f_FM_MAC_ConfigHalfDuplex) (t_Handle h_FmMac, bool newVal);
-+ t_Error (*f_FM_MAC_ConfigLengthCheck) (t_Handle h_FmMac, bool newVal);
-+ t_Error (*f_FM_MAC_ConfigTbiPhyAddr) (t_Handle h_FmMac, uint8_t newVal);
-+ t_Error (*f_FM_MAC_ConfigException) (t_Handle h_FmMac, e_FmMacExceptions, bool enable);
-+ t_Error (*f_FM_MAC_ConfigResetOnInit) (t_Handle h_FmMac, bool enable);
-+#ifdef FM_TX_ECC_FRMS_ERRATA_10GMAC_A004
-+ t_Error (*f_FM_MAC_ConfigSkipFman11Workaround) (t_Handle h_FmMac);
-+#endif /* FM_TX_ECC_FRMS_ERRATA_10GMAC_A004 */
-+
-+ t_Error (*f_FM_MAC_SetException) (t_Handle h_FmMac, e_FmMacExceptions ex, bool enable);
-+
-+ t_Error (*f_FM_MAC_Enable) (t_Handle h_FmMac, e_CommMode mode);
-+ t_Error (*f_FM_MAC_Disable) (t_Handle h_FmMac, e_CommMode mode);
-+ t_Error (*f_FM_MAC_Resume) (t_Handle h_FmMac);
-+ t_Error (*f_FM_MAC_Enable1588TimeStamp) (t_Handle h_FmMac);
-+ t_Error (*f_FM_MAC_Disable1588TimeStamp) (t_Handle h_FmMac);
-+ t_Error (*f_FM_MAC_Reset) (t_Handle h_FmMac, bool wait);
-+
-+ t_Error (*f_FM_MAC_SetTxAutoPauseFrames) (t_Handle h_FmMac,
-+ uint16_t pauseTime);
-+ t_Error (*f_FM_MAC_SetTxPauseFrames) (t_Handle h_FmMac,
-+ uint8_t priority,
-+ uint16_t pauseTime,
-+ uint16_t threshTime);
-+ t_Error (*f_FM_MAC_SetRxIgnorePauseFrames) (t_Handle h_FmMac, bool en);
-+
-+ t_Error (*f_FM_MAC_ResetCounters) (t_Handle h_FmMac);
-+ t_Error (*f_FM_MAC_GetStatistics) (t_Handle h_FmMac, t_FmMacStatistics *p_Statistics);
-+ t_Error (*f_FM_MAC_GetFrameSizeCounters) (t_Handle h_FmMac, t_FmMacFrameSizeCounters *p_FrameSizeCounters, e_CommMode type);
-+
-+ t_Error (*f_FM_MAC_ModifyMacAddr) (t_Handle h_FmMac, t_EnetAddr *p_EnetAddr);
-+ t_Error (*f_FM_MAC_AddHashMacAddr) (t_Handle h_FmMac, t_EnetAddr *p_EnetAddr);
-+ t_Error (*f_FM_MAC_RemoveHashMacAddr) (t_Handle h_FmMac, t_EnetAddr *p_EnetAddr);
-+ t_Error (*f_FM_MAC_AddExactMatchMacAddr) (t_Handle h_FmMac, t_EnetAddr *p_EnetAddr);
-+ t_Error (*f_FM_MAC_RemovelExactMatchMacAddr) (t_Handle h_FmMac, t_EnetAddr *p_EnetAddr);
-+
-+ t_Error (*f_FM_MAC_SetPromiscuous) (t_Handle h_FmMac, bool newVal);
-+ t_Error (*f_FM_MAC_AdjustLink) (t_Handle h_FmMac, e_EnetSpeed speed, bool fullDuplex);
-+ t_Error (*f_FM_MAC_RestartAutoneg) (t_Handle h_FmMac);
-+
-+ t_Error (*f_FM_MAC_SetWakeOnLan) (t_Handle h_FmMac, bool en);
-+
-+ t_Error (*f_FM_MAC_GetId) (t_Handle h_FmMac, uint32_t *macId);
-+
-+ t_Error (*f_FM_MAC_GetVersion) (t_Handle h_FmMac, uint32_t *macVersion);
-+
-+ uint16_t (*f_FM_MAC_GetMaxFrameLength) (t_Handle h_FmMac);
-+
-+ t_Error (*f_FM_MAC_MII_WritePhyReg)(t_Handle h_FmMac, uint8_t phyAddr, uint8_t reg, uint16_t data);
-+ t_Error (*f_FM_MAC_MII_ReadPhyReg)(t_Handle h_FmMac, uint8_t phyAddr, uint8_t reg, uint16_t *p_Data);
-+
-+#if (defined(DEBUG_ERRORS) && (DEBUG_ERRORS > 0))
-+ t_Error (*f_FM_MAC_DumpRegs) (t_Handle h_FmMac);
-+#endif /* (defined(DEBUG_ERRORS) && ... */
-+
-+ t_Handle h_Fm;
-+ t_FmRevisionInfo fmRevInfo;
-+ e_EnetMode enetMode;
-+ uint8_t macId;
-+ bool resetOnInit;
-+ uint16_t clkFreq;
-+} t_FmMacControllerDriver;
-+
-+
-+#if (DPAA_VERSION == 10)
-+t_Handle DTSEC_Config(t_FmMacParams *p_FmMacParam);
-+t_Handle TGEC_Config(t_FmMacParams *p_FmMacParams);
-+#else
-+t_Handle MEMAC_Config(t_FmMacParams *p_FmMacParam);
-+#endif /* (DPAA_VERSION == 10) */
-+uint16_t FM_MAC_GetMaxFrameLength(t_Handle FmMac);
-+
-+
-+/* ........................................................................... */
-+
-+static __inline__ t_EthHashEntry *DequeueAddrFromHashEntry(t_List *p_AddrLst)
-+{
-+ t_EthHashEntry *p_HashEntry = NULL;
-+ if (!LIST_IsEmpty(p_AddrLst))
-+ {
-+ p_HashEntry = ETH_HASH_ENTRY_OBJ(p_AddrLst->p_Next);
-+ LIST_DelAndInit(&p_HashEntry->node);
-+ }
-+ return p_HashEntry;
-+}
-+
-+/* ........................................................................... */
-+
-+static __inline__ void FreeHashTable(t_EthHash *p_Hash)
-+{
-+ t_EthHashEntry *p_HashEntry;
-+ int i = 0;
-+
-+ if (p_Hash)
-+ {
-+ if (p_Hash->p_Lsts)
-+ {
-+ for (i=0; i<p_Hash->size; i++)
-+ {
-+ p_HashEntry = DequeueAddrFromHashEntry(&p_Hash->p_Lsts[i]);
-+ while (p_HashEntry)
-+ {
-+ XX_Free(p_HashEntry);
-+ p_HashEntry = DequeueAddrFromHashEntry(&p_Hash->p_Lsts[i]);
-+ }
-+ }
-+
-+ XX_Free(p_Hash->p_Lsts);
-+ }
-+
-+ XX_Free(p_Hash);
-+ }
-+}
-+
-+/* ........................................................................... */
-+
-+static __inline__ t_EthHash * AllocHashTable(uint16_t size)
-+{
-+ uint32_t i;
-+ t_EthHash *p_Hash;
-+
-+ /* Allocate address hash table */
-+ p_Hash = (t_EthHash *)XX_Malloc(sizeof(t_EthHash));
-+ if (!p_Hash)
-+ {
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("Address hash table"));
-+ return NULL;
-+ }
-+ p_Hash->size = size;
-+
-+ p_Hash->p_Lsts = (t_List *)XX_Malloc(p_Hash->size*sizeof(t_List));
-+ if (!p_Hash->p_Lsts)
-+ {
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("Address hash table"));
-+ XX_Free(p_Hash);
-+ return NULL;
-+ }
-+
-+ for (i=0 ; i<p_Hash->size; i++)
-+ INIT_LIST(&p_Hash->p_Lsts[i]);
-+
-+ return p_Hash;
-+}
-+
-+
-+#endif /* __FM_MAC_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MAC/fman_crc32.c
-@@ -0,0 +1,119 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+#include "fman_crc32.h"
-+#include "common/general.h"
-+
-+
-+/* precomputed CRC values for address hashing */
-+static const uint32_t crc_tbl[256] = {
-+ 0x00000000, 0x77073096, 0xee0e612c, 0x990951ba, 0x076dc419, 0x706af48f,
-+ 0xe963a535, 0x9e6495a3, 0x0edb8832, 0x79dcb8a4, 0xe0d5e91e, 0x97d2d988,
-+ 0x09b64c2b, 0x7eb17cbd, 0xe7b82d07, 0x90bf1d91, 0x1db71064, 0x6ab020f2,
-+ 0xf3b97148, 0x84be41de, 0x1adad47d, 0x6ddde4eb, 0xf4d4b551, 0x83d385c7,
-+ 0x136c9856, 0x646ba8c0, 0xfd62f97a, 0x8a65c9ec, 0x14015c4f, 0x63066cd9,
-+ 0xfa0f3d63, 0x8d080df5, 0x3b6e20c8, 0x4c69105e, 0xd56041e4, 0xa2677172,
-+ 0x3c03e4d1, 0x4b04d447, 0xd20d85fd, 0xa50ab56b, 0x35b5a8fa, 0x42b2986c,
-+ 0xdbbbc9d6, 0xacbcf940, 0x32d86ce3, 0x45df5c75, 0xdcd60dcf, 0xabd13d59,
-+ 0x26d930ac, 0x51de003a, 0xc8d75180, 0xbfd06116, 0x21b4f4b5, 0x56b3c423,
-+ 0xcfba9599, 0xb8bda50f, 0x2802b89e, 0x5f058808, 0xc60cd9b2, 0xb10be924,
-+ 0x2f6f7c87, 0x58684c11, 0xc1611dab, 0xb6662d3d, 0x76dc4190, 0x01db7106,
-+ 0x98d220bc, 0xefd5102a, 0x71b18589, 0x06b6b51f, 0x9fbfe4a5, 0xe8b8d433,
-+ 0x7807c9a2, 0x0f00f934, 0x9609a88e, 0xe10e9818, 0x7f6a0dbb, 0x086d3d2d,
-+ 0x91646c97, 0xe6635c01, 0x6b6b51f4, 0x1c6c6162, 0x856530d8, 0xf262004e,
-+ 0x6c0695ed, 0x1b01a57b, 0x8208f4c1, 0xf50fc457, 0x65b0d9c6, 0x12b7e950,
-+ 0x8bbeb8ea, 0xfcb9887c, 0x62dd1ddf, 0x15da2d49, 0x8cd37cf3, 0xfbd44c65,
-+ 0x4db26158, 0x3ab551ce, 0xa3bc0074, 0xd4bb30e2, 0x4adfa541, 0x3dd895d7,
-+ 0xa4d1c46d, 0xd3d6f4fb, 0x4369e96a, 0x346ed9fc, 0xad678846, 0xda60b8d0,
-+ 0x44042d73, 0x33031de5, 0xaa0a4c5f, 0xdd0d7cc9, 0x5005713c, 0x270241aa,
-+ 0xbe0b1010, 0xc90c2086, 0x5768b525, 0x206f85b3, 0xb966d409, 0xce61e49f,
-+ 0x5edef90e, 0x29d9c998, 0xb0d09822, 0xc7d7a8b4, 0x59b33d17, 0x2eb40d81,
-+ 0xb7bd5c3b, 0xc0ba6cad, 0xedb88320, 0x9abfb3b6, 0x03b6e20c, 0x74b1d29a,
-+ 0xead54739, 0x9dd277af, 0x04db2615, 0x73dc1683, 0xe3630b12, 0x94643b84,
-+ 0x0d6d6a3e, 0x7a6a5aa8, 0xe40ecf0b, 0x9309ff9d, 0x0a00ae27, 0x7d079eb1,
-+ 0xf00f9344, 0x8708a3d2, 0x1e01f268, 0x6906c2fe, 0xf762575d, 0x806567cb,
-+ 0x196c3671, 0x6e6b06e7, 0xfed41b76, 0x89d32be0, 0x10da7a5a, 0x67dd4acc,
-+ 0xf9b9df6f, 0x8ebeeff9, 0x17b7be43, 0x60b08ed5, 0xd6d6a3e8, 0xa1d1937e,
-+ 0x38d8c2c4, 0x4fdff252, 0xd1bb67f1, 0xa6bc5767, 0x3fb506dd, 0x48b2364b,
-+ 0xd80d2bda, 0xaf0a1b4c, 0x36034af6, 0x41047a60, 0xdf60efc3, 0xa867df55,
-+ 0x316e8eef, 0x4669be79, 0xcb61b38c, 0xbc66831a, 0x256fd2a0, 0x5268e236,
-+ 0xcc0c7795, 0xbb0b4703, 0x220216b9, 0x5505262f, 0xc5ba3bbe, 0xb2bd0b28,
-+ 0x2bb45a92, 0x5cb36a04, 0xc2d7ffa7, 0xb5d0cf31, 0x2cd99e8b, 0x5bdeae1d,
-+ 0x9b64c2b0, 0xec63f226, 0x756aa39c, 0x026d930a, 0x9c0906a9, 0xeb0e363f,
-+ 0x72076785, 0x05005713, 0x95bf4a82, 0xe2b87a14, 0x7bb12bae, 0x0cb61b38,
-+ 0x92d28e9b, 0xe5d5be0d, 0x7cdcefb7, 0x0bdbdf21, 0x86d3d2d4, 0xf1d4e242,
-+ 0x68ddb3f8, 0x1fda836e, 0x81be16cd, 0xf6b9265b, 0x6fb077e1, 0x18b74777,
-+ 0x88085ae6, 0xff0f6a70, 0x66063bca, 0x11010b5c, 0x8f659eff, 0xf862ae69,
-+ 0x616bffd3, 0x166ccf45, 0xa00ae278, 0xd70dd2ee, 0x4e048354, 0x3903b3c2,
-+ 0xa7672661, 0xd06016f7, 0x4969474d, 0x3e6e77db, 0xaed16a4a, 0xd9d65adc,
-+ 0x40df0b66, 0x37d83bf0, 0xa9bcae53, 0xdebb9ec5, 0x47b2cf7f, 0x30b5ffe9,
-+ 0xbdbdf21c, 0xcabac28a, 0x53b39330, 0x24b4a3a6, 0xbad03605, 0xcdd70693,
-+ 0x54de5729, 0x23d967bf, 0xb3667a2e, 0xc4614ab8, 0x5d681b02, 0x2a6f2b94,
-+ 0xb40bbe37, 0xc30c8ea1, 0x5a05df1b, 0x2d02ef8d
-+};
-+
-+/* Get the mirrored value of a byte size number. (0x11010011 --> 0x11001011) */
-+static inline uint8_t get_mirror8(uint8_t n)
-+{
-+ uint8_t mirror[16] = {
-+ 0x00, 0x08, 0x04, 0x0c, 0x02, 0x0a, 0x06, 0x0e,
-+ 0x01, 0x09, 0x05, 0x0d, 0x03, 0x0b, 0x07, 0x0f
-+ };
-+ return (uint8_t)(((mirror[n & 0x0f] << 4) | (mirror[n >> 4])));
-+}
-+
-+static inline uint32_t get_mirror32(uint32_t n)
-+{
-+ return ((uint32_t)get_mirror8((uint8_t)(n))<<24) |
-+ ((uint32_t)get_mirror8((uint8_t)(n>>8))<<16) |
-+ ((uint32_t)get_mirror8((uint8_t)(n>>16))<<8) |
-+ ((uint32_t)get_mirror8((uint8_t)(n>>24)));
-+}
-+
-+uint32_t get_mac_addr_crc(uint64_t _addr)
-+{
-+ uint32_t i;
-+ uint8_t data;
-+ uint32_t crc;
-+
-+ /* CRC calculation */
-+ crc = 0xffffffff;
-+ for (i = 0; i < 6; i++) {
-+ data = (uint8_t)(_addr >> ((5-i)*8));
-+ crc = crc ^ data;
-+ crc = crc_tbl[crc&0xff] ^ (crc>>8);
-+ }
-+
-+ crc = get_mirror32(crc);
-+ return crc;
-+}
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MAC/fman_crc32.h
-@@ -0,0 +1,43 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+#ifndef __FMAN_CRC32_H
-+#define __FMAN_CRC32_H
-+
-+#include "common/general.h"
-+
-+
-+uint32_t get_mac_addr_crc(uint64_t _addr);
-+
-+
-+#endif /* __FMAN_CRC32_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MAC/fman_dtsec.c
-@@ -0,0 +1,845 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+#include "fsl_fman_dtsec.h"
-+
-+
-+void fman_dtsec_stop_rx(struct dtsec_regs *regs)
-+{
-+ /* Assert the graceful stop bit */
-+ iowrite32be(ioread32be(&regs->rctrl) | RCTRL_GRS, &regs->rctrl);
-+}
-+
-+void fman_dtsec_stop_tx(struct dtsec_regs *regs)
-+{
-+ /* Assert the graceful stop bit */
-+ iowrite32be(ioread32be(&regs->tctrl) | DTSEC_TCTRL_GTS, &regs->tctrl);
-+}
-+
-+void fman_dtsec_start_tx(struct dtsec_regs *regs)
-+{
-+ /* clear the graceful stop bit */
-+ iowrite32be(ioread32be(&regs->tctrl) & ~DTSEC_TCTRL_GTS, &regs->tctrl);
-+}
-+
-+void fman_dtsec_start_rx(struct dtsec_regs *regs)
-+{
-+ /* clear the graceful stop bit */
-+ iowrite32be(ioread32be(&regs->rctrl) & ~RCTRL_GRS, &regs->rctrl);
-+}
-+
-+void fman_dtsec_defconfig(struct dtsec_cfg *cfg)
-+{
-+ cfg->halfdup_on = DEFAULT_HALFDUP_ON;
-+ cfg->halfdup_retransmit = DEFAULT_HALFDUP_RETRANSMIT;
-+ cfg->halfdup_coll_window = DEFAULT_HALFDUP_COLL_WINDOW;
-+ cfg->halfdup_excess_defer = DEFAULT_HALFDUP_EXCESS_DEFER;
-+ cfg->halfdup_no_backoff = DEFAULT_HALFDUP_NO_BACKOFF;
-+ cfg->halfdup_bp_no_backoff = DEFAULT_HALFDUP_BP_NO_BACKOFF;
-+ cfg->halfdup_alt_backoff_val = DEFAULT_HALFDUP_ALT_BACKOFF_VAL;
-+ cfg->halfdup_alt_backoff_en = DEFAULT_HALFDUP_ALT_BACKOFF_EN;
-+ cfg->rx_drop_bcast = DEFAULT_RX_DROP_BCAST;
-+ cfg->rx_short_frm = DEFAULT_RX_SHORT_FRM;
-+ cfg->rx_len_check = DEFAULT_RX_LEN_CHECK;
-+ cfg->tx_pad_crc = DEFAULT_TX_PAD_CRC;
-+ cfg->tx_crc = DEFAULT_TX_CRC;
-+ cfg->rx_ctrl_acc = DEFAULT_RX_CTRL_ACC;
-+ cfg->tx_pause_time = DEFAULT_TX_PAUSE_TIME;
-+ cfg->tbipa = DEFAULT_TBIPA; /* PHY address 0 is reserved (DPAA RM)*/
-+ cfg->rx_prepend = DEFAULT_RX_PREPEND;
-+ cfg->ptp_tsu_en = DEFAULT_PTP_TSU_EN;
-+ cfg->ptp_exception_en = DEFAULT_PTP_EXCEPTION_EN;
-+ cfg->preamble_len = DEFAULT_PREAMBLE_LEN;
-+ cfg->rx_preamble = DEFAULT_RX_PREAMBLE;
-+ cfg->tx_preamble = DEFAULT_TX_PREAMBLE;
-+ cfg->loopback = DEFAULT_LOOPBACK;
-+ cfg->rx_time_stamp_en = DEFAULT_RX_TIME_STAMP_EN;
-+ cfg->tx_time_stamp_en = DEFAULT_TX_TIME_STAMP_EN;
-+ cfg->rx_flow = DEFAULT_RX_FLOW;
-+ cfg->tx_flow = DEFAULT_TX_FLOW;
-+ cfg->rx_group_hash_exd = DEFAULT_RX_GROUP_HASH_EXD;
-+ cfg->tx_pause_time_extd = DEFAULT_TX_PAUSE_TIME_EXTD;
-+ cfg->rx_promisc = DEFAULT_RX_PROMISC;
-+ cfg->non_back_to_back_ipg1 = DEFAULT_NON_BACK_TO_BACK_IPG1;
-+ cfg->non_back_to_back_ipg2 = DEFAULT_NON_BACK_TO_BACK_IPG2;
-+ cfg->min_ifg_enforcement = DEFAULT_MIN_IFG_ENFORCEMENT;
-+ cfg->back_to_back_ipg = DEFAULT_BACK_TO_BACK_IPG;
-+ cfg->maximum_frame = DEFAULT_MAXIMUM_FRAME;
-+ cfg->tbi_phy_addr = DEFAULT_TBI_PHY_ADDR;
-+ cfg->wake_on_lan = DEFAULT_WAKE_ON_LAN;
-+}
-+
-+int fman_dtsec_init(struct dtsec_regs *regs, struct dtsec_cfg *cfg,
-+ enum enet_interface iface_mode,
-+ enum enet_speed iface_speed,
-+ uint8_t *macaddr,
-+ uint8_t fm_rev_maj,
-+ uint8_t fm_rev_min,
-+ uint32_t exception_mask)
-+{
-+ bool is_rgmii = FALSE;
-+ bool is_sgmii = FALSE;
-+ bool is_qsgmii = FALSE;
-+ int i;
-+ uint32_t tmp;
-+
-+UNUSED(fm_rev_maj);UNUSED(fm_rev_min);
-+
-+ /* let's start with a soft reset */
-+ iowrite32be(MACCFG1_SOFT_RESET, &regs->maccfg1);
-+ iowrite32be(0, &regs->maccfg1);
-+
-+ /*************dtsec_id2******************/
-+ tmp = ioread32be(&regs->tsec_id2);
-+
-+ /* check RGMII support */
-+ if (iface_mode == E_ENET_IF_RGMII ||
-+ iface_mode == E_ENET_IF_RMII)
-+ if (tmp & DTSEC_ID2_INT_REDUCED_OFF)
-+ return -EINVAL;
-+
-+ if (iface_mode == E_ENET_IF_SGMII ||
-+ iface_mode == E_ENET_IF_MII)
-+ if (tmp & DTSEC_ID2_INT_REDUCED_OFF)
-+ return -EINVAL;
-+
-+ /***************ECNTRL************************/
-+
-+ is_rgmii = (bool)((iface_mode == E_ENET_IF_RGMII) ? TRUE : FALSE);
-+ is_sgmii = (bool)((iface_mode == E_ENET_IF_SGMII) ? TRUE : FALSE);
-+ is_qsgmii = (bool)((iface_mode == E_ENET_IF_QSGMII) ? TRUE : FALSE);
-+
-+ tmp = 0;
-+ if (is_rgmii || iface_mode == E_ENET_IF_GMII)
-+ tmp |= DTSEC_ECNTRL_GMIIM;
-+ if (is_sgmii)
-+ tmp |= (DTSEC_ECNTRL_SGMIIM | DTSEC_ECNTRL_TBIM);
-+ if (is_qsgmii)
-+ tmp |= (DTSEC_ECNTRL_SGMIIM | DTSEC_ECNTRL_TBIM |
-+ DTSEC_ECNTRL_QSGMIIM);
-+ if (is_rgmii)
-+ tmp |= DTSEC_ECNTRL_RPM;
-+ if (iface_speed == E_ENET_SPEED_100)
-+ tmp |= DTSEC_ECNTRL_R100M;
-+
-+ iowrite32be(tmp, &regs->ecntrl);
-+ /***************ECNTRL************************/
-+
-+ /***************TCTRL************************/
-+ tmp = 0;
-+ if (cfg->halfdup_on)
-+ tmp |= DTSEC_TCTRL_THDF;
-+ if (cfg->tx_time_stamp_en)
-+ tmp |= DTSEC_TCTRL_TTSE;
-+
-+ iowrite32be(tmp, &regs->tctrl);
-+
-+ /***************TCTRL************************/
-+
-+ /***************PTV************************/
-+ tmp = 0;
-+
-+#ifdef FM_SHORT_PAUSE_TIME_ERRATA_DTSEC1
-+ if ((fm_rev_maj == 1) && (fm_rev_min == 0))
-+ cfg->tx_pause_time += 2;
-+#endif /* FM_SHORT_PAUSE_TIME_ERRATA_DTSEC1 */
-+
-+ if (cfg->tx_pause_time)
-+ tmp |= cfg->tx_pause_time;
-+ if (cfg->tx_pause_time_extd)
-+ tmp |= cfg->tx_pause_time_extd << PTV_PTE_OFST;
-+ iowrite32be(tmp, &regs->ptv);
-+
-+ /***************RCTRL************************/
-+ tmp = 0;
-+ tmp |= ((uint32_t)(cfg->rx_prepend & 0x0000001f)) << 16;
-+ if (cfg->rx_ctrl_acc)
-+ tmp |= RCTRL_CFA;
-+ if (cfg->rx_group_hash_exd)
-+ tmp |= RCTRL_GHTX;
-+ if (cfg->rx_time_stamp_en)
-+ tmp |= RCTRL_RTSE;
-+ if (cfg->rx_drop_bcast)
-+ tmp |= RCTRL_BC_REJ;
-+ if (cfg->rx_short_frm)
-+ tmp |= RCTRL_RSF;
-+ if (cfg->rx_promisc)
-+ tmp |= RCTRL_PROM;
-+
-+ iowrite32be(tmp, &regs->rctrl);
-+ /***************RCTRL************************/
-+
-+ /*
-+ * Assign a Phy Address to the TBI (TBIPA).
-+ * Done also in cases where TBI is not selected to avoid conflict with
-+ * the external PHY's Physical address
-+ */
-+ iowrite32be(cfg->tbipa, &regs->tbipa);
-+
-+ /***************TMR_CTL************************/
-+ iowrite32be(0, &regs->tmr_ctrl);
-+
-+ if (cfg->ptp_tsu_en) {
-+ tmp = 0;
-+ tmp |= TMR_PEVENT_TSRE;
-+ iowrite32be(tmp, &regs->tmr_pevent);
-+
-+ if (cfg->ptp_exception_en) {
-+ tmp = 0;
-+ tmp |= TMR_PEMASK_TSREEN;
-+ iowrite32be(tmp, &regs->tmr_pemask);
-+ }
-+ }
-+
-+ /***************MACCFG1***********************/
-+ tmp = 0;
-+ if (cfg->loopback)
-+ tmp |= MACCFG1_LOOPBACK;
-+ if (cfg->rx_flow)
-+ tmp |= MACCFG1_RX_FLOW;
-+ if (cfg->tx_flow)
-+ tmp |= MACCFG1_TX_FLOW;
-+ iowrite32be(tmp, &regs->maccfg1);
-+
-+ /***************MACCFG1***********************/
-+
-+ /***************MACCFG2***********************/
-+ tmp = 0;
-+
-+ if (iface_speed < E_ENET_SPEED_1000)
-+ tmp |= MACCFG2_NIBBLE_MODE;
-+ else if (iface_speed == E_ENET_SPEED_1000)
-+ tmp |= MACCFG2_BYTE_MODE;
-+
-+ tmp |= ((uint32_t) cfg->preamble_len & 0x0000000f)
-+ << PREAMBLE_LENGTH_SHIFT;
-+
-+ if (cfg->rx_preamble)
-+ tmp |= MACCFG2_PRE_AM_Rx_EN;
-+ if (cfg->tx_preamble)
-+ tmp |= MACCFG2_PRE_AM_Tx_EN;
-+ if (cfg->rx_len_check)
-+ tmp |= MACCFG2_LENGTH_CHECK;
-+ if (cfg->tx_pad_crc)
-+ tmp |= MACCFG2_PAD_CRC_EN;
-+ if (cfg->tx_crc)
-+ tmp |= MACCFG2_CRC_EN;
-+ if (!cfg->halfdup_on)
-+ tmp |= MACCFG2_FULL_DUPLEX;
-+ iowrite32be(tmp, &regs->maccfg2);
-+
-+ /***************MACCFG2***********************/
-+
-+ /***************IPGIFG************************/
-+ tmp = (((cfg->non_back_to_back_ipg1 <<
-+ IPGIFG_NON_BACK_TO_BACK_IPG_1_SHIFT)
-+ & IPGIFG_NON_BACK_TO_BACK_IPG_1)
-+ | ((cfg->non_back_to_back_ipg2 <<
-+ IPGIFG_NON_BACK_TO_BACK_IPG_2_SHIFT)
-+ & IPGIFG_NON_BACK_TO_BACK_IPG_2)
-+ | ((cfg->min_ifg_enforcement <<
-+ IPGIFG_MIN_IFG_ENFORCEMENT_SHIFT)
-+ & IPGIFG_MIN_IFG_ENFORCEMENT)
-+ | (cfg->back_to_back_ipg & IPGIFG_BACK_TO_BACK_IPG));
-+ iowrite32be(tmp, &regs->ipgifg);
-+
-+ /***************IPGIFG************************/
-+
-+ /***************HAFDUP************************/
-+ tmp = 0;
-+
-+ if (cfg->halfdup_alt_backoff_en)
-+ tmp = (uint32_t)(HAFDUP_ALT_BEB |
-+ ((cfg->halfdup_alt_backoff_val & 0x0000000f)
-+ << HAFDUP_ALTERNATE_BEB_TRUNCATION_SHIFT));
-+ if (cfg->halfdup_bp_no_backoff)
-+ tmp |= HAFDUP_BP_NO_BACKOFF;
-+ if (cfg->halfdup_no_backoff)
-+ tmp |= HAFDUP_NO_BACKOFF;
-+ if (cfg->halfdup_excess_defer)
-+ tmp |= HAFDUP_EXCESS_DEFER;
-+ tmp |= ((cfg->halfdup_retransmit << HAFDUP_RETRANSMISSION_MAX_SHIFT)
-+ & HAFDUP_RETRANSMISSION_MAX);
-+ tmp |= (cfg->halfdup_coll_window & HAFDUP_COLLISION_WINDOW);
-+
-+ iowrite32be(tmp, &regs->hafdup);
-+ /***************HAFDUP************************/
-+
-+ /***************MAXFRM************************/
-+ /* Initialize MAXFRM */
-+ iowrite32be(cfg->maximum_frame, &regs->maxfrm);
-+
-+ /***************MAXFRM************************/
-+
-+ /***************CAM1************************/
-+ iowrite32be(0xffffffff, &regs->cam1);
-+ iowrite32be(0xffffffff, &regs->cam2);
-+
-+ /***************IMASK************************/
-+ iowrite32be(exception_mask, &regs->imask);
-+ /***************IMASK************************/
-+
-+ /***************IEVENT************************/
-+ iowrite32be(0xffffffff, &regs->ievent);
-+
-+ /***************MACSTNADDR1/2*****************/
-+
-+ tmp = (uint32_t)((macaddr[5] << 24) |
-+ (macaddr[4] << 16) |
-+ (macaddr[3] << 8) |
-+ macaddr[2]);
-+ iowrite32be(tmp, &regs->macstnaddr1);
-+
-+ tmp = (uint32_t)((macaddr[1] << 24) |
-+ (macaddr[0] << 16));
-+ iowrite32be(tmp, &regs->macstnaddr2);
-+
-+ /***************MACSTNADDR1/2*****************/
-+
-+ /*****************HASH************************/
-+ for (i = 0; i < NUM_OF_HASH_REGS ; i++) {
-+ /* Initialize IADDRx */
-+ iowrite32be(0, &regs->igaddr[i]);
-+ /* Initialize GADDRx */
-+ iowrite32be(0, &regs->gaddr[i]);
-+ }
-+
-+ fman_dtsec_reset_stat(regs);
-+
-+ return 0;
-+}
-+
-+uint16_t fman_dtsec_get_max_frame_len(struct dtsec_regs *regs)
-+{
-+ return (uint16_t)ioread32be(&regs->maxfrm);
-+}
-+
-+void fman_dtsec_set_max_frame_len(struct dtsec_regs *regs, uint16_t length)
-+{
-+ iowrite32be(length, &regs->maxfrm);
-+}
-+
-+void fman_dtsec_set_mac_address(struct dtsec_regs *regs, uint8_t *adr)
-+{
-+ uint32_t tmp;
-+
-+ tmp = (uint32_t)((adr[5] << 24) |
-+ (adr[4] << 16) |
-+ (adr[3] << 8) |
-+ adr[2]);
-+ iowrite32be(tmp, &regs->macstnaddr1);
-+
-+ tmp = (uint32_t)((adr[1] << 24) |
-+ (adr[0] << 16));
-+ iowrite32be(tmp, &regs->macstnaddr2);
-+}
-+
-+void fman_dtsec_get_mac_address(struct dtsec_regs *regs, uint8_t *macaddr)
-+{
-+ uint32_t tmp1, tmp2;
-+
-+ tmp1 = ioread32be(&regs->macstnaddr1);
-+ tmp2 = ioread32be(&regs->macstnaddr2);
-+
-+ macaddr[0] = (uint8_t)((tmp2 & 0x00ff0000) >> 16);
-+ macaddr[1] = (uint8_t)((tmp2 & 0xff000000) >> 24);
-+ macaddr[2] = (uint8_t)(tmp1 & 0x000000ff);
-+ macaddr[3] = (uint8_t)((tmp1 & 0x0000ff00) >> 8);
-+ macaddr[4] = (uint8_t)((tmp1 & 0x00ff0000) >> 16);
-+ macaddr[5] = (uint8_t)((tmp1 & 0xff000000) >> 24);
-+}
-+
-+void fman_dtsec_set_hash_table(struct dtsec_regs *regs, uint32_t crc, bool mcast, bool ghtx)
-+{
-+ int32_t bucket;
-+ if (ghtx)
-+ bucket = (int32_t)((crc >> 23) & 0x1ff);
-+ else {
-+ bucket = (int32_t)((crc >> 24) & 0xff);
-+ /* if !ghtx and mcast the bit must be set in gaddr instead of igaddr. */
-+ if (mcast)
-+ bucket += 0x100;
-+ }
-+ fman_dtsec_set_bucket(regs, bucket, TRUE);
-+}
-+
-+void fman_dtsec_set_bucket(struct dtsec_regs *regs, int bucket, bool enable)
-+{
-+ int reg_idx = (bucket >> 5) & 0xf;
-+ int bit_idx = bucket & 0x1f;
-+ uint32_t bit_mask = 0x80000000 >> bit_idx;
-+ uint32_t *reg;
-+
-+ if (reg_idx > 7)
-+ reg = &regs->gaddr[reg_idx-8];
-+ else
-+ reg = &regs->igaddr[reg_idx];
-+
-+ if (enable)
-+ iowrite32be(ioread32be(reg) | bit_mask, reg);
-+ else
-+ iowrite32be(ioread32be(reg) & (~bit_mask), reg);
-+}
-+
-+void fman_dtsec_reset_filter_table(struct dtsec_regs *regs, bool mcast, bool ucast)
-+{
-+ int i;
-+ bool ghtx;
-+
-+ ghtx = (bool)((ioread32be(&regs->rctrl) & RCTRL_GHTX) ? TRUE : FALSE);
-+
-+ if (ucast || (ghtx && mcast)) {
-+ for (i = 0; i < NUM_OF_HASH_REGS; i++)
-+ iowrite32be(0, &regs->igaddr[i]);
-+ }
-+ if (mcast) {
-+ for (i = 0; i < NUM_OF_HASH_REGS; i++)
-+ iowrite32be(0, &regs->gaddr[i]);
-+ }
-+}
-+
-+int fman_dtsec_set_tbi_phy_addr(struct dtsec_regs *regs,
-+ uint8_t addr)
-+{
-+ if (addr > 0 && addr < 32)
-+ iowrite32be(addr, &regs->tbipa);
-+ else
-+ return -EINVAL;
-+
-+ return 0;
-+}
-+
-+void fman_dtsec_set_wol(struct dtsec_regs *regs, bool en)
-+{
-+ uint32_t tmp;
-+
-+ tmp = ioread32be(&regs->maccfg2);
-+ if (en)
-+ tmp |= MACCFG2_MAGIC_PACKET_EN;
-+ else
-+ tmp &= ~MACCFG2_MAGIC_PACKET_EN;
-+ iowrite32be(tmp, &regs->maccfg2);
-+}
-+
-+int fman_dtsec_adjust_link(struct dtsec_regs *regs,
-+ enum enet_interface iface_mode,
-+ enum enet_speed speed, bool full_dx)
-+{
-+ uint32_t tmp;
-+
-+ UNUSED(iface_mode);
-+
-+ if ((speed == E_ENET_SPEED_1000) && !full_dx)
-+ return -EINVAL;
-+
-+ tmp = ioread32be(&regs->maccfg2);
-+ if (!full_dx)
-+ tmp &= ~MACCFG2_FULL_DUPLEX;
-+ else
-+ tmp |= MACCFG2_FULL_DUPLEX;
-+
-+ tmp &= ~(MACCFG2_NIBBLE_MODE | MACCFG2_BYTE_MODE);
-+ if (speed < E_ENET_SPEED_1000)
-+ tmp |= MACCFG2_NIBBLE_MODE;
-+ else if (speed == E_ENET_SPEED_1000)
-+ tmp |= MACCFG2_BYTE_MODE;
-+ iowrite32be(tmp, &regs->maccfg2);
-+
-+ tmp = ioread32be(&regs->ecntrl);
-+ if (speed == E_ENET_SPEED_100)
-+ tmp |= DTSEC_ECNTRL_R100M;
-+ else
-+ tmp &= ~DTSEC_ECNTRL_R100M;
-+ iowrite32be(tmp, &regs->ecntrl);
-+
-+ return 0;
-+}
-+
-+void fman_dtsec_set_uc_promisc(struct dtsec_regs *regs, bool enable)
-+{
-+ uint32_t tmp;
-+
-+ tmp = ioread32be(&regs->rctrl);
-+
-+ if (enable)
-+ tmp |= RCTRL_UPROM;
-+ else
-+ tmp &= ~RCTRL_UPROM;
-+
-+ iowrite32be(tmp, &regs->rctrl);
-+}
-+
-+void fman_dtsec_set_mc_promisc(struct dtsec_regs *regs, bool enable)
-+{
-+ uint32_t tmp;
-+
-+ tmp = ioread32be(&regs->rctrl);
-+
-+ if (enable)
-+ tmp |= RCTRL_MPROM;
-+ else
-+ tmp &= ~RCTRL_MPROM;
-+
-+ iowrite32be(tmp, &regs->rctrl);
-+}
-+
-+bool fman_dtsec_get_clear_carry_regs(struct dtsec_regs *regs,
-+ uint32_t *car1, uint32_t *car2)
-+{
-+ /* read carry registers */
-+ *car1 = ioread32be(&regs->car1);
-+ *car2 = ioread32be(&regs->car2);
-+ /* clear carry registers */
-+ if (*car1)
-+ iowrite32be(*car1, &regs->car1);
-+ if (*car2)
-+ iowrite32be(*car2, &regs->car2);
-+
-+ return (bool)((*car1 | *car2) ? TRUE : FALSE);
-+}
-+
-+void fman_dtsec_reset_stat(struct dtsec_regs *regs)
-+{
-+ /* clear HW counters */
-+ iowrite32be(ioread32be(&regs->ecntrl) |
-+ DTSEC_ECNTRL_CLRCNT, &regs->ecntrl);
-+}
-+
-+int fman_dtsec_set_stat_level(struct dtsec_regs *regs, enum dtsec_stat_level level)
-+{
-+ switch (level) {
-+ case E_MAC_STAT_NONE:
-+ iowrite32be(0xffffffff, &regs->cam1);
-+ iowrite32be(0xffffffff, &regs->cam2);
-+ iowrite32be(ioread32be(&regs->ecntrl) & ~DTSEC_ECNTRL_STEN,
-+ &regs->ecntrl);
-+ iowrite32be(ioread32be(&regs->imask) & ~DTSEC_IMASK_MSROEN,
-+ &regs->imask);
-+ break;
-+ case E_MAC_STAT_PARTIAL:
-+ iowrite32be(CAM1_ERRORS_ONLY, &regs->cam1);
-+ iowrite32be(CAM2_ERRORS_ONLY, &regs->cam2);
-+ iowrite32be(ioread32be(&regs->ecntrl) | DTSEC_ECNTRL_STEN,
-+ &regs->ecntrl);
-+ iowrite32be(ioread32be(&regs->imask) | DTSEC_IMASK_MSROEN,
-+ &regs->imask);
-+ break;
-+ case E_MAC_STAT_MIB_GRP1:
-+ iowrite32be((uint32_t)~CAM1_MIB_GRP_1, &regs->cam1);
-+ iowrite32be((uint32_t)~CAM2_MIB_GRP_1, &regs->cam2);
-+ iowrite32be(ioread32be(&regs->ecntrl) | DTSEC_ECNTRL_STEN,
-+ &regs->ecntrl);
-+ iowrite32be(ioread32be(&regs->imask) | DTSEC_IMASK_MSROEN,
-+ &regs->imask);
-+ break;
-+ case E_MAC_STAT_FULL:
-+ iowrite32be(0, &regs->cam1);
-+ iowrite32be(0, &regs->cam2);
-+ iowrite32be(ioread32be(&regs->ecntrl) | DTSEC_ECNTRL_STEN,
-+ &regs->ecntrl);
-+ iowrite32be(ioread32be(&regs->imask) | DTSEC_IMASK_MSROEN,
-+ &regs->imask);
-+ break;
-+ default:
-+ return -EINVAL;
-+ }
-+
-+ return 0;
-+}
-+
-+void fman_dtsec_set_ts(struct dtsec_regs *regs, bool en)
-+{
-+ if (en) {
-+ iowrite32be(ioread32be(&regs->rctrl) | RCTRL_RTSE,
-+ &regs->rctrl);
-+ iowrite32be(ioread32be(&regs->tctrl) | DTSEC_TCTRL_TTSE,
-+ &regs->tctrl);
-+ } else {
-+ iowrite32be(ioread32be(&regs->rctrl) & ~RCTRL_RTSE,
-+ &regs->rctrl);
-+ iowrite32be(ioread32be(&regs->tctrl) & ~DTSEC_TCTRL_TTSE,
-+ &regs->tctrl);
-+ }
-+}
-+
-+void fman_dtsec_enable(struct dtsec_regs *regs, bool apply_rx, bool apply_tx)
-+{
-+ uint32_t tmp;
-+
-+ tmp = ioread32be(&regs->maccfg1);
-+
-+ if (apply_rx)
-+ tmp |= MACCFG1_RX_EN ;
-+
-+ if (apply_tx)
-+ tmp |= MACCFG1_TX_EN ;
-+
-+ iowrite32be(tmp, &regs->maccfg1);
-+}
-+
-+void fman_dtsec_clear_addr_in_paddr(struct dtsec_regs *regs, uint8_t paddr_num)
-+{
-+ iowrite32be(0, &regs->macaddr[paddr_num].exact_match1);
-+ iowrite32be(0, &regs->macaddr[paddr_num].exact_match2);
-+}
-+
-+void fman_dtsec_add_addr_in_paddr(struct dtsec_regs *regs,
-+ uint64_t addr,
-+ uint8_t paddr_num)
-+{
-+ uint32_t tmp;
-+
-+ tmp = (uint32_t)(addr);
-+ /* swap */
-+ tmp = (((tmp & 0x000000FF) << 24) |
-+ ((tmp & 0x0000FF00) << 8) |
-+ ((tmp & 0x00FF0000) >> 8) |
-+ ((tmp & 0xFF000000) >> 24));
-+ iowrite32be(tmp, &regs->macaddr[paddr_num].exact_match1);
-+
-+ tmp = (uint32_t)(addr>>32);
-+ /* swap */
-+ tmp = (((tmp & 0x000000FF) << 24) |
-+ ((tmp & 0x0000FF00) << 8) |
-+ ((tmp & 0x00FF0000) >> 8) |
-+ ((tmp & 0xFF000000) >> 24));
-+ iowrite32be(tmp, &regs->macaddr[paddr_num].exact_match2);
-+}
-+
-+void fman_dtsec_disable(struct dtsec_regs *regs, bool apply_rx, bool apply_tx)
-+{
-+ uint32_t tmp;
-+
-+ tmp = ioread32be(&regs->maccfg1);
-+
-+ if (apply_rx)
-+ tmp &= ~MACCFG1_RX_EN;
-+
-+ if (apply_tx)
-+ tmp &= ~MACCFG1_TX_EN;
-+
-+ iowrite32be(tmp, &regs->maccfg1);
-+}
-+
-+void fman_dtsec_set_tx_pause_frames(struct dtsec_regs *regs, uint16_t time)
-+{
-+ uint32_t ptv = 0;
-+
-+ /* fixme: don't enable tx pause for half-duplex */
-+
-+ if (time) {
-+ ptv = ioread32be(&regs->ptv);
-+ ptv &= 0xffff0000;
-+ ptv |= time & 0x0000ffff;
-+ iowrite32be(ptv, &regs->ptv);
-+
-+ /* trigger the transmission of a flow-control pause frame */
-+ iowrite32be(ioread32be(&regs->maccfg1) | MACCFG1_TX_FLOW,
-+ &regs->maccfg1);
-+ } else
-+ iowrite32be(ioread32be(&regs->maccfg1) & ~MACCFG1_TX_FLOW,
-+ &regs->maccfg1);
-+}
-+
-+void fman_dtsec_handle_rx_pause(struct dtsec_regs *regs, bool en)
-+{
-+ uint32_t tmp;
-+
-+ /* todo: check if mac is set to full-duplex */
-+
-+ tmp = ioread32be(&regs->maccfg1);
-+ if (en)
-+ tmp |= MACCFG1_RX_FLOW;
-+ else
-+ tmp &= ~MACCFG1_RX_FLOW;
-+ iowrite32be(tmp, &regs->maccfg1);
-+}
-+
-+uint32_t fman_dtsec_get_rctrl(struct dtsec_regs *regs)
-+{
-+ return ioread32be(&regs->rctrl);
-+}
-+
-+uint32_t fman_dtsec_get_revision(struct dtsec_regs *regs)
-+{
-+ return ioread32be(&regs->tsec_id);
-+}
-+
-+uint32_t fman_dtsec_get_event(struct dtsec_regs *regs, uint32_t ev_mask)
-+{
-+ return ioread32be(&regs->ievent) & ev_mask;
-+}
-+
-+void fman_dtsec_ack_event(struct dtsec_regs *regs, uint32_t ev_mask)
-+{
-+ iowrite32be(ev_mask, &regs->ievent);
-+}
-+
-+uint32_t fman_dtsec_get_interrupt_mask(struct dtsec_regs *regs)
-+{
-+ return ioread32be(&regs->imask);
-+}
-+
-+uint32_t fman_dtsec_check_and_clear_tmr_event(struct dtsec_regs *regs)
-+{
-+ uint32_t event;
-+
-+ event = ioread32be(&regs->tmr_pevent);
-+ event &= ioread32be(&regs->tmr_pemask);
-+
-+ if (event)
-+ iowrite32be(event, &regs->tmr_pevent);
-+ return event;
-+}
-+
-+void fman_dtsec_enable_tmr_interrupt(struct dtsec_regs *regs)
-+{
-+ iowrite32be(ioread32be(&regs->tmr_pemask) | TMR_PEMASK_TSREEN,
-+ &regs->tmr_pemask);
-+}
-+
-+void fman_dtsec_disable_tmr_interrupt(struct dtsec_regs *regs)
-+{
-+ iowrite32be(ioread32be(&regs->tmr_pemask) & ~TMR_PEMASK_TSREEN,
-+ &regs->tmr_pemask);
-+}
-+
-+void fman_dtsec_enable_interrupt(struct dtsec_regs *regs, uint32_t ev_mask)
-+{
-+ iowrite32be(ioread32be(&regs->imask) | ev_mask, &regs->imask);
-+}
-+
-+void fman_dtsec_disable_interrupt(struct dtsec_regs *regs, uint32_t ev_mask)
-+{
-+ iowrite32be(ioread32be(&regs->imask) & ~ev_mask, &regs->imask);
-+}
-+
-+uint32_t fman_dtsec_get_stat_counter(struct dtsec_regs *regs,
-+ enum dtsec_stat_counters reg_name)
-+{
-+ uint32_t ret_val;
-+
-+ switch (reg_name) {
-+ case E_DTSEC_STAT_TR64:
-+ ret_val = ioread32be(&regs->tr64);
-+ break;
-+ case E_DTSEC_STAT_TR127:
-+ ret_val = ioread32be(&regs->tr127);
-+ break;
-+ case E_DTSEC_STAT_TR255:
-+ ret_val = ioread32be(&regs->tr255);
-+ break;
-+ case E_DTSEC_STAT_TR511:
-+ ret_val = ioread32be(&regs->tr511);
-+ break;
-+ case E_DTSEC_STAT_TR1K:
-+ ret_val = ioread32be(&regs->tr1k);
-+ break;
-+ case E_DTSEC_STAT_TRMAX:
-+ ret_val = ioread32be(&regs->trmax);
-+ break;
-+ case E_DTSEC_STAT_TRMGV:
-+ ret_val = ioread32be(&regs->trmgv);
-+ break;
-+ case E_DTSEC_STAT_RBYT:
-+ ret_val = ioread32be(&regs->rbyt);
-+ break;
-+ case E_DTSEC_STAT_RPKT:
-+ ret_val = ioread32be(&regs->rpkt);
-+ break;
-+ case E_DTSEC_STAT_RMCA:
-+ ret_val = ioread32be(&regs->rmca);
-+ break;
-+ case E_DTSEC_STAT_RBCA:
-+ ret_val = ioread32be(&regs->rbca);
-+ break;
-+ case E_DTSEC_STAT_RXPF:
-+ ret_val = ioread32be(&regs->rxpf);
-+ break;
-+ case E_DTSEC_STAT_RALN:
-+ ret_val = ioread32be(&regs->raln);
-+ break;
-+ case E_DTSEC_STAT_RFLR:
-+ ret_val = ioread32be(&regs->rflr);
-+ break;
-+ case E_DTSEC_STAT_RCDE:
-+ ret_val = ioread32be(&regs->rcde);
-+ break;
-+ case E_DTSEC_STAT_RCSE:
-+ ret_val = ioread32be(&regs->rcse);
-+ break;
-+ case E_DTSEC_STAT_RUND:
-+ ret_val = ioread32be(&regs->rund);
-+ break;
-+ case E_DTSEC_STAT_ROVR:
-+ ret_val = ioread32be(&regs->rovr);
-+ break;
-+ case E_DTSEC_STAT_RFRG:
-+ ret_val = ioread32be(&regs->rfrg);
-+ break;
-+ case E_DTSEC_STAT_RJBR:
-+ ret_val = ioread32be(&regs->rjbr);
-+ break;
-+ case E_DTSEC_STAT_RDRP:
-+ ret_val = ioread32be(&regs->rdrp);
-+ break;
-+ case E_DTSEC_STAT_TFCS:
-+ ret_val = ioread32be(&regs->tfcs);
-+ break;
-+ case E_DTSEC_STAT_TBYT:
-+ ret_val = ioread32be(&regs->tbyt);
-+ break;
-+ case E_DTSEC_STAT_TPKT:
-+ ret_val = ioread32be(&regs->tpkt);
-+ break;
-+ case E_DTSEC_STAT_TMCA:
-+ ret_val = ioread32be(&regs->tmca);
-+ break;
-+ case E_DTSEC_STAT_TBCA:
-+ ret_val = ioread32be(&regs->tbca);
-+ break;
-+ case E_DTSEC_STAT_TXPF:
-+ ret_val = ioread32be(&regs->txpf);
-+ break;
-+ case E_DTSEC_STAT_TNCL:
-+ ret_val = ioread32be(&regs->tncl);
-+ break;
-+ case E_DTSEC_STAT_TDRP:
-+ ret_val = ioread32be(&regs->tdrp);
-+ break;
-+ default:
-+ ret_val = 0;
-+ }
-+
-+ return ret_val;
-+}
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MAC/fman_dtsec_mii_acc.c
-@@ -0,0 +1,163 @@
-+/*
-+ * Copyright 2008-2013 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+#include "common/general.h"
-+#include "fsl_fman_dtsec_mii_acc.h"
-+
-+
-+/**
-+ * dtsec_mii_get_div() - calculates the value of the dtsec mii divider
-+ * @dtsec_freq: dtsec clock frequency (in Mhz)
-+ *
-+ * This function calculates the dtsec mii clock divider that determines
-+ * the MII MDC clock. MII MDC clock will be set to work in the range
-+ * of 1.5 to 2.5Mhz
-+ * The output of this function is the value of MIIMCFG[MgmtClk] which
-+ * implicitly determines the divider value.
-+ * Note: the dTSEC system clock is equal to 1/2 of the FMan clock.
-+ *
-+ * The table below which reflects dtsec_mii_get_div() functionality
-+ * shows the relations among dtsec_freq, MgmtClk, actual divider
-+ * and the MII frequency:
-+ *
-+ * dtsec freq MgmtClk div MII freq Mhz
-+ * [0.....80] 1 (1/4)(1/8) [0 to 2.5]
-+ * [81...120] 2 (1/6)(1/8) [1.6 to 2.5]
-+ * [121..160] 3 (1/8)(1/8) [1.8 to 2.5]
-+ * [161..200] 4 (1/10)(1/8) [2.0 to 2.5]
-+ * [201..280] 5 (1/14)(1/8) [1.8 to 2.5]
-+ * [281..400] 6 (1/20)(1/8) [1.1 to 2.5]
-+ * [401..560] 7 (1/28)(1/8) [1.8 to 2.5]
-+ * [560..frq] 7 (1/28)(1/8) [frq/224]
-+ *
-+ * Returns: the MIIMCFG[MgmtClk] appropriate value
-+ */
-+
-+static uint8_t dtsec_mii_get_div(uint16_t dtsec_freq)
-+{
-+ uint16_t mgmt_clk;
-+
-+ if (dtsec_freq < 80) mgmt_clk = 1;
-+ else if (dtsec_freq < 120) mgmt_clk = 2;
-+ else if (dtsec_freq < 160) mgmt_clk = 3;
-+ else if (dtsec_freq < 200) mgmt_clk = 4;
-+ else if (dtsec_freq < 280) mgmt_clk = 5;
-+ else if (dtsec_freq < 400) mgmt_clk = 6;
-+ else mgmt_clk = 7;
-+
-+ return (uint8_t)mgmt_clk;
-+}
-+
-+void fman_dtsec_mii_reset(struct dtsec_mii_reg *regs)
-+{
-+ /* Reset the management interface */
-+ iowrite32be(ioread32be(&regs->miimcfg) | MIIMCFG_RESET_MGMT,
-+ &regs->miimcfg);
-+ iowrite32be(ioread32be(&regs->miimcfg) & ~MIIMCFG_RESET_MGMT,
-+ &regs->miimcfg);
-+}
-+
-+
-+int fman_dtsec_mii_write_reg(struct dtsec_mii_reg *regs, uint8_t addr,
-+ uint8_t reg, uint16_t data, uint16_t dtsec_freq)
-+{
-+ uint32_t tmp;
-+
-+ /* Setup the MII Mgmt clock speed */
-+ iowrite32be((uint32_t)dtsec_mii_get_div(dtsec_freq), &regs->miimcfg);
-+ wmb();
-+
-+ /* Stop the MII management read cycle */
-+ iowrite32be(0, &regs->miimcom);
-+ /* Dummy read to make sure MIIMCOM is written */
-+ tmp = ioread32be(&regs->miimcom);
-+ wmb();
-+
-+ /* Setting up MII Management Address Register */
-+ tmp = (uint32_t)((addr << MIIMADD_PHY_ADDR_SHIFT) | reg);
-+ iowrite32be(tmp, &regs->miimadd);
-+ wmb();
-+
-+ /* Setting up MII Management Control Register with data */
-+ iowrite32be((uint32_t)data, &regs->miimcon);
-+ /* Dummy read to make sure MIIMCON is written */
-+ tmp = ioread32be(&regs->miimcon);
-+ wmb();
-+
-+ /* Wait until MII management write is complete */
-+ /* todo: a timeout could be useful here */
-+ while ((ioread32be(&regs->miimind)) & MIIMIND_BUSY)
-+ /* busy wait */;
-+
-+ return 0;
-+}
-+
-+int fman_dtsec_mii_read_reg(struct dtsec_mii_reg *regs, uint8_t addr,
-+ uint8_t reg, uint16_t *data, uint16_t dtsec_freq)
-+{
-+ uint32_t tmp;
-+
-+ /* Setup the MII Mgmt clock speed */
-+ iowrite32be((uint32_t)dtsec_mii_get_div(dtsec_freq), &regs->miimcfg);
-+ wmb();
-+
-+ /* Setting up the MII Management Address Register */
-+ tmp = (uint32_t)((addr << MIIMADD_PHY_ADDR_SHIFT) | reg);
-+ iowrite32be(tmp, &regs->miimadd);
-+ wmb();
-+
-+ /* Perform an MII management read cycle */
-+ iowrite32be(MIIMCOM_READ_CYCLE, &regs->miimcom);
-+ /* Dummy read to make sure MIIMCOM is written */
-+ tmp = ioread32be(&regs->miimcom);
-+ wmb();
-+
-+ /* Wait until MII management read is complete */
-+ /* todo: a timeout could be useful here */
-+ while ((ioread32be(&regs->miimind)) & MIIMIND_BUSY)
-+ /* busy wait */;
-+
-+ /* Read MII management status */
-+ *data = (uint16_t)ioread32be(&regs->miimstat);
-+ wmb();
-+
-+ iowrite32be(0, &regs->miimcom);
-+ /* Dummy read to make sure MIIMCOM is written */
-+ tmp = ioread32be(&regs->miimcom);
-+
-+ if (*data == 0xffff)
-+ return -ENXIO;
-+
-+ return 0;
-+}
-+
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MAC/fman_memac.c
-@@ -0,0 +1,532 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+#include "fsl_fman_memac.h"
-+
-+
-+uint32_t fman_memac_get_event(struct memac_regs *regs, uint32_t ev_mask)
-+{
-+ return ioread32be(&regs->ievent) & ev_mask;
-+}
-+
-+uint32_t fman_memac_get_interrupt_mask(struct memac_regs *regs)
-+{
-+ return ioread32be(&regs->imask);
-+}
-+
-+void fman_memac_ack_event(struct memac_regs *regs, uint32_t ev_mask)
-+{
-+ iowrite32be(ev_mask, &regs->ievent);
-+}
-+
-+void fman_memac_set_promiscuous(struct memac_regs *regs, bool val)
-+{
-+ uint32_t tmp;
-+
-+ tmp = ioread32be(&regs->command_config);
-+
-+ if (val)
-+ tmp |= CMD_CFG_PROMIS_EN;
-+ else
-+ tmp &= ~CMD_CFG_PROMIS_EN;
-+
-+ iowrite32be(tmp, &regs->command_config);
-+}
-+
-+void fman_memac_clear_addr_in_paddr(struct memac_regs *regs,
-+ uint8_t paddr_num)
-+{
-+ if (paddr_num == 0) {
-+ iowrite32be(0, &regs->mac_addr0.mac_addr_l);
-+ iowrite32be(0, &regs->mac_addr0.mac_addr_u);
-+ } else {
-+ iowrite32be(0x0, &regs->mac_addr[paddr_num - 1].mac_addr_l);
-+ iowrite32be(0x0, &regs->mac_addr[paddr_num - 1].mac_addr_u);
-+ }
-+}
-+
-+void fman_memac_add_addr_in_paddr(struct memac_regs *regs,
-+ uint8_t *adr,
-+ uint8_t paddr_num)
-+{
-+ uint32_t tmp0, tmp1;
-+
-+ tmp0 = (uint32_t)(adr[0] |
-+ adr[1] << 8 |
-+ adr[2] << 16 |
-+ adr[3] << 24);
-+ tmp1 = (uint32_t)(adr[4] | adr[5] << 8);
-+
-+ if (paddr_num == 0) {
-+ iowrite32be(tmp0, &regs->mac_addr0.mac_addr_l);
-+ iowrite32be(tmp1, &regs->mac_addr0.mac_addr_u);
-+ } else {
-+ iowrite32be(tmp0, &regs->mac_addr[paddr_num-1].mac_addr_l);
-+ iowrite32be(tmp1, &regs->mac_addr[paddr_num-1].mac_addr_u);
-+ }
-+}
-+
-+void fman_memac_enable(struct memac_regs *regs, bool apply_rx, bool apply_tx)
-+{
-+ uint32_t tmp;
-+
-+ tmp = ioread32be(&regs->command_config);
-+
-+ if (apply_rx)
-+ tmp |= CMD_CFG_RX_EN;
-+
-+ if (apply_tx)
-+ tmp |= CMD_CFG_TX_EN;
-+
-+ iowrite32be(tmp, &regs->command_config);
-+}
-+
-+void fman_memac_disable(struct memac_regs *regs, bool apply_rx, bool apply_tx)
-+{
-+ uint32_t tmp;
-+
-+ tmp = ioread32be(&regs->command_config);
-+
-+ if (apply_rx)
-+ tmp &= ~CMD_CFG_RX_EN;
-+
-+ if (apply_tx)
-+ tmp &= ~CMD_CFG_TX_EN;
-+
-+ iowrite32be(tmp, &regs->command_config);
-+}
-+
-+void fman_memac_reset_stat(struct memac_regs *regs)
-+{
-+ uint32_t tmp;
-+
-+ tmp = ioread32be(&regs->statn_config);
-+
-+ tmp |= STATS_CFG_CLR;
-+
-+ iowrite32be(tmp, &regs->statn_config);
-+
-+ while (ioread32be(&regs->statn_config) & STATS_CFG_CLR);
-+}
-+
-+void fman_memac_reset(struct memac_regs *regs)
-+{
-+ uint32_t tmp;
-+
-+ tmp = ioread32be(&regs->command_config);
-+
-+ tmp |= CMD_CFG_SW_RESET;
-+
-+ iowrite32be(tmp, &regs->command_config);
-+
-+ while (ioread32be(&regs->command_config) & CMD_CFG_SW_RESET);
-+}
-+
-+int fman_memac_init(struct memac_regs *regs,
-+ struct memac_cfg *cfg,
-+ enum enet_interface enet_interface,
-+ enum enet_speed enet_speed,
-+ bool slow_10g_if,
-+ uint32_t exceptions)
-+{
-+ uint32_t tmp;
-+
-+ /* Config */
-+ tmp = 0;
-+ if (cfg->wan_mode_enable)
-+ tmp |= CMD_CFG_WAN_MODE;
-+ if (cfg->promiscuous_mode_enable)
-+ tmp |= CMD_CFG_PROMIS_EN;
-+ if (cfg->pause_forward_enable)
-+ tmp |= CMD_CFG_PAUSE_FWD;
-+ if (cfg->pause_ignore)
-+ tmp |= CMD_CFG_PAUSE_IGNORE;
-+ if (cfg->tx_addr_ins_enable)
-+ tmp |= CMD_CFG_TX_ADDR_INS;
-+ if (cfg->loopback_enable)
-+ tmp |= CMD_CFG_LOOPBACK_EN;
-+ if (cfg->cmd_frame_enable)
-+ tmp |= CMD_CFG_CNT_FRM_EN;
-+ if (cfg->send_idle_enable)
-+ tmp |= CMD_CFG_SEND_IDLE;
-+ if (cfg->no_length_check_enable)
-+ tmp |= CMD_CFG_NO_LEN_CHK;
-+ if (cfg->rx_sfd_any)
-+ tmp |= CMD_CFG_SFD_ANY;
-+ if (cfg->pad_enable)
-+ tmp |= CMD_CFG_TX_PAD_EN;
-+ if (cfg->wake_on_lan)
-+ tmp |= CMD_CFG_MG;
-+
-+ tmp |= CMD_CFG_CRC_FWD;
-+
-+ iowrite32be(tmp, &regs->command_config);
-+
-+ /* Max Frame Length */
-+ iowrite32be((uint32_t)cfg->max_frame_length, &regs->maxfrm);
-+
-+ /* Pause Time */
-+ iowrite32be((uint32_t)cfg->pause_quanta, &regs->pause_quanta[0]);
-+ iowrite32be((uint32_t)0, &regs->pause_thresh[0]);
-+
-+ /* IF_MODE */
-+ tmp = 0;
-+ switch (enet_interface) {
-+ case E_ENET_IF_XGMII:
-+ case E_ENET_IF_XFI:
-+ tmp |= IF_MODE_XGMII;
-+ break;
-+ default:
-+ tmp |= IF_MODE_GMII;
-+ if (enet_interface == E_ENET_IF_RGMII && !cfg->loopback_enable)
-+ tmp |= IF_MODE_RGMII | IF_MODE_RGMII_AUTO;
-+ }
-+ iowrite32be(tmp, &regs->if_mode);
-+
-+ /* TX_FIFO_SECTIONS */
-+ tmp = 0;
-+ if (enet_interface == E_ENET_IF_XGMII ||
-+ enet_interface == E_ENET_IF_XFI) {
-+ if(slow_10g_if) {
-+ tmp |= (TX_FIFO_SECTIONS_TX_AVAIL_SLOW_10G |
-+ TX_FIFO_SECTIONS_TX_EMPTY_DEFAULT_10G);
-+ } else {
-+ tmp |= (TX_FIFO_SECTIONS_TX_AVAIL_10G |
-+ TX_FIFO_SECTIONS_TX_EMPTY_DEFAULT_10G);
-+ }
-+ } else {
-+ tmp |= (TX_FIFO_SECTIONS_TX_AVAIL_1G |
-+ TX_FIFO_SECTIONS_TX_EMPTY_DEFAULT_1G);
-+ }
-+ iowrite32be(tmp, &regs->tx_fifo_sections);
-+
-+ /* clear all pending events and set-up interrupts */
-+ fman_memac_ack_event(regs, 0xffffffff);
-+ fman_memac_set_exception(regs, exceptions, TRUE);
-+
-+ return 0;
-+}
-+
-+void fman_memac_set_exception(struct memac_regs *regs, uint32_t val, bool enable)
-+{
-+ uint32_t tmp;
-+
-+ tmp = ioread32be(&regs->imask);
-+ if (enable)
-+ tmp |= val;
-+ else
-+ tmp &= ~val;
-+
-+ iowrite32be(tmp, &regs->imask);
-+}
-+
-+void fman_memac_reset_filter_table(struct memac_regs *regs)
-+{
-+ uint32_t i;
-+ for (i = 0; i < 64; i++)
-+ iowrite32be(i & ~HASH_CTRL_MCAST_EN, &regs->hashtable_ctrl);
-+}
-+
-+void fman_memac_set_hash_table_entry(struct memac_regs *regs, uint32_t crc)
-+{
-+ iowrite32be(crc | HASH_CTRL_MCAST_EN, &regs->hashtable_ctrl);
-+}
-+
-+void fman_memac_set_hash_table(struct memac_regs *regs, uint32_t val)
-+{
-+ iowrite32be(val, &regs->hashtable_ctrl);
-+}
-+
-+uint16_t fman_memac_get_max_frame_len(struct memac_regs *regs)
-+{
-+ uint32_t tmp;
-+
-+ tmp = ioread32be(&regs->maxfrm);
-+
-+ return(uint16_t)tmp;
-+}
-+
-+
-+void fman_memac_set_tx_pause_frames(struct memac_regs *regs,
-+ uint8_t priority,
-+ uint16_t pause_time,
-+ uint16_t thresh_time)
-+{
-+ uint32_t tmp;
-+
-+ tmp = ioread32be(&regs->tx_fifo_sections);
-+
-+ if (priority == 0xff) {
-+ GET_TX_EMPTY_DEFAULT_VALUE(tmp);
-+ iowrite32be(tmp, &regs->tx_fifo_sections);
-+
-+ tmp = ioread32be(&regs->command_config);
-+ tmp &= ~CMD_CFG_PFC_MODE;
-+ priority = 0;
-+ } else {
-+ GET_TX_EMPTY_PFC_VALUE(tmp);
-+ iowrite32be(tmp, &regs->tx_fifo_sections);
-+
-+ tmp = ioread32be(&regs->command_config);
-+ tmp |= CMD_CFG_PFC_MODE;
-+ }
-+
-+ iowrite32be(tmp, &regs->command_config);
-+
-+ tmp = ioread32be(&regs->pause_quanta[priority / 2]);
-+ if (priority % 2)
-+ tmp &= 0x0000FFFF;
-+ else
-+ tmp &= 0xFFFF0000;
-+ tmp |= ((uint32_t)pause_time << (16 * (priority % 2)));
-+ iowrite32be(tmp, &regs->pause_quanta[priority / 2]);
-+
-+ tmp = ioread32be(&regs->pause_thresh[priority / 2]);
-+ if (priority % 2)
-+ tmp &= 0x0000FFFF;
-+ else
-+ tmp &= 0xFFFF0000;
-+ tmp |= ((uint32_t)thresh_time<<(16 * (priority % 2)));
-+ iowrite32be(tmp, &regs->pause_thresh[priority / 2]);
-+}
-+
-+void fman_memac_set_rx_ignore_pause_frames(struct memac_regs *regs,bool enable)
-+{
-+ uint32_t tmp;
-+
-+ tmp = ioread32be(&regs->command_config);
-+ if (enable)
-+ tmp |= CMD_CFG_PAUSE_IGNORE;
-+ else
-+ tmp &= ~CMD_CFG_PAUSE_IGNORE;
-+
-+ iowrite32be(tmp, &regs->command_config);
-+}
-+
-+void fman_memac_set_wol(struct memac_regs *regs, bool enable)
-+{
-+ uint32_t tmp;
-+
-+ tmp = ioread32be(&regs->command_config);
-+
-+ if (enable)
-+ tmp |= CMD_CFG_MG;
-+ else
-+ tmp &= ~CMD_CFG_MG;
-+
-+ iowrite32be(tmp, &regs->command_config);
-+}
-+
-+#define GET_MEMAC_CNTR_64(bn) \
-+ (ioread32be(&regs->bn ## _l) | \
-+ ((uint64_t)ioread32be(&regs->bn ## _u) << 32))
-+
-+uint64_t fman_memac_get_counter(struct memac_regs *regs,
-+ enum memac_counters reg_name)
-+{
-+ uint64_t ret_val;
-+
-+ switch (reg_name) {
-+ case E_MEMAC_COUNTER_R64:
-+ ret_val = GET_MEMAC_CNTR_64(r64);
-+ break;
-+ case E_MEMAC_COUNTER_T64:
-+ ret_val = GET_MEMAC_CNTR_64(t64);
-+ break;
-+ case E_MEMAC_COUNTER_R127:
-+ ret_val = GET_MEMAC_CNTR_64(r127);
-+ break;
-+ case E_MEMAC_COUNTER_T127:
-+ ret_val = GET_MEMAC_CNTR_64(t127);
-+ break;
-+ case E_MEMAC_COUNTER_R255:
-+ ret_val = GET_MEMAC_CNTR_64(r255);
-+ break;
-+ case E_MEMAC_COUNTER_T255:
-+ ret_val = GET_MEMAC_CNTR_64(t255);
-+ break;
-+ case E_MEMAC_COUNTER_R511:
-+ ret_val = GET_MEMAC_CNTR_64(r511);
-+ break;
-+ case E_MEMAC_COUNTER_T511:
-+ ret_val = GET_MEMAC_CNTR_64(t511);
-+ break;
-+ case E_MEMAC_COUNTER_R1023:
-+ ret_val = GET_MEMAC_CNTR_64(r1023);
-+ break;
-+ case E_MEMAC_COUNTER_T1023:
-+ ret_val = GET_MEMAC_CNTR_64(t1023);
-+ break;
-+ case E_MEMAC_COUNTER_R1518:
-+ ret_val = GET_MEMAC_CNTR_64(r1518);
-+ break;
-+ case E_MEMAC_COUNTER_T1518:
-+ ret_val = GET_MEMAC_CNTR_64(t1518);
-+ break;
-+ case E_MEMAC_COUNTER_R1519X:
-+ ret_val = GET_MEMAC_CNTR_64(r1519x);
-+ break;
-+ case E_MEMAC_COUNTER_T1519X:
-+ ret_val = GET_MEMAC_CNTR_64(t1519x);
-+ break;
-+ case E_MEMAC_COUNTER_RFRG:
-+ ret_val = GET_MEMAC_CNTR_64(rfrg);
-+ break;
-+ case E_MEMAC_COUNTER_RJBR:
-+ ret_val = GET_MEMAC_CNTR_64(rjbr);
-+ break;
-+ case E_MEMAC_COUNTER_RDRP:
-+ ret_val = GET_MEMAC_CNTR_64(rdrp);
-+ break;
-+ case E_MEMAC_COUNTER_RALN:
-+ ret_val = GET_MEMAC_CNTR_64(raln);
-+ break;
-+ case E_MEMAC_COUNTER_TUND:
-+ ret_val = GET_MEMAC_CNTR_64(tund);
-+ break;
-+ case E_MEMAC_COUNTER_ROVR:
-+ ret_val = GET_MEMAC_CNTR_64(rovr);
-+ break;
-+ case E_MEMAC_COUNTER_RXPF:
-+ ret_val = GET_MEMAC_CNTR_64(rxpf);
-+ break;
-+ case E_MEMAC_COUNTER_TXPF:
-+ ret_val = GET_MEMAC_CNTR_64(txpf);
-+ break;
-+ case E_MEMAC_COUNTER_ROCT:
-+ ret_val = GET_MEMAC_CNTR_64(roct);
-+ break;
-+ case E_MEMAC_COUNTER_RMCA:
-+ ret_val = GET_MEMAC_CNTR_64(rmca);
-+ break;
-+ case E_MEMAC_COUNTER_RBCA:
-+ ret_val = GET_MEMAC_CNTR_64(rbca);
-+ break;
-+ case E_MEMAC_COUNTER_RPKT:
-+ ret_val = GET_MEMAC_CNTR_64(rpkt);
-+ break;
-+ case E_MEMAC_COUNTER_RUCA:
-+ ret_val = GET_MEMAC_CNTR_64(ruca);
-+ break;
-+ case E_MEMAC_COUNTER_RERR:
-+ ret_val = GET_MEMAC_CNTR_64(rerr);
-+ break;
-+ case E_MEMAC_COUNTER_TOCT:
-+ ret_val = GET_MEMAC_CNTR_64(toct);
-+ break;
-+ case E_MEMAC_COUNTER_TMCA:
-+ ret_val = GET_MEMAC_CNTR_64(tmca);
-+ break;
-+ case E_MEMAC_COUNTER_TBCA:
-+ ret_val = GET_MEMAC_CNTR_64(tbca);
-+ break;
-+ case E_MEMAC_COUNTER_TUCA:
-+ ret_val = GET_MEMAC_CNTR_64(tuca);
-+ break;
-+ case E_MEMAC_COUNTER_TERR:
-+ ret_val = GET_MEMAC_CNTR_64(terr);
-+ break;
-+ default:
-+ ret_val = 0;
-+ }
-+
-+ return ret_val;
-+}
-+
-+void fman_memac_adjust_link(struct memac_regs *regs,
-+ enum enet_interface iface_mode,
-+ enum enet_speed speed, bool full_dx)
-+{
-+ uint32_t tmp;
-+
-+ tmp = ioread32be(&regs->if_mode);
-+
-+ if (full_dx)
-+ tmp &= ~IF_MODE_HD;
-+ else
-+ tmp |= IF_MODE_HD;
-+
-+ if (iface_mode == E_ENET_IF_RGMII) {
-+ /* Configure RGMII in manual mode */
-+ tmp &= ~IF_MODE_RGMII_AUTO;
-+ tmp &= ~IF_MODE_RGMII_SP_MASK;
-+
-+ if (full_dx)
-+ tmp |= IF_MODE_RGMII_FD;
-+ else
-+ tmp &= ~IF_MODE_RGMII_FD;
-+
-+ switch (speed) {
-+ case E_ENET_SPEED_1000:
-+ tmp |= IF_MODE_RGMII_1000;
-+ break;
-+ case E_ENET_SPEED_100:
-+ tmp |= IF_MODE_RGMII_100;
-+ break;
-+ case E_ENET_SPEED_10:
-+ tmp |= IF_MODE_RGMII_10;
-+ break;
-+ default:
-+ break;
-+ }
-+ }
-+
-+ iowrite32be(tmp, &regs->if_mode);
-+}
-+
-+void fman_memac_defconfig(struct memac_cfg *cfg)
-+{
-+ cfg->reset_on_init = FALSE;
-+ cfg->wan_mode_enable = FALSE;
-+ cfg->promiscuous_mode_enable = FALSE;
-+ cfg->pause_forward_enable = FALSE;
-+ cfg->pause_ignore = FALSE;
-+ cfg->tx_addr_ins_enable = FALSE;
-+ cfg->loopback_enable = FALSE;
-+ cfg->cmd_frame_enable = FALSE;
-+ cfg->rx_error_discard = FALSE;
-+ cfg->send_idle_enable = FALSE;
-+ cfg->no_length_check_enable = TRUE;
-+ cfg->lgth_check_nostdr = FALSE;
-+ cfg->time_stamp_enable = FALSE;
-+ cfg->tx_ipg_length = DEFAULT_TX_IPG_LENGTH;
-+ cfg->max_frame_length = DEFAULT_FRAME_LENGTH;
-+ cfg->pause_quanta = DEFAULT_PAUSE_QUANTA;
-+ cfg->pad_enable = TRUE;
-+ cfg->phy_tx_ena_on = FALSE;
-+ cfg->rx_sfd_any = FALSE;
-+ cfg->rx_pbl_fwd = FALSE;
-+ cfg->tx_pbl_fwd = FALSE;
-+ cfg->debug_mode = FALSE;
-+ cfg->wake_on_lan = FALSE;
-+}
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MAC/fman_memac_mii_acc.c
-@@ -0,0 +1,213 @@
-+/*
-+ * Copyright 2008-2013 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+#include "fsl_fman_memac_mii_acc.h"
-+
-+static void write_phy_reg_10g(struct memac_mii_access_mem_map *mii_regs,
-+ uint8_t phy_addr, uint8_t reg, uint16_t data)
-+{
-+ uint32_t tmp_reg;
-+
-+ tmp_reg = ioread32be(&mii_regs->mdio_cfg);
-+ /* Leave only MDIO_CLK_DIV bits set on */
-+ tmp_reg &= MDIO_CFG_CLK_DIV_MASK;
-+ /* Set maximum MDIO_HOLD value to allow phy to see
-+ change of data signal */
-+ tmp_reg |= MDIO_CFG_HOLD_MASK;
-+ /* Add 10G interface mode */
-+ tmp_reg |= MDIO_CFG_ENC45;
-+ iowrite32be(tmp_reg, &mii_regs->mdio_cfg);
-+
-+ /* Wait for command completion */
-+ while ((ioread32be(&mii_regs->mdio_cfg)) & MDIO_CFG_BSY)
-+ udelay(1);
-+
-+ /* Specify phy and register to be accessed */
-+ iowrite32be(phy_addr, &mii_regs->mdio_ctrl);
-+ iowrite32be(reg, &mii_regs->mdio_addr);
-+ wmb();
-+
-+ while ((ioread32be(&mii_regs->mdio_cfg)) & MDIO_CFG_BSY)
-+ udelay(1);
-+
-+ /* Write data */
-+ iowrite32be(data, &mii_regs->mdio_data);
-+ wmb();
-+
-+ /* Wait for write transaction end */
-+ while ((ioread32be(&mii_regs->mdio_data)) & MDIO_DATA_BSY)
-+ udelay(1);
-+}
-+
-+static uint32_t read_phy_reg_10g(struct memac_mii_access_mem_map *mii_regs,
-+ uint8_t phy_addr, uint8_t reg, uint16_t *data)
-+{
-+ uint32_t tmp_reg;
-+
-+ tmp_reg = ioread32be(&mii_regs->mdio_cfg);
-+ /* Leave only MDIO_CLK_DIV bits set on */
-+ tmp_reg &= MDIO_CFG_CLK_DIV_MASK;
-+ /* Set maximum MDIO_HOLD value to allow phy to see
-+ change of data signal */
-+ tmp_reg |= MDIO_CFG_HOLD_MASK;
-+ /* Add 10G interface mode */
-+ tmp_reg |= MDIO_CFG_ENC45;
-+ iowrite32be(tmp_reg, &mii_regs->mdio_cfg);
-+
-+ /* Wait for command completion */
-+ while ((ioread32be(&mii_regs->mdio_cfg)) & MDIO_CFG_BSY)
-+ udelay(1);
-+
-+ /* Specify phy and register to be accessed */
-+ iowrite32be(phy_addr, &mii_regs->mdio_ctrl);
-+ iowrite32be(reg, &mii_regs->mdio_addr);
-+ wmb();
-+
-+ while ((ioread32be(&mii_regs->mdio_cfg)) & MDIO_CFG_BSY)
-+ udelay(1);
-+
-+ /* Read cycle */
-+ tmp_reg = phy_addr;
-+ tmp_reg |= MDIO_CTL_READ;
-+ iowrite32be(tmp_reg, &mii_regs->mdio_ctrl);
-+ wmb();
-+
-+ /* Wait for data to be available */
-+ while ((ioread32be(&mii_regs->mdio_data)) & MDIO_DATA_BSY)
-+ udelay(1);
-+
-+ *data = (uint16_t)ioread32be(&mii_regs->mdio_data);
-+
-+ /* Check if there was an error */
-+ return ioread32be(&mii_regs->mdio_cfg);
-+}
-+
-+static void write_phy_reg_1g(struct memac_mii_access_mem_map *mii_regs,
-+ uint8_t phy_addr, uint8_t reg, uint16_t data)
-+{
-+ uint32_t tmp_reg;
-+
-+ /* Leave only MDIO_CLK_DIV and MDIO_HOLD bits set on */
-+ tmp_reg = ioread32be(&mii_regs->mdio_cfg);
-+ tmp_reg &= (MDIO_CFG_CLK_DIV_MASK | MDIO_CFG_HOLD_MASK);
-+ iowrite32be(tmp_reg, &mii_regs->mdio_cfg);
-+
-+ /* Wait for command completion */
-+ while ((ioread32be(&mii_regs->mdio_cfg)) & MDIO_CFG_BSY)
-+ udelay(1);
-+
-+ /* Write transaction */
-+ tmp_reg = (phy_addr << MDIO_CTL_PHY_ADDR_SHIFT);
-+ tmp_reg |= reg;
-+ iowrite32be(tmp_reg, &mii_regs->mdio_ctrl);
-+
-+ while ((ioread32be(&mii_regs->mdio_cfg)) & MDIO_CFG_BSY)
-+ udelay(1);
-+
-+ iowrite32be(data, &mii_regs->mdio_data);
-+
-+ wmb();
-+
-+ /* Wait for write transaction to end */
-+ while ((ioread32be(&mii_regs->mdio_data)) & MDIO_DATA_BSY)
-+ udelay(1);
-+}
-+
-+static uint32_t read_phy_reg_1g(struct memac_mii_access_mem_map *mii_regs,
-+ uint8_t phy_addr, uint8_t reg, uint16_t *data)
-+{
-+ uint32_t tmp_reg;
-+
-+ /* Leave only MDIO_CLK_DIV and MDIO_HOLD bits set on */
-+ tmp_reg = ioread32be(&mii_regs->mdio_cfg);
-+ tmp_reg &= (MDIO_CFG_CLK_DIV_MASK | MDIO_CFG_HOLD_MASK);
-+ iowrite32be(tmp_reg, &mii_regs->mdio_cfg);
-+
-+ /* Wait for command completion */
-+ while ((ioread32be(&mii_regs->mdio_cfg)) & MDIO_CFG_BSY)
-+ udelay(1);
-+
-+ /* Read transaction */
-+ tmp_reg = (phy_addr << MDIO_CTL_PHY_ADDR_SHIFT);
-+ tmp_reg |= reg;
-+ tmp_reg |= MDIO_CTL_READ;
-+ iowrite32be(tmp_reg, &mii_regs->mdio_ctrl);
-+
-+ while ((ioread32be(&mii_regs->mdio_cfg)) & MDIO_CFG_BSY)
-+ udelay(1);
-+
-+ /* Wait for data to be available */
-+ while ((ioread32be(&mii_regs->mdio_data)) & MDIO_DATA_BSY)
-+ udelay(1);
-+
-+ *data = (uint16_t)ioread32be(&mii_regs->mdio_data);
-+
-+ /* Check error */
-+ return ioread32be(&mii_regs->mdio_cfg);
-+}
-+
-+/*****************************************************************************/
-+int fman_memac_mii_write_phy_reg(struct memac_mii_access_mem_map *mii_regs,
-+ uint8_t phy_addr, uint8_t reg, uint16_t data,
-+ enum enet_speed enet_speed)
-+{
-+ /* Figure out interface type - 10G vs 1G.
-+ In 10G interface both phy_addr and devAddr present. */
-+ if (enet_speed == E_ENET_SPEED_10000)
-+ write_phy_reg_10g(mii_regs, phy_addr, reg, data);
-+ else
-+ write_phy_reg_1g(mii_regs, phy_addr, reg, data);
-+
-+ return 0;
-+}
-+
-+/*****************************************************************************/
-+int fman_memac_mii_read_phy_reg(struct memac_mii_access_mem_map *mii_regs,
-+ uint8_t phy_addr, uint8_t reg, uint16_t *data,
-+ enum enet_speed enet_speed)
-+{
-+ uint32_t ans;
-+ /* Figure out interface type - 10G vs 1G.
-+ In 10G interface both phy_addr and devAddr present. */
-+ if (enet_speed == E_ENET_SPEED_10000)
-+ ans = read_phy_reg_10g(mii_regs, phy_addr, reg, data);
-+ else
-+ ans = read_phy_reg_1g(mii_regs, phy_addr, reg, data);
-+
-+ if (ans & MDIO_CFG_READ_ERR)
-+ return -EINVAL;
-+ return 0;
-+}
-+
-+/* ......................................................................... */
-+
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MAC/fman_tgec.c
-@@ -0,0 +1,367 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+#include "fsl_fman_tgec.h"
-+
-+
-+void fman_tgec_set_mac_address(struct tgec_regs *regs, uint8_t *adr)
-+{
-+ uint32_t tmp0, tmp1;
-+
-+ tmp0 = (uint32_t)(adr[0] |
-+ adr[1] << 8 |
-+ adr[2] << 16 |
-+ adr[3] << 24);
-+ tmp1 = (uint32_t)(adr[4] | adr[5] << 8);
-+ iowrite32be(tmp0, &regs->mac_addr_0);
-+ iowrite32be(tmp1, &regs->mac_addr_1);
-+}
-+
-+void fman_tgec_reset_stat(struct tgec_regs *regs)
-+{
-+ uint32_t tmp;
-+
-+ tmp = ioread32be(&regs->command_config);
-+
-+ tmp |= CMD_CFG_STAT_CLR;
-+
-+ iowrite32be(tmp, &regs->command_config);
-+
-+ while (ioread32be(&regs->command_config) & CMD_CFG_STAT_CLR) ;
-+}
-+
-+#define GET_TGEC_CNTR_64(bn) \
-+ (((uint64_t)ioread32be(&regs->bn ## _u) << 32) | \
-+ ioread32be(&regs->bn ## _l))
-+
-+uint64_t fman_tgec_get_counter(struct tgec_regs *regs, enum tgec_counters reg_name)
-+{
-+ uint64_t ret_val;
-+
-+ switch (reg_name) {
-+ case E_TGEC_COUNTER_R64:
-+ ret_val = GET_TGEC_CNTR_64(r64);
-+ break;
-+ case E_TGEC_COUNTER_R127:
-+ ret_val = GET_TGEC_CNTR_64(r127);
-+ break;
-+ case E_TGEC_COUNTER_R255:
-+ ret_val = GET_TGEC_CNTR_64(r255);
-+ break;
-+ case E_TGEC_COUNTER_R511:
-+ ret_val = GET_TGEC_CNTR_64(r511);
-+ break;
-+ case E_TGEC_COUNTER_R1023:
-+ ret_val = GET_TGEC_CNTR_64(r1023);
-+ break;
-+ case E_TGEC_COUNTER_R1518:
-+ ret_val = GET_TGEC_CNTR_64(r1518);
-+ break;
-+ case E_TGEC_COUNTER_R1519X:
-+ ret_val = GET_TGEC_CNTR_64(r1519x);
-+ break;
-+ case E_TGEC_COUNTER_TRFRG:
-+ ret_val = GET_TGEC_CNTR_64(trfrg);
-+ break;
-+ case E_TGEC_COUNTER_TRJBR:
-+ ret_val = GET_TGEC_CNTR_64(trjbr);
-+ break;
-+ case E_TGEC_COUNTER_RDRP:
-+ ret_val = GET_TGEC_CNTR_64(rdrp);
-+ break;
-+ case E_TGEC_COUNTER_RALN:
-+ ret_val = GET_TGEC_CNTR_64(raln);
-+ break;
-+ case E_TGEC_COUNTER_TRUND:
-+ ret_val = GET_TGEC_CNTR_64(trund);
-+ break;
-+ case E_TGEC_COUNTER_TROVR:
-+ ret_val = GET_TGEC_CNTR_64(trovr);
-+ break;
-+ case E_TGEC_COUNTER_RXPF:
-+ ret_val = GET_TGEC_CNTR_64(rxpf);
-+ break;
-+ case E_TGEC_COUNTER_TXPF:
-+ ret_val = GET_TGEC_CNTR_64(txpf);
-+ break;
-+ case E_TGEC_COUNTER_ROCT:
-+ ret_val = GET_TGEC_CNTR_64(roct);
-+ break;
-+ case E_TGEC_COUNTER_RMCA:
-+ ret_val = GET_TGEC_CNTR_64(rmca);
-+ break;
-+ case E_TGEC_COUNTER_RBCA:
-+ ret_val = GET_TGEC_CNTR_64(rbca);
-+ break;
-+ case E_TGEC_COUNTER_RPKT:
-+ ret_val = GET_TGEC_CNTR_64(rpkt);
-+ break;
-+ case E_TGEC_COUNTER_RUCA:
-+ ret_val = GET_TGEC_CNTR_64(ruca);
-+ break;
-+ case E_TGEC_COUNTER_RERR:
-+ ret_val = GET_TGEC_CNTR_64(rerr);
-+ break;
-+ case E_TGEC_COUNTER_TOCT:
-+ ret_val = GET_TGEC_CNTR_64(toct);
-+ break;
-+ case E_TGEC_COUNTER_TMCA:
-+ ret_val = GET_TGEC_CNTR_64(tmca);
-+ break;
-+ case E_TGEC_COUNTER_TBCA:
-+ ret_val = GET_TGEC_CNTR_64(tbca);
-+ break;
-+ case E_TGEC_COUNTER_TUCA:
-+ ret_val = GET_TGEC_CNTR_64(tuca);
-+ break;
-+ case E_TGEC_COUNTER_TERR:
-+ ret_val = GET_TGEC_CNTR_64(terr);
-+ break;
-+ default:
-+ ret_val = 0;
-+ }
-+
-+ return ret_val;
-+}
-+
-+void fman_tgec_enable(struct tgec_regs *regs, bool apply_rx, bool apply_tx)
-+{
-+ uint32_t tmp;
-+
-+ tmp = ioread32be(&regs->command_config);
-+ if (apply_rx)
-+ tmp |= CMD_CFG_RX_EN;
-+ if (apply_tx)
-+ tmp |= CMD_CFG_TX_EN;
-+ iowrite32be(tmp, &regs->command_config);
-+}
-+
-+void fman_tgec_disable(struct tgec_regs *regs, bool apply_rx, bool apply_tx)
-+{
-+ uint32_t tmp_reg_32;
-+
-+ tmp_reg_32 = ioread32be(&regs->command_config);
-+ if (apply_rx)
-+ tmp_reg_32 &= ~CMD_CFG_RX_EN;
-+ if (apply_tx)
-+ tmp_reg_32 &= ~CMD_CFG_TX_EN;
-+ iowrite32be(tmp_reg_32, &regs->command_config);
-+}
-+
-+void fman_tgec_set_promiscuous(struct tgec_regs *regs, bool val)
-+{
-+ uint32_t tmp;
-+
-+ tmp = ioread32be(&regs->command_config);
-+ if (val)
-+ tmp |= CMD_CFG_PROMIS_EN;
-+ else
-+ tmp &= ~CMD_CFG_PROMIS_EN;
-+ iowrite32be(tmp, &regs->command_config);
-+}
-+
-+void fman_tgec_reset_filter_table(struct tgec_regs *regs)
-+{
-+ uint32_t i;
-+ for (i = 0; i < 512; i++)
-+ iowrite32be(i & ~TGEC_HASH_MCAST_EN, &regs->hashtable_ctrl);
-+}
-+
-+void fman_tgec_set_hash_table_entry(struct tgec_regs *regs, uint32_t crc)
-+{
-+ uint32_t hash = (crc >> TGEC_HASH_MCAST_SHIFT) & TGEC_HASH_ADR_MSK; /* Take 9 MSB bits */
-+ iowrite32be(hash | TGEC_HASH_MCAST_EN, &regs->hashtable_ctrl);
-+}
-+
-+void fman_tgec_set_hash_table(struct tgec_regs *regs, uint32_t value)
-+{
-+ iowrite32be(value, &regs->hashtable_ctrl);
-+}
-+
-+void fman_tgec_set_tx_pause_frames(struct tgec_regs *regs, uint16_t pause_time)
-+{
-+ iowrite32be((uint32_t)pause_time, &regs->pause_quant);
-+}
-+
-+void fman_tgec_set_rx_ignore_pause_frames(struct tgec_regs *regs, bool en)
-+{
-+ uint32_t tmp;
-+
-+ tmp = ioread32be(&regs->command_config);
-+ if (en)
-+ tmp |= CMD_CFG_PAUSE_IGNORE;
-+ else
-+ tmp &= ~CMD_CFG_PAUSE_IGNORE;
-+ iowrite32be(tmp, &regs->command_config);
-+}
-+
-+void fman_tgec_enable_1588_time_stamp(struct tgec_regs *regs, bool en)
-+{
-+ uint32_t tmp;
-+
-+ tmp = ioread32be(&regs->command_config);
-+ if (en)
-+ tmp |= CMD_CFG_EN_TIMESTAMP;
-+ else
-+ tmp &= ~CMD_CFG_EN_TIMESTAMP;
-+ iowrite32be(tmp, &regs->command_config);
-+}
-+
-+uint32_t fman_tgec_get_event(struct tgec_regs *regs, uint32_t ev_mask)
-+{
-+ return ioread32be(&regs->ievent) & ev_mask;
-+}
-+
-+void fman_tgec_ack_event(struct tgec_regs *regs, uint32_t ev_mask)
-+{
-+ iowrite32be(ev_mask, &regs->ievent);
-+}
-+
-+uint32_t fman_tgec_get_interrupt_mask(struct tgec_regs *regs)
-+{
-+ return ioread32be(&regs->imask);
-+}
-+
-+void fman_tgec_add_addr_in_paddr(struct tgec_regs *regs, uint8_t *adr)
-+{
-+ uint32_t tmp0, tmp1;
-+
-+ tmp0 = (uint32_t)(adr[0] |
-+ adr[1] << 8 |
-+ adr[2] << 16 |
-+ adr[3] << 24);
-+ tmp1 = (uint32_t)(adr[4] | adr[5] << 8);
-+ iowrite32be(tmp0, &regs->mac_addr_2);
-+ iowrite32be(tmp1, &regs->mac_addr_3);
-+}
-+
-+void fman_tgec_clear_addr_in_paddr(struct tgec_regs *regs)
-+{
-+ iowrite32be(0, &regs->mac_addr_2);
-+ iowrite32be(0, &regs->mac_addr_3);
-+}
-+
-+uint32_t fman_tgec_get_revision(struct tgec_regs *regs)
-+{
-+ return ioread32be(&regs->tgec_id);
-+}
-+
-+void fman_tgec_enable_interrupt(struct tgec_regs *regs, uint32_t ev_mask)
-+{
-+ iowrite32be(ioread32be(&regs->imask) | ev_mask, &regs->imask);
-+}
-+
-+void fman_tgec_disable_interrupt(struct tgec_regs *regs, uint32_t ev_mask)
-+{
-+ iowrite32be(ioread32be(&regs->imask) & ~ev_mask, &regs->imask);
-+}
-+
-+uint16_t fman_tgec_get_max_frame_len(struct tgec_regs *regs)
-+{
-+ return (uint16_t) ioread32be(&regs->maxfrm);
-+}
-+
-+void fman_tgec_defconfig(struct tgec_cfg *cfg)
-+{
-+ cfg->wan_mode_enable = DEFAULT_WAN_MODE_ENABLE;
-+ cfg->promiscuous_mode_enable = DEFAULT_PROMISCUOUS_MODE_ENABLE;
-+ cfg->pause_forward_enable = DEFAULT_PAUSE_FORWARD_ENABLE;
-+ cfg->pause_ignore = DEFAULT_PAUSE_IGNORE;
-+ cfg->tx_addr_ins_enable = DEFAULT_TX_ADDR_INS_ENABLE;
-+ cfg->loopback_enable = DEFAULT_LOOPBACK_ENABLE;
-+ cfg->cmd_frame_enable = DEFAULT_CMD_FRAME_ENABLE;
-+ cfg->rx_error_discard = DEFAULT_RX_ERROR_DISCARD;
-+ cfg->send_idle_enable = DEFAULT_SEND_IDLE_ENABLE;
-+ cfg->no_length_check_enable = DEFAULT_NO_LENGTH_CHECK_ENABLE;
-+ cfg->lgth_check_nostdr = DEFAULT_LGTH_CHECK_NOSTDR;
-+ cfg->time_stamp_enable = DEFAULT_TIME_STAMP_ENABLE;
-+ cfg->tx_ipg_length = DEFAULT_TX_IPG_LENGTH;
-+ cfg->max_frame_length = DEFAULT_MAX_FRAME_LENGTH;
-+ cfg->pause_quant = DEFAULT_PAUSE_QUANT;
-+#ifdef FM_TX_ECC_FRMS_ERRATA_10GMAC_A004
-+ cfg->skip_fman11_workaround = FALSE;
-+#endif /* FM_TX_ECC_FRMS_ERRATA_10GMAC_A004 */
-+}
-+
-+int fman_tgec_init(struct tgec_regs *regs, struct tgec_cfg *cfg,
-+ uint32_t exception_mask)
-+{
-+ uint32_t tmp;
-+
-+ /* Config */
-+ tmp = 0x40; /* CRC forward */
-+ if (cfg->wan_mode_enable)
-+ tmp |= CMD_CFG_WAN_MODE;
-+ if (cfg->promiscuous_mode_enable)
-+ tmp |= CMD_CFG_PROMIS_EN;
-+ if (cfg->pause_forward_enable)
-+ tmp |= CMD_CFG_PAUSE_FWD;
-+ if (cfg->pause_ignore)
-+ tmp |= CMD_CFG_PAUSE_IGNORE;
-+ if (cfg->tx_addr_ins_enable)
-+ tmp |= CMD_CFG_TX_ADDR_INS;
-+ if (cfg->loopback_enable)
-+ tmp |= CMD_CFG_LOOPBACK_EN;
-+ if (cfg->cmd_frame_enable)
-+ tmp |= CMD_CFG_CMD_FRM_EN;
-+ if (cfg->rx_error_discard)
-+ tmp |= CMD_CFG_RX_ER_DISC;
-+ if (cfg->send_idle_enable)
-+ tmp |= CMD_CFG_SEND_IDLE;
-+ if (cfg->no_length_check_enable)
-+ tmp |= CMD_CFG_NO_LEN_CHK;
-+ if (cfg->time_stamp_enable)
-+ tmp |= CMD_CFG_EN_TIMESTAMP;
-+ iowrite32be(tmp, &regs->command_config);
-+
-+ /* Max Frame Length */
-+ iowrite32be((uint32_t)cfg->max_frame_length, &regs->maxfrm);
-+ /* Pause Time */
-+ iowrite32be(cfg->pause_quant, &regs->pause_quant);
-+
-+ /* clear all pending events and set-up interrupts */
-+ fman_tgec_ack_event(regs, 0xffffffff);
-+ fman_tgec_enable_interrupt(regs, exception_mask);
-+
-+ return 0;
-+}
-+
-+void fman_tgec_set_erratum_tx_fifo_corruption_10gmac_a007(struct tgec_regs *regs)
-+{
-+ uint32_t tmp;
-+
-+ /* restore the default tx ipg Length */
-+ tmp = (ioread32be(&regs->tx_ipg_len) & ~TGEC_TX_IPG_LENGTH_MASK) | 12;
-+
-+ iowrite32be(tmp, &regs->tx_ipg_len);
-+}
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MAC/memac.c
-@@ -0,0 +1,1153 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/******************************************************************************
-+ @File memac.c
-+
-+ @Description FM mEMAC driver
-+*//***************************************************************************/
-+
-+#include "std_ext.h"
-+#include "string_ext.h"
-+#include "error_ext.h"
-+#include "xx_ext.h"
-+#include "endian_ext.h"
-+#include "debug_ext.h"
-+
-+#include "fm_common.h"
-+#include "memac.h"
-+
-+
-+/*****************************************************************************/
-+/* Internal routines */
-+/*****************************************************************************/
-+
-+/* ......................................................................... */
-+
-+static uint32_t GetMacAddrHashCode(uint64_t ethAddr)
-+{
-+ uint64_t mask1, mask2;
-+ uint32_t xorVal = 0;
-+ uint8_t i, j;
-+
-+ for (i=0; i<6; i++)
-+ {
-+ mask1 = ethAddr & (uint64_t)0x01;
-+ ethAddr >>= 1;
-+
-+ for (j=0; j<7; j++)
-+ {
-+ mask2 = ethAddr & (uint64_t)0x01;
-+ mask1 ^= mask2;
-+ ethAddr >>= 1;
-+ }
-+
-+ xorVal |= (mask1 << (5-i));
-+ }
-+
-+ return xorVal;
-+}
-+
-+/* ......................................................................... */
-+
-+static void SetupSgmiiInternalPhy(t_Memac *p_Memac, uint8_t phyAddr)
-+{
-+ uint16_t tmpReg16;
-+ e_EnetMode enetMode;
-+
-+ /* In case the higher MACs are used (i.e. the MACs that should support 10G),
-+ speed=10000 is provided for SGMII ports. Temporary modify enet mode
-+ to 1G one, so MII functions can work correctly. */
-+ enetMode = p_Memac->enetMode;
-+
-+ /* SGMII mode + AN enable */
-+ tmpReg16 = PHY_SGMII_IF_MODE_AN | PHY_SGMII_IF_MODE_SGMII;
-+ if ((p_Memac->enetMode) == e_ENET_MODE_SGMII_2500)
-+ tmpReg16 = PHY_SGMII_CR_PHY_RESET | PHY_SGMII_IF_SPEED_GIGABIT | PHY_SGMII_IF_MODE_SGMII;
-+
-+ p_Memac->enetMode = MAKE_ENET_MODE(ENET_INTERFACE_FROM_MODE(p_Memac->enetMode), e_ENET_SPEED_1000);
-+ MEMAC_MII_WritePhyReg(p_Memac, phyAddr, 0x14, tmpReg16);
-+
-+ /* Device ability according to SGMII specification */
-+ tmpReg16 = PHY_SGMII_DEV_ABILITY_SGMII;
-+ MEMAC_MII_WritePhyReg(p_Memac, phyAddr, 0x4, tmpReg16);
-+
-+ /* Adjust link timer for SGMII -
-+ According to Cisco SGMII specification the timer should be 1.6 ms.
-+ The link_timer register is configured in units of the clock.
-+ - When running as 1G SGMII, Serdes clock is 125 MHz, so
-+ unit = 1 / (125*10^6 Hz) = 8 ns.
-+ 1.6 ms in units of 8 ns = 1.6ms / 8ns = 2 * 10^5 = 0x30d40
-+ - When running as 2.5G SGMII, Serdes clock is 312.5 MHz, so
-+ unit = 1 / (312.5*10^6 Hz) = 3.2 ns.
-+ 1.6 ms in units of 3.2 ns = 1.6ms / 3.2ns = 5 * 10^5 = 0x7a120.
-+ Since link_timer value of 1G SGMII will be too short for 2.5 SGMII,
-+ we always set up here a value of 2.5 SGMII. */
-+ MEMAC_MII_WritePhyReg(p_Memac, phyAddr, 0x13, 0x0007);
-+ MEMAC_MII_WritePhyReg(p_Memac, phyAddr, 0x12, 0xa120);
-+
-+ /* Restart AN */
-+ tmpReg16 = PHY_SGMII_CR_DEF_VAL | PHY_SGMII_CR_RESET_AN;
-+ MEMAC_MII_WritePhyReg(p_Memac, phyAddr, 0x0, tmpReg16);
-+
-+ /* Restore original enet mode */
-+ p_Memac->enetMode = enetMode;
-+}
-+
-+/* ......................................................................... */
-+
-+static void SetupSgmiiInternalPhyBaseX(t_Memac *p_Memac, uint8_t phyAddr)
-+{
-+ uint16_t tmpReg16;
-+ e_EnetMode enetMode;
-+
-+ /* In case the higher MACs are used (i.e. the MACs that should support 10G),
-+ speed=10000 is provided for SGMII ports. Temporary modify enet mode
-+ to 1G one, so MII functions can work correctly. */
-+ enetMode = p_Memac->enetMode;
-+ p_Memac->enetMode = MAKE_ENET_MODE(ENET_INTERFACE_FROM_MODE(p_Memac->enetMode), e_ENET_SPEED_1000);
-+
-+ /* 1000BaseX mode */
-+ tmpReg16 = PHY_SGMII_IF_MODE_1000X;
-+ MEMAC_MII_WritePhyReg(p_Memac, phyAddr, 0x14, tmpReg16);
-+
-+ /* AN Device capability */
-+ tmpReg16 = PHY_SGMII_DEV_ABILITY_1000X;
-+ MEMAC_MII_WritePhyReg(p_Memac, phyAddr, 0x4, tmpReg16);
-+
-+ /* Adjust link timer for SGMII -
-+ For Serdes 1000BaseX auto-negotiation the timer should be 10 ms.
-+ The link_timer register is configured in units of the clock.
-+ - When running as 1G SGMII, Serdes clock is 125 MHz, so
-+ unit = 1 / (125*10^6 Hz) = 8 ns.
-+ 10 ms in units of 8 ns = 10ms / 8ns = 1250000 = 0x1312d0
-+ - When running as 2.5G SGMII, Serdes clock is 312.5 MHz, so
-+ unit = 1 / (312.5*10^6 Hz) = 3.2 ns.
-+ 10 ms in units of 3.2 ns = 10ms / 3.2ns = 3125000 = 0x2faf08.
-+ Since link_timer value of 1G SGMII will be too short for 2.5 SGMII,
-+ we always set up here a value of 2.5 SGMII. */
-+ MEMAC_MII_WritePhyReg(p_Memac, phyAddr, 0x13, 0x002f);
-+ MEMAC_MII_WritePhyReg(p_Memac, phyAddr, 0x12, 0xaf08);
-+
-+ /* Restart AN */
-+ tmpReg16 = PHY_SGMII_CR_DEF_VAL | PHY_SGMII_CR_RESET_AN;
-+ MEMAC_MII_WritePhyReg(p_Memac, phyAddr, 0x0, tmpReg16);
-+
-+ /* Restore original enet mode */
-+ p_Memac->enetMode = enetMode;
-+}
-+
-+/* ......................................................................... */
-+
-+static t_Error CheckInitParameters(t_Memac *p_Memac)
-+{
-+ e_FmMacType portType;
-+
-+ portType = ((ENET_SPEED_FROM_MODE(p_Memac->enetMode) < e_ENET_SPEED_10000) ? e_FM_MAC_1G : e_FM_MAC_10G);
-+
-+#if (FM_MAX_NUM_OF_10G_MACS > 0)
-+ if ((portType == e_FM_MAC_10G) && (p_Memac->macId >= FM_MAX_NUM_OF_10G_MACS))
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("10G MAC ID must be less than %d", FM_MAX_NUM_OF_10G_MACS));
-+#endif /* (FM_MAX_NUM_OF_10G_MACS > 0) */
-+
-+ if ((portType == e_FM_MAC_1G) && (p_Memac->macId >= FM_MAX_NUM_OF_1G_MACS))
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("1G MAC ID must be less than %d", FM_MAX_NUM_OF_1G_MACS));
-+ if (p_Memac->addr == 0)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("Ethernet MAC must have a valid MAC address"));
-+ if (!p_Memac->f_Exception)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("Uninitialized f_Exception"));
-+ if (!p_Memac->f_Event)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("Uninitialized f_Event"));
-+#ifdef FM_LEN_CHECK_ERRATA_FMAN_SW002
-+ if (!p_Memac->p_MemacDriverParam->no_length_check_enable)
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, ("LengthCheck!"));
-+#endif /* FM_LEN_CHECK_ERRATA_FMAN_SW002 */
-+
-+ return E_OK;
-+}
-+
-+/* ........................................................................... */
-+
-+static void MemacErrException(t_Handle h_Memac)
-+{
-+ t_Memac *p_Memac = (t_Memac *)h_Memac;
-+ uint32_t event, imask;
-+
-+ event = fman_memac_get_event(p_Memac->p_MemMap, 0xffffffff);
-+ imask = fman_memac_get_interrupt_mask(p_Memac->p_MemMap);
-+
-+ /* Imask include both error and notification/event bits.
-+ Leaving only error bits enabled by imask.
-+ The imask error bits are shifted by 16 bits offset from
-+ their corresponding location in the ievent - hence the >> 16 */
-+ event &= ((imask & MEMAC_ALL_ERRS_IMASK) >> 16);
-+
-+ fman_memac_ack_event(p_Memac->p_MemMap, event);
-+
-+ if (event & MEMAC_IEVNT_TS_ECC_ER)
-+ p_Memac->f_Exception(p_Memac->h_App, e_FM_MAC_EX_TS_FIFO_ECC_ERR);
-+ if (event & MEMAC_IEVNT_TX_ECC_ER)
-+ p_Memac->f_Exception(p_Memac->h_App, e_FM_MAC_EX_10G_1TX_ECC_ER);
-+ if (event & MEMAC_IEVNT_RX_ECC_ER)
-+ p_Memac->f_Exception(p_Memac->h_App, e_FM_MAC_EX_10G_RX_ECC_ER);
-+}
-+
-+static void MemacException(t_Handle h_Memac)
-+{
-+ t_Memac *p_Memac = (t_Memac *)h_Memac;
-+ uint32_t event, imask;
-+
-+ event = fman_memac_get_event(p_Memac->p_MemMap, 0xffffffff);
-+ imask = fman_memac_get_interrupt_mask(p_Memac->p_MemMap);
-+
-+ /* Imask include both error and notification/event bits.
-+ Leaving only error bits enabled by imask.
-+ The imask error bits are shifted by 16 bits offset from
-+ their corresponding location in the ievent - hence the >> 16 */
-+ event &= ((imask & MEMAC_ALL_ERRS_IMASK) >> 16);
-+
-+ fman_memac_ack_event(p_Memac->p_MemMap, event);
-+
-+ if (event & MEMAC_IEVNT_MGI)
-+ p_Memac->f_Exception(p_Memac->h_App, e_FM_MAC_EX_MAGIC_PACKET_INDICATION);
-+}
-+
-+/* ......................................................................... */
-+
-+static void FreeInitResources(t_Memac *p_Memac)
-+{
-+ e_FmMacType portType;
-+
-+ portType =
-+ ((ENET_SPEED_FROM_MODE(p_Memac->enetMode) < e_ENET_SPEED_10000) ? e_FM_MAC_1G : e_FM_MAC_10G);
-+
-+ if (portType == e_FM_MAC_10G)
-+ FmUnregisterIntr(p_Memac->fmMacControllerDriver.h_Fm, e_FM_MOD_10G_MAC, p_Memac->macId, e_FM_INTR_TYPE_ERR);
-+ else
-+ FmUnregisterIntr(p_Memac->fmMacControllerDriver.h_Fm, e_FM_MOD_1G_MAC, p_Memac->macId, e_FM_INTR_TYPE_ERR);
-+
-+ /* release the driver's group hash table */
-+ FreeHashTable(p_Memac->p_MulticastAddrHash);
-+ p_Memac->p_MulticastAddrHash = NULL;
-+
-+ /* release the driver's individual hash table */
-+ FreeHashTable(p_Memac->p_UnicastAddrHash);
-+ p_Memac->p_UnicastAddrHash = NULL;
-+}
-+
-+
-+/*****************************************************************************/
-+/* mEMAC API routines */
-+/*****************************************************************************/
-+
-+/* ......................................................................... */
-+
-+static t_Error MemacEnable(t_Handle h_Memac, e_CommMode mode)
-+{
-+ t_Memac *p_Memac = (t_Memac *)h_Memac;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Memac, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Memac->p_MemacDriverParam, E_INVALID_STATE);
-+
-+ fman_memac_enable(p_Memac->p_MemMap, (mode & e_COMM_MODE_RX), (mode & e_COMM_MODE_TX));
-+
-+ return E_OK;
-+}
-+
-+/* ......................................................................... */
-+
-+static t_Error MemacDisable (t_Handle h_Memac, e_CommMode mode)
-+{
-+ t_Memac *p_Memac = (t_Memac *)h_Memac;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Memac, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Memac->p_MemacDriverParam, E_INVALID_STATE);
-+
-+ fman_memac_disable(p_Memac->p_MemMap, (mode & e_COMM_MODE_RX), (mode & e_COMM_MODE_TX));
-+
-+ return E_OK;
-+}
-+
-+/* ......................................................................... */
-+
-+static t_Error MemacSetPromiscuous(t_Handle h_Memac, bool newVal)
-+{
-+ t_Memac *p_Memac = (t_Memac *)h_Memac;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Memac, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Memac->p_MemacDriverParam, E_INVALID_STATE);
-+
-+ fman_memac_set_promiscuous(p_Memac->p_MemMap, newVal);
-+
-+ return E_OK;
-+}
-+
-+/* .............................................................................. */
-+
-+static t_Error MemacAdjustLink(t_Handle h_Memac, e_EnetSpeed speed, bool fullDuplex)
-+{
-+ t_Memac *p_Memac = (t_Memac *)h_Memac;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Memac, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Memac->p_MemacDriverParam, E_INVALID_STATE);
-+
-+ if ((speed >= e_ENET_SPEED_1000) && (!fullDuplex))
-+ RETURN_ERROR(MAJOR, E_CONFLICT,
-+ ("Ethernet MAC 1G or 10G does not support half-duplex"));
-+
-+ fman_memac_adjust_link(p_Memac->p_MemMap,
-+ (enum enet_interface)ENET_INTERFACE_FROM_MODE(p_Memac->enetMode),
-+ (enum enet_speed)speed,
-+ fullDuplex);
-+ return E_OK;
-+}
-+
-+
-+/*****************************************************************************/
-+/* Memac Configs modification functions */
-+/*****************************************************************************/
-+
-+/* ......................................................................... */
-+
-+static t_Error MemacConfigLoopback(t_Handle h_Memac, bool newVal)
-+{
-+ t_Memac *p_Memac = (t_Memac *)h_Memac;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Memac, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Memac->p_MemacDriverParam, E_INVALID_STATE);
-+
-+ p_Memac->p_MemacDriverParam->loopback_enable = newVal;
-+
-+ return E_OK;
-+}
-+
-+/* ......................................................................... */
-+
-+static t_Error MemacConfigWan(t_Handle h_Memac, bool newVal)
-+{
-+ t_Memac *p_Memac = (t_Memac *)h_Memac;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Memac, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Memac->p_MemacDriverParam, E_INVALID_STATE);
-+
-+ p_Memac->p_MemacDriverParam->wan_mode_enable = newVal;
-+
-+ return E_OK;
-+}
-+
-+/* ......................................................................... */
-+
-+static t_Error MemacConfigMaxFrameLength(t_Handle h_Memac, uint16_t newVal)
-+{
-+ t_Memac *p_Memac = (t_Memac *)h_Memac;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Memac, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Memac->p_MemacDriverParam, E_INVALID_STATE);
-+
-+ p_Memac->p_MemacDriverParam->max_frame_length = newVal;
-+
-+ return E_OK;
-+}
-+
-+/* ......................................................................... */
-+
-+static t_Error MemacConfigPad(t_Handle h_Memac, bool newVal)
-+{
-+ t_Memac *p_Memac = (t_Memac *)h_Memac;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Memac, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Memac->p_MemacDriverParam, E_INVALID_STATE);
-+
-+ p_Memac->p_MemacDriverParam->pad_enable = newVal;
-+
-+ return E_OK;
-+}
-+
-+/* ......................................................................... */
-+
-+static t_Error MemacConfigLengthCheck(t_Handle h_Memac, bool newVal)
-+{
-+ t_Memac *p_Memac = (t_Memac *)h_Memac;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Memac, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Memac->p_MemacDriverParam, E_INVALID_STATE);
-+
-+ p_Memac->p_MemacDriverParam->no_length_check_enable = !newVal;
-+
-+ return E_OK;
-+}
-+
-+/* ......................................................................... */
-+
-+static t_Error MemacConfigException(t_Handle h_Memac, e_FmMacExceptions exception, bool enable)
-+{
-+ t_Memac *p_Memac = (t_Memac *)h_Memac;
-+ uint32_t bitMask = 0;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Memac, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Memac->p_MemacDriverParam, E_INVALID_STATE);
-+
-+ GET_EXCEPTION_FLAG(bitMask, exception);
-+ if (bitMask)
-+ {
-+ if (enable)
-+ p_Memac->exceptions |= bitMask;
-+ else
-+ p_Memac->exceptions &= ~bitMask;
-+ }
-+ else
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("Undefined exception"));
-+
-+ return E_OK;
-+}
-+
-+/* ......................................................................... */
-+
-+static t_Error MemacConfigResetOnInit(t_Handle h_Memac, bool enable)
-+{
-+ t_Memac *p_Memac = (t_Memac *)h_Memac;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Memac, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Memac->p_MemacDriverParam, E_INVALID_STATE);
-+
-+ p_Memac->p_MemacDriverParam->reset_on_init = enable;
-+
-+ return E_OK;
-+}
-+
-+
-+/*****************************************************************************/
-+/* Memac Run Time API functions */
-+/*****************************************************************************/
-+
-+/* ......................................................................... */
-+
-+static t_Error MemacSetTxPauseFrames(t_Handle h_Memac,
-+ uint8_t priority,
-+ uint16_t pauseTime,
-+ uint16_t threshTime)
-+{
-+ t_Memac *p_Memac = (t_Memac *)h_Memac;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Memac, E_INVALID_STATE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Memac->p_MemacDriverParam, E_INVALID_STATE);
-+
-+ if (priority != 0xFF)
-+ {
-+ bool PortConfigured, PreFetchEnabled;
-+
-+ if (FmGetTnumAgingPeriod(p_Memac->fmMacControllerDriver.h_Fm) == 0)
-+ RETURN_ERROR(MAJOR, E_CONFLICT, ("For PFC operation, TNUM aging must be enabled"));
-+
-+ FmGetPortPreFetchConfiguration(p_Memac->fmMacControllerDriver.h_Fm,
-+ p_Memac->fmMacControllerDriver.macId,
-+ &PortConfigured,
-+ &PreFetchEnabled);
-+
-+ if ((ENET_SPEED_FROM_MODE(p_Memac->fmMacControllerDriver.enetMode) == e_ENET_SPEED_1000) && !PortConfigured)
-+ DBG(INFO, ("For PFC correct operation, prefetch must be configured on the FM Tx PORT"));
-+
-+ if ((ENET_SPEED_FROM_MODE(p_Memac->fmMacControllerDriver.enetMode) == e_ENET_SPEED_1000) && PortConfigured && !PreFetchEnabled)
-+ DBG(WARNING, ("For PFC correct operation, prefetch must be configured on the FM Tx PORT"));
-+ }
-+
-+ fman_memac_set_tx_pause_frames(p_Memac->p_MemMap, priority, pauseTime, threshTime);
-+
-+ return E_OK;
-+}
-+
-+/* ......................................................................... */
-+
-+static t_Error MemacSetTxAutoPauseFrames(t_Handle h_Memac,
-+ uint16_t pauseTime)
-+{
-+ return MemacSetTxPauseFrames(h_Memac, FM_MAC_NO_PFC, pauseTime, 0);
-+}
-+
-+/* ......................................................................... */
-+
-+static t_Error MemacSetRxIgnorePauseFrames(t_Handle h_Memac, bool en)
-+{
-+ t_Memac *p_Memac = (t_Memac *)h_Memac;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Memac, E_INVALID_STATE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Memac->p_MemacDriverParam, E_INVALID_STATE);
-+
-+ fman_memac_set_rx_ignore_pause_frames(p_Memac->p_MemMap, en);
-+
-+ return E_OK;
-+}
-+
-+/* ......................................................................... */
-+
-+static t_Error MemacSetWakeOnLan(t_Handle h_Memac, bool en)
-+{
-+ t_Memac *p_Memac = (t_Memac *)h_Memac;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Memac, E_INVALID_STATE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Memac->p_MemacDriverParam, E_INVALID_STATE);
-+
-+ fman_memac_set_wol(p_Memac->p_MemMap, en);
-+
-+ return E_OK;
-+}
-+
-+/* .............................................................................. */
-+
-+static t_Error MemacEnable1588TimeStamp(t_Handle h_Memac)
-+{
-+ t_Memac *p_Memac = (t_Memac *)h_Memac;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Memac, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Memac->p_MemacDriverParam, E_INVALID_STATE);
-+UNUSED(p_Memac);
-+DBG(WARNING, ("mEMAC has 1588 always enabled!"));
-+
-+ return E_OK;
-+}
-+
-+/* Counters handling */
-+/* ......................................................................... */
-+
-+static t_Error MemacGetStatistics(t_Handle h_Memac, t_FmMacStatistics *p_Statistics)
-+{
-+ t_Memac *p_Memac = (t_Memac *)h_Memac;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Memac, E_NULL_POINTER);
-+ SANITY_CHECK_RETURN_ERROR(!p_Memac->p_MemacDriverParam, E_INVALID_STATE);
-+ SANITY_CHECK_RETURN_ERROR(p_Statistics, E_NULL_POINTER);
-+
-+ p_Statistics->eStatPkts64 = fman_memac_get_counter(p_Memac->p_MemMap, E_MEMAC_COUNTER_R64);
-+ p_Statistics->eStatPkts65to127 = fman_memac_get_counter(p_Memac->p_MemMap, E_MEMAC_COUNTER_R127);
-+ p_Statistics->eStatPkts128to255 = fman_memac_get_counter(p_Memac->p_MemMap, E_MEMAC_COUNTER_R255);
-+ p_Statistics->eStatPkts256to511 = fman_memac_get_counter(p_Memac->p_MemMap, E_MEMAC_COUNTER_R511);
-+ p_Statistics->eStatPkts512to1023 = fman_memac_get_counter(p_Memac->p_MemMap, E_MEMAC_COUNTER_R1023);
-+ p_Statistics->eStatPkts1024to1518 = fman_memac_get_counter(p_Memac->p_MemMap, E_MEMAC_COUNTER_R1518);
-+ p_Statistics->eStatPkts1519to1522 = fman_memac_get_counter(p_Memac->p_MemMap, E_MEMAC_COUNTER_R1519X);
-+/* */
-+ p_Statistics->eStatFragments = fman_memac_get_counter(p_Memac->p_MemMap, E_MEMAC_COUNTER_RFRG);
-+ p_Statistics->eStatJabbers = fman_memac_get_counter(p_Memac->p_MemMap, E_MEMAC_COUNTER_RJBR);
-+
-+ p_Statistics->eStatsDropEvents = fman_memac_get_counter(p_Memac->p_MemMap, E_MEMAC_COUNTER_RDRP);
-+ p_Statistics->eStatCRCAlignErrors = fman_memac_get_counter(p_Memac->p_MemMap, E_MEMAC_COUNTER_RALN);
-+
-+ p_Statistics->eStatUndersizePkts = fman_memac_get_counter(p_Memac->p_MemMap, E_MEMAC_COUNTER_TUND);
-+ p_Statistics->eStatOversizePkts = fman_memac_get_counter(p_Memac->p_MemMap, E_MEMAC_COUNTER_ROVR);
-+/* Pause */
-+ p_Statistics->reStatPause = fman_memac_get_counter(p_Memac->p_MemMap, E_MEMAC_COUNTER_RXPF);
-+ p_Statistics->teStatPause = fman_memac_get_counter(p_Memac->p_MemMap, E_MEMAC_COUNTER_TXPF);
-+
-+/* MIB II */
-+ p_Statistics->ifInOctets = fman_memac_get_counter(p_Memac->p_MemMap, E_MEMAC_COUNTER_ROCT);
-+ p_Statistics->ifInUcastPkts = fman_memac_get_counter(p_Memac->p_MemMap, E_MEMAC_COUNTER_RUCA);
-+ p_Statistics->ifInMcastPkts = fman_memac_get_counter(p_Memac->p_MemMap, E_MEMAC_COUNTER_RMCA);
-+ p_Statistics->ifInBcastPkts = fman_memac_get_counter(p_Memac->p_MemMap, E_MEMAC_COUNTER_RBCA);
-+ p_Statistics->ifInPkts = p_Statistics->ifInUcastPkts
-+ + p_Statistics->ifInMcastPkts
-+ + p_Statistics->ifInBcastPkts;
-+ p_Statistics->ifInDiscards = 0;
-+ p_Statistics->ifInErrors = fman_memac_get_counter(p_Memac->p_MemMap, E_MEMAC_COUNTER_RERR);
-+
-+ p_Statistics->ifOutOctets = fman_memac_get_counter(p_Memac->p_MemMap, E_MEMAC_COUNTER_TOCT);
-+ p_Statistics->ifOutUcastPkts = fman_memac_get_counter(p_Memac->p_MemMap, E_MEMAC_COUNTER_TUCA);
-+ p_Statistics->ifOutMcastPkts = fman_memac_get_counter(p_Memac->p_MemMap, E_MEMAC_COUNTER_TMCA);
-+ p_Statistics->ifOutBcastPkts = fman_memac_get_counter(p_Memac->p_MemMap, E_MEMAC_COUNTER_TBCA);
-+ p_Statistics->ifOutPkts = p_Statistics->ifOutUcastPkts
-+ + p_Statistics->ifOutMcastPkts
-+ + p_Statistics->ifOutBcastPkts;
-+ p_Statistics->ifOutDiscards = 0;
-+ p_Statistics->ifOutErrors = fman_memac_get_counter(p_Memac->p_MemMap, E_MEMAC_COUNTER_TERR);
-+
-+ return E_OK;
-+}
-+
-+/* ......................................................................... */
-+
-+static t_Error MemacGetFrameSizeCounters(t_Handle h_Memac, t_FmMacFrameSizeCounters *p_FrameSizeCounters, e_CommMode type)
-+{
-+ t_Memac *p_Memac = (t_Memac *)h_Memac;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Memac, E_NULL_POINTER);
-+ SANITY_CHECK_RETURN_ERROR(!p_Memac->p_MemacDriverParam, E_INVALID_STATE);
-+ SANITY_CHECK_RETURN_ERROR(p_FrameSizeCounters, E_NULL_POINTER);
-+
-+ switch (type)
-+ {
-+ case e_COMM_MODE_NONE:
-+ break;
-+
-+ case e_COMM_MODE_RX:
-+ p_FrameSizeCounters->count_pkts_64 = fman_memac_get_counter(p_Memac->p_MemMap, E_MEMAC_COUNTER_R64);
-+ p_FrameSizeCounters->count_pkts_65_to_127 = fman_memac_get_counter(p_Memac->p_MemMap, E_MEMAC_COUNTER_R127);
-+ p_FrameSizeCounters->count_pkts_128_to_255 = fman_memac_get_counter(p_Memac->p_MemMap, E_MEMAC_COUNTER_R255);
-+ p_FrameSizeCounters->count_pkts_256_to_511 = fman_memac_get_counter(p_Memac->p_MemMap, E_MEMAC_COUNTER_R511);
-+ p_FrameSizeCounters->count_pkts_512_to_1023 = fman_memac_get_counter(p_Memac->p_MemMap, E_MEMAC_COUNTER_R1023);
-+ p_FrameSizeCounters->count_pkts_1024_to_1518 = fman_memac_get_counter(p_Memac->p_MemMap, E_MEMAC_COUNTER_R1518);
-+ p_FrameSizeCounters->count_pkts_1519_to_1522 = fman_memac_get_counter(p_Memac->p_MemMap, E_MEMAC_COUNTER_R1519X);
-+ break;
-+
-+ case e_COMM_MODE_TX:
-+ p_FrameSizeCounters->count_pkts_64 = fman_memac_get_counter(p_Memac->p_MemMap, E_MEMAC_COUNTER_T64);
-+ p_FrameSizeCounters->count_pkts_65_to_127 = fman_memac_get_counter(p_Memac->p_MemMap, E_MEMAC_COUNTER_T127);
-+ p_FrameSizeCounters->count_pkts_128_to_255 = fman_memac_get_counter(p_Memac->p_MemMap, E_MEMAC_COUNTER_T255);
-+ p_FrameSizeCounters->count_pkts_256_to_511 = fman_memac_get_counter(p_Memac->p_MemMap, E_MEMAC_COUNTER_T511);
-+ p_FrameSizeCounters->count_pkts_512_to_1023 = fman_memac_get_counter(p_Memac->p_MemMap, E_MEMAC_COUNTER_T1023);
-+ p_FrameSizeCounters->count_pkts_1024_to_1518 = fman_memac_get_counter(p_Memac->p_MemMap, E_MEMAC_COUNTER_T1518);
-+ p_FrameSizeCounters->count_pkts_1519_to_1522 = fman_memac_get_counter(p_Memac->p_MemMap, E_MEMAC_COUNTER_T1519X);
-+ break;
-+
-+ case e_COMM_MODE_RX_AND_TX:
-+ p_FrameSizeCounters->count_pkts_64 = fman_memac_get_counter(p_Memac->p_MemMap, E_MEMAC_COUNTER_R64)
-+ + fman_memac_get_counter(p_Memac->p_MemMap, E_MEMAC_COUNTER_T64);
-+ p_FrameSizeCounters->count_pkts_65_to_127 = fman_memac_get_counter(p_Memac->p_MemMap, E_MEMAC_COUNTER_R127)
-+ + fman_memac_get_counter(p_Memac->p_MemMap, E_MEMAC_COUNTER_T127);
-+ p_FrameSizeCounters->count_pkts_128_to_255 = fman_memac_get_counter(p_Memac->p_MemMap, E_MEMAC_COUNTER_R255)
-+ + fman_memac_get_counter(p_Memac->p_MemMap, E_MEMAC_COUNTER_T255);
-+ p_FrameSizeCounters->count_pkts_256_to_511 = fman_memac_get_counter(p_Memac->p_MemMap, E_MEMAC_COUNTER_R511)
-+ + fman_memac_get_counter(p_Memac->p_MemMap, E_MEMAC_COUNTER_T511);
-+ p_FrameSizeCounters->count_pkts_512_to_1023 = fman_memac_get_counter(p_Memac->p_MemMap, E_MEMAC_COUNTER_R1023)
-+ + fman_memac_get_counter(p_Memac->p_MemMap, E_MEMAC_COUNTER_T1023);
-+ p_FrameSizeCounters->count_pkts_1024_to_1518 = fman_memac_get_counter(p_Memac->p_MemMap, E_MEMAC_COUNTER_R1518)
-+ + fman_memac_get_counter(p_Memac->p_MemMap, E_MEMAC_COUNTER_T1518);
-+ p_FrameSizeCounters->count_pkts_1519_to_1522 = fman_memac_get_counter(p_Memac->p_MemMap, E_MEMAC_COUNTER_R1519X)
-+ + fman_memac_get_counter(p_Memac->p_MemMap, E_MEMAC_COUNTER_T1519X);
-+ break;
-+ }
-+
-+ return E_OK;
-+}
-+
-+/* ......................................................................... */
-+
-+static t_Error MemacModifyMacAddress (t_Handle h_Memac, t_EnetAddr *p_EnetAddr)
-+{
-+ t_Memac *p_Memac = (t_Memac *)h_Memac;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Memac, E_NULL_POINTER);
-+ SANITY_CHECK_RETURN_ERROR(!p_Memac->p_MemacDriverParam, E_INVALID_STATE);
-+
-+ fman_memac_add_addr_in_paddr(p_Memac->p_MemMap, (uint8_t *)(*p_EnetAddr), 0);
-+
-+ return E_OK;
-+}
-+
-+/* ......................................................................... */
-+
-+static t_Error MemacResetCounters (t_Handle h_Memac)
-+{
-+ t_Memac *p_Memac = (t_Memac *)h_Memac;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Memac, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Memac->p_MemacDriverParam, E_INVALID_STATE);
-+
-+ fman_memac_reset_stat(p_Memac->p_MemMap);
-+
-+ return E_OK;
-+}
-+
-+/* ......................................................................... */
-+
-+static t_Error MemacAddExactMatchMacAddress(t_Handle h_Memac, t_EnetAddr *p_EthAddr)
-+{
-+ t_Memac *p_Memac = (t_Memac *) h_Memac;
-+ uint64_t ethAddr;
-+ uint8_t paddrNum;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Memac, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Memac->p_MemacDriverParam, E_INVALID_STATE);
-+
-+ ethAddr = ENET_ADDR_TO_UINT64(*p_EthAddr);
-+
-+ if (ethAddr & GROUP_ADDRESS)
-+ /* Multicast address has no effect in PADDR */
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("Multicast address"));
-+
-+ /* Make sure no PADDR contains this address */
-+ for (paddrNum = 0; paddrNum < MEMAC_NUM_OF_PADDRS; paddrNum++)
-+ if (p_Memac->indAddrRegUsed[paddrNum])
-+ if (p_Memac->paddr[paddrNum] == ethAddr)
-+ RETURN_ERROR(MAJOR, E_ALREADY_EXISTS, NO_MSG);
-+
-+ /* Find first unused PADDR */
-+ for (paddrNum = 0; paddrNum < MEMAC_NUM_OF_PADDRS; paddrNum++)
-+ if (!(p_Memac->indAddrRegUsed[paddrNum]))
-+ {
-+ /* mark this PADDR as used */
-+ p_Memac->indAddrRegUsed[paddrNum] = TRUE;
-+ /* store address */
-+ p_Memac->paddr[paddrNum] = ethAddr;
-+
-+ /* put in hardware */
-+ fman_memac_add_addr_in_paddr(p_Memac->p_MemMap, (uint8_t*)(*p_EthAddr), paddrNum);
-+ p_Memac->numOfIndAddrInRegs++;
-+
-+ return E_OK;
-+ }
-+
-+ /* No free PADDR */
-+ RETURN_ERROR(MAJOR, E_FULL, NO_MSG);
-+}
-+
-+/* ......................................................................... */
-+
-+static t_Error MemacDelExactMatchMacAddress(t_Handle h_Memac, t_EnetAddr *p_EthAddr)
-+{
-+ t_Memac *p_Memac = (t_Memac *) h_Memac;
-+ uint64_t ethAddr;
-+ uint8_t paddrNum;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Memac, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Memac->p_MemacDriverParam, E_INVALID_STATE);
-+
-+ ethAddr = ENET_ADDR_TO_UINT64(*p_EthAddr);
-+
-+ /* Find used PADDR containing this address */
-+ for (paddrNum = 0; paddrNum < MEMAC_NUM_OF_PADDRS; paddrNum++)
-+ {
-+ if ((p_Memac->indAddrRegUsed[paddrNum]) &&
-+ (p_Memac->paddr[paddrNum] == ethAddr))
-+ {
-+ /* mark this PADDR as not used */
-+ p_Memac->indAddrRegUsed[paddrNum] = FALSE;
-+ /* clear in hardware */
-+ fman_memac_clear_addr_in_paddr(p_Memac->p_MemMap, paddrNum);
-+ p_Memac->numOfIndAddrInRegs--;
-+
-+ return E_OK;
-+ }
-+ }
-+
-+ RETURN_ERROR(MAJOR, E_NOT_FOUND, NO_MSG);
-+}
-+
-+/* ......................................................................... */
-+
-+static t_Error MemacGetId(t_Handle h_Memac, uint32_t *macId)
-+{
-+ t_Memac *p_Memac = (t_Memac *)h_Memac;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Memac, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Memac->p_MemacDriverParam, E_INVALID_STATE);
-+
-+ *macId = p_Memac->macId;
-+
-+ return E_OK;
-+}
-+
-+/* ......................................................................... */
-+
-+
-+static t_Error MemacAddHashMacAddress(t_Handle h_Memac, t_EnetAddr *p_EthAddr)
-+{
-+ t_Memac *p_Memac = (t_Memac *)h_Memac;
-+ t_EthHashEntry *p_HashEntry;
-+ uint32_t hash;
-+ uint64_t ethAddr;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Memac, E_NULL_POINTER);
-+ SANITY_CHECK_RETURN_ERROR(!p_Memac->p_MemacDriverParam, E_INVALID_STATE);
-+
-+ ethAddr = ENET_ADDR_TO_UINT64(*p_EthAddr);
-+
-+ if (!(ethAddr & GROUP_ADDRESS))
-+ /* Unicast addresses not supported in hash */
-+ RETURN_ERROR(MAJOR, E_NOT_SUPPORTED, ("Unicast Address"));
-+
-+ hash = GetMacAddrHashCode(ethAddr) & HASH_CTRL_ADDR_MASK;
-+
-+ /* Create element to be added to the driver hash table */
-+ p_HashEntry = (t_EthHashEntry *)XX_Malloc(sizeof(t_EthHashEntry));
-+ p_HashEntry->addr = ethAddr;
-+ INIT_LIST(&p_HashEntry->node);
-+
-+ LIST_AddToTail(&(p_HashEntry->node), &(p_Memac->p_MulticastAddrHash->p_Lsts[hash]));
-+ fman_memac_set_hash_table(p_Memac->p_MemMap, (hash | HASH_CTRL_MCAST_EN));
-+
-+ return E_OK;
-+}
-+
-+/* ......................................................................... */
-+
-+static t_Error MemacDelHashMacAddress(t_Handle h_Memac, t_EnetAddr *p_EthAddr)
-+{
-+ t_Memac *p_Memac = (t_Memac *)h_Memac;
-+ t_EthHashEntry *p_HashEntry = NULL;
-+ t_List *p_Pos;
-+ uint32_t hash;
-+ uint64_t ethAddr;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Memac, E_NULL_POINTER);
-+ SANITY_CHECK_RETURN_ERROR(!p_Memac->p_MemacDriverParam, E_INVALID_STATE);
-+
-+ ethAddr = ENET_ADDR_TO_UINT64(*p_EthAddr);
-+
-+ hash = GetMacAddrHashCode(ethAddr) & HASH_CTRL_ADDR_MASK;
-+
-+ LIST_FOR_EACH(p_Pos, &(p_Memac->p_MulticastAddrHash->p_Lsts[hash]))
-+ {
-+ p_HashEntry = ETH_HASH_ENTRY_OBJ(p_Pos);
-+ if (p_HashEntry->addr == ethAddr)
-+ {
-+ LIST_DelAndInit(&p_HashEntry->node);
-+ XX_Free(p_HashEntry);
-+ break;
-+ }
-+ }
-+ if (LIST_IsEmpty(&p_Memac->p_MulticastAddrHash->p_Lsts[hash]))
-+ fman_memac_set_hash_table(p_Memac->p_MemMap, (hash & ~HASH_CTRL_MCAST_EN));
-+
-+ return E_OK;
-+}
-+
-+
-+/* ......................................................................... */
-+
-+static t_Error MemacSetException(t_Handle h_Memac, e_FmMacExceptions exception, bool enable)
-+{
-+ t_Memac *p_Memac = (t_Memac *)h_Memac;
-+ uint32_t bitMask = 0;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Memac, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Memac->p_MemacDriverParam, E_INVALID_STATE);
-+
-+ GET_EXCEPTION_FLAG(bitMask, exception);
-+ if (bitMask)
-+ {
-+ if (enable)
-+ p_Memac->exceptions |= bitMask;
-+ else
-+ p_Memac->exceptions &= ~bitMask;
-+ }
-+ else
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("Undefined exception"));
-+
-+ fman_memac_set_exception(p_Memac->p_MemMap, bitMask, enable);
-+
-+ return E_OK;
-+}
-+
-+/* ......................................................................... */
-+
-+static uint16_t MemacGetMaxFrameLength(t_Handle h_Memac)
-+{
-+ t_Memac *p_Memac = (t_Memac *)h_Memac;
-+
-+ SANITY_CHECK_RETURN_VALUE(p_Memac, E_INVALID_HANDLE, 0);
-+ SANITY_CHECK_RETURN_VALUE(!p_Memac->p_MemacDriverParam, E_INVALID_STATE, 0);
-+
-+ return fman_memac_get_max_frame_len(p_Memac->p_MemMap);
-+}
-+
-+static t_Error MemacInitInternalPhy(t_Handle h_Memac)
-+{
-+ t_Memac *p_Memac = (t_Memac *)h_Memac;
-+ uint8_t i, phyAddr;
-+
-+ if (ENET_INTERFACE_FROM_MODE(p_Memac->enetMode) == e_ENET_IF_SGMII)
-+ {
-+ /* Configure internal SGMII PHY */
-+ if (p_Memac->enetMode & ENET_IF_SGMII_BASEX)
-+ SetupSgmiiInternalPhyBaseX(p_Memac, PHY_MDIO_ADDR);
-+ else
-+ SetupSgmiiInternalPhy(p_Memac, PHY_MDIO_ADDR);
-+ }
-+ else if (ENET_INTERFACE_FROM_MODE(p_Memac->enetMode) == e_ENET_IF_QSGMII)
-+ {
-+ /* Configure 4 internal SGMII PHYs */
-+ for (i = 0; i < 4; i++)
-+ {
-+ /* QSGMII PHY address occupies 3 upper bits of 5-bit
-+ phyAddress; the lower 2 bits are used to extend
-+ register address space and access each one of 4
-+ ports inside QSGMII. */
-+ phyAddr = (uint8_t)((PHY_MDIO_ADDR << 2) | i);
-+ if (p_Memac->enetMode & ENET_IF_SGMII_BASEX)
-+ SetupSgmiiInternalPhyBaseX(p_Memac, phyAddr);
-+ else
-+ SetupSgmiiInternalPhy(p_Memac, phyAddr);
-+ }
-+ }
-+ return E_OK;
-+}
-+
-+/*****************************************************************************/
-+/* mEMAC Init & Free API */
-+/*****************************************************************************/
-+
-+/* ......................................................................... */
-+void *g_MemacRegs;
-+static t_Error MemacInit(t_Handle h_Memac)
-+{
-+ t_Memac *p_Memac = (t_Memac *)h_Memac;
-+ struct memac_cfg *p_MemacDriverParam;
-+ enum enet_interface enet_interface;
-+ enum enet_speed enet_speed;
-+ t_EnetAddr ethAddr;
-+ e_FmMacType portType;
-+ t_Error err;
-+ bool slow_10g_if = FALSE;
-+ if (p_Memac->macId == 3) /* This is a quick WA */
-+ g_MemacRegs = p_Memac->p_MemMap;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Memac, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Memac->p_MemacDriverParam, E_INVALID_STATE);
-+ SANITY_CHECK_RETURN_ERROR(p_Memac->fmMacControllerDriver.h_Fm, E_INVALID_HANDLE);
-+
-+ FM_GetRevision(p_Memac->fmMacControllerDriver.h_Fm, &p_Memac->fmMacControllerDriver.fmRevInfo);
-+ if (p_Memac->fmMacControllerDriver.fmRevInfo.majorRev == 6 &&
-+ p_Memac->fmMacControllerDriver.fmRevInfo.minorRev == 4)
-+ slow_10g_if = TRUE;
-+
-+ CHECK_INIT_PARAMETERS(p_Memac, CheckInitParameters);
-+
-+ p_MemacDriverParam = p_Memac->p_MemacDriverParam;
-+
-+ portType =
-+ ((ENET_SPEED_FROM_MODE(p_Memac->enetMode) < e_ENET_SPEED_10000) ? e_FM_MAC_1G : e_FM_MAC_10G);
-+
-+ /* First, reset the MAC if desired. */
-+ if (p_MemacDriverParam->reset_on_init)
-+ fman_memac_reset(p_Memac->p_MemMap);
-+
-+ /* MAC Address */
-+ MAKE_ENET_ADDR_FROM_UINT64(p_Memac->addr, ethAddr);
-+ fman_memac_add_addr_in_paddr(p_Memac->p_MemMap, (uint8_t*)ethAddr, 0);
-+
-+ enet_interface = (enum enet_interface) ENET_INTERFACE_FROM_MODE(p_Memac->enetMode);
-+ enet_speed = (enum enet_speed) ENET_SPEED_FROM_MODE(p_Memac->enetMode);
-+
-+ fman_memac_init(p_Memac->p_MemMap,
-+ p_Memac->p_MemacDriverParam,
-+ enet_interface,
-+ enet_speed,
-+ slow_10g_if,
-+ p_Memac->exceptions);
-+
-+#ifdef FM_RX_FIFO_CORRUPT_ERRATA_10GMAC_A006320
-+ {
-+ uint32_t tmpReg = 0;
-+
-+ FM_GetRevision(p_Memac->fmMacControllerDriver.h_Fm, &p_Memac->fmMacControllerDriver.fmRevInfo);
-+ /* check the FMAN version - the bug exists only in rev1 */
-+ if ((p_Memac->fmMacControllerDriver.fmRevInfo.majorRev == 6) &&
-+ (p_Memac->fmMacControllerDriver.fmRevInfo.minorRev == 0))
-+ {
-+ /* MAC strips CRC from received frames - this workaround should
-+ decrease the likelihood of bug appearance
-+ */
-+ tmpReg = GET_UINT32(p_Memac->p_MemMap->command_config);
-+ tmpReg &= ~CMD_CFG_CRC_FWD;
-+ WRITE_UINT32(p_Memac->p_MemMap->command_config, tmpReg);
-+ /* DBG(WARNING, ("mEMAC strips CRC from received frames as part of A006320 errata workaround"));*/
-+ }
-+ }
-+#endif /* FM_RX_FIFO_CORRUPT_ERRATA_10GMAC_A006320 */
-+
-+ MemacInitInternalPhy(h_Memac);
-+
-+ /* Max Frame Length */
-+ err = FmSetMacMaxFrame(p_Memac->fmMacControllerDriver.h_Fm,
-+ portType,
-+ p_Memac->fmMacControllerDriver.macId,
-+ p_MemacDriverParam->max_frame_length);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, ("settings Mac max frame length is FAILED"));
-+
-+ p_Memac->p_MulticastAddrHash = AllocHashTable(HASH_TABLE_SIZE);
-+ if (!p_Memac->p_MulticastAddrHash)
-+ {
-+ FreeInitResources(p_Memac);
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, ("allocation hash table is FAILED"));
-+ }
-+
-+ p_Memac->p_UnicastAddrHash = AllocHashTable(HASH_TABLE_SIZE);
-+ if (!p_Memac->p_UnicastAddrHash)
-+ {
-+ FreeInitResources(p_Memac);
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, ("allocation hash table is FAILED"));
-+ }
-+
-+ FmRegisterIntr(p_Memac->fmMacControllerDriver.h_Fm,
-+ (portType == e_FM_MAC_10G) ? e_FM_MOD_10G_MAC : e_FM_MOD_1G_MAC,
-+ p_Memac->macId,
-+ e_FM_INTR_TYPE_ERR,
-+ MemacErrException,
-+ p_Memac);
-+
-+ FmRegisterIntr(p_Memac->fmMacControllerDriver.h_Fm,
-+ (portType == e_FM_MAC_10G) ? e_FM_MOD_10G_MAC : e_FM_MOD_1G_MAC,
-+ p_Memac->macId,
-+ e_FM_INTR_TYPE_NORMAL,
-+ MemacException,
-+ p_Memac);
-+
-+ XX_Free(p_MemacDriverParam);
-+ p_Memac->p_MemacDriverParam = NULL;
-+
-+ return E_OK;
-+}
-+
-+/* ......................................................................... */
-+
-+static t_Error MemacFree(t_Handle h_Memac)
-+{
-+ t_Memac *p_Memac = (t_Memac *)h_Memac;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Memac, E_INVALID_HANDLE);
-+
-+ if (p_Memac->p_MemacDriverParam)
-+ {
-+ /* Called after config */
-+ XX_Free(p_Memac->p_MemacDriverParam);
-+ p_Memac->p_MemacDriverParam = NULL;
-+ }
-+ else
-+ /* Called after init */
-+ FreeInitResources(p_Memac);
-+
-+ XX_Free(p_Memac);
-+
-+ return E_OK;
-+}
-+
-+/* ......................................................................... */
-+
-+static void InitFmMacControllerDriver(t_FmMacControllerDriver *p_FmMacControllerDriver)
-+{
-+ p_FmMacControllerDriver->f_FM_MAC_Init = MemacInit;
-+ p_FmMacControllerDriver->f_FM_MAC_Free = MemacFree;
-+
-+ p_FmMacControllerDriver->f_FM_MAC_SetStatistics = NULL;
-+ p_FmMacControllerDriver->f_FM_MAC_ConfigLoopback = MemacConfigLoopback;
-+ p_FmMacControllerDriver->f_FM_MAC_ConfigMaxFrameLength = MemacConfigMaxFrameLength;
-+
-+ p_FmMacControllerDriver->f_FM_MAC_ConfigWan = MemacConfigWan;
-+
-+ p_FmMacControllerDriver->f_FM_MAC_ConfigPadAndCrc = MemacConfigPad;
-+ p_FmMacControllerDriver->f_FM_MAC_ConfigHalfDuplex = NULL; /* half-duplex is detected automatically */
-+ p_FmMacControllerDriver->f_FM_MAC_ConfigLengthCheck = MemacConfigLengthCheck;
-+
-+ p_FmMacControllerDriver->f_FM_MAC_ConfigException = MemacConfigException;
-+ p_FmMacControllerDriver->f_FM_MAC_ConfigResetOnInit = MemacConfigResetOnInit;
-+
-+ p_FmMacControllerDriver->f_FM_MAC_SetException = MemacSetException;
-+
-+ p_FmMacControllerDriver->f_FM_MAC_Enable1588TimeStamp = MemacEnable1588TimeStamp; /* always enabled */
-+ p_FmMacControllerDriver->f_FM_MAC_Disable1588TimeStamp = NULL;
-+
-+ p_FmMacControllerDriver->f_FM_MAC_SetPromiscuous = MemacSetPromiscuous;
-+ p_FmMacControllerDriver->f_FM_MAC_AdjustLink = MemacAdjustLink;
-+ p_FmMacControllerDriver->f_FM_MAC_RestartAutoneg = NULL;
-+
-+ p_FmMacControllerDriver->f_FM_MAC_Enable = MemacEnable;
-+ p_FmMacControllerDriver->f_FM_MAC_Disable = MemacDisable;
-+ p_FmMacControllerDriver->f_FM_MAC_Resume = MemacInitInternalPhy;
-+
-+ p_FmMacControllerDriver->f_FM_MAC_SetTxAutoPauseFrames = MemacSetTxAutoPauseFrames;
-+ p_FmMacControllerDriver->f_FM_MAC_SetTxPauseFrames = MemacSetTxPauseFrames;
-+ p_FmMacControllerDriver->f_FM_MAC_SetRxIgnorePauseFrames = MemacSetRxIgnorePauseFrames;
-+
-+ p_FmMacControllerDriver->f_FM_MAC_SetWakeOnLan = MemacSetWakeOnLan;
-+
-+ p_FmMacControllerDriver->f_FM_MAC_ResetCounters = MemacResetCounters;
-+ p_FmMacControllerDriver->f_FM_MAC_GetStatistics = MemacGetStatistics;
-+ p_FmMacControllerDriver->f_FM_MAC_GetFrameSizeCounters = MemacGetFrameSizeCounters;
-+
-+ p_FmMacControllerDriver->f_FM_MAC_ModifyMacAddr = MemacModifyMacAddress;
-+ p_FmMacControllerDriver->f_FM_MAC_AddHashMacAddr = MemacAddHashMacAddress;
-+ p_FmMacControllerDriver->f_FM_MAC_RemoveHashMacAddr = MemacDelHashMacAddress;
-+ p_FmMacControllerDriver->f_FM_MAC_AddExactMatchMacAddr = MemacAddExactMatchMacAddress;
-+ p_FmMacControllerDriver->f_FM_MAC_RemovelExactMatchMacAddr = MemacDelExactMatchMacAddress;
-+ p_FmMacControllerDriver->f_FM_MAC_GetId = MemacGetId;
-+ p_FmMacControllerDriver->f_FM_MAC_GetVersion = NULL;
-+ p_FmMacControllerDriver->f_FM_MAC_GetMaxFrameLength = MemacGetMaxFrameLength;
-+
-+ p_FmMacControllerDriver->f_FM_MAC_MII_WritePhyReg = MEMAC_MII_WritePhyReg;
-+ p_FmMacControllerDriver->f_FM_MAC_MII_ReadPhyReg = MEMAC_MII_ReadPhyReg;
-+}
-+
-+
-+/*****************************************************************************/
-+/* mEMAC Config Main Entry */
-+/*****************************************************************************/
-+
-+/* ......................................................................... */
-+
-+t_Handle MEMAC_Config(t_FmMacParams *p_FmMacParam)
-+{
-+ t_Memac *p_Memac;
-+ struct memac_cfg *p_MemacDriverParam;
-+ uintptr_t baseAddr;
-+
-+ SANITY_CHECK_RETURN_VALUE(p_FmMacParam, E_NULL_POINTER, NULL);
-+
-+ baseAddr = p_FmMacParam->baseAddr;
-+ /* Allocate memory for the mEMAC data structure */
-+ p_Memac = (t_Memac *)XX_Malloc(sizeof(t_Memac));
-+ if (!p_Memac)
-+ {
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("mEMAC driver structure"));
-+ return NULL;
-+ }
-+ memset(p_Memac, 0, sizeof(t_Memac));
-+ InitFmMacControllerDriver(&p_Memac->fmMacControllerDriver);
-+
-+ /* Allocate memory for the mEMAC driver parameters data structure */
-+ p_MemacDriverParam = (struct memac_cfg *)XX_Malloc(sizeof(struct memac_cfg));
-+ if (!p_MemacDriverParam)
-+ {
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("mEMAC driver parameters"));
-+ XX_Free(p_Memac);
-+ return NULL;
-+ }
-+ memset(p_MemacDriverParam, 0, sizeof(struct memac_cfg));
-+
-+ /* Plant parameter structure pointer */
-+ p_Memac->p_MemacDriverParam = p_MemacDriverParam;
-+
-+ fman_memac_defconfig(p_MemacDriverParam);
-+
-+ p_Memac->addr = ENET_ADDR_TO_UINT64(p_FmMacParam->addr);
-+
-+ p_Memac->p_MemMap = (struct memac_regs *)UINT_TO_PTR(baseAddr);
-+ p_Memac->p_MiiMemMap = (struct memac_mii_access_mem_map*)UINT_TO_PTR(baseAddr + MEMAC_TO_MII_OFFSET);
-+
-+ p_Memac->enetMode = p_FmMacParam->enetMode;
-+ p_Memac->macId = p_FmMacParam->macId;
-+ p_Memac->exceptions = MEMAC_default_exceptions;
-+ p_Memac->f_Exception = p_FmMacParam->f_Exception;
-+ p_Memac->f_Event = p_FmMacParam->f_Event;
-+ p_Memac->h_App = p_FmMacParam->h_App;
-+
-+ return p_Memac;
-+}
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MAC/memac.h
-@@ -0,0 +1,110 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/******************************************************************************
-+ @File memac.h
-+
-+ @Description FM Multirate Ethernet MAC (mEMAC)
-+*//***************************************************************************/
-+#ifndef __MEMAC_H
-+#define __MEMAC_H
-+
-+#include "std_ext.h"
-+#include "error_ext.h"
-+#include "list_ext.h"
-+
-+#include "fsl_fman_memac_mii_acc.h"
-+#include "fm_mac.h"
-+#include "fsl_fman_memac.h"
-+
-+
-+#define MEMAC_default_exceptions \
-+ ((uint32_t)(MEMAC_IMASK_TSECC_ER | MEMAC_IMASK_TECC_ER | MEMAC_IMASK_RECC_ER | MEMAC_IMASK_MGI))
-+
-+#define GET_EXCEPTION_FLAG(bitMask, exception) switch (exception){ \
-+ case e_FM_MAC_EX_10G_1TX_ECC_ER: \
-+ bitMask = MEMAC_IMASK_TECC_ER; break; \
-+ case e_FM_MAC_EX_10G_RX_ECC_ER: \
-+ bitMask = MEMAC_IMASK_RECC_ER; break; \
-+ case e_FM_MAC_EX_TS_FIFO_ECC_ERR: \
-+ bitMask = MEMAC_IMASK_TSECC_ER; break; \
-+ case e_FM_MAC_EX_MAGIC_PACKET_INDICATION: \
-+ bitMask = MEMAC_IMASK_MGI; break; \
-+ default: bitMask = 0;break;}
-+
-+
-+typedef struct
-+{
-+ t_FmMacControllerDriver fmMacControllerDriver; /**< Upper Mac control block */
-+ t_Handle h_App; /**< Handle to the upper layer application */
-+ struct memac_regs *p_MemMap; /**< Pointer to MAC memory mapped registers */
-+ struct memac_mii_access_mem_map *p_MiiMemMap; /**< Pointer to MII memory mapped registers */
-+ uint64_t addr; /**< MAC address of device */
-+ e_EnetMode enetMode; /**< Ethernet physical interface */
-+ t_FmMacExceptionCallback *f_Exception;
-+ int mdioIrq;
-+ t_FmMacExceptionCallback *f_Event;
-+ bool indAddrRegUsed[MEMAC_NUM_OF_PADDRS]; /**< Whether a particular individual address recognition register is being used */
-+ uint64_t paddr[MEMAC_NUM_OF_PADDRS]; /**< MAC address for particular individual address recognition register */
-+ uint8_t numOfIndAddrInRegs; /**< Number of individual addresses in registers for this station. */
-+ t_EthHash *p_MulticastAddrHash; /**< Pointer to driver's global address hash table */
-+ t_EthHash *p_UnicastAddrHash; /**< Pointer to driver's individual address hash table */
-+ bool debugMode;
-+ uint8_t macId;
-+ uint32_t exceptions;
-+ struct memac_cfg *p_MemacDriverParam;
-+} t_Memac;
-+
-+
-+/* Internal PHY access */
-+#define PHY_MDIO_ADDR 0
-+
-+/* Internal PHY Registers - SGMII */
-+#define PHY_SGMII_CR_PHY_RESET 0x8000
-+#define PHY_SGMII_CR_RESET_AN 0x0200
-+#define PHY_SGMII_CR_DEF_VAL 0x1140
-+#define PHY_SGMII_DEV_ABILITY_SGMII 0x4001
-+#define PHY_SGMII_DEV_ABILITY_1000X 0x01A0
-+#define PHY_SGMII_IF_SPEED_GIGABIT 0x0008
-+#define PHY_SGMII_IF_MODE_AN 0x0002
-+#define PHY_SGMII_IF_MODE_SGMII 0x0001
-+#define PHY_SGMII_IF_MODE_1000X 0x0000
-+
-+
-+#define MEMAC_TO_MII_OFFSET 0x030 /* Offset from the MEM map to the MDIO mem map */
-+
-+t_Error MEMAC_MII_WritePhyReg(t_Handle h_Memac, uint8_t phyAddr, uint8_t reg, uint16_t data);
-+t_Error MEMAC_MII_ReadPhyReg(t_Handle h_Memac, uint8_t phyAddr, uint8_t reg, uint16_t *p_Data);
-+
-+
-+#endif /* __MEMAC_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MAC/memac_mii_acc.c
-@@ -0,0 +1,78 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+#include "error_ext.h"
-+#include "std_ext.h"
-+#include "fm_mac.h"
-+#include "memac.h"
-+#include "xx_ext.h"
-+
-+#include "fm_common.h"
-+#include "memac_mii_acc.h"
-+
-+
-+/*****************************************************************************/
-+t_Error MEMAC_MII_WritePhyReg(t_Handle h_Memac,
-+ uint8_t phyAddr,
-+ uint8_t reg,
-+ uint16_t data)
-+{
-+ t_Memac *p_Memac = (t_Memac *)h_Memac;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Memac, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Memac->p_MiiMemMap, E_INVALID_HANDLE);
-+
-+ return (t_Error)fman_memac_mii_write_phy_reg(p_Memac->p_MiiMemMap,
-+ phyAddr,
-+ reg,
-+ data,
-+ (enum enet_speed)ENET_SPEED_FROM_MODE(p_Memac->enetMode));
-+}
-+
-+/*****************************************************************************/
-+t_Error MEMAC_MII_ReadPhyReg(t_Handle h_Memac,
-+ uint8_t phyAddr,
-+ uint8_t reg,
-+ uint16_t *p_Data)
-+{
-+ t_Memac *p_Memac = (t_Memac *)h_Memac;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Memac, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Memac->p_MiiMemMap, E_INVALID_HANDLE);
-+
-+ return fman_memac_mii_read_phy_reg(p_Memac->p_MiiMemMap,
-+ phyAddr,
-+ reg,
-+ p_Data,
-+ (enum enet_speed)ENET_SPEED_FROM_MODE(p_Memac->enetMode));
-+}
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MAC/memac_mii_acc.h
-@@ -0,0 +1,73 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+#ifndef __MEMAC_MII_ACC_H
-+#define __MEMAC_MII_ACC_H
-+
-+#include "std_ext.h"
-+
-+
-+/* MII Management Registers */
-+#define MDIO_CFG_CLK_DIV_MASK 0x0080ff80
-+#define MDIO_CFG_CLK_DIV_SHIFT 7
-+#define MDIO_CFG_HOLD_MASK 0x0000001c
-+#define MDIO_CFG_ENC45 0x00000040
-+#define MDIO_CFG_READ_ERR 0x00000002
-+#define MDIO_CFG_BSY 0x00000001
-+
-+#define MDIO_CTL_PHY_ADDR_SHIFT 5
-+#define MDIO_CTL_READ 0x00008000
-+
-+#define MDIO_DATA_BSY 0x80000000
-+
-+#if defined(__MWERKS__) && !defined(__GNUC__)
-+#pragma pack(push,1)
-+#endif /* defined(__MWERKS__) && ... */
-+
-+/*----------------------------------------------------*/
-+/* MII Configuration Control Memory Map Registers */
-+/*----------------------------------------------------*/
-+typedef struct t_MemacMiiAccessMemMap
-+{
-+ volatile uint32_t mdio_cfg; /* 0x030 */
-+ volatile uint32_t mdio_ctrl; /* 0x034 */
-+ volatile uint32_t mdio_data; /* 0x038 */
-+ volatile uint32_t mdio_addr; /* 0x03c */
-+} t_MemacMiiAccessMemMap ;
-+
-+#if defined(__MWERKS__) && !defined(__GNUC__)
-+#pragma pack(pop)
-+#endif /* defined(__MWERKS__) && ... */
-+
-+
-+#endif /* __MEMAC_MII_ACC_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MAC/tgec.c
-@@ -0,0 +1,1017 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/******************************************************************************
-+ @File tgec.c
-+
-+ @Description FM 10G MAC ...
-+*//***************************************************************************/
-+
-+#include "std_ext.h"
-+#include "string_ext.h"
-+#include "error_ext.h"
-+#include "xx_ext.h"
-+#include "endian_ext.h"
-+#include "debug_ext.h"
-+#include "crc_mac_addr_ext.h"
-+
-+#include "fm_common.h"
-+#include "fsl_fman_tgec.h"
-+#include "tgec.h"
-+
-+
-+/*****************************************************************************/
-+/* Internal routines */
-+/*****************************************************************************/
-+
-+static t_Error CheckInitParameters(t_Tgec *p_Tgec)
-+{
-+ if (ENET_SPEED_FROM_MODE(p_Tgec->enetMode) < e_ENET_SPEED_10000)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("Ethernet 10G MAC driver only support 10G speed"));
-+#if (FM_MAX_NUM_OF_10G_MACS > 0)
-+ if (p_Tgec->macId >= FM_MAX_NUM_OF_10G_MACS)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("macId of 10G can not be greater than 0"));
-+#endif /* (FM_MAX_NUM_OF_10G_MACS > 0) */
-+
-+ if (p_Tgec->addr == 0)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("Ethernet 10G MAC Must have a valid MAC Address"));
-+ if (!p_Tgec->f_Exception)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("uninitialized f_Exception"));
-+ if (!p_Tgec->f_Event)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("uninitialized f_Event"));
-+#ifdef FM_LEN_CHECK_ERRATA_FMAN_SW002
-+ if (!p_Tgec->p_TgecDriverParam->no_length_check_enable)
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, ("LengthCheck!"));
-+#endif /* FM_LEN_CHECK_ERRATA_FMAN_SW002 */
-+ return E_OK;
-+}
-+
-+/* ......................................................................... */
-+
-+static uint32_t GetMacAddrHashCode(uint64_t ethAddr)
-+{
-+ uint32_t crc;
-+
-+ /* CRC calculation */
-+ GET_MAC_ADDR_CRC(ethAddr, crc);
-+
-+ crc = GetMirror32(crc);
-+
-+ return crc;
-+}
-+
-+/* ......................................................................... */
-+
-+static void TgecErrException(t_Handle h_Tgec)
-+{
-+ t_Tgec *p_Tgec = (t_Tgec *)h_Tgec;
-+ uint32_t event;
-+ struct tgec_regs *p_TgecMemMap = p_Tgec->p_MemMap;
-+
-+ /* do not handle MDIO events */
-+ event = fman_tgec_get_event(p_TgecMemMap, ~(TGEC_IMASK_MDIO_SCAN_EVENT | TGEC_IMASK_MDIO_CMD_CMPL));
-+ event &= fman_tgec_get_interrupt_mask(p_TgecMemMap);
-+
-+ fman_tgec_ack_event(p_TgecMemMap, event);
-+
-+ if (event & TGEC_IMASK_REM_FAULT)
-+ p_Tgec->f_Exception(p_Tgec->h_App, e_FM_MAC_EX_10G_REM_FAULT);
-+ if (event & TGEC_IMASK_LOC_FAULT)
-+ p_Tgec->f_Exception(p_Tgec->h_App, e_FM_MAC_EX_10G_LOC_FAULT);
-+ if (event & TGEC_IMASK_TX_ECC_ER)
-+ p_Tgec->f_Exception(p_Tgec->h_App, e_FM_MAC_EX_10G_1TX_ECC_ER);
-+ if (event & TGEC_IMASK_TX_FIFO_UNFL)
-+ p_Tgec->f_Exception(p_Tgec->h_App, e_FM_MAC_EX_10G_TX_FIFO_UNFL);
-+ if (event & TGEC_IMASK_TX_FIFO_OVFL)
-+ p_Tgec->f_Exception(p_Tgec->h_App, e_FM_MAC_EX_10G_TX_FIFO_OVFL);
-+ if (event & TGEC_IMASK_TX_ER)
-+ p_Tgec->f_Exception(p_Tgec->h_App, e_FM_MAC_EX_10G_TX_ER);
-+ if (event & TGEC_IMASK_RX_FIFO_OVFL)
-+ p_Tgec->f_Exception(p_Tgec->h_App, e_FM_MAC_EX_10G_RX_FIFO_OVFL);
-+ if (event & TGEC_IMASK_RX_ECC_ER)
-+ p_Tgec->f_Exception(p_Tgec->h_App, e_FM_MAC_EX_10G_RX_ECC_ER);
-+ if (event & TGEC_IMASK_RX_JAB_FRM)
-+ p_Tgec->f_Exception(p_Tgec->h_App, e_FM_MAC_EX_10G_RX_JAB_FRM);
-+ if (event & TGEC_IMASK_RX_OVRSZ_FRM)
-+ p_Tgec->f_Exception(p_Tgec->h_App, e_FM_MAC_EX_10G_RX_OVRSZ_FRM);
-+ if (event & TGEC_IMASK_RX_RUNT_FRM)
-+ p_Tgec->f_Exception(p_Tgec->h_App, e_FM_MAC_EX_10G_RX_RUNT_FRM);
-+ if (event & TGEC_IMASK_RX_FRAG_FRM)
-+ p_Tgec->f_Exception(p_Tgec->h_App, e_FM_MAC_EX_10G_RX_FRAG_FRM);
-+ if (event & TGEC_IMASK_RX_LEN_ER)
-+ p_Tgec->f_Exception(p_Tgec->h_App, e_FM_MAC_EX_10G_RX_LEN_ER);
-+ if (event & TGEC_IMASK_RX_CRC_ER)
-+ p_Tgec->f_Exception(p_Tgec->h_App, e_FM_MAC_EX_10G_RX_CRC_ER);
-+ if (event & TGEC_IMASK_RX_ALIGN_ER)
-+ p_Tgec->f_Exception(p_Tgec->h_App, e_FM_MAC_EX_10G_RX_ALIGN_ER);
-+}
-+
-+/* ......................................................................... */
-+
-+static void TgecException(t_Handle h_Tgec)
-+{
-+ t_Tgec *p_Tgec = (t_Tgec *)h_Tgec;
-+ uint32_t event;
-+ struct tgec_regs *p_TgecMemMap = p_Tgec->p_MemMap;
-+
-+ /* handle only MDIO events */
-+ event = fman_tgec_get_event(p_TgecMemMap, (TGEC_IMASK_MDIO_SCAN_EVENT | TGEC_IMASK_MDIO_CMD_CMPL));
-+ event &= fman_tgec_get_interrupt_mask(p_TgecMemMap);
-+
-+ fman_tgec_ack_event(p_TgecMemMap, event);
-+
-+ if (event & TGEC_IMASK_MDIO_SCAN_EVENT)
-+ p_Tgec->f_Event(p_Tgec->h_App, e_FM_MAC_EX_10G_MDIO_SCAN_EVENTMDIO);
-+ if (event & TGEC_IMASK_MDIO_CMD_CMPL)
-+ p_Tgec->f_Event(p_Tgec->h_App, e_FM_MAC_EX_10G_MDIO_CMD_CMPL);
-+}
-+
-+/* ......................................................................... */
-+
-+static void FreeInitResources(t_Tgec *p_Tgec)
-+{
-+ if (p_Tgec->mdioIrq != NO_IRQ)
-+ {
-+ XX_DisableIntr(p_Tgec->mdioIrq);
-+ XX_FreeIntr(p_Tgec->mdioIrq);
-+ }
-+
-+ FmUnregisterIntr(p_Tgec->fmMacControllerDriver.h_Fm, e_FM_MOD_10G_MAC, p_Tgec->macId, e_FM_INTR_TYPE_ERR);
-+
-+ /* release the driver's group hash table */
-+ FreeHashTable(p_Tgec->p_MulticastAddrHash);
-+ p_Tgec->p_MulticastAddrHash = NULL;
-+
-+ /* release the driver's individual hash table */
-+ FreeHashTable(p_Tgec->p_UnicastAddrHash);
-+ p_Tgec->p_UnicastAddrHash = NULL;
-+}
-+
-+
-+/*****************************************************************************/
-+/* 10G MAC API routines */
-+/*****************************************************************************/
-+
-+/* ......................................................................... */
-+
-+static t_Error TgecEnable(t_Handle h_Tgec, e_CommMode mode)
-+{
-+ t_Tgec *p_Tgec = (t_Tgec *)h_Tgec;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Tgec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Tgec->p_TgecDriverParam, E_INVALID_STATE);
-+
-+ fman_tgec_enable(p_Tgec->p_MemMap, (mode & e_COMM_MODE_RX), (mode & e_COMM_MODE_TX));
-+
-+ return E_OK;
-+}
-+
-+/* ......................................................................... */
-+
-+static t_Error TgecDisable (t_Handle h_Tgec, e_CommMode mode)
-+{
-+ t_Tgec *p_Tgec = (t_Tgec *)h_Tgec;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Tgec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Tgec->p_TgecDriverParam, E_INVALID_STATE);
-+
-+ fman_tgec_disable(p_Tgec->p_MemMap, (mode & e_COMM_MODE_RX), (mode & e_COMM_MODE_TX));
-+
-+ return E_OK;
-+}
-+
-+/* ......................................................................... */
-+
-+static t_Error TgecSetPromiscuous(t_Handle h_Tgec, bool newVal)
-+{
-+ t_Tgec *p_Tgec = (t_Tgec *)h_Tgec;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Tgec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Tgec->p_TgecDriverParam, E_INVALID_STATE);
-+
-+ fman_tgec_set_promiscuous(p_Tgec->p_MemMap, newVal);
-+
-+ return E_OK;
-+}
-+
-+
-+/*****************************************************************************/
-+/* Tgec Configs modification functions */
-+/*****************************************************************************/
-+
-+/* ......................................................................... */
-+
-+static t_Error TgecConfigLoopback(t_Handle h_Tgec, bool newVal)
-+{
-+ t_Tgec *p_Tgec = (t_Tgec *)h_Tgec;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Tgec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Tgec->p_TgecDriverParam, E_INVALID_STATE);
-+
-+ p_Tgec->p_TgecDriverParam->loopback_enable = newVal;
-+
-+ return E_OK;
-+}
-+
-+/* ......................................................................... */
-+
-+static t_Error TgecConfigWan(t_Handle h_Tgec, bool newVal)
-+{
-+ t_Tgec *p_Tgec = (t_Tgec *)h_Tgec;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Tgec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Tgec->p_TgecDriverParam, E_INVALID_STATE);
-+
-+ p_Tgec->p_TgecDriverParam->wan_mode_enable = newVal;
-+
-+ return E_OK;
-+}
-+
-+/* ......................................................................... */
-+
-+static t_Error TgecConfigMaxFrameLength(t_Handle h_Tgec, uint16_t newVal)
-+{
-+ t_Tgec *p_Tgec = (t_Tgec *)h_Tgec;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Tgec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Tgec->p_TgecDriverParam, E_INVALID_STATE);
-+
-+ p_Tgec->p_TgecDriverParam->max_frame_length = newVal;
-+
-+ return E_OK;
-+}
-+
-+/* ......................................................................... */
-+
-+static t_Error TgecConfigLengthCheck(t_Handle h_Tgec, bool newVal)
-+{
-+ t_Tgec *p_Tgec = (t_Tgec *)h_Tgec;
-+
-+ UNUSED(newVal);
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Tgec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Tgec->p_TgecDriverParam, E_INVALID_STATE);
-+
-+ p_Tgec->p_TgecDriverParam->no_length_check_enable = !newVal;
-+
-+ return E_OK;
-+}
-+
-+/* ......................................................................... */
-+
-+static t_Error TgecConfigException(t_Handle h_Tgec, e_FmMacExceptions exception, bool enable)
-+{
-+ t_Tgec *p_Tgec = (t_Tgec *)h_Tgec;
-+ uint32_t bitMask = 0;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Tgec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Tgec->p_TgecDriverParam, E_INVALID_STATE);
-+
-+ GET_EXCEPTION_FLAG(bitMask, exception);
-+ if (bitMask)
-+ {
-+ if (enable)
-+ p_Tgec->exceptions |= bitMask;
-+ else
-+ p_Tgec->exceptions &= ~bitMask;
-+ }
-+ else
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("Undefined exception"));
-+
-+ return E_OK;
-+}
-+
-+#ifdef FM_TX_ECC_FRMS_ERRATA_10GMAC_A004
-+/* ......................................................................... */
-+
-+static t_Error TgecConfigSkipFman11Workaround(t_Handle h_Tgec)
-+{
-+ t_Tgec *p_Tgec = (t_Tgec *)h_Tgec;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Tgec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Tgec->p_TgecDriverParam, E_INVALID_STATE);
-+
-+ p_Tgec->p_TgecDriverParam->skip_fman11_workaround = TRUE;
-+
-+ return E_OK;
-+}
-+#endif /* FM_TX_ECC_FRMS_ERRATA_10GMAC_A004 */
-+
-+
-+/*****************************************************************************/
-+/* Tgec Run Time API functions */
-+/*****************************************************************************/
-+
-+/* ......................................................................... */
-+/* backward compatibility. will be removed in the future. */
-+static t_Error TgecTxMacPause(t_Handle h_Tgec, uint16_t pauseTime)
-+{
-+ t_Tgec *p_Tgec = (t_Tgec *)h_Tgec;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Tgec, E_INVALID_STATE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Tgec->p_TgecDriverParam, E_INVALID_STATE);
-+ fman_tgec_set_tx_pause_frames(p_Tgec->p_MemMap, pauseTime);
-+
-+
-+ return E_OK;
-+}
-+
-+/* ......................................................................... */
-+
-+static t_Error TgecSetTxPauseFrames(t_Handle h_Tgec,
-+ uint8_t priority,
-+ uint16_t pauseTime,
-+ uint16_t threshTime)
-+{
-+ t_Tgec *p_Tgec = (t_Tgec *)h_Tgec;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Tgec, E_INVALID_STATE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Tgec->p_TgecDriverParam, E_INVALID_STATE);
-+
-+ UNUSED(priority); UNUSED(threshTime);
-+
-+ fman_tgec_set_tx_pause_frames(p_Tgec->p_MemMap, pauseTime);
-+
-+ return E_OK;
-+}
-+
-+/* ......................................................................... */
-+
-+static t_Error TgecRxIgnoreMacPause(t_Handle h_Tgec, bool en)
-+{
-+ t_Tgec *p_Tgec = (t_Tgec *)h_Tgec;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Tgec, E_INVALID_STATE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Tgec->p_TgecDriverParam, E_INVALID_STATE);
-+
-+ fman_tgec_set_rx_ignore_pause_frames(p_Tgec->p_MemMap, en);
-+
-+ return E_OK;
-+}
-+
-+/* ......................................................................... */
-+
-+static t_Error TgecGetStatistics(t_Handle h_Tgec, t_FmMacStatistics *p_Statistics)
-+{
-+ t_Tgec *p_Tgec = (t_Tgec *)h_Tgec;
-+ struct tgec_regs *p_TgecMemMap;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Tgec, E_NULL_POINTER);
-+ SANITY_CHECK_RETURN_ERROR(!p_Tgec->p_TgecDriverParam, E_INVALID_STATE);
-+ SANITY_CHECK_RETURN_ERROR(p_Statistics, E_NULL_POINTER);
-+
-+ p_TgecMemMap = p_Tgec->p_MemMap;
-+
-+ p_Statistics->eStatPkts64 = fman_tgec_get_counter(p_TgecMemMap, E_TGEC_COUNTER_R64);
-+ p_Statistics->eStatPkts65to127 = fman_tgec_get_counter(p_TgecMemMap, E_TGEC_COUNTER_R127);
-+ p_Statistics->eStatPkts128to255 = fman_tgec_get_counter(p_TgecMemMap, E_TGEC_COUNTER_R255);
-+ p_Statistics->eStatPkts256to511 = fman_tgec_get_counter(p_TgecMemMap, E_TGEC_COUNTER_R511);
-+ p_Statistics->eStatPkts512to1023 = fman_tgec_get_counter(p_TgecMemMap, E_TGEC_COUNTER_R1023);
-+ p_Statistics->eStatPkts1024to1518 = fman_tgec_get_counter(p_TgecMemMap, E_TGEC_COUNTER_R1518);
-+ p_Statistics->eStatPkts1519to1522 = fman_tgec_get_counter(p_TgecMemMap, E_TGEC_COUNTER_R1519X);
-+/* */
-+ p_Statistics->eStatFragments = fman_tgec_get_counter(p_TgecMemMap, E_TGEC_COUNTER_TRFRG);
-+ p_Statistics->eStatJabbers = fman_tgec_get_counter(p_TgecMemMap, E_TGEC_COUNTER_TRJBR);
-+
-+ p_Statistics->eStatsDropEvents = fman_tgec_get_counter(p_TgecMemMap, E_TGEC_COUNTER_RDRP);
-+ p_Statistics->eStatCRCAlignErrors = fman_tgec_get_counter(p_TgecMemMap, E_TGEC_COUNTER_RALN);
-+
-+ p_Statistics->eStatUndersizePkts = fman_tgec_get_counter(p_TgecMemMap, E_TGEC_COUNTER_TRUND);
-+ p_Statistics->eStatOversizePkts = fman_tgec_get_counter(p_TgecMemMap, E_TGEC_COUNTER_TROVR);
-+/* Pause */
-+ p_Statistics->reStatPause = fman_tgec_get_counter(p_TgecMemMap, E_TGEC_COUNTER_RXPF);
-+ p_Statistics->teStatPause = fman_tgec_get_counter(p_TgecMemMap, E_TGEC_COUNTER_TXPF);
-+
-+/* MIB II */
-+ p_Statistics->ifInOctets = fman_tgec_get_counter(p_TgecMemMap, E_TGEC_COUNTER_ROCT);
-+ p_Statistics->ifInUcastPkts = fman_tgec_get_counter(p_TgecMemMap, E_TGEC_COUNTER_RUCA);
-+ p_Statistics->ifInMcastPkts = fman_tgec_get_counter(p_TgecMemMap, E_TGEC_COUNTER_RMCA);
-+ p_Statistics->ifInBcastPkts = fman_tgec_get_counter(p_TgecMemMap, E_TGEC_COUNTER_RBCA);
-+ p_Statistics->ifInPkts = p_Statistics->ifInUcastPkts
-+ + p_Statistics->ifInMcastPkts
-+ + p_Statistics->ifInBcastPkts;
-+ p_Statistics->ifInDiscards = 0;
-+ p_Statistics->ifInErrors = fman_tgec_get_counter(p_TgecMemMap, E_TGEC_COUNTER_RERR);
-+
-+ p_Statistics->ifOutOctets = fman_tgec_get_counter(p_TgecMemMap, E_TGEC_COUNTER_TOCT);
-+ p_Statistics->ifOutUcastPkts = fman_tgec_get_counter(p_TgecMemMap, E_TGEC_COUNTER_TUCA);
-+ p_Statistics->ifOutMcastPkts = fman_tgec_get_counter(p_TgecMemMap, E_TGEC_COUNTER_TMCA);
-+ p_Statistics->ifOutBcastPkts = fman_tgec_get_counter(p_TgecMemMap, E_TGEC_COUNTER_TBCA);
-+ p_Statistics->ifOutPkts = p_Statistics->ifOutUcastPkts
-+ + p_Statistics->ifOutMcastPkts
-+ + p_Statistics->ifOutBcastPkts;
-+ p_Statistics->ifOutDiscards = 0;
-+ p_Statistics->ifOutErrors = fman_tgec_get_counter(p_TgecMemMap, E_TGEC_COUNTER_TERR);
-+
-+ return E_OK;
-+}
-+
-+/* ......................................................................... */
-+
-+static t_Error TgecGetFrameSizeCounters(t_Handle h_Tgec, t_FmMacFrameSizeCounters *p_FrameSizeCounters, e_CommMode type)
-+{
-+ t_Tgec *p_Tgec = (t_Tgec *)h_Tgec;
-+ struct tgec_regs *p_TgecMemMap;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Tgec, E_NULL_POINTER);
-+ SANITY_CHECK_RETURN_ERROR(!p_Tgec->p_TgecDriverParam, E_INVALID_STATE);
-+ SANITY_CHECK_RETURN_ERROR(p_FrameSizeCounters, E_NULL_POINTER);
-+
-+ p_TgecMemMap = p_Tgec->p_MemMap;
-+
-+ switch (type)
-+ {
-+ case e_COMM_MODE_NONE:
-+ break;
-+
-+ case e_COMM_MODE_RX:
-+ p_FrameSizeCounters->count_pkts_64 = fman_tgec_get_counter(p_TgecMemMap, E_TGEC_COUNTER_R64);
-+ p_FrameSizeCounters->count_pkts_65_to_127 = fman_tgec_get_counter(p_TgecMemMap, E_TGEC_COUNTER_R127);
-+ p_FrameSizeCounters->count_pkts_128_to_255 = fman_tgec_get_counter(p_TgecMemMap, E_TGEC_COUNTER_R255);
-+ p_FrameSizeCounters->count_pkts_256_to_511 = fman_tgec_get_counter(p_TgecMemMap, E_TGEC_COUNTER_R511);
-+ p_FrameSizeCounters->count_pkts_512_to_1023 = fman_tgec_get_counter(p_TgecMemMap, E_TGEC_COUNTER_R1023);
-+ p_FrameSizeCounters->count_pkts_1024_to_1518 = fman_tgec_get_counter(p_TgecMemMap, E_TGEC_COUNTER_R1518);
-+ p_FrameSizeCounters->count_pkts_1519_to_1522 = fman_tgec_get_counter(p_TgecMemMap, E_TGEC_COUNTER_R1519X);
-+ break;
-+
-+ case e_COMM_MODE_TX:
-+ //Tx counters not supported
-+ break;
-+
-+ case e_COMM_MODE_RX_AND_TX:
-+ //Tx counters not supported
-+ break;
-+ }
-+
-+ return E_OK;
-+}
-+
-+
-+/* ......................................................................... */
-+
-+static t_Error TgecEnable1588TimeStamp(t_Handle h_Tgec)
-+{
-+ t_Tgec *p_Tgec = (t_Tgec *)h_Tgec;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Tgec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Tgec->p_TgecDriverParam, E_INVALID_STATE);
-+
-+ fman_tgec_enable_1588_time_stamp(p_Tgec->p_MemMap, 1);
-+
-+ return E_OK;
-+}
-+
-+/* ......................................................................... */
-+
-+static t_Error TgecDisable1588TimeStamp(t_Handle h_Tgec)
-+{
-+ t_Tgec *p_Tgec = (t_Tgec *)h_Tgec;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Tgec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Tgec->p_TgecDriverParam, E_INVALID_STATE);
-+
-+ fman_tgec_enable_1588_time_stamp(p_Tgec->p_MemMap, 0);
-+
-+ return E_OK;
-+}
-+
-+/* ......................................................................... */
-+
-+static t_Error TgecModifyMacAddress (t_Handle h_Tgec, t_EnetAddr *p_EnetAddr)
-+{
-+ t_Tgec *p_Tgec = (t_Tgec *)h_Tgec;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Tgec, E_NULL_POINTER);
-+ SANITY_CHECK_RETURN_ERROR(!p_Tgec->p_TgecDriverParam, E_INVALID_STATE);
-+
-+ p_Tgec->addr = ENET_ADDR_TO_UINT64(*p_EnetAddr);
-+ fman_tgec_set_mac_address(p_Tgec->p_MemMap, (uint8_t *)(*p_EnetAddr));
-+
-+ return E_OK;
-+}
-+
-+/* ......................................................................... */
-+
-+static t_Error TgecResetCounters (t_Handle h_Tgec)
-+{
-+ t_Tgec *p_Tgec = (t_Tgec *)h_Tgec;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Tgec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Tgec->p_TgecDriverParam, E_INVALID_STATE);
-+
-+ fman_tgec_reset_stat(p_Tgec->p_MemMap);
-+
-+ return E_OK;
-+}
-+
-+/* ......................................................................... */
-+
-+static t_Error TgecAddExactMatchMacAddress(t_Handle h_Tgec, t_EnetAddr *p_EthAddr)
-+{
-+ t_Tgec *p_Tgec = (t_Tgec *) h_Tgec;
-+ uint64_t ethAddr;
-+ uint8_t paddrNum;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Tgec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Tgec->p_TgecDriverParam, E_INVALID_STATE);
-+
-+ ethAddr = ENET_ADDR_TO_UINT64(*p_EthAddr);
-+
-+ if (ethAddr & GROUP_ADDRESS)
-+ /* Multicast address has no effect in PADDR */
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("Multicast address"));
-+
-+ /* Make sure no PADDR contains this address */
-+ for (paddrNum = 0; paddrNum < TGEC_NUM_OF_PADDRS; paddrNum++)
-+ if (p_Tgec->indAddrRegUsed[paddrNum])
-+ if (p_Tgec->paddr[paddrNum] == ethAddr)
-+ RETURN_ERROR(MAJOR, E_ALREADY_EXISTS, NO_MSG);
-+
-+ /* Find first unused PADDR */
-+ for (paddrNum = 0; paddrNum < TGEC_NUM_OF_PADDRS; paddrNum++)
-+ {
-+ if (!(p_Tgec->indAddrRegUsed[paddrNum]))
-+ {
-+ /* mark this PADDR as used */
-+ p_Tgec->indAddrRegUsed[paddrNum] = TRUE;
-+ /* store address */
-+ p_Tgec->paddr[paddrNum] = ethAddr;
-+
-+ /* put in hardware */
-+ fman_tgec_add_addr_in_paddr(p_Tgec->p_MemMap, (uint8_t*)(*p_EthAddr)/* , paddrNum */);
-+ p_Tgec->numOfIndAddrInRegs++;
-+
-+ return E_OK;
-+ }
-+ }
-+
-+ /* No free PADDR */
-+ RETURN_ERROR(MAJOR, E_FULL, NO_MSG);
-+}
-+
-+/* ......................................................................... */
-+
-+static t_Error TgecDelExactMatchMacAddress(t_Handle h_Tgec, t_EnetAddr *p_EthAddr)
-+{
-+ t_Tgec *p_Tgec = (t_Tgec *) h_Tgec;
-+ uint64_t ethAddr;
-+ uint8_t paddrNum;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Tgec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Tgec->p_TgecDriverParam, E_INVALID_STATE);
-+
-+ ethAddr = ENET_ADDR_TO_UINT64(*p_EthAddr);
-+
-+ /* Find used PADDR containing this address */
-+ for (paddrNum = 0; paddrNum < TGEC_NUM_OF_PADDRS; paddrNum++)
-+ {
-+ if ((p_Tgec->indAddrRegUsed[paddrNum]) &&
-+ (p_Tgec->paddr[paddrNum] == ethAddr))
-+ {
-+ /* mark this PADDR as not used */
-+ p_Tgec->indAddrRegUsed[paddrNum] = FALSE;
-+ /* clear in hardware */
-+ fman_tgec_clear_addr_in_paddr(p_Tgec->p_MemMap /*, paddrNum */);
-+ p_Tgec->numOfIndAddrInRegs--;
-+
-+ return E_OK;
-+ }
-+ }
-+
-+ RETURN_ERROR(MAJOR, E_NOT_FOUND, NO_MSG);
-+}
-+
-+/* ......................................................................... */
-+
-+static t_Error TgecAddHashMacAddress(t_Handle h_Tgec, t_EnetAddr *p_EthAddr)
-+{
-+ t_Tgec *p_Tgec = (t_Tgec *)h_Tgec;
-+ t_EthHashEntry *p_HashEntry;
-+ uint32_t crc;
-+ uint32_t hash;
-+ uint64_t ethAddr;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Tgec, E_NULL_POINTER);
-+ SANITY_CHECK_RETURN_ERROR(!p_Tgec->p_TgecDriverParam, E_INVALID_STATE);
-+
-+ ethAddr = ENET_ADDR_TO_UINT64(*p_EthAddr);
-+
-+ if (!(ethAddr & GROUP_ADDRESS))
-+ /* Unicast addresses not supported in hash */
-+ RETURN_ERROR(MAJOR, E_NOT_SUPPORTED, ("Unicast Address"));
-+
-+ /* CRC calculation */
-+ crc = GetMacAddrHashCode(ethAddr);
-+
-+ hash = (crc >> TGEC_HASH_MCAST_SHIFT) & TGEC_HASH_ADR_MSK; /* Take 9 MSB bits */
-+
-+ /* Create element to be added to the driver hash table */
-+ p_HashEntry = (t_EthHashEntry *)XX_Malloc(sizeof(t_EthHashEntry));
-+ p_HashEntry->addr = ethAddr;
-+ INIT_LIST(&p_HashEntry->node);
-+
-+ LIST_AddToTail(&(p_HashEntry->node), &(p_Tgec->p_MulticastAddrHash->p_Lsts[hash]));
-+ fman_tgec_set_hash_table(p_Tgec->p_MemMap, (hash | TGEC_HASH_MCAST_EN));
-+
-+ return E_OK;
-+}
-+
-+/* ......................................................................... */
-+
-+static t_Error TgecDelHashMacAddress(t_Handle h_Tgec, t_EnetAddr *p_EthAddr)
-+{
-+ t_Tgec *p_Tgec = (t_Tgec *)h_Tgec;
-+ t_EthHashEntry *p_HashEntry = NULL;
-+ t_List *p_Pos;
-+ uint32_t crc;
-+ uint32_t hash;
-+ uint64_t ethAddr;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Tgec, E_NULL_POINTER);
-+ SANITY_CHECK_RETURN_ERROR(!p_Tgec->p_TgecDriverParam, E_INVALID_STATE);
-+
-+ ethAddr = ((*(uint64_t *)p_EthAddr) >> 16);
-+
-+ /* CRC calculation */
-+ crc = GetMacAddrHashCode(ethAddr);
-+
-+ hash = (crc >> TGEC_HASH_MCAST_SHIFT) & TGEC_HASH_ADR_MSK; /* Take 9 MSB bits */
-+
-+ LIST_FOR_EACH(p_Pos, &(p_Tgec->p_MulticastAddrHash->p_Lsts[hash]))
-+ {
-+ p_HashEntry = ETH_HASH_ENTRY_OBJ(p_Pos);
-+ if (p_HashEntry->addr == ethAddr)
-+ {
-+ LIST_DelAndInit(&p_HashEntry->node);
-+ XX_Free(p_HashEntry);
-+ break;
-+ }
-+ }
-+ if (LIST_IsEmpty(&p_Tgec->p_MulticastAddrHash->p_Lsts[hash]))
-+ fman_tgec_set_hash_table(p_Tgec->p_MemMap, (hash & ~TGEC_HASH_MCAST_EN));
-+
-+ return E_OK;
-+}
-+
-+/* ......................................................................... */
-+
-+static t_Error TgecGetId(t_Handle h_Tgec, uint32_t *macId)
-+{
-+ t_Tgec *p_Tgec = (t_Tgec *)h_Tgec;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Tgec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Tgec->p_TgecDriverParam, E_INVALID_STATE);
-+
-+ UNUSED(p_Tgec);
-+ UNUSED(macId);
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, ("TgecGetId Not Supported"));
-+}
-+
-+/* ......................................................................... */
-+
-+static t_Error TgecGetVersion(t_Handle h_Tgec, uint32_t *macVersion)
-+{
-+ t_Tgec *p_Tgec = (t_Tgec *)h_Tgec;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Tgec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Tgec->p_TgecDriverParam, E_INVALID_STATE);
-+
-+ *macVersion = fman_tgec_get_revision(p_Tgec->p_MemMap);
-+
-+ return E_OK;
-+}
-+
-+/* ......................................................................... */
-+
-+static t_Error TgecSetExcpetion(t_Handle h_Tgec, e_FmMacExceptions exception, bool enable)
-+{
-+ t_Tgec *p_Tgec = (t_Tgec *)h_Tgec;
-+ uint32_t bitMask = 0;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Tgec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Tgec->p_TgecDriverParam, E_INVALID_STATE);
-+
-+ GET_EXCEPTION_FLAG(bitMask, exception);
-+ if (bitMask)
-+ {
-+ if (enable)
-+ p_Tgec->exceptions |= bitMask;
-+ else
-+ p_Tgec->exceptions &= ~bitMask;
-+ }
-+ else
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("Undefined exception"));
-+
-+ if (enable)
-+ fman_tgec_enable_interrupt(p_Tgec->p_MemMap, bitMask);
-+ else
-+ fman_tgec_disable_interrupt(p_Tgec->p_MemMap, bitMask);
-+
-+ return E_OK;
-+}
-+
-+/* ......................................................................... */
-+
-+static uint16_t TgecGetMaxFrameLength(t_Handle h_Tgec)
-+{
-+ t_Tgec *p_Tgec = (t_Tgec *)h_Tgec;
-+
-+ SANITY_CHECK_RETURN_VALUE(p_Tgec, E_INVALID_HANDLE, 0);
-+ SANITY_CHECK_RETURN_VALUE(!p_Tgec->p_TgecDriverParam, E_INVALID_STATE, 0);
-+
-+ return fman_tgec_get_max_frame_len(p_Tgec->p_MemMap);
-+}
-+
-+/* ......................................................................... */
-+
-+#ifdef FM_TX_ECC_FRMS_ERRATA_10GMAC_A004
-+static t_Error TgecTxEccWorkaround(t_Tgec *p_Tgec)
-+{
-+ t_Error err;
-+
-+#if defined(DEBUG_ERRORS) && (DEBUG_ERRORS > 0)
-+ XX_Print("Applying 10G TX ECC workaround (10GMAC-A004) ... ");
-+#endif /* (DEBUG_ERRORS > 0) */
-+ /* enable and set promiscuous */
-+ fman_tgec_enable(p_Tgec->p_MemMap, TRUE, TRUE);
-+ fman_tgec_set_promiscuous(p_Tgec->p_MemMap, TRUE);
-+ err = Fm10GTxEccWorkaround(p_Tgec->fmMacControllerDriver.h_Fm, p_Tgec->macId);
-+ /* disable */
-+ fman_tgec_set_promiscuous(p_Tgec->p_MemMap, FALSE);
-+ fman_tgec_enable(p_Tgec->p_MemMap, FALSE, FALSE);
-+ fman_tgec_reset_stat(p_Tgec->p_MemMap);
-+ fman_tgec_ack_event(p_Tgec->p_MemMap, 0xffffffff);
-+#if defined(DEBUG_ERRORS) && (DEBUG_ERRORS > 0)
-+ if (err)
-+ XX_Print("FAILED!\n");
-+ else
-+ XX_Print("done.\n");
-+#endif /* (DEBUG_ERRORS > 0) */
-+
-+ return err;
-+}
-+#endif /* FM_TX_ECC_FRMS_ERRATA_10GMAC_A004 */
-+
-+/*****************************************************************************/
-+/* FM Init & Free API */
-+/*****************************************************************************/
-+
-+/* ......................................................................... */
-+
-+static t_Error TgecInit(t_Handle h_Tgec)
-+{
-+ t_Tgec *p_Tgec = (t_Tgec *)h_Tgec;
-+ struct tgec_cfg *p_TgecDriverParam;
-+ t_EnetAddr ethAddr;
-+ t_Error err;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Tgec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Tgec->p_TgecDriverParam, E_INVALID_STATE);
-+ SANITY_CHECK_RETURN_ERROR(p_Tgec->fmMacControllerDriver.h_Fm, E_INVALID_HANDLE);
-+
-+ FM_GetRevision(p_Tgec->fmMacControllerDriver.h_Fm, &p_Tgec->fmMacControllerDriver.fmRevInfo);
-+ CHECK_INIT_PARAMETERS(p_Tgec, CheckInitParameters);
-+
-+ p_TgecDriverParam = p_Tgec->p_TgecDriverParam;
-+
-+ MAKE_ENET_ADDR_FROM_UINT64(p_Tgec->addr, ethAddr);
-+ fman_tgec_set_mac_address(p_Tgec->p_MemMap, (uint8_t *)ethAddr);
-+
-+ /* interrupts */
-+#ifdef FM_10G_REM_N_LCL_FLT_EX_10GMAC_ERRATA_SW005
-+ {
-+ if (p_Tgec->fmMacControllerDriver.fmRevInfo.majorRev <=2)
-+ p_Tgec->exceptions &= ~(TGEC_IMASK_REM_FAULT | TGEC_IMASK_LOC_FAULT);
-+ }
-+#endif /* FM_10G_REM_N_LCL_FLT_EX_10GMAC_ERRATA_SW005 */
-+
-+#ifdef FM_TX_ECC_FRMS_ERRATA_10GMAC_A004
-+ if (!p_Tgec->p_TgecDriverParam->skip_fman11_workaround &&
-+ ((err = TgecTxEccWorkaround(p_Tgec)) != E_OK))
-+ {
-+ FreeInitResources(p_Tgec);
-+ REPORT_ERROR(MINOR, err, ("TgecTxEccWorkaround FAILED"));
-+ }
-+#endif /* FM_TX_ECC_FRMS_ERRATA_10GMAC_A004 */
-+
-+ err = fman_tgec_init(p_Tgec->p_MemMap, p_TgecDriverParam, p_Tgec->exceptions);
-+ if (err)
-+ {
-+ FreeInitResources(p_Tgec);
-+ RETURN_ERROR(MAJOR, err, ("This TGEC version does not support the required i/f mode"));
-+ }
-+
-+ /* Max Frame Length */
-+ err = FmSetMacMaxFrame(p_Tgec->fmMacControllerDriver.h_Fm,
-+ e_FM_MAC_10G,
-+ p_Tgec->fmMacControllerDriver.macId,
-+ p_TgecDriverParam->max_frame_length);
-+ if (err != E_OK)
-+ {
-+ FreeInitResources(p_Tgec);
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+ }
-+/* we consider having no IPC a non crasher... */
-+
-+#ifdef FM_TX_FIFO_CORRUPTION_ERRATA_10GMAC_A007
-+ if (p_Tgec->fmMacControllerDriver.fmRevInfo.majorRev == 2)
-+ fman_tgec_set_erratum_tx_fifo_corruption_10gmac_a007(p_Tgec->p_MemMap);
-+#endif /* FM_TX_FIFO_CORRUPTION_ERRATA_10GMAC_A007 */
-+
-+ p_Tgec->p_MulticastAddrHash = AllocHashTable(HASH_TABLE_SIZE);
-+ if (!p_Tgec->p_MulticastAddrHash)
-+ {
-+ FreeInitResources(p_Tgec);
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, ("allocation hash table is FAILED"));
-+ }
-+
-+ p_Tgec->p_UnicastAddrHash = AllocHashTable(HASH_TABLE_SIZE);
-+ if (!p_Tgec->p_UnicastAddrHash)
-+ {
-+ FreeInitResources(p_Tgec);
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, ("allocation hash table is FAILED"));
-+ }
-+
-+ FmRegisterIntr(p_Tgec->fmMacControllerDriver.h_Fm,
-+ e_FM_MOD_10G_MAC,
-+ p_Tgec->macId,
-+ e_FM_INTR_TYPE_ERR,
-+ TgecErrException,
-+ p_Tgec);
-+ if (p_Tgec->mdioIrq != NO_IRQ)
-+ {
-+ XX_SetIntr(p_Tgec->mdioIrq, TgecException, p_Tgec);
-+ XX_EnableIntr(p_Tgec->mdioIrq);
-+ }
-+
-+ XX_Free(p_TgecDriverParam);
-+ p_Tgec->p_TgecDriverParam = NULL;
-+
-+ return E_OK;
-+}
-+
-+/* ......................................................................... */
-+
-+static t_Error TgecFree(t_Handle h_Tgec)
-+{
-+ t_Tgec *p_Tgec = (t_Tgec *)h_Tgec;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Tgec, E_INVALID_HANDLE);
-+
-+ if (p_Tgec->p_TgecDriverParam)
-+ {
-+ /* Called after config */
-+ XX_Free(p_Tgec->p_TgecDriverParam);
-+ p_Tgec->p_TgecDriverParam = NULL;
-+ }
-+ else
-+ /* Called after init */
-+ FreeInitResources(p_Tgec);
-+
-+ XX_Free(p_Tgec);
-+
-+ return E_OK;
-+}
-+
-+/* ......................................................................... */
-+
-+static void InitFmMacControllerDriver(t_FmMacControllerDriver *p_FmMacControllerDriver)
-+{
-+ p_FmMacControllerDriver->f_FM_MAC_Init = TgecInit;
-+ p_FmMacControllerDriver->f_FM_MAC_Free = TgecFree;
-+
-+ p_FmMacControllerDriver->f_FM_MAC_SetStatistics = NULL;
-+ p_FmMacControllerDriver->f_FM_MAC_ConfigLoopback = TgecConfigLoopback;
-+ p_FmMacControllerDriver->f_FM_MAC_ConfigMaxFrameLength = TgecConfigMaxFrameLength;
-+
-+ p_FmMacControllerDriver->f_FM_MAC_ConfigWan = TgecConfigWan;
-+
-+ p_FmMacControllerDriver->f_FM_MAC_ConfigPadAndCrc = NULL; /* TGEC always works with pad+crc */
-+ p_FmMacControllerDriver->f_FM_MAC_ConfigHalfDuplex = NULL; /* half-duplex is not supported in xgec */
-+ p_FmMacControllerDriver->f_FM_MAC_ConfigLengthCheck = TgecConfigLengthCheck;
-+ p_FmMacControllerDriver->f_FM_MAC_ConfigException = TgecConfigException;
-+ p_FmMacControllerDriver->f_FM_MAC_ConfigResetOnInit = NULL;
-+
-+#ifdef FM_TX_ECC_FRMS_ERRATA_10GMAC_A004
-+ p_FmMacControllerDriver->f_FM_MAC_ConfigSkipFman11Workaround= TgecConfigSkipFman11Workaround;
-+#endif /* FM_TX_ECC_FRMS_ERRATA_10GMAC_A004 */
-+
-+ p_FmMacControllerDriver->f_FM_MAC_SetException = TgecSetExcpetion;
-+
-+ p_FmMacControllerDriver->f_FM_MAC_Enable1588TimeStamp = TgecEnable1588TimeStamp;
-+ p_FmMacControllerDriver->f_FM_MAC_Disable1588TimeStamp = TgecDisable1588TimeStamp;
-+
-+ p_FmMacControllerDriver->f_FM_MAC_SetPromiscuous = TgecSetPromiscuous;
-+ p_FmMacControllerDriver->f_FM_MAC_AdjustLink = NULL;
-+ p_FmMacControllerDriver->f_FM_MAC_SetWakeOnLan = NULL;
-+ p_FmMacControllerDriver->f_FM_MAC_RestartAutoneg = NULL;
-+
-+ p_FmMacControllerDriver->f_FM_MAC_Enable = TgecEnable;
-+ p_FmMacControllerDriver->f_FM_MAC_Disable = TgecDisable;
-+ p_FmMacControllerDriver->f_FM_MAC_Resume = NULL;
-+
-+ p_FmMacControllerDriver->f_FM_MAC_SetTxAutoPauseFrames = TgecTxMacPause;
-+ p_FmMacControllerDriver->f_FM_MAC_SetTxPauseFrames = TgecSetTxPauseFrames;
-+ p_FmMacControllerDriver->f_FM_MAC_SetRxIgnorePauseFrames = TgecRxIgnoreMacPause;
-+
-+ p_FmMacControllerDriver->f_FM_MAC_ResetCounters = TgecResetCounters;
-+ p_FmMacControllerDriver->f_FM_MAC_GetStatistics = TgecGetStatistics;
-+ p_FmMacControllerDriver->f_FM_MAC_GetFrameSizeCounters = TgecGetFrameSizeCounters;
-+
-+ p_FmMacControllerDriver->f_FM_MAC_ModifyMacAddr = TgecModifyMacAddress;
-+ p_FmMacControllerDriver->f_FM_MAC_AddHashMacAddr = TgecAddHashMacAddress;
-+ p_FmMacControllerDriver->f_FM_MAC_RemoveHashMacAddr = TgecDelHashMacAddress;
-+ p_FmMacControllerDriver->f_FM_MAC_AddExactMatchMacAddr = TgecAddExactMatchMacAddress;
-+ p_FmMacControllerDriver->f_FM_MAC_RemovelExactMatchMacAddr = TgecDelExactMatchMacAddress;
-+ p_FmMacControllerDriver->f_FM_MAC_GetId = TgecGetId;
-+ p_FmMacControllerDriver->f_FM_MAC_GetVersion = TgecGetVersion;
-+ p_FmMacControllerDriver->f_FM_MAC_GetMaxFrameLength = TgecGetMaxFrameLength;
-+
-+ p_FmMacControllerDriver->f_FM_MAC_MII_WritePhyReg = TGEC_MII_WritePhyReg;
-+ p_FmMacControllerDriver->f_FM_MAC_MII_ReadPhyReg = TGEC_MII_ReadPhyReg;
-+}
-+
-+
-+/*****************************************************************************/
-+/* Tgec Config Main Entry */
-+/*****************************************************************************/
-+
-+/* ......................................................................... */
-+
-+t_Handle TGEC_Config(t_FmMacParams *p_FmMacParam)
-+{
-+ t_Tgec *p_Tgec;
-+ struct tgec_cfg *p_TgecDriverParam;
-+ uintptr_t baseAddr;
-+
-+ SANITY_CHECK_RETURN_VALUE(p_FmMacParam, E_NULL_POINTER, NULL);
-+
-+ baseAddr = p_FmMacParam->baseAddr;
-+ /* allocate memory for the UCC GETH data structure. */
-+ p_Tgec = (t_Tgec *)XX_Malloc(sizeof(t_Tgec));
-+ if (!p_Tgec)
-+ {
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("10G MAC driver structure"));
-+ return NULL;
-+ }
-+ memset(p_Tgec, 0, sizeof(t_Tgec));
-+ InitFmMacControllerDriver(&p_Tgec->fmMacControllerDriver);
-+
-+ /* allocate memory for the 10G MAC driver parameters data structure. */
-+ p_TgecDriverParam = (struct tgec_cfg *) XX_Malloc(sizeof(struct tgec_cfg));
-+ if (!p_TgecDriverParam)
-+ {
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("10G MAC driver parameters"));
-+ XX_Free(p_Tgec);
-+ return NULL;
-+ }
-+ memset(p_TgecDriverParam, 0, sizeof(struct tgec_cfg));
-+
-+ /* Plant parameter structure pointer */
-+ p_Tgec->p_TgecDriverParam = p_TgecDriverParam;
-+
-+ fman_tgec_defconfig(p_TgecDriverParam);
-+
-+ p_Tgec->p_MemMap = (struct tgec_regs *)UINT_TO_PTR(baseAddr);
-+ p_Tgec->p_MiiMemMap = (t_TgecMiiAccessMemMap *)UINT_TO_PTR(baseAddr + TGEC_TO_MII_OFFSET);
-+ p_Tgec->addr = ENET_ADDR_TO_UINT64(p_FmMacParam->addr);
-+ p_Tgec->enetMode = p_FmMacParam->enetMode;
-+ p_Tgec->macId = p_FmMacParam->macId;
-+ p_Tgec->exceptions = DEFAULT_exceptions;
-+ p_Tgec->mdioIrq = p_FmMacParam->mdioIrq;
-+ p_Tgec->f_Exception = p_FmMacParam->f_Exception;
-+ p_Tgec->f_Event = p_FmMacParam->f_Event;
-+ p_Tgec->h_App = p_FmMacParam->h_App;
-+
-+ return p_Tgec;
-+}
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MAC/tgec.h
-@@ -0,0 +1,151 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/******************************************************************************
-+ @File tgec.h
-+
-+ @Description FM 10G MAC ...
-+*//***************************************************************************/
-+#ifndef __TGEC_H
-+#define __TGEC_H
-+
-+#include "std_ext.h"
-+#include "error_ext.h"
-+#include "list_ext.h"
-+#include "enet_ext.h"
-+
-+#include "tgec_mii_acc.h"
-+#include "fm_mac.h"
-+
-+
-+#define DEFAULT_exceptions \
-+ ((uint32_t)(TGEC_IMASK_MDIO_SCAN_EVENT | \
-+ TGEC_IMASK_REM_FAULT | \
-+ TGEC_IMASK_LOC_FAULT | \
-+ TGEC_IMASK_TX_ECC_ER | \
-+ TGEC_IMASK_TX_FIFO_UNFL | \
-+ TGEC_IMASK_TX_FIFO_OVFL | \
-+ TGEC_IMASK_TX_ER | \
-+ TGEC_IMASK_RX_FIFO_OVFL | \
-+ TGEC_IMASK_RX_ECC_ER | \
-+ TGEC_IMASK_RX_JAB_FRM | \
-+ TGEC_IMASK_RX_OVRSZ_FRM | \
-+ TGEC_IMASK_RX_RUNT_FRM | \
-+ TGEC_IMASK_RX_FRAG_FRM | \
-+ TGEC_IMASK_RX_CRC_ER | \
-+ TGEC_IMASK_RX_ALIGN_ER))
-+
-+#define GET_EXCEPTION_FLAG(bitMask, exception) switch (exception){ \
-+ case e_FM_MAC_EX_10G_MDIO_SCAN_EVENTMDIO: \
-+ bitMask = TGEC_IMASK_MDIO_SCAN_EVENT ; break; \
-+ case e_FM_MAC_EX_10G_MDIO_CMD_CMPL: \
-+ bitMask = TGEC_IMASK_MDIO_CMD_CMPL ; break; \
-+ case e_FM_MAC_EX_10G_REM_FAULT: \
-+ bitMask = TGEC_IMASK_REM_FAULT ; break; \
-+ case e_FM_MAC_EX_10G_LOC_FAULT: \
-+ bitMask = TGEC_IMASK_LOC_FAULT ; break; \
-+ case e_FM_MAC_EX_10G_1TX_ECC_ER: \
-+ bitMask = TGEC_IMASK_TX_ECC_ER ; break; \
-+ case e_FM_MAC_EX_10G_TX_FIFO_UNFL: \
-+ bitMask = TGEC_IMASK_TX_FIFO_UNFL ; break; \
-+ case e_FM_MAC_EX_10G_TX_FIFO_OVFL: \
-+ bitMask = TGEC_IMASK_TX_FIFO_OVFL ; break; \
-+ case e_FM_MAC_EX_10G_TX_ER: \
-+ bitMask = TGEC_IMASK_TX_ER ; break; \
-+ case e_FM_MAC_EX_10G_RX_FIFO_OVFL: \
-+ bitMask = TGEC_IMASK_RX_FIFO_OVFL ; break; \
-+ case e_FM_MAC_EX_10G_RX_ECC_ER: \
-+ bitMask = TGEC_IMASK_RX_ECC_ER ; break; \
-+ case e_FM_MAC_EX_10G_RX_JAB_FRM: \
-+ bitMask = TGEC_IMASK_RX_JAB_FRM ; break; \
-+ case e_FM_MAC_EX_10G_RX_OVRSZ_FRM: \
-+ bitMask = TGEC_IMASK_RX_OVRSZ_FRM ; break; \
-+ case e_FM_MAC_EX_10G_RX_RUNT_FRM: \
-+ bitMask = TGEC_IMASK_RX_RUNT_FRM ; break; \
-+ case e_FM_MAC_EX_10G_RX_FRAG_FRM: \
-+ bitMask = TGEC_IMASK_RX_FRAG_FRM ; break; \
-+ case e_FM_MAC_EX_10G_RX_LEN_ER: \
-+ bitMask = TGEC_IMASK_RX_LEN_ER ; break; \
-+ case e_FM_MAC_EX_10G_RX_CRC_ER: \
-+ bitMask = TGEC_IMASK_RX_CRC_ER ; break; \
-+ case e_FM_MAC_EX_10G_RX_ALIGN_ER: \
-+ bitMask = TGEC_IMASK_RX_ALIGN_ER ; break; \
-+ default: bitMask = 0;break;}
-+
-+#define MAX_PACKET_ALIGNMENT 31
-+#define MAX_INTER_PACKET_GAP 0x7f
-+#define MAX_INTER_PALTERNATE_BEB 0x0f
-+#define MAX_RETRANSMISSION 0x0f
-+#define MAX_COLLISION_WINDOW 0x03ff
-+
-+#define TGEC_NUM_OF_PADDRS 1 /* number of pattern match registers (entries) */
-+
-+#define GROUP_ADDRESS 0x0000010000000000LL /* Group address bit indication */
-+
-+#define HASH_TABLE_SIZE 512 /* Hash table size (= 32 bits * 8 regs) */
-+
-+#define TGEC_TO_MII_OFFSET 0x1030 /* Offset from the MEM map to the MDIO mem map */
-+
-+/* 10-gigabit Ethernet MAC Controller ID (10GEC_ID) */
-+#define TGEC_ID_ID 0xffff0000
-+#define TGEC_ID_MAC_VERSION 0x0000FF00
-+#define TGEC_ID_MAC_REV 0x000000ff
-+
-+
-+typedef struct {
-+ t_FmMacControllerDriver fmMacControllerDriver; /**< Upper Mac control block */
-+ t_Handle h_App; /**< Handle to the upper layer application */
-+ struct tgec_regs *p_MemMap; /**< pointer to 10G memory mapped registers. */
-+ t_TgecMiiAccessMemMap *p_MiiMemMap; /**< pointer to MII memory mapped registers. */
-+ uint64_t addr; /**< MAC address of device; */
-+ e_EnetMode enetMode; /**< Ethernet physical interface */
-+ t_FmMacExceptionCallback *f_Exception;
-+ int mdioIrq;
-+ t_FmMacExceptionCallback *f_Event;
-+ bool indAddrRegUsed[TGEC_NUM_OF_PADDRS]; /**< Whether a particular individual address recognition register is being used */
-+ uint64_t paddr[TGEC_NUM_OF_PADDRS]; /**< MAC address for particular individual address recognition register */
-+ uint8_t numOfIndAddrInRegs; /**< Number of individual addresses in registers for this station. */
-+ t_EthHash *p_MulticastAddrHash; /**< pointer to driver's global address hash table */
-+ t_EthHash *p_UnicastAddrHash; /**< pointer to driver's individual address hash table */
-+ bool debugMode;
-+ uint8_t macId;
-+ uint32_t exceptions;
-+ struct tgec_cfg *p_TgecDriverParam;
-+} t_Tgec;
-+
-+
-+t_Error TGEC_MII_WritePhyReg(t_Handle h_Tgec, uint8_t phyAddr, uint8_t reg, uint16_t data);
-+t_Error TGEC_MII_ReadPhyReg(t_Handle h_Tgec, uint8_t phyAddr, uint8_t reg, uint16_t *p_Data);
-+
-+
-+#endif /* __TGEC_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MAC/tgec_mii_acc.c
-@@ -0,0 +1,139 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+
-+#include "error_ext.h"
-+#include "std_ext.h"
-+#include "fm_mac.h"
-+#include "tgec.h"
-+#include "xx_ext.h"
-+
-+#include "fm_common.h"
-+
-+
-+/*****************************************************************************/
-+t_Error TGEC_MII_WritePhyReg(t_Handle h_Tgec,
-+ uint8_t phyAddr,
-+ uint8_t reg,
-+ uint16_t data)
-+{
-+ t_Tgec *p_Tgec = (t_Tgec *)h_Tgec;
-+ t_TgecMiiAccessMemMap *p_MiiAccess;
-+ uint32_t cfgStatusReg;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Tgec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Tgec->p_MiiMemMap, E_INVALID_HANDLE);
-+
-+ p_MiiAccess = p_Tgec->p_MiiMemMap;
-+
-+ /* Configure MII */
-+ cfgStatusReg = GET_UINT32(p_MiiAccess->mdio_cfg_status);
-+ cfgStatusReg &= ~MIIMCOM_DIV_MASK;
-+ /* (one half of fm clock => 2.5Mhz) */
-+ cfgStatusReg |=((((p_Tgec->fmMacControllerDriver.clkFreq*10)/2)/25) << MIIMCOM_DIV_SHIFT);
-+ WRITE_UINT32(p_MiiAccess->mdio_cfg_status, cfgStatusReg);
-+
-+ while ((GET_UINT32(p_MiiAccess->mdio_cfg_status)) & MIIMIND_BUSY)
-+ XX_UDelay (1);
-+
-+ WRITE_UINT32(p_MiiAccess->mdio_command, phyAddr);
-+
-+ WRITE_UINT32(p_MiiAccess->mdio_regaddr, reg);
-+
-+ CORE_MemoryBarrier();
-+
-+ while ((GET_UINT32(p_MiiAccess->mdio_cfg_status)) & MIIMIND_BUSY)
-+ XX_UDelay (1);
-+
-+ WRITE_UINT32(p_MiiAccess->mdio_data, data);
-+
-+ CORE_MemoryBarrier();
-+
-+ while ((GET_UINT32(p_MiiAccess->mdio_data)) & MIIDATA_BUSY)
-+ XX_UDelay (1);
-+
-+ return E_OK;
-+}
-+
-+/*****************************************************************************/
-+t_Error TGEC_MII_ReadPhyReg(t_Handle h_Tgec,
-+ uint8_t phyAddr,
-+ uint8_t reg,
-+ uint16_t *p_Data)
-+{
-+ t_Tgec *p_Tgec = (t_Tgec *)h_Tgec;
-+ t_TgecMiiAccessMemMap *p_MiiAccess;
-+ uint32_t cfgStatusReg;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Tgec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Tgec->p_MiiMemMap, E_INVALID_HANDLE);
-+
-+ p_MiiAccess = p_Tgec->p_MiiMemMap;
-+
-+ /* Configure MII */
-+ cfgStatusReg = GET_UINT32(p_MiiAccess->mdio_cfg_status);
-+ cfgStatusReg &= ~MIIMCOM_DIV_MASK;
-+ /* (one half of fm clock => 2.5Mhz) */
-+ cfgStatusReg |=((((p_Tgec->fmMacControllerDriver.clkFreq*10)/2)/25) << MIIMCOM_DIV_SHIFT);
-+ WRITE_UINT32(p_MiiAccess->mdio_cfg_status, cfgStatusReg);
-+
-+ while ((GET_UINT32(p_MiiAccess->mdio_cfg_status)) & MIIMIND_BUSY)
-+ XX_UDelay (1);
-+
-+ WRITE_UINT32(p_MiiAccess->mdio_command, phyAddr);
-+
-+ WRITE_UINT32(p_MiiAccess->mdio_regaddr, reg);
-+
-+ CORE_MemoryBarrier();
-+
-+ while ((GET_UINT32(p_MiiAccess->mdio_cfg_status)) & MIIMIND_BUSY)
-+ XX_UDelay (1);
-+
-+ WRITE_UINT32(p_MiiAccess->mdio_command, (uint32_t)(phyAddr | MIIMCOM_READ_CYCLE));
-+
-+ CORE_MemoryBarrier();
-+
-+ while ((GET_UINT32(p_MiiAccess->mdio_data)) & MIIDATA_BUSY)
-+ XX_UDelay (1);
-+
-+ *p_Data = (uint16_t)GET_UINT32(p_MiiAccess->mdio_data);
-+
-+ cfgStatusReg = GET_UINT32(p_MiiAccess->mdio_cfg_status);
-+
-+ if (cfgStatusReg & MIIMIND_READ_ERROR)
-+ RETURN_ERROR(MINOR, E_INVALID_VALUE,
-+ ("Read Error: phyAddr 0x%x, dev 0x%x, reg 0x%x, cfgStatusReg 0x%x",
-+ ((phyAddr & 0xe0)>>5), (phyAddr & 0x1f), reg, cfgStatusReg));
-+
-+ return E_OK;
-+}
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MAC/tgec_mii_acc.h
-@@ -0,0 +1,80 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+#ifndef __TGEC_MII_ACC_H
-+#define __TGEC_MII_ACC_H
-+
-+#include "std_ext.h"
-+
-+
-+/* MII Management Command Register */
-+#define MIIMCOM_READ_POST_INCREMENT 0x00004000
-+#define MIIMCOM_READ_CYCLE 0x00008000
-+#define MIIMCOM_SCAN_CYCLE 0x00000800
-+#define MIIMCOM_PREAMBLE_DISABLE 0x00000400
-+
-+#define MIIMCOM_MDIO_HOLD_1_REG_CLK 0
-+#define MIIMCOM_MDIO_HOLD_2_REG_CLK 1
-+#define MIIMCOM_MDIO_HOLD_3_REG_CLK 2
-+#define MIIMCOM_MDIO_HOLD_4_REG_CLK 3
-+
-+#define MIIMCOM_DIV_MASK 0x0000ff00
-+#define MIIMCOM_DIV_SHIFT 8
-+
-+/* MII Management Indicator Register */
-+#define MIIMIND_BUSY 0x00000001
-+#define MIIMIND_READ_ERROR 0x00000002
-+
-+#define MIIDATA_BUSY 0x80000000
-+
-+#if defined(__MWERKS__) && !defined(__GNUC__)
-+#pragma pack(push,1)
-+#endif /* defined(__MWERKS__) && ... */
-+
-+/*----------------------------------------------------*/
-+/* MII Configuration Control Memory Map Registers */
-+/*----------------------------------------------------*/
-+typedef _Packed struct t_TgecMiiAccessMemMap
-+{
-+ volatile uint32_t mdio_cfg_status; /* 0x030 */
-+ volatile uint32_t mdio_command; /* 0x034 */
-+ volatile uint32_t mdio_data; /* 0x038 */
-+ volatile uint32_t mdio_regaddr; /* 0x03c */
-+} _PackedType t_TgecMiiAccessMemMap ;
-+
-+#if defined(__MWERKS__) && !defined(__GNUC__)
-+#pragma pack(pop)
-+#endif /* defined(__MWERKS__) && ... */
-+
-+
-+#endif /* __TGEC_MII_ACC_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MACSEC/Makefile
-@@ -0,0 +1,15 @@
-+#
-+# Makefile for the Freescale Ethernet controllers
-+#
-+ccflags-y += -DVERSION=\"\"
-+#
-+#Include netcomm SW specific definitions
-+include $(srctree)/drivers/net/ethernet/freescale/sdk_fman/ncsw_config.mk
-+
-+NCSW_FM_INC = $(srctree)/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/inc
-+
-+ccflags-y += -I$(NCSW_FM_INC)
-+
-+obj-y += fsl-ncsw-macsec.o
-+
-+fsl-ncsw-macsec-objs := fm_macsec.o fm_macsec_guest.o fm_macsec_master.o fm_macsec_secy.o
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MACSEC/fm_macsec.c
-@@ -0,0 +1,237 @@
-+/*
-+ * Copyright 2008-2015 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+/******************************************************************************
-+
-+ @File fm_macsec.c
-+
-+ @Description FM MACSEC driver routines implementation.
-+*//***************************************************************************/
-+
-+#include "std_ext.h"
-+#include "error_ext.h"
-+#include "xx_ext.h"
-+#include "string_ext.h"
-+#include "sprint_ext.h"
-+#include "debug_ext.h"
-+
-+#include "fm_macsec.h"
-+
-+
-+/****************************************/
-+/* API Init unit functions */
-+/****************************************/
-+t_Handle FM_MACSEC_Config(t_FmMacsecParams *p_FmMacsecParam)
-+{
-+ t_FmMacsecControllerDriver *p_FmMacsecControllerDriver;
-+
-+ SANITY_CHECK_RETURN_VALUE(p_FmMacsecParam, E_INVALID_HANDLE, NULL);
-+
-+ if (p_FmMacsecParam->guestMode)
-+ p_FmMacsecControllerDriver = (t_FmMacsecControllerDriver *)FM_MACSEC_GUEST_Config(p_FmMacsecParam);
-+ else
-+ p_FmMacsecControllerDriver = (t_FmMacsecControllerDriver *)FM_MACSEC_MASTER_Config(p_FmMacsecParam);
-+
-+ if (!p_FmMacsecControllerDriver)
-+ return NULL;
-+
-+ return (t_Handle)p_FmMacsecControllerDriver;
-+}
-+
-+t_Error FM_MACSEC_Init(t_Handle h_FmMacsec)
-+{
-+ t_FmMacsecControllerDriver *p_FmMacsecControllerDriver = (t_FmMacsecControllerDriver *)h_FmMacsec;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsecControllerDriver, E_INVALID_HANDLE);
-+
-+ if (p_FmMacsecControllerDriver->f_FM_MACSEC_Init)
-+ return p_FmMacsecControllerDriver->f_FM_MACSEC_Init(h_FmMacsec);
-+
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, NO_MSG);
-+}
-+
-+t_Error FM_MACSEC_Free(t_Handle h_FmMacsec)
-+{
-+ t_FmMacsecControllerDriver *p_FmMacsecControllerDriver = (t_FmMacsecControllerDriver *)h_FmMacsec;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsecControllerDriver, E_INVALID_HANDLE);
-+
-+ if (p_FmMacsecControllerDriver->f_FM_MACSEC_Free)
-+ return p_FmMacsecControllerDriver->f_FM_MACSEC_Free(h_FmMacsec);
-+
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, NO_MSG);
-+}
-+
-+t_Error FM_MACSEC_ConfigUnknownSciFrameTreatment(t_Handle h_FmMacsec, e_FmMacsecUnknownSciFrameTreatment treatMode)
-+{
-+ t_FmMacsecControllerDriver *p_FmMacsecControllerDriver = (t_FmMacsecControllerDriver *)h_FmMacsec;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsecControllerDriver, E_INVALID_HANDLE);
-+
-+ if (p_FmMacsecControllerDriver->f_FM_MACSEC_ConfigUnknownSciFrameTreatment)
-+ return p_FmMacsecControllerDriver->f_FM_MACSEC_ConfigUnknownSciFrameTreatment(h_FmMacsec, treatMode);
-+
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, NO_MSG);
-+}
-+
-+t_Error FM_MACSEC_ConfigInvalidTagsFrameTreatment(t_Handle h_FmMacsec, bool deliverUncontrolled)
-+{
-+ t_FmMacsecControllerDriver *p_FmMacsecControllerDriver = (t_FmMacsecControllerDriver *)h_FmMacsec;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsecControllerDriver, E_INVALID_HANDLE);
-+
-+ if (p_FmMacsecControllerDriver->f_FM_MACSEC_ConfigInvalidTagsFrameTreatment)
-+ return p_FmMacsecControllerDriver->f_FM_MACSEC_ConfigInvalidTagsFrameTreatment(h_FmMacsec, deliverUncontrolled);
-+
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, NO_MSG);
-+}
-+
-+t_Error FM_MACSEC_ConfigEncryptWithNoChangedTextFrameTreatment(t_Handle h_FmMacsec, bool discardUncontrolled)
-+{
-+ t_FmMacsecControllerDriver *p_FmMacsecControllerDriver = (t_FmMacsecControllerDriver *)h_FmMacsec;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsecControllerDriver, E_INVALID_HANDLE);
-+
-+ if (p_FmMacsecControllerDriver->f_FM_MACSEC_ConfigEncryptWithNoChangedTextFrameTreatment)
-+ return p_FmMacsecControllerDriver->f_FM_MACSEC_ConfigEncryptWithNoChangedTextFrameTreatment(h_FmMacsec, discardUncontrolled);
-+
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, NO_MSG);
-+}
-+
-+t_Error FM_MACSEC_ConfigUntagFrameTreatment(t_Handle h_FmMacsec, e_FmMacsecUntagFrameTreatment treatMode)
-+{
-+ t_FmMacsecControllerDriver *p_FmMacsecControllerDriver = (t_FmMacsecControllerDriver *)h_FmMacsec;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsecControllerDriver, E_INVALID_HANDLE);
-+
-+ if (p_FmMacsecControllerDriver->f_FM_MACSEC_ConfigUntagFrameTreatment)
-+ return p_FmMacsecControllerDriver->f_FM_MACSEC_ConfigUntagFrameTreatment(h_FmMacsec, treatMode);
-+
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, NO_MSG);
-+}
-+
-+t_Error FM_MACSEC_ConfigPnExhaustionThreshold(t_Handle h_FmMacsec, uint32_t pnExhThr)
-+{
-+ t_FmMacsecControllerDriver *p_FmMacsecControllerDriver = (t_FmMacsecControllerDriver *)h_FmMacsec;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsecControllerDriver, E_INVALID_HANDLE);
-+
-+ if (p_FmMacsecControllerDriver->f_FM_MACSEC_ConfigPnExhaustionThreshold)
-+ return p_FmMacsecControllerDriver->f_FM_MACSEC_ConfigPnExhaustionThreshold(h_FmMacsec, pnExhThr);
-+
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, NO_MSG);
-+}
-+
-+t_Error FM_MACSEC_ConfigKeysUnreadable(t_Handle h_FmMacsec)
-+{
-+ t_FmMacsecControllerDriver *p_FmMacsecControllerDriver = (t_FmMacsecControllerDriver *)h_FmMacsec;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsecControllerDriver, E_INVALID_HANDLE);
-+
-+ if (p_FmMacsecControllerDriver->f_FM_MACSEC_ConfigKeysUnreadable)
-+ return p_FmMacsecControllerDriver->f_FM_MACSEC_ConfigKeysUnreadable(h_FmMacsec);
-+
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, NO_MSG);
-+}
-+
-+t_Error FM_MACSEC_ConfigSectagWithoutSCI(t_Handle h_FmMacsec)
-+{
-+ t_FmMacsecControllerDriver *p_FmMacsecControllerDriver = (t_FmMacsecControllerDriver *)h_FmMacsec;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsecControllerDriver, E_INVALID_HANDLE);
-+
-+ if (p_FmMacsecControllerDriver->f_FM_MACSEC_ConfigSectagWithoutSCI)
-+ return p_FmMacsecControllerDriver->f_FM_MACSEC_ConfigSectagWithoutSCI(h_FmMacsec);
-+
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, NO_MSG);
-+}
-+
-+t_Error FM_MACSEC_ConfigException(t_Handle h_FmMacsec, e_FmMacsecExceptions exception, bool enable)
-+{
-+ t_FmMacsecControllerDriver *p_FmMacsecControllerDriver = (t_FmMacsecControllerDriver *)h_FmMacsec;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsecControllerDriver, E_INVALID_HANDLE);
-+
-+ if (p_FmMacsecControllerDriver->f_FM_MACSEC_ConfigException)
-+ return p_FmMacsecControllerDriver->f_FM_MACSEC_ConfigException(h_FmMacsec, exception, enable);
-+
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, NO_MSG);
-+}
-+
-+t_Error FM_MACSEC_GetRevision(t_Handle h_FmMacsec, uint32_t *p_MacsecRevision)
-+{
-+ t_FmMacsecControllerDriver *p_FmMacsecControllerDriver = (t_FmMacsecControllerDriver *)h_FmMacsec;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsecControllerDriver, E_INVALID_HANDLE);
-+
-+ if (p_FmMacsecControllerDriver->f_FM_MACSEC_GetRevision)
-+ return p_FmMacsecControllerDriver->f_FM_MACSEC_GetRevision(h_FmMacsec, p_MacsecRevision);
-+
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, NO_MSG);
-+}
-+
-+
-+t_Error FM_MACSEC_Enable(t_Handle h_FmMacsec)
-+{
-+ t_FmMacsecControllerDriver *p_FmMacsecControllerDriver = (t_FmMacsecControllerDriver *)h_FmMacsec;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsecControllerDriver, E_INVALID_HANDLE);
-+
-+ if (p_FmMacsecControllerDriver->f_FM_MACSEC_Enable)
-+ return p_FmMacsecControllerDriver->f_FM_MACSEC_Enable(h_FmMacsec);
-+
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, NO_MSG);
-+}
-+
-+t_Error FM_MACSEC_Disable(t_Handle h_FmMacsec)
-+{
-+ t_FmMacsecControllerDriver *p_FmMacsecControllerDriver = (t_FmMacsecControllerDriver *)h_FmMacsec;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsecControllerDriver, E_INVALID_HANDLE);
-+
-+ if (p_FmMacsecControllerDriver->f_FM_MACSEC_Disable)
-+ return p_FmMacsecControllerDriver->f_FM_MACSEC_Disable(h_FmMacsec);
-+
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, NO_MSG);
-+}
-+
-+t_Error FM_MACSEC_SetException(t_Handle h_FmMacsec, e_FmMacsecExceptions exception, bool enable)
-+{
-+ t_FmMacsecControllerDriver *p_FmMacsecControllerDriver = (t_FmMacsecControllerDriver *)h_FmMacsec;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsecControllerDriver, E_INVALID_HANDLE);
-+
-+ if (p_FmMacsecControllerDriver->f_FM_MACSEC_SetException)
-+ return p_FmMacsecControllerDriver->f_FM_MACSEC_SetException(h_FmMacsec, exception, enable);
-+
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, NO_MSG);
-+}
-+
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MACSEC/fm_macsec.h
-@@ -0,0 +1,203 @@
-+/*
-+ * Copyright 2008-2015 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+/******************************************************************************
-+ @File fm_macsec.h
-+
-+ @Description FM MACSEC internal structures and definitions.
-+*//***************************************************************************/
-+#ifndef __FM_MACSEC_H
-+#define __FM_MACSEC_H
-+
-+#include "error_ext.h"
-+#include "std_ext.h"
-+#include "fm_macsec_ext.h"
-+
-+#include "fm_common.h"
-+
-+
-+#define __ERR_MODULE__ MODULE_FM_MACSEC
-+
-+
-+typedef struct
-+{
-+ t_Error (*f_FM_MACSEC_Init) (t_Handle h_FmMacsec);
-+ t_Error (*f_FM_MACSEC_Free) (t_Handle h_FmMacsec);
-+
-+ t_Error (*f_FM_MACSEC_ConfigUnknownSciFrameTreatment) (t_Handle h_FmMacsec, e_FmMacsecUnknownSciFrameTreatment treatMode);
-+ t_Error (*f_FM_MACSEC_ConfigInvalidTagsFrameTreatment) (t_Handle h_FmMacsec, bool deliverUncontrolled);
-+ t_Error (*f_FM_MACSEC_ConfigEncryptWithNoChangedTextFrameTreatment) (t_Handle h_FmMacsec, bool discardUncontrolled);
-+ t_Error (*f_FM_MACSEC_ConfigChangedTextWithNoEncryptFrameTreatment) (t_Handle h_FmMacsec, bool deliverUncontrolled);
-+ t_Error (*f_FM_MACSEC_ConfigUntagFrameTreatment) (t_Handle h_FmMacsec, e_FmMacsecUntagFrameTreatment treatMode);
-+ t_Error (*f_FM_MACSEC_ConfigOnlyScbIsSetFrameTreatment) (t_Handle h_FmMacsec, bool deliverUncontrolled);
-+ t_Error (*f_FM_MACSEC_ConfigPnExhaustionThreshold) (t_Handle h_FmMacsec, uint32_t pnExhThr);
-+ t_Error (*f_FM_MACSEC_ConfigKeysUnreadable) (t_Handle h_FmMacsec);
-+ t_Error (*f_FM_MACSEC_ConfigSectagWithoutSCI) (t_Handle h_FmMacsec);
-+ t_Error (*f_FM_MACSEC_ConfigException) (t_Handle h_FmMacsec, e_FmMacsecExceptions exception, bool enable);
-+
-+ t_Error (*f_FM_MACSEC_GetRevision) (t_Handle h_FmMacsec, uint32_t *p_MacsecRevision);
-+ t_Error (*f_FM_MACSEC_Enable) (t_Handle h_FmMacsec);
-+ t_Error (*f_FM_MACSEC_Disable) (t_Handle h_FmMacsec);
-+ t_Error (*f_FM_MACSEC_SetException) (t_Handle h_FmMacsec, e_FmMacsecExceptions exception, bool enable);
-+
-+} t_FmMacsecControllerDriver;
-+
-+t_Handle FM_MACSEC_GUEST_Config(t_FmMacsecParams *p_FmMacsecParam);
-+t_Handle FM_MACSEC_MASTER_Config(t_FmMacsecParams *p_FmMacsecParams);
-+
-+/***********************************************************************/
-+/* MACSEC internal routines */
-+/***********************************************************************/
-+
-+/**************************************************************************//**
-+
-+ @Group FM_MACSEC_InterModule_grp FM MACSEC Inter-Module Unit
-+
-+ @Description FM MACSEC Inter Module functions -
-+ These are not User API routines but routines that may be called
-+ from other modules. This will be the case in a single core environment,
-+ where instead of using the XX messaging mechanism, the routines may be
-+ called from other modules. In a multicore environment, the other modules may
-+ be run by other cores and therefore these routines may not be called directly.
-+
-+ @{
-+*//***************************************************************************/
-+
-+#define MAX_NUM_OF_SA_PER_SC 4
-+
-+typedef enum
-+{
-+ e_SC_RX = 0,
-+ e_SC_TX
-+} e_ScType;
-+
-+typedef enum
-+{
-+ e_SC_SA_A = 0,
-+ e_SC_SA_B ,
-+ e_SC_SA_C ,
-+ e_SC_SA_D
-+} e_ScSaId;
-+
-+typedef struct
-+{
-+ uint32_t scId;
-+ macsecSCI_t sci;
-+ bool replayProtect;
-+ uint32_t replayWindow;
-+ e_FmMacsecValidFrameBehavior validateFrames;
-+ uint16_t confidentialityOffset;
-+ e_FmMacsecSecYCipherSuite cipherSuite;
-+} t_RxScParams;
-+
-+typedef struct
-+{
-+ uint32_t scId;
-+ macsecSCI_t sci;
-+ bool protectFrames;
-+ e_FmMacsecSciInsertionMode sciInsertionMode;
-+ bool confidentialityEnable;
-+ uint16_t confidentialityOffset;
-+ e_FmMacsecSecYCipherSuite cipherSuite;
-+} t_TxScParams;
-+
-+typedef enum e_FmMacsecGlobalExceptions {
-+ e_FM_MACSEC_EX_TX_SC, /**< Tx Sc 0 frame discarded error. */
-+ e_FM_MACSEC_EX_ECC /**< MACSEC memory ECC multiple-bit error. */
-+} e_FmMacsecGlobalExceptions;
-+
-+typedef enum e_FmMacsecGlobalEvents {
-+ e_FM_MACSEC_EV_TX_SC_NEXT_PN /**< Tx Sc 0 Next Pn exhaustion threshold reached. */
-+} e_FmMacsecGlobalEvents;
-+
-+/**************************************************************************//**
-+ @Description Enum for inter-module interrupts registration
-+*//***************************************************************************/
-+typedef enum e_FmMacsecEventModules{
-+ e_FM_MACSEC_MOD_SC_TX,
-+ e_FM_MACSEC_MOD_DUMMY_LAST
-+} e_FmMacsecEventModules;
-+
-+typedef enum e_FmMacsecInterModuleEvent {
-+ e_FM_MACSEC_EV_SC_TX,
-+ e_FM_MACSEC_EV_ERR_SC_TX,
-+ e_FM_MACSEC_EV_DUMMY_LAST
-+} e_FmMacsecInterModuleEvent;
-+
-+#define NUM_OF_INTER_MODULE_EVENTS (NUM_OF_TX_SC * 2)
-+
-+#define GET_MACSEC_MODULE_EVENT(mod, id, intrType, event) \
-+ switch(mod){ \
-+ case e_FM_MACSEC_MOD_SC_TX: \
-+ event = (intrType == e_FM_INTR_TYPE_ERR) ? \
-+ e_FM_MACSEC_EV_ERR_SC_TX: \
-+ e_FM_MACSEC_EV_SC_TX; \
-+ event += (uint8_t)(2 * id);break; \
-+ break; \
-+ default:event = e_FM_MACSEC_EV_DUMMY_LAST; \
-+ break;}
-+
-+void FmMacsecRegisterIntr(t_Handle h_FmMacsec,
-+ e_FmMacsecEventModules module,
-+ uint8_t modId,
-+ e_FmIntrType intrType,
-+ void (*f_Isr) (t_Handle h_Arg, uint32_t id),
-+ t_Handle h_Arg);
-+
-+void FmMacsecUnregisterIntr(t_Handle h_FmMacsec,
-+ e_FmMacsecEventModules module,
-+ uint8_t modId,
-+ e_FmIntrType intrType);
-+
-+t_Error FmMacsecAllocScs(t_Handle h_FmMacsec, e_ScType type, bool isPtp, uint32_t numOfScs, uint32_t *p_ScIds);
-+t_Error FmMacsecFreeScs(t_Handle h_FmMacsec, e_ScType type, uint32_t numOfScs, uint32_t *p_ScIds);
-+t_Error FmMacsecCreateRxSc(t_Handle h_FmMacsec, t_RxScParams *p_RxScParams);
-+t_Error FmMacsecDeleteRxSc(t_Handle h_FmMacsec, uint32_t scId);
-+t_Error FmMacsecCreateTxSc(t_Handle h_FmMacsec, t_TxScParams *p_RxScParams);
-+t_Error FmMacsecDeleteTxSc(t_Handle h_FmMacsec, uint32_t scId);
-+t_Error FmMacsecCreateRxSa(t_Handle h_FmMacsec, uint32_t scId, e_ScSaId saId, macsecAN_t an, uint32_t lowestPn, macsecSAKey_t key);
-+t_Error FmMacsecCreateTxSa(t_Handle h_FmMacsec, uint32_t scId, e_ScSaId saId, macsecSAKey_t key);
-+t_Error FmMacsecDeleteRxSa(t_Handle h_FmMacsec, uint32_t scId, e_ScSaId saId);
-+t_Error FmMacsecDeleteTxSa(t_Handle h_FmMacsec, uint32_t scId, e_ScSaId saId);
-+t_Error FmMacsecRxSaSetReceive(t_Handle h_FmMacsec, uint32_t scId, e_ScSaId saId, bool enableReceive);
-+t_Error FmMacsecRxSaUpdateNextPn(t_Handle h_FmMacsec, uint32_t scId, e_ScSaId saId, uint32_t updtNextPN);
-+t_Error FmMacsecRxSaUpdateLowestPn(t_Handle h_FmMacsec, uint32_t scId, e_ScSaId saId, uint32_t updtLowestPN);
-+t_Error FmMacsecTxSaSetActive(t_Handle h_FmMacsec, uint32_t scId, e_ScSaId saId, macsecAN_t an);
-+t_Error FmMacsecTxSaGetActive(t_Handle h_FmMacsec, uint32_t scId, macsecAN_t *p_An);
-+t_Error FmMacsecSetPTP(t_Handle h_FmMacsec, bool enable);
-+
-+t_Error FmMacsecSetException(t_Handle h_FmMacsec, e_FmMacsecGlobalExceptions exception, uint32_t scId, bool enable);
-+t_Error FmMacsecSetEvent(t_Handle h_FmMacsec, e_FmMacsecGlobalEvents event, uint32_t scId, bool enable);
-+
-+
-+
-+#endif /* __FM_MACSEC_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MACSEC/fm_macsec_guest.c
-@@ -0,0 +1,59 @@
-+/*
-+ * Copyright 2008-2015 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+/******************************************************************************
-+ @File fm_macsec.c
-+
-+ @Description FM MACSEC driver routines implementation.
-+*//***************************************************************************/
-+
-+#include "std_ext.h"
-+#include "error_ext.h"
-+#include "xx_ext.h"
-+#include "string_ext.h"
-+#include "sprint_ext.h"
-+#include "debug_ext.h"
-+#include "fm_macsec.h"
-+
-+
-+/****************************************/
-+/* static functions */
-+/****************************************/
-+
-+/****************************************/
-+/* API Init unit functions */
-+/****************************************/
-+t_Handle FM_MACSEC_GUEST_Config(t_FmMacsecParams *p_FmMacsecParam)
-+{
-+ UNUSED(p_FmMacsecParam);
-+ return NULL;
-+}
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MACSEC/fm_macsec_master.c
-@@ -0,0 +1,1031 @@
-+/*
-+ * Copyright 2008-2015 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+/******************************************************************************
-+ @File fm_macsec.c
-+
-+ @Description FM MACSEC driver routines implementation.
-+*//***************************************************************************/
-+
-+#include "std_ext.h"
-+#include "error_ext.h"
-+#include "xx_ext.h"
-+#include "string_ext.h"
-+#include "sprint_ext.h"
-+#include "fm_mac_ext.h"
-+
-+#include "fm_macsec_master.h"
-+
-+
-+extern uint16_t FM_MAC_GetMaxFrameLength(t_Handle FmMac);
-+
-+
-+/****************************************/
-+/* static functions */
-+/****************************************/
-+static t_Error CheckFmMacsecParameters(t_FmMacsec *p_FmMacsec)
-+{
-+ if (!p_FmMacsec->f_Exception)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("Exceptions callback not provided"));
-+
-+ return E_OK;
-+}
-+
-+static void UnimplementedIsr(t_Handle h_Arg, uint32_t id)
-+{
-+ UNUSED(h_Arg); UNUSED(id);
-+
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Unimplemented Isr!"));
-+}
-+
-+static void MacsecEventIsr(t_Handle h_FmMacsec)
-+{
-+ t_FmMacsec *p_FmMacsec = (t_FmMacsec*)h_FmMacsec;
-+ uint32_t events,event,i;
-+
-+ SANITY_CHECK_RETURN(p_FmMacsec, E_INVALID_HANDLE);
-+
-+ events = GET_UINT32(p_FmMacsec->p_FmMacsecRegs->evr);
-+ events |= GET_UINT32(p_FmMacsec->p_FmMacsecRegs->ever);
-+ WRITE_UINT32(p_FmMacsec->p_FmMacsecRegs->evr,events);
-+
-+ for (i=0; i<NUM_OF_TX_SC; i++)
-+ if (events & FM_MACSEC_EV_TX_SC_NEXT_PN(i))
-+ {
-+ GET_MACSEC_MODULE_EVENT(e_FM_MACSEC_MOD_SC_TX, i, e_FM_INTR_TYPE_NORMAL, event);
-+ p_FmMacsec->intrMng[event].f_Isr(p_FmMacsec->intrMng[event].h_SrcHandle, i);
-+ }
-+}
-+
-+static void MacsecErrorIsr(t_Handle h_FmMacsec)
-+{
-+ t_FmMacsec *p_FmMacsec = (t_FmMacsec*)h_FmMacsec;
-+ uint32_t errors,error,i;
-+
-+ SANITY_CHECK_RETURN(p_FmMacsec, E_INVALID_HANDLE);
-+
-+ errors = GET_UINT32(p_FmMacsec->p_FmMacsecRegs->err);
-+ errors |= GET_UINT32(p_FmMacsec->p_FmMacsecRegs->erer);
-+ WRITE_UINT32(p_FmMacsec->p_FmMacsecRegs->err,errors);
-+
-+ for (i=0; i<NUM_OF_TX_SC; i++)
-+ if (errors & FM_MACSEC_EX_TX_SC(i))
-+ {
-+ GET_MACSEC_MODULE_EVENT(e_FM_MACSEC_MOD_SC_TX, i, e_FM_INTR_TYPE_ERR, error);
-+ p_FmMacsec->intrMng[error].f_Isr(p_FmMacsec->intrMng[error].h_SrcHandle, i);
-+ }
-+
-+ if (errors & FM_MACSEC_EX_ECC)
-+ {
-+ uint8_t eccType;
-+ uint32_t tmpReg;
-+
-+ tmpReg = GET_UINT32(p_FmMacsec->p_FmMacsecRegs->meec);
-+ ASSERT_COND(tmpReg & MECC_CAP);
-+ eccType = (uint8_t)((tmpReg & MECC_CET) >> MECC_CET_SHIFT);
-+
-+ if (!eccType && (p_FmMacsec->userExceptions & FM_MACSEC_USER_EX_SINGLE_BIT_ECC))
-+ p_FmMacsec->f_Exception(p_FmMacsec->h_App,e_FM_MACSEC_EX_SINGLE_BIT_ECC);
-+ else if (eccType && (p_FmMacsec->userExceptions & FM_MACSEC_USER_EX_MULTI_BIT_ECC))
-+ p_FmMacsec->f_Exception(p_FmMacsec->h_App,e_FM_MACSEC_EX_MULTI_BIT_ECC);
-+ else
-+ WRITE_UINT32(p_FmMacsec->p_FmMacsecRegs->meec,tmpReg);
-+ }
-+}
-+
-+static t_Error MacsecInit(t_Handle h_FmMacsec)
-+{
-+ t_FmMacsec *p_FmMacsec = (t_FmMacsec*)h_FmMacsec;
-+ t_FmMacsecDriverParam *p_FmMacsecDriverParam = NULL;
-+ uint32_t tmpReg,i,macId;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsec->p_FmMacsecDriverParam, E_INVALID_HANDLE);
-+
-+ CHECK_INIT_PARAMETERS(p_FmMacsec, CheckFmMacsecParameters);
-+
-+ p_FmMacsecDriverParam = p_FmMacsec->p_FmMacsecDriverParam;
-+
-+ for (i=0;i<e_FM_MACSEC_EV_DUMMY_LAST;i++)
-+ p_FmMacsec->intrMng[i].f_Isr = UnimplementedIsr;
-+
-+ tmpReg = 0;
-+ tmpReg |= (p_FmMacsecDriverParam->changedTextWithNoEncryptDeliverUncontrolled << CFG_UECT_SHIFT)|
-+ (p_FmMacsecDriverParam->onlyScbIsSetDeliverUncontrolled << CFG_ESCBT_SHIFT) |
-+ (p_FmMacsecDriverParam->unknownSciTreatMode << CFG_USFT_SHIFT) |
-+ (p_FmMacsecDriverParam->invalidTagsDeliverUncontrolled << CFG_ITT_SHIFT) |
-+ (p_FmMacsecDriverParam->encryptWithNoChangedTextDiscardUncontrolled << CFG_KFT_SHIFT) |
-+ (p_FmMacsecDriverParam->untagTreatMode << CFG_UFT_SHIFT) |
-+ (p_FmMacsecDriverParam->keysUnreadable << CFG_KSS_SHIFT) |
-+ (p_FmMacsecDriverParam->reservedSc0 << CFG_S0I_SHIFT) |
-+ (p_FmMacsecDriverParam->byPassMode << CFG_BYPN_SHIFT);
-+ WRITE_UINT32(p_FmMacsec->p_FmMacsecRegs->cfg, tmpReg);
-+
-+ tmpReg = FM_MAC_GetMaxFrameLength(p_FmMacsec->h_FmMac);
-+ /* At least Ethernet FCS (4 bytes) overhead must be subtracted from MFL.
-+ * In addition, the SCI (8 bytes) overhead might be subtracted as well. */
-+ tmpReg -= p_FmMacsecDriverParam->mflSubtract;
-+ WRITE_UINT32(p_FmMacsec->p_FmMacsecRegs->mfl, tmpReg);
-+
-+ WRITE_UINT32(p_FmMacsec->p_FmMacsecRegs->tpnet, p_FmMacsecDriverParam->pnExhThr);
-+
-+ if (!p_FmMacsec->userExceptions)
-+ p_FmMacsec->exceptions &= ~FM_MACSEC_EX_ECC;
-+ WRITE_UINT32(p_FmMacsec->p_FmMacsecRegs->erer, p_FmMacsec->exceptions);
-+
-+ p_FmMacsec->numRxScAvailable = NUM_OF_RX_SC;
-+ if (p_FmMacsecDriverParam->reservedSc0)
-+ p_FmMacsec->numRxScAvailable --;
-+ p_FmMacsec->numTxScAvailable = NUM_OF_TX_SC;
-+
-+ XX_Free(p_FmMacsecDriverParam);
-+ p_FmMacsec->p_FmMacsecDriverParam = NULL;
-+
-+ FM_MAC_GetId(p_FmMacsec->h_FmMac, &macId);
-+ FmRegisterIntr(p_FmMacsec->h_Fm,
-+ e_FM_MOD_MACSEC,
-+ (uint8_t)macId,
-+ e_FM_INTR_TYPE_NORMAL,
-+ MacsecEventIsr,
-+ p_FmMacsec);
-+
-+ FmRegisterIntr(p_FmMacsec->h_Fm,
-+ e_FM_MOD_MACSEC,
-+ 0,
-+ e_FM_INTR_TYPE_ERR,
-+ MacsecErrorIsr,
-+ p_FmMacsec);
-+
-+ return E_OK;
-+}
-+
-+static t_Error MacsecFree(t_Handle h_FmMacsec)
-+{
-+ t_FmMacsec *p_FmMacsec = (t_FmMacsec*)h_FmMacsec;
-+ uint32_t macId;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmMacsec->p_FmMacsecDriverParam, E_INVALID_HANDLE);
-+
-+ FM_MAC_GetId(p_FmMacsec->h_FmMac, &macId);
-+ FmUnregisterIntr(p_FmMacsec->h_Fm,
-+ e_FM_MOD_MACSEC,
-+ (uint8_t)macId,
-+ e_FM_INTR_TYPE_NORMAL);
-+
-+ FmUnregisterIntr(p_FmMacsec->h_Fm,
-+ e_FM_MOD_MACSEC,
-+ 0,
-+ e_FM_INTR_TYPE_ERR);
-+
-+ if (p_FmMacsec->rxScSpinLock)
-+ XX_FreeSpinlock(p_FmMacsec->rxScSpinLock);
-+ if (p_FmMacsec->txScSpinLock)
-+ XX_FreeSpinlock(p_FmMacsec->txScSpinLock);
-+
-+ XX_Free(p_FmMacsec);
-+
-+ return E_OK;
-+}
-+
-+static t_Error MacsecConfigUnknownSciFrameTreatment(t_Handle h_FmMacsec, e_FmMacsecUnknownSciFrameTreatment treatMode)
-+{
-+ t_FmMacsec *p_FmMacsec = (t_FmMacsec*)h_FmMacsec;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsec->p_FmMacsecDriverParam, E_INVALID_HANDLE);
-+
-+ p_FmMacsec->p_FmMacsecDriverParam->unknownSciTreatMode = treatMode;
-+
-+ return E_OK;
-+}
-+
-+static t_Error MacsecConfigInvalidTagsFrameTreatment(t_Handle h_FmMacsec, bool deliverUncontrolled)
-+{
-+ t_FmMacsec *p_FmMacsec = (t_FmMacsec*)h_FmMacsec;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsec->p_FmMacsecDriverParam, E_INVALID_HANDLE);
-+
-+ p_FmMacsec->p_FmMacsecDriverParam->invalidTagsDeliverUncontrolled = deliverUncontrolled;
-+
-+ return E_OK;
-+}
-+
-+static t_Error MacsecConfigChangedTextWithNoEncryptFrameTreatment(t_Handle h_FmMacsec, bool deliverUncontrolled)
-+{
-+ t_FmMacsec *p_FmMacsec = (t_FmMacsec*)h_FmMacsec;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsec->p_FmMacsecDriverParam, E_INVALID_HANDLE);
-+
-+ p_FmMacsec->p_FmMacsecDriverParam->changedTextWithNoEncryptDeliverUncontrolled = deliverUncontrolled;
-+
-+ return E_OK;
-+}
-+
-+static t_Error MacsecConfigOnlyScbIsSetFrameTreatment(t_Handle h_FmMacsec, bool deliverUncontrolled)
-+{
-+ t_FmMacsec *p_FmMacsec = (t_FmMacsec*)h_FmMacsec;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsec->p_FmMacsecDriverParam, E_INVALID_HANDLE);
-+
-+ p_FmMacsec->p_FmMacsecDriverParam->onlyScbIsSetDeliverUncontrolled = deliverUncontrolled;
-+
-+ return E_OK;
-+}
-+
-+static t_Error MacsecConfigEncryptWithNoChangedTextFrameTreatment(t_Handle h_FmMacsec, bool discardUncontrolled)
-+{
-+ t_FmMacsec *p_FmMacsec = (t_FmMacsec*)h_FmMacsec;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsec->p_FmMacsecDriverParam, E_INVALID_HANDLE);
-+
-+ p_FmMacsec->p_FmMacsecDriverParam->encryptWithNoChangedTextDiscardUncontrolled = discardUncontrolled;
-+
-+ return E_OK;
-+}
-+
-+static t_Error MacsecConfigUntagFrameTreatment(t_Handle h_FmMacsec, e_FmMacsecUntagFrameTreatment treatMode)
-+{
-+ t_FmMacsec *p_FmMacsec = (t_FmMacsec*)h_FmMacsec;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsec->p_FmMacsecDriverParam, E_INVALID_HANDLE);
-+
-+ p_FmMacsec->p_FmMacsecDriverParam->untagTreatMode = treatMode;
-+
-+ return E_OK;
-+}
-+
-+static t_Error MacsecConfigPnExhaustionThreshold(t_Handle h_FmMacsec, uint32_t pnExhThr)
-+{
-+ t_FmMacsec *p_FmMacsec = (t_FmMacsec*)h_FmMacsec;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsec->p_FmMacsecDriverParam, E_INVALID_HANDLE);
-+
-+ p_FmMacsec->p_FmMacsecDriverParam->pnExhThr = pnExhThr;
-+
-+ return E_OK;
-+}
-+
-+static t_Error MacsecConfigKeysUnreadable(t_Handle h_FmMacsec)
-+{
-+ t_FmMacsec *p_FmMacsec = (t_FmMacsec*)h_FmMacsec;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsec->p_FmMacsecDriverParam, E_INVALID_HANDLE);
-+
-+ p_FmMacsec->p_FmMacsecDriverParam->keysUnreadable = TRUE;
-+
-+ return E_OK;
-+}
-+
-+static t_Error MacsecConfigSectagWithoutSCI(t_Handle h_FmMacsec)
-+{
-+ t_FmMacsec *p_FmMacsec = (t_FmMacsec*)h_FmMacsec;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsec->p_FmMacsecDriverParam, E_INVALID_HANDLE);
-+
-+ p_FmMacsec->p_FmMacsecDriverParam->sectagOverhead -= MACSEC_SCI_SIZE;
-+ p_FmMacsec->p_FmMacsecDriverParam->mflSubtract += MACSEC_SCI_SIZE;
-+
-+ return E_OK;
-+}
-+
-+static t_Error MacsecConfigException(t_Handle h_FmMacsec, e_FmMacsecExceptions exception, bool enable)
-+{
-+ t_FmMacsec *p_FmMacsec = (t_FmMacsec*)h_FmMacsec;
-+ uint32_t bitMask = 0;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsec->p_FmMacsecDriverParam, E_INVALID_HANDLE);
-+
-+ GET_USER_EXCEPTION_FLAG(bitMask, exception);
-+ if (bitMask)
-+ {
-+ if (enable)
-+ p_FmMacsec->userExceptions |= bitMask;
-+ else
-+ p_FmMacsec->userExceptions &= ~bitMask;
-+ }
-+ else
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("Undefined exception"));
-+
-+ return E_OK;
-+}
-+
-+static t_Error MacsecGetRevision(t_Handle h_FmMacsec, uint32_t *p_MacsecRevision)
-+{
-+ t_FmMacsec *p_FmMacsec = (t_FmMacsec*)h_FmMacsec;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmMacsec->p_FmMacsecDriverParam, E_INVALID_HANDLE);
-+
-+ *p_MacsecRevision = GET_UINT32(p_FmMacsec->p_FmMacsecRegs->ip_rev1);
-+
-+ return E_OK;
-+}
-+
-+static t_Error MacsecEnable(t_Handle h_FmMacsec)
-+{
-+ t_FmMacsec *p_FmMacsec = (t_FmMacsec*)h_FmMacsec;
-+ uint32_t tmpReg;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmMacsec->p_FmMacsecDriverParam, E_INVALID_HANDLE);
-+
-+ tmpReg = GET_UINT32(p_FmMacsec->p_FmMacsecRegs->cfg);
-+ tmpReg |= CFG_BYPN;
-+ WRITE_UINT32(p_FmMacsec->p_FmMacsecRegs->cfg,tmpReg);
-+
-+ return E_OK;
-+}
-+
-+static t_Error MacsecDisable(t_Handle h_FmMacsec)
-+{
-+ t_FmMacsec *p_FmMacsec = (t_FmMacsec*)h_FmMacsec;
-+ uint32_t tmpReg;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmMacsec->p_FmMacsecDriverParam, E_INVALID_HANDLE);
-+
-+ tmpReg = GET_UINT32(p_FmMacsec->p_FmMacsecRegs->cfg);
-+ tmpReg &= ~CFG_BYPN;
-+ WRITE_UINT32(p_FmMacsec->p_FmMacsecRegs->cfg,tmpReg);
-+
-+ return E_OK;
-+}
-+
-+static t_Error MacsecSetException(t_Handle h_FmMacsec, e_FmMacsecExceptions exception, bool enable)
-+{
-+ t_FmMacsec *p_FmMacsec = (t_FmMacsec*)h_FmMacsec;
-+ uint32_t bitMask;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmMacsec->p_FmMacsecDriverParam, E_INVALID_HANDLE);
-+
-+ GET_USER_EXCEPTION_FLAG(bitMask, exception);
-+ if (bitMask)
-+ {
-+ if (enable)
-+ p_FmMacsec->userExceptions |= bitMask;
-+ else
-+ p_FmMacsec->userExceptions &= ~bitMask;
-+ }
-+ else
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("Undefined exception"));
-+
-+ if (!p_FmMacsec->userExceptions)
-+ p_FmMacsec->exceptions &= ~FM_MACSEC_EX_ECC;
-+ else
-+ p_FmMacsec->exceptions |= FM_MACSEC_EX_ECC;
-+ WRITE_UINT32(p_FmMacsec->p_FmMacsecRegs->erer, p_FmMacsec->exceptions);
-+
-+ return E_OK;
-+}
-+
-+static void InitFmMacsecControllerDriver(t_FmMacsecControllerDriver *p_FmMacsecControllerDriver)
-+{
-+ p_FmMacsecControllerDriver->f_FM_MACSEC_Init = MacsecInit;
-+ p_FmMacsecControllerDriver->f_FM_MACSEC_Free = MacsecFree;
-+ p_FmMacsecControllerDriver->f_FM_MACSEC_ConfigUnknownSciFrameTreatment = MacsecConfigUnknownSciFrameTreatment;
-+ p_FmMacsecControllerDriver->f_FM_MACSEC_ConfigInvalidTagsFrameTreatment = MacsecConfigInvalidTagsFrameTreatment;
-+ p_FmMacsecControllerDriver->f_FM_MACSEC_ConfigEncryptWithNoChangedTextFrameTreatment = MacsecConfigEncryptWithNoChangedTextFrameTreatment;
-+ p_FmMacsecControllerDriver->f_FM_MACSEC_ConfigUntagFrameTreatment = MacsecConfigUntagFrameTreatment;
-+ p_FmMacsecControllerDriver->f_FM_MACSEC_ConfigChangedTextWithNoEncryptFrameTreatment = MacsecConfigChangedTextWithNoEncryptFrameTreatment;
-+ p_FmMacsecControllerDriver->f_FM_MACSEC_ConfigOnlyScbIsSetFrameTreatment = MacsecConfigOnlyScbIsSetFrameTreatment;
-+ p_FmMacsecControllerDriver->f_FM_MACSEC_ConfigPnExhaustionThreshold = MacsecConfigPnExhaustionThreshold;
-+ p_FmMacsecControllerDriver->f_FM_MACSEC_ConfigKeysUnreadable = MacsecConfigKeysUnreadable;
-+ p_FmMacsecControllerDriver->f_FM_MACSEC_ConfigSectagWithoutSCI = MacsecConfigSectagWithoutSCI;
-+ p_FmMacsecControllerDriver->f_FM_MACSEC_ConfigException = MacsecConfigException;
-+ p_FmMacsecControllerDriver->f_FM_MACSEC_GetRevision = MacsecGetRevision;
-+ p_FmMacsecControllerDriver->f_FM_MACSEC_Enable = MacsecEnable;
-+ p_FmMacsecControllerDriver->f_FM_MACSEC_Disable = MacsecDisable;
-+ p_FmMacsecControllerDriver->f_FM_MACSEC_SetException = MacsecSetException;
-+}
-+
-+/****************************************/
-+/* Inter-Module functions */
-+/****************************************/
-+
-+void FmMacsecRegisterIntr(t_Handle h_FmMacsec,
-+ e_FmMacsecEventModules module,
-+ uint8_t modId,
-+ e_FmIntrType intrType,
-+ void (*f_Isr) (t_Handle h_Arg, uint32_t id),
-+ t_Handle h_Arg)
-+{
-+ t_FmMacsec *p_FmMacsec = (t_FmMacsec*)h_FmMacsec;
-+ uint8_t event= 0;
-+
-+ SANITY_CHECK_RETURN(p_FmMacsec, E_INVALID_HANDLE);
-+
-+ GET_MACSEC_MODULE_EVENT(module, modId, intrType, event);
-+
-+ ASSERT_COND(event != e_FM_MACSEC_EV_DUMMY_LAST);
-+ p_FmMacsec->intrMng[event].f_Isr = f_Isr;
-+ p_FmMacsec->intrMng[event].h_SrcHandle = h_Arg;
-+}
-+
-+void FmMacsecUnregisterIntr(t_Handle h_FmMacsec,
-+ e_FmMacsecEventModules module,
-+ uint8_t modId,
-+ e_FmIntrType intrType)
-+{
-+ t_FmMacsec *p_FmMacsec = (t_FmMacsec*)h_FmMacsec;
-+ uint8_t event= 0;
-+
-+ SANITY_CHECK_RETURN(p_FmMacsec, E_INVALID_HANDLE);
-+
-+ GET_MACSEC_MODULE_EVENT(module, modId,intrType, event);
-+
-+ ASSERT_COND(event != e_FM_MACSEC_EV_DUMMY_LAST);
-+ p_FmMacsec->intrMng[event].f_Isr = NULL;
-+ p_FmMacsec->intrMng[event].h_SrcHandle = NULL;
-+}
-+
-+t_Error FmMacsecAllocScs(t_Handle h_FmMacsec, e_ScType type, bool isPtp, uint32_t numOfScs, uint32_t *p_ScIds)
-+{
-+ t_FmMacsec *p_FmMacsec = (t_FmMacsec*)h_FmMacsec;
-+ t_Error err = E_OK;
-+ bool *p_ScTable;
-+ uint32_t *p_ScAvailable,i;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_ScIds, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(numOfScs, E_INVALID_HANDLE);
-+
-+ if (type == e_SC_RX)
-+ {
-+ p_ScTable = (bool *)p_FmMacsec->rxScTable;
-+ p_ScAvailable = &p_FmMacsec->numRxScAvailable;
-+ i = (NUM_OF_RX_SC - 1);
-+ }
-+ else
-+ {
-+ p_ScTable = (bool *)p_FmMacsec->txScTable;
-+ p_ScAvailable = &p_FmMacsec->numTxScAvailable;
-+ i = (NUM_OF_TX_SC - 1);
-+
-+ }
-+ if (*p_ScAvailable < numOfScs)
-+ RETURN_ERROR(MINOR, E_NOT_AVAILABLE, ("Not enough SCs available"));
-+
-+ if (isPtp)
-+ {
-+ i = 0;
-+ if (p_ScTable[i])
-+ RETURN_ERROR(MINOR, E_NOT_AVAILABLE, ("Sc 0 Not available"));
-+ }
-+
-+ for (;numOfScs;i--)
-+ {
-+ if (p_ScTable[i])
-+ continue;
-+ numOfScs --;
-+ (*p_ScAvailable)--;
-+ p_ScIds[numOfScs] = i;
-+ p_ScTable[i] = TRUE;
-+ }
-+
-+ return err;
-+}
-+
-+t_Error FmMacsecFreeScs(t_Handle h_FmMacsec, e_ScType type, uint32_t numOfScs, uint32_t *p_ScIds)
-+{
-+ t_FmMacsec *p_FmMacsec = (t_FmMacsec*)h_FmMacsec;
-+ t_Error err = E_OK;
-+ bool *p_ScTable;
-+ uint32_t *p_ScAvailable,maxNumOfSc,i;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_ScIds, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(numOfScs, E_INVALID_HANDLE);
-+
-+ if (type == e_SC_RX)
-+ {
-+ p_ScTable = (bool *)p_FmMacsec->rxScTable;
-+ p_ScAvailable = &p_FmMacsec->numRxScAvailable;
-+ maxNumOfSc = NUM_OF_RX_SC;
-+ }
-+ else
-+ {
-+ p_ScTable = (bool *)p_FmMacsec->txScTable;
-+ p_ScAvailable = &p_FmMacsec->numTxScAvailable;
-+ maxNumOfSc = NUM_OF_TX_SC;
-+ }
-+
-+ if ((*p_ScAvailable + numOfScs) > maxNumOfSc)
-+ RETURN_ERROR(MINOR, E_FULL, ("Too much SCs"));
-+
-+ for (i=0;i<numOfScs;i++)
-+ {
-+ p_ScTable[p_ScIds[i]] = FALSE;
-+ (*p_ScAvailable)++;
-+ }
-+
-+ return err;
-+
-+}
-+
-+t_Error FmMacsecSetPTP(t_Handle h_FmMacsec, bool enable)
-+{
-+ t_FmMacsec *p_FmMacsec = (t_FmMacsec*)h_FmMacsec;
-+ uint32_t tmpReg = 0;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsec, E_INVALID_HANDLE);
-+
-+ tmpReg = GET_UINT32(p_FmMacsec->p_FmMacsecRegs->cfg);
-+ if (enable && (tmpReg & CFG_S0I))
-+ RETURN_ERROR(MINOR, E_INVALID_STATE, ("MACSEC already in point-to-point mode"));
-+
-+ if (enable)
-+ tmpReg |= CFG_S0I;
-+ else
-+ tmpReg &= ~CFG_S0I;
-+ WRITE_UINT32(p_FmMacsec->p_FmMacsecRegs->cfg, tmpReg);
-+
-+ return E_OK;
-+}
-+
-+t_Error FmMacsecCreateRxSc(t_Handle h_FmMacsec, t_RxScParams *p_RxScParams)
-+{
-+ t_FmMacsec *p_FmMacsec = (t_FmMacsec*)h_FmMacsec;
-+ t_Error err = E_OK;
-+ uint32_t tmpReg = 0, intFlags;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_RxScParams, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_RxScParams->scId < NUM_OF_RX_SC, E_INVALID_HANDLE);
-+
-+ intFlags = XX_LockIntrSpinlock(p_FmMacsec->rxScSpinLock);
-+
-+ WRITE_UINT32(p_FmMacsec->p_FmMacsecRegs->rxsca, p_RxScParams->scId);
-+ tmpReg = GET_UINT32(p_FmMacsec->p_FmMacsecRegs->rxsccfg);
-+ if (tmpReg & RX_SCCFG_SCI_EN_MASK)
-+ {
-+ XX_UnlockIntrSpinlock(p_FmMacsec->rxScSpinLock, intFlags);
-+ RETURN_ERROR(MINOR, E_INVALID_STATE, ("Rx Sc %d must be disable",p_RxScParams->scId));
-+ }
-+
-+ WRITE_UINT32(p_FmMacsec->p_FmMacsecRegs->rxsci1h, GET_SCI_FIRST_HALF(p_RxScParams->sci));
-+ WRITE_UINT32(p_FmMacsec->p_FmMacsecRegs->rxsci2h, GET_SCI_SECOND_HALF(p_RxScParams->sci));
-+ tmpReg |= ((p_RxScParams->replayProtect << RX_SCCFG_RP_SHIFT) & RX_SCCFG_RP_MASK);
-+ tmpReg |= ((p_RxScParams->validateFrames << RX_SCCFG_VF_SHIFT) & RX_SCCFG_VF_MASK);
-+ tmpReg |= ((p_RxScParams->confidentialityOffset << RX_SCCFG_CO_SHIFT) & RX_SCCFG_CO_MASK);
-+ tmpReg |= RX_SCCFG_SCI_EN_MASK;
-+ tmpReg |= (p_RxScParams->cipherSuite << RX_SCCFG_CS_SHIFT);
-+ WRITE_UINT32(p_FmMacsec->p_FmMacsecRegs->rxsccfg, tmpReg);
-+
-+ WRITE_UINT32(p_FmMacsec->p_FmMacsecRegs->rpw, p_RxScParams->replayWindow);
-+
-+ XX_UnlockIntrSpinlock(p_FmMacsec->rxScSpinLock, intFlags);
-+
-+ return err;
-+}
-+
-+t_Error FmMacsecDeleteRxSc(t_Handle h_FmMacsec, uint32_t scId)
-+{
-+ t_FmMacsec *p_FmMacsec = (t_FmMacsec*)h_FmMacsec;
-+ t_Error err = E_OK;
-+ uint32_t tmpReg = 0, intFlags;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(scId < NUM_OF_RX_SC, E_INVALID_HANDLE);
-+
-+ intFlags = XX_LockIntrSpinlock(p_FmMacsec->rxScSpinLock);
-+
-+ tmpReg &= ~RX_SCCFG_SCI_EN_MASK;
-+ WRITE_UINT32(p_FmMacsec->p_FmMacsecRegs->rxsca, scId);
-+ WRITE_UINT32(p_FmMacsec->p_FmMacsecRegs->rxsccfg, tmpReg);
-+
-+ XX_UnlockIntrSpinlock(p_FmMacsec->rxScSpinLock, intFlags);
-+
-+ return err;
-+}
-+
-+t_Error FmMacsecCreateTxSc(t_Handle h_FmMacsec, t_TxScParams *p_TxScParams)
-+{
-+ t_FmMacsec *p_FmMacsec = (t_FmMacsec*)h_FmMacsec;
-+ t_Error err = E_OK;
-+ uint32_t tmpReg = 0, intFlags;
-+ bool alwaysIncludeSCI = FALSE, useES = FALSE, useSCB = FALSE;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_TxScParams, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_TxScParams->scId < NUM_OF_TX_SC, E_INVALID_HANDLE);
-+
-+ intFlags = XX_LockIntrSpinlock(p_FmMacsec->txScSpinLock);
-+
-+ WRITE_UINT32(p_FmMacsec->p_FmMacsecRegs->txsca, p_TxScParams->scId);
-+
-+ tmpReg = GET_UINT32(p_FmMacsec->p_FmMacsecRegs->txsccfg);
-+ if (tmpReg & TX_SCCFG_SCE_MASK)
-+ {
-+ XX_UnlockIntrSpinlock(p_FmMacsec->txScSpinLock, intFlags);
-+ RETURN_ERROR(MINOR, E_INVALID_STATE, ("Tx Sc %d must be disable",p_TxScParams->scId));
-+ }
-+
-+ WRITE_UINT32(p_FmMacsec->p_FmMacsecRegs->txsci1h, GET_SCI_FIRST_HALF(p_TxScParams->sci));
-+ WRITE_UINT32(p_FmMacsec->p_FmMacsecRegs->txsci2h, GET_SCI_SECOND_HALF(p_TxScParams->sci));
-+ alwaysIncludeSCI = (p_TxScParams->sciInsertionMode == e_FM_MACSEC_SCI_INSERTION_MODE_EXPLICIT_SECTAG);
-+ useES = (p_TxScParams->sciInsertionMode == e_FM_MACSEC_SCI_INSERTION_MODE_EXPLICIT_MAC_SA);
-+
-+ tmpReg |= ((p_TxScParams->protectFrames << TX_SCCFG_PF_SHIFT) & TX_SCCFG_PF_MASK);
-+ tmpReg |= ((alwaysIncludeSCI << TX_SCCFG_AIS_SHIFT) & TX_SCCFG_AIS_MASK);
-+ tmpReg |= ((useES << TX_SCCFG_UES_SHIFT) & TX_SCCFG_UES_MASK);
-+ tmpReg |= ((useSCB << TX_SCCFG_USCB_SHIFT) & TX_SCCFG_USCB_MASK);
-+ tmpReg |= ((p_TxScParams->confidentialityEnable << TX_SCCFG_CE_SHIFT) & TX_SCCFG_CE_MASK);
-+ tmpReg |= ((p_TxScParams->confidentialityOffset << TX_SCCFG_CO_SHIFT) & TX_SCCFG_CO_MASK);
-+ tmpReg |= TX_SCCFG_SCE_MASK;
-+ tmpReg |= (p_TxScParams->cipherSuite << TX_SCCFG_CS_SHIFT);
-+ WRITE_UINT32(p_FmMacsec->p_FmMacsecRegs->txsccfg, tmpReg);
-+
-+ XX_UnlockIntrSpinlock(p_FmMacsec->txScSpinLock, intFlags);
-+
-+ return err;
-+}
-+
-+t_Error FmMacsecDeleteTxSc(t_Handle h_FmMacsec, uint32_t scId)
-+{
-+ t_FmMacsec *p_FmMacsec = (t_FmMacsec*)h_FmMacsec;
-+ t_Error err = E_OK;
-+ uint32_t tmpReg = 0, intFlags;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(scId < NUM_OF_TX_SC, E_INVALID_HANDLE);
-+
-+ intFlags = XX_LockIntrSpinlock(p_FmMacsec->txScSpinLock);
-+
-+ tmpReg &= ~TX_SCCFG_SCE_MASK;
-+ WRITE_UINT32(p_FmMacsec->p_FmMacsecRegs->txsca, scId);
-+ WRITE_UINT32(p_FmMacsec->p_FmMacsecRegs->txsccfg, tmpReg);
-+
-+ XX_UnlockIntrSpinlock(p_FmMacsec->txScSpinLock, intFlags);
-+
-+ return err;
-+}
-+
-+t_Error FmMacsecCreateRxSa(t_Handle h_FmMacsec, uint32_t scId, e_ScSaId saId, macsecAN_t an, uint32_t lowestPn, macsecSAKey_t key)
-+{
-+ t_FmMacsec *p_FmMacsec = (t_FmMacsec*)h_FmMacsec;
-+ t_Error err = E_OK;
-+ uint32_t tmpReg = 0, intFlags;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(scId < NUM_OF_RX_SC, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(saId < NUM_OF_SA_PER_RX_SC, E_INVALID_HANDLE);
-+
-+ intFlags = XX_LockIntrSpinlock(p_FmMacsec->rxScSpinLock);
-+
-+ WRITE_UINT32(p_FmMacsec->p_FmMacsecRegs->rxsca, scId);
-+ WRITE_UINT32(p_FmMacsec->p_FmMacsecRegs->fmMacsecRxScSa[saId].rxsanpn, DEFAULT_initNextPn);
-+ WRITE_UINT32(p_FmMacsec->p_FmMacsecRegs->fmMacsecRxScSa[saId].rxsalpn, lowestPn);
-+ MemCpy8((void*)p_FmMacsec->p_FmMacsecRegs->fmMacsecRxScSa[saId].rxsak, key, sizeof(macsecSAKey_t));
-+
-+ tmpReg |= RX_SACFG_ACTIVE;
-+ tmpReg |= ((an << RX_SACFG_AN_SHIFT) & RX_SACFG_AN_MASK);
-+ WRITE_UINT32(p_FmMacsec->p_FmMacsecRegs->fmMacsecRxScSa[saId].rxsacs, tmpReg);
-+
-+ XX_UnlockIntrSpinlock(p_FmMacsec->rxScSpinLock, intFlags);
-+
-+ return err;
-+}
-+
-+t_Error FmMacsecCreateTxSa(t_Handle h_FmMacsec, uint32_t scId, e_ScSaId saId, macsecSAKey_t key)
-+{
-+ t_FmMacsec *p_FmMacsec = (t_FmMacsec*)h_FmMacsec;
-+ t_Error err = E_OK;
-+ uint32_t tmpReg = 0, intFlags;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(scId < NUM_OF_RX_SC, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(saId < NUM_OF_SA_PER_TX_SC, E_INVALID_HANDLE);
-+
-+ intFlags = XX_LockIntrSpinlock(p_FmMacsec->txScSpinLock);
-+
-+ WRITE_UINT32(p_FmMacsec->p_FmMacsecRegs->txsca, scId);
-+ WRITE_UINT32(p_FmMacsec->p_FmMacsecRegs->fmMacsecTxScSa[saId].txsanpn, DEFAULT_initNextPn);
-+ MemCpy8((void*)p_FmMacsec->p_FmMacsecRegs->fmMacsecTxScSa[saId].txsak, key, sizeof(macsecSAKey_t));
-+
-+ tmpReg |= TX_SACFG_ACTIVE;
-+ WRITE_UINT32(p_FmMacsec->p_FmMacsecRegs->fmMacsecTxScSa[saId].txsacs, tmpReg);
-+
-+ XX_UnlockIntrSpinlock(p_FmMacsec->txScSpinLock, intFlags);
-+
-+ return err;
-+}
-+
-+t_Error FmMacsecDeleteRxSa(t_Handle h_FmMacsec, uint32_t scId, e_ScSaId saId)
-+{
-+ t_FmMacsec *p_FmMacsec = (t_FmMacsec*)h_FmMacsec;
-+ t_Error err = E_OK;
-+ uint32_t tmpReg = 0, i, intFlags;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(scId < NUM_OF_RX_SC, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(saId < NUM_OF_SA_PER_RX_SC, E_INVALID_HANDLE);
-+
-+ intFlags = XX_LockIntrSpinlock(p_FmMacsec->rxScSpinLock);
-+
-+ WRITE_UINT32(p_FmMacsec->p_FmMacsecRegs->rxsca, scId);
-+ WRITE_UINT32(p_FmMacsec->p_FmMacsecRegs->fmMacsecRxScSa[saId].rxsanpn, 0x0);
-+ WRITE_UINT32(p_FmMacsec->p_FmMacsecRegs->fmMacsecRxScSa[saId].rxsalpn, 0x0);
-+ for (i=0; i<4; i++)
-+ WRITE_UINT32(p_FmMacsec->p_FmMacsecRegs->fmMacsecRxScSa[saId].rxsak[i], 0x0);
-+
-+ tmpReg |= RX_SACFG_ACTIVE;
-+ tmpReg &= ~RX_SACFG_EN_MASK;
-+ WRITE_UINT32(p_FmMacsec->p_FmMacsecRegs->fmMacsecRxScSa[saId].rxsacs, tmpReg);
-+
-+ XX_UnlockIntrSpinlock(p_FmMacsec->rxScSpinLock, intFlags);
-+
-+ return err;
-+}
-+
-+t_Error FmMacsecDeleteTxSa(t_Handle h_FmMacsec, uint32_t scId, e_ScSaId saId)
-+{
-+ t_FmMacsec *p_FmMacsec = (t_FmMacsec*)h_FmMacsec;
-+ t_Error err = E_OK;
-+ uint32_t tmpReg = 0, i, intFlags;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(scId < NUM_OF_RX_SC, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(saId < NUM_OF_SA_PER_TX_SC, E_INVALID_HANDLE);
-+
-+ intFlags = XX_LockIntrSpinlock(p_FmMacsec->txScSpinLock);
-+
-+ WRITE_UINT32(p_FmMacsec->p_FmMacsecRegs->txsca, scId);
-+ WRITE_UINT32(p_FmMacsec->p_FmMacsecRegs->fmMacsecTxScSa[saId].txsanpn, 0x0);
-+ for (i=0; i<4; i++)
-+ WRITE_UINT32(p_FmMacsec->p_FmMacsecRegs->fmMacsecTxScSa[saId].txsak[i], 0x0);
-+
-+ tmpReg |= TX_SACFG_ACTIVE;
-+ WRITE_UINT32(p_FmMacsec->p_FmMacsecRegs->fmMacsecTxScSa[saId].txsacs, tmpReg);
-+
-+ XX_UnlockIntrSpinlock(p_FmMacsec->txScSpinLock, intFlags);
-+
-+ return err;
-+}
-+
-+t_Error FmMacsecRxSaSetReceive(t_Handle h_FmMacsec, uint32_t scId, e_ScSaId saId, bool enableReceive)
-+{
-+ t_FmMacsec *p_FmMacsec = (t_FmMacsec*)h_FmMacsec;
-+ t_Error err = E_OK;
-+ uint32_t tmpReg = 0, intFlags;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(scId < NUM_OF_RX_SC, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(saId < NUM_OF_SA_PER_RX_SC, E_INVALID_HANDLE);
-+
-+ intFlags = XX_LockIntrSpinlock(p_FmMacsec->rxScSpinLock);
-+
-+ WRITE_UINT32(p_FmMacsec->p_FmMacsecRegs->rxsca, scId);
-+ tmpReg = GET_UINT32(p_FmMacsec->p_FmMacsecRegs->fmMacsecRxScSa[saId].rxsacs);
-+ if (enableReceive)
-+ tmpReg |= RX_SACFG_EN_MASK;
-+ else
-+ tmpReg &= ~RX_SACFG_EN_MASK;
-+
-+ WRITE_UINT32(p_FmMacsec->p_FmMacsecRegs->fmMacsecRxScSa[saId].rxsacs, tmpReg);
-+
-+ XX_UnlockIntrSpinlock(p_FmMacsec->rxScSpinLock, intFlags);
-+
-+ return err;
-+}
-+
-+t_Error FmMacsecRxSaUpdateNextPn(t_Handle h_FmMacsec, uint32_t scId, e_ScSaId saId, uint32_t updtNextPN)
-+{
-+ t_FmMacsec *p_FmMacsec = (t_FmMacsec*)h_FmMacsec;
-+ t_Error err = E_OK;
-+ uint32_t intFlags;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(scId < NUM_OF_RX_SC, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(saId < NUM_OF_SA_PER_RX_SC, E_INVALID_HANDLE);
-+
-+ intFlags = XX_LockIntrSpinlock(p_FmMacsec->rxScSpinLock);
-+
-+ WRITE_UINT32(p_FmMacsec->p_FmMacsecRegs->rxsca, scId);
-+ WRITE_UINT32(p_FmMacsec->p_FmMacsecRegs->fmMacsecRxScSa[saId].rxsanpn, updtNextPN);
-+
-+ XX_UnlockIntrSpinlock(p_FmMacsec->rxScSpinLock, intFlags);
-+
-+ return err;
-+}
-+
-+t_Error FmMacsecRxSaUpdateLowestPn(t_Handle h_FmMacsec, uint32_t scId, e_ScSaId saId, uint32_t updtLowestPN)
-+{
-+ t_FmMacsec *p_FmMacsec = (t_FmMacsec*)h_FmMacsec;
-+ t_Error err = E_OK;
-+ uint32_t intFlags;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(scId < NUM_OF_RX_SC, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(saId < NUM_OF_SA_PER_RX_SC, E_INVALID_HANDLE);
-+
-+ intFlags = XX_LockIntrSpinlock(p_FmMacsec->rxScSpinLock);
-+
-+ WRITE_UINT32(p_FmMacsec->p_FmMacsecRegs->rxsca, scId);
-+ WRITE_UINT32(p_FmMacsec->p_FmMacsecRegs->fmMacsecRxScSa[saId].rxsalpn, updtLowestPN);
-+
-+ XX_UnlockIntrSpinlock(p_FmMacsec->rxScSpinLock, intFlags);
-+
-+ return err;
-+}
-+
-+t_Error FmMacsecTxSaSetActive(t_Handle h_FmMacsec, uint32_t scId, e_ScSaId saId, macsecAN_t an)
-+{
-+ t_FmMacsec *p_FmMacsec = (t_FmMacsec*)h_FmMacsec;
-+ t_Error err = E_OK;
-+ uint32_t tmpReg = 0, intFlags;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(scId < NUM_OF_RX_SC, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(saId < NUM_OF_SA_PER_TX_SC, E_INVALID_HANDLE);
-+
-+ intFlags = XX_LockIntrSpinlock(p_FmMacsec->txScSpinLock);
-+
-+ WRITE_UINT32(p_FmMacsec->p_FmMacsecRegs->txsca, scId);
-+
-+ tmpReg = GET_UINT32(p_FmMacsec->p_FmMacsecRegs->txsccfg);
-+
-+ tmpReg |= ((an << TX_SCCFG_AN_SHIFT) & TX_SCCFG_AN_MASK);
-+ tmpReg |= ((saId << TX_SCCFG_ASA_SHIFT) & TX_SCCFG_ASA_MASK);
-+
-+ WRITE_UINT32(p_FmMacsec->p_FmMacsecRegs->txsccfg, tmpReg);
-+
-+ XX_UnlockIntrSpinlock(p_FmMacsec->txScSpinLock, intFlags);
-+
-+ return err;
-+}
-+
-+t_Error FmMacsecTxSaGetActive(t_Handle h_FmMacsec, uint32_t scId, macsecAN_t *p_An)
-+{
-+ t_FmMacsec *p_FmMacsec = (t_FmMacsec*)h_FmMacsec;
-+ t_Error err = E_OK;
-+ uint32_t tmpReg = 0, intFlags;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(scId < NUM_OF_RX_SC, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_An, E_INVALID_HANDLE);
-+
-+ intFlags = XX_LockIntrSpinlock(p_FmMacsec->txScSpinLock);
-+
-+ WRITE_UINT32(p_FmMacsec->p_FmMacsecRegs->txsca, scId);
-+
-+ tmpReg = GET_UINT32(p_FmMacsec->p_FmMacsecRegs->txsccfg);
-+
-+ XX_UnlockIntrSpinlock(p_FmMacsec->txScSpinLock, intFlags);
-+
-+ *p_An = (macsecAN_t)((tmpReg & TX_SCCFG_AN_MASK) >> TX_SCCFG_AN_SHIFT);
-+
-+ return err;
-+}
-+
-+t_Error FmMacsecSetException(t_Handle h_FmMacsec, e_FmMacsecGlobalExceptions exception, uint32_t scId, bool enable)
-+{
-+ t_FmMacsec *p_FmMacsec = (t_FmMacsec*)h_FmMacsec;
-+ uint32_t bitMask;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmMacsec->p_FmMacsecDriverParam, E_INVALID_HANDLE);
-+
-+ GET_EXCEPTION_FLAG(bitMask, exception, scId);
-+ if (bitMask)
-+ {
-+ if (enable)
-+ p_FmMacsec->exceptions |= bitMask;
-+ else
-+ p_FmMacsec->exceptions &= ~bitMask;
-+ }
-+ else
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("Undefined exception"));
-+
-+ WRITE_UINT32(p_FmMacsec->p_FmMacsecRegs->erer, p_FmMacsec->exceptions);
-+
-+ return E_OK;
-+}
-+
-+t_Error FmMacsecSetEvent(t_Handle h_FmMacsec, e_FmMacsecGlobalEvents event, uint32_t scId, bool enable)
-+{
-+ t_FmMacsec *p_FmMacsec = (t_FmMacsec*)h_FmMacsec;
-+ uint32_t bitMask;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmMacsec->p_FmMacsecDriverParam, E_INVALID_HANDLE);
-+
-+ GET_EVENT_FLAG(bitMask, event, scId);
-+ if (bitMask)
-+ {
-+ if (enable)
-+ p_FmMacsec->events |= bitMask;
-+ else
-+ p_FmMacsec->events &= ~bitMask;
-+ }
-+ else
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("Undefined event"));
-+
-+ WRITE_UINT32(p_FmMacsec->p_FmMacsecRegs->ever, p_FmMacsec->events);
-+
-+ return E_OK;
-+}
-+
-+/****************************************/
-+/* API Init unit functions */
-+/****************************************/
-+t_Handle FM_MACSEC_MASTER_Config(t_FmMacsecParams *p_FmMacsecParam)
-+{
-+ t_FmMacsec *p_FmMacsec;
-+ uint32_t macId;
-+
-+ /* Allocate FM MACSEC structure */
-+ p_FmMacsec = (t_FmMacsec *) XX_Malloc(sizeof(t_FmMacsec));
-+ if (!p_FmMacsec)
-+ {
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("FM MACSEC driver structure"));
-+ return NULL;
-+ }
-+ memset(p_FmMacsec, 0, sizeof(t_FmMacsec));
-+ InitFmMacsecControllerDriver(&p_FmMacsec->fmMacsecControllerDriver);
-+
-+ /* Allocate the FM MACSEC driver's parameters structure */
-+ p_FmMacsec->p_FmMacsecDriverParam = (t_FmMacsecDriverParam *)XX_Malloc(sizeof(t_FmMacsecDriverParam));
-+ if (!p_FmMacsec->p_FmMacsecDriverParam)
-+ {
-+ XX_Free(p_FmMacsec);
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("FM MACSEC driver parameters"));
-+ return NULL;
-+ }
-+ memset(p_FmMacsec->p_FmMacsecDriverParam, 0, sizeof(t_FmMacsecDriverParam));
-+
-+ /* Initialize FM MACSEC parameters which will be kept by the driver */
-+ p_FmMacsec->h_Fm = p_FmMacsecParam->h_Fm;
-+ p_FmMacsec->h_FmMac = p_FmMacsecParam->nonGuestParams.h_FmMac;
-+ p_FmMacsec->p_FmMacsecRegs = (t_FmMacsecRegs *)UINT_TO_PTR(p_FmMacsecParam->nonGuestParams.baseAddr);
-+ p_FmMacsec->f_Exception = p_FmMacsecParam->nonGuestParams.f_Exception;
-+ p_FmMacsec->h_App = p_FmMacsecParam->nonGuestParams.h_App;
-+ p_FmMacsec->userExceptions = DEFAULT_userExceptions;
-+ p_FmMacsec->exceptions = DEFAULT_exceptions;
-+ p_FmMacsec->events = DEFAULT_events;
-+ p_FmMacsec->rxScSpinLock = XX_InitSpinlock();
-+ p_FmMacsec->txScSpinLock = XX_InitSpinlock();
-+
-+ /* Initialize FM MACSEC driver parameters parameters (for initialization phase only) */
-+ p_FmMacsec->p_FmMacsecDriverParam->unknownSciTreatMode = DEFAULT_unknownSciFrameTreatment;
-+ p_FmMacsec->p_FmMacsecDriverParam->invalidTagsDeliverUncontrolled = DEFAULT_invalidTagsFrameTreatment;
-+ p_FmMacsec->p_FmMacsecDriverParam->encryptWithNoChangedTextDiscardUncontrolled = DEFAULT_encryptWithNoChangedTextFrameTreatment;
-+ p_FmMacsec->p_FmMacsecDriverParam->untagTreatMode = DEFAULT_untagFrameTreatment;
-+ p_FmMacsec->p_FmMacsecDriverParam->keysUnreadable = DEFAULT_keysUnreadable;
-+ p_FmMacsec->p_FmMacsecDriverParam->reservedSc0 = DEFAULT_sc0ReservedForPTP;
-+ p_FmMacsec->p_FmMacsecDriverParam->byPassMode = !DEFAULT_normalMode;
-+ p_FmMacsec->p_FmMacsecDriverParam->pnExhThr = DEFAULT_pnExhThr;
-+ p_FmMacsec->p_FmMacsecDriverParam->sectagOverhead = DEFAULT_sectagOverhead;
-+ p_FmMacsec->p_FmMacsecDriverParam->mflSubtract = DEFAULT_mflSubtract;
-+ /* build the FM MACSEC master IPC address */
-+ memset(p_FmMacsec->fmMacsecModuleName, 0, (sizeof(char))*MODULE_NAME_SIZE);
-+ FM_MAC_GetId(p_FmMacsec->h_FmMac,&macId);
-+ if (Sprint (p_FmMacsec->fmMacsecModuleName, "FM-%d-MAC-%d-MACSEC-Master",
-+ FmGetId(p_FmMacsec->h_Fm),macId) != 24)
-+ {
-+ XX_Free(p_FmMacsec->p_FmMacsecDriverParam);
-+ XX_Free(p_FmMacsec);
-+ REPORT_ERROR(MAJOR, E_INVALID_STATE, ("Sprint failed"));
-+ return NULL;
-+ }
-+ return p_FmMacsec;
-+}
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MACSEC/fm_macsec_master.h
-@@ -0,0 +1,479 @@
-+/*
-+ * Copyright 2008-2015 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+/******************************************************************************
-+ @File fm_macsec_master.h
-+
-+ @Description FM MACSEC internal structures and definitions.
-+*//***************************************************************************/
-+#ifndef __FM_MACSEC_MASTER_H
-+#define __FM_MACSEC_MASTER_H
-+
-+#include "error_ext.h"
-+#include "std_ext.h"
-+
-+#include "fm_macsec.h"
-+
-+
-+#define MACSEC_ICV_SIZE 16
-+#define MACSEC_SECTAG_SIZE 16
-+#define MACSEC_SCI_SIZE 8
-+#define MACSEC_FCS_SIZE 4
-+
-+/**************************************************************************//**
-+ @Description Exceptions
-+*//***************************************************************************/
-+
-+#define FM_MACSEC_EX_TX_SC_0 0x80000000
-+#define FM_MACSEC_EX_TX_SC(sc) (FM_MACSEC_EX_TX_SC_0 >> (sc))
-+#define FM_MACSEC_EX_ECC 0x00000001
-+
-+#define GET_EXCEPTION_FLAG(bitMask, exception, id) switch (exception){ \
-+ case e_FM_MACSEC_EX_TX_SC: \
-+ bitMask = FM_MACSEC_EX_TX_SC(id); break; \
-+ case e_FM_MACSEC_EX_ECC: \
-+ bitMask = FM_MACSEC_EX_ECC; break; \
-+ default: bitMask = 0;break;}
-+
-+#define FM_MACSEC_USER_EX_SINGLE_BIT_ECC 0x80000000
-+#define FM_MACSEC_USER_EX_MULTI_BIT_ECC 0x40000000
-+
-+#define GET_USER_EXCEPTION_FLAG(bitMask, exception) switch (exception){ \
-+ case e_FM_MACSEC_EX_SINGLE_BIT_ECC: \
-+ bitMask = FM_MACSEC_USER_EX_SINGLE_BIT_ECC; break; \
-+ case e_FM_MACSEC_EX_MULTI_BIT_ECC: \
-+ bitMask = FM_MACSEC_USER_EX_MULTI_BIT_ECC; break; \
-+ default: bitMask = 0;break;}
-+
-+/**************************************************************************//**
-+ @Description Events
-+*//***************************************************************************/
-+
-+#define FM_MACSEC_EV_TX_SC_0_NEXT_PN 0x80000000
-+#define FM_MACSEC_EV_TX_SC_NEXT_PN(sc) (FM_MACSEC_EV_TX_SC_0_NEXT_PN >> (sc))
-+
-+#define GET_EVENT_FLAG(bitMask, event, id) switch (event){ \
-+ case e_FM_MACSEC_EV_TX_SC_NEXT_PN: \
-+ bitMask = FM_MACSEC_EV_TX_SC_NEXT_PN(id); break; \
-+ default: bitMask = 0;break;}
-+
-+/**************************************************************************//**
-+ @Description Defaults
-+*//***************************************************************************/
-+#define DEFAULT_userExceptions (FM_MACSEC_USER_EX_SINGLE_BIT_ECC |\
-+ FM_MACSEC_USER_EX_MULTI_BIT_ECC)
-+
-+#define DEFAULT_exceptions (FM_MACSEC_EX_TX_SC(0) |\
-+ FM_MACSEC_EX_TX_SC(1) |\
-+ FM_MACSEC_EX_TX_SC(2) |\
-+ FM_MACSEC_EX_TX_SC(3) |\
-+ FM_MACSEC_EX_TX_SC(4) |\
-+ FM_MACSEC_EX_TX_SC(5) |\
-+ FM_MACSEC_EX_TX_SC(6) |\
-+ FM_MACSEC_EX_TX_SC(7) |\
-+ FM_MACSEC_EX_TX_SC(8) |\
-+ FM_MACSEC_EX_TX_SC(9) |\
-+ FM_MACSEC_EX_TX_SC(10) |\
-+ FM_MACSEC_EX_TX_SC(11) |\
-+ FM_MACSEC_EX_TX_SC(12) |\
-+ FM_MACSEC_EX_TX_SC(13) |\
-+ FM_MACSEC_EX_TX_SC(14) |\
-+ FM_MACSEC_EX_TX_SC(15) |\
-+ FM_MACSEC_EX_ECC )
-+
-+#define DEFAULT_events (FM_MACSEC_EV_TX_SC_NEXT_PN(0) |\
-+ FM_MACSEC_EV_TX_SC_NEXT_PN(1) |\
-+ FM_MACSEC_EV_TX_SC_NEXT_PN(2) |\
-+ FM_MACSEC_EV_TX_SC_NEXT_PN(3) |\
-+ FM_MACSEC_EV_TX_SC_NEXT_PN(4) |\
-+ FM_MACSEC_EV_TX_SC_NEXT_PN(5) |\
-+ FM_MACSEC_EV_TX_SC_NEXT_PN(6) |\
-+ FM_MACSEC_EV_TX_SC_NEXT_PN(7) |\
-+ FM_MACSEC_EV_TX_SC_NEXT_PN(8) |\
-+ FM_MACSEC_EV_TX_SC_NEXT_PN(9) |\
-+ FM_MACSEC_EV_TX_SC_NEXT_PN(10) |\
-+ FM_MACSEC_EV_TX_SC_NEXT_PN(11) |\
-+ FM_MACSEC_EV_TX_SC_NEXT_PN(12) |\
-+ FM_MACSEC_EV_TX_SC_NEXT_PN(13) |\
-+ FM_MACSEC_EV_TX_SC_NEXT_PN(14) |\
-+ FM_MACSEC_EV_TX_SC_NEXT_PN(15) )
-+
-+#define DEFAULT_unknownSciFrameTreatment e_FM_MACSEC_UNKNOWN_SCI_FRAME_TREATMENT_DISCARD_BOTH
-+#define DEFAULT_invalidTagsFrameTreatment FALSE
-+#define DEFAULT_encryptWithNoChangedTextFrameTreatment FALSE
-+#define DEFAULT_untagFrameTreatment e_FM_MACSEC_UNTAG_FRAME_TREATMENT_DELIVER_UNCONTROLLED_DISCARD_CONTROLLED
-+#define DEFAULT_changedTextWithNoEncryptFrameTreatment FALSE
-+#define DEFAULT_onlyScbIsSetFrameTreatment FALSE
-+#define DEFAULT_keysUnreadable FALSE
-+#define DEFAULT_normalMode TRUE
-+#define DEFAULT_sc0ReservedForPTP FALSE
-+#define DEFAULT_initNextPn 1
-+#define DEFAULT_pnExhThr 0xffffffff
-+#define DEFAULT_sectagOverhead (MACSEC_ICV_SIZE + MACSEC_SECTAG_SIZE)
-+#define DEFAULT_mflSubtract MACSEC_FCS_SIZE
-+
-+
-+/**************************************************************************//**
-+ @Description Memory Mapped Registers
-+*//***************************************************************************/
-+
-+#if defined(__MWERKS__) && !defined(__GNUC__)
-+#pragma pack(push,1)
-+#endif /* defined(__MWERKS__) && ... */
-+
-+typedef _Packed struct
-+{
-+ /* MACsec configuration */
-+ volatile uint32_t cfg; /**< MACsec configuration */
-+ volatile uint32_t et; /**< MACsec EtherType */
-+ volatile uint8_t res1[56]; /**< reserved */
-+ volatile uint32_t mfl; /**< Maximum Frame Length */
-+ volatile uint32_t tpnet; /**< TX Packet Number exhaustion threshold */
-+ volatile uint8_t res2[56]; /**< reserved */
-+ volatile uint32_t rxsca; /**< RX SC access select */
-+ volatile uint8_t res3[60]; /**< reserved */
-+ volatile uint32_t txsca; /**< TX SC access select */
-+ volatile uint8_t res4[60]; /**< reserved */
-+
-+ /* RX configuration, status and statistic */
-+ volatile uint32_t rxsci1h; /**< RX Secure Channel Identifier first half */
-+ volatile uint32_t rxsci2h; /**< RX Secure Channel Identifier second half */
-+ volatile uint8_t res5[8]; /**< reserved */
-+ volatile uint32_t ifio1hs; /**< ifInOctets first half Statistic */
-+ volatile uint32_t ifio2hs; /**< ifInOctets second half Statistic */
-+ volatile uint32_t ifiups; /**< ifInUcastPkts Statistic */
-+ volatile uint8_t res6[4]; /**< reserved */
-+ volatile uint32_t ifimps; /**< ifInMulticastPkts Statistic */
-+ volatile uint32_t ifibps; /**< ifInBroadcastPkts Statistic */
-+ volatile uint32_t rxsccfg; /**< RX Secure Channel configuration */
-+ volatile uint32_t rpw; /**< replayWindow */
-+ volatile uint8_t res7[16]; /**< reserved */
-+ volatile uint32_t inov1hs; /**< InOctetsValidated first half Statistic */
-+ volatile uint32_t inov2hs; /**< InOctetsValidated second half Statistic */
-+ volatile uint32_t inod1hs; /**< InOctetsDecrypted first half Statistic */
-+ volatile uint32_t inod2hs; /**< InOctetsDecrypted second half Statistic */
-+ volatile uint32_t rxscipus; /**< RX Secure Channel InPktsUnchecked Statistic */
-+ volatile uint32_t rxscipds; /**< RX Secure Channel InPktsDelayed Statistic */
-+ volatile uint32_t rxscipls; /**< RX Secure Channel InPktsLate Statistic */
-+ volatile uint8_t res8[4]; /**< reserved */
-+ volatile uint32_t rxaninuss[MAX_NUM_OF_SA_PER_SC]; /**< RX AN 0-3 InNotUsingSA Statistic */
-+ volatile uint32_t rxanipuss[MAX_NUM_OF_SA_PER_SC]; /**< RX AN 0-3 InPktsUnusedSA Statistic */
-+ _Packed struct
-+ {
-+ volatile uint32_t rxsacs; /**< RX Security Association configuration and status */
-+ volatile uint32_t rxsanpn; /**< RX Security Association nextPN */
-+ volatile uint32_t rxsalpn; /**< RX Security Association lowestPN */
-+ volatile uint32_t rxsaipos; /**< RX Security Association InPktsOK Statistic */
-+ volatile uint32_t rxsak[4]; /**< RX Security Association key (128 bit) */
-+ volatile uint32_t rxsah[4]; /**< RX Security Association hash (128 bit) */
-+ volatile uint32_t rxsaipis; /**< RX Security Association InPktsInvalid Statistic */
-+ volatile uint32_t rxsaipnvs; /**< RX Security Association InPktsNotValid Statistic */
-+ volatile uint8_t res9[8]; /**< reserved */
-+ } _PackedType fmMacsecRxScSa[NUM_OF_SA_PER_RX_SC];
-+
-+ /* TX configuration, status and statistic */
-+ volatile uint32_t txsci1h; /**< TX Secure Channel Identifier first half */
-+ volatile uint32_t txsci2h; /**< TX Secure Channel Identifier second half */
-+ volatile uint8_t res10[8]; /**< reserved */
-+ volatile uint32_t ifoo1hs; /**< ifOutOctets first half Statistic */
-+ volatile uint32_t ifoo2hs; /**< ifOutOctets second half Statistic */
-+ volatile uint32_t ifoups; /**< ifOutUcastPkts Statistic */
-+ volatile uint32_t opus; /**< OutPktsUntagged Statistic */
-+ volatile uint32_t ifomps; /**< ifOutMulticastPkts Statistic */
-+ volatile uint32_t ifobps; /**< ifOutBroadcastPkts Statistic */
-+ volatile uint32_t txsccfg; /**< TX Secure Channel configuration */
-+ volatile uint32_t optls; /**< OutPktsTooLong Statistic */
-+ volatile uint8_t res11[16]; /**< reserved */
-+ volatile uint32_t oop1hs; /**< OutOctetsProtected first half Statistic */
-+ volatile uint32_t oop2hs; /**< OutOctetsProtected second half Statistic */
-+ volatile uint32_t ooe1hs; /**< OutOctetsEncrypted first half Statistic */
-+ volatile uint32_t ooe2hs; /**< OutOctetsEncrypted second half Statistic */
-+ volatile uint8_t res12[48]; /**< reserved */
-+ _Packed struct
-+ {
-+ volatile uint32_t txsacs; /**< TX Security Association configuration and status */
-+ volatile uint32_t txsanpn; /**< TX Security Association nextPN */
-+ volatile uint32_t txsaopps; /**< TX Security Association OutPktsProtected Statistic */
-+ volatile uint32_t txsaopes; /**< TX Security Association OutPktsEncrypted Statistic */
-+ volatile uint32_t txsak[4]; /**< TX Security Association key (128 bit) */
-+ volatile uint32_t txsah[4]; /**< TX Security Association hash (128 bit) */
-+ volatile uint8_t res13[16]; /**< reserved */
-+ } _PackedType fmMacsecTxScSa[NUM_OF_SA_PER_TX_SC];
-+ volatile uint8_t res14[248]; /**< reserved */
-+
-+ /* Global configuration and status */
-+ volatile uint32_t ip_rev1; /**< MACsec IP Block Revision 1 register */
-+ volatile uint32_t ip_rev2; /**< MACsec IP Block Revision 2 register */
-+ volatile uint32_t evr; /**< MACsec Event Register */
-+ volatile uint32_t ever; /**< MACsec Event Enable Register */
-+ volatile uint32_t evfr; /**< MACsec Event Force Register */
-+ volatile uint32_t err; /**< MACsec Error Register */
-+ volatile uint32_t erer; /**< MACsec Error Enable Register */
-+ volatile uint32_t erfr; /**< MACsec Error Force Register */
-+ volatile uint8_t res15[40]; /**< reserved */
-+ volatile uint32_t meec; /**< MACsec Memory ECC Error Capture Register */
-+ volatile uint32_t idle; /**< MACsec Idle status Register */
-+ volatile uint8_t res16[184]; /**< reserved */
-+ /* DEBUG */
-+ volatile uint32_t rxec; /**< MACsec RX error capture Register */
-+ volatile uint8_t res17[28]; /**< reserved */
-+ volatile uint32_t txec; /**< MACsec TX error capture Register */
-+ volatile uint8_t res18[220]; /**< reserved */
-+
-+ /* Macsec Rx global statistic */
-+ volatile uint32_t ifiocp1hs; /**< ifInOctetsCp first half Statistic */
-+ volatile uint32_t ifiocp2hs; /**< ifInOctetsCp second half Statistic */
-+ volatile uint32_t ifiupcps; /**< ifInUcastPktsCp Statistic */
-+ volatile uint8_t res19[4]; /**< reserved */
-+ volatile uint32_t ifioup1hs; /**< ifInOctetsUp first half Statistic */
-+ volatile uint32_t ifioup2hs; /**< ifInOctetsUp second half Statistic */
-+ volatile uint32_t ifiupups; /**< ifInUcastPktsUp Statistic */
-+ volatile uint8_t res20[4]; /**< reserved */
-+ volatile uint32_t ifimpcps; /**< ifInMulticastPktsCp Statistic */
-+ volatile uint32_t ifibpcps; /**< ifInBroadcastPktsCp Statistic */
-+ volatile uint32_t ifimpups; /**< ifInMulticastPktsUp Statistic */
-+ volatile uint32_t ifibpups; /**< ifInBroadcastPktsUp Statistic */
-+ volatile uint32_t ipwts; /**< InPktsWithoutTag Statistic */
-+ volatile uint32_t ipkays; /**< InPktsKaY Statistic */
-+ volatile uint32_t ipbts; /**< InPktsBadTag Statistic */
-+ volatile uint32_t ipsnfs; /**< InPktsSCINotFound Statistic */
-+ volatile uint32_t ipuecs; /**< InPktsUnsupportedEC Statistic */
-+ volatile uint32_t ipescbs; /**< InPktsEponSingleCopyBroadcast Statistic */
-+ volatile uint32_t iptls; /**< InPktsTooLong Statistic */
-+ volatile uint8_t res21[52]; /**< reserved */
-+
-+ /* Macsec Tx global statistic */
-+ volatile uint32_t opds; /**< OutPktsDiscarded Statistic */
-+#if (DPAA_VERSION >= 11)
-+ volatile uint8_t res22[124]; /**< reserved */
-+ _Packed struct
-+ {
-+ volatile uint32_t rxsak[8]; /**< RX Security Association key (128/256 bit) */
-+ volatile uint8_t res23[32]; /**< reserved */
-+ } _PackedType rxScSaKey[NUM_OF_SA_PER_RX_SC];
-+ _Packed struct
-+ {
-+ volatile uint32_t txsak[8]; /**< TX Security Association key (128/256 bit) */
-+ volatile uint8_t res24[32]; /**< reserved */
-+ } _PackedType txScSaKey[NUM_OF_SA_PER_TX_SC];
-+#endif /* (DPAA_VERSION >= 11) */
-+} _PackedType t_FmMacsecRegs;
-+
-+#if defined(__MWERKS__) && !defined(__GNUC__)
-+#pragma pack(pop)
-+#endif /* defined(__MWERKS__) && ... */
-+
-+
-+/**************************************************************************//**
-+ @Description General defines
-+*//***************************************************************************/
-+
-+#define SCI_HIGH_MASK 0xffffffff00000000LL
-+#define SCI_LOW_MASK 0x00000000ffffffffLL
-+
-+#define LONG_SHIFT 32
-+
-+#define GET_SCI_FIRST_HALF(sci) (uint32_t)((macsecSCI_t)((macsecSCI_t)(sci) & SCI_HIGH_MASK) >> LONG_SHIFT)
-+#define GET_SCI_SECOND_HALF(sci) (uint32_t)((macsecSCI_t)(sci) & SCI_LOW_MASK)
-+
-+/**************************************************************************//**
-+ @Description Configuration defines
-+*//***************************************************************************/
-+
-+/* masks */
-+#define CFG_UECT 0x00000800
-+#define CFG_ESCBT 0x00000400
-+#define CFG_USFT 0x00000300
-+#define CFG_ITT 0x00000080
-+#define CFG_KFT 0x00000040
-+#define CFG_UFT 0x00000030
-+#define CFG_KSS 0x00000004
-+#define CFG_BYPN 0x00000002
-+#define CFG_S0I 0x00000001
-+
-+#define ET_TYPE 0x0000ffff
-+
-+#define MFL_MAX_LEN 0x0000ffff
-+
-+#define RXSCA_SC_SEL 0x0000000f
-+
-+#define TXSCA_SC_SEL 0x0000000f
-+
-+#define IP_REV_1_IP_ID 0xffff0000
-+#define IP_REV_1_IP_MJ 0x0000ff00
-+#define IP_REV_1_IP_MM 0x000000ff
-+
-+#define IP_REV_2_IP_INT 0x00ff0000
-+#define IP_REV_2_IP_ERR 0x0000ff00
-+#define IP_REV_2_IP_CFG 0x000000ff
-+
-+#define MECC_CAP 0x80000000
-+#define MECC_CET 0x40000000
-+#define MECC_SERCNT 0x00ff0000
-+#define MECC_MEMADDR 0x000001ff
-+
-+/* shifts */
-+#define CFG_UECT_SHIFT (31-20)
-+#define CFG_ESCBT_SHIFT (31-21)
-+#define CFG_USFT_SHIFT (31-23)
-+#define CFG_ITT_SHIFT (31-24)
-+#define CFG_KFT_SHIFT (31-25)
-+#define CFG_UFT_SHIFT (31-27)
-+#define CFG_KSS_SHIFT (31-29)
-+#define CFG_BYPN_SHIFT (31-30)
-+#define CFG_S0I_SHIFT (31-31)
-+
-+#define IP_REV_1_IP_ID_SHIFT (31-15)
-+#define IP_REV_1_IP_MJ_SHIFT (31-23)
-+#define IP_REV_1_IP_MM_SHIFT (31-31)
-+
-+#define IP_REV_2_IP_INT_SHIFT (31-15)
-+#define IP_REV_2_IP_ERR_SHIFT (31-23)
-+#define IP_REV_2_IP_CFG_SHIFT (31-31)
-+
-+#define MECC_CAP_SHIFT (31-0)
-+#define MECC_CET_SHIFT (31-1)
-+#define MECC_SERCNT_SHIFT (31-15)
-+#define MECC_MEMADDR_SHIFT (31-31)
-+
-+/**************************************************************************//**
-+ @Description RX SC defines
-+*//***************************************************************************/
-+
-+/* masks */
-+#define RX_SCCFG_SCI_EN_MASK 0x00000800
-+#define RX_SCCFG_RP_MASK 0x00000400
-+#define RX_SCCFG_VF_MASK 0x00000300
-+#define RX_SCCFG_CO_MASK 0x0000003f
-+
-+/* shifts */
-+#define RX_SCCFG_SCI_EN_SHIFT (31-20)
-+#define RX_SCCFG_RP_SHIFT (31-21)
-+#define RX_SCCFG_VF_SHIFT (31-23)
-+#define RX_SCCFG_CO_SHIFT (31-31)
-+#define RX_SCCFG_CS_SHIFT (31-7)
-+
-+/**************************************************************************//**
-+ @Description RX SA defines
-+*//***************************************************************************/
-+
-+/* masks */
-+#define RX_SACFG_ACTIVE 0x80000000
-+#define RX_SACFG_AN_MASK 0x00000006
-+#define RX_SACFG_EN_MASK 0x00000001
-+
-+/* shifts */
-+#define RX_SACFG_AN_SHIFT (31-30)
-+#define RX_SACFG_EN_SHIFT (31-31)
-+
-+/**************************************************************************//**
-+ @Description TX SC defines
-+*//***************************************************************************/
-+
-+/* masks */
-+#define TX_SCCFG_AN_MASK 0x000c0000
-+#define TX_SCCFG_ASA_MASK 0x00020000
-+#define TX_SCCFG_SCE_MASK 0x00010000
-+#define TX_SCCFG_CO_MASK 0x00003f00
-+#define TX_SCCFG_CE_MASK 0x00000010
-+#define TX_SCCFG_PF_MASK 0x00000008
-+#define TX_SCCFG_AIS_MASK 0x00000004
-+#define TX_SCCFG_UES_MASK 0x00000002
-+#define TX_SCCFG_USCB_MASK 0x00000001
-+
-+/* shifts */
-+#define TX_SCCFG_AN_SHIFT (31-13)
-+#define TX_SCCFG_ASA_SHIFT (31-14)
-+#define TX_SCCFG_SCE_SHIFT (31-15)
-+#define TX_SCCFG_CO_SHIFT (31-23)
-+#define TX_SCCFG_CE_SHIFT (31-27)
-+#define TX_SCCFG_PF_SHIFT (31-28)
-+#define TX_SCCFG_AIS_SHIFT (31-29)
-+#define TX_SCCFG_UES_SHIFT (31-30)
-+#define TX_SCCFG_USCB_SHIFT (31-31)
-+#define TX_SCCFG_CS_SHIFT (31-7)
-+
-+/**************************************************************************//**
-+ @Description TX SA defines
-+*//***************************************************************************/
-+
-+/* masks */
-+#define TX_SACFG_ACTIVE 0x80000000
-+
-+
-+typedef struct
-+{
-+ void (*f_Isr) (t_Handle h_Arg, uint32_t id);
-+ t_Handle h_SrcHandle;
-+} t_FmMacsecIntrSrc;
-+
-+typedef struct
-+{
-+ e_FmMacsecUnknownSciFrameTreatment unknownSciTreatMode;
-+ bool invalidTagsDeliverUncontrolled;
-+ bool changedTextWithNoEncryptDeliverUncontrolled;
-+ bool onlyScbIsSetDeliverUncontrolled;
-+ bool encryptWithNoChangedTextDiscardUncontrolled;
-+ e_FmMacsecUntagFrameTreatment untagTreatMode;
-+ uint32_t pnExhThr;
-+ bool keysUnreadable;
-+ bool byPassMode;
-+ bool reservedSc0;
-+ uint32_t sectagOverhead;
-+ uint32_t mflSubtract;
-+} t_FmMacsecDriverParam;
-+
-+typedef struct
-+{
-+ t_FmMacsecControllerDriver fmMacsecControllerDriver;
-+ t_Handle h_Fm;
-+ t_FmMacsecRegs *p_FmMacsecRegs;
-+ t_Handle h_FmMac; /**< A handle to the FM MAC object related to */
-+ char fmMacsecModuleName[MODULE_NAME_SIZE];
-+ t_FmMacsecIntrSrc intrMng[NUM_OF_INTER_MODULE_EVENTS];
-+ uint32_t events;
-+ uint32_t exceptions;
-+ uint32_t userExceptions;
-+ t_FmMacsecExceptionsCallback *f_Exception; /**< Exception Callback Routine */
-+ t_Handle h_App; /**< A handle to an application layer object; This handle will
-+ be passed by the driver upon calling the above callbacks */
-+ bool rxScTable[NUM_OF_RX_SC];
-+ uint32_t numRxScAvailable;
-+ bool txScTable[NUM_OF_TX_SC];
-+ uint32_t numTxScAvailable;
-+ t_Handle rxScSpinLock;
-+ t_Handle txScSpinLock;
-+ t_FmMacsecDriverParam *p_FmMacsecDriverParam;
-+} t_FmMacsec;
-+
-+
-+#endif /* __FM_MACSEC_MASTER_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MACSEC/fm_macsec_secy.c
-@@ -0,0 +1,883 @@
-+/*
-+ * Copyright 2008-2015 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+/******************************************************************************
-+ @File fm_macsec_secy.c
-+
-+ @Description FM MACSEC SECY driver routines implementation.
-+*//***************************************************************************/
-+
-+#include "std_ext.h"
-+#include "error_ext.h"
-+#include "xx_ext.h"
-+#include "string_ext.h"
-+#include "sprint_ext.h"
-+
-+#include "fm_macsec_secy.h"
-+
-+
-+/****************************************/
-+/* static functions */
-+/****************************************/
-+static void FmMacsecSecYExceptionsIsr(t_Handle h_FmMacsecSecY, uint32_t id)
-+{
-+ t_FmMacsecSecY *p_FmMacsecSecY = (t_FmMacsecSecY *)h_FmMacsecSecY;
-+
-+ UNUSED(id);
-+ SANITY_CHECK_RETURN(p_FmMacsecSecY, E_INVALID_HANDLE);
-+
-+ if (p_FmMacsecSecY->exceptions & FM_MACSEC_SECY_EX_FRAME_DISCARDED)
-+ p_FmMacsecSecY->f_Exception(p_FmMacsecSecY->h_App, e_FM_MACSEC_SECY_EX_FRAME_DISCARDED);
-+}
-+
-+static void FmMacsecSecYEventsIsr(t_Handle h_FmMacsecSecY, uint32_t id)
-+{
-+ t_FmMacsecSecY *p_FmMacsecSecY = (t_FmMacsecSecY *)h_FmMacsecSecY;
-+
-+ UNUSED(id);
-+ SANITY_CHECK_RETURN(p_FmMacsecSecY, E_INVALID_HANDLE);
-+
-+ if (p_FmMacsecSecY->events & FM_MACSEC_SECY_EV_NEXT_PN)
-+ p_FmMacsecSecY->f_Event(p_FmMacsecSecY->h_App, e_FM_MACSEC_SECY_EV_NEXT_PN);
-+}
-+
-+static t_Error CheckFmMacsecSecYParameters(t_FmMacsecSecY *p_FmMacsecSecY)
-+{
-+ if (!p_FmMacsecSecY->f_Exception)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("Exceptions callback not provided"));
-+
-+ if (!p_FmMacsecSecY->f_Event)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("Events callback not provided"));
-+
-+ if (!p_FmMacsecSecY->numOfRxSc)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("Num of Rx Scs must be greater than '0'"));
-+
-+
-+ return E_OK;
-+}
-+
-+static t_Handle FmMacsecSecYCreateSc(t_FmMacsecSecY *p_FmMacsecSecY,
-+ macsecSCI_t sci,
-+ e_FmMacsecSecYCipherSuite cipherSuite,
-+ e_ScType type)
-+{
-+ t_SecYSc *p_ScTable;
-+ void *p_Params;
-+ uint32_t numOfSc,i;
-+ t_Error err = E_OK;
-+ t_RxScParams rxScParams;
-+ t_TxScParams txScParams;
-+
-+ ASSERT_COND(p_FmMacsecSecY);
-+ ASSERT_COND(p_FmMacsecSecY->h_FmMacsec);
-+
-+ if (type == e_SC_RX)
-+ {
-+ memset(&rxScParams, 0, sizeof(rxScParams));
-+ i = (NUM_OF_RX_SC - 1);
-+ p_ScTable = p_FmMacsecSecY->p_RxSc;
-+ numOfSc = p_FmMacsecSecY->numOfRxSc;
-+ rxScParams.confidentialityOffset = p_FmMacsecSecY->confidentialityOffset;
-+ rxScParams.replayProtect = p_FmMacsecSecY->replayProtect;
-+ rxScParams.replayWindow = p_FmMacsecSecY->replayWindow;
-+ rxScParams.validateFrames = p_FmMacsecSecY->validateFrames;
-+ rxScParams.cipherSuite = cipherSuite;
-+ p_Params = &rxScParams;
-+ }
-+ else
-+ {
-+ memset(&txScParams, 0, sizeof(txScParams));
-+ i = (NUM_OF_TX_SC - 1);
-+ p_ScTable = p_FmMacsecSecY->p_TxSc;
-+ numOfSc = p_FmMacsecSecY->numOfTxSc;
-+ txScParams.sciInsertionMode = p_FmMacsecSecY->sciInsertionMode;
-+ txScParams.protectFrames = p_FmMacsecSecY->protectFrames;
-+ txScParams.confidentialityEnable = p_FmMacsecSecY->confidentialityEnable;
-+ txScParams.confidentialityOffset = p_FmMacsecSecY->confidentialityOffset;
-+ txScParams.cipherSuite = cipherSuite;
-+ p_Params = &txScParams;
-+ }
-+
-+ for (i=0;i<numOfSc;i++)
-+ if (!p_ScTable[i].inUse)
-+ break;
-+ if (i == numOfSc)
-+ {
-+ REPORT_ERROR(MAJOR, E_FULL, ("FM MACSEC SECY SC"));
-+ return NULL;
-+ }
-+
-+ if (type == e_SC_RX)
-+ {
-+ ((t_RxScParams *)p_Params)->scId = p_ScTable[i].scId;
-+ ((t_RxScParams *)p_Params)->sci = sci;
-+ if ((err = FmMacsecCreateRxSc(p_FmMacsecSecY->h_FmMacsec, (t_RxScParams *)p_Params)) != E_OK)
-+ {
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("FM MACSEC SECY RX SC"));
-+ return NULL;
-+ }
-+ }
-+ else
-+ {
-+ ((t_TxScParams *)p_Params)->scId = p_ScTable[i].scId;
-+ ((t_TxScParams *)p_Params)->sci = sci;
-+ if ((err = FmMacsecCreateTxSc(p_FmMacsecSecY->h_FmMacsec, (t_TxScParams *)p_Params)) != E_OK)
-+ {
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("FM MACSEC SECY TX SC"));
-+ return NULL;
-+ }
-+ }
-+
-+ p_ScTable[i].inUse = TRUE;
-+ return &p_ScTable[i];
-+}
-+
-+static t_Error FmMacsecSecYDeleteSc(t_FmMacsecSecY *p_FmMacsecSecY, t_SecYSc *p_FmSecYSc, e_ScType type)
-+{
-+ t_Error err = E_OK;
-+
-+ ASSERT_COND(p_FmMacsecSecY);
-+ ASSERT_COND(p_FmMacsecSecY->h_FmMacsec);
-+ ASSERT_COND(p_FmSecYSc);
-+
-+ if (type == e_SC_RX)
-+ {
-+ if ((err = FmMacsecDeleteRxSc(p_FmMacsecSecY->h_FmMacsec, p_FmSecYSc->scId)) != E_OK)
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+ }
-+ else
-+ if ((err = FmMacsecDeleteTxSc(p_FmMacsecSecY->h_FmMacsec, p_FmSecYSc->scId)) != E_OK)
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+
-+ p_FmSecYSc->inUse = FALSE;
-+
-+ return err;
-+}
-+
-+/****************************************/
-+/* API Init unit functions */
-+/****************************************/
-+t_Handle FM_MACSEC_SECY_Config(t_FmMacsecSecYParams *p_FmMacsecSecYParam)
-+{
-+ t_FmMacsecSecY *p_FmMacsecSecY;
-+
-+ /* Allocate FM MACSEC structure */
-+ p_FmMacsecSecY = (t_FmMacsecSecY *) XX_Malloc(sizeof(t_FmMacsecSecY));
-+ if (!p_FmMacsecSecY)
-+ {
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("FM MACSEC SECY driver structure"));
-+ return NULL;
-+ }
-+ memset(p_FmMacsecSecY, 0, sizeof(t_FmMacsecSecY));
-+
-+ /* Allocate the FM MACSEC driver's parameters structure */
-+ p_FmMacsecSecY->p_FmMacsecSecYDriverParam = (t_FmMacsecSecYDriverParam *)XX_Malloc(sizeof(t_FmMacsecSecYDriverParam));
-+ if (!p_FmMacsecSecY->p_FmMacsecSecYDriverParam)
-+ {
-+ XX_Free(p_FmMacsecSecY);
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("FM MACSEC SECY driver parameters"));
-+ return NULL;
-+ }
-+ memset(p_FmMacsecSecY->p_FmMacsecSecYDriverParam, 0, sizeof(t_FmMacsecSecYDriverParam));
-+
-+ /* Initialize FM MACSEC SECY parameters which will be kept by the driver */
-+ p_FmMacsecSecY->h_FmMacsec = p_FmMacsecSecYParam->h_FmMacsec;
-+ p_FmMacsecSecY->f_Event = p_FmMacsecSecYParam->f_Event;
-+ p_FmMacsecSecY->f_Exception = p_FmMacsecSecYParam->f_Exception;
-+ p_FmMacsecSecY->h_App = p_FmMacsecSecYParam->h_App;
-+ p_FmMacsecSecY->confidentialityEnable = DEFAULT_confidentialityEnable;
-+ p_FmMacsecSecY->confidentialityOffset = DEFAULT_confidentialityOffset;
-+ p_FmMacsecSecY->validateFrames = DEFAULT_validateFrames;
-+ p_FmMacsecSecY->replayProtect = DEFAULT_replayEnable;
-+ p_FmMacsecSecY->replayWindow = DEFAULT_replayWindow;
-+ p_FmMacsecSecY->protectFrames = DEFAULT_protectFrames;
-+ p_FmMacsecSecY->sciInsertionMode = DEFAULT_sciInsertionMode;
-+ p_FmMacsecSecY->isPointToPoint = DEFAULT_ptp;
-+ p_FmMacsecSecY->numOfRxSc = p_FmMacsecSecYParam->numReceiveChannels;
-+ p_FmMacsecSecY->numOfTxSc = DEFAULT_numOfTxSc;
-+ p_FmMacsecSecY->exceptions = DEFAULT_exceptions;
-+ p_FmMacsecSecY->events = DEFAULT_events;
-+
-+ memcpy(&p_FmMacsecSecY->p_FmMacsecSecYDriverParam->txScParams,
-+ &p_FmMacsecSecYParam->txScParams,
-+ sizeof(t_FmMacsecSecYSCParams));
-+ return p_FmMacsecSecY;
-+}
-+
-+t_Error FM_MACSEC_SECY_Init(t_Handle h_FmMacsecSecY)
-+{
-+ t_FmMacsecSecY *p_FmMacsecSecY = (t_FmMacsecSecY *)h_FmMacsecSecY;
-+ t_FmMacsecSecYDriverParam *p_FmMacsecSecYDriverParam = NULL;
-+ uint32_t rxScIds[NUM_OF_RX_SC], txScIds[NUM_OF_TX_SC], i, j;
-+ t_Error err;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsecSecY, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsecSecY->p_FmMacsecSecYDriverParam, E_INVALID_HANDLE);
-+
-+ CHECK_INIT_PARAMETERS(p_FmMacsecSecY, CheckFmMacsecSecYParameters);
-+
-+ p_FmMacsecSecYDriverParam = p_FmMacsecSecY->p_FmMacsecSecYDriverParam;
-+
-+ if ((p_FmMacsecSecY->isPointToPoint) &&
-+ ((err = FmMacsecSetPTP(p_FmMacsecSecY->h_FmMacsec, TRUE)) != E_OK))
-+ RETURN_ERROR(MAJOR, err, ("Can't set Poin-to-Point"));
-+
-+ /* Rx Sc Allocation */
-+ p_FmMacsecSecY->p_RxSc = (t_SecYSc *)XX_Malloc(sizeof(t_SecYSc) * p_FmMacsecSecY->numOfRxSc);
-+ if (!p_FmMacsecSecY->p_RxSc)
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, ("FM MACSEC SECY RX SC"));
-+ memset(p_FmMacsecSecY->p_RxSc, 0, sizeof(t_SecYSc) * p_FmMacsecSecY->numOfRxSc);
-+ if ((err = FmMacsecAllocScs(p_FmMacsecSecY->h_FmMacsec, e_SC_RX, p_FmMacsecSecY->isPointToPoint, p_FmMacsecSecY->numOfRxSc, rxScIds)) != E_OK)
-+ {
-+ if (p_FmMacsecSecY->p_TxSc)
-+ XX_Free(p_FmMacsecSecY->p_TxSc);
-+ if (p_FmMacsecSecY->p_RxSc)
-+ XX_Free(p_FmMacsecSecY->p_RxSc);
-+ return ERROR_CODE(err);
-+ }
-+ for (i=0; i<p_FmMacsecSecY->numOfRxSc; i++)
-+ {
-+ p_FmMacsecSecY->p_RxSc[i].scId = rxScIds[i];
-+ p_FmMacsecSecY->p_RxSc[i].type = e_SC_RX;
-+ for (j=0; j<MAX_NUM_OF_SA_PER_SC;j++)
-+ p_FmMacsecSecY->p_RxSc[i].sa[j].saId = (e_ScSaId)SECY_AN_FREE_VALUE;
-+ }
-+
-+ /* Tx Sc Allocation */
-+ p_FmMacsecSecY->p_TxSc = (t_SecYSc *)XX_Malloc(sizeof(t_SecYSc) * p_FmMacsecSecY->numOfTxSc);
-+ if (!p_FmMacsecSecY->p_TxSc)
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, ("FM MACSEC SECY TX SC"));
-+ memset(p_FmMacsecSecY->p_TxSc, 0, sizeof(t_SecYSc) * p_FmMacsecSecY->numOfTxSc);
-+
-+ if ((err = FmMacsecAllocScs(p_FmMacsecSecY->h_FmMacsec, e_SC_TX, p_FmMacsecSecY->isPointToPoint, p_FmMacsecSecY->numOfTxSc, txScIds)) != E_OK)
-+ {
-+ if (p_FmMacsecSecY->p_TxSc)
-+ XX_Free(p_FmMacsecSecY->p_TxSc);
-+ if (p_FmMacsecSecY->p_RxSc)
-+ XX_Free(p_FmMacsecSecY->p_RxSc);
-+ return ERROR_CODE(err);
-+ }
-+ for (i=0; i<p_FmMacsecSecY->numOfTxSc; i++)
-+ {
-+ p_FmMacsecSecY->p_TxSc[i].scId = txScIds[i];
-+ p_FmMacsecSecY->p_TxSc[i].type = e_SC_TX;
-+ for (j=0; j<MAX_NUM_OF_SA_PER_SC;j++)
-+ p_FmMacsecSecY->p_TxSc[i].sa[j].saId = (e_ScSaId)SECY_AN_FREE_VALUE;
-+ FmMacsecRegisterIntr(p_FmMacsecSecY->h_FmMacsec,
-+ e_FM_MACSEC_MOD_SC_TX,
-+ (uint8_t)txScIds[i],
-+ e_FM_INTR_TYPE_ERR,
-+ FmMacsecSecYExceptionsIsr,
-+ p_FmMacsecSecY);
-+ FmMacsecRegisterIntr(p_FmMacsecSecY->h_FmMacsec,
-+ e_FM_MACSEC_MOD_SC_TX,
-+ (uint8_t)txScIds[i],
-+ e_FM_INTR_TYPE_NORMAL,
-+ FmMacsecSecYEventsIsr,
-+ p_FmMacsecSecY);
-+
-+ if (p_FmMacsecSecY->exceptions & FM_MACSEC_SECY_EX_FRAME_DISCARDED)
-+ FmMacsecSetException(p_FmMacsecSecY->h_FmMacsec, e_FM_MACSEC_EX_TX_SC, txScIds[i], TRUE);
-+ if (p_FmMacsecSecY->events & FM_MACSEC_SECY_EV_NEXT_PN)
-+ FmMacsecSetEvent(p_FmMacsecSecY->h_FmMacsec, e_FM_MACSEC_EV_TX_SC_NEXT_PN, txScIds[i], TRUE);
-+ }
-+
-+ FmMacsecSecYCreateSc(p_FmMacsecSecY,
-+ p_FmMacsecSecYDriverParam->txScParams.sci,
-+ p_FmMacsecSecYDriverParam->txScParams.cipherSuite,
-+ e_SC_TX);
-+ XX_Free(p_FmMacsecSecYDriverParam);
-+ p_FmMacsecSecY->p_FmMacsecSecYDriverParam = NULL;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_MACSEC_SECY_Free(t_Handle h_FmMacsecSecY)
-+{
-+ t_FmMacsecSecY *p_FmMacsecSecY = (t_FmMacsecSecY *)h_FmMacsecSecY;
-+ t_Error err = E_OK;
-+ uint32_t rxScIds[NUM_OF_RX_SC], txScIds[NUM_OF_TX_SC], i;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsecSecY, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmMacsecSecY->p_FmMacsecSecYDriverParam, E_INVALID_STATE);
-+
-+ if (p_FmMacsecSecY->isPointToPoint)
-+ FmMacsecSetPTP(p_FmMacsecSecY->h_FmMacsec, FALSE);
-+ if (p_FmMacsecSecY->p_RxSc)
-+ {
-+ for (i=0; i<p_FmMacsecSecY->numOfRxSc; i++)
-+ rxScIds[i] = p_FmMacsecSecY->p_RxSc[i].scId;
-+ if ((err = FmMacsecFreeScs(p_FmMacsecSecY->h_FmMacsec, e_SC_RX, p_FmMacsecSecY->numOfRxSc, rxScIds)) != E_OK)
-+ return ERROR_CODE(err);
-+ XX_Free(p_FmMacsecSecY->p_RxSc);
-+ }
-+ if (p_FmMacsecSecY->p_TxSc)
-+ {
-+ FmMacsecSecYDeleteSc(p_FmMacsecSecY, &p_FmMacsecSecY->p_TxSc[0], e_SC_TX);
-+
-+ for (i=0; i<p_FmMacsecSecY->numOfTxSc; i++) {
-+ txScIds[i] = p_FmMacsecSecY->p_TxSc[i].scId;
-+ FmMacsecUnregisterIntr(p_FmMacsecSecY->h_FmMacsec,
-+ e_FM_MACSEC_MOD_SC_TX,
-+ (uint8_t)txScIds[i],
-+ e_FM_INTR_TYPE_ERR);
-+ FmMacsecUnregisterIntr(p_FmMacsecSecY->h_FmMacsec,
-+ e_FM_MACSEC_MOD_SC_TX,
-+ (uint8_t)txScIds[i],
-+ e_FM_INTR_TYPE_NORMAL);
-+
-+ if (p_FmMacsecSecY->exceptions & FM_MACSEC_SECY_EX_FRAME_DISCARDED)
-+ FmMacsecSetException(p_FmMacsecSecY->h_FmMacsec, e_FM_MACSEC_EX_TX_SC, txScIds[i], FALSE);
-+ if (p_FmMacsecSecY->events & FM_MACSEC_SECY_EV_NEXT_PN)
-+ FmMacsecSetEvent(p_FmMacsecSecY->h_FmMacsec, e_FM_MACSEC_EV_TX_SC_NEXT_PN, txScIds[i], FALSE);
-+ }
-+
-+ if ((err = FmMacsecFreeScs(p_FmMacsecSecY->h_FmMacsec, e_SC_TX, p_FmMacsecSecY->numOfTxSc, txScIds)) != E_OK)
-+ return ERROR_CODE(err);
-+ XX_Free(p_FmMacsecSecY->p_TxSc);
-+ }
-+
-+ XX_Free(p_FmMacsecSecY);
-+
-+ return err;
-+}
-+
-+t_Error FM_MACSEC_SECY_ConfigSciInsertionMode(t_Handle h_FmMacsecSecY, e_FmMacsecSciInsertionMode sciInsertionMode)
-+{
-+ t_FmMacsecSecY *p_FmMacsecSecY = (t_FmMacsecSecY *)h_FmMacsecSecY;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsecSecY, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsecSecY->p_FmMacsecSecYDriverParam, E_INVALID_STATE);
-+
-+ p_FmMacsecSecY->sciInsertionMode = sciInsertionMode;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_MACSEC_SECY_ConfigProtectFrames(t_Handle h_FmMacsecSecY, bool protectFrames)
-+{
-+ t_FmMacsecSecY *p_FmMacsecSecY = (t_FmMacsecSecY *)h_FmMacsecSecY;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsecSecY, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsecSecY->p_FmMacsecSecYDriverParam, E_INVALID_STATE);
-+
-+ p_FmMacsecSecY->protectFrames = protectFrames;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_MACSEC_SECY_ConfigReplayWindow(t_Handle h_FmMacsecSecY, bool replayProtect, uint32_t replayWindow)
-+{
-+ t_FmMacsecSecY *p_FmMacsecSecY = (t_FmMacsecSecY *)h_FmMacsecSecY;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsecSecY, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsecSecY->p_FmMacsecSecYDriverParam, E_INVALID_STATE);
-+
-+ p_FmMacsecSecY->replayProtect = replayProtect;
-+ p_FmMacsecSecY->replayWindow = replayWindow;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_MACSEC_SECY_ConfigValidationMode(t_Handle h_FmMacsecSecY, e_FmMacsecValidFrameBehavior validateFrames)
-+{
-+ t_FmMacsecSecY *p_FmMacsecSecY = (t_FmMacsecSecY *)h_FmMacsecSecY;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsecSecY, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsecSecY->p_FmMacsecSecYDriverParam, E_INVALID_STATE);
-+
-+ p_FmMacsecSecY->validateFrames = validateFrames;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_MACSEC_SECY_ConfigConfidentiality(t_Handle h_FmMacsecSecY, bool confidentialityEnable, uint16_t confidentialityOffset)
-+{
-+ t_FmMacsecSecY *p_FmMacsecSecY = (t_FmMacsecSecY *)h_FmMacsecSecY;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsecSecY, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsecSecY->p_FmMacsecSecYDriverParam, E_INVALID_STATE);
-+
-+ p_FmMacsecSecY->confidentialityEnable = confidentialityEnable;
-+ p_FmMacsecSecY->confidentialityOffset = confidentialityOffset;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_MACSEC_SECY_ConfigPointToPoint(t_Handle h_FmMacsecSecY)
-+{
-+ t_FmMacsecSecY *p_FmMacsecSecY = (t_FmMacsecSecY *)h_FmMacsecSecY;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsecSecY, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsecSecY->p_FmMacsecSecYDriverParam, E_INVALID_STATE);
-+
-+ p_FmMacsecSecY->numOfRxSc = 1;
-+ p_FmMacsecSecY->isPointToPoint = TRUE;
-+ p_FmMacsecSecY->sciInsertionMode = e_FM_MACSEC_SCI_INSERTION_MODE_IMPLICT_PTP;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_MACSEC_SECY_ConfigException(t_Handle h_FmMacsecSecY, e_FmMacsecSecYExceptions exception, bool enable)
-+{
-+ t_FmMacsecSecY *p_FmMacsecSecY = (t_FmMacsecSecY *)h_FmMacsecSecY;
-+ uint32_t bitMask = 0;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsecSecY, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsecSecY->p_FmMacsecSecYDriverParam, E_INVALID_STATE);
-+
-+ GET_EXCEPTION_FLAG(bitMask, exception);
-+ if (bitMask)
-+ {
-+ if (enable)
-+ p_FmMacsecSecY->exceptions |= bitMask;
-+ else
-+ p_FmMacsecSecY->exceptions &= ~bitMask;
-+ }
-+ else
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("Undefined exception"));
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_MACSEC_SECY_ConfigEvent(t_Handle h_FmMacsecSecY, e_FmMacsecSecYEvents event, bool enable)
-+{
-+ t_FmMacsecSecY *p_FmMacsecSecY = (t_FmMacsecSecY *)h_FmMacsecSecY;
-+ uint32_t bitMask = 0;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsecSecY, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsecSecY->p_FmMacsecSecYDriverParam, E_INVALID_STATE);
-+
-+ GET_EVENT_FLAG(bitMask, event);
-+ if (bitMask)
-+ {
-+ if (enable)
-+ p_FmMacsecSecY->events |= bitMask;
-+ else
-+ p_FmMacsecSecY->events &= ~bitMask;
-+ }
-+ else
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("Undefined event"));
-+
-+ return E_OK;
-+}
-+
-+t_Handle FM_MACSEC_SECY_CreateRxSc(t_Handle h_FmMacsecSecY, t_FmMacsecSecYSCParams *p_ScParams)
-+{
-+ t_FmMacsecSecY *p_FmMacsecSecY = (t_FmMacsecSecY *)h_FmMacsecSecY;
-+
-+ SANITY_CHECK_RETURN_VALUE(p_FmMacsecSecY, E_INVALID_HANDLE, NULL);
-+ SANITY_CHECK_RETURN_VALUE(p_FmMacsecSecY->h_FmMacsec, E_INVALID_HANDLE, NULL);
-+ SANITY_CHECK_RETURN_VALUE(p_ScParams, E_NULL_POINTER, NULL);
-+ SANITY_CHECK_RETURN_VALUE(!p_FmMacsecSecY->p_FmMacsecSecYDriverParam, E_INVALID_STATE, NULL);
-+
-+ return FmMacsecSecYCreateSc(p_FmMacsecSecY, p_ScParams->sci, p_ScParams->cipherSuite, e_SC_RX);
-+}
-+
-+t_Error FM_MACSEC_SECY_DeleteRxSc(t_Handle h_FmMacsecSecY, t_Handle h_Sc)
-+{
-+ t_FmMacsecSecY *p_FmMacsecSecY = (t_FmMacsecSecY *)h_FmMacsecSecY;
-+ t_SecYSc *p_FmSecYSc = (t_SecYSc *)h_Sc;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsecSecY, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsecSecY->h_FmMacsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmMacsecSecY->p_FmMacsecSecYDriverParam, E_INVALID_STATE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmSecYSc, E_INVALID_HANDLE);
-+
-+ return FmMacsecSecYDeleteSc(p_FmMacsecSecY, p_FmSecYSc, e_SC_RX);
-+}
-+
-+t_Error FM_MACSEC_SECY_CreateRxSa(t_Handle h_FmMacsecSecY, t_Handle h_Sc, macsecAN_t an, uint32_t lowestPn, macsecSAKey_t key)
-+{
-+ t_FmMacsecSecY *p_FmMacsecSecY = (t_FmMacsecSecY *)h_FmMacsecSecY;
-+ t_SecYSc *p_FmSecYSc = (t_SecYSc *)h_Sc;
-+ t_Error err = E_OK;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsecSecY, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsecSecY->h_FmMacsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmMacsecSecY->p_FmMacsecSecYDriverParam, E_INVALID_STATE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmSecYSc, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(an < MAX_NUM_OF_SA_PER_SC, E_INVALID_STATE);
-+
-+ if (p_FmSecYSc->sa[an].saId != SECY_AN_FREE_VALUE)
-+ RETURN_ERROR(MINOR, E_INVALID_STATE, ("An %d is already assigned",an));
-+
-+ if ((err = FmMacsecCreateRxSa(p_FmMacsecSecY->h_FmMacsec, p_FmSecYSc->scId, (e_ScSaId)p_FmSecYSc->numOfSa, an, lowestPn, key)) != E_OK)
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+
-+ p_FmSecYSc->sa[an].saId = (e_ScSaId)p_FmSecYSc->numOfSa++;
-+ return err;
-+}
-+
-+t_Error FM_MACSEC_SECY_DeleteRxSa(t_Handle h_FmMacsecSecY, t_Handle h_Sc, macsecAN_t an)
-+{
-+ t_FmMacsecSecY *p_FmMacsecSecY = (t_FmMacsecSecY *)h_FmMacsecSecY;
-+ t_SecYSc *p_FmSecYSc = (t_SecYSc *)h_Sc;
-+ t_Error err = E_OK;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsecSecY, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsecSecY->h_FmMacsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmMacsecSecY->p_FmMacsecSecYDriverParam, E_INVALID_STATE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmSecYSc, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(an < MAX_NUM_OF_SA_PER_SC, E_INVALID_STATE);
-+
-+ if (p_FmSecYSc->sa[an].saId == SECY_AN_FREE_VALUE)
-+ RETURN_ERROR(MINOR, E_INVALID_STATE, ("An %d is already deleted",an));
-+
-+ if ((err = FmMacsecDeleteRxSa(p_FmMacsecSecY->h_FmMacsec, p_FmSecYSc->scId, p_FmSecYSc->sa[an].saId)) != E_OK)
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+
-+ p_FmSecYSc->numOfSa--;
-+ p_FmSecYSc->sa[an].saId = (e_ScSaId)SECY_AN_FREE_VALUE;
-+ /* TODO - check if statistics need to be read*/
-+ return err;
-+}
-+
-+t_Error FM_MACSEC_SECY_RxSaEnableReceive(t_Handle h_FmMacsecSecY, t_Handle h_Sc, macsecAN_t an)
-+{
-+ t_FmMacsecSecY *p_FmMacsecSecY = (t_FmMacsecSecY *)h_FmMacsecSecY;
-+ t_SecYSc *p_FmSecYSc = (t_SecYSc *)h_Sc;
-+ t_Error err = E_OK;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsecSecY, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsecSecY->h_FmMacsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmMacsecSecY->p_FmMacsecSecYDriverParam, E_INVALID_STATE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmSecYSc, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(an < MAX_NUM_OF_SA_PER_SC, E_INVALID_STATE);
-+
-+ if (p_FmSecYSc->sa[an].saId == SECY_AN_FREE_VALUE)
-+ RETURN_ERROR(MINOR, E_INVALID_STATE, ("An %d is not configured",an));
-+
-+ if ((err = FmMacsecRxSaSetReceive(p_FmMacsecSecY->h_FmMacsec,p_FmSecYSc->scId, p_FmSecYSc->sa[an].saId, TRUE)) != E_OK)
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+
-+ p_FmSecYSc->sa[an].active = TRUE;
-+ return err;
-+}
-+
-+t_Error FM_MACSEC_SECY_RxSaDisableReceive(t_Handle h_FmMacsecSecY, t_Handle h_Sc, macsecAN_t an)
-+{
-+ t_FmMacsecSecY *p_FmMacsecSecY = (t_FmMacsecSecY *)h_FmMacsecSecY;
-+ t_SecYSc *p_FmSecYSc = (t_SecYSc *)h_Sc;
-+ t_Error err = E_OK;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsecSecY, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsecSecY->h_FmMacsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmMacsecSecY->p_FmMacsecSecYDriverParam, E_INVALID_STATE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmSecYSc, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(an < MAX_NUM_OF_SA_PER_SC, E_INVALID_STATE);
-+
-+ if (p_FmSecYSc->sa[an].saId == SECY_AN_FREE_VALUE)
-+ RETURN_ERROR(MINOR, E_INVALID_STATE, ("An %d is not configured",an));
-+
-+ if ((err = FmMacsecRxSaSetReceive(p_FmMacsecSecY->h_FmMacsec,p_FmSecYSc->scId, p_FmSecYSc->sa[an].saId, FALSE)) != E_OK)
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+
-+ p_FmSecYSc->sa[an].active = FALSE;
-+ return err;
-+}
-+
-+t_Error FM_MACSEC_SECY_RxSaUpdateNextPn(t_Handle h_FmMacsecSecY, t_Handle h_Sc, macsecAN_t an, uint32_t updtNextPN)
-+{
-+ t_FmMacsecSecY *p_FmMacsecSecY = (t_FmMacsecSecY *)h_FmMacsecSecY;
-+ t_SecYSc *p_FmSecYSc = (t_SecYSc *)h_Sc;
-+ t_Error err = E_OK;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsecSecY, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsecSecY->h_FmMacsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmMacsecSecY->p_FmMacsecSecYDriverParam, E_INVALID_STATE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmSecYSc, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(an < MAX_NUM_OF_SA_PER_SC, E_INVALID_STATE);
-+
-+ if (p_FmSecYSc->sa[an].saId == SECY_AN_FREE_VALUE)
-+ RETURN_ERROR(MINOR, E_INVALID_STATE, ("An %d is not configured",an));
-+
-+ if ((err = FmMacsecRxSaUpdateNextPn(p_FmMacsecSecY->h_FmMacsec,p_FmSecYSc->scId, p_FmSecYSc->sa[an].saId, updtNextPN)) != E_OK)
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+
-+ return err;
-+}
-+
-+t_Error FM_MACSEC_SECY_RxSaUpdateLowestPn(t_Handle h_FmMacsecSecY, t_Handle h_Sc, macsecAN_t an, uint32_t updtLowestPN)
-+{
-+ t_FmMacsecSecY *p_FmMacsecSecY = (t_FmMacsecSecY *)h_FmMacsecSecY;
-+ t_SecYSc *p_FmSecYSc = (t_SecYSc *)h_Sc;
-+ t_Error err = E_OK;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsecSecY, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsecSecY->h_FmMacsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmMacsecSecY->p_FmMacsecSecYDriverParam, E_INVALID_STATE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmSecYSc, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(an < MAX_NUM_OF_SA_PER_SC, E_INVALID_STATE);
-+
-+ if (p_FmSecYSc->sa[an].saId == SECY_AN_FREE_VALUE)
-+ RETURN_ERROR(MINOR, E_INVALID_STATE, ("An %d is not configured",an));
-+
-+ if ((err = FmMacsecRxSaUpdateLowestPn(p_FmMacsecSecY->h_FmMacsec,p_FmSecYSc->scId, p_FmSecYSc->sa[an].saId, updtLowestPN)) != E_OK)
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+
-+ return err;
-+}
-+
-+t_Error FM_MACSEC_SECY_RxSaModifyKey(t_Handle h_FmMacsecSecY, t_Handle h_Sc, macsecAN_t an, macsecSAKey_t key)
-+{
-+ t_FmMacsecSecY *p_FmMacsecSecY = (t_FmMacsecSecY *)h_FmMacsecSecY;
-+ t_SecYSc *p_FmSecYSc = (t_SecYSc *)h_Sc;
-+ t_Error err = E_OK;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsecSecY, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsecSecY->h_FmMacsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmMacsecSecY->p_FmMacsecSecYDriverParam, E_INVALID_STATE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmSecYSc, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(an < MAX_NUM_OF_SA_PER_SC, E_INVALID_STATE);
-+
-+ if (p_FmSecYSc->sa[an].saId == SECY_AN_FREE_VALUE)
-+ RETURN_ERROR(MINOR, E_INVALID_STATE, ("An %d is not configured",an));
-+
-+ if (p_FmSecYSc->sa[an].active)
-+ if ((err = FmMacsecRxSaSetReceive(p_FmMacsecSecY->h_FmMacsec, p_FmSecYSc->scId, p_FmSecYSc->sa[an].saId, FALSE)) != E_OK)
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+
-+ /* TODO - statistics should be read */
-+
-+ if ((err = FmMacsecCreateRxSa(p_FmMacsecSecY->h_FmMacsec, p_FmSecYSc->scId, p_FmSecYSc->sa[an].saId, an, 1, key)) != E_OK)
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+
-+ if (p_FmSecYSc->sa[an].active)
-+ if ((err = FmMacsecRxSaSetReceive(p_FmMacsecSecY->h_FmMacsec, p_FmSecYSc->scId, p_FmSecYSc->sa[an].saId, TRUE)) != E_OK)
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+ return err;
-+}
-+
-+
-+t_Error FM_MACSEC_SECY_CreateTxSa(t_Handle h_FmMacsecSecY, macsecAN_t an, macsecSAKey_t key)
-+{
-+ t_FmMacsecSecY *p_FmMacsecSecY = (t_FmMacsecSecY *)h_FmMacsecSecY;
-+ t_SecYSc *p_FmSecYSc;
-+ t_Error err = E_OK;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsecSecY, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsecSecY->h_FmMacsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmMacsecSecY->p_FmMacsecSecYDriverParam, E_INVALID_STATE);
-+ p_FmSecYSc = &p_FmMacsecSecY->p_TxSc[0];
-+ SANITY_CHECK_RETURN_ERROR(p_FmSecYSc, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(an < MAX_NUM_OF_SA_PER_SC, E_INVALID_STATE);
-+
-+ if (p_FmSecYSc->sa[an].saId != SECY_AN_FREE_VALUE)
-+ RETURN_ERROR(MINOR, err, ("An %d is already assigned",an));
-+
-+ if ((err = FmMacsecCreateTxSa(p_FmMacsecSecY->h_FmMacsec,p_FmSecYSc->scId, (e_ScSaId)p_FmSecYSc->numOfSa, key)) != E_OK)
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+
-+ p_FmSecYSc->sa[an].saId = (e_ScSaId)p_FmSecYSc->numOfSa++;
-+ return err;
-+}
-+
-+t_Error FM_MACSEC_SECY_DeleteTxSa(t_Handle h_FmMacsecSecY, macsecAN_t an)
-+{
-+ t_FmMacsecSecY *p_FmMacsecSecY = (t_FmMacsecSecY *)h_FmMacsecSecY;
-+ t_SecYSc *p_FmSecYSc;
-+ t_Error err = E_OK;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsecSecY, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsecSecY->h_FmMacsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmMacsecSecY->p_FmMacsecSecYDriverParam, E_INVALID_STATE);
-+ p_FmSecYSc = &p_FmMacsecSecY->p_TxSc[0];
-+ SANITY_CHECK_RETURN_ERROR(p_FmSecYSc, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(an < MAX_NUM_OF_SA_PER_SC, E_INVALID_STATE);
-+
-+ if (p_FmSecYSc->sa[an].saId == SECY_AN_FREE_VALUE)
-+ RETURN_ERROR(MINOR, E_INVALID_STATE, ("An %d is already deleted",an));
-+
-+ if ((err = FmMacsecDeleteTxSa(p_FmMacsecSecY->h_FmMacsec, p_FmSecYSc->scId, p_FmSecYSc->sa[an].saId)) != E_OK)
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+
-+ p_FmSecYSc->numOfSa--;
-+ p_FmSecYSc->sa[an].saId = (e_ScSaId)SECY_AN_FREE_VALUE;
-+ /* TODO - check if statistics need to be read*/
-+ return err;
-+}
-+
-+t_Error FM_MACSEC_SECY_TxSaModifyKey(t_Handle h_FmMacsecSecY, macsecAN_t nextActiveAn, macsecSAKey_t key)
-+{
-+ t_FmMacsecSecY *p_FmMacsecSecY = (t_FmMacsecSecY *)h_FmMacsecSecY;
-+ t_SecYSc *p_FmSecYSc;
-+ macsecAN_t currentAn;
-+ t_Error err = E_OK;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsecSecY, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsecSecY->h_FmMacsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmMacsecSecY->p_FmMacsecSecYDriverParam, E_INVALID_STATE);
-+ p_FmSecYSc = &p_FmMacsecSecY->p_TxSc[0];
-+ SANITY_CHECK_RETURN_ERROR(p_FmSecYSc, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(nextActiveAn < MAX_NUM_OF_SA_PER_SC, E_INVALID_STATE);
-+
-+ if ((err = FmMacsecTxSaGetActive(p_FmMacsecSecY->h_FmMacsec,
-+ p_FmSecYSc->scId,
-+ &currentAn)) != E_OK)
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+
-+ if ((err = FmMacsecTxSaSetActive(p_FmMacsecSecY->h_FmMacsec,
-+ p_FmSecYSc->scId,
-+ p_FmSecYSc->sa[nextActiveAn].saId,
-+ nextActiveAn)) != E_OK)
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+
-+ /* TODO - statistics should be read */
-+
-+ if ((err = FmMacsecCreateTxSa(p_FmMacsecSecY->h_FmMacsec, p_FmSecYSc->scId, p_FmSecYSc->sa[currentAn].saId, key)) != E_OK)
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+
-+ return err;
-+}
-+
-+t_Error FM_MACSEC_SECY_TxSaSetActive(t_Handle h_FmMacsecSecY, macsecAN_t an)
-+{
-+ t_FmMacsecSecY *p_FmMacsecSecY = (t_FmMacsecSecY *)h_FmMacsecSecY;
-+ t_SecYSc *p_FmSecYSc;
-+ t_Error err = E_OK;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsecSecY, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsecSecY->h_FmMacsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmMacsecSecY->p_FmMacsecSecYDriverParam, E_INVALID_STATE);
-+ p_FmSecYSc = &p_FmMacsecSecY->p_TxSc[0];
-+ SANITY_CHECK_RETURN_ERROR(p_FmSecYSc, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(an < MAX_NUM_OF_SA_PER_SC, E_INVALID_STATE);
-+
-+ if (p_FmSecYSc->sa[an].saId == SECY_AN_FREE_VALUE)
-+ RETURN_ERROR(MINOR, E_INVALID_STATE, ("An %d is not configured",an));
-+
-+ if ((err = FmMacsecTxSaSetActive(p_FmMacsecSecY->h_FmMacsec,
-+ p_FmSecYSc->scId,
-+ p_FmSecYSc->sa[an].saId,
-+ an)) != E_OK)
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+
-+ return err;
-+}
-+
-+t_Error FM_MACSEC_SECY_TxSaGetActive(t_Handle h_FmMacsecSecY, macsecAN_t *p_An)
-+{
-+ t_FmMacsecSecY *p_FmMacsecSecY = (t_FmMacsecSecY *)h_FmMacsecSecY;
-+ t_SecYSc *p_FmSecYSc;
-+ t_Error err = E_OK;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsecSecY, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsecSecY->h_FmMacsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmMacsecSecY->p_FmMacsecSecYDriverParam, E_INVALID_STATE);
-+ p_FmSecYSc = &p_FmMacsecSecY->p_TxSc[0];
-+ SANITY_CHECK_RETURN_ERROR(p_FmSecYSc, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_An, E_INVALID_HANDLE);
-+
-+ if ((err = FmMacsecTxSaGetActive(p_FmMacsecSecY->h_FmMacsec,
-+ p_FmSecYSc->scId,
-+ p_An)) != E_OK)
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+
-+ return err;
-+}
-+
-+t_Error FM_MACSEC_SECY_GetRxScPhysId(t_Handle h_FmMacsecSecY, t_Handle h_Sc, uint32_t *p_ScPhysId)
-+{
-+ t_SecYSc *p_FmSecYSc = (t_SecYSc *)h_Sc;
-+ t_Error err = E_OK;
-+
-+ SANITY_CHECK_RETURN_ERROR(h_FmMacsecSecY, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(((t_FmMacsecSecY *)h_FmMacsecSecY)->h_FmMacsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!((t_FmMacsecSecY *)h_FmMacsecSecY)->p_FmMacsecSecYDriverParam, E_INVALID_STATE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmSecYSc, E_INVALID_HANDLE);
-+#ifdef DISABLE_SANITY_CHECKS
-+ UNUSED(h_FmMacsecSecY);
-+#endif /* DISABLE_SANITY_CHECKS */
-+
-+ *p_ScPhysId = p_FmSecYSc->scId;
-+ return err;
-+}
-+
-+t_Error FM_MACSEC_SECY_GetTxScPhysId(t_Handle h_FmMacsecSecY, uint32_t *p_ScPhysId)
-+{
-+ t_FmMacsecSecY *p_FmMacsecSecY = (t_FmMacsecSecY *)h_FmMacsecSecY;
-+ t_SecYSc *p_FmSecYSc;
-+ t_Error err = E_OK;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsecSecY, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmMacsecSecY->h_FmMacsec, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmMacsecSecY->p_FmMacsecSecYDriverParam, E_INVALID_STATE);
-+ p_FmSecYSc = &p_FmMacsecSecY->p_TxSc[0];
-+ SANITY_CHECK_RETURN_ERROR(p_FmSecYSc, E_INVALID_HANDLE);
-+
-+ *p_ScPhysId = p_FmSecYSc->scId;
-+ return err;
-+}
-+
-+t_Error FM_MACSEC_SECY_SetException(t_Handle h_FmMacsecSecY, e_FmMacsecExceptions exception, bool enable)
-+{
-+ UNUSED(h_FmMacsecSecY);UNUSED(exception);UNUSED(enable);
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, NO_MSG);
-+}
-+
-+t_Error FM_MACSEC_SECY_SetEvent(t_Handle h_FmMacsecSecY, e_FmMacsecSecYEvents event, bool enable)
-+{
-+ UNUSED(h_FmMacsecSecY);UNUSED(event);UNUSED(enable);
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, NO_MSG);
-+}
-+
-+t_Error FM_MACSEC_SECY_GetStatistics(t_Handle h_FmMacsecSecY, t_FmMacsecSecYStatistics *p_Statistics)
-+{
-+ UNUSED(h_FmMacsecSecY);UNUSED(p_Statistics);
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, NO_MSG);
-+}
-+
-+t_Error FM_MACSEC_SECY_RxScGetStatistics(t_Handle h_FmMacsecSecY, t_Handle h_Sc, t_FmMacsecSecYRxScStatistics *p_Statistics)
-+{
-+ UNUSED(h_FmMacsecSecY);UNUSED(h_Sc);UNUSED(p_Statistics);
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, NO_MSG);
-+}
-+
-+t_Error FM_MACSEC_SECY_RxSaGetStatistics(t_Handle h_FmMacsecSecY, t_Handle h_Sc, macsecAN_t an, t_FmMacsecSecYRxSaStatistics *p_Statistics)
-+{
-+ UNUSED(h_FmMacsecSecY);UNUSED(h_Sc);UNUSED(an);UNUSED(p_Statistics);
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, NO_MSG);
-+}
-+
-+t_Error FM_MACSEC_SECY_TxScGetStatistics(t_Handle h_FmMacsecSecY, t_FmMacsecSecYTxScStatistics *p_Statistics)
-+{
-+ UNUSED(h_FmMacsecSecY);UNUSED(p_Statistics);
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, NO_MSG);
-+}
-+
-+t_Error FM_MACSEC_SECY_TxSaGetStatistics(t_Handle h_FmMacsecSecY, macsecAN_t an, t_FmMacsecSecYTxSaStatistics *p_Statistics)
-+{
-+ UNUSED(h_FmMacsecSecY);UNUSED(an);UNUSED(p_Statistics);
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, NO_MSG);
-+}
-+
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MACSEC/fm_macsec_secy.h
-@@ -0,0 +1,144 @@
-+/*
-+ * Copyright 2008-2015 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+/******************************************************************************
-+ @File fm_macsec_secy.h
-+
-+ @Description FM MACSEC SecY internal structures and definitions.
-+*//***************************************************************************/
-+#ifndef __FM_MACSEC_SECY_H
-+#define __FM_MACSEC_SECY_H
-+
-+#include "error_ext.h"
-+#include "std_ext.h"
-+
-+#include "fm_macsec.h"
-+
-+
-+/**************************************************************************//**
-+ @Description Exceptions
-+*//***************************************************************************/
-+
-+#define FM_MACSEC_SECY_EX_FRAME_DISCARDED 0x80000000
-+
-+#define GET_EXCEPTION_FLAG(bitMask, exception) switch (exception){ \
-+ case e_FM_MACSEC_SECY_EX_FRAME_DISCARDED: \
-+ bitMask = FM_MACSEC_SECY_EX_FRAME_DISCARDED; break; \
-+ default: bitMask = 0;break;}
-+
-+/**************************************************************************//**
-+ @Description Events
-+*//***************************************************************************/
-+
-+#define FM_MACSEC_SECY_EV_NEXT_PN 0x80000000
-+
-+#define GET_EVENT_FLAG(bitMask, event) switch (event){ \
-+ case e_FM_MACSEC_SECY_EV_NEXT_PN: \
-+ bitMask = FM_MACSEC_SECY_EV_NEXT_PN; break; \
-+ default: bitMask = 0;break;}
-+
-+/**************************************************************************//**
-+ @Description Defaults
-+*//***************************************************************************/
-+
-+#define DEFAULT_exceptions (FM_MACSEC_SECY_EX_FRAME_DISCARDED)
-+#define DEFAULT_events (FM_MACSEC_SECY_EV_NEXT_PN)
-+#define DEFAULT_numOfTxSc 1
-+#define DEFAULT_confidentialityEnable FALSE
-+#define DEFAULT_confidentialityOffset 0
-+#define DEFAULT_sciInsertionMode e_FM_MACSEC_SCI_INSERTION_MODE_EXPLICIT_SECTAG
-+#define DEFAULT_validateFrames e_FM_MACSEC_VALID_FRAME_BEHAVIOR_STRICT
-+#define DEFAULT_replayEnable FALSE
-+#define DEFAULT_replayWindow 0
-+#define DEFAULT_protectFrames TRUE
-+#define DEFAULT_ptp FALSE
-+
-+/**************************************************************************//**
-+ @Description General defines
-+*//***************************************************************************/
-+
-+#define SECY_AN_FREE_VALUE MAX_NUM_OF_SA_PER_SC
-+
-+
-+typedef struct {
-+ e_ScSaId saId;
-+ bool active;
-+ union {
-+ t_FmMacsecSecYRxSaStatistics rxSaStatistics;
-+ t_FmMacsecSecYTxSaStatistics txSaStatistics;
-+ };
-+} t_SecYSa;
-+
-+typedef struct {
-+ bool inUse;
-+ uint32_t scId;
-+ e_ScType type;
-+ uint8_t numOfSa;
-+ t_SecYSa sa[MAX_NUM_OF_SA_PER_SC];
-+ union {
-+ t_FmMacsecSecYRxScStatistics rxScStatistics;
-+ t_FmMacsecSecYTxScStatistics txScStatistics;
-+ };
-+} t_SecYSc;
-+
-+typedef struct {
-+ t_FmMacsecSecYSCParams txScParams; /**< Tx SC Params */
-+} t_FmMacsecSecYDriverParam;
-+
-+typedef struct {
-+ t_Handle h_FmMacsec;
-+ bool confidentialityEnable; /**< TRUE - confidentiality protection and integrity protection
-+ FALSE - no confidentiality protection, only integrity protection*/
-+ uint16_t confidentialityOffset; /**< The number of initial octets of each MSDU without confidentiality protection
-+ common values are 0, 30, and 50 */
-+ bool replayProtect; /**< replay protection function mode */
-+ uint32_t replayWindow; /**< the size of the replay window */
-+ e_FmMacsecValidFrameBehavior validateFrames; /**< validation function mode */
-+ e_FmMacsecSciInsertionMode sciInsertionMode;
-+ bool protectFrames;
-+ bool isPointToPoint;
-+ e_FmMacsecSecYCipherSuite cipherSuite; /**< Cipher suite to be used for this SecY */
-+ uint32_t numOfRxSc; /**< Number of receive channels */
-+ uint32_t numOfTxSc; /**< Number of transmit channels */
-+ t_SecYSc *p_RxSc;
-+ t_SecYSc *p_TxSc;
-+ uint32_t events;
-+ uint32_t exceptions;
-+ t_FmMacsecSecYExceptionsCallback *f_Exception; /**< TODO */
-+ t_FmMacsecSecYEventsCallback *f_Event; /**< TODO */
-+ t_Handle h_App;
-+ t_FmMacsecSecYStatistics statistics;
-+ t_FmMacsecSecYDriverParam *p_FmMacsecSecYDriverParam;
-+} t_FmMacsecSecY;
-+
-+
-+#endif /* __FM_MACSEC_SECY_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/Makefile
-@@ -0,0 +1,23 @@
-+#
-+# Makefile for the Freescale Ethernet controllers
-+#
-+ccflags-y += -DVERSION=\"\"
-+#
-+#Include netcomm SW specific definitions
-+include $(srctree)/drivers/net/ethernet/freescale/sdk_fman/ncsw_config.mk
-+NCSW_FM_INC = $(srctree)/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/inc
-+
-+ccflags-y += -I$(NCSW_FM_INC)
-+
-+
-+obj-y += fsl-ncsw-PFM1.o
-+
-+fsl-ncsw-PFM1-objs := fm.o fm_muram.o fman.o
-+
-+obj-y += MAC/
-+obj-y += Pcd/
-+obj-y += SP/
-+obj-y += Port/
-+obj-y += HC/
-+obj-y += Rtc/
-+obj-y += MACSEC/
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/Pcd/Makefile
-@@ -0,0 +1,26 @@
-+#
-+# Makefile for the Freescale Ethernet controllers
-+#
-+ccflags-y += -DVERSION=\"\"
-+#
-+#Include netcomm SW specific definitions
-+include $(srctree)/drivers/net/ethernet/freescale/sdk_fman/ncsw_config.mk
-+
-+NCSW_FM_INC = $(srctree)/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/inc
-+
-+ccflags-y += -I$(NCSW_FM_INC)
-+
-+obj-y += fsl-ncsw-Pcd.o
-+
-+fsl-ncsw-Pcd-objs := fman_kg.o fman_prs.o fm_cc.o fm_kg.o fm_pcd.o fm_plcr.o fm_prs.o fm_manip.o
-+
-+ifeq ($(CONFIG_FMAN_V3H),y)
-+fsl-ncsw-Pcd-objs += fm_replic.o
-+endif
-+ifeq ($(CONFIG_FMAN_V3L),y)
-+fsl-ncsw-Pcd-objs += fm_replic.o
-+endif
-+ifeq ($(CONFIG_FMAN_ARM),y)
-+fsl-ncsw-Pcd-objs += fm_replic.o
-+endif
-+
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/Pcd/crc64.h
-@@ -0,0 +1,360 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+ /**************************************************************************//**
-+ @File crc64.h
-+
-+ @Description brief This file contains the CRC64 Table, and __inline__
-+ functions used for calculating crc.
-+*//***************************************************************************/
-+#ifndef __CRC64_H
-+#define __CRC64_H
-+
-+#include "std_ext.h"
-+
-+
-+#define BITS_PER_BYTE 8
-+
-+#define CRC64_EXPON_ECMA_182 0xC96C5795D7870F42ULL
-+#define CRC64_DEFAULT_INITVAL 0xFFFFFFFFFFFFFFFFULL
-+
-+#define CRC64_BYTE_MASK 0xFF
-+#define CRC64_TABLE_ENTRIES ( 1 << BITS_PER_BYTE )
-+#define CRC64_ODD_MASK 1
-+
-+
-+/**
-+ \brief '64 bit crc' Table
-+ */
-+struct crc64_t {
-+ uint64_t initial; /**< Initial seed */
-+ uint64_t table[CRC64_TABLE_ENTRIES]; /**< CRC table entries */
-+};
-+
-+
-+static struct crc64_t CRC64_ECMA_182 = {
-+ CRC64_DEFAULT_INITVAL,
-+ {
-+ 0x0000000000000000ULL,
-+ 0xb32e4cbe03a75f6fULL,
-+ 0xf4843657a840a05bULL,
-+ 0x47aa7ae9abe7ff34ULL,
-+ 0x7bd0c384ff8f5e33ULL,
-+ 0xc8fe8f3afc28015cULL,
-+ 0x8f54f5d357cffe68ULL,
-+ 0x3c7ab96d5468a107ULL,
-+ 0xf7a18709ff1ebc66ULL,
-+ 0x448fcbb7fcb9e309ULL,
-+ 0x0325b15e575e1c3dULL,
-+ 0xb00bfde054f94352ULL,
-+ 0x8c71448d0091e255ULL,
-+ 0x3f5f08330336bd3aULL,
-+ 0x78f572daa8d1420eULL,
-+ 0xcbdb3e64ab761d61ULL,
-+ 0x7d9ba13851336649ULL,
-+ 0xceb5ed8652943926ULL,
-+ 0x891f976ff973c612ULL,
-+ 0x3a31dbd1fad4997dULL,
-+ 0x064b62bcaebc387aULL,
-+ 0xb5652e02ad1b6715ULL,
-+ 0xf2cf54eb06fc9821ULL,
-+ 0x41e11855055bc74eULL,
-+ 0x8a3a2631ae2dda2fULL,
-+ 0x39146a8fad8a8540ULL,
-+ 0x7ebe1066066d7a74ULL,
-+ 0xcd905cd805ca251bULL,
-+ 0xf1eae5b551a2841cULL,
-+ 0x42c4a90b5205db73ULL,
-+ 0x056ed3e2f9e22447ULL,
-+ 0xb6409f5cfa457b28ULL,
-+ 0xfb374270a266cc92ULL,
-+ 0x48190ecea1c193fdULL,
-+ 0x0fb374270a266cc9ULL,
-+ 0xbc9d3899098133a6ULL,
-+ 0x80e781f45de992a1ULL,
-+ 0x33c9cd4a5e4ecdceULL,
-+ 0x7463b7a3f5a932faULL,
-+ 0xc74dfb1df60e6d95ULL,
-+ 0x0c96c5795d7870f4ULL,
-+ 0xbfb889c75edf2f9bULL,
-+ 0xf812f32ef538d0afULL,
-+ 0x4b3cbf90f69f8fc0ULL,
-+ 0x774606fda2f72ec7ULL,
-+ 0xc4684a43a15071a8ULL,
-+ 0x83c230aa0ab78e9cULL,
-+ 0x30ec7c140910d1f3ULL,
-+ 0x86ace348f355aadbULL,
-+ 0x3582aff6f0f2f5b4ULL,
-+ 0x7228d51f5b150a80ULL,
-+ 0xc10699a158b255efULL,
-+ 0xfd7c20cc0cdaf4e8ULL,
-+ 0x4e526c720f7dab87ULL,
-+ 0x09f8169ba49a54b3ULL,
-+ 0xbad65a25a73d0bdcULL,
-+ 0x710d64410c4b16bdULL,
-+ 0xc22328ff0fec49d2ULL,
-+ 0x85895216a40bb6e6ULL,
-+ 0x36a71ea8a7ace989ULL,
-+ 0x0adda7c5f3c4488eULL,
-+ 0xb9f3eb7bf06317e1ULL,
-+ 0xfe5991925b84e8d5ULL,
-+ 0x4d77dd2c5823b7baULL,
-+ 0x64b62bcaebc387a1ULL,
-+ 0xd7986774e864d8ceULL,
-+ 0x90321d9d438327faULL,
-+ 0x231c512340247895ULL,
-+ 0x1f66e84e144cd992ULL,
-+ 0xac48a4f017eb86fdULL,
-+ 0xebe2de19bc0c79c9ULL,
-+ 0x58cc92a7bfab26a6ULL,
-+ 0x9317acc314dd3bc7ULL,
-+ 0x2039e07d177a64a8ULL,
-+ 0x67939a94bc9d9b9cULL,
-+ 0xd4bdd62abf3ac4f3ULL,
-+ 0xe8c76f47eb5265f4ULL,
-+ 0x5be923f9e8f53a9bULL,
-+ 0x1c4359104312c5afULL,
-+ 0xaf6d15ae40b59ac0ULL,
-+ 0x192d8af2baf0e1e8ULL,
-+ 0xaa03c64cb957be87ULL,
-+ 0xeda9bca512b041b3ULL,
-+ 0x5e87f01b11171edcULL,
-+ 0x62fd4976457fbfdbULL,
-+ 0xd1d305c846d8e0b4ULL,
-+ 0x96797f21ed3f1f80ULL,
-+ 0x2557339fee9840efULL,
-+ 0xee8c0dfb45ee5d8eULL,
-+ 0x5da24145464902e1ULL,
-+ 0x1a083bacedaefdd5ULL,
-+ 0xa9267712ee09a2baULL,
-+ 0x955cce7fba6103bdULL,
-+ 0x267282c1b9c65cd2ULL,
-+ 0x61d8f8281221a3e6ULL,
-+ 0xd2f6b4961186fc89ULL,
-+ 0x9f8169ba49a54b33ULL,
-+ 0x2caf25044a02145cULL,
-+ 0x6b055fede1e5eb68ULL,
-+ 0xd82b1353e242b407ULL,
-+ 0xe451aa3eb62a1500ULL,
-+ 0x577fe680b58d4a6fULL,
-+ 0x10d59c691e6ab55bULL,
-+ 0xa3fbd0d71dcdea34ULL,
-+ 0x6820eeb3b6bbf755ULL,
-+ 0xdb0ea20db51ca83aULL,
-+ 0x9ca4d8e41efb570eULL,
-+ 0x2f8a945a1d5c0861ULL,
-+ 0x13f02d374934a966ULL,
-+ 0xa0de61894a93f609ULL,
-+ 0xe7741b60e174093dULL,
-+ 0x545a57dee2d35652ULL,
-+ 0xe21ac88218962d7aULL,
-+ 0x5134843c1b317215ULL,
-+ 0x169efed5b0d68d21ULL,
-+ 0xa5b0b26bb371d24eULL,
-+ 0x99ca0b06e7197349ULL,
-+ 0x2ae447b8e4be2c26ULL,
-+ 0x6d4e3d514f59d312ULL,
-+ 0xde6071ef4cfe8c7dULL,
-+ 0x15bb4f8be788911cULL,
-+ 0xa6950335e42fce73ULL,
-+ 0xe13f79dc4fc83147ULL,
-+ 0x521135624c6f6e28ULL,
-+ 0x6e6b8c0f1807cf2fULL,
-+ 0xdd45c0b11ba09040ULL,
-+ 0x9aefba58b0476f74ULL,
-+ 0x29c1f6e6b3e0301bULL,
-+ 0xc96c5795d7870f42ULL,
-+ 0x7a421b2bd420502dULL,
-+ 0x3de861c27fc7af19ULL,
-+ 0x8ec62d7c7c60f076ULL,
-+ 0xb2bc941128085171ULL,
-+ 0x0192d8af2baf0e1eULL,
-+ 0x4638a2468048f12aULL,
-+ 0xf516eef883efae45ULL,
-+ 0x3ecdd09c2899b324ULL,
-+ 0x8de39c222b3eec4bULL,
-+ 0xca49e6cb80d9137fULL,
-+ 0x7967aa75837e4c10ULL,
-+ 0x451d1318d716ed17ULL,
-+ 0xf6335fa6d4b1b278ULL,
-+ 0xb199254f7f564d4cULL,
-+ 0x02b769f17cf11223ULL,
-+ 0xb4f7f6ad86b4690bULL,
-+ 0x07d9ba1385133664ULL,
-+ 0x4073c0fa2ef4c950ULL,
-+ 0xf35d8c442d53963fULL,
-+ 0xcf273529793b3738ULL,
-+ 0x7c0979977a9c6857ULL,
-+ 0x3ba3037ed17b9763ULL,
-+ 0x888d4fc0d2dcc80cULL,
-+ 0x435671a479aad56dULL,
-+ 0xf0783d1a7a0d8a02ULL,
-+ 0xb7d247f3d1ea7536ULL,
-+ 0x04fc0b4dd24d2a59ULL,
-+ 0x3886b22086258b5eULL,
-+ 0x8ba8fe9e8582d431ULL,
-+ 0xcc0284772e652b05ULL,
-+ 0x7f2cc8c92dc2746aULL,
-+ 0x325b15e575e1c3d0ULL,
-+ 0x8175595b76469cbfULL,
-+ 0xc6df23b2dda1638bULL,
-+ 0x75f16f0cde063ce4ULL,
-+ 0x498bd6618a6e9de3ULL,
-+ 0xfaa59adf89c9c28cULL,
-+ 0xbd0fe036222e3db8ULL,
-+ 0x0e21ac88218962d7ULL,
-+ 0xc5fa92ec8aff7fb6ULL,
-+ 0x76d4de52895820d9ULL,
-+ 0x317ea4bb22bfdfedULL,
-+ 0x8250e80521188082ULL,
-+ 0xbe2a516875702185ULL,
-+ 0x0d041dd676d77eeaULL,
-+ 0x4aae673fdd3081deULL,
-+ 0xf9802b81de97deb1ULL,
-+ 0x4fc0b4dd24d2a599ULL,
-+ 0xfceef8632775faf6ULL,
-+ 0xbb44828a8c9205c2ULL,
-+ 0x086ace348f355aadULL,
-+ 0x34107759db5dfbaaULL,
-+ 0x873e3be7d8faa4c5ULL,
-+ 0xc094410e731d5bf1ULL,
-+ 0x73ba0db070ba049eULL,
-+ 0xb86133d4dbcc19ffULL,
-+ 0x0b4f7f6ad86b4690ULL,
-+ 0x4ce50583738cb9a4ULL,
-+ 0xffcb493d702be6cbULL,
-+ 0xc3b1f050244347ccULL,
-+ 0x709fbcee27e418a3ULL,
-+ 0x3735c6078c03e797ULL,
-+ 0x841b8ab98fa4b8f8ULL,
-+ 0xadda7c5f3c4488e3ULL,
-+ 0x1ef430e13fe3d78cULL,
-+ 0x595e4a08940428b8ULL,
-+ 0xea7006b697a377d7ULL,
-+ 0xd60abfdbc3cbd6d0ULL,
-+ 0x6524f365c06c89bfULL,
-+ 0x228e898c6b8b768bULL,
-+ 0x91a0c532682c29e4ULL,
-+ 0x5a7bfb56c35a3485ULL,
-+ 0xe955b7e8c0fd6beaULL,
-+ 0xaeffcd016b1a94deULL,
-+ 0x1dd181bf68bdcbb1ULL,
-+ 0x21ab38d23cd56ab6ULL,
-+ 0x9285746c3f7235d9ULL,
-+ 0xd52f0e859495caedULL,
-+ 0x6601423b97329582ULL,
-+ 0xd041dd676d77eeaaULL,
-+ 0x636f91d96ed0b1c5ULL,
-+ 0x24c5eb30c5374ef1ULL,
-+ 0x97eba78ec690119eULL,
-+ 0xab911ee392f8b099ULL,
-+ 0x18bf525d915feff6ULL,
-+ 0x5f1528b43ab810c2ULL,
-+ 0xec3b640a391f4fadULL,
-+ 0x27e05a6e926952ccULL,
-+ 0x94ce16d091ce0da3ULL,
-+ 0xd3646c393a29f297ULL,
-+ 0x604a2087398eadf8ULL,
-+ 0x5c3099ea6de60cffULL,
-+ 0xef1ed5546e415390ULL,
-+ 0xa8b4afbdc5a6aca4ULL,
-+ 0x1b9ae303c601f3cbULL,
-+ 0x56ed3e2f9e224471ULL,
-+ 0xe5c372919d851b1eULL,
-+ 0xa26908783662e42aULL,
-+ 0x114744c635c5bb45ULL,
-+ 0x2d3dfdab61ad1a42ULL,
-+ 0x9e13b115620a452dULL,
-+ 0xd9b9cbfcc9edba19ULL,
-+ 0x6a978742ca4ae576ULL,
-+ 0xa14cb926613cf817ULL,
-+ 0x1262f598629ba778ULL,
-+ 0x55c88f71c97c584cULL,
-+ 0xe6e6c3cfcadb0723ULL,
-+ 0xda9c7aa29eb3a624ULL,
-+ 0x69b2361c9d14f94bULL,
-+ 0x2e184cf536f3067fULL,
-+ 0x9d36004b35545910ULL,
-+ 0x2b769f17cf112238ULL,
-+ 0x9858d3a9ccb67d57ULL,
-+ 0xdff2a94067518263ULL,
-+ 0x6cdce5fe64f6dd0cULL,
-+ 0x50a65c93309e7c0bULL,
-+ 0xe388102d33392364ULL,
-+ 0xa4226ac498dedc50ULL,
-+ 0x170c267a9b79833fULL,
-+ 0xdcd7181e300f9e5eULL,
-+ 0x6ff954a033a8c131ULL,
-+ 0x28532e49984f3e05ULL,
-+ 0x9b7d62f79be8616aULL,
-+ 0xa707db9acf80c06dULL,
-+ 0x14299724cc279f02ULL,
-+ 0x5383edcd67c06036ULL,
-+ 0xe0ada17364673f59ULL
-+ }
-+};
-+
-+
-+/**
-+ \brief Initializes the crc seed
-+ */
-+static __inline__ uint64_t crc64_init(void)
-+{
-+ return CRC64_ECMA_182.initial;
-+}
-+
-+/**
-+ \brief Computes 64 bit the crc
-+ \param[in] data Pointer to the Data in the frame
-+ \param[in] len Length of the Data
-+ \param[in] crc seed
-+ \return calculated crc
-+ */
-+static __inline__ uint64_t crc64_compute(void const *data,
-+ uint32_t len,
-+ uint64_t seed)
-+{
-+ uint32_t i;
-+ uint64_t crc = seed;
-+ uint8_t *bdata = (uint8_t *) data;
-+
-+ for (i = 0; i < len; i++)
-+ crc =
-+ CRC64_ECMA_182.
-+ table[(crc ^ *bdata++) & CRC64_BYTE_MASK] ^ (crc >> 8);
-+
-+ return crc;
-+}
-+
-+
-+#endif /* __CRC64_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/Pcd/fm_cc.c
-@@ -0,0 +1,7582 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/******************************************************************************
-+ @File fm_cc.c
-+
-+ @Description FM Coarse Classifier implementation
-+ *//***************************************************************************/
-+#include <linux/math64.h>
-+#include "std_ext.h"
-+#include "error_ext.h"
-+#include "string_ext.h"
-+#include "debug_ext.h"
-+#include "fm_pcd_ext.h"
-+#include "fm_muram_ext.h"
-+
-+#include "fm_common.h"
-+#include "fm_pcd.h"
-+#include "fm_hc.h"
-+#include "fm_cc.h"
-+#include "crc64.h"
-+
-+/****************************************/
-+/* static functions */
-+/****************************************/
-+
-+
-+static t_Error CcRootTryLock(t_Handle h_FmPcdCcTree)
-+{
-+ t_FmPcdCcTree *p_FmPcdCcTree = (t_FmPcdCcTree *)h_FmPcdCcTree;
-+
-+ ASSERT_COND(h_FmPcdCcTree);
-+
-+ if (FmPcdLockTryLock(p_FmPcdCcTree->p_Lock))
-+ return E_OK;
-+
-+ return ERROR_CODE(E_BUSY);
-+}
-+
-+static void CcRootReleaseLock(t_Handle h_FmPcdCcTree)
-+{
-+ t_FmPcdCcTree *p_FmPcdCcTree = (t_FmPcdCcTree *)h_FmPcdCcTree;
-+
-+ ASSERT_COND(h_FmPcdCcTree);
-+
-+ FmPcdLockUnlock(p_FmPcdCcTree->p_Lock);
-+}
-+
-+static void UpdateNodeOwner(t_FmPcdCcNode *p_CcNode, bool add)
-+{
-+ uint32_t intFlags;
-+
-+ ASSERT_COND(p_CcNode);
-+
-+ intFlags = XX_LockIntrSpinlock(p_CcNode->h_Spinlock);
-+
-+ if (add)
-+ p_CcNode->owners++;
-+ else
-+ {
-+ ASSERT_COND(p_CcNode->owners);
-+ p_CcNode->owners--;
-+ }
-+
-+ XX_UnlockIntrSpinlock(p_CcNode->h_Spinlock, intFlags);
-+}
-+
-+static __inline__ t_FmPcdStatsObj* DequeueStatsObj(t_List *p_List)
-+{
-+ t_FmPcdStatsObj *p_StatsObj = NULL;
-+ t_List *p_Next;
-+
-+ if (!LIST_IsEmpty(p_List))
-+ {
-+ p_Next = LIST_FIRST(p_List);
-+ p_StatsObj = LIST_OBJECT(p_Next, t_FmPcdStatsObj, node);
-+ ASSERT_COND(p_StatsObj);
-+ LIST_DelAndInit(p_Next);
-+ }
-+
-+ return p_StatsObj;
-+}
-+
-+static __inline__ void EnqueueStatsObj(t_List *p_List,
-+ t_FmPcdStatsObj *p_StatsObj)
-+{
-+ LIST_AddToTail(&p_StatsObj->node, p_List);
-+}
-+
-+static void FreeStatObjects(t_List *p_List, t_Handle h_FmMuram)
-+{
-+ t_FmPcdStatsObj *p_StatsObj;
-+
-+ while (!LIST_IsEmpty(p_List))
-+ {
-+ p_StatsObj = DequeueStatsObj(p_List);
-+ ASSERT_COND(p_StatsObj);
-+
-+ FM_MURAM_FreeMem(h_FmMuram, p_StatsObj->h_StatsAd);
-+ FM_MURAM_FreeMem(h_FmMuram, p_StatsObj->h_StatsCounters);
-+
-+ XX_Free(p_StatsObj);
-+ }
-+}
-+
-+static t_FmPcdStatsObj* GetStatsObj(t_FmPcdCcNode *p_CcNode)
-+{
-+ t_FmPcdStatsObj* p_StatsObj;
-+ t_Handle h_FmMuram;
-+
-+ ASSERT_COND(p_CcNode);
-+
-+ /* If 'maxNumOfKeys' was passed, all statistics object were preallocated
-+ upon node initialization */
-+ if (p_CcNode->maxNumOfKeys)
-+ {
-+ p_StatsObj = DequeueStatsObj(&p_CcNode->availableStatsLst);
-+
-+ /* Clean statistics counters & ADs */
-+ MemSet8(p_StatsObj->h_StatsAd, 0, FM_PCD_CC_AD_ENTRY_SIZE);
-+ MemSet8(p_StatsObj->h_StatsCounters, 0, p_CcNode->countersArraySize);
-+ }
-+ else
-+ {
-+ h_FmMuram = ((t_FmPcd *)(p_CcNode->h_FmPcd))->h_FmMuram;
-+ ASSERT_COND(h_FmMuram);
-+
-+ p_StatsObj = XX_Malloc(sizeof(t_FmPcdStatsObj));
-+ if (!p_StatsObj)
-+ {
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("statistics object"));
-+ return NULL;
-+ }
-+
-+ p_StatsObj->h_StatsAd = (t_Handle)FM_MURAM_AllocMem(
-+ h_FmMuram, FM_PCD_CC_AD_ENTRY_SIZE, FM_PCD_CC_AD_TABLE_ALIGN);
-+ if (!p_StatsObj->h_StatsAd)
-+ {
-+ XX_Free(p_StatsObj);
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("MURAM allocation for statistics ADs"));
-+ return NULL;
-+ }
-+ MemSet8(p_StatsObj->h_StatsAd, 0, FM_PCD_CC_AD_ENTRY_SIZE);
-+
-+ p_StatsObj->h_StatsCounters = (t_Handle)FM_MURAM_AllocMem(
-+ h_FmMuram, p_CcNode->countersArraySize,
-+ FM_PCD_CC_AD_TABLE_ALIGN);
-+ if (!p_StatsObj->h_StatsCounters)
-+ {
-+ FM_MURAM_FreeMem(h_FmMuram, p_StatsObj->h_StatsAd);
-+ XX_Free(p_StatsObj);
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("MURAM allocation for statistics counters"));
-+ return NULL;
-+ }
-+ MemSet8(p_StatsObj->h_StatsCounters, 0, p_CcNode->countersArraySize);
-+ }
-+
-+ return p_StatsObj;
-+}
-+
-+static void PutStatsObj(t_FmPcdCcNode *p_CcNode, t_FmPcdStatsObj *p_StatsObj)
-+{
-+ t_Handle h_FmMuram;
-+
-+ ASSERT_COND(p_CcNode);
-+ ASSERT_COND(p_StatsObj);
-+
-+ /* If 'maxNumOfKeys' was passed, all statistics object were preallocated
-+ upon node initialization and now will be enqueued back to the list */
-+ if (p_CcNode->maxNumOfKeys)
-+ {
-+ /* Clean statistics counters */
-+ MemSet8(p_StatsObj->h_StatsCounters, 0, p_CcNode->countersArraySize);
-+
-+ /* Clean statistics ADs */
-+ MemSet8(p_StatsObj->h_StatsAd, 0, FM_PCD_CC_AD_ENTRY_SIZE);
-+
-+ EnqueueStatsObj(&p_CcNode->availableStatsLst, p_StatsObj);
-+ }
-+ else
-+ {
-+ h_FmMuram = ((t_FmPcd *)(p_CcNode->h_FmPcd))->h_FmMuram;
-+ ASSERT_COND(h_FmMuram);
-+
-+ FM_MURAM_FreeMem(h_FmMuram, p_StatsObj->h_StatsAd);
-+ FM_MURAM_FreeMem(h_FmMuram, p_StatsObj->h_StatsCounters);
-+
-+ XX_Free(p_StatsObj);
-+ }
-+}
-+
-+static void SetStatsCounters(t_AdOfTypeStats *p_StatsAd,
-+ uint32_t statsCountersAddr)
-+{
-+ uint32_t tmp = (statsCountersAddr & FM_PCD_AD_STATS_COUNTERS_ADDR_MASK);
-+
-+ WRITE_UINT32(p_StatsAd->statsTableAddr, tmp);
-+}
-+
-+
-+static void UpdateStatsAd(t_FmPcdCcStatsParams *p_FmPcdCcStatsParams,
-+ t_Handle h_Ad, uint64_t physicalMuramBase)
-+{
-+ t_AdOfTypeStats *p_StatsAd;
-+ uint32_t statsCountersAddr, nextActionAddr, tmp;
-+#if (DPAA_VERSION >= 11)
-+ uint32_t frameLengthRangesAddr;
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+ p_StatsAd = (t_AdOfTypeStats *)p_FmPcdCcStatsParams->h_StatsAd;
-+
-+ tmp = FM_PCD_AD_STATS_TYPE;
-+
-+#if (DPAA_VERSION >= 11)
-+ if (p_FmPcdCcStatsParams->h_StatsFLRs)
-+ {
-+ frameLengthRangesAddr = (uint32_t)((XX_VirtToPhys(
-+ p_FmPcdCcStatsParams->h_StatsFLRs) - physicalMuramBase));
-+ tmp |= (frameLengthRangesAddr & FM_PCD_AD_STATS_FLR_ADDR_MASK);
-+ }
-+#endif /* (DPAA_VERSION >= 11) */
-+ WRITE_UINT32(p_StatsAd->profileTableAddr, tmp);
-+
-+ nextActionAddr = (uint32_t)((XX_VirtToPhys(h_Ad) - physicalMuramBase));
-+ tmp = 0;
-+ tmp |= (uint32_t)((nextActionAddr << FM_PCD_AD_STATS_NEXT_ACTION_SHIFT)
-+ & FM_PCD_AD_STATS_NEXT_ACTION_MASK);
-+ tmp |= (FM_PCD_AD_STATS_NAD_EN | FM_PCD_AD_STATS_OP_CODE);
-+
-+#if (DPAA_VERSION >= 11)
-+ if (p_FmPcdCcStatsParams->h_StatsFLRs)
-+ tmp |= FM_PCD_AD_STATS_FLR_EN;
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+ WRITE_UINT32(p_StatsAd->nextActionIndx, tmp);
-+
-+ statsCountersAddr = (uint32_t)((XX_VirtToPhys(
-+ p_FmPcdCcStatsParams->h_StatsCounters) - physicalMuramBase));
-+ SetStatsCounters(p_StatsAd, statsCountersAddr);
-+}
-+
-+static void FillAdOfTypeContLookup(t_Handle h_Ad,
-+ t_FmPcdCcStatsParams *p_FmPcdCcStatsParams,
-+ t_Handle h_FmPcd, t_Handle p_CcNode,
-+ t_Handle h_Manip, t_Handle h_FrmReplic)
-+{
-+ t_FmPcdCcNode *p_Node = (t_FmPcdCcNode *)p_CcNode;
-+ t_AdOfTypeContLookup *p_AdContLookup = (t_AdOfTypeContLookup *)h_Ad;
-+ t_Handle h_TmpAd;
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+ uint32_t tmpReg32;
-+ t_Handle p_AdNewPtr = NULL;
-+
-+ UNUSED(h_Manip);
-+ UNUSED(h_FrmReplic);
-+
-+ /* there are 3 cases handled in this routine of building a "Continue lookup" type AD.
-+ * Case 1: No Manip. The action descriptor is built within the match table.
-+ * p_AdResult = p_AdNewPtr;
-+ * Case 2: Manip exists. A new AD is created - p_AdNewPtr. It is initialized
-+ * either in the FmPcdManipUpdateAdResultForCc routine or it was already
-+ * initialized and returned here.
-+ * p_AdResult (within the match table) will be initialized after
-+ * this routine returns and point to the existing AD.
-+ * Case 3: Manip exists. The action descriptor is built within the match table.
-+ * FmPcdManipUpdateAdContLookupForCc returns a NULL p_AdNewPtr.
-+ */
-+
-+ /* As default, the "new" ptr is the current one. i.e. the content of the result
-+ * AD will be written into the match table itself (case (1))*/
-+ p_AdNewPtr = p_AdContLookup;
-+
-+ /* Initialize an action descriptor, if current statistics mode requires an Ad */
-+ if (p_FmPcdCcStatsParams)
-+ {
-+ ASSERT_COND(p_FmPcdCcStatsParams->h_StatsAd);
-+ ASSERT_COND(p_FmPcdCcStatsParams->h_StatsCounters);
-+
-+ /* Swapping addresses between statistics Ad and the current lookup AD */
-+ h_TmpAd = p_FmPcdCcStatsParams->h_StatsAd;
-+ p_FmPcdCcStatsParams->h_StatsAd = h_Ad;
-+ h_Ad = h_TmpAd;
-+
-+ p_AdNewPtr = h_Ad;
-+ p_AdContLookup = h_Ad;
-+
-+ /* Init statistics Ad and connect current lookup AD as 'next action' from statistics Ad */
-+ UpdateStatsAd(p_FmPcdCcStatsParams, h_Ad, p_FmPcd->physicalMuramBase);
-+ }
-+
-+#if DPAA_VERSION >= 11
-+ if (h_Manip && h_FrmReplic)
-+ FmPcdManipUpdateAdContLookupForCc(
-+ h_Manip,
-+ h_Ad,
-+ &p_AdNewPtr,
-+ (uint32_t)((XX_VirtToPhys(
-+ FrmReplicGroupGetSourceTableDescriptor(h_FrmReplic))
-+ - p_FmPcd->physicalMuramBase)));
-+ else
-+ if (h_FrmReplic)
-+ FrmReplicGroupUpdateAd(h_FrmReplic, h_Ad, &p_AdNewPtr);
-+ else
-+#endif /* (DPAA_VERSION >= 11) */
-+ if (h_Manip)
-+ FmPcdManipUpdateAdContLookupForCc(
-+ h_Manip,
-+ h_Ad,
-+ &p_AdNewPtr,
-+
-+#ifdef FM_CAPWAP_SUPPORT
-+ /*no check for opcode of manip - this step can be reached only with capwap_applic_specific*/
-+ (uint32_t)((XX_VirtToPhys(p_Node->h_AdTable) - p_FmPcd->physicalMuramBase))
-+#else /* not FM_CAPWAP_SUPPORT */
-+ (uint32_t)((XX_VirtToPhys(p_Node->h_Ad)
-+ - p_FmPcd->physicalMuramBase))
-+#endif /* not FM_CAPWAP_SUPPORT */
-+ );
-+
-+ /* if (p_AdNewPtr = NULL) --> Done. (case (3)) */
-+ if (p_AdNewPtr)
-+ {
-+ /* cases (1) & (2) */
-+ tmpReg32 = 0;
-+ tmpReg32 |= FM_PCD_AD_CONT_LOOKUP_TYPE;
-+ tmpReg32 |=
-+ p_Node->sizeOfExtraction ? ((p_Node->sizeOfExtraction - 1) << 24) :
-+ 0;
-+ tmpReg32 |= (uint32_t)(XX_VirtToPhys(p_Node->h_AdTable)
-+ - p_FmPcd->physicalMuramBase);
-+ WRITE_UINT32(p_AdContLookup->ccAdBase, tmpReg32);
-+
-+ tmpReg32 = 0;
-+ tmpReg32 |= p_Node->numOfKeys << 24;
-+ tmpReg32 |= (p_Node->lclMask ? FM_PCD_AD_CONT_LOOKUP_LCL_MASK : 0);
-+ tmpReg32 |=
-+ p_Node->h_KeysMatchTable ? (uint32_t)(XX_VirtToPhys(
-+ p_Node->h_KeysMatchTable) - p_FmPcd->physicalMuramBase) :
-+ 0;
-+ WRITE_UINT32(p_AdContLookup->matchTblPtr, tmpReg32);
-+
-+ tmpReg32 = 0;
-+ tmpReg32 |= p_Node->prsArrayOffset << 24;
-+ tmpReg32 |= p_Node->offset << 16;
-+ tmpReg32 |= p_Node->parseCode;
-+ WRITE_UINT32(p_AdContLookup->pcAndOffsets, tmpReg32);
-+
-+ MemCpy8((void*)&p_AdContLookup->gmask, p_Node->p_GlblMask,
-+ CC_GLBL_MASK_SIZE);
-+ }
-+}
-+
-+static t_Error AllocAndFillAdForContLookupManip(t_Handle h_CcNode)
-+{
-+ t_FmPcdCcNode *p_CcNode = (t_FmPcdCcNode *)h_CcNode;
-+ uint32_t intFlags;
-+
-+ ASSERT_COND(p_CcNode);
-+
-+ intFlags = XX_LockIntrSpinlock(p_CcNode->h_Spinlock);
-+
-+ if (!p_CcNode->h_Ad)
-+ {
-+ if (p_CcNode->maxNumOfKeys)
-+ p_CcNode->h_Ad = p_CcNode->h_TmpAd;
-+ else
-+ p_CcNode->h_Ad = (t_Handle)FM_MURAM_AllocMem(
-+ ((t_FmPcd *)(p_CcNode->h_FmPcd))->h_FmMuram,
-+ FM_PCD_CC_AD_ENTRY_SIZE, FM_PCD_CC_AD_TABLE_ALIGN);
-+
-+ XX_UnlockIntrSpinlock(p_CcNode->h_Spinlock, intFlags);
-+
-+ if (!p_CcNode->h_Ad)
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY,
-+ ("MURAM allocation for CC action descriptor"));
-+
-+ MemSet8(p_CcNode->h_Ad, 0, FM_PCD_CC_AD_ENTRY_SIZE);
-+
-+ FillAdOfTypeContLookup(p_CcNode->h_Ad, NULL, p_CcNode->h_FmPcd,
-+ p_CcNode, NULL, NULL);
-+ }
-+ else
-+ XX_UnlockIntrSpinlock(p_CcNode->h_Spinlock, intFlags);
-+
-+ return E_OK;
-+}
-+
-+static t_Error SetRequiredAction1(
-+ t_Handle h_FmPcd, uint32_t requiredAction,
-+ t_FmPcdCcKeyAndNextEngineParams *p_CcKeyAndNextEngineParamsTmp,
-+ t_Handle h_AdTmp, uint16_t numOfEntries, t_Handle h_Tree)
-+{
-+ t_AdOfTypeResult *p_AdTmp = (t_AdOfTypeResult *)h_AdTmp;
-+ uint32_t tmpReg32;
-+ t_Error err;
-+ t_FmPcdCcNode *p_CcNode;
-+ int i = 0;
-+ uint16_t tmp = 0;
-+ uint16_t profileId;
-+ uint8_t relativeSchemeId, physicalSchemeId;
-+ t_CcNodeInformation ccNodeInfo;
-+
-+ for (i = 0; i < numOfEntries; i++)
-+ {
-+ if (i == 0)
-+ h_AdTmp = PTR_MOVE(h_AdTmp, i*FM_PCD_CC_AD_ENTRY_SIZE);
-+ else
-+ h_AdTmp = PTR_MOVE(h_AdTmp, FM_PCD_CC_AD_ENTRY_SIZE);
-+
-+ switch (p_CcKeyAndNextEngineParamsTmp[i].nextEngineParams.nextEngine)
-+ {
-+ case (e_FM_PCD_CC):
-+ if (requiredAction)
-+ {
-+ p_CcNode =
-+ p_CcKeyAndNextEngineParamsTmp[i].nextEngineParams.params.ccParams.h_CcNode;
-+ ASSERT_COND(p_CcNode);
-+ if (p_CcNode->shadowAction == requiredAction)
-+ break;
-+ if ((requiredAction & UPDATE_CC_WITH_TREE)
-+ && !(p_CcNode->shadowAction & UPDATE_CC_WITH_TREE))
-+ {
-+
-+ memset(&ccNodeInfo, 0, sizeof(t_CcNodeInformation));
-+ ccNodeInfo.h_CcNode = h_Tree;
-+ EnqueueNodeInfoToRelevantLst(&p_CcNode->ccTreesLst,
-+ &ccNodeInfo, NULL);
-+ p_CcKeyAndNextEngineParamsTmp[i].shadowAction |=
-+ UPDATE_CC_WITH_TREE;
-+ }
-+ if ((requiredAction & UPDATE_CC_SHADOW_CLEAR)
-+ && !(p_CcNode->shadowAction & UPDATE_CC_SHADOW_CLEAR))
-+ {
-+
-+ p_CcNode->shadowAction = 0;
-+ }
-+
-+ if ((requiredAction & UPDATE_CC_WITH_DELETE_TREE)
-+ && !(p_CcNode->shadowAction
-+ & UPDATE_CC_WITH_DELETE_TREE))
-+ {
-+ DequeueNodeInfoFromRelevantLst(&p_CcNode->ccTreesLst,
-+ h_Tree, NULL);
-+ p_CcKeyAndNextEngineParamsTmp[i].shadowAction |=
-+ UPDATE_CC_WITH_DELETE_TREE;
-+ }
-+ if (p_CcNode->keyAndNextEngineParams[p_CcNode->numOfKeys].nextEngineParams.nextEngine
-+ != e_FM_PCD_INVALID)
-+ tmp = (uint8_t)(p_CcNode->numOfKeys + 1);
-+ else
-+ tmp = p_CcNode->numOfKeys;
-+ err = SetRequiredAction1(h_FmPcd, requiredAction,
-+ p_CcNode->keyAndNextEngineParams,
-+ p_CcNode->h_AdTable, tmp, h_Tree);
-+ if (err != E_OK)
-+ return err;
-+ if (requiredAction != UPDATE_CC_SHADOW_CLEAR)
-+ p_CcNode->shadowAction |= requiredAction;
-+ }
-+ break;
-+
-+ case (e_FM_PCD_KG):
-+ if ((requiredAction & UPDATE_NIA_ENQ_WITHOUT_DMA)
-+ && !(p_CcKeyAndNextEngineParamsTmp[i].shadowAction
-+ & UPDATE_NIA_ENQ_WITHOUT_DMA))
-+ {
-+ physicalSchemeId =
-+ FmPcdKgGetSchemeId(
-+ p_CcKeyAndNextEngineParamsTmp[i].nextEngineParams.params.kgParams.h_DirectScheme);
-+ relativeSchemeId = FmPcdKgGetRelativeSchemeId(
-+ h_FmPcd, physicalSchemeId);
-+ if (relativeSchemeId == FM_PCD_KG_NUM_OF_SCHEMES)
-+ RETURN_ERROR(MAJOR, E_NOT_IN_RANGE, NO_MSG);
-+ if (!FmPcdKgIsSchemeValidSw(
-+ p_CcKeyAndNextEngineParamsTmp[i].nextEngineParams.params.kgParams.h_DirectScheme))
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE,
-+ ("Invalid direct scheme."));
-+ if (!KgIsSchemeAlwaysDirect(h_FmPcd, relativeSchemeId))
-+ RETURN_ERROR(
-+ MAJOR, E_INVALID_STATE,
-+ ("For this action scheme has to be direct."));
-+ err =
-+ FmPcdKgCcGetSetParams(
-+ h_FmPcd,
-+ p_CcKeyAndNextEngineParamsTmp[i].nextEngineParams.params.kgParams.h_DirectScheme,
-+ requiredAction, 0);
-+ if (err != E_OK)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ p_CcKeyAndNextEngineParamsTmp[i].shadowAction |=
-+ requiredAction;
-+ }
-+ break;
-+
-+ case (e_FM_PCD_PLCR):
-+ if ((requiredAction & UPDATE_NIA_ENQ_WITHOUT_DMA)
-+ && !(p_CcKeyAndNextEngineParamsTmp[i].shadowAction
-+ & UPDATE_NIA_ENQ_WITHOUT_DMA))
-+ {
-+ if (!p_CcKeyAndNextEngineParamsTmp[i].nextEngineParams.params.plcrParams.overrideParams)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_NOT_SUPPORTED,
-+ ("In this initialization only overrideFqid can be initialized"));
-+ if (!p_CcKeyAndNextEngineParamsTmp[i].nextEngineParams.params.plcrParams.sharedProfile)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_NOT_SUPPORTED,
-+ ("In this initialization only overrideFqid can be initialized"));
-+ err =
-+ FmPcdPlcrGetAbsoluteIdByProfileParams(
-+ h_FmPcd,
-+ e_FM_PCD_PLCR_SHARED,
-+ NULL,
-+ p_CcKeyAndNextEngineParamsTmp[i].nextEngineParams.params.plcrParams.newRelativeProfileId,
-+ &profileId);
-+ if (err != E_OK)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ err = FmPcdPlcrCcGetSetParams(h_FmPcd, profileId,
-+ requiredAction);
-+ if (err != E_OK)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ p_CcKeyAndNextEngineParamsTmp[i].shadowAction |=
-+ requiredAction;
-+ }
-+ break;
-+
-+ case (e_FM_PCD_DONE):
-+ if ((requiredAction & UPDATE_NIA_ENQ_WITHOUT_DMA)
-+ && !(p_CcKeyAndNextEngineParamsTmp[i].shadowAction
-+ & UPDATE_NIA_ENQ_WITHOUT_DMA))
-+ {
-+ tmpReg32 = GET_UINT32(p_AdTmp->nia);
-+ if ((tmpReg32 & GET_NIA_BMI_AC_ENQ_FRAME(h_FmPcd))
-+ != GET_NIA_BMI_AC_ENQ_FRAME(h_FmPcd))
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_STATE,
-+ ("Next engine was previously assigned not as PCD_DONE"));
-+ tmpReg32 |= NIA_BMI_AC_ENQ_FRAME_WITHOUT_DMA;
-+ WRITE_UINT32(p_AdTmp->nia, tmpReg32);
-+ p_CcKeyAndNextEngineParamsTmp[i].shadowAction |=
-+ requiredAction;
-+ }
-+ break;
-+
-+ default:
-+ break;
-+ }
-+ }
-+
-+ return E_OK;
-+}
-+
-+static t_Error SetRequiredAction(
-+ t_Handle h_FmPcd, uint32_t requiredAction,
-+ t_FmPcdCcKeyAndNextEngineParams *p_CcKeyAndNextEngineParamsTmp,
-+ t_Handle h_AdTmp, uint16_t numOfEntries, t_Handle h_Tree)
-+{
-+ t_Error err = SetRequiredAction1(h_FmPcd, requiredAction,
-+ p_CcKeyAndNextEngineParamsTmp, h_AdTmp,
-+ numOfEntries, h_Tree);
-+ if (err != E_OK)
-+ return err;
-+ return SetRequiredAction1(h_FmPcd, UPDATE_CC_SHADOW_CLEAR,
-+ p_CcKeyAndNextEngineParamsTmp, h_AdTmp,
-+ numOfEntries, h_Tree);
-+}
-+
-+static t_Error ReleaseModifiedDataStructure(
-+ t_Handle h_FmPcd, t_List *h_FmPcdOldPointersLst,
-+ t_List *h_FmPcdNewPointersLst,
-+ t_FmPcdModifyCcKeyAdditionalParams *p_AdditionalParams,
-+ bool useShadowStructs)
-+{
-+ t_List *p_Pos;
-+ t_Error err = E_OK;
-+ t_CcNodeInformation ccNodeInfo, *p_CcNodeInformation;
-+ t_Handle h_Muram;
-+ t_FmPcdCcNode *p_FmPcdCcNextNode, *p_FmPcdCcWorkingOnNode;
-+ t_List *p_UpdateLst;
-+ uint32_t intFlags;
-+
-+ SANITY_CHECK_RETURN_ERROR(h_FmPcd, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_AdditionalParams->h_CurrentNode,
-+ E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(h_FmPcdOldPointersLst, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(h_FmPcdNewPointersLst, E_INVALID_HANDLE);
-+
-+ /* We don't update subtree of the new node with new tree because it was done in the previous stage */
-+ if (p_AdditionalParams->h_NodeForAdd)
-+ {
-+ p_FmPcdCcNextNode = (t_FmPcdCcNode*)p_AdditionalParams->h_NodeForAdd;
-+
-+ if (!p_AdditionalParams->tree)
-+ p_UpdateLst = &p_FmPcdCcNextNode->ccPrevNodesLst;
-+ else
-+ p_UpdateLst = &p_FmPcdCcNextNode->ccTreeIdLst;
-+
-+ p_CcNodeInformation = FindNodeInfoInReleventLst(
-+ p_UpdateLst, p_AdditionalParams->h_CurrentNode,
-+ p_FmPcdCcNextNode->h_Spinlock);
-+
-+ if (p_CcNodeInformation)
-+ p_CcNodeInformation->index++;
-+ else
-+ {
-+ memset(&ccNodeInfo, 0, sizeof(t_CcNodeInformation));
-+ ccNodeInfo.h_CcNode = (t_Handle)p_AdditionalParams->h_CurrentNode;
-+ ccNodeInfo.index = 1;
-+ EnqueueNodeInfoToRelevantLst(p_UpdateLst, &ccNodeInfo,
-+ p_FmPcdCcNextNode->h_Spinlock);
-+ }
-+ if (p_AdditionalParams->h_ManipForAdd)
-+ {
-+ p_CcNodeInformation = FindNodeInfoInReleventLst(
-+ FmPcdManipGetNodeLstPointedOnThisManip(
-+ p_AdditionalParams->h_ManipForAdd),
-+ p_AdditionalParams->h_CurrentNode,
-+ FmPcdManipGetSpinlock(p_AdditionalParams->h_ManipForAdd));
-+
-+ if (p_CcNodeInformation)
-+ p_CcNodeInformation->index++;
-+ else
-+ {
-+ memset(&ccNodeInfo, 0, sizeof(t_CcNodeInformation));
-+ ccNodeInfo.h_CcNode =
-+ (t_Handle)p_AdditionalParams->h_CurrentNode;
-+ ccNodeInfo.index = 1;
-+ EnqueueNodeInfoToRelevantLst(
-+ FmPcdManipGetNodeLstPointedOnThisManip(
-+ p_AdditionalParams->h_ManipForAdd),
-+ &ccNodeInfo,
-+ FmPcdManipGetSpinlock(
-+ p_AdditionalParams->h_ManipForAdd));
-+ }
-+ }
-+ }
-+
-+ if (p_AdditionalParams->h_NodeForRmv)
-+ {
-+ p_FmPcdCcNextNode = (t_FmPcdCcNode*)p_AdditionalParams->h_NodeForRmv;
-+
-+ if (!p_AdditionalParams->tree)
-+ {
-+ p_UpdateLst = &p_FmPcdCcNextNode->ccPrevNodesLst;
-+ p_FmPcdCcWorkingOnNode =
-+ (t_FmPcdCcNode *)(p_AdditionalParams->h_CurrentNode);
-+
-+ for (p_Pos = LIST_FIRST(&p_FmPcdCcWorkingOnNode->ccTreesLst);
-+ p_Pos != (&p_FmPcdCcWorkingOnNode->ccTreesLst); p_Pos =
-+ LIST_NEXT(p_Pos))
-+ {
-+ p_CcNodeInformation = CC_NODE_F_OBJECT(p_Pos);
-+
-+ ASSERT_COND(p_CcNodeInformation->h_CcNode);
-+
-+ err =
-+ SetRequiredAction(
-+ h_FmPcd,
-+ UPDATE_CC_WITH_DELETE_TREE,
-+ &((t_FmPcdCcNode *)(p_AdditionalParams->h_CurrentNode))->keyAndNextEngineParams[p_AdditionalParams->savedKeyIndex],
-+ PTR_MOVE(((t_FmPcdCcNode *)(p_AdditionalParams->h_CurrentNode))->h_AdTable, p_AdditionalParams->savedKeyIndex*FM_PCD_CC_AD_ENTRY_SIZE),
-+ 1, p_CcNodeInformation->h_CcNode);
-+ }
-+ }
-+ else
-+ {
-+ p_UpdateLst = &p_FmPcdCcNextNode->ccTreeIdLst;
-+
-+ err =
-+ SetRequiredAction(
-+ h_FmPcd,
-+ UPDATE_CC_WITH_DELETE_TREE,
-+ &((t_FmPcdCcTree *)(p_AdditionalParams->h_CurrentNode))->keyAndNextEngineParams[p_AdditionalParams->savedKeyIndex],
-+ UINT_TO_PTR(((t_FmPcdCcTree *)(p_AdditionalParams->h_CurrentNode))->ccTreeBaseAddr + p_AdditionalParams->savedKeyIndex*FM_PCD_CC_AD_ENTRY_SIZE),
-+ 1, p_AdditionalParams->h_CurrentNode);
-+ }
-+ if (err)
-+ return err;
-+
-+ /* We remove from the subtree of the removed node tree because it wasn't done in the previous stage
-+ Update ccPrevNodesLst or ccTreeIdLst of the removed node
-+ Update of the node owner */
-+ p_CcNodeInformation = FindNodeInfoInReleventLst(
-+ p_UpdateLst, p_AdditionalParams->h_CurrentNode,
-+ p_FmPcdCcNextNode->h_Spinlock);
-+
-+ ASSERT_COND(p_CcNodeInformation);
-+ ASSERT_COND(p_CcNodeInformation->index);
-+
-+ p_CcNodeInformation->index--;
-+
-+ if (p_CcNodeInformation->index == 0)
-+ DequeueNodeInfoFromRelevantLst(p_UpdateLst,
-+ p_AdditionalParams->h_CurrentNode,
-+ p_FmPcdCcNextNode->h_Spinlock);
-+
-+ UpdateNodeOwner(p_FmPcdCcNextNode, FALSE);
-+
-+ if (p_AdditionalParams->h_ManipForRmv)
-+ {
-+ p_CcNodeInformation = FindNodeInfoInReleventLst(
-+ FmPcdManipGetNodeLstPointedOnThisManip(
-+ p_AdditionalParams->h_ManipForRmv),
-+ p_AdditionalParams->h_CurrentNode,
-+ FmPcdManipGetSpinlock(p_AdditionalParams->h_ManipForRmv));
-+
-+ ASSERT_COND(p_CcNodeInformation);
-+ ASSERT_COND(p_CcNodeInformation->index);
-+
-+ p_CcNodeInformation->index--;
-+
-+ if (p_CcNodeInformation->index == 0)
-+ DequeueNodeInfoFromRelevantLst(
-+ FmPcdManipGetNodeLstPointedOnThisManip(
-+ p_AdditionalParams->h_ManipForRmv),
-+ p_AdditionalParams->h_CurrentNode,
-+ FmPcdManipGetSpinlock(
-+ p_AdditionalParams->h_ManipForRmv));
-+ }
-+ }
-+
-+ if (p_AdditionalParams->h_ManipForRmv)
-+ FmPcdManipUpdateOwner(p_AdditionalParams->h_ManipForRmv, FALSE);
-+
-+ if (p_AdditionalParams->p_StatsObjForRmv)
-+ PutStatsObj((t_FmPcdCcNode *)(p_AdditionalParams->h_CurrentNode),
-+ p_AdditionalParams->p_StatsObjForRmv);
-+
-+#if (DPAA_VERSION >= 11)
-+ if (p_AdditionalParams->h_FrmReplicForRmv)
-+ FrmReplicGroupUpdateOwner(p_AdditionalParams->h_FrmReplicForRmv,
-+ FALSE/* remove */);
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+ if (!useShadowStructs)
-+ {
-+ h_Muram = FmPcdGetMuramHandle(h_FmPcd);
-+ ASSERT_COND(h_Muram);
-+
-+ if ((p_AdditionalParams->tree && !((t_FmPcd *)h_FmPcd)->p_CcShadow)
-+ || (!p_AdditionalParams->tree
-+ && !((t_FmPcdCcNode *)(p_AdditionalParams->h_CurrentNode))->maxNumOfKeys))
-+ {
-+ /* We release new AD which was allocated and updated for copy from to actual AD */
-+ for (p_Pos = LIST_FIRST(h_FmPcdNewPointersLst);
-+ p_Pos != (h_FmPcdNewPointersLst); p_Pos = LIST_NEXT(p_Pos))
-+ {
-+
-+ p_CcNodeInformation = CC_NODE_F_OBJECT(p_Pos);
-+ ASSERT_COND(p_CcNodeInformation->h_CcNode);
-+ FM_MURAM_FreeMem(h_Muram, p_CcNodeInformation->h_CcNode);
-+ }
-+ }
-+
-+ /* Free Old data structure if it has to be freed - new data structure was allocated*/
-+ if (p_AdditionalParams->p_AdTableOld)
-+ FM_MURAM_FreeMem(h_Muram, p_AdditionalParams->p_AdTableOld);
-+
-+ if (p_AdditionalParams->p_KeysMatchTableOld)
-+ FM_MURAM_FreeMem(h_Muram, p_AdditionalParams->p_KeysMatchTableOld);
-+ }
-+
-+ /* Update current modified node with changed fields if it's required*/
-+ if (!p_AdditionalParams->tree)
-+ {
-+ if (p_AdditionalParams->p_AdTableNew)
-+ ((t_FmPcdCcNode *)(p_AdditionalParams->h_CurrentNode))->h_AdTable =
-+ p_AdditionalParams->p_AdTableNew;
-+
-+ if (p_AdditionalParams->p_KeysMatchTableNew)
-+ ((t_FmPcdCcNode *)(p_AdditionalParams->h_CurrentNode))->h_KeysMatchTable =
-+ p_AdditionalParams->p_KeysMatchTableNew;
-+
-+ /* Locking node's spinlock before updating 'keys and next engine' structure,
-+ as it maybe used to retrieve keys statistics */
-+ intFlags =
-+ XX_LockIntrSpinlock(
-+ ((t_FmPcdCcNode *)(p_AdditionalParams->h_CurrentNode))->h_Spinlock);
-+
-+ ((t_FmPcdCcNode *)(p_AdditionalParams->h_CurrentNode))->numOfKeys =
-+ p_AdditionalParams->numOfKeys;
-+
-+ memcpy(((t_FmPcdCcNode *)(p_AdditionalParams->h_CurrentNode))->keyAndNextEngineParams,
-+ &p_AdditionalParams->keyAndNextEngineParams,
-+ sizeof(t_FmPcdCcKeyAndNextEngineParams) * (CC_MAX_NUM_OF_KEYS));
-+
-+ XX_UnlockIntrSpinlock(
-+ ((t_FmPcdCcNode *)(p_AdditionalParams->h_CurrentNode))->h_Spinlock,
-+ intFlags);
-+ }
-+ else
-+ {
-+ uint8_t numEntries =
-+ ((t_FmPcdCcTree *)(p_AdditionalParams->h_CurrentNode))->numOfEntries;
-+ ASSERT_COND(numEntries < FM_PCD_MAX_NUM_OF_CC_GROUPS);
-+ memcpy(&((t_FmPcdCcTree *)(p_AdditionalParams->h_CurrentNode))->keyAndNextEngineParams,
-+ &p_AdditionalParams->keyAndNextEngineParams,
-+ sizeof(t_FmPcdCcKeyAndNextEngineParams) * numEntries);
-+ }
-+
-+ ReleaseLst(h_FmPcdOldPointersLst);
-+ ReleaseLst(h_FmPcdNewPointersLst);
-+
-+ XX_Free(p_AdditionalParams);
-+
-+ return E_OK;
-+}
-+
-+static t_Handle BuildNewAd(
-+ t_Handle h_Ad,
-+ t_FmPcdModifyCcKeyAdditionalParams *p_FmPcdModifyCcKeyAdditionalParams,
-+ t_FmPcdCcNode *p_CcNode,
-+ t_FmPcdCcNextEngineParams *p_FmPcdCcNextEngineParams)
-+{
-+ t_FmPcdCcNode *p_FmPcdCcNodeTmp;
-+ t_Handle h_OrigAd = NULL;
-+
-+ p_FmPcdCcNodeTmp = (t_FmPcdCcNode*)XX_Malloc(sizeof(t_FmPcdCcNode));
-+ if (!p_FmPcdCcNodeTmp)
-+ {
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("p_FmPcdCcNodeTmp"));
-+ return NULL;
-+ }
-+ memset(p_FmPcdCcNodeTmp, 0, sizeof(t_FmPcdCcNode));
-+
-+ p_FmPcdCcNodeTmp->numOfKeys = p_FmPcdModifyCcKeyAdditionalParams->numOfKeys;
-+ p_FmPcdCcNodeTmp->h_KeysMatchTable =
-+ p_FmPcdModifyCcKeyAdditionalParams->p_KeysMatchTableNew;
-+ p_FmPcdCcNodeTmp->h_AdTable =
-+ p_FmPcdModifyCcKeyAdditionalParams->p_AdTableNew;
-+
-+ p_FmPcdCcNodeTmp->lclMask = p_CcNode->lclMask;
-+ p_FmPcdCcNodeTmp->parseCode = p_CcNode->parseCode;
-+ p_FmPcdCcNodeTmp->offset = p_CcNode->offset;
-+ p_FmPcdCcNodeTmp->prsArrayOffset = p_CcNode->prsArrayOffset;
-+ p_FmPcdCcNodeTmp->ctrlFlow = p_CcNode->ctrlFlow;
-+ p_FmPcdCcNodeTmp->ccKeySizeAccExtraction = p_CcNode->ccKeySizeAccExtraction;
-+ p_FmPcdCcNodeTmp->sizeOfExtraction = p_CcNode->sizeOfExtraction;
-+ p_FmPcdCcNodeTmp->glblMaskSize = p_CcNode->glblMaskSize;
-+ p_FmPcdCcNodeTmp->p_GlblMask = p_CcNode->p_GlblMask;
-+
-+ if (p_FmPcdCcNextEngineParams->nextEngine == e_FM_PCD_CC)
-+ {
-+ if (p_FmPcdCcNextEngineParams->h_Manip)
-+ {
-+ h_OrigAd = p_CcNode->h_Ad;
-+ if (AllocAndFillAdForContLookupManip(
-+ p_FmPcdCcNextEngineParams->params.ccParams.h_CcNode)
-+ != E_OK)
-+ {
-+ REPORT_ERROR(MAJOR, E_INVALID_STATE, NO_MSG);
-+ XX_Free(p_FmPcdCcNodeTmp);
-+ return NULL;
-+ }
-+ }
-+ FillAdOfTypeContLookup(h_Ad, NULL, p_CcNode->h_FmPcd, p_FmPcdCcNodeTmp,
-+ h_OrigAd ? NULL : p_FmPcdCcNextEngineParams->h_Manip, NULL);
-+ }
-+
-+#if (DPAA_VERSION >= 11)
-+ if ((p_FmPcdCcNextEngineParams->nextEngine == e_FM_PCD_FR)
-+ && (p_FmPcdCcNextEngineParams->params.frParams.h_FrmReplic))
-+ {
-+ FillAdOfTypeContLookup(
-+ h_Ad, NULL, p_CcNode->h_FmPcd, p_FmPcdCcNodeTmp,
-+ p_FmPcdCcNextEngineParams->h_Manip,
-+ p_FmPcdCcNextEngineParams->params.frParams.h_FrmReplic);
-+ }
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+ XX_Free(p_FmPcdCcNodeTmp);
-+
-+ return E_OK;
-+}
-+
-+static t_Error DynamicChangeHc(
-+ t_Handle h_FmPcd, t_List *h_OldPointersLst, t_List *h_NewPointersLst,
-+ t_FmPcdModifyCcKeyAdditionalParams *p_AdditionalParams,
-+ bool useShadowStructs)
-+{
-+ t_List *p_PosOld, *p_PosNew;
-+ uint32_t oldAdAddrOffset, newAdAddrOffset;
-+ uint16_t i = 0;
-+ t_Error err = E_OK;
-+ uint8_t numOfModifiedPtr;
-+
-+ ASSERT_COND(h_FmPcd);
-+ ASSERT_COND(h_OldPointersLst);
-+ ASSERT_COND(h_NewPointersLst);
-+
-+ numOfModifiedPtr = (uint8_t)LIST_NumOfObjs(h_OldPointersLst);
-+
-+ if (numOfModifiedPtr)
-+ {
-+ p_PosNew = LIST_FIRST(h_NewPointersLst);
-+ p_PosOld = LIST_FIRST(h_OldPointersLst);
-+
-+ /* Retrieve address of new AD */
-+ newAdAddrOffset = FmPcdCcGetNodeAddrOffsetFromNodeInfo(h_FmPcd,
-+ p_PosNew);
-+ if (newAdAddrOffset == (uint32_t)ILLEGAL_BASE)
-+ {
-+ ReleaseModifiedDataStructure(h_FmPcd, h_OldPointersLst,
-+ h_NewPointersLst,
-+ p_AdditionalParams, useShadowStructs);
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("New AD address"));
-+ }
-+
-+ for (i = 0; i < numOfModifiedPtr; i++)
-+ {
-+ /* Retrieve address of current AD */
-+ oldAdAddrOffset = FmPcdCcGetNodeAddrOffsetFromNodeInfo(h_FmPcd,
-+ p_PosOld);
-+ if (oldAdAddrOffset == (uint32_t)ILLEGAL_BASE)
-+ {
-+ ReleaseModifiedDataStructure(h_FmPcd, h_OldPointersLst,
-+ h_NewPointersLst,
-+ p_AdditionalParams,
-+ useShadowStructs);
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("Old AD address"));
-+ }
-+
-+ /* Invoke host command to copy from new AD to old AD */
-+ err = FmHcPcdCcDoDynamicChange(((t_FmPcd *)h_FmPcd)->h_Hc,
-+ oldAdAddrOffset, newAdAddrOffset);
-+ if (err)
-+ {
-+ ReleaseModifiedDataStructure(h_FmPcd, h_OldPointersLst,
-+ h_NewPointersLst,
-+ p_AdditionalParams,
-+ useShadowStructs);
-+ RETURN_ERROR(
-+ MAJOR,
-+ err,
-+ ("For part of nodes changes are done - situation is danger"));
-+ }
-+
-+ p_PosOld = LIST_NEXT(p_PosOld);
-+ }
-+ }
-+ return E_OK;
-+}
-+
-+static t_Error DoDynamicChange(
-+ t_Handle h_FmPcd, t_List *h_OldPointersLst, t_List *h_NewPointersLst,
-+ t_FmPcdModifyCcKeyAdditionalParams *p_AdditionalParams,
-+ bool useShadowStructs)
-+{
-+ t_FmPcdCcNode *p_CcNode =
-+ (t_FmPcdCcNode *)(p_AdditionalParams->h_CurrentNode);
-+ t_List *p_PosNew;
-+ t_CcNodeInformation *p_CcNodeInfo;
-+ t_FmPcdCcNextEngineParams nextEngineParams;
-+ t_Handle h_Ad;
-+ uint32_t keySize;
-+ t_Error err = E_OK;
-+ uint8_t numOfModifiedPtr;
-+
-+ ASSERT_COND(h_FmPcd);
-+
-+ memset(&nextEngineParams, 0, sizeof(t_FmPcdCcNextEngineParams));
-+
-+ numOfModifiedPtr = (uint8_t)LIST_NumOfObjs(h_OldPointersLst);
-+
-+ if (numOfModifiedPtr)
-+ {
-+
-+ p_PosNew = LIST_FIRST(h_NewPointersLst);
-+
-+ /* Invoke host-command to copy from the new Ad to existing Ads */
-+ err = DynamicChangeHc(h_FmPcd, h_OldPointersLst, h_NewPointersLst,
-+ p_AdditionalParams, useShadowStructs);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+
-+ if (useShadowStructs)
-+ {
-+ /* When the host-command above has ended, the old structures are 'free'and we can update
-+ them by copying from the new shadow structures. */
-+ if (p_CcNode->lclMask)
-+ keySize = (uint32_t)(2 * p_CcNode->ccKeySizeAccExtraction);
-+ else
-+ keySize = p_CcNode->ccKeySizeAccExtraction;
-+
-+ MemCpy8(p_AdditionalParams->p_KeysMatchTableOld,
-+ p_AdditionalParams->p_KeysMatchTableNew,
-+ p_CcNode->maxNumOfKeys * keySize * sizeof(uint8_t));
-+
-+ MemCpy8(
-+ p_AdditionalParams->p_AdTableOld,
-+ p_AdditionalParams->p_AdTableNew,
-+ (uint32_t)((p_CcNode->maxNumOfKeys + 1)
-+ * FM_PCD_CC_AD_ENTRY_SIZE));
-+
-+ /* Retrieve the address of the allocated Ad */
-+ p_CcNodeInfo = CC_NODE_F_OBJECT(p_PosNew);
-+ h_Ad = p_CcNodeInfo->h_CcNode;
-+
-+ /* Build a new Ad that holds the old (now updated) structures */
-+ p_AdditionalParams->p_KeysMatchTableNew =
-+ p_AdditionalParams->p_KeysMatchTableOld;
-+ p_AdditionalParams->p_AdTableNew = p_AdditionalParams->p_AdTableOld;
-+
-+ nextEngineParams.nextEngine = e_FM_PCD_CC;
-+ nextEngineParams.params.ccParams.h_CcNode = (t_Handle)p_CcNode;
-+
-+ BuildNewAd(h_Ad, p_AdditionalParams, p_CcNode, &nextEngineParams);
-+
-+ /* HC to copy from the new Ad (old updated structures) to current Ad (uses shadow structures) */
-+ err = DynamicChangeHc(h_FmPcd, h_OldPointersLst, h_NewPointersLst,
-+ p_AdditionalParams, useShadowStructs);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+ }
-+
-+ err = ReleaseModifiedDataStructure(h_FmPcd, h_OldPointersLst,
-+ h_NewPointersLst,
-+ p_AdditionalParams, useShadowStructs);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+
-+ return E_OK;
-+}
-+
-+#ifdef FM_CAPWAP_SUPPORT
-+static bool IsCapwapApplSpecific(t_Handle h_Node)
-+{
-+ t_FmPcdCcNode *p_CcNode = (t_FmPcdCcNode *)h_Node;
-+ bool isManipForCapwapApplSpecificBuild = FALSE;
-+ int i = 0;
-+
-+ ASSERT_COND(h_Node);
-+ /* assumption that this function called only for INDEXED_FLOW_ID - so no miss*/
-+ for (i = 0; i < p_CcNode->numOfKeys; i++)
-+ {
-+ if ( p_CcNode->keyAndNextEngineParams[i].nextEngineParams.h_Manip &&
-+ FmPcdManipIsCapwapApplSpecific(p_CcNode->keyAndNextEngineParams[i].nextEngineParams.h_Manip))
-+ {
-+ isManipForCapwapApplSpecificBuild = TRUE;
-+ break;
-+ }
-+ }
-+ return isManipForCapwapApplSpecificBuild;
-+
-+}
-+#endif /* FM_CAPWAP_SUPPORT */
-+
-+static t_Error CcUpdateParam(
-+ t_Handle h_FmPcd, t_Handle h_PcdParams, t_Handle h_FmPort,
-+ t_FmPcdCcKeyAndNextEngineParams *p_CcKeyAndNextEngineParams,
-+ uint16_t numOfEntries, t_Handle h_Ad, bool validate, uint16_t level,
-+ t_Handle h_FmTree, bool modify)
-+{
-+ t_FmPcdCcNode *p_CcNode;
-+ t_Error err;
-+ uint16_t tmp = 0;
-+ int i = 0;
-+ t_FmPcdCcTree *p_CcTree = (t_FmPcdCcTree *)h_FmTree;
-+
-+ level++;
-+
-+ if (p_CcTree->h_IpReassemblyManip)
-+ {
-+ err = FmPcdManipUpdate(h_FmPcd, h_PcdParams, h_FmPort,
-+ p_CcTree->h_IpReassemblyManip, NULL, validate,
-+ level, h_FmTree, modify);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+
-+ if (p_CcTree->h_CapwapReassemblyManip)
-+ {
-+ err = FmPcdManipUpdate(h_FmPcd, h_PcdParams, h_FmPort,
-+ p_CcTree->h_CapwapReassemblyManip, NULL, validate,
-+ level, h_FmTree, modify);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+
-+ if (numOfEntries)
-+ {
-+ for (i = 0; i < numOfEntries; i++)
-+ {
-+ if (i == 0)
-+ h_Ad = PTR_MOVE(h_Ad, i*FM_PCD_CC_AD_ENTRY_SIZE);
-+ else
-+ h_Ad = PTR_MOVE(h_Ad, FM_PCD_CC_AD_ENTRY_SIZE);
-+
-+ if (p_CcKeyAndNextEngineParams[i].nextEngineParams.nextEngine
-+ == e_FM_PCD_CC)
-+ {
-+ p_CcNode =
-+ p_CcKeyAndNextEngineParams[i].nextEngineParams.params.ccParams.h_CcNode;
-+ ASSERT_COND(p_CcNode);
-+
-+ if (p_CcKeyAndNextEngineParams[i].nextEngineParams.h_Manip)
-+ {
-+ err =
-+ FmPcdManipUpdate(
-+ h_FmPcd,
-+ NULL,
-+ h_FmPort,
-+ p_CcKeyAndNextEngineParams[i].nextEngineParams.h_Manip,
-+ h_Ad, validate, level, h_FmTree, modify);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+
-+ if (p_CcNode->keyAndNextEngineParams[p_CcNode->numOfKeys].nextEngineParams.nextEngine
-+ != e_FM_PCD_INVALID)
-+ tmp = (uint8_t)(p_CcNode->numOfKeys + 1);
-+ else
-+ tmp = p_CcNode->numOfKeys;
-+
-+ err = CcUpdateParam(h_FmPcd, h_PcdParams, h_FmPort,
-+ p_CcNode->keyAndNextEngineParams, tmp,
-+ p_CcNode->h_AdTable, validate, level,
-+ h_FmTree, modify);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+ else
-+ {
-+ if (p_CcKeyAndNextEngineParams[i].nextEngineParams.h_Manip)
-+ {
-+ err =
-+ FmPcdManipUpdate(
-+ h_FmPcd,
-+ NULL,
-+ h_FmPort,
-+ p_CcKeyAndNextEngineParams[i].nextEngineParams.h_Manip,
-+ h_Ad, validate, level, h_FmTree, modify);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+ }
-+ }
-+ }
-+
-+ return E_OK;
-+}
-+
-+static ccPrivateInfo_t IcDefineCode(t_FmPcdCcNodeParams *p_CcNodeParam)
-+{
-+ switch (p_CcNodeParam->extractCcParams.extractNonHdr.action)
-+ {
-+ case (e_FM_PCD_ACTION_EXACT_MATCH):
-+ switch (p_CcNodeParam->extractCcParams.extractNonHdr.src)
-+ {
-+ case (e_FM_PCD_EXTRACT_FROM_KEY):
-+ return CC_PRIVATE_INFO_IC_KEY_EXACT_MATCH;
-+ case (e_FM_PCD_EXTRACT_FROM_HASH):
-+ return CC_PRIVATE_INFO_IC_HASH_EXACT_MATCH;
-+ default:
-+ return CC_PRIVATE_INFO_NONE;
-+ }
-+
-+ case (e_FM_PCD_ACTION_INDEXED_LOOKUP):
-+ switch (p_CcNodeParam->extractCcParams.extractNonHdr.src)
-+ {
-+ case (e_FM_PCD_EXTRACT_FROM_HASH):
-+ return CC_PRIVATE_INFO_IC_HASH_INDEX_LOOKUP;
-+ case (e_FM_PCD_EXTRACT_FROM_FLOW_ID):
-+ return CC_PRIVATE_INFO_IC_DEQ_FQID_INDEX_LOOKUP;
-+ default:
-+ return CC_PRIVATE_INFO_NONE;
-+ }
-+
-+ default:
-+ break;
-+ }
-+
-+ return CC_PRIVATE_INFO_NONE;
-+}
-+
-+static t_CcNodeInformation * DequeueAdditionalInfoFromRelevantLst(
-+ t_List *p_List)
-+{
-+ t_CcNodeInformation *p_CcNodeInfo = NULL;
-+
-+ if (!LIST_IsEmpty(p_List))
-+ {
-+ p_CcNodeInfo = CC_NODE_F_OBJECT(p_List->p_Next);
-+ LIST_DelAndInit(&p_CcNodeInfo->node);
-+ }
-+
-+ return p_CcNodeInfo;
-+}
-+
-+void ReleaseLst(t_List *p_List)
-+{
-+ t_CcNodeInformation *p_CcNodeInfo = NULL;
-+
-+ if (!LIST_IsEmpty(p_List))
-+ {
-+ p_CcNodeInfo = DequeueAdditionalInfoFromRelevantLst(p_List);
-+ while (p_CcNodeInfo)
-+ {
-+ XX_Free(p_CcNodeInfo);
-+ p_CcNodeInfo = DequeueAdditionalInfoFromRelevantLst(p_List);
-+ }
-+ }
-+
-+ LIST_Del(p_List);
-+}
-+
-+static void DeleteNode(t_FmPcdCcNode *p_CcNode)
-+{
-+ uint32_t i;
-+
-+ if (!p_CcNode)
-+ return;
-+
-+ if (p_CcNode->p_GlblMask)
-+ {
-+ XX_Free(p_CcNode->p_GlblMask);
-+ p_CcNode->p_GlblMask = NULL;
-+ }
-+
-+ if (p_CcNode->h_KeysMatchTable)
-+ {
-+ FM_MURAM_FreeMem(FmPcdGetMuramHandle(p_CcNode->h_FmPcd),
-+ p_CcNode->h_KeysMatchTable);
-+ p_CcNode->h_KeysMatchTable = NULL;
-+ }
-+
-+ if (p_CcNode->h_AdTable)
-+ {
-+ FM_MURAM_FreeMem(FmPcdGetMuramHandle(p_CcNode->h_FmPcd),
-+ p_CcNode->h_AdTable);
-+ p_CcNode->h_AdTable = NULL;
-+ }
-+
-+ if (p_CcNode->h_Ad)
-+ {
-+ FM_MURAM_FreeMem(FmPcdGetMuramHandle(p_CcNode->h_FmPcd),
-+ p_CcNode->h_Ad);
-+ p_CcNode->h_Ad = NULL;
-+ p_CcNode->h_TmpAd = NULL;
-+ }
-+
-+ if (p_CcNode->h_StatsFLRs)
-+ {
-+ FM_MURAM_FreeMem(FmPcdGetMuramHandle(p_CcNode->h_FmPcd),
-+ p_CcNode->h_StatsFLRs);
-+ p_CcNode->h_StatsFLRs = NULL;
-+ }
-+
-+ if (p_CcNode->h_Spinlock)
-+ {
-+ XX_FreeSpinlock(p_CcNode->h_Spinlock);
-+ p_CcNode->h_Spinlock = NULL;
-+ }
-+
-+ /* Restore the original counters pointer instead of the mutual pointer (mutual to all hash buckets) */
-+ if (p_CcNode->isHashBucket
-+ && (p_CcNode->statisticsMode != e_FM_PCD_CC_STATS_MODE_NONE))
-+ p_CcNode->keyAndNextEngineParams[p_CcNode->numOfKeys].p_StatsObj->h_StatsCounters =
-+ p_CcNode->h_PrivMissStatsCounters;
-+
-+ /* Releasing all currently used statistics objects, including 'miss' entry */
-+ for (i = 0; i < p_CcNode->numOfKeys + 1; i++)
-+ if (p_CcNode->keyAndNextEngineParams[i].p_StatsObj)
-+ PutStatsObj(p_CcNode,
-+ p_CcNode->keyAndNextEngineParams[i].p_StatsObj);
-+
-+ if (!LIST_IsEmpty(&p_CcNode->availableStatsLst))
-+ {
-+ t_Handle h_FmMuram = FmPcdGetMuramHandle(p_CcNode->h_FmPcd);
-+ ASSERT_COND(h_FmMuram);
-+
-+ FreeStatObjects(&p_CcNode->availableStatsLst, h_FmMuram);
-+ }
-+
-+ LIST_Del(&p_CcNode->availableStatsLst);
-+
-+ ReleaseLst(&p_CcNode->availableStatsLst);
-+ ReleaseLst(&p_CcNode->ccPrevNodesLst);
-+ ReleaseLst(&p_CcNode->ccTreeIdLst);
-+ ReleaseLst(&p_CcNode->ccTreesLst);
-+
-+ XX_Free(p_CcNode);
-+}
-+
-+static void DeleteTree(t_FmPcdCcTree *p_FmPcdTree, t_FmPcd *p_FmPcd)
-+{
-+ if (p_FmPcdTree)
-+ {
-+ if (p_FmPcdTree->ccTreeBaseAddr)
-+ {
-+ FM_MURAM_FreeMem(FmPcdGetMuramHandle(p_FmPcd),
-+ UINT_TO_PTR(p_FmPcdTree->ccTreeBaseAddr));
-+ p_FmPcdTree->ccTreeBaseAddr = 0;
-+ }
-+
-+ ReleaseLst(&p_FmPcdTree->fmPortsLst);
-+
-+ XX_Free(p_FmPcdTree);
-+ }
-+}
-+
-+static void GetCcExtractKeySize(uint8_t parseCodeRealSize,
-+ uint8_t *parseCodeCcSize)
-+{
-+ if ((parseCodeRealSize > 0) && (parseCodeRealSize < 2))
-+ *parseCodeCcSize = 1;
-+ else
-+ if (parseCodeRealSize == 2)
-+ *parseCodeCcSize = 2;
-+ else
-+ if ((parseCodeRealSize > 2) && (parseCodeRealSize <= 4))
-+ *parseCodeCcSize = 4;
-+ else
-+ if ((parseCodeRealSize > 4) && (parseCodeRealSize <= 8))
-+ *parseCodeCcSize = 8;
-+ else
-+ if ((parseCodeRealSize > 8) && (parseCodeRealSize <= 16))
-+ *parseCodeCcSize = 16;
-+ else
-+ if ((parseCodeRealSize > 16)
-+ && (parseCodeRealSize <= 24))
-+ *parseCodeCcSize = 24;
-+ else
-+ if ((parseCodeRealSize > 24)
-+ && (parseCodeRealSize <= 32))
-+ *parseCodeCcSize = 32;
-+ else
-+ if ((parseCodeRealSize > 32)
-+ && (parseCodeRealSize <= 40))
-+ *parseCodeCcSize = 40;
-+ else
-+ if ((parseCodeRealSize > 40)
-+ && (parseCodeRealSize <= 48))
-+ *parseCodeCcSize = 48;
-+ else
-+ if ((parseCodeRealSize > 48)
-+ && (parseCodeRealSize <= 56))
-+ *parseCodeCcSize = 56;
-+ else
-+ *parseCodeCcSize = 0;
-+}
-+
-+static void GetSizeHeaderField(e_NetHeaderType hdr, t_FmPcdFields field,
-+ uint8_t *parseCodeRealSize)
-+{
-+ switch (hdr)
-+ {
-+ case (HEADER_TYPE_ETH):
-+ switch (field.eth)
-+ {
-+ case (NET_HEADER_FIELD_ETH_DA):
-+ *parseCodeRealSize = 6;
-+ break;
-+
-+ case (NET_HEADER_FIELD_ETH_SA):
-+ *parseCodeRealSize = 6;
-+ break;
-+
-+ case (NET_HEADER_FIELD_ETH_TYPE):
-+ *parseCodeRealSize = 2;
-+ break;
-+
-+ default:
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Extraction not supported1"));
-+ *parseCodeRealSize = CC_SIZE_ILLEGAL;
-+ break;
-+ }
-+ break;
-+
-+ case (HEADER_TYPE_PPPoE):
-+ switch (field.pppoe)
-+ {
-+ case (NET_HEADER_FIELD_PPPoE_PID):
-+ *parseCodeRealSize = 2;
-+ break;
-+
-+ default:
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Extraction not supported1"));
-+ *parseCodeRealSize = CC_SIZE_ILLEGAL;
-+ break;
-+ }
-+ break;
-+
-+ case (HEADER_TYPE_VLAN):
-+ switch (field.vlan)
-+ {
-+ case (NET_HEADER_FIELD_VLAN_TCI):
-+ *parseCodeRealSize = 2;
-+ break;
-+
-+ default:
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Extraction not supported2"));
-+ *parseCodeRealSize = CC_SIZE_ILLEGAL;
-+ break;
-+ }
-+ break;
-+
-+ case (HEADER_TYPE_MPLS):
-+ switch (field.mpls)
-+ {
-+ case (NET_HEADER_FIELD_MPLS_LABEL_STACK):
-+ *parseCodeRealSize = 4;
-+ break;
-+
-+ default:
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Extraction not supported3"));
-+ *parseCodeRealSize = CC_SIZE_ILLEGAL;
-+ break;
-+ }
-+ break;
-+
-+ case (HEADER_TYPE_IPv4):
-+ switch (field.ipv4)
-+ {
-+ case (NET_HEADER_FIELD_IPv4_DST_IP):
-+ case (NET_HEADER_FIELD_IPv4_SRC_IP):
-+ *parseCodeRealSize = 4;
-+ break;
-+
-+ case (NET_HEADER_FIELD_IPv4_TOS):
-+ case (NET_HEADER_FIELD_IPv4_PROTO):
-+ *parseCodeRealSize = 1;
-+ break;
-+
-+ case (NET_HEADER_FIELD_IPv4_DST_IP
-+ | NET_HEADER_FIELD_IPv4_SRC_IP):
-+ *parseCodeRealSize = 8;
-+ break;
-+
-+ case (NET_HEADER_FIELD_IPv4_TTL):
-+ *parseCodeRealSize = 1;
-+ break;
-+
-+ default:
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Extraction not supported4"));
-+ *parseCodeRealSize = CC_SIZE_ILLEGAL;
-+ break;
-+ }
-+ break;
-+
-+ case (HEADER_TYPE_IPv6):
-+ switch (field.ipv6)
-+ {
-+ case (NET_HEADER_FIELD_IPv6_VER | NET_HEADER_FIELD_IPv6_FL
-+ | NET_HEADER_FIELD_IPv6_TC):
-+ *parseCodeRealSize = 4;
-+ break;
-+
-+ case (NET_HEADER_FIELD_IPv6_NEXT_HDR):
-+ case (NET_HEADER_FIELD_IPv6_HOP_LIMIT):
-+ *parseCodeRealSize = 1;
-+ break;
-+
-+ case (NET_HEADER_FIELD_IPv6_DST_IP):
-+ case (NET_HEADER_FIELD_IPv6_SRC_IP):
-+ *parseCodeRealSize = 16;
-+ break;
-+
-+ default:
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Extraction not supported5"));
-+ *parseCodeRealSize = CC_SIZE_ILLEGAL;
-+ break;
-+ }
-+ break;
-+
-+ case (HEADER_TYPE_IP):
-+ switch (field.ip)
-+ {
-+ case (NET_HEADER_FIELD_IP_DSCP):
-+ case (NET_HEADER_FIELD_IP_PROTO):
-+ *parseCodeRealSize = 1;
-+ break;
-+
-+ default:
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Extraction not supported5"));
-+ *parseCodeRealSize = CC_SIZE_ILLEGAL;
-+ break;
-+ }
-+ break;
-+
-+ case (HEADER_TYPE_GRE):
-+ switch (field.gre)
-+ {
-+ case (NET_HEADER_FIELD_GRE_TYPE):
-+ *parseCodeRealSize = 2;
-+ break;
-+
-+ default:
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Extraction not supported6"));
-+ *parseCodeRealSize = CC_SIZE_ILLEGAL;
-+ break;
-+ }
-+ break;
-+
-+ case (HEADER_TYPE_MINENCAP):
-+ switch (field.minencap)
-+ {
-+ case (NET_HEADER_FIELD_MINENCAP_TYPE):
-+ *parseCodeRealSize = 1;
-+ break;
-+
-+ case (NET_HEADER_FIELD_MINENCAP_DST_IP):
-+ case (NET_HEADER_FIELD_MINENCAP_SRC_IP):
-+ *parseCodeRealSize = 4;
-+ break;
-+
-+ case (NET_HEADER_FIELD_MINENCAP_SRC_IP
-+ | NET_HEADER_FIELD_MINENCAP_DST_IP):
-+ *parseCodeRealSize = 8;
-+ break;
-+
-+ default:
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Extraction not supported7"));
-+ *parseCodeRealSize = CC_SIZE_ILLEGAL;
-+ break;
-+ }
-+ break;
-+
-+ case (HEADER_TYPE_TCP):
-+ switch (field.tcp)
-+ {
-+ case (NET_HEADER_FIELD_TCP_PORT_SRC):
-+ case (NET_HEADER_FIELD_TCP_PORT_DST):
-+ *parseCodeRealSize = 2;
-+ break;
-+
-+ case (NET_HEADER_FIELD_TCP_PORT_SRC
-+ | NET_HEADER_FIELD_TCP_PORT_DST):
-+ *parseCodeRealSize = 4;
-+ break;
-+
-+ default:
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Extraction not supported8"));
-+ *parseCodeRealSize = CC_SIZE_ILLEGAL;
-+ break;
-+ }
-+ break;
-+
-+ case (HEADER_TYPE_UDP):
-+ switch (field.udp)
-+ {
-+ case (NET_HEADER_FIELD_UDP_PORT_SRC):
-+ case (NET_HEADER_FIELD_UDP_PORT_DST):
-+ *parseCodeRealSize = 2;
-+ break;
-+
-+ case (NET_HEADER_FIELD_UDP_PORT_SRC
-+ | NET_HEADER_FIELD_UDP_PORT_DST):
-+ *parseCodeRealSize = 4;
-+ break;
-+
-+ default:
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Extraction not supported9"));
-+ *parseCodeRealSize = CC_SIZE_ILLEGAL;
-+ break;
-+ }
-+ break;
-+
-+ default:
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Extraction not supported10"));
-+ *parseCodeRealSize = CC_SIZE_ILLEGAL;
-+ break;
-+ }
-+}
-+
-+t_Error ValidateNextEngineParams(
-+ t_Handle h_FmPcd, t_FmPcdCcNextEngineParams *p_FmPcdCcNextEngineParams,
-+ e_FmPcdCcStatsMode statsMode)
-+{
-+ uint16_t absoluteProfileId;
-+ t_Error err = E_OK;
-+ uint8_t relativeSchemeId;
-+
-+ if ((statsMode == e_FM_PCD_CC_STATS_MODE_NONE)
-+ && (p_FmPcdCcNextEngineParams->statisticsEn))
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_CONFLICT,
-+ ("Statistics are requested for a key, but statistics mode was set"
-+ "to 'NONE' upon initialization"));
-+
-+ switch (p_FmPcdCcNextEngineParams->nextEngine)
-+ {
-+ case (e_FM_PCD_INVALID):
-+ err = E_NOT_SUPPORTED;
-+ break;
-+
-+ case (e_FM_PCD_DONE):
-+ if ((p_FmPcdCcNextEngineParams->params.enqueueParams.action
-+ == e_FM_PCD_ENQ_FRAME)
-+ && p_FmPcdCcNextEngineParams->params.enqueueParams.overrideFqid)
-+ {
-+ if (!p_FmPcdCcNextEngineParams->params.enqueueParams.newFqid)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_CONFLICT,
-+ ("When overrideFqid is set, newFqid must not be zero"));
-+ if (p_FmPcdCcNextEngineParams->params.enqueueParams.newFqid
-+ & ~0x00FFFFFF)
-+ RETURN_ERROR(
-+ MAJOR, E_INVALID_VALUE,
-+ ("fqidForCtrlFlow must be between 1 and 2^24-1"));
-+ }
-+ break;
-+
-+ case (e_FM_PCD_KG):
-+ relativeSchemeId =
-+ FmPcdKgGetRelativeSchemeId(
-+ h_FmPcd,
-+ FmPcdKgGetSchemeId(
-+ p_FmPcdCcNextEngineParams->params.kgParams.h_DirectScheme));
-+ if (relativeSchemeId == FM_PCD_KG_NUM_OF_SCHEMES)
-+ RETURN_ERROR(MAJOR, E_NOT_IN_RANGE, NO_MSG);
-+ if (!FmPcdKgIsSchemeValidSw(
-+ p_FmPcdCcNextEngineParams->params.kgParams.h_DirectScheme))
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE,
-+ ("not valid schemeIndex in KG next engine param"));
-+ if (!KgIsSchemeAlwaysDirect(h_FmPcd, relativeSchemeId))
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_STATE,
-+ ("CC Node may point only to a scheme that is always direct."));
-+ break;
-+
-+ case (e_FM_PCD_PLCR):
-+ if (p_FmPcdCcNextEngineParams->params.plcrParams.overrideParams)
-+ {
-+ /* if private policer profile, it may be uninitialized yet, therefore no checks are done at this stage */
-+ if (p_FmPcdCcNextEngineParams->params.plcrParams.sharedProfile)
-+ {
-+ err =
-+ FmPcdPlcrGetAbsoluteIdByProfileParams(
-+ h_FmPcd,
-+ e_FM_PCD_PLCR_SHARED,
-+ NULL,
-+ p_FmPcdCcNextEngineParams->params.plcrParams.newRelativeProfileId,
-+ &absoluteProfileId);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err,
-+ ("Shared profile offset is out of range"));
-+ if (!FmPcdPlcrIsProfileValid(h_FmPcd, absoluteProfileId))
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE,
-+ ("Invalid profile"));
-+ }
-+ }
-+ break;
-+
-+ case (e_FM_PCD_HASH):
-+ p_FmPcdCcNextEngineParams->nextEngine = e_FM_PCD_CC;
-+ case (e_FM_PCD_CC):
-+ if (!p_FmPcdCcNextEngineParams->params.ccParams.h_CcNode)
-+ RETURN_ERROR(MAJOR, E_NULL_POINTER,
-+ ("handler to next Node is NULL"));
-+ break;
-+
-+#if (DPAA_VERSION >= 11)
-+ case (e_FM_PCD_FR):
-+ if (!p_FmPcdCcNextEngineParams->params.frParams.h_FrmReplic)
-+ err = E_NOT_SUPPORTED;
-+ break;
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+ default:
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE,
-+ ("Next engine is not correct"));
-+ }
-+
-+
-+ return err;
-+}
-+
-+static uint8_t GetGenParseCode(e_FmPcdExtractFrom src,
-+ uint32_t offset, bool glblMask,
-+ uint8_t *parseArrayOffset, bool fromIc,
-+ ccPrivateInfo_t icCode)
-+{
-+ if (!fromIc)
-+ {
-+ switch (src)
-+ {
-+ case (e_FM_PCD_EXTRACT_FROM_FRAME_START):
-+ if (glblMask)
-+ return CC_PC_GENERIC_WITH_MASK;
-+ else
-+ return CC_PC_GENERIC_WITHOUT_MASK;
-+
-+ case (e_FM_PCD_EXTRACT_FROM_CURR_END_OF_PARSE):
-+ *parseArrayOffset = CC_PC_PR_NEXT_HEADER_OFFSET;
-+ if (offset)
-+ return CC_PR_OFFSET;
-+ else
-+ return CC_PR_WITHOUT_OFFSET;
-+
-+ default:
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE, ("Illegal 'extract from' src"));
-+ return CC_PC_ILLEGAL;
-+ }
-+ }
-+ else
-+ {
-+ switch (icCode)
-+ {
-+ case (CC_PRIVATE_INFO_IC_KEY_EXACT_MATCH):
-+ *parseArrayOffset = 0x50;
-+ return CC_PC_GENERIC_IC_GMASK;
-+
-+ case (CC_PRIVATE_INFO_IC_HASH_EXACT_MATCH):
-+ *parseArrayOffset = 0x48;
-+ return CC_PC_GENERIC_IC_GMASK;
-+
-+ case (CC_PRIVATE_INFO_IC_HASH_INDEX_LOOKUP):
-+ *parseArrayOffset = 0x48;
-+ return CC_PC_GENERIC_IC_HASH_INDEXED;
-+
-+ case (CC_PRIVATE_INFO_IC_DEQ_FQID_INDEX_LOOKUP):
-+ *parseArrayOffset = 0x16;
-+ return CC_PC_GENERIC_IC_HASH_INDEXED;
-+
-+ default:
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE, ("Illegal 'extract from' src"));
-+ break;
-+ }
-+ }
-+
-+ return CC_PC_ILLEGAL;
-+}
-+
-+static uint8_t GetFullFieldParseCode(e_NetHeaderType hdr, e_FmPcdHdrIndex index,
-+ t_FmPcdFields field)
-+{
-+ switch (hdr)
-+ {
-+ case (HEADER_TYPE_NONE):
-+ ASSERT_COND(FALSE);
-+ return CC_PC_ILLEGAL;
-+
-+ case (HEADER_TYPE_ETH):
-+ switch (field.eth)
-+ {
-+ case (NET_HEADER_FIELD_ETH_DA):
-+ return CC_PC_FF_MACDST;
-+ case (NET_HEADER_FIELD_ETH_SA):
-+ return CC_PC_FF_MACSRC;
-+ case (NET_HEADER_FIELD_ETH_TYPE):
-+ return CC_PC_FF_ETYPE;
-+ default:
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Extraction not supported"));
-+ return CC_PC_ILLEGAL;
-+ }
-+
-+ case (HEADER_TYPE_VLAN):
-+ switch (field.vlan)
-+ {
-+ case (NET_HEADER_FIELD_VLAN_TCI):
-+ if ((index == e_FM_PCD_HDR_INDEX_NONE)
-+ || (index == e_FM_PCD_HDR_INDEX_1))
-+ return CC_PC_FF_TCI1;
-+ if (index == e_FM_PCD_HDR_INDEX_LAST)
-+ return CC_PC_FF_TCI2;
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Extraction not supported"));
-+ return CC_PC_ILLEGAL;
-+ default:
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Extraction not supported"));
-+ return CC_PC_ILLEGAL;
-+ }
-+
-+ case (HEADER_TYPE_MPLS):
-+ switch (field.mpls)
-+ {
-+ case (NET_HEADER_FIELD_MPLS_LABEL_STACK):
-+ if ((index == e_FM_PCD_HDR_INDEX_NONE)
-+ || (index == e_FM_PCD_HDR_INDEX_1))
-+ return CC_PC_FF_MPLS1;
-+ if (index == e_FM_PCD_HDR_INDEX_LAST)
-+ return CC_PC_FF_MPLS_LAST;
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Illegal MPLS index"));
-+ return CC_PC_ILLEGAL;
-+ default:
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Extraction not supported"));
-+ return CC_PC_ILLEGAL;
-+ }
-+
-+ case (HEADER_TYPE_IPv4):
-+ switch (field.ipv4)
-+ {
-+ case (NET_HEADER_FIELD_IPv4_DST_IP):
-+ if ((index == e_FM_PCD_HDR_INDEX_NONE)
-+ || (index == e_FM_PCD_HDR_INDEX_1))
-+ return CC_PC_FF_IPV4DST1;
-+ if (index == e_FM_PCD_HDR_INDEX_2)
-+ return CC_PC_FF_IPV4DST2;
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Illegal IPv4 index"));
-+ return CC_PC_ILLEGAL;
-+ case (NET_HEADER_FIELD_IPv4_TOS):
-+ if ((index == e_FM_PCD_HDR_INDEX_NONE)
-+ || (index == e_FM_PCD_HDR_INDEX_1))
-+ return CC_PC_FF_IPV4IPTOS_TC1;
-+ if (index == e_FM_PCD_HDR_INDEX_2)
-+ return CC_PC_FF_IPV4IPTOS_TC2;
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Illegal IPv4 index"));
-+ return CC_PC_ILLEGAL;
-+ case (NET_HEADER_FIELD_IPv4_PROTO):
-+ if ((index == e_FM_PCD_HDR_INDEX_NONE)
-+ || (index == e_FM_PCD_HDR_INDEX_1))
-+ return CC_PC_FF_IPV4PTYPE1;
-+ if (index == e_FM_PCD_HDR_INDEX_2)
-+ return CC_PC_FF_IPV4PTYPE2;
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Illegal IPv4 index"));
-+ return CC_PC_ILLEGAL;
-+ case (NET_HEADER_FIELD_IPv4_SRC_IP):
-+ if ((index == e_FM_PCD_HDR_INDEX_NONE)
-+ || (index == e_FM_PCD_HDR_INDEX_1))
-+ return CC_PC_FF_IPV4SRC1;
-+ if (index == e_FM_PCD_HDR_INDEX_2)
-+ return CC_PC_FF_IPV4SRC2;
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Illegal IPv4 index"));
-+ return CC_PC_ILLEGAL;
-+ case (NET_HEADER_FIELD_IPv4_SRC_IP
-+ | NET_HEADER_FIELD_IPv4_DST_IP):
-+ if ((index == e_FM_PCD_HDR_INDEX_NONE)
-+ || (index == e_FM_PCD_HDR_INDEX_1))
-+ return CC_PC_FF_IPV4SRC1_IPV4DST1;
-+ if (index == e_FM_PCD_HDR_INDEX_2)
-+ return CC_PC_FF_IPV4SRC2_IPV4DST2;
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Illegal IPv4 index"));
-+ return CC_PC_ILLEGAL;
-+ case (NET_HEADER_FIELD_IPv4_TTL):
-+ return CC_PC_FF_IPV4TTL;
-+ default:
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Extraction not supported"));
-+ return CC_PC_ILLEGAL;
-+ }
-+
-+ case (HEADER_TYPE_IPv6):
-+ switch (field.ipv6)
-+ {
-+ case (NET_HEADER_FIELD_IPv6_VER | NET_HEADER_FIELD_IPv6_FL
-+ | NET_HEADER_FIELD_IPv6_TC):
-+ if ((index == e_FM_PCD_HDR_INDEX_NONE)
-+ || (index == e_FM_PCD_HDR_INDEX_1))
-+ return CC_PC_FF_IPTOS_IPV6TC1_IPV6FLOW1;
-+ if (index == e_FM_PCD_HDR_INDEX_2)
-+ return CC_PC_FF_IPTOS_IPV6TC2_IPV6FLOW2;
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Illegal IPv6 index"));
-+ return CC_PC_ILLEGAL;
-+
-+ case (NET_HEADER_FIELD_IPv6_NEXT_HDR):
-+ if ((index == e_FM_PCD_HDR_INDEX_NONE)
-+ || (index == e_FM_PCD_HDR_INDEX_1))
-+ return CC_PC_FF_IPV6PTYPE1;
-+ if (index == e_FM_PCD_HDR_INDEX_2)
-+ return CC_PC_FF_IPV6PTYPE2;
-+ if (index == e_FM_PCD_HDR_INDEX_LAST)
-+ return CC_PC_FF_IPPID;
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Illegal IPv6 index"));
-+ return CC_PC_ILLEGAL;
-+
-+ case (NET_HEADER_FIELD_IPv6_DST_IP):
-+ if ((index == e_FM_PCD_HDR_INDEX_NONE)
-+ || (index == e_FM_PCD_HDR_INDEX_1))
-+ return CC_PC_FF_IPV6DST1;
-+ if (index == e_FM_PCD_HDR_INDEX_2)
-+ return CC_PC_FF_IPV6DST2;
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Illegal IPv6 index"));
-+ return CC_PC_ILLEGAL;
-+
-+ case (NET_HEADER_FIELD_IPv6_SRC_IP):
-+ if ((index == e_FM_PCD_HDR_INDEX_NONE)
-+ || (index == e_FM_PCD_HDR_INDEX_1))
-+ return CC_PC_FF_IPV6SRC1;
-+ if (index == e_FM_PCD_HDR_INDEX_2)
-+ return CC_PC_FF_IPV6SRC2;
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Illegal IPv6 index"));
-+ return CC_PC_ILLEGAL;
-+
-+ case (NET_HEADER_FIELD_IPv6_HOP_LIMIT):
-+ return CC_PC_FF_IPV6HOP_LIMIT;
-+
-+ default:
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Extraction not supported"));
-+ return CC_PC_ILLEGAL;
-+ }
-+
-+ case (HEADER_TYPE_IP):
-+ switch (field.ip)
-+ {
-+ case (NET_HEADER_FIELD_IP_DSCP):
-+ if ((index == e_FM_PCD_HDR_INDEX_NONE)
-+ || (index == e_FM_PCD_HDR_INDEX_1))
-+ return CC_PC_FF_IPDSCP;
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Illegal IP index"));
-+ return CC_PC_ILLEGAL;
-+
-+ case (NET_HEADER_FIELD_IP_PROTO):
-+ if (index == e_FM_PCD_HDR_INDEX_LAST)
-+ return CC_PC_FF_IPPID;
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Illegal IP index"));
-+ return CC_PC_ILLEGAL;
-+
-+ default:
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Extraction not supported"));
-+ return CC_PC_ILLEGAL;
-+ }
-+
-+ case (HEADER_TYPE_GRE):
-+ switch (field.gre)
-+ {
-+ case (NET_HEADER_FIELD_GRE_TYPE):
-+ return CC_PC_FF_GREPTYPE;
-+
-+ default:
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Extraction not supported"));
-+ return CC_PC_ILLEGAL;
-+ }
-+
-+ case (HEADER_TYPE_MINENCAP):
-+ switch (field.minencap)
-+ {
-+ case (NET_HEADER_FIELD_MINENCAP_TYPE):
-+ return CC_PC_FF_MINENCAP_PTYPE;
-+
-+ case (NET_HEADER_FIELD_MINENCAP_DST_IP):
-+ return CC_PC_FF_MINENCAP_IPDST;
-+
-+ case (NET_HEADER_FIELD_MINENCAP_SRC_IP):
-+ return CC_PC_FF_MINENCAP_IPSRC;
-+
-+ case (NET_HEADER_FIELD_MINENCAP_SRC_IP
-+ | NET_HEADER_FIELD_MINENCAP_DST_IP):
-+ return CC_PC_FF_MINENCAP_IPSRC_IPDST;
-+
-+ default:
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Extraction not supported"));
-+ return CC_PC_ILLEGAL;
-+ }
-+
-+ case (HEADER_TYPE_TCP):
-+ switch (field.tcp)
-+ {
-+ case (NET_HEADER_FIELD_TCP_PORT_SRC):
-+ return CC_PC_FF_L4PSRC;
-+
-+ case (NET_HEADER_FIELD_TCP_PORT_DST):
-+ return CC_PC_FF_L4PDST;
-+
-+ case (NET_HEADER_FIELD_TCP_PORT_DST
-+ | NET_HEADER_FIELD_TCP_PORT_SRC):
-+ return CC_PC_FF_L4PSRC_L4PDST;
-+
-+ default:
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Extraction not supported"));
-+ return CC_PC_ILLEGAL;
-+ }
-+
-+ case (HEADER_TYPE_PPPoE):
-+ switch (field.pppoe)
-+ {
-+ case (NET_HEADER_FIELD_PPPoE_PID):
-+ return CC_PC_FF_PPPPID;
-+
-+ default:
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Extraction not supported"));
-+ return CC_PC_ILLEGAL;
-+ }
-+
-+ case (HEADER_TYPE_UDP):
-+ switch (field.udp)
-+ {
-+ case (NET_HEADER_FIELD_UDP_PORT_SRC):
-+ return CC_PC_FF_L4PSRC;
-+
-+ case (NET_HEADER_FIELD_UDP_PORT_DST):
-+ return CC_PC_FF_L4PDST;
-+
-+ case (NET_HEADER_FIELD_UDP_PORT_DST
-+ | NET_HEADER_FIELD_UDP_PORT_SRC):
-+ return CC_PC_FF_L4PSRC_L4PDST;
-+
-+ default:
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Extraction not supported"));
-+ return CC_PC_ILLEGAL;
-+ }
-+
-+ default:
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Extraction not supported"));
-+ return CC_PC_ILLEGAL;
-+ }
-+}
-+
-+static uint8_t GetPrParseCode(e_NetHeaderType hdr, e_FmPcdHdrIndex hdrIndex,
-+ uint32_t offset, bool glblMask,
-+ uint8_t *parseArrayOffset)
-+{
-+ bool offsetRelevant = FALSE;
-+
-+ if (offset)
-+ offsetRelevant = TRUE;
-+
-+ switch (hdr)
-+ {
-+ case (HEADER_TYPE_NONE):
-+ ASSERT_COND(FALSE);
-+ return CC_PC_ILLEGAL;
-+
-+ case (HEADER_TYPE_ETH):
-+ *parseArrayOffset = (uint8_t)CC_PC_PR_ETH_OFFSET;
-+ break;
-+
-+ case (HEADER_TYPE_USER_DEFINED_SHIM1):
-+ if (offset || glblMask)
-+ *parseArrayOffset = (uint8_t)CC_PC_PR_USER_DEFINED_SHIM1_OFFSET;
-+ else
-+ return CC_PC_PR_SHIM1;
-+ break;
-+
-+ case (HEADER_TYPE_USER_DEFINED_SHIM2):
-+ if (offset || glblMask)
-+ *parseArrayOffset = (uint8_t)CC_PC_PR_USER_DEFINED_SHIM2_OFFSET;
-+ else
-+ return CC_PC_PR_SHIM2;
-+ break;
-+
-+ case (HEADER_TYPE_LLC_SNAP):
-+ *parseArrayOffset = CC_PC_PR_USER_LLC_SNAP_OFFSET;
-+ break;
-+
-+ case (HEADER_TYPE_PPPoE):
-+ *parseArrayOffset = CC_PC_PR_PPPOE_OFFSET;
-+ break;
-+
-+ case (HEADER_TYPE_MPLS):
-+ if ((hdrIndex == e_FM_PCD_HDR_INDEX_NONE)
-+ || (hdrIndex == e_FM_PCD_HDR_INDEX_1))
-+ *parseArrayOffset = CC_PC_PR_MPLS1_OFFSET;
-+ else
-+ if (hdrIndex == e_FM_PCD_HDR_INDEX_LAST)
-+ *parseArrayOffset = CC_PC_PR_MPLS_LAST_OFFSET;
-+ else
-+ {
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Illegal MPLS header index"));
-+ return CC_PC_ILLEGAL;
-+ }
-+ break;
-+
-+ case (HEADER_TYPE_IPv4):
-+ case (HEADER_TYPE_IPv6):
-+ if ((hdrIndex == e_FM_PCD_HDR_INDEX_NONE)
-+ || (hdrIndex == e_FM_PCD_HDR_INDEX_1))
-+ *parseArrayOffset = CC_PC_PR_IP1_OFFSET;
-+ else
-+ if (hdrIndex == e_FM_PCD_HDR_INDEX_2)
-+ *parseArrayOffset = CC_PC_PR_IP_LAST_OFFSET;
-+ else
-+ {
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Illegal IP header index"));
-+ return CC_PC_ILLEGAL;
-+ }
-+ break;
-+
-+ case (HEADER_TYPE_MINENCAP):
-+ *parseArrayOffset = CC_PC_PR_MINENC_OFFSET;
-+ break;
-+
-+ case (HEADER_TYPE_GRE):
-+ *parseArrayOffset = CC_PC_PR_GRE_OFFSET;
-+ break;
-+
-+ case (HEADER_TYPE_TCP):
-+ case (HEADER_TYPE_UDP):
-+ case (HEADER_TYPE_IPSEC_AH):
-+ case (HEADER_TYPE_IPSEC_ESP):
-+ case (HEADER_TYPE_DCCP):
-+ case (HEADER_TYPE_SCTP):
-+ *parseArrayOffset = CC_PC_PR_L4_OFFSET;
-+ break;
-+
-+ default:
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Illegal IP header for this type of operation"));
-+ return CC_PC_ILLEGAL;
-+ }
-+
-+ if (offsetRelevant)
-+ return CC_PR_OFFSET;
-+ else
-+ return CC_PR_WITHOUT_OFFSET;
-+}
-+
-+static uint8_t GetFieldParseCode(e_NetHeaderType hdr, t_FmPcdFields field,
-+ uint32_t offset, uint8_t *parseArrayOffset,
-+ e_FmPcdHdrIndex hdrIndex)
-+{
-+ bool offsetRelevant = FALSE;
-+
-+ if (offset)
-+ offsetRelevant = TRUE;
-+
-+ switch (hdr)
-+ {
-+ case (HEADER_TYPE_NONE):
-+ ASSERT_COND(FALSE);
-+ break;
-+ case (HEADER_TYPE_ETH):
-+ switch (field.eth)
-+ {
-+ case (NET_HEADER_FIELD_ETH_TYPE):
-+ *parseArrayOffset = CC_PC_PR_ETYPE_LAST_OFFSET;
-+ break;
-+
-+ default:
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Extraction not supported"));
-+ return CC_PC_ILLEGAL;
-+ }
-+ break;
-+
-+ case (HEADER_TYPE_VLAN):
-+ switch (field.vlan)
-+ {
-+ case (NET_HEADER_FIELD_VLAN_TCI):
-+ if ((hdrIndex == e_FM_PCD_HDR_INDEX_NONE)
-+ || (hdrIndex == e_FM_PCD_HDR_INDEX_1))
-+ *parseArrayOffset = CC_PC_PR_VLAN1_OFFSET;
-+ else
-+ if (hdrIndex == e_FM_PCD_HDR_INDEX_LAST)
-+ *parseArrayOffset = CC_PC_PR_VLAN2_OFFSET;
-+ break;
-+
-+ default:
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Extraction not supported"));
-+ return CC_PC_ILLEGAL;
-+ }
-+ break;
-+
-+ default:
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Illegal header "));
-+ return CC_PC_ILLEGAL;
-+ }
-+
-+ if (offsetRelevant)
-+ return CC_PR_OFFSET;
-+ else
-+ return CC_PR_WITHOUT_OFFSET;
-+}
-+
-+static void FillAdOfTypeResult(t_Handle h_Ad,
-+ t_FmPcdCcStatsParams *p_FmPcdCcStatsParams,
-+ t_FmPcd *p_FmPcd,
-+ t_FmPcdCcNextEngineParams *p_CcNextEngineParams)
-+{
-+ t_AdOfTypeResult *p_AdResult = (t_AdOfTypeResult *)h_Ad;
-+ t_Handle h_TmpAd;
-+ uint32_t tmp = 0, tmpNia = 0;
-+ uint16_t profileId;
-+ t_Handle p_AdNewPtr = NULL;
-+ t_Error err = E_OK;
-+
-+ /* There are 3 cases handled in this routine of building a "result" type AD.
-+ * Case 1: No Manip. The action descriptor is built within the match table.
-+ * Case 2: Manip exists. A new AD is created - p_AdNewPtr. It is initialized
-+ * either in the FmPcdManipUpdateAdResultForCc routine or it was already
-+ * initialized and returned here.
-+ * p_AdResult (within the match table) will be initialized after
-+ * this routine returns and point to the existing AD.
-+ * Case 3: Manip exists. The action descriptor is built within the match table.
-+ * FmPcdManipUpdateAdResultForCc returns a NULL p_AdNewPtr.
-+ *
-+ * If statistics were enabled and the statistics mode of this node requires
-+ * a statistics Ad, it will be placed after the result Ad and before the
-+ * manip Ad, if manip Ad exists here.
-+ */
-+
-+ /* As default, the "new" ptr is the current one. i.e. the content of the result
-+ * AD will be written into the match table itself (case (1))*/
-+ p_AdNewPtr = p_AdResult;
-+
-+ /* Initialize an action descriptor, if current statistics mode requires an Ad */
-+ if (p_FmPcdCcStatsParams)
-+ {
-+ ASSERT_COND(p_FmPcdCcStatsParams->h_StatsAd);
-+ ASSERT_COND(p_FmPcdCcStatsParams->h_StatsCounters);
-+
-+ /* Swapping addresses between statistics Ad and the current lookup AD addresses */
-+ h_TmpAd = p_FmPcdCcStatsParams->h_StatsAd;
-+ p_FmPcdCcStatsParams->h_StatsAd = h_Ad;
-+ h_Ad = h_TmpAd;
-+
-+ p_AdNewPtr = h_Ad;
-+ p_AdResult = h_Ad;
-+
-+ /* Init statistics Ad and connect current lookup AD as 'next action' from statistics Ad */
-+ UpdateStatsAd(p_FmPcdCcStatsParams, h_Ad, p_FmPcd->physicalMuramBase);
-+ }
-+
-+ /* Create manip and return p_AdNewPtr to either a new descriptor or NULL */
-+ if (p_CcNextEngineParams->h_Manip)
-+ FmPcdManipUpdateAdResultForCc(p_CcNextEngineParams->h_Manip,
-+ p_CcNextEngineParams, h_Ad, &p_AdNewPtr);
-+
-+ /* if (p_AdNewPtr = NULL) --> Done. (case (3)) */
-+ if (p_AdNewPtr)
-+ {
-+ /* case (1) and (2) */
-+ switch (p_CcNextEngineParams->nextEngine)
-+ {
-+ case (e_FM_PCD_DONE):
-+ if (p_CcNextEngineParams->params.enqueueParams.action
-+ == e_FM_PCD_ENQ_FRAME)
-+ {
-+ if (p_CcNextEngineParams->params.enqueueParams.overrideFqid)
-+ {
-+ tmp = FM_PCD_AD_RESULT_CONTRL_FLOW_TYPE;
-+ tmp |=
-+ p_CcNextEngineParams->params.enqueueParams.newFqid;
-+#if (DPAA_VERSION >= 11)
-+ tmp |=
-+ (p_CcNextEngineParams->params.enqueueParams.newRelativeStorageProfileId
-+ & FM_PCD_AD_RESULT_VSP_MASK)
-+ << FM_PCD_AD_RESULT_VSP_SHIFT;
-+#endif /* (DPAA_VERSION >= 11) */
-+ }
-+ else
-+ {
-+ tmp = FM_PCD_AD_RESULT_DATA_FLOW_TYPE;
-+ tmp |= FM_PCD_AD_RESULT_PLCR_DIS;
-+ }
-+ }
-+
-+ if (p_CcNextEngineParams->params.enqueueParams.action
-+ == e_FM_PCD_DROP_FRAME)
-+ tmpNia |= GET_NIA_BMI_AC_DISCARD_FRAME(p_FmPcd);
-+ else
-+ tmpNia |= GET_NIA_BMI_AC_ENQ_FRAME(p_FmPcd);
-+ break;
-+
-+ case (e_FM_PCD_KG):
-+ if (p_CcNextEngineParams->params.kgParams.overrideFqid)
-+ {
-+ tmp = FM_PCD_AD_RESULT_CONTRL_FLOW_TYPE;
-+ tmp |= p_CcNextEngineParams->params.kgParams.newFqid;
-+#if (DPAA_VERSION >= 11)
-+ tmp |=
-+ (p_CcNextEngineParams->params.kgParams.newRelativeStorageProfileId
-+ & FM_PCD_AD_RESULT_VSP_MASK)
-+ << FM_PCD_AD_RESULT_VSP_SHIFT;
-+#endif /* (DPAA_VERSION >= 11) */
-+ }
-+ else
-+ {
-+ tmp = FM_PCD_AD_RESULT_DATA_FLOW_TYPE;
-+ tmp |= FM_PCD_AD_RESULT_PLCR_DIS;
-+ }
-+ tmpNia = NIA_KG_DIRECT;
-+ tmpNia |= NIA_ENG_KG;
-+ tmpNia |= NIA_KG_CC_EN;
-+ tmpNia |= FmPcdKgGetSchemeId(
-+ p_CcNextEngineParams->params.kgParams.h_DirectScheme);
-+ break;
-+
-+ case (e_FM_PCD_PLCR):
-+ if (p_CcNextEngineParams->params.plcrParams.overrideParams)
-+ {
-+ tmp = FM_PCD_AD_RESULT_CONTRL_FLOW_TYPE;
-+
-+ /* if private policer profile, it may be uninitialized yet, therefore no checks are done at this stage */
-+ if (p_CcNextEngineParams->params.plcrParams.sharedProfile)
-+ {
-+ tmpNia |= NIA_PLCR_ABSOLUTE;
-+ err = FmPcdPlcrGetAbsoluteIdByProfileParams(
-+ (t_Handle)p_FmPcd,
-+ e_FM_PCD_PLCR_SHARED,
-+ NULL,
-+ p_CcNextEngineParams->params.plcrParams.newRelativeProfileId,
-+ &profileId);
-+
-+ if (err != E_OK) {
-+ REPORT_ERROR(MAJOR, err, NO_MSG);
-+ return;
-+ }
-+
-+ }
-+ else
-+ profileId =
-+ p_CcNextEngineParams->params.plcrParams.newRelativeProfileId;
-+
-+ tmp |= p_CcNextEngineParams->params.plcrParams.newFqid;
-+#if (DPAA_VERSION >= 11)
-+ tmp |=
-+ (p_CcNextEngineParams->params.plcrParams.newRelativeStorageProfileId
-+ & FM_PCD_AD_RESULT_VSP_MASK)
-+ << FM_PCD_AD_RESULT_VSP_SHIFT;
-+#endif /* (DPAA_VERSION >= 11) */
-+ WRITE_UINT32(
-+ p_AdResult->plcrProfile,
-+ (uint32_t)((uint32_t)profileId << FM_PCD_AD_PROFILEID_FOR_CNTRL_SHIFT));
-+ }
-+ else
-+ tmp = FM_PCD_AD_RESULT_DATA_FLOW_TYPE;
-+
-+ tmpNia |=
-+ NIA_ENG_PLCR
-+ | p_CcNextEngineParams->params.plcrParams.newRelativeProfileId;
-+ break;
-+
-+ default:
-+ return;
-+ }WRITE_UINT32(p_AdResult->fqid, tmp);
-+
-+ if (p_CcNextEngineParams->h_Manip)
-+ {
-+ tmp = GET_UINT32(p_AdResult->plcrProfile);
-+ tmp |= (uint32_t)(XX_VirtToPhys(p_AdNewPtr)
-+ - (p_FmPcd->physicalMuramBase)) >> 4;
-+ WRITE_UINT32(p_AdResult->plcrProfile, tmp);
-+
-+ tmpNia |= FM_PCD_AD_RESULT_EXTENDED_MODE;
-+ tmpNia |= FM_PCD_AD_RESULT_NADEN;
-+ }
-+
-+#if (DPAA_VERSION >= 11)
-+ tmpNia |= FM_PCD_AD_RESULT_NO_OM_VSPE;
-+#endif /* (DPAA_VERSION >= 11) */
-+ WRITE_UINT32(p_AdResult->nia, tmpNia);
-+ }
-+}
-+
-+static t_Error CcUpdateParams(t_Handle h_FmPcd, t_Handle h_PcdParams,
-+ t_Handle h_FmPort, t_Handle h_FmTree,
-+ bool validate)
-+{
-+ t_FmPcdCcTree *p_CcTree = (t_FmPcdCcTree *)h_FmTree;
-+
-+ return CcUpdateParam(h_FmPcd, h_PcdParams, h_FmPort,
-+ p_CcTree->keyAndNextEngineParams,
-+ p_CcTree->numOfEntries,
-+ UINT_TO_PTR(p_CcTree->ccTreeBaseAddr), validate, 0,
-+ h_FmTree, FALSE);
-+}
-+
-+
-+static void ReleaseNewNodeCommonPart(
-+ t_FmPcdModifyCcKeyAdditionalParams *p_AdditionalInfo)
-+{
-+ if (p_AdditionalInfo->p_AdTableNew)
-+ FM_MURAM_FreeMem(
-+ FmPcdGetMuramHandle(
-+ ((t_FmPcdCcNode *)(p_AdditionalInfo->h_CurrentNode))->h_FmPcd),
-+ p_AdditionalInfo->p_AdTableNew);
-+
-+ if (p_AdditionalInfo->p_KeysMatchTableNew)
-+ FM_MURAM_FreeMem(
-+ FmPcdGetMuramHandle(
-+ ((t_FmPcdCcNode *)(p_AdditionalInfo->h_CurrentNode))->h_FmPcd),
-+ p_AdditionalInfo->p_KeysMatchTableNew);
-+}
-+
-+static t_Error UpdateGblMask(t_FmPcdCcNode *p_CcNode, uint8_t keySize,
-+ uint8_t *p_Mask)
-+{
-+ uint8_t prvGlblMaskSize = p_CcNode->glblMaskSize;
-+
-+ if (p_Mask && !p_CcNode->glblMaskUpdated && (keySize <= 4)
-+ && !p_CcNode->lclMask)
-+ {
-+ if (p_CcNode->parseCode && (p_CcNode->parseCode != CC_PC_FF_TCI1)
-+ && (p_CcNode->parseCode != CC_PC_FF_TCI2)
-+ && (p_CcNode->parseCode != CC_PC_FF_MPLS1)
-+ && (p_CcNode->parseCode != CC_PC_FF_MPLS_LAST)
-+ && (p_CcNode->parseCode != CC_PC_FF_IPV4IPTOS_TC1)
-+ && (p_CcNode->parseCode != CC_PC_FF_IPV4IPTOS_TC2)
-+ && (p_CcNode->parseCode != CC_PC_FF_IPTOS_IPV6TC1_IPV6FLOW1)
-+ && (p_CcNode->parseCode != CC_PC_FF_IPDSCP)
-+ && (p_CcNode->parseCode != CC_PC_FF_IPTOS_IPV6TC2_IPV6FLOW2))
-+ {
-+ p_CcNode->glblMaskSize = 0;
-+ p_CcNode->lclMask = TRUE;
-+ }
-+ else
-+ {
-+ memcpy(p_CcNode->p_GlblMask, p_Mask, (sizeof(uint8_t)) * keySize);
-+ p_CcNode->glblMaskUpdated = TRUE;
-+ p_CcNode->glblMaskSize = 4;
-+ }
-+ }
-+ else
-+ if (p_Mask && (keySize <= 4) && !p_CcNode->lclMask)
-+ {
-+ if (memcmp(p_CcNode->p_GlblMask, p_Mask, keySize) != 0)
-+ {
-+ p_CcNode->lclMask = TRUE;
-+ p_CcNode->glblMaskSize = 0;
-+ }
-+ }
-+ else
-+ if (!p_Mask && p_CcNode->glblMaskUpdated && (keySize <= 4))
-+ {
-+ uint32_t tmpMask = 0xffffffff;
-+ if (memcmp(p_CcNode->p_GlblMask, &tmpMask, 4) != 0)
-+ {
-+ p_CcNode->lclMask = TRUE;
-+ p_CcNode->glblMaskSize = 0;
-+ }
-+ }
-+ else
-+ if (p_Mask)
-+ {
-+ p_CcNode->lclMask = TRUE;
-+ p_CcNode->glblMaskSize = 0;
-+ }
-+
-+ /* In static mode (maxNumOfKeys > 0), local mask is supported
-+ only is mask support was enabled at initialization */
-+ if (p_CcNode->maxNumOfKeys && (!p_CcNode->maskSupport) && p_CcNode->lclMask)
-+ {
-+ p_CcNode->lclMask = FALSE;
-+ p_CcNode->glblMaskSize = prvGlblMaskSize;
-+ return ERROR_CODE(E_NOT_SUPPORTED);
-+ }
-+
-+ return E_OK;
-+}
-+
-+static __inline__ t_Handle GetNewAd(t_Handle h_FmPcdCcNodeOrTree, bool isTree)
-+{
-+ t_FmPcd *p_FmPcd;
-+ t_Handle h_Ad;
-+
-+ if (isTree)
-+ p_FmPcd = (t_FmPcd *)(((t_FmPcdCcTree *)h_FmPcdCcNodeOrTree)->h_FmPcd);
-+ else
-+ p_FmPcd = (t_FmPcd *)(((t_FmPcdCcNode *)h_FmPcdCcNodeOrTree)->h_FmPcd);
-+
-+ if ((isTree && p_FmPcd->p_CcShadow)
-+ || (!isTree && ((t_FmPcdCcNode *)h_FmPcdCcNodeOrTree)->maxNumOfKeys))
-+ {
-+ /* The allocated shadow is divided as follows:
-+ 0 . . . 16 . . .
-+ ---------------------------------------------------
-+ | Shadow | Shadow Keys | Shadow Next |
-+ | Ad | Match Table | Engine Table |
-+ | (16 bytes) | (maximal size) | (maximal size) |
-+ ---------------------------------------------------
-+ */
-+ if (!p_FmPcd->p_CcShadow)
-+ {
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("CC Shadow not allocated"));
-+ return NULL;
-+ }
-+
-+ h_Ad = p_FmPcd->p_CcShadow;
-+ }
-+ else
-+ {
-+ h_Ad = (t_Handle)FM_MURAM_AllocMem(FmPcdGetMuramHandle(p_FmPcd),
-+ FM_PCD_CC_AD_ENTRY_SIZE,
-+ FM_PCD_CC_AD_TABLE_ALIGN);
-+ if (!h_Ad)
-+ {
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("MURAM allocation for CC node action descriptor"));
-+ return NULL;
-+ }
-+ }
-+
-+ return h_Ad;
-+}
-+
-+static t_Error BuildNewNodeCommonPart(
-+ t_FmPcdCcNode *p_CcNode, int *size,
-+ t_FmPcdModifyCcKeyAdditionalParams *p_AdditionalInfo)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd *)p_CcNode->h_FmPcd;
-+
-+ if (p_CcNode->lclMask)
-+ *size = 2 * p_CcNode->ccKeySizeAccExtraction;
-+ else
-+ *size = p_CcNode->ccKeySizeAccExtraction;
-+
-+ if (p_CcNode->maxNumOfKeys == 0)
-+ {
-+ p_AdditionalInfo->p_AdTableNew = (t_Handle)FM_MURAM_AllocMem(
-+ FmPcdGetMuramHandle(p_FmPcd),
-+ (uint32_t)((p_AdditionalInfo->numOfKeys + 1)
-+ * FM_PCD_CC_AD_ENTRY_SIZE),
-+ FM_PCD_CC_AD_TABLE_ALIGN);
-+ if (!p_AdditionalInfo->p_AdTableNew)
-+ RETURN_ERROR(
-+ MAJOR, E_NO_MEMORY,
-+ ("MURAM allocation for CC node action descriptors table"));
-+
-+ p_AdditionalInfo->p_KeysMatchTableNew = (t_Handle)FM_MURAM_AllocMem(
-+ FmPcdGetMuramHandle(p_FmPcd),
-+ (uint32_t)(*size * sizeof(uint8_t)
-+ * (p_AdditionalInfo->numOfKeys + 1)),
-+ FM_PCD_CC_KEYS_MATCH_TABLE_ALIGN);
-+ if (!p_AdditionalInfo->p_KeysMatchTableNew)
-+ {
-+ FM_MURAM_FreeMem(FmPcdGetMuramHandle(p_CcNode->h_FmPcd),
-+ p_AdditionalInfo->p_AdTableNew);
-+ p_AdditionalInfo->p_AdTableNew = NULL;
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY,
-+ ("MURAM allocation for CC node key match table"));
-+ }
-+
-+ MemSet8(
-+ (uint8_t*)p_AdditionalInfo->p_AdTableNew,
-+ 0,
-+ (uint32_t)((p_AdditionalInfo->numOfKeys + 1)
-+ * FM_PCD_CC_AD_ENTRY_SIZE));
-+ MemSet8((uint8_t*)p_AdditionalInfo->p_KeysMatchTableNew, 0,
-+ *size * sizeof(uint8_t) * (p_AdditionalInfo->numOfKeys + 1));
-+ }
-+ else
-+ {
-+ /* The allocated shadow is divided as follows:
-+ 0 . . . 16 . . .
-+ ---------------------------------------------------
-+ | Shadow | Shadow Keys | Shadow Next |
-+ | Ad | Match Table | Engine Table |
-+ | (16 bytes) | (maximal size) | (maximal size) |
-+ ---------------------------------------------------
-+ */
-+
-+ if (!p_FmPcd->p_CcShadow)
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, ("CC Shadow not allocated"));
-+
-+ p_AdditionalInfo->p_KeysMatchTableNew =
-+ PTR_MOVE(p_FmPcd->p_CcShadow, FM_PCD_CC_AD_ENTRY_SIZE);
-+ p_AdditionalInfo->p_AdTableNew =
-+ PTR_MOVE(p_AdditionalInfo->p_KeysMatchTableNew, p_CcNode->keysMatchTableMaxSize);
-+
-+ MemSet8(
-+ (uint8_t*)p_AdditionalInfo->p_AdTableNew,
-+ 0,
-+ (uint32_t)((p_CcNode->maxNumOfKeys + 1)
-+ * FM_PCD_CC_AD_ENTRY_SIZE));
-+ MemSet8((uint8_t*)p_AdditionalInfo->p_KeysMatchTableNew, 0,
-+ (*size) * sizeof(uint8_t) * (p_CcNode->maxNumOfKeys));
-+ }
-+
-+ p_AdditionalInfo->p_AdTableOld = p_CcNode->h_AdTable;
-+ p_AdditionalInfo->p_KeysMatchTableOld = p_CcNode->h_KeysMatchTable;
-+
-+ return E_OK;
-+}
-+
-+static t_Error BuildNewNodeAddOrMdfyKeyAndNextEngine(
-+ t_Handle h_FmPcd, t_FmPcdCcNode *p_CcNode, uint16_t keyIndex,
-+ t_FmPcdCcKeyParams *p_KeyParams,
-+ t_FmPcdModifyCcKeyAdditionalParams *p_AdditionalInfo, bool add)
-+{
-+ t_Error err = E_OK;
-+ t_Handle p_AdTableNewTmp, p_KeysMatchTableNewTmp;
-+ t_Handle p_KeysMatchTableOldTmp, p_AdTableOldTmp;
-+ int size;
-+ int i = 0, j = 0;
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+ uint32_t requiredAction = 0;
-+ bool prvLclMask;
-+ t_CcNodeInformation *p_CcNodeInformation;
-+ t_FmPcdCcStatsParams statsParams = { 0 };
-+ t_List *p_Pos;
-+ t_FmPcdStatsObj *p_StatsObj;
-+
-+ /* Check that new NIA is legal */
-+ err = ValidateNextEngineParams(h_FmPcd, &p_KeyParams->ccNextEngineParams,
-+ p_CcNode->statisticsMode);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+
-+ prvLclMask = p_CcNode->lclMask;
-+
-+ /* Check that new key is not require update of localMask */
-+ err = UpdateGblMask(p_CcNode, p_CcNode->ccKeySizeAccExtraction,
-+ p_KeyParams->p_Mask);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, (NO_MSG));
-+
-+ /* Update internal data structure with new next engine for the given index */
-+ memcpy(&p_AdditionalInfo->keyAndNextEngineParams[keyIndex].nextEngineParams,
-+ &p_KeyParams->ccNextEngineParams, sizeof(t_FmPcdCcNextEngineParams));
-+
-+ memcpy(p_AdditionalInfo->keyAndNextEngineParams[keyIndex].key,
-+ p_KeyParams->p_Key, p_CcNode->userSizeOfExtraction);
-+
-+ if ((p_AdditionalInfo->keyAndNextEngineParams[keyIndex].nextEngineParams.nextEngine
-+ == e_FM_PCD_CC)
-+ && p_AdditionalInfo->keyAndNextEngineParams[keyIndex].nextEngineParams.h_Manip)
-+ {
-+ err =
-+ AllocAndFillAdForContLookupManip(
-+ p_AdditionalInfo->keyAndNextEngineParams[keyIndex].nextEngineParams.params.ccParams.h_CcNode);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, (NO_MSG));
-+ }
-+
-+ if (p_KeyParams->p_Mask)
-+ memcpy(p_AdditionalInfo->keyAndNextEngineParams[keyIndex].mask,
-+ p_KeyParams->p_Mask, p_CcNode->userSizeOfExtraction);
-+ else
-+ memset(p_AdditionalInfo->keyAndNextEngineParams[keyIndex].mask, 0xFF,
-+ p_CcNode->userSizeOfExtraction);
-+
-+ /* Update numOfKeys */
-+ if (add)
-+ p_AdditionalInfo->numOfKeys = (uint8_t)(p_CcNode->numOfKeys + 1);
-+ else
-+ p_AdditionalInfo->numOfKeys = (uint8_t)p_CcNode->numOfKeys;
-+
-+ /* Allocate new tables in MURAM: keys match table and action descriptors table */
-+ err = BuildNewNodeCommonPart(p_CcNode, &size, p_AdditionalInfo);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+
-+ /* Check that manip is legal and what requiredAction is necessary for this manip */
-+ if (p_KeyParams->ccNextEngineParams.h_Manip)
-+ {
-+ err = FmPcdManipCheckParamsForCcNextEngine(
-+ &p_KeyParams->ccNextEngineParams, &requiredAction);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, (NO_MSG));
-+ }
-+
-+ p_AdditionalInfo->keyAndNextEngineParams[keyIndex].requiredAction =
-+ requiredAction;
-+ p_AdditionalInfo->keyAndNextEngineParams[keyIndex].requiredAction |=
-+ UPDATE_CC_WITH_TREE;
-+
-+ /* Update new Ad and new Key Table according to new requirement */
-+ i = 0;
-+ for (j = 0; j < p_AdditionalInfo->numOfKeys; j++)
-+ {
-+ p_AdTableNewTmp =
-+ PTR_MOVE(p_AdditionalInfo->p_AdTableNew, j*FM_PCD_CC_AD_ENTRY_SIZE);
-+
-+ if (j == keyIndex)
-+ {
-+ if (p_KeyParams->ccNextEngineParams.statisticsEn)
-+ {
-+ /* Allocate a statistics object that holds statistics AD and counters.
-+ - For added key - New statistics AD and counters pointer need to be allocated
-+ new statistics object. If statistics were enabled, we need to replace the
-+ existing descriptor with a new descriptor with nullified counters.
-+ */
-+ p_StatsObj = GetStatsObj(p_CcNode);
-+ ASSERT_COND(p_StatsObj);
-+
-+ /* Store allocated statistics object */
-+ ASSERT_COND(keyIndex < CC_MAX_NUM_OF_KEYS);
-+ p_AdditionalInfo->keyAndNextEngineParams[keyIndex].p_StatsObj =
-+ p_StatsObj;
-+
-+ statsParams.h_StatsAd = p_StatsObj->h_StatsAd;
-+ statsParams.h_StatsCounters = p_StatsObj->h_StatsCounters;
-+#if (DPAA_VERSION >= 11)
-+ statsParams.h_StatsFLRs = p_CcNode->h_StatsFLRs;
-+
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+ /* Building action descriptor for the received new key */
-+ NextStepAd(p_AdTableNewTmp, &statsParams,
-+ &p_KeyParams->ccNextEngineParams, p_FmPcd);
-+ }
-+ else
-+ {
-+ /* Building action descriptor for the received new key */
-+ NextStepAd(p_AdTableNewTmp, NULL,
-+ &p_KeyParams->ccNextEngineParams, p_FmPcd);
-+ }
-+
-+ /* Copy the received new key into keys match table */
-+ p_KeysMatchTableNewTmp =
-+ PTR_MOVE(p_AdditionalInfo->p_KeysMatchTableNew, j*size*sizeof(uint8_t));
-+
-+ MemCpy8((void*)p_KeysMatchTableNewTmp, p_KeyParams->p_Key,
-+ p_CcNode->userSizeOfExtraction);
-+
-+ /* Update mask for the received new key */
-+ if (p_CcNode->lclMask)
-+ {
-+ if (p_KeyParams->p_Mask)
-+ {
-+ MemCpy8(PTR_MOVE(p_KeysMatchTableNewTmp,
-+ p_CcNode->ccKeySizeAccExtraction),
-+ p_KeyParams->p_Mask,
-+ p_CcNode->userSizeOfExtraction);
-+ }
-+ else
-+ if (p_CcNode->ccKeySizeAccExtraction > 4)
-+ {
-+ MemSet8(PTR_MOVE(p_KeysMatchTableNewTmp,
-+ p_CcNode->ccKeySizeAccExtraction),
-+ 0xff, p_CcNode->userSizeOfExtraction);
-+ }
-+ else
-+ {
-+ MemCpy8(PTR_MOVE(p_KeysMatchTableNewTmp,
-+ p_CcNode->ccKeySizeAccExtraction),
-+ p_CcNode->p_GlblMask,
-+ p_CcNode->userSizeOfExtraction);
-+ }
-+ }
-+
-+ /* If key modification requested, the old entry is omitted and replaced by the new parameters */
-+ if (!add)
-+ i++;
-+ }
-+ else
-+ {
-+ /* Copy existing action descriptors to the newly allocated Ad table */
-+ p_AdTableOldTmp =
-+ PTR_MOVE(p_AdditionalInfo->p_AdTableOld, i*FM_PCD_CC_AD_ENTRY_SIZE);
-+ MemCpy8(p_AdTableNewTmp, p_AdTableOldTmp,
-+ FM_PCD_CC_AD_ENTRY_SIZE);
-+
-+ /* Copy existing keys and their masks to the newly allocated keys match table */
-+ p_KeysMatchTableNewTmp =
-+ PTR_MOVE(p_AdditionalInfo->p_KeysMatchTableNew, j * size * sizeof(uint8_t));
-+ p_KeysMatchTableOldTmp =
-+ PTR_MOVE(p_AdditionalInfo->p_KeysMatchTableOld, i * size * sizeof(uint8_t));
-+
-+ if (p_CcNode->lclMask)
-+ {
-+ if (prvLclMask)
-+ {
-+ MemCpy8(
-+ PTR_MOVE(p_KeysMatchTableNewTmp, p_CcNode->ccKeySizeAccExtraction),
-+ PTR_MOVE(p_KeysMatchTableOldTmp, p_CcNode->ccKeySizeAccExtraction),
-+ p_CcNode->ccKeySizeAccExtraction);
-+ }
-+ else
-+ {
-+ p_KeysMatchTableOldTmp =
-+ PTR_MOVE(p_CcNode->h_KeysMatchTable,
-+ i * (int)p_CcNode->ccKeySizeAccExtraction * sizeof(uint8_t));
-+
-+ if (p_CcNode->ccKeySizeAccExtraction > 4)
-+ {
-+ MemSet8(PTR_MOVE(p_KeysMatchTableNewTmp,
-+ p_CcNode->ccKeySizeAccExtraction),
-+ 0xff, p_CcNode->userSizeOfExtraction);
-+ }
-+ else
-+ {
-+ MemCpy8(PTR_MOVE(p_KeysMatchTableNewTmp,
-+ p_CcNode->ccKeySizeAccExtraction),
-+ p_CcNode->p_GlblMask,
-+ p_CcNode->userSizeOfExtraction);
-+ }
-+ }
-+ }
-+
-+ MemCpy8(p_KeysMatchTableNewTmp, p_KeysMatchTableOldTmp,
-+ p_CcNode->ccKeySizeAccExtraction);
-+
-+ i++;
-+ }
-+ }
-+
-+ /* Miss action descriptor */
-+ p_AdTableNewTmp =
-+ PTR_MOVE(p_AdditionalInfo->p_AdTableNew, j * FM_PCD_CC_AD_ENTRY_SIZE);
-+ p_AdTableOldTmp =
-+ PTR_MOVE(p_AdditionalInfo->p_AdTableOld, i * FM_PCD_CC_AD_ENTRY_SIZE);
-+ MemCpy8(p_AdTableNewTmp, p_AdTableOldTmp, FM_PCD_CC_AD_ENTRY_SIZE);
-+
-+ if (!LIST_IsEmpty(&p_CcNode->ccTreesLst))
-+ {
-+ LIST_FOR_EACH(p_Pos, &p_CcNode->ccTreesLst)
-+ {
-+ p_CcNodeInformation = CC_NODE_F_OBJECT(p_Pos);
-+ ASSERT_COND(p_CcNodeInformation->h_CcNode);
-+ /* Update the manipulation which has to be updated from parameters of the port */
-+ /* It's has to be updated with restrictions defined in the function */
-+ err =
-+ SetRequiredAction(
-+ p_CcNode->h_FmPcd,
-+ p_CcNode->shadowAction
-+ | p_AdditionalInfo->keyAndNextEngineParams[keyIndex].requiredAction,
-+ &p_AdditionalInfo->keyAndNextEngineParams[keyIndex],
-+ PTR_MOVE(p_AdditionalInfo->p_AdTableNew, keyIndex*FM_PCD_CC_AD_ENTRY_SIZE),
-+ 1, p_CcNodeInformation->h_CcNode);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, (NO_MSG));
-+
-+ err =
-+ CcUpdateParam(
-+ p_CcNode->h_FmPcd,
-+ NULL,
-+ NULL,
-+ &p_AdditionalInfo->keyAndNextEngineParams[keyIndex],
-+ 1,
-+ PTR_MOVE(p_AdditionalInfo->p_AdTableNew, keyIndex*FM_PCD_CC_AD_ENTRY_SIZE),
-+ TRUE, p_CcNodeInformation->index,
-+ p_CcNodeInformation->h_CcNode, TRUE);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, (NO_MSG));
-+ }
-+ }
-+
-+ if (p_CcNode->lclMask)
-+ memset(p_CcNode->p_GlblMask, 0xff, CC_GLBL_MASK_SIZE * sizeof(uint8_t));
-+
-+ if (p_KeyParams->ccNextEngineParams.nextEngine == e_FM_PCD_CC)
-+ p_AdditionalInfo->h_NodeForAdd =
-+ p_KeyParams->ccNextEngineParams.params.ccParams.h_CcNode;
-+ if (p_KeyParams->ccNextEngineParams.h_Manip)
-+ p_AdditionalInfo->h_ManipForAdd =
-+ p_KeyParams->ccNextEngineParams.h_Manip;
-+
-+#if (DPAA_VERSION >= 11)
-+ if ((p_KeyParams->ccNextEngineParams.nextEngine == e_FM_PCD_FR)
-+ && (p_KeyParams->ccNextEngineParams.params.frParams.h_FrmReplic))
-+ p_AdditionalInfo->h_FrmReplicForAdd =
-+ p_KeyParams->ccNextEngineParams.params.frParams.h_FrmReplic;
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+ if (!add)
-+ {
-+ if (p_CcNode->keyAndNextEngineParams[keyIndex].nextEngineParams.nextEngine
-+ == e_FM_PCD_CC)
-+ p_AdditionalInfo->h_NodeForRmv =
-+ p_CcNode->keyAndNextEngineParams[keyIndex].nextEngineParams.params.ccParams.h_CcNode;
-+
-+ if (p_CcNode->keyAndNextEngineParams[keyIndex].nextEngineParams.h_Manip)
-+ p_AdditionalInfo->h_ManipForRmv =
-+ p_CcNode->keyAndNextEngineParams[keyIndex].nextEngineParams.h_Manip;
-+
-+ /* If statistics were previously enabled, store the old statistics object to be released */
-+ if (p_CcNode->keyAndNextEngineParams[keyIndex].p_StatsObj)
-+ {
-+ p_AdditionalInfo->p_StatsObjForRmv =
-+ p_CcNode->keyAndNextEngineParams[keyIndex].p_StatsObj;
-+ }
-+
-+#if (DPAA_VERSION >= 11)
-+ if ((p_CcNode->keyAndNextEngineParams[keyIndex].nextEngineParams.nextEngine
-+ == e_FM_PCD_FR)
-+ && (p_CcNode->keyAndNextEngineParams[keyIndex].nextEngineParams.params.frParams.h_FrmReplic))
-+ p_AdditionalInfo->h_FrmReplicForRmv =
-+ p_CcNode->keyAndNextEngineParams[keyIndex].nextEngineParams.params.frParams.h_FrmReplic;
-+#endif /* (DPAA_VERSION >= 11) */
-+ }
-+
-+ return E_OK;
-+}
-+
-+static t_Error BuildNewNodeRemoveKey(
-+ t_FmPcdCcNode *p_CcNode, uint16_t keyIndex,
-+ t_FmPcdModifyCcKeyAdditionalParams *p_AdditionalInfo)
-+{
-+ int i = 0, j = 0;
-+ t_Handle p_AdTableNewTmp, p_KeysMatchTableNewTmp;
-+ t_Handle p_KeysMatchTableOldTmp, p_AdTableOldTmp;
-+ int size;
-+ t_Error err = E_OK;
-+
-+ /*save new numOfKeys*/
-+ p_AdditionalInfo->numOfKeys = (uint16_t)(p_CcNode->numOfKeys - 1);
-+
-+ /*function which allocates in the memory new KeyTbl, AdTbl*/
-+ err = BuildNewNodeCommonPart(p_CcNode, &size, p_AdditionalInfo);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+
-+ /*update new Ad and new Key Table according to new requirement*/
-+ for (i = 0, j = 0; j < p_CcNode->numOfKeys; i++, j++)
-+ {
-+ if (j == keyIndex)
-+ j++;
-+
-+ if (j == p_CcNode->numOfKeys)
-+ break;
-+ p_AdTableNewTmp =
-+ PTR_MOVE(p_AdditionalInfo->p_AdTableNew, i * FM_PCD_CC_AD_ENTRY_SIZE);
-+ p_AdTableOldTmp =
-+ PTR_MOVE(p_AdditionalInfo->p_AdTableOld, j * FM_PCD_CC_AD_ENTRY_SIZE);
-+ MemCpy8(p_AdTableNewTmp, p_AdTableOldTmp, FM_PCD_CC_AD_ENTRY_SIZE);
-+
-+ p_KeysMatchTableOldTmp =
-+ PTR_MOVE(p_AdditionalInfo->p_KeysMatchTableOld, j * size * sizeof(uint8_t));
-+ p_KeysMatchTableNewTmp =
-+ PTR_MOVE(p_AdditionalInfo->p_KeysMatchTableNew, i * size * sizeof(uint8_t));
-+ MemCpy8(p_KeysMatchTableNewTmp, p_KeysMatchTableOldTmp,
-+ size * sizeof(uint8_t));
-+ }
-+
-+ p_AdTableNewTmp =
-+ PTR_MOVE(p_AdditionalInfo->p_AdTableNew, i * FM_PCD_CC_AD_ENTRY_SIZE);
-+ p_AdTableOldTmp =
-+ PTR_MOVE(p_AdditionalInfo->p_AdTableOld, j * FM_PCD_CC_AD_ENTRY_SIZE);
-+ MemCpy8(p_AdTableNewTmp, p_AdTableOldTmp, FM_PCD_CC_AD_ENTRY_SIZE);
-+
-+ if (p_CcNode->keyAndNextEngineParams[keyIndex].nextEngineParams.nextEngine
-+ == e_FM_PCD_CC)
-+ p_AdditionalInfo->h_NodeForRmv =
-+ p_CcNode->keyAndNextEngineParams[keyIndex].nextEngineParams.params.ccParams.h_CcNode;
-+
-+ if (p_CcNode->keyAndNextEngineParams[keyIndex].nextEngineParams.h_Manip)
-+ p_AdditionalInfo->h_ManipForRmv =
-+ p_CcNode->keyAndNextEngineParams[keyIndex].nextEngineParams.h_Manip;
-+
-+ /* If statistics were previously enabled, store the old statistics object to be released */
-+ if (p_CcNode->keyAndNextEngineParams[keyIndex].p_StatsObj)
-+ {
-+ p_AdditionalInfo->p_StatsObjForRmv =
-+ p_CcNode->keyAndNextEngineParams[keyIndex].p_StatsObj;
-+ }
-+
-+#if (DPAA_VERSION >= 11)
-+ if ((p_CcNode->keyAndNextEngineParams[keyIndex].nextEngineParams.nextEngine
-+ == e_FM_PCD_FR)
-+ && (p_CcNode->keyAndNextEngineParams[keyIndex].nextEngineParams.params.frParams.h_FrmReplic))
-+ p_AdditionalInfo->h_FrmReplicForRmv =
-+ p_CcNode->keyAndNextEngineParams[keyIndex].nextEngineParams.params.frParams.h_FrmReplic;
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+ return E_OK;
-+}
-+
-+static t_Error BuildNewNodeModifyKey(
-+ t_FmPcdCcNode *p_CcNode, uint16_t keyIndex, uint8_t *p_Key,
-+ uint8_t *p_Mask, t_FmPcdModifyCcKeyAdditionalParams *p_AdditionalInfo)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd *)p_CcNode->h_FmPcd;
-+ t_Error err = E_OK;
-+ t_Handle p_AdTableNewTmp, p_KeysMatchTableNewTmp;
-+ t_Handle p_KeysMatchTableOldTmp, p_AdTableOldTmp;
-+ int size;
-+ int i = 0, j = 0;
-+ bool prvLclMask;
-+ t_FmPcdStatsObj *p_StatsObj, tmpStatsObj;
-+ p_AdditionalInfo->numOfKeys = p_CcNode->numOfKeys;
-+
-+ prvLclMask = p_CcNode->lclMask;
-+
-+ /* Check that new key is not require update of localMask */
-+ err = UpdateGblMask(p_CcNode, p_CcNode->ccKeySizeAccExtraction, p_Mask);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, (NO_MSG));
-+
-+ /* Update internal data structure with new next engine for the given index */
-+ memcpy(p_AdditionalInfo->keyAndNextEngineParams[keyIndex].key, p_Key,
-+ p_CcNode->userSizeOfExtraction);
-+
-+ if (p_Mask)
-+ memcpy(p_AdditionalInfo->keyAndNextEngineParams[keyIndex].mask, p_Mask,
-+ p_CcNode->userSizeOfExtraction);
-+ else
-+ memset(p_AdditionalInfo->keyAndNextEngineParams[keyIndex].mask, 0xFF,
-+ p_CcNode->userSizeOfExtraction);
-+
-+ /*function which build in the memory new KeyTbl, AdTbl*/
-+ err = BuildNewNodeCommonPart(p_CcNode, &size, p_AdditionalInfo);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+
-+ /*fill the New AdTable and New KeyTable*/
-+ for (j = 0, i = 0; j < p_AdditionalInfo->numOfKeys; j++, i++)
-+ {
-+ p_AdTableNewTmp =
-+ PTR_MOVE(p_AdditionalInfo->p_AdTableNew, j*FM_PCD_CC_AD_ENTRY_SIZE);
-+ p_AdTableOldTmp =
-+ PTR_MOVE(p_AdditionalInfo->p_AdTableOld, i*FM_PCD_CC_AD_ENTRY_SIZE);
-+
-+ MemCpy8(p_AdTableNewTmp, p_AdTableOldTmp, FM_PCD_CC_AD_ENTRY_SIZE);
-+
-+ if (j == keyIndex)
-+ {
-+ ASSERT_COND(keyIndex < CC_MAX_NUM_OF_KEYS);
-+ if (p_CcNode->keyAndNextEngineParams[keyIndex].p_StatsObj)
-+ {
-+ /* As statistics were enabled, we need to update the existing
-+ statistics descriptor with a new nullified counters. */
-+ p_StatsObj = GetStatsObj(p_CcNode);
-+ ASSERT_COND(p_StatsObj);
-+
-+ SetStatsCounters(
-+ p_AdTableNewTmp,
-+ (uint32_t)((XX_VirtToPhys(p_StatsObj->h_StatsCounters)
-+ - p_FmPcd->physicalMuramBase)));
-+
-+ tmpStatsObj.h_StatsAd = p_StatsObj->h_StatsAd;
-+ tmpStatsObj.h_StatsCounters = p_StatsObj->h_StatsCounters;
-+
-+ /* As we need to replace only the counters, we build a new statistics
-+ object that holds the old AD and the new counters - this will be the
-+ currently used statistics object.
-+ The newly allocated AD is not required and may be released back to
-+ the available objects with the previous counters pointer. */
-+ p_StatsObj->h_StatsAd =
-+ p_CcNode->keyAndNextEngineParams[keyIndex].p_StatsObj->h_StatsAd;
-+
-+ p_CcNode->keyAndNextEngineParams[keyIndex].p_StatsObj->h_StatsAd =
-+ tmpStatsObj.h_StatsAd;
-+
-+ /* Store allocated statistics object */
-+ p_AdditionalInfo->keyAndNextEngineParams[keyIndex].p_StatsObj =
-+ p_StatsObj;
-+
-+ /* As statistics were previously enabled, store the old statistics object to be released */
-+ p_AdditionalInfo->p_StatsObjForRmv =
-+ p_CcNode->keyAndNextEngineParams[keyIndex].p_StatsObj;
-+ }
-+
-+ p_KeysMatchTableNewTmp =
-+ PTR_MOVE(p_AdditionalInfo->p_KeysMatchTableNew, j * size * sizeof(uint8_t));
-+
-+ MemCpy8(p_KeysMatchTableNewTmp, p_Key,
-+ p_CcNode->userSizeOfExtraction);
-+
-+ if (p_CcNode->lclMask)
-+ {
-+ if (p_Mask)
-+ MemCpy8(PTR_MOVE(p_KeysMatchTableNewTmp,
-+ p_CcNode->ccKeySizeAccExtraction),
-+ p_Mask, p_CcNode->userSizeOfExtraction);
-+ else
-+ if (p_CcNode->ccKeySizeAccExtraction > 4)
-+ MemSet8(PTR_MOVE(p_KeysMatchTableNewTmp,
-+ p_CcNode->ccKeySizeAccExtraction),
-+ 0xff, p_CcNode->userSizeOfExtraction);
-+ else
-+ MemCpy8(PTR_MOVE(p_KeysMatchTableNewTmp,
-+ p_CcNode->ccKeySizeAccExtraction),
-+ p_CcNode->p_GlblMask,
-+ p_CcNode->userSizeOfExtraction);
-+ }
-+ }
-+ else
-+ {
-+ p_KeysMatchTableNewTmp =
-+ PTR_MOVE(p_AdditionalInfo->p_KeysMatchTableNew, j * size * sizeof(uint8_t));
-+ p_KeysMatchTableOldTmp =
-+ PTR_MOVE(p_CcNode->h_KeysMatchTable, i * size * sizeof(uint8_t));
-+
-+ if (p_CcNode->lclMask)
-+ {
-+ if (prvLclMask)
-+ MemCpy8(
-+ PTR_MOVE(p_KeysMatchTableNewTmp, p_CcNode->ccKeySizeAccExtraction),
-+ PTR_MOVE(p_KeysMatchTableOldTmp, p_CcNode->ccKeySizeAccExtraction),
-+ p_CcNode->userSizeOfExtraction);
-+ else
-+ {
-+ p_KeysMatchTableOldTmp =
-+ PTR_MOVE(p_CcNode->h_KeysMatchTable,
-+ i * (int)p_CcNode->ccKeySizeAccExtraction * sizeof(uint8_t));
-+
-+ if (p_CcNode->ccKeySizeAccExtraction > 4)
-+ MemSet8(PTR_MOVE(p_KeysMatchTableNewTmp,
-+ p_CcNode->ccKeySizeAccExtraction),
-+ 0xff, p_CcNode->userSizeOfExtraction);
-+ else
-+ MemCpy8(
-+ PTR_MOVE(p_KeysMatchTableNewTmp, p_CcNode->ccKeySizeAccExtraction),
-+ p_CcNode->p_GlblMask,
-+ p_CcNode->userSizeOfExtraction);
-+ }
-+ }
-+ MemCpy8((void*)p_KeysMatchTableNewTmp, p_KeysMatchTableOldTmp,
-+ p_CcNode->ccKeySizeAccExtraction);
-+ }
-+ }
-+
-+ p_AdTableNewTmp =
-+ PTR_MOVE(p_AdditionalInfo->p_AdTableNew, j * FM_PCD_CC_AD_ENTRY_SIZE);
-+ p_AdTableOldTmp = PTR_MOVE(p_CcNode->h_AdTable, i * FM_PCD_CC_AD_ENTRY_SIZE);
-+
-+ MemCpy8(p_AdTableNewTmp, p_AdTableOldTmp, FM_PCD_CC_AD_ENTRY_SIZE);
-+
-+ return E_OK;
-+}
-+
-+static t_Error BuildNewNodeModifyNextEngine(
-+ t_Handle h_FmPcd, t_Handle h_FmPcdCcNodeOrTree, uint16_t keyIndex,
-+ t_FmPcdCcNextEngineParams *p_CcNextEngineParams, t_List *h_OldLst,
-+ t_List *h_NewLst, t_FmPcdModifyCcKeyAdditionalParams *p_AdditionalInfo)
-+{
-+ t_Error err = E_OK;
-+ uint32_t requiredAction = 0;
-+ t_List *p_Pos;
-+ t_CcNodeInformation *p_CcNodeInformation, ccNodeInfo;
-+ t_Handle p_Ad;
-+ t_FmPcdCcNode *p_FmPcdCcNode1 = NULL;
-+ t_FmPcdCcTree *p_FmPcdCcTree = NULL;
-+ t_FmPcdStatsObj *p_StatsObj;
-+ t_FmPcdCcStatsParams statsParams = { 0 };
-+
-+ ASSERT_COND(p_CcNextEngineParams);
-+
-+ /* check that new NIA is legal */
-+ if (!p_AdditionalInfo->tree)
-+ err = ValidateNextEngineParams(
-+ h_FmPcd, p_CcNextEngineParams,
-+ ((t_FmPcdCcNode *)h_FmPcdCcNodeOrTree)->statisticsMode);
-+ else
-+ /* Statistics are not supported for CC root */
-+ err = ValidateNextEngineParams(h_FmPcd, p_CcNextEngineParams,
-+ e_FM_PCD_CC_STATS_MODE_NONE);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+
-+ /* Update internal data structure for next engine per index (index - key) */
-+ memcpy(&p_AdditionalInfo->keyAndNextEngineParams[keyIndex].nextEngineParams,
-+ p_CcNextEngineParams, sizeof(t_FmPcdCcNextEngineParams));
-+
-+ /* Check that manip is legal and what requiredAction is necessary for this manip */
-+ if (p_CcNextEngineParams->h_Manip)
-+ {
-+ err = FmPcdManipCheckParamsForCcNextEngine(p_CcNextEngineParams,
-+ &requiredAction);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, (NO_MSG));
-+ }
-+
-+ if (!p_AdditionalInfo->tree)
-+ {
-+ p_FmPcdCcNode1 = (t_FmPcdCcNode *)h_FmPcdCcNodeOrTree;
-+ p_AdditionalInfo->numOfKeys = p_FmPcdCcNode1->numOfKeys;
-+ p_Ad = p_FmPcdCcNode1->h_AdTable;
-+
-+ if (p_FmPcdCcNode1->keyAndNextEngineParams[keyIndex].nextEngineParams.nextEngine
-+ == e_FM_PCD_CC)
-+ p_AdditionalInfo->h_NodeForRmv =
-+ p_FmPcdCcNode1->keyAndNextEngineParams[keyIndex].nextEngineParams.params.ccParams.h_CcNode;
-+
-+ if (p_FmPcdCcNode1->keyAndNextEngineParams[keyIndex].nextEngineParams.h_Manip)
-+ p_AdditionalInfo->h_ManipForRmv =
-+ p_FmPcdCcNode1->keyAndNextEngineParams[keyIndex].nextEngineParams.h_Manip;
-+
-+#if (DPAA_VERSION >= 11)
-+ if ((p_FmPcdCcNode1->keyAndNextEngineParams[keyIndex].nextEngineParams.nextEngine
-+ == e_FM_PCD_FR)
-+ && (p_FmPcdCcNode1->keyAndNextEngineParams[keyIndex].nextEngineParams.params.frParams.h_FrmReplic))
-+ p_AdditionalInfo->h_FrmReplicForRmv =
-+ p_FmPcdCcNode1->keyAndNextEngineParams[keyIndex].nextEngineParams.params.frParams.h_FrmReplic;
-+#endif /* (DPAA_VERSION >= 11) */
-+ }
-+ else
-+ {
-+ p_FmPcdCcTree = (t_FmPcdCcTree *)h_FmPcdCcNodeOrTree;
-+ p_Ad = UINT_TO_PTR(p_FmPcdCcTree->ccTreeBaseAddr);
-+
-+ if (p_FmPcdCcTree->keyAndNextEngineParams[keyIndex].nextEngineParams.nextEngine
-+ == e_FM_PCD_CC)
-+ p_AdditionalInfo->h_NodeForRmv =
-+ p_FmPcdCcTree->keyAndNextEngineParams[keyIndex].nextEngineParams.params.ccParams.h_CcNode;
-+
-+ if (p_FmPcdCcTree->keyAndNextEngineParams[keyIndex].nextEngineParams.h_Manip)
-+ p_AdditionalInfo->h_ManipForRmv =
-+ p_FmPcdCcTree->keyAndNextEngineParams[keyIndex].nextEngineParams.h_Manip;
-+
-+#if (DPAA_VERSION >= 11)
-+ if ((p_FmPcdCcTree->keyAndNextEngineParams[keyIndex].nextEngineParams.nextEngine
-+ == e_FM_PCD_FR)
-+ && (p_FmPcdCcTree->keyAndNextEngineParams[keyIndex].nextEngineParams.params.frParams.h_FrmReplic))
-+ p_AdditionalInfo->h_FrmReplicForRmv =
-+ p_FmPcdCcTree->keyAndNextEngineParams[keyIndex].nextEngineParams.params.frParams.h_FrmReplic;
-+#endif /* (DPAA_VERSION >= 11) */
-+ }
-+
-+ if ((p_CcNextEngineParams->nextEngine == e_FM_PCD_CC)
-+ && p_CcNextEngineParams->h_Manip)
-+ {
-+ err = AllocAndFillAdForContLookupManip(
-+ p_CcNextEngineParams->params.ccParams.h_CcNode);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, (NO_MSG));
-+ }
-+
-+ ASSERT_COND(p_Ad);
-+
-+ memset(&ccNodeInfo, 0, sizeof(t_CcNodeInformation));
-+ ccNodeInfo.h_CcNode = PTR_MOVE(p_Ad, keyIndex * FM_PCD_CC_AD_ENTRY_SIZE);
-+
-+ /* If statistics were enabled, this Ad is the statistics Ad. Need to follow its
-+ nextAction to retrieve the actual Nia-Ad. If statistics should remain enabled,
-+ only the actual Nia-Ad should be modified. */
-+ if ((!p_AdditionalInfo->tree)
-+ && (((t_FmPcdCcNode *)h_FmPcdCcNodeOrTree)->keyAndNextEngineParams[keyIndex].p_StatsObj)
-+ && (p_CcNextEngineParams->statisticsEn))
-+ ccNodeInfo.h_CcNode =
-+ ((t_FmPcdCcNode *)h_FmPcdCcNodeOrTree)->keyAndNextEngineParams[keyIndex].p_StatsObj->h_StatsAd;
-+
-+ EnqueueNodeInfoToRelevantLst(h_OldLst, &ccNodeInfo, NULL);
-+
-+ memset(&ccNodeInfo, 0, sizeof(t_CcNodeInformation));
-+ p_Ad = GetNewAd(h_FmPcdCcNodeOrTree, p_AdditionalInfo->tree);
-+ if (!p_Ad)
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY,
-+ ("MURAM allocation for CC node action descriptor"));
-+ MemSet8((uint8_t *)p_Ad, 0, FM_PCD_CC_AD_ENTRY_SIZE);
-+
-+ /* If statistics were not enabled before, but requested now - Allocate a statistics
-+ object that holds statistics AD and counters. */
-+ if ((!p_AdditionalInfo->tree)
-+ && (!((t_FmPcdCcNode *)h_FmPcdCcNodeOrTree)->keyAndNextEngineParams[keyIndex].p_StatsObj)
-+ && (p_CcNextEngineParams->statisticsEn))
-+ {
-+ p_StatsObj = GetStatsObj((t_FmPcdCcNode *)h_FmPcdCcNodeOrTree);
-+ ASSERT_COND(p_StatsObj);
-+
-+ /* Store allocated statistics object */
-+ p_AdditionalInfo->keyAndNextEngineParams[keyIndex].p_StatsObj =
-+ p_StatsObj;
-+
-+ statsParams.h_StatsAd = p_StatsObj->h_StatsAd;
-+ statsParams.h_StatsCounters = p_StatsObj->h_StatsCounters;
-+
-+#if (DPAA_VERSION >= 11)
-+ statsParams.h_StatsFLRs =
-+ ((t_FmPcdCcNode *)h_FmPcdCcNodeOrTree)->h_StatsFLRs;
-+
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+ NextStepAd(p_Ad, &statsParams, p_CcNextEngineParams, h_FmPcd);
-+ }
-+ else
-+ NextStepAd(p_Ad, NULL, p_CcNextEngineParams, h_FmPcd);
-+
-+ ccNodeInfo.h_CcNode = p_Ad;
-+ EnqueueNodeInfoToRelevantLst(h_NewLst, &ccNodeInfo, NULL);
-+
-+ p_AdditionalInfo->keyAndNextEngineParams[keyIndex].requiredAction =
-+ requiredAction;
-+ p_AdditionalInfo->keyAndNextEngineParams[keyIndex].requiredAction |=
-+ UPDATE_CC_WITH_TREE;
-+
-+ if (!p_AdditionalInfo->tree)
-+ {
-+ ASSERT_COND(p_FmPcdCcNode1);
-+ if (!LIST_IsEmpty(&p_FmPcdCcNode1->ccTreesLst))
-+ {
-+ LIST_FOR_EACH(p_Pos, &p_FmPcdCcNode1->ccTreesLst)
-+ {
-+ p_CcNodeInformation = CC_NODE_F_OBJECT(p_Pos);
-+
-+ ASSERT_COND(p_CcNodeInformation->h_CcNode);
-+ /* Update the manipulation which has to be updated from parameters of the port
-+ it's has to be updated with restrictions defined in the function */
-+
-+ err =
-+ SetRequiredAction(
-+ p_FmPcdCcNode1->h_FmPcd,
-+ p_FmPcdCcNode1->shadowAction
-+ | p_AdditionalInfo->keyAndNextEngineParams[keyIndex].requiredAction,
-+ &p_AdditionalInfo->keyAndNextEngineParams[keyIndex],
-+ p_Ad, 1, p_CcNodeInformation->h_CcNode);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, (NO_MSG));
-+
-+ err = CcUpdateParam(
-+ p_FmPcdCcNode1->h_FmPcd, NULL, NULL,
-+ &p_AdditionalInfo->keyAndNextEngineParams[keyIndex], 1,
-+ p_Ad, TRUE, p_CcNodeInformation->index,
-+ p_CcNodeInformation->h_CcNode, TRUE);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, (NO_MSG));
-+ }
-+ }
-+ }
-+ else
-+ {
-+ ASSERT_COND(p_FmPcdCcTree);
-+
-+ err =
-+ SetRequiredAction(
-+ h_FmPcd,
-+ p_FmPcdCcTree->requiredAction
-+ | p_AdditionalInfo->keyAndNextEngineParams[keyIndex].requiredAction,
-+ &p_AdditionalInfo->keyAndNextEngineParams[keyIndex],
-+ p_Ad, 1, (t_Handle)p_FmPcdCcTree);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, (NO_MSG));
-+
-+ err = CcUpdateParam(h_FmPcd, NULL, NULL,
-+ &p_AdditionalInfo->keyAndNextEngineParams[keyIndex],
-+ 1, p_Ad, TRUE, 0, (t_Handle)p_FmPcdCcTree, TRUE);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, (NO_MSG));
-+ }
-+
-+ if (p_CcNextEngineParams->nextEngine == e_FM_PCD_CC)
-+ p_AdditionalInfo->h_NodeForAdd =
-+ p_CcNextEngineParams->params.ccParams.h_CcNode;
-+ if (p_CcNextEngineParams->h_Manip)
-+ p_AdditionalInfo->h_ManipForAdd = p_CcNextEngineParams->h_Manip;
-+
-+ /* If statistics were previously enabled, but now are disabled,
-+ store the old statistics object to be released */
-+ if ((!p_AdditionalInfo->tree)
-+ && (((t_FmPcdCcNode *)h_FmPcdCcNodeOrTree)->keyAndNextEngineParams[keyIndex].p_StatsObj)
-+ && (!p_CcNextEngineParams->statisticsEn))
-+ {
-+ p_AdditionalInfo->p_StatsObjForRmv =
-+ ((t_FmPcdCcNode *)h_FmPcdCcNodeOrTree)->keyAndNextEngineParams[keyIndex].p_StatsObj;
-+
-+
-+ p_AdditionalInfo->keyAndNextEngineParams[keyIndex].p_StatsObj = NULL;
-+ }
-+#if (DPAA_VERSION >= 11)
-+ if ((p_CcNextEngineParams->nextEngine == e_FM_PCD_FR)
-+ && (p_CcNextEngineParams->params.frParams.h_FrmReplic))
-+ p_AdditionalInfo->h_FrmReplicForAdd =
-+ p_CcNextEngineParams->params.frParams.h_FrmReplic;
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+ return E_OK;
-+}
-+
-+static void UpdateAdPtrOfNodesWhichPointsOnCrntMdfNode(
-+ t_FmPcdCcNode *p_CrntMdfNode, t_List *h_OldLst,
-+ t_FmPcdCcNextEngineParams **p_NextEngineParams)
-+{
-+ t_CcNodeInformation *p_CcNodeInformation;
-+ t_FmPcdCcNode *p_NodePtrOnCurrentMdfNode = NULL;
-+ t_List *p_Pos;
-+ int i = 0;
-+ t_Handle p_AdTablePtOnCrntCurrentMdfNode/*, p_AdTableNewModified*/;
-+ t_CcNodeInformation ccNodeInfo;
-+
-+ LIST_FOR_EACH(p_Pos, &p_CrntMdfNode->ccPrevNodesLst)
-+ {
-+ p_CcNodeInformation = CC_NODE_F_OBJECT(p_Pos);
-+ p_NodePtrOnCurrentMdfNode =
-+ (t_FmPcdCcNode *)p_CcNodeInformation->h_CcNode;
-+
-+ ASSERT_COND(p_NodePtrOnCurrentMdfNode);
-+
-+ /* Search in the previous node which exact index points on this current modified node for getting AD */
-+ for (i = 0; i < p_NodePtrOnCurrentMdfNode->numOfKeys + 1; i++)
-+ {
-+ if (p_NodePtrOnCurrentMdfNode->keyAndNextEngineParams[i].nextEngineParams.nextEngine
-+ == e_FM_PCD_CC)
-+ {
-+ if (p_NodePtrOnCurrentMdfNode->keyAndNextEngineParams[i].nextEngineParams.params.ccParams.h_CcNode
-+ == (t_Handle)p_CrntMdfNode)
-+ {
-+ if (p_NodePtrOnCurrentMdfNode->keyAndNextEngineParams[i].nextEngineParams.h_Manip)
-+ p_AdTablePtOnCrntCurrentMdfNode = p_CrntMdfNode->h_Ad;
-+ else
-+ if (p_NodePtrOnCurrentMdfNode->keyAndNextEngineParams[i].p_StatsObj)
-+ p_AdTablePtOnCrntCurrentMdfNode =
-+ p_NodePtrOnCurrentMdfNode->keyAndNextEngineParams[i].p_StatsObj->h_StatsAd;
-+ else
-+ p_AdTablePtOnCrntCurrentMdfNode =
-+ PTR_MOVE(p_NodePtrOnCurrentMdfNode->h_AdTable, i*FM_PCD_CC_AD_ENTRY_SIZE);
-+
-+ memset(&ccNodeInfo, 0, sizeof(t_CcNodeInformation));
-+ ccNodeInfo.h_CcNode = p_AdTablePtOnCrntCurrentMdfNode;
-+ EnqueueNodeInfoToRelevantLst(h_OldLst, &ccNodeInfo, NULL);
-+
-+ if (!(*p_NextEngineParams))
-+ *p_NextEngineParams =
-+ &p_NodePtrOnCurrentMdfNode->keyAndNextEngineParams[i].nextEngineParams;
-+ }
-+ }
-+ }
-+
-+ ASSERT_COND(i != p_NodePtrOnCurrentMdfNode->numOfKeys);
-+ }
-+}
-+
-+static void UpdateAdPtrOfTreesWhichPointsOnCrntMdfNode(
-+ t_FmPcdCcNode *p_CrntMdfNode, t_List *h_OldLst,
-+ t_FmPcdCcNextEngineParams **p_NextEngineParams)
-+{
-+ t_CcNodeInformation *p_CcNodeInformation;
-+ t_FmPcdCcTree *p_TreePtrOnCurrentMdfNode = NULL;
-+ t_List *p_Pos;
-+ int i = 0;
-+ t_Handle p_AdTableTmp;
-+ t_CcNodeInformation ccNodeInfo;
-+
-+ LIST_FOR_EACH(p_Pos, &p_CrntMdfNode->ccTreeIdLst)
-+ {
-+ p_CcNodeInformation = CC_NODE_F_OBJECT(p_Pos);
-+ p_TreePtrOnCurrentMdfNode =
-+ (t_FmPcdCcTree *)p_CcNodeInformation->h_CcNode;
-+
-+ ASSERT_COND(p_TreePtrOnCurrentMdfNode);
-+
-+ /*search in the trees which exact index points on this current modified node for getting AD */
-+ for (i = 0; i < p_TreePtrOnCurrentMdfNode->numOfEntries; i++)
-+ {
-+ if (p_TreePtrOnCurrentMdfNode->keyAndNextEngineParams[i].nextEngineParams.nextEngine
-+ == e_FM_PCD_CC)
-+ {
-+ if (p_TreePtrOnCurrentMdfNode->keyAndNextEngineParams[i].nextEngineParams.params.ccParams.h_CcNode
-+ == (t_Handle)p_CrntMdfNode)
-+ {
-+ p_AdTableTmp =
-+ UINT_TO_PTR(p_TreePtrOnCurrentMdfNode->ccTreeBaseAddr + i*FM_PCD_CC_AD_ENTRY_SIZE);
-+ memset(&ccNodeInfo, 0, sizeof(t_CcNodeInformation));
-+ ccNodeInfo.h_CcNode = p_AdTableTmp;
-+ EnqueueNodeInfoToRelevantLst(h_OldLst, &ccNodeInfo, NULL);
-+
-+ if (!(*p_NextEngineParams))
-+ *p_NextEngineParams =
-+ &p_TreePtrOnCurrentMdfNode->keyAndNextEngineParams[i].nextEngineParams;
-+ }
-+ }
-+ }
-+
-+ ASSERT_COND(i == p_TreePtrOnCurrentMdfNode->numOfEntries);
-+ }
-+}
-+
-+static t_FmPcdModifyCcKeyAdditionalParams * ModifyNodeCommonPart(
-+ t_Handle h_FmPcdCcNodeOrTree, uint16_t keyIndex,
-+ e_ModifyState modifyState, bool ttlCheck, bool hashCheck, bool tree)
-+{
-+ t_FmPcdModifyCcKeyAdditionalParams *p_FmPcdModifyCcKeyAdditionalParams;
-+ int i = 0, j = 0;
-+ bool wasUpdate = FALSE;
-+ t_FmPcdCcNode *p_CcNode = NULL;
-+ t_FmPcdCcTree *p_FmPcdCcTree;
-+ uint16_t numOfKeys;
-+ t_FmPcdCcKeyAndNextEngineParams *p_KeyAndNextEngineParams;
-+
-+ SANITY_CHECK_RETURN_VALUE(h_FmPcdCcNodeOrTree, E_INVALID_HANDLE, NULL);
-+
-+ if (!tree)
-+ {
-+ p_CcNode = (t_FmPcdCcNode *)h_FmPcdCcNodeOrTree;
-+ numOfKeys = p_CcNode->numOfKeys;
-+
-+ /* node has to be pointed by another node or tree */
-+
-+ p_KeyAndNextEngineParams = (t_FmPcdCcKeyAndNextEngineParams *)XX_Malloc(
-+ sizeof(t_FmPcdCcKeyAndNextEngineParams) * (numOfKeys + 1));
-+ if (!p_KeyAndNextEngineParams)
-+ {
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("Next engine and required action structure"));
-+ return NULL;
-+ }
-+ memcpy(p_KeyAndNextEngineParams, p_CcNode->keyAndNextEngineParams,
-+ (numOfKeys + 1) * sizeof(t_FmPcdCcKeyAndNextEngineParams));
-+
-+ if (ttlCheck)
-+ {
-+ if ((p_CcNode->parseCode == CC_PC_FF_IPV4TTL)
-+ || (p_CcNode->parseCode == CC_PC_FF_IPV6HOP_LIMIT))
-+ {
-+ XX_Free(p_KeyAndNextEngineParams);
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE, ("nodeId of CC_PC_FF_IPV4TTL or CC_PC_FF_IPV6HOP_LIMIT can not be used for this operation"));
-+ return NULL;
-+ }
-+ }
-+
-+ if (hashCheck)
-+ {
-+ if (p_CcNode->parseCode == CC_PC_GENERIC_IC_HASH_INDEXED)
-+ {
-+ XX_Free(p_KeyAndNextEngineParams);
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE, ("nodeId of CC_PC_GENERIC_IC_HASH_INDEXED can not be used for this operation"));
-+ return NULL;
-+ }
-+ }
-+ }
-+ else
-+ {
-+ p_FmPcdCcTree = (t_FmPcdCcTree *)h_FmPcdCcNodeOrTree;
-+ numOfKeys = p_FmPcdCcTree->numOfEntries;
-+
-+ p_KeyAndNextEngineParams = (t_FmPcdCcKeyAndNextEngineParams *)XX_Malloc(
-+ sizeof(t_FmPcdCcKeyAndNextEngineParams)
-+ * FM_PCD_MAX_NUM_OF_CC_GROUPS);
-+ if (!p_KeyAndNextEngineParams)
-+ {
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("Next engine and required action structure"));
-+ return NULL;
-+ }
-+ memcpy(p_KeyAndNextEngineParams,
-+ p_FmPcdCcTree->keyAndNextEngineParams,
-+ FM_PCD_MAX_NUM_OF_CC_GROUPS
-+ * sizeof(t_FmPcdCcKeyAndNextEngineParams));
-+ }
-+
-+ p_FmPcdModifyCcKeyAdditionalParams =
-+ (t_FmPcdModifyCcKeyAdditionalParams *)XX_Malloc(
-+ sizeof(t_FmPcdModifyCcKeyAdditionalParams));
-+ if (!p_FmPcdModifyCcKeyAdditionalParams)
-+ {
-+ XX_Free(p_KeyAndNextEngineParams);
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("Allocation of internal data structure FAILED"));
-+ return NULL;
-+ }
-+ memset(p_FmPcdModifyCcKeyAdditionalParams, 0,
-+ sizeof(t_FmPcdModifyCcKeyAdditionalParams));
-+
-+ p_FmPcdModifyCcKeyAdditionalParams->h_CurrentNode = h_FmPcdCcNodeOrTree;
-+ p_FmPcdModifyCcKeyAdditionalParams->savedKeyIndex = keyIndex;
-+
-+ while (i < numOfKeys)
-+ {
-+ if ((j == keyIndex) && !wasUpdate)
-+ {
-+ if (modifyState == e_MODIFY_STATE_ADD)
-+ j++;
-+ else
-+ if (modifyState == e_MODIFY_STATE_REMOVE)
-+ i++;
-+ wasUpdate = TRUE;
-+ }
-+ else
-+ {
-+ memcpy(&p_FmPcdModifyCcKeyAdditionalParams->keyAndNextEngineParams[j],
-+ p_KeyAndNextEngineParams + i,
-+ sizeof(t_FmPcdCcKeyAndNextEngineParams));
-+ i++;
-+ j++;
-+ }
-+ }
-+
-+ if (keyIndex == numOfKeys)
-+ {
-+ if (modifyState == e_MODIFY_STATE_ADD)
-+ j++;
-+ }
-+
-+ memcpy(&p_FmPcdModifyCcKeyAdditionalParams->keyAndNextEngineParams[j],
-+ p_KeyAndNextEngineParams + numOfKeys,
-+ sizeof(t_FmPcdCcKeyAndNextEngineParams));
-+
-+ XX_Free(p_KeyAndNextEngineParams);
-+
-+ return p_FmPcdModifyCcKeyAdditionalParams;
-+}
-+
-+static t_Error UpdatePtrWhichPointOnCrntMdfNode(
-+ t_FmPcdCcNode *p_CcNode,
-+ t_FmPcdModifyCcKeyAdditionalParams *p_FmPcdModifyCcKeyAdditionalParams,
-+ t_List *h_OldLst, t_List *h_NewLst)
-+{
-+ t_FmPcdCcNextEngineParams *p_NextEngineParams = NULL;
-+ t_CcNodeInformation ccNodeInfo = { 0 };
-+ t_Handle h_NewAd;
-+ t_Handle h_OrigAd = NULL;
-+
-+ /* Building a list of all action descriptors that point to the previous node */
-+ if (!LIST_IsEmpty(&p_CcNode->ccPrevNodesLst))
-+ UpdateAdPtrOfNodesWhichPointsOnCrntMdfNode(p_CcNode, h_OldLst,
-+ &p_NextEngineParams);
-+
-+ if (!LIST_IsEmpty(&p_CcNode->ccTreeIdLst))
-+ UpdateAdPtrOfTreesWhichPointsOnCrntMdfNode(p_CcNode, h_OldLst,
-+ &p_NextEngineParams);
-+
-+ /* This node must be found as next engine of one of its previous nodes or trees*/
-+ if (p_NextEngineParams)
-+ {
-+ /* Building a new action descriptor that points to the modified node */
-+ h_NewAd = GetNewAd(p_CcNode, FALSE);
-+ if (!h_NewAd)
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, NO_MSG);
-+ MemSet8(h_NewAd, 0, FM_PCD_CC_AD_ENTRY_SIZE);
-+
-+ h_OrigAd = p_CcNode->h_Ad;
-+ BuildNewAd(h_NewAd, p_FmPcdModifyCcKeyAdditionalParams, p_CcNode,
-+ p_NextEngineParams);
-+
-+ ccNodeInfo.h_CcNode = h_NewAd;
-+ EnqueueNodeInfoToRelevantLst(h_NewLst, &ccNodeInfo, NULL);
-+
-+ if (p_NextEngineParams->h_Manip && !h_OrigAd)
-+ FmPcdManipUpdateOwner(p_NextEngineParams->h_Manip, FALSE);
-+ }
-+ return E_OK;
-+}
-+
-+static void UpdateCcRootOwner(t_FmPcdCcTree *p_FmPcdCcTree, bool add)
-+{
-+ ASSERT_COND(p_FmPcdCcTree);
-+
-+ /* this routine must be protected by the calling routine! */
-+
-+ if (add)
-+ p_FmPcdCcTree->owners++;
-+ else
-+ {
-+ ASSERT_COND(p_FmPcdCcTree->owners);
-+ p_FmPcdCcTree->owners--;
-+ }
-+}
-+
-+static t_Error CheckAndSetManipParamsWithCcNodeParams(t_FmPcdCcNode *p_CcNode)
-+{
-+ t_Error err = E_OK;
-+ int i = 0;
-+
-+ for (i = 0; i < p_CcNode->numOfKeys; i++)
-+ {
-+ if (p_CcNode->keyAndNextEngineParams[i].nextEngineParams.h_Manip)
-+ {
-+ err =
-+ FmPcdManipCheckParamsWithCcNodeParams(
-+ p_CcNode->keyAndNextEngineParams[i].nextEngineParams.h_Manip,
-+ (t_Handle)p_CcNode);
-+ if (err)
-+ return err;
-+ }
-+ }
-+
-+ return err;
-+}
-+static t_Error ValidateAndCalcStatsParams(t_FmPcdCcNode *p_CcNode,
-+ t_FmPcdCcNodeParams *p_CcNodeParam,
-+ uint32_t *p_NumOfRanges,
-+ uint32_t *p_CountersArraySize)
-+{
-+ e_FmPcdCcStatsMode statisticsMode = p_CcNode->statisticsMode;
-+ uint32_t i;
-+
-+ UNUSED(p_CcNodeParam);
-+
-+ switch (statisticsMode)
-+ {
-+ case e_FM_PCD_CC_STATS_MODE_NONE:
-+ for (i = 0; i < p_CcNode->numOfKeys; i++)
-+ if (p_CcNodeParam->keysParams.keyParams[i].ccNextEngineParams.statisticsEn)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("Statistics cannot be enabled for key %d when statistics mode was set to 'NONE'", i));
-+ return E_OK;
-+
-+ case e_FM_PCD_CC_STATS_MODE_FRAME:
-+ case e_FM_PCD_CC_STATS_MODE_BYTE_AND_FRAME:
-+ *p_NumOfRanges = 1;
-+ *p_CountersArraySize = 2 * FM_PCD_CC_STATS_COUNTER_SIZE;
-+ return E_OK;
-+
-+#if (DPAA_VERSION >= 11)
-+ case e_FM_PCD_CC_STATS_MODE_RMON:
-+ {
-+ uint16_t *p_FrameLengthRanges =
-+ p_CcNodeParam->keysParams.frameLengthRanges;
-+ uint32_t i;
-+
-+ if (p_FrameLengthRanges[0] <= 0)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("Statistics mode"));
-+
-+ if (p_FrameLengthRanges[0] == 0xFFFF)
-+ {
-+ *p_NumOfRanges = 1;
-+ *p_CountersArraySize = 2 * FM_PCD_CC_STATS_COUNTER_SIZE;
-+ return E_OK;
-+ }
-+
-+ for (i = 1; i < FM_PCD_CC_STATS_MAX_NUM_OF_FLR; i++)
-+ {
-+ if (p_FrameLengthRanges[i - 1] >= p_FrameLengthRanges[i])
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("Frame length range must be larger at least by 1 from preceding range"));
-+
-+ /* Stop when last range is reached */
-+ if (p_FrameLengthRanges[i] == 0xFFFF)
-+ break;
-+ }
-+
-+ if ((i >= FM_PCD_CC_STATS_MAX_NUM_OF_FLR)
-+ || (p_FrameLengthRanges[i] != 0xFFFF))
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE,
-+ ("Last Frame length range must be 0xFFFF"));
-+
-+ *p_NumOfRanges = i + 1;
-+
-+ /* Allocate an extra counter for byte count, as counters
-+ array always begins with byte count */
-+ *p_CountersArraySize = (*p_NumOfRanges + 1)
-+ * FM_PCD_CC_STATS_COUNTER_SIZE;
-+
-+ }
-+ return E_OK;
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+ default:
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("Statistics mode"));
-+ }
-+}
-+
-+static t_Error CheckParams(t_Handle h_FmPcd, t_FmPcdCcNodeParams *p_CcNodeParam,
-+ t_FmPcdCcNode *p_CcNode, bool *isKeyTblAlloc)
-+{
-+ int tmp = 0;
-+ t_FmPcdCcKeyParams *p_KeyParams;
-+ t_Error err;
-+ uint32_t requiredAction = 0;
-+
-+ /* Validate statistics parameters */
-+ err = ValidateAndCalcStatsParams(p_CcNode, p_CcNodeParam,
-+ &(p_CcNode->numOfStatsFLRs),
-+ &(p_CcNode->countersArraySize));
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, ("Invalid statistics parameters"));
-+
-+ /* Validate next engine parameters on Miss */
-+ err = ValidateNextEngineParams(
-+ h_FmPcd, &p_CcNodeParam->keysParams.ccNextEngineParamsForMiss,
-+ p_CcNode->statisticsMode);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err,
-+ ("For this node MissNextEngineParams are not valid"));
-+
-+ if (p_CcNodeParam->keysParams.ccNextEngineParamsForMiss.h_Manip)
-+ {
-+ err = FmPcdManipCheckParamsForCcNextEngine(
-+ &p_CcNodeParam->keysParams.ccNextEngineParamsForMiss,
-+ &requiredAction);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, (NO_MSG));
-+ }
-+
-+ memcpy(&p_CcNode->keyAndNextEngineParams[p_CcNode->numOfKeys].nextEngineParams,
-+ &p_CcNodeParam->keysParams.ccNextEngineParamsForMiss,
-+ sizeof(t_FmPcdCcNextEngineParams));
-+
-+ p_CcNode->keyAndNextEngineParams[p_CcNode->numOfKeys].requiredAction =
-+ requiredAction;
-+
-+ if ((p_CcNode->keyAndNextEngineParams[p_CcNode->numOfKeys].nextEngineParams.nextEngine
-+ == e_FM_PCD_CC)
-+ && p_CcNode->keyAndNextEngineParams[p_CcNode->numOfKeys].nextEngineParams.h_Manip)
-+ {
-+ err =
-+ AllocAndFillAdForContLookupManip(
-+ p_CcNode->keyAndNextEngineParams[p_CcNode->numOfKeys].nextEngineParams.params.ccParams.h_CcNode);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, (NO_MSG));
-+ }
-+
-+ for (tmp = 0; tmp < p_CcNode->numOfKeys; tmp++)
-+ {
-+ p_KeyParams = &p_CcNodeParam->keysParams.keyParams[tmp];
-+
-+ if (!p_KeyParams->p_Key)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("p_Key is not initialized"));
-+
-+ err = ValidateNextEngineParams(h_FmPcd,
-+ &p_KeyParams->ccNextEngineParams,
-+ p_CcNode->statisticsMode);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, (NO_MSG));
-+
-+ err = UpdateGblMask(p_CcNode, p_CcNodeParam->keysParams.keySize,
-+ p_KeyParams->p_Mask);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, (NO_MSG));
-+
-+ if (p_KeyParams->ccNextEngineParams.h_Manip)
-+ {
-+ err = FmPcdManipCheckParamsForCcNextEngine(
-+ &p_KeyParams->ccNextEngineParams, &requiredAction);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, (NO_MSG));
-+ }
-+
-+ /* Store 'key' parameters - key, mask (if passed by the user) */
-+ memcpy(p_CcNode->keyAndNextEngineParams[tmp].key, p_KeyParams->p_Key,
-+ p_CcNodeParam->keysParams.keySize);
-+
-+ if (p_KeyParams->p_Mask)
-+ memcpy(p_CcNode->keyAndNextEngineParams[tmp].mask,
-+ p_KeyParams->p_Mask, p_CcNodeParam->keysParams.keySize);
-+ else
-+ memset((void *)(p_CcNode->keyAndNextEngineParams[tmp].mask), 0xFF,
-+ p_CcNodeParam->keysParams.keySize);
-+
-+ /* Store next engine parameters */
-+ memcpy(&p_CcNode->keyAndNextEngineParams[tmp].nextEngineParams,
-+ &p_KeyParams->ccNextEngineParams,
-+ sizeof(t_FmPcdCcNextEngineParams));
-+
-+ p_CcNode->keyAndNextEngineParams[tmp].requiredAction = requiredAction;
-+
-+ if ((p_CcNode->keyAndNextEngineParams[tmp].nextEngineParams.nextEngine
-+ == e_FM_PCD_CC)
-+ && p_CcNode->keyAndNextEngineParams[tmp].nextEngineParams.h_Manip)
-+ {
-+ err =
-+ AllocAndFillAdForContLookupManip(
-+ p_CcNode->keyAndNextEngineParams[tmp].nextEngineParams.params.ccParams.h_CcNode);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, (NO_MSG));
-+ }
-+ }
-+
-+ if (p_CcNode->maxNumOfKeys)
-+ {
-+ if (p_CcNode->maxNumOfKeys < p_CcNode->numOfKeys)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("Number of keys exceed the provided maximal number of keys"));
-+ }
-+
-+ *isKeyTblAlloc = TRUE;
-+
-+ return E_OK;
-+}
-+
-+static t_Error Ipv4TtlOrIpv6HopLimitCheckParams(
-+ t_Handle h_FmPcd, t_FmPcdCcNodeParams *p_CcNodeParam,
-+ t_FmPcdCcNode *p_CcNode, bool *isKeyTblAlloc)
-+{
-+ int tmp = 0;
-+ t_FmPcdCcKeyParams *p_KeyParams;
-+ t_Error err;
-+ uint8_t key = 0x01;
-+ uint32_t requiredAction = 0;
-+
-+ if (p_CcNode->numOfKeys != 1)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("For node of the type IPV4_TTL or IPV6_HOP_LIMIT the maximal supported 'numOfKeys' is 1"));
-+
-+ if ((p_CcNodeParam->keysParams.maxNumOfKeys)
-+ && (p_CcNodeParam->keysParams.maxNumOfKeys != 1))
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("For node of the type IPV4_TTL or IPV6_HOP_LIMIT the maximal supported 'maxNumOfKeys' is 1"));
-+
-+ /* Validate statistics parameters */
-+ err = ValidateAndCalcStatsParams(p_CcNode, p_CcNodeParam,
-+ &(p_CcNode->numOfStatsFLRs),
-+ &(p_CcNode->countersArraySize));
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, ("Invalid statistics parameters"));
-+
-+ err = ValidateNextEngineParams(
-+ h_FmPcd, &p_CcNodeParam->keysParams.ccNextEngineParamsForMiss,
-+ p_CcNodeParam->keysParams.statisticsMode);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err,
-+ ("For this node MissNextEngineParams are not valid"));
-+
-+ if (p_CcNodeParam->keysParams.ccNextEngineParamsForMiss.h_Manip)
-+ {
-+ err = FmPcdManipCheckParamsForCcNextEngine(
-+ &p_CcNodeParam->keysParams.ccNextEngineParamsForMiss,
-+ &requiredAction);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, (NO_MSG));
-+ }
-+
-+ memcpy(&p_CcNode->keyAndNextEngineParams[p_CcNode->numOfKeys].nextEngineParams,
-+ &p_CcNodeParam->keysParams.ccNextEngineParamsForMiss,
-+ sizeof(t_FmPcdCcNextEngineParams));
-+
-+ p_CcNode->keyAndNextEngineParams[p_CcNode->numOfKeys].requiredAction =
-+ requiredAction;
-+
-+ if ((p_CcNode->keyAndNextEngineParams[p_CcNode->numOfKeys].nextEngineParams.nextEngine
-+ == e_FM_PCD_CC)
-+ && p_CcNode->keyAndNextEngineParams[p_CcNode->numOfKeys].nextEngineParams.h_Manip)
-+ {
-+ err =
-+ AllocAndFillAdForContLookupManip(
-+ p_CcNode->keyAndNextEngineParams[p_CcNode->numOfKeys].nextEngineParams.params.ccParams.h_CcNode);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, (NO_MSG));
-+ }
-+
-+ for (tmp = 0; tmp < p_CcNode->numOfKeys; tmp++)
-+ {
-+ p_KeyParams = &p_CcNodeParam->keysParams.keyParams[tmp];
-+
-+ if (p_KeyParams->p_Mask)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("For node of the type IPV4_TTL or IPV6_HOP_LIMIT p_Mask can not be initialized"));
-+
-+ if (memcmp(p_KeyParams->p_Key, &key, 1) != 0)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("For node of the type IPV4_TTL or IPV6_HOP_LIMIT p_Key has to be 1"));
-+
-+ err = ValidateNextEngineParams(h_FmPcd,
-+ &p_KeyParams->ccNextEngineParams,
-+ p_CcNode->statisticsMode);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, (NO_MSG));
-+
-+ if (p_KeyParams->ccNextEngineParams.h_Manip)
-+ {
-+ err = FmPcdManipCheckParamsForCcNextEngine(
-+ &p_KeyParams->ccNextEngineParams, &requiredAction);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, (NO_MSG));
-+ }
-+
-+ /* Store 'key' parameters - key (fixed to 0x01), key size of 1 byte and full mask */
-+ p_CcNode->keyAndNextEngineParams[tmp].key[0] = key;
-+ p_CcNode->keyAndNextEngineParams[tmp].mask[0] = 0xFF;
-+
-+ /* Store NextEngine parameters */
-+ memcpy(&p_CcNode->keyAndNextEngineParams[tmp].nextEngineParams,
-+ &p_KeyParams->ccNextEngineParams,
-+ sizeof(t_FmPcdCcNextEngineParams));
-+
-+ if ((p_CcNode->keyAndNextEngineParams[tmp].nextEngineParams.nextEngine
-+ == e_FM_PCD_CC)
-+ && p_CcNode->keyAndNextEngineParams[tmp].nextEngineParams.h_Manip)
-+ {
-+ err =
-+ AllocAndFillAdForContLookupManip(
-+ p_CcNode->keyAndNextEngineParams[tmp].nextEngineParams.params.ccParams.h_CcNode);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, (NO_MSG));
-+ }
-+ p_CcNode->keyAndNextEngineParams[tmp].requiredAction = requiredAction;
-+ }
-+
-+ *isKeyTblAlloc = FALSE;
-+
-+ return E_OK;
-+}
-+
-+static t_Error IcHashIndexedCheckParams(t_Handle h_FmPcd,
-+ t_FmPcdCcNodeParams *p_CcNodeParam,
-+ t_FmPcdCcNode *p_CcNode,
-+ bool *isKeyTblAlloc)
-+{
-+ int tmp = 0, countOnes = 0;
-+ t_FmPcdCcKeyParams *p_KeyParams;
-+ t_Error err;
-+ uint16_t glblMask = p_CcNodeParam->extractCcParams.extractNonHdr.icIndxMask;
-+ uint16_t countMask = (uint16_t)(glblMask >> 4);
-+ uint32_t requiredAction = 0;
-+
-+ if (glblMask & 0x000f)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE,
-+ ("icIndxMask has to be with last nibble 0"));
-+
-+ while (countMask)
-+ {
-+ countOnes++;
-+ countMask = (uint16_t)(countMask >> 1);
-+ }
-+
-+ if (!POWER_OF_2(p_CcNode->numOfKeys))
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("For Node of the type INDEXED numOfKeys has to be powerOfTwo"));
-+
-+ if (p_CcNode->numOfKeys != ((uint32_t)1 << countOnes))
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("For Node of the type IC_HASH_INDEXED numOfKeys has to be powerOfTwo"));
-+
-+ if (p_CcNodeParam->keysParams.maxNumOfKeys
-+ && (p_CcNodeParam->keysParams.maxNumOfKeys != p_CcNode->numOfKeys))
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("For Node of the type INDEXED 'maxNumOfKeys' should be 0 or equal 'numOfKeys'"));
-+
-+ /* Validate statistics parameters */
-+ err = ValidateAndCalcStatsParams(p_CcNode, p_CcNodeParam,
-+ &(p_CcNode->numOfStatsFLRs),
-+ &(p_CcNode->countersArraySize));
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, ("Invalid statistics parameters"));
-+
-+ err = ValidateNextEngineParams(
-+ h_FmPcd, &p_CcNodeParam->keysParams.ccNextEngineParamsForMiss,
-+ p_CcNode->statisticsMode);
-+ if (GET_ERROR_TYPE(err) != E_NOT_SUPPORTED)
-+ RETURN_ERROR(
-+ MAJOR,
-+ err,
-+ ("MissNextEngineParams for the node of the type IC_INDEX_HASH has to be UnInitialized"));
-+
-+ for (tmp = 0; tmp < p_CcNode->numOfKeys; tmp++)
-+ {
-+ p_KeyParams = &p_CcNodeParam->keysParams.keyParams[tmp];
-+
-+ if (p_KeyParams->p_Mask || p_KeyParams->p_Key)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("For Node of the type IC_HASH_INDEXED p_Key or p_Mask has to be NULL"));
-+
-+ if ((glblMask & (tmp * 16)) == (tmp * 16))
-+ {
-+ err = ValidateNextEngineParams(h_FmPcd,
-+ &p_KeyParams->ccNextEngineParams,
-+ p_CcNode->statisticsMode);
-+ if (err)
-+ RETURN_ERROR(
-+ MAJOR,
-+ err,
-+ ("This index has to be initialized for the node of the type IC_INDEX_HASH according to settings of GlobalMask "));
-+
-+ if (p_KeyParams->ccNextEngineParams.h_Manip)
-+ {
-+ err = FmPcdManipCheckParamsForCcNextEngine(
-+ &p_KeyParams->ccNextEngineParams, &requiredAction);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, (NO_MSG));
-+ p_CcNode->keyAndNextEngineParams[tmp].requiredAction =
-+ requiredAction;
-+ }
-+
-+ memcpy(&p_CcNode->keyAndNextEngineParams[tmp].nextEngineParams,
-+ &p_KeyParams->ccNextEngineParams,
-+ sizeof(t_FmPcdCcNextEngineParams));
-+
-+ if ((p_CcNode->keyAndNextEngineParams[tmp].nextEngineParams.nextEngine
-+ == e_FM_PCD_CC)
-+ && p_CcNode->keyAndNextEngineParams[tmp].nextEngineParams.h_Manip)
-+ {
-+ err =
-+ AllocAndFillAdForContLookupManip(
-+ p_CcNode->keyAndNextEngineParams[tmp].nextEngineParams.params.ccParams.h_CcNode);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, (NO_MSG));
-+ }
-+ }
-+ else
-+ {
-+ err = ValidateNextEngineParams(h_FmPcd,
-+ &p_KeyParams->ccNextEngineParams,
-+ p_CcNode->statisticsMode);
-+ if (GET_ERROR_TYPE(err) != E_NOT_SUPPORTED)
-+ RETURN_ERROR(
-+ MAJOR,
-+ err,
-+ ("This index has to be UnInitialized for the node of the type IC_INDEX_HASH according to settings of GlobalMask"));
-+ }
-+ }
-+
-+ *isKeyTblAlloc = FALSE;
-+ cpu_to_be16s(&glblMask);
-+ memcpy(PTR_MOVE(p_CcNode->p_GlblMask, 2), &glblMask, 2);
-+
-+ return E_OK;
-+}
-+
-+static t_Error ModifyNextEngineParamNode(
-+ t_Handle h_FmPcd, t_Handle h_FmPcdCcNode, uint16_t keyIndex,
-+ t_FmPcdCcNextEngineParams *p_FmPcdCcNextEngineParams)
-+{
-+ t_FmPcdCcNode *p_CcNode = (t_FmPcdCcNode *)h_FmPcdCcNode;
-+ t_FmPcd *p_FmPcd;
-+ t_List h_OldPointersLst, h_NewPointersLst;
-+ t_FmPcdModifyCcKeyAdditionalParams *p_ModifyKeyParams;
-+ t_Error err = E_OK;
-+
-+ SANITY_CHECK_RETURN_ERROR(h_FmPcd, E_INVALID_VALUE);
-+ SANITY_CHECK_RETURN_ERROR(p_CcNode, E_INVALID_HANDLE);
-+
-+ if (keyIndex >= p_CcNode->numOfKeys)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE,
-+ ("keyIndex > previously cleared last index + 1"));
-+
-+ p_FmPcd = (t_FmPcd *)p_CcNode->h_FmPcd;
-+
-+ INIT_LIST(&h_OldPointersLst);
-+ INIT_LIST(&h_NewPointersLst);
-+
-+ p_ModifyKeyParams = ModifyNodeCommonPart(p_CcNode, keyIndex,
-+ e_MODIFY_STATE_CHANGE, FALSE,
-+ FALSE, FALSE);
-+ if (!p_ModifyKeyParams)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, NO_MSG);
-+
-+ if (p_CcNode->maxNumOfKeys
-+ && !TRY_LOCK(p_FmPcd->h_ShadowSpinlock, &p_FmPcd->shadowLock))
-+ {
-+ XX_Free(p_ModifyKeyParams);
-+ return ERROR_CODE(E_BUSY);
-+ }
-+
-+ err = BuildNewNodeModifyNextEngine(h_FmPcd, p_CcNode, keyIndex,
-+ p_FmPcdCcNextEngineParams,
-+ &h_OldPointersLst, &h_NewPointersLst,
-+ p_ModifyKeyParams);
-+ if (err)
-+ {
-+ XX_Free(p_ModifyKeyParams);
-+ if (p_CcNode->maxNumOfKeys)
-+ RELEASE_LOCK(p_FmPcd->shadowLock);
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+
-+ err = DoDynamicChange(p_FmPcd, &h_OldPointersLst, &h_NewPointersLst,
-+ p_ModifyKeyParams, FALSE);
-+
-+ if (p_CcNode->maxNumOfKeys)
-+ RELEASE_LOCK(p_FmPcd->shadowLock);
-+
-+ ReleaseLst(&h_OldPointersLst);
-+ ReleaseLst(&h_NewPointersLst);
-+
-+ return err;
-+}
-+
-+static t_Error FindKeyIndex(t_Handle h_CcNode, uint8_t keySize, uint8_t *p_Key,
-+ uint8_t *p_Mask, uint16_t *p_KeyIndex)
-+{
-+ t_FmPcdCcNode *p_CcNode = (t_FmPcdCcNode *)h_CcNode;
-+ uint8_t tmpMask[FM_PCD_MAX_SIZE_OF_KEY];
-+ uint16_t i;
-+
-+ ASSERT_COND(p_Key);
-+ ASSERT_COND(p_KeyIndex);
-+ ASSERT_COND(keySize < FM_PCD_MAX_SIZE_OF_KEY);
-+
-+ if (keySize != p_CcNode->userSizeOfExtraction)
-+ RETURN_ERROR(
-+ MINOR, E_INVALID_VALUE,
-+ ("Key size doesn't match the extraction size of the node"));
-+
-+ /* If user didn't pass a mask for this key, we'll look for full extraction mask */
-+ if (!p_Mask)
-+ memset(tmpMask, 0xFF, keySize);
-+
-+ for (i = 0; i < p_CcNode->numOfKeys; i++)
-+ {
-+ /* Comparing received key */
-+ if (memcmp(p_Key, p_CcNode->keyAndNextEngineParams[i].key, keySize)
-+ == 0)
-+ {
-+ if (p_Mask)
-+ {
-+ /* If a user passed a mask for this key, it must match to the existing key's mask for a correct match */
-+ if (memcmp(p_Mask, p_CcNode->keyAndNextEngineParams[i].mask,
-+ keySize) == 0)
-+ {
-+ *p_KeyIndex = i;
-+ return E_OK;
-+ }
-+ }
-+ else
-+ {
-+ /* If user didn't pass a mask for this key, check if the existing key mask is full extraction */
-+ if (memcmp(tmpMask, p_CcNode->keyAndNextEngineParams[i].mask,
-+ keySize) == 0)
-+ {
-+ *p_KeyIndex = i;
-+ return E_OK;
-+ }
-+ }
-+ }
-+ }
-+
-+ return ERROR_CODE(E_NOT_FOUND);
-+}
-+
-+static t_Error CalcAndUpdateCcShadow(t_FmPcdCcNode *p_CcNode,
-+ bool isKeyTblAlloc,
-+ uint32_t *p_MatchTableSize,
-+ uint32_t *p_AdTableSize)
-+{
-+ uint32_t shadowSize;
-+ t_Error err;
-+
-+ /* Calculate keys table maximal size - each entry consists of a key and a mask,
-+ (if local mask support is requested) */
-+ *p_MatchTableSize = p_CcNode->ccKeySizeAccExtraction * sizeof(uint8_t)
-+ * p_CcNode->maxNumOfKeys;
-+
-+ if (p_CcNode->maskSupport)
-+ *p_MatchTableSize *= 2;
-+
-+ /* Calculate next action descriptors table, including one more entry for miss */
-+ *p_AdTableSize = (uint32_t)((p_CcNode->maxNumOfKeys + 1)
-+ * FM_PCD_CC_AD_ENTRY_SIZE);
-+
-+ /* Calculate maximal shadow size of this node.
-+ All shadow structures will be used for runtime modifications host command. If
-+ keys table was allocated for this node, the keys table and next engines table may
-+ be modified in run time (entries added or removed), so shadow tables are requires.
-+ Otherwise, the only supported runtime modification is a specific next engine update
-+ and this requires shadow memory of a single AD */
-+
-+ /* Shadow size should be enough to hold the following 3 structures:
-+ * 1 - an action descriptor */
-+ shadowSize = FM_PCD_CC_AD_ENTRY_SIZE;
-+
-+ /* 2 - keys match table, if was allocated for the current node */
-+ if (isKeyTblAlloc)
-+ shadowSize += *p_MatchTableSize;
-+
-+ /* 3 - next action descriptors table */
-+ shadowSize += *p_AdTableSize;
-+
-+ /* Update shadow to the calculated size */
-+ err = FmPcdUpdateCcShadow(p_CcNode->h_FmPcd, (uint32_t)shadowSize,
-+ FM_PCD_CC_AD_TABLE_ALIGN);
-+ if (err != E_OK)
-+ {
-+ DeleteNode(p_CcNode);
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, ("MURAM allocation for CC node shadow"));
-+ }
-+
-+ return E_OK;
-+}
-+
-+static t_Error AllocStatsObjs(t_FmPcdCcNode *p_CcNode)
-+{
-+ t_FmPcdStatsObj *p_StatsObj;
-+ t_Handle h_FmMuram, h_StatsAd, h_StatsCounters;
-+ uint32_t i;
-+
-+ h_FmMuram = FmPcdGetMuramHandle(p_CcNode->h_FmPcd);
-+ if (!h_FmMuram)
-+ RETURN_ERROR(MAJOR, E_INVALID_HANDLE, ("FM MURAM"));
-+
-+ /* Allocate statistics ADs and statistics counter. An extra pair (AD + counters)
-+ will be allocated to support runtime modifications */
-+ for (i = 0; i < p_CcNode->maxNumOfKeys + 2; i++)
-+ {
-+ /* Allocate list object structure */
-+ p_StatsObj = XX_Malloc(sizeof(t_FmPcdStatsObj));
-+ if (!p_StatsObj)
-+ {
-+ FreeStatObjects(&p_CcNode->availableStatsLst, h_FmMuram);
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, ("Statistics object"));
-+ }
-+ memset(p_StatsObj, 0, sizeof(t_FmPcdStatsObj));
-+
-+ /* Allocate statistics AD from MURAM */
-+ h_StatsAd = (t_Handle)FM_MURAM_AllocMem(h_FmMuram,
-+ FM_PCD_CC_AD_ENTRY_SIZE,
-+ FM_PCD_CC_AD_TABLE_ALIGN);
-+ if (!h_StatsAd)
-+ {
-+ FreeStatObjects(&p_CcNode->availableStatsLst, h_FmMuram);
-+ XX_Free(p_StatsObj);
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY,
-+ ("MURAM allocation for statistics ADs"));
-+ }
-+ MemSet8(h_StatsAd, 0, FM_PCD_CC_AD_ENTRY_SIZE);
-+
-+ /* Allocate statistics counters from MURAM */
-+ h_StatsCounters = (t_Handle)FM_MURAM_AllocMem(
-+ h_FmMuram, p_CcNode->countersArraySize,
-+ FM_PCD_CC_AD_TABLE_ALIGN);
-+ if (!h_StatsCounters)
-+ {
-+ FreeStatObjects(&p_CcNode->availableStatsLst, h_FmMuram);
-+ FM_MURAM_FreeMem(h_FmMuram, h_StatsAd);
-+ XX_Free(p_StatsObj);
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY,
-+ ("MURAM allocation for statistics counters"));
-+ }
-+ MemSet8(h_StatsCounters, 0, p_CcNode->countersArraySize);
-+
-+ p_StatsObj->h_StatsAd = h_StatsAd;
-+ p_StatsObj->h_StatsCounters = h_StatsCounters;
-+
-+ EnqueueStatsObj(&p_CcNode->availableStatsLst, p_StatsObj);
-+ }
-+
-+ return E_OK;
-+}
-+
-+static t_Error MatchTableGetKeyStatistics(
-+ t_FmPcdCcNode *p_CcNode, uint16_t keyIndex,
-+ t_FmPcdCcKeyStatistics *p_KeyStatistics)
-+{
-+ uint32_t *p_StatsCounters, i;
-+
-+ if (p_CcNode->statisticsMode == e_FM_PCD_CC_STATS_MODE_NONE)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE,
-+ ("Statistics were not enabled for this match table"));
-+
-+ if (!p_CcNode->keyAndNextEngineParams[keyIndex].p_StatsObj)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE,
-+ ("Statistics were not enabled for this key"));
-+
-+ memset(p_KeyStatistics, 0, sizeof(t_FmPcdCcKeyStatistics));
-+
-+ p_StatsCounters =
-+ p_CcNode->keyAndNextEngineParams[keyIndex].p_StatsObj->h_StatsCounters;
-+ ASSERT_COND(p_StatsCounters);
-+
-+ p_KeyStatistics->byteCount = GET_UINT32(*p_StatsCounters);
-+
-+ for (i = 1; i <= p_CcNode->numOfStatsFLRs; i++)
-+ {
-+ p_StatsCounters =
-+ PTR_MOVE(p_StatsCounters, FM_PCD_CC_STATS_COUNTER_SIZE);
-+
-+ p_KeyStatistics->frameCount += GET_UINT32(*p_StatsCounters);
-+
-+#if (DPAA_VERSION >= 11)
-+ p_KeyStatistics->frameLengthRangeCount[i - 1] =
-+ GET_UINT32(*p_StatsCounters);
-+#endif /* (DPAA_VERSION >= 11) */
-+ }
-+
-+ return E_OK;
-+}
-+
-+static t_Error MatchTableSet(t_Handle h_FmPcd, t_FmPcdCcNode *p_CcNode,
-+ t_FmPcdCcNodeParams *p_CcNodeParam)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd *)h_FmPcd;
-+ t_FmPcdCcNode *p_FmPcdCcNextNode;
-+ t_Error err = E_OK;
-+ uint32_t tmp, keySize;
-+ bool glblMask = FALSE;
-+ t_FmPcdCcKeyParams *p_KeyParams;
-+ t_Handle h_FmMuram, p_KeysMatchTblTmp, p_AdTableTmp;
-+#if (DPAA_VERSION >= 11)
-+ t_Handle h_StatsFLRs;
-+#endif /* (DPAA_VERSION >= 11) */
-+ bool fullField = FALSE;
-+ ccPrivateInfo_t icCode = CC_PRIVATE_INFO_NONE;
-+ bool isKeyTblAlloc, fromIc = FALSE;
-+ uint32_t matchTableSize, adTableSize;
-+ t_CcNodeInformation ccNodeInfo, *p_CcInformation;
-+ t_FmPcdStatsObj *p_StatsObj;
-+ t_FmPcdCcStatsParams statsParams = { 0 };
-+ t_Handle h_Manip;
-+
-+ ASSERT_COND(h_FmPcd);
-+ ASSERT_COND(p_CcNode);
-+ ASSERT_COND(p_CcNodeParam);
-+
-+ p_CcNode->p_GlblMask = (t_Handle)XX_Malloc(
-+ CC_GLBL_MASK_SIZE * sizeof(uint8_t));
-+ memset(p_CcNode->p_GlblMask, 0, CC_GLBL_MASK_SIZE * sizeof(uint8_t));
-+
-+ p_CcNode->h_FmPcd = h_FmPcd;
-+ p_CcNode->numOfKeys = p_CcNodeParam->keysParams.numOfKeys;
-+ p_CcNode->maxNumOfKeys = p_CcNodeParam->keysParams.maxNumOfKeys;
-+ p_CcNode->maskSupport = p_CcNodeParam->keysParams.maskSupport;
-+ p_CcNode->statisticsMode = p_CcNodeParam->keysParams.statisticsMode;
-+
-+ /* For backward compatibility - even if statistics mode is nullified,
-+ we'll fix it to frame mode so we can support per-key request for
-+ statistics using 'statisticsEn' in next engine parameters */
-+ if (!p_CcNode->maxNumOfKeys
-+ && (p_CcNode->statisticsMode == e_FM_PCD_CC_STATS_MODE_NONE))
-+ p_CcNode->statisticsMode = e_FM_PCD_CC_STATS_MODE_FRAME;
-+
-+ h_FmMuram = FmPcdGetMuramHandle(h_FmPcd);
-+ if (!h_FmMuram)
-+ RETURN_ERROR(MAJOR, E_INVALID_HANDLE, ("FM MURAM"));
-+
-+ INIT_LIST(&p_CcNode->ccPrevNodesLst);
-+ INIT_LIST(&p_CcNode->ccTreeIdLst);
-+ INIT_LIST(&p_CcNode->ccTreesLst);
-+ INIT_LIST(&p_CcNode->availableStatsLst);
-+
-+ p_CcNode->h_Spinlock = XX_InitSpinlock();
-+ if (!p_CcNode->h_Spinlock)
-+ {
-+ DeleteNode(p_CcNode);
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, ("CC node spinlock"));
-+ }
-+
-+ if ((p_CcNodeParam->extractCcParams.type == e_FM_PCD_EXTRACT_BY_HDR)
-+ && ((p_CcNodeParam->extractCcParams.extractByHdr.hdr
-+ == HEADER_TYPE_IPv4)
-+ || (p_CcNodeParam->extractCcParams.extractByHdr.hdr
-+ == HEADER_TYPE_IPv6))
-+ && (p_CcNodeParam->extractCcParams.extractByHdr.type
-+ == e_FM_PCD_EXTRACT_FULL_FIELD)
-+ && ((p_CcNodeParam->extractCcParams.extractByHdr.extractByHdrType.fullField.ipv6
-+ == NET_HEADER_FIELD_IPv6_HOP_LIMIT)
-+ || (p_CcNodeParam->extractCcParams.extractByHdr.extractByHdrType.fullField.ipv4
-+ == NET_HEADER_FIELD_IPv4_TTL)))
-+ {
-+ err = Ipv4TtlOrIpv6HopLimitCheckParams(h_FmPcd, p_CcNodeParam, p_CcNode,
-+ &isKeyTblAlloc);
-+ glblMask = FALSE;
-+ }
-+ else
-+ if ((p_CcNodeParam->extractCcParams.type == e_FM_PCD_EXTRACT_NON_HDR)
-+ && ((p_CcNodeParam->extractCcParams.extractNonHdr.src
-+ == e_FM_PCD_EXTRACT_FROM_KEY)
-+ || (p_CcNodeParam->extractCcParams.extractNonHdr.src
-+ == e_FM_PCD_EXTRACT_FROM_HASH)
-+ || (p_CcNodeParam->extractCcParams.extractNonHdr.src
-+ == e_FM_PCD_EXTRACT_FROM_FLOW_ID)))
-+ {
-+ if ((p_CcNodeParam->extractCcParams.extractNonHdr.src
-+ == e_FM_PCD_EXTRACT_FROM_FLOW_ID)
-+ && (p_CcNodeParam->extractCcParams.extractNonHdr.offset != 0))
-+ {
-+ DeleteNode(p_CcNode);
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("In the case of the extraction from e_FM_PCD_EXTRACT_FROM_FLOW_ID offset has to be 0"));
-+ }
-+
-+ icCode = IcDefineCode(p_CcNodeParam);
-+ fromIc = TRUE;
-+ if (icCode == CC_PRIVATE_INFO_NONE)
-+ {
-+ DeleteNode(p_CcNode);
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_STATE,
-+ ("user asked extraction from IC and field in internal context or action wasn't initialized in the right way"));
-+ }
-+
-+ if ((icCode == CC_PRIVATE_INFO_IC_DEQ_FQID_INDEX_LOOKUP)
-+ || (icCode == CC_PRIVATE_INFO_IC_HASH_INDEX_LOOKUP))
-+ {
-+ err = IcHashIndexedCheckParams(h_FmPcd, p_CcNodeParam, p_CcNode,
-+ &isKeyTblAlloc);
-+ glblMask = TRUE;
-+ }
-+ else
-+ {
-+ err = CheckParams(h_FmPcd, p_CcNodeParam, p_CcNode,
-+ &isKeyTblAlloc);
-+ if (p_CcNode->glblMaskSize)
-+ glblMask = TRUE;
-+ }
-+ }
-+ else
-+ {
-+ err = CheckParams(h_FmPcd, p_CcNodeParam, p_CcNode, &isKeyTblAlloc);
-+ if (p_CcNode->glblMaskSize)
-+ glblMask = TRUE;
-+ }
-+
-+ if (err)
-+ {
-+ DeleteNode(p_CcNode);
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+
-+ switch (p_CcNodeParam->extractCcParams.type)
-+ {
-+ case (e_FM_PCD_EXTRACT_BY_HDR):
-+ switch (p_CcNodeParam->extractCcParams.extractByHdr.type)
-+ {
-+ case (e_FM_PCD_EXTRACT_FULL_FIELD):
-+ p_CcNode->parseCode =
-+ GetFullFieldParseCode(
-+ p_CcNodeParam->extractCcParams.extractByHdr.hdr,
-+ p_CcNodeParam->extractCcParams.extractByHdr.hdrIndex,
-+ p_CcNodeParam->extractCcParams.extractByHdr.extractByHdrType.fullField);
-+ GetSizeHeaderField(
-+ p_CcNodeParam->extractCcParams.extractByHdr.hdr,
-+ p_CcNodeParam->extractCcParams.extractByHdr.extractByHdrType.fullField,
-+ &p_CcNode->sizeOfExtraction);
-+ fullField = TRUE;
-+ if ((p_CcNode->parseCode != CC_PC_FF_TCI1)
-+ && (p_CcNode->parseCode != CC_PC_FF_TCI2)
-+ && (p_CcNode->parseCode != CC_PC_FF_MPLS1)
-+ && (p_CcNode->parseCode != CC_PC_FF_MPLS_LAST)
-+ && (p_CcNode->parseCode != CC_PC_FF_IPV4IPTOS_TC1)
-+ && (p_CcNode->parseCode != CC_PC_FF_IPV4IPTOS_TC2)
-+ && (p_CcNode->parseCode
-+ != CC_PC_FF_IPTOS_IPV6TC1_IPV6FLOW1)
-+ && (p_CcNode->parseCode != CC_PC_FF_IPDSCP)
-+ && (p_CcNode->parseCode
-+ != CC_PC_FF_IPTOS_IPV6TC2_IPV6FLOW2)
-+ && glblMask)
-+ {
-+ glblMask = FALSE;
-+ p_CcNode->glblMaskSize = 4;
-+ p_CcNode->lclMask = TRUE;
-+ }
-+ break;
-+
-+ case (e_FM_PCD_EXTRACT_FROM_HDR):
-+ p_CcNode->sizeOfExtraction =
-+ p_CcNodeParam->extractCcParams.extractByHdr.extractByHdrType.fromHdr.size;
-+ p_CcNode->offset =
-+ p_CcNodeParam->extractCcParams.extractByHdr.extractByHdrType.fromHdr.offset;
-+ p_CcNode->userOffset =
-+ p_CcNodeParam->extractCcParams.extractByHdr.extractByHdrType.fromHdr.offset;
-+ p_CcNode->parseCode =
-+ GetPrParseCode(
-+ p_CcNodeParam->extractCcParams.extractByHdr.hdr,
-+ p_CcNodeParam->extractCcParams.extractByHdr.hdrIndex,
-+ p_CcNode->offset, glblMask,
-+ &p_CcNode->prsArrayOffset);
-+ break;
-+
-+ case (e_FM_PCD_EXTRACT_FROM_FIELD):
-+ p_CcNode->offset =
-+ p_CcNodeParam->extractCcParams.extractByHdr.extractByHdrType.fromField.offset;
-+ p_CcNode->userOffset =
-+ p_CcNodeParam->extractCcParams.extractByHdr.extractByHdrType.fromField.offset;
-+ p_CcNode->sizeOfExtraction =
-+ p_CcNodeParam->extractCcParams.extractByHdr.extractByHdrType.fromField.size;
-+ p_CcNode->parseCode =
-+ GetFieldParseCode(
-+ p_CcNodeParam->extractCcParams.extractByHdr.hdr,
-+ p_CcNodeParam->extractCcParams.extractByHdr.extractByHdrType.fromField.field,
-+ p_CcNode->offset,
-+ &p_CcNode->prsArrayOffset,
-+ p_CcNodeParam->extractCcParams.extractByHdr.hdrIndex);
-+ break;
-+
-+ default:
-+ DeleteNode(p_CcNode);
-+ RETURN_ERROR(MAJOR, E_INVALID_SELECTION, NO_MSG);
-+ }
-+ break;
-+
-+ case (e_FM_PCD_EXTRACT_NON_HDR):
-+ /* get the field code for the generic extract */
-+ p_CcNode->sizeOfExtraction =
-+ p_CcNodeParam->extractCcParams.extractNonHdr.size;
-+ p_CcNode->offset =
-+ p_CcNodeParam->extractCcParams.extractNonHdr.offset;
-+ p_CcNode->userOffset =
-+ p_CcNodeParam->extractCcParams.extractNonHdr.offset;
-+ p_CcNode->parseCode = GetGenParseCode(
-+ p_CcNodeParam->extractCcParams.extractNonHdr.src,
-+ p_CcNode->offset, glblMask, &p_CcNode->prsArrayOffset,
-+ fromIc, icCode);
-+
-+ if (p_CcNode->parseCode == CC_PC_GENERIC_IC_HASH_INDEXED)
-+ {
-+ if ((p_CcNode->offset + p_CcNode->sizeOfExtraction) > 8)
-+ {
-+ DeleteNode(p_CcNode);
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_SELECTION,
-+ ("when node of the type CC_PC_GENERIC_IC_HASH_INDEXED offset + size can not be bigger then size of HASH 64 bits (8 bytes)"));
-+ }
-+ }
-+ if ((p_CcNode->parseCode == CC_PC_GENERIC_IC_GMASK)
-+ || (p_CcNode->parseCode == CC_PC_GENERIC_IC_HASH_INDEXED))
-+ {
-+ p_CcNode->offset += p_CcNode->prsArrayOffset;
-+ p_CcNode->prsArrayOffset = 0;
-+ }
-+ break;
-+
-+ default:
-+ DeleteNode(p_CcNode);
-+ RETURN_ERROR(MAJOR, E_INVALID_SELECTION, NO_MSG);
-+ }
-+
-+ if (p_CcNode->parseCode == CC_PC_ILLEGAL)
-+ {
-+ DeleteNode(p_CcNode);
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("illegal extraction type"));
-+ }
-+
-+ if ((p_CcNode->sizeOfExtraction > FM_PCD_MAX_SIZE_OF_KEY)
-+ || !p_CcNode->sizeOfExtraction)
-+ {
-+ DeleteNode(p_CcNode);
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE,
-+ ("sizeOfExatrction can not be greater than 56 and not 0"));
-+ }
-+
-+ if (p_CcNodeParam->keysParams.keySize != p_CcNode->sizeOfExtraction)
-+ {
-+ DeleteNode(p_CcNode);
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE,
-+ ("keySize has to be equal to sizeOfExtraction"));
-+ }
-+
-+ p_CcNode->userSizeOfExtraction = p_CcNode->sizeOfExtraction;
-+
-+ if (!glblMask)
-+ memset(p_CcNode->p_GlblMask, 0xff, CC_GLBL_MASK_SIZE * sizeof(uint8_t));
-+
-+ err = CheckAndSetManipParamsWithCcNodeParams(p_CcNode);
-+ if (err != E_OK)
-+ {
-+ DeleteNode(p_CcNode);
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE,
-+ ("keySize has to be equal to sizeOfExtraction"));
-+ }
-+
-+ /* Calculating matching table entry size by rounding up the user-defined size of extraction to valid entry size */
-+ GetCcExtractKeySize(p_CcNode->sizeOfExtraction,
-+ &p_CcNode->ccKeySizeAccExtraction);
-+
-+ /* If local mask is used, it is stored next to each key in the keys match table */
-+ if (p_CcNode->lclMask)
-+ keySize = (uint32_t)(2 * p_CcNode->ccKeySizeAccExtraction);
-+ else
-+ keySize = p_CcNode->ccKeySizeAccExtraction;
-+
-+ /* Update CC shadow with maximal size required by this node */
-+ if (p_CcNode->maxNumOfKeys)
-+ {
-+ err = CalcAndUpdateCcShadow(p_CcNode, isKeyTblAlloc, &matchTableSize,
-+ &adTableSize);
-+ if (err != E_OK)
-+ {
-+ DeleteNode(p_CcNode);
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+
-+ p_CcNode->keysMatchTableMaxSize = matchTableSize;
-+
-+ if (p_CcNode->statisticsMode != e_FM_PCD_CC_STATS_MODE_NONE)
-+ {
-+ err = AllocStatsObjs(p_CcNode);
-+ if (err != E_OK)
-+ {
-+ DeleteNode(p_CcNode);
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+ }
-+
-+ /* If manipulation will be initialized before this node, it will use the table
-+ descriptor in the AD table of previous node and this node will need an extra
-+ AD as his table descriptor. */
-+ p_CcNode->h_TmpAd = (t_Handle)FM_MURAM_AllocMem(
-+ h_FmMuram, FM_PCD_CC_AD_ENTRY_SIZE, FM_PCD_CC_AD_TABLE_ALIGN);
-+ if (!p_CcNode->h_TmpAd)
-+ {
-+ DeleteNode(p_CcNode);
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY,
-+ ("MURAM allocation for CC action descriptor"));
-+ }
-+ }
-+ else
-+ {
-+ matchTableSize = (uint32_t)(keySize * sizeof(uint8_t)
-+ * (p_CcNode->numOfKeys + 1));
-+ adTableSize = (uint32_t)(FM_PCD_CC_AD_ENTRY_SIZE
-+ * (p_CcNode->numOfKeys + 1));
-+ }
-+
-+#if (DPAA_VERSION >= 11)
-+ switch (p_CcNode->statisticsMode)
-+ {
-+
-+ case e_FM_PCD_CC_STATS_MODE_RMON:
-+ /* If RMON statistics or RMON conditional statistics modes are requested,
-+ allocate frame length ranges array */
-+ p_CcNode->h_StatsFLRs = FM_MURAM_AllocMem(
-+ h_FmMuram,
-+ (uint32_t)(p_CcNode->numOfStatsFLRs)
-+ * FM_PCD_CC_STATS_FLR_SIZE,
-+ FM_PCD_CC_AD_TABLE_ALIGN);
-+
-+ if (!p_CcNode->h_StatsFLRs)
-+ {
-+ DeleteNode(p_CcNode);
-+ RETURN_ERROR(
-+ MAJOR, E_NO_MEMORY,
-+ ("MURAM allocation for CC frame length ranges array"));
-+ }
-+
-+ /* Initialize using value received from the user */
-+ for (tmp = 0; tmp < p_CcNode->numOfStatsFLRs; tmp++)
-+ {
-+ uint16_t flr =
-+ cpu_to_be16(p_CcNodeParam->keysParams.frameLengthRanges[tmp]);
-+
-+ h_StatsFLRs =
-+ PTR_MOVE(p_CcNode->h_StatsFLRs, tmp * FM_PCD_CC_STATS_FLR_SIZE);
-+
-+ MemCpy8(h_StatsFLRs,
-+ &flr,
-+ FM_PCD_CC_STATS_FLR_SIZE);
-+ }
-+ break;
-+
-+ default:
-+ break;
-+ }
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+ /* Allocate keys match table. Not required for some CC nodes, for example for IPv4 TTL
-+ identification, IPv6 hop count identification, etc. */
-+ if (isKeyTblAlloc)
-+ {
-+ p_CcNode->h_KeysMatchTable = (t_Handle)FM_MURAM_AllocMem(
-+ h_FmMuram, matchTableSize, FM_PCD_CC_KEYS_MATCH_TABLE_ALIGN);
-+ if (!p_CcNode->h_KeysMatchTable)
-+ {
-+ DeleteNode(p_CcNode);
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY,
-+ ("MURAM allocation for CC node key match table"));
-+ }
-+ MemSet8((uint8_t *)p_CcNode->h_KeysMatchTable, 0, matchTableSize);
-+ }
-+
-+ /* Allocate action descriptors table */
-+ p_CcNode->h_AdTable = (t_Handle)FM_MURAM_AllocMem(h_FmMuram, adTableSize,
-+ FM_PCD_CC_AD_TABLE_ALIGN);
-+ if (!p_CcNode->h_AdTable)
-+ {
-+ DeleteNode(p_CcNode);
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY,
-+ ("MURAM allocation for CC node action descriptors table"));
-+ }
-+ MemSet8((uint8_t *)p_CcNode->h_AdTable, 0, adTableSize);
-+
-+ p_KeysMatchTblTmp = p_CcNode->h_KeysMatchTable;
-+ p_AdTableTmp = p_CcNode->h_AdTable;
-+
-+ /* For each key, create the key and the next step AD */
-+ for (tmp = 0; tmp < p_CcNode->numOfKeys; tmp++)
-+ {
-+ p_KeyParams = &p_CcNodeParam->keysParams.keyParams[tmp];
-+
-+ if (p_KeysMatchTblTmp)
-+ {
-+ /* Copy the key */
-+ MemCpy8((void*)p_KeysMatchTblTmp, p_KeyParams->p_Key,
-+ p_CcNode->sizeOfExtraction);
-+
-+ /* Copy the key mask or initialize it to 0xFF..F */
-+ if (p_CcNode->lclMask && p_KeyParams->p_Mask)
-+ {
-+ MemCpy8(PTR_MOVE(p_KeysMatchTblTmp,
-+ p_CcNode->ccKeySizeAccExtraction), /* User's size of extraction rounded up to a valid matching table entry size */
-+ p_KeyParams->p_Mask, p_CcNode->sizeOfExtraction); /* Exact size of extraction as received from the user */
-+ }
-+ else
-+ if (p_CcNode->lclMask)
-+ {
-+ MemSet8(PTR_MOVE(p_KeysMatchTblTmp,
-+ p_CcNode->ccKeySizeAccExtraction), /* User's size of extraction rounded up to a valid matching table entry size */
-+ 0xff, p_CcNode->sizeOfExtraction); /* Exact size of extraction as received from the user */
-+ }
-+
-+ p_KeysMatchTblTmp =
-+ PTR_MOVE(p_KeysMatchTblTmp, keySize * sizeof(uint8_t));
-+ }
-+
-+ /* Create the next action descriptor in the match table */
-+ if (p_KeyParams->ccNextEngineParams.statisticsEn)
-+ {
-+ p_StatsObj = GetStatsObj(p_CcNode);
-+ ASSERT_COND(p_StatsObj);
-+
-+ statsParams.h_StatsAd = p_StatsObj->h_StatsAd;
-+ statsParams.h_StatsCounters = p_StatsObj->h_StatsCounters;
-+#if (DPAA_VERSION >= 11)
-+ statsParams.h_StatsFLRs = p_CcNode->h_StatsFLRs;
-+
-+#endif /* (DPAA_VERSION >= 11) */
-+ NextStepAd(p_AdTableTmp, &statsParams,
-+ &p_KeyParams->ccNextEngineParams, p_FmPcd);
-+
-+ p_CcNode->keyAndNextEngineParams[tmp].p_StatsObj = p_StatsObj;
-+ }
-+ else
-+ {
-+ NextStepAd(p_AdTableTmp, NULL, &p_KeyParams->ccNextEngineParams,
-+ p_FmPcd);
-+
-+ p_CcNode->keyAndNextEngineParams[tmp].p_StatsObj = NULL;
-+ }
-+
-+ p_AdTableTmp = PTR_MOVE(p_AdTableTmp, FM_PCD_CC_AD_ENTRY_SIZE);
-+ }
-+
-+ /* Update next engine for the 'miss' entry */
-+ if (p_CcNodeParam->keysParams.ccNextEngineParamsForMiss.statisticsEn)
-+ {
-+ p_StatsObj = GetStatsObj(p_CcNode);
-+ ASSERT_COND(p_StatsObj);
-+
-+ /* All 'bucket' nodes of a hash table should share the same statistics counters,
-+ allocated by the hash table. So, if this node is a bucket of a hash table,
-+ we'll replace the locally allocated counters with the shared counters. */
-+ if (p_CcNode->isHashBucket)
-+ {
-+ ASSERT_COND(p_CcNode->h_MissStatsCounters);
-+
-+ /* Store original counters pointer and replace it with mutual preallocated pointer */
-+ p_CcNode->h_PrivMissStatsCounters = p_StatsObj->h_StatsCounters;
-+ p_StatsObj->h_StatsCounters = p_CcNode->h_MissStatsCounters;
-+ }
-+
-+ statsParams.h_StatsAd = p_StatsObj->h_StatsAd;
-+ statsParams.h_StatsCounters = p_StatsObj->h_StatsCounters;
-+#if (DPAA_VERSION >= 11)
-+ statsParams.h_StatsFLRs = p_CcNode->h_StatsFLRs;
-+
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+ NextStepAd(p_AdTableTmp, &statsParams,
-+ &p_CcNodeParam->keysParams.ccNextEngineParamsForMiss,
-+ p_FmPcd);
-+
-+ p_CcNode->keyAndNextEngineParams[tmp].p_StatsObj = p_StatsObj;
-+ }
-+ else
-+ {
-+ NextStepAd(p_AdTableTmp, NULL,
-+ &p_CcNodeParam->keysParams.ccNextEngineParamsForMiss,
-+ p_FmPcd);
-+
-+ p_CcNode->keyAndNextEngineParams[tmp].p_StatsObj = NULL;
-+ }
-+
-+ /* This parameter will be used to initialize the "key length" field in the action descriptor
-+ that points to this node and it should be 0 for full field extraction */
-+ if (fullField == TRUE)
-+ p_CcNode->sizeOfExtraction = 0;
-+
-+ for (tmp = 0; tmp < MIN(p_CcNode->numOfKeys + 1, CC_MAX_NUM_OF_KEYS); tmp++)
-+ {
-+ if (p_CcNode->keyAndNextEngineParams[tmp].nextEngineParams.nextEngine
-+ == e_FM_PCD_CC)
-+ {
-+ p_FmPcdCcNextNode =
-+ (t_FmPcdCcNode*)p_CcNode->keyAndNextEngineParams[tmp].nextEngineParams.params.ccParams.h_CcNode;
-+ p_CcInformation = FindNodeInfoInReleventLst(
-+ &p_FmPcdCcNextNode->ccPrevNodesLst, (t_Handle)p_CcNode,
-+ p_FmPcdCcNextNode->h_Spinlock);
-+ if (!p_CcInformation)
-+ {
-+ memset(&ccNodeInfo, 0, sizeof(t_CcNodeInformation));
-+ ccNodeInfo.h_CcNode = (t_Handle)p_CcNode;
-+ ccNodeInfo.index = 1;
-+ EnqueueNodeInfoToRelevantLst(&p_FmPcdCcNextNode->ccPrevNodesLst,
-+ &ccNodeInfo,
-+ p_FmPcdCcNextNode->h_Spinlock);
-+ }
-+ else
-+ p_CcInformation->index++;
-+
-+ if (p_CcNode->keyAndNextEngineParams[tmp].nextEngineParams.h_Manip)
-+ {
-+ h_Manip =
-+ p_CcNode->keyAndNextEngineParams[tmp].nextEngineParams.h_Manip;
-+ p_CcInformation = FindNodeInfoInReleventLst(
-+ FmPcdManipGetNodeLstPointedOnThisManip(h_Manip),
-+ (t_Handle)p_CcNode, FmPcdManipGetSpinlock(h_Manip));
-+ if (!p_CcInformation)
-+ {
-+ memset(&ccNodeInfo, 0, sizeof(t_CcNodeInformation));
-+ ccNodeInfo.h_CcNode = (t_Handle)p_CcNode;
-+ ccNodeInfo.index = 1;
-+ EnqueueNodeInfoToRelevantLst(
-+ FmPcdManipGetNodeLstPointedOnThisManip(h_Manip),
-+ &ccNodeInfo, FmPcdManipGetSpinlock(h_Manip));
-+ }
-+ else
-+ p_CcInformation->index++;
-+ }
-+ }
-+ }
-+
-+ p_AdTableTmp = p_CcNode->h_AdTable;
-+
-+ if (!FmPcdLockTryLockAll(h_FmPcd))
-+ {
-+ FM_PCD_MatchTableDelete((t_Handle)p_CcNode);
-+ DBG(TRACE, ("FmPcdLockTryLockAll failed"));
-+ return ERROR_CODE(E_BUSY);
-+ }
-+
-+ /* Required action for each next engine */
-+ for (tmp = 0; tmp < MIN(p_CcNode->numOfKeys + 1, CC_MAX_NUM_OF_KEYS); tmp++)
-+ {
-+ if (p_CcNode->keyAndNextEngineParams[tmp].requiredAction)
-+ {
-+ err = SetRequiredAction(
-+ h_FmPcd,
-+ p_CcNode->keyAndNextEngineParams[tmp].requiredAction,
-+ &p_CcNode->keyAndNextEngineParams[tmp], p_AdTableTmp, 1,
-+ NULL);
-+ if (err)
-+ {
-+ FmPcdLockUnlockAll(h_FmPcd);
-+ FM_PCD_MatchTableDelete((t_Handle)p_CcNode);
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+ p_AdTableTmp = PTR_MOVE(p_AdTableTmp, FM_PCD_CC_AD_ENTRY_SIZE);
-+ }
-+ }
-+
-+ FmPcdLockUnlockAll(h_FmPcd);
-+
-+ return E_OK;
-+}
-+/************************** End of static functions **************************/
-+
-+/*****************************************************************************/
-+/* Inter-module API routines */
-+/*****************************************************************************/
-+
-+t_CcNodeInformation* FindNodeInfoInReleventLst(t_List *p_List, t_Handle h_Info,
-+ t_Handle h_Spinlock)
-+{
-+ t_CcNodeInformation *p_CcInformation;
-+ t_List *p_Pos;
-+ uint32_t intFlags;
-+
-+ intFlags = XX_LockIntrSpinlock(h_Spinlock);
-+
-+ for (p_Pos = LIST_FIRST(p_List); p_Pos != (p_List);
-+ p_Pos = LIST_NEXT(p_Pos))
-+ {
-+ p_CcInformation = CC_NODE_F_OBJECT(p_Pos);
-+
-+ ASSERT_COND(p_CcInformation->h_CcNode);
-+
-+ if (p_CcInformation->h_CcNode == h_Info)
-+ {
-+ XX_UnlockIntrSpinlock(h_Spinlock, intFlags);
-+ return p_CcInformation;
-+ }
-+ }
-+
-+ XX_UnlockIntrSpinlock(h_Spinlock, intFlags);
-+
-+ return NULL;
-+}
-+
-+void EnqueueNodeInfoToRelevantLst(t_List *p_List, t_CcNodeInformation *p_CcInfo,
-+ t_Handle h_Spinlock)
-+{
-+ t_CcNodeInformation *p_CcInformation;
-+ uint32_t intFlags = 0;
-+
-+ p_CcInformation = (t_CcNodeInformation *)XX_Malloc(
-+ sizeof(t_CcNodeInformation));
-+
-+ if (p_CcInformation)
-+ {
-+ memset(p_CcInformation, 0, sizeof(t_CcNodeInformation));
-+ memcpy(p_CcInformation, p_CcInfo, sizeof(t_CcNodeInformation));
-+ INIT_LIST(&p_CcInformation->node);
-+
-+ if (h_Spinlock)
-+ intFlags = XX_LockIntrSpinlock(h_Spinlock);
-+
-+ LIST_AddToTail(&p_CcInformation->node, p_List);
-+
-+ if (h_Spinlock)
-+ XX_UnlockIntrSpinlock(h_Spinlock, intFlags);
-+ }
-+ else
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("CC Node Information"));
-+}
-+
-+void DequeueNodeInfoFromRelevantLst(t_List *p_List, t_Handle h_Info,
-+ t_Handle h_Spinlock)
-+{
-+ t_CcNodeInformation *p_CcInformation = NULL;
-+ uint32_t intFlags = 0;
-+ t_List *p_Pos;
-+
-+ if (h_Spinlock)
-+ intFlags = XX_LockIntrSpinlock(h_Spinlock);
-+
-+ if (LIST_IsEmpty(p_List))
-+ {
-+ XX_RestoreAllIntr(intFlags);
-+ return;
-+ }
-+
-+ for (p_Pos = LIST_FIRST(p_List); p_Pos != (p_List);
-+ p_Pos = LIST_NEXT(p_Pos))
-+ {
-+ p_CcInformation = CC_NODE_F_OBJECT(p_Pos);
-+ ASSERT_COND(p_CcInformation);
-+ ASSERT_COND(p_CcInformation->h_CcNode);
-+ if (p_CcInformation->h_CcNode == h_Info)
-+ break;
-+ }
-+
-+ if (p_CcInformation)
-+ {
-+ LIST_DelAndInit(&p_CcInformation->node);
-+ XX_Free(p_CcInformation);
-+ }
-+
-+ if (h_Spinlock)
-+ XX_UnlockIntrSpinlock(h_Spinlock, intFlags);
-+}
-+
-+void NextStepAd(t_Handle h_Ad, t_FmPcdCcStatsParams *p_FmPcdCcStatsParams,
-+ t_FmPcdCcNextEngineParams *p_FmPcdCcNextEngineParams,
-+ t_FmPcd *p_FmPcd)
-+{
-+ switch (p_FmPcdCcNextEngineParams->nextEngine)
-+ {
-+ case (e_FM_PCD_KG):
-+ case (e_FM_PCD_PLCR):
-+ case (e_FM_PCD_DONE):
-+ /* if NIA is not CC, create a "result" type AD */
-+ FillAdOfTypeResult(h_Ad, p_FmPcdCcStatsParams, p_FmPcd,
-+ p_FmPcdCcNextEngineParams);
-+ break;
-+#if (DPAA_VERSION >= 11)
-+ case (e_FM_PCD_FR):
-+ if (p_FmPcdCcNextEngineParams->params.frParams.h_FrmReplic)
-+ {
-+ FillAdOfTypeContLookup(
-+ h_Ad, p_FmPcdCcStatsParams, p_FmPcd,
-+ p_FmPcdCcNextEngineParams->params.ccParams.h_CcNode,
-+ p_FmPcdCcNextEngineParams->h_Manip,
-+ p_FmPcdCcNextEngineParams->params.frParams.h_FrmReplic);
-+ FrmReplicGroupUpdateOwner(
-+ p_FmPcdCcNextEngineParams->params.frParams.h_FrmReplic,
-+ TRUE/* add */);
-+ }
-+ break;
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+ case (e_FM_PCD_CC):
-+ /* if NIA is not CC, create a TD to continue the CC lookup */
-+ FillAdOfTypeContLookup(
-+ h_Ad, p_FmPcdCcStatsParams, p_FmPcd,
-+ p_FmPcdCcNextEngineParams->params.ccParams.h_CcNode,
-+ p_FmPcdCcNextEngineParams->h_Manip, NULL);
-+
-+ UpdateNodeOwner(p_FmPcdCcNextEngineParams->params.ccParams.h_CcNode,
-+ TRUE);
-+ break;
-+
-+ default:
-+ return;
-+ }
-+}
-+
-+t_Error FmPcdCcTreeAddIPR(t_Handle h_FmPcd, t_Handle h_FmTree,
-+ t_Handle h_NetEnv, t_Handle h_IpReassemblyManip,
-+ bool createSchemes)
-+{
-+ t_FmPcdCcTree *p_FmPcdCcTree = (t_FmPcdCcTree *)h_FmTree;
-+ t_FmPcdCcNextEngineParams nextEngineParams;
-+ t_NetEnvParams netEnvParams;
-+ t_Handle h_Ad;
-+ bool isIpv6Present;
-+ uint8_t ipv4GroupId, ipv6GroupId;
-+ t_Error err;
-+
-+ ASSERT_COND(p_FmPcdCcTree);
-+
-+ /* this routine must be protected by the calling routine! */
-+
-+ memset(&nextEngineParams, 0, sizeof(t_FmPcdCcNextEngineParams));
-+ memset(&netEnvParams, 0, sizeof(t_NetEnvParams));
-+
-+ h_Ad = UINT_TO_PTR(p_FmPcdCcTree->ccTreeBaseAddr);
-+
-+ isIpv6Present = FmPcdManipIpReassmIsIpv6Hdr(h_IpReassemblyManip);
-+
-+ if (isIpv6Present
-+ && (p_FmPcdCcTree->numOfEntries > (FM_PCD_MAX_NUM_OF_CC_GROUPS - 2)))
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("need two free entries for IPR"));
-+
-+ if (p_FmPcdCcTree->numOfEntries > (FM_PCD_MAX_NUM_OF_CC_GROUPS - 1))
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("need two free entries for IPR"));
-+
-+ nextEngineParams.nextEngine = e_FM_PCD_DONE;
-+ nextEngineParams.h_Manip = h_IpReassemblyManip;
-+
-+ /* Lock tree */
-+ err = CcRootTryLock(p_FmPcdCcTree);
-+ if (err)
-+ return ERROR_CODE(E_BUSY);
-+
-+ if (p_FmPcdCcTree->h_IpReassemblyManip == h_IpReassemblyManip)
-+ {
-+ CcRootReleaseLock(p_FmPcdCcTree);
-+ return E_OK;
-+ }
-+
-+ if ((p_FmPcdCcTree->h_IpReassemblyManip)
-+ && (p_FmPcdCcTree->h_IpReassemblyManip != h_IpReassemblyManip))
-+ {
-+ CcRootReleaseLock(p_FmPcdCcTree);
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE,
-+ ("This tree was previously updated with different IPR"));
-+ }
-+
-+ /* Initialize IPR for the first time for this tree */
-+ if (isIpv6Present)
-+ {
-+ ipv6GroupId = p_FmPcdCcTree->numOfGrps++;
-+ p_FmPcdCcTree->fmPcdGroupParam[ipv6GroupId].baseGroupEntry =
-+ (FM_PCD_MAX_NUM_OF_CC_GROUPS - 2);
-+
-+ if (createSchemes)
-+ {
-+ err = FmPcdManipBuildIpReassmScheme(h_FmPcd, h_NetEnv,
-+ p_FmPcdCcTree,
-+ h_IpReassemblyManip, FALSE,
-+ ipv6GroupId);
-+ if (err)
-+ {
-+ p_FmPcdCcTree->numOfGrps--;
-+ CcRootReleaseLock(p_FmPcdCcTree);
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+ }
-+
-+ NextStepAd(
-+ PTR_MOVE(h_Ad, (FM_PCD_MAX_NUM_OF_CC_GROUPS-2) * FM_PCD_CC_AD_ENTRY_SIZE),
-+ NULL, &nextEngineParams, h_FmPcd);
-+ }
-+
-+ ipv4GroupId = p_FmPcdCcTree->numOfGrps++;
-+ p_FmPcdCcTree->fmPcdGroupParam[ipv4GroupId].totalBitsMask = 0;
-+ p_FmPcdCcTree->fmPcdGroupParam[ipv4GroupId].baseGroupEntry =
-+ (FM_PCD_MAX_NUM_OF_CC_GROUPS - 1);
-+
-+ if (createSchemes)
-+ {
-+ err = FmPcdManipBuildIpReassmScheme(h_FmPcd, h_NetEnv, p_FmPcdCcTree,
-+ h_IpReassemblyManip, TRUE,
-+ ipv4GroupId);
-+ if (err)
-+ {
-+ p_FmPcdCcTree->numOfGrps--;
-+ if (isIpv6Present)
-+ {
-+ p_FmPcdCcTree->numOfGrps--;
-+ FmPcdManipDeleteIpReassmSchemes(h_IpReassemblyManip);
-+ }
-+ CcRootReleaseLock(p_FmPcdCcTree);
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+ }
-+
-+ NextStepAd(
-+ PTR_MOVE(h_Ad, (FM_PCD_MAX_NUM_OF_CC_GROUPS-1) * FM_PCD_CC_AD_ENTRY_SIZE),
-+ NULL, &nextEngineParams, h_FmPcd);
-+
-+ p_FmPcdCcTree->h_IpReassemblyManip = h_IpReassemblyManip;
-+
-+ CcRootReleaseLock(p_FmPcdCcTree);
-+
-+ return E_OK;
-+}
-+
-+t_Error FmPcdCcTreeAddCPR(t_Handle h_FmPcd, t_Handle h_FmTree,
-+ t_Handle h_NetEnv, t_Handle h_ReassemblyManip,
-+ bool createSchemes)
-+{
-+ t_FmPcdCcTree *p_FmPcdCcTree = (t_FmPcdCcTree *)h_FmTree;
-+ t_FmPcdCcNextEngineParams nextEngineParams;
-+ t_NetEnvParams netEnvParams;
-+ t_Handle h_Ad;
-+ uint8_t groupId;
-+ t_Error err;
-+
-+ ASSERT_COND(p_FmPcdCcTree);
-+
-+ /* this routine must be protected by the calling routine! */
-+ memset(&nextEngineParams, 0, sizeof(t_FmPcdCcNextEngineParams));
-+ memset(&netEnvParams, 0, sizeof(t_NetEnvParams));
-+
-+ h_Ad = UINT_TO_PTR(p_FmPcdCcTree->ccTreeBaseAddr);
-+
-+ if (p_FmPcdCcTree->numOfEntries > (FM_PCD_MAX_NUM_OF_CC_GROUPS - 1))
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("need one free entries for CPR"));
-+
-+ nextEngineParams.nextEngine = e_FM_PCD_DONE;
-+ nextEngineParams.h_Manip = h_ReassemblyManip;
-+
-+ /* Lock tree */
-+ err = CcRootTryLock(p_FmPcdCcTree);
-+ if (err)
-+ return ERROR_CODE(E_BUSY);
-+
-+ if (p_FmPcdCcTree->h_CapwapReassemblyManip == h_ReassemblyManip)
-+ {
-+ CcRootReleaseLock(p_FmPcdCcTree);
-+ return E_OK;
-+ }
-+
-+ if ((p_FmPcdCcTree->h_CapwapReassemblyManip)
-+ && (p_FmPcdCcTree->h_CapwapReassemblyManip != h_ReassemblyManip))
-+ {
-+ CcRootReleaseLock(p_FmPcdCcTree);
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE,
-+ ("This tree was previously updated with different CPR"));
-+ }
-+
-+ groupId = p_FmPcdCcTree->numOfGrps++;
-+ p_FmPcdCcTree->fmPcdGroupParam[groupId].baseGroupEntry =
-+ (FM_PCD_MAX_NUM_OF_CC_GROUPS - 1);
-+
-+ if (createSchemes)
-+ {
-+ err = FmPcdManipBuildCapwapReassmScheme(h_FmPcd, h_NetEnv,
-+ p_FmPcdCcTree,
-+ h_ReassemblyManip, groupId);
-+ if (err)
-+ {
-+ p_FmPcdCcTree->numOfGrps--;
-+ CcRootReleaseLock(p_FmPcdCcTree);
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+ }
-+
-+ NextStepAd(
-+ PTR_MOVE(h_Ad, (FM_PCD_MAX_NUM_OF_CC_GROUPS-1) * FM_PCD_CC_AD_ENTRY_SIZE),
-+ NULL, &nextEngineParams, h_FmPcd);
-+
-+ p_FmPcdCcTree->h_CapwapReassemblyManip = h_ReassemblyManip;
-+
-+ CcRootReleaseLock(p_FmPcdCcTree);
-+
-+ return E_OK;
-+}
-+
-+t_Handle FmPcdCcTreeGetSavedManipParams(t_Handle h_FmTree)
-+{
-+ t_FmPcdCcTree *p_FmPcdCcTree = (t_FmPcdCcTree *)h_FmTree;
-+
-+ ASSERT_COND(p_FmPcdCcTree);
-+
-+ return p_FmPcdCcTree->h_FmPcdCcSavedManipParams;
-+}
-+
-+void FmPcdCcTreeSetSavedManipParams(t_Handle h_FmTree,
-+ t_Handle h_SavedManipParams)
-+{
-+ t_FmPcdCcTree *p_FmPcdCcTree = (t_FmPcdCcTree *)h_FmTree;
-+
-+ ASSERT_COND(p_FmPcdCcTree);
-+
-+ p_FmPcdCcTree->h_FmPcdCcSavedManipParams = h_SavedManipParams;
-+}
-+
-+uint8_t FmPcdCcGetParseCode(t_Handle h_CcNode)
-+{
-+ t_FmPcdCcNode *p_CcNode = (t_FmPcdCcNode *)h_CcNode;
-+
-+ ASSERT_COND(p_CcNode);
-+
-+ return p_CcNode->parseCode;
-+}
-+
-+uint8_t FmPcdCcGetOffset(t_Handle h_CcNode)
-+{
-+ t_FmPcdCcNode *p_CcNode = (t_FmPcdCcNode *)h_CcNode;
-+
-+ ASSERT_COND(p_CcNode);
-+
-+ return p_CcNode->offset;
-+}
-+
-+uint16_t FmPcdCcGetNumOfKeys(t_Handle h_CcNode)
-+{
-+ t_FmPcdCcNode *p_CcNode = (t_FmPcdCcNode *)h_CcNode;
-+
-+ ASSERT_COND(p_CcNode);
-+
-+ return p_CcNode->numOfKeys;
-+}
-+
-+t_Error FmPcdCcModifyNextEngineParamTree(
-+ t_Handle h_FmPcd, t_Handle h_FmPcdCcTree, uint8_t grpId, uint8_t index,
-+ t_FmPcdCcNextEngineParams *p_FmPcdCcNextEngineParams)
-+{
-+ t_FmPcdCcTree *p_FmPcdCcTree = (t_FmPcdCcTree *)h_FmPcdCcTree;
-+ t_FmPcd *p_FmPcd;
-+ t_List h_OldPointersLst, h_NewPointersLst;
-+ uint16_t keyIndex;
-+ t_FmPcdModifyCcKeyAdditionalParams *p_ModifyKeyParams;
-+ t_Error err = E_OK;
-+
-+ SANITY_CHECK_RETURN_ERROR(h_FmPcd, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(h_FmPcdCcTree, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR((grpId <= 7), E_INVALID_VALUE);
-+
-+ if (grpId >= p_FmPcdCcTree->numOfGrps)
-+ RETURN_ERROR(MAJOR, E_INVALID_HANDLE,
-+ ("grpId you asked > numOfGroup of relevant tree"));
-+
-+ if (index >= p_FmPcdCcTree->fmPcdGroupParam[grpId].numOfEntriesInGroup)
-+ RETURN_ERROR(MAJOR, E_INVALID_HANDLE, ("index > numOfEntriesInGroup"));
-+
-+ p_FmPcd = (t_FmPcd *)h_FmPcd;
-+
-+ INIT_LIST(&h_OldPointersLst);
-+ INIT_LIST(&h_NewPointersLst);
-+
-+ keyIndex = (uint16_t)(p_FmPcdCcTree->fmPcdGroupParam[grpId].baseGroupEntry
-+ + index);
-+
-+ p_ModifyKeyParams = ModifyNodeCommonPart(p_FmPcdCcTree, keyIndex,
-+ e_MODIFY_STATE_CHANGE, FALSE,
-+ FALSE, TRUE);
-+ if (!p_ModifyKeyParams)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, NO_MSG);
-+
-+ p_ModifyKeyParams->tree = TRUE;
-+
-+ if (p_FmPcd->p_CcShadow
-+ && !TRY_LOCK(p_FmPcd->h_ShadowSpinlock, &p_FmPcd->shadowLock))
-+ {
-+ XX_Free(p_ModifyKeyParams);
-+ return ERROR_CODE(E_BUSY);
-+ }
-+
-+ err = BuildNewNodeModifyNextEngine(p_FmPcd, p_FmPcdCcTree, keyIndex,
-+ p_FmPcdCcNextEngineParams,
-+ &h_OldPointersLst, &h_NewPointersLst,
-+ p_ModifyKeyParams);
-+ if (err)
-+ {
-+ XX_Free(p_ModifyKeyParams);
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+
-+ err = DoDynamicChange(p_FmPcd, &h_OldPointersLst, &h_NewPointersLst,
-+ p_ModifyKeyParams, FALSE);
-+
-+ if (p_FmPcd->p_CcShadow)
-+ RELEASE_LOCK(p_FmPcd->shadowLock);
-+
-+ ReleaseLst(&h_OldPointersLst);
-+ ReleaseLst(&h_NewPointersLst);
-+
-+ return err;
-+
-+}
-+
-+t_Error FmPcdCcRemoveKey(t_Handle h_FmPcd, t_Handle h_FmPcdCcNode,
-+ uint16_t keyIndex)
-+{
-+
-+ t_FmPcdCcNode *p_CcNode = (t_FmPcdCcNode *)h_FmPcdCcNode;
-+ t_FmPcd *p_FmPcd;
-+ t_FmPcdModifyCcKeyAdditionalParams *p_ModifyKeyParams;
-+ t_List h_OldPointersLst, h_NewPointersLst;
-+ bool useShadowStructs = FALSE;
-+ t_Error err = E_OK;
-+
-+ if (keyIndex >= p_CcNode->numOfKeys)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE,
-+ ("impossible to remove key when numOfKeys <= keyIndex"));
-+
-+ if (p_CcNode->h_FmPcd != h_FmPcd)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("handler to FmPcd is different from the handle provided at node initialization time"));
-+
-+ p_FmPcd = (t_FmPcd *)p_CcNode->h_FmPcd;
-+
-+ INIT_LIST(&h_OldPointersLst);
-+ INIT_LIST(&h_NewPointersLst);
-+
-+ p_ModifyKeyParams = ModifyNodeCommonPart(p_CcNode, keyIndex,
-+ e_MODIFY_STATE_REMOVE, TRUE, TRUE,
-+ FALSE);
-+ if (!p_ModifyKeyParams)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, NO_MSG);
-+
-+ if (p_CcNode->maxNumOfKeys)
-+ {
-+ if (!TRY_LOCK(p_FmPcd->h_ShadowSpinlock, &p_FmPcd->shadowLock))
-+ {
-+ XX_Free(p_ModifyKeyParams);
-+ return ERROR_CODE(E_BUSY);
-+ }
-+
-+ useShadowStructs = TRUE;
-+ }
-+
-+ err = BuildNewNodeRemoveKey(p_CcNode, keyIndex, p_ModifyKeyParams);
-+ if (err)
-+ {
-+ XX_Free(p_ModifyKeyParams);
-+ if (p_CcNode->maxNumOfKeys)
-+ RELEASE_LOCK(p_FmPcd->shadowLock);
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+
-+ err = UpdatePtrWhichPointOnCrntMdfNode(p_CcNode, p_ModifyKeyParams,
-+ &h_OldPointersLst,
-+ &h_NewPointersLst);
-+ if (err)
-+ {
-+ ReleaseNewNodeCommonPart(p_ModifyKeyParams);
-+ XX_Free(p_ModifyKeyParams);
-+ if (p_CcNode->maxNumOfKeys)
-+ RELEASE_LOCK(p_FmPcd->shadowLock);
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+
-+ err = DoDynamicChange(p_FmPcd, &h_OldPointersLst, &h_NewPointersLst,
-+ p_ModifyKeyParams, useShadowStructs);
-+
-+ if (p_CcNode->maxNumOfKeys)
-+ RELEASE_LOCK(p_FmPcd->shadowLock);
-+
-+ ReleaseLst(&h_OldPointersLst);
-+ ReleaseLst(&h_NewPointersLst);
-+
-+ return err;
-+}
-+
-+t_Error FmPcdCcModifyKey(t_Handle h_FmPcd, t_Handle h_FmPcdCcNode,
-+ uint16_t keyIndex, uint8_t keySize, uint8_t *p_Key,
-+ uint8_t *p_Mask)
-+{
-+ t_FmPcdCcNode *p_CcNode = (t_FmPcdCcNode *)h_FmPcdCcNode;
-+ t_FmPcd *p_FmPcd;
-+ t_List h_OldPointersLst, h_NewPointersLst;
-+ t_FmPcdModifyCcKeyAdditionalParams *p_ModifyKeyParams;
-+ uint16_t tmpKeyIndex;
-+ bool useShadowStructs = FALSE;
-+ t_Error err = E_OK;
-+
-+ if (keyIndex >= p_CcNode->numOfKeys)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE,
-+ ("keyIndex > previously cleared last index + 1"));
-+
-+ if (keySize != p_CcNode->userSizeOfExtraction)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("size for ModifyKey has to be the same as defined in SetNode"));
-+
-+ if (p_CcNode->h_FmPcd != h_FmPcd)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("handler to FmPcd is different from the handle provided at node initialization time"));
-+
-+ err = FindKeyIndex(h_FmPcdCcNode, keySize, p_Key, p_Mask, &tmpKeyIndex);
-+ if (GET_ERROR_TYPE(err) != E_NOT_FOUND)
-+ RETURN_ERROR(
-+ MINOR,
-+ E_ALREADY_EXISTS,
-+ ("The received key and mask pair was already found in the match table of the provided node"));
-+
-+ p_FmPcd = (t_FmPcd *)p_CcNode->h_FmPcd;
-+
-+ INIT_LIST(&h_OldPointersLst);
-+ INIT_LIST(&h_NewPointersLst);
-+
-+ p_ModifyKeyParams = ModifyNodeCommonPart(p_CcNode, keyIndex,
-+ e_MODIFY_STATE_CHANGE, TRUE, TRUE,
-+ FALSE);
-+ if (!p_ModifyKeyParams)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, NO_MSG);
-+
-+ if (p_CcNode->maxNumOfKeys)
-+ {
-+ if (!TRY_LOCK(p_FmPcd->h_ShadowSpinlock, &p_FmPcd->shadowLock))
-+ {
-+ XX_Free(p_ModifyKeyParams);
-+ return ERROR_CODE(E_BUSY);
-+ }
-+
-+ useShadowStructs = TRUE;
-+ }
-+
-+ err = BuildNewNodeModifyKey(p_CcNode, keyIndex, p_Key, p_Mask,
-+ p_ModifyKeyParams);
-+ if (err)
-+ {
-+ XX_Free(p_ModifyKeyParams);
-+ if (p_CcNode->maxNumOfKeys)
-+ RELEASE_LOCK(p_FmPcd->shadowLock);
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+
-+ err = UpdatePtrWhichPointOnCrntMdfNode(p_CcNode, p_ModifyKeyParams,
-+ &h_OldPointersLst,
-+ &h_NewPointersLst);
-+ if (err)
-+ {
-+ ReleaseNewNodeCommonPart(p_ModifyKeyParams);
-+ XX_Free(p_ModifyKeyParams);
-+ if (p_CcNode->maxNumOfKeys)
-+ RELEASE_LOCK(p_FmPcd->shadowLock);
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+
-+ err = DoDynamicChange(p_FmPcd, &h_OldPointersLst, &h_NewPointersLst,
-+ p_ModifyKeyParams, useShadowStructs);
-+
-+ if (p_CcNode->maxNumOfKeys)
-+ RELEASE_LOCK(p_FmPcd->shadowLock);
-+
-+ ReleaseLst(&h_OldPointersLst);
-+ ReleaseLst(&h_NewPointersLst);
-+
-+ return err;
-+}
-+
-+t_Error FmPcdCcModifyMissNextEngineParamNode(
-+ t_Handle h_FmPcd, t_Handle h_FmPcdCcNode,
-+ t_FmPcdCcNextEngineParams *p_FmPcdCcNextEngineParams)
-+{
-+ t_FmPcdCcNode *p_CcNode = (t_FmPcdCcNode *)h_FmPcdCcNode;
-+ t_FmPcd *p_FmPcd;
-+ t_List h_OldPointersLst, h_NewPointersLst;
-+ uint16_t keyIndex;
-+ t_FmPcdModifyCcKeyAdditionalParams *p_ModifyKeyParams;
-+ t_Error err = E_OK;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_CcNode, E_INVALID_VALUE);
-+
-+ keyIndex = p_CcNode->numOfKeys;
-+
-+ p_FmPcd = (t_FmPcd *)p_CcNode->h_FmPcd;
-+
-+ INIT_LIST(&h_OldPointersLst);
-+ INIT_LIST(&h_NewPointersLst);
-+
-+ p_ModifyKeyParams = ModifyNodeCommonPart(p_CcNode, keyIndex,
-+ e_MODIFY_STATE_CHANGE, FALSE, TRUE,
-+ FALSE);
-+ if (!p_ModifyKeyParams)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, NO_MSG);
-+
-+ if (p_CcNode->maxNumOfKeys
-+ && !TRY_LOCK(p_FmPcd->h_ShadowSpinlock, &p_FmPcd->shadowLock))
-+ {
-+ XX_Free(p_ModifyKeyParams);
-+ return ERROR_CODE(E_BUSY);
-+ }
-+
-+ err = BuildNewNodeModifyNextEngine(h_FmPcd, p_CcNode, keyIndex,
-+ p_FmPcdCcNextEngineParams,
-+ &h_OldPointersLst, &h_NewPointersLst,
-+ p_ModifyKeyParams);
-+ if (err)
-+ {
-+ XX_Free(p_ModifyKeyParams);
-+ if (p_CcNode->maxNumOfKeys)
-+ RELEASE_LOCK(p_FmPcd->shadowLock);
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+
-+ err = DoDynamicChange(p_FmPcd, &h_OldPointersLst, &h_NewPointersLst,
-+ p_ModifyKeyParams, FALSE);
-+
-+ if (p_CcNode->maxNumOfKeys)
-+ RELEASE_LOCK(p_FmPcd->shadowLock);
-+
-+ ReleaseLst(&h_OldPointersLst);
-+ ReleaseLst(&h_NewPointersLst);
-+
-+ return err;
-+}
-+
-+t_Error FmPcdCcAddKey(t_Handle h_FmPcd, t_Handle h_FmPcdCcNode,
-+ uint16_t keyIndex, uint8_t keySize,
-+ t_FmPcdCcKeyParams *p_FmPcdCcKeyParams)
-+{
-+ t_FmPcdCcNode *p_CcNode = (t_FmPcdCcNode *)h_FmPcdCcNode;
-+ t_FmPcd *p_FmPcd;
-+ t_FmPcdModifyCcKeyAdditionalParams *p_ModifyKeyParams;
-+ t_List h_OldPointersLst, h_NewPointersLst;
-+ bool useShadowStructs = FALSE;
-+ uint16_t tmpKeyIndex;
-+ t_Error err = E_OK;
-+
-+ if (keyIndex > p_CcNode->numOfKeys)
-+ RETURN_ERROR(MAJOR, E_NOT_IN_RANGE,
-+ ("keyIndex > previously cleared last index + 1"));
-+
-+ if (keySize != p_CcNode->userSizeOfExtraction)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("keySize has to be defined as it was defined in initialization step"));
-+
-+ if (p_CcNode->h_FmPcd != h_FmPcd)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("handler to FmPcd is different from the handle provided at node initialization time"));
-+
-+ if (p_CcNode->maxNumOfKeys)
-+ {
-+ if (p_CcNode->numOfKeys == p_CcNode->maxNumOfKeys)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_FULL,
-+ ("number of keys exceeds the maximal number of keys provided at node initialization time"));
-+ }
-+ else
-+ if (p_CcNode->numOfKeys == FM_PCD_MAX_NUM_OF_KEYS)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("number of keys can not be larger than %d", FM_PCD_MAX_NUM_OF_KEYS));
-+
-+ err = FindKeyIndex(h_FmPcdCcNode, keySize, p_FmPcdCcKeyParams->p_Key,
-+ p_FmPcdCcKeyParams->p_Mask, &tmpKeyIndex);
-+ if (GET_ERROR_TYPE(err) != E_NOT_FOUND)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_ALREADY_EXISTS,
-+ ("The received key and mask pair was already found in the match table of the provided node"));
-+
-+ p_FmPcd = (t_FmPcd *)p_CcNode->h_FmPcd;
-+
-+ INIT_LIST(&h_OldPointersLst);
-+ INIT_LIST(&h_NewPointersLst);
-+
-+ p_ModifyKeyParams = ModifyNodeCommonPart(p_CcNode, keyIndex,
-+ e_MODIFY_STATE_ADD, TRUE, TRUE,
-+ FALSE);
-+ if (!p_ModifyKeyParams)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, NO_MSG);
-+
-+ if (p_CcNode->maxNumOfKeys)
-+ {
-+ if (!TRY_LOCK(p_FmPcd->h_ShadowSpinlock, &p_FmPcd->shadowLock))
-+ {
-+ XX_Free(p_ModifyKeyParams);
-+ return ERROR_CODE(E_BUSY);
-+ }
-+
-+ useShadowStructs = TRUE;
-+ }
-+
-+ err = BuildNewNodeAddOrMdfyKeyAndNextEngine(h_FmPcd, p_CcNode, keyIndex,
-+ p_FmPcdCcKeyParams,
-+ p_ModifyKeyParams, TRUE);
-+ if (err)
-+ {
-+ ReleaseNewNodeCommonPart(p_ModifyKeyParams);
-+ XX_Free(p_ModifyKeyParams);
-+ if (p_CcNode->maxNumOfKeys)
-+ RELEASE_LOCK(p_FmPcd->shadowLock);
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+
-+ err = UpdatePtrWhichPointOnCrntMdfNode(p_CcNode, p_ModifyKeyParams,
-+ &h_OldPointersLst,
-+ &h_NewPointersLst);
-+ if (err)
-+ {
-+ ReleaseNewNodeCommonPart(p_ModifyKeyParams);
-+ XX_Free(p_ModifyKeyParams);
-+ if (p_CcNode->maxNumOfKeys)
-+ RELEASE_LOCK(p_FmPcd->shadowLock);
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+
-+ err = DoDynamicChange(p_FmPcd, &h_OldPointersLst, &h_NewPointersLst,
-+ p_ModifyKeyParams, useShadowStructs);
-+ if (p_CcNode->maxNumOfKeys)
-+ RELEASE_LOCK(p_FmPcd->shadowLock);
-+
-+ ReleaseLst(&h_OldPointersLst);
-+ ReleaseLst(&h_NewPointersLst);
-+
-+ return err;
-+}
-+
-+t_Error FmPcdCcModifyKeyAndNextEngine(t_Handle h_FmPcd, t_Handle h_FmPcdCcNode,
-+ uint16_t keyIndex, uint8_t keySize,
-+ t_FmPcdCcKeyParams *p_FmPcdCcKeyParams)
-+{
-+ t_FmPcdCcNode *p_CcNode = (t_FmPcdCcNode *)h_FmPcdCcNode;
-+ t_FmPcd *p_FmPcd;
-+ t_List h_OldPointersLst, h_NewPointersLst;
-+ t_FmPcdModifyCcKeyAdditionalParams *p_ModifyKeyParams;
-+ uint16_t tmpKeyIndex;
-+ bool useShadowStructs = FALSE;
-+ t_Error err = E_OK;
-+
-+ if (keyIndex > p_CcNode->numOfKeys)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE,
-+ ("keyIndex > previously cleared last index + 1"));
-+
-+ if (keySize != p_CcNode->userSizeOfExtraction)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("keySize has to be defined as it was defined in initialization step"));
-+
-+ if (p_CcNode->h_FmPcd != h_FmPcd)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("handler to FmPcd is different from the handle provided at node initialization time"));
-+
-+ err = FindKeyIndex(h_FmPcdCcNode, keySize, p_FmPcdCcKeyParams->p_Key,
-+ p_FmPcdCcKeyParams->p_Mask, &tmpKeyIndex);
-+ if (GET_ERROR_TYPE(err) != E_NOT_FOUND)
-+ RETURN_ERROR(
-+ MINOR,
-+ E_ALREADY_EXISTS,
-+ ("The received key and mask pair was already found in the match table of the provided node"));
-+
-+ p_FmPcd = (t_FmPcd *)p_CcNode->h_FmPcd;
-+
-+ INIT_LIST(&h_OldPointersLst);
-+ INIT_LIST(&h_NewPointersLst);
-+
-+ p_ModifyKeyParams = ModifyNodeCommonPart(p_CcNode, keyIndex,
-+ e_MODIFY_STATE_CHANGE, TRUE, TRUE,
-+ FALSE);
-+ if (!p_ModifyKeyParams)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, NO_MSG);
-+
-+ if (p_CcNode->maxNumOfKeys)
-+ {
-+ if (!TRY_LOCK(p_FmPcd->h_ShadowSpinlock, &p_FmPcd->shadowLock))
-+ {
-+ XX_Free(p_ModifyKeyParams);
-+ return ERROR_CODE(E_BUSY);
-+ }
-+
-+ useShadowStructs = TRUE;
-+ }
-+
-+ err = BuildNewNodeAddOrMdfyKeyAndNextEngine(h_FmPcd, p_CcNode, keyIndex,
-+ p_FmPcdCcKeyParams,
-+ p_ModifyKeyParams, FALSE);
-+ if (err)
-+ {
-+ ReleaseNewNodeCommonPart(p_ModifyKeyParams);
-+ XX_Free(p_ModifyKeyParams);
-+ if (p_CcNode->maxNumOfKeys)
-+ RELEASE_LOCK(p_FmPcd->shadowLock);
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+
-+ err = UpdatePtrWhichPointOnCrntMdfNode(p_CcNode, p_ModifyKeyParams,
-+ &h_OldPointersLst,
-+ &h_NewPointersLst);
-+ if (err)
-+ {
-+ ReleaseNewNodeCommonPart(p_ModifyKeyParams);
-+ XX_Free(p_ModifyKeyParams);
-+ if (p_CcNode->maxNumOfKeys)
-+ RELEASE_LOCK(p_FmPcd->shadowLock);
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+
-+ err = DoDynamicChange(p_FmPcd, &h_OldPointersLst, &h_NewPointersLst,
-+ p_ModifyKeyParams, useShadowStructs);
-+
-+ if (p_CcNode->maxNumOfKeys)
-+ RELEASE_LOCK(p_FmPcd->shadowLock);
-+
-+ ReleaseLst(&h_OldPointersLst);
-+ ReleaseLst(&h_NewPointersLst);
-+
-+ return err;
-+}
-+
-+uint32_t FmPcdCcGetNodeAddrOffsetFromNodeInfo(t_Handle h_FmPcd,
-+ t_Handle h_Pointer)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd *)h_FmPcd;
-+ t_CcNodeInformation *p_CcNodeInfo;
-+
-+ SANITY_CHECK_RETURN_VALUE(h_FmPcd, E_INVALID_HANDLE,
-+ (uint32_t)ILLEGAL_BASE);
-+
-+ p_CcNodeInfo = CC_NODE_F_OBJECT(h_Pointer);
-+
-+ return (uint32_t)(XX_VirtToPhys(p_CcNodeInfo->h_CcNode)
-+ - p_FmPcd->physicalMuramBase);
-+}
-+
-+t_Error FmPcdCcGetGrpParams(t_Handle h_FmPcdCcTree, uint8_t grpId,
-+ uint32_t *p_GrpBits, uint8_t *p_GrpBase)
-+{
-+ t_FmPcdCcTree *p_FmPcdCcTree = (t_FmPcdCcTree *)h_FmPcdCcTree;
-+
-+ SANITY_CHECK_RETURN_ERROR(h_FmPcdCcTree, E_INVALID_HANDLE);
-+
-+ if (grpId >= p_FmPcdCcTree->numOfGrps)
-+ RETURN_ERROR(MAJOR, E_INVALID_HANDLE,
-+ ("grpId you asked > numOfGroup of relevant tree"));
-+
-+ *p_GrpBits = p_FmPcdCcTree->fmPcdGroupParam[grpId].totalBitsMask;
-+ *p_GrpBase = p_FmPcdCcTree->fmPcdGroupParam[grpId].baseGroupEntry;
-+
-+ return E_OK;
-+}
-+
-+t_Error FmPcdCcBindTree(t_Handle h_FmPcd, t_Handle h_PcdParams,
-+ t_Handle h_FmPcdCcTree, uint32_t *p_Offset,
-+ t_Handle h_FmPort)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+ t_FmPcdCcTree *p_FmPcdCcTree = (t_FmPcdCcTree *)h_FmPcdCcTree;
-+ t_Error err = E_OK;
-+
-+ SANITY_CHECK_RETURN_ERROR(h_FmPcd, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(h_FmPcdCcTree, E_INVALID_HANDLE);
-+
-+ /* this routine must be protected by the calling routine by locking all PCD modules! */
-+
-+ err = CcUpdateParams(h_FmPcd, h_PcdParams, h_FmPort, h_FmPcdCcTree, TRUE);
-+
-+ if (err == E_OK)
-+ UpdateCcRootOwner(p_FmPcdCcTree, TRUE);
-+
-+ *p_Offset = (uint32_t)(XX_VirtToPhys(
-+ UINT_TO_PTR(p_FmPcdCcTree->ccTreeBaseAddr))
-+ - p_FmPcd->physicalMuramBase);
-+
-+ return err;
-+}
-+
-+t_Error FmPcdCcUnbindTree(t_Handle h_FmPcd, t_Handle h_FmPcdCcTree)
-+{
-+ t_FmPcdCcTree *p_FmPcdCcTree = (t_FmPcdCcTree *)h_FmPcdCcTree;
-+
-+ /* this routine must be protected by the calling routine by locking all PCD modules! */
-+
-+ UNUSED(h_FmPcd);
-+
-+ SANITY_CHECK_RETURN_ERROR(h_FmPcdCcTree, E_INVALID_HANDLE);
-+
-+ UpdateCcRootOwner(p_FmPcdCcTree, FALSE);
-+
-+ return E_OK;
-+}
-+
-+t_Error FmPcdCcNodeTreeTryLock(t_Handle h_FmPcd, t_Handle h_FmPcdCcNode,
-+ t_List *p_List)
-+{
-+ t_FmPcdCcNode *p_CcNode = (t_FmPcdCcNode *)h_FmPcdCcNode;
-+ t_List *p_Pos, *p_Tmp;
-+ t_CcNodeInformation *p_CcNodeInfo, nodeInfo;
-+ uint32_t intFlags;
-+ t_Error err = E_OK;
-+
-+ intFlags = FmPcdLock(h_FmPcd);
-+
-+ LIST_FOR_EACH(p_Pos, &p_CcNode->ccTreesLst)
-+ {
-+ p_CcNodeInfo = CC_NODE_F_OBJECT(p_Pos);
-+ ASSERT_COND(p_CcNodeInfo->h_CcNode);
-+
-+ err = CcRootTryLock(p_CcNodeInfo->h_CcNode);
-+
-+ if (err)
-+ {
-+ LIST_FOR_EACH(p_Tmp, &p_CcNode->ccTreesLst)
-+ {
-+ if (p_Tmp == p_Pos)
-+ break;
-+
-+ CcRootReleaseLock(p_CcNodeInfo->h_CcNode);
-+ }
-+ break;
-+ }
-+
-+ memset(&nodeInfo, 0, sizeof(t_CcNodeInformation));
-+ nodeInfo.h_CcNode = p_CcNodeInfo->h_CcNode;
-+ EnqueueNodeInfoToRelevantLst(p_List, &nodeInfo, NULL);
-+ }
-+
-+ FmPcdUnlock(h_FmPcd, intFlags);
-+ CORE_MemoryBarrier();
-+
-+ return err;
-+}
-+
-+void FmPcdCcNodeTreeReleaseLock(t_Handle h_FmPcd, t_List *p_List)
-+{
-+ t_List *p_Pos;
-+ t_CcNodeInformation *p_CcNodeInfo;
-+ t_Handle h_FmPcdCcTree;
-+ uint32_t intFlags;
-+
-+ intFlags = FmPcdLock(h_FmPcd);
-+
-+ LIST_FOR_EACH(p_Pos, p_List)
-+ {
-+ p_CcNodeInfo = CC_NODE_F_OBJECT(p_Pos);
-+ h_FmPcdCcTree = p_CcNodeInfo->h_CcNode;
-+ CcRootReleaseLock(h_FmPcdCcTree);
-+ }
-+
-+ ReleaseLst(p_List);
-+
-+ FmPcdUnlock(h_FmPcd, intFlags);
-+ CORE_MemoryBarrier();
-+}
-+
-+t_Error FmPcdUpdateCcShadow(t_FmPcd *p_FmPcd, uint32_t size, uint32_t align)
-+{
-+ uint32_t intFlags;
-+ uint32_t newSize = 0, newAlign = 0;
-+ bool allocFail = FALSE;
-+
-+ ASSERT_COND(p_FmPcd);
-+
-+ if (!size)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("size must be larger then 0"));
-+
-+ if (!POWER_OF_2(align))
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("alignment must be power of 2"));
-+
-+ newSize = p_FmPcd->ccShadowSize;
-+ newAlign = p_FmPcd->ccShadowAlign;
-+
-+ /* Check if current shadow is large enough to hold the requested size */
-+ if (size > p_FmPcd->ccShadowSize)
-+ newSize = size;
-+
-+ /* Check if current shadow matches the requested alignment */
-+ if (align > p_FmPcd->ccShadowAlign)
-+ newAlign = align;
-+
-+ /* If a bigger shadow size or bigger shadow alignment are required,
-+ a new shadow will be allocated */
-+ if ((newSize != p_FmPcd->ccShadowSize)
-+ || (newAlign != p_FmPcd->ccShadowAlign))
-+ {
-+ intFlags = FmPcdLock(p_FmPcd);
-+
-+ if (p_FmPcd->p_CcShadow)
-+ {
-+ FM_MURAM_FreeMem(FmPcdGetMuramHandle(p_FmPcd), p_FmPcd->p_CcShadow);
-+ p_FmPcd->ccShadowSize = 0;
-+ p_FmPcd->ccShadowAlign = 0;
-+ }
-+
-+ p_FmPcd->p_CcShadow = FM_MURAM_AllocMem(FmPcdGetMuramHandle(p_FmPcd),
-+ newSize, newAlign);
-+ if (!p_FmPcd->p_CcShadow)
-+ {
-+ allocFail = TRUE;
-+
-+ /* If new shadow size allocation failed,
-+ re-allocate with previous parameters */
-+ p_FmPcd->p_CcShadow = FM_MURAM_AllocMem(
-+ FmPcdGetMuramHandle(p_FmPcd), p_FmPcd->ccShadowSize,
-+ p_FmPcd->ccShadowAlign);
-+ }
-+
-+ FmPcdUnlock(p_FmPcd, intFlags);
-+
-+ if (allocFail)
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY,
-+ ("MURAM allocation for CC Shadow memory"));
-+
-+ p_FmPcd->ccShadowSize = newSize;
-+ p_FmPcd->ccShadowAlign = newAlign;
-+ }
-+
-+ return E_OK;
-+}
-+
-+#if (DPAA_VERSION >= 11)
-+void FmPcdCcGetAdTablesThatPointOnReplicGroup(t_Handle h_Node,
-+ t_Handle h_ReplicGroup,
-+ t_List *p_AdTables,
-+ uint32_t *p_NumOfAdTables)
-+{
-+ t_FmPcdCcNode *p_CurrentNode = (t_FmPcdCcNode *)h_Node;
-+ int i = 0;
-+ void * p_AdTable;
-+ t_CcNodeInformation ccNodeInfo;
-+
-+ ASSERT_COND(h_Node);
-+ *p_NumOfAdTables = 0;
-+
-+ /* search in the current node which exact index points on this current replicator group for getting AD */
-+ for (i = 0; i < p_CurrentNode->numOfKeys + 1; i++)
-+ {
-+ if ((p_CurrentNode->keyAndNextEngineParams[i].nextEngineParams.nextEngine
-+ == e_FM_PCD_FR)
-+ && ((p_CurrentNode->keyAndNextEngineParams[i].nextEngineParams.params.frParams.h_FrmReplic
-+ == (t_Handle)h_ReplicGroup)))
-+ {
-+ /* save the current ad table in the list */
-+ /* this entry uses the input replicator group */
-+ p_AdTable =
-+ PTR_MOVE(p_CurrentNode->h_AdTable, i*FM_PCD_CC_AD_ENTRY_SIZE);
-+ memset(&ccNodeInfo, 0, sizeof(t_CcNodeInformation));
-+ ccNodeInfo.h_CcNode = p_AdTable;
-+ EnqueueNodeInfoToRelevantLst(p_AdTables, &ccNodeInfo, NULL);
-+ (*p_NumOfAdTables)++;
-+ }
-+ }
-+
-+ ASSERT_COND(i != p_CurrentNode->numOfKeys);
-+}
-+#endif /* (DPAA_VERSION >= 11) */
-+/*********************** End of inter-module routines ************************/
-+
-+/****************************************/
-+/* API Init unit functions */
-+/****************************************/
-+
-+t_Handle FM_PCD_CcRootBuild(t_Handle h_FmPcd,
-+ t_FmPcdCcTreeParams *p_PcdGroupsParam)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd *)h_FmPcd;
-+ t_Error err = E_OK;
-+ int i = 0, j = 0, k = 0;
-+ t_FmPcdCcTree *p_FmPcdCcTree;
-+ uint8_t numOfEntries;
-+ t_Handle p_CcTreeTmp;
-+ t_FmPcdCcGrpParams *p_FmPcdCcGroupParams;
-+ t_FmPcdCcKeyAndNextEngineParams *p_Params, *p_KeyAndNextEngineParams;
-+ t_NetEnvParams netEnvParams;
-+ uint8_t lastOne = 0;
-+ uint32_t requiredAction = 0;
-+ t_FmPcdCcNode *p_FmPcdCcNextNode;
-+ t_CcNodeInformation ccNodeInfo, *p_CcInformation;
-+
-+ SANITY_CHECK_RETURN_VALUE(h_FmPcd, E_INVALID_HANDLE, NULL);
-+ SANITY_CHECK_RETURN_VALUE(p_PcdGroupsParam, E_INVALID_HANDLE, NULL);
-+
-+ if (p_PcdGroupsParam->numOfGrps > FM_PCD_MAX_NUM_OF_CC_GROUPS)
-+ {
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE, ("numOfGrps should not exceed %d", FM_PCD_MAX_NUM_OF_CC_GROUPS));
-+ return NULL;
-+ }
-+
-+ p_FmPcdCcTree = (t_FmPcdCcTree*)XX_Malloc(sizeof(t_FmPcdCcTree));
-+ if (!p_FmPcdCcTree)
-+ {
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("PCD tree structure"));
-+ return NULL;
-+ }
-+ memset(p_FmPcdCcTree, 0, sizeof(t_FmPcdCcTree));
-+ p_FmPcdCcTree->h_FmPcd = h_FmPcd;
-+
-+ p_Params = (t_FmPcdCcKeyAndNextEngineParams*)XX_Malloc(
-+ FM_PCD_MAX_NUM_OF_CC_GROUPS
-+ * sizeof(t_FmPcdCcKeyAndNextEngineParams));
-+ memset(p_Params,
-+ 0,
-+ FM_PCD_MAX_NUM_OF_CC_GROUPS
-+ * sizeof(t_FmPcdCcKeyAndNextEngineParams));
-+
-+ INIT_LIST(&p_FmPcdCcTree->fmPortsLst);
-+
-+#ifdef FM_CAPWAP_SUPPORT
-+ if ((p_PcdGroupsParam->numOfGrps == 1) &&
-+ (p_PcdGroupsParam->ccGrpParams[0].numOfDistinctionUnits == 0) &&
-+ (p_PcdGroupsParam->ccGrpParams[0].nextEnginePerEntriesInGrp[0].nextEngine == e_FM_PCD_CC) &&
-+ p_PcdGroupsParam->ccGrpParams[0].nextEnginePerEntriesInGrp[0].params.ccParams.h_CcNode &&
-+ IsCapwapApplSpecific(p_PcdGroupsParam->ccGrpParams[0].nextEnginePerEntriesInGrp[0].params.ccParams.h_CcNode))
-+ {
-+ p_PcdGroupsParam->ccGrpParams[0].nextEnginePerEntriesInGrp[0].h_Manip = FmPcdManipApplSpecificBuild();
-+ if (!p_PcdGroupsParam->ccGrpParams[0].nextEnginePerEntriesInGrp[0].h_Manip)
-+ {
-+ DeleteTree(p_FmPcdCcTree,p_FmPcd);
-+ XX_Free(p_Params);
-+ REPORT_ERROR(MAJOR, E_INVALID_STATE, NO_MSG);
-+ return NULL;
-+ }
-+ }
-+#endif /* FM_CAPWAP_SUPPORT */
-+
-+ numOfEntries = 0;
-+ p_FmPcdCcTree->netEnvId = FmPcdGetNetEnvId(p_PcdGroupsParam->h_NetEnv);
-+
-+ for (i = 0; i < p_PcdGroupsParam->numOfGrps; i++)
-+ {
-+ p_FmPcdCcGroupParams = &p_PcdGroupsParam->ccGrpParams[i];
-+
-+ if (p_FmPcdCcGroupParams->numOfDistinctionUnits
-+ > FM_PCD_MAX_NUM_OF_CC_UNITS)
-+ {
-+ DeleteTree(p_FmPcdCcTree, p_FmPcd);
-+ XX_Free(p_Params);
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE,
-+ ("numOfDistinctionUnits (group %d) should not exceed %d", i, FM_PCD_MAX_NUM_OF_CC_UNITS));
-+ return NULL;
-+ }
-+
-+ p_FmPcdCcTree->fmPcdGroupParam[i].baseGroupEntry = numOfEntries;
-+ p_FmPcdCcTree->fmPcdGroupParam[i].numOfEntriesInGroup = (uint8_t)(0x01
-+ << p_FmPcdCcGroupParams->numOfDistinctionUnits);
-+ numOfEntries += p_FmPcdCcTree->fmPcdGroupParam[i].numOfEntriesInGroup;
-+ if (numOfEntries > FM_PCD_MAX_NUM_OF_CC_GROUPS)
-+ {
-+ DeleteTree(p_FmPcdCcTree, p_FmPcd);
-+ XX_Free(p_Params);
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE, ("numOfEntries can not be larger than %d", FM_PCD_MAX_NUM_OF_CC_GROUPS));
-+ return NULL;
-+ }
-+
-+ if (lastOne)
-+ {
-+ if (p_FmPcdCcTree->fmPcdGroupParam[i].numOfEntriesInGroup > lastOne)
-+ {
-+ DeleteTree(p_FmPcdCcTree, p_FmPcd);
-+ XX_Free(p_Params);
-+ REPORT_ERROR(MAJOR, E_CONFLICT, ("numOfEntries per group must be set in descending order"));
-+ return NULL;
-+ }
-+ }
-+
-+ lastOne = p_FmPcdCcTree->fmPcdGroupParam[i].numOfEntriesInGroup;
-+
-+ netEnvParams.netEnvId = p_FmPcdCcTree->netEnvId;
-+ netEnvParams.numOfDistinctionUnits =
-+ p_FmPcdCcGroupParams->numOfDistinctionUnits;
-+
-+ memcpy(netEnvParams.unitIds, &p_FmPcdCcGroupParams->unitIds,
-+ (sizeof(uint8_t)) * p_FmPcdCcGroupParams->numOfDistinctionUnits);
-+
-+ err = PcdGetUnitsVector(p_FmPcd, &netEnvParams);
-+ if (err)
-+ {
-+ DeleteTree(p_FmPcdCcTree, p_FmPcd);
-+ XX_Free(p_Params);
-+ REPORT_ERROR(MAJOR, err, NO_MSG);
-+ return NULL;
-+ }
-+
-+ p_FmPcdCcTree->fmPcdGroupParam[i].totalBitsMask = netEnvParams.vector;
-+ for (j = 0; j < p_FmPcdCcTree->fmPcdGroupParam[i].numOfEntriesInGroup;
-+ j++)
-+ {
-+ err = ValidateNextEngineParams(
-+ h_FmPcd,
-+ &p_FmPcdCcGroupParams->nextEnginePerEntriesInGrp[j],
-+ e_FM_PCD_CC_STATS_MODE_NONE);
-+ if (err)
-+ {
-+ DeleteTree(p_FmPcdCcTree, p_FmPcd);
-+ XX_Free(p_Params);
-+ REPORT_ERROR(MAJOR, err, (NO_MSG));
-+ return NULL;
-+ }
-+
-+ if (p_FmPcdCcGroupParams->nextEnginePerEntriesInGrp[j].h_Manip)
-+ {
-+ err = FmPcdManipCheckParamsForCcNextEngine(
-+ &p_FmPcdCcGroupParams->nextEnginePerEntriesInGrp[j],
-+ &requiredAction);
-+ if (err)
-+ {
-+ DeleteTree(p_FmPcdCcTree, p_FmPcd);
-+ XX_Free(p_Params);
-+ REPORT_ERROR(MAJOR, E_INVALID_STATE, NO_MSG);
-+ return NULL;
-+ }
-+ }
-+ p_KeyAndNextEngineParams = p_Params + k;
-+
-+ memcpy(&p_KeyAndNextEngineParams->nextEngineParams,
-+ &p_FmPcdCcGroupParams->nextEnginePerEntriesInGrp[j],
-+ sizeof(t_FmPcdCcNextEngineParams));
-+
-+ if ((p_KeyAndNextEngineParams->nextEngineParams.nextEngine
-+ == e_FM_PCD_CC)
-+ && p_KeyAndNextEngineParams->nextEngineParams.h_Manip)
-+ {
-+ err =
-+ AllocAndFillAdForContLookupManip(
-+ p_KeyAndNextEngineParams->nextEngineParams.params.ccParams.h_CcNode);
-+ if (err)
-+ {
-+ DeleteTree(p_FmPcdCcTree, p_FmPcd);
-+ XX_Free(p_Params);
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("MURAM allocation for CC Tree"));
-+ return NULL;
-+ }
-+ }
-+
-+ requiredAction |= UPDATE_CC_WITH_TREE;
-+ p_KeyAndNextEngineParams->requiredAction = requiredAction;
-+
-+ k++;
-+ }
-+ }
-+
-+ p_FmPcdCcTree->numOfEntries = (uint8_t)k;
-+ p_FmPcdCcTree->numOfGrps = p_PcdGroupsParam->numOfGrps;
-+
-+ p_FmPcdCcTree->ccTreeBaseAddr =
-+ PTR_TO_UINT(FM_MURAM_AllocMem(FmPcdGetMuramHandle(h_FmPcd),
-+ (uint32_t)( FM_PCD_MAX_NUM_OF_CC_GROUPS * FM_PCD_CC_AD_ENTRY_SIZE),
-+ FM_PCD_CC_TREE_ADDR_ALIGN));
-+ if (!p_FmPcdCcTree->ccTreeBaseAddr)
-+ {
-+ DeleteTree(p_FmPcdCcTree, p_FmPcd);
-+ XX_Free(p_Params);
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("MURAM allocation for CC Tree"));
-+ return NULL;
-+ }
-+ MemSet8(
-+ UINT_TO_PTR(p_FmPcdCcTree->ccTreeBaseAddr), 0,
-+ (uint32_t)(FM_PCD_MAX_NUM_OF_CC_GROUPS * FM_PCD_CC_AD_ENTRY_SIZE));
-+
-+ p_CcTreeTmp = UINT_TO_PTR(p_FmPcdCcTree->ccTreeBaseAddr);
-+
-+ for (i = 0; i < numOfEntries; i++)
-+ {
-+ p_KeyAndNextEngineParams = p_Params + i;
-+
-+ NextStepAd(p_CcTreeTmp, NULL,
-+ &p_KeyAndNextEngineParams->nextEngineParams, p_FmPcd);
-+
-+ p_CcTreeTmp = PTR_MOVE(p_CcTreeTmp, FM_PCD_CC_AD_ENTRY_SIZE);
-+
-+ memcpy(&p_FmPcdCcTree->keyAndNextEngineParams[i],
-+ p_KeyAndNextEngineParams,
-+ sizeof(t_FmPcdCcKeyAndNextEngineParams));
-+
-+ if (p_FmPcdCcTree->keyAndNextEngineParams[i].nextEngineParams.nextEngine
-+ == e_FM_PCD_CC)
-+ {
-+ p_FmPcdCcNextNode =
-+ (t_FmPcdCcNode*)p_FmPcdCcTree->keyAndNextEngineParams[i].nextEngineParams.params.ccParams.h_CcNode;
-+ p_CcInformation = FindNodeInfoInReleventLst(
-+ &p_FmPcdCcNextNode->ccTreeIdLst, (t_Handle)p_FmPcdCcTree,
-+ p_FmPcdCcNextNode->h_Spinlock);
-+
-+ if (!p_CcInformation)
-+ {
-+ memset(&ccNodeInfo, 0, sizeof(t_CcNodeInformation));
-+ ccNodeInfo.h_CcNode = (t_Handle)p_FmPcdCcTree;
-+ ccNodeInfo.index = 1;
-+ EnqueueNodeInfoToRelevantLst(&p_FmPcdCcNextNode->ccTreeIdLst,
-+ &ccNodeInfo,
-+ p_FmPcdCcNextNode->h_Spinlock);
-+ }
-+ else
-+ p_CcInformation->index++;
-+ }
-+ }
-+
-+ FmPcdIncNetEnvOwners(h_FmPcd, p_FmPcdCcTree->netEnvId);
-+ p_CcTreeTmp = UINT_TO_PTR(p_FmPcdCcTree->ccTreeBaseAddr);
-+
-+ if (!FmPcdLockTryLockAll(p_FmPcd))
-+ {
-+ FM_PCD_CcRootDelete(p_FmPcdCcTree);
-+ XX_Free(p_Params);
-+ DBG(TRACE, ("FmPcdLockTryLockAll failed"));
-+ return NULL;
-+ }
-+
-+ for (i = 0; i < numOfEntries; i++)
-+ {
-+ if (p_FmPcdCcTree->keyAndNextEngineParams[i].requiredAction)
-+ {
-+ err = SetRequiredAction(
-+ h_FmPcd,
-+ p_FmPcdCcTree->keyAndNextEngineParams[i].requiredAction,
-+ &p_FmPcdCcTree->keyAndNextEngineParams[i], p_CcTreeTmp, 1,
-+ p_FmPcdCcTree);
-+ if (err)
-+ {
-+ FmPcdLockUnlockAll(p_FmPcd);
-+ FM_PCD_CcRootDelete(p_FmPcdCcTree);
-+ XX_Free(p_Params);
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("No memory"));
-+ return NULL;
-+ }
-+ p_CcTreeTmp = PTR_MOVE(p_CcTreeTmp, FM_PCD_CC_AD_ENTRY_SIZE);
-+ }
-+ }
-+
-+ FmPcdLockUnlockAll(p_FmPcd);
-+ p_FmPcdCcTree->p_Lock = FmPcdAcquireLock(p_FmPcd);
-+ if (!p_FmPcdCcTree->p_Lock)
-+ {
-+ FM_PCD_CcRootDelete(p_FmPcdCcTree);
-+ XX_Free(p_Params);
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("FM CC lock"));
-+ return NULL;
-+ }
-+
-+ XX_Free(p_Params);
-+
-+ return p_FmPcdCcTree;
-+}
-+
-+t_Error FM_PCD_CcRootDelete(t_Handle h_CcTree)
-+{
-+ t_FmPcd *p_FmPcd;
-+ t_FmPcdCcTree *p_CcTree = (t_FmPcdCcTree *)h_CcTree;
-+ int i = 0;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_CcTree, E_INVALID_STATE);
-+ p_FmPcd = (t_FmPcd *)p_CcTree->h_FmPcd;
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd, E_INVALID_HANDLE);
-+
-+ FmPcdDecNetEnvOwners(p_FmPcd, p_CcTree->netEnvId);
-+
-+ if (p_CcTree->owners)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_SELECTION,
-+ ("the tree with this ID can not be removed because this tree is occupied, first - unbind this tree"));
-+
-+ /* Delete ip-reassembly schemes if exist */
-+ if (p_CcTree->h_IpReassemblyManip)
-+ {
-+ FmPcdManipDeleteIpReassmSchemes(p_CcTree->h_IpReassemblyManip);
-+ FmPcdManipUpdateOwner(p_CcTree->h_IpReassemblyManip, FALSE);
-+ }
-+
-+ /* Delete capwap-reassembly schemes if exist */
-+ if (p_CcTree->h_CapwapReassemblyManip)
-+ {
-+ FmPcdManipDeleteCapwapReassmSchemes(p_CcTree->h_CapwapReassemblyManip);
-+ FmPcdManipUpdateOwner(p_CcTree->h_CapwapReassemblyManip, FALSE);
-+ }
-+
-+ for (i = 0; i < p_CcTree->numOfEntries; i++)
-+ {
-+ if (p_CcTree->keyAndNextEngineParams[i].nextEngineParams.nextEngine
-+ == e_FM_PCD_CC)
-+ UpdateNodeOwner(
-+ p_CcTree->keyAndNextEngineParams[i].nextEngineParams.params.ccParams.h_CcNode,
-+ FALSE);
-+
-+ if (p_CcTree->keyAndNextEngineParams[i].nextEngineParams.h_Manip)
-+ FmPcdManipUpdateOwner(
-+ p_CcTree->keyAndNextEngineParams[i].nextEngineParams.h_Manip,
-+ FALSE);
-+
-+#ifdef FM_CAPWAP_SUPPORT
-+ if ((p_CcTree->numOfGrps == 1) &&
-+ (p_CcTree->fmPcdGroupParam[0].numOfEntriesInGroup == 1) &&
-+ (p_CcTree->keyAndNextEngineParams[0].nextEngineParams.nextEngine == e_FM_PCD_CC) &&
-+ p_CcTree->keyAndNextEngineParams[0].nextEngineParams.params.ccParams.h_CcNode &&
-+ IsCapwapApplSpecific(p_CcTree->keyAndNextEngineParams[0].nextEngineParams.params.ccParams.h_CcNode))
-+ {
-+ if (FM_PCD_ManipNodeDelete(p_CcTree->keyAndNextEngineParams[0].nextEngineParams.h_Manip) != E_OK)
-+ return E_INVALID_STATE;
-+ }
-+#endif /* FM_CAPWAP_SUPPORT */
-+
-+#if (DPAA_VERSION >= 11)
-+ if ((p_CcTree->keyAndNextEngineParams[i].nextEngineParams.nextEngine
-+ == e_FM_PCD_FR)
-+ && (p_CcTree->keyAndNextEngineParams[i].nextEngineParams.params.frParams.h_FrmReplic))
-+ FrmReplicGroupUpdateOwner(
-+ p_CcTree->keyAndNextEngineParams[i].nextEngineParams.params.frParams.h_FrmReplic,
-+ FALSE);
-+#endif /* (DPAA_VERSION >= 11) */
-+ }
-+
-+ if (p_CcTree->p_Lock)
-+ FmPcdReleaseLock(p_CcTree->h_FmPcd, p_CcTree->p_Lock);
-+
-+ DeleteTree(p_CcTree, p_FmPcd);
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PCD_CcRootModifyNextEngine(
-+ t_Handle h_CcTree, uint8_t grpId, uint8_t index,
-+ t_FmPcdCcNextEngineParams *p_FmPcdCcNextEngineParams)
-+{
-+ t_FmPcd *p_FmPcd;
-+ t_FmPcdCcTree *p_CcTree = (t_FmPcdCcTree *)h_CcTree;
-+ t_Error err = E_OK;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcdCcNextEngineParams, E_NULL_POINTER);
-+ SANITY_CHECK_RETURN_ERROR(p_CcTree, E_INVALID_STATE);
-+ p_FmPcd = (t_FmPcd *)p_CcTree->h_FmPcd;
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd, E_INVALID_HANDLE);
-+
-+ if (!FmPcdLockTryLockAll(p_FmPcd))
-+ {
-+ DBG(TRACE, ("FmPcdLockTryLockAll failed"));
-+ return ERROR_CODE(E_BUSY);
-+ }
-+
-+ err = FmPcdCcModifyNextEngineParamTree(p_FmPcd, p_CcTree, grpId, index,
-+ p_FmPcdCcNextEngineParams);
-+ FmPcdLockUnlockAll(p_FmPcd);
-+
-+ if (err)
-+ {
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+
-+ return E_OK;
-+}
-+
-+t_Handle FM_PCD_MatchTableSet(t_Handle h_FmPcd,
-+ t_FmPcdCcNodeParams *p_CcNodeParam)
-+{
-+ t_FmPcdCcNode *p_CcNode;
-+ t_Error err;
-+
-+ SANITY_CHECK_RETURN_VALUE(h_FmPcd, E_INVALID_HANDLE, NULL);
-+ SANITY_CHECK_RETURN_VALUE(p_CcNodeParam, E_NULL_POINTER, NULL);
-+
-+ p_CcNode = (t_FmPcdCcNode*)XX_Malloc(sizeof(t_FmPcdCcNode));
-+ if (!p_CcNode)
-+ {
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("No memory"));
-+ return NULL;
-+ }
-+ memset(p_CcNode, 0, sizeof(t_FmPcdCcNode));
-+
-+ err = MatchTableSet(h_FmPcd, p_CcNode, p_CcNodeParam);
-+
-+ switch(GET_ERROR_TYPE(err)
-+) {
-+ case E_OK:
-+ break;
-+
-+ case E_BUSY:
-+ DBG(TRACE, ("E_BUSY error"));
-+ return NULL;
-+
-+ default:
-+ REPORT_ERROR(MAJOR, err, NO_MSG);
-+ return NULL;
-+ }
-+
-+ return p_CcNode;
-+}
-+
-+t_Error FM_PCD_MatchTableDelete(t_Handle h_CcNode)
-+{
-+ t_FmPcdCcNode *p_CcNode = (t_FmPcdCcNode *)h_CcNode;
-+ int i = 0;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_CcNode, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_CcNode->h_FmPcd, E_INVALID_HANDLE);
-+
-+ if (p_CcNode->owners)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_STATE,
-+ ("This node cannot be removed because it is occupied; first unbind this node"));
-+
-+ for (i = 0; i < p_CcNode->numOfKeys; i++)
-+ if (p_CcNode->keyAndNextEngineParams[i].nextEngineParams.nextEngine
-+ == e_FM_PCD_CC)
-+ UpdateNodeOwner(
-+ p_CcNode->keyAndNextEngineParams[i].nextEngineParams.params.ccParams.h_CcNode,
-+ FALSE);
-+
-+ if (p_CcNode->keyAndNextEngineParams[i].nextEngineParams.nextEngine
-+ == e_FM_PCD_CC)
-+ UpdateNodeOwner(
-+ p_CcNode->keyAndNextEngineParams[i].nextEngineParams.params.ccParams.h_CcNode,
-+ FALSE);
-+
-+ /* Handle also Miss entry */
-+ for (i = 0; i < p_CcNode->numOfKeys + 1; i++)
-+ {
-+ if (p_CcNode->keyAndNextEngineParams[i].nextEngineParams.h_Manip)
-+ FmPcdManipUpdateOwner(
-+ p_CcNode->keyAndNextEngineParams[i].nextEngineParams.h_Manip,
-+ FALSE);
-+
-+#if (DPAA_VERSION >= 11)
-+ if ((p_CcNode->keyAndNextEngineParams[i].nextEngineParams.nextEngine
-+ == e_FM_PCD_FR)
-+ && (p_CcNode->keyAndNextEngineParams[i].nextEngineParams.params.frParams.h_FrmReplic))
-+ {
-+ FrmReplicGroupUpdateOwner(
-+ p_CcNode->keyAndNextEngineParams[i].nextEngineParams.params.frParams.h_FrmReplic,
-+ FALSE);
-+ }
-+#endif /* (DPAA_VERSION >= 11) */
-+ }
-+
-+ DeleteNode(p_CcNode);
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PCD_MatchTableAddKey(t_Handle h_CcNode, uint16_t keyIndex,
-+ uint8_t keySize,
-+ t_FmPcdCcKeyParams *p_KeyParams)
-+{
-+ t_FmPcd *p_FmPcd;
-+ t_FmPcdCcNode *p_CcNode = (t_FmPcdCcNode *)h_CcNode;
-+ t_Error err = E_OK;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_KeyParams, E_NULL_POINTER);
-+ SANITY_CHECK_RETURN_ERROR(p_CcNode, E_INVALID_HANDLE);
-+ p_FmPcd = (t_FmPcd *)p_CcNode->h_FmPcd;
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd->h_Hc, E_INVALID_HANDLE);
-+
-+ if (keyIndex == FM_PCD_LAST_KEY_INDEX)
-+ keyIndex = p_CcNode->numOfKeys;
-+
-+ if (!FmPcdLockTryLockAll(p_FmPcd))
-+ {
-+ DBG(TRACE, ("FmPcdLockTryLockAll failed"));
-+ return ERROR_CODE(E_BUSY);
-+ }
-+
-+ err = FmPcdCcAddKey(p_FmPcd, p_CcNode, keyIndex, keySize, p_KeyParams);
-+
-+ FmPcdLockUnlockAll(p_FmPcd);
-+
-+ switch(GET_ERROR_TYPE(err)
-+) {
-+ case E_OK:
-+ return E_OK;
-+
-+ case E_BUSY:
-+ DBG(TRACE, ("E_BUSY error"));
-+ return ERROR_CODE(E_BUSY);
-+
-+ default:
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+}
-+
-+t_Error FM_PCD_MatchTableRemoveKey(t_Handle h_CcNode, uint16_t keyIndex)
-+{
-+ t_FmPcd *p_FmPcd;
-+ t_FmPcdCcNode *p_CcNode = (t_FmPcdCcNode *)h_CcNode;
-+ t_Error err = E_OK;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_CcNode, E_INVALID_HANDLE);
-+ p_FmPcd = (t_FmPcd *)p_CcNode->h_FmPcd;
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd->h_Hc, E_INVALID_HANDLE);
-+
-+ if (!FmPcdLockTryLockAll(p_FmPcd))
-+ {
-+ DBG(TRACE, ("FmPcdLockTryLockAll failed"));
-+ return ERROR_CODE(E_BUSY);
-+ }
-+
-+ err = FmPcdCcRemoveKey(p_FmPcd, p_CcNode, keyIndex);
-+
-+ FmPcdLockUnlockAll(p_FmPcd);
-+
-+ switch(GET_ERROR_TYPE(err)
-+) {
-+ case E_OK:
-+ return E_OK;
-+
-+ case E_BUSY:
-+ DBG(TRACE, ("E_BUSY error"));
-+ return ERROR_CODE(E_BUSY);
-+
-+ default:
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PCD_MatchTableModifyKey(t_Handle h_CcNode, uint16_t keyIndex,
-+ uint8_t keySize, uint8_t *p_Key,
-+ uint8_t *p_Mask)
-+{
-+ t_FmPcd *p_FmPcd;
-+ t_FmPcdCcNode *p_CcNode = (t_FmPcdCcNode *)h_CcNode;
-+ t_Error err = E_OK;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_CcNode, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Key, E_NULL_POINTER);
-+ p_FmPcd = (t_FmPcd *)p_CcNode->h_FmPcd;
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd->h_Hc, E_INVALID_HANDLE);
-+
-+
-+ if (!FmPcdLockTryLockAll(p_FmPcd))
-+ {
-+ DBG(TRACE, ("FmPcdLockTryLockAll failed"));
-+ return ERROR_CODE(E_BUSY);
-+ }
-+
-+ err = FmPcdCcModifyKey(p_FmPcd, p_CcNode, keyIndex, keySize, p_Key, p_Mask);
-+
-+ FmPcdLockUnlockAll(p_FmPcd);
-+
-+ switch(GET_ERROR_TYPE(err)
-+) {
-+ case E_OK:
-+ return E_OK;
-+
-+ case E_BUSY:
-+ DBG(TRACE, ("E_BUSY error"));
-+ return ERROR_CODE(E_BUSY);
-+
-+ default:
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+}
-+
-+t_Error FM_PCD_MatchTableModifyNextEngine(
-+ t_Handle h_CcNode, uint16_t keyIndex,
-+ t_FmPcdCcNextEngineParams *p_FmPcdCcNextEngineParams)
-+{
-+ t_FmPcd *p_FmPcd;
-+ t_FmPcdCcNode *p_CcNode = (t_FmPcdCcNode *)h_CcNode;
-+ t_Error err = E_OK;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcdCcNextEngineParams, E_NULL_POINTER);
-+ SANITY_CHECK_RETURN_ERROR(p_CcNode, E_INVALID_HANDLE);
-+ p_FmPcd = (t_FmPcd *)p_CcNode->h_FmPcd;
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd->h_Hc, E_INVALID_HANDLE);
-+
-+ if (!FmPcdLockTryLockAll(p_FmPcd))
-+ {
-+ DBG(TRACE, ("FmPcdLockTryLockAll failed"));
-+ return ERROR_CODE(E_BUSY);
-+ }
-+
-+ err = ModifyNextEngineParamNode(p_FmPcd, p_CcNode, keyIndex,
-+ p_FmPcdCcNextEngineParams);
-+
-+ FmPcdLockUnlockAll(p_FmPcd);
-+
-+ switch(GET_ERROR_TYPE(err)
-+) {
-+ case E_OK:
-+ return E_OK;
-+
-+ case E_BUSY:
-+ DBG(TRACE, ("E_BUSY error"));
-+ return ERROR_CODE(E_BUSY);
-+
-+ default:
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+}
-+
-+t_Error FM_PCD_MatchTableModifyMissNextEngine(
-+ t_Handle h_CcNode, t_FmPcdCcNextEngineParams *p_FmPcdCcNextEngineParams)
-+{
-+ t_FmPcd *p_FmPcd;
-+ t_FmPcdCcNode *p_CcNode = (t_FmPcdCcNode *)h_CcNode;
-+ t_Error err = E_OK;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcdCcNextEngineParams, E_NULL_POINTER);
-+ SANITY_CHECK_RETURN_ERROR(p_CcNode, E_INVALID_HANDLE);
-+ p_FmPcd = (t_FmPcd *)p_CcNode->h_FmPcd;
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd->h_Hc, E_INVALID_HANDLE);
-+
-+ if (!FmPcdLockTryLockAll(p_FmPcd))
-+ {
-+ DBG(TRACE, ("FmPcdLockTryLockAll failed"));
-+ return ERROR_CODE(E_BUSY);
-+ }
-+
-+ err = FmPcdCcModifyMissNextEngineParamNode(p_FmPcd, p_CcNode,
-+ p_FmPcdCcNextEngineParams);
-+
-+ FmPcdLockUnlockAll(p_FmPcd);
-+
-+ switch(GET_ERROR_TYPE(err)
-+) {
-+ case E_OK:
-+ return E_OK;
-+
-+ case E_BUSY:
-+ DBG(TRACE, ("E_BUSY error"));
-+ return ERROR_CODE(E_BUSY);
-+
-+ default:
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+}
-+
-+t_Error FM_PCD_MatchTableModifyKeyAndNextEngine(t_Handle h_CcNode,
-+ uint16_t keyIndex,
-+ uint8_t keySize,
-+ t_FmPcdCcKeyParams *p_KeyParams)
-+{
-+ t_FmPcd *p_FmPcd;
-+ t_FmPcdCcNode *p_CcNode = (t_FmPcdCcNode *)h_CcNode;
-+ t_Error err = E_OK;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_KeyParams, E_NULL_POINTER);
-+ SANITY_CHECK_RETURN_ERROR(p_CcNode, E_INVALID_HANDLE);
-+ p_FmPcd = (t_FmPcd *)p_CcNode->h_FmPcd;
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd->h_Hc, E_INVALID_HANDLE);
-+
-+ if (!FmPcdLockTryLockAll(p_FmPcd))
-+ {
-+ DBG(TRACE, ("FmPcdLockTryLockAll failed"));
-+ return ERROR_CODE(E_BUSY);
-+ }
-+
-+ err = FmPcdCcModifyKeyAndNextEngine(p_FmPcd, p_CcNode, keyIndex, keySize,
-+ p_KeyParams);
-+
-+ FmPcdLockUnlockAll(p_FmPcd);
-+
-+ switch(GET_ERROR_TYPE(err)
-+) {
-+ case E_OK:
-+ return E_OK;
-+
-+ case E_BUSY:
-+ DBG(TRACE, ("E_BUSY error"));
-+ return ERROR_CODE(E_BUSY);
-+
-+ default:
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+}
-+
-+t_Error FM_PCD_MatchTableFindNRemoveKey(t_Handle h_CcNode, uint8_t keySize,
-+ uint8_t *p_Key, uint8_t *p_Mask)
-+{
-+ t_FmPcd *p_FmPcd;
-+ t_FmPcdCcNode *p_CcNode = (t_FmPcdCcNode *)h_CcNode;
-+ uint16_t keyIndex;
-+ t_Error err;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Key, E_NULL_POINTER);
-+ SANITY_CHECK_RETURN_ERROR(p_CcNode, E_INVALID_HANDLE);
-+ p_FmPcd = (t_FmPcd *)p_CcNode->h_FmPcd;
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd->h_Hc, E_INVALID_HANDLE);
-+
-+ if (!FmPcdLockTryLockAll(p_FmPcd))
-+ {
-+ DBG(TRACE, ("FmPcdLockTryLockAll failed"));
-+ return ERROR_CODE(E_BUSY);
-+ }
-+
-+ err = FindKeyIndex(p_CcNode, keySize, p_Key, p_Mask, &keyIndex);
-+ if (GET_ERROR_TYPE(err) != E_OK)
-+ {
-+ FmPcdLockUnlockAll(p_FmPcd);
-+ RETURN_ERROR(
-+ MAJOR,
-+ err,
-+ ("The received key and mask pair was not found in the match table of the provided node"));
-+ }
-+
-+ err = FmPcdCcRemoveKey(p_FmPcd, p_CcNode, keyIndex);
-+
-+ FmPcdLockUnlockAll(p_FmPcd);
-+
-+ switch(GET_ERROR_TYPE(err)
-+) {
-+ case E_OK:
-+ return E_OK;
-+
-+ case E_BUSY:
-+ DBG(TRACE, ("E_BUSY error"));
-+ return ERROR_CODE(E_BUSY);
-+
-+ default:
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+}
-+
-+t_Error FM_PCD_MatchTableFindNModifyNextEngine(
-+ t_Handle h_CcNode, uint8_t keySize, uint8_t *p_Key, uint8_t *p_Mask,
-+ t_FmPcdCcNextEngineParams *p_FmPcdCcNextEngineParams)
-+{
-+ t_FmPcd *p_FmPcd;
-+ t_FmPcdCcNode *p_CcNode = (t_FmPcdCcNode *)h_CcNode;
-+ uint16_t keyIndex;
-+ t_Error err;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Key, E_NULL_POINTER);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcdCcNextEngineParams, E_NULL_POINTER);
-+ SANITY_CHECK_RETURN_ERROR(p_CcNode, E_INVALID_HANDLE);
-+ p_FmPcd = (t_FmPcd *)p_CcNode->h_FmPcd;
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd->h_Hc, E_INVALID_HANDLE);
-+
-+ if (!FmPcdLockTryLockAll(p_FmPcd))
-+ {
-+ DBG(TRACE, ("FmPcdLockTryLockAll failed"));
-+ return ERROR_CODE(E_BUSY);
-+ }
-+
-+ err = FindKeyIndex(p_CcNode, keySize, p_Key, p_Mask, &keyIndex);
-+ if (GET_ERROR_TYPE(err) != E_OK)
-+ {
-+ FmPcdLockUnlockAll(p_FmPcd);
-+ RETURN_ERROR(
-+ MAJOR,
-+ err,
-+ ("The received key and mask pair was not found in the match table of the provided node"));
-+ }
-+
-+ err = ModifyNextEngineParamNode(p_FmPcd, p_CcNode, keyIndex,
-+ p_FmPcdCcNextEngineParams);
-+
-+ FmPcdLockUnlockAll(p_FmPcd);
-+
-+ switch(GET_ERROR_TYPE(err)
-+) {
-+ case E_OK:
-+ return E_OK;
-+
-+ case E_BUSY:
-+ DBG(TRACE, ("E_BUSY error"));
-+ return ERROR_CODE(E_BUSY);
-+
-+ default:
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+}
-+
-+t_Error FM_PCD_MatchTableFindNModifyKeyAndNextEngine(
-+ t_Handle h_CcNode, uint8_t keySize, uint8_t *p_Key, uint8_t *p_Mask,
-+ t_FmPcdCcKeyParams *p_KeyParams)
-+{
-+ t_FmPcd *p_FmPcd;
-+ t_FmPcdCcNode *p_CcNode = (t_FmPcdCcNode *)h_CcNode;
-+ uint16_t keyIndex;
-+ t_Error err;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Key, E_NULL_POINTER);
-+ SANITY_CHECK_RETURN_ERROR(p_KeyParams, E_NULL_POINTER);
-+ SANITY_CHECK_RETURN_ERROR(p_CcNode, E_INVALID_HANDLE);
-+ p_FmPcd = (t_FmPcd *)p_CcNode->h_FmPcd;
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd->h_Hc, E_INVALID_HANDLE);
-+
-+ if (!FmPcdLockTryLockAll(p_FmPcd))
-+ {
-+ DBG(TRACE, ("FmPcdLockTryLockAll failed"));
-+ return ERROR_CODE(E_BUSY);
-+ }
-+
-+ err = FindKeyIndex(p_CcNode, keySize, p_Key, p_Mask, &keyIndex);
-+ if (GET_ERROR_TYPE(err) != E_OK)
-+ {
-+ FmPcdLockUnlockAll(p_FmPcd);
-+ RETURN_ERROR(
-+ MAJOR,
-+ err,
-+ ("The received key and mask pair was not found in the match table of the provided node"));
-+ }
-+
-+ err = FmPcdCcModifyKeyAndNextEngine(p_FmPcd, h_CcNode, keyIndex, keySize,
-+ p_KeyParams);
-+
-+ FmPcdLockUnlockAll(p_FmPcd);
-+
-+ switch(GET_ERROR_TYPE(err)
-+) {
-+ case E_OK:
-+ return E_OK;
-+
-+ case E_BUSY:
-+ DBG(TRACE, ("E_BUSY error"));
-+ return ERROR_CODE(E_BUSY);
-+
-+ default:
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+}
-+
-+t_Error FM_PCD_MatchTableFindNModifyKey(t_Handle h_CcNode, uint8_t keySize,
-+ uint8_t *p_Key, uint8_t *p_Mask,
-+ uint8_t *p_NewKey, uint8_t *p_NewMask)
-+{
-+ t_FmPcd *p_FmPcd;
-+ t_FmPcdCcNode *p_CcNode = (t_FmPcdCcNode *)h_CcNode;
-+ t_List h_List;
-+ uint16_t keyIndex;
-+ t_Error err;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Key, E_NULL_POINTER);
-+ SANITY_CHECK_RETURN_ERROR(p_NewKey, E_NULL_POINTER);
-+ SANITY_CHECK_RETURN_ERROR(p_CcNode, E_INVALID_HANDLE);
-+ p_FmPcd = (t_FmPcd *)p_CcNode->h_FmPcd;
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd->h_Hc, E_INVALID_HANDLE);
-+
-+ INIT_LIST(&h_List);
-+
-+ err = FmPcdCcNodeTreeTryLock(p_FmPcd, p_CcNode, &h_List);
-+ if (err)
-+ {
-+ DBG(TRACE, ("Node's trees lock failed"));
-+ return ERROR_CODE(E_BUSY);
-+ }
-+
-+ err = FindKeyIndex(p_CcNode, keySize, p_Key, p_Mask, &keyIndex);
-+ if (GET_ERROR_TYPE(err) != E_OK)
-+ {
-+ FmPcdCcNodeTreeReleaseLock(p_FmPcd, &h_List);
-+ RETURN_ERROR(MAJOR, err,
-+ ("The received key and mask pair was not found in the "
-+ "match table of the provided node"));
-+ }
-+
-+ err = FmPcdCcModifyKey(p_FmPcd, p_CcNode, keyIndex, keySize, p_NewKey,
-+ p_NewMask);
-+
-+ FmPcdCcNodeTreeReleaseLock(p_FmPcd, &h_List);
-+
-+ switch(GET_ERROR_TYPE(err)
-+) {
-+ case E_OK:
-+ return E_OK;
-+
-+ case E_BUSY:
-+ DBG(TRACE, ("E_BUSY error"));
-+ return ERROR_CODE(E_BUSY);
-+
-+ default:
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+}
-+
-+t_Error FM_PCD_MatchTableGetNextEngine(
-+ t_Handle h_CcNode, uint16_t keyIndex,
-+ t_FmPcdCcNextEngineParams *p_FmPcdCcNextEngineParams)
-+{
-+ t_FmPcdCcNode *p_CcNode = (t_FmPcdCcNode *)h_CcNode;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_CcNode, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcdCcNextEngineParams, E_NULL_POINTER);
-+
-+ if (keyIndex >= p_CcNode->numOfKeys)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE,
-+ ("keyIndex exceeds current number of keys"));
-+
-+ if (keyIndex > (FM_PCD_MAX_NUM_OF_KEYS - 1))
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("keyIndex can not be larger than %d", (FM_PCD_MAX_NUM_OF_KEYS - 1)));
-+
-+ memcpy(p_FmPcdCcNextEngineParams,
-+ &p_CcNode->keyAndNextEngineParams[keyIndex].nextEngineParams,
-+ sizeof(t_FmPcdCcNextEngineParams));
-+
-+ return E_OK;
-+}
-+
-+
-+uint32_t FM_PCD_MatchTableGetKeyCounter(t_Handle h_CcNode, uint16_t keyIndex)
-+{
-+ t_FmPcdCcNode *p_CcNode = (t_FmPcdCcNode *)h_CcNode;
-+ uint32_t *p_StatsCounters, frameCount;
-+ uint32_t intFlags;
-+
-+ SANITY_CHECK_RETURN_VALUE(p_CcNode, E_INVALID_HANDLE, 0);
-+
-+ if (p_CcNode->statisticsMode == e_FM_PCD_CC_STATS_MODE_NONE)
-+ {
-+ REPORT_ERROR(MAJOR, E_INVALID_STATE, ("Statistics were not enabled for this match table"));
-+ return 0;
-+ }
-+
-+ if ((p_CcNode->statisticsMode != e_FM_PCD_CC_STATS_MODE_FRAME)
-+ && (p_CcNode->statisticsMode
-+ != e_FM_PCD_CC_STATS_MODE_BYTE_AND_FRAME))
-+ {
-+ REPORT_ERROR(MAJOR, E_INVALID_STATE, ("Frame count is not supported in the statistics mode of this match table"));
-+ return 0;
-+ }
-+
-+ intFlags = XX_LockIntrSpinlock(p_CcNode->h_Spinlock);
-+
-+ if (keyIndex >= p_CcNode->numOfKeys)
-+ {
-+ XX_UnlockIntrSpinlock(p_CcNode->h_Spinlock, intFlags);
-+ REPORT_ERROR(MAJOR, E_INVALID_STATE, ("The provided keyIndex exceeds the number of keys in this match table"));
-+ return 0;
-+ }
-+
-+ if (!p_CcNode->keyAndNextEngineParams[keyIndex].p_StatsObj)
-+ {
-+ XX_UnlockIntrSpinlock(p_CcNode->h_Spinlock, intFlags);
-+ REPORT_ERROR(MAJOR, E_INVALID_STATE, ("Statistics were not enabled for this key"));
-+ return 0;
-+ }
-+
-+ p_StatsCounters =
-+ p_CcNode->keyAndNextEngineParams[keyIndex].p_StatsObj->h_StatsCounters;
-+ ASSERT_COND(p_StatsCounters);
-+
-+ /* The first counter is byte counter, so we need to advance to the next counter */
-+ frameCount = GET_UINT32(*(uint32_t *)(PTR_MOVE(p_StatsCounters,
-+ FM_PCD_CC_STATS_COUNTER_SIZE)));
-+
-+ XX_UnlockIntrSpinlock(p_CcNode->h_Spinlock, intFlags);
-+
-+ return frameCount;
-+}
-+
-+t_Error FM_PCD_MatchTableGetKeyStatistics(
-+ t_Handle h_CcNode, uint16_t keyIndex,
-+ t_FmPcdCcKeyStatistics *p_KeyStatistics)
-+{
-+ t_FmPcdCcNode *p_CcNode = (t_FmPcdCcNode *)h_CcNode;
-+ uint32_t intFlags;
-+ t_Error err;
-+
-+ SANITY_CHECK_RETURN_ERROR(h_CcNode, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_KeyStatistics, E_NULL_POINTER);
-+
-+ intFlags = XX_LockIntrSpinlock(p_CcNode->h_Spinlock);
-+
-+ if (keyIndex >= p_CcNode->numOfKeys)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_STATE,
-+ ("The provided keyIndex exceeds the number of keys in this match table"));
-+
-+ err = MatchTableGetKeyStatistics(p_CcNode, keyIndex, p_KeyStatistics);
-+
-+ XX_UnlockIntrSpinlock(p_CcNode->h_Spinlock, intFlags);
-+
-+ if (err != E_OK)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PCD_MatchTableGetMissStatistics(
-+ t_Handle h_CcNode, t_FmPcdCcKeyStatistics *p_MissStatistics)
-+{
-+ t_FmPcdCcNode *p_CcNode = (t_FmPcdCcNode *)h_CcNode;
-+ uint32_t intFlags;
-+ t_Error err;
-+
-+ SANITY_CHECK_RETURN_ERROR(h_CcNode, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_MissStatistics, E_NULL_POINTER);
-+
-+ intFlags = XX_LockIntrSpinlock(p_CcNode->h_Spinlock);
-+
-+ err = MatchTableGetKeyStatistics(p_CcNode, p_CcNode->numOfKeys,
-+ p_MissStatistics);
-+
-+ XX_UnlockIntrSpinlock(p_CcNode->h_Spinlock, intFlags);
-+
-+ if (err != E_OK)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PCD_MatchTableFindNGetKeyStatistics(
-+ t_Handle h_CcNode, uint8_t keySize, uint8_t *p_Key, uint8_t *p_Mask,
-+ t_FmPcdCcKeyStatistics *p_KeyStatistics)
-+{
-+ t_FmPcdCcNode *p_CcNode = (t_FmPcdCcNode *)h_CcNode;
-+ uint16_t keyIndex;
-+ uint32_t intFlags;
-+ t_Error err;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Key, E_NULL_POINTER);
-+ SANITY_CHECK_RETURN_ERROR(p_KeyStatistics, E_NULL_POINTER);
-+
-+ intFlags = XX_LockIntrSpinlock(p_CcNode->h_Spinlock);
-+
-+ err = FindKeyIndex(p_CcNode, keySize, p_Key, p_Mask, &keyIndex);
-+ if (GET_ERROR_TYPE(err) != E_OK)
-+ {
-+ XX_UnlockIntrSpinlock(p_CcNode->h_Spinlock, intFlags);
-+ RETURN_ERROR(MAJOR, err,
-+ ("The received key and mask pair was not found in the "
-+ "match table of the provided node"));
-+ }
-+
-+ ASSERT_COND(keyIndex < p_CcNode->numOfKeys);
-+
-+ err = MatchTableGetKeyStatistics(p_CcNode, keyIndex, p_KeyStatistics);
-+
-+ XX_UnlockIntrSpinlock(p_CcNode->h_Spinlock, intFlags);
-+
-+ if (err != E_OK)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PCD_MatchTableGetIndexedHashBucket(t_Handle h_CcNode,
-+ uint8_t keySize, uint8_t *p_Key,
-+ uint8_t hashShift,
-+ t_Handle *p_CcNodeBucketHandle,
-+ uint8_t *p_BucketIndex,
-+ uint16_t *p_LastIndex)
-+{
-+ t_FmPcdCcNode *p_CcNode = (t_FmPcdCcNode *)h_CcNode;
-+ uint16_t glblMask;
-+ uint64_t crc64 = 0;
-+
-+ SANITY_CHECK_RETURN_ERROR(h_CcNode, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(
-+ p_CcNode->parseCode == CC_PC_GENERIC_IC_HASH_INDEXED,
-+ E_INVALID_STATE);
-+ SANITY_CHECK_RETURN_ERROR(p_Key, E_NULL_POINTER);
-+ SANITY_CHECK_RETURN_ERROR(p_CcNodeBucketHandle, E_NULL_POINTER);
-+
-+ memcpy(&glblMask, PTR_MOVE(p_CcNode->p_GlblMask, 2), 2);
-+ be16_to_cpus(&glblMask);
-+
-+ crc64 = crc64_init();
-+ crc64 = crc64_compute(p_Key, keySize, crc64);
-+ crc64 >>= hashShift;
-+
-+ *p_BucketIndex = (uint8_t)(((crc64 >> (8 * (6 - p_CcNode->userOffset)))
-+ & glblMask) >> 4);
-+ if (*p_BucketIndex >= p_CcNode->numOfKeys)
-+ RETURN_ERROR(MINOR, E_NOT_IN_RANGE, ("bucket index!"));
-+
-+ *p_CcNodeBucketHandle =
-+ p_CcNode->keyAndNextEngineParams[*p_BucketIndex].nextEngineParams.params.ccParams.h_CcNode;
-+ if (!*p_CcNodeBucketHandle)
-+ RETURN_ERROR(MINOR, E_NOT_FOUND, ("bucket!"));
-+
-+ *p_LastIndex = ((t_FmPcdCcNode *)*p_CcNodeBucketHandle)->numOfKeys;
-+
-+ return E_OK;
-+}
-+
-+t_Handle FM_PCD_HashTableSet(t_Handle h_FmPcd, t_FmPcdHashTableParams *p_Param)
-+{
-+ t_FmPcdCcNode *p_CcNodeHashTbl;
-+ t_FmPcdCcNodeParams *p_IndxHashCcNodeParam, *p_ExactMatchCcNodeParam;
-+ t_FmPcdCcNode *p_CcNode;
-+ t_Handle h_MissStatsCounters = NULL;
-+ t_FmPcdCcKeyParams *p_HashKeyParams;
-+ int i;
-+ uint16_t numOfSets, numOfWays, countMask, onesCount = 0;
-+ bool statsEnForMiss = FALSE;
-+ t_Error err;
-+
-+ SANITY_CHECK_RETURN_VALUE(h_FmPcd, E_INVALID_HANDLE, NULL);
-+ SANITY_CHECK_RETURN_VALUE(p_Param, E_NULL_POINTER, NULL);
-+
-+ if (p_Param->maxNumOfKeys == 0)
-+ {
-+ REPORT_ERROR(MINOR, E_INVALID_VALUE, ("Max number of keys must be higher then 0"));
-+ return NULL;
-+ }
-+
-+ if (p_Param->hashResMask == 0)
-+ {
-+ REPORT_ERROR(MINOR, E_INVALID_VALUE, ("Hash result mask must differ from 0"));
-+ return NULL;
-+ }
-+
-+ /*Fix: QorIQ SDK / QSDK-2131*/
-+ if (p_Param->ccNextEngineParamsForMiss.nextEngine == e_FM_PCD_INVALID)
-+ {
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE, ("Next PCD Engine for on-miss entry is invalid. On-miss entry is always required. You can use e_FM_PCD_DONE."));
-+ return NULL;
-+ }
-+
-+#if (DPAA_VERSION >= 11)
-+ if (p_Param->statisticsMode == e_FM_PCD_CC_STATS_MODE_RMON)
-+ {
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE,
-+ ("RMON statistics mode is not supported for hash table"));
-+ return NULL;
-+ }
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+ p_ExactMatchCcNodeParam = (t_FmPcdCcNodeParams*)XX_Malloc(
-+ sizeof(t_FmPcdCcNodeParams));
-+ if (!p_ExactMatchCcNodeParam)
-+ {
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("p_ExactMatchCcNodeParam"));
-+ return NULL;
-+ }
-+ memset(p_ExactMatchCcNodeParam, 0, sizeof(t_FmPcdCcNodeParams));
-+
-+ p_IndxHashCcNodeParam = (t_FmPcdCcNodeParams*)XX_Malloc(
-+ sizeof(t_FmPcdCcNodeParams));
-+ if (!p_IndxHashCcNodeParam)
-+ {
-+ XX_Free(p_ExactMatchCcNodeParam);
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("p_IndxHashCcNodeParam"));
-+ return NULL;
-+ }
-+ memset(p_IndxHashCcNodeParam, 0, sizeof(t_FmPcdCcNodeParams));
-+
-+ /* Calculate number of sets and number of ways of the hash table */
-+ countMask = (uint16_t)(p_Param->hashResMask >> 4);
-+ while (countMask)
-+ {
-+ onesCount++;
-+ countMask = (uint16_t)(countMask >> 1);
-+ }
-+
-+ numOfSets = (uint16_t)(1 << onesCount);
-+ numOfWays = (uint16_t)DIV_CEIL(p_Param->maxNumOfKeys, numOfSets);
-+
-+ if (p_Param->maxNumOfKeys % numOfSets)
-+ DBG(INFO, ("'maxNumOfKeys' is not a multiple of hash number of ways, so number of ways will be rounded up"));
-+
-+ if ((p_Param->statisticsMode == e_FM_PCD_CC_STATS_MODE_FRAME)
-+ || (p_Param->statisticsMode == e_FM_PCD_CC_STATS_MODE_BYTE_AND_FRAME))
-+ {
-+ /* Allocating a statistics counters table that will be used by all
-+ 'miss' entries of the hash table */
-+ h_MissStatsCounters = (t_Handle)FM_MURAM_AllocMem(
-+ FmPcdGetMuramHandle(h_FmPcd), 2 * FM_PCD_CC_STATS_COUNTER_SIZE,
-+ FM_PCD_CC_AD_TABLE_ALIGN);
-+ if (!h_MissStatsCounters)
-+ {
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("MURAM allocation for statistics table for hash miss"));
-+ XX_Free(p_IndxHashCcNodeParam);
-+ XX_Free(p_ExactMatchCcNodeParam);
-+ return NULL;
-+ }
-+ memset(h_MissStatsCounters, 0, (2 * FM_PCD_CC_STATS_COUNTER_SIZE));
-+
-+ /* Always enable statistics for 'miss', so that a statistics AD will be
-+ initialized from the start. We'll store the requested 'statistics enable'
-+ value and it will be used when statistics are read by the user. */
-+ statsEnForMiss = p_Param->ccNextEngineParamsForMiss.statisticsEn;
-+ p_Param->ccNextEngineParamsForMiss.statisticsEn = TRUE;
-+ }
-+
-+ /* Building exact-match node params, will be used to create the hash buckets */
-+ p_ExactMatchCcNodeParam->extractCcParams.type = e_FM_PCD_EXTRACT_NON_HDR;
-+
-+ p_ExactMatchCcNodeParam->extractCcParams.extractNonHdr.src =
-+ e_FM_PCD_EXTRACT_FROM_KEY;
-+ p_ExactMatchCcNodeParam->extractCcParams.extractNonHdr.action =
-+ e_FM_PCD_ACTION_EXACT_MATCH;
-+ p_ExactMatchCcNodeParam->extractCcParams.extractNonHdr.offset = 0;
-+ p_ExactMatchCcNodeParam->extractCcParams.extractNonHdr.size =
-+ p_Param->matchKeySize;
-+
-+ p_ExactMatchCcNodeParam->keysParams.maxNumOfKeys = numOfWays;
-+ p_ExactMatchCcNodeParam->keysParams.maskSupport = FALSE;
-+ p_ExactMatchCcNodeParam->keysParams.statisticsMode =
-+ p_Param->statisticsMode;
-+ p_ExactMatchCcNodeParam->keysParams.numOfKeys = 0;
-+ p_ExactMatchCcNodeParam->keysParams.keySize = p_Param->matchKeySize;
-+ p_ExactMatchCcNodeParam->keysParams.ccNextEngineParamsForMiss =
-+ p_Param->ccNextEngineParamsForMiss;
-+
-+ p_HashKeyParams = p_IndxHashCcNodeParam->keysParams.keyParams;
-+
-+ for (i = 0; i < numOfSets; i++)
-+ {
-+ /* Each exact-match node will be marked as a 'bucket' and provided with
-+ a pointer to statistics counters, to be used for 'miss' entry
-+ statistics */
-+ p_CcNode = (t_FmPcdCcNode *)XX_Malloc(sizeof(t_FmPcdCcNode));
-+ if (!p_CcNode)
-+ break;
-+ memset(p_CcNode, 0, sizeof(t_FmPcdCcNode));
-+
-+ p_CcNode->isHashBucket = TRUE;
-+ p_CcNode->h_MissStatsCounters = h_MissStatsCounters;
-+
-+ err = MatchTableSet(h_FmPcd, p_CcNode, p_ExactMatchCcNodeParam);
-+ if (err)
-+ break;
-+
-+ p_HashKeyParams[i].ccNextEngineParams.nextEngine = e_FM_PCD_CC;
-+ p_HashKeyParams[i].ccNextEngineParams.statisticsEn = FALSE;
-+ p_HashKeyParams[i].ccNextEngineParams.params.ccParams.h_CcNode =
-+ p_CcNode;
-+ }
-+
-+ if (i < numOfSets)
-+ {
-+ for (i = i - 1; i >= 0; i--)
-+ FM_PCD_MatchTableDelete(
-+ p_HashKeyParams[i].ccNextEngineParams.params.ccParams.h_CcNode);
-+
-+ FM_MURAM_FreeMem(FmPcdGetMuramHandle(h_FmPcd), h_MissStatsCounters);
-+
-+ REPORT_ERROR(MAJOR, E_NULL_POINTER, NO_MSG);
-+ XX_Free(p_IndxHashCcNodeParam);
-+ XX_Free(p_ExactMatchCcNodeParam);
-+ return NULL;
-+ }
-+
-+ /* Creating indexed-hash CC node */
-+ p_IndxHashCcNodeParam->extractCcParams.type = e_FM_PCD_EXTRACT_NON_HDR;
-+ p_IndxHashCcNodeParam->extractCcParams.extractNonHdr.src =
-+ e_FM_PCD_EXTRACT_FROM_HASH;
-+ p_IndxHashCcNodeParam->extractCcParams.extractNonHdr.action =
-+ e_FM_PCD_ACTION_INDEXED_LOOKUP;
-+ p_IndxHashCcNodeParam->extractCcParams.extractNonHdr.icIndxMask =
-+ p_Param->hashResMask;
-+ p_IndxHashCcNodeParam->extractCcParams.extractNonHdr.offset =
-+ p_Param->hashShift;
-+ p_IndxHashCcNodeParam->extractCcParams.extractNonHdr.size = 2;
-+
-+ p_IndxHashCcNodeParam->keysParams.maxNumOfKeys = numOfSets;
-+ p_IndxHashCcNodeParam->keysParams.maskSupport = FALSE;
-+ p_IndxHashCcNodeParam->keysParams.statisticsMode =
-+ e_FM_PCD_CC_STATS_MODE_NONE;
-+ /* Number of keys of this node is number of sets of the hash */
-+ p_IndxHashCcNodeParam->keysParams.numOfKeys = numOfSets;
-+ p_IndxHashCcNodeParam->keysParams.keySize = 2;
-+
-+ p_CcNodeHashTbl = FM_PCD_MatchTableSet(h_FmPcd, p_IndxHashCcNodeParam);
-+
-+ if (p_CcNodeHashTbl)
-+ {
-+ p_CcNodeHashTbl->kgHashShift = p_Param->kgHashShift;
-+
-+ /* Storing the allocated counters for buckets 'miss' in the hash table
-+ and if statistics for miss were enabled. */
-+ p_CcNodeHashTbl->h_MissStatsCounters = h_MissStatsCounters;
-+ p_CcNodeHashTbl->statsEnForMiss = statsEnForMiss;
-+ }
-+
-+ XX_Free(p_IndxHashCcNodeParam);
-+ XX_Free(p_ExactMatchCcNodeParam);
-+
-+ return p_CcNodeHashTbl;
-+}
-+
-+t_Error FM_PCD_HashTableDelete(t_Handle h_HashTbl)
-+{
-+ t_FmPcdCcNode *p_HashTbl = (t_FmPcdCcNode *)h_HashTbl;
-+ t_Handle h_FmPcd;
-+ t_Handle *p_HashBuckets, h_MissStatsCounters;
-+ uint16_t i, numOfBuckets;
-+ t_Error err;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_HashTbl, E_INVALID_HANDLE);
-+
-+ /* Store all hash buckets before the hash is freed */
-+ numOfBuckets = p_HashTbl->numOfKeys;
-+
-+ p_HashBuckets = (t_Handle *)XX_Malloc(numOfBuckets * sizeof(t_Handle));
-+ if (!p_HashBuckets)
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, NO_MSG);
-+
-+ for (i = 0; i < numOfBuckets; i++)
-+ p_HashBuckets[i] =
-+ p_HashTbl->keyAndNextEngineParams[i].nextEngineParams.params.ccParams.h_CcNode;
-+
-+ h_FmPcd = p_HashTbl->h_FmPcd;
-+ h_MissStatsCounters = p_HashTbl->h_MissStatsCounters;
-+
-+ /* Free the hash */
-+ err = FM_PCD_MatchTableDelete(p_HashTbl);
-+
-+ /* Free each hash bucket */
-+ for (i = 0; i < numOfBuckets; i++)
-+ err |= FM_PCD_MatchTableDelete(p_HashBuckets[i]);
-+
-+ XX_Free(p_HashBuckets);
-+
-+ /* Free statistics counters for 'miss', if these were allocated */
-+ if (h_MissStatsCounters)
-+ FM_MURAM_FreeMem(FmPcdGetMuramHandle(h_FmPcd), h_MissStatsCounters);
-+
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PCD_HashTableAddKey(t_Handle h_HashTbl, uint8_t keySize,
-+ t_FmPcdCcKeyParams *p_KeyParams)
-+{
-+ t_FmPcdCcNode *p_HashTbl = (t_FmPcdCcNode *)h_HashTbl;
-+ t_Handle h_HashBucket;
-+ uint8_t bucketIndex;
-+ uint16_t lastIndex;
-+ t_Error err;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_HashTbl, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_KeyParams, E_NULL_POINTER);
-+ SANITY_CHECK_RETURN_ERROR(p_KeyParams->p_Key, E_NULL_POINTER);
-+
-+ if (p_KeyParams->p_Mask)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE,
-+ ("Keys masks not supported for hash table"));
-+
-+ err = FM_PCD_MatchTableGetIndexedHashBucket(p_HashTbl, keySize,
-+ p_KeyParams->p_Key,
-+ p_HashTbl->kgHashShift,
-+ &h_HashBucket, &bucketIndex,
-+ &lastIndex);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+
-+ return FM_PCD_MatchTableAddKey(h_HashBucket, FM_PCD_LAST_KEY_INDEX, keySize,
-+ p_KeyParams);
-+}
-+
-+t_Error FM_PCD_HashTableRemoveKey(t_Handle h_HashTbl, uint8_t keySize,
-+ uint8_t *p_Key)
-+{
-+ t_FmPcdCcNode *p_HashTbl = (t_FmPcdCcNode *)h_HashTbl;
-+ t_Handle h_HashBucket;
-+ uint8_t bucketIndex;
-+ uint16_t lastIndex;
-+ t_Error err;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_HashTbl, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Key, E_NULL_POINTER);
-+
-+ err = FM_PCD_MatchTableGetIndexedHashBucket(p_HashTbl, keySize, p_Key,
-+ p_HashTbl->kgHashShift,
-+ &h_HashBucket, &bucketIndex,
-+ &lastIndex);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+
-+ return FM_PCD_MatchTableFindNRemoveKey(h_HashBucket, keySize, p_Key, NULL);
-+}
-+
-+t_Error FM_PCD_HashTableModifyNextEngine(
-+ t_Handle h_HashTbl, uint8_t keySize, uint8_t *p_Key,
-+ t_FmPcdCcNextEngineParams *p_FmPcdCcNextEngineParams)
-+{
-+ t_FmPcdCcNode *p_HashTbl = (t_FmPcdCcNode *)h_HashTbl;
-+ t_Handle h_HashBucket;
-+ uint8_t bucketIndex;
-+ uint16_t lastIndex;
-+ t_Error err;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_HashTbl, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Key, E_NULL_POINTER);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcdCcNextEngineParams, E_NULL_POINTER);
-+
-+ err = FM_PCD_MatchTableGetIndexedHashBucket(p_HashTbl, keySize, p_Key,
-+ p_HashTbl->kgHashShift,
-+ &h_HashBucket, &bucketIndex,
-+ &lastIndex);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+
-+ return FM_PCD_MatchTableFindNModifyNextEngine(h_HashBucket, keySize, p_Key,
-+ NULL,
-+ p_FmPcdCcNextEngineParams);
-+}
-+
-+t_Error FM_PCD_HashTableModifyMissNextEngine(
-+ t_Handle h_HashTbl,
-+ t_FmPcdCcNextEngineParams *p_FmPcdCcNextEngineParams)
-+{
-+ t_FmPcdCcNode *p_HashTbl = (t_FmPcdCcNode *)h_HashTbl;
-+ t_Handle h_HashBucket;
-+ uint8_t i;
-+ bool nullifyMissStats = FALSE;
-+ t_Error err;
-+
-+ SANITY_CHECK_RETURN_ERROR(h_HashTbl, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcdCcNextEngineParams, E_NULL_POINTER);
-+
-+ if ((!p_HashTbl->h_MissStatsCounters)
-+ && (p_FmPcdCcNextEngineParams->statisticsEn))
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_CONFLICT,
-+ ("Statistics are requested for a key, but statistics mode was set"
-+ "to 'NONE' upon initialization"));
-+
-+ if (p_HashTbl->h_MissStatsCounters)
-+ {
-+ if ((!p_HashTbl->statsEnForMiss)
-+ && (p_FmPcdCcNextEngineParams->statisticsEn))
-+ nullifyMissStats = TRUE;
-+
-+ if ((p_HashTbl->statsEnForMiss)
-+ && (!p_FmPcdCcNextEngineParams->statisticsEn))
-+ {
-+ p_HashTbl->statsEnForMiss = FALSE;
-+ p_FmPcdCcNextEngineParams->statisticsEn = TRUE;
-+ }
-+ }
-+
-+ for (i = 0; i < p_HashTbl->numOfKeys; i++)
-+ {
-+ h_HashBucket =
-+ p_HashTbl->keyAndNextEngineParams[i].nextEngineParams.params.ccParams.h_CcNode;
-+
-+ err = FM_PCD_MatchTableModifyMissNextEngine(h_HashBucket,
-+ p_FmPcdCcNextEngineParams);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+
-+ if (nullifyMissStats)
-+ {
-+ memset(p_HashTbl->h_MissStatsCounters, 0,
-+ (2 * FM_PCD_CC_STATS_COUNTER_SIZE));
-+ memset(p_HashTbl->h_MissStatsCounters, 0,
-+ (2 * FM_PCD_CC_STATS_COUNTER_SIZE));
-+ p_HashTbl->statsEnForMiss = TRUE;
-+ }
-+
-+ return E_OK;
-+}
-+
-+
-+t_Error FM_PCD_HashTableGetMissNextEngine(
-+ t_Handle h_HashTbl,
-+ t_FmPcdCcNextEngineParams *p_FmPcdCcNextEngineParams)
-+{
-+ t_FmPcdCcNode *p_HashTbl = (t_FmPcdCcNode *)h_HashTbl;
-+ t_FmPcdCcNode *p_HashBucket;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_HashTbl, E_INVALID_HANDLE);
-+
-+ /* Miss next engine of each bucket was initialized with the next engine of the hash table */
-+ p_HashBucket =
-+ p_HashTbl->keyAndNextEngineParams[0].nextEngineParams.params.ccParams.h_CcNode;
-+
-+ memcpy(p_FmPcdCcNextEngineParams,
-+ &p_HashBucket->keyAndNextEngineParams[p_HashBucket->numOfKeys].nextEngineParams,
-+ sizeof(t_FmPcdCcNextEngineParams));
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PCD_HashTableFindNGetKeyStatistics(
-+ t_Handle h_HashTbl, uint8_t keySize, uint8_t *p_Key,
-+ t_FmPcdCcKeyStatistics *p_KeyStatistics)
-+{
-+ t_FmPcdCcNode *p_HashTbl = (t_FmPcdCcNode *)h_HashTbl;
-+ t_Handle h_HashBucket;
-+ uint8_t bucketIndex;
-+ uint16_t lastIndex;
-+ t_Error err;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_HashTbl, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Key, E_NULL_POINTER);
-+ SANITY_CHECK_RETURN_ERROR(p_KeyStatistics, E_NULL_POINTER);
-+
-+ err = FM_PCD_MatchTableGetIndexedHashBucket(p_HashTbl, keySize, p_Key,
-+ p_HashTbl->kgHashShift,
-+ &h_HashBucket, &bucketIndex,
-+ &lastIndex);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+
-+ return FM_PCD_MatchTableFindNGetKeyStatistics(h_HashBucket, keySize, p_Key,
-+ NULL, p_KeyStatistics);
-+}
-+
-+t_Error FM_PCD_HashTableGetMissStatistics(
-+ t_Handle h_HashTbl, t_FmPcdCcKeyStatistics *p_MissStatistics)
-+{
-+ t_FmPcdCcNode *p_HashTbl = (t_FmPcdCcNode *)h_HashTbl;
-+ t_Handle h_HashBucket;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_HashTbl, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_MissStatistics, E_NULL_POINTER);
-+
-+ if (!p_HashTbl->statsEnForMiss)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE,
-+ ("Statistics were not enabled for miss"));
-+
-+ h_HashBucket =
-+ p_HashTbl->keyAndNextEngineParams[0].nextEngineParams.params.ccParams.h_CcNode;
-+
-+ return FM_PCD_MatchTableGetMissStatistics(h_HashBucket, p_MissStatistics);
-+}
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/Pcd/fm_cc.h
-@@ -0,0 +1,399 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/******************************************************************************
-+ @File fm_cc.h
-+
-+ @Description FM PCD CC ...
-+*//***************************************************************************/
-+#ifndef __FM_CC_H
-+#define __FM_CC_H
-+
-+#include "std_ext.h"
-+#include "error_ext.h"
-+#include "list_ext.h"
-+
-+#include "fm_pcd.h"
-+
-+
-+/***********************************************************************/
-+/* Coarse classification defines */
-+/***********************************************************************/
-+
-+#define CC_MAX_NUM_OF_KEYS (FM_PCD_MAX_NUM_OF_KEYS + 1)
-+
-+#define CC_PC_FF_MACDST 0x00
-+#define CC_PC_FF_MACSRC 0x01
-+#define CC_PC_FF_ETYPE 0x02
-+
-+#define CC_PC_FF_TCI1 0x03
-+#define CC_PC_FF_TCI2 0x04
-+
-+#define CC_PC_FF_MPLS1 0x06
-+#define CC_PC_FF_MPLS_LAST 0x07
-+
-+#define CC_PC_FF_IPV4DST1 0x08
-+#define CC_PC_FF_IPV4DST2 0x16
-+#define CC_PC_FF_IPV4IPTOS_TC1 0x09
-+#define CC_PC_FF_IPV4IPTOS_TC2 0x17
-+#define CC_PC_FF_IPV4PTYPE1 0x0A
-+#define CC_PC_FF_IPV4PTYPE2 0x18
-+#define CC_PC_FF_IPV4SRC1 0x0b
-+#define CC_PC_FF_IPV4SRC2 0x19
-+#define CC_PC_FF_IPV4SRC1_IPV4DST1 0x0c
-+#define CC_PC_FF_IPV4SRC2_IPV4DST2 0x1a
-+#define CC_PC_FF_IPV4TTL 0x29
-+
-+
-+#define CC_PC_FF_IPTOS_IPV6TC1_IPV6FLOW1 0x0d /*TODO - CLASS - what is it? TOS*/
-+#define CC_PC_FF_IPTOS_IPV6TC2_IPV6FLOW2 0x1b
-+#define CC_PC_FF_IPV6PTYPE1 0x0e
-+#define CC_PC_FF_IPV6PTYPE2 0x1c
-+#define CC_PC_FF_IPV6DST1 0x0f
-+#define CC_PC_FF_IPV6DST2 0x1d
-+#define CC_PC_FF_IPV6SRC1 0x10
-+#define CC_PC_FF_IPV6SRC2 0x1e
-+#define CC_PC_FF_IPV6HOP_LIMIT 0x2a
-+#define CC_PC_FF_IPPID 0x24
-+#define CC_PC_FF_IPDSCP 0x76
-+
-+#define CC_PC_FF_GREPTYPE 0x11
-+
-+#define CC_PC_FF_MINENCAP_PTYPE 0x12
-+#define CC_PC_FF_MINENCAP_IPDST 0x13
-+#define CC_PC_FF_MINENCAP_IPSRC 0x14
-+#define CC_PC_FF_MINENCAP_IPSRC_IPDST 0x15
-+
-+#define CC_PC_FF_L4PSRC 0x1f
-+#define CC_PC_FF_L4PDST 0x20
-+#define CC_PC_FF_L4PSRC_L4PDST 0x21
-+
-+#define CC_PC_FF_PPPPID 0x05
-+
-+#define CC_PC_PR_SHIM1 0x22
-+#define CC_PC_PR_SHIM2 0x23
-+
-+#define CC_PC_GENERIC_WITHOUT_MASK 0x27
-+#define CC_PC_GENERIC_WITH_MASK 0x28
-+#define CC_PC_GENERIC_IC_GMASK 0x2B
-+#define CC_PC_GENERIC_IC_HASH_INDEXED 0x2C
-+#define CC_PC_GENERIC_IC_AGING_MASK 0x2D
-+
-+#define CC_PR_OFFSET 0x25
-+#define CC_PR_WITHOUT_OFFSET 0x26
-+
-+#define CC_PC_PR_ETH_OFFSET 19
-+#define CC_PC_PR_USER_DEFINED_SHIM1_OFFSET 16
-+#define CC_PC_PR_USER_DEFINED_SHIM2_OFFSET 17
-+#define CC_PC_PR_USER_LLC_SNAP_OFFSET 20
-+#define CC_PC_PR_VLAN1_OFFSET 21
-+#define CC_PC_PR_VLAN2_OFFSET 22
-+#define CC_PC_PR_PPPOE_OFFSET 24
-+#define CC_PC_PR_MPLS1_OFFSET 25
-+#define CC_PC_PR_MPLS_LAST_OFFSET 26
-+#define CC_PC_PR_IP1_OFFSET 27
-+#define CC_PC_PR_IP_LAST_OFFSET 28
-+#define CC_PC_PR_MINENC_OFFSET 28
-+#define CC_PC_PR_L4_OFFSET 30
-+#define CC_PC_PR_GRE_OFFSET 29
-+#define CC_PC_PR_ETYPE_LAST_OFFSET 23
-+#define CC_PC_PR_NEXT_HEADER_OFFSET 31
-+
-+#define CC_PC_ILLEGAL 0xff
-+#define CC_SIZE_ILLEGAL 0
-+
-+#define FM_PCD_CC_KEYS_MATCH_TABLE_ALIGN 16
-+#define FM_PCD_CC_AD_TABLE_ALIGN 16
-+#define FM_PCD_CC_AD_ENTRY_SIZE 16
-+#define FM_PCD_CC_NUM_OF_KEYS 255
-+#define FM_PCD_CC_TREE_ADDR_ALIGN 256
-+
-+#define FM_PCD_AD_RESULT_CONTRL_FLOW_TYPE 0x00000000
-+#define FM_PCD_AD_RESULT_DATA_FLOW_TYPE 0x80000000
-+#define FM_PCD_AD_RESULT_PLCR_DIS 0x20000000
-+#define FM_PCD_AD_RESULT_EXTENDED_MODE 0x80000000
-+#define FM_PCD_AD_RESULT_NADEN 0x20000000
-+#define FM_PCD_AD_RESULT_STATISTICS_EN 0x40000000
-+
-+#define FM_PCD_AD_CONT_LOOKUP_TYPE 0x40000000
-+#define FM_PCD_AD_CONT_LOOKUP_LCL_MASK 0x00800000
-+
-+#define FM_PCD_AD_STATS_TYPE 0x40000000
-+#define FM_PCD_AD_STATS_FLR_ADDR_MASK 0x00FFFFFF
-+#define FM_PCD_AD_STATS_COUNTERS_ADDR_MASK 0x00FFFFFF
-+#define FM_PCD_AD_STATS_NEXT_ACTION_MASK 0xFFFF0000
-+#define FM_PCD_AD_STATS_NEXT_ACTION_SHIFT 12
-+#define FM_PCD_AD_STATS_NAD_EN 0x00008000
-+#define FM_PCD_AD_STATS_OP_CODE 0x00000036
-+#define FM_PCD_AD_STATS_FLR_EN 0x00004000
-+#define FM_PCD_AD_STATS_COND_EN 0x00002000
-+
-+
-+
-+#define FM_PCD_AD_BYPASS_TYPE 0xc0000000
-+
-+#define FM_PCD_AD_TYPE_MASK 0xc0000000
-+#define FM_PCD_AD_OPCODE_MASK 0x0000000f
-+
-+#define FM_PCD_AD_PROFILEID_FOR_CNTRL_SHIFT 16
-+#if (DPAA_VERSION >= 11)
-+#define FM_PCD_AD_RESULT_VSP_SHIFT 24
-+#define FM_PCD_AD_RESULT_NO_OM_VSPE 0x02000000
-+#define FM_PCD_AD_RESULT_VSP_MASK 0x3f
-+#define FM_PCD_AD_NCSPFQIDM_MASK 0x80000000
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+#define GLBL_MASK_FOR_HASH_INDEXED 0xfff00000
-+#define CC_GLBL_MASK_SIZE 4
-+#define CC_AGING_MASK_SIZE 4
-+
-+typedef uint32_t ccPrivateInfo_t; /**< private info of CC: */
-+
-+#define CC_PRIVATE_INFO_NONE 0
-+#define CC_PRIVATE_INFO_IC_HASH_INDEX_LOOKUP 0x80000000
-+#define CC_PRIVATE_INFO_IC_HASH_EXACT_MATCH 0x40000000
-+#define CC_PRIVATE_INFO_IC_KEY_EXACT_MATCH 0x20000000
-+#define CC_PRIVATE_INFO_IC_DEQ_FQID_INDEX_LOOKUP 0x10000000
-+
-+#define CC_BUILD_AGING_MASK(numOfKeys) ((((1LL << ((numOfKeys) + 1)) - 1)) << (31 - (numOfKeys)))
-+/***********************************************************************/
-+/* Memory map */
-+/***********************************************************************/
-+#if defined(__MWERKS__) && !defined(__GNUC__)
-+#pragma pack(push,1)
-+#endif /* defined(__MWERKS__) && ... */
-+
-+typedef struct
-+{
-+ volatile uint32_t fqid;
-+ volatile uint32_t plcrProfile;
-+ volatile uint32_t nia;
-+ volatile uint32_t res;
-+} t_AdOfTypeResult;
-+
-+typedef struct
-+{
-+ volatile uint32_t ccAdBase;
-+ volatile uint32_t matchTblPtr;
-+ volatile uint32_t pcAndOffsets;
-+ volatile uint32_t gmask;
-+} t_AdOfTypeContLookup;
-+
-+typedef struct
-+{
-+ volatile uint32_t profileTableAddr;
-+ volatile uint32_t reserved;
-+ volatile uint32_t nextActionIndx;
-+ volatile uint32_t statsTableAddr;
-+} t_AdOfTypeStats;
-+
-+typedef union
-+{
-+ volatile t_AdOfTypeResult adResult;
-+ volatile t_AdOfTypeContLookup adContLookup;
-+} t_Ad;
-+
-+#if defined(__MWERKS__) && !defined(__GNUC__)
-+#pragma pack(pop)
-+#endif /* defined(__MWERKS__) && ... */
-+
-+
-+/***********************************************************************/
-+/* Driver's internal structures */
-+/***********************************************************************/
-+
-+typedef struct t_FmPcdStatsObj
-+{
-+ t_Handle h_StatsAd;
-+ t_Handle h_StatsCounters;
-+ t_List node;
-+} t_FmPcdStatsObj;
-+
-+typedef struct
-+{
-+ uint8_t key[FM_PCD_MAX_SIZE_OF_KEY];
-+ uint8_t mask[FM_PCD_MAX_SIZE_OF_KEY];
-+
-+ t_FmPcdCcNextEngineParams nextEngineParams;
-+ uint32_t requiredAction;
-+ uint32_t shadowAction;
-+
-+ t_FmPcdStatsObj *p_StatsObj;
-+
-+} t_FmPcdCcKeyAndNextEngineParams;
-+
-+typedef struct
-+{
-+ t_Handle p_Ad;
-+ e_FmPcdEngine fmPcdEngine;
-+ bool adAllocated;
-+ bool isTree;
-+
-+ uint32_t myInfo;
-+ t_List *h_CcNextNodesLst;
-+ t_Handle h_AdditionalInfo;
-+ t_Handle h_Node;
-+} t_FmPcdModifyCcAdditionalParams;
-+
-+typedef struct
-+{
-+ t_Handle p_AdTableNew;
-+ t_Handle p_KeysMatchTableNew;
-+ t_Handle p_AdTableOld;
-+ t_Handle p_KeysMatchTableOld;
-+ uint16_t numOfKeys;
-+ t_Handle h_CurrentNode;
-+ uint16_t savedKeyIndex;
-+ t_Handle h_NodeForAdd;
-+ t_Handle h_NodeForRmv;
-+ t_Handle h_ManipForRmv;
-+ t_Handle h_ManipForAdd;
-+ t_FmPcdStatsObj *p_StatsObjForRmv;
-+#if (DPAA_VERSION >= 11)
-+ t_Handle h_FrmReplicForAdd;
-+ t_Handle h_FrmReplicForRmv;
-+#endif /* (DPAA_VERSION >= 11) */
-+ bool tree;
-+
-+ t_FmPcdCcKeyAndNextEngineParams keyAndNextEngineParams[CC_MAX_NUM_OF_KEYS];
-+} t_FmPcdModifyCcKeyAdditionalParams;
-+
-+typedef struct
-+{
-+ t_Handle h_Manip;
-+ t_Handle h_CcNode;
-+} t_CcNextEngineInfo;
-+
-+typedef struct
-+{
-+ uint16_t numOfKeys;
-+ uint16_t maxNumOfKeys;
-+
-+ bool maskSupport;
-+ uint32_t keysMatchTableMaxSize;
-+
-+ e_FmPcdCcStatsMode statisticsMode;
-+ uint32_t numOfStatsFLRs;
-+ uint32_t countersArraySize;
-+
-+ bool isHashBucket; /**< Valid for match table node that is a bucket of a hash table only */
-+ t_Handle h_MissStatsCounters; /**< Valid for hash table node and match table that is a bucket;
-+ Holds the statistics counters allocated by the hash table and
-+ are shared by all hash table buckets; */
-+ t_Handle h_PrivMissStatsCounters; /**< Valid for match table node that is a bucket of a hash table only;
-+ Holds the statistics counters that were allocated for this node
-+ and replaced by the shared counters (allocated by the hash table); */
-+ bool statsEnForMiss; /**< Valid for hash table node only; TRUE is statistics are currently
-+ enabled for hash 'miss', FALSE otherwise; This parameter effects the
-+ returned statistics count to user, statistics AD always present for 'miss'
-+ for all hash buckets; */
-+ bool glblMaskUpdated;
-+ t_Handle p_GlblMask;
-+ bool lclMask;
-+ uint8_t parseCode;
-+ uint8_t offset;
-+ uint8_t prsArrayOffset;
-+ bool ctrlFlow;
-+ uint16_t owners;
-+
-+ uint8_t ccKeySizeAccExtraction;
-+ uint8_t sizeOfExtraction;
-+ uint8_t glblMaskSize;
-+
-+ t_Handle h_KeysMatchTable;
-+ t_Handle h_AdTable;
-+ t_Handle h_StatsAds;
-+ t_Handle h_TmpAd;
-+ t_Handle h_Ad;
-+ t_Handle h_StatsFLRs;
-+
-+ t_List availableStatsLst;
-+
-+ t_List ccPrevNodesLst;
-+
-+ t_List ccTreeIdLst;
-+ t_List ccTreesLst;
-+
-+ t_Handle h_FmPcd;
-+ uint32_t shadowAction;
-+ uint8_t userSizeOfExtraction;
-+ uint8_t userOffset;
-+ uint8_t kgHashShift; /* used in hash-table */
-+
-+ t_Handle h_Spinlock;
-+
-+ t_FmPcdCcKeyAndNextEngineParams keyAndNextEngineParams[CC_MAX_NUM_OF_KEYS];
-+} t_FmPcdCcNode;
-+
-+typedef struct
-+{
-+ t_FmPcdCcNode *p_FmPcdCcNode;
-+ bool occupied;
-+ uint16_t owners;
-+ volatile bool lock;
-+} t_FmPcdCcNodeArray;
-+
-+typedef struct
-+{
-+ uint8_t numOfEntriesInGroup;
-+ uint32_t totalBitsMask;
-+ uint8_t baseGroupEntry;
-+} t_FmPcdCcGroupParam;
-+
-+typedef struct
-+{
-+ t_Handle h_FmPcd;
-+ uint8_t netEnvId;
-+ uintptr_t ccTreeBaseAddr;
-+ uint8_t numOfGrps;
-+ t_FmPcdCcGroupParam fmPcdGroupParam[FM_PCD_MAX_NUM_OF_CC_GROUPS];
-+ t_List fmPortsLst;
-+ t_FmPcdLock *p_Lock;
-+ uint8_t numOfEntries;
-+ uint16_t owners;
-+ t_Handle h_FmPcdCcSavedManipParams;
-+ bool modifiedState;
-+ uint32_t requiredAction;
-+ t_Handle h_IpReassemblyManip;
-+ t_Handle h_CapwapReassemblyManip;
-+
-+ t_FmPcdCcKeyAndNextEngineParams keyAndNextEngineParams[FM_PCD_MAX_NUM_OF_CC_GROUPS];
-+} t_FmPcdCcTree;
-+
-+
-+t_Error FmPcdCcNodeTreeTryLock(t_Handle h_FmPcd,t_Handle h_FmPcdCcNode, t_List *p_List);
-+void FmPcdCcNodeTreeReleaseLock(t_Handle h_FmPcd, t_List *p_List);
-+t_Error FmPcdUpdateCcShadow (t_FmPcd *p_FmPcd, uint32_t size, uint32_t align);
-+
-+
-+#endif /* __FM_CC_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/Pcd/fm_kg.c
-@@ -0,0 +1,3242 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/******************************************************************************
-+ @File fm_kg.c
-+
-+ @Description FM PCD ...
-+*//***************************************************************************/
-+#include "std_ext.h"
-+#include "error_ext.h"
-+#include "string_ext.h"
-+#include "debug_ext.h"
-+#include "net_ext.h"
-+#include "fm_port_ext.h"
-+
-+#include "fm_common.h"
-+#include "fm_pcd.h"
-+#include "fm_hc.h"
-+#include "fm_pcd_ipc.h"
-+#include "fm_kg.h"
-+#include "fsl_fman_kg.h"
-+
-+
-+/****************************************/
-+/* static functions */
-+/****************************************/
-+
-+static uint32_t KgHwLock(t_Handle h_FmPcdKg)
-+{
-+ ASSERT_COND(h_FmPcdKg);
-+ return XX_LockIntrSpinlock(((t_FmPcdKg *)h_FmPcdKg)->h_HwSpinlock);
-+}
-+
-+static void KgHwUnlock(t_Handle h_FmPcdKg, uint32_t intFlags)
-+{
-+ ASSERT_COND(h_FmPcdKg);
-+ XX_UnlockIntrSpinlock(((t_FmPcdKg *)h_FmPcdKg)->h_HwSpinlock, intFlags);
-+}
-+
-+static uint32_t KgSchemeLock(t_Handle h_Scheme)
-+{
-+ ASSERT_COND(h_Scheme);
-+ return FmPcdLockSpinlock(((t_FmPcdKgScheme *)h_Scheme)->p_Lock);
-+}
-+
-+static void KgSchemeUnlock(t_Handle h_Scheme, uint32_t intFlags)
-+{
-+ ASSERT_COND(h_Scheme);
-+ FmPcdUnlockSpinlock(((t_FmPcdKgScheme *)h_Scheme)->p_Lock, intFlags);
-+}
-+
-+static bool KgSchemeFlagTryLock(t_Handle h_Scheme)
-+{
-+ ASSERT_COND(h_Scheme);
-+ return FmPcdLockTryLock(((t_FmPcdKgScheme *)h_Scheme)->p_Lock);
-+}
-+
-+static void KgSchemeFlagUnlock(t_Handle h_Scheme)
-+{
-+ ASSERT_COND(h_Scheme);
-+ FmPcdLockUnlock(((t_FmPcdKgScheme *)h_Scheme)->p_Lock);
-+}
-+
-+static t_Error WriteKgarWait(t_FmPcd *p_FmPcd, uint32_t fmkg_ar)
-+{
-+
-+ struct fman_kg_regs *regs = p_FmPcd->p_FmPcdKg->p_FmPcdKgRegs;
-+
-+ if (fman_kg_write_ar_wait(regs, fmkg_ar))
-+ RETURN_ERROR(MINOR, E_INVALID_STATE, ("Keygen scheme access violation"));
-+
-+ return E_OK;
-+}
-+
-+static e_FmPcdKgExtractDfltSelect GetGenericSwDefault(t_FmPcdKgExtractDflt swDefaults[], uint8_t numOfSwDefaults, uint8_t code)
-+{
-+ int i;
-+
-+ switch (code)
-+ {
-+ case (KG_SCH_GEN_PARSE_RESULT_N_FQID):
-+ case (KG_SCH_GEN_DEFAULT):
-+ case (KG_SCH_GEN_NEXTHDR):
-+ for (i=0 ; i<numOfSwDefaults ; i++)
-+ if (swDefaults[i].type == e_FM_PCD_KG_GENERIC_NOT_FROM_DATA)
-+ return swDefaults[i].dfltSelect;
-+ break;
-+ case (KG_SCH_GEN_SHIM1):
-+ case (KG_SCH_GEN_SHIM2):
-+ case (KG_SCH_GEN_IP_PID_NO_V):
-+ case (KG_SCH_GEN_ETH_NO_V):
-+ case (KG_SCH_GEN_SNAP_NO_V):
-+ case (KG_SCH_GEN_VLAN1_NO_V):
-+ case (KG_SCH_GEN_VLAN2_NO_V):
-+ case (KG_SCH_GEN_ETH_TYPE_NO_V):
-+ case (KG_SCH_GEN_PPP_NO_V):
-+ case (KG_SCH_GEN_MPLS1_NO_V):
-+ case (KG_SCH_GEN_MPLS_LAST_NO_V):
-+ case (KG_SCH_GEN_L3_NO_V):
-+ case (KG_SCH_GEN_IP2_NO_V):
-+ case (KG_SCH_GEN_GRE_NO_V):
-+ case (KG_SCH_GEN_L4_NO_V):
-+ for (i=0 ; i<numOfSwDefaults ; i++)
-+ if (swDefaults[i].type == e_FM_PCD_KG_GENERIC_FROM_DATA_NO_V)
-+ return swDefaults[i].dfltSelect;
-+ break;
-+ case (KG_SCH_GEN_START_OF_FRM):
-+ case (KG_SCH_GEN_ETH):
-+ case (KG_SCH_GEN_SNAP):
-+ case (KG_SCH_GEN_VLAN1):
-+ case (KG_SCH_GEN_VLAN2):
-+ case (KG_SCH_GEN_ETH_TYPE):
-+ case (KG_SCH_GEN_PPP):
-+ case (KG_SCH_GEN_MPLS1):
-+ case (KG_SCH_GEN_MPLS2):
-+ case (KG_SCH_GEN_MPLS3):
-+ case (KG_SCH_GEN_MPLS_LAST):
-+ case (KG_SCH_GEN_IPV4):
-+ case (KG_SCH_GEN_IPV6):
-+ case (KG_SCH_GEN_IPV4_TUNNELED):
-+ case (KG_SCH_GEN_IPV6_TUNNELED):
-+ case (KG_SCH_GEN_MIN_ENCAP):
-+ case (KG_SCH_GEN_GRE):
-+ case (KG_SCH_GEN_TCP):
-+ case (KG_SCH_GEN_UDP):
-+ case (KG_SCH_GEN_IPSEC_AH):
-+ case (KG_SCH_GEN_SCTP):
-+ case (KG_SCH_GEN_DCCP):
-+ case (KG_SCH_GEN_IPSEC_ESP):
-+ for (i=0 ; i<numOfSwDefaults ; i++)
-+ if (swDefaults[i].type == e_FM_PCD_KG_GENERIC_FROM_DATA)
-+ return swDefaults[i].dfltSelect;
-+ break;
-+ default:
-+ break;
-+ }
-+
-+ return e_FM_PCD_KG_DFLT_ILLEGAL;
-+}
-+
-+static uint8_t GetGenCode(e_FmPcdExtractFrom src, uint8_t *p_Offset)
-+{
-+ *p_Offset = 0;
-+
-+ switch (src)
-+ {
-+ case (e_FM_PCD_EXTRACT_FROM_FRAME_START):
-+ return KG_SCH_GEN_START_OF_FRM;
-+ case (e_FM_PCD_EXTRACT_FROM_DFLT_VALUE):
-+ return KG_SCH_GEN_DEFAULT;
-+ case (e_FM_PCD_EXTRACT_FROM_PARSE_RESULT):
-+ return KG_SCH_GEN_PARSE_RESULT_N_FQID;
-+ case (e_FM_PCD_EXTRACT_FROM_ENQ_FQID):
-+ *p_Offset = 32;
-+ return KG_SCH_GEN_PARSE_RESULT_N_FQID;
-+ case (e_FM_PCD_EXTRACT_FROM_CURR_END_OF_PARSE):
-+ return KG_SCH_GEN_NEXTHDR;
-+ default:
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE, ("Illegal 'extract from' src"));
-+ return 0;
-+ }
-+}
-+
-+static uint8_t GetGenHdrCode(e_NetHeaderType hdr, e_FmPcdHdrIndex hdrIndex, bool ignoreProtocolValidation)
-+{
-+ if (!ignoreProtocolValidation)
-+ switch (hdr)
-+ {
-+ case (HEADER_TYPE_NONE):
-+ ASSERT_COND(FALSE);
-+ case (HEADER_TYPE_ETH):
-+ return KG_SCH_GEN_ETH;
-+ case (HEADER_TYPE_LLC_SNAP):
-+ return KG_SCH_GEN_SNAP;
-+ case (HEADER_TYPE_PPPoE):
-+ return KG_SCH_GEN_PPP;
-+ case (HEADER_TYPE_MPLS):
-+ if ((hdrIndex == e_FM_PCD_HDR_INDEX_NONE) || (hdrIndex == e_FM_PCD_HDR_INDEX_1))
-+ return KG_SCH_GEN_MPLS1;
-+ if (hdrIndex == e_FM_PCD_HDR_INDEX_2)
-+ return KG_SCH_GEN_MPLS2;
-+ if (hdrIndex == e_FM_PCD_HDR_INDEX_3)
-+ return KG_SCH_GEN_MPLS3;
-+ if (hdrIndex == e_FM_PCD_HDR_INDEX_LAST)
-+ return KG_SCH_GEN_MPLS_LAST;
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Illegal MPLS header index"));
-+ return 0;
-+ case (HEADER_TYPE_IPv4):
-+ if ((hdrIndex == e_FM_PCD_HDR_INDEX_NONE) || (hdrIndex == e_FM_PCD_HDR_INDEX_1))
-+ return KG_SCH_GEN_IPV4;
-+ if ((hdrIndex == e_FM_PCD_HDR_INDEX_2) || (hdrIndex == e_FM_PCD_HDR_INDEX_LAST))
-+ return KG_SCH_GEN_IPV4_TUNNELED;
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Illegal IPv4 header index"));
-+ return 0;
-+ case (HEADER_TYPE_IPv6):
-+ if ((hdrIndex == e_FM_PCD_HDR_INDEX_NONE) || (hdrIndex == e_FM_PCD_HDR_INDEX_1))
-+ return KG_SCH_GEN_IPV6;
-+ if ((hdrIndex == e_FM_PCD_HDR_INDEX_2) || (hdrIndex == e_FM_PCD_HDR_INDEX_LAST))
-+ return KG_SCH_GEN_IPV6_TUNNELED;
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Illegal IPv6 header index"));
-+ return 0;
-+ case (HEADER_TYPE_GRE):
-+ return KG_SCH_GEN_GRE;
-+ case (HEADER_TYPE_TCP):
-+ return KG_SCH_GEN_TCP;
-+ case (HEADER_TYPE_UDP):
-+ return KG_SCH_GEN_UDP;
-+ case (HEADER_TYPE_IPSEC_AH):
-+ return KG_SCH_GEN_IPSEC_AH;
-+ case (HEADER_TYPE_IPSEC_ESP):
-+ return KG_SCH_GEN_IPSEC_ESP;
-+ case (HEADER_TYPE_SCTP):
-+ return KG_SCH_GEN_SCTP;
-+ case (HEADER_TYPE_DCCP):
-+ return KG_SCH_GEN_DCCP;
-+ default:
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Extraction not supported"));
-+ return 0;
-+ }
-+ else
-+ switch (hdr)
-+ {
-+ case (HEADER_TYPE_NONE):
-+ ASSERT_COND(FALSE);
-+ case (HEADER_TYPE_ETH):
-+ return KG_SCH_GEN_ETH_NO_V;
-+ case (HEADER_TYPE_LLC_SNAP):
-+ return KG_SCH_GEN_SNAP_NO_V;
-+ case (HEADER_TYPE_PPPoE):
-+ return KG_SCH_GEN_PPP_NO_V;
-+ case (HEADER_TYPE_MPLS):
-+ if ((hdrIndex == e_FM_PCD_HDR_INDEX_NONE) || (hdrIndex == e_FM_PCD_HDR_INDEX_1))
-+ return KG_SCH_GEN_MPLS1_NO_V;
-+ if (hdrIndex == e_FM_PCD_HDR_INDEX_LAST)
-+ return KG_SCH_GEN_MPLS_LAST_NO_V;
-+ if ((hdrIndex == e_FM_PCD_HDR_INDEX_2) || (hdrIndex == e_FM_PCD_HDR_INDEX_3) )
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Indexed MPLS Extraction not supported"));
-+ else
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Illegal MPLS header index"));
-+ return 0;
-+ case (HEADER_TYPE_IPv4):
-+ case (HEADER_TYPE_IPv6):
-+ if ((hdrIndex == e_FM_PCD_HDR_INDEX_NONE) || (hdrIndex == e_FM_PCD_HDR_INDEX_1))
-+ return KG_SCH_GEN_L3_NO_V;
-+ if ((hdrIndex == e_FM_PCD_HDR_INDEX_2) || (hdrIndex == e_FM_PCD_HDR_INDEX_LAST))
-+ return KG_SCH_GEN_IP2_NO_V;
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Illegal IP header index"));
-+ case (HEADER_TYPE_MINENCAP):
-+ return KG_SCH_GEN_IP2_NO_V;
-+ case (HEADER_TYPE_USER_DEFINED_L3):
-+ return KG_SCH_GEN_L3_NO_V;
-+ case (HEADER_TYPE_GRE):
-+ return KG_SCH_GEN_GRE_NO_V;
-+ case (HEADER_TYPE_TCP):
-+ case (HEADER_TYPE_UDP):
-+ case (HEADER_TYPE_IPSEC_AH):
-+ case (HEADER_TYPE_IPSEC_ESP):
-+ case (HEADER_TYPE_SCTP):
-+ case (HEADER_TYPE_DCCP):
-+ return KG_SCH_GEN_L4_NO_V;
-+ case (HEADER_TYPE_USER_DEFINED_SHIM1):
-+ return KG_SCH_GEN_SHIM1;
-+ case (HEADER_TYPE_USER_DEFINED_SHIM2):
-+ return KG_SCH_GEN_SHIM2;
-+ default:
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Extraction not supported"));
-+ return 0;
-+ }
-+}
-+static t_GenericCodes GetGenFieldCode(e_NetHeaderType hdr, t_FmPcdFields field, bool ignoreProtocolValidation, e_FmPcdHdrIndex hdrIndex)
-+{
-+ if (!ignoreProtocolValidation)
-+ switch (hdr)
-+ {
-+ case (HEADER_TYPE_NONE):
-+ ASSERT_COND(FALSE);
-+ break;
-+ case (HEADER_TYPE_ETH):
-+ switch (field.eth)
-+ {
-+ case (NET_HEADER_FIELD_ETH_TYPE):
-+ return KG_SCH_GEN_ETH_TYPE;
-+ default:
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Extraction not supported"));
-+ return 0;
-+ }
-+ break;
-+ case (HEADER_TYPE_VLAN):
-+ switch (field.vlan)
-+ {
-+ case (NET_HEADER_FIELD_VLAN_TCI):
-+ if ((hdrIndex == e_FM_PCD_HDR_INDEX_NONE) || (hdrIndex == e_FM_PCD_HDR_INDEX_1))
-+ return KG_SCH_GEN_VLAN1;
-+ if (hdrIndex == e_FM_PCD_HDR_INDEX_LAST)
-+ return KG_SCH_GEN_VLAN2;
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Illegal VLAN header index"));
-+ return 0;
-+ }
-+ break;
-+ case (HEADER_TYPE_MPLS):
-+ case (HEADER_TYPE_IPSEC_AH):
-+ case (HEADER_TYPE_IPSEC_ESP):
-+ case (HEADER_TYPE_LLC_SNAP):
-+ case (HEADER_TYPE_PPPoE):
-+ case (HEADER_TYPE_IPv4):
-+ case (HEADER_TYPE_IPv6):
-+ case (HEADER_TYPE_GRE):
-+ case (HEADER_TYPE_MINENCAP):
-+ case (HEADER_TYPE_USER_DEFINED_L3):
-+ case (HEADER_TYPE_TCP):
-+ case (HEADER_TYPE_UDP):
-+ case (HEADER_TYPE_SCTP):
-+ case (HEADER_TYPE_DCCP):
-+ case (HEADER_TYPE_USER_DEFINED_L4):
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Extraction not supported"));
-+ return 0;
-+ default:
-+ break;
-+
-+ }
-+ else
-+ switch (hdr)
-+ {
-+ case (HEADER_TYPE_NONE):
-+ ASSERT_COND(FALSE);
-+ break;
-+ case (HEADER_TYPE_ETH):
-+ switch (field.eth)
-+ {
-+ case (NET_HEADER_FIELD_ETH_TYPE):
-+ return KG_SCH_GEN_ETH_TYPE_NO_V;
-+ default:
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Extraction not supported"));
-+ return 0;
-+ }
-+ break;
-+ case (HEADER_TYPE_VLAN):
-+ switch (field.vlan)
-+ {
-+ case (NET_HEADER_FIELD_VLAN_TCI) :
-+ if ((hdrIndex == e_FM_PCD_HDR_INDEX_NONE) || (hdrIndex == e_FM_PCD_HDR_INDEX_1))
-+ return KG_SCH_GEN_VLAN1_NO_V;
-+ if (hdrIndex == e_FM_PCD_HDR_INDEX_LAST)
-+ return KG_SCH_GEN_VLAN2_NO_V;
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Illegal VLAN header index"));
-+ return 0;
-+ }
-+ break;
-+ case (HEADER_TYPE_IPv4):
-+ switch (field.ipv4)
-+ {
-+ case (NET_HEADER_FIELD_IPv4_PROTO):
-+ return KG_SCH_GEN_IP_PID_NO_V;
-+ default:
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Extraction not supported"));
-+ return 0;
-+ }
-+ break;
-+ case (HEADER_TYPE_IPv6):
-+ switch (field.ipv6)
-+ {
-+ case (NET_HEADER_FIELD_IPv6_NEXT_HDR):
-+ return KG_SCH_GEN_IP_PID_NO_V;
-+ default:
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Extraction not supported"));
-+ return 0;
-+ }
-+ break;
-+ case (HEADER_TYPE_MPLS):
-+ case (HEADER_TYPE_LLC_SNAP):
-+ case (HEADER_TYPE_PPPoE):
-+ case (HEADER_TYPE_GRE):
-+ case (HEADER_TYPE_MINENCAP):
-+ case (HEADER_TYPE_USER_DEFINED_L3):
-+ case (HEADER_TYPE_TCP):
-+ case (HEADER_TYPE_UDP):
-+ case (HEADER_TYPE_IPSEC_AH):
-+ case (HEADER_TYPE_IPSEC_ESP):
-+ case (HEADER_TYPE_SCTP):
-+ case (HEADER_TYPE_DCCP):
-+ case (HEADER_TYPE_USER_DEFINED_L4):
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Extraction not supported"));
-+ return 0;
-+ default:
-+ break;
-+ }
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Header not supported"));
-+ return 0;
-+}
-+
-+static t_KnownFieldsMasks GetKnownProtMask(t_FmPcd *p_FmPcd, e_NetHeaderType hdr, e_FmPcdHdrIndex index, t_FmPcdFields field)
-+{
-+ UNUSED(p_FmPcd);
-+
-+ switch (hdr)
-+ {
-+ case (HEADER_TYPE_NONE):
-+ ASSERT_COND(FALSE);
-+ break;
-+ case (HEADER_TYPE_ETH):
-+ switch (field.eth)
-+ {
-+ case (NET_HEADER_FIELD_ETH_DA):
-+ return KG_SCH_KN_MACDST;
-+ case (NET_HEADER_FIELD_ETH_SA):
-+ return KG_SCH_KN_MACSRC;
-+ case (NET_HEADER_FIELD_ETH_TYPE):
-+ return KG_SCH_KN_ETYPE;
-+ default:
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Extraction not supported"));
-+ return 0;
-+ }
-+ case (HEADER_TYPE_LLC_SNAP):
-+ switch (field.llcSnap)
-+ {
-+ case (NET_HEADER_FIELD_LLC_SNAP_TYPE):
-+ return KG_SCH_KN_ETYPE;
-+ default:
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Extraction not supported"));
-+ return 0;
-+ }
-+ case (HEADER_TYPE_VLAN):
-+ switch (field.vlan)
-+ {
-+ case (NET_HEADER_FIELD_VLAN_TCI):
-+ if ((index == e_FM_PCD_HDR_INDEX_NONE) || (index == e_FM_PCD_HDR_INDEX_1))
-+ return KG_SCH_KN_TCI1;
-+ if (index == e_FM_PCD_HDR_INDEX_LAST)
-+ return KG_SCH_KN_TCI2;
-+ else
-+ {
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Extraction not supported"));
-+ return 0;
-+ }
-+ default:
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Extraction not supported"));
-+ return 0;
-+ }
-+ case (HEADER_TYPE_MPLS):
-+ switch (field.mpls)
-+ {
-+ case (NET_HEADER_FIELD_MPLS_LABEL_STACK):
-+ if ((index == e_FM_PCD_HDR_INDEX_NONE) || (index == e_FM_PCD_HDR_INDEX_1))
-+ return KG_SCH_KN_MPLS1;
-+ if (index == e_FM_PCD_HDR_INDEX_2)
-+ return KG_SCH_KN_MPLS2;
-+ if (index == e_FM_PCD_HDR_INDEX_LAST)
-+ return KG_SCH_KN_MPLS_LAST;
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Illegal MPLS index"));
-+ return 0;
-+ default:
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Extraction not supported"));
-+ return 0;
-+ }
-+ case (HEADER_TYPE_IPv4):
-+ switch (field.ipv4)
-+ {
-+ case (NET_HEADER_FIELD_IPv4_SRC_IP):
-+ if ((index == e_FM_PCD_HDR_INDEX_NONE) || (index == e_FM_PCD_HDR_INDEX_1))
-+ return KG_SCH_KN_IPSRC1;
-+ if ((index == e_FM_PCD_HDR_INDEX_2) || (index == e_FM_PCD_HDR_INDEX_LAST))
-+ return KG_SCH_KN_IPSRC2;
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Illegal IPv4 index"));
-+ return 0;
-+ case (NET_HEADER_FIELD_IPv4_DST_IP):
-+ if ((index == e_FM_PCD_HDR_INDEX_NONE) || (index == e_FM_PCD_HDR_INDEX_1))
-+ return KG_SCH_KN_IPDST1;
-+ if ((index == e_FM_PCD_HDR_INDEX_2) || (index == e_FM_PCD_HDR_INDEX_LAST))
-+ return KG_SCH_KN_IPDST2;
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Illegal IPv4 index"));
-+ return 0;
-+ case (NET_HEADER_FIELD_IPv4_PROTO):
-+ if ((index == e_FM_PCD_HDR_INDEX_NONE) || (index == e_FM_PCD_HDR_INDEX_1))
-+ return KG_SCH_KN_PTYPE1;
-+ if ((index == e_FM_PCD_HDR_INDEX_2) || (index == e_FM_PCD_HDR_INDEX_LAST))
-+ return KG_SCH_KN_PTYPE2;
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Illegal IPv4 index"));
-+ return 0;
-+ case (NET_HEADER_FIELD_IPv4_TOS):
-+ if ((index == e_FM_PCD_HDR_INDEX_NONE) || (index == e_FM_PCD_HDR_INDEX_1))
-+ return KG_SCH_KN_IPTOS_TC1;
-+ if ((index == e_FM_PCD_HDR_INDEX_2) || (index == e_FM_PCD_HDR_INDEX_LAST))
-+ return KG_SCH_KN_IPTOS_TC2;
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Illegal IPv4 index"));
-+ return 0;
-+ default:
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Extraction not supported"));
-+ return 0;
-+ }
-+ case (HEADER_TYPE_IPv6):
-+ switch (field.ipv6)
-+ {
-+ case (NET_HEADER_FIELD_IPv6_SRC_IP):
-+ if ((index == e_FM_PCD_HDR_INDEX_NONE) || (index == e_FM_PCD_HDR_INDEX_1))
-+ return KG_SCH_KN_IPSRC1;
-+ if ((index == e_FM_PCD_HDR_INDEX_2) || (index == e_FM_PCD_HDR_INDEX_LAST))
-+ return KG_SCH_KN_IPSRC2;
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Illegal IPv6 index"));
-+ return 0;
-+ case (NET_HEADER_FIELD_IPv6_DST_IP):
-+ if ((index == e_FM_PCD_HDR_INDEX_NONE) || (index == e_FM_PCD_HDR_INDEX_1))
-+ return KG_SCH_KN_IPDST1;
-+ if ((index == e_FM_PCD_HDR_INDEX_2) || (index == e_FM_PCD_HDR_INDEX_LAST))
-+ return KG_SCH_KN_IPDST2;
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Illegal IPv6 index"));
-+ return 0;
-+ case (NET_HEADER_FIELD_IPv6_NEXT_HDR):
-+ if ((index == e_FM_PCD_HDR_INDEX_NONE) || (index == e_FM_PCD_HDR_INDEX_1))
-+ return KG_SCH_KN_PTYPE1;
-+ if (index == e_FM_PCD_HDR_INDEX_2)
-+ return KG_SCH_KN_PTYPE2;
-+ if (index == e_FM_PCD_HDR_INDEX_LAST)
-+#ifdef FM_KG_NO_IPPID_SUPPORT
-+ if (p_FmPcd->fmRevInfo.majorRev < 6)
-+ return KG_SCH_KN_PTYPE2;
-+#endif /* FM_KG_NO_IPPID_SUPPORT */
-+ return KG_SCH_KN_IPPID;
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Illegal IPv6 index"));
-+ return 0;
-+ case (NET_HEADER_FIELD_IPv6_VER | NET_HEADER_FIELD_IPv6_FL | NET_HEADER_FIELD_IPv6_TC):
-+ if ((index == e_FM_PCD_HDR_INDEX_NONE) || (index == e_FM_PCD_HDR_INDEX_1))
-+ return (KG_SCH_KN_IPV6FL1 | KG_SCH_KN_IPTOS_TC1);
-+ if ((index == e_FM_PCD_HDR_INDEX_2) || (index == e_FM_PCD_HDR_INDEX_LAST))
-+ return (KG_SCH_KN_IPV6FL2 | KG_SCH_KN_IPTOS_TC2);
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Illegal IPv6 index"));
-+ return 0;
-+ case (NET_HEADER_FIELD_IPv6_VER | NET_HEADER_FIELD_IPv6_TC):
-+ if ((index == e_FM_PCD_HDR_INDEX_NONE) || (index == e_FM_PCD_HDR_INDEX_1))
-+ return KG_SCH_KN_IPTOS_TC1;
-+ if ((index == e_FM_PCD_HDR_INDEX_2) || (index == e_FM_PCD_HDR_INDEX_LAST))
-+ return KG_SCH_KN_IPTOS_TC2;
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Illegal IPv6 index"));
-+ return 0;
-+ case (NET_HEADER_FIELD_IPv6_FL):
-+ if ((index == e_FM_PCD_HDR_INDEX_NONE) || (index == e_FM_PCD_HDR_INDEX_1))
-+ return KG_SCH_KN_IPV6FL1;
-+ if ((index == e_FM_PCD_HDR_INDEX_2) || (index == e_FM_PCD_HDR_INDEX_LAST))
-+ return KG_SCH_KN_IPV6FL2;
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Illegal IPv6 index"));
-+ return 0;
-+ default:
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Extraction not supported"));
-+ return 0;
-+ }
-+ case (HEADER_TYPE_GRE):
-+ switch (field.gre)
-+ {
-+ case (NET_HEADER_FIELD_GRE_TYPE):
-+ return KG_SCH_KN_GREPTYPE;
-+ default:
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Extraction not supported"));
-+ return 0;
-+ }
-+ case (HEADER_TYPE_MINENCAP):
-+ switch (field.minencap)
-+ {
-+ case (NET_HEADER_FIELD_MINENCAP_SRC_IP):
-+ return KG_SCH_KN_IPSRC2;
-+ case (NET_HEADER_FIELD_MINENCAP_DST_IP):
-+ return KG_SCH_KN_IPDST2;
-+ case (NET_HEADER_FIELD_MINENCAP_TYPE):
-+ return KG_SCH_KN_PTYPE2;
-+ default:
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Extraction not supported"));
-+ return 0;
-+ }
-+ case (HEADER_TYPE_TCP):
-+ switch (field.tcp)
-+ {
-+ case (NET_HEADER_FIELD_TCP_PORT_SRC):
-+ return KG_SCH_KN_L4PSRC;
-+ case (NET_HEADER_FIELD_TCP_PORT_DST):
-+ return KG_SCH_KN_L4PDST;
-+ case (NET_HEADER_FIELD_TCP_FLAGS):
-+ return KG_SCH_KN_TFLG;
-+ default:
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Extraction not supported"));
-+ return 0;
-+ }
-+ case (HEADER_TYPE_UDP):
-+ switch (field.udp)
-+ {
-+ case (NET_HEADER_FIELD_UDP_PORT_SRC):
-+ return KG_SCH_KN_L4PSRC;
-+ case (NET_HEADER_FIELD_UDP_PORT_DST):
-+ return KG_SCH_KN_L4PDST;
-+ default:
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Extraction not supported"));
-+ return 0;
-+ }
-+ case (HEADER_TYPE_IPSEC_AH):
-+ switch (field.ipsecAh)
-+ {
-+ case (NET_HEADER_FIELD_IPSEC_AH_SPI):
-+ return KG_SCH_KN_IPSEC_SPI;
-+ case (NET_HEADER_FIELD_IPSEC_AH_NH):
-+ return KG_SCH_KN_IPSEC_NH;
-+ default:
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Extraction not supported"));
-+ return 0;
-+ }
-+ case (HEADER_TYPE_IPSEC_ESP):
-+ switch (field.ipsecEsp)
-+ {
-+ case (NET_HEADER_FIELD_IPSEC_ESP_SPI):
-+ return KG_SCH_KN_IPSEC_SPI;
-+ default:
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Extraction not supported"));
-+ return 0;
-+ }
-+ case (HEADER_TYPE_SCTP):
-+ switch (field.sctp)
-+ {
-+ case (NET_HEADER_FIELD_SCTP_PORT_SRC):
-+ return KG_SCH_KN_L4PSRC;
-+ case (NET_HEADER_FIELD_SCTP_PORT_DST):
-+ return KG_SCH_KN_L4PDST;
-+ default:
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Extraction not supported"));
-+ return 0;
-+ }
-+ case (HEADER_TYPE_DCCP):
-+ switch (field.dccp)
-+ {
-+ case (NET_HEADER_FIELD_DCCP_PORT_SRC):
-+ return KG_SCH_KN_L4PSRC;
-+ case (NET_HEADER_FIELD_DCCP_PORT_DST):
-+ return KG_SCH_KN_L4PDST;
-+ default:
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Extraction not supported"));
-+ return 0;
-+ }
-+ case (HEADER_TYPE_PPPoE):
-+ switch (field.pppoe)
-+ {
-+ case (NET_HEADER_FIELD_PPPoE_PID):
-+ return KG_SCH_KN_PPPID;
-+ case (NET_HEADER_FIELD_PPPoE_SID):
-+ return KG_SCH_KN_PPPSID;
-+ default:
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Extraction not supported"));
-+ return 0;
-+ }
-+ default:
-+ break;
-+
-+ }
-+
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Extraction not supported"));
-+ return 0;
-+}
-+
-+
-+static uint8_t GetKnownFieldId(uint32_t bitMask)
-+{
-+ uint8_t cnt = 0;
-+
-+ while (bitMask)
-+ if (bitMask & 0x80000000)
-+ break;
-+ else
-+ {
-+ cnt++;
-+ bitMask <<= 1;
-+ }
-+ return cnt;
-+
-+}
-+
-+static uint8_t GetExtractedOrMask(uint8_t bitOffset, bool fqid)
-+{
-+ uint8_t i, mask, numOfOnesToClear, walking1Mask = 1;
-+
-+ /* bitOffset 1-7 --> mask 0x1-0x7F */
-+ if (bitOffset<8)
-+ {
-+ mask = 0;
-+ for (i = 0 ; i < bitOffset ; i++, walking1Mask <<= 1)
-+ mask |= walking1Mask;
-+ }
-+ else
-+ {
-+ mask = 0xFF;
-+ numOfOnesToClear = 0;
-+ if (fqid && bitOffset>24)
-+ /* bitOffset 25-31 --> mask 0xFE-0x80 */
-+ numOfOnesToClear = (uint8_t)(bitOffset-24);
-+ else
-+ /* bitOffset 9-15 --> mask 0xFE-0x80 */
-+ if (!fqid && bitOffset>8)
-+ numOfOnesToClear = (uint8_t)(bitOffset-8);
-+ for (i = 0 ; i < numOfOnesToClear ; i++, walking1Mask <<= 1)
-+ mask &= ~walking1Mask;
-+ /* bitOffset 8-24 for FQID, 8 for PP --> no mask (0xFF)*/
-+ }
-+ return mask;
-+}
-+
-+static void IncSchemeOwners(t_FmPcd *p_FmPcd, t_FmPcdKgInterModuleBindPortToSchemes *p_BindPort)
-+{
-+ t_FmPcdKg *p_FmPcdKg;
-+ t_FmPcdKgScheme *p_Scheme;
-+ uint32_t intFlags;
-+ uint8_t relativeSchemeId;
-+ int i;
-+
-+ p_FmPcdKg = p_FmPcd->p_FmPcdKg;
-+
-+ /* for each scheme - update owners counters */
-+ for (i = 0; i < p_BindPort->numOfSchemes; i++)
-+ {
-+ relativeSchemeId = FmPcdKgGetRelativeSchemeId(p_FmPcd, p_BindPort->schemesIds[i]);
-+ ASSERT_COND(relativeSchemeId < FM_PCD_KG_NUM_OF_SCHEMES);
-+
-+ p_Scheme = &p_FmPcdKg->schemes[relativeSchemeId];
-+
-+ /* increment owners number */
-+ intFlags = KgSchemeLock(p_Scheme);
-+ p_Scheme->owners++;
-+ KgSchemeUnlock(p_Scheme, intFlags);
-+ }
-+}
-+
-+static void DecSchemeOwners(t_FmPcd *p_FmPcd, t_FmPcdKgInterModuleBindPortToSchemes *p_BindPort)
-+{
-+ t_FmPcdKg *p_FmPcdKg;
-+ t_FmPcdKgScheme *p_Scheme;
-+ uint32_t intFlags;
-+ uint8_t relativeSchemeId;
-+ int i;
-+
-+ p_FmPcdKg = p_FmPcd->p_FmPcdKg;
-+
-+ /* for each scheme - update owners counters */
-+ for (i = 0; i < p_BindPort->numOfSchemes; i++)
-+ {
-+ relativeSchemeId = FmPcdKgGetRelativeSchemeId(p_FmPcd, p_BindPort->schemesIds[i]);
-+ ASSERT_COND(relativeSchemeId < FM_PCD_KG_NUM_OF_SCHEMES);
-+
-+ p_Scheme = &p_FmPcdKg->schemes[relativeSchemeId];
-+
-+ /* increment owners number */
-+ ASSERT_COND(p_Scheme->owners);
-+ intFlags = KgSchemeLock(p_Scheme);
-+ p_Scheme->owners--;
-+ KgSchemeUnlock(p_Scheme, intFlags);
-+ }
-+}
-+
-+static void UpdateRequiredActionFlag(t_FmPcdKgScheme *p_Scheme, bool set)
-+{
-+ /* this routine is locked by the calling routine */
-+ ASSERT_COND(p_Scheme);
-+ ASSERT_COND(p_Scheme->valid);
-+
-+ if (set)
-+ p_Scheme->requiredActionFlag = TRUE;
-+ else
-+ {
-+ p_Scheme->requiredAction = 0;
-+ p_Scheme->requiredActionFlag = FALSE;
-+ }
-+}
-+
-+static t_Error KgWriteSp(t_FmPcd *p_FmPcd, uint8_t hardwarePortId, uint32_t spReg, bool add)
-+{
-+ struct fman_kg_regs *p_KgRegs;
-+
-+ uint32_t tmpKgarReg = 0, intFlags;
-+ t_Error err = E_OK;
-+
-+ /* The calling routine had locked the port, so for each port only one core can access
-+ * (so we don't need a lock here) */
-+
-+ if (p_FmPcd->h_Hc)
-+ return FmHcKgWriteSp(p_FmPcd->h_Hc, hardwarePortId, spReg, add);
-+
-+ p_KgRegs = p_FmPcd->p_FmPcdKg->p_FmPcdKgRegs;
-+
-+ tmpKgarReg = FmPcdKgBuildReadPortSchemeBindActionReg(hardwarePortId);
-+ /* lock a common KG reg */
-+ intFlags = KgHwLock(p_FmPcd->p_FmPcdKg);
-+ err = WriteKgarWait(p_FmPcd, tmpKgarReg);
-+ if (err)
-+ {
-+ KgHwUnlock(p_FmPcd->p_FmPcdKg, intFlags);
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+ }
-+
-+ fman_kg_write_sp(p_KgRegs, spReg, add);
-+
-+ tmpKgarReg = FmPcdKgBuildWritePortSchemeBindActionReg(hardwarePortId);
-+
-+ err = WriteKgarWait(p_FmPcd, tmpKgarReg);
-+ KgHwUnlock(p_FmPcd->p_FmPcdKg, intFlags);
-+ return err;
-+}
-+
-+static t_Error KgWriteCpp(t_FmPcd *p_FmPcd, uint8_t hardwarePortId, uint32_t cppReg)
-+{
-+ struct fman_kg_regs *p_KgRegs;
-+ uint32_t tmpKgarReg, intFlags;
-+ t_Error err;
-+
-+ p_KgRegs = p_FmPcd->p_FmPcdKg->p_FmPcdKgRegs;
-+
-+ if (p_FmPcd->h_Hc)
-+ {
-+ err = FmHcKgWriteCpp(p_FmPcd->h_Hc, hardwarePortId, cppReg);
-+ return err;
-+ }
-+
-+ intFlags = KgHwLock(p_FmPcd->p_FmPcdKg);
-+ fman_kg_write_cpp(p_KgRegs, cppReg);
-+ tmpKgarReg = FmPcdKgBuildWritePortClsPlanBindActionReg(hardwarePortId);
-+ err = WriteKgarWait(p_FmPcd, tmpKgarReg);
-+ KgHwUnlock(p_FmPcd->p_FmPcdKg, intFlags);
-+
-+ return err;
-+}
-+
-+static uint32_t BuildCppReg(t_FmPcd *p_FmPcd, uint8_t clsPlanGrpId)
-+{
-+ uint32_t tmpKgpeCpp;
-+
-+ tmpKgpeCpp = (uint32_t)(p_FmPcd->p_FmPcdKg->clsPlanGrps[clsPlanGrpId].baseEntry / 8);
-+ tmpKgpeCpp |= (uint32_t)(((p_FmPcd->p_FmPcdKg->clsPlanGrps[clsPlanGrpId].sizeOfGrp / 8) - 1) << FM_KG_PE_CPP_MASK_SHIFT);
-+
-+ return tmpKgpeCpp;
-+}
-+
-+static t_Error BindPortToClsPlanGrp(t_FmPcd *p_FmPcd, uint8_t hardwarePortId, uint8_t clsPlanGrpId)
-+{
-+ uint32_t tmpKgpeCpp = 0;
-+
-+ tmpKgpeCpp = BuildCppReg(p_FmPcd, clsPlanGrpId);
-+ return KgWriteCpp(p_FmPcd, hardwarePortId, tmpKgpeCpp);
-+}
-+
-+static void UnbindPortToClsPlanGrp(t_FmPcd *p_FmPcd, uint8_t hardwarePortId)
-+{
-+ KgWriteCpp(p_FmPcd, hardwarePortId, 0);
-+}
-+
-+#if (defined(DEBUG_ERRORS) && (DEBUG_ERRORS > 0))
-+static uint32_t __attribute__((unused)) ReadClsPlanBlockActionReg(uint8_t grpId)
-+{
-+ return (uint32_t)(FM_KG_KGAR_GO |
-+ FM_KG_KGAR_READ |
-+ FM_PCD_KG_KGAR_SEL_CLS_PLAN_ENTRY |
-+ DUMMY_PORT_ID |
-+ ((uint32_t)grpId << FM_PCD_KG_KGAR_NUM_SHIFT) |
-+ FM_PCD_KG_KGAR_WSEL_MASK);
-+
-+ /* if we ever want to write 1 by 1, use:
-+ sel = (uint8_t)(0x01 << (7- (entryId % CLS_PLAN_NUM_PER_GRP)));
-+ */
-+}
-+#endif /* (defined(DEBUG_ERRORS) && ... */
-+
-+static void PcdKgErrorException(t_Handle h_FmPcd)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd *)h_FmPcd;
-+ uint32_t event,schemeIndexes = 0, index = 0;
-+ struct fman_kg_regs *p_KgRegs;
-+
-+ ASSERT_COND(FmIsMaster(p_FmPcd->h_Fm));
-+ p_KgRegs = p_FmPcd->p_FmPcdKg->p_FmPcdKgRegs;
-+ fman_kg_get_event(p_KgRegs, &event, &schemeIndexes);
-+
-+ if (event & FM_EX_KG_DOUBLE_ECC)
-+ p_FmPcd->f_Exception(p_FmPcd->h_App,e_FM_PCD_KG_EXCEPTION_DOUBLE_ECC);
-+ if (event & FM_EX_KG_KEYSIZE_OVERFLOW)
-+ {
-+ if (schemeIndexes)
-+ {
-+ while (schemeIndexes)
-+ {
-+ if (schemeIndexes & 0x1)
-+ p_FmPcd->f_FmPcdIndexedException(p_FmPcd->h_App,e_FM_PCD_KG_EXCEPTION_KEYSIZE_OVERFLOW, (uint16_t)(31 - index));
-+ schemeIndexes >>= 1;
-+ index+=1;
-+ }
-+ }
-+ else /* this should happen only when interrupt is forced. */
-+ p_FmPcd->f_Exception(p_FmPcd->h_App,e_FM_PCD_KG_EXCEPTION_KEYSIZE_OVERFLOW);
-+ }
-+}
-+
-+static t_Error KgInitGuest(t_FmPcd *p_FmPcd)
-+{
-+ t_Error err = E_OK;
-+ t_FmPcdIpcKgSchemesParams kgAlloc;
-+ uint32_t replyLength;
-+ t_FmPcdIpcReply reply;
-+ t_FmPcdIpcMsg msg;
-+
-+ ASSERT_COND(p_FmPcd->guestId != NCSW_MASTER_ID);
-+
-+ /* in GUEST_PARTITION, we use the IPC */
-+ memset(&reply, 0, sizeof(reply));
-+ memset(&msg, 0, sizeof(msg));
-+ memset(&kgAlloc, 0, sizeof(t_FmPcdIpcKgSchemesParams));
-+ kgAlloc.numOfSchemes = p_FmPcd->p_FmPcdKg->numOfSchemes;
-+ kgAlloc.guestId = p_FmPcd->guestId;
-+ msg.msgId = FM_PCD_ALLOC_KG_SCHEMES;
-+ memcpy(msg.msgBody, &kgAlloc, sizeof(kgAlloc));
-+ replyLength = sizeof(uint32_t) + p_FmPcd->p_FmPcdKg->numOfSchemes*sizeof(uint8_t);
-+ if ((err = XX_IpcSendMessage(p_FmPcd->h_IpcSession,
-+ (uint8_t*)&msg,
-+ sizeof(msg.msgId) + sizeof(kgAlloc),
-+ (uint8_t*)&reply,
-+ &replyLength,
-+ NULL,
-+ NULL)) != E_OK)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ if (replyLength != (sizeof(uint32_t) + p_FmPcd->p_FmPcdKg->numOfSchemes*sizeof(uint8_t)))
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("IPC reply length mismatch"));
-+ memcpy(p_FmPcd->p_FmPcdKg->schemesIds, (uint8_t*)(reply.replyBody),p_FmPcd->p_FmPcdKg->numOfSchemes*sizeof(uint8_t));
-+
-+ return (t_Error)reply.error;
-+}
-+
-+static t_Error KgInitMaster(t_FmPcd *p_FmPcd)
-+{
-+ t_Error err = E_OK;
-+ struct fman_kg_regs *p_Regs = p_FmPcd->p_FmPcdKg->p_FmPcdKgRegs;
-+
-+ ASSERT_COND(p_FmPcd->guestId == NCSW_MASTER_ID);
-+
-+ if (p_FmPcd->exceptions & FM_EX_KG_DOUBLE_ECC)
-+ FmEnableRamsEcc(p_FmPcd->h_Fm);
-+
-+ fman_kg_init(p_Regs, p_FmPcd->exceptions, GET_NIA_BMI_AC_ENQ_FRAME(p_FmPcd));
-+
-+ /* register even if no interrupts enabled, to allow future enablement */
-+ FmRegisterIntr(p_FmPcd->h_Fm,
-+ e_FM_MOD_KG,
-+ 0,
-+ e_FM_INTR_TYPE_ERR,
-+ PcdKgErrorException,
-+ p_FmPcd);
-+
-+ fman_kg_enable_scheme_interrupts(p_Regs);
-+
-+ if (p_FmPcd->p_FmPcdKg->numOfSchemes)
-+ {
-+ err = FmPcdKgAllocSchemes(p_FmPcd,
-+ p_FmPcd->p_FmPcdKg->numOfSchemes,
-+ p_FmPcd->guestId,
-+ p_FmPcd->p_FmPcdKg->schemesIds);
-+ if (err)
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+ }
-+
-+ return E_OK;
-+}
-+
-+static void ValidateSchemeSw(t_FmPcdKgScheme *p_Scheme)
-+{
-+ ASSERT_COND(!p_Scheme->valid);
-+ if (p_Scheme->netEnvId != ILLEGAL_NETENV)
-+ FmPcdIncNetEnvOwners(p_Scheme->h_FmPcd, p_Scheme->netEnvId);
-+ p_Scheme->valid = TRUE;
-+}
-+
-+static t_Error InvalidateSchemeSw(t_FmPcdKgScheme *p_Scheme)
-+{
-+ if (p_Scheme->owners)
-+ RETURN_ERROR(MINOR, E_INVALID_STATE, ("Trying to delete a scheme that has ports bound to"));
-+
-+ if (p_Scheme->netEnvId != ILLEGAL_NETENV)
-+ FmPcdDecNetEnvOwners(p_Scheme->h_FmPcd, p_Scheme->netEnvId);
-+ p_Scheme->valid = FALSE;
-+
-+ return E_OK;
-+}
-+
-+static t_Error BuildSchemeRegs(t_FmPcdKgScheme *p_Scheme,
-+ t_FmPcdKgSchemeParams *p_SchemeParams,
-+ struct fman_kg_scheme_regs *p_SchemeRegs)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd *)(p_Scheme->h_FmPcd);
-+ uint32_t grpBits = 0;
-+ uint8_t grpBase;
-+ bool direct=TRUE, absolute=FALSE;
-+ uint16_t profileId=0, numOfProfiles=0, relativeProfileId;
-+ t_Error err = E_OK;
-+ int i = 0;
-+ t_NetEnvParams netEnvParams;
-+ uint32_t tmpReg, fqbTmp = 0, ppcTmp = 0, selectTmp, maskTmp, knownTmp, genTmp;
-+ t_FmPcdKgKeyExtractAndHashParams *p_KeyAndHash = NULL;
-+ uint8_t j, curr, idx;
-+ uint8_t id, shift=0, code=0, offset=0, size=0;
-+ t_FmPcdExtractEntry *p_Extract = NULL;
-+ t_FmPcdKgExtractedOrParams *p_ExtractOr;
-+ bool generic = FALSE;
-+ t_KnownFieldsMasks bitMask;
-+ e_FmPcdKgExtractDfltSelect swDefault = (e_FmPcdKgExtractDfltSelect)0;
-+ t_FmPcdKgSchemesExtracts *p_LocalExtractsArray;
-+ uint8_t numOfSwDefaults = 0;
-+ t_FmPcdKgExtractDflt swDefaults[NUM_OF_SW_DEFAULTS];
-+ uint8_t currGenId = 0;
-+
-+ memset(swDefaults, 0, NUM_OF_SW_DEFAULTS*sizeof(t_FmPcdKgExtractDflt));
-+ memset(p_SchemeRegs, 0, sizeof(struct fman_kg_scheme_regs));
-+
-+ if (p_SchemeParams->netEnvParams.numOfDistinctionUnits > FM_PCD_MAX_NUM_OF_DISTINCTION_UNITS)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE,
-+ ("numOfDistinctionUnits should not exceed %d", FM_PCD_MAX_NUM_OF_DISTINCTION_UNITS));
-+
-+ /* by netEnv parameters, get match vector */
-+ if (!p_SchemeParams->alwaysDirect)
-+ {
-+ p_Scheme->netEnvId = FmPcdGetNetEnvId(p_SchemeParams->netEnvParams.h_NetEnv);
-+ netEnvParams.netEnvId = p_Scheme->netEnvId;
-+ netEnvParams.numOfDistinctionUnits = p_SchemeParams->netEnvParams.numOfDistinctionUnits;
-+ memcpy(netEnvParams.unitIds, p_SchemeParams->netEnvParams.unitIds, (sizeof(uint8_t))*p_SchemeParams->netEnvParams.numOfDistinctionUnits);
-+ err = PcdGetUnitsVector(p_FmPcd, &netEnvParams);
-+ if (err)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, NO_MSG);
-+ p_Scheme->matchVector = netEnvParams.vector;
-+ }
-+ else
-+ {
-+ p_Scheme->matchVector = SCHEME_ALWAYS_DIRECT;
-+ p_Scheme->netEnvId = ILLEGAL_NETENV;
-+ }
-+
-+ if (p_SchemeParams->nextEngine == e_FM_PCD_INVALID)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("Next Engine of the scheme is not Valid"));
-+
-+ if (p_SchemeParams->bypassFqidGeneration)
-+ {
-+#ifdef FM_KG_NO_BYPASS_FQID_GEN
-+ if ((p_FmPcd->fmRevInfo.majorRev != 4) && (p_FmPcd->fmRevInfo.majorRev < 6))
-+ RETURN_ERROR(MAJOR, E_NOT_SUPPORTED, ("bypassFqidGeneration."));
-+#endif /* FM_KG_NO_BYPASS_FQID_GEN */
-+ if (p_SchemeParams->baseFqid)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("baseFqid set for a scheme that does not generate an FQID"));
-+ }
-+ else
-+ if (!p_SchemeParams->baseFqid)
-+ DBG(WARNING, ("baseFqid is 0."));
-+
-+ if (p_SchemeParams->nextEngine == e_FM_PCD_PLCR)
-+ {
-+ direct = p_SchemeParams->kgNextEngineParams.plcrProfile.direct;
-+ p_Scheme->directPlcr = direct;
-+ absolute = (bool)(p_SchemeParams->kgNextEngineParams.plcrProfile.sharedProfile ? TRUE : FALSE);
-+ if (!direct && absolute)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("Indirect policing is not available when profile is shared."));
-+
-+ if (direct)
-+ {
-+ profileId = p_SchemeParams->kgNextEngineParams.plcrProfile.profileSelect.directRelativeProfileId;
-+ numOfProfiles = 1;
-+ }
-+ else
-+ {
-+ profileId = p_SchemeParams->kgNextEngineParams.plcrProfile.profileSelect.indirectProfile.fqidOffsetRelativeProfileIdBase;
-+ shift = p_SchemeParams->kgNextEngineParams.plcrProfile.profileSelect.indirectProfile.fqidOffsetShift;
-+ numOfProfiles = p_SchemeParams->kgNextEngineParams.plcrProfile.profileSelect.indirectProfile.numOfProfiles;
-+ }
-+ }
-+
-+ if (p_SchemeParams->nextEngine == e_FM_PCD_CC)
-+ {
-+#ifdef FM_KG_NO_BYPASS_PLCR_PROFILE_GEN
-+ if ((p_SchemeParams->kgNextEngineParams.cc.plcrNext) && (p_SchemeParams->kgNextEngineParams.cc.bypassPlcrProfileGeneration))
-+ {
-+ if ((p_FmPcd->fmRevInfo.majorRev != 4) && (p_FmPcd->fmRevInfo.majorRev < 6))
-+ RETURN_ERROR(MAJOR, E_NOT_SUPPORTED, ("bypassPlcrProfileGeneration."));
-+ }
-+#endif /* FM_KG_NO_BYPASS_PLCR_PROFILE_GEN */
-+
-+ err = FmPcdCcGetGrpParams(p_SchemeParams->kgNextEngineParams.cc.h_CcTree,
-+ p_SchemeParams->kgNextEngineParams.cc.grpId,
-+ &grpBits,
-+ &grpBase);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ p_Scheme->ccUnits = grpBits;
-+
-+ if ((p_SchemeParams->kgNextEngineParams.cc.plcrNext) &&
-+ (!p_SchemeParams->kgNextEngineParams.cc.bypassPlcrProfileGeneration))
-+ {
-+ if (p_SchemeParams->kgNextEngineParams.cc.plcrProfile.sharedProfile)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("Shared profile may not be used after Coarse classification."));
-+ absolute = FALSE;
-+ direct = p_SchemeParams->kgNextEngineParams.cc.plcrProfile.direct;
-+ if (direct)
-+ {
-+ profileId = p_SchemeParams->kgNextEngineParams.cc.plcrProfile.profileSelect.directRelativeProfileId;
-+ numOfProfiles = 1;
-+ }
-+ else
-+ {
-+ profileId = p_SchemeParams->kgNextEngineParams.cc.plcrProfile.profileSelect.indirectProfile.fqidOffsetRelativeProfileIdBase;
-+ shift = p_SchemeParams->kgNextEngineParams.cc.plcrProfile.profileSelect.indirectProfile.fqidOffsetShift;
-+ numOfProfiles = p_SchemeParams->kgNextEngineParams.cc.plcrProfile.profileSelect.indirectProfile.numOfProfiles;
-+ }
-+ }
-+ }
-+
-+ /* if policer is used directly after KG, or after CC */
-+ if ((p_SchemeParams->nextEngine == e_FM_PCD_PLCR) ||
-+ ((p_SchemeParams->nextEngine == e_FM_PCD_CC) &&
-+ (p_SchemeParams->kgNextEngineParams.cc.plcrNext) &&
-+ (!p_SchemeParams->kgNextEngineParams.cc.bypassPlcrProfileGeneration)))
-+ {
-+ /* if private policer profile, it may be uninitialized yet, therefore no checks are done at this stage */
-+ if (absolute)
-+ {
-+ /* for absolute direct policy only, */
-+ relativeProfileId = profileId;
-+ err = FmPcdPlcrGetAbsoluteIdByProfileParams((t_Handle)p_FmPcd,e_FM_PCD_PLCR_SHARED,NULL, relativeProfileId, &profileId);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, ("Shared profile not valid offset"));
-+ if (!FmPcdPlcrIsProfileValid(p_FmPcd, profileId))
-+ RETURN_ERROR(MINOR, E_INVALID_STATE, ("Shared profile not valid."));
-+ p_Scheme->relativeProfileId = profileId;
-+ }
-+ else
-+ {
-+ /* save relative profile id's for later check */
-+ p_Scheme->nextRelativePlcrProfile = TRUE;
-+ p_Scheme->relativeProfileId = profileId;
-+ p_Scheme->numOfProfiles = numOfProfiles;
-+ }
-+ }
-+ else
-+ {
-+ /* if policer is NOT going to be used after KG at all than if bypassFqidGeneration
-+ is set, we do not need numOfUsedExtractedOrs and hashDistributionNumOfFqids */
-+ if (p_SchemeParams->bypassFqidGeneration && p_SchemeParams->numOfUsedExtractedOrs)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE,
-+ ("numOfUsedExtractedOrs is set in a scheme that does not generate FQID or policer profile ID"));
-+ if (p_SchemeParams->bypassFqidGeneration &&
-+ p_SchemeParams->useHash &&
-+ p_SchemeParams->keyExtractAndHashParams.hashDistributionNumOfFqids)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE,
-+ ("hashDistributionNumOfFqids is set in a scheme that does not generate FQID or policer profile ID"));
-+ }
-+
-+ /* configure all 21 scheme registers */
-+ tmpReg = KG_SCH_MODE_EN;
-+ switch (p_SchemeParams->nextEngine)
-+ {
-+ case (e_FM_PCD_PLCR):
-+ /* add to mode register - NIA */
-+ tmpReg |= KG_SCH_MODE_NIA_PLCR;
-+ tmpReg |= NIA_ENG_PLCR;
-+ tmpReg |= (uint32_t)(p_SchemeParams->kgNextEngineParams.plcrProfile.sharedProfile ? NIA_PLCR_ABSOLUTE:0);
-+ /* initialize policer profile command - */
-+ /* configure kgse_ppc */
-+ if (direct)
-+ /* use profileId as base, other fields are 0 */
-+ p_SchemeRegs->kgse_ppc = (uint32_t)profileId;
-+ else
-+ {
-+ if (shift > MAX_PP_SHIFT)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("fqidOffsetShift may not be larger than %d", MAX_PP_SHIFT));
-+
-+ if (!numOfProfiles || !POWER_OF_2(numOfProfiles))
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("numOfProfiles must not be 0 and must be a power of 2"));
-+
-+ ppcTmp = ((uint32_t)shift << KG_SCH_PP_SHIFT_HIGH_SHIFT) & KG_SCH_PP_SHIFT_HIGH;
-+ ppcTmp |= ((uint32_t)shift << KG_SCH_PP_SHIFT_LOW_SHIFT) & KG_SCH_PP_SHIFT_LOW;
-+ ppcTmp |= ((uint32_t)(numOfProfiles-1) << KG_SCH_PP_MASK_SHIFT);
-+ ppcTmp |= (uint32_t)profileId;
-+
-+ p_SchemeRegs->kgse_ppc = ppcTmp;
-+ }
-+ break;
-+ case (e_FM_PCD_CC):
-+ /* mode reg - define NIA */
-+ tmpReg |= (NIA_ENG_FM_CTL | NIA_FM_CTL_AC_CC);
-+
-+ p_SchemeRegs->kgse_ccbs = grpBits;
-+ tmpReg |= (uint32_t)(grpBase << KG_SCH_MODE_CCOBASE_SHIFT);
-+
-+ if (p_SchemeParams->kgNextEngineParams.cc.plcrNext)
-+ {
-+ if (!p_SchemeParams->kgNextEngineParams.cc.bypassPlcrProfileGeneration)
-+ {
-+ /* find out if absolute or relative */
-+ if (absolute)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("It is illegal to request a shared profile in a scheme that is in a KG->CC->PLCR flow"));
-+ if (direct)
-+ {
-+ /* mask = 0, base = directProfileId */
-+ p_SchemeRegs->kgse_ppc = (uint32_t)profileId;
-+ }
-+ else
-+ {
-+ if (shift > MAX_PP_SHIFT)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("fqidOffsetShift may not be larger than %d", MAX_PP_SHIFT));
-+ if (!numOfProfiles || !POWER_OF_2(numOfProfiles))
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("numOfProfiles must not be 0 and must be a power of 2"));
-+
-+ ppcTmp = ((uint32_t)shift << KG_SCH_PP_SHIFT_HIGH_SHIFT) & KG_SCH_PP_SHIFT_HIGH;
-+ ppcTmp |= ((uint32_t)shift << KG_SCH_PP_SHIFT_LOW_SHIFT) & KG_SCH_PP_SHIFT_LOW;
-+ ppcTmp |= ((uint32_t)(numOfProfiles-1) << KG_SCH_PP_MASK_SHIFT);
-+ ppcTmp |= (uint32_t)profileId;
-+
-+ p_SchemeRegs->kgse_ppc = ppcTmp;
-+ }
-+ }
-+ }
-+ break;
-+ case (e_FM_PCD_DONE):
-+ if (p_SchemeParams->kgNextEngineParams.doneAction == e_FM_PCD_DROP_FRAME)
-+ tmpReg |= GET_NIA_BMI_AC_DISCARD_FRAME(p_FmPcd);
-+ else
-+ tmpReg |= GET_NIA_BMI_AC_ENQ_FRAME(p_FmPcd);
-+ break;
-+ default:
-+ RETURN_ERROR(MAJOR, E_NOT_SUPPORTED, ("Next engine not supported"));
-+ }
-+ p_SchemeRegs->kgse_mode = tmpReg;
-+
-+ p_SchemeRegs->kgse_mv = p_Scheme->matchVector;
-+
-+#if (DPAA_VERSION >= 11)
-+ if (p_SchemeParams->overrideStorageProfile)
-+ {
-+ p_SchemeRegs->kgse_om |= KG_SCH_OM_VSPE;
-+
-+ if (p_SchemeParams->storageProfile.direct)
-+ {
-+ profileId = p_SchemeParams->storageProfile.profileSelect.directRelativeProfileId;
-+ shift = 0;
-+ numOfProfiles = 1;
-+ }
-+ else
-+ {
-+ profileId = p_SchemeParams->storageProfile.profileSelect.indirectProfile.fqidOffsetRelativeProfileIdBase;
-+ shift = p_SchemeParams->storageProfile.profileSelect.indirectProfile.fqidOffsetShift;
-+ numOfProfiles = p_SchemeParams->storageProfile.profileSelect.indirectProfile.numOfProfiles;
-+ }
-+ if (shift > MAX_SP_SHIFT)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("fqidOffsetShift may not be larger than %d", MAX_SP_SHIFT));
-+
-+ if (!numOfProfiles || !POWER_OF_2(numOfProfiles))
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("numOfProfiles must not be 0 and must be a power of 2"));
-+
-+ tmpReg = (uint32_t)shift << KG_SCH_VSP_SHIFT;
-+ tmpReg |= ((uint32_t)(numOfProfiles-1) << KG_SCH_VSP_MASK_SHIFT);
-+ tmpReg |= (uint32_t)profileId;
-+
-+
-+ p_SchemeRegs->kgse_vsp = tmpReg;
-+
-+ p_Scheme->vspe = TRUE;
-+
-+ }
-+ else
-+ p_SchemeRegs->kgse_vsp = KG_SCH_VSP_NO_KSP_EN;
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+ if (p_SchemeParams->useHash)
-+ {
-+ p_KeyAndHash = &p_SchemeParams->keyExtractAndHashParams;
-+
-+ if (p_KeyAndHash->numOfUsedExtracts >= FM_PCD_KG_MAX_NUM_OF_EXTRACTS_PER_KEY)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("numOfUsedExtracts out of range"));
-+
-+ /* configure kgse_dv0 */
-+ p_SchemeRegs->kgse_dv0 = p_KeyAndHash->privateDflt0;
-+
-+ /* configure kgse_dv1 */
-+ p_SchemeRegs->kgse_dv1 = p_KeyAndHash->privateDflt1;
-+
-+ if (!p_SchemeParams->bypassFqidGeneration)
-+ {
-+ if (!p_KeyAndHash->hashDistributionNumOfFqids || !POWER_OF_2(p_KeyAndHash->hashDistributionNumOfFqids))
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("hashDistributionNumOfFqids must not be 0 and must be a power of 2"));
-+ if ((p_KeyAndHash->hashDistributionNumOfFqids-1) & p_SchemeParams->baseFqid)
-+ DBG(WARNING, ("baseFqid unaligned. Distribution may result in less than hashDistributionNumOfFqids queues."));
-+ }
-+
-+ /* configure kgse_ekdv */
-+ tmpReg = 0;
-+ for ( i=0 ;i<p_KeyAndHash->numOfUsedDflts ; i++)
-+ {
-+ switch (p_KeyAndHash->dflts[i].type)
-+ {
-+ case (e_FM_PCD_KG_MAC_ADDR):
-+ tmpReg |= (p_KeyAndHash->dflts[i].dfltSelect << KG_SCH_DEF_MAC_ADDR_SHIFT);
-+ break;
-+ case (e_FM_PCD_KG_TCI):
-+ tmpReg |= (p_KeyAndHash->dflts[i].dfltSelect << KG_SCH_DEF_TCI_SHIFT);
-+ break;
-+ case (e_FM_PCD_KG_ENET_TYPE):
-+ tmpReg |= (p_KeyAndHash->dflts[i].dfltSelect << KG_SCH_DEF_ENET_TYPE_SHIFT);
-+ break;
-+ case (e_FM_PCD_KG_PPP_SESSION_ID):
-+ tmpReg |= (p_KeyAndHash->dflts[i].dfltSelect << KG_SCH_DEF_PPP_SESSION_ID_SHIFT);
-+ break;
-+ case (e_FM_PCD_KG_PPP_PROTOCOL_ID):
-+ tmpReg |= (p_KeyAndHash->dflts[i].dfltSelect << KG_SCH_DEF_PPP_PROTOCOL_ID_SHIFT);
-+ break;
-+ case (e_FM_PCD_KG_MPLS_LABEL):
-+ tmpReg |= (p_KeyAndHash->dflts[i].dfltSelect << KG_SCH_DEF_MPLS_LABEL_SHIFT);
-+ break;
-+ case (e_FM_PCD_KG_IP_ADDR):
-+ tmpReg |= (p_KeyAndHash->dflts[i].dfltSelect << KG_SCH_DEF_IP_ADDR_SHIFT);
-+ break;
-+ case (e_FM_PCD_KG_PROTOCOL_TYPE):
-+ tmpReg |= (p_KeyAndHash->dflts[i].dfltSelect << KG_SCH_DEF_PROTOCOL_TYPE_SHIFT);
-+ break;
-+ case (e_FM_PCD_KG_IP_TOS_TC):
-+ tmpReg |= (p_KeyAndHash->dflts[i].dfltSelect << KG_SCH_DEF_IP_TOS_TC_SHIFT);
-+ break;
-+ case (e_FM_PCD_KG_IPV6_FLOW_LABEL):
-+ tmpReg |= (p_KeyAndHash->dflts[i].dfltSelect << KG_SCH_DEF_L4_PORT_SHIFT);
-+ break;
-+ case (e_FM_PCD_KG_IPSEC_SPI):
-+ tmpReg |= (p_KeyAndHash->dflts[i].dfltSelect << KG_SCH_DEF_IPSEC_SPI_SHIFT);
-+ break;
-+ case (e_FM_PCD_KG_L4_PORT):
-+ tmpReg |= (p_KeyAndHash->dflts[i].dfltSelect << KG_SCH_DEF_L4_PORT_SHIFT);
-+ break;
-+ case (e_FM_PCD_KG_TCP_FLAG):
-+ tmpReg |= (p_KeyAndHash->dflts[i].dfltSelect << KG_SCH_DEF_TCP_FLAG_SHIFT);
-+ break;
-+ case (e_FM_PCD_KG_GENERIC_FROM_DATA):
-+ swDefaults[numOfSwDefaults].type = e_FM_PCD_KG_GENERIC_FROM_DATA;
-+ swDefaults[numOfSwDefaults].dfltSelect = p_KeyAndHash->dflts[i].dfltSelect;
-+ numOfSwDefaults ++;
-+ break;
-+ case (e_FM_PCD_KG_GENERIC_FROM_DATA_NO_V):
-+ swDefaults[numOfSwDefaults].type = e_FM_PCD_KG_GENERIC_FROM_DATA_NO_V;
-+ swDefaults[numOfSwDefaults].dfltSelect = p_KeyAndHash->dflts[i].dfltSelect;
-+ numOfSwDefaults ++;
-+ break;
-+ case (e_FM_PCD_KG_GENERIC_NOT_FROM_DATA):
-+ swDefaults[numOfSwDefaults].type = e_FM_PCD_KG_GENERIC_NOT_FROM_DATA;
-+ swDefaults[numOfSwDefaults].dfltSelect = p_KeyAndHash->dflts[i].dfltSelect;
-+ numOfSwDefaults ++;
-+ break;
-+ default:
-+ RETURN_ERROR(MAJOR, E_INVALID_SELECTION, NO_MSG);
-+ }
-+ }
-+ p_SchemeRegs->kgse_ekdv = tmpReg;
-+
-+ p_LocalExtractsArray = (t_FmPcdKgSchemesExtracts *)XX_Malloc(sizeof(t_FmPcdKgSchemesExtracts));
-+ if (!p_LocalExtractsArray)
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, ("No memory"));
-+
-+ /* configure kgse_ekfc and kgse_gec */
-+ knownTmp = 0;
-+ for ( i=0 ;i<p_KeyAndHash->numOfUsedExtracts ; i++)
-+ {
-+ p_Extract = &p_KeyAndHash->extractArray[i];
-+ switch (p_Extract->type)
-+ {
-+ case (e_FM_PCD_KG_EXTRACT_PORT_PRIVATE_INFO):
-+ knownTmp |= KG_SCH_KN_PORT_ID;
-+ /* save in driver structure */
-+ p_LocalExtractsArray->extractsArray[i].id = GetKnownFieldId(KG_SCH_KN_PORT_ID);
-+ p_LocalExtractsArray->extractsArray[i].known = TRUE;
-+ break;
-+ case (e_FM_PCD_EXTRACT_BY_HDR):
-+ switch (p_Extract->extractByHdr.hdr)
-+ {
-+#if (DPAA_VERSION >= 11) || ((DPAA_VERSION == 10) && defined(FM_CAPWAP_SUPPORT))
-+ case (HEADER_TYPE_UDP_LITE):
-+ p_Extract->extractByHdr.hdr = HEADER_TYPE_UDP;
-+ break;
-+#endif /* (DPAA_VERSION >= 11) || ((DPAA_VERSION == 10) && defined(FM_CAPWAP_SUPPORT)) */
-+ case (HEADER_TYPE_UDP_ENCAP_ESP):
-+ switch (p_Extract->extractByHdr.type)
-+ {
-+ case (e_FM_PCD_EXTRACT_FROM_HDR):
-+ /* case where extraction from ESP only */
-+ if (p_Extract->extractByHdr.extractByHdrType.fromHdr.offset >= UDP_HEADER_SIZE)
-+ {
-+ p_Extract->extractByHdr.hdr = FmPcdGetAliasHdr(p_FmPcd, p_Scheme->netEnvId, HEADER_TYPE_UDP_ENCAP_ESP);
-+ p_Extract->extractByHdr.extractByHdrType.fromHdr.offset -= UDP_HEADER_SIZE;
-+ p_Extract->extractByHdr.ignoreProtocolValidation = TRUE;
-+ }
-+ else
-+ {
-+ p_Extract->extractByHdr.hdr = HEADER_TYPE_UDP;
-+ p_Extract->extractByHdr.ignoreProtocolValidation = FALSE;
-+ }
-+ break;
-+ case (e_FM_PCD_EXTRACT_FROM_FIELD):
-+ switch (p_Extract->extractByHdr.extractByHdrType.fromField.field.udpEncapEsp)
-+ {
-+ case (NET_HEADER_FIELD_UDP_ENCAP_ESP_PORT_SRC):
-+ case (NET_HEADER_FIELD_UDP_ENCAP_ESP_PORT_DST):
-+ case (NET_HEADER_FIELD_UDP_ENCAP_ESP_LEN):
-+ case (NET_HEADER_FIELD_UDP_ENCAP_ESP_CKSUM):
-+ p_Extract->extractByHdr.hdr = HEADER_TYPE_UDP;
-+ break;
-+ case (NET_HEADER_FIELD_UDP_ENCAP_ESP_SPI):
-+ p_Extract->extractByHdr.type = e_FM_PCD_EXTRACT_FROM_HDR;
-+ p_Extract->extractByHdr.hdr = FmPcdGetAliasHdr(p_FmPcd, p_Scheme->netEnvId, HEADER_TYPE_UDP_ENCAP_ESP);
-+ /*p_Extract->extractByHdr.extractByHdrType.fromField.offset += ESP_SPI_OFFSET;*/
-+ p_Extract->extractByHdr.ignoreProtocolValidation = TRUE;
-+ break;
-+ case (NET_HEADER_FIELD_UDP_ENCAP_ESP_SEQUENCE_NUM):
-+ p_Extract->extractByHdr.type = e_FM_PCD_EXTRACT_FROM_HDR;
-+ p_Extract->extractByHdr.hdr = FmPcdGetAliasHdr(p_FmPcd, p_Scheme->netEnvId, HEADER_TYPE_UDP_ENCAP_ESP);
-+ p_Extract->extractByHdr.extractByHdrType.fromField.offset += ESP_SEQ_NUM_OFFSET;
-+ p_Extract->extractByHdr.ignoreProtocolValidation = TRUE;
-+ break;
-+ }
-+ break;
-+ case (e_FM_PCD_EXTRACT_FULL_FIELD):
-+ switch (p_Extract->extractByHdr.extractByHdrType.fullField.udpEncapEsp)
-+ {
-+ case (NET_HEADER_FIELD_UDP_ENCAP_ESP_PORT_SRC):
-+ case (NET_HEADER_FIELD_UDP_ENCAP_ESP_PORT_DST):
-+ case (NET_HEADER_FIELD_UDP_ENCAP_ESP_LEN):
-+ case (NET_HEADER_FIELD_UDP_ENCAP_ESP_CKSUM):
-+ p_Extract->extractByHdr.hdr = HEADER_TYPE_UDP;
-+ break;
-+ case (NET_HEADER_FIELD_UDP_ENCAP_ESP_SPI):
-+ p_Extract->extractByHdr.type = e_FM_PCD_EXTRACT_FROM_HDR;
-+ p_Extract->extractByHdr.hdr = FmPcdGetAliasHdr(p_FmPcd, p_Scheme->netEnvId, HEADER_TYPE_UDP_ENCAP_ESP);
-+ p_Extract->extractByHdr.extractByHdrType.fromHdr.size = ESP_SPI_SIZE;
-+ p_Extract->extractByHdr.extractByHdrType.fromHdr.offset = ESP_SPI_OFFSET;
-+ p_Extract->extractByHdr.ignoreProtocolValidation = TRUE;
-+ break;
-+ case (NET_HEADER_FIELD_UDP_ENCAP_ESP_SEQUENCE_NUM):
-+ p_Extract->extractByHdr.type = e_FM_PCD_EXTRACT_FROM_HDR;
-+ p_Extract->extractByHdr.hdr = FmPcdGetAliasHdr(p_FmPcd, p_Scheme->netEnvId, HEADER_TYPE_UDP_ENCAP_ESP);
-+ p_Extract->extractByHdr.extractByHdrType.fromHdr.size = ESP_SEQ_NUM_SIZE;
-+ p_Extract->extractByHdr.extractByHdrType.fromHdr.offset = ESP_SEQ_NUM_OFFSET;
-+ p_Extract->extractByHdr.ignoreProtocolValidation = TRUE;
-+ break;
-+ }
-+ break;
-+ }
-+ break;
-+ default:
-+ break;
-+ }
-+ switch (p_Extract->extractByHdr.type)
-+ {
-+ case (e_FM_PCD_EXTRACT_FROM_HDR):
-+ generic = TRUE;
-+ /* get the header code for the generic extract */
-+ code = GetGenHdrCode(p_Extract->extractByHdr.hdr, p_Extract->extractByHdr.hdrIndex, p_Extract->extractByHdr.ignoreProtocolValidation);
-+ /* set generic register fields */
-+ offset = p_Extract->extractByHdr.extractByHdrType.fromHdr.offset;
-+ size = p_Extract->extractByHdr.extractByHdrType.fromHdr.size;
-+ break;
-+ case (e_FM_PCD_EXTRACT_FROM_FIELD):
-+ generic = TRUE;
-+ /* get the field code for the generic extract */
-+ code = GetGenFieldCode(p_Extract->extractByHdr.hdr,
-+ p_Extract->extractByHdr.extractByHdrType.fromField.field, p_Extract->extractByHdr.ignoreProtocolValidation,p_Extract->extractByHdr.hdrIndex);
-+ offset = p_Extract->extractByHdr.extractByHdrType.fromField.offset;
-+ size = p_Extract->extractByHdr.extractByHdrType.fromField.size;
-+ break;
-+ case (e_FM_PCD_EXTRACT_FULL_FIELD):
-+ if (!p_Extract->extractByHdr.ignoreProtocolValidation)
-+ {
-+ /* if we have a known field for it - use it, otherwise use generic */
-+ bitMask = GetKnownProtMask(p_FmPcd, p_Extract->extractByHdr.hdr, p_Extract->extractByHdr.hdrIndex,
-+ p_Extract->extractByHdr.extractByHdrType.fullField);
-+ if (bitMask)
-+ {
-+ knownTmp |= bitMask;
-+ /* save in driver structure */
-+ p_LocalExtractsArray->extractsArray[i].id = GetKnownFieldId(bitMask);
-+ p_LocalExtractsArray->extractsArray[i].known = TRUE;
-+ }
-+ else
-+ generic = TRUE;
-+ }
-+ else
-+ generic = TRUE;
-+ if (generic)
-+ {
-+ /* tmp - till we cover more headers under generic */
-+ XX_Free(p_LocalExtractsArray);
-+ RETURN_ERROR(MAJOR, E_NOT_SUPPORTED, ("Full header selection not supported"));
-+ }
-+ break;
-+ default:
-+ XX_Free(p_LocalExtractsArray);
-+ RETURN_ERROR(MAJOR, E_INVALID_SELECTION, NO_MSG);
-+ }
-+ break;
-+ case (e_FM_PCD_EXTRACT_NON_HDR):
-+ /* use generic */
-+ generic = TRUE;
-+ offset = 0;
-+ /* get the field code for the generic extract */
-+ code = GetGenCode(p_Extract->extractNonHdr.src, &offset);
-+ offset += p_Extract->extractNonHdr.offset;
-+ size = p_Extract->extractNonHdr.size;
-+ break;
-+ default:
-+ RETURN_ERROR(MAJOR, E_INVALID_SELECTION, NO_MSG);
-+ }
-+
-+ if (generic)
-+ {
-+ /* set generic register fields */
-+ if (currGenId >= FM_KG_NUM_OF_GENERIC_REGS)
-+ {
-+ XX_Free(p_LocalExtractsArray);
-+ RETURN_ERROR(MAJOR, E_FULL, ("Generic registers are fully used"));
-+ }
-+ if (!code)
-+ {
-+ XX_Free(p_LocalExtractsArray);
-+ RETURN_ERROR(MAJOR, E_NOT_SUPPORTED, NO_MSG);
-+ }
-+
-+ genTmp = KG_SCH_GEN_VALID;
-+ genTmp |= (uint32_t)(code << KG_SCH_GEN_HT_SHIFT);
-+ genTmp |= offset;
-+ if ((size > MAX_KG_SCH_SIZE) || (size < 1))
-+ {
-+ XX_Free(p_LocalExtractsArray);
-+ RETURN_ERROR(MAJOR, E_NOT_SUPPORTED, ("Illegal extraction (size out of range)"));
-+ }
-+ genTmp |= (uint32_t)((size - 1) << KG_SCH_GEN_SIZE_SHIFT);
-+ swDefault = GetGenericSwDefault(swDefaults, numOfSwDefaults, code);
-+ if (swDefault == e_FM_PCD_KG_DFLT_ILLEGAL)
-+ DBG(WARNING, ("No sw default configured"));
-+ else
-+ genTmp |= swDefault << KG_SCH_GEN_DEF_SHIFT;
-+
-+ genTmp |= KG_SCH_GEN_MASK;
-+ p_SchemeRegs->kgse_gec[currGenId] = genTmp;
-+ /* save in driver structure */
-+ p_LocalExtractsArray->extractsArray[i].id = currGenId++;
-+ p_LocalExtractsArray->extractsArray[i].known = FALSE;
-+ generic = FALSE;
-+ }
-+ }
-+ p_SchemeRegs->kgse_ekfc = knownTmp;
-+
-+ selectTmp = 0;
-+ maskTmp = 0xFFFFFFFF;
-+ /* configure kgse_bmch, kgse_bmcl and kgse_fqb */
-+
-+ if (p_KeyAndHash->numOfUsedMasks > FM_PCD_KG_NUM_OF_EXTRACT_MASKS)
-+ {
-+ XX_Free(p_LocalExtractsArray);
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("Only %d masks supported", FM_PCD_KG_NUM_OF_EXTRACT_MASKS));
-+ }
-+ for ( i=0 ;i<p_KeyAndHash->numOfUsedMasks ; i++)
-+ {
-+ /* Get the relative id of the extract (for known 0-0x1f, for generic 0-7) */
-+ id = p_LocalExtractsArray->extractsArray[p_KeyAndHash->masks[i].extractArrayIndex].id;
-+ /* Get the shift of the select field (depending on i) */
-+ GET_MASK_SEL_SHIFT(shift,i);
-+ if (p_LocalExtractsArray->extractsArray[p_KeyAndHash->masks[i].extractArrayIndex].known)
-+ selectTmp |= id << shift;
-+ else
-+ selectTmp |= (id + MASK_FOR_GENERIC_BASE_ID) << shift;
-+
-+ /* Get the shift of the offset field (depending on i) - may
-+ be in kgse_bmch or in kgse_fqb (depending on i) */
-+ GET_MASK_OFFSET_SHIFT(shift,i);
-+ if (i<=1)
-+ selectTmp |= p_KeyAndHash->masks[i].offset << shift;
-+ else
-+ fqbTmp |= p_KeyAndHash->masks[i].offset << shift;
-+
-+ /* Get the shift of the mask field (depending on i) */
-+ GET_MASK_SHIFT(shift,i);
-+ /* pass all bits */
-+ maskTmp |= KG_SCH_BITMASK_MASK << shift;
-+ /* clear bits that need masking */
-+ maskTmp &= ~(0xFF << shift) ;
-+ /* set mask bits */
-+ maskTmp |= (p_KeyAndHash->masks[i].mask << shift) ;
-+ }
-+ p_SchemeRegs->kgse_bmch = selectTmp;
-+ p_SchemeRegs->kgse_bmcl = maskTmp;
-+ /* kgse_fqb will be written t the end of the routine */
-+
-+ /* configure kgse_hc */
-+ if (p_KeyAndHash->hashShift > MAX_HASH_SHIFT)
-+ {
-+ XX_Free(p_LocalExtractsArray);
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("hashShift must not be larger than %d", MAX_HASH_SHIFT));
-+ }
-+ if (p_KeyAndHash->hashDistributionFqidsShift > MAX_DIST_FQID_SHIFT)
-+ {
-+ XX_Free(p_LocalExtractsArray);
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("hashDistributionFqidsShift must not be larger than %d", MAX_DIST_FQID_SHIFT));
-+ }
-+
-+ tmpReg = 0;
-+
-+ tmpReg |= ((p_KeyAndHash->hashDistributionNumOfFqids - 1) << p_KeyAndHash->hashDistributionFqidsShift);
-+ tmpReg |= p_KeyAndHash->hashShift << KG_SCH_HASH_CONFIG_SHIFT_SHIFT;
-+
-+ if (p_KeyAndHash->symmetricHash)
-+ {
-+ if ((!!(p_SchemeRegs->kgse_ekfc & KG_SCH_KN_MACSRC) != !!(p_SchemeRegs->kgse_ekfc & KG_SCH_KN_MACDST)) ||
-+ (!!(p_SchemeRegs->kgse_ekfc & KG_SCH_KN_IPSRC1) != !!(p_SchemeRegs->kgse_ekfc & KG_SCH_KN_IPDST1)) ||
-+ (!!(p_SchemeRegs->kgse_ekfc & KG_SCH_KN_IPSRC2) != !!(p_SchemeRegs->kgse_ekfc & KG_SCH_KN_IPDST2)) ||
-+ (!!(p_SchemeRegs->kgse_ekfc & KG_SCH_KN_L4PSRC) != !!(p_SchemeRegs->kgse_ekfc & KG_SCH_KN_L4PDST)))
-+ {
-+ XX_Free(p_LocalExtractsArray);
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("symmetricHash set but src/dest extractions missing"));
-+ }
-+ tmpReg |= KG_SCH_HASH_CONFIG_SYM;
-+ }
-+ p_SchemeRegs->kgse_hc = tmpReg;
-+
-+ /* build the return array describing the order of the extractions */
-+
-+ /* the last currGenId places of the array
-+ are for generic extracts that are always last.
-+ We now sort for the calculation of the order of the known
-+ extractions we sort the known extracts between orderedArray[0] and
-+ orderedArray[p_KeyAndHash->numOfUsedExtracts - currGenId - 1].
-+ for the calculation of the order of the generic extractions we use:
-+ num_of_generic - currGenId
-+ num_of_known - p_KeyAndHash->numOfUsedExtracts - currGenId
-+ first_generic_index = num_of_known */
-+ curr = 0;
-+ for (i=0;i<p_KeyAndHash->numOfUsedExtracts ; i++)
-+ {
-+ if (p_LocalExtractsArray->extractsArray[i].known)
-+ {
-+ ASSERT_COND(curr<(p_KeyAndHash->numOfUsedExtracts - currGenId));
-+ j = curr;
-+ /* id is the extract id (port id = 0, mac src = 1 etc.). the value in the array is the original
-+ index in the user's extractions array */
-+ /* we compare the id of the current extract with the id of the extract in the orderedArray[j-1]
-+ location */
-+ while ((j > 0) && (p_LocalExtractsArray->extractsArray[i].id <
-+ p_LocalExtractsArray->extractsArray[p_Scheme->orderedArray[j-1]].id))
-+ {
-+ p_Scheme->orderedArray[j] =
-+ p_Scheme->orderedArray[j-1];
-+ j--;
-+ }
-+ p_Scheme->orderedArray[j] = (uint8_t)i;
-+ curr++;
-+ }
-+ else
-+ {
-+ /* index is first_generic_index + generic index (id) */
-+ idx = (uint8_t)(p_KeyAndHash->numOfUsedExtracts - currGenId + p_LocalExtractsArray->extractsArray[i].id);
-+ ASSERT_COND(idx < FM_PCD_KG_MAX_NUM_OF_EXTRACTS_PER_KEY);
-+ p_Scheme->orderedArray[idx]= (uint8_t)i;
-+ }
-+ }
-+ XX_Free(p_LocalExtractsArray);
-+ }
-+ else
-+ {
-+ /* clear all unused registers: */
-+ p_SchemeRegs->kgse_ekfc = 0;
-+ p_SchemeRegs->kgse_ekdv = 0;
-+ p_SchemeRegs->kgse_bmch = 0;
-+ p_SchemeRegs->kgse_bmcl = 0;
-+ p_SchemeRegs->kgse_hc = 0;
-+ p_SchemeRegs->kgse_dv0 = 0;
-+ p_SchemeRegs->kgse_dv1 = 0;
-+ }
-+
-+ if (p_SchemeParams->bypassFqidGeneration)
-+ p_SchemeRegs->kgse_hc |= KG_SCH_HASH_CONFIG_NO_FQID;
-+
-+ /* configure kgse_spc */
-+ if ( p_SchemeParams->schemeCounter.update)
-+ p_SchemeRegs->kgse_spc = p_SchemeParams->schemeCounter.value;
-+
-+
-+ /* check that are enough generic registers */
-+ if (p_SchemeParams->numOfUsedExtractedOrs + currGenId > FM_KG_NUM_OF_GENERIC_REGS)
-+ RETURN_ERROR(MAJOR, E_FULL, ("Generic registers are fully used"));
-+
-+ /* extracted OR mask on Qid */
-+ for ( i=0 ;i<p_SchemeParams->numOfUsedExtractedOrs ; i++)
-+ {
-+
-+ p_Scheme->extractedOrs = TRUE;
-+ /* configure kgse_gec[i] */
-+ p_ExtractOr = &p_SchemeParams->extractedOrs[i];
-+ switch (p_ExtractOr->type)
-+ {
-+ case (e_FM_PCD_KG_EXTRACT_PORT_PRIVATE_INFO):
-+ code = KG_SCH_GEN_PARSE_RESULT_N_FQID;
-+ offset = 0;
-+ break;
-+ case (e_FM_PCD_EXTRACT_BY_HDR):
-+ /* get the header code for the generic extract */
-+ code = GetGenHdrCode(p_ExtractOr->extractByHdr.hdr, p_ExtractOr->extractByHdr.hdrIndex, p_ExtractOr->extractByHdr.ignoreProtocolValidation);
-+ /* set generic register fields */
-+ offset = p_ExtractOr->extractionOffset;
-+ break;
-+ case (e_FM_PCD_EXTRACT_NON_HDR):
-+ /* get the field code for the generic extract */
-+ offset = 0;
-+ code = GetGenCode(p_ExtractOr->src, &offset);
-+ offset += p_ExtractOr->extractionOffset;
-+ break;
-+ default:
-+ RETURN_ERROR(MAJOR, E_INVALID_SELECTION, NO_MSG);
-+ }
-+
-+ /* set generic register fields */
-+ if (!code)
-+ RETURN_ERROR(MAJOR, E_NOT_SUPPORTED, NO_MSG);
-+ genTmp = KG_SCH_GEN_EXTRACT_TYPE | KG_SCH_GEN_VALID;
-+ genTmp |= (uint32_t)(code << KG_SCH_GEN_HT_SHIFT);
-+ genTmp |= offset;
-+ if (!!p_ExtractOr->bitOffsetInFqid == !!p_ExtractOr->bitOffsetInPlcrProfile)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, (" extracted byte must effect either FQID or Policer profile"));
-+
-+ /************************************************************************************
-+ bitOffsetInFqid and bitOffsetInPolicerProfile are translated to rotate parameter
-+ in the following way:
-+
-+ Driver API and implementation:
-+ ==============================
-+ FQID: extracted OR byte may be shifted right 1-31 bits to effect parts of the FQID.
-+ if shifted less than 8 bits, or more than 24 bits a mask is set on the bits that
-+ are not overlapping FQID.
-+ ------------------------
-+ | FQID (24) |
-+ ------------------------
-+ --------
-+ | | extracted OR byte
-+ --------
-+
-+ Policer Profile: extracted OR byte may be shifted right 1-15 bits to effect parts of the
-+ PP id. Unless shifted exactly 8 bits to overlap the PP id, a mask is set on the bits that
-+ are not overlapping PP id.
-+
-+ --------
-+ | PP (8) |
-+ --------
-+ --------
-+ | | extracted OR byte
-+ --------
-+
-+ HW implementation
-+ =================
-+ FQID and PP construct a 32 bit word in the way describe below. Extracted byte is located
-+ as the highest byte of that word and may be rotated to effect any part os the FQID or
-+ the PP.
-+ ------------------------ --------
-+ | FQID (24) || PP (8) |
-+ ------------------------ --------
-+ --------
-+ | | extracted OR byte
-+ --------
-+
-+ ************************************************************************************/
-+
-+ if (p_ExtractOr->bitOffsetInFqid)
-+ {
-+ if (p_ExtractOr->bitOffsetInFqid > MAX_KG_SCH_FQID_BIT_OFFSET )
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("Illegal extraction (bitOffsetInFqid out of range)"));
-+ if (p_ExtractOr->bitOffsetInFqid<8)
-+ genTmp |= (uint32_t)((p_ExtractOr->bitOffsetInFqid+24) << KG_SCH_GEN_SIZE_SHIFT);
-+ else
-+ genTmp |= (uint32_t)((p_ExtractOr->bitOffsetInFqid-8) << KG_SCH_GEN_SIZE_SHIFT);
-+ p_ExtractOr->mask &= GetExtractedOrMask(p_ExtractOr->bitOffsetInFqid, TRUE);
-+ }
-+ else /* effect policer profile */
-+ {
-+ if (p_ExtractOr->bitOffsetInPlcrProfile > MAX_KG_SCH_PP_BIT_OFFSET )
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("Illegal extraction (bitOffsetInPlcrProfile out of range)"));
-+ p_Scheme->bitOffsetInPlcrProfile = p_ExtractOr->bitOffsetInPlcrProfile;
-+ genTmp |= (uint32_t)((p_ExtractOr->bitOffsetInPlcrProfile+16) << KG_SCH_GEN_SIZE_SHIFT);
-+ p_ExtractOr->mask &= GetExtractedOrMask(p_ExtractOr->bitOffsetInPlcrProfile, FALSE);
-+ }
-+
-+ genTmp |= (uint32_t)(p_ExtractOr->extractionOffset << KG_SCH_GEN_DEF_SHIFT);
-+ /* clear bits that need masking */
-+ genTmp &= ~KG_SCH_GEN_MASK ;
-+ /* set mask bits */
-+ genTmp |= (uint32_t)(p_ExtractOr->mask << KG_SCH_GEN_MASK_SHIFT);
-+ p_SchemeRegs->kgse_gec[currGenId++] = genTmp;
-+
-+ }
-+ /* clear all unused GEC registers */
-+ for ( i=currGenId ;i<FM_KG_NUM_OF_GENERIC_REGS ; i++)
-+ p_SchemeRegs->kgse_gec[i] = 0;
-+
-+ /* add base Qid for this scheme */
-+ /* add configuration for kgse_fqb */
-+ if (p_SchemeParams->baseFqid & ~0x00FFFFFF)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("baseFqid must be between 1 and 2^24-1"));
-+
-+ fqbTmp |= p_SchemeParams->baseFqid;
-+ p_SchemeRegs->kgse_fqb = fqbTmp;
-+
-+ p_Scheme->nextEngine = p_SchemeParams->nextEngine;
-+ p_Scheme->doneAction = p_SchemeParams->kgNextEngineParams.doneAction;
-+
-+ return E_OK;
-+}
-+
-+
-+/*****************************************************************************/
-+/* Inter-module API routines */
-+/*****************************************************************************/
-+
-+t_Error FmPcdKgBuildClsPlanGrp(t_Handle h_FmPcd, t_FmPcdKgInterModuleClsPlanGrpParams *p_Grp, t_FmPcdKgInterModuleClsPlanSet *p_ClsPlanSet)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+ t_FmPcdKgClsPlanGrp *p_ClsPlanGrp;
-+ t_FmPcdIpcKgClsPlanParams kgAlloc;
-+ t_Error err = E_OK;
-+ uint32_t oredVectors = 0;
-+ int i, j;
-+
-+ /* this routine is protected by the calling routine ! */
-+ if (p_Grp->numOfOptions >= FM_PCD_MAX_NUM_OF_OPTIONS(FM_PCD_MAX_NUM_OF_CLS_PLANS))
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE,("Too many classification plan basic options selected."));
-+
-+ /* find a new clsPlan group */
-+ for (i = 0; i < FM_MAX_NUM_OF_PORTS; i++)
-+ if (!p_FmPcd->p_FmPcdKg->clsPlanGrps[i].used)
-+ break;
-+ if (i == FM_MAX_NUM_OF_PORTS)
-+ RETURN_ERROR(MAJOR, E_FULL,("No classification plan groups available."));
-+
-+ p_FmPcd->p_FmPcdKg->clsPlanGrps[i].used = TRUE;
-+
-+ p_Grp->clsPlanGrpId = (uint8_t)i;
-+
-+ if (p_Grp->numOfOptions == 0)
-+ p_FmPcd->p_FmPcdKg->emptyClsPlanGrpId = (uint8_t)i;
-+
-+ p_ClsPlanGrp = &p_FmPcd->p_FmPcdKg->clsPlanGrps[i];
-+ p_ClsPlanGrp->netEnvId = p_Grp->netEnvId;
-+ p_ClsPlanGrp->owners = 0;
-+ FmPcdSetClsPlanGrpId(p_FmPcd, p_Grp->netEnvId, p_Grp->clsPlanGrpId);
-+ if (p_Grp->numOfOptions != 0)
-+ FmPcdIncNetEnvOwners(p_FmPcd, p_Grp->netEnvId);
-+
-+ p_ClsPlanGrp->sizeOfGrp = (uint16_t)(1 << p_Grp->numOfOptions);
-+ /* a minimal group of 8 is required */
-+ if (p_ClsPlanGrp->sizeOfGrp < CLS_PLAN_NUM_PER_GRP)
-+ p_ClsPlanGrp->sizeOfGrp = CLS_PLAN_NUM_PER_GRP;
-+ if (p_FmPcd->guestId == NCSW_MASTER_ID)
-+ {
-+ err = KgAllocClsPlanEntries(h_FmPcd, p_ClsPlanGrp->sizeOfGrp, p_FmPcd->guestId, &p_ClsPlanGrp->baseEntry);
-+
-+ if (err)
-+ RETURN_ERROR(MINOR, E_INVALID_STATE, NO_MSG);
-+ }
-+ else
-+ {
-+ t_FmPcdIpcMsg msg;
-+ uint32_t replyLength;
-+ t_FmPcdIpcReply reply;
-+
-+ /* in GUEST_PARTITION, we use the IPC, to also set a private driver group if required */
-+ memset(&reply, 0, sizeof(reply));
-+ memset(&msg, 0, sizeof(msg));
-+ memset(&kgAlloc, 0, sizeof(kgAlloc));
-+ kgAlloc.guestId = p_FmPcd->guestId;
-+ kgAlloc.numOfClsPlanEntries = p_ClsPlanGrp->sizeOfGrp;
-+ msg.msgId = FM_PCD_ALLOC_KG_CLSPLAN;
-+ memcpy(msg.msgBody, &kgAlloc, sizeof(kgAlloc));
-+ replyLength = (sizeof(uint32_t) + sizeof(p_ClsPlanGrp->baseEntry));
-+ if ((err = XX_IpcSendMessage(p_FmPcd->h_IpcSession,
-+ (uint8_t*)&msg,
-+ sizeof(msg.msgId) + sizeof(kgAlloc),
-+ (uint8_t*)&reply,
-+ &replyLength,
-+ NULL,
-+ NULL)) != E_OK)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+
-+ if (replyLength != (sizeof(uint32_t) + sizeof(p_ClsPlanGrp->baseEntry)))
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("IPC reply length mismatch"));
-+ if ((t_Error)reply.error != E_OK)
-+ RETURN_ERROR(MINOR, (t_Error)reply.error, NO_MSG);
-+
-+ p_ClsPlanGrp->baseEntry = *(uint8_t*)(reply.replyBody);
-+ }
-+
-+ /* build classification plan entries parameters */
-+ p_ClsPlanSet->baseEntry = p_ClsPlanGrp->baseEntry;
-+ p_ClsPlanSet->numOfClsPlanEntries = p_ClsPlanGrp->sizeOfGrp;
-+
-+ oredVectors = 0;
-+ for (i = 0; i<p_Grp->numOfOptions; i++)
-+ {
-+ oredVectors |= p_Grp->optVectors[i];
-+ /* save an array of used options - the indexes represent the power of 2 index */
-+ p_ClsPlanGrp->optArray[i] = p_Grp->options[i];
-+ }
-+ /* set the classification plan relevant entries so that all bits
-+ * relevant to the list of options is cleared
-+ */
-+ for (j = 0; j<p_ClsPlanGrp->sizeOfGrp; j++)
-+ p_ClsPlanSet->vectors[j] = ~oredVectors;
-+
-+ for (i = 0; i<p_Grp->numOfOptions; i++)
-+ {
-+ /* option i got the place 2^i in the clsPlan array. all entries that
-+ * have bit i set, should have the vector bit cleared. So each option
-+ * has one location that it is exclusive (1,2,4,8...) and represent the
-+ * presence of that option only, and other locations that represent a
-+ * combination of options.
-+ * e.g:
-+ * If ethernet-BC is option 1 it gets entry 2 in the table. Entry 2
-+ * now represents a frame with ethernet-BC header - so the bit
-+ * representing ethernet-BC should be set and all other option bits
-+ * should be cleared.
-+ * Entries 2,3,6,7,10... also have ethernet-BC and therefore have bit
-+ * vector[1] set, but they also have other bits set:
-+ * 3=1+2, options 0 and 1
-+ * 6=2+4, options 1 and 2
-+ * 7=1+2+4, options 0,1,and 2
-+ * 10=2+8, options 1 and 3
-+ * etc.
-+ * */
-+
-+ /* now for each option (i), we set their bits in all entries (j)
-+ * that contain bit 2^i.
-+ */
-+ for (j = 0; j<p_ClsPlanGrp->sizeOfGrp; j++)
-+ {
-+ if (j & (1<<i))
-+ p_ClsPlanSet->vectors[j] |= p_Grp->optVectors[i];
-+ }
-+ }
-+
-+ return E_OK;
-+}
-+
-+void FmPcdKgDestroyClsPlanGrp(t_Handle h_FmPcd, uint8_t grpId)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+ t_FmPcdIpcKgClsPlanParams kgAlloc;
-+ t_Error err;
-+ t_FmPcdIpcMsg msg;
-+ uint32_t replyLength;
-+ t_FmPcdIpcReply reply;
-+
-+ /* check that no port is bound to this clsPlan */
-+ if (p_FmPcd->p_FmPcdKg->clsPlanGrps[grpId].owners)
-+ {
-+ REPORT_ERROR(MINOR, E_INVALID_STATE, ("Trying to delete a clsPlan grp that has ports bound to"));
-+ return;
-+ }
-+
-+ FmPcdSetClsPlanGrpId(p_FmPcd, p_FmPcd->p_FmPcdKg->clsPlanGrps[grpId].netEnvId, ILLEGAL_CLS_PLAN);
-+
-+ if (grpId == p_FmPcd->p_FmPcdKg->emptyClsPlanGrpId)
-+ p_FmPcd->p_FmPcdKg->emptyClsPlanGrpId = ILLEGAL_CLS_PLAN;
-+ else
-+ FmPcdDecNetEnvOwners(p_FmPcd, p_FmPcd->p_FmPcdKg->clsPlanGrps[grpId].netEnvId);
-+
-+ /* free blocks */
-+ if (p_FmPcd->guestId == NCSW_MASTER_ID)
-+ KgFreeClsPlanEntries(h_FmPcd,
-+ p_FmPcd->p_FmPcdKg->clsPlanGrps[grpId].sizeOfGrp,
-+ p_FmPcd->guestId,
-+ p_FmPcd->p_FmPcdKg->clsPlanGrps[grpId].baseEntry);
-+ else /* in GUEST_PARTITION, we use the IPC, to also set a private driver group if required */
-+ {
-+ memset(&reply, 0, sizeof(reply));
-+ memset(&msg, 0, sizeof(msg));
-+ kgAlloc.guestId = p_FmPcd->guestId;
-+ kgAlloc.numOfClsPlanEntries = p_FmPcd->p_FmPcdKg->clsPlanGrps[grpId].sizeOfGrp;
-+ kgAlloc.clsPlanBase = p_FmPcd->p_FmPcdKg->clsPlanGrps[grpId].baseEntry;
-+ msg.msgId = FM_PCD_FREE_KG_CLSPLAN;
-+ memcpy(msg.msgBody, &kgAlloc, sizeof(kgAlloc));
-+ replyLength = sizeof(uint32_t);
-+ err = XX_IpcSendMessage(p_FmPcd->h_IpcSession,
-+ (uint8_t*)&msg,
-+ sizeof(msg.msgId) + sizeof(kgAlloc),
-+ (uint8_t*)&reply,
-+ &replyLength,
-+ NULL,
-+ NULL);
-+ if (err != E_OK)
-+ {
-+ REPORT_ERROR(MINOR, err, NO_MSG);
-+ return;
-+ }
-+ if (replyLength != sizeof(uint32_t))
-+ {
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE, ("IPC reply length mismatch"));
-+ return;
-+ }
-+ if ((t_Error)reply.error != E_OK)
-+ {
-+ REPORT_ERROR(MAJOR, E_INVALID_STATE, ("Free KG clsPlan failed"));
-+ return;
-+ }
-+ }
-+
-+ /* clear clsPlan driver structure */
-+ memset(&p_FmPcd->p_FmPcdKg->clsPlanGrps[grpId], 0, sizeof(t_FmPcdKgClsPlanGrp));
-+}
-+
-+t_Error FmPcdKgBuildBindPortToSchemes(t_Handle h_FmPcd, t_FmPcdKgInterModuleBindPortToSchemes *p_BindPort, uint32_t *p_SpReg, bool add)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+ uint32_t j, schemesPerPortVector = 0;
-+ t_FmPcdKgScheme *p_Scheme;
-+ uint8_t i, relativeSchemeId;
-+ uint32_t tmp, walking1Mask;
-+ uint8_t swPortIndex = 0;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd->p_FmPcdKg, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmPcd->p_FmPcdDriverParam, E_INVALID_STATE);
-+
-+ /* for each scheme */
-+ for (i = 0; i<p_BindPort->numOfSchemes; i++)
-+ {
-+ relativeSchemeId = FmPcdKgGetRelativeSchemeId(p_FmPcd, p_BindPort->schemesIds[i]);
-+ if (relativeSchemeId >= FM_PCD_KG_NUM_OF_SCHEMES)
-+ RETURN_ERROR(MAJOR, E_NOT_IN_RANGE, NO_MSG);
-+
-+ if (add)
-+ {
-+ p_Scheme = &p_FmPcd->p_FmPcdKg->schemes[relativeSchemeId];
-+ if (!FmPcdKgIsSchemeValidSw(p_Scheme))
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("Requested scheme is invalid."));
-+ /* check netEnvId of the port against the scheme netEnvId */
-+ if ((p_Scheme->netEnvId != p_BindPort->netEnvId) && (p_Scheme->netEnvId != ILLEGAL_NETENV))
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("Port may not be bound to requested scheme - differ in netEnvId"));
-+
-+ /* if next engine is private port policer profile, we need to check that it is valid */
-+ HW_PORT_ID_TO_SW_PORT_INDX(swPortIndex, p_BindPort->hardwarePortId);
-+ if (p_Scheme->nextRelativePlcrProfile)
-+ {
-+ for (j = 0;j<p_Scheme->numOfProfiles;j++)
-+ {
-+ ASSERT_COND(p_FmPcd->p_FmPcdPlcr->portsMapping[swPortIndex].h_FmPort);
-+ if (p_Scheme->relativeProfileId+j >= p_FmPcd->p_FmPcdPlcr->portsMapping[swPortIndex].numOfProfiles)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("Relative profile not in range"));
-+ if (!FmPcdPlcrIsProfileValid(p_FmPcd, (uint16_t)(p_FmPcd->p_FmPcdPlcr->portsMapping[swPortIndex].profilesBase + p_Scheme->relativeProfileId + j)))
-+ RETURN_ERROR(MINOR, E_INVALID_STATE, ("Relative profile not valid."));
-+ }
-+ }
-+ if (!p_BindPort->useClsPlan)
-+ {
-+ /* This check may be redundant as port is a assigned to the whole NetEnv */
-+
-+ /* if this port does not use clsPlan, it may not be bound to schemes with units that contain
-+ cls plan options. Schemes that are used only directly, should not be checked.
-+ it also may not be bound to schemes that go to CC with units that are options - so we OR
-+ the match vector and the grpBits (= ccUnits) */
-+ if ((p_Scheme->matchVector != SCHEME_ALWAYS_DIRECT) || p_Scheme->ccUnits)
-+ {
-+ uint8_t netEnvId;
-+ walking1Mask = 0x80000000;
-+ netEnvId = (p_Scheme->netEnvId == ILLEGAL_NETENV)? p_BindPort->netEnvId:p_Scheme->netEnvId;
-+ tmp = (p_Scheme->matchVector == SCHEME_ALWAYS_DIRECT)? 0:p_Scheme->matchVector;
-+ tmp |= p_Scheme->ccUnits;
-+ while (tmp)
-+ {
-+ if (tmp & walking1Mask)
-+ {
-+ tmp &= ~walking1Mask;
-+ if (!PcdNetEnvIsUnitWithoutOpts(p_FmPcd, netEnvId, walking1Mask))
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("Port (without clsPlan) may not be bound to requested scheme - uses clsPlan options"));
-+ }
-+ walking1Mask >>= 1;
-+ }
-+ }
-+ }
-+ }
-+ /* build vector */
-+ schemesPerPortVector |= 1 << (31 - p_BindPort->schemesIds[i]);
-+ }
-+
-+ *p_SpReg = schemesPerPortVector;
-+
-+ return E_OK;
-+}
-+
-+t_Error FmPcdKgBindPortToSchemes(t_Handle h_FmPcd , t_FmPcdKgInterModuleBindPortToSchemes *p_SchemeBind)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+ uint32_t spReg;
-+ t_Error err = E_OK;
-+
-+ err = FmPcdKgBuildBindPortToSchemes(h_FmPcd, p_SchemeBind, &spReg, TRUE);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+
-+ err = KgWriteSp(p_FmPcd, p_SchemeBind->hardwarePortId, spReg, TRUE);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+
-+ IncSchemeOwners(p_FmPcd, p_SchemeBind);
-+
-+ return E_OK;
-+}
-+
-+t_Error FmPcdKgUnbindPortToSchemes(t_Handle h_FmPcd, t_FmPcdKgInterModuleBindPortToSchemes *p_SchemeBind)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+ uint32_t spReg;
-+ t_Error err = E_OK;
-+
-+ err = FmPcdKgBuildBindPortToSchemes(p_FmPcd, p_SchemeBind, &spReg, FALSE);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+
-+ err = KgWriteSp(p_FmPcd, p_SchemeBind->hardwarePortId, spReg, FALSE);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+
-+ DecSchemeOwners(p_FmPcd, p_SchemeBind);
-+
-+ return E_OK;
-+}
-+
-+bool FmPcdKgIsSchemeValidSw(t_Handle h_Scheme)
-+{
-+ t_FmPcdKgScheme *p_Scheme = (t_FmPcdKgScheme*)h_Scheme;
-+
-+ return p_Scheme->valid;
-+}
-+
-+bool KgIsSchemeAlwaysDirect(t_Handle h_FmPcd, uint8_t schemeId)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+
-+ if (p_FmPcd->p_FmPcdKg->schemes[schemeId].matchVector == SCHEME_ALWAYS_DIRECT)
-+ return TRUE;
-+ else
-+ return FALSE;
-+}
-+
-+t_Error FmPcdKgAllocSchemes(t_Handle h_FmPcd, uint8_t numOfSchemes, uint8_t guestId, uint8_t *p_SchemesIds)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd *)h_FmPcd;
-+ uint8_t i, j;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd->p_FmPcdKg, E_INVALID_HANDLE);
-+
-+ /* This routine is issued only on master core of master partition -
-+ either directly or through IPC, so no need for lock */
-+
-+ for (j = 0, i = 0; i < FM_PCD_KG_NUM_OF_SCHEMES && j < numOfSchemes; i++)
-+ {
-+ if (!p_FmPcd->p_FmPcdKg->schemesMng[i].allocated)
-+ {
-+ p_FmPcd->p_FmPcdKg->schemesMng[i].allocated = TRUE;
-+ p_FmPcd->p_FmPcdKg->schemesMng[i].ownerId = guestId;
-+ p_SchemesIds[j] = i;
-+ j++;
-+ }
-+ }
-+
-+ if (j != numOfSchemes)
-+ {
-+ /* roll back */
-+ for (j--; j; j--)
-+ {
-+ p_FmPcd->p_FmPcdKg->schemesMng[p_SchemesIds[j]].allocated = FALSE;
-+ p_FmPcd->p_FmPcdKg->schemesMng[p_SchemesIds[j]].ownerId = 0;
-+ p_SchemesIds[j] = 0;
-+ }
-+
-+ RETURN_ERROR(MAJOR, E_NOT_AVAILABLE, ("No schemes found"));
-+ }
-+
-+ return E_OK;
-+}
-+
-+t_Error FmPcdKgFreeSchemes(t_Handle h_FmPcd, uint8_t numOfSchemes, uint8_t guestId, uint8_t *p_SchemesIds)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd *)h_FmPcd;
-+ uint8_t i;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd->p_FmPcdKg, E_INVALID_HANDLE);
-+
-+ /* This routine is issued only on master core of master partition -
-+ either directly or through IPC */
-+
-+ for (i = 0; i < numOfSchemes; i++)
-+ {
-+ if (!p_FmPcd->p_FmPcdKg->schemesMng[p_SchemesIds[i]].allocated)
-+ {
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("Scheme was not previously allocated"));
-+ }
-+ if (p_FmPcd->p_FmPcdKg->schemesMng[p_SchemesIds[i]].ownerId != guestId)
-+ {
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("Scheme is not owned by caller. "));
-+ }
-+ p_FmPcd->p_FmPcdKg->schemesMng[p_SchemesIds[i]].allocated = FALSE;
-+ p_FmPcd->p_FmPcdKg->schemesMng[p_SchemesIds[i]].ownerId = 0;
-+ }
-+
-+ return E_OK;
-+}
-+
-+t_Error KgAllocClsPlanEntries(t_Handle h_FmPcd, uint16_t numOfClsPlanEntries, uint8_t guestId, uint8_t *p_First)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd *)h_FmPcd;
-+ uint8_t numOfBlocks, blocksFound=0, first=0;
-+ uint8_t i, j;
-+
-+ /* This routine is issued only on master core of master partition -
-+ either directly or through IPC, so no need for lock */
-+
-+ if (!numOfClsPlanEntries)
-+ return E_OK;
-+
-+ if ((numOfClsPlanEntries % CLS_PLAN_NUM_PER_GRP) || (!POWER_OF_2(numOfClsPlanEntries)))
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("numOfClsPlanEntries must be a power of 2 and divisible by 8"));
-+
-+ numOfBlocks = (uint8_t)(numOfClsPlanEntries/CLS_PLAN_NUM_PER_GRP);
-+
-+ /* try to find consequent blocks */
-+ first = 0;
-+ for (i = 0; i < FM_PCD_MAX_NUM_OF_CLS_PLANS/CLS_PLAN_NUM_PER_GRP;)
-+ {
-+ if (!p_FmPcd->p_FmPcdKg->clsPlanBlocksMng[i].allocated)
-+ {
-+ blocksFound++;
-+ i++;
-+ if (blocksFound == numOfBlocks)
-+ break;
-+ }
-+ else
-+ {
-+ blocksFound = 0;
-+ /* advance i to the next aligned address */
-+ first = i = (uint8_t)(first + numOfBlocks);
-+ }
-+ }
-+
-+ if (blocksFound == numOfBlocks)
-+ {
-+ *p_First = (uint8_t)(first * CLS_PLAN_NUM_PER_GRP);
-+ for (j = first; j < (first + numOfBlocks); j++)
-+ {
-+ p_FmPcd->p_FmPcdKg->clsPlanBlocksMng[j].allocated = TRUE;
-+ p_FmPcd->p_FmPcdKg->clsPlanBlocksMng[j].ownerId = guestId;
-+ }
-+ return E_OK;
-+ }
-+ else
-+ RETURN_ERROR(MINOR, E_FULL, ("No resources for clsPlan"));
-+}
-+
-+void KgFreeClsPlanEntries(t_Handle h_FmPcd, uint16_t numOfClsPlanEntries, uint8_t guestId, uint8_t base)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+ uint8_t numOfBlocks;
-+ uint8_t i, baseBlock;
-+
-+#ifdef DISABLE_ASSERTIONS
-+UNUSED(guestId);
-+#endif /* DISABLE_ASSERTIONS */
-+
-+ /* This routine is issued only on master core of master partition -
-+ either directly or through IPC, so no need for lock */
-+
-+ numOfBlocks = (uint8_t)(numOfClsPlanEntries/CLS_PLAN_NUM_PER_GRP);
-+ ASSERT_COND(!(base%CLS_PLAN_NUM_PER_GRP));
-+
-+ baseBlock = (uint8_t)(base/CLS_PLAN_NUM_PER_GRP);
-+ for (i=baseBlock;i<baseBlock+numOfBlocks;i++)
-+ {
-+ ASSERT_COND(p_FmPcd->p_FmPcdKg->clsPlanBlocksMng[i].allocated);
-+ ASSERT_COND(guestId == p_FmPcd->p_FmPcdKg->clsPlanBlocksMng[i].ownerId);
-+ p_FmPcd->p_FmPcdKg->clsPlanBlocksMng[i].allocated = FALSE;
-+ p_FmPcd->p_FmPcdKg->clsPlanBlocksMng[i].ownerId = 0;
-+ }
-+}
-+
-+void KgEnable(t_FmPcd *p_FmPcd)
-+{
-+ struct fman_kg_regs *p_Regs = p_FmPcd->p_FmPcdKg->p_FmPcdKgRegs;
-+
-+ ASSERT_COND(FmIsMaster(p_FmPcd->h_Fm));
-+ fman_kg_enable(p_Regs);
-+}
-+
-+void KgDisable(t_FmPcd *p_FmPcd)
-+{
-+ struct fman_kg_regs *p_Regs = p_FmPcd->p_FmPcdKg->p_FmPcdKgRegs;
-+
-+ ASSERT_COND(FmIsMaster(p_FmPcd->h_Fm));
-+ fman_kg_disable(p_Regs);
-+}
-+
-+void KgSetClsPlan(t_Handle h_FmPcd, t_FmPcdKgInterModuleClsPlanSet *p_Set)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd *)h_FmPcd;
-+ struct fman_kg_cp_regs *p_FmPcdKgPortRegs;
-+ uint32_t tmpKgarReg = 0, intFlags;
-+ uint16_t i, j;
-+
-+ /* This routine is protected by the calling routine ! */
-+ ASSERT_COND(FmIsMaster(p_FmPcd->h_Fm));
-+ p_FmPcdKgPortRegs = &p_FmPcd->p_FmPcdKg->p_IndirectAccessRegs->clsPlanRegs;
-+
-+ intFlags = KgHwLock(p_FmPcd->p_FmPcdKg);
-+ for (i=p_Set->baseEntry;i<p_Set->baseEntry+p_Set->numOfClsPlanEntries;i+=8)
-+ {
-+ tmpKgarReg = FmPcdKgBuildWriteClsPlanBlockActionReg((uint8_t)(i / CLS_PLAN_NUM_PER_GRP));
-+
-+ for (j = i; j < i+8; j++)
-+ {
-+ ASSERT_COND(IN_RANGE(0, (j - p_Set->baseEntry), FM_PCD_MAX_NUM_OF_CLS_PLANS-1));
-+ WRITE_UINT32(p_FmPcdKgPortRegs->kgcpe[j % CLS_PLAN_NUM_PER_GRP],p_Set->vectors[j - p_Set->baseEntry]);
-+ }
-+
-+ if (WriteKgarWait(p_FmPcd, tmpKgarReg) != E_OK)
-+ {
-+ REPORT_ERROR(MAJOR, E_INVALID_STATE, ("WriteKgarWait FAILED"));
-+ KgHwUnlock(p_FmPcd->p_FmPcdKg, intFlags);
-+ return;
-+ }
-+ }
-+ KgHwUnlock(p_FmPcd->p_FmPcdKg, intFlags);
-+}
-+
-+t_Handle KgConfig( t_FmPcd *p_FmPcd, t_FmPcdParams *p_FmPcdParams)
-+{
-+ t_FmPcdKg *p_FmPcdKg;
-+
-+ UNUSED(p_FmPcd);
-+
-+ if (p_FmPcdParams->numOfSchemes > FM_PCD_KG_NUM_OF_SCHEMES)
-+ {
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE,
-+ ("numOfSchemes should not exceed %d", FM_PCD_KG_NUM_OF_SCHEMES));
-+ return NULL;
-+ }
-+
-+ p_FmPcdKg = (t_FmPcdKg *)XX_Malloc(sizeof(t_FmPcdKg));
-+ if (!p_FmPcdKg)
-+ {
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("FM Keygen allocation FAILED"));
-+ return NULL;
-+ }
-+ memset(p_FmPcdKg, 0, sizeof(t_FmPcdKg));
-+
-+
-+ if (FmIsMaster(p_FmPcd->h_Fm))
-+ {
-+ p_FmPcdKg->p_FmPcdKgRegs = (struct fman_kg_regs *)UINT_TO_PTR(FmGetPcdKgBaseAddr(p_FmPcdParams->h_Fm));
-+ p_FmPcd->exceptions |= DEFAULT_fmPcdKgErrorExceptions;
-+ p_FmPcdKg->p_IndirectAccessRegs = (u_FmPcdKgIndirectAccessRegs *)&p_FmPcdKg->p_FmPcdKgRegs->fmkg_indirect[0];
-+ }
-+
-+ p_FmPcdKg->numOfSchemes = p_FmPcdParams->numOfSchemes;
-+ if ((p_FmPcd->guestId == NCSW_MASTER_ID) && !p_FmPcdKg->numOfSchemes)
-+ {
-+ p_FmPcdKg->numOfSchemes = FM_PCD_KG_NUM_OF_SCHEMES;
-+ DBG(WARNING, ("numOfSchemes was defined 0 by user, re-defined by driver to FM_PCD_KG_NUM_OF_SCHEMES"));
-+ }
-+
-+ p_FmPcdKg->emptyClsPlanGrpId = ILLEGAL_CLS_PLAN;
-+
-+ return p_FmPcdKg;
-+}
-+
-+t_Error KgInit(t_FmPcd *p_FmPcd)
-+{
-+ t_Error err = E_OK;
-+
-+ p_FmPcd->p_FmPcdKg->h_HwSpinlock = XX_InitSpinlock();
-+ if (!p_FmPcd->p_FmPcdKg->h_HwSpinlock)
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, ("FM KG HW spinlock"));
-+
-+ if (p_FmPcd->guestId == NCSW_MASTER_ID)
-+ err = KgInitMaster(p_FmPcd);
-+ else
-+ err = KgInitGuest(p_FmPcd);
-+
-+ if (err != E_OK)
-+ {
-+ if (p_FmPcd->p_FmPcdKg->h_HwSpinlock)
-+ XX_FreeSpinlock(p_FmPcd->p_FmPcdKg->h_HwSpinlock);
-+ }
-+
-+ return err;
-+}
-+
-+t_Error KgFree(t_FmPcd *p_FmPcd)
-+{
-+ t_FmPcdIpcKgSchemesParams kgAlloc;
-+ t_Error err = E_OK;
-+ t_FmPcdIpcMsg msg;
-+ uint32_t replyLength;
-+ t_FmPcdIpcReply reply;
-+
-+ FmUnregisterIntr(p_FmPcd->h_Fm, e_FM_MOD_KG, 0, e_FM_INTR_TYPE_ERR);
-+
-+ if (p_FmPcd->guestId == NCSW_MASTER_ID)
-+ {
-+ err = FmPcdKgFreeSchemes(p_FmPcd,
-+ p_FmPcd->p_FmPcdKg->numOfSchemes,
-+ p_FmPcd->guestId,
-+ p_FmPcd->p_FmPcdKg->schemesIds);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+
-+ if (p_FmPcd->p_FmPcdKg->h_HwSpinlock)
-+ XX_FreeSpinlock(p_FmPcd->p_FmPcdKg->h_HwSpinlock);
-+
-+ return E_OK;
-+ }
-+
-+ /* guest */
-+ memset(&reply, 0, sizeof(reply));
-+ memset(&msg, 0, sizeof(msg));
-+ kgAlloc.numOfSchemes = p_FmPcd->p_FmPcdKg->numOfSchemes;
-+ kgAlloc.guestId = p_FmPcd->guestId;
-+ ASSERT_COND(kgAlloc.numOfSchemes < FM_PCD_KG_NUM_OF_SCHEMES);
-+ memcpy(kgAlloc.schemesIds, p_FmPcd->p_FmPcdKg->schemesIds, (sizeof(uint8_t))*kgAlloc.numOfSchemes);
-+ msg.msgId = FM_PCD_FREE_KG_SCHEMES;
-+ memcpy(msg.msgBody, &kgAlloc, sizeof(kgAlloc));
-+ replyLength = sizeof(uint32_t);
-+ if ((err = XX_IpcSendMessage(p_FmPcd->h_IpcSession,
-+ (uint8_t*)&msg,
-+ sizeof(msg.msgId) + sizeof(kgAlloc),
-+ (uint8_t*)&reply,
-+ &replyLength,
-+ NULL,
-+ NULL)) != E_OK)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ if (replyLength != sizeof(uint32_t))
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("IPC reply length mismatch"));
-+
-+ if (p_FmPcd->p_FmPcdKg->h_HwSpinlock)
-+ XX_FreeSpinlock(p_FmPcd->p_FmPcdKg->h_HwSpinlock);
-+
-+ return (t_Error)reply.error;
-+}
-+
-+t_Error FmPcdKgSetOrBindToClsPlanGrp(t_Handle h_FmPcd, uint8_t hardwarePortId, uint8_t netEnvId, protocolOpt_t *p_OptArray, uint8_t *p_ClsPlanGrpId, bool *p_IsEmptyClsPlanGrp)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd *)h_FmPcd;
-+ t_FmPcdKgInterModuleClsPlanGrpParams grpParams, *p_GrpParams;
-+ t_FmPcdKgClsPlanGrp *p_ClsPlanGrp;
-+ t_FmPcdKgInterModuleClsPlanSet *p_ClsPlanSet;
-+ t_Error err;
-+
-+ /* This function is issued only from FM_PORT_SetPcd which locked all PCD modules,
-+ so no need for lock here */
-+
-+ memset(&grpParams, 0, sizeof(grpParams));
-+ grpParams.clsPlanGrpId = ILLEGAL_CLS_PLAN;
-+ p_GrpParams = &grpParams;
-+
-+ p_GrpParams->netEnvId = netEnvId;
-+
-+ /* Get from the NetEnv the information of the clsPlan (can be already created,
-+ * or needs to build) */
-+ err = PcdGetClsPlanGrpParams(h_FmPcd, p_GrpParams);
-+ if (err)
-+ RETURN_ERROR(MINOR,err,NO_MSG);
-+
-+ if (p_GrpParams->grpExists)
-+ {
-+ /* this group was already updated (at least) in SW */
-+ *p_ClsPlanGrpId = p_GrpParams->clsPlanGrpId;
-+ }
-+ else
-+ {
-+ p_ClsPlanSet = (t_FmPcdKgInterModuleClsPlanSet *)XX_Malloc(sizeof(t_FmPcdKgInterModuleClsPlanSet));
-+ if (!p_ClsPlanSet)
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, ("Classification plan set"));
-+ memset(p_ClsPlanSet, 0, sizeof(t_FmPcdKgInterModuleClsPlanSet));
-+ /* Build (in SW) the clsPlan parameters, including the vectors to be written to HW */
-+ err = FmPcdKgBuildClsPlanGrp(h_FmPcd, p_GrpParams, p_ClsPlanSet);
-+ if (err)
-+ {
-+ XX_Free(p_ClsPlanSet);
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+ }
-+ *p_ClsPlanGrpId = p_GrpParams->clsPlanGrpId;
-+
-+ if (p_FmPcd->h_Hc)
-+ {
-+ /* write clsPlan entries to memory */
-+ err = FmHcPcdKgSetClsPlan(p_FmPcd->h_Hc, p_ClsPlanSet);
-+ if (err)
-+ {
-+ XX_Free(p_ClsPlanSet);
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+ }
-+ else
-+ /* write clsPlan entries to memory */
-+ KgSetClsPlan(p_FmPcd, p_ClsPlanSet);
-+
-+ XX_Free(p_ClsPlanSet);
-+ }
-+
-+ /* Set caller parameters */
-+
-+ /* mark if this is an empty classification group */
-+ if (*p_ClsPlanGrpId == p_FmPcd->p_FmPcdKg->emptyClsPlanGrpId)
-+ *p_IsEmptyClsPlanGrp = TRUE;
-+ else
-+ *p_IsEmptyClsPlanGrp = FALSE;
-+
-+ p_ClsPlanGrp = &p_FmPcd->p_FmPcdKg->clsPlanGrps[*p_ClsPlanGrpId];
-+
-+ /* increment owners number */
-+ p_ClsPlanGrp->owners++;
-+
-+ /* copy options array for port */
-+ memcpy(p_OptArray, &p_FmPcd->p_FmPcdKg->clsPlanGrps[*p_ClsPlanGrpId].optArray, FM_PCD_MAX_NUM_OF_OPTIONS(FM_PCD_MAX_NUM_OF_CLS_PLANS)*sizeof(protocolOpt_t));
-+
-+ /* bind port to the new or existing group */
-+ err = BindPortToClsPlanGrp(p_FmPcd, hardwarePortId, p_GrpParams->clsPlanGrpId);
-+ if (err)
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+
-+ return E_OK;
-+}
-+
-+t_Error FmPcdKgDeleteOrUnbindPortToClsPlanGrp(t_Handle h_FmPcd, uint8_t hardwarePortId, uint8_t clsPlanGrpId)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd *)h_FmPcd;
-+ t_FmPcdKgClsPlanGrp *p_ClsPlanGrp = &p_FmPcd->p_FmPcdKg->clsPlanGrps[clsPlanGrpId];
-+ t_FmPcdKgInterModuleClsPlanSet *p_ClsPlanSet;
-+ t_Error err;
-+
-+ /* This function is issued only from FM_PORT_DeletePcd which locked all PCD modules,
-+ so no need for lock here */
-+
-+ UnbindPortToClsPlanGrp(p_FmPcd, hardwarePortId);
-+
-+ /* decrement owners number */
-+ ASSERT_COND(p_ClsPlanGrp->owners);
-+ p_ClsPlanGrp->owners--;
-+
-+ if (!p_ClsPlanGrp->owners)
-+ {
-+ if (p_FmPcd->h_Hc)
-+ {
-+ err = FmHcPcdKgDeleteClsPlan(p_FmPcd->h_Hc, clsPlanGrpId);
-+ return err;
-+ }
-+ else
-+ {
-+ /* clear clsPlan entries in memory */
-+ p_ClsPlanSet = (t_FmPcdKgInterModuleClsPlanSet *)XX_Malloc(sizeof(t_FmPcdKgInterModuleClsPlanSet));
-+ if (!p_ClsPlanSet)
-+ {
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, ("Classification plan set"));
-+ }
-+ memset(p_ClsPlanSet, 0, sizeof(t_FmPcdKgInterModuleClsPlanSet));
-+
-+ p_ClsPlanSet->baseEntry = p_FmPcd->p_FmPcdKg->clsPlanGrps[clsPlanGrpId].baseEntry;
-+ p_ClsPlanSet->numOfClsPlanEntries = p_FmPcd->p_FmPcdKg->clsPlanGrps[clsPlanGrpId].sizeOfGrp;
-+ KgSetClsPlan(p_FmPcd, p_ClsPlanSet);
-+ XX_Free(p_ClsPlanSet);
-+
-+ FmPcdKgDestroyClsPlanGrp(h_FmPcd, clsPlanGrpId);
-+ }
-+ }
-+ return E_OK;
-+}
-+
-+uint32_t FmPcdKgGetRequiredAction(t_Handle h_FmPcd, uint8_t schemeId)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+ ASSERT_COND(p_FmPcd->p_FmPcdKg->schemes[schemeId].valid);
-+
-+ return p_FmPcd->p_FmPcdKg->schemes[schemeId].requiredAction;
-+}
-+
-+uint32_t FmPcdKgGetRequiredActionFlag(t_Handle h_FmPcd, uint8_t schemeId)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+
-+ ASSERT_COND(p_FmPcd->p_FmPcdKg->schemes[schemeId].valid);
-+
-+ return p_FmPcd->p_FmPcdKg->schemes[schemeId].requiredActionFlag;
-+}
-+
-+bool FmPcdKgIsDirectPlcr(t_Handle h_FmPcd, uint8_t schemeId)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+
-+ ASSERT_COND(p_FmPcd->p_FmPcdKg->schemes[schemeId].valid);
-+
-+ return p_FmPcd->p_FmPcdKg->schemes[schemeId].directPlcr;
-+}
-+
-+
-+uint16_t FmPcdKgGetRelativeProfileId(t_Handle h_FmPcd, uint8_t schemeId)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+
-+ ASSERT_COND(p_FmPcd->p_FmPcdKg->schemes[schemeId].valid);
-+
-+ return p_FmPcd->p_FmPcdKg->schemes[schemeId].relativeProfileId;
-+}
-+
-+bool FmPcdKgIsDistrOnPlcrProfile(t_Handle h_FmPcd, uint8_t schemeId)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+
-+ ASSERT_COND(p_FmPcd->p_FmPcdKg->schemes[schemeId].valid);
-+
-+ if ((p_FmPcd->p_FmPcdKg->schemes[schemeId].extractedOrs &&
-+ p_FmPcd->p_FmPcdKg->schemes[schemeId].bitOffsetInPlcrProfile) ||
-+ p_FmPcd->p_FmPcdKg->schemes[schemeId].nextRelativePlcrProfile)
-+ return TRUE;
-+ else
-+ return FALSE;
-+
-+}
-+
-+e_FmPcdEngine FmPcdKgGetNextEngine(t_Handle h_FmPcd, uint8_t relativeSchemeId)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+
-+ ASSERT_COND(p_FmPcd->p_FmPcdKg->schemes[relativeSchemeId].valid);
-+
-+ return p_FmPcd->p_FmPcdKg->schemes[relativeSchemeId].nextEngine;
-+}
-+
-+e_FmPcdDoneAction FmPcdKgGetDoneAction(t_Handle h_FmPcd, uint8_t schemeId)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+
-+ ASSERT_COND(p_FmPcd->p_FmPcdKg->schemes[schemeId].valid);
-+
-+ return p_FmPcd->p_FmPcdKg->schemes[schemeId].doneAction;
-+}
-+
-+void FmPcdKgUpdateRequiredAction(t_Handle h_Scheme, uint32_t requiredAction)
-+{
-+ t_FmPcdKgScheme *p_Scheme = (t_FmPcdKgScheme *)h_Scheme;
-+
-+ /* this routine is protected by calling routine */
-+
-+ ASSERT_COND(p_Scheme->valid);
-+
-+ p_Scheme->requiredAction |= requiredAction;
-+}
-+
-+bool FmPcdKgHwSchemeIsValid(uint32_t schemeModeReg)
-+{
-+ return (bool)!!(schemeModeReg & KG_SCH_MODE_EN);
-+}
-+
-+uint32_t FmPcdKgBuildWriteSchemeActionReg(uint8_t schemeId, bool updateCounter)
-+{
-+ return (uint32_t)(((uint32_t)schemeId << FM_PCD_KG_KGAR_NUM_SHIFT) |
-+ FM_KG_KGAR_GO |
-+ FM_KG_KGAR_WRITE |
-+ FM_KG_KGAR_SEL_SCHEME_ENTRY |
-+ DUMMY_PORT_ID |
-+ (updateCounter ? FM_KG_KGAR_SCM_WSEL_UPDATE_CNT:0));
-+}
-+
-+uint32_t FmPcdKgBuildReadSchemeActionReg(uint8_t schemeId)
-+{
-+ return (uint32_t)(((uint32_t)schemeId << FM_PCD_KG_KGAR_NUM_SHIFT) |
-+ FM_KG_KGAR_GO |
-+ FM_KG_KGAR_READ |
-+ FM_KG_KGAR_SEL_SCHEME_ENTRY |
-+ DUMMY_PORT_ID |
-+ FM_KG_KGAR_SCM_WSEL_UPDATE_CNT);
-+
-+}
-+
-+uint32_t FmPcdKgBuildWriteClsPlanBlockActionReg(uint8_t grpId)
-+{
-+ return (uint32_t)(FM_KG_KGAR_GO |
-+ FM_KG_KGAR_WRITE |
-+ FM_PCD_KG_KGAR_SEL_CLS_PLAN_ENTRY |
-+ DUMMY_PORT_ID |
-+ ((uint32_t)grpId << FM_PCD_KG_KGAR_NUM_SHIFT) |
-+ FM_PCD_KG_KGAR_WSEL_MASK);
-+
-+ /* if we ever want to write 1 by 1, use:
-+ sel = (uint8_t)(0x01 << (7- (entryId % CLS_PLAN_NUM_PER_GRP)));
-+ */
-+}
-+
-+uint32_t FmPcdKgBuildWritePortSchemeBindActionReg(uint8_t hardwarePortId)
-+{
-+
-+ return (uint32_t)(FM_KG_KGAR_GO |
-+ FM_KG_KGAR_WRITE |
-+ FM_PCD_KG_KGAR_SEL_PORT_ENTRY |
-+ hardwarePortId |
-+ FM_PCD_KG_KGAR_SEL_PORT_WSEL_SP);
-+}
-+
-+uint32_t FmPcdKgBuildReadPortSchemeBindActionReg(uint8_t hardwarePortId)
-+{
-+
-+ return (uint32_t)(FM_KG_KGAR_GO |
-+ FM_KG_KGAR_READ |
-+ FM_PCD_KG_KGAR_SEL_PORT_ENTRY |
-+ hardwarePortId |
-+ FM_PCD_KG_KGAR_SEL_PORT_WSEL_SP);
-+}
-+
-+uint32_t FmPcdKgBuildWritePortClsPlanBindActionReg(uint8_t hardwarePortId)
-+{
-+
-+ return (uint32_t)(FM_KG_KGAR_GO |
-+ FM_KG_KGAR_WRITE |
-+ FM_PCD_KG_KGAR_SEL_PORT_ENTRY |
-+ hardwarePortId |
-+ FM_PCD_KG_KGAR_SEL_PORT_WSEL_CPP);
-+}
-+
-+uint8_t FmPcdKgGetClsPlanGrpBase(t_Handle h_FmPcd, uint8_t clsPlanGrp)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+
-+ return p_FmPcd->p_FmPcdKg->clsPlanGrps[clsPlanGrp].baseEntry;
-+}
-+
-+uint16_t FmPcdKgGetClsPlanGrpSize(t_Handle h_FmPcd, uint8_t clsPlanGrp)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+
-+ return p_FmPcd->p_FmPcdKg->clsPlanGrps[clsPlanGrp].sizeOfGrp;
-+}
-+
-+
-+uint8_t FmPcdKgGetSchemeId(t_Handle h_Scheme)
-+{
-+ return ((t_FmPcdKgScheme*)h_Scheme)->schemeId;
-+
-+}
-+
-+#if (DPAA_VERSION >= 11)
-+bool FmPcdKgGetVspe(t_Handle h_Scheme)
-+{
-+ return ((t_FmPcdKgScheme*)h_Scheme)->vspe;
-+
-+}
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+uint8_t FmPcdKgGetRelativeSchemeId(t_Handle h_FmPcd, uint8_t schemeId)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+ uint8_t i;
-+
-+ for (i = 0;i<p_FmPcd->p_FmPcdKg->numOfSchemes;i++)
-+ if (p_FmPcd->p_FmPcdKg->schemesIds[i] == schemeId)
-+ return i;
-+
-+ if (i == p_FmPcd->p_FmPcdKg->numOfSchemes)
-+ REPORT_ERROR(MAJOR, E_NOT_IN_RANGE, ("Scheme is out of partition range"));
-+
-+ return FM_PCD_KG_NUM_OF_SCHEMES;
-+}
-+
-+t_Handle FmPcdKgGetSchemeHandle(t_Handle h_FmPcd, uint8_t relativeSchemeId)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+
-+ ASSERT_COND(p_FmPcd);
-+
-+ /* check that schemeId is in range */
-+ if (relativeSchemeId >= p_FmPcd->p_FmPcdKg->numOfSchemes)
-+ {
-+ REPORT_ERROR(MAJOR, E_NOT_IN_RANGE, ("relative-scheme-id %d!", relativeSchemeId));
-+ return NULL;
-+ }
-+
-+ if (!FmPcdKgIsSchemeValidSw(&p_FmPcd->p_FmPcdKg->schemes[relativeSchemeId]))
-+ return NULL;
-+
-+ return &p_FmPcd->p_FmPcdKg->schemes[relativeSchemeId];
-+}
-+
-+bool FmPcdKgIsSchemeHasOwners(t_Handle h_Scheme)
-+{
-+ return (((t_FmPcdKgScheme*)h_Scheme)->owners == 0)?FALSE:TRUE;
-+}
-+
-+t_Error FmPcdKgCcGetSetParams(t_Handle h_FmPcd, t_Handle h_Scheme, uint32_t requiredAction, uint32_t value)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+ uint8_t relativeSchemeId, physicalSchemeId;
-+ uint32_t tmpKgarReg, tmpReg32 = 0, intFlags;
-+ t_Error err;
-+ t_FmPcdKgScheme *p_Scheme = (t_FmPcdKgScheme*)h_Scheme;
-+
-+ SANITY_CHECK_RETURN_VALUE(h_FmPcd, E_INVALID_HANDLE, 0);
-+ SANITY_CHECK_RETURN_VALUE(p_FmPcd->p_FmPcdKg, E_INVALID_HANDLE, 0);
-+ SANITY_CHECK_RETURN_VALUE(!p_FmPcd->p_FmPcdDriverParam, E_INVALID_STATE, 0);
-+
-+ /* Calling function locked all PCD modules, so no need to lock here */
-+
-+ if (!FmPcdKgIsSchemeValidSw(h_Scheme))
-+ RETURN_ERROR(MAJOR, E_ALREADY_EXISTS, ("Scheme is Invalid"));
-+
-+ if (p_FmPcd->h_Hc)
-+ {
-+ err = FmHcPcdKgCcGetSetParams(p_FmPcd->h_Hc, h_Scheme, requiredAction, value);
-+
-+ UpdateRequiredActionFlag(h_Scheme,TRUE);
-+ FmPcdKgUpdateRequiredAction(h_Scheme,requiredAction);
-+ return err;
-+ }
-+
-+ physicalSchemeId = p_Scheme->schemeId;
-+
-+ relativeSchemeId = FmPcdKgGetRelativeSchemeId(p_FmPcd, physicalSchemeId);
-+ if (relativeSchemeId >= FM_PCD_KG_NUM_OF_SCHEMES)
-+ RETURN_ERROR(MAJOR, E_NOT_IN_RANGE, NO_MSG);
-+
-+ if (!p_FmPcd->p_FmPcdKg->schemes[relativeSchemeId].requiredActionFlag ||
-+ !(p_FmPcd->p_FmPcdKg->schemes[relativeSchemeId].requiredAction & requiredAction))
-+ {
-+ if (requiredAction & UPDATE_NIA_ENQ_WITHOUT_DMA)
-+ {
-+ switch (p_FmPcd->p_FmPcdKg->schemes[relativeSchemeId].nextEngine)
-+ {
-+ case (e_FM_PCD_DONE):
-+ if (p_FmPcd->p_FmPcdKg->schemes[relativeSchemeId].doneAction == e_FM_PCD_ENQ_FRAME)
-+ {
-+ tmpKgarReg = FmPcdKgBuildReadSchemeActionReg(physicalSchemeId);
-+ intFlags = KgHwLock(p_FmPcd->p_FmPcdKg);
-+ WriteKgarWait(p_FmPcd, tmpKgarReg);
-+ tmpReg32 = GET_UINT32(p_FmPcd->p_FmPcdKg->p_IndirectAccessRegs->schemeRegs.kgse_mode);
-+ ASSERT_COND(tmpReg32 & (NIA_ENG_BMI | NIA_BMI_AC_ENQ_FRAME));
-+ WRITE_UINT32(p_FmPcd->p_FmPcdKg->p_IndirectAccessRegs->schemeRegs.kgse_mode, tmpReg32 | NIA_BMI_AC_ENQ_FRAME_WITHOUT_DMA);
-+ /* call indirect command for scheme write */
-+ tmpKgarReg = FmPcdKgBuildWriteSchemeActionReg(physicalSchemeId, FALSE);
-+ WriteKgarWait(p_FmPcd, tmpKgarReg);
-+ KgHwUnlock(p_FmPcd->p_FmPcdKg, intFlags);
-+ }
-+ break;
-+ case (e_FM_PCD_PLCR):
-+ if (!p_FmPcd->p_FmPcdKg->schemes[relativeSchemeId].directPlcr ||
-+ (p_FmPcd->p_FmPcdKg->schemes[relativeSchemeId].extractedOrs &&
-+ p_FmPcd->p_FmPcdKg->schemes[relativeSchemeId].bitOffsetInPlcrProfile) ||
-+ p_FmPcd->p_FmPcdKg->schemes[relativeSchemeId].nextRelativePlcrProfile)
-+ {
-+ RETURN_ERROR(MAJOR, E_NOT_SUPPORTED, ("In this situation PP can not be with distribution and has to be shared"));
-+ }
-+ err = FmPcdPlcrCcGetSetParams(h_FmPcd, p_FmPcd->p_FmPcdKg->schemes[relativeSchemeId].relativeProfileId, requiredAction);
-+ if (err)
-+ {
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+ break;
-+ default:
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE,("in this situation the next engine after scheme can be or PLCR or ENQ_FRAME"));
-+ }
-+ }
-+ if (requiredAction & UPDATE_KG_NIA_CC_WA)
-+ {
-+ if (p_FmPcd->p_FmPcdKg->schemes[relativeSchemeId].nextEngine == e_FM_PCD_CC)
-+ {
-+ tmpKgarReg = FmPcdKgBuildReadSchemeActionReg(physicalSchemeId);
-+ intFlags = KgHwLock(p_FmPcd->p_FmPcdKg);
-+ WriteKgarWait(p_FmPcd, tmpKgarReg);
-+ tmpReg32 = GET_UINT32(p_FmPcd->p_FmPcdKg->p_IndirectAccessRegs->schemeRegs.kgse_mode);
-+ ASSERT_COND(tmpReg32 & (NIA_ENG_FM_CTL | NIA_FM_CTL_AC_CC));
-+ tmpReg32 &= ~NIA_FM_CTL_AC_CC;
-+ WRITE_UINT32(p_FmPcd->p_FmPcdKg->p_IndirectAccessRegs->schemeRegs.kgse_mode, tmpReg32 | NIA_FM_CTL_AC_PRE_CC);
-+ /* call indirect command for scheme write */
-+ tmpKgarReg = FmPcdKgBuildWriteSchemeActionReg(physicalSchemeId, FALSE);
-+ WriteKgarWait(p_FmPcd, tmpKgarReg);
-+ KgHwUnlock(p_FmPcd->p_FmPcdKg, intFlags);
-+ }
-+ }
-+ if (requiredAction & UPDATE_KG_OPT_MODE)
-+ {
-+ tmpKgarReg = FmPcdKgBuildReadSchemeActionReg(physicalSchemeId);
-+ intFlags = KgHwLock(p_FmPcd->p_FmPcdKg);
-+ WriteKgarWait(p_FmPcd, tmpKgarReg);
-+ WRITE_UINT32(p_FmPcd->p_FmPcdKg->p_IndirectAccessRegs->schemeRegs.kgse_om, value);
-+ /* call indirect command for scheme write */
-+ tmpKgarReg = FmPcdKgBuildWriteSchemeActionReg(physicalSchemeId, FALSE);
-+ WriteKgarWait(p_FmPcd, tmpKgarReg);
-+ KgHwUnlock(p_FmPcd->p_FmPcdKg, intFlags);
-+ }
-+ if (requiredAction & UPDATE_KG_NIA)
-+ {
-+ tmpKgarReg = FmPcdKgBuildReadSchemeActionReg(physicalSchemeId);
-+ intFlags = KgHwLock(p_FmPcd->p_FmPcdKg);
-+ WriteKgarWait(p_FmPcd, tmpKgarReg);
-+ tmpReg32 = GET_UINT32(p_FmPcd->p_FmPcdKg->p_IndirectAccessRegs->schemeRegs.kgse_mode);
-+ tmpReg32 &= ~(NIA_ENG_MASK | NIA_AC_MASK);
-+ tmpReg32 |= value;
-+ WRITE_UINT32(p_FmPcd->p_FmPcdKg->p_IndirectAccessRegs->schemeRegs.kgse_mode, tmpReg32);
-+ /* call indirect command for scheme write */
-+ tmpKgarReg = FmPcdKgBuildWriteSchemeActionReg(physicalSchemeId, FALSE);
-+ WriteKgarWait(p_FmPcd, tmpKgarReg);
-+ KgHwUnlock(p_FmPcd->p_FmPcdKg, intFlags);
-+ }
-+ }
-+
-+ UpdateRequiredActionFlag(h_Scheme, TRUE);
-+ FmPcdKgUpdateRequiredAction(h_Scheme, requiredAction);
-+
-+ return E_OK;
-+}
-+/*********************** End of inter-module routines ************************/
-+
-+
-+/****************************************/
-+/* API routines */
-+/****************************************/
-+
-+t_Handle FM_PCD_KgSchemeSet(t_Handle h_FmPcd, t_FmPcdKgSchemeParams *p_SchemeParams)
-+{
-+ t_FmPcd *p_FmPcd;
-+ struct fman_kg_scheme_regs schemeRegs;
-+ struct fman_kg_scheme_regs *p_MemRegs;
-+ uint8_t i;
-+ t_Error err = E_OK;
-+ uint32_t tmpKgarReg;
-+ uint32_t intFlags;
-+ uint8_t physicalSchemeId, relativeSchemeId = 0;
-+ t_FmPcdKgScheme *p_Scheme;
-+
-+ if (p_SchemeParams->modify)
-+ {
-+ p_Scheme = (t_FmPcdKgScheme *)p_SchemeParams->id.h_Scheme;
-+ p_FmPcd = p_Scheme->h_FmPcd;
-+
-+ SANITY_CHECK_RETURN_VALUE(p_FmPcd, E_INVALID_HANDLE, NULL);
-+ SANITY_CHECK_RETURN_VALUE(p_FmPcd->p_FmPcdKg, E_INVALID_HANDLE, NULL);
-+
-+ if (!FmPcdKgIsSchemeValidSw(p_Scheme))
-+ {
-+ REPORT_ERROR(MAJOR, E_ALREADY_EXISTS,
-+ ("Scheme is invalid"));
-+ return NULL;
-+ }
-+
-+ if (!KgSchemeFlagTryLock(p_Scheme))
-+ {
-+ DBG(TRACE, ("Scheme Try Lock - BUSY"));
-+ /* Signal to caller BUSY condition */
-+ p_SchemeParams->id.h_Scheme = NULL;
-+ return NULL;
-+ }
-+ }
-+ else
-+ {
-+ p_FmPcd = (t_FmPcd*)h_FmPcd;
-+
-+ SANITY_CHECK_RETURN_VALUE(p_FmPcd, E_INVALID_HANDLE, NULL);
-+ SANITY_CHECK_RETURN_VALUE(p_FmPcd->p_FmPcdKg, E_INVALID_HANDLE, NULL);
-+
-+ relativeSchemeId = p_SchemeParams->id.relativeSchemeId;
-+ /* check that schemeId is in range */
-+ if (relativeSchemeId >= p_FmPcd->p_FmPcdKg->numOfSchemes)
-+ {
-+ REPORT_ERROR(MAJOR, E_NOT_IN_RANGE, ("relative-scheme-id %d!", relativeSchemeId));
-+ return NULL;
-+ }
-+
-+ p_Scheme = &p_FmPcd->p_FmPcdKg->schemes[relativeSchemeId];
-+ if (FmPcdKgIsSchemeValidSw(p_Scheme))
-+ {
-+ REPORT_ERROR(MAJOR, E_ALREADY_EXISTS,
-+ ("Scheme id (%d)!", relativeSchemeId));
-+ return NULL;
-+ }
-+ /* Clear all fields, scheme may have beed previously used */
-+ memset(p_Scheme, 0, sizeof(t_FmPcdKgScheme));
-+
-+ p_Scheme->schemeId = p_FmPcd->p_FmPcdKg->schemesIds[relativeSchemeId];
-+ p_Scheme->h_FmPcd = p_FmPcd;
-+
-+ p_Scheme->p_Lock = FmPcdAcquireLock(p_FmPcd);
-+ if (!p_Scheme->p_Lock)
-+ REPORT_ERROR(MAJOR, E_NOT_AVAILABLE, ("FM KG Scheme lock obj!"));
-+ }
-+
-+ err = BuildSchemeRegs((t_Handle)p_Scheme, p_SchemeParams, &schemeRegs);
-+ if (err)
-+ {
-+ REPORT_ERROR(MAJOR, err, NO_MSG);
-+ if (p_SchemeParams->modify)
-+ KgSchemeFlagUnlock(p_Scheme);
-+ if (!p_SchemeParams->modify &&
-+ p_Scheme->p_Lock)
-+ FmPcdReleaseLock(p_FmPcd, p_Scheme->p_Lock);
-+ return NULL;
-+ }
-+
-+ if (p_FmPcd->h_Hc)
-+ {
-+ err = FmHcPcdKgSetScheme(p_FmPcd->h_Hc,
-+ (t_Handle)p_Scheme,
-+ &schemeRegs,
-+ p_SchemeParams->schemeCounter.update);
-+ if (p_SchemeParams->modify)
-+ KgSchemeFlagUnlock(p_Scheme);
-+ if (err)
-+ {
-+ if (!p_SchemeParams->modify &&
-+ p_Scheme->p_Lock)
-+ FmPcdReleaseLock(p_FmPcd, p_Scheme->p_Lock);
-+ return NULL;
-+ }
-+ if (!p_SchemeParams->modify)
-+ ValidateSchemeSw(p_Scheme);
-+ return (t_Handle)p_Scheme;
-+ }
-+
-+ physicalSchemeId = p_Scheme->schemeId;
-+
-+ /* configure all 21 scheme registers */
-+ p_MemRegs = &p_FmPcd->p_FmPcdKg->p_IndirectAccessRegs->schemeRegs;
-+ intFlags = KgHwLock(p_FmPcd->p_FmPcdKg);
-+ WRITE_UINT32(p_MemRegs->kgse_ppc, schemeRegs.kgse_ppc);
-+ WRITE_UINT32(p_MemRegs->kgse_ccbs, schemeRegs.kgse_ccbs);
-+ WRITE_UINT32(p_MemRegs->kgse_mode, schemeRegs.kgse_mode);
-+ WRITE_UINT32(p_MemRegs->kgse_mv, schemeRegs.kgse_mv);
-+ WRITE_UINT32(p_MemRegs->kgse_dv0, schemeRegs.kgse_dv0);
-+ WRITE_UINT32(p_MemRegs->kgse_dv1, schemeRegs.kgse_dv1);
-+ WRITE_UINT32(p_MemRegs->kgse_ekdv, schemeRegs.kgse_ekdv);
-+ WRITE_UINT32(p_MemRegs->kgse_ekfc, schemeRegs.kgse_ekfc);
-+ WRITE_UINT32(p_MemRegs->kgse_bmch, schemeRegs.kgse_bmch);
-+ WRITE_UINT32(p_MemRegs->kgse_bmcl, schemeRegs.kgse_bmcl);
-+ WRITE_UINT32(p_MemRegs->kgse_hc, schemeRegs.kgse_hc);
-+ WRITE_UINT32(p_MemRegs->kgse_spc, schemeRegs.kgse_spc);
-+ WRITE_UINT32(p_MemRegs->kgse_fqb, schemeRegs.kgse_fqb);
-+ WRITE_UINT32(p_MemRegs->kgse_om, schemeRegs.kgse_om);
-+ WRITE_UINT32(p_MemRegs->kgse_vsp, schemeRegs.kgse_vsp);
-+ for (i=0 ; i<FM_KG_NUM_OF_GENERIC_REGS ; i++)
-+ WRITE_UINT32(p_MemRegs->kgse_gec[i], schemeRegs.kgse_gec[i]);
-+
-+ /* call indirect command for scheme write */
-+ tmpKgarReg = FmPcdKgBuildWriteSchemeActionReg(physicalSchemeId, p_SchemeParams->schemeCounter.update);
-+
-+ WriteKgarWait(p_FmPcd, tmpKgarReg);
-+ KgHwUnlock(p_FmPcd->p_FmPcdKg, intFlags);
-+
-+ if (!p_SchemeParams->modify)
-+ ValidateSchemeSw(p_Scheme);
-+ else
-+ KgSchemeFlagUnlock(p_Scheme);
-+
-+ return (t_Handle)p_Scheme;
-+}
-+
-+t_Error FM_PCD_KgSchemeDelete(t_Handle h_Scheme)
-+{
-+ t_FmPcd *p_FmPcd;
-+ uint8_t physicalSchemeId;
-+ uint32_t tmpKgarReg, intFlags;
-+ t_Error err = E_OK;
-+ t_FmPcdKgScheme *p_Scheme = (t_FmPcdKgScheme *)h_Scheme;
-+
-+ SANITY_CHECK_RETURN_ERROR(h_Scheme, E_INVALID_HANDLE);
-+
-+ p_FmPcd = (t_FmPcd*)(p_Scheme->h_FmPcd);
-+
-+ UpdateRequiredActionFlag(h_Scheme, FALSE);
-+
-+ /* check that no port is bound to this scheme */
-+ err = InvalidateSchemeSw(h_Scheme);
-+ if (err)
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+
-+ if (p_FmPcd->h_Hc)
-+ {
-+ err = FmHcPcdKgDeleteScheme(p_FmPcd->h_Hc, h_Scheme);
-+ if (p_Scheme->p_Lock)
-+ FmPcdReleaseLock(p_FmPcd, p_Scheme->p_Lock);
-+ return err;
-+ }
-+
-+ physicalSchemeId = ((t_FmPcdKgScheme *)h_Scheme)->schemeId;
-+
-+ intFlags = KgHwLock(p_FmPcd->p_FmPcdKg);
-+ /* clear mode register, including enable bit */
-+ WRITE_UINT32(p_FmPcd->p_FmPcdKg->p_IndirectAccessRegs->schemeRegs.kgse_mode, 0);
-+
-+ /* call indirect command for scheme write */
-+ tmpKgarReg = FmPcdKgBuildWriteSchemeActionReg(physicalSchemeId, FALSE);
-+
-+ WriteKgarWait(p_FmPcd, tmpKgarReg);
-+ KgHwUnlock(p_FmPcd->p_FmPcdKg, intFlags);
-+
-+ if (p_Scheme->p_Lock)
-+ FmPcdReleaseLock(p_FmPcd, p_Scheme->p_Lock);
-+
-+ return E_OK;
-+}
-+
-+uint32_t FM_PCD_KgSchemeGetCounter(t_Handle h_Scheme)
-+{
-+ t_FmPcd *p_FmPcd;
-+ uint32_t tmpKgarReg, spc, intFlags;
-+ uint8_t physicalSchemeId;
-+
-+ SANITY_CHECK_RETURN_VALUE(h_Scheme, E_INVALID_HANDLE, 0);
-+
-+ p_FmPcd = (t_FmPcd*)(((t_FmPcdKgScheme *)h_Scheme)->h_FmPcd);
-+ if (p_FmPcd->h_Hc)
-+ return FmHcPcdKgGetSchemeCounter(p_FmPcd->h_Hc, h_Scheme);
-+
-+ physicalSchemeId = ((t_FmPcdKgScheme *)h_Scheme)->schemeId;
-+
-+ if (FmPcdKgGetRelativeSchemeId(p_FmPcd, physicalSchemeId) == FM_PCD_KG_NUM_OF_SCHEMES)
-+ REPORT_ERROR(MAJOR, E_NOT_IN_RANGE, NO_MSG);
-+
-+ tmpKgarReg = FmPcdKgBuildReadSchemeActionReg(physicalSchemeId);
-+ intFlags = KgHwLock(p_FmPcd->p_FmPcdKg);
-+ WriteKgarWait(p_FmPcd, tmpKgarReg);
-+ if (!(GET_UINT32(p_FmPcd->p_FmPcdKg->p_IndirectAccessRegs->schemeRegs.kgse_mode) & KG_SCH_MODE_EN))
-+ REPORT_ERROR(MAJOR, E_ALREADY_EXISTS, ("Scheme is Invalid"));
-+ spc = GET_UINT32(p_FmPcd->p_FmPcdKg->p_IndirectAccessRegs->schemeRegs.kgse_spc);
-+ KgHwUnlock(p_FmPcd->p_FmPcdKg, intFlags);
-+
-+ return spc;
-+}
-+
-+t_Error FM_PCD_KgSchemeSetCounter(t_Handle h_Scheme, uint32_t value)
-+{
-+ t_FmPcd *p_FmPcd;
-+ uint32_t tmpKgarReg, intFlags;
-+ uint8_t physicalSchemeId;
-+
-+ SANITY_CHECK_RETURN_VALUE(h_Scheme, E_INVALID_HANDLE, 0);
-+
-+ p_FmPcd = (t_FmPcd*)(((t_FmPcdKgScheme *)h_Scheme)->h_FmPcd);
-+
-+ if (!FmPcdKgIsSchemeValidSw(h_Scheme))
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("Requested scheme is invalid."));
-+
-+ if (p_FmPcd->h_Hc)
-+ return FmHcPcdKgSetSchemeCounter(p_FmPcd->h_Hc, h_Scheme, value);
-+
-+ physicalSchemeId = ((t_FmPcdKgScheme *)h_Scheme)->schemeId;
-+ /* check that schemeId is in range */
-+ if (FmPcdKgGetRelativeSchemeId(p_FmPcd, physicalSchemeId) == FM_PCD_KG_NUM_OF_SCHEMES)
-+ REPORT_ERROR(MAJOR, E_NOT_IN_RANGE, NO_MSG);
-+
-+ /* read specified scheme into scheme registers */
-+ tmpKgarReg = FmPcdKgBuildReadSchemeActionReg(physicalSchemeId);
-+ intFlags = KgHwLock(p_FmPcd->p_FmPcdKg);
-+ WriteKgarWait(p_FmPcd, tmpKgarReg);
-+ if (!(GET_UINT32(p_FmPcd->p_FmPcdKg->p_IndirectAccessRegs->schemeRegs.kgse_mode) & KG_SCH_MODE_EN))
-+ {
-+ KgHwUnlock(p_FmPcd->p_FmPcdKg, intFlags);
-+ RETURN_ERROR(MAJOR, E_ALREADY_EXISTS, ("Scheme is Invalid"));
-+ }
-+
-+ /* change counter value */
-+ WRITE_UINT32(p_FmPcd->p_FmPcdKg->p_IndirectAccessRegs->schemeRegs.kgse_spc, value);
-+
-+ /* call indirect command for scheme write */
-+ tmpKgarReg = FmPcdKgBuildWriteSchemeActionReg(physicalSchemeId, TRUE);
-+
-+ WriteKgarWait(p_FmPcd, tmpKgarReg);
-+ KgHwUnlock(p_FmPcd->p_FmPcdKg, intFlags);
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PCD_KgSetAdditionalDataAfterParsing(t_Handle h_FmPcd, uint8_t payloadOffset)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+ struct fman_kg_regs *p_Regs;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmPcd->p_FmPcdDriverParam, E_NULL_POINTER);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd->p_FmPcdKg, E_NULL_POINTER);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd->p_FmPcdKg->p_FmPcdKgRegs, E_NULL_POINTER);
-+
-+ p_Regs = p_FmPcd->p_FmPcdKg->p_FmPcdKgRegs;
-+ if (!FmIsMaster(p_FmPcd->h_Fm))
-+ RETURN_ERROR(MAJOR, E_NOT_SUPPORTED, ("FM_PCD_KgSetAdditionalDataAfterParsing - guest mode!"));
-+
-+ WRITE_UINT32(p_Regs->fmkg_fdor,payloadOffset);
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PCD_KgSetDfltValue(t_Handle h_FmPcd, uint8_t valueId, uint32_t value)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+ struct fman_kg_regs *p_Regs;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(((valueId == 0) || (valueId == 1)), E_INVALID_VALUE);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmPcd->p_FmPcdDriverParam, E_NULL_POINTER);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd->p_FmPcdKg, E_NULL_POINTER);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd->p_FmPcdKg->p_FmPcdKgRegs, E_NULL_POINTER);
-+
-+ p_Regs = p_FmPcd->p_FmPcdKg->p_FmPcdKgRegs;
-+
-+ if (!FmIsMaster(p_FmPcd->h_Fm))
-+ RETURN_ERROR(MAJOR, E_NOT_SUPPORTED, ("FM_PCD_KgSetDfltValue - guest mode!"));
-+
-+ if (valueId == 0)
-+ WRITE_UINT32(p_Regs->fmkg_gdv0r,value);
-+ else
-+ WRITE_UINT32(p_Regs->fmkg_gdv1r,value);
-+ return E_OK;
-+}
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/Pcd/fm_kg.h
-@@ -0,0 +1,206 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/******************************************************************************
-+ @File fm_kg.h
-+
-+ @Description FM KG private header
-+*//***************************************************************************/
-+#ifndef __FM_KG_H
-+#define __FM_KG_H
-+
-+#include "std_ext.h"
-+
-+/***********************************************************************/
-+/* Keygen defines */
-+/***********************************************************************/
-+/* maskes */
-+#if (DPAA_VERSION >= 11)
-+#define KG_SCH_VSP_SHIFT_MASK 0x0003f000
-+#define KG_SCH_OM_VSPE 0x00000001
-+#define KG_SCH_VSP_NO_KSP_EN 0x80000000
-+
-+#define MAX_SP_SHIFT 23
-+#define KG_SCH_VSP_MASK_SHIFT 12
-+#define KG_SCH_VSP_SHIFT 24
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+typedef uint32_t t_KnownFieldsMasks;
-+#define KG_SCH_KN_PORT_ID 0x80000000
-+#define KG_SCH_KN_MACDST 0x40000000
-+#define KG_SCH_KN_MACSRC 0x20000000
-+#define KG_SCH_KN_TCI1 0x10000000
-+#define KG_SCH_KN_TCI2 0x08000000
-+#define KG_SCH_KN_ETYPE 0x04000000
-+#define KG_SCH_KN_PPPSID 0x02000000
-+#define KG_SCH_KN_PPPID 0x01000000
-+#define KG_SCH_KN_MPLS1 0x00800000
-+#define KG_SCH_KN_MPLS2 0x00400000
-+#define KG_SCH_KN_MPLS_LAST 0x00200000
-+#define KG_SCH_KN_IPSRC1 0x00100000
-+#define KG_SCH_KN_IPDST1 0x00080000
-+#define KG_SCH_KN_PTYPE1 0x00040000
-+#define KG_SCH_KN_IPTOS_TC1 0x00020000
-+#define KG_SCH_KN_IPV6FL1 0x00010000
-+#define KG_SCH_KN_IPSRC2 0x00008000
-+#define KG_SCH_KN_IPDST2 0x00004000
-+#define KG_SCH_KN_PTYPE2 0x00002000
-+#define KG_SCH_KN_IPTOS_TC2 0x00001000
-+#define KG_SCH_KN_IPV6FL2 0x00000800
-+#define KG_SCH_KN_GREPTYPE 0x00000400
-+#define KG_SCH_KN_IPSEC_SPI 0x00000200
-+#define KG_SCH_KN_IPSEC_NH 0x00000100
-+#define KG_SCH_KN_IPPID 0x00000080
-+#define KG_SCH_KN_L4PSRC 0x00000004
-+#define KG_SCH_KN_L4PDST 0x00000002
-+#define KG_SCH_KN_TFLG 0x00000001
-+
-+typedef uint8_t t_GenericCodes;
-+#define KG_SCH_GEN_SHIM1 0x70
-+#define KG_SCH_GEN_DEFAULT 0x10
-+#define KG_SCH_GEN_PARSE_RESULT_N_FQID 0x20
-+#define KG_SCH_GEN_START_OF_FRM 0x40
-+#define KG_SCH_GEN_SHIM2 0x71
-+#define KG_SCH_GEN_IP_PID_NO_V 0x72
-+#define KG_SCH_GEN_ETH 0x03
-+#define KG_SCH_GEN_ETH_NO_V 0x73
-+#define KG_SCH_GEN_SNAP 0x04
-+#define KG_SCH_GEN_SNAP_NO_V 0x74
-+#define KG_SCH_GEN_VLAN1 0x05
-+#define KG_SCH_GEN_VLAN1_NO_V 0x75
-+#define KG_SCH_GEN_VLAN2 0x06
-+#define KG_SCH_GEN_VLAN2_NO_V 0x76
-+#define KG_SCH_GEN_ETH_TYPE 0x07
-+#define KG_SCH_GEN_ETH_TYPE_NO_V 0x77
-+#define KG_SCH_GEN_PPP 0x08
-+#define KG_SCH_GEN_PPP_NO_V 0x78
-+#define KG_SCH_GEN_MPLS1 0x09
-+#define KG_SCH_GEN_MPLS2 0x19
-+#define KG_SCH_GEN_MPLS3 0x29
-+#define KG_SCH_GEN_MPLS1_NO_V 0x79
-+#define KG_SCH_GEN_MPLS_LAST 0x0a
-+#define KG_SCH_GEN_MPLS_LAST_NO_V 0x7a
-+#define KG_SCH_GEN_IPV4 0x0b
-+#define KG_SCH_GEN_IPV6 0x1b
-+#define KG_SCH_GEN_L3_NO_V 0x7b
-+#define KG_SCH_GEN_IPV4_TUNNELED 0x0c
-+#define KG_SCH_GEN_IPV6_TUNNELED 0x1c
-+#define KG_SCH_GEN_MIN_ENCAP 0x2c
-+#define KG_SCH_GEN_IP2_NO_V 0x7c
-+#define KG_SCH_GEN_GRE 0x0d
-+#define KG_SCH_GEN_GRE_NO_V 0x7d
-+#define KG_SCH_GEN_TCP 0x0e
-+#define KG_SCH_GEN_UDP 0x1e
-+#define KG_SCH_GEN_IPSEC_AH 0x2e
-+#define KG_SCH_GEN_SCTP 0x3e
-+#define KG_SCH_GEN_DCCP 0x4e
-+#define KG_SCH_GEN_IPSEC_ESP 0x6e
-+#define KG_SCH_GEN_L4_NO_V 0x7e
-+#define KG_SCH_GEN_NEXTHDR 0x7f
-+/* shifts */
-+#define KG_SCH_PP_SHIFT_HIGH_SHIFT 27
-+#define KG_SCH_PP_SHIFT_LOW_SHIFT 12
-+#define KG_SCH_PP_MASK_SHIFT 16
-+#define KG_SCH_MODE_CCOBASE_SHIFT 24
-+#define KG_SCH_DEF_MAC_ADDR_SHIFT 30
-+#define KG_SCH_DEF_TCI_SHIFT 28
-+#define KG_SCH_DEF_ENET_TYPE_SHIFT 26
-+#define KG_SCH_DEF_PPP_SESSION_ID_SHIFT 24
-+#define KG_SCH_DEF_PPP_PROTOCOL_ID_SHIFT 22
-+#define KG_SCH_DEF_MPLS_LABEL_SHIFT 20
-+#define KG_SCH_DEF_IP_ADDR_SHIFT 18
-+#define KG_SCH_DEF_PROTOCOL_TYPE_SHIFT 16
-+#define KG_SCH_DEF_IP_TOS_TC_SHIFT 14
-+#define KG_SCH_DEF_IPV6_FLOW_LABEL_SHIFT 12
-+#define KG_SCH_DEF_IPSEC_SPI_SHIFT 10
-+#define KG_SCH_DEF_L4_PORT_SHIFT 8
-+#define KG_SCH_DEF_TCP_FLAG_SHIFT 6
-+#define KG_SCH_HASH_CONFIG_SHIFT_SHIFT 24
-+#define KG_SCH_GEN_MASK_SHIFT 16
-+#define KG_SCH_GEN_HT_SHIFT 8
-+#define KG_SCH_GEN_SIZE_SHIFT 24
-+#define KG_SCH_GEN_DEF_SHIFT 29
-+#define FM_PCD_KG_KGAR_NUM_SHIFT 16
-+
-+/* others */
-+#define NUM_OF_SW_DEFAULTS 3
-+#define MAX_PP_SHIFT 23
-+#define MAX_KG_SCH_SIZE 16
-+#define MASK_FOR_GENERIC_BASE_ID 0x20
-+#define MAX_HASH_SHIFT 40
-+#define MAX_KG_SCH_FQID_BIT_OFFSET 31
-+#define MAX_KG_SCH_PP_BIT_OFFSET 15
-+#define MAX_DIST_FQID_SHIFT 23
-+
-+#define GET_MASK_SEL_SHIFT(shift,i) \
-+switch (i) { \
-+ case (0):shift = 26;break; \
-+ case (1):shift = 20;break; \
-+ case (2):shift = 10;break; \
-+ case (3):shift = 4;break; \
-+ default: \
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, NO_MSG); \
-+}
-+
-+#define GET_MASK_OFFSET_SHIFT(shift,i) \
-+switch (i) { \
-+ case (0):shift = 16;break; \
-+ case (1):shift = 0;break; \
-+ case (2):shift = 28;break; \
-+ case (3):shift = 24;break; \
-+ default: \
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, NO_MSG); \
-+}
-+
-+#define GET_MASK_SHIFT(shift,i) \
-+switch (i) { \
-+ case (0):shift = 24;break; \
-+ case (1):shift = 16;break; \
-+ case (2):shift = 8;break; \
-+ case (3):shift = 0;break; \
-+ default: \
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, NO_MSG); \
-+}
-+
-+/***********************************************************************/
-+/* Keygen defines */
-+/***********************************************************************/
-+
-+#define KG_DOUBLE_MEANING_REGS_OFFSET 0x100
-+#define NO_VALIDATION 0x70
-+#define KG_ACTION_REG_TO 1024
-+#define KG_MAX_PROFILE 255
-+#define SCHEME_ALWAYS_DIRECT 0xFFFFFFFF
-+
-+
-+#endif /* __FM_KG_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/Pcd/fm_manip.c
-@@ -0,0 +1,5571 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/******************************************************************************
-+ @File fm_manip.c
-+
-+ @Description FM PCD manip ...
-+ *//***************************************************************************/
-+#include "std_ext.h"
-+#include "error_ext.h"
-+#include "string_ext.h"
-+#include "debug_ext.h"
-+#include "fm_pcd_ext.h"
-+#include "fm_port_ext.h"
-+#include "fm_muram_ext.h"
-+#include "memcpy_ext.h"
-+
-+#include "fm_common.h"
-+#include "fm_hc.h"
-+#include "fm_manip.h"
-+
-+/****************************************/
-+/* static functions */
-+/****************************************/
-+static t_Handle GetManipInfo(t_FmPcdManip *p_Manip, e_ManipInfo manipInfo)
-+{
-+ t_FmPcdManip *p_CurManip = p_Manip;
-+
-+ if (!MANIP_IS_UNIFIED(p_Manip))
-+ p_CurManip = p_Manip;
-+ else
-+ {
-+ /* go to first unified */
-+ while (MANIP_IS_UNIFIED_NON_FIRST(p_CurManip))
-+ p_CurManip = p_CurManip->h_PrevManip;
-+ }
-+
-+ switch (manipInfo)
-+ {
-+ case (e_MANIP_HMCT):
-+ return p_CurManip->p_Hmct;
-+ case (e_MANIP_HMTD):
-+ return p_CurManip->h_Ad;
-+ case (e_MANIP_HANDLER_TABLE_OWNER):
-+ return (t_Handle)p_CurManip;
-+ default:
-+ return NULL;
-+ }
-+}
-+
-+static uint16_t GetHmctSize(t_FmPcdManip *p_Manip)
-+{
-+ uint16_t size = 0;
-+ t_FmPcdManip *p_CurManip = p_Manip;
-+
-+ if (!MANIP_IS_UNIFIED(p_Manip))
-+ return p_Manip->tableSize;
-+
-+ /* accumulate sizes, starting with the first node */
-+ while (MANIP_IS_UNIFIED_NON_FIRST(p_CurManip))
-+ p_CurManip = p_CurManip->h_PrevManip;
-+
-+ while (MANIP_IS_UNIFIED_NON_LAST(p_CurManip))
-+ {
-+ size += p_CurManip->tableSize;
-+ p_CurManip = (t_FmPcdManip *)p_CurManip->h_NextManip;
-+ }
-+ size += p_CurManip->tableSize; /* add last size */
-+
-+ return (size);
-+}
-+
-+static uint16_t GetDataSize(t_FmPcdManip *p_Manip)
-+{
-+ uint16_t size = 0;
-+ t_FmPcdManip *p_CurManip = p_Manip;
-+
-+ if (!MANIP_IS_UNIFIED(p_Manip))
-+ return p_Manip->dataSize;
-+
-+ /* accumulate sizes, starting with the first node */
-+ while (MANIP_IS_UNIFIED_NON_FIRST(p_CurManip))
-+ p_CurManip = p_CurManip->h_PrevManip;
-+
-+ while (MANIP_IS_UNIFIED_NON_LAST(p_CurManip))
-+ {
-+ size += p_CurManip->dataSize;
-+ p_CurManip = (t_FmPcdManip *)p_CurManip->h_NextManip;
-+ }
-+ size += p_CurManip->dataSize; /* add last size */
-+
-+ return (size);
-+}
-+
-+static t_Error CalculateTableSize(t_FmPcdManipParams *p_FmPcdManipParams,
-+ uint16_t *p_TableSize, uint8_t *p_DataSize)
-+{
-+ uint8_t localDataSize, remain, tableSize = 0, dataSize = 0;
-+
-+ if (p_FmPcdManipParams->u.hdr.rmv)
-+ {
-+ switch (p_FmPcdManipParams->u.hdr.rmvParams.type)
-+ {
-+ case (e_FM_PCD_MANIP_RMV_GENERIC):
-+ tableSize += HMCD_BASIC_SIZE;
-+ break;
-+ case (e_FM_PCD_MANIP_RMV_BY_HDR):
-+ switch (p_FmPcdManipParams->u.hdr.rmvParams.u.byHdr.type)
-+ {
-+ case (e_FM_PCD_MANIP_RMV_BY_HDR_SPECIFIC_L2):
-+#if (DPAA_VERSION >= 11)
-+ case (e_FM_PCD_MANIP_RMV_BY_HDR_CAPWAP):
-+ case (e_FM_PCD_MANIP_RMV_BY_HDR_FROM_START):
-+#endif /* (DPAA_VERSION >= 11) */
-+ tableSize += HMCD_BASIC_SIZE;
-+ break;
-+ default:
-+ RETURN_ERROR(MINOR, E_INVALID_SELECTION,
-+ ("Unknown byHdr.type"));
-+ }
-+ break;
-+ default:
-+ RETURN_ERROR(MINOR, E_INVALID_SELECTION,
-+ ("Unknown rmvParams.type"));
-+ }
-+ }
-+
-+ if (p_FmPcdManipParams->u.hdr.insrt)
-+ {
-+ switch (p_FmPcdManipParams->u.hdr.insrtParams.type)
-+ {
-+ case (e_FM_PCD_MANIP_INSRT_GENERIC):
-+ remain =
-+ (uint8_t)(p_FmPcdManipParams->u.hdr.insrtParams.u.generic.size
-+ % 4);
-+ if (remain)
-+ localDataSize =
-+ (uint8_t)(p_FmPcdManipParams->u.hdr.insrtParams.u.generic.size
-+ + 4 - remain);
-+ else
-+ localDataSize =
-+ p_FmPcdManipParams->u.hdr.insrtParams.u.generic.size;
-+ tableSize += (uint8_t)(HMCD_BASIC_SIZE + localDataSize);
-+ break;
-+ case (e_FM_PCD_MANIP_INSRT_BY_HDR):
-+ {
-+ switch (p_FmPcdManipParams->u.hdr.insrtParams.u.byHdr.type)
-+ {
-+
-+ case (e_FM_PCD_MANIP_INSRT_BY_HDR_SPECIFIC_L2):
-+ tableSize += HMCD_BASIC_SIZE + HMCD_PTR_SIZE;
-+ switch (p_FmPcdManipParams->u.hdr.insrtParams.u.byHdr.u.specificL2Params.specificL2)
-+ {
-+ case (e_FM_PCD_MANIP_HDR_INSRT_MPLS):
-+ case (e_FM_PCD_MANIP_HDR_INSRT_PPPOE):
-+ dataSize +=
-+ p_FmPcdManipParams->u.hdr.insrtParams.u.byHdr.u.specificL2Params.size;
-+ break;
-+ default:
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, NO_MSG);
-+ }
-+ break;
-+#if (DPAA_VERSION >= 11)
-+ case (e_FM_PCD_MANIP_INSRT_BY_HDR_IP):
-+ tableSize +=
-+ (HMCD_BASIC_SIZE + HMCD_PTR_SIZE
-+ + HMCD_PARAM_SIZE
-+ + p_FmPcdManipParams->u.hdr.insrtParams.u.byHdr.u.ipParams.insrt.size);
-+ dataSize += 2;
-+ break;
-+
-+ case (e_FM_PCD_MANIP_INSRT_BY_HDR_UDP):
-+ case (e_FM_PCD_MANIP_INSRT_BY_HDR_UDP_LITE):
-+ tableSize += (HMCD_BASIC_SIZE + HMCD_L4_HDR_SIZE);
-+
-+ break;
-+
-+ case (e_FM_PCD_MANIP_INSRT_BY_HDR_CAPWAP):
-+ tableSize +=
-+ (HMCD_BASIC_SIZE
-+ + p_FmPcdManipParams->u.hdr.insrtParams.u.byHdr.u.insrt.size);
-+ break;
-+#endif /* (DPAA_VERSION >= 11) */
-+ default:
-+ RETURN_ERROR(MINOR, E_INVALID_SELECTION,
-+ ("Unknown byHdr.type"));
-+ }
-+ }
-+ break;
-+ default:
-+ RETURN_ERROR(MINOR, E_INVALID_SELECTION,
-+ ("Unknown insrtParams.type"));
-+ }
-+ }
-+
-+ if (p_FmPcdManipParams->u.hdr.fieldUpdate)
-+ {
-+ switch (p_FmPcdManipParams->u.hdr.fieldUpdateParams.type)
-+ {
-+ case (e_FM_PCD_MANIP_HDR_FIELD_UPDATE_VLAN):
-+ tableSize += HMCD_BASIC_SIZE;
-+ if (p_FmPcdManipParams->u.hdr.fieldUpdateParams.u.vlan.updateType
-+ == e_FM_PCD_MANIP_HDR_FIELD_UPDATE_DSCP_TO_VLAN)
-+ {
-+ tableSize += HMCD_PTR_SIZE;
-+ dataSize += DSCP_TO_VLAN_TABLE_SIZE;
-+ }
-+ break;
-+ case (e_FM_PCD_MANIP_HDR_FIELD_UPDATE_IPV4):
-+ tableSize += HMCD_BASIC_SIZE;
-+ if (p_FmPcdManipParams->u.hdr.fieldUpdateParams.u.ipv4.validUpdates
-+ & HDR_MANIP_IPV4_ID)
-+ {
-+ tableSize += HMCD_PARAM_SIZE;
-+ dataSize += 2;
-+ }
-+ if (p_FmPcdManipParams->u.hdr.fieldUpdateParams.u.ipv4.validUpdates
-+ & HDR_MANIP_IPV4_SRC)
-+ tableSize += HMCD_IPV4_ADDR_SIZE;
-+ if (p_FmPcdManipParams->u.hdr.fieldUpdateParams.u.ipv4.validUpdates
-+ & HDR_MANIP_IPV4_DST)
-+ tableSize += HMCD_IPV4_ADDR_SIZE;
-+ break;
-+ case (e_FM_PCD_MANIP_HDR_FIELD_UPDATE_IPV6):
-+ tableSize += HMCD_BASIC_SIZE;
-+ if (p_FmPcdManipParams->u.hdr.fieldUpdateParams.u.ipv4.validUpdates
-+ & HDR_MANIP_IPV6_SRC)
-+ tableSize += HMCD_IPV6_ADDR_SIZE;
-+ if (p_FmPcdManipParams->u.hdr.fieldUpdateParams.u.ipv4.validUpdates
-+ & HDR_MANIP_IPV6_DST)
-+ tableSize += HMCD_IPV6_ADDR_SIZE;
-+ break;
-+ case (e_FM_PCD_MANIP_HDR_FIELD_UPDATE_TCP_UDP):
-+ if (p_FmPcdManipParams->u.hdr.fieldUpdateParams.u.tcpUdp.validUpdates
-+ == HDR_MANIP_TCP_UDP_CHECKSUM)
-+ /* we implement this case with the update-checksum descriptor */
-+ tableSize += HMCD_BASIC_SIZE;
-+ else
-+ /* we implement this case with the TCP/UDP-update descriptor */
-+ tableSize += HMCD_BASIC_SIZE + HMCD_PARAM_SIZE;
-+ break;
-+ default:
-+ RETURN_ERROR(MINOR, E_INVALID_SELECTION,
-+ ("Unknown fieldUpdateParams.type"));
-+ }
-+ }
-+
-+ if (p_FmPcdManipParams->u.hdr.custom)
-+ {
-+ switch (p_FmPcdManipParams->u.hdr.customParams.type)
-+ {
-+ case (e_FM_PCD_MANIP_HDR_CUSTOM_IP_REPLACE):
-+ {
-+ tableSize += HMCD_BASIC_SIZE + HMCD_PARAM_SIZE + HMCD_PARAM_SIZE;
-+ dataSize +=
-+ p_FmPcdManipParams->u.hdr.customParams.u.ipHdrReplace.hdrSize;
-+ if ((p_FmPcdManipParams->u.hdr.customParams.u.ipHdrReplace.replaceType
-+ == e_FM_PCD_MANIP_HDR_CUSTOM_REPLACE_IPV6_BY_IPV4)
-+ && (p_FmPcdManipParams->u.hdr.customParams.u.ipHdrReplace.updateIpv4Id))
-+ dataSize += 2;
-+ }
-+ break;
-+ case (e_FM_PCD_MANIP_HDR_CUSTOM_GEN_FIELD_REPLACE):
-+ tableSize += HMCD_BASIC_SIZE + HMCD_PARAM_SIZE;
-+ break;
-+ default:
-+ RETURN_ERROR(MINOR, E_INVALID_SELECTION,
-+ ("Unknown customParams.type"));
-+ }
-+ }
-+
-+ *p_TableSize = tableSize;
-+ *p_DataSize = dataSize;
-+
-+ return E_OK;
-+}
-+
-+static t_Error GetPrOffsetByHeaderOrField(t_FmManipHdrInfo *p_HdrInfo,
-+ uint8_t *parseArrayOffset)
-+{
-+ e_NetHeaderType hdr = p_HdrInfo->hdr;
-+ e_FmPcdHdrIndex hdrIndex = p_HdrInfo->hdrIndex;
-+ bool byField = p_HdrInfo->byField;
-+ t_FmPcdFields field;
-+
-+ if (byField)
-+ field = p_HdrInfo->fullField;
-+
-+ if (byField)
-+ {
-+ switch (hdr)
-+ {
-+ case (HEADER_TYPE_ETH):
-+ switch (field.eth)
-+ {
-+ case (NET_HEADER_FIELD_ETH_TYPE):
-+ *parseArrayOffset = CC_PC_PR_ETYPE_LAST_OFFSET;
-+ break;
-+ default:
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_NOT_SUPPORTED,
-+ ("Header manipulation of the type Ethernet with this field not supported"));
-+ }
-+ break;
-+ case (HEADER_TYPE_VLAN):
-+ switch (field.vlan)
-+ {
-+ case (NET_HEADER_FIELD_VLAN_TCI):
-+ if ((hdrIndex == e_FM_PCD_HDR_INDEX_NONE)
-+ || (hdrIndex == e_FM_PCD_HDR_INDEX_1))
-+ *parseArrayOffset = CC_PC_PR_VLAN1_OFFSET;
-+ else
-+ if (hdrIndex == e_FM_PCD_HDR_INDEX_LAST)
-+ *parseArrayOffset = CC_PC_PR_VLAN2_OFFSET;
-+ break;
-+ default:
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_NOT_SUPPORTED,
-+ ("Header manipulation of the type VLAN with this field not supported"));
-+ }
-+ break;
-+ default:
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_NOT_SUPPORTED,
-+ ("Header manipulation of this header by field not supported"));
-+ }
-+ }
-+ else
-+ {
-+ switch (hdr)
-+ {
-+ case (HEADER_TYPE_ETH):
-+ *parseArrayOffset = (uint8_t)CC_PC_PR_ETH_OFFSET;
-+ break;
-+ case (HEADER_TYPE_USER_DEFINED_SHIM1):
-+ *parseArrayOffset = (uint8_t)CC_PC_PR_USER_DEFINED_SHIM1_OFFSET;
-+ break;
-+ case (HEADER_TYPE_USER_DEFINED_SHIM2):
-+ *parseArrayOffset = (uint8_t)CC_PC_PR_USER_DEFINED_SHIM2_OFFSET;
-+ break;
-+ case (HEADER_TYPE_LLC_SNAP):
-+ *parseArrayOffset = CC_PC_PR_USER_LLC_SNAP_OFFSET;
-+ break;
-+ case (HEADER_TYPE_PPPoE):
-+ *parseArrayOffset = CC_PC_PR_PPPOE_OFFSET;
-+ break;
-+ case (HEADER_TYPE_MPLS):
-+ if ((hdrIndex == e_FM_PCD_HDR_INDEX_NONE)
-+ || (hdrIndex == e_FM_PCD_HDR_INDEX_1))
-+ *parseArrayOffset = CC_PC_PR_MPLS1_OFFSET;
-+ else
-+ if (hdrIndex == e_FM_PCD_HDR_INDEX_LAST)
-+ *parseArrayOffset = CC_PC_PR_MPLS_LAST_OFFSET;
-+ break;
-+ case (HEADER_TYPE_IPv4):
-+ case (HEADER_TYPE_IPv6):
-+ if ((hdrIndex == e_FM_PCD_HDR_INDEX_NONE)
-+ || (hdrIndex == e_FM_PCD_HDR_INDEX_1))
-+ *parseArrayOffset = CC_PC_PR_IP1_OFFSET;
-+ else
-+ if (hdrIndex == e_FM_PCD_HDR_INDEX_2)
-+ *parseArrayOffset = CC_PC_PR_IP_LAST_OFFSET;
-+ break;
-+ case (HEADER_TYPE_MINENCAP):
-+ *parseArrayOffset = CC_PC_PR_MINENC_OFFSET;
-+ break;
-+ case (HEADER_TYPE_GRE):
-+ *parseArrayOffset = CC_PC_PR_GRE_OFFSET;
-+ break;
-+ case (HEADER_TYPE_TCP):
-+ case (HEADER_TYPE_UDP):
-+ case (HEADER_TYPE_IPSEC_AH):
-+ case (HEADER_TYPE_IPSEC_ESP):
-+ case (HEADER_TYPE_DCCP):
-+ case (HEADER_TYPE_SCTP):
-+ *parseArrayOffset = CC_PC_PR_L4_OFFSET;
-+ break;
-+ case (HEADER_TYPE_CAPWAP):
-+ case (HEADER_TYPE_CAPWAP_DTLS):
-+ *parseArrayOffset = CC_PC_PR_NEXT_HEADER_OFFSET;
-+ break;
-+ default:
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_NOT_SUPPORTED,
-+ ("Header manipulation of this header is not supported"));
-+ }
-+ }
-+ return E_OK;
-+}
-+
-+static t_Error BuildHmct(t_FmPcdManip *p_Manip,
-+ t_FmPcdManipParams *p_FmPcdManipParams,
-+ uint8_t *p_DestHmct, uint8_t *p_DestData, bool new)
-+{
-+ uint32_t *p_TmpHmct = (uint32_t*)p_DestHmct, *p_LocalData;
-+ uint32_t tmpReg = 0, *p_Last = NULL, tmp_ipv6_addr;
-+ uint8_t remain, i, size = 0, origSize, *p_UsrData = NULL, *p_TmpData =
-+ p_DestData;
-+ t_Handle h_FmPcd = p_Manip->h_FmPcd;
-+ uint8_t j = 0;
-+
-+ if (p_FmPcdManipParams->u.hdr.rmv)
-+ {
-+ if (p_FmPcdManipParams->u.hdr.rmvParams.type
-+ == e_FM_PCD_MANIP_RMV_GENERIC)
-+ {
-+ /* initialize HMCD */
-+ tmpReg = (uint32_t)(HMCD_OPCODE_GENERIC_RMV) << HMCD_OC_SHIFT;
-+ /* tmp, should be conditional */
-+ tmpReg |= p_FmPcdManipParams->u.hdr.rmvParams.u.generic.offset
-+ << HMCD_RMV_OFFSET_SHIFT;
-+ tmpReg |= p_FmPcdManipParams->u.hdr.rmvParams.u.generic.size
-+ << HMCD_RMV_SIZE_SHIFT;
-+ }
-+ else
-+ if (p_FmPcdManipParams->u.hdr.rmvParams.type
-+ == e_FM_PCD_MANIP_RMV_BY_HDR)
-+ {
-+ switch (p_FmPcdManipParams->u.hdr.rmvParams.u.byHdr.type)
-+ {
-+ case (e_FM_PCD_MANIP_RMV_BY_HDR_SPECIFIC_L2):
-+ {
-+ uint8_t hmcdOpt;
-+
-+ /* initialize HMCD */
-+ tmpReg = (uint32_t)(HMCD_OPCODE_L2_RMV) << HMCD_OC_SHIFT;
-+
-+ switch (p_FmPcdManipParams->u.hdr.rmvParams.u.byHdr.u.specificL2)
-+ {
-+ case (e_FM_PCD_MANIP_HDR_RMV_ETHERNET):
-+ hmcdOpt = HMCD_RMV_L2_ETHERNET;
-+ break;
-+ case (e_FM_PCD_MANIP_HDR_RMV_STACKED_QTAGS):
-+ hmcdOpt = HMCD_RMV_L2_STACKED_QTAGS;
-+ break;
-+ case (e_FM_PCD_MANIP_HDR_RMV_ETHERNET_AND_MPLS):
-+ hmcdOpt = HMCD_RMV_L2_ETHERNET_AND_MPLS;
-+ break;
-+ case (e_FM_PCD_MANIP_HDR_RMV_MPLS):
-+ hmcdOpt = HMCD_RMV_L2_MPLS;
-+ break;
-+ case (e_FM_PCD_MANIP_HDR_RMV_PPPOE):
-+ hmcdOpt = HMCD_RMV_L2_PPPOE;
-+ break;
-+ default:
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, NO_MSG);
-+ }
-+ tmpReg |= hmcdOpt << HMCD_L2_MODE_SHIFT;
-+ break;
-+ }
-+#if (DPAA_VERSION >= 11)
-+ case (e_FM_PCD_MANIP_RMV_BY_HDR_CAPWAP):
-+ tmpReg = (uint32_t)(HMCD_OPCODE_CAPWAP_RMV)
-+ << HMCD_OC_SHIFT;
-+ break;
-+ case (e_FM_PCD_MANIP_RMV_BY_HDR_FROM_START):
-+ {
-+ uint8_t prsArrayOffset;
-+ t_Error err = E_OK;
-+
-+ tmpReg = (uint32_t)(HMCD_OPCODE_RMV_TILL)
-+ << HMCD_OC_SHIFT;
-+
-+ err =
-+ GetPrOffsetByHeaderOrField(
-+ &p_FmPcdManipParams->u.hdr.rmvParams.u.byHdr.u.hdrInfo,
-+ &prsArrayOffset);
-+ ASSERT_COND(!err);
-+ /* was previously checked */
-+
-+ tmpReg |= ((uint32_t)prsArrayOffset << 16);
-+ }
-+ break;
-+#endif /* (DPAA_VERSION >= 11) */
-+ default:
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED,
-+ ("manip header remove by hdr type!"));
-+ }
-+ }
-+
-+ WRITE_UINT32(*p_TmpHmct, tmpReg);
-+ /* save a pointer to the "last" indication word */
-+ p_Last = p_TmpHmct;
-+ /* advance to next command */
-+ p_TmpHmct += HMCD_BASIC_SIZE / 4;
-+ }
-+
-+ if (p_FmPcdManipParams->u.hdr.insrt)
-+ {
-+ if (p_FmPcdManipParams->u.hdr.insrtParams.type
-+ == e_FM_PCD_MANIP_INSRT_GENERIC)
-+ {
-+ /* initialize HMCD */
-+ if (p_FmPcdManipParams->u.hdr.insrtParams.u.generic.replace)
-+ tmpReg = (uint32_t)(HMCD_OPCODE_GENERIC_REPLACE)
-+ << HMCD_OC_SHIFT;
-+ else
-+ tmpReg = (uint32_t)(HMCD_OPCODE_GENERIC_INSRT) << HMCD_OC_SHIFT;
-+
-+ tmpReg |= p_FmPcdManipParams->u.hdr.insrtParams.u.generic.offset
-+ << HMCD_INSRT_OFFSET_SHIFT;
-+ tmpReg |= p_FmPcdManipParams->u.hdr.insrtParams.u.generic.size
-+ << HMCD_INSRT_SIZE_SHIFT;
-+
-+ size = p_FmPcdManipParams->u.hdr.insrtParams.u.generic.size;
-+ p_UsrData = p_FmPcdManipParams->u.hdr.insrtParams.u.generic.p_Data;
-+
-+ WRITE_UINT32(*p_TmpHmct, tmpReg);
-+ /* save a pointer to the "last" indication word */
-+ p_Last = p_TmpHmct;
-+
-+ p_TmpHmct += HMCD_BASIC_SIZE / 4;
-+
-+ /* initialize data to be inserted */
-+ /* if size is not a multiple of 4, padd with 0's */
-+ origSize = size;
-+ remain = (uint8_t)(size % 4);
-+ if (remain)
-+ {
-+ size += (uint8_t)(4 - remain);
-+ p_LocalData = (uint32_t *)XX_Malloc(size);
-+ memset((uint8_t *)p_LocalData, 0, size);
-+ memcpy((uint8_t *)p_LocalData, p_UsrData, origSize);
-+ }
-+ else
-+ p_LocalData = (uint32_t*)p_UsrData;
-+
-+ /* initialize data and advance pointer to next command */
-+ MemCpy8(p_TmpHmct, p_LocalData, size);
-+ p_TmpHmct += size / sizeof(uint32_t);
-+
-+ if (remain)
-+ XX_Free(p_LocalData);
-+ }
-+
-+ else
-+ if (p_FmPcdManipParams->u.hdr.insrtParams.type
-+ == e_FM_PCD_MANIP_INSRT_BY_HDR)
-+ {
-+ switch (p_FmPcdManipParams->u.hdr.insrtParams.u.byHdr.type)
-+ {
-+ case (e_FM_PCD_MANIP_INSRT_BY_HDR_SPECIFIC_L2):
-+ {
-+ uint8_t hmcdOpt;
-+
-+ /* initialize HMCD */
-+ tmpReg = (uint32_t)(HMCD_OPCODE_L2_INSRT)
-+ << HMCD_OC_SHIFT;
-+
-+ switch (p_FmPcdManipParams->u.hdr.insrtParams.u.byHdr.u.specificL2Params.specificL2)
-+ {
-+ case (e_FM_PCD_MANIP_HDR_INSRT_MPLS):
-+ if (p_FmPcdManipParams->u.hdr.insrtParams.u.byHdr.u.specificL2Params.update)
-+ hmcdOpt = HMCD_INSRT_N_UPDATE_L2_MPLS;
-+ else
-+ hmcdOpt = HMCD_INSRT_L2_MPLS;
-+ break;
-+ case (e_FM_PCD_MANIP_HDR_INSRT_PPPOE):
-+ hmcdOpt = HMCD_INSRT_L2_PPPOE;
-+ break;
-+ default:
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, NO_MSG);
-+ }
-+ tmpReg |= hmcdOpt << HMCD_L2_MODE_SHIFT;
-+
-+ WRITE_UINT32(*p_TmpHmct, tmpReg);
-+ /* save a pointer to the "last" indication word */
-+ p_Last = p_TmpHmct;
-+
-+ p_TmpHmct += HMCD_BASIC_SIZE / 4;
-+
-+ /* set size and pointer of user's data */
-+ size =
-+ (uint8_t)p_FmPcdManipParams->u.hdr.insrtParams.u.byHdr.u.specificL2Params.size;
-+
-+ ASSERT_COND(p_TmpData);
-+ MemCpy8(
-+ p_TmpData,
-+ p_FmPcdManipParams->u.hdr.insrtParams.u.byHdr.u.specificL2Params.p_Data,
-+ size);
-+ tmpReg =
-+ (size << HMCD_INSRT_L2_SIZE_SHIFT)
-+ | (uint32_t)(XX_VirtToPhys(p_TmpData)
-+ - (((t_FmPcd*)h_FmPcd)->physicalMuramBase));
-+ WRITE_UINT32(*p_TmpHmct, tmpReg);
-+ p_TmpHmct += HMCD_PTR_SIZE / 4;
-+ p_TmpData += size;
-+ }
-+ break;
-+#if (DPAA_VERSION >= 11)
-+ case (e_FM_PCD_MANIP_INSRT_BY_HDR_IP):
-+ tmpReg = (uint32_t)(HMCD_OPCODE_IP_INSRT)
-+ << HMCD_OC_SHIFT;
-+ if (p_FmPcdManipParams->u.hdr.insrtParams.u.byHdr.u.ipParams.calcL4Checksum)
-+ tmpReg |= HMCD_IP_L4_CS_CALC;
-+ if (p_FmPcdManipParams->u.hdr.insrtParams.u.byHdr.u.ipParams.mappingMode
-+ == e_FM_PCD_MANIP_HDR_QOS_MAPPING_AS_IS)
-+ tmpReg |= HMCD_IP_OR_QOS;
-+ tmpReg |=
-+ p_FmPcdManipParams->u.hdr.insrtParams.u.byHdr.u.ipParams.lastPidOffset
-+ & HMCD_IP_LAST_PID_MASK;
-+ tmpReg |=
-+ ((p_FmPcdManipParams->u.hdr.insrtParams.u.byHdr.u.ipParams.insrt.size
-+ << HMCD_IP_SIZE_SHIFT)
-+ & HMCD_IP_SIZE_MASK);
-+ if (p_FmPcdManipParams->u.hdr.insrtParams.u.byHdr.u.ipParams.dontFragOverwrite)
-+ tmpReg |= HMCD_IP_DF_MODE;
-+
-+ WRITE_UINT32(*p_TmpHmct, tmpReg);
-+
-+ /* save a pointer to the "last" indication word */
-+ p_Last = p_TmpHmct;
-+
-+ p_TmpHmct += HMCD_BASIC_SIZE / 4;
-+
-+ /* set IP id */
-+ ASSERT_COND(p_TmpData);
-+ WRITE_UINT16(
-+ *(uint16_t*)p_TmpData,
-+ p_FmPcdManipParams->u.hdr.insrtParams.u.byHdr.u.ipParams.id);
-+ WRITE_UINT32(
-+ *p_TmpHmct,
-+ (uint32_t)(XX_VirtToPhys(p_TmpData) - (((t_FmPcd*)p_Manip->h_FmPcd)->physicalMuramBase)));
-+ p_TmpData += 2;
-+ p_TmpHmct += HMCD_PTR_SIZE / 4;
-+
-+ WRITE_UINT8(*p_TmpHmct, p_FmPcdManipParams->u.hdr.insrtParams.u.byHdr.u.ipParams.lastDstOffset);
-+ p_TmpHmct += HMCD_PARAM_SIZE / 4;
-+
-+ MemCpy8(
-+ p_TmpHmct,
-+ p_FmPcdManipParams->u.hdr.insrtParams.u.byHdr.u.ipParams.insrt.p_Data,
-+ p_FmPcdManipParams->u.hdr.insrtParams.u.byHdr.u.ipParams.insrt.size);
-+ p_TmpHmct +=
-+ p_FmPcdManipParams->u.hdr.insrtParams.u.byHdr.u.ipParams.insrt.size
-+ / 4;
-+ break;
-+ case (e_FM_PCD_MANIP_INSRT_BY_HDR_UDP_LITE):
-+ tmpReg = HMCD_INSRT_UDP_LITE;
-+ case (e_FM_PCD_MANIP_INSRT_BY_HDR_UDP):
-+ tmpReg |= (uint32_t)(HMCD_OPCODE_UDP_INSRT)
-+ << HMCD_OC_SHIFT;
-+
-+ WRITE_UINT32(*p_TmpHmct, tmpReg);
-+
-+ /* save a pointer to the "last" indication word */
-+ p_Last = p_TmpHmct;
-+
-+ p_TmpHmct += HMCD_BASIC_SIZE / 4;
-+
-+ MemCpy8(
-+ p_TmpHmct,
-+ p_FmPcdManipParams->u.hdr.insrtParams.u.byHdr.u.insrt.p_Data,
-+ p_FmPcdManipParams->u.hdr.insrtParams.u.byHdr.u.insrt.size);
-+ p_TmpHmct +=
-+ p_FmPcdManipParams->u.hdr.insrtParams.u.byHdr.u.insrt.size
-+ / 4;
-+ break;
-+ case (e_FM_PCD_MANIP_INSRT_BY_HDR_CAPWAP):
-+ tmpReg = (uint32_t)(HMCD_OPCODE_CAPWAP_INSRT)
-+ << HMCD_OC_SHIFT;
-+ tmpReg |= HMCD_CAPWAP_INSRT;
-+
-+ WRITE_UINT32(*p_TmpHmct, tmpReg);
-+
-+ /* save a pointer to the "last" indication word */
-+ p_Last = p_TmpHmct;
-+
-+ p_TmpHmct += HMCD_BASIC_SIZE / 4;
-+
-+ MemCpy8(
-+ p_TmpHmct,
-+ p_FmPcdManipParams->u.hdr.insrtParams.u.byHdr.u.insrt.p_Data,
-+ p_FmPcdManipParams->u.hdr.insrtParams.u.byHdr.u.insrt.size);
-+ p_TmpHmct +=
-+ p_FmPcdManipParams->u.hdr.insrtParams.u.byHdr.u.insrt.size
-+ / 4;
-+ break;
-+#endif /* (DPAA_VERSION >= 11) */
-+ default:
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED,
-+ ("manip header insert by header type!"));
-+
-+ }
-+ }
-+ }
-+
-+ if (p_FmPcdManipParams->u.hdr.fieldUpdate)
-+ {
-+ switch (p_FmPcdManipParams->u.hdr.fieldUpdateParams.type)
-+ {
-+ case (e_FM_PCD_MANIP_HDR_FIELD_UPDATE_VLAN):
-+ /* set opcode */
-+ tmpReg = (uint32_t)(HMCD_OPCODE_VLAN_PRI_UPDATE)
-+ << HMCD_OC_SHIFT;
-+
-+ /* set mode & table pointer */
-+ if (p_FmPcdManipParams->u.hdr.fieldUpdateParams.u.vlan.updateType
-+ == e_FM_PCD_MANIP_HDR_FIELD_UPDATE_DSCP_TO_VLAN)
-+ {
-+ /* set Mode */
-+ tmpReg |= (uint32_t)(HMCD_VLAN_PRI_UPDATE_DSCP_TO_VPRI)
-+ << HMCD_VLAN_PRI_REP_MODE_SHIFT;
-+ /* set VPRI default */
-+ tmpReg |=
-+ p_FmPcdManipParams->u.hdr.fieldUpdateParams.u.vlan.u.dscpToVpri.vpriDefVal;
-+ WRITE_UINT32(*p_TmpHmct, tmpReg);
-+ /* save a pointer to the "last" indication word */
-+ p_Last = p_TmpHmct;
-+ /* write the table pointer into the Manip descriptor */
-+ p_TmpHmct += HMCD_BASIC_SIZE / 4;
-+
-+ tmpReg = 0;
-+ ASSERT_COND(p_TmpData);
-+ for (i = 0; i < HMCD_DSCP_VALUES; i++)
-+ {
-+ /* first we build from each 8 values a 32bit register */
-+ tmpReg |=
-+ (p_FmPcdManipParams->u.hdr.fieldUpdateParams.u.vlan.u.dscpToVpri.dscpToVpriTable[i])
-+ << (32 - 4 * (j + 1));
-+ j++;
-+ /* Than we write this register to the next table word
-+ * (i=7-->word 0, i=15-->word 1,... i=63-->word 7) */
-+ if ((i % 8) == 7)
-+ {
-+ WRITE_UINT32(*((uint32_t*)p_TmpData + (i+1)/8-1),
-+ tmpReg);
-+ tmpReg = 0;
-+ j = 0;
-+ }
-+ }
-+
-+ WRITE_UINT32(
-+ *p_TmpHmct,
-+ (uint32_t)(XX_VirtToPhys(p_TmpData) - (((t_FmPcd*)h_FmPcd)->physicalMuramBase)));
-+ p_TmpHmct += HMCD_PTR_SIZE / 4;
-+
-+ p_TmpData += DSCP_TO_VLAN_TABLE_SIZE;
-+ }
-+ else
-+ if (p_FmPcdManipParams->u.hdr.fieldUpdateParams.u.vlan.updateType
-+ == e_FM_PCD_MANIP_HDR_FIELD_UPDATE_VLAN_VPRI)
-+ {
-+ /* set Mode */
-+ /* line commented out as it has no-side-effect ('0' value). */
-+ /*tmpReg |= HMCD_VLAN_PRI_UPDATE << HMCD_VLAN_PRI_REP_MODE_SHIFT*/;
-+ /* set VPRI parameter */
-+ tmpReg |=
-+ p_FmPcdManipParams->u.hdr.fieldUpdateParams.u.vlan.u.vpri;
-+ WRITE_UINT32(*p_TmpHmct, tmpReg);
-+ /* save a pointer to the "last" indication word */
-+ p_Last = p_TmpHmct;
-+ p_TmpHmct += HMCD_BASIC_SIZE / 4;
-+ }
-+ break;
-+
-+ case (e_FM_PCD_MANIP_HDR_FIELD_UPDATE_IPV4):
-+ /* set opcode */
-+ tmpReg = (uint32_t)(HMCD_OPCODE_IPV4_UPDATE) << HMCD_OC_SHIFT;
-+ if (p_FmPcdManipParams->u.hdr.fieldUpdateParams.u.ipv4.validUpdates
-+ & HDR_MANIP_IPV4_TTL)
-+ tmpReg |= HMCD_IPV4_UPDATE_TTL;
-+ if (p_FmPcdManipParams->u.hdr.fieldUpdateParams.u.ipv4.validUpdates
-+ & HDR_MANIP_IPV4_TOS)
-+ {
-+ tmpReg |= HMCD_IPV4_UPDATE_TOS;
-+ tmpReg |=
-+ p_FmPcdManipParams->u.hdr.fieldUpdateParams.u.ipv4.tos
-+ << HMCD_IPV4_UPDATE_TOS_SHIFT;
-+ }
-+ if (p_FmPcdManipParams->u.hdr.fieldUpdateParams.u.ipv4.validUpdates
-+ & HDR_MANIP_IPV4_ID)
-+ tmpReg |= HMCD_IPV4_UPDATE_ID;
-+ if (p_FmPcdManipParams->u.hdr.fieldUpdateParams.u.ipv4.validUpdates
-+ & HDR_MANIP_IPV4_SRC)
-+ tmpReg |= HMCD_IPV4_UPDATE_SRC;
-+ if (p_FmPcdManipParams->u.hdr.fieldUpdateParams.u.ipv4.validUpdates
-+ & HDR_MANIP_IPV4_DST)
-+ tmpReg |= HMCD_IPV4_UPDATE_DST;
-+ /* write the first 4 bytes of the descriptor */
-+ WRITE_UINT32(*p_TmpHmct, tmpReg);
-+ /* save a pointer to the "last" indication word */
-+ p_Last = p_TmpHmct;
-+
-+ p_TmpHmct += HMCD_BASIC_SIZE / 4;
-+
-+ if (p_FmPcdManipParams->u.hdr.fieldUpdateParams.u.ipv4.validUpdates
-+ & HDR_MANIP_IPV4_ID)
-+ {
-+ ASSERT_COND(p_TmpData);
-+ WRITE_UINT16(
-+ *(uint16_t*)p_TmpData,
-+ p_FmPcdManipParams->u.hdr.fieldUpdateParams.u.ipv4.id);
-+ WRITE_UINT32(
-+ *p_TmpHmct,
-+ (uint32_t)(XX_VirtToPhys(p_TmpData) - (((t_FmPcd*)p_Manip->h_FmPcd)->physicalMuramBase)));
-+ p_TmpData += 2;
-+ p_TmpHmct += HMCD_PTR_SIZE / 4;
-+ }
-+
-+ if (p_FmPcdManipParams->u.hdr.fieldUpdateParams.u.ipv4.validUpdates
-+ & HDR_MANIP_IPV4_SRC)
-+ {
-+ WRITE_UINT32(
-+ *p_TmpHmct,
-+ p_FmPcdManipParams->u.hdr.fieldUpdateParams.u.ipv4.src);
-+ p_TmpHmct += HMCD_IPV4_ADDR_SIZE / 4;
-+ }
-+
-+ if (p_FmPcdManipParams->u.hdr.fieldUpdateParams.u.ipv4.validUpdates
-+ & HDR_MANIP_IPV4_DST)
-+ {
-+ WRITE_UINT32(
-+ *p_TmpHmct,
-+ p_FmPcdManipParams->u.hdr.fieldUpdateParams.u.ipv4.dst);
-+ p_TmpHmct += HMCD_IPV4_ADDR_SIZE / 4;
-+ }
-+ break;
-+
-+ case (e_FM_PCD_MANIP_HDR_FIELD_UPDATE_IPV6):
-+ /* set opcode */
-+ tmpReg = (uint32_t)(HMCD_OPCODE_IPV6_UPDATE) << HMCD_OC_SHIFT;
-+ if (p_FmPcdManipParams->u.hdr.fieldUpdateParams.u.ipv6.validUpdates
-+ & HDR_MANIP_IPV6_HL)
-+ tmpReg |= HMCD_IPV6_UPDATE_HL;
-+ if (p_FmPcdManipParams->u.hdr.fieldUpdateParams.u.ipv6.validUpdates
-+ & HDR_MANIP_IPV6_TC)
-+ {
-+ tmpReg |= HMCD_IPV6_UPDATE_TC;
-+ tmpReg |=
-+ p_FmPcdManipParams->u.hdr.fieldUpdateParams.u.ipv6.trafficClass
-+ << HMCD_IPV6_UPDATE_TC_SHIFT;
-+ }
-+ if (p_FmPcdManipParams->u.hdr.fieldUpdateParams.u.ipv6.validUpdates
-+ & HDR_MANIP_IPV6_SRC)
-+ tmpReg |= HMCD_IPV6_UPDATE_SRC;
-+ if (p_FmPcdManipParams->u.hdr.fieldUpdateParams.u.ipv6.validUpdates
-+ & HDR_MANIP_IPV6_DST)
-+ tmpReg |= HMCD_IPV6_UPDATE_DST;
-+ /* write the first 4 bytes of the descriptor */
-+ WRITE_UINT32(*p_TmpHmct, tmpReg);
-+ /* save a pointer to the "last" indication word */
-+ p_Last = p_TmpHmct;
-+
-+ p_TmpHmct += HMCD_BASIC_SIZE / 4;
-+ if (p_FmPcdManipParams->u.hdr.fieldUpdateParams.u.ipv6.validUpdates
-+ & HDR_MANIP_IPV6_SRC)
-+ {
-+ for (i = 0; i < NET_HEADER_FIELD_IPv6_ADDR_SIZE; i += 4)
-+ {
-+ memcpy(&tmp_ipv6_addr,
-+ &p_FmPcdManipParams->u.hdr.fieldUpdateParams.u.ipv6.src[i],
-+ sizeof(uint32_t));
-+ WRITE_UINT32(*p_TmpHmct, tmp_ipv6_addr);
-+ p_TmpHmct += HMCD_PTR_SIZE / 4;
-+ }
-+ }
-+ if (p_FmPcdManipParams->u.hdr.fieldUpdateParams.u.ipv6.validUpdates
-+ & HDR_MANIP_IPV6_DST)
-+ {
-+ for (i = 0; i < NET_HEADER_FIELD_IPv6_ADDR_SIZE; i += 4)
-+ {
-+ memcpy(&tmp_ipv6_addr,
-+ &p_FmPcdManipParams->u.hdr.fieldUpdateParams.u.ipv6.dst[i],
-+ sizeof(uint32_t));
-+ WRITE_UINT32(*p_TmpHmct, tmp_ipv6_addr);
-+ p_TmpHmct += HMCD_PTR_SIZE / 4;
-+ }
-+ }
-+ break;
-+
-+ case (e_FM_PCD_MANIP_HDR_FIELD_UPDATE_TCP_UDP):
-+ if (p_FmPcdManipParams->u.hdr.fieldUpdateParams.u.tcpUdp.validUpdates
-+ == HDR_MANIP_TCP_UDP_CHECKSUM)
-+ {
-+ /* we implement this case with the update-checksum descriptor */
-+ /* set opcode */
-+ tmpReg = (uint32_t)(HMCD_OPCODE_TCP_UDP_CHECKSUM)
-+ << HMCD_OC_SHIFT;
-+ /* write the first 4 bytes of the descriptor */
-+ WRITE_UINT32(*p_TmpHmct, tmpReg);
-+ /* save a pointer to the "last" indication word */
-+ p_Last = p_TmpHmct;
-+
-+ p_TmpHmct += HMCD_BASIC_SIZE / 4;
-+ }
-+ else
-+ {
-+ /* we implement this case with the TCP/UDP update descriptor */
-+ /* set opcode */
-+ tmpReg = (uint32_t)(HMCD_OPCODE_TCP_UDP_UPDATE)
-+ << HMCD_OC_SHIFT;
-+ if (p_FmPcdManipParams->u.hdr.fieldUpdateParams.u.tcpUdp.validUpdates
-+ & HDR_MANIP_TCP_UDP_DST)
-+ tmpReg |= HMCD_TCP_UDP_UPDATE_DST;
-+ if (p_FmPcdManipParams->u.hdr.fieldUpdateParams.u.tcpUdp.validUpdates
-+ & HDR_MANIP_TCP_UDP_SRC)
-+ tmpReg |= HMCD_TCP_UDP_UPDATE_SRC;
-+ /* write the first 4 bytes of the descriptor */
-+ WRITE_UINT32(*p_TmpHmct, tmpReg);
-+ /* save a pointer to the "last" indication word */
-+ p_Last = p_TmpHmct;
-+
-+ p_TmpHmct += HMCD_BASIC_SIZE / 4;
-+
-+ tmpReg = 0;
-+ if (p_FmPcdManipParams->u.hdr.fieldUpdateParams.u.tcpUdp.validUpdates
-+ & HDR_MANIP_TCP_UDP_SRC)
-+ tmpReg |=
-+ ((uint32_t)p_FmPcdManipParams->u.hdr.fieldUpdateParams.u.tcpUdp.src)
-+ << HMCD_TCP_UDP_UPDATE_SRC_SHIFT;
-+ if (p_FmPcdManipParams->u.hdr.fieldUpdateParams.u.tcpUdp.validUpdates
-+ & HDR_MANIP_TCP_UDP_DST)
-+ tmpReg |=
-+ ((uint32_t)p_FmPcdManipParams->u.hdr.fieldUpdateParams.u.tcpUdp.dst);
-+ WRITE_UINT32(*p_TmpHmct, tmpReg);
-+ p_TmpHmct += HMCD_PTR_SIZE / 4;
-+ }
-+ break;
-+
-+ default:
-+ RETURN_ERROR(MINOR, E_INVALID_SELECTION,
-+ ("Unknown fieldUpdateParams.type"));
-+ }
-+ }
-+
-+ if (p_FmPcdManipParams->u.hdr.custom)
-+ {
-+ switch (p_FmPcdManipParams->u.hdr.customParams.type)
-+ {
-+ case (e_FM_PCD_MANIP_HDR_CUSTOM_IP_REPLACE):
-+ /* set opcode */
-+ tmpReg = (uint32_t)(HMCD_OPCODE_REPLACE_IP) << HMCD_OC_SHIFT;
-+
-+ if (p_FmPcdManipParams->u.hdr.customParams.u.ipHdrReplace.decTtlHl)
-+ tmpReg |= HMCD_IP_REPLACE_TTL_HL;
-+ if (p_FmPcdManipParams->u.hdr.customParams.u.ipHdrReplace.replaceType
-+ == e_FM_PCD_MANIP_HDR_CUSTOM_REPLACE_IPV4_BY_IPV6)
-+ /* line commented out as it has no-side-effect ('0' value). */
-+ /*tmpReg |= HMCD_IP_REPLACE_REPLACE_IPV4*/;
-+ else
-+ if (p_FmPcdManipParams->u.hdr.customParams.u.ipHdrReplace.replaceType
-+ == e_FM_PCD_MANIP_HDR_CUSTOM_REPLACE_IPV6_BY_IPV4)
-+ {
-+ tmpReg |= HMCD_IP_REPLACE_REPLACE_IPV6;
-+ if (p_FmPcdManipParams->u.hdr.customParams.u.ipHdrReplace.updateIpv4Id)
-+ tmpReg |= HMCD_IP_REPLACE_ID;
-+ }
-+ else
-+ RETURN_ERROR(
-+ MINOR,
-+ E_NOT_SUPPORTED,
-+ ("One flag out of HDR_MANIP_IP_REPLACE_IPV4, HDR_MANIP_IP_REPLACE_IPV6 - must be set."));
-+
-+ /* write the first 4 bytes of the descriptor */
-+ WRITE_UINT32(*p_TmpHmct, tmpReg);
-+ /* save a pointer to the "last" indication word */
-+ p_Last = p_TmpHmct;
-+
-+ p_TmpHmct += HMCD_BASIC_SIZE / 4;
-+
-+ size =
-+ p_FmPcdManipParams->u.hdr.customParams.u.ipHdrReplace.hdrSize;
-+ ASSERT_COND(p_TmpData);
-+ MemCpy8(
-+ p_TmpData,
-+ p_FmPcdManipParams->u.hdr.customParams.u.ipHdrReplace.hdr,
-+ size);
-+ tmpReg = (uint32_t)(size << HMCD_IP_REPLACE_L3HDRSIZE_SHIFT);
-+ tmpReg |= (uint32_t)(XX_VirtToPhys(p_TmpData)
-+ - (((t_FmPcd*)h_FmPcd)->physicalMuramBase));
-+ WRITE_UINT32(*p_TmpHmct, tmpReg);
-+ p_TmpHmct += HMCD_PTR_SIZE / 4;
-+ p_TmpData += size;
-+
-+ if ((p_FmPcdManipParams->u.hdr.customParams.u.ipHdrReplace.replaceType
-+ == e_FM_PCD_MANIP_HDR_CUSTOM_REPLACE_IPV6_BY_IPV4)
-+ && (p_FmPcdManipParams->u.hdr.customParams.u.ipHdrReplace.updateIpv4Id))
-+ {
-+ WRITE_UINT16(
-+ *(uint16_t*)p_TmpData,
-+ p_FmPcdManipParams->u.hdr.customParams.u.ipHdrReplace.id);
-+ WRITE_UINT32(
-+ *p_TmpHmct,
-+ (uint32_t)(XX_VirtToPhys(p_TmpData) - (((t_FmPcd*)h_FmPcd)->physicalMuramBase)));
-+ p_TmpData += 2;
-+ }
-+ p_TmpHmct += HMCD_PTR_SIZE / 4;
-+ break;
-+ case (e_FM_PCD_MANIP_HDR_CUSTOM_GEN_FIELD_REPLACE):
-+ /* set opcode */
-+ tmpReg = (uint32_t)(HMCD_OPCODE_GEN_FIELD_REPLACE) << HMCD_OC_SHIFT;
-+ tmpReg |= p_FmPcdManipParams->u.hdr.customParams.u.genFieldReplace.size << HMCD_GEN_FIELD_SIZE_SHIFT;
-+ tmpReg |= p_FmPcdManipParams->u.hdr.customParams.u.genFieldReplace.srcOffset << HMCD_GEN_FIELD_SRC_OFF_SHIFT;
-+ tmpReg |= p_FmPcdManipParams->u.hdr.customParams.u.genFieldReplace.dstOffset << HMCD_GEN_FIELD_DST_OFF_SHIFT;
-+ if (p_FmPcdManipParams->u.hdr.customParams.u.genFieldReplace.mask)
-+ tmpReg |= HMCD_GEN_FIELD_MASK_EN;
-+
-+ /* write the first 4 bytes of the descriptor */
-+ WRITE_UINT32(*p_TmpHmct, tmpReg);
-+ /* save a pointer to the "last" indication word */
-+ p_Last = p_TmpHmct;
-+
-+ p_TmpHmct += HMCD_BASIC_SIZE/4;
-+
-+ if (p_FmPcdManipParams->u.hdr.customParams.u.genFieldReplace.mask)
-+ {
-+ tmpReg = p_FmPcdManipParams->u.hdr.customParams.u.genFieldReplace.mask << HMCD_GEN_FIELD_MASK_SHIFT;
-+ tmpReg |= p_FmPcdManipParams->u.hdr.customParams.u.genFieldReplace.maskOffset << HMCD_GEN_FIELD_MASK_OFF_SHIFT;
-+ /* write the next 4 bytes of the descriptor */
-+ WRITE_UINT32(*p_TmpHmct, tmpReg);
-+ }
-+ p_TmpHmct += HMCD_PARAM_SIZE/4;
-+ break;
-+ default:
-+ RETURN_ERROR(MINOR, E_INVALID_SELECTION,
-+ ("Unknown customParams.type"));
-+ }
-+ }
-+
-+ /* If this node has a nextManip, and no parsing is required, the old table must be copied to the new table
-+ the old table and should be freed */
-+ if (p_FmPcdManipParams->h_NextManip
-+ && (p_Manip->nextManipType == e_FM_PCD_MANIP_HDR)
-+ && (MANIP_DONT_REPARSE(p_Manip)))
-+ {
-+ if (new)
-+ {
-+ /* If this is the first time this manip is created we need to free unused memory. If it
-+ * is a dynamic changes case, the memory used is either the CC shadow or the existing
-+ * table - no allocation, no free */
-+ MANIP_UPDATE_UNIFIED_POSITION(p_FmPcdManipParams->h_NextManip);
-+
-+ p_Manip->unifiedPosition = e_MANIP_UNIFIED_FIRST;
-+ }
-+ }
-+ else
-+ {
-+ ASSERT_COND(p_Last);
-+ /* set the "last" indication on the last command of the current table */
-+ WRITE_UINT32(*p_Last, GET_UINT32(*p_Last) | HMCD_LAST);
-+ }
-+
-+ return E_OK;
-+}
-+
-+static t_Error CreateManipActionNew(t_FmPcdManip *p_Manip,
-+ t_FmPcdManipParams *p_FmPcdManipParams)
-+{
-+ t_FmPcdManip *p_CurManip;
-+ t_Error err;
-+ uint32_t nextSize = 0, totalSize;
-+ uint16_t tmpReg;
-+ uint8_t *p_OldHmct, *p_TmpHmctPtr, *p_TmpDataPtr;
-+
-+ /* set Manip structure */
-+
-+ p_Manip->dontParseAfterManip =
-+ p_FmPcdManipParams->u.hdr.dontParseAfterManip;
-+
-+ if (p_FmPcdManipParams->h_NextManip)
-+ { /* Next Header manipulation exists */
-+ p_Manip->nextManipType = MANIP_GET_TYPE(p_FmPcdManipParams->h_NextManip);
-+
-+ if ((p_Manip->nextManipType == e_FM_PCD_MANIP_HDR) && p_Manip->dontParseAfterManip)
-+ nextSize = (uint32_t)(GetHmctSize(p_FmPcdManipParams->h_NextManip)
-+ + GetDataSize(p_FmPcdManipParams->h_NextManip));
-+ else /* either parsing is required or next manip is Frag; no table merging. */
-+ p_Manip->cascaded = TRUE;
-+ /* pass up the "cascaded" attribute. The whole chain is cascaded
-+ * if something is cascaded along the way. */
-+ if (MANIP_IS_CASCADED(p_FmPcdManipParams->h_NextManip))
-+ p_Manip->cascaded = TRUE;
-+ }
-+
-+ /* Allocate new table */
-+ /* calculate table size according to manip parameters */
-+ err = CalculateTableSize(p_FmPcdManipParams, &p_Manip->tableSize,
-+ &p_Manip->dataSize);
-+ if (err)
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+
-+ totalSize = (uint16_t)(p_Manip->tableSize + p_Manip->dataSize + nextSize);
-+
-+ p_Manip->p_Hmct = (uint8_t*)FM_MURAM_AllocMem(
-+ ((t_FmPcd *)p_Manip->h_FmPcd)->h_FmMuram, totalSize, 4);
-+ if (!p_Manip->p_Hmct)
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, ("MURAM alloc failed"));
-+
-+ if (p_Manip->dataSize)
-+ p_Manip->p_Data =
-+ (uint8_t*)PTR_MOVE(p_Manip->p_Hmct, (p_Manip->tableSize + nextSize));
-+
-+ /* update shadow size to allow runtime replacement of Header manipulation */
-+ /* The allocated shadow is divided as follows:
-+ 0 . . . 16 . . .
-+ --------------------------------
-+ | Shadow | Shadow HMTD |
-+ | HMTD | Match Table |
-+ | (16 bytes) | (maximal size) |
-+ --------------------------------
-+ */
-+
-+ err = FmPcdUpdateCcShadow(p_Manip->h_FmPcd, (uint32_t)(totalSize + 16),
-+ (uint16_t)FM_PCD_CC_AD_TABLE_ALIGN);
-+ if (err != E_OK)
-+ {
-+ FM_MURAM_FreeMem(p_Manip->h_FmPcd, p_Manip->p_Hmct);
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY,
-+ ("MURAM allocation for HdrManip node shadow"));
-+ }
-+
-+ if (p_FmPcdManipParams->h_NextManip
-+ && (p_Manip->nextManipType == e_FM_PCD_MANIP_HDR)
-+ && (MANIP_DONT_REPARSE(p_Manip)))
-+ {
-+ p_OldHmct = (uint8_t *)GetManipInfo(p_FmPcdManipParams->h_NextManip,
-+ e_MANIP_HMCT);
-+ p_CurManip = p_FmPcdManipParams->h_NextManip;
-+ /* Run till the last Manip (which is the first to configure) */
-+ while (MANIP_IS_UNIFIED_NON_LAST(p_CurManip))
-+ p_CurManip = p_CurManip->h_NextManip;
-+
-+ while (p_CurManip)
-+ {
-+ /* If this is a unified table, point to the part of the table
-+ * which is the relative offset in HMCT.
-+ */
-+ p_TmpHmctPtr = (uint8_t*)PTR_MOVE(p_Manip->p_Hmct,
-+ (p_Manip->tableSize +
-+ (PTR_TO_UINT(p_CurManip->p_Hmct) -
-+ PTR_TO_UINT(p_OldHmct))));
-+ if (p_CurManip->p_Data)
-+ p_TmpDataPtr = (uint8_t*)PTR_MOVE(p_Manip->p_Hmct,
-+ (p_Manip->tableSize +
-+ (PTR_TO_UINT(p_CurManip->p_Data) -
-+ PTR_TO_UINT(p_OldHmct))));
-+ else
-+ p_TmpDataPtr = NULL;
-+
-+ BuildHmct(p_CurManip, &p_CurManip->manipParams, p_TmpHmctPtr,
-+ p_TmpDataPtr, FALSE);
-+ /* update old manip table pointer */
-+ MANIP_SET_HMCT_PTR(p_CurManip, p_TmpHmctPtr);
-+ MANIP_SET_DATA_PTR(p_CurManip, p_TmpDataPtr);
-+
-+ p_CurManip = p_CurManip->h_PrevManip;
-+ }
-+ /* We copied the HMCT to create a new large HMCT so we can free the old one */
-+ FM_MURAM_FreeMem(MANIP_GET_MURAM(p_FmPcdManipParams->h_NextManip),
-+ p_OldHmct);
-+ }
-+
-+ /* Fill table */
-+ err = BuildHmct(p_Manip, p_FmPcdManipParams, p_Manip->p_Hmct,
-+ p_Manip->p_Data, TRUE);
-+ if (err)
-+ {
-+ FM_MURAM_FreeMem(p_Manip->h_FmPcd, p_Manip->p_Hmct);
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+ }
-+
-+ /* Build HMTD (table descriptor) */
-+ tmpReg = HMTD_CFG_TYPE; /* NADEN = 0 */
-+
-+ /* add parseAfterManip */
-+ if (!p_Manip->dontParseAfterManip)
-+ tmpReg |= HMTD_CFG_PRS_AFTER_HM;
-+
-+ /* create cascade */
-+ /*if (p_FmPcdManipParams->h_NextManip
-+ && (!MANIP_DONT_REPARSE(p_Manip) || (p_Manip->nextManipType != e_FM_PCD_MANIP_HDR)))*/
-+ if (p_Manip->cascaded)
-+ {
-+ uint16_t nextAd;
-+ /* indicate that there's another HM table descriptor */
-+ tmpReg |= HMTD_CFG_NEXT_AD_EN;
-+ /* get address of next HMTD (table descriptor; h_Ad).
-+ * If the next HMTD was removed due to table unifing, get the address
-+ * of the "next next" as written in the h_Ad of the next h_Manip node.
-+ */
-+ if (p_Manip->unifiedPosition != e_MANIP_UNIFIED_FIRST)
-+ nextAd = (uint16_t)((uint32_t)(XX_VirtToPhys(MANIP_GET_HMTD_PTR(p_FmPcdManipParams->h_NextManip)) - (((t_FmPcd*)p_Manip->h_FmPcd)->physicalMuramBase)) >> 4);
-+ else
-+ nextAd = ((t_Hmtd *)((t_FmPcdManip *)p_FmPcdManipParams->h_NextManip)->h_Ad)->nextAdIdx;
-+
-+ WRITE_UINT16(((t_Hmtd *)p_Manip->h_Ad)->nextAdIdx, nextAd);
-+ }
-+
-+ WRITE_UINT16(((t_Hmtd *)p_Manip->h_Ad)->cfg, tmpReg);
-+ WRITE_UINT32(
-+ ((t_Hmtd *)p_Manip->h_Ad)->hmcdBasePtr,
-+ (uint32_t)(XX_VirtToPhys(p_Manip->p_Hmct) - (((t_FmPcd*)p_Manip->h_FmPcd)->physicalMuramBase)));
-+
-+ WRITE_UINT8(((t_Hmtd *)p_Manip->h_Ad)->opCode, HMAN_OC);
-+
-+ if (p_Manip->unifiedPosition == e_MANIP_UNIFIED_FIRST)
-+ {
-+ /* The HMTD of the next Manip is never going to be used */
-+ if (((t_FmPcdManip *)p_FmPcdManipParams->h_NextManip)->muramAllocate)
-+ FM_MURAM_FreeMem(
-+ ((t_FmPcd *)((t_FmPcdManip *)p_FmPcdManipParams->h_NextManip)->h_FmPcd)->h_FmMuram,
-+ ((t_FmPcdManip *)p_FmPcdManipParams->h_NextManip)->h_Ad);
-+ else
-+ XX_Free(((t_FmPcdManip *)p_FmPcdManipParams->h_NextManip)->h_Ad);
-+ ((t_FmPcdManip *)p_FmPcdManipParams->h_NextManip)->h_Ad = NULL;
-+ }
-+
-+ return E_OK;
-+}
-+
-+static t_Error CreateManipActionShadow(t_FmPcdManip *p_Manip,
-+ t_FmPcdManipParams *p_FmPcdManipParams)
-+{
-+ uint8_t *p_WholeHmct, *p_TmpHmctPtr, newDataSize, *p_TmpDataPtr = NULL;
-+ uint16_t newSize;
-+ t_FmPcd *p_FmPcd = (t_FmPcd *)p_Manip->h_FmPcd;
-+ t_Error err;
-+ t_FmPcdManip *p_CurManip = p_Manip;
-+
-+ err = CalculateTableSize(p_FmPcdManipParams, &newSize, &newDataSize);
-+ if (err)
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+
-+ /* check coherency of new table parameters */
-+ if (newSize > p_Manip->tableSize)
-+ RETURN_ERROR(
-+ MINOR,
-+ E_INVALID_VALUE,
-+ ("New Hdr Manip configuration requires larger size than current one (command table)."));
-+ if (newDataSize > p_Manip->dataSize)
-+ RETURN_ERROR(
-+ MINOR,
-+ E_INVALID_VALUE,
-+ ("New Hdr Manip configuration requires larger size than current one (data)."));
-+ if (p_FmPcdManipParams->h_NextManip)
-+ RETURN_ERROR(
-+ MINOR, E_INVALID_VALUE,
-+ ("New Hdr Manip configuration can not contain h_NextManip."));
-+ if (MANIP_IS_UNIFIED(p_Manip) && (newSize != p_Manip->tableSize))
-+ RETURN_ERROR(
-+ MINOR,
-+ E_INVALID_VALUE,
-+ ("New Hdr Manip configuration in a chained manipulation requires different size than current one."));
-+ if (p_Manip->dontParseAfterManip
-+ != p_FmPcdManipParams->u.hdr.dontParseAfterManip)
-+ RETURN_ERROR(
-+ MINOR,
-+ E_INVALID_VALUE,
-+ ("New Hdr Manip configuration differs in dontParseAfterManip value."));
-+
-+ p_Manip->tableSize = newSize;
-+ p_Manip->dataSize = newDataSize;
-+
-+ /* Build the new table in the shadow */
-+ if (!MANIP_IS_UNIFIED(p_Manip))
-+ {
-+ p_TmpHmctPtr = (uint8_t*)PTR_MOVE(p_FmPcd->p_CcShadow, 16);
-+ if (p_Manip->p_Data)
-+ p_TmpDataPtr =
-+ (uint8_t*)PTR_MOVE(p_TmpHmctPtr,
-+ (PTR_TO_UINT(p_Manip->p_Data) - PTR_TO_UINT(p_Manip->p_Hmct)));
-+
-+ BuildHmct(p_Manip, p_FmPcdManipParams, p_TmpHmctPtr, p_Manip->p_Data,
-+ FALSE);
-+ }
-+ else
-+ {
-+ p_WholeHmct = (uint8_t *)GetManipInfo(p_Manip, e_MANIP_HMCT);
-+ ASSERT_COND(p_WholeHmct);
-+
-+ /* Run till the last Manip (which is the first to configure) */
-+ while (MANIP_IS_UNIFIED_NON_LAST(p_CurManip))
-+ p_CurManip = p_CurManip->h_NextManip;
-+
-+ while (p_CurManip)
-+ {
-+ /* If this is a non-head node in a unified table, point to the part of the shadow
-+ * which is the relative offset in HMCT.
-+ * else, point to the beginning of the
-+ * shadow table (we save 16 for the HMTD.
-+ */
-+ p_TmpHmctPtr =
-+ (uint8_t*)PTR_MOVE(p_FmPcd->p_CcShadow,
-+ (16 + PTR_TO_UINT(p_CurManip->p_Hmct) - PTR_TO_UINT(p_WholeHmct)));
-+ if (p_CurManip->p_Data)
-+ p_TmpDataPtr =
-+ (uint8_t*)PTR_MOVE(p_FmPcd->p_CcShadow,
-+ (16 + PTR_TO_UINT(p_CurManip->p_Data) - PTR_TO_UINT(p_WholeHmct)));
-+
-+ BuildHmct(p_CurManip, &p_CurManip->manipParams, p_TmpHmctPtr,
-+ p_TmpDataPtr, FALSE);
-+ p_CurManip = p_CurManip->h_PrevManip;
-+ }
-+ }
-+
-+ return E_OK;
-+}
-+
-+static t_Error CreateManipActionBackToOrig(
-+ t_FmPcdManip *p_Manip, t_FmPcdManipParams *p_FmPcdManipParams)
-+{
-+ uint8_t *p_WholeHmct = NULL, *p_TmpHmctPtr, *p_TmpDataPtr;
-+ t_FmPcdManip *p_CurManip = p_Manip;
-+
-+ /* Build the new table in the shadow */
-+ if (!MANIP_IS_UNIFIED(p_Manip))
-+ BuildHmct(p_Manip, p_FmPcdManipParams, p_Manip->p_Hmct, p_Manip->p_Data,
-+ FALSE);
-+ else
-+ {
-+ p_WholeHmct = (uint8_t *)GetManipInfo(p_Manip, e_MANIP_HMCT);
-+ ASSERT_COND(p_WholeHmct);
-+
-+ /* Run till the last Manip (which is the first to configure) */
-+ while (MANIP_IS_UNIFIED_NON_LAST(p_CurManip))
-+ p_CurManip = p_CurManip->h_NextManip;
-+
-+ while (p_CurManip)
-+ {
-+ /* If this is a unified table, point to the part of the table
-+ * which is the relative offset in HMCT.
-+ */
-+ p_TmpHmctPtr = p_CurManip->p_Hmct; /*- (uint32_t)p_WholeHmct*/
-+ p_TmpDataPtr = p_CurManip->p_Data; /*- (uint32_t)p_WholeHmct*/
-+
-+ BuildHmct(p_CurManip, &p_CurManip->manipParams, p_TmpHmctPtr,
-+ p_TmpDataPtr, FALSE);
-+
-+ p_CurManip = p_CurManip->h_PrevManip;
-+ }
-+ }
-+
-+ return E_OK;
-+}
-+
-+#if (defined(FM_CAPWAP_SUPPORT) && (DPAA_VERSION == 10))
-+static t_Error UpdateManipIc(t_Handle h_Manip, uint8_t icOffset)
-+{
-+ t_FmPcdManip *p_Manip = (t_FmPcdManip *)h_Manip;
-+ t_Handle p_Ad;
-+ uint32_t tmpReg32 = 0;
-+ SANITY_CHECK_RETURN_ERROR(h_Manip, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Manip->h_Ad, E_INVALID_HANDLE);
-+
-+ switch (p_Manip->opcode)
-+ {
-+ case (HMAN_OC_MV_INT_FRAME_HDR_FROM_FRM_TO_BUFFER_PREFFIX):
-+ p_Ad = (t_AdOfTypeContLookup *)p_Manip->h_Ad;
-+ if (p_Manip->updateParams & INTERNAL_CONTEXT_OFFSET)
-+ {
-+ tmpReg32 =
-+ *(uint32_t *)&((t_AdOfTypeContLookup *)p_Ad)->pcAndOffsets;
-+ tmpReg32 |= (uint32_t)((uint32_t)icOffset << 16);
-+ *(uint32_t *)&((t_AdOfTypeContLookup *)p_Ad)->pcAndOffsets =
-+ tmpReg32;
-+ p_Manip->updateParams &= ~INTERNAL_CONTEXT_OFFSET;
-+ p_Manip->icOffset = icOffset;
-+ }
-+ else
-+ {
-+ if (p_Manip->icOffset != icOffset)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("this manipulation was updated previously by different value"););
-+ }
-+ break;
-+ case (HMAN_OC_CAPWAP_RMV_DTLS_IF_EXIST):
-+ if (p_Manip->h_Frag)
-+ {
-+ if (p_Manip->updateParams & INTERNAL_CONTEXT_OFFSET)
-+ {
-+ p_Ad = (t_AdOfTypeContLookup *)p_Manip->h_Ad;
-+ tmpReg32 |= GET_UINT32(((t_AdOfTypeContLookup *)p_Ad)->pcAndOffsets);
-+ tmpReg32 |= (uint32_t)((uint32_t)icOffset << 16);
-+ WRITE_UINT32(((t_AdOfTypeContLookup *)p_Ad)->pcAndOffsets, tmpReg32);
-+ p_Manip->updateParams &= ~INTERNAL_CONTEXT_OFFSET;
-+ p_Manip->icOffset = icOffset;
-+ }
-+ else
-+ {
-+ if (p_Manip->icOffset != icOffset)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("this manipulation was updated previousely by different value"););
-+ }
-+ }
-+ break;
-+ }
-+
-+ return E_OK;
-+}
-+
-+static t_Error UpdateInitMvIntFrameHeaderFromFrameToBufferPrefix(
-+ t_Handle h_FmPort, t_FmPcdManip *p_Manip, t_Handle h_Ad, bool validate)
-+{
-+
-+ t_AdOfTypeContLookup *p_Ad = (t_AdOfTypeContLookup *)h_Ad;
-+ t_FmPortGetSetCcParams fmPortGetSetCcParams;
-+ t_Error err;
-+ uint32_t tmpReg32;
-+
-+ memset(&fmPortGetSetCcParams, 0, sizeof(t_FmPortGetSetCcParams));
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Manip, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(
-+ (p_Manip->opcode & HMAN_OC_MV_INT_FRAME_HDR_FROM_FRM_TO_BUFFER_PREFFIX),
-+ E_INVALID_STATE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Manip->muramAllocate, E_INVALID_STATE);
-+
-+ if (p_Manip->updateParams)
-+ {
-+ if ((!(p_Manip->updateParams & OFFSET_OF_PR))
-+ || (p_Manip->shadowUpdateParams & OFFSET_OF_PR))
-+ RETURN_ERROR(
-+ MAJOR, E_INVALID_STATE,
-+ ("in this stage parameters from Port has not be updated"));
-+
-+ fmPortGetSetCcParams.getCcParams.type = p_Manip->updateParams;
-+ fmPortGetSetCcParams.setCcParams.type = UPDATE_PSO;
-+ fmPortGetSetCcParams.setCcParams.psoSize = 16;
-+
-+ err = FmPortGetSetCcParams(h_FmPort, &fmPortGetSetCcParams);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ if (fmPortGetSetCcParams.getCcParams.type & OFFSET_OF_PR)
-+ RETURN_ERROR(
-+ MAJOR, E_INVALID_STATE,
-+ ("Parser result offset wasn't configured previousely"));
-+#ifdef FM_LOCKUP_ALIGNMENT_ERRATA_FMAN_SW004
-+ ASSERT_COND(!(fmPortGetSetCcParams.getCcParams.prOffset % 16));
-+#endif
-+ }
-+ else
-+ if (validate)
-+ {
-+ if ((!(p_Manip->shadowUpdateParams & OFFSET_OF_PR))
-+ || (p_Manip->updateParams & OFFSET_OF_PR))
-+ RETURN_ERROR(
-+ MAJOR, E_INVALID_STATE,
-+ ("in this stage parameters from Port has be updated"));
-+ fmPortGetSetCcParams.getCcParams.type = p_Manip->shadowUpdateParams;
-+ fmPortGetSetCcParams.setCcParams.type = UPDATE_PSO;
-+ fmPortGetSetCcParams.setCcParams.psoSize = 16;
-+
-+ err = FmPortGetSetCcParams(h_FmPort, &fmPortGetSetCcParams);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ if (fmPortGetSetCcParams.getCcParams.type & OFFSET_OF_PR)
-+ RETURN_ERROR(
-+ MAJOR, E_INVALID_STATE,
-+ ("Parser result offset wasn't configured previousely"));
-+
-+ }
-+
-+ ASSERT_COND(p_Ad);
-+
-+ if (p_Manip->updateParams & OFFSET_OF_PR)
-+ {
-+ tmpReg32 = 0;
-+ tmpReg32 |= fmPortGetSetCcParams.getCcParams.prOffset;
-+ WRITE_UINT32(p_Ad->matchTblPtr,
-+ (GET_UINT32(p_Ad->matchTblPtr) | tmpReg32));
-+ p_Manip->updateParams &= ~OFFSET_OF_PR;
-+ p_Manip->shadowUpdateParams |= OFFSET_OF_PR;
-+ }
-+ else
-+ if (validate)
-+ {
-+ tmpReg32 = GET_UINT32(p_Ad->matchTblPtr);
-+ if ((uint8_t)tmpReg32 != fmPortGetSetCcParams.getCcParams.prOffset)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_STATE,
-+ ("this manipulation was updated previousely by different value"););
-+ }
-+
-+ return E_OK;
-+}
-+
-+static t_Error UpdateModifyCapwapFragmenation(t_FmPcdManip *p_Manip, t_Handle h_Ad, bool validate,t_Handle h_FmTree)
-+{
-+ t_AdOfTypeContLookup *p_Ad = (t_AdOfTypeContLookup *)h_Ad;
-+ t_FmPcdCcSavedManipParams *p_SavedManipParams = NULL;
-+ uint32_t tmpReg32 = 0;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Manip,E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Manip->h_Frag,E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Manip->frag,E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(((p_Manip->opcode == HMAN_OC_CAPWAP_FRAGMENTATION) || (p_Manip->opcode == HMAN_OC_INSRT_HDR_BY_TEMPL_N_OR_FRAG_AFTER)), E_INVALID_STATE);
-+
-+ p_Ad = (t_AdOfTypeContLookup *)p_Manip->h_Frag;
-+
-+ if (p_Manip->updateParams)
-+ {
-+
-+ if ((!(p_Manip->updateParams & OFFSET_OF_DATA)) ||
-+ ((p_Manip->shadowUpdateParams & OFFSET_OF_DATA)))
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("in this stage parameters from Port has not be updated"));
-+ p_SavedManipParams = FmPcdCcTreeGetSavedManipParams(h_FmTree);
-+ if (!p_SavedManipParams)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("for this manipulation tree has to be configured previosely with this type"));
-+ p_Manip->capwapFragParams.dataOffset = p_SavedManipParams->capwapParams.dataOffset;
-+
-+ tmpReg32 = GET_UINT32(p_Ad->pcAndOffsets);
-+ tmpReg32 |= ((uint32_t)p_Manip->capwapFragParams.dataOffset<< 16);
-+ WRITE_UINT32(p_Ad->pcAndOffsets,tmpReg32);
-+
-+ p_Manip->updateParams &= ~OFFSET_OF_DATA;
-+ p_Manip->shadowUpdateParams |= OFFSET_OF_DATA;
-+ }
-+ else if (validate)
-+ {
-+
-+ p_SavedManipParams = FmPcdCcTreeGetSavedManipParams(h_FmTree);
-+ if (!p_SavedManipParams)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("for this manipulation tree has to be configured previosely with this type"));
-+ if (p_Manip->capwapFragParams.dataOffset != p_SavedManipParams->capwapParams.dataOffset)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("this manipulation was updated previousely by different value"));
-+ }
-+
-+ return E_OK;
-+}
-+
-+static t_Error UpdateInitCapwapFragmentation(t_Handle h_FmPort,
-+ t_FmPcdManip *p_Manip,
-+ t_Handle h_Ad,
-+ bool validate,
-+ t_Handle h_FmTree)
-+{
-+ t_AdOfTypeContLookup *p_Ad;
-+ t_FmPortGetSetCcParams fmPortGetSetCcParams;
-+ t_Error err;
-+ uint32_t tmpReg32 = 0;
-+ t_FmPcdCcSavedManipParams *p_SavedManipParams;
-+
-+ UNUSED(h_Ad);
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Manip,E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Manip->h_Frag,E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Manip->frag,E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(((p_Manip->opcode == HMAN_OC_CAPWAP_FRAGMENTATION) ||
-+ (p_Manip->opcode == HMAN_OC_INSRT_HDR_BY_TEMPL_N_OR_FRAG_AFTER)), E_INVALID_STATE);
-+
-+ p_Ad = (t_AdOfTypeContLookup *)p_Manip->h_Frag;
-+
-+ if (p_Manip->updateParams)
-+ {
-+ if ((!(p_Manip->updateParams & OFFSET_OF_DATA)) ||
-+ ((p_Manip->shadowUpdateParams & OFFSET_OF_DATA)))
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("in this stage parameters from Port has not be updated"));
-+ fmPortGetSetCcParams.getCcParams.type = p_Manip->updateParams;
-+ fmPortGetSetCcParams.setCcParams.type = UPDATE_NIA_PNEN | UPDATE_FMFP_PRC_WITH_ONE_RISC_ONLY;
-+ fmPortGetSetCcParams.setCcParams.nia = NIA_FM_CTL_AC_POP_TO_N_STEP | NIA_ENG_FM_CTL;
-+ /* For CAPWAP Rassembly used FMAN_CTRL2 hardcoded - so for fragmentation its better to use FMAN_CTRL1 */
-+ fmPortGetSetCcParams.setCcParams.orFmanCtrl = FPM_PORT_FM_CTL1;
-+
-+ err = FmPortGetSetCcParams(h_FmPort, &fmPortGetSetCcParams);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+
-+ if (fmPortGetSetCcParams.getCcParams.type & OFFSET_OF_DATA)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("Data offset wasn't configured previousely"));
-+
-+ p_SavedManipParams = (t_FmPcdCcSavedManipParams *)XX_Malloc(sizeof(t_FmPcdCcSavedManipParams));
-+ p_SavedManipParams->capwapParams.dataOffset = fmPortGetSetCcParams.getCcParams.dataOffset;
-+
-+#ifdef FM_LOCKUP_ALIGNMENT_ERRATA_FMAN_SW004
-+ ASSERT_COND(!(p_SavedManipParams->capwapParams.dataOffset % 16));
-+#endif /* FM_LOCKUP_ALIGNMENT_ERRATA_FMAN_SW004 */
-+
-+ FmPcdCcTreeSetSavedManipParams(h_FmTree, (t_Handle)p_SavedManipParams);
-+ }
-+ else if (validate)
-+ {
-+ if ((!(p_Manip->shadowUpdateParams & OFFSET_OF_DATA)) ||
-+ ((p_Manip->updateParams & OFFSET_OF_DATA)))
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("in this stage parameters from Port has be updated"));
-+ fmPortGetSetCcParams.getCcParams.type = p_Manip->shadowUpdateParams;
-+ fmPortGetSetCcParams.setCcParams.type = UPDATE_NIA_PNEN | UPDATE_FMFP_PRC_WITH_ONE_RISC_ONLY;
-+ fmPortGetSetCcParams.setCcParams.nia = NIA_FM_CTL_AC_POP_TO_N_STEP | NIA_ENG_FM_CTL;
-+ err = FmPortGetSetCcParams(h_FmPort, &fmPortGetSetCcParams);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+
-+ if (fmPortGetSetCcParams.getCcParams.type & OFFSET_OF_DATA)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("Data offset wasn't configured previousely"));
-+ }
-+
-+ if (p_Manip->updateParams)
-+ {
-+ tmpReg32 = GET_UINT32(p_Ad->pcAndOffsets);
-+ tmpReg32 |= ((uint32_t)fmPortGetSetCcParams.getCcParams.dataOffset<< 16);
-+ WRITE_UINT32(p_Ad->pcAndOffsets,tmpReg32);
-+
-+ p_Manip->updateParams &= ~OFFSET_OF_DATA;
-+ p_Manip->shadowUpdateParams |= OFFSET_OF_DATA;
-+ p_Manip->capwapFragParams.dataOffset = fmPortGetSetCcParams.getCcParams.dataOffset;
-+ }
-+ else if (validate)
-+ {
-+ if (p_Manip->capwapFragParams.dataOffset != fmPortGetSetCcParams.getCcParams.dataOffset)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("this manipulation was updated previousely by different value"));
-+ }
-+
-+ return E_OK;
-+}
-+
-+static t_Error UpdateInitCapwapReasm(t_Handle h_FmPcd,
-+ t_Handle h_FmPort,
-+ t_FmPcdManip *p_Manip,
-+ t_Handle h_Ad,
-+ bool validate)
-+{
-+ t_CapwapReasmPram *p_ReassmTbl;
-+ t_Error err;
-+ t_FmPortGetSetCcParams fmPortGetSetCcParams;
-+ uint8_t i = 0;
-+ uint16_t size;
-+ uint32_t tmpReg32;
-+ t_FmPcd *p_FmPcd = (t_FmPcd *)h_FmPcd;
-+ t_FmPcdCcCapwapReassmTimeoutParams ccCapwapReassmTimeoutParams;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Manip,E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Manip->h_Frag,E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Manip->frag,E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR((p_Manip->opcode == HMAN_OC_CAPWAP_RMV_DTLS_IF_EXIST), E_INVALID_STATE);
-+ SANITY_CHECK_RETURN_ERROR(h_FmPcd,E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd->h_Hc,E_INVALID_HANDLE);
-+
-+ if (p_Manip->h_FmPcd != h_FmPcd)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE,
-+ ("handler of PCD previously was initiated by different value"));
-+
-+ UNUSED(h_Ad);
-+
-+ memset(&fmPortGetSetCcParams, 0, sizeof(t_FmPortGetSetCcParams));
-+ p_ReassmTbl = (t_CapwapReasmPram *)p_Manip->h_Frag;
-+
-+ if (p_Manip->updateParams)
-+ {
-+ if ((!(p_Manip->updateParams & NUM_OF_TASKS) &&
-+ !(p_Manip->updateParams & OFFSET_OF_DATA) &&
-+ !(p_Manip->updateParams & OFFSET_OF_PR) &&
-+ !(p_Manip->updateParams & HW_PORT_ID)) ||
-+ ((p_Manip->shadowUpdateParams & NUM_OF_TASKS) ||
-+ (p_Manip->shadowUpdateParams & OFFSET_OF_DATA) || (p_Manip->shadowUpdateParams & OFFSET_OF_PR) ||
-+ (p_Manip->shadowUpdateParams & HW_PORT_ID)))
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("in this stage parameters from Port has not be updated"));
-+
-+ fmPortGetSetCcParams.getCcParams.type = p_Manip->updateParams;
-+ fmPortGetSetCcParams.setCcParams.type = UPDATE_NIA_PNEN;
-+ fmPortGetSetCcParams.setCcParams.nia = NIA_FM_CTL_AC_POP_TO_N_STEP | NIA_ENG_FM_CTL;
-+
-+ err = FmPortGetSetCcParams(h_FmPort, &fmPortGetSetCcParams);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+
-+ if (fmPortGetSetCcParams.getCcParams.type & NUM_OF_TASKS)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("Num of tasks wasn't configured previousely"));
-+ if (fmPortGetSetCcParams.getCcParams.type & OFFSET_OF_DATA)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("offset of the data wasn't configured previousely"));
-+ if (fmPortGetSetCcParams.getCcParams.type & HW_PORT_ID)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("hwPortId wasn't updated"));
-+#ifdef FM_LOCKUP_ALIGNMENT_ERRATA_FMAN_SW004
-+ ASSERT_COND((fmPortGetSetCcParams.getCcParams.dataOffset % 16) == 0);
-+#endif /* FM_LOCKUP_ALIGNMENT_ERRATA_FMAN_SW004 */
-+ }
-+ else if (validate)
-+ {
-+ if ((!(p_Manip->shadowUpdateParams & NUM_OF_TASKS) &&
-+ !(p_Manip->shadowUpdateParams & OFFSET_OF_DATA) &&
-+ !(p_Manip->shadowUpdateParams & OFFSET_OF_PR) &&
-+ !(p_Manip->shadowUpdateParams & HW_PORT_ID)) &&
-+ ((p_Manip->updateParams & NUM_OF_TASKS) ||
-+ (p_Manip->updateParams & OFFSET_OF_DATA) || (p_Manip->updateParams & OFFSET_OF_PR) ||
-+ (p_Manip->updateParams & HW_PORT_ID)))
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("in this stage parameters from Port has be updated"));
-+
-+ fmPortGetSetCcParams.getCcParams.type = p_Manip->shadowUpdateParams;
-+ fmPortGetSetCcParams.setCcParams.type = UPDATE_NIA_PNEN;
-+ fmPortGetSetCcParams.setCcParams.nia = NIA_FM_CTL_AC_POP_TO_N_STEP | NIA_ENG_FM_CTL;
-+
-+ err = FmPortGetSetCcParams(h_FmPort, &fmPortGetSetCcParams);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+
-+ if (fmPortGetSetCcParams.getCcParams.type & NUM_OF_TASKS)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("NumOfTasks wasn't configured previously"));
-+ if (fmPortGetSetCcParams.getCcParams.type & OFFSET_OF_DATA)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("offset of the data wasn't configured previously"));
-+ if (fmPortGetSetCcParams.getCcParams.type & HW_PORT_ID)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("hwPortId wasn't updated"));
-+ }
-+
-+ if (p_Manip->updateParams)
-+ {
-+ if (p_Manip->updateParams & NUM_OF_TASKS)
-+ {
-+ /*recommendation of Microcode team - (maxNumFramesInProcess * 2) */
-+ size = (uint16_t)(p_Manip->capwapFragParams.maxNumFramesInProcess*2 + fmPortGetSetCcParams.getCcParams.numOfTasks);
-+ if (size > 255)
-+ RETURN_ERROR(MAJOR,E_INVALID_VALUE, ("numOfOpenReassmEntries + numOfTasks per port can not be greater than 256"));
-+
-+ p_Manip->capwapFragParams.numOfTasks = fmPortGetSetCcParams.getCcParams.numOfTasks;
-+
-+ /*p_ReassmFrmDescrIndxPoolTbl*/
-+ p_Manip->capwapFragParams.p_ReassmFrmDescrIndxPoolTbl =
-+ (t_Handle)FM_MURAM_AllocMem(p_FmPcd->h_FmMuram,
-+ (uint32_t)(size + 1),
-+ 4);
-+ if (!p_Manip->capwapFragParams.p_ReassmFrmDescrIndxPoolTbl)
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, ("MURAM alloc for CAPWAP Reassembly frame buffer index pool table"));
-+
-+ MemSet8(p_Manip->capwapFragParams.p_ReassmFrmDescrIndxPoolTbl, 0, (uint32_t)(size + 1));
-+
-+ for ( i = 0; i < size; i++)
-+ WRITE_UINT8(*(uint8_t *)PTR_MOVE(p_Manip->capwapFragParams.p_ReassmFrmDescrIndxPoolTbl, i), (uint8_t)(i+1));
-+
-+ tmpReg32 = (uint32_t)(XX_VirtToPhys(p_Manip->capwapFragParams.p_ReassmFrmDescrIndxPoolTbl) - p_FmPcd->physicalMuramBase);
-+
-+ WRITE_UINT32(p_ReassmTbl->reasmFrmDescIndexPoolTblPtr, tmpReg32);
-+
-+ /*p_ReassmFrmDescrPoolTbl*/
-+ p_Manip->capwapFragParams.p_ReassmFrmDescrPoolTbl =
-+ (t_Handle)FM_MURAM_AllocMem(p_FmPcd->h_FmMuram,
-+ (uint32_t)((size + 1) * FM_PCD_MANIP_CAPWAP_REASM_RFD_SIZE),
-+ 4);
-+
-+ if (!p_Manip->capwapFragParams.p_ReassmFrmDescrPoolTbl)
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, ("MURAM alloc for CAPWAP Reassembly frame buffer pool table"));
-+
-+ MemSet8(p_Manip->capwapFragParams.p_ReassmFrmDescrPoolTbl, 0, (uint32_t)((size +1)* FM_PCD_MANIP_CAPWAP_REASM_RFD_SIZE));
-+
-+ tmpReg32 = (uint32_t)(XX_VirtToPhys(p_Manip->capwapFragParams.p_ReassmFrmDescrPoolTbl) - p_FmPcd->physicalMuramBase);
-+
-+ WRITE_UINT32(p_ReassmTbl->reasmFrmDescPoolTblPtr, tmpReg32);
-+
-+ /*p_TimeOutTbl*/
-+
-+ p_Manip->capwapFragParams.p_TimeOutTbl =
-+ (t_Handle)FM_MURAM_AllocMem(p_FmPcd->h_FmMuram,
-+ (uint32_t)((size + 1)* FM_PCD_MANIP_CAPWAP_REASM_TIME_OUT_ENTRY_SIZE),
-+ 4);
-+
-+ if (!p_Manip->capwapFragParams.p_TimeOutTbl)
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, ("MURAM alloc for CAPWAP Reassembly timeout table"));
-+
-+ MemSet8(p_Manip->capwapFragParams.p_TimeOutTbl, 0, (uint16_t)((size + 1)*FM_PCD_MANIP_CAPWAP_REASM_TIME_OUT_ENTRY_SIZE));
-+
-+ tmpReg32 = (uint32_t)(XX_VirtToPhys(p_Manip->capwapFragParams.p_TimeOutTbl) - p_FmPcd->physicalMuramBase);
-+ WRITE_UINT32(p_ReassmTbl->timeOutTblPtr, tmpReg32);
-+
-+ p_Manip->updateParams &= ~NUM_OF_TASKS;
-+ p_Manip->shadowUpdateParams |= NUM_OF_TASKS;
-+ }
-+
-+ if (p_Manip->updateParams & OFFSET_OF_DATA)
-+ {
-+ p_Manip->capwapFragParams.dataOffset = fmPortGetSetCcParams.getCcParams.dataOffset;
-+ tmpReg32 = GET_UINT32(p_ReassmTbl->mode);
-+ tmpReg32|= p_Manip->capwapFragParams.dataOffset;
-+ WRITE_UINT32(p_ReassmTbl->mode, tmpReg32);
-+ p_Manip->updateParams &= ~OFFSET_OF_DATA;
-+ p_Manip->shadowUpdateParams |= OFFSET_OF_DATA;
-+ }
-+
-+ if (!(fmPortGetSetCcParams.getCcParams.type & OFFSET_OF_PR))
-+ {
-+ p_Manip->capwapFragParams.prOffset = fmPortGetSetCcParams.getCcParams.prOffset;
-+
-+ tmpReg32 = GET_UINT32(p_ReassmTbl->mode);
-+ tmpReg32|= FM_PCD_MANIP_CAPWAP_REASM_PR_COPY;
-+ WRITE_UINT32(p_ReassmTbl->mode, tmpReg32);
-+
-+ tmpReg32 = GET_UINT32(p_ReassmTbl->intStatsTblPtr);
-+ tmpReg32 |= (uint32_t)p_Manip->capwapFragParams.prOffset << 24;
-+ WRITE_UINT32(p_ReassmTbl->intStatsTblPtr, tmpReg32);
-+ p_Manip->updateParams &= ~OFFSET_OF_PR;
-+ p_Manip->shadowUpdateParams |= OFFSET_OF_PR;
-+ }
-+ else
-+ {
-+ p_Manip->capwapFragParams.prOffset = 0xff;
-+ p_Manip->updateParams &= ~OFFSET_OF_PR;
-+ p_Manip->shadowUpdateParams |= OFFSET_OF_PR;
-+ }
-+
-+ p_Manip->capwapFragParams.hwPortId = fmPortGetSetCcParams.getCcParams.hardwarePortId;
-+ p_Manip->updateParams &= ~HW_PORT_ID;
-+ p_Manip->shadowUpdateParams |= HW_PORT_ID;
-+
-+ /*timeout hc */
-+ ccCapwapReassmTimeoutParams.fqidForTimeOutFrames = p_Manip->capwapFragParams.fqidForTimeOutFrames;
-+ ccCapwapReassmTimeoutParams.portIdAndCapwapReassmTbl = (uint32_t)p_Manip->capwapFragParams.hwPortId << 24;
-+ ccCapwapReassmTimeoutParams.portIdAndCapwapReassmTbl |= (uint32_t)((XX_VirtToPhys(p_ReassmTbl) - p_FmPcd->physicalMuramBase));
-+ ccCapwapReassmTimeoutParams.timeoutRequestTime = (((uint32_t)1<<p_Manip->capwapFragParams.bitFor1Micro) * p_Manip->capwapFragParams.timeoutRoutineRequestTime)/2;
-+ return FmHcPcdCcCapwapTimeoutReassm(p_FmPcd->h_Hc,&ccCapwapReassmTimeoutParams);
-+ }
-+
-+ else if (validate)
-+ {
-+ if (fmPortGetSetCcParams.getCcParams.hardwarePortId != p_Manip->capwapFragParams.hwPortId)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("Reassembly manipulation previously was assigned to another port"));
-+ if (fmPortGetSetCcParams.getCcParams.numOfTasks != p_Manip->capwapFragParams.numOfTasks)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("numOfTasks for this manipulation previously was defined by another value "));
-+
-+ if (!(fmPortGetSetCcParams.getCcParams.type & OFFSET_OF_PR))
-+ {
-+ if (p_Manip->capwapFragParams.prOffset != fmPortGetSetCcParams.getCcParams.prOffset)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("Parse result offset previously was defined by another value "));
-+ }
-+ else
-+ {
-+ if (p_Manip->capwapFragParams.prOffset != 0xff)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("Parse result offset previously was defined by another value "));
-+ }
-+ if (fmPortGetSetCcParams.getCcParams.dataOffset != p_Manip->capwapFragParams.dataOffset)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("Data offset previously was defined by another value "));
-+ }
-+
-+ return E_OK;
-+}
-+#endif /* (defined(FM_CAPWAP_SUPPORT) && (DPAA_VERSION == 10)) */
-+
-+t_Error FmPcdRegisterReassmPort(t_Handle h_FmPcd, t_Handle h_ReasmCommonPramTbl)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+ t_FmPcdCcReassmTimeoutParams ccReassmTimeoutParams = { 0 };
-+ t_Error err = E_OK;
-+ uint8_t result;
-+ uint32_t bitFor1Micro, tsbs, log2num;
-+
-+ ASSERT_COND(p_FmPcd);
-+ ASSERT_COND(h_ReasmCommonPramTbl);
-+
-+ bitFor1Micro = FmGetTimeStampScale(p_FmPcd->h_Fm);
-+ if (bitFor1Micro == 0)
-+ RETURN_ERROR(MAJOR, E_NOT_AVAILABLE, ("Timestamp scale"));
-+
-+ bitFor1Micro = 32 - bitFor1Micro;
-+ LOG2(FM_PCD_MANIP_REASM_TIMEOUT_THREAD_THRESH, log2num);
-+ tsbs = bitFor1Micro - log2num;
-+
-+ ccReassmTimeoutParams.iprcpt = (uint32_t)(XX_VirtToPhys(
-+ h_ReasmCommonPramTbl) - p_FmPcd->physicalMuramBase);
-+ ccReassmTimeoutParams.tsbs = (uint8_t)tsbs;
-+ ccReassmTimeoutParams.activate = TRUE;
-+ if ((err = FmHcPcdCcTimeoutReassm(p_FmPcd->h_Hc, &ccReassmTimeoutParams,
-+ &result)) != E_OK)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+
-+ switch (result)
-+ {
-+ case (0):
-+ return E_OK;
-+ case (1):
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, ("failed to allocate TNUM"));
-+ case (2):
-+ RETURN_ERROR(
-+ MAJOR, E_NO_MEMORY,
-+ ("failed to allocate internal buffer from the HC-Port"));
-+ case (3):
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE,
-+ ("'Disable Timeout Task' with invalid IPRCPT"));
-+ case (4):
-+ RETURN_ERROR(MAJOR, E_FULL, ("too many timeout tasks"));
-+ case (5):
-+ RETURN_ERROR(MAJOR, E_INVALID_SELECTION, ("invalid sub command"));
-+ default:
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, NO_MSG);
-+ }
-+ return E_OK;
-+}
-+
-+static t_Error CreateReassCommonTable(t_FmPcdManip *p_Manip)
-+{
-+ uint32_t tmpReg32 = 0, i, bitFor1Micro;
-+ uint64_t tmpReg64, size;
-+ t_FmPcd *p_FmPcd = (t_FmPcd *)p_Manip->h_FmPcd;
-+ t_Error err = E_OK;
-+
-+ bitFor1Micro = FmGetTimeStampScale(p_FmPcd->h_Fm);
-+ if (bitFor1Micro == 0)
-+ RETURN_ERROR(MAJOR, E_NOT_AVAILABLE, ("Timestamp scale"));
-+
-+ /* Allocation of the Reassembly Common Parameters table. This table is located in the
-+ MURAM. Its size is 64 bytes and its base address should be 8-byte aligned. */
-+ p_Manip->reassmParams.p_ReassCommonTbl =
-+ (t_ReassCommonTbl *)FM_MURAM_AllocMem(
-+ p_FmPcd->h_FmMuram,
-+ FM_PCD_MANIP_REASM_COMMON_PARAM_TABLE_SIZE,
-+ FM_PCD_MANIP_REASM_COMMON_PARAM_TABLE_ALIGN);
-+
-+ if (!p_Manip->reassmParams.p_ReassCommonTbl)
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY,
-+ ("MURAM alloc for Reassembly common parameters table"));
-+
-+ MemSet8(p_Manip->reassmParams.p_ReassCommonTbl, 0,
-+ FM_PCD_MANIP_REASM_COMMON_PARAM_TABLE_SIZE);
-+
-+ /* Setting the TimeOut Mode.*/
-+ tmpReg32 = 0;
-+ if (p_Manip->reassmParams.timeOutMode
-+ == e_FM_PCD_MANIP_TIME_OUT_BETWEEN_FRAMES)
-+ tmpReg32 |= FM_PCD_MANIP_REASM_TIME_OUT_BETWEEN_FRAMES;
-+
-+ /* Setting TimeOut FQID - Frames that time out are enqueued to this FQID.
-+ In order to cause TimeOut frames to be discarded, this queue should be configured accordingly*/
-+ tmpReg32 |= p_Manip->reassmParams.fqidForTimeOutFrames;
-+ WRITE_UINT32(p_Manip->reassmParams.p_ReassCommonTbl->timeoutModeAndFqid,
-+ tmpReg32);
-+
-+ /* Calculation the size of IP Reassembly Frame Descriptor - number of frames that are allowed to be reassembled simultaneously + 129.*/
-+ size = p_Manip->reassmParams.maxNumFramesInProcess + 129;
-+
-+ /*Allocation of IP Reassembly Frame Descriptor Indexes Pool - This pool resides in the MURAM */
-+ p_Manip->reassmParams.reassFrmDescrIndxPoolTblAddr =
-+ PTR_TO_UINT(FM_MURAM_AllocMem(p_FmPcd->h_FmMuram,
-+ (uint32_t)(size * 2),
-+ 256));
-+ if (!p_Manip->reassmParams.reassFrmDescrIndxPoolTblAddr)
-+ RETURN_ERROR(
-+ MAJOR, E_NO_MEMORY,
-+ ("MURAM alloc for Reassembly frame descriptor indexes pool"));
-+
-+ MemSet8(UINT_TO_PTR(p_Manip->reassmParams.reassFrmDescrIndxPoolTblAddr),
-+ 0, (uint32_t)(size * 2));
-+
-+ /* The entries in IP Reassembly Frame Descriptor Indexes Pool contains indexes starting with 1 up to
-+ the maximum number of frames that are allowed to be reassembled simultaneously + 128.
-+ The last entry in this pool must contain the index zero*/
-+ for (i = 0; i < (size - 1); i++)
-+ WRITE_UINT16(
-+ *(uint16_t *)PTR_MOVE(UINT_TO_PTR(p_Manip->reassmParams.reassFrmDescrIndxPoolTblAddr), (i<<1)),
-+ (uint16_t)(i+1));
-+
-+ /* Sets the IP Reassembly Frame Descriptor Indexes Pool offset from MURAM */
-+ tmpReg32 = (uint32_t)(XX_VirtToPhys(
-+ UINT_TO_PTR(p_Manip->reassmParams.reassFrmDescrIndxPoolTblAddr))
-+ - p_FmPcd->physicalMuramBase);
-+ WRITE_UINT32(
-+ p_Manip->reassmParams.p_ReassCommonTbl->reassFrmDescIndexPoolTblPtr,
-+ tmpReg32);
-+
-+ /* Allocation of the Reassembly Frame Descriptors Pool - This pool resides in external memory.
-+ The number of entries in this pool should be equal to the number of entries in IP Reassembly Frame Descriptor Indexes Pool.*/
-+ p_Manip->reassmParams.reassFrmDescrPoolTblAddr =
-+ PTR_TO_UINT(XX_MallocSmart((uint32_t)(size * 64), p_Manip->reassmParams.dataMemId, 64));
-+
-+ if (!p_Manip->reassmParams.reassFrmDescrPoolTblAddr)
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, ("Memory allocation FAILED"));
-+
-+ MemSet8(UINT_TO_PTR(p_Manip->reassmParams.reassFrmDescrPoolTblAddr), 0,
-+ (uint32_t)(size * 64));
-+
-+ /* Sets the Reassembly Frame Descriptors Pool and liodn offset*/
-+ tmpReg64 = (uint64_t)(XX_VirtToPhys(
-+ UINT_TO_PTR(p_Manip->reassmParams.reassFrmDescrPoolTblAddr)));
-+ tmpReg64 |= ((uint64_t)(p_Manip->reassmParams.dataLiodnOffset
-+ & FM_PCD_MANIP_REASM_LIODN_MASK)
-+ << (uint64_t)FM_PCD_MANIP_REASM_LIODN_SHIFT);
-+ tmpReg64 |= ((uint64_t)(p_Manip->reassmParams.dataLiodnOffset
-+ & FM_PCD_MANIP_REASM_ELIODN_MASK)
-+ << (uint64_t)FM_PCD_MANIP_REASM_ELIODN_SHIFT);
-+ WRITE_UINT32(
-+ p_Manip->reassmParams.p_ReassCommonTbl->liodnAndReassFrmDescPoolPtrHi,
-+ (uint32_t)(tmpReg64 >> 32));
-+ WRITE_UINT32(
-+ p_Manip->reassmParams.p_ReassCommonTbl->reassFrmDescPoolPtrLow,
-+ (uint32_t)tmpReg64);
-+
-+ /*Allocation of the TimeOut table - This table resides in the MURAM.
-+ The number of entries in this table is identical to the number of entries in the Reassembly Frame Descriptors Pool*/
-+ p_Manip->reassmParams.timeOutTblAddr =
-+ PTR_TO_UINT(FM_MURAM_AllocMem(p_FmPcd->h_FmMuram, (uint32_t)(size * 8),8));
-+
-+ if (!p_Manip->reassmParams.timeOutTblAddr)
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY,
-+ ("MURAM alloc for Reassembly timeout table"));
-+
-+ MemSet8(UINT_TO_PTR(p_Manip->reassmParams.timeOutTblAddr), 0,
-+ (uint16_t)(size * 8));
-+
-+ /* Sets the TimeOut table offset from MURAM */
-+ tmpReg32 = (uint32_t)(XX_VirtToPhys(
-+ UINT_TO_PTR(p_Manip->reassmParams.timeOutTblAddr))
-+ - p_FmPcd->physicalMuramBase);
-+ WRITE_UINT32(p_Manip->reassmParams.p_ReassCommonTbl->timeOutTblPtr,
-+ tmpReg32);
-+
-+ /* Sets the Expiration Delay */
-+ tmpReg32 = 0;
-+ tmpReg32 |= (((uint32_t)(1 << bitFor1Micro))
-+ * p_Manip->reassmParams.timeoutThresholdForReassmProcess);
-+ WRITE_UINT32(p_Manip->reassmParams.p_ReassCommonTbl->expirationDelay,
-+ tmpReg32);
-+
-+ err = FmPcdRegisterReassmPort(p_FmPcd,
-+ p_Manip->reassmParams.p_ReassCommonTbl);
-+ if (err != E_OK)
-+ {
-+ FM_MURAM_FreeMem(p_FmPcd->h_FmMuram,
-+ p_Manip->reassmParams.p_ReassCommonTbl);
-+ RETURN_ERROR(MAJOR, err, ("port registration"));
-+ }
-+
-+ return err;
-+}
-+
-+static t_Error CreateReassTable(t_FmPcdManip *p_Manip, e_NetHeaderType hdr)
-+{
-+ t_FmPcd *p_FmPcd = p_Manip->h_FmPcd;
-+ uint32_t tmpReg32, autoLearnHashTblSize;
-+ uint32_t numOfWays, setSize, setSizeCode, keySize;
-+ uint32_t waySize, numOfSets, numOfEntries;
-+ uint64_t tmpReg64;
-+ uint16_t minFragSize;
-+ uint16_t maxReassemSize;
-+ uintptr_t *p_AutoLearnHashTblAddr, *p_AutoLearnSetLockTblAddr;
-+ t_ReassTbl **p_ReassTbl;
-+
-+ switch (hdr)
-+ {
-+ case HEADER_TYPE_IPv4:
-+ p_ReassTbl = &p_Manip->reassmParams.ip.p_Ipv4ReassTbl;
-+ p_AutoLearnHashTblAddr =
-+ &p_Manip->reassmParams.ip.ipv4AutoLearnHashTblAddr;
-+ p_AutoLearnSetLockTblAddr =
-+ &p_Manip->reassmParams.ip.ipv4AutoLearnSetLockTblAddr;
-+ minFragSize = p_Manip->reassmParams.ip.minFragSize[0];
-+ maxReassemSize = 0;
-+ numOfWays = p_Manip->reassmParams.ip.numOfFramesPerHashEntry[0];
-+ keySize = 4 + 4 + 1 + 2; /* 3-tuple + IP-Id */
-+ break;
-+ case HEADER_TYPE_IPv6:
-+ p_ReassTbl = &p_Manip->reassmParams.ip.p_Ipv6ReassTbl;
-+ p_AutoLearnHashTblAddr =
-+ &p_Manip->reassmParams.ip.ipv6AutoLearnHashTblAddr;
-+ p_AutoLearnSetLockTblAddr =
-+ &p_Manip->reassmParams.ip.ipv6AutoLearnSetLockTblAddr;
-+ minFragSize = p_Manip->reassmParams.ip.minFragSize[1];
-+ maxReassemSize = 0;
-+ numOfWays = p_Manip->reassmParams.ip.numOfFramesPerHashEntry[1];
-+ keySize = 16 + 16 + 4; /* 2-tuple + IP-Id */
-+ if (numOfWays > e_FM_PCD_MANIP_SIX_WAYS_HASH)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("num of ways"));
-+ break;
-+ case HEADER_TYPE_CAPWAP:
-+ p_ReassTbl = &p_Manip->reassmParams.capwap.p_ReassTbl;
-+ p_AutoLearnHashTblAddr =
-+ &p_Manip->reassmParams.capwap.autoLearnHashTblAddr;
-+ p_AutoLearnSetLockTblAddr =
-+ &p_Manip->reassmParams.capwap.autoLearnSetLockTblAddr;
-+ minFragSize = 0;
-+ maxReassemSize = p_Manip->reassmParams.capwap.maxRessembledsSize;
-+ numOfWays = p_Manip->reassmParams.capwap.numOfFramesPerHashEntry;
-+ keySize = 4;
-+ break;
-+ default:
-+ RETURN_ERROR(MAJOR, E_NOT_SUPPORTED, ("header type"));
-+ }
-+ keySize += 2; /* 2 bytes reserved for RFDIndex */
-+#if (DPAA_VERSION >= 11)
-+ keySize += 2; /* 2 bytes reserved */
-+#endif /* (DPAA_VERSION >= 11) */
-+ waySize = ROUND_UP(keySize, 8);
-+
-+ /* Allocates the Reassembly Parameters Table - This table is located in the MURAM.*/
-+ *p_ReassTbl = (t_ReassTbl *)FM_MURAM_AllocMem(
-+ p_FmPcd->h_FmMuram, FM_PCD_MANIP_REASM_TABLE_SIZE,
-+ FM_PCD_MANIP_REASM_TABLE_ALIGN);
-+ if (!*p_ReassTbl)
-+ RETURN_ERROR( MAJOR, E_NO_MEMORY,
-+ ("MURAM alloc for Reassembly specific parameters table"));
-+ memset(*p_ReassTbl, 0, sizeof(t_ReassTbl));
-+
-+ /* Sets the Reassembly common Parameters table offset from MURAM in the Reassembly Table descriptor*/
-+ tmpReg32 = (uint32_t)(XX_VirtToPhys(p_Manip->reassmParams.p_ReassCommonTbl)
-+ - p_FmPcd->physicalMuramBase);
-+ WRITE_UINT32((*p_ReassTbl)->reassCommonPrmTblPtr, tmpReg32);
-+
-+ /* Calculate set size (set size is rounded-up to next power of 2) */
-+ NEXT_POWER_OF_2(numOfWays * waySize, setSize);
-+
-+ /* Get set size code */
-+ LOG2(setSize, setSizeCode);
-+
-+ /* Sets ways number and set size code */
-+ WRITE_UINT16((*p_ReassTbl)->waysNumAndSetSize,
-+ (uint16_t)((numOfWays << 8) | setSizeCode));
-+
-+ /* It is recommended that the total number of entries in this table
-+ (number of sets * number of ways) will be twice the number of frames that
-+ are expected to be reassembled simultaneously.*/
-+ numOfEntries = (uint32_t)(p_Manip->reassmParams.maxNumFramesInProcess * 2);
-+
-+ /* sets number calculation - number of entries = number of sets * number of ways */
-+ numOfSets = numOfEntries / numOfWays;
-+
-+ /* Sets AutoLearnHashKeyMask*/
-+ NEXT_POWER_OF_2(numOfSets, numOfSets);
-+
-+ WRITE_UINT16((*p_ReassTbl)->autoLearnHashKeyMask,
-+ (uint16_t)(numOfSets - 1));
-+
-+ /* Allocation of Reassembly Automatic Learning Hash Table - This table resides in external memory.
-+ The size of this table is determined by the number of sets and the set size.
-+ Table size = set size * number of sets
-+ This table base address should be aligned to SetSize.*/
-+ autoLearnHashTblSize = numOfSets * setSize;
-+
-+ *p_AutoLearnHashTblAddr =
-+ PTR_TO_UINT(XX_MallocSmart(autoLearnHashTblSize, p_Manip->reassmParams.dataMemId, setSize));
-+ if (!*p_AutoLearnHashTblAddr)
-+ {
-+ FM_MURAM_FreeMem(p_FmPcd->h_FmMuram, *p_ReassTbl);
-+ *p_ReassTbl = NULL;
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, ("Memory allocation FAILED"));
-+ }
-+ MemSet8(UINT_TO_PTR(*p_AutoLearnHashTblAddr), 0, autoLearnHashTblSize);
-+
-+ /* Sets the Reassembly Automatic Learning Hash Table and liodn offset */
-+ tmpReg64 = ((uint64_t)(p_Manip->reassmParams.dataLiodnOffset
-+ & FM_PCD_MANIP_REASM_LIODN_MASK)
-+ << (uint64_t)FM_PCD_MANIP_REASM_LIODN_SHIFT);
-+ tmpReg64 |= ((uint64_t)(p_Manip->reassmParams.dataLiodnOffset
-+ & FM_PCD_MANIP_REASM_ELIODN_MASK)
-+ << (uint64_t)FM_PCD_MANIP_REASM_ELIODN_SHIFT);
-+ tmpReg64 |= XX_VirtToPhys(UINT_TO_PTR(*p_AutoLearnHashTblAddr));
-+ WRITE_UINT32( (*p_ReassTbl)->liodnAlAndAutoLearnHashTblPtrHi,
-+ (uint32_t)(tmpReg64 >> 32));
-+ WRITE_UINT32((*p_ReassTbl)->autoLearnHashTblPtrLow, (uint32_t)tmpReg64);
-+
-+ /* Allocation of the Set Lock table - This table resides in external memory
-+ The size of this table is (number of sets in the Reassembly Automatic Learning Hash table)*4 bytes.
-+ This table resides in external memory and its base address should be 4-byte aligned */
-+ *p_AutoLearnSetLockTblAddr =
-+ PTR_TO_UINT(XX_MallocSmart((uint32_t)(numOfSets * 4), p_Manip->reassmParams.dataMemId, 4));
-+ if (!*p_AutoLearnSetLockTblAddr)
-+ {
-+ FM_MURAM_FreeMem(p_FmPcd->h_FmMuram, *p_ReassTbl);
-+ *p_ReassTbl = NULL;
-+ XX_FreeSmart(UINT_TO_PTR(*p_AutoLearnHashTblAddr));
-+ *p_AutoLearnHashTblAddr = 0;
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, ("Memory allocation FAILED"));
-+ }
-+ MemSet8(UINT_TO_PTR(*p_AutoLearnSetLockTblAddr), 0, (numOfSets * 4));
-+
-+ /* sets Set Lock table pointer and liodn offset*/
-+ tmpReg64 = ((uint64_t)(p_Manip->reassmParams.dataLiodnOffset
-+ & FM_PCD_MANIP_REASM_LIODN_MASK)
-+ << (uint64_t)FM_PCD_MANIP_REASM_LIODN_SHIFT);
-+ tmpReg64 |= ((uint64_t)(p_Manip->reassmParams.dataLiodnOffset
-+ & FM_PCD_MANIP_REASM_ELIODN_MASK)
-+ << (uint64_t)FM_PCD_MANIP_REASM_ELIODN_SHIFT);
-+ tmpReg64 |= XX_VirtToPhys(UINT_TO_PTR(*p_AutoLearnSetLockTblAddr));
-+ WRITE_UINT32( (*p_ReassTbl)->liodnSlAndAutoLearnSetLockTblPtrHi,
-+ (uint32_t)(tmpReg64 >> 32));
-+ WRITE_UINT32((*p_ReassTbl)->autoLearnSetLockTblPtrLow, (uint32_t)tmpReg64);
-+
-+ /* Sets user's requested minimum fragment size (in Bytes) for First/Middle fragment */
-+ WRITE_UINT16((*p_ReassTbl)->minFragSize, minFragSize);
-+
-+ WRITE_UINT16((*p_ReassTbl)->maxReassemblySize, maxReassemSize);
-+
-+ return E_OK;
-+}
-+
-+static t_Error UpdateInitReasm(t_Handle h_FmPcd, t_Handle h_PcdParams,
-+ t_Handle h_FmPort, t_FmPcdManip *p_Manip,
-+ t_Handle h_Ad, bool validate)
-+{
-+ t_FmPortGetSetCcParams fmPortGetSetCcParams;
-+ uint32_t tmpReg32;
-+ t_Error err;
-+ t_FmPortPcdParams *p_PcdParams = (t_FmPortPcdParams *)h_PcdParams;
-+#if (DPAA_VERSION >= 11)
-+ t_FmPcdCtrlParamsPage *p_ParamsPage;
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Manip, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Manip->frag, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(
-+ (p_Manip->opcode == HMAN_OC_IP_REASSEMBLY) || (p_Manip->opcode == HMAN_OC_CAPWAP_REASSEMBLY),
-+ E_INVALID_STATE);
-+ SANITY_CHECK_RETURN_ERROR(h_FmPcd, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Manip->updateParams || h_PcdParams,
-+ E_INVALID_HANDLE);
-+
-+ UNUSED(h_Ad);
-+
-+ if (!p_Manip->updateParams)
-+ return E_OK;
-+
-+ if (p_Manip->h_FmPcd != h_FmPcd)
-+ RETURN_ERROR(
-+ MAJOR, E_INVALID_STATE,
-+ ("handler of PCD previously was initiated by different value"));
-+
-+ if (p_Manip->updateParams)
-+ {
-+ if ((!(p_Manip->updateParams
-+ & (NUM_OF_TASKS | NUM_OF_EXTRA_TASKS | DISCARD_MASK)))
-+ || ((p_Manip->shadowUpdateParams
-+ & (NUM_OF_TASKS | NUM_OF_EXTRA_TASKS | DISCARD_MASK))))
-+ RETURN_ERROR(
-+ MAJOR, E_INVALID_STATE,
-+ ("in this stage parameters from Port has not be updated"));
-+
-+ fmPortGetSetCcParams.setCcParams.type = 0;
-+ if (p_Manip->opcode == HMAN_OC_CAPWAP_REASSEMBLY)
-+ {
-+ fmPortGetSetCcParams.setCcParams.type |= UPDATE_OFP_DPTE;
-+ fmPortGetSetCcParams.setCcParams.ofpDpde = 0xF;
-+ }
-+ fmPortGetSetCcParams.getCcParams.type = p_Manip->updateParams | FM_REV;
-+ if ((err = FmPortGetSetCcParams(h_FmPort, &fmPortGetSetCcParams))
-+ != E_OK)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ if (fmPortGetSetCcParams.getCcParams.type
-+ & (NUM_OF_TASKS | NUM_OF_EXTRA_TASKS | DISCARD_MASK | FM_REV))
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE,
-+ ("offset of the data wasn't configured previously"));
-+ if (p_Manip->updateParams
-+ & (NUM_OF_TASKS | NUM_OF_EXTRA_TASKS | DISCARD_MASK))
-+ {
-+ t_FmPcd *p_FmPcd = (t_FmPcd *)h_FmPcd;
-+ uint8_t *p_Ptr, i, totalNumOfTnums;
-+
-+ totalNumOfTnums =
-+ (uint8_t)(fmPortGetSetCcParams.getCcParams.numOfTasks
-+ + fmPortGetSetCcParams.getCcParams.numOfExtraTasks);
-+
-+ p_Manip->reassmParams.internalBufferPoolAddr =
-+ PTR_TO_UINT(FM_MURAM_AllocMem(p_FmPcd->h_FmMuram,
-+ (uint32_t)(totalNumOfTnums * BMI_FIFO_UNITS),
-+ BMI_FIFO_UNITS));
-+ if (!p_Manip->reassmParams.internalBufferPoolAddr)
-+ RETURN_ERROR(
-+ MAJOR, E_NO_MEMORY,
-+ ("MURAM alloc for Reassembly internal buffers pool"));
-+ MemSet8(
-+ UINT_TO_PTR(p_Manip->reassmParams.internalBufferPoolAddr),
-+ 0, (uint32_t)(totalNumOfTnums * BMI_FIFO_UNITS));
-+
-+ p_Manip->reassmParams.internalBufferPoolManagementIndexAddr =
-+ PTR_TO_UINT(FM_MURAM_AllocMem(p_FmPcd->h_FmMuram,
-+ (uint32_t)(5 + totalNumOfTnums),
-+ 4));
-+ if (!p_Manip->reassmParams.internalBufferPoolManagementIndexAddr)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_NO_MEMORY,
-+ ("MURAM alloc for Reassembly internal buffers management"));
-+
-+ p_Ptr =
-+ (uint8_t*)UINT_TO_PTR(p_Manip->reassmParams.internalBufferPoolManagementIndexAddr);
-+ WRITE_UINT32(
-+ *(uint32_t*)p_Ptr,
-+ (uint32_t)(XX_VirtToPhys(UINT_TO_PTR(p_Manip->reassmParams.internalBufferPoolAddr)) - p_FmPcd->physicalMuramBase));
-+ for (i = 0, p_Ptr += 4; i < totalNumOfTnums; i++, p_Ptr++)
-+ WRITE_UINT8(*p_Ptr, i);
-+ WRITE_UINT8(*p_Ptr, 0xFF);
-+
-+ tmpReg32 =
-+ (4 << FM_PCD_MANIP_REASM_COMMON_INT_BUFFER_IDX_SHIFT)
-+ | ((uint32_t)(XX_VirtToPhys(
-+ UINT_TO_PTR(p_Manip->reassmParams.internalBufferPoolManagementIndexAddr))
-+ - p_FmPcd->physicalMuramBase));
-+ WRITE_UINT32(
-+ p_Manip->reassmParams.p_ReassCommonTbl->internalBufferManagement,
-+ tmpReg32);
-+
-+ p_Manip->updateParams &= ~(NUM_OF_TASKS | NUM_OF_EXTRA_TASKS
-+ | DISCARD_MASK);
-+ p_Manip->shadowUpdateParams |= (NUM_OF_TASKS | NUM_OF_EXTRA_TASKS
-+ | DISCARD_MASK);
-+ }
-+ }
-+
-+ if (p_Manip->opcode == HMAN_OC_CAPWAP_REASSEMBLY)
-+ {
-+ if (p_Manip->reassmParams.capwap.h_Scheme)
-+ {
-+ p_PcdParams->p_KgParams->h_Schemes[p_PcdParams->p_KgParams->numOfSchemes] =
-+ p_Manip->reassmParams.capwap.h_Scheme;
-+ p_PcdParams->p_KgParams->numOfSchemes++;
-+ }
-+
-+ }
-+ else
-+ {
-+ if (p_Manip->reassmParams.ip.h_Ipv4Scheme)
-+ {
-+ p_PcdParams->p_KgParams->h_Schemes[p_PcdParams->p_KgParams->numOfSchemes] =
-+ p_Manip->reassmParams.ip.h_Ipv4Scheme;
-+ p_PcdParams->p_KgParams->numOfSchemes++;
-+ }
-+ if (p_Manip->reassmParams.ip.h_Ipv6Scheme)
-+ {
-+ p_PcdParams->p_KgParams->h_Schemes[p_PcdParams->p_KgParams->numOfSchemes] =
-+ p_Manip->reassmParams.ip.h_Ipv6Scheme;
-+ p_PcdParams->p_KgParams->numOfSchemes++;
-+ }
-+#if (DPAA_VERSION >= 11)
-+ if (fmPortGetSetCcParams.getCcParams.revInfo.majorRev >= 6)
-+ {
-+ if ((err = FmPortSetGprFunc(h_FmPort, e_FM_PORT_GPR_MURAM_PAGE,
-+ (void**)&p_ParamsPage)) != E_OK)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+
-+ tmpReg32 = NIA_ENG_KG;
-+ if (p_Manip->reassmParams.ip.h_Ipv4Scheme)
-+ {
-+ tmpReg32 |= NIA_KG_DIRECT;
-+ tmpReg32 |= NIA_KG_CC_EN;
-+ tmpReg32 |= FmPcdKgGetSchemeId(
-+ p_Manip->reassmParams.ip.h_Ipv4Scheme);
-+ WRITE_UINT32(p_ParamsPage->iprIpv4Nia, tmpReg32);
-+ }
-+ if (p_Manip->reassmParams.ip.h_Ipv6Scheme)
-+ {
-+ tmpReg32 &= ~NIA_AC_MASK;
-+ tmpReg32 |= NIA_KG_DIRECT;
-+ tmpReg32 |= NIA_KG_CC_EN;
-+ tmpReg32 |= FmPcdKgGetSchemeId(
-+ p_Manip->reassmParams.ip.h_Ipv6Scheme);
-+ WRITE_UINT32(p_ParamsPage->iprIpv6Nia, tmpReg32);
-+ }
-+ }
-+#else
-+ if (fmPortGetSetCcParams.getCcParams.revInfo.majorRev < 6)
-+ {
-+ WRITE_UINT32(p_Manip->reassmParams.p_ReassCommonTbl->discardMask,
-+ fmPortGetSetCcParams.getCcParams.discardMask);
-+ }
-+#endif /* (DPAA_VERSION >= 11) */
-+ }
-+ return E_OK;
-+}
-+
-+#if (DPAA_VERSION == 10)
-+static t_Error FmPcdFragHcScratchPoolFill(t_Handle h_FmPcd, uint8_t scratchBpid)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+ t_FmPcdCcFragScratchPoolCmdParams fmPcdCcFragScratchPoolCmdParams;
-+ t_Error err;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd, E_INVALID_HANDLE);
-+
-+ memset(&fmPcdCcFragScratchPoolCmdParams, 0, sizeof(t_FmPcdCcFragScratchPoolCmdParams));
-+
-+ fmPcdCcFragScratchPoolCmdParams.numOfBuffers = NUM_OF_SCRATCH_POOL_BUFFERS;
-+ fmPcdCcFragScratchPoolCmdParams.bufferPoolId = scratchBpid;
-+ if ((err = FmHcPcdCcIpFragScratchPollCmd(p_FmPcd->h_Hc, TRUE, &fmPcdCcFragScratchPoolCmdParams)) != E_OK)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+
-+ if (fmPcdCcFragScratchPoolCmdParams.numOfBuffers != 0)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("Fill scratch pool failed,"
-+ "Failed to release %d buffers to the BM (missing FBPRs)",
-+ fmPcdCcFragScratchPoolCmdParams.numOfBuffers));
-+
-+ return E_OK;
-+}
-+
-+static t_Error FmPcdFragHcScratchPoolEmpty(t_Handle h_FmPcd, uint8_t scratchBpid)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+ t_FmPcdCcFragScratchPoolCmdParams fmPcdCcFragScratchPoolCmdParams;
-+ t_Error err;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd, E_INVALID_HANDLE);
-+
-+ memset(&fmPcdCcFragScratchPoolCmdParams, 0, sizeof(t_FmPcdCcFragScratchPoolCmdParams));
-+
-+ fmPcdCcFragScratchPoolCmdParams.bufferPoolId = scratchBpid;
-+ if ((err = FmHcPcdCcIpFragScratchPollCmd(p_FmPcd->h_Hc, FALSE, &fmPcdCcFragScratchPoolCmdParams)) != E_OK)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+
-+ return E_OK;
-+}
-+#endif /* (DPAA_VERSION == 10) */
-+
-+static void ReleaseManipHandler(t_FmPcdManip *p_Manip, t_FmPcd *p_FmPcd)
-+{
-+ if (p_Manip->h_Ad)
-+ {
-+ if (p_Manip->muramAllocate)
-+ FM_MURAM_FreeMem(p_FmPcd->h_FmMuram, p_Manip->h_Ad);
-+ else
-+ XX_Free(p_Manip->h_Ad);
-+ p_Manip->h_Ad = NULL;
-+ }
-+ if (p_Manip->p_Template)
-+ {
-+ FM_MURAM_FreeMem(p_FmPcd->h_FmMuram, p_Manip->p_Template);
-+ p_Manip->p_Template = NULL;
-+ }
-+#if (defined(FM_CAPWAP_SUPPORT) && (DPAA_VERSION == 10))
-+ if (p_Manip->h_Frag)
-+ {
-+ if (p_Manip->capwapFragParams.p_AutoLearnHashTbl)
-+ FM_MURAM_FreeMem(p_FmPcd->h_FmMuram,
-+ p_Manip->capwapFragParams.p_AutoLearnHashTbl);
-+ if (p_Manip->capwapFragParams.p_ReassmFrmDescrPoolTbl)
-+ FM_MURAM_FreeMem(p_FmPcd->h_FmMuram,
-+ p_Manip->capwapFragParams.p_ReassmFrmDescrPoolTbl);
-+ if (p_Manip->capwapFragParams.p_ReassmFrmDescrIndxPoolTbl)
-+ FM_MURAM_FreeMem(p_FmPcd->h_FmMuram,
-+ p_Manip->capwapFragParams.p_ReassmFrmDescrIndxPoolTbl);
-+ if (p_Manip->capwapFragParams.p_TimeOutTbl)
-+ FM_MURAM_FreeMem(p_FmPcd->h_FmMuram,
-+ p_Manip->capwapFragParams.p_TimeOutTbl);
-+ FM_MURAM_FreeMem(p_FmPcd->h_FmMuram, p_Manip->h_Frag);
-+
-+ }
-+#endif /* (defined(FM_CAPWAP_SUPPORT) && (DPAA_VERSION == 10)) */
-+ if (p_Manip->frag)
-+ {
-+ if (p_Manip->fragParams.p_Frag)
-+ {
-+#if (DPAA_VERSION == 10)
-+ FmPcdFragHcScratchPoolEmpty((t_Handle)p_FmPcd, p_Manip->fragParams.scratchBpid);
-+#endif /* (DPAA_VERSION == 10) */
-+
-+ FM_MURAM_FreeMem(p_FmPcd->h_FmMuram, p_Manip->fragParams.p_Frag);
-+ }
-+ }
-+ else
-+ if (p_Manip->reassm)
-+ {
-+ FmPcdUnregisterReassmPort(p_FmPcd,
-+ p_Manip->reassmParams.p_ReassCommonTbl);
-+
-+ if (p_Manip->reassmParams.timeOutTblAddr)
-+ FM_MURAM_FreeMem(
-+ p_FmPcd->h_FmMuram,
-+ UINT_TO_PTR(p_Manip->reassmParams.timeOutTblAddr));
-+ if (p_Manip->reassmParams.reassFrmDescrPoolTblAddr)
-+ XX_FreeSmart(
-+ UINT_TO_PTR(p_Manip->reassmParams.reassFrmDescrPoolTblAddr));
-+ if (p_Manip->reassmParams.p_ReassCommonTbl)
-+ FM_MURAM_FreeMem(p_FmPcd->h_FmMuram,
-+ p_Manip->reassmParams.p_ReassCommonTbl);
-+ if (p_Manip->reassmParams.reassFrmDescrIndxPoolTblAddr)
-+ FM_MURAM_FreeMem(
-+ p_FmPcd->h_FmMuram,
-+ UINT_TO_PTR(p_Manip->reassmParams.reassFrmDescrIndxPoolTblAddr));
-+ if (p_Manip->reassmParams.internalBufferPoolManagementIndexAddr)
-+ FM_MURAM_FreeMem(
-+ p_FmPcd->h_FmMuram,
-+ UINT_TO_PTR(p_Manip->reassmParams.internalBufferPoolManagementIndexAddr));
-+ if (p_Manip->reassmParams.internalBufferPoolAddr)
-+ FM_MURAM_FreeMem(
-+ p_FmPcd->h_FmMuram,
-+ UINT_TO_PTR(p_Manip->reassmParams.internalBufferPoolAddr));
-+ if (p_Manip->reassmParams.hdr == HEADER_TYPE_CAPWAP)
-+ {
-+
-+ }
-+ else
-+ {
-+ if (p_Manip->reassmParams.ip.ipv4AutoLearnHashTblAddr)
-+ XX_FreeSmart(
-+ UINT_TO_PTR(p_Manip->reassmParams.ip.ipv4AutoLearnHashTblAddr));
-+ if (p_Manip->reassmParams.ip.ipv6AutoLearnHashTblAddr)
-+ XX_FreeSmart(
-+ UINT_TO_PTR(p_Manip->reassmParams.ip.ipv6AutoLearnHashTblAddr));
-+ if (p_Manip->reassmParams.ip.ipv4AutoLearnSetLockTblAddr)
-+ XX_FreeSmart(
-+ UINT_TO_PTR(p_Manip->reassmParams.ip.ipv4AutoLearnSetLockTblAddr));
-+ if (p_Manip->reassmParams.ip.ipv6AutoLearnSetLockTblAddr)
-+ XX_FreeSmart(
-+ UINT_TO_PTR(p_Manip->reassmParams.ip.ipv6AutoLearnSetLockTblAddr));
-+ if (p_Manip->reassmParams.ip.p_Ipv4ReassTbl)
-+ FM_MURAM_FreeMem(p_FmPcd->h_FmMuram,
-+ p_Manip->reassmParams.ip.p_Ipv4ReassTbl);
-+ if (p_Manip->reassmParams.ip.p_Ipv6ReassTbl)
-+ FM_MURAM_FreeMem(p_FmPcd->h_FmMuram,
-+ p_Manip->reassmParams.ip.p_Ipv6ReassTbl);
-+ if (p_Manip->reassmParams.ip.h_Ipv6Ad)
-+ XX_FreeSmart(p_Manip->reassmParams.ip.h_Ipv6Ad);
-+ if (p_Manip->reassmParams.ip.h_Ipv4Ad)
-+ XX_FreeSmart(p_Manip->reassmParams.ip.h_Ipv4Ad);
-+ }
-+ }
-+
-+ if (p_Manip->p_StatsTbl)
-+ FM_MURAM_FreeMem(p_FmPcd->h_FmMuram, p_Manip->p_StatsTbl);
-+}
-+
-+#if (defined(FM_CAPWAP_SUPPORT) && (DPAA_VERSION == 10))
-+static t_Error CheckManipParamsAndSetType(t_FmPcdManip *p_Manip, t_FmPcdManipParams *p_ManipParams)
-+{
-+ if (p_ManipParams->u.hdr.rmv)
-+ {
-+ switch (p_ManipParams->u.hdr.rmvParams.type)
-+ {
-+ case (e_FM_PCD_MANIP_RMV_BY_HDR):
-+ switch (p_ManipParams->u.hdr.rmvParams.u.byHdr.type)
-+ {
-+ case (e_FM_PCD_MANIP_RMV_BY_HDR_FROM_START) :
-+ if (p_ManipParams->u.hdr.rmvParams.u.byHdr.u.fromStartByHdr.include)
-+ {
-+ switch (p_ManipParams->u.hdr.rmvParams.u.byHdr.u.fromStartByHdr.hdrInfo.hdr)
-+ {
-+ case (HEADER_TYPE_CAPWAP_DTLS) :
-+ p_Manip->opcode = HMAN_OC_CAPWAP_RMV_DTLS_IF_EXIST;
-+ p_Manip->muramAllocate = TRUE;
-+ if (p_ManipParams->u.hdr.insrt)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("for CAPWAP_DTLS_HDR remove can not be insrt manipualtion after"));
-+ if (p_ManipParams->fragOrReasm)
-+ {
-+ if (!p_ManipParams->fragOrReasmParams.frag)
-+ {
-+ switch (p_ManipParams->fragOrReasmParams.hdr)
-+ {
-+ case (HEADER_TYPE_CAPWAP):
-+ p_Manip->opcode = HMAN_OC_CAPWAP_REASSEMBLY;
-+ break;
-+ default:
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("unsupported header for Reassembly"));
-+ }
-+ }
-+ else
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("for this type of manipulation frag can not be TRUE"));
-+ }
-+ break;
-+ default:
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("non valid net header of remove location"));
-+ }
-+ }
-+ else
-+ {
-+ switch (p_ManipParams->u.hdr.rmvParams.u.byHdr.u.fromStartByHdr.hdrInfo.hdr)
-+ {
-+ case (HEADER_TYPE_CAPWAP_DTLS) :
-+ case (HEADER_TYPE_CAPWAP) :
-+ if (p_ManipParams->fragOrReasm || p_ManipParams->u.hdr.insrt)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("for the type of remove e_FM_PCD_MANIP_RMV_FROM_START_OF_FRAME_TILL_CAPWAP can not be insert or fragOrReasm TRUE"));
-+ p_Manip->opcode = HMAN_OC_RMV_N_OR_INSRT_INT_FRM_HDR;
-+ p_Manip->muramAllocate = TRUE;
-+ p_ManipParams->u.hdr.insrt = TRUE; //internal frame header
-+ break;
-+ default :
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("invalid type of remove manipulation"));
-+ }
-+ }
-+ break;
-+ default :
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("invalid type of remove manipulation"));
-+ }
-+ break;
-+ default:
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("invalid type of remove manipulation"));
-+ }
-+ }
-+ else if (p_ManipParams->u.hdr.insrt)
-+ {
-+ switch (p_ManipParams->u.hdr.insrtParams.type)
-+ {
-+ case (e_FM_PCD_MANIP_INSRT_BY_TEMPLATE) :
-+
-+ p_Manip->opcode = HMAN_OC_INSRT_HDR_BY_TEMPL_N_OR_FRAG_AFTER;
-+ p_Manip->muramAllocate = FALSE;
-+ if (p_ManipParams->fragOrReasm)
-+ {
-+ if (p_ManipParams->fragOrReasmParams.frag)
-+ {
-+ switch (p_ManipParams->fragOrReasmParams.hdr)
-+ {
-+ case (HEADER_TYPE_CAPWAP):
-+ p_Manip->opcode = HMAN_OC_CAPWAP_FRAGMENTATION;
-+ break;
-+ default:
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("Invalid header for fragmentation"));
-+ }
-+ }
-+ else
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE,("can not reach this point"));
-+ }
-+ break;
-+
-+ default:
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("for only isert manipulation unsupported type"));
-+ }
-+ }
-+ else if (p_ManipParams->fragOrReasm)
-+ {
-+ if (p_ManipParams->fragOrReasmParams.frag)
-+ {
-+ switch (p_ManipParams->fragOrReasmParams.hdr)
-+ {
-+ case (HEADER_TYPE_CAPWAP):
-+ p_Manip->opcode = HMAN_OC_CAPWAP_FRAGMENTATION;
-+ p_Manip->muramAllocate = FALSE;
-+ break;
-+ default:
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("Unsupported header for fragmentation"));
-+ }
-+ }
-+ else
-+ {
-+ switch (p_ManipParams->fragOrReasmParams.hdr)
-+ {
-+ case (HEADER_TYPE_CAPWAP):
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("Reassembly has to be with additional operation - rmv = TRUE, type of remove - e_FM_PCD_MANIP_RMV_FROM_START_OF_FRAME_INCLUDE_SPECIFIC_LOCATION,type = e_FM_PCD_MANIP_LOC_BY_HDR, hdr = HEADER_TYPE_CAPWAP_DTLS"));
-+ default:
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("Unsupported header for reassembly"));
-+ }
-+ }
-+
-+ }
-+ else
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("User didn't ask for any manipulation"));
-+
-+ p_Manip->insrt = p_ManipParams->u.hdr.insrt;
-+ p_Manip->rmv = p_ManipParams->u.hdr.rmv;
-+
-+ return E_OK;
-+}
-+
-+#else /* not (defined(FM_CAPWAP_SUPPORT) && (DPAA_VERSION == 10)) */
-+static t_Error CheckManipParamsAndSetType(t_FmPcdManip *p_Manip,
-+ t_FmPcdManipParams *p_ManipParams)
-+{
-+ switch (p_ManipParams->type)
-+ {
-+ case e_FM_PCD_MANIP_HDR:
-+ /* Check that next-manip is not already used */
-+ if (p_ManipParams->h_NextManip)
-+ {
-+ if (!MANIP_IS_FIRST(p_ManipParams->h_NextManip))
-+ RETURN_ERROR(
-+ MAJOR, E_INVALID_STATE,
-+ ("h_NextManip is already a part of another chain"));
-+ if ((MANIP_GET_TYPE(p_ManipParams->h_NextManip)
-+ != e_FM_PCD_MANIP_HDR) &&
-+ (MANIP_GET_TYPE(p_ManipParams->h_NextManip)
-+ != e_FM_PCD_MANIP_FRAG))
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_NOT_SUPPORTED,
-+ ("For a Header Manipulation node - no support of h_NextManip of type other than Header Manipulation or Fragmentation."));
-+ }
-+
-+ if (p_ManipParams->u.hdr.rmv)
-+ {
-+ switch (p_ManipParams->u.hdr.rmvParams.type)
-+ {
-+ case (e_FM_PCD_MANIP_RMV_BY_HDR):
-+ switch (p_ManipParams->u.hdr.rmvParams.u.byHdr.type)
-+ {
-+ case (e_FM_PCD_MANIP_RMV_BY_HDR_SPECIFIC_L2):
-+ break;
-+#if (DPAA_VERSION >= 11)
-+ case (e_FM_PCD_MANIP_RMV_BY_HDR_CAPWAP):
-+ break;
-+ case (e_FM_PCD_MANIP_RMV_BY_HDR_FROM_START):
-+ {
-+ t_Error err;
-+ uint8_t prsArrayOffset;
-+
-+ err =
-+ GetPrOffsetByHeaderOrField(
-+ &p_ManipParams->u.hdr.rmvParams.u.byHdr.u.hdrInfo,
-+ &prsArrayOffset);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ break;
-+ }
-+#endif /* (DPAA_VERSION >= 11) */
-+ default:
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_STATE,
-+ ("invalid type of remove manipulation"));
-+ }
-+ break;
-+ case (e_FM_PCD_MANIP_RMV_GENERIC):
-+ break;
-+ default:
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE,
-+ ("invalid type of remove manipulation"));
-+ }
-+ p_Manip->opcode = HMAN_OC;
-+ p_Manip->muramAllocate = TRUE;
-+ p_Manip->rmv = TRUE;
-+ }
-+ else
-+ if (p_ManipParams->u.hdr.insrt)
-+ {
-+ switch (p_ManipParams->u.hdr.insrtParams.type)
-+ {
-+ case (e_FM_PCD_MANIP_INSRT_BY_HDR):
-+ {
-+ switch (p_ManipParams->u.hdr.insrtParams.u.byHdr.type)
-+ {
-+ case (e_FM_PCD_MANIP_INSRT_BY_HDR_SPECIFIC_L2):
-+ /* nothing to check */
-+ break;
-+#if (DPAA_VERSION >= 11)
-+ case (e_FM_PCD_MANIP_INSRT_BY_HDR_IP):
-+ if (p_ManipParams->u.hdr.insrtParams.u.byHdr.u.ipParams.insrt.size
-+ % 4)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("IP inserted header must be of size which is a multiple of four bytes"));
-+ break;
-+ case (e_FM_PCD_MANIP_INSRT_BY_HDR_CAPWAP):
-+ if (p_ManipParams->u.hdr.insrtParams.u.byHdr.u.insrt.size
-+ % 4)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("CAPWAP inserted header must be of size which is a multiple of four bytes"));
-+ break;
-+ case (e_FM_PCD_MANIP_INSRT_BY_HDR_UDP):
-+ case (e_FM_PCD_MANIP_INSRT_BY_HDR_UDP_LITE):
-+ if (p_ManipParams->u.hdr.insrtParams.u.byHdr.u.insrt.size
-+ != 8)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("Inserted header must be of size 8"));
-+ break;
-+#endif /* (DPAA_VERSION >= 11) */
-+ default:
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_STATE,
-+ ("unsupported insert by header type"));
-+ }
-+ }
-+ case (e_FM_PCD_MANIP_INSRT_GENERIC):
-+ break;
-+ default:
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_STATE,
-+ ("for only insert manipulation unsupported type"));
-+ }
-+ p_Manip->opcode = HMAN_OC;
-+ p_Manip->muramAllocate = TRUE;
-+ p_Manip->insrt = TRUE;
-+ }
-+ else
-+ if (p_ManipParams->u.hdr.fieldUpdate)
-+ {
-+ /* Check parameters */
-+ if (p_ManipParams->u.hdr.fieldUpdateParams.type
-+ == e_FM_PCD_MANIP_HDR_FIELD_UPDATE_VLAN)
-+ {
-+ if ((p_ManipParams->u.hdr.fieldUpdateParams.u.vlan.updateType
-+ == e_FM_PCD_MANIP_HDR_FIELD_UPDATE_VLAN_VPRI)
-+ && (p_ManipParams->u.hdr.fieldUpdateParams.u.vlan.u.vpri
-+ > 7))
-+ RETURN_ERROR(
-+ MAJOR, E_INVALID_VALUE,
-+ ("vpri should get values of 0-7 "));
-+ if (p_ManipParams->u.hdr.fieldUpdateParams.u.vlan.updateType
-+ == e_FM_PCD_MANIP_HDR_FIELD_UPDATE_DSCP_TO_VLAN)
-+ {
-+ int i;
-+
-+ if (p_ManipParams->u.hdr.fieldUpdateParams.u.vlan.u.dscpToVpri.vpriDefVal
-+ > 7)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("vpriDefVal should get values of 0-7 "));
-+ for (i = 0; i < FM_PCD_MANIP_DSCP_TO_VLAN_TRANS;
-+ i++)
-+ if (p_ManipParams->u.hdr.fieldUpdateParams.u.vlan.u.dscpToVpri.dscpToVpriTable[i]
-+ & 0xf0)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("dscpToVpriTabl value out of range (0-15)"));
-+ }
-+
-+ }
-+
-+ p_Manip->opcode = HMAN_OC;
-+ p_Manip->muramAllocate = TRUE;
-+ p_Manip->fieldUpdate = TRUE;
-+ }
-+ else
-+ if (p_ManipParams->u.hdr.custom)
-+ {
-+ if (p_ManipParams->u.hdr.customParams.type == e_FM_PCD_MANIP_HDR_CUSTOM_GEN_FIELD_REPLACE)
-+ {
-+
-+ if ((p_ManipParams->u.hdr.customParams.u.genFieldReplace.size == 0) ||
-+ (p_ManipParams->u.hdr.customParams.u.genFieldReplace.size > 8))
-+ RETURN_ERROR(
-+ MAJOR, E_INVALID_VALUE,
-+ ("size should get values of 1-8 "));
-+
-+ if (p_ManipParams->u.hdr.customParams.u.genFieldReplace.srcOffset > 7)
-+ RETURN_ERROR(
-+ MAJOR, E_INVALID_VALUE,
-+ ("srcOffset should be <= 7"));
-+
-+ if ((p_ManipParams->u.hdr.customParams.u.genFieldReplace.srcOffset +
-+ p_ManipParams->u.hdr.customParams.u.genFieldReplace.size) > 8)
-+ RETURN_ERROR(
-+ MAJOR, E_INVALID_VALUE,
-+ ("(srcOffset + size) should be <= 8"));
-+
-+ if ((p_ManipParams->u.hdr.customParams.u.genFieldReplace.dstOffset +
-+ p_ManipParams->u.hdr.customParams.u.genFieldReplace.size) > 256)
-+ RETURN_ERROR(
-+ MAJOR, E_INVALID_VALUE,
-+ ("(dstOffset + size) should be <= 256"));
-+
-+ }
-+
-+ p_Manip->opcode = HMAN_OC;
-+ p_Manip->muramAllocate = TRUE;
-+ p_Manip->custom = TRUE;
-+ }
-+ break;
-+ case e_FM_PCD_MANIP_REASSEM:
-+ if (p_ManipParams->h_NextManip)
-+ RETURN_ERROR(MAJOR, E_NOT_SUPPORTED,
-+ ("next manip with reassembly"));
-+ switch (p_ManipParams->u.reassem.hdr)
-+ {
-+ case (HEADER_TYPE_IPv4):
-+ p_Manip->reassmParams.hdr = HEADER_TYPE_IPv4;
-+ p_Manip->opcode = HMAN_OC_IP_REASSEMBLY;
-+ break;
-+ case (HEADER_TYPE_IPv6):
-+ p_Manip->reassmParams.hdr = HEADER_TYPE_IPv6;
-+ p_Manip->opcode = HMAN_OC_IP_REASSEMBLY;
-+ break;
-+#if (DPAA_VERSION >= 11)
-+ case (HEADER_TYPE_CAPWAP):
-+ p_Manip->reassmParams.hdr = HEADER_TYPE_CAPWAP;
-+ p_Manip->opcode = HMAN_OC_CAPWAP_REASSEMBLY;
-+ break;
-+#endif /* (DPAA_VERSION >= 11) */
-+ default:
-+ RETURN_ERROR(MAJOR, E_NOT_SUPPORTED,
-+ ("header for reassembly"));
-+ }
-+ break;
-+ case e_FM_PCD_MANIP_FRAG:
-+ if (p_ManipParams->h_NextManip)
-+ RETURN_ERROR(MAJOR, E_NOT_SUPPORTED,
-+ ("next manip with fragmentation"));
-+ switch (p_ManipParams->u.frag.hdr)
-+ {
-+ case (HEADER_TYPE_IPv4):
-+ case (HEADER_TYPE_IPv6):
-+ p_Manip->opcode = HMAN_OC_IP_FRAGMENTATION;
-+ break;
-+#if (DPAA_VERSION >= 11)
-+ case (HEADER_TYPE_CAPWAP):
-+ p_Manip->opcode = HMAN_OC_CAPWAP_FRAGMENTATION;
-+ break;
-+#endif /* (DPAA_VERSION >= 11) */
-+ default:
-+ RETURN_ERROR(MAJOR, E_NOT_SUPPORTED,
-+ ("header for fragmentation"));
-+ }
-+ p_Manip->muramAllocate = TRUE;
-+ break;
-+ case e_FM_PCD_MANIP_SPECIAL_OFFLOAD:
-+ switch (p_ManipParams->u.specialOffload.type)
-+ {
-+ case (e_FM_PCD_MANIP_SPECIAL_OFFLOAD_IPSEC):
-+ p_Manip->opcode = HMAN_OC_IPSEC_MANIP;
-+ p_Manip->muramAllocate = TRUE;
-+ break;
-+#if (DPAA_VERSION >= 11)
-+ case (e_FM_PCD_MANIP_SPECIAL_OFFLOAD_CAPWAP):
-+ p_Manip->opcode = HMAN_OC_CAPWAP_MANIP;
-+ p_Manip->muramAllocate = TRUE;
-+ break;
-+#endif /* (DPAA_VERSION >= 11) */
-+ default:
-+ RETURN_ERROR(MAJOR, E_NOT_SUPPORTED,
-+ ("special offload type"));
-+ }
-+ break;
-+ default:
-+ RETURN_ERROR(MAJOR, E_NOT_SUPPORTED, ("manip type"));
-+ }
-+
-+ return E_OK;
-+}
-+#endif /* not (defined(FM_CAPWAP_SUPPORT) && (DPAA_VERSION == 10)) */
-+
-+#if (defined(FM_CAPWAP_SUPPORT) && (DPAA_VERSION == 10))
-+
-+static t_Error UpdateIndxStats(t_Handle h_FmPcd,
-+ t_Handle h_FmPort,
-+ t_FmPcdManip *p_Manip)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd *)h_FmPcd;
-+ uint32_t tmpReg32 = 0;
-+ t_AdOfTypeContLookup *p_Ad;
-+ t_FmPortGetSetCcParams fmPortGetSetCcParams;
-+ t_Error err;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Manip,E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Manip->h_Ad,E_INVALID_HANDLE);
-+
-+ p_Ad = (t_AdOfTypeContLookup *)p_Manip->h_Ad;
-+ if (p_Manip->h_FmPcd != h_FmPcd)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE,
-+ ("handler of PCD previously was initiated by different value"));
-+
-+ memset(&fmPortGetSetCcParams, 0, sizeof(t_FmPortGetSetCcParams));
-+
-+ if (!p_Manip->p_StatsTbl)
-+ {
-+
-+ fmPortGetSetCcParams.setCcParams.type = UPDATE_NIA_PNDN;
-+ fmPortGetSetCcParams.setCcParams.nia = NIA_FM_CTL_AC_CC;
-+ err = FmPortGetSetCcParams(h_FmPort, &fmPortGetSetCcParams);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+
-+ tmpReg32 = GET_UINT32(p_Ad->ccAdBase);
-+
-+ p_Manip->p_StatsTbl =
-+ (t_Handle)FM_MURAM_AllocMem(p_FmPcd->h_FmMuram,
-+ (uint32_t)p_Manip->owner * FM_PCD_MANIP_INDEXED_STATS_ENTRY_SIZE,
-+ 4);
-+ if (!p_Manip->p_StatsTbl)
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, ("MURAM alloc for Manipulation indexed statistics table"));
-+
-+ MemSet8(p_Manip->p_StatsTbl, 0, (uint32_t)(p_Manip->owner * 4));
-+
-+ tmpReg32 |= (uint32_t)(XX_VirtToPhys(p_Manip->p_StatsTbl) - p_FmPcd->physicalMuramBase);
-+
-+ if (p_Manip->cnia)
-+ tmpReg32 |= FM_PCD_MANIP_INDEXED_STATS_CNIA;
-+
-+ tmpReg32 |= FM_PCD_MANIP_INDEXED_STATS_DPD;
-+ WRITE_UINT32(p_Ad->ccAdBase, tmpReg32);
-+ }
-+ else
-+ {
-+ fmPortGetSetCcParams.setCcParams.type = UPDATE_NIA_PNDN;
-+ fmPortGetSetCcParams.setCcParams.nia = NIA_FM_CTL_AC_CC;
-+ err = FmPortGetSetCcParams(h_FmPort, &fmPortGetSetCcParams);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+
-+ return E_OK;
-+}
-+
-+static t_Error RmvHdrTillSpecLocNOrInsrtIntFrmHdr(t_FmPcdManipHdrRmvParams *p_ManipParams, t_FmPcdManip *p_Manip)
-+{
-+ t_AdOfTypeContLookup *p_Ad;
-+ uint32_t tmpReg32 = 0;
-+ uint8_t prsArrayOffset = 0;
-+ t_Error err;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Manip,E_NULL_POINTER);
-+ SANITY_CHECK_RETURN_ERROR(p_ManipParams,E_NULL_POINTER);
-+ SANITY_CHECK_RETURN_ERROR(p_Manip->h_Ad,E_INVALID_HANDLE);
-+
-+ p_Ad = (t_AdOfTypeContLookup *)p_Manip->h_Ad;
-+ if (p_Manip->rmv)
-+ {
-+ err = GetPrOffsetByHeaderOrField(&p_ManipParams->u.byHdr.u.fromStartByHdr.hdrInfo, &prsArrayOffset);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+
-+ tmpReg32 |= (uint32_t)prsArrayOffset << 24;
-+ tmpReg32 |= HMAN_RMV_HDR;
-+ }
-+
-+ if (p_Manip->insrt)
-+ tmpReg32 |= HMAN_INSRT_INT_FRM_HDR;
-+
-+ tmpReg32 |= (uint32_t)HMAN_OC_RMV_N_OR_INSRT_INT_FRM_HDR;
-+
-+ WRITE_UINT32(p_Ad->pcAndOffsets, tmpReg32);
-+
-+ tmpReg32 = 0;
-+ tmpReg32 |= FM_PCD_AD_CONT_LOOKUP_TYPE;
-+ WRITE_UINT32(p_Ad->ccAdBase, tmpReg32);
-+
-+ return E_OK;
-+}
-+
-+static t_Error MvIntFrameHeaderFromFrameToBufferPrefix(t_FmPcdManip *p_Manip,
-+ bool caamUsed)
-+{
-+ t_AdOfTypeContLookup *p_Ad = (t_AdOfTypeContLookup *)p_Manip->h_Ad;
-+ uint32_t tmpReg32 = 0;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Ad, E_INVALID_HANDLE);
-+
-+ p_Manip->updateParams |= OFFSET_OF_PR | INTERNAL_CONTEXT_OFFSET;
-+
-+ tmpReg32 = 0;
-+ tmpReg32 |= FM_PCD_AD_CONT_LOOKUP_TYPE;
-+ *(uint32_t *)&p_Ad->ccAdBase = tmpReg32;
-+
-+ tmpReg32 = 0;
-+ tmpReg32 |= HMAN_OC_MV_INT_FRAME_HDR_FROM_FRM_TO_BUFFER_PREFFIX;
-+ tmpReg32 |= (uint32_t)0x16 << 16;
-+ *(uint32_t *)&p_Ad->pcAndOffsets = tmpReg32;
-+
-+ if (caamUsed)
-+ *(uint32_t *)&p_Ad->gmask = 0xf0000000;
-+
-+ return E_OK;
-+}
-+
-+static t_Error CapwapRmvDtlsHdr(t_FmPcd *p_FmPcd, t_FmPcdManip *p_Manip)
-+{
-+ t_AdOfTypeContLookup *p_Ad;
-+ uint32_t tmpReg32 = 0;
-+ t_Error err = E_OK;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Manip->h_Ad,E_INVALID_HANDLE);
-+
-+ p_Ad = (t_AdOfTypeContLookup *)p_Manip->h_Ad;
-+
-+ tmpReg32 = 0;
-+ tmpReg32 |= (uint32_t)HMAN_OC_CAPWAP_RMV_DTLS_IF_EXIST;
-+ WRITE_UINT32(p_Ad->pcAndOffsets, tmpReg32);
-+
-+ tmpReg32 = 0;
-+ tmpReg32 |= FM_PCD_AD_CONT_LOOKUP_TYPE;
-+
-+
-+ if (p_Manip->h_Frag)
-+ {
-+ p_Manip->updateParams |= INTERNAL_CONTEXT_OFFSET;
-+ tmpReg32 |= (uint32_t)(XX_VirtToPhys(p_Manip->h_Frag) - (p_FmPcd->physicalMuramBase));
-+ }
-+
-+ WRITE_UINT32(p_Ad->ccAdBase, tmpReg32);
-+
-+ return err;
-+}
-+
-+static t_Error CapwapReassembly(t_CapwapReassemblyParams *p_ManipParams,
-+ t_FmPcdManip *p_Manip,
-+ t_FmPcd *p_FmPcd,
-+ uint8_t poolId)
-+{
-+ t_Handle p_Table;
-+ uint32_t tmpReg32 = 0;
-+ int i = 0;
-+ uint8_t log2Num;
-+ uint8_t numOfSets;
-+ uint32_t j = 0;
-+ uint32_t bitFor1Micro;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Manip->h_Ad, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd->h_Hc, E_INVALID_HANDLE);
-+
-+ if (!p_FmPcd->h_Hc)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE,("hc port has to be initialized in this mode"));
-+ if (!POWER_OF_2(p_ManipParams->timeoutRoutineRequestTime))
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("timeoutRoutineRequestTime has to be power of 2"));
-+ if (!POWER_OF_2(p_ManipParams->maxNumFramesInProcess))
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE,("maxNumFramesInProcess has to be power of 2"));
-+ if (!p_ManipParams->timeoutRoutineRequestTime && p_ManipParams->timeoutThresholdForReassmProcess)
-+ DBG(WARNING, ("if timeoutRoutineRequestTime 0, timeoutThresholdForReassmProcess is uselessly"));
-+ if (p_ManipParams->numOfFramesPerHashEntry == e_FM_PCD_MANIP_FOUR_WAYS_HASH)
-+ {
-+ if ((p_ManipParams->maxNumFramesInProcess < 4) ||
-+ (p_ManipParams->maxNumFramesInProcess > 512))
-+ RETURN_ERROR(MAJOR,E_INVALID_VALUE, ("In the case of numOfFramesPerHashEntry = e_FM_PCD_MANIP_EIGHT_WAYS_HASH maxNumFramesInProcess has to be in the range 4-512"));
-+ }
-+ else
-+ {
-+ if ((p_ManipParams->maxNumFramesInProcess < 8) ||
-+ (p_ManipParams->maxNumFramesInProcess > 2048))
-+ RETURN_ERROR(MAJOR,E_INVALID_VALUE, ("In the case of numOfFramesPerHashEntry = e_FM_PCD_MANIP_FOUR_WAYS_HASH maxNumFramesInProcess has to be in the range 8-2048"));
-+ }
-+
-+ bitFor1Micro = FmGetTimeStampScale(p_FmPcd->h_Fm);
-+ if (bitFor1Micro == 0)
-+ RETURN_ERROR(MAJOR, E_NOT_AVAILABLE, ("Timestamp scale"));
-+
-+ p_Manip->updateParams |= (NUM_OF_TASKS | OFFSET_OF_PR | OFFSET_OF_DATA | HW_PORT_ID);
-+
-+ p_Manip->h_Frag = (t_Handle)FM_MURAM_AllocMem(p_FmPcd->h_FmMuram,
-+ FM_PCD_MANIP_CAPWAP_REASM_TABLE_SIZE,
-+ FM_PCD_MANIP_CAPWAP_REASM_TABLE_ALIGN);
-+ if (!p_Manip->h_Frag)
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, ("MURAM alloc CAPWAP reassembly parameters table"));
-+
-+ MemSet8(p_Manip->h_Frag, 0, FM_PCD_MANIP_CAPWAP_REASM_TABLE_SIZE);
-+
-+ p_Table = (t_CapwapReasmPram *)p_Manip->h_Frag;
-+
-+ p_Manip->capwapFragParams.p_AutoLearnHashTbl =
-+ (t_Handle)FM_MURAM_AllocMem(p_FmPcd->h_FmMuram,
-+ (uint32_t)(p_ManipParams->maxNumFramesInProcess * 2 * FM_PCD_MANIP_CAPWAP_REASM_AUTO_LEARNING_HASH_ENTRY_SIZE),
-+ FM_PCD_MANIP_CAPWAP_REASM_TABLE_ALIGN);
-+
-+ if (!p_Manip->capwapFragParams.p_AutoLearnHashTbl)
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY,("MURAM alloc for CAPWAP automatic learning hash table"));
-+
-+ MemSet8(p_Manip->capwapFragParams.p_AutoLearnHashTbl, 0, (uint32_t)(p_ManipParams->maxNumFramesInProcess * 2 * FM_PCD_MANIP_CAPWAP_REASM_AUTO_LEARNING_HASH_ENTRY_SIZE));
-+
-+ tmpReg32 = (uint32_t)(XX_VirtToPhys(p_Manip->capwapFragParams.p_AutoLearnHashTbl) - p_FmPcd->physicalMuramBase);
-+
-+ WRITE_UINT32(((t_CapwapReasmPram *)p_Table)->autoLearnHashTblPtr, tmpReg32);
-+
-+ tmpReg32 = 0;
-+ if (p_ManipParams->timeOutMode == e_FM_PCD_MANIP_TIME_OUT_BETWEEN_FRAMES)
-+ tmpReg32 |= FM_PCD_MANIP_CAPWAP_REASM_TIME_OUT_BETWEEN_FRAMES;
-+ if (p_ManipParams->haltOnDuplicationFrag)
-+ tmpReg32 |= FM_PCD_MANIP_CAPWAP_REASM_HALT_ON_DUPLICATE_FRAG;
-+ if (p_ManipParams->numOfFramesPerHashEntry == e_FM_PCD_MANIP_EIGHT_WAYS_HASH)
-+ {
-+ i = 8;
-+ tmpReg32 |= FM_PCD_MANIP_CAPWAP_REASM_AUTOMATIC_LEARNIN_HASH_8_WAYS;
-+ }
-+ else
-+ i = 4;
-+
-+ numOfSets = (uint8_t)((p_ManipParams->maxNumFramesInProcess * 2) / i);
-+ LOG2(numOfSets, log2Num);
-+ tmpReg32 |= (uint32_t)(log2Num - 1) << 24;
-+
-+ WRITE_UINT32(((t_CapwapReasmPram *)p_Table)->mode, tmpReg32);
-+
-+ for (j=0; j<p_ManipParams->maxNumFramesInProcess*2; j++)
-+ if (((j / i) % 2)== 0)
-+ WRITE_UINT32(*(uint32_t *)PTR_MOVE(p_Manip->capwapFragParams.p_AutoLearnHashTbl, j * FM_PCD_MANIP_CAPWAP_REASM_AUTO_LEARNING_HASH_ENTRY_SIZE), 0x80000000);
-+
-+ tmpReg32 = 0x00008000;
-+ tmpReg32 |= (uint32_t)poolId << 16;
-+ WRITE_UINT32(((t_CapwapReasmPram *)p_Table)->bufferPoolIdAndRisc1SetIndexes, tmpReg32);
-+ WRITE_UINT32(((t_CapwapReasmPram *)p_Table)->risc23SetIndexes, 0x80008000);
-+ WRITE_UINT32(((t_CapwapReasmPram *)p_Table)->risc4SetIndexesAndExtendedStatsTblPtr, 0x80000000);
-+
-+ p_Manip->capwapFragParams.maxNumFramesInProcess = p_ManipParams->maxNumFramesInProcess;
-+
-+ p_Manip->capwapFragParams.sgBpid = poolId;
-+
-+ p_Manip->capwapFragParams.fqidForTimeOutFrames = p_ManipParams->fqidForTimeOutFrames;
-+ p_Manip->capwapFragParams.timeoutRoutineRequestTime = p_ManipParams->timeoutRoutineRequestTime;
-+ p_Manip->capwapFragParams.bitFor1Micro = bitFor1Micro;
-+
-+ tmpReg32 = 0;
-+ tmpReg32 |= (((uint32_t)1<<p_Manip->capwapFragParams.bitFor1Micro) * p_ManipParams->timeoutThresholdForReassmProcess);
-+ WRITE_UINT32(((t_CapwapReasmPram *)p_Table)->expirationDelay, tmpReg32);
-+
-+ return E_OK;
-+}
-+
-+static t_Error CapwapFragmentation(t_CapwapFragmentationParams *p_ManipParams,
-+ t_FmPcdManip *p_Manip,
-+ t_FmPcd *p_FmPcd,
-+ uint8_t poolId)
-+{
-+ t_AdOfTypeContLookup *p_Ad;
-+ uint32_t tmpReg32 = 0;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Manip->h_Ad,E_INVALID_HANDLE);
-+
-+ p_Manip->updateParams |= OFFSET_OF_DATA;
-+
-+ p_Manip->frag = TRUE;
-+
-+ p_Manip->h_Frag = (t_Handle)FM_MURAM_AllocMem(p_FmPcd->h_FmMuram,
-+ FM_PCD_CC_AD_ENTRY_SIZE,
-+ FM_PCD_CC_AD_TABLE_ALIGN);
-+ if (!p_Manip->h_Frag)
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, ("MURAM alloc for CAPWAP fragmentation table descriptor"));
-+
-+ MemSet8(p_Manip->h_Frag, 0, FM_PCD_CC_AD_ENTRY_SIZE);
-+
-+ p_Ad = (t_AdOfTypeContLookup *)p_Manip->h_Frag;
-+
-+ tmpReg32 = 0;
-+ tmpReg32 |= (uint32_t)HMAN_OC_CAPWAP_FRAGMENTATION;
-+
-+ if (p_ManipParams->headerOptionsCompr)
-+ tmpReg32 |= FM_PCD_MANIP_CAPWAP_FRAG_COMPR_OPTION_FIELD_EN;
-+ tmpReg32 |= ((uint32_t)poolId << 8);
-+ WRITE_UINT32(p_Ad->pcAndOffsets, tmpReg32);
-+
-+ tmpReg32 = 0;
-+ tmpReg32 |= FM_PCD_AD_CONT_LOOKUP_TYPE;
-+ WRITE_UINT32(p_Ad->ccAdBase, tmpReg32);
-+
-+ p_Manip->sizeForFragmentation = p_ManipParams->sizeForFragmentation;
-+ p_Manip->capwapFragParams.sgBpid = poolId;
-+
-+ return E_OK;
-+}
-+
-+static t_Error IndxStats(t_FmPcdStatsParams *p_StatsParams,t_FmPcdManip *p_Manip,t_FmPcd *p_FmPcd)
-+{
-+ t_AdOfTypeContLookup *p_Ad;
-+ uint32_t tmpReg32 = 0;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Manip->h_Ad,E_INVALID_HANDLE);
-+
-+ UNUSED(p_FmPcd);
-+
-+ p_Ad = (t_AdOfTypeContLookup *)p_Manip->h_Ad;
-+
-+ tmpReg32 = 0;
-+ tmpReg32 |= (uint32_t)HMAN_OC_CAPWAP_INDEXED_STATS;
-+ if (p_StatsParams->type == e_FM_PCD_STATS_PER_FLOWID)
-+ tmpReg32 |= (uint32_t)0x16 << 16;
-+ WRITE_UINT32(p_Ad->pcAndOffsets, tmpReg32);
-+
-+ tmpReg32 = 0;
-+ tmpReg32 |= FM_PCD_AD_CONT_LOOKUP_TYPE;
-+ WRITE_UINT32(p_Ad->ccAdBase, tmpReg32);
-+
-+ return E_OK;
-+}
-+
-+static t_Error InsrtHdrByTempl(t_FmPcdManipHdrInsrtParams *p_ManipParams, t_FmPcdManip *p_Manip, t_FmPcd *p_FmPcd)
-+{
-+ t_FmPcdManipHdrInsrtByTemplateParams *p_InsrtByTemplate = &p_ManipParams->u.byTemplate;
-+ uint8_t tmpReg8 = 0xff;
-+ t_AdOfTypeContLookup *p_Ad;
-+ bool ipModify = FALSE;
-+ uint32_t tmpReg32 = 0, tmpRegNia = 0;
-+ uint16_t tmpReg16 = 0;
-+ t_Error err = E_OK;
-+ uint8_t extraAddedBytes = 0, blockSize = 0, extraAddedBytesAlignedToBlockSize = 0, log2Num = 0;
-+ uint8_t *p_Template = NULL;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_ManipParams,E_NULL_POINTER);
-+ SANITY_CHECK_RETURN_ERROR(p_Manip,E_NULL_POINTER);
-+ SANITY_CHECK_RETURN_ERROR(p_Manip->h_Ad,E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd,E_NULL_POINTER);
-+
-+ p_Ad = (t_AdOfTypeContLookup *)p_Manip->h_Ad;
-+ if (p_Manip->insrt)
-+ {
-+ if ((!p_InsrtByTemplate->size && p_InsrtByTemplate->modifyOuterIp) ||
-+ (!p_InsrtByTemplate->size && p_InsrtByTemplate->modifyOuterVlan))
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("Inconsistent parameters : asking for header template modifications with no template for insertion (template size)"));
-+
-+ if (p_InsrtByTemplate->size && p_InsrtByTemplate->modifyOuterIp && (p_InsrtByTemplate->size <= p_InsrtByTemplate->modifyOuterIpParams.ipOuterOffset))
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("Inconsistent parameters : size of template < ipOuterOffset"));
-+
-+ if (p_InsrtByTemplate->size > 128)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("Size of header template for insertion can not be more than 128"));
-+
-+ if (p_InsrtByTemplate->size)
-+ {
-+ p_Manip->p_Template = (uint8_t *)FM_MURAM_AllocMem(p_FmPcd->h_FmMuram,
-+ p_InsrtByTemplate->size,
-+ FM_PCD_CC_AD_TABLE_ALIGN);
-+ if(!p_Manip->p_Template)
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, ("Memory allocation in MURAM FAILED"));
-+
-+ tmpReg32 = (uint32_t)(XX_VirtToPhys(p_Manip->p_Template) - (p_FmPcd->physicalMuramBase));
-+ tmpReg32 |= (uint32_t)p_InsrtByTemplate->size << 24;
-+ *(uint32_t *)&p_Ad->matchTblPtr = tmpReg32;
-+ }
-+
-+ tmpReg32 = 0;
-+
-+ p_Template = (uint8_t *)XX_Malloc(p_InsrtByTemplate->size * sizeof(uint8_t));
-+
-+ if (!p_Template)
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, ("XX_Malloc allocation FAILED"));
-+
-+ memcpy(p_Template, p_InsrtByTemplate->hdrTemplate, p_InsrtByTemplate->size * sizeof(uint8_t));
-+
-+ if (p_InsrtByTemplate->modifyOuterIp)
-+ {
-+ ipModify = TRUE;
-+
-+ tmpReg8 = (uint8_t)p_Template[p_InsrtByTemplate->modifyOuterIpParams.ipOuterOffset];
-+
-+ if((tmpReg8 & 0xf0) == 0x40)
-+ tmpReg8 = 4;
-+ else if((tmpReg8 & 0xf0) == 0x60)
-+ tmpReg8 = 6;
-+ else
-+ tmpReg8 = 0xff;
-+
-+ if (tmpReg8 != 0xff)
-+ {
-+ if(p_InsrtByTemplate->modifyOuterIpParams.dscpEcn & 0xff00)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("Inconsistent parameters : IPV4 present in header template, dscpEcn has to be only 1 byte"));
-+ if(p_InsrtByTemplate->modifyOuterIpParams.recalculateLength)
-+ {
-+
-+ if((p_InsrtByTemplate->modifyOuterIpParams.recalculateLengthParams.extraBytesAddedAlignedToBlockSize + p_InsrtByTemplate->modifyOuterIpParams.recalculateLengthParams.extraBytesAddedNotAlignedToBlockSize) > 255)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("extra Byte added can not be more than 256 bytes"));
-+ extraAddedBytes = (uint8_t) (p_InsrtByTemplate->modifyOuterIpParams.recalculateLengthParams.extraBytesAddedAlignedToBlockSize + p_InsrtByTemplate->modifyOuterIpParams.recalculateLengthParams.extraBytesAddedNotAlignedToBlockSize);
-+ blockSize = p_InsrtByTemplate->modifyOuterIpParams.recalculateLengthParams.blockSize;
-+ extraAddedBytesAlignedToBlockSize = p_InsrtByTemplate->modifyOuterIpParams.recalculateLengthParams.extraBytesAddedAlignedToBlockSize;
-+ /*IP header template - IP totalLength -
-+ (1 byte) extraByteForIp = headerTemplateSize - ipOffset + insertedBytesAfterThisStage ,
-+ in the case of SEC insertedBytesAfterThisStage - SEC trailer (21/31) + header(13)
-+ second byte - extraByteForIp = headerTemplate - ipOffset + insertedBytesAfterThisStage*/
-+ }
-+ if (blockSize)
-+ {
-+ if (!POWER_OF_2(blockSize))
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("inputFrmPaddingUpToBlockSize has to be power of 2"));
-+ }
-+
-+ }
-+ if (tmpReg8 == 4)
-+ {
-+ if ((IPv4_HDRCHECKSUM_FIELD_OFFSET_FROM_IP + p_InsrtByTemplate->modifyOuterIpParams.ipOuterOffset) > p_InsrtByTemplate->size)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("Inconsistent parameters : IP present in header template, user asked for IP modifications but ipOffset + ipTotalLengthFieldOffset in header template bigger than template size"));
-+
-+ p_Template[p_InsrtByTemplate->modifyOuterIpParams.ipOuterOffset + IPv4_DSCECN_FIELD_OFFSET_FROM_IP] = (uint8_t)p_InsrtByTemplate->modifyOuterIpParams.dscpEcn;
-+
-+ if (blockSize)
-+ blockSize -= 1;
-+
-+ if ((p_InsrtByTemplate->size - p_InsrtByTemplate->modifyOuterIpParams.ipOuterOffset + extraAddedBytes) > 255)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("p_InsrtByTemplate->size - p_InsrtByTemplate->modifyOuterIpParams.ipOuterOffset + extraAddedBytes has to be less than 255"));
-+
-+ p_Template[p_InsrtByTemplate->modifyOuterIpParams.ipOuterOffset + IPv4_TOTALLENGTH_FIELD_OFFSET_FROM_IP + 1] = blockSize; // IPV6 - in AD instead of SEQ IND
-+ p_Template[p_InsrtByTemplate->modifyOuterIpParams.ipOuterOffset + IPv4_TOTALLENGTH_FIELD_OFFSET_FROM_IP] = (uint8_t)(p_InsrtByTemplate->size - p_InsrtByTemplate->modifyOuterIpParams.ipOuterOffset + extraAddedBytes);// for IPV6 decrement additional 40 bytes of IPV6 heade size
-+
-+ p_Template[p_InsrtByTemplate->modifyOuterIpParams.ipOuterOffset + IPv4_ID_FIELD_OFFSET_FROM_IP] = 0x00;
-+ p_Template[p_InsrtByTemplate->modifyOuterIpParams.ipOuterOffset + IPv4_ID_FIELD_OFFSET_FROM_IP + 1] = extraAddedBytesAlignedToBlockSize;
-+
-+ /*IP header template - relevant only for ipv4 CheckSum = 0*/
-+ p_Template[p_InsrtByTemplate->modifyOuterIpParams.ipOuterOffset + IPv4_HDRCHECKSUM_FIELD_OFFSET_FROM_IP] = 0x00;
-+ p_Template[p_InsrtByTemplate->modifyOuterIpParams.ipOuterOffset + IPv4_HDRCHECKSUM_FIELD_OFFSET_FROM_IP + 1] = 0x00;
-+
-+ /*UDP checksum has to be 0*/
-+ if (p_InsrtByTemplate->modifyOuterIpParams.udpPresent)
-+ {
-+ if ((p_InsrtByTemplate->modifyOuterIpParams.udpOffset + UDP_CHECKSUM_FIELD_OFFSET_FROM_UDP + UDP_CHECKSUM_FIELD_SIZE) > p_InsrtByTemplate->size)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("Inconsistent parameters : UDP present according to user but (UDP offset + UDP header size) < size of header template"));
-+
-+ p_Template[p_InsrtByTemplate->modifyOuterIpParams.udpOffset + UDP_CHECKSUM_FIELD_OFFSET_FROM_UDP ] = 0x00;
-+ p_Template[p_InsrtByTemplate->modifyOuterIpParams.udpOffset + UDP_CHECKSUM_FIELD_OFFSET_FROM_UDP + 1] = 0x00;
-+
-+ }
-+
-+ if (p_InsrtByTemplate->modifyOuterIpParams.ipIdentGenId > 7)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("ipIdentGenId has to be one out of 8 sequence number generators (0 - 7) for IP identification field"));
-+
-+ tmpRegNia |= (uint32_t)p_InsrtByTemplate->modifyOuterIpParams.ipIdentGenId<<24;
-+ }
-+ else if (tmpReg8 == 6)
-+ {
-+ /*TODO - add check for maximum value of blockSize;*/
-+ if (blockSize)
-+ LOG2(blockSize, log2Num);
-+ tmpRegNia |= (uint32_t)log2Num << 24;
-+
-+ // for IPV6 decrement additional 40 bytes of IPV6 heade size - because IPV6 header size is not included in payloadLength
-+ p_Template[p_InsrtByTemplate->modifyOuterIpParams.ipOuterOffset + IPv6_PAYLOAD_LENGTH_OFFSET_FROM_IP] = (uint8_t)(p_InsrtByTemplate->size - p_InsrtByTemplate->modifyOuterIpParams.ipOuterOffset + extraAddedBytes - 40);
-+ p_Template[p_InsrtByTemplate->modifyOuterIpParams.ipOuterOffset + IPv6_PAYLOAD_LENGTH_OFFSET_FROM_IP + 1] = extraAddedBytesAlignedToBlockSize;
-+ if (p_InsrtByTemplate->modifyOuterIpParams.udpPresent)
-+ {
-+ if ((p_InsrtByTemplate->modifyOuterIpParams.udpOffset + UDP_CHECKSUM_FIELD_OFFSET_FROM_UDP + UDP_CHECKSUM_FIELD_SIZE) > p_InsrtByTemplate->size)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("Inconsistent parameters : UDP present according to user but (UDP offset + UDP header size) < size of header template"));
-+ if (p_Template[p_InsrtByTemplate->modifyOuterIpParams.ipOuterOffset + IPv6_NEXT_HEADER_OFFSET_FROM_IP] != 0x88)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("OUr suppport is only IPv6/UDPLite"));
-+ p_Template[p_InsrtByTemplate->modifyOuterIpParams.udpOffset + UDP_LENGTH_FIELD_OFFSET_FROM_UDP] = 0x00;
-+ p_Template[p_InsrtByTemplate->modifyOuterIpParams.udpOffset + UDP_LENGTH_FIELD_OFFSET_FROM_UDP + 1] = 0x08;
-+ p_Template[p_InsrtByTemplate->modifyOuterIpParams.udpOffset + UDP_CHECKSUM_FIELD_OFFSET_FROM_UDP] = 0x00;
-+ p_Template[p_InsrtByTemplate->modifyOuterIpParams.udpOffset + UDP_CHECKSUM_FIELD_OFFSET_FROM_UDP + 1] = 0x00;
-+ }
-+ }
-+ else
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("IP version supported only IPV4"));
-+ }
-+
-+ tmpReg32 = tmpReg16 = tmpReg8 = 0;
-+ /*TODO - check it*/
-+ if (p_InsrtByTemplate->modifyOuterVlan)
-+ {
-+ if (p_InsrtByTemplate->modifyOuterVlanParams.vpri & ~0x07)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE,("Inconsistent parameters : user asked for VLAN modifications but VPRI more than 3 bits"));
-+
-+ memcpy(&tmpReg16, &p_Template[VLAN_TAG_FIELD_OFFSET_FROM_ETH], 2*(sizeof(uint8_t)));
-+ if ((tmpReg16 != 0x9100) && (tmpReg16!= 0x9200) && (tmpReg16 != 0x8100))
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE,("Inconsistent parameters : user asked for VLAN modifications but Tag Protocol identifier is not VLAN "));
-+
-+ memcpy(&tmpReg8, &p_Template[14],1*(sizeof(uint8_t)));
-+ tmpReg8 &= 0x1f;
-+ tmpReg8 |= (uint8_t)(p_InsrtByTemplate->modifyOuterVlanParams.vpri << 5);
-+
-+ p_Template[14] = tmpReg8;
-+ }
-+
-+ MemCpy8(p_Manip->p_Template, p_Template, p_InsrtByTemplate->size);
-+
-+ XX_Free(p_Template);
-+ }
-+
-+ tmpReg32 = 0;
-+ if (p_Manip->h_Frag)
-+ {
-+ tmpRegNia |= (uint32_t)(XX_VirtToPhys(p_Manip->h_Frag) - (p_FmPcd->physicalMuramBase));
-+ tmpReg32 |= (uint32_t)p_Manip->sizeForFragmentation << 16;
-+ }
-+ else
-+ tmpReg32 = 0xffff0000;
-+
-+ if (ipModify)
-+ tmpReg32 |= (uint32_t)p_InsrtByTemplate->modifyOuterIpParams.ipOuterOffset << 8;
-+ else
-+ tmpReg32 |= (uint32_t)0x0000ff00;
-+
-+ tmpReg32 |= (uint32_t)HMAN_OC_INSRT_HDR_BY_TEMPL_N_OR_FRAG_AFTER;
-+ *(uint32_t *)&p_Ad->pcAndOffsets = tmpReg32;
-+
-+ tmpRegNia |= FM_PCD_AD_CONT_LOOKUP_TYPE;
-+ *(uint32_t *)&p_Ad->ccAdBase = tmpRegNia;
-+
-+ return err;
-+}
-+
-+static t_Error CheckStatsParamsAndSetType(t_FmPcdManip *p_Manip, t_FmPcdStatsParams *p_StatsParams)
-+{
-+
-+ switch (p_StatsParams->type)
-+ {
-+ case (e_FM_PCD_STATS_PER_FLOWID):
-+ p_Manip->opcode = HMAN_OC_CAPWAP_INDEXED_STATS;
-+ p_Manip->muramAllocate = TRUE;
-+ break;
-+ default:
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("Unsupported statistics type"));
-+ }
-+
-+ return E_OK;
-+}
-+#endif /* (defined(FM_CAPWAP_SUPPORT) && (DPAA_VERSION == 10)) */
-+
-+static t_Error FillReassmManipParams(t_FmPcdManip *p_Manip, e_NetHeaderType hdr)
-+{
-+ t_AdOfTypeContLookup *p_Ad;
-+ t_FmPcd *p_FmPcd = (t_FmPcd *)p_Manip->h_FmPcd;
-+ uint32_t tmpReg32;
-+ t_Error err = E_OK;
-+
-+ /* Creates the Reassembly Parameters table. It contains parameters that are specific to either the IPv4 reassembly
-+ function or to the IPv6 reassembly function. If both IPv4 reassembly and IPv6 reassembly are required, then
-+ two separate IP Reassembly Parameter tables are required.*/
-+ if ((err = CreateReassTable(p_Manip, hdr)) != E_OK)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+
-+ /* Sets the first Ad register (ccAdBase) - Action Descriptor Type and Pointer to the Reassembly Parameters Table offset from MURAM*/
-+ tmpReg32 = 0;
-+ tmpReg32 |= FM_PCD_AD_CONT_LOOKUP_TYPE;
-+
-+ /* Gets the required Action descriptor table pointer */
-+ switch (hdr)
-+ {
-+ case HEADER_TYPE_IPv4:
-+ p_Ad = (t_AdOfTypeContLookup *)p_Manip->reassmParams.ip.h_Ipv4Ad;
-+ tmpReg32 |= (uint32_t)(XX_VirtToPhys(
-+ p_Manip->reassmParams.ip.p_Ipv4ReassTbl)
-+ - (p_FmPcd->physicalMuramBase));
-+ break;
-+ case HEADER_TYPE_IPv6:
-+ p_Ad = (t_AdOfTypeContLookup *)p_Manip->reassmParams.ip.h_Ipv6Ad;
-+ tmpReg32 |= (uint32_t)(XX_VirtToPhys(
-+ p_Manip->reassmParams.ip.p_Ipv6ReassTbl)
-+ - (p_FmPcd->physicalMuramBase));
-+ break;
-+ case HEADER_TYPE_CAPWAP:
-+ p_Ad = (t_AdOfTypeContLookup *)p_Manip->reassmParams.capwap.h_Ad;
-+ tmpReg32 |= (uint32_t)(XX_VirtToPhys(
-+ p_Manip->reassmParams.capwap.p_ReassTbl)
-+ - (p_FmPcd->physicalMuramBase));
-+ break;
-+ default:
-+ RETURN_ERROR(MAJOR, E_NOT_SUPPORTED, ("header type"));
-+ }
-+
-+ WRITE_UINT32(p_Ad->ccAdBase, tmpReg32);
-+
-+ /* Sets the second Ad register (matchTblPtr) - Buffer pool ID (BPID for V2) and Scatter/Gather table offset*/
-+ /* mark the Scatter/Gather table offset to be set later on when the port will be known */
-+ p_Manip->updateParams = (NUM_OF_TASKS | NUM_OF_EXTRA_TASKS | DISCARD_MASK);
-+
-+ if ((hdr == HEADER_TYPE_IPv6) || (hdr == HEADER_TYPE_IPv4))
-+ {
-+#if (DPAA_VERSION == 10)
-+ tmpReg32 = (uint32_t)(p_Manip->reassmParams.sgBpid << 8);
-+ WRITE_UINT32(p_Ad->matchTblPtr, tmpReg32);
-+#endif /* (DPAA_VERSION == 10) */
-+#if (DPAA_VERSION >= 11)
-+ if (p_Manip->reassmParams.ip.nonConsistentSpFqid != 0)
-+ {
-+ tmpReg32 = FM_PCD_AD_NCSPFQIDM_MASK
-+ | (uint32_t)(p_Manip->reassmParams.ip.nonConsistentSpFqid);
-+ WRITE_UINT32(p_Ad->gmask, tmpReg32);
-+ }
-+#endif /* (DPAA_VERSION >= 11) */
-+ /* Sets the third Ad register (pcAndOffsets)- IP Reassemble Operation Code*/
-+ tmpReg32 = 0;
-+ tmpReg32 |= (uint32_t)HMAN_OC_IP_REASSEMBLY;
-+ }
-+#if (DPAA_VERSION >= 11)
-+ else
-+ if (hdr == HEADER_TYPE_CAPWAP)
-+ {
-+ tmpReg32 = 0;
-+ tmpReg32 |= (uint32_t)HMAN_OC_CAPWAP_REASSEMBLY;
-+ }
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+ WRITE_UINT32(p_Ad->pcAndOffsets, tmpReg32);
-+
-+ p_Manip->reassm = TRUE;
-+
-+ return E_OK;
-+}
-+
-+static t_Error SetIpv4ReassmManip(t_FmPcdManip *p_Manip)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd *)p_Manip->h_FmPcd;
-+
-+ /* Allocation if IPv4 Action descriptor */
-+ p_Manip->reassmParams.ip.h_Ipv4Ad = (t_Handle)XX_MallocSmart(
-+ FM_PCD_CC_AD_ENTRY_SIZE, p_Manip->reassmParams.dataMemId,
-+ FM_PCD_CC_AD_TABLE_ALIGN);
-+ if (!p_Manip->reassmParams.ip.h_Ipv4Ad)
-+ {
-+ ReleaseManipHandler(p_Manip, p_FmPcd);
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY,
-+ ("Allocation of IPv4 table descriptor"));
-+ }
-+
-+ memset(p_Manip->reassmParams.ip.h_Ipv4Ad, 0, FM_PCD_CC_AD_ENTRY_SIZE);
-+
-+ /* Fill reassembly manipulation parameter in the IP Reassembly Action Descriptor */
-+ return FillReassmManipParams(p_Manip, HEADER_TYPE_IPv4);
-+}
-+
-+static t_Error SetIpv6ReassmManip(t_FmPcdManip *p_Manip)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd *)p_Manip->h_FmPcd;
-+
-+ /* Allocation if IPv6 Action descriptor */
-+ p_Manip->reassmParams.ip.h_Ipv6Ad = (t_Handle)XX_MallocSmart(
-+ FM_PCD_CC_AD_ENTRY_SIZE, p_Manip->reassmParams.dataMemId,
-+ FM_PCD_CC_AD_TABLE_ALIGN);
-+ if (!p_Manip->reassmParams.ip.h_Ipv6Ad)
-+ {
-+ ReleaseManipHandler(p_Manip, p_FmPcd);
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY,
-+ ("Allocation of IPv6 table descriptor"));
-+ }
-+
-+ memset(p_Manip->reassmParams.ip.h_Ipv6Ad, 0, FM_PCD_CC_AD_ENTRY_SIZE);
-+
-+ /* Fill reassembly manipulation parameter in the IP Reassembly Action Descriptor */
-+ return FillReassmManipParams(p_Manip, HEADER_TYPE_IPv6);
-+}
-+
-+static t_Error IpReassembly(t_FmPcdManipReassemParams *p_ManipReassmParams,
-+ t_FmPcdManip *p_Manip)
-+{
-+ uint32_t maxSetNumber = 10000;
-+ t_FmPcdManipReassemIpParams reassmManipParams =
-+ p_ManipReassmParams->u.ipReassem;
-+ t_Error res;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Manip->h_FmPcd, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(((t_FmPcd *)p_Manip->h_FmPcd)->h_Hc,
-+ E_INVALID_HANDLE);
-+
-+ /* Check validation of user's parameter.*/
-+ if ((reassmManipParams.timeoutThresholdForReassmProcess < 1000)
-+ || (reassmManipParams.timeoutThresholdForReassmProcess > 8000000))
-+ RETURN_ERROR(
-+ MAJOR, E_INVALID_VALUE,
-+ ("timeoutThresholdForReassmProcess should be 1msec - 8sec"));
-+ /* It is recommended that the total number of entries in this table (number of sets * number of ways)
-+ will be twice the number of frames that are expected to be reassembled simultaneously.*/
-+ if (reassmManipParams.maxNumFramesInProcess
-+ > (reassmManipParams.maxNumFramesInProcess * maxSetNumber / 2))
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("maxNumFramesInProcess has to be less than (maximun set number * number of ways / 2)"));
-+
-+ if ((p_ManipReassmParams->hdr == HEADER_TYPE_IPv6)
-+ && (reassmManipParams.minFragSize[1] < 256))
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("minFragSize[1] must be >= 256"));
-+
-+ /* Saves user's reassembly manipulation parameters */
-+ p_Manip->reassmParams.ip.relativeSchemeId[0] =
-+ reassmManipParams.relativeSchemeId[0];
-+ p_Manip->reassmParams.ip.relativeSchemeId[1] =
-+ reassmManipParams.relativeSchemeId[1];
-+ p_Manip->reassmParams.ip.numOfFramesPerHashEntry[0] =
-+ reassmManipParams.numOfFramesPerHashEntry[0];
-+ p_Manip->reassmParams.ip.numOfFramesPerHashEntry[1] =
-+ reassmManipParams.numOfFramesPerHashEntry[1];
-+ p_Manip->reassmParams.ip.minFragSize[0] = reassmManipParams.minFragSize[0];
-+ p_Manip->reassmParams.ip.minFragSize[1] = reassmManipParams.minFragSize[1];
-+ p_Manip->reassmParams.maxNumFramesInProcess =
-+ reassmManipParams.maxNumFramesInProcess;
-+ p_Manip->reassmParams.timeOutMode = reassmManipParams.timeOutMode;
-+ p_Manip->reassmParams.fqidForTimeOutFrames =
-+ reassmManipParams.fqidForTimeOutFrames;
-+ p_Manip->reassmParams.timeoutThresholdForReassmProcess =
-+ reassmManipParams.timeoutThresholdForReassmProcess;
-+ p_Manip->reassmParams.dataMemId = reassmManipParams.dataMemId;
-+ p_Manip->reassmParams.dataLiodnOffset = reassmManipParams.dataLiodnOffset;
-+#if (DPAA_VERSION == 10)
-+ p_Manip->reassmParams.sgBpid = reassmManipParams.sgBpid;
-+#endif /* (DPAA_VERSION == 10) */
-+#if (DPAA_VERSION >= 11)
-+ if (reassmManipParams.nonConsistentSpFqid != 0)
-+ {
-+ p_Manip->reassmParams.ip.nonConsistentSpFqid =
-+ reassmManipParams.nonConsistentSpFqid;
-+ }
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+ /* Creates and initializes the IP Reassembly common parameter table */
-+ CreateReassCommonTable(p_Manip);
-+
-+ /* Creation of IPv4 reassembly manipulation */
-+ if ((p_Manip->reassmParams.hdr == HEADER_TYPE_IPv6)
-+ || (p_Manip->reassmParams.hdr == HEADER_TYPE_IPv4))
-+ {
-+ res = SetIpv4ReassmManip(p_Manip);
-+ if (res != E_OK)
-+ return res;
-+ }
-+
-+ /* Creation of IPv6 reassembly manipulation */
-+ if (p_Manip->reassmParams.hdr == HEADER_TYPE_IPv6)
-+ {
-+ res = SetIpv6ReassmManip(p_Manip);
-+ if (res != E_OK)
-+ return res;
-+ }
-+
-+ return E_OK;
-+}
-+
-+static void setIpReassmSchemeParams(t_FmPcd* p_FmPcd,
-+ t_FmPcdKgSchemeParams *p_Scheme,
-+ t_Handle h_CcTree, bool ipv4,
-+ uint8_t groupId)
-+{
-+ uint32_t j;
-+ uint8_t res;
-+
-+ /* Configures scheme's network environment parameters */
-+ p_Scheme->netEnvParams.numOfDistinctionUnits = 2;
-+ if (ipv4)
-+ res = FmPcdNetEnvGetUnitId(
-+ p_FmPcd, FmPcdGetNetEnvId(p_Scheme->netEnvParams.h_NetEnv),
-+ HEADER_TYPE_IPv4, FALSE, 0);
-+ else
-+ res = FmPcdNetEnvGetUnitId(
-+ p_FmPcd, FmPcdGetNetEnvId(p_Scheme->netEnvParams.h_NetEnv),
-+ HEADER_TYPE_IPv6, FALSE, 0);
-+ ASSERT_COND(res != FM_PCD_MAX_NUM_OF_DISTINCTION_UNITS);
-+ p_Scheme->netEnvParams.unitIds[0] = res;
-+
-+ res = FmPcdNetEnvGetUnitId(
-+ p_FmPcd, FmPcdGetNetEnvId(p_Scheme->netEnvParams.h_NetEnv),
-+ HEADER_TYPE_USER_DEFINED_SHIM2, FALSE, 0);
-+ ASSERT_COND(res != FM_PCD_MAX_NUM_OF_DISTINCTION_UNITS);
-+ p_Scheme->netEnvParams.unitIds[1] = res;
-+
-+ /* Configures scheme's next engine parameters*/
-+ p_Scheme->nextEngine = e_FM_PCD_CC;
-+ p_Scheme->kgNextEngineParams.cc.h_CcTree = h_CcTree;
-+ p_Scheme->kgNextEngineParams.cc.grpId = groupId;
-+ p_Scheme->useHash = TRUE;
-+
-+ /* Configures scheme's key*/
-+ if (ipv4 == TRUE)
-+ {
-+ p_Scheme->keyExtractAndHashParams.numOfUsedExtracts = 4;
-+ p_Scheme->keyExtractAndHashParams.extractArray[0].type =
-+ e_FM_PCD_EXTRACT_BY_HDR;
-+ p_Scheme->keyExtractAndHashParams.extractArray[0].extractByHdr.type =
-+ e_FM_PCD_EXTRACT_FULL_FIELD;
-+ p_Scheme->keyExtractAndHashParams.extractArray[0].extractByHdr.hdr =
-+ HEADER_TYPE_IPv4;
-+ p_Scheme->keyExtractAndHashParams.extractArray[0].extractByHdr.extractByHdrType.fullField.ipv4 =
-+ NET_HEADER_FIELD_IPv4_DST_IP;
-+ p_Scheme->keyExtractAndHashParams.extractArray[1].type =
-+ e_FM_PCD_EXTRACT_BY_HDR;
-+ p_Scheme->keyExtractAndHashParams.extractArray[1].extractByHdr.type =
-+ e_FM_PCD_EXTRACT_FULL_FIELD;
-+ p_Scheme->keyExtractAndHashParams.extractArray[1].extractByHdr.hdr =
-+ HEADER_TYPE_IPv4;
-+ p_Scheme->keyExtractAndHashParams.extractArray[1].extractByHdr.extractByHdrType.fullField.ipv4 =
-+ NET_HEADER_FIELD_IPv4_SRC_IP;
-+ p_Scheme->keyExtractAndHashParams.extractArray[2].type =
-+ e_FM_PCD_EXTRACT_BY_HDR;
-+ p_Scheme->keyExtractAndHashParams.extractArray[2].extractByHdr.type =
-+ e_FM_PCD_EXTRACT_FULL_FIELD;
-+ p_Scheme->keyExtractAndHashParams.extractArray[2].extractByHdr.hdr =
-+ HEADER_TYPE_IPv4;
-+ p_Scheme->keyExtractAndHashParams.extractArray[2].extractByHdr.extractByHdrType.fullField.ipv4 =
-+ NET_HEADER_FIELD_IPv4_PROTO;
-+ p_Scheme->keyExtractAndHashParams.extractArray[3].type =
-+ e_FM_PCD_EXTRACT_BY_HDR;
-+ p_Scheme->keyExtractAndHashParams.extractArray[3].extractByHdr.hdr =
-+ HEADER_TYPE_IPv4;
-+ p_Scheme->keyExtractAndHashParams.extractArray[3].extractByHdr.type =
-+ e_FM_PCD_EXTRACT_FROM_HDR;
-+ p_Scheme->keyExtractAndHashParams.extractArray[3].extractByHdr.ignoreProtocolValidation =
-+ FALSE;
-+ p_Scheme->keyExtractAndHashParams.extractArray[3].extractByHdr.extractByHdrType.fromHdr.size =
-+ 2;
-+ p_Scheme->keyExtractAndHashParams.extractArray[3].extractByHdr.extractByHdrType.fromHdr.offset =
-+ 4;
-+ }
-+ else /* IPv6 */
-+ {
-+ p_Scheme->keyExtractAndHashParams.numOfUsedExtracts = 3;
-+ p_Scheme->keyExtractAndHashParams.extractArray[0].type =
-+ e_FM_PCD_EXTRACT_BY_HDR;
-+ p_Scheme->keyExtractAndHashParams.extractArray[0].extractByHdr.type =
-+ e_FM_PCD_EXTRACT_FULL_FIELD;
-+ p_Scheme->keyExtractAndHashParams.extractArray[0].extractByHdr.hdr =
-+ HEADER_TYPE_IPv6;
-+ p_Scheme->keyExtractAndHashParams.extractArray[0].extractByHdr.extractByHdrType.fullField.ipv6 =
-+ NET_HEADER_FIELD_IPv6_DST_IP;
-+ p_Scheme->keyExtractAndHashParams.extractArray[1].type =
-+ e_FM_PCD_EXTRACT_BY_HDR;
-+ p_Scheme->keyExtractAndHashParams.extractArray[1].extractByHdr.type =
-+ e_FM_PCD_EXTRACT_FULL_FIELD;
-+ p_Scheme->keyExtractAndHashParams.extractArray[1].extractByHdr.hdr =
-+ HEADER_TYPE_IPv6;
-+ p_Scheme->keyExtractAndHashParams.extractArray[1].extractByHdr.extractByHdrType.fullField.ipv6 =
-+ NET_HEADER_FIELD_IPv6_SRC_IP;
-+ p_Scheme->keyExtractAndHashParams.extractArray[2].type =
-+ e_FM_PCD_EXTRACT_BY_HDR;
-+ p_Scheme->keyExtractAndHashParams.extractArray[2].extractByHdr.hdr =
-+ HEADER_TYPE_USER_DEFINED_SHIM2;
-+ p_Scheme->keyExtractAndHashParams.extractArray[2].extractByHdr.type =
-+ e_FM_PCD_EXTRACT_FROM_HDR;
-+ p_Scheme->keyExtractAndHashParams.extractArray[2].extractByHdr.extractByHdrType.fromHdr.size =
-+ 4;
-+ p_Scheme->keyExtractAndHashParams.extractArray[2].extractByHdr.extractByHdrType.fromHdr.offset =
-+ 4;
-+ p_Scheme->keyExtractAndHashParams.extractArray[2].extractByHdr.ignoreProtocolValidation =
-+ TRUE;
-+ }
-+
-+ p_Scheme->keyExtractAndHashParams.privateDflt0 = 0x01020304;
-+ p_Scheme->keyExtractAndHashParams.privateDflt1 = 0x11121314;
-+ p_Scheme->keyExtractAndHashParams.numOfUsedDflts =
-+ FM_PCD_KG_NUM_OF_DEFAULT_GROUPS;
-+ for (j = 0; j < FM_PCD_KG_NUM_OF_DEFAULT_GROUPS; j++)
-+ {
-+ p_Scheme->keyExtractAndHashParams.dflts[j].type =
-+ (e_FmPcdKgKnownFieldsDfltTypes)j; /* all types */
-+ p_Scheme->keyExtractAndHashParams.dflts[j].dfltSelect =
-+ e_FM_PCD_KG_DFLT_GBL_0;
-+ }
-+}
-+
-+static t_Error IpReassemblyStats(t_FmPcdManip *p_Manip,
-+ t_FmPcdManipReassemIpStats *p_Stats)
-+{
-+ ASSERT_COND(p_Manip);
-+ ASSERT_COND(p_Stats);
-+ ASSERT_COND(p_Manip->reassmParams.p_ReassCommonTbl);
-+
-+ p_Stats->timeout =
-+ GET_UINT32(p_Manip->reassmParams.p_ReassCommonTbl->totalTimeOutCounter);
-+ p_Stats->rfdPoolBusy =
-+ GET_UINT32(p_Manip->reassmParams.p_ReassCommonTbl->totalRfdPoolBusyCounter);
-+ p_Stats->internalBufferBusy =
-+ GET_UINT32(p_Manip->reassmParams.p_ReassCommonTbl->totalInternalBufferBusy);
-+ p_Stats->externalBufferBusy =
-+ GET_UINT32(p_Manip->reassmParams.p_ReassCommonTbl->totalExternalBufferBusy);
-+ p_Stats->sgFragments =
-+ GET_UINT32(p_Manip->reassmParams.p_ReassCommonTbl->totalSgFragmentCounter);
-+ p_Stats->dmaSemaphoreDepletion =
-+ GET_UINT32(p_Manip->reassmParams.p_ReassCommonTbl->totalDmaSemaphoreDepletionCounter);
-+#if (DPAA_VERSION >= 11)
-+ p_Stats->nonConsistentSp =
-+ GET_UINT32(p_Manip->reassmParams.p_ReassCommonTbl->totalNCSPCounter);
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+ if (p_Manip->reassmParams.ip.p_Ipv4ReassTbl)
-+ {
-+ p_Stats->specificHdrStatistics[0].successfullyReassembled =
-+ GET_UINT32(p_Manip->reassmParams.ip.p_Ipv4ReassTbl->totalSuccessfullyReasmFramesCounter);
-+ p_Stats->specificHdrStatistics[0].validFragments =
-+ GET_UINT32(p_Manip->reassmParams.ip.p_Ipv4ReassTbl->totalValidFragmentCounter);
-+ p_Stats->specificHdrStatistics[0].processedFragments =
-+ GET_UINT32(p_Manip->reassmParams.ip.p_Ipv4ReassTbl->totalProcessedFragCounter);
-+ p_Stats->specificHdrStatistics[0].malformedFragments =
-+ GET_UINT32(p_Manip->reassmParams.ip.p_Ipv4ReassTbl->totalMalformdFragCounter);
-+ p_Stats->specificHdrStatistics[0].autoLearnBusy =
-+ GET_UINT32(p_Manip->reassmParams.ip.p_Ipv4ReassTbl->totalSetBusyCounter);
-+ p_Stats->specificHdrStatistics[0].discardedFragments =
-+ GET_UINT32(p_Manip->reassmParams.ip.p_Ipv4ReassTbl->totalDiscardedFragsCounter);
-+ p_Stats->specificHdrStatistics[0].moreThan16Fragments =
-+ GET_UINT32(p_Manip->reassmParams.ip.p_Ipv4ReassTbl->totalMoreThan16FramesCounter);
-+ }
-+ if (p_Manip->reassmParams.ip.p_Ipv6ReassTbl)
-+ {
-+ p_Stats->specificHdrStatistics[1].successfullyReassembled =
-+ GET_UINT32(p_Manip->reassmParams.ip.p_Ipv6ReassTbl->totalSuccessfullyReasmFramesCounter);
-+ p_Stats->specificHdrStatistics[1].validFragments =
-+ GET_UINT32(p_Manip->reassmParams.ip.p_Ipv6ReassTbl->totalValidFragmentCounter);
-+ p_Stats->specificHdrStatistics[1].processedFragments =
-+ GET_UINT32(p_Manip->reassmParams.ip.p_Ipv6ReassTbl->totalProcessedFragCounter);
-+ p_Stats->specificHdrStatistics[1].malformedFragments =
-+ GET_UINT32(p_Manip->reassmParams.ip.p_Ipv6ReassTbl->totalMalformdFragCounter);
-+ p_Stats->specificHdrStatistics[1].autoLearnBusy =
-+ GET_UINT32(p_Manip->reassmParams.ip.p_Ipv6ReassTbl->totalSetBusyCounter);
-+ p_Stats->specificHdrStatistics[1].discardedFragments =
-+ GET_UINT32(p_Manip->reassmParams.ip.p_Ipv6ReassTbl->totalDiscardedFragsCounter);
-+ p_Stats->specificHdrStatistics[1].moreThan16Fragments =
-+ GET_UINT32(p_Manip->reassmParams.ip.p_Ipv6ReassTbl->totalMoreThan16FramesCounter);
-+ }
-+ return E_OK;
-+}
-+
-+static t_Error IpFragmentationStats(t_FmPcdManip *p_Manip,
-+ t_FmPcdManipFragIpStats *p_Stats)
-+{
-+ t_AdOfTypeContLookup *p_Ad;
-+
-+ ASSERT_COND(p_Manip);
-+ ASSERT_COND(p_Stats);
-+ ASSERT_COND(p_Manip->h_Ad);
-+ ASSERT_COND(p_Manip->fragParams.p_Frag);
-+
-+ p_Ad = (t_AdOfTypeContLookup *)p_Manip->h_Ad;
-+
-+ p_Stats->totalFrames = GET_UINT32(p_Ad->gmask);
-+ p_Stats->fragmentedFrames = GET_UINT32(p_Manip->fragParams.p_Frag->ccAdBase)
-+ & 0x00ffffff;
-+ p_Stats->generatedFragments =
-+ GET_UINT32(p_Manip->fragParams.p_Frag->matchTblPtr);
-+
-+ return E_OK;
-+}
-+
-+static t_Error IpFragmentation(t_FmPcdManipFragIpParams *p_ManipParams,
-+ t_FmPcdManip *p_Manip)
-+{
-+ uint32_t pcAndOffsetsReg = 0, ccAdBaseReg = 0, gmaskReg = 0;
-+ t_FmPcd *p_FmPcd;
-+#if (DPAA_VERSION == 10)
-+ t_Error err = E_OK;
-+#endif /* (DPAA_VERSION == 10) */
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Manip->h_Ad, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_ManipParams->sizeForFragmentation != 0xFFFF,
-+ E_INVALID_VALUE);
-+
-+ p_FmPcd = p_Manip->h_FmPcd;
-+ /* Allocation of fragmentation Action Descriptor */
-+ p_Manip->fragParams.p_Frag = (t_AdOfTypeContLookup *)FM_MURAM_AllocMem(
-+ p_FmPcd->h_FmMuram, FM_PCD_CC_AD_ENTRY_SIZE,
-+ FM_PCD_CC_AD_TABLE_ALIGN);
-+ if (!p_Manip->fragParams.p_Frag)
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY,
-+ ("MURAM alloc for Fragmentation table descriptor"));
-+ MemSet8(p_Manip->fragParams.p_Frag, 0, FM_PCD_CC_AD_ENTRY_SIZE);
-+
-+ /* Prepare the third Ad register (pcAndOffsets)- OperationCode */
-+ pcAndOffsetsReg = (uint32_t)HMAN_OC_IP_FRAGMENTATION;
-+
-+ /* Prepare the first Ad register (ccAdBase) - Don't frag action and Action descriptor type*/
-+ ccAdBaseReg = FM_PCD_AD_CONT_LOOKUP_TYPE;
-+ ccAdBaseReg |= (p_ManipParams->dontFragAction
-+ << FM_PCD_MANIP_IP_FRAG_DF_SHIFT);
-+
-+
-+ /* Set Scatter/Gather BPid */
-+ if (p_ManipParams->sgBpidEn)
-+ {
-+ ccAdBaseReg |= FM_PCD_MANIP_IP_FRAG_SG_BDID_EN;
-+ pcAndOffsetsReg |= ((p_ManipParams->sgBpid
-+ << FM_PCD_MANIP_IP_FRAG_SG_BDID_SHIFT)
-+ & FM_PCD_MANIP_IP_FRAG_SG_BDID_MASK);
-+ }
-+
-+ /* Prepare the first Ad register (gmask) - scratch buffer pool id and Pointer to fragment ID */
-+ gmaskReg = (uint32_t)(XX_VirtToPhys(UINT_TO_PTR(p_FmPcd->ipv6FrameIdAddr))
-+ - p_FmPcd->physicalMuramBase);
-+#if (DPAA_VERSION == 10)
-+ gmaskReg |= p_ManipParams->scratchBpid << FM_PCD_MANIP_IP_FRAG_SCRATCH_BPID;
-+#else
-+ gmaskReg |= (0xFF) << FM_PCD_MANIP_IP_FRAG_SCRATCH_BPID;
-+#endif /* (DPAA_VERSION == 10) */
-+
-+ /* Set all Ad registers */
-+ WRITE_UINT32(p_Manip->fragParams.p_Frag->pcAndOffsets, pcAndOffsetsReg);
-+ WRITE_UINT32(p_Manip->fragParams.p_Frag->ccAdBase, ccAdBaseReg);
-+ WRITE_UINT32(p_Manip->fragParams.p_Frag->gmask, gmaskReg);
-+
-+ /* Saves user's fragmentation manipulation parameters */
-+ p_Manip->frag = TRUE;
-+ p_Manip->sizeForFragmentation = p_ManipParams->sizeForFragmentation;
-+
-+#if (DPAA_VERSION == 10)
-+ p_Manip->fragParams.scratchBpid = p_ManipParams->scratchBpid;
-+
-+ /* scratch buffer pool initialization */
-+ if ((err = FmPcdFragHcScratchPoolFill((t_Handle)p_FmPcd, p_ManipParams->scratchBpid)) != E_OK)
-+ {
-+ FM_MURAM_FreeMem(p_FmPcd->h_FmMuram, p_Manip->fragParams.p_Frag);
-+ p_Manip->fragParams.p_Frag = NULL;
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+#endif /* (DPAA_VERSION == 10) */
-+
-+ return E_OK;
-+}
-+
-+static t_Error IPManip(t_FmPcdManip *p_Manip)
-+{
-+ t_Error err = E_OK;
-+ t_FmPcd *p_FmPcd;
-+ t_AdOfTypeContLookup *p_Ad;
-+ uint32_t tmpReg32 = 0, tmpRegNia = 0;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Manip, E_INVALID_HANDLE);
-+ p_FmPcd = p_Manip->h_FmPcd;
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd, E_INVALID_HANDLE);
-+
-+ p_Ad = (t_AdOfTypeContLookup *)p_Manip->h_Ad;
-+
-+ tmpReg32 = FM_PCD_MANIP_IP_NO_FRAGMENTATION;
-+ if (p_Manip->frag == TRUE)
-+ {
-+ tmpRegNia = (uint32_t)(XX_VirtToPhys(p_Manip->fragParams.p_Frag)
-+ - (p_FmPcd->physicalMuramBase));
-+ tmpReg32 = (uint32_t)p_Manip->sizeForFragmentation
-+ << FM_PCD_MANIP_IP_MTU_SHIFT;
-+ }
-+
-+ tmpRegNia |= FM_PCD_AD_CONT_LOOKUP_TYPE;
-+ tmpReg32 |= HMAN_OC_IP_MANIP;
-+
-+#if (DPAA_VERSION >= 11)
-+ tmpRegNia |= FM_PCD_MANIP_IP_CNIA;
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+ WRITE_UINT32(p_Ad->pcAndOffsets, tmpReg32);
-+ WRITE_UINT32(p_Ad->ccAdBase, tmpRegNia);
-+ WRITE_UINT32(p_Ad->gmask, 0);
-+ /* Total frame counter - MUST be initialized to zero.*/
-+
-+ return err;
-+}
-+
-+static t_Error UpdateInitIpFrag(t_Handle h_FmPcd, t_Handle h_PcdParams,
-+ t_Handle h_FmPort, t_FmPcdManip *p_Manip,
-+ t_Handle h_Ad, bool validate)
-+{
-+ t_FmPortGetSetCcParams fmPortGetSetCcParams;
-+ t_Error err;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Manip, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR((p_Manip->opcode == HMAN_OC_IP_FRAGMENTATION),
-+ E_INVALID_STATE);
-+ SANITY_CHECK_RETURN_ERROR(h_FmPcd, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(h_FmPort, E_INVALID_HANDLE);
-+
-+ UNUSED(h_FmPcd);
-+ UNUSED(h_Ad);
-+ UNUSED(h_PcdParams);
-+ UNUSED(validate);
-+ UNUSED(p_Manip);
-+
-+ fmPortGetSetCcParams.setCcParams.type = 0;
-+ fmPortGetSetCcParams.getCcParams.type = MANIP_EXTRA_SPACE;
-+ if ((err = FmPortGetSetCcParams(h_FmPort, &fmPortGetSetCcParams)) != E_OK)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+
-+ if (!fmPortGetSetCcParams.getCcParams.internalBufferOffset)
-+ DBG(WARNING, ("manipExtraSpace must be larger than '0'"));
-+
-+ return E_OK;
-+}
-+
-+static t_Error IPSecManip(t_FmPcdManipParams *p_ManipParams,
-+ t_FmPcdManip *p_Manip)
-+{
-+ t_AdOfTypeContLookup *p_Ad;
-+ t_FmPcdManipSpecialOffloadIPSecParams *p_IPSecParams;
-+ t_Error err = E_OK;
-+ uint32_t tmpReg32 = 0;
-+ uint32_t power;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Manip, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_ManipParams, E_INVALID_HANDLE);
-+
-+ p_IPSecParams = &p_ManipParams->u.specialOffload.u.ipsec;
-+
-+ SANITY_CHECK_RETURN_ERROR(
-+ !p_IPSecParams->variableIpHdrLen || p_IPSecParams->decryption,
-+ E_INVALID_VALUE);
-+ SANITY_CHECK_RETURN_ERROR(
-+ !p_IPSecParams->variableIpVersion || !p_IPSecParams->decryption,
-+ E_INVALID_VALUE);
-+ SANITY_CHECK_RETURN_ERROR(
-+ !p_IPSecParams->variableIpVersion || p_IPSecParams->outerIPHdrLen,
-+ E_INVALID_VALUE);
-+ SANITY_CHECK_RETURN_ERROR(
-+ !p_IPSecParams->arwSize || p_IPSecParams->arwAddr,
-+ E_INVALID_VALUE);
-+ SANITY_CHECK_RETURN_ERROR(
-+ !p_IPSecParams->arwSize || p_IPSecParams->decryption,
-+ E_INVALID_VALUE);
-+ SANITY_CHECK_RETURN_ERROR((p_IPSecParams->arwSize % 16) == 0, E_INVALID_VALUE);
-+
-+ p_Ad = (t_AdOfTypeContLookup *)p_Manip->h_Ad;
-+
-+ tmpReg32 |= FM_PCD_AD_CONT_LOOKUP_TYPE;
-+ tmpReg32 |= (p_IPSecParams->decryption) ? FM_PCD_MANIP_IPSEC_DEC : 0;
-+ tmpReg32 |= (p_IPSecParams->ecnCopy) ? FM_PCD_MANIP_IPSEC_ECN_EN : 0;
-+ tmpReg32 |= (p_IPSecParams->dscpCopy) ? FM_PCD_MANIP_IPSEC_DSCP_EN : 0;
-+ tmpReg32 |=
-+ (p_IPSecParams->variableIpHdrLen) ? FM_PCD_MANIP_IPSEC_VIPL_EN : 0;
-+ tmpReg32 |=
-+ (p_IPSecParams->variableIpVersion) ? FM_PCD_MANIP_IPSEC_VIPV_EN : 0;
-+ if (p_IPSecParams->arwSize)
-+ tmpReg32 |= (uint32_t)((XX_VirtToPhys(UINT_TO_PTR(p_IPSecParams->arwAddr))-FM_MM_MURAM)
-+ & (FM_MURAM_SIZE-1));
-+ WRITE_UINT32(p_Ad->ccAdBase, tmpReg32);
-+
-+ tmpReg32 = 0;
-+ if (p_IPSecParams->arwSize) {
-+ NEXT_POWER_OF_2((p_IPSecParams->arwSize + 32), power);
-+ LOG2(power, power);
-+ tmpReg32 = (p_IPSecParams->arwSize | (power - 5)) << FM_PCD_MANIP_IPSEC_ARW_SIZE_SHIFT;
-+ }
-+
-+ if (p_ManipParams->h_NextManip)
-+ tmpReg32 |=
-+ (uint32_t)(XX_VirtToPhys(((t_FmPcdManip *)p_ManipParams->h_NextManip)->h_Ad)-
-+ (((t_FmPcd *)p_Manip->h_FmPcd)->physicalMuramBase)) >> 4;
-+ WRITE_UINT32(p_Ad->matchTblPtr, tmpReg32);
-+
-+ tmpReg32 = HMAN_OC_IPSEC_MANIP;
-+ tmpReg32 |= p_IPSecParams->outerIPHdrLen
-+ << FM_PCD_MANIP_IPSEC_IP_HDR_LEN_SHIFT;
-+ if (p_ManipParams->h_NextManip)
-+ tmpReg32 |= FM_PCD_MANIP_IPSEC_NADEN;
-+ WRITE_UINT32(p_Ad->pcAndOffsets, tmpReg32);
-+
-+ return err;
-+}
-+
-+static t_Error SetCapwapReassmManip(t_FmPcdManip *p_Manip)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd *)p_Manip->h_FmPcd;
-+
-+ /* Allocation if CAPWAP Action descriptor */
-+ p_Manip->reassmParams.capwap.h_Ad = (t_Handle)XX_MallocSmart(
-+ FM_PCD_CC_AD_ENTRY_SIZE, p_Manip->reassmParams.dataMemId,
-+ FM_PCD_CC_AD_TABLE_ALIGN);
-+ if (!p_Manip->reassmParams.capwap.h_Ad)
-+ {
-+ ReleaseManipHandler(p_Manip, p_FmPcd);
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY,
-+ ("Allocation of CAPWAP table descriptor"));
-+ }
-+
-+ memset(p_Manip->reassmParams.capwap.h_Ad, 0, FM_PCD_CC_AD_ENTRY_SIZE);
-+
-+ /* Fill reassembly manipulation parameter in the Reassembly Action Descriptor */
-+ return FillReassmManipParams(p_Manip, HEADER_TYPE_CAPWAP);
-+}
-+
-+static void setCapwapReassmSchemeParams(t_FmPcd* p_FmPcd,
-+ t_FmPcdKgSchemeParams *p_Scheme,
-+ t_Handle h_CcTree, uint8_t groupId)
-+{
-+ uint8_t res;
-+
-+ /* Configures scheme's network environment parameters */
-+ p_Scheme->netEnvParams.numOfDistinctionUnits = 1;
-+ res = FmPcdNetEnvGetUnitId(
-+ p_FmPcd, FmPcdGetNetEnvId(p_Scheme->netEnvParams.h_NetEnv),
-+ HEADER_TYPE_USER_DEFINED_SHIM2, FALSE, 0);
-+ ASSERT_COND(res != FM_PCD_MAX_NUM_OF_DISTINCTION_UNITS);
-+ p_Scheme->netEnvParams.unitIds[0] = res;
-+
-+ /* Configures scheme's next engine parameters*/
-+ p_Scheme->nextEngine = e_FM_PCD_CC;
-+ p_Scheme->kgNextEngineParams.cc.h_CcTree = h_CcTree;
-+ p_Scheme->kgNextEngineParams.cc.grpId = groupId;
-+ p_Scheme->useHash = TRUE;
-+
-+ /* Configures scheme's key*/
-+ p_Scheme->keyExtractAndHashParams.numOfUsedExtracts = 2;
-+ p_Scheme->keyExtractAndHashParams.extractArray[0].type =
-+ e_FM_PCD_EXTRACT_NON_HDR;
-+ p_Scheme->keyExtractAndHashParams.extractArray[0].extractNonHdr.src =
-+ e_FM_PCD_EXTRACT_FROM_PARSE_RESULT;
-+ p_Scheme->keyExtractAndHashParams.extractArray[0].extractNonHdr.action =
-+ e_FM_PCD_ACTION_NONE;
-+ p_Scheme->keyExtractAndHashParams.extractArray[0].extractNonHdr.offset = 20;
-+ p_Scheme->keyExtractAndHashParams.extractArray[0].extractNonHdr.size = 4;
-+ p_Scheme->keyExtractAndHashParams.extractArray[1].type =
-+ e_FM_PCD_EXTRACT_NON_HDR;
-+ p_Scheme->keyExtractAndHashParams.extractArray[1].extractNonHdr.src =
-+ e_FM_PCD_EXTRACT_FROM_DFLT_VALUE;
-+ p_Scheme->keyExtractAndHashParams.extractArray[1].extractNonHdr.action =
-+ e_FM_PCD_ACTION_NONE;
-+ p_Scheme->keyExtractAndHashParams.extractArray[1].extractNonHdr.offset = 0;
-+ p_Scheme->keyExtractAndHashParams.extractArray[1].extractNonHdr.size = 1;
-+
-+ p_Scheme->keyExtractAndHashParams.privateDflt0 = 0x0;
-+ p_Scheme->keyExtractAndHashParams.privateDflt1 = 0x0;
-+ p_Scheme->keyExtractAndHashParams.numOfUsedDflts = 1;
-+ p_Scheme->keyExtractAndHashParams.dflts[0].type = e_FM_PCD_KG_GENERIC_NOT_FROM_DATA;
-+ p_Scheme->keyExtractAndHashParams.dflts[0].dfltSelect = e_FM_PCD_KG_DFLT_PRIVATE_0;
-+}
-+
-+#if (DPAA_VERSION >= 11)
-+static t_Error CapwapReassemblyStats(t_FmPcdManip *p_Manip,
-+ t_FmPcdManipReassemCapwapStats *p_Stats)
-+{
-+ ASSERT_COND(p_Manip);
-+ ASSERT_COND(p_Stats);
-+ ASSERT_COND(p_Manip->reassmParams.p_ReassCommonTbl);
-+
-+ p_Stats->timeout =
-+ GET_UINT32(p_Manip->reassmParams.p_ReassCommonTbl->totalTimeOutCounter);
-+ p_Stats->rfdPoolBusy =
-+ GET_UINT32(p_Manip->reassmParams.p_ReassCommonTbl->totalRfdPoolBusyCounter);
-+ p_Stats->internalBufferBusy =
-+ GET_UINT32(p_Manip->reassmParams.p_ReassCommonTbl->totalInternalBufferBusy);
-+ p_Stats->externalBufferBusy =
-+ GET_UINT32(p_Manip->reassmParams.p_ReassCommonTbl->totalExternalBufferBusy);
-+ p_Stats->sgFragments =
-+ GET_UINT32(p_Manip->reassmParams.p_ReassCommonTbl->totalSgFragmentCounter);
-+ p_Stats->dmaSemaphoreDepletion =
-+ GET_UINT32(p_Manip->reassmParams.p_ReassCommonTbl->totalDmaSemaphoreDepletionCounter);
-+ p_Stats->exceedMaxReassemblyFrameLen =
-+ GET_UINT32(p_Manip->reassmParams.p_ReassCommonTbl->totalNCSPCounter);
-+
-+ p_Stats->successfullyReassembled =
-+ GET_UINT32(p_Manip->reassmParams.capwap.p_ReassTbl->totalSuccessfullyReasmFramesCounter);
-+ p_Stats->validFragments =
-+ GET_UINT32(p_Manip->reassmParams.capwap.p_ReassTbl->totalValidFragmentCounter);
-+ p_Stats->processedFragments =
-+ GET_UINT32(p_Manip->reassmParams.capwap.p_ReassTbl->totalProcessedFragCounter);
-+ p_Stats->malformedFragments =
-+ GET_UINT32(p_Manip->reassmParams.capwap.p_ReassTbl->totalMalformdFragCounter);
-+ p_Stats->autoLearnBusy =
-+ GET_UINT32(p_Manip->reassmParams.capwap.p_ReassTbl->totalSetBusyCounter);
-+ p_Stats->discardedFragments =
-+ GET_UINT32(p_Manip->reassmParams.capwap.p_ReassTbl->totalDiscardedFragsCounter);
-+ p_Stats->moreThan16Fragments =
-+ GET_UINT32(p_Manip->reassmParams.capwap.p_ReassTbl->totalMoreThan16FramesCounter);
-+
-+ return E_OK;
-+}
-+
-+static t_Error CapwapFragmentationStats(t_FmPcdManip *p_Manip,
-+ t_FmPcdManipFragCapwapStats *p_Stats)
-+{
-+ t_AdOfTypeContLookup *p_Ad;
-+
-+ ASSERT_COND(p_Manip);
-+ ASSERT_COND(p_Stats);
-+ ASSERT_COND(p_Manip->h_Ad);
-+ ASSERT_COND(p_Manip->fragParams.p_Frag);
-+
-+ p_Ad = (t_AdOfTypeContLookup *)p_Manip->h_Ad;
-+
-+ p_Stats->totalFrames = GET_UINT32(p_Ad->gmask);
-+
-+ return E_OK;
-+}
-+
-+static t_Error CapwapReassembly(t_FmPcdManipReassemParams *p_ManipReassmParams,
-+ t_FmPcdManip *p_Manip)
-+{
-+ uint32_t maxSetNumber = 10000;
-+ t_FmPcdManipReassemCapwapParams reassmManipParams =
-+ p_ManipReassmParams->u.capwapReassem;
-+ t_Error res;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Manip->h_FmPcd, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(((t_FmPcd *)p_Manip->h_FmPcd)->h_Hc,
-+ E_INVALID_HANDLE);
-+
-+ /* Check validation of user's parameter.*/
-+ if ((reassmManipParams.timeoutThresholdForReassmProcess < 1000)
-+ || (reassmManipParams.timeoutThresholdForReassmProcess > 8000000))
-+ RETURN_ERROR(
-+ MAJOR, E_INVALID_VALUE,
-+ ("timeoutThresholdForReassmProcess should be 1msec - 8sec"));
-+ /* It is recommended that the total number of entries in this table (number of sets * number of ways)
-+ will be twice the number of frames that are expected to be reassembled simultaneously.*/
-+ if (reassmManipParams.maxNumFramesInProcess
-+ > (reassmManipParams.maxNumFramesInProcess * maxSetNumber / 2))
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("maxNumFramesInProcess has to be less than (maximun set number * number of ways / 2)"));
-+
-+ /* Saves user's reassembly manipulation parameters */
-+ p_Manip->reassmParams.capwap.relativeSchemeId =
-+ reassmManipParams.relativeSchemeId;
-+ p_Manip->reassmParams.capwap.numOfFramesPerHashEntry =
-+ reassmManipParams.numOfFramesPerHashEntry;
-+ p_Manip->reassmParams.capwap.maxRessembledsSize =
-+ reassmManipParams.maxReassembledFrameLength;
-+ p_Manip->reassmParams.maxNumFramesInProcess =
-+ reassmManipParams.maxNumFramesInProcess;
-+ p_Manip->reassmParams.timeOutMode = reassmManipParams.timeOutMode;
-+ p_Manip->reassmParams.fqidForTimeOutFrames =
-+ reassmManipParams.fqidForTimeOutFrames;
-+ p_Manip->reassmParams.timeoutThresholdForReassmProcess =
-+ reassmManipParams.timeoutThresholdForReassmProcess;
-+ p_Manip->reassmParams.dataMemId = reassmManipParams.dataMemId;
-+ p_Manip->reassmParams.dataLiodnOffset = reassmManipParams.dataLiodnOffset;
-+
-+ /* Creates and initializes the Reassembly common parameter table */
-+ CreateReassCommonTable(p_Manip);
-+
-+ res = SetCapwapReassmManip(p_Manip);
-+ if (res != E_OK)
-+ return res;
-+
-+ return E_OK;
-+}
-+
-+static t_Error CapwapFragmentation(t_FmPcdManipFragCapwapParams *p_ManipParams,
-+ t_FmPcdManip *p_Manip)
-+{
-+ t_FmPcd *p_FmPcd;
-+ t_AdOfTypeContLookup *p_Ad;
-+ uint32_t pcAndOffsetsReg = 0, ccAdBaseReg = 0, gmaskReg = 0;
-+ uint32_t tmpReg32 = 0, tmpRegNia = 0;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Manip->h_Ad, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_ManipParams->sizeForFragmentation != 0xFFFF,
-+ E_INVALID_VALUE);
-+ p_FmPcd = p_Manip->h_FmPcd;
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd, E_INVALID_HANDLE);
-+
-+ /* Allocation of fragmentation Action Descriptor */
-+ p_Manip->fragParams.p_Frag = (t_AdOfTypeContLookup *)FM_MURAM_AllocMem(
-+ p_FmPcd->h_FmMuram, FM_PCD_CC_AD_ENTRY_SIZE,
-+ FM_PCD_CC_AD_TABLE_ALIGN);
-+ if (!p_Manip->fragParams.p_Frag)
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY,
-+ ("MURAM alloc for Fragmentation table descriptor"));
-+ MemSet8(p_Manip->fragParams.p_Frag, 0, FM_PCD_CC_AD_ENTRY_SIZE);
-+
-+ /* Prepare the third Ad register (pcAndOffsets)- OperationCode */
-+ pcAndOffsetsReg = (uint32_t)HMAN_OC_CAPWAP_FRAGMENTATION;
-+
-+ /* Prepare the first Ad register (ccAdBase) - Don't frag action and Action descriptor type*/
-+ ccAdBaseReg = FM_PCD_AD_CONT_LOOKUP_TYPE;
-+ ccAdBaseReg |=
-+ (p_ManipParams->compressModeEn) ? FM_PCD_MANIP_CAPWAP_FRAG_COMPRESS_EN :
-+ 0;
-+
-+ /* Set Scatter/Gather BPid */
-+ if (p_ManipParams->sgBpidEn)
-+ {
-+ ccAdBaseReg |= FM_PCD_MANIP_CAPWAP_FRAG_SG_BDID_EN;
-+ pcAndOffsetsReg |= ((p_ManipParams->sgBpid
-+ << FM_PCD_MANIP_CAPWAP_FRAG_SG_BDID_SHIFT)
-+ & FM_PCD_MANIP_CAPWAP_FRAG_SG_BDID_MASK);
-+ }
-+
-+ /* Prepare the first Ad register (gmask) - scratch buffer pool id and Pointer to fragment ID */
-+ gmaskReg = (uint32_t)(XX_VirtToPhys(UINT_TO_PTR(p_FmPcd->capwapFrameIdAddr))
-+ - p_FmPcd->physicalMuramBase);
-+ gmaskReg |= (0xFF) << FM_PCD_MANIP_IP_FRAG_SCRATCH_BPID;
-+
-+ /* Set all Ad registers */
-+ WRITE_UINT32(p_Manip->fragParams.p_Frag->pcAndOffsets, pcAndOffsetsReg);
-+ WRITE_UINT32(p_Manip->fragParams.p_Frag->ccAdBase, ccAdBaseReg);
-+ WRITE_UINT32(p_Manip->fragParams.p_Frag->gmask, gmaskReg);
-+
-+ /* Saves user's fragmentation manipulation parameters */
-+ p_Manip->frag = TRUE;
-+ p_Manip->sizeForFragmentation = p_ManipParams->sizeForFragmentation;
-+
-+ p_Ad = (t_AdOfTypeContLookup *)p_Manip->h_Ad;
-+
-+ tmpRegNia = (uint32_t)(XX_VirtToPhys(p_Manip->fragParams.p_Frag)
-+ - (p_FmPcd->physicalMuramBase));
-+ tmpReg32 = (uint32_t)p_Manip->sizeForFragmentation
-+ << FM_PCD_MANIP_CAPWAP_FRAG_CHECK_MTU_SHIFT;
-+
-+ tmpRegNia |= FM_PCD_AD_CONT_LOOKUP_TYPE;
-+ tmpReg32 |= HMAN_OC_CAPWAP_FRAG_CHECK;
-+
-+ tmpRegNia |= FM_PCD_MANIP_CAPWAP_FRAG_CHECK_CNIA;
-+
-+ WRITE_UINT32(p_Ad->pcAndOffsets, tmpReg32);
-+ WRITE_UINT32(p_Ad->ccAdBase, tmpRegNia);
-+ WRITE_UINT32(p_Ad->gmask, 0);
-+ /* Total frame counter - MUST be initialized to zero.*/
-+
-+ return E_OK;
-+}
-+
-+static t_Error UpdateInitCapwapFrag(t_Handle h_FmPcd, t_Handle h_PcdParams,
-+ t_Handle h_FmPort, t_FmPcdManip *p_Manip,
-+ t_Handle h_Ad, bool validate)
-+{
-+ t_FmPortGetSetCcParams fmPortGetSetCcParams;
-+ t_Error err;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Manip, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR((p_Manip->opcode == HMAN_OC_CAPWAP_FRAGMENTATION),
-+ E_INVALID_STATE);
-+ SANITY_CHECK_RETURN_ERROR(h_FmPcd, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(h_FmPort, E_INVALID_HANDLE);
-+
-+ UNUSED(h_FmPcd);
-+ UNUSED(h_Ad);
-+ UNUSED(h_PcdParams);
-+ UNUSED(validate);
-+ UNUSED(p_Manip);
-+
-+ fmPortGetSetCcParams.setCcParams.type = 0;
-+ fmPortGetSetCcParams.getCcParams.type = MANIP_EXTRA_SPACE;
-+ if ((err = FmPortGetSetCcParams(h_FmPort, &fmPortGetSetCcParams)) != E_OK)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+
-+ if (!fmPortGetSetCcParams.getCcParams.internalBufferOffset)
-+ DBG(WARNING, ("manipExtraSpace must be larger than '0'"));
-+
-+ return E_OK;
-+}
-+
-+static t_Error CapwapManip(t_FmPcdManipParams *p_ManipParams,
-+ t_FmPcdManip *p_Manip)
-+{
-+ t_AdOfTypeContLookup *p_Ad;
-+ t_FmPcdManipSpecialOffloadCapwapParams *p_Params;
-+ t_Error err = E_OK;
-+ uint32_t tmpReg32 = 0;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Manip, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_ManipParams, E_INVALID_HANDLE);
-+
-+ p_Params = &p_ManipParams->u.specialOffload.u.capwap;
-+
-+ p_Ad = (t_AdOfTypeContLookup *)p_Manip->h_Ad;
-+ tmpReg32 |= FM_PCD_AD_CONT_LOOKUP_TYPE;
-+ tmpReg32 |= (p_Params->dtls) ? FM_PCD_MANIP_CAPWAP_DTLS : 0;
-+ /* TODO - add 'qosSrc' */
-+ WRITE_UINT32(p_Ad->ccAdBase, tmpReg32);
-+
-+ tmpReg32 = HMAN_OC_CAPWAP_MANIP;
-+ if (p_ManipParams->h_NextManip)
-+ {
-+ WRITE_UINT32(
-+ p_Ad->matchTblPtr,
-+ (uint32_t)(XX_VirtToPhys(((t_FmPcdManip *)p_ManipParams->h_NextManip)->h_Ad)- (((t_FmPcd *)p_Manip->h_FmPcd)->physicalMuramBase)) >> 4);
-+
-+ tmpReg32 |= FM_PCD_MANIP_CAPWAP_NADEN;
-+ }
-+
-+ WRITE_UINT32(p_Ad->pcAndOffsets, tmpReg32);
-+
-+ return err;
-+}
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+static t_Handle ManipOrStatsSetNode(t_Handle h_FmPcd, t_Handle *p_Params,
-+ bool stats)
-+{
-+ t_FmPcdManip *p_Manip;
-+ t_Error err;
-+ t_FmPcd *p_FmPcd = (t_FmPcd *)h_FmPcd;
-+
-+ p_Manip = (t_FmPcdManip*)XX_Malloc(sizeof(t_FmPcdManip));
-+ if (!p_Manip)
-+ {
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("No memory"));
-+ return NULL;
-+ }
-+ memset(p_Manip, 0, sizeof(t_FmPcdManip));
-+
-+ p_Manip->type = ((t_FmPcdManipParams *)p_Params)->type;
-+ memcpy((uint8_t*)&p_Manip->manipParams, p_Params,
-+ sizeof(p_Manip->manipParams));
-+
-+ if (!stats)
-+ err = CheckManipParamsAndSetType(p_Manip,
-+ (t_FmPcdManipParams *)p_Params);
-+#if (defined(FM_CAPWAP_SUPPORT) && (DPAA_VERSION == 10))
-+ else
-+ err = CheckStatsParamsAndSetType(p_Manip, (t_FmPcdStatsParams *)p_Params);
-+#else /* not (defined(FM_CAPWAP_SUPPORT) && (DPAA_VERSION == 10)) */
-+ else
-+ {
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Statistics node!"));
-+ XX_Free(p_Manip);
-+ return NULL;
-+ }
-+#endif /* (defined(FM_CAPWAP_SUPPORT) && (DPAA_VERSION == 10)) */
-+ if (err)
-+ {
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE, ("Invalid header manipulation type"));
-+ XX_Free(p_Manip);
-+ return NULL;
-+ }
-+
-+ if ((p_Manip->opcode != HMAN_OC_IP_REASSEMBLY) && (p_Manip->opcode != HMAN_OC_CAPWAP_REASSEMBLY))
-+ {
-+ /* In Case of reassembly manipulation the reassembly action descriptor will
-+ be defines later on */
-+ if (p_Manip->muramAllocate)
-+ {
-+ p_Manip->h_Ad = (t_Handle)FM_MURAM_AllocMem(
-+ p_FmPcd->h_FmMuram, FM_PCD_CC_AD_ENTRY_SIZE,
-+ FM_PCD_CC_AD_TABLE_ALIGN);
-+ if (!p_Manip->h_Ad)
-+ {
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("MURAM alloc for Manipulation action descriptor"));
-+ ReleaseManipHandler(p_Manip, p_FmPcd);
-+ XX_Free(p_Manip);
-+ return NULL;
-+ }
-+
-+ MemSet8(p_Manip->h_Ad, 0, FM_PCD_CC_AD_ENTRY_SIZE);
-+ }
-+ else
-+ {
-+ p_Manip->h_Ad = (t_Handle)XX_Malloc(
-+ FM_PCD_CC_AD_ENTRY_SIZE * sizeof(uint8_t));
-+ if (!p_Manip->h_Ad)
-+ {
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("Allocation of Manipulation action descriptor"));
-+ ReleaseManipHandler(p_Manip, p_FmPcd);
-+ XX_Free(p_Manip);
-+ return NULL;
-+ }
-+
-+ memset(p_Manip->h_Ad, 0, FM_PCD_CC_AD_ENTRY_SIZE * sizeof(uint8_t));
-+ }
-+ }
-+
-+ p_Manip->h_FmPcd = h_FmPcd;
-+
-+ return p_Manip;
-+}
-+
-+static void UpdateAdPtrOfNodesWhichPointsOnCrntMdfManip(
-+ t_FmPcdManip *p_CrntMdfManip, t_List *h_NodesLst)
-+{
-+ t_CcNodeInformation *p_CcNodeInformation;
-+ t_FmPcdCcNode *p_NodePtrOnCurrentMdfManip = NULL;
-+ t_List *p_Pos;
-+ int i = 0;
-+ t_Handle p_AdTablePtOnCrntCurrentMdfNode/*, p_AdTableNewModified*/;
-+ t_CcNodeInformation ccNodeInfo;
-+
-+ LIST_FOR_EACH(p_Pos, &p_CrntMdfManip->nodesLst)
-+ {
-+ p_CcNodeInformation = CC_NODE_F_OBJECT(p_Pos);
-+ p_NodePtrOnCurrentMdfManip =
-+ (t_FmPcdCcNode *)p_CcNodeInformation->h_CcNode;
-+
-+ ASSERT_COND(p_NodePtrOnCurrentMdfManip);
-+
-+ /* Search in the previous node which exact index points on this current modified node for getting AD */
-+ for (i = 0; i < p_NodePtrOnCurrentMdfManip->numOfKeys + 1; i++)
-+ {
-+ if (p_NodePtrOnCurrentMdfManip->keyAndNextEngineParams[i].nextEngineParams.nextEngine
-+ == e_FM_PCD_CC)
-+ {
-+ if (p_NodePtrOnCurrentMdfManip->keyAndNextEngineParams[i].nextEngineParams.h_Manip
-+ == (t_Handle)p_CrntMdfManip)
-+ {
-+ if (p_NodePtrOnCurrentMdfManip->keyAndNextEngineParams[i].p_StatsObj)
-+ p_AdTablePtOnCrntCurrentMdfNode =
-+ p_NodePtrOnCurrentMdfManip->keyAndNextEngineParams[i].p_StatsObj->h_StatsAd;
-+ else
-+ p_AdTablePtOnCrntCurrentMdfNode =
-+ PTR_MOVE(p_NodePtrOnCurrentMdfManip->h_AdTable, i*FM_PCD_CC_AD_ENTRY_SIZE);
-+
-+ memset(&ccNodeInfo, 0, sizeof(t_CcNodeInformation));
-+ ccNodeInfo.h_CcNode = p_AdTablePtOnCrntCurrentMdfNode;
-+ EnqueueNodeInfoToRelevantLst(h_NodesLst, &ccNodeInfo, NULL);
-+ }
-+ }
-+ }
-+
-+ ASSERT_COND(i != p_NodePtrOnCurrentMdfManip->numOfKeys);
-+ }
-+}
-+
-+static void BuildHmtd(uint8_t *p_Dest, uint8_t *p_Src, uint8_t *p_Hmcd,
-+ t_FmPcd *p_FmPcd)
-+{
-+ t_Error err;
-+
-+ /* Copy the HMTD */
-+ MemCpy8(p_Dest, (uint8_t*)p_Src, 16);
-+ /* Replace the HMCT table pointer */
-+ WRITE_UINT32(
-+ ((t_Hmtd *)p_Dest)->hmcdBasePtr,
-+ (uint32_t)(XX_VirtToPhys(p_Hmcd) - ((t_FmPcd*)p_FmPcd)->physicalMuramBase));
-+ /* Call Host Command to replace HMTD by a new HMTD */
-+ err = FmHcPcdCcDoDynamicChange(
-+ p_FmPcd->h_Hc,
-+ (uint32_t)(XX_VirtToPhys(p_Src) - p_FmPcd->physicalMuramBase),
-+ (uint32_t)(XX_VirtToPhys(p_Dest) - p_FmPcd->physicalMuramBase));
-+ if (err)
-+ REPORT_ERROR(MINOR, err, ("Failed in dynamic manip change, continued to the rest of the owners."));
-+}
-+
-+static t_Error FmPcdManipInitUpdate(t_Handle h_FmPcd, t_Handle h_PcdParams,
-+ t_Handle h_FmPort, t_Handle h_Manip,
-+ t_Handle h_Ad, bool validate, int level,
-+ t_Handle h_FmTree)
-+{
-+ t_FmPcdManip *p_Manip = (t_FmPcdManip *)h_Manip;
-+ t_Error err = E_OK;
-+
-+ SANITY_CHECK_RETURN_ERROR(h_Manip, E_INVALID_HANDLE);
-+
-+ UNUSED(level);
-+ UNUSED(h_FmTree);
-+
-+ switch (p_Manip->opcode)
-+ {
-+#if (defined(FM_CAPWAP_SUPPORT) && (DPAA_VERSION == 10))
-+ case (HMAN_OC_MV_INT_FRAME_HDR_FROM_FRM_TO_BUFFER_PREFFIX):
-+ err = UpdateInitMvIntFrameHeaderFromFrameToBufferPrefix(h_FmPort,
-+ p_Manip,
-+ h_Ad,
-+ validate);
-+ break;
-+ case (HMAN_OC_INSRT_HDR_BY_TEMPL_N_OR_FRAG_AFTER):
-+ if (!p_Manip->h_Frag)
-+ break;
-+ case (HMAN_OC_CAPWAP_FRAGMENTATION):
-+ err = UpdateInitCapwapFragmentation(h_FmPort, p_Manip, h_Ad, validate, h_FmTree);
-+ break;
-+ case (HMAN_OC_CAPWAP_RMV_DTLS_IF_EXIST):
-+ if (p_Manip->h_Frag)
-+ err = UpdateInitCapwapReasm(h_FmPcd, h_FmPort, p_Manip, h_Ad, validate);
-+ break;
-+ case (HMAN_OC_CAPWAP_INDEXED_STATS):
-+ err = UpdateIndxStats(h_FmPcd, h_FmPort, p_Manip);
-+ break;
-+#endif /* (defined(FM_CAPWAP_SUPPORT) && (DPAA_VERSION == 10)) */
-+ case (HMAN_OC_IP_REASSEMBLY):
-+ err = UpdateInitReasm(h_FmPcd, h_PcdParams, h_FmPort, p_Manip, h_Ad,
-+ validate);
-+ break;
-+ case (HMAN_OC_IP_FRAGMENTATION):
-+ err = UpdateInitIpFrag(h_FmPcd, h_PcdParams, h_FmPort, p_Manip,
-+ h_Ad, validate);
-+ break;
-+#if (DPAA_VERSION >= 11)
-+ case (HMAN_OC_CAPWAP_FRAGMENTATION):
-+ err = UpdateInitCapwapFrag(h_FmPcd, h_PcdParams, h_FmPort, p_Manip,
-+ h_Ad, validate);
-+ break;
-+ case (HMAN_OC_CAPWAP_REASSEMBLY):
-+ err = UpdateInitReasm(h_FmPcd, h_PcdParams, h_FmPort, p_Manip, h_Ad,
-+ validate);
-+ break;
-+#endif /* (DPAA_VERSION >= 11) */
-+ default:
-+ return E_OK;
-+ }
-+
-+ return err;
-+}
-+
-+static t_Error FmPcdManipModifyUpdate(t_Handle h_Manip, t_Handle h_Ad,
-+ bool validate, int level,
-+ t_Handle h_FmTree)
-+{
-+
-+ t_FmPcdManip *p_Manip = (t_FmPcdManip *)h_Manip;
-+ t_Error err = E_OK;
-+
-+ UNUSED(level);
-+
-+ switch (p_Manip->opcode)
-+ {
-+#if (defined(FM_CAPWAP_SUPPORT) && (DPAA_VERSION == 10))
-+ case (HMAN_OC_MV_INT_FRAME_HDR_FROM_FRM_TO_BUFFER_PREFFIX):
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_STATE,
-+ ("modify node with this type of manipulation is not suppported"));
-+ case (HMAN_OC_CAPWAP_RMV_DTLS_IF_EXIST):
-+
-+ if (p_Manip->h_Frag)
-+ {
-+ if (!(p_Manip->shadowUpdateParams & NUM_OF_TASKS)
-+ && !(p_Manip->shadowUpdateParams & OFFSET_OF_DATA)
-+ && !(p_Manip->shadowUpdateParams & OFFSET_OF_PR))
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_STATE,
-+ ("modify node with this type of manipulation requires manipulation be updated previously in SetPcd function"));
-+ }
-+ break;
-+ case (HMAN_OC_INSRT_HDR_BY_TEMPL_N_OR_FRAG_AFTER):
-+ if (p_Manip->h_Frag)
-+ err = UpdateModifyCapwapFragmenation(p_Manip, h_Ad, validate, h_FmTree);
-+ break;
-+#endif /* (defined(FM_CAPWAP_SUPPORT) && (DPAA_VERSION == 10)) */
-+ default:
-+ return E_OK;
-+ }
-+
-+ return err;
-+}
-+
-+/*****************************************************************************/
-+/* Inter-module API routines */
-+/*****************************************************************************/
-+
-+t_Error FmPcdManipUpdate(t_Handle h_FmPcd, t_Handle h_PcdParams,
-+ t_Handle h_FmPort, t_Handle h_Manip, t_Handle h_Ad,
-+ bool validate, int level, t_Handle h_FmTree,
-+ bool modify)
-+{
-+ t_Error err;
-+
-+ if (!modify)
-+ err = FmPcdManipInitUpdate(h_FmPcd, h_PcdParams, h_FmPort, h_Manip,
-+ h_Ad, validate, level, h_FmTree);
-+ else
-+ err = FmPcdManipModifyUpdate(h_Manip, h_Ad, validate, level, h_FmTree);
-+
-+ return err;
-+}
-+
-+void FmPcdManipUpdateOwner(t_Handle h_Manip, bool add)
-+{
-+
-+ uint32_t intFlags;
-+
-+ intFlags = XX_LockIntrSpinlock(((t_FmPcdManip *)h_Manip)->h_Spinlock);
-+ if (add)
-+ ((t_FmPcdManip *)h_Manip)->owner++;
-+ else
-+ {
-+ ASSERT_COND(((t_FmPcdManip *)h_Manip)->owner);
-+ ((t_FmPcdManip *)h_Manip)->owner--;
-+ }
-+ XX_UnlockIntrSpinlock(((t_FmPcdManip *)h_Manip)->h_Spinlock, intFlags);
-+}
-+
-+t_List *FmPcdManipGetNodeLstPointedOnThisManip(t_Handle h_Manip)
-+{
-+ ASSERT_COND(h_Manip);
-+ return &((t_FmPcdManip *)h_Manip)->nodesLst;
-+}
-+
-+t_List *FmPcdManipGetSpinlock(t_Handle h_Manip)
-+{
-+ ASSERT_COND(h_Manip);
-+ return ((t_FmPcdManip *)h_Manip)->h_Spinlock;
-+}
-+
-+t_Error FmPcdManipCheckParamsForCcNextEngine(
-+ t_FmPcdCcNextEngineParams *p_FmPcdCcNextEngineParams,
-+ uint32_t *requiredAction)
-+{
-+ t_FmPcdManip *p_Manip;
-+#if (defined(FM_CAPWAP_SUPPORT) && (DPAA_VERSION == 10))
-+ t_Error err = E_OK;
-+#endif /* (defined(FM_CAPWAP_SUPPORT) && (DPAA_VERSION == 10))*/
-+ bool pointFromCc = TRUE;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcdCcNextEngineParams, E_NULL_POINTER);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcdCcNextEngineParams->h_Manip,
-+ E_NULL_POINTER);
-+
-+ p_Manip = (t_FmPcdManip *)(p_FmPcdCcNextEngineParams->h_Manip);
-+ *requiredAction = 0;
-+
-+ while (p_Manip)
-+ {
-+ switch (p_Manip->opcode)
-+ {
-+#if (defined(FM_CAPWAP_SUPPORT) && (DPAA_VERSION == 10))
-+ case (HMAN_OC_CAPWAP_INDEXED_STATS):
-+ if (p_FmPcdCcNextEngineParams->nextEngine != e_FM_PCD_DONE)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("For this type of header manipulation has to be nextEngine e_FM_PCD_DONE"));
-+ if (p_FmPcdCcNextEngineParams->params.enqueueParams.overrideFqid)
-+ p_Manip->cnia = TRUE;
-+ case (HMAN_OC_CAPWAP_RMV_DTLS_IF_EXIST):
-+ *requiredAction = UPDATE_NIA_ENQ_WITHOUT_DMA;
-+ case (HMAN_OC_RMV_N_OR_INSRT_INT_FRM_HDR):
-+ p_Manip->ownerTmp++;
-+ break;
-+ case (HMAN_OC_INSRT_HDR_BY_TEMPL_N_OR_FRAG_AFTER):
-+ if ((p_FmPcdCcNextEngineParams->nextEngine != e_FM_PCD_DONE)
-+ && !p_FmPcdCcNextEngineParams->params.enqueueParams.overrideFqid)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_STATE,
-+ ("For this type of header manipulation has to be nextEngine e_FM_PCD_DONE with fqidForCtrlFlow FALSE"));
-+ p_Manip->ownerTmp++;
-+ break;
-+ case (HMAN_OC_MV_INT_FRAME_HDR_FROM_FRM_TO_BUFFER_PREFFIX):
-+ if ((p_FmPcdCcNextEngineParams->nextEngine != e_FM_PCD_CC)
-+ && (FmPcdCcGetParseCode(p_FmPcdCcNextEngineParams->params.ccParams.h_CcNode)
-+ != CC_PC_GENERIC_IC_HASH_INDEXED))
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("For this type of header manipulation next engine has to be CC and action = e_FM_PCD_ACTION_INDEXED_LOOKUP"));
-+ err = UpdateManipIc(p_FmPcdCcNextEngineParams->h_Manip,
-+ FmPcdCcGetOffset(p_FmPcdCcNextEngineParams->params.ccParams.h_CcNode));
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ *requiredAction = UPDATE_NIA_ENQ_WITHOUT_DMA;
-+ break;
-+ #endif /* (defined(FM_CAPWAP_SUPPORT) && (DPAA_VERSION == 10)) */
-+ case (HMAN_OC_IP_FRAGMENTATION):
-+ case (HMAN_OC_IP_REASSEMBLY):
-+#if (DPAA_VERSION >= 11)
-+ case (HMAN_OC_CAPWAP_REASSEMBLY):
-+ case (HMAN_OC_CAPWAP_FRAGMENTATION):
-+#endif /* (DPAA_VERSION >= 11) */
-+ if (p_FmPcdCcNextEngineParams->nextEngine != e_FM_PCD_DONE)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_STATE,
-+ ("For this type of header manipulation has to be nextEngine e_FM_PCD_DONE"));
-+ p_Manip->ownerTmp++;
-+ break;
-+ case (HMAN_OC_IPSEC_MANIP):
-+#if (DPAA_VERSION >= 11)
-+ case (HMAN_OC_CAPWAP_MANIP):
-+#endif /* (DPAA_VERSION >= 11) */
-+ p_Manip->ownerTmp++;
-+ break;
-+ case (HMAN_OC):
-+ if ((p_FmPcdCcNextEngineParams->nextEngine == e_FM_PCD_CC)
-+ && MANIP_IS_CASCADED(p_Manip))
-+ RETURN_ERROR(
-+ MINOR,
-+ E_INVALID_STATE,
-+ ("Can't have a cascaded manipulation when and Next Engine is CC"));
-+ if (!MANIP_IS_FIRST(p_Manip) && pointFromCc)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_STATE,
-+ ("h_Manip is already used and may not be shared (no sharing of non-head manip nodes)"));
-+ break;
-+ default:
-+ RETURN_ERROR(
-+ MAJOR, E_INVALID_STATE,
-+ ("invalid type of header manipulation for this state"));
-+ }
-+ p_Manip = p_Manip->h_NextManip;
-+ pointFromCc = FALSE;
-+ }
-+ return E_OK;
-+}
-+
-+
-+t_Error FmPcdManipCheckParamsWithCcNodeParams(t_Handle h_Manip,
-+ t_Handle h_FmPcdCcNode)
-+{
-+ t_FmPcdManip *p_Manip = (t_FmPcdManip *)h_Manip;
-+ t_Error err = E_OK;
-+
-+ SANITY_CHECK_RETURN_ERROR(h_Manip, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(h_FmPcdCcNode, E_INVALID_HANDLE);
-+
-+ switch (p_Manip->opcode)
-+ {
-+#if (defined(FM_CAPWAP_SUPPORT) && (DPAA_VERSION == 10))
-+ case (HMAN_OC_CAPWAP_INDEXED_STATS):
-+ if (p_Manip->ownerTmp != FmPcdCcGetNumOfKeys(h_FmPcdCcNode))
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("The manipulation of the type statistics flowId if exist has to be pointed by all numOfKeys"));
-+ break;
-+ case (HMAN_OC_CAPWAP_RMV_DTLS_IF_EXIST):
-+ if (p_Manip->h_Frag)
-+ {
-+ if (p_Manip->ownerTmp != FmPcdCcGetNumOfKeys(h_FmPcdCcNode))
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("The manipulation of the type remove DTLS if exist has to be pointed by all numOfKeys"));
-+ err = UpdateManipIc(h_Manip, FmPcdCcGetOffset(h_FmPcdCcNode));
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+ break;
-+#endif /* (defined(FM_CAPWAP_SUPPORT) && (DPAA_VERSION == 10)) */
-+ default:
-+ break;
-+ }
-+
-+ return err;
-+}
-+
-+void FmPcdManipUpdateAdResultForCc(
-+ t_Handle h_Manip, t_FmPcdCcNextEngineParams *p_CcNextEngineParams,
-+ t_Handle p_Ad, t_Handle *p_AdNewPtr)
-+{
-+ t_FmPcdManip *p_Manip = (t_FmPcdManip *)h_Manip;
-+
-+ /* This routine creates a Manip AD and can return in "p_AdNewPtr"
-+ * either the new descriptor or NULL if it writes the Manip AD into p_AD (into the match table) */
-+
-+ ASSERT_COND(p_Manip);
-+ ASSERT_COND(p_CcNextEngineParams);
-+ ASSERT_COND(p_Ad);
-+ ASSERT_COND(p_AdNewPtr);
-+
-+ FmPcdManipUpdateOwner(h_Manip, TRUE);
-+
-+ /* According to "type", either build & initialize a new AD (p_AdNew) or initialize
-+ * p_Ad ( the AD in the match table) and set p_AdNew = NULL. */
-+ switch (p_Manip->opcode)
-+ {
-+#if (defined(FM_CAPWAP_SUPPORT) && (DPAA_VERSION == 10))
-+ case (HMAN_OC_RMV_N_OR_INSRT_INT_FRM_HDR):
-+ case (HMAN_OC_CAPWAP_RMV_DTLS_IF_EXIST):
-+ case (HMAN_OC_CAPWAP_INDEXED_STATS):
-+ *p_AdNewPtr = p_Manip->h_Ad;
-+ break;
-+ case (HMAN_OC_INSRT_HDR_BY_TEMPL_N_OR_FRAG_AFTER):
-+ case (HMAN_OC_CAPWAP_FRAGMENTATION):
-+ WRITE_UINT32(((t_AdOfTypeResult *)p_Ad)->fqid,
-+ ((t_AdOfTypeResult *)(p_Manip->h_Ad))->fqid);
-+ WRITE_UINT32(((t_AdOfTypeResult *)p_Ad)->plcrProfile,
-+ ((t_AdOfTypeResult *)(p_Manip->h_Ad))->plcrProfile);
-+ WRITE_UINT32(((t_AdOfTypeResult *)p_Ad)->nia,
-+ ((t_AdOfTypeResult *)(p_Manip->h_Ad))->nia);
-+ *p_AdNewPtr = NULL;
-+ break;
-+#endif /* (defined(FM_CAPWAP_SUPPORT) && (DPAA_VERSION == 10)) */
-+ case (HMAN_OC_IPSEC_MANIP):
-+#if (DPAA_VERSION >= 11)
-+ case (HMAN_OC_CAPWAP_MANIP):
-+#endif /* (DPAA_VERSION >= 11) */
-+ *p_AdNewPtr = p_Manip->h_Ad;
-+ break;
-+ case (HMAN_OC_IP_FRAGMENTATION):
-+#if (DPAA_VERSION >= 11)
-+ case (HMAN_OC_CAPWAP_FRAGMENTATION):
-+#endif /* (DPAA_VERSION >= 11) */
-+ if ((p_CcNextEngineParams->nextEngine == e_FM_PCD_DONE)
-+ && (!p_CcNextEngineParams->params.enqueueParams.overrideFqid))
-+ {
-+ memcpy((uint8_t *)p_Ad, (uint8_t *)p_Manip->h_Ad,
-+ sizeof(t_AdOfTypeContLookup));
-+#if (DPAA_VERSION >= 11)
-+ WRITE_UINT32(
-+ ((t_AdOfTypeContLookup *)p_Ad)->ccAdBase,
-+ GET_UINT32(((t_AdOfTypeContLookup *)p_Ad)->ccAdBase) & ~FM_PCD_MANIP_IP_CNIA);
-+#endif /* (DPAA_VERSION >= 11) */
-+ *p_AdNewPtr = NULL;
-+ }
-+ else
-+ *p_AdNewPtr = p_Manip->h_Ad;
-+ break;
-+ case (HMAN_OC_IP_REASSEMBLY):
-+ if (FmPcdManipIpReassmIsIpv6Hdr(p_Manip))
-+ {
-+ if (!p_Manip->reassmParams.ip.ipv6Assigned)
-+ {
-+ *p_AdNewPtr = p_Manip->reassmParams.ip.h_Ipv6Ad;
-+ p_Manip->reassmParams.ip.ipv6Assigned = TRUE;
-+ FmPcdManipUpdateOwner(h_Manip, FALSE);
-+ }
-+ else
-+ {
-+ *p_AdNewPtr = p_Manip->reassmParams.ip.h_Ipv4Ad;
-+ p_Manip->reassmParams.ip.ipv6Assigned = FALSE;
-+ }
-+ }
-+ else
-+ *p_AdNewPtr = p_Manip->reassmParams.ip.h_Ipv4Ad;
-+ memcpy((uint8_t *)p_Ad, (uint8_t *)*p_AdNewPtr,
-+ sizeof(t_AdOfTypeContLookup));
-+ *p_AdNewPtr = NULL;
-+ break;
-+#if (DPAA_VERSION >= 11)
-+ case (HMAN_OC_CAPWAP_REASSEMBLY):
-+ *p_AdNewPtr = p_Manip->reassmParams.capwap.h_Ad;
-+ memcpy((uint8_t *)p_Ad, (uint8_t *)*p_AdNewPtr,
-+ sizeof(t_AdOfTypeContLookup));
-+ *p_AdNewPtr = NULL;
-+ break;
-+#endif /* (DPAA_VERSION >= 11) */
-+ case (HMAN_OC):
-+ /* Allocate and initialize HMTD */
-+ *p_AdNewPtr = p_Manip->h_Ad;
-+ break;
-+ default:
-+ break;
-+ }
-+}
-+
-+void FmPcdManipUpdateAdContLookupForCc(t_Handle h_Manip, t_Handle p_Ad,
-+ t_Handle *p_AdNewPtr,
-+ uint32_t adTableOffset)
-+{
-+ t_FmPcdManip *p_Manip = (t_FmPcdManip *)h_Manip;
-+
-+ /* This routine creates a Manip AD and can return in "p_AdNewPtr"
-+ * either the new descriptor or NULL if it writes the Manip AD into p_AD (into the match table) */
-+ ASSERT_COND(p_Manip);
-+
-+ FmPcdManipUpdateOwner(h_Manip, TRUE);
-+
-+ switch (p_Manip->opcode)
-+ {
-+#if (defined(FM_CAPWAP_SUPPORT) && (DPAA_VERSION == 10))
-+ case (HMAN_OC_MV_INT_FRAME_HDR_FROM_FRM_TO_BUFFER_PREFFIX):
-+ WRITE_UINT32(((t_AdOfTypeContLookup *)p_Ad)->ccAdBase,
-+ ((t_AdOfTypeContLookup *)(p_Manip->h_Ad))->ccAdBase);
-+ WRITE_UINT32(
-+ ((t_AdOfTypeContLookup *)p_Ad)->matchTblPtr,
-+ ((t_AdOfTypeContLookup *)(p_Manip->h_Ad))->matchTblPtr);
-+ WRITE_UINT32(
-+ ((t_AdOfTypeContLookup *)p_Ad)->pcAndOffsets,
-+ ((t_AdOfTypeContLookup *)(p_Manip->h_Ad))->pcAndOffsets);
-+ WRITE_UINT32(((t_AdOfTypeContLookup *)p_Ad)->gmask,
-+ ((t_AdOfTypeContLookup *)(p_Manip->h_Ad))->gmask);
-+ WRITE_UINT32(
-+ ((t_AdOfTypeContLookup *)p_Ad)->ccAdBase,
-+ (GET_UINT32(((t_AdOfTypeContLookup *)p_Ad)->ccAdBase) | adTableOffset));
-+ *p_AdNewPtr = NULL;
-+ break;
-+#endif /* (defined(FM_CAPWAP_SUPPORT) && (DPAA_VERSION == 10)) */
-+ case (HMAN_OC):
-+ /* Initialize HMTD within the match table*/
-+ MemSet8(p_Ad, 0, FM_PCD_CC_AD_ENTRY_SIZE);
-+ /* copy the existing HMTD *//* ask Alla - memcpy??? */
-+ memcpy((uint8_t*)p_Ad, p_Manip->h_Ad, sizeof(t_Hmtd));
-+ /* update NADEN to be "1"*/
-+ WRITE_UINT16(
-+ ((t_Hmtd *)p_Ad)->cfg,
-+ (uint16_t)(GET_UINT16(((t_Hmtd *)p_Ad)->cfg) | HMTD_CFG_NEXT_AD_EN));
-+ /* update next action descriptor */
-+ WRITE_UINT16(((t_Hmtd *)p_Ad)->nextAdIdx,
-+ (uint16_t)(adTableOffset >> 4));
-+ /* mark that Manip's HMTD is not used */
-+ *p_AdNewPtr = NULL;
-+ break;
-+
-+ default:
-+ break;
-+ }
-+}
-+
-+t_Error FmPcdManipBuildIpReassmScheme(t_FmPcd *p_FmPcd, t_Handle h_NetEnv,
-+ t_Handle h_CcTree, t_Handle h_Manip,
-+ bool isIpv4, uint8_t groupId)
-+{
-+ t_FmPcdManip *p_Manip = (t_FmPcdManip *)h_Manip;
-+ t_FmPcdKgSchemeParams *p_SchemeParams = NULL;
-+ t_Handle h_Scheme;
-+
-+ ASSERT_COND(p_FmPcd);
-+ ASSERT_COND(h_NetEnv);
-+ ASSERT_COND(p_Manip);
-+
-+ /* scheme was already build, no need to check for IPv6 */
-+ if (p_Manip->reassmParams.ip.h_Ipv4Scheme)
-+ return E_OK;
-+
-+ if (isIpv4) {
-+ h_Scheme = FmPcdKgGetSchemeHandle(p_FmPcd, p_Manip->reassmParams.ip.relativeSchemeId[0]);
-+ if (h_Scheme) {
-+ /* scheme was found */
-+ p_Manip->reassmParams.ip.h_Ipv4Scheme = h_Scheme;
-+ return E_OK;
-+ }
-+ } else {
-+ h_Scheme = FmPcdKgGetSchemeHandle(p_FmPcd, p_Manip->reassmParams.ip.relativeSchemeId[1]);
-+ if (h_Scheme) {
-+ /* scheme was found */
-+ p_Manip->reassmParams.ip.h_Ipv6Scheme = h_Scheme;
-+ return E_OK;
-+ }
-+ }
-+
-+ p_SchemeParams = XX_Malloc(sizeof(t_FmPcdKgSchemeParams));
-+ if (!p_SchemeParams)
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY,
-+ ("Memory allocation failed for scheme"));
-+
-+ /* Configures the IPv4 or IPv6 scheme*/
-+ memset(p_SchemeParams, 0, sizeof(t_FmPcdKgSchemeParams));
-+ p_SchemeParams->netEnvParams.h_NetEnv = h_NetEnv;
-+ p_SchemeParams->id.relativeSchemeId = (uint8_t)(
-+ (isIpv4 == TRUE) ? p_Manip->reassmParams.ip.relativeSchemeId[0] :
-+ p_Manip->reassmParams.ip.relativeSchemeId[1]);
-+ p_SchemeParams->schemeCounter.update = TRUE;
-+#if (DPAA_VERSION >= 11)
-+ p_SchemeParams->alwaysDirect = TRUE;
-+ p_SchemeParams->bypassFqidGeneration = TRUE;
-+#else
-+ p_SchemeParams->keyExtractAndHashParams.hashDistributionNumOfFqids = 1;
-+ p_SchemeParams->baseFqid = 0xFFFFFF; /*TODO- baseFqid*/
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+ setIpReassmSchemeParams(p_FmPcd, p_SchemeParams, h_CcTree, isIpv4, groupId);
-+
-+ /* Sets the new scheme */
-+ if (isIpv4)
-+ p_Manip->reassmParams.ip.h_Ipv4Scheme = FM_PCD_KgSchemeSet(
-+ p_FmPcd, p_SchemeParams);
-+ else
-+ p_Manip->reassmParams.ip.h_Ipv6Scheme = FM_PCD_KgSchemeSet(
-+ p_FmPcd, p_SchemeParams);
-+
-+ XX_Free(p_SchemeParams);
-+
-+ return E_OK;
-+}
-+
-+t_Error FmPcdManipDeleteIpReassmSchemes(t_Handle h_Manip)
-+{
-+ t_FmPcdManip *p_Manip = (t_FmPcdManip *)h_Manip;
-+
-+ ASSERT_COND(p_Manip);
-+
-+ if ((p_Manip->reassmParams.ip.h_Ipv4Scheme) &&
-+ !FmPcdKgIsSchemeHasOwners(p_Manip->reassmParams.ip.h_Ipv4Scheme))
-+ FM_PCD_KgSchemeDelete(p_Manip->reassmParams.ip.h_Ipv4Scheme);
-+
-+ if ((p_Manip->reassmParams.ip.h_Ipv6Scheme) &&
-+ !FmPcdKgIsSchemeHasOwners(p_Manip->reassmParams.ip.h_Ipv6Scheme))
-+ FM_PCD_KgSchemeDelete(p_Manip->reassmParams.ip.h_Ipv6Scheme);
-+
-+ return E_OK;
-+}
-+
-+bool FmPcdManipIpReassmIsIpv6Hdr(t_Handle h_Manip)
-+{
-+ t_FmPcdManip *p_Manip = (t_FmPcdManip *)h_Manip;
-+
-+ ASSERT_COND(p_Manip);
-+
-+ return (p_Manip->reassmParams.hdr == HEADER_TYPE_IPv6);
-+}
-+
-+t_Error FmPcdManipBuildCapwapReassmScheme(t_FmPcd *p_FmPcd, t_Handle h_NetEnv,
-+ t_Handle h_CcTree, t_Handle h_Manip,
-+ uint8_t groupId)
-+{
-+ t_FmPcdManip *p_Manip = (t_FmPcdManip *)h_Manip;
-+ t_FmPcdKgSchemeParams *p_SchemeParams = NULL;
-+
-+ ASSERT_COND(p_FmPcd);
-+ ASSERT_COND(h_NetEnv);
-+ ASSERT_COND(p_Manip);
-+
-+ /* scheme was already build, no need to check for IPv6 */
-+ if (p_Manip->reassmParams.capwap.h_Scheme)
-+ return E_OK;
-+
-+ p_SchemeParams = XX_Malloc(sizeof(t_FmPcdKgSchemeParams));
-+ if (!p_SchemeParams)
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY,
-+ ("Memory allocation failed for scheme"));
-+
-+ memset(p_SchemeParams, 0, sizeof(t_FmPcdKgSchemeParams));
-+ p_SchemeParams->netEnvParams.h_NetEnv = h_NetEnv;
-+ p_SchemeParams->id.relativeSchemeId =
-+ (uint8_t)p_Manip->reassmParams.capwap.relativeSchemeId;
-+ p_SchemeParams->schemeCounter.update = TRUE;
-+ p_SchemeParams->bypassFqidGeneration = TRUE;
-+
-+ setCapwapReassmSchemeParams(p_FmPcd, p_SchemeParams, h_CcTree, groupId);
-+
-+ p_Manip->reassmParams.capwap.h_Scheme = FM_PCD_KgSchemeSet(p_FmPcd,
-+ p_SchemeParams);
-+
-+ XX_Free(p_SchemeParams);
-+
-+ return E_OK;
-+}
-+
-+t_Error FmPcdManipDeleteCapwapReassmSchemes(t_Handle h_Manip)
-+{
-+ t_FmPcdManip *p_Manip = (t_FmPcdManip *)h_Manip;
-+
-+ ASSERT_COND(p_Manip);
-+
-+ if (p_Manip->reassmParams.capwap.h_Scheme)
-+ FM_PCD_KgSchemeDelete(p_Manip->reassmParams.capwap.h_Scheme);
-+
-+ return E_OK;
-+}
-+
-+#if (defined(FM_CAPWAP_SUPPORT) && (DPAA_VERSION == 10))
-+t_Handle FmPcdManipApplSpecificBuild(void)
-+{
-+ t_FmPcdManip *p_Manip;
-+
-+ p_Manip = (t_FmPcdManip*)XX_Malloc(sizeof(t_FmPcdManip));
-+ if (!p_Manip)
-+ {
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("No memory"));
-+ return NULL;
-+ }
-+ memset(p_Manip, 0, sizeof(t_FmPcdManip));
-+
-+ p_Manip->opcode = HMAN_OC_MV_INT_FRAME_HDR_FROM_FRM_TO_BUFFER_PREFFIX;
-+ p_Manip->muramAllocate = FALSE;
-+
-+ p_Manip->h_Ad = (t_Handle)XX_Malloc(FM_PCD_CC_AD_ENTRY_SIZE * sizeof(uint8_t));
-+ if (!p_Manip->h_Ad)
-+ {
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("Allocation of Manipulation action descriptor"));
-+ XX_Free(p_Manip);
-+ return NULL;
-+ }
-+
-+ memset(p_Manip->h_Ad, 0, FM_PCD_CC_AD_ENTRY_SIZE * sizeof(uint8_t));
-+
-+ /*treatFdStatusFieldsAsErrors = TRUE hardcoded - assumption its always come after CAAM*/
-+ /*Application specific = type of flowId index, move internal frame header from data to IC,
-+ SEC errors check*/
-+ if (MvIntFrameHeaderFromFrameToBufferPrefix(p_Manip, TRUE)!= E_OK)
-+ {
-+ XX_Free(p_Manip->h_Ad);
-+ XX_Free(p_Manip);
-+ return NULL;
-+ }
-+ return p_Manip;
-+}
-+
-+bool FmPcdManipIsCapwapApplSpecific(t_Handle h_Manip)
-+{
-+ t_FmPcdManip *p_Manip = (t_FmPcdManip *)h_Manip;
-+ ASSERT_COND(h_Manip);
-+
-+ return (bool)((p_Manip->opcode == HMAN_OC_CAPWAP_RMV_DTLS_IF_EXIST) ? TRUE : FALSE);
-+}
-+#endif /* (defined(FM_CAPWAP_SUPPORT) && (DPAA_VERSION == 10)) */
-+/*********************** End of inter-module routines ************************/
-+
-+/****************************************/
-+/* API Init unit functions */
-+/****************************************/
-+
-+t_Handle FM_PCD_ManipNodeSet(t_Handle h_FmPcd,
-+ t_FmPcdManipParams *p_ManipParams)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd *)h_FmPcd;
-+ t_FmPcdManip *p_Manip;
-+ t_Error err;
-+
-+ SANITY_CHECK_RETURN_VALUE(h_FmPcd, E_INVALID_HANDLE, NULL);
-+ SANITY_CHECK_RETURN_VALUE(p_ManipParams, E_INVALID_HANDLE, NULL);
-+
-+ p_Manip = ManipOrStatsSetNode(h_FmPcd, (t_Handle)p_ManipParams, FALSE);
-+ if (!p_Manip)
-+ return NULL;
-+
-+ if (((p_Manip->opcode == HMAN_OC_IP_REASSEMBLY)
-+ || (p_Manip->opcode == HMAN_OC_IP_FRAGMENTATION)
-+ || (p_Manip->opcode == HMAN_OC)
-+ || (p_Manip->opcode == HMAN_OC_IPSEC_MANIP)
-+#if (DPAA_VERSION >= 11)
-+ || (p_Manip->opcode == HMAN_OC_CAPWAP_MANIP)
-+ || (p_Manip->opcode == HMAN_OC_CAPWAP_FRAGMENTATION)
-+ || (p_Manip->opcode == HMAN_OC_CAPWAP_REASSEMBLY)
-+#endif /* (DPAA_VERSION >= 11) */
-+ ) && (!FmPcdIsAdvancedOffloadSupported(p_FmPcd)))
-+ {
-+ REPORT_ERROR(MAJOR, E_INVALID_STATE, ("Advanced-offload must be enabled"));
-+ XX_Free(p_Manip);
-+ return NULL;
-+ }
-+ p_Manip->h_Spinlock = XX_InitSpinlock();
-+ if (!p_Manip->h_Spinlock)
-+ {
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE, ("UNSUPPORTED HEADER MANIPULATION TYPE"));
-+ ReleaseManipHandler(p_Manip, p_FmPcd);
-+ XX_Free(p_Manip);
-+ return NULL;
-+ }INIT_LIST(&p_Manip->nodesLst);
-+
-+ switch (p_Manip->opcode)
-+ {
-+ case (HMAN_OC_IP_REASSEMBLY):
-+ /* IpReassembly */
-+ err = IpReassembly(&p_ManipParams->u.reassem, p_Manip);
-+ break;
-+ case (HMAN_OC_IP_FRAGMENTATION):
-+ /* IpFragmentation */
-+ err = IpFragmentation(&p_ManipParams->u.frag.u.ipFrag, p_Manip);
-+ if (err)
-+ break;
-+ err = IPManip(p_Manip);
-+ break;
-+ case (HMAN_OC_IPSEC_MANIP):
-+ err = IPSecManip(p_ManipParams, p_Manip);
-+ break;
-+#if (DPAA_VERSION >= 11)
-+ case (HMAN_OC_CAPWAP_REASSEMBLY):
-+ /* CapwapReassembly */
-+ err = CapwapReassembly(&p_ManipParams->u.reassem, p_Manip);
-+ break;
-+ case (HMAN_OC_CAPWAP_FRAGMENTATION):
-+ /* CapwapFragmentation */
-+ err = CapwapFragmentation(&p_ManipParams->u.frag.u.capwapFrag,
-+ p_Manip);
-+ break;
-+ case (HMAN_OC_CAPWAP_MANIP):
-+ err = CapwapManip(p_ManipParams, p_Manip);
-+ break;
-+#endif /* (DPAA_VERSION >= 11) */
-+#if (defined(FM_CAPWAP_SUPPORT) && (DPAA_VERSION == 10))
-+ case (HMAN_OC_RMV_N_OR_INSRT_INT_FRM_HDR):
-+ /* HmanType1 */
-+ err = RmvHdrTillSpecLocNOrInsrtIntFrmHdr(&p_ManipParams->u.hdr.rmvParams, p_Manip);
-+ break;
-+ case (HMAN_OC_CAPWAP_FRAGMENTATION):
-+ err = CapwapFragmentation(&p_ManipParams->fragOrReasmParams.u.capwapFragParams,
-+ p_Manip,
-+ p_FmPcd,
-+ p_ManipParams->fragOrReasmParams.sgBpid);
-+ if (err)
-+ {
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE, ("UNSUPPORTED HEADER MANIPULATION TYPE"));
-+ ReleaseManipHandler(p_Manip, p_FmPcd);
-+ XX_Free(p_Manip);
-+ return NULL;
-+ }
-+ if (p_Manip->insrt)
-+ p_Manip->opcode = HMAN_OC_INSRT_HDR_BY_TEMPL_N_OR_FRAG_AFTER;
-+ case (HMAN_OC_INSRT_HDR_BY_TEMPL_N_OR_FRAG_AFTER):
-+ /* HmanType2 + if user asked only for fragmentation still need to allocate HmanType2 */
-+ err = InsrtHdrByTempl(&p_ManipParams->u.hdr.insrtParams, p_Manip, p_FmPcd);
-+ break;
-+ case (HMAN_OC_CAPWAP_REASSEMBLY):
-+ err = CapwapReassembly(&p_ManipParams->fragOrReasmParams.u.capwapReasmParams,
-+ p_Manip,
-+ p_FmPcd,
-+ p_ManipParams->fragOrReasmParams.sgBpid);
-+ if (err)
-+ {
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE, ("UNSUPPORTED HEADER MANIPULATION TYPE"));
-+ ReleaseManipHandler(p_Manip, p_FmPcd);
-+ XX_Free(p_Manip);
-+ return NULL;
-+ }
-+ if (p_Manip->rmv)
-+ p_Manip->opcode = HMAN_OC_CAPWAP_RMV_DTLS_IF_EXIST;
-+ case (HMAN_OC_CAPWAP_RMV_DTLS_IF_EXIST):
-+ /*CAPWAP decapsulation + if user asked only for reassembly still need to allocate CAPWAP decapsulation*/
-+ err = CapwapRmvDtlsHdr(p_FmPcd, p_Manip);
-+ break;
-+ case (HMAN_OC_MV_INT_FRAME_HDR_FROM_FRM_TO_BUFFER_PREFFIX):
-+ /*Application Specific type 1*/
-+ err = MvIntFrameHeaderFromFrameToBufferPrefix(p_Manip, TRUE);
-+ break;
-+#endif /* (defined(FM_CAPWAP_SUPPORT) && (DPAA_VERSION == 10)) */
-+ case (HMAN_OC):
-+ /* New Manip */
-+ err = CreateManipActionNew(p_Manip, p_ManipParams);
-+ break;
-+ default:
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE, ("UNSUPPORTED HEADER MANIPULATION TYPE"));
-+ ReleaseManipHandler(p_Manip, p_FmPcd);
-+ XX_Free(p_Manip);
-+ return NULL;
-+ }
-+
-+ if (err)
-+ {
-+ REPORT_ERROR(MAJOR, err, NO_MSG);
-+ ReleaseManipHandler(p_Manip, p_FmPcd);
-+ XX_Free(p_Manip);
-+ return NULL;
-+ }
-+
-+ if (p_ManipParams->h_NextManip)
-+ {
-+ /* in the check routine we've verified that h_NextManip has no owners
-+ * and that only supported types are allowed. */
-+ p_Manip->h_NextManip = p_ManipParams->h_NextManip;
-+ /* save a "prev" pointer in h_NextManip */
-+ MANIP_SET_PREV(p_Manip->h_NextManip, p_Manip);
-+ FmPcdManipUpdateOwner(p_Manip->h_NextManip, TRUE);
-+ }
-+
-+ return p_Manip;
-+}
-+
-+t_Error FM_PCD_ManipNodeReplace(t_Handle h_Manip,
-+ t_FmPcdManipParams *p_ManipParams)
-+{
-+ t_FmPcdManip *p_Manip = (t_FmPcdManip *)h_Manip, *p_FirstManip;
-+ t_FmPcd *p_FmPcd = (t_FmPcd *)(p_Manip->h_FmPcd);
-+ t_Error err;
-+ uint8_t *p_WholeHmct = NULL, *p_ShadowHmct = NULL, *p_Hmtd = NULL;
-+ t_List lstOfNodeshichPointsOnCrntMdfManip, *p_Pos;
-+ t_CcNodeInformation *p_CcNodeInfo;
-+ SANITY_CHECK_RETURN_ERROR(h_Manip, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_ManipParams, E_INVALID_HANDLE);
-+
-+ INIT_LIST(&lstOfNodeshichPointsOnCrntMdfManip);
-+
-+ if ((p_ManipParams->type != e_FM_PCD_MANIP_HDR)
-+ || (p_Manip->type != e_FM_PCD_MANIP_HDR))
-+ RETURN_ERROR(
-+ MINOR,
-+ E_NOT_SUPPORTED,
-+ ("FM_PCD_ManipNodeReplace Functionality supported only for Header Manipulation."));
-+
-+ ASSERT_COND(p_Manip->opcode == HMAN_OC);
-+ ASSERT_COND(p_Manip->manipParams.h_NextManip == p_Manip->h_NextManip);
-+ memcpy((uint8_t*)&p_Manip->manipParams, p_ManipParams,
-+ sizeof(p_Manip->manipParams));
-+ p_Manip->manipParams.h_NextManip = p_Manip->h_NextManip;
-+
-+ /* The replacement of the HdrManip depends on the node type.*/
-+ /*
-+ * (1) If this is an independent node, all its owners should be updated.
-+ *
-+ * (2) If it is the head of a cascaded chain (it does not have a "prev" but
-+ * it has a "next" and it has a "cascaded" indication), the next
-+ * node remains unchanged, and the behavior is as in (1).
-+ *
-+ * (3) If it is not the head, but a part of a cascaded chain, in can be
-+ * also replaced as a regular node with just one owner.
-+ *
-+ * (4) If it is a part of a chain implemented as a unified table, the
-+ * whole table is replaced and the owners of the head node must be updated.
-+ *
-+ */
-+ /* lock shadow */
-+ if (!p_FmPcd->p_CcShadow)
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, ("CC Shadow not allocated"));
-+
-+ if (!TRY_LOCK(p_FmPcd->h_ShadowSpinlock, &p_FmPcd->shadowLock))
-+ return ERROR_CODE(E_BUSY);
-+
-+ /* this routine creates a new manip action in the CC Shadow. */
-+ err = CreateManipActionShadow(p_Manip, p_ManipParams);
-+ if (err)
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+
-+ /* If the owners list is empty (these are NOT the "owners" counter, but pointers from CC)
-+ * replace only HMTD and no lcok is required. Otherwise
-+ * lock the whole PCD
-+ * In case 4 MANIP_IS_UNIFIED_NON_FIRST(p_Manip) - Use the head node instead. */
-+ if (!FmPcdLockTryLockAll(p_FmPcd))
-+ {
-+ DBG(TRACE, ("FmPcdLockTryLockAll failed"));
-+ return ERROR_CODE(E_BUSY);
-+ }
-+
-+ p_ShadowHmct = (uint8_t*)PTR_MOVE(p_FmPcd->p_CcShadow, 16);
-+
-+ p_FirstManip = (t_FmPcdManip*)GetManipInfo(p_Manip,
-+ e_MANIP_HANDLER_TABLE_OWNER);
-+ ASSERT_COND(p_FirstManip);
-+
-+ if (!LIST_IsEmpty(&p_FirstManip->nodesLst))
-+ UpdateAdPtrOfNodesWhichPointsOnCrntMdfManip(
-+ p_FirstManip, &lstOfNodeshichPointsOnCrntMdfManip);
-+
-+ p_Hmtd = (uint8_t *)GetManipInfo(p_Manip, e_MANIP_HMTD);
-+ ASSERT_COND(p_Hmtd);
-+ BuildHmtd(p_FmPcd->p_CcShadow, (uint8_t *)p_Hmtd, p_ShadowHmct,
-+ ((t_FmPcd*)(p_Manip->h_FmPcd)));
-+
-+ LIST_FOR_EACH(p_Pos, &lstOfNodeshichPointsOnCrntMdfManip)
-+ {
-+ p_CcNodeInfo = CC_NODE_F_OBJECT(p_Pos);
-+ BuildHmtd(p_FmPcd->p_CcShadow, (uint8_t *)p_CcNodeInfo->h_CcNode,
-+ p_ShadowHmct, ((t_FmPcd*)(p_Manip->h_FmPcd)));
-+ }
-+
-+ p_WholeHmct = (uint8_t *)GetManipInfo(p_Manip, e_MANIP_HMCT);
-+ ASSERT_COND(p_WholeHmct);
-+
-+ /* re-build the HMCT n the original location */
-+ err = CreateManipActionBackToOrig(p_Manip, p_ManipParams);
-+ if (err)
-+ {
-+ RELEASE_LOCK(p_FmPcd->shadowLock);
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+ }
-+
-+ p_Hmtd = (uint8_t *)GetManipInfo(p_Manip, e_MANIP_HMTD);
-+ ASSERT_COND(p_Hmtd);
-+ BuildHmtd(p_FmPcd->p_CcShadow, (uint8_t *)p_Hmtd, p_WholeHmct,
-+ ((t_FmPcd*)p_Manip->h_FmPcd));
-+
-+ /* If LIST > 0, create a list of p_Ad's that point to the HMCT. Join also t_HMTD to this list.
-+ * For each p_Hmct (from list+fixed):
-+ * call Host Command to replace HMTD by a new one */LIST_FOR_EACH(p_Pos, &lstOfNodeshichPointsOnCrntMdfManip)
-+ {
-+ p_CcNodeInfo = CC_NODE_F_OBJECT(p_Pos);
-+ BuildHmtd(p_FmPcd->p_CcShadow, (uint8_t *)p_CcNodeInfo->h_CcNode,
-+ p_WholeHmct, ((t_FmPcd*)(p_Manip->h_FmPcd)));
-+ }
-+
-+
-+ ReleaseLst(&lstOfNodeshichPointsOnCrntMdfManip);
-+
-+ FmPcdLockUnlockAll(p_FmPcd);
-+
-+ /* unlock shadow */
-+ RELEASE_LOCK(p_FmPcd->shadowLock);
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PCD_ManipNodeDelete(t_Handle h_ManipNode)
-+{
-+ t_FmPcdManip *p_Manip = (t_FmPcdManip *)h_ManipNode;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Manip, E_INVALID_HANDLE);
-+
-+ if (p_Manip->owner)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_STATE,
-+ ("This manipulation node not be removed because this node is occupied, first - unbind this node "));
-+
-+ if (p_Manip->h_NextManip)
-+ {
-+ MANIP_SET_PREV(p_Manip->h_NextManip, NULL);
-+ FmPcdManipUpdateOwner(p_Manip->h_NextManip, FALSE);
-+ }
-+
-+ if (p_Manip->p_Hmct
-+ && (MANIP_IS_UNIFIED_FIRST(p_Manip) || !MANIP_IS_UNIFIED(p_Manip)))
-+ FM_MURAM_FreeMem(((t_FmPcd *)p_Manip->h_FmPcd)->h_FmMuram,
-+ p_Manip->p_Hmct);
-+
-+ if (p_Manip->h_Spinlock)
-+ {
-+ XX_FreeSpinlock(p_Manip->h_Spinlock);
-+ p_Manip->h_Spinlock = NULL;
-+ }
-+
-+ ReleaseManipHandler(p_Manip, p_Manip->h_FmPcd);
-+
-+ XX_Free(h_ManipNode);
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PCD_ManipGetStatistics(t_Handle h_ManipNode,
-+ t_FmPcdManipStats *p_FmPcdManipStats)
-+{
-+ t_FmPcdManip *p_Manip = (t_FmPcdManip *)h_ManipNode;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Manip, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcdManipStats, E_NULL_POINTER);
-+
-+ switch (p_Manip->opcode)
-+ {
-+ case (HMAN_OC_IP_REASSEMBLY):
-+ return IpReassemblyStats(p_Manip,
-+ &p_FmPcdManipStats->u.reassem.u.ipReassem);
-+ case (HMAN_OC_IP_FRAGMENTATION):
-+ return IpFragmentationStats(p_Manip,
-+ &p_FmPcdManipStats->u.frag.u.ipFrag);
-+#if (DPAA_VERSION >= 11)
-+ case (HMAN_OC_CAPWAP_REASSEMBLY):
-+ return CapwapReassemblyStats(
-+ p_Manip, &p_FmPcdManipStats->u.reassem.u.capwapReassem);
-+ case (HMAN_OC_CAPWAP_FRAGMENTATION):
-+ return CapwapFragmentationStats(
-+ p_Manip, &p_FmPcdManipStats->u.frag.u.capwapFrag);
-+#endif /* (DPAA_VERSION >= 11) */
-+ default:
-+ RETURN_ERROR(MAJOR, E_NOT_SUPPORTED,
-+ ("no statistics to this type of manip"));
-+ }
-+
-+ return E_OK;
-+}
-+
-+#if (defined(FM_CAPWAP_SUPPORT) && (DPAA_VERSION == 10))
-+t_Handle FM_PCD_StatisticsSetNode(t_Handle h_FmPcd, t_FmPcdStatsParams *p_StatsParams)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd *)h_FmPcd;
-+ t_FmPcdManip *p_Manip;
-+ t_Error err;
-+
-+ SANITY_CHECK_RETURN_VALUE(h_FmPcd,E_INVALID_HANDLE,NULL);
-+ SANITY_CHECK_RETURN_VALUE(p_StatsParams,E_INVALID_HANDLE,NULL);
-+
-+ p_Manip = ManipOrStatsSetNode(h_FmPcd, (t_Handle)p_StatsParams, TRUE);
-+ if (!p_Manip)
-+ return NULL;
-+
-+ switch (p_Manip->opcode)
-+ {
-+ case (HMAN_OC_CAPWAP_INDEXED_STATS):
-+ /* Indexed statistics */
-+ err = IndxStats(p_StatsParams, p_Manip, p_FmPcd);
-+ break;
-+ default:
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE, ("UNSUPPORTED Statistics type"));
-+ ReleaseManipHandler(p_Manip, p_FmPcd);
-+ XX_Free(p_Manip);
-+ return NULL;
-+ }
-+
-+ if (err)
-+ {
-+ REPORT_ERROR(MAJOR, err, NO_MSG);
-+ ReleaseManipHandler(p_Manip, p_FmPcd);
-+ XX_Free(p_Manip);
-+ return NULL;
-+ }
-+
-+ return p_Manip;
-+}
-+#endif /* (defined(FM_CAPWAP_SUPPORT) && (DPAA_VERSION == 10)) */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/Pcd/fm_manip.h
-@@ -0,0 +1,555 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/******************************************************************************
-+ @File fm_manip.h
-+
-+ @Description FM PCD manip...
-+*//***************************************************************************/
-+#ifndef __FM_MANIP_H
-+#define __FM_MANIP_H
-+
-+#include "std_ext.h"
-+#include "error_ext.h"
-+#include "list_ext.h"
-+
-+#include "fm_cc.h"
-+
-+
-+/***********************************************************************/
-+/* Header manipulations defines */
-+/***********************************************************************/
-+
-+#define NUM_OF_SCRATCH_POOL_BUFFERS 1000 /*TODO - Change it!!*/
-+
-+#if (defined(FM_CAPWAP_SUPPORT) && (DPAA_VERSION == 10))
-+#define HMAN_OC_RMV_N_OR_INSRT_INT_FRM_HDR 0x2e
-+#define HMAN_OC_INSRT_HDR_BY_TEMPL_N_OR_FRAG_AFTER 0x31
-+#define HMAN_OC_MV_INT_FRAME_HDR_FROM_FRM_TO_BUFFER_PREFFIX 0x2f
-+#define HMAN_OC_CAPWAP_RMV_DTLS_IF_EXIST 0x30
-+#define HMAN_OC_CAPWAP_REASSEMBLY 0x11 /* dummy */
-+#define HMAN_OC_CAPWAP_INDEXED_STATS 0x32 /* dummy */
-+#define HMAN_OC_CAPWAP_FRAGMENTATION 0x33
-+#else
-+#define HMAN_OC_CAPWAP_MANIP 0x2F
-+#define HMAN_OC_CAPWAP_FRAG_CHECK 0x2E
-+#define HMAN_OC_CAPWAP_FRAGMENTATION 0x33
-+#define HMAN_OC_CAPWAP_REASSEMBLY 0x30
-+#endif /* (defined(FM_CAPWAP_SUPPORT) && (DPAA_VERSION == 10)) */
-+#define HMAN_OC_IP_MANIP 0x34
-+#define HMAN_OC_IP_FRAGMENTATION 0x74
-+#define HMAN_OC_IP_REASSEMBLY 0xB4
-+#define HMAN_OC_IPSEC_MANIP 0xF4
-+#define HMAN_OC 0x35
-+
-+#if (defined(FM_CAPWAP_SUPPORT) && (DPAA_VERSION == 10))
-+#define HMAN_RMV_HDR 0x80000000
-+#define HMAN_INSRT_INT_FRM_HDR 0x40000000
-+
-+#define UDP_CHECKSUM_FIELD_OFFSET_FROM_UDP 6
-+#define UDP_CHECKSUM_FIELD_SIZE 2
-+#define UDP_LENGTH_FIELD_OFFSET_FROM_UDP 4
-+
-+#define IPv4_DSCECN_FIELD_OFFSET_FROM_IP 1
-+#define IPv4_TOTALLENGTH_FIELD_OFFSET_FROM_IP 2
-+#define IPv4_HDRCHECKSUM_FIELD_OFFSET_FROM_IP 10
-+#define VLAN_TAG_FIELD_OFFSET_FROM_ETH 12
-+#define IPv4_ID_FIELD_OFFSET_FROM_IP 4
-+
-+#define IPv6_PAYLOAD_LENGTH_OFFSET_FROM_IP 4
-+#define IPv6_NEXT_HEADER_OFFSET_FROM_IP 6
-+
-+#define FM_PCD_MANIP_CAPWAP_REASM_TABLE_SIZE 0x80
-+#define FM_PCD_MANIP_CAPWAP_REASM_TABLE_ALIGN 8
-+#define FM_PCD_MANIP_CAPWAP_REASM_RFD_SIZE 32
-+#define FM_PCD_MANIP_CAPWAP_REASM_AUTO_LEARNING_HASH_ENTRY_SIZE 4
-+#define FM_PCD_MANIP_CAPWAP_REASM_TIME_OUT_ENTRY_SIZE 8
-+
-+
-+#define FM_PCD_MANIP_CAPWAP_REASM_TIME_OUT_BETWEEN_FRAMES 0x40000000
-+#define FM_PCD_MANIP_CAPWAP_REASM_HALT_ON_DUPLICATE_FRAG 0x10000000
-+#define FM_PCD_MANIP_CAPWAP_REASM_AUTOMATIC_LEARNIN_HASH_8_WAYS 0x08000000
-+#define FM_PCD_MANIP_CAPWAP_REASM_PR_COPY 0x00800000
-+
-+#define FM_PCD_MANIP_CAPWAP_FRAG_COMPR_OPTION_FIELD_EN 0x80000000
-+
-+#define FM_PCD_MANIP_INDEXED_STATS_ENTRY_SIZE 4
-+#define FM_PCD_MANIP_INDEXED_STATS_CNIA 0x20000000
-+#define FM_PCD_MANIP_INDEXED_STATS_DPD 0x10000000
-+#endif /* (defined(FM_CAPWAP_SUPPORT) && (DPAA_VERSION == 10)) */
-+
-+#if (DPAA_VERSION >= 11)
-+#define FM_PCD_MANIP_CAPWAP_DTLS 0x00040000
-+#define FM_PCD_MANIP_CAPWAP_NADEN 0x20000000
-+
-+#define FM_PCD_MANIP_CAPWAP_FRAG_CHECK_MTU_SHIFT 16
-+#define FM_PCD_MANIP_CAPWAP_FRAG_CHECK_NO_FRAGMENTATION 0xFFFF0000
-+#define FM_PCD_MANIP_CAPWAP_FRAG_CHECK_CNIA 0x20000000
-+
-+#define FM_PCD_MANIP_CAPWAP_FRAG_COMPRESS_EN 0x04000000
-+#define FM_PCD_MANIP_CAPWAP_FRAG_SCRATCH_BPID 24
-+#define FM_PCD_MANIP_CAPWAP_FRAG_SG_BDID_EN 0x08000000
-+#define FM_PCD_MANIP_CAPWAP_FRAG_SG_BDID_MASK 0xFF000000
-+#define FM_PCD_MANIP_CAPWAP_FRAG_SG_BDID_SHIFT 24
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+#define FM_PCD_MANIP_REASM_TABLE_SIZE 0x40
-+#define FM_PCD_MANIP_REASM_TABLE_ALIGN 8
-+
-+#define FM_PCD_MANIP_REASM_COMMON_PARAM_TABLE_SIZE 64
-+#define FM_PCD_MANIP_REASM_COMMON_PARAM_TABLE_ALIGN 8
-+#define FM_PCD_MANIP_REASM_TIME_OUT_BETWEEN_FRAMES 0x80000000
-+#define FM_PCD_MANIP_REASM_COUPLING_ENABLE 0x40000000
-+#define FM_PCD_MANIP_REASM_COUPLING_MASK 0xFF000000
-+#define FM_PCD_MANIP_REASM_COUPLING_SHIFT 24
-+#define FM_PCD_MANIP_REASM_LIODN_MASK 0x0000003F
-+#define FM_PCD_MANIP_REASM_LIODN_SHIFT 56
-+#define FM_PCD_MANIP_REASM_ELIODN_MASK 0x000003c0
-+#define FM_PCD_MANIP_REASM_ELIODN_SHIFT 38
-+#define FM_PCD_MANIP_REASM_COMMON_INT_BUFFER_IDX_MASK 0x000000FF
-+#define FM_PCD_MANIP_REASM_COMMON_INT_BUFFER_IDX_SHIFT 24
-+#define FM_PCD_MANIP_REASM_TIMEOUT_THREAD_THRESH 1024
-+
-+#define FM_PCD_MANIP_IP_MTU_SHIFT 16
-+#define FM_PCD_MANIP_IP_NO_FRAGMENTATION 0xFFFF0000
-+#define FM_PCD_MANIP_IP_CNIA 0x20000000
-+
-+#define FM_PCD_MANIP_IP_FRAG_DF_SHIFT 28
-+#define FM_PCD_MANIP_IP_FRAG_SCRATCH_BPID 24
-+#define FM_PCD_MANIP_IP_FRAG_SG_BDID_EN 0x08000000
-+#define FM_PCD_MANIP_IP_FRAG_SG_BDID_MASK 0xFF000000
-+#define FM_PCD_MANIP_IP_FRAG_SG_BDID_SHIFT 24
-+
-+#define FM_PCD_MANIP_IPSEC_DEC 0x10000000
-+#define FM_PCD_MANIP_IPSEC_VIPV_EN 0x08000000
-+#define FM_PCD_MANIP_IPSEC_ECN_EN 0x04000000
-+#define FM_PCD_MANIP_IPSEC_DSCP_EN 0x02000000
-+#define FM_PCD_MANIP_IPSEC_VIPL_EN 0x01000000
-+#define FM_PCD_MANIP_IPSEC_NADEN 0x20000000
-+
-+#define FM_PCD_MANIP_IPSEC_IP_HDR_LEN_MASK 0x00FF0000
-+#define FM_PCD_MANIP_IPSEC_IP_HDR_LEN_SHIFT 16
-+
-+#define FM_PCD_MANIP_IPSEC_ARW_SIZE_MASK 0xFFFF0000
-+#define FM_PCD_MANIP_IPSEC_ARW_SIZE_SHIFT 16
-+
-+#define e_FM_MANIP_IP_INDX 1
-+
-+#define HMCD_OPCODE_GENERIC_RMV 0x01
-+#define HMCD_OPCODE_GENERIC_INSRT 0x02
-+#define HMCD_OPCODE_GENERIC_REPLACE 0x05
-+#define HMCD_OPCODE_L2_RMV 0x08
-+#define HMCD_OPCODE_L2_INSRT 0x09
-+#define HMCD_OPCODE_VLAN_PRI_UPDATE 0x0B
-+#define HMCD_OPCODE_IPV4_UPDATE 0x0C
-+#define HMCD_OPCODE_IPV6_UPDATE 0x10
-+#define HMCD_OPCODE_TCP_UDP_UPDATE 0x0E
-+#define HMCD_OPCODE_TCP_UDP_CHECKSUM 0x14
-+#define HMCD_OPCODE_REPLACE_IP 0x12
-+#define HMCD_OPCODE_RMV_TILL 0x15
-+#define HMCD_OPCODE_UDP_INSRT 0x16
-+#define HMCD_OPCODE_IP_INSRT 0x17
-+#define HMCD_OPCODE_CAPWAP_RMV 0x18
-+#define HMCD_OPCODE_CAPWAP_INSRT 0x18
-+#define HMCD_OPCODE_GEN_FIELD_REPLACE 0x19
-+
-+#define HMCD_LAST 0x00800000
-+
-+#define HMCD_DSCP_VALUES 64
-+
-+#define HMCD_BASIC_SIZE 4
-+#define HMCD_PTR_SIZE 4
-+#define HMCD_PARAM_SIZE 4
-+#define HMCD_IPV4_ADDR_SIZE 4
-+#define HMCD_IPV6_ADDR_SIZE 0x10
-+#define HMCD_L4_HDR_SIZE 8
-+
-+#define HMCD_CAPWAP_INSRT 0x00010000
-+#define HMCD_INSRT_UDP_LITE 0x00010000
-+#define HMCD_IP_ID_MASK 0x0000FFFF
-+#define HMCD_IP_SIZE_MASK 0x0000FF00
-+#define HMCD_IP_SIZE_SHIFT 8
-+#define HMCD_IP_LAST_PID_MASK 0x000000FF
-+#define HMCD_IP_OR_QOS 0x00010000
-+#define HMCD_IP_L4_CS_CALC 0x00040000
-+#define HMCD_IP_DF_MODE 0x00400000
-+
-+
-+#define HMCD_OC_SHIFT 24
-+
-+#define HMCD_RMV_OFFSET_SHIFT 0
-+#define HMCD_RMV_SIZE_SHIFT 8
-+
-+#define HMCD_INSRT_OFFSET_SHIFT 0
-+#define HMCD_INSRT_SIZE_SHIFT 8
-+
-+#define HMTD_CFG_TYPE 0x4000
-+#define HMTD_CFG_EXT_HMCT 0x0080
-+#define HMTD_CFG_PRS_AFTER_HM 0x0040
-+#define HMTD_CFG_NEXT_AD_EN 0x0020
-+
-+#define HMCD_RMV_L2_ETHERNET 0
-+#define HMCD_RMV_L2_STACKED_QTAGS 1
-+#define HMCD_RMV_L2_ETHERNET_AND_MPLS 2
-+#define HMCD_RMV_L2_MPLS 3
-+#define HMCD_RMV_L2_PPPOE 4
-+
-+#define HMCD_INSRT_L2_MPLS 0
-+#define HMCD_INSRT_N_UPDATE_L2_MPLS 1
-+#define HMCD_INSRT_L2_PPPOE 2
-+#define HMCD_INSRT_L2_SIZE_SHIFT 24
-+
-+#define HMCD_L2_MODE_SHIFT 16
-+
-+#define HMCD_VLAN_PRI_REP_MODE_SHIFT 16
-+#define HMCD_VLAN_PRI_UPDATE 0
-+#define HMCD_VLAN_PRI_UPDATE_DSCP_TO_VPRI 1
-+
-+#define HMCD_IPV4_UPDATE_TTL 0x00000001
-+#define HMCD_IPV4_UPDATE_TOS 0x00000002
-+#define HMCD_IPV4_UPDATE_DST 0x00000020
-+#define HMCD_IPV4_UPDATE_SRC 0x00000040
-+#define HMCD_IPV4_UPDATE_ID 0x00000080
-+#define HMCD_IPV4_UPDATE_TOS_SHIFT 8
-+
-+#define HMCD_IPV6_UPDATE_HL 0x00000001
-+#define HMCD_IPV6_UPDATE_TC 0x00000002
-+#define HMCD_IPV6_UPDATE_DST 0x00000040
-+#define HMCD_IPV6_UPDATE_SRC 0x00000080
-+#define HMCD_IPV6_UPDATE_TC_SHIFT 8
-+
-+#define HMCD_TCP_UDP_UPDATE_DST 0x00004000
-+#define HMCD_TCP_UDP_UPDATE_SRC 0x00008000
-+#define HMCD_TCP_UDP_UPDATE_SRC_SHIFT 16
-+
-+#define HMCD_IP_REPLACE_REPLACE_IPV4 0x00000000
-+#define HMCD_IP_REPLACE_REPLACE_IPV6 0x00010000
-+#define HMCD_IP_REPLACE_TTL_HL 0x00200000
-+#define HMCD_IP_REPLACE_ID 0x00400000
-+
-+#define HMCD_IP_REPLACE_L3HDRSIZE_SHIFT 24
-+
-+#define HMCD_GEN_FIELD_SIZE_SHIFT 16
-+#define HMCD_GEN_FIELD_SRC_OFF_SHIFT 8
-+#define HMCD_GEN_FIELD_DST_OFF_SHIFT 0
-+#define HMCD_GEN_FIELD_MASK_EN 0x00400000
-+
-+#define HMCD_GEN_FIELD_MASK_OFF_SHIFT 16
-+#define HMCD_GEN_FIELD_MASK_SHIFT 24
-+
-+#define DSCP_TO_VLAN_TABLE_SIZE 32
-+
-+#define MANIP_GET_HMCT_SIZE(h_Manip) (((t_FmPcdManip *)h_Manip)->tableSize)
-+#define MANIP_GET_DATA_SIZE(h_Manip) (((t_FmPcdManip *)h_Manip)->dataSize)
-+
-+#define MANIP_GET_HMCT_PTR(h_Manip) (((t_FmPcdManip *)h_Manip)->p_Hmct)
-+#define MANIP_GET_DATA_PTR(h_Manip) (((t_FmPcdManip *)h_Manip)->p_Data)
-+
-+#define MANIP_SET_HMCT_PTR(h_Manip, h_NewPtr) (((t_FmPcdManip *)h_Manip)->p_Hmct = h_NewPtr)
-+#define MANIP_SET_DATA_PTR(h_Manip, h_NewPtr) (((t_FmPcdManip *)h_Manip)->p_Data = h_NewPtr)
-+
-+#define MANIP_GET_HMTD_PTR(h_Manip) (((t_FmPcdManip *)h_Manip)->h_Ad)
-+#define MANIP_DONT_REPARSE(h_Manip) (((t_FmPcdManip *)h_Manip)->dontParseAfterManip)
-+#define MANIP_SET_PREV(h_Manip, h_Prev) (((t_FmPcdManip *)h_Manip)->h_PrevManip = h_Prev)
-+#define MANIP_GET_OWNERS(h_Manip) (((t_FmPcdManip *)h_Manip)->owner)
-+#define MANIP_GET_TYPE(h_Manip) (((t_FmPcdManip *)h_Manip)->type)
-+#define MANIP_SET_UNIFIED_TBL_PTR_INDICATION(h_Manip) (((t_FmPcdManip *)h_Manip)->unifiedTablePtr = TRUE)
-+#define MANIP_GET_MURAM(h_Manip) (((t_FmPcd *)((t_FmPcdManip *)h_Manip)->h_FmPcd)->h_FmMuram)
-+#define MANIP_FREE_HMTD(h_Manip) \
-+ {if (((t_FmPcdManip *)h_Manip)->muramAllocate) \
-+ FM_MURAM_FreeMem(((t_FmPcd *)((t_FmPcdManip *)h_Manip)->h_FmPcd)->h_FmMuram, ((t_FmPcdManip *)h_Manip)->h_Ad);\
-+ else \
-+ XX_Free(((t_FmPcdManip *)h_Manip)->h_Ad); \
-+ ((t_FmPcdManip *)h_Manip)->h_Ad = NULL; \
-+ }
-+/* position regarding Manip SW structure */
-+#define MANIP_IS_FIRST(h_Manip) (!(((t_FmPcdManip *)h_Manip)->h_PrevManip))
-+#define MANIP_IS_CASCADED(h_Manip) (((t_FmPcdManip *)h_Manip)->cascaded)
-+#define MANIP_IS_UNIFIED(h_Manip) (!(((t_FmPcdManip *)h_Manip)->unifiedPosition == e_MANIP_UNIFIED_NONE))
-+#define MANIP_IS_UNIFIED_NON_FIRST(h_Manip) ((((t_FmPcdManip *)h_Manip)->unifiedPosition == e_MANIP_UNIFIED_MID) || \
-+ (((t_FmPcdManip *)h_Manip)->unifiedPosition == e_MANIP_UNIFIED_LAST))
-+#define MANIP_IS_UNIFIED_NON_LAST(h_Manip) ((((t_FmPcdManip *)h_Manip)->unifiedPosition == e_MANIP_UNIFIED_FIRST) ||\
-+ (((t_FmPcdManip *)h_Manip)->unifiedPosition == e_MANIP_UNIFIED_MID))
-+#define MANIP_IS_UNIFIED_FIRST(h_Manip) (((t_FmPcdManip *)h_Manip)->unifiedPosition == e_MANIP_UNIFIED_FIRST)
-+#define MANIP_IS_UNIFIED_LAST(h_Manip) (((t_FmPcdManip *)h_Manip)->unifiedPosition == e_MANIP_UNIFIED_LAST)
-+
-+#define MANIP_UPDATE_UNIFIED_POSITION(h_Manip) (((t_FmPcdManip *)h_Manip)->unifiedPosition = \
-+ (((t_FmPcdManip *)h_Manip)->unifiedPosition == e_MANIP_UNIFIED_NONE)? \
-+ e_MANIP_UNIFIED_LAST : e_MANIP_UNIFIED_MID)
-+
-+typedef enum e_ManipUnifiedPosition {
-+ e_MANIP_UNIFIED_NONE = 0,
-+ e_MANIP_UNIFIED_FIRST,
-+ e_MANIP_UNIFIED_MID,
-+ e_MANIP_UNIFIED_LAST
-+} e_ManipUnifiedPosition;
-+
-+typedef enum e_ManipInfo {
-+ e_MANIP_HMTD,
-+ e_MANIP_HMCT,
-+ e_MANIP_HANDLER_TABLE_OWNER
-+}e_ManipInfo;
-+/***********************************************************************/
-+/* Memory map */
-+/***********************************************************************/
-+#if defined(__MWERKS__) && !defined(__GNUC__)
-+#pragma pack(push,1)
-+#endif /* defined(__MWERKS__) && ... */
-+
-+#if (defined(FM_CAPWAP_SUPPORT) && (DPAA_VERSION == 10))
-+typedef struct t_CapwapReasmPram {
-+ volatile uint32_t mode;
-+ volatile uint32_t autoLearnHashTblPtr;
-+ volatile uint32_t intStatsTblPtr;
-+ volatile uint32_t reasmFrmDescPoolTblPtr;
-+ volatile uint32_t reasmFrmDescIndexPoolTblPtr;
-+ volatile uint32_t timeOutTblPtr;
-+ volatile uint32_t bufferPoolIdAndRisc1SetIndexes;
-+ volatile uint32_t risc23SetIndexes;
-+ volatile uint32_t risc4SetIndexesAndExtendedStatsTblPtr;
-+ volatile uint32_t extendedStatsTblPtr;
-+ volatile uint32_t expirationDelay;
-+ volatile uint32_t totalProcessedFragCounter;
-+ volatile uint32_t totalUnsuccessfulReasmFramesCounter;
-+ volatile uint32_t totalDuplicatedFragCounter;
-+ volatile uint32_t totalMalformdFragCounter;
-+ volatile uint32_t totalTimeOutCounter;
-+ volatile uint32_t totalSetBusyCounter;
-+ volatile uint32_t totalRfdPoolBusyCounter;
-+ volatile uint32_t totalDiscardedFragsCounter;
-+ volatile uint32_t totalMoreThan16FramesCounter;
-+ volatile uint32_t internalBufferBusy;
-+ volatile uint32_t externalBufferBusy;
-+ volatile uint32_t reserved1[4];
-+} t_CapwapReasmPram;
-+#endif /* (defined(FM_CAPWAP_SUPPORT) && (DPAA_VERSION == 10)) */
-+
-+typedef _Packed struct t_ReassTbl {
-+ volatile uint16_t waysNumAndSetSize;
-+ volatile uint16_t autoLearnHashKeyMask;
-+ volatile uint32_t reassCommonPrmTblPtr;
-+ volatile uint32_t liodnAlAndAutoLearnHashTblPtrHi;
-+ volatile uint32_t autoLearnHashTblPtrLow;
-+ volatile uint32_t liodnSlAndAutoLearnSetLockTblPtrHi;
-+ volatile uint32_t autoLearnSetLockTblPtrLow;
-+ volatile uint16_t minFragSize; /* Not relevant for CAPWAP*/
-+ volatile uint16_t maxReassemblySize; /* Only relevant for CAPWAP*/
-+ volatile uint32_t totalSuccessfullyReasmFramesCounter;
-+ volatile uint32_t totalValidFragmentCounter;
-+ volatile uint32_t totalProcessedFragCounter;
-+ volatile uint32_t totalMalformdFragCounter;
-+ volatile uint32_t totalSetBusyCounter;
-+ volatile uint32_t totalDiscardedFragsCounter;
-+ volatile uint32_t totalMoreThan16FramesCounter;
-+ volatile uint32_t reserved2[2];
-+} _PackedType t_ReassTbl;
-+
-+typedef struct t_ReassCommonTbl {
-+ volatile uint32_t timeoutModeAndFqid;
-+ volatile uint32_t reassFrmDescIndexPoolTblPtr;
-+ volatile uint32_t liodnAndReassFrmDescPoolPtrHi;
-+ volatile uint32_t reassFrmDescPoolPtrLow;
-+ volatile uint32_t timeOutTblPtr;
-+ volatile uint32_t expirationDelay;
-+ volatile uint32_t internalBufferManagement;
-+ volatile uint32_t reserved2;
-+ volatile uint32_t totalTimeOutCounter;
-+ volatile uint32_t totalRfdPoolBusyCounter;
-+ volatile uint32_t totalInternalBufferBusy;
-+ volatile uint32_t totalExternalBufferBusy;
-+ volatile uint32_t totalSgFragmentCounter;
-+ volatile uint32_t totalDmaSemaphoreDepletionCounter;
-+ volatile uint32_t totalNCSPCounter;
-+ volatile uint32_t discardMask;
-+} t_ReassCommonTbl;
-+
-+typedef _Packed struct t_Hmtd {
-+ volatile uint16_t cfg;
-+ volatile uint8_t eliodnOffset;
-+ volatile uint8_t extHmcdBasePtrHi;
-+ volatile uint32_t hmcdBasePtr;
-+ volatile uint16_t nextAdIdx;
-+ volatile uint8_t res1;
-+ volatile uint8_t opCode;
-+ volatile uint32_t res2;
-+} _PackedType t_Hmtd;
-+
-+#if defined(__MWERKS__) && !defined(__GNUC__)
-+#pragma pack(pop)
-+#endif /* defined(__MWERKS__) && ... */
-+
-+
-+/***********************************************************************/
-+/* Driver's internal structures */
-+/***********************************************************************/
-+#if (defined(FM_CAPWAP_SUPPORT) && (DPAA_VERSION == 10))
-+typedef struct
-+{
-+ t_Handle p_AutoLearnHashTbl;
-+ t_Handle p_ReassmFrmDescrPoolTbl;
-+ t_Handle p_ReassmFrmDescrIndxPoolTbl;
-+ t_Handle p_TimeOutTbl;
-+ uint16_t maxNumFramesInProcess;
-+ uint8_t numOfTasks;
-+ //uint8_t poolId;
-+ uint8_t prOffset;
-+ uint16_t dataOffset;
-+ uint8_t sgBpid;
-+ uint8_t hwPortId;
-+ uint32_t fqidForTimeOutFrames;
-+ uint32_t timeoutRoutineRequestTime;
-+ uint32_t bitFor1Micro;
-+} t_CapwapFragParams;
-+#endif /* (defined(FM_CAPWAP_SUPPORT) && (DPAA_VERSION == 10)) */
-+
-+typedef struct
-+{
-+ t_AdOfTypeContLookup *p_Frag;
-+#if (DPAA_VERSION == 10)
-+ uint8_t scratchBpid;
-+#endif /* (DPAA_VERSION == 10) */
-+} t_FragParams;
-+
-+typedef struct t_ReassmParams
-+{
-+ e_NetHeaderType hdr; /* Header selection */
-+ t_ReassCommonTbl *p_ReassCommonTbl;
-+ uintptr_t reassFrmDescrIndxPoolTblAddr;
-+ uintptr_t reassFrmDescrPoolTblAddr;
-+ uintptr_t timeOutTblAddr;
-+ uintptr_t internalBufferPoolManagementIndexAddr;
-+ uintptr_t internalBufferPoolAddr;
-+ uint32_t maxNumFramesInProcess;
-+ uint8_t sgBpid;
-+ uint8_t dataMemId;
-+ uint16_t dataLiodnOffset;
-+ uint32_t fqidForTimeOutFrames;
-+ e_FmPcdManipReassemTimeOutMode timeOutMode;
-+ uint32_t timeoutThresholdForReassmProcess;
-+ union {
-+ struct {
-+ t_Handle h_Ipv4Ad;
-+ t_Handle h_Ipv6Ad;
-+ bool ipv6Assigned;
-+ t_ReassTbl *p_Ipv4ReassTbl;
-+ t_ReassTbl *p_Ipv6ReassTbl;
-+ uintptr_t ipv4AutoLearnHashTblAddr;
-+ uintptr_t ipv6AutoLearnHashTblAddr;
-+ uintptr_t ipv4AutoLearnSetLockTblAddr;
-+ uintptr_t ipv6AutoLearnSetLockTblAddr;
-+ uint16_t minFragSize[2];
-+ e_FmPcdManipReassemWaysNumber numOfFramesPerHashEntry[2];
-+ uint8_t relativeSchemeId[2];
-+ t_Handle h_Ipv4Scheme;
-+ t_Handle h_Ipv6Scheme;
-+ uint32_t nonConsistentSpFqid;
-+ } ip;
-+ struct {
-+ t_Handle h_Ad;
-+ t_ReassTbl *p_ReassTbl;
-+ uintptr_t autoLearnHashTblAddr;
-+ uintptr_t autoLearnSetLockTblAddr;
-+ uint16_t maxRessembledsSize;
-+ e_FmPcdManipReassemWaysNumber numOfFramesPerHashEntry;
-+ uint8_t relativeSchemeId;
-+ t_Handle h_Scheme;
-+ } capwap;
-+ };
-+} t_ReassmParams;
-+
-+typedef struct{
-+ e_FmPcdManipType type;
-+ t_FmPcdManipParams manipParams;
-+ bool muramAllocate;
-+ t_Handle h_Ad;
-+ uint32_t opcode;
-+ bool rmv;
-+ bool insrt;
-+ t_Handle h_NextManip;
-+ t_Handle h_PrevManip;
-+ e_FmPcdManipType nextManipType;
-+ /* HdrManip parameters*/
-+ uint8_t *p_Hmct;
-+ uint8_t *p_Data;
-+ bool dontParseAfterManip;
-+ bool fieldUpdate;
-+ bool custom;
-+ uint16_t tableSize;
-+ uint8_t dataSize;
-+ bool cascaded;
-+ e_ManipUnifiedPosition unifiedPosition;
-+ /* end HdrManip */
-+ uint8_t *p_Template;
-+ uint16_t owner;
-+ uint32_t updateParams;
-+ uint32_t shadowUpdateParams;
-+ bool frag;
-+ bool reassm;
-+ uint16_t sizeForFragmentation;
-+#if (defined(FM_CAPWAP_SUPPORT) && (DPAA_VERSION == 10))
-+ t_Handle h_Frag;
-+ t_CapwapFragParams capwapFragParams;
-+#endif /* (defined(FM_CAPWAP_SUPPORT) && (DPAA_VERSION == 10)) */
-+ union {
-+ t_ReassmParams reassmParams;
-+ t_FragParams fragParams;
-+ };
-+ uint8_t icOffset;
-+ uint16_t ownerTmp;
-+ bool cnia;
-+ t_Handle p_StatsTbl;
-+ t_Handle h_FmPcd;
-+ t_List nodesLst;
-+ t_Handle h_Spinlock;
-+} t_FmPcdManip;
-+
-+typedef struct t_FmPcdCcSavedManipParams
-+{
-+ union
-+ {
-+ struct
-+ {
-+ uint16_t dataOffset;
-+ //uint8_t poolId;
-+ }capwapParams;
-+ struct
-+ {
-+ uint16_t dataOffset;
-+ uint8_t poolId;
-+ }ipParams;
-+ };
-+
-+} t_FmPcdCcSavedManipParams;
-+
-+
-+#endif /* __FM_MANIP_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/Pcd/fm_pcd.c
-@@ -0,0 +1,2095 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/******************************************************************************
-+ @File fm_pcd.c
-+
-+ @Description FM PCD ...
-+*//***************************************************************************/
-+#include "std_ext.h"
-+#include "error_ext.h"
-+#include "string_ext.h"
-+#include "xx_ext.h"
-+#include "sprint_ext.h"
-+#include "debug_ext.h"
-+#include "net_ext.h"
-+#include "fm_ext.h"
-+#include "fm_pcd_ext.h"
-+
-+#include "fm_common.h"
-+#include "fm_pcd.h"
-+#include "fm_pcd_ipc.h"
-+#include "fm_hc.h"
-+#include "fm_muram_ext.h"
-+
-+
-+/****************************************/
-+/* static functions */
-+/****************************************/
-+
-+static t_Error CheckFmPcdParameters(t_FmPcd *p_FmPcd)
-+{
-+ if (!p_FmPcd->h_Fm)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("h_Fm has to be initialized"));
-+
-+ if (p_FmPcd->guestId == NCSW_MASTER_ID)
-+ {
-+ if (p_FmPcd->p_FmPcdKg && !p_FmPcd->p_FmPcdKg->p_FmPcdKgRegs)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("Something WRONG"));
-+
-+ if (p_FmPcd->p_FmPcdPlcr && !p_FmPcd->p_FmPcdPlcr->p_FmPcdPlcrRegs)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("Something WRONG"));
-+
-+ if (!p_FmPcd->f_Exception)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("f_FmPcdExceptions has to be initialized"));
-+
-+ if ((!p_FmPcd->f_FmPcdIndexedException) && (p_FmPcd->p_FmPcdPlcr || p_FmPcd->p_FmPcdKg))
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("f_FmPcdIndexedException has to be initialized"));
-+
-+ if (p_FmPcd->p_FmPcdDriverParam->prsMaxParseCycleLimit > PRS_MAX_CYCLE_LIMIT)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("prsMaxParseCycleLimit has to be less than 8191"));
-+ }
-+
-+ return E_OK;
-+}
-+
-+static volatile bool blockingFlag = FALSE;
-+static void IpcMsgCompletionCB(t_Handle h_FmPcd,
-+ uint8_t *p_Msg,
-+ uint8_t *p_Reply,
-+ uint32_t replyLength,
-+ t_Error status)
-+{
-+ UNUSED(h_FmPcd);UNUSED(p_Msg);UNUSED(p_Reply);UNUSED(replyLength);UNUSED(status);
-+ blockingFlag = FALSE;
-+}
-+
-+static t_Error IpcMsgHandlerCB(t_Handle h_FmPcd,
-+ uint8_t *p_Msg,
-+ uint32_t msgLength,
-+ uint8_t *p_Reply,
-+ uint32_t *p_ReplyLength)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+ t_Error err = E_OK;
-+ t_FmPcdIpcMsg *p_IpcMsg = (t_FmPcdIpcMsg*)p_Msg;
-+ t_FmPcdIpcReply *p_IpcReply = (t_FmPcdIpcReply*)p_Reply;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR((msgLength >= sizeof(uint32_t)), E_INVALID_VALUE);
-+
-+#ifdef DISABLE_SANITY_CHECKS
-+ UNUSED(msgLength);
-+#endif /* DISABLE_SANITY_CHECKS */
-+
-+ ASSERT_COND(p_Msg);
-+
-+ memset(p_IpcReply, 0, (sizeof(uint8_t) * FM_PCD_MAX_REPLY_SIZE));
-+ *p_ReplyLength = 0;
-+
-+ switch (p_IpcMsg->msgId)
-+ {
-+ case (FM_PCD_MASTER_IS_ALIVE):
-+ *(uint8_t*)(p_IpcReply->replyBody) = 1;
-+ p_IpcReply->error = E_OK;
-+ *p_ReplyLength = sizeof(uint32_t) + sizeof(uint8_t);
-+ break;
-+ case (FM_PCD_MASTER_IS_ENABLED):
-+ /* count partitions registrations */
-+ if (p_FmPcd->enabled)
-+ p_FmPcd->numOfEnabledGuestPartitionsPcds++;
-+ *(uint8_t*)(p_IpcReply->replyBody) = (uint8_t)p_FmPcd->enabled;
-+ p_IpcReply->error = E_OK;
-+ *p_ReplyLength = sizeof(uint32_t) + sizeof(uint8_t);
-+ break;
-+ case (FM_PCD_GUEST_DISABLE):
-+ if (p_FmPcd->numOfEnabledGuestPartitionsPcds)
-+ {
-+ p_FmPcd->numOfEnabledGuestPartitionsPcds--;
-+ p_IpcReply->error = E_OK;
-+ }
-+ else
-+ {
-+ REPORT_ERROR(MINOR, E_INVALID_STATE,("Trying to disable an unregistered partition"));
-+ p_IpcReply->error = E_INVALID_STATE;
-+ }
-+ *p_ReplyLength = sizeof(uint32_t);
-+ break;
-+ case (FM_PCD_GET_COUNTER):
-+ {
-+ e_FmPcdCounters inCounter;
-+ uint32_t outCounter;
-+
-+ memcpy((uint8_t*)&inCounter, p_IpcMsg->msgBody, sizeof(uint32_t));
-+ outCounter = FM_PCD_GetCounter(h_FmPcd, inCounter);
-+ memcpy(p_IpcReply->replyBody, (uint8_t*)&outCounter, sizeof(uint32_t));
-+ p_IpcReply->error = E_OK;
-+ *p_ReplyLength = sizeof(uint32_t) + sizeof(uint32_t);
-+ break;
-+ }
-+ case (FM_PCD_ALLOC_KG_SCHEMES):
-+ {
-+ t_FmPcdIpcKgSchemesParams ipcSchemesParams;
-+
-+ memcpy((uint8_t*)&ipcSchemesParams, p_IpcMsg->msgBody, sizeof(t_FmPcdIpcKgSchemesParams));
-+ err = FmPcdKgAllocSchemes(h_FmPcd,
-+ ipcSchemesParams.numOfSchemes,
-+ ipcSchemesParams.guestId,
-+ p_IpcReply->replyBody);
-+ p_IpcReply->error = err;
-+ *p_ReplyLength = sizeof(uint32_t) + ipcSchemesParams.numOfSchemes*sizeof(uint8_t);
-+ break;
-+ }
-+ case (FM_PCD_FREE_KG_SCHEMES):
-+ {
-+ t_FmPcdIpcKgSchemesParams ipcSchemesParams;
-+
-+ memcpy((uint8_t*)&ipcSchemesParams, p_IpcMsg->msgBody, sizeof(t_FmPcdIpcKgSchemesParams));
-+ err = FmPcdKgFreeSchemes(h_FmPcd,
-+ ipcSchemesParams.numOfSchemes,
-+ ipcSchemesParams.guestId,
-+ ipcSchemesParams.schemesIds);
-+ p_IpcReply->error = err;
-+ *p_ReplyLength = sizeof(uint32_t);
-+ break;
-+ }
-+ case (FM_PCD_ALLOC_KG_CLSPLAN):
-+ {
-+ t_FmPcdIpcKgClsPlanParams ipcKgClsPlanParams;
-+
-+ memcpy((uint8_t*)&ipcKgClsPlanParams, p_IpcMsg->msgBody, sizeof(t_FmPcdIpcKgClsPlanParams));
-+ err = KgAllocClsPlanEntries(h_FmPcd,
-+ ipcKgClsPlanParams.numOfClsPlanEntries,
-+ ipcKgClsPlanParams.guestId,
-+ p_IpcReply->replyBody);
-+ p_IpcReply->error = err;
-+ *p_ReplyLength = sizeof(uint32_t) + sizeof(uint8_t);
-+ break;
-+ }
-+ case (FM_PCD_FREE_KG_CLSPLAN):
-+ {
-+ t_FmPcdIpcKgClsPlanParams ipcKgClsPlanParams;
-+
-+ memcpy((uint8_t*)&ipcKgClsPlanParams, p_IpcMsg->msgBody, sizeof(t_FmPcdIpcKgClsPlanParams));
-+ KgFreeClsPlanEntries(h_FmPcd,
-+ ipcKgClsPlanParams.numOfClsPlanEntries,
-+ ipcKgClsPlanParams.guestId,
-+ ipcKgClsPlanParams.clsPlanBase);
-+ *p_ReplyLength = sizeof(uint32_t);
-+ break;
-+ }
-+ case (FM_PCD_ALLOC_PROFILES):
-+ {
-+ t_FmIpcResourceAllocParams ipcAllocParams;
-+ uint16_t base;
-+ memcpy(&ipcAllocParams, p_IpcMsg->msgBody, sizeof(t_FmIpcResourceAllocParams));
-+ base = PlcrAllocProfilesForPartition(h_FmPcd,
-+ ipcAllocParams.base,
-+ ipcAllocParams.num,
-+ ipcAllocParams.guestId);
-+ memcpy(p_IpcReply->replyBody, (uint16_t*)&base, sizeof(uint16_t));
-+ *p_ReplyLength = sizeof(uint32_t) + sizeof(uint16_t);
-+ break;
-+ }
-+ case (FM_PCD_FREE_PROFILES):
-+ {
-+ t_FmIpcResourceAllocParams ipcAllocParams;
-+ memcpy(&ipcAllocParams, p_IpcMsg->msgBody, sizeof(t_FmIpcResourceAllocParams));
-+ PlcrFreeProfilesForPartition(h_FmPcd,
-+ ipcAllocParams.base,
-+ ipcAllocParams.num,
-+ ipcAllocParams.guestId);
-+ break;
-+ }
-+ case (FM_PCD_SET_PORT_PROFILES):
-+ {
-+ t_FmIpcResourceAllocParams ipcAllocParams;
-+ memcpy(&ipcAllocParams, p_IpcMsg->msgBody, sizeof(t_FmIpcResourceAllocParams));
-+ PlcrSetPortProfiles(h_FmPcd,
-+ ipcAllocParams.guestId,
-+ ipcAllocParams.num,
-+ ipcAllocParams.base);
-+ break;
-+ }
-+ case (FM_PCD_CLEAR_PORT_PROFILES):
-+ {
-+ t_FmIpcResourceAllocParams ipcAllocParams;
-+ memcpy(&ipcAllocParams, p_IpcMsg->msgBody, sizeof(t_FmIpcResourceAllocParams));
-+ PlcrClearPortProfiles(h_FmPcd,
-+ ipcAllocParams.guestId);
-+ break;
-+ }
-+ case (FM_PCD_GET_SW_PRS_OFFSET):
-+ {
-+ t_FmPcdIpcSwPrsLable ipcSwPrsLable;
-+ uint32_t swPrsOffset;
-+
-+ memcpy((uint8_t*)&ipcSwPrsLable, p_IpcMsg->msgBody, sizeof(t_FmPcdIpcSwPrsLable));
-+ swPrsOffset =
-+ FmPcdGetSwPrsOffset(h_FmPcd,
-+ (e_NetHeaderType)ipcSwPrsLable.enumHdr,
-+ ipcSwPrsLable.indexPerHdr);
-+ memcpy(p_IpcReply->replyBody, (uint8_t*)&swPrsOffset, sizeof(uint32_t));
-+ *p_ReplyLength = sizeof(uint32_t) + sizeof(uint32_t);
-+ break;
-+ }
-+ case (FM_PCD_PRS_INC_PORT_STATS):
-+ {
-+ t_FmPcdIpcPrsIncludePort ipcPrsIncludePort;
-+
-+ memcpy((uint8_t*)&ipcPrsIncludePort, p_IpcMsg->msgBody, sizeof(t_FmPcdIpcPrsIncludePort));
-+ PrsIncludePortInStatistics(h_FmPcd,
-+ ipcPrsIncludePort.hardwarePortId,
-+ ipcPrsIncludePort.include);
-+ break;
-+ }
-+ default:
-+ *p_ReplyLength = 0;
-+ RETURN_ERROR(MINOR, E_INVALID_SELECTION, ("command not found!!!"));
-+ }
-+ return E_OK;
-+}
-+
-+static uint32_t NetEnvLock(t_Handle h_NetEnv)
-+{
-+ ASSERT_COND(h_NetEnv);
-+ return XX_LockIntrSpinlock(((t_FmPcdNetEnv*)h_NetEnv)->h_Spinlock);
-+}
-+
-+static void NetEnvUnlock(t_Handle h_NetEnv, uint32_t intFlags)
-+{
-+ ASSERT_COND(h_NetEnv);
-+ XX_UnlockIntrSpinlock(((t_FmPcdNetEnv*)h_NetEnv)->h_Spinlock, intFlags);
-+}
-+
-+static void EnqueueLockToFreeLst(t_FmPcd *p_FmPcd, t_FmPcdLock *p_Lock)
-+{
-+ uint32_t intFlags;
-+
-+ intFlags = XX_LockIntrSpinlock(p_FmPcd->h_Spinlock);
-+ LIST_AddToTail(&p_Lock->node, &p_FmPcd->freeLocksLst);
-+ XX_UnlockIntrSpinlock(p_FmPcd->h_Spinlock, intFlags);
-+}
-+
-+static t_FmPcdLock * DequeueLockFromFreeLst(t_FmPcd *p_FmPcd)
-+{
-+ t_FmPcdLock *p_Lock = NULL;
-+ uint32_t intFlags;
-+
-+ intFlags = XX_LockIntrSpinlock(p_FmPcd->h_Spinlock);
-+ if (!LIST_IsEmpty(&p_FmPcd->freeLocksLst))
-+ {
-+ p_Lock = FM_PCD_LOCK_OBJ(p_FmPcd->freeLocksLst.p_Next);
-+ LIST_DelAndInit(&p_Lock->node);
-+ }
-+ if (p_FmPcd->h_Spinlock)
-+ XX_UnlockIntrSpinlock(p_FmPcd->h_Spinlock, intFlags);
-+
-+ return p_Lock;
-+}
-+
-+static void EnqueueLockToAcquiredLst(t_FmPcd *p_FmPcd, t_FmPcdLock *p_Lock)
-+{
-+ uint32_t intFlags;
-+
-+ intFlags = XX_LockIntrSpinlock(p_FmPcd->h_Spinlock);
-+ LIST_AddToTail(&p_Lock->node, &p_FmPcd->acquiredLocksLst);
-+ XX_UnlockIntrSpinlock(p_FmPcd->h_Spinlock, intFlags);
-+}
-+
-+static t_Error FillFreeLocksLst(t_FmPcd *p_FmPcd)
-+{
-+ t_FmPcdLock *p_Lock;
-+ int i;
-+
-+ for (i=0; i<10; i++)
-+ {
-+ p_Lock = (t_FmPcdLock *)XX_Malloc(sizeof(t_FmPcdLock));
-+ if (!p_Lock)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("FM-PCD lock obj!"));
-+ memset(p_Lock, 0, sizeof(t_FmPcdLock));
-+ INIT_LIST(&p_Lock->node);
-+ p_Lock->h_Spinlock = XX_InitSpinlock();
-+ if (!p_Lock->h_Spinlock)
-+ {
-+ XX_Free(p_Lock);
-+ RETURN_ERROR(MINOR, E_INVALID_STATE, ("FM-PCD spinlock obj!"));
-+ }
-+ EnqueueLockToFreeLst(p_FmPcd, p_Lock);
-+ }
-+
-+ return E_OK;
-+}
-+
-+static void ReleaseFreeLocksLst(t_FmPcd *p_FmPcd)
-+{
-+ t_FmPcdLock *p_Lock;
-+
-+ p_Lock = DequeueLockFromFreeLst(p_FmPcd);
-+ while (p_Lock)
-+ {
-+ XX_FreeSpinlock(p_Lock->h_Spinlock);
-+ XX_Free(p_Lock);
-+ p_Lock = DequeueLockFromFreeLst(p_FmPcd);
-+ }
-+}
-+
-+
-+
-+/*****************************************************************************/
-+/* Inter-module API routines */
-+/*****************************************************************************/
-+
-+void FmPcdSetClsPlanGrpId(t_FmPcd *p_FmPcd, uint8_t netEnvId, uint8_t clsPlanGrpId)
-+{
-+ ASSERT_COND(p_FmPcd);
-+ p_FmPcd->netEnvs[netEnvId].clsPlanGrpId = clsPlanGrpId;
-+}
-+
-+t_Error PcdGetClsPlanGrpParams(t_FmPcd *p_FmPcd, t_FmPcdKgInterModuleClsPlanGrpParams *p_GrpParams)
-+{
-+ uint8_t netEnvId = p_GrpParams->netEnvId;
-+ int i, k, j;
-+
-+ ASSERT_COND(p_FmPcd);
-+ if (p_FmPcd->netEnvs[netEnvId].clsPlanGrpId != ILLEGAL_CLS_PLAN)
-+ {
-+ p_GrpParams->grpExists = TRUE;
-+ p_GrpParams->clsPlanGrpId = p_FmPcd->netEnvs[netEnvId].clsPlanGrpId;
-+ return E_OK;
-+ }
-+
-+ for (i=0; ((i < FM_PCD_MAX_NUM_OF_DISTINCTION_UNITS) &&
-+ (p_FmPcd->netEnvs[netEnvId].units[i].hdrs[0].hdr != HEADER_TYPE_NONE)); i++)
-+ {
-+ for (k=0; ((k < FM_PCD_MAX_NUM_OF_INTERCHANGEABLE_HDRS) &&
-+ (p_FmPcd->netEnvs[netEnvId].units[i].hdrs[k].hdr != HEADER_TYPE_NONE)); k++)
-+ {
-+ /* if an option exists, add it to the opts list */
-+ if (p_FmPcd->netEnvs[netEnvId].units[i].hdrs[k].opt)
-+ {
-+ /* check if this option already exists, add if it doesn't */
-+ for (j = 0;j<p_GrpParams->numOfOptions;j++)
-+ {
-+ if (p_GrpParams->options[j] == p_FmPcd->netEnvs[netEnvId].units[i].hdrs[k].opt)
-+ break;
-+ }
-+ p_GrpParams->optVectors[j] |= p_FmPcd->netEnvs[netEnvId].unitsVectors[i];
-+ if (j == p_GrpParams->numOfOptions)
-+ {
-+ p_GrpParams->options[p_GrpParams->numOfOptions] = p_FmPcd->netEnvs[netEnvId].units[i].hdrs[k].opt;
-+ p_GrpParams->numOfOptions++;
-+ }
-+ }
-+ }
-+ }
-+
-+ if (p_GrpParams->numOfOptions == 0)
-+ {
-+ if (p_FmPcd->p_FmPcdKg->emptyClsPlanGrpId != ILLEGAL_CLS_PLAN)
-+ {
-+ p_GrpParams->grpExists = TRUE;
-+ p_GrpParams->clsPlanGrpId = p_FmPcd->p_FmPcdKg->emptyClsPlanGrpId;
-+ }
-+ }
-+
-+ return E_OK;
-+
-+}
-+
-+t_Error PcdGetVectorForOpt(t_FmPcd *p_FmPcd, uint8_t netEnvId, protocolOpt_t opt, uint32_t *p_Vector)
-+{
-+ uint8_t j,k;
-+
-+ *p_Vector = 0;
-+
-+ ASSERT_COND(p_FmPcd);
-+ for (j=0; ((j < FM_PCD_MAX_NUM_OF_DISTINCTION_UNITS) &&
-+ (p_FmPcd->netEnvs[netEnvId].units[j].hdrs[0].hdr != HEADER_TYPE_NONE)); j++)
-+ {
-+ for (k=0; ((k < FM_PCD_MAX_NUM_OF_INTERCHANGEABLE_HDRS) &&
-+ (p_FmPcd->netEnvs[netEnvId].units[j].hdrs[k].hdr != HEADER_TYPE_NONE)); k++)
-+ {
-+ if (p_FmPcd->netEnvs[netEnvId].units[j].hdrs[k].opt == opt)
-+ *p_Vector |= p_FmPcd->netEnvs[netEnvId].unitsVectors[j];
-+ }
-+ }
-+
-+ if (!*p_Vector)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("Requested option was not defined for this Network Environment Characteristics module"));
-+ else
-+ return E_OK;
-+}
-+
-+t_Error PcdGetUnitsVector(t_FmPcd *p_FmPcd, t_NetEnvParams *p_Params)
-+{
-+ int i;
-+
-+ ASSERT_COND(p_FmPcd);
-+ ASSERT_COND(p_Params->netEnvId < FM_MAX_NUM_OF_PORTS);
-+
-+ p_Params->vector = 0;
-+ for (i=0; i<p_Params->numOfDistinctionUnits ;i++)
-+ {
-+ if (p_FmPcd->netEnvs[p_Params->netEnvId].units[p_Params->unitIds[i]].hdrs[0].hdr == HEADER_TYPE_NONE)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("Requested unit was not defined for this Network Environment Characteristics module"));
-+ ASSERT_COND(p_FmPcd->netEnvs[p_Params->netEnvId].unitsVectors[p_Params->unitIds[i]]);
-+ p_Params->vector |= p_FmPcd->netEnvs[p_Params->netEnvId].unitsVectors[p_Params->unitIds[i]];
-+ }
-+
-+ return E_OK;
-+}
-+
-+bool PcdNetEnvIsUnitWithoutOpts(t_FmPcd *p_FmPcd, uint8_t netEnvId, uint32_t unitVector)
-+{
-+ int i=0, k;
-+
-+ ASSERT_COND(p_FmPcd);
-+ /* check whether a given unit may be used by non-clsPlan users. */
-+ /* first, recognize the unit by its vector */
-+ while (p_FmPcd->netEnvs[netEnvId].units[i].hdrs[0].hdr != HEADER_TYPE_NONE)
-+ {
-+ if (p_FmPcd->netEnvs[netEnvId].unitsVectors[i] == unitVector)
-+ {
-+ for (k=0;
-+ ((k < FM_PCD_MAX_NUM_OF_INTERCHANGEABLE_HDRS) &&
-+ (p_FmPcd->netEnvs[netEnvId].units[i].hdrs[k].hdr != HEADER_TYPE_NONE));
-+ k++)
-+ /* check that no option exists */
-+ if ((protocolOpt_t)p_FmPcd->netEnvs[netEnvId].units[i].hdrs[k].opt)
-+ return FALSE;
-+ break;
-+ }
-+ i++;
-+ }
-+ /* assert that a unit was found to mach the vector */
-+ ASSERT_COND(p_FmPcd->netEnvs[netEnvId].units[i].hdrs[0].hdr != HEADER_TYPE_NONE);
-+
-+ return TRUE;
-+}
-+bool FmPcdNetEnvIsHdrExist(t_Handle h_FmPcd, uint8_t netEnvId, e_NetHeaderType hdr)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+ int i, k;
-+
-+ ASSERT_COND(p_FmPcd);
-+
-+ for (i=0; ((i < FM_PCD_MAX_NUM_OF_DISTINCTION_UNITS) &&
-+ (p_FmPcd->netEnvs[netEnvId].units[i].hdrs[0].hdr != HEADER_TYPE_NONE)); i++)
-+ {
-+ for (k=0; ((k < FM_PCD_MAX_NUM_OF_INTERCHANGEABLE_HDRS) &&
-+ (p_FmPcd->netEnvs[netEnvId].units[i].hdrs[k].hdr != HEADER_TYPE_NONE)); k++)
-+ if (p_FmPcd->netEnvs[netEnvId].units[i].hdrs[k].hdr == hdr)
-+ return TRUE;
-+ }
-+ for (i=0; ((i < FM_PCD_MAX_NUM_OF_ALIAS_HDRS) &&
-+ (p_FmPcd->netEnvs[netEnvId].aliasHdrs[i].hdr != HEADER_TYPE_NONE)); i++)
-+ {
-+ if (p_FmPcd->netEnvs[netEnvId].aliasHdrs[i].hdr == hdr)
-+ return TRUE;
-+ }
-+
-+ return FALSE;
-+}
-+
-+uint8_t FmPcdNetEnvGetUnitId(t_FmPcd *p_FmPcd, uint8_t netEnvId, e_NetHeaderType hdr, bool interchangeable, protocolOpt_t opt)
-+{
-+ uint8_t i, k;
-+
-+ ASSERT_COND(p_FmPcd);
-+
-+ if (interchangeable)
-+ {
-+ for (i=0; (i < FM_PCD_MAX_NUM_OF_DISTINCTION_UNITS) &&
-+ (p_FmPcd->netEnvs[netEnvId].units[i].hdrs[0].hdr != HEADER_TYPE_NONE); i++)
-+ {
-+ for (k=0; (k < FM_PCD_MAX_NUM_OF_INTERCHANGEABLE_HDRS) &&
-+ (p_FmPcd->netEnvs[netEnvId].units[i].hdrs[k].hdr != HEADER_TYPE_NONE); k++)
-+ {
-+ if ((p_FmPcd->netEnvs[netEnvId].units[i].hdrs[k].hdr == hdr) &&
-+ (p_FmPcd->netEnvs[netEnvId].units[i].hdrs[k].opt == opt))
-+
-+ return i;
-+ }
-+ }
-+ }
-+ else
-+ {
-+ for (i=0; (i < FM_PCD_MAX_NUM_OF_DISTINCTION_UNITS) &&
-+ (p_FmPcd->netEnvs[netEnvId].units[i].hdrs[0].hdr != HEADER_TYPE_NONE); i++)
-+ if ((p_FmPcd->netEnvs[netEnvId].units[i].hdrs[0].hdr == hdr) &&
-+ (p_FmPcd->netEnvs[netEnvId].units[i].hdrs[0].opt == opt) &&
-+ (p_FmPcd->netEnvs[netEnvId].units[i].hdrs[1].hdr == HEADER_TYPE_NONE))
-+ return i;
-+
-+ for (i=0; (i < FM_PCD_MAX_NUM_OF_ALIAS_HDRS) &&
-+ (p_FmPcd->netEnvs[netEnvId].aliasHdrs[i].hdr != HEADER_TYPE_NONE); i++)
-+ if ((p_FmPcd->netEnvs[netEnvId].aliasHdrs[i].hdr == hdr) &&
-+ (p_FmPcd->netEnvs[netEnvId].aliasHdrs[i].opt == opt))
-+ return p_FmPcd->netEnvs[netEnvId].aliasHdrs[i].aliasHdr;
-+ }
-+
-+ return FM_PCD_MAX_NUM_OF_DISTINCTION_UNITS;
-+}
-+
-+t_Error FmPcdUnregisterReassmPort(t_Handle h_FmPcd, t_Handle h_ReasmCommonPramTbl)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+ t_FmPcdCcReassmTimeoutParams ccReassmTimeoutParams = {0};
-+ uint8_t result;
-+ t_Error err = E_OK;
-+
-+ ASSERT_COND(p_FmPcd);
-+ ASSERT_COND(h_ReasmCommonPramTbl);
-+
-+ ccReassmTimeoutParams.iprcpt = (uint32_t)(XX_VirtToPhys(h_ReasmCommonPramTbl) - p_FmPcd->physicalMuramBase);
-+ ccReassmTimeoutParams.activate = FALSE; /*Disable Timeout Task*/
-+
-+ if ((err = FmHcPcdCcTimeoutReassm(p_FmPcd->h_Hc, &ccReassmTimeoutParams, &result)) != E_OK)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+
-+ switch (result)
-+ {
-+ case (0):
-+ return E_OK;
-+ case (1):
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, (""));
-+ case (2):
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, (""));
-+ case (3):
-+ RETURN_ERROR(MAJOR, E_INVALID_HANDLE, ("Disable Timeout Task with invalid IPRCPT"));
-+ default:
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, NO_MSG);
-+ }
-+
-+ return E_OK;
-+}
-+
-+e_NetHeaderType FmPcdGetAliasHdr(t_FmPcd *p_FmPcd, uint8_t netEnvId, e_NetHeaderType hdr)
-+{
-+ int i;
-+
-+ ASSERT_COND(p_FmPcd);
-+ ASSERT_COND(netEnvId < FM_MAX_NUM_OF_PORTS);
-+
-+ for (i=0; (i < FM_PCD_MAX_NUM_OF_ALIAS_HDRS)
-+ && (p_FmPcd->netEnvs[netEnvId].aliasHdrs[i].hdr != HEADER_TYPE_NONE); i++)
-+ {
-+ if (p_FmPcd->netEnvs[netEnvId].aliasHdrs[i].hdr == hdr)
-+ return p_FmPcd->netEnvs[netEnvId].aliasHdrs[i].aliasHdr;
-+ }
-+
-+ return HEADER_TYPE_NONE;
-+}
-+
-+void FmPcdPortRegister(t_Handle h_FmPcd, t_Handle h_FmPort, uint8_t hardwarePortId)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+ uint16_t swPortIndex = 0;
-+
-+ ASSERT_COND(h_FmPcd);
-+ HW_PORT_ID_TO_SW_PORT_INDX(swPortIndex, hardwarePortId);
-+ p_FmPcd->p_FmPcdPlcr->portsMapping[swPortIndex].h_FmPort = h_FmPort;
-+}
-+
-+uint32_t FmPcdGetLcv(t_Handle h_FmPcd, uint32_t netEnvId, uint8_t hdrNum)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+
-+ ASSERT_COND(h_FmPcd);
-+ return p_FmPcd->netEnvs[netEnvId].lcvs[hdrNum];
-+}
-+
-+uint32_t FmPcdGetMacsecLcv(t_Handle h_FmPcd, uint32_t netEnvId)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+
-+ ASSERT_COND(h_FmPcd);
-+ return p_FmPcd->netEnvs[netEnvId].macsecVector;
-+}
-+
-+uint8_t FmPcdGetNetEnvId(t_Handle h_NetEnv)
-+{
-+ return ((t_FmPcdNetEnv*)h_NetEnv)->netEnvId;
-+}
-+
-+void FmPcdIncNetEnvOwners(t_Handle h_FmPcd, uint8_t netEnvId)
-+{
-+ uint32_t intFlags;
-+
-+ ASSERT_COND(h_FmPcd);
-+
-+ intFlags = NetEnvLock(&((t_FmPcd*)h_FmPcd)->netEnvs[netEnvId]);
-+ ((t_FmPcd*)h_FmPcd)->netEnvs[netEnvId].owners++;
-+ NetEnvUnlock(&((t_FmPcd*)h_FmPcd)->netEnvs[netEnvId], intFlags);
-+}
-+
-+void FmPcdDecNetEnvOwners(t_Handle h_FmPcd, uint8_t netEnvId)
-+{
-+ uint32_t intFlags;
-+
-+ ASSERT_COND(h_FmPcd);
-+ ASSERT_COND(((t_FmPcd*)h_FmPcd)->netEnvs[netEnvId].owners);
-+
-+ intFlags = NetEnvLock(&((t_FmPcd*)h_FmPcd)->netEnvs[netEnvId]);
-+ ((t_FmPcd*)h_FmPcd)->netEnvs[netEnvId].owners--;
-+ NetEnvUnlock(&((t_FmPcd*)h_FmPcd)->netEnvs[netEnvId], intFlags);
-+}
-+
-+uint32_t FmPcdLock(t_Handle h_FmPcd)
-+{
-+ ASSERT_COND(h_FmPcd);
-+ return XX_LockIntrSpinlock(((t_FmPcd*)h_FmPcd)->h_Spinlock);
-+}
-+
-+void FmPcdUnlock(t_Handle h_FmPcd, uint32_t intFlags)
-+{
-+ ASSERT_COND(h_FmPcd);
-+ XX_UnlockIntrSpinlock(((t_FmPcd*)h_FmPcd)->h_Spinlock, intFlags);
-+}
-+
-+t_FmPcdLock * FmPcdAcquireLock(t_Handle h_FmPcd)
-+{
-+ t_FmPcdLock *p_Lock;
-+ ASSERT_COND(h_FmPcd);
-+ p_Lock = DequeueLockFromFreeLst((t_FmPcd*)h_FmPcd);
-+ if (!p_Lock)
-+ {
-+ FillFreeLocksLst(h_FmPcd);
-+ p_Lock = DequeueLockFromFreeLst((t_FmPcd*)h_FmPcd);
-+ }
-+
-+ if (p_Lock)
-+ EnqueueLockToAcquiredLst((t_FmPcd*)h_FmPcd, p_Lock);
-+ return p_Lock;
-+}
-+
-+void FmPcdReleaseLock(t_Handle h_FmPcd, t_FmPcdLock *p_Lock)
-+{
-+ uint32_t intFlags;
-+ ASSERT_COND(h_FmPcd);
-+ intFlags = FmPcdLock(h_FmPcd);
-+ LIST_DelAndInit(&p_Lock->node);
-+ FmPcdUnlock(h_FmPcd, intFlags);
-+ EnqueueLockToFreeLst((t_FmPcd*)h_FmPcd, p_Lock);
-+}
-+
-+bool FmPcdLockTryLockAll(t_Handle h_FmPcd)
-+{
-+ uint32_t intFlags;
-+ t_List *p_Pos, *p_SavedPos=NULL;
-+
-+ ASSERT_COND(h_FmPcd);
-+ intFlags = FmPcdLock(h_FmPcd);
-+ LIST_FOR_EACH(p_Pos, &((t_FmPcd*)h_FmPcd)->acquiredLocksLst)
-+ {
-+ t_FmPcdLock *p_Lock = FM_PCD_LOCK_OBJ(p_Pos);
-+ if (!FmPcdLockTryLock(p_Lock))
-+ {
-+ p_SavedPos = p_Pos;
-+ break;
-+ }
-+ }
-+ if (p_SavedPos)
-+ {
-+ LIST_FOR_EACH(p_Pos, &((t_FmPcd*)h_FmPcd)->acquiredLocksLst)
-+ {
-+ t_FmPcdLock *p_Lock = FM_PCD_LOCK_OBJ(p_Pos);
-+ if (p_Pos == p_SavedPos)
-+ break;
-+ FmPcdLockUnlock(p_Lock);
-+ }
-+ }
-+ FmPcdUnlock(h_FmPcd, intFlags);
-+
-+ CORE_MemoryBarrier();
-+
-+ if (p_SavedPos)
-+ return FALSE;
-+
-+ return TRUE;
-+}
-+
-+void FmPcdLockUnlockAll(t_Handle h_FmPcd)
-+{
-+ uint32_t intFlags;
-+ t_List *p_Pos;
-+
-+ ASSERT_COND(h_FmPcd);
-+ intFlags = FmPcdLock(h_FmPcd);
-+ LIST_FOR_EACH(p_Pos, &((t_FmPcd*)h_FmPcd)->acquiredLocksLst)
-+ {
-+ t_FmPcdLock *p_Lock = FM_PCD_LOCK_OBJ(p_Pos);
-+ p_Lock->flag = FALSE;
-+ }
-+ FmPcdUnlock(h_FmPcd, intFlags);
-+
-+ CORE_MemoryBarrier();
-+}
-+
-+t_Error FmPcdHcSync(t_Handle h_FmPcd)
-+{
-+ ASSERT_COND(h_FmPcd);
-+ ASSERT_COND(((t_FmPcd*)h_FmPcd)->h_Hc);
-+
-+ return FmHcPcdSync(((t_FmPcd*)h_FmPcd)->h_Hc);
-+}
-+
-+t_Handle FmPcdGetHcHandle(t_Handle h_FmPcd)
-+{
-+ ASSERT_COND(h_FmPcd);
-+ return ((t_FmPcd*)h_FmPcd)->h_Hc;
-+}
-+
-+bool FmPcdIsAdvancedOffloadSupported(t_Handle h_FmPcd)
-+{
-+ ASSERT_COND(h_FmPcd);
-+ return ((t_FmPcd*)h_FmPcd)->advancedOffloadSupport;
-+}
-+/*********************** End of inter-module routines ************************/
-+
-+
-+/****************************************/
-+/* API Init unit functions */
-+/****************************************/
-+
-+t_Handle FM_PCD_Config(t_FmPcdParams *p_FmPcdParams)
-+{
-+ t_FmPcd *p_FmPcd = NULL;
-+ t_FmPhysAddr physicalMuramBase;
-+ uint8_t i;
-+
-+ SANITY_CHECK_RETURN_VALUE(p_FmPcdParams, E_INVALID_HANDLE,NULL);
-+
-+ p_FmPcd = (t_FmPcd *) XX_Malloc(sizeof(t_FmPcd));
-+ if (!p_FmPcd)
-+ {
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("FM PCD"));
-+ return NULL;
-+ }
-+ memset(p_FmPcd, 0, sizeof(t_FmPcd));
-+
-+ p_FmPcd->p_FmPcdDriverParam = (t_FmPcdDriverParam *) XX_Malloc(sizeof(t_FmPcdDriverParam));
-+ if (!p_FmPcd->p_FmPcdDriverParam)
-+ {
-+ XX_Free(p_FmPcd);
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("FM PCD Driver Param"));
-+ return NULL;
-+ }
-+ memset(p_FmPcd->p_FmPcdDriverParam, 0, sizeof(t_FmPcdDriverParam));
-+
-+ p_FmPcd->h_Fm = p_FmPcdParams->h_Fm;
-+ p_FmPcd->guestId = FmGetGuestId(p_FmPcd->h_Fm);
-+ p_FmPcd->h_FmMuram = FmGetMuramHandle(p_FmPcd->h_Fm);
-+ if (p_FmPcd->h_FmMuram)
-+ {
-+ FmGetPhysicalMuramBase(p_FmPcdParams->h_Fm, &physicalMuramBase);
-+ p_FmPcd->physicalMuramBase = (uint64_t)((uint64_t)(&physicalMuramBase)->low | ((uint64_t)(&physicalMuramBase)->high << 32));
-+ }
-+
-+ for (i = 0; i<FM_MAX_NUM_OF_PORTS; i++)
-+ p_FmPcd->netEnvs[i].clsPlanGrpId = ILLEGAL_CLS_PLAN;
-+
-+ if (p_FmPcdParams->useHostCommand)
-+ {
-+ t_FmHcParams hcParams;
-+
-+ memset(&hcParams, 0, sizeof(hcParams));
-+ hcParams.h_Fm = p_FmPcd->h_Fm;
-+ hcParams.h_FmPcd = (t_Handle)p_FmPcd;
-+ memcpy((uint8_t*)&hcParams.params, (uint8_t*)&p_FmPcdParams->hc, sizeof(t_FmPcdHcParams));
-+ p_FmPcd->h_Hc = FmHcConfigAndInit(&hcParams);
-+ if (!p_FmPcd->h_Hc)
-+ {
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("FM PCD HC"));
-+ FM_PCD_Free(p_FmPcd);
-+ return NULL;
-+ }
-+ }
-+ else if (p_FmPcd->guestId != NCSW_MASTER_ID)
-+ REPORT_ERROR(MAJOR, E_INVALID_STATE, ("No Host Command defined for a guest partition."));
-+
-+ if (p_FmPcdParams->kgSupport)
-+ {
-+ p_FmPcd->p_FmPcdKg = (t_FmPcdKg *)KgConfig(p_FmPcd, p_FmPcdParams);
-+ if (!p_FmPcd->p_FmPcdKg)
-+ {
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("FM PCD Keygen"));
-+ FM_PCD_Free(p_FmPcd);
-+ return NULL;
-+ }
-+ }
-+
-+ if (p_FmPcdParams->plcrSupport)
-+ {
-+ p_FmPcd->p_FmPcdPlcr = (t_FmPcdPlcr *)PlcrConfig(p_FmPcd, p_FmPcdParams);
-+ if (!p_FmPcd->p_FmPcdPlcr)
-+ {
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("FM PCD Policer"));
-+ FM_PCD_Free(p_FmPcd);
-+ return NULL;
-+ }
-+ }
-+
-+ if (p_FmPcdParams->prsSupport)
-+ {
-+ p_FmPcd->p_FmPcdPrs = (t_FmPcdPrs *)PrsConfig(p_FmPcd, p_FmPcdParams);
-+ if (!p_FmPcd->p_FmPcdPrs)
-+ {
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("FM PCD Parser"));
-+ FM_PCD_Free(p_FmPcd);
-+ return NULL;
-+ }
-+ }
-+
-+ p_FmPcd->h_Spinlock = XX_InitSpinlock();
-+ if (!p_FmPcd->h_Spinlock)
-+ {
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("FM PCD spinlock"));
-+ FM_PCD_Free(p_FmPcd);
-+ return NULL;
-+ }
-+ INIT_LIST(&p_FmPcd->freeLocksLst);
-+ INIT_LIST(&p_FmPcd->acquiredLocksLst);
-+
-+ p_FmPcd->numOfEnabledGuestPartitionsPcds = 0;
-+
-+ p_FmPcd->f_Exception = p_FmPcdParams->f_Exception;
-+ p_FmPcd->f_FmPcdIndexedException = p_FmPcdParams->f_ExceptionId;
-+ p_FmPcd->h_App = p_FmPcdParams->h_App;
-+
-+ p_FmPcd->p_CcShadow = NULL;
-+ p_FmPcd->ccShadowSize = 0;
-+ p_FmPcd->ccShadowAlign = 0;
-+
-+ p_FmPcd->h_ShadowSpinlock = XX_InitSpinlock();
-+ if (!p_FmPcd->h_ShadowSpinlock)
-+ {
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("FM PCD shadow spinlock"));
-+ FM_PCD_Free(p_FmPcd);
-+ return NULL;
-+ }
-+
-+ return p_FmPcd;
-+}
-+
-+t_Error FM_PCD_Init(t_Handle h_FmPcd)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+ t_Error err = E_OK;
-+ t_FmPcdIpcMsg msg;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd->p_FmPcdDriverParam, E_INVALID_HANDLE);
-+
-+ FM_GetRevision(p_FmPcd->h_Fm, &p_FmPcd->fmRevInfo);
-+
-+ if (p_FmPcd->guestId != NCSW_MASTER_ID)
-+ {
-+ memset(p_FmPcd->fmPcdIpcHandlerModuleName, 0, (sizeof(char)) * MODULE_NAME_SIZE);
-+ if (Sprint (p_FmPcd->fmPcdIpcHandlerModuleName, "FM_PCD_%d_%d", FmGetId(p_FmPcd->h_Fm), NCSW_MASTER_ID) != 10)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("Sprint failed"));
-+ memset(p_FmPcd->fmPcdModuleName, 0, (sizeof(char)) * MODULE_NAME_SIZE);
-+ if (Sprint (p_FmPcd->fmPcdModuleName, "FM_PCD_%d_%d",FmGetId(p_FmPcd->h_Fm), p_FmPcd->guestId) != (p_FmPcd->guestId<10 ? 10:11))
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("Sprint failed"));
-+
-+ p_FmPcd->h_IpcSession = XX_IpcInitSession(p_FmPcd->fmPcdIpcHandlerModuleName, p_FmPcd->fmPcdModuleName);
-+ if (p_FmPcd->h_IpcSession)
-+ {
-+ t_FmPcdIpcReply reply;
-+ uint32_t replyLength;
-+ uint8_t isMasterAlive = 0;
-+
-+ memset(&msg, 0, sizeof(msg));
-+ memset(&reply, 0, sizeof(reply));
-+ msg.msgId = FM_PCD_MASTER_IS_ALIVE;
-+ msg.msgBody[0] = p_FmPcd->guestId;
-+ blockingFlag = TRUE;
-+
-+ do
-+ {
-+ replyLength = sizeof(uint32_t) + sizeof(isMasterAlive);
-+ if ((err = XX_IpcSendMessage(p_FmPcd->h_IpcSession,
-+ (uint8_t*)&msg,
-+ sizeof(msg.msgId)+sizeof(p_FmPcd->guestId),
-+ (uint8_t*)&reply,
-+ &replyLength,
-+ IpcMsgCompletionCB,
-+ h_FmPcd)) != E_OK)
-+ REPORT_ERROR(MAJOR, err, NO_MSG);
-+ while (blockingFlag) ;
-+ if (replyLength != (sizeof(uint32_t) + sizeof(isMasterAlive)))
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE, ("IPC reply length mismatch"));
-+ isMasterAlive = *(uint8_t*)(reply.replyBody);
-+ } while (!isMasterAlive);
-+ }
-+ }
-+
-+ CHECK_INIT_PARAMETERS(p_FmPcd, CheckFmPcdParameters);
-+
-+ if (p_FmPcd->p_FmPcdKg)
-+ {
-+ err = KgInit(p_FmPcd);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+
-+ if (p_FmPcd->p_FmPcdPlcr)
-+ {
-+ err = PlcrInit(p_FmPcd);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+
-+ if (p_FmPcd->p_FmPcdPrs)
-+ {
-+ err = PrsInit(p_FmPcd);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+
-+ if (p_FmPcd->guestId == NCSW_MASTER_ID)
-+ {
-+ /* register to inter-core messaging mechanism */
-+ memset(p_FmPcd->fmPcdModuleName, 0, (sizeof(char)) * MODULE_NAME_SIZE);
-+ if (Sprint (p_FmPcd->fmPcdModuleName, "FM_PCD_%d_%d",FmGetId(p_FmPcd->h_Fm),NCSW_MASTER_ID) != 10)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("Sprint failed"));
-+ err = XX_IpcRegisterMsgHandler(p_FmPcd->fmPcdModuleName, IpcMsgHandlerCB, p_FmPcd, FM_PCD_MAX_REPLY_SIZE);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+
-+ /* IPv6 Frame-Id used for fragmentation */
-+ p_FmPcd->ipv6FrameIdAddr = PTR_TO_UINT(FM_MURAM_AllocMem(p_FmPcd->h_FmMuram, 4, 4));
-+ if (!p_FmPcd->ipv6FrameIdAddr)
-+ {
-+ FM_PCD_Free(p_FmPcd);
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, ("MURAM allocation for IPv6 Frame-Id"));
-+ }
-+ IOMemSet32(UINT_TO_PTR(p_FmPcd->ipv6FrameIdAddr), 0, 4);
-+
-+ /* CAPWAP Frame-Id used for fragmentation */
-+ p_FmPcd->capwapFrameIdAddr = PTR_TO_UINT(FM_MURAM_AllocMem(p_FmPcd->h_FmMuram, 2, 4));
-+ if (!p_FmPcd->capwapFrameIdAddr)
-+ {
-+ FM_PCD_Free(p_FmPcd);
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, ("MURAM allocation for CAPWAP Frame-Id"));
-+ }
-+ IOMemSet32(UINT_TO_PTR(p_FmPcd->capwapFrameIdAddr), 0, 2);
-+
-+ XX_Free(p_FmPcd->p_FmPcdDriverParam);
-+ p_FmPcd->p_FmPcdDriverParam = NULL;
-+
-+ FmRegisterPcd(p_FmPcd->h_Fm, p_FmPcd);
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PCD_Free(t_Handle h_FmPcd)
-+{
-+ t_FmPcd *p_FmPcd =(t_FmPcd *)h_FmPcd;
-+ t_Error err = E_OK;
-+
-+ if (p_FmPcd->ipv6FrameIdAddr)
-+ FM_MURAM_FreeMem(p_FmPcd->h_FmMuram, UINT_TO_PTR(p_FmPcd->ipv6FrameIdAddr));
-+
-+ if (p_FmPcd->capwapFrameIdAddr)
-+ FM_MURAM_FreeMem(p_FmPcd->h_FmMuram, UINT_TO_PTR(p_FmPcd->capwapFrameIdAddr));
-+
-+ if (p_FmPcd->enabled)
-+ FM_PCD_Disable(p_FmPcd);
-+
-+ if (p_FmPcd->p_FmPcdDriverParam)
-+ {
-+ XX_Free(p_FmPcd->p_FmPcdDriverParam);
-+ p_FmPcd->p_FmPcdDriverParam = NULL;
-+ }
-+
-+ if (p_FmPcd->p_FmPcdKg)
-+ {
-+ if ((err = KgFree(p_FmPcd)) != E_OK)
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+ XX_Free(p_FmPcd->p_FmPcdKg);
-+ p_FmPcd->p_FmPcdKg = NULL;
-+ }
-+
-+ if (p_FmPcd->p_FmPcdPlcr)
-+ {
-+ PlcrFree(p_FmPcd);
-+ XX_Free(p_FmPcd->p_FmPcdPlcr);
-+ p_FmPcd->p_FmPcdPlcr = NULL;
-+ }
-+
-+ if (p_FmPcd->p_FmPcdPrs)
-+ {
-+ if (p_FmPcd->guestId == NCSW_MASTER_ID)
-+ PrsFree(p_FmPcd);
-+ XX_Free(p_FmPcd->p_FmPcdPrs);
-+ p_FmPcd->p_FmPcdPrs = NULL;
-+ }
-+
-+ if (p_FmPcd->h_Hc)
-+ {
-+ FmHcFree(p_FmPcd->h_Hc);
-+ p_FmPcd->h_Hc = NULL;
-+ }
-+
-+ XX_IpcUnregisterMsgHandler(p_FmPcd->fmPcdModuleName);
-+
-+ FmUnregisterPcd(p_FmPcd->h_Fm);
-+
-+ ReleaseFreeLocksLst(p_FmPcd);
-+
-+ if (p_FmPcd->h_Spinlock)
-+ XX_FreeSpinlock(p_FmPcd->h_Spinlock);
-+
-+ if (p_FmPcd->h_ShadowSpinlock)
-+ XX_FreeSpinlock(p_FmPcd->h_ShadowSpinlock);
-+
-+ XX_Free(p_FmPcd);
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PCD_ConfigException(t_Handle h_FmPcd, e_FmPcdExceptions exception, bool enable)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+ uint32_t bitMask = 0;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd, E_INVALID_HANDLE);
-+
-+ if (p_FmPcd->guestId != NCSW_MASTER_ID)
-+ RETURN_ERROR(MAJOR, E_NOT_SUPPORTED, ("FM_PCD_ConfigException - guest mode!"));
-+
-+ GET_FM_PCD_EXCEPTION_FLAG(bitMask, exception);
-+ if (bitMask)
-+ {
-+ if (enable)
-+ p_FmPcd->exceptions |= bitMask;
-+ else
-+ p_FmPcd->exceptions &= ~bitMask;
-+ }
-+ else
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("Undefined exception"));
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PCD_ConfigHcFramesDataMemory(t_Handle h_FmPcd, uint8_t memId)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd->h_Hc, E_INVALID_HANDLE);
-+
-+ return FmHcSetFramesDataMemory(p_FmPcd->h_Hc, memId);
-+}
-+
-+t_Error FM_PCD_Enable(t_Handle h_FmPcd)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+ t_Error err = E_OK;
-+
-+ SANITY_CHECK_RETURN_ERROR(h_FmPcd, E_INVALID_HANDLE);
-+
-+ if (p_FmPcd->enabled)
-+ return E_OK;
-+
-+ if ((p_FmPcd->guestId != NCSW_MASTER_ID) &&
-+ p_FmPcd->h_IpcSession)
-+ {
-+ uint8_t enabled;
-+ t_FmPcdIpcMsg msg;
-+ t_FmPcdIpcReply reply;
-+ uint32_t replyLength;
-+
-+ memset(&reply, 0, sizeof(reply));
-+ memset(&msg, 0, sizeof(msg));
-+ msg.msgId = FM_PCD_MASTER_IS_ENABLED;
-+ replyLength = sizeof(uint32_t) + sizeof(enabled);
-+ if ((err = XX_IpcSendMessage(p_FmPcd->h_IpcSession,
-+ (uint8_t*)&msg,
-+ sizeof(msg.msgId),
-+ (uint8_t*)&reply,
-+ &replyLength,
-+ NULL,
-+ NULL)) != E_OK)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ if (replyLength != sizeof(uint32_t) + sizeof(enabled))
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("IPC reply length mismatch"));
-+ p_FmPcd->enabled = (bool)!!(*(uint8_t*)(reply.replyBody));
-+ if (!p_FmPcd->enabled)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("FM-PCD master should be enabled first!"));
-+
-+ return E_OK;
-+ }
-+ else if (p_FmPcd->guestId != NCSW_MASTER_ID)
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED,
-+ ("running in guest-mode without IPC!"));
-+
-+ if (p_FmPcd->p_FmPcdKg)
-+ KgEnable(p_FmPcd);
-+
-+ if (p_FmPcd->p_FmPcdPlcr)
-+ PlcrEnable(p_FmPcd);
-+
-+ if (p_FmPcd->p_FmPcdPrs)
-+ PrsEnable(p_FmPcd);
-+
-+ p_FmPcd->enabled = TRUE;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PCD_Disable(t_Handle h_FmPcd)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+ t_Error err = E_OK;
-+
-+ SANITY_CHECK_RETURN_ERROR(h_FmPcd, E_INVALID_HANDLE);
-+
-+ if (!p_FmPcd->enabled)
-+ return E_OK;
-+
-+ if ((p_FmPcd->guestId != NCSW_MASTER_ID) &&
-+ p_FmPcd->h_IpcSession)
-+ {
-+ t_FmPcdIpcMsg msg;
-+ t_FmPcdIpcReply reply;
-+ uint32_t replyLength;
-+
-+ memset(&reply, 0, sizeof(reply));
-+ memset(&msg, 0, sizeof(msg));
-+ msg.msgId = FM_PCD_GUEST_DISABLE;
-+ replyLength = sizeof(uint32_t);
-+ if ((err = XX_IpcSendMessage(p_FmPcd->h_IpcSession,
-+ (uint8_t*)&msg,
-+ sizeof(msg.msgId),
-+ (uint8_t*)&reply,
-+ &replyLength,
-+ NULL,
-+ NULL)) != E_OK)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ if (replyLength != sizeof(uint32_t))
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("IPC reply length mismatch"));
-+ if (reply.error == E_OK)
-+ p_FmPcd->enabled = FALSE;
-+
-+ return (t_Error)(reply.error);
-+ }
-+ else if (p_FmPcd->guestId != NCSW_MASTER_ID)
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED,
-+ ("running in guest-mode without IPC!"));
-+
-+ if (p_FmPcd->numOfEnabledGuestPartitionsPcds != 0)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE,
-+ ("Trying to disable a master partition PCD while"
-+ "guest partitions are still enabled!"));
-+
-+ if (p_FmPcd->p_FmPcdKg)
-+ KgDisable(p_FmPcd);
-+
-+ if (p_FmPcd->p_FmPcdPlcr)
-+ PlcrDisable(p_FmPcd);
-+
-+ if (p_FmPcd->p_FmPcdPrs)
-+ PrsDisable(p_FmPcd);
-+
-+ p_FmPcd->enabled = FALSE;
-+
-+ return E_OK;
-+}
-+
-+t_Handle FM_PCD_NetEnvCharacteristicsSet(t_Handle h_FmPcd, t_FmPcdNetEnvParams *p_NetEnvParams)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+ uint32_t intFlags, specialUnits = 0;
-+ uint8_t bitId = 0;
-+ uint8_t i, j, k;
-+ uint8_t netEnvCurrId;
-+ uint8_t ipsecAhUnit = 0,ipsecEspUnit = 0;
-+ bool ipsecAhExists = FALSE, ipsecEspExists = FALSE, shim1Selected = FALSE;
-+ uint8_t hdrNum;
-+ t_FmPcdNetEnvParams *p_ModifiedNetEnvParams;
-+
-+ SANITY_CHECK_RETURN_VALUE(h_FmPcd, E_INVALID_STATE, NULL);
-+ SANITY_CHECK_RETURN_VALUE(!p_FmPcd->p_FmPcdDriverParam, E_INVALID_STATE, NULL);
-+ SANITY_CHECK_RETURN_VALUE(p_NetEnvParams, E_NULL_POINTER, NULL);
-+
-+ intFlags = FmPcdLock(p_FmPcd);
-+
-+ /* find a new netEnv */
-+ for (i = 0; i < FM_MAX_NUM_OF_PORTS; i++)
-+ if (!p_FmPcd->netEnvs[i].used)
-+ break;
-+
-+ if (i== FM_MAX_NUM_OF_PORTS)
-+ {
-+ REPORT_ERROR(MAJOR, E_FULL,("No more than %d netEnv's allowed.", FM_MAX_NUM_OF_PORTS));
-+ FmPcdUnlock(p_FmPcd, intFlags);
-+ return NULL;
-+ }
-+
-+ p_FmPcd->netEnvs[i].used = TRUE;
-+ FmPcdUnlock(p_FmPcd, intFlags);
-+
-+ /* As anyone doesn't have handle of this netEnv yet, no need
-+ to protect it with spinlocks */
-+
-+ p_ModifiedNetEnvParams = (t_FmPcdNetEnvParams *)XX_Malloc(sizeof(t_FmPcdNetEnvParams));
-+ if (!p_ModifiedNetEnvParams)
-+ {
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("FmPcdNetEnvParams"));
-+ return NULL;
-+ }
-+
-+ memcpy(p_ModifiedNetEnvParams, p_NetEnvParams, sizeof(t_FmPcdNetEnvParams));
-+ p_NetEnvParams = p_ModifiedNetEnvParams;
-+
-+ netEnvCurrId = (uint8_t)i;
-+
-+ /* clear from previous use */
-+ memset(&p_FmPcd->netEnvs[netEnvCurrId].units, 0, FM_PCD_MAX_NUM_OF_DISTINCTION_UNITS * sizeof(t_FmPcdIntDistinctionUnit));
-+ memset(&p_FmPcd->netEnvs[netEnvCurrId].aliasHdrs, 0, FM_PCD_MAX_NUM_OF_ALIAS_HDRS * sizeof(t_FmPcdNetEnvAliases));
-+ memcpy(&p_FmPcd->netEnvs[netEnvCurrId].units, p_NetEnvParams->units, p_NetEnvParams->numOfDistinctionUnits*sizeof(t_FmPcdIntDistinctionUnit));
-+
-+ p_FmPcd->netEnvs[netEnvCurrId].netEnvId = netEnvCurrId;
-+ p_FmPcd->netEnvs[netEnvCurrId].h_FmPcd = p_FmPcd;
-+
-+ p_FmPcd->netEnvs[netEnvCurrId].clsPlanGrpId = ILLEGAL_CLS_PLAN;
-+
-+ /* check that header with opt is not interchanged with the same header */
-+ for (i = 0; (i < FM_PCD_MAX_NUM_OF_DISTINCTION_UNITS)
-+ && (p_FmPcd->netEnvs[netEnvCurrId].units[i].hdrs[0].hdr != HEADER_TYPE_NONE); i++)
-+ {
-+ for (k = 0; (k < FM_PCD_MAX_NUM_OF_INTERCHANGEABLE_HDRS)
-+ && (p_FmPcd->netEnvs[netEnvCurrId].units[i].hdrs[k].hdr != HEADER_TYPE_NONE); k++)
-+ {
-+ /* if an option exists, check that other headers are not the same header
-+ without option */
-+ if (p_FmPcd->netEnvs[netEnvCurrId].units[i].hdrs[k].opt)
-+ {
-+ for (j = 0; (j < FM_PCD_MAX_NUM_OF_INTERCHANGEABLE_HDRS)
-+ && (p_FmPcd->netEnvs[netEnvCurrId].units[i].hdrs[j].hdr != HEADER_TYPE_NONE); j++)
-+ {
-+ if ((p_FmPcd->netEnvs[netEnvCurrId].units[i].hdrs[j].hdr == p_FmPcd->netEnvs[netEnvCurrId].units[i].hdrs[k].hdr) &&
-+ !p_FmPcd->netEnvs[netEnvCurrId].units[i].hdrs[j].opt)
-+ {
-+ REPORT_ERROR(MINOR, E_FULL,
-+ ("Illegal unit - header with opt may not be interchangeable with the same header without opt"));
-+ XX_Free(p_ModifiedNetEnvParams);
-+ return NULL;
-+ }
-+ }
-+ }
-+ }
-+ }
-+
-+ /* Specific headers checking */
-+ for (i = 0; (i < FM_PCD_MAX_NUM_OF_DISTINCTION_UNITS)
-+ && (p_FmPcd->netEnvs[netEnvCurrId].units[i].hdrs[0].hdr != HEADER_TYPE_NONE); i++)
-+ {
-+ for (k = 0; (k < FM_PCD_MAX_NUM_OF_INTERCHANGEABLE_HDRS)
-+ && (p_FmPcd->netEnvs[netEnvCurrId].units[i].hdrs[k].hdr != HEADER_TYPE_NONE); k++)
-+ {
-+ /* Some headers pairs may not be defined on different units as the parser
-+ doesn't distinguish */
-+ /* IPSEC_AH and IPSEC_SPI can't be 2 units, */
-+ /* check that header with opt is not interchanged with the same header */
-+ if (p_FmPcd->netEnvs[netEnvCurrId].units[i].hdrs[k].hdr == HEADER_TYPE_IPSEC_AH)
-+ {
-+ if (ipsecEspExists && (ipsecEspUnit != i))
-+ {
-+ REPORT_ERROR(MINOR, E_INVALID_STATE, ("HEADER_TYPE_IPSEC_AH and HEADER_TYPE_IPSEC_ESP may not be defined in separate units"));
-+ XX_Free(p_ModifiedNetEnvParams);
-+ return NULL;
-+ }
-+ else
-+ {
-+ ipsecAhUnit = i;
-+ ipsecAhExists = TRUE;
-+ }
-+ }
-+ if (p_FmPcd->netEnvs[netEnvCurrId].units[i].hdrs[k].hdr == HEADER_TYPE_IPSEC_ESP)
-+ {
-+ if (ipsecAhExists && (ipsecAhUnit != i))
-+ {
-+ REPORT_ERROR(MINOR, E_INVALID_STATE, ("HEADER_TYPE_IPSEC_AH and HEADER_TYPE_IPSEC_ESP may not be defined in separate units"));
-+ XX_Free(p_ModifiedNetEnvParams);
-+ return NULL;
-+ }
-+ else
-+ {
-+ ipsecEspUnit = i;
-+ ipsecEspExists = TRUE;
-+ }
-+ }
-+ /* ENCAP_ESP */
-+ if (p_FmPcd->netEnvs[netEnvCurrId].units[i].hdrs[k].hdr == HEADER_TYPE_UDP_ENCAP_ESP)
-+ {
-+ /* IPSec UDP encapsulation is currently set to use SHIM1 */
-+ p_FmPcd->netEnvs[netEnvCurrId].aliasHdrs[specialUnits].hdr = HEADER_TYPE_UDP_ENCAP_ESP;
-+ p_FmPcd->netEnvs[netEnvCurrId].aliasHdrs[specialUnits++].aliasHdr = HEADER_TYPE_USER_DEFINED_SHIM1;
-+ p_FmPcd->netEnvs[netEnvCurrId].units[i].hdrs[k].hdr = HEADER_TYPE_USER_DEFINED_SHIM1;
-+ p_FmPcd->netEnvs[netEnvCurrId].units[i].hdrs[k].opt = 0;
-+ }
-+#if (DPAA_VERSION >= 11) || ((DPAA_VERSION == 10) && defined(FM_CAPWAP_SUPPORT))
-+ /* UDP_LITE */
-+ if (p_FmPcd->netEnvs[netEnvCurrId].units[i].hdrs[k].hdr == HEADER_TYPE_UDP_LITE)
-+ {
-+ p_FmPcd->netEnvs[netEnvCurrId].aliasHdrs[specialUnits].hdr = HEADER_TYPE_UDP_LITE;
-+ p_FmPcd->netEnvs[netEnvCurrId].aliasHdrs[specialUnits++].aliasHdr = HEADER_TYPE_UDP;
-+ p_FmPcd->netEnvs[netEnvCurrId].units[i].hdrs[k].hdr = HEADER_TYPE_UDP;
-+ p_FmPcd->netEnvs[netEnvCurrId].units[i].hdrs[k].opt = 0;
-+ }
-+#endif /* (DPAA_VERSION >= 11) || ((DPAA_VERSION == 10) && defined(FM_CAPWAP_SUPPORT)) */
-+
-+ /* IP FRAG */
-+ if ((p_FmPcd->netEnvs[netEnvCurrId].units[i].hdrs[k].hdr == HEADER_TYPE_IPv4) &&
-+ (p_FmPcd->netEnvs[netEnvCurrId].units[i].hdrs[k].opt == IPV4_FRAG_1))
-+ {
-+ /* If IPv4+Frag, we need to set 2 units - SHIM 2 and IPv4. We first set SHIM2, and than check if
-+ * IPv4 exists. If so we don't need to set an extra unit
-+ * We consider as "having IPv4" any IPv4 without interchangable headers
-+ * but including any options. */
-+ p_FmPcd->netEnvs[netEnvCurrId].aliasHdrs[specialUnits].hdr = HEADER_TYPE_IPv4;
-+ p_FmPcd->netEnvs[netEnvCurrId].aliasHdrs[specialUnits].opt = IPV4_FRAG_1;
-+ p_FmPcd->netEnvs[netEnvCurrId].aliasHdrs[specialUnits++].aliasHdr = HEADER_TYPE_USER_DEFINED_SHIM2;
-+ p_FmPcd->netEnvs[netEnvCurrId].units[i].hdrs[k].hdr = HEADER_TYPE_USER_DEFINED_SHIM2;
-+ p_FmPcd->netEnvs[netEnvCurrId].units[i].hdrs[k].opt = 0;
-+
-+ /* check if IPv4 header exists by itself */
-+ if (FmPcdNetEnvGetUnitId(p_FmPcd, netEnvCurrId, HEADER_TYPE_IPv4, FALSE, 0) == FM_PCD_MAX_NUM_OF_DISTINCTION_UNITS)
-+ {
-+ p_FmPcd->netEnvs[netEnvCurrId].units[p_NetEnvParams->numOfDistinctionUnits].hdrs[0].hdr = HEADER_TYPE_IPv4;
-+ p_FmPcd->netEnvs[netEnvCurrId].units[p_NetEnvParams->numOfDistinctionUnits++].hdrs[0].opt = 0;
-+ }
-+ }
-+ if ((p_FmPcd->netEnvs[netEnvCurrId].units[i].hdrs[k].hdr == HEADER_TYPE_IPv6) &&
-+ (p_FmPcd->netEnvs[netEnvCurrId].units[i].hdrs[k].opt == IPV6_FRAG_1))
-+ {
-+ /* If IPv6+Frag, we need to set 2 units - SHIM 2 and IPv6. We first set SHIM2, and than check if
-+ * IPv4 exists. If so we don't need to set an extra unit
-+ * We consider as "having IPv6" any IPv6 without interchangable headers
-+ * but including any options. */
-+ p_FmPcd->netEnvs[netEnvCurrId].aliasHdrs[specialUnits].hdr = HEADER_TYPE_IPv6;
-+ p_FmPcd->netEnvs[netEnvCurrId].aliasHdrs[specialUnits].opt = IPV6_FRAG_1;
-+ p_FmPcd->netEnvs[netEnvCurrId].aliasHdrs[specialUnits++].aliasHdr = HEADER_TYPE_USER_DEFINED_SHIM2;
-+ p_FmPcd->netEnvs[netEnvCurrId].units[i].hdrs[k].hdr = HEADER_TYPE_USER_DEFINED_SHIM2;
-+ p_FmPcd->netEnvs[netEnvCurrId].units[i].hdrs[k].opt = 0;
-+
-+ /* check if IPv6 header exists by itself */
-+ if (FmPcdNetEnvGetUnitId(p_FmPcd, netEnvCurrId, HEADER_TYPE_IPv6, FALSE, 0) == FM_PCD_MAX_NUM_OF_DISTINCTION_UNITS)
-+ {
-+ p_FmPcd->netEnvs[netEnvCurrId].units[p_NetEnvParams->numOfDistinctionUnits].hdrs[0].hdr = HEADER_TYPE_IPv6;
-+ p_FmPcd->netEnvs[netEnvCurrId].units[p_NetEnvParams->numOfDistinctionUnits++].hdrs[0].opt = 0;
-+ }
-+ }
-+#if (DPAA_VERSION >= 11)
-+ /* CAPWAP FRAG */
-+ if ((p_FmPcd->netEnvs[netEnvCurrId].units[i].hdrs[k].hdr == HEADER_TYPE_CAPWAP) &&
-+ (p_FmPcd->netEnvs[netEnvCurrId].units[i].hdrs[k].opt == CAPWAP_FRAG_1))
-+ {
-+ p_FmPcd->netEnvs[netEnvCurrId].aliasHdrs[specialUnits].hdr = HEADER_TYPE_CAPWAP;
-+ p_FmPcd->netEnvs[netEnvCurrId].aliasHdrs[specialUnits].opt = CAPWAP_FRAG_1;
-+ p_FmPcd->netEnvs[netEnvCurrId].aliasHdrs[specialUnits++].aliasHdr = HEADER_TYPE_USER_DEFINED_SHIM2;
-+ p_FmPcd->netEnvs[netEnvCurrId].units[i].hdrs[k].hdr = HEADER_TYPE_USER_DEFINED_SHIM2;
-+ p_FmPcd->netEnvs[netEnvCurrId].units[i].hdrs[k].opt = 0;
-+ }
-+#endif /* (DPAA_VERSION >= 11) */
-+ }
-+ }
-+
-+ /* if private header (shim), check that no other headers specified */
-+ for (i = 0; (i < FM_PCD_MAX_NUM_OF_DISTINCTION_UNITS)
-+ && (p_FmPcd->netEnvs[netEnvCurrId].units[i].hdrs[0].hdr != HEADER_TYPE_NONE); i++)
-+ {
-+ if (IS_PRIVATE_HEADER(p_FmPcd->netEnvs[netEnvCurrId].units[i].hdrs[0].hdr))
-+ if (p_FmPcd->netEnvs[netEnvCurrId].units[i].hdrs[1].hdr != HEADER_TYPE_NONE)
-+ {
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("SHIM header may not be interchanged with other headers"));
-+ XX_Free(p_ModifiedNetEnvParams);
-+ return NULL;
-+ }
-+ }
-+
-+ for (i = 0; i < p_NetEnvParams->numOfDistinctionUnits; i++)
-+ {
-+ if (IS_PRIVATE_HEADER(p_FmPcd->netEnvs[netEnvCurrId].units[i].hdrs[0].hdr))
-+ switch (p_FmPcd->netEnvs[netEnvCurrId].units[i].hdrs[0].hdr)
-+ {
-+ case (HEADER_TYPE_USER_DEFINED_SHIM1):
-+ if (shim1Selected)
-+ {
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("SHIM header cannot be selected with UDP_IPSEC_ESP"));
-+ XX_Free(p_ModifiedNetEnvParams);
-+ return NULL;
-+ }
-+ shim1Selected = TRUE;
-+ p_FmPcd->netEnvs[netEnvCurrId].unitsVectors[i] = 0x00000001;
-+ break;
-+ case (HEADER_TYPE_USER_DEFINED_SHIM2):
-+ p_FmPcd->netEnvs[netEnvCurrId].unitsVectors[i] = 0x00000002;
-+ break;
-+ default:
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Requested SHIM not supported"));
-+ }
-+ else
-+ {
-+ p_FmPcd->netEnvs[netEnvCurrId].unitsVectors[i] = (uint32_t)(0x80000000 >> bitId++);
-+
-+ if (IS_SPECIAL_HEADER(p_FmPcd->netEnvs[netEnvCurrId].units[i].hdrs[0].hdr))
-+ p_FmPcd->netEnvs[netEnvCurrId].macsecVector = p_FmPcd->netEnvs[netEnvCurrId].unitsVectors[i];
-+ }
-+ }
-+
-+ /* define a set of hardware parser LCV's according to the defined netenv */
-+
-+ /* set an array of LCV's for each header in the netEnv */
-+ for (i = 0; (i < FM_PCD_MAX_NUM_OF_DISTINCTION_UNITS)
-+ && (p_FmPcd->netEnvs[netEnvCurrId].units[i].hdrs[0].hdr != HEADER_TYPE_NONE); i++)
-+ {
-+ /* private headers have no LCV in the hard parser */
-+ if (!IS_PRIVATE_HEADER(p_FmPcd->netEnvs[netEnvCurrId].units[i].hdrs[0].hdr))
-+ {
-+ for (k = 0; (k < FM_PCD_MAX_NUM_OF_INTERCHANGEABLE_HDRS)
-+ && (p_FmPcd->netEnvs[netEnvCurrId].units[i].hdrs[k].hdr != HEADER_TYPE_NONE); k++)
-+ {
-+ hdrNum = GetPrsHdrNum(p_FmPcd->netEnvs[netEnvCurrId].units[i].hdrs[k].hdr);
-+ if ((hdrNum == ILLEGAL_HDR_NUM) || (hdrNum == NO_HDR_NUM))
-+ {
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, NO_MSG);
-+ XX_Free(p_ModifiedNetEnvParams);
-+ return NULL;
-+ }
-+ p_FmPcd->netEnvs[netEnvCurrId].lcvs[hdrNum] |= p_FmPcd->netEnvs[netEnvCurrId].unitsVectors[i];
-+ }
-+ }
-+ }
-+ XX_Free(p_ModifiedNetEnvParams);
-+
-+ p_FmPcd->netEnvs[netEnvCurrId].h_Spinlock = XX_InitSpinlock();
-+ if (!p_FmPcd->netEnvs[netEnvCurrId].h_Spinlock)
-+ {
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("FM Pcd NetEnv spinlock"));
-+ return NULL;
-+ }
-+ return &p_FmPcd->netEnvs[netEnvCurrId];
-+}
-+
-+t_Error FM_PCD_NetEnvCharacteristicsDelete(t_Handle h_NetEnv)
-+{
-+ t_FmPcdNetEnv *p_NetEnv = (t_FmPcdNetEnv*)h_NetEnv;
-+ t_FmPcd *p_FmPcd = p_NetEnv->h_FmPcd;
-+ uint32_t intFlags;
-+ uint8_t netEnvId = p_NetEnv->netEnvId;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd, E_INVALID_STATE);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmPcd->p_FmPcdDriverParam, E_INVALID_STATE);
-+
-+ /* check that no port is bound to this netEnv */
-+ if (p_FmPcd->netEnvs[netEnvId].owners)
-+ {
-+ RETURN_ERROR(MINOR, E_INVALID_STATE,
-+ ("Trying to delete a netEnv that has ports/schemes/trees/clsPlanGrps bound to"));
-+ }
-+
-+ intFlags = FmPcdLock(p_FmPcd);
-+
-+ p_FmPcd->netEnvs[netEnvId].used = FALSE;
-+ p_FmPcd->netEnvs[netEnvId].clsPlanGrpId = ILLEGAL_CLS_PLAN;
-+
-+ memset(p_FmPcd->netEnvs[netEnvId].units, 0, sizeof(t_FmPcdIntDistinctionUnit)*FM_PCD_MAX_NUM_OF_DISTINCTION_UNITS);
-+ memset(p_FmPcd->netEnvs[netEnvId].unitsVectors, 0, sizeof(uint32_t)*FM_PCD_MAX_NUM_OF_DISTINCTION_UNITS);
-+ memset(p_FmPcd->netEnvs[netEnvId].lcvs, 0, sizeof(uint32_t)*FM_PCD_PRS_NUM_OF_HDRS);
-+
-+ if (p_FmPcd->netEnvs[netEnvId].h_Spinlock)
-+ XX_FreeSpinlock(p_FmPcd->netEnvs[netEnvId].h_Spinlock);
-+
-+ FmPcdUnlock(p_FmPcd, intFlags);
-+ return E_OK;
-+}
-+
-+void FM_PCD_HcTxConf(t_Handle h_FmPcd, t_DpaaFD *p_Fd)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+
-+ SANITY_CHECK_RETURN(h_FmPcd, E_INVALID_STATE);
-+
-+ FmHcTxConf(p_FmPcd->h_Hc, p_Fd);
-+}
-+
-+t_Error FM_PCD_SetAdvancedOffloadSupport(t_Handle h_FmPcd)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+ t_FmCtrlCodeRevisionInfo revInfo;
-+ t_Error err;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmPcd->p_FmPcdDriverParam, E_INVALID_STATE);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmPcd->enabled, E_INVALID_STATE);
-+
-+ if ((err = FM_GetFmanCtrlCodeRevision(p_FmPcd->h_Fm, &revInfo)) != E_OK)
-+ {
-+ DBG(WARNING, ("FM in guest-mode without IPC, can't validate firmware revision."));
-+ revInfo.packageRev = IP_OFFLOAD_PACKAGE_NUMBER;
-+ }
-+ if (!IS_OFFLOAD_PACKAGE(revInfo.packageRev))
-+ RETURN_ERROR(MAJOR, E_NOT_SUPPORTED, ("Fman ctrl code package"));
-+
-+ if (!p_FmPcd->h_Hc)
-+ RETURN_ERROR(MAJOR, E_INVALID_HANDLE, ("HC must be initialized in this mode"));
-+
-+ p_FmPcd->advancedOffloadSupport = TRUE;
-+
-+ return E_OK;
-+}
-+
-+uint32_t FM_PCD_GetCounter(t_Handle h_FmPcd, e_FmPcdCounters counter)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+ uint32_t outCounter = 0;
-+ t_Error err;
-+
-+ SANITY_CHECK_RETURN_VALUE(h_FmPcd, E_INVALID_HANDLE, 0);
-+ SANITY_CHECK_RETURN_VALUE(!p_FmPcd->p_FmPcdDriverParam, E_INVALID_STATE, 0);
-+
-+ switch (counter)
-+ {
-+ case (e_FM_PCD_KG_COUNTERS_TOTAL):
-+ if (!p_FmPcd->p_FmPcdKg)
-+ {
-+ REPORT_ERROR(MAJOR, E_INVALID_STATE, ("KeyGen is not activated"));
-+ return 0;
-+ }
-+ if ((p_FmPcd->guestId != NCSW_MASTER_ID) &&
-+ !p_FmPcd->p_FmPcdKg->p_FmPcdKgRegs &&
-+ !p_FmPcd->h_IpcSession)
-+ {
-+ REPORT_ERROR(MINOR, E_NOT_SUPPORTED,
-+ ("running in guest-mode without neither IPC nor mapped register!"));
-+ return 0;
-+ }
-+ break;
-+
-+ case (e_FM_PCD_PLCR_COUNTERS_YELLOW):
-+ case (e_FM_PCD_PLCR_COUNTERS_RED):
-+ case (e_FM_PCD_PLCR_COUNTERS_RECOLORED_TO_RED):
-+ case (e_FM_PCD_PLCR_COUNTERS_RECOLORED_TO_YELLOW):
-+ case (e_FM_PCD_PLCR_COUNTERS_TOTAL):
-+ case (e_FM_PCD_PLCR_COUNTERS_LENGTH_MISMATCH):
-+ if (!p_FmPcd->p_FmPcdPlcr)
-+ {
-+ REPORT_ERROR(MAJOR, E_INVALID_STATE, ("Policer is not activated"));
-+ return 0;
-+ }
-+ if ((p_FmPcd->guestId != NCSW_MASTER_ID) &&
-+ !p_FmPcd->p_FmPcdPlcr->p_FmPcdPlcrRegs &&
-+ !p_FmPcd->h_IpcSession)
-+ {
-+ REPORT_ERROR(MINOR, E_NOT_SUPPORTED,
-+ ("running in \"guest-mode\" without neither IPC nor mapped register!"));
-+ return 0;
-+ }
-+
-+ /* check that counters are enabled */
-+ if (p_FmPcd->p_FmPcdPlcr->p_FmPcdPlcrRegs &&
-+ !(GET_UINT32(p_FmPcd->p_FmPcdPlcr->p_FmPcdPlcrRegs->fmpl_gcr) & FM_PCD_PLCR_GCR_STEN))
-+ {
-+ REPORT_ERROR(MINOR, E_INVALID_STATE, ("Requested counter was not enabled"));
-+ return 0;
-+ }
-+ ASSERT_COND(p_FmPcd->p_FmPcdPlcr->p_FmPcdPlcrRegs ||
-+ ((p_FmPcd->guestId != NCSW_MASTER_ID) && p_FmPcd->h_IpcSession));
-+ break;
-+
-+ case (e_FM_PCD_PRS_COUNTERS_PARSE_DISPATCH):
-+ case (e_FM_PCD_PRS_COUNTERS_L2_PARSE_RESULT_RETURNED):
-+ case (e_FM_PCD_PRS_COUNTERS_L3_PARSE_RESULT_RETURNED):
-+ case (e_FM_PCD_PRS_COUNTERS_L4_PARSE_RESULT_RETURNED):
-+ case (e_FM_PCD_PRS_COUNTERS_SHIM_PARSE_RESULT_RETURNED):
-+ case (e_FM_PCD_PRS_COUNTERS_L2_PARSE_RESULT_RETURNED_WITH_ERR):
-+ case (e_FM_PCD_PRS_COUNTERS_L3_PARSE_RESULT_RETURNED_WITH_ERR):
-+ case (e_FM_PCD_PRS_COUNTERS_L4_PARSE_RESULT_RETURNED_WITH_ERR):
-+ case (e_FM_PCD_PRS_COUNTERS_SHIM_PARSE_RESULT_RETURNED_WITH_ERR):
-+ case (e_FM_PCD_PRS_COUNTERS_SOFT_PRS_CYCLES):
-+ case (e_FM_PCD_PRS_COUNTERS_SOFT_PRS_STALL_CYCLES):
-+ case (e_FM_PCD_PRS_COUNTERS_HARD_PRS_CYCLE_INCL_STALL_CYCLES):
-+ case (e_FM_PCD_PRS_COUNTERS_MURAM_READ_CYCLES):
-+ case (e_FM_PCD_PRS_COUNTERS_MURAM_READ_STALL_CYCLES):
-+ case (e_FM_PCD_PRS_COUNTERS_MURAM_WRITE_CYCLES):
-+ case (e_FM_PCD_PRS_COUNTERS_MURAM_WRITE_STALL_CYCLES):
-+ case (e_FM_PCD_PRS_COUNTERS_FPM_COMMAND_STALL_CYCLES):
-+ if (!p_FmPcd->p_FmPcdPrs)
-+ {
-+ REPORT_ERROR(MAJOR, E_INVALID_STATE, ("Parser is not activated"));
-+ return 0;
-+ }
-+ if ((p_FmPcd->guestId != NCSW_MASTER_ID) &&
-+ !p_FmPcd->p_FmPcdPrs->p_FmPcdPrsRegs &&
-+ !p_FmPcd->h_IpcSession)
-+ {
-+ REPORT_ERROR(MINOR, E_NOT_SUPPORTED,
-+ ("running in guest-mode without neither IPC nor mapped register!"));
-+ return 0;
-+ }
-+ break;
-+ default:
-+ REPORT_ERROR(MAJOR, E_INVALID_STATE, ("Unsupported type of counter"));
-+ return 0;
-+ }
-+
-+ if ((p_FmPcd->guestId != NCSW_MASTER_ID) &&
-+ p_FmPcd->h_IpcSession)
-+ {
-+ t_FmPcdIpcMsg msg;
-+ t_FmPcdIpcReply reply;
-+ uint32_t replyLength;
-+
-+ memset(&msg, 0, sizeof(msg));
-+ memset(&reply, 0, sizeof(reply));
-+ msg.msgId = FM_PCD_GET_COUNTER;
-+ memcpy(msg.msgBody, (uint8_t *)&counter, sizeof(uint32_t));
-+ replyLength = sizeof(uint32_t) + sizeof(uint32_t);
-+ if ((err = XX_IpcSendMessage(p_FmPcd->h_IpcSession,
-+ (uint8_t*)&msg,
-+ sizeof(msg.msgId) +sizeof(uint32_t),
-+ (uint8_t*)&reply,
-+ &replyLength,
-+ NULL,
-+ NULL)) != E_OK)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ if (replyLength != sizeof(uint32_t) + sizeof(uint32_t))
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("IPC reply length mismatch"));
-+
-+ memcpy((uint8_t*)&outCounter, reply.replyBody, sizeof(uint32_t));
-+ return outCounter;
-+ }
-+
-+ switch (counter)
-+ {
-+ /* Parser statistics */
-+ case (e_FM_PCD_PRS_COUNTERS_PARSE_DISPATCH):
-+ return GET_UINT32(p_FmPcd->p_FmPcdPrs->p_FmPcdPrsRegs->fmpr_pds);
-+ case (e_FM_PCD_PRS_COUNTERS_L2_PARSE_RESULT_RETURNED):
-+ return GET_UINT32(p_FmPcd->p_FmPcdPrs->p_FmPcdPrsRegs->fmpr_l2rrs);
-+ case (e_FM_PCD_PRS_COUNTERS_L3_PARSE_RESULT_RETURNED):
-+ return GET_UINT32(p_FmPcd->p_FmPcdPrs->p_FmPcdPrsRegs->fmpr_l3rrs);
-+ case (e_FM_PCD_PRS_COUNTERS_L4_PARSE_RESULT_RETURNED):
-+ return GET_UINT32(p_FmPcd->p_FmPcdPrs->p_FmPcdPrsRegs->fmpr_l4rrs);
-+ case (e_FM_PCD_PRS_COUNTERS_SHIM_PARSE_RESULT_RETURNED):
-+ return GET_UINT32(p_FmPcd->p_FmPcdPrs->p_FmPcdPrsRegs->fmpr_srrs);
-+ case (e_FM_PCD_PRS_COUNTERS_L2_PARSE_RESULT_RETURNED_WITH_ERR):
-+ return GET_UINT32(p_FmPcd->p_FmPcdPrs->p_FmPcdPrsRegs->fmpr_l2rres);
-+ case (e_FM_PCD_PRS_COUNTERS_L3_PARSE_RESULT_RETURNED_WITH_ERR):
-+ return GET_UINT32(p_FmPcd->p_FmPcdPrs->p_FmPcdPrsRegs->fmpr_l3rres);
-+ case (e_FM_PCD_PRS_COUNTERS_L4_PARSE_RESULT_RETURNED_WITH_ERR):
-+ return GET_UINT32(p_FmPcd->p_FmPcdPrs->p_FmPcdPrsRegs->fmpr_l4rres);
-+ case (e_FM_PCD_PRS_COUNTERS_SHIM_PARSE_RESULT_RETURNED_WITH_ERR):
-+ return GET_UINT32(p_FmPcd->p_FmPcdPrs->p_FmPcdPrsRegs->fmpr_srres);
-+ case (e_FM_PCD_PRS_COUNTERS_SOFT_PRS_CYCLES):
-+ return GET_UINT32(p_FmPcd->p_FmPcdPrs->p_FmPcdPrsRegs->fmpr_spcs);
-+ case (e_FM_PCD_PRS_COUNTERS_SOFT_PRS_STALL_CYCLES):
-+ return GET_UINT32(p_FmPcd->p_FmPcdPrs->p_FmPcdPrsRegs->fmpr_spscs);
-+ case (e_FM_PCD_PRS_COUNTERS_HARD_PRS_CYCLE_INCL_STALL_CYCLES):
-+ return GET_UINT32(p_FmPcd->p_FmPcdPrs->p_FmPcdPrsRegs->fmpr_hxscs);
-+ case (e_FM_PCD_PRS_COUNTERS_MURAM_READ_CYCLES):
-+ return GET_UINT32(p_FmPcd->p_FmPcdPrs->p_FmPcdPrsRegs->fmpr_mrcs);
-+ case (e_FM_PCD_PRS_COUNTERS_MURAM_READ_STALL_CYCLES):
-+ return GET_UINT32(p_FmPcd->p_FmPcdPrs->p_FmPcdPrsRegs->fmpr_mrscs);
-+ case (e_FM_PCD_PRS_COUNTERS_MURAM_WRITE_CYCLES):
-+ return GET_UINT32(p_FmPcd->p_FmPcdPrs->p_FmPcdPrsRegs->fmpr_mwcs);
-+ case (e_FM_PCD_PRS_COUNTERS_MURAM_WRITE_STALL_CYCLES):
-+ return GET_UINT32(p_FmPcd->p_FmPcdPrs->p_FmPcdPrsRegs->fmpr_mwscs);
-+ case (e_FM_PCD_PRS_COUNTERS_FPM_COMMAND_STALL_CYCLES):
-+ return GET_UINT32(p_FmPcd->p_FmPcdPrs->p_FmPcdPrsRegs->fmpr_fcscs);
-+ case (e_FM_PCD_KG_COUNTERS_TOTAL):
-+ return GET_UINT32(p_FmPcd->p_FmPcdKg->p_FmPcdKgRegs->fmkg_tpc);
-+
-+ /* Policer statistics */
-+ case (e_FM_PCD_PLCR_COUNTERS_YELLOW):
-+ return GET_UINT32(p_FmPcd->p_FmPcdPlcr->p_FmPcdPlcrRegs->fmpl_ypcnt);
-+ case (e_FM_PCD_PLCR_COUNTERS_RED):
-+ return GET_UINT32(p_FmPcd->p_FmPcdPlcr->p_FmPcdPlcrRegs->fmpl_rpcnt);
-+ case (e_FM_PCD_PLCR_COUNTERS_RECOLORED_TO_RED):
-+ return GET_UINT32(p_FmPcd->p_FmPcdPlcr->p_FmPcdPlcrRegs->fmpl_rrpcnt);
-+ case (e_FM_PCD_PLCR_COUNTERS_RECOLORED_TO_YELLOW):
-+ return GET_UINT32(p_FmPcd->p_FmPcdPlcr->p_FmPcdPlcrRegs->fmpl_rypcnt);
-+ case (e_FM_PCD_PLCR_COUNTERS_TOTAL):
-+ return GET_UINT32(p_FmPcd->p_FmPcdPlcr->p_FmPcdPlcrRegs->fmpl_tpcnt);
-+ case (e_FM_PCD_PLCR_COUNTERS_LENGTH_MISMATCH):
-+ return GET_UINT32(p_FmPcd->p_FmPcdPlcr->p_FmPcdPlcrRegs->fmpl_flmcnt);
-+ }
-+ return 0;
-+}
-+
-+t_Error FM_PCD_SetException(t_Handle h_FmPcd, e_FmPcdExceptions exception, bool enable)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+ uint32_t bitMask = 0, tmpReg;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmPcd->p_FmPcdDriverParam, E_INVALID_STATE);
-+
-+ if (p_FmPcd->guestId != NCSW_MASTER_ID)
-+ RETURN_ERROR(MAJOR, E_NOT_SUPPORTED, ("FM_PCD_SetException - guest mode!"));
-+
-+ GET_FM_PCD_EXCEPTION_FLAG(bitMask, exception);
-+
-+ if (bitMask)
-+ {
-+ if (enable)
-+ p_FmPcd->exceptions |= bitMask;
-+ else
-+ p_FmPcd->exceptions &= ~bitMask;
-+
-+ switch (exception)
-+ {
-+ case (e_FM_PCD_KG_EXCEPTION_DOUBLE_ECC):
-+ case (e_FM_PCD_KG_EXCEPTION_KEYSIZE_OVERFLOW):
-+ if (!p_FmPcd->p_FmPcdKg)
-+ RETURN_ERROR(MINOR, E_INVALID_STATE, ("Can't ask for this interrupt - keygen is not working"));
-+ break;
-+ case (e_FM_PCD_PLCR_EXCEPTION_DOUBLE_ECC):
-+ case (e_FM_PCD_PLCR_EXCEPTION_INIT_ENTRY_ERROR):
-+ case (e_FM_PCD_PLCR_EXCEPTION_PRAM_SELF_INIT_COMPLETE):
-+ case (e_FM_PCD_PLCR_EXCEPTION_ATOMIC_ACTION_COMPLETE):
-+ if (!p_FmPcd->p_FmPcdPlcr)
-+ RETURN_ERROR(MINOR, E_INVALID_STATE, ("Can't ask for this interrupt - policer is not working"));
-+ break;
-+ case (e_FM_PCD_PRS_EXCEPTION_DOUBLE_ECC):
-+ case (e_FM_PCD_PRS_EXCEPTION_SINGLE_ECC):
-+ if (!p_FmPcd->p_FmPcdPrs)
-+ RETURN_ERROR(MINOR, E_INVALID_STATE, ("Can't ask for this interrupt - parser is not working"));
-+ break;
-+ }
-+
-+ switch (exception)
-+ {
-+ case (e_FM_PCD_KG_EXCEPTION_DOUBLE_ECC):
-+ tmpReg = GET_UINT32(p_FmPcd->p_FmPcdKg->p_FmPcdKgRegs->fmkg_eeer);
-+ if (enable)
-+ tmpReg |= FM_EX_KG_DOUBLE_ECC;
-+ else
-+ tmpReg &= ~FM_EX_KG_DOUBLE_ECC;
-+ WRITE_UINT32(p_FmPcd->p_FmPcdKg->p_FmPcdKgRegs->fmkg_eeer, tmpReg);
-+ break;
-+ case (e_FM_PCD_KG_EXCEPTION_KEYSIZE_OVERFLOW):
-+ tmpReg = GET_UINT32(p_FmPcd->p_FmPcdKg->p_FmPcdKgRegs->fmkg_eeer);
-+ if (enable)
-+ tmpReg |= FM_EX_KG_KEYSIZE_OVERFLOW;
-+ else
-+ tmpReg &= ~FM_EX_KG_KEYSIZE_OVERFLOW;
-+ WRITE_UINT32(p_FmPcd->p_FmPcdKg->p_FmPcdKgRegs->fmkg_eeer, tmpReg);
-+ break;
-+ case (e_FM_PCD_PRS_EXCEPTION_DOUBLE_ECC):
-+ tmpReg = GET_UINT32(p_FmPcd->p_FmPcdPrs->p_FmPcdPrsRegs->fmpr_perer);
-+ if (enable)
-+ tmpReg |= FM_PCD_PRS_DOUBLE_ECC;
-+ else
-+ tmpReg &= ~FM_PCD_PRS_DOUBLE_ECC;
-+ WRITE_UINT32(p_FmPcd->p_FmPcdPrs->p_FmPcdPrsRegs->fmpr_perer, tmpReg);
-+ break;
-+ case (e_FM_PCD_PRS_EXCEPTION_SINGLE_ECC):
-+ tmpReg = GET_UINT32(p_FmPcd->p_FmPcdPrs->p_FmPcdPrsRegs->fmpr_pever);
-+ if (enable)
-+ tmpReg |= FM_PCD_PRS_SINGLE_ECC;
-+ else
-+ tmpReg &= ~FM_PCD_PRS_SINGLE_ECC;
-+ WRITE_UINT32(p_FmPcd->p_FmPcdPrs->p_FmPcdPrsRegs->fmpr_pever, tmpReg);
-+ break;
-+ case (e_FM_PCD_PLCR_EXCEPTION_DOUBLE_ECC):
-+ tmpReg = GET_UINT32(p_FmPcd->p_FmPcdPlcr->p_FmPcdPlcrRegs->fmpl_eier);
-+ if (enable)
-+ tmpReg |= FM_PCD_PLCR_DOUBLE_ECC;
-+ else
-+ tmpReg &= ~FM_PCD_PLCR_DOUBLE_ECC;
-+ WRITE_UINT32(p_FmPcd->p_FmPcdPlcr->p_FmPcdPlcrRegs->fmpl_eier, tmpReg);
-+ break;
-+ case (e_FM_PCD_PLCR_EXCEPTION_INIT_ENTRY_ERROR):
-+ tmpReg = GET_UINT32(p_FmPcd->p_FmPcdPlcr->p_FmPcdPlcrRegs->fmpl_eier);
-+ if (enable)
-+ tmpReg |= FM_PCD_PLCR_INIT_ENTRY_ERROR;
-+ else
-+ tmpReg &= ~FM_PCD_PLCR_INIT_ENTRY_ERROR;
-+ WRITE_UINT32(p_FmPcd->p_FmPcdPlcr->p_FmPcdPlcrRegs->fmpl_eier, tmpReg);
-+ break;
-+ case (e_FM_PCD_PLCR_EXCEPTION_PRAM_SELF_INIT_COMPLETE):
-+ tmpReg = GET_UINT32(p_FmPcd->p_FmPcdPlcr->p_FmPcdPlcrRegs->fmpl_ier);
-+ if (enable)
-+ tmpReg |= FM_PCD_PLCR_PRAM_SELF_INIT_COMPLETE;
-+ else
-+ tmpReg &= ~FM_PCD_PLCR_PRAM_SELF_INIT_COMPLETE;
-+ WRITE_UINT32(p_FmPcd->p_FmPcdPlcr->p_FmPcdPlcrRegs->fmpl_ier, tmpReg);
-+ break;
-+ case (e_FM_PCD_PLCR_EXCEPTION_ATOMIC_ACTION_COMPLETE):
-+ tmpReg = GET_UINT32(p_FmPcd->p_FmPcdPlcr->p_FmPcdPlcrRegs->fmpl_ier);
-+ if (enable)
-+ tmpReg |= FM_PCD_PLCR_ATOMIC_ACTION_COMPLETE;
-+ else
-+ tmpReg &= ~FM_PCD_PLCR_ATOMIC_ACTION_COMPLETE;
-+ WRITE_UINT32(p_FmPcd->p_FmPcdPlcr->p_FmPcdPlcrRegs->fmpl_ier, tmpReg);
-+ break;
-+ }
-+ /* for ECC exceptions driver automatically enables ECC mechanism, if disabled.
-+ Driver may disable them automatically, depending on driver's status */
-+ if (enable && ((exception == e_FM_PCD_KG_EXCEPTION_DOUBLE_ECC) |
-+ (exception == e_FM_PCD_PLCR_EXCEPTION_DOUBLE_ECC) |
-+ (exception == e_FM_PCD_PRS_EXCEPTION_DOUBLE_ECC) |
-+ (exception == e_FM_PCD_PRS_EXCEPTION_SINGLE_ECC)))
-+ FmEnableRamsEcc(p_FmPcd->h_Fm);
-+ if (!enable && ((exception == e_FM_PCD_KG_EXCEPTION_DOUBLE_ECC) |
-+ (exception == e_FM_PCD_PLCR_EXCEPTION_DOUBLE_ECC) |
-+ (exception == e_FM_PCD_PRS_EXCEPTION_DOUBLE_ECC) |
-+ (exception == e_FM_PCD_PRS_EXCEPTION_SINGLE_ECC)))
-+ FmDisableRamsEcc(p_FmPcd->h_Fm);
-+ }
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PCD_ForceIntr (t_Handle h_FmPcd, e_FmPcdExceptions exception)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+
-+ SANITY_CHECK_RETURN_ERROR(h_FmPcd, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmPcd->p_FmPcdDriverParam, E_INVALID_STATE);
-+
-+ if (p_FmPcd->guestId != NCSW_MASTER_ID)
-+ RETURN_ERROR(MAJOR, E_NOT_SUPPORTED, ("FM_PCD_ForceIntr - guest mode!"));
-+
-+ switch (exception)
-+ {
-+ case (e_FM_PCD_KG_EXCEPTION_DOUBLE_ECC):
-+ case (e_FM_PCD_KG_EXCEPTION_KEYSIZE_OVERFLOW):
-+ if (!p_FmPcd->p_FmPcdKg)
-+ RETURN_ERROR(MINOR, E_INVALID_STATE, ("Can't ask for this interrupt - keygen is not working"));
-+ break;
-+ case (e_FM_PCD_PLCR_EXCEPTION_DOUBLE_ECC):
-+ case (e_FM_PCD_PLCR_EXCEPTION_INIT_ENTRY_ERROR):
-+ case (e_FM_PCD_PLCR_EXCEPTION_PRAM_SELF_INIT_COMPLETE):
-+ case (e_FM_PCD_PLCR_EXCEPTION_ATOMIC_ACTION_COMPLETE):
-+ if (!p_FmPcd->p_FmPcdPlcr)
-+ RETURN_ERROR(MINOR, E_INVALID_STATE, ("Can't ask for this interrupt - policer is not working"));
-+ break;
-+ case (e_FM_PCD_PRS_EXCEPTION_DOUBLE_ECC):
-+ case (e_FM_PCD_PRS_EXCEPTION_SINGLE_ECC):
-+ if (!p_FmPcd->p_FmPcdPrs)
-+ RETURN_ERROR(MINOR, E_INVALID_STATE, ("Can't ask for this interrupt -parsrer is not working"));
-+ break;
-+ default:
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("Invalid interrupt requested"));
-+ }
-+ switch (exception)
-+ {
-+ case e_FM_PCD_PRS_EXCEPTION_DOUBLE_ECC:
-+ if (!(p_FmPcd->exceptions & FM_PCD_EX_PRS_DOUBLE_ECC))
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, ("The selected exception is masked"));
-+ break;
-+ case e_FM_PCD_PRS_EXCEPTION_SINGLE_ECC:
-+ if (!(p_FmPcd->exceptions & FM_PCD_EX_PRS_SINGLE_ECC))
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, ("The selected exception is masked"));
-+ break;
-+ case e_FM_PCD_KG_EXCEPTION_DOUBLE_ECC:
-+ if (!(p_FmPcd->exceptions & FM_EX_KG_DOUBLE_ECC))
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, ("The selected exception is masked"));
-+ WRITE_UINT32(p_FmPcd->p_FmPcdKg->p_FmPcdKgRegs->fmkg_feer, FM_EX_KG_DOUBLE_ECC);
-+ break;
-+ case e_FM_PCD_KG_EXCEPTION_KEYSIZE_OVERFLOW:
-+ if (!(p_FmPcd->exceptions & FM_EX_KG_KEYSIZE_OVERFLOW))
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, ("The selected exception is masked"));
-+ WRITE_UINT32(p_FmPcd->p_FmPcdKg->p_FmPcdKgRegs->fmkg_feer, FM_EX_KG_KEYSIZE_OVERFLOW);
-+ break;
-+ case e_FM_PCD_PLCR_EXCEPTION_DOUBLE_ECC:
-+ if (!(p_FmPcd->exceptions & FM_PCD_EX_PLCR_DOUBLE_ECC))
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, ("The selected exception is masked"));
-+ WRITE_UINT32(p_FmPcd->p_FmPcdPlcr->p_FmPcdPlcrRegs->fmpl_eifr, FM_PCD_PLCR_DOUBLE_ECC);
-+ break;
-+ case e_FM_PCD_PLCR_EXCEPTION_INIT_ENTRY_ERROR:
-+ if (!(p_FmPcd->exceptions & FM_PCD_EX_PLCR_INIT_ENTRY_ERROR))
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, ("The selected exception is masked"));
-+ WRITE_UINT32(p_FmPcd->p_FmPcdPlcr->p_FmPcdPlcrRegs->fmpl_eifr, FM_PCD_PLCR_INIT_ENTRY_ERROR);
-+ break;
-+ case e_FM_PCD_PLCR_EXCEPTION_PRAM_SELF_INIT_COMPLETE:
-+ if (!(p_FmPcd->exceptions & FM_PCD_EX_PLCR_PRAM_SELF_INIT_COMPLETE))
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, ("The selected exception is masked"));
-+ WRITE_UINT32(p_FmPcd->p_FmPcdPlcr->p_FmPcdPlcrRegs->fmpl_ifr, FM_PCD_PLCR_PRAM_SELF_INIT_COMPLETE);
-+ break;
-+ case e_FM_PCD_PLCR_EXCEPTION_ATOMIC_ACTION_COMPLETE:
-+ if (!(p_FmPcd->exceptions & FM_PCD_EX_PLCR_ATOMIC_ACTION_COMPLETE))
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, ("The selected exception is masked"));
-+ WRITE_UINT32(p_FmPcd->p_FmPcdPlcr->p_FmPcdPlcrRegs->fmpl_ifr, FM_PCD_PLCR_ATOMIC_ACTION_COMPLETE);
-+ break;
-+ }
-+
-+ return E_OK;
-+}
-+
-+
-+t_Error FM_PCD_ModifyCounter(t_Handle h_FmPcd, e_FmPcdCounters counter, uint32_t value)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+
-+ SANITY_CHECK_RETURN_ERROR(h_FmPcd, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmPcd->p_FmPcdDriverParam, E_INVALID_STATE);
-+
-+ if (p_FmPcd->guestId != NCSW_MASTER_ID)
-+ RETURN_ERROR(MAJOR, E_NOT_SUPPORTED, ("FM_PCD_ModifyCounter - guest mode!"));
-+
-+ switch (counter)
-+ {
-+ case (e_FM_PCD_KG_COUNTERS_TOTAL):
-+ if (!p_FmPcd->p_FmPcdKg)
-+ RETURN_ERROR(MINOR, E_INVALID_STATE, ("Invalid counters - KeyGen is not working"));
-+ break;
-+ case (e_FM_PCD_PLCR_COUNTERS_YELLOW):
-+ case (e_FM_PCD_PLCR_COUNTERS_RED):
-+ case (e_FM_PCD_PLCR_COUNTERS_RECOLORED_TO_RED):
-+ case (e_FM_PCD_PLCR_COUNTERS_RECOLORED_TO_YELLOW):
-+ case (e_FM_PCD_PLCR_COUNTERS_TOTAL):
-+ case (e_FM_PCD_PLCR_COUNTERS_LENGTH_MISMATCH):
-+ if (!p_FmPcd->p_FmPcdPlcr)
-+ RETURN_ERROR(MINOR, E_INVALID_STATE, ("Invalid counters - Policer is not working"));
-+ if (!(GET_UINT32(p_FmPcd->p_FmPcdPlcr->p_FmPcdPlcrRegs->fmpl_gcr) & FM_PCD_PLCR_GCR_STEN))
-+ RETURN_ERROR(MINOR, E_INVALID_STATE, ("Requested counter was not enabled"));
-+ break;
-+ case (e_FM_PCD_PRS_COUNTERS_PARSE_DISPATCH):
-+ case (e_FM_PCD_PRS_COUNTERS_L2_PARSE_RESULT_RETURNED):
-+ case (e_FM_PCD_PRS_COUNTERS_L3_PARSE_RESULT_RETURNED):
-+ case (e_FM_PCD_PRS_COUNTERS_L4_PARSE_RESULT_RETURNED):
-+ case (e_FM_PCD_PRS_COUNTERS_SHIM_PARSE_RESULT_RETURNED):
-+ case (e_FM_PCD_PRS_COUNTERS_L2_PARSE_RESULT_RETURNED_WITH_ERR):
-+ case (e_FM_PCD_PRS_COUNTERS_L3_PARSE_RESULT_RETURNED_WITH_ERR):
-+ case (e_FM_PCD_PRS_COUNTERS_L4_PARSE_RESULT_RETURNED_WITH_ERR):
-+ case (e_FM_PCD_PRS_COUNTERS_SHIM_PARSE_RESULT_RETURNED_WITH_ERR):
-+ case (e_FM_PCD_PRS_COUNTERS_SOFT_PRS_CYCLES):
-+ case (e_FM_PCD_PRS_COUNTERS_SOFT_PRS_STALL_CYCLES):
-+ case (e_FM_PCD_PRS_COUNTERS_HARD_PRS_CYCLE_INCL_STALL_CYCLES):
-+ case (e_FM_PCD_PRS_COUNTERS_MURAM_READ_CYCLES):
-+ case (e_FM_PCD_PRS_COUNTERS_MURAM_READ_STALL_CYCLES):
-+ case (e_FM_PCD_PRS_COUNTERS_MURAM_WRITE_CYCLES):
-+ case (e_FM_PCD_PRS_COUNTERS_MURAM_WRITE_STALL_CYCLES):
-+ case (e_FM_PCD_PRS_COUNTERS_FPM_COMMAND_STALL_CYCLES):
-+ if (!p_FmPcd->p_FmPcdPrs)
-+ RETURN_ERROR(MINOR, E_INVALID_STATE, ("Unsupported type of counter"));
-+ break;
-+ default:
-+ RETURN_ERROR(MINOR, E_INVALID_STATE, ("Unsupported type of counter"));
-+ }
-+ switch (counter)
-+ {
-+ case (e_FM_PCD_PRS_COUNTERS_PARSE_DISPATCH):
-+ WRITE_UINT32(p_FmPcd->p_FmPcdPrs->p_FmPcdPrsRegs->fmpr_pds, value);
-+ break;
-+ case (e_FM_PCD_PRS_COUNTERS_L2_PARSE_RESULT_RETURNED):
-+ WRITE_UINT32(p_FmPcd->p_FmPcdPrs->p_FmPcdPrsRegs->fmpr_l2rrs, value);
-+ break;
-+ case (e_FM_PCD_PRS_COUNTERS_L3_PARSE_RESULT_RETURNED):
-+ WRITE_UINT32(p_FmPcd->p_FmPcdPrs->p_FmPcdPrsRegs->fmpr_l3rrs, value);
-+ break;
-+ case (e_FM_PCD_PRS_COUNTERS_L4_PARSE_RESULT_RETURNED):
-+ WRITE_UINT32(p_FmPcd->p_FmPcdPrs->p_FmPcdPrsRegs->fmpr_l4rrs, value);
-+ break;
-+ case (e_FM_PCD_PRS_COUNTERS_SHIM_PARSE_RESULT_RETURNED):
-+ WRITE_UINT32(p_FmPcd->p_FmPcdPrs->p_FmPcdPrsRegs->fmpr_srrs, value);
-+ break;
-+ case (e_FM_PCD_PRS_COUNTERS_L2_PARSE_RESULT_RETURNED_WITH_ERR):
-+ WRITE_UINT32(p_FmPcd->p_FmPcdPrs->p_FmPcdPrsRegs->fmpr_l2rres, value);
-+ break;
-+ case (e_FM_PCD_PRS_COUNTERS_L3_PARSE_RESULT_RETURNED_WITH_ERR):
-+ WRITE_UINT32(p_FmPcd->p_FmPcdPrs->p_FmPcdPrsRegs->fmpr_l3rres, value);
-+ break;
-+ case (e_FM_PCD_PRS_COUNTERS_L4_PARSE_RESULT_RETURNED_WITH_ERR):
-+ WRITE_UINT32(p_FmPcd->p_FmPcdPrs->p_FmPcdPrsRegs->fmpr_l4rres, value);
-+ break;
-+ case (e_FM_PCD_PRS_COUNTERS_SHIM_PARSE_RESULT_RETURNED_WITH_ERR):
-+ WRITE_UINT32(p_FmPcd->p_FmPcdPrs->p_FmPcdPrsRegs->fmpr_srres, value);
-+ break;
-+ case (e_FM_PCD_PRS_COUNTERS_SOFT_PRS_CYCLES):
-+ WRITE_UINT32(p_FmPcd->p_FmPcdPrs->p_FmPcdPrsRegs->fmpr_spcs, value);
-+ break;
-+ case (e_FM_PCD_PRS_COUNTERS_SOFT_PRS_STALL_CYCLES):
-+ WRITE_UINT32(p_FmPcd->p_FmPcdPrs->p_FmPcdPrsRegs->fmpr_spscs, value);
-+ break;
-+ case (e_FM_PCD_PRS_COUNTERS_HARD_PRS_CYCLE_INCL_STALL_CYCLES):
-+ WRITE_UINT32(p_FmPcd->p_FmPcdPrs->p_FmPcdPrsRegs->fmpr_hxscs, value);
-+ break;
-+ case (e_FM_PCD_PRS_COUNTERS_MURAM_READ_CYCLES):
-+ WRITE_UINT32(p_FmPcd->p_FmPcdPrs->p_FmPcdPrsRegs->fmpr_mrcs, value);
-+ break;
-+ case (e_FM_PCD_PRS_COUNTERS_MURAM_READ_STALL_CYCLES):
-+ WRITE_UINT32(p_FmPcd->p_FmPcdPrs->p_FmPcdPrsRegs->fmpr_mrscs, value);
-+ break;
-+ case (e_FM_PCD_PRS_COUNTERS_MURAM_WRITE_CYCLES):
-+ WRITE_UINT32(p_FmPcd->p_FmPcdPrs->p_FmPcdPrsRegs->fmpr_mwcs, value);
-+ break;
-+ case (e_FM_PCD_PRS_COUNTERS_MURAM_WRITE_STALL_CYCLES):
-+ WRITE_UINT32(p_FmPcd->p_FmPcdPrs->p_FmPcdPrsRegs->fmpr_mwscs, value);
-+ break;
-+ case (e_FM_PCD_PRS_COUNTERS_FPM_COMMAND_STALL_CYCLES):
-+ WRITE_UINT32(p_FmPcd->p_FmPcdPrs->p_FmPcdPrsRegs->fmpr_fcscs, value);
-+ break;
-+ case (e_FM_PCD_KG_COUNTERS_TOTAL):
-+ WRITE_UINT32(p_FmPcd->p_FmPcdKg->p_FmPcdKgRegs->fmkg_tpc,value);
-+ break;
-+
-+ /*Policer counters*/
-+ case (e_FM_PCD_PLCR_COUNTERS_YELLOW):
-+ WRITE_UINT32(p_FmPcd->p_FmPcdPlcr->p_FmPcdPlcrRegs->fmpl_ypcnt, value);
-+ break;
-+ case (e_FM_PCD_PLCR_COUNTERS_RED):
-+ WRITE_UINT32(p_FmPcd->p_FmPcdPlcr->p_FmPcdPlcrRegs->fmpl_rpcnt, value);
-+ break;
-+ case (e_FM_PCD_PLCR_COUNTERS_RECOLORED_TO_RED):
-+ WRITE_UINT32(p_FmPcd->p_FmPcdPlcr->p_FmPcdPlcrRegs->fmpl_rrpcnt, value);
-+ break;
-+ case (e_FM_PCD_PLCR_COUNTERS_RECOLORED_TO_YELLOW):
-+ WRITE_UINT32(p_FmPcd->p_FmPcdPlcr->p_FmPcdPlcrRegs->fmpl_rypcnt, value);
-+ break;
-+ case (e_FM_PCD_PLCR_COUNTERS_TOTAL):
-+ WRITE_UINT32(p_FmPcd->p_FmPcdPlcr->p_FmPcdPlcrRegs->fmpl_tpcnt, value);
-+ break;
-+ case (e_FM_PCD_PLCR_COUNTERS_LENGTH_MISMATCH):
-+ WRITE_UINT32(p_FmPcd->p_FmPcdPlcr->p_FmPcdPlcrRegs->fmpl_flmcnt, value);
-+ break;
-+ }
-+
-+ return E_OK;
-+}
-+
-+t_Handle FM_PCD_GetHcPort(t_Handle h_FmPcd)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+ return FmHcGetPort(p_FmPcd->h_Hc);
-+}
-+
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/Pcd/fm_pcd.h
-@@ -0,0 +1,543 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/******************************************************************************
-+ @File fm_pcd.h
-+
-+ @Description FM PCD ...
-+*//***************************************************************************/
-+#ifndef __FM_PCD_H
-+#define __FM_PCD_H
-+
-+#include "std_ext.h"
-+#include "error_ext.h"
-+#include "list_ext.h"
-+#include "fm_pcd_ext.h"
-+#include "fm_common.h"
-+#include "fsl_fman_prs.h"
-+#include "fsl_fman_kg.h"
-+
-+#define __ERR_MODULE__ MODULE_FM_PCD
-+
-+
-+/****************************/
-+/* Defaults */
-+/****************************/
-+#define DEFAULT_plcrAutoRefresh FALSE
-+#define DEFAULT_fmPcdKgErrorExceptions (FM_EX_KG_DOUBLE_ECC | FM_EX_KG_KEYSIZE_OVERFLOW)
-+#define DEFAULT_fmPcdPlcrErrorExceptions (FM_PCD_EX_PLCR_DOUBLE_ECC | FM_PCD_EX_PLCR_INIT_ENTRY_ERROR)
-+#define DEFAULT_fmPcdPlcrExceptions 0
-+#define DEFAULT_fmPcdPrsErrorExceptions (FM_PCD_EX_PRS_DOUBLE_ECC)
-+
-+#define DEFAULT_fmPcdPrsExceptions FM_PCD_EX_PRS_SINGLE_ECC
-+#define DEFAULT_numOfUsedProfilesPerWindow 16
-+#define DEFAULT_numOfSharedPlcrProfiles 4
-+
-+/****************************/
-+/* Network defines */
-+/****************************/
-+#define UDP_HEADER_SIZE 8
-+
-+#define ESP_SPI_OFFSET 0
-+#define ESP_SPI_SIZE 4
-+#define ESP_SEQ_NUM_OFFSET ESP_SPI_SIZE
-+#define ESP_SEQ_NUM_SIZE 4
-+
-+/****************************/
-+/* General defines */
-+/****************************/
-+#define ILLEGAL_CLS_PLAN 0xff
-+#define ILLEGAL_NETENV 0xff
-+
-+#define FM_PCD_MAX_NUM_OF_ALIAS_HDRS 3
-+
-+/****************************/
-+/* Error defines */
-+/****************************/
-+
-+#define FM_PCD_EX_PLCR_DOUBLE_ECC 0x20000000
-+#define FM_PCD_EX_PLCR_INIT_ENTRY_ERROR 0x10000000
-+#define FM_PCD_EX_PLCR_PRAM_SELF_INIT_COMPLETE 0x08000000
-+#define FM_PCD_EX_PLCR_ATOMIC_ACTION_COMPLETE 0x04000000
-+
-+#define GET_FM_PCD_EXCEPTION_FLAG(bitMask, exception) \
-+switch (exception){ \
-+ case e_FM_PCD_KG_EXCEPTION_DOUBLE_ECC: \
-+ bitMask = FM_EX_KG_DOUBLE_ECC; break; \
-+ case e_FM_PCD_PLCR_EXCEPTION_DOUBLE_ECC: \
-+ bitMask = FM_PCD_EX_PLCR_DOUBLE_ECC; break; \
-+ case e_FM_PCD_KG_EXCEPTION_KEYSIZE_OVERFLOW: \
-+ bitMask = FM_EX_KG_KEYSIZE_OVERFLOW; break; \
-+ case e_FM_PCD_PLCR_EXCEPTION_INIT_ENTRY_ERROR: \
-+ bitMask = FM_PCD_EX_PLCR_INIT_ENTRY_ERROR; break; \
-+ case e_FM_PCD_PLCR_EXCEPTION_PRAM_SELF_INIT_COMPLETE: \
-+ bitMask = FM_PCD_EX_PLCR_PRAM_SELF_INIT_COMPLETE; break; \
-+ case e_FM_PCD_PLCR_EXCEPTION_ATOMIC_ACTION_COMPLETE: \
-+ bitMask = FM_PCD_EX_PLCR_ATOMIC_ACTION_COMPLETE; break; \
-+ case e_FM_PCD_PRS_EXCEPTION_DOUBLE_ECC: \
-+ bitMask = FM_PCD_EX_PRS_DOUBLE_ECC; break; \
-+ case e_FM_PCD_PRS_EXCEPTION_SINGLE_ECC: \
-+ bitMask = FM_PCD_EX_PRS_SINGLE_ECC; break; \
-+ default: bitMask = 0;break;}
-+
-+/***********************************************************************/
-+/* Policer defines */
-+/***********************************************************************/
-+#define FM_PCD_PLCR_GCR_STEN 0x40000000
-+#define FM_PCD_PLCR_DOUBLE_ECC 0x80000000
-+#define FM_PCD_PLCR_INIT_ENTRY_ERROR 0x40000000
-+#define FM_PCD_PLCR_PRAM_SELF_INIT_COMPLETE 0x80000000
-+#define FM_PCD_PLCR_ATOMIC_ACTION_COMPLETE 0x40000000
-+
-+/***********************************************************************/
-+/* Memory map */
-+/***********************************************************************/
-+#if defined(__MWERKS__) && !defined(__GNUC__)
-+#pragma pack(push,1)
-+#endif /* defined(__MWERKS__) && ... */
-+
-+
-+typedef struct {
-+/* General Configuration and Status Registers */
-+ volatile uint32_t fmpl_gcr; /* 0x000 FMPL_GCR - FM Policer General Configuration */
-+ volatile uint32_t fmpl_gsr; /* 0x004 FMPL_GSR - FM Policer Global Status Register */
-+ volatile uint32_t fmpl_evr; /* 0x008 FMPL_EVR - FM Policer Event Register */
-+ volatile uint32_t fmpl_ier; /* 0x00C FMPL_IER - FM Policer Interrupt Enable Register */
-+ volatile uint32_t fmpl_ifr; /* 0x010 FMPL_IFR - FM Policer Interrupt Force Register */
-+ volatile uint32_t fmpl_eevr; /* 0x014 FMPL_EEVR - FM Policer Error Event Register */
-+ volatile uint32_t fmpl_eier; /* 0x018 FMPL_EIER - FM Policer Error Interrupt Enable Register */
-+ volatile uint32_t fmpl_eifr; /* 0x01C FMPL_EIFR - FM Policer Error Interrupt Force Register */
-+/* Global Statistic Counters */
-+ volatile uint32_t fmpl_rpcnt; /* 0x020 FMPL_RPC - FM Policer RED Packets Counter */
-+ volatile uint32_t fmpl_ypcnt; /* 0x024 FMPL_YPC - FM Policer YELLOW Packets Counter */
-+ volatile uint32_t fmpl_rrpcnt; /* 0x028 FMPL_RRPC - FM Policer Recolored RED Packet Counter */
-+ volatile uint32_t fmpl_rypcnt; /* 0x02C FMPL_RYPC - FM Policer Recolored YELLOW Packet Counter */
-+ volatile uint32_t fmpl_tpcnt; /* 0x030 FMPL_TPC - FM Policer Total Packet Counter */
-+ volatile uint32_t fmpl_flmcnt; /* 0x034 FMPL_FLMC - FM Policer Frame Length Mismatch Counter */
-+ volatile uint32_t fmpl_res0[21]; /* 0x038 - 0x08B Reserved */
-+/* Profile RAM Access Registers */
-+ volatile uint32_t fmpl_par; /* 0x08C FMPL_PAR - FM Policer Profile Action Register*/
-+ t_FmPcdPlcrProfileRegs profileRegs;
-+/* Error Capture Registers */
-+ volatile uint32_t fmpl_serc; /* 0x100 FMPL_SERC - FM Policer Soft Error Capture */
-+ volatile uint32_t fmpl_upcr; /* 0x104 FMPL_UPCR - FM Policer Uninitialized Profile Capture Register */
-+ volatile uint32_t fmpl_res2; /* 0x108 Reserved */
-+/* Debug Registers */
-+ volatile uint32_t fmpl_res3[61]; /* 0x10C-0x200 Reserved Debug*/
-+/* Profile Selection Mapping Registers Per Port-ID (n=1-11, 16) */
-+ volatile uint32_t fmpl_dpmr; /* 0x200 FMPL_DPMR - FM Policer Default Mapping Register */
-+ volatile uint32_t fmpl_pmr[63]; /*+default 0x204-0x2FF FMPL_PMR1 - FMPL_PMR63, - FM Policer Profile Mapping Registers.
-+ (for port-ID 1-11, only for supported Port-ID registers) */
-+} t_FmPcdPlcrRegs;
-+
-+#if defined(__MWERKS__) && !defined(__GNUC__)
-+#pragma pack(pop)
-+#endif /* defined(__MWERKS__) && ... */
-+
-+
-+/***********************************************************************/
-+/* Driver's internal structures */
-+/***********************************************************************/
-+
-+typedef struct {
-+ bool known;
-+ uint8_t id;
-+} t_FmPcdKgSchemesExtractsEntry;
-+
-+typedef struct {
-+ t_FmPcdKgSchemesExtractsEntry extractsArray[FM_PCD_KG_MAX_NUM_OF_EXTRACTS_PER_KEY];
-+} t_FmPcdKgSchemesExtracts;
-+
-+typedef struct {
-+ t_Handle h_Manip;
-+ bool keepRes;
-+ e_FmPcdEngine nextEngine;
-+ uint8_t parseCode;
-+} t_FmPcdInfoForManip;
-+
-+/**************************************************************************//**
-+ @Description A structure of parameters to communicate
-+ between the port and PCD regarding the KG scheme.
-+*//***************************************************************************/
-+typedef struct {
-+ uint8_t netEnvId; /* in */
-+ uint8_t numOfDistinctionUnits; /* in */
-+ uint8_t unitIds[FM_PCD_MAX_NUM_OF_DISTINCTION_UNITS]; /* in */
-+ uint32_t vector; /* out */
-+} t_NetEnvParams;
-+
-+typedef struct {
-+ bool allocated;
-+ uint8_t ownerId; /* guestId for KG in multi-partition only.
-+ portId for PLCR in any environment */
-+} t_FmPcdAllocMng;
-+
-+typedef struct {
-+ volatile bool lock;
-+ bool used;
-+ uint8_t owners;
-+ uint8_t netEnvId;
-+ uint8_t guestId;
-+ uint8_t baseEntry;
-+ uint16_t sizeOfGrp;
-+ protocolOpt_t optArray[FM_PCD_MAX_NUM_OF_OPTIONS(FM_PCD_MAX_NUM_OF_CLS_PLANS)];
-+} t_FmPcdKgClsPlanGrp;
-+
-+typedef struct {
-+ t_Handle h_FmPcd;
-+ uint8_t schemeId;
-+ t_FmPcdLock *p_Lock;
-+ bool valid;
-+ uint8_t netEnvId;
-+ uint8_t owners;
-+ uint32_t matchVector;
-+ uint32_t ccUnits;
-+ bool nextRelativePlcrProfile;
-+ uint16_t relativeProfileId;
-+ uint16_t numOfProfiles;
-+ t_FmPcdKgKeyOrder orderedArray;
-+ e_FmPcdEngine nextEngine;
-+ e_FmPcdDoneAction doneAction;
-+ bool requiredActionFlag;
-+ uint32_t requiredAction;
-+ bool extractedOrs;
-+ uint8_t bitOffsetInPlcrProfile;
-+ bool directPlcr;
-+#if (DPAA_VERSION >= 11)
-+ bool vspe;
-+#endif
-+} t_FmPcdKgScheme;
-+
-+typedef union {
-+ struct fman_kg_scheme_regs schemeRegs;
-+ struct fman_kg_pe_regs portRegs;
-+ struct fman_kg_cp_regs clsPlanRegs;
-+} u_FmPcdKgIndirectAccessRegs;
-+
-+typedef struct {
-+ struct fman_kg_regs *p_FmPcdKgRegs;
-+ uint32_t schemeExceptionsBitMask;
-+ uint8_t numOfSchemes;
-+ t_Handle h_HwSpinlock;
-+ uint8_t schemesIds[FM_PCD_KG_NUM_OF_SCHEMES];
-+ t_FmPcdKgScheme schemes[FM_PCD_KG_NUM_OF_SCHEMES];
-+ t_FmPcdKgClsPlanGrp clsPlanGrps[FM_MAX_NUM_OF_PORTS];
-+ uint8_t emptyClsPlanGrpId;
-+ t_FmPcdAllocMng schemesMng[FM_PCD_KG_NUM_OF_SCHEMES]; /* only for MASTER ! */
-+ t_FmPcdAllocMng clsPlanBlocksMng[FM_PCD_MAX_NUM_OF_CLS_PLANS/CLS_PLAN_NUM_PER_GRP];
-+ u_FmPcdKgIndirectAccessRegs *p_IndirectAccessRegs;
-+} t_FmPcdKg;
-+
-+typedef struct {
-+ uint16_t profilesBase;
-+ uint16_t numOfProfiles;
-+ t_Handle h_FmPort;
-+} t_FmPcdPlcrMapParam;
-+
-+typedef struct {
-+ uint16_t absoluteProfileId;
-+ t_Handle h_FmPcd;
-+ bool valid;
-+ t_FmPcdLock *p_Lock;
-+ t_FmPcdAllocMng profilesMng;
-+ bool requiredActionFlag;
-+ uint32_t requiredAction;
-+ e_FmPcdEngine nextEngineOnGreen; /**< Green next engine type */
-+ u_FmPcdPlcrNextEngineParams paramsOnGreen; /**< Green next engine params */
-+
-+ e_FmPcdEngine nextEngineOnYellow; /**< Yellow next engine type */
-+ u_FmPcdPlcrNextEngineParams paramsOnYellow; /**< Yellow next engine params */
-+
-+ e_FmPcdEngine nextEngineOnRed; /**< Red next engine type */
-+ u_FmPcdPlcrNextEngineParams paramsOnRed; /**< Red next engine params */
-+} t_FmPcdPlcrProfile;
-+
-+typedef struct {
-+ t_FmPcdPlcrRegs *p_FmPcdPlcrRegs;
-+ uint16_t partPlcrProfilesBase;
-+ uint16_t partNumOfPlcrProfiles;
-+ t_FmPcdPlcrProfile profiles[FM_PCD_PLCR_NUM_ENTRIES];
-+ uint16_t numOfSharedProfiles;
-+ uint16_t sharedProfilesIds[FM_PCD_PLCR_NUM_ENTRIES];
-+ t_FmPcdPlcrMapParam portsMapping[FM_MAX_NUM_OF_PORTS];
-+ t_Handle h_HwSpinlock;
-+ t_Handle h_SwSpinlock;
-+} t_FmPcdPlcr;
-+
-+typedef struct {
-+ uint32_t *p_SwPrsCode;
-+ uint32_t *p_CurrSwPrs;
-+ uint8_t currLabel;
-+ struct fman_prs_regs *p_FmPcdPrsRegs;
-+ t_FmPcdPrsLabelParams labelsTable[FM_PCD_PRS_NUM_OF_LABELS];
-+ uint32_t fmPcdPrsPortIdStatistics;
-+} t_FmPcdPrs;
-+
-+typedef struct {
-+ struct {
-+ e_NetHeaderType hdr;
-+ protocolOpt_t opt; /* only one option !! */
-+ } hdrs[FM_PCD_MAX_NUM_OF_INTERCHANGEABLE_HDRS];
-+} t_FmPcdIntDistinctionUnit;
-+
-+typedef struct {
-+ e_NetHeaderType hdr;
-+ protocolOpt_t opt; /* only one option !! */
-+ e_NetHeaderType aliasHdr;
-+} t_FmPcdNetEnvAliases;
-+
-+typedef struct {
-+ uint8_t netEnvId;
-+ t_Handle h_FmPcd;
-+ t_Handle h_Spinlock;
-+ bool used;
-+ uint8_t owners;
-+ uint8_t clsPlanGrpId;
-+ t_FmPcdIntDistinctionUnit units[FM_PCD_MAX_NUM_OF_DISTINCTION_UNITS];
-+ uint32_t unitsVectors[FM_PCD_MAX_NUM_OF_DISTINCTION_UNITS];
-+ uint32_t lcvs[FM_PCD_PRS_NUM_OF_HDRS];
-+ uint32_t macsecVector;
-+ t_FmPcdNetEnvAliases aliasHdrs[FM_PCD_MAX_NUM_OF_ALIAS_HDRS];
-+} t_FmPcdNetEnv;
-+
-+typedef struct {
-+ struct fman_prs_cfg dfltCfg;
-+ bool plcrAutoRefresh;
-+ uint16_t prsMaxParseCycleLimit;
-+} t_FmPcdDriverParam;
-+
-+typedef struct {
-+ t_Handle h_Fm;
-+ t_Handle h_FmMuram;
-+ t_FmRevisionInfo fmRevInfo;
-+
-+ uint64_t physicalMuramBase;
-+
-+ t_Handle h_Spinlock;
-+ t_List freeLocksLst;
-+ t_List acquiredLocksLst;
-+
-+ t_Handle h_IpcSession; /* relevant for guest only */
-+ bool enabled;
-+ uint8_t guestId; /**< Guest Partition Id */
-+ uint8_t numOfEnabledGuestPartitionsPcds;
-+ char fmPcdModuleName[MODULE_NAME_SIZE];
-+ char fmPcdIpcHandlerModuleName[MODULE_NAME_SIZE]; /* relevant for guest only - this is the master's name */
-+ t_FmPcdNetEnv netEnvs[FM_MAX_NUM_OF_PORTS];
-+ t_FmPcdKg *p_FmPcdKg;
-+ t_FmPcdPlcr *p_FmPcdPlcr;
-+ t_FmPcdPrs *p_FmPcdPrs;
-+
-+ void *p_CcShadow; /**< CC MURAM shadow */
-+ uint32_t ccShadowSize;
-+ uint32_t ccShadowAlign;
-+ volatile bool shadowLock;
-+ t_Handle h_ShadowSpinlock;
-+
-+ t_Handle h_Hc;
-+
-+ uint32_t exceptions;
-+ t_FmPcdExceptionCallback *f_Exception;
-+ t_FmPcdIdExceptionCallback *f_FmPcdIndexedException;
-+ t_Handle h_App;
-+ uintptr_t ipv6FrameIdAddr;
-+ uintptr_t capwapFrameIdAddr;
-+ bool advancedOffloadSupport;
-+
-+ t_FmPcdDriverParam *p_FmPcdDriverParam;
-+} t_FmPcd;
-+
-+#if (DPAA_VERSION >= 11)
-+typedef uint8_t t_FmPcdFrmReplicUpdateType;
-+#define FRM_REPLIC_UPDATE_COUNTER 0x01
-+#define FRM_REPLIC_UPDATE_INFO 0x02
-+#endif /* (DPAA_VERSION >= 11) */
-+/***********************************************************************/
-+/* PCD internal routines */
-+/***********************************************************************/
-+
-+t_Error PcdGetVectorForOpt(t_FmPcd *p_FmPcd, uint8_t netEnvId, protocolOpt_t opt, uint32_t *p_Vector);
-+t_Error PcdGetUnitsVector(t_FmPcd *p_FmPcd, t_NetEnvParams *p_Params);
-+bool PcdNetEnvIsUnitWithoutOpts(t_FmPcd *p_FmPcd, uint8_t netEnvId, uint32_t unitVector);
-+t_Error PcdGetClsPlanGrpParams(t_FmPcd *p_FmPcd, t_FmPcdKgInterModuleClsPlanGrpParams *p_GrpParams);
-+void FmPcdSetClsPlanGrpId(t_FmPcd *p_FmPcd, uint8_t netEnvId, uint8_t clsPlanGrpId);
-+e_NetHeaderType FmPcdGetAliasHdr(t_FmPcd *p_FmPcd, uint8_t netEnvId, e_NetHeaderType hdr);
-+uint8_t FmPcdNetEnvGetUnitIdForSingleHdr(t_FmPcd *p_FmPcd, uint8_t netEnvId, e_NetHeaderType hdr);
-+uint8_t FmPcdNetEnvGetUnitId(t_FmPcd *p_FmPcd, uint8_t netEnvId, e_NetHeaderType hdr, bool interchangeable, protocolOpt_t opt);
-+
-+t_Error FmPcdManipBuildIpReassmScheme(t_FmPcd *p_FmPcd, t_Handle h_NetEnv, t_Handle h_CcTree, t_Handle h_Manip, bool isIpv4, uint8_t groupId);
-+t_Error FmPcdManipDeleteIpReassmSchemes(t_Handle h_Manip);
-+t_Error FmPcdManipBuildCapwapReassmScheme(t_FmPcd *p_FmPcd, t_Handle h_NetEnv, t_Handle h_CcTree, t_Handle h_Manip, uint8_t groupId);
-+t_Error FmPcdManipDeleteCapwapReassmSchemes(t_Handle h_Manip);
-+bool FmPcdManipIpReassmIsIpv6Hdr(t_Handle h_Manip);
-+
-+t_Handle KgConfig( t_FmPcd *p_FmPcd, t_FmPcdParams *p_FmPcdParams);
-+t_Error KgInit(t_FmPcd *p_FmPcd);
-+t_Error KgFree(t_FmPcd *p_FmPcd);
-+void KgSetClsPlan(t_Handle h_FmPcd, t_FmPcdKgInterModuleClsPlanSet *p_Set);
-+bool KgIsSchemeAlwaysDirect(t_Handle h_FmPcd, uint8_t schemeId);
-+void KgEnable(t_FmPcd *p_FmPcd);
-+void KgDisable(t_FmPcd *p_FmPcd);
-+t_Error KgAllocClsPlanEntries(t_Handle h_FmPcd, uint16_t numOfClsPlanEntries, uint8_t guestId, uint8_t *p_First);
-+void KgFreeClsPlanEntries(t_Handle h_FmPcd, uint16_t numOfClsPlanEntries, uint8_t guestId, uint8_t base);
-+
-+/* only for MULTI partittion */
-+t_Error FmPcdKgAllocSchemes(t_Handle h_FmPcd, uint8_t numOfSchemes, uint8_t guestId, uint8_t *p_SchemesIds);
-+t_Error FmPcdKgFreeSchemes(t_Handle h_FmPcd, uint8_t numOfSchemes, uint8_t guestId, uint8_t *p_SchemesIds);
-+/* only for SINGLE partittion */
-+t_Error KgBindPortToSchemes(t_Handle h_FmPcd , uint8_t hardwarePortId, uint32_t spReg);
-+
-+t_FmPcdLock *FmPcdAcquireLock(t_Handle h_FmPcd);
-+void FmPcdReleaseLock(t_Handle h_FmPcd, t_FmPcdLock *p_Lock);
-+
-+t_Handle PlcrConfig(t_FmPcd *p_FmPcd, t_FmPcdParams *p_FmPcdParams);
-+t_Error PlcrInit(t_FmPcd *p_FmPcd);
-+t_Error PlcrFree(t_FmPcd *p_FmPcd);
-+void PlcrEnable(t_FmPcd *p_FmPcd);
-+void PlcrDisable(t_FmPcd *p_FmPcd);
-+uint16_t PlcrAllocProfilesForPartition(t_FmPcd *p_FmPcd, uint16_t base, uint16_t numOfProfiles, uint8_t guestId);
-+void PlcrFreeProfilesForPartition(t_FmPcd *p_FmPcd, uint16_t base, uint16_t numOfProfiles, uint8_t guestId);
-+t_Error PlcrSetPortProfiles(t_FmPcd *p_FmPcd,
-+ uint8_t hardwarePortId,
-+ uint16_t numOfProfiles,
-+ uint16_t base);
-+t_Error PlcrClearPortProfiles(t_FmPcd *p_FmPcd, uint8_t hardwarePortId);
-+
-+t_Handle PrsConfig(t_FmPcd *p_FmPcd,t_FmPcdParams *p_FmPcdParams);
-+t_Error PrsInit(t_FmPcd *p_FmPcd);
-+void PrsEnable(t_FmPcd *p_FmPcd);
-+void PrsDisable(t_FmPcd *p_FmPcd);
-+void PrsFree(t_FmPcd *p_FmPcd );
-+t_Error PrsIncludePortInStatistics(t_FmPcd *p_FmPcd, uint8_t hardwarePortId, bool include);
-+
-+t_Error FmPcdCcGetGrpParams(t_Handle treeId, uint8_t grpId, uint32_t *p_GrpBits, uint8_t *p_GrpBase);
-+uint8_t FmPcdCcGetOffset(t_Handle h_CcNode);
-+uint8_t FmPcdCcGetParseCode(t_Handle h_CcNode);
-+uint16_t FmPcdCcGetNumOfKeys(t_Handle h_CcNode);
-+t_Error ValidateNextEngineParams(t_Handle h_FmPcd, t_FmPcdCcNextEngineParams *p_FmPcdCcNextEngineParams, e_FmPcdCcStatsMode supportedStatsMode);
-+
-+void FmPcdManipUpdateOwner(t_Handle h_Manip, bool add);
-+t_Error FmPcdManipCheckParamsForCcNextEngine(t_FmPcdCcNextEngineParams *p_InfoForManip, uint32_t *requiredAction);
-+void FmPcdManipUpdateAdResultForCc(t_Handle h_Manip,
-+ t_FmPcdCcNextEngineParams *p_CcNextEngineParams,
-+ t_Handle p_Ad,
-+ t_Handle *p_AdNewPtr);
-+void FmPcdManipUpdateAdContLookupForCc(t_Handle h_Manip, t_Handle p_Ad, t_Handle *p_AdNew, uint32_t adTableOffset);
-+void FmPcdManipUpdateOwner(t_Handle h_Manip, bool add);
-+t_Error FmPcdManipCheckParamsWithCcNodeParams(t_Handle h_Manip, t_Handle h_FmPcdCcNode);
-+#ifdef FM_CAPWAP_SUPPORT
-+t_Handle FmPcdManipApplSpecificBuild(void);
-+bool FmPcdManipIsCapwapApplSpecific(t_Handle h_Manip);
-+#endif /* FM_CAPWAP_SUPPORT */
-+#if (DPAA_VERSION >= 11)
-+void * FrmReplicGroupGetSourceTableDescriptor(t_Handle h_ReplicGroup);
-+void FrmReplicGroupUpdateOwner(t_Handle h_ReplicGroup, bool add);
-+void FrmReplicGroupUpdateAd(t_Handle h_ReplicGroup, void *p_Ad, t_Handle *h_AdNew);
-+
-+void FmPcdCcGetAdTablesThatPointOnReplicGroup(t_Handle h_Node,
-+ t_Handle h_ReplicGroup,
-+ t_List *p_AdTables,
-+ uint32_t *p_NumOfAdTables);
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+void EnqueueNodeInfoToRelevantLst(t_List *p_List, t_CcNodeInformation *p_CcInfo, t_Handle h_Spinlock);
-+void DequeueNodeInfoFromRelevantLst(t_List *p_List, t_Handle h_Info, t_Handle h_Spinlock);
-+t_CcNodeInformation* FindNodeInfoInReleventLst(t_List *p_List, t_Handle h_Info, t_Handle h_Spinlock);
-+t_List *FmPcdManipGetSpinlock(t_Handle h_Manip);
-+t_List *FmPcdManipGetNodeLstPointedOnThisManip(t_Handle h_Manip);
-+
-+typedef struct
-+{
-+ t_Handle h_StatsAd;
-+ t_Handle h_StatsCounters;
-+#if (DPAA_VERSION >= 11)
-+ t_Handle h_StatsFLRs;
-+#endif /* (DPAA_VERSION >= 11) */
-+} t_FmPcdCcStatsParams;
-+
-+void NextStepAd(t_Handle h_Ad,
-+ t_FmPcdCcStatsParams *p_FmPcdCcStatsParams,
-+ t_FmPcdCcNextEngineParams *p_FmPcdCcNextEngineParams,
-+ t_FmPcd *p_FmPcd);
-+void ReleaseLst(t_List *p_List);
-+
-+static __inline__ t_Handle FmPcdGetMuramHandle(t_Handle h_FmPcd)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+ ASSERT_COND(p_FmPcd);
-+ return p_FmPcd->h_FmMuram;
-+}
-+
-+static __inline__ uint64_t FmPcdGetMuramPhysBase(t_Handle h_FmPcd)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+ ASSERT_COND(p_FmPcd);
-+ return p_FmPcd->physicalMuramBase;
-+}
-+
-+static __inline__ uint32_t FmPcdLockSpinlock(t_FmPcdLock *p_Lock)
-+{
-+ ASSERT_COND(p_Lock);
-+ return XX_LockIntrSpinlock(p_Lock->h_Spinlock);
-+}
-+
-+static __inline__ void FmPcdUnlockSpinlock(t_FmPcdLock *p_Lock, uint32_t flags)
-+{
-+ ASSERT_COND(p_Lock);
-+ XX_UnlockIntrSpinlock(p_Lock->h_Spinlock, flags);
-+}
-+
-+static __inline__ bool FmPcdLockTryLock(t_FmPcdLock *p_Lock)
-+{
-+ uint32_t intFlags;
-+
-+ ASSERT_COND(p_Lock);
-+ intFlags = XX_LockIntrSpinlock(p_Lock->h_Spinlock);
-+ if (p_Lock->flag)
-+ {
-+ XX_UnlockIntrSpinlock(p_Lock->h_Spinlock, intFlags);
-+ return FALSE;
-+ }
-+ p_Lock->flag = TRUE;
-+ XX_UnlockIntrSpinlock(p_Lock->h_Spinlock, intFlags);
-+ return TRUE;
-+}
-+
-+static __inline__ void FmPcdLockUnlock(t_FmPcdLock *p_Lock)
-+{
-+ ASSERT_COND(p_Lock);
-+ p_Lock->flag = FALSE;
-+}
-+
-+
-+#endif /* __FM_PCD_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/Pcd/fm_pcd_ipc.h
-@@ -0,0 +1,280 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/**************************************************************************//**
-+ @File fm_pcd_ipc.h
-+
-+ @Description FM PCD Inter-Partition prototypes, structures and definitions.
-+*//***************************************************************************/
-+#ifndef __FM_PCD_IPC_H
-+#define __FM_PCD_IPC_H
-+
-+#include "std_ext.h"
-+
-+
-+/**************************************************************************//**
-+ @Group FM_grp Frame Manager API
-+
-+ @Description FM API functions, definitions and enums
-+
-+ @{
-+*//***************************************************************************/
-+
-+
-+#if defined(__MWERKS__) && !defined(__GNUC__)
-+#pragma pack(push,1)
-+#endif /* defined(__MWERKS__) && ... */
-+
-+/**************************************************************************//**
-+ @Description Structure for getting a sw parser address according to a label
-+ Fields commented 'IN' are passed by the port module to be used
-+ by the FM module.
-+ Fields commented 'OUT' will be filled by FM before returning to port.
-+*//***************************************************************************/
-+typedef _Packed struct t_FmPcdIpcSwPrsLable
-+{
-+ uint32_t enumHdr; /**< IN. The existence of this header will invoke
-+ the sw parser code. */
-+ uint8_t indexPerHdr; /**< IN. Normally 0, if more than one sw parser
-+ attachments for the same header, use this
-+
-+ index to distinguish between them. */
-+} _PackedType t_FmPcdIpcSwPrsLable;
-+
-+/**************************************************************************//**
-+ @Description Structure for port-PCD communication.
-+ Fields commented 'IN' are passed by the port module to be used
-+ by the FM module.
-+ Fields commented 'OUT' will be filled by FM before returning to port.
-+ Some fields are optional (depending on configuration) and
-+ will be analized by the port and FM modules accordingly.
-+*//***************************************************************************/
-+
-+typedef struct t_FmPcdIpcKgSchemesParams
-+{
-+ uint8_t guestId;
-+ uint8_t numOfSchemes;
-+ uint8_t schemesIds[FM_PCD_KG_NUM_OF_SCHEMES];
-+} _PackedType t_FmPcdIpcKgSchemesParams;
-+
-+typedef struct t_FmPcdIpcKgClsPlanParams
-+{
-+ uint8_t guestId;
-+ uint16_t numOfClsPlanEntries;
-+ uint8_t clsPlanBase;
-+} _PackedType t_FmPcdIpcKgClsPlanParams;
-+
-+typedef _Packed struct t_FmPcdIpcPrsIncludePort
-+{
-+ uint8_t hardwarePortId;
-+ bool include;
-+} _PackedType t_FmPcdIpcPrsIncludePort;
-+
-+
-+#define FM_PCD_MAX_REPLY_SIZE 16
-+#define FM_PCD_MAX_MSG_SIZE 36
-+#define FM_PCD_MAX_REPLY_BODY_SIZE 36
-+
-+typedef _Packed struct {
-+ uint32_t msgId;
-+ uint8_t msgBody[FM_PCD_MAX_MSG_SIZE];
-+} _PackedType t_FmPcdIpcMsg;
-+
-+typedef _Packed struct t_FmPcdIpcReply {
-+ uint32_t error;
-+ uint8_t replyBody[FM_PCD_MAX_REPLY_BODY_SIZE];
-+} _PackedType t_FmPcdIpcReply;
-+
-+typedef _Packed struct t_FmIpcResourceAllocParams {
-+ uint8_t guestId;
-+ uint16_t base;
-+ uint16_t num;
-+}_PackedType t_FmIpcResourceAllocParams;
-+
-+#if defined(__MWERKS__) && !defined(__GNUC__)
-+#pragma pack(pop)
-+#endif /* defined(__MWERKS__) && ... */
-+
-+
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_ALLOC_KG_SCHEMES
-+
-+ @Description Used by FM PCD front-end in order to allocate KG resources
-+
-+ @Param[in/out] t_FmPcdIpcKgAllocParams Pointer
-+*//***************************************************************************/
-+#define FM_PCD_ALLOC_KG_SCHEMES 3
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_FREE_KG_SCHEMES
-+
-+ @Description Used by FM PCD front-end in order to Free KG resources
-+
-+ @Param[in/out] t_FmPcdIpcKgSchemesParams Pointer
-+*//***************************************************************************/
-+#define FM_PCD_FREE_KG_SCHEMES 4
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_ALLOC_PROFILES
-+
-+ @Description Used by FM PCD front-end in order to allocate Policer profiles
-+
-+ @Param[in/out] t_FmIpcResourceAllocParams Pointer
-+*//***************************************************************************/
-+#define FM_PCD_ALLOC_PROFILES 5
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_FREE_PROFILES
-+
-+ @Description Used by FM PCD front-end in order to Free Policer profiles
-+
-+ @Param[in/out] t_FmIpcResourceAllocParams Pointer
-+*//***************************************************************************/
-+#define FM_PCD_FREE_PROFILES 6
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_SET_PORT_PROFILES
-+
-+ @Description Used by FM PCD front-end in order to allocate Policer profiles
-+ for specific port
-+
-+ @Param[in/out] t_FmIpcResourceAllocParams Pointer
-+*//***************************************************************************/
-+#define FM_PCD_SET_PORT_PROFILES 7
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_CLEAR_PORT_PROFILES
-+
-+ @Description Used by FM PCD front-end in order to allocate Policer profiles
-+ for specific port
-+
-+ @Param[in/out] t_FmIpcResourceAllocParams Pointer
-+*//***************************************************************************/
-+#define FM_PCD_CLEAR_PORT_PROFILES 8
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_GET_PHYS_MURAM_BASE
-+
-+ @Description Used by FM PCD front-end in order to get MURAM base address
-+
-+ @Param[in/out] t_FmPcdIcPhysAddr Pointer
-+*//***************************************************************************/
-+#define FM_PCD_GET_PHYS_MURAM_BASE 9
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_GET_SW_PRS_OFFSET
-+
-+ @Description Used by FM front-end to get the SW parser offset of the start of
-+ code relevant to a given label.
-+
-+ @Param[in/out] t_FmPcdIpcSwPrsLable Pointer
-+*//***************************************************************************/
-+#define FM_PCD_GET_SW_PRS_OFFSET 10
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_MASTER_IS_ENABLED
-+
-+ @Description Used by FM front-end in order to verify
-+ PCD enablement.
-+
-+ @Param[in] bool Pointer
-+*//***************************************************************************/
-+#define FM_PCD_MASTER_IS_ENABLED 15
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_GUEST_DISABLE
-+
-+ @Description Used by FM front-end to inform back-end when
-+ front-end PCD is disabled
-+
-+ @Param[in] None
-+*//***************************************************************************/
-+#define FM_PCD_GUEST_DISABLE 16
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_FREE_KG_CLSPLAN
-+
-+ @Description Used by FM PCD front-end in order to Free KG classification plan entries
-+
-+ @Param[in/out] t_FmPcdIpcKgClsPlanParams Pointer
-+*//***************************************************************************/
-+#define FM_PCD_FREE_KG_CLSPLAN 22
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_ALLOC_KG_CLSPLAN
-+
-+ @Description Used by FM PCD front-end in order to allocate KG classification plan entries
-+
-+ @Param[in/out] t_FmPcdIpcKgClsPlanParams Pointer
-+*//***************************************************************************/
-+#define FM_PCD_ALLOC_KG_CLSPLAN 23
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_MASTER_IS_ALIVE
-+
-+ @Description Used by FM front-end to check that back-end exists
-+
-+ @Param[in] None
-+*//***************************************************************************/
-+#define FM_PCD_MASTER_IS_ALIVE 24
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_GET_COUNTER
-+
-+ @Description Used by FM front-end to read PCD counters
-+
-+ @Param[in/out] t_FmPcdIpcGetCounter Pointer
-+*//***************************************************************************/
-+#define FM_PCD_GET_COUNTER 25
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_PRS_INC_PORT_STATS
-+
-+ @Description Used by FM front-end to set/clear statistics for port
-+
-+ @Param[in/out] t_FmPcdIpcPrsIncludePort Pointer
-+*//***************************************************************************/
-+#define FM_PCD_PRS_INC_PORT_STATS 26
-+
-+#if (DPAA_VERSION >= 11)
-+/* TODO - doc */
-+#define FM_PCD_ALLOC_SP 27
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+
-+/** @} */ /* end of FM_PCD_IPC_grp group */
-+/** @} */ /* end of FM_grp group */
-+
-+
-+#endif /* __FM_PCD_IPC_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/Pcd/fm_plcr.c
-@@ -0,0 +1,1847 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/******************************************************************************
-+ @File fm_plcr.c
-+
-+ @Description FM PCD POLICER...
-+*//***************************************************************************/
-+#include <linux/math64.h>
-+#include "std_ext.h"
-+#include "error_ext.h"
-+#include "string_ext.h"
-+#include "debug_ext.h"
-+#include "net_ext.h"
-+#include "fm_ext.h"
-+
-+#include "fm_common.h"
-+#include "fm_pcd.h"
-+#include "fm_hc.h"
-+#include "fm_pcd_ipc.h"
-+#include "fm_plcr.h"
-+
-+
-+/****************************************/
-+/* static functions */
-+/****************************************/
-+
-+static uint32_t PlcrProfileLock(t_Handle h_Profile)
-+{
-+ ASSERT_COND(h_Profile);
-+ return FmPcdLockSpinlock(((t_FmPcdPlcrProfile *)h_Profile)->p_Lock);
-+}
-+
-+static void PlcrProfileUnlock(t_Handle h_Profile, uint32_t intFlags)
-+{
-+ ASSERT_COND(h_Profile);
-+ FmPcdUnlockSpinlock(((t_FmPcdPlcrProfile *)h_Profile)->p_Lock, intFlags);
-+}
-+
-+static bool PlcrProfileFlagTryLock(t_Handle h_Profile)
-+{
-+ ASSERT_COND(h_Profile);
-+ return FmPcdLockTryLock(((t_FmPcdPlcrProfile *)h_Profile)->p_Lock);
-+}
-+
-+static void PlcrProfileFlagUnlock(t_Handle h_Profile)
-+{
-+ ASSERT_COND(h_Profile);
-+ FmPcdLockUnlock(((t_FmPcdPlcrProfile *)h_Profile)->p_Lock);
-+}
-+
-+static uint32_t PlcrHwLock(t_Handle h_FmPcdPlcr)
-+{
-+ ASSERT_COND(h_FmPcdPlcr);
-+ return XX_LockIntrSpinlock(((t_FmPcdPlcr*)h_FmPcdPlcr)->h_HwSpinlock);
-+}
-+
-+static void PlcrHwUnlock(t_Handle h_FmPcdPlcr, uint32_t intFlags)
-+{
-+ ASSERT_COND(h_FmPcdPlcr);
-+ XX_UnlockIntrSpinlock(((t_FmPcdPlcr*)h_FmPcdPlcr)->h_HwSpinlock, intFlags);
-+}
-+
-+static uint32_t PlcrSwLock(t_Handle h_FmPcdPlcr)
-+{
-+ ASSERT_COND(h_FmPcdPlcr);
-+ return XX_LockIntrSpinlock(((t_FmPcdPlcr*)h_FmPcdPlcr)->h_SwSpinlock);
-+}
-+
-+static void PlcrSwUnlock(t_Handle h_FmPcdPlcr, uint32_t intFlags)
-+{
-+ ASSERT_COND(h_FmPcdPlcr);
-+ XX_UnlockIntrSpinlock(((t_FmPcdPlcr*)h_FmPcdPlcr)->h_SwSpinlock, intFlags);
-+}
-+
-+static bool IsProfileShared(t_Handle h_FmPcd, uint16_t absoluteProfileId)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+ uint16_t i;
-+
-+ SANITY_CHECK_RETURN_VALUE(p_FmPcd, E_INVALID_HANDLE, FALSE);
-+
-+ for (i=0;i<p_FmPcd->p_FmPcdPlcr->numOfSharedProfiles;i++)
-+ if (p_FmPcd->p_FmPcdPlcr->sharedProfilesIds[i] == absoluteProfileId)
-+ return TRUE;
-+ return FALSE;
-+}
-+
-+static t_Error SetProfileNia(t_FmPcd *p_FmPcd, e_FmPcdEngine nextEngine, u_FmPcdPlcrNextEngineParams *p_NextEngineParams, uint32_t *nextAction)
-+{
-+ uint32_t nia;
-+ uint16_t absoluteProfileId;
-+ uint8_t relativeSchemeId, physicalSchemeId;
-+
-+ nia = FM_PCD_PLCR_NIA_VALID;
-+
-+ switch (nextEngine)
-+ {
-+ case e_FM_PCD_DONE :
-+ switch (p_NextEngineParams->action)
-+ {
-+ case e_FM_PCD_DROP_FRAME :
-+ nia |= GET_NIA_BMI_AC_DISCARD_FRAME(p_FmPcd);
-+ break;
-+ case e_FM_PCD_ENQ_FRAME:
-+ nia |= GET_NIA_BMI_AC_ENQ_FRAME(p_FmPcd);
-+ break;
-+ default:
-+ RETURN_ERROR(MAJOR, E_INVALID_SELECTION, NO_MSG);
-+ }
-+ break;
-+ case e_FM_PCD_KG:
-+ physicalSchemeId = FmPcdKgGetSchemeId(p_NextEngineParams->h_DirectScheme);
-+ relativeSchemeId = FmPcdKgGetRelativeSchemeId(p_FmPcd, physicalSchemeId);
-+ if (relativeSchemeId >= FM_PCD_KG_NUM_OF_SCHEMES)
-+ RETURN_ERROR(MAJOR, E_NOT_IN_RANGE, NO_MSG);
-+ if (!FmPcdKgIsSchemeValidSw(p_NextEngineParams->h_DirectScheme))
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("Invalid direct scheme."));
-+ if (!KgIsSchemeAlwaysDirect(p_FmPcd, relativeSchemeId))
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("Policer Profile may point only to a scheme that is always direct."));
-+ nia |= NIA_ENG_KG | NIA_KG_DIRECT | physicalSchemeId;
-+ break;
-+ case e_FM_PCD_PLCR:
-+ absoluteProfileId = ((t_FmPcdPlcrProfile *)p_NextEngineParams->h_Profile)->absoluteProfileId;
-+ if (!IsProfileShared(p_FmPcd, absoluteProfileId))
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("Next profile must be a shared profile"));
-+ if (!FmPcdPlcrIsProfileValid(p_FmPcd, absoluteProfileId))
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("Invalid profile "));
-+ nia |= NIA_ENG_PLCR | NIA_PLCR_ABSOLUTE | absoluteProfileId;
-+ break;
-+ default:
-+ RETURN_ERROR(MAJOR, E_INVALID_SELECTION, NO_MSG);
-+ }
-+
-+ *nextAction = nia;
-+
-+ return E_OK;
-+}
-+
-+static uint32_t CalcFPP(uint32_t fpp)
-+{
-+ if (fpp > 15)
-+ return 15 - (0x1f - fpp);
-+ else
-+ return 16 + fpp;
-+}
-+
-+static void GetInfoRateReg(e_FmPcdPlcrRateMode rateMode,
-+ uint32_t rate,
-+ uint64_t tsuInTenthNano,
-+ uint32_t fppShift,
-+ uint64_t *p_Integer,
-+ uint64_t *p_Fraction)
-+{
-+ uint64_t tmp, div;
-+
-+ if (rateMode == e_FM_PCD_PLCR_BYTE_MODE)
-+ {
-+ /* now we calculate the initial integer for the bigger rate */
-+ /* from Kbps to Bytes/TSU */
-+ tmp = (uint64_t)rate;
-+ tmp *= 1000; /* kb --> b */
-+ tmp *= tsuInTenthNano; /* bps --> bpTsu(in 10nano) */
-+
-+ div = 1000000000; /* nano */
-+ div *= 10; /* 10 nano */
-+ div *= 8; /* bit to byte */
-+ }
-+ else
-+ {
-+ /* now we calculate the initial integer for the bigger rate */
-+ /* from Kbps to Bytes/TSU */
-+ tmp = (uint64_t)rate;
-+ tmp *= tsuInTenthNano; /* bps --> bpTsu(in 10nano) */
-+
-+ div = 1000000000; /* nano */
-+ div *= 10; /* 10 nano */
-+ }
-+ *p_Integer = div64_u64(tmp<<fppShift, div);
-+
-+ /* for calculating the fraction, we will recalculate cir and deduct the integer.
-+ * For precision, we will multiply by 2^16. we do not divid back, since we write
-+ * this value as fraction - see spec.
-+ */
-+ *p_Fraction = div64_u64(((tmp<<fppShift)<<16) - ((*p_Integer<<16)*div), div);
-+}
-+
-+/* .......... */
-+
-+static void CalcRates(uint32_t bitFor1Micro,
-+ t_FmPcdPlcrNonPassthroughAlgParams *p_NonPassthroughAlgParam,
-+ uint32_t *cir,
-+ uint32_t *cbs,
-+ uint32_t *pir_eir,
-+ uint32_t *pbs_ebs,
-+ uint32_t *fpp)
-+{
-+ uint64_t integer, fraction;
-+ uint32_t temp, tsuInTenthNanos;
-+ uint8_t fppShift=0;
-+
-+ /* we want the tsu to count 10 nano for better precision normally tsu is 3.9 nano, now we will get 39 */
-+ tsuInTenthNanos = (uint32_t)(1000*10/(1 << bitFor1Micro));
-+
-+ /* we choose the faster rate to calibrate fpp */
-+ /* The meaning of this step:
-+ * when fppShift is 0 it means all TS bits are treated as integer and TSU is the TS LSB count.
-+ * In this configuration we calculate the integer and fraction that represent the higher infoRate
-+ * When this is done, we can tell where we have "spare" unused bits and optimize the division of TS
-+ * into "integer" and "fraction" where the logic is - as many bits as possible for integer at
-+ * high rate, as many bits as possible for fraction at low rate.
-+ */
-+ if (p_NonPassthroughAlgParam->committedInfoRate > p_NonPassthroughAlgParam->peakOrExcessInfoRate)
-+ GetInfoRateReg(p_NonPassthroughAlgParam->rateMode, p_NonPassthroughAlgParam->committedInfoRate, tsuInTenthNanos, 0, &integer, &fraction);
-+ else
-+ GetInfoRateReg(p_NonPassthroughAlgParam->rateMode, p_NonPassthroughAlgParam->peakOrExcessInfoRate, tsuInTenthNanos, 0, &integer, &fraction);
-+
-+ /* we shift integer, as in cir/pir it is represented by the MSB 16 bits, and
-+ * the LSB bits are for the fraction */
-+ temp = (uint32_t)((integer<<16) & 0x00000000FFFFFFFF);
-+ /* temp is effected by the rate. For low rates it may be as low as 0, and then we'll
-+ * take max FP = 31.
-+ * For high rates it will never exceed the 32 bit reg (after the 16 shift), as it is
-+ * limited by the 10G physical port.
-+ */
-+ if (temp != 0)
-+ {
-+ /* In this case, the largest rate integer is non 0, if it does not occupy all (high) 16
-+ * bits of the PIR_EIR we can use this fact and enlarge it to occupy all 16 bits.
-+ * The logic is to have as many bits for integer in the higher rates, but if we have "0"s
-+ * in the integer part of the cir/pir register, than these bits are wasted. So we want
-+ * to use these bits for the fraction. in this way we will have for fraction - the number
-+ * of "0" bits and the rest - for integer.
-+ * In other words: For each bit we shift it in PIR_EIR, we move the FP in the TS
-+ * one bit to the left - preserving the relationship and achieving more bits
-+ * for integer in the TS.
-+ */
-+
-+ /* count zeroes left of the higher used bit (in order to shift the value such that
-+ * unused bits may be used for fraction).
-+ */
-+ while ((temp & 0x80000000) == 0)
-+ {
-+ temp = temp << 1;
-+ fppShift++;
-+ }
-+ if (fppShift > 15)
-+ {
-+ REPORT_ERROR(MAJOR, E_INVALID_SELECTION, ("timeStampPeriod to Information rate ratio is too small"));
-+ return;
-+ }
-+ }
-+ else
-+ {
-+ temp = (uint32_t)fraction; /* fraction will alyas be smaller than 2^16 */
-+ if (!temp)
-+ /* integer and fraction are 0, we set FP to its max val */
-+ fppShift = 31;
-+ else
-+ {
-+ /* integer was 0 but fraction is not. FP is 16 for the fraction,
-+ * + all left zeroes of the fraction. */
-+ fppShift=16;
-+ /* count zeroes left of the higher used bit (in order to shift the value such that
-+ * unused bits may be used for fraction).
-+ */
-+ while ((temp & 0x8000) == 0)
-+ {
-+ temp = temp << 1;
-+ fppShift++;
-+ }
-+ }
-+ }
-+
-+ /*
-+ * This means that the FM TS register will now be used so that 'fppShift' bits are for
-+ * fraction and the rest for integer */
-+ /* now we re-calculate cir and pir_eir with the calculated FP */
-+ GetInfoRateReg(p_NonPassthroughAlgParam->rateMode, p_NonPassthroughAlgParam->committedInfoRate, tsuInTenthNanos, fppShift, &integer, &fraction);
-+ *cir = (uint32_t)(integer << 16 | (fraction & 0xFFFF));
-+ GetInfoRateReg(p_NonPassthroughAlgParam->rateMode, p_NonPassthroughAlgParam->peakOrExcessInfoRate, tsuInTenthNanos, fppShift, &integer, &fraction);
-+ *pir_eir = (uint32_t)(integer << 16 | (fraction & 0xFFFF));
-+
-+ *cbs = p_NonPassthroughAlgParam->committedBurstSize;
-+ *pbs_ebs = p_NonPassthroughAlgParam->peakOrExcessBurstSize;
-+
-+ /* convert FP as it should be written to reg.
-+ * 0-15 --> 16-31
-+ * 16-31 --> 0-15
-+ */
-+ *fpp = CalcFPP(fppShift);
-+}
-+
-+static void WritePar(t_FmPcd *p_FmPcd, uint32_t par)
-+{
-+ t_FmPcdPlcrRegs *p_FmPcdPlcrRegs = p_FmPcd->p_FmPcdPlcr->p_FmPcdPlcrRegs;
-+
-+ ASSERT_COND(FmIsMaster(p_FmPcd->h_Fm));
-+ WRITE_UINT32(p_FmPcdPlcrRegs->fmpl_par, par);
-+
-+ while (GET_UINT32(p_FmPcdPlcrRegs->fmpl_par) & FM_PCD_PLCR_PAR_GO) ;
-+}
-+
-+static t_Error BuildProfileRegs(t_FmPcd *p_FmPcd,
-+ t_FmPcdPlcrProfileParams *p_ProfileParams,
-+ t_FmPcdPlcrProfileRegs *p_PlcrRegs)
-+{
-+ t_Error err = E_OK;
-+ uint32_t pemode, gnia, ynia, rnia, bitFor1Micro;
-+
-+ ASSERT_COND(p_FmPcd);
-+
-+ bitFor1Micro = FmGetTimeStampScale(p_FmPcd->h_Fm);
-+ if (bitFor1Micro == 0)
-+ RETURN_ERROR(MAJOR, E_NOT_AVAILABLE, ("Timestamp scale"));
-+
-+/* Set G, Y, R Nia */
-+ err = SetProfileNia(p_FmPcd, p_ProfileParams->nextEngineOnGreen, &(p_ProfileParams->paramsOnGreen), &gnia);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ err = SetProfileNia(p_FmPcd, p_ProfileParams->nextEngineOnYellow, &(p_ProfileParams->paramsOnYellow), &ynia);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ err = SetProfileNia(p_FmPcd, p_ProfileParams->nextEngineOnRed, &(p_ProfileParams->paramsOnRed), &rnia);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+
-+/* Mode fmpl_pemode */
-+ pemode = FM_PCD_PLCR_PEMODE_PI;
-+
-+ switch (p_ProfileParams->algSelection)
-+ {
-+ case e_FM_PCD_PLCR_PASS_THROUGH:
-+ p_PlcrRegs->fmpl_pecir = 0;
-+ p_PlcrRegs->fmpl_pecbs = 0;
-+ p_PlcrRegs->fmpl_pepepir_eir = 0;
-+ p_PlcrRegs->fmpl_pepbs_ebs = 0;
-+ p_PlcrRegs->fmpl_pelts = 0;
-+ p_PlcrRegs->fmpl_pects = 0;
-+ p_PlcrRegs->fmpl_pepts_ets = 0;
-+ pemode &= ~FM_PCD_PLCR_PEMODE_ALG_MASK;
-+ switch (p_ProfileParams->colorMode)
-+ {
-+ case e_FM_PCD_PLCR_COLOR_BLIND:
-+ pemode |= FM_PCD_PLCR_PEMODE_CBLND;
-+ switch (p_ProfileParams->color.dfltColor)
-+ {
-+ case e_FM_PCD_PLCR_GREEN:
-+ pemode &= ~FM_PCD_PLCR_PEMODE_DEFC_MASK;
-+ break;
-+ case e_FM_PCD_PLCR_YELLOW:
-+ pemode |= FM_PCD_PLCR_PEMODE_DEFC_Y;
-+ break;
-+ case e_FM_PCD_PLCR_RED:
-+ pemode |= FM_PCD_PLCR_PEMODE_DEFC_R;
-+ break;
-+ case e_FM_PCD_PLCR_OVERRIDE:
-+ pemode |= FM_PCD_PLCR_PEMODE_DEFC_OVERRIDE;
-+ break;
-+ default:
-+ RETURN_ERROR(MAJOR, E_INVALID_SELECTION, NO_MSG);
-+ }
-+
-+ break;
-+ case e_FM_PCD_PLCR_COLOR_AWARE:
-+ pemode &= ~FM_PCD_PLCR_PEMODE_CBLND;
-+ break;
-+ default:
-+ RETURN_ERROR(MAJOR, E_INVALID_SELECTION, NO_MSG);
-+ }
-+ break;
-+
-+ case e_FM_PCD_PLCR_RFC_2698:
-+ /* Select algorithm MODE[ALG] = "01" */
-+ pemode |= FM_PCD_PLCR_PEMODE_ALG_RFC2698;
-+ if (p_ProfileParams->nonPassthroughAlgParams.committedInfoRate > p_ProfileParams->nonPassthroughAlgParams.peakOrExcessInfoRate)
-+ RETURN_ERROR(MAJOR, E_INVALID_SELECTION, ("in RFC2698 Peak rate must be equal or larger than committedInfoRate."));
-+ goto cont_rfc;
-+ case e_FM_PCD_PLCR_RFC_4115:
-+ /* Select algorithm MODE[ALG] = "10" */
-+ pemode |= FM_PCD_PLCR_PEMODE_ALG_RFC4115;
-+cont_rfc:
-+ /* Select Color-Blind / Color-Aware operation (MODE[CBLND]) */
-+ switch (p_ProfileParams->colorMode)
-+ {
-+ case e_FM_PCD_PLCR_COLOR_BLIND:
-+ pemode |= FM_PCD_PLCR_PEMODE_CBLND;
-+ break;
-+ case e_FM_PCD_PLCR_COLOR_AWARE:
-+ pemode &= ~FM_PCD_PLCR_PEMODE_CBLND;
-+ /*In color aware more select override color interpretation (MODE[OVCLR]) */
-+ switch (p_ProfileParams->color.override)
-+ {
-+ case e_FM_PCD_PLCR_GREEN:
-+ pemode &= ~FM_PCD_PLCR_PEMODE_OVCLR_MASK;
-+ break;
-+ case e_FM_PCD_PLCR_YELLOW:
-+ pemode |= FM_PCD_PLCR_PEMODE_OVCLR_Y;
-+ break;
-+ case e_FM_PCD_PLCR_RED:
-+ pemode |= FM_PCD_PLCR_PEMODE_OVCLR_R;
-+ break;
-+ case e_FM_PCD_PLCR_OVERRIDE:
-+ pemode |= FM_PCD_PLCR_PEMODE_OVCLR_G_NC;
-+ break;
-+ default:
-+ RETURN_ERROR(MAJOR, E_INVALID_SELECTION, NO_MSG);
-+ }
-+ break;
-+ default:
-+ RETURN_ERROR(MAJOR, E_INVALID_SELECTION, NO_MSG);
-+ }
-+ /* Select Measurement Unit Mode to BYTE or PACKET (MODE[PKT]) */
-+ switch (p_ProfileParams->nonPassthroughAlgParams.rateMode)
-+ {
-+ case e_FM_PCD_PLCR_BYTE_MODE :
-+ pemode &= ~FM_PCD_PLCR_PEMODE_PKT;
-+ switch (p_ProfileParams->nonPassthroughAlgParams.byteModeParams.frameLengthSelection)
-+ {
-+ case e_FM_PCD_PLCR_L2_FRM_LEN:
-+ pemode |= FM_PCD_PLCR_PEMODE_FLS_L2;
-+ break;
-+ case e_FM_PCD_PLCR_L3_FRM_LEN:
-+ pemode |= FM_PCD_PLCR_PEMODE_FLS_L3;
-+ break;
-+ case e_FM_PCD_PLCR_L4_FRM_LEN:
-+ pemode |= FM_PCD_PLCR_PEMODE_FLS_L4;
-+ break;
-+ case e_FM_PCD_PLCR_FULL_FRM_LEN:
-+ pemode |= FM_PCD_PLCR_PEMODE_FLS_FULL;
-+ break;
-+ default:
-+ RETURN_ERROR(MAJOR, E_INVALID_SELECTION, NO_MSG);
-+ }
-+ switch (p_ProfileParams->nonPassthroughAlgParams.byteModeParams.rollBackFrameSelection)
-+ {
-+ case e_FM_PCD_PLCR_ROLLBACK_L2_FRM_LEN:
-+ pemode &= ~FM_PCD_PLCR_PEMODE_RBFLS;
-+ break;
-+ case e_FM_PCD_PLCR_ROLLBACK_FULL_FRM_LEN:
-+ pemode |= FM_PCD_PLCR_PEMODE_RBFLS;
-+ break;
-+ default:
-+ RETURN_ERROR(MAJOR, E_INVALID_SELECTION, NO_MSG);
-+ }
-+ break;
-+ case e_FM_PCD_PLCR_PACKET_MODE :
-+ pemode |= FM_PCD_PLCR_PEMODE_PKT;
-+ break;
-+ default:
-+ RETURN_ERROR(MAJOR, E_INVALID_SELECTION, NO_MSG);
-+ }
-+ /* Select timeStamp floating point position (MODE[FPP]) to fit the actual traffic rates. For PACKET
-+ mode with low traffic rates move the fixed point to the left to increase fraction accuracy. For BYTE
-+ mode with high traffic rates move the fixed point to the right to increase integer accuracy. */
-+
-+ /* Configure Traffic Parameters*/
-+ {
-+ uint32_t cir=0, cbs=0, pir_eir=0, pbs_ebs=0, fpp=0;
-+
-+ CalcRates(bitFor1Micro, &p_ProfileParams->nonPassthroughAlgParams, &cir, &cbs, &pir_eir, &pbs_ebs, &fpp);
-+
-+ /* Set Committed Information Rate (CIR) */
-+ p_PlcrRegs->fmpl_pecir = cir;
-+ /* Set Committed Burst Size (CBS). */
-+ p_PlcrRegs->fmpl_pecbs = cbs;
-+ /* Set Peak Information Rate (PIR_EIR used as PIR) */
-+ p_PlcrRegs->fmpl_pepepir_eir = pir_eir;
-+ /* Set Peak Burst Size (PBS_EBS used as PBS) */
-+ p_PlcrRegs->fmpl_pepbs_ebs = pbs_ebs;
-+
-+ /* Initialize the Metering Buckets to be full (write them with 0xFFFFFFFF. */
-+ /* Peak Rate Token Bucket Size (PTS_ETS used as PTS) */
-+ p_PlcrRegs->fmpl_pepts_ets = 0xFFFFFFFF;
-+ /* Committed Rate Token Bucket Size (CTS) */
-+ p_PlcrRegs->fmpl_pects = 0xFFFFFFFF;
-+
-+ /* Set the FPP based on calculation */
-+ pemode |= (fpp << FM_PCD_PLCR_PEMODE_FPP_SHIFT);
-+ }
-+ break; /* FM_PCD_PLCR_PEMODE_ALG_RFC2698 , FM_PCD_PLCR_PEMODE_ALG_RFC4115 */
-+ default:
-+ RETURN_ERROR(MAJOR, E_INVALID_SELECTION, NO_MSG);
-+ }
-+
-+ p_PlcrRegs->fmpl_pemode = pemode;
-+
-+ p_PlcrRegs->fmpl_pegnia = gnia;
-+ p_PlcrRegs->fmpl_peynia = ynia;
-+ p_PlcrRegs->fmpl_pernia = rnia;
-+
-+ /* Zero Counters */
-+ p_PlcrRegs->fmpl_pegpc = 0;
-+ p_PlcrRegs->fmpl_peypc = 0;
-+ p_PlcrRegs->fmpl_perpc = 0;
-+ p_PlcrRegs->fmpl_perypc = 0;
-+ p_PlcrRegs->fmpl_perrpc = 0;
-+
-+ return E_OK;
-+}
-+
-+static t_Error AllocSharedProfiles(t_FmPcd *p_FmPcd, uint16_t numOfProfiles, uint16_t *profilesIds)
-+{
-+ uint32_t profilesFound;
-+ uint16_t i, k=0;
-+ uint32_t intFlags;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd, E_INVALID_HANDLE);
-+
-+ if (!numOfProfiles)
-+ return E_OK;
-+
-+ if (numOfProfiles>FM_PCD_PLCR_NUM_ENTRIES)
-+ RETURN_ERROR(MINOR, E_INVALID_VALUE, ("numProfiles is too big."));
-+
-+ intFlags = PlcrSwLock(p_FmPcd->p_FmPcdPlcr);
-+ /* Find numOfProfiles free profiles (may be spread) */
-+ profilesFound = 0;
-+ for (i=0;i<FM_PCD_PLCR_NUM_ENTRIES; i++)
-+ if (!p_FmPcd->p_FmPcdPlcr->profiles[i].profilesMng.allocated)
-+ {
-+ profilesFound++;
-+ profilesIds[k] = i;
-+ k++;
-+ if (profilesFound == numOfProfiles)
-+ break;
-+ }
-+
-+ if (profilesFound != numOfProfiles)
-+ {
-+ PlcrSwUnlock(p_FmPcd->p_FmPcdPlcr, intFlags);
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE,NO_MSG);
-+ }
-+
-+ for (i = 0;i<k;i++)
-+ {
-+ p_FmPcd->p_FmPcdPlcr->profiles[profilesIds[i]].profilesMng.allocated = TRUE;
-+ p_FmPcd->p_FmPcdPlcr->profiles[profilesIds[i]].profilesMng.ownerId = 0;
-+ }
-+ PlcrSwUnlock(p_FmPcd->p_FmPcdPlcr, intFlags);
-+
-+ return E_OK;
-+}
-+
-+static void FreeSharedProfiles(t_FmPcd *p_FmPcd, uint16_t numOfProfiles, uint16_t *profilesIds)
-+{
-+ uint16_t i;
-+
-+ SANITY_CHECK_RETURN(p_FmPcd, E_INVALID_HANDLE);
-+
-+ ASSERT_COND(numOfProfiles);
-+
-+ for (i=0; i < numOfProfiles; i++)
-+ {
-+ ASSERT_COND(p_FmPcd->p_FmPcdPlcr->profiles[profilesIds[i]].profilesMng.allocated);
-+ p_FmPcd->p_FmPcdPlcr->profiles[profilesIds[i]].profilesMng.allocated = FALSE;
-+ p_FmPcd->p_FmPcdPlcr->profiles[profilesIds[i]].profilesMng.ownerId = p_FmPcd->guestId;
-+ }
-+}
-+
-+static void UpdateRequiredActionFlag(t_Handle h_FmPcd, uint16_t absoluteProfileId, bool set)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+
-+ /* this routine is protected by calling routine */
-+
-+ ASSERT_COND(p_FmPcd->p_FmPcdPlcr->profiles[absoluteProfileId].valid);
-+
-+ if (set)
-+ p_FmPcd->p_FmPcdPlcr->profiles[absoluteProfileId].requiredActionFlag = TRUE;
-+ else
-+ {
-+ p_FmPcd->p_FmPcdPlcr->profiles[absoluteProfileId].requiredAction = 0;
-+ p_FmPcd->p_FmPcdPlcr->profiles[absoluteProfileId].requiredActionFlag = FALSE;
-+ }
-+}
-+
-+/*********************************************/
-+/*............Policer Exception..............*/
-+/*********************************************/
-+static void EventsCB(t_Handle h_FmPcd)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd *)h_FmPcd;
-+ uint32_t event, mask, force;
-+
-+ ASSERT_COND(FmIsMaster(p_FmPcd->h_Fm));
-+ event = GET_UINT32(p_FmPcd->p_FmPcdPlcr->p_FmPcdPlcrRegs->fmpl_evr);
-+ mask = GET_UINT32(p_FmPcd->p_FmPcdPlcr->p_FmPcdPlcrRegs->fmpl_ier);
-+
-+ event &= mask;
-+
-+ /* clear the forced events */
-+ force = GET_UINT32(p_FmPcd->p_FmPcdPlcr->p_FmPcdPlcrRegs->fmpl_ifr);
-+ if (force & event)
-+ WRITE_UINT32(p_FmPcd->p_FmPcdPlcr->p_FmPcdPlcrRegs->fmpl_ifr, force & ~event);
-+
-+
-+ WRITE_UINT32(p_FmPcd->p_FmPcdPlcr->p_FmPcdPlcrRegs->fmpl_evr, event);
-+
-+ if (event & FM_PCD_PLCR_PRAM_SELF_INIT_COMPLETE)
-+ p_FmPcd->f_Exception(p_FmPcd->h_App,e_FM_PCD_PLCR_EXCEPTION_PRAM_SELF_INIT_COMPLETE);
-+ if (event & FM_PCD_PLCR_ATOMIC_ACTION_COMPLETE)
-+ p_FmPcd->f_Exception(p_FmPcd->h_App,e_FM_PCD_PLCR_EXCEPTION_ATOMIC_ACTION_COMPLETE);
-+}
-+
-+/* ..... */
-+
-+static void ErrorExceptionsCB(t_Handle h_FmPcd)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd *)h_FmPcd;
-+ uint32_t event, force, captureReg, mask;
-+
-+ ASSERT_COND(FmIsMaster(p_FmPcd->h_Fm));
-+ event = GET_UINT32(p_FmPcd->p_FmPcdPlcr->p_FmPcdPlcrRegs->fmpl_eevr);
-+ mask = GET_UINT32(p_FmPcd->p_FmPcdPlcr->p_FmPcdPlcrRegs->fmpl_eier);
-+
-+ event &= mask;
-+
-+ /* clear the forced events */
-+ force = GET_UINT32(p_FmPcd->p_FmPcdPlcr->p_FmPcdPlcrRegs->fmpl_eifr);
-+ if (force & event)
-+ WRITE_UINT32(p_FmPcd->p_FmPcdPlcr->p_FmPcdPlcrRegs->fmpl_eifr, force & ~event);
-+
-+ WRITE_UINT32(p_FmPcd->p_FmPcdPlcr->p_FmPcdPlcrRegs->fmpl_eevr, event);
-+
-+ if (event & FM_PCD_PLCR_DOUBLE_ECC)
-+ p_FmPcd->f_Exception(p_FmPcd->h_App,e_FM_PCD_PLCR_EXCEPTION_DOUBLE_ECC);
-+ if (event & FM_PCD_PLCR_INIT_ENTRY_ERROR)
-+ {
-+ captureReg = GET_UINT32(p_FmPcd->p_FmPcdPlcr->p_FmPcdPlcrRegs->fmpl_upcr);
-+ /*ASSERT_COND(captureReg & PLCR_ERR_UNINIT_CAP);
-+ p_UnInitCapt->profileNum = (uint8_t)(captureReg & PLCR_ERR_UNINIT_NUM_MASK);
-+ p_UnInitCapt->portId = (uint8_t)((captureReg & PLCR_ERR_UNINIT_PID_MASK) >>PLCR_ERR_UNINIT_PID_SHIFT) ;
-+ p_UnInitCapt->absolute = (bool)(captureReg & PLCR_ERR_UNINIT_ABSOLUTE_MASK);*/
-+ p_FmPcd->f_FmPcdIndexedException(p_FmPcd->h_App,e_FM_PCD_PLCR_EXCEPTION_INIT_ENTRY_ERROR,(uint16_t)(captureReg & PLCR_ERR_UNINIT_NUM_MASK));
-+ WRITE_UINT32(p_FmPcd->p_FmPcdPlcr->p_FmPcdPlcrRegs->fmpl_upcr, PLCR_ERR_UNINIT_CAP);
-+ }
-+}
-+
-+
-+/*****************************************************************************/
-+/* Inter-module API routines */
-+/*****************************************************************************/
-+
-+t_Handle PlcrConfig(t_FmPcd *p_FmPcd, t_FmPcdParams *p_FmPcdParams)
-+{
-+ t_FmPcdPlcr *p_FmPcdPlcr;
-+ uint16_t i=0;
-+
-+ UNUSED(p_FmPcd);
-+ UNUSED(p_FmPcdParams);
-+
-+ p_FmPcdPlcr = (t_FmPcdPlcr *) XX_Malloc(sizeof(t_FmPcdPlcr));
-+ if (!p_FmPcdPlcr)
-+ {
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("FM Policer structure allocation FAILED"));
-+ return NULL;
-+ }
-+ memset(p_FmPcdPlcr, 0, sizeof(t_FmPcdPlcr));
-+ if (p_FmPcd->guestId == NCSW_MASTER_ID)
-+ {
-+ p_FmPcdPlcr->p_FmPcdPlcrRegs = (t_FmPcdPlcrRegs *)UINT_TO_PTR(FmGetPcdPlcrBaseAddr(p_FmPcdParams->h_Fm));
-+ p_FmPcd->p_FmPcdDriverParam->plcrAutoRefresh = DEFAULT_plcrAutoRefresh;
-+ p_FmPcd->exceptions |= (DEFAULT_fmPcdPlcrExceptions | DEFAULT_fmPcdPlcrErrorExceptions);
-+ }
-+
-+ p_FmPcdPlcr->numOfSharedProfiles = DEFAULT_numOfSharedPlcrProfiles;
-+
-+ p_FmPcdPlcr->partPlcrProfilesBase = p_FmPcdParams->partPlcrProfilesBase;
-+ p_FmPcdPlcr->partNumOfPlcrProfiles = p_FmPcdParams->partNumOfPlcrProfiles;
-+ /* for backward compatabilty. if no policer profile, will set automatically to the max */
-+ if ((p_FmPcd->guestId == NCSW_MASTER_ID) &&
-+ (p_FmPcdPlcr->partNumOfPlcrProfiles == 0))
-+ p_FmPcdPlcr->partNumOfPlcrProfiles = FM_PCD_PLCR_NUM_ENTRIES;
-+
-+ for (i=0; i<FM_PCD_PLCR_NUM_ENTRIES; i++)
-+ p_FmPcdPlcr->profiles[i].profilesMng.ownerId = (uint8_t)ILLEGAL_BASE;
-+
-+ return p_FmPcdPlcr;
-+}
-+
-+t_Error PlcrInit(t_FmPcd *p_FmPcd)
-+{
-+ t_FmPcdDriverParam *p_Param = p_FmPcd->p_FmPcdDriverParam;
-+ t_FmPcdPlcr *p_FmPcdPlcr = p_FmPcd->p_FmPcdPlcr;
-+ t_FmPcdPlcrRegs *p_Regs = p_FmPcd->p_FmPcdPlcr->p_FmPcdPlcrRegs;
-+ t_Error err = E_OK;
-+ uint32_t tmpReg32 = 0;
-+ uint16_t base;
-+
-+ if ((p_FmPcdPlcr->partPlcrProfilesBase + p_FmPcdPlcr->partNumOfPlcrProfiles) > FM_PCD_PLCR_NUM_ENTRIES)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("partPlcrProfilesBase+partNumOfPlcrProfiles out of range!!!"));
-+
-+ p_FmPcdPlcr->h_HwSpinlock = XX_InitSpinlock();
-+ if (!p_FmPcdPlcr->h_HwSpinlock)
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, ("FM Policer HW spinlock"));
-+
-+ p_FmPcdPlcr->h_SwSpinlock = XX_InitSpinlock();
-+ if (!p_FmPcdPlcr->h_SwSpinlock)
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, ("FM Policer SW spinlock"));
-+
-+ base = PlcrAllocProfilesForPartition(p_FmPcd,
-+ p_FmPcdPlcr->partPlcrProfilesBase,
-+ p_FmPcdPlcr->partNumOfPlcrProfiles,
-+ p_FmPcd->guestId);
-+ if (base == (uint16_t)ILLEGAL_BASE)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, NO_MSG);
-+
-+ if (p_FmPcdPlcr->numOfSharedProfiles)
-+ {
-+ err = AllocSharedProfiles(p_FmPcd,
-+ p_FmPcdPlcr->numOfSharedProfiles,
-+ p_FmPcdPlcr->sharedProfilesIds);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err,NO_MSG);
-+ }
-+
-+ if (p_FmPcd->guestId != NCSW_MASTER_ID)
-+ return E_OK;
-+
-+ /**********************FMPL_GCR******************/
-+ tmpReg32 = 0;
-+ tmpReg32 |= FM_PCD_PLCR_GCR_STEN;
-+ if (p_Param->plcrAutoRefresh)
-+ tmpReg32 |= FM_PCD_PLCR_GCR_DAR;
-+ tmpReg32 |= GET_NIA_BMI_AC_ENQ_FRAME(p_FmPcd);
-+
-+ WRITE_UINT32(p_Regs->fmpl_gcr, tmpReg32);
-+ /**********************FMPL_GCR******************/
-+
-+ /**********************FMPL_EEVR******************/
-+ WRITE_UINT32(p_Regs->fmpl_eevr, (FM_PCD_PLCR_DOUBLE_ECC | FM_PCD_PLCR_INIT_ENTRY_ERROR));
-+ /**********************FMPL_EEVR******************/
-+ /**********************FMPL_EIER******************/
-+ tmpReg32 = 0;
-+ if (p_FmPcd->exceptions & FM_PCD_EX_PLCR_DOUBLE_ECC)
-+ {
-+ FmEnableRamsEcc(p_FmPcd->h_Fm);
-+ tmpReg32 |= FM_PCD_PLCR_DOUBLE_ECC;
-+ }
-+ if (p_FmPcd->exceptions & FM_PCD_EX_PLCR_INIT_ENTRY_ERROR)
-+ tmpReg32 |= FM_PCD_PLCR_INIT_ENTRY_ERROR;
-+ WRITE_UINT32(p_Regs->fmpl_eier, tmpReg32);
-+ /**********************FMPL_EIER******************/
-+
-+ /**********************FMPL_EVR******************/
-+ WRITE_UINT32(p_Regs->fmpl_evr, (FM_PCD_PLCR_PRAM_SELF_INIT_COMPLETE | FM_PCD_PLCR_ATOMIC_ACTION_COMPLETE));
-+ /**********************FMPL_EVR******************/
-+ /**********************FMPL_IER******************/
-+ tmpReg32 = 0;
-+ if (p_FmPcd->exceptions & FM_PCD_EX_PLCR_PRAM_SELF_INIT_COMPLETE)
-+ tmpReg32 |= FM_PCD_PLCR_PRAM_SELF_INIT_COMPLETE;
-+ if (p_FmPcd->exceptions & FM_PCD_EX_PLCR_ATOMIC_ACTION_COMPLETE)
-+ tmpReg32 |= FM_PCD_PLCR_ATOMIC_ACTION_COMPLETE;
-+ WRITE_UINT32(p_Regs->fmpl_ier, tmpReg32);
-+ /**********************FMPL_IER******************/
-+
-+ /* register even if no interrupts enabled, to allow future enablement */
-+ FmRegisterIntr(p_FmPcd->h_Fm,
-+ e_FM_MOD_PLCR,
-+ 0,
-+ e_FM_INTR_TYPE_ERR,
-+ ErrorExceptionsCB,
-+ p_FmPcd);
-+ FmRegisterIntr(p_FmPcd->h_Fm,
-+ e_FM_MOD_PLCR,
-+ 0,
-+ e_FM_INTR_TYPE_NORMAL,
-+ EventsCB,
-+ p_FmPcd);
-+
-+ /* driver initializes one DFLT profile at the last entry*/
-+ /**********************FMPL_DPMR******************/
-+ tmpReg32 = 0;
-+ WRITE_UINT32(p_Regs->fmpl_dpmr, tmpReg32);
-+ p_FmPcd->p_FmPcdPlcr->profiles[0].profilesMng.allocated = TRUE;
-+
-+ return E_OK;
-+}
-+
-+t_Error PlcrFree(t_FmPcd *p_FmPcd)
-+{
-+ FmUnregisterIntr(p_FmPcd->h_Fm, e_FM_MOD_PLCR, 0, e_FM_INTR_TYPE_ERR);
-+ FmUnregisterIntr(p_FmPcd->h_Fm, e_FM_MOD_PLCR, 0, e_FM_INTR_TYPE_NORMAL);
-+
-+ if (p_FmPcd->p_FmPcdPlcr->numOfSharedProfiles)
-+ FreeSharedProfiles(p_FmPcd,
-+ p_FmPcd->p_FmPcdPlcr->numOfSharedProfiles,
-+ p_FmPcd->p_FmPcdPlcr->sharedProfilesIds);
-+
-+ if (p_FmPcd->p_FmPcdPlcr->partNumOfPlcrProfiles)
-+ PlcrFreeProfilesForPartition(p_FmPcd,
-+ p_FmPcd->p_FmPcdPlcr->partPlcrProfilesBase,
-+ p_FmPcd->p_FmPcdPlcr->partNumOfPlcrProfiles,
-+ p_FmPcd->guestId);
-+
-+ if (p_FmPcd->p_FmPcdPlcr->h_SwSpinlock)
-+ XX_FreeSpinlock(p_FmPcd->p_FmPcdPlcr->h_SwSpinlock);
-+
-+ if (p_FmPcd->p_FmPcdPlcr->h_HwSpinlock)
-+ XX_FreeSpinlock(p_FmPcd->p_FmPcdPlcr->h_HwSpinlock);
-+
-+ return E_OK;
-+}
-+
-+void PlcrEnable(t_FmPcd *p_FmPcd)
-+{
-+ t_FmPcdPlcrRegs *p_Regs = p_FmPcd->p_FmPcdPlcr->p_FmPcdPlcrRegs;
-+
-+ WRITE_UINT32(p_Regs->fmpl_gcr, GET_UINT32(p_Regs->fmpl_gcr) | FM_PCD_PLCR_GCR_EN);
-+}
-+
-+void PlcrDisable(t_FmPcd *p_FmPcd)
-+{
-+ t_FmPcdPlcrRegs *p_Regs = p_FmPcd->p_FmPcdPlcr->p_FmPcdPlcrRegs;
-+
-+ WRITE_UINT32(p_Regs->fmpl_gcr, GET_UINT32(p_Regs->fmpl_gcr) & ~FM_PCD_PLCR_GCR_EN);
-+}
-+
-+uint16_t PlcrAllocProfilesForPartition(t_FmPcd *p_FmPcd, uint16_t base, uint16_t numOfProfiles, uint8_t guestId)
-+{
-+ uint32_t intFlags;
-+ uint16_t profilesFound = 0;
-+ int i = 0;
-+
-+ ASSERT_COND(p_FmPcd);
-+ ASSERT_COND(p_FmPcd->p_FmPcdPlcr);
-+
-+ if (!numOfProfiles)
-+ return 0;
-+
-+ if ((numOfProfiles > FM_PCD_PLCR_NUM_ENTRIES) ||
-+ (base + numOfProfiles > FM_PCD_PLCR_NUM_ENTRIES))
-+ return (uint16_t)ILLEGAL_BASE;
-+
-+ if (p_FmPcd->h_IpcSession)
-+ {
-+ t_FmIpcResourceAllocParams ipcAllocParams;
-+ t_FmPcdIpcMsg msg;
-+ t_FmPcdIpcReply reply;
-+ t_Error err;
-+ uint32_t replyLength;
-+
-+ memset(&msg, 0, sizeof(msg));
-+ memset(&reply, 0, sizeof(reply));
-+ memset(&ipcAllocParams, 0, sizeof(t_FmIpcResourceAllocParams));
-+ ipcAllocParams.guestId = p_FmPcd->guestId;
-+ ipcAllocParams.num = p_FmPcd->p_FmPcdPlcr->partNumOfPlcrProfiles;
-+ ipcAllocParams.base = p_FmPcd->p_FmPcdPlcr->partPlcrProfilesBase;
-+ msg.msgId = FM_PCD_ALLOC_PROFILES;
-+ memcpy(msg.msgBody, &ipcAllocParams, sizeof(t_FmIpcResourceAllocParams));
-+ replyLength = sizeof(uint32_t) + sizeof(uint16_t);
-+ err = XX_IpcSendMessage(p_FmPcd->h_IpcSession,
-+ (uint8_t*)&msg,
-+ sizeof(msg.msgId) + sizeof(t_FmIpcResourceAllocParams),
-+ (uint8_t*)&reply,
-+ &replyLength,
-+ NULL,
-+ NULL);
-+ if ((err != E_OK) ||
-+ (replyLength != (sizeof(uint32_t) + sizeof(uint16_t))))
-+ {
-+ REPORT_ERROR(MAJOR, err, NO_MSG);
-+ return (uint16_t)ILLEGAL_BASE;
-+ }
-+ else
-+ memcpy((uint8_t*)&p_FmPcd->p_FmPcdPlcr->partPlcrProfilesBase, reply.replyBody, sizeof(uint16_t));
-+ if (p_FmPcd->p_FmPcdPlcr->partPlcrProfilesBase == (uint16_t)ILLEGAL_BASE)
-+ {
-+ REPORT_ERROR(MAJOR, err, NO_MSG);
-+ return (uint16_t)ILLEGAL_BASE;
-+ }
-+ }
-+ else if (p_FmPcd->guestId != NCSW_MASTER_ID)
-+ {
-+ DBG(WARNING, ("FM Guest mode, without IPC - can't validate Policer-profiles range!"));
-+ return (uint16_t)ILLEGAL_BASE;
-+ }
-+
-+ intFlags = XX_LockIntrSpinlock(p_FmPcd->h_Spinlock);
-+ for (i=base; i<(base+numOfProfiles); i++)
-+ if (p_FmPcd->p_FmPcdPlcr->profiles[i].profilesMng.ownerId == (uint8_t)ILLEGAL_BASE)
-+ profilesFound++;
-+ else
-+ break;
-+
-+ if (profilesFound == numOfProfiles)
-+ for (i=base; i<(base+numOfProfiles); i++)
-+ p_FmPcd->p_FmPcdPlcr->profiles[i].profilesMng.ownerId = guestId;
-+ else
-+ {
-+ XX_UnlockIntrSpinlock(p_FmPcd->h_Spinlock, intFlags);
-+ return (uint16_t)ILLEGAL_BASE;
-+ }
-+ XX_UnlockIntrSpinlock(p_FmPcd->h_Spinlock, intFlags);
-+
-+ return base;
-+}
-+
-+void PlcrFreeProfilesForPartition(t_FmPcd *p_FmPcd, uint16_t base, uint16_t numOfProfiles, uint8_t guestId)
-+{
-+ int i = 0;
-+
-+ ASSERT_COND(p_FmPcd);
-+ ASSERT_COND(p_FmPcd->p_FmPcdPlcr);
-+
-+ if (p_FmPcd->h_IpcSession)
-+ {
-+ t_FmIpcResourceAllocParams ipcAllocParams;
-+ t_FmPcdIpcMsg msg;
-+ t_Error err;
-+
-+ memset(&msg, 0, sizeof(msg));
-+ memset(&ipcAllocParams, 0, sizeof(t_FmIpcResourceAllocParams));
-+ ipcAllocParams.guestId = p_FmPcd->guestId;
-+ ipcAllocParams.num = p_FmPcd->p_FmPcdPlcr->partNumOfPlcrProfiles;
-+ ipcAllocParams.base = p_FmPcd->p_FmPcdPlcr->partPlcrProfilesBase;
-+ msg.msgId = FM_PCD_FREE_PROFILES;
-+ memcpy(msg.msgBody, &ipcAllocParams, sizeof(t_FmIpcResourceAllocParams));
-+ err = XX_IpcSendMessage(p_FmPcd->h_IpcSession,
-+ (uint8_t*)&msg,
-+ sizeof(msg.msgId) + sizeof(t_FmIpcResourceAllocParams),
-+ NULL,
-+ NULL,
-+ NULL,
-+ NULL);
-+ if (err != E_OK)
-+ REPORT_ERROR(MAJOR, err, NO_MSG);
-+ return;
-+ }
-+ else if (p_FmPcd->guestId != NCSW_MASTER_ID)
-+ {
-+ DBG(WARNING, ("FM Guest mode, without IPC - can't validate Policer-profiles range!"));
-+ return;
-+ }
-+
-+ for (i=base; i<(base+numOfProfiles); i++)
-+ {
-+ if (p_FmPcd->p_FmPcdPlcr->profiles[i].profilesMng.ownerId == guestId)
-+ p_FmPcd->p_FmPcdPlcr->profiles[i].profilesMng.ownerId = (uint8_t)ILLEGAL_BASE;
-+ else
-+ DBG(WARNING, ("Request for freeing storage profile window which wasn't allocated to this partition"));
-+ }
-+}
-+
-+t_Error PlcrSetPortProfiles(t_FmPcd *p_FmPcd,
-+ uint8_t hardwarePortId,
-+ uint16_t numOfProfiles,
-+ uint16_t base)
-+{
-+ t_FmPcdPlcrRegs *p_Regs = p_FmPcd->p_FmPcdPlcr->p_FmPcdPlcrRegs;
-+ uint32_t log2Num, tmpReg32;
-+
-+ if ((p_FmPcd->guestId != NCSW_MASTER_ID) &&
-+ !p_Regs &&
-+ p_FmPcd->h_IpcSession)
-+ {
-+ t_FmIpcResourceAllocParams ipcAllocParams;
-+ t_FmPcdIpcMsg msg;
-+ t_Error err;
-+
-+ memset(&msg, 0, sizeof(msg));
-+ memset(&ipcAllocParams, 0, sizeof(t_FmIpcResourceAllocParams));
-+ ipcAllocParams.guestId = hardwarePortId;
-+ ipcAllocParams.num = numOfProfiles;
-+ ipcAllocParams.base = base;
-+ msg.msgId = FM_PCD_SET_PORT_PROFILES;
-+ memcpy(msg.msgBody, &ipcAllocParams, sizeof(t_FmIpcResourceAllocParams));
-+ err = XX_IpcSendMessage(p_FmPcd->h_IpcSession,
-+ (uint8_t*)&msg,
-+ sizeof(msg.msgId) + sizeof(t_FmIpcResourceAllocParams),
-+ NULL,
-+ NULL,
-+ NULL,
-+ NULL);
-+ if (err != E_OK)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ return E_OK;
-+ }
-+ else if (!p_Regs)
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED,
-+ ("Either IPC or 'baseAddress' is required!"));
-+
-+ ASSERT_COND(IN_RANGE(1, hardwarePortId, 63));
-+
-+ if (GET_UINT32(p_Regs->fmpl_pmr[hardwarePortId-1]) & FM_PCD_PLCR_PMR_V)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE,
-+ ("The requesting port has already an allocated profiles window."));
-+
-+ /**********************FMPL_PMRx******************/
-+ LOG2((uint64_t)numOfProfiles, log2Num);
-+ tmpReg32 = base;
-+ tmpReg32 |= log2Num << 16;
-+ tmpReg32 |= FM_PCD_PLCR_PMR_V;
-+ WRITE_UINT32(p_Regs->fmpl_pmr[hardwarePortId-1], tmpReg32);
-+
-+ return E_OK;
-+}
-+
-+t_Error PlcrClearPortProfiles(t_FmPcd *p_FmPcd, uint8_t hardwarePortId)
-+{
-+ t_FmPcdPlcrRegs *p_Regs = p_FmPcd->p_FmPcdPlcr->p_FmPcdPlcrRegs;
-+
-+ if ((p_FmPcd->guestId != NCSW_MASTER_ID) &&
-+ !p_Regs &&
-+ p_FmPcd->h_IpcSession)
-+ {
-+ t_FmIpcResourceAllocParams ipcAllocParams;
-+ t_FmPcdIpcMsg msg;
-+ t_Error err;
-+
-+ memset(&msg, 0, sizeof(msg));
-+ memset(&ipcAllocParams, 0, sizeof(t_FmIpcResourceAllocParams));
-+ ipcAllocParams.guestId = hardwarePortId;
-+ msg.msgId = FM_PCD_CLEAR_PORT_PROFILES;
-+ memcpy(msg.msgBody, &ipcAllocParams, sizeof(t_FmIpcResourceAllocParams));
-+ err = XX_IpcSendMessage(p_FmPcd->h_IpcSession,
-+ (uint8_t*)&msg,
-+ sizeof(msg.msgId) + sizeof(t_FmIpcResourceAllocParams),
-+ NULL,
-+ NULL,
-+ NULL,
-+ NULL);
-+ if (err != E_OK)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ return E_OK;
-+ }
-+ else if (!p_Regs)
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED,
-+ ("Either IPC or 'baseAddress' is required!"));
-+
-+ ASSERT_COND(IN_RANGE(1, hardwarePortId, 63));
-+ WRITE_UINT32(p_Regs->fmpl_pmr[hardwarePortId-1], 0);
-+
-+ return E_OK;
-+}
-+
-+t_Error FmPcdPlcrAllocProfiles(t_Handle h_FmPcd, uint8_t hardwarePortId, uint16_t numOfProfiles)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+ t_Error err = E_OK;
-+ uint32_t profilesFound;
-+ uint32_t intFlags;
-+ uint16_t i, first, swPortIndex = 0;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd, E_INVALID_HANDLE);
-+
-+ if (!numOfProfiles)
-+ return E_OK;
-+
-+ ASSERT_COND(hardwarePortId);
-+
-+ if (numOfProfiles>FM_PCD_PLCR_NUM_ENTRIES)
-+ RETURN_ERROR(MINOR, E_INVALID_VALUE, ("numProfiles is too big."));
-+
-+ if (!POWER_OF_2(numOfProfiles))
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("numProfiles must be a power of 2."));
-+
-+ first = 0;
-+ profilesFound = 0;
-+ intFlags = PlcrSwLock(p_FmPcd->p_FmPcdPlcr);
-+
-+ for (i=0; i<FM_PCD_PLCR_NUM_ENTRIES; )
-+ {
-+ if (!p_FmPcd->p_FmPcdPlcr->profiles[i].profilesMng.allocated)
-+ {
-+ profilesFound++;
-+ i++;
-+ if (profilesFound == numOfProfiles)
-+ break;
-+ }
-+ else
-+ {
-+ profilesFound = 0;
-+ /* advance i to the next aligned address */
-+ i = first = (uint16_t)(first + numOfProfiles);
-+ }
-+ }
-+
-+ if (profilesFound == numOfProfiles)
-+ {
-+ for (i=first; i<first + numOfProfiles; i++)
-+ {
-+ p_FmPcd->p_FmPcdPlcr->profiles[i].profilesMng.allocated = TRUE;
-+ p_FmPcd->p_FmPcdPlcr->profiles[i].profilesMng.ownerId = hardwarePortId;
-+ }
-+ }
-+ else
-+ {
-+ PlcrSwUnlock(p_FmPcd->p_FmPcdPlcr, intFlags);
-+ RETURN_ERROR(MINOR, E_FULL, ("No profiles."));
-+ }
-+ PlcrSwUnlock(p_FmPcd->p_FmPcdPlcr, intFlags);
-+
-+ err = PlcrSetPortProfiles(p_FmPcd, hardwarePortId, numOfProfiles, first);
-+ if (err)
-+ {
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+
-+ HW_PORT_ID_TO_SW_PORT_INDX(swPortIndex, hardwarePortId);
-+
-+ p_FmPcd->p_FmPcdPlcr->portsMapping[swPortIndex].numOfProfiles = numOfProfiles;
-+ p_FmPcd->p_FmPcdPlcr->portsMapping[swPortIndex].profilesBase = first;
-+
-+ return E_OK;
-+}
-+
-+t_Error FmPcdPlcrFreeProfiles(t_Handle h_FmPcd, uint8_t hardwarePortId)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+ t_Error err = E_OK;
-+ uint32_t intFlags;
-+ uint16_t i, swPortIndex = 0;
-+
-+ ASSERT_COND(IN_RANGE(1, hardwarePortId, 63));
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmPcd->p_FmPcdDriverParam, E_INVALID_HANDLE);
-+
-+ HW_PORT_ID_TO_SW_PORT_INDX(swPortIndex, hardwarePortId);
-+
-+ err = PlcrClearPortProfiles(p_FmPcd, hardwarePortId);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err,NO_MSG);
-+
-+ intFlags = PlcrSwLock(p_FmPcd->p_FmPcdPlcr);
-+ for (i=p_FmPcd->p_FmPcdPlcr->portsMapping[swPortIndex].profilesBase;
-+ i<(p_FmPcd->p_FmPcdPlcr->portsMapping[swPortIndex].profilesBase +
-+ p_FmPcd->p_FmPcdPlcr->portsMapping[swPortIndex].numOfProfiles);
-+ i++)
-+ {
-+ ASSERT_COND(p_FmPcd->p_FmPcdPlcr->profiles[i].profilesMng.ownerId == hardwarePortId);
-+ ASSERT_COND(p_FmPcd->p_FmPcdPlcr->profiles[i].profilesMng.allocated);
-+
-+ p_FmPcd->p_FmPcdPlcr->profiles[i].profilesMng.allocated = FALSE;
-+ p_FmPcd->p_FmPcdPlcr->profiles[i].profilesMng.ownerId = p_FmPcd->guestId;
-+ }
-+ PlcrSwUnlock(p_FmPcd->p_FmPcdPlcr, intFlags);
-+
-+ p_FmPcd->p_FmPcdPlcr->portsMapping[swPortIndex].numOfProfiles = 0;
-+ p_FmPcd->p_FmPcdPlcr->portsMapping[swPortIndex].profilesBase = 0;
-+
-+ return E_OK;
-+}
-+
-+t_Error FmPcdPlcrCcGetSetParams(t_Handle h_FmPcd, uint16_t profileIndx ,uint32_t requiredAction)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd *)h_FmPcd;
-+ t_FmPcdPlcr *p_FmPcdPlcr = p_FmPcd->p_FmPcdPlcr;
-+ t_FmPcdPlcrRegs *p_FmPcdPlcrRegs = p_FmPcdPlcr->p_FmPcdPlcrRegs;
-+ uint32_t tmpReg32, intFlags;
-+ t_Error err;
-+
-+ /* Calling function locked all PCD modules, so no need to lock here */
-+
-+ if (profileIndx >= FM_PCD_PLCR_NUM_ENTRIES)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE,("Policer profile out of range"));
-+
-+ if (!FmPcdPlcrIsProfileValid(p_FmPcd, profileIndx))
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE,("Policer profile is not valid"));
-+
-+ /*intFlags = PlcrProfileLock(&p_FmPcd->p_FmPcdPlcr->profiles[profileIndx]);*/
-+
-+ if (p_FmPcd->h_Hc)
-+ {
-+ err = FmHcPcdPlcrCcGetSetParams(p_FmPcd->h_Hc, profileIndx, requiredAction);
-+
-+ UpdateRequiredActionFlag(p_FmPcd, profileIndx, TRUE);
-+ FmPcdPlcrUpdateRequiredAction(p_FmPcd, profileIndx, requiredAction);
-+
-+ /*PlcrProfileUnlock(&p_FmPcd->p_FmPcdPlcr->profiles[profileIndx], intFlags);*/
-+ return err;
-+ }
-+
-+ /* lock the HW because once we read the registers we don't want them to be changed
-+ * by another access. (We can copy to a tmp location and release the lock!) */
-+
-+ intFlags = PlcrHwLock(p_FmPcdPlcr);
-+ WritePar(p_FmPcd, FmPcdPlcrBuildReadPlcrActionReg(profileIndx));
-+
-+ if (!p_FmPcd->p_FmPcdPlcr->profiles[profileIndx].requiredActionFlag ||
-+ !(p_FmPcd->p_FmPcdPlcr->profiles[profileIndx].requiredAction & requiredAction))
-+ {
-+ if (requiredAction & UPDATE_NIA_ENQ_WITHOUT_DMA)
-+ {
-+ if ((p_FmPcd->p_FmPcdPlcr->profiles[profileIndx].nextEngineOnGreen!= e_FM_PCD_DONE) ||
-+ (p_FmPcd->p_FmPcdPlcr->profiles[profileIndx].nextEngineOnYellow!= e_FM_PCD_DONE) ||
-+ (p_FmPcd->p_FmPcdPlcr->profiles[profileIndx].nextEngineOnRed!= e_FM_PCD_DONE))
-+ {
-+ PlcrHwUnlock(p_FmPcdPlcr, intFlags);
-+ /*PlcrProfileUnlock(&p_FmPcd->p_FmPcdPlcr->profiles[profileIndx], intFlags);*/
-+ RETURN_ERROR (MAJOR, E_OK, ("In this case the next engine can be e_FM_PCD_DONE"));
-+ }
-+
-+ if (p_FmPcd->p_FmPcdPlcr->profiles[profileIndx].paramsOnGreen.action == e_FM_PCD_ENQ_FRAME)
-+ {
-+ tmpReg32 = GET_UINT32(p_FmPcdPlcrRegs->profileRegs.fmpl_pegnia);
-+ if (!(tmpReg32 & (NIA_ENG_BMI | NIA_BMI_AC_ENQ_FRAME)))
-+ {
-+ PlcrHwUnlock(p_FmPcdPlcr, intFlags);
-+ /*PlcrProfileUnlock(&p_FmPcd->p_FmPcdPlcr->profiles[profileIndx], intFlags);*/
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("Next engine of this policer profile has to be assigned to FM_PCD_DONE"));
-+ }
-+ tmpReg32 |= NIA_BMI_AC_ENQ_FRAME_WITHOUT_DMA;
-+ WRITE_UINT32(p_FmPcdPlcrRegs->profileRegs.fmpl_pegnia, tmpReg32);
-+ tmpReg32 = FmPcdPlcrBuildWritePlcrActionReg(profileIndx);
-+ tmpReg32 |= FM_PCD_PLCR_PAR_PWSEL_PEGNIA;
-+ WritePar(p_FmPcd, tmpReg32);
-+ }
-+
-+ if (p_FmPcd->p_FmPcdPlcr->profiles[profileIndx].paramsOnYellow.action == e_FM_PCD_ENQ_FRAME)
-+ {
-+ tmpReg32 = GET_UINT32(p_FmPcdPlcrRegs->profileRegs.fmpl_peynia);
-+ if (!(tmpReg32 & (NIA_ENG_BMI | NIA_BMI_AC_ENQ_FRAME)))
-+ {
-+ PlcrHwUnlock(p_FmPcdPlcr, intFlags);
-+ /*PlcrProfileUnlock(&p_FmPcd->p_FmPcdPlcr->profiles[profileIndx], intFlags);*/
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("Next engine of this policer profile has to be assigned to FM_PCD_DONE"));
-+ }
-+ tmpReg32 |= NIA_BMI_AC_ENQ_FRAME_WITHOUT_DMA;
-+ WRITE_UINT32(p_FmPcdPlcrRegs->profileRegs.fmpl_peynia, tmpReg32);
-+ tmpReg32 = FmPcdPlcrBuildWritePlcrActionReg(profileIndx);
-+ tmpReg32 |= FM_PCD_PLCR_PAR_PWSEL_PEYNIA;
-+ WritePar(p_FmPcd, tmpReg32);
-+ PlcrHwUnlock(p_FmPcdPlcr, intFlags);
-+ }
-+
-+ if (p_FmPcd->p_FmPcdPlcr->profiles[profileIndx].paramsOnRed.action == e_FM_PCD_ENQ_FRAME)
-+ {
-+ tmpReg32 = GET_UINT32(p_FmPcdPlcrRegs->profileRegs.fmpl_pernia);
-+ if (!(tmpReg32 & (NIA_ENG_BMI | NIA_BMI_AC_ENQ_FRAME)))
-+ {
-+ PlcrHwUnlock(p_FmPcdPlcr, intFlags);
-+ /*PlcrProfileUnlock(&p_FmPcd->p_FmPcdPlcr->profiles[profileIndx], intFlags);*/
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("Next engine of this policer profile has to be assigned to FM_PCD_DONE"));
-+ }
-+ tmpReg32 |= NIA_BMI_AC_ENQ_FRAME_WITHOUT_DMA;
-+ WRITE_UINT32(p_FmPcdPlcrRegs->profileRegs.fmpl_pernia, tmpReg32);
-+ tmpReg32 = FmPcdPlcrBuildWritePlcrActionReg(profileIndx);
-+ tmpReg32 |= FM_PCD_PLCR_PAR_PWSEL_PERNIA;
-+ WritePar(p_FmPcd, tmpReg32);
-+
-+ }
-+ }
-+ }
-+ PlcrHwUnlock(p_FmPcdPlcr, intFlags);
-+
-+ UpdateRequiredActionFlag(p_FmPcd, profileIndx, TRUE);
-+ FmPcdPlcrUpdateRequiredAction(p_FmPcd, profileIndx, requiredAction);
-+
-+ /*PlcrProfileUnlock(&p_FmPcd->p_FmPcdPlcr->profiles[profileIndx], intFlags);*/
-+
-+ return E_OK;
-+}
-+
-+uint32_t FmPcdPlcrGetRequiredActionFlag(t_Handle h_FmPcd, uint16_t absoluteProfileId)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+
-+ ASSERT_COND(p_FmPcd->p_FmPcdPlcr->profiles[absoluteProfileId].valid);
-+
-+ return p_FmPcd->p_FmPcdPlcr->profiles[absoluteProfileId].requiredActionFlag;
-+}
-+
-+uint32_t FmPcdPlcrGetRequiredAction(t_Handle h_FmPcd, uint16_t absoluteProfileId)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+
-+ ASSERT_COND(p_FmPcd->p_FmPcdPlcr->profiles[absoluteProfileId].valid);
-+
-+ return p_FmPcd->p_FmPcdPlcr->profiles[absoluteProfileId].requiredAction;
-+}
-+
-+bool FmPcdPlcrIsProfileValid(t_Handle h_FmPcd, uint16_t absoluteProfileId)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+ t_FmPcdPlcr *p_FmPcdPlcr = p_FmPcd->p_FmPcdPlcr;
-+
-+ ASSERT_COND(absoluteProfileId < FM_PCD_PLCR_NUM_ENTRIES);
-+
-+ return p_FmPcdPlcr->profiles[absoluteProfileId].valid;
-+}
-+
-+void FmPcdPlcrValidateProfileSw(t_Handle h_FmPcd, uint16_t absoluteProfileId)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+ uint32_t intFlags;
-+
-+ ASSERT_COND(!p_FmPcd->p_FmPcdPlcr->profiles[absoluteProfileId].valid);
-+
-+ intFlags = PlcrProfileLock(&p_FmPcd->p_FmPcdPlcr->profiles[absoluteProfileId]);
-+ p_FmPcd->p_FmPcdPlcr->profiles[absoluteProfileId].valid = TRUE;
-+ PlcrProfileUnlock(&p_FmPcd->p_FmPcdPlcr->profiles[absoluteProfileId], intFlags);
-+}
-+
-+void FmPcdPlcrInvalidateProfileSw(t_Handle h_FmPcd, uint16_t absoluteProfileId)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+ uint32_t intFlags;
-+
-+ ASSERT_COND(p_FmPcd->p_FmPcdPlcr->profiles[absoluteProfileId].valid);
-+
-+ intFlags = PlcrProfileLock(&p_FmPcd->p_FmPcdPlcr->profiles[absoluteProfileId]);
-+ p_FmPcd->p_FmPcdPlcr->profiles[absoluteProfileId].valid = FALSE;
-+ PlcrProfileUnlock(&p_FmPcd->p_FmPcdPlcr->profiles[absoluteProfileId], intFlags);
-+}
-+
-+uint16_t FmPcdPlcrProfileGetAbsoluteId(t_Handle h_Profile)
-+{
-+ return ((t_FmPcdPlcrProfile*)h_Profile)->absoluteProfileId;
-+}
-+
-+t_Error FmPcdPlcrGetAbsoluteIdByProfileParams(t_Handle h_FmPcd,
-+ e_FmPcdProfileTypeSelection profileType,
-+ t_Handle h_FmPort,
-+ uint16_t relativeProfile,
-+ uint16_t *p_AbsoluteId)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+ t_FmPcdPlcr *p_FmPcdPlcr = p_FmPcd->p_FmPcdPlcr;
-+ uint8_t i;
-+
-+ switch (profileType)
-+ {
-+ case e_FM_PCD_PLCR_PORT_PRIVATE:
-+ /* get port PCD id from port handle */
-+ for (i=0;i<FM_MAX_NUM_OF_PORTS;i++)
-+ if (p_FmPcd->p_FmPcdPlcr->portsMapping[i].h_FmPort == h_FmPort)
-+ break;
-+ if (i == FM_MAX_NUM_OF_PORTS)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE , ("Invalid port handle."));
-+
-+ if (!p_FmPcd->p_FmPcdPlcr->portsMapping[i].numOfProfiles)
-+ RETURN_ERROR(MAJOR, E_INVALID_SELECTION , ("Port has no allocated profiles"));
-+ if (relativeProfile >= p_FmPcd->p_FmPcdPlcr->portsMapping[i].numOfProfiles)
-+ RETURN_ERROR(MAJOR, E_INVALID_SELECTION , ("Profile id is out of range"));
-+ *p_AbsoluteId = (uint16_t)(p_FmPcd->p_FmPcdPlcr->portsMapping[i].profilesBase + relativeProfile);
-+ break;
-+ case e_FM_PCD_PLCR_SHARED:
-+ if (relativeProfile >= p_FmPcdPlcr->numOfSharedProfiles)
-+ RETURN_ERROR(MAJOR, E_INVALID_SELECTION , ("Profile id is out of range"));
-+ *p_AbsoluteId = (uint16_t)(p_FmPcdPlcr->sharedProfilesIds[relativeProfile]);
-+ break;
-+ default:
-+ RETURN_ERROR(MAJOR, E_INVALID_SELECTION, ("Invalid policer profile type"));
-+ }
-+
-+ return E_OK;
-+}
-+
-+uint16_t FmPcdPlcrGetPortProfilesBase(t_Handle h_FmPcd, uint8_t hardwarePortId)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd *)h_FmPcd;
-+ uint16_t swPortIndex = 0;
-+
-+ HW_PORT_ID_TO_SW_PORT_INDX(swPortIndex, hardwarePortId);
-+
-+ return p_FmPcd->p_FmPcdPlcr->portsMapping[swPortIndex].profilesBase;
-+}
-+
-+uint16_t FmPcdPlcrGetPortNumOfProfiles(t_Handle h_FmPcd, uint8_t hardwarePortId)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd *)h_FmPcd;
-+ uint16_t swPortIndex = 0;
-+
-+ HW_PORT_ID_TO_SW_PORT_INDX(swPortIndex, hardwarePortId);
-+
-+ return p_FmPcd->p_FmPcdPlcr->portsMapping[swPortIndex].numOfProfiles;
-+
-+}
-+uint32_t FmPcdPlcrBuildWritePlcrActionReg(uint16_t absoluteProfileId)
-+{
-+ return (uint32_t)(FM_PCD_PLCR_PAR_GO |
-+ ((uint32_t)absoluteProfileId << FM_PCD_PLCR_PAR_PNUM_SHIFT));
-+}
-+
-+uint32_t FmPcdPlcrBuildWritePlcrActionRegs(uint16_t absoluteProfileId)
-+{
-+ return (uint32_t)(FM_PCD_PLCR_PAR_GO |
-+ ((uint32_t)absoluteProfileId << FM_PCD_PLCR_PAR_PNUM_SHIFT) |
-+ FM_PCD_PLCR_PAR_PWSEL_MASK);
-+}
-+
-+bool FmPcdPlcrHwProfileIsValid(uint32_t profileModeReg)
-+{
-+
-+ if (profileModeReg & FM_PCD_PLCR_PEMODE_PI)
-+ return TRUE;
-+ else
-+ return FALSE;
-+}
-+
-+uint32_t FmPcdPlcrBuildReadPlcrActionReg(uint16_t absoluteProfileId)
-+{
-+ return (uint32_t)(FM_PCD_PLCR_PAR_GO |
-+ FM_PCD_PLCR_PAR_R |
-+ ((uint32_t)absoluteProfileId << FM_PCD_PLCR_PAR_PNUM_SHIFT) |
-+ FM_PCD_PLCR_PAR_PWSEL_MASK);
-+}
-+
-+uint32_t FmPcdPlcrBuildCounterProfileReg(e_FmPcdPlcrProfileCounters counter)
-+{
-+ switch (counter)
-+ {
-+ case (e_FM_PCD_PLCR_PROFILE_GREEN_PACKET_TOTAL_COUNTER):
-+ return FM_PCD_PLCR_PAR_PWSEL_PEGPC;
-+ case (e_FM_PCD_PLCR_PROFILE_YELLOW_PACKET_TOTAL_COUNTER):
-+ return FM_PCD_PLCR_PAR_PWSEL_PEYPC;
-+ case (e_FM_PCD_PLCR_PROFILE_RED_PACKET_TOTAL_COUNTER) :
-+ return FM_PCD_PLCR_PAR_PWSEL_PERPC;
-+ case (e_FM_PCD_PLCR_PROFILE_RECOLOURED_YELLOW_PACKET_TOTAL_COUNTER) :
-+ return FM_PCD_PLCR_PAR_PWSEL_PERYPC;
-+ case (e_FM_PCD_PLCR_PROFILE_RECOLOURED_RED_PACKET_TOTAL_COUNTER) :
-+ return FM_PCD_PLCR_PAR_PWSEL_PERRPC;
-+ default:
-+ REPORT_ERROR(MAJOR, E_INVALID_SELECTION, NO_MSG);
-+ return 0;
-+ }
-+}
-+
-+uint32_t FmPcdPlcrBuildNiaProfileReg(bool green, bool yellow, bool red)
-+{
-+
-+ uint32_t tmpReg32 = 0;
-+
-+ if (green)
-+ tmpReg32 |= FM_PCD_PLCR_PAR_PWSEL_PEGNIA;
-+ if (yellow)
-+ tmpReg32 |= FM_PCD_PLCR_PAR_PWSEL_PEYNIA;
-+ if (red)
-+ tmpReg32 |= FM_PCD_PLCR_PAR_PWSEL_PERNIA;
-+
-+ return tmpReg32;
-+}
-+
-+void FmPcdPlcrUpdateRequiredAction(t_Handle h_FmPcd, uint16_t absoluteProfileId, uint32_t requiredAction)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+
-+ /* this routine is protected by calling routine */
-+
-+ ASSERT_COND(p_FmPcd->p_FmPcdPlcr->profiles[absoluteProfileId].valid);
-+
-+ p_FmPcd->p_FmPcdPlcr->profiles[absoluteProfileId].requiredAction |= requiredAction;
-+}
-+
-+/*********************** End of inter-module routines ************************/
-+
-+
-+/**************************************************/
-+/*............Policer API.........................*/
-+/**************************************************/
-+
-+t_Error FM_PCD_ConfigPlcrAutoRefreshMode(t_Handle h_FmPcd, bool enable)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd->p_FmPcdDriverParam, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd->p_FmPcdPlcr, E_INVALID_HANDLE);
-+
-+ if (!FmIsMaster(p_FmPcd->h_Fm))
-+ RETURN_ERROR(MAJOR, E_NOT_SUPPORTED, ("FM_PCD_ConfigPlcrAutoRefreshMode - guest mode!"));
-+
-+ p_FmPcd->p_FmPcdDriverParam->plcrAutoRefresh = enable;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PCD_ConfigPlcrNumOfSharedProfiles(t_Handle h_FmPcd, uint16_t numOfSharedPlcrProfiles)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd->p_FmPcdDriverParam, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd->p_FmPcdPlcr, E_INVALID_HANDLE);
-+
-+ p_FmPcd->p_FmPcdPlcr->numOfSharedProfiles = numOfSharedPlcrProfiles;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PCD_SetPlcrStatistics(t_Handle h_FmPcd, bool enable)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+ uint32_t tmpReg32;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmPcd->p_FmPcdDriverParam, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd->p_FmPcdPlcr, E_INVALID_HANDLE);
-+
-+ if (!FmIsMaster(p_FmPcd->h_Fm))
-+ RETURN_ERROR(MAJOR, E_NOT_SUPPORTED, ("FM_PCD_SetPlcrStatistics - guest mode!"));
-+
-+ tmpReg32 = GET_UINT32(p_FmPcd->p_FmPcdPlcr->p_FmPcdPlcrRegs->fmpl_gcr);
-+ if (enable)
-+ tmpReg32 |= FM_PCD_PLCR_GCR_STEN;
-+ else
-+ tmpReg32 &= ~FM_PCD_PLCR_GCR_STEN;
-+
-+ WRITE_UINT32(p_FmPcd->p_FmPcdPlcr->p_FmPcdPlcrRegs->fmpl_gcr, tmpReg32);
-+ return E_OK;
-+}
-+
-+t_Handle FM_PCD_PlcrProfileSet(t_Handle h_FmPcd,
-+ t_FmPcdPlcrProfileParams *p_ProfileParams)
-+{
-+ t_FmPcd *p_FmPcd;
-+ t_FmPcdPlcrRegs *p_FmPcdPlcrRegs;
-+ t_FmPcdPlcrProfileRegs plcrProfileReg;
-+ uint32_t intFlags;
-+ uint16_t absoluteProfileId;
-+ t_Error err = E_OK;
-+ uint32_t tmpReg32;
-+ t_FmPcdPlcrProfile *p_Profile;
-+
-+ SANITY_CHECK_RETURN_VALUE(h_FmPcd, E_INVALID_HANDLE, NULL);
-+
-+ if (p_ProfileParams->modify)
-+ {
-+ p_Profile = (t_FmPcdPlcrProfile *)p_ProfileParams->id.h_Profile;
-+ p_FmPcd = p_Profile->h_FmPcd;
-+ absoluteProfileId = p_Profile->absoluteProfileId;
-+ if (absoluteProfileId >= FM_PCD_PLCR_NUM_ENTRIES)
-+ {
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE, ("profileId too Big "));
-+ return NULL;
-+ }
-+
-+ SANITY_CHECK_RETURN_VALUE(p_FmPcd->p_FmPcdPlcr, E_INVALID_HANDLE, NULL);
-+
-+ /* Try lock profile using flag */
-+ if (!PlcrProfileFlagTryLock(p_Profile))
-+ {
-+ DBG(TRACE, ("Profile Try Lock - BUSY"));
-+ /* Signal to caller BUSY condition */
-+ p_ProfileParams->id.h_Profile = NULL;
-+ return NULL;
-+ }
-+ }
-+ else
-+ {
-+ p_FmPcd = (t_FmPcd*)h_FmPcd;
-+
-+ SANITY_CHECK_RETURN_VALUE(p_FmPcd->p_FmPcdPlcr, E_INVALID_HANDLE, NULL);
-+
-+ /* SMP: needs to be protected only if another core now changes the windows */
-+ err = FmPcdPlcrGetAbsoluteIdByProfileParams(h_FmPcd,
-+ p_ProfileParams->id.newParams.profileType,
-+ p_ProfileParams->id.newParams.h_FmPort,
-+ p_ProfileParams->id.newParams.relativeProfileId,
-+ &absoluteProfileId);
-+ if (err)
-+ {
-+ REPORT_ERROR(MAJOR, err, NO_MSG);
-+ return NULL;
-+ }
-+
-+ if (absoluteProfileId >= FM_PCD_PLCR_NUM_ENTRIES)
-+ {
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE, ("profileId too Big "));
-+ return NULL;
-+ }
-+
-+ if (FmPcdPlcrIsProfileValid(p_FmPcd, absoluteProfileId))
-+ {
-+ REPORT_ERROR(MAJOR, E_ALREADY_EXISTS, ("Policer Profile is already used"));
-+ return NULL;
-+ }
-+
-+ /* initialize profile struct */
-+ p_Profile = &p_FmPcd->p_FmPcdPlcr->profiles[absoluteProfileId];
-+
-+ p_Profile->h_FmPcd = p_FmPcd;
-+ p_Profile->absoluteProfileId = absoluteProfileId;
-+
-+ p_Profile->p_Lock = FmPcdAcquireLock(p_FmPcd);
-+ if (!p_Profile->p_Lock)
-+ REPORT_ERROR(MAJOR, E_NOT_AVAILABLE, ("FM Policer Profile lock obj!"));
-+ }
-+
-+ SANITY_CHECK_RETURN_VALUE(!p_FmPcd->p_FmPcdDriverParam, E_INVALID_STATE, NULL);
-+
-+ p_Profile->nextEngineOnGreen = p_ProfileParams->nextEngineOnGreen;
-+ memcpy(&p_Profile->paramsOnGreen, &(p_ProfileParams->paramsOnGreen), sizeof(u_FmPcdPlcrNextEngineParams));
-+
-+ p_Profile->nextEngineOnYellow = p_ProfileParams->nextEngineOnYellow;
-+ memcpy(&p_Profile->paramsOnYellow, &(p_ProfileParams->paramsOnYellow), sizeof(u_FmPcdPlcrNextEngineParams));
-+
-+ p_Profile->nextEngineOnRed = p_ProfileParams->nextEngineOnRed;
-+ memcpy(&p_Profile->paramsOnRed, &(p_ProfileParams->paramsOnRed), sizeof(u_FmPcdPlcrNextEngineParams));
-+
-+ memset(&plcrProfileReg, 0, sizeof(t_FmPcdPlcrProfileRegs));
-+
-+ /* build the policer profile registers */
-+ err = BuildProfileRegs(h_FmPcd, p_ProfileParams, &plcrProfileReg);
-+ if (err)
-+ {
-+ REPORT_ERROR(MAJOR, err, NO_MSG);
-+ if (p_ProfileParams->modify)
-+ /* unlock */
-+ PlcrProfileFlagUnlock(p_Profile);
-+ if (!p_ProfileParams->modify &&
-+ p_Profile->p_Lock)
-+ /* release allocated Profile lock */
-+ FmPcdReleaseLock(p_FmPcd, p_Profile->p_Lock);
-+ return NULL;
-+ }
-+
-+ if (p_FmPcd->h_Hc)
-+ {
-+ err = FmHcPcdPlcrSetProfile(p_FmPcd->h_Hc, (t_Handle)p_Profile, &plcrProfileReg);
-+ if (p_ProfileParams->modify)
-+ PlcrProfileFlagUnlock(p_Profile);
-+ if (err)
-+ {
-+ /* release the allocated scheme lock */
-+ if (!p_ProfileParams->modify &&
-+ p_Profile->p_Lock)
-+ FmPcdReleaseLock(p_FmPcd, p_Profile->p_Lock);
-+
-+ return NULL;
-+ }
-+ if (!p_ProfileParams->modify)
-+ FmPcdPlcrValidateProfileSw(p_FmPcd,absoluteProfileId);
-+ return (t_Handle)p_Profile;
-+ }
-+
-+ p_FmPcdPlcrRegs = p_FmPcd->p_FmPcdPlcr->p_FmPcdPlcrRegs;
-+ SANITY_CHECK_RETURN_VALUE(p_FmPcdPlcrRegs, E_INVALID_HANDLE, NULL);
-+
-+ intFlags = PlcrHwLock(p_FmPcd->p_FmPcdPlcr);
-+ WRITE_UINT32(p_FmPcdPlcrRegs->profileRegs.fmpl_pemode , plcrProfileReg.fmpl_pemode);
-+ WRITE_UINT32(p_FmPcdPlcrRegs->profileRegs.fmpl_pegnia , plcrProfileReg.fmpl_pegnia);
-+ WRITE_UINT32(p_FmPcdPlcrRegs->profileRegs.fmpl_peynia , plcrProfileReg.fmpl_peynia);
-+ WRITE_UINT32(p_FmPcdPlcrRegs->profileRegs.fmpl_pernia , plcrProfileReg.fmpl_pernia);
-+ WRITE_UINT32(p_FmPcdPlcrRegs->profileRegs.fmpl_pecir , plcrProfileReg.fmpl_pecir);
-+ WRITE_UINT32(p_FmPcdPlcrRegs->profileRegs.fmpl_pecbs , plcrProfileReg.fmpl_pecbs);
-+ WRITE_UINT32(p_FmPcdPlcrRegs->profileRegs.fmpl_pepepir_eir,plcrProfileReg.fmpl_pepepir_eir);
-+ WRITE_UINT32(p_FmPcdPlcrRegs->profileRegs.fmpl_pepbs_ebs,plcrProfileReg.fmpl_pepbs_ebs);
-+ WRITE_UINT32(p_FmPcdPlcrRegs->profileRegs.fmpl_pelts , plcrProfileReg.fmpl_pelts);
-+ WRITE_UINT32(p_FmPcdPlcrRegs->profileRegs.fmpl_pects , plcrProfileReg.fmpl_pects);
-+ WRITE_UINT32(p_FmPcdPlcrRegs->profileRegs.fmpl_pepts_ets,plcrProfileReg.fmpl_pepts_ets);
-+ WRITE_UINT32(p_FmPcdPlcrRegs->profileRegs.fmpl_pegpc , plcrProfileReg.fmpl_pegpc);
-+ WRITE_UINT32(p_FmPcdPlcrRegs->profileRegs.fmpl_peypc , plcrProfileReg.fmpl_peypc);
-+ WRITE_UINT32(p_FmPcdPlcrRegs->profileRegs.fmpl_perpc , plcrProfileReg.fmpl_perpc);
-+ WRITE_UINT32(p_FmPcdPlcrRegs->profileRegs.fmpl_perypc , plcrProfileReg.fmpl_perypc);
-+ WRITE_UINT32(p_FmPcdPlcrRegs->profileRegs.fmpl_perrpc , plcrProfileReg.fmpl_perrpc);
-+
-+ tmpReg32 = FmPcdPlcrBuildWritePlcrActionRegs(absoluteProfileId);
-+ WritePar(p_FmPcd, tmpReg32);
-+
-+ PlcrHwUnlock(p_FmPcd->p_FmPcdPlcr, intFlags);
-+
-+ if (!p_ProfileParams->modify)
-+ FmPcdPlcrValidateProfileSw(p_FmPcd,absoluteProfileId);
-+ else
-+ PlcrProfileFlagUnlock(p_Profile);
-+
-+ return (t_Handle)p_Profile;
-+}
-+
-+t_Error FM_PCD_PlcrProfileDelete(t_Handle h_Profile)
-+{
-+ t_FmPcdPlcrProfile *p_Profile = (t_FmPcdPlcrProfile*)h_Profile;
-+ t_FmPcd *p_FmPcd;
-+ uint16_t profileIndx;
-+ uint32_t tmpReg32, intFlags;
-+ t_Error err;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Profile, E_INVALID_HANDLE);
-+ p_FmPcd = p_Profile->h_FmPcd;
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd, E_INVALID_HANDLE);
-+
-+ profileIndx = p_Profile->absoluteProfileId;
-+
-+ UpdateRequiredActionFlag(p_FmPcd, profileIndx, FALSE);
-+
-+ FmPcdPlcrInvalidateProfileSw(p_FmPcd,profileIndx);
-+
-+ if (p_FmPcd->h_Hc)
-+ {
-+ err = FmHcPcdPlcrDeleteProfile(p_FmPcd->h_Hc, h_Profile);
-+ if (p_Profile->p_Lock)
-+ /* release allocated Profile lock */
-+ FmPcdReleaseLock(p_FmPcd, p_Profile->p_Lock);
-+
-+ return err;
-+ }
-+
-+ intFlags = PlcrHwLock(p_FmPcd->p_FmPcdPlcr);
-+ WRITE_UINT32(p_FmPcd->p_FmPcdPlcr->p_FmPcdPlcrRegs->profileRegs.fmpl_pemode, ~FM_PCD_PLCR_PEMODE_PI);
-+
-+ tmpReg32 = FmPcdPlcrBuildWritePlcrActionRegs(profileIndx);
-+ WritePar(p_FmPcd, tmpReg32);
-+ PlcrHwUnlock(p_FmPcd->p_FmPcdPlcr, intFlags);
-+
-+
-+ if (p_Profile->p_Lock)
-+ /* release allocated Profile lock */
-+ FmPcdReleaseLock(p_FmPcd, p_Profile->p_Lock);
-+
-+ /* we do not memset profile as all its fields are being re-initialized at "set",
-+ * plus its allocation information is still valid. */
-+ return E_OK;
-+}
-+
-+/***************************************************/
-+/*............Policer Profile Counter..............*/
-+/***************************************************/
-+uint32_t FM_PCD_PlcrProfileGetCounter(t_Handle h_Profile, e_FmPcdPlcrProfileCounters counter)
-+{
-+ t_FmPcdPlcrProfile *p_Profile = (t_FmPcdPlcrProfile*)h_Profile;
-+ t_FmPcd *p_FmPcd;
-+ uint16_t profileIndx;
-+ uint32_t intFlags, counterVal = 0;
-+ t_FmPcdPlcrRegs *p_FmPcdPlcrRegs;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Profile, E_INVALID_HANDLE);
-+ p_FmPcd = p_Profile->h_FmPcd;
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd, E_INVALID_HANDLE);
-+
-+ if (p_FmPcd->h_Hc)
-+ return FmHcPcdPlcrGetProfileCounter(p_FmPcd->h_Hc, h_Profile, counter);
-+
-+ p_FmPcdPlcrRegs = p_FmPcd->p_FmPcdPlcr->p_FmPcdPlcrRegs;
-+ SANITY_CHECK_RETURN_VALUE(p_FmPcdPlcrRegs, E_INVALID_HANDLE, 0);
-+
-+ profileIndx = p_Profile->absoluteProfileId;
-+
-+ if (profileIndx >= FM_PCD_PLCR_NUM_ENTRIES)
-+ {
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE, ("profileId too Big "));
-+ return 0;
-+ }
-+ intFlags = PlcrHwLock(p_FmPcd->p_FmPcdPlcr);
-+ WritePar(p_FmPcd, FmPcdPlcrBuildReadPlcrActionReg(profileIndx));
-+
-+ switch (counter)
-+ {
-+ case e_FM_PCD_PLCR_PROFILE_GREEN_PACKET_TOTAL_COUNTER:
-+ counterVal = (GET_UINT32(p_FmPcdPlcrRegs->profileRegs.fmpl_pegpc));
-+ break;
-+ case e_FM_PCD_PLCR_PROFILE_YELLOW_PACKET_TOTAL_COUNTER:
-+ counterVal = GET_UINT32(p_FmPcdPlcrRegs->profileRegs.fmpl_peypc);
-+ break;
-+ case e_FM_PCD_PLCR_PROFILE_RED_PACKET_TOTAL_COUNTER:
-+ counterVal = GET_UINT32(p_FmPcdPlcrRegs->profileRegs.fmpl_perpc);
-+ break;
-+ case e_FM_PCD_PLCR_PROFILE_RECOLOURED_YELLOW_PACKET_TOTAL_COUNTER:
-+ counterVal = GET_UINT32(p_FmPcdPlcrRegs->profileRegs.fmpl_perypc);
-+ break;
-+ case e_FM_PCD_PLCR_PROFILE_RECOLOURED_RED_PACKET_TOTAL_COUNTER:
-+ counterVal = GET_UINT32(p_FmPcdPlcrRegs->profileRegs.fmpl_perrpc);
-+ break;
-+ default:
-+ REPORT_ERROR(MAJOR, E_INVALID_SELECTION, NO_MSG);
-+ break;
-+ }
-+ PlcrHwUnlock(p_FmPcd->p_FmPcdPlcr, intFlags);
-+
-+ return counterVal;
-+}
-+
-+t_Error FM_PCD_PlcrProfileSetCounter(t_Handle h_Profile, e_FmPcdPlcrProfileCounters counter, uint32_t value)
-+{
-+ t_FmPcdPlcrProfile *p_Profile = (t_FmPcdPlcrProfile*)h_Profile;
-+ t_FmPcd *p_FmPcd;
-+ uint16_t profileIndx;
-+ uint32_t tmpReg32, intFlags;
-+ t_FmPcdPlcrRegs *p_FmPcdPlcrRegs;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Profile, E_INVALID_HANDLE);
-+
-+ p_FmPcd = p_Profile->h_FmPcd;
-+ profileIndx = p_Profile->absoluteProfileId;
-+
-+ if (p_FmPcd->h_Hc)
-+ return FmHcPcdPlcrSetProfileCounter(p_FmPcd->h_Hc, h_Profile, counter, value);
-+
-+ p_FmPcdPlcrRegs = p_FmPcd->p_FmPcdPlcr->p_FmPcdPlcrRegs;
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcdPlcrRegs, E_INVALID_HANDLE);
-+
-+ intFlags = PlcrHwLock(p_FmPcd->p_FmPcdPlcr);
-+ switch (counter)
-+ {
-+ case e_FM_PCD_PLCR_PROFILE_GREEN_PACKET_TOTAL_COUNTER:
-+ WRITE_UINT32(p_FmPcdPlcrRegs->profileRegs.fmpl_pegpc, value);
-+ break;
-+ case e_FM_PCD_PLCR_PROFILE_YELLOW_PACKET_TOTAL_COUNTER:
-+ WRITE_UINT32(p_FmPcdPlcrRegs->profileRegs.fmpl_peypc, value);
-+ break;
-+ case e_FM_PCD_PLCR_PROFILE_RED_PACKET_TOTAL_COUNTER:
-+ WRITE_UINT32(p_FmPcdPlcrRegs->profileRegs.fmpl_perpc, value);
-+ break;
-+ case e_FM_PCD_PLCR_PROFILE_RECOLOURED_YELLOW_PACKET_TOTAL_COUNTER:
-+ WRITE_UINT32(p_FmPcdPlcrRegs->profileRegs.fmpl_perypc ,value);
-+ break;
-+ case e_FM_PCD_PLCR_PROFILE_RECOLOURED_RED_PACKET_TOTAL_COUNTER:
-+ WRITE_UINT32(p_FmPcdPlcrRegs->profileRegs.fmpl_perrpc ,value);
-+ break;
-+ default:
-+ PlcrHwUnlock(p_FmPcd->p_FmPcdPlcr, intFlags);
-+ RETURN_ERROR(MAJOR, E_INVALID_SELECTION, NO_MSG);
-+ }
-+
-+ /* Activate the atomic write action by writing FMPL_PAR with: GO=1, RW=1, PSI=0, PNUM =
-+ * Profile Number, PWSEL=0xFFFF (select all words).
-+ */
-+ tmpReg32 = FmPcdPlcrBuildWritePlcrActionReg(profileIndx);
-+ tmpReg32 |= FmPcdPlcrBuildCounterProfileReg(counter);
-+ WritePar(p_FmPcd, tmpReg32);
-+ PlcrHwUnlock(p_FmPcd->p_FmPcdPlcr, intFlags);
-+
-+ return E_OK;
-+}
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/Pcd/fm_plcr.h
-@@ -0,0 +1,165 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/******************************************************************************
-+ @File fm_plcr.h
-+
-+ @Description FM Policer private header
-+*//***************************************************************************/
-+#ifndef __FM_PLCR_H
-+#define __FM_PLCR_H
-+
-+#include "std_ext.h"
-+
-+
-+/***********************************************************************/
-+/* Policer defines */
-+/***********************************************************************/
-+
-+#define FM_PCD_PLCR_PAR_GO 0x80000000
-+#define FM_PCD_PLCR_PAR_PWSEL_MASK 0x0000FFFF
-+#define FM_PCD_PLCR_PAR_R 0x40000000
-+
-+/* shifts */
-+#define FM_PCD_PLCR_PAR_PNUM_SHIFT 16
-+
-+/* masks */
-+#define FM_PCD_PLCR_PEMODE_PI 0x80000000
-+#define FM_PCD_PLCR_PEMODE_CBLND 0x40000000
-+#define FM_PCD_PLCR_PEMODE_ALG_MASK 0x30000000
-+#define FM_PCD_PLCR_PEMODE_ALG_RFC2698 0x10000000
-+#define FM_PCD_PLCR_PEMODE_ALG_RFC4115 0x20000000
-+#define FM_PCD_PLCR_PEMODE_DEFC_MASK 0x0C000000
-+#define FM_PCD_PLCR_PEMODE_DEFC_Y 0x04000000
-+#define FM_PCD_PLCR_PEMODE_DEFC_R 0x08000000
-+#define FM_PCD_PLCR_PEMODE_DEFC_OVERRIDE 0x0C000000
-+#define FM_PCD_PLCR_PEMODE_OVCLR_MASK 0x03000000
-+#define FM_PCD_PLCR_PEMODE_OVCLR_Y 0x01000000
-+#define FM_PCD_PLCR_PEMODE_OVCLR_R 0x02000000
-+#define FM_PCD_PLCR_PEMODE_OVCLR_G_NC 0x03000000
-+#define FM_PCD_PLCR_PEMODE_PKT 0x00800000
-+#define FM_PCD_PLCR_PEMODE_FPP_MASK 0x001F0000
-+#define FM_PCD_PLCR_PEMODE_FPP_SHIFT 16
-+#define FM_PCD_PLCR_PEMODE_FLS_MASK 0x0000F000
-+#define FM_PCD_PLCR_PEMODE_FLS_L2 0x00003000
-+#define FM_PCD_PLCR_PEMODE_FLS_L3 0x0000B000
-+#define FM_PCD_PLCR_PEMODE_FLS_L4 0x0000E000
-+#define FM_PCD_PLCR_PEMODE_FLS_FULL 0x0000F000
-+#define FM_PCD_PLCR_PEMODE_RBFLS 0x00000800
-+#define FM_PCD_PLCR_PEMODE_TRA 0x00000004
-+#define FM_PCD_PLCR_PEMODE_TRB 0x00000002
-+#define FM_PCD_PLCR_PEMODE_TRC 0x00000001
-+#define FM_PCD_PLCR_DOUBLE_ECC 0x80000000
-+#define FM_PCD_PLCR_INIT_ENTRY_ERROR 0x40000000
-+#define FM_PCD_PLCR_PRAM_SELF_INIT_COMPLETE 0x80000000
-+#define FM_PCD_PLCR_ATOMIC_ACTION_COMPLETE 0x40000000
-+
-+#define FM_PCD_PLCR_NIA_VALID 0x80000000
-+
-+#define FM_PCD_PLCR_GCR_EN 0x80000000
-+#define FM_PCD_PLCR_GCR_STEN 0x40000000
-+#define FM_PCD_PLCR_GCR_DAR 0x20000000
-+#define FM_PCD_PLCR_GCR_DEFNIA 0x00FFFFFF
-+#define FM_PCD_PLCR_NIA_ABS 0x00000100
-+
-+#define FM_PCD_PLCR_GSR_BSY 0x80000000
-+#define FM_PCD_PLCR_GSR_DQS 0x60000000
-+#define FM_PCD_PLCR_GSR_RPB 0x20000000
-+#define FM_PCD_PLCR_GSR_FQS 0x0C000000
-+#define FM_PCD_PLCR_GSR_LPALG 0x0000C000
-+#define FM_PCD_PLCR_GSR_LPCA 0x00003000
-+#define FM_PCD_PLCR_GSR_LPNUM 0x000000FF
-+
-+#define FM_PCD_PLCR_EVR_PSIC 0x80000000
-+#define FM_PCD_PLCR_EVR_AAC 0x40000000
-+
-+#define FM_PCD_PLCR_PAR_PSI 0x20000000
-+#define FM_PCD_PLCR_PAR_PNUM 0x00FF0000
-+/* PWSEL Selctive select options */
-+#define FM_PCD_PLCR_PAR_PWSEL_PEMODE 0x00008000 /* 0 */
-+#define FM_PCD_PLCR_PAR_PWSEL_PEGNIA 0x00004000 /* 1 */
-+#define FM_PCD_PLCR_PAR_PWSEL_PEYNIA 0x00002000 /* 2 */
-+#define FM_PCD_PLCR_PAR_PWSEL_PERNIA 0x00001000 /* 3 */
-+#define FM_PCD_PLCR_PAR_PWSEL_PECIR 0x00000800 /* 4 */
-+#define FM_PCD_PLCR_PAR_PWSEL_PECBS 0x00000400 /* 5 */
-+#define FM_PCD_PLCR_PAR_PWSEL_PEPIR_EIR 0x00000200 /* 6 */
-+#define FM_PCD_PLCR_PAR_PWSEL_PEPBS_EBS 0x00000100 /* 7 */
-+#define FM_PCD_PLCR_PAR_PWSEL_PELTS 0x00000080 /* 8 */
-+#define FM_PCD_PLCR_PAR_PWSEL_PECTS 0x00000040 /* 9 */
-+#define FM_PCD_PLCR_PAR_PWSEL_PEPTS_ETS 0x00000020 /* 10 */
-+#define FM_PCD_PLCR_PAR_PWSEL_PEGPC 0x00000010 /* 11 */
-+#define FM_PCD_PLCR_PAR_PWSEL_PEYPC 0x00000008 /* 12 */
-+#define FM_PCD_PLCR_PAR_PWSEL_PERPC 0x00000004 /* 13 */
-+#define FM_PCD_PLCR_PAR_PWSEL_PERYPC 0x00000002 /* 14 */
-+#define FM_PCD_PLCR_PAR_PWSEL_PERRPC 0x00000001 /* 15 */
-+
-+#define FM_PCD_PLCR_PAR_PMR_BRN_1TO1 0x0000 /* - Full bit replacement. {PBNUM[0:N-1]
-+ 1-> 2^N specific locations. */
-+#define FM_PCD_PLCR_PAR_PMR_BRN_2TO2 0x1 /* - {PBNUM[0:N-2],PNUM[N-1]}.
-+ 2-> 2^(N-1) base locations. */
-+#define FM_PCD_PLCR_PAR_PMR_BRN_4TO4 0x2 /* - {PBNUM[0:N-3],PNUM[N-2:N-1]}.
-+ 4-> 2^(N-2) base locations. */
-+#define FM_PCD_PLCR_PAR_PMR_BRN_8TO8 0x3 /* - {PBNUM[0:N-4],PNUM[N-3:N-1]}.
-+ 8->2^(N-3) base locations. */
-+#define FM_PCD_PLCR_PAR_PMR_BRN_16TO16 0x4 /* - {PBNUM[0:N-5],PNUM[N-4:N-1]}.
-+ 16-> 2^(N-4) base locations. */
-+#define FM_PCD_PLCR_PAR_PMR_BRN_32TO32 0x5 /* {PBNUM[0:N-6],PNUM[N-5:N-1]}.
-+ 32-> 2^(N-5) base locations. */
-+#define FM_PCD_PLCR_PAR_PMR_BRN_64TO64 0x6 /* {PBNUM[0:N-7],PNUM[N-6:N-1]}.
-+ 64-> 2^(N-6) base locations. */
-+#define FM_PCD_PLCR_PAR_PMR_BRN_128TO128 0x7 /* {PBNUM[0:N-8],PNUM[N-7:N-1]}.
-+ 128-> 2^(N-7) base locations. */
-+#define FM_PCD_PLCR_PAR_PMR_BRN_256TO256 0x8 /* - No bit replacement for N=8. {PNUM[N-8:N-1]}.
-+ When N=8 this option maps all 256 profiles by the DISPATCH bus into one group. */
-+
-+#define FM_PCD_PLCR_PMR_V 0x80000000
-+#define PLCR_ERR_ECC_CAP 0x80000000
-+#define PLCR_ERR_ECC_TYPE_DOUBLE 0x40000000
-+#define PLCR_ERR_ECC_PNUM_MASK 0x00000FF0
-+#define PLCR_ERR_ECC_OFFSET_MASK 0x0000000F
-+
-+#define PLCR_ERR_UNINIT_CAP 0x80000000
-+#define PLCR_ERR_UNINIT_NUM_MASK 0x000000FF
-+#define PLCR_ERR_UNINIT_PID_MASK 0x003f0000
-+#define PLCR_ERR_UNINIT_ABSOLUTE_MASK 0x00008000
-+
-+/* shifts */
-+#define PLCR_ERR_ECC_PNUM_SHIFT 4
-+#define PLCR_ERR_UNINIT_PID_SHIFT 16
-+
-+#define FM_PCD_PLCR_PMR_BRN_SHIFT 16
-+
-+#define PLCR_PORT_WINDOW_SIZE(hardwarePortId)
-+
-+
-+#endif /* __FM_PLCR_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/Pcd/fm_prs.c
-@@ -0,0 +1,423 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/******************************************************************************
-+ @File fm_pcd.c
-+
-+ @Description FM PCD ...
-+*//***************************************************************************/
-+#include <linux/math64.h>
-+#include "std_ext.h"
-+#include "error_ext.h"
-+#include "string_ext.h"
-+#include "debug_ext.h"
-+#include "net_ext.h"
-+
-+#include "fm_common.h"
-+#include "fm_pcd.h"
-+#include "fm_pcd_ipc.h"
-+#include "fm_prs.h"
-+#include "fsl_fman_prs.h"
-+
-+
-+static void PcdPrsErrorException(t_Handle h_FmPcd)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd *)h_FmPcd;
-+ uint32_t event, ev_mask;
-+ struct fman_prs_regs *PrsRegs = (struct fman_prs_regs *)p_FmPcd->p_FmPcdPrs->p_FmPcdPrsRegs;
-+
-+ ASSERT_COND(p_FmPcd->guestId == NCSW_MASTER_ID);
-+ ev_mask = fman_prs_get_err_ev_mask(PrsRegs);
-+
-+ event = fman_prs_get_err_event(PrsRegs, ev_mask);
-+
-+ fman_prs_ack_err_event(PrsRegs, event);
-+
-+ DBG(TRACE, ("parser error - 0x%08x\n",event));
-+
-+ if(event & FM_PCD_PRS_DOUBLE_ECC)
-+ p_FmPcd->f_Exception(p_FmPcd->h_App,e_FM_PCD_PRS_EXCEPTION_DOUBLE_ECC);
-+}
-+
-+static void PcdPrsException(t_Handle h_FmPcd)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd *)h_FmPcd;
-+ uint32_t event, ev_mask;
-+ struct fman_prs_regs *PrsRegs = (struct fman_prs_regs *)p_FmPcd->p_FmPcdPrs->p_FmPcdPrsRegs;
-+
-+ ASSERT_COND(p_FmPcd->guestId == NCSW_MASTER_ID);
-+ ev_mask = fman_prs_get_expt_ev_mask(PrsRegs);
-+ event = fman_prs_get_expt_event(PrsRegs, ev_mask);
-+
-+ ASSERT_COND(event & FM_PCD_PRS_SINGLE_ECC);
-+
-+ DBG(TRACE, ("parser event - 0x%08x\n",event));
-+
-+ fman_prs_ack_expt_event(PrsRegs, event);
-+
-+ p_FmPcd->f_Exception(p_FmPcd->h_App,e_FM_PCD_PRS_EXCEPTION_SINGLE_ECC);
-+}
-+
-+t_Handle PrsConfig(t_FmPcd *p_FmPcd,t_FmPcdParams *p_FmPcdParams)
-+{
-+ t_FmPcdPrs *p_FmPcdPrs;
-+ uintptr_t baseAddr;
-+
-+ UNUSED(p_FmPcd);
-+ UNUSED(p_FmPcdParams);
-+
-+ p_FmPcdPrs = (t_FmPcdPrs *) XX_Malloc(sizeof(t_FmPcdPrs));
-+ if (!p_FmPcdPrs)
-+ {
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("FM Parser structure allocation FAILED"));
-+ return NULL;
-+ }
-+ memset(p_FmPcdPrs, 0, sizeof(t_FmPcdPrs));
-+ fman_prs_defconfig(&p_FmPcd->p_FmPcdDriverParam->dfltCfg);
-+
-+ if (p_FmPcd->guestId == NCSW_MASTER_ID)
-+ {
-+ baseAddr = FmGetPcdPrsBaseAddr(p_FmPcdParams->h_Fm);
-+ p_FmPcdPrs->p_SwPrsCode = (uint32_t *)UINT_TO_PTR(baseAddr);
-+ p_FmPcdPrs->p_FmPcdPrsRegs = (struct fman_prs_regs *)UINT_TO_PTR(baseAddr + PRS_REGS_OFFSET);
-+ }
-+
-+ p_FmPcdPrs->fmPcdPrsPortIdStatistics = p_FmPcd->p_FmPcdDriverParam->dfltCfg.port_id_stat;
-+ p_FmPcd->p_FmPcdDriverParam->prsMaxParseCycleLimit = p_FmPcd->p_FmPcdDriverParam->dfltCfg.max_prs_cyc_lim;
-+ p_FmPcd->exceptions |= p_FmPcd->p_FmPcdDriverParam->dfltCfg.prs_exceptions;
-+
-+ return p_FmPcdPrs;
-+}
-+
-+#if ((DPAA_VERSION == 10) && defined(FM_CAPWAP_SUPPORT))
-+ static uint8_t swPrsPatch[] = SW_PRS_UDP_LITE_PATCH;
-+#else
-+ static uint8_t swPrsPatch[] = SW_PRS_OFFLOAD_PATCH;
-+#endif /* FM_CAPWAP_SUPPORT */
-+
-+t_Error PrsInit(t_FmPcd *p_FmPcd)
-+{
-+ t_FmPcdDriverParam *p_Param = p_FmPcd->p_FmPcdDriverParam;
-+ uint32_t *p_TmpCode;
-+ uint32_t *p_LoadTarget = (uint32_t *)PTR_MOVE(p_FmPcd->p_FmPcdPrs->p_SwPrsCode,
-+ FM_PCD_SW_PRS_SIZE-FM_PCD_PRS_SW_PATCHES_SIZE);
-+ struct fman_prs_regs *PrsRegs = (struct fman_prs_regs *)p_FmPcd->p_FmPcdPrs->p_FmPcdPrsRegs;
-+ uint32_t i;
-+
-+ ASSERT_COND(sizeof(swPrsPatch) <= (FM_PCD_PRS_SW_PATCHES_SIZE-FM_PCD_PRS_SW_TAIL_SIZE));
-+
-+ /* nothing to do in guest-partition */
-+ if (p_FmPcd->guestId != NCSW_MASTER_ID)
-+ return E_OK;
-+
-+ p_TmpCode = (uint32_t *)XX_MallocSmart(ROUND_UP(sizeof(swPrsPatch),4), 0, sizeof(uint32_t));
-+ if (!p_TmpCode)
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, ("Tmp Sw-Parser code allocation FAILED"));
-+ memset((uint8_t *)p_TmpCode, 0, ROUND_UP(sizeof(swPrsPatch),4));
-+ memcpy((uint8_t *)p_TmpCode, (uint8_t *)swPrsPatch, sizeof(swPrsPatch));
-+
-+ fman_prs_init(PrsRegs, &p_Param->dfltCfg);
-+
-+ /* register even if no interrupts enabled, to allow future enablement */
-+ FmRegisterIntr(p_FmPcd->h_Fm, e_FM_MOD_PRS, 0, e_FM_INTR_TYPE_ERR, PcdPrsErrorException, p_FmPcd);
-+
-+ /* register even if no interrupts enabled, to allow future enablement */
-+ FmRegisterIntr(p_FmPcd->h_Fm, e_FM_MOD_PRS, 0, e_FM_INTR_TYPE_NORMAL, PcdPrsException, p_FmPcd);
-+
-+ if(p_FmPcd->exceptions & FM_PCD_EX_PRS_SINGLE_ECC)
-+ FmEnableRamsEcc(p_FmPcd->h_Fm);
-+
-+ if(p_FmPcd->exceptions & FM_PCD_EX_PRS_DOUBLE_ECC)
-+ FmEnableRamsEcc(p_FmPcd->h_Fm);
-+
-+ /* load sw parser Ip-Frag patch */
-+ for (i=0; i<DIV_CEIL(sizeof(swPrsPatch), 4); i++)
-+ WRITE_UINT32(p_LoadTarget[i], GET_UINT32(p_TmpCode[i]));
-+
-+ XX_FreeSmart(p_TmpCode);
-+
-+ return E_OK;
-+}
-+
-+void PrsFree(t_FmPcd *p_FmPcd)
-+{
-+ ASSERT_COND(p_FmPcd->guestId == NCSW_MASTER_ID);
-+ FmUnregisterIntr(p_FmPcd->h_Fm, e_FM_MOD_PRS, 0, e_FM_INTR_TYPE_ERR);
-+ /* register even if no interrupts enabled, to allow future enablement */
-+ FmUnregisterIntr(p_FmPcd->h_Fm, e_FM_MOD_PRS, 0, e_FM_INTR_TYPE_NORMAL);
-+}
-+
-+void PrsEnable(t_FmPcd *p_FmPcd)
-+{
-+ struct fman_prs_regs *PrsRegs = (struct fman_prs_regs *)p_FmPcd->p_FmPcdPrs->p_FmPcdPrsRegs;
-+
-+ ASSERT_COND(p_FmPcd->guestId == NCSW_MASTER_ID);
-+ fman_prs_enable(PrsRegs);
-+}
-+
-+void PrsDisable(t_FmPcd *p_FmPcd)
-+{
-+ struct fman_prs_regs *PrsRegs = (struct fman_prs_regs *)p_FmPcd->p_FmPcdPrs->p_FmPcdPrsRegs;
-+
-+ ASSERT_COND(p_FmPcd->guestId == NCSW_MASTER_ID);
-+ fman_prs_disable(PrsRegs);
-+}
-+
-+int PrsIsEnabled(t_FmPcd *p_FmPcd)
-+{
-+ struct fman_prs_regs *PrsRegs = (struct fman_prs_regs *)p_FmPcd->p_FmPcdPrs->p_FmPcdPrsRegs;
-+
-+ ASSERT_COND(p_FmPcd->guestId == NCSW_MASTER_ID);
-+ return fman_prs_is_enabled(PrsRegs);
-+}
-+
-+t_Error PrsIncludePortInStatistics(t_FmPcd *p_FmPcd, uint8_t hardwarePortId, bool include)
-+{
-+ struct fman_prs_regs *PrsRegs;
-+ uint32_t bitMask = 0;
-+ uint8_t prsPortId;
-+
-+ SANITY_CHECK_RETURN_ERROR((hardwarePortId >=1 && hardwarePortId <= 16), E_INVALID_VALUE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd->p_FmPcdPrs, E_INVALID_HANDLE);
-+
-+ PrsRegs = (struct fman_prs_regs *)p_FmPcd->p_FmPcdPrs->p_FmPcdPrsRegs;
-+
-+ GET_FM_PCD_PRS_PORT_ID(prsPortId, hardwarePortId);
-+ GET_FM_PCD_INDEX_FLAG(bitMask, prsPortId);
-+
-+ if (include)
-+ p_FmPcd->p_FmPcdPrs->fmPcdPrsPortIdStatistics |= bitMask;
-+ else
-+ p_FmPcd->p_FmPcdPrs->fmPcdPrsPortIdStatistics &= ~bitMask;
-+
-+ fman_prs_set_stst_port_msk(PrsRegs,
-+ p_FmPcd->p_FmPcdPrs->fmPcdPrsPortIdStatistics);
-+
-+ return E_OK;
-+}
-+
-+t_Error FmPcdPrsIncludePortInStatistics(t_Handle h_FmPcd, uint8_t hardwarePortId, bool include)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd *)h_FmPcd;
-+ t_Error err;
-+
-+ SANITY_CHECK_RETURN_ERROR((hardwarePortId >=1 && hardwarePortId <= 16), E_INVALID_VALUE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd->p_FmPcdPrs, E_INVALID_HANDLE);
-+
-+ if ((p_FmPcd->guestId != NCSW_MASTER_ID) &&
-+ p_FmPcd->h_IpcSession)
-+ {
-+ t_FmPcdIpcPrsIncludePort prsIncludePortParams;
-+ t_FmPcdIpcMsg msg;
-+
-+ prsIncludePortParams.hardwarePortId = hardwarePortId;
-+ prsIncludePortParams.include = include;
-+ memset(&msg, 0, sizeof(msg));
-+ msg.msgId = FM_PCD_PRS_INC_PORT_STATS;
-+ memcpy(msg.msgBody, &prsIncludePortParams, sizeof(prsIncludePortParams));
-+ err = XX_IpcSendMessage(p_FmPcd->h_IpcSession,
-+ (uint8_t*)&msg,
-+ sizeof(msg.msgId) +sizeof(prsIncludePortParams),
-+ NULL,
-+ NULL,
-+ NULL,
-+ NULL);
-+ if (err != E_OK)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ return E_OK;
-+ }
-+ else if (p_FmPcd->guestId != NCSW_MASTER_ID)
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED,
-+ ("running in guest-mode without IPC!"));
-+
-+ return PrsIncludePortInStatistics(p_FmPcd, hardwarePortId, include);
-+}
-+
-+uint32_t FmPcdGetSwPrsOffset(t_Handle h_FmPcd, e_NetHeaderType hdr, uint8_t indexPerHdr)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd *)h_FmPcd;
-+ t_FmPcdPrsLabelParams *p_Label;
-+ int i;
-+
-+ SANITY_CHECK_RETURN_VALUE(p_FmPcd, E_INVALID_HANDLE, 0);
-+ SANITY_CHECK_RETURN_VALUE(!p_FmPcd->p_FmPcdDriverParam, E_INVALID_HANDLE, 0);
-+
-+ if ((p_FmPcd->guestId != NCSW_MASTER_ID) &&
-+ p_FmPcd->h_IpcSession)
-+ {
-+ t_Error err = E_OK;
-+ t_FmPcdIpcSwPrsLable labelParams;
-+ t_FmPcdIpcMsg msg;
-+ uint32_t prsOffset = 0;
-+ t_FmPcdIpcReply reply;
-+ uint32_t replyLength;
-+
-+ memset(&reply, 0, sizeof(reply));
-+ memset(&msg, 0, sizeof(msg));
-+ labelParams.enumHdr = (uint32_t)hdr;
-+ labelParams.indexPerHdr = indexPerHdr;
-+ msg.msgId = FM_PCD_GET_SW_PRS_OFFSET;
-+ memcpy(msg.msgBody, &labelParams, sizeof(labelParams));
-+ replyLength = sizeof(uint32_t) + sizeof(uint32_t);
-+ err = XX_IpcSendMessage(p_FmPcd->h_IpcSession,
-+ (uint8_t*)&msg,
-+ sizeof(msg.msgId) +sizeof(labelParams),
-+ (uint8_t*)&reply,
-+ &replyLength,
-+ NULL,
-+ NULL);
-+ if (err != E_OK)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ if (replyLength != sizeof(uint32_t) + sizeof(uint32_t))
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("IPC reply length mismatch"));
-+
-+ memcpy((uint8_t*)&prsOffset, reply.replyBody, sizeof(uint32_t));
-+ return prsOffset;
-+ }
-+ else if (p_FmPcd->guestId != NCSW_MASTER_ID)
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED,
-+ ("running in guest-mode without IPC!"));
-+
-+ ASSERT_COND(p_FmPcd->p_FmPcdPrs->currLabel < FM_PCD_PRS_NUM_OF_LABELS);
-+
-+ for (i=0; i<p_FmPcd->p_FmPcdPrs->currLabel; i++)
-+ {
-+ p_Label = &p_FmPcd->p_FmPcdPrs->labelsTable[i];
-+
-+ if ((hdr == p_Label->hdr) && (indexPerHdr == p_Label->indexPerHdr))
-+ return p_Label->instructionOffset;
-+ }
-+
-+ REPORT_ERROR(MAJOR, E_NOT_FOUND, ("Sw Parser attachment Not found"));
-+ return (uint32_t)ILLEGAL_BASE;
-+}
-+
-+void FM_PCD_SetPrsStatistics(t_Handle h_FmPcd, bool enable)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+ struct fman_prs_regs *PrsRegs;
-+
-+ SANITY_CHECK_RETURN(p_FmPcd, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN(p_FmPcd->p_FmPcdPrs, E_INVALID_HANDLE);
-+
-+ PrsRegs = (struct fman_prs_regs *)p_FmPcd->p_FmPcdPrs->p_FmPcdPrsRegs;
-+
-+
-+ if(p_FmPcd->guestId != NCSW_MASTER_ID)
-+ {
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("FM_PCD_SetPrsStatistics - guest mode!"));
-+ return;
-+ }
-+
-+ fman_prs_set_stst(PrsRegs, enable);
-+}
-+
-+t_Error FM_PCD_PrsLoadSw(t_Handle h_FmPcd, t_FmPcdPrsSwParams *p_SwPrs)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+ uint32_t *p_LoadTarget;
-+ uint32_t *p_TmpCode;
-+ int i;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmPcd->p_FmPcdDriverParam, E_INVALID_STATE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd->p_FmPcdPrs, E_INVALID_STATE);
-+ SANITY_CHECK_RETURN_ERROR(p_SwPrs, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmPcd->enabled, E_INVALID_HANDLE);
-+
-+ if (p_FmPcd->guestId != NCSW_MASTER_ID)
-+ RETURN_ERROR(MAJOR, E_NOT_SUPPORTED, ("FM in guest-mode!"));
-+
-+ if (!p_SwPrs->override)
-+ {
-+ if(p_FmPcd->p_FmPcdPrs->p_CurrSwPrs > p_FmPcd->p_FmPcdPrs->p_SwPrsCode + p_SwPrs->base*2/4)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("SW parser base must be larger than current loaded code"));
-+ }
-+ else
-+ p_FmPcd->p_FmPcdPrs->currLabel = 0;
-+
-+ if (p_SwPrs->size > FM_PCD_SW_PRS_SIZE - FM_PCD_PRS_SW_TAIL_SIZE - p_SwPrs->base*2)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("p_SwPrs->size may not be larger than MAX_SW_PRS_CODE_SIZE"));
-+
-+ if (p_FmPcd->p_FmPcdPrs->currLabel + p_SwPrs->numOfLabels > FM_PCD_PRS_NUM_OF_LABELS)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("Exceeded number of labels allowed "));
-+
-+ p_TmpCode = (uint32_t *)XX_MallocSmart(ROUND_UP(p_SwPrs->size,4), 0, sizeof(uint32_t));
-+ if (!p_TmpCode)
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, ("Tmp Sw-Parser code allocation FAILED"));
-+ memset((uint8_t *)p_TmpCode, 0, ROUND_UP(p_SwPrs->size,4));
-+ memcpy((uint8_t *)p_TmpCode, p_SwPrs->p_Code, p_SwPrs->size);
-+
-+ /* save sw parser labels */
-+ memcpy(&p_FmPcd->p_FmPcdPrs->labelsTable[p_FmPcd->p_FmPcdPrs->currLabel],
-+ p_SwPrs->labelsTable,
-+ p_SwPrs->numOfLabels*sizeof(t_FmPcdPrsLabelParams));
-+ p_FmPcd->p_FmPcdPrs->currLabel += p_SwPrs->numOfLabels;
-+
-+ /* load sw parser code */
-+ p_LoadTarget = p_FmPcd->p_FmPcdPrs->p_SwPrsCode + p_SwPrs->base*2/4;
-+
-+ for(i=0; i<DIV_CEIL(p_SwPrs->size, 4); i++)
-+ WRITE_UINT32(p_LoadTarget[i], GET_UINT32(p_TmpCode[i]));
-+
-+ p_FmPcd->p_FmPcdPrs->p_CurrSwPrs =
-+ p_FmPcd->p_FmPcdPrs->p_SwPrsCode + p_SwPrs->base*2/4 + ROUND_UP(p_SwPrs->size,4);
-+
-+ /* copy data parameters */
-+ for (i=0;i<FM_PCD_PRS_NUM_OF_HDRS;i++)
-+ WRITE_UINT32(*(p_FmPcd->p_FmPcdPrs->p_SwPrsCode+PRS_SW_DATA/4+i), p_SwPrs->swPrsDataParams[i]);
-+
-+ /* Clear last 4 bytes */
-+ WRITE_UINT32(*(p_FmPcd->p_FmPcdPrs->p_SwPrsCode+(PRS_SW_DATA-FM_PCD_PRS_SW_TAIL_SIZE)/4), 0);
-+
-+ XX_FreeSmart(p_TmpCode);
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PCD_ConfigPrsMaxCycleLimit(t_Handle h_FmPcd,uint16_t value)
-+{
-+ t_FmPcd *p_FmPcd = (t_FmPcd*)h_FmPcd;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPcd->p_FmPcdDriverParam, E_INVALID_HANDLE);
-+
-+ if(p_FmPcd->guestId != NCSW_MASTER_ID)
-+ RETURN_ERROR(MAJOR, E_NOT_SUPPORTED, ("FM_PCD_ConfigPrsMaxCycleLimit - guest mode!"));
-+
-+ p_FmPcd->p_FmPcdDriverParam->prsMaxParseCycleLimit = value;
-+
-+ return E_OK;
-+}
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/Pcd/fm_prs.h
-@@ -0,0 +1,316 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/******************************************************************************
-+ @File fm_prs.h
-+
-+ @Description FM Parser private header
-+ *//***************************************************************************/
-+#ifndef __FM_PRS_H
-+#define __FM_PRS_H
-+
-+#include "std_ext.h"
-+
-+/***********************************************************************/
-+/* SW parser IP_FRAG patch */
-+/***********************************************************************/
-+
-+#if ((DPAA_VERSION == 10) && defined(FM_CAPWAP_SUPPORT))
-+#define SW_PRS_UDP_LITE_PATCH \
-+{\
-+ 0x31,0x52,0x00,0xDA,0xFC,0x00,0x00,0x00,0x00,0x00, \
-+ 0x00,0x00,0x50,0x2C,0x40,0x00,0x31,0x92,0x50,0x2C, \
-+ 0x00,0x88,0x18,0x2F,0x00,0x01,0x1B,0xFE,0x18,0x71, \
-+ 0x02,0x1F,0x00,0x08,0x00,0x83,0x02,0x1F,0x00,0x20, \
-+ 0x28,0x1B,0x00,0x05,0x29,0x1F,0x30,0xD0,0x60,0x4F, \
-+ 0x00,0x07,0x00,0x05,0x00,0x00,0xC3,0x8F,0x00,0x52, \
-+ 0x00,0x01,0x07,0x01,0x60,0x3B,0x00,0x00,0x30,0xD0, \
-+ 0x00,0xDA,0x00,0x01,0x00,0x00,0x00,0x00,0x00,0x00, \
-+ 0x40,0x4C,0x00,0x00,0x02,0x8F,0x00,0x00,0x30,0xF2, \
-+ 0x00,0x06,0x18,0x5D,0x00,0x00,0x9F,0xFF,0x30,0xF2, \
-+ 0x00,0x06,0x29,0x1E,0x07,0x08,0x30,0xD0,0x00,0x52, \
-+ 0x00,0x08,0x28,0x1A,0x60,0x37,0x00,0x00,0x30,0xF2, \
-+ 0x18,0x5D,0x06,0x00,0x29,0x1E,0x30,0xF2,0x2F,0x0E, \
-+ 0x30,0x72,0x00,0x00,0x9B,0x8F,0x00,0x06,0x2F,0x0E, \
-+ 0x32,0xF1,0x32,0xB0,0x00,0x4F,0x00,0x57,0x00,0x28, \
-+ 0x00,0x00,0x97,0x9E,0x00,0x4E,0x30,0x72,0x00,0x06, \
-+ 0x2F,0x0E,0x32,0xC1,0x32,0xF0,0x00,0x4A,0x00,0x80, \
-+ 0x00,0x02,0x00,0x00,0x97,0x9E,0x40,0x7E,0x00,0x08, \
-+ 0x08,0x16,0x00,0x54,0x00,0x01,0x1B,0xFE,0x00,0x00, \
-+ 0x9F,0x9E,0x40,0xB3,0x00,0x00,0x02,0x1F,0x00,0x08, \
-+ 0x28,0x1B,0x30,0x73,0x29,0x1F,0x30,0xD0,0x60,0x9F, \
-+ 0x00,0x07,0x00,0x05,0x00,0x00,0xC3,0x8F,0x00,0x52, \
-+ 0x00,0x01,0x07,0x01,0x60,0x8B,0x00,0x00,0x30,0xD0, \
-+ 0x00,0xDA,0x00,0x01,0x00,0x00,0x00,0x00,0x00,0x00, \
-+ 0x40,0x9C,0x00,0x00,0x02,0x8F,0x00,0x00,0x30,0xF2, \
-+ 0x00,0x06,0x18,0xAD,0x00,0x00,0x9F,0xFF,0x30,0xF2, \
-+ 0x00,0x06,0x29,0x1E,0x07,0x08,0x30,0xD0,0x00,0x52, \
-+ 0x00,0x08,0x28,0x1A,0x60,0x87,0x00,0x00,0x30,0xF2, \
-+ 0x18,0xAD,0x06,0x00,0x29,0x1E,0x30,0xF2,0x50,0xB3, \
-+ 0xFF,0xFF,0x18,0xB8,0x08,0x16,0x00,0x54,0x00,0x01, \
-+ 0x1B,0xFE,0x18,0xC5,0x32,0xF1,0x28,0x5D,0x32,0xF1, \
-+ 0x00,0x55,0x00,0x08,0x28,0x5F,0x00,0x00,0x8F,0x9F, \
-+ 0x29,0x33,0x08,0x16,0x00,0x49,0x00,0x01,0x1B,0xFF, \
-+ 0x00,0x01,0x1B,0xFF \
-+}
-+#endif /* ((DPAA_VERSION == 10) && defined(FM_CAPWAP_SUPPORT)) */
-+
-+#if (DPAA_VERSION == 10)
-+/* Version: 106.1.9 */
-+#define SW_PRS_OFFLOAD_PATCH \
-+{ \
-+ 0x31,0x52,0x00,0xDA,0x0A,0x00,0x00,0x00,0x00,0x00, \
-+ 0x00,0x00,0x43,0x0A,0x00,0x00,0x00,0x01,0x1B,0xFE, \
-+ 0x00,0x00,0x99,0x00,0x53,0x13,0x00,0x00,0x00,0x00, \
-+ 0x9F,0x98,0x53,0x13,0x00,0x00,0x1B,0x23,0x33,0xF1, \
-+ 0x00,0xF9,0x00,0x01,0x00,0x00,0x00,0x00,0x00,0x00, \
-+ 0x28,0x7F,0x00,0x03,0x00,0x02,0x00,0x00,0x00,0x01, \
-+ 0x32,0xC1,0x32,0xF0,0x00,0x4A,0x00,0x80,0x1F,0xFF, \
-+ 0x00,0x01,0x1B,0xFE,0x31,0x52,0x00,0xDA,0x06,0x00, \
-+ 0x00,0x00,0x00,0x00,0x00,0x00,0x43,0x2F,0x00,0x00, \
-+ 0x00,0x01,0x1B,0xFE,0x31,0x52,0x00,0xDA,0x00,0x40, \
-+ 0x00,0x00,0x00,0x00,0x00,0x00,0x53,0x95,0x00,0x00, \
-+ 0x00,0x00,0x9B,0x8F,0x2F,0x0F,0x32,0xC1,0x00,0x55, \
-+ 0x00,0x28,0x28,0x43,0x30,0x7E,0x43,0x45,0x00,0x00, \
-+ 0x30,0x7E,0x43,0x45,0x00,0x3C,0x1B,0x5D,0x32,0x11, \
-+ 0x32,0xC0,0x00,0x4F,0x00,0x81,0x00,0x00,0x83,0x8F, \
-+ 0x2F,0x0F,0x06,0x00,0x32,0x11,0x32,0xC0,0x00,0x4F, \
-+ 0x00,0x55,0x00,0x01,0x00,0x81,0x32,0x11,0x00,0x00, \
-+ 0x83,0x8E,0x00,0x50,0x00,0x01,0x01,0x04,0x00,0x4D, \
-+ 0x28,0x43,0x06,0x00,0x1B,0x3E,0x30,0x7E,0x53,0x79, \
-+ 0x00,0x2B,0x32,0x11,0x32,0xC0,0x00,0x4F,0x00,0x81, \
-+ 0x00,0x00,0x87,0x8F,0x28,0x23,0x06,0x00,0x32,0x11, \
-+ 0x32,0xC0,0x00,0x4F,0x00,0x55,0x00,0x01,0x00,0x81, \
-+ 0x32,0x11,0x00,0x00,0x83,0x8E,0x00,0x50,0x00,0x01, \
-+ 0x01,0x04,0x00,0x4D,0x28,0x43,0x06,0x00,0x00,0x01, \
-+ 0x1B,0xFE,0x00,0x00,0x9B,0x8E,0x53,0x90,0x00,0x00, \
-+ 0x06,0x29,0x00,0x00,0x83,0x8F,0x28,0x23,0x06,0x00, \
-+ 0x06,0x29,0x32,0xC1,0x00,0x55,0x00,0x28,0x00,0x00, \
-+ 0x83,0x8E,0x00,0x50,0x00,0x01,0x01,0x04,0x00,0x4D, \
-+ 0x28,0x43,0x06,0x00,0x00,0x01,0x1B,0xFE,0x32,0xC1, \
-+ 0x00,0x55,0x00,0x28,0x28,0x43,0x1B,0xCF,0x00,0x00, \
-+ 0x9B,0x8F,0x2F,0x0F,0x32,0xC1,0x00,0x55,0x00,0x28, \
-+ 0x28,0x43,0x30,0x7E,0x43,0xBF,0x00,0x2C,0x32,0x11, \
-+ 0x32,0xC0,0x00,0x4F,0x00,0x81,0x00,0x00,0x87,0x8F, \
-+ 0x28,0x23,0x06,0x00,0x32,0x11,0x32,0xC0,0x00,0x4F, \
-+ 0x00,0x81,0x00,0x00,0x83,0x8F,0x2F,0x0F,0x06,0x00, \
-+ 0x32,0x11,0x32,0xC0,0x00,0x4F,0x00,0x55,0x00,0x01, \
-+ 0x00,0x81,0x32,0x11,0x00,0x00,0x83,0x8E,0x00,0x50, \
-+ 0x00,0x01,0x01,0x04,0x00,0x4D,0x28,0x43,0x06,0x00, \
-+ 0x1B,0x9C,0x33,0xF1,0x00,0xF9,0x00,0x01,0x00,0x00, \
-+ 0x00,0x00,0x00,0x00,0x28,0x7F,0x00,0x03,0x00,0x02, \
-+ 0x00,0x00,0x00,0x01,0x32,0xC1,0x32,0xF0,0x00,0x4A, \
-+ 0x00,0x80,0x1F,0xFF,0x00,0x01,0x1B,0xFE, \
-+}
-+
-+#else
-+#define SW_PRS_OFFLOAD_PATCH \
-+{ \
-+ 0x31,0x52,0x00,0xDA,0x0E,0x4F,0x00,0x00,0x00,0x00, \
-+ 0x00,0x00,0x51,0x16,0x08,0x4B,0x31,0x53,0x00,0xFB, \
-+ 0xFF,0xF0,0x00,0x00,0x00,0x00,0x00,0x00,0x29,0x2B, \
-+ 0x33,0xF1,0x00,0xFB,0x00,0xDF,0x00,0x00,0x00,0x00, \
-+ 0x00,0x00,0x28,0x7F,0x31,0x52,0x00,0xDA,0x0A,0x00, \
-+ 0x00,0x00,0x00,0x00,0x00,0x00,0x41,0x20,0x00,0x00, \
-+ 0x00,0x01,0x1B,0xFE,0x00,0x00,0x99,0x00,0x51,0x29, \
-+ 0x00,0x00,0x00,0x00,0x9F,0x98,0x51,0x29,0x00,0x00, \
-+ 0x19,0x44,0x09,0x5F,0x00,0x20,0x00,0x00,0x09,0x4F, \
-+ 0x00,0x20,0x00,0x00,0x34,0xB7,0x00,0xF9,0x00,0x00, \
-+ 0x01,0x00,0x00,0x00,0x00,0x00,0x2B,0x97,0x31,0xB3, \
-+ 0x29,0x8F,0x33,0xF1,0x00,0xF9,0x00,0x01,0x00,0x00, \
-+ 0x00,0x00,0x00,0x00,0x28,0x7F,0x00,0x03,0x00,0x02, \
-+ 0x00,0x00,0x00,0x01,0x1B,0xFE,0x00,0x01,0x1B,0xFE, \
-+ 0x31,0x52,0x00,0xDA,0xFC,0x00,0x00,0x00,0x00,0x00, \
-+ 0x00,0x00,0x51,0x52,0x40,0x00,0x31,0x92,0x51,0x52, \
-+ 0x00,0x88,0x19,0x55,0x08,0x05,0x00,0x00,0x19,0x99, \
-+ 0x02,0x1F,0x00,0x08,0x00,0x83,0x02,0x1F,0x00,0x20, \
-+ 0x28,0x1B,0x00,0x05,0x29,0x1F,0x30,0xD0,0x61,0x75, \
-+ 0x00,0x07,0x00,0x05,0x00,0x00,0xC3,0x8F,0x00,0x52, \
-+ 0x00,0x01,0x07,0x01,0x61,0x61,0x00,0x00,0x30,0xD0, \
-+ 0x00,0xDA,0x00,0x01,0x00,0x00,0x00,0x00,0x00,0x00, \
-+ 0x41,0x72,0x00,0x00,0x02,0x8F,0x00,0x00,0x30,0xF2, \
-+ 0x00,0x06,0x19,0x83,0x00,0x00,0x9F,0xFF,0x30,0xF2, \
-+ 0x00,0x06,0x29,0x1E,0x07,0x08,0x30,0xD0,0x00,0x52, \
-+ 0x00,0x08,0x28,0x1A,0x61,0x5D,0x00,0x00,0x30,0xF2, \
-+ 0x19,0x83,0x06,0x00,0x29,0x1E,0x30,0xF2,0x29,0x0E, \
-+ 0x30,0x72,0x00,0x00,0x9B,0x8F,0x00,0x06,0x29,0x0E, \
-+ 0x32,0xF1,0x32,0xB0,0x00,0x4F,0x00,0x57,0x00,0x28, \
-+ 0x00,0x00,0x97,0x9E,0x00,0x4E,0x30,0x72,0x00,0x06, \
-+ 0x29,0x0E,0x08,0x05,0x00,0x01,0x31,0x52,0x00,0xDA, \
-+ 0x0E,0x4F,0x00,0x00,0x00,0x00,0x00,0x00,0x51,0xAF, \
-+ 0x04,0x4B,0x31,0x53,0x00,0xFB,0xFF,0xF0,0x00,0x00, \
-+ 0x00,0x00,0x00,0x00,0x29,0x2B,0x33,0xF1,0x00,0xFB, \
-+ 0x00,0xDF,0x00,0x00,0x00,0x00,0x00,0x00,0x28,0x7F, \
-+ 0x31,0x52,0x00,0xDA,0x06,0x00,0x00,0x00,0x00,0x00, \
-+ 0x00,0x00,0x41,0xB9,0x00,0x00,0x00,0x01,0x1B,0xFE, \
-+ 0x31,0x52,0x00,0xDA,0x00,0x40,0x00,0x00,0x00,0x00, \
-+ 0x00,0x00,0x42,0x06,0x00,0x00,0x00,0x00,0x9B,0x8F, \
-+ 0x28,0x01,0x32,0xC1,0x00,0x55,0x00,0x28,0x28,0x43, \
-+ 0x30,0x00,0x41,0xEB,0x00,0x2C,0x32,0x11,0x32,0xC0, \
-+ 0x00,0x4F,0x00,0x81,0x00,0x00,0x87,0x8F,0x28,0x23, \
-+ 0x06,0x00,0x32,0x11,0x32,0xC0,0x00,0x4F,0x00,0x81, \
-+ 0x00,0x00,0x83,0x8F,0x28,0x01,0x06,0x00,0x32,0x11, \
-+ 0x32,0xC0,0x00,0x4F,0x00,0x55,0x00,0x01,0x00,0x81, \
-+ 0x32,0x11,0x00,0x00,0x83,0x8E,0x00,0x50,0x00,0x01, \
-+ 0x01,0x04,0x00,0x4D,0x28,0x43,0x06,0x00,0x19,0xC8, \
-+ 0x09,0x5F,0x00,0x20,0x00,0x00,0x09,0x4F,0x00,0x20, \
-+ 0x00,0x00,0x34,0xB7,0x00,0xF9,0x00,0x00,0x01,0x00, \
-+ 0x00,0x00,0x00,0x00,0x2B,0x97,0x31,0xB3,0x29,0x8F, \
-+ 0x33,0xF1,0x00,0xF9,0x00,0x01,0x00,0x00,0x00,0x00, \
-+ 0x00,0x00,0x28,0x7F,0x00,0x03,0x00,0x02,0x00,0x00, \
-+ 0x00,0x01,0x1B,0xFE,0x30,0x50,0x52,0x0B,0x00,0x00, \
-+ 0x00,0x01,0x1B,0xFE,0x32,0xF1,0x32,0xC0,0x00,0x4F, \
-+ 0x00,0x81,0x00,0x02,0x00,0x00,0x97,0x9E,0x42,0x18, \
-+ 0x00,0x08,0x08,0x16,0x00,0x54,0x00,0x01,0x1B,0xFE, \
-+ 0x00,0x00,0x9F,0x9E,0x42,0x4D,0x00,0x00,0x02,0x1F, \
-+ 0x00,0x08,0x28,0x1B,0x30,0x73,0x29,0x1F,0x30,0xD0, \
-+ 0x62,0x39,0x00,0x07,0x00,0x05,0x00,0x00,0xC3,0x8F, \
-+ 0x00,0x52,0x00,0x01,0x07,0x01,0x62,0x25,0x00,0x00, \
-+ 0x30,0xD0,0x00,0xDA,0x00,0x01,0x00,0x00,0x00,0x00, \
-+ 0x00,0x00,0x42,0x36,0x00,0x00,0x02,0x8F,0x00,0x00, \
-+ 0x30,0xF2,0x00,0x06,0x1A,0x47,0x00,0x00,0x9F,0xFF, \
-+ 0x30,0xF2,0x00,0x06,0x29,0x1E,0x07,0x08,0x30,0xD0, \
-+ 0x00,0x52,0x00,0x08,0x28,0x1A,0x62,0x21,0x00,0x00, \
-+ 0x30,0xF2,0x1A,0x47,0x06,0x00,0x29,0x1E,0x30,0xF2, \
-+ 0x52,0x4D,0xFF,0xFF,0x1A,0x52,0x08,0x16,0x00,0x54, \
-+ 0x00,0x01,0x1B,0xFE,0x1A,0x5F,0x32,0xF1,0x28,0x5D, \
-+ 0x32,0xF1,0x00,0x55,0x00,0x08,0x28,0x5F,0x00,0x00, \
-+ 0x8F,0x9F,0x29,0x33,0x08,0x16,0x00,0x49,0x00,0x01, \
-+ 0x1B,0xFF,0x00,0x01,0x1B,0xFF,0x31,0x52,0x00,0xDA, \
-+ 0xFC,0x00,0x00,0x00,0x00,0x00,0x00,0x00,0x52,0x6D, \
-+ 0x40,0x00,0x31,0x92,0x52,0x6D,0x00,0x88,0x1A,0x70, \
-+ 0x08,0x05,0x00,0x00,0x1A,0xB4,0x02,0x1F,0x00,0x08, \
-+ 0x00,0x83,0x02,0x1F,0x00,0x20,0x28,0x1B,0x00,0x05, \
-+ 0x29,0x1F,0x30,0xD0,0x62,0x90,0x00,0x07,0x00,0x05, \
-+ 0x00,0x00,0xC3,0x8F,0x00,0x52,0x00,0x01,0x07,0x01, \
-+ 0x62,0x7C,0x00,0x00,0x30,0xD0,0x00,0xDA,0x00,0x01, \
-+ 0x00,0x00,0x00,0x00,0x00,0x00,0x42,0x8D,0x00,0x00, \
-+ 0x02,0x8F,0x00,0x00,0x30,0xF2,0x00,0x06,0x1A,0x9E, \
-+ 0x00,0x00,0x9F,0xFF,0x30,0xF2,0x00,0x06,0x29,0x1E, \
-+ 0x07,0x08,0x30,0xD0,0x00,0x52,0x00,0x08,0x28,0x1A, \
-+ 0x62,0x78,0x00,0x00,0x30,0xF2,0x1A,0x9E,0x06,0x00, \
-+ 0x29,0x1E,0x30,0xF2,0x29,0x0E,0x30,0x72,0x00,0x00, \
-+ 0x9B,0x8F,0x00,0x06,0x29,0x0E,0x32,0xF1,0x32,0xB0, \
-+ 0x00,0x4F,0x00,0x57,0x00,0x28,0x00,0x00,0x97,0x9E, \
-+ 0x00,0x4E,0x30,0x72,0x00,0x06,0x29,0x0E,0x08,0x05, \
-+ 0x00,0x01,0x31,0x52,0x00,0xDA,0x0E,0x4F,0x00,0x00, \
-+ 0x00,0x00,0x00,0x00,0x52,0xCA,0x04,0x4B,0x31,0x53, \
-+ 0x00,0xFB,0xFF,0xF0,0x00,0x00,0x00,0x00,0x00,0x00, \
-+ 0x29,0x2B,0x33,0xF1,0x00,0xFB,0x00,0xDF,0x00,0x00, \
-+ 0x00,0x00,0x00,0x00,0x28,0x7F,0x31,0x52,0x00,0xDA, \
-+ 0x06,0x00,0x00,0x00,0x00,0x00,0x00,0x00,0x42,0xD4, \
-+ 0x00,0x00,0x00,0x01,0x1B,0xFE,0x31,0x52,0x00,0xDA, \
-+ 0x00,0x40,0x00,0x00,0x00,0x00,0x00,0x00,0x53,0x37, \
-+ 0x00,0x00,0x00,0x00,0x9B,0x8F,0x28,0x01,0x32,0xC1, \
-+ 0x00,0x55,0x00,0x28,0x28,0x43,0x30,0x00,0x42,0xEA, \
-+ 0x00,0x00,0x30,0x00,0x42,0xEA,0x00,0x3C,0x1B,0x02, \
-+ 0x32,0x11,0x32,0xC0,0x00,0x4F,0x00,0x81,0x00,0x00, \
-+ 0x83,0x8F,0x28,0x01,0x06,0x00,0x32,0x11,0x32,0xC0, \
-+ 0x00,0x4F,0x00,0x55,0x00,0x01,0x00,0x81,0x32,0x11, \
-+ 0x00,0x00,0x83,0x8E,0x00,0x50,0x00,0x01,0x01,0x04, \
-+ 0x00,0x4D,0x28,0x43,0x06,0x00,0x1A,0xE3,0x30,0x00, \
-+ 0x43,0x20,0x00,0x2B,0x00,0x00,0x9B,0x8E,0x43,0x0E, \
-+ 0x00,0x00,0x32,0xC1,0x00,0x55,0x00,0x28,0x28,0x43, \
-+ 0x1B,0x1F,0x06,0x29,0x00,0x00,0x83,0x8F,0x28,0x23, \
-+ 0x06,0x00,0x06,0x29,0x32,0xC1,0x00,0x55,0x00,0x28, \
-+ 0x00,0x00,0x83,0x8E,0x00,0x50,0x00,0x01,0x01,0x04, \
-+ 0x00,0x4D,0x28,0x43,0x06,0x00,0x1B,0x37,0x32,0x11, \
-+ 0x32,0xC0,0x00,0x4F,0x00,0x81,0x00,0x00,0x87,0x8F, \
-+ 0x28,0x23,0x06,0x00,0x32,0x11,0x32,0xC0,0x00,0x4F, \
-+ 0x00,0x55,0x00,0x01,0x00,0x81,0x32,0x11,0x00,0x00, \
-+ 0x83,0x8E,0x00,0x50,0x00,0x01,0x01,0x04,0x00,0x4D, \
-+ 0x28,0x43,0x06,0x00,0x30,0x50,0x53,0x3C,0x00,0x00, \
-+ 0x00,0x01,0x1B,0xFE,0x32,0xF1,0x32,0xC0,0x00,0x4F, \
-+ 0x00,0x81,0x00,0x02,0x00,0x00,0x97,0x9E,0x43,0x49, \
-+ 0x00,0x08,0x08,0x16,0x00,0x54,0x00,0x01,0x1B,0xFE, \
-+ 0x00,0x00,0x9F,0x9E,0x43,0x7E,0x00,0x00,0x02,0x1F, \
-+ 0x00,0x08,0x28,0x1B,0x30,0x73,0x29,0x1F,0x30,0xD0, \
-+ 0x63,0x6A,0x00,0x07,0x00,0x05,0x00,0x00,0xC3,0x8F, \
-+ 0x00,0x52,0x00,0x01,0x07,0x01,0x63,0x56,0x00,0x00, \
-+ 0x30,0xD0,0x00,0xDA,0x00,0x01,0x00,0x00,0x00,0x00, \
-+ 0x00,0x00,0x43,0x67,0x00,0x00,0x02,0x8F,0x00,0x00, \
-+ 0x30,0xF2,0x00,0x06,0x1B,0x78,0x00,0x00,0x9F,0xFF, \
-+ 0x30,0xF2,0x00,0x06,0x29,0x1E,0x07,0x08,0x30,0xD0, \
-+ 0x00,0x52,0x00,0x08,0x28,0x1A,0x63,0x52,0x00,0x00, \
-+ 0x30,0xF2,0x1B,0x78,0x06,0x00,0x29,0x1E,0x30,0xF2, \
-+ 0x53,0x7E,0xFF,0xFF,0x1B,0x83,0x08,0x16,0x00,0x54, \
-+ 0x00,0x01,0x1B,0xFE,0x1B,0x90,0x32,0xF1,0x28,0x5D, \
-+ 0x32,0xF1,0x00,0x55,0x00,0x08,0x28,0x5F,0x00,0x00, \
-+ 0x8F,0x9F,0x29,0x33,0x08,0x16,0x00,0x49,0x00,0x01, \
-+ 0x1B,0xFF,0x00,0x01,0x1B,0xFF,0x08,0x07,0x00,0x02, \
-+ 0x00,0x00,0x8D,0x80,0x53,0x9C,0x00,0x01,0x30,0x71, \
-+ 0x00,0x55,0x00,0x01,0x28,0x0F,0x00,0x00,0x8D,0x00, \
-+ 0x53,0xA4,0x00,0x01,0x30,0x71,0x00,0x55,0x00,0x01, \
-+ 0x28,0x0F,0x00,0x00,0x83,0x8E,0x53,0xB9,0x00,0x00, \
-+ 0x00,0x00,0x86,0x08,0x30,0x71,0x00,0x7B,0x03,0xB9, \
-+ 0x33,0xB4,0x00,0xDA,0xFF,0xFF,0x00,0x0F,0x00,0x00, \
-+ 0x00,0x00,0x00,0x00,0x86,0x09,0x01,0x03,0x00,0x7D, \
-+ 0x03,0xB9,0x1B,0xC8,0x33,0xD1,0x00,0xF9,0x00,0x10, \
-+ 0x00,0x00,0x00,0x00,0x00,0x00,0x28,0x7B,0x09,0x5F, \
-+ 0x00,0x1A,0x00,0x00,0x09,0x4F,0x00,0x1A,0x00,0x00, \
-+ 0x00,0x01,0x1B,0xFF,0x00,0x00,0x8C,0x00,0x53,0xF0, \
-+ 0x00,0x01,0x34,0xF5,0x00,0xFB,0xFF,0xFF,0x00,0x7F, \
-+ 0x00,0x00,0x00,0x00,0x2A,0x9F,0x00,0x00,0x93,0x8F, \
-+ 0x28,0x49,0x00,0x00,0x97,0x8F,0x28,0x4B,0x34,0x61, \
-+ 0x28,0x4D,0x34,0x71,0x28,0x4F,0x34,0xB7,0x00,0xF9, \
-+ 0x00,0x00,0x01,0x00,0x00,0x00,0x00,0x00,0x2B,0x97, \
-+ 0x33,0xF1,0x00,0xF9,0x00,0x01,0x00,0x00,0x00,0x00, \
-+ 0x00,0x00,0x28,0x7F,0x00,0x03,0x00,0x02,0x00,0x00, \
-+ 0x00,0x01,0x1B,0xFF,0x00,0x01,0x1B,0xFF, \
-+}
-+#endif /* (DPAA_VERSION == 10) */
-+
-+/****************************/
-+/* Parser defines */
-+/****************************/
-+#define FM_PCD_PRS_SW_TAIL_SIZE 4 /**< Number of bytes that must be cleared at
-+ the end of the SW parser area */
-+
-+/* masks */
-+#define PRS_ERR_CAP 0x80000000
-+#define PRS_ERR_TYPE_DOUBLE 0x40000000
-+#define PRS_ERR_SINGLE_ECC_CNT_MASK 0x00FF0000
-+#define PRS_ERR_ADDR_MASK 0x000001FF
-+
-+/* others */
-+#define PRS_MAX_CYCLE_LIMIT 8191
-+#define PRS_SW_DATA 0x00000800
-+#define PRS_REGS_OFFSET 0x00000840
-+
-+#define GET_FM_PCD_PRS_PORT_ID(prsPortId,hardwarePortId) \
-+ prsPortId = (uint8_t)(hardwarePortId & 0x0f)
-+
-+#define GET_FM_PCD_INDEX_FLAG(bitMask, prsPortId) \
-+ bitMask = 0x80000000>>prsPortId
-+
-+#endif /* __FM_PRS_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/Pcd/fm_replic.c
-@@ -0,0 +1,984 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/******************************************************************************
-+ @File fm_replic.c
-+
-+ @Description FM frame replicator
-+*//***************************************************************************/
-+#include "std_ext.h"
-+#include "error_ext.h"
-+#include "string_ext.h"
-+#include "debug_ext.h"
-+#include "fm_pcd_ext.h"
-+#include "fm_muram_ext.h"
-+#include "fm_common.h"
-+#include "fm_hc.h"
-+#include "fm_replic.h"
-+#include "fm_cc.h"
-+#include "list_ext.h"
-+
-+
-+/****************************************/
-+/* static functions */
-+/****************************************/
-+static uint8_t GetMemberPosition(t_FmPcdFrmReplicGroup *p_ReplicGroup,
-+ uint32_t memberIndex,
-+ bool isAddOperation)
-+{
-+ uint8_t memberPosition;
-+ uint32_t lastMemberIndex;
-+
-+ ASSERT_COND(p_ReplicGroup);
-+
-+ /* the last member index is different between add and remove operation -
-+ in case of remove - this is exactly the last member index
-+ in case of add - this is the last member index + 1 - e.g.
-+ if we have 4 members, the index of the actual last member is 3(because the
-+ index starts from 0) therefore in order to add a new member as the last
-+ member we shall use memberIndex = 4 and not 3
-+ */
-+ if (isAddOperation)
-+ lastMemberIndex = p_ReplicGroup->numOfEntries;
-+ else
-+ lastMemberIndex = p_ReplicGroup->numOfEntries-1;
-+
-+ /* last */
-+ if (memberIndex == lastMemberIndex)
-+ memberPosition = FRM_REPLIC_LAST_MEMBER_INDEX;
-+ else
-+ {
-+ /* first */
-+ if (memberIndex == 0)
-+ memberPosition = FRM_REPLIC_FIRST_MEMBER_INDEX;
-+ else
-+ {
-+ /* middle */
-+ ASSERT_COND(memberIndex < lastMemberIndex);
-+ memberPosition = FRM_REPLIC_MIDDLE_MEMBER_INDEX;
-+ }
-+ }
-+ return memberPosition;
-+}
-+
-+static t_Error MemberCheckParams(t_Handle h_FmPcd,
-+ t_FmPcdCcNextEngineParams *p_MemberParams)
-+{
-+ t_Error err;
-+
-+
-+ if ((p_MemberParams->nextEngine != e_FM_PCD_DONE) &&
-+ (p_MemberParams->nextEngine != e_FM_PCD_KG) &&
-+ (p_MemberParams->nextEngine != e_FM_PCD_PLCR))
-+ RETURN_ERROR(MAJOR, E_NOT_SUPPORTED, ("Next engine of a member should be MatchTable(cc) or Done or Policer"));
-+
-+ /* check the regular parameters of the next engine */
-+ err = ValidateNextEngineParams(h_FmPcd, p_MemberParams, e_FM_PCD_CC_STATS_MODE_NONE);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, ("member next engine parameters"));
-+
-+ return E_OK;
-+}
-+
-+static t_Error CheckParams(t_Handle h_FmPcd,
-+ t_FmPcdFrmReplicGroupParams *p_ReplicGroupParam)
-+{
-+ int i;
-+ t_Error err;
-+
-+ /* check that max num of entries is at least 2 */
-+ if (!IN_RANGE(2, p_ReplicGroupParam->maxNumOfEntries, FM_PCD_FRM_REPLIC_MAX_NUM_OF_ENTRIES))
-+ RETURN_ERROR(MAJOR, E_NOT_IN_RANGE, ("maxNumOfEntries in the frame replicator parameters should be 2-%d",FM_PCD_FRM_REPLIC_MAX_NUM_OF_ENTRIES));
-+
-+ /* check that number of entries is greater than zero */
-+ if (!p_ReplicGroupParam->numOfEntries)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("numOFEntries in the frame replicator group should be greater than zero"));
-+
-+ /* check that max num of entries is equal or greater than number of entries */
-+ if (p_ReplicGroupParam->maxNumOfEntries < p_ReplicGroupParam->numOfEntries)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("maxNumOfEntries should be equal or greater than numOfEntries"));
-+
-+ for (i=0; i<p_ReplicGroupParam->numOfEntries; i++)
-+ {
-+ err = MemberCheckParams(h_FmPcd, &p_ReplicGroupParam->nextEngineParams[i]);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, ("member check parameters"));
-+ }
-+ return E_OK;
-+}
-+
-+static t_FmPcdFrmReplicMember *GetAvailableMember(t_FmPcdFrmReplicGroup *p_ReplicGroup)
-+{
-+ t_FmPcdFrmReplicMember *p_ReplicMember = NULL;
-+ t_List *p_Next;
-+
-+ if (!LIST_IsEmpty(&p_ReplicGroup->availableMembersList))
-+ {
-+ p_Next = LIST_FIRST(&p_ReplicGroup->availableMembersList);
-+ p_ReplicMember = LIST_OBJECT(p_Next, t_FmPcdFrmReplicMember, node);
-+ ASSERT_COND(p_ReplicMember);
-+ LIST_DelAndInit(p_Next);
-+ }
-+ return p_ReplicMember;
-+}
-+
-+static void PutAvailableMember(t_FmPcdFrmReplicGroup *p_ReplicGroup,
-+ t_FmPcdFrmReplicMember *p_ReplicMember)
-+{
-+ LIST_AddToTail(&p_ReplicMember->node, &p_ReplicGroup->availableMembersList);
-+}
-+
-+static void AddMemberToList(t_FmPcdFrmReplicGroup *p_ReplicGroup,
-+ t_FmPcdFrmReplicMember *p_CurrentMember,
-+ t_List *p_ListHead)
-+{
-+ LIST_Add(&p_CurrentMember->node, p_ListHead);
-+
-+ p_ReplicGroup->numOfEntries++;
-+}
-+
-+static void RemoveMemberFromList(t_FmPcdFrmReplicGroup *p_ReplicGroup,
-+ t_FmPcdFrmReplicMember *p_CurrentMember)
-+{
-+ ASSERT_COND(p_ReplicGroup->numOfEntries);
-+ LIST_DelAndInit(&p_CurrentMember->node);
-+ p_ReplicGroup->numOfEntries--;
-+}
-+
-+static void LinkSourceToMember(t_FmPcdFrmReplicGroup *p_ReplicGroup,
-+ t_AdOfTypeContLookup *p_SourceTd,
-+ t_FmPcdFrmReplicMember *p_ReplicMember)
-+{
-+ t_FmPcd *p_FmPcd;
-+
-+ ASSERT_COND(p_SourceTd);
-+ ASSERT_COND(p_ReplicMember);
-+ ASSERT_COND(p_ReplicGroup);
-+ ASSERT_COND(p_ReplicGroup->h_FmPcd);
-+
-+ /* Link the first member in the group to the source TD */
-+ p_FmPcd = p_ReplicGroup->h_FmPcd;
-+
-+ WRITE_UINT32(p_SourceTd->matchTblPtr,
-+ (uint32_t)(XX_VirtToPhys(p_ReplicMember->p_MemberAd) -
-+ p_FmPcd->physicalMuramBase));
-+}
-+
-+static void LinkMemberToMember(t_FmPcdFrmReplicGroup *p_ReplicGroup,
-+ t_FmPcdFrmReplicMember *p_CurrentMember,
-+ t_FmPcdFrmReplicMember *p_NextMember)
-+{
-+ t_AdOfTypeResult *p_CurrReplicAd = (t_AdOfTypeResult*)p_CurrentMember->p_MemberAd;
-+ t_AdOfTypeResult *p_NextReplicAd = NULL;
-+ t_FmPcd *p_FmPcd;
-+ uint32_t offset = 0;
-+
-+ /* Check if the next member exists or it's NULL (- means that this is the last member) */
-+ if (p_NextMember)
-+ {
-+ p_NextReplicAd = (t_AdOfTypeResult*)p_NextMember->p_MemberAd;
-+ p_FmPcd = p_ReplicGroup->h_FmPcd;
-+ offset = (XX_VirtToPhys(p_NextReplicAd) - (p_FmPcd->physicalMuramBase));
-+ offset = ((offset>>NEXT_FRM_REPLIC_ADDR_SHIFT)<< NEXT_FRM_REPLIC_MEMBER_INDEX_SHIFT);
-+ }
-+
-+ /* link the current AD to point to the AD of the next member */
-+ WRITE_UINT32(p_CurrReplicAd->res, offset);
-+}
-+
-+static t_Error ModifyDescriptor(t_FmPcdFrmReplicGroup *p_ReplicGroup,
-+ void *p_OldDescriptor,
-+ void *p_NewDescriptor)
-+{
-+ t_Handle h_Hc;
-+ t_Error err;
-+ t_FmPcd *p_FmPcd;
-+
-+ ASSERT_COND(p_ReplicGroup);
-+ ASSERT_COND(p_ReplicGroup->h_FmPcd);
-+ ASSERT_COND(p_OldDescriptor);
-+ ASSERT_COND(p_NewDescriptor);
-+
-+ p_FmPcd = p_ReplicGroup->h_FmPcd;
-+ h_Hc = FmPcdGetHcHandle(p_FmPcd);
-+ if (!h_Hc)
-+ RETURN_ERROR(MAJOR, E_INVALID_HANDLE, ("Host command"));
-+
-+ err = FmHcPcdCcDoDynamicChange(h_Hc,
-+ (uint32_t)(XX_VirtToPhys(p_OldDescriptor) - p_FmPcd->physicalMuramBase),
-+ (uint32_t)(XX_VirtToPhys(p_NewDescriptor) - p_FmPcd->physicalMuramBase));
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, ("Dynamic change host command"));
-+
-+ return E_OK;
-+}
-+
-+static void FillReplicAdOfTypeResult(void *p_ReplicAd, bool last)
-+{
-+ t_AdOfTypeResult *p_CurrReplicAd = (t_AdOfTypeResult*)p_ReplicAd;
-+ uint32_t tmp;
-+
-+ tmp = GET_UINT32(p_CurrReplicAd->plcrProfile);
-+ if (last)
-+ /* clear the NL bit in case it's the last member in the group*/
-+ WRITE_UINT32(p_CurrReplicAd->plcrProfile,(tmp & ~FRM_REPLIC_NL_BIT));
-+ else
-+ /* set the NL bit in case it's not the last member in the group */
-+ WRITE_UINT32(p_CurrReplicAd->plcrProfile, (tmp |FRM_REPLIC_NL_BIT));
-+
-+ /* set FR bit in the action descriptor */
-+ tmp = GET_UINT32(p_CurrReplicAd->nia);
-+ WRITE_UINT32(p_CurrReplicAd->nia,
-+ (tmp | FRM_REPLIC_FR_BIT | FM_PCD_AD_RESULT_EXTENDED_MODE ));
-+}
-+
-+static void BuildSourceTd(void *p_Ad)
-+{
-+ t_AdOfTypeContLookup *p_SourceTd;
-+
-+ ASSERT_COND(p_Ad);
-+
-+ p_SourceTd = (t_AdOfTypeContLookup *)p_Ad;
-+
-+ IOMemSet32((uint8_t*)p_SourceTd, 0, FM_PCD_CC_AD_ENTRY_SIZE);
-+
-+ /* initialize the source table descriptor */
-+ WRITE_UINT32(p_SourceTd->ccAdBase, FM_PCD_AD_CONT_LOOKUP_TYPE);
-+ WRITE_UINT32(p_SourceTd->pcAndOffsets, FRM_REPLIC_SOURCE_TD_OPCODE);
-+}
-+
-+static t_Error BuildShadowAndModifyDescriptor(t_FmPcdFrmReplicGroup *p_ReplicGroup,
-+ t_FmPcdFrmReplicMember *p_NextMember,
-+ t_FmPcdFrmReplicMember *p_CurrentMember,
-+ bool sourceDescriptor,
-+ bool last)
-+{
-+ t_FmPcd *p_FmPcd;
-+ t_FmPcdFrmReplicMember shadowMember;
-+ t_Error err;
-+
-+ ASSERT_COND(p_ReplicGroup);
-+ ASSERT_COND(p_ReplicGroup->h_FmPcd);
-+
-+ p_FmPcd = p_ReplicGroup->h_FmPcd;
-+ ASSERT_COND(p_FmPcd->p_CcShadow);
-+
-+ if (!TRY_LOCK(p_FmPcd->h_ShadowSpinlock, &p_FmPcd->shadowLock))
-+ return ERROR_CODE(E_BUSY);
-+
-+ if (sourceDescriptor)
-+ {
-+ BuildSourceTd(p_FmPcd->p_CcShadow);
-+ LinkSourceToMember(p_ReplicGroup, p_FmPcd->p_CcShadow, p_NextMember);
-+
-+ /* Modify the source table descriptor according to the prepared shadow descriptor */
-+ err = ModifyDescriptor(p_ReplicGroup,
-+ p_ReplicGroup->p_SourceTd,
-+ p_FmPcd->p_CcShadow/* new prepared source td */);
-+
-+ RELEASE_LOCK(p_FmPcd->shadowLock);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, ("Modify source Descriptor in BuildShadowAndModifyDescriptor"));
-+
-+ }
-+ else
-+ {
-+ IO2IOCpy32(p_FmPcd->p_CcShadow,
-+ p_CurrentMember->p_MemberAd,
-+ FM_PCD_CC_AD_ENTRY_SIZE);
-+
-+ /* update the last bit in the shadow ad */
-+ FillReplicAdOfTypeResult(p_FmPcd->p_CcShadow, last);
-+
-+ shadowMember.p_MemberAd = p_FmPcd->p_CcShadow;
-+
-+ /* update the next FR member index */
-+ LinkMemberToMember(p_ReplicGroup, &shadowMember, p_NextMember);
-+
-+ /* Modify the next member according to the prepared shadow descriptor */
-+ err = ModifyDescriptor(p_ReplicGroup,
-+ p_CurrentMember->p_MemberAd,
-+ p_FmPcd->p_CcShadow);
-+
-+ RELEASE_LOCK(p_FmPcd->shadowLock);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, ("Modify Descriptor in BuildShadowAndModifyDescriptor"));
-+ }
-+
-+
-+ return E_OK;
-+}
-+
-+static t_FmPcdFrmReplicMember* GetMemberByIndex(t_FmPcdFrmReplicGroup *p_ReplicGroup,
-+ uint16_t memberIndex)
-+{
-+ int i=0;
-+ t_List *p_Pos;
-+ t_FmPcdFrmReplicMember *p_Member = NULL;
-+
-+ LIST_FOR_EACH(p_Pos, &p_ReplicGroup->membersList)
-+ {
-+ if (i == memberIndex)
-+ {
-+ p_Member = LIST_OBJECT(p_Pos, t_FmPcdFrmReplicMember, node);
-+ return p_Member;
-+ }
-+ i++;
-+ }
-+ return p_Member;
-+}
-+
-+static t_Error AllocMember(t_FmPcdFrmReplicGroup *p_ReplicGroup)
-+{
-+ t_FmPcdFrmReplicMember *p_CurrentMember;
-+ t_Handle h_Muram;
-+
-+ ASSERT_COND(p_ReplicGroup);
-+
-+ h_Muram = FmPcdGetMuramHandle(p_ReplicGroup->h_FmPcd);
-+ ASSERT_COND(h_Muram);
-+
-+ /* Initialize an internal structure of a member to add to the available members list */
-+ p_CurrentMember = (t_FmPcdFrmReplicMember *)XX_Malloc(sizeof(t_FmPcdFrmReplicMember));
-+ if (!p_CurrentMember)
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, ("Frame replicator member"));
-+
-+ memset(p_CurrentMember, 0 ,sizeof(t_FmPcdFrmReplicMember));
-+
-+ /* Allocate the member AD */
-+ p_CurrentMember->p_MemberAd =
-+ (t_AdOfTypeResult*)FM_MURAM_AllocMem(h_Muram,
-+ FM_PCD_CC_AD_ENTRY_SIZE,
-+ FM_PCD_CC_AD_TABLE_ALIGN);
-+ if (!p_CurrentMember->p_MemberAd)
-+ {
-+ XX_Free(p_CurrentMember);
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, ("member AD table"));
-+ }
-+ IOMemSet32((uint8_t*)p_CurrentMember->p_MemberAd, 0, FM_PCD_CC_AD_ENTRY_SIZE);
-+
-+ /* Add the new member to the available members list */
-+ LIST_AddToTail(&p_CurrentMember->node, &(p_ReplicGroup->availableMembersList));
-+
-+ return E_OK;
-+}
-+
-+static t_FmPcdFrmReplicMember* InitMember(t_FmPcdFrmReplicGroup *p_ReplicGroup,
-+ t_FmPcdCcNextEngineParams *p_MemberParams,
-+ bool last)
-+{
-+ t_FmPcdFrmReplicMember *p_CurrentMember = NULL;
-+
-+ ASSERT_COND(p_ReplicGroup);
-+
-+ /* Get an available member from the internal members list */
-+ p_CurrentMember = GetAvailableMember(p_ReplicGroup);
-+ if (!p_CurrentMember)
-+ {
-+ REPORT_ERROR(MAJOR, E_NOT_FOUND, ("Available member"));
-+ return NULL;
-+ }
-+ p_CurrentMember->h_Manip = NULL;
-+
-+ /* clear the Ad of the new member */
-+ IOMemSet32((uint8_t*)p_CurrentMember->p_MemberAd, 0, FM_PCD_CC_AD_ENTRY_SIZE);
-+
-+ INIT_LIST(&p_CurrentMember->node);
-+
-+ /* Initialize the Ad of the member */
-+ NextStepAd(p_CurrentMember->p_MemberAd,
-+ NULL,
-+ p_MemberParams,
-+ p_ReplicGroup->h_FmPcd);
-+
-+ /* save Manip handle (for free needs) */
-+ if (p_MemberParams->h_Manip)
-+ p_CurrentMember->h_Manip = p_MemberParams->h_Manip;
-+
-+ /* Initialize the relevant frame replicator fields in the AD */
-+ FillReplicAdOfTypeResult(p_CurrentMember->p_MemberAd, last);
-+
-+ return p_CurrentMember;
-+}
-+
-+static void FreeMember(t_FmPcdFrmReplicGroup *p_ReplicGroup,
-+ t_FmPcdFrmReplicMember *p_Member)
-+{
-+ /* Note: Can't free the member AD just returns the member to the available
-+ member list - therefore only memset the AD */
-+
-+ /* zero the AD */
-+ IOMemSet32(p_Member->p_MemberAd, 0, FM_PCD_CC_AD_ENTRY_SIZE);
-+
-+
-+ /* return the member to the available members list */
-+ PutAvailableMember(p_ReplicGroup, p_Member);
-+}
-+
-+static t_Error RemoveMember(t_FmPcdFrmReplicGroup *p_ReplicGroup,
-+ uint16_t memberIndex)
-+{
-+ t_FmPcd *p_FmPcd = NULL;
-+ t_FmPcdFrmReplicMember *p_CurrentMember = NULL, *p_PreviousMember = NULL, *p_NextMember = NULL;
-+ t_Error err;
-+ uint8_t memberPosition;
-+
-+ p_FmPcd = p_ReplicGroup->h_FmPcd;
-+ ASSERT_COND(p_FmPcd);
-+ UNUSED(p_FmPcd);
-+
-+ p_CurrentMember = GetMemberByIndex(p_ReplicGroup, memberIndex);
-+ ASSERT_COND(p_CurrentMember);
-+
-+ /* determine the member position in the group */
-+ memberPosition = GetMemberPosition(p_ReplicGroup,
-+ memberIndex,
-+ FALSE/*remove operation*/);
-+
-+ switch (memberPosition)
-+ {
-+ case FRM_REPLIC_FIRST_MEMBER_INDEX:
-+ p_NextMember = GetMemberByIndex(p_ReplicGroup, (uint16_t)(memberIndex+1));
-+ ASSERT_COND(p_NextMember);
-+
-+ /* update the source td itself by using a host command */
-+ err = BuildShadowAndModifyDescriptor(p_ReplicGroup,
-+ p_NextMember,
-+ NULL,
-+ TRUE/*sourceDescriptor*/,
-+ FALSE/*last*/);
-+ break;
-+
-+ case FRM_REPLIC_MIDDLE_MEMBER_INDEX:
-+ p_PreviousMember = GetMemberByIndex(p_ReplicGroup, (uint16_t)(memberIndex-1));
-+ ASSERT_COND(p_PreviousMember);
-+
-+ p_NextMember = GetMemberByIndex(p_ReplicGroup, (uint16_t)(memberIndex+1));
-+ ASSERT_COND(p_NextMember);
-+
-+ err = BuildShadowAndModifyDescriptor(p_ReplicGroup,
-+ p_NextMember,
-+ p_PreviousMember,
-+ FALSE/*sourceDescriptor*/,
-+ FALSE/*last*/);
-+
-+ break;
-+
-+ case FRM_REPLIC_LAST_MEMBER_INDEX:
-+ p_PreviousMember = GetMemberByIndex(p_ReplicGroup, (uint16_t)(memberIndex-1));
-+ ASSERT_COND(p_PreviousMember);
-+
-+ err = BuildShadowAndModifyDescriptor(p_ReplicGroup,
-+ NULL,
-+ p_PreviousMember,
-+ FALSE/*sourceDescriptor*/,
-+ TRUE/*last*/);
-+ break;
-+
-+ default:
-+ RETURN_ERROR(MAJOR, E_INVALID_SELECTION, ("member position in remove member"));
-+ }
-+
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+
-+ if (p_CurrentMember->h_Manip)
-+ {
-+ FmPcdManipUpdateOwner(p_CurrentMember->h_Manip, FALSE);
-+ p_CurrentMember->h_Manip = NULL;
-+ }
-+
-+ /* remove the member from the driver internal members list */
-+ RemoveMemberFromList(p_ReplicGroup, p_CurrentMember);
-+
-+ /* return the member to the available members list */
-+ FreeMember(p_ReplicGroup, p_CurrentMember);
-+
-+ return E_OK;
-+}
-+
-+static void DeleteGroup(t_FmPcdFrmReplicGroup *p_ReplicGroup)
-+{
-+ int i, j;
-+ t_Handle h_Muram;
-+ t_FmPcdFrmReplicMember *p_Member, *p_CurrentMember;
-+
-+ if (p_ReplicGroup)
-+ {
-+ ASSERT_COND(p_ReplicGroup->h_FmPcd);
-+ h_Muram = FmPcdGetMuramHandle(p_ReplicGroup->h_FmPcd);
-+ ASSERT_COND(h_Muram);
-+
-+ /* free the source table descriptor */
-+ if (p_ReplicGroup->p_SourceTd)
-+ {
-+ FM_MURAM_FreeMem(h_Muram, p_ReplicGroup->p_SourceTd);
-+ p_ReplicGroup->p_SourceTd = NULL;
-+ }
-+
-+ /* Remove all members from the members linked list (hw and sw) and
-+ return the members to the available members list */
-+ if (p_ReplicGroup->numOfEntries)
-+ {
-+ j = p_ReplicGroup->numOfEntries-1;
-+
-+ /* manually removal of the member because there are no owners of
-+ this group */
-+ for (i=j; i>=0; i--)
-+ {
-+ p_CurrentMember = GetMemberByIndex(p_ReplicGroup, (uint16_t)i/*memberIndex*/);
-+ ASSERT_COND(p_CurrentMember);
-+
-+ if (p_CurrentMember->h_Manip)
-+ {
-+ FmPcdManipUpdateOwner(p_CurrentMember->h_Manip, FALSE);
-+ p_CurrentMember->h_Manip = NULL;
-+ }
-+
-+ /* remove the member from the internal driver members list */
-+ RemoveMemberFromList(p_ReplicGroup, p_CurrentMember);
-+
-+ /* return the member to the available members list */
-+ FreeMember(p_ReplicGroup, p_CurrentMember);
-+ }
-+ }
-+
-+ /* Free members AD */
-+ for (i=0; i<p_ReplicGroup->maxNumOfEntries; i++)
-+ {
-+ p_Member = GetAvailableMember(p_ReplicGroup);
-+ ASSERT_COND(p_Member);
-+ if (p_Member->p_MemberAd)
-+ {
-+ FM_MURAM_FreeMem(h_Muram, p_Member->p_MemberAd);
-+ p_Member->p_MemberAd = NULL;
-+ }
-+ XX_Free(p_Member);
-+ }
-+
-+ /* release the group lock */
-+ if (p_ReplicGroup->p_Lock)
-+ FmPcdReleaseLock(p_ReplicGroup->h_FmPcd, p_ReplicGroup->p_Lock);
-+
-+ /* free the replicator group */
-+ XX_Free(p_ReplicGroup);
-+ }
-+}
-+
-+
-+/*****************************************************************************/
-+/* Inter-module API routines */
-+/*****************************************************************************/
-+
-+/* NOTE: the inter-module routines are locked by cc in case of using them */
-+void * FrmReplicGroupGetSourceTableDescriptor(t_Handle h_ReplicGroup)
-+{
-+ t_FmPcdFrmReplicGroup *p_ReplicGroup = (t_FmPcdFrmReplicGroup *)h_ReplicGroup;
-+ ASSERT_COND(p_ReplicGroup);
-+
-+ return (p_ReplicGroup->p_SourceTd);
-+}
-+
-+void FrmReplicGroupUpdateAd(t_Handle h_ReplicGroup,
-+ void *p_Ad,
-+ t_Handle *h_AdNew)
-+{
-+ t_FmPcdFrmReplicGroup *p_ReplicGroup = (t_FmPcdFrmReplicGroup *)h_ReplicGroup;
-+ t_AdOfTypeResult *p_AdResult = (t_AdOfTypeResult*)p_Ad;
-+ t_FmPcd *p_FmPcd;
-+
-+ ASSERT_COND(p_ReplicGroup);
-+ p_FmPcd = p_ReplicGroup->h_FmPcd;
-+
-+ /* build a bypass ad */
-+ WRITE_UINT32(p_AdResult->fqid, FM_PCD_AD_BYPASS_TYPE |
-+ (uint32_t)((XX_VirtToPhys(p_ReplicGroup->p_SourceTd)) - p_FmPcd->physicalMuramBase));
-+
-+ *h_AdNew = NULL;
-+}
-+
-+void FrmReplicGroupUpdateOwner(t_Handle h_ReplicGroup,
-+ bool add)
-+{
-+ t_FmPcdFrmReplicGroup *p_ReplicGroup = (t_FmPcdFrmReplicGroup *)h_ReplicGroup;
-+ ASSERT_COND(p_ReplicGroup);
-+
-+ /* update the group owner counter */
-+ if (add)
-+ p_ReplicGroup->owners++;
-+ else
-+ {
-+ ASSERT_COND(p_ReplicGroup->owners);
-+ p_ReplicGroup->owners--;
-+ }
-+}
-+
-+t_Error FrmReplicGroupTryLock(t_Handle h_ReplicGroup)
-+{
-+ t_FmPcdFrmReplicGroup *p_ReplicGroup = (t_FmPcdFrmReplicGroup *)h_ReplicGroup;
-+
-+ ASSERT_COND(h_ReplicGroup);
-+
-+ if (FmPcdLockTryLock(p_ReplicGroup->p_Lock))
-+ return E_OK;
-+
-+ return ERROR_CODE(E_BUSY);
-+}
-+
-+void FrmReplicGroupUnlock(t_Handle h_ReplicGroup)
-+{
-+ t_FmPcdFrmReplicGroup *p_ReplicGroup = (t_FmPcdFrmReplicGroup *)h_ReplicGroup;
-+
-+ ASSERT_COND(h_ReplicGroup);
-+
-+ FmPcdLockUnlock(p_ReplicGroup->p_Lock);
-+}
-+/*********************** End of inter-module routines ************************/
-+
-+
-+/****************************************/
-+/* API Init unit functions */
-+/****************************************/
-+t_Handle FM_PCD_FrmReplicSetGroup(t_Handle h_FmPcd,
-+ t_FmPcdFrmReplicGroupParams *p_ReplicGroupParam)
-+{
-+ t_FmPcdFrmReplicGroup *p_ReplicGroup;
-+ t_FmPcdFrmReplicMember *p_CurrentMember, *p_NextMember = NULL;
-+ int i;
-+ t_Error err;
-+ bool last = FALSE;
-+ t_Handle h_Muram;
-+
-+ SANITY_CHECK_RETURN_VALUE(h_FmPcd, E_INVALID_HANDLE, NULL);
-+ SANITY_CHECK_RETURN_VALUE(p_ReplicGroupParam, E_INVALID_HANDLE, NULL);
-+
-+ if (!FmPcdIsAdvancedOffloadSupported(h_FmPcd))
-+ {
-+ REPORT_ERROR(MAJOR, E_INVALID_STATE, ("Advanced-offload must be enabled"));
-+ return NULL;
-+ }
-+
-+ err = CheckParams(h_FmPcd, p_ReplicGroupParam);
-+ if (err)
-+ {
-+ REPORT_ERROR(MAJOR, err, (NO_MSG));
-+ return NULL;
-+ }
-+
-+ p_ReplicGroup = (t_FmPcdFrmReplicGroup*)XX_Malloc(sizeof(t_FmPcdFrmReplicGroup));
-+ if (!p_ReplicGroup)
-+ {
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("No memory"));
-+ return NULL;
-+ }
-+ memset(p_ReplicGroup, 0, sizeof(t_FmPcdFrmReplicGroup));
-+
-+ /* initialize lists for internal driver use */
-+ INIT_LIST(&p_ReplicGroup->availableMembersList);
-+ INIT_LIST(&p_ReplicGroup->membersList);
-+
-+ p_ReplicGroup->h_FmPcd = h_FmPcd;
-+
-+ h_Muram = FmPcdGetMuramHandle(p_ReplicGroup->h_FmPcd);
-+ ASSERT_COND(h_Muram);
-+
-+ /* initialize the group lock */
-+ p_ReplicGroup->p_Lock = FmPcdAcquireLock(p_ReplicGroup->h_FmPcd);
-+ if (!p_ReplicGroup->p_Lock)
-+ {
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("Replic group lock"));
-+ DeleteGroup(p_ReplicGroup);
-+ return NULL;
-+ }
-+
-+ /* Allocate the frame replicator source table descriptor */
-+ p_ReplicGroup->p_SourceTd =
-+ (t_Handle)FM_MURAM_AllocMem(h_Muram,
-+ FM_PCD_CC_AD_ENTRY_SIZE,
-+ FM_PCD_CC_AD_TABLE_ALIGN);
-+ if (!p_ReplicGroup->p_SourceTd)
-+ {
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("frame replicator source table descriptor"));
-+ DeleteGroup(p_ReplicGroup);
-+ return NULL;
-+ }
-+
-+ /* update the shadow size - required for the host commands */
-+ err = FmPcdUpdateCcShadow(p_ReplicGroup->h_FmPcd,
-+ FM_PCD_CC_AD_ENTRY_SIZE,
-+ FM_PCD_CC_AD_TABLE_ALIGN);
-+ if (err)
-+ {
-+ REPORT_ERROR(MAJOR, err, ("Update CC shadow"));
-+ DeleteGroup(p_ReplicGroup);
-+ return NULL;
-+ }
-+
-+ p_ReplicGroup->maxNumOfEntries = p_ReplicGroupParam->maxNumOfEntries;
-+
-+ /* Allocate the maximal number of members ADs and Statistics AD for the group
-+ It prevents allocation of Muram in run-time */
-+ for (i=0; i<p_ReplicGroup->maxNumOfEntries; i++)
-+ {
-+ err = AllocMember(p_ReplicGroup);
-+ if (err)
-+ {
-+ REPORT_ERROR(MAJOR, err, ("allocate a new member"));
-+ DeleteGroup(p_ReplicGroup);
-+ return NULL;
-+ }
-+ }
-+
-+ /* Initialize the members linked lists:
-+ (hw - the one that is used by the FMan controller and
-+ sw - the one that is managed by the driver internally) */
-+ for (i=(p_ReplicGroupParam->numOfEntries-1); i>=0; i--)
-+ {
-+ /* check if this is the last member in the group */
-+ if (i == (p_ReplicGroupParam->numOfEntries-1))
-+ last = TRUE;
-+ else
-+ last = FALSE;
-+
-+ /* Initialize a new member */
-+ p_CurrentMember = InitMember(p_ReplicGroup,
-+ &(p_ReplicGroupParam->nextEngineParams[i]),
-+ last);
-+ if (!p_CurrentMember)
-+ {
-+ REPORT_ERROR(MAJOR, E_INVALID_HANDLE, ("No available member"));
-+ DeleteGroup(p_ReplicGroup);
-+ return NULL;
-+ }
-+
-+ /* Build the members group - link two consecutive members in the hw linked list */
-+ LinkMemberToMember(p_ReplicGroup, p_CurrentMember, p_NextMember);
-+
-+ /* update the driver internal members list to be compatible to the hw members linked list */
-+ AddMemberToList(p_ReplicGroup, p_CurrentMember, &p_ReplicGroup->membersList);
-+
-+ p_NextMember = p_CurrentMember;
-+ }
-+
-+ /* initialize the source table descriptor */
-+ BuildSourceTd(p_ReplicGroup->p_SourceTd);
-+
-+ /* link the source table descriptor to point to the first member in the group */
-+ LinkSourceToMember(p_ReplicGroup, p_ReplicGroup->p_SourceTd, p_NextMember);
-+
-+ return p_ReplicGroup;
-+}
-+
-+t_Error FM_PCD_FrmReplicDeleteGroup(t_Handle h_ReplicGroup)
-+{
-+ t_FmPcdFrmReplicGroup *p_ReplicGroup = (t_FmPcdFrmReplicGroup *)h_ReplicGroup;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_ReplicGroup, E_INVALID_HANDLE);
-+
-+ if (p_ReplicGroup->owners)
-+ RETURN_ERROR(MAJOR,
-+ E_INVALID_STATE,
-+ ("the group has owners and can't be deleted"));
-+
-+ DeleteGroup(p_ReplicGroup);
-+
-+ return E_OK;
-+}
-+
-+
-+/*****************************************************************************/
-+/* API Run-time Frame replicator Control unit functions */
-+/*****************************************************************************/
-+t_Error FM_PCD_FrmReplicAddMember(t_Handle h_ReplicGroup,
-+ uint16_t memberIndex,
-+ t_FmPcdCcNextEngineParams *p_MemberParams)
-+{
-+ t_FmPcdFrmReplicGroup *p_ReplicGroup = (t_FmPcdFrmReplicGroup*) h_ReplicGroup;
-+ t_FmPcdFrmReplicMember *p_NewMember, *p_CurrentMember = NULL, *p_PreviousMember = NULL;
-+ t_Error err;
-+ uint8_t memberPosition;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_ReplicGroup, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_MemberParams, E_INVALID_HANDLE);
-+
-+ /* group lock */
-+ err = FrmReplicGroupTryLock(p_ReplicGroup);
-+ if (GET_ERROR_TYPE(err) == E_BUSY)
-+ return ERROR_CODE(E_BUSY);
-+
-+ if (memberIndex > p_ReplicGroup->numOfEntries)
-+ {
-+ /* unlock */
-+ FrmReplicGroupUnlock(p_ReplicGroup);
-+ RETURN_ERROR(MAJOR, E_INVALID_SELECTION,
-+ ("memberIndex is greater than the members in the list"));
-+ }
-+
-+ if (memberIndex >= p_ReplicGroup->maxNumOfEntries)
-+ {
-+ /* unlock */
-+ FrmReplicGroupUnlock(p_ReplicGroup);
-+ RETURN_ERROR(MAJOR, E_INVALID_SELECTION, ("memberIndex is greater than the allowed number of members in the group"));
-+ }
-+
-+ if ((p_ReplicGroup->numOfEntries + 1) > FM_PCD_FRM_REPLIC_MAX_NUM_OF_ENTRIES)
-+ {
-+ /* unlock */
-+ FrmReplicGroupUnlock(p_ReplicGroup);
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE,
-+ ("numOfEntries with new entry can not be larger than %d\n",
-+ FM_PCD_FRM_REPLIC_MAX_NUM_OF_ENTRIES));
-+ }
-+
-+ err = MemberCheckParams(p_ReplicGroup->h_FmPcd, p_MemberParams);
-+ if (err)
-+ {
-+ /* unlock */
-+ FrmReplicGroupUnlock(p_ReplicGroup);
-+ RETURN_ERROR(MAJOR, err, ("member check parameters in add operation"));
-+ }
-+ /* determine the member position in the group */
-+ memberPosition = GetMemberPosition(p_ReplicGroup,
-+ memberIndex,
-+ TRUE/* add operation */);
-+
-+ /* Initialize a new member */
-+ p_NewMember = InitMember(p_ReplicGroup,
-+ p_MemberParams,
-+ (memberPosition == FRM_REPLIC_LAST_MEMBER_INDEX ? TRUE : FALSE));
-+ if (!p_NewMember)
-+ {
-+ /* unlock */
-+ FrmReplicGroupUnlock(p_ReplicGroup);
-+ RETURN_ERROR(MAJOR, E_INVALID_HANDLE, ("No available member"));
-+ }
-+
-+ switch (memberPosition)
-+ {
-+ case FRM_REPLIC_FIRST_MEMBER_INDEX:
-+ p_CurrentMember = GetMemberByIndex(p_ReplicGroup, memberIndex);
-+ ASSERT_COND(p_CurrentMember);
-+
-+ LinkMemberToMember(p_ReplicGroup, p_NewMember, p_CurrentMember);
-+
-+ /* update the internal group source TD */
-+ LinkSourceToMember(p_ReplicGroup,
-+ p_ReplicGroup->p_SourceTd,
-+ p_NewMember);
-+
-+ /* add member to the internal sw member list */
-+ AddMemberToList(p_ReplicGroup,
-+ p_NewMember,
-+ &p_ReplicGroup->membersList);
-+ break;
-+
-+ case FRM_REPLIC_MIDDLE_MEMBER_INDEX:
-+ p_CurrentMember = GetMemberByIndex(p_ReplicGroup, memberIndex);
-+ ASSERT_COND(p_CurrentMember);
-+
-+ p_PreviousMember = GetMemberByIndex(p_ReplicGroup, (uint16_t)(memberIndex-1));
-+ ASSERT_COND(p_PreviousMember);
-+
-+ LinkMemberToMember(p_ReplicGroup, p_NewMember, p_CurrentMember);
-+ LinkMemberToMember(p_ReplicGroup, p_PreviousMember, p_NewMember);
-+
-+ AddMemberToList(p_ReplicGroup, p_NewMember, &p_PreviousMember->node);
-+ break;
-+
-+ case FRM_REPLIC_LAST_MEMBER_INDEX:
-+ p_PreviousMember = GetMemberByIndex(p_ReplicGroup, (uint16_t)(memberIndex-1));
-+ ASSERT_COND(p_PreviousMember);
-+
-+ LinkMemberToMember(p_ReplicGroup, p_PreviousMember, p_NewMember);
-+ FillReplicAdOfTypeResult(p_PreviousMember->p_MemberAd, FALSE/*last*/);
-+
-+ /* add the new member to the internal sw member list */
-+ AddMemberToList(p_ReplicGroup, p_NewMember, &p_PreviousMember->node);
-+ break;
-+
-+ default:
-+ /* unlock */
-+ FrmReplicGroupUnlock(p_ReplicGroup);
-+ RETURN_ERROR(MAJOR, E_INVALID_SELECTION, ("member position in add member"));
-+
-+ }
-+
-+ /* unlock */
-+ FrmReplicGroupUnlock(p_ReplicGroup);
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PCD_FrmReplicRemoveMember(t_Handle h_ReplicGroup,
-+ uint16_t memberIndex)
-+{
-+ t_FmPcdFrmReplicGroup *p_ReplicGroup = (t_FmPcdFrmReplicGroup*) h_ReplicGroup;
-+ t_Error err;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_ReplicGroup, E_INVALID_HANDLE);
-+
-+ /* lock */
-+ err = FrmReplicGroupTryLock(p_ReplicGroup);
-+ if (GET_ERROR_TYPE(err) == E_BUSY)
-+ return ERROR_CODE(E_BUSY);
-+
-+ if (memberIndex >= p_ReplicGroup->numOfEntries)
-+ RETURN_ERROR(MAJOR, E_INVALID_SELECTION, ("member index to remove"));
-+
-+ /* Design decision: group must contain at least one member
-+ No possibility to remove the last member from the group */
-+ if (p_ReplicGroup->numOfEntries == 1)
-+ RETURN_ERROR(MAJOR, E_CONFLICT, ("Can't remove the last member. At least one member should be related to a group."));
-+
-+ err = RemoveMember(p_ReplicGroup, memberIndex);
-+
-+ /* unlock */
-+ FrmReplicGroupUnlock(p_ReplicGroup);
-+
-+ switch (GET_ERROR_TYPE(err))
-+ {
-+ case E_OK:
-+ return E_OK;
-+
-+ case E_BUSY:
-+ DBG(TRACE, ("E_BUSY error"));
-+ return ERROR_CODE(E_BUSY);
-+
-+ default:
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+}
-+
-+/*********************** End of API routines ************************/
-+
-+
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/Pcd/fm_replic.h
-@@ -0,0 +1,101 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/******************************************************************************
-+ @File fm_replic.h
-+
-+ @Description FM frame replicator
-+*//***************************************************************************/
-+#ifndef __FM_REPLIC_H
-+#define __FM_REPLIC_H
-+
-+#include "std_ext.h"
-+#include "error_ext.h"
-+
-+
-+#define FRM_REPLIC_SOURCE_TD_OPCODE 0x75
-+#define NEXT_FRM_REPLIC_ADDR_SHIFT 4
-+#define NEXT_FRM_REPLIC_MEMBER_INDEX_SHIFT 16
-+#define FRM_REPLIC_FR_BIT 0x08000000
-+#define FRM_REPLIC_NL_BIT 0x10000000
-+#define FRM_REPLIC_INVALID_MEMBER_INDEX 0xffff
-+#define FRM_REPLIC_FIRST_MEMBER_INDEX 0
-+
-+#define FRM_REPLIC_MIDDLE_MEMBER_INDEX 1
-+#define FRM_REPLIC_LAST_MEMBER_INDEX 2
-+
-+#define SOURCE_TD_ITSELF_OPTION 0x01
-+#define SOURCE_TD_COPY_OPTION 0x02
-+#define SOURCE_TD_ITSELF_AND_COPY_OPTION SOURCE_TD_ITSELF_OPTION | SOURCE_TD_COPY_OPTION
-+#define SOURCE_TD_NONE 0x04
-+
-+/*typedef enum e_SourceTdOption
-+{
-+ e_SOURCE_TD_NONE = 0,
-+ e_SOURCE_TD_ITSELF_OPTION = 1,
-+ e_SOURCE_TD_COPY_OPTION = 2,
-+ e_SOURCE_TD_ITSELF_AND_COPY_OPTION = e_SOURCE_TD_ITSELF_OPTION | e_SOURCE_TD_COPY_OPTION
-+} e_SourceTdOption;
-+*/
-+
-+typedef struct
-+{
-+ volatile uint32_t type;
-+ volatile uint32_t frGroupPointer;
-+ volatile uint32_t operationCode;
-+ volatile uint32_t reserved;
-+} t_FrmReplicGroupSourceAd;
-+
-+typedef struct t_FmPcdFrmReplicMember
-+{
-+ void *p_MemberAd; /**< pointer to the member AD */
-+ void *p_StatisticsAd;/**< pointer to the statistics AD of the member */
-+ t_Handle h_Manip; /**< manip handle - need for free routines */
-+ t_List node;
-+} t_FmPcdFrmReplicMember;
-+
-+typedef struct t_FmPcdFrmReplicGroup
-+{
-+ t_Handle h_FmPcd;
-+
-+ uint8_t maxNumOfEntries;/**< maximal number of members in the group */
-+ uint8_t numOfEntries; /**< actual number of members in the group */
-+ uint16_t owners; /**< how many keys share this frame replicator group */
-+ void *p_SourceTd; /**< pointer to the frame replicator source table descriptor */
-+ t_List membersList; /**< the members list - should reflect the order of the members as in the hw linked list*/
-+ t_List availableMembersList;/**< list of all the available members in the group */
-+ t_FmPcdLock *p_Lock;
-+} t_FmPcdFrmReplicGroup;
-+
-+
-+#endif /* __FM_REPLIC_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/Pcd/fman_kg.c
-@@ -0,0 +1,888 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#include "fsl_fman_kg.h"
-+
-+/****************************************/
-+/* static functions */
-+/****************************************/
-+
-+
-+static uint32_t build_ar_bind_scheme(uint8_t hwport_id, bool write)
-+{
-+ uint32_t rw;
-+
-+ rw = write ? (uint32_t)FM_KG_KGAR_WRITE : (uint32_t)FM_KG_KGAR_READ;
-+
-+ return (uint32_t)(FM_KG_KGAR_GO |
-+ rw |
-+ FM_PCD_KG_KGAR_SEL_PORT_ENTRY |
-+ hwport_id |
-+ FM_PCD_KG_KGAR_SEL_PORT_WSEL_SP);
-+}
-+
-+static void clear_pe_all_scheme(struct fman_kg_regs *regs, uint8_t hwport_id)
-+{
-+ uint32_t ar;
-+
-+ fman_kg_write_sp(regs, 0xffffffff, 0);
-+
-+ ar = build_ar_bind_scheme(hwport_id, TRUE);
-+ fman_kg_write_ar_wait(regs, ar);
-+}
-+
-+static uint32_t build_ar_bind_cls_plan(uint8_t hwport_id, bool write)
-+{
-+ uint32_t rw;
-+
-+ rw = write ? (uint32_t)FM_KG_KGAR_WRITE : (uint32_t)FM_KG_KGAR_READ;
-+
-+ return (uint32_t)(FM_KG_KGAR_GO |
-+ rw |
-+ FM_PCD_KG_KGAR_SEL_PORT_ENTRY |
-+ hwport_id |
-+ FM_PCD_KG_KGAR_SEL_PORT_WSEL_CPP);
-+}
-+
-+static void clear_pe_all_cls_plan(struct fman_kg_regs *regs, uint8_t hwport_id)
-+{
-+ uint32_t ar;
-+
-+ fman_kg_write_cpp(regs, 0);
-+
-+ ar = build_ar_bind_cls_plan(hwport_id, TRUE);
-+ fman_kg_write_ar_wait(regs, ar);
-+}
-+
-+static uint8_t get_gen_ht_code(enum fman_kg_gen_extract_src src,
-+ bool no_validation,
-+ uint8_t *offset)
-+{
-+ int code;
-+
-+ switch (src) {
-+ case E_FMAN_KG_GEN_EXTRACT_ETH:
-+ code = no_validation ? 0x73 : 0x3;
-+ break;
-+
-+ case E_FMAN_KG_GEN_EXTRACT_ETYPE:
-+ code = no_validation ? 0x77 : 0x7;
-+ break;
-+
-+ case E_FMAN_KG_GEN_EXTRACT_SNAP:
-+ code = no_validation ? 0x74 : 0x4;
-+ break;
-+
-+ case E_FMAN_KG_GEN_EXTRACT_VLAN_TCI_1:
-+ code = no_validation ? 0x75 : 0x5;
-+ break;
-+
-+ case E_FMAN_KG_GEN_EXTRACT_VLAN_TCI_N:
-+ code = no_validation ? 0x76 : 0x6;
-+ break;
-+
-+ case E_FMAN_KG_GEN_EXTRACT_PPPoE:
-+ code = no_validation ? 0x78 : 0x8;
-+ break;
-+
-+ case E_FMAN_KG_GEN_EXTRACT_MPLS_1:
-+ code = no_validation ? 0x79 : 0x9;
-+ break;
-+
-+ case E_FMAN_KG_GEN_EXTRACT_MPLS_2:
-+ code = no_validation ? FM_KG_SCH_GEN_HT_INVALID : 0x19;
-+ break;
-+
-+ case E_FMAN_KG_GEN_EXTRACT_MPLS_3:
-+ code = no_validation ? FM_KG_SCH_GEN_HT_INVALID : 0x29;
-+ break;
-+
-+ case E_FMAN_KG_GEN_EXTRACT_MPLS_N:
-+ code = no_validation ? 0x7a : 0xa;
-+ break;
-+
-+ case E_FMAN_KG_GEN_EXTRACT_IPv4_1:
-+ code = no_validation ? 0x7b : 0xb;
-+ break;
-+
-+ case E_FMAN_KG_GEN_EXTRACT_IPv6_1:
-+ code = no_validation ? 0x7b : 0x1b;
-+ break;
-+
-+ case E_FMAN_KG_GEN_EXTRACT_IPv4_2:
-+ code = no_validation ? 0x7c : 0xc;
-+ break;
-+
-+ case E_FMAN_KG_GEN_EXTRACT_IPv6_2:
-+ code = no_validation ? 0x7c : 0x1c;
-+ break;
-+
-+ case E_FMAN_KG_GEN_EXTRACT_MINENCAP:
-+ code = no_validation ? 0x7c : 0x2c;
-+ break;
-+
-+ case E_FMAN_KG_GEN_EXTRACT_IP_PID:
-+ code = no_validation ? 0x72 : 0x2;
-+ break;
-+
-+ case E_FMAN_KG_GEN_EXTRACT_GRE:
-+ code = no_validation ? 0x7d : 0xd;
-+ break;
-+
-+ case E_FMAN_KG_GEN_EXTRACT_TCP:
-+ code = no_validation ? 0x7e : 0xe;
-+ break;
-+
-+ case E_FMAN_KG_GEN_EXTRACT_UDP:
-+ code = no_validation ? 0x7e : 0x1e;
-+ break;
-+
-+ case E_FMAN_KG_GEN_EXTRACT_SCTP:
-+ code = no_validation ? 0x7e : 0x3e;
-+ break;
-+
-+ case E_FMAN_KG_GEN_EXTRACT_DCCP:
-+ code = no_validation ? 0x7e : 0x4e;
-+ break;
-+
-+ case E_FMAN_KG_GEN_EXTRACT_IPSEC_AH:
-+ code = no_validation ? 0x7e : 0x2e;
-+ break;
-+
-+ case E_FMAN_KG_GEN_EXTRACT_IPSEC_ESP:
-+ code = no_validation ? 0x7e : 0x6e;
-+ break;
-+
-+ case E_FMAN_KG_GEN_EXTRACT_SHIM_1:
-+ code = 0x70;
-+ break;
-+
-+ case E_FMAN_KG_GEN_EXTRACT_SHIM_2:
-+ code = 0x71;
-+ break;
-+
-+ case E_FMAN_KG_GEN_EXTRACT_FROM_DFLT:
-+ code = 0x10;
-+ break;
-+
-+ case E_FMAN_KG_GEN_EXTRACT_FROM_FRAME_START:
-+ code = 0x40;
-+ break;
-+
-+ case E_FMAN_KG_GEN_EXTRACT_FROM_PARSE_RESULT:
-+ code = 0x20;
-+ break;
-+
-+ case E_FMAN_KG_GEN_EXTRACT_FROM_END_OF_PARSE:
-+ code = 0x7f;
-+ break;
-+
-+ case E_FMAN_KG_GEN_EXTRACT_FROM_FQID:
-+ code = 0x20;
-+ *offset += 0x20;
-+ break;
-+
-+ default:
-+ code = FM_KG_SCH_GEN_HT_INVALID;
-+ }
-+
-+ return (uint8_t)code;
-+}
-+
-+static uint32_t build_ar_scheme(uint8_t scheme,
-+ uint8_t hwport_id,
-+ bool update_counter,
-+ bool write)
-+{
-+ uint32_t rw;
-+
-+ rw = (uint32_t)(write ? FM_KG_KGAR_WRITE : FM_KG_KGAR_READ);
-+
-+ return (uint32_t)(FM_KG_KGAR_GO |
-+ rw |
-+ FM_KG_KGAR_SEL_SCHEME_ENTRY |
-+ hwport_id |
-+ ((uint32_t)scheme << FM_KG_KGAR_NUM_SHIFT) |
-+ (update_counter ? FM_KG_KGAR_SCM_WSEL_UPDATE_CNT : 0));
-+}
-+
-+static uint32_t build_ar_cls_plan(uint8_t grp,
-+ uint8_t entries_mask,
-+ uint8_t hwport_id,
-+ bool write)
-+{
-+ uint32_t rw;
-+
-+ rw = (uint32_t)(write ? FM_KG_KGAR_WRITE : FM_KG_KGAR_READ);
-+
-+ return (uint32_t)(FM_KG_KGAR_GO |
-+ rw |
-+ FM_PCD_KG_KGAR_SEL_CLS_PLAN_ENTRY |
-+ hwport_id |
-+ ((uint32_t)grp << FM_KG_KGAR_NUM_SHIFT) |
-+ ((uint32_t)entries_mask << FM_KG_KGAR_WSEL_SHIFT));
-+}
-+
-+int fman_kg_write_ar_wait(struct fman_kg_regs *regs, uint32_t fmkg_ar)
-+{
-+ iowrite32be(fmkg_ar, &regs->fmkg_ar);
-+ /* Wait for GO to be idle and read error */
-+ while ((fmkg_ar = ioread32be(&regs->fmkg_ar)) & FM_KG_KGAR_GO) ;
-+ if (fmkg_ar & FM_PCD_KG_KGAR_ERR)
-+ return -EINVAL;
-+ return 0;
-+}
-+
-+void fman_kg_write_sp(struct fman_kg_regs *regs, uint32_t sp, bool add)
-+{
-+
-+ struct fman_kg_pe_regs *kgpe_regs;
-+ uint32_t tmp;
-+
-+ kgpe_regs = (struct fman_kg_pe_regs *)&(regs->fmkg_indirect[0]);
-+ tmp = ioread32be(&kgpe_regs->fmkg_pe_sp);
-+
-+ if (add)
-+ tmp |= sp;
-+ else /* clear */
-+ tmp &= ~sp;
-+
-+ iowrite32be(tmp, &kgpe_regs->fmkg_pe_sp);
-+
-+}
-+
-+void fman_kg_write_cpp(struct fman_kg_regs *regs, uint32_t cpp)
-+{
-+ struct fman_kg_pe_regs *kgpe_regs;
-+
-+ kgpe_regs = (struct fman_kg_pe_regs *)&(regs->fmkg_indirect[0]);
-+
-+ iowrite32be(cpp, &kgpe_regs->fmkg_pe_cpp);
-+}
-+
-+void fman_kg_get_event(struct fman_kg_regs *regs,
-+ uint32_t *event,
-+ uint32_t *scheme_idx)
-+{
-+ uint32_t mask, force;
-+
-+ *event = ioread32be(&regs->fmkg_eer);
-+ mask = ioread32be(&regs->fmkg_eeer);
-+ *scheme_idx = ioread32be(&regs->fmkg_seer);
-+ *scheme_idx &= ioread32be(&regs->fmkg_seeer);
-+
-+ *event &= mask;
-+
-+ /* clear the forced events */
-+ force = ioread32be(&regs->fmkg_feer);
-+ if (force & *event)
-+ iowrite32be(force & ~*event ,&regs->fmkg_feer);
-+
-+ iowrite32be(*event, &regs->fmkg_eer);
-+ iowrite32be(*scheme_idx, &regs->fmkg_seer);
-+}
-+
-+
-+void fman_kg_init(struct fman_kg_regs *regs,
-+ uint32_t exceptions,
-+ uint32_t dflt_nia)
-+{
-+ uint32_t tmp;
-+ int i;
-+
-+ iowrite32be(FM_EX_KG_DOUBLE_ECC | FM_EX_KG_KEYSIZE_OVERFLOW,
-+ &regs->fmkg_eer);
-+
-+ tmp = 0;
-+ if (exceptions & FM_EX_KG_DOUBLE_ECC)
-+ tmp |= FM_EX_KG_DOUBLE_ECC;
-+
-+ if (exceptions & FM_EX_KG_KEYSIZE_OVERFLOW)
-+ tmp |= FM_EX_KG_KEYSIZE_OVERFLOW;
-+
-+ iowrite32be(tmp, &regs->fmkg_eeer);
-+ iowrite32be(0, &regs->fmkg_fdor);
-+ iowrite32be(0, &regs->fmkg_gdv0r);
-+ iowrite32be(0, &regs->fmkg_gdv1r);
-+ iowrite32be(dflt_nia, &regs->fmkg_gcr);
-+
-+ /* Clear binding between ports to schemes and classification plans
-+ * so that all ports are not bound to any scheme/classification plan */
-+ for (i = 0; i < FMAN_MAX_NUM_OF_HW_PORTS; i++) {
-+ clear_pe_all_scheme(regs, (uint8_t)i);
-+ clear_pe_all_cls_plan(regs, (uint8_t)i);
-+ }
-+}
-+
-+void fman_kg_enable_scheme_interrupts(struct fman_kg_regs *regs)
-+{
-+ /* enable and enable all scheme interrupts */
-+ iowrite32be(0xFFFFFFFF, &regs->fmkg_seer);
-+ iowrite32be(0xFFFFFFFF, &regs->fmkg_seeer);
-+}
-+
-+void fman_kg_enable(struct fman_kg_regs *regs)
-+{
-+ iowrite32be(ioread32be(&regs->fmkg_gcr) | FM_KG_KGGCR_EN,
-+ &regs->fmkg_gcr);
-+}
-+
-+void fman_kg_disable(struct fman_kg_regs *regs)
-+{
-+ iowrite32be(ioread32be(&regs->fmkg_gcr) & ~FM_KG_KGGCR_EN,
-+ &regs->fmkg_gcr);
-+}
-+
-+void fman_kg_set_data_after_prs(struct fman_kg_regs *regs, uint8_t offset)
-+{
-+ iowrite32be(offset, &regs->fmkg_fdor);
-+}
-+
-+void fman_kg_set_dflt_val(struct fman_kg_regs *regs,
-+ uint8_t def_id,
-+ uint32_t val)
-+{
-+ if(def_id == 0)
-+ iowrite32be(val, &regs->fmkg_gdv0r);
-+ else
-+ iowrite32be(val, &regs->fmkg_gdv1r);
-+}
-+
-+
-+void fman_kg_set_exception(struct fman_kg_regs *regs,
-+ uint32_t exception,
-+ bool enable)
-+{
-+ uint32_t tmp;
-+
-+ tmp = ioread32be(&regs->fmkg_eeer);
-+
-+ if (enable) {
-+ tmp |= exception;
-+ } else {
-+ tmp &= ~exception;
-+ }
-+
-+ iowrite32be(tmp, &regs->fmkg_eeer);
-+}
-+
-+void fman_kg_get_exception(struct fman_kg_regs *regs,
-+ uint32_t *events,
-+ uint32_t *scheme_ids,
-+ bool clear)
-+{
-+ uint32_t mask;
-+
-+ *events = ioread32be(&regs->fmkg_eer);
-+ mask = ioread32be(&regs->fmkg_eeer);
-+ *events &= mask;
-+
-+ *scheme_ids = 0;
-+
-+ if (*events & FM_EX_KG_KEYSIZE_OVERFLOW) {
-+ *scheme_ids = ioread32be(&regs->fmkg_seer);
-+ mask = ioread32be(&regs->fmkg_seeer);
-+ *scheme_ids &= mask;
-+ }
-+
-+ if (clear) {
-+ iowrite32be(*scheme_ids, &regs->fmkg_seer);
-+ iowrite32be(*events, &regs->fmkg_eer);
-+ }
-+}
-+
-+void fman_kg_get_capture(struct fman_kg_regs *regs,
-+ struct fman_kg_ex_ecc_attr *ecc_attr,
-+ bool clear)
-+{
-+ uint32_t tmp;
-+
-+ tmp = ioread32be(&regs->fmkg_serc);
-+
-+ if (tmp & KG_FMKG_SERC_CAP) {
-+ /* Captured data is valid */
-+ ecc_attr->valid = TRUE;
-+ ecc_attr->double_ecc =
-+ (bool)((tmp & KG_FMKG_SERC_CET) ? TRUE : FALSE);
-+ ecc_attr->single_ecc_count =
-+ (uint8_t)((tmp & KG_FMKG_SERC_CNT_MSK) >>
-+ KG_FMKG_SERC_CNT_SHIFT);
-+ ecc_attr->addr = (uint16_t)(tmp & KG_FMKG_SERC_ADDR_MSK);
-+
-+ if (clear)
-+ iowrite32be(KG_FMKG_SERC_CAP, &regs->fmkg_serc);
-+ } else {
-+ /* No ECC error is captured */
-+ ecc_attr->valid = FALSE;
-+ }
-+}
-+
-+int fman_kg_build_scheme(struct fman_kg_scheme_params *params,
-+ struct fman_kg_scheme_regs *scheme_regs)
-+{
-+ struct fman_kg_extract_params *extract_params;
-+ struct fman_kg_gen_extract_params *gen_params;
-+ uint32_t tmp_reg, i, select, mask, fqb;
-+ uint8_t offset, shift, ht;
-+
-+ /* Zero out all registers so no need to care about unused ones */
-+ memset(scheme_regs, 0, sizeof(struct fman_kg_scheme_regs));
-+
-+ /* Mode register */
-+ tmp_reg = fm_kg_build_nia(params->next_engine,
-+ params->next_engine_action);
-+ if (tmp_reg == KG_NIA_INVALID) {
-+ return -EINVAL;
-+ }
-+
-+ if (params->next_engine == E_FMAN_PCD_PLCR) {
-+ tmp_reg |= FMAN_KG_SCH_MODE_NIA_PLCR;
-+ }
-+ else if (params->next_engine == E_FMAN_PCD_CC) {
-+ tmp_reg |= (uint32_t)params->cc_params.base_offset <<
-+ FMAN_KG_SCH_MODE_CCOBASE_SHIFT;
-+ }
-+
-+ tmp_reg |= FMAN_KG_SCH_MODE_EN;
-+ scheme_regs->kgse_mode = tmp_reg;
-+
-+ /* Match vector */
-+ scheme_regs->kgse_mv = params->match_vector;
-+
-+ extract_params = &params->extract_params;
-+
-+ /* Scheme default values registers */
-+ scheme_regs->kgse_dv0 = extract_params->def_scheme_0;
-+ scheme_regs->kgse_dv1 = extract_params->def_scheme_1;
-+
-+ /* Extract Known Fields Command register */
-+ scheme_regs->kgse_ekfc = extract_params->known_fields;
-+
-+ /* Entry Extract Known Default Value register */
-+ tmp_reg = 0;
-+ tmp_reg |= extract_params->known_fields_def.mac_addr <<
-+ FMAN_KG_SCH_DEF_MAC_ADDR_SHIFT;
-+ tmp_reg |= extract_params->known_fields_def.vlan_tci <<
-+ FMAN_KG_SCH_DEF_VLAN_TCI_SHIFT;
-+ tmp_reg |= extract_params->known_fields_def.etype <<
-+ FMAN_KG_SCH_DEF_ETYPE_SHIFT;
-+ tmp_reg |= extract_params->known_fields_def.ppp_sid <<
-+ FMAN_KG_SCH_DEF_PPP_SID_SHIFT;
-+ tmp_reg |= extract_params->known_fields_def.ppp_pid <<
-+ FMAN_KG_SCH_DEF_PPP_PID_SHIFT;
-+ tmp_reg |= extract_params->known_fields_def.mpls <<
-+ FMAN_KG_SCH_DEF_MPLS_SHIFT;
-+ tmp_reg |= extract_params->known_fields_def.ip_addr <<
-+ FMAN_KG_SCH_DEF_IP_ADDR_SHIFT;
-+ tmp_reg |= extract_params->known_fields_def.ptype <<
-+ FMAN_KG_SCH_DEF_PTYPE_SHIFT;
-+ tmp_reg |= extract_params->known_fields_def.ip_tos_tc <<
-+ FMAN_KG_SCH_DEF_IP_TOS_TC_SHIFT;
-+ tmp_reg |= extract_params->known_fields_def.ipv6_fl <<
-+ FMAN_KG_SCH_DEF_IPv6_FL_SHIFT;
-+ tmp_reg |= extract_params->known_fields_def.ipsec_spi <<
-+ FMAN_KG_SCH_DEF_IPSEC_SPI_SHIFT;
-+ tmp_reg |= extract_params->known_fields_def.l4_port <<
-+ FMAN_KG_SCH_DEF_L4_PORT_SHIFT;
-+ tmp_reg |= extract_params->known_fields_def.tcp_flg <<
-+ FMAN_KG_SCH_DEF_TCP_FLG_SHIFT;
-+
-+ scheme_regs->kgse_ekdv = tmp_reg;
-+
-+ /* Generic extract registers */
-+ if (extract_params->gen_extract_num > FM_KG_NUM_OF_GENERIC_REGS) {
-+ return -EINVAL;
-+ }
-+
-+ for (i = 0; i < extract_params->gen_extract_num; i++) {
-+ gen_params = extract_params->gen_extract + i;
-+
-+ tmp_reg = FMAN_KG_SCH_GEN_VALID;
-+ tmp_reg |= (uint32_t)gen_params->def_val <<
-+ FMAN_KG_SCH_GEN_DEF_SHIFT;
-+
-+ if (gen_params->type == E_FMAN_KG_HASH_EXTRACT) {
-+ if ((gen_params->extract > FMAN_KG_SCH_GEN_SIZE_MAX) ||
-+ (gen_params->extract == 0)) {
-+ return -EINVAL;
-+ }
-+ } else {
-+ tmp_reg |= FMAN_KG_SCH_GEN_OR;
-+ }
-+
-+ tmp_reg |= (uint32_t)gen_params->extract <<
-+ FMAN_KG_SCH_GEN_SIZE_SHIFT;
-+ tmp_reg |= (uint32_t)gen_params->mask <<
-+ FMAN_KG_SCH_GEN_MASK_SHIFT;
-+
-+ offset = gen_params->offset;
-+ ht = get_gen_ht_code(gen_params->src,
-+ gen_params->no_validation,
-+ &offset);
-+ tmp_reg |= (uint32_t)ht << FMAN_KG_SCH_GEN_HT_SHIFT;
-+ tmp_reg |= offset;
-+
-+ scheme_regs->kgse_gec[i] = tmp_reg;
-+ }
-+
-+ /* Masks registers */
-+ if (extract_params->masks_num > FM_KG_EXTRACT_MASKS_NUM) {
-+ return -EINVAL;
-+ }
-+
-+ select = 0;
-+ mask = 0;
-+ fqb = 0;
-+ for (i = 0; i < extract_params->masks_num; i++) {
-+ /* MCSx fields */
-+ KG_GET_MASK_SEL_SHIFT(shift, i);
-+ if (extract_params->masks[i].is_known) {
-+ /* Mask known field */
-+ select |= extract_params->masks[i].field_or_gen_idx <<
-+ shift;
-+ } else {
-+ /* Mask generic extract */
-+ select |= (extract_params->masks[i].field_or_gen_idx +
-+ FM_KG_MASK_SEL_GEN_BASE) << shift;
-+ }
-+
-+ /* MOx fields - spread between se_bmch and se_fqb registers */
-+ KG_GET_MASK_OFFSET_SHIFT(shift, i);
-+ if (i < 2) {
-+ select |= (uint32_t)extract_params->masks[i].offset <<
-+ shift;
-+ } else {
-+ fqb |= (uint32_t)extract_params->masks[i].offset <<
-+ shift;
-+ }
-+
-+ /* BMx fields */
-+ KG_GET_MASK_SHIFT(shift, i);
-+ mask |= (uint32_t)extract_params->masks[i].mask << shift;
-+ }
-+
-+ /* Finish with rest of BMx fileds -
-+ * don't mask bits for unused masks by setting
-+ * corresponding BMx field = 0xFF */
-+ for (i = extract_params->masks_num; i < FM_KG_EXTRACT_MASKS_NUM; i++) {
-+ KG_GET_MASK_SHIFT(shift, i);
-+ mask |= 0xFF << shift;
-+ }
-+
-+ scheme_regs->kgse_bmch = select;
-+ scheme_regs->kgse_bmcl = mask;
-+
-+ /* Finish with FQB register initialization.
-+ * Check fqid is 24-bit value. */
-+ if (params->base_fqid & ~0x00FFFFFF) {
-+ return -EINVAL;
-+ }
-+
-+ fqb |= params->base_fqid;
-+ scheme_regs->kgse_fqb = fqb;
-+
-+ /* Hash Configuration register */
-+ tmp_reg = 0;
-+ if (params->hash_params.use_hash) {
-+ /* Check hash mask is 24-bit value */
-+ if (params->hash_params.mask & ~0x00FFFFFF) {
-+ return -EINVAL;
-+ }
-+
-+ /* Hash function produces 64-bit value, 24 bits of that
-+ * are used to generate fq_id and policer profile.
-+ * Thus, maximal shift is 40 bits to allow 24 bits out of 64.
-+ */
-+ if (params->hash_params.shift_r > FMAN_KG_SCH_HASH_HSHIFT_MAX) {
-+ return -EINVAL;
-+ }
-+
-+ tmp_reg |= params->hash_params.mask;
-+ tmp_reg |= (uint32_t)params->hash_params.shift_r <<
-+ FMAN_KG_SCH_HASH_HSHIFT_SHIFT;
-+
-+ if (params->hash_params.sym) {
-+ tmp_reg |= FMAN_KG_SCH_HASH_SYM;
-+ }
-+
-+ }
-+
-+ if (params->bypass_fqid_gen) {
-+ tmp_reg |= FMAN_KG_SCH_HASH_NO_FQID_GEN;
-+ }
-+
-+ scheme_regs->kgse_hc = tmp_reg;
-+
-+ /* Policer Profile register */
-+ if (params->policer_params.bypass_pp_gen) {
-+ tmp_reg = 0;
-+ } else {
-+ /* Lower 8 bits of 24-bits extracted from hash result
-+ * are used for policer profile generation.
-+ * That leaves maximum shift value = 23. */
-+ if (params->policer_params.shift > FMAN_KG_SCH_PP_SHIFT_MAX) {
-+ return -EINVAL;
-+ }
-+
-+ tmp_reg = params->policer_params.base;
-+ tmp_reg |= ((uint32_t)params->policer_params.shift <<
-+ FMAN_KG_SCH_PP_SH_SHIFT) &
-+ FMAN_KG_SCH_PP_SH_MASK;
-+ tmp_reg |= ((uint32_t)params->policer_params.shift <<
-+ FMAN_KG_SCH_PP_SL_SHIFT) &
-+ FMAN_KG_SCH_PP_SL_MASK;
-+ tmp_reg |= (uint32_t)params->policer_params.mask <<
-+ FMAN_KG_SCH_PP_MASK_SHIFT;
-+ }
-+
-+ scheme_regs->kgse_ppc = tmp_reg;
-+
-+ /* Coarse Classification Bit Select register */
-+ if (params->next_engine == E_FMAN_PCD_CC) {
-+ scheme_regs->kgse_ccbs = params->cc_params.qlcv_bits_sel;
-+ }
-+
-+ /* Packets Counter register */
-+ if (params->update_counter) {
-+ scheme_regs->kgse_spc = params->counter_value;
-+ }
-+
-+ return 0;
-+}
-+
-+int fman_kg_write_scheme(struct fman_kg_regs *regs,
-+ uint8_t scheme_id,
-+ uint8_t hwport_id,
-+ struct fman_kg_scheme_regs *scheme_regs,
-+ bool update_counter)
-+{
-+ struct fman_kg_scheme_regs *kgse_regs;
-+ uint32_t tmp_reg;
-+ int err, i;
-+
-+ /* Write indirect scheme registers */
-+ kgse_regs = (struct fman_kg_scheme_regs *)&(regs->fmkg_indirect[0]);
-+
-+ iowrite32be(scheme_regs->kgse_mode, &kgse_regs->kgse_mode);
-+ iowrite32be(scheme_regs->kgse_ekfc, &kgse_regs->kgse_ekfc);
-+ iowrite32be(scheme_regs->kgse_ekdv, &kgse_regs->kgse_ekdv);
-+ iowrite32be(scheme_regs->kgse_bmch, &kgse_regs->kgse_bmch);
-+ iowrite32be(scheme_regs->kgse_bmcl, &kgse_regs->kgse_bmcl);
-+ iowrite32be(scheme_regs->kgse_fqb, &kgse_regs->kgse_fqb);
-+ iowrite32be(scheme_regs->kgse_hc, &kgse_regs->kgse_hc);
-+ iowrite32be(scheme_regs->kgse_ppc, &kgse_regs->kgse_ppc);
-+ iowrite32be(scheme_regs->kgse_spc, &kgse_regs->kgse_spc);
-+ iowrite32be(scheme_regs->kgse_dv0, &kgse_regs->kgse_dv0);
-+ iowrite32be(scheme_regs->kgse_dv1, &kgse_regs->kgse_dv1);
-+ iowrite32be(scheme_regs->kgse_ccbs, &kgse_regs->kgse_ccbs);
-+ iowrite32be(scheme_regs->kgse_mv, &kgse_regs->kgse_mv);
-+
-+ for (i = 0 ; i < FM_KG_NUM_OF_GENERIC_REGS ; i++)
-+ iowrite32be(scheme_regs->kgse_gec[i], &kgse_regs->kgse_gec[i]);
-+
-+ /* Write AR (Action register) */
-+ tmp_reg = build_ar_scheme(scheme_id, hwport_id, update_counter, TRUE);
-+ err = fman_kg_write_ar_wait(regs, tmp_reg);
-+ return err;
-+}
-+
-+int fman_kg_delete_scheme(struct fman_kg_regs *regs,
-+ uint8_t scheme_id,
-+ uint8_t hwport_id)
-+{
-+ struct fman_kg_scheme_regs *kgse_regs;
-+ uint32_t tmp_reg;
-+ int err, i;
-+
-+ kgse_regs = (struct fman_kg_scheme_regs *)&(regs->fmkg_indirect[0]);
-+
-+ /* Clear all registers including enable bit in mode register */
-+ for (i = 0; i < (sizeof(struct fman_kg_scheme_regs)) / 4; ++i) {
-+ iowrite32be(0, ((uint32_t *)kgse_regs + i));
-+ }
-+
-+ /* Write AR (Action register) */
-+ tmp_reg = build_ar_scheme(scheme_id, hwport_id, FALSE, TRUE);
-+ err = fman_kg_write_ar_wait(regs, tmp_reg);
-+ return err;
-+}
-+
-+int fman_kg_get_scheme_counter(struct fman_kg_regs *regs,
-+ uint8_t scheme_id,
-+ uint8_t hwport_id,
-+ uint32_t *counter)
-+{
-+ struct fman_kg_scheme_regs *kgse_regs;
-+ uint32_t tmp_reg;
-+ int err;
-+
-+ kgse_regs = (struct fman_kg_scheme_regs *)&(regs->fmkg_indirect[0]);
-+
-+ tmp_reg = build_ar_scheme(scheme_id, hwport_id, TRUE, FALSE);
-+ err = fman_kg_write_ar_wait(regs, tmp_reg);
-+
-+ if (err != 0)
-+ return err;
-+
-+ *counter = ioread32be(&kgse_regs->kgse_spc);
-+
-+ return 0;
-+}
-+
-+int fman_kg_set_scheme_counter(struct fman_kg_regs *regs,
-+ uint8_t scheme_id,
-+ uint8_t hwport_id,
-+ uint32_t counter)
-+{
-+ struct fman_kg_scheme_regs *kgse_regs;
-+ uint32_t tmp_reg;
-+ int err;
-+
-+ kgse_regs = (struct fman_kg_scheme_regs *)&(regs->fmkg_indirect[0]);
-+
-+ tmp_reg = build_ar_scheme(scheme_id, hwport_id, TRUE, FALSE);
-+
-+ err = fman_kg_write_ar_wait(regs, tmp_reg);
-+ if (err != 0)
-+ return err;
-+
-+ /* Keygen indirect access memory contains all scheme_id registers
-+ * by now. Change only counter value. */
-+ iowrite32be(counter, &kgse_regs->kgse_spc);
-+
-+ /* Write back scheme registers */
-+ tmp_reg = build_ar_scheme(scheme_id, hwport_id, TRUE, TRUE);
-+ err = fman_kg_write_ar_wait(regs, tmp_reg);
-+
-+ return err;
-+}
-+
-+uint32_t fman_kg_get_schemes_total_counter(struct fman_kg_regs *regs)
-+{
-+ return ioread32be(&regs->fmkg_tpc);
-+}
-+
-+int fman_kg_build_cls_plan(struct fman_kg_cls_plan_params *params,
-+ struct fman_kg_cp_regs *cls_plan_regs)
-+{
-+ uint8_t entries_set, entry_bit;
-+ int i;
-+
-+ /* Zero out all group's register */
-+ memset(cls_plan_regs, 0, sizeof(struct fman_kg_cp_regs));
-+
-+ /* Go over all classification entries in params->entries_mask and
-+ * configure the corresponding cpe register */
-+ entries_set = params->entries_mask;
-+ for (i = 0; entries_set; i++) {
-+ entry_bit = (uint8_t)(0x80 >> i);
-+ if ((entry_bit & entries_set) == 0)
-+ continue;
-+ entries_set ^= entry_bit;
-+ cls_plan_regs->kgcpe[i] = params->mask_vector[i];
-+ }
-+
-+ return 0;
-+}
-+
-+int fman_kg_write_cls_plan(struct fman_kg_regs *regs,
-+ uint8_t grp_id,
-+ uint8_t entries_mask,
-+ uint8_t hwport_id,
-+ struct fman_kg_cp_regs *cls_plan_regs)
-+{
-+ struct fman_kg_cp_regs *kgcpe_regs;
-+ uint32_t tmp_reg;
-+ int i, err;
-+
-+ /* Check group index is valid and the group isn't empty */
-+ if (grp_id >= FM_KG_CLS_PLAN_GRPS_NUM)
-+ return -EINVAL;
-+
-+ /* Write indirect classification plan registers */
-+ kgcpe_regs = (struct fman_kg_cp_regs *)&(regs->fmkg_indirect[0]);
-+
-+ for (i = 0; i < FM_KG_NUM_CLS_PLAN_ENTR; i++) {
-+ iowrite32be(cls_plan_regs->kgcpe[i], &kgcpe_regs->kgcpe[i]);
-+ }
-+
-+ tmp_reg = build_ar_cls_plan(grp_id, entries_mask, hwport_id, TRUE);
-+ err = fman_kg_write_ar_wait(regs, tmp_reg);
-+ return err;
-+}
-+
-+int fman_kg_write_bind_schemes(struct fman_kg_regs *regs,
-+ uint8_t hwport_id,
-+ uint32_t schemes)
-+{
-+ struct fman_kg_pe_regs *kg_pe_regs;
-+ uint32_t tmp_reg;
-+ int err;
-+
-+ kg_pe_regs = (struct fman_kg_pe_regs *)&(regs->fmkg_indirect[0]);
-+
-+ iowrite32be(schemes, &kg_pe_regs->fmkg_pe_sp);
-+
-+ tmp_reg = build_ar_bind_scheme(hwport_id, TRUE);
-+ err = fman_kg_write_ar_wait(regs, tmp_reg);
-+ return err;
-+}
-+
-+int fman_kg_build_bind_cls_plans(uint8_t grp_base,
-+ uint8_t grp_mask,
-+ uint32_t *bind_cls_plans)
-+{
-+ /* Check grp_base and grp_mask are 5-bits values */
-+ if ((grp_base & ~0x0000001F) || (grp_mask & ~0x0000001F))
-+ return -EINVAL;
-+
-+ *bind_cls_plans = (uint32_t) ((grp_mask << FMAN_KG_PE_CPP_MASK_SHIFT) | grp_base);
-+ return 0;
-+}
-+
-+
-+int fman_kg_write_bind_cls_plans(struct fman_kg_regs *regs,
-+ uint8_t hwport_id,
-+ uint32_t bind_cls_plans)
-+{
-+ struct fman_kg_pe_regs *kg_pe_regs;
-+ uint32_t tmp_reg;
-+ int err;
-+
-+ kg_pe_regs = (struct fman_kg_pe_regs *)&(regs->fmkg_indirect[0]);
-+
-+ iowrite32be(bind_cls_plans, &kg_pe_regs->fmkg_pe_cpp);
-+
-+ tmp_reg = build_ar_bind_cls_plan(hwport_id, TRUE);
-+ err = fman_kg_write_ar_wait(regs, tmp_reg);
-+ return err;
-+}
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/Pcd/fman_prs.c
-@@ -0,0 +1,129 @@
-+/*
-+ * Copyright 2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#include "fsl_fman_prs.h"
-+
-+uint32_t fman_prs_get_err_event(struct fman_prs_regs *regs, uint32_t ev_mask)
-+{
-+ return ioread32be(&regs->fmpr_perr) & ev_mask;
-+}
-+
-+uint32_t fman_prs_get_err_ev_mask(struct fman_prs_regs *regs)
-+{
-+ return ioread32be(&regs->fmpr_perer);
-+}
-+
-+void fman_prs_ack_err_event(struct fman_prs_regs *regs, uint32_t event)
-+{
-+ iowrite32be(event, &regs->fmpr_perr);
-+}
-+
-+uint32_t fman_prs_get_expt_event(struct fman_prs_regs *regs, uint32_t ev_mask)
-+{
-+ return ioread32be(&regs->fmpr_pevr) & ev_mask;
-+}
-+
-+uint32_t fman_prs_get_expt_ev_mask(struct fman_prs_regs *regs)
-+{
-+ return ioread32be(&regs->fmpr_pever);
-+}
-+
-+void fman_prs_ack_expt_event(struct fman_prs_regs *regs, uint32_t event)
-+{
-+ iowrite32be(event, &regs->fmpr_pevr);
-+}
-+
-+void fman_prs_defconfig(struct fman_prs_cfg *cfg)
-+{
-+ cfg->port_id_stat = 0;
-+ cfg->max_prs_cyc_lim = DEFAULT_MAX_PRS_CYC_LIM;
-+ cfg->prs_exceptions = 0x03000000;
-+}
-+
-+int fman_prs_init(struct fman_prs_regs *regs, struct fman_prs_cfg *cfg)
-+{
-+ uint32_t tmp;
-+
-+ iowrite32be(cfg->max_prs_cyc_lim, &regs->fmpr_rpclim);
-+ iowrite32be((FM_PCD_PRS_SINGLE_ECC | FM_PCD_PRS_PORT_IDLE_STS),
-+ &regs->fmpr_pevr);
-+
-+ if (cfg->prs_exceptions & FM_PCD_EX_PRS_SINGLE_ECC)
-+ iowrite32be(FM_PCD_PRS_SINGLE_ECC, &regs->fmpr_pever);
-+ else
-+ iowrite32be(0, &regs->fmpr_pever);
-+
-+ iowrite32be(FM_PCD_PRS_DOUBLE_ECC, &regs->fmpr_perr);
-+
-+ tmp = 0;
-+ if (cfg->prs_exceptions & FM_PCD_EX_PRS_DOUBLE_ECC)
-+ tmp |= FM_PCD_PRS_DOUBLE_ECC;
-+ iowrite32be(tmp, &regs->fmpr_perer);
-+
-+ iowrite32be(cfg->port_id_stat, &regs->fmpr_ppsc);
-+
-+ return 0;
-+}
-+
-+void fman_prs_enable(struct fman_prs_regs *regs)
-+{
-+ uint32_t tmp;
-+
-+ tmp = ioread32be(&regs->fmpr_rpimac) | FM_PCD_PRS_RPIMAC_EN;
-+ iowrite32be(tmp, &regs->fmpr_rpimac);
-+}
-+
-+void fman_prs_disable(struct fman_prs_regs *regs)
-+{
-+ uint32_t tmp;
-+
-+ tmp = ioread32be(&regs->fmpr_rpimac) & ~FM_PCD_PRS_RPIMAC_EN;
-+ iowrite32be(tmp, &regs->fmpr_rpimac);
-+}
-+
-+int fman_prs_is_enabled(struct fman_prs_regs *regs)
-+{
-+ return ioread32be(&regs->fmpr_rpimac) & FM_PCD_PRS_RPIMAC_EN;
-+}
-+
-+void fman_prs_set_stst_port_msk(struct fman_prs_regs *regs, uint32_t pid_msk)
-+{
-+ iowrite32be(pid_msk, &regs->fmpr_ppsc);
-+}
-+
-+void fman_prs_set_stst(struct fman_prs_regs *regs, bool enable)
-+{
-+ if (enable)
-+ iowrite32be(FM_PCD_PRS_PPSC_ALL_PORTS, &regs->fmpr_ppsc);
-+ else
-+ iowrite32be(0, &regs->fmpr_ppsc);
-+}
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/Port/Makefile
-@@ -0,0 +1,15 @@
-+#
-+# Makefile for the Freescale Ethernet controllers
-+#
-+ccflags-y += -DVERSION=\"\"
-+#
-+#Include netcomm SW specific definitions
-+include $(srctree)/drivers/net/ethernet/freescale/sdk_fman/ncsw_config.mk
-+
-+NCSW_FM_INC = $(srctree)/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/inc
-+
-+ccflags-y += -I$(NCSW_FM_INC)
-+
-+obj-y += fsl-ncsw-Pcd.o
-+
-+fsl-ncsw-Pcd-objs := fm_port.o fm_port_im.o fman_port.o
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/Port/fm_port.c
-@@ -0,0 +1,6436 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/******************************************************************************
-+ @File fm_port.c
-+
-+ @Description FM driver routines implementation.
-+ *//***************************************************************************/
-+#include "error_ext.h"
-+#include "std_ext.h"
-+#include "string_ext.h"
-+#include "sprint_ext.h"
-+#include "debug_ext.h"
-+#include "fm_muram_ext.h"
-+
-+#include "fman_common.h"
-+#include "fm_port.h"
-+#include "fm_port_dsar.h"
-+#include "common/general.h"
-+
-+/****************************************/
-+/* static functions */
-+/****************************************/
-+static t_Error FmPortConfigAutoResForDeepSleepSupport1(t_FmPort *p_FmPort);
-+
-+static t_Error CheckInitParameters(t_FmPort *p_FmPort)
-+{
-+ t_FmPortDriverParam *p_Params = p_FmPort->p_FmPortDriverParam;
-+ struct fman_port_cfg *p_DfltConfig = &p_Params->dfltCfg;
-+ t_Error ans = E_OK;
-+ uint32_t unusedMask;
-+
-+ if (p_FmPort->imEn)
-+ {
-+ if (p_FmPort->portType == e_FM_PORT_TYPE_RX_10G)
-+ if (p_FmPort->p_FmPortDriverParam->dfltCfg.tx_fifo_deq_pipeline_depth
-+ > 2)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("fifoDeqPipelineDepth for IM 10G can't be larger than 2"));
-+
-+ if ((ans = FmPortImCheckInitParameters(p_FmPort)) != E_OK)
-+ return ERROR_CODE(ans);
-+ }
-+ else
-+ {
-+ /****************************************/
-+ /* Rx only */
-+ /****************************************/
-+ if ((p_FmPort->portType == e_FM_PORT_TYPE_RX)
-+ || (p_FmPort->portType == e_FM_PORT_TYPE_RX_10G))
-+ {
-+ /* external buffer pools */
-+ if (!p_Params->extBufPools.numOfPoolsUsed)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("extBufPools.numOfPoolsUsed=0. At least one buffer pool must be defined"));
-+
-+ if (FmSpCheckBufPoolsParams(&p_Params->extBufPools,
-+ p_Params->p_BackupBmPools,
-+ &p_Params->bufPoolDepletion) != E_OK)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, NO_MSG);
-+
-+ /* Check that part of IC that needs copying is small enough to enter start margin */
-+ if (p_Params->intContext.size
-+ && (p_Params->intContext.size
-+ + p_Params->intContext.extBufOffset
-+ > p_Params->bufMargins.startMargins))
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE,
-+ ("intContext.size is larger than start margins"));
-+
-+ if ((p_Params->liodnOffset != (uint16_t)DPAA_LIODN_DONT_OVERRIDE)
-+ && (p_Params->liodnOffset & ~FM_LIODN_OFFSET_MASK))
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("liodnOffset is larger than %d", FM_LIODN_OFFSET_MASK+1));
-+
-+#ifdef FM_NO_BACKUP_POOLS
-+ if ((p_FmPort->fmRevInfo.majorRev != 4) && (p_FmPort->fmRevInfo.majorRev < 6))
-+ if (p_FmPort->p_FmPortDriverParam->p_BackupBmPools)
-+ RETURN_ERROR(MAJOR, E_NOT_SUPPORTED, ("BackupBmPools"));
-+#endif /* FM_NO_BACKUP_POOLS */
-+ }
-+
-+ /****************************************/
-+ /* Non Rx ports */
-+ /****************************************/
-+ else
-+ {
-+ if (p_Params->deqSubPortal >= FM_MAX_NUM_OF_SUB_PORTALS)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ (" deqSubPortal has to be in the range of 0 - %d", FM_MAX_NUM_OF_SUB_PORTALS));
-+
-+ /* to protect HW internal-context from overwrite */
-+ if ((p_Params->intContext.size)
-+ && (p_Params->intContext.intContextOffset
-+ < MIN_TX_INT_OFFSET))
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("non-Rx intContext.intContextOffset can't be smaller than %d", MIN_TX_INT_OFFSET));
-+
-+ if ((p_FmPort->portType == e_FM_PORT_TYPE_TX)
-+ || (p_FmPort->portType == e_FM_PORT_TYPE_TX_10G)
-+ /* in O/H DEFAULT_notSupported indicates that it is not supported and should not be checked */
-+ || (p_FmPort->p_FmPortDriverParam->dfltCfg.tx_fifo_deq_pipeline_depth
-+ != DEFAULT_notSupported))
-+ {
-+ /* Check that not larger than 8 */
-+ if ((!p_FmPort->p_FmPortDriverParam->dfltCfg.tx_fifo_deq_pipeline_depth)
-+ || (p_FmPort->p_FmPortDriverParam->dfltCfg.tx_fifo_deq_pipeline_depth
-+ > MAX_FIFO_PIPELINE_DEPTH))
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("fifoDeqPipelineDepth can't be larger than %d", MAX_FIFO_PIPELINE_DEPTH));
-+ }
-+ }
-+
-+ /****************************************/
-+ /* Rx Or Offline Parsing */
-+ /****************************************/
-+ if ((p_FmPort->portType == e_FM_PORT_TYPE_RX)
-+ || (p_FmPort->portType == e_FM_PORT_TYPE_RX_10G)
-+ || (p_FmPort->portType == e_FM_PORT_TYPE_OH_OFFLINE_PARSING))
-+ {
-+ if (!p_Params->dfltFqid)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE,
-+ ("dfltFqid must be between 1 and 2^24-1"));
-+#if defined(FM_CAPWAP_SUPPORT) && defined(FM_LOCKUP_ALIGNMENT_ERRATA_FMAN_SW004)
-+ if (p_FmPort->p_FmPortDriverParam->bufferPrefixContent.manipExtraSpace % 16)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("bufferPrefixContent.manipExtraSpace has to be devidable by 16"));
-+#endif /* defined(FM_CAPWAP_SUPPORT) && ... */
-+ }
-+
-+ /****************************************/
-+ /* All ports */
-+ /****************************************/
-+ /* common BMI registers values */
-+ /* Check that Queue Id is not larger than 2^24, and is not 0 */
-+ if ((p_Params->errFqid & ~0x00FFFFFF) || !p_Params->errFqid)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE,
-+ ("errFqid must be between 1 and 2^24-1"));
-+ if (p_Params->dfltFqid & ~0x00FFFFFF)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE,
-+ ("dfltFqid must be between 1 and 2^24-1"));
-+ }
-+
-+ /****************************************/
-+ /* Rx only */
-+ /****************************************/
-+ if ((p_FmPort->portType == e_FM_PORT_TYPE_RX)
-+ || (p_FmPort->portType == e_FM_PORT_TYPE_RX_10G))
-+ {
-+ if (p_DfltConfig->rx_pri_elevation % BMI_FIFO_UNITS)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("rxFifoPriElevationLevel has to be divisible by %d", BMI_FIFO_UNITS));
-+ if ((p_DfltConfig->rx_pri_elevation < BMI_FIFO_UNITS)
-+ || (p_DfltConfig->rx_pri_elevation > MAX_PORT_FIFO_SIZE))
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("rxFifoPriElevationLevel has to be in the range of 256 - %d", MAX_PORT_FIFO_SIZE));
-+ if (p_DfltConfig->rx_fifo_thr % BMI_FIFO_UNITS)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("rxFifoThreshold has to be divisible by %d", BMI_FIFO_UNITS));
-+ if ((p_DfltConfig->rx_fifo_thr < BMI_FIFO_UNITS)
-+ || (p_DfltConfig->rx_fifo_thr > MAX_PORT_FIFO_SIZE))
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("rxFifoThreshold has to be in the range of 256 - %d", MAX_PORT_FIFO_SIZE));
-+
-+ /* Check that not larger than 16 */
-+ if (p_DfltConfig->rx_cut_end_bytes > FRAME_END_DATA_SIZE)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("cutBytesFromEnd can't be larger than %d", FRAME_END_DATA_SIZE));
-+
-+ if (FmSpCheckBufMargins(&p_Params->bufMargins) != E_OK)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, NO_MSG);
-+
-+ /* extra FIFO size (allowed only to Rx ports) */
-+ if (p_Params->setSizeOfFifo
-+ && (p_FmPort->fifoBufs.extra % BMI_FIFO_UNITS))
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("fifoBufs.extra has to be divisible by %d", BMI_FIFO_UNITS));
-+
-+ if (p_Params->bufPoolDepletion.poolsGrpModeEnable
-+ && !p_Params->bufPoolDepletion.numOfPools)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("bufPoolDepletion.numOfPools can not be 0 when poolsGrpModeEnable=TRUE"));
-+#ifdef FM_CSI_CFED_LIMIT
-+ if (p_FmPort->fmRevInfo.majorRev == 4)
-+ {
-+ /* Check that not larger than 16 */
-+ if (p_DfltConfig->rx_cut_end_bytes + p_DfltConfig->checksum_bytes_ignore > FRAME_END_DATA_SIZE)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("cheksumLastBytesIgnore + cutBytesFromEnd can't be larger than %d", FRAME_END_DATA_SIZE));
-+ }
-+#endif /* FM_CSI_CFED_LIMIT */
-+ }
-+
-+ /****************************************/
-+ /* Non Rx ports */
-+ /****************************************/
-+ /* extra FIFO size (allowed only to Rx ports) */
-+ else
-+ if (p_FmPort->fifoBufs.extra)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE,
-+ (" No fifoBufs.extra for non Rx ports"));
-+
-+ /****************************************/
-+ /* Tx only */
-+ /****************************************/
-+ if ((p_FmPort->portType == e_FM_PORT_TYPE_TX)
-+ || (p_FmPort->portType == e_FM_PORT_TYPE_TX_10G))
-+ {
-+ if (p_DfltConfig->tx_fifo_min_level % BMI_FIFO_UNITS)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("txFifoMinFillLevel has to be divisible by %d", BMI_FIFO_UNITS));
-+ if (p_DfltConfig->tx_fifo_min_level > (MAX_PORT_FIFO_SIZE - 256))
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("txFifoMinFillLevel has to be in the range of 0 - %d", (MAX_PORT_FIFO_SIZE - 256)));
-+ if (p_DfltConfig->tx_fifo_low_comf_level % BMI_FIFO_UNITS)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("txFifoLowComfLevel has to be divisible by %d", BMI_FIFO_UNITS));
-+ if ((p_DfltConfig->tx_fifo_low_comf_level < BMI_FIFO_UNITS)
-+ || (p_DfltConfig->tx_fifo_low_comf_level > MAX_PORT_FIFO_SIZE))
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("txFifoLowComfLevel has to be in the range of 256 - %d", MAX_PORT_FIFO_SIZE));
-+
-+ if (p_FmPort->portType == e_FM_PORT_TYPE_TX)
-+ if (p_FmPort->p_FmPortDriverParam->dfltCfg.tx_fifo_deq_pipeline_depth
-+ > 2)
-+ RETURN_ERROR(
-+ MAJOR, E_INVALID_VALUE,
-+ ("fifoDeqPipelineDepth for 1G can't be larger than 2"));
-+ }
-+
-+ /****************************************/
-+ /* Non Tx Ports */
-+ /****************************************/
-+ /* If discard override was selected , no frames may be discarded. */
-+ else
-+ if (p_DfltConfig->discard_override && p_Params->errorsToDiscard)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_CONFLICT,
-+ ("errorsToDiscard is not empty, but frmDiscardOverride selected (all discarded frames to be enqueued to error queue)."));
-+
-+ /****************************************/
-+ /* Rx and Offline parsing */
-+ /****************************************/
-+ if ((p_FmPort->portType == e_FM_PORT_TYPE_RX)
-+ || (p_FmPort->portType == e_FM_PORT_TYPE_RX_10G)
-+ || (p_FmPort->portType == e_FM_PORT_TYPE_OH_OFFLINE_PARSING))
-+ {
-+ if (p_FmPort->portType == e_FM_PORT_TYPE_OH_OFFLINE_PARSING)
-+ unusedMask = BMI_STATUS_OP_MASK_UNUSED;
-+ else
-+ unusedMask = BMI_STATUS_RX_MASK_UNUSED;
-+
-+ /* Check that no common bits with BMI_STATUS_MASK_UNUSED */
-+ if (p_Params->errorsToDiscard & unusedMask)
-+ RETURN_ERROR(MAJOR, E_INVALID_SELECTION,
-+ ("errorsToDiscard contains undefined bits"));
-+ }
-+
-+ /****************************************/
-+ /* Offline Ports */
-+ /****************************************/
-+#ifdef FM_OP_OPEN_DMA_MIN_LIMIT
-+ if ((p_FmPort->fmRevInfo.majorRev >= 6)
-+ && (p_FmPort->portType == e_FM_PORT_TYPE_OH_OFFLINE_PARSING)
-+ && p_Params->setNumOfOpenDmas
-+ && (p_FmPort->openDmas.num < MIN_NUM_OF_OP_DMAS))
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("For Offline port, openDmas.num can't be smaller than %d", MIN_NUM_OF_OP_DMAS));
-+#endif /* FM_OP_OPEN_DMA_MIN_LIMIT */
-+
-+ /****************************************/
-+ /* Offline & HC Ports */
-+ /****************************************/
-+ if ((p_FmPort->portType == e_FM_PORT_TYPE_OH_OFFLINE_PARSING)
-+ || (p_FmPort->portType == e_FM_PORT_TYPE_OH_HOST_COMMAND))
-+ {
-+#ifndef FM_FRAME_END_PARAMS_FOR_OP
-+ if ((p_FmPort->fmRevInfo.majorRev < 6) &&
-+ (p_FmPort->p_FmPortDriverParam->cheksumLastBytesIgnore != DEFAULT_notSupported))
-+ /* this is an indication that user called config for this mode which is not supported in this integration */
-+ RETURN_ERROR(MAJOR, E_NOT_SUPPORTED, ("cheksumLastBytesIgnore is available for Rx & Tx ports only"));
-+#endif /* !FM_FRAME_END_PARAMS_FOR_OP */
-+
-+#ifndef FM_DEQ_PIPELINE_PARAMS_FOR_OP
-+ if ((!((p_FmPort->fmRevInfo.majorRev == 4) ||
-+ (p_FmPort->fmRevInfo.majorRev >= 6))) &&
-+ (p_FmPort->p_FmPortDriverParam->dfltCfg.tx_fifo_deq_pipeline_depth != DEFAULT_notSupported))
-+ /* this is an indication that user called config for this mode which is not supported in this integration */
-+ RETURN_ERROR(MAJOR, E_INVALID_OPERATION, ("fifoDeqPipelineDepth is available for Tx ports only"));
-+#endif /* !FM_DEQ_PIPELINE_PARAMS_FOR_OP */
-+ }
-+
-+ /****************************************/
-+ /* All ports */
-+ /****************************************/
-+ /* Check that not larger than 16 */
-+ if ((p_Params->cheksumLastBytesIgnore > FRAME_END_DATA_SIZE)
-+ && ((p_Params->cheksumLastBytesIgnore != DEFAULT_notSupported)))
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("cheksumLastBytesIgnore can't be larger than %d", FRAME_END_DATA_SIZE));
-+
-+ if (FmSpCheckIntContextParams(&p_Params->intContext) != E_OK)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, NO_MSG);
-+
-+ /* common BMI registers values */
-+ if (p_Params->setNumOfTasks
-+ && ((!p_FmPort->tasks.num)
-+ || (p_FmPort->tasks.num > MAX_NUM_OF_TASKS)))
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE,
-+ ("tasks.num can't be larger than %d", MAX_NUM_OF_TASKS));
-+ if (p_Params->setNumOfTasks
-+ && (p_FmPort->tasks.extra > MAX_NUM_OF_EXTRA_TASKS))
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("tasks.extra can't be larger than %d", MAX_NUM_OF_EXTRA_TASKS));
-+ if (p_Params->setNumOfOpenDmas
-+ && ((!p_FmPort->openDmas.num)
-+ || (p_FmPort->openDmas.num > MAX_NUM_OF_DMAS)))
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE,
-+ ("openDmas.num can't be larger than %d", MAX_NUM_OF_DMAS));
-+ if (p_Params->setNumOfOpenDmas
-+ && (p_FmPort->openDmas.extra > MAX_NUM_OF_EXTRA_DMAS))
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("openDmas.extra can't be larger than %d", MAX_NUM_OF_EXTRA_DMAS));
-+ if (p_Params->setSizeOfFifo
-+ && (!p_FmPort->fifoBufs.num
-+ || (p_FmPort->fifoBufs.num > MAX_PORT_FIFO_SIZE)))
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("fifoBufs.num has to be in the range of 256 - %d", MAX_PORT_FIFO_SIZE));
-+ if (p_Params->setSizeOfFifo && (p_FmPort->fifoBufs.num % BMI_FIFO_UNITS))
-+ RETURN_ERROR(
-+ MAJOR, E_INVALID_VALUE,
-+ ("fifoBufs.num has to be divisible by %d", BMI_FIFO_UNITS));
-+
-+#ifdef FM_QMI_NO_DEQ_OPTIONS_SUPPORT
-+ if (p_FmPort->fmRevInfo.majorRev == 4)
-+ if (p_FmPort->p_FmPortDriverParam->deqPrefetchOption != DEFAULT_notSupported)
-+ /* this is an indication that user called config for this mode which is not supported in this integration */
-+ RETURN_ERROR(MAJOR, E_INVALID_OPERATION, ("deqPrefetchOption"));
-+#endif /* FM_QMI_NO_DEQ_OPTIONS_SUPPORT */
-+
-+ return E_OK;
-+}
-+
-+static t_Error VerifySizeOfFifo(t_FmPort *p_FmPort)
-+{
-+ uint32_t minFifoSizeRequired = 0, optFifoSizeForB2B = 0;
-+
-+ /*************************/
-+ /* TX PORTS */
-+ /*************************/
-+ if ((p_FmPort->portType == e_FM_PORT_TYPE_TX)
-+ || (p_FmPort->portType == e_FM_PORT_TYPE_TX_10G))
-+ {
-+ minFifoSizeRequired =
-+ (uint32_t)(ROUND_UP(p_FmPort->maxFrameLength, BMI_FIFO_UNITS)
-+ + (3 * BMI_FIFO_UNITS));
-+ if (!p_FmPort->imEn)
-+ minFifoSizeRequired +=
-+ p_FmPort->p_FmPortDriverParam->dfltCfg.tx_fifo_deq_pipeline_depth
-+ * BMI_FIFO_UNITS;
-+
-+ optFifoSizeForB2B = minFifoSizeRequired;
-+
-+ /* Add some margin for back-to-back capability to improve performance,
-+ allows the hardware to pipeline new frame dma while the previous
-+ frame not yet transmitted. */
-+ if (p_FmPort->portType == e_FM_PORT_TYPE_TX_10G)
-+ optFifoSizeForB2B += 3 * BMI_FIFO_UNITS;
-+ else
-+ optFifoSizeForB2B += 2 * BMI_FIFO_UNITS;
-+ }
-+
-+ /*************************/
-+ /* RX IM PORTS */
-+ /*************************/
-+ else
-+ if (((p_FmPort->portType == e_FM_PORT_TYPE_RX)
-+ || (p_FmPort->portType == e_FM_PORT_TYPE_RX_10G))
-+ && p_FmPort->imEn)
-+ {
-+ optFifoSizeForB2B =
-+ minFifoSizeRequired =
-+ (uint32_t)(ROUND_UP(p_FmPort->maxFrameLength, BMI_FIFO_UNITS)
-+ + (4 * BMI_FIFO_UNITS));
-+ }
-+
-+ /*************************/
-+ /* RX non-IM PORTS */
-+ /*************************/
-+ else
-+ if (((p_FmPort->portType == e_FM_PORT_TYPE_RX)
-+ || (p_FmPort->portType == e_FM_PORT_TYPE_RX_10G))
-+ && !p_FmPort->imEn)
-+ {
-+ if (p_FmPort->fmRevInfo.majorRev == 4)
-+ {
-+ if (p_FmPort->rxPoolsParams.numOfPools == 1)
-+ minFifoSizeRequired = 8 * BMI_FIFO_UNITS;
-+ else
-+ minFifoSizeRequired =
-+ (uint32_t)(ROUND_UP(p_FmPort->rxPoolsParams.secondLargestBufSize, BMI_FIFO_UNITS)
-+ + (7 * BMI_FIFO_UNITS));
-+ }
-+ else
-+ {
-+#if (DPAA_VERSION >= 11)
-+ minFifoSizeRequired =
-+ (uint32_t)(ROUND_UP(p_FmPort->maxFrameLength, BMI_FIFO_UNITS)
-+ + (5 * BMI_FIFO_UNITS));
-+ /* 4 according to spec + 1 for FOF>0 */
-+#else
-+ minFifoSizeRequired = (uint32_t)
-+ (ROUND_UP(MIN(p_FmPort->maxFrameLength, p_FmPort->rxPoolsParams.largestBufSize), BMI_FIFO_UNITS)
-+ + (7*BMI_FIFO_UNITS));
-+#endif /* (DPAA_VERSION >= 11) */
-+ }
-+
-+ optFifoSizeForB2B = minFifoSizeRequired;
-+
-+ /* Add some margin for back-to-back capability to improve performance,
-+ allows the hardware to pipeline new frame dma while the previous
-+ frame not yet transmitted. */
-+ if (p_FmPort->portType == e_FM_PORT_TYPE_RX_10G)
-+ optFifoSizeForB2B += 8 * BMI_FIFO_UNITS;
-+ else
-+ optFifoSizeForB2B += 3 * BMI_FIFO_UNITS;
-+ }
-+
-+ /* For O/H ports, check fifo size and update if necessary */
-+ else
-+ if ((p_FmPort->portType == e_FM_PORT_TYPE_OH_OFFLINE_PARSING)
-+ || (p_FmPort->portType == e_FM_PORT_TYPE_OH_HOST_COMMAND))
-+ {
-+#if (DPAA_VERSION >= 11)
-+ optFifoSizeForB2B =
-+ minFifoSizeRequired =
-+ (uint32_t)(ROUND_UP(p_FmPort->maxFrameLength, BMI_FIFO_UNITS)
-+ + ((p_FmPort->p_FmPortDriverParam->dfltCfg.tx_fifo_deq_pipeline_depth
-+ + 5) * BMI_FIFO_UNITS));
-+ /* 4 according to spec + 1 for FOF>0 */
-+#else
-+ optFifoSizeForB2B = minFifoSizeRequired = (uint32_t)((p_FmPort->tasks.num + 2) * BMI_FIFO_UNITS);
-+#endif /* (DPAA_VERSION >= 11) */
-+ }
-+
-+ ASSERT_COND(minFifoSizeRequired > 0);
-+ ASSERT_COND(optFifoSizeForB2B >= minFifoSizeRequired);
-+
-+ /* Verify the size */
-+ if (p_FmPort->fifoBufs.num < minFifoSizeRequired)
-+ DBG(INFO,
-+ ("FIFO size is %d and should be enlarged to %d bytes",p_FmPort->fifoBufs.num, minFifoSizeRequired));
-+ else if (p_FmPort->fifoBufs.num < optFifoSizeForB2B)
-+ DBG(INFO,
-+ ("For back-to-back frames processing, FIFO size is %d and needs to enlarge to %d bytes", p_FmPort->fifoBufs.num, optFifoSizeForB2B));
-+
-+ return E_OK;
-+}
-+
-+static void FmPortDriverParamFree(t_FmPort *p_FmPort)
-+{
-+ if (p_FmPort->p_FmPortDriverParam)
-+ {
-+ XX_Free(p_FmPort->p_FmPortDriverParam);
-+ p_FmPort->p_FmPortDriverParam = NULL;
-+ }
-+}
-+
-+static t_Error SetExtBufferPools(t_FmPort *p_FmPort)
-+{
-+ t_FmExtPools *p_ExtBufPools = &p_FmPort->p_FmPortDriverParam->extBufPools;
-+ t_FmBufPoolDepletion *p_BufPoolDepletion =
-+ &p_FmPort->p_FmPortDriverParam->bufPoolDepletion;
-+ uint8_t orderedArray[FM_PORT_MAX_NUM_OF_EXT_POOLS];
-+ uint16_t sizesArray[BM_MAX_NUM_OF_POOLS];
-+ int i = 0, j = 0, err;
-+ struct fman_port_bpools bpools;
-+
-+ memset(&orderedArray, 0, sizeof(uint8_t) * FM_PORT_MAX_NUM_OF_EXT_POOLS);
-+ memset(&sizesArray, 0, sizeof(uint16_t) * BM_MAX_NUM_OF_POOLS);
-+ memcpy(&p_FmPort->extBufPools, p_ExtBufPools, sizeof(t_FmExtPools));
-+
-+ FmSpSetBufPoolsInAscOrderOfBufSizes(p_ExtBufPools, orderedArray,
-+ sizesArray);
-+
-+ /* Prepare flibs bpools structure */
-+ memset(&bpools, 0, sizeof(struct fman_port_bpools));
-+ bpools.count = p_ExtBufPools->numOfPoolsUsed;
-+ bpools.counters_enable = TRUE;
-+ for (i = 0; i < p_ExtBufPools->numOfPoolsUsed; i++)
-+ {
-+ bpools.bpool[i].bpid = orderedArray[i];
-+ bpools.bpool[i].size = sizesArray[orderedArray[i]];
-+ /* functionality available only for some derivatives (limited by config) */
-+ if (p_FmPort->p_FmPortDriverParam->p_BackupBmPools)
-+ for (j = 0;
-+ j
-+ < p_FmPort->p_FmPortDriverParam->p_BackupBmPools->numOfBackupPools;
-+ j++)
-+ if (orderedArray[i]
-+ == p_FmPort->p_FmPortDriverParam->p_BackupBmPools->poolIds[j])
-+ {
-+ bpools.bpool[i].is_backup = TRUE;
-+ break;
-+ }
-+ }
-+
-+ /* save pools parameters for later use */
-+ p_FmPort->rxPoolsParams.numOfPools = p_ExtBufPools->numOfPoolsUsed;
-+ p_FmPort->rxPoolsParams.largestBufSize =
-+ sizesArray[orderedArray[p_ExtBufPools->numOfPoolsUsed - 1]];
-+ p_FmPort->rxPoolsParams.secondLargestBufSize =
-+ sizesArray[orderedArray[p_ExtBufPools->numOfPoolsUsed - 2]];
-+
-+ /* FMBM_RMPD reg. - pool depletion */
-+ if (p_BufPoolDepletion->poolsGrpModeEnable)
-+ {
-+ bpools.grp_bp_depleted_num = p_BufPoolDepletion->numOfPools;
-+ for (i = 0; i < BM_MAX_NUM_OF_POOLS; i++)
-+ {
-+ if (p_BufPoolDepletion->poolsToConsider[i])
-+ {
-+ for (j = 0; j < p_ExtBufPools->numOfPoolsUsed; j++)
-+ {
-+ if (i == orderedArray[j])
-+ {
-+ bpools.bpool[j].grp_bp_depleted = TRUE;
-+ break;
-+ }
-+ }
-+ }
-+ }
-+ }
-+
-+ if (p_BufPoolDepletion->singlePoolModeEnable)
-+ {
-+ for (i = 0; i < BM_MAX_NUM_OF_POOLS; i++)
-+ {
-+ if (p_BufPoolDepletion->poolsToConsiderForSingleMode[i])
-+ {
-+ for (j = 0; j < p_ExtBufPools->numOfPoolsUsed; j++)
-+ {
-+ if (i == orderedArray[j])
-+ {
-+ bpools.bpool[j].single_bp_depleted = TRUE;
-+ break;
-+ }
-+ }
-+ }
-+ }
-+ }
-+
-+#if (DPAA_VERSION >= 11)
-+ /* fill QbbPEV */
-+ if (p_BufPoolDepletion->poolsGrpModeEnable
-+ || p_BufPoolDepletion->singlePoolModeEnable)
-+ {
-+ for (i = 0; i < FM_MAX_NUM_OF_PFC_PRIORITIES; i++)
-+ {
-+ if (p_BufPoolDepletion->pfcPrioritiesEn[i] == TRUE)
-+ {
-+ bpools.bpool[i].pfc_priorities_en = TRUE;
-+ }
-+ }
-+ }
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+ /* Issue flibs function */
-+ err = fman_port_set_bpools(&p_FmPort->port, &bpools);
-+ if (err != 0)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("fman_port_set_bpools"));
-+
-+ if (p_FmPort->p_FmPortDriverParam->p_BackupBmPools)
-+ XX_Free(p_FmPort->p_FmPortDriverParam->p_BackupBmPools);
-+
-+ return E_OK;
-+}
-+
-+static t_Error ClearPerfCnts(t_FmPort *p_FmPort)
-+{
-+ if (p_FmPort->portType != e_FM_PORT_TYPE_OH_OFFLINE_PARSING)
-+ FM_PORT_ModifyCounter(p_FmPort, e_FM_PORT_COUNTERS_QUEUE_UTIL, 0);
-+ FM_PORT_ModifyCounter(p_FmPort, e_FM_PORT_COUNTERS_TASK_UTIL, 0);
-+ FM_PORT_ModifyCounter(p_FmPort, e_FM_PORT_COUNTERS_DMA_UTIL, 0);
-+ FM_PORT_ModifyCounter(p_FmPort, e_FM_PORT_COUNTERS_FIFO_UTIL, 0);
-+ return E_OK;
-+}
-+
-+static t_Error InitLowLevelDriver(t_FmPort *p_FmPort)
-+{
-+ t_FmPortDriverParam *p_DriverParams = p_FmPort->p_FmPortDriverParam;
-+ struct fman_port_params portParams;
-+ uint32_t tmpVal;
-+ t_Error err;
-+
-+ /* Set up flibs parameters and issue init function */
-+
-+ memset(&portParams, 0, sizeof(struct fman_port_params));
-+ portParams.discard_mask = p_DriverParams->errorsToDiscard;
-+ portParams.dflt_fqid = p_DriverParams->dfltFqid;
-+ portParams.err_fqid = p_DriverParams->errFqid;
-+ portParams.deq_sp = p_DriverParams->deqSubPortal;
-+ portParams.dont_release_buf = p_DriverParams->dontReleaseBuf;
-+ switch (p_FmPort->portType)
-+ {
-+ case (e_FM_PORT_TYPE_RX_10G):
-+ case (e_FM_PORT_TYPE_RX):
-+ portParams.err_mask = (RX_ERRS_TO_ENQ & ~portParams.discard_mask);
-+ if (!p_FmPort->imEn)
-+ {
-+ if (p_DriverParams->forwardReuseIntContext)
-+ p_DriverParams->dfltCfg.rx_fd_bits =
-+ (uint8_t)(BMI_PORT_RFNE_FRWD_RPD >> 24);
-+ }
-+ break;
-+
-+ case (e_FM_PORT_TYPE_OH_OFFLINE_PARSING):
-+ portParams.err_mask = (OP_ERRS_TO_ENQ & ~portParams.discard_mask);
-+ break;
-+ break;
-+
-+ default:
-+ break;
-+ }
-+
-+ tmpVal =
-+ (uint32_t)(
-+ (p_FmPort->internalBufferOffset % OFFSET_UNITS) ? (p_FmPort->internalBufferOffset
-+ / OFFSET_UNITS + 1) :
-+ (p_FmPort->internalBufferOffset / OFFSET_UNITS));
-+ p_FmPort->internalBufferOffset = (uint8_t)(tmpVal * OFFSET_UNITS);
-+ p_DriverParams->dfltCfg.int_buf_start_margin =
-+ p_FmPort->internalBufferOffset;
-+
-+ p_DriverParams->dfltCfg.ext_buf_start_margin =
-+ p_DriverParams->bufMargins.startMargins;
-+ p_DriverParams->dfltCfg.ext_buf_end_margin =
-+ p_DriverParams->bufMargins.endMargins;
-+
-+ p_DriverParams->dfltCfg.ic_ext_offset =
-+ p_DriverParams->intContext.extBufOffset;
-+ p_DriverParams->dfltCfg.ic_int_offset =
-+ p_DriverParams->intContext.intContextOffset;
-+ p_DriverParams->dfltCfg.ic_size = p_DriverParams->intContext.size;
-+
-+ p_DriverParams->dfltCfg.stats_counters_enable = TRUE;
-+ p_DriverParams->dfltCfg.perf_counters_enable = TRUE;
-+ p_DriverParams->dfltCfg.queue_counters_enable = TRUE;
-+
-+ p_DriverParams->dfltCfg.perf_cnt_params.task_val =
-+ (uint8_t)p_FmPort->tasks.num;
-+ if (p_FmPort->portType == e_FM_PORT_TYPE_OH_OFFLINE_PARSING ||
-+ p_FmPort->portType == e_FM_PORT_TYPE_OH_HOST_COMMAND)p_DriverParams->dfltCfg.perf_cnt_params.queue_val = 0;
-+ else
-+ p_DriverParams->dfltCfg.perf_cnt_params.queue_val = 1;
-+ p_DriverParams->dfltCfg.perf_cnt_params.dma_val =
-+ (uint8_t)p_FmPort->openDmas.num;
-+ p_DriverParams->dfltCfg.perf_cnt_params.fifo_val = p_FmPort->fifoBufs.num;
-+
-+ if (0
-+ != fman_port_init(&p_FmPort->port, &p_DriverParams->dfltCfg,
-+ &portParams))
-+ RETURN_ERROR(MAJOR, E_NO_DEVICE, ("fman_port_init"));
-+
-+ if (p_FmPort->imEn && ((err = FmPortImInit(p_FmPort)) != E_OK))
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ else
-+ {
-+ // from QMIInit
-+ if ((p_FmPort->portType != e_FM_PORT_TYPE_RX_10G)
-+ && (p_FmPort->portType != e_FM_PORT_TYPE_RX))
-+ {
-+ if (p_DriverParams->deqPrefetchOption == e_FM_PORT_DEQ_NO_PREFETCH)
-+ FmSetPortPreFetchConfiguration(p_FmPort->h_Fm, p_FmPort->portId,
-+ FALSE);
-+ else
-+ FmSetPortPreFetchConfiguration(p_FmPort->h_Fm, p_FmPort->portId,
-+ TRUE);
-+ }
-+ }
-+ /* The code bellow is a trick so the FM will not release the buffer
-+ to BM nor will try to enqueue the frame to QM */
-+ if (((p_FmPort->portType == e_FM_PORT_TYPE_TX_10G)
-+ || (p_FmPort->portType == e_FM_PORT_TYPE_TX)) && (!p_FmPort->imEn))
-+ {
-+ if (!p_DriverParams->dfltFqid && p_DriverParams->dontReleaseBuf)
-+ {
-+ /* override fmbm_tcfqid 0 with a false non-0 value. This will force FM to
-+ * act according to tfene. Otherwise, if fmbm_tcfqid is 0 the FM will release
-+ * buffers to BM regardless of fmbm_tfene
-+ */
-+ WRITE_UINT32(p_FmPort->port.bmi_regs->tx.fmbm_tcfqid, 0xFFFFFF);
-+ WRITE_UINT32(p_FmPort->port.bmi_regs->tx.fmbm_tfene,
-+ NIA_ENG_BMI | NIA_BMI_AC_TX_RELEASE);
-+ }
-+ }
-+
-+ return E_OK;
-+}
-+
-+static bool CheckRxBmiCounter(t_FmPort *p_FmPort, e_FmPortCounters counter)
-+{
-+ UNUSED(p_FmPort);
-+
-+ switch (counter)
-+ {
-+ case (e_FM_PORT_COUNTERS_CYCLE):
-+ case (e_FM_PORT_COUNTERS_TASK_UTIL):
-+ case (e_FM_PORT_COUNTERS_QUEUE_UTIL):
-+ case (e_FM_PORT_COUNTERS_DMA_UTIL):
-+ case (e_FM_PORT_COUNTERS_FIFO_UTIL):
-+ case (e_FM_PORT_COUNTERS_RX_PAUSE_ACTIVATION):
-+ case (e_FM_PORT_COUNTERS_FRAME):
-+ case (e_FM_PORT_COUNTERS_DISCARD_FRAME):
-+ case (e_FM_PORT_COUNTERS_RX_BAD_FRAME):
-+ case (e_FM_PORT_COUNTERS_RX_LARGE_FRAME):
-+ case (e_FM_PORT_COUNTERS_RX_FILTER_FRAME):
-+ case (e_FM_PORT_COUNTERS_RX_LIST_DMA_ERR):
-+ case (e_FM_PORT_COUNTERS_RX_OUT_OF_BUFFERS_DISCARD):
-+ case (e_FM_PORT_COUNTERS_DEALLOC_BUF):
-+ case (e_FM_PORT_COUNTERS_PREPARE_TO_ENQUEUE_COUNTER):
-+ return TRUE;
-+ default:
-+ return FALSE;
-+ }
-+}
-+
-+static bool CheckTxBmiCounter(t_FmPort *p_FmPort, e_FmPortCounters counter)
-+{
-+ UNUSED(p_FmPort);
-+
-+ switch (counter)
-+ {
-+ case (e_FM_PORT_COUNTERS_CYCLE):
-+ case (e_FM_PORT_COUNTERS_TASK_UTIL):
-+ case (e_FM_PORT_COUNTERS_QUEUE_UTIL):
-+ case (e_FM_PORT_COUNTERS_DMA_UTIL):
-+ case (e_FM_PORT_COUNTERS_FIFO_UTIL):
-+ case (e_FM_PORT_COUNTERS_FRAME):
-+ case (e_FM_PORT_COUNTERS_DISCARD_FRAME):
-+ case (e_FM_PORT_COUNTERS_LENGTH_ERR):
-+ case (e_FM_PORT_COUNTERS_UNSUPPRTED_FORMAT):
-+ case (e_FM_PORT_COUNTERS_DEALLOC_BUF):
-+ return TRUE;
-+ default:
-+ return FALSE;
-+ }
-+}
-+
-+static bool CheckOhBmiCounter(t_FmPort *p_FmPort, e_FmPortCounters counter)
-+{
-+ switch (counter)
-+ {
-+ case (e_FM_PORT_COUNTERS_CYCLE):
-+ case (e_FM_PORT_COUNTERS_TASK_UTIL):
-+ case (e_FM_PORT_COUNTERS_DMA_UTIL):
-+ case (e_FM_PORT_COUNTERS_FIFO_UTIL):
-+ case (e_FM_PORT_COUNTERS_FRAME):
-+ case (e_FM_PORT_COUNTERS_DISCARD_FRAME):
-+ case (e_FM_PORT_COUNTERS_RX_LIST_DMA_ERR):
-+ case (e_FM_PORT_COUNTERS_WRED_DISCARD):
-+ case (e_FM_PORT_COUNTERS_LENGTH_ERR):
-+ case (e_FM_PORT_COUNTERS_UNSUPPRTED_FORMAT):
-+ case (e_FM_PORT_COUNTERS_DEALLOC_BUF):
-+ return TRUE;
-+ case (e_FM_PORT_COUNTERS_RX_FILTER_FRAME):
-+ if (p_FmPort->portType == e_FM_PORT_TYPE_OH_HOST_COMMAND)
-+ return FALSE;
-+ else
-+ return TRUE;
-+ default:
-+ return FALSE;
-+ }
-+}
-+
-+static t_Error BmiPortCheckAndGetCounterType(
-+ t_FmPort *p_FmPort, e_FmPortCounters counter,
-+ enum fman_port_stats_counters *p_StatsType,
-+ enum fman_port_perf_counters *p_PerfType, bool *p_IsStats)
-+{
-+ volatile uint32_t *p_Reg;
-+ bool isValid;
-+
-+ switch (p_FmPort->portType)
-+ {
-+ case (e_FM_PORT_TYPE_RX_10G):
-+ case (e_FM_PORT_TYPE_RX):
-+ p_Reg = &p_FmPort->port.bmi_regs->rx.fmbm_rstc;
-+ isValid = CheckRxBmiCounter(p_FmPort, counter);
-+ break;
-+ case (e_FM_PORT_TYPE_TX_10G):
-+ case (e_FM_PORT_TYPE_TX):
-+ p_Reg = &p_FmPort->port.bmi_regs->tx.fmbm_tstc;
-+ isValid = CheckTxBmiCounter(p_FmPort, counter);
-+ break;
-+ case (e_FM_PORT_TYPE_OH_OFFLINE_PARSING):
-+ case (e_FM_PORT_TYPE_OH_HOST_COMMAND):
-+ p_Reg = &p_FmPort->port.bmi_regs->oh.fmbm_ostc;
-+ isValid = CheckOhBmiCounter(p_FmPort, counter);
-+ break;
-+ default:
-+ RETURN_ERROR(MINOR, E_INVALID_STATE, ("Unsupported port type"));
-+ }
-+
-+ if (!isValid)
-+ RETURN_ERROR(MINOR, E_INVALID_STATE,
-+ ("Requested counter is not available for this port type"));
-+
-+ /* check that counters are enabled */
-+ switch (counter)
-+ {
-+ case (e_FM_PORT_COUNTERS_CYCLE):
-+ case (e_FM_PORT_COUNTERS_TASK_UTIL):
-+ case (e_FM_PORT_COUNTERS_QUEUE_UTIL):
-+ case (e_FM_PORT_COUNTERS_DMA_UTIL):
-+ case (e_FM_PORT_COUNTERS_FIFO_UTIL):
-+ case (e_FM_PORT_COUNTERS_RX_PAUSE_ACTIVATION):
-+ /* performance counters - may be read when disabled */
-+ *p_IsStats = FALSE;
-+ break;
-+ case (e_FM_PORT_COUNTERS_FRAME):
-+ case (e_FM_PORT_COUNTERS_DISCARD_FRAME):
-+ case (e_FM_PORT_COUNTERS_DEALLOC_BUF):
-+ case (e_FM_PORT_COUNTERS_RX_BAD_FRAME):
-+ case (e_FM_PORT_COUNTERS_RX_LARGE_FRAME):
-+ case (e_FM_PORT_COUNTERS_RX_FILTER_FRAME):
-+ case (e_FM_PORT_COUNTERS_RX_LIST_DMA_ERR):
-+ case (e_FM_PORT_COUNTERS_RX_OUT_OF_BUFFERS_DISCARD):
-+ case (e_FM_PORT_COUNTERS_LENGTH_ERR):
-+ case (e_FM_PORT_COUNTERS_UNSUPPRTED_FORMAT):
-+ case (e_FM_PORT_COUNTERS_WRED_DISCARD):
-+ *p_IsStats = TRUE;
-+ if (!(GET_UINT32(*p_Reg) & BMI_COUNTERS_EN))
-+ RETURN_ERROR(MINOR, E_INVALID_STATE,
-+ ("Requested counter was not enabled"));
-+ break;
-+ default:
-+ break;
-+ }
-+
-+ /* Set counter */
-+ switch (counter)
-+ {
-+ case (e_FM_PORT_COUNTERS_CYCLE):
-+ *p_PerfType = E_FMAN_PORT_PERF_CNT_CYCLE;
-+ break;
-+ case (e_FM_PORT_COUNTERS_TASK_UTIL):
-+ *p_PerfType = E_FMAN_PORT_PERF_CNT_TASK_UTIL;
-+ break;
-+ case (e_FM_PORT_COUNTERS_QUEUE_UTIL):
-+ *p_PerfType = E_FMAN_PORT_PERF_CNT_QUEUE_UTIL;
-+ break;
-+ case (e_FM_PORT_COUNTERS_DMA_UTIL):
-+ *p_PerfType = E_FMAN_PORT_PERF_CNT_DMA_UTIL;
-+ break;
-+ case (e_FM_PORT_COUNTERS_FIFO_UTIL):
-+ *p_PerfType = E_FMAN_PORT_PERF_CNT_FIFO_UTIL;
-+ break;
-+ case (e_FM_PORT_COUNTERS_RX_PAUSE_ACTIVATION):
-+ *p_PerfType = E_FMAN_PORT_PERF_CNT_RX_PAUSE;
-+ break;
-+ case (e_FM_PORT_COUNTERS_FRAME):
-+ *p_StatsType = E_FMAN_PORT_STATS_CNT_FRAME;
-+ break;
-+ case (e_FM_PORT_COUNTERS_DISCARD_FRAME):
-+ *p_StatsType = E_FMAN_PORT_STATS_CNT_DISCARD;
-+ break;
-+ case (e_FM_PORT_COUNTERS_DEALLOC_BUF):
-+ *p_StatsType = E_FMAN_PORT_STATS_CNT_DEALLOC_BUF;
-+ break;
-+ case (e_FM_PORT_COUNTERS_RX_BAD_FRAME):
-+ *p_StatsType = E_FMAN_PORT_STATS_CNT_RX_BAD_FRAME;
-+ break;
-+ case (e_FM_PORT_COUNTERS_RX_LARGE_FRAME):
-+ *p_StatsType = E_FMAN_PORT_STATS_CNT_RX_LARGE_FRAME;
-+ break;
-+ case (e_FM_PORT_COUNTERS_RX_OUT_OF_BUFFERS_DISCARD):
-+ *p_StatsType = E_FMAN_PORT_STATS_CNT_RX_OUT_OF_BUF;
-+ break;
-+ case (e_FM_PORT_COUNTERS_RX_FILTER_FRAME):
-+ *p_StatsType = E_FMAN_PORT_STATS_CNT_FILTERED_FRAME;
-+ break;
-+ case (e_FM_PORT_COUNTERS_RX_LIST_DMA_ERR):
-+ *p_StatsType = E_FMAN_PORT_STATS_CNT_DMA_ERR;
-+ break;
-+ case (e_FM_PORT_COUNTERS_WRED_DISCARD):
-+ *p_StatsType = E_FMAN_PORT_STATS_CNT_WRED_DISCARD;
-+ break;
-+ case (e_FM_PORT_COUNTERS_LENGTH_ERR):
-+ *p_StatsType = E_FMAN_PORT_STATS_CNT_LEN_ERR;
-+ break;
-+ case (e_FM_PORT_COUNTERS_UNSUPPRTED_FORMAT):
-+ *p_StatsType = E_FMAN_PORT_STATS_CNT_UNSUPPORTED_FORMAT;
-+ break;
-+ default:
-+ break;
-+ }
-+
-+ return E_OK;
-+}
-+
-+static t_Error AdditionalPrsParams(t_FmPort *p_FmPort,
-+ t_FmPcdPrsAdditionalHdrParams *p_HdrParams,
-+ uint32_t *p_SoftSeqAttachReg)
-+{
-+ uint8_t hdrNum, Ipv4HdrNum;
-+ u_FmPcdHdrPrsOpts *p_prsOpts;
-+ uint32_t tmpReg = *p_SoftSeqAttachReg, tmpPrsOffset;
-+
-+ if (IS_PRIVATE_HEADER(p_HdrParams->hdr)
-+ || IS_SPECIAL_HEADER(p_HdrParams->hdr))
-+ RETURN_ERROR(
-+ MAJOR, E_NOT_SUPPORTED,
-+ ("No additional parameters for private or special headers."));
-+
-+ if (p_HdrParams->errDisable)
-+ tmpReg |= PRS_HDR_ERROR_DIS;
-+
-+ /* Set parser options */
-+ if (p_HdrParams->usePrsOpts)
-+ {
-+ p_prsOpts = &p_HdrParams->prsOpts;
-+ switch (p_HdrParams->hdr)
-+ {
-+ case (HEADER_TYPE_MPLS):
-+ if (p_prsOpts->mplsPrsOptions.labelInterpretationEnable)
-+ tmpReg |= PRS_HDR_MPLS_LBL_INTER_EN;
-+ hdrNum = GetPrsHdrNum(p_prsOpts->mplsPrsOptions.nextParse);
-+ if (hdrNum == ILLEGAL_HDR_NUM)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, NO_MSG);
-+ Ipv4HdrNum = GetPrsHdrNum(HEADER_TYPE_IPv4);
-+ if (hdrNum < Ipv4HdrNum)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE,
-+ ("Header must be equal or higher than IPv4"));
-+ tmpReg |= ((uint32_t)hdrNum * PRS_HDR_ENTRY_SIZE)
-+ << PRS_HDR_MPLS_NEXT_HDR_SHIFT;
-+ break;
-+ case (HEADER_TYPE_PPPoE):
-+ if (p_prsOpts->pppoePrsOptions.enableMTUCheck)
-+ tmpReg |= PRS_HDR_PPPOE_MTU_CHECK_EN;
-+ break;
-+ case (HEADER_TYPE_IPv6):
-+ if (p_prsOpts->ipv6PrsOptions.routingHdrEnable)
-+ tmpReg |= PRS_HDR_IPV6_ROUTE_HDR_EN;
-+ break;
-+ case (HEADER_TYPE_TCP):
-+ if (p_prsOpts->tcpPrsOptions.padIgnoreChecksum)
-+ tmpReg |= PRS_HDR_TCP_PAD_REMOVAL;
-+ else
-+ tmpReg &= ~PRS_HDR_TCP_PAD_REMOVAL;
-+ break;
-+ case (HEADER_TYPE_UDP):
-+ if (p_prsOpts->udpPrsOptions.padIgnoreChecksum)
-+ tmpReg |= PRS_HDR_UDP_PAD_REMOVAL;
-+ else
-+ tmpReg &= ~PRS_HDR_UDP_PAD_REMOVAL;
-+ break;
-+ default:
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("Invalid header"));
-+ }
-+ }
-+
-+ /* set software parsing (address is divided in 2 since parser uses 2 byte access. */
-+ if (p_HdrParams->swPrsEnable)
-+ {
-+ tmpPrsOffset = FmPcdGetSwPrsOffset(p_FmPort->h_FmPcd, p_HdrParams->hdr,
-+ p_HdrParams->indexPerHdr);
-+ if (tmpPrsOffset == ILLEGAL_BASE)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, NO_MSG);
-+ tmpReg |= (PRS_HDR_SW_PRS_EN | tmpPrsOffset);
-+ }
-+ *p_SoftSeqAttachReg = tmpReg;
-+
-+ return E_OK;
-+}
-+
-+static uint32_t GetPortSchemeBindParams(
-+ t_Handle h_FmPort, t_FmPcdKgInterModuleBindPortToSchemes *p_SchemeBind)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+ uint32_t walking1Mask = 0x80000000, tmp;
-+ uint8_t idx = 0;
-+
-+ p_SchemeBind->netEnvId = p_FmPort->netEnvId;
-+ p_SchemeBind->hardwarePortId = p_FmPort->hardwarePortId;
-+ p_SchemeBind->useClsPlan = p_FmPort->useClsPlan;
-+ p_SchemeBind->numOfSchemes = 0;
-+ tmp = p_FmPort->schemesPerPortVector;
-+ if (tmp)
-+ {
-+ while (tmp)
-+ {
-+ if (tmp & walking1Mask)
-+ {
-+ p_SchemeBind->schemesIds[p_SchemeBind->numOfSchemes] = idx;
-+ p_SchemeBind->numOfSchemes++;
-+ tmp &= ~walking1Mask;
-+ }
-+ walking1Mask >>= 1;
-+ idx++;
-+ }
-+ }
-+
-+ return tmp;
-+}
-+
-+static void FmPortCheckNApplyMacsec(t_Handle h_FmPort)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+ volatile uint32_t *p_BmiCfgReg = NULL;
-+ uint32_t macsecEn = BMI_PORT_CFG_EN_MACSEC;
-+ uint32_t lcv, walking1Mask = 0x80000000;
-+ uint8_t cnt = 0;
-+
-+ ASSERT_COND(p_FmPort);
-+ ASSERT_COND(p_FmPort->h_FmPcd);
-+ ASSERT_COND(!p_FmPort->p_FmPortDriverParam);
-+
-+ if ((p_FmPort->portType != e_FM_PORT_TYPE_RX_10G)
-+ && (p_FmPort->portType != e_FM_PORT_TYPE_RX))
-+ return;
-+
-+ p_BmiCfgReg = &p_FmPort->port.bmi_regs->rx.fmbm_rcfg;
-+ /* get LCV for MACSEC */
-+ if ((lcv = FmPcdGetMacsecLcv(p_FmPort->h_FmPcd, p_FmPort->netEnvId))
-+ != 0)
-+ {
-+ while (!(lcv & walking1Mask))
-+ {
-+ cnt++;
-+ walking1Mask >>= 1;
-+ }
-+
-+ macsecEn |= (uint32_t)cnt << BMI_PORT_CFG_MS_SEL_SHIFT;
-+ WRITE_UINT32(*p_BmiCfgReg, GET_UINT32(*p_BmiCfgReg) | macsecEn);
-+ }
-+}
-+
-+static t_Error SetPcd(t_FmPort *p_FmPort, t_FmPortPcdParams *p_PcdParams)
-+{
-+ t_Error err = E_OK;
-+ uint32_t tmpReg;
-+ volatile uint32_t *p_BmiNia = NULL;
-+ volatile uint32_t *p_BmiPrsNia = NULL;
-+ volatile uint32_t *p_BmiPrsStartOffset = NULL;
-+ volatile uint32_t *p_BmiInitPrsResult = NULL;
-+ volatile uint32_t *p_BmiCcBase = NULL;
-+ uint16_t hdrNum, L3HdrNum, greHdrNum;
-+ int i;
-+ bool isEmptyClsPlanGrp;
-+ uint32_t tmpHxs[FM_PCD_PRS_NUM_OF_HDRS];
-+ uint16_t absoluteProfileId;
-+ uint8_t physicalSchemeId;
-+ uint32_t ccTreePhysOffset;
-+ t_FmPcdKgInterModuleBindPortToSchemes schemeBind;
-+ uint32_t initialSwPrs = 0;
-+
-+ ASSERT_COND(p_FmPort);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmPort->p_FmPortDriverParam, E_INVALID_STATE);
-+
-+ if (p_FmPort->imEn)
-+ RETURN_ERROR(MAJOR, E_INVALID_OPERATION,
-+ ("available for non-independant mode ports only"));
-+
-+ if ((p_FmPort->portType != e_FM_PORT_TYPE_RX_10G)
-+ && (p_FmPort->portType != e_FM_PORT_TYPE_RX)
-+ && (p_FmPort->portType != e_FM_PORT_TYPE_OH_OFFLINE_PARSING))
-+ RETURN_ERROR(MAJOR, E_INVALID_OPERATION,
-+ ("available for Rx and offline parsing ports only"));
-+
-+ p_FmPort->netEnvId = FmPcdGetNetEnvId(p_PcdParams->h_NetEnv);
-+
-+ p_FmPort->pcdEngines = 0;
-+
-+ /* initialize p_FmPort->pcdEngines field in port's structure */
-+ switch (p_PcdParams->pcdSupport)
-+ {
-+ case (e_FM_PORT_PCD_SUPPORT_NONE):
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_STATE,
-+ ("No PCD configuration required if e_FM_PORT_PCD_SUPPORT_NONE selected"));
-+ case (e_FM_PORT_PCD_SUPPORT_PRS_ONLY):
-+ p_FmPort->pcdEngines |= FM_PCD_PRS;
-+ break;
-+ case (e_FM_PORT_PCD_SUPPORT_PLCR_ONLY):
-+ p_FmPort->pcdEngines |= FM_PCD_PLCR;
-+ break;
-+ case (e_FM_PORT_PCD_SUPPORT_PRS_AND_PLCR):
-+ p_FmPort->pcdEngines |= FM_PCD_PRS;
-+ p_FmPort->pcdEngines |= FM_PCD_PLCR;
-+ break;
-+ case (e_FM_PORT_PCD_SUPPORT_PRS_AND_KG):
-+ p_FmPort->pcdEngines |= FM_PCD_PRS;
-+ p_FmPort->pcdEngines |= FM_PCD_KG;
-+ break;
-+ case (e_FM_PORT_PCD_SUPPORT_PRS_AND_KG_AND_CC):
-+ p_FmPort->pcdEngines |= FM_PCD_PRS;
-+ p_FmPort->pcdEngines |= FM_PCD_CC;
-+ p_FmPort->pcdEngines |= FM_PCD_KG;
-+ break;
-+ case (e_FM_PORT_PCD_SUPPORT_PRS_AND_KG_AND_CC_AND_PLCR):
-+ p_FmPort->pcdEngines |= FM_PCD_PRS;
-+ p_FmPort->pcdEngines |= FM_PCD_KG;
-+ p_FmPort->pcdEngines |= FM_PCD_CC;
-+ p_FmPort->pcdEngines |= FM_PCD_PLCR;
-+ break;
-+ case (e_FM_PORT_PCD_SUPPORT_PRS_AND_CC):
-+ p_FmPort->pcdEngines |= FM_PCD_PRS;
-+ p_FmPort->pcdEngines |= FM_PCD_CC;
-+ break;
-+ case (e_FM_PORT_PCD_SUPPORT_PRS_AND_CC_AND_PLCR):
-+ p_FmPort->pcdEngines |= FM_PCD_PRS;
-+ p_FmPort->pcdEngines |= FM_PCD_CC;
-+ p_FmPort->pcdEngines |= FM_PCD_PLCR;
-+ break;
-+ case (e_FM_PORT_PCD_SUPPORT_PRS_AND_KG_AND_PLCR):
-+ p_FmPort->pcdEngines |= FM_PCD_PRS;
-+ p_FmPort->pcdEngines |= FM_PCD_KG;
-+ p_FmPort->pcdEngines |= FM_PCD_PLCR;
-+ break;
-+ case (e_FM_PORT_PCD_SUPPORT_CC_ONLY):
-+ p_FmPort->pcdEngines |= FM_PCD_CC;
-+ break;
-+#ifdef FM_CAPWAP_SUPPORT
-+ case (e_FM_PORT_PCD_SUPPORT_CC_AND_KG):
-+ p_FmPort->pcdEngines |= FM_PCD_CC;
-+ p_FmPort->pcdEngines |= FM_PCD_KG;
-+ break;
-+ case (e_FM_PORT_PCD_SUPPORT_CC_AND_KG_AND_PLCR):
-+ p_FmPort->pcdEngines |= FM_PCD_CC;
-+ p_FmPort->pcdEngines |= FM_PCD_KG;
-+ p_FmPort->pcdEngines |= FM_PCD_PLCR;
-+ break;
-+#endif /* FM_CAPWAP_SUPPORT */
-+
-+ default:
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("invalid pcdSupport"));
-+ }
-+
-+ if ((p_FmPort->pcdEngines & FM_PCD_PRS)
-+ && (p_PcdParams->p_PrsParams->numOfHdrsWithAdditionalParams
-+ > FM_PCD_PRS_NUM_OF_HDRS))
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("Port parser numOfHdrsWithAdditionalParams may not exceed %d", FM_PCD_PRS_NUM_OF_HDRS));
-+
-+ /* check that parameters exist for each and only each defined engine */
-+ if ((!!(p_FmPort->pcdEngines & FM_PCD_PRS) != !!p_PcdParams->p_PrsParams)
-+ || (!!(p_FmPort->pcdEngines & FM_PCD_KG)
-+ != !!p_PcdParams->p_KgParams)
-+ || (!!(p_FmPort->pcdEngines & FM_PCD_CC)
-+ != !!p_PcdParams->p_CcParams))
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_STATE,
-+ ("PCD initialization structure is not consistent with pcdSupport"));
-+
-+ /* get PCD registers pointers */
-+ switch (p_FmPort->portType)
-+ {
-+ case (e_FM_PORT_TYPE_RX_10G):
-+ case (e_FM_PORT_TYPE_RX):
-+ p_BmiNia = &p_FmPort->port.bmi_regs->rx.fmbm_rfne;
-+ p_BmiPrsNia = &p_FmPort->port.bmi_regs->rx.fmbm_rfpne;
-+ p_BmiPrsStartOffset = &p_FmPort->port.bmi_regs->rx.fmbm_rpso;
-+ p_BmiInitPrsResult = &p_FmPort->port.bmi_regs->rx.fmbm_rprai[0];
-+ p_BmiCcBase = &p_FmPort->port.bmi_regs->rx.fmbm_rccb;
-+ break;
-+ case (e_FM_PORT_TYPE_OH_OFFLINE_PARSING):
-+ p_BmiNia = &p_FmPort->port.bmi_regs->oh.fmbm_ofne;
-+ p_BmiPrsNia = &p_FmPort->port.bmi_regs->oh.fmbm_ofpne;
-+ p_BmiPrsStartOffset = &p_FmPort->port.bmi_regs->oh.fmbm_opso;
-+ p_BmiInitPrsResult = &p_FmPort->port.bmi_regs->oh.fmbm_oprai[0];
-+ p_BmiCcBase = &p_FmPort->port.bmi_regs->oh.fmbm_occb;
-+ break;
-+ default:
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("Invalid port type"));
-+ }
-+
-+ /* set PCD port parameter */
-+ if (p_FmPort->pcdEngines & FM_PCD_CC)
-+ {
-+ err = FmPcdCcBindTree(p_FmPort->h_FmPcd, p_PcdParams,
-+ p_PcdParams->p_CcParams->h_CcTree,
-+ &ccTreePhysOffset, p_FmPort);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+
-+ WRITE_UINT32(*p_BmiCcBase, ccTreePhysOffset);
-+ p_FmPort->ccTreeId = p_PcdParams->p_CcParams->h_CcTree;
-+ }
-+
-+ if (p_FmPort->pcdEngines & FM_PCD_KG)
-+ {
-+ if (p_PcdParams->p_KgParams->numOfSchemes == 0)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("For ports using Keygen, at least one scheme must be bound. "));
-+
-+ err = FmPcdKgSetOrBindToClsPlanGrp(p_FmPort->h_FmPcd,
-+ p_FmPort->hardwarePortId,
-+ p_FmPort->netEnvId,
-+ p_FmPort->optArray,
-+ &p_FmPort->clsPlanGrpId,
-+ &isEmptyClsPlanGrp);
-+ if (err)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE,
-+ ("FmPcdKgSetOrBindToClsPlanGrp failed. "));
-+
-+ p_FmPort->useClsPlan = !isEmptyClsPlanGrp;
-+
-+ schemeBind.netEnvId = p_FmPort->netEnvId;
-+ schemeBind.hardwarePortId = p_FmPort->hardwarePortId;
-+ schemeBind.numOfSchemes = p_PcdParams->p_KgParams->numOfSchemes;
-+ schemeBind.useClsPlan = p_FmPort->useClsPlan;
-+
-+ /* for each scheme */
-+ for (i = 0; i < p_PcdParams->p_KgParams->numOfSchemes; i++)
-+ {
-+ ASSERT_COND(p_PcdParams->p_KgParams->h_Schemes[i]);
-+ physicalSchemeId = FmPcdKgGetSchemeId(
-+ p_PcdParams->p_KgParams->h_Schemes[i]);
-+ schemeBind.schemesIds[i] = physicalSchemeId;
-+ /* build vector */
-+ p_FmPort->schemesPerPortVector |= 1
-+ << (31 - (uint32_t)physicalSchemeId);
-+#if (DPAA_VERSION >= 11)
-+ /*because of the state that VSPE is defined per port - all PCD path should be according to this requirement
-+ if !VSPE - in port, for relevant scheme VSPE can not be set*/
-+ if (!p_FmPort->vspe
-+ && FmPcdKgGetVspe((p_PcdParams->p_KgParams->h_Schemes[i])))
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE,
-+ ("VSPE is not at port level"));
-+#endif /* (DPAA_VERSION >= 11) */
-+ }
-+
-+ err = FmPcdKgBindPortToSchemes(p_FmPort->h_FmPcd, &schemeBind);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+
-+ /***************************/
-+ /* configure NIA after BMI */
-+ /***************************/
-+ /* rfne may contain FDCS bits, so first we read them. */
-+ p_FmPort->savedBmiNia = GET_UINT32(*p_BmiNia) & BMI_RFNE_FDCS_MASK;
-+
-+ /* If policer is used directly after BMI or PRS */
-+ if ((p_FmPort->pcdEngines & FM_PCD_PLCR)
-+ && ((p_PcdParams->pcdSupport == e_FM_PORT_PCD_SUPPORT_PLCR_ONLY)
-+ || (p_PcdParams->pcdSupport
-+ == e_FM_PORT_PCD_SUPPORT_PRS_AND_PLCR)))
-+ {
-+ if (!p_PcdParams->p_PlcrParams->h_Profile)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE,
-+ ("Profile should be initialized"));
-+
-+ absoluteProfileId = (uint16_t)FmPcdPlcrProfileGetAbsoluteId(
-+ p_PcdParams->p_PlcrParams->h_Profile);
-+
-+ if (!FmPcdPlcrIsProfileValid(p_FmPort->h_FmPcd, absoluteProfileId))
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE,
-+ ("Private port profile not valid."));
-+
-+ tmpReg = (uint32_t)(absoluteProfileId | NIA_PLCR_ABSOLUTE);
-+
-+ if (p_FmPort->pcdEngines & FM_PCD_PRS) /* e_FM_PCD_SUPPORT_PRS_AND_PLCR */
-+ /* update BMI HPNIA */
-+ WRITE_UINT32(*p_BmiPrsNia, (uint32_t)(NIA_ENG_PLCR | tmpReg));
-+ else
-+ /* e_FM_PCD_SUPPORT_PLCR_ONLY */
-+ /* update BMI NIA */
-+ p_FmPort->savedBmiNia |= (uint32_t)(NIA_ENG_PLCR);
-+ }
-+
-+ /* if CC is used directly after BMI */
-+ if ((p_PcdParams->pcdSupport == e_FM_PORT_PCD_SUPPORT_CC_ONLY)
-+#ifdef FM_CAPWAP_SUPPORT
-+ || (p_PcdParams->pcdSupport == e_FM_PORT_PCD_SUPPORT_CC_AND_KG)
-+ || (p_PcdParams->pcdSupport == e_FM_PORT_PCD_SUPPORT_CC_AND_KG_AND_PLCR)
-+#endif /* FM_CAPWAP_SUPPORT */
-+ )
-+ {
-+ if (p_FmPort->portType != e_FM_PORT_TYPE_OH_OFFLINE_PARSING)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_OPERATION,
-+ ("e_FM_PORT_PCD_SUPPORT_CC_xx available for offline parsing ports only"));
-+ p_FmPort->savedBmiNia |= (uint32_t)(NIA_ENG_FM_CTL | NIA_FM_CTL_AC_CC);
-+ /* check that prs start offset == RIM[FOF] */
-+ }
-+
-+ if (p_FmPort->pcdEngines & FM_PCD_PRS)
-+ {
-+ ASSERT_COND(p_PcdParams->p_PrsParams);
-+#if (DPAA_VERSION >= 11)
-+ if (p_PcdParams->p_PrsParams->firstPrsHdr == HEADER_TYPE_CAPWAP)
-+ hdrNum = OFFLOAD_SW_PATCH_CAPWAP_LABEL;
-+ else
-+ {
-+#endif /* (DPAA_VERSION >= 11) */
-+ /* if PRS is used it is always first */
-+ hdrNum = GetPrsHdrNum(p_PcdParams->p_PrsParams->firstPrsHdr);
-+ if (hdrNum == ILLEGAL_HDR_NUM)
-+ RETURN_ERROR(MAJOR, E_NOT_SUPPORTED, ("Unsupported header."));
-+#if (DPAA_VERSION >= 11)
-+ }
-+#endif /* (DPAA_VERSION >= 11) */
-+ p_FmPort->savedBmiNia |= (uint32_t)(NIA_ENG_PRS | (uint32_t)(hdrNum));
-+ /* set after parser NIA */
-+ tmpReg = 0;
-+ switch (p_PcdParams->pcdSupport)
-+ {
-+ case (e_FM_PORT_PCD_SUPPORT_PRS_ONLY):
-+ WRITE_UINT32(*p_BmiPrsNia,
-+ GET_NIA_BMI_AC_ENQ_FRAME(p_FmPort->h_FmPcd));
-+ break;
-+ case (e_FM_PORT_PCD_SUPPORT_PRS_AND_KG_AND_CC):
-+ case (e_FM_PORT_PCD_SUPPORT_PRS_AND_KG_AND_CC_AND_PLCR):
-+ tmpReg = NIA_KG_CC_EN;
-+ case (e_FM_PORT_PCD_SUPPORT_PRS_AND_KG):
-+ case (e_FM_PORT_PCD_SUPPORT_PRS_AND_KG_AND_PLCR):
-+ if (p_PcdParams->p_KgParams->directScheme)
-+ {
-+ physicalSchemeId = FmPcdKgGetSchemeId(
-+ p_PcdParams->p_KgParams->h_DirectScheme);
-+ /* check that this scheme was bound to this port */
-+ for (i = 0; i < p_PcdParams->p_KgParams->numOfSchemes; i++)
-+ if (p_PcdParams->p_KgParams->h_DirectScheme
-+ == p_PcdParams->p_KgParams->h_Schemes[i])
-+ break;
-+ if (i == p_PcdParams->p_KgParams->numOfSchemes)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("Direct scheme is not one of the port selected schemes."));
-+ tmpReg |= (uint32_t)(NIA_KG_DIRECT | physicalSchemeId);
-+ }
-+ WRITE_UINT32(*p_BmiPrsNia, NIA_ENG_KG | tmpReg);
-+ break;
-+ case (e_FM_PORT_PCD_SUPPORT_PRS_AND_CC):
-+ case (e_FM_PORT_PCD_SUPPORT_PRS_AND_CC_AND_PLCR):
-+ WRITE_UINT32(*p_BmiPrsNia,
-+ (uint32_t)(NIA_ENG_FM_CTL | NIA_FM_CTL_AC_CC));
-+ break;
-+ case (e_FM_PORT_PCD_SUPPORT_PRS_AND_PLCR):
-+ break;
-+ default:
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("Invalid PCD support"));
-+ }
-+
-+ /* set start parsing offset */
-+ WRITE_UINT32(*p_BmiPrsStartOffset,
-+ p_PcdParams->p_PrsParams->parsingOffset);
-+
-+ /************************************/
-+ /* Parser port parameters */
-+ /************************************/
-+ /* stop before configuring */
-+ WRITE_UINT32(p_FmPort->p_FmPortPrsRegs->pcac, PRS_CAC_STOP);
-+ /* wait for parser to be in idle state */
-+ while (GET_UINT32(p_FmPort->p_FmPortPrsRegs->pcac) & PRS_CAC_ACTIVE)
-+ ;
-+
-+ /* set soft seq attachment register */
-+ memset(tmpHxs, 0, FM_PCD_PRS_NUM_OF_HDRS * sizeof(uint32_t));
-+
-+ /* set protocol options */
-+ for (i = 0; p_FmPort->optArray[i]; i++)
-+ switch (p_FmPort->optArray[i])
-+ {
-+ case (ETH_BROADCAST):
-+ hdrNum = GetPrsHdrNum(HEADER_TYPE_ETH);
-+ tmpHxs[hdrNum] |= (i + 1) << PRS_HDR_ETH_BC_SHIFT;
-+ break;
-+ case (ETH_MULTICAST):
-+ hdrNum = GetPrsHdrNum(HEADER_TYPE_ETH);
-+ tmpHxs[hdrNum] |= (i + 1) << PRS_HDR_ETH_MC_SHIFT;
-+ break;
-+ case (VLAN_STACKED):
-+ hdrNum = GetPrsHdrNum(HEADER_TYPE_VLAN);
-+ tmpHxs[hdrNum] |= (i + 1) << PRS_HDR_VLAN_STACKED_SHIFT;
-+ break;
-+ case (MPLS_STACKED):
-+ hdrNum = GetPrsHdrNum(HEADER_TYPE_MPLS);
-+ tmpHxs[hdrNum] |= (i + 1) << PRS_HDR_MPLS_STACKED_SHIFT;
-+ break;
-+ case (IPV4_BROADCAST_1):
-+ hdrNum = GetPrsHdrNum(HEADER_TYPE_IPv4);
-+ tmpHxs[hdrNum] |= (i + 1) << PRS_HDR_IPV4_1_BC_SHIFT;
-+ break;
-+ case (IPV4_MULTICAST_1):
-+ hdrNum = GetPrsHdrNum(HEADER_TYPE_IPv4);
-+ tmpHxs[hdrNum] |= (i + 1) << PRS_HDR_IPV4_1_MC_SHIFT;
-+ break;
-+ case (IPV4_UNICAST_2):
-+ hdrNum = GetPrsHdrNum(HEADER_TYPE_IPv4);
-+ tmpHxs[hdrNum] |= (i + 1) << PRS_HDR_IPV4_2_UC_SHIFT;
-+ break;
-+ case (IPV4_MULTICAST_BROADCAST_2):
-+ hdrNum = GetPrsHdrNum(HEADER_TYPE_IPv4);
-+ tmpHxs[hdrNum] |= (i + 1) << PRS_HDR_IPV4_2_MC_BC_SHIFT;
-+ break;
-+ case (IPV6_MULTICAST_1):
-+ hdrNum = GetPrsHdrNum(HEADER_TYPE_IPv6);
-+ tmpHxs[hdrNum] |= (i + 1) << PRS_HDR_IPV6_1_MC_SHIFT;
-+ break;
-+ case (IPV6_UNICAST_2):
-+ hdrNum = GetPrsHdrNum(HEADER_TYPE_IPv6);
-+ tmpHxs[hdrNum] |= (i + 1) << PRS_HDR_IPV6_2_UC_SHIFT;
-+ break;
-+ case (IPV6_MULTICAST_2):
-+ hdrNum = GetPrsHdrNum(HEADER_TYPE_IPv6);
-+ tmpHxs[hdrNum] |= (i + 1) << PRS_HDR_IPV6_2_MC_SHIFT;
-+ break;
-+ }
-+
-+ if (FmPcdNetEnvIsHdrExist(p_FmPort->h_FmPcd, p_FmPort->netEnvId,
-+ HEADER_TYPE_UDP_ENCAP_ESP))
-+ {
-+ if (p_PcdParams->p_PrsParams->numOfHdrsWithAdditionalParams == FM_PCD_PRS_NUM_OF_HDRS)
-+ RETURN_ERROR(
-+ MINOR, E_INVALID_VALUE,
-+ ("If HEADER_TYPE_UDP_ENCAP_ESP is used, numOfHdrsWithAdditionalParams may be up to FM_PCD_PRS_NUM_OF_HDRS - 1"));
-+
-+ p_PcdParams->p_PrsParams->additionalParams[p_PcdParams->p_PrsParams->numOfHdrsWithAdditionalParams].hdr =
-+ HEADER_TYPE_UDP;
-+ p_PcdParams->p_PrsParams->additionalParams[p_PcdParams->p_PrsParams->numOfHdrsWithAdditionalParams].swPrsEnable =
-+ TRUE;
-+ p_PcdParams->p_PrsParams->numOfHdrsWithAdditionalParams++;
-+ }
-+
-+ /* set MPLS default next header - HW reset workaround */
-+ hdrNum = GetPrsHdrNum(HEADER_TYPE_MPLS);
-+ tmpHxs[hdrNum] |= PRS_HDR_MPLS_LBL_INTER_EN;
-+ L3HdrNum = GetPrsHdrNum(HEADER_TYPE_USER_DEFINED_L3);
-+ tmpHxs[hdrNum] |= (uint32_t)L3HdrNum << PRS_HDR_MPLS_NEXT_HDR_SHIFT;
-+
-+ /* for GRE, disable errors */
-+ greHdrNum = GetPrsHdrNum(HEADER_TYPE_GRE);
-+ tmpHxs[greHdrNum] |= PRS_HDR_ERROR_DIS;
-+
-+ /* For UDP remove PAD from L4 checksum calculation */
-+ hdrNum = GetPrsHdrNum(HEADER_TYPE_UDP);
-+ tmpHxs[hdrNum] |= PRS_HDR_UDP_PAD_REMOVAL;
-+ /* For TCP remove PAD from L4 checksum calculation */
-+ hdrNum = GetPrsHdrNum(HEADER_TYPE_TCP);
-+ tmpHxs[hdrNum] |= PRS_HDR_TCP_PAD_REMOVAL;
-+
-+ /* config additional params for specific headers */
-+ for (i = 0; i < p_PcdParams->p_PrsParams->numOfHdrsWithAdditionalParams;
-+ i++)
-+ {
-+ /* case for using sw parser as the initial NIA address, before
-+ * HW parsing
-+ */
-+ if ((p_PcdParams->p_PrsParams->additionalParams[i].hdr == HEADER_TYPE_NONE) &&
-+ p_PcdParams->p_PrsParams->additionalParams[i].swPrsEnable)
-+ {
-+ initialSwPrs = FmPcdGetSwPrsOffset(p_FmPort->h_FmPcd, HEADER_TYPE_NONE,
-+ p_PcdParams->p_PrsParams->additionalParams[i].indexPerHdr);
-+ if (initialSwPrs == ILLEGAL_BASE)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, NO_MSG);
-+
-+ /* clear parser first HXS */
-+ p_FmPort->savedBmiNia &= ~BMI_RFNE_HXS_MASK; /* 0x000000FF */
-+ /* rewrite with soft parser start */
-+ p_FmPort->savedBmiNia |= initialSwPrs;
-+ continue;
-+ }
-+
-+ hdrNum =
-+ GetPrsHdrNum(p_PcdParams->p_PrsParams->additionalParams[i].hdr);
-+ if (hdrNum == ILLEGAL_HDR_NUM)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, NO_MSG);
-+ if (hdrNum == NO_HDR_NUM)
-+ RETURN_ERROR(
-+ MAJOR, E_INVALID_VALUE,
-+ ("Private headers may not use additional parameters"));
-+
-+ err = AdditionalPrsParams(
-+ p_FmPort, &p_PcdParams->p_PrsParams->additionalParams[i],
-+ &tmpHxs[hdrNum]);
-+ if (err)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, NO_MSG);
-+ }
-+
-+ /* Check if ip-reassembly port - need to link sw-parser code */
-+ if (p_FmPort->h_IpReassemblyManip)
-+ {
-+ /* link to sw parser code for IP Frag - only if no other code is applied. */
-+ hdrNum = GetPrsHdrNum(HEADER_TYPE_IPv4);
-+ if (!(tmpHxs[hdrNum] & PRS_HDR_SW_PRS_EN))
-+ tmpHxs[hdrNum] |= (PRS_HDR_SW_PRS_EN | OFFLOAD_SW_PATCH_IPv4_IPR_LABEL);
-+ hdrNum = GetPrsHdrNum(HEADER_TYPE_IPv6);
-+ if (!(tmpHxs[hdrNum] & PRS_HDR_SW_PRS_EN))
-+ tmpHxs[hdrNum] |= (PRS_HDR_SW_PRS_EN | OFFLOAD_SW_PATCH_IPv6_IPR_LABEL);
-+ } else {
-+ if (FmPcdNetEnvIsHdrExist(p_FmPort->h_FmPcd, p_FmPort->netEnvId, HEADER_TYPE_UDP_LITE))
-+ {
-+ hdrNum = GetPrsHdrNum(HEADER_TYPE_IPv6);
-+ if (!(tmpHxs[hdrNum] & PRS_HDR_SW_PRS_EN))
-+ tmpHxs[hdrNum] |= (PRS_HDR_SW_PRS_EN | OFFLOAD_SW_PATCH_IPv6_IPF_LABEL);
-+ } else if ((FmPcdIsAdvancedOffloadSupported(p_FmPort->h_FmPcd)
-+ && (p_FmPort->portType == e_FM_PORT_TYPE_OH_OFFLINE_PARSING)))
-+ {
-+ hdrNum = GetPrsHdrNum(HEADER_TYPE_IPv6);
-+ if (!(tmpHxs[hdrNum] & PRS_HDR_SW_PRS_EN))
-+ tmpHxs[hdrNum] |= (PRS_HDR_SW_PRS_EN | OFFLOAD_SW_PATCH_IPv6_IPF_LABEL);
-+ }
-+ }
-+
-+#if ((DPAA_VERSION == 10) && defined(FM_CAPWAP_SUPPORT))
-+ if (FmPcdNetEnvIsHdrExist(p_FmPort->h_FmPcd, p_FmPort->netEnvId,
-+ HEADER_TYPE_UDP_LITE))
-+ {
-+ /* link to sw parser code for udp lite - only if no other code is applied. */
-+ hdrNum = GetPrsHdrNum(HEADER_TYPE_IPv6);
-+ if (!(tmpHxs[hdrNum] & PRS_HDR_SW_PRS_EN))
-+ tmpHxs[hdrNum] |= (PRS_HDR_SW_PRS_EN | UDP_LITE_SW_PATCH_LABEL);
-+ }
-+#endif /* ((DPAA_VERSION == 10) && defined(FM_CAPWAP_SUPPORT)) */
-+ for (i = 0; i < FM_PCD_PRS_NUM_OF_HDRS; i++)
-+ {
-+ /* For all header set LCV as taken from netEnv*/
-+ WRITE_UINT32(
-+ p_FmPort->p_FmPortPrsRegs->hdrs[i].lcv,
-+ FmPcdGetLcv(p_FmPort->h_FmPcd, p_FmPort->netEnvId, (uint8_t)i));
-+ /* set HXS register according to default+Additional params+protocol options */
-+ WRITE_UINT32(p_FmPort->p_FmPortPrsRegs->hdrs[i].softSeqAttach,
-+ tmpHxs[i]);
-+ }
-+
-+ /* set tpid. */
-+ tmpReg = PRS_TPID_DFLT;
-+ if (p_PcdParams->p_PrsParams->setVlanTpid1)
-+ {
-+ tmpReg &= PRS_TPID2_MASK;
-+ tmpReg |= (uint32_t)p_PcdParams->p_PrsParams->vlanTpid1
-+ << PRS_PCTPID_SHIFT;
-+ }
-+ if (p_PcdParams->p_PrsParams->setVlanTpid2)
-+ {
-+ tmpReg &= PRS_TPID1_MASK;
-+ tmpReg |= (uint32_t)p_PcdParams->p_PrsParams->vlanTpid2;
-+ }WRITE_UINT32(p_FmPort->p_FmPortPrsRegs->pctpid, tmpReg);
-+
-+ /* enable parser */
-+ WRITE_UINT32(p_FmPort->p_FmPortPrsRegs->pcac, 0);
-+
-+ if (p_PcdParams->p_PrsParams->prsResultPrivateInfo)
-+ p_FmPort->privateInfo =
-+ p_PcdParams->p_PrsParams->prsResultPrivateInfo;
-+
-+ } /* end parser */
-+ else {
-+ if (FmPcdIsAdvancedOffloadSupported(p_FmPort->h_FmPcd)
-+ && (p_FmPort->portType == e_FM_PORT_TYPE_OH_OFFLINE_PARSING))
-+ {
-+ hdrNum = GetPrsHdrNum(HEADER_TYPE_IPv6);
-+ WRITE_UINT32(p_FmPort->p_FmPortPrsRegs->hdrs[hdrNum].softSeqAttach,
-+ (PRS_HDR_SW_PRS_EN | OFFLOAD_SW_PATCH_IPv6_IPF_LABEL));
-+ }
-+
-+ WRITE_UINT32(*p_BmiPrsStartOffset, 0);
-+
-+ p_FmPort->privateInfo = 0;
-+ }
-+
-+ FmPortCheckNApplyMacsec(p_FmPort);
-+
-+ WRITE_UINT32(
-+ *p_BmiPrsStartOffset,
-+ GET_UINT32(*p_BmiPrsStartOffset) + p_FmPort->internalBufferOffset);
-+
-+ /* set initial parser result - used for all engines */
-+ for (i = 0; i < FM_PORT_PRS_RESULT_NUM_OF_WORDS; i++)
-+ {
-+ if (!i)
-+ WRITE_UINT32(
-+ *(p_BmiInitPrsResult),
-+ (uint32_t)(((uint32_t)p_FmPort->privateInfo << BMI_PR_PORTID_SHIFT) | BMI_PRS_RESULT_HIGH));
-+ else
-+ {
-+ if (i < FM_PORT_PRS_RESULT_NUM_OF_WORDS / 2)
-+ WRITE_UINT32(*(p_BmiInitPrsResult+i), BMI_PRS_RESULT_HIGH);
-+ else
-+ WRITE_UINT32(*(p_BmiInitPrsResult+i), BMI_PRS_RESULT_LOW);
-+ }
-+ }
-+
-+ return E_OK;
-+}
-+
-+static t_Error DeletePcd(t_FmPort *p_FmPort)
-+{
-+ t_Error err = E_OK;
-+ volatile uint32_t *p_BmiNia = NULL;
-+ volatile uint32_t *p_BmiPrsStartOffset = NULL;
-+
-+ ASSERT_COND(p_FmPort);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmPort->p_FmPortDriverParam, E_INVALID_STATE);
-+
-+ if (p_FmPort->imEn)
-+ RETURN_ERROR(MAJOR, E_INVALID_OPERATION,
-+ ("available for non-independant mode ports only"));
-+
-+ if ((p_FmPort->portType != e_FM_PORT_TYPE_RX_10G)
-+ && (p_FmPort->portType != e_FM_PORT_TYPE_RX)
-+ && (p_FmPort->portType != e_FM_PORT_TYPE_OH_OFFLINE_PARSING))
-+ RETURN_ERROR( MAJOR, E_INVALID_OPERATION,
-+ ("available for Rx and offline parsing ports only"));
-+
-+ if (!p_FmPort->pcdEngines)
-+ RETURN_ERROR(MAJOR, E_INVALID_OPERATION, ("called for non PCD port"));
-+
-+ /* get PCD registers pointers */
-+ switch (p_FmPort->portType)
-+ {
-+ case (e_FM_PORT_TYPE_RX_10G):
-+ case (e_FM_PORT_TYPE_RX):
-+ p_BmiNia = &p_FmPort->port.bmi_regs->rx.fmbm_rfne;
-+ p_BmiPrsStartOffset = &p_FmPort->port.bmi_regs->rx.fmbm_rpso;
-+ break;
-+ case (e_FM_PORT_TYPE_OH_OFFLINE_PARSING):
-+ p_BmiNia = &p_FmPort->port.bmi_regs->oh.fmbm_ofne;
-+ p_BmiPrsStartOffset = &p_FmPort->port.bmi_regs->oh.fmbm_opso;
-+ break;
-+ default:
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("Invalid port type"));
-+ }
-+
-+ if ((GET_UINT32(*p_BmiNia) & GET_NO_PCD_NIA_BMI_AC_ENQ_FRAME())
-+ != GET_NO_PCD_NIA_BMI_AC_ENQ_FRAME())
-+ RETURN_ERROR(MAJOR, E_INVALID_OPERATION,
-+ ("port has to be detached previousely"));
-+
-+ WRITE_UINT32(*p_BmiPrsStartOffset, 0);
-+
-+ /* "cut" PCD out of the port's flow - go to BMI */
-+ /* WRITE_UINT32(*p_BmiNia, (p_FmPort->savedBmiNia & BMI_RFNE_FDCS_MASK) | (NIA_ENG_BMI | NIA_BMI_AC_ENQ_FRAME)); */
-+
-+ if (p_FmPort->pcdEngines & FM_PCD_PRS)
-+ {
-+ /* stop parser */
-+ WRITE_UINT32(p_FmPort->p_FmPortPrsRegs->pcac, PRS_CAC_STOP);
-+ /* wait for parser to be in idle state */
-+ while (GET_UINT32(p_FmPort->p_FmPortPrsRegs->pcac) & PRS_CAC_ACTIVE)
-+ ;
-+ }
-+
-+ if (p_FmPort->pcdEngines & FM_PCD_KG)
-+ {
-+ t_FmPcdKgInterModuleBindPortToSchemes schemeBind;
-+
-+ /* unbind all schemes */
-+ p_FmPort->schemesPerPortVector = GetPortSchemeBindParams(p_FmPort,
-+ &schemeBind);
-+
-+ err = FmPcdKgUnbindPortToSchemes(p_FmPort->h_FmPcd, &schemeBind);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+
-+ err = FmPcdKgDeleteOrUnbindPortToClsPlanGrp(p_FmPort->h_FmPcd,
-+ p_FmPort->hardwarePortId,
-+ p_FmPort->clsPlanGrpId);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ p_FmPort->useClsPlan = FALSE;
-+ }
-+
-+ if (p_FmPort->pcdEngines & FM_PCD_CC)
-+ {
-+ /* unbind - we need to get the treeId too */
-+ err = FmPcdCcUnbindTree(p_FmPort->h_FmPcd, p_FmPort->ccTreeId);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+
-+ p_FmPort->pcdEngines = 0;
-+
-+ return E_OK;
-+}
-+
-+static t_Error AttachPCD(t_FmPort *p_FmPort)
-+{
-+ volatile uint32_t *p_BmiNia = NULL;
-+
-+ ASSERT_COND(p_FmPort);
-+
-+ /* get PCD registers pointers */
-+ if (p_FmPort->portType == e_FM_PORT_TYPE_OH_OFFLINE_PARSING)
-+ p_BmiNia = &p_FmPort->port.bmi_regs->oh.fmbm_ofne;
-+ else
-+ p_BmiNia = &p_FmPort->port.bmi_regs->rx.fmbm_rfne;
-+
-+ /* check that current NIA is BMI to BMI */
-+ if ((GET_UINT32(*p_BmiNia) & ~BMI_RFNE_FDCS_MASK)
-+ != GET_NO_PCD_NIA_BMI_AC_ENQ_FRAME())
-+ RETURN_ERROR( MAJOR, E_INVALID_OPERATION,
-+ ("may be called only for ports in BMI-to-BMI state."));
-+
-+ if (p_FmPort->requiredAction & UPDATE_FMFP_PRC_WITH_ONE_RISC_ONLY)
-+ if (FmSetNumOfRiscsPerPort(p_FmPort->h_Fm, p_FmPort->hardwarePortId, 1,
-+ p_FmPort->orFmanCtrl) != E_OK)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, NO_MSG);
-+
-+ if (p_FmPort->requiredAction & UPDATE_NIA_CMNE)
-+ {
-+ if (p_FmPort->portType == e_FM_PORT_TYPE_OH_OFFLINE_PARSING)
-+ WRITE_UINT32(p_FmPort->port.bmi_regs->oh.fmbm_ocmne,
-+ p_FmPort->savedBmiCmne);
-+ else
-+ WRITE_UINT32(p_FmPort->port.bmi_regs->rx.fmbm_rcmne,
-+ p_FmPort->savedBmiCmne);
-+ }
-+
-+ if (p_FmPort->requiredAction & UPDATE_NIA_PNEN)
-+ WRITE_UINT32(p_FmPort->p_FmPortQmiRegs->fmqm_pnen,
-+ p_FmPort->savedQmiPnen);
-+
-+ if (p_FmPort->requiredAction & UPDATE_NIA_FENE)
-+ {
-+ if (p_FmPort->portType == e_FM_PORT_TYPE_OH_OFFLINE_PARSING)
-+ WRITE_UINT32(p_FmPort->port.bmi_regs->oh.fmbm_ofene,
-+ p_FmPort->savedBmiFene);
-+ else
-+ WRITE_UINT32(p_FmPort->port.bmi_regs->rx.fmbm_rfene,
-+ p_FmPort->savedBmiFene);
-+ }
-+
-+ if (p_FmPort->requiredAction & UPDATE_NIA_FPNE)
-+ {
-+ if (p_FmPort->portType == e_FM_PORT_TYPE_OH_OFFLINE_PARSING)
-+ WRITE_UINT32(p_FmPort->port.bmi_regs->oh.fmbm_ofpne,
-+ p_FmPort->savedBmiFpne);
-+ else
-+ WRITE_UINT32(p_FmPort->port.bmi_regs->rx.fmbm_rfpne,
-+ p_FmPort->savedBmiFpne);
-+ }
-+
-+ if (p_FmPort->requiredAction & UPDATE_OFP_DPTE)
-+ {
-+ ASSERT_COND(p_FmPort->portType == e_FM_PORT_TYPE_OH_OFFLINE_PARSING);
-+
-+ WRITE_UINT32(p_FmPort->port.bmi_regs->oh.fmbm_ofp,
-+ p_FmPort->savedBmiOfp);
-+ }
-+
-+ WRITE_UINT32(*p_BmiNia, p_FmPort->savedBmiNia);
-+
-+ if (p_FmPort->requiredAction & UPDATE_NIA_PNDN)
-+ {
-+ p_FmPort->origNonRxQmiRegsPndn =
-+ GET_UINT32(p_FmPort->port.qmi_regs->fmqm_pndn);
-+ WRITE_UINT32(p_FmPort->port.qmi_regs->fmqm_pndn,
-+ p_FmPort->savedNonRxQmiRegsPndn);
-+ }
-+
-+ return E_OK;
-+}
-+
-+static t_Error DetachPCD(t_FmPort *p_FmPort)
-+{
-+ volatile uint32_t *p_BmiNia = NULL;
-+
-+ ASSERT_COND(p_FmPort);
-+
-+ /* get PCD registers pointers */
-+ if (p_FmPort->requiredAction & UPDATE_NIA_PNDN)
-+ WRITE_UINT32(p_FmPort->port.qmi_regs->fmqm_pndn,
-+ p_FmPort->origNonRxQmiRegsPndn);
-+
-+ if (p_FmPort->portType == e_FM_PORT_TYPE_OH_OFFLINE_PARSING)
-+ p_BmiNia = &p_FmPort->port.bmi_regs->oh.fmbm_ofne;
-+ else
-+ p_BmiNia = &p_FmPort->port.bmi_regs->rx.fmbm_rfne;
-+
-+ WRITE_UINT32(
-+ *p_BmiNia,
-+ (p_FmPort->savedBmiNia & BMI_RFNE_FDCS_MASK) | GET_NO_PCD_NIA_BMI_AC_ENQ_FRAME());
-+
-+ if (FmPcdGetHcHandle(p_FmPort->h_FmPcd))
-+ FmPcdHcSync(p_FmPort->h_FmPcd);
-+
-+ if (p_FmPort->requiredAction & UPDATE_NIA_FENE)
-+ {
-+ if (p_FmPort->portType == e_FM_PORT_TYPE_OH_OFFLINE_PARSING)
-+ WRITE_UINT32(p_FmPort->port.bmi_regs->oh.fmbm_ofene,
-+ NIA_ENG_QMI_ENQ | NIA_ORDER_RESTOR);
-+ else
-+ WRITE_UINT32(p_FmPort->port.bmi_regs->rx.fmbm_rfene,
-+ NIA_ENG_QMI_ENQ | NIA_ORDER_RESTOR);
-+ }
-+
-+ if (p_FmPort->requiredAction & UPDATE_NIA_PNEN)
-+ WRITE_UINT32(p_FmPort->port.qmi_regs->fmqm_pnen,
-+ NIA_ENG_BMI | NIA_BMI_AC_RELEASE);
-+
-+ if (p_FmPort->requiredAction & UPDATE_FMFP_PRC_WITH_ONE_RISC_ONLY)
-+ if (FmSetNumOfRiscsPerPort(p_FmPort->h_Fm, p_FmPort->hardwarePortId, 2,
-+ p_FmPort->orFmanCtrl) != E_OK)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, NO_MSG);
-+
-+ p_FmPort->requiredAction = 0;
-+
-+ return E_OK;
-+}
-+
-+/*****************************************************************************/
-+/* Inter-module API routines */
-+/*****************************************************************************/
-+void FmPortSetMacsecCmd(t_Handle h_FmPort, uint8_t dfltSci)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+ volatile uint32_t *p_BmiCfgReg = NULL;
-+ uint32_t tmpReg;
-+
-+ SANITY_CHECK_RETURN(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN(p_FmPort->p_FmPortDriverParam, E_INVALID_STATE);
-+
-+ if ((p_FmPort->portType != e_FM_PORT_TYPE_TX_10G)
-+ && (p_FmPort->portType != e_FM_PORT_TYPE_TX))
-+ {
-+ REPORT_ERROR(MAJOR, E_INVALID_OPERATION, ("The routine is relevant for Tx ports only"));
-+ return;
-+ }
-+
-+ p_BmiCfgReg = &p_FmPort->port.bmi_regs->tx.fmbm_tfca;
-+ tmpReg = GET_UINT32(*p_BmiCfgReg) & ~BMI_CMD_ATTR_MACCMD_MASK;
-+ tmpReg |= BMI_CMD_ATTR_MACCMD_SECURED;
-+ tmpReg |= (((uint32_t)dfltSci << BMI_CMD_ATTR_MACCMD_SC_SHIFT)
-+ & BMI_CMD_ATTR_MACCMD_SC_MASK);
-+
-+ WRITE_UINT32(*p_BmiCfgReg, tmpReg);
-+}
-+
-+uint8_t FmPortGetNetEnvId(t_Handle h_FmPort)
-+{
-+ return ((t_FmPort*)h_FmPort)->netEnvId;
-+}
-+
-+uint8_t FmPortGetHardwarePortId(t_Handle h_FmPort)
-+{
-+ return ((t_FmPort*)h_FmPort)->hardwarePortId;
-+}
-+
-+uint32_t FmPortGetPcdEngines(t_Handle h_FmPort)
-+{
-+ return ((t_FmPort*)h_FmPort)->pcdEngines;
-+}
-+
-+#if (DPAA_VERSION >= 11)
-+t_Error FmPortSetGprFunc(t_Handle h_FmPort, e_FmPortGprFuncType gprFunc,
-+ void **p_Value)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+ uint32_t muramPageOffset;
-+
-+ ASSERT_COND(p_FmPort);
-+ ASSERT_COND(p_Value);
-+
-+ if (p_FmPort->gprFunc != e_FM_PORT_GPR_EMPTY)
-+ {
-+ if (p_FmPort->gprFunc != gprFunc)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE,
-+ ("gpr was assigned with different func"));
-+ }
-+ else
-+ {
-+ switch (gprFunc)
-+ {
-+ case (e_FM_PORT_GPR_MURAM_PAGE):
-+ p_FmPort->p_ParamsPage = FM_MURAM_AllocMem(p_FmPort->h_FmMuram,
-+ 256, 8);
-+ if (!p_FmPort->p_ParamsPage)
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, ("MURAM alloc for page"));
-+
-+ IOMemSet32(p_FmPort->p_ParamsPage, 0, 256);
-+ muramPageOffset =
-+ (uint32_t)(XX_VirtToPhys(p_FmPort->p_ParamsPage)
-+ - p_FmPort->fmMuramPhysBaseAddr);
-+ switch (p_FmPort->portType)
-+ {
-+ case (e_FM_PORT_TYPE_RX_10G):
-+ case (e_FM_PORT_TYPE_RX):
-+ WRITE_UINT32(
-+ p_FmPort->p_FmPortBmiRegs->rxPortBmiRegs.fmbm_rgpr,
-+ muramPageOffset);
-+ break;
-+ case (e_FM_PORT_TYPE_OH_OFFLINE_PARSING):
-+ WRITE_UINT32(
-+ p_FmPort->p_FmPortBmiRegs->ohPortBmiRegs.fmbm_ogpr,
-+ muramPageOffset);
-+ break;
-+ default:
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE,
-+ ("Invalid port type"));
-+ }
-+ break;
-+ default:
-+ RETURN_ERROR(MAJOR, E_INVALID_SELECTION, NO_MSG);
-+ }
-+ p_FmPort->gprFunc = gprFunc;
-+ }
-+
-+ switch (p_FmPort->gprFunc)
-+ {
-+ case (e_FM_PORT_GPR_MURAM_PAGE):
-+ *p_Value = p_FmPort->p_ParamsPage;
-+ break;
-+ default:
-+ RETURN_ERROR(MAJOR, E_INVALID_SELECTION, NO_MSG);
-+ }
-+
-+ return E_OK;
-+}
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+t_Error FmPortGetSetCcParams(t_Handle h_FmPort,
-+ t_FmPortGetSetCcParams *p_CcParams)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+ int tmpInt;
-+ volatile uint32_t *p_BmiPrsStartOffset = NULL;
-+
-+ /* this function called from Cc for pass and receive parameters port params between CC and PORT*/
-+
-+ if ((p_CcParams->getCcParams.type & OFFSET_OF_PR)
-+ && (p_FmPort->bufferOffsets.prsResultOffset != ILLEGAL_BASE))
-+ {
-+ p_CcParams->getCcParams.prOffset =
-+ (uint8_t)p_FmPort->bufferOffsets.prsResultOffset;
-+ p_CcParams->getCcParams.type &= ~OFFSET_OF_PR;
-+ }
-+ if (p_CcParams->getCcParams.type & HW_PORT_ID)
-+ {
-+ p_CcParams->getCcParams.hardwarePortId =
-+ (uint8_t)p_FmPort->hardwarePortId;
-+ p_CcParams->getCcParams.type &= ~HW_PORT_ID;
-+ }
-+ if ((p_CcParams->getCcParams.type & OFFSET_OF_DATA)
-+ && (p_FmPort->bufferOffsets.dataOffset != ILLEGAL_BASE))
-+ {
-+ p_CcParams->getCcParams.dataOffset =
-+ (uint16_t)p_FmPort->bufferOffsets.dataOffset;
-+ p_CcParams->getCcParams.type &= ~OFFSET_OF_DATA;
-+ }
-+ if (p_CcParams->getCcParams.type & NUM_OF_TASKS)
-+ {
-+ p_CcParams->getCcParams.numOfTasks = (uint8_t)p_FmPort->tasks.num;
-+ p_CcParams->getCcParams.type &= ~NUM_OF_TASKS;
-+ }
-+ if (p_CcParams->getCcParams.type & NUM_OF_EXTRA_TASKS)
-+ {
-+ p_CcParams->getCcParams.numOfExtraTasks =
-+ (uint8_t)p_FmPort->tasks.extra;
-+ p_CcParams->getCcParams.type &= ~NUM_OF_EXTRA_TASKS;
-+ }
-+ if (p_CcParams->getCcParams.type & FM_REV)
-+ {
-+ p_CcParams->getCcParams.revInfo.majorRev = p_FmPort->fmRevInfo.majorRev;
-+ p_CcParams->getCcParams.revInfo.minorRev = p_FmPort->fmRevInfo.minorRev;
-+ p_CcParams->getCcParams.type &= ~FM_REV;
-+ }
-+ if (p_CcParams->getCcParams.type & DISCARD_MASK)
-+ {
-+ if (p_FmPort->portType == e_FM_PORT_TYPE_OH_OFFLINE_PARSING)
-+ p_CcParams->getCcParams.discardMask =
-+ GET_UINT32(p_FmPort->p_FmPortBmiRegs->ohPortBmiRegs.fmbm_ofsdm);
-+ else
-+ p_CcParams->getCcParams.discardMask =
-+ GET_UINT32(p_FmPort->p_FmPortBmiRegs->rxPortBmiRegs.fmbm_rfsdm);
-+ p_CcParams->getCcParams.type &= ~DISCARD_MASK;
-+ }
-+ if (p_CcParams->getCcParams.type & MANIP_EXTRA_SPACE)
-+ {
-+ p_CcParams->getCcParams.internalBufferOffset =
-+ p_FmPort->internalBufferOffset;
-+ p_CcParams->getCcParams.type &= ~MANIP_EXTRA_SPACE;
-+ }
-+ if (p_CcParams->getCcParams.type & GET_NIA_FPNE)
-+ {
-+ if (p_FmPort->portType == e_FM_PORT_TYPE_OH_OFFLINE_PARSING)
-+ p_CcParams->getCcParams.nia =
-+ GET_UINT32(p_FmPort->port.bmi_regs->oh.fmbm_ofpne);
-+ else
-+ p_CcParams->getCcParams.nia =
-+ GET_UINT32(p_FmPort->port.bmi_regs->rx.fmbm_rfpne);
-+ p_CcParams->getCcParams.type &= ~GET_NIA_FPNE;
-+ }
-+ if (p_CcParams->getCcParams.type & GET_NIA_PNDN)
-+ {
-+ if (p_FmPort->portType != e_FM_PORT_TYPE_OH_OFFLINE_PARSING)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("Invalid port type"));
-+ p_CcParams->getCcParams.nia =
-+ GET_UINT32(p_FmPort->p_FmPortQmiRegs->nonRxQmiRegs.fmqm_pndn);
-+ p_CcParams->getCcParams.type &= ~GET_NIA_PNDN;
-+ }
-+
-+ if ((p_CcParams->setCcParams.type & UPDATE_FMFP_PRC_WITH_ONE_RISC_ONLY)
-+ && !(p_FmPort->requiredAction & UPDATE_FMFP_PRC_WITH_ONE_RISC_ONLY))
-+ {
-+ p_FmPort->requiredAction |= UPDATE_FMFP_PRC_WITH_ONE_RISC_ONLY;
-+ p_FmPort->orFmanCtrl = p_CcParams->setCcParams.orFmanCtrl;
-+ }
-+
-+ if ((p_CcParams->setCcParams.type & UPDATE_NIA_PNEN)
-+ && !(p_FmPort->requiredAction & UPDATE_NIA_PNEN))
-+ {
-+ p_FmPort->savedQmiPnen = p_CcParams->setCcParams.nia;
-+ p_FmPort->requiredAction |= UPDATE_NIA_PNEN;
-+ }
-+ else
-+ if (p_CcParams->setCcParams.type & UPDATE_NIA_PNEN)
-+ {
-+ if (p_FmPort->savedQmiPnen != p_CcParams->setCcParams.nia)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE,
-+ ("PNEN was defined previously different"));
-+ }
-+
-+ if ((p_CcParams->setCcParams.type & UPDATE_NIA_PNDN)
-+ && !(p_FmPort->requiredAction & UPDATE_NIA_PNDN))
-+ {
-+ p_FmPort->savedNonRxQmiRegsPndn = p_CcParams->setCcParams.nia;
-+ p_FmPort->requiredAction |= UPDATE_NIA_PNDN;
-+ }
-+ else
-+ if (p_CcParams->setCcParams.type & UPDATE_NIA_PNDN)
-+ {
-+ if (p_FmPort->savedNonRxQmiRegsPndn != p_CcParams->setCcParams.nia)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE,
-+ ("PNDN was defined previously different"));
-+ }
-+
-+ if ((p_CcParams->setCcParams.type & UPDATE_NIA_FENE)
-+ && (p_CcParams->setCcParams.overwrite
-+ || !(p_FmPort->requiredAction & UPDATE_NIA_FENE)))
-+ {
-+ p_FmPort->savedBmiFene = p_CcParams->setCcParams.nia;
-+ p_FmPort->requiredAction |= UPDATE_NIA_FENE;
-+ }
-+ else
-+ if (p_CcParams->setCcParams.type & UPDATE_NIA_FENE)
-+ {
-+ if (p_FmPort->savedBmiFene != p_CcParams->setCcParams.nia)
-+ RETURN_ERROR( MAJOR, E_INVALID_STATE,
-+ ("xFENE was defined previously different"));
-+ }
-+
-+ if ((p_CcParams->setCcParams.type & UPDATE_NIA_FPNE)
-+ && !(p_FmPort->requiredAction & UPDATE_NIA_FPNE))
-+ {
-+ p_FmPort->savedBmiFpne = p_CcParams->setCcParams.nia;
-+ p_FmPort->requiredAction |= UPDATE_NIA_FPNE;
-+ }
-+ else
-+ if (p_CcParams->setCcParams.type & UPDATE_NIA_FPNE)
-+ {
-+ if (p_FmPort->savedBmiFpne != p_CcParams->setCcParams.nia)
-+ RETURN_ERROR( MAJOR, E_INVALID_STATE,
-+ ("xFPNE was defined previously different"));
-+ }
-+
-+ if ((p_CcParams->setCcParams.type & UPDATE_NIA_CMNE)
-+ && !(p_FmPort->requiredAction & UPDATE_NIA_CMNE))
-+ {
-+ p_FmPort->savedBmiCmne = p_CcParams->setCcParams.nia;
-+ p_FmPort->requiredAction |= UPDATE_NIA_CMNE;
-+ }
-+ else
-+ if (p_CcParams->setCcParams.type & UPDATE_NIA_CMNE)
-+ {
-+ if (p_FmPort->savedBmiCmne != p_CcParams->setCcParams.nia)
-+ RETURN_ERROR( MAJOR, E_INVALID_STATE,
-+ ("xCMNE was defined previously different"));
-+ }
-+
-+ if ((p_CcParams->setCcParams.type & UPDATE_PSO)
-+ && !(p_FmPort->requiredAction & UPDATE_PSO))
-+ {
-+ /* get PCD registers pointers */
-+ switch (p_FmPort->portType)
-+ {
-+ case (e_FM_PORT_TYPE_RX_10G):
-+ case (e_FM_PORT_TYPE_RX):
-+ p_BmiPrsStartOffset = &p_FmPort->port.bmi_regs->rx.fmbm_rpso;
-+ break;
-+ case (e_FM_PORT_TYPE_OH_OFFLINE_PARSING):
-+ p_BmiPrsStartOffset = &p_FmPort->port.bmi_regs->oh.fmbm_opso;
-+ break;
-+ default:
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("Invalid port type"));
-+ }
-+
-+ /* set start parsing offset */
-+ tmpInt = (int)GET_UINT32(*p_BmiPrsStartOffset)
-+ + p_CcParams->setCcParams.psoSize;
-+ if (tmpInt > 0)
-+ WRITE_UINT32(*p_BmiPrsStartOffset, (uint32_t)tmpInt);
-+
-+ p_FmPort->requiredAction |= UPDATE_PSO;
-+ p_FmPort->savedPrsStartOffset = p_CcParams->setCcParams.psoSize;
-+ }
-+ else
-+ if (p_CcParams->setCcParams.type & UPDATE_PSO)
-+ {
-+ if (p_FmPort->savedPrsStartOffset
-+ != p_CcParams->setCcParams.psoSize)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_STATE,
-+ ("parser start offset was defoned previousley different"));
-+ }
-+
-+ if ((p_CcParams->setCcParams.type & UPDATE_OFP_DPTE)
-+ && !(p_FmPort->requiredAction & UPDATE_OFP_DPTE))
-+ {
-+ if (p_FmPort->portType != e_FM_PORT_TYPE_OH_OFFLINE_PARSING)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("Invalid port type"));
-+ p_FmPort->savedBmiOfp = GET_UINT32(p_FmPort->port.bmi_regs->oh.fmbm_ofp);
-+ p_FmPort->savedBmiOfp &= ~BMI_FIFO_PIPELINE_DEPTH_MASK;
-+ p_FmPort->savedBmiOfp |= p_CcParams->setCcParams.ofpDpde
-+ << BMI_FIFO_PIPELINE_DEPTH_SHIFT;
-+ p_FmPort->requiredAction |= UPDATE_OFP_DPTE;
-+ }
-+
-+ return E_OK;
-+}
-+/*********************** End of inter-module routines ************************/
-+
-+/****************************************/
-+/* API Init unit functions */
-+/****************************************/
-+
-+t_Handle FM_PORT_Config(t_FmPortParams *p_FmPortParams)
-+{
-+ t_FmPort *p_FmPort;
-+ uintptr_t baseAddr = p_FmPortParams->baseAddr;
-+ uint32_t tmpReg;
-+
-+ /* Allocate FM structure */
-+ p_FmPort = (t_FmPort *)XX_Malloc(sizeof(t_FmPort));
-+ if (!p_FmPort)
-+ {
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("FM Port driver structure"));
-+ return NULL;
-+ }
-+ memset(p_FmPort, 0, sizeof(t_FmPort));
-+
-+ /* Allocate the FM driver's parameters structure */
-+ p_FmPort->p_FmPortDriverParam = (t_FmPortDriverParam *)XX_Malloc(
-+ sizeof(t_FmPortDriverParam));
-+ if (!p_FmPort->p_FmPortDriverParam)
-+ {
-+ XX_Free(p_FmPort);
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("FM Port driver parameters"));
-+ return NULL;
-+ }
-+ memset(p_FmPort->p_FmPortDriverParam, 0, sizeof(t_FmPortDriverParam));
-+
-+ /* Initialize FM port parameters which will be kept by the driver */
-+ p_FmPort->portType = p_FmPortParams->portType;
-+ p_FmPort->portId = p_FmPortParams->portId;
-+ p_FmPort->pcdEngines = FM_PCD_NONE;
-+ p_FmPort->f_Exception = p_FmPortParams->f_Exception;
-+ p_FmPort->h_App = p_FmPortParams->h_App;
-+ p_FmPort->h_Fm = p_FmPortParams->h_Fm;
-+
-+ /* get FM revision */
-+ FM_GetRevision(p_FmPort->h_Fm, &p_FmPort->fmRevInfo);
-+
-+ /* calculate global portId number */
-+ p_FmPort->hardwarePortId = SwPortIdToHwPortId(p_FmPort->portType,
-+ p_FmPortParams->portId,
-+ p_FmPort->fmRevInfo.majorRev,
-+ p_FmPort->fmRevInfo.minorRev);
-+
-+ if (p_FmPort->fmRevInfo.majorRev >= 6)
-+ {
-+ if ((p_FmPort->portType == e_FM_PORT_TYPE_OH_HOST_COMMAND)
-+ && (p_FmPortParams->portId != FM_OH_PORT_ID))
-+ DBG(WARNING,
-+ ("Port ID %d is recommended for HC port. Overwriting HW defaults to be suitable for HC.",
-+ FM_OH_PORT_ID));
-+
-+ if ((p_FmPort->portType == e_FM_PORT_TYPE_OH_OFFLINE_PARSING)
-+ && (p_FmPortParams->portId == FM_OH_PORT_ID))
-+ DBG(WARNING, ("Use non-zero portId for OP port due to insufficient resources on portId 0."));
-+ }
-+
-+ /* Set up FM port parameters for initialization phase only */
-+
-+ /* First, fill in flibs struct */
-+ fman_port_defconfig(&p_FmPort->p_FmPortDriverParam->dfltCfg,
-+ (enum fman_port_type)p_FmPort->portType);
-+ /* Overwrite some integration specific parameters */
-+ p_FmPort->p_FmPortDriverParam->dfltCfg.rx_pri_elevation =
-+ DEFAULT_PORT_rxFifoPriElevationLevel;
-+ p_FmPort->p_FmPortDriverParam->dfltCfg.rx_fifo_thr =
-+ DEFAULT_PORT_rxFifoThreshold;
-+
-+#if defined(FM_OP_NO_VSP_NO_RELEASE_ERRATA_FMAN_A006675) || defined(FM_ERROR_VSP_NO_MATCH_SW006)
-+ p_FmPort->p_FmPortDriverParam->dfltCfg.errata_A006675 = TRUE;
-+#else
-+ p_FmPort->p_FmPortDriverParam->dfltCfg.errata_A006675 = FALSE;
-+#endif
-+ if ((p_FmPort->fmRevInfo.majorRev == 6)
-+ && (p_FmPort->fmRevInfo.minorRev == 0))
-+ p_FmPort->p_FmPortDriverParam->dfltCfg.errata_A006320 = TRUE;
-+ else
-+ p_FmPort->p_FmPortDriverParam->dfltCfg.errata_A006320 = FALSE;
-+
-+ /* Excessive Threshold register - exists for pre-FMv3 chips only */
-+ if (p_FmPort->fmRevInfo.majorRev < 6)
-+ {
-+#ifdef FM_NO_RESTRICT_ON_ACCESS_RSRC
-+ p_FmPort->p_FmPortDriverParam->dfltCfg.excessive_threshold_register =
-+ TRUE;
-+#endif
-+ p_FmPort->p_FmPortDriverParam->dfltCfg.fmbm_rebm_has_sgd = FALSE;
-+ p_FmPort->p_FmPortDriverParam->dfltCfg.fmbm_tfne_has_features = FALSE;
-+ }
-+ else
-+ {
-+ p_FmPort->p_FmPortDriverParam->dfltCfg.excessive_threshold_register =
-+ FALSE;
-+ p_FmPort->p_FmPortDriverParam->dfltCfg.fmbm_rebm_has_sgd = TRUE;
-+ p_FmPort->p_FmPortDriverParam->dfltCfg.fmbm_tfne_has_features = TRUE;
-+ }
-+ if (p_FmPort->fmRevInfo.majorRev == 4)
-+ p_FmPort->p_FmPortDriverParam->dfltCfg.qmi_deq_options_support = FALSE;
-+ else
-+ p_FmPort->p_FmPortDriverParam->dfltCfg.qmi_deq_options_support = TRUE;
-+
-+ /* Continue with other parameters */
-+ p_FmPort->p_FmPortDriverParam->baseAddr = baseAddr;
-+ /* set memory map pointers */
-+ p_FmPort->p_FmPortQmiRegs =
-+ (t_FmPortQmiRegs *)UINT_TO_PTR(baseAddr + QMI_PORT_REGS_OFFSET);
-+ p_FmPort->p_FmPortBmiRegs =
-+ (u_FmPortBmiRegs *)UINT_TO_PTR(baseAddr + BMI_PORT_REGS_OFFSET);
-+ p_FmPort->p_FmPortPrsRegs =
-+ (t_FmPortPrsRegs *)UINT_TO_PTR(baseAddr + PRS_PORT_REGS_OFFSET);
-+
-+ p_FmPort->p_FmPortDriverParam->bufferPrefixContent.privDataSize =
-+ DEFAULT_PORT_bufferPrefixContent_privDataSize;
-+ p_FmPort->p_FmPortDriverParam->bufferPrefixContent.passPrsResult =
-+ DEFAULT_PORT_bufferPrefixContent_passPrsResult;
-+ p_FmPort->p_FmPortDriverParam->bufferPrefixContent.passTimeStamp =
-+ DEFAULT_PORT_bufferPrefixContent_passTimeStamp;
-+ p_FmPort->p_FmPortDriverParam->bufferPrefixContent.passAllOtherPCDInfo =
-+ DEFAULT_PORT_bufferPrefixContent_passTimeStamp;
-+ p_FmPort->p_FmPortDriverParam->bufferPrefixContent.dataAlign =
-+ DEFAULT_PORT_bufferPrefixContent_dataAlign;
-+ /* p_FmPort->p_FmPortDriverParam->dmaSwapData = (e_FmDmaSwapOption)DEFAULT_PORT_dmaSwapData;
-+ p_FmPort->p_FmPortDriverParam->dmaIntContextCacheAttr = (e_FmDmaCacheOption)DEFAULT_PORT_dmaIntContextCacheAttr;
-+ p_FmPort->p_FmPortDriverParam->dmaHeaderCacheAttr = (e_FmDmaCacheOption)DEFAULT_PORT_dmaHeaderCacheAttr;
-+ p_FmPort->p_FmPortDriverParam->dmaScatterGatherCacheAttr = (e_FmDmaCacheOption)DEFAULT_PORT_dmaScatterGatherCacheAttr;
-+ p_FmPort->p_FmPortDriverParam->dmaWriteOptimize = DEFAULT_PORT_dmaWriteOptimize;
-+ */
-+ p_FmPort->p_FmPortDriverParam->liodnBase = p_FmPortParams->liodnBase;
-+ p_FmPort->p_FmPortDriverParam->cheksumLastBytesIgnore =
-+ DEFAULT_PORT_cheksumLastBytesIgnore;
-+
-+ p_FmPort->maxFrameLength = DEFAULT_PORT_maxFrameLength;
-+ /* resource distribution. */
-+ p_FmPort->fifoBufs.num = DEFAULT_PORT_numOfFifoBufs(p_FmPort->portType)
-+ * BMI_FIFO_UNITS;
-+ p_FmPort->fifoBufs.extra = DEFAULT_PORT_extraNumOfFifoBufs
-+ * BMI_FIFO_UNITS;
-+ p_FmPort->openDmas.num = DEFAULT_PORT_numOfOpenDmas(p_FmPort->portType);
-+ p_FmPort->openDmas.extra =
-+ DEFAULT_PORT_extraNumOfOpenDmas(p_FmPort->portType);
-+ p_FmPort->tasks.num = DEFAULT_PORT_numOfTasks(p_FmPort->portType);
-+ p_FmPort->tasks.extra = DEFAULT_PORT_extraNumOfTasks(p_FmPort->portType);
-+
-+
-+#ifdef FM_HEAVY_TRAFFIC_SEQUENCER_HANG_ERRATA_FMAN_A006981
-+ if ((p_FmPort->fmRevInfo.majorRev == 6)
-+ && (p_FmPort->fmRevInfo.minorRev == 0)
-+ && ((p_FmPort->portType == e_FM_PORT_TYPE_OH_OFFLINE_PARSING)
-+ || (p_FmPort->portType == e_FM_PORT_TYPE_TX)))
-+ {
-+ p_FmPort->openDmas.num = 16;
-+ p_FmPort->openDmas.extra = 0;
-+ }
-+#endif /* FM_HEAVY_TRAFFIC_SEQUENCER_HANG_ERRATA_FMAN_A006981 */
-+
-+ /* Port type specific initialization: */
-+ switch (p_FmPort->portType)
-+ {
-+ case (e_FM_PORT_TYPE_RX):
-+ case (e_FM_PORT_TYPE_RX_10G):
-+ /* Initialize FM port parameters for initialization phase only */
-+ p_FmPort->p_FmPortDriverParam->cutBytesFromEnd =
-+ DEFAULT_PORT_cutBytesFromEnd;
-+ p_FmPort->p_FmPortDriverParam->enBufPoolDepletion = FALSE;
-+ p_FmPort->p_FmPortDriverParam->frmDiscardOverride =
-+ DEFAULT_PORT_frmDiscardOverride;
-+
-+ tmpReg =
-+ GET_UINT32(p_FmPort->p_FmPortBmiRegs->rxPortBmiRegs.fmbm_rfp);
-+ p_FmPort->p_FmPortDriverParam->rxFifoPriElevationLevel =
-+ (((tmpReg & BMI_RX_FIFO_PRI_ELEVATION_MASK)
-+ >> BMI_RX_FIFO_PRI_ELEVATION_SHIFT) + 1)
-+ * BMI_FIFO_UNITS;
-+ p_FmPort->p_FmPortDriverParam->rxFifoThreshold = (((tmpReg
-+ & BMI_RX_FIFO_THRESHOLD_MASK)
-+ >> BMI_RX_FIFO_THRESHOLD_SHIFT) + 1) * BMI_FIFO_UNITS;
-+
-+ p_FmPort->p_FmPortDriverParam->bufMargins.endMargins =
-+ DEFAULT_PORT_BufMargins_endMargins;
-+ p_FmPort->p_FmPortDriverParam->errorsToDiscard =
-+ DEFAULT_PORT_errorsToDiscard;
-+ p_FmPort->p_FmPortDriverParam->forwardReuseIntContext =
-+ DEFAULT_PORT_forwardIntContextReuse;
-+#if (DPAA_VERSION >= 11)
-+ p_FmPort->p_FmPortDriverParam->noScatherGather =
-+ DEFAULT_PORT_noScatherGather;
-+#endif /* (DPAA_VERSION >= 11) */
-+ break;
-+
-+ case (e_FM_PORT_TYPE_TX):
-+ p_FmPort->p_FmPortDriverParam->dontReleaseBuf = FALSE;
-+#ifdef FM_WRONG_RESET_VALUES_ERRATA_FMAN_A005127
-+ tmpReg = 0x00001013;
-+ WRITE_UINT32( p_FmPort->p_FmPortBmiRegs->txPortBmiRegs.fmbm_tfp,
-+ tmpReg);
-+#endif /* FM_WRONG_RESET_VALUES_ERRATA_FMAN_A005127 */
-+ case (e_FM_PORT_TYPE_TX_10G):
-+ tmpReg =
-+ GET_UINT32(p_FmPort->p_FmPortBmiRegs->txPortBmiRegs.fmbm_tfp);
-+ p_FmPort->p_FmPortDriverParam->txFifoMinFillLevel = ((tmpReg
-+ & BMI_TX_FIFO_MIN_FILL_MASK)
-+ >> BMI_TX_FIFO_MIN_FILL_SHIFT) * BMI_FIFO_UNITS;
-+ p_FmPort->p_FmPortDriverParam->dfltCfg.tx_fifo_deq_pipeline_depth =
-+ (uint8_t)(((tmpReg & BMI_FIFO_PIPELINE_DEPTH_MASK)
-+ >> BMI_FIFO_PIPELINE_DEPTH_SHIFT) + 1);
-+ p_FmPort->p_FmPortDriverParam->txFifoLowComfLevel = (((tmpReg
-+ & BMI_TX_LOW_COMF_MASK) >> BMI_TX_LOW_COMF_SHIFT) + 1)
-+ * BMI_FIFO_UNITS;
-+
-+ p_FmPort->p_FmPortDriverParam->deqType = DEFAULT_PORT_deqType;
-+ p_FmPort->p_FmPortDriverParam->deqPrefetchOption =
-+ DEFAULT_PORT_deqPrefetchOption;
-+ p_FmPort->p_FmPortDriverParam->deqHighPriority =
-+ (bool)((p_FmPort->portType == e_FM_PORT_TYPE_TX) ? DEFAULT_PORT_deqHighPriority_1G :
-+ DEFAULT_PORT_deqHighPriority_10G);
-+ p_FmPort->p_FmPortDriverParam->deqByteCnt =
-+ (uint16_t)(
-+ (p_FmPort->portType == e_FM_PORT_TYPE_TX) ? DEFAULT_PORT_deqByteCnt_1G :
-+ DEFAULT_PORT_deqByteCnt_10G);
-+ break;
-+ case (e_FM_PORT_TYPE_OH_OFFLINE_PARSING):
-+ p_FmPort->p_FmPortDriverParam->errorsToDiscard =
-+ DEFAULT_PORT_errorsToDiscard;
-+#if (DPAA_VERSION >= 11)
-+ p_FmPort->p_FmPortDriverParam->noScatherGather =
-+ DEFAULT_PORT_noScatherGather;
-+#endif /* (DPAA_VERSION >= 11) */
-+ case (e_FM_PORT_TYPE_OH_HOST_COMMAND):
-+ p_FmPort->p_FmPortDriverParam->deqPrefetchOption =
-+ DEFAULT_PORT_deqPrefetchOption_HC;
-+ p_FmPort->p_FmPortDriverParam->deqHighPriority =
-+ DEFAULT_PORT_deqHighPriority_1G;
-+ p_FmPort->p_FmPortDriverParam->deqType = DEFAULT_PORT_deqType;
-+ p_FmPort->p_FmPortDriverParam->deqByteCnt =
-+ DEFAULT_PORT_deqByteCnt_1G;
-+
-+ tmpReg =
-+ GET_UINT32(p_FmPort->p_FmPortBmiRegs->ohPortBmiRegs.fmbm_ofp);
-+ p_FmPort->p_FmPortDriverParam->dfltCfg.tx_fifo_deq_pipeline_depth =
-+ (uint8_t)(((tmpReg & BMI_FIFO_PIPELINE_DEPTH_MASK)
-+ >> BMI_FIFO_PIPELINE_DEPTH_SHIFT) + 1);
-+ if ((p_FmPort->portType == e_FM_PORT_TYPE_OH_HOST_COMMAND)
-+ && (p_FmPortParams->portId != FM_OH_PORT_ID))
-+ {
-+ /* Overwrite HC defaults */
-+ p_FmPort->p_FmPortDriverParam->dfltCfg.tx_fifo_deq_pipeline_depth =
-+ DEFAULT_PORT_fifoDeqPipelineDepth_OH;
-+ }
-+
-+#ifndef FM_FRAME_END_PARAMS_FOR_OP
-+ if (p_FmPort->fmRevInfo.majorRev < 6)
-+ p_FmPort->p_FmPortDriverParam->cheksumLastBytesIgnore = DEFAULT_notSupported;
-+#endif /* !FM_FRAME_END_PARAMS_FOR_OP */
-+
-+#ifndef FM_DEQ_PIPELINE_PARAMS_FOR_OP
-+ if (!((p_FmPort->fmRevInfo.majorRev == 4) ||
-+ (p_FmPort->fmRevInfo.majorRev >= 6)))
-+ p_FmPort->p_FmPortDriverParam->dfltCfg.tx_fifo_deq_pipeline_depth = DEFAULT_notSupported;
-+#endif /* !FM_DEQ_PIPELINE_PARAMS_FOR_OP */
-+ break;
-+
-+ default:
-+ XX_Free(p_FmPort->p_FmPortDriverParam);
-+ XX_Free(p_FmPort);
-+ REPORT_ERROR(MAJOR, E_INVALID_STATE, ("Invalid port type"));
-+ return NULL;
-+ }
-+#ifdef FM_QMI_NO_DEQ_OPTIONS_SUPPORT
-+ if (p_FmPort->fmRevInfo.majorRev == 4)
-+ p_FmPort->p_FmPortDriverParam->deqPrefetchOption = (e_FmPortDeqPrefetchOption)DEFAULT_notSupported;
-+#endif /* FM_QMI_NO_DEQ_OPTIONS_SUPPORT */
-+
-+ p_FmPort->imEn = p_FmPortParams->independentModeEnable;
-+
-+ if (p_FmPort->imEn)
-+ {
-+ if ((p_FmPort->portType == e_FM_PORT_TYPE_TX)
-+ || (p_FmPort->portType == e_FM_PORT_TYPE_TX_10G))
-+ p_FmPort->p_FmPortDriverParam->dfltCfg.tx_fifo_deq_pipeline_depth =
-+ DEFAULT_PORT_fifoDeqPipelineDepth_IM;
-+ FmPortConfigIM(p_FmPort, p_FmPortParams);
-+ }
-+ else
-+ {
-+ switch (p_FmPort->portType)
-+ {
-+ case (e_FM_PORT_TYPE_RX):
-+ case (e_FM_PORT_TYPE_RX_10G):
-+ /* Initialize FM port parameters for initialization phase only */
-+ memcpy(&p_FmPort->p_FmPortDriverParam->extBufPools,
-+ &p_FmPortParams->specificParams.rxParams.extBufPools,
-+ sizeof(t_FmExtPools));
-+ p_FmPort->p_FmPortDriverParam->errFqid =
-+ p_FmPortParams->specificParams.rxParams.errFqid;
-+ p_FmPort->p_FmPortDriverParam->dfltFqid =
-+ p_FmPortParams->specificParams.rxParams.dfltFqid;
-+ p_FmPort->p_FmPortDriverParam->liodnOffset =
-+ p_FmPortParams->specificParams.rxParams.liodnOffset;
-+ break;
-+ case (e_FM_PORT_TYPE_OH_OFFLINE_PARSING):
-+ case (e_FM_PORT_TYPE_TX):
-+ case (e_FM_PORT_TYPE_TX_10G):
-+ case (e_FM_PORT_TYPE_OH_HOST_COMMAND):
-+ p_FmPort->p_FmPortDriverParam->errFqid =
-+ p_FmPortParams->specificParams.nonRxParams.errFqid;
-+ p_FmPort->p_FmPortDriverParam->deqSubPortal =
-+ (uint8_t)(p_FmPortParams->specificParams.nonRxParams.qmChannel
-+ & QMI_DEQ_CFG_SUBPORTAL_MASK);
-+ p_FmPort->p_FmPortDriverParam->dfltFqid =
-+ p_FmPortParams->specificParams.nonRxParams.dfltFqid;
-+ break;
-+ default:
-+ XX_Free(p_FmPort->p_FmPortDriverParam);
-+ XX_Free(p_FmPort);
-+ REPORT_ERROR(MAJOR, E_INVALID_STATE, ("Invalid port type"));
-+ return NULL;
-+ }
-+ }
-+
-+ memset(p_FmPort->name, 0, (sizeof(char)) * MODULE_NAME_SIZE);
-+ if (Sprint(
-+ p_FmPort->name,
-+ "FM-%d-port-%s-%d",
-+ FmGetId(p_FmPort->h_Fm),
-+ ((p_FmPort->portType == e_FM_PORT_TYPE_OH_OFFLINE_PARSING
-+ || (p_FmPort->portType == e_FM_PORT_TYPE_OH_HOST_COMMAND)) ? "OH" :
-+ (p_FmPort->portType == e_FM_PORT_TYPE_RX ? "1g-RX" :
-+ (p_FmPort->portType == e_FM_PORT_TYPE_TX ? "1g-TX" :
-+ (p_FmPort->portType
-+ == e_FM_PORT_TYPE_RX_10G ? "10g-RX" :
-+ "10g-TX")))),
-+ p_FmPort->portId) == 0)
-+ {
-+ XX_Free(p_FmPort->p_FmPortDriverParam);
-+ XX_Free(p_FmPort);
-+ REPORT_ERROR(MAJOR, E_INVALID_STATE, ("Sprint failed"));
-+ return NULL;
-+ }
-+
-+ p_FmPort->h_Spinlock = XX_InitSpinlock();
-+ if (!p_FmPort->h_Spinlock)
-+ {
-+ XX_Free(p_FmPort->p_FmPortDriverParam);
-+ XX_Free(p_FmPort);
-+ REPORT_ERROR(MAJOR, E_INVALID_STATE, ("Sprint failed"));
-+ return NULL;
-+ }
-+
-+ return p_FmPort;
-+}
-+
-+t_FmPort *rx_port = 0;
-+t_FmPort *tx_port = 0;
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_Init
-+
-+ @Description Initializes the FM module
-+
-+ @Param[in] h_FmPort - FM module descriptor
-+
-+ @Return E_OK on success; Error code otherwise.
-+ *//***************************************************************************/
-+t_Error FM_PORT_Init(t_Handle h_FmPort)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+ t_FmPortDriverParam *p_DriverParams;
-+ t_Error errCode;
-+ t_FmInterModulePortInitParams fmParams;
-+ t_FmRevisionInfo revInfo;
-+
-+ SANITY_CHECK_RETURN_ERROR(h_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort->p_FmPortDriverParam, E_INVALID_HANDLE);
-+
-+ errCode = FmSpBuildBufferStructure(
-+ &p_FmPort->p_FmPortDriverParam->intContext,
-+ &p_FmPort->p_FmPortDriverParam->bufferPrefixContent,
-+ &p_FmPort->p_FmPortDriverParam->bufMargins,
-+ &p_FmPort->bufferOffsets, &p_FmPort->internalBufferOffset);
-+ if (errCode != E_OK)
-+ RETURN_ERROR(MAJOR, errCode, NO_MSG);
-+#ifdef FM_HEAVY_TRAFFIC_HANG_ERRATA_FMAN_A005669
-+ if ((p_FmPort->p_FmPortDriverParam->bcbWorkaround) &&
-+ (p_FmPort->portType == e_FM_PORT_TYPE_RX))
-+ {
-+ p_FmPort->p_FmPortDriverParam->errorsToDiscard |= FM_PORT_FRM_ERR_PHYSICAL;
-+ if (!p_FmPort->fifoBufs.num)
-+ p_FmPort->fifoBufs.num = DEFAULT_PORT_numOfFifoBufs(p_FmPort->portType)*BMI_FIFO_UNITS;
-+ p_FmPort->fifoBufs.num += 4*KILOBYTE;
-+ }
-+#endif /* FM_HEAVY_TRAFFIC_HANG_ERRATA_FMAN_A005669 */
-+
-+ CHECK_INIT_PARAMETERS(p_FmPort, CheckInitParameters);
-+
-+ p_DriverParams = p_FmPort->p_FmPortDriverParam;
-+
-+ /* Set up flibs port structure */
-+ memset(&p_FmPort->port, 0, sizeof(struct fman_port));
-+ p_FmPort->port.type = (enum fman_port_type)p_FmPort->portType;
-+ FM_GetRevision(p_FmPort->h_Fm, &revInfo);
-+ p_FmPort->port.fm_rev_maj = revInfo.majorRev;
-+ p_FmPort->port.fm_rev_min = revInfo.minorRev;
-+ p_FmPort->port.bmi_regs =
-+ (union fman_port_bmi_regs *)UINT_TO_PTR(p_DriverParams->baseAddr + BMI_PORT_REGS_OFFSET);
-+ p_FmPort->port.qmi_regs =
-+ (struct fman_port_qmi_regs *)UINT_TO_PTR(p_DriverParams->baseAddr + QMI_PORT_REGS_OFFSET);
-+ p_FmPort->port.ext_pools_num = (uint8_t)((revInfo.majorRev == 4) ? 4 : 8);
-+ p_FmPort->port.im_en = p_FmPort->imEn;
-+ p_FmPort->p_FmPortPrsRegs =
-+ (t_FmPortPrsRegs *)UINT_TO_PTR(p_DriverParams->baseAddr + PRS_PORT_REGS_OFFSET);
-+
-+ if (((p_FmPort->portType == e_FM_PORT_TYPE_RX_10G)
-+ || (p_FmPort->portType == e_FM_PORT_TYPE_RX)) && !p_FmPort->imEn)
-+ {
-+ /* Call the external Buffer routine which also checks fifo
-+ size and updates it if necessary */
-+ /* define external buffer pools and pool depletion*/
-+ errCode = SetExtBufferPools(p_FmPort);
-+ if (errCode)
-+ RETURN_ERROR(MAJOR, errCode, NO_MSG);
-+ /* check if the largest external buffer pool is large enough */
-+ if (p_DriverParams->bufMargins.startMargins + MIN_EXT_BUF_SIZE
-+ + p_DriverParams->bufMargins.endMargins
-+ > p_FmPort->rxPoolsParams.largestBufSize)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("bufMargins.startMargins (%d) + minimum buf size (64) + bufMargins.endMargins (%d) is larger than maximum external buffer size (%d)", p_DriverParams->bufMargins.startMargins, p_DriverParams->bufMargins.endMargins, p_FmPort->rxPoolsParams.largestBufSize));
-+ }
-+ if (p_FmPort->portType == e_FM_PORT_TYPE_OH_OFFLINE_PARSING)
-+ {
-+ {
-+#ifdef FM_NO_OP_OBSERVED_POOLS
-+ t_FmRevisionInfo revInfo;
-+
-+ FM_GetRevision(p_FmPort->h_Fm, &revInfo);
-+ if ((revInfo.majorRev == 4) && (p_DriverParams->enBufPoolDepletion))
-+#endif /* FM_NO_OP_OBSERVED_POOLS */
-+ {
-+ /* define external buffer pools */
-+ errCode = SetExtBufferPools(p_FmPort);
-+ if (errCode)
-+ RETURN_ERROR(MAJOR, errCode, NO_MSG);
-+ }
-+ }
-+ }
-+
-+ /************************************************************/
-+ /* Call FM module routine for communicating parameters */
-+ /************************************************************/
-+ memset(&fmParams, 0, sizeof(fmParams));
-+ fmParams.hardwarePortId = p_FmPort->hardwarePortId;
-+ fmParams.portType = (e_FmPortType)p_FmPort->portType;
-+ fmParams.numOfTasks = (uint8_t)p_FmPort->tasks.num;
-+ fmParams.numOfExtraTasks = (uint8_t)p_FmPort->tasks.extra;
-+ fmParams.numOfOpenDmas = (uint8_t)p_FmPort->openDmas.num;
-+ fmParams.numOfExtraOpenDmas = (uint8_t)p_FmPort->openDmas.extra;
-+
-+ if (p_FmPort->fifoBufs.num)
-+ {
-+ errCode = VerifySizeOfFifo(p_FmPort);
-+ if (errCode != E_OK)
-+ RETURN_ERROR(MAJOR, errCode, NO_MSG);
-+ }
-+ fmParams.sizeOfFifo = p_FmPort->fifoBufs.num;
-+ fmParams.extraSizeOfFifo = p_FmPort->fifoBufs.extra;
-+ fmParams.independentMode = p_FmPort->imEn;
-+ fmParams.liodnOffset = p_DriverParams->liodnOffset;
-+ fmParams.liodnBase = p_DriverParams->liodnBase;
-+ fmParams.deqPipelineDepth =
-+ p_FmPort->p_FmPortDriverParam->dfltCfg.tx_fifo_deq_pipeline_depth;
-+ fmParams.maxFrameLength = p_FmPort->maxFrameLength;
-+#ifndef FM_DEQ_PIPELINE_PARAMS_FOR_OP
-+ if ((p_FmPort->portType == e_FM_PORT_TYPE_OH_OFFLINE_PARSING) ||
-+ (p_FmPort->portType == e_FM_PORT_TYPE_OH_HOST_COMMAND))
-+ {
-+ if (!((p_FmPort->fmRevInfo.majorRev == 4) ||
-+ (p_FmPort->fmRevInfo.majorRev >= 6)))
-+ /* HC ports do not have fifoDeqPipelineDepth, but it is needed only
-+ * for deq threshold calculation.
-+ */
-+ fmParams.deqPipelineDepth = 2;
-+ }
-+#endif /* !FM_DEQ_PIPELINE_PARAMS_FOR_OP */
-+
-+ errCode = FmGetSetPortParams(p_FmPort->h_Fm, &fmParams);
-+ if (errCode)
-+ RETURN_ERROR(MAJOR, errCode, NO_MSG);
-+
-+ /* get params for use in init */
-+ p_FmPort->fmMuramPhysBaseAddr =
-+ (uint64_t)((uint64_t)(fmParams.fmMuramPhysBaseAddr.low)
-+ | ((uint64_t)(fmParams.fmMuramPhysBaseAddr.high) << 32));
-+ p_FmPort->h_FmMuram = FmGetMuramHandle(p_FmPort->h_Fm);
-+
-+ errCode = InitLowLevelDriver(p_FmPort);
-+ if (errCode != E_OK)
-+ RETURN_ERROR(MAJOR, errCode, NO_MSG);
-+
-+ FmPortDriverParamFree(p_FmPort);
-+
-+#if (DPAA_VERSION >= 11)
-+ if ((p_FmPort->portType == e_FM_PORT_TYPE_RX_10G)
-+ || (p_FmPort->portType == e_FM_PORT_TYPE_RX)
-+ || (p_FmPort->portType == e_FM_PORT_TYPE_OH_OFFLINE_PARSING))
-+ {
-+ t_FmPcdCtrlParamsPage *p_ParamsPage;
-+
-+ FmPortSetGprFunc(p_FmPort, e_FM_PORT_GPR_MURAM_PAGE,
-+ (void**)&p_ParamsPage);
-+ ASSERT_COND(p_ParamsPage);
-+
-+ WRITE_UINT32(p_ParamsPage->misc, FM_CTL_PARAMS_PAGE_ALWAYS_ON);
-+#ifdef FM_OP_NO_VSP_NO_RELEASE_ERRATA_FMAN_A006675
-+ if (p_FmPort->portType == e_FM_PORT_TYPE_OH_OFFLINE_PARSING)
-+ {
-+ WRITE_UINT32(
-+ p_ParamsPage->misc,
-+ (GET_UINT32(p_ParamsPage->misc) | FM_CTL_PARAMS_PAGE_OP_FIX_EN));
-+ WRITE_UINT32(
-+ p_ParamsPage->discardMask,
-+ GET_UINT32(p_FmPort->p_FmPortBmiRegs->ohPortBmiRegs.fmbm_ofsdm));
-+ }
-+#endif /* FM_OP_NO_VSP_NO_RELEASE_ERRATA_FMAN_A006675 */
-+#ifdef FM_ERROR_VSP_NO_MATCH_SW006
-+ if (p_FmPort->portType == e_FM_PORT_TYPE_OH_OFFLINE_PARSING)
-+ WRITE_UINT32(
-+ p_ParamsPage->errorsDiscardMask,
-+ (GET_UINT32(p_FmPort->p_FmPortBmiRegs->ohPortBmiRegs.fmbm_ofsdm) | GET_UINT32(p_FmPort->p_FmPortBmiRegs->ohPortBmiRegs.fmbm_ofsem)));
-+ else
-+ WRITE_UINT32(
-+ p_ParamsPage->errorsDiscardMask,
-+ (GET_UINT32(p_FmPort->p_FmPortBmiRegs->rxPortBmiRegs.fmbm_rfsdm) | GET_UINT32(p_FmPort->p_FmPortBmiRegs->rxPortBmiRegs.fmbm_rfsem)));
-+#endif /* FM_ERROR_VSP_NO_MATCH_SW006 */
-+ }
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+ if (p_FmPort->deepSleepVars.autoResMaxSizes)
-+ FmPortConfigAutoResForDeepSleepSupport1(p_FmPort);
-+ return E_OK;
-+}
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_Free
-+
-+ @Description Frees all resources that were assigned to FM module.
-+
-+ Calling this routine invalidates the descriptor.
-+
-+ @Param[in] h_FmPort - FM module descriptor
-+
-+ @Return E_OK on success; Error code otherwise.
-+ *//***************************************************************************/
-+t_Error FM_PORT_Free(t_Handle h_FmPort)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+ t_FmInterModulePortFreeParams fmParams;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+
-+ if (p_FmPort->pcdEngines)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_STATE,
-+ ("Trying to free a port with PCD. FM_PORT_DeletePCD must be called first."));
-+
-+ if (p_FmPort->enabled)
-+ {
-+ if (FM_PORT_Disable(p_FmPort) != E_OK)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("FM_PORT_Disable FAILED"));
-+ }
-+
-+ if (p_FmPort->imEn)
-+ FmPortImFree(p_FmPort);
-+
-+ FmPortDriverParamFree(p_FmPort);
-+
-+ memset(&fmParams, 0, sizeof(fmParams));
-+ fmParams.hardwarePortId = p_FmPort->hardwarePortId;
-+ fmParams.portType = (e_FmPortType)p_FmPort->portType;
-+ fmParams.deqPipelineDepth =
-+ p_FmPort->p_FmPortDriverParam->dfltCfg.tx_fifo_deq_pipeline_depth;
-+
-+ FmFreePortParams(p_FmPort->h_Fm, &fmParams);
-+
-+#if (DPAA_VERSION >= 11)
-+ if (FmVSPFreeForPort(p_FmPort->h_Fm, p_FmPort->portType, p_FmPort->portId)
-+ != E_OK)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("VSP free of port FAILED"));
-+
-+ if (p_FmPort->p_ParamsPage)
-+ FM_MURAM_FreeMem(p_FmPort->h_FmMuram, p_FmPort->p_ParamsPage);
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+ if (p_FmPort->h_Spinlock)
-+ XX_FreeSpinlock(p_FmPort->h_Spinlock);
-+
-+ XX_Free(p_FmPort);
-+
-+ return E_OK;
-+}
-+
-+/*************************************************/
-+/* API Advanced Init unit functions */
-+/*************************************************/
-+
-+t_Error FM_PORT_ConfigNumOfOpenDmas(t_Handle h_FmPort, t_FmPortRsrc *p_OpenDmas)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort->p_FmPortDriverParam, E_INVALID_HANDLE);
-+
-+ p_FmPort->p_FmPortDriverParam->setNumOfOpenDmas = TRUE;
-+ memcpy(&p_FmPort->openDmas, p_OpenDmas, sizeof(t_FmPortRsrc));
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PORT_ConfigNumOfTasks(t_Handle h_FmPort, t_FmPortRsrc *p_NumOfTasks)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort->p_FmPortDriverParam, E_INVALID_HANDLE);
-+
-+ memcpy(&p_FmPort->tasks, p_NumOfTasks, sizeof(t_FmPortRsrc));
-+ p_FmPort->p_FmPortDriverParam->setNumOfTasks = TRUE;
-+ return E_OK;
-+}
-+
-+t_Error FM_PORT_ConfigSizeOfFifo(t_Handle h_FmPort, t_FmPortRsrc *p_SizeOfFifo)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort->p_FmPortDriverParam, E_INVALID_HANDLE);
-+
-+ p_FmPort->p_FmPortDriverParam->setSizeOfFifo = TRUE;
-+ memcpy(&p_FmPort->fifoBufs, p_SizeOfFifo, sizeof(t_FmPortRsrc));
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PORT_ConfigDeqHighPriority(t_Handle h_FmPort, bool highPri)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort->p_FmPortDriverParam, E_INVALID_HANDLE);
-+ if ((p_FmPort->portType == e_FM_PORT_TYPE_RX_10G)
-+ || (p_FmPort->portType == e_FM_PORT_TYPE_RX))
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, ("not available for Rx ports"));
-+
-+ p_FmPort->p_FmPortDriverParam->dfltCfg.deq_high_pri = highPri;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PORT_ConfigDeqType(t_Handle h_FmPort, e_FmPortDeqType deqType)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort->p_FmPortDriverParam, E_INVALID_HANDLE);
-+ if ((p_FmPort->portType == e_FM_PORT_TYPE_RX_10G)
-+ || (p_FmPort->portType == e_FM_PORT_TYPE_RX))
-+ RETURN_ERROR(MAJOR, E_INVALID_OPERATION,
-+ ("not available for Rx ports"));
-+
-+ p_FmPort->p_FmPortDriverParam->dfltCfg.deq_type =
-+ (enum fman_port_deq_type)deqType;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PORT_ConfigDeqPrefetchOption(
-+ t_Handle h_FmPort, e_FmPortDeqPrefetchOption deqPrefetchOption)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort->p_FmPortDriverParam, E_INVALID_HANDLE);
-+ if ((p_FmPort->portType == e_FM_PORT_TYPE_RX_10G)
-+ || (p_FmPort->portType == e_FM_PORT_TYPE_RX))
-+ RETURN_ERROR(MAJOR, E_INVALID_OPERATION,
-+ ("not available for Rx ports"));
-+ p_FmPort->p_FmPortDriverParam->dfltCfg.deq_prefetch_opt =
-+ (enum fman_port_deq_prefetch)deqPrefetchOption;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PORT_ConfigBackupPools(t_Handle h_FmPort,
-+ t_FmBackupBmPools *p_BackupBmPools)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort->p_FmPortDriverParam, E_INVALID_HANDLE);
-+ if ((p_FmPort->portType != e_FM_PORT_TYPE_RX_10G)
-+ && (p_FmPort->portType != e_FM_PORT_TYPE_RX))
-+ RETURN_ERROR(MAJOR, E_INVALID_OPERATION,
-+ ("available for Rx ports only"));
-+
-+ p_FmPort->p_FmPortDriverParam->p_BackupBmPools =
-+ (t_FmBackupBmPools *)XX_Malloc(sizeof(t_FmBackupBmPools));
-+ if (!p_FmPort->p_FmPortDriverParam->p_BackupBmPools)
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, ("p_BackupBmPools allocation failed"));
-+ memcpy(p_FmPort->p_FmPortDriverParam->p_BackupBmPools, p_BackupBmPools,
-+ sizeof(t_FmBackupBmPools));
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PORT_ConfigDeqByteCnt(t_Handle h_FmPort, uint16_t deqByteCnt)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort->p_FmPortDriverParam, E_INVALID_HANDLE);
-+ if ((p_FmPort->portType == e_FM_PORT_TYPE_RX_10G)
-+ || (p_FmPort->portType == e_FM_PORT_TYPE_RX))
-+ RETURN_ERROR(MAJOR, E_INVALID_OPERATION,
-+ ("not available for Rx ports"));
-+
-+ p_FmPort->p_FmPortDriverParam->dfltCfg.deq_byte_cnt = deqByteCnt;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PORT_ConfigBufferPrefixContent(
-+ t_Handle h_FmPort, t_FmBufferPrefixContent *p_FmBufferPrefixContent)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort->p_FmPortDriverParam, E_INVALID_HANDLE);
-+
-+ memcpy(&p_FmPort->p_FmPortDriverParam->bufferPrefixContent,
-+ p_FmBufferPrefixContent, sizeof(t_FmBufferPrefixContent));
-+ /* if dataAlign was not initialized by user, we return to driver's default */
-+ if (!p_FmPort->p_FmPortDriverParam->bufferPrefixContent.dataAlign)
-+ p_FmPort->p_FmPortDriverParam->bufferPrefixContent.dataAlign =
-+ DEFAULT_PORT_bufferPrefixContent_dataAlign;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PORT_ConfigCheksumLastBytesIgnore(t_Handle h_FmPort,
-+ uint8_t checksumLastBytesIgnore)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort->p_FmPortDriverParam, E_INVALID_HANDLE);
-+
-+ p_FmPort->p_FmPortDriverParam->dfltCfg.checksum_bytes_ignore =
-+ checksumLastBytesIgnore;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PORT_ConfigCutBytesFromEnd(t_Handle h_FmPort,
-+ uint8_t cutBytesFromEnd)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort->p_FmPortDriverParam, E_INVALID_HANDLE);
-+ if ((p_FmPort->portType != e_FM_PORT_TYPE_RX_10G)
-+ && (p_FmPort->portType != e_FM_PORT_TYPE_RX))
-+ RETURN_ERROR(MAJOR, E_INVALID_OPERATION,
-+ ("available for Rx ports only"));
-+
-+ p_FmPort->p_FmPortDriverParam->dfltCfg.rx_cut_end_bytes = cutBytesFromEnd;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PORT_ConfigPoolDepletion(t_Handle h_FmPort,
-+ t_FmBufPoolDepletion *p_BufPoolDepletion)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort->p_FmPortDriverParam, E_INVALID_HANDLE);
-+ if ((p_FmPort->portType != e_FM_PORT_TYPE_RX_10G)
-+ && (p_FmPort->portType != e_FM_PORT_TYPE_RX))
-+ RETURN_ERROR(MAJOR, E_INVALID_OPERATION,
-+ ("available for Rx ports only"));
-+
-+ p_FmPort->p_FmPortDriverParam->enBufPoolDepletion = TRUE;
-+ memcpy(&p_FmPort->p_FmPortDriverParam->bufPoolDepletion, p_BufPoolDepletion,
-+ sizeof(t_FmBufPoolDepletion));
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PORT_ConfigObservedPoolDepletion(
-+ t_Handle h_FmPort,
-+ t_FmPortObservedBufPoolDepletion *p_FmPortObservedBufPoolDepletion)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort->p_FmPortDriverParam, E_INVALID_HANDLE);
-+ if (p_FmPort->portType != e_FM_PORT_TYPE_OH_OFFLINE_PARSING)
-+ RETURN_ERROR(MAJOR, E_INVALID_OPERATION,
-+ ("available for OP ports only"));
-+
-+ p_FmPort->p_FmPortDriverParam->enBufPoolDepletion = TRUE;
-+ memcpy(&p_FmPort->p_FmPortDriverParam->bufPoolDepletion,
-+ &p_FmPortObservedBufPoolDepletion->poolDepletionParams,
-+ sizeof(t_FmBufPoolDepletion));
-+ memcpy(&p_FmPort->p_FmPortDriverParam->extBufPools,
-+ &p_FmPortObservedBufPoolDepletion->poolsParams,
-+ sizeof(t_FmExtPools));
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PORT_ConfigExtBufPools(t_Handle h_FmPort, t_FmExtPools *p_FmExtPools)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort->p_FmPortDriverParam, E_INVALID_HANDLE);
-+
-+ if (p_FmPort->portType != e_FM_PORT_TYPE_OH_OFFLINE_PARSING)
-+ RETURN_ERROR(MAJOR, E_INVALID_OPERATION,
-+ ("available for OP ports only"));
-+
-+ memcpy(&p_FmPort->p_FmPortDriverParam->extBufPools, p_FmExtPools,
-+ sizeof(t_FmExtPools));
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PORT_ConfigDontReleaseTxBufToBM(t_Handle h_FmPort)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort->p_FmPortDriverParam, E_INVALID_HANDLE);
-+ if ((p_FmPort->portType != e_FM_PORT_TYPE_TX_10G)
-+ && (p_FmPort->portType != e_FM_PORT_TYPE_TX))
-+ RETURN_ERROR(MAJOR, E_INVALID_OPERATION,
-+ ("available for Tx ports only"));
-+
-+ p_FmPort->p_FmPortDriverParam->dontReleaseBuf = TRUE;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PORT_ConfigDfltColor(t_Handle h_FmPort, e_FmPortColor color)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort->p_FmPortDriverParam, E_INVALID_HANDLE);
-+ p_FmPort->p_FmPortDriverParam->dfltCfg.color = (enum fman_port_color)color;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PORT_ConfigSyncReq(t_Handle h_FmPort, bool syncReq)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort->p_FmPortDriverParam, E_INVALID_HANDLE);
-+
-+ if ((p_FmPort->portType == e_FM_PORT_TYPE_TX_10G)
-+ || (p_FmPort->portType == e_FM_PORT_TYPE_TX))
-+ RETURN_ERROR(MAJOR, E_INVALID_OPERATION,
-+ ("Not available for Tx ports"));
-+
-+ p_FmPort->p_FmPortDriverParam->dfltCfg.sync_req = syncReq;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PORT_ConfigFrmDiscardOverride(t_Handle h_FmPort, bool override)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort->p_FmPortDriverParam, E_INVALID_HANDLE);
-+ if ((p_FmPort->portType == e_FM_PORT_TYPE_TX_10G)
-+ || (p_FmPort->portType == e_FM_PORT_TYPE_TX))
-+ RETURN_ERROR(MAJOR, E_INVALID_OPERATION,
-+ ("Not available for Tx ports"));
-+
-+ p_FmPort->p_FmPortDriverParam->dfltCfg.discard_override = override;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PORT_ConfigErrorsToDiscard(t_Handle h_FmPort,
-+ fmPortFrameErrSelect_t errs)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort->p_FmPortDriverParam, E_INVALID_HANDLE);
-+ if ((p_FmPort->portType != e_FM_PORT_TYPE_RX_10G)
-+ && (p_FmPort->portType != e_FM_PORT_TYPE_RX)
-+ && (p_FmPort->portType != e_FM_PORT_TYPE_OH_OFFLINE_PARSING))
-+ RETURN_ERROR( MAJOR, E_INVALID_OPERATION,
-+ ("available for Rx and offline parsing ports only"));
-+
-+ p_FmPort->p_FmPortDriverParam->errorsToDiscard = errs;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PORT_ConfigDmaSwapData(t_Handle h_FmPort, e_FmDmaSwapOption swapData)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort->p_FmPortDriverParam, E_INVALID_HANDLE);
-+
-+ p_FmPort->p_FmPortDriverParam->dfltCfg.dma_swap_data =
-+ (enum fman_port_dma_swap)swapData;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PORT_ConfigDmaIcCacheAttr(t_Handle h_FmPort,
-+ e_FmDmaCacheOption intContextCacheAttr)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort->p_FmPortDriverParam, E_INVALID_HANDLE);
-+
-+ p_FmPort->p_FmPortDriverParam->dfltCfg.dma_ic_stash_on =
-+ (bool)(intContextCacheAttr == e_FM_DMA_STASH);
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PORT_ConfigDmaHdrAttr(t_Handle h_FmPort,
-+ e_FmDmaCacheOption headerCacheAttr)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort->p_FmPortDriverParam, E_INVALID_HANDLE);
-+
-+ p_FmPort->p_FmPortDriverParam->dfltCfg.dma_header_stash_on =
-+ (bool)(headerCacheAttr == e_FM_DMA_STASH);
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PORT_ConfigDmaScatterGatherAttr(
-+ t_Handle h_FmPort, e_FmDmaCacheOption scatterGatherCacheAttr)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort->p_FmPortDriverParam, E_INVALID_HANDLE);
-+
-+ p_FmPort->p_FmPortDriverParam->dfltCfg.dma_sg_stash_on =
-+ (bool)(scatterGatherCacheAttr == e_FM_DMA_STASH);
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PORT_ConfigDmaWriteOptimize(t_Handle h_FmPort, bool optimize)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort->p_FmPortDriverParam, E_INVALID_HANDLE);
-+
-+ if ((p_FmPort->portType == e_FM_PORT_TYPE_TX_10G)
-+ || (p_FmPort->portType == e_FM_PORT_TYPE_TX))
-+ RETURN_ERROR(MAJOR, E_INVALID_OPERATION,
-+ ("Not available for Tx ports"));
-+
-+ p_FmPort->p_FmPortDriverParam->dfltCfg.dma_write_optimize = optimize;
-+
-+ return E_OK;
-+}
-+
-+#if (DPAA_VERSION >= 11)
-+t_Error FM_PORT_ConfigNoScatherGather(t_Handle h_FmPort, bool noScatherGather)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+
-+ UNUSED(noScatherGather);
-+ UNUSED(p_FmPort);
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort->p_FmPortDriverParam, E_INVALID_HANDLE);
-+
-+ p_FmPort->p_FmPortDriverParam->noScatherGather = noScatherGather;
-+
-+ return E_OK;
-+}
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+t_Error FM_PORT_ConfigForwardReuseIntContext(t_Handle h_FmPort,
-+ bool forwardReuse)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort->p_FmPortDriverParam, E_INVALID_HANDLE);
-+
-+ if ((p_FmPort->portType != e_FM_PORT_TYPE_RX_10G)
-+ && (p_FmPort->portType != e_FM_PORT_TYPE_RX))
-+ RETURN_ERROR(MAJOR, E_INVALID_OPERATION,
-+ ("available for Rx ports only"));
-+
-+ p_FmPort->p_FmPortDriverParam->forwardReuseIntContext = forwardReuse;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PORT_ConfigMaxFrameLength(t_Handle h_FmPort, uint16_t length)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort->p_FmPortDriverParam, E_INVALID_HANDLE);
-+
-+ p_FmPort->maxFrameLength = length;
-+
-+ return E_OK;
-+}
-+
-+#ifdef FM_HEAVY_TRAFFIC_HANG_ERRATA_FMAN_A005669
-+t_Error FM_PORT_ConfigBCBWorkaround(t_Handle h_FmPort)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort->p_FmPortDriverParam, E_INVALID_HANDLE);
-+
-+ p_FmPort->p_FmPortDriverParam->bcbWorkaround = TRUE;
-+
-+ return E_OK;
-+}
-+#endif /* FM_HEAVY_TRAFFIC_HANG_ERRATA_FMAN_A005669 */
-+
-+/****************************************************/
-+/* Hidden-DEBUG Only API */
-+/****************************************************/
-+
-+t_Error FM_PORT_ConfigTxFifoMinFillLevel(t_Handle h_FmPort,
-+ uint32_t minFillLevel)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort->p_FmPortDriverParam, E_INVALID_HANDLE);
-+ if ((p_FmPort->portType != e_FM_PORT_TYPE_TX_10G)
-+ && (p_FmPort->portType != e_FM_PORT_TYPE_TX))
-+ RETURN_ERROR(MAJOR, E_INVALID_OPERATION,
-+ ("available for Tx ports only"));
-+
-+ p_FmPort->p_FmPortDriverParam->dfltCfg.tx_fifo_min_level = minFillLevel;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PORT_ConfigFifoDeqPipelineDepth(t_Handle h_FmPort,
-+ uint8_t deqPipelineDepth)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort->p_FmPortDriverParam, E_INVALID_HANDLE);
-+
-+ if ((p_FmPort->portType == e_FM_PORT_TYPE_RX_10G)
-+ || (p_FmPort->portType == e_FM_PORT_TYPE_RX))
-+ RETURN_ERROR(MAJOR, E_INVALID_OPERATION,
-+ ("Not available for Rx ports"));
-+
-+ if (p_FmPort->imEn)
-+ RETURN_ERROR(MAJOR, E_INVALID_OPERATION,
-+ ("Not available for IM ports!"));
-+
-+ p_FmPort->p_FmPortDriverParam->dfltCfg.tx_fifo_deq_pipeline_depth =
-+ deqPipelineDepth;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PORT_ConfigTxFifoLowComfLevel(t_Handle h_FmPort,
-+ uint32_t fifoLowComfLevel)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort->p_FmPortDriverParam, E_INVALID_HANDLE);
-+ if ((p_FmPort->portType != e_FM_PORT_TYPE_TX_10G)
-+ && (p_FmPort->portType != e_FM_PORT_TYPE_TX))
-+ RETURN_ERROR(MAJOR, E_INVALID_OPERATION,
-+ ("available for Tx ports only"));
-+
-+ p_FmPort->p_FmPortDriverParam->dfltCfg.tx_fifo_low_comf_level =
-+ fifoLowComfLevel;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PORT_ConfigRxFifoThreshold(t_Handle h_FmPort, uint32_t fifoThreshold)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort->p_FmPortDriverParam, E_INVALID_HANDLE);
-+ if ((p_FmPort->portType != e_FM_PORT_TYPE_RX_10G)
-+ && (p_FmPort->portType != e_FM_PORT_TYPE_RX))
-+ RETURN_ERROR(MAJOR, E_INVALID_OPERATION,
-+ ("available for Rx ports only"));
-+
-+ p_FmPort->p_FmPortDriverParam->dfltCfg.rx_fifo_thr = fifoThreshold;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PORT_ConfigRxFifoPriElevationLevel(t_Handle h_FmPort,
-+ uint32_t priElevationLevel)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort->p_FmPortDriverParam, E_INVALID_HANDLE);
-+ if ((p_FmPort->portType != e_FM_PORT_TYPE_RX_10G)
-+ && (p_FmPort->portType != e_FM_PORT_TYPE_RX))
-+ RETURN_ERROR(MAJOR, E_INVALID_OPERATION,
-+ ("available for Rx ports only"));
-+
-+ p_FmPort->p_FmPortDriverParam->dfltCfg.rx_pri_elevation = priElevationLevel;
-+
-+ return E_OK;
-+}
-+/****************************************************/
-+/* API Run-time Control unit functions */
-+/****************************************************/
-+
-+t_Error FM_PORT_SetNumOfOpenDmas(t_Handle h_FmPort,
-+ t_FmPortRsrc *p_NumOfOpenDmas)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+ t_Error err;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmPort->p_FmPortDriverParam, E_INVALID_HANDLE);
-+
-+ if ((!p_NumOfOpenDmas->num) || (p_NumOfOpenDmas->num > MAX_NUM_OF_DMAS))
-+ RETURN_ERROR( MAJOR, E_INVALID_VALUE,
-+ ("openDmas-num can't be larger than %d", MAX_NUM_OF_DMAS));
-+ if (p_NumOfOpenDmas->extra > MAX_NUM_OF_EXTRA_DMAS)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("openDmas-extra can't be larger than %d", MAX_NUM_OF_EXTRA_DMAS));
-+ err = FmSetNumOfOpenDmas(p_FmPort->h_Fm, p_FmPort->hardwarePortId,
-+ (uint8_t*)&p_NumOfOpenDmas->num,
-+ (uint8_t*)&p_NumOfOpenDmas->extra, FALSE);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+
-+ memcpy(&p_FmPort->openDmas, p_NumOfOpenDmas, sizeof(t_FmPortRsrc));
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PORT_SetNumOfTasks(t_Handle h_FmPort, t_FmPortRsrc *p_NumOfTasks)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+ t_Error err;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmPort->p_FmPortDriverParam, E_INVALID_HANDLE);
-+
-+ /* only driver uses host command port, so ASSERT rather than RETURN_ERROR */
-+ ASSERT_COND(p_FmPort->portType != e_FM_PORT_TYPE_OH_HOST_COMMAND);
-+
-+ if ((!p_NumOfTasks->num) || (p_NumOfTasks->num > MAX_NUM_OF_TASKS))
-+ RETURN_ERROR(
-+ MAJOR, E_INVALID_VALUE,
-+ ("NumOfTasks-num can't be larger than %d", MAX_NUM_OF_TASKS));
-+ if (p_NumOfTasks->extra > MAX_NUM_OF_EXTRA_TASKS)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("NumOfTasks-extra can't be larger than %d", MAX_NUM_OF_EXTRA_TASKS));
-+
-+ err = FmSetNumOfTasks(p_FmPort->h_Fm, p_FmPort->hardwarePortId,
-+ (uint8_t*)&p_NumOfTasks->num,
-+ (uint8_t*)&p_NumOfTasks->extra, FALSE);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+
-+ /* update driver's struct */
-+ memcpy(&p_FmPort->tasks, p_NumOfTasks, sizeof(t_FmPortRsrc));
-+ return E_OK;
-+}
-+
-+t_Error FM_PORT_SetSizeOfFifo(t_Handle h_FmPort, t_FmPortRsrc *p_SizeOfFifo)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+ t_Error err;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmPort->p_FmPortDriverParam, E_INVALID_HANDLE);
-+
-+ if (!p_SizeOfFifo->num || (p_SizeOfFifo->num > MAX_PORT_FIFO_SIZE))
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("SizeOfFifo-num has to be in the range of 256 - %d", MAX_PORT_FIFO_SIZE));
-+ if (p_SizeOfFifo->num % BMI_FIFO_UNITS)
-+ RETURN_ERROR(
-+ MAJOR, E_INVALID_VALUE,
-+ ("SizeOfFifo-num has to be divisible by %d", BMI_FIFO_UNITS));
-+ if ((p_FmPort->portType == e_FM_PORT_TYPE_RX)
-+ || (p_FmPort->portType == e_FM_PORT_TYPE_RX_10G))
-+ {
-+ /* extra FIFO size (allowed only to Rx ports) */
-+ if (p_SizeOfFifo->extra % BMI_FIFO_UNITS)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("SizeOfFifo-extra has to be divisible by %d", BMI_FIFO_UNITS));
-+ }
-+ else
-+ if (p_SizeOfFifo->extra)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE,
-+ (" No SizeOfFifo-extra for non Rx ports"));
-+
-+ memcpy(&p_FmPort->fifoBufs, p_SizeOfFifo, sizeof(t_FmPortRsrc));
-+
-+ /* we do not change user's parameter */
-+ err = VerifySizeOfFifo(p_FmPort);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+
-+ err = FmSetSizeOfFifo(p_FmPort->h_Fm, p_FmPort->hardwarePortId,
-+ &p_SizeOfFifo->num, &p_SizeOfFifo->extra, FALSE);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+
-+ return E_OK;
-+}
-+
-+uint32_t FM_PORT_GetBufferDataOffset(t_Handle h_FmPort)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+
-+ SANITY_CHECK_RETURN_VALUE(p_FmPort, E_INVALID_HANDLE, 0);
-+ SANITY_CHECK_RETURN_VALUE(!p_FmPort->p_FmPortDriverParam, E_INVALID_STATE,
-+ 0);
-+
-+ return p_FmPort->bufferOffsets.dataOffset;
-+}
-+
-+uint8_t * FM_PORT_GetBufferICInfo(t_Handle h_FmPort, char *p_Data)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+
-+ SANITY_CHECK_RETURN_VALUE(p_FmPort, E_INVALID_HANDLE, NULL);
-+ SANITY_CHECK_RETURN_VALUE(!p_FmPort->p_FmPortDriverParam, E_INVALID_STATE,
-+ NULL);
-+
-+ if (p_FmPort->bufferOffsets.pcdInfoOffset == ILLEGAL_BASE)
-+ return NULL;
-+
-+ return (uint8_t *)PTR_MOVE(p_Data, p_FmPort->bufferOffsets.pcdInfoOffset);
-+}
-+
-+t_FmPrsResult * FM_PORT_GetBufferPrsResult(t_Handle h_FmPort, char *p_Data)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+
-+ SANITY_CHECK_RETURN_VALUE(p_FmPort, E_INVALID_HANDLE, NULL);
-+ SANITY_CHECK_RETURN_VALUE(!p_FmPort->p_FmPortDriverParam, E_INVALID_STATE,
-+ NULL);
-+
-+ if (p_FmPort->bufferOffsets.prsResultOffset == ILLEGAL_BASE)
-+ return NULL;
-+
-+ return (t_FmPrsResult *)PTR_MOVE(p_Data, p_FmPort->bufferOffsets.prsResultOffset);
-+}
-+
-+uint64_t * FM_PORT_GetBufferTimeStamp(t_Handle h_FmPort, char *p_Data)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+
-+ SANITY_CHECK_RETURN_VALUE(p_FmPort, E_INVALID_HANDLE, NULL);
-+ SANITY_CHECK_RETURN_VALUE(!p_FmPort->p_FmPortDriverParam, E_INVALID_STATE,
-+ NULL);
-+
-+ if (p_FmPort->bufferOffsets.timeStampOffset == ILLEGAL_BASE)
-+ return NULL;
-+
-+ return (uint64_t *)PTR_MOVE(p_Data, p_FmPort->bufferOffsets.timeStampOffset);
-+}
-+
-+uint8_t * FM_PORT_GetBufferHashResult(t_Handle h_FmPort, char *p_Data)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+
-+ SANITY_CHECK_RETURN_VALUE(p_FmPort, E_INVALID_HANDLE, NULL);
-+ SANITY_CHECK_RETURN_VALUE(!p_FmPort->p_FmPortDriverParam, E_INVALID_STATE,
-+ NULL);
-+
-+ if (p_FmPort->bufferOffsets.hashResultOffset == ILLEGAL_BASE)
-+ return NULL;
-+
-+ return (uint8_t *)PTR_MOVE(p_Data, p_FmPort->bufferOffsets.hashResultOffset);
-+}
-+
-+t_Error FM_PORT_Disable(t_Handle h_FmPort)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+ int err;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmPort->p_FmPortDriverParam, E_INVALID_STATE);
-+
-+ if (p_FmPort->imEn)
-+ FmPortImDisable(p_FmPort);
-+
-+ err = fman_port_disable(&p_FmPort->port);
-+ if (err == -EBUSY)
-+ {
-+ DBG(WARNING, ("%s: BMI or QMI is Busy. Port forced down",
-+ p_FmPort->name));
-+ }
-+ else
-+ if (err != 0)
-+ {
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("fman_port_disable"));
-+ }
-+
-+ p_FmPort->enabled = FALSE;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PORT_Enable(t_Handle h_FmPort)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+ int err;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmPort->p_FmPortDriverParam, E_INVALID_STATE);
-+
-+ /* Used by FM_PORT_Free routine as indication
-+ if to disable port. Thus set it to TRUE prior
-+ to enabling itself. This way if part of enable
-+ process fails there will be still things
-+ to disable during Free. For example, if BMI
-+ enable succeeded but QMI failed, still BMI
-+ needs to be disabled by Free. */
-+ p_FmPort->enabled = TRUE;
-+
-+ if (p_FmPort->imEn)
-+ FmPortImEnable(p_FmPort);
-+
-+ err = fman_port_enable(&p_FmPort->port);
-+ if (err != 0)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("fman_port_enable"));
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PORT_SetRateLimit(t_Handle h_FmPort, t_FmPortRateLimit *p_RateLimit)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+ uint8_t factor, countUnitBit;
-+ uint16_t baseGran;
-+ struct fman_port_rate_limiter params;
-+ int err;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmPort->p_FmPortDriverParam, E_INVALID_HANDLE);
-+
-+ switch (p_FmPort->portType)
-+ {
-+ case (e_FM_PORT_TYPE_TX_10G):
-+ case (e_FM_PORT_TYPE_TX):
-+ baseGran = BMI_RATE_LIMIT_GRAN_TX;
-+ break;
-+ case (e_FM_PORT_TYPE_OH_OFFLINE_PARSING):
-+ baseGran = BMI_RATE_LIMIT_GRAN_OP;
-+ break;
-+ default:
-+ RETURN_ERROR( MAJOR, E_INVALID_OPERATION,
-+ ("available for Tx and Offline parsing ports only"));
-+ }
-+
-+ countUnitBit = (uint8_t)FmGetTimeStampScale(p_FmPort->h_Fm); /* TimeStamp per nano seconds units */
-+ /* normally, we use 1 usec as the reference count */
-+ factor = 1;
-+ /* if ratelimit is too small for a 1usec factor, multiply the factor */
-+ while (p_RateLimit->rateLimit < baseGran / factor)
-+ {
-+ if (countUnitBit == 31)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("Rate limit is too small"));
-+
-+ countUnitBit++;
-+ factor <<= 1;
-+ }
-+ /* if ratelimit is too large for a 1usec factor, it is also larger than max rate*/
-+ if (p_RateLimit->rateLimit
-+ > ((uint32_t)baseGran * (1 << 10) * (uint32_t)factor))
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("Rate limit is too large"));
-+
-+ if (!p_RateLimit->maxBurstSize
-+ || (p_RateLimit->maxBurstSize > BMI_RATE_LIMIT_MAX_BURST_SIZE))
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("maxBurstSize must be between 1K and %dk", BMI_RATE_LIMIT_MAX_BURST_SIZE));
-+
-+ params.count_1micro_bit = (uint8_t)FmGetTimeStampScale(p_FmPort->h_Fm);
-+ params.high_burst_size_gran = FALSE;
-+ params.burst_size = p_RateLimit->maxBurstSize;
-+ params.rate = p_RateLimit->rateLimit;
-+ params.rate_factor = E_FMAN_PORT_RATE_DOWN_NONE;
-+
-+ if (p_FmPort->portType == e_FM_PORT_TYPE_OH_OFFLINE_PARSING)
-+ {
-+#ifndef FM_NO_ADVANCED_RATE_LIMITER
-+
-+ if ((p_FmPort->fmRevInfo.majorRev == 4)
-+ || (p_FmPort->fmRevInfo.majorRev >= 6))
-+ {
-+ params.high_burst_size_gran = TRUE;
-+ }
-+ else
-+#endif /* ! FM_NO_ADVANCED_RATE_LIMITER */
-+ {
-+ if (p_RateLimit->rateLimitDivider
-+ != e_FM_PORT_DUAL_RATE_LIMITER_NONE)
-+ RETURN_ERROR(MAJOR, E_NOT_SUPPORTED,
-+ ("FM_PORT_ConfigDualRateLimitScaleDown"));
-+
-+ if (p_RateLimit->maxBurstSize % 1000)
-+ {
-+ p_RateLimit->maxBurstSize =
-+ (uint16_t)((p_RateLimit->maxBurstSize / 1000) + 1);
-+ DBG(WARNING, ("rateLimit.maxBurstSize rounded up to %d", (p_RateLimit->maxBurstSize/1000+1)*1000));
-+ }
-+ else
-+ p_RateLimit->maxBurstSize = (uint16_t)(p_RateLimit->maxBurstSize
-+ / 1000);
-+ }
-+ params.rate_factor =
-+ (enum fman_port_rate_limiter_scale_down)p_RateLimit->rateLimitDivider;
-+ params.burst_size = p_RateLimit->maxBurstSize;
-+ }
-+
-+ err = fman_port_set_rate_limiter(&p_FmPort->port, &params);
-+ if (err != 0)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("fman_port_set_rate_limiter"));
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PORT_DeleteRateLimit(t_Handle h_FmPort)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+ int err;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmPort->p_FmPortDriverParam, E_INVALID_HANDLE);
-+
-+ if ((p_FmPort->portType == e_FM_PORT_TYPE_RX_10G)
-+ || (p_FmPort->portType == e_FM_PORT_TYPE_RX)
-+ || (p_FmPort->portType == e_FM_PORT_TYPE_OH_HOST_COMMAND))
-+ RETURN_ERROR( MAJOR, E_INVALID_OPERATION,
-+ ("available for Tx and Offline parsing ports only"));
-+
-+ err = fman_port_delete_rate_limiter(&p_FmPort->port);
-+ if (err != 0)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("fman_port_set_rate_limiter"));
-+ return E_OK;
-+}
-+
-+t_Error FM_PORT_SetPfcPrioritiesMappingToQmanWQ(t_Handle h_FmPort, uint8_t prio,
-+ uint8_t wq)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+ uint32_t tmpReg;
-+ uint32_t wqTmpReg;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmPort->p_FmPortDriverParam, E_INVALID_STATE);
-+
-+ if ((p_FmPort->portType != e_FM_PORT_TYPE_TX)
-+ && (p_FmPort->portType != e_FM_PORT_TYPE_TX_10G))
-+ RETURN_ERROR(MAJOR, E_INVALID_OPERATION,
-+ ("PFC mapping is available for Tx ports only"));
-+
-+ if (prio > 7)
-+ RETURN_ERROR(MAJOR, E_NOT_IN_RANGE,
-+ ("PFC priority (%d) is out of range (0-7)", prio));
-+ if (wq > 7)
-+ RETURN_ERROR(MAJOR, E_NOT_IN_RANGE,
-+ ("WQ (%d) is out of range (0-7)", wq));
-+
-+ tmpReg = GET_UINT32(p_FmPort->p_FmPortBmiRegs->txPortBmiRegs.fmbm_tpfcm[0]);
-+ tmpReg &= ~(0xf << ((7 - prio) * 4));
-+ wqTmpReg = ((uint32_t)wq << ((7 - prio) * 4));
-+ tmpReg |= wqTmpReg;
-+
-+ WRITE_UINT32(p_FmPort->p_FmPortBmiRegs->txPortBmiRegs.fmbm_tpfcm[0],
-+ tmpReg);
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PORT_SetFrameQueueCounters(t_Handle h_FmPort, bool enable)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmPort->p_FmPortDriverParam, E_INVALID_STATE);
-+
-+ fman_port_set_queue_cnt_mode(&p_FmPort->port, enable);
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PORT_SetPerformanceCounters(t_Handle h_FmPort, bool enable)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+ int err;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmPort->p_FmPortDriverParam, E_INVALID_STATE);
-+
-+ err = fman_port_set_perf_cnt_mode(&p_FmPort->port, enable);
-+ if (err != 0)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("fman_port_set_perf_cnt_mode"));
-+ return E_OK;
-+}
-+
-+t_Error FM_PORT_SetPerformanceCountersParams(
-+ t_Handle h_FmPort, t_FmPortPerformanceCnt *p_FmPortPerformanceCnt)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+ struct fman_port_perf_cnt_params params;
-+ int err;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+
-+ /* check parameters */
-+ if (!p_FmPortPerformanceCnt->taskCompVal
-+ || (p_FmPortPerformanceCnt->taskCompVal > p_FmPort->tasks.num))
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("taskCompVal (%d) has to be in the range of 1 - %d (current value)!", p_FmPortPerformanceCnt->taskCompVal, p_FmPort->tasks.num));
-+ if (!p_FmPortPerformanceCnt->dmaCompVal
-+ || (p_FmPortPerformanceCnt->dmaCompVal > p_FmPort->openDmas.num))
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("dmaCompVal (%d) has to be in the range of 1 - %d (current value)!", p_FmPortPerformanceCnt->dmaCompVal, p_FmPort->openDmas.num));
-+ if (!p_FmPortPerformanceCnt->fifoCompVal
-+ || (p_FmPortPerformanceCnt->fifoCompVal > p_FmPort->fifoBufs.num))
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("fifoCompVal (%d) has to be in the range of 256 - %d (current value)!", p_FmPortPerformanceCnt->fifoCompVal, p_FmPort->fifoBufs.num));
-+ if (p_FmPortPerformanceCnt->fifoCompVal % BMI_FIFO_UNITS)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("fifoCompVal (%d) has to be divisible by %d", p_FmPortPerformanceCnt->fifoCompVal, BMI_FIFO_UNITS));
-+
-+ switch (p_FmPort->portType)
-+ {
-+ case (e_FM_PORT_TYPE_RX_10G):
-+ case (e_FM_PORT_TYPE_RX):
-+ if (!p_FmPortPerformanceCnt->queueCompVal
-+ || (p_FmPortPerformanceCnt->queueCompVal
-+ > MAX_PERFORMANCE_RX_QUEUE_COMP))
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("performanceCnt.queueCompVal for Rx has to be in the range of 1 - %d", MAX_PERFORMANCE_RX_QUEUE_COMP));
-+ break;
-+ case (e_FM_PORT_TYPE_TX_10G):
-+ case (e_FM_PORT_TYPE_TX):
-+ if (!p_FmPortPerformanceCnt->queueCompVal
-+ || (p_FmPortPerformanceCnt->queueCompVal
-+ > MAX_PERFORMANCE_TX_QUEUE_COMP))
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("performanceCnt.queueCompVal for Tx has to be in the range of 1 - %d", MAX_PERFORMANCE_TX_QUEUE_COMP));
-+ break;
-+ case (e_FM_PORT_TYPE_OH_OFFLINE_PARSING):
-+ case (e_FM_PORT_TYPE_OH_HOST_COMMAND):
-+ if (p_FmPortPerformanceCnt->queueCompVal)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("performanceCnt.queueCompVal is not relevant for H/O ports."));
-+ break;
-+ default:
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("Invalid port type"));
-+ }
-+
-+ params.task_val = p_FmPortPerformanceCnt->taskCompVal;
-+ params.queue_val = p_FmPortPerformanceCnt->queueCompVal;
-+ params.dma_val = p_FmPortPerformanceCnt->dmaCompVal;
-+ params.fifo_val = p_FmPortPerformanceCnt->fifoCompVal;
-+
-+ err = fman_port_set_perf_cnt_params(&p_FmPort->port, &params);
-+ if (err != 0)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("fman_port_set_perf_cnt_params"));
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PORT_AnalyzePerformanceParams(t_Handle h_FmPort)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+ t_FmPortPerformanceCnt currParams, savedParams;
-+ t_Error err;
-+ bool underTest, failed = FALSE;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+
-+ XX_Print("Analyzing Performance parameters for port (type %d, id%d)\n",
-+ p_FmPort->portType, p_FmPort->portId);
-+
-+ currParams.taskCompVal = (uint8_t)p_FmPort->tasks.num;
-+ if ((p_FmPort->portType == e_FM_PORT_TYPE_OH_OFFLINE_PARSING)
-+ || (p_FmPort->portType == e_FM_PORT_TYPE_OH_HOST_COMMAND))
-+ currParams.queueCompVal = 0;
-+ else
-+ currParams.queueCompVal = 1;
-+ currParams.dmaCompVal = (uint8_t)p_FmPort->openDmas.num;
-+ currParams.fifoCompVal = p_FmPort->fifoBufs.num;
-+
-+ FM_PORT_SetPerformanceCounters(p_FmPort, FALSE);
-+ ClearPerfCnts(p_FmPort);
-+ if ((err = FM_PORT_SetPerformanceCountersParams(p_FmPort, &currParams))
-+ != E_OK)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ FM_PORT_SetPerformanceCounters(p_FmPort, TRUE);
-+ XX_UDelay(1000000);
-+ FM_PORT_SetPerformanceCounters(p_FmPort, FALSE);
-+ if (FM_PORT_GetCounter(p_FmPort, e_FM_PORT_COUNTERS_TASK_UTIL))
-+ {
-+ XX_Print(
-+ "Max num of defined port tasks (%d) utilized - Please enlarge\n",
-+ p_FmPort->tasks.num);
-+ failed = TRUE;
-+ }
-+ if (FM_PORT_GetCounter(p_FmPort, e_FM_PORT_COUNTERS_DMA_UTIL))
-+ {
-+ XX_Print(
-+ "Max num of defined port openDmas (%d) utilized - Please enlarge\n",
-+ p_FmPort->openDmas.num);
-+ failed = TRUE;
-+ }
-+ if (FM_PORT_GetCounter(p_FmPort, e_FM_PORT_COUNTERS_FIFO_UTIL))
-+ {
-+ XX_Print(
-+ "Max size of defined port fifo (%d) utilized - Please enlarge\n",
-+ p_FmPort->fifoBufs.num);
-+ failed = TRUE;
-+ }
-+ if (failed)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, NO_MSG);
-+
-+ memset(&savedParams, 0, sizeof(savedParams));
-+ while (TRUE)
-+ {
-+ underTest = FALSE;
-+ if ((currParams.taskCompVal != 1) && !savedParams.taskCompVal)
-+ {
-+ currParams.taskCompVal--;
-+ underTest = TRUE;
-+ }
-+ if ((currParams.dmaCompVal != 1) && !savedParams.dmaCompVal)
-+ {
-+ currParams.dmaCompVal--;
-+ underTest = TRUE;
-+ }
-+ if ((currParams.fifoCompVal != BMI_FIFO_UNITS)
-+ && !savedParams.fifoCompVal)
-+ {
-+ currParams.fifoCompVal -= BMI_FIFO_UNITS;
-+ underTest = TRUE;
-+ }
-+ if (!underTest)
-+ break;
-+
-+ ClearPerfCnts(p_FmPort);
-+ if ((err = FM_PORT_SetPerformanceCountersParams(p_FmPort, &currParams))
-+ != E_OK)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ FM_PORT_SetPerformanceCounters(p_FmPort, TRUE);
-+ XX_UDelay(1000000);
-+ FM_PORT_SetPerformanceCounters(p_FmPort, FALSE);
-+
-+ if (!savedParams.taskCompVal
-+ && FM_PORT_GetCounter(p_FmPort, e_FM_PORT_COUNTERS_TASK_UTIL))
-+ savedParams.taskCompVal = (uint8_t)(currParams.taskCompVal + 2);
-+ if (!savedParams.dmaCompVal
-+ && FM_PORT_GetCounter(p_FmPort, e_FM_PORT_COUNTERS_DMA_UTIL))
-+ savedParams.dmaCompVal = (uint8_t)(currParams.dmaCompVal + 2);
-+ if (!savedParams.fifoCompVal
-+ && FM_PORT_GetCounter(p_FmPort, e_FM_PORT_COUNTERS_FIFO_UTIL))
-+ savedParams.fifoCompVal = currParams.fifoCompVal
-+ + (2 * BMI_FIFO_UNITS);
-+ }
-+
-+ XX_Print("best vals: tasks %d, dmas %d, fifos %d\n",
-+ savedParams.taskCompVal, savedParams.dmaCompVal,
-+ savedParams.fifoCompVal);
-+ return E_OK;
-+}
-+
-+t_Error FM_PORT_SetStatisticsCounters(t_Handle h_FmPort, bool enable)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+ int err;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmPort->p_FmPortDriverParam, E_INVALID_STATE);
-+
-+ err = fman_port_set_stats_cnt_mode(&p_FmPort->port, enable);
-+ if (err != 0)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("fman_port_set_stats_cnt_mode"));
-+ return E_OK;
-+}
-+
-+t_Error FM_PORT_SetErrorsRoute(t_Handle h_FmPort, fmPortFrameErrSelect_t errs)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+ volatile uint32_t *p_ErrDiscard = NULL;
-+ int err;
-+
-+ UNUSED(p_ErrDiscard);
-+ err = fman_port_set_err_mask(&p_FmPort->port, (uint32_t)errs);
-+ if (err != 0)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("fman_port_set_err_mask"));
-+
-+#ifdef FM_ERROR_VSP_NO_MATCH_SW006
-+ if (p_FmPort->fmRevInfo.majorRev >= 6)
-+ {
-+ t_FmPcdCtrlParamsPage *p_ParamsPage;
-+
-+ FmPortSetGprFunc(p_FmPort, e_FM_PORT_GPR_MURAM_PAGE,
-+ (void**)&p_ParamsPage);
-+ ASSERT_COND(p_ParamsPage);
-+ switch (p_FmPort->portType)
-+ {
-+ case (e_FM_PORT_TYPE_RX_10G):
-+ case (e_FM_PORT_TYPE_RX):
-+ p_ErrDiscard =
-+ &p_FmPort->p_FmPortBmiRegs->rxPortBmiRegs.fmbm_rfsdm;
-+ break;
-+ case (e_FM_PORT_TYPE_OH_OFFLINE_PARSING):
-+ p_ErrDiscard =
-+ &p_FmPort->p_FmPortBmiRegs->ohPortBmiRegs.fmbm_ofsdm;
-+ break;
-+ default:
-+ RETURN_ERROR(
-+ MAJOR, E_INVALID_OPERATION,
-+ ("available for Rx and offline parsing ports only"));
-+ }
-+ WRITE_UINT32(p_ParamsPage->errorsDiscardMask,
-+ GET_UINT32(*p_ErrDiscard) | errs);
-+ }
-+#endif /* FM_ERROR_VSP_NO_MATCH_SW006 */
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PORT_SetAllocBufCounter(t_Handle h_FmPort, uint8_t poolId,
-+ bool enable)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+ int err;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(poolId<BM_MAX_NUM_OF_POOLS, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmPort->p_FmPortDriverParam, E_INVALID_STATE);
-+
-+ if ((p_FmPort->portType != e_FM_PORT_TYPE_RX_10G)
-+ && (p_FmPort->portType != e_FM_PORT_TYPE_RX))
-+ RETURN_ERROR(MAJOR, E_INVALID_OPERATION,
-+ ("available for Rx ports only"));
-+
-+ err = fman_port_set_bpool_cnt_mode(&p_FmPort->port, poolId, enable);
-+ if (err != 0)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("fman_port_set_bpool_cnt_mode"));
-+ return E_OK;
-+}
-+
-+t_Error FM_PORT_GetBmiCounters(t_Handle h_FmPort, t_FmPortBmiStats *p_BmiStats)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+
-+ if ((p_FmPort->portType == e_FM_PORT_TYPE_RX)
-+ || (p_FmPort->portType == e_FM_PORT_TYPE_RX_10G)){
-+ p_BmiStats->cntCycle =
-+ FM_PORT_GetCounter(h_FmPort, e_FM_PORT_COUNTERS_CYCLE);
-+ /* fmbm_rccn */
-+ p_BmiStats->cntTaskUtil =
-+ FM_PORT_GetCounter(h_FmPort, e_FM_PORT_COUNTERS_TASK_UTIL);
-+ /* fmbm_rtuc */
-+ p_BmiStats->cntQueueUtil =
-+ FM_PORT_GetCounter(h_FmPort, e_FM_PORT_COUNTERS_QUEUE_UTIL);
-+ /* fmbm_rrquc */
-+ p_BmiStats->cntDmaUtil =
-+ FM_PORT_GetCounter(h_FmPort, e_FM_PORT_COUNTERS_DMA_UTIL);
-+ /* fmbm_rduc */
-+ p_BmiStats->cntFifoUtil =
-+ FM_PORT_GetCounter(h_FmPort, e_FM_PORT_COUNTERS_FIFO_UTIL);
-+ /* fmbm_rfuc */
-+ p_BmiStats->cntRxPauseActivation =
-+ FM_PORT_GetCounter(h_FmPort, e_FM_PORT_COUNTERS_RX_PAUSE_ACTIVATION);
-+ /* fmbm_rpac */
-+ p_BmiStats->cntFrame =
-+ FM_PORT_GetCounter(h_FmPort, e_FM_PORT_COUNTERS_FRAME);
-+ /* fmbm_rfrc */
-+ p_BmiStats->cntDiscardFrame =
-+ FM_PORT_GetCounter(h_FmPort, e_FM_PORT_COUNTERS_DISCARD_FRAME);
-+ /* fmbm_rfdc */
-+ p_BmiStats->cntDeallocBuf =
-+ FM_PORT_GetCounter(h_FmPort, e_FM_PORT_COUNTERS_DEALLOC_BUF);
-+ /* fmbm_rbdc */
-+ p_BmiStats->cntRxBadFrame =
-+ FM_PORT_GetCounter(h_FmPort, e_FM_PORT_COUNTERS_RX_BAD_FRAME);
-+ /* fmbm_rfbc */
-+ p_BmiStats->cntRxLargeFrame =
-+ FM_PORT_GetCounter(h_FmPort, e_FM_PORT_COUNTERS_RX_LARGE_FRAME);
-+ /* fmbm_rlfc */
-+ p_BmiStats->cntRxFilterFrame =
-+ FM_PORT_GetCounter(h_FmPort, e_FM_PORT_COUNTERS_RX_FILTER_FRAME);
-+ /* fmbm_rffc */
-+ p_BmiStats->cntRxListDmaErr =
-+ FM_PORT_GetCounter(h_FmPort, e_FM_PORT_COUNTERS_RX_LIST_DMA_ERR);
-+ /* fmbm_rfldec */
-+ p_BmiStats->cntRxOutOfBuffersDiscard =
-+ FM_PORT_GetCounter(h_FmPort, e_FM_PORT_COUNTERS_RX_OUT_OF_BUFFERS_DISCARD);
-+ /* fmbm_rodc */
-+ p_BmiStats->cntWredDiscard = 0;
-+ p_BmiStats->cntLengthErr = 0;
-+ p_BmiStats->cntUnsupportedFormat = 0;
-+ }
-+ else if ((p_FmPort->portType == e_FM_PORT_TYPE_TX)
-+ || (p_FmPort->portType == e_FM_PORT_TYPE_TX_10G)){
-+ p_BmiStats->cntCycle =
-+ FM_PORT_GetCounter(h_FmPort, e_FM_PORT_COUNTERS_CYCLE);
-+ /* fmbm_tccn */
-+ p_BmiStats->cntTaskUtil =
-+ FM_PORT_GetCounter(h_FmPort, e_FM_PORT_COUNTERS_TASK_UTIL);
-+ /* fmbm_ttuc */
-+ p_BmiStats->cntQueueUtil =
-+ FM_PORT_GetCounter(h_FmPort, e_FM_PORT_COUNTERS_QUEUE_UTIL);
-+ /* fmbm_ttcquc */
-+ p_BmiStats->cntDmaUtil =
-+ FM_PORT_GetCounter(h_FmPort, e_FM_PORT_COUNTERS_DMA_UTIL);
-+ /* fmbm_tduc */
-+ p_BmiStats->cntFifoUtil =
-+ FM_PORT_GetCounter(h_FmPort, e_FM_PORT_COUNTERS_FIFO_UTIL);
-+ /* fmbm_tfuc */
-+ p_BmiStats->cntRxPauseActivation = 0;
-+ p_BmiStats->cntFrame =
-+ FM_PORT_GetCounter(h_FmPort, e_FM_PORT_COUNTERS_FRAME);
-+ /* fmbm_tfrc */
-+ p_BmiStats->cntDiscardFrame =
-+ FM_PORT_GetCounter(h_FmPort, e_FM_PORT_COUNTERS_DISCARD_FRAME);
-+ /* fmbm_tfdc */
-+ p_BmiStats->cntDeallocBuf =
-+ FM_PORT_GetCounter(h_FmPort, e_FM_PORT_COUNTERS_DEALLOC_BUF);
-+ /* fmbm_tbdc */
-+ p_BmiStats->cntRxBadFrame = 0;
-+ p_BmiStats->cntRxLargeFrame = 0;
-+ p_BmiStats->cntRxFilterFrame = 0;
-+ p_BmiStats->cntRxListDmaErr = 0;
-+ p_BmiStats->cntRxOutOfBuffersDiscard = 0;
-+ p_BmiStats->cntWredDiscard = 0;
-+ p_BmiStats->cntLengthErr =
-+ FM_PORT_GetCounter(h_FmPort, e_FM_PORT_COUNTERS_LENGTH_ERR);
-+ /* fmbm_tfledc */
-+ p_BmiStats->cntUnsupportedFormat =
-+ FM_PORT_GetCounter(h_FmPort, e_FM_PORT_COUNTERS_UNSUPPRTED_FORMAT);
-+ /* fmbm_tfufdc */
-+ }
-+ else if (p_FmPort->portType == e_FM_PORT_TYPE_OH_OFFLINE_PARSING) {
-+ p_BmiStats->cntCycle =
-+ FM_PORT_GetCounter(h_FmPort, e_FM_PORT_COUNTERS_CYCLE);
-+ /* fmbm_occn */
-+ p_BmiStats->cntTaskUtil =
-+ FM_PORT_GetCounter(h_FmPort, e_FM_PORT_COUNTERS_TASK_UTIL);
-+ /* fmbm_otuc */
-+ p_BmiStats->cntQueueUtil = 0;
-+ p_BmiStats->cntDmaUtil =
-+ FM_PORT_GetCounter(h_FmPort, e_FM_PORT_COUNTERS_DMA_UTIL);
-+ /* fmbm_oduc */
-+ p_BmiStats->cntFifoUtil =
-+ FM_PORT_GetCounter(h_FmPort, e_FM_PORT_COUNTERS_FIFO_UTIL);
-+ /* fmbm_ofuc*/
-+ p_BmiStats->cntRxPauseActivation = 0;
-+ p_BmiStats->cntFrame =
-+ FM_PORT_GetCounter(h_FmPort, e_FM_PORT_COUNTERS_FRAME);
-+ /* fmbm_ofrc */
-+ p_BmiStats->cntDiscardFrame =
-+ FM_PORT_GetCounter(h_FmPort, e_FM_PORT_COUNTERS_DISCARD_FRAME);
-+ /* fmbm_ofdc */
-+ p_BmiStats->cntDeallocBuf =
-+ FM_PORT_GetCounter(h_FmPort, e_FM_PORT_COUNTERS_DEALLOC_BUF);
-+ /* fmbm_obdc*/
-+ p_BmiStats->cntRxBadFrame = 0;
-+ p_BmiStats->cntRxLargeFrame = 0;
-+ p_BmiStats->cntRxFilterFrame =
-+ FM_PORT_GetCounter(h_FmPort, e_FM_PORT_COUNTERS_RX_FILTER_FRAME);
-+ /* fmbm_offc */
-+ p_BmiStats->cntRxListDmaErr =
-+ FM_PORT_GetCounter(h_FmPort, e_FM_PORT_COUNTERS_RX_LIST_DMA_ERR);
-+ /* fmbm_ofldec */
-+ p_BmiStats->cntRxOutOfBuffersDiscard =
-+ FM_PORT_GetCounter(h_FmPort, e_FM_PORT_COUNTERS_RX_OUT_OF_BUFFERS_DISCARD);
-+ /* fmbm_rodc */
-+ p_BmiStats->cntWredDiscard =
-+ FM_PORT_GetCounter(h_FmPort, e_FM_PORT_COUNTERS_WRED_DISCARD);
-+ /* fmbm_ofwdc */
-+ p_BmiStats->cntLengthErr =
-+ FM_PORT_GetCounter(h_FmPort, e_FM_PORT_COUNTERS_LENGTH_ERR);
-+ /* fmbm_ofledc */
-+ p_BmiStats->cntUnsupportedFormat =
-+ FM_PORT_GetCounter(h_FmPort, e_FM_PORT_COUNTERS_UNSUPPRTED_FORMAT);
-+ /* fmbm_ofufdc */
-+ }
-+ return E_OK;
-+}
-+
-+uint32_t FM_PORT_GetCounter(t_Handle h_FmPort, e_FmPortCounters counter)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+ bool bmiCounter = FALSE;
-+ enum fman_port_stats_counters statsType;
-+ enum fman_port_perf_counters perfType;
-+ enum fman_port_qmi_counters queueType;
-+ bool isStats;
-+ t_Error errCode;
-+
-+ SANITY_CHECK_RETURN_VALUE(p_FmPort, E_INVALID_HANDLE, 0);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmPort->p_FmPortDriverParam, E_INVALID_STATE);
-+
-+ switch (counter)
-+ {
-+ case (e_FM_PORT_COUNTERS_DEQ_TOTAL):
-+ case (e_FM_PORT_COUNTERS_DEQ_FROM_DEFAULT):
-+ case (e_FM_PORT_COUNTERS_DEQ_CONFIRM):
-+ /* check that counter is available for the port type */
-+ if ((p_FmPort->portType == e_FM_PORT_TYPE_RX)
-+ || (p_FmPort->portType == e_FM_PORT_TYPE_RX_10G))
-+ {
-+ REPORT_ERROR(MINOR, E_INVALID_STATE,
-+ ("Requested counter is not available for Rx ports"));
-+ return 0;
-+ }
-+ bmiCounter = FALSE;
-+ break;
-+ case (e_FM_PORT_COUNTERS_ENQ_TOTAL):
-+ bmiCounter = FALSE;
-+ break;
-+ default: /* BMI counters (or error - will be checked in BMI routine )*/
-+ bmiCounter = TRUE;
-+ break;
-+ }
-+
-+ if (bmiCounter)
-+ {
-+ errCode = BmiPortCheckAndGetCounterType(p_FmPort, counter, &statsType,
-+ &perfType, &isStats);
-+ if (errCode != E_OK)
-+ {
-+ REPORT_ERROR(MINOR, errCode, NO_MSG);
-+ return 0;
-+ }
-+ if (isStats)
-+ return fman_port_get_stats_counter(&p_FmPort->port, statsType);
-+ else
-+ return fman_port_get_perf_counter(&p_FmPort->port, perfType);
-+ }
-+ else /* QMI counter */
-+ {
-+ /* check that counters are enabled */
-+ if (!(GET_UINT32(p_FmPort->port.qmi_regs->fmqm_pnc)
-+ & QMI_PORT_CFG_EN_COUNTERS))
-+
-+ {
-+ REPORT_ERROR(MINOR, E_INVALID_STATE, ("Requested counter was not enabled"));
-+ return 0;
-+ }
-+
-+ /* Set counter */
-+ switch (counter)
-+ {
-+ case (e_FM_PORT_COUNTERS_ENQ_TOTAL):
-+ queueType = E_FMAN_PORT_ENQ_TOTAL;
-+ break;
-+ case (e_FM_PORT_COUNTERS_DEQ_TOTAL):
-+ queueType = E_FMAN_PORT_DEQ_TOTAL;
-+ break;
-+ case (e_FM_PORT_COUNTERS_DEQ_FROM_DEFAULT):
-+ queueType = E_FMAN_PORT_DEQ_FROM_DFLT;
-+ break;
-+ case (e_FM_PORT_COUNTERS_DEQ_CONFIRM):
-+ queueType = E_FMAN_PORT_DEQ_CONFIRM;
-+ break;
-+ default:
-+ REPORT_ERROR(MINOR, E_INVALID_STATE, ("Requested counter is not available"));
-+ return 0;
-+ }
-+
-+ return fman_port_get_qmi_counter(&p_FmPort->port, queueType);
-+ }
-+
-+ return 0;
-+}
-+
-+t_Error FM_PORT_ModifyCounter(t_Handle h_FmPort, e_FmPortCounters counter,
-+ uint32_t value)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+ bool bmiCounter = FALSE;
-+ enum fman_port_stats_counters statsType;
-+ enum fman_port_perf_counters perfType;
-+ enum fman_port_qmi_counters queueType;
-+ bool isStats;
-+ t_Error errCode;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmPort->p_FmPortDriverParam, E_INVALID_STATE);
-+
-+ switch (counter)
-+ {
-+ case (e_FM_PORT_COUNTERS_DEQ_TOTAL):
-+ case (e_FM_PORT_COUNTERS_DEQ_FROM_DEFAULT):
-+ case (e_FM_PORT_COUNTERS_DEQ_CONFIRM):
-+ /* check that counter is available for the port type */
-+ if ((p_FmPort->portType == e_FM_PORT_TYPE_RX)
-+ || (p_FmPort->portType == e_FM_PORT_TYPE_RX_10G))
-+ RETURN_ERROR(
-+ MINOR, E_INVALID_STATE,
-+ ("Requested counter is not available for Rx ports"));
-+ case (e_FM_PORT_COUNTERS_ENQ_TOTAL):
-+ bmiCounter = FALSE;
-+ break;
-+ default: /* BMI counters (or error - will be checked in BMI routine )*/
-+ bmiCounter = TRUE;
-+ break;
-+ }
-+
-+ if (bmiCounter)
-+ {
-+ errCode = BmiPortCheckAndGetCounterType(p_FmPort, counter, &statsType,
-+ &perfType, &isStats);
-+ if (errCode != E_OK)
-+ {
-+ RETURN_ERROR(MINOR, errCode, NO_MSG);
-+ }
-+ if (isStats)
-+ fman_port_set_stats_counter(&p_FmPort->port, statsType, value);
-+ else
-+ fman_port_set_perf_counter(&p_FmPort->port, perfType, value);
-+ }
-+ else /* QMI counter */
-+ {
-+ /* check that counters are enabled */
-+ if (!(GET_UINT32(p_FmPort->port.qmi_regs->fmqm_pnc)
-+ & QMI_PORT_CFG_EN_COUNTERS))
-+ {
-+ RETURN_ERROR(MINOR, E_INVALID_STATE,
-+ ("Requested counter was not enabled"));
-+ }
-+
-+ /* Set counter */
-+ switch (counter)
-+ {
-+ case (e_FM_PORT_COUNTERS_ENQ_TOTAL):
-+ queueType = E_FMAN_PORT_ENQ_TOTAL;
-+ break;
-+ case (e_FM_PORT_COUNTERS_DEQ_TOTAL):
-+ queueType = E_FMAN_PORT_DEQ_TOTAL;
-+ break;
-+ case (e_FM_PORT_COUNTERS_DEQ_FROM_DEFAULT):
-+ queueType = E_FMAN_PORT_DEQ_FROM_DFLT;
-+ break;
-+ case (e_FM_PORT_COUNTERS_DEQ_CONFIRM):
-+ queueType = E_FMAN_PORT_DEQ_CONFIRM;
-+ break;
-+ default:
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE,
-+ ("Requested counter is not available"));
-+ }
-+
-+ fman_port_set_qmi_counter(&p_FmPort->port, queueType, value);
-+ }
-+
-+ return E_OK;
-+}
-+
-+uint32_t FM_PORT_GetAllocBufCounter(t_Handle h_FmPort, uint8_t poolId)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+
-+ SANITY_CHECK_RETURN_VALUE(p_FmPort, E_INVALID_HANDLE, 0);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmPort->p_FmPortDriverParam, E_INVALID_STATE);
-+
-+ if ((p_FmPort->portType != e_FM_PORT_TYPE_RX)
-+ && (p_FmPort->portType == e_FM_PORT_TYPE_RX_10G))
-+ {
-+ REPORT_ERROR(MINOR, E_INVALID_STATE, ("Requested counter is not available for non-Rx ports"));
-+ return 0;
-+ }
-+ return fman_port_get_bpool_counter(&p_FmPort->port, poolId);
-+}
-+
-+t_Error FM_PORT_ModifyAllocBufCounter(t_Handle h_FmPort, uint8_t poolId,
-+ uint32_t value)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort *)h_FmPort;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmPort->p_FmPortDriverParam, E_INVALID_STATE);
-+
-+ if ((p_FmPort->portType != e_FM_PORT_TYPE_RX)
-+ && (p_FmPort->portType == e_FM_PORT_TYPE_RX_10G))
-+ RETURN_ERROR( MINOR, E_INVALID_STATE,
-+ ("Requested counter is not available for non-Rx ports"));
-+
-+ fman_port_set_bpool_counter(&p_FmPort->port, poolId, value);
-+ return E_OK;
-+}
-+bool FM_PORT_IsStalled(t_Handle h_FmPort)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+ t_Error err;
-+ bool isStalled;
-+
-+ SANITY_CHECK_RETURN_VALUE(p_FmPort, E_INVALID_HANDLE, FALSE);
-+ SANITY_CHECK_RETURN_VALUE(!p_FmPort->p_FmPortDriverParam, E_INVALID_STATE,
-+ FALSE);
-+
-+ err = FmIsPortStalled(p_FmPort->h_Fm, p_FmPort->hardwarePortId, &isStalled);
-+ if (err != E_OK)
-+ {
-+ REPORT_ERROR(MAJOR, err, NO_MSG);
-+ return TRUE;
-+ }
-+ return isStalled;
-+}
-+
-+t_Error FM_PORT_ReleaseStalled(t_Handle h_FmPort)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmPort->p_FmPortDriverParam, E_INVALID_STATE);
-+
-+ return FmResumeStalledPort(p_FmPort->h_Fm, p_FmPort->hardwarePortId);
-+}
-+
-+t_Error FM_PORT_SetRxL4ChecksumVerify(t_Handle h_FmPort, bool l4Checksum)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+ int err;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmPort->p_FmPortDriverParam, E_INVALID_STATE);
-+
-+ if ((p_FmPort->portType != e_FM_PORT_TYPE_RX_10G)
-+ && (p_FmPort->portType != e_FM_PORT_TYPE_RX))
-+ RETURN_ERROR(MAJOR, E_INVALID_OPERATION,
-+ ("available for Rx ports only"));
-+
-+ if (l4Checksum)
-+ err = fman_port_modify_rx_fd_bits(
-+ &p_FmPort->port, (uint8_t)(BMI_PORT_RFNE_FRWD_DCL4C >> 24),
-+ TRUE);
-+ else
-+ err = fman_port_modify_rx_fd_bits(
-+ &p_FmPort->port, (uint8_t)(BMI_PORT_RFNE_FRWD_DCL4C >> 24),
-+ FALSE);
-+ if (err != 0)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("fman_port_modify_rx_fd_bits"));
-+
-+ return E_OK;
-+}
-+
-+/*****************************************************************************/
-+/* API Run-time PCD Control unit functions */
-+/*****************************************************************************/
-+
-+#if (DPAA_VERSION >= 11)
-+t_Error FM_PORT_VSPAlloc(t_Handle h_FmPort, t_FmPortVSPAllocParams *p_VSPParams)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+ t_Error err = E_OK;
-+ volatile uint32_t *p_BmiStorageProfileId = NULL, *p_BmiVspe = NULL;
-+ uint32_t tmpReg = 0, tmp = 0;
-+ uint16_t hwStoragePrflId;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort->h_Fm, E_INVALID_HANDLE);
-+ /*for numOfProfiles = 0 don't call this function*/
-+ SANITY_CHECK_RETURN_ERROR(p_VSPParams->numOfProfiles, E_INVALID_VALUE);
-+ /*dfltRelativeId should be in the range of numOfProfiles*/
-+ SANITY_CHECK_RETURN_ERROR(
-+ p_VSPParams->dfltRelativeId < p_VSPParams->numOfProfiles,
-+ E_INVALID_VALUE);
-+ /*p_FmPort should be from Rx type or OP*/
-+ SANITY_CHECK_RETURN_ERROR(
-+ ((p_FmPort->portType == e_FM_PORT_TYPE_RX_10G) || (p_FmPort->portType == e_FM_PORT_TYPE_RX) || (p_FmPort->portType == e_FM_PORT_TYPE_OH_OFFLINE_PARSING)),
-+ E_INVALID_VALUE);
-+ /*port should be disabled*/
-+ SANITY_CHECK_RETURN_ERROR(!p_FmPort->enabled, E_INVALID_STATE);
-+ /*if its called for Rx port relevant Tx Port should be passed (initialized) too and it should be disabled*/
-+ SANITY_CHECK_RETURN_ERROR(
-+ ((p_VSPParams->h_FmTxPort && !((t_FmPort *)(p_VSPParams->h_FmTxPort))->enabled) || (p_FmPort->portType == e_FM_PORT_TYPE_OH_OFFLINE_PARSING)),
-+ E_INVALID_VALUE);
-+ /*should be called before SetPCD - this port should be without PCD*/
-+ SANITY_CHECK_RETURN_ERROR(!p_FmPort->pcdEngines, E_INVALID_STATE);
-+
-+ /*alloc window of VSPs for this port*/
-+ err = FmVSPAllocForPort(p_FmPort->h_Fm, p_FmPort->portType,
-+ p_FmPort->portId, p_VSPParams->numOfProfiles);
-+ if (err != E_OK)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+
-+ /*get absolute VSP ID for dfltRelative*/
-+ err = FmVSPGetAbsoluteProfileId(p_FmPort->h_Fm, p_FmPort->portType,
-+ p_FmPort->portId,
-+ p_VSPParams->dfltRelativeId,
-+ &hwStoragePrflId);
-+ if (err != E_OK)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+
-+ /*fill relevant registers for p_FmPort and relative TxPort in the case p_FmPort from Rx type*/
-+ switch (p_FmPort->portType)
-+ {
-+ case (e_FM_PORT_TYPE_RX_10G):
-+ case (e_FM_PORT_TYPE_RX):
-+ p_BmiStorageProfileId =
-+ &(((t_FmPort *)(p_VSPParams->h_FmTxPort))->p_FmPortBmiRegs->txPortBmiRegs.fmbm_tcfqid);
-+ p_BmiVspe =
-+ &(((t_FmPort *)(p_VSPParams->h_FmTxPort))->p_FmPortBmiRegs->txPortBmiRegs.fmbm_tfne);
-+
-+ tmpReg = GET_UINT32(*p_BmiStorageProfileId) & ~BMI_SP_ID_MASK;
-+ tmpReg |= (uint32_t)hwStoragePrflId << BMI_SP_ID_SHIFT;
-+ WRITE_UINT32(*p_BmiStorageProfileId, tmpReg);
-+
-+ tmpReg = GET_UINT32(*p_BmiVspe);
-+ WRITE_UINT32(*p_BmiVspe, tmpReg | BMI_SP_EN);
-+
-+ p_BmiStorageProfileId =
-+ &p_FmPort->p_FmPortBmiRegs->rxPortBmiRegs.fmbm_rfqid;
-+ p_BmiVspe = &p_FmPort->p_FmPortBmiRegs->rxPortBmiRegs.fmbm_rpp;
-+ hwStoragePrflId = p_VSPParams->dfltRelativeId;
-+ break;
-+
-+ case (e_FM_PORT_TYPE_OH_OFFLINE_PARSING):
-+ tmpReg = NIA_ENG_BMI | NIA_BMI_AC_FETCH_ALL_FRAME;
-+ WRITE_UINT32( p_FmPort->p_FmPortQmiRegs->nonRxQmiRegs.fmqm_pndn,
-+ tmpReg);
-+
-+ p_BmiStorageProfileId =
-+ &p_FmPort->p_FmPortBmiRegs->ohPortBmiRegs.fmbm_ofqid;
-+ p_BmiVspe = &p_FmPort->p_FmPortBmiRegs->ohPortBmiRegs.fmbm_opp;
-+ tmp |= BMI_EBD_EN;
-+ break;
-+
-+ default:
-+ RETURN_ERROR( MAJOR, E_INVALID_OPERATION,
-+ ("available for Rx and offline parsing ports only"));
-+ }
-+
-+ p_FmPort->vspe = TRUE;
-+ p_FmPort->dfltRelativeId = p_VSPParams->dfltRelativeId;
-+
-+ tmpReg = GET_UINT32(*p_BmiStorageProfileId) & ~BMI_SP_ID_MASK;
-+ tmpReg |= (uint32_t)hwStoragePrflId << BMI_SP_ID_SHIFT;
-+ WRITE_UINT32(*p_BmiStorageProfileId, tmpReg);
-+
-+ tmpReg = GET_UINT32(*p_BmiVspe);
-+ WRITE_UINT32(*p_BmiVspe, tmpReg | BMI_SP_EN | tmp);
-+ return E_OK;
-+}
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+t_Error FM_PORT_PcdPlcrAllocProfiles(t_Handle h_FmPort, uint16_t numOfProfiles)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+ t_Error err = E_OK;
-+
-+ p_FmPort->h_FmPcd = FmGetPcdHandle(p_FmPort->h_Fm);
-+ ASSERT_COND(p_FmPort->h_FmPcd);
-+
-+ if (!TRY_LOCK(p_FmPort->h_Spinlock, &p_FmPort->lock))
-+ {
-+ DBG(TRACE, ("FM Port Try Lock - BUSY"));
-+ return ERROR_CODE(E_BUSY);
-+ }
-+
-+ if (numOfProfiles)
-+ {
-+ err = FmPcdPlcrAllocProfiles(p_FmPort->h_FmPcd,
-+ p_FmPort->hardwarePortId, numOfProfiles);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+ /* set the port handle within the PCD policer, even if no profiles defined */
-+ FmPcdPortRegister(p_FmPort->h_FmPcd, h_FmPort, p_FmPort->hardwarePortId);
-+
-+ RELEASE_LOCK(p_FmPort->lock);
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PORT_PcdPlcrFreeProfiles(t_Handle h_FmPort)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+ t_Error err = E_OK;
-+
-+ if (!TRY_LOCK(p_FmPort->h_Spinlock, &p_FmPort->lock))
-+ {
-+ DBG(TRACE, ("FM Port Try Lock - BUSY"));
-+ return ERROR_CODE(E_BUSY);
-+ }
-+
-+ err = FmPcdPlcrFreeProfiles(p_FmPort->h_FmPcd, p_FmPort->hardwarePortId);
-+
-+ RELEASE_LOCK(p_FmPort->lock);
-+
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PORT_PcdKgModifyInitialScheme(t_Handle h_FmPort,
-+ t_FmPcdKgSchemeSelect *p_FmPcdKgScheme)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+ volatile uint32_t *p_BmiHpnia = NULL;
-+ uint32_t tmpReg;
-+ uint8_t relativeSchemeId;
-+ uint8_t physicalSchemeId;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmPort->p_FmPortDriverParam, E_INVALID_STATE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort->pcdEngines & FM_PCD_KG,
-+ E_INVALID_STATE);
-+
-+ tmpReg = (uint32_t)((p_FmPort->pcdEngines & FM_PCD_CC) ? NIA_KG_CC_EN : 0);
-+ switch (p_FmPort->portType)
-+ {
-+ case (e_FM_PORT_TYPE_RX_10G):
-+ case (e_FM_PORT_TYPE_RX):
-+ p_BmiHpnia = &p_FmPort->port.bmi_regs->rx.fmbm_rfpne;
-+ break;
-+ case (e_FM_PORT_TYPE_OH_OFFLINE_PARSING):
-+ p_BmiHpnia = &p_FmPort->port.bmi_regs->oh.fmbm_ofpne;
-+ break;
-+ default:
-+ RETURN_ERROR( MAJOR, E_INVALID_OPERATION,
-+ ("available for Rx and offline parsing ports only"));
-+ }
-+
-+ if (!TRY_LOCK(p_FmPort->h_Spinlock, &p_FmPort->lock))
-+ {
-+ DBG(TRACE, ("FM Port Try Lock - BUSY"));
-+ return ERROR_CODE(E_BUSY);
-+ }
-+
-+ /* if we want to change to direct scheme, we need to check that this scheme is valid */
-+ if (p_FmPcdKgScheme->direct)
-+ {
-+ physicalSchemeId = FmPcdKgGetSchemeId(p_FmPcdKgScheme->h_DirectScheme);
-+ /* check that this scheme is bound to this port */
-+ if (!(p_FmPort->schemesPerPortVector
-+ & (uint32_t)(1 << (31 - (uint32_t)physicalSchemeId))))
-+ {
-+ RELEASE_LOCK(p_FmPort->lock);
-+ RETURN_ERROR(
-+ MAJOR, E_INVALID_STATE,
-+ ("called with a scheme that is not bound to this port"));
-+ }
-+
-+ relativeSchemeId = FmPcdKgGetRelativeSchemeId(p_FmPort->h_FmPcd,
-+ physicalSchemeId);
-+ if (relativeSchemeId >= FM_PCD_KG_NUM_OF_SCHEMES)
-+ {
-+ RELEASE_LOCK(p_FmPort->lock);
-+ RETURN_ERROR(MAJOR, E_NOT_IN_RANGE,
-+ ("called with invalid Scheme "));
-+ }
-+
-+ if (!FmPcdKgIsSchemeValidSw(p_FmPcdKgScheme->h_DirectScheme))
-+ {
-+ RELEASE_LOCK(p_FmPort->lock);
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE,
-+ ("called with uninitialized Scheme "));
-+ }
-+
-+ WRITE_UINT32(
-+ *p_BmiHpnia,
-+ NIA_ENG_KG | tmpReg | NIA_KG_DIRECT | (uint32_t)physicalSchemeId);
-+ }
-+ else
-+ /* change to indirect scheme */
-+ WRITE_UINT32(*p_BmiHpnia, NIA_ENG_KG | tmpReg);
-+ RELEASE_LOCK(p_FmPort->lock);
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PORT_PcdPlcrModifyInitialProfile(t_Handle h_FmPort,
-+ t_Handle h_Profile)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+ volatile uint32_t *p_BmiNia;
-+ volatile uint32_t *p_BmiHpnia;
-+ uint32_t tmpReg;
-+ uint16_t absoluteProfileId = FmPcdPlcrProfileGetAbsoluteId(h_Profile);
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmPort->p_FmPortDriverParam, E_INVALID_STATE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort->pcdEngines & FM_PCD_PLCR,
-+ E_INVALID_STATE);
-+
-+ /* check relevance of this routine - only when policer is used
-+ directly after BMI or Parser */
-+ if ((p_FmPort->pcdEngines & FM_PCD_KG)
-+ || (p_FmPort->pcdEngines & FM_PCD_CC))
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_STATE,
-+ ("relevant only when PCD support mode is e_FM_PCD_SUPPORT_PLCR_ONLY or e_FM_PCD_SUPPORT_PRS_AND_PLCR"));
-+
-+ switch (p_FmPort->portType)
-+ {
-+ case (e_FM_PORT_TYPE_RX_10G):
-+ case (e_FM_PORT_TYPE_RX):
-+ p_BmiNia = &p_FmPort->port.bmi_regs->rx.fmbm_rfne;
-+ p_BmiHpnia = &p_FmPort->port.bmi_regs->rx.fmbm_rfpne;
-+ tmpReg = GET_UINT32(*p_BmiNia) & BMI_RFNE_FDCS_MASK;
-+ break;
-+ case (e_FM_PORT_TYPE_OH_OFFLINE_PARSING):
-+ p_BmiNia = &p_FmPort->port.bmi_regs->oh.fmbm_ofne;
-+ p_BmiHpnia = &p_FmPort->port.bmi_regs->oh.fmbm_ofpne;
-+ tmpReg = 0;
-+ break;
-+ default:
-+ RETURN_ERROR( MAJOR, E_INVALID_OPERATION,
-+ ("available for Rx and offline parsing ports only"));
-+ }
-+
-+ if (!TRY_LOCK(p_FmPort->h_Spinlock, &p_FmPort->lock))
-+ {
-+ DBG(TRACE, ("FM Port Try Lock - BUSY"));
-+ return ERROR_CODE(E_BUSY);
-+ }
-+
-+ if (!FmPcdPlcrIsProfileValid(p_FmPort->h_FmPcd, absoluteProfileId))
-+ {
-+ RELEASE_LOCK(p_FmPort->lock);
-+ RETURN_ERROR(MAJOR, E_INVALID_OPERATION, ("Invalid profile"));
-+ }
-+
-+ tmpReg |= (uint32_t)(NIA_ENG_PLCR | NIA_PLCR_ABSOLUTE | absoluteProfileId);
-+
-+ if (p_FmPort->pcdEngines & FM_PCD_PRS) /* e_FM_PCD_SUPPORT_PRS_AND_PLCR */
-+ {
-+ /* update BMI HPNIA */
-+ WRITE_UINT32(*p_BmiHpnia, tmpReg);
-+ }
-+ else /* e_FM_PCD_SUPPORT_PLCR_ONLY */
-+ {
-+ /* rfne may contain FDCS bits, so first we read them. */
-+ tmpReg |= (GET_UINT32(*p_BmiNia) & BMI_RFNE_FDCS_MASK);
-+ /* update BMI NIA */
-+ WRITE_UINT32(*p_BmiNia, tmpReg);
-+ }RELEASE_LOCK(p_FmPort->lock);
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PORT_PcdCcModifyTree(t_Handle h_FmPort, t_Handle h_CcTree)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+ t_Error err = E_OK;
-+ volatile uint32_t *p_BmiCcBase = NULL;
-+ volatile uint32_t *p_BmiNia = NULL;
-+ uint32_t ccTreePhysOffset;
-+
-+ SANITY_CHECK_RETURN_ERROR(h_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(h_CcTree, E_INVALID_HANDLE);
-+
-+ if (p_FmPort->imEn)
-+ RETURN_ERROR(MAJOR, E_INVALID_OPERATION,
-+ ("available for non-independent mode ports only"));
-+
-+ /* get PCD registers pointers */
-+ switch (p_FmPort->portType)
-+ {
-+ case (e_FM_PORT_TYPE_RX_10G):
-+ case (e_FM_PORT_TYPE_RX):
-+ p_BmiNia = &p_FmPort->port.bmi_regs->rx.fmbm_rfne;
-+ break;
-+ case (e_FM_PORT_TYPE_OH_OFFLINE_PARSING):
-+ p_BmiNia = &p_FmPort->port.bmi_regs->oh.fmbm_ofne;
-+ break;
-+ default:
-+ RETURN_ERROR( MAJOR, E_INVALID_OPERATION,
-+ ("available for Rx and offline parsing ports only"));
-+ }
-+
-+ /* check that current NIA is BMI to BMI */
-+ if ((GET_UINT32(*p_BmiNia) & ~BMI_RFNE_FDCS_MASK)
-+ != GET_NIA_BMI_AC_ENQ_FRAME(p_FmPort->h_FmPcd))
-+ RETURN_ERROR( MAJOR, E_INVALID_OPERATION,
-+ ("may be called only for ports in BMI-to-BMI state."));
-+
-+ if (p_FmPort->pcdEngines & FM_PCD_CC)
-+ {
-+ if (p_FmPort->h_IpReassemblyManip)
-+ {
-+ err = FmPcdCcTreeAddIPR(p_FmPort->h_FmPcd, h_CcTree, NULL,
-+ p_FmPort->h_IpReassemblyManip, FALSE);
-+ if (err != E_OK)
-+ {
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+ }
-+ else
-+ if (p_FmPort->h_CapwapReassemblyManip)
-+ {
-+ err = FmPcdCcTreeAddCPR(p_FmPort->h_FmPcd, h_CcTree, NULL,
-+ p_FmPort->h_CapwapReassemblyManip,
-+ FALSE);
-+ if (err != E_OK)
-+ {
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+ }
-+ switch (p_FmPort->portType)
-+ {
-+ case (e_FM_PORT_TYPE_RX_10G):
-+ case (e_FM_PORT_TYPE_RX):
-+ p_BmiCcBase = &p_FmPort->port.bmi_regs->rx.fmbm_rccb;
-+ break;
-+ case (e_FM_PORT_TYPE_OH_OFFLINE_PARSING):
-+ p_BmiCcBase = &p_FmPort->port.bmi_regs->oh.fmbm_occb;
-+ break;
-+ default:
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("Invalid port type"));
-+ }
-+
-+ if (!TRY_LOCK(p_FmPort->h_Spinlock, &p_FmPort->lock))
-+ {
-+ DBG(TRACE, ("FM Port Try Lock - BUSY"));
-+ return ERROR_CODE(E_BUSY);
-+ }
-+ err = FmPcdCcBindTree(p_FmPort->h_FmPcd, NULL, h_CcTree,
-+ &ccTreePhysOffset, h_FmPort);
-+ if (err)
-+ {
-+ RELEASE_LOCK(p_FmPort->lock);
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }WRITE_UINT32(*p_BmiCcBase, ccTreePhysOffset);
-+
-+ p_FmPort->ccTreeId = h_CcTree;
-+ RELEASE_LOCK(p_FmPort->lock);
-+ }
-+ else
-+ RETURN_ERROR( MAJOR, E_INVALID_STATE,
-+ ("Coarse Classification not defined for this port."));
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PORT_AttachPCD(t_Handle h_FmPort)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+ t_Error err = E_OK;
-+
-+ SANITY_CHECK_RETURN_ERROR(h_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmPort->p_FmPortDriverParam, E_INVALID_STATE);
-+
-+ if (p_FmPort->imEn)
-+ RETURN_ERROR(MAJOR, E_INVALID_OPERATION,
-+ ("available for non-independent mode ports only"));
-+
-+ if ((p_FmPort->portType != e_FM_PORT_TYPE_RX_10G)
-+ && (p_FmPort->portType != e_FM_PORT_TYPE_RX)
-+ && (p_FmPort->portType != e_FM_PORT_TYPE_OH_OFFLINE_PARSING))
-+ RETURN_ERROR( MAJOR, E_INVALID_OPERATION,
-+ ("available for Rx and offline parsing ports only"));
-+
-+ if (!TRY_LOCK(p_FmPort->h_Spinlock, &p_FmPort->lock))
-+ {
-+ DBG(TRACE, ("FM Port Try Lock - BUSY"));
-+ return ERROR_CODE(E_BUSY);
-+ }
-+
-+ if (p_FmPort->h_ReassemblyTree)
-+ p_FmPort->pcdEngines |= FM_PCD_CC;
-+
-+ err = AttachPCD(h_FmPort);
-+ RELEASE_LOCK(p_FmPort->lock);
-+
-+ return err;
-+}
-+
-+t_Error FM_PORT_DetachPCD(t_Handle h_FmPort)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+ t_Error err = E_OK;
-+
-+ SANITY_CHECK_RETURN_ERROR(h_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmPort->p_FmPortDriverParam, E_INVALID_STATE);
-+
-+ if (p_FmPort->imEn)
-+ RETURN_ERROR(MAJOR, E_INVALID_OPERATION,
-+ ("available for non-independent mode ports only"));
-+
-+ if ((p_FmPort->portType != e_FM_PORT_TYPE_RX_10G)
-+ && (p_FmPort->portType != e_FM_PORT_TYPE_RX)
-+ && (p_FmPort->portType != e_FM_PORT_TYPE_OH_OFFLINE_PARSING))
-+ RETURN_ERROR( MAJOR, E_INVALID_OPERATION,
-+ ("available for Rx and offline parsing ports only"));
-+
-+ if (!TRY_LOCK(p_FmPort->h_Spinlock, &p_FmPort->lock))
-+ {
-+ DBG(TRACE, ("FM Port Try Lock - BUSY"));
-+ return ERROR_CODE(E_BUSY);
-+ }
-+
-+ err = DetachPCD(h_FmPort);
-+ if (err != E_OK)
-+ {
-+ RELEASE_LOCK(p_FmPort->lock);
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+
-+ if (p_FmPort->h_ReassemblyTree)
-+ p_FmPort->pcdEngines &= ~FM_PCD_CC;
-+ RELEASE_LOCK(p_FmPort->lock);
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PORT_SetPCD(t_Handle h_FmPort, t_FmPortPcdParams *p_PcdParam)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+ t_Error err = E_OK;
-+ t_FmPortPcdParams modifiedPcdParams, *p_PcdParams;
-+ t_FmPcdCcTreeParams *p_FmPcdCcTreeParams;
-+ t_FmPortPcdCcParams fmPortPcdCcParams;
-+ t_FmPortGetSetCcParams fmPortGetSetCcParams;
-+
-+ SANITY_CHECK_RETURN_ERROR(h_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_PcdParam, E_NULL_POINTER);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmPort->p_FmPortDriverParam, E_INVALID_STATE);
-+
-+ if (p_FmPort->imEn)
-+ RETURN_ERROR(MAJOR, E_INVALID_OPERATION,
-+ ("available for non-independent mode ports only"));
-+
-+ if ((p_FmPort->portType != e_FM_PORT_TYPE_RX_10G)
-+ && (p_FmPort->portType != e_FM_PORT_TYPE_RX)
-+ && (p_FmPort->portType != e_FM_PORT_TYPE_OH_OFFLINE_PARSING))
-+ RETURN_ERROR( MAJOR, E_INVALID_OPERATION,
-+ ("available for Rx and offline parsing ports only"));
-+
-+ if (!TRY_LOCK(p_FmPort->h_Spinlock, &p_FmPort->lock))
-+ {
-+ DBG(TRACE, ("FM Port Try Lock - BUSY"));
-+ return ERROR_CODE(E_BUSY);
-+ }
-+
-+ p_FmPort->h_FmPcd = FmGetPcdHandle(p_FmPort->h_Fm);
-+ ASSERT_COND(p_FmPort->h_FmPcd);
-+
-+ if (p_PcdParam->p_CcParams && !p_PcdParam->p_CcParams->h_CcTree)
-+ RETURN_ERROR(MAJOR, E_INVALID_HANDLE,
-+ ("Tree handle must be given if CC is required"));
-+
-+ memcpy(&modifiedPcdParams, p_PcdParam, sizeof(t_FmPortPcdParams));
-+ p_PcdParams = &modifiedPcdParams;
-+ if ((p_PcdParams->h_IpReassemblyManip)
-+#if (DPAA_VERSION >= 11)
-+ || (p_PcdParams->h_CapwapReassemblyManip)
-+#endif /* (DPAA_VERSION >= 11) */
-+ )
-+ {
-+ if ((p_PcdParams->pcdSupport != e_FM_PORT_PCD_SUPPORT_PRS_AND_KG)
-+ && (p_PcdParams->pcdSupport
-+ != e_FM_PORT_PCD_SUPPORT_PRS_AND_KG_AND_CC)
-+ && (p_PcdParams->pcdSupport
-+ != e_FM_PORT_PCD_SUPPORT_PRS_AND_KG_AND_CC_AND_PLCR)
-+ && (p_PcdParams->pcdSupport
-+ != e_FM_PORT_PCD_SUPPORT_PRS_AND_KG_AND_PLCR))
-+ {
-+ RELEASE_LOCK(p_FmPort->lock);
-+ RETURN_ERROR( MAJOR, E_INVALID_STATE,
-+ ("pcdSupport must have KG for supporting Reassembly"));
-+ }
-+ p_FmPort->h_IpReassemblyManip = p_PcdParams->h_IpReassemblyManip;
-+#if (DPAA_VERSION >= 11)
-+ if ((p_PcdParams->h_IpReassemblyManip)
-+ && (p_PcdParams->h_CapwapReassemblyManip))
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE,
-+ ("Either IP-R or CAPWAP-R is allowed"));
-+ if ((p_PcdParams->h_CapwapReassemblyManip)
-+ && (p_FmPort->portType != e_FM_PORT_TYPE_OH_OFFLINE_PARSING))
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE,
-+ ("CAPWAP-R is allowed only on offline-port"));
-+ if (p_PcdParams->h_CapwapReassemblyManip)
-+ p_FmPort->h_CapwapReassemblyManip =
-+ p_PcdParams->h_CapwapReassemblyManip;
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+ if (!p_PcdParams->p_CcParams)
-+ {
-+ if (!((p_PcdParams->pcdSupport == e_FM_PORT_PCD_SUPPORT_PRS_AND_KG)
-+ || (p_PcdParams->pcdSupport
-+ == e_FM_PORT_PCD_SUPPORT_PRS_AND_KG_AND_PLCR)))
-+ {
-+ RELEASE_LOCK(p_FmPort->lock);
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_STATE,
-+ ("PCD initialization structure is not consistent with pcdSupport"));
-+ }
-+
-+ /* No user-tree, need to build internal tree */
-+ p_FmPcdCcTreeParams = (t_FmPcdCcTreeParams*)XX_Malloc(
-+ sizeof(t_FmPcdCcTreeParams));
-+ if (!p_FmPcdCcTreeParams)
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, ("p_FmPcdCcTreeParams"));
-+ memset(p_FmPcdCcTreeParams, 0, sizeof(t_FmPcdCcTreeParams));
-+ p_FmPcdCcTreeParams->h_NetEnv = p_PcdParams->h_NetEnv;
-+ p_FmPort->h_ReassemblyTree = FM_PCD_CcRootBuild(
-+ p_FmPort->h_FmPcd, p_FmPcdCcTreeParams);
-+
-+ if (!p_FmPort->h_ReassemblyTree)
-+ {
-+ RELEASE_LOCK(p_FmPort->lock);
-+ XX_Free(p_FmPcdCcTreeParams);
-+ RETURN_ERROR( MAJOR, E_INVALID_HANDLE,
-+ ("FM_PCD_CcBuildTree for Reassembly failed"));
-+ }
-+ if (p_PcdParams->pcdSupport == e_FM_PORT_PCD_SUPPORT_PRS_AND_KG)
-+ p_PcdParams->pcdSupport =
-+ e_FM_PORT_PCD_SUPPORT_PRS_AND_KG_AND_CC;
-+ else
-+ p_PcdParams->pcdSupport =
-+ e_FM_PORT_PCD_SUPPORT_PRS_AND_KG_AND_CC_AND_PLCR;
-+
-+ memset(&fmPortPcdCcParams, 0, sizeof(t_FmPortPcdCcParams));
-+ fmPortPcdCcParams.h_CcTree = p_FmPort->h_ReassemblyTree;
-+ p_PcdParams->p_CcParams = &fmPortPcdCcParams;
-+ XX_Free(p_FmPcdCcTreeParams);
-+ }
-+
-+ if (p_FmPort->h_IpReassemblyManip)
-+ err = FmPcdCcTreeAddIPR(p_FmPort->h_FmPcd,
-+ p_PcdParams->p_CcParams->h_CcTree,
-+ p_PcdParams->h_NetEnv,
-+ p_FmPort->h_IpReassemblyManip, TRUE);
-+#if (DPAA_VERSION >= 11)
-+ else
-+ if (p_FmPort->h_CapwapReassemblyManip)
-+ err = FmPcdCcTreeAddCPR(p_FmPort->h_FmPcd,
-+ p_PcdParams->p_CcParams->h_CcTree,
-+ p_PcdParams->h_NetEnv,
-+ p_FmPort->h_CapwapReassemblyManip,
-+ TRUE);
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+ if (err != E_OK)
-+ {
-+ if (p_FmPort->h_ReassemblyTree)
-+ {
-+ FM_PCD_CcRootDelete(p_FmPort->h_ReassemblyTree);
-+ p_FmPort->h_ReassemblyTree = NULL;
-+ }RELEASE_LOCK(p_FmPort->lock);
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+ }
-+
-+ if (!FmPcdLockTryLockAll(p_FmPort->h_FmPcd))
-+ {
-+ if (p_FmPort->h_ReassemblyTree)
-+ {
-+ FM_PCD_CcRootDelete(p_FmPort->h_ReassemblyTree);
-+ p_FmPort->h_ReassemblyTree = NULL;
-+ }RELEASE_LOCK(p_FmPort->lock);
-+ DBG(TRACE, ("Try LockAll - BUSY"));
-+ return ERROR_CODE(E_BUSY);
-+ }
-+
-+ err = SetPcd(h_FmPort, p_PcdParams);
-+ if (err)
-+ {
-+ if (p_FmPort->h_ReassemblyTree)
-+ {
-+ FM_PCD_CcRootDelete(p_FmPort->h_ReassemblyTree);
-+ p_FmPort->h_ReassemblyTree = NULL;
-+ }
-+ FmPcdLockUnlockAll(p_FmPort->h_FmPcd);
-+ RELEASE_LOCK(p_FmPort->lock);
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+
-+ if ((p_FmPort->pcdEngines & FM_PCD_PRS)
-+ && (p_PcdParams->p_PrsParams->includeInPrsStatistics))
-+ {
-+ err = FmPcdPrsIncludePortInStatistics(p_FmPort->h_FmPcd,
-+ p_FmPort->hardwarePortId, TRUE);
-+ if (err)
-+ {
-+ DeletePcd(p_FmPort);
-+ if (p_FmPort->h_ReassemblyTree)
-+ {
-+ FM_PCD_CcRootDelete(p_FmPort->h_ReassemblyTree);
-+ p_FmPort->h_ReassemblyTree = NULL;
-+ }
-+ FmPcdLockUnlockAll(p_FmPort->h_FmPcd);
-+ RELEASE_LOCK(p_FmPort->lock);
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+ p_FmPort->includeInPrsStatistics = TRUE;
-+ }
-+
-+ FmPcdIncNetEnvOwners(p_FmPort->h_FmPcd, p_FmPort->netEnvId);
-+
-+ if (FmPcdIsAdvancedOffloadSupported(p_FmPort->h_FmPcd))
-+ {
-+ memset(&fmPortGetSetCcParams, 0, sizeof(t_FmPortGetSetCcParams));
-+
-+ if (p_FmPort->portType == e_FM_PORT_TYPE_OH_OFFLINE_PARSING)
-+ {
-+#ifdef FM_KG_ERASE_FLOW_ID_ERRATA_FMAN_SW004
-+ if ((p_FmPort->fmRevInfo.majorRev < 6) &&
-+ (p_FmPort->pcdEngines & FM_PCD_KG))
-+ {
-+ int i;
-+ for (i = 0; i<p_PcdParams->p_KgParams->numOfSchemes; i++)
-+ /* The following function must be locked */
-+ FmPcdKgCcGetSetParams(p_FmPort->h_FmPcd,
-+ p_PcdParams->p_KgParams->h_Schemes[i],
-+ UPDATE_KG_NIA_CC_WA,
-+ 0);
-+ }
-+#endif /* FM_KG_ERASE_FLOW_ID_ERRATA_FMAN_SW004 */
-+
-+#if (DPAA_VERSION >= 11)
-+ {
-+ t_FmPcdCtrlParamsPage *p_ParamsPage;
-+
-+ FmPortSetGprFunc(p_FmPort, e_FM_PORT_GPR_MURAM_PAGE,
-+ (void**)&p_ParamsPage);
-+ ASSERT_COND(p_ParamsPage);
-+ WRITE_UINT32(p_ParamsPage->postBmiFetchNia,
-+ p_FmPort->savedBmiNia);
-+ }
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+ /* Set post-bmi-fetch nia */
-+ p_FmPort->savedBmiNia &= BMI_RFNE_FDCS_MASK;
-+ p_FmPort->savedBmiNia |= (NIA_FM_CTL_AC_POST_BMI_FETCH
-+ | NIA_ENG_FM_CTL);
-+
-+ /* Set pre-bmi-fetch nia */
-+ fmPortGetSetCcParams.setCcParams.type = UPDATE_NIA_PNDN;
-+#if (DPAA_VERSION >= 11)
-+ fmPortGetSetCcParams.setCcParams.nia =
-+ (NIA_FM_CTL_AC_PRE_BMI_FETCH_FULL_FRAME | NIA_ENG_FM_CTL);
-+#else
-+ fmPortGetSetCcParams.setCcParams.nia = (NIA_FM_CTL_AC_PRE_BMI_FETCH_HEADER | NIA_ENG_FM_CTL);
-+#endif /* (DPAA_VERSION >= 11) */
-+ if ((err = FmPortGetSetCcParams(p_FmPort, &fmPortGetSetCcParams))
-+ != E_OK)
-+ {
-+ DeletePcd(p_FmPort);
-+ if (p_FmPort->h_ReassemblyTree)
-+ {
-+ FM_PCD_CcRootDelete(p_FmPort->h_ReassemblyTree);
-+ p_FmPort->h_ReassemblyTree = NULL;
-+ }
-+ FmPcdLockUnlockAll(p_FmPort->h_FmPcd);
-+ RELEASE_LOCK(p_FmPort->lock);
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+ }
-+
-+ FmPcdLockUnlockAll(p_FmPort->h_FmPcd);
-+
-+ /* Set pop-to-next-step nia */
-+#if (DPAA_VERSION == 10)
-+ if (p_FmPort->fmRevInfo.majorRev < 6)
-+ {
-+ fmPortGetSetCcParams.setCcParams.type = UPDATE_NIA_PNEN;
-+ fmPortGetSetCcParams.setCcParams.nia = NIA_FM_CTL_AC_POP_TO_N_STEP | NIA_ENG_FM_CTL;
-+ }
-+ else
-+ {
-+#endif /* (DPAA_VERSION == 10) */
-+ fmPortGetSetCcParams.getCcParams.type = GET_NIA_FPNE;
-+#if (DPAA_VERSION == 10)
-+ }
-+#endif /* (DPAA_VERSION == 10) */
-+ if ((err = FmPortGetSetCcParams(h_FmPort, &fmPortGetSetCcParams))
-+ != E_OK)
-+ {
-+ DeletePcd(p_FmPort);
-+ if (p_FmPort->h_ReassemblyTree)
-+ {
-+ FM_PCD_CcRootDelete(p_FmPort->h_ReassemblyTree);
-+ p_FmPort->h_ReassemblyTree = NULL;
-+ }RELEASE_LOCK(p_FmPort->lock);
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+
-+ /* Set post-bmi-prepare-to-enq nia */
-+ fmPortGetSetCcParams.setCcParams.type = UPDATE_NIA_FENE;
-+ fmPortGetSetCcParams.setCcParams.nia = (NIA_FM_CTL_AC_POST_BMI_ENQ
-+ | NIA_ENG_FM_CTL);
-+ if ((err = FmPortGetSetCcParams(h_FmPort, &fmPortGetSetCcParams))
-+ != E_OK)
-+ {
-+ DeletePcd(p_FmPort);
-+ if (p_FmPort->h_ReassemblyTree)
-+ {
-+ FM_PCD_CcRootDelete(p_FmPort->h_ReassemblyTree);
-+ p_FmPort->h_ReassemblyTree = NULL;
-+ }RELEASE_LOCK(p_FmPort->lock);
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+
-+ if ((p_FmPort->h_IpReassemblyManip)
-+ || (p_FmPort->h_CapwapReassemblyManip))
-+ {
-+#if (DPAA_VERSION == 10)
-+ if (p_FmPort->fmRevInfo.majorRev < 6)
-+ {
-+ /* Overwrite post-bmi-prepare-to-enq nia */
-+ fmPortGetSetCcParams.setCcParams.type = UPDATE_NIA_FENE;
-+ fmPortGetSetCcParams.setCcParams.nia = (NIA_FM_CTL_AC_POST_BMI_ENQ_ORR | NIA_ENG_FM_CTL | NIA_ORDER_RESTOR);
-+ fmPortGetSetCcParams.setCcParams.overwrite = TRUE;
-+ }
-+ else
-+ {
-+#endif /* (DPAA_VERSION == 10) */
-+ /* Set the ORR bit (for order-restoration) */
-+ fmPortGetSetCcParams.setCcParams.type = UPDATE_NIA_FPNE;
-+ fmPortGetSetCcParams.setCcParams.nia =
-+ fmPortGetSetCcParams.getCcParams.nia | NIA_ORDER_RESTOR;
-+#if (DPAA_VERSION == 10)
-+ }
-+#endif /* (DPAA_VERSION == 10) */
-+ if ((err = FmPortGetSetCcParams(h_FmPort, &fmPortGetSetCcParams))
-+ != E_OK)
-+ {
-+ DeletePcd(p_FmPort);
-+ if (p_FmPort->h_ReassemblyTree)
-+ {
-+ FM_PCD_CcRootDelete(p_FmPort->h_ReassemblyTree);
-+ p_FmPort->h_ReassemblyTree = NULL;
-+ }RELEASE_LOCK(p_FmPort->lock);
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+ }
-+ }
-+ else
-+ FmPcdLockUnlockAll(p_FmPort->h_FmPcd);
-+
-+#if (DPAA_VERSION >= 11)
-+ {
-+ t_FmPcdCtrlParamsPage *p_ParamsPage;
-+
-+ memset(&fmPortGetSetCcParams, 0, sizeof(t_FmPortGetSetCcParams));
-+
-+ fmPortGetSetCcParams.setCcParams.type = UPDATE_NIA_CMNE;
-+ if (FmPcdIsAdvancedOffloadSupported(p_FmPort->h_FmPcd))
-+ fmPortGetSetCcParams.setCcParams.nia = NIA_FM_CTL_AC_POP_TO_N_STEP
-+ | NIA_ENG_FM_CTL;
-+ else
-+ fmPortGetSetCcParams.setCcParams.nia =
-+ NIA_FM_CTL_AC_NO_IPACC_POP_TO_N_STEP | NIA_ENG_FM_CTL;
-+ if ((err = FmPortGetSetCcParams(h_FmPort, &fmPortGetSetCcParams))
-+ != E_OK)
-+ {
-+ DeletePcd(p_FmPort);
-+ if (p_FmPort->h_ReassemblyTree)
-+ {
-+ FM_PCD_CcRootDelete(p_FmPort->h_ReassemblyTree);
-+ p_FmPort->h_ReassemblyTree = NULL;
-+ }RELEASE_LOCK(p_FmPort->lock);
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+
-+ FmPortSetGprFunc(p_FmPort, e_FM_PORT_GPR_MURAM_PAGE,
-+ (void**)&p_ParamsPage);
-+ ASSERT_COND(p_ParamsPage);
-+
-+ if (FmPcdIsAdvancedOffloadSupported(p_FmPort->h_FmPcd))
-+ WRITE_UINT32(
-+ p_ParamsPage->misc,
-+ GET_UINT32(p_ParamsPage->misc) | FM_CTL_PARAMS_PAGE_OFFLOAD_SUPPORT_EN);
-+
-+ if ((p_FmPort->h_IpReassemblyManip)
-+ || (p_FmPort->h_CapwapReassemblyManip))
-+ {
-+ if (p_FmPort->portType == e_FM_PORT_TYPE_OH_OFFLINE_PARSING)
-+ WRITE_UINT32(
-+ p_ParamsPage->discardMask,
-+ GET_UINT32(p_FmPort->p_FmPortBmiRegs->ohPortBmiRegs.fmbm_ofsdm));
-+ else
-+ WRITE_UINT32(
-+ p_ParamsPage->discardMask,
-+ GET_UINT32(p_FmPort->p_FmPortBmiRegs->rxPortBmiRegs.fmbm_rfsdm));
-+ }
-+#ifdef FM_ERROR_VSP_NO_MATCH_SW006
-+ if (p_FmPort->vspe)
-+ WRITE_UINT32(
-+ p_ParamsPage->misc,
-+ GET_UINT32(p_ParamsPage->misc) | (p_FmPort->dfltRelativeId & FM_CTL_PARAMS_PAGE_ERROR_VSP_MASK));
-+#endif /* FM_ERROR_VSP_NO_MATCH_SW006 */
-+ }
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+ err = AttachPCD(h_FmPort);
-+ if (err)
-+ {
-+ DeletePcd(p_FmPort);
-+ if (p_FmPort->h_ReassemblyTree)
-+ {
-+ FM_PCD_CcRootDelete(p_FmPort->h_ReassemblyTree);
-+ p_FmPort->h_ReassemblyTree = NULL;
-+ }RELEASE_LOCK(p_FmPort->lock);
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+
-+ RELEASE_LOCK(p_FmPort->lock);
-+
-+ return err;
-+}
-+
-+t_Error FM_PORT_DeletePCD(t_Handle h_FmPort)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+ t_Error err = E_OK;
-+
-+ SANITY_CHECK_RETURN_ERROR(h_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmPort->p_FmPortDriverParam, E_INVALID_STATE);
-+
-+ if (p_FmPort->imEn)
-+ RETURN_ERROR(MAJOR, E_INVALID_OPERATION,
-+ ("available for non-independant mode ports only"));
-+
-+ if ((p_FmPort->portType != e_FM_PORT_TYPE_RX_10G)
-+ && (p_FmPort->portType != e_FM_PORT_TYPE_RX)
-+ && (p_FmPort->portType != e_FM_PORT_TYPE_OH_OFFLINE_PARSING))
-+ RETURN_ERROR( MAJOR, E_INVALID_OPERATION,
-+ ("available for Rx and offline parsing ports only"));
-+
-+ if (!TRY_LOCK(p_FmPort->h_Spinlock, &p_FmPort->lock))
-+ {
-+ DBG(TRACE, ("FM Port Try Lock - BUSY"));
-+ return ERROR_CODE(E_BUSY);
-+ }
-+
-+ err = DetachPCD(h_FmPort);
-+ if (err)
-+ {
-+ RELEASE_LOCK(p_FmPort->lock);
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+
-+ FmPcdDecNetEnvOwners(p_FmPort->h_FmPcd, p_FmPort->netEnvId);
-+
-+ /* we do it anyway, instead of checking if included */
-+ if ((p_FmPort->pcdEngines & FM_PCD_PRS) && p_FmPort->includeInPrsStatistics)
-+ {
-+ FmPcdPrsIncludePortInStatistics(p_FmPort->h_FmPcd,
-+ p_FmPort->hardwarePortId, FALSE);
-+ p_FmPort->includeInPrsStatistics = FALSE;
-+ }
-+
-+ if (!FmPcdLockTryLockAll(p_FmPort->h_FmPcd))
-+ {
-+ RELEASE_LOCK(p_FmPort->lock);
-+ DBG(TRACE, ("Try LockAll - BUSY"));
-+ return ERROR_CODE(E_BUSY);
-+ }
-+
-+ err = DeletePcd(h_FmPort);
-+ FmPcdLockUnlockAll(p_FmPort->h_FmPcd);
-+ if (err)
-+ {
-+ RELEASE_LOCK(p_FmPort->lock);
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+
-+ if (p_FmPort->h_ReassemblyTree)
-+ {
-+ err = FM_PCD_CcRootDelete(p_FmPort->h_ReassemblyTree);
-+ if (err)
-+ {
-+ RELEASE_LOCK(p_FmPort->lock);
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+ p_FmPort->h_ReassemblyTree = NULL;
-+ }RELEASE_LOCK(p_FmPort->lock);
-+
-+ return err;
-+}
-+
-+t_Error FM_PORT_PcdKgBindSchemes(t_Handle h_FmPort,
-+ t_FmPcdPortSchemesParams *p_PortScheme)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+ t_FmPcdKgInterModuleBindPortToSchemes schemeBind;
-+ t_Error err = E_OK;
-+ uint32_t tmpScmVec = 0;
-+ int i;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmPort->p_FmPortDriverParam, E_INVALID_STATE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort->pcdEngines & FM_PCD_KG,
-+ E_INVALID_STATE);
-+
-+ schemeBind.netEnvId = p_FmPort->netEnvId;
-+ schemeBind.hardwarePortId = p_FmPort->hardwarePortId;
-+ schemeBind.numOfSchemes = p_PortScheme->numOfSchemes;
-+ schemeBind.useClsPlan = p_FmPort->useClsPlan;
-+ for (i = 0; i < schemeBind.numOfSchemes; i++)
-+ {
-+ schemeBind.schemesIds[i] = FmPcdKgGetSchemeId(
-+ p_PortScheme->h_Schemes[i]);
-+ /* build vector */
-+ tmpScmVec |= 1 << (31 - (uint32_t)schemeBind.schemesIds[i]);
-+ }
-+
-+ if (!TRY_LOCK(p_FmPort->h_Spinlock, &p_FmPort->lock))
-+ {
-+ DBG(TRACE, ("FM Port Try Lock - BUSY"));
-+ return ERROR_CODE(E_BUSY);
-+ }
-+
-+ err = FmPcdKgBindPortToSchemes(p_FmPort->h_FmPcd, &schemeBind);
-+ if (err == E_OK)
-+ p_FmPort->schemesPerPortVector |= tmpScmVec;
-+
-+#ifdef FM_KG_ERASE_FLOW_ID_ERRATA_FMAN_SW004
-+ if ((FmPcdIsAdvancedOffloadSupported(p_FmPort->h_FmPcd)) &&
-+ (p_FmPort->portType == e_FM_PORT_TYPE_OH_OFFLINE_PARSING) &&
-+ (p_FmPort->fmRevInfo.majorRev < 6))
-+ {
-+ for (i=0; i<p_PortScheme->numOfSchemes; i++)
-+ FmPcdKgCcGetSetParams(p_FmPort->h_FmPcd, p_PortScheme->h_Schemes[i], UPDATE_KG_NIA_CC_WA, 0);
-+ }
-+#endif /* FM_KG_ERASE_FLOW_ID_ERRATA_FMAN_SW004 */
-+
-+ RELEASE_LOCK(p_FmPort->lock);
-+
-+ return err;
-+}
-+
-+t_Error FM_PORT_PcdKgUnbindSchemes(t_Handle h_FmPort,
-+ t_FmPcdPortSchemesParams *p_PortScheme)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+ t_FmPcdKgInterModuleBindPortToSchemes schemeBind;
-+ t_Error err = E_OK;
-+ uint32_t tmpScmVec = 0;
-+ int i;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmPort->p_FmPortDriverParam, E_INVALID_STATE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort->pcdEngines & FM_PCD_KG,
-+ E_INVALID_STATE);
-+
-+ schemeBind.netEnvId = p_FmPort->netEnvId;
-+ schemeBind.hardwarePortId = p_FmPort->hardwarePortId;
-+ schemeBind.numOfSchemes = p_PortScheme->numOfSchemes;
-+ for (i = 0; i < schemeBind.numOfSchemes; i++)
-+ {
-+ schemeBind.schemesIds[i] = FmPcdKgGetSchemeId(
-+ p_PortScheme->h_Schemes[i]);
-+ /* build vector */
-+ tmpScmVec |= 1 << (31 - (uint32_t)schemeBind.schemesIds[i]);
-+ }
-+
-+ if (!TRY_LOCK(p_FmPort->h_Spinlock, &p_FmPort->lock))
-+ {
-+ DBG(TRACE, ("FM Port Try Lock - BUSY"));
-+ return ERROR_CODE(E_BUSY);
-+ }
-+
-+ err = FmPcdKgUnbindPortToSchemes(p_FmPort->h_FmPcd, &schemeBind);
-+ if (err == E_OK)
-+ p_FmPort->schemesPerPortVector &= ~tmpScmVec;
-+ RELEASE_LOCK(p_FmPort->lock);
-+
-+ return err;
-+}
-+
-+t_Error FM_PORT_AddCongestionGrps(t_Handle h_FmPort,
-+ t_FmPortCongestionGrps *p_CongestionGrps)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+ uint8_t priorityTmpArray[FM_PORT_NUM_OF_CONGESTION_GRPS];
-+ uint8_t mod, index;
-+ uint32_t i, grpsMap[FMAN_PORT_CG_MAP_NUM];
-+ int err;
-+#if (DPAA_VERSION >= 11)
-+ int j;
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+
-+ /* un-necessary check of the indexes; probably will be needed in the future when there
-+ will be more CGs available ....
-+ for (i=0; i<p_CongestionGrps->numOfCongestionGrpsToConsider; i++)
-+ if (p_CongestionGrps->congestionGrpsToConsider[i] >= FM_PORT_NUM_OF_CONGESTION_GRPS)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("CG id!"));
-+ */
-+
-+#ifdef FM_NO_OP_OBSERVED_CGS
-+ if ((p_FmPort->fmRevInfo.majorRev != 4) &&
-+ (p_FmPort->fmRevInfo.majorRev < 6))
-+ {
-+ if ((p_FmPort->portType != e_FM_PORT_TYPE_RX_10G) &&
-+ (p_FmPort->portType != e_FM_PORT_TYPE_RX))
-+ RETURN_ERROR(MAJOR, E_NOT_SUPPORTED, ("Available for Rx ports only"));
-+ }
-+ else
-+#endif /* FM_NO_OP_OBSERVED_CGS */
-+ if ((p_FmPort->portType != e_FM_PORT_TYPE_RX_10G)
-+ && (p_FmPort->portType != e_FM_PORT_TYPE_RX)
-+ && (p_FmPort->portType != e_FM_PORT_TYPE_OH_OFFLINE_PARSING))
-+ RETURN_ERROR(MAJOR, E_NOT_SUPPORTED,
-+ ("Available for Rx & OP ports only"));
-+
-+ /* Prepare groups map array */
-+ memset(grpsMap, 0, FMAN_PORT_CG_MAP_NUM * sizeof(uint32_t));
-+ for (i = 0; i < p_CongestionGrps->numOfCongestionGrpsToConsider; i++)
-+ {
-+ index = (uint8_t)(p_CongestionGrps->congestionGrpsToConsider[i] / 32);
-+ mod = (uint8_t)(p_CongestionGrps->congestionGrpsToConsider[i] % 32);
-+ if (p_FmPort->fmRevInfo.majorRev != 4)
-+ grpsMap[7 - index] |= (uint32_t)(1 << mod);
-+ else
-+ grpsMap[0] |= (uint32_t)(1 << mod);
-+ }
-+
-+ memset(&priorityTmpArray, 0,
-+ FM_PORT_NUM_OF_CONGESTION_GRPS * sizeof(uint8_t));
-+
-+ for (i = 0; i < p_CongestionGrps->numOfCongestionGrpsToConsider; i++)
-+ {
-+#if (DPAA_VERSION >= 11)
-+ for (j = 0; j < FM_MAX_NUM_OF_PFC_PRIORITIES; j++)
-+ if (p_CongestionGrps->pfcPrioritiesEn[i][j])
-+ priorityTmpArray[p_CongestionGrps->congestionGrpsToConsider[i]] |=
-+ (0x01 << (FM_MAX_NUM_OF_PFC_PRIORITIES - j - 1));
-+#endif /* (DPAA_VERSION >= 11) */
-+ }
-+
-+#if (DPAA_VERSION >= 11)
-+ for (i = 0; i < FM_PORT_NUM_OF_CONGESTION_GRPS; i++)
-+ {
-+ err = FmSetCongestionGroupPFCpriority(p_FmPort->h_Fm, i,
-+ priorityTmpArray[i]);
-+ if (err)
-+ return err;
-+ }
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+ err = fman_port_add_congestion_grps(&p_FmPort->port, grpsMap);
-+ if (err != 0)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("fman_port_add_congestion_grps"));
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PORT_RemoveCongestionGrps(t_Handle h_FmPort,
-+ t_FmPortCongestionGrps *p_CongestionGrps)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+ uint8_t mod, index;
-+ uint32_t i, grpsMap[FMAN_PORT_CG_MAP_NUM];
-+ int err;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+
-+ {
-+#ifdef FM_NO_OP_OBSERVED_CGS
-+ t_FmRevisionInfo revInfo;
-+
-+ FM_GetRevision(p_FmPort->h_Fm, &revInfo);
-+ if (revInfo.majorRev != 4)
-+ {
-+ if ((p_FmPort->portType != e_FM_PORT_TYPE_RX_10G) &&
-+ (p_FmPort->portType != e_FM_PORT_TYPE_RX))
-+ RETURN_ERROR(MAJOR, E_NOT_SUPPORTED, ("Available for Rx ports only"));
-+ }
-+ else
-+#endif /* FM_NO_OP_OBSERVED_CGS */
-+ if ((p_FmPort->portType != e_FM_PORT_TYPE_RX_10G)
-+ && (p_FmPort->portType != e_FM_PORT_TYPE_RX)
-+ && (p_FmPort->portType != e_FM_PORT_TYPE_OH_OFFLINE_PARSING))
-+ RETURN_ERROR(MAJOR, E_NOT_SUPPORTED,
-+ ("Available for Rx & OP ports only"));
-+ }
-+
-+ /* Prepare groups map array */
-+ memset(grpsMap, 0, FMAN_PORT_CG_MAP_NUM * sizeof(uint32_t));
-+ for (i = 0; i < p_CongestionGrps->numOfCongestionGrpsToConsider; i++)
-+ {
-+ index = (uint8_t)(p_CongestionGrps->congestionGrpsToConsider[i] / 32);
-+ mod = (uint8_t)(p_CongestionGrps->congestionGrpsToConsider[i] % 32);
-+ if (p_FmPort->fmRevInfo.majorRev != 4)
-+ grpsMap[7 - index] |= (uint32_t)(1 << mod);
-+ else
-+ grpsMap[0] |= (uint32_t)(1 << mod);
-+ }
-+
-+#if (DPAA_VERSION >= 11)
-+ for (i = 0; i < p_CongestionGrps->numOfCongestionGrpsToConsider; i++)
-+ {
-+ t_Error err = FmSetCongestionGroupPFCpriority(
-+ p_FmPort->h_Fm, p_CongestionGrps->congestionGrpsToConsider[i],
-+ 0);
-+ if (err)
-+ return err;
-+ }
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+ err = fman_port_remove_congestion_grps(&p_FmPort->port, grpsMap);
-+ if (err != 0)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE,
-+ ("fman_port_remove_congestion_grps"));
-+ return E_OK;
-+}
-+
-+#if (DPAA_VERSION >= 11)
-+t_Error FM_PORT_GetIPv4OptionsCount(t_Handle h_FmPort,
-+ uint32_t *p_Ipv4OptionsCount)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(
-+ (p_FmPort->portType == e_FM_PORT_TYPE_OH_OFFLINE_PARSING),
-+ E_INVALID_VALUE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort->p_ParamsPage, E_INVALID_STATE);
-+ SANITY_CHECK_RETURN_ERROR(p_Ipv4OptionsCount, E_NULL_POINTER);
-+
-+ *p_Ipv4OptionsCount = GET_UINT32(p_FmPort->p_ParamsPage->ipfOptionsCounter);
-+
-+ return E_OK;
-+}
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+t_Error FM_PORT_ConfigDsarSupport(t_Handle h_FmPortRx,
-+ t_FmPortDsarTablesSizes *params)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort *)h_FmPortRx;
-+ p_FmPort->deepSleepVars.autoResMaxSizes = XX_Malloc(
-+ sizeof(struct t_FmPortDsarTablesSizes));
-+ memcpy(p_FmPort->deepSleepVars.autoResMaxSizes, params,
-+ sizeof(struct t_FmPortDsarTablesSizes));
-+ return E_OK;
-+}
-+
-+static t_Error FmPortConfigAutoResForDeepSleepSupport1(t_FmPort *p_FmPort)
-+{
-+ uint32_t *param_page;
-+ t_FmPortDsarTablesSizes *params = p_FmPort->deepSleepVars.autoResMaxSizes;
-+ t_ArCommonDesc *ArCommonDescPtr;
-+ uint32_t size = sizeof(t_ArCommonDesc);
-+ // ARP
-+ // should put here if (params->max_num_of_arp_entries)?
-+ size = ROUND_UP(size,4);
-+ size += sizeof(t_DsarArpDescriptor);
-+ size += sizeof(t_DsarArpBindingEntry) * params->maxNumOfArpEntries;
-+ size += sizeof(t_DsarArpStatistics);
-+ //ICMPV4
-+ size = ROUND_UP(size,4);
-+ size += sizeof(t_DsarIcmpV4Descriptor);
-+ size += sizeof(t_DsarIcmpV4BindingEntry) * params->maxNumOfEchoIpv4Entries;
-+ size += sizeof(t_DsarIcmpV4Statistics);
-+ //ICMPV6
-+ size = ROUND_UP(size,4);
-+ size += sizeof(t_DsarIcmpV6Descriptor);
-+ size += sizeof(t_DsarIcmpV6BindingEntry) * params->maxNumOfEchoIpv6Entries;
-+ size += sizeof(t_DsarIcmpV6Statistics);
-+ //ND
-+ size = ROUND_UP(size,4);
-+ size += sizeof(t_DsarNdDescriptor);
-+ size += sizeof(t_DsarIcmpV6BindingEntry) * params->maxNumOfNdpEntries;
-+ size += sizeof(t_DsarIcmpV6Statistics);
-+ //SNMP
-+ size = ROUND_UP(size,4);
-+ size += sizeof(t_DsarSnmpDescriptor);
-+ size += sizeof(t_DsarSnmpIpv4AddrTblEntry)
-+ * params->maxNumOfSnmpIPV4Entries;
-+ size += sizeof(t_DsarSnmpIpv6AddrTblEntry)
-+ * params->maxNumOfSnmpIPV6Entries;
-+ size += sizeof(t_OidsTblEntry) * params->maxNumOfSnmpOidEntries;
-+ size += params->maxNumOfSnmpOidChar;
-+ size += sizeof(t_DsarIcmpV6Statistics);
-+ //filters
-+ size = ROUND_UP(size,4);
-+ size += params->maxNumOfIpProtFiltering;
-+ size = ROUND_UP(size,4);
-+ size += params->maxNumOfUdpPortFiltering * sizeof(t_PortTblEntry);
-+ size = ROUND_UP(size,4);
-+ size += params->maxNumOfTcpPortFiltering * sizeof(t_PortTblEntry);
-+
-+ // add here for more protocols
-+
-+ // statistics
-+ size = ROUND_UP(size,4);
-+ size += sizeof(t_ArStatistics);
-+
-+ ArCommonDescPtr = FM_MURAM_AllocMem(p_FmPort->h_FmMuram, size, 0x10);
-+
-+ param_page =
-+ XX_PhysToVirt(
-+ p_FmPort->fmMuramPhysBaseAddr
-+ + GET_UINT32(p_FmPort->p_FmPortBmiRegs->rxPortBmiRegs.fmbm_rgpr));
-+ WRITE_UINT32(
-+ *param_page,
-+ (uint32_t)(XX_VirtToPhys(ArCommonDescPtr) - p_FmPort->fmMuramPhysBaseAddr));
-+ return E_OK;
-+}
-+
-+t_FmPortDsarTablesSizes* FM_PORT_GetDsarTablesMaxSizes(t_Handle h_FmPortRx)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort *)h_FmPortRx;
-+ return p_FmPort->deepSleepVars.autoResMaxSizes;
-+}
-+
-+struct arOffsets
-+{
-+ uint32_t arp;
-+ uint32_t nd;
-+ uint32_t icmpv4;
-+ uint32_t icmpv6;
-+ uint32_t snmp;
-+ uint32_t stats;
-+ uint32_t filtIp;
-+ uint32_t filtUdp;
-+ uint32_t filtTcp;
-+};
-+
-+static uint32_t AR_ComputeOffsets(struct arOffsets* of,
-+ struct t_FmPortDsarParams *params,
-+ t_FmPort *p_FmPort)
-+{
-+ uint32_t size = sizeof(t_ArCommonDesc);
-+ // ARP
-+ if (params->p_AutoResArpInfo)
-+ {
-+ size = ROUND_UP(size,4);
-+ of->arp = size;
-+ size += sizeof(t_DsarArpDescriptor);
-+ size += sizeof(t_DsarArpBindingEntry)
-+ * params->p_AutoResArpInfo->tableSize;
-+ size += sizeof(t_DsarArpStatistics);
-+ }
-+ // ICMPV4
-+ if (params->p_AutoResEchoIpv4Info)
-+ {
-+ size = ROUND_UP(size,4);
-+ of->icmpv4 = size;
-+ size += sizeof(t_DsarIcmpV4Descriptor);
-+ size += sizeof(t_DsarIcmpV4BindingEntry)
-+ * params->p_AutoResEchoIpv4Info->tableSize;
-+ size += sizeof(t_DsarIcmpV4Statistics);
-+ }
-+ // ICMPV6
-+ if (params->p_AutoResEchoIpv6Info)
-+ {
-+ size = ROUND_UP(size,4);
-+ of->icmpv6 = size;
-+ size += sizeof(t_DsarIcmpV6Descriptor);
-+ size += sizeof(t_DsarIcmpV6BindingEntry)
-+ * params->p_AutoResEchoIpv6Info->tableSize;
-+ size += sizeof(t_DsarIcmpV6Statistics);
-+ }
-+ // ND
-+ if (params->p_AutoResNdpInfo)
-+ {
-+ size = ROUND_UP(size,4);
-+ of->nd = size;
-+ size += sizeof(t_DsarNdDescriptor);
-+ size += sizeof(t_DsarIcmpV6BindingEntry)
-+ * (params->p_AutoResNdpInfo->tableSizeAssigned
-+ + params->p_AutoResNdpInfo->tableSizeTmp);
-+ size += sizeof(t_DsarIcmpV6Statistics);
-+ }
-+ // SNMP
-+ if (params->p_AutoResSnmpInfo)
-+ {
-+ size = ROUND_UP(size,4);
-+ of->snmp = size;
-+ size += sizeof(t_DsarSnmpDescriptor);
-+ size += sizeof(t_DsarSnmpIpv4AddrTblEntry)
-+ * params->p_AutoResSnmpInfo->numOfIpv4Addresses;
-+ size += sizeof(t_DsarSnmpIpv6AddrTblEntry)
-+ * params->p_AutoResSnmpInfo->numOfIpv6Addresses;
-+ size += sizeof(t_OidsTblEntry) * params->p_AutoResSnmpInfo->oidsTblSize;
-+ size += p_FmPort->deepSleepVars.autoResMaxSizes->maxNumOfSnmpOidChar;
-+ size += sizeof(t_DsarIcmpV6Statistics);
-+ }
-+ //filters
-+ size = ROUND_UP(size,4);
-+ if (params->p_AutoResFilteringInfo)
-+ {
-+ of->filtIp = size;
-+ size += params->p_AutoResFilteringInfo->ipProtTableSize;
-+ size = ROUND_UP(size,4);
-+ of->filtUdp = size;
-+ size += params->p_AutoResFilteringInfo->udpPortsTableSize
-+ * sizeof(t_PortTblEntry);
-+ size = ROUND_UP(size,4);
-+ of->filtTcp = size;
-+ size += params->p_AutoResFilteringInfo->tcpPortsTableSize
-+ * sizeof(t_PortTblEntry);
-+ }
-+ // add here for more protocols
-+ // statistics
-+ size = ROUND_UP(size,4);
-+ of->stats = size;
-+ size += sizeof(t_ArStatistics);
-+ return size;
-+}
-+
-+uint32_t* ARDesc;
-+void PrsEnable(t_Handle p_FmPcd);
-+void PrsDisable(t_Handle p_FmPcd);
-+int PrsIsEnabled(t_Handle p_FmPcd);
-+t_Handle FM_PCD_GetHcPort(t_Handle h_FmPcd);
-+
-+static t_Error DsarCheckParams(t_FmPortDsarParams *params,
-+ t_FmPortDsarTablesSizes *sizes)
-+{
-+ bool macInit = FALSE;
-+ uint8_t mac[6];
-+ int i = 0;
-+
-+ // check table sizes
-+ if (params->p_AutoResArpInfo
-+ && sizes->maxNumOfArpEntries < params->p_AutoResArpInfo->tableSize)
-+ RETURN_ERROR(
-+ MAJOR, E_INVALID_VALUE,
-+ ("DSAR: Arp table size exceeds the configured maximum size."));
-+ if (params->p_AutoResEchoIpv4Info
-+ && sizes->maxNumOfEchoIpv4Entries
-+ < params->p_AutoResEchoIpv4Info->tableSize)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("DSAR: EchoIpv4 table size exceeds the configured maximum size."));
-+ if (params->p_AutoResNdpInfo
-+ && sizes->maxNumOfNdpEntries
-+ < params->p_AutoResNdpInfo->tableSizeAssigned
-+ + params->p_AutoResNdpInfo->tableSizeTmp)
-+ RETURN_ERROR(
-+ MAJOR, E_INVALID_VALUE,
-+ ("DSAR: NDP table size exceeds the configured maximum size."));
-+ if (params->p_AutoResEchoIpv6Info
-+ && sizes->maxNumOfEchoIpv6Entries
-+ < params->p_AutoResEchoIpv6Info->tableSize)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("DSAR: EchoIpv6 table size exceeds the configured maximum size."));
-+ if (params->p_AutoResSnmpInfo
-+ && sizes->maxNumOfSnmpOidEntries
-+ < params->p_AutoResSnmpInfo->oidsTblSize)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("DSAR: Snmp Oid table size exceeds the configured maximum size."));
-+ if (params->p_AutoResSnmpInfo
-+ && sizes->maxNumOfSnmpIPV4Entries
-+ < params->p_AutoResSnmpInfo->numOfIpv4Addresses)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("DSAR: Snmp ipv4 table size exceeds the configured maximum size."));
-+ if (params->p_AutoResSnmpInfo
-+ && sizes->maxNumOfSnmpIPV6Entries
-+ < params->p_AutoResSnmpInfo->numOfIpv6Addresses)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("DSAR: Snmp ipv6 table size exceeds the configured maximum size."));
-+ if (params->p_AutoResFilteringInfo)
-+ {
-+ if (sizes->maxNumOfIpProtFiltering
-+ < params->p_AutoResFilteringInfo->ipProtTableSize)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("DSAR: ip filter table size exceeds the configured maximum size."));
-+ if (sizes->maxNumOfTcpPortFiltering
-+ < params->p_AutoResFilteringInfo->udpPortsTableSize)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("DSAR: udp filter table size exceeds the configured maximum size."));
-+ if (sizes->maxNumOfUdpPortFiltering
-+ < params->p_AutoResFilteringInfo->tcpPortsTableSize)
-+ RETURN_ERROR(
-+ MAJOR,
-+ E_INVALID_VALUE,
-+ ("DSAR: tcp filter table size exceeds the configured maximum size."));
-+ }
-+ /* check only 1 MAC address is configured (this is what ucode currently supports) */
-+ if (params->p_AutoResArpInfo && params->p_AutoResArpInfo->tableSize)
-+ {
-+ memcpy(mac, params->p_AutoResArpInfo->p_AutoResTable[0].mac, 6);
-+ i = 1;
-+ macInit = TRUE;
-+
-+ for (; i < params->p_AutoResArpInfo->tableSize; i++)
-+ if (memcmp(mac, params->p_AutoResArpInfo->p_AutoResTable[i].mac, 6))
-+ RETURN_ERROR(
-+ MAJOR, E_INVALID_VALUE,
-+ ("DSAR: Only 1 mac address is currently supported."));
-+ }
-+ if (params->p_AutoResEchoIpv4Info
-+ && params->p_AutoResEchoIpv4Info->tableSize)
-+ {
-+ i = 0;
-+ if (!macInit)
-+ {
-+ memcpy(mac, params->p_AutoResEchoIpv4Info->p_AutoResTable[0].mac,
-+ 6);
-+ i = 1;
-+ macInit = TRUE;
-+ }
-+ for (; i < params->p_AutoResEchoIpv4Info->tableSize; i++)
-+ if (memcmp(mac,
-+ params->p_AutoResEchoIpv4Info->p_AutoResTable[i].mac, 6))
-+ RETURN_ERROR(
-+ MAJOR, E_INVALID_VALUE,
-+ ("DSAR: Only 1 mac address is currently supported."));
-+ }
-+ if (params->p_AutoResEchoIpv6Info
-+ && params->p_AutoResEchoIpv6Info->tableSize)
-+ {
-+ i = 0;
-+ if (!macInit)
-+ {
-+ memcpy(mac, params->p_AutoResEchoIpv6Info->p_AutoResTable[0].mac,
-+ 6);
-+ i = 1;
-+ macInit = TRUE;
-+ }
-+ for (; i < params->p_AutoResEchoIpv6Info->tableSize; i++)
-+ if (memcmp(mac,
-+ params->p_AutoResEchoIpv6Info->p_AutoResTable[i].mac, 6))
-+ RETURN_ERROR(
-+ MAJOR, E_INVALID_VALUE,
-+ ("DSAR: Only 1 mac address is currently supported."));
-+ }
-+ if (params->p_AutoResNdpInfo && params->p_AutoResNdpInfo->tableSizeAssigned)
-+ {
-+ i = 0;
-+ if (!macInit)
-+ {
-+ memcpy(mac, params->p_AutoResNdpInfo->p_AutoResTableAssigned[0].mac,
-+ 6);
-+ i = 1;
-+ macInit = TRUE;
-+ }
-+ for (; i < params->p_AutoResNdpInfo->tableSizeAssigned; i++)
-+ if (memcmp(mac,
-+ params->p_AutoResNdpInfo->p_AutoResTableAssigned[i].mac,
-+ 6))
-+ RETURN_ERROR(
-+ MAJOR, E_INVALID_VALUE,
-+ ("DSAR: Only 1 mac address is currently supported."));
-+ }
-+ if (params->p_AutoResNdpInfo && params->p_AutoResNdpInfo->tableSizeTmp)
-+ {
-+ i = 0;
-+ if (!macInit)
-+ {
-+ memcpy(mac, params->p_AutoResNdpInfo->p_AutoResTableTmp[0].mac, 6);
-+ i = 1;
-+ }
-+ for (; i < params->p_AutoResNdpInfo->tableSizeTmp; i++)
-+ if (memcmp(mac, params->p_AutoResNdpInfo->p_AutoResTableTmp[i].mac,
-+ 6))
-+ RETURN_ERROR(
-+ MAJOR, E_INVALID_VALUE,
-+ ("DSAR: Only 1 mac address is currently supported."));
-+ }
-+ return E_OK;
-+}
-+
-+static int GetBERLen(uint8_t* buf)
-+{
-+ if (*buf & 0x80)
-+ {
-+ if ((*buf & 0x7F) == 1)
-+ return buf[1];
-+ else
-+ return *(uint16_t*)&buf[1]; // assuming max len is 2
-+ }
-+ else
-+ return buf[0];
-+}
-+#define TOTAL_BER_LEN(len) (len < 128) ? len + 2 : len + 3
-+
-+#define SCFG_FMCLKDPSLPCR_ADDR 0xFFE0FC00C
-+#define SCFG_FMCLKDPSLPCR_DS_VAL 0x08402000
-+#define SCFG_FMCLKDPSLPCR_NORMAL_VAL 0x00402000
-+static int fm_soc_suspend(void)
-+{
-+ uint32_t *fmclk, tmp32;
-+ fmclk = ioremap(SCFG_FMCLKDPSLPCR_ADDR, 4);
-+ tmp32 = GET_UINT32(*fmclk);
-+ WRITE_UINT32(*fmclk, SCFG_FMCLKDPSLPCR_DS_VAL);
-+ tmp32 = GET_UINT32(*fmclk);
-+ iounmap(fmclk);
-+ return 0;
-+}
-+
-+void fm_clk_down(void)
-+{
-+ uint32_t *fmclk, tmp32;
-+ fmclk = ioremap(SCFG_FMCLKDPSLPCR_ADDR, 4);
-+ tmp32 = GET_UINT32(*fmclk);
-+ WRITE_UINT32(*fmclk, SCFG_FMCLKDPSLPCR_DS_VAL | 0x40000000);
-+ tmp32 = GET_UINT32(*fmclk);
-+ iounmap(fmclk);
-+}
-+
-+t_Error FM_PORT_EnterDsar(t_Handle h_FmPortRx, t_FmPortDsarParams *params)
-+{
-+ int i, j;
-+ t_Error err;
-+ uint32_t nia;
-+ t_FmPort *p_FmPort = (t_FmPort *)h_FmPortRx;
-+ t_FmPort *p_FmPortTx = (t_FmPort *)params->h_FmPortTx;
-+ t_DsarArpDescriptor *ArpDescriptor;
-+ t_DsarIcmpV4Descriptor* ICMPV4Descriptor;
-+ t_DsarIcmpV6Descriptor* ICMPV6Descriptor;
-+ t_DsarNdDescriptor* NDDescriptor;
-+
-+ uint64_t fmMuramVirtBaseAddr = (uint64_t)PTR_TO_UINT(XX_PhysToVirt(p_FmPort->fmMuramPhysBaseAddr));
-+ uint32_t *param_page = XX_PhysToVirt(p_FmPort->fmMuramPhysBaseAddr + GET_UINT32(p_FmPort->p_FmPortBmiRegs->rxPortBmiRegs.fmbm_rgpr));
-+ t_ArCommonDesc *ArCommonDescPtr = (t_ArCommonDesc*)(XX_PhysToVirt(p_FmPort->fmMuramPhysBaseAddr + GET_UINT32(*param_page)));
-+ struct arOffsets* of;
-+ uint8_t tmp = 0;
-+ t_FmGetSetParams fmGetSetParams;
-+ memset(&fmGetSetParams, 0, sizeof (t_FmGetSetParams));
-+ fmGetSetParams.setParams.type = UPDATE_FPM_BRKC_SLP;
-+ fmGetSetParams.setParams.sleep = 1;
-+
-+ err = DsarCheckParams(params, p_FmPort->deepSleepVars.autoResMaxSizes);
-+ if (err != E_OK)
-+ return err;
-+
-+ p_FmPort->deepSleepVars.autoResOffsets = XX_Malloc(sizeof(struct arOffsets));
-+ of = (struct arOffsets *)p_FmPort->deepSleepVars.autoResOffsets;
-+ IOMemSet32(ArCommonDescPtr, 0, AR_ComputeOffsets(of, params, p_FmPort));
-+
-+ // common
-+ WRITE_UINT8(ArCommonDescPtr->arTxPort, p_FmPortTx->hardwarePortId);
-+ nia = GET_UINT32(p_FmPort->p_FmPortBmiRegs->rxPortBmiRegs.fmbm_rfne); // bmi nia
-+ if ((nia & 0x007C0000) == 0x00440000) // bmi nia is parser
-+ WRITE_UINT32(ArCommonDescPtr->activeHPNIA, GET_UINT32(p_FmPort->p_FmPortBmiRegs->rxPortBmiRegs.fmbm_rfpne));
-+ else
-+ WRITE_UINT32(ArCommonDescPtr->activeHPNIA, nia);
-+ WRITE_UINT16(ArCommonDescPtr->snmpPort, 161);
-+
-+ // ARP
-+ if (params->p_AutoResArpInfo)
-+ {
-+ t_DsarArpBindingEntry* arp_bindings;
-+ ArpDescriptor = (t_DsarArpDescriptor*)(PTR_TO_UINT(ArCommonDescPtr) + of->arp);
-+ WRITE_UINT32(ArCommonDescPtr->p_ArpDescriptor, PTR_TO_UINT(ArpDescriptor) - fmMuramVirtBaseAddr);
-+ arp_bindings = (t_DsarArpBindingEntry*)(PTR_TO_UINT(ArpDescriptor) + sizeof(t_DsarArpDescriptor));
-+ if (params->p_AutoResArpInfo->enableConflictDetection)
-+ WRITE_UINT16(ArpDescriptor->control, 1);
-+ else
-+ WRITE_UINT16(ArpDescriptor->control, 0);
-+ if (params->p_AutoResArpInfo->tableSize)
-+ {
-+ t_FmPortDsarArpEntry* arp_entry = params->p_AutoResArpInfo->p_AutoResTable;
-+ WRITE_UINT16(*(uint16_t*)&ArCommonDescPtr->macStationAddr[0], *(uint16_t*)&arp_entry[0].mac[0]);
-+ WRITE_UINT32(*(uint32_t*)&ArCommonDescPtr->macStationAddr[2], *(uint32_t*)&arp_entry[0].mac[2]);
-+ WRITE_UINT16(ArpDescriptor->numOfBindings, params->p_AutoResArpInfo->tableSize);
-+
-+ for (i = 0; i < params->p_AutoResArpInfo->tableSize; i++)
-+ {
-+ WRITE_UINT32(arp_bindings[i].ipv4Addr, arp_entry[i].ipAddress);
-+ if (arp_entry[i].isVlan)
-+ WRITE_UINT16(arp_bindings[i].vlanId, arp_entry[i].vid & 0xFFF);
-+ }
-+ WRITE_UINT32(ArpDescriptor->p_Bindings, PTR_TO_UINT(arp_bindings) - fmMuramVirtBaseAddr);
-+ }
-+ WRITE_UINT32(ArpDescriptor->p_Statistics, PTR_TO_UINT(arp_bindings) +
-+ sizeof(t_DsarArpBindingEntry) * params->p_AutoResArpInfo->tableSize - fmMuramVirtBaseAddr);
-+ }
-+
-+ // ICMPV4
-+ if (params->p_AutoResEchoIpv4Info)
-+ {
-+ t_DsarIcmpV4BindingEntry* icmpv4_bindings;
-+ ICMPV4Descriptor = (t_DsarIcmpV4Descriptor*)(PTR_TO_UINT(ArCommonDescPtr) + of->icmpv4);
-+ WRITE_UINT32(ArCommonDescPtr->p_IcmpV4Descriptor, PTR_TO_UINT(ICMPV4Descriptor) - fmMuramVirtBaseAddr);
-+ icmpv4_bindings = (t_DsarIcmpV4BindingEntry*)(PTR_TO_UINT(ICMPV4Descriptor) + sizeof(t_DsarIcmpV4Descriptor));
-+ WRITE_UINT16(ICMPV4Descriptor->control, 0);
-+ if (params->p_AutoResEchoIpv4Info->tableSize)
-+ {
-+ t_FmPortDsarArpEntry* arp_entry = params->p_AutoResEchoIpv4Info->p_AutoResTable;
-+ WRITE_UINT16(*(uint16_t*)&ArCommonDescPtr->macStationAddr[0], *(uint16_t*)&arp_entry[0].mac[0]);
-+ WRITE_UINT32(*(uint32_t*)&ArCommonDescPtr->macStationAddr[2], *(uint32_t*)&arp_entry[0].mac[2]);
-+ WRITE_UINT16(ICMPV4Descriptor->numOfBindings, params->p_AutoResEchoIpv4Info->tableSize);
-+
-+ for (i = 0; i < params->p_AutoResEchoIpv4Info->tableSize; i++)
-+ {
-+ WRITE_UINT32(icmpv4_bindings[i].ipv4Addr, arp_entry[i].ipAddress);
-+ if (arp_entry[i].isVlan)
-+ WRITE_UINT16(icmpv4_bindings[i].vlanId, arp_entry[i].vid & 0xFFF);
-+ }
-+ WRITE_UINT32(ICMPV4Descriptor->p_Bindings, PTR_TO_UINT(icmpv4_bindings) - fmMuramVirtBaseAddr);
-+ }
-+ WRITE_UINT32(ICMPV4Descriptor->p_Statistics, PTR_TO_UINT(icmpv4_bindings) +
-+ sizeof(t_DsarIcmpV4BindingEntry) * params->p_AutoResEchoIpv4Info->tableSize - fmMuramVirtBaseAddr);
-+ }
-+
-+ // ICMPV6
-+ if (params->p_AutoResEchoIpv6Info)
-+ {
-+ t_DsarIcmpV6BindingEntry* icmpv6_bindings;
-+ ICMPV6Descriptor = (t_DsarIcmpV6Descriptor*)(PTR_TO_UINT(ArCommonDescPtr) + of->icmpv6);
-+ WRITE_UINT32(ArCommonDescPtr->p_IcmpV6Descriptor, PTR_TO_UINT(ICMPV6Descriptor) - fmMuramVirtBaseAddr);
-+ icmpv6_bindings = (t_DsarIcmpV6BindingEntry*)(PTR_TO_UINT(ICMPV6Descriptor) + sizeof(t_DsarIcmpV6Descriptor));
-+ WRITE_UINT16(ICMPV6Descriptor->control, 0);
-+ if (params->p_AutoResEchoIpv6Info->tableSize)
-+ {
-+ t_FmPortDsarNdpEntry* ndp_entry = params->p_AutoResEchoIpv6Info->p_AutoResTable;
-+ WRITE_UINT16(*(uint16_t*)&ArCommonDescPtr->macStationAddr[0], *(uint16_t*)&ndp_entry[0].mac[0]);
-+ WRITE_UINT32(*(uint32_t*)&ArCommonDescPtr->macStationAddr[2], *(uint32_t*)&ndp_entry[0].mac[2]);
-+ WRITE_UINT16(ICMPV6Descriptor->numOfBindings, params->p_AutoResEchoIpv6Info->tableSize);
-+
-+ for (i = 0; i < params->p_AutoResEchoIpv6Info->tableSize; i++)
-+ {
-+ for (j = 0; j < 4; j++)
-+ WRITE_UINT32(icmpv6_bindings[i].ipv6Addr[j], ndp_entry[i].ipAddress[j]);
-+ if (ndp_entry[i].isVlan)
-+ WRITE_UINT16(*(uint16_t*)&icmpv6_bindings[i].ipv6Addr[4], ndp_entry[i].vid & 0xFFF); // writing vlan
-+ }
-+ WRITE_UINT32(ICMPV6Descriptor->p_Bindings, PTR_TO_UINT(icmpv6_bindings) - fmMuramVirtBaseAddr);
-+ }
-+ WRITE_UINT32(ICMPV6Descriptor->p_Statistics, PTR_TO_UINT(icmpv6_bindings) +
-+ sizeof(t_DsarIcmpV6BindingEntry) * params->p_AutoResEchoIpv6Info->tableSize - fmMuramVirtBaseAddr);
-+ }
-+
-+ // ND
-+ if (params->p_AutoResNdpInfo)
-+ {
-+ t_DsarIcmpV6BindingEntry* icmpv6_bindings;
-+ NDDescriptor = (t_DsarNdDescriptor*)(PTR_TO_UINT(ArCommonDescPtr) + of->nd);
-+ WRITE_UINT32(ArCommonDescPtr->p_NdDescriptor, PTR_TO_UINT(NDDescriptor) - fmMuramVirtBaseAddr);
-+ icmpv6_bindings = (t_DsarIcmpV6BindingEntry*)(PTR_TO_UINT(NDDescriptor) + sizeof(t_DsarNdDescriptor));
-+ if (params->p_AutoResNdpInfo->enableConflictDetection)
-+ WRITE_UINT16(NDDescriptor->control, 1);
-+ else
-+ WRITE_UINT16(NDDescriptor->control, 0);
-+ if (params->p_AutoResNdpInfo->tableSizeAssigned + params->p_AutoResNdpInfo->tableSizeTmp)
-+ {
-+ t_FmPortDsarNdpEntry* ndp_entry = params->p_AutoResNdpInfo->p_AutoResTableAssigned;
-+ WRITE_UINT16(*(uint16_t*)&ArCommonDescPtr->macStationAddr[0], *(uint16_t*)&ndp_entry[0].mac[0]);
-+ WRITE_UINT32(*(uint32_t*)&ArCommonDescPtr->macStationAddr[2], *(uint32_t*)&ndp_entry[0].mac[2]);
-+ WRITE_UINT16(NDDescriptor->numOfBindings, params->p_AutoResNdpInfo->tableSizeAssigned
-+ + params->p_AutoResNdpInfo->tableSizeTmp);
-+
-+ for (i = 0; i < params->p_AutoResNdpInfo->tableSizeAssigned; i++)
-+ {
-+ for (j = 0; j < 4; j++)
-+ WRITE_UINT32(icmpv6_bindings[i].ipv6Addr[j], ndp_entry[i].ipAddress[j]);
-+ if (ndp_entry[i].isVlan)
-+ WRITE_UINT16(*(uint16_t*)&icmpv6_bindings[i].ipv6Addr[4], ndp_entry[i].vid & 0xFFF); // writing vlan
-+ }
-+ ndp_entry = params->p_AutoResNdpInfo->p_AutoResTableTmp;
-+ for (i = 0; i < params->p_AutoResNdpInfo->tableSizeTmp; i++)
-+ {
-+ for (j = 0; j < 4; j++)
-+ WRITE_UINT32(icmpv6_bindings[i + params->p_AutoResNdpInfo->tableSizeAssigned].ipv6Addr[j], ndp_entry[i].ipAddress[j]);
-+ if (ndp_entry[i].isVlan)
-+ WRITE_UINT16(*(uint16_t*)&icmpv6_bindings[i + params->p_AutoResNdpInfo->tableSizeAssigned].ipv6Addr[4], ndp_entry[i].vid & 0xFFF); // writing vlan
-+ }
-+ WRITE_UINT32(NDDescriptor->p_Bindings, PTR_TO_UINT(icmpv6_bindings) - fmMuramVirtBaseAddr);
-+ }
-+ WRITE_UINT32(NDDescriptor->p_Statistics, PTR_TO_UINT(icmpv6_bindings) + sizeof(t_DsarIcmpV6BindingEntry)
-+ * (params->p_AutoResNdpInfo->tableSizeAssigned + params->p_AutoResNdpInfo->tableSizeTmp)
-+ - fmMuramVirtBaseAddr);
-+ WRITE_UINT32(NDDescriptor->solicitedAddr, 0xFFFFFFFF);
-+ }
-+
-+ // SNMP
-+ if (params->p_AutoResSnmpInfo)
-+ {
-+ t_FmPortDsarSnmpInfo *snmpSrc = params->p_AutoResSnmpInfo;
-+ t_DsarSnmpIpv4AddrTblEntry* snmpIpv4Addr;
-+ t_DsarSnmpIpv6AddrTblEntry* snmpIpv6Addr;
-+ t_OidsTblEntry* snmpOid;
-+ uint8_t *charPointer;
-+ int len;
-+ t_DsarSnmpDescriptor* SnmpDescriptor = (t_DsarSnmpDescriptor*)(PTR_TO_UINT(ArCommonDescPtr) + of->snmp);
-+ WRITE_UINT32(ArCommonDescPtr->p_SnmpDescriptor, PTR_TO_UINT(SnmpDescriptor) - fmMuramVirtBaseAddr);
-+ WRITE_UINT16(SnmpDescriptor->control, snmpSrc->control);
-+ WRITE_UINT16(SnmpDescriptor->maxSnmpMsgLength, snmpSrc->maxSnmpMsgLength);
-+ snmpIpv4Addr = (t_DsarSnmpIpv4AddrTblEntry*)(PTR_TO_UINT(SnmpDescriptor) + sizeof(t_DsarSnmpDescriptor));
-+ if (snmpSrc->numOfIpv4Addresses)
-+ {
-+ t_FmPortDsarSnmpIpv4AddrTblEntry* snmpIpv4AddrSrc = snmpSrc->p_Ipv4AddrTbl;
-+ WRITE_UINT16(SnmpDescriptor->numOfIpv4Addresses, snmpSrc->numOfIpv4Addresses);
-+ for (i = 0; i < snmpSrc->numOfIpv4Addresses; i++)
-+ {
-+ WRITE_UINT32(snmpIpv4Addr[i].ipv4Addr, snmpIpv4AddrSrc[i].ipv4Addr);
-+ if (snmpIpv4AddrSrc[i].isVlan)
-+ WRITE_UINT16(snmpIpv4Addr[i].vlanId, snmpIpv4AddrSrc[i].vid & 0xFFF);
-+ }
-+ WRITE_UINT32(SnmpDescriptor->p_Ipv4AddrTbl, PTR_TO_UINT(snmpIpv4Addr) - fmMuramVirtBaseAddr);
-+ }
-+ snmpIpv6Addr = (t_DsarSnmpIpv6AddrTblEntry*)(PTR_TO_UINT(snmpIpv4Addr)
-+ + sizeof(t_DsarSnmpIpv4AddrTblEntry) * snmpSrc->numOfIpv4Addresses);
-+ if (snmpSrc->numOfIpv6Addresses)
-+ {
-+ t_FmPortDsarSnmpIpv6AddrTblEntry* snmpIpv6AddrSrc = snmpSrc->p_Ipv6AddrTbl;
-+ WRITE_UINT16(SnmpDescriptor->numOfIpv6Addresses, snmpSrc->numOfIpv6Addresses);
-+ for (i = 0; i < snmpSrc->numOfIpv6Addresses; i++)
-+ {
-+ for (j = 0; j < 4; j++)
-+ WRITE_UINT32(snmpIpv6Addr[i].ipv6Addr[j], snmpIpv6AddrSrc[i].ipv6Addr[j]);
-+ if (snmpIpv6AddrSrc[i].isVlan)
-+ WRITE_UINT16(snmpIpv6Addr[i].vlanId, snmpIpv6AddrSrc[i].vid & 0xFFF);
-+ }
-+ WRITE_UINT32(SnmpDescriptor->p_Ipv6AddrTbl, PTR_TO_UINT(snmpIpv6Addr) - fmMuramVirtBaseAddr);
-+ }
-+ snmpOid = (t_OidsTblEntry*)(PTR_TO_UINT(snmpIpv6Addr)
-+ + sizeof(t_DsarSnmpIpv6AddrTblEntry) * snmpSrc->numOfIpv6Addresses);
-+ charPointer = (uint8_t*)(PTR_TO_UINT(snmpOid)
-+ + sizeof(t_OidsTblEntry) * snmpSrc->oidsTblSize);
-+ len = TOTAL_BER_LEN(GetBERLen(&snmpSrc->p_RdOnlyCommunityStr[1]));
-+ Mem2IOCpy32(charPointer, snmpSrc->p_RdOnlyCommunityStr, len);
-+ WRITE_UINT32(SnmpDescriptor->p_RdOnlyCommunityStr, PTR_TO_UINT(charPointer) - fmMuramVirtBaseAddr);
-+ charPointer += len;
-+ len = TOTAL_BER_LEN(GetBERLen(&snmpSrc->p_RdWrCommunityStr[1]));
-+ Mem2IOCpy32(charPointer, snmpSrc->p_RdWrCommunityStr, len);
-+ WRITE_UINT32(SnmpDescriptor->p_RdWrCommunityStr, PTR_TO_UINT(charPointer) - fmMuramVirtBaseAddr);
-+ charPointer += len;
-+ WRITE_UINT32(SnmpDescriptor->oidsTblSize, snmpSrc->oidsTblSize);
-+ WRITE_UINT32(SnmpDescriptor->p_OidsTbl, PTR_TO_UINT(snmpOid) - fmMuramVirtBaseAddr);
-+ for (i = 0; i < snmpSrc->oidsTblSize; i++)
-+ {
-+ WRITE_UINT16(snmpOid->oidSize, snmpSrc->p_OidsTbl[i].oidSize);
-+ WRITE_UINT16(snmpOid->resSize, snmpSrc->p_OidsTbl[i].resSize);
-+ Mem2IOCpy32(charPointer, snmpSrc->p_OidsTbl[i].oidVal, snmpSrc->p_OidsTbl[i].oidSize);
-+ WRITE_UINT32(snmpOid->p_Oid, PTR_TO_UINT(charPointer) - fmMuramVirtBaseAddr);
-+ charPointer += snmpSrc->p_OidsTbl[i].oidSize;
-+ if (snmpSrc->p_OidsTbl[i].resSize <= 4)
-+ WRITE_UINT32(snmpOid->resValOrPtr, *snmpSrc->p_OidsTbl[i].resVal);
-+ else
-+ {
-+ Mem2IOCpy32(charPointer, snmpSrc->p_OidsTbl[i].resVal, snmpSrc->p_OidsTbl[i].resSize);
-+ WRITE_UINT32(snmpOid->resValOrPtr, PTR_TO_UINT(charPointer) - fmMuramVirtBaseAddr);
-+ charPointer += snmpSrc->p_OidsTbl[i].resSize;
-+ }
-+ snmpOid++;
-+ }
-+ charPointer = UINT_TO_PTR(ROUND_UP(PTR_TO_UINT(charPointer),4));
-+ WRITE_UINT32(SnmpDescriptor->p_Statistics, PTR_TO_UINT(charPointer) - fmMuramVirtBaseAddr);
-+ }
-+
-+ // filtering
-+ if (params->p_AutoResFilteringInfo)
-+ {
-+ if (params->p_AutoResFilteringInfo->ipProtPassOnHit)
-+ tmp |= IP_PROT_TBL_PASS_MASK;
-+ if (params->p_AutoResFilteringInfo->udpPortPassOnHit)
-+ tmp |= UDP_PORT_TBL_PASS_MASK;
-+ if (params->p_AutoResFilteringInfo->tcpPortPassOnHit)
-+ tmp |= TCP_PORT_TBL_PASS_MASK;
-+ WRITE_UINT8(ArCommonDescPtr->filterControl, tmp);
-+ WRITE_UINT16(ArCommonDescPtr->tcpControlPass, params->p_AutoResFilteringInfo->tcpFlagsMask);
-+
-+ // ip filtering
-+ if (params->p_AutoResFilteringInfo->ipProtTableSize)
-+ {
-+ uint8_t* ip_tbl = (uint8_t*)(PTR_TO_UINT(ArCommonDescPtr) + of->filtIp);
-+ WRITE_UINT8(ArCommonDescPtr->ipProtocolTblSize, params->p_AutoResFilteringInfo->ipProtTableSize);
-+ for (i = 0; i < params->p_AutoResFilteringInfo->ipProtTableSize; i++)
-+ WRITE_UINT8(ip_tbl[i], params->p_AutoResFilteringInfo->p_IpProtTablePtr[i]);
-+ WRITE_UINT32(ArCommonDescPtr->p_IpProtocolFiltTbl, PTR_TO_UINT(ip_tbl) - fmMuramVirtBaseAddr);
-+ }
-+
-+ // udp filtering
-+ if (params->p_AutoResFilteringInfo->udpPortsTableSize)
-+ {
-+ t_PortTblEntry* udp_tbl = (t_PortTblEntry*)(PTR_TO_UINT(ArCommonDescPtr) + of->filtUdp);
-+ WRITE_UINT8(ArCommonDescPtr->udpPortTblSize, params->p_AutoResFilteringInfo->udpPortsTableSize);
-+ for (i = 0; i < params->p_AutoResFilteringInfo->udpPortsTableSize; i++)
-+ {
-+ WRITE_UINT32(udp_tbl[i].Ports,
-+ (params->p_AutoResFilteringInfo->p_UdpPortsTablePtr[i].srcPort << 16) +
-+ params->p_AutoResFilteringInfo->p_UdpPortsTablePtr[i].dstPort);
-+ WRITE_UINT32(udp_tbl[i].PortsMask,
-+ (params->p_AutoResFilteringInfo->p_UdpPortsTablePtr[i].srcPortMask << 16) +
-+ params->p_AutoResFilteringInfo->p_UdpPortsTablePtr[i].dstPortMask);
-+ }
-+ WRITE_UINT32(ArCommonDescPtr->p_UdpPortFiltTbl, PTR_TO_UINT(udp_tbl) - fmMuramVirtBaseAddr);
-+ }
-+
-+ // tcp filtering
-+ if (params->p_AutoResFilteringInfo->tcpPortsTableSize)
-+ {
-+ t_PortTblEntry* tcp_tbl = (t_PortTblEntry*)(PTR_TO_UINT(ArCommonDescPtr) + of->filtTcp);
-+ WRITE_UINT8(ArCommonDescPtr->tcpPortTblSize, params->p_AutoResFilteringInfo->tcpPortsTableSize);
-+ for (i = 0; i < params->p_AutoResFilteringInfo->tcpPortsTableSize; i++)
-+ {
-+ WRITE_UINT32(tcp_tbl[i].Ports,
-+ (params->p_AutoResFilteringInfo->p_TcpPortsTablePtr[i].srcPort << 16) +
-+ params->p_AutoResFilteringInfo->p_TcpPortsTablePtr[i].dstPort);
-+ WRITE_UINT32(tcp_tbl[i].PortsMask,
-+ (params->p_AutoResFilteringInfo->p_TcpPortsTablePtr[i].srcPortMask << 16) +
-+ params->p_AutoResFilteringInfo->p_TcpPortsTablePtr[i].dstPortMask);
-+ }
-+ WRITE_UINT32(ArCommonDescPtr->p_TcpPortFiltTbl, PTR_TO_UINT(tcp_tbl) - fmMuramVirtBaseAddr);
-+ }
-+ }
-+ // common stats
-+ WRITE_UINT32(ArCommonDescPtr->p_ArStats, PTR_TO_UINT(ArCommonDescPtr) + of->stats - fmMuramVirtBaseAddr);
-+
-+ // get into Deep Sleep sequence:
-+
-+ // Ensures that FMan do not enter the idle state. This is done by programing
-+ // FMDPSLPCR[FM_STOP] to one.
-+ fm_soc_suspend();
-+
-+ ARDesc = UINT_TO_PTR(XX_VirtToPhys(ArCommonDescPtr));
-+ return E_OK;
-+
-+}
-+
-+void FM_ChangeClock(t_Handle h_Fm, int hardwarePortId);
-+t_Error FM_PORT_EnterDsarFinal(t_Handle h_DsarRxPort, t_Handle h_DsarTxPort)
-+{
-+ t_FmGetSetParams fmGetSetParams;
-+ t_FmPort *p_FmPort = (t_FmPort *)h_DsarRxPort;
-+ t_FmPort *p_FmPortTx = (t_FmPort *)h_DsarTxPort;
-+ t_Handle *h_FmPcd = FmGetPcd(p_FmPort->h_Fm);
-+ t_FmPort *p_FmPortHc = FM_PCD_GetHcPort(h_FmPcd);
-+ memset(&fmGetSetParams, 0, sizeof (t_FmGetSetParams));
-+ fmGetSetParams.setParams.type = UPDATE_FM_CLD;
-+ FmGetSetParams(p_FmPort->h_Fm, &fmGetSetParams);
-+
-+ /* Issue graceful stop to HC port */
-+ FM_PORT_Disable(p_FmPortHc);
-+
-+ // config tx port
-+ p_FmPort->deepSleepVars.fmbm_tcfg = GET_UINT32(p_FmPortTx->p_FmPortBmiRegs->txPortBmiRegs.fmbm_tcfg);
-+ WRITE_UINT32(p_FmPortTx->p_FmPortBmiRegs->txPortBmiRegs.fmbm_tcfg, GET_UINT32(p_FmPortTx->p_FmPortBmiRegs->txPortBmiRegs.fmbm_tcfg) | BMI_PORT_CFG_IM | BMI_PORT_CFG_EN);
-+ // ????
-+ p_FmPort->deepSleepVars.fmbm_tcmne = GET_UINT32(p_FmPortTx->p_FmPortBmiRegs->txPortBmiRegs.fmbm_tcmne);
-+ WRITE_UINT32(p_FmPortTx->p_FmPortBmiRegs->txPortBmiRegs.fmbm_tcmne, 0xE);
-+ // Stage 7:echo
-+ p_FmPort->deepSleepVars.fmbm_rfpne = GET_UINT32(p_FmPort->p_FmPortBmiRegs->rxPortBmiRegs.fmbm_rfpne);
-+ WRITE_UINT32(p_FmPort->p_FmPortBmiRegs->rxPortBmiRegs.fmbm_rfpne, 0x2E);
-+ if (!PrsIsEnabled(h_FmPcd))
-+ {
-+ p_FmPort->deepSleepVars.dsarEnabledParser = TRUE;
-+ PrsEnable(h_FmPcd);
-+ }
-+ else
-+ p_FmPort->deepSleepVars.dsarEnabledParser = FALSE;
-+
-+ p_FmPort->deepSleepVars.fmbm_rfne = GET_UINT32(p_FmPort->p_FmPortBmiRegs->rxPortBmiRegs.fmbm_rfne);
-+ WRITE_UINT32(p_FmPort->p_FmPortBmiRegs->rxPortBmiRegs.fmbm_rfne, 0x440000);
-+
-+ // save rcfg for restoring: accumulate mode is changed by ucode
-+ p_FmPort->deepSleepVars.fmbm_rcfg = GET_UINT32(p_FmPort->port.bmi_regs->rx.fmbm_rcfg);
-+ WRITE_UINT32(p_FmPort->port.bmi_regs->rx.fmbm_rcfg, p_FmPort->deepSleepVars.fmbm_rcfg | BMI_PORT_CFG_AM);
-+ memset(&fmGetSetParams, 0, sizeof (t_FmGetSetParams));
-+ fmGetSetParams.setParams.type = UPDATE_FPM_BRKC_SLP;
-+ fmGetSetParams.setParams.sleep = 1;
-+ FmGetSetParams(p_FmPort->h_Fm, &fmGetSetParams);
-+
-+// ***** issue external request sync command
-+ memset(&fmGetSetParams, 0, sizeof (t_FmGetSetParams));
-+ fmGetSetParams.setParams.type = UPDATE_FPM_EXTC;
-+ FmGetSetParams(p_FmPort->h_Fm, &fmGetSetParams);
-+ // get
-+ memset(&fmGetSetParams, 0, sizeof (t_FmGetSetParams));
-+ fmGetSetParams.getParams.type = GET_FMFP_EXTC;
-+ FmGetSetParams(p_FmPort->h_Fm, &fmGetSetParams);
-+ if (fmGetSetParams.getParams.fmfp_extc != 0)
-+ {
-+ // clear
-+ memset(&fmGetSetParams, 0, sizeof (t_FmGetSetParams));
-+ fmGetSetParams.setParams.type = UPDATE_FPM_EXTC_CLEAR;
-+ FmGetSetParams(p_FmPort->h_Fm, &fmGetSetParams);
-+}
-+
-+ memset(&fmGetSetParams, 0, sizeof (t_FmGetSetParams));
-+ fmGetSetParams.getParams.type = GET_FMFP_EXTC | GET_FM_NPI;
-+ do
-+ {
-+ FmGetSetParams(p_FmPort->h_Fm, &fmGetSetParams);
-+ } while (fmGetSetParams.getParams.fmfp_extc != 0 && fmGetSetParams.getParams.fm_npi == 0);
-+ if (fmGetSetParams.getParams.fm_npi != 0)
-+ XX_Print("FM: Sync did not finish\n");
-+
-+ // check that all stoped
-+ memset(&fmGetSetParams, 0, sizeof (t_FmGetSetParams));
-+ fmGetSetParams.getParams.type = GET_FMQM_GS | GET_FM_NPI;
-+ FmGetSetParams(p_FmPort->h_Fm, &fmGetSetParams);
-+ while (fmGetSetParams.getParams.fmqm_gs & 0xF0000000)
-+ FmGetSetParams(p_FmPort->h_Fm, &fmGetSetParams);
-+ if (fmGetSetParams.getParams.fmqm_gs == 0 && fmGetSetParams.getParams.fm_npi == 0)
-+ XX_Print("FM: Sleeping\n");
-+// FM_ChangeClock(p_FmPort->h_Fm, p_FmPort->hardwarePortId);
-+
-+ return E_OK;
-+}
-+
-+EXPORT_SYMBOL(FM_PORT_EnterDsarFinal);
-+
-+void FM_PORT_Dsar_DumpRegs()
-+{
-+ uint32_t* hh = XX_PhysToVirt(PTR_TO_UINT(ARDesc));
-+ DUMP_MEMORY(hh, 0x220);
-+}
-+
-+void FM_PORT_ExitDsar(t_Handle h_FmPortRx, t_Handle h_FmPortTx)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort *)h_FmPortRx;
-+ t_FmPort *p_FmPortTx = (t_FmPort *)h_FmPortTx;
-+ t_Handle *h_FmPcd = FmGetPcd(p_FmPort->h_Fm);
-+ t_FmPort *p_FmPortHc = FM_PCD_GetHcPort(h_FmPcd);
-+ t_FmGetSetParams fmGetSetParams;
-+ memset(&fmGetSetParams, 0, sizeof (t_FmGetSetParams));
-+ fmGetSetParams.setParams.type = UPDATE_FPM_BRKC_SLP;
-+ fmGetSetParams.setParams.sleep = 0;
-+ if (p_FmPort->deepSleepVars.autoResOffsets)
-+ {
-+ XX_Free(p_FmPort->deepSleepVars.autoResOffsets);
-+ p_FmPort->deepSleepVars.autoResOffsets = 0;
-+ }
-+
-+ if (p_FmPort->deepSleepVars.dsarEnabledParser)
-+ PrsDisable(FmGetPcd(p_FmPort->h_Fm));
-+ WRITE_UINT32(p_FmPort->p_FmPortBmiRegs->rxPortBmiRegs.fmbm_rfpne, p_FmPort->deepSleepVars.fmbm_rfpne);
-+ WRITE_UINT32(p_FmPort->p_FmPortBmiRegs->rxPortBmiRegs.fmbm_rfne, p_FmPort->deepSleepVars.fmbm_rfne);
-+ WRITE_UINT32(p_FmPort->p_FmPortBmiRegs->rxPortBmiRegs.fmbm_rcfg, p_FmPort->deepSleepVars.fmbm_rcfg);
-+ FmGetSetParams(p_FmPort->h_Fm, &fmGetSetParams);
-+ WRITE_UINT32(p_FmPortTx->p_FmPortBmiRegs->txPortBmiRegs.fmbm_tcmne, p_FmPort->deepSleepVars.fmbm_tcmne);
-+ WRITE_UINT32(p_FmPortTx->p_FmPortBmiRegs->txPortBmiRegs.fmbm_tcfg, p_FmPort->deepSleepVars.fmbm_tcfg);
-+ FM_PORT_Enable(p_FmPortHc);
-+}
-+
-+bool FM_PORT_IsInDsar(t_Handle h_FmPort)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort *)h_FmPort;
-+ return PTR_TO_UINT(p_FmPort->deepSleepVars.autoResOffsets);
-+}
-+
-+t_Error FM_PORT_GetDsarStats(t_Handle h_FmPortRx, t_FmPortDsarStats *stats)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort *)h_FmPortRx;
-+ struct arOffsets *of = (struct arOffsets*)p_FmPort->deepSleepVars.autoResOffsets;
-+ uint8_t* fmMuramVirtBaseAddr = XX_PhysToVirt(p_FmPort->fmMuramPhysBaseAddr);
-+ uint32_t *param_page = XX_PhysToVirt(p_FmPort->fmMuramPhysBaseAddr + GET_UINT32(p_FmPort->p_FmPortBmiRegs->rxPortBmiRegs.fmbm_rgpr));
-+ t_ArCommonDesc *ArCommonDescPtr = (t_ArCommonDesc*)(XX_PhysToVirt(p_FmPort->fmMuramPhysBaseAddr + GET_UINT32(*param_page)));
-+ t_DsarArpDescriptor *ArpDescriptor = (t_DsarArpDescriptor*)(PTR_TO_UINT(ArCommonDescPtr) + of->arp);
-+ t_DsarArpStatistics* arp_stats = (t_DsarArpStatistics*)(PTR_TO_UINT(ArpDescriptor->p_Statistics) + fmMuramVirtBaseAddr);
-+ t_DsarIcmpV4Descriptor* ICMPV4Descriptor = (t_DsarIcmpV4Descriptor*)(PTR_TO_UINT(ArCommonDescPtr) + of->icmpv4);
-+ t_DsarIcmpV4Statistics* icmpv4_stats = (t_DsarIcmpV4Statistics*)(PTR_TO_UINT(ICMPV4Descriptor->p_Statistics) + fmMuramVirtBaseAddr);
-+ t_DsarNdDescriptor* NDDescriptor = (t_DsarNdDescriptor*)(PTR_TO_UINT(ArCommonDescPtr) + of->nd);
-+ t_NdStatistics* nd_stats = (t_NdStatistics*)(PTR_TO_UINT(NDDescriptor->p_Statistics) + fmMuramVirtBaseAddr);
-+ t_DsarIcmpV6Descriptor* ICMPV6Descriptor = (t_DsarIcmpV6Descriptor*)(PTR_TO_UINT(ArCommonDescPtr) + of->icmpv6);
-+ t_DsarIcmpV6Statistics* icmpv6_stats = (t_DsarIcmpV6Statistics*)(PTR_TO_UINT(ICMPV6Descriptor->p_Statistics) + fmMuramVirtBaseAddr);
-+ t_DsarSnmpDescriptor* SnmpDescriptor = (t_DsarSnmpDescriptor*)(PTR_TO_UINT(ArCommonDescPtr) + of->snmp);
-+ t_DsarSnmpStatistics* snmp_stats = (t_DsarSnmpStatistics*)(PTR_TO_UINT(SnmpDescriptor->p_Statistics) + fmMuramVirtBaseAddr);
-+ stats->arpArCnt = arp_stats->arCnt;
-+ stats->echoIcmpv4ArCnt = icmpv4_stats->arCnt;
-+ stats->ndpArCnt = nd_stats->arCnt;
-+ stats->echoIcmpv6ArCnt = icmpv6_stats->arCnt;
-+ stats->snmpGetCnt = snmp_stats->snmpGetReqCnt;
-+ stats->snmpGetNextCnt = snmp_stats->snmpGetNextReqCnt;
-+ return E_OK;
-+}
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/Port/fm_port.h
-@@ -0,0 +1,999 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/******************************************************************************
-+ @File fm_port.h
-+
-+ @Description FM Port internal structures and definitions.
-+*//***************************************************************************/
-+#ifndef __FM_PORT_H
-+#define __FM_PORT_H
-+
-+#include "error_ext.h"
-+#include "std_ext.h"
-+#include "fm_port_ext.h"
-+
-+#include "fm_common.h"
-+#include "fm_sp_common.h"
-+#include "fsl_fman_sp.h"
-+#include "fm_port_ext.h"
-+#include "fsl_fman_port.h"
-+
-+#define __ERR_MODULE__ MODULE_FM_PORT
-+
-+
-+#define MIN_EXT_BUF_SIZE 64
-+#define DATA_ALIGNMENT 64
-+#define MAX_LIODN_OFFSET 64
-+#define MAX_PORT_FIFO_SIZE MIN(BMI_MAX_FIFO_SIZE, 1024*BMI_FIFO_UNITS)
-+
-+/**************************************************************************//**
-+ @Description Memory Map defines
-+*//***************************************************************************/
-+#define BMI_PORT_REGS_OFFSET 0
-+#define QMI_PORT_REGS_OFFSET 0x400
-+#define PRS_PORT_REGS_OFFSET 0x800
-+
-+/**************************************************************************//**
-+ @Description defaults
-+*//***************************************************************************/
-+#define DEFAULT_PORT_deqHighPriority_1G FALSE
-+#define DEFAULT_PORT_deqHighPriority_10G TRUE
-+#define DEFAULT_PORT_deqType e_FM_PORT_DEQ_TYPE1
-+#define DEFAULT_PORT_deqPrefetchOption e_FM_PORT_DEQ_FULL_PREFETCH
-+#define DEFAULT_PORT_deqPrefetchOption_HC e_FM_PORT_DEQ_NO_PREFETCH
-+#define DEFAULT_PORT_deqByteCnt_10G 0x1400
-+#define DEFAULT_PORT_deqByteCnt_1G 0x400
-+#define DEFAULT_PORT_bufferPrefixContent_privDataSize DEFAULT_FM_SP_bufferPrefixContent_privDataSize
-+#define DEFAULT_PORT_bufferPrefixContent_passPrsResult DEFAULT_FM_SP_bufferPrefixContent_passPrsResult
-+#define DEFAULT_PORT_bufferPrefixContent_passTimeStamp DEFAULT_FM_SP_bufferPrefixContent_passTimeStamp
-+#define DEFAULT_PORT_bufferPrefixContent_allOtherPCDInfo DEFAULT_FM_SP_bufferPrefixContent_allOtherPCDInfo
-+#define DEFAULT_PORT_bufferPrefixContent_dataAlign DEFAULT_FM_SP_bufferPrefixContent_dataAlign
-+#define DEFAULT_PORT_cheksumLastBytesIgnore 0
-+#define DEFAULT_PORT_cutBytesFromEnd 4
-+#define DEFAULT_PORT_fifoDeqPipelineDepth_IM 2
-+
-+#define DEFAULT_PORT_frmDiscardOverride FALSE
-+
-+#define DEFAULT_PORT_dmaSwapData (e_FmDmaSwapOption)DEFAULT_FMAN_SP_DMA_SWAP_DATA
-+#define DEFAULT_PORT_dmaIntContextCacheAttr (e_FmDmaCacheOption)DEFAULT_FMAN_SP_DMA_INT_CONTEXT_CACHE_ATTR
-+#define DEFAULT_PORT_dmaHeaderCacheAttr (e_FmDmaCacheOption)DEFAULT_FMAN_SP_DMA_HEADER_CACHE_ATTR
-+#define DEFAULT_PORT_dmaScatterGatherCacheAttr (e_FmDmaCacheOption)DEFAULT_FMAN_SP_DMA_SCATTER_GATHER_CACHE_ATTR
-+#define DEFAULT_PORT_dmaWriteOptimize DEFAULT_FMAN_SP_DMA_WRITE_OPTIMIZE
-+
-+#define DEFAULT_PORT_noScatherGather DEFAULT_FMAN_SP_NO_SCATTER_GATHER
-+#define DEFAULT_PORT_forwardIntContextReuse FALSE
-+#define DEFAULT_PORT_BufMargins_startMargins 32
-+#define DEFAULT_PORT_BufMargins_endMargins 0
-+#define DEFAULT_PORT_syncReq TRUE
-+#define DEFAULT_PORT_syncReqForHc FALSE
-+#define DEFAULT_PORT_color e_FM_PORT_COLOR_GREEN
-+#define DEFAULT_PORT_errorsToDiscard FM_PORT_FRM_ERR_CLS_DISCARD
-+/* #define DEFAULT_PORT_dualRateLimitScaleDown e_FM_PORT_DUAL_RATE_LIMITER_NONE */
-+/* #define DEFAULT_PORT_rateLimitBurstSizeHighGranularity FALSE */
-+#define DEFAULT_PORT_exception IM_EV_BSY
-+#define DEFAULT_PORT_maxFrameLength 9600
-+
-+#define DEFAULT_notSupported 0xff
-+
-+#if (DPAA_VERSION < 11)
-+#define DEFAULT_PORT_rxFifoPriElevationLevel MAX_PORT_FIFO_SIZE
-+#define DEFAULT_PORT_rxFifoThreshold (MAX_PORT_FIFO_SIZE*3/4)
-+
-+#define DEFAULT_PORT_txFifoMinFillLevel 0
-+#define DEFAULT_PORT_txFifoLowComfLevel (5*KILOBYTE)
-+#define DEFAULT_PORT_fifoDeqPipelineDepth_1G 1
-+#define DEFAULT_PORT_fifoDeqPipelineDepth_10G 4
-+
-+#define DEFAULT_PORT_fifoDeqPipelineDepth_OH 2
-+
-+/* Host command port MUST NOT be changed to more than 1 !!! */
-+#define DEFAULT_PORT_numOfTasks(type) \
-+ (uint32_t)((((type) == e_FM_PORT_TYPE_RX_10G) || \
-+ ((type) == e_FM_PORT_TYPE_TX_10G)) ? 16 : \
-+ ((((type) == e_FM_PORT_TYPE_RX) || \
-+ ((type) == e_FM_PORT_TYPE_TX) || \
-+ ((type) == e_FM_PORT_TYPE_OH_OFFLINE_PARSING)) ? 3 : 1))
-+
-+#define DEFAULT_PORT_extraNumOfTasks(type) \
-+ (uint32_t)(((type) == e_FM_PORT_TYPE_RX_10G) ? 8 : \
-+ (((type) == e_FM_PORT_TYPE_RX) ? 2 : 0))
-+
-+#define DEFAULT_PORT_numOfOpenDmas(type) \
-+ (uint32_t)((((type) == e_FM_PORT_TYPE_TX_10G) || \
-+ ((type) == e_FM_PORT_TYPE_RX_10G)) ? 8 : 1 )
-+
-+#define DEFAULT_PORT_extraNumOfOpenDmas(type) \
-+ (uint32_t)(((type) == e_FM_PORT_TYPE_RX_10G) ? 8 : \
-+ (((type) == e_FM_PORT_TYPE_RX) ? 1 : 0))
-+
-+#define DEFAULT_PORT_numOfFifoBufs(type) \
-+ (uint32_t)((((type) == e_FM_PORT_TYPE_RX_10G) || \
-+ ((type) == e_FM_PORT_TYPE_TX_10G)) ? 48 : \
-+ ((type) == e_FM_PORT_TYPE_RX) ? 45 : \
-+ ((type) == e_FM_PORT_TYPE_TX) ? 44 : 8)
-+
-+#define DEFAULT_PORT_extraNumOfFifoBufs 0
-+
-+#else /* (DPAA_VERSION < 11) */
-+/* Defaults are registers' reset values */
-+#define DEFAULT_PORT_rxFifoPriElevationLevel MAX_PORT_FIFO_SIZE
-+#define DEFAULT_PORT_rxFifoThreshold MAX_PORT_FIFO_SIZE
-+
-+#define DEFAULT_PORT_txFifoMinFillLevel 0
-+#define DEFAULT_PORT_txFifoLowComfLevel (5 * KILOBYTE)
-+#define DEFAULT_PORT_fifoDeqPipelineDepth_1G 2
-+#define DEFAULT_PORT_fifoDeqPipelineDepth_10G 4
-+
-+#define DEFAULT_PORT_fifoDeqPipelineDepth_OH 2
-+
-+#define DEFAULT_PORT_numOfTasks(type) \
-+ (uint32_t)((((type) == e_FM_PORT_TYPE_RX_10G) || \
-+ ((type) == e_FM_PORT_TYPE_TX_10G)) ? 14 : \
-+ (((type) == e_FM_PORT_TYPE_RX) || \
-+ ((type) == e_FM_PORT_TYPE_TX)) ? 4 : \
-+ ((type) == e_FM_PORT_TYPE_OH_OFFLINE_PARSING) ? 6 : 1)
-+
-+#define DEFAULT_PORT_extraNumOfTasks(type) 0
-+
-+#define DEFAULT_PORT_numOfOpenDmas(type) \
-+ (uint32_t)(((type) == e_FM_PORT_TYPE_RX_10G) ? 8 : \
-+ ((type) == e_FM_PORT_TYPE_TX_10G) ? 12 : \
-+ ((type) == e_FM_PORT_TYPE_RX) ? 2 : \
-+ ((type) == e_FM_PORT_TYPE_TX) ? 3 : \
-+ ((type) == e_FM_PORT_TYPE_OH_HOST_COMMAND) ? 2 : 4)
-+
-+#define DEFAULT_PORT_extraNumOfOpenDmas(type) 0
-+
-+#define DEFAULT_PORT_numOfFifoBufs(type) \
-+ (uint32_t) (((type) == e_FM_PORT_TYPE_RX_10G) ? 96 : \
-+ ((type) == e_FM_PORT_TYPE_TX_10G) ? 64 : \
-+ ((type) == e_FM_PORT_TYPE_OH_HOST_COMMAND) ? 10 : 50)
-+
-+#define DEFAULT_PORT_extraNumOfFifoBufs 0
-+
-+#endif /* (DPAA_VERSION < 11) */
-+
-+#define DEFAULT_PORT_txBdRingLength 16
-+#define DEFAULT_PORT_rxBdRingLength 128
-+#define DEFAULT_PORT_ImfwExtStructsMemId 0
-+#define DEFAULT_PORT_ImfwExtStructsMemAttr MEMORY_ATTR_CACHEABLE
-+
-+#define FM_PORT_CG_REG_NUM(_cgId) (((FM_PORT_NUM_OF_CONGESTION_GRPS/32)-1)-_cgId/32)
-+
-+/**************************************************************************//**
-+ @Collection PCD Engines
-+*//***************************************************************************/
-+typedef uint32_t fmPcdEngines_t; /**< options as defined below: */
-+
-+#define FM_PCD_NONE 0 /**< No PCD Engine indicated */
-+#define FM_PCD_PRS 0x80000000 /**< Parser indicated */
-+#define FM_PCD_KG 0x40000000 /**< Keygen indicated */
-+#define FM_PCD_CC 0x20000000 /**< Coarse classification indicated */
-+#define FM_PCD_PLCR 0x10000000 /**< Policer indicated */
-+#define FM_PCD_MANIP 0x08000000 /**< Manipulation indicated */
-+/* @} */
-+
-+#define FM_PORT_MAX_NUM_OF_EXT_POOLS_ALL_INTEGRATIONS 8
-+#define FM_PORT_MAX_NUM_OF_CONGESTION_GRPS_ALL_INTEGRATIONS 256
-+#define FM_PORT_CG_REG_NUM(_cgId) (((FM_PORT_NUM_OF_CONGESTION_GRPS/32)-1)-_cgId/32)
-+
-+#define FM_OH_PORT_ID 0
-+
-+/***********************************************************************/
-+/* SW parser OFFLOAD labels (offsets) */
-+/***********************************************************************/
-+#if (DPAA_VERSION == 10)
-+#define OFFLOAD_SW_PATCH_IPv4_IPR_LABEL 0x300
-+#define OFFLOAD_SW_PATCH_IPv6_IPR_LABEL 0x325
-+#define OFFLOAD_SW_PATCH_IPv6_IPF_LABEL 0x325
-+#else
-+#define OFFLOAD_SW_PATCH_IPv4_IPR_LABEL 0x100
-+/* Will be used for:
-+ * 1. identify fragments
-+ * 2. udp-lite
-+ */
-+#define OFFLOAD_SW_PATCH_IPv6_IPR_LABEL 0x146
-+/* Will be used for:
-+ * 1. will identify the fragmentable area
-+ * 2. udp-lite
-+ */
-+#define OFFLOAD_SW_PATCH_IPv6_IPF_LABEL 0x261
-+#define OFFLOAD_SW_PATCH_CAPWAP_LABEL 0x38d
-+#endif /* (DPAA_VERSION == 10) */
-+
-+#if ((DPAA_VERSION == 10) && defined(FM_CAPWAP_SUPPORT))
-+#define UDP_LITE_SW_PATCH_LABEL 0x2E0
-+#endif /* ((DPAA_VERSION == 10) && defined(FM_CAPWAP_SUPPORT)) */
-+
-+
-+/**************************************************************************//**
-+ @Description Memory Mapped Registers
-+*//***************************************************************************/
-+
-+#if defined(__MWERKS__) && !defined(__GNUC__)
-+#pragma pack(push,1)
-+#endif /* defined(__MWERKS__) && ... */
-+
-+typedef struct
-+{
-+ volatile uint32_t fmbm_rcfg; /**< Rx Configuration */
-+ volatile uint32_t fmbm_rst; /**< Rx Status */
-+ volatile uint32_t fmbm_rda; /**< Rx DMA attributes*/
-+ volatile uint32_t fmbm_rfp; /**< Rx FIFO Parameters*/
-+ volatile uint32_t fmbm_rfed; /**< Rx Frame End Data*/
-+ volatile uint32_t fmbm_ricp; /**< Rx Internal Context Parameters*/
-+ volatile uint32_t fmbm_rim; /**< Rx Internal Buffer Margins*/
-+ volatile uint32_t fmbm_rebm; /**< Rx External Buffer Margins*/
-+ volatile uint32_t fmbm_rfne; /**< Rx Frame Next Engine*/
-+ volatile uint32_t fmbm_rfca; /**< Rx Frame Command Attributes.*/
-+ volatile uint32_t fmbm_rfpne; /**< Rx Frame Parser Next Engine*/
-+ volatile uint32_t fmbm_rpso; /**< Rx Parse Start Offset*/
-+ volatile uint32_t fmbm_rpp; /**< Rx Policer Profile */
-+ volatile uint32_t fmbm_rccb; /**< Rx Coarse Classification Base */
-+ volatile uint32_t fmbm_reth; /**< Rx Excessive Threshold */
-+ volatile uint32_t reserved1[0x01];/**< (0x03C) */
-+ volatile uint32_t fmbm_rprai[FM_PORT_PRS_RESULT_NUM_OF_WORDS];
-+ /**< Rx Parse Results Array Initialization*/
-+ volatile uint32_t fmbm_rfqid; /**< Rx Frame Queue ID*/
-+ volatile uint32_t fmbm_refqid; /**< Rx Error Frame Queue ID*/
-+ volatile uint32_t fmbm_rfsdm; /**< Rx Frame Status Discard Mask*/
-+ volatile uint32_t fmbm_rfsem; /**< Rx Frame Status Error Mask*/
-+ volatile uint32_t fmbm_rfene; /**< Rx Frame Enqueue Next Engine */
-+ volatile uint32_t reserved2[0x02];/**< (0x074-0x078) */
-+ volatile uint32_t fmbm_rcmne; /**< Rx Frame Continuous Mode Next Engine */
-+ volatile uint32_t reserved3[0x20];/**< (0x080 0x0FF) */
-+ volatile uint32_t fmbm_ebmpi[FM_PORT_MAX_NUM_OF_EXT_POOLS_ALL_INTEGRATIONS];
-+ /**< Buffer Manager pool Information-*/
-+ volatile uint32_t fmbm_acnt[FM_PORT_MAX_NUM_OF_EXT_POOLS_ALL_INTEGRATIONS];
-+ /**< Allocate Counter-*/
-+ volatile uint32_t reserved4[0x08];
-+ /**< 0x130/0x140 - 0x15F reserved -*/
-+ volatile uint32_t fmbm_rcgm[FM_PORT_MAX_NUM_OF_CONGESTION_GRPS_ALL_INTEGRATIONS/32];
-+ /**< Congestion Group Map*/
-+ volatile uint32_t fmbm_rmpd; /**< BM Pool Depletion */
-+ volatile uint32_t reserved5[0x1F];/**< (0x184 0x1FF) */
-+ volatile uint32_t fmbm_rstc; /**< Rx Statistics Counters*/
-+ volatile uint32_t fmbm_rfrc; /**< Rx Frame Counter*/
-+ volatile uint32_t fmbm_rfbc; /**< Rx Bad Frames Counter*/
-+ volatile uint32_t fmbm_rlfc; /**< Rx Large Frames Counter*/
-+ volatile uint32_t fmbm_rffc; /**< Rx Filter Frames Counter*/
-+ volatile uint32_t fmbm_rfcd; /**< Rx Frame Discard Counter*/
-+ volatile uint32_t fmbm_rfldec; /**< Rx Frames List DMA Error Counter*/
-+ volatile uint32_t fmbm_rodc; /**< Rx Out of Buffers Discard Counter-*/
-+ volatile uint32_t fmbm_rbdc; /**< Rx Buffers Deallocate Counter-*/
-+ volatile uint32_t fmbm_rpec; /**< Rx RX Prepare to enqueue Counter-*/
-+ volatile uint32_t reserved6[0x16];/**< (0x228 0x27F) */
-+ volatile uint32_t fmbm_rpc; /**< Rx Performance Counters*/
-+ volatile uint32_t fmbm_rpcp; /**< Rx Performance Count Parameters*/
-+ volatile uint32_t fmbm_rccn; /**< Rx Cycle Counter*/
-+ volatile uint32_t fmbm_rtuc; /**< Rx Tasks Utilization Counter*/
-+ volatile uint32_t fmbm_rrquc; /**< Rx Receive Queue Utilization Counter*/
-+ volatile uint32_t fmbm_rduc; /**< Rx DMA Utilization Counter*/
-+ volatile uint32_t fmbm_rfuc; /**< Rx FIFO Utilization Counter*/
-+ volatile uint32_t fmbm_rpac; /**< Rx Pause Activation Counter*/
-+ volatile uint32_t reserved7[0x18];/**< (0x2A0-0x2FF) */
-+ volatile uint32_t fmbm_rdcfg[0x3];/**< Rx Debug-*/
-+ volatile uint32_t fmbm_rgpr; /**< Rx General Purpose Register. */
-+ volatile uint32_t reserved8[0x3a];/**< (0x310-0x3FF) */
-+} t_FmPortRxBmiRegs;
-+
-+typedef struct
-+{
-+ volatile uint32_t fmbm_tcfg; /**< Tx Configuration */
-+ volatile uint32_t fmbm_tst; /**< Tx Status */
-+ volatile uint32_t fmbm_tda; /**< Tx DMA attributes */
-+ volatile uint32_t fmbm_tfp; /**< Tx FIFO Parameters */
-+ volatile uint32_t fmbm_tfed; /**< Tx Frame End Data */
-+ volatile uint32_t fmbm_ticp; /**< Tx Internal Context Parameters */
-+ volatile uint32_t fmbm_tfdne; /**< Tx Frame Dequeue Next Engine. */
-+ volatile uint32_t fmbm_tfca; /**< Tx Frame Command attribute. */
-+ volatile uint32_t fmbm_tcfqid; /**< Tx Confirmation Frame Queue ID. */
-+ volatile uint32_t fmbm_tfeqid; /**< Tx Frame Error Queue ID */
-+ volatile uint32_t fmbm_tfene; /**< Tx Frame Enqueue Next Engine */
-+ volatile uint32_t fmbm_trlmts; /**< Tx Rate Limiter Scale */
-+ volatile uint32_t fmbm_trlmt; /**< Tx Rate Limiter */
-+ volatile uint32_t fmbm_tccb; /**< Tx Coarse Classification Base */
-+ volatile uint32_t reserved0[0x0e];/**< (0x038-0x070) */
-+ volatile uint32_t fmbm_tfne; /**< Tx Frame Next Engine */
-+ volatile uint32_t fmbm_tpfcm[0x02];/**< Tx Priority based Flow Control (PFC) Mapping */
-+ volatile uint32_t fmbm_tcmne; /**< Tx Frame Continuous Mode Next Engine */
-+ volatile uint32_t reserved2[0x60];/**< (0x080-0x200) */
-+ volatile uint32_t fmbm_tstc; /**< Tx Statistics Counters */
-+ volatile uint32_t fmbm_tfrc; /**< Tx Frame Counter */
-+ volatile uint32_t fmbm_tfdc; /**< Tx Frames Discard Counter */
-+ volatile uint32_t fmbm_tfledc; /**< Tx Frame Length error discard counter */
-+ volatile uint32_t fmbm_tfufdc; /**< Tx Frame unsupported format discard Counter */
-+ volatile uint32_t fmbm_tbdc; /**< Tx Buffers Deallocate Counter */
-+ volatile uint32_t reserved3[0x1A];/**< (0x218-0x280) */
-+ volatile uint32_t fmbm_tpc; /**< Tx Performance Counters*/
-+ volatile uint32_t fmbm_tpcp; /**< Tx Performance Count Parameters*/
-+ volatile uint32_t fmbm_tccn; /**< Tx Cycle Counter*/
-+ volatile uint32_t fmbm_ttuc; /**< Tx Tasks Utilization Counter*/
-+ volatile uint32_t fmbm_ttcquc; /**< Tx Transmit Confirm Queue Utilization Counter*/
-+ volatile uint32_t fmbm_tduc; /**< Tx DMA Utilization Counter*/
-+ volatile uint32_t fmbm_tfuc; /**< Tx FIFO Utilization Counter*/
-+ volatile uint32_t reserved4[16]; /**< (0x29C-0x2FF) */
-+ volatile uint32_t fmbm_tdcfg[0x3];/**< Tx Debug-*/
-+ volatile uint32_t fmbm_tgpr; /**< O/H General Purpose Register */
-+ volatile uint32_t reserved5[0x3a];/**< (0x310-0x3FF) */
-+} t_FmPortTxBmiRegs;
-+
-+typedef struct
-+{
-+ volatile uint32_t fmbm_ocfg; /**< O/H Configuration */
-+ volatile uint32_t fmbm_ost; /**< O/H Status */
-+ volatile uint32_t fmbm_oda; /**< O/H DMA attributes */
-+ volatile uint32_t fmbm_oicp; /**< O/H Internal Context Parameters */
-+ volatile uint32_t fmbm_ofdne; /**< O/H Frame Dequeue Next Engine */
-+ volatile uint32_t fmbm_ofne; /**< O/H Frame Next Engine */
-+ volatile uint32_t fmbm_ofca; /**< O/H Frame Command Attributes. */
-+ volatile uint32_t fmbm_ofpne; /**< O/H Frame Parser Next Engine */
-+ volatile uint32_t fmbm_opso; /**< O/H Parse Start Offset */
-+ volatile uint32_t fmbm_opp; /**< O/H Policer Profile */
-+ volatile uint32_t fmbm_occb; /**< O/H Coarse Classification base */
-+ volatile uint32_t fmbm_oim; /**< O/H Internal margins*/
-+ volatile uint32_t fmbm_ofp; /**< O/H Fifo Parameters*/
-+ volatile uint32_t fmbm_ofed; /**< O/H Frame End Data*/
-+ volatile uint32_t reserved0[2]; /**< (0x038 - 0x03F) */
-+ volatile uint32_t fmbm_oprai[FM_PORT_PRS_RESULT_NUM_OF_WORDS];
-+ /**< O/H Parse Results Array Initialization */
-+ volatile uint32_t fmbm_ofqid; /**< O/H Frame Queue ID */
-+ volatile uint32_t fmbm_oefqid; /**< O/H Error Frame Queue ID */
-+ volatile uint32_t fmbm_ofsdm; /**< O/H Frame Status Discard Mask */
-+ volatile uint32_t fmbm_ofsem; /**< O/H Frame Status Error Mask */
-+ volatile uint32_t fmbm_ofene; /**< O/H Frame Enqueue Next Engine */
-+ volatile uint32_t fmbm_orlmts; /**< O/H Rate Limiter Scale */
-+ volatile uint32_t fmbm_orlmt; /**< O/H Rate Limiter */
-+ volatile uint32_t fmbm_ocmne; /**< O/H Continuous Mode Next Engine */
-+ volatile uint32_t reserved1[0x20];/**< (0x080 - 0x0FF) */
-+ volatile uint32_t fmbm_oebmpi[2]; /**< Buffer Manager Observed Pool Information */
-+ volatile uint32_t reserved2[0x16];/**< (0x108 - 0x15F) */
-+ volatile uint32_t fmbm_ocgm; /**< Observed Congestion Group Map */
-+ volatile uint32_t reserved3[0x7]; /**< (0x164 - 0x17F) */
-+ volatile uint32_t fmbm_ompd; /**< Observed BMan Pool Depletion */
-+ volatile uint32_t reserved4[0x1F];/**< (0x184 - 0x1FF) */
-+ volatile uint32_t fmbm_ostc; /**< O/H Statistics Counters */
-+ volatile uint32_t fmbm_ofrc; /**< O/H Frame Counter */
-+ volatile uint32_t fmbm_ofdc; /**< O/H Frames Discard Counter */
-+ volatile uint32_t fmbm_ofledc; /**< O/H Frames Length Error Discard Counter */
-+ volatile uint32_t fmbm_ofufdc; /**< O/H Frames Unsupported Format Discard Counter */
-+ volatile uint32_t fmbm_offc; /**< O/H Filter Frames Counter */
-+ volatile uint32_t fmbm_ofwdc; /**< - Rx Frames WRED Discard Counter */
-+ volatile uint32_t fmbm_ofldec; /**< O/H Frames List DMA Error Counter */
-+ volatile uint32_t fmbm_obdc; /**< O/H Buffers Deallocate Counter */
-+ volatile uint32_t fmbm_oodc; /**< O/H Out of Buffers Discard Counter */
-+ volatile uint32_t fmbm_opec; /**< O/H Prepare to enqueue Counter */
-+ volatile uint32_t reserved5[0x15];/**< ( - 0x27F) */
-+ volatile uint32_t fmbm_opc; /**< O/H Performance Counters */
-+ volatile uint32_t fmbm_opcp; /**< O/H Performance Count Parameters */
-+ volatile uint32_t fmbm_occn; /**< O/H Cycle Counter */
-+ volatile uint32_t fmbm_otuc; /**< O/H Tasks Utilization Counter */
-+ volatile uint32_t fmbm_oduc; /**< O/H DMA Utilization Counter */
-+ volatile uint32_t fmbm_ofuc; /**< O/H FIFO Utilization Counter */
-+ volatile uint32_t reserved6[26]; /**< (0x298-0x2FF) */
-+ volatile uint32_t fmbm_odcfg[0x3];/**< O/H Debug (only 1 in P1023) */
-+ volatile uint32_t fmbm_ogpr; /**< O/H General Purpose Register. */
-+ volatile uint32_t reserved7[0x3a];/**< (0x310 0x3FF) */
-+} t_FmPortOhBmiRegs;
-+
-+typedef union
-+{
-+ t_FmPortRxBmiRegs rxPortBmiRegs;
-+ t_FmPortTxBmiRegs txPortBmiRegs;
-+ t_FmPortOhBmiRegs ohPortBmiRegs;
-+} u_FmPortBmiRegs;
-+
-+typedef struct
-+{
-+ volatile uint32_t reserved1[2]; /**< 0xn024 - 0x02B */
-+ volatile uint32_t fmqm_pndn; /**< PortID n Dequeue NIA Register */
-+ volatile uint32_t fmqm_pndc; /**< PortID n Dequeue Config Register */
-+ volatile uint32_t fmqm_pndtfc; /**< PortID n Dequeue Total Frame Counter */
-+ volatile uint32_t fmqm_pndfdc; /**< PortID n Dequeue FQID from Default Counter */
-+ volatile uint32_t fmqm_pndcc; /**< PortID n Dequeue Confirm Counter */
-+} t_FmPortNonRxQmiRegs;
-+
-+typedef struct
-+{
-+ volatile uint32_t fmqm_pnc; /**< PortID n Configuration Register */
-+ volatile uint32_t fmqm_pns; /**< PortID n Status Register */
-+ volatile uint32_t fmqm_pnts; /**< PortID n Task Status Register */
-+ volatile uint32_t reserved0[4]; /**< 0xn00C - 0xn01B */
-+ volatile uint32_t fmqm_pnen; /**< PortID n Enqueue NIA Register */
-+ volatile uint32_t fmqm_pnetfc; /**< PortID n Enqueue Total Frame Counter */
-+ t_FmPortNonRxQmiRegs nonRxQmiRegs; /**< Registers for Tx Hc & Op ports */
-+} t_FmPortQmiRegs;
-+
-+typedef struct
-+{
-+ struct
-+ {
-+ volatile uint32_t softSeqAttach; /**< Soft Sequence Attachment */
-+ volatile uint32_t lcv; /**< Line-up Enable Confirmation Mask */
-+ } hdrs[FM_PCD_PRS_NUM_OF_HDRS];
-+ volatile uint32_t reserved0[0xde];
-+ volatile uint32_t pcac; /**< Parse Internal Memory Configuration Access Control Register */
-+ volatile uint32_t pctpid; /**< Parse Internal Memory Configured TPID Register */
-+} t_FmPortPrsRegs;
-+
-+/**************************************************************************//*
-+ @Description Basic buffer descriptor (BD) structure
-+*//***************************************************************************/
-+typedef _Packed struct
-+{
-+ volatile uint16_t status;
-+ volatile uint16_t length;
-+ volatile uint8_t reserved0[0x6];
-+ volatile uint8_t reserved1[0x1];
-+ volatile t_FmPhysAddr buff;
-+} _PackedType t_FmImBd;
-+
-+typedef _Packed struct
-+{
-+ volatile uint16_t gen; /**< tbd */
-+ volatile uint8_t reserved0[0x1];
-+ volatile t_FmPhysAddr bdRingBase; /**< tbd */
-+ volatile uint16_t bdRingSize; /**< tbd */
-+ volatile uint16_t offsetIn; /**< tbd */
-+ volatile uint16_t offsetOut; /**< tbd */
-+ volatile uint8_t reserved1[0x12]; /**< 0x0e - 0x1f */
-+} _PackedType t_FmPortImQd;
-+
-+typedef _Packed struct
-+{
-+ volatile uint32_t mode; /**< Mode register */
-+ volatile uint32_t rxQdPtr; /**< tbd */
-+ volatile uint32_t txQdPtr; /**< tbd */
-+ volatile uint16_t mrblr; /**< tbd */
-+ volatile uint16_t rxQdBsyCnt; /**< tbd */
-+ volatile uint8_t reserved0[0x10]; /**< 0x10 - 0x1f */
-+ t_FmPortImQd rxQd;
-+ t_FmPortImQd txQd;
-+ volatile uint8_t reserved1[0xa0]; /**< 0x60 - 0xff */
-+} _PackedType t_FmPortImPram;
-+
-+#if defined(__MWERKS__) && !defined(__GNUC__)
-+#pragma pack(pop)
-+#endif /* defined(__MWERKS__) && ... */
-+
-+
-+/**************************************************************************//**
-+ @Description Registers bit fields
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Description BMI defines
-+*//***************************************************************************/
-+#if (DPAA_VERSION >= 11)
-+#define BMI_SP_ID_MASK 0xff000000
-+#define BMI_SP_ID_SHIFT 24
-+#define BMI_SP_EN 0x01000000
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+#define BMI_PORT_CFG_EN 0x80000000
-+#define BMI_PORT_CFG_EN_MACSEC 0x00800000
-+#define BMI_PORT_CFG_FDOVR 0x02000000
-+#define BMI_PORT_CFG_IM 0x01000000
-+#define BMI_PORT_CFG_AM 0x00000040
-+#define BMI_PORT_STATUS_BSY 0x80000000
-+#define BMI_COUNTERS_EN 0x80000000
-+
-+#define BMI_PORT_RFNE_FRWD_DCL4C 0x10000000
-+#define BMI_PORT_RFNE_FRWD_RPD 0x40000000
-+#define BMI_RFNE_FDCS_MASK 0xFF000000
-+#define BMI_RFNE_HXS_MASK 0x000000FF
-+
-+#define BMI_CMD_MR_LEAC 0x00200000
-+#define BMI_CMD_MR_SLEAC 0x00100000
-+#define BMI_CMD_MR_MA 0x00080000
-+#define BMI_CMD_MR_DEAS 0x00040000
-+#define BMI_CMD_RX_MR_DEF (BMI_CMD_MR_LEAC | \
-+ BMI_CMD_MR_SLEAC | \
-+ BMI_CMD_MR_MA | \
-+ BMI_CMD_MR_DEAS)
-+#define BMI_CMD_ATTR_ORDER 0x80000000
-+#define BMI_CMD_ATTR_SYNC 0x02000000
-+#define BMI_CMD_ATTR_MODE_MISS_ALLIGN_ADDR_EN 0x00080000
-+#define BMI_CMD_ATTR_MACCMD_MASK 0x0000ff00
-+#define BMI_CMD_ATTR_MACCMD_OVERRIDE 0x00008000
-+#define BMI_CMD_ATTR_MACCMD_SECURED 0x00001000
-+#define BMI_CMD_ATTR_MACCMD_SC_MASK 0x00000f00
-+
-+#define BMI_EXT_BUF_POOL_ID_MASK 0x003F0000
-+#define BMI_STATUS_RX_MASK_UNUSED (uint32_t)(~(FM_PORT_FRM_ERR_DMA | \
-+ FM_PORT_FRM_ERR_PHYSICAL | \
-+ FM_PORT_FRM_ERR_SIZE | \
-+ FM_PORT_FRM_ERR_CLS_DISCARD | \
-+ FM_PORT_FRM_ERR_EXTRACTION | \
-+ FM_PORT_FRM_ERR_NO_SCHEME | \
-+ FM_PORT_FRM_ERR_COLOR_RED | \
-+ FM_PORT_FRM_ERR_COLOR_YELLOW | \
-+ FM_PORT_FRM_ERR_ILL_PLCR | \
-+ FM_PORT_FRM_ERR_PLCR_FRAME_LEN | \
-+ FM_PORT_FRM_ERR_PRS_TIMEOUT | \
-+ FM_PORT_FRM_ERR_PRS_ILL_INSTRUCT | \
-+ FM_PORT_FRM_ERR_BLOCK_LIMIT_EXCEEDED | \
-+ FM_PORT_FRM_ERR_PRS_HDR_ERR | \
-+ FM_PORT_FRM_ERR_IPRE | \
-+ FM_PORT_FRM_ERR_IPR_NCSP | \
-+ FM_PORT_FRM_ERR_KEYSIZE_OVERFLOW))
-+
-+#define BMI_STATUS_OP_MASK_UNUSED (uint32_t)(BMI_STATUS_RX_MASK_UNUSED & \
-+ ~(FM_PORT_FRM_ERR_LENGTH | \
-+ FM_PORT_FRM_ERR_NON_FM | \
-+ FM_PORT_FRM_ERR_UNSUPPORTED_FORMAT))
-+
-+#define BMI_RATE_LIMIT_EN 0x80000000
-+#define BMI_RATE_LIMIT_BURST_SIZE_GRAN 0x80000000
-+#define BMI_RATE_LIMIT_SCALE_BY_2 0x00000001
-+#define BMI_RATE_LIMIT_SCALE_BY_4 0x00000002
-+#define BMI_RATE_LIMIT_SCALE_BY_8 0x00000003
-+
-+#define BMI_RX_FIFO_THRESHOLD_BC 0x80000000
-+
-+#define BMI_PRS_RESULT_HIGH 0x00000000
-+#define BMI_PRS_RESULT_LOW 0xFFFFFFFF
-+
-+
-+#define RX_ERRS_TO_ENQ (FM_PORT_FRM_ERR_DMA | \
-+ FM_PORT_FRM_ERR_PHYSICAL | \
-+ FM_PORT_FRM_ERR_SIZE | \
-+ FM_PORT_FRM_ERR_EXTRACTION | \
-+ FM_PORT_FRM_ERR_NO_SCHEME | \
-+ FM_PORT_FRM_ERR_ILL_PLCR | \
-+ FM_PORT_FRM_ERR_PLCR_FRAME_LEN | \
-+ FM_PORT_FRM_ERR_PRS_TIMEOUT | \
-+ FM_PORT_FRM_ERR_PRS_ILL_INSTRUCT | \
-+ FM_PORT_FRM_ERR_BLOCK_LIMIT_EXCEEDED | \
-+ FM_PORT_FRM_ERR_PRS_HDR_ERR | \
-+ FM_PORT_FRM_ERR_KEYSIZE_OVERFLOW | \
-+ FM_PORT_FRM_ERR_IPRE)
-+
-+#define OP_ERRS_TO_ENQ (RX_ERRS_TO_ENQ | \
-+ FM_PORT_FRM_ERR_LENGTH | \
-+ FM_PORT_FRM_ERR_NON_FM | \
-+ FM_PORT_FRM_ERR_UNSUPPORTED_FORMAT)
-+
-+
-+#define BMI_RX_FIFO_PRI_ELEVATION_MASK 0x03FF0000
-+#define BMI_RX_FIFO_THRESHOLD_MASK 0x000003FF
-+#define BMI_TX_FIFO_MIN_FILL_MASK 0x03FF0000
-+#define BMI_FIFO_PIPELINE_DEPTH_MASK 0x0000F000
-+#define BMI_TX_LOW_COMF_MASK 0x000003FF
-+
-+/* shifts */
-+#define BMI_PORT_CFG_MS_SEL_SHIFT 16
-+#define BMI_DMA_ATTR_IC_CACHE_SHIFT FMAN_SP_DMA_ATTR_IC_CACHE_SHIFT
-+#define BMI_DMA_ATTR_HDR_CACHE_SHIFT FMAN_SP_DMA_ATTR_HDR_CACHE_SHIFT
-+#define BMI_DMA_ATTR_SG_CACHE_SHIFT FMAN_SP_DMA_ATTR_SG_CACHE_SHIFT
-+
-+#define BMI_IM_FOF_SHIFT 28
-+#define BMI_PR_PORTID_SHIFT 24
-+
-+#define BMI_RX_FIFO_PRI_ELEVATION_SHIFT 16
-+#define BMI_RX_FIFO_THRESHOLD_SHIFT 0
-+
-+#define BMI_RX_FRAME_END_CS_IGNORE_SHIFT 24
-+#define BMI_RX_FRAME_END_CUT_SHIFT 16
-+
-+#define BMI_IC_SIZE_SHIFT FMAN_SP_IC_SIZE_SHIFT
-+
-+#define BMI_INT_BUF_MARG_SHIFT 28
-+
-+#define BMI_EXT_BUF_MARG_END_SHIFT FMAN_SP_EXT_BUF_MARG_END_SHIFT
-+
-+#define BMI_CMD_ATTR_COLOR_SHIFT 26
-+#define BMI_CMD_ATTR_COM_MODE_SHIFT 16
-+#define BMI_CMD_ATTR_MACCMD_SHIFT 8
-+#define BMI_CMD_ATTR_MACCMD_OVERRIDE_SHIFT 15
-+#define BMI_CMD_ATTR_MACCMD_SECURED_SHIFT 12
-+#define BMI_CMD_ATTR_MACCMD_SC_SHIFT 8
-+
-+#define BMI_POOL_DEP_NUM_OF_POOLS_VECTOR_SHIFT 24
-+
-+#define BMI_TX_FIFO_MIN_FILL_SHIFT 16
-+#define BMI_TX_LOW_COMF_SHIFT 0
-+
-+#define BMI_PERFORMANCE_TASK_COMP_SHIFT 24
-+#define BMI_PERFORMANCE_PORT_COMP_SHIFT 16
-+#define BMI_PERFORMANCE_DMA_COMP_SHIFT 12
-+#define BMI_PERFORMANCE_FIFO_COMP_SHIFT 0
-+
-+#define BMI_MAX_BURST_SHIFT 16
-+#define BMI_COUNT_RATE_UNIT_SHIFT 16
-+
-+/* sizes */
-+#define FRAME_END_DATA_SIZE 16
-+#define FRAME_OFFSET_UNITS 16
-+#define MIN_TX_INT_OFFSET 16
-+#define MAX_FRAME_OFFSET 64
-+#define MAX_FIFO_PIPELINE_DEPTH 8
-+#define MAX_PERFORMANCE_TASK_COMP 64
-+#define MAX_PERFORMANCE_TX_QUEUE_COMP 8
-+#define MAX_PERFORMANCE_RX_QUEUE_COMP 64
-+#define MAX_PERFORMANCE_DMA_COMP 16
-+#define MAX_NUM_OF_TASKS 64
-+#define MAX_NUM_OF_EXTRA_TASKS 8
-+#define MAX_NUM_OF_DMAS 16
-+#define MAX_NUM_OF_EXTRA_DMAS 8
-+#define MAX_BURST_SIZE 1024
-+#define MIN_NUM_OF_OP_DMAS 2
-+
-+
-+/**************************************************************************//**
-+ @Description QMI defines
-+*//***************************************************************************/
-+/* masks */
-+#define QMI_PORT_CFG_EN 0x80000000
-+#define QMI_PORT_CFG_EN_COUNTERS 0x10000000
-+#define QMI_PORT_STATUS_DEQ_TNUM_BSY 0x80000000
-+#define QMI_PORT_STATUS_DEQ_FD_BSY 0x20000000
-+
-+#define QMI_DEQ_CFG_PREFETCH_NO_TNUM 0x02000000
-+#define QMI_DEQ_CFG_PREFETCH_WAITING_TNUM 0
-+#define QMI_DEQ_CFG_PREFETCH_1_FRAME 0
-+#define QMI_DEQ_CFG_PREFETCH_3_FRAMES 0x01000000
-+
-+#define QMI_DEQ_CFG_PRI 0x80000000
-+#define QMI_DEQ_CFG_TYPE1 0x10000000
-+#define QMI_DEQ_CFG_TYPE2 0x20000000
-+#define QMI_DEQ_CFG_TYPE3 0x30000000
-+
-+#define QMI_DEQ_CFG_SUBPORTAL_MASK 0x1f
-+#define QMI_DEQ_CFG_SUBPORTAL_SHIFT 20
-+
-+/**************************************************************************//**
-+ @Description PARSER defines
-+*//***************************************************************************/
-+/* masks */
-+#define PRS_HDR_ERROR_DIS 0x00000800
-+#define PRS_HDR_SW_PRS_EN 0x00000400
-+#define PRS_CP_OFFSET_MASK 0x0000000F
-+#define PRS_TPID1_MASK 0xFFFF0000
-+#define PRS_TPID2_MASK 0x0000FFFF
-+#define PRS_TPID_DFLT 0x91009100
-+
-+#define PRS_HDR_MPLS_LBL_INTER_EN 0x00200000
-+#define PRS_HDR_IPV6_ROUTE_HDR_EN 0x00008000
-+#define PRS_HDR_PPPOE_MTU_CHECK_EN 0x80000000
-+#define PRS_HDR_UDP_PAD_REMOVAL 0x80000000
-+#define PRS_HDR_TCP_PAD_REMOVAL 0x80000000
-+#define PRS_CAC_STOP 0x00000001
-+#define PRS_CAC_ACTIVE 0x00000100
-+
-+/* shifts */
-+#define PRS_PCTPID_SHIFT 16
-+#define PRS_HDR_MPLS_NEXT_HDR_SHIFT 22
-+#define PRS_HDR_ETH_BC_SHIFT 28
-+#define PRS_HDR_ETH_MC_SHIFT 24
-+#define PRS_HDR_VLAN_STACKED_SHIFT 16
-+#define PRS_HDR_MPLS_STACKED_SHIFT 16
-+#define PRS_HDR_IPV4_1_BC_SHIFT 28
-+#define PRS_HDR_IPV4_1_MC_SHIFT 24
-+#define PRS_HDR_IPV4_2_UC_SHIFT 20
-+#define PRS_HDR_IPV4_2_MC_BC_SHIFT 16
-+#define PRS_HDR_IPV6_1_MC_SHIFT 24
-+#define PRS_HDR_IPV6_2_UC_SHIFT 20
-+#define PRS_HDR_IPV6_2_MC_SHIFT 16
-+
-+#define PRS_HDR_ETH_BC_MASK 0x0fffffff
-+#define PRS_HDR_ETH_MC_MASK 0xf0ffffff
-+#define PRS_HDR_VLAN_STACKED_MASK 0xfff0ffff
-+#define PRS_HDR_MPLS_STACKED_MASK 0xfff0ffff
-+#define PRS_HDR_IPV4_1_BC_MASK 0x0fffffff
-+#define PRS_HDR_IPV4_1_MC_MASK 0xf0ffffff
-+#define PRS_HDR_IPV4_2_UC_MASK 0xff0fffff
-+#define PRS_HDR_IPV4_2_MC_BC_MASK 0xfff0ffff
-+#define PRS_HDR_IPV6_1_MC_MASK 0xf0ffffff
-+#define PRS_HDR_IPV6_2_UC_MASK 0xff0fffff
-+#define PRS_HDR_IPV6_2_MC_MASK 0xfff0ffff
-+
-+/* others */
-+#define PRS_HDR_ENTRY_SIZE 8
-+#define DEFAULT_CLS_PLAN_VECTOR 0xFFFFFFFF
-+
-+#define IPSEC_SW_PATCH_START 0x20
-+#define SCTP_SW_PATCH_START 0x4D
-+#define DCCP_SW_PATCH_START 0x41
-+
-+/**************************************************************************//**
-+ @Description IM defines
-+*//***************************************************************************/
-+#define BD_R_E 0x80000000
-+#define BD_L 0x08000000
-+
-+#define BD_RX_CRE 0x00080000
-+#define BD_RX_FTL 0x00040000
-+#define BD_RX_FTS 0x00020000
-+#define BD_RX_OV 0x00010000
-+
-+#define BD_RX_ERRORS (BD_RX_CRE | BD_RX_FTL | BD_RX_FTS | BD_RX_OV)
-+
-+#define FM_IM_SIZEOF_BD sizeof(t_FmImBd)
-+
-+#define BD_STATUS_MASK 0xffff0000
-+#define BD_LENGTH_MASK 0x0000ffff
-+
-+#define BD_STATUS_AND_LENGTH_SET(bd, val) WRITE_UINT32(*(volatile uint32_t*)(bd), (val))
-+
-+#define BD_STATUS_AND_LENGTH(bd) GET_UINT32(*(volatile uint32_t*)(bd))
-+
-+#define BD_GET(id) &p_FmPort->im.p_BdRing[id]
-+
-+#define IM_ILEGAL_BD_ID 0xffff
-+
-+/* others */
-+#define IM_PRAM_ALIGN 0x100
-+
-+/* masks */
-+#define IM_MODE_GBL 0x20000000
-+#define IM_MODE_BO_MASK 0x18000000
-+#define IM_MODE_BO_SHIFT 3
-+#define IM_MODE_GRC_STP 0x00800000
-+
-+#define IM_MODE_SET_BO(val) (uint32_t)((val << (31-IM_MODE_BO_SHIFT)) & IM_MODE_BO_MASK)
-+
-+#define IM_RXQD_BSYINTM 0x0008
-+#define IM_RXQD_RXFINTM 0x0010
-+#define IM_RXQD_FPMEVT_SEL_MASK 0x0003
-+
-+#define IM_EV_BSY 0x40000000
-+#define IM_EV_RX 0x80000000
-+
-+
-+/**************************************************************************//**
-+ @Description Additional defines
-+*//***************************************************************************/
-+
-+typedef struct {
-+ t_Handle h_FmMuram;
-+ t_FmPortImPram *p_FmPortImPram;
-+ uint8_t fwExtStructsMemId;
-+ uint32_t fwExtStructsMemAttr;
-+ uint16_t bdRingSize;
-+ t_FmImBd *p_BdRing;
-+ t_Handle *p_BdShadow;
-+ uint16_t currBdId;
-+ uint16_t firstBdOfFrameId;
-+
-+ /* Rx port parameters */
-+ uint8_t dataMemId; /**< Memory partition ID for data buffers */
-+ uint32_t dataMemAttributes; /**< Memory attributes for data buffers */
-+ t_BufferPoolInfo rxPool;
-+ uint16_t mrblr;
-+ uint16_t rxFrameAccumLength;
-+ t_FmPortImRxStoreCallback *f_RxStore;
-+
-+ /* Tx port parameters */
-+ uint32_t txFirstBdStatus;
-+ t_FmPortImTxConfCallback *f_TxConf;
-+} t_FmMacIm;
-+
-+
-+typedef struct {
-+ struct fman_port_cfg dfltCfg;
-+ uint32_t dfltFqid;
-+ uint32_t confFqid;
-+ uint32_t errFqid;
-+ uintptr_t baseAddr;
-+ uint8_t deqSubPortal;
-+ bool deqHighPriority;
-+ e_FmPortDeqType deqType;
-+ e_FmPortDeqPrefetchOption deqPrefetchOption;
-+ uint16_t deqByteCnt;
-+ uint8_t cheksumLastBytesIgnore;
-+ uint8_t cutBytesFromEnd;
-+ t_FmBufPoolDepletion bufPoolDepletion;
-+ uint8_t pipelineDepth;
-+ uint16_t fifoLowComfLevel;
-+ bool frmDiscardOverride;
-+ bool enRateLimit;
-+ t_FmPortRateLimit rateLimit;
-+ e_FmPortDualRateLimiterScaleDown rateLimitDivider;
-+ bool enBufPoolDepletion;
-+ uint16_t liodnOffset;
-+ uint16_t liodnBase;
-+ t_FmExtPools extBufPools;
-+ e_FmDmaSwapOption dmaSwapData;
-+ e_FmDmaCacheOption dmaIntContextCacheAttr;
-+ e_FmDmaCacheOption dmaHeaderCacheAttr;
-+ e_FmDmaCacheOption dmaScatterGatherCacheAttr;
-+ bool dmaReadOptimize;
-+ bool dmaWriteOptimize;
-+ uint32_t txFifoMinFillLevel;
-+ uint32_t txFifoLowComfLevel;
-+ uint32_t rxFifoPriElevationLevel;
-+ uint32_t rxFifoThreshold;
-+ t_FmSpBufMargins bufMargins;
-+ t_FmSpIntContextDataCopy intContext;
-+ bool syncReq;
-+ e_FmPortColor color;
-+ fmPortFrameErrSelect_t errorsToDiscard;
-+ fmPortFrameErrSelect_t errorsToEnq;
-+ bool forwardReuseIntContext;
-+ t_FmBufferPrefixContent bufferPrefixContent;
-+ t_FmBackupBmPools *p_BackupBmPools;
-+ bool dontReleaseBuf;
-+ bool setNumOfTasks;
-+ bool setNumOfOpenDmas;
-+ bool setSizeOfFifo;
-+#if (DPAA_VERSION >= 11)
-+ bool noScatherGather;
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+#ifdef FM_HEAVY_TRAFFIC_HANG_ERRATA_FMAN_A005669
-+ bool bcbWorkaround;
-+#endif /* FM_HEAVY_TRAFFIC_HANG_ERRATA_FMAN_A005669 */
-+} t_FmPortDriverParam;
-+
-+
-+typedef struct t_FmPortRxPoolsParams
-+{
-+ uint8_t numOfPools;
-+ uint16_t secondLargestBufSize;
-+ uint16_t largestBufSize;
-+} t_FmPortRxPoolsParams;
-+
-+typedef struct t_FmPortDsarVars {
-+ t_Handle *autoResOffsets;
-+ t_FmPortDsarTablesSizes *autoResMaxSizes;
-+ uint32_t fmbm_tcfg;
-+ uint32_t fmbm_tcmne;
-+ uint32_t fmbm_rfne;
-+ uint32_t fmbm_rfpne;
-+ uint32_t fmbm_rcfg;
-+ bool dsarEnabledParser;
-+} t_FmPortDsarVars;
-+typedef struct {
-+ struct fman_port port;
-+ t_Handle h_Fm;
-+ t_Handle h_FmPcd;
-+ t_Handle h_FmMuram;
-+ t_FmRevisionInfo fmRevInfo;
-+ uint8_t portId;
-+ e_FmPortType portType;
-+ int enabled;
-+ char name[MODULE_NAME_SIZE];
-+ uint8_t hardwarePortId;
-+ uint16_t fmClkFreq;
-+ t_FmPortQmiRegs *p_FmPortQmiRegs;
-+ u_FmPortBmiRegs *p_FmPortBmiRegs;
-+ t_FmPortPrsRegs *p_FmPortPrsRegs;
-+ fmPcdEngines_t pcdEngines;
-+ uint32_t savedBmiNia;
-+ uint8_t netEnvId;
-+ uint32_t optArray[FM_PCD_MAX_NUM_OF_OPTIONS(FM_PCD_MAX_NUM_OF_CLS_PLANS)];
-+ uint32_t lcvs[FM_PCD_PRS_NUM_OF_HDRS];
-+ uint8_t privateInfo;
-+ uint32_t schemesPerPortVector;
-+ bool useClsPlan;
-+ uint8_t clsPlanGrpId;
-+ t_Handle ccTreeId;
-+ t_Handle completeArg;
-+ void (*f_Complete)(t_Handle arg);
-+ t_FmSpBufferOffsets bufferOffsets;
-+ /* Independent-Mode parameters support */
-+ bool imEn;
-+ t_FmMacIm im;
-+ volatile bool lock;
-+ t_Handle h_Spinlock;
-+ t_FmPortExceptionCallback *f_Exception;
-+ t_Handle h_App;
-+ uint8_t internalBufferOffset;
-+ uint8_t fmanCtrlEventId;
-+ uint32_t exceptions;
-+ bool polling;
-+ t_FmExtPools extBufPools;
-+ uint32_t requiredAction;
-+ uint32_t savedQmiPnen;
-+ uint32_t savedBmiFene;
-+ uint32_t savedBmiFpne;
-+ uint32_t savedBmiCmne;
-+ uint32_t savedBmiOfp;
-+ uint32_t savedNonRxQmiRegsPndn;
-+ uint32_t origNonRxQmiRegsPndn;
-+ int savedPrsStartOffset;
-+ bool includeInPrsStatistics;
-+ uint16_t maxFrameLength;
-+ t_FmFmanCtrl orFmanCtrl;
-+ t_FmPortRsrc openDmas;
-+ t_FmPortRsrc tasks;
-+ t_FmPortRsrc fifoBufs;
-+ t_FmPortRxPoolsParams rxPoolsParams;
-+// bool explicitUserSizeOfFifo;
-+ t_Handle h_IpReassemblyManip;
-+ t_Handle h_CapwapReassemblyManip;
-+ t_Handle h_ReassemblyTree;
-+ uint64_t fmMuramPhysBaseAddr;
-+#if (DPAA_VERSION >= 11)
-+ bool vspe;
-+ uint8_t dfltRelativeId;
-+ e_FmPortGprFuncType gprFunc;
-+ t_FmPcdCtrlParamsPage *p_ParamsPage;
-+#endif /* (DPAA_VERSION >= 11) */
-+ t_FmPortDsarVars deepSleepVars;
-+ t_FmPortDriverParam *p_FmPortDriverParam;
-+} t_FmPort;
-+
-+
-+void FmPortConfigIM (t_FmPort *p_FmPort, t_FmPortParams *p_FmPortParams);
-+t_Error FmPortImCheckInitParameters(t_FmPort *p_FmPort);
-+
-+t_Error FmPortImInit(t_FmPort *p_FmPort);
-+void FmPortImFree(t_FmPort *p_FmPort);
-+
-+t_Error FmPortImEnable (t_FmPort *p_FmPort);
-+t_Error FmPortImDisable (t_FmPort *p_FmPort);
-+t_Error FmPortImRx (t_FmPort *p_FmPort);
-+
-+void FmPortSetMacsecLcv(t_Handle h_FmPort);
-+void FmPortSetMacsecCmd(t_Handle h_FmPort, uint8_t dfltSci);
-+
-+
-+t_Error FM_PORT_SetNumOfOpenDmas(t_Handle h_FmPort, t_FmPortRsrc *p_NumOfOpenDmas);
-+t_Error FM_PORT_SetNumOfTasks(t_Handle h_FmPort, t_FmPortRsrc *p_NumOfTasks);
-+t_Error FM_PORT_SetSizeOfFifo(t_Handle h_FmPort, t_FmPortRsrc *p_SizeOfFifo);
-+
-+static __inline__ uint8_t * BdBufferGet (t_PhysToVirt *f_PhysToVirt, t_FmImBd *p_Bd)
-+{
-+ uint64_t physAddr = (uint64_t)((uint64_t)GET_UINT8(p_Bd->buff.high) << 32);
-+ physAddr |= GET_UINT32(p_Bd->buff.low);
-+
-+ return (uint8_t *)f_PhysToVirt((physAddress_t)(physAddr));
-+}
-+
-+static __inline__ void SET_ADDR(volatile t_FmPhysAddr *fmPhysAddr, uint64_t value)
-+{
-+ WRITE_UINT8(fmPhysAddr->high,(uint8_t)((value & 0x000000ff00000000LL) >> 32));
-+ WRITE_UINT32(fmPhysAddr->low,(uint32_t)value);
-+}
-+
-+static __inline__ void BdBufferSet(t_VirtToPhys *f_VirtToPhys, t_FmImBd *p_Bd, uint8_t *p_Buffer)
-+{
-+ uint64_t physAddr = (uint64_t)(f_VirtToPhys(p_Buffer));
-+ SET_ADDR(&p_Bd->buff, physAddr);
-+}
-+
-+static __inline__ uint16_t GetNextBdId(t_FmPort *p_FmPort, uint16_t id)
-+{
-+ if (id < p_FmPort->im.bdRingSize-1)
-+ return (uint16_t)(id+1);
-+ else
-+ return 0;
-+}
-+
-+void FM_PORT_Dsar_DumpRegs(void);
-+
-+
-+#endif /* __FM_PORT_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/Port/fm_port_dsar.h
-@@ -0,0 +1,494 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+/**************************************************************************//**
-+ @File fm_port_dsar.h
-+
-+ @Description Deep Sleep Auto Response project - common module header file.
-+
-+ Author - Eyal Harari
-+
-+ @Cautions See the FMan Controller spec and design document for more information.
-+*//***************************************************************************/
-+
-+#ifndef __FM_PORT_DSAR_H_
-+#define __FM_PORT_DSAR_H_
-+
-+#define DSAR_GETSER_MASK 0xFF0000FF
-+
-+#if defined(__MWERKS__) && !defined(__GNUC__)
-+#pragma pack(push,1)
-+#endif /* defined(__MWERKS__) && ... */
-+
-+/**************************************************************************//**
-+ @Description Deep Sleep Auto Response VLAN-IPv4 Binding Table (for ARP/ICMPv4)
-+ Refer to the FMan Controller spec for more details.
-+*//***************************************************************************/
-+typedef _Packed struct
-+{
-+ uint32_t ipv4Addr; /*!< 32 bit IPv4 Address. */
-+ uint16_t vlanId; /*!< 12 bits VLAN ID. The 4 left-most bits should be cleared */
-+ /*!< This field should be 0x0000 for an entry with no VLAN tag or a null VLAN ID. */
-+ uint16_t reserved;
-+} _PackedType t_DsarArpBindingEntry;
-+
-+/**************************************************************************//**
-+ @Description Deep Sleep Auto Response Address Resolution Protocol Statistics Descriptor
-+ Refer to the FMan Controller spec for more details.
-+ 0x00 INVAL_CNT Invalid ARP IPv4-Ethernet counter
-+ 0x04 ECHO_CNT Echo counter
-+ 0x08 CD_CNT Conflict Detection counter
-+ 0x0C AR_CNT Auto-Response counter
-+ 0x10 RATM_CNT Replies Addressed To Me counter
-+ 0x14 UKOP_CNT Unknown Operation counter
-+ 0x18 NMTP_CNT Not my TPA counter
-+ 0x1C NMVLAN_CNT Not My VLAN counter
-+*//***************************************************************************/
-+typedef _Packed struct
-+{
-+ uint32_t invalCnt; /**< Invalid ARP IPv4-Ethernet counter. */
-+ uint32_t echoCnt; /**< Echo counter. */
-+ uint32_t cdCnt; /**< Conflict Detection counter. */
-+ uint32_t arCnt; /**< Auto-Response counter. */
-+ uint32_t ratmCnt; /**< Replies Addressed To Me counter. */
-+ uint32_t ukopCnt; /**< Unknown Operation counter. */
-+ uint32_t nmtpCnt; /**< Not my TPA counter. */
-+ uint32_t nmVlanCnt; /**< Not My VLAN counter */
-+} _PackedType t_DsarArpStatistics;
-+
-+
-+/**************************************************************************//**
-+ @Description Deep Sleep Auto Response Address Resolution Protocol Descriptor
-+ 0x0 0-15 Control bits [0-15]. Bit 15 = CDEN.
-+ 0x2 0-15 NumOfBindings Number of entries in the binding list.
-+ 0x4 0-15 BindingsPointer Bindings Pointer. This points to an IPv4-MAC Addresses Bindings list.
-+ 0x6 0-15
-+ 0x8 0-15 StatisticsPointer Statistics Pointer. This field points to the ARP Descriptors statistics data structure.
-+ 0xA 0-15
-+ 0xC 0-15 Reserved Reserved. Must be cleared.
-+ 0xE 015
-+
-+*//***************************************************************************/
-+typedef _Packed struct
-+{
-+ uint16_t control; /** Control bits [0-15]. Bit 15 = CDEN */
-+ uint16_t numOfBindings; /**< Number of VLAN-IPv4 */
-+ uint32_t p_Bindings; /**< VLAN-IPv4 Bindings table pointer. */
-+ uint32_t p_Statistics; /**< Statistics Data Structure pointer. */
-+ uint32_t reserved1; /**< Reserved. */
-+} _PackedType t_DsarArpDescriptor;
-+
-+
-+/**************************************************************************//**
-+ @Description Deep Sleep Auto Response VLAN-IPv4 Binding Table (for ARP/ICMPv4)
-+ Refer to the FMan Controller spec for more details.
-+*//***************************************************************************/
-+typedef _Packed struct
-+{
-+ uint32_t ipv4Addr; /*!< 32 bit IPv4 Address. */
-+ uint16_t vlanId; /*!< 12 bits VLAN ID. The 4 left-most bits should be cleared */
-+ /*!< This field should be 0x0000 for an entry with no VLAN tag or a null VLAN ID. */
-+ uint16_t reserved;
-+} _PackedType t_DsarIcmpV4BindingEntry;
-+
-+/**************************************************************************//**
-+ @Description Deep Sleep Auto Response ICMPv4 Statistics Descriptor
-+ Refer to the FMan Controller spec for more details.
-+ 0x00 INVAL_CNT Invalid ICMPv4 header counter
-+ 0x04 NMVLAN_CNT Not My VLAN counter
-+ 0x08 NMIP_CNT Not My IP counter
-+ 0x0C AR_CNT Auto-Response counter
-+ 0x10 CSERR_CNT Checksum Error counter
-+ 0x14 Reserved Reserved
-+ 0x18 Reserved Reserved
-+ 0x1C Reserved Reserved
-+
-+*//***************************************************************************/
-+typedef _Packed struct
-+{
-+ uint32_t invalCnt; /**< Invalid ICMPv4 Echo counter. */
-+ uint32_t nmVlanCnt; /**< Not My VLAN counter */
-+ uint32_t nmIpCnt; /**< Not My IP counter */
-+ uint32_t arCnt; /**< Auto-Response counter */
-+ uint32_t cserrCnt; /**< Checksum Error counter */
-+ uint32_t reserved0; /**< Reserved */
-+ uint32_t reserved1; /**< Reserved */
-+ uint32_t reserved2; /**< Reserved */
-+} _PackedType t_DsarIcmpV4Statistics;
-+
-+
-+
-+/**************************************************************************//**
-+ @Description Deep Sleep Auto Response ICMPv4 Descriptor
-+ 0x0 0-15 Control bits [0-15]
-+ 0x2 0-15 NumOfBindings Number of entries in the binding list.
-+ 0x4 0-15 BindingsPointer Bindings Pointer. This points to an VLAN-IPv4 Addresses Bindings list.
-+ 0x6 0-15
-+ 0x8 0-15 StatisticsPointer Statistics Pointer. This field points to the ICMPv4 statistics data structure.
-+ 0xA 0-15
-+ 0xC 0-15 Reserved Reserved. Must be cleared.
-+ 0xE 015
-+
-+*//***************************************************************************/
-+typedef _Packed struct
-+{
-+ uint16_t control; /** Control bits [0-15]. */
-+ uint16_t numOfBindings; /**< Number of VLAN-IPv4 */
-+ uint32_t p_Bindings; /**< VLAN-IPv4 Bindings table pointer. */
-+ uint32_t p_Statistics; /**< Statistics Data Structure pointer. */
-+ uint32_t reserved1; /**< Reserved. */
-+} _PackedType t_DsarIcmpV4Descriptor;
-+
-+/**************************************************************************//**
-+ @Description Deep Sleep Auto Response VLAN-IPv4 Binding Table (for ARP/ICMPv4)
-+ The 4 left-most bits (15:12) of the VlanId parameter are control flags.
-+ Flags[3:1] (VlanId[15:13]): Reserved, should be cleared.
-+ Flags[0] (VlanId[12]): Temporary address.
-+ • 0 - Assigned IP address.
-+ • 1- Temporary (tentative) IP address.
-+ Refer to the FMan Controller spec for more details.
-+*//***************************************************************************/
-+typedef _Packed struct
-+{
-+ uint32_t ipv6Addr[4]; /*!< 3 * 32 bit IPv4 Address. */
-+ uint16_t resFlags:4; /*!< reserved flags. should be cleared */
-+ uint16_t vlanId:12; /*!< 12 bits VLAN ID. */
-+ /*!< This field should be 0x000 for an entry with no VLAN tag or a null VLAN ID. */
-+ uint16_t reserved;
-+} _PackedType t_DsarIcmpV6BindingEntry;
-+
-+/**************************************************************************//**
-+ @Description Deep Sleep Auto Response ICMPv4 Statistics Descriptor
-+ Refer to the FMan Controller spec for more details.
-+ 0x00 INVAL_CNT Invalid ICMPv4 header counter
-+ 0x04 NMVLAN_CNT Not My VLAN counter
-+ 0x08 NMIP_CNT Not My IP counter
-+ 0x0C AR_CNT Auto-Response counter
-+ 0x10 CSERR_CNT Checksum Error counter
-+ 0x14 MCAST_CNT Multicast counter
-+ 0x18 Reserved Reserved
-+ 0x1C Reserved Reserved
-+
-+*//***************************************************************************/
-+typedef _Packed struct
-+{
-+ uint32_t invalCnt; /**< Invalid ICMPv4 Echo counter. */
-+ uint32_t nmVlanCnt; /**< Not My VLAN counter */
-+ uint32_t nmIpCnt; /**< Not My IP counter */
-+ uint32_t arCnt; /**< Auto-Response counter */
-+ uint32_t reserved1; /**< Reserved */
-+ uint32_t reserved2; /**< Reserved */
-+ uint32_t reserved3; /**< Reserved */
-+ uint32_t reserved4; /**< Reserved */
-+} _PackedType t_DsarIcmpV6Statistics;
-+
-+/**************************************************************************//**
-+ @Description Deep Sleep Auto Response Neighbor Discovery Statistics Descriptor
-+ 0x00 INVAL_CNT Invalid Neighbor Discovery message counter
-+ 0x04 NMVLAN_CNT Not My VLAN counter
-+ 0x08 NMIP_CNT Not My IP counter
-+ 0x0C AR_CNT Auto-Response counter
-+ 0x10 CSERR_CNT Checksum Error counter
-+ 0x14 USADVERT_CNT Unsolicited Neighbor Advertisements counter
-+ 0x18 NMMCAST_CNT Not My Multicast group counter
-+ 0x1C NSLLA_CNT No Source Link-Layer Address counter. Indicates that there was a match on a Target
-+ Address of a packet that its source IP address is a unicast address, but the ICMPv6
-+ Source Link-layer Address option is omitted
-+*//***************************************************************************/
-+typedef _Packed struct
-+{
-+ uint32_t invalCnt; /**< Invalid ICMPv4 Echo counter. */
-+ uint32_t nmVlanCnt; /**< Not My VLAN counter */
-+ uint32_t nmIpCnt; /**< Not My IP counter */
-+ uint32_t arCnt; /**< Auto-Response counter */
-+ uint32_t reserved1; /**< Reserved */
-+ uint32_t usadvertCnt; /**< Unsolicited Neighbor Advertisements counter */
-+ uint32_t nmmcastCnt; /**< Not My Multicast group counter */
-+ uint32_t nsllaCnt; /**< No Source Link-Layer Address counter */
-+} _PackedType t_NdStatistics;
-+
-+/**************************************************************************//**
-+ @Description Deep Sleep Auto Response ICMPv6 Descriptor
-+ 0x0 0-15 Control bits [0-15]
-+ 0x2 0-15 NumOfBindings Number of entries in the binding list.
-+ 0x4 0-15 BindingsPointer Bindings Pointer. This points to an VLAN-IPv4 Addresses Bindings list.
-+ 0x6 0-15
-+ 0x8 0-15 StatisticsPointer Statistics Pointer. This field points to the ICMPv4 statistics data structure.
-+ 0xA 0-15
-+ 0xC 0-15 Reserved Reserved. Must be cleared.
-+ 0xE 015
-+
-+*//***************************************************************************/
-+typedef _Packed struct
-+{
-+ uint16_t control; /** Control bits [0-15]. */
-+ uint16_t numOfBindings; /**< Number of VLAN-IPv6 */
-+ uint32_t p_Bindings; /**< VLAN-IPv4 Bindings table pointer. */
-+ uint32_t p_Statistics; /**< Statistics Data Structure pointer. */
-+ uint32_t reserved1; /**< Reserved. */
-+} _PackedType t_DsarIcmpV6Descriptor;
-+
-+
-+/**************************************************************************//**
-+ @Description Internet Control Message Protocol (ICMPv6) Echo message header
-+ The fields names are taken from RFC 4443.
-+*//***************************************************************************/
-+/* 0 1 2 3 */
-+/* 0 1 2 3 4 5 6 7 8 9 0 1 2 3 4 5 6 7 8 9 0 1 2 3 4 5 6 7 8 9 0 1 */
-+/* +-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+ */
-+/* | Type | Code | Checksum | */
-+/* +-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+ */
-+/* | Identifier | Sequence Number | */
-+/* +-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+ */
-+/* | Data ... */
-+/* +-+-+-+-+- */
-+typedef _Packed struct
-+{
-+ uint8_t type;
-+ uint8_t code;
-+ uint16_t checksum;
-+ uint16_t identifier;
-+ uint16_t sequenceNumber;
-+} _PackedType t_IcmpV6EchoHdr;
-+
-+/**************************************************************************//**
-+ @Description Internet Control Message Protocol (ICMPv6)
-+ Neighbor Solicitation/Advertisement header
-+ The fields names are taken from RFC 4861.
-+ The R/S/O fields are valid for Neighbor Advertisement only
-+*//***************************************************************************/
-+/* 0 1 2 3
-+ * 0 1 2 3 4 5 6 7 8 9 0 1 2 3 4 5 6 7 8 9 0 1 2 3 4 5 6 7 8 9 0 1
-+ * +-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
-+ * | Type | Code | Checksum |
-+ * +-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
-+ * |R|S|O| Reserved |
-+ * +-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
-+ * | |
-+ * + +
-+ * | |
-+ * + Target Address +
-+ * | |
-+ * + +
-+ * | |
-+ * +-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
-+ * | Options ...
-+ * +-+-+-+-+-+-+-+-+-+-+-+-
-+ *
-+ * Options Format:
-+ * 0 1 2 3
-+ * 0 1 2 3 4 5 6 7 8 9 0 1 2 3 4 5 6 7 8 9 0 1 2 3 4 5 6 7 8 9 0 1
-+ * +-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
-+ * | Type | Length | Link-Layer Address ... |
-+ * +-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
-+ * | Link-Layer Address |
-+ * +-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
-+*/
-+typedef _Packed struct
-+{
-+ uint8_t type;
-+ uint8_t code;
-+ uint16_t checksum;
-+ uint32_t router:1;
-+ uint32_t solicited:1;
-+ uint32_t override:1;
-+ uint32_t reserved:29;
-+ uint32_t targetAddr[4];
-+ uint8_t optionType;
-+ uint8_t optionLength;
-+ uint8_t linkLayerAddr[6];
-+} _PackedType t_IcmpV6NdHdr;
-+
-+/**************************************************************************//**
-+ @Description Deep Sleep Auto Response ICMPv6 Descriptor
-+ 0x0 0-15 Control bits [0-15]
-+ 0x2 0-15 NumOfBindings Number of entries in the binding list.
-+ 0x4 0-15 BindingsPointer Bindings Pointer. This points to an VLAN-IPv4 Addresses Bindings list.
-+ 0x6 0-15
-+ 0x8 0-15 StatisticsPointer Statistics Pointer. This field points to the ICMPv4 statistics data structure.
-+ 0xA 0-15
-+ 0xC 0-15 Reserved Reserved. Must be cleared.
-+ 0xE 015
-+
-+*//***************************************************************************/
-+typedef _Packed struct
-+{
-+ uint16_t control; /** Control bits [0-15]. */
-+ uint16_t numOfBindings; /**< Number of VLAN-IPv6 */
-+ uint32_t p_Bindings; /**< VLAN-IPv4 Bindings table pointer. */
-+ uint32_t p_Statistics; /**< Statistics Data Structure pointer. */
-+ uint32_t solicitedAddr; /**< Solicited Node Multicast Group Address */
-+} _PackedType t_DsarNdDescriptor;
-+
-+/**************************************************************************//**
-+@Description Deep Sleep Auto Response SNMP OIDs table entry
-+
-+*//***************************************************************************/
-+typedef struct {
-+ uint16_t oidSize; /**< Size in octets of the OID. */
-+ uint16_t resSize; /**< Size in octets of the value that is attached to the OID. */
-+ uint32_t p_Oid; /**< Pointer to the OID. OID is encoded in BER but type and length are excluded. */
-+ uint32_t resValOrPtr; /**< Value (for up to 4 octets) or pointer to the Value. Encoded in BER. */
-+ uint32_t reserved;
-+} t_OidsTblEntry;
-+
-+/**************************************************************************//**
-+ @Description Deep Sleep Auto Response SNMP IPv4 Addresses Table Entry
-+ Refer to the FMan Controller spec for more details.
-+*//***************************************************************************/
-+typedef struct
-+{
-+ uint32_t ipv4Addr; /*!< 32 bit IPv4 Address. */
-+ uint16_t vlanId; /*!< 12 bits VLAN ID. The 4 left-most bits should be cleared */
-+ /*!< This field should be 0x0000 for an entry with no VLAN tag or a null VLAN ID. */
-+ uint16_t reserved;
-+} t_DsarSnmpIpv4AddrTblEntry;
-+
-+/**************************************************************************//**
-+ @Description Deep Sleep Auto Response SNMP IPv6 Addresses Table Entry
-+ Refer to the FMan Controller spec for more details.
-+*//***************************************************************************/
-+#pragma pack(push,1)
-+typedef struct
-+{
-+ uint32_t ipv6Addr[4]; /*!< 4 * 32 bit IPv6 Address. */
-+ uint16_t vlanId; /*!< 12 bits VLAN ID. The 4 left-most bits should be cleared */
-+ /*!< This field should be 0x0000 for an entry with no VLAN tag or a null VLAN ID. */
-+ uint16_t reserved;
-+} t_DsarSnmpIpv6AddrTblEntry;
-+#pragma pack(pop)
-+
-+/**************************************************************************//**
-+@Description Deep Sleep Auto Response SNMP statistics table
-+
-+*//***************************************************************************/
-+typedef struct {
-+ uint32_t snmpErrCnt; /**< Counts SNMP errors (wrong version, BER encoding, format). */
-+ uint32_t snmpCommunityErrCnt; /**< Counts messages that were dropped due to insufficient permission. */
-+ uint32_t snmpTotalDiscardCnt; /**< Counts any message that was dropped. */
-+ uint32_t snmpGetReqCnt; /**< Counts the number of get-request messages */
-+ uint32_t snmpGetNextReqCnt; /**< Counts the number of get-next-request messages */
-+} t_DsarSnmpStatistics;
-+
-+/**************************************************************************//**
-+ @Description Deep Sleep Auto Response SNMP Descriptor
-+
-+*//***************************************************************************/
-+typedef struct
-+{
-+ uint16_t control; /**< Control bits [0-15]. */
-+ uint16_t maxSnmpMsgLength; /**< Maximal allowed SNMP message length. */
-+ uint16_t numOfIpv4Addresses; /**< Number of entries in IPv4 addresses table. */
-+ uint16_t numOfIpv6Addresses; /**< Number of entries in IPv6 addresses table. */
-+ uint32_t p_Ipv4AddrTbl; /**< Pointer to IPv4 addresses table. */
-+ uint32_t p_Ipv6AddrTbl; /**< Pointer to IPv6 addresses table. */
-+ uint32_t p_RdOnlyCommunityStr; /**< Pointer to the Read Only Community String. */
-+ uint32_t p_RdWrCommunityStr; /**< Pointer to the Read Write Community String. */
-+ uint32_t p_OidsTbl; /**< Pointer to OIDs table. */
-+ uint32_t oidsTblSize; /**< Number of entries in OIDs table. */
-+ uint32_t p_Statistics; /**< Pointer to SNMP statistics table. */
-+} t_DsarSnmpDescriptor;
-+
-+/**************************************************************************//**
-+@Description Deep Sleep Auto Response (Common) Statistics
-+
-+*//***************************************************************************/
-+typedef _Packed struct {
-+ uint32_t dsarDiscarded;
-+ uint32_t dsarErrDiscarded;
-+ uint32_t dsarFragDiscarded;
-+ uint32_t dsarTunnelDiscarded;
-+ uint32_t dsarArpDiscarded;
-+ uint32_t dsarIpDiscarded;
-+ uint32_t dsarTcpDiscarded;
-+ uint32_t dsarUdpDiscarded;
-+ uint32_t dsarIcmpV6ChecksumErr; /* ICMPv6 Checksum Error counter */
-+ uint32_t dsarIcmpV6OtherType; /* ICMPv6 'Other' type (not Echo or Neighbor Solicitaion/Advertisement counter */
-+ uint32_t dsarIcmpV4OtherType; /* ICMPv4 'Other' type (not Echo) counter */
-+} _PackedType t_ArStatistics;
-+
-+
-+/**************************************************************************//**
-+@Description Deep Sleep Auto Response TCP/UDP port filter table entry
-+
-+*//***************************************************************************/
-+typedef _Packed struct {
-+ uint32_t Ports;
-+ uint32_t PortsMask;
-+} _PackedType t_PortTblEntry;
-+
-+
-+
-+/**************************************************************************//**
-+@Description Deep Sleep Auto Response Common Parameters Descriptor
-+
-+*//***************************************************************************/
-+typedef _Packed struct {
-+ uint8_t arTxPort; /* 0x00 0-7 Auto Response Transmit Port number */
-+ uint8_t controlBits; /* 0x00 8-15 Auto Response control bits */
-+ uint16_t res1; /* 0x00 16-31 Reserved */
-+ uint32_t activeHPNIA; /* 0x04 0-31 Active mode Hardware Parser NIA */
-+ uint16_t snmpPort; /* 0x08 0-15 SNMP Port. */
-+ uint8_t macStationAddr[6]; /* 0x08 16-31 and 0x0C 0-31 MAC Station Address */
-+ uint8_t res2; /* 0x10 0-7 Reserved */
-+ uint8_t filterControl; /* 0x10 8-15 Filtering Control Bits. */
-+ uint16_t tcpControlPass; /* 0x10 16-31 TCP control pass flags */
-+ uint8_t ipProtocolTblSize; /* 0x14 0-7 IP Protocol Table Size. */
-+ uint8_t udpPortTblSize; /* 0x14 8-15 UDP Port Table Size. */
-+ uint8_t tcpPortTblSize; /* 0x14 16-23 TCP Port Table Size. */
-+ uint8_t res3; /* 0x14 24-31 Reserved */
-+ uint32_t p_IpProtocolFiltTbl; /* 0x18 0-31 Pointer to IP Protocol Filter Table */
-+ uint32_t p_UdpPortFiltTbl; /* 0x1C 0-31 Pointer to UDP Port Filter Table */
-+ uint32_t p_TcpPortFiltTbl; /* 0x20 0-31 Pointer to TCP Port Filter Table */
-+ uint32_t res4; /* 0x24 Reserved */
-+ uint32_t p_ArpDescriptor; /* 0x28 0-31 ARP Descriptor Pointer. */
-+ uint32_t p_NdDescriptor; /* 0x2C 0-31 Neighbor Discovery Descriptor. */
-+ uint32_t p_IcmpV4Descriptor; /* 0x30 0-31 ICMPv4 Descriptor pointer. */
-+ uint32_t p_IcmpV6Descriptor; /* 0x34 0-31 ICMPv6 Descriptor pointer. */
-+ uint32_t p_SnmpDescriptor; /* 0x38 0-31 SNMP Descriptor pointer. */
-+ uint32_t p_ArStats; /* 0x3C 0-31 Pointer to Auto Response Statistics */
-+} _PackedType t_ArCommonDesc;
-+
-+#if defined(__MWERKS__) && !defined(__GNUC__)
-+#pragma pack(pop)
-+#endif /* defined(__MWERKS__) && ... */
-+
-+/* t_ArCommonDesc.filterControl bits */
-+#define IP_PROT_TBL_PASS_MASK 0x08
-+#define UDP_PORT_TBL_PASS_MASK 0x04
-+#define TCP_PORT_TBL_PASS_MASK 0x02
-+
-+/* Offset of TCF flags within TCP packet */
-+#define TCP_FLAGS_OFFSET 12
-+
-+
-+#endif /* __FM_PORT_DSAR_H_ */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/Port/fm_port_im.c
-@@ -0,0 +1,753 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/******************************************************************************
-+ @File fm_port_im.c
-+
-+ @Description FM Port Independent-Mode ...
-+*//***************************************************************************/
-+#include "std_ext.h"
-+#include "string_ext.h"
-+#include "error_ext.h"
-+#include "memcpy_ext.h"
-+#include "fm_muram_ext.h"
-+
-+#include "fm_port.h"
-+
-+
-+#define TX_CONF_STATUS_UNSENT 0x1
-+
-+
-+typedef enum e_TxConfType
-+{
-+ e_TX_CONF_TYPE_CHECK = 0 /**< check if all the buffers were touched by the muxator, no confirmation callback */
-+ ,e_TX_CONF_TYPE_CALLBACK = 1 /**< confirm to user all the available sent buffers */
-+ ,e_TX_CONF_TYPE_FLUSH = 3 /**< confirm all buffers plus the unsent one with an appropriate status */
-+} e_TxConfType;
-+
-+
-+static void ImException(t_Handle h_FmPort, uint32_t event)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+
-+ ASSERT_COND(((event & (IM_EV_RX | IM_EV_BSY)) && FmIsMaster(p_FmPort->h_Fm)) ||
-+ !FmIsMaster(p_FmPort->h_Fm));
-+
-+ if (event & IM_EV_RX)
-+ FmPortImRx(p_FmPort);
-+ if ((event & IM_EV_BSY) && p_FmPort->f_Exception)
-+ p_FmPort->f_Exception(p_FmPort->h_App, e_FM_PORT_EXCEPTION_IM_BUSY);
-+}
-+
-+
-+static t_Error TxConf(t_FmPort *p_FmPort, e_TxConfType confType)
-+{
-+ t_Error retVal = E_BUSY;
-+ uint32_t bdStatus;
-+ uint16_t savedStartBdId, confBdId;
-+
-+ ASSERT_COND(p_FmPort);
-+
-+ /*
-+ if (confType==e_TX_CONF_TYPE_CHECK)
-+ return (WfqEntryIsQueueEmpty(p_FmPort->im.h_WfqEntry) ? E_OK : E_BUSY);
-+ */
-+
-+ confBdId = savedStartBdId = p_FmPort->im.currBdId;
-+ bdStatus = BD_STATUS_AND_LENGTH(BD_GET(confBdId));
-+
-+ /* If R bit is set, we don't enter, or we break.
-+ we run till we get to R, or complete the loop */
-+ while ((!(bdStatus & BD_R_E) || (confType == e_TX_CONF_TYPE_FLUSH)) && (retVal != E_OK))
-+ {
-+ if (confType & e_TX_CONF_TYPE_CALLBACK) /* if it is confirmation with user callbacks */
-+ BD_STATUS_AND_LENGTH_SET(BD_GET(confBdId), 0);
-+
-+ /* case 1: R bit is 0 and Length is set -> confirm! */
-+ if ((confType & e_TX_CONF_TYPE_CALLBACK) && (bdStatus & BD_LENGTH_MASK))
-+ {
-+ if (p_FmPort->im.f_TxConf)
-+ {
-+ if ((confType == e_TX_CONF_TYPE_FLUSH) && (bdStatus & BD_R_E))
-+ p_FmPort->im.f_TxConf(p_FmPort->h_App,
-+ BdBufferGet(XX_PhysToVirt, BD_GET(confBdId)),
-+ TX_CONF_STATUS_UNSENT,
-+ p_FmPort->im.p_BdShadow[confBdId]);
-+ else
-+ p_FmPort->im.f_TxConf(p_FmPort->h_App,
-+ BdBufferGet(XX_PhysToVirt, BD_GET(confBdId)),
-+ 0,
-+ p_FmPort->im.p_BdShadow[confBdId]);
-+ }
-+ }
-+ /* case 2: R bit is 0 and Length is 0 -> not used yet, nop! */
-+
-+ confBdId = GetNextBdId(p_FmPort, confBdId);
-+ if (confBdId == savedStartBdId)
-+ retVal = E_OK;
-+ bdStatus = BD_STATUS_AND_LENGTH(BD_GET(confBdId));
-+ }
-+
-+ return retVal;
-+}
-+
-+t_Error FmPortImEnable(t_FmPort *p_FmPort)
-+{
-+ uint32_t tmpReg = GET_UINT32(p_FmPort->im.p_FmPortImPram->mode);
-+ WRITE_UINT32(p_FmPort->im.p_FmPortImPram->mode, (uint32_t)(tmpReg & ~IM_MODE_GRC_STP));
-+ return E_OK;
-+}
-+
-+t_Error FmPortImDisable(t_FmPort *p_FmPort)
-+{
-+ uint32_t tmpReg = GET_UINT32(p_FmPort->im.p_FmPortImPram->mode);
-+ WRITE_UINT32(p_FmPort->im.p_FmPortImPram->mode, (uint32_t)(tmpReg | IM_MODE_GRC_STP));
-+ return E_OK;
-+}
-+
-+t_Error FmPortImRx(t_FmPort *p_FmPort)
-+{
-+ t_Handle h_CurrUserPriv, h_NewUserPriv;
-+ uint32_t bdStatus;
-+ volatile uint8_t buffPos;
-+ uint16_t length;
-+ uint16_t errors;
-+ uint8_t *p_CurData, *p_Data;
-+ uint32_t flags;
-+
-+ ASSERT_COND(p_FmPort);
-+
-+ flags = XX_LockIntrSpinlock(p_FmPort->h_Spinlock);
-+ if (p_FmPort->lock)
-+ {
-+ XX_UnlockIntrSpinlock(p_FmPort->h_Spinlock, flags);
-+ return E_OK;
-+ }
-+ p_FmPort->lock = TRUE;
-+ XX_UnlockIntrSpinlock(p_FmPort->h_Spinlock, flags);
-+
-+ bdStatus = BD_STATUS_AND_LENGTH(BD_GET(p_FmPort->im.currBdId));
-+
-+ while (!(bdStatus & BD_R_E)) /* while there is data in the Rx BD */
-+ {
-+ if ((p_Data = p_FmPort->im.rxPool.f_GetBuf(p_FmPort->im.rxPool.h_BufferPool, &h_NewUserPriv)) == NULL)
-+ {
-+ p_FmPort->lock = FALSE;
-+ RETURN_ERROR(MAJOR, E_NOT_AVAILABLE, ("Data buffer"));
-+ }
-+
-+ if (p_FmPort->im.firstBdOfFrameId == IM_ILEGAL_BD_ID)
-+ p_FmPort->im.firstBdOfFrameId = p_FmPort->im.currBdId;
-+
-+ p_CurData = BdBufferGet(p_FmPort->im.rxPool.f_PhysToVirt, BD_GET(p_FmPort->im.currBdId));
-+ h_CurrUserPriv = p_FmPort->im.p_BdShadow[p_FmPort->im.currBdId];
-+ length = (uint16_t)((bdStatus & BD_L) ?
-+ ((bdStatus & BD_LENGTH_MASK) - p_FmPort->im.rxFrameAccumLength):
-+ (bdStatus & BD_LENGTH_MASK));
-+ p_FmPort->im.rxFrameAccumLength += length;
-+
-+ /* determine whether buffer is first, last, first and last (single */
-+ /* buffer frame) or middle (not first and not last) */
-+ buffPos = (uint8_t)((p_FmPort->im.currBdId == p_FmPort->im.firstBdOfFrameId) ?
-+ ((bdStatus & BD_L) ? SINGLE_BUF : FIRST_BUF) :
-+ ((bdStatus & BD_L) ? LAST_BUF : MIDDLE_BUF));
-+
-+ if (bdStatus & BD_L)
-+ {
-+ p_FmPort->im.rxFrameAccumLength = 0;
-+ p_FmPort->im.firstBdOfFrameId = IM_ILEGAL_BD_ID;
-+ }
-+
-+ BdBufferSet(p_FmPort->im.rxPool.f_VirtToPhys, BD_GET(p_FmPort->im.currBdId), p_Data);
-+
-+ BD_STATUS_AND_LENGTH_SET(BD_GET(p_FmPort->im.currBdId), BD_R_E);
-+
-+ errors = (uint16_t)((bdStatus & BD_RX_ERRORS) >> 16);
-+ p_FmPort->im.p_BdShadow[p_FmPort->im.currBdId] = h_NewUserPriv;
-+
-+ p_FmPort->im.currBdId = GetNextBdId(p_FmPort, p_FmPort->im.currBdId);
-+ WRITE_UINT16(p_FmPort->im.p_FmPortImPram->rxQd.offsetOut, (uint16_t)(p_FmPort->im.currBdId<<4));
-+ /* Pass the buffer if one of the conditions is true:
-+ - There are no errors
-+ - This is a part of a larger frame ( the application has already received some buffers ) */
-+ if ((buffPos != SINGLE_BUF) || !errors)
-+ {
-+ if (p_FmPort->im.f_RxStore(p_FmPort->h_App,
-+ p_CurData,
-+ length,
-+ errors,
-+ buffPos,
-+ h_CurrUserPriv) == e_RX_STORE_RESPONSE_PAUSE)
-+ break;
-+ }
-+ else if (p_FmPort->im.rxPool.f_PutBuf(p_FmPort->im.rxPool.h_BufferPool,
-+ p_CurData,
-+ h_CurrUserPriv))
-+ {
-+ p_FmPort->lock = FALSE;
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("Failed freeing data buffer"));
-+ }
-+
-+ bdStatus = BD_STATUS_AND_LENGTH(BD_GET(p_FmPort->im.currBdId));
-+ }
-+ p_FmPort->lock = FALSE;
-+ return E_OK;
-+}
-+
-+void FmPortConfigIM (t_FmPort *p_FmPort, t_FmPortParams *p_FmPortParams)
-+{
-+ ASSERT_COND(p_FmPort);
-+
-+ SANITY_CHECK_RETURN(p_FmPort->p_FmPortDriverParam, E_INVALID_HANDLE);
-+
-+ p_FmPort->im.h_FmMuram = p_FmPortParams->specificParams.imRxTxParams.h_FmMuram;
-+ p_FmPort->p_FmPortDriverParam->liodnOffset = p_FmPortParams->specificParams.imRxTxParams.liodnOffset;
-+ p_FmPort->im.dataMemId = p_FmPortParams->specificParams.imRxTxParams.dataMemId;
-+ p_FmPort->im.dataMemAttributes = p_FmPortParams->specificParams.imRxTxParams.dataMemAttributes;
-+
-+ p_FmPort->im.fwExtStructsMemId = DEFAULT_PORT_ImfwExtStructsMemId;
-+ p_FmPort->im.fwExtStructsMemAttr = DEFAULT_PORT_ImfwExtStructsMemAttr;
-+
-+ if ((p_FmPort->portType == e_FM_PORT_TYPE_RX) ||
-+ (p_FmPort->portType == e_FM_PORT_TYPE_RX_10G))
-+ {
-+ p_FmPort->im.rxPool.h_BufferPool = p_FmPortParams->specificParams.imRxTxParams.rxPoolParams.h_BufferPool;
-+ p_FmPort->im.rxPool.f_GetBuf = p_FmPortParams->specificParams.imRxTxParams.rxPoolParams.f_GetBuf;
-+ p_FmPort->im.rxPool.f_PutBuf = p_FmPortParams->specificParams.imRxTxParams.rxPoolParams.f_PutBuf;
-+ p_FmPort->im.rxPool.bufferSize = p_FmPortParams->specificParams.imRxTxParams.rxPoolParams.bufferSize;
-+ p_FmPort->im.rxPool.f_PhysToVirt = p_FmPortParams->specificParams.imRxTxParams.rxPoolParams.f_PhysToVirt;
-+ if (!p_FmPort->im.rxPool.f_PhysToVirt)
-+ p_FmPort->im.rxPool.f_PhysToVirt = XX_PhysToVirt;
-+ p_FmPort->im.rxPool.f_VirtToPhys = p_FmPortParams->specificParams.imRxTxParams.rxPoolParams.f_VirtToPhys;
-+ if (!p_FmPort->im.rxPool.f_VirtToPhys)
-+ p_FmPort->im.rxPool.f_VirtToPhys = XX_VirtToPhys;
-+ p_FmPort->im.f_RxStore = p_FmPortParams->specificParams.imRxTxParams.f_RxStore;
-+
-+ p_FmPort->im.mrblr = 0x8000;
-+ while (p_FmPort->im.mrblr)
-+ {
-+ if (p_FmPort->im.rxPool.bufferSize & p_FmPort->im.mrblr)
-+ break;
-+ p_FmPort->im.mrblr >>= 1;
-+ }
-+ if (p_FmPort->im.mrblr != p_FmPort->im.rxPool.bufferSize)
-+ DBG(WARNING, ("Max-Rx-Buffer-Length set to %d", p_FmPort->im.mrblr));
-+ p_FmPort->im.bdRingSize = DEFAULT_PORT_rxBdRingLength;
-+ p_FmPort->exceptions = DEFAULT_PORT_exception;
-+ if (FmIsMaster(p_FmPort->h_Fm))
-+ p_FmPort->polling = FALSE;
-+ else
-+ p_FmPort->polling = TRUE;
-+ p_FmPort->fmanCtrlEventId = (uint8_t)NO_IRQ;
-+ }
-+ else
-+ {
-+ p_FmPort->im.f_TxConf = p_FmPortParams->specificParams.imRxTxParams.f_TxConf;
-+
-+ p_FmPort->im.bdRingSize = DEFAULT_PORT_txBdRingLength;
-+ }
-+}
-+
-+t_Error FmPortImCheckInitParameters(t_FmPort *p_FmPort)
-+{
-+ if ((p_FmPort->portType != e_FM_PORT_TYPE_RX) &&
-+ (p_FmPort->portType != e_FM_PORT_TYPE_RX_10G) &&
-+ (p_FmPort->portType != e_FM_PORT_TYPE_TX) &&
-+ (p_FmPort->portType != e_FM_PORT_TYPE_TX_10G))
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, NO_MSG);
-+
-+ if ((p_FmPort->portType == e_FM_PORT_TYPE_RX) ||
-+ (p_FmPort->portType == e_FM_PORT_TYPE_RX_10G))
-+ {
-+ if (!POWER_OF_2(p_FmPort->im.mrblr))
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("max Rx buffer length must be power of 2!!!"));
-+ if (p_FmPort->im.mrblr < 256)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("max Rx buffer length must at least 256!!!"));
-+ if (p_FmPort->p_FmPortDriverParam->liodnOffset & ~FM_LIODN_OFFSET_MASK)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("liodnOffset is larger than %d", FM_LIODN_OFFSET_MASK+1));
-+ }
-+
-+ return E_OK;
-+}
-+
-+t_Error FmPortImInit(t_FmPort *p_FmPort)
-+{
-+ t_FmImBd *p_Bd=NULL;
-+ t_Handle h_BufContext;
-+ uint64_t tmpPhysBase;
-+ uint16_t log2Num;
-+ uint8_t *p_Data/*, *p_Tmp*/;
-+ int i;
-+ t_Error err;
-+ uint16_t tmpReg16;
-+ uint32_t tmpReg32;
-+
-+ ASSERT_COND(p_FmPort);
-+
-+ p_FmPort->im.p_FmPortImPram =
-+ (t_FmPortImPram *)FM_MURAM_AllocMem(p_FmPort->im.h_FmMuram, sizeof(t_FmPortImPram), IM_PRAM_ALIGN);
-+ if (!p_FmPort->im.p_FmPortImPram)
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, ("Independent-Mode Parameter-RAM!!!"));
-+ WRITE_BLOCK(p_FmPort->im.p_FmPortImPram, 0, sizeof(t_FmPortImPram));
-+
-+ if ((p_FmPort->portType == e_FM_PORT_TYPE_RX) ||
-+ (p_FmPort->portType == e_FM_PORT_TYPE_RX_10G))
-+ {
-+ p_FmPort->im.p_BdRing =
-+ (t_FmImBd *)XX_MallocSmart((uint32_t)(sizeof(t_FmImBd)*p_FmPort->im.bdRingSize),
-+ p_FmPort->im.fwExtStructsMemId,
-+ 4);
-+ if (!p_FmPort->im.p_BdRing)
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, ("Independent-Mode Rx BD ring!!!"));
-+ IOMemSet32(p_FmPort->im.p_BdRing, 0, (uint32_t)(sizeof(t_FmImBd)*p_FmPort->im.bdRingSize));
-+
-+ p_FmPort->im.p_BdShadow = (t_Handle *)XX_Malloc((uint32_t)(sizeof(t_Handle)*p_FmPort->im.bdRingSize));
-+ if (!p_FmPort->im.p_BdShadow)
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, ("Independent-Mode Rx BD shadow!!!"));
-+ memset(p_FmPort->im.p_BdShadow, 0, (uint32_t)(sizeof(t_Handle)*p_FmPort->im.bdRingSize));
-+
-+ /* Initialize the Rx-BD ring */
-+ for (i=0; i<p_FmPort->im.bdRingSize; i++)
-+ {
-+ p_Bd = BD_GET(i);
-+ BD_STATUS_AND_LENGTH_SET (p_Bd, BD_R_E);
-+
-+ if ((p_Data = p_FmPort->im.rxPool.f_GetBuf(p_FmPort->im.rxPool.h_BufferPool, &h_BufContext)) == NULL)
-+ RETURN_ERROR(MAJOR, E_NOT_AVAILABLE, ("Data buffer"));
-+ BdBufferSet(p_FmPort->im.rxPool.f_VirtToPhys, p_Bd, p_Data);
-+ p_FmPort->im.p_BdShadow[i] = h_BufContext;
-+ }
-+
-+ if ((p_FmPort->im.dataMemAttributes & MEMORY_ATTR_CACHEABLE) ||
-+ (p_FmPort->im.fwExtStructsMemAttr & MEMORY_ATTR_CACHEABLE))
-+ WRITE_UINT32(p_FmPort->im.p_FmPortImPram->mode, IM_MODE_GBL | IM_MODE_SET_BO(2));
-+ else
-+ WRITE_UINT32(p_FmPort->im.p_FmPortImPram->mode, IM_MODE_SET_BO(2));
-+
-+ WRITE_UINT32(p_FmPort->im.p_FmPortImPram->rxQdPtr,
-+ (uint32_t)((uint64_t)(XX_VirtToPhys(p_FmPort->im.p_FmPortImPram)) -
-+ p_FmPort->fmMuramPhysBaseAddr + 0x20));
-+
-+ LOG2((uint64_t)p_FmPort->im.mrblr, log2Num);
-+ WRITE_UINT16(p_FmPort->im.p_FmPortImPram->mrblr, log2Num);
-+
-+ /* Initialize Rx QD */
-+ tmpPhysBase = (uint64_t)(XX_VirtToPhys(p_FmPort->im.p_BdRing));
-+ SET_ADDR(&p_FmPort->im.p_FmPortImPram->rxQd.bdRingBase, tmpPhysBase);
-+ WRITE_UINT16(p_FmPort->im.p_FmPortImPram->rxQd.bdRingSize, (uint16_t)(sizeof(t_FmImBd)*p_FmPort->im.bdRingSize));
-+
-+ /* Update the IM PRAM address in the BMI */
-+ WRITE_UINT32(p_FmPort->p_FmPortBmiRegs->rxPortBmiRegs.fmbm_rfqid,
-+ (uint32_t)((uint64_t)(XX_VirtToPhys(p_FmPort->im.p_FmPortImPram)) -
-+ p_FmPort->fmMuramPhysBaseAddr));
-+ if (!p_FmPort->polling || p_FmPort->exceptions)
-+ {
-+ /* Allocate, configure and register interrupts */
-+ err = FmAllocFmanCtrlEventReg(p_FmPort->h_Fm, &p_FmPort->fmanCtrlEventId);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+
-+ ASSERT_COND(!(p_FmPort->fmanCtrlEventId & ~IM_RXQD_FPMEVT_SEL_MASK));
-+ tmpReg16 = (uint16_t)(p_FmPort->fmanCtrlEventId & IM_RXQD_FPMEVT_SEL_MASK);
-+ tmpReg32 = 0;
-+
-+ if (p_FmPort->exceptions & IM_EV_BSY)
-+ {
-+ tmpReg16 |= IM_RXQD_BSYINTM;
-+ tmpReg32 |= IM_EV_BSY;
-+ }
-+ if (!p_FmPort->polling)
-+ {
-+ tmpReg16 |= IM_RXQD_RXFINTM;
-+ tmpReg32 |= IM_EV_RX;
-+ }
-+ WRITE_UINT16(p_FmPort->im.p_FmPortImPram->rxQd.gen, tmpReg16);
-+
-+ FmRegisterFmanCtrlIntr(p_FmPort->h_Fm, p_FmPort->fmanCtrlEventId, ImException , (t_Handle)p_FmPort);
-+
-+ FmSetFmanCtrlIntr(p_FmPort->h_Fm, p_FmPort->fmanCtrlEventId, tmpReg32);
-+ }
-+ else
-+ p_FmPort->fmanCtrlEventId = (uint8_t)NO_IRQ;
-+ }
-+ else
-+ {
-+ p_FmPort->im.p_BdRing = (t_FmImBd *)XX_MallocSmart((uint32_t)(sizeof(t_FmImBd)*p_FmPort->im.bdRingSize), p_FmPort->im.fwExtStructsMemId, 4);
-+ if (!p_FmPort->im.p_BdRing)
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, ("Independent-Mode Tx BD ring!!!"));
-+ IOMemSet32(p_FmPort->im.p_BdRing, 0, (uint32_t)(sizeof(t_FmImBd)*p_FmPort->im.bdRingSize));
-+
-+ p_FmPort->im.p_BdShadow = (t_Handle *)XX_Malloc((uint32_t)(sizeof(t_Handle)*p_FmPort->im.bdRingSize));
-+ if (!p_FmPort->im.p_BdShadow)
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, ("Independent-Mode Rx BD shadow!!!"));
-+ memset(p_FmPort->im.p_BdShadow, 0, (uint32_t)(sizeof(t_Handle)*p_FmPort->im.bdRingSize));
-+ p_FmPort->im.firstBdOfFrameId = IM_ILEGAL_BD_ID;
-+
-+ if ((p_FmPort->im.dataMemAttributes & MEMORY_ATTR_CACHEABLE) ||
-+ (p_FmPort->im.fwExtStructsMemAttr & MEMORY_ATTR_CACHEABLE))
-+ WRITE_UINT32(p_FmPort->im.p_FmPortImPram->mode, IM_MODE_GBL | IM_MODE_SET_BO(2));
-+ else
-+ WRITE_UINT32(p_FmPort->im.p_FmPortImPram->mode, IM_MODE_SET_BO(2));
-+
-+ WRITE_UINT32(p_FmPort->im.p_FmPortImPram->txQdPtr,
-+ (uint32_t)((uint64_t)(XX_VirtToPhys(p_FmPort->im.p_FmPortImPram)) -
-+ p_FmPort->fmMuramPhysBaseAddr + 0x40));
-+
-+ /* Initialize Tx QD */
-+ tmpPhysBase = (uint64_t)(XX_VirtToPhys(p_FmPort->im.p_BdRing));
-+ SET_ADDR(&p_FmPort->im.p_FmPortImPram->txQd.bdRingBase, tmpPhysBase);
-+ WRITE_UINT16(p_FmPort->im.p_FmPortImPram->txQd.bdRingSize, (uint16_t)(sizeof(t_FmImBd)*p_FmPort->im.bdRingSize));
-+
-+ /* Update the IM PRAM address in the BMI */
-+ WRITE_UINT32(p_FmPort->p_FmPortBmiRegs->txPortBmiRegs.fmbm_tcfqid,
-+ (uint32_t)((uint64_t)(XX_VirtToPhys(p_FmPort->im.p_FmPortImPram)) -
-+ p_FmPort->fmMuramPhysBaseAddr));
-+ }
-+
-+
-+ return E_OK;
-+}
-+
-+void FmPortImFree(t_FmPort *p_FmPort)
-+{
-+ uint32_t bdStatus;
-+ uint8_t *p_CurData;
-+
-+ ASSERT_COND(p_FmPort);
-+ ASSERT_COND(p_FmPort->im.p_FmPortImPram);
-+
-+ if ((p_FmPort->portType == e_FM_PORT_TYPE_RX) ||
-+ (p_FmPort->portType == e_FM_PORT_TYPE_RX_10G))
-+ {
-+ if (!p_FmPort->polling || p_FmPort->exceptions)
-+ {
-+ /* Deallocate and unregister interrupts */
-+ FmSetFmanCtrlIntr(p_FmPort->h_Fm, p_FmPort->fmanCtrlEventId, 0);
-+
-+ FmFreeFmanCtrlEventReg(p_FmPort->h_Fm, p_FmPort->fmanCtrlEventId);
-+
-+ WRITE_UINT16(p_FmPort->im.p_FmPortImPram->rxQd.gen, 0);
-+
-+ FmUnregisterFmanCtrlIntr(p_FmPort->h_Fm, p_FmPort->fmanCtrlEventId);
-+ }
-+ /* Try first clean what has received */
-+ FmPortImRx(p_FmPort);
-+
-+ /* Now, get rid of the the empty buffer! */
-+ bdStatus = BD_STATUS_AND_LENGTH(BD_GET(p_FmPort->im.currBdId));
-+
-+ while (bdStatus & BD_R_E) /* while there is data in the Rx BD */
-+ {
-+ p_CurData = BdBufferGet(p_FmPort->im.rxPool.f_PhysToVirt, BD_GET(p_FmPort->im.currBdId));
-+
-+ BdBufferSet(p_FmPort->im.rxPool.f_VirtToPhys, BD_GET(p_FmPort->im.currBdId), NULL);
-+ BD_STATUS_AND_LENGTH_SET(BD_GET(p_FmPort->im.currBdId), 0);
-+
-+ p_FmPort->im.rxPool.f_PutBuf(p_FmPort->im.rxPool.h_BufferPool,
-+ p_CurData,
-+ p_FmPort->im.p_BdShadow[p_FmPort->im.currBdId]);
-+
-+ p_FmPort->im.currBdId = GetNextBdId(p_FmPort, p_FmPort->im.currBdId);
-+ bdStatus = BD_STATUS_AND_LENGTH(BD_GET(p_FmPort->im.currBdId));
-+ }
-+ }
-+ else
-+ TxConf(p_FmPort, e_TX_CONF_TYPE_FLUSH);
-+
-+ FM_MURAM_FreeMem(p_FmPort->im.h_FmMuram, p_FmPort->im.p_FmPortImPram);
-+
-+ if (p_FmPort->im.p_BdShadow)
-+ XX_Free(p_FmPort->im.p_BdShadow);
-+
-+ if (p_FmPort->im.p_BdRing)
-+ XX_FreeSmart(p_FmPort->im.p_BdRing);
-+}
-+
-+
-+t_Error FM_PORT_ConfigIMMaxRxBufLength(t_Handle h_FmPort, uint16_t newVal)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort->imEn, E_INVALID_STATE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort->p_FmPortDriverParam, E_INVALID_HANDLE);
-+
-+ p_FmPort->im.mrblr = newVal;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PORT_ConfigIMRxBdRingLength(t_Handle h_FmPort, uint16_t newVal)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort->imEn, E_INVALID_STATE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort->p_FmPortDriverParam, E_INVALID_HANDLE);
-+
-+ p_FmPort->im.bdRingSize = newVal;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PORT_ConfigIMTxBdRingLength(t_Handle h_FmPort, uint16_t newVal)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort->imEn, E_INVALID_STATE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort->p_FmPortDriverParam, E_INVALID_HANDLE);
-+
-+ p_FmPort->im.bdRingSize = newVal;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PORT_ConfigIMFmanCtrlExternalStructsMemory(t_Handle h_FmPort,
-+ uint8_t memId,
-+ uint32_t memAttributes)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort->imEn, E_INVALID_STATE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort->p_FmPortDriverParam, E_INVALID_HANDLE);
-+
-+ p_FmPort->im.fwExtStructsMemId = memId;
-+ p_FmPort->im.fwExtStructsMemAttr = memAttributes;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PORT_ConfigIMPolling(t_Handle h_FmPort)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort->imEn, E_INVALID_STATE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort->p_FmPortDriverParam, E_INVALID_HANDLE);
-+
-+ if ((p_FmPort->portType != e_FM_PORT_TYPE_RX_10G) && (p_FmPort->portType != e_FM_PORT_TYPE_RX))
-+ RETURN_ERROR(MAJOR, E_INVALID_OPERATION, ("Available for Rx ports only"));
-+
-+ if (!FmIsMaster(p_FmPort->h_Fm))
-+ RETURN_ERROR(MAJOR, E_INVALID_OPERATION, ("Available on master-partition only;"
-+ "in guest-partitions, IM is always in polling!"));
-+
-+ p_FmPort->polling = TRUE;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PORT_SetIMExceptions(t_Handle h_FmPort, e_FmPortExceptions exception, bool enable)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+ t_Error err;
-+ uint16_t tmpReg16;
-+ uint32_t tmpReg32;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort->imEn, E_INVALID_STATE);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmPort->p_FmPortDriverParam, E_INVALID_HANDLE);
-+
-+ if (exception == e_FM_PORT_EXCEPTION_IM_BUSY)
-+ {
-+ if (enable)
-+ {
-+ p_FmPort->exceptions |= IM_EV_BSY;
-+ if (p_FmPort->fmanCtrlEventId == (uint8_t)NO_IRQ)
-+ {
-+ /* Allocate, configure and register interrupts */
-+ err = FmAllocFmanCtrlEventReg(p_FmPort->h_Fm, &p_FmPort->fmanCtrlEventId);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ ASSERT_COND(!(p_FmPort->fmanCtrlEventId & ~IM_RXQD_FPMEVT_SEL_MASK));
-+
-+ FmRegisterFmanCtrlIntr(p_FmPort->h_Fm, p_FmPort->fmanCtrlEventId, ImException, (t_Handle)p_FmPort);
-+ tmpReg16 = (uint16_t)((p_FmPort->fmanCtrlEventId & IM_RXQD_FPMEVT_SEL_MASK) | IM_RXQD_BSYINTM);
-+ tmpReg32 = IM_EV_BSY;
-+ }
-+ else
-+ {
-+ tmpReg16 = (uint16_t)(GET_UINT16(p_FmPort->im.p_FmPortImPram->rxQd.gen) | IM_RXQD_BSYINTM);
-+ tmpReg32 = FmGetFmanCtrlIntr(p_FmPort->h_Fm, p_FmPort->fmanCtrlEventId) | IM_EV_BSY;
-+ }
-+
-+ WRITE_UINT16(p_FmPort->im.p_FmPortImPram->rxQd.gen, tmpReg16);
-+ FmSetFmanCtrlIntr(p_FmPort->h_Fm, p_FmPort->fmanCtrlEventId, tmpReg32);
-+ }
-+ else
-+ {
-+ p_FmPort->exceptions &= ~IM_EV_BSY;
-+ if (!p_FmPort->exceptions && p_FmPort->polling)
-+ {
-+ FmFreeFmanCtrlEventReg(p_FmPort->h_Fm, p_FmPort->fmanCtrlEventId);
-+ FmUnregisterFmanCtrlIntr(p_FmPort->h_Fm, p_FmPort->fmanCtrlEventId);
-+ FmSetFmanCtrlIntr(p_FmPort->h_Fm, p_FmPort->fmanCtrlEventId, 0);
-+ WRITE_UINT16(p_FmPort->im.p_FmPortImPram->rxQd.gen, 0);
-+ p_FmPort->fmanCtrlEventId = (uint8_t)NO_IRQ;
-+ }
-+ else
-+ {
-+ tmpReg16 = (uint16_t)(GET_UINT16(p_FmPort->im.p_FmPortImPram->rxQd.gen) & ~IM_RXQD_BSYINTM);
-+ WRITE_UINT16(p_FmPort->im.p_FmPortImPram->rxQd.gen, tmpReg16);
-+ tmpReg32 = FmGetFmanCtrlIntr(p_FmPort->h_Fm, p_FmPort->fmanCtrlEventId) & ~IM_EV_BSY;
-+ FmSetFmanCtrlIntr(p_FmPort->h_Fm, p_FmPort->fmanCtrlEventId, tmpReg32);
-+ }
-+ }
-+ }
-+ else
-+ RETURN_ERROR(MINOR, E_INVALID_SELECTION, ("Invalid exception."));
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_PORT_ImTx( t_Handle h_FmPort,
-+ uint8_t *p_Data,
-+ uint16_t length,
-+ bool lastBuffer,
-+ t_Handle h_BufContext)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+ uint16_t nextBdId;
-+ uint32_t bdStatus, nextBdStatus;
-+ bool firstBuffer;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort->imEn, E_INVALID_STATE);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmPort->p_FmPortDriverParam, E_INVALID_HANDLE);
-+
-+ bdStatus = BD_STATUS_AND_LENGTH(BD_GET(p_FmPort->im.currBdId));
-+ nextBdId = GetNextBdId(p_FmPort, p_FmPort->im.currBdId);
-+ nextBdStatus = BD_STATUS_AND_LENGTH(BD_GET(nextBdId));
-+
-+ if (!(bdStatus & BD_R_E) && !(nextBdStatus & BD_R_E))
-+ {
-+ /* Confirm the current BD - BD is available */
-+ if ((bdStatus & BD_LENGTH_MASK) && (p_FmPort->im.f_TxConf))
-+ p_FmPort->im.f_TxConf (p_FmPort->h_App,
-+ BdBufferGet(XX_PhysToVirt, BD_GET(p_FmPort->im.currBdId)),
-+ 0,
-+ p_FmPort->im.p_BdShadow[p_FmPort->im.currBdId]);
-+
-+ bdStatus = length;
-+
-+ /* if this is the first BD of a frame */
-+ if (p_FmPort->im.firstBdOfFrameId == IM_ILEGAL_BD_ID)
-+ {
-+ firstBuffer = TRUE;
-+ p_FmPort->im.txFirstBdStatus = (bdStatus | BD_R_E);
-+
-+ if (!lastBuffer)
-+ p_FmPort->im.firstBdOfFrameId = p_FmPort->im.currBdId;
-+ }
-+ else
-+ firstBuffer = FALSE;
-+
-+ BdBufferSet(XX_VirtToPhys, BD_GET(p_FmPort->im.currBdId), p_Data);
-+ p_FmPort->im.p_BdShadow[p_FmPort->im.currBdId] = h_BufContext;
-+
-+ /* deal with last */
-+ if (lastBuffer)
-+ {
-+ /* if single buffer frame */
-+ if (firstBuffer)
-+ BD_STATUS_AND_LENGTH_SET(BD_GET(p_FmPort->im.currBdId), p_FmPort->im.txFirstBdStatus | BD_L);
-+ else
-+ {
-+ /* Set the last BD of the frame */
-+ BD_STATUS_AND_LENGTH_SET (BD_GET(p_FmPort->im.currBdId), (bdStatus | BD_R_E | BD_L));
-+ /* Set the first BD of the frame */
-+ BD_STATUS_AND_LENGTH_SET(BD_GET(p_FmPort->im.firstBdOfFrameId), p_FmPort->im.txFirstBdStatus);
-+ p_FmPort->im.firstBdOfFrameId = IM_ILEGAL_BD_ID;
-+ }
-+ WRITE_UINT16(p_FmPort->im.p_FmPortImPram->txQd.offsetIn, (uint16_t)(GetNextBdId(p_FmPort, p_FmPort->im.currBdId)<<4));
-+ }
-+ else if (!firstBuffer) /* mid frame buffer */
-+ BD_STATUS_AND_LENGTH_SET (BD_GET(p_FmPort->im.currBdId), bdStatus | BD_R_E);
-+
-+ p_FmPort->im.currBdId = GetNextBdId(p_FmPort, p_FmPort->im.currBdId);
-+ }
-+ else
-+ {
-+ /* Discard current frame. Return error. */
-+ if (p_FmPort->im.firstBdOfFrameId != IM_ILEGAL_BD_ID)
-+ {
-+ /* Error: No free BD */
-+ /* Response: Discard current frame. Return error. */
-+ uint16_t cleanBdId = p_FmPort->im.firstBdOfFrameId;
-+
-+ ASSERT_COND(p_FmPort->im.firstBdOfFrameId != p_FmPort->im.currBdId);
-+
-+ /* Since firstInFrame is not NULL, one buffer at least has already been
-+ inserted into the BD ring. Using do-while covers the situation of a
-+ frame spanned throughout the whole Tx BD ring (p_CleanBd is incremented
-+ prior to testing whether or not it's equal to TxBd). */
-+ do
-+ {
-+ BD_STATUS_AND_LENGTH_SET(BD_GET(cleanBdId), 0);
-+ /* Advance BD pointer */
-+ cleanBdId = GetNextBdId(p_FmPort, cleanBdId);
-+ } while (cleanBdId != p_FmPort->im.currBdId);
-+
-+ p_FmPort->im.currBdId = cleanBdId;
-+ p_FmPort->im.firstBdOfFrameId = IM_ILEGAL_BD_ID;
-+ }
-+
-+ return ERROR_CODE(E_FULL);
-+ }
-+
-+ return E_OK;
-+}
-+
-+void FM_PORT_ImTxConf(t_Handle h_FmPort)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+
-+ SANITY_CHECK_RETURN(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN(p_FmPort->imEn, E_INVALID_STATE);
-+ SANITY_CHECK_RETURN(!p_FmPort->p_FmPortDriverParam, E_INVALID_HANDLE);
-+
-+ TxConf(p_FmPort, e_TX_CONF_TYPE_CALLBACK);
-+}
-+
-+t_Error FM_PORT_ImRx(t_Handle h_FmPort)
-+{
-+ t_FmPort *p_FmPort = (t_FmPort*)h_FmPort;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmPort->imEn, E_INVALID_STATE);
-+ SANITY_CHECK_RETURN_ERROR(!p_FmPort->p_FmPortDriverParam, E_INVALID_HANDLE);
-+
-+ return FmPortImRx(p_FmPort);
-+}
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/Port/fman_port.c
-@@ -0,0 +1,1568 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+#include "common/general.h"
-+
-+#include "fman_common.h"
-+#include "fsl_fman_port.h"
-+
-+
-+/* problem Eyal: the following should not be here*/
-+#define NIA_FM_CTL_AC_NO_IPACC_PRE_BMI_ENQ_FRAME 0x00000028
-+
-+static uint32_t get_no_pcd_nia_bmi_ac_enc_frame(struct fman_port_cfg *cfg)
-+{
-+ if (cfg->errata_A006675)
-+ return NIA_ENG_FM_CTL |
-+ NIA_FM_CTL_AC_NO_IPACC_PRE_BMI_ENQ_FRAME;
-+ else
-+ return NIA_ENG_BMI | NIA_BMI_AC_ENQ_FRAME;
-+}
-+
-+static int init_bmi_rx(struct fman_port *port,
-+ struct fman_port_cfg *cfg,
-+ struct fman_port_params *params)
-+{
-+ struct fman_port_rx_bmi_regs *regs = &port->bmi_regs->rx;
-+ uint32_t tmp;
-+
-+ /* Rx Configuration register */
-+ tmp = 0;
-+ if (port->im_en)
-+ tmp |= BMI_PORT_CFG_IM;
-+ else if (cfg->discard_override)
-+ tmp |= BMI_PORT_CFG_FDOVR;
-+ iowrite32be(tmp, &regs->fmbm_rcfg);
-+
-+ /* DMA attributes */
-+ tmp = (uint32_t)cfg->dma_swap_data << BMI_DMA_ATTR_SWP_SHIFT;
-+ if (cfg->dma_ic_stash_on)
-+ tmp |= BMI_DMA_ATTR_IC_STASH_ON;
-+ if (cfg->dma_header_stash_on)
-+ tmp |= BMI_DMA_ATTR_HDR_STASH_ON;
-+ if (cfg->dma_sg_stash_on)
-+ tmp |= BMI_DMA_ATTR_SG_STASH_ON;
-+ if (cfg->dma_write_optimize)
-+ tmp |= BMI_DMA_ATTR_WRITE_OPTIMIZE;
-+ iowrite32be(tmp, &regs->fmbm_rda);
-+
-+ /* Rx FIFO parameters */
-+ tmp = (cfg->rx_pri_elevation / FMAN_PORT_BMI_FIFO_UNITS - 1) <<
-+ BMI_RX_FIFO_PRI_ELEVATION_SHIFT;
-+ tmp |= cfg->rx_fifo_thr / FMAN_PORT_BMI_FIFO_UNITS - 1;
-+ iowrite32be(tmp, &regs->fmbm_rfp);
-+
-+ if (cfg->excessive_threshold_register)
-+ /* always allow access to the extra resources */
-+ iowrite32be(BMI_RX_FIFO_THRESHOLD_ETHE, &regs->fmbm_reth);
-+
-+ /* Frame end data */
-+ tmp = (uint32_t)cfg->checksum_bytes_ignore <<
-+ BMI_RX_FRAME_END_CS_IGNORE_SHIFT;
-+ tmp |= (uint32_t)cfg->rx_cut_end_bytes <<
-+ BMI_RX_FRAME_END_CUT_SHIFT;
-+ if (cfg->errata_A006320)
-+ tmp &= 0xffe0ffff;
-+ iowrite32be(tmp, &regs->fmbm_rfed);
-+
-+ /* Internal context parameters */
-+ tmp = ((uint32_t)cfg->ic_ext_offset / FMAN_PORT_IC_OFFSET_UNITS) <<
-+ BMI_IC_TO_EXT_SHIFT;
-+ tmp |= ((uint32_t)cfg->ic_int_offset / FMAN_PORT_IC_OFFSET_UNITS) <<
-+ BMI_IC_FROM_INT_SHIFT;
-+ tmp |= cfg->ic_size / FMAN_PORT_IC_OFFSET_UNITS;
-+ iowrite32be(tmp, &regs->fmbm_ricp);
-+
-+ /* Internal buffer offset */
-+ tmp = ((uint32_t)cfg->int_buf_start_margin / FMAN_PORT_IC_OFFSET_UNITS)
-+ << BMI_INT_BUF_MARG_SHIFT;
-+ iowrite32be(tmp, &regs->fmbm_rim);
-+
-+ /* External buffer margins */
-+ if (!port->im_en)
-+ {
-+ tmp = (uint32_t)cfg->ext_buf_start_margin <<
-+ BMI_EXT_BUF_MARG_START_SHIFT;
-+ tmp |= (uint32_t)cfg->ext_buf_end_margin;
-+ if (cfg->fmbm_rebm_has_sgd && cfg->no_scatter_gather)
-+ tmp |= BMI_SG_DISABLE;
-+ iowrite32be(tmp, &regs->fmbm_rebm);
-+ }
-+
-+ /* Frame attributes */
-+ tmp = BMI_CMD_RX_MR_DEF;
-+ if (!port->im_en)
-+ {
-+ tmp |= BMI_CMD_ATTR_ORDER;
-+ tmp |= (uint32_t)cfg->color << BMI_CMD_ATTR_COLOR_SHIFT;
-+ if (cfg->sync_req)
-+ tmp |= BMI_CMD_ATTR_SYNC;
-+ }
-+ iowrite32be(tmp, &regs->fmbm_rfca);
-+
-+ /* NIA */
-+ if (port->im_en)
-+ tmp = NIA_ENG_FM_CTL | NIA_FM_CTL_AC_IND_MODE_RX;
-+ else
-+ {
-+ tmp = (uint32_t)cfg->rx_fd_bits << BMI_NEXT_ENG_FD_BITS_SHIFT;
-+ tmp |= get_no_pcd_nia_bmi_ac_enc_frame(cfg);
-+ }
-+ iowrite32be(tmp, &regs->fmbm_rfne);
-+
-+ /* Enqueue NIA */
-+ iowrite32be(NIA_ENG_QMI_ENQ | NIA_ORDER_RESTOR, &regs->fmbm_rfene);
-+
-+ /* Default/error queues */
-+ if (!port->im_en)
-+ {
-+ iowrite32be((params->dflt_fqid & 0x00FFFFFF), &regs->fmbm_rfqid);
-+ iowrite32be((params->err_fqid & 0x00FFFFFF), &regs->fmbm_refqid);
-+ }
-+
-+ /* Discard/error masks */
-+ iowrite32be(params->discard_mask, &regs->fmbm_rfsdm);
-+ iowrite32be(params->err_mask, &regs->fmbm_rfsem);
-+
-+ /* Statistics counters */
-+ tmp = 0;
-+ if (cfg->stats_counters_enable)
-+ tmp = BMI_COUNTERS_EN;
-+ iowrite32be(tmp, &regs->fmbm_rstc);
-+
-+ /* Performance counters */
-+ fman_port_set_perf_cnt_params(port, &cfg->perf_cnt_params);
-+ tmp = 0;
-+ if (cfg->perf_counters_enable)
-+ tmp = BMI_COUNTERS_EN;
-+ iowrite32be(tmp, &regs->fmbm_rpc);
-+
-+ return 0;
-+}
-+
-+static int init_bmi_tx(struct fman_port *port,
-+ struct fman_port_cfg *cfg,
-+ struct fman_port_params *params)
-+{
-+ struct fman_port_tx_bmi_regs *regs = &port->bmi_regs->tx;
-+ uint32_t tmp;
-+
-+ /* Tx Configuration register */
-+ tmp = 0;
-+ if (port->im_en)
-+ tmp |= BMI_PORT_CFG_IM;
-+ iowrite32be(tmp, &regs->fmbm_tcfg);
-+
-+ /* DMA attributes */
-+ tmp = (uint32_t)cfg->dma_swap_data << BMI_DMA_ATTR_SWP_SHIFT;
-+ if (cfg->dma_ic_stash_on)
-+ tmp |= BMI_DMA_ATTR_IC_STASH_ON;
-+ if (cfg->dma_header_stash_on)
-+ tmp |= BMI_DMA_ATTR_HDR_STASH_ON;
-+ if (cfg->dma_sg_stash_on)
-+ tmp |= BMI_DMA_ATTR_SG_STASH_ON;
-+ iowrite32be(tmp, &regs->fmbm_tda);
-+
-+ /* Tx FIFO parameters */
-+ tmp = (cfg->tx_fifo_min_level / FMAN_PORT_BMI_FIFO_UNITS) <<
-+ BMI_TX_FIFO_MIN_FILL_SHIFT;
-+ tmp |= ((uint32_t)cfg->tx_fifo_deq_pipeline_depth - 1) <<
-+ BMI_FIFO_PIPELINE_DEPTH_SHIFT;
-+ tmp |= (uint32_t)(cfg->tx_fifo_low_comf_level /
-+ FMAN_PORT_BMI_FIFO_UNITS - 1);
-+ iowrite32be(tmp, &regs->fmbm_tfp);
-+
-+ /* Frame end data */
-+ tmp = (uint32_t)cfg->checksum_bytes_ignore <<
-+ BMI_FRAME_END_CS_IGNORE_SHIFT;
-+ iowrite32be(tmp, &regs->fmbm_tfed);
-+
-+ /* Internal context parameters */
-+ if (!port->im_en)
-+ {
-+ tmp = ((uint32_t)cfg->ic_ext_offset / FMAN_PORT_IC_OFFSET_UNITS) <<
-+ BMI_IC_TO_EXT_SHIFT;
-+ tmp |= ((uint32_t)cfg->ic_int_offset / FMAN_PORT_IC_OFFSET_UNITS) <<
-+ BMI_IC_FROM_INT_SHIFT;
-+ tmp |= cfg->ic_size / FMAN_PORT_IC_OFFSET_UNITS;
-+ iowrite32be(tmp, &regs->fmbm_ticp);
-+ }
-+ /* Frame attributes */
-+ tmp = BMI_CMD_TX_MR_DEF;
-+ if (port->im_en)
-+ tmp |= BMI_CMD_MR_DEAS;
-+ else
-+ {
-+ tmp |= BMI_CMD_ATTR_ORDER;
-+ tmp |= (uint32_t)cfg->color << BMI_CMD_ATTR_COLOR_SHIFT;
-+ }
-+ iowrite32be(tmp, &regs->fmbm_tfca);
-+
-+ /* Dequeue NIA + enqueue NIA */
-+ if (port->im_en)
-+ {
-+ iowrite32be(NIA_ENG_FM_CTL | NIA_FM_CTL_AC_IND_MODE_TX, &regs->fmbm_tfdne);
-+ iowrite32be(NIA_ENG_FM_CTL | NIA_FM_CTL_AC_IND_MODE_TX, &regs->fmbm_tfene);
-+ }
-+ else
-+ {
-+ iowrite32be(NIA_ENG_QMI_DEQ, &regs->fmbm_tfdne);
-+ iowrite32be(NIA_ENG_QMI_ENQ | NIA_ORDER_RESTOR, &regs->fmbm_tfene);
-+ if (cfg->fmbm_tfne_has_features)
-+ iowrite32be(!params->dflt_fqid ?
-+ BMI_EBD_EN | NIA_BMI_AC_FETCH_ALL_FRAME :
-+ NIA_BMI_AC_FETCH_ALL_FRAME, &regs->fmbm_tfne);
-+ if (!params->dflt_fqid && params->dont_release_buf)
-+ {
-+ iowrite32be(0x00FFFFFF, &regs->fmbm_tcfqid);
-+ iowrite32be(NIA_ENG_BMI | NIA_BMI_AC_TX_RELEASE, &regs->fmbm_tfene);
-+ if (cfg->fmbm_tfne_has_features)
-+ iowrite32be(ioread32be(&regs->fmbm_tfne) & ~BMI_EBD_EN, &regs->fmbm_tfne);
-+ }
-+ }
-+
-+ /* Confirmation/error queues */
-+ if (!port->im_en)
-+ {
-+ if (params->dflt_fqid || !params->dont_release_buf)
-+ iowrite32be(params->dflt_fqid & 0x00FFFFFF, &regs->fmbm_tcfqid);
-+ iowrite32be((params->err_fqid & 0x00FFFFFF), &regs->fmbm_tefqid);
-+ }
-+ /* Statistics counters */
-+ tmp = 0;
-+ if (cfg->stats_counters_enable)
-+ tmp = BMI_COUNTERS_EN;
-+ iowrite32be(tmp, &regs->fmbm_tstc);
-+
-+ /* Performance counters */
-+ fman_port_set_perf_cnt_params(port, &cfg->perf_cnt_params);
-+ tmp = 0;
-+ if (cfg->perf_counters_enable)
-+ tmp = BMI_COUNTERS_EN;
-+ iowrite32be(tmp, &regs->fmbm_tpc);
-+
-+ return 0;
-+}
-+
-+static int init_bmi_oh(struct fman_port *port,
-+ struct fman_port_cfg *cfg,
-+ struct fman_port_params *params)
-+{
-+ struct fman_port_oh_bmi_regs *regs = &port->bmi_regs->oh;
-+ uint32_t tmp;
-+
-+ /* OP Configuration register */
-+ tmp = 0;
-+ if (cfg->discard_override)
-+ tmp |= BMI_PORT_CFG_FDOVR;
-+ iowrite32be(tmp, &regs->fmbm_ocfg);
-+
-+ /* DMA attributes */
-+ tmp = (uint32_t)cfg->dma_swap_data << BMI_DMA_ATTR_SWP_SHIFT;
-+ if (cfg->dma_ic_stash_on)
-+ tmp |= BMI_DMA_ATTR_IC_STASH_ON;
-+ if (cfg->dma_header_stash_on)
-+ tmp |= BMI_DMA_ATTR_HDR_STASH_ON;
-+ if (cfg->dma_sg_stash_on)
-+ tmp |= BMI_DMA_ATTR_SG_STASH_ON;
-+ if (cfg->dma_write_optimize)
-+ tmp |= BMI_DMA_ATTR_WRITE_OPTIMIZE;
-+ iowrite32be(tmp, &regs->fmbm_oda);
-+
-+ /* Tx FIFO parameters */
-+ tmp = ((uint32_t)cfg->tx_fifo_deq_pipeline_depth - 1) <<
-+ BMI_FIFO_PIPELINE_DEPTH_SHIFT;
-+ iowrite32be(tmp, &regs->fmbm_ofp);
-+
-+ /* Internal context parameters */
-+ tmp = ((uint32_t)cfg->ic_ext_offset / FMAN_PORT_IC_OFFSET_UNITS) <<
-+ BMI_IC_TO_EXT_SHIFT;
-+ tmp |= ((uint32_t)cfg->ic_int_offset / FMAN_PORT_IC_OFFSET_UNITS) <<
-+ BMI_IC_FROM_INT_SHIFT;
-+ tmp |= cfg->ic_size / FMAN_PORT_IC_OFFSET_UNITS;
-+ iowrite32be(tmp, &regs->fmbm_oicp);
-+
-+ /* Frame attributes */
-+ tmp = BMI_CMD_OP_MR_DEF;
-+ tmp |= (uint32_t)cfg->color << BMI_CMD_ATTR_COLOR_SHIFT;
-+ if (cfg->sync_req)
-+ tmp |= BMI_CMD_ATTR_SYNC;
-+ if (port->type == E_FMAN_PORT_TYPE_OP)
-+ tmp |= BMI_CMD_ATTR_ORDER;
-+ iowrite32be(tmp, &regs->fmbm_ofca);
-+
-+ /* Internal buffer offset */
-+ tmp = ((uint32_t)cfg->int_buf_start_margin / FMAN_PORT_IC_OFFSET_UNITS)
-+ << BMI_INT_BUF_MARG_SHIFT;
-+ iowrite32be(tmp, &regs->fmbm_oim);
-+
-+ /* Dequeue NIA */
-+ iowrite32be(NIA_ENG_QMI_DEQ, &regs->fmbm_ofdne);
-+
-+ /* NIA and Enqueue NIA */
-+ if (port->type == E_FMAN_PORT_TYPE_HC) {
-+ iowrite32be(NIA_ENG_FM_CTL | NIA_FM_CTL_AC_HC,
-+ &regs->fmbm_ofne);
-+ iowrite32be(NIA_ENG_QMI_ENQ, &regs->fmbm_ofene);
-+ } else {
-+ iowrite32be(get_no_pcd_nia_bmi_ac_enc_frame(cfg),
-+ &regs->fmbm_ofne);
-+ iowrite32be(NIA_ENG_QMI_ENQ | NIA_ORDER_RESTOR,
-+ &regs->fmbm_ofene);
-+ }
-+
-+ /* Default/error queues */
-+ iowrite32be((params->dflt_fqid & 0x00FFFFFF), &regs->fmbm_ofqid);
-+ iowrite32be((params->err_fqid & 0x00FFFFFF), &regs->fmbm_oefqid);
-+
-+ /* Discard/error masks */
-+ if (port->type == E_FMAN_PORT_TYPE_OP) {
-+ iowrite32be(params->discard_mask, &regs->fmbm_ofsdm);
-+ iowrite32be(params->err_mask, &regs->fmbm_ofsem);
-+ }
-+
-+ /* Statistics counters */
-+ tmp = 0;
-+ if (cfg->stats_counters_enable)
-+ tmp = BMI_COUNTERS_EN;
-+ iowrite32be(tmp, &regs->fmbm_ostc);
-+
-+ /* Performance counters */
-+ fman_port_set_perf_cnt_params(port, &cfg->perf_cnt_params);
-+ tmp = 0;
-+ if (cfg->perf_counters_enable)
-+ tmp = BMI_COUNTERS_EN;
-+ iowrite32be(tmp, &regs->fmbm_opc);
-+
-+ return 0;
-+}
-+
-+static int init_qmi(struct fman_port *port,
-+ struct fman_port_cfg *cfg,
-+ struct fman_port_params *params)
-+{
-+ struct fman_port_qmi_regs *regs = port->qmi_regs;
-+ uint32_t tmp;
-+
-+ tmp = 0;
-+ if (cfg->queue_counters_enable)
-+ tmp |= QMI_PORT_CFG_EN_COUNTERS;
-+ iowrite32be(tmp, &regs->fmqm_pnc);
-+
-+ /* Rx port configuration */
-+ if ((port->type == E_FMAN_PORT_TYPE_RX) ||
-+ (port->type == E_FMAN_PORT_TYPE_RX_10G)) {
-+ /* Enqueue NIA */
-+ iowrite32be(NIA_ENG_BMI | NIA_BMI_AC_RELEASE, &regs->fmqm_pnen);
-+ return 0;
-+ }
-+
-+ /* Continue with Tx and O/H port configuration */
-+ if ((port->type == E_FMAN_PORT_TYPE_TX) ||
-+ (port->type == E_FMAN_PORT_TYPE_TX_10G)) {
-+ /* Enqueue NIA */
-+ iowrite32be(NIA_ENG_BMI | NIA_BMI_AC_TX_RELEASE,
-+ &regs->fmqm_pnen);
-+ /* Dequeue NIA */
-+ iowrite32be(NIA_ENG_BMI | NIA_BMI_AC_TX, &regs->fmqm_pndn);
-+ } else {
-+ /* Enqueue NIA */
-+ iowrite32be(NIA_ENG_BMI | NIA_BMI_AC_RELEASE, &regs->fmqm_pnen);
-+ /* Dequeue NIA */
-+ iowrite32be(NIA_ENG_BMI | NIA_BMI_AC_FETCH, &regs->fmqm_pndn);
-+ }
-+
-+ /* Dequeue Configuration register */
-+ tmp = 0;
-+ if (cfg->deq_high_pri)
-+ tmp |= QMI_DEQ_CFG_PRI;
-+
-+ switch (cfg->deq_type) {
-+ case E_FMAN_PORT_DEQ_BY_PRI:
-+ tmp |= QMI_DEQ_CFG_TYPE1;
-+ break;
-+ case E_FMAN_PORT_DEQ_ACTIVE_FQ:
-+ tmp |= QMI_DEQ_CFG_TYPE2;
-+ break;
-+ case E_FMAN_PORT_DEQ_ACTIVE_FQ_NO_ICS:
-+ tmp |= QMI_DEQ_CFG_TYPE3;
-+ break;
-+ default:
-+ return -EINVAL;
-+ }
-+
-+ if (cfg->qmi_deq_options_support) {
-+ if ((port->type == E_FMAN_PORT_TYPE_HC) &&
-+ (cfg->deq_prefetch_opt != E_FMAN_PORT_DEQ_NO_PREFETCH))
-+ return -EINVAL;
-+
-+ switch (cfg->deq_prefetch_opt) {
-+ case E_FMAN_PORT_DEQ_NO_PREFETCH:
-+ break;
-+ case E_FMAN_PORT_DEQ_PART_PREFETCH:
-+ tmp |= QMI_DEQ_CFG_PREFETCH_PARTIAL;
-+ break;
-+ case E_FMAN_PORT_DEQ_FULL_PREFETCH:
-+ tmp |= QMI_DEQ_CFG_PREFETCH_FULL;
-+ break;
-+ default:
-+ return -EINVAL;
-+ }
-+ }
-+ tmp |= (uint32_t)(params->deq_sp & QMI_DEQ_CFG_SP_MASK) <<
-+ QMI_DEQ_CFG_SP_SHIFT;
-+ tmp |= cfg->deq_byte_cnt;
-+ iowrite32be(tmp, &regs->fmqm_pndc);
-+
-+ return 0;
-+}
-+
-+static void get_rx_stats_reg(struct fman_port *port,
-+ enum fman_port_stats_counters counter,
-+ uint32_t **stats_reg)
-+{
-+ struct fman_port_rx_bmi_regs *regs = &port->bmi_regs->rx;
-+
-+ switch (counter) {
-+ case E_FMAN_PORT_STATS_CNT_FRAME:
-+ *stats_reg = &regs->fmbm_rfrc;
-+ break;
-+ case E_FMAN_PORT_STATS_CNT_DISCARD:
-+ *stats_reg = &regs->fmbm_rfdc;
-+ break;
-+ case E_FMAN_PORT_STATS_CNT_DEALLOC_BUF:
-+ *stats_reg = &regs->fmbm_rbdc;
-+ break;
-+ case E_FMAN_PORT_STATS_CNT_RX_BAD_FRAME:
-+ *stats_reg = &regs->fmbm_rfbc;
-+ break;
-+ case E_FMAN_PORT_STATS_CNT_RX_LARGE_FRAME:
-+ *stats_reg = &regs->fmbm_rlfc;
-+ break;
-+ case E_FMAN_PORT_STATS_CNT_RX_OUT_OF_BUF:
-+ *stats_reg = &regs->fmbm_rodc;
-+ break;
-+ case E_FMAN_PORT_STATS_CNT_FILTERED_FRAME:
-+ *stats_reg = &regs->fmbm_rffc;
-+ break;
-+ case E_FMAN_PORT_STATS_CNT_DMA_ERR:
-+ *stats_reg = &regs->fmbm_rfldec;
-+ break;
-+ default:
-+ *stats_reg = NULL;
-+ }
-+}
-+
-+static void get_tx_stats_reg(struct fman_port *port,
-+ enum fman_port_stats_counters counter,
-+ uint32_t **stats_reg)
-+{
-+ struct fman_port_tx_bmi_regs *regs = &port->bmi_regs->tx;
-+
-+ switch (counter) {
-+ case E_FMAN_PORT_STATS_CNT_FRAME:
-+ *stats_reg = &regs->fmbm_tfrc;
-+ break;
-+ case E_FMAN_PORT_STATS_CNT_DISCARD:
-+ *stats_reg = &regs->fmbm_tfdc;
-+ break;
-+ case E_FMAN_PORT_STATS_CNT_DEALLOC_BUF:
-+ *stats_reg = &regs->fmbm_tbdc;
-+ break;
-+ case E_FMAN_PORT_STATS_CNT_LEN_ERR:
-+ *stats_reg = &regs->fmbm_tfledc;
-+ break;
-+ case E_FMAN_PORT_STATS_CNT_UNSUPPORTED_FORMAT:
-+ *stats_reg = &regs->fmbm_tfufdc;
-+ break;
-+ default:
-+ *stats_reg = NULL;
-+ }
-+}
-+
-+static void get_oh_stats_reg(struct fman_port *port,
-+ enum fman_port_stats_counters counter,
-+ uint32_t **stats_reg)
-+{
-+ struct fman_port_oh_bmi_regs *regs = &port->bmi_regs->oh;
-+
-+ switch (counter) {
-+ case E_FMAN_PORT_STATS_CNT_FRAME:
-+ *stats_reg = &regs->fmbm_ofrc;
-+ break;
-+ case E_FMAN_PORT_STATS_CNT_DISCARD:
-+ *stats_reg = &regs->fmbm_ofdc;
-+ break;
-+ case E_FMAN_PORT_STATS_CNT_DEALLOC_BUF:
-+ *stats_reg = &regs->fmbm_obdc;
-+ break;
-+ case E_FMAN_PORT_STATS_CNT_FILTERED_FRAME:
-+ *stats_reg = &regs->fmbm_offc;
-+ break;
-+ case E_FMAN_PORT_STATS_CNT_DMA_ERR:
-+ *stats_reg = &regs->fmbm_ofldec;
-+ break;
-+ case E_FMAN_PORT_STATS_CNT_LEN_ERR:
-+ *stats_reg = &regs->fmbm_ofledc;
-+ break;
-+ case E_FMAN_PORT_STATS_CNT_UNSUPPORTED_FORMAT:
-+ *stats_reg = &regs->fmbm_ofufdc;
-+ break;
-+ case E_FMAN_PORT_STATS_CNT_WRED_DISCARD:
-+ *stats_reg = &regs->fmbm_ofwdc;
-+ break;
-+ default:
-+ *stats_reg = NULL;
-+ }
-+}
-+
-+static void get_rx_perf_reg(struct fman_port *port,
-+ enum fman_port_perf_counters counter,
-+ uint32_t **perf_reg)
-+{
-+ struct fman_port_rx_bmi_regs *regs = &port->bmi_regs->rx;
-+
-+ switch (counter) {
-+ case E_FMAN_PORT_PERF_CNT_CYCLE:
-+ *perf_reg = &regs->fmbm_rccn;
-+ break;
-+ case E_FMAN_PORT_PERF_CNT_TASK_UTIL:
-+ *perf_reg = &regs->fmbm_rtuc;
-+ break;
-+ case E_FMAN_PORT_PERF_CNT_QUEUE_UTIL:
-+ *perf_reg = &regs->fmbm_rrquc;
-+ break;
-+ case E_FMAN_PORT_PERF_CNT_DMA_UTIL:
-+ *perf_reg = &regs->fmbm_rduc;
-+ break;
-+ case E_FMAN_PORT_PERF_CNT_FIFO_UTIL:
-+ *perf_reg = &regs->fmbm_rfuc;
-+ break;
-+ case E_FMAN_PORT_PERF_CNT_RX_PAUSE:
-+ *perf_reg = &regs->fmbm_rpac;
-+ break;
-+ default:
-+ *perf_reg = NULL;
-+ }
-+}
-+
-+static void get_tx_perf_reg(struct fman_port *port,
-+ enum fman_port_perf_counters counter,
-+ uint32_t **perf_reg)
-+{
-+ struct fman_port_tx_bmi_regs *regs = &port->bmi_regs->tx;
-+
-+ switch (counter) {
-+ case E_FMAN_PORT_PERF_CNT_CYCLE:
-+ *perf_reg = &regs->fmbm_tccn;
-+ break;
-+ case E_FMAN_PORT_PERF_CNT_TASK_UTIL:
-+ *perf_reg = &regs->fmbm_ttuc;
-+ break;
-+ case E_FMAN_PORT_PERF_CNT_QUEUE_UTIL:
-+ *perf_reg = &regs->fmbm_ttcquc;
-+ break;
-+ case E_FMAN_PORT_PERF_CNT_DMA_UTIL:
-+ *perf_reg = &regs->fmbm_tduc;
-+ break;
-+ case E_FMAN_PORT_PERF_CNT_FIFO_UTIL:
-+ *perf_reg = &regs->fmbm_tfuc;
-+ break;
-+ default:
-+ *perf_reg = NULL;
-+ }
-+}
-+
-+static void get_oh_perf_reg(struct fman_port *port,
-+ enum fman_port_perf_counters counter,
-+ uint32_t **perf_reg)
-+{
-+ struct fman_port_oh_bmi_regs *regs = &port->bmi_regs->oh;
-+
-+ switch (counter) {
-+ case E_FMAN_PORT_PERF_CNT_CYCLE:
-+ *perf_reg = &regs->fmbm_occn;
-+ break;
-+ case E_FMAN_PORT_PERF_CNT_TASK_UTIL:
-+ *perf_reg = &regs->fmbm_otuc;
-+ break;
-+ case E_FMAN_PORT_PERF_CNT_DMA_UTIL:
-+ *perf_reg = &regs->fmbm_oduc;
-+ break;
-+ case E_FMAN_PORT_PERF_CNT_FIFO_UTIL:
-+ *perf_reg = &regs->fmbm_ofuc;
-+ break;
-+ default:
-+ *perf_reg = NULL;
-+ }
-+}
-+
-+static void get_qmi_counter_reg(struct fman_port *port,
-+ enum fman_port_qmi_counters counter,
-+ uint32_t **queue_reg)
-+{
-+ struct fman_port_qmi_regs *regs = port->qmi_regs;
-+
-+ switch (counter) {
-+ case E_FMAN_PORT_ENQ_TOTAL:
-+ *queue_reg = &regs->fmqm_pnetfc;
-+ break;
-+ case E_FMAN_PORT_DEQ_TOTAL:
-+ if ((port->type == E_FMAN_PORT_TYPE_RX) ||
-+ (port->type == E_FMAN_PORT_TYPE_RX_10G))
-+ /* Counter not available for Rx ports */
-+ *queue_reg = NULL;
-+ else
-+ *queue_reg = &regs->fmqm_pndtfc;
-+ break;
-+ case E_FMAN_PORT_DEQ_FROM_DFLT:
-+ if ((port->type == E_FMAN_PORT_TYPE_RX) ||
-+ (port->type == E_FMAN_PORT_TYPE_RX_10G))
-+ /* Counter not available for Rx ports */
-+ *queue_reg = NULL;
-+ else
-+ *queue_reg = &regs->fmqm_pndfdc;
-+ break;
-+ case E_FMAN_PORT_DEQ_CONFIRM:
-+ if ((port->type == E_FMAN_PORT_TYPE_RX) ||
-+ (port->type == E_FMAN_PORT_TYPE_RX_10G))
-+ /* Counter not available for Rx ports */
-+ *queue_reg = NULL;
-+ else
-+ *queue_reg = &regs->fmqm_pndcc;
-+ break;
-+ default:
-+ *queue_reg = NULL;
-+ }
-+}
-+
-+void fman_port_defconfig(struct fman_port_cfg *cfg, enum fman_port_type type)
-+{
-+ cfg->dma_swap_data = E_FMAN_PORT_DMA_NO_SWAP;
-+ cfg->dma_ic_stash_on = FALSE;
-+ cfg->dma_header_stash_on = FALSE;
-+ cfg->dma_sg_stash_on = FALSE;
-+ cfg->dma_write_optimize = TRUE;
-+ cfg->color = E_FMAN_PORT_COLOR_GREEN;
-+ cfg->discard_override = FALSE;
-+ cfg->checksum_bytes_ignore = 0;
-+ cfg->rx_cut_end_bytes = 4;
-+ cfg->rx_pri_elevation = ((0x3FF + 1) * FMAN_PORT_BMI_FIFO_UNITS);
-+ cfg->rx_fifo_thr = ((0x3FF + 1) * FMAN_PORT_BMI_FIFO_UNITS);
-+ cfg->rx_fd_bits = 0;
-+ cfg->ic_ext_offset = 0;
-+ cfg->ic_int_offset = 0;
-+ cfg->ic_size = 0;
-+ cfg->int_buf_start_margin = 0;
-+ cfg->ext_buf_start_margin = 0;
-+ cfg->ext_buf_end_margin = 0;
-+ cfg->tx_fifo_min_level = 0;
-+ cfg->tx_fifo_low_comf_level = (5 * KILOBYTE);
-+ cfg->stats_counters_enable = TRUE;
-+ cfg->perf_counters_enable = TRUE;
-+ cfg->deq_type = E_FMAN_PORT_DEQ_BY_PRI;
-+
-+ if (type == E_FMAN_PORT_TYPE_HC) {
-+ cfg->sync_req = FALSE;
-+ cfg->deq_prefetch_opt = E_FMAN_PORT_DEQ_NO_PREFETCH;
-+ } else {
-+ cfg->sync_req = TRUE;
-+ cfg->deq_prefetch_opt = E_FMAN_PORT_DEQ_FULL_PREFETCH;
-+ }
-+
-+ if (type == E_FMAN_PORT_TYPE_TX_10G) {
-+ cfg->tx_fifo_deq_pipeline_depth = 4;
-+ cfg->deq_high_pri = TRUE;
-+ cfg->deq_byte_cnt = 0x1400;
-+ } else {
-+ if ((type == E_FMAN_PORT_TYPE_HC) ||
-+ (type == E_FMAN_PORT_TYPE_OP))
-+ cfg->tx_fifo_deq_pipeline_depth = 2;
-+ else
-+ cfg->tx_fifo_deq_pipeline_depth = 1;
-+
-+ cfg->deq_high_pri = FALSE;
-+ cfg->deq_byte_cnt = 0x400;
-+ }
-+ cfg->no_scatter_gather = DEFAULT_FMAN_SP_NO_SCATTER_GATHER;
-+}
-+
-+static uint8_t fman_port_find_bpool(struct fman_port *port, uint8_t bpid)
-+{
-+ uint32_t *bp_reg, tmp;
-+ uint8_t i, id;
-+
-+ /* Find the pool */
-+ bp_reg = port->bmi_regs->rx.fmbm_ebmpi;
-+ for (i = 0;
-+ (i < port->ext_pools_num && (i < FMAN_PORT_MAX_EXT_POOLS_NUM));
-+ i++) {
-+ tmp = ioread32be(&bp_reg[i]);
-+ id = (uint8_t)((tmp & BMI_EXT_BUF_POOL_ID_MASK) >>
-+ BMI_EXT_BUF_POOL_ID_SHIFT);
-+
-+ if (id == bpid)
-+ break;
-+ }
-+
-+ return i;
-+}
-+
-+int fman_port_init(struct fman_port *port,
-+ struct fman_port_cfg *cfg,
-+ struct fman_port_params *params)
-+{
-+ int err;
-+
-+ /* Init BMI registers */
-+ switch (port->type) {
-+ case E_FMAN_PORT_TYPE_RX:
-+ case E_FMAN_PORT_TYPE_RX_10G:
-+ err = init_bmi_rx(port, cfg, params);
-+ break;
-+ case E_FMAN_PORT_TYPE_TX:
-+ case E_FMAN_PORT_TYPE_TX_10G:
-+ err = init_bmi_tx(port, cfg, params);
-+ break;
-+ case E_FMAN_PORT_TYPE_OP:
-+ case E_FMAN_PORT_TYPE_HC:
-+ err = init_bmi_oh(port, cfg, params);
-+ break;
-+ default:
-+ return -EINVAL;
-+ }
-+
-+ if (err)
-+ return err;
-+
-+ /* Init QMI registers */
-+ if (!port->im_en)
-+ {
-+ err = init_qmi(port, cfg, params);
-+ return err;
-+ }
-+ return 0;
-+}
-+
-+int fman_port_enable(struct fman_port *port)
-+{
-+ uint32_t *bmi_cfg_reg, tmp;
-+ bool rx_port;
-+
-+ switch (port->type) {
-+ case E_FMAN_PORT_TYPE_RX:
-+ case E_FMAN_PORT_TYPE_RX_10G:
-+ bmi_cfg_reg = &port->bmi_regs->rx.fmbm_rcfg;
-+ rx_port = TRUE;
-+ break;
-+ case E_FMAN_PORT_TYPE_TX:
-+ case E_FMAN_PORT_TYPE_TX_10G:
-+ bmi_cfg_reg = &port->bmi_regs->tx.fmbm_tcfg;
-+ rx_port = FALSE;
-+ break;
-+ case E_FMAN_PORT_TYPE_OP:
-+ case E_FMAN_PORT_TYPE_HC:
-+ bmi_cfg_reg = &port->bmi_regs->oh.fmbm_ocfg;
-+ rx_port = FALSE;
-+ break;
-+ default:
-+ return -EINVAL;
-+ }
-+
-+ /* Enable QMI */
-+ if (!rx_port) {
-+ tmp = ioread32be(&port->qmi_regs->fmqm_pnc) | QMI_PORT_CFG_EN;
-+ iowrite32be(tmp, &port->qmi_regs->fmqm_pnc);
-+ }
-+
-+ /* Enable BMI */
-+ tmp = ioread32be(bmi_cfg_reg) | BMI_PORT_CFG_EN;
-+ iowrite32be(tmp, bmi_cfg_reg);
-+
-+ return 0;
-+}
-+
-+int fman_port_disable(const struct fman_port *port)
-+{
-+ uint32_t *bmi_cfg_reg, *bmi_status_reg, tmp;
-+ bool rx_port, failure = FALSE;
-+ int count;
-+
-+ switch (port->type) {
-+ case E_FMAN_PORT_TYPE_RX:
-+ case E_FMAN_PORT_TYPE_RX_10G:
-+ bmi_cfg_reg = &port->bmi_regs->rx.fmbm_rcfg;
-+ bmi_status_reg = &port->bmi_regs->rx.fmbm_rst;
-+ rx_port = TRUE;
-+ break;
-+ case E_FMAN_PORT_TYPE_TX:
-+ case E_FMAN_PORT_TYPE_TX_10G:
-+ bmi_cfg_reg = &port->bmi_regs->tx.fmbm_tcfg;
-+ bmi_status_reg = &port->bmi_regs->tx.fmbm_tst;
-+ rx_port = FALSE;
-+ break;
-+ case E_FMAN_PORT_TYPE_OP:
-+ case E_FMAN_PORT_TYPE_HC:
-+ bmi_cfg_reg = &port->bmi_regs->oh.fmbm_ocfg;
-+ bmi_status_reg = &port->bmi_regs->oh.fmbm_ost;
-+ rx_port = FALSE;
-+ break;
-+ default:
-+ return -EINVAL;
-+ }
-+
-+ /* Disable QMI */
-+ if (!rx_port) {
-+ tmp = ioread32be(&port->qmi_regs->fmqm_pnc) & ~QMI_PORT_CFG_EN;
-+ iowrite32be(tmp, &port->qmi_regs->fmqm_pnc);
-+
-+ /* Wait for QMI to finish FD handling */
-+ count = 100;
-+ do {
-+ udelay(10);
-+ tmp = ioread32be(&port->qmi_regs->fmqm_pns);
-+ } while ((tmp & QMI_PORT_STATUS_DEQ_FD_BSY) && --count);
-+
-+ if (count == 0)
-+ {
-+ /* Timeout */
-+ failure = TRUE;
-+ }
-+ }
-+
-+ /* Disable BMI */
-+ tmp = ioread32be(bmi_cfg_reg) & ~BMI_PORT_CFG_EN;
-+ iowrite32be(tmp, bmi_cfg_reg);
-+
-+ /* Wait for graceful stop end */
-+ count = 500;
-+ do {
-+ udelay(10);
-+ tmp = ioread32be(bmi_status_reg);
-+ } while ((tmp & BMI_PORT_STATUS_BSY) && --count);
-+
-+ if (count == 0)
-+ {
-+ /* Timeout */
-+ failure = TRUE;
-+ }
-+
-+ if (failure)
-+ return -EBUSY;
-+
-+ return 0;
-+}
-+
-+int fman_port_set_bpools(const struct fman_port *port,
-+ const struct fman_port_bpools *bp)
-+{
-+ uint32_t tmp, *bp_reg, *bp_depl_reg;
-+ uint8_t i, max_bp_num;
-+ bool grp_depl_used = FALSE, rx_port;
-+
-+ switch (port->type) {
-+ case E_FMAN_PORT_TYPE_RX:
-+ case E_FMAN_PORT_TYPE_RX_10G:
-+ max_bp_num = port->ext_pools_num;
-+ rx_port = TRUE;
-+ bp_reg = port->bmi_regs->rx.fmbm_ebmpi;
-+ bp_depl_reg = &port->bmi_regs->rx.fmbm_mpd;
-+ break;
-+ case E_FMAN_PORT_TYPE_OP:
-+ if (port->fm_rev_maj != 4)
-+ return -EINVAL;
-+ max_bp_num = FMAN_PORT_OBS_EXT_POOLS_NUM;
-+ rx_port = FALSE;
-+ bp_reg = port->bmi_regs->oh.fmbm_oebmpi;
-+ bp_depl_reg = &port->bmi_regs->oh.fmbm_ompd;
-+ break;
-+ default:
-+ return -EINVAL;
-+ }
-+
-+ if (rx_port) {
-+ /* Check buffers are provided in ascending order */
-+ for (i = 0;
-+ (i < (bp->count-1) && (i < FMAN_PORT_MAX_EXT_POOLS_NUM - 1));
-+ i++) {
-+ if (bp->bpool[i].size > bp->bpool[i+1].size)
-+ return -EINVAL;
-+ }
-+ }
-+
-+ /* Set up external buffers pools */
-+ for (i = 0; i < bp->count; i++) {
-+ tmp = BMI_EXT_BUF_POOL_VALID;
-+ tmp |= ((uint32_t)bp->bpool[i].bpid <<
-+ BMI_EXT_BUF_POOL_ID_SHIFT) & BMI_EXT_BUF_POOL_ID_MASK;
-+
-+ if (rx_port) {
-+ if (bp->counters_enable)
-+ tmp |= BMI_EXT_BUF_POOL_EN_COUNTER;
-+
-+ if (bp->bpool[i].is_backup)
-+ tmp |= BMI_EXT_BUF_POOL_BACKUP;
-+
-+ tmp |= (uint32_t)bp->bpool[i].size;
-+ }
-+
-+ iowrite32be(tmp, &bp_reg[i]);
-+ }
-+
-+ /* Clear unused pools */
-+ for (i = bp->count; i < max_bp_num; i++)
-+ iowrite32be(0, &bp_reg[i]);
-+
-+ /* Pools depletion */
-+ tmp = 0;
-+ for (i = 0; i < FMAN_PORT_MAX_EXT_POOLS_NUM; i++) {
-+ if (bp->bpool[i].grp_bp_depleted) {
-+ grp_depl_used = TRUE;
-+ tmp |= 0x80000000 >> i;
-+ }
-+
-+ if (bp->bpool[i].single_bp_depleted)
-+ tmp |= 0x80 >> i;
-+
-+ if (bp->bpool[i].pfc_priorities_en)
-+ tmp |= 0x0100 << i;
-+ }
-+
-+ if (grp_depl_used)
-+ tmp |= ((uint32_t)bp->grp_bp_depleted_num - 1) <<
-+ BMI_POOL_DEP_NUM_OF_POOLS_SHIFT;
-+
-+ iowrite32be(tmp, bp_depl_reg);
-+ return 0;
-+}
-+
-+int fman_port_set_rate_limiter(struct fman_port *port,
-+ struct fman_port_rate_limiter *rate_limiter)
-+{
-+ uint32_t *rate_limit_reg, *rate_limit_scale_reg;
-+ uint32_t granularity, tmp;
-+ uint8_t usec_bit, factor;
-+
-+ switch (port->type) {
-+ case E_FMAN_PORT_TYPE_TX:
-+ case E_FMAN_PORT_TYPE_TX_10G:
-+ rate_limit_reg = &port->bmi_regs->tx.fmbm_trlmt;
-+ rate_limit_scale_reg = &port->bmi_regs->tx.fmbm_trlmts;
-+ granularity = BMI_RATE_LIMIT_GRAN_TX;
-+ break;
-+ case E_FMAN_PORT_TYPE_OP:
-+ rate_limit_reg = &port->bmi_regs->oh.fmbm_orlmt;
-+ rate_limit_scale_reg = &port->bmi_regs->oh.fmbm_orlmts;
-+ granularity = BMI_RATE_LIMIT_GRAN_OP;
-+ break;
-+ default:
-+ return -EINVAL;
-+ }
-+
-+ /* Factor is per 1 usec count */
-+ factor = 1;
-+ usec_bit = rate_limiter->count_1micro_bit;
-+
-+ /* If rate limit is too small for an 1usec factor, adjust timestamp
-+ * scale and multiply the factor */
-+ while (rate_limiter->rate < (granularity / factor)) {
-+ if (usec_bit == 31)
-+ /* Can't configure rate limiter - rate is too small */
-+ return -EINVAL;
-+
-+ usec_bit++;
-+ factor <<= 1;
-+ }
-+
-+ /* Figure out register value. The "while" above quarantees that
-+ * (rate_limiter->rate * factor / granularity) >= 1 */
-+ tmp = (uint32_t)(rate_limiter->rate * factor / granularity - 1);
-+
-+ /* Check rate limit isn't too large */
-+ if (tmp >= BMI_RATE_LIMIT_MAX_RATE_IN_GRAN_UNITS)
-+ return -EINVAL;
-+
-+ /* Check burst size is in allowed range */
-+ if ((rate_limiter->burst_size == 0) ||
-+ (rate_limiter->burst_size >
-+ BMI_RATE_LIMIT_MAX_BURST_SIZE))
-+ return -EINVAL;
-+
-+ tmp |= (uint32_t)(rate_limiter->burst_size - 1) <<
-+ BMI_RATE_LIMIT_MAX_BURST_SHIFT;
-+
-+ if ((port->type == E_FMAN_PORT_TYPE_OP) &&
-+ (port->fm_rev_maj == 4)) {
-+ if (rate_limiter->high_burst_size_gran)
-+ tmp |= BMI_RATE_LIMIT_HIGH_BURST_SIZE_GRAN;
-+ }
-+
-+ iowrite32be(tmp, rate_limit_reg);
-+
-+ /* Set up rate limiter scale register */
-+ tmp = BMI_RATE_LIMIT_SCALE_EN;
-+ tmp |= (31 - (uint32_t)usec_bit) << BMI_RATE_LIMIT_SCALE_TSBS_SHIFT;
-+
-+ if ((port->type == E_FMAN_PORT_TYPE_OP) &&
-+ (port->fm_rev_maj == 4))
-+ tmp |= rate_limiter->rate_factor;
-+
-+ iowrite32be(tmp, rate_limit_scale_reg);
-+
-+ return 0;
-+}
-+
-+int fman_port_delete_rate_limiter(struct fman_port *port)
-+{
-+ uint32_t *rate_limit_scale_reg;
-+
-+ switch (port->type) {
-+ case E_FMAN_PORT_TYPE_TX:
-+ case E_FMAN_PORT_TYPE_TX_10G:
-+ rate_limit_scale_reg = &port->bmi_regs->tx.fmbm_trlmts;
-+ break;
-+ case E_FMAN_PORT_TYPE_OP:
-+ rate_limit_scale_reg = &port->bmi_regs->oh.fmbm_orlmts;
-+ break;
-+ default:
-+ return -EINVAL;
-+ }
-+
-+ iowrite32be(0, rate_limit_scale_reg);
-+ return 0;
-+}
-+
-+int fman_port_set_err_mask(struct fman_port *port, uint32_t err_mask)
-+{
-+ uint32_t *err_mask_reg;
-+
-+ /* Obtain register address */
-+ switch (port->type) {
-+ case E_FMAN_PORT_TYPE_RX:
-+ case E_FMAN_PORT_TYPE_RX_10G:
-+ err_mask_reg = &port->bmi_regs->rx.fmbm_rfsem;
-+ break;
-+ case E_FMAN_PORT_TYPE_OP:
-+ err_mask_reg = &port->bmi_regs->oh.fmbm_ofsem;
-+ break;
-+ default:
-+ return -EINVAL;
-+ }
-+
-+ iowrite32be(err_mask, err_mask_reg);
-+ return 0;
-+}
-+
-+int fman_port_set_discard_mask(struct fman_port *port, uint32_t discard_mask)
-+{
-+ uint32_t *discard_mask_reg;
-+
-+ /* Obtain register address */
-+ switch (port->type) {
-+ case E_FMAN_PORT_TYPE_RX:
-+ case E_FMAN_PORT_TYPE_RX_10G:
-+ discard_mask_reg = &port->bmi_regs->rx.fmbm_rfsdm;
-+ break;
-+ case E_FMAN_PORT_TYPE_OP:
-+ discard_mask_reg = &port->bmi_regs->oh.fmbm_ofsdm;
-+ break;
-+ default:
-+ return -EINVAL;
-+ }
-+
-+ iowrite32be(discard_mask, discard_mask_reg);
-+ return 0;
-+}
-+
-+int fman_port_modify_rx_fd_bits(struct fman_port *port,
-+ uint8_t rx_fd_bits,
-+ bool add)
-+{
-+ uint32_t tmp;
-+
-+ switch (port->type) {
-+ case E_FMAN_PORT_TYPE_RX:
-+ case E_FMAN_PORT_TYPE_RX_10G:
-+ break;
-+ default:
-+ return -EINVAL;
-+ }
-+
-+ tmp = ioread32be(&port->bmi_regs->rx.fmbm_rfne);
-+
-+ if (add)
-+ tmp |= (uint32_t)rx_fd_bits << BMI_NEXT_ENG_FD_BITS_SHIFT;
-+ else
-+ tmp &= ~((uint32_t)rx_fd_bits << BMI_NEXT_ENG_FD_BITS_SHIFT);
-+
-+ iowrite32be(tmp, &port->bmi_regs->rx.fmbm_rfne);
-+ return 0;
-+}
-+
-+int fman_port_set_perf_cnt_params(struct fman_port *port,
-+ struct fman_port_perf_cnt_params *params)
-+{
-+ uint32_t *pcp_reg, tmp;
-+
-+ /* Obtain register address and check parameters are in range */
-+ switch (port->type) {
-+ case E_FMAN_PORT_TYPE_RX:
-+ case E_FMAN_PORT_TYPE_RX_10G:
-+ pcp_reg = &port->bmi_regs->rx.fmbm_rpcp;
-+ if ((params->queue_val == 0) ||
-+ (params->queue_val > MAX_PERFORMANCE_RX_QUEUE_COMP))
-+ return -EINVAL;
-+ break;
-+ case E_FMAN_PORT_TYPE_TX:
-+ case E_FMAN_PORT_TYPE_TX_10G:
-+ pcp_reg = &port->bmi_regs->tx.fmbm_tpcp;
-+ if ((params->queue_val == 0) ||
-+ (params->queue_val > MAX_PERFORMANCE_TX_QUEUE_COMP))
-+ return -EINVAL;
-+ break;
-+ case E_FMAN_PORT_TYPE_OP:
-+ case E_FMAN_PORT_TYPE_HC:
-+ pcp_reg = &port->bmi_regs->oh.fmbm_opcp;
-+ if (params->queue_val != 0)
-+ return -EINVAL;
-+ break;
-+ default:
-+ return -EINVAL;
-+ }
-+
-+ if ((params->task_val == 0) ||
-+ (params->task_val > MAX_PERFORMANCE_TASK_COMP))
-+ return -EINVAL;
-+ if ((params->dma_val == 0) ||
-+ (params->dma_val > MAX_PERFORMANCE_DMA_COMP))
-+ return -EINVAL;
-+ if ((params->fifo_val == 0) ||
-+ ((params->fifo_val / FMAN_PORT_BMI_FIFO_UNITS) >
-+ MAX_PERFORMANCE_FIFO_COMP))
-+ return -EINVAL;
-+ tmp = (uint32_t)(params->task_val - 1) <<
-+ BMI_PERFORMANCE_TASK_COMP_SHIFT;
-+ tmp |= (uint32_t)(params->dma_val - 1) <<
-+ BMI_PERFORMANCE_DMA_COMP_SHIFT;
-+ tmp |= (uint32_t)(params->fifo_val / FMAN_PORT_BMI_FIFO_UNITS - 1);
-+
-+ switch (port->type) {
-+ case E_FMAN_PORT_TYPE_RX:
-+ case E_FMAN_PORT_TYPE_RX_10G:
-+ case E_FMAN_PORT_TYPE_TX:
-+ case E_FMAN_PORT_TYPE_TX_10G:
-+ tmp |= (uint32_t)(params->queue_val - 1) <<
-+ BMI_PERFORMANCE_QUEUE_COMP_SHIFT;
-+ break;
-+ default:
-+ break;
-+ }
-+
-+
-+ iowrite32be(tmp, pcp_reg);
-+ return 0;
-+}
-+
-+int fman_port_set_stats_cnt_mode(struct fman_port *port, bool enable)
-+{
-+ uint32_t *stats_reg, tmp;
-+
-+ switch (port->type) {
-+ case E_FMAN_PORT_TYPE_RX:
-+ case E_FMAN_PORT_TYPE_RX_10G:
-+ stats_reg = &port->bmi_regs->rx.fmbm_rstc;
-+ break;
-+ case E_FMAN_PORT_TYPE_TX:
-+ case E_FMAN_PORT_TYPE_TX_10G:
-+ stats_reg = &port->bmi_regs->tx.fmbm_tstc;
-+ break;
-+ case E_FMAN_PORT_TYPE_OP:
-+ case E_FMAN_PORT_TYPE_HC:
-+ stats_reg = &port->bmi_regs->oh.fmbm_ostc;
-+ break;
-+ default:
-+ return -EINVAL;
-+ }
-+
-+ tmp = ioread32be(stats_reg);
-+
-+ if (enable)
-+ tmp |= BMI_COUNTERS_EN;
-+ else
-+ tmp &= ~BMI_COUNTERS_EN;
-+
-+ iowrite32be(tmp, stats_reg);
-+ return 0;
-+}
-+
-+int fman_port_set_perf_cnt_mode(struct fman_port *port, bool enable)
-+{
-+ uint32_t *stats_reg, tmp;
-+
-+ switch (port->type) {
-+ case E_FMAN_PORT_TYPE_RX:
-+ case E_FMAN_PORT_TYPE_RX_10G:
-+ stats_reg = &port->bmi_regs->rx.fmbm_rpc;
-+ break;
-+ case E_FMAN_PORT_TYPE_TX:
-+ case E_FMAN_PORT_TYPE_TX_10G:
-+ stats_reg = &port->bmi_regs->tx.fmbm_tpc;
-+ break;
-+ case E_FMAN_PORT_TYPE_OP:
-+ case E_FMAN_PORT_TYPE_HC:
-+ stats_reg = &port->bmi_regs->oh.fmbm_opc;
-+ break;
-+ default:
-+ return -EINVAL;
-+ }
-+
-+ tmp = ioread32be(stats_reg);
-+
-+ if (enable)
-+ tmp |= BMI_COUNTERS_EN;
-+ else
-+ tmp &= ~BMI_COUNTERS_EN;
-+
-+ iowrite32be(tmp, stats_reg);
-+ return 0;
-+}
-+
-+int fman_port_set_queue_cnt_mode(struct fman_port *port, bool enable)
-+{
-+ uint32_t tmp;
-+
-+ tmp = ioread32be(&port->qmi_regs->fmqm_pnc);
-+
-+ if (enable)
-+ tmp |= QMI_PORT_CFG_EN_COUNTERS;
-+ else
-+ tmp &= ~QMI_PORT_CFG_EN_COUNTERS;
-+
-+ iowrite32be(tmp, &port->qmi_regs->fmqm_pnc);
-+ return 0;
-+}
-+
-+int fman_port_set_bpool_cnt_mode(struct fman_port *port,
-+ uint8_t bpid,
-+ bool enable)
-+{
-+ uint8_t index;
-+ uint32_t tmp;
-+
-+ switch (port->type) {
-+ case E_FMAN_PORT_TYPE_RX:
-+ case E_FMAN_PORT_TYPE_RX_10G:
-+ break;
-+ default:
-+ return -EINVAL;
-+ }
-+
-+ /* Find the pool */
-+ index = fman_port_find_bpool(port, bpid);
-+ if (index == port->ext_pools_num || index == FMAN_PORT_MAX_EXT_POOLS_NUM)
-+ /* Not found */
-+ return -EINVAL;
-+
-+ tmp = ioread32be(&port->bmi_regs->rx.fmbm_ebmpi[index]);
-+
-+ if (enable)
-+ tmp |= BMI_EXT_BUF_POOL_EN_COUNTER;
-+ else
-+ tmp &= ~BMI_EXT_BUF_POOL_EN_COUNTER;
-+
-+ iowrite32be(tmp, &port->bmi_regs->rx.fmbm_ebmpi[index]);
-+ return 0;
-+}
-+
-+uint32_t fman_port_get_stats_counter(struct fman_port *port,
-+ enum fman_port_stats_counters counter)
-+{
-+ uint32_t *stats_reg, ret_val;
-+
-+ switch (port->type) {
-+ case E_FMAN_PORT_TYPE_RX:
-+ case E_FMAN_PORT_TYPE_RX_10G:
-+ get_rx_stats_reg(port, counter, &stats_reg);
-+ break;
-+ case E_FMAN_PORT_TYPE_TX:
-+ case E_FMAN_PORT_TYPE_TX_10G:
-+ get_tx_stats_reg(port, counter, &stats_reg);
-+ break;
-+ case E_FMAN_PORT_TYPE_OP:
-+ case E_FMAN_PORT_TYPE_HC:
-+ get_oh_stats_reg(port, counter, &stats_reg);
-+ break;
-+ default:
-+ stats_reg = NULL;
-+ }
-+
-+ if (stats_reg == NULL)
-+ return 0;
-+
-+ ret_val = ioread32be(stats_reg);
-+ return ret_val;
-+}
-+
-+void fman_port_set_stats_counter(struct fman_port *port,
-+ enum fman_port_stats_counters counter,
-+ uint32_t value)
-+{
-+ uint32_t *stats_reg;
-+
-+ switch (port->type) {
-+ case E_FMAN_PORT_TYPE_RX:
-+ case E_FMAN_PORT_TYPE_RX_10G:
-+ get_rx_stats_reg(port, counter, &stats_reg);
-+ break;
-+ case E_FMAN_PORT_TYPE_TX:
-+ case E_FMAN_PORT_TYPE_TX_10G:
-+ get_tx_stats_reg(port, counter, &stats_reg);
-+ break;
-+ case E_FMAN_PORT_TYPE_OP:
-+ case E_FMAN_PORT_TYPE_HC:
-+ get_oh_stats_reg(port, counter, &stats_reg);
-+ break;
-+ default:
-+ stats_reg = NULL;
-+ }
-+
-+ if (stats_reg == NULL)
-+ return;
-+
-+ iowrite32be(value, stats_reg);
-+}
-+
-+uint32_t fman_port_get_perf_counter(struct fman_port *port,
-+ enum fman_port_perf_counters counter)
-+{
-+ uint32_t *perf_reg, ret_val;
-+
-+ switch (port->type) {
-+ case E_FMAN_PORT_TYPE_RX:
-+ case E_FMAN_PORT_TYPE_RX_10G:
-+ get_rx_perf_reg(port, counter, &perf_reg);
-+ break;
-+ case E_FMAN_PORT_TYPE_TX:
-+ case E_FMAN_PORT_TYPE_TX_10G:
-+ get_tx_perf_reg(port, counter, &perf_reg);
-+ break;
-+ case E_FMAN_PORT_TYPE_OP:
-+ case E_FMAN_PORT_TYPE_HC:
-+ get_oh_perf_reg(port, counter, &perf_reg);
-+ break;
-+ default:
-+ perf_reg = NULL;
-+ }
-+
-+ if (perf_reg == NULL)
-+ return 0;
-+
-+ ret_val = ioread32be(perf_reg);
-+ return ret_val;
-+}
-+
-+void fman_port_set_perf_counter(struct fman_port *port,
-+ enum fman_port_perf_counters counter,
-+ uint32_t value)
-+{
-+ uint32_t *perf_reg;
-+
-+ switch (port->type) {
-+ case E_FMAN_PORT_TYPE_RX:
-+ case E_FMAN_PORT_TYPE_RX_10G:
-+ get_rx_perf_reg(port, counter, &perf_reg);
-+ break;
-+ case E_FMAN_PORT_TYPE_TX:
-+ case E_FMAN_PORT_TYPE_TX_10G:
-+ get_tx_perf_reg(port, counter, &perf_reg);
-+ break;
-+ case E_FMAN_PORT_TYPE_OP:
-+ case E_FMAN_PORT_TYPE_HC:
-+ get_oh_perf_reg(port, counter, &perf_reg);
-+ break;
-+ default:
-+ perf_reg = NULL;
-+ }
-+
-+ if (perf_reg == NULL)
-+ return;
-+
-+ iowrite32be(value, perf_reg);
-+}
-+
-+uint32_t fman_port_get_qmi_counter(struct fman_port *port,
-+ enum fman_port_qmi_counters counter)
-+{
-+ uint32_t *queue_reg, ret_val;
-+
-+ get_qmi_counter_reg(port, counter, &queue_reg);
-+
-+ if (queue_reg == NULL)
-+ return 0;
-+
-+ ret_val = ioread32be(queue_reg);
-+ return ret_val;
-+}
-+
-+void fman_port_set_qmi_counter(struct fman_port *port,
-+ enum fman_port_qmi_counters counter,
-+ uint32_t value)
-+{
-+ uint32_t *queue_reg;
-+
-+ get_qmi_counter_reg(port, counter, &queue_reg);
-+
-+ if (queue_reg == NULL)
-+ return;
-+
-+ iowrite32be(value, queue_reg);
-+}
-+
-+uint32_t fman_port_get_bpool_counter(struct fman_port *port, uint8_t bpid)
-+{
-+ uint8_t index;
-+ uint32_t ret_val;
-+
-+ switch (port->type) {
-+ case E_FMAN_PORT_TYPE_RX:
-+ case E_FMAN_PORT_TYPE_RX_10G:
-+ break;
-+ default:
-+ return 0;
-+ }
-+
-+ /* Find the pool */
-+ index = fman_port_find_bpool(port, bpid);
-+ if (index == port->ext_pools_num || index == FMAN_PORT_MAX_EXT_POOLS_NUM)
-+ /* Not found */
-+ return 0;
-+
-+ ret_val = ioread32be(&port->bmi_regs->rx.fmbm_acnt[index]);
-+ return ret_val;
-+}
-+
-+void fman_port_set_bpool_counter(struct fman_port *port,
-+ uint8_t bpid,
-+ uint32_t value)
-+{
-+ uint8_t index;
-+
-+ switch (port->type) {
-+ case E_FMAN_PORT_TYPE_RX:
-+ case E_FMAN_PORT_TYPE_RX_10G:
-+ break;
-+ default:
-+ return;
-+ }
-+
-+ /* Find the pool */
-+ index = fman_port_find_bpool(port, bpid);
-+ if (index == port->ext_pools_num || index == FMAN_PORT_MAX_EXT_POOLS_NUM)
-+ /* Not found */
-+ return;
-+
-+ iowrite32be(value, &port->bmi_regs->rx.fmbm_acnt[index]);
-+}
-+
-+int fman_port_add_congestion_grps(struct fman_port *port,
-+ uint32_t grps_map[FMAN_PORT_CG_MAP_NUM])
-+{
-+ int i;
-+ uint32_t tmp, *grp_map_reg;
-+ uint8_t max_grp_map_num;
-+
-+ switch (port->type) {
-+ case E_FMAN_PORT_TYPE_RX:
-+ case E_FMAN_PORT_TYPE_RX_10G:
-+ if (port->fm_rev_maj == 4)
-+ max_grp_map_num = 1;
-+ else
-+ max_grp_map_num = FMAN_PORT_CG_MAP_NUM;
-+ grp_map_reg = port->bmi_regs->rx.fmbm_rcgm;
-+ break;
-+ case E_FMAN_PORT_TYPE_OP:
-+ max_grp_map_num = 1;
-+ if (port->fm_rev_maj != 4)
-+ return -EINVAL;
-+ grp_map_reg = port->bmi_regs->oh.fmbm_ocgm;
-+ break;
-+ default:
-+ return -EINVAL;
-+ }
-+
-+ for (i = (max_grp_map_num - 1); i >= 0; i--) {
-+ if (grps_map[i] == 0)
-+ continue;
-+ tmp = ioread32be(&grp_map_reg[i]);
-+ tmp |= grps_map[i];
-+ iowrite32be(tmp, &grp_map_reg[i]);
-+ }
-+
-+ return 0;
-+}
-+
-+int fman_port_remove_congestion_grps(struct fman_port *port,
-+ uint32_t grps_map[FMAN_PORT_CG_MAP_NUM])
-+{
-+ int i;
-+ uint32_t tmp, *grp_map_reg;
-+ uint8_t max_grp_map_num;
-+
-+ switch (port->type) {
-+ case E_FMAN_PORT_TYPE_RX:
-+ case E_FMAN_PORT_TYPE_RX_10G:
-+ if (port->fm_rev_maj == 4)
-+ max_grp_map_num = 1;
-+ else
-+ max_grp_map_num = FMAN_PORT_CG_MAP_NUM;
-+ grp_map_reg = port->bmi_regs->rx.fmbm_rcgm;
-+ break;
-+ case E_FMAN_PORT_TYPE_OP:
-+ max_grp_map_num = 1;
-+ if (port->fm_rev_maj != 4)
-+ return -EINVAL;
-+ grp_map_reg = port->bmi_regs->oh.fmbm_ocgm;
-+ break;
-+ default:
-+ return -EINVAL;
-+ }
-+
-+ for (i = (max_grp_map_num - 1); i >= 0; i--) {
-+ if (grps_map[i] == 0)
-+ continue;
-+ tmp = ioread32be(&grp_map_reg[i]);
-+ tmp &= ~grps_map[i];
-+ iowrite32be(tmp, &grp_map_reg[i]);
-+ }
-+ return 0;
-+}
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/Rtc/Makefile
-@@ -0,0 +1,15 @@
-+#
-+# Makefile for the Freescale Ethernet controllers
-+#
-+ccflags-y += -DVERSION=\"\"
-+#
-+#Include netcomm SW specific definitions
-+include $(srctree)/drivers/net/ethernet/freescale/sdk_fman/ncsw_config.mk
-+
-+NCSW_FM_INC = $(srctree)/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/inc
-+
-+ccflags-y += -I$(NCSW_FM_INC)
-+
-+obj-y += fsl-ncsw-RTC.o
-+
-+fsl-ncsw-RTC-objs := fm_rtc.o fman_rtc.o
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/Rtc/fm_rtc.c
-@@ -0,0 +1,692 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/******************************************************************************
-+ @File fm_rtc.c
-+
-+ @Description FM RTC driver implementation.
-+
-+ @Cautions None
-+*//***************************************************************************/
-+#include <linux/math64.h>
-+#include "error_ext.h"
-+#include "debug_ext.h"
-+#include "string_ext.h"
-+#include "part_ext.h"
-+#include "xx_ext.h"
-+#include "ncsw_ext.h"
-+
-+#include "fm_rtc.h"
-+#include "fm_common.h"
-+
-+
-+
-+/*****************************************************************************/
-+static t_Error CheckInitParameters(t_FmRtc *p_Rtc)
-+{
-+ struct rtc_cfg *p_RtcDriverParam = p_Rtc->p_RtcDriverParam;
-+ int i;
-+
-+ if ((p_RtcDriverParam->src_clk != E_FMAN_RTC_SOURCE_CLOCK_EXTERNAL) &&
-+ (p_RtcDriverParam->src_clk != E_FMAN_RTC_SOURCE_CLOCK_SYSTEM) &&
-+ (p_RtcDriverParam->src_clk != E_FMAN_RTC_SOURCE_CLOCK_OSCILATOR))
-+ RETURN_ERROR(MAJOR, E_INVALID_CLOCK, ("Source clock undefined"));
-+
-+ if (p_Rtc->outputClockDivisor == 0)
-+ {
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE,
-+ ("Divisor for output clock (should be positive)"));
-+ }
-+
-+ for (i=0; i < FM_RTC_NUM_OF_ALARMS; i++)
-+ {
-+ if ((p_RtcDriverParam->alarm_polarity[i] != E_FMAN_RTC_ALARM_POLARITY_ACTIVE_LOW) &&
-+ (p_RtcDriverParam->alarm_polarity[i] != E_FMAN_RTC_ALARM_POLARITY_ACTIVE_HIGH))
-+ {
-+ RETURN_ERROR(MAJOR, E_INVALID_SELECTION, ("Alarm %d signal polarity", i));
-+ }
-+ }
-+ for (i=0; i < FM_RTC_NUM_OF_EXT_TRIGGERS; i++)
-+ {
-+ if ((p_RtcDriverParam->trigger_polarity[i] != E_FMAN_RTC_TRIGGER_ON_FALLING_EDGE) &&
-+ (p_RtcDriverParam->trigger_polarity[i] != E_FMAN_RTC_TRIGGER_ON_RISING_EDGE))
-+ {
-+ RETURN_ERROR(MAJOR, E_INVALID_SELECTION, ("Trigger %d signal polarity", i));
-+ }
-+ }
-+
-+ return E_OK;
-+}
-+
-+/*****************************************************************************/
-+static void RtcExceptions(t_Handle h_FmRtc)
-+{
-+ t_FmRtc *p_Rtc = (t_FmRtc *)h_FmRtc;
-+ struct rtc_regs *p_MemMap;
-+ register uint32_t events;
-+
-+ ASSERT_COND(p_Rtc);
-+ p_MemMap = p_Rtc->p_MemMap;
-+
-+ events = fman_rtc_check_and_clear_event(p_MemMap);
-+ if (events & FMAN_RTC_TMR_TEVENT_ALM1)
-+ {
-+ if (p_Rtc->alarmParams[0].clearOnExpiration)
-+ {
-+ fman_rtc_set_timer_alarm_l(p_MemMap, 0, 0);
-+ fman_rtc_disable_interupt(p_MemMap, FMAN_RTC_TMR_TEVENT_ALM1);
-+ }
-+ ASSERT_COND(p_Rtc->alarmParams[0].f_AlarmCallback);
-+ p_Rtc->alarmParams[0].f_AlarmCallback(p_Rtc->h_App, 0);
-+ }
-+ if (events & FMAN_RTC_TMR_TEVENT_ALM2)
-+ {
-+ if (p_Rtc->alarmParams[1].clearOnExpiration)
-+ {
-+ fman_rtc_set_timer_alarm_l(p_MemMap, 1, 0);
-+ fman_rtc_disable_interupt(p_MemMap, FMAN_RTC_TMR_TEVENT_ALM2);
-+ }
-+ ASSERT_COND(p_Rtc->alarmParams[1].f_AlarmCallback);
-+ p_Rtc->alarmParams[1].f_AlarmCallback(p_Rtc->h_App, 1);
-+ }
-+ if (events & FMAN_RTC_TMR_TEVENT_PP1)
-+ {
-+ ASSERT_COND(p_Rtc->periodicPulseParams[0].f_PeriodicPulseCallback);
-+ p_Rtc->periodicPulseParams[0].f_PeriodicPulseCallback(p_Rtc->h_App, 0);
-+ }
-+ if (events & FMAN_RTC_TMR_TEVENT_PP2)
-+ {
-+ ASSERT_COND(p_Rtc->periodicPulseParams[1].f_PeriodicPulseCallback);
-+ p_Rtc->periodicPulseParams[1].f_PeriodicPulseCallback(p_Rtc->h_App, 1);
-+ }
-+ if (events & FMAN_RTC_TMR_TEVENT_ETS1)
-+ {
-+ ASSERT_COND(p_Rtc->externalTriggerParams[0].f_ExternalTriggerCallback);
-+ p_Rtc->externalTriggerParams[0].f_ExternalTriggerCallback(p_Rtc->h_App, 0);
-+ }
-+ if (events & FMAN_RTC_TMR_TEVENT_ETS2)
-+ {
-+ ASSERT_COND(p_Rtc->externalTriggerParams[1].f_ExternalTriggerCallback);
-+ p_Rtc->externalTriggerParams[1].f_ExternalTriggerCallback(p_Rtc->h_App, 1);
-+ }
-+}
-+
-+
-+/*****************************************************************************/
-+t_Handle FM_RTC_Config(t_FmRtcParams *p_FmRtcParam)
-+{
-+ t_FmRtc *p_Rtc;
-+
-+ SANITY_CHECK_RETURN_VALUE(p_FmRtcParam, E_NULL_POINTER, NULL);
-+
-+ /* Allocate memory for the FM RTC driver parameters */
-+ p_Rtc = (t_FmRtc *)XX_Malloc(sizeof(t_FmRtc));
-+ if (!p_Rtc)
-+ {
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("FM RTC driver structure"));
-+ return NULL;
-+ }
-+
-+ memset(p_Rtc, 0, sizeof(t_FmRtc));
-+
-+ /* Allocate memory for the FM RTC driver parameters */
-+ p_Rtc->p_RtcDriverParam = (struct rtc_cfg *)XX_Malloc(sizeof(struct rtc_cfg));
-+ if (!p_Rtc->p_RtcDriverParam)
-+ {
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("FM RTC driver parameters"));
-+ XX_Free(p_Rtc);
-+ return NULL;
-+ }
-+
-+ memset(p_Rtc->p_RtcDriverParam, 0, sizeof(struct rtc_cfg));
-+
-+ /* Store RTC configuration parameters */
-+ p_Rtc->h_Fm = p_FmRtcParam->h_Fm;
-+
-+ /* Set default RTC configuration parameters */
-+ fman_rtc_defconfig(p_Rtc->p_RtcDriverParam);
-+
-+ p_Rtc->outputClockDivisor = DEFAULT_OUTPUT_CLOCK_DIVISOR;
-+ p_Rtc->p_RtcDriverParam->bypass = DEFAULT_BYPASS;
-+ p_Rtc->clockPeriodNanoSec = DEFAULT_CLOCK_PERIOD; /* 1 usec */
-+
-+
-+ /* Store RTC parameters in the RTC control structure */
-+ p_Rtc->p_MemMap = (struct rtc_regs *)UINT_TO_PTR(p_FmRtcParam->baseAddress);
-+ p_Rtc->h_App = p_FmRtcParam->h_App;
-+
-+ return p_Rtc;
-+}
-+
-+/*****************************************************************************/
-+t_Error FM_RTC_Init(t_Handle h_FmRtc)
-+{
-+ t_FmRtc *p_Rtc = (t_FmRtc *)h_FmRtc;
-+ struct rtc_cfg *p_RtcDriverParam;
-+ struct rtc_regs *p_MemMap;
-+ uint32_t freqCompensation = 0;
-+ uint64_t tmpDouble;
-+ bool init_freq_comp = FALSE;
-+
-+ p_RtcDriverParam = p_Rtc->p_RtcDriverParam;
-+ p_MemMap = p_Rtc->p_MemMap;
-+
-+ if (CheckInitParameters(p_Rtc)!=E_OK)
-+ RETURN_ERROR(MAJOR, E_CONFLICT,
-+ ("Init Parameters are not Valid"));
-+
-+ /* TODO check that no timestamping MACs are working in this stage. */
-+
-+ /* find source clock frequency in Mhz */
-+ if (p_Rtc->p_RtcDriverParam->src_clk != E_FMAN_RTC_SOURCE_CLOCK_SYSTEM)
-+ p_Rtc->srcClkFreqMhz = p_Rtc->p_RtcDriverParam->ext_src_clk_freq;
-+ else
-+ p_Rtc->srcClkFreqMhz = (uint32_t)(FmGetMacClockFreq(p_Rtc->h_Fm));
-+
-+ /* if timer in Master mode Initialize TMR_CTRL */
-+ /* We want the counter (TMR_CNT) to count in nano-seconds */
-+ if (!p_RtcDriverParam->timer_slave_mode && p_Rtc->p_RtcDriverParam->bypass)
-+ p_Rtc->clockPeriodNanoSec = (1000 / p_Rtc->srcClkFreqMhz);
-+ else
-+ {
-+ /* Initialize TMR_ADD with the initial frequency compensation value:
-+ freqCompensation = (2^32 / frequency ratio) */
-+ /* frequency ratio = sorce clock/rtc clock =
-+ * (p_Rtc->srcClkFreqMhz*1000000))/ 1/(p_Rtc->clockPeriodNanoSec * 1000000000) */
-+ init_freq_comp = TRUE;
-+ freqCompensation = (uint32_t)DIV_CEIL(ACCUMULATOR_OVERFLOW * 1000,
-+ p_Rtc->clockPeriodNanoSec * p_Rtc->srcClkFreqMhz);
-+ }
-+
-+ /* check the legality of the relation between source and destination clocks */
-+ /* should be larger than 1.0001 */
-+ tmpDouble = 10000 * (uint64_t)p_Rtc->clockPeriodNanoSec * (uint64_t)p_Rtc->srcClkFreqMhz;
-+ if ((tmpDouble) <= 10001)
-+ RETURN_ERROR(MAJOR, E_CONFLICT,
-+ ("Invalid relation between source and destination clocks. Should be larger than 1.0001"));
-+
-+ fman_rtc_init(p_RtcDriverParam,
-+ p_MemMap,
-+ FM_RTC_NUM_OF_ALARMS,
-+ FM_RTC_NUM_OF_PERIODIC_PULSES,
-+ FM_RTC_NUM_OF_EXT_TRIGGERS,
-+ init_freq_comp,
-+ freqCompensation,
-+ p_Rtc->outputClockDivisor);
-+
-+ /* Register the FM RTC interrupt */
-+ FmRegisterIntr(p_Rtc->h_Fm, e_FM_MOD_TMR, 0, e_FM_INTR_TYPE_NORMAL, RtcExceptions , p_Rtc);
-+
-+ /* Free parameters structures */
-+ XX_Free(p_Rtc->p_RtcDriverParam);
-+ p_Rtc->p_RtcDriverParam = NULL;
-+
-+ return E_OK;
-+}
-+
-+/*****************************************************************************/
-+t_Error FM_RTC_Free(t_Handle h_FmRtc)
-+{
-+ t_FmRtc *p_Rtc = (t_FmRtc *)h_FmRtc;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Rtc, E_INVALID_HANDLE);
-+
-+ if (p_Rtc->p_RtcDriverParam)
-+ {
-+ XX_Free(p_Rtc->p_RtcDriverParam);
-+ }
-+ else
-+ {
-+ FM_RTC_Disable(h_FmRtc);
-+ }
-+
-+ /* Unregister FM RTC interrupt */
-+ FmUnregisterIntr(p_Rtc->h_Fm, e_FM_MOD_TMR, 0, e_FM_INTR_TYPE_NORMAL);
-+ XX_Free(p_Rtc);
-+
-+ return E_OK;
-+}
-+
-+/*****************************************************************************/
-+t_Error FM_RTC_ConfigSourceClock(t_Handle h_FmRtc,
-+ e_FmSrcClk srcClk,
-+ uint32_t freqInMhz)
-+{
-+ t_FmRtc *p_Rtc = (t_FmRtc *)h_FmRtc;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Rtc, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Rtc->p_RtcDriverParam, E_INVALID_STATE);
-+
-+ p_Rtc->p_RtcDriverParam->src_clk = (enum fman_src_clock)srcClk;
-+ if (srcClk != e_FM_RTC_SOURCE_CLOCK_SYSTEM)
-+ p_Rtc->p_RtcDriverParam->ext_src_clk_freq = freqInMhz;
-+
-+ return E_OK;
-+}
-+
-+/*****************************************************************************/
-+t_Error FM_RTC_ConfigPeriod(t_Handle h_FmRtc, uint32_t period)
-+{
-+ t_FmRtc *p_Rtc = (t_FmRtc *)h_FmRtc;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Rtc, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Rtc->p_RtcDriverParam, E_INVALID_STATE);
-+
-+ p_Rtc->clockPeriodNanoSec = period;
-+
-+ return E_OK;
-+}
-+
-+/*****************************************************************************/
-+t_Error FM_RTC_ConfigFrequencyBypass(t_Handle h_FmRtc, bool enabled)
-+{
-+ t_FmRtc *p_Rtc = (t_FmRtc *)h_FmRtc;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Rtc, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Rtc->p_RtcDriverParam, E_INVALID_STATE);
-+
-+ p_Rtc->p_RtcDriverParam->bypass = enabled;
-+
-+ return E_OK;
-+}
-+
-+/*****************************************************************************/
-+t_Error FM_RTC_ConfigInvertedInputClockPhase(t_Handle h_FmRtc, bool inverted)
-+{
-+ t_FmRtc *p_Rtc = (t_FmRtc *)h_FmRtc;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Rtc, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Rtc->p_RtcDriverParam, E_INVALID_STATE);
-+
-+ p_Rtc->p_RtcDriverParam->invert_input_clk_phase = inverted;
-+
-+ return E_OK;
-+}
-+
-+/*****************************************************************************/
-+t_Error FM_RTC_ConfigInvertedOutputClockPhase(t_Handle h_FmRtc, bool inverted)
-+{
-+ t_FmRtc *p_Rtc = (t_FmRtc *)h_FmRtc;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Rtc, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Rtc->p_RtcDriverParam, E_INVALID_STATE);
-+
-+ p_Rtc->p_RtcDriverParam->invert_output_clk_phase = inverted;
-+
-+ return E_OK;
-+}
-+
-+/*****************************************************************************/
-+t_Error FM_RTC_ConfigOutputClockDivisor(t_Handle h_FmRtc, uint16_t divisor)
-+{
-+ t_FmRtc *p_Rtc = (t_FmRtc *)h_FmRtc;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Rtc, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Rtc->p_RtcDriverParam, E_INVALID_STATE);
-+
-+ p_Rtc->outputClockDivisor = divisor;
-+
-+ return E_OK;
-+}
-+
-+/*****************************************************************************/
-+t_Error FM_RTC_ConfigPulseRealignment(t_Handle h_FmRtc, bool enable)
-+{
-+ t_FmRtc *p_Rtc = (t_FmRtc *)h_FmRtc;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Rtc, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Rtc->p_RtcDriverParam, E_INVALID_STATE);
-+
-+ p_Rtc->p_RtcDriverParam->pulse_realign = enable;
-+
-+ return E_OK;
-+}
-+
-+/*****************************************************************************/
-+t_Error FM_RTC_ConfigAlarmPolarity(t_Handle h_FmRtc,
-+ uint8_t alarmId,
-+ e_FmRtcAlarmPolarity alarmPolarity)
-+{
-+ t_FmRtc *p_Rtc = (t_FmRtc *)h_FmRtc;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Rtc, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Rtc->p_RtcDriverParam, E_INVALID_STATE);
-+
-+ if (alarmId >= FM_RTC_NUM_OF_ALARMS)
-+ RETURN_ERROR(MAJOR, E_INVALID_SELECTION, ("Alarm ID"));
-+
-+ p_Rtc->p_RtcDriverParam->alarm_polarity[alarmId] =
-+ (enum fman_rtc_alarm_polarity)alarmPolarity;
-+
-+ return E_OK;
-+}
-+
-+/*****************************************************************************/
-+t_Error FM_RTC_ConfigExternalTriggerPolarity(t_Handle h_FmRtc,
-+ uint8_t triggerId,
-+ e_FmRtcTriggerPolarity triggerPolarity)
-+{
-+ t_FmRtc *p_Rtc = (t_FmRtc *)h_FmRtc;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Rtc, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Rtc->p_RtcDriverParam, E_INVALID_STATE);
-+
-+ if (triggerId >= FM_RTC_NUM_OF_EXT_TRIGGERS)
-+ {
-+ RETURN_ERROR(MAJOR, E_INVALID_SELECTION, ("External trigger ID"));
-+ }
-+
-+ p_Rtc->p_RtcDriverParam->trigger_polarity[triggerId] =
-+ (enum fman_rtc_trigger_polarity)triggerPolarity;
-+
-+ return E_OK;
-+}
-+
-+/*****************************************************************************/
-+t_Error FM_RTC_Enable(t_Handle h_FmRtc, bool resetClock)
-+{
-+ t_FmRtc *p_Rtc = (t_FmRtc *)h_FmRtc;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Rtc, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Rtc->p_RtcDriverParam, E_INVALID_STATE);
-+
-+ fman_rtc_enable(p_Rtc->p_MemMap, resetClock);
-+ return E_OK;
-+}
-+
-+/*****************************************************************************/
-+t_Error FM_RTC_Disable(t_Handle h_FmRtc)
-+{
-+ t_FmRtc *p_Rtc = (t_FmRtc *)h_FmRtc;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Rtc, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Rtc->p_RtcDriverParam, E_INVALID_STATE);
-+
-+ /* TODO A check must be added here, that no timestamping MAC's
-+ * are working in this stage. */
-+ fman_rtc_disable(p_Rtc->p_MemMap);
-+
-+ return E_OK;
-+}
-+
-+/*****************************************************************************/
-+t_Error FM_RTC_SetClockOffset(t_Handle h_FmRtc, int64_t offset)
-+{
-+ t_FmRtc *p_Rtc = (t_FmRtc *)h_FmRtc;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Rtc, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Rtc->p_RtcDriverParam, E_INVALID_STATE);
-+
-+ fman_rtc_set_timer_offset(p_Rtc->p_MemMap, offset);
-+ return E_OK;
-+}
-+
-+/*****************************************************************************/
-+t_Error FM_RTC_SetAlarm(t_Handle h_FmRtc, t_FmRtcAlarmParams *p_FmRtcAlarmParams)
-+{
-+ t_FmRtc *p_Rtc = (t_FmRtc *)h_FmRtc;
-+ uint64_t tmpAlarm;
-+ bool enable = FALSE;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Rtc, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Rtc->p_RtcDriverParam, E_INVALID_STATE);
-+
-+ if (p_FmRtcAlarmParams->alarmId >= FM_RTC_NUM_OF_ALARMS)
-+ {
-+ RETURN_ERROR(MAJOR, E_INVALID_SELECTION, ("Alarm ID"));
-+ }
-+
-+ if (p_FmRtcAlarmParams->alarmTime < p_Rtc->clockPeriodNanoSec)
-+ RETURN_ERROR(MAJOR, E_INVALID_SELECTION,
-+ ("Alarm time must be equal or larger than RTC period - %d nanoseconds",
-+ p_Rtc->clockPeriodNanoSec));
-+ tmpAlarm = p_FmRtcAlarmParams->alarmTime;
-+ if (do_div(tmpAlarm, p_Rtc->clockPeriodNanoSec))
-+ RETURN_ERROR(MAJOR, E_INVALID_SELECTION,
-+ ("Alarm time must be a multiple of RTC period - %d nanoseconds",
-+ p_Rtc->clockPeriodNanoSec));
-+
-+ if (p_FmRtcAlarmParams->f_AlarmCallback)
-+ {
-+ p_Rtc->alarmParams[p_FmRtcAlarmParams->alarmId].f_AlarmCallback = p_FmRtcAlarmParams->f_AlarmCallback;
-+ p_Rtc->alarmParams[p_FmRtcAlarmParams->alarmId].clearOnExpiration = p_FmRtcAlarmParams->clearOnExpiration;
-+ enable = TRUE;
-+ }
-+
-+ fman_rtc_set_alarm(p_Rtc->p_MemMap, p_FmRtcAlarmParams->alarmId, (unsigned long)tmpAlarm, enable);
-+
-+ return E_OK;
-+}
-+
-+/*****************************************************************************/
-+t_Error FM_RTC_SetPeriodicPulse(t_Handle h_FmRtc, t_FmRtcPeriodicPulseParams *p_FmRtcPeriodicPulseParams)
-+{
-+ t_FmRtc *p_Rtc = (t_FmRtc *)h_FmRtc;
-+ bool enable = FALSE;
-+ uint64_t tmpFiper;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Rtc, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Rtc->p_RtcDriverParam, E_INVALID_STATE);
-+
-+ if (p_FmRtcPeriodicPulseParams->periodicPulseId >= FM_RTC_NUM_OF_PERIODIC_PULSES)
-+ {
-+ RETURN_ERROR(MAJOR, E_INVALID_SELECTION, ("Periodic pulse ID"));
-+ }
-+ if (fman_rtc_is_enabled(p_Rtc->p_MemMap))
-+ RETURN_ERROR(MAJOR, E_INVALID_SELECTION, ("Can't set Periodic pulse when RTC is enabled."));
-+ if (p_FmRtcPeriodicPulseParams->periodicPulsePeriod < p_Rtc->clockPeriodNanoSec)
-+ RETURN_ERROR(MAJOR, E_INVALID_SELECTION,
-+ ("Periodic pulse must be equal or larger than RTC period - %d nanoseconds",
-+ p_Rtc->clockPeriodNanoSec));
-+ tmpFiper = p_FmRtcPeriodicPulseParams->periodicPulsePeriod;
-+ if (do_div(tmpFiper, p_Rtc->clockPeriodNanoSec))
-+ RETURN_ERROR(MAJOR, E_INVALID_SELECTION,
-+ ("Periodic pulse must be a multiple of RTC period - %d nanoseconds",
-+ p_Rtc->clockPeriodNanoSec));
-+ if (tmpFiper & 0xffffffff00000000LL)
-+ RETURN_ERROR(MAJOR, E_INVALID_SELECTION,
-+ ("Periodic pulse/RTC Period must be smaller than 4294967296",
-+ p_Rtc->clockPeriodNanoSec));
-+
-+ if (p_FmRtcPeriodicPulseParams->f_PeriodicPulseCallback)
-+ {
-+ p_Rtc->periodicPulseParams[p_FmRtcPeriodicPulseParams->periodicPulseId].f_PeriodicPulseCallback =
-+ p_FmRtcPeriodicPulseParams->f_PeriodicPulseCallback;
-+ enable = TRUE;
-+ }
-+ fman_rtc_set_periodic_pulse(p_Rtc->p_MemMap, p_FmRtcPeriodicPulseParams->periodicPulseId, (uint32_t)tmpFiper, enable);
-+ return E_OK;
-+}
-+
-+/*****************************************************************************/
-+t_Error FM_RTC_ClearPeriodicPulse(t_Handle h_FmRtc, uint8_t periodicPulseId)
-+{
-+ t_FmRtc *p_Rtc = (t_FmRtc *)h_FmRtc;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Rtc, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Rtc->p_RtcDriverParam, E_INVALID_STATE);
-+
-+ if (periodicPulseId >= FM_RTC_NUM_OF_PERIODIC_PULSES)
-+ {
-+ RETURN_ERROR(MAJOR, E_INVALID_SELECTION, ("Periodic pulse ID"));
-+ }
-+
-+ p_Rtc->periodicPulseParams[periodicPulseId].f_PeriodicPulseCallback = NULL;
-+ fman_rtc_clear_periodic_pulse(p_Rtc->p_MemMap, periodicPulseId);
-+
-+ return E_OK;
-+}
-+
-+/*****************************************************************************/
-+t_Error FM_RTC_SetExternalTrigger(t_Handle h_FmRtc, t_FmRtcExternalTriggerParams *p_FmRtcExternalTriggerParams)
-+{
-+ t_FmRtc *p_Rtc = (t_FmRtc *)h_FmRtc;
-+ bool enable = FALSE;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Rtc, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Rtc->p_RtcDriverParam, E_INVALID_STATE);
-+
-+ if (p_FmRtcExternalTriggerParams->externalTriggerId >= FM_RTC_NUM_OF_EXT_TRIGGERS)
-+ {
-+ RETURN_ERROR(MAJOR, E_INVALID_SELECTION, ("External Trigger ID"));
-+ }
-+
-+ if (p_FmRtcExternalTriggerParams->f_ExternalTriggerCallback)
-+ {
-+ p_Rtc->externalTriggerParams[p_FmRtcExternalTriggerParams->externalTriggerId].f_ExternalTriggerCallback = p_FmRtcExternalTriggerParams->f_ExternalTriggerCallback;
-+ enable = TRUE;
-+ }
-+
-+ fman_rtc_set_ext_trigger(p_Rtc->p_MemMap, p_FmRtcExternalTriggerParams->externalTriggerId, enable, p_FmRtcExternalTriggerParams->usePulseAsInput);
-+ return E_OK;
-+}
-+
-+/*****************************************************************************/
-+t_Error FM_RTC_ClearExternalTrigger(t_Handle h_FmRtc, uint8_t externalTriggerId)
-+{
-+ t_FmRtc *p_Rtc = (t_FmRtc *)h_FmRtc;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Rtc, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Rtc->p_RtcDriverParam, E_INVALID_STATE);
-+
-+ if (externalTriggerId >= FM_RTC_NUM_OF_EXT_TRIGGERS)
-+ RETURN_ERROR(MAJOR, E_INVALID_SELECTION, ("External Trigger ID"));
-+
-+ p_Rtc->externalTriggerParams[externalTriggerId].f_ExternalTriggerCallback = NULL;
-+
-+ fman_rtc_clear_external_trigger(p_Rtc->p_MemMap, externalTriggerId);
-+
-+ return E_OK;
-+}
-+
-+/*****************************************************************************/
-+t_Error FM_RTC_GetExternalTriggerTimeStamp(t_Handle h_FmRtc,
-+ uint8_t triggerId,
-+ uint64_t *p_TimeStamp)
-+{
-+ t_FmRtc *p_Rtc = (t_FmRtc *)h_FmRtc;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Rtc, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Rtc->p_RtcDriverParam, E_INVALID_STATE);
-+
-+ if (triggerId >= FM_RTC_NUM_OF_EXT_TRIGGERS)
-+ RETURN_ERROR(MAJOR, E_INVALID_SELECTION, ("External trigger ID"));
-+
-+ *p_TimeStamp = fman_rtc_get_trigger_stamp(p_Rtc->p_MemMap, triggerId)*p_Rtc->clockPeriodNanoSec;
-+
-+ return E_OK;
-+}
-+
-+/*****************************************************************************/
-+t_Error FM_RTC_GetCurrentTime(t_Handle h_FmRtc, uint64_t *p_Ts)
-+{
-+ t_FmRtc *p_Rtc = (t_FmRtc *)h_FmRtc;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Rtc, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Rtc->p_RtcDriverParam, E_INVALID_STATE);
-+
-+ *p_Ts = fman_rtc_get_timer(p_Rtc->p_MemMap)*p_Rtc->clockPeriodNanoSec;
-+
-+ return E_OK;
-+}
-+
-+/*****************************************************************************/
-+t_Error FM_RTC_SetCurrentTime(t_Handle h_FmRtc, uint64_t ts)
-+{
-+ t_FmRtc *p_Rtc = (t_FmRtc *)h_FmRtc;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Rtc, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Rtc->p_RtcDriverParam, E_INVALID_STATE);
-+
-+ do_div(ts, p_Rtc->clockPeriodNanoSec);
-+ fman_rtc_set_timer(p_Rtc->p_MemMap, (int64_t)ts);
-+
-+ return E_OK;
-+}
-+
-+/*****************************************************************************/
-+t_Error FM_RTC_GetFreqCompensation(t_Handle h_FmRtc, uint32_t *p_Compensation)
-+{
-+ t_FmRtc *p_Rtc = (t_FmRtc *)h_FmRtc;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Rtc, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Rtc->p_RtcDriverParam, E_INVALID_STATE);
-+
-+ *p_Compensation = fman_rtc_get_frequency_compensation(p_Rtc->p_MemMap);
-+
-+ return E_OK;
-+}
-+
-+/*****************************************************************************/
-+t_Error FM_RTC_SetFreqCompensation(t_Handle h_FmRtc, uint32_t freqCompensation)
-+{
-+ t_FmRtc *p_Rtc = (t_FmRtc *)h_FmRtc;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Rtc, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Rtc->p_RtcDriverParam, E_INVALID_STATE);
-+
-+ /* set the new freqCompensation */
-+ fman_rtc_set_frequency_compensation(p_Rtc->p_MemMap, freqCompensation);
-+
-+ return E_OK;
-+}
-+
-+#ifdef CONFIG_PTP_1588_CLOCK_DPAA
-+/*****************************************************************************/
-+t_Error FM_RTC_EnableInterrupt(t_Handle h_FmRtc, uint32_t events)
-+{
-+ t_FmRtc *p_Rtc = (t_FmRtc *)h_FmRtc;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Rtc, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Rtc->p_RtcDriverParam, E_INVALID_STATE);
-+
-+ /* enable interrupt */
-+ fman_rtc_enable_interupt(p_Rtc->p_MemMap, events);
-+
-+ return E_OK;
-+}
-+
-+/*****************************************************************************/
-+t_Error FM_RTC_DisableInterrupt(t_Handle h_FmRtc, uint32_t events)
-+{
-+ t_FmRtc *p_Rtc = (t_FmRtc *)h_FmRtc;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Rtc, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Rtc->p_RtcDriverParam, E_INVALID_STATE);
-+
-+ /* disable interrupt */
-+ fman_rtc_disable_interupt(p_Rtc->p_MemMap, events);
-+
-+ return E_OK;
-+}
-+#endif
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/Rtc/fm_rtc.h
-@@ -0,0 +1,96 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/******************************************************************************
-+ @File fm_rtc.h
-+
-+ @Description Memory map and internal definitions for FM RTC IEEE1588 Timer driver.
-+
-+ @Cautions None
-+*//***************************************************************************/
-+
-+#ifndef __FM_RTC_H__
-+#define __FM_RTC_H__
-+
-+#include "std_ext.h"
-+#include "fm_rtc_ext.h"
-+
-+
-+#define __ERR_MODULE__ MODULE_FM_RTC
-+
-+/* General definitions */
-+
-+#define ACCUMULATOR_OVERFLOW ((uint64_t)(1LL << 32))
-+#define DEFAULT_OUTPUT_CLOCK_DIVISOR 0x00000002
-+#define DEFAULT_BYPASS FALSE
-+#define DEFAULT_CLOCK_PERIOD 1000
-+
-+
-+
-+typedef struct t_FmRtcAlarm
-+{
-+ t_FmRtcExceptionsCallback *f_AlarmCallback;
-+ bool clearOnExpiration;
-+} t_FmRtcAlarm;
-+
-+typedef struct t_FmRtcPeriodicPulse
-+{
-+ t_FmRtcExceptionsCallback *f_PeriodicPulseCallback;
-+} t_FmRtcPeriodicPulse;
-+
-+typedef struct t_FmRtcExternalTrigger
-+{
-+ t_FmRtcExceptionsCallback *f_ExternalTriggerCallback;
-+} t_FmRtcExternalTrigger;
-+
-+
-+/**************************************************************************//**
-+ @Description RTC FM driver control structure.
-+*//***************************************************************************/
-+typedef struct t_FmRtc
-+{
-+ t_Part *p_Part; /**< Pointer to the integration device */
-+ t_Handle h_Fm;
-+ t_Handle h_App; /**< Application handle */
-+ struct rtc_regs *p_MemMap;
-+ uint32_t clockPeriodNanoSec; /**< RTC clock period in nano-seconds (for FS mode) */
-+ uint32_t srcClkFreqMhz;
-+ uint16_t outputClockDivisor; /**< Output clock divisor (for FS mode) */
-+ t_FmRtcAlarm alarmParams[FM_RTC_NUM_OF_ALARMS];
-+ t_FmRtcPeriodicPulse periodicPulseParams[FM_RTC_NUM_OF_PERIODIC_PULSES];
-+ t_FmRtcExternalTrigger externalTriggerParams[FM_RTC_NUM_OF_EXT_TRIGGERS];
-+ struct rtc_cfg *p_RtcDriverParam; /**< RTC Driver parameters (for Init phase) */
-+} t_FmRtc;
-+
-+
-+#endif /* __FM_RTC_H__ */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/Rtc/fman_rtc.c
-@@ -0,0 +1,334 @@
-+/*
-+ * Copyright 2008-2013 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#include "fsl_fman_rtc.h"
-+
-+void fman_rtc_defconfig(struct rtc_cfg *cfg)
-+{
-+ int i;
-+ cfg->src_clk = DEFAULT_SRC_CLOCK;
-+ cfg->invert_input_clk_phase = DEFAULT_INVERT_INPUT_CLK_PHASE;
-+ cfg->invert_output_clk_phase = DEFAULT_INVERT_OUTPUT_CLK_PHASE;
-+ cfg->pulse_realign = DEFAULT_PULSE_REALIGN;
-+ for (i = 0; i < FMAN_RTC_MAX_NUM_OF_ALARMS; i++)
-+ cfg->alarm_polarity[i] = DEFAULT_ALARM_POLARITY;
-+ for (i = 0; i < FMAN_RTC_MAX_NUM_OF_EXT_TRIGGERS; i++)
-+ cfg->trigger_polarity[i] = DEFAULT_TRIGGER_POLARITY;
-+}
-+
-+uint32_t fman_rtc_get_events(struct rtc_regs *regs)
-+{
-+ return ioread32be(&regs->tmr_tevent);
-+}
-+
-+uint32_t fman_rtc_get_event(struct rtc_regs *regs, uint32_t ev_mask)
-+{
-+ return ioread32be(&regs->tmr_tevent) & ev_mask;
-+}
-+
-+uint32_t fman_rtc_get_interrupt_mask(struct rtc_regs *regs)
-+{
-+ return ioread32be(&regs->tmr_temask);
-+}
-+
-+void fman_rtc_set_interrupt_mask(struct rtc_regs *regs, uint32_t mask)
-+{
-+ iowrite32be(mask, &regs->tmr_temask);
-+}
-+
-+void fman_rtc_ack_event(struct rtc_regs *regs, uint32_t events)
-+{
-+ iowrite32be(events, &regs->tmr_tevent);
-+}
-+
-+uint32_t fman_rtc_check_and_clear_event(struct rtc_regs *regs)
-+{
-+ uint32_t event;
-+
-+ event = ioread32be(&regs->tmr_tevent);
-+ event &= ioread32be(&regs->tmr_temask);
-+
-+ if (event)
-+ iowrite32be(event, &regs->tmr_tevent);
-+ return event;
-+}
-+
-+uint32_t fman_rtc_get_frequency_compensation(struct rtc_regs *regs)
-+{
-+ return ioread32be(&regs->tmr_add);
-+}
-+
-+void fman_rtc_set_frequency_compensation(struct rtc_regs *regs, uint32_t val)
-+{
-+ iowrite32be(val, &regs->tmr_add);
-+}
-+
-+void fman_rtc_enable_interupt(struct rtc_regs *regs, uint32_t events)
-+{
-+ fman_rtc_set_interrupt_mask(regs, fman_rtc_get_interrupt_mask(regs) | events);
-+}
-+
-+void fman_rtc_disable_interupt(struct rtc_regs *regs, uint32_t events)
-+{
-+ fman_rtc_set_interrupt_mask(regs, fman_rtc_get_interrupt_mask(regs) & ~events);
-+}
-+
-+void fman_rtc_set_timer_alarm_l(struct rtc_regs *regs, int index, uint32_t val)
-+{
-+ iowrite32be(val, &regs->tmr_alarm[index].tmr_alarm_l);
-+}
-+
-+void fman_rtc_set_timer_fiper(struct rtc_regs *regs, int index, uint32_t val)
-+{
-+ iowrite32be(val, &regs->tmr_fiper[index]);
-+}
-+
-+void fman_rtc_set_timer_alarm(struct rtc_regs *regs, int index, int64_t val)
-+{
-+ iowrite32be((uint32_t)val, &regs->tmr_alarm[index].tmr_alarm_l);
-+ iowrite32be((uint32_t)(val >> 32), &regs->tmr_alarm[index].tmr_alarm_h);
-+}
-+
-+void fman_rtc_set_timer_offset(struct rtc_regs *regs, int64_t val)
-+{
-+ iowrite32be((uint32_t)val, &regs->tmr_off_l);
-+ iowrite32be((uint32_t)(val >> 32), &regs->tmr_off_h);
-+}
-+
-+uint64_t fman_rtc_get_trigger_stamp(struct rtc_regs *regs, int id)
-+{
-+ uint64_t time;
-+ /* TMR_CNT_L must be read first to get an accurate value */
-+ time = (uint64_t)ioread32be(&regs->tmr_etts[id].tmr_etts_l);
-+ time |= ((uint64_t)ioread32be(&regs->tmr_etts[id].tmr_etts_h)
-+ << 32);
-+
-+ return time;
-+}
-+
-+uint32_t fman_rtc_get_timer_ctrl(struct rtc_regs *regs)
-+{
-+ return ioread32be(&regs->tmr_ctrl);
-+}
-+
-+void fman_rtc_set_timer_ctrl(struct rtc_regs *regs, uint32_t val)
-+{
-+ iowrite32be(val, &regs->tmr_ctrl);
-+}
-+
-+void fman_rtc_timers_soft_reset(struct rtc_regs *regs)
-+{
-+ fman_rtc_set_timer_ctrl(regs, FMAN_RTC_TMR_CTRL_TMSR);
-+ udelay(10);
-+ fman_rtc_set_timer_ctrl(regs, 0);
-+}
-+
-+void fman_rtc_init(struct rtc_cfg *cfg, struct rtc_regs *regs, int num_alarms,
-+ int num_fipers, int num_ext_triggers, bool init_freq_comp,
-+ uint32_t freq_compensation, uint32_t output_clock_divisor)
-+{
-+ uint32_t tmr_ctrl;
-+ int i;
-+
-+ fman_rtc_timers_soft_reset(regs);
-+
-+ /* Set the source clock */
-+ switch (cfg->src_clk) {
-+ case E_FMAN_RTC_SOURCE_CLOCK_SYSTEM:
-+ tmr_ctrl = FMAN_RTC_TMR_CTRL_CKSEL_MAC_CLK;
-+ break;
-+ case E_FMAN_RTC_SOURCE_CLOCK_OSCILATOR:
-+ tmr_ctrl = FMAN_RTC_TMR_CTRL_CKSEL_OSC_CLK;
-+ break;
-+ default:
-+ /* Use a clock from the External TMR reference clock.*/
-+ tmr_ctrl = FMAN_RTC_TMR_CTRL_CKSEL_EXT_CLK;
-+ break;
-+ }
-+
-+ /* whatever period the user picked, the timestamp will advance in '1'
-+ * every time the period passed. */
-+ tmr_ctrl |= ((1 << FMAN_RTC_TMR_CTRL_TCLK_PERIOD_SHIFT) &
-+ FMAN_RTC_TMR_CTRL_TCLK_PERIOD_MASK);
-+
-+ if (cfg->invert_input_clk_phase)
-+ tmr_ctrl |= FMAN_RTC_TMR_CTRL_CIPH;
-+ if (cfg->invert_output_clk_phase)
-+ tmr_ctrl |= FMAN_RTC_TMR_CTRL_COPH;
-+
-+ for (i = 0; i < num_alarms; i++) {
-+ if (cfg->alarm_polarity[i] ==
-+ E_FMAN_RTC_ALARM_POLARITY_ACTIVE_LOW)
-+ tmr_ctrl |= (FMAN_RTC_TMR_CTRL_ALMP1 >> i);
-+ }
-+
-+ for (i = 0; i < num_ext_triggers; i++)
-+ if (cfg->trigger_polarity[i] ==
-+ E_FMAN_RTC_TRIGGER_ON_FALLING_EDGE)
-+ tmr_ctrl |= (FMAN_RTC_TMR_CTRL_ETEP1 << i);
-+
-+ if (!cfg->timer_slave_mode && cfg->bypass)
-+ tmr_ctrl |= FMAN_RTC_TMR_CTRL_BYP;
-+
-+ fman_rtc_set_timer_ctrl(regs, tmr_ctrl);
-+ if (init_freq_comp)
-+ fman_rtc_set_frequency_compensation(regs, freq_compensation);
-+
-+ /* Clear TMR_ALARM registers */
-+ for (i = 0; i < num_alarms; i++)
-+ fman_rtc_set_timer_alarm(regs, i, 0xFFFFFFFFFFFFFFFFLL);
-+
-+ /* Clear TMR_TEVENT */
-+ fman_rtc_ack_event(regs, FMAN_RTC_TMR_TEVENT_ALL);
-+
-+ /* Initialize TMR_TEMASK */
-+ fman_rtc_set_interrupt_mask(regs, 0);
-+
-+ /* Clear TMR_FIPER registers */
-+ for (i = 0; i < num_fipers; i++)
-+ fman_rtc_set_timer_fiper(regs, i, 0xFFFFFFFF);
-+
-+ /* Initialize TMR_PRSC */
-+ iowrite32be(output_clock_divisor, &regs->tmr_prsc);
-+
-+ /* Clear TMR_OFF */
-+ fman_rtc_set_timer_offset(regs, 0);
-+}
-+
-+bool fman_rtc_is_enabled(struct rtc_regs *regs)
-+{
-+ return (bool)(fman_rtc_get_timer_ctrl(regs) & FMAN_RTC_TMR_CTRL_TE);
-+}
-+
-+void fman_rtc_enable(struct rtc_regs *regs, bool reset_clock)
-+{
-+ uint32_t tmr_ctrl = fman_rtc_get_timer_ctrl(regs);
-+
-+ /* TODO check that no timestamping MACs are working in this stage. */
-+ if (reset_clock) {
-+ fman_rtc_set_timer_ctrl(regs, (tmr_ctrl | FMAN_RTC_TMR_CTRL_TMSR));
-+
-+ udelay(10);
-+ /* Clear TMR_OFF */
-+ fman_rtc_set_timer_offset(regs, 0);
-+ }
-+
-+ fman_rtc_set_timer_ctrl(regs, (tmr_ctrl | FMAN_RTC_TMR_CTRL_TE));
-+}
-+
-+void fman_rtc_disable(struct rtc_regs *regs)
-+{
-+ fman_rtc_set_timer_ctrl(regs, (fman_rtc_get_timer_ctrl(regs)
-+ & ~(FMAN_RTC_TMR_CTRL_TE)));
-+}
-+
-+void fman_rtc_clear_periodic_pulse(struct rtc_regs *regs, int id)
-+{
-+ uint32_t tmp_reg;
-+ if (id == 0)
-+ tmp_reg = FMAN_RTC_TMR_TEVENT_PP1;
-+ else
-+ tmp_reg = FMAN_RTC_TMR_TEVENT_PP2;
-+ fman_rtc_disable_interupt(regs, tmp_reg);
-+
-+ tmp_reg = fman_rtc_get_timer_ctrl(regs);
-+ if (tmp_reg & FMAN_RTC_TMR_CTRL_FS)
-+ fman_rtc_set_timer_ctrl(regs, tmp_reg & ~FMAN_RTC_TMR_CTRL_FS);
-+
-+ fman_rtc_set_timer_fiper(regs, id, 0xFFFFFFFF);
-+}
-+
-+void fman_rtc_clear_external_trigger(struct rtc_regs *regs, int id)
-+{
-+ uint32_t tmpReg, tmp_ctrl;
-+
-+ if (id == 0)
-+ tmpReg = FMAN_RTC_TMR_TEVENT_ETS1;
-+ else
-+ tmpReg = FMAN_RTC_TMR_TEVENT_ETS2;
-+ fman_rtc_disable_interupt(regs, tmpReg);
-+
-+ if (id == 0)
-+ tmpReg = FMAN_RTC_TMR_CTRL_PP1L;
-+ else
-+ tmpReg = FMAN_RTC_TMR_CTRL_PP2L;
-+ tmp_ctrl = fman_rtc_get_timer_ctrl(regs);
-+ if (tmp_ctrl & tmpReg)
-+ fman_rtc_set_timer_ctrl(regs, tmp_ctrl & ~tmpReg);
-+}
-+
-+void fman_rtc_set_alarm(struct rtc_regs *regs, int id, uint32_t val, bool enable)
-+{
-+ uint32_t tmpReg;
-+ fman_rtc_set_timer_alarm(regs, id, val);
-+ if (enable) {
-+ if (id == 0)
-+ tmpReg = FMAN_RTC_TMR_TEVENT_ALM1;
-+ else
-+ tmpReg = FMAN_RTC_TMR_TEVENT_ALM2;
-+ fman_rtc_enable_interupt(regs, tmpReg);
-+ }
-+}
-+
-+void fman_rtc_set_periodic_pulse(struct rtc_regs *regs, int id, uint32_t val,
-+ bool enable)
-+{
-+ uint32_t tmpReg;
-+ fman_rtc_set_timer_fiper(regs, id, val);
-+ if (enable) {
-+ if (id == 0)
-+ tmpReg = FMAN_RTC_TMR_TEVENT_PP1;
-+ else
-+ tmpReg = FMAN_RTC_TMR_TEVENT_PP2;
-+ fman_rtc_enable_interupt(regs, tmpReg);
-+ }
-+}
-+
-+void fman_rtc_set_ext_trigger(struct rtc_regs *regs, int id, bool enable,
-+ bool use_pulse_as_input)
-+{
-+ uint32_t tmpReg;
-+ if (enable) {
-+ if (id == 0)
-+ tmpReg = FMAN_RTC_TMR_TEVENT_ETS1;
-+ else
-+ tmpReg = FMAN_RTC_TMR_TEVENT_ETS2;
-+ fman_rtc_enable_interupt(regs, tmpReg);
-+ }
-+ if (use_pulse_as_input) {
-+ if (id == 0)
-+ tmpReg = FMAN_RTC_TMR_CTRL_PP1L;
-+ else
-+ tmpReg = FMAN_RTC_TMR_CTRL_PP2L;
-+ fman_rtc_set_timer_ctrl(regs, fman_rtc_get_timer_ctrl(regs) | tmpReg);
-+ }
-+}
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/SP/Makefile
-@@ -0,0 +1,15 @@
-+#
-+# Makefile for the Freescale Ethernet controllers
-+#
-+ccflags-y += -DVERSION=\"\"
-+#
-+#Include netcomm SW specific definitions
-+include $(srctree)/drivers/net/ethernet/freescale/sdk_fman/ncsw_config.mk
-+
-+NCSW_FM_INC = $(srctree)/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/inc
-+
-+ccflags-y += -I$(NCSW_FM_INC)
-+
-+obj-y += fsl-ncsw-sp.o
-+
-+fsl-ncsw-sp-objs := fm_sp.o fman_sp.o
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/SP/fm_sp.c
-@@ -0,0 +1,757 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/******************************************************************************
-+ @File fm_sp.c
-+
-+ @Description FM PCD Storage profile ...
-+*//***************************************************************************/
-+
-+#include "std_ext.h"
-+#include "error_ext.h"
-+#include "string_ext.h"
-+#include "debug_ext.h"
-+#include "net_ext.h"
-+
-+#include "fm_vsp_ext.h"
-+#include "fm_sp.h"
-+#include "fm_common.h"
-+#include "fsl_fman_sp.h"
-+
-+
-+#if (DPAA_VERSION >= 11)
-+static t_Error CheckParamsGeneratedInternally(t_FmVspEntry *p_FmVspEntry)
-+{
-+ t_Error err = E_OK;
-+
-+ if ((err = FmSpCheckIntContextParams(&p_FmVspEntry->intContext))!= E_OK)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ if ((err = FmSpCheckBufMargins(&p_FmVspEntry->bufMargins)) != E_OK)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ return err;
-+
-+}
-+
-+static t_Error CheckParams(t_FmVspEntry *p_FmVspEntry)
-+{
-+ t_Error err = E_OK;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmVspEntry, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmVspEntry->p_FmVspEntryDriverParams, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmVspEntry->h_Fm, E_INVALID_HANDLE);
-+
-+ if ((err = FmSpCheckBufPoolsParams(&p_FmVspEntry->p_FmVspEntryDriverParams->extBufPools,
-+ p_FmVspEntry->p_FmVspEntryDriverParams->p_BackupBmPools,
-+ p_FmVspEntry->p_FmVspEntryDriverParams->p_BufPoolDepletion)) != E_OK)
-+
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+
-+ if (p_FmVspEntry->p_FmVspEntryDriverParams->liodnOffset & ~FM_LIODN_OFFSET_MASK)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("liodnOffset is larger than %d", FM_LIODN_OFFSET_MASK+1));
-+
-+ err = FmVSPCheckRelativeProfile(p_FmVspEntry->h_Fm,
-+ p_FmVspEntry->portType,
-+ p_FmVspEntry->portId,
-+ p_FmVspEntry->relativeProfileId);
-+
-+ return err;
-+}
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+
-+/*****************************************************************************/
-+/* Inter-module API routines */
-+/*****************************************************************************/
-+void FmSpSetBufPoolsInAscOrderOfBufSizes(t_FmExtPools *p_FmExtPools,
-+ uint8_t *orderedArray,
-+ uint16_t *sizesArray)
-+{
-+ uint16_t bufSize = 0;
-+ int i=0, j=0, k=0;
-+
-+ /* First we copy the external buffers pools information to an ordered local array */
-+ for (i=0;i<p_FmExtPools->numOfPoolsUsed;i++)
-+ {
-+ /* get pool size */
-+ bufSize = p_FmExtPools->extBufPool[i].size;
-+
-+ /* keep sizes in an array according to poolId for direct access */
-+ sizesArray[p_FmExtPools->extBufPool[i].id] = bufSize;
-+
-+ /* save poolId in an ordered array according to size */
-+ for (j=0;j<=i;j++)
-+ {
-+ /* this is the next free place in the array */
-+ if (j==i)
-+ orderedArray[i] = p_FmExtPools->extBufPool[i].id;
-+ else
-+ {
-+ /* find the right place for this poolId */
-+ if (bufSize < sizesArray[orderedArray[j]])
-+ {
-+ /* move the poolIds one place ahead to make room for this poolId */
-+ for (k=i;k>j;k--)
-+ orderedArray[k] = orderedArray[k-1];
-+
-+ /* now k==j, this is the place for the new size */
-+ orderedArray[k] = p_FmExtPools->extBufPool[i].id;
-+ break;
-+ }
-+ }
-+ }
-+ }
-+}
-+
-+t_Error FmSpCheckBufPoolsParams(t_FmExtPools *p_FmExtPools,
-+ t_FmBackupBmPools *p_FmBackupBmPools,
-+ t_FmBufPoolDepletion *p_FmBufPoolDepletion)
-+{
-+
-+ int i = 0, j = 0;
-+ bool found;
-+ uint8_t count = 0;
-+
-+ if (p_FmExtPools)
-+ {
-+ if (p_FmExtPools->numOfPoolsUsed > FM_PORT_MAX_NUM_OF_EXT_POOLS)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("numOfPoolsUsed can't be larger than %d", FM_PORT_MAX_NUM_OF_EXT_POOLS));
-+
-+ for (i=0;i<p_FmExtPools->numOfPoolsUsed;i++)
-+ {
-+ if (p_FmExtPools->extBufPool[i].id >= BM_MAX_NUM_OF_POOLS)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("extBufPools.extBufPool[%d].id can't be larger than %d", i, BM_MAX_NUM_OF_POOLS));
-+ if (!p_FmExtPools->extBufPool[i].size)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("extBufPools.extBufPool[%d].size is 0", i));
-+ }
-+ }
-+ if (!p_FmExtPools && (p_FmBackupBmPools || p_FmBufPoolDepletion))
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("backupBmPools ot bufPoolDepletion can not be defined without external pools"));
-+
-+ /* backup BM pools indication is valid only for some chip derivatives
-+ (limited by the config routine) */
-+ if (p_FmBackupBmPools)
-+ {
-+ if (p_FmBackupBmPools->numOfBackupPools >= p_FmExtPools->numOfPoolsUsed)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("p_BackupBmPools must be smaller than extBufPools.numOfPoolsUsed"));
-+ found = FALSE;
-+ for (i = 0;i<p_FmBackupBmPools->numOfBackupPools;i++)
-+ {
-+
-+ for (j=0;j<p_FmExtPools->numOfPoolsUsed;j++)
-+ {
-+ if (p_FmBackupBmPools->poolIds[i] == p_FmExtPools->extBufPool[j].id)
-+ {
-+ found = TRUE;
-+ break;
-+ }
-+ }
-+ if (!found)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("All p_BackupBmPools.poolIds must be included in extBufPools.extBufPool[n].id"));
-+ else
-+ found = FALSE;
-+ }
-+ }
-+
-+ /* up to extBufPools.numOfPoolsUsed pools may be defined */
-+ if (p_FmBufPoolDepletion && p_FmBufPoolDepletion->poolsGrpModeEnable)
-+ {
-+ if ((p_FmBufPoolDepletion->numOfPools > p_FmExtPools->numOfPoolsUsed))
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("bufPoolDepletion.numOfPools can't be larger than %d and can't be larger than numOfPoolsUsed", FM_PORT_MAX_NUM_OF_EXT_POOLS));
-+
-+ if (!p_FmBufPoolDepletion->numOfPools)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("bufPoolDepletion.numOfPoolsToConsider can not be 0 when poolsGrpModeEnable=TRUE"));
-+
-+ found = FALSE;
-+ count = 0;
-+ /* for each pool that is in poolsToConsider, check if it is defined
-+ in extBufPool */
-+ for (i=0;i<BM_MAX_NUM_OF_POOLS;i++)
-+ {
-+ if (p_FmBufPoolDepletion->poolsToConsider[i])
-+ {
-+ for (j=0;j<p_FmExtPools->numOfPoolsUsed;j++)
-+ {
-+ if (i == p_FmExtPools->extBufPool[j].id)
-+ {
-+ found = TRUE;
-+ count++;
-+ break;
-+ }
-+ }
-+ if (!found)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("Pools selected for depletion are not used."));
-+ else
-+ found = FALSE;
-+ }
-+ }
-+ /* check that the number of pools that we have checked is equal to the number announced by the user */
-+ if (count != p_FmBufPoolDepletion->numOfPools)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("bufPoolDepletion.numOfPools is larger than the number of pools defined."));
-+ }
-+
-+ if (p_FmBufPoolDepletion && p_FmBufPoolDepletion->singlePoolModeEnable)
-+ {
-+ /* calculate vector for number of pools depletion */
-+ found = FALSE;
-+ count = 0;
-+ for (i=0;i<BM_MAX_NUM_OF_POOLS;i++)
-+ {
-+ if (p_FmBufPoolDepletion->poolsToConsiderForSingleMode[i])
-+ {
-+ for (j=0;j<p_FmExtPools->numOfPoolsUsed;j++)
-+ {
-+ if (i == p_FmExtPools->extBufPool[j].id)
-+ {
-+ found = TRUE;
-+ count++;
-+ break;
-+ }
-+ }
-+ if (!found)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("Pools selected for depletion are not used."));
-+ else
-+ found = FALSE;
-+ }
-+ }
-+ if (!count)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("No pools defined for single buffer mode pool depletion."));
-+ }
-+
-+ return E_OK;
-+}
-+
-+t_Error FmSpCheckIntContextParams(t_FmSpIntContextDataCopy *p_FmSpIntContextDataCopy)
-+{
-+ /* Check that divisible by 16 and not larger than 240 */
-+ if (p_FmSpIntContextDataCopy->intContextOffset >MAX_INT_OFFSET)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("intContext.intContextOffset can't be larger than %d", MAX_INT_OFFSET));
-+ if (p_FmSpIntContextDataCopy->intContextOffset % OFFSET_UNITS)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("intContext.intContextOffset has to be divisible by %d", OFFSET_UNITS));
-+
-+ /* check that ic size+ic internal offset, does not exceed ic block size */
-+ if (p_FmSpIntContextDataCopy->size + p_FmSpIntContextDataCopy->intContextOffset > MAX_IC_SIZE)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("intContext.size + intContext.intContextOffset has to be smaller than %d", MAX_IC_SIZE));
-+ /* Check that divisible by 16 and not larger than 256 */
-+ if (p_FmSpIntContextDataCopy->size % OFFSET_UNITS)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("intContext.size has to be divisible by %d", OFFSET_UNITS));
-+
-+ /* Check that divisible by 16 and not larger than 4K */
-+ if (p_FmSpIntContextDataCopy->extBufOffset > MAX_EXT_OFFSET)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("intContext.extBufOffset can't be larger than %d", MAX_EXT_OFFSET));
-+ if (p_FmSpIntContextDataCopy->extBufOffset % OFFSET_UNITS)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("intContext.extBufOffset has to be divisible by %d", OFFSET_UNITS));
-+
-+ return E_OK;
-+}
-+
-+t_Error FmSpCheckBufMargins(t_FmSpBufMargins *p_FmSpBufMargins)
-+{
-+ /* Check the margin definition */
-+ if (p_FmSpBufMargins->startMargins > MAX_EXT_BUFFER_OFFSET)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("bufMargins.startMargins can't be larger than %d", MAX_EXT_BUFFER_OFFSET));
-+ if (p_FmSpBufMargins->endMargins > MAX_EXT_BUFFER_OFFSET)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("bufMargins.endMargins can't be larger than %d", MAX_EXT_BUFFER_OFFSET));
-+
-+ return E_OK;
-+}
-+
-+t_Error FmSpBuildBufferStructure(t_FmSpIntContextDataCopy *p_FmSpIntContextDataCopy,
-+ t_FmBufferPrefixContent *p_BufferPrefixContent,
-+ t_FmSpBufMargins *p_FmSpBufMargins,
-+ t_FmSpBufferOffsets *p_FmSpBufferOffsets,
-+ uint8_t *internalBufferOffset)
-+{
-+ uint32_t tmp;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmSpIntContextDataCopy, E_INVALID_VALUE);
-+ ASSERT_COND(p_FmSpIntContextDataCopy);
-+ ASSERT_COND(p_BufferPrefixContent);
-+ ASSERT_COND(p_FmSpBufMargins);
-+ ASSERT_COND(p_FmSpBufferOffsets);
-+
-+ /* Align start of internal context data to 16 byte */
-+ p_FmSpIntContextDataCopy->extBufOffset =
-+ (uint16_t)((p_BufferPrefixContent->privDataSize & (OFFSET_UNITS-1)) ?
-+ ((p_BufferPrefixContent->privDataSize + OFFSET_UNITS) & ~(uint16_t)(OFFSET_UNITS-1)) :
-+ p_BufferPrefixContent->privDataSize);
-+
-+ /* Translate margin and intContext params to FM parameters */
-+ /* Initialize with illegal value. Later we'll set legal values. */
-+ p_FmSpBufferOffsets->prsResultOffset = (uint32_t)ILLEGAL_BASE;
-+ p_FmSpBufferOffsets->timeStampOffset = (uint32_t)ILLEGAL_BASE;
-+ p_FmSpBufferOffsets->hashResultOffset= (uint32_t)ILLEGAL_BASE;
-+ p_FmSpBufferOffsets->pcdInfoOffset = (uint32_t)ILLEGAL_BASE;
-+
-+ /* Internally the driver supports 4 options
-+ 1. prsResult/timestamp/hashResult selection (in fact 8 options, but for simplicity we'll
-+ relate to it as 1).
-+ 2. All IC context (from AD) not including debug.*/
-+
-+ /* This 'if' covers option 2. We copy from beginning of context. */
-+ if (p_BufferPrefixContent->passAllOtherPCDInfo)
-+ {
-+ p_FmSpIntContextDataCopy->size = 128; /* must be aligned to 16 */
-+ /* Start copying data after 16 bytes (FD) from the beginning of the internal context */
-+ p_FmSpIntContextDataCopy->intContextOffset = 16;
-+
-+ if (p_BufferPrefixContent->passAllOtherPCDInfo)
-+ p_FmSpBufferOffsets->pcdInfoOffset = p_FmSpIntContextDataCopy->extBufOffset;
-+ if (p_BufferPrefixContent->passPrsResult)
-+ p_FmSpBufferOffsets->prsResultOffset =
-+ (uint32_t)(p_FmSpIntContextDataCopy->extBufOffset + 16);
-+ if (p_BufferPrefixContent->passTimeStamp)
-+ p_FmSpBufferOffsets->timeStampOffset =
-+ (uint32_t)(p_FmSpIntContextDataCopy->extBufOffset + 48);
-+ if (p_BufferPrefixContent->passHashResult)
-+ p_FmSpBufferOffsets->hashResultOffset =
-+ (uint32_t)(p_FmSpIntContextDataCopy->extBufOffset + 56);
-+ }
-+ else
-+ {
-+ /* This case covers the options under 1 */
-+ /* Copy size must be in 16-byte granularity. */
-+ p_FmSpIntContextDataCopy->size =
-+ (uint16_t)((p_BufferPrefixContent->passPrsResult ? 32 : 0) +
-+ ((p_BufferPrefixContent->passTimeStamp ||
-+ p_BufferPrefixContent->passHashResult) ? 16 : 0));
-+
-+ /* Align start of internal context data to 16 byte */
-+ p_FmSpIntContextDataCopy->intContextOffset =
-+ (uint8_t)(p_BufferPrefixContent->passPrsResult ? 32 :
-+ ((p_BufferPrefixContent->passTimeStamp ||
-+ p_BufferPrefixContent->passHashResult) ? 64 : 0));
-+
-+ if (p_BufferPrefixContent->passPrsResult)
-+ p_FmSpBufferOffsets->prsResultOffset = p_FmSpIntContextDataCopy->extBufOffset;
-+ if (p_BufferPrefixContent->passTimeStamp)
-+ p_FmSpBufferOffsets->timeStampOffset = p_BufferPrefixContent->passPrsResult ?
-+ (p_FmSpIntContextDataCopy->extBufOffset + sizeof(t_FmPrsResult)) :
-+ p_FmSpIntContextDataCopy->extBufOffset;
-+ if (p_BufferPrefixContent->passHashResult)
-+ /* If PR is not requested, whether TS is requested or not, IC will be copied from TS */
-+ p_FmSpBufferOffsets->hashResultOffset = p_BufferPrefixContent->passPrsResult ?
-+ (p_FmSpIntContextDataCopy->extBufOffset + sizeof(t_FmPrsResult) + 8) :
-+ p_FmSpIntContextDataCopy->extBufOffset + 8;
-+ }
-+
-+ if (p_FmSpIntContextDataCopy->size)
-+ p_FmSpBufMargins->startMargins =
-+ (uint16_t)(p_FmSpIntContextDataCopy->extBufOffset +
-+ p_FmSpIntContextDataCopy->size);
-+ else
-+ /* No Internal Context passing, STartMargin is immediately after privateInfo */
-+ p_FmSpBufMargins->startMargins = p_BufferPrefixContent->privDataSize;
-+
-+ /* save extra space for manip in both external and internal buffers */
-+ if (p_BufferPrefixContent->manipExtraSpace)
-+ {
-+ uint8_t extraSpace;
-+#ifdef FM_CAPWAP_SUPPORT
-+ if ((p_BufferPrefixContent->manipExtraSpace + CAPWAP_FRAG_EXTRA_SPACE) >= 256)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE,
-+ ("p_BufferPrefixContent->manipExtraSpace should be less than %d",
-+ 256-CAPWAP_FRAG_EXTRA_SPACE));
-+ extraSpace = (uint8_t)(p_BufferPrefixContent->manipExtraSpace + CAPWAP_FRAG_EXTRA_SPACE);
-+#else
-+ extraSpace = p_BufferPrefixContent->manipExtraSpace;
-+#endif /* FM_CAPWAP_SUPPORT */
-+ p_FmSpBufferOffsets->manipOffset = p_FmSpBufMargins->startMargins;
-+ p_FmSpBufMargins->startMargins += extraSpace;
-+ *internalBufferOffset = extraSpace;
-+ }
-+
-+ /* align data start */
-+ tmp = (uint32_t)(p_FmSpBufMargins->startMargins % p_BufferPrefixContent->dataAlign);
-+ if (tmp)
-+ p_FmSpBufMargins->startMargins += (p_BufferPrefixContent->dataAlign-tmp);
-+ p_FmSpBufferOffsets->dataOffset = p_FmSpBufMargins->startMargins;
-+
-+ return E_OK;
-+}
-+/*********************** End of inter-module routines ************************/
-+
-+
-+#if (DPAA_VERSION >= 11)
-+/*****************************************************************************/
-+/* API routines */
-+/*****************************************************************************/
-+t_Handle FM_VSP_Config(t_FmVspParams *p_FmVspParams)
-+{
-+ t_FmVspEntry *p_FmVspEntry = NULL;
-+ struct fm_storage_profile_params fm_vsp_params;
-+
-+ p_FmVspEntry = (t_FmVspEntry *)XX_Malloc(sizeof(t_FmVspEntry));
-+ if (!p_FmVspEntry)
-+ {
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("p_StorageProfile allocation failed"));
-+ return NULL;
-+ }
-+ memset(p_FmVspEntry, 0, sizeof(t_FmVspEntry));
-+
-+ p_FmVspEntry->p_FmVspEntryDriverParams = (t_FmVspEntryDriverParams *)XX_Malloc(sizeof(t_FmVspEntryDriverParams));
-+ if (!p_FmVspEntry->p_FmVspEntryDriverParams)
-+ {
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("p_StorageProfile allocation failed"));
-+ XX_Free(p_FmVspEntry);
-+ return NULL;
-+ }
-+ memset(p_FmVspEntry->p_FmVspEntryDriverParams, 0, sizeof(t_FmVspEntryDriverParams));
-+ fman_vsp_defconfig(&fm_vsp_params);
-+ p_FmVspEntry->p_FmVspEntryDriverParams->dmaHeaderCacheAttr = fm_vsp_params.header_cache_attr;
-+ p_FmVspEntry->p_FmVspEntryDriverParams->dmaIntContextCacheAttr = fm_vsp_params.int_context_cache_attr;
-+ p_FmVspEntry->p_FmVspEntryDriverParams->dmaScatterGatherCacheAttr = fm_vsp_params.scatter_gather_cache_attr;
-+ p_FmVspEntry->p_FmVspEntryDriverParams->dmaSwapData = fm_vsp_params.dma_swap_data;
-+ p_FmVspEntry->p_FmVspEntryDriverParams->dmaWriteOptimize = fm_vsp_params.dma_write_optimize;
-+ p_FmVspEntry->p_FmVspEntryDriverParams->noScatherGather = fm_vsp_params.no_scather_gather;
-+ p_FmVspEntry->p_FmVspEntryDriverParams->bufferPrefixContent.privDataSize = DEFAULT_FM_SP_bufferPrefixContent_privDataSize;
-+ p_FmVspEntry->p_FmVspEntryDriverParams->bufferPrefixContent.passPrsResult= DEFAULT_FM_SP_bufferPrefixContent_passPrsResult;
-+ p_FmVspEntry->p_FmVspEntryDriverParams->bufferPrefixContent.passTimeStamp= DEFAULT_FM_SP_bufferPrefixContent_passTimeStamp;
-+ p_FmVspEntry->p_FmVspEntryDriverParams->bufferPrefixContent.passAllOtherPCDInfo
-+ = DEFAULT_FM_SP_bufferPrefixContent_passTimeStamp;
-+ p_FmVspEntry->p_FmVspEntryDriverParams->bufferPrefixContent.dataAlign = DEFAULT_FM_SP_bufferPrefixContent_dataAlign;
-+ p_FmVspEntry->p_FmVspEntryDriverParams->liodnOffset = p_FmVspParams->liodnOffset;
-+
-+ memcpy(&p_FmVspEntry->p_FmVspEntryDriverParams->extBufPools, &p_FmVspParams->extBufPools, sizeof(t_FmExtPools));
-+ p_FmVspEntry->h_Fm = p_FmVspParams->h_Fm;
-+ p_FmVspEntry->portType = p_FmVspParams->portParams.portType;
-+ p_FmVspEntry->portId = p_FmVspParams->portParams.portId;
-+
-+ p_FmVspEntry->relativeProfileId = p_FmVspParams->relativeProfileId;
-+
-+ return p_FmVspEntry;
-+}
-+
-+t_Error FM_VSP_Init(t_Handle h_FmVsp)
-+{
-+
-+ t_FmVspEntry *p_FmVspEntry = (t_FmVspEntry *)h_FmVsp;
-+ struct fm_storage_profile_params fm_vsp_params;
-+ uint8_t orderedArray[FM_PORT_MAX_NUM_OF_EXT_POOLS];
-+ uint16_t sizesArray[BM_MAX_NUM_OF_POOLS];
-+ t_Error err;
-+ uint16_t absoluteProfileId = 0;
-+ int i = 0;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmVspEntry, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmVspEntry->p_FmVspEntryDriverParams,E_INVALID_HANDLE);
-+
-+ CHECK_INIT_PARAMETERS(p_FmVspEntry, CheckParams);
-+
-+ memset(&orderedArray, 0, sizeof(uint8_t) * FM_PORT_MAX_NUM_OF_EXT_POOLS);
-+ memset(&sizesArray, 0, sizeof(uint16_t) * BM_MAX_NUM_OF_POOLS);
-+
-+ err = FmSpBuildBufferStructure(&p_FmVspEntry->intContext,
-+ &p_FmVspEntry->p_FmVspEntryDriverParams->bufferPrefixContent,
-+ &p_FmVspEntry->bufMargins,
-+ &p_FmVspEntry->bufferOffsets,
-+ &p_FmVspEntry->internalBufferOffset);
-+ if (err != E_OK)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+
-+
-+ err = CheckParamsGeneratedInternally(p_FmVspEntry);
-+ if (err != E_OK)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+
-+
-+ p_FmVspEntry->p_FmSpRegsBase =
-+ (struct fm_pcd_storage_profile_regs *)FmGetVSPBaseAddr(p_FmVspEntry->h_Fm);
-+ if (!p_FmVspEntry->p_FmSpRegsBase)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("impossible to initialize SpRegsBase"));
-+
-+ /* order external buffer pools in ascending order of buffer pools sizes */
-+ FmSpSetBufPoolsInAscOrderOfBufSizes(&(p_FmVspEntry->p_FmVspEntryDriverParams)->extBufPools,
-+ orderedArray,
-+ sizesArray);
-+
-+ p_FmVspEntry->extBufPools.numOfPoolsUsed =
-+ p_FmVspEntry->p_FmVspEntryDriverParams->extBufPools.numOfPoolsUsed;
-+ for (i = 0; i < p_FmVspEntry->extBufPools.numOfPoolsUsed; i++)
-+ {
-+ p_FmVspEntry->extBufPools.extBufPool[i].id = orderedArray[i];
-+ p_FmVspEntry->extBufPools.extBufPool[i].size = sizesArray[orderedArray[i]];
-+ }
-+
-+ /* on user responsibility to fill it according requirement */
-+ memset(&fm_vsp_params, 0, sizeof(struct fm_storage_profile_params));
-+ fm_vsp_params.dma_swap_data = p_FmVspEntry->p_FmVspEntryDriverParams->dmaSwapData;
-+ fm_vsp_params.int_context_cache_attr = p_FmVspEntry->p_FmVspEntryDriverParams->dmaIntContextCacheAttr;
-+ fm_vsp_params.header_cache_attr = p_FmVspEntry->p_FmVspEntryDriverParams->dmaHeaderCacheAttr;
-+ fm_vsp_params.scatter_gather_cache_attr = p_FmVspEntry->p_FmVspEntryDriverParams->dmaScatterGatherCacheAttr;
-+ fm_vsp_params.dma_write_optimize = p_FmVspEntry->p_FmVspEntryDriverParams->dmaWriteOptimize;
-+ fm_vsp_params.liodn_offset = p_FmVspEntry->p_FmVspEntryDriverParams->liodnOffset;
-+ fm_vsp_params.no_scather_gather = p_FmVspEntry->p_FmVspEntryDriverParams->noScatherGather;
-+
-+ if (p_FmVspEntry->p_FmVspEntryDriverParams->p_BufPoolDepletion)
-+ {
-+ fm_vsp_params.buf_pool_depletion.buf_pool_depletion_enabled = TRUE;
-+ fm_vsp_params.buf_pool_depletion.pools_grp_mode_enable = p_FmVspEntry->p_FmVspEntryDriverParams->p_BufPoolDepletion->poolsGrpModeEnable;
-+ fm_vsp_params.buf_pool_depletion.num_pools = p_FmVspEntry->p_FmVspEntryDriverParams->p_BufPoolDepletion->numOfPools;
-+ fm_vsp_params.buf_pool_depletion.pools_to_consider = p_FmVspEntry->p_FmVspEntryDriverParams->p_BufPoolDepletion->poolsToConsider;
-+ fm_vsp_params.buf_pool_depletion.single_pool_mode_enable = p_FmVspEntry->p_FmVspEntryDriverParams->p_BufPoolDepletion->singlePoolModeEnable;
-+ fm_vsp_params.buf_pool_depletion.pools_to_consider_for_single_mode = p_FmVspEntry->p_FmVspEntryDriverParams->p_BufPoolDepletion->poolsToConsiderForSingleMode;
-+ fm_vsp_params.buf_pool_depletion.has_pfc_priorities = TRUE;
-+ fm_vsp_params.buf_pool_depletion.pfc_priorities_en = p_FmVspEntry->p_FmVspEntryDriverParams->p_BufPoolDepletion->pfcPrioritiesEn;
-+ }
-+ else
-+ fm_vsp_params.buf_pool_depletion.buf_pool_depletion_enabled = FALSE;
-+
-+ if (p_FmVspEntry->p_FmVspEntryDriverParams->p_BackupBmPools)
-+ {
-+ fm_vsp_params.backup_pools.num_backup_pools = p_FmVspEntry->p_FmVspEntryDriverParams->p_BackupBmPools->numOfBackupPools;
-+ fm_vsp_params.backup_pools.pool_ids = p_FmVspEntry->p_FmVspEntryDriverParams->p_BackupBmPools->poolIds;
-+ }
-+ else
-+ fm_vsp_params.backup_pools.num_backup_pools = 0;
-+
-+ fm_vsp_params.fm_ext_pools.num_pools_used = p_FmVspEntry->extBufPools.numOfPoolsUsed;
-+ fm_vsp_params.fm_ext_pools.ext_buf_pool = (struct fman_ext_pool_params*)&p_FmVspEntry->extBufPools.extBufPool;
-+ fm_vsp_params.buf_margins = (struct fman_sp_buf_margins*)&p_FmVspEntry->bufMargins;
-+ fm_vsp_params.int_context = (struct fman_sp_int_context_data_copy*)&p_FmVspEntry->intContext;
-+
-+ /* no check on err - it was checked earlier */
-+ FmVSPGetAbsoluteProfileId(p_FmVspEntry->h_Fm,
-+ p_FmVspEntry->portType,
-+ p_FmVspEntry->portId,
-+ p_FmVspEntry->relativeProfileId,
-+ &absoluteProfileId);
-+
-+ ASSERT_COND(p_FmVspEntry->p_FmSpRegsBase);
-+ ASSERT_COND(fm_vsp_params.int_context);
-+ ASSERT_COND(fm_vsp_params.buf_margins);
-+ ASSERT_COND((absoluteProfileId <= FM_VSP_MAX_NUM_OF_ENTRIES));
-+
-+ /* Set all registers related to VSP */
-+ fman_vsp_init(p_FmVspEntry->p_FmSpRegsBase, absoluteProfileId, &fm_vsp_params,FM_PORT_MAX_NUM_OF_EXT_POOLS, BM_MAX_NUM_OF_POOLS, FM_MAX_NUM_OF_PFC_PRIORITIES);
-+
-+ p_FmVspEntry->absoluteSpId = absoluteProfileId;
-+
-+ if (p_FmVspEntry->p_FmVspEntryDriverParams)
-+ XX_Free(p_FmVspEntry->p_FmVspEntryDriverParams);
-+ p_FmVspEntry->p_FmVspEntryDriverParams = NULL;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_VSP_Free(t_Handle h_FmVsp)
-+{
-+ t_FmVspEntry *p_FmVspEntry = (t_FmVspEntry *)h_FmVsp;
-+ SANITY_CHECK_RETURN_ERROR(h_FmVsp, E_INVALID_HANDLE);
-+ XX_Free(p_FmVspEntry);
-+ return E_OK;
-+}
-+
-+t_Error FM_VSP_ConfigBufferPrefixContent(t_Handle h_FmVsp, t_FmBufferPrefixContent *p_FmBufferPrefixContent)
-+{
-+ t_FmVspEntry *p_FmVspEntry = (t_FmVspEntry*)h_FmVsp;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmVspEntry, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmVspEntry->p_FmVspEntryDriverParams, E_INVALID_HANDLE);
-+
-+ memcpy(&p_FmVspEntry->p_FmVspEntryDriverParams->bufferPrefixContent, p_FmBufferPrefixContent, sizeof(t_FmBufferPrefixContent));
-+ /* if dataAlign was not initialized by user, we return to driver's default */
-+ if (!p_FmVspEntry->p_FmVspEntryDriverParams->bufferPrefixContent.dataAlign)
-+ p_FmVspEntry->p_FmVspEntryDriverParams->bufferPrefixContent.dataAlign = DEFAULT_FM_SP_bufferPrefixContent_dataAlign;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_VSP_ConfigDmaSwapData(t_Handle h_FmVsp, e_FmDmaSwapOption swapData)
-+{
-+ t_FmVspEntry *p_FmVspEntry = (t_FmVspEntry*)h_FmVsp;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmVspEntry, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmVspEntry->p_FmVspEntryDriverParams, E_INVALID_HANDLE);
-+
-+ p_FmVspEntry->p_FmVspEntryDriverParams->dmaSwapData = swapData;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_VSP_ConfigDmaIcCacheAttr(t_Handle h_FmVsp, e_FmDmaCacheOption intContextCacheAttr)
-+{
-+ t_FmVspEntry *p_FmVspEntry = (t_FmVspEntry*)h_FmVsp;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmVspEntry, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmVspEntry->p_FmVspEntryDriverParams, E_INVALID_HANDLE);
-+
-+ p_FmVspEntry->p_FmVspEntryDriverParams->dmaIntContextCacheAttr = intContextCacheAttr;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_VSP_ConfigDmaHdrAttr(t_Handle h_FmVsp, e_FmDmaCacheOption headerCacheAttr)
-+{
-+ t_FmVspEntry *p_FmVspEntry = (t_FmVspEntry*)h_FmVsp;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmVspEntry, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmVspEntry->p_FmVspEntryDriverParams, E_INVALID_HANDLE);
-+
-+ p_FmVspEntry->p_FmVspEntryDriverParams->dmaHeaderCacheAttr = headerCacheAttr;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_VSP_ConfigDmaScatterGatherAttr(t_Handle h_FmVsp, e_FmDmaCacheOption scatterGatherCacheAttr)
-+{
-+ t_FmVspEntry *p_FmVspEntry = (t_FmVspEntry*)h_FmVsp;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmVspEntry, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmVspEntry->p_FmVspEntryDriverParams, E_INVALID_HANDLE);
-+
-+ p_FmVspEntry->p_FmVspEntryDriverParams->dmaScatterGatherCacheAttr = scatterGatherCacheAttr;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_VSP_ConfigDmaWriteOptimize(t_Handle h_FmVsp, bool optimize)
-+{
-+ t_FmVspEntry *p_FmVspEntry = (t_FmVspEntry*)h_FmVsp;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmVspEntry, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmVspEntry->p_FmVspEntryDriverParams, E_INVALID_HANDLE);
-+
-+
-+ p_FmVspEntry->p_FmVspEntryDriverParams->dmaWriteOptimize = optimize;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_VSP_ConfigNoScatherGather(t_Handle h_FmVsp, bool noScatherGather)
-+{
-+ t_FmVspEntry *p_FmVspEntry = (t_FmVspEntry*)h_FmVsp;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_FmVspEntry, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmVspEntry->p_FmVspEntryDriverParams, E_INVALID_HANDLE);
-+
-+
-+ p_FmVspEntry->p_FmVspEntryDriverParams->noScatherGather = noScatherGather;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_VSP_ConfigPoolDepletion(t_Handle h_FmVsp, t_FmBufPoolDepletion *p_BufPoolDepletion)
-+{
-+ t_FmVspEntry *p_FmVspEntry = (t_FmVspEntry*)h_FmVsp;
-+
-+ SANITY_CHECK_RETURN_ERROR(h_FmVsp, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmVspEntry->p_FmVspEntryDriverParams, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_BufPoolDepletion, E_INVALID_HANDLE);
-+
-+ p_FmVspEntry->p_FmVspEntryDriverParams->p_BufPoolDepletion = (t_FmBufPoolDepletion *)XX_Malloc(sizeof(t_FmBufPoolDepletion));
-+ if (!p_FmVspEntry->p_FmVspEntryDriverParams->p_BufPoolDepletion)
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, ("p_BufPoolDepletion allocation failed"));
-+ memcpy(p_FmVspEntry->p_FmVspEntryDriverParams->p_BufPoolDepletion, p_BufPoolDepletion, sizeof(t_FmBufPoolDepletion));
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_VSP_ConfigBackupPools(t_Handle h_FmVsp, t_FmBackupBmPools *p_BackupBmPools)
-+{
-+ t_FmVspEntry *p_FmVspEntry = (t_FmVspEntry*)h_FmVsp;
-+
-+ SANITY_CHECK_RETURN_ERROR(h_FmVsp, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmVspEntry->p_FmVspEntryDriverParams, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_BackupBmPools, E_INVALID_HANDLE);
-+
-+ p_FmVspEntry->p_FmVspEntryDriverParams->p_BackupBmPools = (t_FmBackupBmPools *)XX_Malloc(sizeof(t_FmBackupBmPools));
-+ if (!p_FmVspEntry->p_FmVspEntryDriverParams->p_BackupBmPools)
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, ("p_BackupBmPools allocation failed"));
-+ memcpy(p_FmVspEntry->p_FmVspEntryDriverParams->p_BackupBmPools, p_BackupBmPools, sizeof(t_FmBackupBmPools));
-+
-+ return E_OK;
-+}
-+
-+uint32_t FM_VSP_GetBufferDataOffset(t_Handle h_FmVsp)
-+{
-+ t_FmVspEntry *p_FmVspEntry = (t_FmVspEntry*)h_FmVsp;
-+
-+ SANITY_CHECK_RETURN_VALUE(p_FmVspEntry, E_INVALID_HANDLE, 0);
-+ SANITY_CHECK_RETURN_VALUE(!p_FmVspEntry->p_FmVspEntryDriverParams, E_INVALID_STATE, 0);
-+
-+ return p_FmVspEntry->bufferOffsets.dataOffset;
-+}
-+
-+uint8_t * FM_VSP_GetBufferICInfo(t_Handle h_FmVsp, char *p_Data)
-+{
-+ t_FmVspEntry *p_FmVspEntry = (t_FmVspEntry*)h_FmVsp;
-+
-+ SANITY_CHECK_RETURN_VALUE(p_FmVspEntry, E_INVALID_HANDLE, NULL);
-+ SANITY_CHECK_RETURN_VALUE(!p_FmVspEntry->p_FmVspEntryDriverParams, E_INVALID_STATE, NULL);
-+
-+ if (p_FmVspEntry->bufferOffsets.pcdInfoOffset == ILLEGAL_BASE)
-+ return NULL;
-+
-+ return (uint8_t *)PTR_MOVE(p_Data, p_FmVspEntry->bufferOffsets.pcdInfoOffset);
-+}
-+
-+t_FmPrsResult * FM_VSP_GetBufferPrsResult(t_Handle h_FmVsp, char *p_Data)
-+{
-+ t_FmVspEntry *p_FmVspEntry = (t_FmVspEntry*)h_FmVsp;
-+
-+ SANITY_CHECK_RETURN_VALUE(p_FmVspEntry, E_INVALID_HANDLE, NULL);
-+ SANITY_CHECK_RETURN_VALUE(!p_FmVspEntry->p_FmVspEntryDriverParams, E_INVALID_STATE, NULL);
-+
-+ if (p_FmVspEntry->bufferOffsets.prsResultOffset == ILLEGAL_BASE)
-+ return NULL;
-+
-+ return (t_FmPrsResult *)PTR_MOVE(p_Data, p_FmVspEntry->bufferOffsets.prsResultOffset);
-+}
-+
-+uint64_t * FM_VSP_GetBufferTimeStamp(t_Handle h_FmVsp, char *p_Data)
-+{
-+ t_FmVspEntry *p_FmVspEntry = (t_FmVspEntry*)h_FmVsp;
-+
-+ SANITY_CHECK_RETURN_VALUE(p_FmVspEntry, E_INVALID_HANDLE, NULL);
-+ SANITY_CHECK_RETURN_VALUE(!p_FmVspEntry->p_FmVspEntryDriverParams, E_INVALID_STATE, NULL);
-+
-+ if (p_FmVspEntry->bufferOffsets.timeStampOffset == ILLEGAL_BASE)
-+ return NULL;
-+
-+ return (uint64_t *)PTR_MOVE(p_Data, p_FmVspEntry->bufferOffsets.timeStampOffset);
-+}
-+
-+uint8_t * FM_VSP_GetBufferHashResult(t_Handle h_FmVsp, char *p_Data)
-+{
-+ t_FmVspEntry *p_FmVspEntry = (t_FmVspEntry*)h_FmVsp;
-+
-+ SANITY_CHECK_RETURN_VALUE(p_FmVspEntry, E_INVALID_HANDLE, NULL);
-+ SANITY_CHECK_RETURN_VALUE(!p_FmVspEntry->p_FmVspEntryDriverParams, E_INVALID_STATE, NULL);
-+
-+ if (p_FmVspEntry->bufferOffsets.hashResultOffset == ILLEGAL_BASE)
-+ return NULL;
-+
-+ return (uint8_t *)PTR_MOVE(p_Data, p_FmVspEntry->bufferOffsets.hashResultOffset);
-+}
-+
-+#endif /* (DPAA_VERSION >= 11) */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/SP/fm_sp.h
-@@ -0,0 +1,85 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/******************************************************************************
-+ @File fm_sp.h
-+
-+ @Description FM SP ...
-+*//***************************************************************************/
-+#ifndef __FM_SP_H
-+#define __FM_SP_H
-+
-+#include "std_ext.h"
-+#include "error_ext.h"
-+#include "list_ext.h"
-+
-+#include "fm_sp_common.h"
-+#include "fm_common.h"
-+
-+
-+#define __ERR_MODULE__ MODULE_FM_SP
-+
-+typedef struct {
-+ t_FmBufferPrefixContent bufferPrefixContent;
-+ e_FmDmaSwapOption dmaSwapData;
-+ e_FmDmaCacheOption dmaIntContextCacheAttr;
-+ e_FmDmaCacheOption dmaHeaderCacheAttr;
-+ e_FmDmaCacheOption dmaScatterGatherCacheAttr;
-+ bool dmaWriteOptimize;
-+ uint16_t liodnOffset;
-+ bool noScatherGather;
-+ t_FmBufPoolDepletion *p_BufPoolDepletion;
-+ t_FmBackupBmPools *p_BackupBmPools;
-+ t_FmExtPools extBufPools;
-+} t_FmVspEntryDriverParams;
-+
-+typedef struct {
-+ bool valid;
-+ volatile bool lock;
-+ uint8_t pointedOwners;
-+ uint16_t absoluteSpId;
-+ uint8_t internalBufferOffset;
-+ t_FmSpBufMargins bufMargins;
-+ t_FmSpIntContextDataCopy intContext;
-+ t_FmSpBufferOffsets bufferOffsets;
-+ t_Handle h_Fm;
-+ e_FmPortType portType; /**< Port type */
-+ uint8_t portId; /**< Port Id - relative to type */
-+ uint8_t relativeProfileId;
-+ struct fm_pcd_storage_profile_regs *p_FmSpRegsBase;
-+ t_FmExtPools extBufPools;
-+ t_FmVspEntryDriverParams *p_FmVspEntryDriverParams;
-+} t_FmVspEntry;
-+
-+
-+#endif /* __FM_SP_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/SP/fman_sp.c
-@@ -0,0 +1,197 @@
-+/*
-+ * Copyright 2013 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#include "fsl_fman_sp.h"
-+
-+
-+uint32_t fman_vsp_get_statistics(struct fm_pcd_storage_profile_regs *regs,
-+ uint16_t index)
-+{
-+ struct fm_pcd_storage_profile_regs *sp_regs;
-+ sp_regs = &regs[index];
-+ return ioread32be(&sp_regs->fm_sp_acnt);
-+}
-+
-+void fman_vsp_set_statistics(struct fm_pcd_storage_profile_regs *regs,
-+ uint16_t index, uint32_t value)
-+{
-+ struct fm_pcd_storage_profile_regs *sp_regs;
-+ sp_regs = &regs[index];
-+ iowrite32be(value, &sp_regs->fm_sp_acnt);
-+}
-+
-+void fman_vsp_defconfig(struct fm_storage_profile_params *cfg)
-+{
-+ cfg->dma_swap_data =
-+ DEFAULT_FMAN_SP_DMA_SWAP_DATA;
-+ cfg->int_context_cache_attr =
-+ DEFAULT_FMAN_SP_DMA_INT_CONTEXT_CACHE_ATTR;
-+ cfg->header_cache_attr =
-+ DEFAULT_FMAN_SP_DMA_HEADER_CACHE_ATTR;
-+ cfg->scatter_gather_cache_attr =
-+ DEFAULT_FMAN_SP_DMA_SCATTER_GATHER_CACHE_ATTR;
-+ cfg->dma_write_optimize =
-+ DEFAULT_FMAN_SP_DMA_WRITE_OPTIMIZE;
-+ cfg->no_scather_gather =
-+ DEFAULT_FMAN_SP_NO_SCATTER_GATHER;
-+}
-+
-+static inline uint32_t calc_vec_dep(int max_pools, bool *pools,
-+ struct fman_ext_pools *ext_buf_pools, uint32_t mask)
-+{
-+ int i, j;
-+ uint32_t vector = 0;
-+ for (i = 0; i < max_pools; i++)
-+ if (pools[i])
-+ for (j = 0; j < ext_buf_pools->num_pools_used; j++)
-+ if (i == ext_buf_pools->ext_buf_pool[j].id) {
-+ vector |= mask >> j;
-+ break;
-+ }
-+ return vector;
-+}
-+
-+void fman_vsp_init(struct fm_pcd_storage_profile_regs *regs,
-+ uint16_t index, struct fm_storage_profile_params *fm_vsp_params,
-+ int port_max_num_of_ext_pools, int bm_max_num_of_pools,
-+ int max_num_of_pfc_priorities)
-+{
-+ int i = 0, j = 0;
-+ struct fm_pcd_storage_profile_regs *sp_regs;
-+ uint32_t tmp_reg, vector;
-+ struct fman_ext_pools *ext_buf_pools = &fm_vsp_params->fm_ext_pools;
-+ struct fman_buf_pool_depletion *buf_pool_depletion =
-+ &fm_vsp_params->buf_pool_depletion;
-+ struct fman_backup_bm_pools *backup_pools =
-+ &fm_vsp_params->backup_pools;
-+ struct fman_sp_int_context_data_copy *int_context_data_copy =
-+ fm_vsp_params->int_context;
-+ struct fman_sp_buf_margins *external_buffer_margins =
-+ fm_vsp_params->buf_margins;
-+ bool no_scather_gather = fm_vsp_params->no_scather_gather;
-+ uint16_t liodn_offset = fm_vsp_params->liodn_offset;
-+
-+ sp_regs = &regs[index];
-+
-+ /* fill external buffers manager pool information register*/
-+ for (i = 0; i < ext_buf_pools->num_pools_used; i++) {
-+ tmp_reg = FMAN_SP_EXT_BUF_POOL_VALID |
-+ FMAN_SP_EXT_BUF_POOL_EN_COUNTER;
-+ tmp_reg |= ((uint32_t)ext_buf_pools->ext_buf_pool[i].id <<
-+ FMAN_SP_EXT_BUF_POOL_ID_SHIFT);
-+ tmp_reg |= ext_buf_pools->ext_buf_pool[i].size;
-+ /* functionality available only for some deriviatives
-+ (limited by config) */
-+ for (j = 0; j < backup_pools->num_backup_pools; j++)
-+ if (ext_buf_pools->ext_buf_pool[i].id ==
-+ backup_pools->pool_ids[j]) {
-+ tmp_reg |= FMAN_SP_EXT_BUF_POOL_BACKUP;
-+ break;
-+ }
-+ iowrite32be(tmp_reg, &sp_regs->fm_sp_ebmpi[i]);
-+ }
-+
-+ /* clear unused pools */
-+ for (i = ext_buf_pools->num_pools_used;
-+ i < port_max_num_of_ext_pools; i++)
-+ iowrite32be(0, &sp_regs->fm_sp_ebmpi[i]);
-+
-+ /* fill pool depletion register*/
-+ tmp_reg = 0;
-+ if (buf_pool_depletion->buf_pool_depletion_enabled && buf_pool_depletion->pools_grp_mode_enable) {
-+ /* calculate vector for number of pools depletion */
-+ vector = calc_vec_dep(bm_max_num_of_pools, buf_pool_depletion->
-+ pools_to_consider, ext_buf_pools, 0x80000000);
-+
-+ /* configure num of pools and vector for number of pools mode */
-+ tmp_reg |= (((uint32_t)buf_pool_depletion->num_pools - 1) <<
-+ FMAN_SP_POOL_DEP_NUM_OF_POOLS_SHIFT);
-+ tmp_reg |= vector;
-+ }
-+
-+ if (buf_pool_depletion->buf_pool_depletion_enabled && buf_pool_depletion->single_pool_mode_enable) {
-+ /* calculate vector for number of pools depletion */
-+ vector = calc_vec_dep(bm_max_num_of_pools, buf_pool_depletion->
-+ pools_to_consider_for_single_mode,
-+ ext_buf_pools, 0x00000080);
-+
-+ /* configure num of pools and vector for number of pools mode */
-+ tmp_reg |= vector;
-+ }
-+
-+ /* fill QbbPEV */
-+ if (buf_pool_depletion->buf_pool_depletion_enabled) {
-+ vector = 0;
-+ for (i = 0; i < max_num_of_pfc_priorities; i++)
-+ if (buf_pool_depletion->pfc_priorities_en[i] == TRUE)
-+ vector |= 0x00000100 << i;
-+ tmp_reg |= vector;
-+ }
-+ iowrite32be(tmp_reg, &sp_regs->fm_sp_mpd);
-+
-+ /* fill dma attributes register */
-+ tmp_reg = 0;
-+ tmp_reg |= (uint32_t)fm_vsp_params->dma_swap_data <<
-+ FMAN_SP_DMA_ATTR_SWP_SHIFT;
-+ tmp_reg |= (uint32_t)fm_vsp_params->int_context_cache_attr <<
-+ FMAN_SP_DMA_ATTR_IC_CACHE_SHIFT;
-+ tmp_reg |= (uint32_t)fm_vsp_params->header_cache_attr <<
-+ FMAN_SP_DMA_ATTR_HDR_CACHE_SHIFT;
-+ tmp_reg |= (uint32_t)fm_vsp_params->scatter_gather_cache_attr <<
-+ FMAN_SP_DMA_ATTR_SG_CACHE_SHIFT;
-+ if (fm_vsp_params->dma_write_optimize)
-+ tmp_reg |= FMAN_SP_DMA_ATTR_WRITE_OPTIMIZE;
-+ iowrite32be(tmp_reg, &sp_regs->fm_sp_da);
-+
-+ /* IC parameters - fill internal context parameters register */
-+ tmp_reg = 0;
-+ tmp_reg |= (((uint32_t)int_context_data_copy->ext_buf_offset/
-+ OFFSET_UNITS) << FMAN_SP_IC_TO_EXT_SHIFT);
-+ tmp_reg |= (((uint32_t)int_context_data_copy->int_context_offset/
-+ OFFSET_UNITS) << FMAN_SP_IC_FROM_INT_SHIFT);
-+ tmp_reg |= (((uint32_t)int_context_data_copy->size/OFFSET_UNITS) <<
-+ FMAN_SP_IC_SIZE_SHIFT);
-+ iowrite32be(tmp_reg, &sp_regs->fm_sp_icp);
-+
-+ /* buffer margins - fill external buffer margins register */
-+ tmp_reg = 0;
-+ tmp_reg |= (((uint32_t)external_buffer_margins->start_margins) <<
-+ FMAN_SP_EXT_BUF_MARG_START_SHIFT);
-+ tmp_reg |= (((uint32_t)external_buffer_margins->end_margins) <<
-+ FMAN_SP_EXT_BUF_MARG_END_SHIFT);
-+ if (no_scather_gather)
-+ tmp_reg |= FMAN_SP_SG_DISABLE;
-+ iowrite32be(tmp_reg, &sp_regs->fm_sp_ebm);
-+
-+ /* buffer margins - fill spliodn register */
-+ iowrite32be(liodn_offset, &sp_regs->fm_sp_spliodn);
-+}
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/fm.c
-@@ -0,0 +1,5216 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/******************************************************************************
-+ @File fm.c
-+
-+ @Description FM driver routines implementation.
-+*//***************************************************************************/
-+#include "std_ext.h"
-+#include "error_ext.h"
-+#include "xx_ext.h"
-+#include "string_ext.h"
-+#include "sprint_ext.h"
-+#include "debug_ext.h"
-+#include "fm_muram_ext.h"
-+#include <linux/math64.h>
-+
-+#include "fm_common.h"
-+#include "fm_ipc.h"
-+#include "fm.h"
-+#ifndef CONFIG_FMAN_ARM
-+#include <linux/fsl/svr.h>
-+#endif
-+#include "fsl_fman.h"
-+
-+
-+/****************************************/
-+/* static functions */
-+/****************************************/
-+
-+static volatile bool blockingFlag = FALSE;
-+static void IpcMsgCompletionCB(t_Handle h_Fm,
-+ uint8_t *p_Msg,
-+ uint8_t *p_Reply,
-+ uint32_t replyLength,
-+ t_Error status)
-+{
-+ UNUSED(h_Fm);UNUSED(p_Msg);UNUSED(p_Reply);UNUSED(replyLength);UNUSED(status);
-+ blockingFlag = FALSE;
-+}
-+
-+static void FreeInitResources(t_Fm *p_Fm)
-+{
-+ if (p_Fm->camBaseAddr)
-+ FM_MURAM_FreeMem(p_Fm->h_FmMuram, UINT_TO_PTR(p_Fm->camBaseAddr));
-+ if (p_Fm->fifoBaseAddr)
-+ FM_MURAM_FreeMem(p_Fm->h_FmMuram, UINT_TO_PTR(p_Fm->fifoBaseAddr));
-+ if (p_Fm->resAddr)
-+ FM_MURAM_FreeMem(p_Fm->h_FmMuram, UINT_TO_PTR(p_Fm->resAddr));
-+}
-+
-+static bool IsFmanCtrlCodeLoaded(t_Fm *p_Fm)
-+{
-+ t_FMIramRegs *p_Iram;
-+
-+ ASSERT_COND(p_Fm);
-+ p_Iram = (t_FMIramRegs *)UINT_TO_PTR(p_Fm->baseAddr + FM_MM_IMEM);
-+
-+ return (bool)!!(GET_UINT32(p_Iram->iready) & IRAM_READY);
-+}
-+
-+static t_Error CheckFmParameters(t_Fm *p_Fm)
-+{
-+ if (IsFmanCtrlCodeLoaded(p_Fm) && !p_Fm->resetOnInit)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("Old FMan CTRL code is loaded; FM must be reset!"));
-+#if (DPAA_VERSION < 11)
-+ if (!p_Fm->p_FmDriverParam->dma_axi_dbg_num_of_beats ||
-+ (p_Fm->p_FmDriverParam->dma_axi_dbg_num_of_beats > DMA_MODE_MAX_AXI_DBG_NUM_OF_BEATS))
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE,
-+ ("axiDbgNumOfBeats has to be in the range 1 - %d", DMA_MODE_MAX_AXI_DBG_NUM_OF_BEATS));
-+#endif /* (DPAA_VERSION < 11) */
-+ if (p_Fm->p_FmDriverParam->dma_cam_num_of_entries % DMA_CAM_UNITS)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("dma_cam_num_of_entries has to be divisble by %d", DMA_CAM_UNITS));
-+// if (!p_Fm->p_FmDriverParam->dma_cam_num_of_entries || (p_Fm->p_FmDriverParam->dma_cam_num_of_entries > DMA_MODE_MAX_CAM_NUM_OF_ENTRIES))
-+// RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("dma_cam_num_of_entries has to be in the range 1 - %d", DMA_MODE_MAX_CAM_NUM_OF_ENTRIES));
-+ if (p_Fm->p_FmDriverParam->dma_comm_qtsh_asrt_emer > DMA_THRESH_MAX_COMMQ)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("dma_comm_qtsh_asrt_emer can not be larger than %d", DMA_THRESH_MAX_COMMQ));
-+ if (p_Fm->p_FmDriverParam->dma_comm_qtsh_clr_emer > DMA_THRESH_MAX_COMMQ)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("dma_comm_qtsh_clr_emer can not be larger than %d", DMA_THRESH_MAX_COMMQ));
-+ if (p_Fm->p_FmDriverParam->dma_comm_qtsh_clr_emer >= p_Fm->p_FmDriverParam->dma_comm_qtsh_asrt_emer)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("dma_comm_qtsh_clr_emer must be smaller than dma_comm_qtsh_asrt_emer"));
-+#if (DPAA_VERSION < 11)
-+ if (p_Fm->p_FmDriverParam->dma_read_buf_tsh_asrt_emer > DMA_THRESH_MAX_BUF)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("dma_read_buf_tsh_asrt_emer can not be larger than %d", DMA_THRESH_MAX_BUF));
-+ if (p_Fm->p_FmDriverParam->dma_read_buf_tsh_clr_emer > DMA_THRESH_MAX_BUF)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("dma_read_buf_tsh_clr_emer can not be larger than %d", DMA_THRESH_MAX_BUF));
-+ if (p_Fm->p_FmDriverParam->dma_read_buf_tsh_clr_emer >= p_Fm->p_FmDriverParam->dma_read_buf_tsh_asrt_emer)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("dma_read_buf_tsh_clr_emer must be smaller than dma_read_buf_tsh_asrt_emer"));
-+ if (p_Fm->p_FmDriverParam->dma_write_buf_tsh_asrt_emer > DMA_THRESH_MAX_BUF)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("dma_write_buf_tsh_asrt_emer can not be larger than %d", DMA_THRESH_MAX_BUF));
-+ if (p_Fm->p_FmDriverParam->dma_write_buf_tsh_clr_emer > DMA_THRESH_MAX_BUF)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("dma_write_buf_tsh_clr_emer can not be larger than %d", DMA_THRESH_MAX_BUF));
-+ if (p_Fm->p_FmDriverParam->dma_write_buf_tsh_clr_emer >= p_Fm->p_FmDriverParam->dma_write_buf_tsh_asrt_emer)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("dma_write_buf_tsh_clr_emer must be smaller than dma_write_buf_tsh_asrt_emer"));
-+#else /* (DPAA_VERSION >= 11) */
-+ if ((p_Fm->p_FmDriverParam->dma_dbg_cnt_mode == E_FMAN_DMA_DBG_CNT_INT_READ_EM)||
-+ (p_Fm->p_FmDriverParam->dma_dbg_cnt_mode == E_FMAN_DMA_DBG_CNT_INT_WRITE_EM) ||
-+ (p_Fm->p_FmDriverParam->dma_dbg_cnt_mode == E_FMAN_DMA_DBG_CNT_RAW_WAR_PROT))
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("dma_dbg_cnt_mode value not supported by this integration."));
-+ if ((p_Fm->p_FmDriverParam->dma_emergency_bus_select == FM_DMA_MURAM_READ_EMERGENCY)||
-+ (p_Fm->p_FmDriverParam->dma_emergency_bus_select == FM_DMA_MURAM_WRITE_EMERGENCY))
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("emergencyBusSelect value not supported by this integration."));
-+ if (p_Fm->p_FmDriverParam->dma_stop_on_bus_error)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("dma_stop_on_bus_error not supported by this integration."));
-+#ifdef FM_AID_MODE_NO_TNUM_SW005
-+ if (p_Fm->p_FmDriverParam->dma_aid_mode != E_FMAN_DMA_AID_OUT_PORT_ID)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("dma_aid_mode not supported by this integration."));
-+#endif /* FM_AID_MODE_NO_TNUM_SW005 */
-+ if (p_Fm->p_FmDriverParam->dma_axi_dbg_num_of_beats)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("dma_axi_dbg_num_of_beats not supported by this integration."));
-+#endif /* (DPAA_VERSION < 11) */
-+
-+ if (!p_Fm->p_FmStateStruct->fmClkFreq)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("fmClkFreq must be set."));
-+ if (USEC_TO_CLK(p_Fm->p_FmDriverParam->dma_watchdog, p_Fm->p_FmStateStruct->fmClkFreq) > DMA_MAX_WATCHDOG)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE,
-+ ("dma_watchdog depends on FM clock. dma_watchdog(in microseconds) * clk (in Mhz), may not exceed 0x08x", DMA_MAX_WATCHDOG));
-+
-+#if (DPAA_VERSION >= 11)
-+ if ((p_Fm->partVSPBase + p_Fm->partNumOfVSPs) > FM_VSP_MAX_NUM_OF_ENTRIES)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("partVSPBase+partNumOfVSPs out of range!!!"));
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+ if (p_Fm->p_FmStateStruct->totalFifoSize % BMI_FIFO_UNITS)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("totalFifoSize number has to be divisible by %d", BMI_FIFO_UNITS));
-+ if (!p_Fm->p_FmStateStruct->totalFifoSize ||
-+ (p_Fm->p_FmStateStruct->totalFifoSize > BMI_MAX_FIFO_SIZE))
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE,
-+ ("totalFifoSize (currently defined as %d) has to be in the range of 256 to %d",
-+ p_Fm->p_FmStateStruct->totalFifoSize,
-+ BMI_MAX_FIFO_SIZE));
-+ if (!p_Fm->p_FmStateStruct->totalNumOfTasks ||
-+ (p_Fm->p_FmStateStruct->totalNumOfTasks > BMI_MAX_NUM_OF_TASKS))
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("totalNumOfTasks number has to be in the range 1 - %d", BMI_MAX_NUM_OF_TASKS));
-+
-+#ifdef FM_HAS_TOTAL_DMAS
-+ if (!p_Fm->p_FmStateStruct->maxNumOfOpenDmas ||
-+ (p_Fm->p_FmStateStruct->maxNumOfOpenDmas > BMI_MAX_NUM_OF_DMAS))
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("maxNumOfOpenDmas number has to be in the range 1 - %d", BMI_MAX_NUM_OF_DMAS));
-+#endif /* FM_HAS_TOTAL_DMAS */
-+
-+ if (p_Fm->p_FmDriverParam->disp_limit_tsh > FPM_MAX_DISP_LIMIT)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("disp_limit_tsh can't be greater than %d", FPM_MAX_DISP_LIMIT));
-+
-+ if (!p_Fm->f_Exception)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("Exceptions callback not provided"));
-+ if (!p_Fm->f_BusError)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("Exceptions callback not provided"));
-+
-+#ifdef FM_NO_WATCHDOG
-+ if ((p_Fm->p_FmStateStruct->revInfo.majorRev == 2) &&
-+ (p_Fm->p_FmDriverParam->dma_watchdog))
-+ RETURN_ERROR(MAJOR, E_NOT_SUPPORTED, ("watchdog!"));
-+#endif /* FM_NO_WATCHDOG */
-+
-+#ifdef FM_ECC_HALT_NO_SYNC_ERRATA_10GMAC_A008
-+ if ((p_Fm->p_FmStateStruct->revInfo.majorRev < 6) &&
-+ (p_Fm->p_FmDriverParam->halt_on_unrecov_ecc_err))
-+ RETURN_ERROR(MAJOR, E_NOT_SUPPORTED, ("HaltOnEccError!"));
-+#endif /* FM_ECC_HALT_NO_SYNC_ERRATA_10GMAC_A008 */
-+
-+#ifdef FM_NO_TNUM_AGING
-+ if ((p_Fm->p_FmStateStruct->revInfo.majorRev != 4) &&
-+ (p_Fm->p_FmStateStruct->revInfo.majorRev < 6))
-+ if (p_Fm->p_FmDriverParam->tnum_aging_period)
-+ RETURN_ERROR(MAJOR, E_NOT_SUPPORTED, ("Tnum aging!"));
-+#endif /* FM_NO_TNUM_AGING */
-+
-+ /* check that user did not set revision-dependent exceptions */
-+#ifdef FM_NO_DISPATCH_RAM_ECC
-+ if ((p_Fm->p_FmStateStruct->revInfo.majorRev != 4) &&
-+ (p_Fm->p_FmStateStruct->revInfo.majorRev < 6))
-+ if (p_Fm->userSetExceptions & FM_EX_BMI_DISPATCH_RAM_ECC)
-+ RETURN_ERROR(MAJOR, E_NOT_SUPPORTED, ("exception e_FM_EX_BMI_DISPATCH_RAM_ECC!"));
-+#endif /* FM_NO_DISPATCH_RAM_ECC */
-+
-+#ifdef FM_QMI_NO_ECC_EXCEPTIONS
-+ if (p_Fm->p_FmStateStruct->revInfo.majorRev == 4)
-+ if (p_Fm->userSetExceptions & (FM_EX_QMI_SINGLE_ECC | FM_EX_QMI_DOUBLE_ECC))
-+ RETURN_ERROR(MAJOR, E_NOT_SUPPORTED, ("exception e_FM_EX_QMI_SINGLE_ECC/e_FM_EX_QMI_DOUBLE_ECC!"));
-+#endif /* FM_QMI_NO_ECC_EXCEPTIONS */
-+
-+#ifdef FM_QMI_NO_SINGLE_ECC_EXCEPTION
-+ if (p_Fm->p_FmStateStruct->revInfo.majorRev >= 6)
-+ if (p_Fm->userSetExceptions & FM_EX_QMI_SINGLE_ECC)
-+ RETURN_ERROR(MAJOR, E_NOT_SUPPORTED, ("exception e_FM_EX_QMI_SINGLE_ECC!"));
-+#endif /* FM_QMI_NO_SINGLE_ECC_EXCEPTION */
-+
-+ return E_OK;
-+}
-+
-+
-+static void SendIpcIsr(t_Fm *p_Fm, uint32_t macEvent, uint32_t pendingReg)
-+{
-+ ASSERT_COND(p_Fm->guestId == NCSW_MASTER_ID);
-+
-+ if (p_Fm->intrMng[macEvent].guestId == NCSW_MASTER_ID)
-+ p_Fm->intrMng[macEvent].f_Isr(p_Fm->intrMng[macEvent].h_SrcHandle);
-+
-+ /* If the MAC is running on guest-partition and we have IPC session with it,
-+ we inform him about the event through IPC; otherwise, we ignore the event. */
-+ else if (p_Fm->h_IpcSessions[p_Fm->intrMng[macEvent].guestId])
-+ {
-+ t_Error err;
-+ t_FmIpcIsr fmIpcIsr;
-+ t_FmIpcMsg msg;
-+
-+ memset(&msg, 0, sizeof(msg));
-+ msg.msgId = FM_GUEST_ISR;
-+ fmIpcIsr.pendingReg = pendingReg;
-+ fmIpcIsr.boolErr = FALSE;
-+ memcpy(msg.msgBody, &fmIpcIsr, sizeof(fmIpcIsr));
-+ err = XX_IpcSendMessage(p_Fm->h_IpcSessions[p_Fm->intrMng[macEvent].guestId],
-+ (uint8_t*)&msg,
-+ sizeof(msg.msgId) + sizeof(fmIpcIsr),
-+ NULL,
-+ NULL,
-+ NULL,
-+ NULL);
-+ if (err != E_OK)
-+ REPORT_ERROR(MINOR, err, NO_MSG);
-+ }
-+ else
-+ DBG(TRACE, ("FM Guest mode, without IPC - can't call ISR!"));
-+}
-+
-+static void BmiErrEvent(t_Fm *p_Fm)
-+{
-+ uint32_t event;
-+ struct fman_bmi_regs *bmi_rg = p_Fm->p_FmBmiRegs;
-+
-+
-+ event = fman_get_bmi_err_event(bmi_rg);
-+
-+ if (event & BMI_ERR_INTR_EN_STORAGE_PROFILE_ECC)
-+ p_Fm->f_Exception(p_Fm->h_App,e_FM_EX_BMI_STORAGE_PROFILE_ECC);
-+ if (event & BMI_ERR_INTR_EN_LIST_RAM_ECC)
-+ p_Fm->f_Exception(p_Fm->h_App,e_FM_EX_BMI_LIST_RAM_ECC);
-+ if (event & BMI_ERR_INTR_EN_STATISTICS_RAM_ECC)
-+ p_Fm->f_Exception(p_Fm->h_App,e_FM_EX_BMI_STATISTICS_RAM_ECC);
-+ if (event & BMI_ERR_INTR_EN_DISPATCH_RAM_ECC)
-+ p_Fm->f_Exception(p_Fm->h_App,e_FM_EX_BMI_DISPATCH_RAM_ECC);
-+}
-+
-+static void QmiErrEvent(t_Fm *p_Fm)
-+{
-+ uint32_t event;
-+ struct fman_qmi_regs *qmi_rg = p_Fm->p_FmQmiRegs;
-+
-+ event = fman_get_qmi_err_event(qmi_rg);
-+
-+ if (event & QMI_ERR_INTR_EN_DOUBLE_ECC)
-+ p_Fm->f_Exception(p_Fm->h_App,e_FM_EX_QMI_DOUBLE_ECC);
-+ if (event & QMI_ERR_INTR_EN_DEQ_FROM_DEF)
-+ p_Fm->f_Exception(p_Fm->h_App,e_FM_EX_QMI_DEQ_FROM_UNKNOWN_PORTID);
-+}
-+
-+static void DmaErrEvent(t_Fm *p_Fm)
-+{
-+ uint32_t status, com_id;
-+ uint8_t tnum;
-+ uint8_t hardwarePortId;
-+ uint8_t relativePortId;
-+ uint16_t liodn;
-+ struct fman_dma_regs *dma_rg = p_Fm->p_FmDmaRegs;
-+
-+ status = fman_get_dma_err_event(dma_rg);
-+
-+ if (status & DMA_STATUS_BUS_ERR)
-+ {
-+ com_id = fman_get_dma_com_id(dma_rg);
-+ hardwarePortId = (uint8_t)(((com_id & DMA_TRANSFER_PORTID_MASK) >> DMA_TRANSFER_PORTID_SHIFT));
-+ ASSERT_COND(IN_RANGE(1, hardwarePortId, 63));
-+ HW_PORT_ID_TO_SW_PORT_ID(relativePortId, hardwarePortId);
-+ tnum = (uint8_t)((com_id & DMA_TRANSFER_TNUM_MASK) >> DMA_TRANSFER_TNUM_SHIFT);
-+ liodn = (uint16_t)(com_id & DMA_TRANSFER_LIODN_MASK);
-+ ASSERT_COND(p_Fm->p_FmStateStruct->portsTypes[hardwarePortId] != e_FM_PORT_TYPE_DUMMY);
-+ p_Fm->f_BusError(p_Fm->h_App,
-+ p_Fm->p_FmStateStruct->portsTypes[hardwarePortId],
-+ relativePortId,
-+ fman_get_dma_addr(dma_rg),
-+ tnum,
-+ liodn);
-+ }
-+ if (status & DMA_STATUS_FM_SPDAT_ECC)
-+ p_Fm->f_Exception(p_Fm->h_App, e_FM_EX_DMA_SINGLE_PORT_ECC);
-+ if (status & DMA_STATUS_READ_ECC)
-+ p_Fm->f_Exception(p_Fm->h_App, e_FM_EX_DMA_READ_ECC);
-+ if (status & DMA_STATUS_SYSTEM_WRITE_ECC)
-+ p_Fm->f_Exception(p_Fm->h_App, e_FM_EX_DMA_SYSTEM_WRITE_ECC);
-+ if (status & DMA_STATUS_FM_WRITE_ECC)
-+ p_Fm->f_Exception(p_Fm->h_App, e_FM_EX_DMA_FM_WRITE_ECC);
-+ }
-+
-+static void FpmErrEvent(t_Fm *p_Fm)
-+{
-+ uint32_t event;
-+ struct fman_fpm_regs *fpm_rg = p_Fm->p_FmFpmRegs;
-+
-+ event = fman_get_fpm_err_event(fpm_rg);
-+
-+ if ((event & FPM_EV_MASK_DOUBLE_ECC) && (event & FPM_EV_MASK_DOUBLE_ECC_EN))
-+ p_Fm->f_Exception(p_Fm->h_App,e_FM_EX_FPM_DOUBLE_ECC);
-+ if ((event & FPM_EV_MASK_STALL) && (event & FPM_EV_MASK_STALL_EN))
-+ p_Fm->f_Exception(p_Fm->h_App,e_FM_EX_FPM_STALL_ON_TASKS);
-+ if ((event & FPM_EV_MASK_SINGLE_ECC) && (event & FPM_EV_MASK_SINGLE_ECC_EN))
-+ p_Fm->f_Exception(p_Fm->h_App,e_FM_EX_FPM_SINGLE_ECC);
-+}
-+
-+static void MuramErrIntr(t_Fm *p_Fm)
-+{
-+ uint32_t event;
-+ struct fman_fpm_regs *fpm_rg = p_Fm->p_FmFpmRegs;
-+
-+ event = fman_get_muram_err_event(fpm_rg);
-+
-+ if (event & FPM_RAM_MURAM_ECC)
-+ p_Fm->f_Exception(p_Fm->h_App, e_FM_EX_MURAM_ECC);
-+}
-+
-+static void IramErrIntr(t_Fm *p_Fm)
-+{
-+ uint32_t event;
-+ struct fman_fpm_regs *fpm_rg = p_Fm->p_FmFpmRegs;
-+
-+ event = fman_get_iram_err_event(fpm_rg);
-+
-+ if (event & FPM_RAM_IRAM_ECC)
-+ p_Fm->f_Exception(p_Fm->h_App, e_FM_EX_IRAM_ECC);
-+}
-+
-+static void QmiEvent(t_Fm *p_Fm)
-+{
-+ uint32_t event;
-+ struct fman_qmi_regs *qmi_rg = p_Fm->p_FmQmiRegs;
-+
-+ event = fman_get_qmi_event(qmi_rg);
-+
-+ if (event & QMI_INTR_EN_SINGLE_ECC)
-+ p_Fm->f_Exception(p_Fm->h_App,e_FM_EX_QMI_SINGLE_ECC);
-+}
-+
-+static void UnimplementedIsr(t_Handle h_Arg)
-+{
-+ UNUSED(h_Arg);
-+
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Unimplemented ISR!"));
-+}
-+
-+static void UnimplementedFmanCtrlIsr(t_Handle h_Arg, uint32_t event)
-+{
-+ UNUSED(h_Arg); UNUSED(event);
-+
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Unimplemented FmCtl ISR!"));
-+}
-+
-+static void EnableTimeStamp(t_Fm *p_Fm)
-+{
-+ struct fman_fpm_regs *fpm_rg = p_Fm->p_FmFpmRegs;
-+
-+ ASSERT_COND(p_Fm->p_FmStateStruct);
-+ ASSERT_COND(p_Fm->p_FmStateStruct->count1MicroBit);
-+
-+ fman_enable_time_stamp(fpm_rg, p_Fm->p_FmStateStruct->count1MicroBit, p_Fm->p_FmStateStruct->fmClkFreq);
-+
-+ p_Fm->p_FmStateStruct->enabledTimeStamp = TRUE;
-+}
-+
-+static t_Error ClearIRam(t_Fm *p_Fm)
-+{
-+ t_FMIramRegs *p_Iram;
-+ int i;
-+ int iram_size;
-+
-+ ASSERT_COND(p_Fm);
-+ p_Iram = (t_FMIramRegs *)UINT_TO_PTR(p_Fm->baseAddr + FM_MM_IMEM);
-+ iram_size = FM_IRAM_SIZE(p_Fm->p_FmStateStruct->revInfo.majorRev,p_Fm->p_FmStateStruct->revInfo.minorRev);
-+
-+ /* Enable the auto-increment */
-+ WRITE_UINT32(p_Iram->iadd, IRAM_IADD_AIE);
-+ while (GET_UINT32(p_Iram->iadd) != IRAM_IADD_AIE) ;
-+
-+ for (i=0; i < (iram_size/4); i++)
-+ WRITE_UINT32(p_Iram->idata, 0xffffffff);
-+
-+ WRITE_UINT32(p_Iram->iadd, iram_size - 4);
-+ CORE_MemoryBarrier();
-+ while (GET_UINT32(p_Iram->idata) != 0xffffffff) ;
-+
-+ return E_OK;
-+}
-+
-+static t_Error LoadFmanCtrlCode(t_Fm *p_Fm)
-+{
-+ t_FMIramRegs *p_Iram;
-+ int i;
-+ uint32_t tmp;
-+ uint8_t compTo16;
-+
-+ ASSERT_COND(p_Fm);
-+ p_Iram = (t_FMIramRegs *)UINT_TO_PTR(p_Fm->baseAddr + FM_MM_IMEM);
-+
-+ /* Enable the auto-increment */
-+ WRITE_UINT32(p_Iram->iadd, IRAM_IADD_AIE);
-+ while (GET_UINT32(p_Iram->iadd) != IRAM_IADD_AIE) ;
-+
-+ for (i=0; i < (p_Fm->firmware.size / 4); i++)
-+ WRITE_UINT32(p_Iram->idata, p_Fm->firmware.p_Code[i]);
-+
-+ compTo16 = (uint8_t)(p_Fm->firmware.size % 16);
-+ if (compTo16)
-+ for (i=0; i < ((16-compTo16) / 4); i++)
-+ WRITE_UINT32(p_Iram->idata, 0xffffffff);
-+
-+ WRITE_UINT32(p_Iram->iadd,p_Fm->firmware.size-4);
-+ while (GET_UINT32(p_Iram->iadd) != (p_Fm->firmware.size-4)) ;
-+
-+ /* verify that writing has completed */
-+ while (GET_UINT32(p_Iram->idata) != p_Fm->firmware.p_Code[(p_Fm->firmware.size / 4)-1]) ;
-+
-+ if (p_Fm->fwVerify)
-+ {
-+ WRITE_UINT32(p_Iram->iadd, IRAM_IADD_AIE);
-+ while (GET_UINT32(p_Iram->iadd) != IRAM_IADD_AIE) ;
-+ for (i=0; i < (p_Fm->firmware.size / 4); i++)
-+ {
-+ tmp = GET_UINT32(p_Iram->idata);
-+ if (tmp != p_Fm->firmware.p_Code[i])
-+ RETURN_ERROR(MAJOR, E_WRITE_FAILED,
-+ ("UCode write error : write 0x%x, read 0x%x",
-+ p_Fm->firmware.p_Code[i],tmp));
-+ }
-+ WRITE_UINT32(p_Iram->iadd, 0x0);
-+ }
-+
-+ /* Enable patch from IRAM */
-+ WRITE_UINT32(p_Iram->iready, IRAM_READY);
-+ XX_UDelay(1000);
-+
-+ DBG(INFO, ("FMan-Controller code (ver %d.%d.%d) loaded to IRAM.",
-+ ((uint16_t *)p_Fm->firmware.p_Code)[2],
-+ ((uint8_t *)p_Fm->firmware.p_Code)[6],
-+ ((uint8_t *)p_Fm->firmware.p_Code)[7]));
-+
-+ return E_OK;
-+}
-+
-+#ifdef FM_UCODE_NOT_RESET_ERRATA_BUGZILLA6173
-+static t_Error FwNotResetErratumBugzilla6173WA(t_Fm *p_Fm)
-+{
-+ t_FMIramRegs *p_Iram = (t_FMIramRegs *)UINT_TO_PTR(p_Fm->baseAddr + FM_MM_IMEM);
-+ uint32_t tmpReg;
-+ uint32_t savedSpliodn[63];
-+
-+ /* write to IRAM first location the debug instruction */
-+ WRITE_UINT32(p_Iram->iadd, 0);
-+ while (GET_UINT32(p_Iram->iadd) != 0) ;
-+ WRITE_UINT32(p_Iram->idata, FM_FW_DEBUG_INSTRUCTION);
-+
-+ WRITE_UINT32(p_Iram->iadd, 0);
-+ while (GET_UINT32(p_Iram->iadd) != 0) ;
-+ while (GET_UINT32(p_Iram->idata) != FM_FW_DEBUG_INSTRUCTION) ;
-+
-+ /* Enable patch from IRAM */
-+ WRITE_UINT32(p_Iram->iready, IRAM_READY);
-+ CORE_MemoryBarrier();
-+ XX_UDelay(100);
-+ IO2MemCpy32((uint8_t *)savedSpliodn,
-+ (uint8_t *)p_Fm->p_FmBmiRegs->fmbm_spliodn,
-+ 63*sizeof(uint32_t));
-+
-+ /* reset FMAN */
-+ WRITE_UINT32(p_Fm->p_FmFpmRegs->fm_rstc, FPM_RSTC_FM_RESET);
-+ CORE_MemoryBarrier();
-+ XX_UDelay(100);
-+
-+ /* verify breakpoint debug status register */
-+ tmpReg = GET_UINT32(*(uint32_t *)UINT_TO_PTR(p_Fm->baseAddr + FM_DEBUG_STATUS_REGISTER_OFFSET));
-+ if (!tmpReg)
-+ REPORT_ERROR(MAJOR, E_INVALID_STATE, ("Invalid debug status register value is '0'"));
-+
-+ /*************************************/
-+ /* Load FMan-Controller code to IRAM */
-+ /*************************************/
-+ ClearIRam(p_Fm);
-+ if (p_Fm->firmware.p_Code &&
-+ (LoadFmanCtrlCode(p_Fm) != E_OK))
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, NO_MSG);
-+ XX_UDelay(100);
-+
-+ /* reset FMAN again to start the microcode */
-+ WRITE_UINT32(p_Fm->p_FmFpmRegs->fm_rstc, FPM_RSTC_FM_RESET);
-+ CORE_MemoryBarrier();
-+ XX_UDelay(100);
-+ Mem2IOCpy32((uint8_t *)p_Fm->p_FmBmiRegs->fmbm_spliodn,
-+ (uint8_t *)savedSpliodn,
-+ 63*sizeof(uint32_t));
-+
-+ if (fman_is_qmi_halt_not_busy_state(p_Fm->p_FmQmiRegs))
-+ {
-+ fman_resume(p_Fm->p_FmFpmRegs);
-+ CORE_MemoryBarrier();
-+ XX_UDelay(100);
-+ }
-+
-+ return E_OK;
-+}
-+#endif /* FM_UCODE_NOT_RESET_ERRATA_BUGZILLA6173 */
-+
-+static void GuestErrorIsr(t_Fm *p_Fm, uint32_t pending)
-+{
-+#define FM_G_CALL_1G_MAC_ERR_ISR(_id) \
-+do { \
-+ p_Fm->intrMng[(e_FmInterModuleEvent)(e_FM_EV_ERR_1G_MAC0+_id)].f_Isr(p_Fm->intrMng[(e_FmInterModuleEvent)(e_FM_EV_ERR_1G_MAC0+_id)].h_SrcHandle);\
-+} while (0)
-+#define FM_G_CALL_10G_MAC_ERR_ISR(_id) \
-+do { \
-+ p_Fm->intrMng[(e_FmInterModuleEvent)(e_FM_EV_ERR_10G_MAC0+_id)].f_Isr(p_Fm->intrMng[(e_FmInterModuleEvent)(e_FM_EV_ERR_10G_MAC0+_id)].h_SrcHandle);\
-+} while (0)
-+
-+ /* error interrupts */
-+ if (pending & ERR_INTR_EN_1G_MAC0)
-+ FM_G_CALL_1G_MAC_ERR_ISR(0);
-+ if (pending & ERR_INTR_EN_1G_MAC1)
-+ FM_G_CALL_1G_MAC_ERR_ISR(1);
-+ if (pending & ERR_INTR_EN_1G_MAC2)
-+ FM_G_CALL_1G_MAC_ERR_ISR(2);
-+ if (pending & ERR_INTR_EN_1G_MAC3)
-+ FM_G_CALL_1G_MAC_ERR_ISR(3);
-+ if (pending & ERR_INTR_EN_1G_MAC4)
-+ FM_G_CALL_1G_MAC_ERR_ISR(4);
-+ if (pending & ERR_INTR_EN_1G_MAC5)
-+ FM_G_CALL_1G_MAC_ERR_ISR(5);
-+ if (pending & ERR_INTR_EN_1G_MAC6)
-+ FM_G_CALL_1G_MAC_ERR_ISR(6);
-+ if (pending & ERR_INTR_EN_1G_MAC7)
-+ FM_G_CALL_1G_MAC_ERR_ISR(7);
-+ if (pending & ERR_INTR_EN_10G_MAC0)
-+ FM_G_CALL_10G_MAC_ERR_ISR(0);
-+ if (pending & ERR_INTR_EN_10G_MAC1)
-+ FM_G_CALL_10G_MAC_ERR_ISR(1);
-+}
-+
-+static void GuestEventIsr(t_Fm *p_Fm, uint32_t pending)
-+{
-+#define FM_G_CALL_1G_MAC_ISR(_id) \
-+do { \
-+ p_Fm->intrMng[(e_FmInterModuleEvent)(e_FM_EV_1G_MAC0+_id)].f_Isr(p_Fm->intrMng[(e_FmInterModuleEvent)(e_FM_EV_1G_MAC0+_id)].h_SrcHandle);\
-+} while (0)
-+#define FM_G_CALL_10G_MAC_ISR(_id) \
-+do { \
-+ p_Fm->intrMng[(e_FmInterModuleEvent)(e_FM_EV_10G_MAC0+_id)].f_Isr(p_Fm->intrMng[(e_FmInterModuleEvent)(e_FM_EV_10G_MAC0+_id)].h_SrcHandle);\
-+} while (0)
-+
-+ if (pending & INTR_EN_1G_MAC0)
-+ FM_G_CALL_1G_MAC_ISR(0);
-+ if (pending & INTR_EN_1G_MAC1)
-+ FM_G_CALL_1G_MAC_ISR(1);
-+ if (pending & INTR_EN_1G_MAC2)
-+ FM_G_CALL_1G_MAC_ISR(2);
-+ if (pending & INTR_EN_1G_MAC3)
-+ FM_G_CALL_1G_MAC_ISR(3);
-+ if (pending & INTR_EN_1G_MAC4)
-+ FM_G_CALL_1G_MAC_ISR(4);
-+ if (pending & INTR_EN_1G_MAC5)
-+ FM_G_CALL_1G_MAC_ISR(5);
-+ if (pending & INTR_EN_1G_MAC6)
-+ FM_G_CALL_1G_MAC_ISR(6);
-+ if (pending & INTR_EN_1G_MAC7)
-+ FM_G_CALL_1G_MAC_ISR(7);
-+ if (pending & INTR_EN_10G_MAC0)
-+ FM_G_CALL_10G_MAC_ISR(0);
-+ if (pending & INTR_EN_10G_MAC1)
-+ FM_G_CALL_10G_MAC_ISR(1);
-+ if (pending & INTR_EN_TMR)
-+ p_Fm->intrMng[e_FM_EV_TMR].f_Isr(p_Fm->intrMng[e_FM_EV_TMR].h_SrcHandle);
-+}
-+
-+#if (DPAA_VERSION >= 11)
-+static t_Error SetVSPWindow(t_Handle h_Fm,
-+ uint8_t hardwarePortId,
-+ uint8_t baseStorageProfile,
-+ uint8_t log2NumOfProfiles)
-+{
-+ t_Fm *p_Fm = (t_Fm *)h_Fm;
-+
-+ ASSERT_COND(h_Fm);
-+ ASSERT_COND(IN_RANGE(1, hardwarePortId, 63));
-+
-+ if ((p_Fm->guestId != NCSW_MASTER_ID) &&
-+ !p_Fm->p_FmBmiRegs &&
-+ p_Fm->h_IpcSessions[0])
-+ {
-+ t_FmIpcVspSetPortWindow fmIpcVspSetPortWindow;
-+ t_FmIpcMsg msg;
-+ t_Error err = E_OK;
-+
-+ memset(&msg, 0, sizeof(msg));
-+ memset(&fmIpcVspSetPortWindow, 0, sizeof(t_FmIpcVspSetPortWindow));
-+ fmIpcVspSetPortWindow.hardwarePortId = hardwarePortId;
-+ fmIpcVspSetPortWindow.baseStorageProfile = baseStorageProfile;
-+ fmIpcVspSetPortWindow.log2NumOfProfiles = log2NumOfProfiles;
-+ msg.msgId = FM_VSP_SET_PORT_WINDOW;
-+ memcpy(msg.msgBody, &fmIpcVspSetPortWindow, sizeof(t_FmIpcVspSetPortWindow));
-+
-+ err = XX_IpcSendMessage(p_Fm->h_IpcSessions[0],
-+ (uint8_t*)&msg,
-+ sizeof(msg.msgId),
-+ NULL,
-+ NULL,
-+ NULL,
-+ NULL);
-+ if (err != E_OK)
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+ return E_OK;
-+ }
-+ else if (!p_Fm->p_FmBmiRegs)
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED,
-+ ("Either IPC or 'baseAddress' is required!"));
-+
-+ fman_set_vsp_window(p_Fm->p_FmBmiRegs,
-+ hardwarePortId,
-+ baseStorageProfile,
-+ log2NumOfProfiles);
-+
-+ return E_OK;
-+}
-+
-+static uint8_t AllocVSPsForPartition(t_Handle h_Fm, uint8_t base, uint8_t numOfProfiles, uint8_t guestId)
-+{
-+ t_Fm *p_Fm = (t_Fm *)h_Fm;
-+ uint8_t profilesFound = 0;
-+ int i = 0;
-+ uint32_t intFlags;
-+
-+ if (!numOfProfiles)
-+ return E_OK;
-+
-+ if ((numOfProfiles > FM_VSP_MAX_NUM_OF_ENTRIES) ||
-+ (base + numOfProfiles > FM_VSP_MAX_NUM_OF_ENTRIES))
-+ return (uint8_t)ILLEGAL_BASE;
-+
-+ if (p_Fm->h_IpcSessions[0])
-+ {
-+ t_FmIpcResourceAllocParams ipcAllocParams;
-+ t_FmIpcMsg msg;
-+ t_FmIpcReply reply;
-+ t_Error err;
-+ uint32_t replyLength;
-+
-+ memset(&msg, 0, sizeof(msg));
-+ memset(&reply, 0, sizeof(reply));
-+ memset(&ipcAllocParams, 0, sizeof(t_FmIpcResourceAllocParams));
-+ ipcAllocParams.guestId = p_Fm->guestId;
-+ ipcAllocParams.num = p_Fm->partNumOfVSPs;
-+ ipcAllocParams.base = p_Fm->partVSPBase;
-+ msg.msgId = FM_VSP_ALLOC;
-+ memcpy(msg.msgBody, &ipcAllocParams, sizeof(t_FmIpcResourceAllocParams));
-+ replyLength = sizeof(uint32_t) + sizeof(uint8_t);
-+ err = XX_IpcSendMessage(p_Fm->h_IpcSessions[0],
-+ (uint8_t*)&msg,
-+ sizeof(msg.msgId) + sizeof(t_FmIpcResourceAllocParams),
-+ (uint8_t*)&reply,
-+ &replyLength,
-+ NULL,
-+ NULL);
-+ if ((err != E_OK) ||
-+ (replyLength != (sizeof(uint32_t) + sizeof(uint8_t))))
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ else
-+ memcpy((uint8_t*)&p_Fm->partVSPBase, reply.replyBody, sizeof(uint8_t));
-+ if (p_Fm->partVSPBase == (uint8_t)(ILLEGAL_BASE))
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+ if (p_Fm->guestId != NCSW_MASTER_ID)
-+ {
-+ DBG(WARNING, ("FM Guest mode, without IPC - can't validate VSP range!"));
-+ return (uint8_t)ILLEGAL_BASE;
-+ }
-+
-+ intFlags = XX_LockIntrSpinlock(p_Fm->h_Spinlock);
-+ for (i = base; i < base + numOfProfiles; i++)
-+ if (p_Fm->p_FmSp->profiles[i].profilesMng.ownerId == (uint8_t)ILLEGAL_BASE)
-+ profilesFound++;
-+ else
-+ break;
-+
-+ if (profilesFound == numOfProfiles)
-+ for (i = base; i<base + numOfProfiles; i++)
-+ p_Fm->p_FmSp->profiles[i].profilesMng.ownerId = guestId;
-+ else
-+ {
-+ XX_UnlockIntrSpinlock(p_Fm->h_Spinlock, intFlags);
-+ return (uint8_t)ILLEGAL_BASE;
-+ }
-+ XX_UnlockIntrSpinlock(p_Fm->h_Spinlock, intFlags);
-+
-+ return base;
-+}
-+
-+static void FreeVSPsForPartition(t_Handle h_Fm, uint8_t base, uint8_t numOfProfiles, uint8_t guestId)
-+{
-+ t_Fm *p_Fm = (t_Fm *)h_Fm;
-+ int i = 0;
-+
-+ ASSERT_COND(p_Fm);
-+
-+ if (p_Fm->h_IpcSessions[0])
-+ {
-+ t_FmIpcResourceAllocParams ipcAllocParams;
-+ t_FmIpcMsg msg;
-+ t_FmIpcReply reply;
-+ uint32_t replyLength;
-+ t_Error err;
-+
-+ memset(&msg, 0, sizeof(msg));
-+ memset(&reply, 0, sizeof(reply));
-+ memset(&ipcAllocParams, 0, sizeof(t_FmIpcResourceAllocParams));
-+ ipcAllocParams.guestId = p_Fm->guestId;
-+ ipcAllocParams.num = p_Fm->partNumOfVSPs;
-+ ipcAllocParams.base = p_Fm->partVSPBase;
-+ msg.msgId = FM_VSP_FREE;
-+ memcpy(msg.msgBody, &ipcAllocParams, sizeof(t_FmIpcResourceAllocParams));
-+ replyLength = sizeof(uint32_t) + sizeof(uint8_t);
-+ err = XX_IpcSendMessage(p_Fm->h_IpcSessions[0],
-+ (uint8_t*)&msg,
-+ sizeof(msg.msgId) + sizeof(t_FmIpcResourceAllocParams),
-+ (uint8_t*)&reply,
-+ &replyLength,
-+ NULL,
-+ NULL);
-+ if (err != E_OK)
-+ REPORT_ERROR(MAJOR, err, NO_MSG);
-+ return;
-+ }
-+ if (p_Fm->guestId != NCSW_MASTER_ID)
-+ {
-+ DBG(WARNING, ("FM Guest mode, without IPC - can't validate VSP range!"));
-+ return;
-+ }
-+
-+ ASSERT_COND(p_Fm->p_FmSp);
-+
-+ for (i=base; i<numOfProfiles; i++)
-+ {
-+ if (p_Fm->p_FmSp->profiles[i].profilesMng.ownerId == guestId)
-+ p_Fm->p_FmSp->profiles[i].profilesMng.ownerId = (uint8_t)ILLEGAL_BASE;
-+ else
-+ DBG(WARNING, ("Request for freeing storage profile window which wasn't allocated to this partition"));
-+ }
-+}
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+static t_Error FmGuestHandleIpcMsgCB(t_Handle h_Fm,
-+ uint8_t *p_Msg,
-+ uint32_t msgLength,
-+ uint8_t *p_Reply,
-+ uint32_t *p_ReplyLength)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+ t_FmIpcMsg *p_IpcMsg = (t_FmIpcMsg*)p_Msg;
-+
-+ UNUSED(p_Reply);
-+ SANITY_CHECK_RETURN_ERROR(p_Fm, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR((msgLength > sizeof(uint32_t)), E_INVALID_VALUE);
-+
-+#ifdef DISABLE_SANITY_CHECKS
-+ UNUSED(msgLength);
-+#endif /* DISABLE_SANITY_CHECKS */
-+
-+ ASSERT_COND(p_Msg);
-+
-+ *p_ReplyLength = 0;
-+
-+ switch (p_IpcMsg->msgId)
-+ {
-+ case (FM_GUEST_ISR):
-+ {
-+ t_FmIpcIsr ipcIsr;
-+
-+ memcpy((uint8_t*)&ipcIsr, p_IpcMsg->msgBody, sizeof(t_FmIpcIsr));
-+ if (ipcIsr.boolErr)
-+ GuestErrorIsr(p_Fm, ipcIsr.pendingReg);
-+ else
-+ GuestEventIsr(p_Fm, ipcIsr.pendingReg);
-+ break;
-+ }
-+ default:
-+ *p_ReplyLength = 0;
-+ RETURN_ERROR(MINOR, E_INVALID_SELECTION, ("command not found!!!"));
-+ }
-+ return E_OK;
-+}
-+
-+static t_Error FmHandleIpcMsgCB(t_Handle h_Fm,
-+ uint8_t *p_Msg,
-+ uint32_t msgLength,
-+ uint8_t *p_Reply,
-+ uint32_t *p_ReplyLength)
-+{
-+ t_Error err;
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+ t_FmIpcMsg *p_IpcMsg = (t_FmIpcMsg*)p_Msg;
-+ t_FmIpcReply *p_IpcReply = (t_FmIpcReply*)p_Reply;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Fm, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR((msgLength >= sizeof(uint32_t)), E_INVALID_VALUE);
-+
-+#ifdef DISABLE_SANITY_CHECKS
-+ UNUSED(msgLength);
-+#endif /* DISABLE_SANITY_CHECKS */
-+
-+ ASSERT_COND(p_IpcMsg);
-+
-+ memset(p_IpcReply, 0, (sizeof(uint8_t) * FM_IPC_MAX_REPLY_SIZE));
-+ *p_ReplyLength = 0;
-+
-+ switch (p_IpcMsg->msgId)
-+ {
-+ case (FM_GET_SET_PORT_PARAMS):
-+ {
-+ t_FmIpcPortInInitParams ipcInitParams;
-+ t_FmInterModulePortInitParams initParams;
-+ t_FmIpcPortOutInitParams ipcOutInitParams;
-+
-+ memcpy((uint8_t*)&ipcInitParams, p_IpcMsg->msgBody, sizeof(t_FmIpcPortInInitParams));
-+ initParams.hardwarePortId = ipcInitParams.hardwarePortId;
-+ initParams.portType = (e_FmPortType)ipcInitParams.enumPortType;
-+ initParams.independentMode = (bool)(ipcInitParams.boolIndependentMode);
-+ initParams.liodnOffset = ipcInitParams.liodnOffset;
-+ initParams.numOfTasks = ipcInitParams.numOfTasks;
-+ initParams.numOfExtraTasks = ipcInitParams.numOfExtraTasks;
-+ initParams.numOfOpenDmas = ipcInitParams.numOfOpenDmas;
-+ initParams.numOfExtraOpenDmas = ipcInitParams.numOfExtraOpenDmas;
-+ initParams.sizeOfFifo = ipcInitParams.sizeOfFifo;
-+ initParams.extraSizeOfFifo = ipcInitParams.extraSizeOfFifo;
-+ initParams.deqPipelineDepth = ipcInitParams.deqPipelineDepth;
-+ initParams.maxFrameLength = ipcInitParams.maxFrameLength;
-+ initParams.liodnBase = ipcInitParams.liodnBase;
-+
-+ p_IpcReply->error = (uint32_t)FmGetSetPortParams(h_Fm, &initParams);
-+
-+ ipcOutInitParams.ipcPhysAddr.high = initParams.fmMuramPhysBaseAddr.high;
-+ ipcOutInitParams.ipcPhysAddr.low = initParams.fmMuramPhysBaseAddr.low;
-+ ipcOutInitParams.sizeOfFifo = initParams.sizeOfFifo;
-+ ipcOutInitParams.extraSizeOfFifo = initParams.extraSizeOfFifo;
-+ ipcOutInitParams.numOfTasks = initParams.numOfTasks;
-+ ipcOutInitParams.numOfExtraTasks = initParams.numOfExtraTasks;
-+ ipcOutInitParams.numOfOpenDmas = initParams.numOfOpenDmas;
-+ ipcOutInitParams.numOfExtraOpenDmas = initParams.numOfExtraOpenDmas;
-+ memcpy(p_IpcReply->replyBody, (uint8_t*)&ipcOutInitParams, sizeof(ipcOutInitParams));
-+ *p_ReplyLength = sizeof(uint32_t) + sizeof(t_FmIpcPortOutInitParams);
-+ break;
-+ }
-+ case (FM_SET_SIZE_OF_FIFO):
-+ {
-+ t_FmIpcPortRsrcParams ipcPortRsrcParams;
-+
-+ memcpy((uint8_t*)&ipcPortRsrcParams, p_IpcMsg->msgBody, sizeof(t_FmIpcPortRsrcParams));
-+ p_IpcReply->error = (uint32_t)FmSetSizeOfFifo(h_Fm,
-+ ipcPortRsrcParams.hardwarePortId,
-+ &ipcPortRsrcParams.val,
-+ &ipcPortRsrcParams.extra,
-+ (bool)ipcPortRsrcParams.boolInitialConfig);
-+ *p_ReplyLength = sizeof(uint32_t);
-+ break;
-+ }
-+ case (FM_SET_NUM_OF_TASKS):
-+ {
-+ t_FmIpcPortRsrcParams ipcPortRsrcParams;
-+
-+ memcpy((uint8_t*)&ipcPortRsrcParams, p_IpcMsg->msgBody, sizeof(t_FmIpcPortRsrcParams));
-+ p_IpcReply->error = (uint32_t)FmSetNumOfTasks(h_Fm, ipcPortRsrcParams.hardwarePortId,
-+ (uint8_t*)&ipcPortRsrcParams.val,
-+ (uint8_t*)&ipcPortRsrcParams.extra,
-+ (bool)ipcPortRsrcParams.boolInitialConfig);
-+ *p_ReplyLength = sizeof(uint32_t);
-+ break;
-+ }
-+ case (FM_SET_NUM_OF_OPEN_DMAS):
-+ {
-+ t_FmIpcPortRsrcParams ipcPortRsrcParams;
-+
-+ memcpy((uint8_t*)&ipcPortRsrcParams, p_IpcMsg->msgBody, sizeof(t_FmIpcPortRsrcParams));
-+ p_IpcReply->error = (uint32_t)FmSetNumOfOpenDmas(h_Fm, ipcPortRsrcParams.hardwarePortId,
-+ (uint8_t*)&ipcPortRsrcParams.val,
-+ (uint8_t*)&ipcPortRsrcParams.extra,
-+ (bool)ipcPortRsrcParams.boolInitialConfig);
-+ *p_ReplyLength = sizeof(uint32_t);
-+ break;
-+ }
-+ case (FM_RESUME_STALLED_PORT):
-+ *p_ReplyLength = sizeof(uint32_t);
-+ p_IpcReply->error = (uint32_t)FmResumeStalledPort(h_Fm, p_IpcMsg->msgBody[0]);
-+ break;
-+ case (FM_MASTER_IS_ALIVE):
-+ {
-+ uint8_t guestId = p_IpcMsg->msgBody[0];
-+ /* build the FM master partition IPC address */
-+ memset(p_Fm->fmIpcHandlerModuleName[guestId], 0, (sizeof(char)) * MODULE_NAME_SIZE);
-+ if (Sprint (p_Fm->fmIpcHandlerModuleName[guestId], "FM_%d_%d",p_Fm->p_FmStateStruct->fmId, guestId) != (guestId<10 ? 6:7))
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("Sprint failed"));
-+ p_Fm->h_IpcSessions[guestId] = XX_IpcInitSession(p_Fm->fmIpcHandlerModuleName[guestId], p_Fm->fmModuleName);
-+ if (p_Fm->h_IpcSessions[guestId] == NULL)
-+ RETURN_ERROR(MAJOR, E_NOT_AVAILABLE, ("FM Master IPC session for guest %d", guestId));
-+ *(uint8_t*)(p_IpcReply->replyBody) = 1;
-+ *p_ReplyLength = sizeof(uint32_t) + sizeof(uint8_t);
-+ break;
-+ }
-+ case (FM_IS_PORT_STALLED):
-+ {
-+ bool tmp;
-+
-+ p_IpcReply->error = (uint32_t)FmIsPortStalled(h_Fm, p_IpcMsg->msgBody[0], &tmp);
-+ *(uint8_t*)(p_IpcReply->replyBody) = (uint8_t)tmp;
-+ *p_ReplyLength = sizeof(uint32_t) + sizeof(uint8_t);
-+ break;
-+ }
-+ case (FM_RESET_MAC):
-+ {
-+ t_FmIpcMacParams ipcMacParams;
-+
-+ memcpy((uint8_t*)&ipcMacParams, p_IpcMsg->msgBody, sizeof(t_FmIpcMacParams));
-+ p_IpcReply->error = (uint32_t)FmResetMac(p_Fm,
-+ (e_FmMacType)(ipcMacParams.enumType),
-+ ipcMacParams.id);
-+ *p_ReplyLength = sizeof(uint32_t);
-+ break;
-+ }
-+ case (FM_SET_MAC_MAX_FRAME):
-+ {
-+ t_FmIpcMacMaxFrameParams ipcMacMaxFrameParams;
-+
-+ memcpy((uint8_t*)&ipcMacMaxFrameParams, p_IpcMsg->msgBody, sizeof(t_FmIpcMacMaxFrameParams));
-+ err = FmSetMacMaxFrame(p_Fm,
-+ (e_FmMacType)(ipcMacMaxFrameParams.macParams.enumType),
-+ ipcMacMaxFrameParams.macParams.id,
-+ ipcMacMaxFrameParams.maxFrameLength);
-+ if (err != E_OK)
-+ REPORT_ERROR(MINOR, err, NO_MSG);
-+ break;
-+ }
-+#if (DPAA_VERSION >= 11)
-+ case (FM_VSP_ALLOC) :
-+ {
-+ t_FmIpcResourceAllocParams ipcAllocParams;
-+ uint8_t vspBase;
-+ memcpy(&ipcAllocParams, p_IpcMsg->msgBody, sizeof(t_FmIpcResourceAllocParams));
-+ vspBase = AllocVSPsForPartition(h_Fm, (uint8_t)ipcAllocParams.base, (uint8_t)ipcAllocParams.num, ipcAllocParams.guestId);
-+ memcpy(p_IpcReply->replyBody, (uint8_t*)&vspBase, sizeof(uint8_t));
-+ *p_ReplyLength = sizeof(uint32_t) + sizeof(uint8_t);
-+ break;
-+ }
-+ case (FM_VSP_FREE) :
-+ {
-+ t_FmIpcResourceAllocParams ipcAllocParams;
-+ memcpy(&ipcAllocParams, p_IpcMsg->msgBody, sizeof(t_FmIpcResourceAllocParams));
-+ FreeVSPsForPartition(h_Fm, (uint8_t)ipcAllocParams.base, (uint8_t)ipcAllocParams.num, ipcAllocParams.guestId);
-+ break;
-+ }
-+ case (FM_VSP_SET_PORT_WINDOW) :
-+ {
-+ t_FmIpcVspSetPortWindow ipcVspSetPortWindow;
-+ memcpy(&ipcVspSetPortWindow, p_IpcMsg->msgBody, sizeof(t_FmIpcVspSetPortWindow));
-+ err = SetVSPWindow(h_Fm,
-+ ipcVspSetPortWindow.hardwarePortId,
-+ ipcVspSetPortWindow.baseStorageProfile,
-+ ipcVspSetPortWindow.log2NumOfProfiles);
-+ return err;
-+ }
-+ case (FM_SET_CONG_GRP_PFC_PRIO) :
-+ {
-+ t_FmIpcSetCongestionGroupPfcPriority fmIpcSetCongestionGroupPfcPriority;
-+ memcpy(&fmIpcSetCongestionGroupPfcPriority, p_IpcMsg->msgBody, sizeof(t_FmIpcSetCongestionGroupPfcPriority));
-+ err = FmSetCongestionGroupPFCpriority(h_Fm,
-+ fmIpcSetCongestionGroupPfcPriority.congestionGroupId,
-+ fmIpcSetCongestionGroupPfcPriority.priorityBitMap);
-+ return err;
-+ }
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+ case (FM_FREE_PORT):
-+ {
-+ t_FmInterModulePortFreeParams portParams;
-+ t_FmIpcPortFreeParams ipcPortParams;
-+
-+ memcpy((uint8_t*)&ipcPortParams, p_IpcMsg->msgBody, sizeof(t_FmIpcPortFreeParams));
-+ portParams.hardwarePortId = ipcPortParams.hardwarePortId;
-+ portParams.portType = (e_FmPortType)(ipcPortParams.enumPortType);
-+ portParams.deqPipelineDepth = ipcPortParams.deqPipelineDepth;
-+ FmFreePortParams(h_Fm, &portParams);
-+ break;
-+ }
-+ case (FM_REGISTER_INTR):
-+ {
-+ t_FmIpcRegisterIntr ipcRegIntr;
-+
-+ memcpy((uint8_t*)&ipcRegIntr, p_IpcMsg->msgBody, sizeof(ipcRegIntr));
-+ p_Fm->intrMng[ipcRegIntr.event].guestId = ipcRegIntr.guestId;
-+ break;
-+ }
-+ case (FM_GET_PARAMS):
-+ {
-+ t_FmIpcParams ipcParams;
-+
-+ /* Get clock frequency */
-+ ipcParams.fmClkFreq = p_Fm->p_FmStateStruct->fmClkFreq;
-+ ipcParams.fmMacClkFreq = p_Fm->p_FmStateStruct->fmMacClkFreq;
-+
-+ fman_get_revision(p_Fm->p_FmFpmRegs,&ipcParams.majorRev,&ipcParams.minorRev);
-+
-+ memcpy(p_IpcReply->replyBody, (uint8_t*)&ipcParams, sizeof(t_FmIpcParams));
-+ *p_ReplyLength = sizeof(uint32_t) + sizeof(t_FmIpcParams);
-+ break;
-+ }
-+ case (FM_GET_FMAN_CTRL_CODE_REV):
-+ {
-+ t_FmCtrlCodeRevisionInfo fmanCtrlRevInfo;
-+ t_FmIpcFmanCtrlCodeRevisionInfo ipcRevInfo;
-+
-+ p_IpcReply->error = (uint32_t)FM_GetFmanCtrlCodeRevision(h_Fm, &fmanCtrlRevInfo);
-+ ipcRevInfo.packageRev = fmanCtrlRevInfo.packageRev;
-+ ipcRevInfo.majorRev = fmanCtrlRevInfo.majorRev;
-+ ipcRevInfo.minorRev = fmanCtrlRevInfo.minorRev;
-+ memcpy(p_IpcReply->replyBody, (uint8_t*)&ipcRevInfo, sizeof(t_FmIpcFmanCtrlCodeRevisionInfo));
-+ *p_ReplyLength = sizeof(uint32_t) + sizeof(t_FmIpcFmanCtrlCodeRevisionInfo);
-+ break;
-+ }
-+
-+ case (FM_DMA_STAT):
-+ {
-+ t_FmDmaStatus dmaStatus;
-+ t_FmIpcDmaStatus ipcDmaStatus;
-+
-+ FM_GetDmaStatus(h_Fm, &dmaStatus);
-+ ipcDmaStatus.boolCmqNotEmpty = (uint8_t)dmaStatus.cmqNotEmpty;
-+ ipcDmaStatus.boolBusError = (uint8_t)dmaStatus.busError;
-+ ipcDmaStatus.boolReadBufEccError = (uint8_t)dmaStatus.readBufEccError;
-+ ipcDmaStatus.boolWriteBufEccSysError = (uint8_t)dmaStatus.writeBufEccSysError;
-+ ipcDmaStatus.boolWriteBufEccFmError = (uint8_t)dmaStatus.writeBufEccFmError;
-+ ipcDmaStatus.boolSinglePortEccError = (uint8_t)dmaStatus.singlePortEccError;
-+ memcpy(p_IpcReply->replyBody, (uint8_t*)&ipcDmaStatus, sizeof(t_FmIpcDmaStatus));
-+ *p_ReplyLength = sizeof(uint32_t) + sizeof(t_FmIpcDmaStatus);
-+ break;
-+ }
-+ case (FM_ALLOC_FMAN_CTRL_EVENT_REG):
-+ p_IpcReply->error = (uint32_t)FmAllocFmanCtrlEventReg(h_Fm, (uint8_t*)p_IpcReply->replyBody);
-+ *p_ReplyLength = sizeof(uint32_t) + sizeof(uint8_t);
-+ break;
-+ case (FM_FREE_FMAN_CTRL_EVENT_REG):
-+ FmFreeFmanCtrlEventReg(h_Fm, p_IpcMsg->msgBody[0]);
-+ break;
-+ case (FM_GET_TIMESTAMP_SCALE):
-+ {
-+ uint32_t timeStamp = FmGetTimeStampScale(h_Fm);
-+
-+ memcpy(p_IpcReply->replyBody, (uint8_t*)&timeStamp, sizeof(uint32_t));
-+ *p_ReplyLength = sizeof(uint32_t) + sizeof(uint32_t);
-+ break;
-+ }
-+ case (FM_GET_COUNTER):
-+ {
-+ e_FmCounters inCounter;
-+ uint32_t outCounter;
-+
-+ memcpy((uint8_t*)&inCounter, p_IpcMsg->msgBody, sizeof(uint32_t));
-+ outCounter = FM_GetCounter(h_Fm, inCounter);
-+ memcpy(p_IpcReply->replyBody, (uint8_t*)&outCounter, sizeof(uint32_t));
-+ *p_ReplyLength = sizeof(uint32_t) + sizeof(uint32_t);
-+ break;
-+ }
-+ case (FM_SET_FMAN_CTRL_EVENTS_ENABLE):
-+ {
-+ t_FmIpcFmanEvents ipcFmanEvents;
-+
-+ memcpy((uint8_t*)&ipcFmanEvents, p_IpcMsg->msgBody, sizeof(t_FmIpcFmanEvents));
-+ FmSetFmanCtrlIntr(h_Fm,
-+ ipcFmanEvents.eventRegId,
-+ ipcFmanEvents.enableEvents);
-+ break;
-+ }
-+ case (FM_GET_FMAN_CTRL_EVENTS_ENABLE):
-+ {
-+ uint32_t tmp = FmGetFmanCtrlIntr(h_Fm, p_IpcMsg->msgBody[0]);
-+
-+ memcpy(p_IpcReply->replyBody, (uint8_t*)&tmp, sizeof(uint32_t));
-+ *p_ReplyLength = sizeof(uint32_t) + sizeof(uint32_t);
-+ break;
-+ }
-+ case (FM_GET_PHYS_MURAM_BASE):
-+ {
-+ t_FmPhysAddr physAddr;
-+ t_FmIpcPhysAddr ipcPhysAddr;
-+
-+ FmGetPhysicalMuramBase(h_Fm, &physAddr);
-+ ipcPhysAddr.high = physAddr.high;
-+ ipcPhysAddr.low = physAddr.low;
-+ memcpy(p_IpcReply->replyBody, (uint8_t*)&ipcPhysAddr, sizeof(t_FmIpcPhysAddr));
-+ *p_ReplyLength = sizeof(uint32_t) + sizeof(t_FmIpcPhysAddr);
-+ break;
-+ }
-+ case (FM_ENABLE_RAM_ECC):
-+ {
-+ if (((err = FM_EnableRamsEcc(h_Fm)) != E_OK) ||
-+ ((err = FM_SetException(h_Fm, e_FM_EX_IRAM_ECC, TRUE)) != E_OK) ||
-+ ((err = FM_SetException(h_Fm, e_FM_EX_MURAM_ECC, TRUE)) != E_OK))
-+#if (!(defined(DEBUG_ERRORS)) || (DEBUG_ERRORS == 0))
-+ UNUSED(err);
-+#else
-+ REPORT_ERROR(MINOR, err, NO_MSG);
-+#endif /* (!(defined(DEBUG_ERRORS)) || (DEBUG_ERRORS == 0)) */
-+ break;
-+ }
-+ case (FM_DISABLE_RAM_ECC):
-+ {
-+
-+ if (((err = FM_SetException(h_Fm, e_FM_EX_IRAM_ECC, FALSE)) != E_OK) ||
-+ ((err = FM_SetException(h_Fm, e_FM_EX_MURAM_ECC, FALSE)) != E_OK) ||
-+ ((err = FM_DisableRamsEcc(h_Fm)) != E_OK))
-+#if (!(defined(DEBUG_ERRORS)) || (DEBUG_ERRORS == 0))
-+ UNUSED(err);
-+#else
-+ REPORT_ERROR(MINOR, err, NO_MSG);
-+#endif /* (!(defined(DEBUG_ERRORS)) || (DEBUG_ERRORS == 0)) */
-+ break;
-+ }
-+ case (FM_SET_NUM_OF_FMAN_CTRL):
-+ {
-+ t_FmIpcPortNumOfFmanCtrls ipcPortNumOfFmanCtrls;
-+
-+ memcpy((uint8_t*)&ipcPortNumOfFmanCtrls, p_IpcMsg->msgBody, sizeof(t_FmIpcPortNumOfFmanCtrls));
-+ err = FmSetNumOfRiscsPerPort(h_Fm,
-+ ipcPortNumOfFmanCtrls.hardwarePortId,
-+ ipcPortNumOfFmanCtrls.numOfFmanCtrls,
-+ ipcPortNumOfFmanCtrls.orFmanCtrl);
-+ if (err != E_OK)
-+ REPORT_ERROR(MINOR, err, NO_MSG);
-+ break;
-+ }
-+#ifdef FM_TX_ECC_FRMS_ERRATA_10GMAC_A004
-+ case (FM_10G_TX_ECC_WA):
-+ p_IpcReply->error = (uint32_t)Fm10GTxEccWorkaround(h_Fm, p_IpcMsg->msgBody[0]);
-+ *p_ReplyLength = sizeof(uint32_t);
-+ break;
-+#endif /* FM_TX_ECC_FRMS_ERRATA_10GMAC_A004 */
-+ default:
-+ *p_ReplyLength = 0;
-+ RETURN_ERROR(MINOR, E_INVALID_SELECTION, ("command not found!!!"));
-+ }
-+ return E_OK;
-+}
-+
-+
-+/****************************************/
-+/* Inter-Module functions */
-+/****************************************/
-+#ifdef FM_TX_ECC_FRMS_ERRATA_10GMAC_A004
-+t_Error Fm10GTxEccWorkaround(t_Handle h_Fm, uint8_t macId)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+ t_Error err = E_OK;
-+ t_FmIpcMsg msg;
-+ t_FmIpcReply reply;
-+ uint32_t replyLength;
-+ uint8_t rxHardwarePortId, txHardwarePortId;
-+ struct fman_fpm_regs *fpm_rg = p_Fm->p_FmFpmRegs;
-+
-+ if (p_Fm->guestId != NCSW_MASTER_ID)
-+ {
-+ memset(&msg, 0, sizeof(msg));
-+ memset(&reply, 0, sizeof(reply));
-+ msg.msgId = FM_10G_TX_ECC_WA;
-+ msg.msgBody[0] = macId;
-+ replyLength = sizeof(uint32_t);
-+ if ((err = XX_IpcSendMessage(p_Fm->h_IpcSessions[0],
-+ (uint8_t*)&msg,
-+ sizeof(msg.msgId)+sizeof(macId),
-+ (uint8_t*)&reply,
-+ &replyLength,
-+ NULL,
-+ NULL)) != E_OK)
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+ if (replyLength != sizeof(uint32_t))
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("IPC reply length mismatch"));
-+ return (t_Error)(reply.error);
-+ }
-+
-+ SANITY_CHECK_RETURN_ERROR((macId == 0), E_NOT_SUPPORTED);
-+ SANITY_CHECK_RETURN_ERROR(IsFmanCtrlCodeLoaded(p_Fm), E_INVALID_STATE);
-+
-+ rxHardwarePortId = SwPortIdToHwPortId(e_FM_PORT_TYPE_RX_10G,
-+ macId,
-+ p_Fm->p_FmStateStruct->revInfo.majorRev,
-+ p_Fm->p_FmStateStruct->revInfo.minorRev);
-+ txHardwarePortId = SwPortIdToHwPortId(e_FM_PORT_TYPE_TX_10G,
-+ macId,
-+ p_Fm->p_FmStateStruct->revInfo.majorRev,
-+ p_Fm->p_FmStateStruct->revInfo.minorRev);
-+ if ((p_Fm->p_FmStateStruct->portsTypes[rxHardwarePortId] != e_FM_PORT_TYPE_DUMMY) ||
-+ (p_Fm->p_FmStateStruct->portsTypes[txHardwarePortId] != e_FM_PORT_TYPE_DUMMY))
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE,
-+ ("MAC should be initialized prior to Rx and Tx ports!"));
-+
-+ return fman_set_erratum_10gmac_a004_wa(fpm_rg);
-+}
-+#endif /* FM_TX_ECC_FRMS_ERRATA_10GMAC_A004 */
-+
-+uint16_t FmGetTnumAgingPeriod(t_Handle h_Fm)
-+{
-+ t_Fm *p_Fm = (t_Fm *)h_Fm;
-+
-+ SANITY_CHECK_RETURN_VALUE(p_Fm, E_INVALID_HANDLE, 0);
-+ SANITY_CHECK_RETURN_VALUE(!p_Fm->p_FmDriverParam, E_INVALID_STATE, 0);
-+
-+ return p_Fm->tnumAgingPeriod;
-+}
-+
-+t_Error FmSetPortPreFetchConfiguration(t_Handle h_Fm,
-+ uint8_t portNum,
-+ bool preFetchConfigured)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Fm, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Fm->p_FmDriverParam, E_INVALID_STATE);
-+
-+ p_Fm->portsPreFetchConfigured[portNum] = TRUE;
-+ p_Fm->portsPreFetchValue[portNum] = preFetchConfigured;
-+
-+ return E_OK;
-+}
-+
-+t_Error FmGetPortPreFetchConfiguration(t_Handle h_Fm,
-+ uint8_t portNum,
-+ bool *p_PortConfigured,
-+ bool *p_PreFetchConfigured)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Fm, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Fm->p_FmDriverParam, E_INVALID_STATE);
-+
-+ /* If the prefetch wasn't configured yet (not enable or disabled)
-+ we return the value TRUE as it was already configured */
-+ if (!p_Fm->portsPreFetchConfigured[portNum])
-+ {
-+ *p_PortConfigured = FALSE;
-+ *p_PreFetchConfigured = FALSE;
-+ }
-+ else
-+ {
-+ *p_PortConfigured = TRUE;
-+ *p_PreFetchConfigured = (p_Fm->portsPreFetchConfigured[portNum]);
-+ }
-+
-+ return E_OK;
-+}
-+
-+t_Error FmSetCongestionGroupPFCpriority(t_Handle h_Fm,
-+ uint32_t congestionGroupId,
-+ uint8_t priorityBitMap)
-+{
-+ t_Fm *p_Fm = (t_Fm *)h_Fm;
-+ uint32_t regNum;
-+
-+ ASSERT_COND(h_Fm);
-+
-+ if (congestionGroupId > FM_PORT_NUM_OF_CONGESTION_GRPS)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE,
-+ ("Congestion group ID bigger than %d",
-+ FM_PORT_NUM_OF_CONGESTION_GRPS));
-+
-+ if (p_Fm->guestId == NCSW_MASTER_ID)
-+ {
-+ ASSERT_COND(p_Fm->baseAddr);
-+ regNum = (FM_PORT_NUM_OF_CONGESTION_GRPS - 1 - congestionGroupId) / 4;
-+ fman_set_congestion_group_pfc_priority((uint32_t *)((p_Fm->baseAddr+FM_MM_CGP)),
-+ congestionGroupId,
-+ priorityBitMap,
-+ regNum);
-+ }
-+ else if (p_Fm->h_IpcSessions[0])
-+ {
-+ t_Error err;
-+ t_FmIpcMsg msg;
-+ t_FmIpcSetCongestionGroupPfcPriority fmIpcSetCongestionGroupPfcPriority;
-+
-+ memset(&msg, 0, sizeof(msg));
-+ memset(&fmIpcSetCongestionGroupPfcPriority, 0, sizeof(t_FmIpcSetCongestionGroupPfcPriority));
-+ fmIpcSetCongestionGroupPfcPriority.congestionGroupId = congestionGroupId;
-+ fmIpcSetCongestionGroupPfcPriority.priorityBitMap = priorityBitMap;
-+
-+ msg.msgId = FM_SET_CONG_GRP_PFC_PRIO;
-+ memcpy(msg.msgBody, &fmIpcSetCongestionGroupPfcPriority, sizeof(t_FmIpcSetCongestionGroupPfcPriority));
-+
-+ err = XX_IpcSendMessage(p_Fm->h_IpcSessions[0],
-+ (uint8_t*)&msg,
-+ sizeof(msg.msgId),
-+ NULL,
-+ NULL,
-+ NULL,
-+ NULL);
-+ if (err != E_OK)
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+ }
-+ else
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("guest without IPC!"));
-+
-+ return E_OK;
-+}
-+
-+uintptr_t FmGetPcdPrsBaseAddr(t_Handle h_Fm)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+
-+ SANITY_CHECK_RETURN_VALUE(p_Fm, E_INVALID_HANDLE, 0);
-+
-+ if (!p_Fm->baseAddr)
-+ {
-+ REPORT_ERROR(MAJOR, E_INVALID_STATE,
-+ ("No base-addr; probably Guest with IPC!"));
-+ return 0;
-+ }
-+
-+ return (p_Fm->baseAddr + FM_MM_PRS);
-+}
-+
-+uintptr_t FmGetPcdKgBaseAddr(t_Handle h_Fm)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+
-+ SANITY_CHECK_RETURN_VALUE(p_Fm, E_INVALID_HANDLE, 0);
-+
-+ if (!p_Fm->baseAddr)
-+ {
-+ REPORT_ERROR(MAJOR, E_INVALID_STATE,
-+ ("No base-addr; probably Guest with IPC!"));
-+ return 0;
-+ }
-+
-+ return (p_Fm->baseAddr + FM_MM_KG);
-+}
-+
-+uintptr_t FmGetPcdPlcrBaseAddr(t_Handle h_Fm)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+
-+ SANITY_CHECK_RETURN_VALUE(p_Fm, E_INVALID_HANDLE, 0);
-+
-+ if (!p_Fm->baseAddr)
-+ {
-+ REPORT_ERROR(MAJOR, E_INVALID_STATE,
-+ ("No base-addr; probably Guest with IPC!"));
-+ return 0;
-+ }
-+
-+ return (p_Fm->baseAddr + FM_MM_PLCR);
-+}
-+
-+#if (DPAA_VERSION >= 11)
-+uintptr_t FmGetVSPBaseAddr(t_Handle h_Fm)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+
-+ SANITY_CHECK_RETURN_VALUE(p_Fm, E_INVALID_HANDLE, 0);
-+
-+ return p_Fm->vspBaseAddr;
-+}
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+t_Handle FmGetMuramHandle(t_Handle h_Fm)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+
-+ SANITY_CHECK_RETURN_VALUE(p_Fm, E_INVALID_HANDLE, NULL);
-+
-+ return (p_Fm->h_FmMuram);
-+}
-+
-+void FmGetPhysicalMuramBase(t_Handle h_Fm, t_FmPhysAddr *p_FmPhysAddr)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+
-+ if (p_Fm->fmMuramPhysBaseAddr)
-+ {
-+ /* General FM driver initialization */
-+ p_FmPhysAddr->low = (uint32_t)p_Fm->fmMuramPhysBaseAddr;
-+ p_FmPhysAddr->high = (uint8_t)((p_Fm->fmMuramPhysBaseAddr & 0x000000ff00000000LL) >> 32);
-+ return;
-+ }
-+
-+ ASSERT_COND(p_Fm->guestId != NCSW_MASTER_ID);
-+
-+ if (p_Fm->h_IpcSessions[0])
-+ {
-+ t_Error err;
-+ t_FmIpcMsg msg;
-+ t_FmIpcReply reply;
-+ uint32_t replyLength;
-+ t_FmIpcPhysAddr ipcPhysAddr;
-+
-+ memset(&msg, 0, sizeof(msg));
-+ memset(&reply, 0, sizeof(reply));
-+ msg.msgId = FM_GET_PHYS_MURAM_BASE;
-+ replyLength = sizeof(uint32_t) + sizeof(t_FmPhysAddr);
-+ err = XX_IpcSendMessage(p_Fm->h_IpcSessions[0],
-+ (uint8_t*)&msg,
-+ sizeof(msg.msgId),
-+ (uint8_t*)&reply,
-+ &replyLength,
-+ NULL,
-+ NULL);
-+ if (err != E_OK)
-+ {
-+ REPORT_ERROR(MINOR, err, NO_MSG);
-+ return;
-+ }
-+ if (replyLength != (sizeof(uint32_t) + sizeof(t_FmPhysAddr)))
-+ {
-+ REPORT_ERROR(MINOR, E_INVALID_VALUE,("IPC reply length mismatch"));
-+ return;
-+ }
-+ memcpy((uint8_t*)&ipcPhysAddr, reply.replyBody, sizeof(t_FmIpcPhysAddr));
-+ p_FmPhysAddr->high = ipcPhysAddr.high;
-+ p_FmPhysAddr->low = ipcPhysAddr.low;
-+ }
-+ else
-+ REPORT_ERROR(MINOR, E_NOT_SUPPORTED,
-+ ("running in guest-mode without neither IPC nor mapped register!"));
-+}
-+
-+#if (DPAA_VERSION >= 11)
-+t_Error FmVSPAllocForPort (t_Handle h_Fm,
-+ e_FmPortType portType,
-+ uint8_t portId,
-+ uint8_t numOfVSPs)
-+{
-+ t_Fm *p_Fm = (t_Fm *)h_Fm;
-+ t_Error err = E_OK;
-+ uint32_t profilesFound, intFlags;
-+ uint8_t first, i;
-+ uint8_t log2Num;
-+ uint8_t swPortIndex=0, hardwarePortId;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Fm, E_INVALID_HANDLE);
-+
-+ if (!numOfVSPs)
-+ return E_OK;
-+
-+ if (numOfVSPs > FM_VSP_MAX_NUM_OF_ENTRIES)
-+ RETURN_ERROR(MINOR, E_INVALID_VALUE, ("numProfiles can not be bigger than %d.",FM_VSP_MAX_NUM_OF_ENTRIES));
-+
-+ if (!POWER_OF_2(numOfVSPs))
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("numProfiles must be a power of 2."));
-+
-+ LOG2((uint64_t)numOfVSPs, log2Num);
-+
-+ if ((log2Num == 0) || (p_Fm->partVSPBase == 0))
-+ first = 0;
-+ else
-+ first = 1<<log2Num;
-+
-+ if (first > (p_Fm->partVSPBase + p_Fm->partNumOfVSPs))
-+ RETURN_ERROR(MINOR, E_INVALID_VALUE, ("can not allocate storage profile port window"));
-+
-+ if (first < p_Fm->partVSPBase)
-+ while (first < p_Fm->partVSPBase)
-+ first = first + numOfVSPs;
-+
-+ if ((first + numOfVSPs) > (p_Fm->partVSPBase + p_Fm->partNumOfVSPs))
-+ RETURN_ERROR(MINOR, E_INVALID_VALUE, ("can not allocate storage profile port window"));
-+
-+ intFlags = XX_LockIntrSpinlock(p_Fm->h_Spinlock);
-+ profilesFound = 0;
-+ for (i=first; i < p_Fm->partVSPBase + p_Fm->partNumOfVSPs; )
-+ {
-+ if (!p_Fm->p_FmSp->profiles[i].profilesMng.allocated)
-+ {
-+ profilesFound++;
-+ i++;
-+ if (profilesFound == numOfVSPs)
-+ break;
-+ }
-+ else
-+ {
-+ profilesFound = 0;
-+ /* advance i to the next aligned address */
-+ first = i = (uint8_t)(first + numOfVSPs);
-+ }
-+ }
-+ if (profilesFound == numOfVSPs)
-+ for (i = first; i<first + numOfVSPs; i++)
-+ p_Fm->p_FmSp->profiles[i].profilesMng.allocated = TRUE;
-+ else
-+ {
-+ XX_UnlockIntrSpinlock(p_Fm->h_Spinlock, intFlags);
-+ RETURN_ERROR(MINOR, E_FULL, ("No profiles."));
-+ }
-+
-+ hardwarePortId = SwPortIdToHwPortId(portType,
-+ portId,
-+ p_Fm->p_FmStateStruct->revInfo.majorRev,
-+ p_Fm->p_FmStateStruct->revInfo.minorRev);
-+ HW_PORT_ID_TO_SW_PORT_INDX(swPortIndex, hardwarePortId);
-+
-+ p_Fm->p_FmSp->portsMapping[swPortIndex].numOfProfiles = numOfVSPs;
-+ p_Fm->p_FmSp->portsMapping[swPortIndex].profilesBase = first;
-+
-+ if ((err = SetVSPWindow(h_Fm,hardwarePortId, first,log2Num)) != E_OK)
-+ for (i = first; i < first + numOfVSPs; i++)
-+ p_Fm->p_FmSp->profiles[i].profilesMng.allocated = FALSE;
-+
-+ XX_UnlockIntrSpinlock(p_Fm->h_Spinlock, intFlags);
-+
-+ return err;
-+}
-+
-+t_Error FmVSPFreeForPort(t_Handle h_Fm,
-+ e_FmPortType portType,
-+ uint8_t portId)
-+{
-+ t_Fm *p_Fm = (t_Fm *)h_Fm;
-+ uint8_t swPortIndex=0, hardwarePortId, first, numOfVSPs, i;
-+ uint32_t intFlags;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Fm, E_INVALID_HANDLE);
-+
-+ hardwarePortId = SwPortIdToHwPortId(portType,
-+ portId,
-+ p_Fm->p_FmStateStruct->revInfo.majorRev,
-+ p_Fm->p_FmStateStruct->revInfo.minorRev);
-+ HW_PORT_ID_TO_SW_PORT_INDX(swPortIndex, hardwarePortId);
-+
-+ numOfVSPs = (uint8_t)p_Fm->p_FmSp->portsMapping[swPortIndex].numOfProfiles;
-+ first = (uint8_t)p_Fm->p_FmSp->portsMapping[swPortIndex].profilesBase;
-+
-+ intFlags = XX_LockIntrSpinlock(p_Fm->h_Spinlock);
-+ for (i = first; i < first + numOfVSPs; i++)
-+ p_Fm->p_FmSp->profiles[i].profilesMng.allocated = FALSE;
-+ XX_UnlockIntrSpinlock(p_Fm->h_Spinlock, intFlags);
-+
-+ p_Fm->p_FmSp->portsMapping[swPortIndex].numOfProfiles = 0;
-+ p_Fm->p_FmSp->portsMapping[swPortIndex].profilesBase = 0;
-+
-+ return E_OK;
-+}
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+t_Error FmAllocFmanCtrlEventReg(t_Handle h_Fm, uint8_t *p_EventId)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+ uint8_t i;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Fm, E_INVALID_HANDLE);
-+
-+ if ((p_Fm->guestId != NCSW_MASTER_ID) &&
-+ p_Fm->h_IpcSessions[0])
-+ {
-+ t_Error err;
-+ t_FmIpcMsg msg;
-+ t_FmIpcReply reply;
-+ uint32_t replyLength;
-+
-+ memset(&msg, 0, sizeof(msg));
-+ memset(&reply, 0, sizeof(reply));
-+ msg.msgId = FM_ALLOC_FMAN_CTRL_EVENT_REG;
-+ replyLength = sizeof(uint32_t) + sizeof(uint8_t);
-+ if ((err = XX_IpcSendMessage(p_Fm->h_IpcSessions[0],
-+ (uint8_t*)&msg,
-+ sizeof(msg.msgId),
-+ (uint8_t*)&reply,
-+ &replyLength,
-+ NULL,
-+ NULL)) != E_OK)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+
-+ if (replyLength != (sizeof(uint32_t) + sizeof(uint8_t)))
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("IPC reply length mismatch"));
-+
-+ *p_EventId = *(uint8_t*)(reply.replyBody);
-+
-+ return (t_Error)(reply.error);
-+ }
-+ else if (p_Fm->guestId != NCSW_MASTER_ID)
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED,
-+ ("running in guest-mode without IPC!"));
-+
-+ for (i=0;i<FM_NUM_OF_FMAN_CTRL_EVENT_REGS;i++)
-+ if (!p_Fm->usedEventRegs[i])
-+ {
-+ p_Fm->usedEventRegs[i] = TRUE;
-+ *p_EventId = i;
-+ break;
-+ }
-+
-+ if (i==FM_NUM_OF_FMAN_CTRL_EVENT_REGS)
-+ RETURN_ERROR(MAJOR, E_BUSY, ("No resource - FMan controller event register."));
-+
-+ return E_OK;
-+}
-+
-+void FmFreeFmanCtrlEventReg(t_Handle h_Fm, uint8_t eventId)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+
-+ SANITY_CHECK_RETURN(p_Fm, E_INVALID_HANDLE);
-+
-+ if ((p_Fm->guestId != NCSW_MASTER_ID) &&
-+ p_Fm->h_IpcSessions[0])
-+ {
-+ t_Error err;
-+ t_FmIpcMsg msg;
-+
-+ memset(&msg, 0, sizeof(msg));
-+ msg.msgId = FM_FREE_FMAN_CTRL_EVENT_REG;
-+ msg.msgBody[0] = eventId;
-+ err = XX_IpcSendMessage(p_Fm->h_IpcSessions[0],
-+ (uint8_t*)&msg,
-+ sizeof(msg.msgId)+sizeof(eventId),
-+ NULL,
-+ NULL,
-+ NULL,
-+ NULL);
-+ if (err != E_OK)
-+ REPORT_ERROR(MINOR, err, NO_MSG);
-+ return;
-+ }
-+ else if (p_Fm->guestId != NCSW_MASTER_ID)
-+ {
-+ REPORT_ERROR(MINOR, E_NOT_SUPPORTED,
-+ ("running in guest-mode without IPC!"));
-+ return;
-+ }
-+
-+ ((t_Fm*)h_Fm)->usedEventRegs[eventId] = FALSE;
-+}
-+
-+void FmSetFmanCtrlIntr(t_Handle h_Fm, uint8_t eventRegId, uint32_t enableEvents)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+ struct fman_fpm_regs *fpm_rg = p_Fm->p_FmFpmRegs;
-+
-+ if ((p_Fm->guestId != NCSW_MASTER_ID) &&
-+ !p_Fm->p_FmFpmRegs &&
-+ p_Fm->h_IpcSessions[0])
-+ {
-+ t_FmIpcFmanEvents fmanCtrl;
-+ t_Error err;
-+ t_FmIpcMsg msg;
-+
-+ fmanCtrl.eventRegId = eventRegId;
-+ fmanCtrl.enableEvents = enableEvents;
-+ memset(&msg, 0, sizeof(msg));
-+ msg.msgId = FM_SET_FMAN_CTRL_EVENTS_ENABLE;
-+ memcpy(msg.msgBody, &fmanCtrl, sizeof(fmanCtrl));
-+ err = XX_IpcSendMessage(p_Fm->h_IpcSessions[0],
-+ (uint8_t*)&msg,
-+ sizeof(msg.msgId)+sizeof(fmanCtrl),
-+ NULL,
-+ NULL,
-+ NULL,
-+ NULL);
-+ if (err != E_OK)
-+ REPORT_ERROR(MINOR, err, NO_MSG);
-+ return;
-+ }
-+ else if (!p_Fm->p_FmFpmRegs)
-+ {
-+ REPORT_ERROR(MINOR, E_NOT_SUPPORTED,
-+ ("Either IPC or 'baseAddress' is required!"));
-+ return;
-+ }
-+
-+ ASSERT_COND(eventRegId < FM_NUM_OF_FMAN_CTRL_EVENT_REGS);
-+ fman_set_ctrl_intr(fpm_rg, eventRegId, enableEvents);
-+}
-+
-+uint32_t FmGetFmanCtrlIntr(t_Handle h_Fm, uint8_t eventRegId)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+ struct fman_fpm_regs *fpm_rg = p_Fm->p_FmFpmRegs;
-+
-+ if ((p_Fm->guestId != NCSW_MASTER_ID) &&
-+ !p_Fm->p_FmFpmRegs &&
-+ p_Fm->h_IpcSessions[0])
-+ {
-+ t_Error err;
-+ t_FmIpcMsg msg;
-+ t_FmIpcReply reply;
-+ uint32_t replyLength, ctrlIntr;
-+
-+ memset(&msg, 0, sizeof(msg));
-+ memset(&reply, 0, sizeof(reply));
-+ msg.msgId = FM_GET_FMAN_CTRL_EVENTS_ENABLE;
-+ msg.msgBody[0] = eventRegId;
-+ replyLength = sizeof(uint32_t) + sizeof(uint32_t);
-+ err = XX_IpcSendMessage(p_Fm->h_IpcSessions[0],
-+ (uint8_t*)&msg,
-+ sizeof(msg.msgId)+sizeof(eventRegId),
-+ (uint8_t*)&reply,
-+ &replyLength,
-+ NULL,
-+ NULL);
-+ if (err != E_OK)
-+ {
-+ REPORT_ERROR(MINOR, err, NO_MSG);
-+ return 0;
-+ }
-+ if (replyLength != (sizeof(uint32_t) + sizeof(uint32_t)))
-+ {
-+ REPORT_ERROR(MINOR, E_INVALID_VALUE, ("IPC reply length mismatch"));
-+ return 0;
-+ }
-+ memcpy((uint8_t*)&ctrlIntr, reply.replyBody, sizeof(uint32_t));
-+ return ctrlIntr;
-+ }
-+ else if (!p_Fm->p_FmFpmRegs)
-+ {
-+ REPORT_ERROR(MINOR, E_NOT_SUPPORTED,
-+ ("Either IPC or 'baseAddress' is required!"));
-+ return 0;
-+ }
-+
-+ return fman_get_ctrl_intr(fpm_rg, eventRegId);
-+}
-+
-+void FmRegisterIntr(t_Handle h_Fm,
-+ e_FmEventModules module,
-+ uint8_t modId,
-+ e_FmIntrType intrType,
-+ void (*f_Isr) (t_Handle h_Arg),
-+ t_Handle h_Arg)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+ int event = 0;
-+
-+ ASSERT_COND(h_Fm);
-+
-+ GET_FM_MODULE_EVENT(module, modId, intrType, event);
-+ ASSERT_COND(event < e_FM_EV_DUMMY_LAST);
-+
-+ /* register in local FM structure */
-+ p_Fm->intrMng[event].f_Isr = f_Isr;
-+ p_Fm->intrMng[event].h_SrcHandle = h_Arg;
-+
-+ if ((p_Fm->guestId != NCSW_MASTER_ID) &&
-+ p_Fm->h_IpcSessions[0])
-+ {
-+ t_FmIpcRegisterIntr fmIpcRegisterIntr;
-+ t_Error err;
-+ t_FmIpcMsg msg;
-+
-+ /* register in Master FM structure */
-+ fmIpcRegisterIntr.event = (uint32_t)event;
-+ fmIpcRegisterIntr.guestId = p_Fm->guestId;
-+ memset(&msg, 0, sizeof(msg));
-+ msg.msgId = FM_REGISTER_INTR;
-+ memcpy(msg.msgBody, &fmIpcRegisterIntr, sizeof(fmIpcRegisterIntr));
-+ err = XX_IpcSendMessage(p_Fm->h_IpcSessions[0],
-+ (uint8_t*)&msg,
-+ sizeof(msg.msgId) + sizeof(fmIpcRegisterIntr),
-+ NULL,
-+ NULL,
-+ NULL,
-+ NULL);
-+ if (err != E_OK)
-+ REPORT_ERROR(MINOR, err, NO_MSG);
-+ }
-+ else if (p_Fm->guestId != NCSW_MASTER_ID)
-+ REPORT_ERROR(MINOR, E_NOT_SUPPORTED,
-+ ("running in guest-mode without IPC!"));
-+}
-+
-+void FmUnregisterIntr(t_Handle h_Fm,
-+ e_FmEventModules module,
-+ uint8_t modId,
-+ e_FmIntrType intrType)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+ int event = 0;
-+
-+ ASSERT_COND(h_Fm);
-+
-+ GET_FM_MODULE_EVENT(module, modId,intrType, event);
-+ ASSERT_COND(event < e_FM_EV_DUMMY_LAST);
-+
-+ p_Fm->intrMng[event].f_Isr = UnimplementedIsr;
-+ p_Fm->intrMng[event].h_SrcHandle = NULL;
-+}
-+
-+void FmRegisterFmanCtrlIntr(t_Handle h_Fm, uint8_t eventRegId, void (*f_Isr) (t_Handle h_Arg, uint32_t event), t_Handle h_Arg)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+
-+ ASSERT_COND(eventRegId<FM_NUM_OF_FMAN_CTRL_EVENT_REGS);
-+
-+ if (p_Fm->guestId != NCSW_MASTER_ID)
-+ {
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("FM in guest-mode"));
-+ return;
-+ }
-+
-+ p_Fm->fmanCtrlIntr[eventRegId].f_Isr = f_Isr;
-+ p_Fm->fmanCtrlIntr[eventRegId].h_SrcHandle = h_Arg;
-+}
-+
-+void FmUnregisterFmanCtrlIntr(t_Handle h_Fm, uint8_t eventRegId)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+
-+ ASSERT_COND(eventRegId<FM_NUM_OF_FMAN_CTRL_EVENT_REGS);
-+
-+ if (p_Fm->guestId != NCSW_MASTER_ID)
-+ {
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("FM in guest-mode"));
-+ return;
-+ }
-+
-+ p_Fm->fmanCtrlIntr[eventRegId].f_Isr = UnimplementedFmanCtrlIsr;
-+ p_Fm->fmanCtrlIntr[eventRegId].h_SrcHandle = NULL;
-+}
-+
-+void FmRegisterPcd(t_Handle h_Fm, t_Handle h_FmPcd)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+
-+ if (p_Fm->h_Pcd)
-+ REPORT_ERROR(MAJOR, E_ALREADY_EXISTS, ("PCD already set"));
-+
-+ p_Fm->h_Pcd = h_FmPcd;
-+}
-+
-+void FmUnregisterPcd(t_Handle h_Fm)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+
-+ if (!p_Fm->h_Pcd)
-+ REPORT_ERROR(MAJOR, E_NOT_FOUND, ("PCD handle!"));
-+
-+ p_Fm->h_Pcd = NULL;
-+}
-+
-+t_Handle FmGetPcdHandle(t_Handle h_Fm)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+
-+ return p_Fm->h_Pcd;
-+}
-+
-+uint8_t FmGetId(t_Handle h_Fm)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+
-+ SANITY_CHECK_RETURN_VALUE(p_Fm, E_INVALID_HANDLE, 0xff);
-+
-+ return p_Fm->p_FmStateStruct->fmId;
-+}
-+
-+t_Error FmReset(t_Handle h_Fm)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Fm, E_INVALID_HANDLE);
-+
-+ WRITE_UINT32(p_Fm->p_FmFpmRegs->fm_rstc, FPM_RSTC_FM_RESET);
-+ CORE_MemoryBarrier();
-+ XX_UDelay(100);
-+
-+ return E_OK;
-+}
-+
-+t_Error FmSetNumOfRiscsPerPort(t_Handle h_Fm,
-+ uint8_t hardwarePortId,
-+ uint8_t numOfFmanCtrls,
-+ t_FmFmanCtrl orFmanCtrl)
-+{
-+
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+ struct fman_fpm_regs *fpm_rg;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Fm, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(((numOfFmanCtrls > 0) && (numOfFmanCtrls < 3)) , E_INVALID_HANDLE);
-+
-+ fpm_rg = p_Fm->p_FmFpmRegs;
-+ if ((p_Fm->guestId != NCSW_MASTER_ID) &&
-+ !p_Fm->p_FmFpmRegs &&
-+ p_Fm->h_IpcSessions[0])
-+ {
-+ t_Error err;
-+ t_FmIpcPortNumOfFmanCtrls params;
-+ t_FmIpcMsg msg;
-+
-+ memset(&msg, 0, sizeof(msg));
-+ params.hardwarePortId = hardwarePortId;
-+ params.numOfFmanCtrls = numOfFmanCtrls;
-+ params.orFmanCtrl = orFmanCtrl;
-+ msg.msgId = FM_SET_NUM_OF_FMAN_CTRL;
-+ memcpy(msg.msgBody, &params, sizeof(params));
-+ err = XX_IpcSendMessage(p_Fm->h_IpcSessions[0],
-+ (uint8_t*)&msg,
-+ sizeof(msg.msgId) +sizeof(params),
-+ NULL,
-+ NULL,
-+ NULL,
-+ NULL);
-+ if (err != E_OK)
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+ return E_OK;
-+ }
-+ else if (!p_Fm->p_FmFpmRegs)
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED,
-+ ("Either IPC or 'baseAddress' is required!"));
-+
-+ fman_set_num_of_riscs_per_port(fpm_rg, hardwarePortId, numOfFmanCtrls, orFmanCtrl);
-+
-+ return E_OK;
-+}
-+
-+t_Error FmGetSetPortParams(t_Handle h_Fm, t_FmInterModulePortInitParams *p_PortParams)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+ t_Error err;
-+ uint32_t intFlags;
-+ uint8_t hardwarePortId = p_PortParams->hardwarePortId, macId;
-+ struct fman_rg fman_rg;
-+
-+ fman_rg.bmi_rg = p_Fm->p_FmBmiRegs;
-+ fman_rg.qmi_rg = p_Fm->p_FmQmiRegs;
-+ fman_rg.fpm_rg = p_Fm->p_FmFpmRegs;
-+ fman_rg.dma_rg = p_Fm->p_FmDmaRegs;
-+
-+ if (p_Fm->guestId != NCSW_MASTER_ID)
-+ {
-+ t_FmIpcPortInInitParams portInParams;
-+ t_FmIpcPortOutInitParams portOutParams;
-+ t_FmIpcMsg msg;
-+ t_FmIpcReply reply;
-+ uint32_t replyLength;
-+
-+ portInParams.hardwarePortId = p_PortParams->hardwarePortId;
-+ portInParams.enumPortType = (uint32_t)p_PortParams->portType;
-+ portInParams.boolIndependentMode= (uint8_t)p_PortParams->independentMode;
-+ portInParams.liodnOffset = p_PortParams->liodnOffset;
-+ portInParams.numOfTasks = p_PortParams->numOfTasks;
-+ portInParams.numOfExtraTasks = p_PortParams->numOfExtraTasks;
-+ portInParams.numOfOpenDmas = p_PortParams->numOfOpenDmas;
-+ portInParams.numOfExtraOpenDmas = p_PortParams->numOfExtraOpenDmas;
-+ portInParams.sizeOfFifo = p_PortParams->sizeOfFifo;
-+ portInParams.extraSizeOfFifo = p_PortParams->extraSizeOfFifo;
-+ portInParams.deqPipelineDepth = p_PortParams->deqPipelineDepth;
-+ portInParams.maxFrameLength = p_PortParams->maxFrameLength;
-+ portInParams.liodnBase = p_PortParams->liodnBase;
-+
-+ memset(&msg, 0, sizeof(msg));
-+ memset(&reply, 0, sizeof(reply));
-+ msg.msgId = FM_GET_SET_PORT_PARAMS;
-+ memcpy(msg.msgBody, &portInParams, sizeof(portInParams));
-+ replyLength = (sizeof(uint32_t) + sizeof(t_FmIpcPortOutInitParams));
-+ if ((err = XX_IpcSendMessage(p_Fm->h_IpcSessions[0],
-+ (uint8_t*)&msg,
-+ sizeof(msg.msgId) +sizeof(portInParams),
-+ (uint8_t*)&reply,
-+ &replyLength,
-+ NULL,
-+ NULL)) != E_OK)
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+ if (replyLength != (sizeof(uint32_t) + sizeof(t_FmIpcPortOutInitParams)))
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("IPC reply length mismatch"));
-+ memcpy((uint8_t*)&portOutParams, reply.replyBody, sizeof(t_FmIpcPortOutInitParams));
-+
-+ p_PortParams->fmMuramPhysBaseAddr.high = portOutParams.ipcPhysAddr.high;
-+ p_PortParams->fmMuramPhysBaseAddr.low = portOutParams.ipcPhysAddr.low;
-+ p_PortParams->numOfTasks = portOutParams.numOfTasks;
-+ p_PortParams->numOfExtraTasks = portOutParams.numOfExtraTasks;
-+ p_PortParams->numOfOpenDmas = portOutParams.numOfOpenDmas;
-+ p_PortParams->numOfExtraOpenDmas = portOutParams.numOfExtraOpenDmas;
-+ p_PortParams->sizeOfFifo = portOutParams.sizeOfFifo;
-+ p_PortParams->extraSizeOfFifo = portOutParams.extraSizeOfFifo;
-+
-+ return (t_Error)(reply.error);
-+ }
-+
-+ ASSERT_COND(IN_RANGE(1, hardwarePortId, 63));
-+
-+ intFlags = XX_LockIntrSpinlock(p_Fm->h_Spinlock);
-+ if (p_PortParams->independentMode)
-+ {
-+ /* set port parameters */
-+ p_Fm->independentMode = p_PortParams->independentMode;
-+ /* disable dispatch limit */
-+ fman_qmi_disable_dispatch_limit(fman_rg.fpm_rg);
-+ }
-+
-+ if (p_PortParams->portType == e_FM_PORT_TYPE_OH_HOST_COMMAND)
-+ {
-+ if (p_Fm->hcPortInitialized)
-+ {
-+ XX_UnlockIntrSpinlock(p_Fm->h_Spinlock, intFlags);
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("Only one host command port is allowed."));
-+ }
-+ else
-+ p_Fm->hcPortInitialized = TRUE;
-+ }
-+ p_Fm->p_FmStateStruct->portsTypes[hardwarePortId] = p_PortParams->portType;
-+
-+ err = FmSetNumOfTasks(p_Fm, hardwarePortId, &p_PortParams->numOfTasks, &p_PortParams->numOfExtraTasks, TRUE);
-+ if (err)
-+ {
-+ XX_UnlockIntrSpinlock(p_Fm->h_Spinlock, intFlags);
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+
-+#ifdef FM_QMI_NO_DEQ_OPTIONS_SUPPORT
-+ if (p_Fm->p_FmStateStruct->revInfo.majorRev != 4)
-+#endif /* FM_QMI_NO_DEQ_OPTIONS_SUPPORT */
-+ if ((p_PortParams->portType != e_FM_PORT_TYPE_RX) &&
-+ (p_PortParams->portType != e_FM_PORT_TYPE_RX_10G))
-+ /* for transmit & O/H ports */
-+ {
-+ uint8_t enqTh;
-+ uint8_t deqTh;
-+
-+ /* update qmi ENQ/DEQ threshold */
-+ p_Fm->p_FmStateStruct->accumulatedNumOfDeqTnums += p_PortParams->deqPipelineDepth;
-+ enqTh = fman_get_qmi_enq_th(fman_rg.qmi_rg);
-+ /* if enqTh is too big, we reduce it to the max value that is still OK */
-+ if (enqTh >= (QMI_MAX_NUM_OF_TNUMS - p_Fm->p_FmStateStruct->accumulatedNumOfDeqTnums))
-+ {
-+ enqTh = (uint8_t)(QMI_MAX_NUM_OF_TNUMS - p_Fm->p_FmStateStruct->accumulatedNumOfDeqTnums - 1);
-+ fman_set_qmi_enq_th(fman_rg.qmi_rg, enqTh);
-+ }
-+
-+ deqTh = fman_get_qmi_deq_th(fman_rg.qmi_rg);
-+ /* if deqTh is too small, we enlarge it to the min value that is still OK.
-+ deqTh may not be larger than 63 (QMI_MAX_NUM_OF_TNUMS-1). */
-+ if ((deqTh <= p_Fm->p_FmStateStruct->accumulatedNumOfDeqTnums) && (deqTh < QMI_MAX_NUM_OF_TNUMS-1))
-+ {
-+ deqTh = (uint8_t)(p_Fm->p_FmStateStruct->accumulatedNumOfDeqTnums + 1);
-+ fman_set_qmi_deq_th(fman_rg.qmi_rg, deqTh);
-+ }
-+ }
-+
-+#ifdef FM_LOW_END_RESTRICTION
-+ if ((hardwarePortId==0x1) || (hardwarePortId==0x29))
-+ {
-+ if (p_Fm->p_FmStateStruct->lowEndRestriction)
-+ {
-+ XX_UnlockIntrSpinlock(p_Fm->h_Spinlock, intFlags);
-+ RETURN_ERROR(MAJOR, E_NOT_AVAILABLE, ("OP #0 cannot work with Tx Port #1."));
-+ }
-+ else
-+ p_Fm->p_FmStateStruct->lowEndRestriction = TRUE;
-+ }
-+#endif /* FM_LOW_END_RESTRICTION */
-+
-+ err = FmSetSizeOfFifo(p_Fm,
-+ hardwarePortId,
-+ &p_PortParams->sizeOfFifo,
-+ &p_PortParams->extraSizeOfFifo,
-+ TRUE);
-+ if (err)
-+ {
-+ XX_UnlockIntrSpinlock(p_Fm->h_Spinlock, intFlags);
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+
-+ err = FmSetNumOfOpenDmas(p_Fm,
-+ hardwarePortId,
-+ &p_PortParams->numOfOpenDmas,
-+ &p_PortParams->numOfExtraOpenDmas,
-+ TRUE);
-+ if (err)
-+ {
-+ XX_UnlockIntrSpinlock(p_Fm->h_Spinlock, intFlags);
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+
-+ fman_set_liodn_per_port(&fman_rg,
-+ hardwarePortId,
-+ p_PortParams->liodnBase,
-+ p_PortParams->liodnOffset);
-+
-+ if (p_Fm->p_FmStateStruct->revInfo.majorRev < 6)
-+ fman_set_order_restoration_per_port(fman_rg.fpm_rg,
-+ hardwarePortId,
-+ p_PortParams->independentMode,
-+ !!((p_PortParams->portType==e_FM_PORT_TYPE_RX) || (p_PortParams->portType==e_FM_PORT_TYPE_RX_10G)));
-+
-+ HW_PORT_ID_TO_SW_PORT_ID(macId, hardwarePortId);
-+
-+#if defined(FM_MAX_NUM_OF_10G_MACS) && (FM_MAX_NUM_OF_10G_MACS)
-+ if ((p_PortParams->portType == e_FM_PORT_TYPE_TX_10G) ||
-+ (p_PortParams->portType == e_FM_PORT_TYPE_RX_10G))
-+ {
-+ ASSERT_COND(macId < FM_MAX_NUM_OF_10G_MACS);
-+ if (p_PortParams->maxFrameLength >= p_Fm->p_FmStateStruct->macMaxFrameLengths10G[macId])
-+ p_Fm->p_FmStateStruct->portMaxFrameLengths10G[macId] = p_PortParams->maxFrameLength;
-+ else
-+ RETURN_ERROR(MINOR, E_INVALID_VALUE, ("Port maxFrameLength is smaller than MAC current MTU"));
-+ }
-+ else
-+#endif /* defined(FM_MAX_NUM_OF_10G_MACS) && ... */
-+ if ((p_PortParams->portType == e_FM_PORT_TYPE_TX) ||
-+ (p_PortParams->portType == e_FM_PORT_TYPE_RX))
-+ {
-+ ASSERT_COND(macId < FM_MAX_NUM_OF_1G_MACS);
-+ if (p_PortParams->maxFrameLength >= p_Fm->p_FmStateStruct->macMaxFrameLengths1G[macId])
-+ p_Fm->p_FmStateStruct->portMaxFrameLengths1G[macId] = p_PortParams->maxFrameLength;
-+ else
-+ RETURN_ERROR(MINOR, E_INVALID_VALUE, ("Port maxFrameLength is smaller than MAC current MTU"));
-+ }
-+
-+ FmGetPhysicalMuramBase(p_Fm, &p_PortParams->fmMuramPhysBaseAddr);
-+ XX_UnlockIntrSpinlock(p_Fm->h_Spinlock, intFlags);
-+
-+ return E_OK;
-+}
-+
-+void FmFreePortParams(t_Handle h_Fm,t_FmInterModulePortFreeParams *p_PortParams)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+ uint32_t intFlags;
-+ uint8_t hardwarePortId = p_PortParams->hardwarePortId;
-+ uint8_t numOfTasks, numOfDmas, macId;
-+ uint16_t sizeOfFifo;
-+ t_Error err;
-+ t_FmIpcPortFreeParams portParams;
-+ t_FmIpcMsg msg;
-+ struct fman_qmi_regs *qmi_rg = p_Fm->p_FmQmiRegs;
-+ struct fman_bmi_regs *bmi_rg = p_Fm->p_FmBmiRegs;
-+
-+ if (p_Fm->guestId != NCSW_MASTER_ID)
-+ {
-+ portParams.hardwarePortId = p_PortParams->hardwarePortId;
-+ portParams.enumPortType = (uint32_t)p_PortParams->portType;
-+ portParams.deqPipelineDepth = p_PortParams->deqPipelineDepth;
-+ memset(&msg, 0, sizeof(msg));
-+ msg.msgId = FM_FREE_PORT;
-+ memcpy(msg.msgBody, &portParams, sizeof(portParams));
-+ err = XX_IpcSendMessage(p_Fm->h_IpcSessions[0],
-+ (uint8_t*)&msg,
-+ sizeof(msg.msgId)+sizeof(portParams),
-+ NULL,
-+ NULL,
-+ NULL,
-+ NULL);
-+ if (err != E_OK)
-+ REPORT_ERROR(MINOR, err, NO_MSG);
-+ return;
-+ }
-+
-+ ASSERT_COND(IN_RANGE(1, hardwarePortId, 63));
-+
-+ intFlags = XX_LockIntrSpinlock(p_Fm->h_Spinlock);
-+
-+ if (p_PortParams->portType == e_FM_PORT_TYPE_OH_HOST_COMMAND)
-+ {
-+ ASSERT_COND(p_Fm->hcPortInitialized);
-+ p_Fm->hcPortInitialized = FALSE;
-+ }
-+
-+ p_Fm->p_FmStateStruct->portsTypes[hardwarePortId] = e_FM_PORT_TYPE_DUMMY;
-+
-+ /* free numOfTasks */
-+ numOfTasks = fman_get_num_of_tasks(bmi_rg, hardwarePortId);
-+ ASSERT_COND(p_Fm->p_FmStateStruct->accumulatedNumOfTasks >= numOfTasks);
-+ p_Fm->p_FmStateStruct->accumulatedNumOfTasks -= numOfTasks;
-+
-+ /* free numOfOpenDmas */
-+ numOfDmas = fman_get_num_of_dmas(bmi_rg, hardwarePortId);
-+ ASSERT_COND(p_Fm->p_FmStateStruct->accumulatedNumOfOpenDmas >= numOfDmas);
-+ p_Fm->p_FmStateStruct->accumulatedNumOfOpenDmas -= numOfDmas;
-+
-+#ifdef FM_HAS_TOTAL_DMAS
-+ if (p_Fm->p_FmStateStruct->revInfo.majorRev < 6)
-+ {
-+ /* update total num of DMA's with committed number of open DMAS, and max uncommitted pool. */
-+ fman_set_num_of_open_dmas(bmi_rg,
-+ hardwarePortId,
-+ 1,
-+ 0,
-+ (uint8_t)(p_Fm->p_FmStateStruct->accumulatedNumOfOpenDmas + p_Fm->p_FmStateStruct->extraOpenDmasPoolSize));
-+ }
-+#endif /* FM_HAS_TOTAL_DMAS */
-+
-+ /* free sizeOfFifo */
-+ sizeOfFifo = fman_get_size_of_fifo(bmi_rg, hardwarePortId);
-+ ASSERT_COND(p_Fm->p_FmStateStruct->accumulatedFifoSize >= (sizeOfFifo * BMI_FIFO_UNITS));
-+ p_Fm->p_FmStateStruct->accumulatedFifoSize -= (sizeOfFifo * BMI_FIFO_UNITS);
-+
-+#ifdef FM_QMI_NO_DEQ_OPTIONS_SUPPORT
-+ if (p_Fm->p_FmStateStruct->revInfo.majorRev != 4)
-+#endif /* FM_QMI_NO_DEQ_OPTIONS_SUPPORT */
-+ if ((p_PortParams->portType != e_FM_PORT_TYPE_RX) &&
-+ (p_PortParams->portType != e_FM_PORT_TYPE_RX_10G))
-+ /* for transmit & O/H ports */
-+ {
-+ uint8_t enqTh;
-+ uint8_t deqTh;
-+
-+ /* update qmi ENQ/DEQ threshold */
-+ p_Fm->p_FmStateStruct->accumulatedNumOfDeqTnums -= p_PortParams->deqPipelineDepth;
-+
-+ /* p_Fm->p_FmStateStruct->accumulatedNumOfDeqTnums is now smaller,
-+ so we can enlarge enqTh */
-+ enqTh = (uint8_t)(QMI_MAX_NUM_OF_TNUMS - p_Fm->p_FmStateStruct->accumulatedNumOfDeqTnums - 1);
-+
-+ /* p_Fm->p_FmStateStruct->accumulatedNumOfDeqTnums is now smaller,
-+ so we can reduce deqTh */
-+ deqTh = (uint8_t)(p_Fm->p_FmStateStruct->accumulatedNumOfDeqTnums + 1);
-+
-+ fman_set_qmi_enq_th(qmi_rg, enqTh);
-+ fman_set_qmi_deq_th(qmi_rg, deqTh);
-+ }
-+
-+ HW_PORT_ID_TO_SW_PORT_ID(macId, hardwarePortId);
-+
-+#if defined(FM_MAX_NUM_OF_10G_MACS) && (FM_MAX_NUM_OF_10G_MACS)
-+ if ((p_PortParams->portType == e_FM_PORT_TYPE_TX_10G) ||
-+ (p_PortParams->portType == e_FM_PORT_TYPE_RX_10G))
-+ {
-+ ASSERT_COND(macId < FM_MAX_NUM_OF_10G_MACS);
-+ p_Fm->p_FmStateStruct->portMaxFrameLengths10G[macId] = 0;
-+ }
-+ else
-+#endif /* defined(FM_MAX_NUM_OF_10G_MACS) && ... */
-+ if ((p_PortParams->portType == e_FM_PORT_TYPE_TX) ||
-+ (p_PortParams->portType == e_FM_PORT_TYPE_RX))
-+ {
-+ ASSERT_COND(macId < FM_MAX_NUM_OF_1G_MACS);
-+ p_Fm->p_FmStateStruct->portMaxFrameLengths1G[macId] = 0;
-+ }
-+
-+#ifdef FM_LOW_END_RESTRICTION
-+ if ((hardwarePortId==0x1) || (hardwarePortId==0x29))
-+ p_Fm->p_FmStateStruct->lowEndRestriction = FALSE;
-+#endif /* FM_LOW_END_RESTRICTION */
-+ XX_UnlockIntrSpinlock(p_Fm->h_Spinlock, intFlags);
-+}
-+
-+t_Error FmIsPortStalled(t_Handle h_Fm, uint8_t hardwarePortId, bool *p_IsStalled)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+ t_Error err;
-+ t_FmIpcMsg msg;
-+ t_FmIpcReply reply;
-+ uint32_t replyLength;
-+ struct fman_fpm_regs *fpm_rg = p_Fm->p_FmFpmRegs;
-+
-+ if ((p_Fm->guestId != NCSW_MASTER_ID) &&
-+ !p_Fm->baseAddr &&
-+ p_Fm->h_IpcSessions[0])
-+ {
-+ memset(&msg, 0, sizeof(msg));
-+ memset(&reply, 0, sizeof(reply));
-+ msg.msgId = FM_IS_PORT_STALLED;
-+ msg.msgBody[0] = hardwarePortId;
-+ replyLength = sizeof(uint32_t) + sizeof(uint8_t);
-+ err = XX_IpcSendMessage(p_Fm->h_IpcSessions[0],
-+ (uint8_t*)&msg,
-+ sizeof(msg.msgId)+sizeof(hardwarePortId),
-+ (uint8_t*)&reply,
-+ &replyLength,
-+ NULL,
-+ NULL);
-+ if (err != E_OK)
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+ if (replyLength != (sizeof(uint32_t) + sizeof(uint8_t)))
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("IPC reply length mismatch"));
-+
-+ *p_IsStalled = (bool)!!(*(uint8_t*)(reply.replyBody));
-+
-+ return (t_Error)(reply.error);
-+ }
-+ else if (!p_Fm->baseAddr)
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED,
-+ ("Either IPC or 'baseAddress' is required!"));
-+
-+ *p_IsStalled = fman_is_port_stalled(fpm_rg, hardwarePortId);
-+
-+ return E_OK;
-+}
-+
-+t_Error FmResumeStalledPort(t_Handle h_Fm, uint8_t hardwarePortId)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+ t_Error err;
-+ bool isStalled;
-+ struct fman_fpm_regs *fpm_rg = p_Fm->p_FmFpmRegs;
-+
-+ if ((p_Fm->guestId != NCSW_MASTER_ID) &&
-+ !p_Fm->baseAddr &&
-+ p_Fm->h_IpcSessions[0])
-+ {
-+ t_FmIpcMsg msg;
-+ t_FmIpcReply reply;
-+ uint32_t replyLength;
-+
-+ memset(&msg, 0, sizeof(msg));
-+ memset(&reply, 0, sizeof(reply));
-+ msg.msgId = FM_RESUME_STALLED_PORT;
-+ msg.msgBody[0] = hardwarePortId;
-+ replyLength = sizeof(uint32_t);
-+ err = XX_IpcSendMessage(p_Fm->h_IpcSessions[0],
-+ (uint8_t*)&msg,
-+ sizeof(msg.msgId) + sizeof(hardwarePortId),
-+ (uint8_t*)&reply,
-+ &replyLength,
-+ NULL,
-+ NULL);
-+ if (err != E_OK)
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+ if (replyLength != sizeof(uint32_t))
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("IPC reply length mismatch"));
-+ return (t_Error)(reply.error);
-+ }
-+ else if (!p_Fm->baseAddr)
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED,
-+ ("Either IPC or 'baseAddress' is required!"));
-+
-+ if (p_Fm->p_FmStateStruct->revInfo.majorRev >= 6)
-+ RETURN_ERROR(MINOR, E_NOT_AVAILABLE, ("Not available for this FM revision!"));
-+
-+ /* Get port status */
-+ err = FmIsPortStalled(h_Fm, hardwarePortId, &isStalled);
-+ if (err)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("Can't get port status"));
-+ if (!isStalled)
-+ return E_OK;
-+
-+ fman_resume_stalled_port(fpm_rg, hardwarePortId);
-+
-+ return E_OK;
-+}
-+
-+t_Error FmResetMac(t_Handle h_Fm, e_FmMacType type, uint8_t macId)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+ t_Error err;
-+ struct fman_fpm_regs *fpm_rg = p_Fm->p_FmFpmRegs;
-+
-+#if (DPAA_VERSION >= 11)
-+ if (p_Fm->p_FmStateStruct->revInfo.majorRev >= 6)
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED,
-+ ("FMan MAC reset!"));
-+#endif /*(DPAA_VERSION >= 11)*/
-+
-+ if ((p_Fm->guestId != NCSW_MASTER_ID) &&
-+ !p_Fm->baseAddr &&
-+ p_Fm->h_IpcSessions[0])
-+ {
-+ t_FmIpcMacParams macParams;
-+ t_FmIpcMsg msg;
-+ t_FmIpcReply reply;
-+ uint32_t replyLength;
-+
-+ memset(&msg, 0, sizeof(msg));
-+ memset(&reply, 0, sizeof(reply));
-+ macParams.id = macId;
-+ macParams.enumType = (uint32_t)type;
-+ msg.msgId = FM_RESET_MAC;
-+ memcpy(msg.msgBody, &macParams, sizeof(macParams));
-+ replyLength = sizeof(uint32_t);
-+ err = XX_IpcSendMessage(p_Fm->h_IpcSessions[0],
-+ (uint8_t*)&msg,
-+ sizeof(msg.msgId)+sizeof(macParams),
-+ (uint8_t*)&reply,
-+ &replyLength,
-+ NULL,
-+ NULL);
-+ if (err != E_OK)
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+ if (replyLength != sizeof(uint32_t))
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("IPC reply length mismatch"));
-+ return (t_Error)(reply.error);
-+ }
-+ else if (!p_Fm->baseAddr)
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED,
-+ ("Either IPC or 'baseAddress' is required!"));
-+
-+ err = (t_Error)fman_reset_mac(fpm_rg, macId, !!(type == e_FM_MAC_10G));
-+
-+ if (err == -EBUSY)
-+ return ERROR_CODE(E_TIMEOUT);
-+ else if (err)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("Illegal MAC ID"));
-+
-+ return E_OK;
-+}
-+
-+t_Error FmSetMacMaxFrame(t_Handle h_Fm, e_FmMacType type, uint8_t macId, uint16_t mtu)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+
-+ if ((p_Fm->guestId != NCSW_MASTER_ID) &&
-+ p_Fm->h_IpcSessions[0])
-+ {
-+ t_FmIpcMacMaxFrameParams macMaxFrameLengthParams;
-+ t_Error err;
-+ t_FmIpcMsg msg;
-+
-+ memset(&msg, 0, sizeof(msg));
-+ macMaxFrameLengthParams.macParams.id = macId;
-+ macMaxFrameLengthParams.macParams.enumType = (uint32_t)type;
-+ macMaxFrameLengthParams.maxFrameLength = (uint16_t)mtu;
-+ msg.msgId = FM_SET_MAC_MAX_FRAME;
-+ memcpy(msg.msgBody, &macMaxFrameLengthParams, sizeof(macMaxFrameLengthParams));
-+ err = XX_IpcSendMessage(p_Fm->h_IpcSessions[0],
-+ (uint8_t*)&msg,
-+ sizeof(msg.msgId)+sizeof(macMaxFrameLengthParams),
-+ NULL,
-+ NULL,
-+ NULL,
-+ NULL);
-+ if (err != E_OK)
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+ return E_OK;
-+ }
-+ else if (p_Fm->guestId != NCSW_MASTER_ID)
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED,
-+ ("running in guest-mode without IPC!"));
-+
-+ /* if port is already initialized, check that MaxFrameLength is smaller
-+ * or equal to the port's max */
-+#if (defined(FM_MAX_NUM_OF_10G_MACS) && (FM_MAX_NUM_OF_10G_MACS))
-+ if (type == e_FM_MAC_10G)
-+ {
-+ if ((!p_Fm->p_FmStateStruct->portMaxFrameLengths10G[macId])
-+ || (p_Fm->p_FmStateStruct->portMaxFrameLengths10G[macId] &&
-+ (mtu <= p_Fm->p_FmStateStruct->portMaxFrameLengths10G[macId])))
-+ p_Fm->p_FmStateStruct->macMaxFrameLengths10G[macId] = mtu;
-+ else
-+ RETURN_ERROR(MINOR, E_INVALID_VALUE, ("MAC maxFrameLength is larger than Port maxFrameLength"));
-+
-+ }
-+ else
-+#else
-+ UNUSED(type);
-+#endif /* (defined(FM_MAX_NUM_OF_10G_MACS) && ... */
-+ if ((!p_Fm->p_FmStateStruct->portMaxFrameLengths1G[macId])
-+ || (p_Fm->p_FmStateStruct->portMaxFrameLengths1G[macId] &&
-+ (mtu <= p_Fm->p_FmStateStruct->portMaxFrameLengths1G[macId])))
-+ p_Fm->p_FmStateStruct->macMaxFrameLengths1G[macId] = mtu;
-+ else
-+ RETURN_ERROR(MINOR, E_INVALID_VALUE, ("MAC maxFrameLength is larger than Port maxFrameLength"));
-+
-+ return E_OK;
-+}
-+
-+uint16_t FmGetClockFreq(t_Handle h_Fm)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+
-+ /* for multicore environment: this depends on the
-+ * fact that fmClkFreq was properly initialized at "init". */
-+ return p_Fm->p_FmStateStruct->fmClkFreq;
-+}
-+
-+uint16_t FmGetMacClockFreq(t_Handle h_Fm)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+
-+ return p_Fm->p_FmStateStruct->fmMacClkFreq;
-+}
-+
-+uint32_t FmGetTimeStampScale(t_Handle h_Fm)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+
-+ if ((p_Fm->guestId != NCSW_MASTER_ID) &&
-+ !p_Fm->baseAddr &&
-+ p_Fm->h_IpcSessions[0])
-+ {
-+ t_Error err;
-+ t_FmIpcMsg msg;
-+ t_FmIpcReply reply;
-+ uint32_t replyLength, timeStamp;
-+
-+ memset(&msg, 0, sizeof(msg));
-+ memset(&reply, 0, sizeof(reply));
-+ msg.msgId = FM_GET_TIMESTAMP_SCALE;
-+ replyLength = sizeof(uint32_t) + sizeof(uint32_t);
-+ if ((err = XX_IpcSendMessage(p_Fm->h_IpcSessions[0],
-+ (uint8_t*)&msg,
-+ sizeof(msg.msgId),
-+ (uint8_t*)&reply,
-+ &replyLength,
-+ NULL,
-+ NULL)) != E_OK)
-+ {
-+ REPORT_ERROR(MAJOR, err, NO_MSG);
-+ return 0;
-+ }
-+ if (replyLength != (sizeof(uint32_t) + sizeof(uint32_t)))
-+ {
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE, ("IPC reply length mismatch"));
-+ return 0;
-+ }
-+
-+ memcpy((uint8_t*)&timeStamp, reply.replyBody, sizeof(uint32_t));
-+ return timeStamp;
-+ }
-+ else if ((p_Fm->guestId != NCSW_MASTER_ID) &&
-+ p_Fm->baseAddr)
-+ {
-+ if (!(GET_UINT32(p_Fm->p_FmFpmRegs->fmfp_tsc1) & FPM_TS_CTL_EN))
-+ {
-+ REPORT_ERROR(MAJOR, E_INVALID_STATE, ("timestamp is not enabled!"));
-+ return 0;
-+ }
-+ }
-+ else if (p_Fm->guestId != NCSW_MASTER_ID)
-+ DBG(WARNING, ("No IPC - can't validate FM if timestamp enabled."));
-+
-+ return p_Fm->p_FmStateStruct->count1MicroBit;
-+}
-+
-+t_Error FmEnableRamsEcc(t_Handle h_Fm)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Fm, E_INVALID_HANDLE);
-+
-+ p_Fm->p_FmStateStruct->ramsEccOwners++;
-+ p_Fm->p_FmStateStruct->internalCall = TRUE;
-+
-+ return FM_EnableRamsEcc(p_Fm);
-+}
-+
-+t_Error FmDisableRamsEcc(t_Handle h_Fm)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Fm, E_INVALID_HANDLE);
-+
-+ ASSERT_COND(p_Fm->p_FmStateStruct->ramsEccOwners);
-+ p_Fm->p_FmStateStruct->ramsEccOwners--;
-+
-+ if (p_Fm->p_FmStateStruct->ramsEccOwners==0)
-+ {
-+ p_Fm->p_FmStateStruct->internalCall = TRUE;
-+ return FM_DisableRamsEcc(p_Fm);
-+ }
-+
-+ return E_OK;
-+}
-+
-+uint8_t FmGetGuestId(t_Handle h_Fm)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+
-+ return p_Fm->guestId;
-+}
-+
-+bool FmIsMaster(t_Handle h_Fm)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+
-+ return (p_Fm->guestId == NCSW_MASTER_ID);
-+}
-+
-+t_Error FmSetSizeOfFifo(t_Handle h_Fm,
-+ uint8_t hardwarePortId,
-+ uint32_t *p_SizeOfFifo,
-+ uint32_t *p_ExtraSizeOfFifo,
-+ bool initialConfig)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+ t_FmIpcPortRsrcParams rsrcParams;
-+ t_Error err;
-+ struct fman_bmi_regs *bmi_rg = p_Fm->p_FmBmiRegs;
-+ uint32_t sizeOfFifo = *p_SizeOfFifo, extraSizeOfFifo = *p_ExtraSizeOfFifo;
-+ uint16_t currentVal = 0, currentExtraVal = 0;
-+
-+ if ((p_Fm->guestId != NCSW_MASTER_ID) &&
-+ !p_Fm->baseAddr &&
-+ p_Fm->h_IpcSessions[0])
-+ {
-+ t_FmIpcMsg msg;
-+ t_FmIpcReply reply;
-+ uint32_t replyLength;
-+
-+ rsrcParams.hardwarePortId = hardwarePortId;
-+ rsrcParams.val = sizeOfFifo;
-+ rsrcParams.extra = extraSizeOfFifo;
-+ rsrcParams.boolInitialConfig = (uint8_t)initialConfig;
-+
-+ memset(&msg, 0, sizeof(msg));
-+ memset(&reply, 0, sizeof(reply));
-+ msg.msgId = FM_SET_SIZE_OF_FIFO;
-+ memcpy(msg.msgBody, &rsrcParams, sizeof(rsrcParams));
-+ replyLength = sizeof(uint32_t);
-+ if ((err = XX_IpcSendMessage(p_Fm->h_IpcSessions[0],
-+ (uint8_t*)&msg,
-+ sizeof(msg.msgId) + sizeof(rsrcParams),
-+ (uint8_t*)&reply,
-+ &replyLength,
-+ NULL,
-+ NULL)) != E_OK)
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+ if (replyLength != sizeof(uint32_t))
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("IPC reply length mismatch"));
-+ return (t_Error)(reply.error);
-+ }
-+ else if ((p_Fm->guestId != NCSW_MASTER_ID) &&
-+ p_Fm->baseAddr)
-+ {
-+ DBG(WARNING, ("No IPC - can't validate FM total-fifo size."));
-+ fman_set_size_of_fifo(bmi_rg, hardwarePortId, sizeOfFifo, extraSizeOfFifo);
-+ }
-+ else if (p_Fm->guestId != NCSW_MASTER_ID)
-+ RETURN_ERROR(MAJOR, E_NOT_SUPPORTED,
-+ ("running in guest-mode without neither IPC nor mapped register!"));
-+
-+ if (!initialConfig)
-+ {
-+ /* !initialConfig - runtime change of existing value.
-+ * - read the current FIFO and extra FIFO size */
-+ currentExtraVal = fman_get_size_of_extra_fifo(bmi_rg, hardwarePortId);
-+ currentVal = fman_get_size_of_fifo(bmi_rg, hardwarePortId);
-+ }
-+
-+ if (extraSizeOfFifo > currentExtraVal)
-+ {
-+ if (extraSizeOfFifo && !p_Fm->p_FmStateStruct->extraFifoPoolSize)
-+ /* if this is the first time a port requires extraFifoPoolSize, the total extraFifoPoolSize
-+ * must be initialized to 1 buffer per port
-+ */
-+ p_Fm->p_FmStateStruct->extraFifoPoolSize = FM_MAX_NUM_OF_RX_PORTS*BMI_FIFO_UNITS;
-+
-+ p_Fm->p_FmStateStruct->extraFifoPoolSize = MAX(p_Fm->p_FmStateStruct->extraFifoPoolSize, extraSizeOfFifo);
-+ }
-+
-+ /* check that there are enough uncommitted fifo size */
-+ if ((p_Fm->p_FmStateStruct->accumulatedFifoSize - currentVal + sizeOfFifo) >
-+ (p_Fm->p_FmStateStruct->totalFifoSize - p_Fm->p_FmStateStruct->extraFifoPoolSize)){
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE,
-+ ("Port request fifo size + accumulated size > total FIFO size:"));
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE,
-+ ("port 0x%x requested %d bytes, extra size = %d, accumulated size = %d total size = %d",
-+ hardwarePortId, sizeOfFifo, p_Fm->p_FmStateStruct->extraFifoPoolSize,
-+ p_Fm->p_FmStateStruct->accumulatedFifoSize,
-+ p_Fm->p_FmStateStruct->totalFifoSize));
-+ }
-+ else
-+ {
-+ /* update accumulated */
-+ ASSERT_COND(p_Fm->p_FmStateStruct->accumulatedFifoSize >= currentVal);
-+ p_Fm->p_FmStateStruct->accumulatedFifoSize -= currentVal;
-+ p_Fm->p_FmStateStruct->accumulatedFifoSize += sizeOfFifo;
-+ fman_set_size_of_fifo(bmi_rg, hardwarePortId, sizeOfFifo, extraSizeOfFifo);
-+ }
-+
-+ return E_OK;
-+}
-+
-+t_Error FmSetNumOfTasks(t_Handle h_Fm,
-+ uint8_t hardwarePortId,
-+ uint8_t *p_NumOfTasks,
-+ uint8_t *p_NumOfExtraTasks,
-+ bool initialConfig)
-+{
-+ t_Fm *p_Fm = (t_Fm *)h_Fm;
-+ t_Error err;
-+ struct fman_bmi_regs *bmi_rg = p_Fm->p_FmBmiRegs;
-+ uint8_t currentVal = 0, currentExtraVal = 0, numOfTasks = *p_NumOfTasks, numOfExtraTasks = *p_NumOfExtraTasks;
-+
-+ ASSERT_COND(IN_RANGE(1, hardwarePortId, 63));
-+
-+ if ((p_Fm->guestId != NCSW_MASTER_ID) &&
-+ !p_Fm->baseAddr &&
-+ p_Fm->h_IpcSessions[0])
-+ {
-+ t_FmIpcPortRsrcParams rsrcParams;
-+ t_FmIpcMsg msg;
-+ t_FmIpcReply reply;
-+ uint32_t replyLength;
-+
-+ rsrcParams.hardwarePortId = hardwarePortId;
-+ rsrcParams.val = numOfTasks;
-+ rsrcParams.extra = numOfExtraTasks;
-+ rsrcParams.boolInitialConfig = (uint8_t)initialConfig;
-+
-+ memset(&msg, 0, sizeof(msg));
-+ memset(&reply, 0, sizeof(reply));
-+ msg.msgId = FM_SET_NUM_OF_TASKS;
-+ memcpy(msg.msgBody, &rsrcParams, sizeof(rsrcParams));
-+ replyLength = sizeof(uint32_t);
-+ if ((err = XX_IpcSendMessage(p_Fm->h_IpcSessions[0],
-+ (uint8_t*)&msg,
-+ sizeof(msg.msgId) + sizeof(rsrcParams),
-+ (uint8_t*)&reply,
-+ &replyLength,
-+ NULL,
-+ NULL)) != E_OK)
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+ if (replyLength != sizeof(uint32_t))
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("IPC reply length mismatch"));
-+ return (t_Error)(reply.error);
-+ }
-+ else if ((p_Fm->guestId != NCSW_MASTER_ID) &&
-+ p_Fm->baseAddr)
-+ {
-+ DBG(WARNING, ("No IPC - can't validate FM total-num-of-tasks."));
-+ fman_set_num_of_tasks(bmi_rg, hardwarePortId, numOfTasks, numOfExtraTasks);
-+ }
-+ else if (p_Fm->guestId != NCSW_MASTER_ID)
-+ RETURN_ERROR(MAJOR, E_NOT_SUPPORTED,
-+ ("running in guest-mode without neither IPC nor mapped register!"));
-+
-+ if (!initialConfig)
-+ {
-+ /* !initialConfig - runtime change of existing value.
-+ * - read the current number of tasks */
-+ currentVal = fman_get_num_of_tasks(bmi_rg, hardwarePortId);
-+ currentExtraVal = fman_get_num_extra_tasks(bmi_rg, hardwarePortId);
-+ }
-+
-+ if (numOfExtraTasks > currentExtraVal)
-+ p_Fm->p_FmStateStruct->extraTasksPoolSize =
-+ (uint8_t)MAX(p_Fm->p_FmStateStruct->extraTasksPoolSize, numOfExtraTasks);
-+
-+ /* check that there are enough uncommitted tasks */
-+ if ((p_Fm->p_FmStateStruct->accumulatedNumOfTasks - currentVal + numOfTasks) >
-+ (p_Fm->p_FmStateStruct->totalNumOfTasks - p_Fm->p_FmStateStruct->extraTasksPoolSize))
-+ RETURN_ERROR(MAJOR, E_NOT_AVAILABLE,
-+ ("Requested numOfTasks and extra tasks pool for fm%d exceed total numOfTasks.",
-+ p_Fm->p_FmStateStruct->fmId));
-+ else
-+ {
-+ ASSERT_COND(p_Fm->p_FmStateStruct->accumulatedNumOfTasks >= currentVal);
-+ /* update accumulated */
-+ p_Fm->p_FmStateStruct->accumulatedNumOfTasks -= currentVal;
-+ p_Fm->p_FmStateStruct->accumulatedNumOfTasks += numOfTasks;
-+ fman_set_num_of_tasks(bmi_rg, hardwarePortId, numOfTasks, numOfExtraTasks);
-+ }
-+
-+ return E_OK;
-+}
-+
-+t_Error FmSetNumOfOpenDmas(t_Handle h_Fm,
-+ uint8_t hardwarePortId,
-+ uint8_t *p_NumOfOpenDmas,
-+ uint8_t *p_NumOfExtraOpenDmas,
-+ bool initialConfig)
-+
-+{
-+ t_Fm *p_Fm = (t_Fm *)h_Fm;
-+ t_Error err;
-+ struct fman_bmi_regs *bmi_rg = p_Fm->p_FmBmiRegs;
-+ uint8_t numOfOpenDmas = *p_NumOfOpenDmas, numOfExtraOpenDmas = *p_NumOfExtraOpenDmas;
-+ uint8_t totalNumDmas = 0, currentVal = 0, currentExtraVal = 0;
-+
-+ ASSERT_COND(IN_RANGE(1, hardwarePortId, 63));
-+
-+ if ((p_Fm->guestId != NCSW_MASTER_ID) &&
-+ !p_Fm->baseAddr &&
-+ p_Fm->h_IpcSessions[0])
-+ {
-+ t_FmIpcPortRsrcParams rsrcParams;
-+ t_FmIpcMsg msg;
-+ t_FmIpcReply reply;
-+ uint32_t replyLength;
-+
-+ rsrcParams.hardwarePortId = hardwarePortId;
-+ rsrcParams.val = numOfOpenDmas;
-+ rsrcParams.extra = numOfExtraOpenDmas;
-+ rsrcParams.boolInitialConfig = (uint8_t)initialConfig;
-+
-+ memset(&msg, 0, sizeof(msg));
-+ memset(&reply, 0, sizeof(reply));
-+ msg.msgId = FM_SET_NUM_OF_OPEN_DMAS;
-+ memcpy(msg.msgBody, &rsrcParams, sizeof(rsrcParams));
-+ replyLength = sizeof(uint32_t);
-+ if ((err = XX_IpcSendMessage(p_Fm->h_IpcSessions[0],
-+ (uint8_t*)&msg,
-+ sizeof(msg.msgId) + sizeof(rsrcParams),
-+ (uint8_t*)&reply,
-+ &replyLength,
-+ NULL,
-+ NULL)) != E_OK)
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+ if (replyLength != sizeof(uint32_t))
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("IPC reply length mismatch"));
-+ return (t_Error)(reply.error);
-+ }
-+#ifdef FM_HAS_TOTAL_DMAS
-+ else if (p_Fm->guestId != NCSW_MASTER_ID)
-+ RETURN_ERROR(MAJOR, E_NOT_SUPPORTED, ("running in guest-mode without IPC!"));
-+#else
-+ else if ((p_Fm->guestId != NCSW_MASTER_ID) &&
-+ p_Fm->baseAddr &&
-+ (p_Fm->p_FmStateStruct->revInfo.majorRev >= 6))
-+ {
-+ /*DBG(WARNING, ("No IPC - can't validate FM total-num-of-dmas."));*/
-+
-+ if (!numOfOpenDmas)
-+ {
-+ /* first config without explic it value: Do Nothing - reset value shouldn't be
-+ changed, read register for port save */
-+ *p_NumOfOpenDmas = fman_get_num_of_dmas(bmi_rg, hardwarePortId);
-+ *p_NumOfExtraOpenDmas = fman_get_num_extra_dmas(bmi_rg, hardwarePortId);
-+ }
-+ else
-+ /* whether it is the first time with explicit value, or runtime "set" - write register */
-+ fman_set_num_of_open_dmas(bmi_rg,
-+ hardwarePortId,
-+ numOfOpenDmas,
-+ numOfExtraOpenDmas,
-+ p_Fm->p_FmStateStruct->accumulatedNumOfOpenDmas + p_Fm->p_FmStateStruct->extraOpenDmasPoolSize);
-+ }
-+ else if (p_Fm->guestId != NCSW_MASTER_ID)
-+ RETURN_ERROR(MAJOR, E_NOT_SUPPORTED,
-+ ("running in guest-mode without neither IPC nor mapped register!"));
-+#endif /* FM_HAS_TOTAL_DMAS */
-+
-+ if (!initialConfig)
-+ {
-+ /* !initialConfig - runtime change of existing value.
-+ * - read the current number of open Dma's */
-+ currentExtraVal = fman_get_num_extra_dmas(bmi_rg, hardwarePortId);
-+ currentVal = fman_get_num_of_dmas(bmi_rg, hardwarePortId);
-+ }
-+
-+#ifdef FM_NO_GUARANTEED_RESET_VALUES
-+ /* it's illegal to be in a state where this is not the first set and no value is specified */
-+ ASSERT_COND(initialConfig || numOfOpenDmas);
-+ if (!numOfOpenDmas)
-+ {
-+ /* !numOfOpenDmas - first configuration according to values in regs.
-+ * - read the current number of open Dma's */
-+ currentExtraVal = fman_get_num_extra_dmas(bmi_rg, hardwarePortId);
-+ currentVal = fman_get_num_of_dmas(bmi_rg, hardwarePortId);
-+ /* This is the first configuration and user did not specify value (!numOfOpenDmas),
-+ * reset values will be used and we just save these values for resource management */
-+ p_Fm->p_FmStateStruct->extraOpenDmasPoolSize =
-+ (uint8_t)MAX(p_Fm->p_FmStateStruct->extraOpenDmasPoolSize, currentExtraVal);
-+ p_Fm->p_FmStateStruct->accumulatedNumOfOpenDmas += currentVal;
-+ *p_NumOfOpenDmas = currentVal;
-+ *p_NumOfExtraOpenDmas = currentExtraVal;
-+ return E_OK;
-+ }
-+#endif /* FM_NO_GUARANTEED_RESET_VALUES */
-+
-+ if (numOfExtraOpenDmas > currentExtraVal)
-+ p_Fm->p_FmStateStruct->extraOpenDmasPoolSize =
-+ (uint8_t)MAX(p_Fm->p_FmStateStruct->extraOpenDmasPoolSize, numOfExtraOpenDmas);
-+
-+#ifdef FM_HAS_TOTAL_DMAS
-+ if ((p_Fm->p_FmStateStruct->revInfo.majorRev < 6) &&
-+ (p_Fm->p_FmStateStruct->accumulatedNumOfOpenDmas - currentVal + numOfOpenDmas >
-+ p_Fm->p_FmStateStruct->maxNumOfOpenDmas))
-+ RETURN_ERROR(MAJOR, E_NOT_AVAILABLE,
-+ ("Requested numOfOpenDmas for fm%d exceeds total numOfOpenDmas.",
-+ p_Fm->p_FmStateStruct->fmId));
-+#else
-+ if ((p_Fm->p_FmStateStruct->revInfo.majorRev >= 6) &&
-+#ifdef FM_HEAVY_TRAFFIC_SEQUENCER_HANG_ERRATA_FMAN_A006981
-+ !((p_Fm->p_FmStateStruct->revInfo.majorRev == 6) &&
-+ (p_Fm->p_FmStateStruct->revInfo.minorRev == 0)) &&
-+#endif /* FM_HEAVY_TRAFFIC_SEQUENCER_HANG_ERRATA_FMAN_A006981 */
-+ (p_Fm->p_FmStateStruct->accumulatedNumOfOpenDmas - currentVal + numOfOpenDmas > DMA_THRESH_MAX_COMMQ + 1))
-+ RETURN_ERROR(MAJOR, E_NOT_AVAILABLE,
-+ ("Requested numOfOpenDmas for fm%d exceeds DMA Command queue (%d)",
-+ p_Fm->p_FmStateStruct->fmId, DMA_THRESH_MAX_COMMQ+1));
-+#endif /* FM_HAS_TOTAL_DMAS */
-+ else
-+ {
-+ ASSERT_COND(p_Fm->p_FmStateStruct->accumulatedNumOfOpenDmas >= currentVal);
-+ /* update acummulated */
-+ p_Fm->p_FmStateStruct->accumulatedNumOfOpenDmas -= currentVal;
-+ p_Fm->p_FmStateStruct->accumulatedNumOfOpenDmas += numOfOpenDmas;
-+
-+#ifdef FM_HAS_TOTAL_DMAS
-+ if (p_Fm->p_FmStateStruct->revInfo.majorRev < 6)
-+ totalNumDmas = (uint8_t)(p_Fm->p_FmStateStruct->accumulatedNumOfOpenDmas + p_Fm->p_FmStateStruct->extraOpenDmasPoolSize);
-+#endif /* FM_HAS_TOTAL_DMAS */
-+ fman_set_num_of_open_dmas(bmi_rg,
-+ hardwarePortId,
-+ numOfOpenDmas,
-+ numOfExtraOpenDmas,
-+ totalNumDmas);
-+ }
-+
-+ return E_OK;
-+}
-+
-+#if (DPAA_VERSION >= 11)
-+t_Error FmVSPCheckRelativeProfile(t_Handle h_Fm,
-+ e_FmPortType portType,
-+ uint8_t portId,
-+ uint16_t relativeProfile)
-+{
-+ t_Fm *p_Fm;
-+ t_FmSp *p_FmPcdSp;
-+ uint8_t swPortIndex=0, hardwarePortId;
-+
-+ ASSERT_COND(h_Fm);
-+ p_Fm = (t_Fm*)h_Fm;
-+
-+ hardwarePortId = SwPortIdToHwPortId(portType,
-+ portId,
-+ p_Fm->p_FmStateStruct->revInfo.majorRev,
-+ p_Fm->p_FmStateStruct->revInfo.minorRev);
-+ ASSERT_COND(hardwarePortId);
-+ HW_PORT_ID_TO_SW_PORT_INDX(swPortIndex, hardwarePortId);
-+
-+ p_FmPcdSp = p_Fm->p_FmSp;
-+ ASSERT_COND(p_FmPcdSp);
-+
-+ if (!p_FmPcdSp->portsMapping[swPortIndex].numOfProfiles)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE , ("Port has no allocated profiles"));
-+ if (relativeProfile >= p_FmPcdSp->portsMapping[swPortIndex].numOfProfiles)
-+ RETURN_ERROR(MAJOR, E_NOT_IN_RANGE , ("Profile id is out of range"));
-+
-+ return E_OK;
-+}
-+
-+t_Error FmVSPGetAbsoluteProfileId(t_Handle h_Fm,
-+ e_FmPortType portType,
-+ uint8_t portId,
-+ uint16_t relativeProfile,
-+ uint16_t *p_AbsoluteId)
-+{
-+ t_Fm *p_Fm;
-+ t_FmSp *p_FmPcdSp;
-+ uint8_t swPortIndex=0, hardwarePortId;
-+ t_Error err;
-+
-+ ASSERT_COND(h_Fm);
-+ p_Fm = (t_Fm*)h_Fm;
-+
-+ err = FmVSPCheckRelativeProfile(h_Fm, portType, portId, relativeProfile);
-+ if (err != E_OK)
-+ return err;
-+
-+ hardwarePortId = SwPortIdToHwPortId(portType,
-+ portId,
-+ p_Fm->p_FmStateStruct->revInfo.majorRev,
-+ p_Fm->p_FmStateStruct->revInfo.minorRev);
-+ ASSERT_COND(hardwarePortId);
-+ HW_PORT_ID_TO_SW_PORT_INDX(swPortIndex, hardwarePortId);
-+
-+ p_FmPcdSp = p_Fm->p_FmSp;
-+ ASSERT_COND(p_FmPcdSp);
-+
-+ *p_AbsoluteId = (uint16_t)(p_FmPcdSp->portsMapping[swPortIndex].profilesBase + relativeProfile);
-+
-+ return E_OK;
-+}
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+static t_Error InitFmDma(t_Fm *p_Fm)
-+{
-+ t_Error err;
-+
-+ err = (t_Error)fman_dma_init(p_Fm->p_FmDmaRegs, p_Fm->p_FmDriverParam);
-+ if (err != E_OK)
-+ return err;
-+
-+ /* Allocate MURAM for CAM */
-+ p_Fm->camBaseAddr = PTR_TO_UINT(FM_MURAM_AllocMem(p_Fm->h_FmMuram,
-+ (uint32_t)(p_Fm->p_FmDriverParam->dma_cam_num_of_entries*DMA_CAM_SIZEOF_ENTRY),
-+ DMA_CAM_ALIGN));
-+ if (!p_Fm->camBaseAddr)
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, ("MURAM alloc for DMA CAM failed"));
-+
-+ WRITE_BLOCK(UINT_TO_PTR(p_Fm->camBaseAddr),
-+ 0,
-+ (uint32_t)(p_Fm->p_FmDriverParam->dma_cam_num_of_entries*DMA_CAM_SIZEOF_ENTRY));
-+
-+ if (p_Fm->p_FmStateStruct->revInfo.majorRev == 2)
-+ {
-+ FM_MURAM_FreeMem(p_Fm->h_FmMuram, UINT_TO_PTR(p_Fm->camBaseAddr));
-+
-+ p_Fm->camBaseAddr = PTR_TO_UINT(FM_MURAM_AllocMem(p_Fm->h_FmMuram,
-+ (uint32_t)(p_Fm->p_FmDriverParam->dma_cam_num_of_entries*72 + 128),
-+ 64));
-+ if (!p_Fm->camBaseAddr)
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, ("MURAM alloc for DMA CAM failed"));
-+
-+ WRITE_BLOCK(UINT_TO_PTR(p_Fm->camBaseAddr),
-+ 0,
-+ (uint32_t)(p_Fm->p_FmDriverParam->dma_cam_num_of_entries*72 + 128));
-+
-+ switch(p_Fm->p_FmDriverParam->dma_cam_num_of_entries)
-+ {
-+ case (8):
-+ WRITE_UINT32(*(uint32_t*)p_Fm->camBaseAddr, 0xff000000);
-+ break;
-+ case (16):
-+ WRITE_UINT32(*(uint32_t*)p_Fm->camBaseAddr, 0xffff0000);
-+ break;
-+ case (24):
-+ WRITE_UINT32(*(uint32_t*)p_Fm->camBaseAddr, 0xffffff00);
-+ break;
-+ case (32):
-+ WRITE_UINT32(*(uint32_t*)p_Fm->camBaseAddr, 0xffffffff);
-+ break;
-+ default:
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("wrong dma_cam_num_of_entries"));
-+ }
-+ }
-+
-+ p_Fm->p_FmDriverParam->cam_base_addr =
-+ (uint32_t)(XX_VirtToPhys(UINT_TO_PTR(p_Fm->camBaseAddr)) - p_Fm->fmMuramPhysBaseAddr);
-+
-+ return E_OK;
-+}
-+
-+static t_Error InitFmFpm(t_Fm *p_Fm)
-+{
-+ return (t_Error)fman_fpm_init(p_Fm->p_FmFpmRegs, p_Fm->p_FmDriverParam);
-+}
-+
-+static t_Error InitFmBmi(t_Fm *p_Fm)
-+{
-+ return (t_Error)fman_bmi_init(p_Fm->p_FmBmiRegs, p_Fm->p_FmDriverParam);
-+}
-+
-+static t_Error InitFmQmi(t_Fm *p_Fm)
-+{
-+ return (t_Error)fman_qmi_init(p_Fm->p_FmQmiRegs, p_Fm->p_FmDriverParam);
-+}
-+
-+static t_Error InitGuestMode(t_Fm *p_Fm)
-+{
-+ t_Error err = E_OK;
-+ int i;
-+ t_FmIpcMsg msg;
-+ t_FmIpcReply reply;
-+ uint32_t replyLength;
-+
-+ ASSERT_COND(p_Fm);
-+ ASSERT_COND(p_Fm->guestId != NCSW_MASTER_ID);
-+
-+ /* build the FM guest partition IPC address */
-+ if (Sprint (p_Fm->fmModuleName, "FM_%d_%d",p_Fm->p_FmStateStruct->fmId, p_Fm->guestId) != (p_Fm->guestId<10 ? 6:7))
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("Sprint failed"));
-+
-+ /* build the FM master partition IPC address */
-+ memset(p_Fm->fmIpcHandlerModuleName, 0, (sizeof(char)) * MODULE_NAME_SIZE);
-+ if (Sprint (p_Fm->fmIpcHandlerModuleName[0], "FM_%d_%d",p_Fm->p_FmStateStruct->fmId, NCSW_MASTER_ID) != 6)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("Sprint failed"));
-+
-+ for (i=0;i<e_FM_EV_DUMMY_LAST;i++)
-+ p_Fm->intrMng[i].f_Isr = UnimplementedIsr;
-+
-+ p_Fm->h_IpcSessions[0] = XX_IpcInitSession(p_Fm->fmIpcHandlerModuleName[0], p_Fm->fmModuleName);
-+ if (p_Fm->h_IpcSessions[0])
-+ {
-+ uint8_t isMasterAlive;
-+ t_FmIpcParams ipcParams;
-+
-+ err = XX_IpcRegisterMsgHandler(p_Fm->fmModuleName, FmGuestHandleIpcMsgCB, p_Fm, FM_IPC_MAX_REPLY_SIZE);
-+ if (err)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+
-+ memset(&msg, 0, sizeof(msg));
-+ memset(&reply, 0, sizeof(reply));
-+ msg.msgId = FM_MASTER_IS_ALIVE;
-+ msg.msgBody[0] = p_Fm->guestId;
-+ replyLength = sizeof(uint32_t) + sizeof(uint8_t);
-+ do
-+ {
-+ blockingFlag = TRUE;
-+ if ((err = XX_IpcSendMessage(p_Fm->h_IpcSessions[0],
-+ (uint8_t*)&msg,
-+ sizeof(msg.msgId)+sizeof(p_Fm->guestId),
-+ (uint8_t*)&reply,
-+ &replyLength,
-+ IpcMsgCompletionCB,
-+ p_Fm)) != E_OK)
-+ REPORT_ERROR(MINOR, err, NO_MSG);
-+ while (blockingFlag) ;
-+ if (replyLength != (sizeof(uint32_t) + sizeof(uint8_t)))
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE, ("IPC reply length mismatch"));
-+ isMasterAlive = *(uint8_t*)(reply.replyBody);
-+ } while (!isMasterAlive);
-+
-+ /* read FM parameters and save */
-+ memset(&msg, 0, sizeof(msg));
-+ memset(&reply, 0, sizeof(reply));
-+ msg.msgId = FM_GET_PARAMS;
-+ replyLength = sizeof(uint32_t) + sizeof(t_FmIpcParams);
-+ if ((err = XX_IpcSendMessage(p_Fm->h_IpcSessions[0],
-+ (uint8_t*)&msg,
-+ sizeof(msg.msgId),
-+ (uint8_t*)&reply,
-+ &replyLength,
-+ NULL,
-+ NULL)) != E_OK)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ if (replyLength != (sizeof(uint32_t) + sizeof(t_FmIpcParams)))
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("IPC reply length mismatch"));
-+ memcpy((uint8_t*)&ipcParams, reply.replyBody, sizeof(t_FmIpcParams));
-+
-+ p_Fm->p_FmStateStruct->fmClkFreq = ipcParams.fmClkFreq;
-+ p_Fm->p_FmStateStruct->fmMacClkFreq = ipcParams.fmMacClkFreq;
-+ p_Fm->p_FmStateStruct->revInfo.majorRev = ipcParams.majorRev;
-+ p_Fm->p_FmStateStruct->revInfo.minorRev = ipcParams.minorRev;
-+ }
-+ else
-+ {
-+ DBG(WARNING, ("FM Guest mode - without IPC"));
-+ if (!p_Fm->p_FmStateStruct->fmClkFreq)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("No fmClkFreq configured for guest without IPC"));
-+ if (p_Fm->baseAddr)
-+ {
-+ fman_get_revision(p_Fm->p_FmFpmRegs,
-+ &p_Fm->p_FmStateStruct->revInfo.majorRev,
-+ &p_Fm->p_FmStateStruct->revInfo.minorRev);
-+
-+ }
-+ }
-+
-+#if (DPAA_VERSION >= 11)
-+ p_Fm->partVSPBase = AllocVSPsForPartition(p_Fm, p_Fm->partVSPBase, p_Fm->partNumOfVSPs, p_Fm->guestId);
-+ if (p_Fm->partVSPBase == (uint8_t)(ILLEGAL_BASE))
-+ DBG(WARNING, ("partition VSPs allocation is FAILED"));
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+ /* General FM driver initialization */
-+ if (p_Fm->baseAddr)
-+ p_Fm->fmMuramPhysBaseAddr =
-+ (uint64_t)(XX_VirtToPhys(UINT_TO_PTR(p_Fm->baseAddr + FM_MM_MURAM)));
-+
-+ XX_Free(p_Fm->p_FmDriverParam);
-+ p_Fm->p_FmDriverParam = NULL;
-+
-+ if ((p_Fm->guestId == NCSW_MASTER_ID) ||
-+ (p_Fm->h_IpcSessions[0]))
-+ {
-+ FM_DisableRamsEcc(p_Fm);
-+ FmMuramClear(p_Fm->h_FmMuram);
-+ FM_EnableRamsEcc(p_Fm);
-+ }
-+
-+ return E_OK;
-+}
-+
-+static __inline__ enum fman_exceptions FmanExceptionTrans(e_FmExceptions exception)
-+{
-+ switch (exception) {
-+ case e_FM_EX_DMA_BUS_ERROR:
-+ return E_FMAN_EX_DMA_BUS_ERROR;
-+ case e_FM_EX_DMA_READ_ECC:
-+ return E_FMAN_EX_DMA_READ_ECC;
-+ case e_FM_EX_DMA_SYSTEM_WRITE_ECC:
-+ return E_FMAN_EX_DMA_SYSTEM_WRITE_ECC;
-+ case e_FM_EX_DMA_FM_WRITE_ECC:
-+ return E_FMAN_EX_DMA_FM_WRITE_ECC;
-+ case e_FM_EX_FPM_STALL_ON_TASKS:
-+ return E_FMAN_EX_FPM_STALL_ON_TASKS;
-+ case e_FM_EX_FPM_SINGLE_ECC:
-+ return E_FMAN_EX_FPM_SINGLE_ECC;
-+ case e_FM_EX_FPM_DOUBLE_ECC:
-+ return E_FMAN_EX_FPM_DOUBLE_ECC;
-+ case e_FM_EX_QMI_SINGLE_ECC:
-+ return E_FMAN_EX_QMI_SINGLE_ECC;
-+ case e_FM_EX_QMI_DOUBLE_ECC:
-+ return E_FMAN_EX_QMI_DOUBLE_ECC;
-+ case e_FM_EX_QMI_DEQ_FROM_UNKNOWN_PORTID:
-+ return E_FMAN_EX_QMI_DEQ_FROM_UNKNOWN_PORTID;
-+ case e_FM_EX_BMI_LIST_RAM_ECC:
-+ return E_FMAN_EX_BMI_LIST_RAM_ECC;
-+ case e_FM_EX_BMI_STORAGE_PROFILE_ECC:
-+ return E_FMAN_EX_BMI_STORAGE_PROFILE_ECC;
-+ case e_FM_EX_BMI_STATISTICS_RAM_ECC:
-+ return E_FMAN_EX_BMI_STATISTICS_RAM_ECC;
-+ case e_FM_EX_BMI_DISPATCH_RAM_ECC:
-+ return E_FMAN_EX_BMI_DISPATCH_RAM_ECC;
-+ case e_FM_EX_IRAM_ECC:
-+ return E_FMAN_EX_IRAM_ECC;
-+ case e_FM_EX_MURAM_ECC:
-+ return E_FMAN_EX_MURAM_ECC;
-+ default:
-+ return E_FMAN_EX_DMA_BUS_ERROR;
-+ }
-+}
-+
-+uint8_t SwPortIdToHwPortId(e_FmPortType type, uint8_t relativePortId, uint8_t majorRev, uint8_t minorRev)
-+{
-+ switch (type)
-+ {
-+ case (e_FM_PORT_TYPE_OH_OFFLINE_PARSING):
-+ case (e_FM_PORT_TYPE_OH_HOST_COMMAND):
-+ CHECK_PORT_ID_OH_PORTS(relativePortId);
-+ return (uint8_t)(BASE_OH_PORTID + (relativePortId));
-+ case (e_FM_PORT_TYPE_RX):
-+ CHECK_PORT_ID_1G_RX_PORTS(relativePortId);
-+ return (uint8_t)(BASE_1G_RX_PORTID + (relativePortId));
-+ case (e_FM_PORT_TYPE_RX_10G):
-+ /* The 10G port in T1024 (FMan Version 6.4) is the first port.
-+ * This is the reason why the 1G port offset is used.
-+ */
-+ if (majorRev == 6 && minorRev == 4)
-+ {
-+ CHECK_PORT_ID_1G_RX_PORTS(relativePortId);
-+ return (uint8_t)(BASE_1G_RX_PORTID + (relativePortId));
-+ }
-+ else
-+ {
-+ CHECK_PORT_ID_10G_RX_PORTS(relativePortId);
-+ return (uint8_t)(BASE_10G_RX_PORTID + (relativePortId));
-+ }
-+ case (e_FM_PORT_TYPE_TX):
-+ CHECK_PORT_ID_1G_TX_PORTS(relativePortId);
-+ return (uint8_t)(BASE_1G_TX_PORTID + (relativePortId));
-+ case (e_FM_PORT_TYPE_TX_10G):
-+ /* The 10G port in T1024 (FMan Version 6.4) is the first port.
-+ * This is the reason why the 1G port offset is used.
-+ */
-+ if (majorRev == 6 && minorRev == 4)
-+ {
-+ CHECK_PORT_ID_1G_TX_PORTS(relativePortId);
-+ return (uint8_t)(BASE_1G_TX_PORTID + (relativePortId));
-+ }
-+ else
-+ {
-+ CHECK_PORT_ID_10G_TX_PORTS(relativePortId);
-+ return (uint8_t)(BASE_10G_TX_PORTID + (relativePortId));
-+ }
-+ default:
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE, ("Illegal port type"));
-+ return 0;
-+ }
-+}
-+
-+#if (defined(DEBUG_ERRORS) && (DEBUG_ERRORS > 0))
-+t_Error FmDumpPortRegs (t_Handle h_Fm, uint8_t hardwarePortId)
-+{
-+ t_Fm *p_Fm = (t_Fm *)h_Fm;
-+
-+ DECLARE_DUMP;
-+
-+ ASSERT_COND(IN_RANGE(1, hardwarePortId, 63));
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Fm, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(((p_Fm->guestId == NCSW_MASTER_ID) ||
-+ p_Fm->baseAddr), E_INVALID_OPERATION);
-+
-+ DUMP_TITLE(&p_Fm->p_FmBmiRegs->fmbm_pp[hardwarePortId-1], ("fmbm_pp for port %u", (hardwarePortId)));
-+ DUMP_MEMORY(&p_Fm->p_FmBmiRegs->fmbm_pp[hardwarePortId-1], sizeof(uint32_t));
-+
-+ DUMP_TITLE(&p_Fm->p_FmBmiRegs->fmbm_pfs[hardwarePortId-1], ("fmbm_pfs for port %u", (hardwarePortId )));
-+ DUMP_MEMORY(&p_Fm->p_FmBmiRegs->fmbm_pfs[hardwarePortId-1], sizeof(uint32_t));
-+
-+ DUMP_TITLE(&p_Fm->p_FmBmiRegs->fmbm_spliodn[hardwarePortId-1], ("fmbm_spliodn for port %u", (hardwarePortId)));
-+ DUMP_MEMORY(&p_Fm->p_FmBmiRegs->fmbm_spliodn[hardwarePortId-1], sizeof(uint32_t));
-+
-+ DUMP_TITLE(&p_Fm->p_FmFpmRegs->fmfp_ps[hardwarePortId], ("fmfp_ps for port %u", (hardwarePortId)));
-+ DUMP_MEMORY(&p_Fm->p_FmFpmRegs->fmfp_ps[hardwarePortId], sizeof(uint32_t));
-+
-+ DUMP_TITLE(&p_Fm->p_FmDmaRegs->fmdmplr[hardwarePortId/2], ("fmdmplr for port %u", (hardwarePortId)));
-+ DUMP_MEMORY(&p_Fm->p_FmDmaRegs->fmdmplr[hardwarePortId/2], sizeof(uint32_t));
-+
-+ return E_OK;
-+}
-+#endif /* (defined(DEBUG_ERRORS) && (DEBUG_ERRORS > 0)) */
-+
-+
-+/*****************************************************************************/
-+/* API Init unit functions */
-+/*****************************************************************************/
-+t_Handle FM_Config(t_FmParams *p_FmParam)
-+{
-+ t_Fm *p_Fm;
-+ uint8_t i;
-+ uintptr_t baseAddr;
-+
-+ SANITY_CHECK_RETURN_VALUE(p_FmParam, E_NULL_POINTER, NULL);
-+ SANITY_CHECK_RETURN_VALUE(((p_FmParam->firmware.p_Code && p_FmParam->firmware.size) ||
-+ (!p_FmParam->firmware.p_Code && !p_FmParam->firmware.size)),
-+ E_INVALID_VALUE, NULL);
-+
-+ baseAddr = p_FmParam->baseAddr;
-+
-+ /* Allocate FM structure */
-+ p_Fm = (t_Fm *) XX_Malloc(sizeof(t_Fm));
-+ if (!p_Fm)
-+ {
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("FM driver structure"));
-+ return NULL;
-+ }
-+ memset(p_Fm, 0, sizeof(t_Fm));
-+
-+ p_Fm->p_FmStateStruct = (t_FmStateStruct *) XX_Malloc(sizeof(t_FmStateStruct));
-+ if (!p_Fm->p_FmStateStruct)
-+ {
-+ XX_Free(p_Fm);
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("FM Status structure"));
-+ return NULL;
-+ }
-+ memset(p_Fm->p_FmStateStruct, 0, sizeof(t_FmStateStruct));
-+
-+ /* Initialize FM parameters which will be kept by the driver */
-+ p_Fm->p_FmStateStruct->fmId = p_FmParam->fmId;
-+ p_Fm->guestId = p_FmParam->guestId;
-+
-+ for (i=0; i<FM_MAX_NUM_OF_HW_PORT_IDS; i++)
-+ p_Fm->p_FmStateStruct->portsTypes[i] = e_FM_PORT_TYPE_DUMMY;
-+
-+ /* Allocate the FM driver's parameters structure */
-+ p_Fm->p_FmDriverParam = (struct fman_cfg *)XX_Malloc(sizeof(struct fman_cfg));
-+ if (!p_Fm->p_FmDriverParam)
-+ {
-+ XX_Free(p_Fm->p_FmStateStruct);
-+ XX_Free(p_Fm);
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("FM driver parameters"));
-+ return NULL;
-+ }
-+ memset(p_Fm->p_FmDriverParam, 0, sizeof(struct fman_cfg));
-+
-+#if (DPAA_VERSION >= 11)
-+ p_Fm->p_FmSp = (t_FmSp *)XX_Malloc(sizeof(t_FmSp));
-+ if (!p_Fm->p_FmSp)
-+ {
-+ XX_Free(p_Fm->p_FmDriverParam);
-+ XX_Free(p_Fm->p_FmStateStruct);
-+ XX_Free(p_Fm);
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("allocation for internal data structure failed"));
-+ return NULL;
-+ }
-+ memset(p_Fm->p_FmSp, 0, sizeof(t_FmSp));
-+
-+ for (i=0; i<FM_VSP_MAX_NUM_OF_ENTRIES; i++)
-+ p_Fm->p_FmSp->profiles[i].profilesMng.ownerId = (uint8_t)ILLEGAL_BASE;
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+ /* Initialize FM parameters which will be kept by the driver */
-+ p_Fm->p_FmStateStruct->fmId = p_FmParam->fmId;
-+ p_Fm->h_FmMuram = p_FmParam->h_FmMuram;
-+ p_Fm->h_App = p_FmParam->h_App;
-+ p_Fm->p_FmStateStruct->fmClkFreq = p_FmParam->fmClkFreq;
-+ p_Fm->p_FmStateStruct->fmMacClkFreq = p_FmParam->fmClkFreq / ((!p_FmParam->fmMacClkRatio)? 2: p_FmParam->fmMacClkRatio);
-+ p_Fm->f_Exception = p_FmParam->f_Exception;
-+ p_Fm->f_BusError = p_FmParam->f_BusError;
-+ p_Fm->p_FmFpmRegs = (struct fman_fpm_regs *)UINT_TO_PTR(baseAddr + FM_MM_FPM);
-+ p_Fm->p_FmBmiRegs = (struct fman_bmi_regs *)UINT_TO_PTR(baseAddr + FM_MM_BMI);
-+ p_Fm->p_FmQmiRegs = (struct fman_qmi_regs *)UINT_TO_PTR(baseAddr + FM_MM_QMI);
-+ p_Fm->p_FmDmaRegs = (struct fman_dma_regs *)UINT_TO_PTR(baseAddr + FM_MM_DMA);
-+ p_Fm->p_FmRegs = (struct fman_regs *)UINT_TO_PTR(baseAddr + FM_MM_BMI);
-+ p_Fm->baseAddr = baseAddr;
-+ p_Fm->p_FmStateStruct->irq = p_FmParam->irq;
-+ p_Fm->p_FmStateStruct->errIrq = p_FmParam->errIrq;
-+ p_Fm->hcPortInitialized = FALSE;
-+ p_Fm->independentMode = FALSE;
-+
-+ p_Fm->h_Spinlock = XX_InitSpinlock();
-+ if (!p_Fm->h_Spinlock)
-+ {
-+ XX_Free(p_Fm->p_FmDriverParam);
-+ XX_Free(p_Fm->p_FmStateStruct);
-+ XX_Free(p_Fm);
-+ REPORT_ERROR(MAJOR, E_INVALID_STATE, ("can't allocate spinlock!"));
-+ return NULL;
-+ }
-+
-+#if (DPAA_VERSION >= 11)
-+ p_Fm->partVSPBase = p_FmParam->partVSPBase;
-+ p_Fm->partNumOfVSPs = p_FmParam->partNumOfVSPs;
-+ p_Fm->vspBaseAddr = p_FmParam->vspBaseAddr;
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+ fman_defconfig(p_Fm->p_FmDriverParam,
-+ !!(p_Fm->guestId == NCSW_MASTER_ID));
-+/* overide macros dependent parameters */
-+#ifdef FM_PEDANTIC_DMA
-+ p_Fm->p_FmDriverParam->pedantic_dma = TRUE;
-+ p_Fm->p_FmDriverParam->dma_aid_override = TRUE;
-+#endif /* FM_PEDANTIC_DMA */
-+#ifndef FM_QMI_NO_DEQ_OPTIONS_SUPPORT
-+ p_Fm->p_FmDriverParam->qmi_deq_option_support = TRUE;
-+#endif /* !FM_QMI_NO_DEQ_OPTIONS_SUPPORT */
-+
-+ p_Fm->p_FmStateStruct->ramsEccEnable = FALSE;
-+ p_Fm->p_FmStateStruct->extraFifoPoolSize = 0;
-+ p_Fm->p_FmStateStruct->exceptions = DEFAULT_exceptions;
-+ p_Fm->resetOnInit = DEFAULT_resetOnInit;
-+ p_Fm->f_ResetOnInitOverride = DEFAULT_resetOnInitOverrideCallback;
-+ p_Fm->fwVerify = DEFAULT_VerifyUcode;
-+ p_Fm->firmware.size = p_FmParam->firmware.size;
-+ if (p_Fm->firmware.size)
-+ {
-+ p_Fm->firmware.p_Code = (uint32_t *)XX_Malloc(p_Fm->firmware.size);
-+ if (!p_Fm->firmware.p_Code)
-+ {
-+ XX_FreeSpinlock(p_Fm->h_Spinlock);
-+ XX_Free(p_Fm->p_FmStateStruct);
-+ XX_Free(p_Fm->p_FmDriverParam);
-+ XX_Free(p_Fm);
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("FM firmware code"));
-+ return NULL;
-+ }
-+ memcpy(p_Fm->firmware.p_Code, p_FmParam->firmware.p_Code ,p_Fm->firmware.size);
-+ }
-+
-+ if (p_Fm->guestId != NCSW_MASTER_ID)
-+ return p_Fm;
-+
-+ /* read revision */
-+ /* Chip dependent, will be configured in Init */
-+ fman_get_revision(p_Fm->p_FmFpmRegs,
-+ &p_Fm->p_FmStateStruct->revInfo.majorRev,
-+ &p_Fm->p_FmStateStruct->revInfo.minorRev);
-+
-+#ifdef FM_AID_MODE_NO_TNUM_SW005
-+ if (p_Fm->p_FmStateStruct->revInfo.majorRev >= 6)
-+ p_Fm->p_FmDriverParam->dma_aid_mode = e_FM_DMA_AID_OUT_PORT_ID;
-+#endif /* FM_AID_MODE_NO_TNUM_SW005 */
-+#ifndef FM_QMI_NO_DEQ_OPTIONS_SUPPORT
-+ if (p_Fm->p_FmStateStruct->revInfo.majorRev != 4)
-+ p_Fm->p_FmDriverParam->qmi_def_tnums_thresh = QMI_DEF_TNUMS_THRESH;
-+#endif /* FM_QMI_NO_DEQ_OPTIONS_SUPPORT */
-+
-+ p_Fm->p_FmStateStruct->totalFifoSize = 0;
-+ p_Fm->p_FmStateStruct->totalNumOfTasks =
-+ DEFAULT_totalNumOfTasks(p_Fm->p_FmStateStruct->revInfo.majorRev,
-+ p_Fm->p_FmStateStruct->revInfo.minorRev);
-+
-+#ifdef FM_HAS_TOTAL_DMAS
-+ p_Fm->p_FmStateStruct->maxNumOfOpenDmas = BMI_MAX_NUM_OF_DMAS;
-+#endif /* FM_HAS_TOTAL_DMAS */
-+#if (DPAA_VERSION < 11)
-+ p_Fm->p_FmDriverParam->dma_comm_qtsh_clr_emer = DEFAULT_dmaCommQLow;
-+ p_Fm->p_FmDriverParam->dma_comm_qtsh_asrt_emer = DEFAULT_dmaCommQHigh;
-+ p_Fm->p_FmDriverParam->dma_cam_num_of_entries = DEFAULT_dmaCamNumOfEntries;
-+ p_Fm->p_FmDriverParam->dma_read_buf_tsh_clr_emer = DEFAULT_dmaReadIntBufLow;
-+ p_Fm->p_FmDriverParam->dma_read_buf_tsh_asrt_emer = DEFAULT_dmaReadIntBufHigh;
-+ p_Fm->p_FmDriverParam->dma_write_buf_tsh_clr_emer = DEFAULT_dmaWriteIntBufLow;
-+ p_Fm->p_FmDriverParam->dma_write_buf_tsh_asrt_emer = DEFAULT_dmaWriteIntBufHigh;
-+ p_Fm->p_FmDriverParam->dma_axi_dbg_num_of_beats = DEFAULT_axiDbgNumOfBeats;
-+#endif /* (DPAA_VERSION < 11) */
-+#ifdef FM_NO_TNUM_AGING
-+ p_Fm->p_FmDriverParam->tnum_aging_period = 0;
-+#endif
-+ p_Fm->tnumAgingPeriod = p_Fm->p_FmDriverParam->tnum_aging_period;
-+
-+ return p_Fm;
-+}
-+
-+/**************************************************************************//**
-+ @Function FM_Init
-+
-+ @Description Initializes the FM module
-+
-+ @Param[in] h_Fm - FM module descriptor
-+
-+ @Return E_OK on success; Error code otherwise.
-+*//***************************************************************************/
-+t_Error FM_Init(t_Handle h_Fm)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+ struct fman_cfg *p_FmDriverParam = NULL;
-+ t_Error err = E_OK;
-+ int i;
-+ t_FmRevisionInfo revInfo;
-+ struct fman_rg fman_rg;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Fm, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Fm->p_FmDriverParam, E_INVALID_HANDLE);
-+
-+ fman_rg.bmi_rg = p_Fm->p_FmBmiRegs;
-+ fman_rg.qmi_rg = p_Fm->p_FmQmiRegs;
-+ fman_rg.fpm_rg = p_Fm->p_FmFpmRegs;
-+ fman_rg.dma_rg = p_Fm->p_FmDmaRegs;
-+
-+ p_Fm->p_FmStateStruct->count1MicroBit = FM_TIMESTAMP_1_USEC_BIT;
-+ p_Fm->p_FmDriverParam->num_of_fman_ctrl_evnt_regs = FM_NUM_OF_FMAN_CTRL_EVENT_REGS;
-+
-+ if (p_Fm->guestId != NCSW_MASTER_ID)
-+ return InitGuestMode(p_Fm);
-+
-+ /* if user didn't configured totalFifoSize - (totalFifoSize=0) we configure default
-+ * according to chip. otherwise, we use user's configuration.
-+ */
-+ if (p_Fm->p_FmStateStruct->totalFifoSize == 0)
-+ p_Fm->p_FmStateStruct->totalFifoSize = DEFAULT_totalFifoSize(p_Fm->p_FmStateStruct->revInfo.majorRev,
-+ p_Fm->p_FmStateStruct->revInfo.minorRev);
-+
-+ CHECK_INIT_PARAMETERS(p_Fm, CheckFmParameters);
-+
-+ p_FmDriverParam = p_Fm->p_FmDriverParam;
-+
-+ FM_GetRevision(p_Fm, &revInfo);
-+
-+ /* clear revision-dependent non existing exception */
-+#ifdef FM_NO_DISPATCH_RAM_ECC
-+ if ((revInfo.majorRev != 4) &&
-+ (revInfo.majorRev < 6))
-+ p_Fm->p_FmStateStruct->exceptions &= ~FM_EX_BMI_DISPATCH_RAM_ECC;
-+#endif /* FM_NO_DISPATCH_RAM_ECC */
-+
-+#ifdef FM_QMI_NO_ECC_EXCEPTIONS
-+ if (revInfo.majorRev == 4)
-+ p_Fm->p_FmStateStruct->exceptions &= ~(FM_EX_QMI_SINGLE_ECC | FM_EX_QMI_DOUBLE_ECC);
-+#endif /* FM_QMI_NO_ECC_EXCEPTIONS */
-+
-+#ifdef FM_QMI_NO_SINGLE_ECC_EXCEPTION
-+ if (revInfo.majorRev >= 6)
-+ p_Fm->p_FmStateStruct->exceptions &= ~FM_EX_QMI_SINGLE_ECC;
-+#endif /* FM_QMI_NO_SINGLE_ECC_EXCEPTION */
-+
-+ FmMuramClear(p_Fm->h_FmMuram);
-+
-+ /* clear CPG */
-+ IOMemSet32(UINT_TO_PTR(p_Fm->baseAddr + FM_MM_CGP), 0, FM_PORT_NUM_OF_CONGESTION_GRPS);
-+
-+ /* add to the default exceptions the user's definitions */
-+ p_Fm->p_FmStateStruct->exceptions |= p_Fm->userSetExceptions;
-+
-+ /* Reset the FM if required */
-+ if (p_Fm->resetOnInit)
-+ {
-+#ifdef FM_UCODE_NOT_RESET_ERRATA_BUGZILLA6173
-+ if ((err = FwNotResetErratumBugzilla6173WA(p_Fm)) != E_OK)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+#else /* not FM_UCODE_NOT_RESET_ERRATA_BUGZILLA6173 */
-+
-+ if (p_Fm->f_ResetOnInitOverride)
-+ {
-+ /* Perform user specific FMan reset */
-+ p_Fm->f_ResetOnInitOverride(h_Fm);
-+ }
-+ else
-+ {
-+ /* Perform FMan reset */
-+ FmReset(h_Fm);
-+ }
-+
-+ if (fman_is_qmi_halt_not_busy_state(p_Fm->p_FmQmiRegs))
-+ {
-+ fman_resume(p_Fm->p_FmFpmRegs);
-+ XX_UDelay(100);
-+ }
-+#endif /* not FM_UCODE_NOT_RESET_ERRATA_BUGZILLA6173 */
-+ }
-+
-+#ifdef FM_UCODE_NOT_RESET_ERRATA_BUGZILLA6173
-+ if (!p_Fm->resetOnInit) /* Skip operations done in errata workaround */
-+ {
-+#endif /* FM_UCODE_NOT_RESET_ERRATA_BUGZILLA6173 */
-+ /* Load FMan-Controller code to IRAM */
-+
-+ ClearIRam(p_Fm);
-+
-+ if (p_Fm->firmware.p_Code && (LoadFmanCtrlCode(p_Fm) != E_OK))
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, NO_MSG);
-+#ifdef FM_UCODE_NOT_RESET_ERRATA_BUGZILLA6173
-+ }
-+#endif /* FM_UCODE_NOT_RESET_ERRATA_BUGZILLA6173 */
-+
-+#ifdef FM_CAPWAP_SUPPORT
-+ /* save first 256 byte in MURAM */
-+ p_Fm->resAddr = PTR_TO_UINT(FM_MURAM_AllocMem(p_Fm->h_FmMuram, 256, 0));
-+ if (!p_Fm->resAddr)
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, ("MURAM alloc for reserved Area failed"));
-+
-+ WRITE_BLOCK(UINT_TO_PTR(p_Fm->resAddr), 0, 256);
-+#endif /* FM_CAPWAP_SUPPORT */
-+
-+#if (DPAA_VERSION >= 11)
-+ p_Fm->partVSPBase = AllocVSPsForPartition(h_Fm, p_Fm->partVSPBase, p_Fm->partNumOfVSPs, p_Fm->guestId);
-+ if (p_Fm->partVSPBase == (uint8_t)(ILLEGAL_BASE))
-+ DBG(WARNING, ("partition VSPs allocation is FAILED"));
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+ /* General FM driver initialization */
-+ p_Fm->fmMuramPhysBaseAddr =
-+ (uint64_t)(XX_VirtToPhys(UINT_TO_PTR(p_Fm->baseAddr + FM_MM_MURAM)));
-+
-+ for (i=0;i<e_FM_EV_DUMMY_LAST;i++)
-+ p_Fm->intrMng[i].f_Isr = UnimplementedIsr;
-+ for (i=0;i<FM_NUM_OF_FMAN_CTRL_EVENT_REGS;i++)
-+ p_Fm->fmanCtrlIntr[i].f_Isr = UnimplementedFmanCtrlIsr;
-+
-+ p_FmDriverParam->exceptions = p_Fm->p_FmStateStruct->exceptions;
-+
-+ /**********************/
-+ /* Init DMA Registers */
-+ /**********************/
-+ err = InitFmDma(p_Fm);
-+ if (err != E_OK)
-+ {
-+ FreeInitResources(p_Fm);
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+
-+ /**********************/
-+ /* Init FPM Registers */
-+ /**********************/
-+ err = InitFmFpm(p_Fm);
-+ if (err != E_OK)
-+ {
-+ FreeInitResources(p_Fm);
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+
-+ /* define common resources */
-+ /* allocate MURAM for FIFO according to total size */
-+ p_Fm->fifoBaseAddr = PTR_TO_UINT(FM_MURAM_AllocMem(p_Fm->h_FmMuram,
-+ p_Fm->p_FmStateStruct->totalFifoSize,
-+ BMI_FIFO_ALIGN));
-+ if (!p_Fm->fifoBaseAddr)
-+ {
-+ FreeInitResources(p_Fm);
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, ("MURAM alloc for BMI FIFO failed"));
-+ }
-+
-+ p_FmDriverParam->fifo_base_addr = (uint32_t)(XX_VirtToPhys(UINT_TO_PTR(p_Fm->fifoBaseAddr)) - p_Fm->fmMuramPhysBaseAddr);
-+ p_FmDriverParam->total_fifo_size = p_Fm->p_FmStateStruct->totalFifoSize;
-+ p_FmDriverParam->total_num_of_tasks = p_Fm->p_FmStateStruct->totalNumOfTasks;
-+ p_FmDriverParam->clk_freq = p_Fm->p_FmStateStruct->fmClkFreq;
-+
-+ /**********************/
-+ /* Init BMI Registers */
-+ /**********************/
-+ err = InitFmBmi(p_Fm);
-+ if (err != E_OK)
-+ {
-+ FreeInitResources(p_Fm);
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+
-+ /**********************/
-+ /* Init QMI Registers */
-+ /**********************/
-+ err = InitFmQmi(p_Fm);
-+ if (err != E_OK)
-+ {
-+ FreeInitResources(p_Fm);
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+
-+ /* build the FM master partition IPC address */
-+ if (Sprint (p_Fm->fmModuleName, "FM_%d_%d",p_Fm->p_FmStateStruct->fmId, NCSW_MASTER_ID) != 6)
-+ {
-+ FreeInitResources(p_Fm);
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("Sprint failed"));
-+ }
-+
-+ err = XX_IpcRegisterMsgHandler(p_Fm->fmModuleName, FmHandleIpcMsgCB, p_Fm, FM_IPC_MAX_REPLY_SIZE);
-+ if (err)
-+ {
-+ FreeInitResources(p_Fm);
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+ }
-+
-+ /* Register the FM interrupts handlers */
-+ if (p_Fm->p_FmStateStruct->irq != NO_IRQ)
-+ {
-+ XX_SetIntr(p_Fm->p_FmStateStruct->irq, FM_EventIsr, p_Fm);
-+ XX_EnableIntr(p_Fm->p_FmStateStruct->irq);
-+ }
-+
-+ if (p_Fm->p_FmStateStruct->errIrq != NO_IRQ)
-+ {
-+ XX_SetIntr(p_Fm->p_FmStateStruct->errIrq, (void (*) (t_Handle))FM_ErrorIsr, p_Fm);
-+ XX_EnableIntr(p_Fm->p_FmStateStruct->errIrq);
-+ }
-+
-+ err = (t_Error)fman_enable(&fman_rg , p_FmDriverParam);
-+ if (err != E_OK)
-+ return err; /* FIXME */
-+
-+ EnableTimeStamp(p_Fm);
-+
-+ if (p_Fm->firmware.p_Code)
-+ {
-+ XX_Free(p_Fm->firmware.p_Code);
-+ p_Fm->firmware.p_Code = NULL;
-+ }
-+
-+ XX_Free(p_Fm->p_FmDriverParam);
-+ p_Fm->p_FmDriverParam = NULL;
-+
-+ return E_OK;
-+}
-+
-+/**************************************************************************//**
-+ @Function FM_Free
-+
-+ @Description Frees all resources that were assigned to FM module.
-+
-+ Calling this routine invalidates the descriptor.
-+
-+ @Param[in] h_Fm - FM module descriptor
-+
-+ @Return E_OK on success; Error code otherwise.
-+*//***************************************************************************/
-+t_Error FM_Free(t_Handle h_Fm)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+ struct fman_rg fman_rg;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Fm, E_INVALID_HANDLE);
-+
-+ fman_rg.bmi_rg = p_Fm->p_FmBmiRegs;
-+ fman_rg.qmi_rg = p_Fm->p_FmQmiRegs;
-+ fman_rg.fpm_rg = p_Fm->p_FmFpmRegs;
-+ fman_rg.dma_rg = p_Fm->p_FmDmaRegs;
-+
-+ if (p_Fm->guestId != NCSW_MASTER_ID)
-+ {
-+#if (DPAA_VERSION >= 11)
-+ FreeVSPsForPartition(h_Fm, p_Fm->partVSPBase, p_Fm->partNumOfVSPs, p_Fm->guestId);
-+
-+ if (p_Fm->p_FmSp)
-+ {
-+ XX_Free(p_Fm->p_FmSp);
-+ p_Fm->p_FmSp = NULL;
-+ }
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+ if (p_Fm->fmModuleName[0] != 0)
-+ XX_IpcUnregisterMsgHandler(p_Fm->fmModuleName);
-+
-+ if (!p_Fm->recoveryMode)
-+ XX_Free(p_Fm->p_FmStateStruct);
-+
-+ XX_Free(p_Fm);
-+
-+ return E_OK;
-+ }
-+
-+ fman_free_resources(&fman_rg);
-+
-+ if ((p_Fm->guestId == NCSW_MASTER_ID) && (p_Fm->fmModuleName[0] != 0))
-+ XX_IpcUnregisterMsgHandler(p_Fm->fmModuleName);
-+
-+ if (p_Fm->p_FmStateStruct)
-+ {
-+ if (p_Fm->p_FmStateStruct->irq != NO_IRQ)
-+ {
-+ XX_DisableIntr(p_Fm->p_FmStateStruct->irq);
-+ XX_FreeIntr(p_Fm->p_FmStateStruct->irq);
-+ }
-+ if (p_Fm->p_FmStateStruct->errIrq != NO_IRQ)
-+ {
-+ XX_DisableIntr(p_Fm->p_FmStateStruct->errIrq);
-+ XX_FreeIntr(p_Fm->p_FmStateStruct->errIrq);
-+ }
-+ }
-+
-+#if (DPAA_VERSION >= 11)
-+ FreeVSPsForPartition(h_Fm, p_Fm->partVSPBase, p_Fm->partNumOfVSPs, p_Fm->guestId);
-+
-+ if (p_Fm->p_FmSp)
-+ {
-+ XX_Free(p_Fm->p_FmSp);
-+ p_Fm->p_FmSp = NULL;
-+ }
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+ if (p_Fm->h_Spinlock)
-+ XX_FreeSpinlock(p_Fm->h_Spinlock);
-+
-+ if (p_Fm->p_FmDriverParam)
-+ {
-+ if (p_Fm->firmware.p_Code)
-+ XX_Free(p_Fm->firmware.p_Code);
-+ XX_Free(p_Fm->p_FmDriverParam);
-+ p_Fm->p_FmDriverParam = NULL;
-+ }
-+
-+ FreeInitResources(p_Fm);
-+
-+ if (!p_Fm->recoveryMode && p_Fm->p_FmStateStruct)
-+ XX_Free(p_Fm->p_FmStateStruct);
-+
-+ XX_Free(p_Fm);
-+
-+ return E_OK;
-+}
-+
-+/*************************************************/
-+/* API Advanced Init unit functions */
-+/*************************************************/
-+
-+t_Error FM_ConfigResetOnInit(t_Handle h_Fm, bool enable)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Fm, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Fm->p_FmDriverParam, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR((p_Fm->guestId == NCSW_MASTER_ID), E_NOT_SUPPORTED);
-+
-+ p_Fm->resetOnInit = enable;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_ConfigResetOnInitOverrideCallback(t_Handle h_Fm, t_FmResetOnInitOverrideCallback *f_ResetOnInitOverride)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Fm, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Fm->p_FmDriverParam, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR((p_Fm->guestId == NCSW_MASTER_ID), E_NOT_SUPPORTED);
-+
-+ p_Fm->f_ResetOnInitOverride = f_ResetOnInitOverride;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_ConfigTotalFifoSize(t_Handle h_Fm, uint32_t totalFifoSize)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Fm, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Fm->p_FmDriverParam, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR((p_Fm->guestId == NCSW_MASTER_ID), E_NOT_SUPPORTED);
-+
-+ p_Fm->p_FmStateStruct->totalFifoSize = totalFifoSize;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_ConfigDmaCacheOverride(t_Handle h_Fm, e_FmDmaCacheOverride cacheOverride)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+ enum fman_dma_cache_override fsl_cache_override;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Fm, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Fm->p_FmDriverParam, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR((p_Fm->guestId == NCSW_MASTER_ID), E_NOT_SUPPORTED);
-+
-+ FMAN_CACHE_OVERRIDE_TRANS(fsl_cache_override, cacheOverride)
-+ p_Fm->p_FmDriverParam->dma_cache_override = fsl_cache_override;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_ConfigDmaAidOverride(t_Handle h_Fm, bool aidOverride)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Fm, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Fm->p_FmDriverParam, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR((p_Fm->guestId == NCSW_MASTER_ID), E_NOT_SUPPORTED);
-+
-+ p_Fm->p_FmDriverParam->dma_aid_override = aidOverride;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_ConfigDmaAidMode(t_Handle h_Fm, e_FmDmaAidMode aidMode)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+ enum fman_dma_aid_mode fsl_aid_mode;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Fm, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Fm->p_FmDriverParam, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR((p_Fm->guestId == NCSW_MASTER_ID), E_NOT_SUPPORTED);
-+
-+ FMAN_AID_MODE_TRANS(fsl_aid_mode, aidMode);
-+ p_Fm->p_FmDriverParam->dma_aid_mode = fsl_aid_mode;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_ConfigDmaAxiDbgNumOfBeats(t_Handle h_Fm, uint8_t axiDbgNumOfBeats)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Fm, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Fm->p_FmDriverParam, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR((p_Fm->guestId == NCSW_MASTER_ID), E_NOT_SUPPORTED);
-+
-+#if (DPAA_VERSION >= 11)
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, ("Not available for this FM revision!"));
-+#else
-+ p_Fm->p_FmDriverParam->dma_axi_dbg_num_of_beats = axiDbgNumOfBeats;
-+
-+ return E_OK;
-+#endif /* (DPAA_VERSION >= 11) */
-+}
-+
-+t_Error FM_ConfigDmaCamNumOfEntries(t_Handle h_Fm, uint8_t numOfEntries)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Fm, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Fm->p_FmDriverParam, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR((p_Fm->guestId == NCSW_MASTER_ID), E_NOT_SUPPORTED);
-+
-+ p_Fm->p_FmDriverParam->dma_cam_num_of_entries = numOfEntries;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_ConfigDmaDbgCounter(t_Handle h_Fm, e_FmDmaDbgCntMode fmDmaDbgCntMode)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+ enum fman_dma_dbg_cnt_mode fsl_dma_dbg_cnt;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Fm, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Fm->p_FmDriverParam, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR((p_Fm->guestId == NCSW_MASTER_ID), E_NOT_SUPPORTED);
-+
-+ FMAN_DMA_DBG_CNT_TRANS(fsl_dma_dbg_cnt, fmDmaDbgCntMode);
-+ p_Fm->p_FmDriverParam->dma_dbg_cnt_mode = fsl_dma_dbg_cnt;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_ConfigDmaStopOnBusErr(t_Handle h_Fm, bool stop)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Fm, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Fm->p_FmDriverParam, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR((p_Fm->guestId == NCSW_MASTER_ID), E_NOT_SUPPORTED);
-+
-+ p_Fm->p_FmDriverParam->dma_stop_on_bus_error = stop;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_ConfigDmaEmergency(t_Handle h_Fm, t_FmDmaEmergency *p_Emergency)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+ enum fman_dma_emergency_level fsl_dma_emer;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Fm, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Fm->p_FmDriverParam, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR((p_Fm->guestId == NCSW_MASTER_ID), E_NOT_SUPPORTED);
-+
-+ FMAN_DMA_EMER_TRANS(fsl_dma_emer, p_Emergency->emergencyLevel);
-+ p_Fm->p_FmDriverParam->dma_en_emergency = TRUE;
-+ p_Fm->p_FmDriverParam->dma_emergency_bus_select = (uint32_t)p_Emergency->emergencyBusSelect;
-+ p_Fm->p_FmDriverParam->dma_emergency_level = fsl_dma_emer;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_ConfigDmaEmergencySmoother(t_Handle h_Fm, uint32_t emergencyCnt)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Fm, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Fm->p_FmDriverParam, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR((p_Fm->guestId == NCSW_MASTER_ID), E_NOT_SUPPORTED);
-+
-+ p_Fm->p_FmDriverParam->dma_en_emergency_smoother = TRUE;
-+ p_Fm->p_FmDriverParam->dma_emergency_switch_counter = emergencyCnt;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_ConfigDmaErr(t_Handle h_Fm, e_FmDmaErr dmaErr)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+ enum fman_dma_err fsl_dma_err;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Fm, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Fm->p_FmDriverParam, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR((p_Fm->guestId == NCSW_MASTER_ID), E_NOT_SUPPORTED);
-+
-+ FMAN_DMA_ERR_TRANS(fsl_dma_err, dmaErr);
-+ p_Fm->p_FmDriverParam->dma_err = fsl_dma_err;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_ConfigCatastrophicErr(t_Handle h_Fm, e_FmCatastrophicErr catastrophicErr)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+ enum fman_catastrophic_err fsl_catastrophic_err;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Fm, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Fm->p_FmDriverParam, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR((p_Fm->guestId == NCSW_MASTER_ID), E_NOT_SUPPORTED);
-+
-+ FMAN_CATASTROPHIC_ERR_TRANS(fsl_catastrophic_err, catastrophicErr);
-+ p_Fm->p_FmDriverParam->catastrophic_err = fsl_catastrophic_err;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_ConfigEnableMuramTestMode(t_Handle h_Fm)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Fm, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Fm->p_FmDriverParam, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR((p_Fm->guestId == NCSW_MASTER_ID), E_NOT_SUPPORTED);
-+
-+ if (p_Fm->p_FmStateStruct->revInfo.majorRev >= 6)
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, ("Not available for this FM revision!"));
-+
-+ p_Fm->p_FmDriverParam->en_muram_test_mode = TRUE;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_ConfigEnableIramTestMode(t_Handle h_Fm)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Fm, E_INVALID_HANDLE );
-+ SANITY_CHECK_RETURN_ERROR(p_Fm->p_FmDriverParam, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR((p_Fm->guestId == NCSW_MASTER_ID), E_NOT_SUPPORTED);
-+
-+ if (p_Fm->p_FmStateStruct->revInfo.majorRev >= 6)
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, ("Not available for this FM revision!"));
-+
-+ p_Fm->p_FmDriverParam->en_iram_test_mode = TRUE;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_ConfigHaltOnExternalActivation(t_Handle h_Fm, bool enable)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Fm, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Fm->p_FmDriverParam, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR((p_Fm->guestId == NCSW_MASTER_ID), E_NOT_SUPPORTED);
-+
-+ p_Fm->p_FmDriverParam->halt_on_external_activ = enable;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_ConfigHaltOnUnrecoverableEccError(t_Handle h_Fm, bool enable)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Fm, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Fm->p_FmDriverParam, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR((p_Fm->guestId == NCSW_MASTER_ID), E_NOT_SUPPORTED);
-+
-+ if (p_Fm->p_FmStateStruct->revInfo.majorRev >= 6)
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, ("Not available for this FM revision!"));
-+
-+ p_Fm->p_FmDriverParam->halt_on_unrecov_ecc_err = enable;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_ConfigException(t_Handle h_Fm, e_FmExceptions exception, bool enable)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+ uint32_t bitMask = 0;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Fm, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Fm->p_FmDriverParam, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR((p_Fm->guestId == NCSW_MASTER_ID), E_NOT_SUPPORTED);
-+
-+ GET_EXCEPTION_FLAG(bitMask, exception);
-+ if (bitMask)
-+ {
-+ if (enable)
-+ p_Fm->userSetExceptions |= bitMask;
-+ else
-+ p_Fm->p_FmStateStruct->exceptions &= ~bitMask;
-+ }
-+ else
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("Undefined exception"));
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_ConfigExternalEccRamsEnable(t_Handle h_Fm, bool enable)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Fm, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Fm->p_FmDriverParam, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR((p_Fm->guestId == NCSW_MASTER_ID), E_NOT_SUPPORTED);
-+
-+ p_Fm->p_FmDriverParam->external_ecc_rams_enable = enable;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_ConfigTnumAgingPeriod(t_Handle h_Fm, uint16_t tnumAgingPeriod)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Fm, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Fm->p_FmDriverParam, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR((p_Fm->guestId == NCSW_MASTER_ID), E_NOT_SUPPORTED);
-+
-+ p_Fm->p_FmDriverParam->tnum_aging_period = tnumAgingPeriod;
-+ p_Fm->tnumAgingPeriod = p_Fm->p_FmDriverParam->tnum_aging_period;
-+
-+ return E_OK;
-+}
-+
-+/****************************************************/
-+/* Hidden-DEBUG Only API */
-+/****************************************************/
-+
-+t_Error FM_ConfigThresholds(t_Handle h_Fm, t_FmThresholds *p_FmThresholds)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Fm, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Fm->p_FmDriverParam, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR((p_Fm->guestId == NCSW_MASTER_ID), E_NOT_SUPPORTED);
-+
-+ p_Fm->p_FmDriverParam->disp_limit_tsh = p_FmThresholds->dispLimit;
-+ p_Fm->p_FmDriverParam->prs_disp_tsh = p_FmThresholds->prsDispTh;
-+ p_Fm->p_FmDriverParam->plcr_disp_tsh = p_FmThresholds->plcrDispTh;
-+ p_Fm->p_FmDriverParam->kg_disp_tsh = p_FmThresholds->kgDispTh;
-+ p_Fm->p_FmDriverParam->bmi_disp_tsh = p_FmThresholds->bmiDispTh;
-+ p_Fm->p_FmDriverParam->qmi_enq_disp_tsh = p_FmThresholds->qmiEnqDispTh;
-+ p_Fm->p_FmDriverParam->qmi_deq_disp_tsh = p_FmThresholds->qmiDeqDispTh;
-+ p_Fm->p_FmDriverParam->fm_ctl1_disp_tsh = p_FmThresholds->fmCtl1DispTh;
-+ p_Fm->p_FmDriverParam->fm_ctl2_disp_tsh = p_FmThresholds->fmCtl2DispTh;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_ConfigDmaSosEmergencyThreshold(t_Handle h_Fm, uint32_t dmaSosEmergency)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Fm, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Fm->p_FmDriverParam, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR((p_Fm->guestId == NCSW_MASTER_ID), E_NOT_SUPPORTED);
-+
-+ p_Fm->p_FmDriverParam->dma_sos_emergency = dmaSosEmergency;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_ConfigDmaWriteBufThresholds(t_Handle h_Fm, t_FmDmaThresholds *p_FmDmaThresholds)
-+
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Fm, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Fm->p_FmDriverParam, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR((p_Fm->guestId == NCSW_MASTER_ID), E_NOT_SUPPORTED);
-+
-+#if (DPAA_VERSION >= 11)
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, ("Not available for this FM revision!"));
-+#else
-+ p_Fm->p_FmDriverParam->dma_write_buf_tsh_asrt_emer = p_FmDmaThresholds->assertEmergency;
-+ p_Fm->p_FmDriverParam->dma_write_buf_tsh_clr_emer = p_FmDmaThresholds->clearEmergency;
-+
-+ return E_OK;
-+#endif
-+}
-+
-+t_Error FM_ConfigDmaCommQThresholds(t_Handle h_Fm, t_FmDmaThresholds *p_FmDmaThresholds)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Fm, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Fm->p_FmDriverParam, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR((p_Fm->guestId == NCSW_MASTER_ID), E_NOT_SUPPORTED);
-+
-+ p_Fm->p_FmDriverParam->dma_comm_qtsh_asrt_emer = p_FmDmaThresholds->assertEmergency;
-+ p_Fm->p_FmDriverParam->dma_comm_qtsh_clr_emer = p_FmDmaThresholds->clearEmergency;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_ConfigDmaReadBufThresholds(t_Handle h_Fm, t_FmDmaThresholds *p_FmDmaThresholds)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Fm, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Fm->p_FmDriverParam, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR((p_Fm->guestId == NCSW_MASTER_ID), E_NOT_SUPPORTED);
-+
-+#if (DPAA_VERSION >= 11)
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, ("Not available for this FM revision!"));
-+#else
-+ p_Fm->p_FmDriverParam->dma_read_buf_tsh_clr_emer = p_FmDmaThresholds->clearEmergency;
-+ p_Fm->p_FmDriverParam->dma_read_buf_tsh_asrt_emer = p_FmDmaThresholds->assertEmergency;
-+
-+ return E_OK;
-+#endif
-+}
-+
-+t_Error FM_ConfigDmaWatchdog(t_Handle h_Fm, uint32_t watchdogValue)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Fm, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Fm->p_FmDriverParam, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR((p_Fm->guestId == NCSW_MASTER_ID), E_NOT_SUPPORTED);
-+
-+ p_Fm->p_FmDriverParam->dma_watchdog = watchdogValue;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_ConfigEnableCounters(t_Handle h_Fm)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Fm, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_Fm->p_FmDriverParam, E_INVALID_HANDLE);
-+UNUSED(p_Fm);
-+
-+ return E_OK;
-+}
-+
-+t_Error FmGetSetParams(t_Handle h_Fm, t_FmGetSetParams *p_Params)
-+{
-+ t_Fm* p_Fm = (t_Fm*)h_Fm;
-+ if (p_Params->setParams.type & UPDATE_FM_CLD)
-+ {
-+ WRITE_UINT32(p_Fm->p_FmFpmRegs->fm_cld, GET_UINT32(
-+ p_Fm->p_FmFpmRegs->fm_cld) | 0x00000800);
-+ }
-+ if (p_Params->setParams.type & CLEAR_IRAM_READY)
-+ {
-+ t_FMIramRegs *p_Iram = (t_FMIramRegs *)UINT_TO_PTR(p_Fm->baseAddr + FM_MM_IMEM);
-+ WRITE_UINT32(p_Iram->iready,GET_UINT32(p_Iram->iready) & ~IRAM_READY);
-+ }
-+ if (p_Params->setParams.type & UPDATE_FPM_EXTC)
-+ WRITE_UINT32(p_Fm->p_FmFpmRegs->fmfp_extc,0x80000000);
-+ if (p_Params->setParams.type & UPDATE_FPM_EXTC_CLEAR)
-+ WRITE_UINT32(p_Fm->p_FmFpmRegs->fmfp_extc,0x00800000);
-+ if (p_Params->setParams.type & UPDATE_FPM_BRKC_SLP)
-+ {
-+ if (p_Params->setParams.sleep)
-+ WRITE_UINT32(p_Fm->p_FmFpmRegs->fmfp_brkc, GET_UINT32(
-+ p_Fm->p_FmFpmRegs->fmfp_brkc) | FPM_BRKC_SLP);
-+ else
-+ WRITE_UINT32(p_Fm->p_FmFpmRegs->fmfp_brkc, GET_UINT32(
-+ p_Fm->p_FmFpmRegs->fmfp_brkc) & ~FPM_BRKC_SLP);
-+ }
-+ if (p_Params->getParams.type & GET_FM_CLD)
-+ p_Params->getParams.fm_cld = GET_UINT32(p_Fm->p_FmFpmRegs->fm_cld);
-+ if (p_Params->getParams.type & GET_FMQM_GS)
-+ p_Params->getParams.fmqm_gs = GET_UINT32(p_Fm->p_FmQmiRegs->fmqm_gs);
-+ if (p_Params->getParams.type & GET_FM_NPI)
-+ p_Params->getParams.fm_npi = GET_UINT32(p_Fm->p_FmFpmRegs->fm_npi);
-+ if (p_Params->getParams.type & GET_FMFP_EXTC)
-+ p_Params->getParams.fmfp_extc = GET_UINT32(p_Fm->p_FmFpmRegs->fmfp_extc);
-+ return E_OK;
-+}
-+
-+
-+/****************************************************/
-+/* API Run-time Control uint functions */
-+/****************************************************/
-+void FM_EventIsr(t_Handle h_Fm)
-+{
-+#define FM_M_CALL_1G_MAC_ISR(_id) \
-+ { \
-+ if (p_Fm->guestId != p_Fm->intrMng[(e_FmInterModuleEvent)(e_FM_EV_1G_MAC0+_id)].guestId) \
-+ SendIpcIsr(p_Fm, (e_FmInterModuleEvent)(e_FM_EV_1G_MAC0+_id), pending); \
-+ else \
-+ p_Fm->intrMng[(e_FmInterModuleEvent)(e_FM_EV_1G_MAC0+_id)].f_Isr(p_Fm->intrMng[(e_FmInterModuleEvent)(e_FM_EV_1G_MAC0+_id)].h_SrcHandle);\
-+ }
-+#define FM_M_CALL_10G_MAC_ISR(_id) \
-+ { \
-+ if (p_Fm->guestId != p_Fm->intrMng[(e_FmInterModuleEvent)(e_FM_EV_10G_MAC0+_id)].guestId) \
-+ SendIpcIsr(p_Fm, (e_FmInterModuleEvent)(e_FM_EV_10G_MAC0+_id), pending); \
-+ else \
-+ p_Fm->intrMng[(e_FmInterModuleEvent)(e_FM_EV_10G_MAC0+_id)].f_Isr(p_Fm->intrMng[(e_FmInterModuleEvent)(e_FM_EV_10G_MAC0+_id)].h_SrcHandle);\
-+ }
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+ uint32_t pending, event;
-+ struct fman_fpm_regs *fpm_rg;
-+
-+ SANITY_CHECK_RETURN(p_Fm, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN(!p_Fm->p_FmDriverParam, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN((p_Fm->guestId == NCSW_MASTER_ID), E_NOT_SUPPORTED);
-+
-+ fpm_rg = p_Fm->p_FmFpmRegs;
-+
-+ /* normal interrupts */
-+ pending = fman_get_normal_pending(fpm_rg);
-+ if (!pending)
-+ return;
-+ if (pending & INTR_EN_WAKEUP) // this is a wake up from sleep interrupt
-+ {
-+ t_FmGetSetParams fmGetSetParams;
-+ memset(&fmGetSetParams, 0, sizeof (t_FmGetSetParams));
-+ fmGetSetParams.setParams.type = UPDATE_FPM_BRKC_SLP;
-+ fmGetSetParams.setParams.sleep = 0;
-+ FmGetSetParams(h_Fm, &fmGetSetParams);
-+ }
-+ if (pending & INTR_EN_QMI)
-+ QmiEvent(p_Fm);
-+ if (pending & INTR_EN_PRS)
-+ p_Fm->intrMng[e_FM_EV_PRS].f_Isr(p_Fm->intrMng[e_FM_EV_PRS].h_SrcHandle);
-+ if (pending & INTR_EN_PLCR)
-+ p_Fm->intrMng[e_FM_EV_PLCR].f_Isr(p_Fm->intrMng[e_FM_EV_PLCR].h_SrcHandle);
-+ if (pending & INTR_EN_TMR)
-+ p_Fm->intrMng[e_FM_EV_TMR].f_Isr(p_Fm->intrMng[e_FM_EV_TMR].h_SrcHandle);
-+
-+ /* MAC events may belong to different partitions */
-+ if (pending & INTR_EN_1G_MAC0)
-+ FM_M_CALL_1G_MAC_ISR(0);
-+ if (pending & INTR_EN_1G_MAC1)
-+ FM_M_CALL_1G_MAC_ISR(1);
-+ if (pending & INTR_EN_1G_MAC2)
-+ FM_M_CALL_1G_MAC_ISR(2);
-+ if (pending & INTR_EN_1G_MAC3)
-+ FM_M_CALL_1G_MAC_ISR(3);
-+ if (pending & INTR_EN_1G_MAC4)
-+ FM_M_CALL_1G_MAC_ISR(4);
-+ if (pending & INTR_EN_1G_MAC5)
-+ FM_M_CALL_1G_MAC_ISR(5);
-+ if (pending & INTR_EN_1G_MAC6)
-+ FM_M_CALL_1G_MAC_ISR(6);
-+ if (pending & INTR_EN_1G_MAC7)
-+ FM_M_CALL_1G_MAC_ISR(7);
-+ if (pending & INTR_EN_10G_MAC0)
-+ FM_M_CALL_10G_MAC_ISR(0);
-+ if (pending & INTR_EN_10G_MAC1)
-+ FM_M_CALL_10G_MAC_ISR(1);
-+
-+ /* IM port events may belong to different partitions */
-+ if (pending & INTR_EN_REV0)
-+ {
-+ event = fman_get_controller_event(fpm_rg, 0);
-+ if (p_Fm->guestId != p_Fm->intrMng[e_FM_EV_FMAN_CTRL_0].guestId)
-+ /*TODO IPC ISR For Fman Ctrl */
-+ ASSERT_COND(0);
-+ /* SendIpcIsr(p_Fm, e_FM_EV_FMAN_CTRL_0, pending); */
-+ else
-+ p_Fm->fmanCtrlIntr[0].f_Isr(p_Fm->fmanCtrlIntr[0].h_SrcHandle, event);
-+
-+ }
-+ if (pending & INTR_EN_REV1)
-+ {
-+ event = fman_get_controller_event(fpm_rg, 1);
-+ if (p_Fm->guestId != p_Fm->intrMng[e_FM_EV_FMAN_CTRL_1].guestId)
-+ /*TODO IPC ISR For Fman Ctrl */
-+ ASSERT_COND(0);
-+ /* SendIpcIsr(p_Fm, e_FM_EV_FMAN_CTRL_1, pending); */
-+ else
-+ p_Fm->fmanCtrlIntr[1].f_Isr(p_Fm->fmanCtrlIntr[1].h_SrcHandle, event);
-+ }
-+ if (pending & INTR_EN_REV2)
-+ {
-+ event = fman_get_controller_event(fpm_rg, 2);
-+ if (p_Fm->guestId != p_Fm->intrMng[e_FM_EV_FMAN_CTRL_2].guestId)
-+ /*TODO IPC ISR For Fman Ctrl */
-+ ASSERT_COND(0);
-+ /* SendIpcIsr(p_Fm, e_FM_EV_FMAN_CTRL_2, pending); */
-+ else
-+ p_Fm->fmanCtrlIntr[2].f_Isr(p_Fm->fmanCtrlIntr[2].h_SrcHandle, event);
-+ }
-+ if (pending & INTR_EN_REV3)
-+ {
-+ event = fman_get_controller_event(fpm_rg, 3);
-+ if (p_Fm->guestId != p_Fm->intrMng[e_FM_EV_FMAN_CTRL_3].guestId)
-+ /*TODO IPC ISR For Fman Ctrl */
-+ ASSERT_COND(0);
-+ /* SendIpcIsr(p_Fm, e_FM_EV_FMAN_CTRL_2, pendin3); */
-+ else
-+ p_Fm->fmanCtrlIntr[3].f_Isr(p_Fm->fmanCtrlIntr[3].h_SrcHandle, event);
-+ }
-+#ifdef FM_MACSEC_SUPPORT
-+ if (pending & INTR_EN_MACSEC_MAC0)
-+ {
-+ if (p_Fm->guestId != p_Fm->intrMng[e_FM_EV_MACSEC_MAC0].guestId)
-+ SendIpcIsr(p_Fm, e_FM_EV_MACSEC_MAC0, pending);
-+ else
-+ p_Fm->intrMng[e_FM_EV_MACSEC_MAC0].f_Isr(p_Fm->intrMng[e_FM_EV_MACSEC_MAC0].h_SrcHandle);
-+ }
-+#endif /* FM_MACSEC_SUPPORT */
-+}
-+
-+t_Error FM_ErrorIsr(t_Handle h_Fm)
-+{
-+#define FM_M_CALL_1G_MAC_ERR_ISR(_id) \
-+ { \
-+ if (p_Fm->guestId != p_Fm->intrMng[(e_FmInterModuleEvent)(e_FM_EV_ERR_1G_MAC0+_id)].guestId) \
-+ SendIpcIsr(p_Fm, (e_FmInterModuleEvent)(e_FM_EV_ERR_1G_MAC0+_id), pending); \
-+ else \
-+ p_Fm->intrMng[(e_FmInterModuleEvent)(e_FM_EV_ERR_1G_MAC0+_id)].f_Isr(p_Fm->intrMng[(e_FmInterModuleEvent)(e_FM_EV_ERR_1G_MAC0+_id)].h_SrcHandle);\
-+ }
-+#define FM_M_CALL_10G_MAC_ERR_ISR(_id) \
-+ { \
-+ if (p_Fm->guestId != p_Fm->intrMng[(e_FmInterModuleEvent)(e_FM_EV_ERR_10G_MAC0+_id)].guestId) \
-+ SendIpcIsr(p_Fm, (e_FmInterModuleEvent)(e_FM_EV_ERR_10G_MAC0+_id), pending); \
-+ else \
-+ p_Fm->intrMng[(e_FmInterModuleEvent)(e_FM_EV_ERR_10G_MAC0+_id)].f_Isr(p_Fm->intrMng[(e_FmInterModuleEvent)(e_FM_EV_ERR_10G_MAC0+_id)].h_SrcHandle);\
-+ }
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+ uint32_t pending;
-+ struct fman_fpm_regs *fpm_rg;
-+
-+ SANITY_CHECK_RETURN_ERROR(h_Fm, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Fm->p_FmDriverParam, E_INVALID_STATE);
-+ SANITY_CHECK_RETURN_ERROR((p_Fm->guestId == NCSW_MASTER_ID), E_NOT_SUPPORTED);
-+
-+ fpm_rg = p_Fm->p_FmFpmRegs;
-+
-+ /* error interrupts */
-+ pending = fman_get_fpm_error_interrupts(fpm_rg);
-+ if (!pending)
-+ return ERROR_CODE(E_EMPTY);
-+
-+ if (pending & ERR_INTR_EN_BMI)
-+ BmiErrEvent(p_Fm);
-+ if (pending & ERR_INTR_EN_QMI)
-+ QmiErrEvent(p_Fm);
-+ if (pending & ERR_INTR_EN_FPM)
-+ FpmErrEvent(p_Fm);
-+ if (pending & ERR_INTR_EN_DMA)
-+ DmaErrEvent(p_Fm);
-+ if (pending & ERR_INTR_EN_IRAM)
-+ IramErrIntr(p_Fm);
-+ if (pending & ERR_INTR_EN_MURAM)
-+ MuramErrIntr(p_Fm);
-+ if (pending & ERR_INTR_EN_PRS)
-+ p_Fm->intrMng[e_FM_EV_ERR_PRS].f_Isr(p_Fm->intrMng[e_FM_EV_ERR_PRS].h_SrcHandle);
-+ if (pending & ERR_INTR_EN_PLCR)
-+ p_Fm->intrMng[e_FM_EV_ERR_PLCR].f_Isr(p_Fm->intrMng[e_FM_EV_ERR_PLCR].h_SrcHandle);
-+ if (pending & ERR_INTR_EN_KG)
-+ p_Fm->intrMng[e_FM_EV_ERR_KG].f_Isr(p_Fm->intrMng[e_FM_EV_ERR_KG].h_SrcHandle);
-+
-+ /* MAC events may belong to different partitions */
-+ if (pending & ERR_INTR_EN_1G_MAC0)
-+ FM_M_CALL_1G_MAC_ERR_ISR(0);
-+ if (pending & ERR_INTR_EN_1G_MAC1)
-+ FM_M_CALL_1G_MAC_ERR_ISR(1);
-+ if (pending & ERR_INTR_EN_1G_MAC2)
-+ FM_M_CALL_1G_MAC_ERR_ISR(2);
-+ if (pending & ERR_INTR_EN_1G_MAC3)
-+ FM_M_CALL_1G_MAC_ERR_ISR(3);
-+ if (pending & ERR_INTR_EN_1G_MAC4)
-+ FM_M_CALL_1G_MAC_ERR_ISR(4);
-+ if (pending & ERR_INTR_EN_1G_MAC5)
-+ FM_M_CALL_1G_MAC_ERR_ISR(5);
-+ if (pending & ERR_INTR_EN_1G_MAC6)
-+ FM_M_CALL_1G_MAC_ERR_ISR(6);
-+ if (pending & ERR_INTR_EN_1G_MAC7)
-+ FM_M_CALL_1G_MAC_ERR_ISR(7);
-+ if (pending & ERR_INTR_EN_10G_MAC0)
-+ FM_M_CALL_10G_MAC_ERR_ISR(0);
-+ if (pending & ERR_INTR_EN_10G_MAC1)
-+ FM_M_CALL_10G_MAC_ERR_ISR(1);
-+
-+#ifdef FM_MACSEC_SUPPORT
-+ if (pending & ERR_INTR_EN_MACSEC_MAC0)
-+ {
-+ if (p_Fm->guestId != p_Fm->intrMng[e_FM_EV_ERR_MACSEC_MAC0].guestId)
-+ SendIpcIsr(p_Fm, e_FM_EV_ERR_MACSEC_MAC0, pending);
-+ else
-+ p_Fm->intrMng[e_FM_EV_ERR_MACSEC_MAC0].f_Isr(p_Fm->intrMng[e_FM_EV_ERR_MACSEC_MAC0].h_SrcHandle);
-+ }
-+#endif /* FM_MACSEC_SUPPORT */
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_SetPortsBandwidth(t_Handle h_Fm, t_FmPortsBandwidthParams *p_PortsBandwidth)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+ int i;
-+ uint8_t sum;
-+ uint8_t hardwarePortId;
-+ uint8_t weights[64];
-+ uint8_t weight, maxPercent = 0;
-+ struct fman_bmi_regs *bmi_rg;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Fm, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Fm->p_FmDriverParam, E_INVALID_STATE);
-+ SANITY_CHECK_RETURN_ERROR((p_Fm->guestId == NCSW_MASTER_ID), E_NOT_SUPPORTED);
-+
-+ bmi_rg = p_Fm->p_FmBmiRegs;
-+
-+ memset(weights, 0, (sizeof(uint8_t) * 64));
-+
-+ /* check that all ports add up to 100% */
-+ sum = 0;
-+ for (i=0; i < p_PortsBandwidth->numOfPorts; i++)
-+ sum +=p_PortsBandwidth->portsBandwidths[i].bandwidth;
-+ if (sum != 100)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("Sum of ports bandwidth differ from 100%"));
-+
-+ /* find highest percent */
-+ for (i=0; i < p_PortsBandwidth->numOfPorts; i++)
-+ {
-+ if (p_PortsBandwidth->portsBandwidths[i].bandwidth > maxPercent)
-+ maxPercent = p_PortsBandwidth->portsBandwidths[i].bandwidth;
-+ }
-+
-+ ASSERT_COND(maxPercent > 0); /* guaranteed by sum = 100 */
-+
-+ /* calculate weight for each port */
-+ for (i=0; i < p_PortsBandwidth->numOfPorts; i++)
-+ {
-+ weight = (uint8_t)((p_PortsBandwidth->portsBandwidths[i].bandwidth * PORT_MAX_WEIGHT ) / maxPercent);
-+ /* we want even division between 1-to-PORT_MAX_WEIGHT. so if exact division
-+ is not reached, we round up so that:
-+ 0 until maxPercent/PORT_MAX_WEIGHT get "1"
-+ maxPercent/PORT_MAX_WEIGHT+1 until (maxPercent/PORT_MAX_WEIGHT)*2 get "2"
-+ ...
-+ maxPercent - maxPercent/PORT_MAX_WEIGHT until maxPercent get "PORT_MAX_WEIGHT: */
-+ if ((uint8_t)((p_PortsBandwidth->portsBandwidths[i].bandwidth * PORT_MAX_WEIGHT ) % maxPercent))
-+ weight++;
-+
-+ /* find the location of this port within the register */
-+ hardwarePortId =
-+ SwPortIdToHwPortId(p_PortsBandwidth->portsBandwidths[i].type,
-+ p_PortsBandwidth->portsBandwidths[i].relativePortId,
-+ p_Fm->p_FmStateStruct->revInfo.majorRev,
-+ p_Fm->p_FmStateStruct->revInfo.minorRev);
-+
-+ ASSERT_COND(IN_RANGE(1, hardwarePortId, 63));
-+ weights[hardwarePortId] = weight;
-+ }
-+
-+ fman_set_ports_bandwidth(bmi_rg, weights);
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_EnableRamsEcc(t_Handle h_Fm)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+ struct fman_fpm_regs *fpm_rg;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Fm, E_INVALID_HANDLE);
-+
-+ fpm_rg = p_Fm->p_FmFpmRegs;
-+
-+ if (p_Fm->guestId != NCSW_MASTER_ID)
-+ {
-+ t_FmIpcMsg msg;
-+ t_Error err;
-+
-+ memset(&msg, 0, sizeof(msg));
-+ msg.msgId = FM_ENABLE_RAM_ECC;
-+ err = XX_IpcSendMessage(p_Fm->h_IpcSessions[0],
-+ (uint8_t*)&msg,
-+ sizeof(msg.msgId),
-+ NULL,
-+ NULL,
-+ NULL,
-+ NULL);
-+ if (err != E_OK)
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+ return E_OK;
-+ }
-+
-+ if (!p_Fm->p_FmStateStruct->internalCall)
-+ p_Fm->p_FmStateStruct->explicitEnable = TRUE;
-+ p_Fm->p_FmStateStruct->internalCall = FALSE;
-+
-+ if (p_Fm->p_FmStateStruct->ramsEccEnable)
-+ return E_OK;
-+ else
-+ {
-+ fman_enable_rams_ecc(fpm_rg);
-+ p_Fm->p_FmStateStruct->ramsEccEnable = TRUE;
-+ }
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_DisableRamsEcc(t_Handle h_Fm)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+ bool explicitDisable = FALSE;
-+ struct fman_fpm_regs *fpm_rg;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Fm, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Fm->p_FmDriverParam, E_INVALID_HANDLE);
-+
-+ fpm_rg = p_Fm->p_FmFpmRegs;
-+
-+ if (p_Fm->guestId != NCSW_MASTER_ID)
-+ {
-+ t_Error err;
-+ t_FmIpcMsg msg;
-+
-+ memset(&msg, 0, sizeof(msg));
-+ msg.msgId = FM_DISABLE_RAM_ECC;
-+ if ((err = XX_IpcSendMessage(p_Fm->h_IpcSessions[0],
-+ (uint8_t*)&msg,
-+ sizeof(msg.msgId),
-+ NULL,
-+ NULL,
-+ NULL,
-+ NULL)) != E_OK)
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+ return E_OK;
-+ }
-+
-+ if (!p_Fm->p_FmStateStruct->internalCall)
-+ explicitDisable = TRUE;
-+ p_Fm->p_FmStateStruct->internalCall = FALSE;
-+
-+ /* if rams are already disabled, or if rams were explicitly enabled and are
-+ currently called indirectly (not explicitly), ignore this call. */
-+ if (!p_Fm->p_FmStateStruct->ramsEccEnable ||
-+ (p_Fm->p_FmStateStruct->explicitEnable && !explicitDisable))
-+ return E_OK;
-+ else
-+ {
-+ if (p_Fm->p_FmStateStruct->explicitEnable)
-+ /* This is the case were both explicit are TRUE.
-+ Turn off this flag for cases were following ramsEnable
-+ routines are called */
-+ p_Fm->p_FmStateStruct->explicitEnable = FALSE;
-+
-+ fman_enable_rams_ecc(fpm_rg);
-+ p_Fm->p_FmStateStruct->ramsEccEnable = FALSE;
-+ }
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_SetException(t_Handle h_Fm, e_FmExceptions exception, bool enable)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+ uint32_t bitMask = 0;
-+ enum fman_exceptions fslException;
-+ struct fman_rg fman_rg;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Fm, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Fm->p_FmDriverParam, E_INVALID_STATE);
-+
-+ fman_rg.bmi_rg = p_Fm->p_FmBmiRegs;
-+ fman_rg.qmi_rg = p_Fm->p_FmQmiRegs;
-+ fman_rg.fpm_rg = p_Fm->p_FmFpmRegs;
-+ fman_rg.dma_rg = p_Fm->p_FmDmaRegs;
-+
-+ GET_EXCEPTION_FLAG(bitMask, exception);
-+ if (bitMask)
-+ {
-+ if (enable)
-+ p_Fm->p_FmStateStruct->exceptions |= bitMask;
-+ else
-+ p_Fm->p_FmStateStruct->exceptions &= ~bitMask;
-+
-+ fslException = FmanExceptionTrans(exception);
-+
-+ return (t_Error)fman_set_exception(&fman_rg,
-+ fslException,
-+ enable);
-+ }
-+ else
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("Undefined exception"));
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_GetRevision(t_Handle h_Fm, t_FmRevisionInfo *p_FmRevisionInfo)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+
-+ p_FmRevisionInfo->majorRev = p_Fm->p_FmStateStruct->revInfo.majorRev;
-+ p_FmRevisionInfo->minorRev = p_Fm->p_FmStateStruct->revInfo.minorRev;
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_GetFmanCtrlCodeRevision(t_Handle h_Fm, t_FmCtrlCodeRevisionInfo *p_RevisionInfo)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+ t_FMIramRegs *p_Iram;
-+ uint32_t revInfo;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Fm, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_RevisionInfo, E_NULL_POINTER);
-+
-+ if ((p_Fm->guestId != NCSW_MASTER_ID) &&
-+ p_Fm->h_IpcSessions[0])
-+ {
-+ t_Error err;
-+ t_FmIpcMsg msg;
-+ t_FmIpcReply reply;
-+ uint32_t replyLength;
-+ t_FmIpcFmanCtrlCodeRevisionInfo ipcRevInfo;
-+
-+ memset(&msg, 0, sizeof(msg));
-+ memset(&reply, 0, sizeof(reply));
-+ msg.msgId = FM_GET_FMAN_CTRL_CODE_REV;
-+ replyLength = sizeof(uint32_t) + sizeof(t_FmCtrlCodeRevisionInfo);
-+ if ((err = XX_IpcSendMessage(p_Fm->h_IpcSessions[0],
-+ (uint8_t*)&msg,
-+ sizeof(msg.msgId),
-+ (uint8_t*)&reply,
-+ &replyLength,
-+ NULL,
-+ NULL)) != E_OK)
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+ if (replyLength != (sizeof(uint32_t) + sizeof(t_FmCtrlCodeRevisionInfo)))
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("IPC reply length mismatch"));
-+ memcpy((uint8_t*)&ipcRevInfo, reply.replyBody, sizeof(t_FmCtrlCodeRevisionInfo));
-+ p_RevisionInfo->packageRev = ipcRevInfo.packageRev;
-+ p_RevisionInfo->majorRev = ipcRevInfo.majorRev;
-+ p_RevisionInfo->minorRev = ipcRevInfo.minorRev;
-+ return (t_Error)(reply.error);
-+ }
-+ else if (p_Fm->guestId != NCSW_MASTER_ID)
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED,
-+ ("running in guest-mode without IPC!"));
-+
-+ p_Iram = (t_FMIramRegs *)UINT_TO_PTR(p_Fm->baseAddr + FM_MM_IMEM);
-+ WRITE_UINT32(p_Iram->iadd, 0x4);
-+ while (GET_UINT32(p_Iram->iadd) != 0x4) ;
-+ revInfo = GET_UINT32(p_Iram->idata);
-+ p_RevisionInfo->packageRev = (uint16_t)((revInfo & 0xFFFF0000) >> 16);
-+ p_RevisionInfo->majorRev = (uint8_t)((revInfo & 0x0000FF00) >> 8);
-+ p_RevisionInfo->minorRev = (uint8_t)(revInfo & 0x000000FF);
-+
-+ return E_OK;
-+}
-+
-+uint32_t FM_GetCounter(t_Handle h_Fm, e_FmCounters counter)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+ t_Error err;
-+ uint32_t counterValue;
-+ struct fman_rg fman_rg;
-+ enum fman_counters fsl_counter;
-+
-+ SANITY_CHECK_RETURN_VALUE(p_Fm, E_INVALID_HANDLE, 0);
-+ SANITY_CHECK_RETURN_VALUE(!p_Fm->p_FmDriverParam, E_INVALID_STATE, 0);
-+
-+ fman_rg.bmi_rg = p_Fm->p_FmBmiRegs;
-+ fman_rg.qmi_rg = p_Fm->p_FmQmiRegs;
-+ fman_rg.fpm_rg = p_Fm->p_FmFpmRegs;
-+ fman_rg.dma_rg = p_Fm->p_FmDmaRegs;
-+
-+ if ((p_Fm->guestId != NCSW_MASTER_ID) &&
-+ !p_Fm->baseAddr &&
-+ p_Fm->h_IpcSessions[0])
-+ {
-+ t_FmIpcMsg msg;
-+ t_FmIpcReply reply;
-+ uint32_t replyLength, outCounter;
-+
-+ memset(&msg, 0, sizeof(msg));
-+ memset(&reply, 0, sizeof(reply));
-+ msg.msgId = FM_GET_COUNTER;
-+ memcpy(msg.msgBody, (uint8_t *)&counter, sizeof(uint32_t));
-+ replyLength = sizeof(uint32_t) + sizeof(uint32_t);
-+ err = XX_IpcSendMessage(p_Fm->h_IpcSessions[0],
-+ (uint8_t*)&msg,
-+ sizeof(msg.msgId) +sizeof(counterValue),
-+ (uint8_t*)&reply,
-+ &replyLength,
-+ NULL,
-+ NULL);
-+ if (err != E_OK)
-+ {
-+ REPORT_ERROR(MAJOR, err, NO_MSG);
-+ return 0;
-+ }
-+ if (replyLength != (sizeof(uint32_t) + sizeof(uint32_t)))
-+ {
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE, ("IPC reply length mismatch"));
-+ return 0;
-+ }
-+
-+ memcpy((uint8_t*)&outCounter, reply.replyBody, sizeof(uint32_t));
-+ return outCounter;
-+ }
-+ else if (!p_Fm->baseAddr)
-+ {
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Either IPC or 'baseAddress' is required!"));
-+ return 0;
-+ }
-+
-+ /* When applicable (when there is an 'enable counters' bit,
-+ check that counters are enabled */
-+ switch (counter)
-+ {
-+ case (e_FM_COUNTERS_DEQ_1):
-+ case (e_FM_COUNTERS_DEQ_2):
-+ case (e_FM_COUNTERS_DEQ_3):
-+ if ((p_Fm->p_FmStateStruct->revInfo.majorRev == 4) ||
-+ (p_Fm->p_FmStateStruct->revInfo.majorRev >= 6))
-+ {
-+ REPORT_ERROR(MAJOR, E_NOT_SUPPORTED, ("Requested counter not supported"));
-+ return 0;
-+ }
-+ case (e_FM_COUNTERS_ENQ_TOTAL_FRAME):
-+ case (e_FM_COUNTERS_DEQ_TOTAL_FRAME):
-+ case (e_FM_COUNTERS_DEQ_0):
-+ case (e_FM_COUNTERS_DEQ_FROM_DEFAULT):
-+ case (e_FM_COUNTERS_DEQ_FROM_CONTEXT):
-+ case (e_FM_COUNTERS_DEQ_FROM_FD):
-+ case (e_FM_COUNTERS_DEQ_CONFIRM):
-+ if (!(GET_UINT32(p_Fm->p_FmQmiRegs->fmqm_gc) & QMI_CFG_EN_COUNTERS))
-+ {
-+ REPORT_ERROR(MAJOR, E_INVALID_STATE, ("Requested counter was not enabled"));
-+ return 0;
-+ }
-+ break;
-+ default:
-+ break;
-+ }
-+
-+ FMAN_COUNTERS_TRANS(fsl_counter, counter);
-+ return fman_get_counter(&fman_rg, fsl_counter);
-+}
-+
-+t_Error FM_ModifyCounter(t_Handle h_Fm, e_FmCounters counter, uint32_t val)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+ struct fman_rg fman_rg;
-+ enum fman_counters fsl_counter;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Fm, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Fm->p_FmDriverParam, E_INVALID_STATE);
-+
-+ fman_rg.bmi_rg = p_Fm->p_FmBmiRegs;
-+ fman_rg.qmi_rg = p_Fm->p_FmQmiRegs;
-+ fman_rg.fpm_rg = p_Fm->p_FmFpmRegs;
-+ fman_rg.dma_rg = p_Fm->p_FmDmaRegs;
-+
-+ FMAN_COUNTERS_TRANS(fsl_counter, counter);
-+ return (t_Error)fman_modify_counter(&fman_rg, fsl_counter, val);
-+}
-+
-+void FM_SetDmaEmergency(t_Handle h_Fm, e_FmDmaMuramPort muramPort, bool enable)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+ struct fman_dma_regs *dma_rg;
-+
-+ SANITY_CHECK_RETURN(p_Fm, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN(!p_Fm->p_FmDriverParam, E_INVALID_STATE);
-+
-+ dma_rg = p_Fm->p_FmDmaRegs;
-+
-+ fman_set_dma_emergency(dma_rg, !!(muramPort==e_FM_DMA_MURAM_PORT_WRITE), enable);
-+}
-+
-+void FM_SetDmaExtBusPri(t_Handle h_Fm, e_FmDmaExtBusPri pri)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+ struct fman_dma_regs *dma_rg;
-+
-+ SANITY_CHECK_RETURN(p_Fm, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN(!p_Fm->p_FmDriverParam, E_INVALID_STATE);
-+
-+ dma_rg = p_Fm->p_FmDmaRegs;
-+
-+ fman_set_dma_ext_bus_pri(dma_rg, pri);
-+}
-+
-+void FM_GetDmaStatus(t_Handle h_Fm, t_FmDmaStatus *p_FmDmaStatus)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+ uint32_t dmaStatus;
-+ struct fman_dma_regs *dma_rg;
-+
-+ SANITY_CHECK_RETURN(p_Fm, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN(!p_Fm->p_FmDriverParam, E_INVALID_STATE);
-+
-+ dma_rg = p_Fm->p_FmDmaRegs;
-+
-+ if ((p_Fm->guestId != NCSW_MASTER_ID) &&
-+ !p_Fm->baseAddr &&
-+ p_Fm->h_IpcSessions[0])
-+ {
-+ t_FmIpcDmaStatus ipcDmaStatus;
-+ t_FmIpcMsg msg;
-+ t_FmIpcReply reply;
-+ t_Error err;
-+ uint32_t replyLength;
-+
-+ memset(&msg, 0, sizeof(msg));
-+ memset(&reply, 0, sizeof(reply));
-+ msg.msgId = FM_DMA_STAT;
-+ replyLength = sizeof(uint32_t) + sizeof(t_FmIpcDmaStatus);
-+ err = XX_IpcSendMessage(p_Fm->h_IpcSessions[0],
-+ (uint8_t*)&msg,
-+ sizeof(msg.msgId),
-+ (uint8_t*)&reply,
-+ &replyLength,
-+ NULL,
-+ NULL);
-+ if (err != E_OK)
-+ {
-+ REPORT_ERROR(MINOR, err, NO_MSG);
-+ return;
-+ }
-+ if (replyLength != (sizeof(uint32_t) + sizeof(t_FmIpcDmaStatus)))
-+ {
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE, ("IPC reply length mismatch"));
-+ return;
-+ }
-+ memcpy((uint8_t*)&ipcDmaStatus, reply.replyBody, sizeof(t_FmIpcDmaStatus));
-+
-+ p_FmDmaStatus->cmqNotEmpty = (bool)ipcDmaStatus.boolCmqNotEmpty; /**< Command queue is not empty */
-+ p_FmDmaStatus->busError = (bool)ipcDmaStatus.boolBusError; /**< Bus error occurred */
-+ p_FmDmaStatus->readBufEccError = (bool)ipcDmaStatus.boolReadBufEccError; /**< Double ECC error on buffer Read */
-+ p_FmDmaStatus->writeBufEccSysError =(bool)ipcDmaStatus.boolWriteBufEccSysError; /**< Double ECC error on buffer write from system side */
-+ p_FmDmaStatus->writeBufEccFmError = (bool)ipcDmaStatus.boolWriteBufEccFmError; /**< Double ECC error on buffer write from FM side */
-+ p_FmDmaStatus->singlePortEccError = (bool)ipcDmaStatus.boolSinglePortEccError; /**< Double ECC error on buffer write from FM side */
-+ return;
-+ }
-+ else if (!p_Fm->baseAddr)
-+ {
-+ REPORT_ERROR(MINOR, E_NOT_SUPPORTED,
-+ ("Either IPC or 'baseAddress' is required!"));
-+ return;
-+ }
-+
-+ dmaStatus = fman_get_dma_status(dma_rg);
-+
-+ p_FmDmaStatus->cmqNotEmpty = (bool)(dmaStatus & DMA_STATUS_CMD_QUEUE_NOT_EMPTY);
-+ p_FmDmaStatus->busError = (bool)(dmaStatus & DMA_STATUS_BUS_ERR);
-+ if (p_Fm->p_FmStateStruct->revInfo.majorRev >= 6)
-+ p_FmDmaStatus->singlePortEccError = (bool)(dmaStatus & DMA_STATUS_FM_SPDAT_ECC);
-+ else
-+ {
-+ p_FmDmaStatus->readBufEccError = (bool)(dmaStatus & DMA_STATUS_READ_ECC);
-+ p_FmDmaStatus->writeBufEccSysError = (bool)(dmaStatus & DMA_STATUS_SYSTEM_WRITE_ECC);
-+ p_FmDmaStatus->writeBufEccFmError = (bool)(dmaStatus & DMA_STATUS_FM_WRITE_ECC);
-+ }
-+}
-+
-+void FM_Resume(t_Handle h_Fm)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+ struct fman_fpm_regs *fpm_rg;
-+
-+ SANITY_CHECK_RETURN(p_Fm, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN(!p_Fm->p_FmDriverParam, E_INVALID_STATE);
-+ SANITY_CHECK_RETURN((p_Fm->guestId == NCSW_MASTER_ID), E_NOT_SUPPORTED);
-+
-+ fpm_rg = p_Fm->p_FmFpmRegs;
-+
-+ fman_resume(fpm_rg);
-+}
-+
-+t_Error FM_GetSpecialOperationCoding(t_Handle h_Fm,
-+ fmSpecialOperations_t spOper,
-+ uint8_t *p_SpOperCoding)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+ t_FmCtrlCodeRevisionInfo revInfo;
-+ t_Error err;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Fm, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Fm->p_FmDriverParam, E_INVALID_STATE);
-+ SANITY_CHECK_RETURN_ERROR(p_SpOperCoding, E_NULL_POINTER);
-+
-+ if (!spOper)
-+ {
-+ *p_SpOperCoding = 0;
-+ return E_OK;
-+ }
-+
-+ if ((err = FM_GetFmanCtrlCodeRevision(p_Fm, &revInfo)) != E_OK)
-+ {
-+ DBG(WARNING, ("FM in guest-mode without IPC, can't validate firmware revision."));
-+ revInfo.packageRev = IP_OFFLOAD_PACKAGE_NUMBER;
-+ }
-+ else if (!IS_OFFLOAD_PACKAGE(revInfo.packageRev))
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, ("Fman ctrl code package"));
-+
-+ switch (spOper)
-+ {
-+ case (FM_SP_OP_CAPWAP_DTLS_DEC):
-+ *p_SpOperCoding = 9;
-+ break;
-+ case (FM_SP_OP_CAPWAP_DTLS_ENC):
-+ *p_SpOperCoding = 10;
-+ break;
-+ case (FM_SP_OP_IPSEC|FM_SP_OP_IPSEC_UPDATE_UDP_LEN|FM_SP_OP_IPSEC_MANIP):
-+ case (FM_SP_OP_IPSEC|FM_SP_OP_IPSEC_UPDATE_UDP_LEN|FM_SP_OP_IPSEC_MANIP|FM_SP_OP_RPD):
-+ *p_SpOperCoding = 5;
-+ break;
-+ case (FM_SP_OP_IPSEC|FM_SP_OP_IPSEC_MANIP):
-+ case (FM_SP_OP_IPSEC|FM_SP_OP_IPSEC_MANIP|FM_SP_OP_RPD):
-+ *p_SpOperCoding = 6;
-+ break;
-+ case (FM_SP_OP_IPSEC|FM_SP_OP_IPSEC_UPDATE_UDP_LEN|FM_SP_OP_RPD):
-+ *p_SpOperCoding = 3;
-+ break;
-+ case (FM_SP_OP_IPSEC|FM_SP_OP_IPSEC_UPDATE_UDP_LEN):
-+ *p_SpOperCoding = 1;
-+ break;
-+ case (FM_SP_OP_IPSEC|FM_SP_OP_IPSEC_UPDATE_UDP_LEN|FM_SP_OP_IPSEC_NO_ETH_HDR):
-+ *p_SpOperCoding = 12;
-+ break;
-+ case (FM_SP_OP_IPSEC|FM_SP_OP_RPD):
-+ *p_SpOperCoding = 4;
-+ break;
-+ case (FM_SP_OP_IPSEC):
-+ *p_SpOperCoding = 2;
-+ break;
-+ case (FM_SP_OP_DCL4C):
-+ *p_SpOperCoding = 7;
-+ break;
-+ case (FM_SP_OP_CLEAR_RPD):
-+ *p_SpOperCoding = 8;
-+ break;
-+ default:
-+ RETURN_ERROR(MINOR, E_INVALID_VALUE, NO_MSG);
-+ }
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_CtrlMonStart(t_Handle h_Fm)
-+{
-+ t_Fm *p_Fm = (t_Fm *)h_Fm;
-+ t_FmTrbRegs *p_MonRegs;
-+ uint8_t i;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Fm, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Fm->p_FmDriverParam, E_INVALID_STATE);
-+ SANITY_CHECK_RETURN_ERROR((p_Fm->guestId == NCSW_MASTER_ID), E_NOT_SUPPORTED);
-+
-+ WRITE_UINT32(p_Fm->p_FmFpmRegs->fmfp_brkc,
-+ GET_UINT32(p_Fm->p_FmFpmRegs->fmfp_brkc) | FPM_BRKC_RDBG);
-+
-+ for (i = 0; i < FM_NUM_OF_CTRL; i++)
-+ {
-+ p_MonRegs = (t_FmTrbRegs *)UINT_TO_PTR(p_Fm->baseAddr + FM_MM_TRB(i));
-+
-+ /* Reset control registers */
-+ WRITE_UINT32(p_MonRegs->tcrh, TRB_TCRH_RESET);
-+ WRITE_UINT32(p_MonRegs->tcrl, TRB_TCRL_RESET);
-+
-+ /* Configure: counter #1 counts all stalls in risc - ldsched stall
-+ counter #2 counts all stalls in risc - other stall*/
-+ WRITE_UINT32(p_MonRegs->tcrl, TRB_TCRL_RESET | TRB_TCRL_UTIL);
-+
-+ /* Enable monitoring */
-+ WRITE_UINT32(p_MonRegs->tcrh, TRB_TCRH_ENABLE_COUNTERS);
-+ }
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_CtrlMonStop(t_Handle h_Fm)
-+{
-+ t_Fm *p_Fm = (t_Fm *)h_Fm;
-+ t_FmTrbRegs *p_MonRegs;
-+ uint8_t i;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Fm, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Fm->p_FmDriverParam, E_INVALID_STATE);
-+ SANITY_CHECK_RETURN_ERROR((p_Fm->guestId == NCSW_MASTER_ID), E_NOT_SUPPORTED);
-+
-+ for (i = 0; i < FM_NUM_OF_CTRL; i++)
-+ {
-+ p_MonRegs = (t_FmTrbRegs *)UINT_TO_PTR(p_Fm->baseAddr + FM_MM_TRB(i));
-+ WRITE_UINT32(p_MonRegs->tcrh, TRB_TCRH_DISABLE_COUNTERS);
-+ }
-+
-+ WRITE_UINT32(p_Fm->p_FmFpmRegs->fmfp_brkc,
-+ GET_UINT32(p_Fm->p_FmFpmRegs->fmfp_brkc) & ~FPM_BRKC_RDBG);
-+
-+ return E_OK;
-+}
-+
-+t_Error FM_CtrlMonGetCounters(t_Handle h_Fm, uint8_t fmCtrlIndex, t_FmCtrlMon *p_Mon)
-+{
-+ t_Fm *p_Fm = (t_Fm *)h_Fm;
-+ t_FmTrbRegs *p_MonRegs;
-+ uint64_t clkCnt, utilValue, effValue;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Fm, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Fm->p_FmDriverParam, E_INVALID_STATE);
-+ SANITY_CHECK_RETURN_ERROR((p_Fm->guestId == NCSW_MASTER_ID), E_NOT_SUPPORTED);
-+ SANITY_CHECK_RETURN_ERROR(p_Mon, E_NULL_POINTER);
-+
-+ if (fmCtrlIndex >= FM_NUM_OF_CTRL)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("FM Controller index"));
-+
-+ p_MonRegs = (t_FmTrbRegs *)UINT_TO_PTR(p_Fm->baseAddr + FM_MM_TRB(fmCtrlIndex));
-+
-+ clkCnt = (uint64_t)
-+ ((uint64_t)GET_UINT32(p_MonRegs->tpcch) << 32 | GET_UINT32(p_MonRegs->tpccl));
-+
-+ utilValue = (uint64_t)
-+ ((uint64_t)GET_UINT32(p_MonRegs->tpc1h) << 32 | GET_UINT32(p_MonRegs->tpc1l));
-+
-+ effValue = (uint64_t)
-+ ((uint64_t)GET_UINT32(p_MonRegs->tpc2h) << 32 | GET_UINT32(p_MonRegs->tpc2l));
-+
-+ p_Mon->percentCnt[0] = (uint8_t)div64_u64((clkCnt - utilValue) * 100, clkCnt);
-+ if (clkCnt != utilValue)
-+ p_Mon->percentCnt[1] = (uint8_t)div64_u64(((clkCnt - utilValue) - effValue) * 100, clkCnt - utilValue);
-+ else
-+ p_Mon->percentCnt[1] = 0;
-+
-+ return E_OK;
-+}
-+
-+t_Handle FM_GetMuramHandle(t_Handle h_Fm)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+
-+ SANITY_CHECK_RETURN_VALUE(p_Fm, E_INVALID_HANDLE, NULL);
-+
-+ return (p_Fm->h_FmMuram);
-+}
-+
-+/****************************************************/
-+/* Hidden-DEBUG Only API */
-+/****************************************************/
-+t_Error FM_ForceIntr (t_Handle h_Fm, e_FmExceptions exception)
-+{
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+ enum fman_exceptions fslException;
-+ struct fman_rg fman_rg;
-+
-+ SANITY_CHECK_RETURN_ERROR(p_Fm, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(!p_Fm->p_FmDriverParam, E_INVALID_STATE);
-+
-+ fman_rg.bmi_rg = p_Fm->p_FmBmiRegs;
-+ fman_rg.qmi_rg = p_Fm->p_FmQmiRegs;
-+ fman_rg.fpm_rg = p_Fm->p_FmFpmRegs;
-+ fman_rg.dma_rg = p_Fm->p_FmDmaRegs;
-+
-+ switch (exception)
-+ {
-+ case e_FM_EX_QMI_DEQ_FROM_UNKNOWN_PORTID:
-+ if (!(p_Fm->p_FmStateStruct->exceptions & FM_EX_QMI_DEQ_FROM_UNKNOWN_PORTID))
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, ("The selected exception is masked"));
-+ break;
-+ case e_FM_EX_QMI_SINGLE_ECC:
-+ if (p_Fm->p_FmStateStruct->revInfo.majorRev >= 6)
-+ RETURN_ERROR(MAJOR, E_NOT_SUPPORTED, ("e_FM_EX_QMI_SINGLE_ECC not supported on this integration."));
-+
-+ if (!(p_Fm->p_FmStateStruct->exceptions & FM_EX_QMI_SINGLE_ECC))
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, ("The selected exception is masked"));
-+ break;
-+ case e_FM_EX_QMI_DOUBLE_ECC:
-+ if (!(p_Fm->p_FmStateStruct->exceptions & FM_EX_QMI_DOUBLE_ECC))
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, ("The selected exception is masked"));
-+ break;
-+ case e_FM_EX_BMI_LIST_RAM_ECC:
-+ if (!(p_Fm->p_FmStateStruct->exceptions & FM_EX_BMI_LIST_RAM_ECC))
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, ("The selected exception is masked"));
-+ break;
-+ case e_FM_EX_BMI_STORAGE_PROFILE_ECC:
-+ if (!(p_Fm->p_FmStateStruct->exceptions & FM_EX_BMI_STORAGE_PROFILE_ECC))
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, ("The selected exception is masked"));
-+ break;
-+ case e_FM_EX_BMI_STATISTICS_RAM_ECC:
-+ if (!(p_Fm->p_FmStateStruct->exceptions & FM_EX_BMI_STATISTICS_RAM_ECC))
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, ("The selected exception is masked"));
-+ break;
-+ case e_FM_EX_BMI_DISPATCH_RAM_ECC:
-+ if (!(p_Fm->p_FmStateStruct->exceptions & FM_EX_BMI_DISPATCH_RAM_ECC))
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, ("The selected exception is masked"));
-+ break;
-+ default:
-+ RETURN_ERROR(MINOR, E_NOT_SUPPORTED, ("The selected exception may not be forced"));
-+ }
-+
-+ fslException = FmanExceptionTrans(exception);
-+ fman_force_intr (&fman_rg, fslException);
-+
-+ return E_OK;
-+}
-+
-+t_Handle FmGetPcd(t_Handle h_Fm)
-+{
-+ return ((t_Fm*)h_Fm)->h_Pcd;
-+}
-+#if (DPAA_VERSION >= 11)
-+extern void *g_MemacRegs;
-+void fm_clk_down(void);
-+uint32_t fman_memac_get_event(void *regs, uint32_t ev_mask);
-+void FM_ChangeClock(t_Handle h_Fm, int hardwarePortId)
-+{
-+ int macId;
-+ uint32_t event, rcr;
-+ t_Fm *p_Fm = (t_Fm*)h_Fm;
-+ rcr = GET_UINT32(p_Fm->p_FmFpmRegs->fm_rcr);
-+ rcr |= 0x04000000;
-+ WRITE_UINT32(p_Fm->p_FmFpmRegs->fm_rcr, rcr);
-+
-+ HW_PORT_ID_TO_SW_PORT_ID(macId, hardwarePortId);
-+ do
-+ {
-+ event = fman_memac_get_event(g_MemacRegs, 0xFFFFFFFF);
-+ } while ((event & 0x00000020) == 0);
-+ fm_clk_down();
-+ rcr = GET_UINT32(p_Fm->p_FmFpmRegs->fm_rcr);
-+ rcr &= ~0x04000000;
-+ WRITE_UINT32(p_Fm->p_FmFpmRegs->fm_rcr, rcr);
-+}
-+#endif
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/fm.h
-@@ -0,0 +1,648 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/******************************************************************************
-+ @File fm.h
-+
-+ @Description FM internal structures and definitions.
-+*//***************************************************************************/
-+#ifndef __FM_H
-+#define __FM_H
-+
-+#include "error_ext.h"
-+#include "std_ext.h"
-+#include "fm_ext.h"
-+#include "fm_ipc.h"
-+
-+#include "fsl_fman.h"
-+
-+#define __ERR_MODULE__ MODULE_FM
-+
-+#define FM_MAX_NUM_OF_HW_PORT_IDS 64
-+#define FM_MAX_NUM_OF_GUESTS 100
-+
-+/**************************************************************************//**
-+ @Description Exceptions
-+*//***************************************************************************/
-+#define FM_EX_DMA_BUS_ERROR 0x80000000 /**< DMA bus error. */
-+#define FM_EX_DMA_READ_ECC 0x40000000
-+#define FM_EX_DMA_SYSTEM_WRITE_ECC 0x20000000
-+#define FM_EX_DMA_FM_WRITE_ECC 0x10000000
-+#define FM_EX_FPM_STALL_ON_TASKS 0x08000000 /**< Stall of tasks on FPM */
-+#define FM_EX_FPM_SINGLE_ECC 0x04000000 /**< Single ECC on FPM */
-+#define FM_EX_FPM_DOUBLE_ECC 0x02000000
-+#define FM_EX_QMI_SINGLE_ECC 0x01000000 /**< Single ECC on FPM */
-+#define FM_EX_QMI_DEQ_FROM_UNKNOWN_PORTID 0x00800000 /**< Dequeu from default queue id */
-+#define FM_EX_QMI_DOUBLE_ECC 0x00400000
-+#define FM_EX_BMI_LIST_RAM_ECC 0x00200000
-+#define FM_EX_BMI_STORAGE_PROFILE_ECC 0x00100000
-+#define FM_EX_BMI_STATISTICS_RAM_ECC 0x00080000
-+#define FM_EX_IRAM_ECC 0x00040000
-+#define FM_EX_MURAM_ECC 0x00020000
-+#define FM_EX_BMI_DISPATCH_RAM_ECC 0x00010000
-+#define FM_EX_DMA_SINGLE_PORT_ECC 0x00008000
-+
-+#define DMA_EMSR_EMSTR_MASK 0x0000FFFF
-+
-+#define DMA_THRESH_COMMQ_MASK 0xFF000000
-+#define DMA_THRESH_READ_INT_BUF_MASK 0x007F0000
-+#define DMA_THRESH_WRITE_INT_BUF_MASK 0x0000007F
-+
-+#define GET_EXCEPTION_FLAG(bitMask, exception) \
-+switch (exception){ \
-+ case e_FM_EX_DMA_BUS_ERROR: \
-+ bitMask = FM_EX_DMA_BUS_ERROR; break; \
-+ case e_FM_EX_DMA_SINGLE_PORT_ECC: \
-+ bitMask = FM_EX_DMA_SINGLE_PORT_ECC; break; \
-+ case e_FM_EX_DMA_READ_ECC: \
-+ bitMask = FM_EX_DMA_READ_ECC; break; \
-+ case e_FM_EX_DMA_SYSTEM_WRITE_ECC: \
-+ bitMask = FM_EX_DMA_SYSTEM_WRITE_ECC; break; \
-+ case e_FM_EX_DMA_FM_WRITE_ECC: \
-+ bitMask = FM_EX_DMA_FM_WRITE_ECC; break; \
-+ case e_FM_EX_FPM_STALL_ON_TASKS: \
-+ bitMask = FM_EX_FPM_STALL_ON_TASKS; break; \
-+ case e_FM_EX_FPM_SINGLE_ECC: \
-+ bitMask = FM_EX_FPM_SINGLE_ECC; break; \
-+ case e_FM_EX_FPM_DOUBLE_ECC: \
-+ bitMask = FM_EX_FPM_DOUBLE_ECC; break; \
-+ case e_FM_EX_QMI_SINGLE_ECC: \
-+ bitMask = FM_EX_QMI_SINGLE_ECC; break; \
-+ case e_FM_EX_QMI_DOUBLE_ECC: \
-+ bitMask = FM_EX_QMI_DOUBLE_ECC; break; \
-+ case e_FM_EX_QMI_DEQ_FROM_UNKNOWN_PORTID: \
-+ bitMask = FM_EX_QMI_DEQ_FROM_UNKNOWN_PORTID; break; \
-+ case e_FM_EX_BMI_LIST_RAM_ECC: \
-+ bitMask = FM_EX_BMI_LIST_RAM_ECC; break; \
-+ case e_FM_EX_BMI_STORAGE_PROFILE_ECC: \
-+ bitMask = FM_EX_BMI_STORAGE_PROFILE_ECC; break; \
-+ case e_FM_EX_BMI_STATISTICS_RAM_ECC: \
-+ bitMask = FM_EX_BMI_STATISTICS_RAM_ECC; break; \
-+ case e_FM_EX_BMI_DISPATCH_RAM_ECC: \
-+ bitMask = FM_EX_BMI_DISPATCH_RAM_ECC; break; \
-+ case e_FM_EX_IRAM_ECC: \
-+ bitMask = FM_EX_IRAM_ECC; break; \
-+ case e_FM_EX_MURAM_ECC: \
-+ bitMask = FM_EX_MURAM_ECC; break; \
-+ default: bitMask = 0;break; \
-+}
-+
-+#define GET_FM_MODULE_EVENT(_mod, _id, _intrType, _event) \
-+ switch (_mod) { \
-+ case e_FM_MOD_PRS: \
-+ if (_id) _event = e_FM_EV_DUMMY_LAST; \
-+ else _event = (_intrType == e_FM_INTR_TYPE_ERR) ? e_FM_EV_ERR_PRS : e_FM_EV_PRS; \
-+ break; \
-+ case e_FM_MOD_KG: \
-+ if (_id) _event = e_FM_EV_DUMMY_LAST; \
-+ else _event = (_intrType == e_FM_INTR_TYPE_ERR) ? e_FM_EV_ERR_KG : e_FM_EV_DUMMY_LAST; \
-+ break; \
-+ case e_FM_MOD_PLCR: \
-+ if (_id) _event = e_FM_EV_DUMMY_LAST; \
-+ else _event = (_intrType == e_FM_INTR_TYPE_ERR) ? e_FM_EV_ERR_PLCR : e_FM_EV_PLCR; \
-+ break; \
-+ case e_FM_MOD_TMR: \
-+ if (_id) _event = e_FM_EV_DUMMY_LAST; \
-+ else _event = (_intrType == e_FM_INTR_TYPE_ERR) ? e_FM_EV_DUMMY_LAST : e_FM_EV_TMR; \
-+ break; \
-+ case e_FM_MOD_10G_MAC: \
-+ if (_id >= FM_MAX_NUM_OF_10G_MACS) _event = e_FM_EV_DUMMY_LAST; \
-+ else _event = (_intrType == e_FM_INTR_TYPE_ERR) ? (e_FM_EV_ERR_10G_MAC0 + _id) : (e_FM_EV_10G_MAC0 + _id); \
-+ break; \
-+ case e_FM_MOD_1G_MAC: \
-+ if (_id >= FM_MAX_NUM_OF_1G_MACS) _event = e_FM_EV_DUMMY_LAST; \
-+ else _event = (_intrType == e_FM_INTR_TYPE_ERR) ? (e_FM_EV_ERR_1G_MAC0 + _id) : (e_FM_EV_1G_MAC0 + _id); \
-+ break; \
-+ case e_FM_MOD_MACSEC: \
-+ switch (_id){ \
-+ case (0): _event = (_intrType == e_FM_INTR_TYPE_ERR) ? e_FM_EV_ERR_MACSEC_MAC0:e_FM_EV_MACSEC_MAC0; \
-+ break; \
-+ } \
-+ break; \
-+ case e_FM_MOD_FMAN_CTRL: \
-+ if (_intrType == e_FM_INTR_TYPE_ERR) _event = e_FM_EV_DUMMY_LAST; \
-+ else _event = (e_FM_EV_FMAN_CTRL_0 + _id); \
-+ break; \
-+ default: _event = e_FM_EV_DUMMY_LAST; \
-+ break; \
-+ }
-+
-+#define FMAN_CACHE_OVERRIDE_TRANS(fsl_cache_override, _cache_override) \
-+ switch (_cache_override){ \
-+ case e_FM_DMA_NO_CACHE_OR: \
-+ fsl_cache_override = E_FMAN_DMA_NO_CACHE_OR; break; \
-+ case e_FM_DMA_NO_STASH_DATA: \
-+ fsl_cache_override = E_FMAN_DMA_NO_STASH_DATA; break; \
-+ case e_FM_DMA_MAY_STASH_DATA: \
-+ fsl_cache_override = E_FMAN_DMA_MAY_STASH_DATA; break; \
-+ case e_FM_DMA_STASH_DATA: \
-+ fsl_cache_override = E_FMAN_DMA_STASH_DATA; break; \
-+ default: \
-+ fsl_cache_override = E_FMAN_DMA_NO_CACHE_OR; break; \
-+ }
-+
-+#define FMAN_AID_MODE_TRANS(fsl_aid_mode, _aid_mode) \
-+ switch (_aid_mode){ \
-+ case e_FM_DMA_AID_OUT_PORT_ID: \
-+ fsl_aid_mode = E_FMAN_DMA_AID_OUT_PORT_ID; break; \
-+ case e_FM_DMA_AID_OUT_TNUM: \
-+ fsl_aid_mode = E_FMAN_DMA_AID_OUT_TNUM; break; \
-+ default: \
-+ fsl_aid_mode = E_FMAN_DMA_AID_OUT_PORT_ID; break; \
-+ }
-+
-+#define FMAN_DMA_DBG_CNT_TRANS(fsl_dma_dbg_cnt, _dma_dbg_cnt) \
-+ switch (_dma_dbg_cnt){ \
-+ case e_FM_DMA_DBG_NO_CNT: \
-+ fsl_dma_dbg_cnt = E_FMAN_DMA_DBG_NO_CNT; break; \
-+ case e_FM_DMA_DBG_CNT_DONE: \
-+ fsl_dma_dbg_cnt = E_FMAN_DMA_DBG_CNT_DONE; break; \
-+ case e_FM_DMA_DBG_CNT_COMM_Q_EM: \
-+ fsl_dma_dbg_cnt = E_FMAN_DMA_DBG_CNT_COMM_Q_EM; break; \
-+ case e_FM_DMA_DBG_CNT_INT_READ_EM: \
-+ fsl_dma_dbg_cnt = E_FMAN_DMA_DBG_CNT_INT_READ_EM; break; \
-+ case e_FM_DMA_DBG_CNT_INT_WRITE_EM: \
-+ fsl_dma_dbg_cnt = E_FMAN_DMA_DBG_CNT_INT_WRITE_EM ; break; \
-+ case e_FM_DMA_DBG_CNT_FPM_WAIT: \
-+ fsl_dma_dbg_cnt = E_FMAN_DMA_DBG_CNT_FPM_WAIT ; break; \
-+ case e_FM_DMA_DBG_CNT_SIGLE_BIT_ECC: \
-+ fsl_dma_dbg_cnt = E_FMAN_DMA_DBG_CNT_SIGLE_BIT_ECC ; break; \
-+ case e_FM_DMA_DBG_CNT_RAW_WAR_PROT: \
-+ fsl_dma_dbg_cnt = E_FMAN_DMA_DBG_CNT_RAW_WAR_PROT ; break; \
-+ default: \
-+ fsl_dma_dbg_cnt = E_FMAN_DMA_DBG_NO_CNT; break; \
-+ }
-+
-+#define FMAN_DMA_EMER_TRANS(fsl_dma_emer, _dma_emer) \
-+ switch (_dma_emer){ \
-+ case e_FM_DMA_EM_EBS: \
-+ fsl_dma_emer = E_FMAN_DMA_EM_EBS; break; \
-+ case e_FM_DMA_EM_SOS: \
-+ fsl_dma_emer = E_FMAN_DMA_EM_SOS; break; \
-+ default: \
-+ fsl_dma_emer = E_FMAN_DMA_EM_EBS; break; \
-+ }
-+
-+#define FMAN_DMA_ERR_TRANS(fsl_dma_err, _dma_err) \
-+ switch (_dma_err){ \
-+ case e_FM_DMA_ERR_CATASTROPHIC: \
-+ fsl_dma_err = E_FMAN_DMA_ERR_CATASTROPHIC; break; \
-+ case e_FM_DMA_ERR_REPORT: \
-+ fsl_dma_err = E_FMAN_DMA_ERR_REPORT; break; \
-+ default: \
-+ fsl_dma_err = E_FMAN_DMA_ERR_CATASTROPHIC; break; \
-+ }
-+
-+#define FMAN_CATASTROPHIC_ERR_TRANS(fsl_catastrophic_err, _catastrophic_err) \
-+ switch (_catastrophic_err){ \
-+ case e_FM_CATASTROPHIC_ERR_STALL_PORT: \
-+ fsl_catastrophic_err = E_FMAN_CATAST_ERR_STALL_PORT; break; \
-+ case e_FM_CATASTROPHIC_ERR_STALL_TASK: \
-+ fsl_catastrophic_err = E_FMAN_CATAST_ERR_STALL_TASK; break; \
-+ default: \
-+ fsl_catastrophic_err = E_FMAN_CATAST_ERR_STALL_PORT; break; \
-+ }
-+
-+#define FMAN_COUNTERS_TRANS(fsl_counters, _counters) \
-+ switch (_counters){ \
-+ case e_FM_COUNTERS_ENQ_TOTAL_FRAME: \
-+ fsl_counters = E_FMAN_COUNTERS_ENQ_TOTAL_FRAME; break; \
-+ case e_FM_COUNTERS_DEQ_TOTAL_FRAME: \
-+ fsl_counters = E_FMAN_COUNTERS_DEQ_TOTAL_FRAME; break; \
-+ case e_FM_COUNTERS_DEQ_0: \
-+ fsl_counters = E_FMAN_COUNTERS_DEQ_0; break; \
-+ case e_FM_COUNTERS_DEQ_1: \
-+ fsl_counters = E_FMAN_COUNTERS_DEQ_1; break; \
-+ case e_FM_COUNTERS_DEQ_2: \
-+ fsl_counters = E_FMAN_COUNTERS_DEQ_2; break; \
-+ case e_FM_COUNTERS_DEQ_3: \
-+ fsl_counters = E_FMAN_COUNTERS_DEQ_3; break; \
-+ case e_FM_COUNTERS_DEQ_FROM_DEFAULT: \
-+ fsl_counters = E_FMAN_COUNTERS_DEQ_FROM_DEFAULT; break; \
-+ case e_FM_COUNTERS_DEQ_FROM_CONTEXT: \
-+ fsl_counters = E_FMAN_COUNTERS_DEQ_FROM_CONTEXT; break; \
-+ case e_FM_COUNTERS_DEQ_FROM_FD: \
-+ fsl_counters = E_FMAN_COUNTERS_DEQ_FROM_FD; break; \
-+ case e_FM_COUNTERS_DEQ_CONFIRM: \
-+ fsl_counters = E_FMAN_COUNTERS_DEQ_CONFIRM; break; \
-+ default: \
-+ fsl_counters = E_FMAN_COUNTERS_ENQ_TOTAL_FRAME; break; \
-+ }
-+
-+/**************************************************************************//**
-+ @Description defaults
-+*//***************************************************************************/
-+#define DEFAULT_exceptions (FM_EX_DMA_BUS_ERROR |\
-+ FM_EX_DMA_READ_ECC |\
-+ FM_EX_DMA_SYSTEM_WRITE_ECC |\
-+ FM_EX_DMA_FM_WRITE_ECC |\
-+ FM_EX_FPM_STALL_ON_TASKS |\
-+ FM_EX_FPM_SINGLE_ECC |\
-+ FM_EX_FPM_DOUBLE_ECC |\
-+ FM_EX_QMI_DEQ_FROM_UNKNOWN_PORTID|\
-+ FM_EX_BMI_LIST_RAM_ECC |\
-+ FM_EX_BMI_STORAGE_PROFILE_ECC |\
-+ FM_EX_BMI_STATISTICS_RAM_ECC |\
-+ FM_EX_IRAM_ECC |\
-+ FM_EX_MURAM_ECC |\
-+ FM_EX_BMI_DISPATCH_RAM_ECC |\
-+ FM_EX_QMI_DOUBLE_ECC |\
-+ FM_EX_QMI_SINGLE_ECC)
-+
-+#define DEFAULT_eccEnable FALSE
-+#ifdef FM_PEDANTIC_DMA
-+#define DEFAULT_aidOverride TRUE
-+#else
-+#define DEFAULT_aidOverride FALSE
-+#endif /* FM_PEDANTIC_DMA */
-+#define DEFAULT_aidMode e_FM_DMA_AID_OUT_TNUM
-+#define DEFAULT_dmaStopOnBusError FALSE
-+#define DEFAULT_stopAtBusError FALSE
-+#define DEFAULT_axiDbgNumOfBeats 1
-+#define DEFAULT_dmaReadIntBufLow ((DMA_THRESH_MAX_BUF+1)/2)
-+#define DEFAULT_dmaReadIntBufHigh ((DMA_THRESH_MAX_BUF+1)*3/4)
-+#define DEFAULT_dmaWriteIntBufLow ((DMA_THRESH_MAX_BUF+1)/2)
-+#define DEFAULT_dmaWriteIntBufHigh ((DMA_THRESH_MAX_BUF+1)*3/4)
-+#define DEFAULT_catastrophicErr e_FM_CATASTROPHIC_ERR_STALL_PORT
-+#define DEFAULT_dmaErr e_FM_DMA_ERR_CATASTROPHIC
-+#define DEFAULT_resetOnInit FALSE
-+#define DEFAULT_resetOnInitOverrideCallback NULL
-+#define DEFAULT_haltOnExternalActivation FALSE /* do not change! if changed, must be disabled for rev1 ! */
-+#define DEFAULT_haltOnUnrecoverableEccError FALSE /* do not change! if changed, must be disabled for rev1 ! */
-+#define DEFAULT_externalEccRamsEnable FALSE
-+#define DEFAULT_VerifyUcode FALSE
-+
-+#if (DPAA_VERSION < 11)
-+#define DEFAULT_totalFifoSize(major, minor) \
-+ (((major == 2) || (major == 5)) ? \
-+ (100*KILOBYTE) : ((major == 4) ? \
-+ (49*KILOBYTE) : (122*KILOBYTE)))
-+#define DEFAULT_totalNumOfTasks(major, minor) \
-+ BMI_MAX_NUM_OF_TASKS
-+
-+#define DEFAULT_dmaCommQLow ((DMA_THRESH_MAX_COMMQ+1)/2)
-+#define DEFAULT_dmaCommQHigh ((DMA_THRESH_MAX_COMMQ+1)*3/4)
-+#define DEFAULT_cacheOverride e_FM_DMA_NO_CACHE_OR
-+#define DEFAULT_dmaCamNumOfEntries 32
-+#define DEFAULT_dmaDbgCntMode e_FM_DMA_DBG_NO_CNT
-+#define DEFAULT_dmaEnEmergency FALSE
-+#define DEFAULT_dmaSosEmergency 0
-+#define DEFAULT_dmaWatchdog 0 /* disabled */
-+#define DEFAULT_dmaEnEmergencySmoother FALSE
-+#define DEFAULT_dmaEmergencySwitchCounter 0
-+
-+#define DEFAULT_dispLimit 0
-+#define DEFAULT_prsDispTh 16
-+#define DEFAULT_plcrDispTh 16
-+#define DEFAULT_kgDispTh 16
-+#define DEFAULT_bmiDispTh 16
-+#define DEFAULT_qmiEnqDispTh 16
-+#define DEFAULT_qmiDeqDispTh 16
-+#define DEFAULT_fmCtl1DispTh 16
-+#define DEFAULT_fmCtl2DispTh 16
-+
-+#else /* (DPAA_VERSION < 11) */
-+/* Defaults are registers' reset values */
-+#define DEFAULT_totalFifoSize(major, minor) \
-+ (((major == 6) && ((minor == 1) || (minor == 4))) ? \
-+ (156*KILOBYTE) : (295*KILOBYTE))
-+
-+/* According to the default value of FMBM_CFG2[TNTSKS] */
-+#define DEFAULT_totalNumOfTasks(major, minor) \
-+ (((major == 6) && ((minor == 1) || (minor == 4))) ? 59 : 124)
-+
-+#define DEFAULT_dmaCommQLow 0x2A
-+#define DEFAULT_dmaCommQHigh 0x3F
-+#define DEFAULT_cacheOverride e_FM_DMA_NO_CACHE_OR
-+#define DEFAULT_dmaCamNumOfEntries 64
-+#define DEFAULT_dmaDbgCntMode e_FM_DMA_DBG_NO_CNT
-+#define DEFAULT_dmaEnEmergency FALSE
-+#define DEFAULT_dmaSosEmergency 0
-+#define DEFAULT_dmaWatchdog 0 /* disabled */
-+#define DEFAULT_dmaEnEmergencySmoother FALSE
-+#define DEFAULT_dmaEmergencySwitchCounter 0
-+
-+#define DEFAULT_dispLimit 0
-+#define DEFAULT_prsDispTh 16
-+#define DEFAULT_plcrDispTh 16
-+#define DEFAULT_kgDispTh 16
-+#define DEFAULT_bmiDispTh 16
-+#define DEFAULT_qmiEnqDispTh 16
-+#define DEFAULT_qmiDeqDispTh 16
-+#define DEFAULT_fmCtl1DispTh 16
-+#define DEFAULT_fmCtl2DispTh 16
-+#endif /* (DPAA_VERSION < 11) */
-+
-+#define FM_TIMESTAMP_1_USEC_BIT 8
-+
-+/**************************************************************************//**
-+ @Collection Defines used for enabling/disabling FM interrupts
-+ @{
-+*//***************************************************************************/
-+#define ERR_INTR_EN_DMA 0x00010000
-+#define ERR_INTR_EN_FPM 0x80000000
-+#define ERR_INTR_EN_BMI 0x00800000
-+#define ERR_INTR_EN_QMI 0x00400000
-+#define ERR_INTR_EN_PRS 0x00200000
-+#define ERR_INTR_EN_KG 0x00100000
-+#define ERR_INTR_EN_PLCR 0x00080000
-+#define ERR_INTR_EN_MURAM 0x00040000
-+#define ERR_INTR_EN_IRAM 0x00020000
-+#define ERR_INTR_EN_10G_MAC0 0x00008000
-+#define ERR_INTR_EN_10G_MAC1 0x00000040
-+#define ERR_INTR_EN_1G_MAC0 0x00004000
-+#define ERR_INTR_EN_1G_MAC1 0x00002000
-+#define ERR_INTR_EN_1G_MAC2 0x00001000
-+#define ERR_INTR_EN_1G_MAC3 0x00000800
-+#define ERR_INTR_EN_1G_MAC4 0x00000400
-+#define ERR_INTR_EN_1G_MAC5 0x00000200
-+#define ERR_INTR_EN_1G_MAC6 0x00000100
-+#define ERR_INTR_EN_1G_MAC7 0x00000080
-+#define ERR_INTR_EN_MACSEC_MAC0 0x00000001
-+
-+#define INTR_EN_QMI 0x40000000
-+#define INTR_EN_PRS 0x20000000
-+#define INTR_EN_WAKEUP 0x10000000
-+#define INTR_EN_PLCR 0x08000000
-+#define INTR_EN_1G_MAC0 0x00080000
-+#define INTR_EN_1G_MAC1 0x00040000
-+#define INTR_EN_1G_MAC2 0x00020000
-+#define INTR_EN_1G_MAC3 0x00010000
-+#define INTR_EN_1G_MAC4 0x00000040
-+#define INTR_EN_1G_MAC5 0x00000020
-+#define INTR_EN_1G_MAC6 0x00000008
-+#define INTR_EN_1G_MAC7 0x00000002
-+#define INTR_EN_10G_MAC0 0x00200000
-+#define INTR_EN_10G_MAC1 0x00100000
-+#define INTR_EN_REV0 0x00008000
-+#define INTR_EN_REV1 0x00004000
-+#define INTR_EN_REV2 0x00002000
-+#define INTR_EN_REV3 0x00001000
-+#define INTR_EN_BRK 0x00000080
-+#define INTR_EN_TMR 0x01000000
-+#define INTR_EN_MACSEC_MAC0 0x00000001
-+/* @} */
-+
-+/**************************************************************************//**
-+ @Description Memory Mapped Registers
-+*//***************************************************************************/
-+
-+#if defined(__MWERKS__) && !defined(__GNUC__)
-+#pragma pack(push,1)
-+#endif /* defined(__MWERKS__) && ... */
-+
-+typedef struct
-+{
-+ volatile uint32_t iadd; /**< FM IRAM instruction address register */
-+ volatile uint32_t idata; /**< FM IRAM instruction data register */
-+ volatile uint32_t itcfg; /**< FM IRAM timing config register */
-+ volatile uint32_t iready; /**< FM IRAM ready register */
-+ volatile uint32_t res[0x1FFFC];
-+} t_FMIramRegs;
-+
-+/* Trace buffer registers -
-+ each FM Controller has its own trace buffer residing at FM_MM_TRB(fmCtrlIndex) offset */
-+typedef struct t_FmTrbRegs
-+{
-+ volatile uint32_t tcrh;
-+ volatile uint32_t tcrl;
-+ volatile uint32_t tesr;
-+ volatile uint32_t tecr0h;
-+ volatile uint32_t tecr0l;
-+ volatile uint32_t terf0h;
-+ volatile uint32_t terf0l;
-+ volatile uint32_t tecr1h;
-+ volatile uint32_t tecr1l;
-+ volatile uint32_t terf1h;
-+ volatile uint32_t terf1l;
-+ volatile uint32_t tpcch;
-+ volatile uint32_t tpccl;
-+ volatile uint32_t tpc1h;
-+ volatile uint32_t tpc1l;
-+ volatile uint32_t tpc2h;
-+ volatile uint32_t tpc2l;
-+ volatile uint32_t twdimr;
-+ volatile uint32_t twicvr;
-+ volatile uint32_t tar;
-+ volatile uint32_t tdr;
-+ volatile uint32_t tsnum1;
-+ volatile uint32_t tsnum2;
-+ volatile uint32_t tsnum3;
-+ volatile uint32_t tsnum4;
-+} t_FmTrbRegs;
-+
-+#if defined(__MWERKS__) && !defined(__GNUC__)
-+#pragma pack(pop)
-+#endif /* defined(__MWERKS__) && ... */
-+
-+/**************************************************************************//**
-+ @Description General defines
-+*//***************************************************************************/
-+#define FM_DEBUG_STATUS_REGISTER_OFFSET 0x000d1084UL
-+#define FM_FW_DEBUG_INSTRUCTION 0x6ffff805UL
-+
-+/**************************************************************************//**
-+ @Description FPM defines
-+*//***************************************************************************/
-+/* masks */
-+#define FPM_BRKC_RDBG 0x00000200
-+#define FPM_BRKC_SLP 0x00000800
-+/**************************************************************************//**
-+ @Description BMI defines
-+*//***************************************************************************/
-+/* masks */
-+#define BMI_INIT_START 0x80000000
-+#define BMI_ERR_INTR_EN_STORAGE_PROFILE_ECC 0x80000000
-+#define BMI_ERR_INTR_EN_LIST_RAM_ECC 0x40000000
-+#define BMI_ERR_INTR_EN_STATISTICS_RAM_ECC 0x20000000
-+#define BMI_ERR_INTR_EN_DISPATCH_RAM_ECC 0x10000000
-+/**************************************************************************//**
-+ @Description QMI defines
-+*//***************************************************************************/
-+/* masks */
-+#define QMI_ERR_INTR_EN_DOUBLE_ECC 0x80000000
-+#define QMI_ERR_INTR_EN_DEQ_FROM_DEF 0x40000000
-+#define QMI_INTR_EN_SINGLE_ECC 0x80000000
-+
-+/**************************************************************************//**
-+ @Description IRAM defines
-+*//***************************************************************************/
-+/* masks */
-+#define IRAM_IADD_AIE 0x80000000
-+#define IRAM_READY 0x80000000
-+
-+/**************************************************************************//**
-+ @Description TRB defines
-+*//***************************************************************************/
-+/* masks */
-+#define TRB_TCRH_RESET 0x04000000
-+#define TRB_TCRH_ENABLE_COUNTERS 0x84008000
-+#define TRB_TCRH_DISABLE_COUNTERS 0x8400C000
-+#define TRB_TCRL_RESET 0x20000000
-+#define TRB_TCRL_UTIL 0x00000460
-+typedef struct {
-+ void (*f_Isr) (t_Handle h_Arg, uint32_t event);
-+ t_Handle h_SrcHandle;
-+} t_FmanCtrlIntrSrc;
-+
-+
-+typedef void (t_FmanCtrlIsr)( t_Handle h_Fm, uint32_t event);
-+
-+typedef struct
-+{
-+/***************************/
-+/* Master/Guest parameters */
-+/***************************/
-+ uint8_t fmId;
-+ e_FmPortType portsTypes[FM_MAX_NUM_OF_HW_PORT_IDS];
-+ uint16_t fmClkFreq;
-+ uint16_t fmMacClkFreq;
-+ t_FmRevisionInfo revInfo;
-+/**************************/
-+/* Master Only parameters */
-+/**************************/
-+ bool enabledTimeStamp;
-+ uint8_t count1MicroBit;
-+ uint8_t totalNumOfTasks;
-+ uint32_t totalFifoSize;
-+ uint8_t maxNumOfOpenDmas;
-+ uint8_t accumulatedNumOfTasks;
-+ uint32_t accumulatedFifoSize;
-+ uint8_t accumulatedNumOfOpenDmas;
-+ uint8_t accumulatedNumOfDeqTnums;
-+#ifdef FM_LOW_END_RESTRICTION
-+ bool lowEndRestriction;
-+#endif /* FM_LOW_END_RESTRICTION */
-+ uint32_t exceptions;
-+ int irq;
-+ int errIrq;
-+ bool ramsEccEnable;
-+ bool explicitEnable;
-+ bool internalCall;
-+ uint8_t ramsEccOwners;
-+ uint32_t extraFifoPoolSize;
-+ uint8_t extraTasksPoolSize;
-+ uint8_t extraOpenDmasPoolSize;
-+#if defined(FM_MAX_NUM_OF_10G_MACS) && (FM_MAX_NUM_OF_10G_MACS)
-+ uint16_t portMaxFrameLengths10G[FM_MAX_NUM_OF_10G_MACS];
-+ uint16_t macMaxFrameLengths10G[FM_MAX_NUM_OF_10G_MACS];
-+#endif /* defined(FM_MAX_NUM_OF_10G_MACS) && ... */
-+ uint16_t portMaxFrameLengths1G[FM_MAX_NUM_OF_1G_MACS];
-+ uint16_t macMaxFrameLengths1G[FM_MAX_NUM_OF_1G_MACS];
-+} t_FmStateStruct;
-+
-+#if (DPAA_VERSION >= 11)
-+typedef struct t_FmMapParam {
-+ uint16_t profilesBase;
-+ uint16_t numOfProfiles;
-+ t_Handle h_FmPort;
-+} t_FmMapParam;
-+
-+typedef struct t_FmAllocMng {
-+ bool allocated;
-+ uint8_t ownerId; /* guestId for KG in multi-partition only,
-+ portId for PLCR in any environment */
-+} t_FmAllocMng;
-+
-+typedef struct t_FmPcdSpEntry {
-+ bool valid;
-+ t_FmAllocMng profilesMng;
-+} t_FmPcdSpEntry;
-+
-+typedef struct t_FmSp {
-+ void *p_FmPcdStoragePrflRegs;
-+ t_FmPcdSpEntry profiles[FM_VSP_MAX_NUM_OF_ENTRIES];
-+ t_FmMapParam portsMapping[FM_MAX_NUM_OF_PORTS];
-+} t_FmSp;
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+typedef struct t_Fm
-+{
-+/***************************/
-+/* Master/Guest parameters */
-+/***************************/
-+/* locals for recovery */
-+ uintptr_t baseAddr;
-+
-+/* un-needed for recovery */
-+ t_Handle h_Pcd;
-+ char fmModuleName[MODULE_NAME_SIZE];
-+ char fmIpcHandlerModuleName[FM_MAX_NUM_OF_GUESTS][MODULE_NAME_SIZE];
-+ t_Handle h_IpcSessions[FM_MAX_NUM_OF_GUESTS];
-+ t_FmIntrSrc intrMng[e_FM_EV_DUMMY_LAST]; /* FM exceptions user callback */
-+ uint8_t guestId;
-+/**************************/
-+/* Master Only parameters */
-+/**************************/
-+/* locals for recovery */
-+ struct fman_fpm_regs *p_FmFpmRegs;
-+ struct fman_bmi_regs *p_FmBmiRegs;
-+ struct fman_qmi_regs *p_FmQmiRegs;
-+ struct fman_dma_regs *p_FmDmaRegs;
-+ struct fman_regs *p_FmRegs;
-+ t_FmExceptionsCallback *f_Exception;
-+ t_FmBusErrorCallback *f_BusError;
-+ t_Handle h_App; /* Application handle */
-+ t_Handle h_Spinlock;
-+ bool recoveryMode;
-+ t_FmStateStruct *p_FmStateStruct;
-+ uint16_t tnumAgingPeriod;
-+#if (DPAA_VERSION >= 11)
-+ t_FmSp *p_FmSp;
-+ uint8_t partNumOfVSPs;
-+ uint8_t partVSPBase;
-+ uintptr_t vspBaseAddr;
-+#endif /* (DPAA_VERSION >= 11) */
-+ bool portsPreFetchConfigured[FM_MAX_NUM_OF_HW_PORT_IDS]; /* Prefetch configration per Tx-port */
-+ bool portsPreFetchValue[FM_MAX_NUM_OF_HW_PORT_IDS]; /* Prefetch configration per Tx-port */
-+
-+/* un-needed for recovery */
-+ struct fman_cfg *p_FmDriverParam;
-+ t_Handle h_FmMuram;
-+ uint64_t fmMuramPhysBaseAddr;
-+ bool independentMode;
-+ bool hcPortInitialized;
-+ uintptr_t camBaseAddr; /* save for freeing */
-+ uintptr_t resAddr;
-+ uintptr_t fifoBaseAddr; /* save for freeing */
-+ t_FmanCtrlIntrSrc fmanCtrlIntr[FM_NUM_OF_FMAN_CTRL_EVENT_REGS]; /* FM exceptions user callback */
-+ bool usedEventRegs[FM_NUM_OF_FMAN_CTRL_EVENT_REGS];
-+ t_FmFirmwareParams firmware;
-+ bool fwVerify;
-+ bool resetOnInit;
-+ t_FmResetOnInitOverrideCallback *f_ResetOnInitOverride;
-+ uint32_t userSetExceptions;
-+} t_Fm;
-+
-+
-+#endif /* __FM_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/fm_ipc.h
-@@ -0,0 +1,465 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/**************************************************************************//**
-+ @File fm_ipc.h
-+
-+ @Description FM Inter-Partition prototypes, structures and definitions.
-+*//***************************************************************************/
-+#ifndef __FM_IPC_H
-+#define __FM_IPC_H
-+
-+#include "error_ext.h"
-+#include "std_ext.h"
-+
-+
-+/**************************************************************************//**
-+ @Group FM_grp Frame Manager API
-+
-+ @Description FM API functions, definitions and enums
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Group FM_IPC_grp FM Inter-Partition messaging Unit
-+
-+ @Description FM Inter-Partition messaging unit API definitions and enums.
-+
-+ @{
-+*//***************************************************************************/
-+
-+#if defined(__MWERKS__) && !defined(__GNUC__)
-+#pragma pack(push,1)
-+#endif /* defined(__MWERKS__) && ... */
-+
-+/**************************************************************************//**
-+ @Description enum for defining MAC types
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Description A structure of parameters for specifying a MAC.
-+*//***************************************************************************/
-+typedef _Packed struct
-+{
-+ uint8_t id;
-+ uint32_t enumType;
-+} _PackedType t_FmIpcMacParams;
-+
-+/**************************************************************************//**
-+ @Description A structure of parameters for specifying a MAC.
-+*//***************************************************************************/
-+typedef _Packed struct
-+{
-+ t_FmIpcMacParams macParams;
-+ uint16_t maxFrameLength;
-+} _PackedType t_FmIpcMacMaxFrameParams;
-+
-+/**************************************************************************//**
-+ @Description FM physical Address
-+*//***************************************************************************/
-+typedef _Packed struct t_FmIpcPhysAddr
-+{
-+ volatile uint8_t high;
-+ volatile uint32_t low;
-+} _PackedType t_FmIpcPhysAddr;
-+
-+
-+typedef _Packed struct t_FmIpcPortOutInitParams {
-+ uint8_t numOfTasks; /**< OUT */
-+ uint8_t numOfExtraTasks; /**< OUT */
-+ uint8_t numOfOpenDmas; /**< OUT */
-+ uint8_t numOfExtraOpenDmas; /**< OUT */
-+ uint32_t sizeOfFifo; /**< OUT */
-+ uint32_t extraSizeOfFifo; /**< OUT */
-+ t_FmIpcPhysAddr ipcPhysAddr; /**< OUT */
-+} _PackedType t_FmIpcPortOutInitParams;
-+
-+/**************************************************************************//**
-+ @Description Structure for IPC communication during FM_PORT_Init.
-+*//***************************************************************************/
-+typedef _Packed struct t_FmIpcPortInInitParams {
-+ uint8_t hardwarePortId; /**< IN. port Id */
-+ uint32_t enumPortType; /**< IN. Port type */
-+ uint8_t boolIndependentMode;/**< IN. TRUE if FM Port operates in independent mode */
-+ uint16_t liodnOffset; /**< IN. Port's requested resource */
-+ uint8_t numOfTasks; /**< IN. Port's requested resource */
-+ uint8_t numOfExtraTasks; /**< IN. Port's requested resource */
-+ uint8_t numOfOpenDmas; /**< IN. Port's requested resource */
-+ uint8_t numOfExtraOpenDmas; /**< IN. Port's requested resource */
-+ uint32_t sizeOfFifo; /**< IN. Port's requested resource */
-+ uint32_t extraSizeOfFifo; /**< IN. Port's requested resource */
-+ uint8_t deqPipelineDepth; /**< IN. Port's requested resource */
-+ uint16_t maxFrameLength; /**< IN. Port's max frame length. */
-+ uint16_t liodnBase; /**< IN. Irrelevant for P4080 rev 1.
-+ LIODN base for this port, to be
-+ used together with LIODN offset. */
-+} _PackedType t_FmIpcPortInInitParams;
-+
-+
-+/**************************************************************************//**
-+ @Description Structure for IPC communication between port and FM
-+ regarding tasks and open DMA resources management.
-+*//***************************************************************************/
-+typedef _Packed struct t_FmIpcPortRsrcParams {
-+ uint8_t hardwarePortId; /**< IN. port Id */
-+ uint32_t val; /**< IN. Port's requested resource */
-+ uint32_t extra; /**< IN. Port's requested resource */
-+ uint8_t boolInitialConfig;
-+} _PackedType t_FmIpcPortRsrcParams;
-+
-+
-+/**************************************************************************//**
-+ @Description Structure for IPC communication between port and FM
-+ regarding tasks and open DMA resources management.
-+*//***************************************************************************/
-+typedef _Packed struct t_FmIpcPortFifoParams {
-+ t_FmIpcPortRsrcParams rsrcParams;
-+ uint32_t enumPortType;
-+ uint8_t boolIndependentMode;
-+ uint8_t deqPipelineDepth;
-+ uint8_t numOfPools;
-+ uint16_t secondLargestBufSize;
-+ uint16_t largestBufSize;
-+ uint8_t boolInitialConfig;
-+} _PackedType t_FmIpcPortFifoParams;
-+
-+/**************************************************************************//**
-+ @Description Structure for port-FM communication during FM_PORT_Free.
-+*//***************************************************************************/
-+typedef _Packed struct t_FmIpcPortFreeParams {
-+ uint8_t hardwarePortId; /**< IN. port Id */
-+ uint32_t enumPortType; /**< IN. Port type */
-+ uint8_t deqPipelineDepth; /**< IN. Port's requested resource */
-+} _PackedType t_FmIpcPortFreeParams;
-+
-+/**************************************************************************//**
-+ @Description Structure for defining DMA status
-+*//***************************************************************************/
-+typedef _Packed struct t_FmIpcDmaStatus {
-+ uint8_t boolCmqNotEmpty; /**< Command queue is not empty */
-+ uint8_t boolBusError; /**< Bus error occurred */
-+ uint8_t boolReadBufEccError; /**< Double ECC error on buffer Read */
-+ uint8_t boolWriteBufEccSysError; /**< Double ECC error on buffer write from system side */
-+ uint8_t boolWriteBufEccFmError; /**< Double ECC error on buffer write from FM side */
-+ uint8_t boolSinglePortEccError; /**< Single port ECC error from FM side */
-+} _PackedType t_FmIpcDmaStatus;
-+
-+typedef _Packed struct t_FmIpcRegisterIntr
-+{
-+ uint8_t guestId; /* IN */
-+ uint32_t event; /* IN */
-+} _PackedType t_FmIpcRegisterIntr;
-+
-+typedef _Packed struct t_FmIpcIsr
-+{
-+ uint8_t boolErr; /* IN */
-+ uint32_t pendingReg; /* IN */
-+} _PackedType t_FmIpcIsr;
-+
-+/**************************************************************************//**
-+ @Description structure for returning FM parameters
-+*//***************************************************************************/
-+typedef _Packed struct t_FmIpcParams {
-+ uint16_t fmClkFreq; /**< OUT: FM Clock frequency */
-+ uint16_t fmMacClkFreq; /**< OUT: FM MAC clock frequence */
-+ uint8_t majorRev; /**< OUT: FM Major revision */
-+ uint8_t minorRev; /**< OUT: FM Minor revision */
-+} _PackedType t_FmIpcParams;
-+
-+
-+/**************************************************************************//**
-+ @Description structure for returning Fman Ctrl Code revision information
-+*//***************************************************************************/
-+typedef _Packed struct t_FmIpcFmanCtrlCodeRevisionInfo {
-+ uint16_t packageRev; /**< OUT: Package revision */
-+ uint8_t majorRev; /**< OUT: Major revision */
-+ uint8_t minorRev; /**< OUT: Minor revision */
-+} _PackedType t_FmIpcFmanCtrlCodeRevisionInfo;
-+
-+/**************************************************************************//**
-+ @Description Structure for defining Fm number of Fman controlers
-+*//***************************************************************************/
-+typedef _Packed struct t_FmIpcPortNumOfFmanCtrls {
-+ uint8_t hardwarePortId; /**< IN. port Id */
-+ uint8_t numOfFmanCtrls; /**< IN. Port type */
-+ t_FmFmanCtrl orFmanCtrl; /**< IN. fman controller for order restoration*/
-+} t_FmIpcPortNumOfFmanCtrls;
-+
-+/**************************************************************************//**
-+ @Description structure for setting Fman contriller events
-+*//***************************************************************************/
-+typedef _Packed struct t_FmIpcFmanEvents {
-+ uint8_t eventRegId; /**< IN: Fman controller event register id */
-+ uint32_t enableEvents; /**< IN/OUT: required enabled events mask */
-+} _PackedType t_FmIpcFmanEvents;
-+
-+typedef _Packed struct t_FmIpcResourceAllocParams {
-+ uint8_t guestId;
-+ uint16_t base;
-+ uint16_t num;
-+}_PackedType t_FmIpcResourceAllocParams;
-+
-+typedef _Packed struct t_FmIpcVspSetPortWindow {
-+ uint8_t hardwarePortId;
-+ uint8_t baseStorageProfile;
-+ uint8_t log2NumOfProfiles;
-+}_PackedType t_FmIpcVspSetPortWindow;
-+
-+typedef _Packed struct t_FmIpcSetCongestionGroupPfcPriority {
-+ uint32_t congestionGroupId;
-+ uint8_t priorityBitMap;
-+}_PackedType t_FmIpcSetCongestionGroupPfcPriority;
-+
-+#define FM_IPC_MAX_REPLY_BODY_SIZE 20
-+#define FM_IPC_MAX_REPLY_SIZE (FM_IPC_MAX_REPLY_BODY_SIZE + sizeof(uint32_t))
-+#define FM_IPC_MAX_MSG_SIZE 30
-+
-+typedef _Packed struct t_FmIpcMsg
-+{
-+ uint32_t msgId;
-+ uint8_t msgBody[FM_IPC_MAX_MSG_SIZE];
-+} _PackedType t_FmIpcMsg;
-+
-+typedef _Packed struct t_FmIpcReply
-+{
-+ uint32_t error;
-+ uint8_t replyBody[FM_IPC_MAX_REPLY_BODY_SIZE];
-+} _PackedType t_FmIpcReply;
-+
-+#if defined(__MWERKS__) && !defined(__GNUC__)
-+#pragma pack(pop)
-+#endif /* defined(__MWERKS__) && ... */
-+
-+
-+/***************************************************************************/
-+/************************ FRONT-END-TO-BACK-END*****************************/
-+/***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Function FM_GET_TIMESTAMP_SCALE
-+
-+ @Description Used by FM front-end.
-+
-+ @Param[out] uint32_t Pointer
-+*//***************************************************************************/
-+#define FM_GET_TIMESTAMP_SCALE 1
-+
-+/**************************************************************************//**
-+ @Function FM_GET_COUNTER
-+
-+ @Description Used by FM front-end.
-+
-+ @Param[in/out] t_FmIpcGetCounter Pointer
-+*//***************************************************************************/
-+#define FM_GET_COUNTER 2
-+
-+/**************************************************************************//**
-+ @Function FM_GET_SET_PORT_PARAMS
-+
-+ @Description Used by FM front-end for the PORT module in order to set and get
-+ parameters in/from master FM module on FM PORT initialization time.
-+
-+ @Param[in/out] t_FmIcPortInitParams Pointer
-+*//***************************************************************************/
-+#define FM_GET_SET_PORT_PARAMS 4
-+
-+/**************************************************************************//**
-+ @Function FM_FREE_PORT
-+
-+ @Description Used by FM front-end for the PORT module when a port is freed
-+ to free all FM PORT resources.
-+
-+ @Param[in] uint8_t Pointer
-+*//***************************************************************************/
-+#define FM_FREE_PORT 5
-+
-+/**************************************************************************//**
-+ @Function FM_RESET_MAC
-+
-+ @Description Used by front-end for the MAC module to reset the MAC registers
-+
-+ @Param[in] t_FmIpcMacParams Pointer .
-+*//***************************************************************************/
-+#define FM_RESET_MAC 6
-+
-+/**************************************************************************//**
-+ @Function FM_RESUME_STALLED_PORT
-+
-+ @Description Used by FM front-end for the PORT module in order to
-+ release a stalled FM Port.
-+
-+ @Param[in] uint8_t Pointer
-+*//***************************************************************************/
-+#define FM_RESUME_STALLED_PORT 7
-+
-+/**************************************************************************//**
-+ @Function FM_IS_PORT_STALLED
-+
-+ @Description Used by FM front-end for the PORT module in order to check whether
-+ an FM port is stalled.
-+
-+ @Param[in/out] t_FmIcPortIsStalled Pointer
-+*//***************************************************************************/
-+#define FM_IS_PORT_STALLED 8
-+
-+/**************************************************************************//**
-+ @Function FM_GET_PARAMS
-+
-+ @Description Used by FM front-end for the PORT module in order to dump
-+ return FM parameters.
-+
-+ @Param[in] uint8_t Pointer
-+*//***************************************************************************/
-+#define FM_GET_PARAMS 10
-+
-+/**************************************************************************//**
-+ @Function FM_REGISTER_INTR
-+
-+ @Description Used by FM front-end to register an interrupt handler to
-+ be called upon interrupt for guest.
-+
-+ @Param[out] t_FmIpcRegisterIntr Pointer
-+*//***************************************************************************/
-+#define FM_REGISTER_INTR 11
-+
-+/**************************************************************************//**
-+ @Function FM_DMA_STAT
-+
-+ @Description Used by FM front-end to read the FM DMA status.
-+
-+ @Param[out] t_FmIpcDmaStatus Pointer
-+*//***************************************************************************/
-+#define FM_DMA_STAT 13
-+
-+/**************************************************************************//**
-+ @Function FM_ALLOC_FMAN_CTRL_EVENT_REG
-+
-+ @Description Used by FM front-end to allocate event register.
-+
-+ @Param[out] Event register id Pointer
-+*//***************************************************************************/
-+#define FM_ALLOC_FMAN_CTRL_EVENT_REG 14
-+
-+/**************************************************************************//**
-+ @Function FM_FREE_FMAN_CTRL_EVENT_REG
-+
-+ @Description Used by FM front-end to free locate event register.
-+
-+ @Param[in] uint8_t Pointer - Event register id
-+*//***************************************************************************/
-+#define FM_FREE_FMAN_CTRL_EVENT_REG 15
-+
-+/**************************************************************************//**
-+ @Function FM_SET_FMAN_CTRL_EVENTS_ENABLE
-+
-+ @Description Used by FM front-end to enable events in the FPM
-+ Fman controller event register.
-+
-+ @Param[in] t_FmIpcFmanEvents Pointer
-+*//***************************************************************************/
-+#define FM_SET_FMAN_CTRL_EVENTS_ENABLE 16
-+
-+/**************************************************************************//**
-+ @Function FM_SET_FMAN_CTRL_EVENTS_ENABLE
-+
-+ @Description Used by FM front-end to enable events in the FPM
-+ Fman controller event register.
-+
-+ @Param[in/out] t_FmIpcFmanEvents Pointer
-+*//***************************************************************************/
-+#define FM_GET_FMAN_CTRL_EVENTS_ENABLE 17
-+
-+/**************************************************************************//**
-+ @Function FM_SET_MAC_MAX_FRAME
-+
-+ @Description Used by FM front-end to set MAC's MTU/RTU's in
-+ back-end.
-+
-+ @Param[in/out] t_FmIpcMacMaxFrameParams Pointer
-+*//***************************************************************************/
-+#define FM_SET_MAC_MAX_FRAME 18
-+
-+/**************************************************************************//**
-+ @Function FM_GET_PHYS_MURAM_BASE
-+
-+ @Description Used by FM front-end in order to get MURAM base address
-+
-+ @Param[in/out] t_FmIpcPhysAddr Pointer
-+*//***************************************************************************/
-+#define FM_GET_PHYS_MURAM_BASE 19
-+
-+/**************************************************************************//**
-+ @Function FM_MASTER_IS_ALIVE
-+
-+ @Description Used by FM front-end in order to verify Master is up
-+
-+ @Param[in/out] bool
-+*//***************************************************************************/
-+#define FM_MASTER_IS_ALIVE 20
-+
-+#define FM_ENABLE_RAM_ECC 21
-+#define FM_DISABLE_RAM_ECC 22
-+#define FM_SET_NUM_OF_FMAN_CTRL 23
-+#define FM_SET_SIZE_OF_FIFO 24
-+#define FM_SET_NUM_OF_TASKS 25
-+#define FM_SET_NUM_OF_OPEN_DMAS 26
-+#define FM_VSP_ALLOC 27
-+#define FM_VSP_FREE 28
-+#define FM_VSP_SET_PORT_WINDOW 29
-+#define FM_GET_FMAN_CTRL_CODE_REV 30
-+#define FM_SET_CONG_GRP_PFC_PRIO 31
-+#ifdef FM_TX_ECC_FRMS_ERRATA_10GMAC_A004
-+#define FM_10G_TX_ECC_WA 100
-+#endif /* FM_TX_ECC_FRMS_ERRATA_10GMAC_A004 */
-+
-+/***************************************************************************/
-+/************************ BACK-END-TO-FRONT-END*****************************/
-+/***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Function FM_GUEST_ISR
-+
-+ @Description Used by FM back-end to report an interrupt to the front-end.
-+
-+ @Param[out] t_FmIpcIsr Pointer
-+*//***************************************************************************/
-+#define FM_GUEST_ISR 1
-+
-+
-+
-+/** @} */ /* end of FM_IPC_grp group */
-+/** @} */ /* end of FM_grp group */
-+
-+
-+#endif /* __FM_IPC_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/fm_muram.c
-@@ -0,0 +1,174 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/******************************************************************************
-+ @File FM_muram.c
-+
-+ @Description FM MURAM ...
-+*//***************************************************************************/
-+#include "error_ext.h"
-+#include "std_ext.h"
-+#include "mm_ext.h"
-+#include "string_ext.h"
-+#include "sprint_ext.h"
-+#include "fm_muram_ext.h"
-+#include "fm_common.h"
-+
-+#define __ERR_MODULE__ MODULE_FM_MURAM
-+
-+
-+typedef struct
-+{
-+ t_Handle h_Mem;
-+ uintptr_t baseAddr;
-+ uint32_t size;
-+} t_FmMuram;
-+
-+
-+void FmMuramClear(t_Handle h_FmMuram)
-+{
-+ t_FmMuram *p_FmMuram = ( t_FmMuram *)h_FmMuram;
-+
-+ SANITY_CHECK_RETURN(h_FmMuram, E_INVALID_HANDLE);
-+ IOMemSet32(UINT_TO_PTR(p_FmMuram->baseAddr), 0, p_FmMuram->size);
-+}
-+
-+
-+t_Handle FM_MURAM_ConfigAndInit(uintptr_t baseAddress, uint32_t size)
-+{
-+ t_Handle h_Mem;
-+ t_FmMuram *p_FmMuram;
-+
-+ if (!baseAddress)
-+ {
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE, ("baseAddress 0 is not supported"));
-+ return NULL;
-+ }
-+
-+ if (baseAddress%4)
-+ {
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE, ("baseAddress not 4 bytes aligned!"));
-+ return NULL;
-+ }
-+
-+ /* Allocate FM MURAM structure */
-+ p_FmMuram = (t_FmMuram *) XX_Malloc(sizeof(t_FmMuram));
-+ if (!p_FmMuram)
-+ {
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("FM MURAM driver structure"));
-+ return NULL;
-+ }
-+ memset(p_FmMuram, 0, sizeof(t_FmMuram));
-+
-+
-+ if ((MM_Init(&h_Mem, baseAddress, size) != E_OK) || (!h_Mem))
-+ {
-+ XX_Free(p_FmMuram);
-+ REPORT_ERROR(MAJOR, E_INVALID_HANDLE, ("FM-MURAM partition!!!"));
-+ return NULL;
-+ }
-+
-+ /* Initialize FM MURAM parameters which will be kept by the driver */
-+ p_FmMuram->baseAddr = baseAddress;
-+ p_FmMuram->size = size;
-+ p_FmMuram->h_Mem = h_Mem;
-+
-+ return p_FmMuram;
-+}
-+
-+t_Error FM_MURAM_Free(t_Handle h_FmMuram)
-+{
-+ t_FmMuram *p_FmMuram = ( t_FmMuram *)h_FmMuram;
-+
-+ if (p_FmMuram->h_Mem)
-+ MM_Free(p_FmMuram->h_Mem);
-+
-+ XX_Free(h_FmMuram);
-+
-+ return E_OK;
-+}
-+
-+void * FM_MURAM_AllocMem(t_Handle h_FmMuram, uint32_t size, uint32_t align)
-+{
-+ t_FmMuram *p_FmMuram = ( t_FmMuram *)h_FmMuram;
-+ uintptr_t addr;
-+
-+ SANITY_CHECK_RETURN_VALUE(h_FmMuram, E_INVALID_HANDLE, NULL);
-+ SANITY_CHECK_RETURN_VALUE(p_FmMuram->h_Mem, E_INVALID_HANDLE, NULL);
-+
-+ addr = (uintptr_t)MM_Get(p_FmMuram->h_Mem, size, align ,"FM MURAM");
-+
-+ if (addr == ILLEGAL_BASE)
-+ return NULL;
-+
-+ return UINT_TO_PTR(addr);
-+}
-+
-+void * FM_MURAM_AllocMemForce(t_Handle h_FmMuram, uint64_t base, uint32_t size)
-+{
-+ t_FmMuram *p_FmMuram = ( t_FmMuram *)h_FmMuram;
-+ uintptr_t addr;
-+
-+ SANITY_CHECK_RETURN_VALUE(h_FmMuram, E_INVALID_HANDLE, NULL);
-+ SANITY_CHECK_RETURN_VALUE(p_FmMuram->h_Mem, E_INVALID_HANDLE, NULL);
-+
-+ addr = (uintptr_t)MM_GetForce(p_FmMuram->h_Mem, base, size, "FM MURAM");
-+
-+ if (addr == ILLEGAL_BASE)
-+ return NULL;
-+
-+ return UINT_TO_PTR(addr);
-+}
-+
-+t_Error FM_MURAM_FreeMem(t_Handle h_FmMuram, void *ptr)
-+{
-+ t_FmMuram *p_FmMuram = ( t_FmMuram *)h_FmMuram;
-+
-+ SANITY_CHECK_RETURN_ERROR(h_FmMuram, E_INVALID_HANDLE);
-+ SANITY_CHECK_RETURN_ERROR(p_FmMuram->h_Mem, E_INVALID_HANDLE);
-+
-+ if (MM_Put(p_FmMuram->h_Mem, PTR_TO_UINT(ptr)) == 0)
-+ RETURN_ERROR(MINOR, E_INVALID_ADDRESS, ("memory pointer!!!"));
-+
-+ return E_OK;
-+}
-+
-+uint64_t FM_MURAM_GetFreeMemSize(t_Handle h_FmMuram)
-+{
-+ t_FmMuram *p_FmMuram = ( t_FmMuram *)h_FmMuram;
-+
-+ SANITY_CHECK_RETURN_VALUE(h_FmMuram, E_INVALID_HANDLE, 0);
-+ SANITY_CHECK_RETURN_VALUE(p_FmMuram->h_Mem, E_INVALID_HANDLE, 0);
-+
-+ return MM_GetFreeMemSize(p_FmMuram->h_Mem);
-+}
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/fman.c
-@@ -0,0 +1,1398 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+#include <linux/math64.h>
-+#include "fsl_fman.h"
-+#include "dpaa_integration_ext.h"
-+
-+uint32_t fman_get_bmi_err_event(struct fman_bmi_regs *bmi_rg)
-+{
-+ uint32_t event, mask, force;
-+
-+ event = ioread32be(&bmi_rg->fmbm_ievr);
-+ mask = ioread32be(&bmi_rg->fmbm_ier);
-+ event &= mask;
-+ /* clear the forced events */
-+ force = ioread32be(&bmi_rg->fmbm_ifr);
-+ if (force & event)
-+ iowrite32be(force & ~event, &bmi_rg->fmbm_ifr);
-+ /* clear the acknowledged events */
-+ iowrite32be(event, &bmi_rg->fmbm_ievr);
-+ return event;
-+}
-+
-+uint32_t fman_get_qmi_err_event(struct fman_qmi_regs *qmi_rg)
-+{
-+ uint32_t event, mask, force;
-+
-+ event = ioread32be(&qmi_rg->fmqm_eie);
-+ mask = ioread32be(&qmi_rg->fmqm_eien);
-+ event &= mask;
-+
-+ /* clear the forced events */
-+ force = ioread32be(&qmi_rg->fmqm_eif);
-+ if (force & event)
-+ iowrite32be(force & ~event, &qmi_rg->fmqm_eif);
-+ /* clear the acknowledged events */
-+ iowrite32be(event, &qmi_rg->fmqm_eie);
-+ return event;
-+}
-+
-+uint32_t fman_get_dma_com_id(struct fman_dma_regs *dma_rg)
-+{
-+ return ioread32be(&dma_rg->fmdmtcid);
-+}
-+
-+uint64_t fman_get_dma_addr(struct fman_dma_regs *dma_rg)
-+{
-+ uint64_t addr;
-+
-+ addr = (uint64_t)ioread32be(&dma_rg->fmdmtal);
-+ addr |= ((uint64_t)(ioread32be(&dma_rg->fmdmtah)) << 32);
-+
-+ return addr;
-+}
-+
-+uint32_t fman_get_dma_err_event(struct fman_dma_regs *dma_rg)
-+{
-+ uint32_t status, mask;
-+
-+ status = ioread32be(&dma_rg->fmdmsr);
-+ mask = ioread32be(&dma_rg->fmdmmr);
-+
-+ /* clear DMA_STATUS_BUS_ERR if mask has no DMA_MODE_BER */
-+ if ((mask & DMA_MODE_BER) != DMA_MODE_BER)
-+ status &= ~DMA_STATUS_BUS_ERR;
-+
-+ /* clear relevant bits if mask has no DMA_MODE_ECC */
-+ if ((mask & DMA_MODE_ECC) != DMA_MODE_ECC)
-+ status &= ~(DMA_STATUS_FM_SPDAT_ECC |
-+ DMA_STATUS_READ_ECC |
-+ DMA_STATUS_SYSTEM_WRITE_ECC |
-+ DMA_STATUS_FM_WRITE_ECC);
-+
-+ /* clear set events */
-+ iowrite32be(status, &dma_rg->fmdmsr);
-+
-+ return status;
-+}
-+
-+uint32_t fman_get_fpm_err_event(struct fman_fpm_regs *fpm_rg)
-+{
-+ uint32_t event;
-+
-+ event = ioread32be(&fpm_rg->fmfp_ee);
-+ /* clear the all occurred events */
-+ iowrite32be(event, &fpm_rg->fmfp_ee);
-+ return event;
-+}
-+
-+uint32_t fman_get_muram_err_event(struct fman_fpm_regs *fpm_rg)
-+{
-+ uint32_t event, mask;
-+
-+ event = ioread32be(&fpm_rg->fm_rcr);
-+ mask = ioread32be(&fpm_rg->fm_rie);
-+
-+ /* clear MURAM event bit (do not clear IRAM event) */
-+ iowrite32be(event & ~FPM_RAM_IRAM_ECC, &fpm_rg->fm_rcr);
-+
-+ if ((mask & FPM_MURAM_ECC_ERR_EX_EN))
-+ return event;
-+ else
-+ return 0;
-+}
-+
-+uint32_t fman_get_iram_err_event(struct fman_fpm_regs *fpm_rg)
-+{
-+ uint32_t event, mask;
-+
-+ event = ioread32be(&fpm_rg->fm_rcr) ;
-+ mask = ioread32be(&fpm_rg->fm_rie);
-+ /* clear IRAM event bit (do not clear MURAM event) */
-+ iowrite32be(event & ~FPM_RAM_MURAM_ECC,
-+ &fpm_rg->fm_rcr);
-+
-+ if ((mask & FPM_IRAM_ECC_ERR_EX_EN))
-+ return event;
-+ else
-+ return 0;
-+}
-+
-+uint32_t fman_get_qmi_event(struct fman_qmi_regs *qmi_rg)
-+{
-+ uint32_t event, mask, force;
-+
-+ event = ioread32be(&qmi_rg->fmqm_ie);
-+ mask = ioread32be(&qmi_rg->fmqm_ien);
-+ event &= mask;
-+ /* clear the forced events */
-+ force = ioread32be(&qmi_rg->fmqm_if);
-+ if (force & event)
-+ iowrite32be(force & ~event, &qmi_rg->fmqm_if);
-+ /* clear the acknowledged events */
-+ iowrite32be(event, &qmi_rg->fmqm_ie);
-+ return event;
-+}
-+
-+void fman_enable_time_stamp(struct fman_fpm_regs *fpm_rg,
-+ uint8_t count1ubit,
-+ uint16_t fm_clk_freq)
-+{
-+ uint32_t tmp;
-+ uint64_t frac;
-+ uint32_t intgr;
-+ uint32_t ts_freq = (uint32_t)(1 << count1ubit); /* in Mhz */
-+
-+ /* configure timestamp so that bit 8 will count 1 microsecond
-+ * Find effective count rate at TIMESTAMP least significant bits:
-+ * Effective_Count_Rate = 1MHz x 2^8 = 256MHz
-+ * Find frequency ratio between effective count rate and the clock:
-+ * Effective_Count_Rate / CLK e.g. for 600 MHz clock:
-+ * 256/600 = 0.4266666... */
-+
-+ intgr = ts_freq / fm_clk_freq;
-+ /* we multiply by 2^16 to keep the fraction of the division
-+ * we do not div back, since we write this value as a fraction
-+ * see spec */
-+
-+ frac = ((uint64_t)ts_freq << 16) - ((uint64_t)intgr << 16) * fm_clk_freq;
-+ /* we check remainder of the division in order to round up if not int */
-+ if (do_div(frac, fm_clk_freq))
-+ frac++;
-+
-+ tmp = (intgr << FPM_TS_INT_SHIFT) | (uint16_t)frac;
-+ iowrite32be(tmp, &fpm_rg->fmfp_tsc2);
-+
-+ /* enable timestamp with original clock */
-+ iowrite32be(FPM_TS_CTL_EN, &fpm_rg->fmfp_tsc1);
-+}
-+
-+uint32_t fman_get_fpm_error_interrupts(struct fman_fpm_regs *fpm_rg)
-+{
-+ return ioread32be(&fpm_rg->fm_epi);
-+}
-+
-+
-+int fman_set_erratum_10gmac_a004_wa(struct fman_fpm_regs *fpm_rg)
-+{
-+ int timeout = 100;
-+
-+ iowrite32be(0x40000000, &fpm_rg->fmfp_extc);
-+
-+ while ((ioread32be(&fpm_rg->fmfp_extc) & 0x40000000) && --timeout)
-+ udelay(10);
-+
-+ if (!timeout)
-+ return -EBUSY;
-+ return 0;
-+}
-+
-+void fman_set_ctrl_intr(struct fman_fpm_regs *fpm_rg,
-+ uint8_t event_reg_id,
-+ uint32_t enable_events)
-+{
-+ iowrite32be(enable_events, &fpm_rg->fmfp_cee[event_reg_id]);
-+}
-+
-+uint32_t fman_get_ctrl_intr(struct fman_fpm_regs *fpm_rg, uint8_t event_reg_id)
-+{
-+ return ioread32be(&fpm_rg->fmfp_cee[event_reg_id]);
-+}
-+
-+void fman_set_num_of_riscs_per_port(struct fman_fpm_regs *fpm_rg,
-+ uint8_t port_id,
-+ uint8_t num_fman_ctrls,
-+ uint32_t or_fman_ctrl)
-+{
-+ uint32_t tmp = 0;
-+
-+ tmp = (uint32_t)(port_id << FPM_PORT_FM_CTL_PORTID_SHIFT);
-+ /*TODO - maybe to put CTL# according to another criteria*/
-+ if (num_fman_ctrls == 2)
-+ tmp = FPM_PRT_FM_CTL2 | FPM_PRT_FM_CTL1;
-+ /* order restoration */
-+ tmp |= (or_fman_ctrl << FPM_PRC_ORA_FM_CTL_SEL_SHIFT) | or_fman_ctrl;
-+
-+ iowrite32be(tmp, &fpm_rg->fmfp_prc);
-+}
-+
-+void fman_set_order_restoration_per_port(struct fman_fpm_regs *fpm_rg,
-+ uint8_t port_id,
-+ bool independent_mode,
-+ bool is_rx_port)
-+{
-+ uint32_t tmp = 0;
-+
-+ tmp = (uint32_t)(port_id << FPM_PORT_FM_CTL_PORTID_SHIFT);
-+ if (independent_mode) {
-+ if (is_rx_port)
-+ tmp |= (FPM_PRT_FM_CTL1 <<
-+ FPM_PRC_ORA_FM_CTL_SEL_SHIFT) | FPM_PRT_FM_CTL1;
-+ else
-+ tmp |= (FPM_PRT_FM_CTL2 <<
-+ FPM_PRC_ORA_FM_CTL_SEL_SHIFT) | FPM_PRT_FM_CTL2;
-+ } else {
-+ tmp |= (FPM_PRT_FM_CTL2|FPM_PRT_FM_CTL1);
-+
-+ /* order restoration */
-+ if (port_id % 2)
-+ tmp |= (FPM_PRT_FM_CTL1 <<
-+ FPM_PRC_ORA_FM_CTL_SEL_SHIFT);
-+ else
-+ tmp |= (FPM_PRT_FM_CTL2 <<
-+ FPM_PRC_ORA_FM_CTL_SEL_SHIFT);
-+ }
-+ iowrite32be(tmp, &fpm_rg->fmfp_prc);
-+}
-+
-+uint8_t fman_get_qmi_deq_th(struct fman_qmi_regs *qmi_rg)
-+{
-+ return (uint8_t)ioread32be(&qmi_rg->fmqm_gc);
-+}
-+
-+uint8_t fman_get_qmi_enq_th(struct fman_qmi_regs *qmi_rg)
-+{
-+ return (uint8_t)(ioread32be(&qmi_rg->fmqm_gc) >> 8);
-+}
-+
-+void fman_set_qmi_enq_th(struct fman_qmi_regs *qmi_rg, uint8_t val)
-+{
-+ uint32_t tmp_reg;
-+
-+ tmp_reg = ioread32be(&qmi_rg->fmqm_gc);
-+ tmp_reg &= ~QMI_CFG_ENQ_MASK;
-+ tmp_reg |= ((uint32_t)val << 8);
-+ iowrite32be(tmp_reg, &qmi_rg->fmqm_gc);
-+}
-+
-+void fman_set_qmi_deq_th(struct fman_qmi_regs *qmi_rg, uint8_t val)
-+{
-+ uint32_t tmp_reg;
-+
-+ tmp_reg = ioread32be(&qmi_rg->fmqm_gc);
-+ tmp_reg &= ~QMI_CFG_DEQ_MASK;
-+ tmp_reg |= (uint32_t)val;
-+ iowrite32be(tmp_reg, &qmi_rg->fmqm_gc);
-+}
-+
-+void fman_qmi_disable_dispatch_limit(struct fman_fpm_regs *fpm_rg)
-+{
-+ iowrite32be(0, &fpm_rg->fmfp_mxd);
-+}
-+
-+void fman_set_liodn_per_port(struct fman_rg *fman_rg, uint8_t port_id,
-+ uint16_t liodn_base,
-+ uint16_t liodn_ofst)
-+{
-+ uint32_t tmp;
-+
-+ if ((port_id > 63) || (port_id < 1))
-+ return;
-+
-+ /* set LIODN base for this port */
-+ tmp = ioread32be(&fman_rg->dma_rg->fmdmplr[port_id / 2]);
-+ if (port_id % 2) {
-+ tmp &= ~FM_LIODN_BASE_MASK;
-+ tmp |= (uint32_t)liodn_base;
-+ } else {
-+ tmp &= ~(FM_LIODN_BASE_MASK << DMA_LIODN_SHIFT);
-+ tmp |= (uint32_t)liodn_base << DMA_LIODN_SHIFT;
-+ }
-+ iowrite32be(tmp, &fman_rg->dma_rg->fmdmplr[port_id / 2]);
-+ iowrite32be((uint32_t)liodn_ofst,
-+ &fman_rg->bmi_rg->fmbm_spliodn[port_id - 1]);
-+}
-+
-+bool fman_is_port_stalled(struct fman_fpm_regs *fpm_rg, uint8_t port_id)
-+{
-+ return (bool)!!(ioread32be(&fpm_rg->fmfp_ps[port_id]) & FPM_PS_STALLED);
-+}
-+
-+void fman_resume_stalled_port(struct fman_fpm_regs *fpm_rg, uint8_t port_id)
-+{
-+ uint32_t tmp;
-+
-+ tmp = (uint32_t)((port_id << FPM_PORT_FM_CTL_PORTID_SHIFT) |
-+ FPM_PRC_REALSE_STALLED);
-+ iowrite32be(tmp, &fpm_rg->fmfp_prc);
-+}
-+
-+int fman_reset_mac(struct fman_fpm_regs *fpm_rg, uint8_t mac_id, bool is_10g)
-+{
-+ uint32_t msk, timeout = 100;
-+
-+ /* Get the relevant bit mask */
-+ if (is_10g) {
-+ switch (mac_id) {
-+ case(0):
-+ msk = FPM_RSTC_10G0_RESET;
-+ break;
-+ case(1):
-+ msk = FPM_RSTC_10G1_RESET;
-+ break;
-+ default:
-+ return -EINVAL;
-+ }
-+ } else {
-+ switch (mac_id) {
-+ case(0):
-+ msk = FPM_RSTC_1G0_RESET;
-+ break;
-+ case(1):
-+ msk = FPM_RSTC_1G1_RESET;
-+ break;
-+ case(2):
-+ msk = FPM_RSTC_1G2_RESET;
-+ break;
-+ case(3):
-+ msk = FPM_RSTC_1G3_RESET;
-+ break;
-+ case(4):
-+ msk = FPM_RSTC_1G4_RESET;
-+ break;
-+ case (5):
-+ msk = FPM_RSTC_1G5_RESET;
-+ break;
-+ case (6):
-+ msk = FPM_RSTC_1G6_RESET;
-+ break;
-+ case (7):
-+ msk = FPM_RSTC_1G7_RESET;
-+ break;
-+ default:
-+ return -EINVAL;
-+ }
-+ }
-+ /* reset */
-+ iowrite32be(msk, &fpm_rg->fm_rstc);
-+ while ((ioread32be(&fpm_rg->fm_rstc) & msk) && --timeout)
-+ udelay(10);
-+
-+ if (!timeout)
-+ return -EBUSY;
-+ return 0;
-+}
-+
-+uint16_t fman_get_size_of_fifo(struct fman_bmi_regs *bmi_rg, uint8_t port_id)
-+{
-+ uint32_t tmp_reg;
-+
-+ if ((port_id > 63) || (port_id < 1))
-+ return 0;
-+
-+ tmp_reg = ioread32be(&bmi_rg->fmbm_pfs[port_id - 1]);
-+ return (uint16_t)((tmp_reg & BMI_FIFO_SIZE_MASK) + 1);
-+}
-+
-+uint32_t fman_get_total_fifo_size(struct fman_bmi_regs *bmi_rg)
-+{
-+ uint32_t reg, res;
-+
-+ reg = ioread32be(&bmi_rg->fmbm_cfg1);
-+ res = (reg >> BMI_CFG1_FIFO_SIZE_SHIFT) & 0x3ff;
-+ return res * FMAN_BMI_FIFO_UNITS;
-+}
-+
-+uint16_t fman_get_size_of_extra_fifo(struct fman_bmi_regs *bmi_rg,
-+ uint8_t port_id)
-+{
-+ uint32_t tmp_reg;
-+
-+ if ((port_id > 63) || (port_id < 1))
-+ return 0;
-+
-+ tmp_reg = ioread32be(&bmi_rg->fmbm_pfs[port_id-1]);
-+ return (uint16_t)((tmp_reg & BMI_EXTRA_FIFO_SIZE_MASK) >>
-+ BMI_EXTRA_FIFO_SIZE_SHIFT);
-+}
-+
-+void fman_set_size_of_fifo(struct fman_bmi_regs *bmi_rg,
-+ uint8_t port_id,
-+ uint32_t sz_fifo,
-+ uint32_t extra_sz_fifo)
-+{
-+ uint32_t tmp;
-+
-+ if ((port_id > 63) || (port_id < 1))
-+ return;
-+
-+ /* calculate reg */
-+ tmp = (uint32_t)((sz_fifo / FMAN_BMI_FIFO_UNITS - 1) |
-+ ((extra_sz_fifo / FMAN_BMI_FIFO_UNITS) <<
-+ BMI_EXTRA_FIFO_SIZE_SHIFT));
-+ iowrite32be(tmp, &bmi_rg->fmbm_pfs[port_id - 1]);
-+}
-+
-+uint8_t fman_get_num_of_tasks(struct fman_bmi_regs *bmi_rg, uint8_t port_id)
-+{
-+ uint32_t tmp;
-+
-+ if ((port_id > 63) || (port_id < 1))
-+ return 0;
-+
-+ tmp = ioread32be(&bmi_rg->fmbm_pp[port_id - 1]);
-+ return (uint8_t)(((tmp & BMI_NUM_OF_TASKS_MASK) >>
-+ BMI_NUM_OF_TASKS_SHIFT) + 1);
-+}
-+
-+uint8_t fman_get_num_extra_tasks(struct fman_bmi_regs *bmi_rg, uint8_t port_id)
-+{
-+ uint32_t tmp;
-+
-+ if ((port_id > 63) || (port_id < 1))
-+ return 0;
-+
-+ tmp = ioread32be(&bmi_rg->fmbm_pp[port_id - 1]);
-+ return (uint8_t)((tmp & BMI_NUM_OF_EXTRA_TASKS_MASK) >>
-+ BMI_EXTRA_NUM_OF_TASKS_SHIFT);
-+}
-+
-+void fman_set_num_of_tasks(struct fman_bmi_regs *bmi_rg,
-+ uint8_t port_id,
-+ uint8_t num_tasks,
-+ uint8_t num_extra_tasks)
-+{
-+ uint32_t tmp;
-+
-+ if ((port_id > 63) || (port_id < 1))
-+ return;
-+
-+ /* calculate reg */
-+ tmp = ioread32be(&bmi_rg->fmbm_pp[port_id - 1]) &
-+ ~(BMI_NUM_OF_TASKS_MASK | BMI_NUM_OF_EXTRA_TASKS_MASK);
-+ tmp |= (uint32_t)(((num_tasks - 1) << BMI_NUM_OF_TASKS_SHIFT) |
-+ (num_extra_tasks << BMI_EXTRA_NUM_OF_TASKS_SHIFT));
-+ iowrite32be(tmp, &bmi_rg->fmbm_pp[port_id - 1]);
-+}
-+
-+uint8_t fman_get_num_of_dmas(struct fman_bmi_regs *bmi_rg, uint8_t port_id)
-+{
-+ uint32_t tmp;
-+
-+ if ((port_id > 63) || (port_id < 1))
-+ return 0;
-+
-+ tmp = ioread32be(&bmi_rg->fmbm_pp[port_id - 1]);
-+ return (uint8_t)(((tmp & BMI_NUM_OF_DMAS_MASK) >>
-+ BMI_NUM_OF_DMAS_SHIFT) + 1);
-+}
-+
-+uint8_t fman_get_num_extra_dmas(struct fman_bmi_regs *bmi_rg, uint8_t port_id)
-+{
-+ uint32_t tmp;
-+
-+ if ((port_id > 63) || (port_id < 1))
-+ return 0;
-+
-+ tmp = ioread32be(&bmi_rg->fmbm_pp[port_id - 1]);
-+ return (uint8_t)((tmp & BMI_NUM_OF_EXTRA_DMAS_MASK) >>
-+ BMI_EXTRA_NUM_OF_DMAS_SHIFT);
-+}
-+
-+void fman_set_num_of_open_dmas(struct fman_bmi_regs *bmi_rg,
-+ uint8_t port_id,
-+ uint8_t num_open_dmas,
-+ uint8_t num_extra_open_dmas,
-+ uint8_t total_num_dmas)
-+{
-+ uint32_t tmp = 0;
-+
-+ if ((port_id > 63) || (port_id < 1))
-+ return;
-+
-+ /* calculate reg */
-+ tmp = ioread32be(&bmi_rg->fmbm_pp[port_id - 1]) &
-+ ~(BMI_NUM_OF_DMAS_MASK | BMI_NUM_OF_EXTRA_DMAS_MASK);
-+ tmp |= (uint32_t)(((num_open_dmas-1) << BMI_NUM_OF_DMAS_SHIFT) |
-+ (num_extra_open_dmas << BMI_EXTRA_NUM_OF_DMAS_SHIFT));
-+ iowrite32be(tmp, &bmi_rg->fmbm_pp[port_id - 1]);
-+
-+ /* update total num of DMA's with committed number of open DMAS,
-+ * and max uncommitted pool. */
-+ if (total_num_dmas)
-+ {
-+ tmp = ioread32be(&bmi_rg->fmbm_cfg2) & ~BMI_CFG2_DMAS_MASK;
-+ tmp |= (uint32_t)(total_num_dmas - 1) << BMI_CFG2_DMAS_SHIFT;
-+ iowrite32be(tmp, &bmi_rg->fmbm_cfg2);
-+ }
-+}
-+
-+void fman_set_vsp_window(struct fman_bmi_regs *bmi_rg,
-+ uint8_t port_id,
-+ uint8_t base_storage_profile,
-+ uint8_t log2_num_of_profiles)
-+{
-+ uint32_t tmp = 0;
-+ if ((port_id > 63) || (port_id < 1))
-+ return;
-+
-+ tmp = ioread32be(&bmi_rg->fmbm_spliodn[port_id-1]);
-+ tmp |= (uint32_t)((uint32_t)base_storage_profile & 0x3f) << 16;
-+ tmp |= (uint32_t)log2_num_of_profiles << 28;
-+ iowrite32be(tmp, &bmi_rg->fmbm_spliodn[port_id-1]);
-+}
-+
-+void fman_set_congestion_group_pfc_priority(uint32_t *cpg_rg,
-+ uint32_t congestion_group_id,
-+ uint8_t priority_bit_map,
-+ uint32_t reg_num)
-+{
-+ uint32_t offset, tmp = 0;
-+
-+ offset = (congestion_group_id%4)*8;
-+
-+ tmp = ioread32be(&cpg_rg[reg_num]);
-+ tmp &= ~(0xFF<<offset);
-+ tmp |= (uint32_t)priority_bit_map << offset;
-+
-+ iowrite32be(tmp,&cpg_rg[reg_num]);
-+}
-+
-+/*****************************************************************************/
-+/* API Init unit functions */
-+/*****************************************************************************/
-+void fman_defconfig(struct fman_cfg *cfg, bool is_master)
-+{
-+ memset(cfg, 0, sizeof(struct fman_cfg));
-+
-+ cfg->catastrophic_err = DEFAULT_CATASTROPHIC_ERR;
-+ cfg->dma_err = DEFAULT_DMA_ERR;
-+ cfg->halt_on_external_activ = DEFAULT_HALT_ON_EXTERNAL_ACTIVATION;
-+ cfg->halt_on_unrecov_ecc_err = DEFAULT_HALT_ON_UNRECOVERABLE_ECC_ERROR;
-+ cfg->en_iram_test_mode = FALSE;
-+ cfg->en_muram_test_mode = FALSE;
-+ cfg->external_ecc_rams_enable = DEFAULT_EXTERNAL_ECC_RAMS_ENABLE;
-+
-+ if (!is_master)
-+ return;
-+
-+ cfg->dma_aid_override = DEFAULT_AID_OVERRIDE;
-+ cfg->dma_aid_mode = DEFAULT_AID_MODE;
-+ cfg->dma_comm_qtsh_clr_emer = DEFAULT_DMA_COMM_Q_LOW;
-+ cfg->dma_comm_qtsh_asrt_emer = DEFAULT_DMA_COMM_Q_HIGH;
-+ cfg->dma_cache_override = DEFAULT_CACHE_OVERRIDE;
-+ cfg->dma_cam_num_of_entries = DEFAULT_DMA_CAM_NUM_OF_ENTRIES;
-+ cfg->dma_dbg_cnt_mode = DEFAULT_DMA_DBG_CNT_MODE;
-+ cfg->dma_en_emergency = DEFAULT_DMA_EN_EMERGENCY;
-+ cfg->dma_sos_emergency = DEFAULT_DMA_SOS_EMERGENCY;
-+ cfg->dma_watchdog = DEFAULT_DMA_WATCHDOG;
-+ cfg->dma_en_emergency_smoother = DEFAULT_DMA_EN_EMERGENCY_SMOOTHER;
-+ cfg->dma_emergency_switch_counter = DEFAULT_DMA_EMERGENCY_SWITCH_COUNTER;
-+ cfg->disp_limit_tsh = DEFAULT_DISP_LIMIT;
-+ cfg->prs_disp_tsh = DEFAULT_PRS_DISP_TH;
-+ cfg->plcr_disp_tsh = DEFAULT_PLCR_DISP_TH;
-+ cfg->kg_disp_tsh = DEFAULT_KG_DISP_TH;
-+ cfg->bmi_disp_tsh = DEFAULT_BMI_DISP_TH;
-+ cfg->qmi_enq_disp_tsh = DEFAULT_QMI_ENQ_DISP_TH;
-+ cfg->qmi_deq_disp_tsh = DEFAULT_QMI_DEQ_DISP_TH;
-+ cfg->fm_ctl1_disp_tsh = DEFAULT_FM_CTL1_DISP_TH;
-+ cfg->fm_ctl2_disp_tsh = DEFAULT_FM_CTL2_DISP_TH;
-+
-+ cfg->pedantic_dma = FALSE;
-+ cfg->tnum_aging_period = DEFAULT_TNUM_AGING_PERIOD;
-+ cfg->dma_stop_on_bus_error = FALSE;
-+ cfg->qmi_deq_option_support = FALSE;
-+}
-+
-+void fman_regconfig(struct fman_rg *fman_rg, struct fman_cfg *cfg)
-+{
-+ uint32_t tmp_reg;
-+
-+ /* read the values from the registers as they are initialized by the HW with
-+ * the required values.
-+ */
-+ tmp_reg = ioread32be(&fman_rg->bmi_rg->fmbm_cfg1);
-+ cfg->total_fifo_size =
-+ (((tmp_reg & BMI_TOTAL_FIFO_SIZE_MASK) >> BMI_CFG1_FIFO_SIZE_SHIFT) + 1) * FMAN_BMI_FIFO_UNITS;
-+
-+ tmp_reg = ioread32be(&fman_rg->bmi_rg->fmbm_cfg2);
-+ cfg->total_num_of_tasks =
-+ (uint8_t)(((tmp_reg & BMI_TOTAL_NUM_OF_TASKS_MASK) >> BMI_CFG2_TASKS_SHIFT) + 1);
-+
-+ tmp_reg = ioread32be(&fman_rg->dma_rg->fmdmtr);
-+ cfg->dma_comm_qtsh_asrt_emer = (uint8_t)(tmp_reg >> DMA_THRESH_COMMQ_SHIFT);
-+
-+ tmp_reg = ioread32be(&fman_rg->dma_rg->fmdmhy);
-+ cfg->dma_comm_qtsh_clr_emer = (uint8_t)(tmp_reg >> DMA_THRESH_COMMQ_SHIFT);
-+
-+ tmp_reg = ioread32be(&fman_rg->dma_rg->fmdmmr);
-+ cfg->dma_cache_override = (enum fman_dma_cache_override)((tmp_reg & DMA_MODE_CACHE_OR_MASK) >> DMA_MODE_CACHE_OR_SHIFT);
-+ cfg->dma_cam_num_of_entries = (uint8_t)((((tmp_reg & DMA_MODE_CEN_MASK) >> DMA_MODE_CEN_SHIFT) +1)*DMA_CAM_UNITS);
-+ cfg->dma_aid_override = (bool)((tmp_reg & DMA_MODE_AID_OR)? TRUE:FALSE);
-+ cfg->dma_dbg_cnt_mode = (enum fman_dma_dbg_cnt_mode)((tmp_reg & DMA_MODE_DBG_MASK) >> DMA_MODE_DBG_SHIFT);
-+ cfg->dma_en_emergency = (bool)((tmp_reg & DMA_MODE_EB)? TRUE : FALSE);
-+
-+ tmp_reg = ioread32be(&fman_rg->fpm_rg->fmfp_mxd);
-+ cfg->disp_limit_tsh = (uint8_t)((tmp_reg & FPM_DISP_LIMIT_MASK) >> FPM_DISP_LIMIT_SHIFT);
-+
-+ tmp_reg = ioread32be(&fman_rg->fpm_rg->fmfp_dist1);
-+ cfg->prs_disp_tsh = (uint8_t)((tmp_reg & FPM_THR1_PRS_MASK ) >> FPM_THR1_PRS_SHIFT);
-+ cfg->plcr_disp_tsh = (uint8_t)((tmp_reg & FPM_THR1_KG_MASK ) >> FPM_THR1_KG_SHIFT);
-+ cfg->kg_disp_tsh = (uint8_t)((tmp_reg & FPM_THR1_PLCR_MASK ) >> FPM_THR1_PLCR_SHIFT);
-+ cfg->bmi_disp_tsh = (uint8_t)((tmp_reg & FPM_THR1_BMI_MASK ) >> FPM_THR1_BMI_SHIFT);
-+
-+ tmp_reg = ioread32be(&fman_rg->fpm_rg->fmfp_dist2);
-+ cfg->qmi_enq_disp_tsh = (uint8_t)((tmp_reg & FPM_THR2_QMI_ENQ_MASK ) >> FPM_THR2_QMI_ENQ_SHIFT);
-+ cfg->qmi_deq_disp_tsh = (uint8_t)((tmp_reg & FPM_THR2_QMI_DEQ_MASK ) >> FPM_THR2_QMI_DEQ_SHIFT);
-+ cfg->fm_ctl1_disp_tsh = (uint8_t)((tmp_reg & FPM_THR2_FM_CTL1_MASK ) >> FPM_THR2_FM_CTL1_SHIFT);
-+ cfg->fm_ctl2_disp_tsh = (uint8_t)((tmp_reg & FPM_THR2_FM_CTL2_MASK ) >> FPM_THR2_FM_CTL2_SHIFT);
-+
-+ tmp_reg = ioread32be(&fman_rg->dma_rg->fmdmsetr);
-+ cfg->dma_sos_emergency = tmp_reg;
-+
-+ tmp_reg = ioread32be(&fman_rg->dma_rg->fmdmwcr);
-+ cfg->dma_watchdog = tmp_reg/cfg->clk_freq;
-+
-+ tmp_reg = ioread32be(&fman_rg->dma_rg->fmdmemsr);
-+ cfg->dma_en_emergency_smoother = (bool)((tmp_reg & DMA_EMSR_EMSTR_MASK)? TRUE : FALSE);
-+ cfg->dma_emergency_switch_counter = (tmp_reg & DMA_EMSR_EMSTR_MASK);
-+}
-+
-+void fman_reset(struct fman_fpm_regs *fpm_rg)
-+{
-+ iowrite32be(FPM_RSTC_FM_RESET, &fpm_rg->fm_rstc);
-+}
-+
-+/**************************************************************************//**
-+ @Function FM_Init
-+
-+ @Description Initializes the FM module
-+
-+ @Param[in] h_Fm - FM module descriptor
-+
-+ @Return E_OK on success; Error code otherwise.
-+*//***************************************************************************/
-+int fman_dma_init(struct fman_dma_regs *dma_rg, struct fman_cfg *cfg)
-+{
-+ uint32_t tmp_reg;
-+
-+ /**********************/
-+ /* Init DMA Registers */
-+ /**********************/
-+ /* clear status reg events */
-+ /* oren - check!!! */
-+ tmp_reg = (DMA_STATUS_BUS_ERR | DMA_STATUS_READ_ECC |
-+ DMA_STATUS_SYSTEM_WRITE_ECC | DMA_STATUS_FM_WRITE_ECC);
-+ iowrite32be(ioread32be(&dma_rg->fmdmsr) | tmp_reg,
-+ &dma_rg->fmdmsr);
-+
-+ /* configure mode register */
-+ tmp_reg = 0;
-+ tmp_reg |= cfg->dma_cache_override << DMA_MODE_CACHE_OR_SHIFT;
-+ if (cfg->dma_aid_override)
-+ tmp_reg |= DMA_MODE_AID_OR;
-+ if (cfg->exceptions & FMAN_EX_DMA_BUS_ERROR)
-+ tmp_reg |= DMA_MODE_BER;
-+ if ((cfg->exceptions & FMAN_EX_DMA_SYSTEM_WRITE_ECC) |
-+ (cfg->exceptions & FMAN_EX_DMA_READ_ECC) |
-+ (cfg->exceptions & FMAN_EX_DMA_FM_WRITE_ECC))
-+ tmp_reg |= DMA_MODE_ECC;
-+ if (cfg->dma_stop_on_bus_error)
-+ tmp_reg |= DMA_MODE_SBER;
-+ if(cfg->dma_axi_dbg_num_of_beats)
-+ tmp_reg |= (uint32_t)(DMA_MODE_AXI_DBG_MASK &
-+ ((cfg->dma_axi_dbg_num_of_beats - 1) << DMA_MODE_AXI_DBG_SHIFT));
-+
-+ if (cfg->dma_en_emergency) {
-+ tmp_reg |= cfg->dma_emergency_bus_select;
-+ tmp_reg |= cfg->dma_emergency_level << DMA_MODE_EMER_LVL_SHIFT;
-+ if (cfg->dma_en_emergency_smoother)
-+ iowrite32be(cfg->dma_emergency_switch_counter,
-+ &dma_rg->fmdmemsr);
-+ }
-+ tmp_reg |= ((cfg->dma_cam_num_of_entries / DMA_CAM_UNITS) - 1) <<
-+ DMA_MODE_CEN_SHIFT;
-+ tmp_reg |= DMA_MODE_SECURE_PROT;
-+ tmp_reg |= cfg->dma_dbg_cnt_mode << DMA_MODE_DBG_SHIFT;
-+ tmp_reg |= cfg->dma_aid_mode << DMA_MODE_AID_MODE_SHIFT;
-+
-+ if (cfg->pedantic_dma)
-+ tmp_reg |= DMA_MODE_EMER_READ;
-+
-+ iowrite32be(tmp_reg, &dma_rg->fmdmmr);
-+
-+ /* configure thresholds register */
-+ tmp_reg = ((uint32_t)cfg->dma_comm_qtsh_asrt_emer <<
-+ DMA_THRESH_COMMQ_SHIFT) |
-+ ((uint32_t)cfg->dma_read_buf_tsh_asrt_emer <<
-+ DMA_THRESH_READ_INT_BUF_SHIFT) |
-+ ((uint32_t)cfg->dma_write_buf_tsh_asrt_emer);
-+
-+ iowrite32be(tmp_reg, &dma_rg->fmdmtr);
-+
-+ /* configure hysteresis register */
-+ tmp_reg = ((uint32_t)cfg->dma_comm_qtsh_clr_emer <<
-+ DMA_THRESH_COMMQ_SHIFT) |
-+ ((uint32_t)cfg->dma_read_buf_tsh_clr_emer <<
-+ DMA_THRESH_READ_INT_BUF_SHIFT) |
-+ ((uint32_t)cfg->dma_write_buf_tsh_clr_emer);
-+
-+ iowrite32be(tmp_reg, &dma_rg->fmdmhy);
-+
-+ /* configure emergency threshold */
-+ iowrite32be(cfg->dma_sos_emergency, &dma_rg->fmdmsetr);
-+
-+ /* configure Watchdog */
-+ iowrite32be((cfg->dma_watchdog * cfg->clk_freq),
-+ &dma_rg->fmdmwcr);
-+
-+ iowrite32be(cfg->cam_base_addr, &dma_rg->fmdmebcr);
-+
-+ return 0;
-+}
-+
-+int fman_fpm_init(struct fman_fpm_regs *fpm_rg, struct fman_cfg *cfg)
-+{
-+ uint32_t tmp_reg;
-+ int i;
-+
-+ /**********************/
-+ /* Init FPM Registers */
-+ /**********************/
-+ tmp_reg = (uint32_t)(cfg->disp_limit_tsh << FPM_DISP_LIMIT_SHIFT);
-+ iowrite32be(tmp_reg, &fpm_rg->fmfp_mxd);
-+
-+ tmp_reg = (((uint32_t)cfg->prs_disp_tsh << FPM_THR1_PRS_SHIFT) |
-+ ((uint32_t)cfg->kg_disp_tsh << FPM_THR1_KG_SHIFT) |
-+ ((uint32_t)cfg->plcr_disp_tsh << FPM_THR1_PLCR_SHIFT) |
-+ ((uint32_t)cfg->bmi_disp_tsh << FPM_THR1_BMI_SHIFT));
-+ iowrite32be(tmp_reg, &fpm_rg->fmfp_dist1);
-+
-+ tmp_reg = (((uint32_t)cfg->qmi_enq_disp_tsh << FPM_THR2_QMI_ENQ_SHIFT) |
-+ ((uint32_t)cfg->qmi_deq_disp_tsh << FPM_THR2_QMI_DEQ_SHIFT) |
-+ ((uint32_t)cfg->fm_ctl1_disp_tsh << FPM_THR2_FM_CTL1_SHIFT) |
-+ ((uint32_t)cfg->fm_ctl2_disp_tsh << FPM_THR2_FM_CTL2_SHIFT));
-+ iowrite32be(tmp_reg, &fpm_rg->fmfp_dist2);
-+
-+ /* define exceptions and error behavior */
-+ tmp_reg = 0;
-+ /* Clear events */
-+ tmp_reg |= (FPM_EV_MASK_STALL | FPM_EV_MASK_DOUBLE_ECC |
-+ FPM_EV_MASK_SINGLE_ECC);
-+ /* enable interrupts */
-+ if (cfg->exceptions & FMAN_EX_FPM_STALL_ON_TASKS)
-+ tmp_reg |= FPM_EV_MASK_STALL_EN;
-+ if (cfg->exceptions & FMAN_EX_FPM_SINGLE_ECC)
-+ tmp_reg |= FPM_EV_MASK_SINGLE_ECC_EN;
-+ if (cfg->exceptions & FMAN_EX_FPM_DOUBLE_ECC)
-+ tmp_reg |= FPM_EV_MASK_DOUBLE_ECC_EN;
-+ tmp_reg |= (cfg->catastrophic_err << FPM_EV_MASK_CAT_ERR_SHIFT);
-+ tmp_reg |= (cfg->dma_err << FPM_EV_MASK_DMA_ERR_SHIFT);
-+ if (!cfg->halt_on_external_activ)
-+ tmp_reg |= FPM_EV_MASK_EXTERNAL_HALT;
-+ if (!cfg->halt_on_unrecov_ecc_err)
-+ tmp_reg |= FPM_EV_MASK_ECC_ERR_HALT;
-+ iowrite32be(tmp_reg, &fpm_rg->fmfp_ee);
-+
-+ /* clear all fmCtls event registers */
-+ for (i = 0; i < cfg->num_of_fman_ctrl_evnt_regs; i++)
-+ iowrite32be(0xFFFFFFFF, &fpm_rg->fmfp_cev[i]);
-+
-+ /* RAM ECC - enable and clear events*/
-+ /* first we need to clear all parser memory,
-+ * as it is uninitialized and may cause ECC errors */
-+ /* event bits */
-+ tmp_reg = (FPM_RAM_MURAM_ECC | FPM_RAM_IRAM_ECC);
-+ /* Rams enable not effected by RCR bit, but by a COP configuration */
-+ if (cfg->external_ecc_rams_enable)
-+ tmp_reg |= FPM_RAM_RAMS_ECC_EN_SRC_SEL;
-+
-+ /* enable test mode */
-+ if (cfg->en_muram_test_mode)
-+ tmp_reg |= FPM_RAM_MURAM_TEST_ECC;
-+ if (cfg->en_iram_test_mode)
-+ tmp_reg |= FPM_RAM_IRAM_TEST_ECC;
-+ iowrite32be(tmp_reg, &fpm_rg->fm_rcr);
-+
-+ tmp_reg = 0;
-+ if (cfg->exceptions & FMAN_EX_IRAM_ECC) {
-+ tmp_reg |= FPM_IRAM_ECC_ERR_EX_EN;
-+ fman_enable_rams_ecc(fpm_rg);
-+ }
-+ if (cfg->exceptions & FMAN_EX_NURAM_ECC) {
-+ tmp_reg |= FPM_MURAM_ECC_ERR_EX_EN;
-+ fman_enable_rams_ecc(fpm_rg);
-+ }
-+ iowrite32be(tmp_reg, &fpm_rg->fm_rie);
-+
-+ return 0;
-+}
-+
-+int fman_bmi_init(struct fman_bmi_regs *bmi_rg, struct fman_cfg *cfg)
-+{
-+ uint32_t tmp_reg;
-+
-+ /**********************/
-+ /* Init BMI Registers */
-+ /**********************/
-+
-+ /* define common resources */
-+ tmp_reg = cfg->fifo_base_addr;
-+ tmp_reg = tmp_reg / BMI_FIFO_ALIGN;
-+
-+ tmp_reg |= ((cfg->total_fifo_size / FMAN_BMI_FIFO_UNITS - 1) <<
-+ BMI_CFG1_FIFO_SIZE_SHIFT);
-+ iowrite32be(tmp_reg, &bmi_rg->fmbm_cfg1);
-+
-+ tmp_reg = ((uint32_t)(cfg->total_num_of_tasks - 1) <<
-+ BMI_CFG2_TASKS_SHIFT);
-+ /* num of DMA's will be dynamically updated when each port is set */
-+ iowrite32be(tmp_reg, &bmi_rg->fmbm_cfg2);
-+
-+ /* define unmaskable exceptions, enable and clear events */
-+ tmp_reg = 0;
-+ iowrite32be(BMI_ERR_INTR_EN_LIST_RAM_ECC |
-+ BMI_ERR_INTR_EN_STORAGE_PROFILE_ECC |
-+ BMI_ERR_INTR_EN_STATISTICS_RAM_ECC |
-+ BMI_ERR_INTR_EN_DISPATCH_RAM_ECC,
-+ &bmi_rg->fmbm_ievr);
-+
-+ if (cfg->exceptions & FMAN_EX_BMI_LIST_RAM_ECC)
-+ tmp_reg |= BMI_ERR_INTR_EN_LIST_RAM_ECC;
-+ if (cfg->exceptions & FMAN_EX_BMI_PIPELINE_ECC)
-+ tmp_reg |= BMI_ERR_INTR_EN_STORAGE_PROFILE_ECC;
-+ if (cfg->exceptions & FMAN_EX_BMI_STATISTICS_RAM_ECC)
-+ tmp_reg |= BMI_ERR_INTR_EN_STATISTICS_RAM_ECC;
-+ if (cfg->exceptions & FMAN_EX_BMI_DISPATCH_RAM_ECC)
-+ tmp_reg |= BMI_ERR_INTR_EN_DISPATCH_RAM_ECC;
-+ iowrite32be(tmp_reg, &bmi_rg->fmbm_ier);
-+
-+ return 0;
-+}
-+
-+int fman_qmi_init(struct fman_qmi_regs *qmi_rg, struct fman_cfg *cfg)
-+{
-+ uint32_t tmp_reg;
-+ uint16_t period_in_fm_clocks;
-+ uint8_t remainder;
-+ /**********************/
-+ /* Init QMI Registers */
-+ /**********************/
-+ /* Clear error interrupt events */
-+
-+ iowrite32be(QMI_ERR_INTR_EN_DOUBLE_ECC | QMI_ERR_INTR_EN_DEQ_FROM_DEF,
-+ &qmi_rg->fmqm_eie);
-+ tmp_reg = 0;
-+ if (cfg->exceptions & FMAN_EX_QMI_DEQ_FROM_UNKNOWN_PORTID)
-+ tmp_reg |= QMI_ERR_INTR_EN_DEQ_FROM_DEF;
-+ if (cfg->exceptions & FMAN_EX_QMI_DOUBLE_ECC)
-+ tmp_reg |= QMI_ERR_INTR_EN_DOUBLE_ECC;
-+ /* enable events */
-+ iowrite32be(tmp_reg, &qmi_rg->fmqm_eien);
-+
-+ if (cfg->tnum_aging_period) {
-+ /* tnum_aging_period is in units of usec, p_FmClockFreq in Mhz */
-+ period_in_fm_clocks = (uint16_t)
-+ (cfg->tnum_aging_period * cfg->clk_freq);
-+ /* period_in_fm_clocks must be a 64 multiply */
-+ remainder = (uint8_t)(period_in_fm_clocks % 64);
-+ if (remainder)
-+ tmp_reg = (uint32_t)((period_in_fm_clocks / 64) + 1);
-+ else{
-+ tmp_reg = (uint32_t)(period_in_fm_clocks / 64);
-+ if (!tmp_reg)
-+ tmp_reg = 1;
-+ }
-+ tmp_reg <<= QMI_TAPC_TAP;
-+ iowrite32be(tmp_reg, &qmi_rg->fmqm_tapc);
-+ }
-+ tmp_reg = 0;
-+ /* Clear interrupt events */
-+ iowrite32be(QMI_INTR_EN_SINGLE_ECC, &qmi_rg->fmqm_ie);
-+ if (cfg->exceptions & FMAN_EX_QMI_SINGLE_ECC)
-+ tmp_reg |= QMI_INTR_EN_SINGLE_ECC;
-+ /* enable events */
-+ iowrite32be(tmp_reg, &qmi_rg->fmqm_ien);
-+
-+ return 0;
-+}
-+
-+int fman_enable(struct fman_rg *fman_rg, struct fman_cfg *cfg)
-+{
-+ uint32_t cfg_reg = 0;
-+
-+ /**********************/
-+ /* Enable all modules */
-+ /**********************/
-+ /* clear & enable global counters - calculate reg and save for later,
-+ because it's the same reg for QMI enable */
-+ cfg_reg = QMI_CFG_EN_COUNTERS;
-+ if (cfg->qmi_deq_option_support)
-+ cfg_reg |= (uint32_t)(((cfg->qmi_def_tnums_thresh) << 8) |
-+ (uint32_t)cfg->qmi_def_tnums_thresh);
-+
-+ iowrite32be(BMI_INIT_START, &fman_rg->bmi_rg->fmbm_init);
-+ iowrite32be(cfg_reg | QMI_CFG_ENQ_EN | QMI_CFG_DEQ_EN,
-+ &fman_rg->qmi_rg->fmqm_gc);
-+
-+ return 0;
-+}
-+
-+void fman_free_resources(struct fman_rg *fman_rg)
-+{
-+ /* disable BMI and QMI */
-+ iowrite32be(0, &fman_rg->bmi_rg->fmbm_init);
-+ iowrite32be(0, &fman_rg->qmi_rg->fmqm_gc);
-+
-+ /* release BMI resources */
-+ iowrite32be(0, &fman_rg->bmi_rg->fmbm_cfg2);
-+ iowrite32be(0, &fman_rg->bmi_rg->fmbm_cfg1);
-+
-+ /* disable ECC */
-+ iowrite32be(0, &fman_rg->fpm_rg->fm_rcr);
-+}
-+
-+/****************************************************/
-+/* API Run-time Control uint functions */
-+/****************************************************/
-+uint32_t fman_get_normal_pending(struct fman_fpm_regs *fpm_rg)
-+{
-+ return ioread32be(&fpm_rg->fm_npi);
-+}
-+
-+uint32_t fman_get_controller_event(struct fman_fpm_regs *fpm_rg, uint8_t reg_id)
-+{
-+ uint32_t event;
-+
-+ event = ioread32be(&fpm_rg->fmfp_fcev[reg_id]) &
-+ ioread32be(&fpm_rg->fmfp_cee[reg_id]);
-+ iowrite32be(event, &fpm_rg->fmfp_cev[reg_id]);
-+
-+ return event;
-+}
-+
-+uint32_t fman_get_error_pending(struct fman_fpm_regs *fpm_rg)
-+{
-+ return ioread32be(&fpm_rg->fm_epi);
-+}
-+
-+void fman_set_ports_bandwidth(struct fman_bmi_regs *bmi_rg, uint8_t *weights)
-+{
-+ int i;
-+ uint8_t shift;
-+ uint32_t tmp = 0;
-+
-+ for (i = 0; i < 64; i++) {
-+ if (weights[i] > 1) { /* no need to write 1 since it is 0 */
-+ /* Add this port to tmp_reg */
-+ /* (each 8 ports result in one register)*/
-+ shift = (uint8_t)(32 - 4 * ((i % 8) + 1));
-+ tmp |= ((weights[i] - 1) << shift);
-+ }
-+ if (i % 8 == 7) { /* last in this set */
-+ iowrite32be(tmp, &bmi_rg->fmbm_arb[i / 8]);
-+ tmp = 0;
-+ }
-+ }
-+}
-+
-+void fman_enable_rams_ecc(struct fman_fpm_regs *fpm_rg)
-+{
-+ uint32_t tmp;
-+
-+ tmp = ioread32be(&fpm_rg->fm_rcr);
-+ if (tmp & FPM_RAM_RAMS_ECC_EN_SRC_SEL)
-+ iowrite32be(tmp | FPM_RAM_IRAM_ECC_EN,
-+ &fpm_rg->fm_rcr);
-+ else
-+ iowrite32be(tmp | FPM_RAM_RAMS_ECC_EN |
-+ FPM_RAM_IRAM_ECC_EN,
-+ &fpm_rg->fm_rcr);
-+}
-+
-+void fman_disable_rams_ecc(struct fman_fpm_regs *fpm_rg)
-+{
-+ uint32_t tmp;
-+
-+ tmp = ioread32be(&fpm_rg->fm_rcr);
-+ if (tmp & FPM_RAM_RAMS_ECC_EN_SRC_SEL)
-+ iowrite32be(tmp & ~FPM_RAM_IRAM_ECC_EN,
-+ &fpm_rg->fm_rcr);
-+ else
-+ iowrite32be(tmp & ~(FPM_RAM_RAMS_ECC_EN | FPM_RAM_IRAM_ECC_EN),
-+ &fpm_rg->fm_rcr);
-+}
-+
-+int fman_set_exception(struct fman_rg *fman_rg,
-+ enum fman_exceptions exception,
-+ bool enable)
-+{
-+ uint32_t tmp;
-+
-+ switch (exception) {
-+ case(E_FMAN_EX_DMA_BUS_ERROR):
-+ tmp = ioread32be(&fman_rg->dma_rg->fmdmmr);
-+ if (enable)
-+ tmp |= DMA_MODE_BER;
-+ else
-+ tmp &= ~DMA_MODE_BER;
-+ /* disable bus error */
-+ iowrite32be(tmp, &fman_rg->dma_rg->fmdmmr);
-+ break;
-+ case(E_FMAN_EX_DMA_READ_ECC):
-+ case(E_FMAN_EX_DMA_SYSTEM_WRITE_ECC):
-+ case(E_FMAN_EX_DMA_FM_WRITE_ECC):
-+ tmp = ioread32be(&fman_rg->dma_rg->fmdmmr);
-+ if (enable)
-+ tmp |= DMA_MODE_ECC;
-+ else
-+ tmp &= ~DMA_MODE_ECC;
-+ iowrite32be(tmp, &fman_rg->dma_rg->fmdmmr);
-+ break;
-+ case(E_FMAN_EX_FPM_STALL_ON_TASKS):
-+ tmp = ioread32be(&fman_rg->fpm_rg->fmfp_ee);
-+ if (enable)
-+ tmp |= FPM_EV_MASK_STALL_EN;
-+ else
-+ tmp &= ~FPM_EV_MASK_STALL_EN;
-+ iowrite32be(tmp, &fman_rg->fpm_rg->fmfp_ee);
-+ break;
-+ case(E_FMAN_EX_FPM_SINGLE_ECC):
-+ tmp = ioread32be(&fman_rg->fpm_rg->fmfp_ee);
-+ if (enable)
-+ tmp |= FPM_EV_MASK_SINGLE_ECC_EN;
-+ else
-+ tmp &= ~FPM_EV_MASK_SINGLE_ECC_EN;
-+ iowrite32be(tmp, &fman_rg->fpm_rg->fmfp_ee);
-+ break;
-+ case(E_FMAN_EX_FPM_DOUBLE_ECC):
-+ tmp = ioread32be(&fman_rg->fpm_rg->fmfp_ee);
-+ if (enable)
-+ tmp |= FPM_EV_MASK_DOUBLE_ECC_EN;
-+ else
-+ tmp &= ~FPM_EV_MASK_DOUBLE_ECC_EN;
-+ iowrite32be(tmp, &fman_rg->fpm_rg->fmfp_ee);
-+ break;
-+ case(E_FMAN_EX_QMI_SINGLE_ECC):
-+ tmp = ioread32be(&fman_rg->qmi_rg->fmqm_ien);
-+ if (enable)
-+ tmp |= QMI_INTR_EN_SINGLE_ECC;
-+ else
-+ tmp &= ~QMI_INTR_EN_SINGLE_ECC;
-+ iowrite32be(tmp, &fman_rg->qmi_rg->fmqm_ien);
-+ break;
-+ case(E_FMAN_EX_QMI_DOUBLE_ECC):
-+ tmp = ioread32be(&fman_rg->qmi_rg->fmqm_eien);
-+ if (enable)
-+ tmp |= QMI_ERR_INTR_EN_DOUBLE_ECC;
-+ else
-+ tmp &= ~QMI_ERR_INTR_EN_DOUBLE_ECC;
-+ iowrite32be(tmp, &fman_rg->qmi_rg->fmqm_eien);
-+ break;
-+ case(E_FMAN_EX_QMI_DEQ_FROM_UNKNOWN_PORTID):
-+ tmp = ioread32be(&fman_rg->qmi_rg->fmqm_eien);
-+ if (enable)
-+ tmp |= QMI_ERR_INTR_EN_DEQ_FROM_DEF;
-+ else
-+ tmp &= ~QMI_ERR_INTR_EN_DEQ_FROM_DEF;
-+ iowrite32be(tmp, &fman_rg->qmi_rg->fmqm_eien);
-+ break;
-+ case(E_FMAN_EX_BMI_LIST_RAM_ECC):
-+ tmp = ioread32be(&fman_rg->bmi_rg->fmbm_ier);
-+ if (enable)
-+ tmp |= BMI_ERR_INTR_EN_LIST_RAM_ECC;
-+ else
-+ tmp &= ~BMI_ERR_INTR_EN_LIST_RAM_ECC;
-+ iowrite32be(tmp, &fman_rg->bmi_rg->fmbm_ier);
-+ break;
-+ case(E_FMAN_EX_BMI_STORAGE_PROFILE_ECC):
-+ tmp = ioread32be(&fman_rg->bmi_rg->fmbm_ier);
-+ if (enable)
-+ tmp |= BMI_ERR_INTR_EN_STORAGE_PROFILE_ECC;
-+ else
-+ tmp &= ~BMI_ERR_INTR_EN_STORAGE_PROFILE_ECC;
-+ iowrite32be(tmp, &fman_rg->bmi_rg->fmbm_ier);
-+ break;
-+ case(E_FMAN_EX_BMI_STATISTICS_RAM_ECC):
-+ tmp = ioread32be(&fman_rg->bmi_rg->fmbm_ier);
-+ if (enable)
-+ tmp |= BMI_ERR_INTR_EN_STATISTICS_RAM_ECC;
-+ else
-+ tmp &= ~BMI_ERR_INTR_EN_STATISTICS_RAM_ECC;
-+ iowrite32be(tmp, &fman_rg->bmi_rg->fmbm_ier);
-+ break;
-+ case(E_FMAN_EX_BMI_DISPATCH_RAM_ECC):
-+ tmp = ioread32be(&fman_rg->bmi_rg->fmbm_ier);
-+ if (enable)
-+ tmp |= BMI_ERR_INTR_EN_DISPATCH_RAM_ECC;
-+ else
-+ tmp &= ~BMI_ERR_INTR_EN_DISPATCH_RAM_ECC;
-+ iowrite32be(tmp, &fman_rg->bmi_rg->fmbm_ier);
-+ break;
-+ case(E_FMAN_EX_IRAM_ECC):
-+ tmp = ioread32be(&fman_rg->fpm_rg->fm_rie);
-+ if (enable) {
-+ /* enable ECC if not enabled */
-+ fman_enable_rams_ecc(fman_rg->fpm_rg);
-+ /* enable ECC interrupts */
-+ tmp |= FPM_IRAM_ECC_ERR_EX_EN;
-+ } else {
-+ /* ECC mechanism may be disabled,
-+ * depending on driver status */
-+ fman_disable_rams_ecc(fman_rg->fpm_rg);
-+ tmp &= ~FPM_IRAM_ECC_ERR_EX_EN;
-+ }
-+ iowrite32be(tmp, &fman_rg->fpm_rg->fm_rie);
-+ break;
-+ case(E_FMAN_EX_MURAM_ECC):
-+ tmp = ioread32be(&fman_rg->fpm_rg->fm_rie);
-+ if (enable) {
-+ /* enable ECC if not enabled */
-+ fman_enable_rams_ecc(fman_rg->fpm_rg);
-+ /* enable ECC interrupts */
-+ tmp |= FPM_MURAM_ECC_ERR_EX_EN;
-+ } else {
-+ /* ECC mechanism may be disabled,
-+ * depending on driver status */
-+ fman_disable_rams_ecc(fman_rg->fpm_rg);
-+ tmp &= ~FPM_MURAM_ECC_ERR_EX_EN;
-+ }
-+ iowrite32be(tmp, &fman_rg->fpm_rg->fm_rie);
-+ break;
-+ default:
-+ return -EINVAL;
-+ }
-+ return 0;
-+}
-+
-+void fman_get_revision(struct fman_fpm_regs *fpm_rg,
-+ uint8_t *major,
-+ uint8_t *minor)
-+{
-+ uint32_t tmp;
-+
-+ tmp = ioread32be(&fpm_rg->fm_ip_rev_1);
-+ *major = (uint8_t)((tmp & FPM_REV1_MAJOR_MASK) >> FPM_REV1_MAJOR_SHIFT);
-+ *minor = (uint8_t)((tmp & FPM_REV1_MINOR_MASK) >> FPM_REV1_MINOR_SHIFT);
-+
-+}
-+
-+uint32_t fman_get_counter(struct fman_rg *fman_rg,
-+ enum fman_counters reg_name)
-+{
-+ uint32_t ret_val;
-+
-+ switch (reg_name) {
-+ case(E_FMAN_COUNTERS_ENQ_TOTAL_FRAME):
-+ ret_val = ioread32be(&fman_rg->qmi_rg->fmqm_etfc);
-+ break;
-+ case(E_FMAN_COUNTERS_DEQ_TOTAL_FRAME):
-+ ret_val = ioread32be(&fman_rg->qmi_rg->fmqm_dtfc);
-+ break;
-+ case(E_FMAN_COUNTERS_DEQ_0):
-+ ret_val = ioread32be(&fman_rg->qmi_rg->fmqm_dc0);
-+ break;
-+ case(E_FMAN_COUNTERS_DEQ_1):
-+ ret_val = ioread32be(&fman_rg->qmi_rg->fmqm_dc1);
-+ break;
-+ case(E_FMAN_COUNTERS_DEQ_2):
-+ ret_val = ioread32be(&fman_rg->qmi_rg->fmqm_dc2);
-+ break;
-+ case(E_FMAN_COUNTERS_DEQ_3):
-+ ret_val = ioread32be(&fman_rg->qmi_rg->fmqm_dc3);
-+ break;
-+ case(E_FMAN_COUNTERS_DEQ_FROM_DEFAULT):
-+ ret_val = ioread32be(&fman_rg->qmi_rg->fmqm_dfdc);
-+ break;
-+ case(E_FMAN_COUNTERS_DEQ_FROM_CONTEXT):
-+ ret_val = ioread32be(&fman_rg->qmi_rg->fmqm_dfcc);
-+ break;
-+ case(E_FMAN_COUNTERS_DEQ_FROM_FD):
-+ ret_val = ioread32be(&fman_rg->qmi_rg->fmqm_dffc);
-+ break;
-+ case(E_FMAN_COUNTERS_DEQ_CONFIRM):
-+ ret_val = ioread32be(&fman_rg->qmi_rg->fmqm_dcc);
-+ break;
-+ default:
-+ ret_val = 0;
-+ }
-+ return ret_val;
-+}
-+
-+int fman_modify_counter(struct fman_rg *fman_rg,
-+ enum fman_counters reg_name,
-+ uint32_t val)
-+{
-+ /* When applicable (when there is an 'enable counters' bit,
-+ * check that counters are enabled */
-+ switch (reg_name) {
-+ case(E_FMAN_COUNTERS_ENQ_TOTAL_FRAME):
-+ case(E_FMAN_COUNTERS_DEQ_TOTAL_FRAME):
-+ case(E_FMAN_COUNTERS_DEQ_0):
-+ case(E_FMAN_COUNTERS_DEQ_1):
-+ case(E_FMAN_COUNTERS_DEQ_2):
-+ case(E_FMAN_COUNTERS_DEQ_3):
-+ case(E_FMAN_COUNTERS_DEQ_FROM_DEFAULT):
-+ case(E_FMAN_COUNTERS_DEQ_FROM_CONTEXT):
-+ case(E_FMAN_COUNTERS_DEQ_FROM_FD):
-+ case(E_FMAN_COUNTERS_DEQ_CONFIRM):
-+ if (!(ioread32be(&fman_rg->qmi_rg->fmqm_gc) &
-+ QMI_CFG_EN_COUNTERS))
-+ return -EINVAL;
-+ break;
-+ default:
-+ break;
-+ }
-+ /* Set counter */
-+ switch (reg_name) {
-+ case(E_FMAN_COUNTERS_ENQ_TOTAL_FRAME):
-+ iowrite32be(val, &fman_rg->qmi_rg->fmqm_etfc);
-+ break;
-+ case(E_FMAN_COUNTERS_DEQ_TOTAL_FRAME):
-+ iowrite32be(val, &fman_rg->qmi_rg->fmqm_dtfc);
-+ break;
-+ case(E_FMAN_COUNTERS_DEQ_0):
-+ iowrite32be(val, &fman_rg->qmi_rg->fmqm_dc0);
-+ break;
-+ case(E_FMAN_COUNTERS_DEQ_1):
-+ iowrite32be(val, &fman_rg->qmi_rg->fmqm_dc1);
-+ break;
-+ case(E_FMAN_COUNTERS_DEQ_2):
-+ iowrite32be(val, &fman_rg->qmi_rg->fmqm_dc2);
-+ break;
-+ case(E_FMAN_COUNTERS_DEQ_3):
-+ iowrite32be(val, &fman_rg->qmi_rg->fmqm_dc3);
-+ break;
-+ case(E_FMAN_COUNTERS_DEQ_FROM_DEFAULT):
-+ iowrite32be(val, &fman_rg->qmi_rg->fmqm_dfdc);
-+ break;
-+ case(E_FMAN_COUNTERS_DEQ_FROM_CONTEXT):
-+ iowrite32be(val, &fman_rg->qmi_rg->fmqm_dfcc);
-+ break;
-+ case(E_FMAN_COUNTERS_DEQ_FROM_FD):
-+ iowrite32be(val, &fman_rg->qmi_rg->fmqm_dffc);
-+ break;
-+ case(E_FMAN_COUNTERS_DEQ_CONFIRM):
-+ iowrite32be(val, &fman_rg->qmi_rg->fmqm_dcc);
-+ break;
-+ case(E_FMAN_COUNTERS_SEMAPHOR_ENTRY_FULL_REJECT):
-+ iowrite32be(val, &fman_rg->dma_rg->fmdmsefrc);
-+ break;
-+ case(E_FMAN_COUNTERS_SEMAPHOR_QUEUE_FULL_REJECT):
-+ iowrite32be(val, &fman_rg->dma_rg->fmdmsqfrc);
-+ break;
-+ case(E_FMAN_COUNTERS_SEMAPHOR_SYNC_REJECT):
-+ iowrite32be(val, &fman_rg->dma_rg->fmdmssrc);
-+ break;
-+ default:
-+ break;
-+ }
-+ return 0;
-+}
-+
-+void fman_set_dma_emergency(struct fman_dma_regs *dma_rg,
-+ bool is_write,
-+ bool enable)
-+{
-+ uint32_t msk;
-+
-+ msk = (uint32_t)(is_write ? DMA_MODE_EMER_WRITE : DMA_MODE_EMER_READ);
-+
-+ if (enable)
-+ iowrite32be(ioread32be(&dma_rg->fmdmmr) | msk,
-+ &dma_rg->fmdmmr);
-+ else /* disable */
-+ iowrite32be(ioread32be(&dma_rg->fmdmmr) & ~msk,
-+ &dma_rg->fmdmmr);
-+}
-+
-+void fman_set_dma_ext_bus_pri(struct fman_dma_regs *dma_rg, uint32_t pri)
-+{
-+ uint32_t tmp;
-+
-+ tmp = ioread32be(&dma_rg->fmdmmr) |
-+ (pri << DMA_MODE_BUS_PRI_SHIFT);
-+
-+ iowrite32be(tmp, &dma_rg->fmdmmr);
-+}
-+
-+uint32_t fman_get_dma_status(struct fman_dma_regs *dma_rg)
-+{
-+ return ioread32be(&dma_rg->fmdmsr);
-+}
-+
-+void fman_force_intr(struct fman_rg *fman_rg,
-+ enum fman_exceptions exception)
-+{
-+ switch (exception) {
-+ case E_FMAN_EX_QMI_DEQ_FROM_UNKNOWN_PORTID:
-+ iowrite32be(QMI_ERR_INTR_EN_DEQ_FROM_DEF,
-+ &fman_rg->qmi_rg->fmqm_eif);
-+ break;
-+ case E_FMAN_EX_QMI_SINGLE_ECC:
-+ iowrite32be(QMI_INTR_EN_SINGLE_ECC,
-+ &fman_rg->qmi_rg->fmqm_if);
-+ break;
-+ case E_FMAN_EX_QMI_DOUBLE_ECC:
-+ iowrite32be(QMI_ERR_INTR_EN_DOUBLE_ECC,
-+ &fman_rg->qmi_rg->fmqm_eif);
-+ break;
-+ case E_FMAN_EX_BMI_LIST_RAM_ECC:
-+ iowrite32be(BMI_ERR_INTR_EN_LIST_RAM_ECC,
-+ &fman_rg->bmi_rg->fmbm_ifr);
-+ break;
-+ case E_FMAN_EX_BMI_STORAGE_PROFILE_ECC:
-+ iowrite32be(BMI_ERR_INTR_EN_STORAGE_PROFILE_ECC,
-+ &fman_rg->bmi_rg->fmbm_ifr);
-+ break;
-+ case E_FMAN_EX_BMI_STATISTICS_RAM_ECC:
-+ iowrite32be(BMI_ERR_INTR_EN_STATISTICS_RAM_ECC,
-+ &fman_rg->bmi_rg->fmbm_ifr);
-+ break;
-+ case E_FMAN_EX_BMI_DISPATCH_RAM_ECC:
-+ iowrite32be(BMI_ERR_INTR_EN_DISPATCH_RAM_ECC,
-+ &fman_rg->bmi_rg->fmbm_ifr);
-+ break;
-+ default:
-+ break;
-+ }
-+}
-+
-+bool fman_is_qmi_halt_not_busy_state(struct fman_qmi_regs *qmi_rg)
-+{
-+ return (bool)!!(ioread32be(&qmi_rg->fmqm_gs) & QMI_GS_HALT_NOT_BUSY);
-+}
-+void fman_resume(struct fman_fpm_regs *fpm_rg)
-+{
-+ uint32_t tmp;
-+
-+ tmp = ioread32be(&fpm_rg->fmfp_ee);
-+ /* clear tmp_reg event bits in order not to clear standing events */
-+ tmp &= ~(FPM_EV_MASK_DOUBLE_ECC |
-+ FPM_EV_MASK_STALL |
-+ FPM_EV_MASK_SINGLE_ECC);
-+ tmp |= FPM_EV_MASK_RELEASE_FM;
-+
-+ iowrite32be(tmp, &fpm_rg->fmfp_ee);
-+}
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/inc/fm_common.h
-@@ -0,0 +1,1214 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/******************************************************************************
-+ @File fm_common.h
-+
-+ @Description FM internal structures and definitions.
-+*//***************************************************************************/
-+#ifndef __FM_COMMON_H
-+#define __FM_COMMON_H
-+
-+#include "error_ext.h"
-+#include "std_ext.h"
-+#include "fm_pcd_ext.h"
-+#include "fm_ext.h"
-+#include "fm_port_ext.h"
-+
-+
-+#define e_FM_PORT_TYPE_OH_HOST_COMMAND e_FM_PORT_TYPE_DUMMY
-+
-+#define CLS_PLAN_NUM_PER_GRP 8
-+
-+#define IP_OFFLOAD_PACKAGE_NUMBER 106
-+#define CAPWAP_OFFLOAD_PACKAGE_NUMBER 108
-+#define IS_OFFLOAD_PACKAGE(num) ((num == IP_OFFLOAD_PACKAGE_NUMBER) || (num == CAPWAP_OFFLOAD_PACKAGE_NUMBER))
-+
-+
-+
-+/**************************************************************************//**
-+ @Description Modules registers offsets
-+*//***************************************************************************/
-+#define FM_MM_MURAM 0x00000000
-+#define FM_MM_BMI 0x00080000
-+#define FM_MM_QMI 0x00080400
-+#define FM_MM_PRS 0x000c7000
-+#define FM_MM_KG 0x000C1000
-+#define FM_MM_DMA 0x000C2000
-+#define FM_MM_FPM 0x000C3000
-+#define FM_MM_PLCR 0x000C0000
-+#define FM_MM_IMEM 0x000C4000
-+#define FM_MM_CGP 0x000DB000
-+#define FM_MM_TRB(i) (0x000D0200 + 0x400 * (i))
-+#if (DPAA_VERSION >= 11)
-+#define FM_MM_SP 0x000dc000
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+
-+/**************************************************************************//**
-+ @Description Enum for inter-module interrupts registration
-+*//***************************************************************************/
-+typedef enum e_FmEventModules{
-+ e_FM_MOD_PRS, /**< Parser event */
-+ e_FM_MOD_KG, /**< Keygen event */
-+ e_FM_MOD_PLCR, /**< Policer event */
-+ e_FM_MOD_10G_MAC, /**< 10G MAC event */
-+ e_FM_MOD_1G_MAC, /**< 1G MAC event */
-+ e_FM_MOD_TMR, /**< Timer event */
-+ e_FM_MOD_FMAN_CTRL, /**< FMAN Controller Timer event */
-+ e_FM_MOD_MACSEC,
-+ e_FM_MOD_DUMMY_LAST
-+} e_FmEventModules;
-+
-+/**************************************************************************//**
-+ @Description Enum for interrupts types
-+*//***************************************************************************/
-+typedef enum e_FmIntrType {
-+ e_FM_INTR_TYPE_ERR,
-+ e_FM_INTR_TYPE_NORMAL
-+} e_FmIntrType;
-+
-+/**************************************************************************//**
-+ @Description Enum for inter-module interrupts registration
-+*//***************************************************************************/
-+typedef enum e_FmInterModuleEvent
-+{
-+ e_FM_EV_PRS = 0, /**< Parser event */
-+ e_FM_EV_ERR_PRS, /**< Parser error event */
-+ e_FM_EV_KG, /**< Keygen event */
-+ e_FM_EV_ERR_KG, /**< Keygen error event */
-+ e_FM_EV_PLCR, /**< Policer event */
-+ e_FM_EV_ERR_PLCR, /**< Policer error event */
-+ e_FM_EV_ERR_10G_MAC0, /**< 10G MAC 0 error event */
-+ e_FM_EV_ERR_10G_MAC1, /**< 10G MAC 1 error event */
-+ e_FM_EV_ERR_1G_MAC0, /**< 1G MAC 0 error event */
-+ e_FM_EV_ERR_1G_MAC1, /**< 1G MAC 1 error event */
-+ e_FM_EV_ERR_1G_MAC2, /**< 1G MAC 2 error event */
-+ e_FM_EV_ERR_1G_MAC3, /**< 1G MAC 3 error event */
-+ e_FM_EV_ERR_1G_MAC4, /**< 1G MAC 4 error event */
-+ e_FM_EV_ERR_1G_MAC5, /**< 1G MAC 5 error event */
-+ e_FM_EV_ERR_1G_MAC6, /**< 1G MAC 6 error event */
-+ e_FM_EV_ERR_1G_MAC7, /**< 1G MAC 7 error event */
-+ e_FM_EV_ERR_MACSEC_MAC0,
-+ e_FM_EV_TMR, /**< Timer event */
-+ e_FM_EV_10G_MAC0, /**< 10G MAC 0 event (Magic packet detection)*/
-+ e_FM_EV_10G_MAC1, /**< 10G MAC 1 event (Magic packet detection)*/
-+ e_FM_EV_1G_MAC0, /**< 1G MAC 0 event (Magic packet detection)*/
-+ e_FM_EV_1G_MAC1, /**< 1G MAC 1 event (Magic packet detection)*/
-+ e_FM_EV_1G_MAC2, /**< 1G MAC 2 (Magic packet detection)*/
-+ e_FM_EV_1G_MAC3, /**< 1G MAC 3 (Magic packet detection)*/
-+ e_FM_EV_1G_MAC4, /**< 1G MAC 4 (Magic packet detection)*/
-+ e_FM_EV_1G_MAC5, /**< 1G MAC 5 (Magic packet detection)*/
-+ e_FM_EV_1G_MAC6, /**< 1G MAC 6 (Magic packet detection)*/
-+ e_FM_EV_1G_MAC7, /**< 1G MAC 7 (Magic packet detection)*/
-+ e_FM_EV_MACSEC_MAC0, /**< MACSEC MAC 0 event */
-+ e_FM_EV_FMAN_CTRL_0, /**< Fman controller event 0 */
-+ e_FM_EV_FMAN_CTRL_1, /**< Fman controller event 1 */
-+ e_FM_EV_FMAN_CTRL_2, /**< Fman controller event 2 */
-+ e_FM_EV_FMAN_CTRL_3, /**< Fman controller event 3 */
-+ e_FM_EV_DUMMY_LAST
-+} e_FmInterModuleEvent;
-+
-+
-+#if defined(__MWERKS__) && !defined(__GNUC__)
-+#pragma pack(push,1)
-+#endif /* defined(__MWERKS__) && ... */
-+
-+/**************************************************************************//**
-+ @Description PCD KG scheme registers
-+*//***************************************************************************/
-+typedef _Packed struct t_FmPcdPlcrProfileRegs {
-+ volatile uint32_t fmpl_pemode; /* 0x090 FMPL_PEMODE - FM Policer Profile Entry Mode*/
-+ volatile uint32_t fmpl_pegnia; /* 0x094 FMPL_PEGNIA - FM Policer Profile Entry GREEN Next Invoked Action*/
-+ volatile uint32_t fmpl_peynia; /* 0x098 FMPL_PEYNIA - FM Policer Profile Entry YELLOW Next Invoked Action*/
-+ volatile uint32_t fmpl_pernia; /* 0x09C FMPL_PERNIA - FM Policer Profile Entry RED Next Invoked Action*/
-+ volatile uint32_t fmpl_pecir; /* 0x0A0 FMPL_PECIR - FM Policer Profile Entry Committed Information Rate*/
-+ volatile uint32_t fmpl_pecbs; /* 0x0A4 FMPL_PECBS - FM Policer Profile Entry Committed Burst Size*/
-+ volatile uint32_t fmpl_pepepir_eir; /* 0x0A8 FMPL_PEPIR_EIR - FM Policer Profile Entry Peak/Excess Information Rate*/
-+ volatile uint32_t fmpl_pepbs_ebs; /* 0x0AC FMPL_PEPBS_EBS - FM Policer Profile Entry Peak/Excess Information Rate*/
-+ volatile uint32_t fmpl_pelts; /* 0x0B0 FMPL_PELTS - FM Policer Profile Entry Last TimeStamp*/
-+ volatile uint32_t fmpl_pects; /* 0x0B4 FMPL_PECTS - FM Policer Profile Entry Committed Token Status*/
-+ volatile uint32_t fmpl_pepts_ets; /* 0x0B8 FMPL_PEPTS_ETS - FM Policer Profile Entry Peak/Excess Token Status*/
-+ volatile uint32_t fmpl_pegpc; /* 0x0BC FMPL_PEGPC - FM Policer Profile Entry GREEN Packet Counter*/
-+ volatile uint32_t fmpl_peypc; /* 0x0C0 FMPL_PEYPC - FM Policer Profile Entry YELLOW Packet Counter*/
-+ volatile uint32_t fmpl_perpc; /* 0x0C4 FMPL_PERPC - FM Policer Profile Entry RED Packet Counter */
-+ volatile uint32_t fmpl_perypc; /* 0x0C8 FMPL_PERYPC - FM Policer Profile Entry Recolored YELLOW Packet Counter*/
-+ volatile uint32_t fmpl_perrpc; /* 0x0CC FMPL_PERRPC - FM Policer Profile Entry Recolored RED Packet Counter*/
-+ volatile uint32_t fmpl_res1[12]; /* 0x0D0-0x0FF Reserved */
-+} _PackedType t_FmPcdPlcrProfileRegs;
-+
-+
-+typedef _Packed struct t_FmPcdCcCapwapReassmTimeoutParams {
-+ volatile uint32_t portIdAndCapwapReassmTbl;
-+ volatile uint32_t fqidForTimeOutFrames;
-+ volatile uint32_t timeoutRequestTime;
-+}_PackedType t_FmPcdCcCapwapReassmTimeoutParams;
-+
-+/**************************************************************************//**
-+ @Description PCD CTRL Parameters Page
-+*//***************************************************************************/
-+typedef _Packed struct t_FmPcdCtrlParamsPage {
-+ volatile uint8_t reserved0[16];
-+ volatile uint32_t iprIpv4Nia;
-+ volatile uint32_t iprIpv6Nia;
-+ volatile uint8_t reserved1[24];
-+ volatile uint32_t ipfOptionsCounter;
-+ volatile uint8_t reserved2[12];
-+ volatile uint32_t misc;
-+ volatile uint32_t errorsDiscardMask;
-+ volatile uint32_t discardMask;
-+ volatile uint8_t reserved3[4];
-+ volatile uint32_t postBmiFetchNia;
-+ volatile uint8_t reserved4[172];
-+} _PackedType t_FmPcdCtrlParamsPage;
-+
-+
-+
-+#if defined(__MWERKS__) && !defined(__GNUC__)
-+#pragma pack(pop)
-+#endif /* defined(__MWERKS__) && ... */
-+
-+
-+/*for UNDER_CONSTRUCTION_FM_RMU_USE_SEC its defined in fm_ext.h*/
-+typedef uint32_t t_FmFmanCtrl;
-+
-+#define FPM_PORT_FM_CTL1 0x00000001
-+#define FPM_PORT_FM_CTL2 0x00000002
-+
-+
-+
-+typedef struct t_FmPcdCcFragScratchPoolCmdParams {
-+ uint32_t numOfBuffers;
-+ uint8_t bufferPoolId;
-+} t_FmPcdCcFragScratchPoolCmdParams;
-+
-+typedef struct t_FmPcdCcReassmTimeoutParams {
-+ bool activate;
-+ uint8_t tsbs;
-+ uint32_t iprcpt;
-+} t_FmPcdCcReassmTimeoutParams;
-+
-+typedef struct {
-+ uint8_t baseEntry;
-+ uint16_t numOfClsPlanEntries;
-+ uint32_t vectors[FM_PCD_MAX_NUM_OF_CLS_PLANS];
-+} t_FmPcdKgInterModuleClsPlanSet;
-+
-+/**************************************************************************//**
-+ @Description Structure for binding a port to keygen schemes.
-+*//***************************************************************************/
-+typedef struct t_FmPcdKgInterModuleBindPortToSchemes {
-+ uint8_t hardwarePortId;
-+ uint8_t netEnvId;
-+ bool useClsPlan; /**< TRUE if this port uses the clsPlan mechanism */
-+ uint8_t numOfSchemes;
-+ uint8_t schemesIds[FM_PCD_KG_NUM_OF_SCHEMES];
-+} t_FmPcdKgInterModuleBindPortToSchemes;
-+
-+typedef struct {
-+ uint32_t nextCcNodeInfo;
-+ t_List node;
-+} t_CcNodeInfo;
-+
-+typedef struct
-+{
-+ t_Handle h_CcNode;
-+ uint16_t index;
-+ t_List node;
-+}t_CcNodeInformation;
-+#define CC_NODE_F_OBJECT(ptr) LIST_OBJECT(ptr, t_CcNodeInformation, node)
-+
-+typedef enum e_ModifyState
-+{
-+ e_MODIFY_STATE_ADD = 0,
-+ e_MODIFY_STATE_REMOVE,
-+ e_MODIFY_STATE_CHANGE
-+} e_ModifyState;
-+
-+typedef struct
-+{
-+ t_Handle h_Manip;
-+ t_List node;
-+}t_ManipInfo;
-+#define CC_NEXT_NODE_F_OBJECT(ptr) LIST_OBJECT(ptr, t_CcNodeInfo, node)
-+
-+typedef struct {
-+ uint32_t type;
-+ uint8_t prOffset;
-+ uint16_t dataOffset;
-+ uint8_t internalBufferOffset;
-+ uint8_t numOfTasks;
-+ uint8_t numOfExtraTasks;
-+ uint8_t hardwarePortId;
-+ t_FmRevisionInfo revInfo;
-+ uint32_t nia;
-+ uint32_t discardMask;
-+} t_GetCcParams;
-+
-+typedef struct {
-+ uint32_t type;
-+ int psoSize;
-+ uint32_t nia;
-+ t_FmFmanCtrl orFmanCtrl;
-+ bool overwrite;
-+ uint8_t ofpDpde;
-+} t_SetCcParams;
-+
-+typedef struct {
-+ t_GetCcParams getCcParams;
-+ t_SetCcParams setCcParams;
-+} t_FmPortGetSetCcParams;
-+
-+typedef struct {
-+ uint32_t type;
-+ bool sleep;
-+} t_FmSetParams;
-+
-+typedef struct {
-+ uint32_t type;
-+ uint32_t fmqm_gs;
-+ uint32_t fm_npi;
-+ uint32_t fm_cld;
-+ uint32_t fmfp_extc;
-+} t_FmGetParams;
-+
-+typedef struct {
-+ t_FmSetParams setParams;
-+ t_FmGetParams getParams;
-+} t_FmGetSetParams;
-+
-+t_Error FmGetSetParams(t_Handle h_Fm, t_FmGetSetParams *p_Params);
-+
-+static __inline__ bool TRY_LOCK(t_Handle h_Spinlock, volatile bool *p_Flag)
-+{
-+ uint32_t intFlags;
-+ if (h_Spinlock)
-+ intFlags = XX_LockIntrSpinlock(h_Spinlock);
-+ else
-+ intFlags = XX_DisableAllIntr();
-+
-+ if (*p_Flag)
-+ {
-+ if (h_Spinlock)
-+ XX_UnlockIntrSpinlock(h_Spinlock, intFlags);
-+ else
-+ XX_RestoreAllIntr(intFlags);
-+ return FALSE;
-+ }
-+ *p_Flag = TRUE;
-+
-+ if (h_Spinlock)
-+ XX_UnlockIntrSpinlock(h_Spinlock, intFlags);
-+ else
-+ XX_RestoreAllIntr(intFlags);
-+
-+ return TRUE;
-+}
-+
-+#define RELEASE_LOCK(_flag) _flag = FALSE;
-+
-+/**************************************************************************//**
-+ @Collection Defines used for manipulation CC and BMI
-+ @{
-+*//***************************************************************************/
-+#define INTERNAL_CONTEXT_OFFSET 0x80000000
-+#define OFFSET_OF_PR 0x40000000
-+#define MANIP_EXTRA_SPACE 0x20000000
-+#define NUM_OF_TASKS 0x10000000
-+#define OFFSET_OF_DATA 0x08000000
-+#define HW_PORT_ID 0x04000000
-+#define FM_REV 0x02000000
-+#define GET_NIA_FPNE 0x01000000
-+#define GET_NIA_PNDN 0x00800000
-+#define NUM_OF_EXTRA_TASKS 0x00400000
-+#define DISCARD_MASK 0x00200000
-+
-+#define UPDATE_NIA_PNEN 0x80000000
-+#define UPDATE_PSO 0x40000000
-+#define UPDATE_NIA_PNDN 0x20000000
-+#define UPDATE_FMFP_PRC_WITH_ONE_RISC_ONLY 0x10000000
-+#define UPDATE_OFP_DPTE 0x08000000
-+#define UPDATE_NIA_FENE 0x04000000
-+#define UPDATE_NIA_CMNE 0x02000000
-+#define UPDATE_NIA_FPNE 0x01000000
-+/* @} */
-+
-+/**************************************************************************//**
-+ @Collection Defines used for manipulation CC and CC
-+ @{
-+*//***************************************************************************/
-+#define UPDATE_NIA_ENQ_WITHOUT_DMA 0x80000000
-+#define UPDATE_CC_WITH_TREE 0x40000000
-+#define UPDATE_CC_WITH_DELETE_TREE 0x20000000
-+#define UPDATE_KG_NIA_CC_WA 0x10000000
-+#define UPDATE_KG_OPT_MODE 0x08000000
-+#define UPDATE_KG_NIA 0x04000000
-+#define UPDATE_CC_SHADOW_CLEAR 0x02000000
-+/* @} */
-+
-+#define UPDATE_FPM_BRKC_SLP 0x80000000
-+#define UPDATE_FPM_EXTC 0x40000000
-+#define UPDATE_FPM_EXTC_CLEAR 0x20000000
-+#define GET_FMQM_GS 0x10000000
-+#define GET_FM_NPI 0x08000000
-+#define GET_FMFP_EXTC 0x04000000
-+#define CLEAR_IRAM_READY 0x02000000
-+#define UPDATE_FM_CLD 0x01000000
-+#define GET_FM_CLD 0x00800000
-+#define FM_MAX_NUM_OF_PORTS (FM_MAX_NUM_OF_OH_PORTS + \
-+ FM_MAX_NUM_OF_1G_RX_PORTS + \
-+ FM_MAX_NUM_OF_10G_RX_PORTS + \
-+ FM_MAX_NUM_OF_1G_TX_PORTS + \
-+ FM_MAX_NUM_OF_10G_TX_PORTS)
-+
-+#define MODULE_NAME_SIZE 30
-+#define DUMMY_PORT_ID 0
-+
-+#define FM_LIODN_OFFSET_MASK 0x3FF
-+
-+/**************************************************************************//**
-+ @Description NIA Description
-+*//***************************************************************************/
-+#define NIA_ENG_MASK 0x007C0000
-+#define NIA_AC_MASK 0x0003ffff
-+
-+#define NIA_ORDER_RESTOR 0x00800000
-+#define NIA_ENG_FM_CTL 0x00000000
-+#define NIA_ENG_PRS 0x00440000
-+#define NIA_ENG_KG 0x00480000
-+#define NIA_ENG_PLCR 0x004C0000
-+#define NIA_ENG_BMI 0x00500000
-+#define NIA_ENG_QMI_ENQ 0x00540000
-+#define NIA_ENG_QMI_DEQ 0x00580000
-+
-+#define NIA_FM_CTL_AC_CC 0x00000006
-+#define NIA_FM_CTL_AC_HC 0x0000000C
-+#define NIA_FM_CTL_AC_IND_MODE_TX 0x00000008
-+#define NIA_FM_CTL_AC_IND_MODE_RX 0x0000000A
-+#define NIA_FM_CTL_AC_POP_TO_N_STEP 0x0000000e
-+#define NIA_FM_CTL_AC_PRE_BMI_FETCH_HEADER 0x00000010
-+#define NIA_FM_CTL_AC_PRE_BMI_FETCH_FULL_FRAME 0x00000018
-+#define NIA_FM_CTL_AC_POST_BMI_FETCH 0x00000012
-+#define NIA_FM_CTL_AC_PRE_BMI_ENQ_FRAME 0x0000001A
-+#define NIA_FM_CTL_AC_PRE_BMI_DISCARD_FRAME 0x0000001E
-+#define NIA_FM_CTL_AC_POST_BMI_ENQ_ORR 0x00000014
-+#define NIA_FM_CTL_AC_POST_BMI_ENQ 0x00000022
-+#define NIA_FM_CTL_AC_PRE_CC 0x00000020
-+#define NIA_FM_CTL_AC_POST_TX 0x00000024
-+/* V3 only */
-+#define NIA_FM_CTL_AC_NO_IPACC_PRE_BMI_ENQ_FRAME 0x00000028
-+#define NIA_FM_CTL_AC_NO_IPACC_PRE_BMI_DISCARD_FRAME 0x0000002A
-+#define NIA_FM_CTL_AC_NO_IPACC_POP_TO_N_STEP 0x0000002C
-+
-+#define NIA_BMI_AC_ENQ_FRAME 0x00000002
-+#define NIA_BMI_AC_TX_RELEASE 0x000002C0
-+#define NIA_BMI_AC_RELEASE 0x000000C0
-+#define NIA_BMI_AC_DISCARD 0x000000C1
-+#define NIA_BMI_AC_TX 0x00000274
-+#define NIA_BMI_AC_FETCH 0x00000208
-+#define NIA_BMI_AC_MASK 0x000003FF
-+
-+#define NIA_KG_DIRECT 0x00000100
-+#define NIA_KG_CC_EN 0x00000200
-+#define NIA_PLCR_ABSOLUTE 0x00008000
-+
-+#define NIA_BMI_AC_ENQ_FRAME_WITHOUT_DMA 0x00000202
-+
-+#if defined(FM_OP_NO_VSP_NO_RELEASE_ERRATA_FMAN_A006675) || defined(FM_ERROR_VSP_NO_MATCH_SW006)
-+#define GET_NIA_BMI_AC_ENQ_FRAME(h_FmPcd) \
-+ (uint32_t)((FmPcdIsAdvancedOffloadSupported(h_FmPcd)) ? \
-+ (NIA_ENG_FM_CTL | NIA_FM_CTL_AC_PRE_BMI_ENQ_FRAME) : \
-+ (NIA_ENG_FM_CTL | NIA_FM_CTL_AC_NO_IPACC_PRE_BMI_ENQ_FRAME))
-+#define GET_NIA_BMI_AC_DISCARD_FRAME(h_FmPcd) \
-+ (uint32_t)((FmPcdIsAdvancedOffloadSupported(h_FmPcd)) ? \
-+ (NIA_ENG_FM_CTL | NIA_FM_CTL_AC_PRE_BMI_DISCARD_FRAME) : \
-+ (NIA_ENG_FM_CTL | NIA_FM_CTL_AC_NO_IPACC_PRE_BMI_DISCARD_FRAME))
-+#define GET_NO_PCD_NIA_BMI_AC_ENQ_FRAME() \
-+ (NIA_ENG_FM_CTL | NIA_FM_CTL_AC_NO_IPACC_PRE_BMI_ENQ_FRAME)
-+#else
-+#define GET_NIA_BMI_AC_ENQ_FRAME(h_FmPcd) \
-+ (uint32_t)((FmPcdIsAdvancedOffloadSupported(h_FmPcd)) ? \
-+ (NIA_ENG_FM_CTL | NIA_FM_CTL_AC_PRE_BMI_ENQ_FRAME) : \
-+ (NIA_ENG_BMI | NIA_BMI_AC_ENQ_FRAME))
-+#define GET_NIA_BMI_AC_DISCARD_FRAME(h_FmPcd) \
-+ (uint32_t)((FmPcdIsAdvancedOffloadSupported(h_FmPcd)) ? \
-+ (NIA_ENG_FM_CTL | NIA_FM_CTL_AC_PRE_BMI_DISCARD_FRAME) : \
-+ (NIA_ENG_BMI | NIA_BMI_AC_DISCARD))
-+#define GET_NO_PCD_NIA_BMI_AC_ENQ_FRAME() \
-+ (NIA_ENG_BMI | NIA_BMI_AC_ENQ_FRAME)
-+#endif /* defined(FM_OP_NO_VSP_NO_RELEASE_ERRATA_FMAN_A006675) || ... */
-+
-+/**************************************************************************//**
-+ @Description CTRL Parameters Page defines
-+*//***************************************************************************/
-+#define FM_CTL_PARAMS_PAGE_OP_FIX_EN 0x80000000
-+#define FM_CTL_PARAMS_PAGE_OFFLOAD_SUPPORT_EN 0x40000000
-+#define FM_CTL_PARAMS_PAGE_ALWAYS_ON 0x00000100
-+
-+#define FM_CTL_PARAMS_PAGE_ERROR_VSP_MASK 0x0000003f
-+
-+/**************************************************************************//**
-+ @Description Port Id defines
-+*//***************************************************************************/
-+#if (DPAA_VERSION == 10)
-+#define BASE_OH_PORTID 1
-+#else
-+#define BASE_OH_PORTID 2
-+#endif /* (DPAA_VERSION == 10) */
-+#define BASE_1G_RX_PORTID 8
-+#define BASE_10G_RX_PORTID 0x10
-+#define BASE_1G_TX_PORTID 0x28
-+#define BASE_10G_TX_PORTID 0x30
-+
-+#define FM_PCD_PORT_OH_BASE_INDX 0
-+#define FM_PCD_PORT_1G_RX_BASE_INDX (FM_PCD_PORT_OH_BASE_INDX+FM_MAX_NUM_OF_OH_PORTS)
-+#define FM_PCD_PORT_10G_RX_BASE_INDX (FM_PCD_PORT_1G_RX_BASE_INDX+FM_MAX_NUM_OF_1G_RX_PORTS)
-+#define FM_PCD_PORT_1G_TX_BASE_INDX (FM_PCD_PORT_10G_RX_BASE_INDX+FM_MAX_NUM_OF_10G_RX_PORTS)
-+#define FM_PCD_PORT_10G_TX_BASE_INDX (FM_PCD_PORT_1G_TX_BASE_INDX+FM_MAX_NUM_OF_1G_TX_PORTS)
-+
-+#if (FM_MAX_NUM_OF_OH_PORTS > 0)
-+#define CHECK_PORT_ID_OH_PORTS(_relativePortId) \
-+ if ((_relativePortId) >= FM_MAX_NUM_OF_OH_PORTS) \
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE, ("Illegal OH_PORT port id"))
-+#else
-+#define CHECK_PORT_ID_OH_PORTS(_relativePortId) \
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE, ("Illegal OH_PORT port id"))
-+#endif
-+#if (FM_MAX_NUM_OF_1G_RX_PORTS > 0)
-+#define CHECK_PORT_ID_1G_RX_PORTS(_relativePortId) \
-+ if ((_relativePortId) >= FM_MAX_NUM_OF_1G_RX_PORTS) \
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE, ("Illegal 1G_RX_PORT port id"))
-+#else
-+#define CHECK_PORT_ID_1G_RX_PORTS(_relativePortId) \
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE, ("Illegal 1G_RX_PORT port id"))
-+#endif
-+#if (FM_MAX_NUM_OF_10G_RX_PORTS > 0)
-+#define CHECK_PORT_ID_10G_RX_PORTS(_relativePortId) \
-+ if ((_relativePortId) >= FM_MAX_NUM_OF_10G_RX_PORTS) \
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE, ("Illegal 10G_RX_PORT port id"))
-+#else
-+#define CHECK_PORT_ID_10G_RX_PORTS(_relativePortId) \
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE, ("Illegal 10G_RX_PORT port id"))
-+#endif
-+#if (FM_MAX_NUM_OF_1G_TX_PORTS > 0)
-+#define CHECK_PORT_ID_1G_TX_PORTS(_relativePortId) \
-+ if ((_relativePortId) >= FM_MAX_NUM_OF_1G_TX_PORTS) \
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE, ("Illegal 1G_TX_PORT port id"))
-+#else
-+#define CHECK_PORT_ID_1G_TX_PORTS(_relativePortId) \
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE, ("Illegal 1G_TX_PORT port id"))
-+#endif
-+#if (FM_MAX_NUM_OF_10G_TX_PORTS > 0)
-+#define CHECK_PORT_ID_10G_TX_PORTS(_relativePortId) \
-+ if ((_relativePortId) >= FM_MAX_NUM_OF_10G_TX_PORTS) \
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE, ("Illegal 10G_TX_PORT port id"))
-+#else
-+#define CHECK_PORT_ID_10G_TX_PORTS(_relativePortId) \
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE, ("Illegal 10G_TX_PORT port id"))
-+#endif
-+
-+uint8_t SwPortIdToHwPortId(e_FmPortType type, uint8_t relativePortId, uint8_t majorRev, uint8_t minorRev);
-+
-+#define HW_PORT_ID_TO_SW_PORT_ID(_relativePortId, hardwarePortId) \
-+{ if (((hardwarePortId) >= BASE_OH_PORTID) && \
-+ ((hardwarePortId) < BASE_OH_PORTID+FM_MAX_NUM_OF_OH_PORTS)) \
-+ _relativePortId = (uint8_t)((hardwarePortId)-BASE_OH_PORTID); \
-+ else if (((hardwarePortId) >= BASE_10G_TX_PORTID) && \
-+ ((hardwarePortId) < BASE_10G_TX_PORTID+FM_MAX_NUM_OF_10G_TX_PORTS)) \
-+ _relativePortId = (uint8_t)((hardwarePortId)-BASE_10G_TX_PORTID); \
-+ else if (((hardwarePortId) >= BASE_1G_TX_PORTID) && \
-+ ((hardwarePortId) < BASE_1G_TX_PORTID+FM_MAX_NUM_OF_1G_TX_PORTS)) \
-+ _relativePortId = (uint8_t)((hardwarePortId)-BASE_1G_TX_PORTID); \
-+ else if (((hardwarePortId) >= BASE_10G_RX_PORTID) && \
-+ ((hardwarePortId) < BASE_10G_RX_PORTID+FM_MAX_NUM_OF_10G_RX_PORTS)) \
-+ _relativePortId = (uint8_t)((hardwarePortId)-BASE_10G_RX_PORTID); \
-+ else if (((hardwarePortId) >= BASE_1G_RX_PORTID) && \
-+ ((hardwarePortId) < BASE_1G_RX_PORTID+FM_MAX_NUM_OF_1G_RX_PORTS)) \
-+ _relativePortId = (uint8_t)((hardwarePortId)-BASE_1G_RX_PORTID); \
-+ else { \
-+ _relativePortId = (uint8_t)DUMMY_PORT_ID; \
-+ ASSERT_COND(TRUE); \
-+ } \
-+}
-+
-+#define HW_PORT_ID_TO_SW_PORT_INDX(swPortIndex, hardwarePortId) \
-+do { \
-+ if (((hardwarePortId) >= BASE_OH_PORTID) && ((hardwarePortId) < BASE_OH_PORTID+FM_MAX_NUM_OF_OH_PORTS)) \
-+ swPortIndex = (uint8_t)((hardwarePortId)-BASE_OH_PORTID+FM_PCD_PORT_OH_BASE_INDX); \
-+ else if (((hardwarePortId) >= BASE_1G_RX_PORTID) && \
-+ ((hardwarePortId) < BASE_1G_RX_PORTID+FM_MAX_NUM_OF_1G_RX_PORTS)) \
-+ swPortIndex = (uint8_t)((hardwarePortId)-BASE_1G_RX_PORTID+FM_PCD_PORT_1G_RX_BASE_INDX); \
-+ else if (((hardwarePortId) >= BASE_10G_RX_PORTID) && \
-+ ((hardwarePortId) < BASE_10G_RX_PORTID+FM_MAX_NUM_OF_10G_RX_PORTS)) \
-+ swPortIndex = (uint8_t)((hardwarePortId)-BASE_10G_RX_PORTID+FM_PCD_PORT_10G_RX_BASE_INDX); \
-+ else if (((hardwarePortId) >= BASE_1G_TX_PORTID) && \
-+ ((hardwarePortId) < BASE_1G_TX_PORTID+FM_MAX_NUM_OF_1G_TX_PORTS)) \
-+ swPortIndex = (uint8_t)((hardwarePortId)-BASE_1G_TX_PORTID+FM_PCD_PORT_1G_TX_BASE_INDX); \
-+ else if (((hardwarePortId) >= BASE_10G_TX_PORTID) && \
-+ ((hardwarePortId) < BASE_10G_TX_PORTID+FM_MAX_NUM_OF_10G_TX_PORTS)) \
-+ swPortIndex = (uint8_t)((hardwarePortId)-BASE_10G_TX_PORTID+FM_PCD_PORT_10G_TX_BASE_INDX); \
-+ else ASSERT_COND(FALSE); \
-+} while (0)
-+
-+#define SW_PORT_INDX_TO_HW_PORT_ID(hardwarePortId, swPortIndex) \
-+do { \
-+ if (((swPortIndex) >= FM_PCD_PORT_OH_BASE_INDX) && ((swPortIndex) < FM_PCD_PORT_1G_RX_BASE_INDX)) \
-+ hardwarePortId = (uint8_t)((swPortIndex)-FM_PCD_PORT_OH_BASE_INDX+BASE_OH_PORTID); \
-+ else if (((swPortIndex) >= FM_PCD_PORT_1G_RX_BASE_INDX) && ((swPortIndex) < FM_PCD_PORT_10G_RX_BASE_INDX)) \
-+ hardwarePortId = (uint8_t)((swPortIndex)-FM_PCD_PORT_1G_RX_BASE_INDX+BASE_1G_RX_PORTID); \
-+ else if (((swPortIndex) >= FM_PCD_PORT_10G_RX_BASE_INDX) && ((swPortIndex) < FM_MAX_NUM_OF_PORTS)) \
-+ hardwarePortId = (uint8_t)((swPortIndex)-FM_PCD_PORT_10G_RX_BASE_INDX+BASE_10G_RX_PORTID); \
-+ else if (((swPortIndex) >= FM_PCD_PORT_1G_TX_BASE_INDX) && ((swPortIndex) < FM_PCD_PORT_10G_TX_BASE_INDX)) \
-+ hardwarePortId = (uint8_t)((swPortIndex)-FM_PCD_PORT_1G_TX_BASE_INDX+BASE_1G_TX_PORTID); \
-+ else if (((swPortIndex) >= FM_PCD_PORT_10G_TX_BASE_INDX) && ((swPortIndex) < FM_MAX_NUM_OF_PORTS)) \
-+ hardwarePortId = (uint8_t)((swPortIndex)-FM_PCD_PORT_10G_TX_BASE_INDX+BASE_10G_TX_PORTID); \
-+ else ASSERT_COND(FALSE); \
-+} while (0)
-+
-+#define BMI_MAX_FIFO_SIZE (FM_MURAM_SIZE)
-+#define BMI_FIFO_UNITS 0x100
-+
-+typedef struct {
-+ void (*f_Isr) (t_Handle h_Arg);
-+ t_Handle h_SrcHandle;
-+ uint8_t guestId;
-+} t_FmIntrSrc;
-+
-+#define ILLEGAL_HDR_NUM 0xFF
-+#define NO_HDR_NUM FM_PCD_PRS_NUM_OF_HDRS
-+
-+#define IS_PRIVATE_HEADER(hdr) (((hdr) == HEADER_TYPE_USER_DEFINED_SHIM1) || \
-+ ((hdr) == HEADER_TYPE_USER_DEFINED_SHIM2))
-+#define IS_SPECIAL_HEADER(hdr) ((hdr) == HEADER_TYPE_MACSEC)
-+
-+static __inline__ uint8_t GetPrsHdrNum(e_NetHeaderType hdr)
-+{
-+ switch (hdr)
-+ { case (HEADER_TYPE_ETH): return 0;
-+ case (HEADER_TYPE_LLC_SNAP): return 1;
-+ case (HEADER_TYPE_VLAN): return 2;
-+ case (HEADER_TYPE_PPPoE): return 3;
-+ case (HEADER_TYPE_PPP): return 3;
-+ case (HEADER_TYPE_MPLS): return 4;
-+ case (HEADER_TYPE_IPv4): return 5;
-+ case (HEADER_TYPE_IPv6): return 6;
-+ case (HEADER_TYPE_GRE): return 7;
-+ case (HEADER_TYPE_MINENCAP): return 8;
-+ case (HEADER_TYPE_USER_DEFINED_L3): return 9;
-+ case (HEADER_TYPE_TCP): return 10;
-+ case (HEADER_TYPE_UDP): return 11;
-+ case (HEADER_TYPE_IPSEC_AH):
-+ case (HEADER_TYPE_IPSEC_ESP): return 12;
-+ case (HEADER_TYPE_SCTP): return 13;
-+ case (HEADER_TYPE_DCCP): return 14;
-+ case (HEADER_TYPE_USER_DEFINED_L4): return 15;
-+ case (HEADER_TYPE_USER_DEFINED_SHIM1):
-+ case (HEADER_TYPE_USER_DEFINED_SHIM2):
-+ case (HEADER_TYPE_MACSEC): return NO_HDR_NUM;
-+ default:
-+ return ILLEGAL_HDR_NUM;
-+ }
-+}
-+
-+#define FM_PCD_MAX_NUM_OF_OPTIONS(clsPlanEntries) ((clsPlanEntries==256)? 8:((clsPlanEntries==128)? 7: ((clsPlanEntries==64)? 6: ((clsPlanEntries==32)? 5:0))))
-+
-+
-+/**************************************************************************//**
-+ @Description A structure for initializing a keygen classification plan group
-+*//***************************************************************************/
-+typedef struct t_FmPcdKgInterModuleClsPlanGrpParams {
-+ uint8_t netEnvId; /* IN */
-+ bool grpExists; /* OUT (unused in FmPcdKgBuildClsPlanGrp)*/
-+ uint8_t clsPlanGrpId; /* OUT */
-+ bool emptyClsPlanGrp; /* OUT */
-+ uint8_t numOfOptions; /* OUT in FmPcdGetSetClsPlanGrpParams IN in FmPcdKgBuildClsPlanGrp*/
-+ protocolOpt_t options[FM_PCD_MAX_NUM_OF_OPTIONS(FM_PCD_MAX_NUM_OF_CLS_PLANS)];
-+ /* OUT in FmPcdGetSetClsPlanGrpParams IN in FmPcdKgBuildClsPlanGrp*/
-+ uint32_t optVectors[FM_PCD_MAX_NUM_OF_OPTIONS(FM_PCD_MAX_NUM_OF_CLS_PLANS)];
-+ /* OUT in FmPcdGetSetClsPlanGrpParams IN in FmPcdKgBuildClsPlanGrp*/
-+} t_FmPcdKgInterModuleClsPlanGrpParams;
-+
-+typedef struct t_FmPcdLock {
-+ t_Handle h_Spinlock;
-+ volatile bool flag;
-+ t_List node;
-+} t_FmPcdLock;
-+#define FM_PCD_LOCK_OBJ(ptr) LIST_OBJECT(ptr, t_FmPcdLock, node)
-+
-+
-+typedef t_Error (t_FmPortGetSetCcParamsCallback) (t_Handle h_FmPort,
-+ t_FmPortGetSetCcParams *p_FmPortGetSetCcParams);
-+
-+
-+/***********************************************************************/
-+/* Common API for FM-PCD module */
-+/***********************************************************************/
-+t_Handle FmPcdGetHcHandle(t_Handle h_FmPcd);
-+uint32_t FmPcdGetSwPrsOffset(t_Handle h_FmPcd, e_NetHeaderType hdr, uint8_t indexPerHdr);
-+uint32_t FmPcdGetLcv(t_Handle h_FmPcd, uint32_t netEnvId, uint8_t hdrNum);
-+uint32_t FmPcdGetMacsecLcv(t_Handle h_FmPcd, uint32_t netEnvId);
-+void FmPcdIncNetEnvOwners(t_Handle h_FmPcd, uint8_t netEnvId);
-+void FmPcdDecNetEnvOwners(t_Handle h_FmPcd, uint8_t netEnvId);
-+uint8_t FmPcdGetNetEnvId(t_Handle h_NetEnv);
-+void FmPcdPortRegister(t_Handle h_FmPcd, t_Handle h_FmPort, uint8_t hardwarePortId);
-+uint32_t FmPcdLock(t_Handle h_FmPcd);
-+void FmPcdUnlock(t_Handle h_FmPcd, uint32_t intFlags);
-+bool FmPcdNetEnvIsHdrExist(t_Handle h_FmPcd, uint8_t netEnvId, e_NetHeaderType hdr);
-+t_Error FmPcdFragHcScratchPoolInit(t_Handle h_FmPcd, uint8_t scratchBpid);
-+t_Error FmPcdRegisterReassmPort(t_Handle h_FmPcd, t_Handle h_IpReasmCommonPramTbl);
-+t_Error FmPcdUnregisterReassmPort(t_Handle h_FmPcd, t_Handle h_IpReasmCommonPramTbl);
-+bool FmPcdIsAdvancedOffloadSupported(t_Handle h_FmPcd);
-+bool FmPcdLockTryLockAll(t_Handle h_FmPcd);
-+void FmPcdLockUnlockAll(t_Handle h_FmPcd);
-+t_Error FmPcdHcSync(t_Handle h_FmPcd);
-+t_Handle FmGetPcd(t_Handle h_Fm);
-+/***********************************************************************/
-+/* Common API for FM-PCD KG module */
-+/***********************************************************************/
-+uint8_t FmPcdKgGetClsPlanGrpBase(t_Handle h_FmPcd, uint8_t clsPlanGrp);
-+uint16_t FmPcdKgGetClsPlanGrpSize(t_Handle h_FmPcd, uint8_t clsPlanGrp);
-+t_Error FmPcdKgBuildClsPlanGrp(t_Handle h_FmPcd, t_FmPcdKgInterModuleClsPlanGrpParams *p_Grp, t_FmPcdKgInterModuleClsPlanSet *p_ClsPlanSet);
-+
-+uint8_t FmPcdKgGetSchemeId(t_Handle h_Scheme);
-+#if (DPAA_VERSION >= 11)
-+bool FmPcdKgGetVspe(t_Handle h_Scheme);
-+#endif /* (DPAA_VERSION >= 11) */
-+uint8_t FmPcdKgGetRelativeSchemeId(t_Handle h_FmPcd, uint8_t schemeId);
-+void FmPcdKgDestroyClsPlanGrp(t_Handle h_FmPcd, uint8_t grpId);
-+t_Error FmPcdKgCheckInvalidateSchemeSw(t_Handle h_Scheme);
-+t_Error FmPcdKgBuildBindPortToSchemes(t_Handle h_FmPcd , t_FmPcdKgInterModuleBindPortToSchemes *p_BindPortToSchemes, uint32_t *p_SpReg, bool add);
-+bool FmPcdKgHwSchemeIsValid(uint32_t schemeModeReg);
-+uint32_t FmPcdKgBuildWriteSchemeActionReg(uint8_t schemeId, bool updateCounter);
-+uint32_t FmPcdKgBuildReadSchemeActionReg(uint8_t schemeId);
-+uint32_t FmPcdKgBuildWriteClsPlanBlockActionReg(uint8_t grpId);
-+uint32_t FmPcdKgBuildWritePortSchemeBindActionReg(uint8_t hardwarePortId);
-+uint32_t FmPcdKgBuildReadPortSchemeBindActionReg(uint8_t hardwarePortId);
-+uint32_t FmPcdKgBuildWritePortClsPlanBindActionReg(uint8_t hardwarePortId);
-+bool FmPcdKgIsSchemeValidSw(t_Handle h_Scheme);
-+
-+t_Error FmPcdKgBindPortToSchemes(t_Handle h_FmPcd , t_FmPcdKgInterModuleBindPortToSchemes *p_SchemeBind);
-+t_Error FmPcdKgUnbindPortToSchemes(t_Handle h_FmPcd , t_FmPcdKgInterModuleBindPortToSchemes *p_SchemeBind);
-+uint32_t FmPcdKgGetRequiredAction(t_Handle h_FmPcd, uint8_t schemeId);
-+uint32_t FmPcdKgGetRequiredActionFlag(t_Handle h_FmPcd, uint8_t schemeId);
-+e_FmPcdDoneAction FmPcdKgGetDoneAction(t_Handle h_FmPcd, uint8_t schemeId);
-+e_FmPcdEngine FmPcdKgGetNextEngine(t_Handle h_FmPcd, uint8_t schemeId);
-+void FmPcdKgUpdateRequiredAction(t_Handle h_Scheme, uint32_t requiredAction);
-+bool FmPcdKgIsDirectPlcr(t_Handle h_FmPcd, uint8_t schemeId);
-+bool FmPcdKgIsDistrOnPlcrProfile(t_Handle h_FmPcd, uint8_t schemeId);
-+uint16_t FmPcdKgGetRelativeProfileId(t_Handle h_FmPcd, uint8_t schemeId);
-+t_Handle FmPcdKgGetSchemeHandle(t_Handle h_FmPcd, uint8_t relativeSchemeId);
-+bool FmPcdKgIsSchemeHasOwners(t_Handle h_Scheme);
-+t_Error FmPcdKgCcGetSetParams(t_Handle h_FmPcd, t_Handle h_Scheme, uint32_t requiredAction, uint32_t value);
-+t_Error FmPcdKgSetOrBindToClsPlanGrp(t_Handle h_FmPcd, uint8_t hardwarePortId, uint8_t netEnvId, protocolOpt_t *p_OptArray, uint8_t *p_ClsPlanGrpId, bool *p_IsEmptyClsPlanGrp);
-+t_Error FmPcdKgDeleteOrUnbindPortToClsPlanGrp(t_Handle h_FmPcd, uint8_t hardwarePortId, uint8_t clsPlanGrpId);
-+
-+/***********************************************************************/
-+/* Common API for FM-PCD parser module */
-+/***********************************************************************/
-+t_Error FmPcdPrsIncludePortInStatistics(t_Handle p_FmPcd, uint8_t hardwarePortId, bool include);
-+
-+/***********************************************************************/
-+/* Common API for FM-PCD policer module */
-+/***********************************************************************/
-+t_Error FmPcdPlcrAllocProfiles(t_Handle h_FmPcd, uint8_t hardwarePortId, uint16_t numOfProfiles);
-+t_Error FmPcdPlcrFreeProfiles(t_Handle h_FmPcd, uint8_t hardwarePortId);
-+bool FmPcdPlcrIsProfileValid(t_Handle h_FmPcd, uint16_t absoluteProfileId);
-+uint16_t FmPcdPlcrGetPortProfilesBase(t_Handle h_FmPcd, uint8_t hardwarePortId);
-+uint16_t FmPcdPlcrGetPortNumOfProfiles(t_Handle h_FmPcd, uint8_t hardwarePortId);
-+uint32_t FmPcdPlcrBuildWritePlcrActionRegs(uint16_t absoluteProfileId);
-+uint32_t FmPcdPlcrBuildCounterProfileReg(e_FmPcdPlcrProfileCounters counter);
-+uint32_t FmPcdPlcrBuildWritePlcrActionReg(uint16_t absoluteProfileId);
-+uint32_t FmPcdPlcrBuildReadPlcrActionReg(uint16_t absoluteProfileId);
-+uint16_t FmPcdPlcrProfileGetAbsoluteId(t_Handle h_Profile);
-+t_Error FmPcdPlcrGetAbsoluteIdByProfileParams(t_Handle h_FmPcd,
-+ e_FmPcdProfileTypeSelection profileType,
-+ t_Handle h_FmPort,
-+ uint16_t relativeProfile,
-+ uint16_t *p_AbsoluteId);
-+void FmPcdPlcrInvalidateProfileSw(t_Handle h_FmPcd, uint16_t absoluteProfileId);
-+void FmPcdPlcrValidateProfileSw(t_Handle h_FmPcd, uint16_t absoluteProfileId);
-+bool FmPcdPlcrHwProfileIsValid(uint32_t profileModeReg);
-+uint32_t FmPcdPlcrGetRequiredAction(t_Handle h_FmPcd, uint16_t absoluteProfileId);
-+uint32_t FmPcdPlcrGetRequiredActionFlag(t_Handle h_FmPcd, uint16_t absoluteProfileId);
-+uint32_t FmPcdPlcrBuildNiaProfileReg(bool green, bool yellow, bool red);
-+void FmPcdPlcrUpdateRequiredAction(t_Handle h_FmPcd, uint16_t absoluteProfileId, uint32_t requiredAction);
-+t_Error FmPcdPlcrCcGetSetParams(t_Handle h_FmPcd, uint16_t profileIndx,uint32_t requiredAction);
-+
-+/***********************************************************************/
-+/* Common API for FM-PCD CC module */
-+/***********************************************************************/
-+uint8_t FmPcdCcGetParseCode(t_Handle h_CcNode);
-+uint8_t FmPcdCcGetOffset(t_Handle h_CcNode);
-+t_Error FmPcdCcRemoveKey(t_Handle h_FmPcd, t_Handle h_FmPcdCcNode, uint16_t keyIndex);
-+t_Error FmPcdCcAddKey(t_Handle h_FmPcd, t_Handle h_CcNode, uint16_t keyIndex, uint8_t keySize, t_FmPcdCcKeyParams *p_FmPCdCcKeyParams);
-+t_Error FmPcdCcModifyKey(t_Handle h_FmPcd, t_Handle h_CcNode, uint16_t keyIndex, uint8_t keySize, uint8_t *p_Key, uint8_t *p_Mask);
-+t_Error FmPcdCcModifyKeyAndNextEngine(t_Handle h_FmPcd, t_Handle h_FmPcdCcNode, uint16_t keyIndex, uint8_t keySize, t_FmPcdCcKeyParams *p_FmPcdCcKeyParams);
-+t_Error FmPcdCcModifyMissNextEngineParamNode(t_Handle h_FmPcd,t_Handle h_FmPcdCcNode, t_FmPcdCcNextEngineParams *p_FmPcdCcNextEngineParams);
-+t_Error FmPcdCcModifyNextEngineParamTree(t_Handle h_FmPcd, t_Handle h_FmPcdCcTree, uint8_t grpId, uint8_t index, t_FmPcdCcNextEngineParams *p_FmPcdCcNextEngineParams);
-+uint32_t FmPcdCcGetNodeAddrOffsetFromNodeInfo(t_Handle h_FmPcd, t_Handle h_Pointer);
-+t_Handle FmPcdCcTreeGetSavedManipParams(t_Handle h_FmTree);
-+void FmPcdCcTreeSetSavedManipParams(t_Handle h_FmTree, t_Handle h_SavedManipParams);
-+t_Error FmPcdCcTreeAddIPR(t_Handle h_FmPcd, t_Handle h_FmTree, t_Handle h_NetEnv, t_Handle h_ReassemblyManip, bool schemes);
-+t_Error FmPcdCcTreeAddCPR(t_Handle h_FmPcd, t_Handle h_FmTree, t_Handle h_NetEnv, t_Handle h_ReassemblyManip, bool schemes);
-+t_Error FmPcdCcBindTree(t_Handle h_FmPcd, t_Handle h_PcdParams, t_Handle h_CcTree, uint32_t *p_Offset,t_Handle h_FmPort);
-+t_Error FmPcdCcUnbindTree(t_Handle h_FmPcd, t_Handle h_CcTree);
-+
-+/***********************************************************************/
-+/* Common API for FM-PCD Manip module */
-+/***********************************************************************/
-+t_Error FmPcdManipUpdate(t_Handle h_FmPcd, t_Handle h_PcdParams, t_Handle h_FmPort, t_Handle h_Manip, t_Handle h_Ad, bool validate, int level, t_Handle h_FmTree, bool modify);
-+
-+/***********************************************************************/
-+/* Common API for FM-Port module */
-+/***********************************************************************/
-+#if (DPAA_VERSION >= 11)
-+typedef enum e_FmPortGprFuncType
-+{
-+ e_FM_PORT_GPR_EMPTY = 0,
-+ e_FM_PORT_GPR_MURAM_PAGE
-+} e_FmPortGprFuncType;
-+
-+t_Error FmPortSetGprFunc(t_Handle h_FmPort, e_FmPortGprFuncType gprFunc, void **p_Value);
-+#endif /* DPAA_VERSION >= 11) */
-+t_Error FmGetSetParams(t_Handle h_Fm, t_FmGetSetParams *p_FmGetSetParams);
-+t_Error FmPortGetSetCcParams(t_Handle h_FmPort, t_FmPortGetSetCcParams *p_FmPortGetSetCcParams);
-+uint8_t FmPortGetNetEnvId(t_Handle h_FmPort);
-+uint8_t FmPortGetHardwarePortId(t_Handle h_FmPort);
-+uint32_t FmPortGetPcdEngines(t_Handle h_FmPort);
-+void FmPortPcdKgSwUnbindClsPlanGrp (t_Handle h_FmPort);
-+
-+
-+#if (DPAA_VERSION >= 11)
-+t_Error FmPcdFrmReplicUpdate(t_Handle h_FmPcd, t_Handle h_FmPort, t_Handle h_FrmReplic);
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+/**************************************************************************//**
-+ @Function FmRegisterIntr
-+
-+ @Description Used to register an inter-module event handler to be processed by FM
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+ @Param[in] mod The module that causes the event
-+ @Param[in] modId Module id - if more than 1 instansiation of this
-+ mode exists,0 otherwise.
-+ @Param[in] intrType Interrupt type (error/normal) selection.
-+ @Param[in] f_Isr The interrupt service routine.
-+ @Param[in] h_Arg Argument to be passed to f_Isr.
-+
-+ @Return None.
-+*//***************************************************************************/
-+void FmRegisterIntr(t_Handle h_Fm,
-+ e_FmEventModules mod,
-+ uint8_t modId,
-+ e_FmIntrType intrType,
-+ void (*f_Isr) (t_Handle h_Arg),
-+ t_Handle h_Arg);
-+
-+/**************************************************************************//**
-+ @Function FmUnregisterIntr
-+
-+ @Description Used to un-register an inter-module event handler that was processed by FM
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+ @Param[in] mod The module that causes the event
-+ @Param[in] modId Module id - if more than 1 instansiation of this
-+ mode exists,0 otherwise.
-+ @Param[in] intrType Interrupt type (error/normal) selection.
-+
-+ @Return None.
-+*//***************************************************************************/
-+void FmUnregisterIntr(t_Handle h_Fm,
-+ e_FmEventModules mod,
-+ uint8_t modId,
-+ e_FmIntrType intrType);
-+
-+/**************************************************************************//**
-+ @Function FmRegisterFmCtlIntr
-+
-+ @Description Used to register to one of the fmCtl events in the FM module
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+ @Param[in] eventRegId FmCtl event id (0-7).
-+ @Param[in] f_Isr The interrupt service routine.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_Init().
-+*//***************************************************************************/
-+void FmRegisterFmCtlIntr(t_Handle h_Fm, uint8_t eventRegId, void (*f_Isr) (t_Handle h_Fm, uint32_t event));
-+
-+
-+/**************************************************************************//**
-+ @Description enum for defining MAC types
-+*//***************************************************************************/
-+typedef enum e_FmMacType {
-+ e_FM_MAC_10G = 0, /**< 10G MAC */
-+ e_FM_MAC_1G /**< 1G MAC */
-+} e_FmMacType;
-+
-+/**************************************************************************//**
-+ @Description Structure for port-FM communication during FM_PORT_Init.
-+ Fields commented 'IN' are passed by the port module to be used
-+ by the FM module.
-+ Fields commented 'OUT' will be filled by FM before returning to port.
-+ Some fields are optional (depending on configuration) and
-+ will be analized by the port and FM modules accordingly.
-+*//***************************************************************************/
-+typedef struct t_FmInterModulePortInitParams {
-+ uint8_t hardwarePortId; /**< IN. port Id */
-+ e_FmPortType portType; /**< IN. Port type */
-+ bool independentMode; /**< IN. TRUE if FM Port operates in independent mode */
-+ uint16_t liodnOffset; /**< IN. Port's requested resource */
-+ uint8_t numOfTasks; /**< IN. Port's requested resource */
-+ uint8_t numOfExtraTasks; /**< IN. Port's requested resource */
-+ uint8_t numOfOpenDmas; /**< IN. Port's requested resource */
-+ uint8_t numOfExtraOpenDmas; /**< IN. Port's requested resource */
-+ uint32_t sizeOfFifo; /**< IN. Port's requested resource */
-+ uint32_t extraSizeOfFifo; /**< IN. Port's requested resource */
-+ uint8_t deqPipelineDepth; /**< IN. Port's requested resource */
-+ uint16_t maxFrameLength; /**< IN. Port's max frame length. */
-+ uint16_t liodnBase; /**< IN. Irrelevant for P4080 rev 1.
-+ LIODN base for this port, to be
-+ used together with LIODN offset. */
-+ t_FmPhysAddr fmMuramPhysBaseAddr;/**< OUT. FM-MURAM physical address*/
-+} t_FmInterModulePortInitParams;
-+
-+/**************************************************************************//**
-+ @Description Structure for port-FM communication during FM_PORT_Free.
-+*//***************************************************************************/
-+typedef struct t_FmInterModulePortFreeParams {
-+ uint8_t hardwarePortId; /**< IN. port Id */
-+ e_FmPortType portType; /**< IN. Port type */
-+ uint8_t deqPipelineDepth; /**< IN. Port's requested resource */
-+} t_FmInterModulePortFreeParams;
-+
-+/**************************************************************************//**
-+ @Function FmGetPcdPrsBaseAddr
-+
-+ @Description Get the base address of the Parser from the FM module
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+
-+ @Return Base address.
-+*//***************************************************************************/
-+uintptr_t FmGetPcdPrsBaseAddr(t_Handle h_Fm);
-+
-+/**************************************************************************//**
-+ @Function FmGetPcdKgBaseAddr
-+
-+ @Description Get the base address of the Keygen from the FM module
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+
-+ @Return Base address.
-+*//***************************************************************************/
-+uintptr_t FmGetPcdKgBaseAddr(t_Handle h_Fm);
-+
-+/**************************************************************************//**
-+ @Function FmGetPcdPlcrBaseAddr
-+
-+ @Description Get the base address of the Policer from the FM module
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+
-+ @Return Base address.
-+*//***************************************************************************/
-+uintptr_t FmGetPcdPlcrBaseAddr(t_Handle h_Fm);
-+
-+/**************************************************************************//**
-+ @Function FmGetMuramHandle
-+
-+ @Description Get the handle of the MURAM from the FM module
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+
-+ @Return MURAM module handle.
-+*//***************************************************************************/
-+t_Handle FmGetMuramHandle(t_Handle h_Fm);
-+
-+/**************************************************************************//**
-+ @Function FmGetPhysicalMuramBase
-+
-+ @Description Get the physical base address of the MURAM from the FM module
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+ @Param[in] fmPhysAddr Physical MURAM base
-+
-+ @Return Physical base address.
-+*//***************************************************************************/
-+void FmGetPhysicalMuramBase(t_Handle h_Fm, t_FmPhysAddr *fmPhysAddr);
-+
-+/**************************************************************************//**
-+ @Function FmGetTimeStampScale
-+
-+ @Description Used internally by other modules in order to get the timeStamp
-+ period as requested by the application.
-+
-+ This function returns bit number that is incremented every 1 usec.
-+ To calculate timestamp period in nsec, use
-+ 1000 / (1 << FmGetTimeStampScale()).
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+
-+ @Return Bit that counts 1 usec.
-+
-+ @Cautions Allowed only following FM_Init().
-+*//***************************************************************************/
-+uint32_t FmGetTimeStampScale(t_Handle h_Fm);
-+
-+/**************************************************************************//**
-+ @Function FmResumeStalledPort
-+
-+ @Description Used internally by FM port to release a stalled port.
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+ @Param[in] hardwarePortId HW port id.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_Init().
-+*//***************************************************************************/
-+t_Error FmResumeStalledPort(t_Handle h_Fm, uint8_t hardwarePortId);
-+
-+/**************************************************************************//**
-+ @Function FmIsPortStalled
-+
-+ @Description Used internally by FM port to read the port's status.
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+ @Param[in] hardwarePortId HW port id.
-+ @Param[in] p_IsStalled A pointer to the boolean port stalled state
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_Init().
-+*//***************************************************************************/
-+t_Error FmIsPortStalled(t_Handle h_Fm, uint8_t hardwarePortId, bool *p_IsStalled);
-+
-+/**************************************************************************//**
-+ @Function FmResetMac
-+
-+ @Description Used by MAC driver to reset the MAC registers
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+ @Param[in] type MAC type.
-+ @Param[in] macId MAC id - according to type.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_Init().
-+*//***************************************************************************/
-+t_Error FmResetMac(t_Handle h_Fm, e_FmMacType type, uint8_t macId);
-+
-+/**************************************************************************//**
-+ @Function FmGetClockFreq
-+
-+ @Description Used by MAC driver to get the FM clock frequency
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+
-+ @Return clock-freq on success; 0 otherwise.
-+
-+ @Cautions Allowed only following FM_Init().
-+*//***************************************************************************/
-+uint16_t FmGetClockFreq(t_Handle h_Fm);
-+
-+/**************************************************************************//**
-+ @Function FmGetMacClockFreq
-+
-+ @Description Used by MAC driver to get the MAC clock frequency
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+
-+ @Return clock-freq on success; 0 otherwise.
-+
-+ @Cautions Allowed only following FM_Init().
-+*//***************************************************************************/
-+uint16_t FmGetMacClockFreq(t_Handle h_Fm);
-+
-+/**************************************************************************//**
-+ @Function FmGetId
-+
-+ @Description Used by PCD driver to read rhe FM id
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_Init().
-+*//***************************************************************************/
-+uint8_t FmGetId(t_Handle h_Fm);
-+
-+/**************************************************************************//**
-+ @Function FmReset
-+
-+ @Description Used to reset the FM
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+
-+ @Return E_OK on success; Error code otherwise.
-+*//***************************************************************************/
-+t_Error FmReset(t_Handle h_Fm);
-+
-+/**************************************************************************//**
-+ @Function FmGetSetPortParams
-+
-+ @Description Used by FM-PORT driver to pass and receive parameters between
-+ PORT and FM modules.
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+ @Param[in,out] p_PortParams A structure of FM Port parameters.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_Init().
-+*//***************************************************************************/
-+t_Error FmGetSetPortParams(t_Handle h_Fm,t_FmInterModulePortInitParams *p_PortParams);
-+
-+/**************************************************************************//**
-+ @Function FmFreePortParams
-+
-+ @Description Used by FM-PORT driver to free port's resources within the FM.
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+ @Param[in,out] p_PortParams A structure of FM Port parameters.
-+
-+ @Return None.
-+
-+ @Cautions Allowed only following FM_Init().
-+*//***************************************************************************/
-+void FmFreePortParams(t_Handle h_Fm,t_FmInterModulePortFreeParams *p_PortParams);
-+
-+/**************************************************************************//**
-+ @Function FmSetNumOfRiscsPerPort
-+
-+ @Description Used by FM-PORT driver to pass parameter between
-+ PORT and FM modules for working with number of RISC..
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+ @Param[in] hardwarePortId hardware port Id.
-+ @Param[in] numOfFmanCtrls number of Fman Controllers.
-+ @Param[in] orFmanCtrl Fman Controller for order restoration.
-+
-+ @Return None.
-+
-+ @Cautions Allowed only following FM_Init().
-+*//***************************************************************************/
-+t_Error FmSetNumOfRiscsPerPort(t_Handle h_Fm, uint8_t hardwarePortId, uint8_t numOfFmanCtrls, t_FmFmanCtrl orFmanCtrl);
-+
-+#if (defined(DEBUG_ERRORS) && (DEBUG_ERRORS > 0))
-+/**************************************************************************//*
-+ @Function FmDumpPortRegs
-+
-+ @Description Dumps FM port registers which are part of FM common registers
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+ @Param[in] hardwarePortId HW port id.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only FM_Init().
-+*//***************************************************************************/
-+t_Error FmDumpPortRegs(t_Handle h_Fm,uint8_t hardwarePortId);
-+#endif /* (defined(DEBUG_ERRORS) && ... */
-+
-+void FmRegisterPcd(t_Handle h_Fm, t_Handle h_FmPcd);
-+void FmUnregisterPcd(t_Handle h_Fm);
-+t_Handle FmGetPcdHandle(t_Handle h_Fm);
-+t_Error FmEnableRamsEcc(t_Handle h_Fm);
-+t_Error FmDisableRamsEcc(t_Handle h_Fm);
-+void FmGetRevision(t_Handle h_Fm, t_FmRevisionInfo *p_FmRevisionInfo);
-+t_Error FmAllocFmanCtrlEventReg(t_Handle h_Fm, uint8_t *p_EventId);
-+void FmFreeFmanCtrlEventReg(t_Handle h_Fm, uint8_t eventId);
-+void FmSetFmanCtrlIntr(t_Handle h_Fm, uint8_t eventRegId, uint32_t enableEvents);
-+uint32_t FmGetFmanCtrlIntr(t_Handle h_Fm, uint8_t eventRegId);
-+void FmRegisterFmanCtrlIntr(t_Handle h_Fm, uint8_t eventRegId, void (*f_Isr) (t_Handle h_Fm, uint32_t event), t_Handle h_Arg);
-+void FmUnregisterFmanCtrlIntr(t_Handle h_Fm, uint8_t eventRegId);
-+t_Error FmSetMacMaxFrame(t_Handle h_Fm, e_FmMacType type, uint8_t macId, uint16_t mtu);
-+bool FmIsMaster(t_Handle h_Fm);
-+uint8_t FmGetGuestId(t_Handle h_Fm);
-+uint16_t FmGetTnumAgingPeriod(t_Handle h_Fm);
-+t_Error FmSetPortPreFetchConfiguration(t_Handle h_Fm, uint8_t portNum, bool preFetchConfigured);
-+t_Error FmGetPortPreFetchConfiguration(t_Handle h_Fm, uint8_t portNum, bool *p_PortConfigured, bool *p_PreFetchConfigured);
-+
-+
-+#ifdef FM_TX_ECC_FRMS_ERRATA_10GMAC_A004
-+t_Error Fm10GTxEccWorkaround(t_Handle h_Fm, uint8_t macId);
-+#endif /* FM_TX_ECC_FRMS_ERRATA_10GMAC_A004 */
-+
-+void FmMuramClear(t_Handle h_FmMuram);
-+t_Error FmSetNumOfOpenDmas(t_Handle h_Fm,
-+ uint8_t hardwarePortId,
-+ uint8_t *p_NumOfOpenDmas,
-+ uint8_t *p_NumOfExtraOpenDmas,
-+ bool initialConfig);
-+t_Error FmSetNumOfTasks(t_Handle h_Fm,
-+ uint8_t hardwarePortId,
-+ uint8_t *p_NumOfTasks,
-+ uint8_t *p_NumOfExtraTasks,
-+ bool initialConfig);
-+t_Error FmSetSizeOfFifo(t_Handle h_Fm,
-+ uint8_t hardwarePortId,
-+ uint32_t *p_SizeOfFifo,
-+ uint32_t *p_ExtraSizeOfFifo,
-+ bool initialConfig);
-+
-+t_Error FmSetCongestionGroupPFCpriority(t_Handle h_Fm,
-+ uint32_t congestionGroupId,
-+ uint8_t priorityBitMap);
-+
-+#if (DPAA_VERSION >= 11)
-+t_Error FmVSPAllocForPort(t_Handle h_Fm,
-+ e_FmPortType portType,
-+ uint8_t portId,
-+ uint8_t numOfStorageProfiles);
-+
-+t_Error FmVSPFreeForPort(t_Handle h_Fm,
-+ e_FmPortType portType,
-+ uint8_t portId);
-+
-+t_Error FmVSPGetAbsoluteProfileId(t_Handle h_Fm,
-+ e_FmPortType portType,
-+ uint8_t portId,
-+ uint16_t relativeProfile,
-+ uint16_t *p_AbsoluteId);
-+t_Error FmVSPCheckRelativeProfile(t_Handle h_Fm,
-+ e_FmPortType portType,
-+ uint8_t portId,
-+ uint16_t relativeProfile);
-+
-+uintptr_t FmGetVSPBaseAddr(t_Handle h_Fm);
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+
-+#endif /* __FM_COMMON_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/inc/fm_hc.h
-@@ -0,0 +1,93 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+#ifndef __FM_HC_H
-+#define __FM_HC_H
-+
-+#include "std_ext.h"
-+#include "error_ext.h"
-+#include "fsl_fman_kg.h"
-+
-+#define __ERR_MODULE__ MODULE_FM_PCD
-+
-+
-+typedef struct t_FmHcParams {
-+ t_Handle h_Fm;
-+ t_Handle h_FmPcd;
-+ t_FmPcdHcParams params;
-+} t_FmHcParams;
-+
-+
-+t_Handle FmHcConfigAndInit(t_FmHcParams *p_FmHcParams);
-+void FmHcFree(t_Handle h_FmHc);
-+t_Error FmHcSetFramesDataMemory(t_Handle h_FmHc,
-+ uint8_t memId);
-+t_Error FmHcDumpRegs(t_Handle h_FmHc);
-+
-+void FmHcTxConf(t_Handle h_FmHc, t_DpaaFD *p_Fd);
-+
-+t_Error FmHcPcdKgSetScheme(t_Handle h_FmHc,
-+ t_Handle h_Scheme,
-+ struct fman_kg_scheme_regs *p_SchemeRegs,
-+ bool updateCounter);
-+t_Error FmHcPcdKgDeleteScheme(t_Handle h_FmHc, t_Handle h_Scheme);
-+t_Error FmHcPcdCcCapwapTimeoutReassm(t_Handle h_FmHc, t_FmPcdCcCapwapReassmTimeoutParams *p_CcCapwapReassmTimeoutParams );
-+t_Error FmHcPcdCcIpFragScratchPollCmd(t_Handle h_FmHc, bool fill, t_FmPcdCcFragScratchPoolCmdParams *p_FmPcdCcFragScratchPoolCmdParams);
-+t_Error FmHcPcdCcTimeoutReassm(t_Handle h_FmHc, t_FmPcdCcReassmTimeoutParams *p_CcReassmTimeoutParams, uint8_t *p_Result);
-+t_Error FmHcPcdKgSetClsPlan(t_Handle h_FmHc, t_FmPcdKgInterModuleClsPlanSet *p_Set);
-+t_Error FmHcPcdKgDeleteClsPlan(t_Handle h_FmHc, uint8_t clsPlanGrpId);
-+
-+t_Error FmHcPcdKgSetSchemeCounter(t_Handle h_FmHc, t_Handle h_Scheme, uint32_t value);
-+uint32_t FmHcPcdKgGetSchemeCounter(t_Handle h_FmHc, t_Handle h_Scheme);
-+
-+t_Error FmHcPcdCcDoDynamicChange(t_Handle h_FmHc, uint32_t oldAdAddrOffset, uint32_t newAdAddrOffset);
-+
-+t_Error FmHcPcdPlcrSetProfile(t_Handle h_FmHc, t_Handle h_Profile, t_FmPcdPlcrProfileRegs *p_PlcrRegs);
-+t_Error FmHcPcdPlcrDeleteProfile(t_Handle h_FmHc, t_Handle h_Profile);
-+
-+t_Error FmHcPcdPlcrSetProfileCounter(t_Handle h_FmHc, t_Handle h_Profile, e_FmPcdPlcrProfileCounters counter, uint32_t value);
-+uint32_t FmHcPcdPlcrGetProfileCounter(t_Handle h_FmHc, t_Handle h_Profile, e_FmPcdPlcrProfileCounters counter);
-+
-+t_Error FmHcKgWriteSp(t_Handle h_FmHc, uint8_t hardwarePortId, uint32_t spReg, bool add);
-+t_Error FmHcKgWriteCpp(t_Handle h_FmHc, uint8_t hardwarePortId, uint32_t cppReg);
-+
-+t_Error FmHcPcdKgCcGetSetParams(t_Handle h_FmHc, t_Handle h_Scheme, uint32_t requiredAction, uint32_t value);
-+t_Error FmHcPcdPlcrCcGetSetParams(t_Handle h_FmHc,uint16_t absoluteProfileId, uint32_t requiredAction);
-+
-+t_Error FmHcPcdSync(t_Handle h_FmHc);
-+t_Handle FmHcGetPort(t_Handle h_FmHc);
-+
-+
-+
-+
-+#endif /* __FM_HC_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/inc/fm_sp_common.h
-@@ -0,0 +1,117 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/******************************************************************************
-+ @File fm_sp_common.h
-+
-+ @Description FM SP ...
-+*//***************************************************************************/
-+#ifndef __FM_SP_COMMON_H
-+#define __FM_SP_COMMON_H
-+
-+#include "std_ext.h"
-+#include "error_ext.h"
-+#include "list_ext.h"
-+
-+#include "fm_ext.h"
-+#include "fm_pcd_ext.h"
-+#include "fsl_fman.h"
-+
-+/**************************************************************************//**
-+ @Description defaults
-+*//***************************************************************************/
-+#define DEFAULT_FM_SP_bufferPrefixContent_privDataSize 0
-+#define DEFAULT_FM_SP_bufferPrefixContent_passPrsResult FALSE
-+#define DEFAULT_FM_SP_bufferPrefixContent_passTimeStamp FALSE
-+#define DEFAULT_FM_SP_bufferPrefixContent_allOtherPCDInfo FALSE
-+#define DEFAULT_FM_SP_bufferPrefixContent_dataAlign 64
-+
-+/**************************************************************************//**
-+ @Description structure for defining internal context copying
-+*//***************************************************************************/
-+typedef struct
-+{
-+ uint16_t extBufOffset; /**< Offset in External buffer to which internal
-+ context is copied to (Rx) or taken from (Tx, Op). */
-+ uint8_t intContextOffset; /**< Offset within internal context to copy from
-+ (Rx) or to copy to (Tx, Op). */
-+ uint16_t size; /**< Internal offset size to be copied */
-+} t_FmSpIntContextDataCopy;
-+
-+/**************************************************************************//**
-+ @Description struct for defining external buffer margins
-+*//***************************************************************************/
-+typedef struct {
-+ uint16_t startMargins; /**< Number of bytes to be left at the beginning
-+ of the external buffer (must be divisible by 16) */
-+ uint16_t endMargins; /**< number of bytes to be left at the end
-+ of the external buffer(must be divisible by 16) */
-+} t_FmSpBufMargins;
-+
-+typedef struct {
-+ uint32_t dataOffset;
-+ uint32_t prsResultOffset;
-+ uint32_t timeStampOffset;
-+ uint32_t hashResultOffset;
-+ uint32_t pcdInfoOffset;
-+ uint32_t manipOffset;
-+} t_FmSpBufferOffsets;
-+
-+
-+t_Error FmSpBuildBufferStructure(t_FmSpIntContextDataCopy *p_FmPortIntContextDataCopy,
-+ t_FmBufferPrefixContent *p_BufferPrefixContent,
-+ t_FmSpBufMargins *p_FmPortBufMargins,
-+ t_FmSpBufferOffsets *p_FmPortBufferOffsets,
-+ uint8_t *internalBufferOffset);
-+
-+t_Error FmSpCheckIntContextParams(t_FmSpIntContextDataCopy *p_FmSpIntContextDataCopy);
-+t_Error FmSpCheckBufPoolsParams(t_FmExtPools *p_FmExtPools,
-+ t_FmBackupBmPools *p_FmBackupBmPools,
-+ t_FmBufPoolDepletion *p_FmBufPoolDepletion);
-+t_Error FmSpCheckBufMargins(t_FmSpBufMargins *p_FmSpBufMargins);
-+void FmSpSetBufPoolsInAscOrderOfBufSizes(t_FmExtPools *p_FmExtPools, uint8_t *orderedArray, uint16_t *sizesArray);
-+
-+t_Error FmPcdSpAllocProfiles(t_Handle h_FmPcd,
-+ uint8_t hardwarePortId,
-+ uint16_t numOfStorageProfiles,
-+ uint16_t *base,
-+ uint8_t *log2Num);
-+t_Error FmPcdSpGetAbsoluteProfileId(t_Handle h_FmPcd,
-+ t_Handle h_FmPort,
-+ uint16_t relativeProfile,
-+ uint16_t *p_AbsoluteId);
-+void SpInvalidateProfileSw(t_Handle h_FmPcd, uint16_t absoluteProfileId);
-+void SpValidateProfileSw(t_Handle h_FmPcd, uint16_t absoluteProfileId);
-+
-+
-+#endif /* __FM_SP_COMMON_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/etc/Makefile
-@@ -0,0 +1,12 @@
-+#
-+# Makefile for the Freescale Ethernet controllers
-+#
-+ccflags-y += -DVERSION=\"\"
-+#
-+#Include netcomm SW specific definitions
-+
-+include $(srctree)/drivers/net/ethernet/freescale/sdk_fman/ncsw_config.mk
-+
-+obj-y += fsl-ncsw-etc.o
-+
-+fsl-ncsw-etc-objs := mm.o memcpy.o sprint.o list.o error.o
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/etc/error.c
-@@ -0,0 +1,95 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/*
-+
-+ @File error.c
-+
-+ @Description General errors and events reporting utilities.
-+*//***************************************************************************/
-+#if (defined(DEBUG_ERRORS) && (DEBUG_ERRORS > 0))
-+#include "error_ext.h"
-+
-+
-+const char *dbgLevelStrings[] =
-+{
-+ "CRITICAL"
-+ ,"MAJOR"
-+ ,"MINOR"
-+ ,"WARNING"
-+ ,"INFO"
-+ ,"TRACE"
-+};
-+
-+
-+char * ErrTypeStrings (e_ErrorType err)
-+{
-+ switch (err)
-+ {
-+ case (E_OK): return "OK";
-+ case (E_WRITE_FAILED): return "Write Access Failed";
-+ case (E_NO_DEVICE): return "No Device";
-+ case (E_NOT_AVAILABLE): return "Resource Is Unavailable";
-+ case (E_NO_MEMORY): return "Memory Allocation Failed";
-+ case (E_INVALID_ADDRESS): return "Invalid Address";
-+ case (E_BUSY): return "Resource Is Busy";
-+ case (E_ALREADY_EXISTS): return "Resource Already Exists";
-+ case (E_INVALID_OPERATION): return "Invalid Operation";
-+ case (E_INVALID_VALUE): return "Invalid Value";
-+ case (E_NOT_IN_RANGE): return "Value Out Of Range";
-+ case (E_NOT_SUPPORTED): return "Unsupported Operation";
-+ case (E_INVALID_STATE): return "Invalid State";
-+ case (E_INVALID_HANDLE): return "Invalid Handle";
-+ case (E_INVALID_ID): return "Invalid ID";
-+ case (E_NULL_POINTER): return "Unexpected NULL Pointer";
-+ case (E_INVALID_SELECTION): return "Invalid Selection";
-+ case (E_INVALID_COMM_MODE): return "Invalid Communication Mode";
-+ case (E_INVALID_MEMORY_TYPE): return "Invalid Memory Type";
-+ case (E_INVALID_CLOCK): return "Invalid Clock";
-+ case (E_CONFLICT): return "Conflict In Settings";
-+ case (E_NOT_ALIGNED): return "Incorrect Alignment";
-+ case (E_NOT_FOUND): return "Resource Not Found";
-+ case (E_FULL): return "Resource Is Full";
-+ case (E_EMPTY): return "Resource Is Empty";
-+ case (E_ALREADY_FREE): return "Resource Already Free";
-+ case (E_READ_FAILED): return "Read Access Failed";
-+ case (E_INVALID_FRAME): return "Invalid Frame";
-+ case (E_SEND_FAILED): return "Send Operation Failed";
-+ case (E_RECEIVE_FAILED): return "Receive Operation Failed";
-+ case (E_TIMEOUT): return "Operation Timed Out";
-+ default:
-+ break;
-+ }
-+ return NULL;
-+}
-+#endif /* (defined(DEBUG_ERRORS) && (DEBUG_ERRORS > 0)) */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/etc/list.c
-@@ -0,0 +1,71 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/**************************************************************************//**
-+
-+ @File list.c
-+
-+ @Description Implementation of list.
-+*//***************************************************************************/
-+#include "std_ext.h"
-+#include "list_ext.h"
-+
-+
-+void LIST_Append(t_List *p_NewList, t_List *p_Head)
-+{
-+ t_List *p_First = LIST_FIRST(p_NewList);
-+
-+ if (p_First != p_NewList)
-+ {
-+ t_List *p_Last = LIST_LAST(p_NewList);
-+ t_List *p_Cur = LIST_NEXT(p_Head);
-+
-+ LIST_PREV(p_First) = p_Head;
-+ LIST_FIRST(p_Head) = p_First;
-+ LIST_NEXT(p_Last) = p_Cur;
-+ LIST_LAST(p_Cur) = p_Last;
-+ }
-+}
-+
-+
-+int LIST_NumOfObjs(t_List *p_List)
-+{
-+ t_List *p_Tmp;
-+ int numOfObjs = 0;
-+
-+ if (!LIST_IsEmpty(p_List))
-+ LIST_FOR_EACH(p_Tmp, p_List)
-+ numOfObjs++;
-+
-+ return numOfObjs;
-+}
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/etc/memcpy.c
-@@ -0,0 +1,620 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+
-+#include "std_ext.h"
-+#include "xx_ext.h"
-+#include "memcpy_ext.h"
-+
-+void * MemCpy8(void* pDst, void* pSrc, uint32_t size)
-+{
-+ int i;
-+
-+ for(i = 0; i < size; ++i)
-+ *(((uint8_t*)(pDst)) + i) = *(((uint8_t*)(pSrc)) + i);
-+
-+ return pDst;
-+}
-+
-+void * MemSet8(void* pDst, int c, uint32_t size)
-+{
-+ int i;
-+
-+ for(i = 0; i < size; ++i)
-+ *(((uint8_t*)(pDst)) + i) = (uint8_t)(c);
-+
-+ return pDst;
-+}
-+
-+void * MemCpy32(void* pDst,void* pSrc, uint32_t size)
-+{
-+ uint32_t leftAlign;
-+ uint32_t rightAlign;
-+ uint32_t lastWord;
-+ uint32_t currWord;
-+ uint32_t *p_Src32;
-+ uint32_t *p_Dst32;
-+ uint8_t *p_Src8;
-+ uint8_t *p_Dst8;
-+
-+ p_Src8 = (uint8_t*)(pSrc);
-+ p_Dst8 = (uint8_t*)(pDst);
-+ /* first copy byte by byte till the source first alignment
-+ * this step is necessary to ensure we do not even try to access
-+ * data which is before the source buffer, hence it is not ours.
-+ */
-+ while((PTR_TO_UINT(p_Src8) & 3) && size) /* (pSrc mod 4) > 0 and size > 0 */
-+ {
-+ *p_Dst8++ = *p_Src8++;
-+ size--;
-+ }
-+
-+ /* align destination (possibly disaligning source)*/
-+ while((PTR_TO_UINT(p_Dst8) & 3) && size) /* (pDst mod 4) > 0 and size > 0 */
-+ {
-+ *p_Dst8++ = *p_Src8++;
-+ size--;
-+ }
-+
-+ /* dest is aligned and source is not necessarily aligned */
-+ leftAlign = (uint32_t)((PTR_TO_UINT(p_Src8) & 3) << 3); /* leftAlign = (pSrc mod 4)*8 */
-+ rightAlign = 32 - leftAlign;
-+
-+
-+ if (leftAlign == 0)
-+ {
-+ /* source is also aligned */
-+ p_Src32 = (uint32_t*)(p_Src8);
-+ p_Dst32 = (uint32_t*)(p_Dst8);
-+ while (size >> 2) /* size >= 4 */
-+ {
-+ *p_Dst32++ = *p_Src32++;
-+ size -= 4;
-+ }
-+ p_Src8 = (uint8_t*)(p_Src32);
-+ p_Dst8 = (uint8_t*)(p_Dst32);
-+ }
-+ else
-+ {
-+ /* source is not aligned (destination is aligned)*/
-+ p_Src32 = (uint32_t*)(p_Src8 - (leftAlign >> 3));
-+ p_Dst32 = (uint32_t*)(p_Dst8);
-+ lastWord = *p_Src32++;
-+ while(size >> 3) /* size >= 8 */
-+ {
-+ currWord = *p_Src32;
-+ *p_Dst32 = (lastWord << leftAlign) | (currWord >> rightAlign);
-+ lastWord = currWord;
-+ p_Src32++;
-+ p_Dst32++;
-+ size -= 4;
-+ }
-+ p_Dst8 = (uint8_t*)(p_Dst32);
-+ p_Src8 = (uint8_t*)(p_Src32) - 4 + (leftAlign >> 3);
-+ }
-+
-+ /* complete the left overs */
-+ while (size--)
-+ *p_Dst8++ = *p_Src8++;
-+
-+ return pDst;
-+}
-+
-+void * IO2IOCpy32(void* pDst,void* pSrc, uint32_t size)
-+{
-+ uint32_t leftAlign;
-+ uint32_t rightAlign;
-+ uint32_t lastWord;
-+ uint32_t currWord;
-+ uint32_t *p_Src32;
-+ uint32_t *p_Dst32;
-+ uint8_t *p_Src8;
-+ uint8_t *p_Dst8;
-+
-+ p_Src8 = (uint8_t*)(pSrc);
-+ p_Dst8 = (uint8_t*)(pDst);
-+ /* first copy byte by byte till the source first alignment
-+ * this step is necessary to ensure we do not even try to access
-+ * data which is before the source buffer, hence it is not ours.
-+ */
-+ while((PTR_TO_UINT(p_Src8) & 3) && size) /* (pSrc mod 4) > 0 and size > 0 */
-+ {
-+ WRITE_UINT8(*p_Dst8, GET_UINT8(*p_Src8));
-+ p_Dst8++;p_Src8++;
-+ size--;
-+ }
-+
-+ /* align destination (possibly disaligning source)*/
-+ while((PTR_TO_UINT(p_Dst8) & 3) && size) /* (pDst mod 4) > 0 and size > 0 */
-+ {
-+ WRITE_UINT8(*p_Dst8, GET_UINT8(*p_Src8));
-+ p_Dst8++;p_Src8++;
-+ size--;
-+ }
-+
-+ /* dest is aligned and source is not necessarily aligned */
-+ leftAlign = (uint32_t)((PTR_TO_UINT(p_Src8) & 3) << 3); /* leftAlign = (pSrc mod 4)*8 */
-+ rightAlign = 32 - leftAlign;
-+
-+ if (leftAlign == 0)
-+ {
-+ /* source is also aligned */
-+ p_Src32 = (uint32_t*)(p_Src8);
-+ p_Dst32 = (uint32_t*)(p_Dst8);
-+ while (size >> 2) /* size >= 4 */
-+ {
-+ WRITE_UINT32(*p_Dst32, GET_UINT32(*p_Src32));
-+ p_Dst32++;p_Src32++;
-+ size -= 4;
-+ }
-+ p_Src8 = (uint8_t*)(p_Src32);
-+ p_Dst8 = (uint8_t*)(p_Dst32);
-+ }
-+ else
-+ {
-+ /* source is not aligned (destination is aligned)*/
-+ p_Src32 = (uint32_t*)(p_Src8 - (leftAlign >> 3));
-+ p_Dst32 = (uint32_t*)(p_Dst8);
-+ lastWord = GET_UINT32(*p_Src32);
-+ p_Src32++;
-+ while(size >> 3) /* size >= 8 */
-+ {
-+ currWord = GET_UINT32(*p_Src32);
-+ WRITE_UINT32(*p_Dst32, (lastWord << leftAlign) | (currWord >> rightAlign));
-+ lastWord = currWord;
-+ p_Src32++;p_Dst32++;
-+ size -= 4;
-+ }
-+ p_Dst8 = (uint8_t*)(p_Dst32);
-+ p_Src8 = (uint8_t*)(p_Src32) - 4 + (leftAlign >> 3);
-+ }
-+
-+ /* complete the left overs */
-+ while (size--)
-+ {
-+ WRITE_UINT8(*p_Dst8, GET_UINT8(*p_Src8));
-+ p_Dst8++;p_Src8++;
-+ }
-+
-+ return pDst;
-+}
-+
-+void * Mem2IOCpy32(void* pDst,void* pSrc, uint32_t size)
-+{
-+ uint32_t leftAlign;
-+ uint32_t rightAlign;
-+ uint32_t lastWord;
-+ uint32_t currWord;
-+ uint32_t *p_Src32;
-+ uint32_t *p_Dst32;
-+ uint8_t *p_Src8;
-+ uint8_t *p_Dst8;
-+
-+ p_Src8 = (uint8_t*)(pSrc);
-+ p_Dst8 = (uint8_t*)(pDst);
-+ /* first copy byte by byte till the source first alignment
-+ * this step is necessary to ensure we do not even try to access
-+ * data which is before the source buffer, hence it is not ours.
-+ */
-+ while((PTR_TO_UINT(p_Src8) & 3) && size) /* (pSrc mod 4) > 0 and size > 0 */
-+ {
-+ WRITE_UINT8(*p_Dst8, *p_Src8);
-+ p_Dst8++;p_Src8++;
-+ size--;
-+ }
-+
-+ /* align destination (possibly disaligning source)*/
-+ while((PTR_TO_UINT(p_Dst8) & 3) && size) /* (pDst mod 4) > 0 and size > 0 */
-+ {
-+ WRITE_UINT8(*p_Dst8, *p_Src8);
-+ p_Dst8++;p_Src8++;
-+ size--;
-+ }
-+
-+ /* dest is aligned and source is not necessarily aligned */
-+ leftAlign = (uint32_t)((PTR_TO_UINT(p_Src8) & 3) << 3); /* leftAlign = (pSrc mod 4)*8 */
-+ rightAlign = 32 - leftAlign;
-+
-+ if (leftAlign == 0)
-+ {
-+ /* source is also aligned */
-+ p_Src32 = (uint32_t*)(p_Src8);
-+ p_Dst32 = (uint32_t*)(p_Dst8);
-+ while (size >> 2) /* size >= 4 */
-+ {
-+ WRITE_UINT32(*p_Dst32, *p_Src32);
-+ p_Dst32++;p_Src32++;
-+ size -= 4;
-+ }
-+ p_Src8 = (uint8_t*)(p_Src32);
-+ p_Dst8 = (uint8_t*)(p_Dst32);
-+ }
-+ else
-+ {
-+ /* source is not aligned (destination is aligned)*/
-+ p_Src32 = (uint32_t*)(p_Src8 - (leftAlign >> 3));
-+ p_Dst32 = (uint32_t*)(p_Dst8);
-+ lastWord = *p_Src32++;
-+ while(size >> 3) /* size >= 8 */
-+ {
-+ currWord = *p_Src32;
-+ WRITE_UINT32(*p_Dst32, (lastWord << leftAlign) | (currWord >> rightAlign));
-+ lastWord = currWord;
-+ p_Src32++;p_Dst32++;
-+ size -= 4;
-+ }
-+ p_Dst8 = (uint8_t*)(p_Dst32);
-+ p_Src8 = (uint8_t*)(p_Src32) - 4 + (leftAlign >> 3);
-+ }
-+
-+ /* complete the left overs */
-+ while (size--)
-+ {
-+ WRITE_UINT8(*p_Dst8, *p_Src8);
-+ p_Dst8++;p_Src8++;
-+ }
-+
-+ return pDst;
-+}
-+
-+void * IO2MemCpy32(void* pDst,void* pSrc, uint32_t size)
-+{
-+ uint32_t leftAlign;
-+ uint32_t rightAlign;
-+ uint32_t lastWord;
-+ uint32_t currWord;
-+ uint32_t *p_Src32;
-+ uint32_t *p_Dst32;
-+ uint8_t *p_Src8;
-+ uint8_t *p_Dst8;
-+
-+ p_Src8 = (uint8_t*)(pSrc);
-+ p_Dst8 = (uint8_t*)(pDst);
-+ /* first copy byte by byte till the source first alignment
-+ * this step is necessary to ensure we do not even try to access
-+ * data which is before the source buffer, hence it is not ours.
-+ */
-+ while((PTR_TO_UINT(p_Src8) & 3) && size) /* (pSrc mod 4) > 0 and size > 0 */
-+ {
-+ *p_Dst8 = GET_UINT8(*p_Src8);
-+ p_Dst8++;p_Src8++;
-+ size--;
-+ }
-+
-+ /* align destination (possibly disaligning source)*/
-+ while((PTR_TO_UINT(p_Dst8) & 3) && size) /* (pDst mod 4) > 0 and size > 0 */
-+ {
-+ *p_Dst8 = GET_UINT8(*p_Src8);
-+ p_Dst8++;p_Src8++;
-+ size--;
-+ }
-+
-+ /* dest is aligned and source is not necessarily aligned */
-+ leftAlign = (uint32_t)((PTR_TO_UINT(p_Src8) & 3) << 3); /* leftAlign = (pSrc mod 4)*8 */
-+ rightAlign = 32 - leftAlign;
-+
-+ if (leftAlign == 0)
-+ {
-+ /* source is also aligned */
-+ p_Src32 = (uint32_t*)(p_Src8);
-+ p_Dst32 = (uint32_t*)(p_Dst8);
-+ while (size >> 2) /* size >= 4 */
-+ {
-+ *p_Dst32 = GET_UINT32(*p_Src32);
-+ p_Dst32++;p_Src32++;
-+ size -= 4;
-+ }
-+ p_Src8 = (uint8_t*)(p_Src32);
-+ p_Dst8 = (uint8_t*)(p_Dst32);
-+ }
-+ else
-+ {
-+ /* source is not aligned (destination is aligned)*/
-+ p_Src32 = (uint32_t*)(p_Src8 - (leftAlign >> 3));
-+ p_Dst32 = (uint32_t*)(p_Dst8);
-+ lastWord = GET_UINT32(*p_Src32);
-+ p_Src32++;
-+ while(size >> 3) /* size >= 8 */
-+ {
-+ currWord = GET_UINT32(*p_Src32);
-+ *p_Dst32 = (lastWord << leftAlign) | (currWord >> rightAlign);
-+ lastWord = currWord;
-+ p_Src32++;p_Dst32++;
-+ size -= 4;
-+ }
-+ p_Dst8 = (uint8_t*)(p_Dst32);
-+ p_Src8 = (uint8_t*)(p_Src32) - 4 + (leftAlign >> 3);
-+ }
-+
-+ /* complete the left overs */
-+ while (size--)
-+ {
-+ *p_Dst8 = GET_UINT8(*p_Src8);
-+ p_Dst8++;p_Src8++;
-+ }
-+
-+ return pDst;
-+}
-+
-+void * MemCpy64(void* pDst,void* pSrc, uint32_t size)
-+{
-+ uint32_t leftAlign;
-+ uint32_t rightAlign;
-+ uint64_t lastWord;
-+ uint64_t currWord;
-+ uint64_t *pSrc64;
-+ uint64_t *pDst64;
-+ uint8_t *p_Src8;
-+ uint8_t *p_Dst8;
-+
-+ p_Src8 = (uint8_t*)(pSrc);
-+ p_Dst8 = (uint8_t*)(pDst);
-+ /* first copy byte by byte till the source first alignment
-+ * this step is necessarily to ensure we do not even try to access
-+ * data which is before the source buffer, hence it is not ours.
-+ */
-+ while((PTR_TO_UINT(p_Src8) & 7) && size) /* (pSrc mod 8) > 0 and size > 0 */
-+ {
-+ *p_Dst8++ = *p_Src8++;
-+ size--;
-+ }
-+
-+ /* align destination (possibly disaligning source)*/
-+ while((PTR_TO_UINT(p_Dst8) & 7) && size) /* (pDst mod 8) > 0 and size > 0 */
-+ {
-+ *p_Dst8++ = *p_Src8++;
-+ size--;
-+ }
-+
-+ /* dest is aligned and source is not necessarily aligned */
-+ leftAlign = (uint32_t)((PTR_TO_UINT(p_Src8) & 7) << 3); /* leftAlign = (pSrc mod 8)*8 */
-+ rightAlign = 64 - leftAlign;
-+
-+
-+ if (leftAlign == 0)
-+ {
-+ /* source is also aligned */
-+ pSrc64 = (uint64_t*)(p_Src8);
-+ pDst64 = (uint64_t*)(p_Dst8);
-+ while (size >> 3) /* size >= 8 */
-+ {
-+ *pDst64++ = *pSrc64++;
-+ size -= 8;
-+ }
-+ p_Src8 = (uint8_t*)(pSrc64);
-+ p_Dst8 = (uint8_t*)(pDst64);
-+ }
-+ else
-+ {
-+ /* source is not aligned (destination is aligned)*/
-+ pSrc64 = (uint64_t*)(p_Src8 - (leftAlign >> 3));
-+ pDst64 = (uint64_t*)(p_Dst8);
-+ lastWord = *pSrc64++;
-+ while(size >> 4) /* size >= 16 */
-+ {
-+ currWord = *pSrc64;
-+ *pDst64 = (lastWord << leftAlign) | (currWord >> rightAlign);
-+ lastWord = currWord;
-+ pSrc64++;
-+ pDst64++;
-+ size -= 8;
-+ }
-+ p_Dst8 = (uint8_t*)(pDst64);
-+ p_Src8 = (uint8_t*)(pSrc64) - 8 + (leftAlign >> 3);
-+ }
-+
-+ /* complete the left overs */
-+ while (size--)
-+ *p_Dst8++ = *p_Src8++;
-+
-+ return pDst;
-+}
-+
-+void * MemSet32(void* pDst, uint8_t val, uint32_t size)
-+{
-+ uint32_t val32;
-+ uint32_t *p_Dst32;
-+ uint8_t *p_Dst8;
-+
-+ p_Dst8 = (uint8_t*)(pDst);
-+
-+ /* generate four 8-bit val's in 32-bit container */
-+ val32 = (uint32_t) val;
-+ val32 |= (val32 << 8);
-+ val32 |= (val32 << 16);
-+
-+ /* align destination to 32 */
-+ while((PTR_TO_UINT(p_Dst8) & 3) && size) /* (pDst mod 4) > 0 and size > 0 */
-+ {
-+ *p_Dst8++ = val;
-+ size--;
-+ }
-+
-+ /* 32-bit chunks */
-+ p_Dst32 = (uint32_t*)(p_Dst8);
-+ while (size >> 2) /* size >= 4 */
-+ {
-+ *p_Dst32++ = val32;
-+ size -= 4;
-+ }
-+
-+ /* complete the leftovers */
-+ p_Dst8 = (uint8_t*)(p_Dst32);
-+ while (size--)
-+ *p_Dst8++ = val;
-+
-+ return pDst;
-+}
-+
-+void * IOMemSet32(void* pDst, uint8_t val, uint32_t size)
-+{
-+ uint32_t val32;
-+ uint32_t *p_Dst32;
-+ uint8_t *p_Dst8;
-+
-+ p_Dst8 = (uint8_t*)(pDst);
-+
-+ /* generate four 8-bit val's in 32-bit container */
-+ val32 = (uint32_t) val;
-+ val32 |= (val32 << 8);
-+ val32 |= (val32 << 16);
-+
-+ /* align destination to 32 */
-+ while((PTR_TO_UINT(p_Dst8) & 3) && size) /* (pDst mod 4) > 0 and size > 0 */
-+ {
-+ WRITE_UINT8(*p_Dst8, val);
-+ p_Dst8++;
-+ size--;
-+ }
-+
-+ /* 32-bit chunks */
-+ p_Dst32 = (uint32_t*)(p_Dst8);
-+ while (size >> 2) /* size >= 4 */
-+ {
-+ WRITE_UINT32(*p_Dst32, val32);
-+ p_Dst32++;
-+ size -= 4;
-+ }
-+
-+ /* complete the leftovers */
-+ p_Dst8 = (uint8_t*)(p_Dst32);
-+ while (size--)
-+ {
-+ WRITE_UINT8(*p_Dst8, val);
-+ p_Dst8++;
-+ }
-+
-+ return pDst;
-+}
-+
-+void * MemSet64(void* pDst, uint8_t val, uint32_t size)
-+{
-+ uint64_t val64;
-+ uint64_t *pDst64;
-+ uint8_t *p_Dst8;
-+
-+ p_Dst8 = (uint8_t*)(pDst);
-+
-+ /* generate four 8-bit val's in 32-bit container */
-+ val64 = (uint64_t) val;
-+ val64 |= (val64 << 8);
-+ val64 |= (val64 << 16);
-+ val64 |= (val64 << 24);
-+ val64 |= (val64 << 32);
-+
-+ /* align destination to 64 */
-+ while((PTR_TO_UINT(p_Dst8) & 7) && size) /* (pDst mod 8) > 0 and size > 0 */
-+ {
-+ *p_Dst8++ = val;
-+ size--;
-+ }
-+
-+ /* 64-bit chunks */
-+ pDst64 = (uint64_t*)(p_Dst8);
-+ while (size >> 4) /* size >= 8 */
-+ {
-+ *pDst64++ = val64;
-+ size -= 8;
-+ }
-+
-+ /* complete the leftovers */
-+ p_Dst8 = (uint8_t*)(pDst64);
-+ while (size--)
-+ *p_Dst8++ = val;
-+
-+ return pDst;
-+}
-+
-+void MemDisp(uint8_t *p, int size)
-+{
-+ uint32_t space = (uint32_t)(PTR_TO_UINT(p) & 0x3);
-+ uint8_t *p_Limit;
-+
-+ if (space)
-+ {
-+ p_Limit = (p - space + 4);
-+
-+ XX_Print("0x%08X: ", (p - space));
-+
-+ while (space--)
-+ {
-+ XX_Print("--");
-+ }
-+ while (size && (p < p_Limit))
-+ {
-+ XX_Print("%02x", *(uint8_t*)p);
-+ size--;
-+ p++;
-+ }
-+
-+ XX_Print(" ");
-+ p_Limit += 12;
-+
-+ while ((size > 3) && (p < p_Limit))
-+ {
-+ XX_Print("%08x ", *(uint32_t*)p);
-+ size -= 4;
-+ p += 4;
-+ }
-+ XX_Print("\r\n");
-+ }
-+
-+ while (size > 15)
-+ {
-+ XX_Print("0x%08X: %08x %08x %08x %08x\r\n",
-+ p, *(uint32_t *)p, *(uint32_t *)(p + 4),
-+ *(uint32_t *)(p + 8), *(uint32_t *)(p + 12));
-+ size -= 16;
-+ p += 16;
-+ }
-+
-+ if (size)
-+ {
-+ XX_Print("0x%08X: ", p);
-+
-+ while (size > 3)
-+ {
-+ XX_Print("%08x ", *(uint32_t *)p);
-+ size -= 4;
-+ p += 4;
-+ }
-+ while (size)
-+ {
-+ XX_Print("%02x", *(uint8_t *)p);
-+ size--;
-+ p++;
-+ }
-+
-+ XX_Print("\r\n");
-+ }
-+}
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/etc/mm.c
-@@ -0,0 +1,1155 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+#include "string_ext.h"
-+#include "error_ext.h"
-+#include "std_ext.h"
-+#include "part_ext.h"
-+#include "xx_ext.h"
-+
-+#include "mm.h"
-+
-+
-+
-+
-+/**********************************************************************
-+ * MM internal routines set *
-+ **********************************************************************/
-+
-+/****************************************************************
-+ * Routine: CreateBusyBlock
-+ *
-+ * Description:
-+ * Initializes a new busy block of "size" bytes and started
-+ * rom "base" address. Each busy block has a name that
-+ * specified the purpose of the memory allocation.
-+ *
-+ * Arguments:
-+ * base - base address of the busy block
-+ * size - size of the busy block
-+ * name - name that specified the busy block
-+ *
-+ * Return value:
-+ * A pointer to new created structure returned on success;
-+ * Otherwise, NULL.
-+ ****************************************************************/
-+static t_BusyBlock * CreateBusyBlock(uint64_t base, uint64_t size, char *name)
-+{
-+ t_BusyBlock *p_BusyBlock;
-+ uint32_t n;
-+
-+ p_BusyBlock = (t_BusyBlock *)XX_Malloc(sizeof(t_BusyBlock));
-+ if ( !p_BusyBlock )
-+ {
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, NO_MSG);
-+ return NULL;
-+ }
-+
-+ p_BusyBlock->base = base;
-+ p_BusyBlock->end = base + size;
-+
-+ n = strlen(name);
-+ if (n >= MM_MAX_NAME_LEN)
-+ n = MM_MAX_NAME_LEN - 1;
-+ strncpy(p_BusyBlock->name, name, MM_MAX_NAME_LEN-1);
-+ p_BusyBlock->name[n] = '\0';
-+ p_BusyBlock->p_Next = 0;
-+
-+ return p_BusyBlock;
-+}
-+
-+/****************************************************************
-+ * Routine: CreateNewBlock
-+ *
-+ * Description:
-+ * Initializes a new memory block of "size" bytes and started
-+ * from "base" address.
-+ *
-+ * Arguments:
-+ * base - base address of the memory block
-+ * size - size of the memory block
-+ *
-+ * Return value:
-+ * A pointer to new created structure returned on success;
-+ * Otherwise, NULL.
-+ ****************************************************************/
-+static t_MemBlock * CreateNewBlock(uint64_t base, uint64_t size)
-+{
-+ t_MemBlock *p_MemBlock;
-+
-+ p_MemBlock = (t_MemBlock *)XX_Malloc(sizeof(t_MemBlock));
-+ if ( !p_MemBlock )
-+ {
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, NO_MSG);
-+ return NULL;
-+ }
-+
-+ p_MemBlock->base = base;
-+ p_MemBlock->end = base+size;
-+ p_MemBlock->p_Next = 0;
-+
-+ return p_MemBlock;
-+}
-+
-+/****************************************************************
-+ * Routine: CreateFreeBlock
-+ *
-+ * Description:
-+ * Initializes a new free block of of "size" bytes and
-+ * started from "base" address.
-+ *
-+ * Arguments:
-+ * base - base address of the free block
-+ * size - size of the free block
-+ *
-+ * Return value:
-+ * A pointer to new created structure returned on success;
-+ * Otherwise, NULL.
-+ ****************************************************************/
-+static t_FreeBlock * CreateFreeBlock(uint64_t base, uint64_t size)
-+{
-+ t_FreeBlock *p_FreeBlock;
-+
-+ p_FreeBlock = (t_FreeBlock *)XX_Malloc(sizeof(t_FreeBlock));
-+ if ( !p_FreeBlock )
-+ {
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, NO_MSG);
-+ return NULL;
-+ }
-+
-+ p_FreeBlock->base = base;
-+ p_FreeBlock->end = base + size;
-+ p_FreeBlock->p_Next = 0;
-+
-+ return p_FreeBlock;
-+}
-+
-+/****************************************************************
-+ * Routine: AddFree
-+ *
-+ * Description:
-+ * Adds a new free block to the free lists. It updates each
-+ * free list to include a new free block.
-+ * Note, that all free block in each free list are ordered
-+ * by their base address.
-+ *
-+ * Arguments:
-+ * p_MM - pointer to the MM object
-+ * base - base address of a given free block
-+ * end - end address of a given free block
-+ *
-+ * Return value:
-+ *
-+ *
-+ ****************************************************************/
-+static t_Error AddFree(t_MM *p_MM, uint64_t base, uint64_t end)
-+{
-+ t_FreeBlock *p_PrevB, *p_CurrB, *p_NewB;
-+ uint64_t alignment;
-+ uint64_t alignBase;
-+ int i;
-+
-+ /* Updates free lists to include a just released block */
-+ for (i=0; i <= MM_MAX_ALIGNMENT; i++)
-+ {
-+ p_PrevB = p_NewB = 0;
-+ p_CurrB = p_MM->freeBlocks[i];
-+
-+ alignment = (uint64_t)(0x1 << i);
-+ alignBase = MAKE_ALIGNED(base, alignment);
-+
-+ /* Goes to the next free list if there is no block to free */
-+ if (alignBase >= end)
-+ continue;
-+
-+ /* Looks for a free block that should be updated */
-+ while ( p_CurrB )
-+ {
-+ if ( alignBase <= p_CurrB->end )
-+ {
-+ if ( end > p_CurrB->end )
-+ {
-+ t_FreeBlock *p_NextB;
-+ while ( p_CurrB->p_Next && end > p_CurrB->p_Next->end )
-+ {
-+ p_NextB = p_CurrB->p_Next;
-+ p_CurrB->p_Next = p_CurrB->p_Next->p_Next;
-+ XX_Free(p_NextB);
-+ }
-+
-+ p_NextB = p_CurrB->p_Next;
-+ if ( !p_NextB || (p_NextB && end < p_NextB->base) )
-+ {
-+ p_CurrB->end = end;
-+ }
-+ else
-+ {
-+ p_CurrB->end = p_NextB->end;
-+ p_CurrB->p_Next = p_NextB->p_Next;
-+ XX_Free(p_NextB);
-+ }
-+ }
-+ else if ( (end < p_CurrB->base) && ((end-alignBase) >= alignment) )
-+ {
-+ if ((p_NewB = CreateFreeBlock(alignBase, end-alignBase)) == NULL)
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, NO_MSG);
-+
-+ p_NewB->p_Next = p_CurrB;
-+ if (p_PrevB)
-+ p_PrevB->p_Next = p_NewB;
-+ else
-+ p_MM->freeBlocks[i] = p_NewB;
-+ break;
-+ }
-+
-+ if ((alignBase < p_CurrB->base) && (end >= p_CurrB->base))
-+ {
-+ p_CurrB->base = alignBase;
-+ }
-+
-+ /* if size of the free block is less then alignment
-+ * deletes that free block from the free list. */
-+ if ( (p_CurrB->end - p_CurrB->base) < alignment)
-+ {
-+ if ( p_PrevB )
-+ p_PrevB->p_Next = p_CurrB->p_Next;
-+ else
-+ p_MM->freeBlocks[i] = p_CurrB->p_Next;
-+ XX_Free(p_CurrB);
-+ p_CurrB = NULL;
-+ }
-+ break;
-+ }
-+ else
-+ {
-+ p_PrevB = p_CurrB;
-+ p_CurrB = p_CurrB->p_Next;
-+ }
-+ }
-+
-+ /* If no free block found to be updated, insert a new free block
-+ * to the end of the free list.
-+ */
-+ if ( !p_CurrB && ((((uint64_t)(end-base)) & ((uint64_t)(alignment-1))) == 0) )
-+ {
-+ if ((p_NewB = CreateFreeBlock(alignBase, end-base)) == NULL)
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, NO_MSG);
-+
-+ if (p_PrevB)
-+ p_PrevB->p_Next = p_NewB;
-+ else
-+ p_MM->freeBlocks[i] = p_NewB;
-+ }
-+
-+ /* Update boundaries of the new free block */
-+ if ((alignment == 1) && !p_NewB)
-+ {
-+ if ( p_CurrB && base > p_CurrB->base )
-+ base = p_CurrB->base;
-+ if ( p_CurrB && end < p_CurrB->end )
-+ end = p_CurrB->end;
-+ }
-+ }
-+
-+ return (E_OK);
-+}
-+
-+/****************************************************************
-+ * Routine: CutFree
-+ *
-+ * Description:
-+ * Cuts a free block from holdBase to holdEnd from the free lists.
-+ * That is, it updates all free lists of the MM object do
-+ * not include a block of memory from holdBase to holdEnd.
-+ * For each free lists it seek for a free block that holds
-+ * either holdBase or holdEnd. If such block is found it updates it.
-+ *
-+ * Arguments:
-+ * p_MM - pointer to the MM object
-+ * holdBase - base address of the allocated block
-+ * holdEnd - end address of the allocated block
-+ *
-+ * Return value:
-+ * E_OK is returned on success,
-+ * otherwise returns an error code.
-+ *
-+ ****************************************************************/
-+static t_Error CutFree(t_MM *p_MM, uint64_t holdBase, uint64_t holdEnd)
-+{
-+ t_FreeBlock *p_PrevB, *p_CurrB, *p_NewB;
-+ uint64_t alignBase, base, end;
-+ uint64_t alignment;
-+ int i;
-+
-+ for (i=0; i <= MM_MAX_ALIGNMENT; i++)
-+ {
-+ p_PrevB = p_NewB = 0;
-+ p_CurrB = p_MM->freeBlocks[i];
-+
-+ alignment = (uint64_t)(0x1 << i);
-+ alignBase = MAKE_ALIGNED(holdEnd, alignment);
-+
-+ while ( p_CurrB )
-+ {
-+ base = p_CurrB->base;
-+ end = p_CurrB->end;
-+
-+ if ( (holdBase <= base) && (holdEnd <= end) && (holdEnd > base) )
-+ {
-+ if ( alignBase >= end ||
-+ (alignBase < end && ((end-alignBase) < alignment)) )
-+ {
-+ if (p_PrevB)
-+ p_PrevB->p_Next = p_CurrB->p_Next;
-+ else
-+ p_MM->freeBlocks[i] = p_CurrB->p_Next;
-+ XX_Free(p_CurrB);
-+ }
-+ else
-+ {
-+ p_CurrB->base = alignBase;
-+ }
-+ break;
-+ }
-+ else if ( (holdBase > base) && (holdEnd <= end) )
-+ {
-+ if ( (holdBase-base) >= alignment )
-+ {
-+ if ( (alignBase < end) && ((end-alignBase) >= alignment) )
-+ {
-+ if ((p_NewB = CreateFreeBlock(alignBase, end-alignBase)) == NULL)
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, NO_MSG);
-+ p_NewB->p_Next = p_CurrB->p_Next;
-+ p_CurrB->p_Next = p_NewB;
-+ }
-+ p_CurrB->end = holdBase;
-+ }
-+ else if ( (alignBase < end) && ((end-alignBase) >= alignment) )
-+ {
-+ p_CurrB->base = alignBase;
-+ }
-+ else
-+ {
-+ if (p_PrevB)
-+ p_PrevB->p_Next = p_CurrB->p_Next;
-+ else
-+ p_MM->freeBlocks[i] = p_CurrB->p_Next;
-+ XX_Free(p_CurrB);
-+ }
-+ break;
-+ }
-+ else
-+ {
-+ p_PrevB = p_CurrB;
-+ p_CurrB = p_CurrB->p_Next;
-+ }
-+ }
-+ }
-+
-+ return (E_OK);
-+}
-+
-+/****************************************************************
-+ * Routine: AddBusy
-+ *
-+ * Description:
-+ * Adds a new busy block to the list of busy blocks. Note,
-+ * that all busy blocks are ordered by their base address in
-+ * the busy list.
-+ *
-+ * Arguments:
-+ * MM - handler to the MM object
-+ * p_NewBusyB - pointer to the a busy block
-+ *
-+ * Return value:
-+ * None.
-+ *
-+ ****************************************************************/
-+static void AddBusy(t_MM *p_MM, t_BusyBlock *p_NewBusyB)
-+{
-+ t_BusyBlock *p_CurrBusyB, *p_PrevBusyB;
-+
-+ /* finds a place of a new busy block in the list of busy blocks */
-+ p_PrevBusyB = 0;
-+ p_CurrBusyB = p_MM->busyBlocks;
-+
-+ while ( p_CurrBusyB && p_NewBusyB->base > p_CurrBusyB->base )
-+ {
-+ p_PrevBusyB = p_CurrBusyB;
-+ p_CurrBusyB = p_CurrBusyB->p_Next;
-+ }
-+
-+ /* insert the new busy block into the list of busy blocks */
-+ if ( p_CurrBusyB )
-+ p_NewBusyB->p_Next = p_CurrBusyB;
-+ if ( p_PrevBusyB )
-+ p_PrevBusyB->p_Next = p_NewBusyB;
-+ else
-+ p_MM->busyBlocks = p_NewBusyB;
-+}
-+
-+/****************************************************************
-+ * Routine: CutBusy
-+ *
-+ * Description:
-+ * Cuts a block from base to end from the list of busy blocks.
-+ * This is done by updating the list of busy blocks do not
-+ * include a given block, that block is going to be free. If a
-+ * given block is a part of some other busy block, so that
-+ * busy block is updated. If there are number of busy blocks
-+ * included in the given block, so all that blocks are removed
-+ * from the busy list and the end blocks are updated.
-+ * If the given block devides some block into two parts, a new
-+ * busy block is added to the busy list.
-+ *
-+ * Arguments:
-+ * p_MM - pointer to the MM object
-+ * base - base address of a given busy block
-+ * end - end address of a given busy block
-+ *
-+ * Return value:
-+ * E_OK on success, E_NOMEMORY otherwise.
-+ *
-+ ****************************************************************/
-+static t_Error CutBusy(t_MM *p_MM, uint64_t base, uint64_t end)
-+{
-+ t_BusyBlock *p_CurrB, *p_PrevB, *p_NewB;
-+
-+ p_CurrB = p_MM->busyBlocks;
-+ p_PrevB = p_NewB = 0;
-+
-+ while ( p_CurrB )
-+ {
-+ if ( base < p_CurrB->end )
-+ {
-+ if ( end > p_CurrB->end )
-+ {
-+ t_BusyBlock *p_NextB;
-+ while ( p_CurrB->p_Next && end >= p_CurrB->p_Next->end )
-+ {
-+ p_NextB = p_CurrB->p_Next;
-+ p_CurrB->p_Next = p_CurrB->p_Next->p_Next;
-+ XX_Free(p_NextB);
-+ }
-+
-+ p_NextB = p_CurrB->p_Next;
-+ if ( p_NextB && end > p_NextB->base )
-+ {
-+ p_NextB->base = end;
-+ }
-+ }
-+
-+ if ( base <= p_CurrB->base )
-+ {
-+ if ( end < p_CurrB->end && end > p_CurrB->base )
-+ {
-+ p_CurrB->base = end;
-+ }
-+ else if ( end >= p_CurrB->end )
-+ {
-+ if ( p_PrevB )
-+ p_PrevB->p_Next = p_CurrB->p_Next;
-+ else
-+ p_MM->busyBlocks = p_CurrB->p_Next;
-+ XX_Free(p_CurrB);
-+ }
-+ }
-+ else
-+ {
-+ if ( end < p_CurrB->end && end > p_CurrB->base )
-+ {
-+ if ((p_NewB = CreateBusyBlock(end,
-+ p_CurrB->end-end,
-+ p_CurrB->name)) == NULL)
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, NO_MSG);
-+ p_NewB->p_Next = p_CurrB->p_Next;
-+ p_CurrB->p_Next = p_NewB;
-+ }
-+ p_CurrB->end = base;
-+ }
-+ break;
-+ }
-+ else
-+ {
-+ p_PrevB = p_CurrB;
-+ p_CurrB = p_CurrB->p_Next;
-+ }
-+ }
-+
-+ return (E_OK);
-+}
-+
-+/****************************************************************
-+ * Routine: MmGetGreaterAlignment
-+ *
-+ * Description:
-+ * Allocates a block of memory according to the given size
-+ * and the alignment. That routine is called from the MM_Get
-+ * routine if the required alignment is greater then MM_MAX_ALIGNMENT.
-+ * In that case, it goes over free blocks of 64 byte align list
-+ * and checks if it has the required size of bytes of the required
-+ * alignment. If no blocks found returns ILLEGAL_BASE.
-+ * After the block is found and data is allocated, it calls
-+ * the internal CutFree routine to update all free lists
-+ * do not include a just allocated block. Of course, each
-+ * free list contains a free blocks with the same alignment.
-+ * It is also creates a busy block that holds
-+ * information about an allocated block.
-+ *
-+ * Arguments:
-+ * MM - handle to the MM object
-+ * size - size of the MM
-+ * alignment - index as a power of two defines
-+ * a required alignment that is greater then 64.
-+ * name - the name that specifies an allocated block.
-+ *
-+ * Return value:
-+ * base address of an allocated block.
-+ * ILLEGAL_BASE if can't allocate a block
-+ *
-+ ****************************************************************/
-+static uint64_t MmGetGreaterAlignment(t_MM *p_MM, uint64_t size, uint64_t alignment, char* name)
-+{
-+ t_FreeBlock *p_FreeB;
-+ t_BusyBlock *p_NewBusyB;
-+ uint64_t holdBase, holdEnd, alignBase = 0;
-+
-+ /* goes over free blocks of the 64 byte alignment list
-+ and look for a block of the suitable size and
-+ base address according to the alignment. */
-+ p_FreeB = p_MM->freeBlocks[MM_MAX_ALIGNMENT];
-+
-+ while ( p_FreeB )
-+ {
-+ alignBase = MAKE_ALIGNED(p_FreeB->base, alignment);
-+
-+ /* the block is found if the aligned base inside the block
-+ * and has the anough size. */
-+ if ( alignBase >= p_FreeB->base &&
-+ alignBase < p_FreeB->end &&
-+ size <= (p_FreeB->end - alignBase) )
-+ break;
-+ else
-+ p_FreeB = p_FreeB->p_Next;
-+ }
-+
-+ /* If such block isn't found */
-+ if ( !p_FreeB )
-+ return (uint64_t)(ILLEGAL_BASE);
-+
-+ holdBase = alignBase;
-+ holdEnd = alignBase + size;
-+
-+ /* init a new busy block */
-+ if ((p_NewBusyB = CreateBusyBlock(holdBase, size, name)) == NULL)
-+ return (uint64_t)(ILLEGAL_BASE);
-+
-+ /* calls Update routine to update a lists of free blocks */
-+ if ( CutFree ( p_MM, holdBase, holdEnd ) != E_OK )
-+ {
-+ XX_Free(p_NewBusyB);
-+ return (uint64_t)(ILLEGAL_BASE);
-+ }
-+
-+ /* insert the new busy block into the list of busy blocks */
-+ AddBusy ( p_MM, p_NewBusyB );
-+
-+ return (holdBase);
-+}
-+
-+
-+/**********************************************************************
-+ * MM API routines set *
-+ **********************************************************************/
-+
-+/*****************************************************************************/
-+t_Error MM_Init(t_Handle *h_MM, uint64_t base, uint64_t size)
-+{
-+ t_MM *p_MM;
-+ uint64_t newBase, newSize;
-+ int i;
-+
-+ if (!size)
-+ {
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("Size (should be positive)"));
-+ }
-+
-+ /* Initializes a new MM object */
-+ p_MM = (t_MM *)XX_Malloc(sizeof(t_MM));
-+ if (!p_MM)
-+ {
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, NO_MSG);
-+ }
-+
-+ p_MM->h_Spinlock = XX_InitSpinlock();
-+ if (!p_MM->h_Spinlock)
-+ {
-+ XX_Free(p_MM);
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, ("MM spinlock!"));
-+ }
-+
-+ /* Initializes counter of free memory to total size */
-+ p_MM->freeMemSize = size;
-+
-+ /* A busy list is empty */
-+ p_MM->busyBlocks = 0;
-+
-+ /* Initializes a new memory block */
-+ if ((p_MM->memBlocks = CreateNewBlock(base, size)) == NULL)
-+ {
-+ MM_Free(p_MM);
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, NO_MSG);
-+ }
-+
-+ /* Initializes a new free block for each free list*/
-+ for (i=0; i <= MM_MAX_ALIGNMENT; i++)
-+ {
-+ newBase = MAKE_ALIGNED( base, (0x1 << i) );
-+ newSize = size - (newBase - base);
-+
-+ if ((p_MM->freeBlocks[i] = CreateFreeBlock(newBase, newSize)) == NULL)
-+ {
-+ MM_Free(p_MM);
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, NO_MSG);
-+ }
-+ }
-+
-+ *h_MM = p_MM;
-+
-+ return (E_OK);
-+}
-+
-+/*****************************************************************************/
-+void MM_Free(t_Handle h_MM)
-+{
-+ t_MM *p_MM = (t_MM *)h_MM;
-+ t_MemBlock *p_MemBlock;
-+ t_BusyBlock *p_BusyBlock;
-+ t_FreeBlock *p_FreeBlock;
-+ void *p_Block;
-+ int i;
-+
-+ ASSERT_COND(p_MM);
-+
-+ /* release memory allocated for busy blocks */
-+ p_BusyBlock = p_MM->busyBlocks;
-+ while ( p_BusyBlock )
-+ {
-+ p_Block = p_BusyBlock;
-+ p_BusyBlock = p_BusyBlock->p_Next;
-+ XX_Free(p_Block);
-+ }
-+
-+ /* release memory allocated for free blocks */
-+ for (i=0; i <= MM_MAX_ALIGNMENT; i++)
-+ {
-+ p_FreeBlock = p_MM->freeBlocks[i];
-+ while ( p_FreeBlock )
-+ {
-+ p_Block = p_FreeBlock;
-+ p_FreeBlock = p_FreeBlock->p_Next;
-+ XX_Free(p_Block);
-+ }
-+ }
-+
-+ /* release memory allocated for memory blocks */
-+ p_MemBlock = p_MM->memBlocks;
-+ while ( p_MemBlock )
-+ {
-+ p_Block = p_MemBlock;
-+ p_MemBlock = p_MemBlock->p_Next;
-+ XX_Free(p_Block);
-+ }
-+
-+ if (p_MM->h_Spinlock)
-+ XX_FreeSpinlock(p_MM->h_Spinlock);
-+
-+ /* release memory allocated for MM object itself */
-+ XX_Free(p_MM);
-+}
-+
-+/*****************************************************************************/
-+uint64_t MM_Get(t_Handle h_MM, uint64_t size, uint64_t alignment, char* name)
-+{
-+ t_MM *p_MM = (t_MM *)h_MM;
-+ t_FreeBlock *p_FreeB;
-+ t_BusyBlock *p_NewBusyB;
-+ uint64_t holdBase, holdEnd, j, i = 0;
-+ uint32_t intFlags;
-+
-+ SANITY_CHECK_RETURN_VALUE(p_MM, E_INVALID_HANDLE, (uint64_t)ILLEGAL_BASE);
-+
-+ /* checks that alignment value is greater then zero */
-+ if (alignment == 0)
-+ {
-+ alignment = 1;
-+ }
-+
-+ j = alignment;
-+
-+ /* checks if alignment is a power of two, if it correct and if the
-+ required size is multiple of the given alignment. */
-+ while ((j & 0x1) == 0)
-+ {
-+ i++;
-+ j = j >> 1;
-+ }
-+
-+ /* if the given alignment isn't power of two, returns an error */
-+ if (j != 1)
-+ {
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE, ("alignment (should be power of 2)"));
-+ return (uint64_t)ILLEGAL_BASE;
-+ }
-+
-+ if (i > MM_MAX_ALIGNMENT)
-+ {
-+ return (MmGetGreaterAlignment(p_MM, size, alignment, name));
-+ }
-+
-+ intFlags = XX_LockIntrSpinlock(p_MM->h_Spinlock);
-+ /* look for a block of the size greater or equal to the required size. */
-+ p_FreeB = p_MM->freeBlocks[i];
-+ while ( p_FreeB && (p_FreeB->end - p_FreeB->base) < size )
-+ p_FreeB = p_FreeB->p_Next;
-+
-+ /* If such block is found */
-+ if ( !p_FreeB )
-+ {
-+ XX_UnlockIntrSpinlock(p_MM->h_Spinlock, intFlags);
-+ return (uint64_t)(ILLEGAL_BASE);
-+ }
-+
-+ holdBase = p_FreeB->base;
-+ holdEnd = holdBase + size;
-+
-+ /* init a new busy block */
-+ if ((p_NewBusyB = CreateBusyBlock(holdBase, size, name)) == NULL)
-+ {
-+ XX_UnlockIntrSpinlock(p_MM->h_Spinlock, intFlags);
-+ return (uint64_t)(ILLEGAL_BASE);
-+ }
-+
-+ /* calls Update routine to update a lists of free blocks */
-+ if ( CutFree ( p_MM, holdBase, holdEnd ) != E_OK )
-+ {
-+ XX_UnlockIntrSpinlock(p_MM->h_Spinlock, intFlags);
-+ XX_Free(p_NewBusyB);
-+ return (uint64_t)(ILLEGAL_BASE);
-+ }
-+
-+ /* Decreasing the allocated memory size from free memory size */
-+ p_MM->freeMemSize -= size;
-+
-+ /* insert the new busy block into the list of busy blocks */
-+ AddBusy ( p_MM, p_NewBusyB );
-+ XX_UnlockIntrSpinlock(p_MM->h_Spinlock, intFlags);
-+
-+ return (holdBase);
-+}
-+
-+/*****************************************************************************/
-+uint64_t MM_GetForce(t_Handle h_MM, uint64_t base, uint64_t size, char* name)
-+{
-+ t_MM *p_MM = (t_MM *)h_MM;
-+ t_FreeBlock *p_FreeB;
-+ t_BusyBlock *p_NewBusyB;
-+ uint32_t intFlags;
-+ bool blockIsFree = FALSE;
-+
-+ ASSERT_COND(p_MM);
-+
-+ intFlags = XX_LockIntrSpinlock(p_MM->h_Spinlock);
-+ p_FreeB = p_MM->freeBlocks[0]; /* The biggest free blocks are in the
-+ free list with alignment 1 */
-+
-+ while ( p_FreeB )
-+ {
-+ if ( base >= p_FreeB->base && (base+size) <= p_FreeB->end )
-+ {
-+ blockIsFree = TRUE;
-+ break;
-+ }
-+ else
-+ p_FreeB = p_FreeB->p_Next;
-+ }
-+
-+ if ( !blockIsFree )
-+ {
-+ XX_UnlockIntrSpinlock(p_MM->h_Spinlock, intFlags);
-+ return (uint64_t)(ILLEGAL_BASE);
-+ }
-+
-+ /* init a new busy block */
-+ if ((p_NewBusyB = CreateBusyBlock(base, size, name)) == NULL)
-+ {
-+ XX_UnlockIntrSpinlock(p_MM->h_Spinlock, intFlags);
-+ return (uint64_t)(ILLEGAL_BASE);
-+ }
-+
-+ /* calls Update routine to update a lists of free blocks */
-+ if ( CutFree ( p_MM, base, base+size ) != E_OK )
-+ {
-+ XX_UnlockIntrSpinlock(p_MM->h_Spinlock, intFlags);
-+ XX_Free(p_NewBusyB);
-+ return (uint64_t)(ILLEGAL_BASE);
-+ }
-+
-+ /* Decreasing the allocated memory size from free memory size */
-+ p_MM->freeMemSize -= size;
-+
-+ /* insert the new busy block into the list of busy blocks */
-+ AddBusy ( p_MM, p_NewBusyB );
-+ XX_UnlockIntrSpinlock(p_MM->h_Spinlock, intFlags);
-+
-+ return (base);
-+}
-+
-+/*****************************************************************************/
-+uint64_t MM_GetForceMin(t_Handle h_MM, uint64_t size, uint64_t alignment, uint64_t min, char* name)
-+{
-+ t_MM *p_MM = (t_MM *)h_MM;
-+ t_FreeBlock *p_FreeB;
-+ t_BusyBlock *p_NewBusyB;
-+ uint64_t holdBase, holdEnd, j = alignment, i=0;
-+ uint32_t intFlags;
-+
-+ ASSERT_COND(p_MM);
-+
-+ /* checks if alignment is a power of two, if it correct and if the
-+ required size is multiple of the given alignment. */
-+ while ((j & 0x1) == 0)
-+ {
-+ i++;
-+ j = j >> 1;
-+ }
-+
-+ if ( (j != 1) || (i > MM_MAX_ALIGNMENT) )
-+ {
-+ return (uint64_t)(ILLEGAL_BASE);
-+ }
-+
-+ intFlags = XX_LockIntrSpinlock(p_MM->h_Spinlock);
-+ p_FreeB = p_MM->freeBlocks[i];
-+
-+ /* look for the first block that contains the minimum
-+ base address. If the whole required size may be fit
-+ into it, use that block, otherwise look for the next
-+ block of size greater or equal to the required size. */
-+ while ( p_FreeB && (min >= p_FreeB->end))
-+ p_FreeB = p_FreeB->p_Next;
-+
-+ /* If such block is found */
-+ if ( !p_FreeB )
-+ {
-+ XX_UnlockIntrSpinlock(p_MM->h_Spinlock, intFlags);
-+ return (uint64_t)(ILLEGAL_BASE);
-+ }
-+
-+ /* if this block is large enough, use this block */
-+ holdBase = ( min <= p_FreeB->base ) ? p_FreeB->base : min;
-+ if ((holdBase + size) <= p_FreeB->end )
-+ {
-+ holdEnd = holdBase + size;
-+ }
-+ else
-+ {
-+ p_FreeB = p_FreeB->p_Next;
-+ while ( p_FreeB && ((p_FreeB->end - p_FreeB->base) < size) )
-+ p_FreeB = p_FreeB->p_Next;
-+
-+ /* If such block is found */
-+ if ( !p_FreeB )
-+ {
-+ XX_UnlockIntrSpinlock(p_MM->h_Spinlock, intFlags);
-+ return (uint64_t)(ILLEGAL_BASE);
-+ }
-+
-+ holdBase = p_FreeB->base;
-+ holdEnd = holdBase + size;
-+ }
-+
-+ /* init a new busy block */
-+ if ((p_NewBusyB = CreateBusyBlock(holdBase, size, name)) == NULL)
-+ {
-+ XX_UnlockIntrSpinlock(p_MM->h_Spinlock, intFlags);
-+ return (uint64_t)(ILLEGAL_BASE);
-+ }
-+
-+ /* calls Update routine to update a lists of free blocks */
-+ if ( CutFree( p_MM, holdBase, holdEnd ) != E_OK )
-+ {
-+ XX_UnlockIntrSpinlock(p_MM->h_Spinlock, intFlags);
-+ XX_Free(p_NewBusyB);
-+ return (uint64_t)(ILLEGAL_BASE);
-+ }
-+
-+ /* Decreasing the allocated memory size from free memory size */
-+ p_MM->freeMemSize -= size;
-+
-+ /* insert the new busy block into the list of busy blocks */
-+ AddBusy( p_MM, p_NewBusyB );
-+ XX_UnlockIntrSpinlock(p_MM->h_Spinlock, intFlags);
-+
-+ return (holdBase);
-+}
-+
-+/*****************************************************************************/
-+uint64_t MM_Put(t_Handle h_MM, uint64_t base)
-+{
-+ t_MM *p_MM = (t_MM *)h_MM;
-+ t_BusyBlock *p_BusyB, *p_PrevBusyB;
-+ uint64_t size;
-+ uint32_t intFlags;
-+
-+ ASSERT_COND(p_MM);
-+
-+ /* Look for a busy block that have the given base value.
-+ * That block will be returned back to the memory.
-+ */
-+ p_PrevBusyB = 0;
-+
-+ intFlags = XX_LockIntrSpinlock(p_MM->h_Spinlock);
-+ p_BusyB = p_MM->busyBlocks;
-+ while ( p_BusyB && base != p_BusyB->base )
-+ {
-+ p_PrevBusyB = p_BusyB;
-+ p_BusyB = p_BusyB->p_Next;
-+ }
-+
-+ if ( !p_BusyB )
-+ {
-+ XX_UnlockIntrSpinlock(p_MM->h_Spinlock, intFlags);
-+ return (uint64_t)(0);
-+ }
-+
-+ if ( AddFree( p_MM, p_BusyB->base, p_BusyB->end ) != E_OK )
-+ {
-+ XX_UnlockIntrSpinlock(p_MM->h_Spinlock, intFlags);
-+ return (uint64_t)(0);
-+ }
-+
-+ /* removes a busy block form the list of busy blocks */
-+ if ( p_PrevBusyB )
-+ p_PrevBusyB->p_Next = p_BusyB->p_Next;
-+ else
-+ p_MM->busyBlocks = p_BusyB->p_Next;
-+
-+ size = p_BusyB->end - p_BusyB->base;
-+
-+ /* Adding the deallocated memory size to free memory size */
-+ p_MM->freeMemSize += size;
-+
-+ XX_Free(p_BusyB);
-+ XX_UnlockIntrSpinlock(p_MM->h_Spinlock, intFlags);
-+
-+ return (size);
-+}
-+
-+/*****************************************************************************/
-+uint64_t MM_PutForce(t_Handle h_MM, uint64_t base, uint64_t size)
-+{
-+ t_MM *p_MM = (t_MM *)h_MM;
-+ uint64_t end = base + size;
-+ uint32_t intFlags;
-+
-+ ASSERT_COND(p_MM);
-+
-+ intFlags = XX_LockIntrSpinlock(p_MM->h_Spinlock);
-+
-+ if ( CutBusy( p_MM, base, end ) != E_OK )
-+ {
-+ XX_UnlockIntrSpinlock(p_MM->h_Spinlock, intFlags);
-+ return (uint64_t)(0);
-+ }
-+
-+ if ( AddFree ( p_MM, base, end ) != E_OK )
-+ {
-+ XX_UnlockIntrSpinlock(p_MM->h_Spinlock, intFlags);
-+ return (uint64_t)(0);
-+ }
-+
-+ /* Adding the deallocated memory size to free memory size */
-+ p_MM->freeMemSize += size;
-+
-+ XX_UnlockIntrSpinlock(p_MM->h_Spinlock, intFlags);
-+
-+ return (size);
-+}
-+
-+/*****************************************************************************/
-+t_Error MM_Add(t_Handle h_MM, uint64_t base, uint64_t size)
-+{
-+ t_MM *p_MM = (t_MM *)h_MM;
-+ t_MemBlock *p_MemB, *p_NewMemB;
-+ t_Error errCode;
-+ uint32_t intFlags;
-+
-+ ASSERT_COND(p_MM);
-+
-+ /* find a last block in the list of memory blocks to insert a new
-+ * memory block
-+ */
-+ intFlags = XX_LockIntrSpinlock(p_MM->h_Spinlock);
-+
-+ p_MemB = p_MM->memBlocks;
-+ while ( p_MemB->p_Next )
-+ {
-+ if ( base >= p_MemB->base && base < p_MemB->end )
-+ {
-+ XX_UnlockIntrSpinlock(p_MM->h_Spinlock, intFlags);
-+ RETURN_ERROR(MAJOR, E_ALREADY_EXISTS, NO_MSG);
-+ }
-+ p_MemB = p_MemB->p_Next;
-+ }
-+ /* check for a last memory block */
-+ if ( base >= p_MemB->base && base < p_MemB->end )
-+ {
-+ XX_UnlockIntrSpinlock(p_MM->h_Spinlock, intFlags);
-+ RETURN_ERROR(MAJOR, E_ALREADY_EXISTS, NO_MSG);
-+ }
-+
-+ /* create a new memory block */
-+ if ((p_NewMemB = CreateNewBlock(base, size)) == NULL)
-+ {
-+ XX_UnlockIntrSpinlock(p_MM->h_Spinlock, intFlags);
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, NO_MSG);
-+ }
-+
-+ /* append a new memory block to the end of the list of memory blocks */
-+ p_MemB->p_Next = p_NewMemB;
-+
-+ /* add a new free block to the free lists */
-+ errCode = AddFree(p_MM, base, base+size);
-+ if (errCode)
-+ {
-+ XX_UnlockIntrSpinlock(p_MM->h_Spinlock, intFlags);
-+ p_MemB->p_Next = 0;
-+ XX_Free(p_NewMemB);
-+ return ((t_Error)errCode);
-+ }
-+
-+ /* Adding the new block size to free memory size */
-+ p_MM->freeMemSize += size;
-+
-+ XX_UnlockIntrSpinlock(p_MM->h_Spinlock, intFlags);
-+
-+ return (E_OK);
-+}
-+
-+/*****************************************************************************/
-+uint64_t MM_GetMemBlock(t_Handle h_MM, int index)
-+{
-+ t_MM *p_MM = (t_MM*)h_MM;
-+ t_MemBlock *p_MemBlock;
-+ int i;
-+
-+ ASSERT_COND(p_MM);
-+
-+ p_MemBlock = p_MM->memBlocks;
-+ for (i=0; i < index; i++)
-+ p_MemBlock = p_MemBlock->p_Next;
-+
-+ if ( p_MemBlock )
-+ return (p_MemBlock->base);
-+ else
-+ return (uint64_t)ILLEGAL_BASE;
-+}
-+
-+/*****************************************************************************/
-+uint64_t MM_GetBase(t_Handle h_MM)
-+{
-+ t_MM *p_MM = (t_MM*)h_MM;
-+ t_MemBlock *p_MemBlock;
-+
-+ ASSERT_COND(p_MM);
-+
-+ p_MemBlock = p_MM->memBlocks;
-+ return p_MemBlock->base;
-+}
-+
-+/*****************************************************************************/
-+bool MM_InRange(t_Handle h_MM, uint64_t addr)
-+{
-+ t_MM *p_MM = (t_MM*)h_MM;
-+ t_MemBlock *p_MemBlock;
-+
-+ ASSERT_COND(p_MM);
-+
-+ p_MemBlock = p_MM->memBlocks;
-+
-+ if ((addr >= p_MemBlock->base) && (addr < p_MemBlock->end))
-+ return TRUE;
-+ else
-+ return FALSE;
-+}
-+
-+/*****************************************************************************/
-+uint64_t MM_GetFreeMemSize(t_Handle h_MM)
-+{
-+ t_MM *p_MM = (t_MM*)h_MM;
-+
-+ ASSERT_COND(p_MM);
-+
-+ return p_MM->freeMemSize;
-+}
-+
-+/*****************************************************************************/
-+void MM_Dump(t_Handle h_MM)
-+{
-+ t_MM *p_MM = (t_MM *)h_MM;
-+ t_FreeBlock *p_FreeB;
-+ t_BusyBlock *p_BusyB;
-+ int i;
-+
-+ p_BusyB = p_MM->busyBlocks;
-+ XX_Print("List of busy blocks:\n");
-+ while (p_BusyB)
-+ {
-+ XX_Print("\t0x%p: (%s: b=0x%llx, e=0x%llx)\n", p_BusyB, p_BusyB->name, p_BusyB->base, p_BusyB->end );
-+ p_BusyB = p_BusyB->p_Next;
-+ }
-+
-+ XX_Print("\nLists of free blocks according to alignment:\n");
-+ for (i=0; i <= MM_MAX_ALIGNMENT; i++)
-+ {
-+ XX_Print("%d alignment:\n", (0x1 << i));
-+ p_FreeB = p_MM->freeBlocks[i];
-+ while (p_FreeB)
-+ {
-+ XX_Print("\t0x%p: (b=0x%llx, e=0x%llx)\n", p_FreeB, p_FreeB->base, p_FreeB->end);
-+ p_FreeB = p_FreeB->p_Next;
-+ }
-+ XX_Print("\n");
-+ }
-+}
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/etc/mm.h
-@@ -0,0 +1,105 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/****************************************************************
-+ *
-+ * File: mm.h
-+ *
-+ *
-+ * Description:
-+ * MM (Memory Management) object definitions.
-+ * It also includes definitions of the Free Block, Busy Block
-+ * and Memory Block structures used by the MM object.
-+ *
-+ ****************************************************************/
-+
-+#ifndef __MM_H
-+#define __MM_H
-+
-+
-+#include "mm_ext.h"
-+
-+#define __ERR_MODULE__ MODULE_MM
-+
-+
-+#define MAKE_ALIGNED(addr, align) \
-+ (((uint64_t)(addr) + ((align) - 1)) & (~(((uint64_t)align) - 1)))
-+
-+
-+/* t_MemBlock data structure defines parameters of the Memory Block */
-+typedef struct t_MemBlock
-+{
-+ struct t_MemBlock *p_Next; /* Pointer to the next memory block */
-+
-+ uint64_t base; /* Base address of the memory block */
-+ uint64_t end; /* End address of the memory block */
-+} t_MemBlock;
-+
-+
-+/* t_FreeBlock data structure defines parameters of the Free Block */
-+typedef struct t_FreeBlock
-+{
-+ struct t_FreeBlock *p_Next; /* Pointer to the next free block */
-+
-+ uint64_t base; /* Base address of the block */
-+ uint64_t end; /* End address of the block */
-+} t_FreeBlock;
-+
-+
-+/* t_BusyBlock data structure defines parameters of the Busy Block */
-+typedef struct t_BusyBlock
-+{
-+ struct t_BusyBlock *p_Next; /* Pointer to the next free block */
-+
-+ uint64_t base; /* Base address of the block */
-+ uint64_t end; /* End address of the block */
-+ char name[MM_MAX_NAME_LEN]; /* That block of memory was allocated for
-+ something specified by the Name */
-+} t_BusyBlock;
-+
-+
-+/* t_MM data structure defines parameters of the MM object */
-+typedef struct t_MM
-+{
-+ t_Handle h_Spinlock;
-+
-+ t_MemBlock *memBlocks; /* List of memory blocks (Memory list) */
-+ t_BusyBlock *busyBlocks; /* List of busy blocks (Busy list) */
-+ t_FreeBlock *freeBlocks[MM_MAX_ALIGNMENT + 1];
-+ /* Alignment lists of free blocks (Free lists) */
-+
-+ uint64_t freeMemSize; /* Total size of free memory (in bytes) */
-+} t_MM;
-+
-+
-+#endif /* __MM_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/etc/sprint.c
-@@ -0,0 +1,81 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/*------------------------------------------------------*/
-+/* File: sprint.c */
-+/* */
-+/* Description: */
-+/* Debug routines (externals) */
-+/*------------------------------------------------------*/
-+#include "string_ext.h"
-+#include "stdlib_ext.h"
-+#include "stdarg_ext.h"
-+#include "sprint_ext.h"
-+#include "std_ext.h"
-+#include "xx_ext.h"
-+
-+
-+int Sprint(char * buf, const char *fmt, ...)
-+{
-+ va_list args;
-+ int i;
-+
-+ va_start(args, fmt);
-+ i=vsprintf(buf,fmt,args);
-+ va_end(args);
-+ return i;
-+}
-+
-+int Snprint(char * buf, uint32_t size, const char *fmt, ...)
-+{
-+ va_list args;
-+ int i;
-+
-+ va_start(args, fmt);
-+ i=vsnprintf(buf,size,fmt,args);
-+ va_end(args);
-+ return i;
-+}
-+
-+#ifndef NCSW_VXWORKS
-+int Sscan(const char * buf, const char * fmt, ...)
-+{
-+ va_list args;
-+ int i;
-+
-+ va_start(args,fmt);
-+ i = vsscanf(buf,fmt,args);
-+ va_end(args);
-+ return i;
-+}
-+#endif /* NCSW_VXWORKS */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/fmanv3h_dflags.h
-@@ -0,0 +1,57 @@
-+/*
-+ * Copyright 2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#ifndef __dflags_h
-+#define __dflags_h
-+
-+
-+#define NCSW_LINUX
-+
-+#define T4240
-+#define NCSW_PPC_CORE
-+
-+#define DEBUG_ERRORS 1
-+
-+#if defined(DEBUG)
-+#define DEBUG_GLOBAL_LEVEL REPORT_LEVEL_INFO
-+
-+#define DEBUG_XX_MALLOC
-+#define DEBUG_MEM_LEAKS
-+
-+#else
-+#define DEBUG_GLOBAL_LEVEL REPORT_LEVEL_WARNING
-+#endif /* (DEBUG) */
-+
-+#define REPORT_EVENTS 1
-+#define EVENT_GLOBAL_LEVEL REPORT_LEVEL_MINOR
-+
-+#endif /* __dflags_h */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/fmanv3l_dflags.h
-@@ -0,0 +1,56 @@
-+/*
-+ * Copyright 2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#ifndef __dflags_h
-+#define __dflags_h
-+
-+
-+#define NCSW_LINUX
-+
-+#define NCSW_PPC_CORE
-+
-+#define DEBUG_ERRORS 1
-+
-+#if defined(DEBUG)
-+#define DEBUG_GLOBAL_LEVEL REPORT_LEVEL_INFO
-+
-+#define DEBUG_XX_MALLOC
-+#define DEBUG_MEM_LEAKS
-+
-+#else
-+#define DEBUG_GLOBAL_LEVEL REPORT_LEVEL_WARNING
-+#endif /* (DEBUG) */
-+
-+#define REPORT_EVENTS 1
-+#define EVENT_GLOBAL_LEVEL REPORT_LEVEL_MINOR
-+
-+#endif /* __dflags_h */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/Peripherals/crc_mac_addr_ext.h
-@@ -0,0 +1,364 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/*------------------------------------------------------*/
-+/* */
-+/* File: crc_mac_addr_ext.h */
-+/* */
-+/* Description: */
-+/* Define a macro that calculate the crc value of */
-+/* an Ethernet MAC address (48 bitd address */
-+/*------------------------------------------------------*/
-+
-+#ifndef __crc_mac_addr_ext_h
-+#define __crc_mac_addr_ext_h
-+
-+#include "std_ext.h"
-+
-+
-+static uint32_t crc_table[256] =
-+{
-+ 0x00000000,
-+ 0x77073096,
-+ 0xee0e612c,
-+ 0x990951ba,
-+ 0x076dc419,
-+ 0x706af48f,
-+ 0xe963a535,
-+ 0x9e6495a3,
-+ 0x0edb8832,
-+ 0x79dcb8a4,
-+ 0xe0d5e91e,
-+ 0x97d2d988,
-+ 0x09b64c2b,
-+ 0x7eb17cbd,
-+ 0xe7b82d07,
-+ 0x90bf1d91,
-+ 0x1db71064,
-+ 0x6ab020f2,
-+ 0xf3b97148,
-+ 0x84be41de,
-+ 0x1adad47d,
-+ 0x6ddde4eb,
-+ 0xf4d4b551,
-+ 0x83d385c7,
-+ 0x136c9856,
-+ 0x646ba8c0,
-+ 0xfd62f97a,
-+ 0x8a65c9ec,
-+ 0x14015c4f,
-+ 0x63066cd9,
-+ 0xfa0f3d63,
-+ 0x8d080df5,
-+ 0x3b6e20c8,
-+ 0x4c69105e,
-+ 0xd56041e4,
-+ 0xa2677172,
-+ 0x3c03e4d1,
-+ 0x4b04d447,
-+ 0xd20d85fd,
-+ 0xa50ab56b,
-+ 0x35b5a8fa,
-+ 0x42b2986c,
-+ 0xdbbbc9d6,
-+ 0xacbcf940,
-+ 0x32d86ce3,
-+ 0x45df5c75,
-+ 0xdcd60dcf,
-+ 0xabd13d59,
-+ 0x26d930ac,
-+ 0x51de003a,
-+ 0xc8d75180,
-+ 0xbfd06116,
-+ 0x21b4f4b5,
-+ 0x56b3c423,
-+ 0xcfba9599,
-+ 0xb8bda50f,
-+ 0x2802b89e,
-+ 0x5f058808,
-+ 0xc60cd9b2,
-+ 0xb10be924,
-+ 0x2f6f7c87,
-+ 0x58684c11,
-+ 0xc1611dab,
-+ 0xb6662d3d,
-+ 0x76dc4190,
-+ 0x01db7106,
-+ 0x98d220bc,
-+ 0xefd5102a,
-+ 0x71b18589,
-+ 0x06b6b51f,
-+ 0x9fbfe4a5,
-+ 0xe8b8d433,
-+ 0x7807c9a2,
-+ 0x0f00f934,
-+ 0x9609a88e,
-+ 0xe10e9818,
-+ 0x7f6a0dbb,
-+ 0x086d3d2d,
-+ 0x91646c97,
-+ 0xe6635c01,
-+ 0x6b6b51f4,
-+ 0x1c6c6162,
-+ 0x856530d8,
-+ 0xf262004e,
-+ 0x6c0695ed,
-+ 0x1b01a57b,
-+ 0x8208f4c1,
-+ 0xf50fc457,
-+ 0x65b0d9c6,
-+ 0x12b7e950,
-+ 0x8bbeb8ea,
-+ 0xfcb9887c,
-+ 0x62dd1ddf,
-+ 0x15da2d49,
-+ 0x8cd37cf3,
-+ 0xfbd44c65,
-+ 0x4db26158,
-+ 0x3ab551ce,
-+ 0xa3bc0074,
-+ 0xd4bb30e2,
-+ 0x4adfa541,
-+ 0x3dd895d7,
-+ 0xa4d1c46d,
-+ 0xd3d6f4fb,
-+ 0x4369e96a,
-+ 0x346ed9fc,
-+ 0xad678846,
-+ 0xda60b8d0,
-+ 0x44042d73,
-+ 0x33031de5,
-+ 0xaa0a4c5f,
-+ 0xdd0d7cc9,
-+ 0x5005713c,
-+ 0x270241aa,
-+ 0xbe0b1010,
-+ 0xc90c2086,
-+ 0x5768b525,
-+ 0x206f85b3,
-+ 0xb966d409,
-+ 0xce61e49f,
-+ 0x5edef90e,
-+ 0x29d9c998,
-+ 0xb0d09822,
-+ 0xc7d7a8b4,
-+ 0x59b33d17,
-+ 0x2eb40d81,
-+ 0xb7bd5c3b,
-+ 0xc0ba6cad,
-+ 0xedb88320,
-+ 0x9abfb3b6,
-+ 0x03b6e20c,
-+ 0x74b1d29a,
-+ 0xead54739,
-+ 0x9dd277af,
-+ 0x04db2615,
-+ 0x73dc1683,
-+ 0xe3630b12,
-+ 0x94643b84,
-+ 0x0d6d6a3e,
-+ 0x7a6a5aa8,
-+ 0xe40ecf0b,
-+ 0x9309ff9d,
-+ 0x0a00ae27,
-+ 0x7d079eb1,
-+ 0xf00f9344,
-+ 0x8708a3d2,
-+ 0x1e01f268,
-+ 0x6906c2fe,
-+ 0xf762575d,
-+ 0x806567cb,
-+ 0x196c3671,
-+ 0x6e6b06e7,
-+ 0xfed41b76,
-+ 0x89d32be0,
-+ 0x10da7a5a,
-+ 0x67dd4acc,
-+ 0xf9b9df6f,
-+ 0x8ebeeff9,
-+ 0x17b7be43,
-+ 0x60b08ed5,
-+ 0xd6d6a3e8,
-+ 0xa1d1937e,
-+ 0x38d8c2c4,
-+ 0x4fdff252,
-+ 0xd1bb67f1,
-+ 0xa6bc5767,
-+ 0x3fb506dd,
-+ 0x48b2364b,
-+ 0xd80d2bda,
-+ 0xaf0a1b4c,
-+ 0x36034af6,
-+ 0x41047a60,
-+ 0xdf60efc3,
-+ 0xa867df55,
-+ 0x316e8eef,
-+ 0x4669be79,
-+ 0xcb61b38c,
-+ 0xbc66831a,
-+ 0x256fd2a0,
-+ 0x5268e236,
-+ 0xcc0c7795,
-+ 0xbb0b4703,
-+ 0x220216b9,
-+ 0x5505262f,
-+ 0xc5ba3bbe,
-+ 0xb2bd0b28,
-+ 0x2bb45a92,
-+ 0x5cb36a04,
-+ 0xc2d7ffa7,
-+ 0xb5d0cf31,
-+ 0x2cd99e8b,
-+ 0x5bdeae1d,
-+ 0x9b64c2b0,
-+ 0xec63f226,
-+ 0x756aa39c,
-+ 0x026d930a,
-+ 0x9c0906a9,
-+ 0xeb0e363f,
-+ 0x72076785,
-+ 0x05005713,
-+ 0x95bf4a82,
-+ 0xe2b87a14,
-+ 0x7bb12bae,
-+ 0x0cb61b38,
-+ 0x92d28e9b,
-+ 0xe5d5be0d,
-+ 0x7cdcefb7,
-+ 0x0bdbdf21,
-+ 0x86d3d2d4,
-+ 0xf1d4e242,
-+ 0x68ddb3f8,
-+ 0x1fda836e,
-+ 0x81be16cd,
-+ 0xf6b9265b,
-+ 0x6fb077e1,
-+ 0x18b74777,
-+ 0x88085ae6,
-+ 0xff0f6a70,
-+ 0x66063bca,
-+ 0x11010b5c,
-+ 0x8f659eff,
-+ 0xf862ae69,
-+ 0x616bffd3,
-+ 0x166ccf45,
-+ 0xa00ae278,
-+ 0xd70dd2ee,
-+ 0x4e048354,
-+ 0x3903b3c2,
-+ 0xa7672661,
-+ 0xd06016f7,
-+ 0x4969474d,
-+ 0x3e6e77db,
-+ 0xaed16a4a,
-+ 0xd9d65adc,
-+ 0x40df0b66,
-+ 0x37d83bf0,
-+ 0xa9bcae53,
-+ 0xdebb9ec5,
-+ 0x47b2cf7f,
-+ 0x30b5ffe9,
-+ 0xbdbdf21c,
-+ 0xcabac28a,
-+ 0x53b39330,
-+ 0x24b4a3a6,
-+ 0xbad03605,
-+ 0xcdd70693,
-+ 0x54de5729,
-+ 0x23d967bf,
-+ 0xb3667a2e,
-+ 0xc4614ab8,
-+ 0x5d681b02,
-+ 0x2a6f2b94,
-+ 0xb40bbe37,
-+ 0xc30c8ea1,
-+ 0x5a05df1b,
-+ 0x2d02ef8d
-+};
-+
-+
-+#define GET_MAC_ADDR_CRC(addr, crc) \
-+{ \
-+ uint32_t i; \
-+ uint8_t data; \
-+ \
-+ /* CRC calculation */ \
-+ crc = 0xffffffff; \
-+ for (i=0; i < 6; i++) \
-+ { \
-+ data = (uint8_t)(addr >> ((5-i)*8)); \
-+ crc = crc^data; \
-+ crc = crc_table[crc&0xff] ^ (crc>>8); \
-+ } \
-+} \
-+
-+/* Define a macro for getting the mirrored value of */
-+/* a byte size number. (0x11010011 --> 0x11001011) */
-+/* Sometimes the mirrored value of the CRC is required */
-+static __inline__ uint8_t GetMirror(uint8_t n)
-+{
-+ uint8_t mirror[16] =
-+ {
-+ 0x00,
-+ 0x08,
-+ 0x04,
-+ 0x0c,
-+ 0x02,
-+ 0x0a,
-+ 0x06,
-+ 0x0e,
-+ 0x01,
-+ 0x09,
-+ 0x05,
-+ 0x0d,
-+ 0x03,
-+ 0x0b,
-+ 0x07,
-+ 0x0f
-+ };
-+ return ((uint8_t)(((mirror[n & 0x0f] << 4) | (mirror[n >> 4]))));
-+}
-+
-+static __inline__ uint32_t GetMirror32(uint32_t n)
-+{
-+ return (((uint32_t)GetMirror((uint8_t)(n))<<24) |
-+ ((uint32_t)GetMirror((uint8_t)(n>>8))<<16) |
-+ ((uint32_t)GetMirror((uint8_t)(n>>16))<<8) |
-+ ((uint32_t)GetMirror((uint8_t)(n>>24))));
-+}
-+
-+#define MIRROR GetMirror
-+#define MIRROR_32 GetMirror32
-+
-+
-+#endif /* __crc_mac_addr_ext_h */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/Peripherals/dpaa_ext.h
-@@ -0,0 +1,210 @@
-+/* Copyright (c) 2008-2012 Freescale Semiconductor, Inc
-+ * All rights reserved.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/**************************************************************************//**
-+ @File dpaa_ext.h
-+
-+ @Description DPAA Application Programming Interface.
-+*//***************************************************************************/
-+#ifndef __DPAA_EXT_H
-+#define __DPAA_EXT_H
-+
-+#include "std_ext.h"
-+#include "error_ext.h"
-+
-+
-+/**************************************************************************//**
-+ @Group DPAA_grp Data Path Acceleration Architecture API
-+
-+ @Description DPAA API functions, definitions and enums.
-+
-+ @{
-+*//***************************************************************************/
-+
-+#if defined(__MWERKS__) && !defined(__GNUC__)
-+#pragma pack(push,1)
-+#endif /* defined(__MWERKS__) && ... */
-+
-+/**************************************************************************//**
-+ @Description Frame descriptor
-+*//***************************************************************************/
-+typedef _Packed struct t_DpaaFD {
-+#if __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
-+ volatile uint8_t liodn;
-+ volatile uint8_t bpid;
-+ volatile uint8_t elion;
-+ volatile uint8_t addrh;
-+ volatile uint32_t addrl;
-+#else
-+ volatile uint32_t addrl;
-+ volatile uint8_t addrh;
-+ volatile uint8_t elion;
-+ volatile uint8_t bpid;
-+ volatile uint8_t liodn;
-+ #endif
-+ volatile uint32_t length; /**< Frame length */
-+ volatile uint32_t status; /**< FD status */
-+} _PackedType t_DpaaFD;
-+
-+/**************************************************************************//**
-+ @Description enum for defining frame format
-+*//***************************************************************************/
-+typedef enum e_DpaaFDFormatType {
-+ e_DPAA_FD_FORMAT_TYPE_SHORT_SBSF = 0x0, /**< Simple frame Single buffer; Offset and
-+ small length (9b OFFSET, 20b LENGTH) */
-+ e_DPAA_FD_FORMAT_TYPE_LONG_SBSF = 0x2, /**< Simple frame, single buffer; big length
-+ (29b LENGTH ,No OFFSET) */
-+ e_DPAA_FD_FORMAT_TYPE_SHORT_MBSF = 0x4, /**< Simple frame, Scatter Gather table; Offset
-+ and small length (9b OFFSET, 20b LENGTH) */
-+ e_DPAA_FD_FORMAT_TYPE_LONG_MBSF = 0x6, /**< Simple frame, Scatter Gather table;
-+ big length (29b LENGTH ,No OFFSET) */
-+ e_DPAA_FD_FORMAT_TYPE_COMPOUND = 0x1, /**< Compound Frame (29b CONGESTION-WEIGHT
-+ No LENGTH or OFFSET) */
-+ e_DPAA_FD_FORMAT_TYPE_DUMMY
-+} e_DpaaFDFormatType;
-+
-+/**************************************************************************//**
-+ @Collection Frame descriptor macros
-+*//***************************************************************************/
-+#define DPAA_FD_DD_MASK 0xc0000000 /**< FD DD field mask */
-+#define DPAA_FD_PID_MASK 0x3f000000 /**< FD PID field mask */
-+#define DPAA_FD_ELIODN_MASK 0x0000f000 /**< FD ELIODN field mask */
-+#define DPAA_FD_BPID_MASK 0x00ff0000 /**< FD BPID field mask */
-+#define DPAA_FD_ADDRH_MASK 0x000000ff /**< FD ADDRH field mask */
-+#define DPAA_FD_ADDRL_MASK 0xffffffff /**< FD ADDRL field mask */
-+#define DPAA_FD_FORMAT_MASK 0xe0000000 /**< FD FORMAT field mask */
-+#define DPAA_FD_OFFSET_MASK 0x1ff00000 /**< FD OFFSET field mask */
-+#define DPAA_FD_LENGTH_MASK 0x000fffff /**< FD LENGTH field mask */
-+
-+#define DPAA_FD_GET_ADDRH(fd) ((t_DpaaFD *)fd)->addrh /**< Macro to get FD ADDRH field */
-+#define DPAA_FD_GET_ADDRL(fd) ((t_DpaaFD *)fd)->addrl /**< Macro to get FD ADDRL field */
-+#define DPAA_FD_GET_PHYS_ADDR(fd) ((physAddress_t)(((uint64_t)DPAA_FD_GET_ADDRH(fd) << 32) | (uint64_t)DPAA_FD_GET_ADDRL(fd))) /**< Macro to get FD ADDR field */
-+#define DPAA_FD_GET_FORMAT(fd) ((((t_DpaaFD *)fd)->length & DPAA_FD_FORMAT_MASK) >> (31-2)) /**< Macro to get FD FORMAT field */
-+#define DPAA_FD_GET_OFFSET(fd) ((((t_DpaaFD *)fd)->length & DPAA_FD_OFFSET_MASK) >> (31-11)) /**< Macro to get FD OFFSET field */
-+#define DPAA_FD_GET_LENGTH(fd) (((t_DpaaFD *)fd)->length & DPAA_FD_LENGTH_MASK) /**< Macro to get FD LENGTH field */
-+#define DPAA_FD_GET_STATUS(fd) ((t_DpaaFD *)fd)->status /**< Macro to get FD STATUS field */
-+#define DPAA_FD_GET_ADDR(fd) XX_PhysToVirt(DPAA_FD_GET_PHYS_ADDR(fd)) /**< Macro to get FD ADDR (virtual) */
-+
-+#define DPAA_FD_SET_ADDRH(fd,val) ((t_DpaaFD *)fd)->addrh = (val) /**< Macro to set FD ADDRH field */
-+#define DPAA_FD_SET_ADDRL(fd,val) ((t_DpaaFD *)fd)->addrl = (val) /**< Macro to set FD ADDRL field */
-+#define DPAA_FD_SET_ADDR(fd,val) \
-+do { \
-+ uint64_t physAddr = (uint64_t)(XX_VirtToPhys(val)); \
-+ DPAA_FD_SET_ADDRH(fd, ((uint32_t)(physAddr >> 32))); \
-+ DPAA_FD_SET_ADDRL(fd, (uint32_t)physAddr); \
-+} while (0) /**< Macro to set FD ADDR field */
-+#define DPAA_FD_SET_FORMAT(fd,val) (((t_DpaaFD *)fd)->length = ((((t_DpaaFD *)fd)->length & ~DPAA_FD_FORMAT_MASK) | (((val) << (31-2))& DPAA_FD_FORMAT_MASK))) /**< Macro to set FD FORMAT field */
-+#define DPAA_FD_SET_OFFSET(fd,val) (((t_DpaaFD *)fd)->length = ((((t_DpaaFD *)fd)->length & ~DPAA_FD_OFFSET_MASK) | (((val) << (31-11))& DPAA_FD_OFFSET_MASK) )) /**< Macro to set FD OFFSET field */
-+#define DPAA_FD_SET_LENGTH(fd,val) (((t_DpaaFD *)fd)->length = (((t_DpaaFD *)fd)->length & ~DPAA_FD_LENGTH_MASK) | ((val) & DPAA_FD_LENGTH_MASK)) /**< Macro to set FD LENGTH field */
-+#define DPAA_FD_SET_STATUS(fd,val) ((t_DpaaFD *)fd)->status = (val) /**< Macro to set FD STATUS field */
-+/* @} */
-+
-+/**************************************************************************//**
-+ @Description Frame Scatter/Gather Table Entry
-+*//***************************************************************************/
-+typedef _Packed struct t_DpaaSGTE {
-+ volatile uint32_t addrh; /**< Buffer Address high */
-+ volatile uint32_t addrl; /**< Buffer Address low */
-+ volatile uint32_t length; /**< Buffer length */
-+ volatile uint32_t offset; /**< SGTE offset */
-+} _PackedType t_DpaaSGTE;
-+
-+#define DPAA_NUM_OF_SG_TABLE_ENTRY 16
-+
-+/**************************************************************************//**
-+ @Description Frame Scatter/Gather Table
-+*//***************************************************************************/
-+typedef _Packed struct t_DpaaSGT {
-+ t_DpaaSGTE tableEntry[DPAA_NUM_OF_SG_TABLE_ENTRY];
-+ /**< Structure that holds information about
-+ a single S/G entry. */
-+} _PackedType t_DpaaSGT;
-+
-+/**************************************************************************//**
-+ @Description Compound Frame Table
-+*//***************************************************************************/
-+typedef _Packed struct t_DpaaCompTbl {
-+ t_DpaaSGTE outputBuffInfo; /**< Structure that holds information about
-+ the compound-frame output buffer;
-+ NOTE: this may point to a S/G table */
-+ t_DpaaSGTE inputBuffInfo; /**< Structure that holds information about
-+ the compound-frame input buffer;
-+ NOTE: this may point to a S/G table */
-+} _PackedType t_DpaaCompTbl;
-+
-+/**************************************************************************//**
-+ @Collection Frame Scatter/Gather Table Entry macros
-+*//***************************************************************************/
-+#define DPAA_SGTE_ADDRH_MASK 0x000000ff /**< SGTE ADDRH field mask */
-+#define DPAA_SGTE_ADDRL_MASK 0xffffffff /**< SGTE ADDRL field mask */
-+#define DPAA_SGTE_E_MASK 0x80000000 /**< SGTE Extension field mask */
-+#define DPAA_SGTE_F_MASK 0x40000000 /**< SGTE Final field mask */
-+#define DPAA_SGTE_LENGTH_MASK 0x3fffffff /**< SGTE LENGTH field mask */
-+#define DPAA_SGTE_BPID_MASK 0x00ff0000 /**< SGTE BPID field mask */
-+#define DPAA_SGTE_OFFSET_MASK 0x00001fff /**< SGTE OFFSET field mask */
-+
-+#define DPAA_SGTE_GET_ADDRH(sgte) (((t_DpaaSGTE *)sgte)->addrh & DPAA_SGTE_ADDRH_MASK) /**< Macro to get SGTE ADDRH field */
-+#define DPAA_SGTE_GET_ADDRL(sgte) ((t_DpaaSGTE *)sgte)->addrl /**< Macro to get SGTE ADDRL field */
-+#define DPAA_SGTE_GET_PHYS_ADDR(sgte) ((physAddress_t)(((uint64_t)DPAA_SGTE_GET_ADDRH(sgte) << 32) | (uint64_t)DPAA_SGTE_GET_ADDRL(sgte))) /**< Macro to get FD ADDR field */
-+#define DPAA_SGTE_GET_EXTENSION(sgte) ((((t_DpaaSGTE *)sgte)->length & DPAA_SGTE_E_MASK) >> (31-0)) /**< Macro to get SGTE EXTENSION field */
-+#define DPAA_SGTE_GET_FINAL(sgte) ((((t_DpaaSGTE *)sgte)->length & DPAA_SGTE_F_MASK) >> (31-1)) /**< Macro to get SGTE FINAL field */
-+#define DPAA_SGTE_GET_LENGTH(sgte) (((t_DpaaSGTE *)sgte)->length & DPAA_SGTE_LENGTH_MASK) /**< Macro to get SGTE LENGTH field */
-+#define DPAA_SGTE_GET_BPID(sgte) ((((t_DpaaSGTE *)sgte)->offset & DPAA_SGTE_BPID_MASK) >> (31-15)) /**< Macro to get SGTE BPID field */
-+#define DPAA_SGTE_GET_OFFSET(sgte) (((t_DpaaSGTE *)sgte)->offset & DPAA_SGTE_OFFSET_MASK) /**< Macro to get SGTE OFFSET field */
-+#define DPAA_SGTE_GET_ADDR(sgte) XX_PhysToVirt(DPAA_SGTE_GET_PHYS_ADDR(sgte))
-+
-+#define DPAA_SGTE_SET_ADDRH(sgte,val) (((t_DpaaSGTE *)sgte)->addrh = ((((t_DpaaSGTE *)sgte)->addrh & ~DPAA_SGTE_ADDRH_MASK) | ((val) & DPAA_SGTE_ADDRH_MASK))) /**< Macro to set SGTE ADDRH field */
-+#define DPAA_SGTE_SET_ADDRL(sgte,val) ((t_DpaaSGTE *)sgte)->addrl = (val) /**< Macro to set SGTE ADDRL field */
-+#define DPAA_SGTE_SET_ADDR(sgte,val) \
-+do { \
-+ uint64_t physAddr = (uint64_t)(XX_VirtToPhys(val)); \
-+ DPAA_SGTE_SET_ADDRH(sgte, ((uint32_t)(physAddr >> 32))); \
-+ DPAA_SGTE_SET_ADDRL(sgte, (uint32_t)physAddr); \
-+} while (0) /**< Macro to set SGTE ADDR field */
-+#define DPAA_SGTE_SET_EXTENSION(sgte,val) (((t_DpaaSGTE *)sgte)->length = ((((t_DpaaSGTE *)sgte)->length & ~DPAA_SGTE_E_MASK) | (((val) << (31-0))& DPAA_SGTE_E_MASK))) /**< Macro to set SGTE EXTENSION field */
-+#define DPAA_SGTE_SET_FINAL(sgte,val) (((t_DpaaSGTE *)sgte)->length = ((((t_DpaaSGTE *)sgte)->length & ~DPAA_SGTE_F_MASK) | (((val) << (31-1))& DPAA_SGTE_F_MASK))) /**< Macro to set SGTE FINAL field */
-+#define DPAA_SGTE_SET_LENGTH(sgte,val) (((t_DpaaSGTE *)sgte)->length = (((t_DpaaSGTE *)sgte)->length & ~DPAA_SGTE_LENGTH_MASK) | ((val) & DPAA_SGTE_LENGTH_MASK)) /**< Macro to set SGTE LENGTH field */
-+#define DPAA_SGTE_SET_BPID(sgte,val) (((t_DpaaSGTE *)sgte)->offset = ((((t_DpaaSGTE *)sgte)->offset & ~DPAA_SGTE_BPID_MASK) | (((val) << (31-15))& DPAA_SGTE_BPID_MASK))) /**< Macro to set SGTE BPID field */
-+#define DPAA_SGTE_SET_OFFSET(sgte,val) (((t_DpaaSGTE *)sgte)->offset = ((((t_DpaaSGTE *)sgte)->offset & ~DPAA_SGTE_OFFSET_MASK) | (((val) << (31-31))& DPAA_SGTE_OFFSET_MASK) )) /**< Macro to set SGTE OFFSET field */
-+/* @} */
-+
-+#if defined(__MWERKS__) && !defined(__GNUC__)
-+#pragma pack(pop)
-+#endif /* defined(__MWERKS__) && ... */
-+
-+#define DPAA_LIODN_DONT_OVERRIDE (-1)
-+
-+/** @} */ /* end of DPAA_grp group */
-+
-+
-+#endif /* __DPAA_EXT_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/Peripherals/fm_ext.h
-@@ -0,0 +1,1731 @@
-+/* Copyright (c) 2008-2012 Freescale Semiconductor, Inc
-+ * All rights reserved.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/**************************************************************************//**
-+ @File fm_ext.h
-+
-+ @Description FM Application Programming Interface.
-+*//***************************************************************************/
-+#ifndef __FM_EXT
-+#define __FM_EXT
-+
-+#include "error_ext.h"
-+#include "std_ext.h"
-+#include "dpaa_ext.h"
-+#include "fsl_fman_sp.h"
-+
-+/**************************************************************************//**
-+ @Group FM_grp Frame Manager API
-+
-+ @Description FM API functions, definitions and enums.
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Group FM_lib_grp FM library
-+
-+ @Description FM API functions, definitions and enums.
-+
-+ The FM module is the main driver module and is a mandatory module
-+ for FM driver users. This module must be initialized first prior
-+ to any other drivers modules.
-+ The FM is a "singleton" module. It is responsible of the common
-+ HW modules: FPM, DMA, common QMI and common BMI initializations and
-+ run-time control routines. This module must be initialized always
-+ when working with any of the FM modules.
-+ NOTE - We assume that the FM library will be initialized only by core No. 0!
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Description Enum for defining port types
-+*//***************************************************************************/
-+typedef enum e_FmPortType {
-+ e_FM_PORT_TYPE_OH_OFFLINE_PARSING = 0, /**< Offline parsing port */
-+ e_FM_PORT_TYPE_RX, /**< 1G Rx port */
-+ e_FM_PORT_TYPE_RX_10G, /**< 10G Rx port */
-+ e_FM_PORT_TYPE_TX, /**< 1G Tx port */
-+ e_FM_PORT_TYPE_TX_10G, /**< 10G Tx port */
-+ e_FM_PORT_TYPE_DUMMY
-+} e_FmPortType;
-+
-+/**************************************************************************//**
-+ @Collection General FM defines
-+*//***************************************************************************/
-+#define FM_MAX_NUM_OF_PARTITIONS 64 /**< Maximum number of partitions */
-+#define FM_PHYS_ADDRESS_SIZE 6 /**< FM Physical address size */
-+/* @} */
-+
-+
-+#if defined(__MWERKS__) && !defined(__GNUC__)
-+#pragma pack(push,1)
-+#endif /* defined(__MWERKS__) && ... */
-+
-+/**************************************************************************//**
-+ @Description FM physical Address
-+*//***************************************************************************/
-+typedef _Packed struct t_FmPhysAddr {
-+ volatile uint8_t high; /**< High part of the physical address */
-+ volatile uint32_t low; /**< Low part of the physical address */
-+} _PackedType t_FmPhysAddr;
-+
-+/**************************************************************************//**
-+ @Description Parse results memory layout
-+*//***************************************************************************/
-+typedef _Packed struct t_FmPrsResult {
-+ volatile uint8_t lpid; /**< Logical port id */
-+ volatile uint8_t shimr; /**< Shim header result */
-+ volatile uint16_t l2r; /**< Layer 2 result */
-+ volatile uint16_t l3r; /**< Layer 3 result */
-+ volatile uint8_t l4r; /**< Layer 4 result */
-+ volatile uint8_t cplan; /**< Classification plan id */
-+ volatile uint16_t nxthdr; /**< Next Header */
-+ volatile uint16_t cksum; /**< Running-sum */
-+ volatile uint16_t flags_frag_off; /**< Flags & fragment-offset field of the last IP-header */
-+ volatile uint8_t route_type; /**< Routing type field of a IPv6 routing extension header */
-+ volatile uint8_t rhp_ip_valid; /**< Routing Extension Header Present; last bit is IP valid */
-+ volatile uint8_t shim_off[2]; /**< Shim offset */
-+ volatile uint8_t ip_pid_off; /**< IP PID (last IP-proto) offset */
-+ volatile uint8_t eth_off; /**< ETH offset */
-+ volatile uint8_t llc_snap_off; /**< LLC_SNAP offset */
-+ volatile uint8_t vlan_off[2]; /**< VLAN offset */
-+ volatile uint8_t etype_off; /**< ETYPE offset */
-+ volatile uint8_t pppoe_off; /**< PPP offset */
-+ volatile uint8_t mpls_off[2]; /**< MPLS offset */
-+ volatile uint8_t ip_off[2]; /**< IP offset */
-+ volatile uint8_t gre_off; /**< GRE offset */
-+ volatile uint8_t l4_off; /**< Layer 4 offset */
-+ volatile uint8_t nxthdr_off; /**< Parser end point */
-+} _PackedType t_FmPrsResult;
-+
-+/**************************************************************************//**
-+ @Collection FM Parser results
-+*//***************************************************************************/
-+#define FM_PR_L2_VLAN_STACK 0x00000100 /**< Parse Result: VLAN stack */
-+#define FM_PR_L2_ETHERNET 0x00008000 /**< Parse Result: Ethernet*/
-+#define FM_PR_L2_VLAN 0x00004000 /**< Parse Result: VLAN */
-+#define FM_PR_L2_LLC_SNAP 0x00002000 /**< Parse Result: LLC_SNAP */
-+#define FM_PR_L2_MPLS 0x00001000 /**< Parse Result: MPLS */
-+#define FM_PR_L2_PPPoE 0x00000800 /**< Parse Result: PPPoE */
-+/* @} */
-+
-+/**************************************************************************//**
-+ @Collection FM Frame descriptor macros
-+*//***************************************************************************/
-+#define FM_FD_CMD_FCO 0x80000000 /**< Frame queue Context Override */
-+#define FM_FD_CMD_RPD 0x40000000 /**< Read Prepended Data */
-+#define FM_FD_CMD_UPD 0x20000000 /**< Update Prepended Data */
-+#define FM_FD_CMD_DTC 0x10000000 /**< Do L4 Checksum */
-+#define FM_FD_CMD_DCL4C 0x10000000 /**< Didn't calculate L4 Checksum */
-+#define FM_FD_CMD_CFQ 0x00ffffff /**< Confirmation Frame Queue */
-+
-+#define FM_FD_ERR_UNSUPPORTED_FORMAT 0x04000000 /**< Not for Rx-Port! Unsupported Format */
-+#define FM_FD_ERR_LENGTH 0x02000000 /**< Not for Rx-Port! Length Error */
-+#define FM_FD_ERR_DMA 0x01000000 /**< DMA Data error */
-+
-+#define FM_FD_IPR 0x00000001 /**< IPR frame (not error) */
-+
-+#define FM_FD_ERR_IPR_NCSP (0x00100000 | FM_FD_IPR) /**< IPR non-consistent-sp */
-+#define FM_FD_ERR_IPR (0x00200000 | FM_FD_IPR) /**< IPR error */
-+#define FM_FD_ERR_IPR_TO (0x00300000 | FM_FD_IPR) /**< IPR timeout */
-+
-+#ifdef FM_CAPWAP_SUPPORT
-+#define FM_FD_ERR_CRE 0x00200000
-+#define FM_FD_ERR_CHE 0x00100000
-+#endif /* FM_CAPWAP_SUPPORT */
-+
-+#define FM_FD_ERR_PHYSICAL 0x00080000 /**< Rx FIFO overflow, FCS error, code error, running disparity
-+ error (SGMII and TBI modes), FIFO parity error. PHY
-+ Sequence error, PHY error control character detected. */
-+#define FM_FD_ERR_SIZE 0x00040000 /**< Frame too long OR Frame size exceeds max_length_frame */
-+#define FM_FD_ERR_CLS_DISCARD 0x00020000 /**< classification discard */
-+#define FM_FD_ERR_EXTRACTION 0x00008000 /**< Extract Out of Frame */
-+#define FM_FD_ERR_NO_SCHEME 0x00004000 /**< No Scheme Selected */
-+#define FM_FD_ERR_KEYSIZE_OVERFLOW 0x00002000 /**< Keysize Overflow */
-+#define FM_FD_ERR_COLOR_RED 0x00000800 /**< Frame color is red */
-+#define FM_FD_ERR_COLOR_YELLOW 0x00000400 /**< Frame color is yellow */
-+#define FM_FD_ERR_ILL_PLCR 0x00000200 /**< Illegal Policer Profile selected */
-+#define FM_FD_ERR_PLCR_FRAME_LEN 0x00000100 /**< Policer frame length error */
-+#define FM_FD_ERR_PRS_TIMEOUT 0x00000080 /**< Parser Time out Exceed */
-+#define FM_FD_ERR_PRS_ILL_INSTRUCT 0x00000040 /**< Invalid Soft Parser instruction */
-+#define FM_FD_ERR_PRS_HDR_ERR 0x00000020 /**< Header error was identified during parsing */
-+#define FM_FD_ERR_BLOCK_LIMIT_EXCEEDED 0x00000008 /**< Frame parsed beyind 256 first bytes */
-+
-+#define FM_FD_TX_STATUS_ERR_MASK (FM_FD_ERR_UNSUPPORTED_FORMAT | \
-+ FM_FD_ERR_LENGTH | \
-+ FM_FD_ERR_DMA) /**< TX Error FD bits */
-+
-+#define FM_FD_RX_STATUS_ERR_MASK (FM_FD_ERR_UNSUPPORTED_FORMAT | \
-+ FM_FD_ERR_LENGTH | \
-+ FM_FD_ERR_DMA | \
-+ FM_FD_ERR_IPR | \
-+ FM_FD_ERR_IPR_TO | \
-+ FM_FD_ERR_IPR_NCSP | \
-+ FM_FD_ERR_PHYSICAL | \
-+ FM_FD_ERR_SIZE | \
-+ FM_FD_ERR_CLS_DISCARD | \
-+ FM_FD_ERR_COLOR_RED | \
-+ FM_FD_ERR_COLOR_YELLOW | \
-+ FM_FD_ERR_ILL_PLCR | \
-+ FM_FD_ERR_PLCR_FRAME_LEN | \
-+ FM_FD_ERR_EXTRACTION | \
-+ FM_FD_ERR_NO_SCHEME | \
-+ FM_FD_ERR_KEYSIZE_OVERFLOW | \
-+ FM_FD_ERR_PRS_TIMEOUT | \
-+ FM_FD_ERR_PRS_ILL_INSTRUCT | \
-+ FM_FD_ERR_PRS_HDR_ERR | \
-+ FM_FD_ERR_BLOCK_LIMIT_EXCEEDED) /**< RX Error FD bits */
-+
-+#define FM_FD_RX_STATUS_ERR_NON_FM 0x00400000 /**< non Frame-Manager error */
-+/* @} */
-+
-+/**************************************************************************//**
-+ @Description Context A
-+*//***************************************************************************/
-+typedef _Packed struct t_FmContextA {
-+ volatile uint32_t command; /**< ContextA Command */
-+ volatile uint8_t res0[4]; /**< ContextA Reserved bits */
-+} _PackedType t_FmContextA;
-+
-+/**************************************************************************//**
-+ @Description Context B
-+*//***************************************************************************/
-+typedef uint32_t t_FmContextB;
-+
-+/**************************************************************************//**
-+ @Collection Special Operation options
-+*//***************************************************************************/
-+typedef uint32_t fmSpecialOperations_t; /**< typedef for defining Special Operation options */
-+
-+#define FM_SP_OP_IPSEC 0x80000000 /**< activate features that related to IPSec (e.g fix Eth-type) */
-+#define FM_SP_OP_IPSEC_UPDATE_UDP_LEN 0x40000000 /**< update the UDP-Len after Encryption */
-+#define FM_SP_OP_IPSEC_MANIP 0x20000000 /**< handle the IPSec-manip options */
-+#define FM_SP_OP_RPD 0x10000000 /**< Set the RPD bit */
-+#define FM_SP_OP_DCL4C 0x08000000 /**< Set the DCL4C bit */
-+#define FM_SP_OP_CHECK_SEC_ERRORS 0x04000000 /**< Check SEC errors */
-+#define FM_SP_OP_CLEAR_RPD 0x02000000 /**< Clear the RPD bit */
-+#define FM_SP_OP_CAPWAP_DTLS_ENC 0x01000000 /**< activate features that related to CAPWAP-DTLS post Encryption */
-+#define FM_SP_OP_CAPWAP_DTLS_DEC 0x00800000 /**< activate features that related to CAPWAP-DTLS post Decryption */
-+#define FM_SP_OP_IPSEC_NO_ETH_HDR 0x00400000 /**< activate features that related to IPSec without Eth hdr */
-+/* @} */
-+
-+/**************************************************************************//**
-+ @Collection Context A macros
-+*//***************************************************************************/
-+#define FM_CONTEXTA_OVERRIDE_MASK 0x80000000
-+#define FM_CONTEXTA_ICMD_MASK 0x40000000
-+#define FM_CONTEXTA_A1_VALID_MASK 0x20000000
-+#define FM_CONTEXTA_MACCMD_MASK 0x00ff0000
-+#define FM_CONTEXTA_MACCMD_VALID_MASK 0x00800000
-+#define FM_CONTEXTA_MACCMD_SECURED_MASK 0x00100000
-+#define FM_CONTEXTA_MACCMD_SC_MASK 0x000f0000
-+#define FM_CONTEXTA_A1_MASK 0x0000ffff
-+
-+#define FM_CONTEXTA_GET_OVERRIDE(contextA) ((((t_FmContextA *)contextA)->command & FM_CONTEXTA_OVERRIDE_MASK) >> (31-0))
-+#define FM_CONTEXTA_GET_ICMD(contextA) ((((t_FmContextA *)contextA)->command & FM_CONTEXTA_ICMD_MASK) >> (31-1))
-+#define FM_CONTEXTA_GET_A1_VALID(contextA) ((((t_FmContextA *)contextA)->command & FM_CONTEXTA_A1_VALID_MASK) >> (31-2))
-+#define FM_CONTEXTA_GET_A1(contextA) ((((t_FmContextA *)contextA)->command & FM_CONTEXTA_A1_MASK) >> (31-31))
-+#define FM_CONTEXTA_GET_MACCMD(contextA) ((((t_FmContextA *)contextA)->command & FM_CONTEXTA_MACCMD_MASK) >> (31-15))
-+#define FM_CONTEXTA_GET_MACCMD_VALID(contextA) ((((t_FmContextA *)contextA)->command & FM_CONTEXTA_MACCMD_VALID_MASK) >> (31-8))
-+#define FM_CONTEXTA_GET_MACCMD_SECURED(contextA) ((((t_FmContextA *)contextA)->command & FM_CONTEXTA_MACCMD_SECURED_MASK) >> (31-11))
-+#define FM_CONTEXTA_GET_MACCMD_SECURE_CHANNEL(contextA) ((((t_FmContextA *)contextA)->command & FM_CONTEXTA_MACCMD_SC_MASK) >> (31-15))
-+
-+#define FM_CONTEXTA_SET_OVERRIDE(contextA,val) (((t_FmContextA *)contextA)->command = (uint32_t)((((t_FmContextA *)contextA)->command & ~FM_CONTEXTA_OVERRIDE_MASK) | (((uint32_t)(val) << (31-0)) & FM_CONTEXTA_OVERRIDE_MASK) ))
-+#define FM_CONTEXTA_SET_ICMD(contextA,val) (((t_FmContextA *)contextA)->command = (uint32_t)((((t_FmContextA *)contextA)->command & ~FM_CONTEXTA_ICMD_MASK) | (((val) << (31-1)) & FM_CONTEXTA_ICMD_MASK) ))
-+#define FM_CONTEXTA_SET_A1_VALID(contextA,val) (((t_FmContextA *)contextA)->command = (uint32_t)((((t_FmContextA *)contextA)->command & ~FM_CONTEXTA_A1_VALID_MASK) | (((val) << (31-2)) & FM_CONTEXTA_A1_VALID_MASK) ))
-+#define FM_CONTEXTA_SET_A1(contextA,val) (((t_FmContextA *)contextA)->command = (uint32_t)((((t_FmContextA *)contextA)->command & ~FM_CONTEXTA_A1_MASK) | (((val) << (31-31)) & FM_CONTEXTA_A1_MASK) ))
-+#define FM_CONTEXTA_SET_MACCMD(contextA,val) (((t_FmContextA *)contextA)->command = (uint32_t)((((t_FmContextA *)contextA)->command & ~FM_CONTEXTA_MACCMD_MASK) | (((val) << (31-15)) & FM_CONTEXTA_MACCMD_MASK) ))
-+#define FM_CONTEXTA_SET_MACCMD_VALID(contextA,val) (((t_FmContextA *)contextA)->command = (uint32_t)((((t_FmContextA *)contextA)->command & ~FM_CONTEXTA_MACCMD_VALID_MASK) | (((val) << (31-8)) & FM_CONTEXTA_MACCMD_VALID_MASK) ))
-+#define FM_CONTEXTA_SET_MACCMD_SECURED(contextA,val) (((t_FmContextA *)contextA)->command = (uint32_t)((((t_FmContextA *)contextA)->command & ~FM_CONTEXTA_MACCMD_SECURED_MASK) | (((val) << (31-11)) & FM_CONTEXTA_MACCMD_SECURED_MASK) ))
-+#define FM_CONTEXTA_SET_MACCMD_SECURE_CHANNEL(contextA,val) (((t_FmContextA *)contextA)->command = (uint32_t)((((t_FmContextA *)contextA)->command & ~FM_CONTEXTA_MACCMD_SC_MASK) | (((val) << (31-15)) & FM_CONTEXTA_MACCMD_SC_MASK) ))
-+/* @} */
-+
-+/**************************************************************************//**
-+ @Collection Context B macros
-+*//***************************************************************************/
-+#define FM_CONTEXTB_FQID_MASK 0x00ffffff
-+
-+#define FM_CONTEXTB_GET_FQID(contextB) (*((t_FmContextB *)contextB) & FM_CONTEXTB_FQID_MASK)
-+#define FM_CONTEXTB_SET_FQID(contextB,val) (*((t_FmContextB *)contextB) = ((*((t_FmContextB *)contextB) & ~FM_CONTEXTB_FQID_MASK) | ((val) & FM_CONTEXTB_FQID_MASK)))
-+/* @} */
-+
-+#if defined(__MWERKS__) && !defined(__GNUC__)
-+#pragma pack(pop)
-+#endif /* defined(__MWERKS__) && ... */
-+
-+
-+/**************************************************************************//**
-+ @Description FM Exceptions
-+*//***************************************************************************/
-+typedef enum e_FmExceptions {
-+ e_FM_EX_DMA_BUS_ERROR = 0, /**< DMA bus error. */
-+ e_FM_EX_DMA_READ_ECC, /**< Read Buffer ECC error (Valid for FM rev < 6)*/
-+ e_FM_EX_DMA_SYSTEM_WRITE_ECC, /**< Write Buffer ECC error on system side (Valid for FM rev < 6)*/
-+ e_FM_EX_DMA_FM_WRITE_ECC, /**< Write Buffer ECC error on FM side (Valid for FM rev < 6)*/
-+ e_FM_EX_DMA_SINGLE_PORT_ECC, /**< Single Port ECC error on FM side (Valid for FM rev > 6)*/
-+ e_FM_EX_FPM_STALL_ON_TASKS, /**< Stall of tasks on FPM */
-+ e_FM_EX_FPM_SINGLE_ECC, /**< Single ECC on FPM. */
-+ e_FM_EX_FPM_DOUBLE_ECC, /**< Double ECC error on FPM ram access */
-+ e_FM_EX_QMI_SINGLE_ECC, /**< Single ECC on QMI. */
-+ e_FM_EX_QMI_DOUBLE_ECC, /**< Double bit ECC occurred on QMI */
-+ e_FM_EX_QMI_DEQ_FROM_UNKNOWN_PORTID,/**< Dequeue from unknown port id */
-+ e_FM_EX_BMI_LIST_RAM_ECC, /**< Linked List RAM ECC error */
-+ e_FM_EX_BMI_STORAGE_PROFILE_ECC, /**< Storage Profile ECC Error */
-+ e_FM_EX_BMI_STATISTICS_RAM_ECC, /**< Statistics Count RAM ECC Error Enable */
-+ e_FM_EX_BMI_DISPATCH_RAM_ECC, /**< Dispatch RAM ECC Error Enable */
-+ e_FM_EX_IRAM_ECC, /**< Double bit ECC occurred on IRAM*/
-+ e_FM_EX_MURAM_ECC /**< Double bit ECC occurred on MURAM*/
-+} e_FmExceptions;
-+
-+/**************************************************************************//**
-+ @Description Enum for defining port DMA swap mode
-+*//***************************************************************************/
-+typedef enum e_FmDmaSwapOption {
-+ e_FM_DMA_NO_SWP = FMAN_DMA_NO_SWP, /**< No swap, transfer data as is.*/
-+ e_FM_DMA_SWP_PPC_LE = FMAN_DMA_SWP_PPC_LE, /**< The transferred data should be swapped
-+ in PowerPc Little Endian mode. */
-+ e_FM_DMA_SWP_BE = FMAN_DMA_SWP_BE /**< The transferred data should be swapped
-+ in Big Endian mode */
-+} e_FmDmaSwapOption;
-+
-+/**************************************************************************//**
-+ @Description Enum for defining port DMA cache attributes
-+*//***************************************************************************/
-+typedef enum e_FmDmaCacheOption {
-+ e_FM_DMA_NO_STASH = FMAN_DMA_NO_STASH, /**< Cacheable, no Allocate (No Stashing) */
-+ e_FM_DMA_STASH = FMAN_DMA_STASH /**< Cacheable and Allocate (Stashing on) */
-+} e_FmDmaCacheOption;
-+
-+
-+/**************************************************************************//**
-+ @Group FM_init_grp FM Initialization Unit
-+
-+ @Description FM Initialization Unit
-+
-+ Initialization Flow
-+ Initialization of the FM Module will be carried out by the application
-+ according to the following sequence:
-+ - Calling the configuration routine with basic parameters.
-+ - Calling the advance initialization routines to change driver's defaults.
-+ - Calling the initialization routine.
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Function t_FmExceptionsCallback
-+
-+ @Description Exceptions user callback routine, will be called upon an
-+ exception passing the exception identification.
-+
-+ @Param[in] h_App - User's application descriptor.
-+ @Param[in] exception - The exception.
-+*//***************************************************************************/
-+typedef void (t_FmExceptionsCallback)(t_Handle h_App,
-+ e_FmExceptions exception);
-+
-+
-+/**************************************************************************//**
-+ @Function t_FmBusErrorCallback
-+
-+ @Description Bus error user callback routine, will be called upon a
-+ bus error, passing parameters describing the errors and the owner.
-+
-+ @Param[in] h_App - User's application descriptor.
-+ @Param[in] portType - Port type (e_FmPortType)
-+ @Param[in] portId - Port id - relative to type.
-+ @Param[in] addr - Address that caused the error
-+ @Param[in] tnum - Owner of error
-+ @Param[in] liodn - Logical IO device number
-+*//***************************************************************************/
-+typedef void (t_FmBusErrorCallback) (t_Handle h_App,
-+ e_FmPortType portType,
-+ uint8_t portId,
-+ uint64_t addr,
-+ uint8_t tnum,
-+ uint16_t liodn);
-+
-+/**************************************************************************//**
-+ @Description A structure for defining buffer prefix area content.
-+*//***************************************************************************/
-+typedef struct t_FmBufferPrefixContent {
-+ uint16_t privDataSize; /**< Number of bytes to be left at the beginning
-+ of the external buffer; Note that the private-area will
-+ start from the base of the buffer address. */
-+ bool passPrsResult; /**< TRUE to pass the parse result to/from the FM;
-+ User may use FM_PORT_GetBufferPrsResult() in order to
-+ get the parser-result from a buffer. */
-+ bool passTimeStamp; /**< TRUE to pass the timeStamp to/from the FM
-+ User may use FM_PORT_GetBufferTimeStamp() in order to
-+ get the parser-result from a buffer. */
-+ bool passHashResult; /**< TRUE to pass the KG hash result to/from the FM
-+ User may use FM_PORT_GetBufferHashResult() in order to
-+ get the parser-result from a buffer. */
-+ bool passAllOtherPCDInfo;/**< Add all other Internal-Context information:
-+ AD, hash-result, key, etc. */
-+ uint16_t dataAlign; /**< 0 to use driver's default alignment [DEFAULT_FM_SP_bufferPrefixContent_dataAlign],
-+ other value for selecting a data alignment (must be a power of 2);
-+ if write optimization is used, must be >= 16. */
-+ uint8_t manipExtraSpace; /**< Maximum extra size needed (insertion-size minus removal-size);
-+ Note that this field impacts the size of the buffer-prefix
-+ (i.e. it pushes the data offset);
-+ This field is irrelevant if DPAA_VERSION==10 */
-+} t_FmBufferPrefixContent;
-+
-+/**************************************************************************//**
-+ @Description A structure of information about each of the external
-+ buffer pools used by a port or storage-profile.
-+*//***************************************************************************/
-+typedef struct t_FmExtPoolParams {
-+ uint8_t id; /**< External buffer pool id */
-+ uint16_t size; /**< External buffer pool buffer size */
-+} t_FmExtPoolParams;
-+
-+/**************************************************************************//**
-+ @Description A structure for informing the driver about the external
-+ buffer pools allocated in the BM and used by a port or a
-+ storage-profile.
-+*//***************************************************************************/
-+typedef struct t_FmExtPools {
-+ uint8_t numOfPoolsUsed; /**< Number of pools use by this port */
-+ t_FmExtPoolParams extBufPool[FM_PORT_MAX_NUM_OF_EXT_POOLS];
-+ /**< Parameters for each port */
-+} t_FmExtPools;
-+
-+/**************************************************************************//**
-+ @Description A structure for defining backup BM Pools.
-+*//***************************************************************************/
-+typedef struct t_FmBackupBmPools {
-+ uint8_t numOfBackupPools; /**< Number of BM backup pools -
-+ must be smaller than the total number of
-+ pools defined for the specified port.*/
-+ uint8_t poolIds[FM_PORT_MAX_NUM_OF_EXT_POOLS];
-+ /**< numOfBackupPools pool id's, specifying which
-+ pools should be used only as backup. Pool
-+ id's specified here must be a subset of the
-+ pools used by the specified port.*/
-+} t_FmBackupBmPools;
-+
-+/**************************************************************************//**
-+ @Description A structure for defining BM pool depletion criteria
-+*//***************************************************************************/
-+typedef struct t_FmBufPoolDepletion {
-+ bool poolsGrpModeEnable; /**< select mode in which pause frames will be sent after
-+ a number of pools (all together!) are depleted */
-+ uint8_t numOfPools; /**< the number of depleted pools that will invoke
-+ pause frames transmission. */
-+ bool poolsToConsider[BM_MAX_NUM_OF_POOLS];
-+ /**< For each pool, TRUE if it should be considered for
-+ depletion (Note - this pool must be used by this port!). */
-+ bool singlePoolModeEnable; /**< select mode in which pause frames will be sent after
-+ a single-pool is depleted; */
-+ bool poolsToConsiderForSingleMode[BM_MAX_NUM_OF_POOLS];
-+ /**< For each pool, TRUE if it should be considered for
-+ depletion (Note - this pool must be used by this port!) */
-+#if (DPAA_VERSION >= 11)
-+ bool pfcPrioritiesEn[FM_MAX_NUM_OF_PFC_PRIORITIES];
-+ /**< This field is used by the MAC as the Priority Enable Vector in the PFC frame which is transmitted */
-+#endif /* (DPAA_VERSION >= 11) */
-+} t_FmBufPoolDepletion;
-+
-+/**************************************************************************//**
-+ @Description A Structure for defining Ucode patch for loading.
-+*//***************************************************************************/
-+typedef struct t_FmFirmwareParams {
-+ uint32_t size; /**< Size of uCode */
-+ uint32_t *p_Code; /**< A pointer to the uCode */
-+} t_FmFirmwareParams;
-+
-+/**************************************************************************//**
-+ @Description A Structure for defining FM initialization parameters
-+*//***************************************************************************/
-+typedef struct t_FmParams {
-+ uint8_t fmId; /**< Index of the FM */
-+ uint8_t guestId; /**< FM Partition Id */
-+ uintptr_t baseAddr; /**< A pointer to base of memory mapped FM registers (virtual);
-+ this field is optional when the FM runs in "guest-mode"
-+ (i.e. guestId != NCSW_MASTER_ID); in that case, the driver will
-+ use the memory-map instead of calling the IPC where possible;
-+ NOTE that this should include ALL common registers of the FM including
-+ the PCD registers area (i.e. until the VSP pages - 880KB). */
-+ t_Handle h_FmMuram; /**< A handle of an initialized MURAM object,
-+ to be used by the FM. */
-+ uint16_t fmClkFreq; /**< In Mhz;
-+ Relevant when FM not runs in "guest-mode". */
-+ uint16_t fmMacClkRatio; /**< FM MAC Clock ratio, for backward comparability:
-+ when fmMacClkRatio = 0, ratio is 2:1
-+ when fmMacClkRatio = 1, ratio is 1:1 */
-+ t_FmExceptionsCallback *f_Exception; /**< An application callback routine to handle exceptions;
-+ Relevant when FM not runs in "guest-mode". */
-+ t_FmBusErrorCallback *f_BusError; /**< An application callback routine to handle exceptions;
-+ Relevant when FM not runs in "guest-mode". */
-+ t_Handle h_App; /**< A handle to an application layer object; This handle will
-+ be passed by the driver upon calling the above callbacks;
-+ Relevant when FM not runs in "guest-mode". */
-+ int irq; /**< FM interrupt source for normal events;
-+ Relevant when FM not runs in "guest-mode". */
-+ int errIrq; /**< FM interrupt source for errors;
-+ Relevant when FM not runs in "guest-mode". */
-+ t_FmFirmwareParams firmware; /**< The firmware parameters structure;
-+ Relevant when FM not runs in "guest-mode". */
-+
-+#if (DPAA_VERSION >= 11)
-+ uintptr_t vspBaseAddr; /**< A pointer to base of memory mapped FM VSP registers (virtual);
-+ i.e. up to 24KB, depending on the specific chip. */
-+ uint8_t partVSPBase; /**< The first Virtual-Storage-Profile-id dedicated to this partition.
-+ NOTE: this parameter relevant only when working with multiple partitions. */
-+ uint8_t partNumOfVSPs; /**< Number of VSPs dedicated to this partition.
-+ NOTE: this parameter relevant only when working with multiple partitions. */
-+#endif /* (DPAA_VERSION >= 11) */
-+} t_FmParams;
-+
-+
-+/**************************************************************************//**
-+ @Function FM_Config
-+
-+ @Description Creates the FM module and returns its handle (descriptor).
-+ This descriptor must be passed as first parameter to all other
-+ FM function calls.
-+
-+ No actual initialization or configuration of FM hardware is
-+ done by this routine. All FM parameters get default values that
-+ may be changed by calling one or more of the advance config routines.
-+
-+ @Param[in] p_FmParams - A pointer to a data structure of mandatory FM parameters
-+
-+ @Return A handle to the FM object, or NULL for Failure.
-+*//***************************************************************************/
-+t_Handle FM_Config(t_FmParams *p_FmParams);
-+
-+/**************************************************************************//**
-+ @Function FM_Init
-+
-+ @Description Initializes the FM module by defining the software structure
-+ and configuring the hardware registers.
-+
-+ @Param[in] h_Fm - FM module descriptor
-+
-+ @Return E_OK on success; Error code otherwise.
-+*//***************************************************************************/
-+t_Error FM_Init(t_Handle h_Fm);
-+
-+/**************************************************************************//**
-+ @Function FM_Free
-+
-+ @Description Frees all resources that were assigned to FM module.
-+
-+ Calling this routine invalidates the descriptor.
-+
-+ @Param[in] h_Fm - FM module descriptor
-+
-+ @Return E_OK on success; Error code otherwise.
-+*//***************************************************************************/
-+t_Error FM_Free(t_Handle h_Fm);
-+
-+
-+/**************************************************************************//**
-+ @Group FM_advanced_init_grp FM Advanced Configuration Unit
-+
-+ @Description Advanced configuration routines are optional routines that may
-+ be called in order to change the default driver settings.
-+
-+ Note: Advanced configuration routines are not available for guest partition.
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Description Enum for selecting DMA debug mode
-+*//***************************************************************************/
-+typedef enum e_FmDmaDbgCntMode {
-+ e_FM_DMA_DBG_NO_CNT = 0, /**< No counting */
-+ e_FM_DMA_DBG_CNT_DONE, /**< Count DONE commands */
-+ e_FM_DMA_DBG_CNT_COMM_Q_EM, /**< count command queue emergency signals */
-+ e_FM_DMA_DBG_CNT_INT_READ_EM, /**< Count Internal Read buffer emergency signal */
-+ e_FM_DMA_DBG_CNT_INT_WRITE_EM, /**< Count Internal Write buffer emergency signal */
-+ e_FM_DMA_DBG_CNT_FPM_WAIT, /**< Count FPM WAIT signal */
-+ e_FM_DMA_DBG_CNT_SIGLE_BIT_ECC, /**< Single bit ECC errors. */
-+ e_FM_DMA_DBG_CNT_RAW_WAR_PROT /**< Number of times there was a need for RAW & WAR protection. */
-+} e_FmDmaDbgCntMode;
-+
-+/**************************************************************************//**
-+ @Description Enum for selecting DMA Cache Override
-+*//***************************************************************************/
-+typedef enum e_FmDmaCacheOverride {
-+ e_FM_DMA_NO_CACHE_OR = 0, /**< No override of the Cache field */
-+ e_FM_DMA_NO_STASH_DATA, /**< Data should not be stashed in system level cache */
-+ e_FM_DMA_MAY_STASH_DATA, /**< Data may be stashed in system level cache */
-+ e_FM_DMA_STASH_DATA /**< Data should be stashed in system level cache */
-+} e_FmDmaCacheOverride;
-+
-+/**************************************************************************//**
-+ @Description Enum for selecting DMA External Bus Priority
-+*//***************************************************************************/
-+typedef enum e_FmDmaExtBusPri {
-+ e_FM_DMA_EXT_BUS_NORMAL = 0, /**< Normal priority */
-+ e_FM_DMA_EXT_BUS_EBS, /**< AXI extended bus service priority */
-+ e_FM_DMA_EXT_BUS_SOS, /**< AXI sos priority */
-+ e_FM_DMA_EXT_BUS_EBS_AND_SOS /**< AXI ebs + sos priority */
-+} e_FmDmaExtBusPri;
-+
-+/**************************************************************************//**
-+ @Description Enum for choosing the field that will be output on AID
-+*//***************************************************************************/
-+typedef enum e_FmDmaAidMode {
-+ e_FM_DMA_AID_OUT_PORT_ID = 0, /**< 4 LSB of PORT_ID */
-+ e_FM_DMA_AID_OUT_TNUM /**< 4 LSB of TNUM */
-+} e_FmDmaAidMode;
-+
-+/**************************************************************************//**
-+ @Description Enum for selecting FPM Catastrophic error behavior
-+*//***************************************************************************/
-+typedef enum e_FmCatastrophicErr {
-+ e_FM_CATASTROPHIC_ERR_STALL_PORT = 0, /**< Port_ID is stalled (only reset can release it) */
-+ e_FM_CATASTROPHIC_ERR_STALL_TASK /**< Only erroneous task is stalled */
-+} e_FmCatastrophicErr;
-+
-+/**************************************************************************//**
-+ @Description Enum for selecting FPM DMA Error behavior
-+*//***************************************************************************/
-+typedef enum e_FmDmaErr {
-+ e_FM_DMA_ERR_CATASTROPHIC = 0, /**< Dma error is treated as a catastrophic
-+ error (e_FmCatastrophicErr)*/
-+ e_FM_DMA_ERR_REPORT /**< Dma error is just reported */
-+} e_FmDmaErr;
-+
-+/**************************************************************************//**
-+ @Description Enum for selecting DMA Emergency level by BMI emergency signal
-+*//***************************************************************************/
-+typedef enum e_FmDmaEmergencyLevel {
-+ e_FM_DMA_EM_EBS = 0, /**< EBS emergency */
-+ e_FM_DMA_EM_SOS /**< SOS emergency */
-+} e_FmDmaEmergencyLevel;
-+
-+/**************************************************************************//**
-+ @Collection Enum for selecting DMA Emergency options
-+*//***************************************************************************/
-+typedef uint32_t fmEmergencyBus_t; /**< DMA emergency options */
-+
-+#define FM_DMA_MURAM_READ_EMERGENCY 0x00800000 /**< Enable emergency for MURAM1 */
-+#define FM_DMA_MURAM_WRITE_EMERGENCY 0x00400000 /**< Enable emergency for MURAM2 */
-+#define FM_DMA_EXT_BUS_EMERGENCY 0x00100000 /**< Enable emergency for external bus */
-+/* @} */
-+
-+/**************************************************************************//**
-+ @Description A structure for defining DMA emergency level
-+*//***************************************************************************/
-+typedef struct t_FmDmaEmergency {
-+ fmEmergencyBus_t emergencyBusSelect; /**< An OR of the busses where emergency
-+ should be enabled */
-+ e_FmDmaEmergencyLevel emergencyLevel; /**< EBS/SOS */
-+} t_FmDmaEmergency;
-+
-+/**************************************************************************//*
-+ @Description structure for defining FM threshold
-+*//***************************************************************************/
-+typedef struct t_FmThresholds {
-+ uint8_t dispLimit; /**< The number of times a frames may
-+ be passed in the FM before assumed to
-+ be looping. */
-+ uint8_t prsDispTh; /**< This is the number pf packets that may be
-+ queued in the parser dispatch queue*/
-+ uint8_t plcrDispTh; /**< This is the number pf packets that may be
-+ queued in the policer dispatch queue*/
-+ uint8_t kgDispTh; /**< This is the number pf packets that may be
-+ queued in the keygen dispatch queue*/
-+ uint8_t bmiDispTh; /**< This is the number pf packets that may be
-+ queued in the BMI dispatch queue*/
-+ uint8_t qmiEnqDispTh; /**< This is the number pf packets that may be
-+ queued in the QMI enqueue dispatch queue*/
-+ uint8_t qmiDeqDispTh; /**< This is the number pf packets that may be
-+ queued in the QMI dequeue dispatch queue*/
-+ uint8_t fmCtl1DispTh; /**< This is the number pf packets that may be
-+ queued in fmCtl1 dispatch queue*/
-+ uint8_t fmCtl2DispTh; /**< This is the number pf packets that may be
-+ queued in fmCtl2 dispatch queue*/
-+} t_FmThresholds;
-+
-+/**************************************************************************//*
-+ @Description structure for defining DMA thresholds
-+*//***************************************************************************/
-+typedef struct t_FmDmaThresholds {
-+ uint8_t assertEmergency; /**< When this value is reached,
-+ assert emergency (Threshold)*/
-+ uint8_t clearEmergency; /**< After emergency is asserted, it is held
-+ until this value is reached (Hystheresis) */
-+} t_FmDmaThresholds;
-+
-+/**************************************************************************//**
-+ @Function t_FmResetOnInitOverrideCallback
-+
-+ @Description FMan specific reset on init user callback routine,
-+ will be used to override the standard FMan reset on init procedure
-+
-+ @Param[in] h_Fm - FMan handler
-+*//***************************************************************************/
-+typedef void (t_FmResetOnInitOverrideCallback)(t_Handle h_Fm);
-+
-+/**************************************************************************//**
-+ @Function FM_ConfigResetOnInit
-+
-+ @Description Define whether to reset the FM before initialization.
-+ Change the default configuration [DEFAULT_resetOnInit].
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+ @Param[in] enable When TRUE, FM will be reset before any initialization.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_Config() and before FM_Init().
-+ This routine should NOT be called from guest-partition
-+ (i.e. guestId != NCSW_MASTER_ID)
-+*//***************************************************************************/
-+t_Error FM_ConfigResetOnInit(t_Handle h_Fm, bool enable);
-+
-+/**************************************************************************//**
-+ @Function FM_ConfigResetOnInitOverrideCallback
-+
-+ @Description Define a special reset of FM before initialization.
-+ Change the default configuration [DEFAULT_resetOnInitOverrideCallback].
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+ @Param[in] f_ResetOnInitOverride FM specific reset on init user callback routine.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_Config() and before FM_Init().
-+ This routine should NOT be called from guest-partition
-+ (i.e. guestId != NCSW_MASTER_ID)
-+*//***************************************************************************/
-+t_Error FM_ConfigResetOnInitOverrideCallback(t_Handle h_Fm, t_FmResetOnInitOverrideCallback *f_ResetOnInitOverride);
-+
-+/**************************************************************************//**
-+ @Function FM_ConfigTotalFifoSize
-+
-+ @Description Define Total FIFO size for the whole FM.
-+ Calling this routine changes the total Fifo size in the internal driver
-+ data base from its default configuration [DEFAULT_totalFifoSize]
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+ @Param[in] totalFifoSize The selected new value.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_Config() and before FM_Init().
-+ This routine should NOT be called from guest-partition
-+ (i.e. guestId != NCSW_MASTER_ID)
-+*//***************************************************************************/
-+t_Error FM_ConfigTotalFifoSize(t_Handle h_Fm, uint32_t totalFifoSize);
-+
-+ /**************************************************************************//**
-+ @Function FM_ConfigDmaCacheOverride
-+
-+ @Description Define cache override mode.
-+ Calling this routine changes the cache override mode
-+ in the internal driver data base from its default configuration [DEFAULT_cacheOverride]
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+ @Param[in] cacheOverride The selected new value.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_Config() and before FM_Init().
-+ This routine should NOT be called from guest-partition
-+ (i.e. guestId != NCSW_MASTER_ID)
-+*//***************************************************************************/
-+t_Error FM_ConfigDmaCacheOverride(t_Handle h_Fm, e_FmDmaCacheOverride cacheOverride);
-+
-+/**************************************************************************//**
-+ @Function FM_ConfigDmaAidOverride
-+
-+ @Description Define DMA AID override mode.
-+ Calling this routine changes the AID override mode
-+ in the internal driver data base from its default configuration [DEFAULT_aidOverride]
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+ @Param[in] aidOverride The selected new value.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_Config() and before FM_Init().
-+ This routine should NOT be called from guest-partition
-+ (i.e. guestId != NCSW_MASTER_ID)
-+*//***************************************************************************/
-+t_Error FM_ConfigDmaAidOverride(t_Handle h_Fm, bool aidOverride);
-+
-+/**************************************************************************//**
-+ @Function FM_ConfigDmaAidMode
-+
-+ @Description Define DMA AID mode.
-+ Calling this routine changes the AID mode in the internal
-+ driver data base from its default configuration [DEFAULT_aidMode]
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+ @Param[in] aidMode The selected new value.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_Config() and before FM_Init().
-+ This routine should NOT be called from guest-partition
-+ (i.e. guestId != NCSW_MASTER_ID)
-+*//***************************************************************************/
-+t_Error FM_ConfigDmaAidMode(t_Handle h_Fm, e_FmDmaAidMode aidMode);
-+
-+/**************************************************************************//**
-+ @Function FM_ConfigDmaAxiDbgNumOfBeats
-+
-+ @Description Define DMA AXI number of beats.
-+ Calling this routine changes the AXI number of beats in the internal
-+ driver data base from its default configuration [DEFAULT_axiDbgNumOfBeats]
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+ @Param[in] axiDbgNumOfBeats The selected new value.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_Config() and before FM_Init().
-+ This routine should NOT be called from guest-partition
-+ (i.e. guestId != NCSW_MASTER_ID)
-+*//***************************************************************************/
-+t_Error FM_ConfigDmaAxiDbgNumOfBeats(t_Handle h_Fm, uint8_t axiDbgNumOfBeats);
-+
-+/**************************************************************************//**
-+ @Function FM_ConfigDmaCamNumOfEntries
-+
-+ @Description Define number of CAM entries.
-+ Calling this routine changes the number of CAM entries in the internal
-+ driver data base from its default configuration [DEFAULT_dmaCamNumOfEntries].
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+ @Param[in] numOfEntries The selected new value.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_Config() and before FM_Init().
-+ This routine should NOT be called from guest-partition
-+ (i.e. guestId != NCSW_MASTER_ID)
-+*//***************************************************************************/
-+t_Error FM_ConfigDmaCamNumOfEntries(t_Handle h_Fm, uint8_t numOfEntries);
-+
-+/**************************************************************************//**
-+ @Function FM_ConfigEnableCounters
-+
-+ @Description Obsolete, always return E_OK.
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+
-+ @Return E_OK on success; Error code otherwise.
-+*//***************************************************************************/
-+t_Error FM_ConfigEnableCounters(t_Handle h_Fm);
-+
-+/**************************************************************************//**
-+ @Function FM_ConfigDmaDbgCounter
-+
-+ @Description Define DMA debug counter.
-+ Calling this routine changes the number of the DMA debug counter in the internal
-+ driver data base from its default configuration [DEFAULT_dmaDbgCntMode].
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+ @Param[in] fmDmaDbgCntMode An enum selecting the debug counter mode.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_Config() and before FM_Init().
-+ This routine should NOT be called from guest-partition
-+ (i.e. guestId != NCSW_MASTER_ID)
-+*//***************************************************************************/
-+t_Error FM_ConfigDmaDbgCounter(t_Handle h_Fm, e_FmDmaDbgCntMode fmDmaDbgCntMode);
-+
-+/**************************************************************************//**
-+ @Function FM_ConfigDmaStopOnBusErr
-+
-+ @Description Define bus error behavior.
-+ Calling this routine changes the bus error behavior definition
-+ in the internal driver data base from its default
-+ configuration [DEFAULT_dmaStopOnBusError].
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+ @Param[in] stop TRUE to stop on bus error, FALSE to continue.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_Config() and before FM_Init().
-+ Only if bus error is enabled.
-+ This routine should NOT be called from guest-partition
-+ (i.e. guestId != NCSW_MASTER_ID)
-+*//***************************************************************************/
-+t_Error FM_ConfigDmaStopOnBusErr(t_Handle h_Fm, bool stop);
-+
-+/**************************************************************************//**
-+ @Function FM_ConfigDmaEmergency
-+
-+ @Description Define DMA emergency.
-+ Calling this routine changes the DMA emergency definition
-+ in the internal driver data base from its default
-+ configuration where's it's disabled.
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+ @Param[in] p_Emergency An OR mask of all required options.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_Config() and before FM_Init().
-+ This routine should NOT be called from guest-partition
-+ (i.e. guestId != NCSW_MASTER_ID)
-+*//***************************************************************************/
-+t_Error FM_ConfigDmaEmergency(t_Handle h_Fm, t_FmDmaEmergency *p_Emergency);
-+
-+/**************************************************************************//**
-+ @Function FM_ConfigDmaErr
-+
-+ @Description DMA error treatment.
-+ Calling this routine changes the DMA error treatment
-+ in the internal driver data base from its default
-+ configuration [DEFAULT_dmaErr].
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+ @Param[in] dmaErr The selected new choice.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_Config() and before FM_Init().
-+ This routine should NOT be called from guest-partition
-+ (i.e. guestId != NCSW_MASTER_ID)
-+*//***************************************************************************/
-+t_Error FM_ConfigDmaErr(t_Handle h_Fm, e_FmDmaErr dmaErr);
-+
-+/**************************************************************************//**
-+ @Function FM_ConfigCatastrophicErr
-+
-+ @Description Define FM behavior on catastrophic error.
-+ Calling this routine changes the FM behavior on catastrophic
-+ error in the internal driver data base from its default
-+ [DEFAULT_catastrophicErr].
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+ @Param[in] catastrophicErr The selected new choice.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_Config() and before FM_Init().
-+ This routine should NOT be called from guest-partition
-+ (i.e. guestId != NCSW_MASTER_ID)
-+*//***************************************************************************/
-+t_Error FM_ConfigCatastrophicErr(t_Handle h_Fm, e_FmCatastrophicErr catastrophicErr);
-+
-+/**************************************************************************//**
-+ @Function FM_ConfigEnableMuramTestMode
-+
-+ @Description Enable MURAM test mode.
-+ Calling this routine changes the internal driver data base
-+ from its default selection of test mode where it's disabled.
-+ This routine is only avaiable on old FM revisions (FMan v2).
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_Config() and before FM_Init().
-+ This routine should NOT be called from guest-partition
-+ (i.e. guestId != NCSW_MASTER_ID)
-+*//***************************************************************************/
-+t_Error FM_ConfigEnableMuramTestMode(t_Handle h_Fm);
-+
-+/**************************************************************************//**
-+ @Function FM_ConfigEnableIramTestMode
-+
-+ @Description Enable IRAM test mode.
-+ Calling this routine changes the internal driver data base
-+ from its default selection of test mode where it's disabled.
-+ This routine is only avaiable on old FM revisions (FMan v2).
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_Config() and before FM_Init().
-+ This routine should NOT be called from guest-partition
-+ (i.e. guestId != NCSW_MASTER_ID)
-+*//***************************************************************************/
-+t_Error FM_ConfigEnableIramTestMode(t_Handle h_Fm);
-+
-+/**************************************************************************//**
-+ @Function FM_ConfigHaltOnExternalActivation
-+
-+ @Description Define FM behavior on external halt activation.
-+ Calling this routine changes the FM behavior on external halt
-+ activation in the internal driver data base from its default
-+ [DEFAULT_haltOnExternalActivation].
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+ @Param[in] enable TRUE to enable halt on external halt
-+ activation.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_Config() and before FM_Init().
-+ This routine should NOT be called from guest-partition
-+ (i.e. guestId != NCSW_MASTER_ID)
-+*//***************************************************************************/
-+t_Error FM_ConfigHaltOnExternalActivation(t_Handle h_Fm, bool enable);
-+
-+/**************************************************************************//**
-+ @Function FM_ConfigHaltOnUnrecoverableEccError
-+
-+ @Description Define FM behavior on external halt activation.
-+ Calling this routine changes the FM behavior on unrecoverable
-+ ECC error in the internal driver data base from its default
-+ [DEFAULT_haltOnUnrecoverableEccError].
-+ This routine is only avaiable on old FM revisions (FMan v2).
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+ @Param[in] enable TRUE to enable halt on unrecoverable Ecc error
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_Config() and before FM_Init().
-+ This routine should NOT be called from guest-partition
-+ (i.e. guestId != NCSW_MASTER_ID)
-+*//***************************************************************************/
-+t_Error FM_ConfigHaltOnUnrecoverableEccError(t_Handle h_Fm, bool enable);
-+
-+/**************************************************************************//**
-+ @Function FM_ConfigException
-+
-+ @Description Define FM exceptions.
-+ Calling this routine changes the exceptions defaults in the
-+ internal driver data base where all exceptions are enabled.
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+ @Param[in] exception The exception to be selected.
-+ @Param[in] enable TRUE to enable interrupt, FALSE to mask it.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_Config() and before FM_Init().
-+ This routine should NOT be called from guest-partition
-+ (i.e. guestId != NCSW_MASTER_ID)
-+*//***************************************************************************/
-+t_Error FM_ConfigException(t_Handle h_Fm, e_FmExceptions exception, bool enable);
-+
-+/**************************************************************************//**
-+ @Function FM_ConfigExternalEccRamsEnable
-+
-+ @Description Select external ECC enabling.
-+ Calling this routine changes the ECC enabling control in the internal
-+ driver data base from its default [DEFAULT_externalEccRamsEnable].
-+ When this option is enabled Rams ECC enabling is not effected
-+ by FM_EnableRamsEcc/FM_DisableRamsEcc, but by a JTAG.
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+ @Param[in] enable TRUE to enable this option.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_Config() and before FM_Init().
-+ This routine should NOT be called from guest-partition
-+ (i.e. guestId != NCSW_MASTER_ID)
-+*//***************************************************************************/
-+t_Error FM_ConfigExternalEccRamsEnable(t_Handle h_Fm, bool enable);
-+
-+/**************************************************************************//**
-+ @Function FM_ConfigTnumAgingPeriod
-+
-+ @Description Define Tnum aging period.
-+ Calling this routine changes the Tnum aging of dequeue TNUMs
-+ in the QMI in the internal driver data base from its default
-+ [DEFAULT_tnumAgingPeriod].
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+ @Param[in] tnumAgingPeriod Tnum Aging Period in microseconds.
-+ Note that period is recalculated in units of
-+ 64 FM clocks. Driver will pick the closest
-+ possible period.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_Config() and before FM_Init().
-+ This routine should NOT be called from guest-partition
-+ (i.e. guestId != NCSW_MASTER_ID)
-+ NOTE that if some MAC is configured for PFC, '0' value is NOT
-+ allowed.
-+*//***************************************************************************/
-+t_Error FM_ConfigTnumAgingPeriod(t_Handle h_Fm, uint16_t tnumAgingPeriod);
-+
-+/**************************************************************************//*
-+ @Function FM_ConfigDmaEmergencySmoother
-+
-+ @Description Define DMA emergency smoother.
-+ Calling this routine changes the definition of the minimum
-+ amount of DATA beats transferred on the AXI READ and WRITE
-+ ports before lowering the emergency level.
-+ By default smoother is disabled.
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+ @Param[in] emergencyCnt emergency switching counter.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_Config() and before FM_Init().
-+ This routine should NOT be called from guest-partition
-+ (i.e. guestId != NCSW_MASTER_ID)
-+*//***************************************************************************/
-+t_Error FM_ConfigDmaEmergencySmoother(t_Handle h_Fm, uint32_t emergencyCnt);
-+
-+/**************************************************************************//*
-+ @Function FM_ConfigThresholds
-+
-+ @Description Calling this routine changes the internal driver data base
-+ from its default FM threshold configuration:
-+ dispLimit: [DEFAULT_dispLimit]
-+ prsDispTh: [DEFAULT_prsDispTh]
-+ plcrDispTh: [DEFAULT_plcrDispTh]
-+ kgDispTh: [DEFAULT_kgDispTh]
-+ bmiDispTh: [DEFAULT_bmiDispTh]
-+ qmiEnqDispTh: [DEFAULT_qmiEnqDispTh]
-+ qmiDeqDispTh: [DEFAULT_qmiDeqDispTh]
-+ fmCtl1DispTh: [DEFAULT_fmCtl1DispTh]
-+ fmCtl2DispTh: [DEFAULT_fmCtl2DispTh]
-+
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+ @Param[in] p_FmThresholds A structure of threshold parameters.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_Config() and before FM_Init().
-+ This routine should NOT be called from guest-partition
-+ (i.e. guestId != NCSW_MASTER_ID)
-+*//***************************************************************************/
-+t_Error FM_ConfigThresholds(t_Handle h_Fm, t_FmThresholds *p_FmThresholds);
-+
-+/**************************************************************************//*
-+ @Function FM_ConfigDmaSosEmergencyThreshold
-+
-+ @Description Calling this routine changes the internal driver data base
-+ from its default dma SOS emergency configuration [DEFAULT_dmaSosEmergency]
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+ @Param[in] dmaSosEmergency The selected new value.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_Config() and before FM_Init().
-+ This routine should NOT be called from guest-partition
-+ (i.e. guestId != NCSW_MASTER_ID)
-+*//***************************************************************************/
-+t_Error FM_ConfigDmaSosEmergencyThreshold(t_Handle h_Fm, uint32_t dmaSosEmergency);
-+
-+/**************************************************************************//*
-+ @Function FM_ConfigDmaWriteBufThresholds
-+
-+ @Description Calling this routine changes the internal driver data base
-+ from its default configuration of DMA write buffer threshold
-+ assertEmergency: [DEFAULT_dmaWriteIntBufLow]
-+ clearEmergency: [DEFAULT_dmaWriteIntBufHigh]
-+ This routine is only avaiable on old FM revisions (FMan v2).
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+ @Param[in] p_FmDmaThresholds A structure of thresholds to define emergency behavior -
-+ When 'assertEmergency' value is reached, emergency is asserted,
-+ then it is held until 'clearEmergency' value is reached.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_Config() and before FM_Init().
-+ This routine should NOT be called from guest-partition
-+ (i.e. guestId != NCSW_MASTER_ID)
-+*//***************************************************************************/
-+t_Error FM_ConfigDmaWriteBufThresholds(t_Handle h_Fm, t_FmDmaThresholds *p_FmDmaThresholds);
-+
-+ /**************************************************************************//*
-+ @Function FM_ConfigDmaCommQThresholds
-+
-+ @Description Calling this routine changes the internal driver data base
-+ from its default configuration of DMA command queue threshold
-+ assertEmergency: [DEFAULT_dmaCommQLow]
-+ clearEmergency: [DEFAULT_dmaCommQHigh]
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+ @Param[in] p_FmDmaThresholds A structure of thresholds to define emergency behavior -
-+ When 'assertEmergency' value is reached, emergency is asserted,
-+ then it is held until 'clearEmergency' value is reached..
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_Config() and before FM_Init().
-+ This routine should NOT be called from guest-partition
-+ (i.e. guestId != NCSW_MASTER_ID)
-+*//***************************************************************************/
-+t_Error FM_ConfigDmaCommQThresholds(t_Handle h_Fm, t_FmDmaThresholds *p_FmDmaThresholds);
-+
-+/**************************************************************************//*
-+ @Function FM_ConfigDmaReadBufThresholds
-+
-+ @Description Calling this routine changes the internal driver data base
-+ from its default configuration of DMA read buffer threshold
-+ assertEmergency: [DEFAULT_dmaReadIntBufLow]
-+ clearEmergency: [DEFAULT_dmaReadIntBufHigh]
-+ This routine is only avaiable on old FM revisions (FMan v2).
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+ @Param[in] p_FmDmaThresholds A structure of thresholds to define emergency behavior -
-+ When 'assertEmergency' value is reached, emergency is asserted,
-+ then it is held until 'clearEmergency' value is reached..
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_Config() and before FM_Init().
-+ This routine should NOT be called from guest-partition
-+ (i.e. guestId != NCSW_MASTER_ID)
-+*//***************************************************************************/
-+t_Error FM_ConfigDmaReadBufThresholds(t_Handle h_Fm, t_FmDmaThresholds *p_FmDmaThresholds);
-+
-+/**************************************************************************//*
-+ @Function FM_ConfigDmaWatchdog
-+
-+ @Description Calling this routine changes the internal driver data base
-+ from its default watchdog configuration, which is disabled
-+ [DEFAULT_dmaWatchdog].
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+ @Param[in] watchDogValue The selected new value - in microseconds.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_Config() and before FM_Init().
-+ This routine should NOT be called from guest-partition
-+ (i.e. guestId != NCSW_MASTER_ID)
-+*//***************************************************************************/
-+t_Error FM_ConfigDmaWatchdog(t_Handle h_Fm, uint32_t watchDogValue);
-+
-+/** @} */ /* end of FM_advanced_init_grp group */
-+/** @} */ /* end of FM_init_grp group */
-+
-+
-+/**************************************************************************//**
-+ @Group FM_runtime_control_grp FM Runtime Control Unit
-+
-+ @Description FM Runtime control unit API functions, definitions and enums.
-+ The FM driver provides a set of control routines.
-+ These routines may only be called after the module was fully
-+ initialized (both configuration and initialization routines were
-+ called). They are typically used to get information from hardware
-+ (status, counters/statistics, revision etc.), to modify a current
-+ state or to force/enable a required action. Run-time control may
-+ be called whenever necessary and as many times as needed.
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Collection General FM defines.
-+*//***************************************************************************/
-+#define FM_MAX_NUM_OF_VALID_PORTS (FM_MAX_NUM_OF_OH_PORTS + \
-+ FM_MAX_NUM_OF_1G_RX_PORTS + \
-+ FM_MAX_NUM_OF_10G_RX_PORTS + \
-+ FM_MAX_NUM_OF_1G_TX_PORTS + \
-+ FM_MAX_NUM_OF_10G_TX_PORTS) /**< Number of available FM ports */
-+/* @} */
-+
-+/**************************************************************************//*
-+ @Description A Structure for Port bandwidth requirement. Port is identified
-+ by type and relative id.
-+*//***************************************************************************/
-+typedef struct t_FmPortBandwidth {
-+ e_FmPortType type; /**< FM port type */
-+ uint8_t relativePortId; /**< Type relative port id */
-+ uint8_t bandwidth; /**< bandwidth - (in term of percents) */
-+} t_FmPortBandwidth;
-+
-+/**************************************************************************//*
-+ @Description A Structure containing an array of Port bandwidth requirements.
-+ The user should state the ports requiring bandwidth in terms of
-+ percentage - i.e. all port's bandwidths in the array must add
-+ up to 100.
-+*//***************************************************************************/
-+typedef struct t_FmPortsBandwidthParams {
-+ uint8_t numOfPorts; /**< The number of relevant ports, which is the
-+ number of valid entries in the array below */
-+ t_FmPortBandwidth portsBandwidths[FM_MAX_NUM_OF_VALID_PORTS];
-+ /**< for each port, it's bandwidth (all port's
-+ bandwidths must add up to 100.*/
-+} t_FmPortsBandwidthParams;
-+
-+/**************************************************************************//**
-+ @Description DMA Emergency control on MURAM
-+*//***************************************************************************/
-+typedef enum e_FmDmaMuramPort {
-+ e_FM_DMA_MURAM_PORT_WRITE, /**< MURAM write port */
-+ e_FM_DMA_MURAM_PORT_READ /**< MURAM read port */
-+} e_FmDmaMuramPort;
-+
-+/**************************************************************************//**
-+ @Description Enum for defining FM counters
-+*//***************************************************************************/
-+typedef enum e_FmCounters {
-+ e_FM_COUNTERS_ENQ_TOTAL_FRAME = 0, /**< QMI total enqueued frames counter */
-+ e_FM_COUNTERS_DEQ_TOTAL_FRAME, /**< QMI total dequeued frames counter */
-+ e_FM_COUNTERS_DEQ_0, /**< QMI 0 frames from QMan counter */
-+ e_FM_COUNTERS_DEQ_1, /**< QMI 1 frames from QMan counter */
-+ e_FM_COUNTERS_DEQ_2, /**< QMI 2 frames from QMan counter */
-+ e_FM_COUNTERS_DEQ_3, /**< QMI 3 frames from QMan counter */
-+ e_FM_COUNTERS_DEQ_FROM_DEFAULT, /**< QMI dequeue from default queue counter */
-+ e_FM_COUNTERS_DEQ_FROM_CONTEXT, /**< QMI dequeue from FQ context counter */
-+ e_FM_COUNTERS_DEQ_FROM_FD, /**< QMI dequeue from FD command field counter */
-+ e_FM_COUNTERS_DEQ_CONFIRM /**< QMI dequeue confirm counter */
-+} e_FmCounters;
-+
-+/**************************************************************************//**
-+ @Description A Structure for returning FM revision information
-+*//***************************************************************************/
-+typedef struct t_FmRevisionInfo {
-+ uint8_t majorRev; /**< Major revision */
-+ uint8_t minorRev; /**< Minor revision */
-+} t_FmRevisionInfo;
-+
-+/**************************************************************************//**
-+ @Description A Structure for returning FM ctrl code revision information
-+*//***************************************************************************/
-+typedef struct t_FmCtrlCodeRevisionInfo {
-+ uint16_t packageRev; /**< Package revision */
-+ uint8_t majorRev; /**< Major revision */
-+ uint8_t minorRev; /**< Minor revision */
-+} t_FmCtrlCodeRevisionInfo;
-+
-+/**************************************************************************//**
-+ @Description A Structure for defining DMA status
-+*//***************************************************************************/
-+typedef struct t_FmDmaStatus {
-+ bool cmqNotEmpty; /**< Command queue is not empty */
-+ bool busError; /**< Bus error occurred */
-+ bool readBufEccError; /**< Double ECC error on buffer Read (Valid for FM rev < 6)*/
-+ bool writeBufEccSysError; /**< Double ECC error on buffer write from system side (Valid for FM rev < 6)*/
-+ bool writeBufEccFmError; /**< Double ECC error on buffer write from FM side (Valid for FM rev < 6) */
-+ bool singlePortEccError; /**< Single Port ECC error from FM side (Valid for FM rev >= 6)*/
-+} t_FmDmaStatus;
-+
-+/**************************************************************************//**
-+ @Description A Structure for obtaining FM controller monitor values
-+*//***************************************************************************/
-+typedef struct t_FmCtrlMon {
-+ uint8_t percentCnt[2]; /**< Percentage value */
-+} t_FmCtrlMon;
-+
-+
-+#if (defined(DEBUG_ERRORS) && (DEBUG_ERRORS > 0))
-+/**************************************************************************//**
-+ @Function FM_DumpRegs
-+
-+ @Description Dumps all FM registers
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+
-+ @Return E_OK on success;
-+
-+ @Cautions Allowed only following FM_Init().
-+*//***************************************************************************/
-+t_Error FM_DumpRegs(t_Handle h_Fm);
-+#endif /* (defined(DEBUG_ERRORS) && ... */
-+
-+/**************************************************************************//**
-+ @Function FM_SetException
-+
-+ @Description Calling this routine enables/disables the specified exception.
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+ @Param[in] exception The exception to be selected.
-+ @Param[in] enable TRUE to enable interrupt, FALSE to mask it.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_Init().
-+ This routine should NOT be called from guest-partition
-+ (i.e. guestId != NCSW_MASTER_ID)
-+*//***************************************************************************/
-+t_Error FM_SetException(t_Handle h_Fm, e_FmExceptions exception, bool enable);
-+
-+/**************************************************************************//**
-+ @Function FM_EnableRamsEcc
-+
-+ @Description Enables ECC mechanism for all the different FM RAM's; E.g. IRAM,
-+ MURAM, Parser, Keygen, Policer, etc.
-+ Note:
-+ If FM_ConfigExternalEccRamsEnable was called to enable external
-+ setting of ECC, this routine effects IRAM ECC only.
-+ This routine is also called by the driver if an ECC exception is
-+ enabled.
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_Config() and before FM_Init().
-+ This routine should NOT be called from guest-partition
-+ (i.e. guestId != NCSW_MASTER_ID)
-+*//***************************************************************************/
-+t_Error FM_EnableRamsEcc(t_Handle h_Fm);
-+
-+/**************************************************************************//**
-+ @Function FM_DisableRamsEcc
-+
-+ @Description Disables ECC mechanism for all the different FM RAM's; E.g. IRAM,
-+ MURAM, Parser, Keygen, Policer, etc.
-+ Note:
-+ If FM_ConfigExternalEccRamsEnable was called to enable external
-+ setting of ECC, this routine effects IRAM ECC only.
-+ In opposed to FM_EnableRamsEcc, this routine must be called
-+ explicitly to disable all Rams ECC.
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_Config() and before FM_Init().
-+ This routine should NOT be called from guest-partition
-+ (i.e. guestId != NCSW_MASTER_ID)
-+*//***************************************************************************/
-+t_Error FM_DisableRamsEcc(t_Handle h_Fm);
-+
-+/**************************************************************************//**
-+ @Function FM_GetRevision
-+
-+ @Description Returns the FM revision
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+ @Param[out] p_FmRevisionInfo A structure of revision information parameters.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_Init().
-+*//***************************************************************************/
-+t_Error FM_GetRevision(t_Handle h_Fm, t_FmRevisionInfo *p_FmRevisionInfo);
-+
-+/**************************************************************************//**
-+ @Function FM_GetFmanCtrlCodeRevision
-+
-+ @Description Returns the Fman controller code revision
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+ @Param[out] p_RevisionInfo A structure of revision information parameters.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_Init().
-+*//***************************************************************************/
-+t_Error FM_GetFmanCtrlCodeRevision(t_Handle h_Fm, t_FmCtrlCodeRevisionInfo *p_RevisionInfo);
-+
-+/**************************************************************************//**
-+ @Function FM_GetCounter
-+
-+ @Description Reads one of the FM counters.
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+ @Param[in] counter The requested counter.
-+
-+ @Return Counter's current value.
-+
-+ @Cautions Allowed only following FM_Init().
-+ Note that it is user's responsibility to call this routine only
-+ for enabled counters, and there will be no indication if a
-+ disabled counter is accessed.
-+*//***************************************************************************/
-+uint32_t FM_GetCounter(t_Handle h_Fm, e_FmCounters counter);
-+
-+/**************************************************************************//**
-+ @Function FM_ModifyCounter
-+
-+ @Description Sets a value to an enabled counter. Use "0" to reset the counter.
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+ @Param[in] counter The requested counter.
-+ @Param[in] val The requested value to be written into the counter.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_Init().
-+ This routine should NOT be called from guest-partition
-+ (i.e. guestId != NCSW_MASTER_ID)
-+*//***************************************************************************/
-+t_Error FM_ModifyCounter(t_Handle h_Fm, e_FmCounters counter, uint32_t val);
-+
-+/**************************************************************************//**
-+ @Function FM_Resume
-+
-+ @Description Release FM after halt FM command or after unrecoverable ECC error.
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_Init().
-+ This routine should NOT be called from guest-partition
-+ (i.e. guestId != NCSW_MASTER_ID)
-+*//***************************************************************************/
-+void FM_Resume(t_Handle h_Fm);
-+
-+/**************************************************************************//**
-+ @Function FM_SetDmaEmergency
-+
-+ @Description Manual emergency set
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+ @Param[in] muramPort MURAM direction select.
-+ @Param[in] enable TRUE to manually enable emergency, FALSE to disable.
-+
-+ @Return None.
-+
-+ @Cautions Allowed only following FM_Init().
-+ This routine should NOT be called from guest-partition
-+ (i.e. guestId != NCSW_MASTER_ID)
-+*//***************************************************************************/
-+void FM_SetDmaEmergency(t_Handle h_Fm, e_FmDmaMuramPort muramPort, bool enable);
-+
-+/**************************************************************************//**
-+ @Function FM_SetDmaExtBusPri
-+
-+ @Description Set the DMA external bus priority
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+ @Param[in] pri External bus priority select
-+
-+ @Return None.
-+
-+ @Cautions Allowed only following FM_Init().
-+ This routine should NOT be called from guest-partition
-+ (i.e. guestId != NCSW_MASTER_ID)
-+*//***************************************************************************/
-+void FM_SetDmaExtBusPri(t_Handle h_Fm, e_FmDmaExtBusPri pri);
-+
-+/**************************************************************************//**
-+ @Function FM_GetDmaStatus
-+
-+ @Description Reads the DMA current status
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+ @Param[out] p_FmDmaStatus A structure of DMA status parameters.
-+
-+ @Cautions Allowed only following FM_Init().
-+*//***************************************************************************/
-+void FM_GetDmaStatus(t_Handle h_Fm, t_FmDmaStatus *p_FmDmaStatus);
-+
-+/**************************************************************************//**
-+ @Function FM_ErrorIsr
-+
-+ @Description FM interrupt-service-routine for errors.
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+
-+ @Return E_OK on success; E_EMPTY if no errors found in register, other
-+ error code otherwise.
-+
-+ @Cautions Allowed only following FM_Init().
-+ This routine should NOT be called from guest-partition
-+ (i.e. guestId != NCSW_MASTER_ID)
-+*//***************************************************************************/
-+t_Error FM_ErrorIsr(t_Handle h_Fm);
-+
-+/**************************************************************************//**
-+ @Function FM_EventIsr
-+
-+ @Description FM interrupt-service-routine for normal events.
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+
-+ @Cautions Allowed only following FM_Init().
-+ This routine should NOT be called from guest-partition
-+ (i.e. guestId != NCSW_MASTER_ID)
-+*//***************************************************************************/
-+void FM_EventIsr(t_Handle h_Fm);
-+
-+/**************************************************************************//**
-+ @Function FM_GetSpecialOperationCoding
-+
-+ @Description Return a specific coding according to the input mask.
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+ @Param[in] spOper special operation mask.
-+ @Param[out] p_SpOperCoding special operation code.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_Init().
-+*//***************************************************************************/
-+t_Error FM_GetSpecialOperationCoding(t_Handle h_Fm,
-+ fmSpecialOperations_t spOper,
-+ uint8_t *p_SpOperCoding);
-+
-+/**************************************************************************//**
-+ @Function FM_CtrlMonStart
-+
-+ @Description Start monitoring utilization of all available FM controllers.
-+
-+ In order to obtain FM controllers utilization the following sequence
-+ should be used:
-+ -# FM_CtrlMonStart()
-+ -# FM_CtrlMonStop()
-+ -# FM_CtrlMonGetCounters() - issued for each FM controller
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_Init().
-+ This routine should NOT be called from guest-partition
-+ (i.e. guestId != NCSW_MASTER_ID).
-+*//***************************************************************************/
-+t_Error FM_CtrlMonStart(t_Handle h_Fm);
-+
-+/**************************************************************************//**
-+ @Function FM_CtrlMonStop
-+
-+ @Description Stop monitoring utilization of all available FM controllers.
-+
-+ In order to obtain FM controllers utilization the following sequence
-+ should be used:
-+ -# FM_CtrlMonStart()
-+ -# FM_CtrlMonStop()
-+ -# FM_CtrlMonGetCounters() - issued for each FM controller
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_Init().
-+ This routine should NOT be called from guest-partition
-+ (i.e. guestId != NCSW_MASTER_ID).
-+*//***************************************************************************/
-+t_Error FM_CtrlMonStop(t_Handle h_Fm);
-+
-+/**************************************************************************//**
-+ @Function FM_CtrlMonGetCounters
-+
-+ @Description Obtain FM controller utilization parameters.
-+
-+ In order to obtain FM controllers utilization the following sequence
-+ should be used:
-+ -# FM_CtrlMonStart()
-+ -# FM_CtrlMonStop()
-+ -# FM_CtrlMonGetCounters() - issued for each FM controller
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+ @Param[in] fmCtrlIndex FM Controller index for that utilization results
-+ are requested.
-+ @Param[in] p_Mon Pointer to utilization results structure.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_Init().
-+ This routine should NOT be called from guest-partition
-+ (i.e. guestId != NCSW_MASTER_ID).
-+*//***************************************************************************/
-+t_Error FM_CtrlMonGetCounters(t_Handle h_Fm, uint8_t fmCtrlIndex, t_FmCtrlMon *p_Mon);
-+
-+
-+/**************************************************************************//*
-+ @Function FM_ForceIntr
-+
-+ @Description Causes an interrupt event on the requested source.
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+ @Param[in] exception An exception to be forced.
-+
-+ @Return E_OK on success; Error code if the exception is not enabled,
-+ or is not able to create interrupt.
-+
-+ @Cautions Allowed only following FM_Init().
-+ This routine should NOT be called from guest-partition
-+ (i.e. guestId != NCSW_MASTER_ID)
-+*//***************************************************************************/
-+t_Error FM_ForceIntr (t_Handle h_Fm, e_FmExceptions exception);
-+
-+/**************************************************************************//*
-+ @Function FM_SetPortsBandwidth
-+
-+ @Description Sets relative weights between ports when accessing common resources.
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+ @Param[in] p_PortsBandwidth A structure of ports bandwidths in percentage, i.e.
-+ total must equal 100.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_Init().
-+ This routine should NOT be called from guest-partition
-+ (i.e. guestId != NCSW_MASTER_ID)
-+*//***************************************************************************/
-+t_Error FM_SetPortsBandwidth(t_Handle h_Fm, t_FmPortsBandwidthParams *p_PortsBandwidth);
-+
-+/**************************************************************************//*
-+ @Function FM_GetMuramHandle
-+
-+ @Description Gets the corresponding MURAM handle
-+
-+ @Param[in] h_Fm A handle to an FM Module.
-+
-+ @Return MURAM handle; NULL otherwise.
-+
-+ @Cautions Allowed only following FM_Init().
-+ This routine should NOT be called from guest-partition
-+ (i.e. guestId != NCSW_MASTER_ID)
-+*//***************************************************************************/
-+t_Handle FM_GetMuramHandle(t_Handle h_Fm);
-+
-+/** @} */ /* end of FM_runtime_control_grp group */
-+/** @} */ /* end of FM_lib_grp group */
-+/** @} */ /* end of FM_grp group */
-+
-+
-+#ifdef NCSW_BACKWARD_COMPATIBLE_API
-+typedef t_FmFirmwareParams t_FmPcdFirmwareParams;
-+typedef t_FmBufferPrefixContent t_FmPortBufferPrefixContent;
-+typedef t_FmExtPoolParams t_FmPortExtPoolParams;
-+typedef t_FmExtPools t_FmPortExtPools;
-+typedef t_FmBackupBmPools t_FmPortBackupBmPools;
-+typedef t_FmBufPoolDepletion t_FmPortBufPoolDepletion;
-+typedef e_FmDmaSwapOption e_FmPortDmaSwapOption;
-+typedef e_FmDmaCacheOption e_FmPortDmaCacheOption;
-+
-+#define FM_CONTEXTA_GET_OVVERIDE FM_CONTEXTA_GET_OVERRIDE
-+#define FM_CONTEXTA_SET_OVVERIDE FM_CONTEXTA_SET_OVERRIDE
-+
-+#define e_FM_EX_BMI_PIPELINE_ECC e_FM_EX_BMI_STORAGE_PROFILE_ECC
-+#define e_FM_PORT_DMA_NO_SWP e_FM_DMA_NO_SWP
-+#define e_FM_PORT_DMA_SWP_PPC_LE e_FM_DMA_SWP_PPC_LE
-+#define e_FM_PORT_DMA_SWP_BE e_FM_DMA_SWP_BE
-+#define e_FM_PORT_DMA_NO_STASH e_FM_DMA_NO_STASH
-+#define e_FM_PORT_DMA_STASH e_FM_DMA_STASH
-+#endif /* NCSW_BACKWARD_COMPATIBLE_API */
-+
-+
-+#endif /* __FM_EXT */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/Peripherals/fm_mac_ext.h
-@@ -0,0 +1,887 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/**************************************************************************//**
-+ @File fm_mac_ext.h
-+
-+ @Description FM MAC ...
-+*//***************************************************************************/
-+#ifndef __FM_MAC_EXT_H
-+#define __FM_MAC_EXT_H
-+
-+#include "std_ext.h"
-+#include "enet_ext.h"
-+
-+
-+/**************************************************************************//**
-+
-+ @Group FM_grp Frame Manager API
-+
-+ @Description FM API functions, definitions and enums
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Group FM_mac_grp FM MAC
-+
-+ @Description FM MAC API functions, definitions and enums
-+
-+ @{
-+*//***************************************************************************/
-+
-+#define FM_MAC_NO_PFC 0xff
-+
-+
-+/**************************************************************************//**
-+ @Description FM MAC Exceptions
-+*//***************************************************************************/
-+typedef enum e_FmMacExceptions {
-+ e_FM_MAC_EX_10G_MDIO_SCAN_EVENTMDIO = 0 /**< 10GEC MDIO scan event interrupt */
-+ ,e_FM_MAC_EX_10G_MDIO_CMD_CMPL /**< 10GEC MDIO command completion interrupt */
-+ ,e_FM_MAC_EX_10G_REM_FAULT /**< 10GEC, mEMAC Remote fault interrupt */
-+ ,e_FM_MAC_EX_10G_LOC_FAULT /**< 10GEC, mEMAC Local fault interrupt */
-+ ,e_FM_MAC_EX_10G_1TX_ECC_ER /**< 10GEC, mEMAC Transmit frame ECC error interrupt */
-+ ,e_FM_MAC_EX_10G_TX_FIFO_UNFL /**< 10GEC, mEMAC Transmit FIFO underflow interrupt */
-+ ,e_FM_MAC_EX_10G_TX_FIFO_OVFL /**< 10GEC, mEMAC Transmit FIFO overflow interrupt */
-+ ,e_FM_MAC_EX_10G_TX_ER /**< 10GEC Transmit frame error interrupt */
-+ ,e_FM_MAC_EX_10G_RX_FIFO_OVFL /**< 10GEC, mEMAC Receive FIFO overflow interrupt */
-+ ,e_FM_MAC_EX_10G_RX_ECC_ER /**< 10GEC, mEMAC Receive frame ECC error interrupt */
-+ ,e_FM_MAC_EX_10G_RX_JAB_FRM /**< 10GEC Receive jabber frame interrupt */
-+ ,e_FM_MAC_EX_10G_RX_OVRSZ_FRM /**< 10GEC Receive oversized frame interrupt */
-+ ,e_FM_MAC_EX_10G_RX_RUNT_FRM /**< 10GEC Receive runt frame interrupt */
-+ ,e_FM_MAC_EX_10G_RX_FRAG_FRM /**< 10GEC Receive fragment frame interrupt */
-+ ,e_FM_MAC_EX_10G_RX_LEN_ER /**< 10GEC Receive payload length error interrupt */
-+ ,e_FM_MAC_EX_10G_RX_CRC_ER /**< 10GEC Receive CRC error interrupt */
-+ ,e_FM_MAC_EX_10G_RX_ALIGN_ER /**< 10GEC Receive alignment error interrupt */
-+ ,e_FM_MAC_EX_1G_BAB_RX /**< dTSEC Babbling receive error */
-+ ,e_FM_MAC_EX_1G_RX_CTL /**< dTSEC Receive control (pause frame) interrupt */
-+ ,e_FM_MAC_EX_1G_GRATEFUL_TX_STP_COMPLET /**< dTSEC Graceful transmit stop complete */
-+ ,e_FM_MAC_EX_1G_BAB_TX /**< dTSEC Babbling transmit error */
-+ ,e_FM_MAC_EX_1G_TX_CTL /**< dTSEC Transmit control (pause frame) interrupt */
-+ ,e_FM_MAC_EX_1G_TX_ERR /**< dTSEC Transmit error */
-+ ,e_FM_MAC_EX_1G_LATE_COL /**< dTSEC Late collision */
-+ ,e_FM_MAC_EX_1G_COL_RET_LMT /**< dTSEC Collision retry limit */
-+ ,e_FM_MAC_EX_1G_TX_FIFO_UNDRN /**< dTSEC Transmit FIFO underrun */
-+ ,e_FM_MAC_EX_1G_MAG_PCKT /**< dTSEC Magic Packet detection */
-+ ,e_FM_MAC_EX_1G_MII_MNG_RD_COMPLET /**< dTSEC MII management read completion */
-+ ,e_FM_MAC_EX_1G_MII_MNG_WR_COMPLET /**< dTSEC MII management write completion */
-+ ,e_FM_MAC_EX_1G_GRATEFUL_RX_STP_COMPLET /**< dTSEC Graceful receive stop complete */
-+ ,e_FM_MAC_EX_1G_TX_DATA_ERR /**< dTSEC Internal data error on transmit */
-+ ,e_FM_MAC_EX_1G_RX_DATA_ERR /**< dTSEC Internal data error on receive */
-+ ,e_FM_MAC_EX_1G_1588_TS_RX_ERR /**< dTSEC Time-Stamp Receive Error */
-+ ,e_FM_MAC_EX_1G_RX_MIB_CNT_OVFL /**< dTSEC MIB counter overflow */
-+ ,e_FM_MAC_EX_TS_FIFO_ECC_ERR /**< mEMAC Time-stamp FIFO ECC error interrupt;
-+ not supported on T4240/B4860 rev1 chips */
-+ ,e_FM_MAC_EX_MAGIC_PACKET_INDICATION = e_FM_MAC_EX_1G_MAG_PCKT
-+ /**< mEMAC Magic Packet Indication Interrupt */
-+} e_FmMacExceptions;
-+
-+/**************************************************************************//**
-+ @Description TM MAC statistics level
-+*//***************************************************************************/
-+typedef enum e_FmMacStatisticsLevel {
-+ e_FM_MAC_NONE_STATISTICS = 0, /**< No statistics */
-+ e_FM_MAC_PARTIAL_STATISTICS, /**< Only error counters are available; Optimized for performance */
-+ e_FM_MAC_FULL_STATISTICS /**< All counters available; Not optimized for performance */
-+} e_FmMacStatisticsLevel;
-+
-+
-+#if (DPAA_VERSION >= 11)
-+/**************************************************************************//**
-+ @Description Priority Flow Control Parameters
-+*//***************************************************************************/
-+typedef struct t_FmMacPfcParams {
-+ bool pfcEnable; /**< Enable/Disable PFC */
-+
-+ uint16_t pauseQuanta[FM_MAX_NUM_OF_PFC_PRIORITIES]; /**< Pause Quanta per priority to be sent in a pause frame. Each quanta represents a 512 bit-times*/
-+
-+ uint16_t pauseThresholdQuanta[FM_MAX_NUM_OF_PFC_PRIORITIES];/**< Pause threshold per priority, when timer passes this threshold time a PFC frames is sent again if the port is still congested or BM pool in depletion*/
-+
-+
-+} t_FmMacPfcParams;
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+/**************************************************************************//**
-+ @Function t_FmMacExceptionCallback
-+
-+ @Description Fm Mac Exception Callback from FM MAC to the user
-+
-+ @Param[in] h_App - Handle to the upper layer handler
-+
-+ @Param[in] exceptions - The exception that occurred
-+
-+ @Return void.
-+*//***************************************************************************/
-+typedef void (t_FmMacExceptionCallback)(t_Handle h_App, e_FmMacExceptions exceptions);
-+
-+
-+/**************************************************************************//**
-+ @Description TM MAC statistics rfc3635
-+*//***************************************************************************/
-+typedef struct t_FmMacStatistics {
-+/* RMON */
-+ uint64_t eStatPkts64; /**< r-10G tr-DT 64 byte frame counter */
-+ uint64_t eStatPkts65to127; /**< r-10G 65 to 127 byte frame counter */
-+ uint64_t eStatPkts128to255; /**< r-10G 128 to 255 byte frame counter */
-+ uint64_t eStatPkts256to511; /**< r-10G 256 to 511 byte frame counter */
-+ uint64_t eStatPkts512to1023; /**< r-10G 512 to 1023 byte frame counter */
-+ uint64_t eStatPkts1024to1518; /**< r-10G 1024 to 1518 byte frame counter */
-+ uint64_t eStatPkts1519to1522; /**< r-10G 1519 to 1522 byte good frame count */
-+/* */
-+ uint64_t eStatFragments; /**< Total number of packets that were less than 64 octets long with a wrong CRC.*/
-+ uint64_t eStatJabbers; /**< Total number of packets longer than valid maximum length octets */
-+ uint64_t eStatsDropEvents; /**< number of dropped packets due to internal errors of the MAC Client (during receive). */
-+ uint64_t eStatCRCAlignErrors; /**< Incremented when frames of correct length but with CRC error are received.*/
-+ uint64_t eStatUndersizePkts; /**< Incremented for frames under 64 bytes with a valid FCS and otherwise well formed;
-+ This count does not include range length errors */
-+ uint64_t eStatOversizePkts; /**< Incremented for frames which exceed 1518 (non VLAN) or 1522 (VLAN) and contains
-+ a valid FCS and otherwise well formed */
-+/* Pause */
-+ uint64_t teStatPause; /**< Pause MAC Control received */
-+ uint64_t reStatPause; /**< Pause MAC Control sent */
-+/* MIB II */
-+ uint64_t ifInOctets; /**< Total number of byte received. */
-+ uint64_t ifInPkts; /**< Total number of packets received.*/
-+ uint64_t ifInUcastPkts; /**< Total number of unicast frame received;
-+ NOTE: this counter is not supported on dTSEC MAC */
-+ uint64_t ifInMcastPkts; /**< Total number of multicast frame received*/
-+ uint64_t ifInBcastPkts; /**< Total number of broadcast frame received */
-+ uint64_t ifInDiscards; /**< Frames received, but discarded due to problems within the MAC RX. */
-+ uint64_t ifInErrors; /**< Number of frames received with error:
-+ - FIFO Overflow Error
-+ - CRC Error
-+ - Frame Too Long Error
-+ - Alignment Error
-+ - The dedicated Error Code (0xfe, not a code error) was received */
-+ uint64_t ifOutOctets; /**< Total number of byte sent. */
-+ uint64_t ifOutPkts; /**< Total number of packets sent .*/
-+ uint64_t ifOutUcastPkts; /**< Total number of unicast frame sent;
-+ NOTE: this counter is not supported on dTSEC MAC */
-+ uint64_t ifOutMcastPkts; /**< Total number of multicast frame sent */
-+ uint64_t ifOutBcastPkts; /**< Total number of multicast frame sent */
-+ uint64_t ifOutDiscards; /**< Frames received, but discarded due to problems within the MAC TX N/A!.*/
-+ uint64_t ifOutErrors; /**< Number of frames transmitted with error:
-+ - FIFO Overflow Error
-+ - FIFO Underflow Error
-+ - Other */
-+} t_FmMacStatistics;
-+
-+/**************************************************************************//**
-+ @Description FM MAC Frame Size Counters
-+*//***************************************************************************/
-+typedef struct t_FmMacFrameSizeCounters {
-+
-+ uint64_t count_pkts_64; /**< 64 byte frame counter */
-+ uint64_t count_pkts_65_to_127; /**< 65 to 127 byte frame counter */
-+ uint64_t count_pkts_128_to_255; /**< 128 to 255 byte frame counter */
-+ uint64_t count_pkts_256_to_511; /**< 256 to 511 byte frame counter */
-+ uint64_t count_pkts_512_to_1023; /**< 512 to 1023 byte frame counter */
-+ uint64_t count_pkts_1024_to_1518; /**< 1024 to 1518 byte frame counter */
-+ uint64_t count_pkts_1519_to_1522; /**< 1519 to 1522 byte good frame count */
-+} t_FmMacFrameSizeCounters;
-+
-+/**************************************************************************//**
-+ @Group FM_mac_init_grp FM MAC Initialization Unit
-+
-+ @Description FM MAC Initialization Unit
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Description FM MAC config input
-+*//***************************************************************************/
-+typedef struct t_FmMacParams {
-+ uintptr_t baseAddr; /**< Base of memory mapped FM MAC registers */
-+ t_EnetAddr addr; /**< MAC address of device; First octet is sent first */
-+ uint8_t macId; /**< MAC ID;
-+ numbering of dTSEC and 1G-mEMAC:
-+ 0 - FM_MAX_NUM_OF_1G_MACS;
-+ numbering of 10G-MAC (TGEC) and 10G-mEMAC:
-+ 0 - FM_MAX_NUM_OF_10G_MACS */
-+ e_EnetMode enetMode; /**< Ethernet operation mode (MAC-PHY interface and speed);
-+ Note that the speed should indicate the maximum rate that
-+ this MAC should support rather than the actual speed;
-+ i.e. user should use the FM_MAC_AdjustLink() routine to
-+ provide accurate speed;
-+ In case of mEMAC RGMII mode, the MAC is configured to RGMII
-+ automatic mode, where actual speed/duplex mode information
-+ is provided by PHY automatically in-band; FM_MAC_AdjustLink()
-+ function should be used to switch to manual RGMII speed/duplex mode
-+ configuration if RGMII PHY doesn't support in-band status signaling;
-+ In addition, in mEMAC, in case where user is using the higher MACs
-+ (i.e. the MACs that should support 10G), user should pass here
-+ speed=10000 even if the interface is not allowing that (e.g. SGMII). */
-+ t_Handle h_Fm; /**< A handle to the FM object this port related to */
-+ int mdioIrq; /**< MDIO exceptions interrupt source - not valid for all
-+ MACs; MUST be set to 'NO_IRQ' for MACs that don't have
-+ mdio-irq, or for polling */
-+ t_FmMacExceptionCallback *f_Event; /**< MDIO Events Callback Routine */
-+ t_FmMacExceptionCallback *f_Exception; /**< Exception Callback Routine */
-+ t_Handle h_App; /**< A handle to an application layer object; This handle will
-+ be passed by the driver upon calling the above callbacks */
-+} t_FmMacParams;
-+
-+
-+/**************************************************************************//**
-+ @Function FM_MAC_Config
-+
-+ @Description Creates descriptor for the FM MAC module.
-+
-+ The routine returns a handle (descriptor) to the FM MAC object.
-+ This descriptor must be passed as first parameter to all other
-+ FM MAC function calls.
-+
-+ No actual initialization or configuration of FM MAC hardware is
-+ done by this routine.
-+
-+ @Param[in] p_FmMacParam - Pointer to data structure of parameters
-+
-+ @Retval Handle to FM MAC object, or NULL for Failure.
-+*//***************************************************************************/
-+t_Handle FM_MAC_Config(t_FmMacParams *p_FmMacParam);
-+
-+/**************************************************************************//**
-+ @Function FM_MAC_Init
-+
-+ @Description Initializes the FM MAC module
-+
-+ @Param[in] h_FmMac - FM module descriptor
-+
-+ @Return E_OK on success; Error code otherwise.
-+*//***************************************************************************/
-+t_Error FM_MAC_Init(t_Handle h_FmMac);
-+
-+/**************************************************************************//**
-+ @Function FM_Free
-+
-+ @Description Frees all resources that were assigned to FM MAC module.
-+
-+ Calling this routine invalidates the descriptor.
-+
-+ @Param[in] h_FmMac - FM module descriptor
-+
-+ @Return E_OK on success; Error code otherwise.
-+*//***************************************************************************/
-+t_Error FM_MAC_Free(t_Handle h_FmMac);
-+
-+
-+/**************************************************************************//**
-+ @Group FM_mac_advanced_init_grp FM MAC Advanced Configuration Unit
-+
-+ @Description Configuration functions used to change default values.
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Function FM_MAC_ConfigResetOnInit
-+
-+ @Description Tell the driver whether to reset the FM MAC before initialization or
-+ not. It changes the default configuration [DEFAULT_resetOnInit].
-+
-+ @Param[in] h_FmMac A handle to a FM MAC Module.
-+ @Param[in] enable When TRUE, FM will be reset before any initialization.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MAC_Config() and before FM_MAC_Init().
-+*//***************************************************************************/
-+t_Error FM_MAC_ConfigResetOnInit(t_Handle h_FmMac, bool enable);
-+
-+/**************************************************************************//**
-+ @Function FM_MAC_ConfigLoopback
-+
-+ @Description Enable/Disable internal loopback mode
-+
-+ @Param[in] h_FmMac A handle to a FM MAC Module.
-+ @Param[in] enable TRUE to enable or FALSE to disable.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MAC_Config() and before FM_MAC_Init().
-+*//***************************************************************************/
-+t_Error FM_MAC_ConfigLoopback(t_Handle h_FmMac, bool enable);
-+
-+/**************************************************************************//**
-+ @Function FM_MAC_ConfigMaxFrameLength
-+
-+ @Description Setup maximum Rx Frame Length (in 1G MAC, effects also Tx)
-+
-+ @Param[in] h_FmMac A handle to a FM MAC Module.
-+ @Param[in] newVal MAX Frame length
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MAC_Config() and before FM_MAC_Init().
-+*//***************************************************************************/
-+t_Error FM_MAC_ConfigMaxFrameLength(t_Handle h_FmMac, uint16_t newVal);
-+
-+/**************************************************************************//**
-+ @Function FM_MAC_ConfigWan
-+
-+ @Description ENABLE WAN mode in 10G-MAC
-+
-+ @Param[in] h_FmMac A handle to a FM MAC Module.
-+ @Param[in] enable TRUE to enable or FALSE to disable.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MAC_Config() and before FM_MAC_Init().
-+*//***************************************************************************/
-+t_Error FM_MAC_ConfigWan(t_Handle h_FmMac, bool enable);
-+
-+/**************************************************************************//**
-+ @Function FM_MAC_ConfigPadAndCrc
-+
-+ @Description Config PAD and CRC mode
-+
-+ @Param[in] h_FmMac A handle to a FM MAC Module.
-+ @Param[in] enable TRUE to enable or FALSE to disable.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MAC_Config() and before FM_MAC_Init().
-+ Not supported on 10G-MAC (i.e. CRC & PAD are added automatically
-+ by HW); on mEMAC, this routine supports only PAD (i.e. CRC is
-+ added automatically by HW).
-+*//***************************************************************************/
-+t_Error FM_MAC_ConfigPadAndCrc(t_Handle h_FmMac, bool enable);
-+
-+/**************************************************************************//**
-+ @Function FM_MAC_ConfigHalfDuplex
-+
-+ @Description Config Half Duplex Mode
-+
-+ @Param[in] h_FmMac A handle to a FM MAC Module.
-+ @Param[in] enable TRUE to enable or FALSE to disable.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MAC_Config() and before FM_MAC_Init().
-+*//***************************************************************************/
-+t_Error FM_MAC_ConfigHalfDuplex(t_Handle h_FmMac, bool enable);
-+
-+/**************************************************************************//**
-+ @Function FM_MAC_ConfigTbiPhyAddr
-+
-+ @Description Configures the address of internal TBI PHY.
-+
-+ @Param[in] h_FmMac A handle to a FM MAC Module.
-+ @Param[in] newVal TBI PHY address (1-31).
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MAC_Config() and before FM_MAC_Init().
-+*//***************************************************************************/
-+t_Error FM_MAC_ConfigTbiPhyAddr(t_Handle h_FmMac, uint8_t newVal);
-+
-+/**************************************************************************//**
-+ @Function FM_MAC_ConfigLengthCheck
-+
-+ @Description Configure the frame length checking.
-+
-+ @Param[in] h_FmMac A handle to a FM MAC Module.
-+ @Param[in] enable TRUE to enable or FALSE to disable.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MAC_Config() and before FM_MAC_Init().
-+*//***************************************************************************/
-+t_Error FM_MAC_ConfigLengthCheck(t_Handle h_FmMac, bool enable);
-+
-+/**************************************************************************//**
-+ @Function FM_MAC_ConfigException
-+
-+ @Description Change Exception selection from default
-+
-+ @Param[in] h_FmMac A handle to a FM MAC Module.
-+ @Param[in] ex Type of the desired exceptions
-+ @Param[in] enable TRUE to enable the specified exception, FALSE to disable it.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MAC_Config() and before FM_MAC_Init().
-+*//***************************************************************************/
-+t_Error FM_MAC_ConfigException(t_Handle h_FmMac, e_FmMacExceptions ex, bool enable);
-+
-+#ifdef FM_TX_ECC_FRMS_ERRATA_10GMAC_A004
-+t_Error FM_MAC_ConfigSkipFman11Workaround (t_Handle h_FmMac);
-+#endif /* FM_TX_ECC_FRMS_ERRATA_10GMAC_A004 */
-+/** @} */ /* end of FM_mac_advanced_init_grp group */
-+/** @} */ /* end of FM_mac_init_grp group */
-+
-+
-+/**************************************************************************//**
-+ @Group FM_mac_runtime_control_grp FM MAC Runtime Control Unit
-+
-+ @Description FM MAC Runtime control unit API functions, definitions and enums.
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Function FM_MAC_Enable
-+
-+ @Description Enable the MAC
-+
-+ @Param[in] h_FmMac A handle to a FM MAC Module.
-+ @Param[in] mode Mode of operation (RX, TX, Both)
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MAC_Init().
-+*//***************************************************************************/
-+t_Error FM_MAC_Enable(t_Handle h_FmMac, e_CommMode mode);
-+
-+/**************************************************************************//**
-+ @Function FM_MAC_Disable
-+
-+ @Description DISABLE the MAC
-+
-+ @Param[in] h_FmMac A handle to a FM MAC Module.
-+ @Param[in] mode Define what part to Disable (RX, TX or BOTH)
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MAC_Init().
-+*//***************************************************************************/
-+t_Error FM_MAC_Disable(t_Handle h_FmMac, e_CommMode mode);
-+
-+/**************************************************************************//**
-+ @Function FM_MAC_Resume
-+
-+ @Description Re-init the MAC after suspend
-+
-+ @Param[in] h_FmMac A handle to a FM MAC Module.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MAC_Init().
-+*//***************************************************************************/
-+t_Error FM_MAC_Resume(t_Handle h_FmMac);
-+
-+/**************************************************************************//**
-+ @Function FM_MAC_Enable1588TimeStamp
-+
-+ @Description Enables the TSU operation.
-+
-+ @Param[in] h_Fm - Handle to the PTP as returned from the FM_MAC_PtpConfig.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MAC_Init().
-+*//***************************************************************************/
-+t_Error FM_MAC_Enable1588TimeStamp(t_Handle h_Fm);
-+
-+/**************************************************************************//**
-+ @Function FM_MAC_Disable1588TimeStamp
-+
-+ @Description Disables the TSU operation.
-+
-+ @Param[in] h_Fm - Handle to the PTP as returned from the FM_MAC_PtpConfig.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MAC_Init().
-+*//***************************************************************************/
-+t_Error FM_MAC_Disable1588TimeStamp(t_Handle h_Fm);
-+
-+/**************************************************************************//**
-+ @Function FM_MAC_SetTxAutoPauseFrames
-+
-+ @Description Enable/Disable transmission of Pause-Frames.
-+ The routine changes the default configuration [DEFAULT_TX_PAUSE_TIME].
-+
-+ @Param[in] h_FmMac - A handle to a FM MAC Module.
-+ @Param[in] pauseTime - Pause quanta value used with transmitted pause frames.
-+ Each quanta represents a 512 bit-times; Note that '0'
-+ as an input here will be used as disabling the
-+ transmission of the pause-frames.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MAC_Init().
-+*//***************************************************************************/
-+t_Error FM_MAC_SetTxAutoPauseFrames(t_Handle h_FmMac,
-+ uint16_t pauseTime);
-+
-+ /**************************************************************************//**
-+ @Function FM_MAC_SetTxPauseFrames
-+
-+ @Description Enable/Disable transmission of Pause-Frames.
-+ The routine changes the default configuration:
-+ pause-time - [DEFAULT_TX_PAUSE_TIME]
-+ threshold-time - [0]
-+
-+ @Param[in] h_FmMac - A handle to a FM MAC Module.
-+ @Param[in] priority - the PFC class of service; use 'FM_MAC_NO_PFC'
-+ to indicate legacy pause support (i.e. no PFC).
-+ @Param[in] pauseTime - Pause quanta value used with transmitted pause frames.
-+ Each quanta represents a 512 bit-times;
-+ Note that '0' as an input here will be used as disabling the
-+ transmission of the pause-frames.
-+ @Param[in] threshTime - Pause Threshold equanta value used by the MAC to retransmit pause frame.
-+ if the situation causing a pause frame to be sent didn't finish when the timer
-+ reached the threshold quanta, the MAC will retransmit the pause frame.
-+ Each quanta represents a 512 bit-times.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MAC_Init().
-+ In order for PFC to work properly the user must configure
-+ TNUM-aging in the tx-port it is recommended that pre-fetch and
-+ rate limit in the tx port should be disabled;
-+ PFC is supported only on new mEMAC; i.e. in MACs that don't have
-+ PFC support (10G-MAC and dTSEC), user should use 'FM_MAC_NO_PFC'
-+ in the 'priority' field.
-+*//***************************************************************************/
-+t_Error FM_MAC_SetTxPauseFrames(t_Handle h_FmMac,
-+ uint8_t priority,
-+ uint16_t pauseTime,
-+ uint16_t threshTime);
-+
-+/**************************************************************************//**
-+ @Function FM_MAC_SetRxIgnorePauseFrames
-+
-+ @Description Enable/Disable ignoring of Pause-Frames.
-+
-+ @Param[in] h_FmMac - A handle to a FM MAC Module.
-+ @Param[in] en - boolean indicates whether to ignore the incoming pause
-+ frames or not.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MAC_Init().
-+*//***************************************************************************/
-+t_Error FM_MAC_SetRxIgnorePauseFrames(t_Handle h_FmMac, bool en);
-+
-+/**************************************************************************//**
-+ @Function FM_MAC_SetWakeOnLan
-+
-+ @Description Enable/Disable Wake On Lan support
-+
-+ @Param[in] h_FmMac - A handle to a FM MAC Module.
-+ @Param[in] en - boolean indicates whether to enable Wake On Lan
-+ support or not.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MAC_Init().
-+*//***************************************************************************/
-+t_Error FM_MAC_SetWakeOnLan(t_Handle h_FmMac, bool en);
-+
-+/**************************************************************************//**
-+ @Function FM_MAC_ResetCounters
-+
-+ @Description reset all statistics counters
-+
-+ @Param[in] h_FmMac - A handle to a FM MAC Module.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MAC_Init().
-+*//***************************************************************************/
-+t_Error FM_MAC_ResetCounters(t_Handle h_FmMac);
-+
-+/**************************************************************************//**
-+ @Function FM_MAC_SetException
-+
-+ @Description Enable/Disable a specific Exception
-+
-+ @Param[in] h_FmMac - A handle to a FM MAC Module.
-+ @Param[in] ex - Type of the desired exceptions
-+ @Param[in] enable - TRUE to enable the specified exception, FALSE to disable it.
-+
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MAC_Init().
-+*//***************************************************************************/
-+t_Error FM_MAC_SetException(t_Handle h_FmMac, e_FmMacExceptions ex, bool enable);
-+
-+/**************************************************************************//**
-+ @Function FM_MAC_SetStatistics
-+
-+ @Description Define Statistics level.
-+ Where applicable, the routine also enables the MIB counters
-+ overflow interrupt in order to keep counters accurate
-+ and account for overflows.
-+ This routine is relevant only for dTSEC.
-+
-+ @Param[in] h_FmMac - A handle to a FM MAC Module.
-+ @Param[in] statisticsLevel - Full statistics level provides all standard counters but may
-+ reduce performance. Partial statistics provides only special
-+ event counters (errors etc.). If selected, regular counters (such as
-+ byte/packet) will be invalid and will return -1.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MAC_Init().
-+*//***************************************************************************/
-+t_Error FM_MAC_SetStatistics(t_Handle h_FmMac, e_FmMacStatisticsLevel statisticsLevel);
-+
-+/**************************************************************************//**
-+ @Function FM_MAC_GetStatistics
-+
-+ @Description get all statistics counters
-+
-+ @Param[in] h_FmMac - A handle to a FM MAC Module.
-+ @Param[in] p_Statistics - Structure with statistics
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_Init().
-+*//***************************************************************************/
-+t_Error FM_MAC_GetStatistics(t_Handle h_FmMac, t_FmMacStatistics *p_Statistics);
-+
-+/**************************************************************************//**
-+ @Function FM_MAC_GetFrameSizeCounters
-+
-+ @Description get MAC statistics counters for different frame size
-+
-+ @Param[in] h_FmMac - A handle to a FM MAC Module.
-+ @Param[in] p_FrameSizeCounters - Structure with counters
-+ @Param[in] type - Type of counters to be read
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_Init().
-+*//***************************************************************************/
-+t_Error FM_MAC_GetFrameSizeCounters(t_Handle h_FmMac, t_FmMacFrameSizeCounters *p_FrameSizeCounters, e_CommMode type);
-+
-+/**************************************************************************//**
-+ @Function FM_MAC_ModifyMacAddr
-+
-+ @Description Replace the main MAC Address
-+
-+ @Param[in] h_FmMac - A handle to a FM Module.
-+ @Param[in] p_EnetAddr - Ethernet Mac address
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only after FM_MAC_Init().
-+*//***************************************************************************/
-+t_Error FM_MAC_ModifyMacAddr(t_Handle h_FmMac, t_EnetAddr *p_EnetAddr);
-+
-+/**************************************************************************//**
-+ @Function FM_MAC_AddHashMacAddr
-+
-+ @Description Add an Address to the hash table. This is for filter purpose only.
-+
-+ @Param[in] h_FmMac - A handle to a FM Module.
-+ @Param[in] p_EnetAddr - Ethernet Mac address
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MAC_Init(). It is a filter only address.
-+ @Cautions Some address need to be filterd out in upper FM blocks.
-+*//***************************************************************************/
-+t_Error FM_MAC_AddHashMacAddr(t_Handle h_FmMac, t_EnetAddr *p_EnetAddr);
-+
-+/**************************************************************************//**
-+ @Function FM_MAC_RemoveHashMacAddr
-+
-+ @Description Delete an Address to the hash table. This is for filter purpose only.
-+
-+ @Param[in] h_FmMac - A handle to a FM Module.
-+ @Param[in] p_EnetAddr - Ethernet Mac address
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MAC_Init().
-+*//***************************************************************************/
-+t_Error FM_MAC_RemoveHashMacAddr(t_Handle h_FmMac, t_EnetAddr *p_EnetAddr);
-+
-+/**************************************************************************//**
-+ @Function FM_MAC_AddExactMatchMacAddr
-+
-+ @Description Add a unicast or multicast mac address for exact-match filtering
-+ (8 on dTSEC, 2 for 10G-MAC)
-+
-+ @Param[in] h_FmMac - A handle to a FM Module.
-+ @Param[in] p_EnetAddr - MAC Address to ADD
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only after FM_MAC_Init().
-+*//***************************************************************************/
-+t_Error FM_MAC_AddExactMatchMacAddr(t_Handle h_FmMac, t_EnetAddr *p_EnetAddr);
-+
-+/**************************************************************************//**
-+ @Function FM_MAC_RemovelExactMatchMacAddr
-+
-+ @Description Remove a uni cast or multi cast mac address.
-+
-+ @Param[in] h_FmMac - A handle to a FM Module.
-+ @Param[in] p_EnetAddr - MAC Address to remove
-+
-+ @Return E_OK on success; Error code otherwise..
-+
-+ @Cautions Allowed only after FM_MAC_Init().
-+*//***************************************************************************/
-+t_Error FM_MAC_RemovelExactMatchMacAddr(t_Handle h_FmMac, t_EnetAddr *p_EnetAddr);
-+
-+/**************************************************************************//**
-+ @Function FM_MAC_SetPromiscuous
-+
-+ @Description Enable/Disable MAC Promiscuous mode for ALL mac addresses.
-+
-+ @Param[in] h_FmMac - A handle to a FM MAC Module.
-+ @Param[in] enable - TRUE to enable or FALSE to disable.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only after FM_MAC_Init().
-+*//***************************************************************************/
-+t_Error FM_MAC_SetPromiscuous(t_Handle h_FmMac, bool enable);
-+
-+/**************************************************************************//**
-+ @Function FM_MAC_AdjustLink
-+
-+ @Description Adjusts the Ethernet link with new speed/duplex setup.
-+ This routine is relevant for dTSEC and mEMAC.
-+ In case of mEMAC, this routine is also used for manual
-+ re-configuration of RGMII speed and duplex mode for
-+ RGMII PHYs not supporting in-band status information
-+ to MAC.
-+
-+ @Param[in] h_FmMac - A handle to a FM Module.
-+ @Param[in] speed - Ethernet speed.
-+ @Param[in] fullDuplex - TRUE for full-duplex mode;
-+ FALSE for half-duplex mode.
-+
-+ @Return E_OK on success; Error code otherwise.
-+*//***************************************************************************/
-+t_Error FM_MAC_AdjustLink(t_Handle h_FmMac, e_EnetSpeed speed, bool fullDuplex);
-+
-+/**************************************************************************//**
-+ @Function FM_MAC_RestartAutoneg
-+
-+ @Description Restarts the auto-negotiation process.
-+ When auto-negotiation process is invoked under traffic the
-+ auto-negotiation process between the internal SGMII PHY and the
-+ external PHY does not always complete successfully. Calling this
-+ function will restart the auto-negotiation process that will end
-+ successfully. It is recommended to call this function after issuing
-+ auto-negotiation restart command to the Eth Phy.
-+ This routine is relevant only for dTSEC.
-+
-+ @Param[in] h_FmMac - A handle to a FM Module.
-+
-+ @Return E_OK on success; Error code otherwise.
-+*//***************************************************************************/
-+t_Error FM_MAC_RestartAutoneg(t_Handle h_FmMac);
-+
-+/**************************************************************************//**
-+ @Function FM_MAC_GetId
-+
-+ @Description Return the MAC ID
-+
-+ @Param[in] h_FmMac - A handle to a FM Module.
-+ @Param[out] p_MacId - MAC ID of device
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only after FM_MAC_Init().
-+*//***************************************************************************/
-+t_Error FM_MAC_GetId(t_Handle h_FmMac, uint32_t *p_MacId);
-+
-+/**************************************************************************//**
-+ @Function FM_MAC_GetVesrion
-+
-+ @Description Return Mac HW chip version
-+
-+ @Param[in] h_FmMac - A handle to a FM Module.
-+ @Param[out] p_MacVresion - Mac version as defined by the chip
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only after FM_MAC_Init().
-+*//***************************************************************************/
-+t_Error FM_MAC_GetVesrion(t_Handle h_FmMac, uint32_t *p_MacVresion);
-+
-+/**************************************************************************//**
-+ @Function FM_MAC_MII_WritePhyReg
-+
-+ @Description Write data into Phy Register
-+
-+ @Param[in] h_FmMac - A handle to a FM Module.
-+ @Param[in] phyAddr - Phy Address on the MII bus
-+ @Param[in] reg - Register Number.
-+ @Param[in] data - Data to write.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only after FM_MAC_Init().
-+*//***************************************************************************/
-+t_Error FM_MAC_MII_WritePhyReg(t_Handle h_FmMac, uint8_t phyAddr, uint8_t reg, uint16_t data);
-+
-+/**************************************************************************//**
-+ @Function FM_MAC_MII_ReadPhyReg
-+
-+ @Description Read data from Phy Register
-+
-+ @Param[in] h_FmMac - A handle to a FM Module.
-+ @Param[in] phyAddr - Phy Address on the MII bus
-+ @Param[in] reg - Register Number.
-+ @Param[out] p_Data - Data from PHY.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only after FM_MAC_Init().
-+*//***************************************************************************/
-+t_Error FM_MAC_MII_ReadPhyReg(t_Handle h_FmMac, uint8_t phyAddr, uint8_t reg, uint16_t *p_Data);
-+
-+#if (defined(DEBUG_ERRORS) && (DEBUG_ERRORS > 0))
-+/**************************************************************************//**
-+ @Function FM_MAC_DumpRegs
-+
-+ @Description Dump internal registers
-+
-+ @Param[in] h_FmMac - A handle to a FM Module.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only after FM_MAC_Init().
-+*//***************************************************************************/
-+t_Error FM_MAC_DumpRegs(t_Handle h_FmMac);
-+#endif /* (defined(DEBUG_ERRORS) && ... */
-+
-+/** @} */ /* end of FM_mac_runtime_control_grp group */
-+/** @} */ /* end of FM_mac_grp group */
-+/** @} */ /* end of FM_grp group */
-+
-+
-+#endif /* __FM_MAC_EXT_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/Peripherals/fm_macsec_ext.h
-@@ -0,0 +1,1271 @@
-+/*
-+ * Copyright 2008-2015 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+/**************************************************************************//**
-+ @File fm_macsec_ext.h
-+
-+ @Description FM MACSEC ...
-+*//***************************************************************************/
-+#ifndef __FM_MACSEC_EXT_H
-+#define __FM_MACSEC_EXT_H
-+
-+#include "std_ext.h"
-+
-+
-+/**************************************************************************//**
-+ @Group FM_grp Frame Manager API
-+
-+ @Description FM API functions, definitions and enums
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Group FM_MACSEC_grp FM MACSEC
-+
-+ @Description FM MACSEC API functions, definitions and enums
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Description MACSEC Exceptions
-+*//***************************************************************************/
-+typedef enum e_FmMacsecExceptions {
-+ e_FM_MACSEC_EX_SINGLE_BIT_ECC, /**< Single bit ECC error */
-+ e_FM_MACSEC_EX_MULTI_BIT_ECC /**< Multi bit ECC error */
-+} e_FmMacsecExceptions;
-+
-+
-+/**************************************************************************//**
-+ @Group FM_MACSEC_init_grp FM-MACSEC Initialization Unit
-+
-+ @Description FM MACSEC Initialization Unit
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Function t_FmMacsecExceptionsCallback
-+
-+ @Description Exceptions user callback routine, will be called upon an
-+ exception passing the exception identification.
-+
-+ @Param[in] h_App A handle to an application layer object; This handle
-+ will be passed by the driver upon calling this callback.
-+ @Param[in] exception The exception.
-+*//***************************************************************************/
-+typedef void (t_FmMacsecExceptionsCallback) ( t_Handle h_App,
-+ e_FmMacsecExceptions exception);
-+
-+
-+/**************************************************************************//**
-+ @Description FM MACSEC config input
-+*//***************************************************************************/
-+typedef struct t_FmMacsecParams {
-+ t_Handle h_Fm; /**< A handle to the FM object related to */
-+ bool guestMode; /**< Partition-id */
-+ union {
-+ struct {
-+ uint8_t fmMacId; /**< FM MAC id */
-+ } guestParams;
-+
-+ struct {
-+ uintptr_t baseAddr; /**< Base of memory mapped FM MACSEC registers */
-+ t_Handle h_FmMac; /**< A handle to the FM MAC object related to */
-+ t_FmMacsecExceptionsCallback *f_Exception; /**< Exception Callback Routine */
-+ t_Handle h_App; /**< A handle to an application layer object; This handle will
-+ be passed by the driver upon calling the above callbacks */
-+ } nonGuestParams;
-+ };
-+} t_FmMacsecParams;
-+
-+/**************************************************************************//**
-+ @Function FM_MACSEC_Config
-+
-+ @Description Creates descriptor for the FM MACSEC module;
-+
-+ The routine returns a handle (descriptor) to the FM MACSEC object;
-+ This descriptor must be passed as first parameter to all other
-+ FM MACSEC function calls;
-+
-+ No actual initialization or configuration of FM MACSEC hardware is
-+ done by this routine.
-+
-+ @Param[in] p_FmMacsecParam Pointer to data structure of parameters.
-+
-+ @Retval Handle to FM MACSEC object, or NULL for Failure.
-+*//***************************************************************************/
-+t_Handle FM_MACSEC_Config(t_FmMacsecParams *p_FmMacsecParam);
-+
-+/**************************************************************************//**
-+ @Function FM_MACSEC_Init
-+
-+ @Description Initializes the FM MACSEC module.
-+
-+ @Param[in] h_FmMacsec FM MACSEC module descriptor.
-+
-+ @Return E_OK on success; Error code otherwise.
-+*//***************************************************************************/
-+t_Error FM_MACSEC_Init(t_Handle h_FmMacsec);
-+
-+/**************************************************************************//**
-+ @Function FM_MACSEC_Free
-+
-+ @Description Frees all resources that were assigned to FM MACSEC module;
-+
-+ Calling this routine invalidates the descriptor.
-+
-+ @Param[in] h_FmMacsec FM MACSEC module descriptor.
-+
-+ @Return E_OK on success; Error code otherwise.
-+*//***************************************************************************/
-+t_Error FM_MACSEC_Free(t_Handle h_FmMacsec);
-+
-+
-+/**************************************************************************//**
-+ @Group FM_MACSEC_advanced_init_grp FM-MACSEC Advanced Configuration Unit
-+
-+ @Description Configuration functions used to change default values.
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Description enum for unknown sci frame treatment
-+*//***************************************************************************/
-+typedef enum e_FmMacsecUnknownSciFrameTreatment {
-+ e_FM_MACSEC_UNKNOWN_SCI_FRAME_TREATMENT_DISCARD_BOTH = 0, /**< Controlled port - Strict mode */
-+ e_FM_MACSEC_UNKNOWN_SCI_FRAME_TREATMENT_DISCARD_UNCONTROLLED_DELIVER_OR_DISCARD_CONTROLLED, /**< If C bit clear deliver on controlled port, else discard
-+ Controlled port - Check or Disable mode */
-+ e_FM_MACSEC_UNKNOWN_SCI_FRAME_TREATMENT_DELIVER_UNCONTROLLED_DISCARD_CONTROLLED, /**< Controlled port - Strict mode */
-+ e_FM_MACSEC_UNKNOWN_SCI_FRAME_TREATMENT_DELIVER_OR_DISCARD_UNCONTROLLED_DELIVER_OR_DISCARD_CONTROLLED /**< If C bit set deliver on uncontrolled port and discard on controlled port,
-+ else discard on uncontrolled port and deliver on controlled port
-+ Controlled port - Check or Disable mode */
-+} e_FmMacsecUnknownSciFrameTreatment;
-+
-+/**************************************************************************//**
-+ @Description enum for untag frame treatment
-+*//***************************************************************************/
-+typedef enum e_FmMacsecUntagFrameTreatment {
-+ e_FM_MACSEC_UNTAG_FRAME_TREATMENT_DELIVER_UNCONTROLLED_DISCARD_CONTROLLED = 0, /**< Controlled port - Strict mode */
-+ e_FM_MACSEC_UNTAG_FRAME_TREATMENT_DISCARD_BOTH, /**< Controlled port - Strict mode */
-+ e_FM_MACSEC_UNTAG_FRAME_TREATMENT_DISCARD_UNCONTROLLED_DELIVER_CONTROLLED_UNMODIFIED /**< Controlled port - Strict mode */
-+} e_FmMacsecUntagFrameTreatment;
-+
-+/**************************************************************************//**
-+ @Function FM_MACSEC_ConfigUnknownSciFrameTreatment
-+
-+ @Description Change the treatment for received frames with unknown sci from its default
-+ configuration [DEFAULT_unknownSciFrameTreatment].
-+
-+ @Param[in] h_FmMacsec FM MACSEC module descriptor.
-+ @Param[in] treatMode The selected mode.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MACSEC_Config() and before FM_MACSEC_Init().
-+*//***************************************************************************/
-+t_Error FM_MACSEC_ConfigUnknownSciFrameTreatment(t_Handle h_FmMacsec, e_FmMacsecUnknownSciFrameTreatment treatMode);
-+
-+/**************************************************************************//**
-+ @Function FM_MACSEC_ConfigInvalidTagsFrameTreatment
-+
-+ @Description Change the treatment for received frames with invalid tags or
-+ a zero value PN or an invalid ICV from its default configuration
-+ [DEFAULT_invalidTagsFrameTreatment].
-+
-+ @Param[in] h_FmMacsec FM MACSEC module descriptor.
-+ @Param[in] deliverUncontrolled If True deliver on the uncontrolled port, else discard;
-+ In both cases discard on the controlled port;
-+ this provide Strict, Check or Disable mode.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MACSEC_Config() and before FM_MACSEC_Init().
-+*//***************************************************************************/
-+t_Error FM_MACSEC_ConfigInvalidTagsFrameTreatment(t_Handle h_FmMacsec, bool deliverUncontrolled);
-+
-+/**************************************************************************//**
-+ @Function FM_MACSEC_ConfigEncryptWithNoChangedTextFrameTreatment
-+
-+ @Description Change the treatment for received frames with the Encryption bit
-+ set and the Changed Text bit clear from its default configuration
-+ [DEFAULT_encryptWithNoChangedTextFrameTreatment].
-+
-+ @Param[in] h_FmMacsec FM MACSEC module descriptor.
-+ @Param[in] discardUncontrolled If True discard on the uncontrolled port, else deliver;
-+ In both cases discard on the controlled port;
-+ this provide Strict, Check or Disable mode.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MACSEC_Config() and before FM_MACSEC_Init().
-+*//***************************************************************************/
-+t_Error FM_MACSEC_ConfigEncryptWithNoChangedTextFrameTreatment(t_Handle h_FmMacsec, bool discardUncontrolled);
-+
-+/**************************************************************************//**
-+ @Function FM_MACSEC_ConfigChangedTextWithNoEncryptFrameTreatment
-+
-+ @Description Change the treatment for received frames with the Encryption bit
-+ clear and the Changed Text bit set from its default configuration
-+ [DEFAULT_changedTextWithNoEncryptFrameTreatment].
-+
-+ @Param[in] h_FmMacsec FM MACSEC module descriptor.
-+ @Param[in] deliverUncontrolled If True deliver on the uncontrolled port, else discard;
-+ In both cases discard on the controlled port;
-+ this provide Strict, Check or Disable mode.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MACSEC_Config() and before FM_MACSEC_Init().
-+*//***************************************************************************/
-+t_Error FM_MACSEC_ConfigChangedTextWithNoEncryptFrameTreatment(t_Handle h_FmMacsec, bool deliverUncontrolled);
-+
-+/**************************************************************************//**
-+ @Function FM_MACSEC_ConfigUntagFrameTreatment
-+
-+ @Description Change the treatment for received frames without the MAC security tag (SecTAG)
-+ from its default configuration [DEFAULT_untagFrameTreatment].
-+
-+ @Param[in] h_FmMacsec FM MACSEC module descriptor.
-+ @Param[in] treatMode The selected mode.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MACSEC_Config() and before FM_MACSEC_Init().
-+*//***************************************************************************/
-+t_Error FM_MACSEC_ConfigUntagFrameTreatment(t_Handle h_FmMacsec, e_FmMacsecUntagFrameTreatment treatMode);
-+
-+/**************************************************************************//**
-+ @Function FM_MACSEC_ConfigOnlyScbIsSetFrameTreatment
-+
-+ @Description Change the treatment for received frames with only SCB bit set
-+ from its default configuration [DEFAULT_onlyScbIsSetFrameTreatment].
-+
-+ @Param[in] h_FmMacsec FM MACSEC module descriptor.
-+ @Param[in] deliverUncontrolled If True deliver on the uncontrolled port, else discard;
-+ In both cases discard on the controlled port;
-+ this provide Strict, Check or Disable mode.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MACSEC_Config() and before FM_MACSEC_Init().
-+*//***************************************************************************/
-+t_Error FM_MACSEC_ConfigOnlyScbIsSetFrameTreatment(t_Handle h_FmMacsec, bool deliverUncontrolled);
-+
-+/**************************************************************************//**
-+ @Function FM_MACSEC_ConfigPnExhaustionThreshold
-+
-+ @Description It's provide the ability to configure a PN exhaustion threshold;
-+ When the NextPn crosses this value an interrupt event
-+ is asserted to warn that the active SA should re-key.
-+
-+ @Param[in] h_FmMacsec FM MACSEC module descriptor.
-+ @Param[in] pnExhThr If the threshold is reached, an interrupt event
-+ is asserted to re-key.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MACSEC_Config() and before FM_MACSEC_Init().
-+*//***************************************************************************/
-+t_Error FM_MACSEC_ConfigPnExhaustionThreshold(t_Handle h_FmMacsec, uint32_t pnExhThr);
-+
-+/**************************************************************************//**
-+ @Function FM_MACSEC_ConfigKeysUnreadable
-+
-+ @Description Turn on privacy mode; All the keys and their hash values can't be read any more;
-+ Can not be cleared unless hard reset.
-+
-+ @Param[in] h_FmMacsec FM MACSEC module descriptor.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MACSEC_Config() and before FM_MACSEC_Init().
-+*//***************************************************************************/
-+t_Error FM_MACSEC_ConfigKeysUnreadable(t_Handle h_FmMacsec);
-+
-+/**************************************************************************//**
-+ @Function FM_MACSEC_ConfigSectagWithoutSCI
-+
-+ @Description Promise that all generated Sectag will be without SCI included.
-+
-+ @Param[in] h_FmMacsec FM MACSEC module descriptor.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MACSEC_Config() and before FM_MACSEC_Init().
-+*//***************************************************************************/
-+t_Error FM_MACSEC_ConfigSectagWithoutSCI(t_Handle h_FmMacsec);
-+
-+/**************************************************************************//**
-+ @Function FM_MACSEC_ConfigException
-+
-+ @Description Calling this routine changes the internal driver data base
-+ from its default selection of exceptions enablement;
-+ By default all exceptions are enabled.
-+
-+ @Param[in] h_FmMacsec FM MACSEC module descriptor.
-+ @Param[in] exception The exception to be selected.
-+ @Param[in] enable TRUE to enable interrupt, FALSE to mask it.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MACSEC_Config() and before FM_MACSEC_Init().
-+*//***************************************************************************/
-+t_Error FM_MACSEC_ConfigException(t_Handle h_FmMacsec, e_FmMacsecExceptions exception, bool enable);
-+
-+/** @} */ /* end of FM_MACSEC_advanced_init_grp group */
-+/** @} */ /* end of FM_MACSEC_init_grp group */
-+
-+
-+/**************************************************************************//**
-+ @Group FM_MACSEC_runtime_control_grp FM-MACSEC Runtime Control Data Unit
-+
-+ @Description FM MACSEC runtime control data unit API functions, definitions and enums.
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Function FM_MACSEC_GetRevision
-+
-+ @Description Return MACSEC HW chip revision
-+
-+ @Param[in] h_FmMacsec FM MACSEC module descriptor.
-+ @Param[out] p_MacsecRevision MACSEC revision as defined by the chip.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only after FM_MACSEC_Init().
-+*//***************************************************************************/
-+t_Error FM_MACSEC_GetRevision(t_Handle h_FmMacsec, uint32_t *p_MacsecRevision);
-+
-+/**************************************************************************//**
-+ @Function FM_MACSEC_Enable
-+
-+ @Description This routine should be called after MACSEC is initialized for enabling all
-+ MACSEC engines according to their existing configuration.
-+
-+ @Param[in] h_FmMacsec FM MACSEC module descriptor.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MACSEC_Init() and when MACSEC is disabled.
-+*//***************************************************************************/
-+t_Error FM_MACSEC_Enable(t_Handle h_FmMacsec);
-+
-+/**************************************************************************//**
-+ @Function FM_MACSEC_Disable
-+
-+ @Description This routine may be called when MACSEC is enabled in order to
-+ disable all MACSEC engines; The MACSEC is working in bypass mode.
-+
-+ @Param[in] h_FmMacsec FM MACSEC module descriptor.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MACSEC_Init() and when MACSEC is enabled.
-+*//***************************************************************************/
-+t_Error FM_MACSEC_Disable(t_Handle h_FmMacsec);
-+
-+/**************************************************************************//**
-+ @Function FM_MACSEC_SetException
-+
-+ @Description Calling this routine enables/disables the specified exception.
-+
-+ @Param[in] h_FmMacsec FM MACSEC module descriptor.
-+ @Param[in] exception The exception to be selected.
-+ @Param[in] enable TRUE to enable interrupt, FALSE to mask it.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MACSEC_Init().
-+*//***************************************************************************/
-+t_Error FM_MACSEC_SetException(t_Handle h_FmMacsec, e_FmMacsecExceptions exception, bool enable);
-+
-+#if (defined(DEBUG_ERRORS) && (DEBUG_ERRORS > 0))
-+/**************************************************************************//**
-+ @Function FM_MACSEC_DumpRegs
-+
-+ @Description Dump internal registers.
-+
-+ @Param[in] h_FmMacsec - FM MACSEC module descriptor.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only after FM_MACSEC_Init().
-+*//***************************************************************************/
-+t_Error FM_MACSEC_DumpRegs(t_Handle h_FmMacsec);
-+#endif /* (defined(DEBUG_ERRORS) && ... */
-+
-+#ifdef VERIFICATION_SUPPORT
-+/********************* VERIFICATION ONLY ********************************/
-+/**************************************************************************//**
-+ @Function FM_MACSEC_BackdoorSet
-+
-+ @Description Set register of the MACSEC memory map
-+
-+ @Param[in] h_FmMacsec FM MACSEC module descriptor.
-+ @Param[out] offset Register offset.
-+ @Param[out] value Value to write.
-+
-+
-+ @Return None
-+
-+ @Cautions Allowed only following FM_MACSEC_Init().
-+*//***************************************************************************/
-+t_Error FM_MACSEC_BackdoorSet(t_Handle h_FmMacsec, uint32_t offset, uint32_t value);
-+
-+/**************************************************************************//**
-+ @Function FM_MACSEC_BackdoorGet
-+
-+ @Description Read from register of the MACSEC memory map.
-+
-+ @Param[in] h_FmMacsec FM MACSEC module descriptor.
-+ @Param[out] offset Register offset.
-+
-+ @Return Value read
-+
-+ @Cautions Allowed only following FM_MACSEC_Init().
-+*//***************************************************************************/
-+uint32_t FM_MACSEC_BackdoorGet(t_Handle h_FmMacsec, uint32_t offset);
-+#endif /* VERIFICATION_SUPPORT */
-+
-+/** @} */ /* end of FM_MACSEC_runtime_control_grp group */
-+
-+
-+/**************************************************************************//**
-+ @Group FM_MACSEC_SECY_grp FM-MACSEC SecY
-+
-+ @Description FM-MACSEC SecY API functions, definitions and enums
-+
-+ @{
-+*//***************************************************************************/
-+
-+typedef uint8_t macsecSAKey_t[32];
-+typedef uint64_t macsecSCI_t;
-+typedef uint8_t macsecAN_t;
-+
-+/**************************************************************************//**
-+@Description MACSEC SECY Cipher Suite
-+*//***************************************************************************/
-+typedef enum e_FmMacsecSecYCipherSuite {
-+ e_FM_MACSEC_SECY_GCM_AES_128 = 0, /**< GCM-AES-128 */
-+#if (DPAA_VERSION >= 11)
-+ e_FM_MACSEC_SECY_GCM_AES_256 /**< GCM-AES-256 */
-+#endif /* (DPAA_VERSION >= 11) */
-+} e_FmMacsecSecYCipherSuite;
-+
-+/**************************************************************************//**
-+ @Description MACSEC SECY Exceptions
-+*//***************************************************************************/
-+typedef enum e_FmMacsecSecYExceptions {
-+ e_FM_MACSEC_SECY_EX_FRAME_DISCARDED /**< Frame Discarded */
-+} e_FmMacsecSecYExceptions;
-+
-+/**************************************************************************//**
-+ @Description MACSEC SECY Events
-+*//***************************************************************************/
-+typedef enum e_FmMacsecSecYEvents {
-+ e_FM_MACSEC_SECY_EV_NEXT_PN /**< Next Packet Number exhaustion threshold reached */
-+} e_FmMacsecSecYEvents;
-+
-+/**************************************************************************//**
-+ @Collection MACSEC SECY Frame Discarded Descriptor error
-+*//***************************************************************************/
-+typedef uint8_t macsecTxScFrameDiscardedErrSelect_t; /**< typedef for defining Frame Discarded Descriptor errors */
-+
-+#define FM_MACSEC_SECY_TX_SC_FRM_DISCAR_ERR_NEXT_PN_ZERO 0x8000 /**< NextPn == 0 */
-+#define FM_MACSEC_SECY_TX_SC_FRM_DISCAR_ERR_SC_DISBALE 0x4000 /**< SC is disable */
-+/* @} */
-+
-+/**************************************************************************//**
-+ @Function t_FmMacsecSecYExceptionsCallback
-+
-+ @Description Exceptions user callback routine, will be called upon an
-+ exception passing the exception identification.
-+
-+ @Param[in] h_App A handle to an application layer object; This handle
-+ will be passed by the driver upon calling this callback.
-+ @Param[in] exception The exception.
-+*//***************************************************************************/
-+typedef void (t_FmMacsecSecYExceptionsCallback) ( t_Handle h_App,
-+ e_FmMacsecSecYExceptions exception);
-+
-+/**************************************************************************//**
-+ @Function t_FmMacsecSecYEventsCallback
-+
-+ @Description Events user callback routine, will be called upon an
-+ event passing the event identification.
-+
-+ @Param[in] h_App A handle to an application layer object; This handle
-+ will be passed by the driver upon calling this callback.
-+ @Param[in] event The event.
-+*//***************************************************************************/
-+typedef void (t_FmMacsecSecYEventsCallback) ( t_Handle h_App,
-+ e_FmMacsecSecYEvents event);
-+
-+/**************************************************************************//**
-+ @Description RFC2863 MIB
-+*//***************************************************************************/
-+typedef struct t_MIBStatistics {
-+ uint64_t ifInOctets; /**< Total number of byte received */
-+ uint64_t ifInPkts; /**< Total number of packets received */
-+ uint64_t ifInMcastPkts; /**< Total number of multicast frame received */
-+ uint64_t ifInBcastPkts; /**< Total number of broadcast frame received */
-+ uint64_t ifInDiscards; /**< Frames received, but discarded due to problems within the MAC RX :
-+ - InPktsNoTag,
-+ - InPktsLate,
-+ - InPktsOverrun */
-+ uint64_t ifInErrors; /**< Number of frames received with error:
-+ - InPktsBadTag,
-+ - InPktsNoSCI,
-+ - InPktsNotUsingSA
-+ - InPktsNotValid */
-+ uint64_t ifOutOctets; /**< Total number of byte sent */
-+ uint64_t ifOutPkts; /**< Total number of packets sent */
-+ uint64_t ifOutMcastPkts; /**< Total number of multicast frame sent */
-+ uint64_t ifOutBcastPkts; /**< Total number of multicast frame sent */
-+ uint64_t ifOutDiscards; /**< Frames received, but discarded due to problems within the MAC TX N/A! */
-+ uint64_t ifOutErrors; /**< Number of frames transmitted with error:
-+ - FIFO Overflow Error
-+ - FIFO Underflow Error
-+ - Other */
-+} t_MIBStatistics;
-+
-+/**************************************************************************//**
-+ @Description MACSEC SecY Rx SA Statistics
-+*//***************************************************************************/
-+typedef struct t_FmMacsecSecYRxSaStatistics {
-+ uint32_t inPktsOK; /**< The number of frames with resolved SCI, have passed all
-+ frame validation frame validation with the validateFrame not set to disable */
-+ uint32_t inPktsInvalid; /**< The number of frames with resolved SCI, that have failed frame
-+ validation with the validateFrame set to check */
-+ uint32_t inPktsNotValid; /**< The number of frames with resolved SCI, discarded on the controlled port,
-+ that have failed frame validation with the validateFrame set to strict or the c bit is set */
-+ uint32_t inPktsNotUsingSA; /**< The number of frames received with resolved SCI and discarded on disabled or
-+ not provisioned SA with validateFrame in the strict mode or the C bit is set */
-+ uint32_t inPktsUnusedSA; /**< The number of frames received with resolved SCI on disabled or not provisioned SA
-+ with validateFrame not in the strict mode and the C bit is cleared */
-+} t_FmMacsecSecYRxSaStatistics;
-+
-+/**************************************************************************//**
-+ @Description MACSEC SecY Tx SA Statistics
-+*//***************************************************************************/
-+typedef struct t_FmMacsecSecYTxSaStatistics {
-+ uint64_t outPktsProtected; /**< The number of frames, that the user of the controlled port requested to
-+ be transmitted, which were integrity protected */
-+ uint64_t outPktsEncrypted; /**< The number of frames, that the user of the controlled port requested to
-+ be transmitted, which were confidentiality protected */
-+} t_FmMacsecSecYTxSaStatistics;
-+
-+/**************************************************************************//**
-+ @Description MACSEC SecY Rx SC Statistics
-+*//***************************************************************************/
-+typedef struct t_FmMacsecSecYRxScStatistics {
-+ uint64_t inPktsUnchecked; /**< The number of frames with resolved SCI, delivered to the user of a controlled port,
-+ that are not validated with the validateFrame set to disable */
-+ uint64_t inPktsDelayed; /**< The number of frames with resolved SCI, delivered to the user of a controlled port,
-+ that have their PN smaller than the lowest_PN with the validateFrame set to
-+ disable or replayProtect disabled */
-+ uint64_t inPktsLate; /**< The number of frames with resolved SCI, discarded on the controlled port,
-+ that have their PN smaller than the lowest_PN with the validateFrame set to
-+ Check or Strict and replayProtect enabled */
-+ uint64_t inPktsOK; /**< The number of frames with resolved SCI, have passed all
-+ frame validation frame validation with the validateFrame not set to disable */
-+ uint64_t inPktsInvalid; /**< The number of frames with resolved SCI, that have failed frame
-+ validation with the validateFrame set to check */
-+ uint64_t inPktsNotValid; /**< The number of frames with resolved SCI, discarded on the controlled port,
-+ that have failed frame validation with the validateFrame set to strict or the c bit is set */
-+ uint64_t inPktsNotUsingSA; /**< The number of frames received with resolved SCI and discarded on disabled or
-+ not provisioned SA with validateFrame in the strict mode or the C bit is set */
-+ uint64_t inPktsUnusedSA; /**< The number of frames received with resolved SCI on disabled or not provisioned SA
-+ with validateFrame not in the strict mode and the C bit is cleared */
-+} t_FmMacsecSecYRxScStatistics;
-+
-+/**************************************************************************//**
-+ @Description MACSEC SecY Tx SC Statistics
-+*//***************************************************************************/
-+typedef struct t_FmMacsecSecYTxScStatistics {
-+ uint64_t outPktsProtected; /**< The number of frames, that the user of the controlled port requested to
-+ be transmitted, which were integrity protected */
-+ uint64_t outPktsEncrypted; /**< The number of frames, that the user of the controlled port requested to
-+ be transmitted, which were confidentiality protected */
-+} t_FmMacsecSecYTxScStatistics;
-+
-+/**************************************************************************//**
-+ @Description MACSEC SecY Statistics
-+*//***************************************************************************/
-+typedef struct t_FmMacsecSecYStatistics {
-+ t_MIBStatistics mibCtrlStatistics; /**< Controlled port MIB statistics */
-+ t_MIBStatistics mibNonCtrlStatistics; /**< Uncontrolled port MIB statistics */
-+/* Frame verification statistics */
-+ uint64_t inPktsUntagged; /**< The number of received packets without the MAC security tag
-+ (SecTAG) with validateFrames which is not in the strict mode */
-+ uint64_t inPktsNoTag; /**< The number of received packets discarded without the
-+ MAC security tag (SecTAG) with validateFrames which is in the strict mode */
-+ uint64_t inPktsBadTag; /**< The number of received packets discarded with an invalid
-+ SecTAG or a zero value PN or an invalid ICV */
-+ uint64_t inPktsUnknownSCI; /**< The number of received packets with unknown SCI with the
-+ condition : validateFrames is not in the strict mode and the
-+ C bit in the SecTAG is not set */
-+ uint64_t inPktsNoSCI; /**< The number of received packets discarded with unknown SCI
-+ information with the condition : validateFrames is in the strict mode
-+ or the C bit in the SecTAG is set */
-+ uint64_t inPktsOverrun; /**< The number of packets discarded because the number of
-+ received packets exceeded the cryptographic performance capabilities */
-+/* Frame validation statistics */
-+ uint64_t inOctetsValidated; /**< The number of octets of plaintext recovered from received frames with
-+ resolved SCI that were integrity protected but not encrypted */
-+ uint64_t inOctetsDecrypted; /**< The number of octets of plaintext recovered from received frames with
-+ resolved SCI that were integrity protected and encrypted */
-+/* Frame generation statistics */
-+ uint64_t outPktsUntagged; /**< The number of frames, that the user of the controlled port requested to
-+ be transmitted, with protectFrame false */
-+ uint64_t outPktsTooLong; /**< The number of frames, that the user of the controlled port requested to
-+ be transmitted, discarded due to length being larger than Maximum Frame Length (MACSEC_MFL) */
-+/* Frame protection statistics */
-+ uint64_t outOctetsProtected; /**< The number of octets of User Data in transmitted frames that were
-+ integrity protected but not encrypted */
-+ uint64_t outOctetsEncrypted; /**< The number of octets of User Data in transmitted frames that were
-+ both integrity protected and encrypted */
-+} t_FmMacsecSecYStatistics;
-+
-+
-+/**************************************************************************//**
-+ @Description MACSEC SecY SC Params
-+*//***************************************************************************/
-+typedef struct t_FmMacsecSecYSCParams {
-+ macsecSCI_t sci; /**< The secure channel identification of the SC */
-+ e_FmMacsecSecYCipherSuite cipherSuite; /**< Cipher suite to be used for the SC */
-+} t_FmMacsecSecYSCParams;
-+
-+/**************************************************************************//**
-+ @Group FM_MACSEC_SECY_init_grp FM-MACSEC SecY Initialization Unit
-+
-+ @Description FM-MACSEC SecY Initialization Unit
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Description enum for validate frames
-+*//***************************************************************************/
-+typedef enum e_FmMacsecValidFrameBehavior {
-+ e_FM_MACSEC_VALID_FRAME_BEHAVIOR_DISABLE = 0, /**< disable the validation function */
-+ e_FM_MACSEC_VALID_FRAME_BEHAVIOR_CHECK, /**< enable the validation function but only for checking
-+ without filtering out invalid frames */
-+ e_FM_MACSEC_VALID_FRAME_BEHAVIOR_STRICT /**< enable the validation function and also strictly filter
-+ out those invalid frames */
-+} e_FmMacsecValidFrameBehavior;
-+
-+/**************************************************************************//**
-+ @Description enum for sci insertion
-+*//***************************************************************************/
-+typedef enum e_FmMacsecSciInsertionMode {
-+ e_FM_MACSEC_SCI_INSERTION_MODE_EXPLICIT_SECTAG = 0, /**< explicit sci in the sectag */
-+ e_FM_MACSEC_SCI_INSERTION_MODE_EXPLICIT_MAC_SA, /**< mac sa is overwritten with the sci*/
-+ e_FM_MACSEC_SCI_INSERTION_MODE_IMPLICT_PTP /**< implicit point-to-point sci (pre-shared) */
-+} e_FmMacsecSciInsertionMode;
-+
-+/**************************************************************************//**
-+ @Description FM MACSEC SecY config input
-+*//***************************************************************************/
-+typedef struct t_FmMacsecSecYParams {
-+ t_Handle h_FmMacsec; /**< A handle to the FM MACSEC object */
-+ t_FmMacsecSecYSCParams txScParams; /**< Tx SC Params */
-+ uint32_t numReceiveChannels; /**< Number of receive channels dedicated to this SecY */
-+ t_FmMacsecSecYExceptionsCallback *f_Exception; /**< Callback routine to be called by the driver upon SecY exception */
-+ t_FmMacsecSecYEventsCallback *f_Event; /**< Callback routine to be called by the driver upon SecY event */
-+ t_Handle h_App; /**< A handle to an application layer object; This handle will
-+ be passed by the driver upon calling the above callbacks */
-+} t_FmMacsecSecYParams;
-+
-+/**************************************************************************//**
-+ @Function FM_MACSEC_SECY_Config
-+
-+ @Description Creates descriptor for the FM MACSEC SECY module;
-+
-+ The routine returns a handle (descriptor) to the FM MACSEC SECY object;
-+ This descriptor must be passed as first parameter to all other
-+ FM MACSEC SECY function calls;
-+ No actual initialization or configuration of FM MACSEC SecY hardware is
-+ done by this routine.
-+
-+ @Param[in] p_FmMacsecSecYParam Pointer to data structure of parameters.
-+
-+ @Return Handle to FM MACSEC SECY object, or NULL for Failure.
-+*//***************************************************************************/
-+t_Handle FM_MACSEC_SECY_Config(t_FmMacsecSecYParams *p_FmMacsecSecYParam);
-+
-+/**************************************************************************//**
-+ @Function FM_MACSEC_SECY_Init
-+
-+ @Description Initializes the FM MACSEC SECY module.
-+
-+ @Param[in] h_FmMacsecSecY FM MACSEC SECY module descriptor.
-+
-+ @Return E_OK on success; Error code otherwise.
-+*//***************************************************************************/
-+t_Error FM_MACSEC_SECY_Init(t_Handle h_FmMacsecSecY);
-+
-+/**************************************************************************//**
-+ @Function FM_MACSEC_SECY_Free
-+
-+ @Description Frees all resources that were assigned to FM MACSEC SECY module.
-+
-+ Calling this routine invalidates the descriptor.
-+
-+ @Param[in] h_FmMacsecSecY FM MACSEC SECY module descriptor.
-+
-+ @Return E_OK on success; Error code otherwise.
-+*//***************************************************************************/
-+t_Error FM_MACSEC_SECY_Free(t_Handle h_FmMacsecSecY);
-+
-+/**************************************************************************//**
-+ @Group FM_MACSEC_SECY_advanced_init_grp FM-MACSEC SecY Advanced Configuration Unit
-+
-+ @Description Configuration functions used to change default values.
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Function FM_MACSEC_SECY_ConfigSciInsertionMode
-+
-+ @Description Calling this routine changes the SCI-insertion-mode in the
-+ internal driver data base from its default configuration
-+ [DEFAULT_sciInsertionMode]
-+
-+ @Param[in] h_FmMacsecSecY FM MACSEC SECY module descriptor.
-+ @Param[in] sciInsertionMode Sci insertion mode
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MACSEC_SECY_Config() and before FM_MACSEC_SECY_Init();
-+
-+*//***************************************************************************/
-+t_Error FM_MACSEC_SECY_ConfigSciInsertionMode(t_Handle h_FmMacsecSecY, e_FmMacsecSciInsertionMode sciInsertionMode);
-+
-+/**************************************************************************//**
-+ @Function FM_MACSEC_SECY_ConfigProtectFrames
-+
-+ @Description Calling this routine changes the protect-frame mode in the
-+ internal driver data base from its default configuration
-+ [DEFAULT_protectFrames]
-+
-+ @Param[in] h_FmMacsecSecY FM MACSEC SECY module descriptor.
-+ @Param[in] protectFrames If FALSE, frames are transmitted without modification
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MACSEC_SECY_Config() and before FM_MACSEC_SECY_Init();
-+
-+*//***************************************************************************/
-+t_Error FM_MACSEC_SECY_ConfigProtectFrames(t_Handle h_FmMacsecSecY, bool protectFrames);
-+
-+/**************************************************************************//**
-+ @Function FM_MACSEC_SECY_ConfigReplayWindow
-+
-+ @Description Calling this routine changes the replay-window settings in the
-+ internal driver data base from its default configuration
-+ [DEFAULT_replayEnable], [DEFAULT_replayWindow]
-+
-+ @Param[in] h_FmMacsecSecY FM MACSEC SECY module descriptor.
-+ @Param[in] replayProtect; Replay protection function mode
-+ @Param[in] replayWindow; The size of the replay window
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MACSEC_SECY_Config() and before FM_MACSEC_SECY_Init();
-+
-+*//***************************************************************************/
-+t_Error FM_MACSEC_SECY_ConfigReplayWindow(t_Handle h_FmMacsecSecY, bool replayProtect, uint32_t replayWindow);
-+
-+/**************************************************************************//**
-+ @Function FM_MACSEC_SECY_ConfigValidationMode
-+
-+ @Description Calling this routine changes the frame-validation-behavior mode
-+ in the internal driver data base from its default configuration
-+ [DEFAULT_validateFrames]
-+
-+ @Param[in] h_FmMacsecSecY FM MACSEC SECY module descriptor.
-+ @Param[in] validateFrames Validation function mode
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MACSEC_SECY_Config() and before FM_MACSEC_SECY_Init();
-+
-+*//***************************************************************************/
-+t_Error FM_MACSEC_SECY_ConfigValidationMode(t_Handle h_FmMacsecSecY, e_FmMacsecValidFrameBehavior validateFrames);
-+
-+/**************************************************************************//**
-+ @Function FM_MACSEC_SECY_ConfigConfidentiality
-+
-+ @Description Calling this routine changes the confidentiality settings in the
-+ internal driver data base from its default configuration
-+ [DEFAULT_confidentialityEnable], [DEFAULT_confidentialityOffset]
-+
-+ @Param[in] h_FmMacsecSecY FM MACSEC SECY module descriptor.
-+ @Param[in] confidentialityEnable TRUE - confidentiality protection and integrity protection
-+ FALSE - no confidentiality protection, only integrity protection
-+ @Param[in] confidentialityOffset The number of initial octets of each MSDU without confidentiality protection
-+ common values are 0, 30, and 50
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MACSEC_SECY_Config() and before FM_MACSEC_SECY_Init();
-+
-+*//***************************************************************************/
-+t_Error FM_MACSEC_SECY_ConfigConfidentiality(t_Handle h_FmMacsecSecY, bool confidentialityEnable, uint16_t confidentialityOffset);
-+
-+/**************************************************************************//**
-+ @Function FM_MACSEC_SECY_ConfigPointToPoint
-+
-+ @Description configure this SecY to work in point-to-point mode, means that
-+ it will have only one rx sc;
-+
-+ @Param[in] h_FmMacsecSecY FM MACSEC SECY module descriptor.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MACSEC_SECY_Config() and before FM_MACSEC_SECY_Init();
-+ Can be called only once in a system; only the first secY that will call this
-+ routine will be able to operate in Point-To-Point mode.
-+*//***************************************************************************/
-+t_Error FM_MACSEC_SECY_ConfigPointToPoint(t_Handle h_FmMacsecSecY);
-+
-+/**************************************************************************//**
-+ @Function FM_MACSEC_SECY_ConfigException
-+
-+ @Description Calling this routine changes the internal driver data base
-+ from its default selection of exceptions enablement;
-+ By default all exceptions are enabled.
-+
-+ @Param[in] h_FmMacsecSecY FM MACSEC SECY module descriptor.
-+ @Param[in] exception The exception to be selected.
-+ @Param[in] enable TRUE to enable interrupt, FALSE to mask it.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MACSEC_SECY_Config() and before FM_MACSEC_SECY_Init().
-+*//***************************************************************************/
-+t_Error FM_MACSEC_SECY_ConfigException(t_Handle h_FmMacsecSecY, e_FmMacsecSecYExceptions exception, bool enable);
-+
-+/**************************************************************************//**
-+ @Function FM_MACSEC_SECY_ConfigEvent
-+
-+ @Description Calling this routine changes the internal driver data base
-+ from its default selection of events enablement;
-+ By default all events are enabled.
-+
-+ @Param[in] h_FmMacsecSecY FM MACSEC SECY module descriptor.
-+ @Param[in] event The event to be selected.
-+ @Param[in] enable TRUE to enable interrupt, FALSE to mask it.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MACSEC_SECY_Config() and before FM_MACSEC_SECY_Init().
-+*//***************************************************************************/
-+t_Error FM_MACSEC_SECY_ConfigEvent(t_Handle h_FmMacsecSecY, e_FmMacsecSecYEvents event, bool enable);
-+
-+/** @} */ /* end of FM_MACSEC_SECY_advanced_init_grp group */
-+/** @} */ /* end of FM_MACSEC_SECY_init_grp group */
-+
-+
-+/**************************************************************************//**
-+ @Group FM_MACSEC_SECY_runtime_control_grp FM-MACSEC SecY Runtime Control Unit
-+
-+ @Description FM MACSEC SECY Runtime control unit API functions, definitions and enums.
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Function FM_MACSEC_SECY_CreateRxSc
-+
-+ @Description Create a receive secure channel.
-+
-+ @Param[in] h_FmMacsecSecY FM MACSEC SECY module descriptor.
-+ @Param[in] scParams secure channel params.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MACSEC_SECY_Init().
-+*//***************************************************************************/
-+t_Handle FM_MACSEC_SECY_CreateRxSc(t_Handle h_FmMacsecSecY, t_FmMacsecSecYSCParams *p_ScParams);
-+
-+/**************************************************************************//**
-+ @Function FM_MACSEC_SECY_DeleteRxSc
-+
-+ @Description Deleting an initialized secure channel.
-+
-+ @Param[in] h_FmMacsecSecY FM MACSEC SECY module descriptor.
-+ @Param[in] h_Sc SC handle as returned by FM_MACSEC_SECY_CreateRxSc.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MACSEC_SECY_CreateRxSc().
-+*//***************************************************************************/
-+t_Error FM_MACSEC_SECY_DeleteRxSc(t_Handle h_FmMacsecSecY, t_Handle h_Sc);
-+
-+/**************************************************************************//**
-+ @Function FM_MACSEC_SECY_CreateRxSa
-+
-+ @Description Create a receive secure association for the secure channel;
-+ the SA cannot be used to receive frames until FM_MACSEC_SECY_RxSaEnableReceive is called.
-+
-+ @Param[in] h_FmMacsecSecY FM MACSEC SECY module descriptor.
-+ @Param[in] h_Sc SC handle as returned by FM_MACSEC_SECY_CreateRxSc.
-+ @Param[in] an association number represent the SA.
-+ @Param[in] lowestPn the lowest acceptable PN value for a received frame.
-+ @Param[in] key the desired key for this SA.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MACSEC_SECY_CreateRxSc().
-+*//***************************************************************************/
-+t_Error FM_MACSEC_SECY_CreateRxSa(t_Handle h_FmMacsecSecY, t_Handle h_Sc, macsecAN_t an, uint32_t lowestPn, macsecSAKey_t key);
-+
-+/**************************************************************************//**
-+ @Function FM_MACSEC_SECY_DeleteRxSa
-+
-+ @Description Deleting an initialized secure association.
-+
-+ @Param[in] h_FmMacsecSecY FM MACSEC SECY module descriptor.
-+ @Param[in] h_Sc SC handle as returned by FM_MACSEC_SECY_CreateRxSc.
-+ @Param[in] an association number represent the SA.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MACSEC_SECY_Init().
-+*//***************************************************************************/
-+t_Error FM_MACSEC_SECY_DeleteRxSa(t_Handle h_FmMacsecSecY, t_Handle h_Sc, macsecAN_t an);
-+
-+/**************************************************************************//**
-+ @Function FM_MACSEC_SECY_RxSaEnableReceive
-+
-+ @Description Enabling the SA to receive frames.
-+
-+ @Param[in] h_FmMacsecSecY FM MACSEC SECY module descriptor.
-+ @Param[in] h_Sc SC handle as returned by FM_MACSEC_SECY_CreateRxSc.
-+ @Param[in] an association number represent the SA.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MACSEC_SECY_CreateRxSa().
-+*//***************************************************************************/
-+t_Error FM_MACSEC_SECY_RxSaEnableReceive(t_Handle h_FmMacsecSecY, t_Handle h_Sc, macsecAN_t an);
-+
-+/**************************************************************************//**
-+ @Function FM_MACSEC_SECY_RxSaDisableReceive
-+
-+ @Description Disabling the SA from receive frames.
-+
-+ @Param[in] h_FmMacsecSecY FM MACSEC SECY module descriptor.
-+ @Param[in] h_Sc SC handle as returned by FM_MACSEC_SECY_CreateRxSc.
-+ @Param[in] an association number represent the SA.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MACSEC_SECY_CreateRxSa().
-+*//***************************************************************************/
-+t_Error FM_MACSEC_SECY_RxSaDisableReceive(t_Handle h_FmMacsecSecY, t_Handle h_Sc, macsecAN_t an);
-+
-+/**************************************************************************//**
-+ @Function FM_MACSEC_SECY_RxSaUpdateNextPn
-+
-+ @Description Update the next packet number expected on RX;
-+ The value of nextPN shall be set to the greater of its existing value and the
-+ supplied of updtNextPN (802.1AE-2006 10.7.15).
-+
-+ @Param[in] h_FmMacsecSecY FM MACSEC SECY module descriptor.
-+ @Param[in] h_Sc SC handle as returned by FM_MACSEC_SECY_CreateRxSc.
-+ @Param[in] an association number represent the SA.
-+ @Param[in] updtNextPN the next PN value for a received frame.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MACSEC_SECY_CreateRxSa().
-+*//***************************************************************************/
-+t_Error FM_MACSEC_SECY_RxSaUpdateNextPn(t_Handle h_FmMacsecSecY, t_Handle h_Sc, macsecAN_t an, uint32_t updtNextPN);
-+
-+/**************************************************************************//**
-+ @Function FM_MACSEC_SECY_RxSaUpdateLowestPn
-+
-+ @Description Update the lowest packet number expected on RX;
-+ The value of lowestPN shall be set to the greater of its existing value and the
-+ supplied of updtLowestPN (802.1AE-2006 10.7.15).
-+
-+ @Param[in] h_FmMacsecSecY FM MACSEC SECY module descriptor.
-+ @Param[in] h_Sc SC handle as returned by FM_MACSEC_SECY_CreateRxSc.
-+ @Param[in] an association number represent the SA.
-+ @Param[in] updtLowestPN the lowest PN acceptable value for a received frame.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MACSEC_SECY_CreateRxSa().
-+*//***************************************************************************/
-+t_Error FM_MACSEC_SECY_RxSaUpdateLowestPn(t_Handle h_FmMacsecSecY, t_Handle h_Sc, macsecAN_t an, uint32_t updtLowestPN);
-+
-+/**************************************************************************//**
-+ @Function FM_MACSEC_SECY_RxSaModifyKey
-+
-+ @Description Modify the current key of the SA with a new one.
-+
-+ @Param[in] h_FmMacsecSecY FM MACSEC SECY module descriptor.
-+ @Param[in] h_Sc SC handle as returned by FM_MACSEC_SECY_CreateRxSc.
-+ @Param[in] an association number represent the SA.
-+ @Param[in] key new key to replace the current key.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MACSEC_SECY_CreateRxSa().
-+*//***************************************************************************/
-+t_Error FM_MACSEC_SECY_RxSaModifyKey(t_Handle h_FmMacsecSecY, t_Handle h_Sc, macsecAN_t an, macsecSAKey_t key);
-+
-+/**************************************************************************//**
-+ @Function FM_MACSEC_SECY_CreateTxSa
-+
-+ @Description Create a transmit secure association for the secure channel;
-+ the SA cannot be used to transmit frames until FM_MACSEC_SECY_TxSaSetActivate is called;
-+ Only one SA can be active at a time.
-+
-+ @Param[in] h_FmMacsecSecY FM MACSEC SECY module descriptor.
-+ @Param[in] an association number represent the SA.
-+ @Param[in] key the desired key for this SA.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MACSEC_SECY_Init().
-+*//***************************************************************************/
-+t_Error FM_MACSEC_SECY_CreateTxSa(t_Handle h_FmMacsecSecY, macsecAN_t an, macsecSAKey_t key);
-+
-+/**************************************************************************//**
-+ @Function FM_MACSEC_SECY_DeleteTxSa
-+
-+ @Description Deleting an initialized secure association.
-+
-+ @Param[in] h_FmMacsecSecY FM MACSEC SECY module descriptor.
-+ @Param[in] an association number represent the SA.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MACSEC_SECY_Init().
-+*//***************************************************************************/
-+t_Error FM_MACSEC_SECY_DeleteTxSa(t_Handle h_FmMacsecSecY, macsecAN_t an);
-+
-+/**************************************************************************//**
-+ @Function FM_MACSEC_SECY_TxSaModifyKey
-+
-+ @Description Modify the key of the inactive SA with a new one.
-+
-+ @Param[in] h_FmMacsecSecY FM MACSEC SECY module descriptor.
-+ @Param[in] nextActiveAn association number represent the next SA to be activated.
-+ @Param[in] key new key to replace the current key.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MACSEC_SECY_Init().
-+*//***************************************************************************/
-+t_Error FM_MACSEC_SECY_TxSaModifyKey(t_Handle h_FmMacsecSecY, macsecAN_t nextActiveAn, macsecSAKey_t key);
-+
-+/**************************************************************************//**
-+ @Function FM_MACSEC_SECY_TxSaSetActive
-+
-+ @Description Set this SA to the active SA to be used on TX for SC;
-+ only one SA can be active at a time.
-+
-+ @Param[in] h_FmMacsecSecY FM MACSEC SECY module descriptor.
-+ @Param[in] an association number represent the SA.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MACSEC_SECY_Init().
-+*//***************************************************************************/
-+t_Error FM_MACSEC_SECY_TxSaSetActive(t_Handle h_FmMacsecSecY, macsecAN_t an);
-+
-+/**************************************************************************//**
-+ @Function FM_MACSEC_SECY_TxSaGetActive
-+
-+ @Description Get the active SA that being used for TX.
-+
-+ @Param[in] h_FmMacsecSecY FM MACSEC SECY module descriptor.
-+ @Param[out] p_An the active an.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MACSEC_SECY_Init().
-+*//***************************************************************************/
-+t_Error FM_MACSEC_SECY_TxSaGetActive(t_Handle h_FmMacsecSecY, macsecAN_t *p_An);
-+
-+/**************************************************************************//**
-+ @Function FM_MACSEC_SECY_GetStatistics
-+
-+ @Description get all statistics counters.
-+
-+ @Param[in] h_FmMacsecSecY FM MACSEC SECY module descriptor.
-+ @Param[in] p_Statistics Structure with statistics.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MACSEC_SECY_Init().
-+*//***************************************************************************/
-+t_Error FM_MACSEC_SECY_GetStatistics(t_Handle h_FmMacsecSecY, t_FmMacsecSecYStatistics *p_Statistics);
-+
-+/**************************************************************************//**
-+ @Function FM_MACSEC_SECY_RxScGetStatistics
-+
-+ @Description get all statistics counters.
-+
-+ @Param[in] h_FmMacsecSecY FM MACSEC SECY module descriptor.
-+ @Param[in] h_Sc Rx Sc handle.
-+ @Param[in] p_Statistics Structure with statistics.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MACSEC_SECY_Init().
-+*//***************************************************************************/
-+t_Error FM_MACSEC_SECY_RxScGetStatistics(t_Handle h_FmMacsecSecY, t_Handle h_Sc, t_FmMacsecSecYRxScStatistics *p_Statistics);
-+
-+/**************************************************************************//**
-+ @Function FM_MACSEC_SECY_RxSaGetStatistics
-+
-+ @Description get all statistics counters
-+
-+ @Param[in] h_FmMacsecSecY FM MACSEC SECY module descriptor.
-+ @Param[in] h_Sc Rx Sc handle.
-+ @Param[in] an association number represent the SA.
-+ @Param[in] p_Statistics Structure with statistics.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MACSEC_SECY_Init().
-+*//***************************************************************************/
-+t_Error FM_MACSEC_SECY_RxSaGetStatistics(t_Handle h_FmMacsecSecY, t_Handle h_Sc, macsecAN_t an, t_FmMacsecSecYRxSaStatistics *p_Statistics);
-+
-+/**************************************************************************//**
-+ @Function FM_MACSEC_SECY_TxScGetStatistics
-+
-+ @Description get all statistics counters.
-+
-+ @Param[in] h_FmMacsecSecY FM MACSEC SECY module descriptor.
-+ @Param[in] p_Statistics Structure with statistics.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MACSEC_SECY_Init().
-+*//***************************************************************************/
-+t_Error FM_MACSEC_SECY_TxScGetStatistics(t_Handle h_FmMacsecSecY, t_FmMacsecSecYTxScStatistics *p_Statistics);
-+
-+/**************************************************************************//**
-+ @Function FM_MACSEC_SECY_TxSaGetStatistics
-+
-+ @Description get all statistics counters.
-+
-+ @Param[in] h_FmMacsecSecY FM MACSEC SECY module descriptor.
-+ @Param[in] an association number represent the SA.
-+ @Param[in] p_Statistics Structure with statistics.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MACSEC_SECY_Init().
-+*//***************************************************************************/
-+t_Error FM_MACSEC_SECY_TxSaGetStatistics(t_Handle h_FmMacsecSecY, macsecAN_t an, t_FmMacsecSecYTxSaStatistics *p_Statistics);
-+
-+/**************************************************************************//**
-+ @Function FM_MACSEC_SECY_SetException
-+
-+ @Description Calling this routine enables/disables the specified exception.
-+
-+ @Param[in] h_FmMacsecSecY FM MACSEC SECY module descriptor.
-+ @Param[in] exception The exception to be selected.
-+ @Param[in] enable TRUE to enable interrupt, FALSE to mask it.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MACSEC_SECY_Init().
-+*//***************************************************************************/
-+t_Error FM_MACSEC_SECY_SetException(t_Handle h_FmMacsecSecY, e_FmMacsecExceptions exception, bool enable);
-+
-+/**************************************************************************//**
-+ @Function FM_MACSEC_SECY_SetEvent
-+
-+ @Description Calling this routine enables/disables the specified event.
-+
-+ @Param[in] h_FmMacsecSecY FM MACSEC SECY module descriptor.
-+ @Param[in] event The event to be selected.
-+ @Param[in] enable TRUE to enable interrupt, FALSE to mask it.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MACSEC_SECY_Config() and before FM_MACSEC_SECY_Init().
-+*//***************************************************************************/
-+t_Error FM_MACSEC_SECY_SetEvent(t_Handle h_FmMacsecSecY, e_FmMacsecSecYEvents event, bool enable);
-+
-+/**************************************************************************//**
-+ @Function FM_MACSEC_SECY_GetRxScPhysId
-+
-+ @Description return the physical id of the Secure Channel.
-+
-+ @Param[in] h_FmMacsecSecY FM MACSEC SECY module descriptor.
-+ @Param[in] h_Sc SC handle as returned by FM_MACSEC_SECY_CreateRxSc.
-+ @Param[out] p_ScPhysId the SC physical id.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MACSEC_SECY_CreateRxSc().
-+*//***************************************************************************/
-+t_Error FM_MACSEC_SECY_GetRxScPhysId(t_Handle h_FmMacsecSecY, t_Handle h_Sc, uint32_t *p_ScPhysId);
-+
-+/**************************************************************************//**
-+ @Function FM_MACSEC_SECY_GetTxScPhysId
-+
-+ @Description return the physical id of the Secure Channel.
-+
-+ @Param[in] h_FmMacsecSecY FM MACSEC SECY module descriptor.
-+ @Param[out] p_ScPhysId the SC physical id.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MACSEC_SECY_Init().
-+*//***************************************************************************/
-+t_Error FM_MACSEC_SECY_GetTxScPhysId(t_Handle h_FmMacsecSecY, uint32_t *p_ScPhysId);
-+
-+/** @} */ /* end of FM_MACSEC_SECY_runtime_control_grp group */
-+/** @} */ /* end of FM_MACSEC_SECY_grp group */
-+/** @} */ /* end of FM_MACSEC_grp group */
-+/** @} */ /* end of FM_grp group */
-+
-+
-+#endif /* __FM_MACSEC_EXT_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/Peripherals/fm_muram_ext.h
-@@ -0,0 +1,170 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/**************************************************************************//**
-+ @File fm_muram_ext.h
-+
-+ @Description FM MURAM Application Programming Interface.
-+*//***************************************************************************/
-+#ifndef __FM_MURAM_EXT
-+#define __FM_MURAM_EXT
-+
-+#include "error_ext.h"
-+#include "std_ext.h"
-+
-+
-+/**************************************************************************//**
-+
-+ @Group FM_grp Frame Manager API
-+
-+ @Description FM API functions, definitions and enums
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Group FM_muram_grp FM MURAM
-+
-+ @Description FM MURAM API functions, definitions and enums
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Group FM_muram_init_grp FM MURAM Initialization Unit
-+
-+ @Description FM MURAM initialization API functions, definitions and enums
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Function FM_MURAM_ConfigAndInit
-+
-+ @Description Creates partition in the MURAM.
-+
-+ The routine returns a handle (descriptor) to the MURAM partition.
-+ This descriptor must be passed as first parameter to all other
-+ FM-MURAM function calls.
-+
-+ No actual initialization or configuration of FM_MURAM hardware is
-+ done by this routine.
-+
-+ @Param[in] baseAddress - Pointer to base of memory mapped FM-MURAM.
-+ @Param[in] size - Size of the FM-MURAM partition.
-+
-+ @Return Handle to FM-MURAM object, or NULL for Failure.
-+*//***************************************************************************/
-+t_Handle FM_MURAM_ConfigAndInit(uintptr_t baseAddress, uint32_t size);
-+
-+/**************************************************************************//**
-+ @Function FM_MURAM_Free
-+
-+ @Description Frees all resources that were assigned to FM-MURAM module.
-+
-+ Calling this routine invalidates the descriptor.
-+
-+ @Param[in] h_FmMuram - FM-MURAM module descriptor.
-+
-+ @Return E_OK on success; Error code otherwise.
-+*//***************************************************************************/
-+t_Error FM_MURAM_Free(t_Handle h_FmMuram);
-+
-+/** @} */ /* end of FM_muram_init_grp group */
-+
-+
-+/**************************************************************************//**
-+ @Group FM_muram_ctrl_grp FM MURAM Control Unit
-+
-+ @Description FM MURAM control API functions, definitions and enums
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Function FM_MURAM_AllocMem
-+
-+ @Description Allocate some memory from FM-MURAM partition.
-+
-+ @Param[in] h_FmMuram - FM-MURAM module descriptor.
-+ @Param[in] size - size of the memory to be allocated.
-+ @Param[in] align - Alignment of the memory.
-+
-+ @Return address of the allocated memory; NULL otherwise.
-+*//***************************************************************************/
-+void * FM_MURAM_AllocMem(t_Handle h_FmMuram, uint32_t size, uint32_t align);
-+
-+/**************************************************************************//**
-+ @Function FM_MURAM_AllocMemForce
-+
-+ @Description Allocate some specific memory from FM-MURAM partition (according
-+ to base).
-+
-+ @Param[in] h_FmMuram - FM-MURAM module descriptor.
-+ @Param[in] base - the desired base-address to be allocated.
-+ @Param[in] size - size of the memory to be allocated.
-+
-+ @Return address of the allocated memory; NULL otherwise.
-+*//***************************************************************************/
-+void * FM_MURAM_AllocMemForce(t_Handle h_FmMuram, uint64_t base, uint32_t size);
-+
-+/**************************************************************************//**
-+ @Function FM_MURAM_FreeMem
-+
-+ @Description Free an allocated memory from FM-MURAM partition.
-+
-+ @Param[in] h_FmMuram - FM-MURAM module descriptor.
-+ @Param[in] ptr - A pointer to an allocated memory.
-+
-+ @Return E_OK on success; Error code otherwise.
-+*//***************************************************************************/
-+t_Error FM_MURAM_FreeMem(t_Handle h_FmMuram, void *ptr);
-+
-+/**************************************************************************//**
-+ @Function FM_MURAM_GetFreeMemSize
-+
-+ @Description Returns the size (in bytes) of free MURAM memory.
-+
-+ @Param[in] h_FmMuram - FM-MURAM module descriptor.
-+
-+ @Return Free MURAM memory size in bytes.
-+*//***************************************************************************/
-+uint64_t FM_MURAM_GetFreeMemSize(t_Handle h_FmMuram);
-+
-+/** @} */ /* end of FM_muram_ctrl_grp group */
-+/** @} */ /* end of FM_muram_grp group */
-+/** @} */ /* end of FM_grp group */
-+
-+
-+
-+#endif /* __FM_MURAM_EXT */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/Peripherals/fm_pcd_ext.h
-@@ -0,0 +1,3974 @@
-+/* Copyright (c) 2008-2012 Freescale Semiconductor, Inc
-+ * All rights reserved.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/**************************************************************************//**
-+ @File fm_pcd_ext.h
-+
-+ @Description FM PCD API definitions
-+*//***************************************************************************/
-+#ifndef __FM_PCD_EXT
-+#define __FM_PCD_EXT
-+
-+#include "std_ext.h"
-+#include "net_ext.h"
-+#include "list_ext.h"
-+#include "fm_ext.h"
-+#include "fsl_fman_kg.h"
-+
-+
-+/**************************************************************************//**
-+ @Group FM_grp Frame Manager API
-+
-+ @Description Frame Manager Application Programming Interface
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Group FM_PCD_grp FM PCD
-+
-+ @Description Frame Manager PCD (Parse-Classify-Distribute) API.
-+
-+ The FM PCD module is responsible for the initialization of all
-+ global classifying FM modules. This includes the parser general and
-+ common registers, the key generator global and common registers,
-+ and the policer global and common registers.
-+ In addition, the FM PCD SW module will initialize all required
-+ key generator schemes, coarse classification flows, and policer
-+ profiles. When FM module is configured to work with one of these
-+ entities, it will register to it using the FM PORT API. The PCD
-+ module will manage the PCD resources - i.e. resource management of
-+ KeyGen schemes, etc.
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Collection General PCD defines
-+*//***************************************************************************/
-+#define FM_PCD_MAX_NUM_OF_PRIVATE_HDRS 2 /**< Number of units/headers saved for user */
-+
-+#define FM_PCD_PRS_NUM_OF_HDRS 16 /**< Number of headers supported by HW parser */
-+#define FM_PCD_MAX_NUM_OF_DISTINCTION_UNITS (32 - FM_PCD_MAX_NUM_OF_PRIVATE_HDRS)
-+ /**< Number of distinction units is limited by
-+ register size (32 bits) minus reserved bits
-+ for private headers. */
-+#define FM_PCD_MAX_NUM_OF_INTERCHANGEABLE_HDRS 4 /**< Maximum number of interchangeable headers
-+ in a distinction unit */
-+#define FM_PCD_KG_NUM_OF_GENERIC_REGS FM_KG_NUM_OF_GENERIC_REGS /**< Total number of generic KeyGen registers */
-+#define FM_PCD_KG_MAX_NUM_OF_EXTRACTS_PER_KEY 35 /**< Max number allowed on any configuration;
-+ For HW implementation reasons, in most
-+ cases less than this will be allowed; The
-+ driver will return an initialization error
-+ if resource is unavailable. */
-+#define FM_PCD_KG_NUM_OF_EXTRACT_MASKS 4 /**< Total number of masks allowed on KeyGen extractions. */
-+#define FM_PCD_KG_NUM_OF_DEFAULT_GROUPS 16 /**< Number of default value logical groups */
-+
-+#define FM_PCD_PRS_NUM_OF_LABELS 32 /**< Maximum number of SW parser labels */
-+#define FM_SW_PRS_MAX_IMAGE_SIZE (FM_PCD_SW_PRS_SIZE /*- FM_PCD_PRS_SW_OFFSET -FM_PCD_PRS_SW_TAIL_SIZE*/-FM_PCD_PRS_SW_PATCHES_SIZE)
-+ /**< Maximum size of SW parser code */
-+
-+#define FM_PCD_MAX_MANIP_INSRT_TEMPLATE_SIZE 128 /**< Maximum size of insertion template for
-+ insert manipulation */
-+
-+#if (DPAA_VERSION >= 11)
-+#define FM_PCD_FRM_REPLIC_MAX_NUM_OF_ENTRIES 64 /**< Maximum possible entries for frame replicator group */
-+#endif /* (DPAA_VERSION >= 11) */
-+/* @} */
-+
-+
-+/**************************************************************************//**
-+ @Group FM_PCD_init_grp FM PCD Initialization Unit
-+
-+ @Description Frame Manager PCD Initialization Unit API
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Description PCD counters
-+*//***************************************************************************/
-+typedef enum e_FmPcdCounters {
-+ e_FM_PCD_KG_COUNTERS_TOTAL, /**< KeyGen counter */
-+ e_FM_PCD_PLCR_COUNTERS_RED, /**< Policer counter - counts the total number of RED packets that exit the Policer. */
-+ e_FM_PCD_PLCR_COUNTERS_YELLOW, /**< Policer counter - counts the total number of YELLOW packets that exit the Policer. */
-+ e_FM_PCD_PLCR_COUNTERS_RECOLORED_TO_RED, /**< Policer counter - counts the number of packets that changed color to RED by the Policer;
-+ This is a subset of e_FM_PCD_PLCR_COUNTERS_RED packet count, indicating active color changes. */
-+ e_FM_PCD_PLCR_COUNTERS_RECOLORED_TO_YELLOW, /**< Policer counter - counts the number of packets that changed color to YELLOW by the Policer;
-+ This is a subset of e_FM_PCD_PLCR_COUNTERS_YELLOW packet count, indicating active color changes. */
-+ e_FM_PCD_PLCR_COUNTERS_TOTAL, /**< Policer counter - counts the total number of packets passed in the Policer. */
-+ e_FM_PCD_PLCR_COUNTERS_LENGTH_MISMATCH, /**< Policer counter - counts the number of packets with length mismatch. */
-+ e_FM_PCD_PRS_COUNTERS_PARSE_DISPATCH, /**< Parser counter - counts the number of times the parser block is dispatched. */
-+ e_FM_PCD_PRS_COUNTERS_L2_PARSE_RESULT_RETURNED, /**< Parser counter - counts the number of times L2 parse result is returned (including errors). */
-+ e_FM_PCD_PRS_COUNTERS_L3_PARSE_RESULT_RETURNED, /**< Parser counter - counts the number of times L3 parse result is returned (including errors). */
-+ e_FM_PCD_PRS_COUNTERS_L4_PARSE_RESULT_RETURNED, /**< Parser counter - counts the number of times L4 parse result is returned (including errors). */
-+ e_FM_PCD_PRS_COUNTERS_SHIM_PARSE_RESULT_RETURNED, /**< Parser counter - counts the number of times SHIM parse result is returned (including errors). */
-+ e_FM_PCD_PRS_COUNTERS_L2_PARSE_RESULT_RETURNED_WITH_ERR, /**< Parser counter - counts the number of times L2 parse result is returned with errors. */
-+ e_FM_PCD_PRS_COUNTERS_L3_PARSE_RESULT_RETURNED_WITH_ERR, /**< Parser counter - counts the number of times L3 parse result is returned with errors. */
-+ e_FM_PCD_PRS_COUNTERS_L4_PARSE_RESULT_RETURNED_WITH_ERR, /**< Parser counter - counts the number of times L4 parse result is returned with errors. */
-+ e_FM_PCD_PRS_COUNTERS_SHIM_PARSE_RESULT_RETURNED_WITH_ERR, /**< Parser counter - counts the number of times SHIM parse result is returned with errors. */
-+ e_FM_PCD_PRS_COUNTERS_SOFT_PRS_CYCLES, /**< Parser counter - counts the number of cycles spent executing soft parser instruction (including stall cycles). */
-+ e_FM_PCD_PRS_COUNTERS_SOFT_PRS_STALL_CYCLES, /**< Parser counter - counts the number of cycles stalled waiting for parser internal memory reads while executing soft parser instruction. */
-+ e_FM_PCD_PRS_COUNTERS_HARD_PRS_CYCLE_INCL_STALL_CYCLES, /**< Parser counter - counts the number of cycles spent executing hard parser (including stall cycles). */
-+ e_FM_PCD_PRS_COUNTERS_MURAM_READ_CYCLES, /**< MURAM counter - counts the number of cycles while performing FMan Memory read. */
-+ e_FM_PCD_PRS_COUNTERS_MURAM_READ_STALL_CYCLES, /**< MURAM counter - counts the number of cycles stalled while performing FMan Memory read. */
-+ e_FM_PCD_PRS_COUNTERS_MURAM_WRITE_CYCLES, /**< MURAM counter - counts the number of cycles while performing FMan Memory write. */
-+ e_FM_PCD_PRS_COUNTERS_MURAM_WRITE_STALL_CYCLES, /**< MURAM counter - counts the number of cycles stalled while performing FMan Memory write. */
-+ e_FM_PCD_PRS_COUNTERS_FPM_COMMAND_STALL_CYCLES /**< FPM counter - counts the number of cycles stalled while performing a FPM Command. */
-+} e_FmPcdCounters;
-+
-+/**************************************************************************//**
-+ @Description PCD interrupts
-+*//***************************************************************************/
-+typedef enum e_FmPcdExceptions {
-+ e_FM_PCD_KG_EXCEPTION_DOUBLE_ECC, /**< KeyGen double-bit ECC error is detected on internal memory read access. */
-+ e_FM_PCD_KG_EXCEPTION_KEYSIZE_OVERFLOW, /**< KeyGen scheme configuration error indicating a key size larger than 56 bytes. */
-+ e_FM_PCD_PLCR_EXCEPTION_DOUBLE_ECC, /**< Policer double-bit ECC error has been detected on PRAM read access. */
-+ e_FM_PCD_PLCR_EXCEPTION_INIT_ENTRY_ERROR, /**< Policer access to a non-initialized profile has been detected. */
-+ e_FM_PCD_PLCR_EXCEPTION_PRAM_SELF_INIT_COMPLETE, /**< Policer RAM self-initialization complete */
-+ e_FM_PCD_PLCR_EXCEPTION_ATOMIC_ACTION_COMPLETE, /**< Policer atomic action complete */
-+ e_FM_PCD_PRS_EXCEPTION_DOUBLE_ECC, /**< Parser double-bit ECC error */
-+ e_FM_PCD_PRS_EXCEPTION_SINGLE_ECC /**< Parser single-bit ECC error */
-+} e_FmPcdExceptions;
-+
-+
-+/**************************************************************************//**
-+ @Description Exceptions user callback routine, will be called upon an
-+ exception passing the exception identification.
-+
-+ @Param[in] h_App - User's application descriptor.
-+ @Param[in] exception - The exception.
-+ *//***************************************************************************/
-+typedef void (t_FmPcdExceptionCallback) (t_Handle h_App, e_FmPcdExceptions exception);
-+
-+/**************************************************************************//**
-+ @Description Exceptions user callback routine, will be called upon an exception
-+ passing the exception identification.
-+
-+ @Param[in] h_App - User's application descriptor.
-+ @Param[in] exception - The exception.
-+ @Param[in] index - id of the relevant source (may be scheme or profile id).
-+ *//***************************************************************************/
-+typedef void (t_FmPcdIdExceptionCallback) ( t_Handle h_App,
-+ e_FmPcdExceptions exception,
-+ uint16_t index);
-+
-+/**************************************************************************//**
-+ @Description A callback for enqueuing frame onto a QM queue.
-+
-+ @Param[in] h_QmArg - Application's handle passed to QM module on enqueue.
-+ @Param[in] p_Fd - Frame descriptor for the frame.
-+
-+ @Return E_OK on success; Error code otherwise.
-+ *//***************************************************************************/
-+typedef t_Error (t_FmPcdQmEnqueueCallback) (t_Handle h_QmArg, void *p_Fd);
-+
-+/**************************************************************************//**
-+ @Description Host-Command parameters structure.
-+
-+ When using Host command for PCD functionalities, a dedicated port
-+ must be used. If this routine is called for a PCD in a single partition
-+ environment, or it is the Master partition in a Multi-partition
-+ environment, The port will be initialized by the PCD driver
-+ initialization routine.
-+ *//***************************************************************************/
-+typedef struct t_FmPcdHcParams {
-+ uintptr_t portBaseAddr; /**< Virtual Address of Host-Command Port memory mapped registers.*/
-+ uint8_t portId; /**< Port Id (0-6 relative to Host-Command/Offline-Parsing ports);
-+ NOTE: When configuring Host Command port for
-+ FMANv3 devices (DPAA_VERSION 11 and higher),
-+ portId=0 MUST be used. */
-+ uint16_t liodnBase; /**< LIODN base for this port, to be used together with LIODN offset
-+ (irrelevant for P4080 revision 1.0) */
-+ uint32_t errFqid; /**< Host-Command Port error queue Id. */
-+ uint32_t confFqid; /**< Host-Command Port confirmation queue Id. */
-+ uint32_t qmChannel; /**< QM channel dedicated to this Host-Command port;
-+ will be used by the FM for dequeue. */
-+ t_FmPcdQmEnqueueCallback *f_QmEnqueue; /**< Callback routine for enqueuing a frame to the QM */
-+ t_Handle h_QmArg; /**< Application's handle passed to QM module on enqueue */
-+} t_FmPcdHcParams;
-+
-+/**************************************************************************//**
-+ @Description The main structure for PCD initialization
-+ *//***************************************************************************/
-+typedef struct t_FmPcdParams {
-+ bool prsSupport; /**< TRUE if Parser will be used for any of the FM ports. */
-+ bool ccSupport; /**< TRUE if Coarse Classification will be used for any
-+ of the FM ports. */
-+ bool kgSupport; /**< TRUE if KeyGen will be used for any of the FM ports. */
-+ bool plcrSupport; /**< TRUE if Policer will be used for any of the FM ports. */
-+ t_Handle h_Fm; /**< A handle to the FM module. */
-+ uint8_t numOfSchemes; /**< Number of schemes dedicated to this partition.
-+ this parameter is relevant if 'kgSupport'=TRUE. */
-+ bool useHostCommand; /**< Optional for single partition, Mandatory for Multi partition */
-+ t_FmPcdHcParams hc; /**< Host Command parameters, relevant only if 'useHostCommand'=TRUE;
-+ Relevant when FM not runs in "guest-mode". */
-+
-+ t_FmPcdExceptionCallback *f_Exception; /**< Callback routine for general PCD exceptions;
-+ Relevant when FM not runs in "guest-mode". */
-+ t_FmPcdIdExceptionCallback *f_ExceptionId; /**< Callback routine for specific KeyGen scheme or
-+ Policer profile exceptions;
-+ Relevant when FM not runs in "guest-mode". */
-+ t_Handle h_App; /**< A handle to an application layer object; This handle will
-+ be passed by the driver upon calling the above callbacks;
-+ Relevant when FM not runs in "guest-mode". */
-+ uint8_t partPlcrProfilesBase; /**< The first policer-profile-id dedicated to this partition.
-+ this parameter is relevant if 'plcrSupport'=TRUE.
-+ NOTE: this parameter relevant only when working with multiple partitions. */
-+ uint16_t partNumOfPlcrProfiles; /**< Number of policer-profiles dedicated to this partition.
-+ this parameter is relevant if 'plcrSupport'=TRUE.
-+ NOTE: this parameter relevant only when working with multiple partitions. */
-+} t_FmPcdParams;
-+
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_Config
-+
-+ @Description Basic configuration of the PCD module.
-+ Creates descriptor for the FM PCD module.
-+
-+ @Param[in] p_FmPcdParams A structure of parameters for the initialization of PCD.
-+
-+ @Return A handle to the initialized module.
-+*//***************************************************************************/
-+t_Handle FM_PCD_Config(t_FmPcdParams *p_FmPcdParams);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_Init
-+
-+ @Description Initialization of the PCD module.
-+
-+ @Param[in] h_FmPcd - FM PCD module descriptor.
-+
-+ @Return E_OK on success; Error code otherwise.
-+*//***************************************************************************/
-+t_Error FM_PCD_Init(t_Handle h_FmPcd);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_Free
-+
-+ @Description Frees all resources that were assigned to FM module.
-+
-+ Calling this routine invalidates the descriptor.
-+
-+ @Param[in] h_FmPcd - FM PCD module descriptor.
-+
-+ @Return E_OK on success; Error code otherwise.
-+*//***************************************************************************/
-+t_Error FM_PCD_Free(t_Handle h_FmPcd);
-+
-+/**************************************************************************//**
-+ @Group FM_PCD_advanced_cfg_grp FM PCD Advanced Configuration Unit
-+
-+ @Description Frame Manager PCD Advanced Configuration API.
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_ConfigException
-+
-+ @Description Calling this routine changes the internal driver data base
-+ from its default selection of exceptions enabling.
-+ [DEFAULT_numOfSharedPlcrProfiles].
-+
-+ @Param[in] h_FmPcd FM PCD module descriptor.
-+ @Param[in] exception The exception to be selected.
-+ @Param[in] enable TRUE to enable interrupt, FALSE to mask it.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions This routine should NOT be called from guest-partition
-+ (i.e. guestId != NCSW_MASTER_ID)
-+*//***************************************************************************/
-+t_Error FM_PCD_ConfigException(t_Handle h_FmPcd, e_FmPcdExceptions exception, bool enable);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_ConfigHcFramesDataMemory
-+
-+ @Description Configures memory-partition-id for FMan-Controller Host-Command
-+ frames. Calling this routine changes the internal driver data
-+ base from its default configuration [0].
-+
-+ @Param[in] h_FmPcd FM PCD module descriptor.
-+ @Param[in] memId Memory partition ID.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions This routine may be called only if 'useHostCommand' was TRUE
-+ when FM_PCD_Config() routine was called.
-+*//***************************************************************************/
-+t_Error FM_PCD_ConfigHcFramesDataMemory(t_Handle h_FmPcd, uint8_t memId);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_ConfigPlcrNumOfSharedProfiles
-+
-+ @Description Calling this routine changes the internal driver data base
-+ from its default selection of exceptions enablement.
-+ [DEFAULT_numOfSharedPlcrProfiles].
-+
-+ @Param[in] h_FmPcd FM PCD module descriptor.
-+ @Param[in] numOfSharedPlcrProfiles Number of profiles to
-+ be shared between ports on this partition
-+
-+ @Return E_OK on success; Error code otherwise.
-+*//***************************************************************************/
-+t_Error FM_PCD_ConfigPlcrNumOfSharedProfiles(t_Handle h_FmPcd, uint16_t numOfSharedPlcrProfiles);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_ConfigPlcrAutoRefreshMode
-+
-+ @Description Calling this routine changes the internal driver data base
-+ from its default selection of exceptions enablement.
-+ By default auto-refresh is [DEFAULT_plcrAutoRefresh].
-+
-+ @Param[in] h_FmPcd FM PCD module descriptor.
-+ @Param[in] enable TRUE to enable, FALSE to disable
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions This routine should NOT be called from guest-partition
-+ (i.e. guestId != NCSW_MASTER_ID)
-+*//***************************************************************************/
-+t_Error FM_PCD_ConfigPlcrAutoRefreshMode(t_Handle h_FmPcd, bool enable);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_ConfigPrsMaxCycleLimit
-+
-+ @Description Calling this routine changes the internal data structure for
-+ the maximum parsing time from its default value
-+ [DEFAULT_MAX_PRS_CYC_LIM].
-+
-+ @Param[in] h_FmPcd FM PCD module descriptor.
-+ @Param[in] value 0 to disable the mechanism, or new
-+ maximum parsing time.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions This routine should NOT be called from guest-partition
-+ (i.e. guestId != NCSW_MASTER_ID)
-+*//***************************************************************************/
-+t_Error FM_PCD_ConfigPrsMaxCycleLimit(t_Handle h_FmPcd,uint16_t value);
-+
-+/** @} */ /* end of FM_PCD_advanced_cfg_grp group */
-+/** @} */ /* end of FM_PCD_init_grp group */
-+
-+
-+/**************************************************************************//**
-+ @Group FM_PCD_Runtime_grp FM PCD Runtime Unit
-+
-+ @Description Frame Manager PCD Runtime Unit API
-+
-+ The runtime control allows creation of PCD infrastructure modules
-+ such as Network Environment Characteristics, Classification Plan
-+ Groups and Coarse Classification Trees.
-+ It also allows on-the-fly initialization, modification and removal
-+ of PCD modules such as KeyGen schemes, coarse classification nodes
-+ and Policer profiles.
-+
-+ In order to explain the programming model of the PCD driver interface
-+ a few terms should be explained, and will be used below.
-+ - Distinction Header - One of the 16 protocols supported by the FM parser,
-+ or one of the SHIM headers (1 or 2). May be a header with a special
-+ option (see below).
-+ - Interchangeable Headers Group - This is a group of Headers recognized
-+ by either one of them. For example, if in a specific context the user
-+ chooses to treat IPv4 and IPV6 in the same way, they may create an
-+ interchangeable Headers Unit consisting of these 2 headers.
-+ - A Distinction Unit - a Distinction Header or an Interchangeable Headers
-+ Group.
-+ - Header with special option - applies to Ethernet, MPLS, VLAN, IPv4 and
-+ IPv6, includes multicast, broadcast and other protocol specific options.
-+ In terms of hardware it relates to the options available in the classification
-+ plan.
-+ - Network Environment Characteristics - a set of Distinction Units that define
-+ the total recognizable header selection for a certain environment. This is
-+ NOT the list of all headers that will ever appear in a flow, but rather
-+ everything that needs distinction in a flow, where distinction is made by KeyGen
-+ schemes and coarse classification action descriptors.
-+
-+ The PCD runtime modules initialization is done in stages. The first stage after
-+ initializing the PCD module itself is to establish a Network Flows Environment
-+ Definition. The application may choose to establish one or more such environments.
-+ Later, when needed, the application will have to state, for some of its modules,
-+ to which single environment it belongs.
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Description A structure for SW parser labels
-+ *//***************************************************************************/
-+typedef struct t_FmPcdPrsLabelParams {
-+ uint32_t instructionOffset; /**< SW parser label instruction offset (2 bytes
-+ resolution), relative to Parser RAM. */
-+ e_NetHeaderType hdr; /**< The existence of this header will invoke
-+ the SW parser code; Use HEADER_TYPE_NONE
-+ to indicate that sw parser is to run
-+ independent of the existence of any protocol
-+ (run before HW parser). */
-+ uint8_t indexPerHdr; /**< Normally 0, if more than one SW parser
-+ attachments for the same header, use this
-+ index to distinguish between them. */
-+} t_FmPcdPrsLabelParams;
-+
-+/**************************************************************************//**
-+ @Description A structure for SW parser
-+ *//***************************************************************************/
-+typedef struct t_FmPcdPrsSwParams {
-+ bool override; /**< FALSE to invoke a check that nothing else
-+ was loaded to this address, including
-+ internal patches.
-+ TRUE to override any existing code.*/
-+ uint32_t size; /**< SW parser code size */
-+ uint16_t base; /**< SW parser base (in instruction counts!
-+ must be larger than 0x20)*/
-+ uint8_t *p_Code; /**< SW parser code */
-+ uint32_t swPrsDataParams[FM_PCD_PRS_NUM_OF_HDRS];
-+ /**< SW parser data (parameters) */
-+ uint8_t numOfLabels; /**< Number of labels for SW parser. */
-+ t_FmPcdPrsLabelParams labelsTable[FM_PCD_PRS_NUM_OF_LABELS];
-+ /**< SW parser labels table, containing
-+ numOfLabels entries */
-+} t_FmPcdPrsSwParams;
-+
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_Enable
-+
-+ @Description This routine should be called after PCD is initialized for enabling all
-+ PCD engines according to their existing configuration.
-+
-+ @Param[in] h_FmPcd FM PCD module descriptor.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_Init() and when PCD is disabled.
-+*//***************************************************************************/
-+t_Error FM_PCD_Enable(t_Handle h_FmPcd);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_Disable
-+
-+ @Description This routine may be called when PCD is enabled in order to
-+ disable all PCD engines. It may be called
-+ only when none of the ports in the system are using the PCD.
-+
-+ @Param[in] h_FmPcd FM PCD module descriptor.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_Init() and when PCD is enabled.
-+*//***************************************************************************/
-+t_Error FM_PCD_Disable(t_Handle h_FmPcd);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_GetCounter
-+
-+ @Description Reads one of the FM PCD counters.
-+
-+ @Param[in] h_FmPcd FM PCD module descriptor.
-+ @Param[in] counter The requested counter.
-+
-+ @Return Counter's current value.
-+
-+ @Cautions Allowed only following FM_PCD_Init().
-+ Note that it is user's responsibility to call this routine only
-+ for enabled counters, and there will be no indication if a
-+ disabled counter is accessed.
-+*//***************************************************************************/
-+uint32_t FM_PCD_GetCounter(t_Handle h_FmPcd, e_FmPcdCounters counter);
-+
-+/**************************************************************************//**
-+@Function FM_PCD_PrsLoadSw
-+
-+@Description This routine may be called in order to load software parsing code.
-+
-+
-+@Param[in] h_FmPcd FM PCD module descriptor.
-+@Param[in] p_SwPrs A pointer to a structure of software
-+ parser parameters, including the software
-+ parser image.
-+
-+@Return E_OK on success; Error code otherwise.
-+
-+@Cautions Allowed only following FM_PCD_Init() and when PCD is disabled.
-+ This routine should NOT be called from guest-partition
-+ (i.e. guestId != NCSW_MASTER_ID)
-+*//***************************************************************************/
-+t_Error FM_PCD_PrsLoadSw(t_Handle h_FmPcd, t_FmPcdPrsSwParams *p_SwPrs);
-+
-+/**************************************************************************//**
-+@Function FM_PCD_SetAdvancedOffloadSupport
-+
-+@Description This routine must be called in order to support the following features:
-+ IP-fragmentation, IP-reassembly, IPsec, Header-manipulation, frame-replicator.
-+
-+@Param[in] h_FmPcd FM PCD module descriptor.
-+
-+@Return E_OK on success; Error code otherwise.
-+
-+@Cautions Allowed only following FM_PCD_Init() and when PCD is disabled.
-+ This routine should NOT be called from guest-partition
-+ (i.e. guestId != NCSW_MASTER_ID)
-+*//***************************************************************************/
-+t_Error FM_PCD_SetAdvancedOffloadSupport(t_Handle h_FmPcd);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_KgSetDfltValue
-+
-+ @Description Calling this routine sets a global default value to be used
-+ by the KeyGen when parser does not recognize a required
-+ field/header.
-+ By default default values are 0.
-+
-+ @Param[in] h_FmPcd FM PCD module descriptor.
-+ @Param[in] valueId 0,1 - one of 2 global default values.
-+ @Param[in] value The requested default value.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_Init() and when PCD is disabled.
-+ This routine should NOT be called from guest-partition
-+ (i.e. guestId != NCSW_MASTER_ID)
-+*//***************************************************************************/
-+t_Error FM_PCD_KgSetDfltValue(t_Handle h_FmPcd, uint8_t valueId, uint32_t value);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_KgSetAdditionalDataAfterParsing
-+
-+ @Description Calling this routine allows the KeyGen to access data past
-+ the parser finishing point.
-+
-+ @Param[in] h_FmPcd FM PCD module descriptor.
-+ @Param[in] payloadOffset the number of bytes beyond the parser location.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_Init() and when PCD is disabled.
-+ This routine should NOT be called from guest-partition
-+ (i.e. guestId != NCSW_MASTER_ID)
-+*//***************************************************************************/
-+t_Error FM_PCD_KgSetAdditionalDataAfterParsing(t_Handle h_FmPcd, uint8_t payloadOffset);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_SetException
-+
-+ @Description Calling this routine enables/disables PCD interrupts.
-+
-+ @Param[in] h_FmPcd FM PCD module descriptor.
-+ @Param[in] exception The exception to be selected.
-+ @Param[in] enable TRUE to enable interrupt, FALSE to mask it.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_Init().
-+ This routine should NOT be called from guest-partition
-+ (i.e. guestId != NCSW_MASTER_ID)
-+*//***************************************************************************/
-+t_Error FM_PCD_SetException(t_Handle h_FmPcd, e_FmPcdExceptions exception, bool enable);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_ModifyCounter
-+
-+ @Description Sets a value to an enabled counter. Use "0" to reset the counter.
-+
-+ @Param[in] h_FmPcd FM PCD module descriptor.
-+ @Param[in] counter The requested counter.
-+ @Param[in] value The requested value to be written into the counter.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_Init().
-+ This routine should NOT be called from guest-partition
-+ (i.e. guestId != NCSW_MASTER_ID)
-+*//***************************************************************************/
-+t_Error FM_PCD_ModifyCounter(t_Handle h_FmPcd, e_FmPcdCounters counter, uint32_t value);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_SetPlcrStatistics
-+
-+ @Description This routine may be used to enable/disable policer statistics
-+ counter. By default the statistics is enabled.
-+
-+ @Param[in] h_FmPcd FM PCD module descriptor
-+ @Param[in] enable TRUE to enable, FALSE to disable.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_Init().
-+ This routine should NOT be called from guest-partition
-+ (i.e. guestId != NCSW_MASTER_ID)
-+*//***************************************************************************/
-+t_Error FM_PCD_SetPlcrStatistics(t_Handle h_FmPcd, bool enable);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_SetPrsStatistics
-+
-+ @Description Defines whether to gather parser statistics including all ports.
-+
-+ @Param[in] h_FmPcd FM PCD module descriptor.
-+ @Param[in] enable TRUE to enable, FALSE to disable.
-+
-+ @Return None
-+
-+ @Cautions Allowed only following FM_PCD_Init().
-+ This routine should NOT be called from guest-partition
-+ (i.e. guestId != NCSW_MASTER_ID)
-+*//***************************************************************************/
-+void FM_PCD_SetPrsStatistics(t_Handle h_FmPcd, bool enable);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_HcTxConf
-+
-+ @Description This routine should be called to confirm frames that were
-+ received on the HC confirmation queue.
-+
-+ @Param[in] h_FmPcd A handle to an FM PCD Module.
-+ @Param[in] p_Fd Frame descriptor of the received frame.
-+
-+ @Cautions Allowed only following FM_PCD_Init(). Allowed only if 'useHostCommand'
-+ option was selected in the initialization.
-+*//***************************************************************************/
-+void FM_PCD_HcTxConf(t_Handle h_FmPcd, t_DpaaFD *p_Fd);
-+
-+/**************************************************************************//*
-+ @Function FM_PCD_ForceIntr
-+
-+ @Description Causes an interrupt event on the requested source.
-+
-+ @Param[in] h_FmPcd FM PCD module descriptor.
-+ @Param[in] exception An exception to be forced.
-+
-+ @Return E_OK on success; Error code if the exception is not enabled,
-+ or is not able to create interrupt.
-+
-+ @Cautions Allowed only following FM_PCD_Init().
-+ This routine should NOT be called from guest-partition
-+ (i.e. guestId != NCSW_MASTER_ID)
-+*//***************************************************************************/
-+t_Error FM_PCD_ForceIntr (t_Handle h_FmPcd, e_FmPcdExceptions exception);
-+
-+#if (defined(DEBUG_ERRORS) && (DEBUG_ERRORS > 0))
-+/**************************************************************************//**
-+ @Function FM_PCD_DumpRegs
-+
-+ @Description Dumps all PCD registers
-+
-+ @Param[in] h_FmPcd A handle to an FM PCD Module.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_Init().
-+ NOTE: this routine may be called only for FM in master mode
-+ (i.e. 'guestId'=NCSW_MASTER_ID) or in a case that the registers
-+ are mapped.
-+*//***************************************************************************/
-+t_Error FM_PCD_DumpRegs(t_Handle h_FmPcd);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_KgDumpRegs
-+
-+ @Description Dumps all PCD KG registers
-+
-+ @Param[in] h_FmPcd A handle to an FM PCD Module.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_Init().
-+ NOTE: this routine may be called only for FM in master mode
-+ (i.e. 'guestId'=NCSW_MASTER_ID) or in a case that the registers
-+ are mapped.
-+*//***************************************************************************/
-+t_Error FM_PCD_KgDumpRegs(t_Handle h_FmPcd);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_PlcrDumpRegs
-+
-+ @Description Dumps all PCD Policer registers
-+
-+ @Param[in] h_FmPcd A handle to an FM PCD Module.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_Init().
-+ NOTE: this routine may be called only for FM in master mode
-+ (i.e. 'guestId'=NCSW_MASTER_ID) or in a case that the registers
-+ are mapped.
-+*//***************************************************************************/
-+t_Error FM_PCD_PlcrDumpRegs(t_Handle h_FmPcd);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_PlcrProfileDumpRegs
-+
-+ @Description Dumps all PCD Policer profile registers
-+
-+ @Param[in] h_Profile A handle to a Policer profile.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_Init().
-+ NOTE: this routine may be called only for FM in master mode
-+ (i.e. 'guestId'=NCSW_MASTER_ID) or in a case that the registers
-+ are mapped.
-+*//***************************************************************************/
-+t_Error FM_PCD_PlcrProfileDumpRegs(t_Handle h_Profile);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_PrsDumpRegs
-+
-+ @Description Dumps all PCD Parser registers
-+
-+ @Param[in] h_FmPcd A handle to an FM PCD Module.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_Init().
-+ NOTE: this routine may be called only for FM in master mode
-+ (i.e. 'guestId'=NCSW_MASTER_ID) or in a case that the registers
-+ are mapped.
-+*//***************************************************************************/
-+t_Error FM_PCD_PrsDumpRegs(t_Handle h_FmPcd);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_HcDumpRegs
-+
-+ @Description Dumps HC Port registers
-+
-+ @Param[in] h_FmPcd A handle to an FM PCD Module.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_Init().
-+ NOTE: this routine may be called only for FM in master mode
-+ (i.e. 'guestId'=NCSW_MASTER_ID).
-+*//***************************************************************************/
-+t_Error FM_PCD_HcDumpRegs(t_Handle h_FmPcd);
-+#endif /* (defined(DEBUG_ERRORS) && ... */
-+
-+
-+
-+/**************************************************************************//**
-+ KeyGen FM_PCD_Runtime_build_grp FM PCD Runtime Building Unit
-+
-+ @Description Frame Manager PCD Runtime Building API
-+
-+ This group contains routines for setting, deleting and modifying
-+ PCD resources, for defining the total PCD tree.
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Collection Definitions of coarse classification
-+ parameters as required by KeyGen (when coarse classification
-+ is the next engine after this scheme).
-+*//***************************************************************************/
-+#define FM_PCD_MAX_NUM_OF_CC_TREES 8
-+#define FM_PCD_MAX_NUM_OF_CC_GROUPS 16
-+#define FM_PCD_MAX_NUM_OF_CC_UNITS 4
-+#define FM_PCD_MAX_NUM_OF_KEYS 256
-+#define FM_PCD_MAX_NUM_OF_FLOWS (4*KILOBYTE)
-+#define FM_PCD_MAX_SIZE_OF_KEY 56
-+#define FM_PCD_MAX_NUM_OF_CC_ENTRIES_IN_GRP 16
-+#define FM_PCD_LAST_KEY_INDEX 0xffff
-+
-+#define FM_PCD_MAX_NUM_OF_CC_NODES 255 /* Obsolete, not used - will be removed in the future */
-+/* @} */
-+
-+/**************************************************************************//**
-+ @Collection A set of definitions to allow protocol
-+ special option description.
-+*//***************************************************************************/
-+typedef uint32_t protocolOpt_t; /**< A general type to define a protocol option. */
-+
-+typedef protocolOpt_t ethProtocolOpt_t; /**< Ethernet protocol options. */
-+#define ETH_BROADCAST 0x80000000 /**< Ethernet Broadcast. */
-+#define ETH_MULTICAST 0x40000000 /**< Ethernet Multicast. */
-+
-+typedef protocolOpt_t vlanProtocolOpt_t; /**< VLAN protocol options. */
-+#define VLAN_STACKED 0x20000000 /**< Stacked VLAN. */
-+
-+typedef protocolOpt_t mplsProtocolOpt_t; /**< MPLS protocol options. */
-+#define MPLS_STACKED 0x10000000 /**< Stacked MPLS. */
-+
-+typedef protocolOpt_t ipv4ProtocolOpt_t; /**< IPv4 protocol options. */
-+#define IPV4_BROADCAST_1 0x08000000 /**< IPv4 Broadcast. */
-+#define IPV4_MULTICAST_1 0x04000000 /**< IPv4 Multicast. */
-+#define IPV4_UNICAST_2 0x02000000 /**< Tunneled IPv4 - Unicast. */
-+#define IPV4_MULTICAST_BROADCAST_2 0x01000000 /**< Tunneled IPv4 - Broadcast/Multicast. */
-+
-+#define IPV4_FRAG_1 0x00000008 /**< IPV4 reassembly option.
-+ IPV4 Reassembly manipulation requires network
-+ environment with IPV4 header and IPV4_FRAG_1 option */
-+
-+typedef protocolOpt_t ipv6ProtocolOpt_t; /**< IPv6 protocol options. */
-+#define IPV6_MULTICAST_1 0x00800000 /**< IPv6 Multicast. */
-+#define IPV6_UNICAST_2 0x00400000 /**< Tunneled IPv6 - Unicast. */
-+#define IPV6_MULTICAST_2 0x00200000 /**< Tunneled IPv6 - Multicast. */
-+
-+#define IPV6_FRAG_1 0x00000004 /**< IPV6 reassembly option.
-+ IPV6 Reassembly manipulation requires network
-+ environment with IPV6 header and IPV6_FRAG_1 option;
-+ in case where fragment found, the fragment-extension offset
-+ may be found at 'shim2' (in parser-result). */
-+#if (DPAA_VERSION >= 11)
-+typedef protocolOpt_t capwapProtocolOpt_t; /**< CAPWAP protocol options. */
-+#define CAPWAP_FRAG_1 0x00000008 /**< CAPWAP reassembly option.
-+ CAPWAP Reassembly manipulation requires network
-+ environment with CAPWAP header and CAPWAP_FRAG_1 option;
-+ in case where fragment found, the fragment-extension offset
-+ may be found at 'shim2' (in parser-result). */
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+
-+/* @} */
-+
-+#define FM_PCD_MANIP_MAX_HDR_SIZE 256
-+#define FM_PCD_MANIP_DSCP_TO_VLAN_TRANS 64
-+
-+/**************************************************************************//**
-+ @Collection A set of definitions to support Header Manipulation selection.
-+*//***************************************************************************/
-+typedef uint32_t hdrManipFlags_t; /**< A general type to define a HMan update command flags. */
-+
-+typedef hdrManipFlags_t ipv4HdrManipUpdateFlags_t; /**< IPv4 protocol HMan update command flags. */
-+
-+#define HDR_MANIP_IPV4_TOS 0x80000000 /**< update TOS with the given value ('tos' field
-+ of t_FmPcdManipHdrFieldUpdateIpv4) */
-+#define HDR_MANIP_IPV4_ID 0x40000000 /**< update IP ID with the given value ('id' field
-+ of t_FmPcdManipHdrFieldUpdateIpv4) */
-+#define HDR_MANIP_IPV4_TTL 0x20000000 /**< Decrement TTL by 1 */
-+#define HDR_MANIP_IPV4_SRC 0x10000000 /**< update IP source address with the given value
-+ ('src' field of t_FmPcdManipHdrFieldUpdateIpv4) */
-+#define HDR_MANIP_IPV4_DST 0x08000000 /**< update IP destination address with the given value
-+ ('dst' field of t_FmPcdManipHdrFieldUpdateIpv4) */
-+
-+typedef hdrManipFlags_t ipv6HdrManipUpdateFlags_t; /**< IPv6 protocol HMan update command flags. */
-+
-+#define HDR_MANIP_IPV6_TC 0x80000000 /**< update Traffic Class address with the given value
-+ ('trafficClass' field of t_FmPcdManipHdrFieldUpdateIpv6) */
-+#define HDR_MANIP_IPV6_HL 0x40000000 /**< Decrement Hop Limit by 1 */
-+#define HDR_MANIP_IPV6_SRC 0x20000000 /**< update IP source address with the given value
-+ ('src' field of t_FmPcdManipHdrFieldUpdateIpv6) */
-+#define HDR_MANIP_IPV6_DST 0x10000000 /**< update IP destination address with the given value
-+ ('dst' field of t_FmPcdManipHdrFieldUpdateIpv6) */
-+
-+typedef hdrManipFlags_t tcpUdpHdrManipUpdateFlags_t;/**< TCP/UDP protocol HMan update command flags. */
-+
-+#define HDR_MANIP_TCP_UDP_SRC 0x80000000 /**< update TCP/UDP source address with the given value
-+ ('src' field of t_FmPcdManipHdrFieldUpdateTcpUdp) */
-+#define HDR_MANIP_TCP_UDP_DST 0x40000000 /**< update TCP/UDP destination address with the given value
-+ ('dst' field of t_FmPcdManipHdrFieldUpdateTcpUdp) */
-+#define HDR_MANIP_TCP_UDP_CHECKSUM 0x20000000 /**< update TCP/UDP checksum */
-+
-+/* @} */
-+
-+/**************************************************************************//**
-+ @Description A type used for returning the order of the key extraction.
-+ each value in this array represents the index of the extraction
-+ command as defined by the user in the initialization extraction array.
-+ The valid size of this array is the user define number of extractions
-+ required (also marked by the second '0' in this array).
-+*//***************************************************************************/
-+typedef uint8_t t_FmPcdKgKeyOrder [FM_PCD_KG_MAX_NUM_OF_EXTRACTS_PER_KEY];
-+
-+/**************************************************************************//**
-+ @Description All PCD engines
-+*//***************************************************************************/
-+typedef enum e_FmPcdEngine {
-+ e_FM_PCD_INVALID = 0, /**< Invalid PCD engine */
-+ e_FM_PCD_DONE, /**< No PCD Engine indicated */
-+ e_FM_PCD_KG, /**< KeyGen */
-+ e_FM_PCD_CC, /**< Coarse classifier */
-+ e_FM_PCD_PLCR, /**< Policer */
-+ e_FM_PCD_PRS, /**< Parser */
-+#if (DPAA_VERSION >= 11)
-+ e_FM_PCD_FR, /**< Frame-Replicator */
-+#endif /* (DPAA_VERSION >= 11) */
-+ e_FM_PCD_HASH /**< Hash table */
-+} e_FmPcdEngine;
-+
-+/**************************************************************************//**
-+ @Description Enumeration type for selecting extraction by header types
-+*//***************************************************************************/
-+typedef enum e_FmPcdExtractByHdrType {
-+ e_FM_PCD_EXTRACT_FROM_HDR, /**< Extract bytes from header */
-+ e_FM_PCD_EXTRACT_FROM_FIELD, /**< Extract bytes from header field */
-+ e_FM_PCD_EXTRACT_FULL_FIELD /**< Extract a full field */
-+} e_FmPcdExtractByHdrType;
-+
-+/**************************************************************************//**
-+ @Description Enumeration type for selecting extraction source
-+ (when it is not the header)
-+*//***************************************************************************/
-+typedef enum e_FmPcdExtractFrom {
-+ e_FM_PCD_EXTRACT_FROM_FRAME_START, /**< KG & CC: Extract from beginning of frame */
-+ e_FM_PCD_EXTRACT_FROM_DFLT_VALUE, /**< KG only: Extract from a default value */
-+ e_FM_PCD_EXTRACT_FROM_CURR_END_OF_PARSE, /**< KG & CC: Extract from the point where parsing had finished */
-+ e_FM_PCD_EXTRACT_FROM_KEY, /**< CC only: Field where saved KEY */
-+ e_FM_PCD_EXTRACT_FROM_HASH, /**< CC only: Field where saved HASH */
-+ e_FM_PCD_EXTRACT_FROM_PARSE_RESULT, /**< KG only: Extract from the parser result */
-+ e_FM_PCD_EXTRACT_FROM_ENQ_FQID, /**< KG & CC: Extract from enqueue FQID */
-+ e_FM_PCD_EXTRACT_FROM_FLOW_ID /**< CC only: Field where saved Dequeue FQID */
-+} e_FmPcdExtractFrom;
-+
-+/**************************************************************************//**
-+ @Description Enumeration type for selecting extraction type
-+*//***************************************************************************/
-+typedef enum e_FmPcdExtractType {
-+ e_FM_PCD_EXTRACT_BY_HDR, /**< Extract according to header */
-+ e_FM_PCD_EXTRACT_NON_HDR, /**< Extract from data that is not the header */
-+ e_FM_PCD_KG_EXTRACT_PORT_PRIVATE_INFO /**< Extract private info as specified by user */
-+} e_FmPcdExtractType;
-+
-+/**************************************************************************//**
-+ @Description Enumeration type for selecting default extraction value
-+*//***************************************************************************/
-+typedef enum e_FmPcdKgExtractDfltSelect {
-+ e_FM_PCD_KG_DFLT_GBL_0, /**< Default selection is KG register 0 */
-+ e_FM_PCD_KG_DFLT_GBL_1, /**< Default selection is KG register 1 */
-+ e_FM_PCD_KG_DFLT_PRIVATE_0, /**< Default selection is a per scheme register 0 */
-+ e_FM_PCD_KG_DFLT_PRIVATE_1, /**< Default selection is a per scheme register 1 */
-+ e_FM_PCD_KG_DFLT_ILLEGAL /**< Illegal selection */
-+} e_FmPcdKgExtractDfltSelect;
-+
-+/**************************************************************************//**
-+ @Description Enumeration type defining all default groups - each group shares
-+ a default value, one of four user-initialized values.
-+*//***************************************************************************/
-+typedef enum e_FmPcdKgKnownFieldsDfltTypes {
-+ e_FM_PCD_KG_MAC_ADDR, /**< MAC Address */
-+ e_FM_PCD_KG_TCI, /**< TCI field */
-+ e_FM_PCD_KG_ENET_TYPE, /**< ENET Type */
-+ e_FM_PCD_KG_PPP_SESSION_ID, /**< PPP Session id */
-+ e_FM_PCD_KG_PPP_PROTOCOL_ID, /**< PPP Protocol id */
-+ e_FM_PCD_KG_MPLS_LABEL, /**< MPLS label */
-+ e_FM_PCD_KG_IP_ADDR, /**< IP address */
-+ e_FM_PCD_KG_PROTOCOL_TYPE, /**< Protocol type */
-+ e_FM_PCD_KG_IP_TOS_TC, /**< TOS or TC */
-+ e_FM_PCD_KG_IPV6_FLOW_LABEL, /**< IPV6 flow label */
-+ e_FM_PCD_KG_IPSEC_SPI, /**< IPSEC SPI */
-+ e_FM_PCD_KG_L4_PORT, /**< L4 Port */
-+ e_FM_PCD_KG_TCP_FLAG, /**< TCP Flag */
-+ e_FM_PCD_KG_GENERIC_FROM_DATA, /**< grouping implemented by SW,
-+ any data extraction that is not the full
-+ field described above */
-+ e_FM_PCD_KG_GENERIC_FROM_DATA_NO_V, /**< grouping implemented by SW,
-+ any data extraction without validation */
-+ e_FM_PCD_KG_GENERIC_NOT_FROM_DATA /**< grouping implemented by SW,
-+ extraction from parser result or
-+ direct use of default value */
-+} e_FmPcdKgKnownFieldsDfltTypes;
-+
-+/**************************************************************************//**
-+ @Description Enumeration type for defining header index for scenarios with
-+ multiple (tunneled) headers
-+*//***************************************************************************/
-+typedef enum e_FmPcdHdrIndex {
-+ e_FM_PCD_HDR_INDEX_NONE = 0, /**< used when multiple headers not used, also
-+ to specify regular IP (not tunneled). */
-+ e_FM_PCD_HDR_INDEX_1, /**< may be used for VLAN, MPLS, tunneled IP */
-+ e_FM_PCD_HDR_INDEX_2, /**< may be used for MPLS, tunneled IP */
-+ e_FM_PCD_HDR_INDEX_3, /**< may be used for MPLS */
-+ e_FM_PCD_HDR_INDEX_LAST = 0xFF /**< may be used for VLAN, MPLS */
-+} e_FmPcdHdrIndex;
-+
-+/**************************************************************************//**
-+ @Description Enumeration type for selecting the policer profile functional type
-+*//***************************************************************************/
-+typedef enum e_FmPcdProfileTypeSelection {
-+ e_FM_PCD_PLCR_PORT_PRIVATE, /**< Port dedicated profile */
-+ e_FM_PCD_PLCR_SHARED /**< Shared profile (shared within partition) */
-+} e_FmPcdProfileTypeSelection;
-+
-+/**************************************************************************//**
-+ @Description Enumeration type for selecting the policer profile algorithm
-+*//***************************************************************************/
-+typedef enum e_FmPcdPlcrAlgorithmSelection {
-+ e_FM_PCD_PLCR_PASS_THROUGH, /**< Policer pass through */
-+ e_FM_PCD_PLCR_RFC_2698, /**< Policer algorithm RFC 2698 */
-+ e_FM_PCD_PLCR_RFC_4115 /**< Policer algorithm RFC 4115 */
-+} e_FmPcdPlcrAlgorithmSelection;
-+
-+/**************************************************************************//**
-+ @Description Enumeration type for selecting a policer profile color mode
-+*//***************************************************************************/
-+typedef enum e_FmPcdPlcrColorMode {
-+ e_FM_PCD_PLCR_COLOR_BLIND, /**< Color blind */
-+ e_FM_PCD_PLCR_COLOR_AWARE /**< Color aware */
-+} e_FmPcdPlcrColorMode;
-+
-+/**************************************************************************//**
-+ @Description Enumeration type for selecting a policer profile color
-+*//***************************************************************************/
-+typedef enum e_FmPcdPlcrColor {
-+ e_FM_PCD_PLCR_GREEN, /**< Green color code */
-+ e_FM_PCD_PLCR_YELLOW, /**< Yellow color code */
-+ e_FM_PCD_PLCR_RED, /**< Red color code */
-+ e_FM_PCD_PLCR_OVERRIDE /**< Color override code */
-+} e_FmPcdPlcrColor;
-+
-+/**************************************************************************//**
-+ @Description Enumeration type for selecting the policer profile packet frame length selector
-+*//***************************************************************************/
-+typedef enum e_FmPcdPlcrFrameLengthSelect {
-+ e_FM_PCD_PLCR_L2_FRM_LEN, /**< L2 frame length */
-+ e_FM_PCD_PLCR_L3_FRM_LEN, /**< L3 frame length */
-+ e_FM_PCD_PLCR_L4_FRM_LEN, /**< L4 frame length */
-+ e_FM_PCD_PLCR_FULL_FRM_LEN /**< Full frame length */
-+} e_FmPcdPlcrFrameLengthSelect;
-+
-+/**************************************************************************//**
-+ @Description Enumeration type for selecting roll-back frame
-+*//***************************************************************************/
-+typedef enum e_FmPcdPlcrRollBackFrameSelect {
-+ e_FM_PCD_PLCR_ROLLBACK_L2_FRM_LEN, /**< Roll-back L2 frame length */
-+ e_FM_PCD_PLCR_ROLLBACK_FULL_FRM_LEN /**< Roll-back Full frame length */
-+} e_FmPcdPlcrRollBackFrameSelect;
-+
-+/**************************************************************************//**
-+ @Description Enumeration type for selecting the policer profile packet or byte mode
-+*//***************************************************************************/
-+typedef enum e_FmPcdPlcrRateMode {
-+ e_FM_PCD_PLCR_BYTE_MODE, /**< Byte mode */
-+ e_FM_PCD_PLCR_PACKET_MODE /**< Packet mode */
-+} e_FmPcdPlcrRateMode;
-+
-+/**************************************************************************//**
-+ @Description Enumeration type for defining action of frame
-+*//***************************************************************************/
-+typedef enum e_FmPcdDoneAction {
-+ e_FM_PCD_ENQ_FRAME = 0, /**< Enqueue frame */
-+ e_FM_PCD_DROP_FRAME /**< Mark this frame as error frame and continue
-+ to error flow; 'FM_PORT_FRM_ERR_CLS_DISCARD'
-+ flag will be set for this frame. */
-+} e_FmPcdDoneAction;
-+
-+/**************************************************************************//**
-+ @Description Enumeration type for selecting the policer counter
-+*//***************************************************************************/
-+typedef enum e_FmPcdPlcrProfileCounters {
-+ e_FM_PCD_PLCR_PROFILE_GREEN_PACKET_TOTAL_COUNTER, /**< Green packets counter */
-+ e_FM_PCD_PLCR_PROFILE_YELLOW_PACKET_TOTAL_COUNTER, /**< Yellow packets counter */
-+ e_FM_PCD_PLCR_PROFILE_RED_PACKET_TOTAL_COUNTER, /**< Red packets counter */
-+ e_FM_PCD_PLCR_PROFILE_RECOLOURED_YELLOW_PACKET_TOTAL_COUNTER, /**< Recolored yellow packets counter */
-+ e_FM_PCD_PLCR_PROFILE_RECOLOURED_RED_PACKET_TOTAL_COUNTER /**< Recolored red packets counter */
-+} e_FmPcdPlcrProfileCounters;
-+
-+/**************************************************************************//**
-+ @Description Enumeration type for selecting the PCD action after extraction
-+*//***************************************************************************/
-+typedef enum e_FmPcdAction {
-+ e_FM_PCD_ACTION_NONE, /**< NONE */
-+ e_FM_PCD_ACTION_EXACT_MATCH, /**< Exact match on the selected extraction */
-+ e_FM_PCD_ACTION_INDEXED_LOOKUP /**< Indexed lookup on the selected extraction */
-+} e_FmPcdAction;
-+
-+/**************************************************************************//**
-+ @Description Enumeration type for selecting type of insert manipulation
-+*//***************************************************************************/
-+typedef enum e_FmPcdManipHdrInsrtType {
-+ e_FM_PCD_MANIP_INSRT_GENERIC, /**< Insert according to offset & size */
-+ e_FM_PCD_MANIP_INSRT_BY_HDR, /**< Insert according to protocol */
-+#if ((DPAA_VERSION == 10) && defined(FM_CAPWAP_SUPPORT))
-+ e_FM_PCD_MANIP_INSRT_BY_TEMPLATE /**< Insert template to start of frame */
-+#endif /* ((DPAA_VERSION == 10) && defined(FM_CAPWAP_SUPPORT)) */
-+} e_FmPcdManipHdrInsrtType;
-+
-+/**************************************************************************//**
-+ @Description Enumeration type for selecting type of remove manipulation
-+*//***************************************************************************/
-+typedef enum e_FmPcdManipHdrRmvType {
-+ e_FM_PCD_MANIP_RMV_GENERIC, /**< Remove according to offset & size */
-+ e_FM_PCD_MANIP_RMV_BY_HDR /**< Remove according to offset & size */
-+} e_FmPcdManipHdrRmvType;
-+
-+/**************************************************************************//**
-+ @Description Enumeration type for selecting specific L2 fields removal
-+*//***************************************************************************/
-+typedef enum e_FmPcdManipHdrRmvSpecificL2 {
-+ e_FM_PCD_MANIP_HDR_RMV_ETHERNET, /**< Ethernet/802.3 MAC */
-+ e_FM_PCD_MANIP_HDR_RMV_STACKED_QTAGS, /**< stacked QTags */
-+ e_FM_PCD_MANIP_HDR_RMV_ETHERNET_AND_MPLS, /**< MPLS and Ethernet/802.3 MAC header until
-+ the header which follows the MPLS header */
-+ e_FM_PCD_MANIP_HDR_RMV_MPLS, /**< Remove MPLS header (Unlimited MPLS labels) */
-+ e_FM_PCD_MANIP_HDR_RMV_PPPOE /**< Remove the PPPoE header and PPP protocol field. */
-+} e_FmPcdManipHdrRmvSpecificL2;
-+
-+/**************************************************************************//**
-+ @Description Enumeration type for selecting specific fields updates
-+*//***************************************************************************/
-+typedef enum e_FmPcdManipHdrFieldUpdateType {
-+ e_FM_PCD_MANIP_HDR_FIELD_UPDATE_VLAN, /**< VLAN updates */
-+ e_FM_PCD_MANIP_HDR_FIELD_UPDATE_IPV4, /**< IPV4 updates */
-+ e_FM_PCD_MANIP_HDR_FIELD_UPDATE_IPV6, /**< IPV6 updates */
-+ e_FM_PCD_MANIP_HDR_FIELD_UPDATE_TCP_UDP, /**< TCP_UDP updates */
-+} e_FmPcdManipHdrFieldUpdateType;
-+
-+/**************************************************************************//**
-+ @Description Enumeration type for selecting VLAN updates
-+*//***************************************************************************/
-+typedef enum e_FmPcdManipHdrFieldUpdateVlan {
-+ e_FM_PCD_MANIP_HDR_FIELD_UPDATE_VLAN_VPRI, /**< Replace VPri of outer most VLAN tag. */
-+ e_FM_PCD_MANIP_HDR_FIELD_UPDATE_DSCP_TO_VLAN /**< DSCP to VLAN priority bits translation */
-+} e_FmPcdManipHdrFieldUpdateVlan;
-+
-+/**************************************************************************//**
-+ @Description Enumeration type for selecting specific L2 header insertion
-+*//***************************************************************************/
-+typedef enum e_FmPcdManipHdrInsrtSpecificL2 {
-+ e_FM_PCD_MANIP_HDR_INSRT_MPLS, /**< Insert MPLS header (Unlimited MPLS labels) */
-+ e_FM_PCD_MANIP_HDR_INSRT_PPPOE /**< Insert PPPOE */
-+} e_FmPcdManipHdrInsrtSpecificL2;
-+
-+#if (DPAA_VERSION >= 11)
-+/**************************************************************************//**
-+ @Description Enumeration type for selecting QoS mapping mode
-+
-+ Note: In all cases except 'e_FM_PCD_MANIP_HDR_QOS_MAPPING_NONE'
-+ User should instruct the port to read the hash-result
-+*//***************************************************************************/
-+typedef enum e_FmPcdManipHdrQosMappingMode {
-+ e_FM_PCD_MANIP_HDR_QOS_MAPPING_NONE = 0, /**< No mapping, QoS field will not be changed */
-+ e_FM_PCD_MANIP_HDR_QOS_MAPPING_AS_IS, /**< QoS field will be overwritten by the last byte in the hash-result. */
-+} e_FmPcdManipHdrQosMappingMode;
-+
-+/**************************************************************************//**
-+ @Description Enumeration type for selecting QoS source
-+
-+ Note: In all cases except 'e_FM_PCD_MANIP_HDR_QOS_SRC_NONE'
-+ User should left room for the hash-result on input/output buffer
-+ and instruct the port to read/write the hash-result to the buffer (RPD should be set)
-+*//***************************************************************************/
-+typedef enum e_FmPcdManipHdrQosSrc {
-+ e_FM_PCD_MANIP_HDR_QOS_SRC_NONE = 0, /**< TODO */
-+ e_FM_PCD_MANIP_HDR_QOS_SRC_USER_DEFINED, /**< QoS will be taken from the last byte in the hash-result. */
-+} e_FmPcdManipHdrQosSrc;
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+/**************************************************************************//**
-+ @Description Enumeration type for selecting type of header insertion
-+*//***************************************************************************/
-+typedef enum e_FmPcdManipHdrInsrtByHdrType {
-+ e_FM_PCD_MANIP_INSRT_BY_HDR_SPECIFIC_L2, /**< Specific L2 fields insertion */
-+#if (DPAA_VERSION >= 11)
-+ e_FM_PCD_MANIP_INSRT_BY_HDR_IP, /**< IP insertion */
-+ e_FM_PCD_MANIP_INSRT_BY_HDR_UDP, /**< UDP insertion */
-+ e_FM_PCD_MANIP_INSRT_BY_HDR_UDP_LITE, /**< UDP lite insertion */
-+ e_FM_PCD_MANIP_INSRT_BY_HDR_CAPWAP /**< CAPWAP insertion */
-+#endif /* (DPAA_VERSION >= 11) */
-+} e_FmPcdManipHdrInsrtByHdrType;
-+
-+/**************************************************************************//**
-+ @Description Enumeration type for selecting specific customCommand
-+*//***************************************************************************/
-+typedef enum e_FmPcdManipHdrCustomType {
-+ e_FM_PCD_MANIP_HDR_CUSTOM_IP_REPLACE, /**< Replace IPv4/IPv6 */
-+ e_FM_PCD_MANIP_HDR_CUSTOM_GEN_FIELD_REPLACE, /**< Replace IPv4/IPv6 */
-+} e_FmPcdManipHdrCustomType;
-+
-+/**************************************************************************//**
-+ @Description Enumeration type for selecting specific customCommand
-+*//***************************************************************************/
-+typedef enum e_FmPcdManipHdrCustomIpReplace {
-+ e_FM_PCD_MANIP_HDR_CUSTOM_REPLACE_IPV4_BY_IPV6, /**< Replace IPv4 by IPv6 */
-+ e_FM_PCD_MANIP_HDR_CUSTOM_REPLACE_IPV6_BY_IPV4 /**< Replace IPv6 by IPv4 */
-+} e_FmPcdManipHdrCustomIpReplace;
-+
-+/**************************************************************************//**
-+ @Description Enumeration type for selecting type of header removal
-+*//***************************************************************************/
-+typedef enum e_FmPcdManipHdrRmvByHdrType {
-+ e_FM_PCD_MANIP_RMV_BY_HDR_SPECIFIC_L2 = 0, /**< Specific L2 fields removal */
-+#if (DPAA_VERSION >= 11)
-+ e_FM_PCD_MANIP_RMV_BY_HDR_CAPWAP, /**< CAPWAP removal */
-+#endif /* (DPAA_VERSION >= 11) */
-+#if (DPAA_VERSION >= 11) || ((DPAA_VERSION == 10) && defined(FM_CAPWAP_SUPPORT))
-+ e_FM_PCD_MANIP_RMV_BY_HDR_FROM_START, /**< Locate from data that is not the header */
-+#endif /* (DPAA_VERSION >= 11) || ((DPAA_VERSION == 10) && defined(FM_CAPWAP_SUPPORT)) */
-+} e_FmPcdManipHdrRmvByHdrType;
-+
-+/**************************************************************************//**
-+ @Description Enumeration type for selecting type of timeout mode
-+*//***************************************************************************/
-+typedef enum e_FmPcdManipReassemTimeOutMode {
-+ e_FM_PCD_MANIP_TIME_OUT_BETWEEN_FRAMES, /**< Limits the time of the reassembly process
-+ from the first fragment to the last */
-+ e_FM_PCD_MANIP_TIME_OUT_BETWEEN_FRAG /**< Limits the time of receiving the fragment */
-+} e_FmPcdManipReassemTimeOutMode;
-+
-+/**************************************************************************//**
-+ @Description Enumeration type for selecting type of WaysNumber mode
-+*//***************************************************************************/
-+typedef enum e_FmPcdManipReassemWaysNumber {
-+ e_FM_PCD_MANIP_ONE_WAY_HASH = 1, /**< One way hash */
-+ e_FM_PCD_MANIP_TWO_WAYS_HASH, /**< Two ways hash */
-+ e_FM_PCD_MANIP_THREE_WAYS_HASH, /**< Three ways hash */
-+ e_FM_PCD_MANIP_FOUR_WAYS_HASH, /**< Four ways hash */
-+ e_FM_PCD_MANIP_FIVE_WAYS_HASH, /**< Five ways hash */
-+ e_FM_PCD_MANIP_SIX_WAYS_HASH, /**< Six ways hash */
-+ e_FM_PCD_MANIP_SEVEN_WAYS_HASH, /**< Seven ways hash */
-+ e_FM_PCD_MANIP_EIGHT_WAYS_HASH /**< Eight ways hash */
-+} e_FmPcdManipReassemWaysNumber;
-+
-+#if ((DPAA_VERSION == 10) && defined(FM_CAPWAP_SUPPORT))
-+/**************************************************************************//**
-+ @Description Enumeration type for selecting type of statistics mode
-+*//***************************************************************************/
-+typedef enum e_FmPcdStatsType {
-+ e_FM_PCD_STATS_PER_FLOWID = 0 /**< Flow ID is used as index for getting statistics */
-+} e_FmPcdStatsType;
-+#endif /* ((DPAA_VERSION == 10) && defined(FM_CAPWAP_SUPPORT)) */
-+
-+/**************************************************************************//**
-+ @Description Enumeration type for selecting manipulation type
-+*//***************************************************************************/
-+typedef enum e_FmPcdManipType {
-+ e_FM_PCD_MANIP_HDR = 0, /**< Header manipulation */
-+ e_FM_PCD_MANIP_REASSEM, /**< Reassembly */
-+ e_FM_PCD_MANIP_FRAG, /**< Fragmentation */
-+ e_FM_PCD_MANIP_SPECIAL_OFFLOAD /**< Special Offloading */
-+} e_FmPcdManipType;
-+
-+/**************************************************************************//**
-+ @Description Enumeration type for selecting type of statistics mode
-+*//***************************************************************************/
-+typedef enum e_FmPcdCcStatsMode {
-+ e_FM_PCD_CC_STATS_MODE_NONE = 0, /**< No statistics support */
-+ e_FM_PCD_CC_STATS_MODE_FRAME, /**< Frame count statistics */
-+ e_FM_PCD_CC_STATS_MODE_BYTE_AND_FRAME, /**< Byte and frame count statistics */
-+#if (DPAA_VERSION >= 11)
-+ e_FM_PCD_CC_STATS_MODE_RMON, /**< Byte and frame length range count statistics;
-+ This mode is supported only on B4860 device */
-+#endif /* (DPAA_VERSION >= 11) */
-+} e_FmPcdCcStatsMode;
-+
-+/**************************************************************************//**
-+ @Description Enumeration type for determining the action in case an IP packet
-+ is larger than MTU but its DF (Don't Fragment) bit is set.
-+*//***************************************************************************/
-+typedef enum e_FmPcdManipDontFragAction {
-+ e_FM_PCD_MANIP_DISCARD_PACKET = 0, /**< Discard packet */
-+ e_FM_PCD_MANIP_ENQ_TO_ERR_Q_OR_DISCARD_PACKET = e_FM_PCD_MANIP_DISCARD_PACKET,
-+ /**< Obsolete, cannot enqueue to error queue;
-+ In practice, selects to discard packets;
-+ Will be removed in the future */
-+ e_FM_PCD_MANIP_FRAGMENT_PACKET, /**< Fragment packet and continue normal processing */
-+ e_FM_PCD_MANIP_CONTINUE_WITHOUT_FRAG /**< Continue normal processing without fragmenting the packet */
-+} e_FmPcdManipDontFragAction;
-+
-+/**************************************************************************//**
-+ @Description Enumeration type for selecting type of special offload manipulation
-+*//***************************************************************************/
-+typedef enum e_FmPcdManipSpecialOffloadType {
-+ e_FM_PCD_MANIP_SPECIAL_OFFLOAD_IPSEC, /**< IPSec offload manipulation */
-+#if (DPAA_VERSION >= 11)
-+ e_FM_PCD_MANIP_SPECIAL_OFFLOAD_CAPWAP /**< CAPWAP offload manipulation */
-+#endif /* (DPAA_VERSION >= 11) */
-+} e_FmPcdManipSpecialOffloadType;
-+
-+
-+/**************************************************************************//**
-+ @Description A Union of protocol dependent special options
-+*//***************************************************************************/
-+typedef union u_FmPcdHdrProtocolOpt {
-+ ethProtocolOpt_t ethOpt; /**< Ethernet options */
-+ vlanProtocolOpt_t vlanOpt; /**< VLAN options */
-+ mplsProtocolOpt_t mplsOpt; /**< MPLS options */
-+ ipv4ProtocolOpt_t ipv4Opt; /**< IPv4 options */
-+ ipv6ProtocolOpt_t ipv6Opt; /**< IPv6 options */
-+#if (DPAA_VERSION >= 11)
-+ capwapProtocolOpt_t capwapOpt; /**< CAPWAP options */
-+#endif /* (DPAA_VERSION >= 11) */
-+} u_FmPcdHdrProtocolOpt;
-+
-+/**************************************************************************//**
-+ @Description A union holding protocol fields
-+
-+
-+ Fields supported as "full fields":
-+ HEADER_TYPE_ETH:
-+ NET_HEADER_FIELD_ETH_DA
-+ NET_HEADER_FIELD_ETH_SA
-+ NET_HEADER_FIELD_ETH_TYPE
-+
-+ HEADER_TYPE_LLC_SNAP:
-+ NET_HEADER_FIELD_LLC_SNAP_TYPE
-+
-+ HEADER_TYPE_VLAN:
-+ NET_HEADER_FIELD_VLAN_TCI
-+ (index may apply:
-+ e_FM_PCD_HDR_INDEX_NONE/e_FM_PCD_HDR_INDEX_1,
-+ e_FM_PCD_HDR_INDEX_LAST)
-+
-+ HEADER_TYPE_MPLS:
-+ NET_HEADER_FIELD_MPLS_LABEL_STACK
-+ (index may apply:
-+ e_FM_PCD_HDR_INDEX_NONE/e_FM_PCD_HDR_INDEX_1,
-+ e_FM_PCD_HDR_INDEX_2,
-+ e_FM_PCD_HDR_INDEX_LAST)
-+
-+ HEADER_TYPE_IPv4:
-+ NET_HEADER_FIELD_IPv4_SRC_IP
-+ NET_HEADER_FIELD_IPv4_DST_IP
-+ NET_HEADER_FIELD_IPv4_PROTO
-+ NET_HEADER_FIELD_IPv4_TOS
-+ (index may apply:
-+ e_FM_PCD_HDR_INDEX_NONE/e_FM_PCD_HDR_INDEX_1,
-+ e_FM_PCD_HDR_INDEX_2/e_FM_PCD_HDR_INDEX_LAST)
-+
-+ HEADER_TYPE_IPv6:
-+ NET_HEADER_FIELD_IPv6_SRC_IP
-+ NET_HEADER_FIELD_IPv6_DST_IP
-+ NET_HEADER_FIELD_IPv6_NEXT_HDR
-+ NET_HEADER_FIELD_IPv6_VER | NET_HEADER_FIELD_IPv6_FL | NET_HEADER_FIELD_IPv6_TC (must come together!)
-+ (index may apply:
-+ e_FM_PCD_HDR_INDEX_NONE/e_FM_PCD_HDR_INDEX_1,
-+ e_FM_PCD_HDR_INDEX_2/e_FM_PCD_HDR_INDEX_LAST)
-+
-+ (Note that starting from DPAA 1-1, NET_HEADER_FIELD_IPv6_NEXT_HDR applies to
-+ the last next header indication, meaning the next L4, which may be
-+ present at the Ipv6 last extension. On earlier revisions this field
-+ applies to the Next-Header field of the main IPv6 header)
-+
-+ HEADER_TYPE_IP:
-+ NET_HEADER_FIELD_IP_PROTO
-+ (index may apply:
-+ e_FM_PCD_HDR_INDEX_LAST)
-+ NET_HEADER_FIELD_IP_DSCP
-+ (index may apply:
-+ e_FM_PCD_HDR_INDEX_NONE/e_FM_PCD_HDR_INDEX_1)
-+ HEADER_TYPE_GRE:
-+ NET_HEADER_FIELD_GRE_TYPE
-+
-+ HEADER_TYPE_MINENCAP
-+ NET_HEADER_FIELD_MINENCAP_SRC_IP
-+ NET_HEADER_FIELD_MINENCAP_DST_IP
-+ NET_HEADER_FIELD_MINENCAP_TYPE
-+
-+ HEADER_TYPE_TCP:
-+ NET_HEADER_FIELD_TCP_PORT_SRC
-+ NET_HEADER_FIELD_TCP_PORT_DST
-+ NET_HEADER_FIELD_TCP_FLAGS
-+
-+ HEADER_TYPE_UDP:
-+ NET_HEADER_FIELD_UDP_PORT_SRC
-+ NET_HEADER_FIELD_UDP_PORT_DST
-+
-+ HEADER_TYPE_UDP_LITE:
-+ NET_HEADER_FIELD_UDP_LITE_PORT_SRC
-+ NET_HEADER_FIELD_UDP_LITE_PORT_DST
-+
-+ HEADER_TYPE_IPSEC_AH:
-+ NET_HEADER_FIELD_IPSEC_AH_SPI
-+ NET_HEADER_FIELD_IPSEC_AH_NH
-+
-+ HEADER_TYPE_IPSEC_ESP:
-+ NET_HEADER_FIELD_IPSEC_ESP_SPI
-+
-+ HEADER_TYPE_SCTP:
-+ NET_HEADER_FIELD_SCTP_PORT_SRC
-+ NET_HEADER_FIELD_SCTP_PORT_DST
-+
-+ HEADER_TYPE_DCCP:
-+ NET_HEADER_FIELD_DCCP_PORT_SRC
-+ NET_HEADER_FIELD_DCCP_PORT_DST
-+
-+ HEADER_TYPE_PPPoE:
-+ NET_HEADER_FIELD_PPPoE_PID
-+ NET_HEADER_FIELD_PPPoE_SID
-+
-+ *****************************************************************
-+ Fields supported as "from fields":
-+ HEADER_TYPE_ETH (with or without validation):
-+ NET_HEADER_FIELD_ETH_TYPE
-+
-+ HEADER_TYPE_VLAN (with or without validation):
-+ NET_HEADER_FIELD_VLAN_TCI
-+ (index may apply:
-+ e_FM_PCD_HDR_INDEX_NONE/e_FM_PCD_HDR_INDEX_1,
-+ e_FM_PCD_HDR_INDEX_LAST)
-+
-+ HEADER_TYPE_IPv4 (without validation):
-+ NET_HEADER_FIELD_IPv4_PROTO
-+ (index may apply:
-+ e_FM_PCD_HDR_INDEX_NONE/e_FM_PCD_HDR_INDEX_1,
-+ e_FM_PCD_HDR_INDEX_2/e_FM_PCD_HDR_INDEX_LAST)
-+
-+ HEADER_TYPE_IPv6 (without validation):
-+ NET_HEADER_FIELD_IPv6_NEXT_HDR
-+ (index may apply:
-+ e_FM_PCD_HDR_INDEX_NONE/e_FM_PCD_HDR_INDEX_1,
-+ e_FM_PCD_HDR_INDEX_2/e_FM_PCD_HDR_INDEX_LAST)
-+
-+*//***************************************************************************/
-+typedef union t_FmPcdFields {
-+ headerFieldEth_t eth; /**< Ethernet */
-+ headerFieldVlan_t vlan; /**< VLAN */
-+ headerFieldLlcSnap_t llcSnap; /**< LLC SNAP */
-+ headerFieldPppoe_t pppoe; /**< PPPoE */
-+ headerFieldMpls_t mpls; /**< MPLS */
-+ headerFieldIp_t ip; /**< IP */
-+ headerFieldIpv4_t ipv4; /**< IPv4 */
-+ headerFieldIpv6_t ipv6; /**< IPv6 */
-+ headerFieldUdp_t udp; /**< UDP */
-+ headerFieldUdpLite_t udpLite; /**< UDP Lite */
-+ headerFieldTcp_t tcp; /**< TCP */
-+ headerFieldSctp_t sctp; /**< SCTP */
-+ headerFieldDccp_t dccp; /**< DCCP */
-+ headerFieldGre_t gre; /**< GRE */
-+ headerFieldMinencap_t minencap; /**< Minimal Encapsulation */
-+ headerFieldIpsecAh_t ipsecAh; /**< IPSec AH */
-+ headerFieldIpsecEsp_t ipsecEsp; /**< IPSec ESP */
-+ headerFieldUdpEncapEsp_t udpEncapEsp; /**< UDP Encapsulation ESP */
-+} t_FmPcdFields;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining header extraction for key generation
-+*//***************************************************************************/
-+typedef struct t_FmPcdFromHdr {
-+ uint8_t size; /**< Size in byte */
-+ uint8_t offset; /**< Byte offset */
-+} t_FmPcdFromHdr;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining field extraction for key generation
-+*//***************************************************************************/
-+typedef struct t_FmPcdFromField {
-+ t_FmPcdFields field; /**< Field selection */
-+ uint8_t size; /**< Size in byte */
-+ uint8_t offset; /**< Byte offset */
-+} t_FmPcdFromField;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining a single network environment unit
-+
-+ A distinction unit should be defined if it will later be used
-+ by one or more PCD engines to distinguish between flows.
-+*//***************************************************************************/
-+typedef struct t_FmPcdDistinctionUnit {
-+ struct {
-+ e_NetHeaderType hdr; /**< One of the headers supported by the FM */
-+ u_FmPcdHdrProtocolOpt opt; /**< Select only one option ! */
-+ } hdrs[FM_PCD_MAX_NUM_OF_INTERCHANGEABLE_HDRS];
-+} t_FmPcdDistinctionUnit;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining all different distinction units supported
-+ by a specific PCD Network Environment Characteristics module.
-+
-+ Each unit represent a protocol or a group of protocols that may
-+ be used later by the different PCD engines to distinguish
-+ between flows.
-+*//***************************************************************************/
-+typedef struct t_FmPcdNetEnvParams {
-+ uint8_t numOfDistinctionUnits; /**< Number of different units to be identified */
-+ t_FmPcdDistinctionUnit units[FM_PCD_MAX_NUM_OF_DISTINCTION_UNITS]; /**< An array of numOfDistinctionUnits of the
-+ different units to be identified */
-+} t_FmPcdNetEnvParams;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining a single extraction action when
-+ creating a key
-+*//***************************************************************************/
-+typedef struct t_FmPcdExtractEntry {
-+ e_FmPcdExtractType type; /**< Extraction type select */
-+ union {
-+ struct {
-+ e_NetHeaderType hdr; /**< Header selection */
-+ bool ignoreProtocolValidation;
-+ /**< Ignore protocol validation */
-+ e_FmPcdHdrIndex hdrIndex; /**< Relevant only for MPLS, VLAN and tunneled
-+ IP. Otherwise should be cleared. */
-+ e_FmPcdExtractByHdrType type; /**< Header extraction type select */
-+ union {
-+ t_FmPcdFromHdr fromHdr; /**< Extract bytes from header parameters */
-+ t_FmPcdFromField fromField; /**< Extract bytes from field parameters */
-+ t_FmPcdFields fullField; /**< Extract full filed parameters */
-+ } extractByHdrType;
-+ } extractByHdr; /**< used when type = e_FM_PCD_KG_EXTRACT_BY_HDR */
-+ struct {
-+ e_FmPcdExtractFrom src; /**< Non-header extraction source */
-+ e_FmPcdAction action; /**< Relevant for CC Only */
-+ uint16_t icIndxMask; /**< Relevant only for CC when
-+ action = e_FM_PCD_ACTION_INDEXED_LOOKUP;
-+ Note that the number of bits that are set within
-+ this mask must be log2 of the CC-node 'numOfKeys'.
-+ Note that the mask cannot be set on the lower bits. */
-+ uint8_t offset; /**< Byte offset */
-+ uint8_t size; /**< Size in byte */
-+ } extractNonHdr; /**< used when type = e_FM_PCD_KG_EXTRACT_NON_HDR */
-+ };
-+} t_FmPcdExtractEntry;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining masks for each extracted field in the key.
-+*//***************************************************************************/
-+typedef struct t_FmPcdKgExtractMask {
-+ uint8_t extractArrayIndex; /**< Index in the extraction array, as initialized by user */
-+ uint8_t offset; /**< Byte offset */
-+ uint8_t mask; /**< A byte mask (selected bits will be used) */
-+} t_FmPcdKgExtractMask;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining default selection per groups of fields
-+*//***************************************************************************/
-+typedef struct t_FmPcdKgExtractDflt {
-+ e_FmPcdKgKnownFieldsDfltTypes type; /**< Default type select */
-+ e_FmPcdKgExtractDfltSelect dfltSelect; /**< Default register select */
-+} t_FmPcdKgExtractDflt;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining key extraction and hashing
-+*//***************************************************************************/
-+typedef struct t_FmPcdKgKeyExtractAndHashParams {
-+ uint32_t privateDflt0; /**< Scheme default register 0 */
-+ uint32_t privateDflt1; /**< Scheme default register 1 */
-+ uint8_t numOfUsedExtracts; /**< defines the valid size of the following array */
-+ t_FmPcdExtractEntry extractArray [FM_PCD_KG_MAX_NUM_OF_EXTRACTS_PER_KEY]; /**< An array of extractions definition. */
-+ uint8_t numOfUsedDflts; /**< defines the valid size of the following array */
-+ t_FmPcdKgExtractDflt dflts[FM_PCD_KG_NUM_OF_DEFAULT_GROUPS];
-+ /**< For each extraction used in this scheme, specify the required
-+ default register to be used when header is not found.
-+ types not specified in this array will get undefined value. */
-+ uint8_t numOfUsedMasks; /**< defines the valid size of the following array */
-+ t_FmPcdKgExtractMask masks[FM_PCD_KG_NUM_OF_EXTRACT_MASKS];
-+ uint8_t hashShift; /**< hash result right shift. Select the 24 bits out of the 64 hash
-+ result. 0 means using the 24 LSB's, otherwise use the
-+ 24 LSB's after shifting right.*/
-+ uint32_t hashDistributionNumOfFqids; /**< must be > 1 and a power of 2. Represents the range
-+ of queues for the key and hash functionality */
-+ uint8_t hashDistributionFqidsShift; /**< selects the FQID bits that will be effected by the hash */
-+ bool symmetricHash; /**< TRUE to generate the same hash for frames with swapped source and
-+ destination fields on all layers; If TRUE, driver will check that for
-+ all layers, if SRC extraction is selected, DST extraction must also be
-+ selected, and vice versa. */
-+} t_FmPcdKgKeyExtractAndHashParams;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining a single FQID mask (extracted OR).
-+*//***************************************************************************/
-+typedef struct t_FmPcdKgExtractedOrParams {
-+ e_FmPcdExtractType type; /**< Extraction type select */
-+ union {
-+ struct { /**< used when type = e_FM_PCD_KG_EXTRACT_BY_HDR */
-+ e_NetHeaderType hdr;
-+ e_FmPcdHdrIndex hdrIndex; /**< Relevant only for MPLS, VLAN and tunneled
-+ IP. Otherwise should be cleared.*/
-+ bool ignoreProtocolValidation;
-+ /**< continue extraction even if protocol is not recognized */
-+ } extractByHdr; /**< Header to extract by */
-+ e_FmPcdExtractFrom src; /**< used when type = e_FM_PCD_KG_EXTRACT_NON_HDR */
-+ };
-+ uint8_t extractionOffset; /**< Offset for extraction (in bytes). */
-+ e_FmPcdKgExtractDfltSelect dfltValue; /**< Select register from which extraction is taken if
-+ field not found */
-+ uint8_t mask; /**< Extraction mask (specified bits are used) */
-+ uint8_t bitOffsetInFqid; /**< 0-31, Selects which bits of the 24 FQID bits to effect using
-+ the extracted byte; Assume byte is placed as the 8 MSB's in
-+ a 32 bit word where the lower bits
-+ are the FQID; i.e if bitOffsetInFqid=1 than its LSB
-+ will effect the FQID MSB, if bitOffsetInFqid=24 than the
-+ extracted byte will effect the 8 LSB's of the FQID,
-+ if bitOffsetInFqid=31 than the byte's MSB will effect
-+ the FQID's LSB; 0 means - no effect on FQID;
-+ Note that one, and only one of
-+ bitOffsetInFqid or bitOffsetInPlcrProfile must be set (i.e,
-+ extracted byte must effect either FQID or Policer profile).*/
-+ uint8_t bitOffsetInPlcrProfile;
-+ /**< 0-15, Selects which bits of the 8 policer profile id bits to
-+ effect using the extracted byte; Assume byte is placed
-+ as the 8 MSB's in a 16 bit word where the lower bits
-+ are the policer profile id; i.e if bitOffsetInPlcrProfile=1
-+ than its LSB will effect the profile MSB, if bitOffsetInFqid=8
-+ than the extracted byte will effect the whole policer profile id,
-+ if bitOffsetInFqid=15 than the byte's MSB will effect
-+ the Policer Profile id's LSB;
-+ 0 means - no effect on policer profile; Note that one, and only one of
-+ bitOffsetInFqid or bitOffsetInPlcrProfile must be set (i.e,
-+ extracted byte must effect either FQID or Policer profile).*/
-+} t_FmPcdKgExtractedOrParams;
-+
-+/**************************************************************************//**
-+ @Description Parameters for configuring a scheme counter
-+*//***************************************************************************/
-+typedef struct t_FmPcdKgSchemeCounter {
-+ bool update; /**< FALSE to keep the current counter state
-+ and continue from that point, TRUE to update/reset
-+ the counter when the scheme is written. */
-+ uint32_t value; /**< If update=TRUE, this value will be written into the
-+ counter. clear this field to reset the counter. */
-+} t_FmPcdKgSchemeCounter;
-+
-+/**************************************************************************//**
-+ @Description Parameters for configuring a policer profile for a KeyGen scheme
-+ (when policer is the next engine after this scheme).
-+*//***************************************************************************/
-+typedef struct t_FmPcdKgPlcrProfile {
-+ bool sharedProfile; /**< TRUE if this profile is shared between ports
-+ (managed by master partition); Must not be TRUE
-+ if profile is after Coarse Classification*/
-+ bool direct; /**< if TRUE, directRelativeProfileId only selects the profile
-+ id, if FALSE fqidOffsetRelativeProfileIdBase is used
-+ together with fqidOffsetShift and numOfProfiles
-+ parameters, to define a range of profiles from
-+ which the KeyGen result will determine the
-+ destination policer profile. */
-+ union {
-+ uint16_t directRelativeProfileId; /**< Used if 'direct' is TRUE, to select policer profile.
-+ should indicate the policer profile offset within the
-+ port's policer profiles or shared window. */
-+ struct {
-+ uint8_t fqidOffsetShift; /**< Shift on the KeyGen create FQID offset (i.e. not the
-+ final FQID - without the FQID base). */
-+ uint8_t fqidOffsetRelativeProfileIdBase;
-+ /**< The base of the FMan Port's relative Storage-Profile ID;
-+ this value will be "OR'ed" with the KeyGen create FQID
-+ offset (i.e. not the final FQID - without the FQID base);
-+ the final result should indicate the Storage-Profile offset
-+ within the FMan Port's relative Storage-Profiles window/
-+ (or the SHARED window depends on 'sharedProfile'). */
-+ uint8_t numOfProfiles; /**< Range of profiles starting at base */
-+ } indirectProfile; /**< Indirect profile parameters */
-+ } profileSelect; /**< Direct/indirect profile selection and parameters */
-+} t_FmPcdKgPlcrProfile;
-+
-+#if (DPAA_VERSION >= 11)
-+/**************************************************************************//**
-+ @Description Parameters for configuring a storage profile for a KeyGen scheme.
-+*//***************************************************************************/
-+typedef struct t_FmPcdKgStorageProfile {
-+ bool direct; /**< If TRUE, directRelativeProfileId only selects the
-+ profile id;
-+ If FALSE, fqidOffsetRelativeProfileIdBase is used
-+ together with fqidOffsetShift and numOfProfiles
-+ parameters to define a range of profiles from which
-+ the KeyGen result will determine the destination
-+ storage profile. */
-+ union {
-+ uint16_t directRelativeProfileId; /**< Used when 'direct' is TRUE, to select a storage profile;
-+ should indicate the storage profile offset within the
-+ port's storage profiles window. */
-+ struct {
-+ uint8_t fqidOffsetShift; /**< Shift on the KeyGen create FQID offset (i.e. not the
-+ final FQID - without the FQID base). */
-+ uint8_t fqidOffsetRelativeProfileIdBase;
-+ /**< The base of the FMan Port's relative Storage-Profile ID;
-+ this value will be "OR'ed" with the KeyGen create FQID
-+ offset (i.e. not the final FQID - without the FQID base);
-+ the final result should indicate the Storage-Profile offset
-+ within the FMan Port's relative Storage-Profiles window. */
-+ uint8_t numOfProfiles; /**< Range of profiles starting at base. */
-+ } indirectProfile; /**< Indirect profile parameters. */
-+ } profileSelect; /**< Direct/indirect profile selection and parameters. */
-+} t_FmPcdKgStorageProfile;
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining CC as the next engine after KeyGen
-+*//***************************************************************************/
-+typedef struct t_FmPcdKgCc {
-+ t_Handle h_CcTree; /**< A handle to a CC Tree */
-+ uint8_t grpId; /**< CC group id within the CC tree */
-+ bool plcrNext; /**< TRUE if after CC, in case of data frame,
-+ policing is required. */
-+ bool bypassPlcrProfileGeneration; /**< TRUE to bypass KeyGen policer profile generation;
-+ selected profile is the one set at port initialization. */
-+ t_FmPcdKgPlcrProfile plcrProfile; /**< Valid only if plcrNext = TRUE and
-+ bypassPlcrProfileGeneration = FALSE */
-+} t_FmPcdKgCc;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining initializing a KeyGen scheme
-+*//***************************************************************************/
-+typedef struct t_FmPcdKgSchemeParams {
-+ bool modify; /**< TRUE to change an existing scheme */
-+ union
-+ {
-+ uint8_t relativeSchemeId; /**< if modify=FALSE:Partition relative scheme id */
-+ t_Handle h_Scheme; /**< if modify=TRUE: a handle of the existing scheme */
-+ } id;
-+ bool alwaysDirect; /**< This scheme is reached only directly, i.e. no need
-+ for match vector; KeyGen will ignore it when matching */
-+ struct { /**< HL Relevant only if alwaysDirect = FALSE */
-+ t_Handle h_NetEnv; /**< A handle to the Network environment as returned
-+ by FM_PCD_NetEnvCharacteristicsSet() */
-+ uint8_t numOfDistinctionUnits; /**< Number of NetEnv units listed in unitIds array */
-+ uint8_t unitIds[FM_PCD_MAX_NUM_OF_DISTINCTION_UNITS];
-+ /**< Indexes as passed to SetNetEnvCharacteristics array*/
-+ } netEnvParams;
-+ bool useHash; /**< use the KeyGen Hash functionality */
-+ t_FmPcdKgKeyExtractAndHashParams keyExtractAndHashParams;
-+ /**< used only if useHash = TRUE */
-+ bool bypassFqidGeneration; /**< Normally - FALSE, TRUE to avoid FQID update in the IC;
-+ In such a case FQID after KeyGen will be the default FQID
-+ defined for the relevant port, or the FQID defined by CC
-+ in cases where CC was the previous engine. */
-+ uint32_t baseFqid; /**< Base FQID; Relevant only if bypassFqidGeneration = FALSE;
-+ If hash is used and an even distribution is expected
-+ according to hashDistributionNumOfFqids, baseFqid must be aligned to
-+ hashDistributionNumOfFqids. */
-+ uint8_t numOfUsedExtractedOrs; /**< Number of FQID masks listed in extractedOrs array */
-+ t_FmPcdKgExtractedOrParams extractedOrs[FM_PCD_KG_NUM_OF_GENERIC_REGS];
-+ /**< FM_PCD_KG_NUM_OF_GENERIC_REGS
-+ registers are shared between qidMasks
-+ functionality and some of the extraction
-+ actions; Normally only some will be used
-+ for qidMask. Driver will return error if
-+ resource is full at initialization time. */
-+
-+#if (DPAA_VERSION >= 11)
-+ bool overrideStorageProfile; /**< TRUE if KeyGen override previously decided storage profile */
-+ t_FmPcdKgStorageProfile storageProfile; /**< Used when overrideStorageProfile TRUE */
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+ e_FmPcdEngine nextEngine; /**< may be BMI, PLCR or CC */
-+ union { /**< depends on nextEngine */
-+ e_FmPcdDoneAction doneAction; /**< Used when next engine is BMI (done) */
-+ t_FmPcdKgPlcrProfile plcrProfile; /**< Used when next engine is PLCR */
-+ t_FmPcdKgCc cc; /**< Used when next engine is CC */
-+ } kgNextEngineParams;
-+ t_FmPcdKgSchemeCounter schemeCounter; /**< A structure of parameters for updating
-+ the scheme counter */
-+} t_FmPcdKgSchemeParams;
-+
-+/**************************************************************************//**
-+ @Collection Definitions for CC statistics
-+*//***************************************************************************/
-+#if (DPAA_VERSION >= 11)
-+#define FM_PCD_CC_STATS_MAX_NUM_OF_FLR 10 /* Maximal supported number of frame length ranges */
-+#define FM_PCD_CC_STATS_FLR_SIZE 2 /* Size in bytes of a frame length range limit */
-+#endif /* (DPAA_VERSION >= 11) */
-+#define FM_PCD_CC_STATS_COUNTER_SIZE 4 /* Size in bytes of a frame length range counter */
-+/* @} */
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining CC as the next engine after a CC node.
-+*//***************************************************************************/
-+typedef struct t_FmPcdCcNextCcParams {
-+ t_Handle h_CcNode; /**< A handle of the next CC node */
-+} t_FmPcdCcNextCcParams;
-+
-+#if (DPAA_VERSION >= 11)
-+/**************************************************************************//**
-+ @Description Parameters for defining Frame replicator as the next engine after a CC node.
-+*//***************************************************************************/
-+typedef struct t_FmPcdCcNextFrParams {
-+ t_Handle h_FrmReplic; /**< A handle of the next frame replicator group */
-+} t_FmPcdCcNextFrParams;
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining Policer as the next engine after a CC node.
-+*//***************************************************************************/
-+typedef struct t_FmPcdCcNextPlcrParams {
-+ bool overrideParams; /**< TRUE if CC override previously decided parameters*/
-+ bool sharedProfile; /**< Relevant only if overrideParams=TRUE:
-+ TRUE if this profile is shared between ports */
-+ uint16_t newRelativeProfileId; /**< Relevant only if overrideParams=TRUE:
-+ (otherwise profile id is taken from KeyGen);
-+ This parameter should indicate the policer
-+ profile offset within the port's
-+ policer profiles or from SHARED window.*/
-+ uint32_t newFqid; /**< Relevant only if overrideParams=TRUE:
-+ FQID for enqueuing the frame;
-+ In earlier chips if policer next engine is KEYGEN,
-+ this parameter can be 0, because the KEYGEN
-+ always decides the enqueue FQID.*/
-+#if (DPAA_VERSION >= 11)
-+ uint8_t newRelativeStorageProfileId;
-+ /**< Indicates the relative storage profile offset within
-+ the port's storage profiles window;
-+ Relevant only if the port was configured with VSP. */
-+#endif /* (DPAA_VERSION >= 11) */
-+} t_FmPcdCcNextPlcrParams;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining enqueue as the next action after a CC node.
-+*//***************************************************************************/
-+typedef struct t_FmPcdCcNextEnqueueParams {
-+ e_FmPcdDoneAction action; /**< Action - when next engine is BMI (done) */
-+ bool overrideFqid; /**< TRUE if CC override previously decided fqid and vspid,
-+ relevant if action = e_FM_PCD_ENQ_FRAME */
-+ uint32_t newFqid; /**< Valid if overrideFqid=TRUE, FQID for enqueuing the frame
-+ (otherwise FQID is taken from KeyGen),
-+ relevant if action = e_FM_PCD_ENQ_FRAME */
-+#if (DPAA_VERSION >= 11)
-+ uint8_t newRelativeStorageProfileId;
-+ /**< Valid if overrideFqid=TRUE, Indicates the relative virtual
-+ storage profile offset within the port's storage profiles
-+ window; Relevant only if the port was configured with VSP. */
-+#endif /* (DPAA_VERSION >= 11) */
-+} t_FmPcdCcNextEnqueueParams;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining KeyGen as the next engine after a CC node.
-+*//***************************************************************************/
-+typedef struct t_FmPcdCcNextKgParams {
-+ bool overrideFqid; /**< TRUE if CC override previously decided fqid and vspid,
-+ Note - this parameters irrelevant for earlier chips */
-+ uint32_t newFqid; /**< Valid if overrideFqid=TRUE, FQID for enqueuing the frame
-+ (otherwise FQID is taken from KeyGen),
-+ Note - this parameters irrelevant for earlier chips */
-+#if (DPAA_VERSION >= 11)
-+ uint8_t newRelativeStorageProfileId;
-+ /**< Valid if overrideFqid=TRUE, Indicates the relative virtual
-+ storage profile offset within the port's storage profiles
-+ window; Relevant only if the port was configured with VSP. */
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+ t_Handle h_DirectScheme; /**< Direct scheme handle to go to. */
-+} t_FmPcdCcNextKgParams;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining the next engine after a CC node.
-+*//***************************************************************************/
-+typedef struct t_FmPcdCcNextEngineParams {
-+ e_FmPcdEngine nextEngine; /**< User has to initialize parameters
-+ according to nextEngine definition */
-+ union {
-+ t_FmPcdCcNextCcParams ccParams; /**< Parameters in case next engine is CC */
-+ t_FmPcdCcNextPlcrParams plcrParams; /**< Parameters in case next engine is PLCR */
-+ t_FmPcdCcNextEnqueueParams enqueueParams; /**< Parameters in case next engine is BMI */
-+ t_FmPcdCcNextKgParams kgParams; /**< Parameters in case next engine is KG */
-+#if (DPAA_VERSION >= 11)
-+ t_FmPcdCcNextFrParams frParams; /**< Parameters in case next engine is FR */
-+#endif /* (DPAA_VERSION >= 11) */
-+ } params; /**< union used for all the next-engine parameters options */
-+
-+ t_Handle h_Manip; /**< Handle to Manipulation object.
-+ Relevant if next engine is of type result
-+ (e_FM_PCD_PLCR, e_FM_PCD_KG, e_FM_PCD_DONE) */
-+
-+ bool statisticsEn; /**< If TRUE, statistics counters are incremented
-+ for each frame passing through this
-+ Coarse Classification entry. */
-+} t_FmPcdCcNextEngineParams;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining a single CC key
-+*//***************************************************************************/
-+typedef struct t_FmPcdCcKeyParams {
-+ uint8_t *p_Key; /**< Relevant only if 'action' = e_FM_PCD_ACTION_EXACT_MATCH;
-+ pointer to the key of the size defined in keySize */
-+ uint8_t *p_Mask; /**< Relevant only if 'action' = e_FM_PCD_ACTION_EXACT_MATCH;
-+ pointer to the Mask per key of the size defined
-+ in keySize. p_Key and p_Mask (if defined) has to be
-+ of the same size defined in the keySize;
-+ NOTE that if this value is equal for all entries whithin
-+ this table, the driver will automatically use global-mask
-+ (i.e. one common mask for all entries) instead of private
-+ one; that is done in order to spare some memory and for
-+ better performance. */
-+ t_FmPcdCcNextEngineParams ccNextEngineParams;
-+ /**< parameters for the next for the defined Key in
-+ the p_Key */
-+} t_FmPcdCcKeyParams;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining CC keys parameters
-+ The driver supports two methods for CC node allocation: dynamic and static.
-+ Static mode was created in order to prevent runtime alloc/free
-+ of FMan memory (MURAM), which may cause fragmentation; in this mode,
-+ the driver automatically allocates the memory according to
-+ 'maxNumOfKeys' parameter. The driver calculates the maximal memory
-+ size that may be used for this CC-Node taking into consideration
-+ 'maskSupport' and 'statisticsMode' parameters.
-+ When 'action' = e_FM_PCD_ACTION_INDEXED_LOOKUP in the extraction
-+ parameters of this node, 'maxNumOfKeys' must be equal to 'numOfKeys'.
-+ In dynamic mode, 'maxNumOfKeys' must be zero. At initialization,
-+ all required structures are allocated according to 'numOfKeys'
-+ parameter. During runtime modification, these structures are
-+ re-allocated according to the updated number of keys.
-+
-+ Please note that 'action' and 'icIndxMask' mentioned in the
-+ specific parameter explanations are passed in the extraction
-+ parameters of the node (fields of extractCcParams.extractNonHdr).
-+*//***************************************************************************/
-+typedef struct t_KeysParams {
-+ uint16_t maxNumOfKeys; /**< Maximum number of keys that will (ever) be used in this CC-Node;
-+ A value of zero may be used for dynamic memory allocation. */
-+ bool maskSupport; /**< This parameter is relevant only if a node is initialized with
-+ 'action' = e_FM_PCD_ACTION_EXACT_MATCH and maxNumOfKeys > 0;
-+ Should be TRUE to reserve table memory for key masks, even if
-+ initial keys do not contain masks, or if the node was initialized
-+ as 'empty' (without keys); this will allow user to add keys with
-+ masks at runtime.
-+ NOTE that if user want to use only global-masks (i.e. one common mask
-+ for all the entries within this table, this parameter should set to 'FALSE'. */
-+ e_FmPcdCcStatsMode statisticsMode; /**< Determines the supported statistics mode for all node's keys.
-+ To enable statistics gathering, statistics should be enabled per
-+ every key, using 'statisticsEn' in next engine parameters structure
-+ of that key;
-+ If 'maxNumOfKeys' is set, all required structures will be
-+ preallocated for all keys. */
-+#if (DPAA_VERSION >= 11)
-+ uint16_t frameLengthRanges[FM_PCD_CC_STATS_MAX_NUM_OF_FLR];
-+ /**< Relevant only for 'RMON' statistics mode
-+ (this feature is supported only on B4860 device);
-+ Holds a list of programmable thresholds - for each received frame,
-+ its length in bytes is examined against these range thresholds and
-+ the appropriate counter is incremented by 1 - for example, to belong
-+ to range i, the following should hold:
-+ range i-1 threshold < frame length <= range i threshold
-+ Each range threshold must be larger then its preceding range
-+ threshold, and last range threshold must be 0xFFFF. */
-+#endif /* (DPAA_VERSION >= 11) */
-+ uint16_t numOfKeys; /**< Number of initial keys;
-+ Note that in case of 'action' = e_FM_PCD_ACTION_INDEXED_LOOKUP,
-+ this field should be power-of-2 of the number of bits that are
-+ set in 'icIndxMask'. */
-+ uint8_t keySize; /**< Size of key - for extraction of type FULL_FIELD, 'keySize' has
-+ to be the standard size of the selected key; For other extraction
-+ types, 'keySize' has to be as size of extraction; When 'action' =
-+ e_FM_PCD_ACTION_INDEXED_LOOKUP, 'keySize' must be 2. */
-+ t_FmPcdCcKeyParams keyParams[FM_PCD_MAX_NUM_OF_KEYS];
-+ /**< An array with 'numOfKeys' entries, each entry specifies the
-+ corresponding key parameters;
-+ When 'action' = e_FM_PCD_ACTION_EXACT_MATCH, this value must not
-+ exceed 255 (FM_PCD_MAX_NUM_OF_KEYS-1) as the last entry is saved
-+ for the 'miss' entry. */
-+ t_FmPcdCcNextEngineParams ccNextEngineParamsForMiss;
-+ /**< Parameters for defining the next engine when a key is not matched;
-+ Not relevant if action = e_FM_PCD_ACTION_INDEXED_LOOKUP. */
-+} t_KeysParams;
-+
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining a CC node
-+*//***************************************************************************/
-+typedef struct t_FmPcdCcNodeParams {
-+ t_FmPcdExtractEntry extractCcParams; /**< Extraction parameters */
-+ t_KeysParams keysParams; /**< Keys definition matching the selected extraction */
-+} t_FmPcdCcNodeParams;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining a hash table
-+*//***************************************************************************/
-+typedef struct t_FmPcdHashTableParams {
-+ uint16_t maxNumOfKeys; /**< Maximum Number Of Keys that will (ever) be used in this Hash-table */
-+ e_FmPcdCcStatsMode statisticsMode; /**< If not e_FM_PCD_CC_STATS_MODE_NONE, the required structures for the
-+ requested statistics mode will be allocated according to maxNumOfKeys. */
-+ uint8_t kgHashShift; /**< KG-Hash-shift as it was configured in the KG-scheme
-+ that leads to this hash-table. */
-+ uint16_t hashResMask; /**< Mask that will be used on the hash-result;
-+ The number-of-sets for this hash will be calculated
-+ as (2^(number of bits set in 'hashResMask'));
-+ The 4 lower bits must be cleared. */
-+ uint8_t hashShift; /**< Byte offset from the beginning of the KeyGen hash result to the
-+ 2-bytes to be used as hash index. */
-+ uint8_t matchKeySize; /**< Size of the exact match keys held by the hash buckets */
-+
-+ t_FmPcdCcNextEngineParams ccNextEngineParamsForMiss; /**< Parameters for defining the next engine when a key is not matched */
-+
-+} t_FmPcdHashTableParams;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining a CC tree group.
-+
-+ This structure defines a CC group in terms of NetEnv units
-+ and the action to be taken in each case. The unitIds list must
-+ be given in order from low to high indices.
-+
-+ t_FmPcdCcNextEngineParams is a list of 2^numOfDistinctionUnits
-+ structures where each defines the next action to be taken for
-+ each units combination. for example:
-+ numOfDistinctionUnits = 2
-+ unitIds = {1,3}
-+ p_NextEnginePerEntriesInGrp[0] = t_FmPcdCcNextEngineParams for the case that
-+ unit 1 - not found; unit 3 - not found;
-+ p_NextEnginePerEntriesInGrp[1] = t_FmPcdCcNextEngineParams for the case that
-+ unit 1 - not found; unit 3 - found;
-+ p_NextEnginePerEntriesInGrp[2] = t_FmPcdCcNextEngineParams for the case that
-+ unit 1 - found; unit 3 - not found;
-+ p_NextEnginePerEntriesInGrp[3] = t_FmPcdCcNextEngineParams for the case that
-+ unit 1 - found; unit 3 - found;
-+*//***************************************************************************/
-+typedef struct t_FmPcdCcGrpParams {
-+ uint8_t numOfDistinctionUnits; /**< Up to 4 */
-+ uint8_t unitIds[FM_PCD_MAX_NUM_OF_CC_UNITS];
-+ /**< Indices of the units as defined in
-+ FM_PCD_NetEnvCharacteristicsSet() */
-+ t_FmPcdCcNextEngineParams nextEnginePerEntriesInGrp[FM_PCD_MAX_NUM_OF_CC_ENTRIES_IN_GRP];
-+ /**< Maximum entries per group is 16 */
-+} t_FmPcdCcGrpParams;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining CC tree groups
-+*//***************************************************************************/
-+typedef struct t_FmPcdCcTreeParams {
-+ t_Handle h_NetEnv; /**< A handle to the Network environment as returned
-+ by FM_PCD_NetEnvCharacteristicsSet() */
-+ uint8_t numOfGrps; /**< Number of CC groups within the CC tree */
-+ t_FmPcdCcGrpParams ccGrpParams[FM_PCD_MAX_NUM_OF_CC_GROUPS];
-+ /**< Parameters for each group. */
-+} t_FmPcdCcTreeParams;
-+
-+
-+/**************************************************************************//**
-+ @Description CC key statistics structure
-+*//***************************************************************************/
-+typedef struct t_FmPcdCcKeyStatistics {
-+ uint32_t byteCount; /**< This counter reflects byte count of frames that
-+ were matched by this key. */
-+ uint32_t frameCount; /**< This counter reflects count of frames that
-+ were matched by this key. */
-+#if (DPAA_VERSION >= 11)
-+ uint32_t frameLengthRangeCount[FM_PCD_CC_STATS_MAX_NUM_OF_FLR];
-+ /**< These counters reflect how many frames matched
-+ this key in 'RMON' statistics mode:
-+ Each counter holds the number of frames of a
-+ specific frames length range, according to the
-+ ranges provided at initialization. */
-+#endif /* (DPAA_VERSION >= 11) */
-+} t_FmPcdCcKeyStatistics;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining policer byte rate
-+*//***************************************************************************/
-+typedef struct t_FmPcdPlcrByteRateModeParams {
-+ e_FmPcdPlcrFrameLengthSelect frameLengthSelection; /**< Frame length selection */
-+ e_FmPcdPlcrRollBackFrameSelect rollBackFrameSelection; /**< relevant option only e_FM_PCD_PLCR_L2_FRM_LEN,
-+ e_FM_PCD_PLCR_FULL_FRM_LEN */
-+} t_FmPcdPlcrByteRateModeParams;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining the policer profile (based on
-+ RFC-2698 or RFC-4115 attributes).
-+*//***************************************************************************/
-+typedef struct t_FmPcdPlcrNonPassthroughAlgParams {
-+ e_FmPcdPlcrRateMode rateMode; /**< Byte mode or Packet mode */
-+ t_FmPcdPlcrByteRateModeParams byteModeParams; /**< Valid for Byte NULL for Packet */
-+ uint32_t committedInfoRate; /**< KBits/Second or Packets/Second */
-+ uint32_t committedBurstSize; /**< Bytes/Packets */
-+ uint32_t peakOrExcessInfoRate; /**< KBits/Second or Packets/Second */
-+ uint32_t peakOrExcessBurstSize; /**< Bytes/Packets */
-+} t_FmPcdPlcrNonPassthroughAlgParams;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining the next engine after policer
-+*//***************************************************************************/
-+typedef union u_FmPcdPlcrNextEngineParams {
-+ e_FmPcdDoneAction action; /**< Action - when next engine is BMI (done) */
-+ t_Handle h_Profile; /**< Policer profile handle - used when next engine
-+ is Policer, must be a SHARED profile */
-+ t_Handle h_DirectScheme; /**< Direct scheme select - when next engine is KeyGen */
-+} u_FmPcdPlcrNextEngineParams;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining the policer profile entry
-+*//***************************************************************************/
-+typedef struct t_FmPcdPlcrProfileParams {
-+ bool modify; /**< TRUE to change an existing profile */
-+ union {
-+ struct {
-+ e_FmPcdProfileTypeSelection profileType; /**< Type of policer profile */
-+ t_Handle h_FmPort; /**< Relevant for per-port profiles only */
-+ uint16_t relativeProfileId; /**< Profile id - relative to shared group or to port */
-+ } newParams; /**< use it when modify = FALSE */
-+ t_Handle h_Profile; /**< A handle to a profile - use it when modify=TRUE */
-+ } id;
-+ e_FmPcdPlcrAlgorithmSelection algSelection; /**< Profile Algorithm PASS_THROUGH, RFC_2698, RFC_4115 */
-+ e_FmPcdPlcrColorMode colorMode; /**< COLOR_BLIND, COLOR_AWARE */
-+
-+ union {
-+ e_FmPcdPlcrColor dfltColor; /**< For Color-Blind Pass-Through mode; the policer will re-color
-+ any incoming packet with the default value. */
-+ e_FmPcdPlcrColor override; /**< For Color-Aware modes; the profile response to a
-+ pre-color value of 2'b11. */
-+ } color;
-+
-+ t_FmPcdPlcrNonPassthroughAlgParams nonPassthroughAlgParams; /**< RFC2698 or RFC4115 parameters */
-+
-+ e_FmPcdEngine nextEngineOnGreen; /**< Next engine for green-colored frames */
-+ u_FmPcdPlcrNextEngineParams paramsOnGreen; /**< Next engine parameters for green-colored frames */
-+
-+ e_FmPcdEngine nextEngineOnYellow; /**< Next engine for yellow-colored frames */
-+ u_FmPcdPlcrNextEngineParams paramsOnYellow; /**< Next engine parameters for yellow-colored frames */
-+
-+ e_FmPcdEngine nextEngineOnRed; /**< Next engine for red-colored frames */
-+ u_FmPcdPlcrNextEngineParams paramsOnRed; /**< Next engine parameters for red-colored frames */
-+
-+ bool trapProfileOnFlowA; /**< Obsolete - do not use */
-+ bool trapProfileOnFlowB; /**< Obsolete - do not use */
-+ bool trapProfileOnFlowC; /**< Obsolete - do not use */
-+} t_FmPcdPlcrProfileParams;
-+
-+/**************************************************************************//**
-+ @Description Parameters for selecting a location for requested manipulation
-+*//***************************************************************************/
-+typedef struct t_FmManipHdrInfo {
-+ e_NetHeaderType hdr; /**< Header selection */
-+ e_FmPcdHdrIndex hdrIndex; /**< Relevant only for MPLS, VLAN and tunneled IP. Otherwise should be cleared. */
-+ bool byField; /**< TRUE if the location of manipulation is according to some field in the specific header*/
-+ t_FmPcdFields fullField; /**< Relevant only when byField = TRUE: Extract field */
-+} t_FmManipHdrInfo;
-+
-+#if ((DPAA_VERSION == 10) && defined(FM_CAPWAP_SUPPORT))
-+/**************************************************************************//**
-+ @Description Parameters for defining an insertion manipulation
-+ of type e_FM_PCD_MANIP_INSRT_TO_START_OF_FRAME_TEMPLATE
-+*//***************************************************************************/
-+typedef struct t_FmPcdManipHdrInsrtByTemplateParams {
-+ uint8_t size; /**< Size of insert template to the start of the frame. */
-+ uint8_t hdrTemplate[FM_PCD_MAX_MANIP_INSRT_TEMPLATE_SIZE];
-+ /**< Array of the insertion template. */
-+
-+ bool modifyOuterIp; /**< TRUE if user want to modify some fields in outer IP. */
-+ struct {
-+ uint16_t ipOuterOffset; /**< Offset of outer IP in the insert template, relevant if modifyOuterIp = TRUE.*/
-+ uint16_t dscpEcn; /**< value of dscpEcn in IP outer, relevant if modifyOuterIp = TRUE.
-+ in IPV4 dscpEcn only byte - it has to be adjusted to the right*/
-+ bool udpPresent; /**< TRUE if UDP is present in the insert template, relevant if modifyOuterIp = TRUE.*/
-+ uint8_t udpOffset; /**< Offset in the insert template of UDP, relevant if modifyOuterIp = TRUE and udpPresent=TRUE.*/
-+ uint8_t ipIdentGenId; /**< Used by FMan-CTRL to calculate IP-identification field,relevant if modifyOuterIp = TRUE.*/
-+ bool recalculateLength; /**< TRUE if recalculate length has to be performed due to the engines in the path which can change the frame later, relevant if modifyOuterIp = TRUE.*/
-+ struct {
-+ uint8_t blockSize; /**< The CAAM block-size; Used by FMan-CTRL to calculate the IP Total Length field.*/
-+ uint8_t extraBytesAddedAlignedToBlockSize; /**< Used by FMan-CTRL to calculate the IP Total Length field and UDP length*/
-+ uint8_t extraBytesAddedNotAlignedToBlockSize;/**< Used by FMan-CTRL to calculate the IP Total Length field and UDP length.*/
-+ } recalculateLengthParams; /**< Recalculate length parameters - relevant if modifyOuterIp = TRUE and recalculateLength = TRUE */
-+ } modifyOuterIpParams; /**< Outer IP modification parameters - ignored if modifyOuterIp is FALSE */
-+
-+ bool modifyOuterVlan; /**< TRUE if user wants to modify VPri field in the outer VLAN header*/
-+ struct {
-+ uint8_t vpri; /**< Value of VPri, relevant if modifyOuterVlan = TRUE
-+ VPri only 3 bits, it has to be adjusted to the right*/
-+ } modifyOuterVlanParams;
-+} t_FmPcdManipHdrInsrtByTemplateParams;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining CAPWAP fragmentation
-+*//***************************************************************************/
-+typedef struct t_CapwapFragmentationParams {
-+ uint16_t sizeForFragmentation; /**< if length of the frame is greater than this value, CAPWAP fragmentation will be executed.*/
-+ bool headerOptionsCompr; /**< TRUE - first fragment include the CAPWAP header options field,
-+ and all other fragments exclude the CAPWAP options field,
-+ FALSE - all fragments include CAPWAP header options field. */
-+} t_CapwapFragmentationParams;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining CAPWAP reassembly
-+*//***************************************************************************/
-+typedef struct t_CapwapReassemblyParams {
-+ uint16_t maxNumFramesInProcess; /**< Number of frames which can be reassembled concurrently; must be power of 2.
-+ In case numOfFramesPerHashEntry == e_FM_PCD_MANIP_FOUR_WAYS_HASH,
-+ maxNumFramesInProcess has to be in the range of 4 - 512,
-+ In case numOfFramesPerHashEntry == e_FM_PCD_MANIP_EIGHT_WAYS_HASH,
-+ maxNumFramesInProcess has to be in the range of 8 - 2048 */
-+ bool haltOnDuplicationFrag; /**< If TRUE, reassembly process will be halted due to duplicated fragment,
-+ and all processed fragments will be enqueued with error indication;
-+ If FALSE, only duplicated fragments will be enqueued with error indication. */
-+
-+ e_FmPcdManipReassemTimeOutMode timeOutMode; /**< Expiration delay initialized by the reassembly process */
-+ uint32_t fqidForTimeOutFrames; /**< FQID in which time out frames will enqueue during Time Out Process */
-+ uint32_t timeoutRoutineRequestTime;
-+ /**< Represents the time interval in microseconds between consecutive
-+ timeout routine requests It has to be power of 2. */
-+ uint32_t timeoutThresholdForReassmProcess;
-+ /**< Time interval (microseconds) for marking frames in process as too old;
-+ Frames in process are those for which at least one fragment was received
-+ but not all fragments. */
-+
-+ e_FmPcdManipReassemWaysNumber numOfFramesPerHashEntry;/**< Number of frames per hash entry (needed for the reassembly process) */
-+} t_CapwapReassemblyParams;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining fragmentation/reassembly manipulation
-+*//***************************************************************************/
-+typedef struct t_FmPcdManipFragOrReasmParams {
-+ bool frag; /**< TRUE if using the structure for fragmentation,
-+ otherwise this structure is used for reassembly */
-+ uint8_t sgBpid; /**< Scatter/Gather buffer pool id;
-+ Same LIODN number is used for these buffers as for
-+ the received frames buffers, so buffers of this pool
-+ need to be allocated in the same memory area as the
-+ received buffers. If the received buffers arrive
-+ from different sources, the Scatter/Gather BP id
-+ should be mutual to all these sources. */
-+ e_NetHeaderType hdr; /**< Header selection */
-+ union {
-+ t_CapwapFragmentationParams capwapFragParams; /**< Structure for CAPWAP fragmentation,
-+ relevant if 'frag' = TRUE, 'hdr' = HEADER_TYPE_CAPWAP */
-+ t_CapwapReassemblyParams capwapReasmParams; /**< Structure for CAPWAP reassembly,
-+ relevant if 'frag' = FALSE, 'hdr' = HEADER_TYPE_CAPWAP */
-+ } u;
-+} t_FmPcdManipFragOrReasmParams;
-+#endif /* ((DPAA_VERSION == 10) && defined(FM_CAPWAP_SUPPORT)) */
-+
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining header removal by header type
-+*//***************************************************************************/
-+typedef struct t_FmPcdManipHdrRmvByHdrParams {
-+ e_FmPcdManipHdrRmvByHdrType type; /**< Selection of header removal location */
-+ union {
-+#if ((DPAA_VERSION == 10) && defined(FM_CAPWAP_SUPPORT))
-+ struct {
-+ bool include; /**< If FALSE, remove until the specified header (not including the header);
-+ If TRUE, remove also the specified header. */
-+ t_FmManipHdrInfo hdrInfo;
-+ } fromStartByHdr; /**< Relevant when type = e_FM_PCD_MANIP_RMV_BY_HDR_FROM_START */
-+#endif /* (DPAA_VERSION >= 11) || ... */
-+#if (DPAA_VERSION >= 11)
-+ t_FmManipHdrInfo hdrInfo; /**< Relevant when type = e_FM_PCD_MANIP_RMV_BY_HDR_FROM_START */
-+#endif /* (DPAA_VERSION >= 11) */
-+ e_FmPcdManipHdrRmvSpecificL2 specificL2; /**< Relevant when type = e_FM_PCD_MANIP_BY_HDR_SPECIFIC_L2;
-+ Defines which L2 headers to remove. */
-+ } u;
-+} t_FmPcdManipHdrRmvByHdrParams;
-+
-+/**************************************************************************//**
-+ @Description Parameters for configuring IP fragmentation manipulation
-+
-+ Restrictions:
-+ - IP Fragmentation output fragments must not be forwarded to application directly.
-+ - Maximum number of fragments per frame is 16.
-+ - Fragmentation of IP fragments is not supported.
-+ - IPv4 packets containing header Option fields are fragmented by copying all option
-+ fields to each fragment, regardless of the copy bit value.
-+ - Transmit confirmation is not supported.
-+ - Fragmentation after SEC can't handle S/G frames.
-+ - Fragmentation nodes must be set as the last PCD action (i.e. the
-+ corresponding CC node key must have next engine set to e_FM_PCD_DONE).
-+ - Only BMan buffers shall be used for frames to be fragmented.
-+ - IPF does not support VSP. Therefore, on the same port where we have IPF
-+ we cannot support VSP.
-+ - NOTE: The following comment is relevant only for FMAN v3 devices: IPF
-+ does not support VSP. Therefore, on the same port where we have IPF we
-+ cannot support VSP.
-+*//***************************************************************************/
-+typedef struct t_FmPcdManipFragIpParams {
-+ uint16_t sizeForFragmentation; /**< If length of the frame is greater than this value,
-+ IP fragmentation will be executed.*/
-+#if (DPAA_VERSION == 10)
-+ uint8_t scratchBpid; /**< Absolute buffer pool id according to BM configuration.*/
-+#endif /* (DPAA_VERSION == 10) */
-+ bool sgBpidEn; /**< Enable a dedicated buffer pool id for the Scatter/Gather buffer allocation;
-+ If disabled, the Scatter/Gather buffer will be allocated from the same pool as the
-+ received frame's buffer. */
-+ uint8_t sgBpid; /**< Scatter/Gather buffer pool id;
-+ This parameters is relevant when 'sgBpidEn=TRUE';
-+ Same LIODN number is used for these buffers as for the received frames buffers, so buffers
-+ of this pool need to be allocated in the same memory area as the received buffers.
-+ If the received buffers arrive from different sources, the Scatter/Gather BP id should be
-+ mutual to all these sources. */
-+ e_FmPcdManipDontFragAction dontFragAction; /**< Don't Fragment Action - If an IP packet is larger
-+ than MTU and its DF bit is set, then this field will
-+ determine the action to be taken.*/
-+} t_FmPcdManipFragIpParams;
-+
-+/**************************************************************************//**
-+ @Description Parameters for configuring IP reassembly manipulation.
-+
-+ This is a common structure for both IPv4 and IPv6 reassembly
-+ manipulation. For reassembly of both IPv4 and IPv6, make sure to
-+ set the 'hdr' field in t_FmPcdManipReassemParams to HEADER_TYPE_IPv6.
-+
-+ Restrictions:
-+ - Application must define at least one scheme to catch the reassembled frames.
-+ - Maximum number of fragments per frame is 16.
-+ - Reassembly of IPv4 fragments containing Option fields is supported.
-+
-+*//***************************************************************************/
-+typedef struct t_FmPcdManipReassemIpParams {
-+ uint8_t relativeSchemeId[2]; /**< Partition relative scheme id:
-+ relativeSchemeId[0] - Relative scheme ID for IPV4 Reassembly manipulation;
-+ relativeSchemeId[1] - Relative scheme ID for IPV6 Reassembly manipulation;
-+ NOTE: The following comment is relevant only for FMAN v2 devices:
-+ Relative scheme ID for IPv4/IPv6 Reassembly manipulation must be smaller than
-+ the user schemes id to ensure that the reassembly schemes will be first match;
-+ Rest schemes, if defined, should have higher relative scheme ID. */
-+#if (DPAA_VERSION >= 11)
-+ uint32_t nonConsistentSpFqid; /**< In case that other fragments of the frame corresponds to different storage
-+ profile than the opening fragment (Non-Consistent-SP state)
-+ then one of two possible scenarios occurs:
-+ if 'nonConsistentSpFqid != 0', the reassembled frame will be enqueued to
-+ this fqid, otherwise a 'Non Consistent SP' bit will be set in the FD[status].*/
-+#else
-+ uint8_t sgBpid; /**< Buffer pool id for the S/G frame created by the reassembly process */
-+#endif /* (DPAA_VERSION >= 11) */
-+ uint8_t dataMemId; /**< Memory partition ID for the IPR's external tables structure */
-+ uint16_t dataLiodnOffset; /**< LIODN offset for access the IPR's external tables structure. */
-+ uint16_t minFragSize[2]; /**< Minimum fragment size:
-+ minFragSize[0] - for ipv4, minFragSize[1] - for ipv6 */
-+ e_FmPcdManipReassemWaysNumber numOfFramesPerHashEntry[2];
-+ /**< Number of frames per hash entry needed for reassembly process:
-+ numOfFramesPerHashEntry[0] - for ipv4 (max value is e_FM_PCD_MANIP_EIGHT_WAYS_HASH);
-+ numOfFramesPerHashEntry[1] - for ipv6 (max value is e_FM_PCD_MANIP_SIX_WAYS_HASH). */
-+ uint16_t maxNumFramesInProcess; /**< Number of frames which can be processed by Reassembly in the same time;
-+ Must be power of 2;
-+ In the case numOfFramesPerHashEntry == e_FM_PCD_MANIP_FOUR_WAYS_HASH,
-+ maxNumFramesInProcess has to be in the range of 4 - 512;
-+ In the case numOfFramesPerHashEntry == e_FM_PCD_MANIP_EIGHT_WAYS_HASH,
-+ maxNumFramesInProcess has to be in the range of 8 - 2048. */
-+ e_FmPcdManipReassemTimeOutMode timeOutMode; /**< Expiration delay initialized by Reassembly process */
-+ uint32_t fqidForTimeOutFrames; /**< FQID in which time out frames will enqueue during Time Out Process;
-+ Recommended value for this field is 0; in this way timed-out frames will be discarded */
-+ uint32_t timeoutThresholdForReassmProcess;
-+ /**< Represents the time interval in microseconds which defines
-+ if opened frame (at least one fragment was processed but not all the fragments)is found as too old*/
-+} t_FmPcdManipReassemIpParams;
-+
-+/**************************************************************************//**
-+ @Description structure for defining IPSEC manipulation
-+*//***************************************************************************/
-+typedef struct t_FmPcdManipSpecialOffloadIPSecParams {
-+ bool decryption; /**< TRUE if being used in decryption direction;
-+ FALSE if being used in encryption direction. */
-+ bool ecnCopy; /**< TRUE to copy the ECN bits from inner/outer to outer/inner
-+ (direction depends on the 'decryption' field). */
-+ bool dscpCopy; /**< TRUE to copy the DSCP bits from inner/outer to outer/inner
-+ (direction depends on the 'decryption' field). */
-+ bool variableIpHdrLen; /**< TRUE for supporting variable IP header length in decryption. */
-+ bool variableIpVersion; /**< TRUE for supporting both IP version on the same SA in encryption */
-+ uint8_t outerIPHdrLen; /**< if 'variableIpVersion == TRUE' then this field must be set to non-zero value;
-+ It is specifies the length of the outer IP header that was configured in the
-+ corresponding SA. */
-+ uint16_t arwSize; /**< if <> '0' then will perform ARW check for this SA;
-+ The value must be a multiplication of 16 */
-+ uintptr_t arwAddr; /**< if arwSize <> '0' then this field must be set to non-zero value;
-+ MUST be allocated from FMAN's MURAM that the post-sec op-port belongs to;
-+ Must be 4B aligned. Required MURAM size is 'NEXT_POWER_OF_2(arwSize+32))/8+4' Bytes */
-+} t_FmPcdManipSpecialOffloadIPSecParams;
-+
-+#if (DPAA_VERSION >= 11)
-+/**************************************************************************//**
-+ @Description Parameters for configuring CAPWAP fragmentation manipulation
-+
-+ Restrictions:
-+ - Maximum number of fragments per frame is 16.
-+ - Transmit confirmation is not supported.
-+ - Fragmentation nodes must be set as the last PCD action (i.e. the
-+ corresponding CC node key must have next engine set to e_FM_PCD_DONE).
-+ - Only BMan buffers shall be used for frames to be fragmented.
-+ - NOTE: The following comment is relevant only for FMAN v3 devices: IPF
-+ does not support VSP. Therefore, on the same port where we have IPF we
-+ cannot support VSP.
-+*//***************************************************************************/
-+typedef struct t_FmPcdManipFragCapwapParams {
-+ uint16_t sizeForFragmentation; /**< If length of the frame is greater than this value,
-+ CAPWAP fragmentation will be executed.*/
-+ bool sgBpidEn; /**< Enable a dedicated buffer pool id for the Scatter/Gather buffer allocation;
-+ If disabled, the Scatter/Gather buffer will be allocated from the same pool as the
-+ received frame's buffer. */
-+ uint8_t sgBpid; /**< Scatter/Gather buffer pool id;
-+ This parameters is relevant when 'sgBpidEn=TRUE';
-+ Same LIODN number is used for these buffers as for the received frames buffers, so buffers
-+ of this pool need to be allocated in the same memory area as the received buffers.
-+ If the received buffers arrive from different sources, the Scatter/Gather BP id should be
-+ mutual to all these sources. */
-+ bool compressModeEn; /**< CAPWAP Header Options Compress Enable mode;
-+ When this mode is enabled then only the first fragment include the CAPWAP header options
-+ field (if user provides it in the input frame) and all other fragments exclude the CAPWAP
-+ options field (CAPWAP header is updated accordingly).*/
-+} t_FmPcdManipFragCapwapParams;
-+
-+/**************************************************************************//**
-+ @Description Parameters for configuring CAPWAP reassembly manipulation.
-+
-+ Restrictions:
-+ - Application must define one scheme to catch the reassembled frames.
-+ - Maximum number of fragments per frame is 16.
-+
-+*//***************************************************************************/
-+typedef struct t_FmPcdManipReassemCapwapParams {
-+ uint8_t relativeSchemeId; /**< Partition relative scheme id;
-+ NOTE: this id must be smaller than the user schemes id to ensure that the reassembly scheme will be first match;
-+ Rest schemes, if defined, should have higher relative scheme ID. */
-+ uint8_t dataMemId; /**< Memory partition ID for the IPR's external tables structure */
-+ uint16_t dataLiodnOffset; /**< LIODN offset for access the IPR's external tables structure. */
-+ uint16_t maxReassembledFrameLength;/**< The maximum CAPWAP reassembled frame length in bytes;
-+ If maxReassembledFrameLength == 0, any successful reassembled frame length is
-+ considered as a valid length;
-+ if maxReassembledFrameLength > 0, a successful reassembled frame which its length
-+ exceeds this value is considered as an error frame (FD status[CRE] bit is set). */
-+ e_FmPcdManipReassemWaysNumber numOfFramesPerHashEntry;
-+ /**< Number of frames per hash entry needed for reassembly process */
-+ uint16_t maxNumFramesInProcess; /**< Number of frames which can be processed by reassembly in the same time;
-+ Must be power of 2;
-+ In the case numOfFramesPerHashEntry == e_FM_PCD_MANIP_FOUR_WAYS_HASH,
-+ maxNumFramesInProcess has to be in the range of 4 - 512;
-+ In the case numOfFramesPerHashEntry == e_FM_PCD_MANIP_EIGHT_WAYS_HASH,
-+ maxNumFramesInProcess has to be in the range of 8 - 2048. */
-+ e_FmPcdManipReassemTimeOutMode timeOutMode; /**< Expiration delay initialized by Reassembly process */
-+ uint32_t fqidForTimeOutFrames; /**< FQID in which time out frames will enqueue during Time Out Process;
-+ Recommended value for this field is 0; in this way timed-out frames will be discarded */
-+ uint32_t timeoutThresholdForReassmProcess;
-+ /**< Represents the time interval in microseconds which defines
-+ if opened frame (at least one fragment was processed but not all the fragments)is found as too old*/
-+} t_FmPcdManipReassemCapwapParams;
-+
-+/**************************************************************************//**
-+ @Description structure for defining CAPWAP manipulation
-+*//***************************************************************************/
-+typedef struct t_FmPcdManipSpecialOffloadCapwapParams {
-+ bool dtls; /**< TRUE if continue to SEC DTLS encryption */
-+ e_FmPcdManipHdrQosSrc qosSrc; /**< TODO */
-+} t_FmPcdManipSpecialOffloadCapwapParams;
-+
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining special offload manipulation
-+*//***************************************************************************/
-+typedef struct t_FmPcdManipSpecialOffloadParams {
-+ e_FmPcdManipSpecialOffloadType type; /**< Type of special offload manipulation */
-+ union
-+ {
-+ t_FmPcdManipSpecialOffloadIPSecParams ipsec; /**< Parameters for IPSec; Relevant when
-+ type = e_FM_PCD_MANIP_SPECIAL_OFFLOAD_IPSEC */
-+#if (DPAA_VERSION >= 11)
-+ t_FmPcdManipSpecialOffloadCapwapParams capwap; /**< Parameters for CAPWAP; Relevant when
-+ type = e_FM_PCD_MANIP_SPECIAL_OFFLOAD_CAPWAP */
-+#endif /* (DPAA_VERSION >= 11) */
-+ } u;
-+} t_FmPcdManipSpecialOffloadParams;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining insertion manipulation
-+*//***************************************************************************/
-+typedef struct t_FmPcdManipHdrInsrt {
-+ uint8_t size; /**< size of inserted section */
-+ uint8_t *p_Data; /**< data to be inserted */
-+} t_FmPcdManipHdrInsrt;
-+
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining generic removal manipulation
-+*//***************************************************************************/
-+typedef struct t_FmPcdManipHdrRmvGenericParams {
-+ uint8_t offset; /**< Offset from beginning of header to the start
-+ location of the removal */
-+ uint8_t size; /**< Size of removed section */
-+} t_FmPcdManipHdrRmvGenericParams;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining generic insertion manipulation
-+*//***************************************************************************/
-+typedef struct t_FmPcdManipHdrInsrtGenericParams {
-+ uint8_t offset; /**< Offset from beginning of header to the start
-+ location of the insertion */
-+ uint8_t size; /**< Size of inserted section */
-+ bool replace; /**< TRUE to override (replace) existing data at
-+ 'offset', FALSE to insert */
-+ uint8_t *p_Data; /**< Pointer to data to be inserted */
-+} t_FmPcdManipHdrInsrtGenericParams;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining header manipulation VLAN DSCP To Vpri translation
-+*//***************************************************************************/
-+typedef struct t_FmPcdManipHdrFieldUpdateVlanDscpToVpri {
-+ uint8_t dscpToVpriTable[FM_PCD_MANIP_DSCP_TO_VLAN_TRANS];
-+ /**< A table of VPri values for each DSCP value;
-+ The index is the DSCP value (0-0x3F) and the
-+ value is the corresponding VPRI (0-15). */
-+ uint8_t vpriDefVal; /**< 0-7, Relevant only if if updateType =
-+ e_FM_PCD_MANIP_HDR_FIELD_UPDATE_DSCP_TO_VLAN,
-+ this field is the Q Tag default value if the
-+ IP header is not found. */
-+} t_FmPcdManipHdrFieldUpdateVlanDscpToVpri;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining header manipulation VLAN fields updates
-+*//***************************************************************************/
-+typedef struct t_FmPcdManipHdrFieldUpdateVlan {
-+ e_FmPcdManipHdrFieldUpdateVlan updateType; /**< Selects VLAN update type */
-+ union {
-+ uint8_t vpri; /**< 0-7, Relevant only if If updateType =
-+ e_FM_PCD_MANIP_HDR_FIELD_UPDATE_VLAN_PRI, this
-+ is the new VLAN pri. */
-+ t_FmPcdManipHdrFieldUpdateVlanDscpToVpri dscpToVpri; /**< Parameters structure, Relevant only if updateType
-+ = e_FM_PCD_MANIP_HDR_FIELD_UPDATE_DSCP_TO_VLAN. */
-+ } u;
-+} t_FmPcdManipHdrFieldUpdateVlan;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining header manipulation IPV4 fields updates
-+*//***************************************************************************/
-+typedef struct t_FmPcdManipHdrFieldUpdateIpv4 {
-+ ipv4HdrManipUpdateFlags_t validUpdates; /**< ORed flag, selecting the required updates */
-+ uint8_t tos; /**< 8 bit New TOS; Relevant if validUpdates contains
-+ HDR_MANIP_IPV4_TOS */
-+ uint16_t id; /**< 16 bit New IP ID; Relevant only if validUpdates
-+ contains HDR_MANIP_IPV4_ID */
-+ uint32_t src; /**< 32 bit New IP SRC; Relevant only if validUpdates
-+ contains HDR_MANIP_IPV4_SRC */
-+ uint32_t dst; /**< 32 bit New IP DST; Relevant only if validUpdates
-+ contains HDR_MANIP_IPV4_DST */
-+} t_FmPcdManipHdrFieldUpdateIpv4;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining header manipulation IPV6 fields updates
-+*//***************************************************************************/
-+typedef struct t_FmPcdManipHdrFieldUpdateIpv6 {
-+ ipv6HdrManipUpdateFlags_t validUpdates; /**< ORed flag, selecting the required updates */
-+ uint8_t trafficClass; /**< 8 bit New Traffic Class; Relevant if validUpdates contains
-+ HDR_MANIP_IPV6_TC */
-+ uint8_t src[NET_HEADER_FIELD_IPv6_ADDR_SIZE];
-+ /**< 16 byte new IP SRC; Relevant only if validUpdates
-+ contains HDR_MANIP_IPV6_SRC */
-+ uint8_t dst[NET_HEADER_FIELD_IPv6_ADDR_SIZE];
-+ /**< 16 byte new IP DST; Relevant only if validUpdates
-+ contains HDR_MANIP_IPV6_DST */
-+} t_FmPcdManipHdrFieldUpdateIpv6;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining header manipulation TCP/UDP fields updates
-+*//***************************************************************************/
-+typedef struct t_FmPcdManipHdrFieldUpdateTcpUdp {
-+ tcpUdpHdrManipUpdateFlags_t validUpdates; /**< ORed flag, selecting the required updates */
-+ uint16_t src; /**< 16 bit New TCP/UDP SRC; Relevant only if validUpdates
-+ contains HDR_MANIP_TCP_UDP_SRC */
-+ uint16_t dst; /**< 16 bit New TCP/UDP DST; Relevant only if validUpdates
-+ contains HDR_MANIP_TCP_UDP_DST */
-+} t_FmPcdManipHdrFieldUpdateTcpUdp;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining header manipulation fields updates
-+*//***************************************************************************/
-+typedef struct t_FmPcdManipHdrFieldUpdateParams {
-+ e_FmPcdManipHdrFieldUpdateType type; /**< Type of header field update manipulation */
-+ union {
-+ t_FmPcdManipHdrFieldUpdateVlan vlan; /**< Parameters for VLAN update. Relevant when
-+ type = e_FM_PCD_MANIP_HDR_FIELD_UPDATE_VLAN */
-+ t_FmPcdManipHdrFieldUpdateIpv4 ipv4; /**< Parameters for IPv4 update. Relevant when
-+ type = e_FM_PCD_MANIP_HDR_FIELD_UPDATE_IPV4 */
-+ t_FmPcdManipHdrFieldUpdateIpv6 ipv6; /**< Parameters for IPv6 update. Relevant when
-+ type = e_FM_PCD_MANIP_HDR_FIELD_UPDATE_IPV6 */
-+ t_FmPcdManipHdrFieldUpdateTcpUdp tcpUdp; /**< Parameters for TCP/UDP update. Relevant when
-+ type = e_FM_PCD_MANIP_HDR_FIELD_UPDATE_TCP_UDP */
-+ } u;
-+} t_FmPcdManipHdrFieldUpdateParams;
-+
-+
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining custom header manipulation for generic field replacement
-+*//***************************************************************************/
-+typedef struct t_FmPcdManipHdrCustomGenFieldReplace {
-+ uint8_t srcOffset; /**< Location of new data - Offset from
-+ Parse Result (>= 16, srcOffset+size <= 32, ) */
-+ uint8_t dstOffset; /**< Location of data to be overwritten - Offset from
-+ start of frame (dstOffset + size <= 256). */
-+ uint8_t size; /**< The number of bytes (<=16) to be replaced */
-+ uint8_t mask; /**< Optional 1 byte mask. Set to select bits for
-+ replacement (1 - bit will be replaced);
-+ Clear to use field as is. */
-+ uint8_t maskOffset; /**< Relevant if mask != 0;
-+ Mask offset within the replaces "size" */
-+} t_FmPcdManipHdrCustomGenFieldReplace;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining custom header manipulation for IP replacement
-+*//***************************************************************************/
-+typedef struct t_FmPcdManipHdrCustomIpHdrReplace {
-+ e_FmPcdManipHdrCustomIpReplace replaceType; /**< Selects replace update type */
-+ bool decTtlHl; /**< Decrement TTL (IPV4) or Hop limit (IPV6) by 1 */
-+ bool updateIpv4Id; /**< Relevant when replaceType =
-+ e_FM_PCD_MANIP_HDR_CUSTOM_REPLACE_IPV6_BY_IPV4 */
-+ uint16_t id; /**< 16 bit New IP ID; Relevant only if
-+ updateIpv4Id = TRUE */
-+ uint8_t hdrSize; /**< The size of the new IP header */
-+ uint8_t hdr[FM_PCD_MANIP_MAX_HDR_SIZE];
-+ /**< The new IP header */
-+} t_FmPcdManipHdrCustomIpHdrReplace;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining custom header manipulation
-+*//***************************************************************************/
-+typedef struct t_FmPcdManipHdrCustomParams {
-+ e_FmPcdManipHdrCustomType type; /**< Type of header field update manipulation */
-+ union {
-+ t_FmPcdManipHdrCustomIpHdrReplace ipHdrReplace; /**< Parameters IP header replacement */
-+ t_FmPcdManipHdrCustomGenFieldReplace genFieldReplace; /**< Parameters IP header replacement */
-+ } u;
-+} t_FmPcdManipHdrCustomParams;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining specific L2 insertion manipulation
-+*//***************************************************************************/
-+typedef struct t_FmPcdManipHdrInsrtSpecificL2Params {
-+ e_FmPcdManipHdrInsrtSpecificL2 specificL2; /**< Selects which L2 headers to insert */
-+ bool update; /**< TRUE to update MPLS header */
-+ uint8_t size; /**< size of inserted section */
-+ uint8_t *p_Data; /**< data to be inserted */
-+} t_FmPcdManipHdrInsrtSpecificL2Params;
-+
-+#if (DPAA_VERSION >= 11)
-+/**************************************************************************//**
-+ @Description Parameters for defining IP insertion manipulation
-+*//***************************************************************************/
-+typedef struct t_FmPcdManipHdrInsrtIpParams {
-+ bool calcL4Checksum; /**< Calculate L4 checksum. */
-+ e_FmPcdManipHdrQosMappingMode mappingMode; /**< TODO */
-+ uint8_t lastPidOffset; /**< the offset of the last Protocol within
-+ the inserted header */
-+ uint16_t id; /**< 16 bit New IP ID */
-+ bool dontFragOverwrite;
-+ /**< IPv4 only. DF is overwritten with the hash-result next-to-last byte.
-+ * This byte is configured to be overwritten when RPD is set. */
-+ uint8_t lastDstOffset;
-+ /**< IPv6 only. if routing extension exist, user should set the offset of the destination address
-+ * in order to calculate UDP checksum pseudo header;
-+ * Otherwise set it to '0'. */
-+ t_FmPcdManipHdrInsrt insrt; /**< size and data to be inserted. */
-+} t_FmPcdManipHdrInsrtIpParams;
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining header insertion manipulation by header type
-+*//***************************************************************************/
-+typedef struct t_FmPcdManipHdrInsrtByHdrParams {
-+ e_FmPcdManipHdrInsrtByHdrType type; /**< Selects manipulation type */
-+ union {
-+
-+ t_FmPcdManipHdrInsrtSpecificL2Params specificL2Params;
-+ /**< Used when type = e_FM_PCD_MANIP_INSRT_BY_HDR_SPECIFIC_L2:
-+ Selects which L2 headers to insert */
-+#if (DPAA_VERSION >= 11)
-+ t_FmPcdManipHdrInsrtIpParams ipParams; /**< Used when type = e_FM_PCD_MANIP_INSRT_BY_HDR_IP */
-+ t_FmPcdManipHdrInsrt insrt; /**< Used when type is one of e_FM_PCD_MANIP_INSRT_BY_HDR_UDP,
-+ e_FM_PCD_MANIP_INSRT_BY_HDR_UDP_LITE, or
-+ e_FM_PCD_MANIP_INSRT_BY_HDR_CAPWAP */
-+#endif /* (DPAA_VERSION >= 11) */
-+ } u;
-+} t_FmPcdManipHdrInsrtByHdrParams;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining header insertion manipulation
-+*//***************************************************************************/
-+typedef struct t_FmPcdManipHdrInsrtParams {
-+ e_FmPcdManipHdrInsrtType type; /**< Type of insertion manipulation */
-+ union {
-+ t_FmPcdManipHdrInsrtByHdrParams byHdr; /**< Parameters for defining header insertion manipulation by header type,
-+ relevant if 'type' = e_FM_PCD_MANIP_INSRT_BY_HDR */
-+ t_FmPcdManipHdrInsrtGenericParams generic; /**< Parameters for defining generic header insertion manipulation,
-+ relevant if 'type' = e_FM_PCD_MANIP_INSRT_GENERIC */
-+#if ((DPAA_VERSION == 10) && defined(FM_CAPWAP_SUPPORT))
-+ t_FmPcdManipHdrInsrtByTemplateParams byTemplate; /**< Parameters for defining header insertion manipulation by template,
-+ relevant if 'type' = e_FM_PCD_MANIP_INSRT_BY_TEMPLATE */
-+#endif /* ((DPAA_VERSION == 10) && defined(FM_CAPWAP_SUPPORT)) */
-+ } u;
-+} t_FmPcdManipHdrInsrtParams;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining header removal manipulation
-+*//***************************************************************************/
-+typedef struct t_FmPcdManipHdrRmvParams {
-+ e_FmPcdManipHdrRmvType type; /**< Type of header removal manipulation */
-+ union {
-+ t_FmPcdManipHdrRmvByHdrParams byHdr; /**< Parameters for defining header removal manipulation by header type,
-+ relevant if type = e_FM_PCD_MANIP_RMV_BY_HDR */
-+ t_FmPcdManipHdrRmvGenericParams generic; /**< Parameters for defining generic header removal manipulation,
-+ relevant if type = e_FM_PCD_MANIP_RMV_GENERIC */
-+ } u;
-+} t_FmPcdManipHdrRmvParams;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining header manipulation node
-+*//***************************************************************************/
-+typedef struct t_FmPcdManipHdrParams {
-+ bool rmv; /**< TRUE, to define removal manipulation */
-+ t_FmPcdManipHdrRmvParams rmvParams; /**< Parameters for removal manipulation, relevant if 'rmv' = TRUE */
-+
-+ bool insrt; /**< TRUE, to define insertion manipulation */
-+ t_FmPcdManipHdrInsrtParams insrtParams; /**< Parameters for insertion manipulation, relevant if 'insrt' = TRUE */
-+
-+ bool fieldUpdate; /**< TRUE, to define field update manipulation */
-+ t_FmPcdManipHdrFieldUpdateParams fieldUpdateParams; /**< Parameters for field update manipulation, relevant if 'fieldUpdate' = TRUE */
-+
-+ bool custom; /**< TRUE, to define custom manipulation */
-+ t_FmPcdManipHdrCustomParams customParams; /**< Parameters for custom manipulation, relevant if 'custom' = TRUE */
-+
-+ bool dontParseAfterManip;/**< TRUE to de-activate the parser after the manipulation defined in this node.
-+ Restrictions:
-+ 1. MUST be set if the next engine after the CC is not another CC node
-+ (but rather Policer or Keygen), and this is the last (no h_NextManip) in a chain
-+ of manipulation nodes. This includes single nodes (i.e. no h_NextManip and
-+ also never pointed as h_NextManip of other manipulation nodes)
-+ 2. MUST be set if the next engine after the CC is another CC node, and
-+ this is NOT the last manipulation node (i.e. it has h_NextManip).*/
-+} t_FmPcdManipHdrParams;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining fragmentation manipulation
-+*//***************************************************************************/
-+typedef struct t_FmPcdManipFragParams {
-+ e_NetHeaderType hdr; /**< Header selection */
-+ union {
-+#if (DPAA_VERSION >= 11)
-+ t_FmPcdManipFragCapwapParams capwapFrag; /**< Parameters for defining CAPWAP fragmentation,
-+ relevant if 'hdr' = HEADER_TYPE_CAPWAP */
-+#endif /* (DPAA_VERSION >= 11) */
-+ t_FmPcdManipFragIpParams ipFrag; /**< Parameters for defining IP fragmentation,
-+ relevant if 'hdr' = HEADER_TYPE_Ipv4 or HEADER_TYPE_Ipv6 */
-+ } u;
-+} t_FmPcdManipFragParams;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining reassembly manipulation
-+*//***************************************************************************/
-+typedef struct t_FmPcdManipReassemParams {
-+ e_NetHeaderType hdr; /**< Header selection */
-+ union {
-+#if (DPAA_VERSION >= 11)
-+ t_FmPcdManipReassemCapwapParams capwapReassem; /**< Parameters for defining CAPWAP reassembly,
-+ relevant if 'hdr' = HEADER_TYPE_CAPWAP */
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+ t_FmPcdManipReassemIpParams ipReassem; /**< Parameters for defining IP reassembly,
-+ relevant if 'hdr' = HEADER_TYPE_Ipv4 or HEADER_TYPE_Ipv6 */
-+ } u;
-+} t_FmPcdManipReassemParams;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining a manipulation node
-+*//***************************************************************************/
-+typedef struct t_FmPcdManipParams {
-+ e_FmPcdManipType type; /**< Selects type of manipulation node */
-+ union{
-+ t_FmPcdManipHdrParams hdr; /**< Parameters for defining header manipulation node */
-+ t_FmPcdManipReassemParams reassem; /**< Parameters for defining reassembly manipulation node */
-+ t_FmPcdManipFragParams frag; /**< Parameters for defining fragmentation manipulation node */
-+ t_FmPcdManipSpecialOffloadParams specialOffload; /**< Parameters for defining special offload manipulation node */
-+ } u;
-+
-+ t_Handle h_NextManip; /**< Supported for Header Manipulation only;
-+ Handle to another (previously defined) manipulation node;
-+ Allows concatenation of manipulation actions;
-+ This parameter is optional and may be NULL. */
-+#if ((DPAA_VERSION == 10) && defined(FM_CAPWAP_SUPPORT))
-+ bool fragOrReasm; /**< TRUE, if defined fragmentation/reassembly manipulation */
-+ t_FmPcdManipFragOrReasmParams fragOrReasmParams; /**< Parameters for fragmentation/reassembly manipulation,
-+ relevant if fragOrReasm = TRUE */
-+#endif /* ((DPAA_VERSION == 10) && defined(FM_CAPWAP_SUPPORT)) */
-+} t_FmPcdManipParams;
-+
-+/**************************************************************************//**
-+ @Description Structure for retrieving IP reassembly statistics
-+*//***************************************************************************/
-+typedef struct t_FmPcdManipReassemIpStats {
-+ /* common counters for both IPv4 and IPv6 */
-+ uint32_t timeout; /**< Counts the number of timeout occurrences */
-+ uint32_t rfdPoolBusy; /**< Counts the number of failed attempts to allocate
-+ a Reassembly Frame Descriptor */
-+ uint32_t internalBufferBusy; /**< Counts the number of times an internal buffer busy occurred */
-+ uint32_t externalBufferBusy; /**< Counts the number of times external buffer busy occurred */
-+ uint32_t sgFragments; /**< Counts the number of Scatter/Gather fragments */
-+ uint32_t dmaSemaphoreDepletion; /**< Counts the number of failed attempts to allocate a DMA semaphore */
-+#if (DPAA_VERSION >= 11)
-+ uint32_t nonConsistentSp; /**< Counts the number of Non Consistent Storage Profile events for
-+ successfully reassembled frames */
-+#endif /* (DPAA_VERSION >= 11) */
-+ struct {
-+ uint32_t successfullyReassembled; /**< Counts the number of successfully reassembled frames */
-+ uint32_t validFragments; /**< Counts the total number of valid fragments that
-+ have been processed for all frames */
-+ uint32_t processedFragments; /**< Counts the number of processed fragments
-+ (valid and error fragments) for all frames */
-+ uint32_t malformedFragments; /**< Counts the number of malformed fragments processed for all frames */
-+ uint32_t discardedFragments; /**< Counts the number of fragments discarded by the reassembly process */
-+ uint32_t autoLearnBusy; /**< Counts the number of times a busy condition occurs when attempting
-+ to access an IP-Reassembly Automatic Learning Hash set */
-+ uint32_t moreThan16Fragments; /**< Counts the fragment occurrences in which the number of fragments-per-frame
-+ exceeds 16 */
-+ } specificHdrStatistics[2]; /**< slot '0' is for IPv4, slot '1' is for IPv6 */
-+} t_FmPcdManipReassemIpStats;
-+
-+/**************************************************************************//**
-+ @Description Structure for retrieving IP fragmentation statistics
-+*//***************************************************************************/
-+typedef struct t_FmPcdManipFragIpStats {
-+ uint32_t totalFrames; /**< Number of frames that passed through the manipulation node */
-+ uint32_t fragmentedFrames; /**< Number of frames that were fragmented */
-+ uint32_t generatedFragments; /**< Number of fragments that were generated */
-+} t_FmPcdManipFragIpStats;
-+
-+#if (DPAA_VERSION >= 11)
-+/**************************************************************************//**
-+ @Description Structure for retrieving CAPWAP reassembly statistics
-+*//***************************************************************************/
-+typedef struct t_FmPcdManipReassemCapwapStats {
-+ uint32_t timeout; /**< Counts the number of timeout occurrences */
-+ uint32_t rfdPoolBusy; /**< Counts the number of failed attempts to allocate
-+ a Reassembly Frame Descriptor */
-+ uint32_t internalBufferBusy; /**< Counts the number of times an internal buffer busy occurred */
-+ uint32_t externalBufferBusy; /**< Counts the number of times external buffer busy occurred */
-+ uint32_t sgFragments; /**< Counts the number of Scatter/Gather fragments */
-+ uint32_t dmaSemaphoreDepletion; /**< Counts the number of failed attempts to allocate a DMA semaphore */
-+ uint32_t successfullyReassembled; /**< Counts the number of successfully reassembled frames */
-+ uint32_t validFragments; /**< Counts the total number of valid fragments that
-+ have been processed for all frames */
-+ uint32_t processedFragments; /**< Counts the number of processed fragments
-+ (valid and error fragments) for all frames */
-+ uint32_t malformedFragments; /**< Counts the number of malformed fragments processed for all frames */
-+ uint32_t autoLearnBusy; /**< Counts the number of times a busy condition occurs when attempting
-+ to access an Reassembly Automatic Learning Hash set */
-+ uint32_t discardedFragments; /**< Counts the number of fragments discarded by the reassembly process */
-+ uint32_t moreThan16Fragments; /**< Counts the fragment occurrences in which the number of fragments-per-frame
-+ exceeds 16 */
-+ uint32_t exceedMaxReassemblyFrameLen;/**< ounts the number of times that a successful reassembled frame
-+ length exceeds MaxReassembledFrameLength value */
-+} t_FmPcdManipReassemCapwapStats;
-+
-+/**************************************************************************//**
-+ @Description Structure for retrieving CAPWAP fragmentation statistics
-+*//***************************************************************************/
-+typedef struct t_FmPcdManipFragCapwapStats {
-+ uint32_t totalFrames; /**< Number of frames that passed through the manipulation node */
-+ uint32_t fragmentedFrames; /**< Number of frames that were fragmented */
-+ uint32_t generatedFragments; /**< Number of fragments that were generated */
-+#if (defined(DEBUG_ERRORS) && (DEBUG_ERRORS > 0))
-+ uint8_t sgAllocationFailure; /**< Number of allocation failure of s/g buffers */
-+#endif /* (defined(DEBUG_ERRORS) && (DEBUG_ERRORS > 0)) */
-+} t_FmPcdManipFragCapwapStats;
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+/**************************************************************************//**
-+ @Description Structure for retrieving reassembly statistics
-+*//***************************************************************************/
-+typedef struct t_FmPcdManipReassemStats {
-+ union {
-+ t_FmPcdManipReassemIpStats ipReassem; /**< Structure for IP reassembly statistics */
-+#if (DPAA_VERSION >= 11)
-+ t_FmPcdManipReassemCapwapStats capwapReassem; /**< Structure for CAPWAP reassembly statistics */
-+#endif /* (DPAA_VERSION >= 11) */
-+ } u;
-+} t_FmPcdManipReassemStats;
-+
-+/**************************************************************************//**
-+ @Description Structure for retrieving fragmentation statistics
-+*//***************************************************************************/
-+typedef struct t_FmPcdManipFragStats {
-+ union {
-+ t_FmPcdManipFragIpStats ipFrag; /**< Structure for IP fragmentation statistics */
-+#if (DPAA_VERSION >= 11)
-+ t_FmPcdManipFragCapwapStats capwapFrag; /**< Structure for CAPWAP fragmentation statistics */
-+#endif /* (DPAA_VERSION >= 11) */
-+ } u;
-+} t_FmPcdManipFragStats;
-+
-+/**************************************************************************//**
-+ @Description Structure for selecting manipulation statistics
-+*//***************************************************************************/
-+typedef struct t_FmPcdManipStats {
-+ union {
-+ t_FmPcdManipReassemStats reassem; /**< Structure for reassembly statistics */
-+ t_FmPcdManipFragStats frag; /**< Structure for fragmentation statistics */
-+ } u;
-+} t_FmPcdManipStats;
-+
-+#if (DPAA_VERSION >= 11)
-+/**************************************************************************//**
-+ @Description Parameters for defining frame replicator group and its members
-+*//***************************************************************************/
-+typedef struct t_FmPcdFrmReplicGroupParams {
-+ uint8_t maxNumOfEntries; /**< Maximal number of members in the group;
-+ Must be at least 2. */
-+ uint8_t numOfEntries; /**< Number of members in the group;
-+ Must be at least 1. */
-+ t_FmPcdCcNextEngineParams nextEngineParams[FM_PCD_FRM_REPLIC_MAX_NUM_OF_ENTRIES];
-+ /**< Array of members' parameters */
-+} t_FmPcdFrmReplicGroupParams;
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+#if ((DPAA_VERSION == 10) && defined(FM_CAPWAP_SUPPORT))
-+/**************************************************************************//**
-+ @Description structure for defining statistics node
-+*//***************************************************************************/
-+typedef struct t_FmPcdStatsParams {
-+ e_FmPcdStatsType type; /**< type of statistics node */
-+} t_FmPcdStatsParams;
-+#endif /* ((DPAA_VERSION == 10) && defined(FM_CAPWAP_SUPPORT)) */
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_NetEnvCharacteristicsSet
-+
-+ @Description Define a set of Network Environment Characteristics.
-+
-+ When setting an environment it is important to understand its
-+ application. It is not meant to describe the flows that will run
-+ on the ports using this environment, but what the user means TO DO
-+ with the PCD mechanisms in order to parse-classify-distribute those
-+ frames.
-+ By specifying a distinction unit, the user means it would use that option
-+ for distinction between frames at either a KeyGen scheme or a coarse
-+ classification action descriptor. Using interchangeable headers to define a
-+ unit means that the user is indifferent to which of the interchangeable
-+ headers is present in the frame, and wants the distinction to be based
-+ on the presence of either one of them.
-+
-+ Depending on context, there are limitations to the use of environments. A
-+ port using the PCD functionality is bound to an environment. Some or even
-+ all ports may share an environment but also an environment per port is
-+ possible. When initializing a scheme, a classification plan group (see below),
-+ or a coarse classification tree, one of the initialized environments must be
-+ stated and related to. When a port is bound to a scheme, a classification
-+ plan group, or a coarse classification tree, it MUST be bound to the same
-+ environment.
-+
-+ The different PCD modules, may relate (for flows definition) ONLY on
-+ distinction units as defined by their environment. When initializing a
-+ scheme for example, it may not choose to select IPV4 as a match for
-+ recognizing flows unless it was defined in the relating environment. In
-+ fact, to guide the user through the configuration of the PCD, each module's
-+ characterization in terms of flows is not done using protocol names, but using
-+ environment indexes.
-+
-+ In terms of HW implementation, the list of distinction units sets the LCV vectors
-+ and later used for match vector, classification plan vectors and coarse classification
-+ indexing.
-+
-+ @Param[in] h_FmPcd FM PCD module descriptor.
-+ @Param[in] p_NetEnvParams A structure of parameters for the initialization of
-+ the network environment.
-+
-+ @Return A handle to the initialized object on success; NULL code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_Init().
-+*//***************************************************************************/
-+t_Handle FM_PCD_NetEnvCharacteristicsSet(t_Handle h_FmPcd, t_FmPcdNetEnvParams *p_NetEnvParams);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_NetEnvCharacteristicsDelete
-+
-+ @Description Deletes a set of Network Environment Characteristics.
-+
-+ @Param[in] h_NetEnv A handle to the Network environment.
-+
-+ @Return E_OK on success; Error code otherwise.
-+*//***************************************************************************/
-+t_Error FM_PCD_NetEnvCharacteristicsDelete(t_Handle h_NetEnv);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_KgSchemeSet
-+
-+ @Description Initializing or modifying and enabling a scheme for the KeyGen.
-+ This routine should be called for adding or modifying a scheme.
-+ When a scheme needs modifying, the API requires that it will be
-+ rewritten. In such a case 'modify' should be TRUE. If the
-+ routine is called for a valid scheme and 'modify' is FALSE,
-+ it will return error.
-+
-+ @Param[in] h_FmPcd If this is a new scheme - A handle to an FM PCD Module.
-+ Otherwise NULL (ignored by driver).
-+ @Param[in,out] p_SchemeParams A structure of parameters for defining the scheme
-+
-+ @Return A handle to the initialized scheme on success; NULL code otherwise.
-+ When used as "modify" (rather than for setting a new scheme),
-+ p_SchemeParams->id.h_Scheme will return NULL if action fails due to scheme
-+ BUSY state.
-+
-+ @Cautions Allowed only following FM_PCD_Init().
-+*//***************************************************************************/
-+t_Handle FM_PCD_KgSchemeSet(t_Handle h_FmPcd,
-+ t_FmPcdKgSchemeParams *p_SchemeParams);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_KgSchemeDelete
-+
-+ @Description Deleting an initialized scheme.
-+
-+ @Param[in] h_Scheme scheme handle as returned by FM_PCD_KgSchemeSet()
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_Init() & FM_PCD_KgSchemeSet().
-+*//***************************************************************************/
-+t_Error FM_PCD_KgSchemeDelete(t_Handle h_Scheme);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_KgSchemeGetCounter
-+
-+ @Description Reads scheme packet counter.
-+
-+ @Param[in] h_Scheme scheme handle as returned by FM_PCD_KgSchemeSet().
-+
-+ @Return Counter's current value.
-+
-+ @Cautions Allowed only following FM_PCD_Init() & FM_PCD_KgSchemeSet().
-+*//***************************************************************************/
-+uint32_t FM_PCD_KgSchemeGetCounter(t_Handle h_Scheme);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_KgSchemeSetCounter
-+
-+ @Description Writes scheme packet counter.
-+
-+ @Param[in] h_Scheme scheme handle as returned by FM_PCD_KgSchemeSet().
-+ @Param[in] value New scheme counter value - typically '0' for
-+ resetting the counter.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_Init() & FM_PCD_KgSchemeSet().
-+*//***************************************************************************/
-+t_Error FM_PCD_KgSchemeSetCounter(t_Handle h_Scheme, uint32_t value);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_PlcrProfileSet
-+
-+ @Description Sets a profile entry in the policer profile table.
-+ The routine overrides any existing value.
-+
-+ @Param[in] h_FmPcd A handle to an FM PCD Module.
-+ @Param[in] p_Profile A structure of parameters for defining a
-+ policer profile entry.
-+
-+ @Return A handle to the initialized object on success; NULL code otherwise.
-+ When used as "modify" (rather than for setting a new profile),
-+ p_Profile->id.h_Profile will return NULL if action fails due to profile
-+ BUSY state.
-+ @Cautions Allowed only following FM_PCD_Init().
-+*//***************************************************************************/
-+t_Handle FM_PCD_PlcrProfileSet(t_Handle h_FmPcd,
-+ t_FmPcdPlcrProfileParams *p_Profile);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_PlcrProfileDelete
-+
-+ @Description Delete a profile entry in the policer profile table.
-+ The routine set entry to invalid.
-+
-+ @Param[in] h_Profile A handle to the profile.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_Init().
-+*//***************************************************************************/
-+t_Error FM_PCD_PlcrProfileDelete(t_Handle h_Profile);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_PlcrProfileGetCounter
-+
-+ @Description Sets an entry in the classification plan.
-+ The routine overrides any existing value.
-+
-+ @Param[in] h_Profile A handle to the profile.
-+ @Param[in] counter Counter selector.
-+
-+ @Return specific counter value.
-+
-+ @Cautions Allowed only following FM_PCD_Init().
-+*//***************************************************************************/
-+uint32_t FM_PCD_PlcrProfileGetCounter(t_Handle h_Profile,
-+ e_FmPcdPlcrProfileCounters counter);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_PlcrProfileSetCounter
-+
-+ @Description Sets an entry in the classification plan.
-+ The routine overrides any existing value.
-+
-+ @Param[in] h_Profile A handle to the profile.
-+ @Param[in] counter Counter selector.
-+ @Param[in] value value to set counter with.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_Init().
-+*//***************************************************************************/
-+t_Error FM_PCD_PlcrProfileSetCounter(t_Handle h_Profile,
-+ e_FmPcdPlcrProfileCounters counter,
-+ uint32_t value);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_CcRootBuild
-+
-+ @Description This routine must be called to define a complete coarse
-+ classification tree. This is the way to define coarse
-+ classification to a certain flow - the KeyGen schemes
-+ may point only to trees defined in this way.
-+
-+ @Param[in] h_FmPcd FM PCD module descriptor.
-+ @Param[in] p_Params A structure of parameters to define the tree.
-+
-+ @Return A handle to the initialized object on success; NULL code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_Init().
-+*//***************************************************************************/
-+t_Handle FM_PCD_CcRootBuild (t_Handle h_FmPcd,
-+ t_FmPcdCcTreeParams *p_Params);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_CcRootDelete
-+
-+ @Description Deleting an built tree.
-+
-+ @Param[in] h_CcTree A handle to a CC tree.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_Init().
-+*//***************************************************************************/
-+t_Error FM_PCD_CcRootDelete(t_Handle h_CcTree);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_CcRootModifyNextEngine
-+
-+ @Description Modify the Next Engine Parameters in the entry of the tree.
-+
-+ @Param[in] h_CcTree A handle to the tree
-+ @Param[in] grpId A Group index in the tree
-+ @Param[in] index Entry index in the group defined by grpId
-+ @Param[in] p_FmPcdCcNextEngineParams Pointer to new next engine parameters
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_CcBuildTree().
-+*//***************************************************************************/
-+t_Error FM_PCD_CcRootModifyNextEngine(t_Handle h_CcTree,
-+ uint8_t grpId,
-+ uint8_t index,
-+ t_FmPcdCcNextEngineParams *p_FmPcdCcNextEngineParams);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_MatchTableSet
-+
-+ @Description This routine should be called for each CC (coarse classification)
-+ node. The whole CC tree should be built bottom up so that each
-+ node points to already defined nodes.
-+
-+ @Param[in] h_FmPcd FM PCD module descriptor.
-+ @Param[in] p_Param A structure of parameters defining the CC node
-+
-+ @Return A handle to the initialized object on success; NULL code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_Init().
-+*//***************************************************************************/
-+t_Handle FM_PCD_MatchTableSet(t_Handle h_FmPcd, t_FmPcdCcNodeParams *p_Param);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_MatchTableDelete
-+
-+ @Description Deleting an built node.
-+
-+ @Param[in] h_CcNode A handle to a CC node.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_Init().
-+*//***************************************************************************/
-+t_Error FM_PCD_MatchTableDelete(t_Handle h_CcNode);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_MatchTableModifyMissNextEngine
-+
-+ @Description Modify the Next Engine Parameters of the Miss key case of the node.
-+
-+ @Param[in] h_CcNode A handle to the node
-+ @Param[in] p_FmPcdCcNextEngineParams Parameters for defining next engine
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_MatchTableSet();
-+ Not relevant in the case the node is of type 'INDEXED_LOOKUP'.
-+ When configuring nextEngine = e_FM_PCD_CC, note that
-+ p_FmPcdCcNextEngineParams->ccParams.h_CcNode must be different
-+ from the currently changed table.
-+
-+*//***************************************************************************/
-+t_Error FM_PCD_MatchTableModifyMissNextEngine(t_Handle h_CcNode,
-+ t_FmPcdCcNextEngineParams *p_FmPcdCcNextEngineParams);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_MatchTableRemoveKey
-+
-+ @Description Remove the key (including next engine parameters of this key)
-+ defined by the index of the relevant node.
-+
-+ @Param[in] h_CcNode A handle to the node
-+ @Param[in] keyIndex Key index for removing
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_MatchTableSet() was called for this
-+ node and the nodes that lead to it.
-+*//***************************************************************************/
-+t_Error FM_PCD_MatchTableRemoveKey(t_Handle h_CcNode, uint16_t keyIndex);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_MatchTableAddKey
-+
-+ @Description Add the key (including next engine parameters of this key in the
-+ index defined by the keyIndex. Note that 'FM_PCD_LAST_KEY_INDEX'
-+ may be used by user that don't care about the position of the
-+ key in the table - in that case, the key will be automatically
-+ added by the driver in the last available entry.
-+
-+ @Param[in] h_CcNode A handle to the node
-+ @Param[in] keyIndex Key index for adding.
-+ @Param[in] keySize Key size of added key
-+ @Param[in] p_KeyParams A pointer to the parameters includes
-+ new key with Next Engine Parameters
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_MatchTableSet() was called for this
-+ node and the nodes that lead to it.
-+*//***************************************************************************/
-+t_Error FM_PCD_MatchTableAddKey(t_Handle h_CcNode,
-+ uint16_t keyIndex,
-+ uint8_t keySize,
-+ t_FmPcdCcKeyParams *p_KeyParams);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_MatchTableModifyNextEngine
-+
-+ @Description Modify the Next Engine Parameters in the relevant key entry of the node.
-+
-+ @Param[in] h_CcNode A handle to the node
-+ @Param[in] keyIndex Key index for Next Engine modifications
-+ @Param[in] p_FmPcdCcNextEngineParams Parameters for defining next engine
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_MatchTableSet().
-+ When configuring nextEngine = e_FM_PCD_CC, note that
-+ p_FmPcdCcNextEngineParams->ccParams.h_CcNode must be different
-+ from the currently changed table.
-+
-+*//***************************************************************************/
-+t_Error FM_PCD_MatchTableModifyNextEngine(t_Handle h_CcNode,
-+ uint16_t keyIndex,
-+ t_FmPcdCcNextEngineParams *p_FmPcdCcNextEngineParams);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_MatchTableModifyKeyAndNextEngine
-+
-+ @Description Modify the key and Next Engine Parameters of this key in the
-+ index defined by the keyIndex.
-+
-+ @Param[in] h_CcNode A handle to the node
-+ @Param[in] keyIndex Key index for adding
-+ @Param[in] keySize Key size of added key
-+ @Param[in] p_KeyParams A pointer to the parameters includes
-+ modified key and modified Next Engine Parameters
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_MatchTableSet() was called for this
-+ node and the nodes that lead to it.
-+ When configuring nextEngine = e_FM_PCD_CC, note that
-+ p_FmPcdCcNextEngineParams->ccParams.h_CcNode must be different
-+ from the currently changed table.
-+*//***************************************************************************/
-+t_Error FM_PCD_MatchTableModifyKeyAndNextEngine(t_Handle h_CcNode,
-+ uint16_t keyIndex,
-+ uint8_t keySize,
-+ t_FmPcdCcKeyParams *p_KeyParams);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_MatchTableModifyKey
-+
-+ @Description Modify the key in the index defined by the keyIndex.
-+
-+ @Param[in] h_CcNode A handle to the node
-+ @Param[in] keyIndex Key index for adding
-+ @Param[in] keySize Key size of added key
-+ @Param[in] p_Key A pointer to the new key
-+ @Param[in] p_Mask A pointer to the new mask if relevant,
-+ otherwise pointer to NULL
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_MatchTableSet() was called for this
-+ node and the nodes that lead to it.
-+*//***************************************************************************/
-+t_Error FM_PCD_MatchTableModifyKey(t_Handle h_CcNode,
-+ uint16_t keyIndex,
-+ uint8_t keySize,
-+ uint8_t *p_Key,
-+ uint8_t *p_Mask);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_MatchTableFindNRemoveKey
-+
-+ @Description Remove the key (including next engine parameters of this key)
-+ defined by the key and mask. Note that this routine will search
-+ the node to locate the index of the required key (& mask) to remove.
-+
-+ @Param[in] h_CcNode A handle to the node
-+ @Param[in] keySize Key size of the one to remove.
-+ @Param[in] p_Key A pointer to the requested key to remove.
-+ @Param[in] p_Mask A pointer to the mask if relevant,
-+ otherwise pointer to NULL
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_MatchTableSet() was called for this
-+ node and the nodes that lead to it.
-+*//***************************************************************************/
-+t_Error FM_PCD_MatchTableFindNRemoveKey(t_Handle h_CcNode,
-+ uint8_t keySize,
-+ uint8_t *p_Key,
-+ uint8_t *p_Mask);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_MatchTableFindNModifyNextEngine
-+
-+ @Description Modify the Next Engine Parameters in the relevant key entry of
-+ the node. Note that this routine will search the node to locate
-+ the index of the required key (& mask) to modify.
-+
-+ @Param[in] h_CcNode A handle to the node
-+ @Param[in] keySize Key size of the one to modify.
-+ @Param[in] p_Key A pointer to the requested key to modify.
-+ @Param[in] p_Mask A pointer to the mask if relevant,
-+ otherwise pointer to NULL
-+ @Param[in] p_FmPcdCcNextEngineParams Parameters for defining next engine
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_MatchTableSet().
-+ When configuring nextEngine = e_FM_PCD_CC, note that
-+ p_FmPcdCcNextEngineParams->ccParams.h_CcNode must be different
-+ from the currently changed table.
-+*//***************************************************************************/
-+t_Error FM_PCD_MatchTableFindNModifyNextEngine(t_Handle h_CcNode,
-+ uint8_t keySize,
-+ uint8_t *p_Key,
-+ uint8_t *p_Mask,
-+ t_FmPcdCcNextEngineParams *p_FmPcdCcNextEngineParams);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_MatchTableFindNModifyKeyAndNextEngine
-+
-+ @Description Modify the key and Next Engine Parameters of this key in the
-+ index defined by the keyIndex. Note that this routine will search
-+ the node to locate the index of the required key (& mask) to modify.
-+
-+ @Param[in] h_CcNode A handle to the node
-+ @Param[in] keySize Key size of the one to modify.
-+ @Param[in] p_Key A pointer to the requested key to modify.
-+ @Param[in] p_Mask A pointer to the mask if relevant,
-+ otherwise pointer to NULL
-+ @Param[in] p_KeyParams A pointer to the parameters includes
-+ modified key and modified Next Engine Parameters
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_MatchTableSet() was called for this
-+ node and the nodes that lead to it.
-+ When configuring nextEngine = e_FM_PCD_CC, note that
-+ p_FmPcdCcNextEngineParams->ccParams.h_CcNode must be different
-+ from the currently changed table.
-+*//***************************************************************************/
-+t_Error FM_PCD_MatchTableFindNModifyKeyAndNextEngine(t_Handle h_CcNode,
-+ uint8_t keySize,
-+ uint8_t *p_Key,
-+ uint8_t *p_Mask,
-+ t_FmPcdCcKeyParams *p_KeyParams);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_MatchTableFindNModifyKey
-+
-+ @Description Modify the key in the index defined by the keyIndex. Note that
-+ this routine will search the node to locate the index of the
-+ required key (& mask) to modify.
-+
-+ @Param[in] h_CcNode A handle to the node
-+ @Param[in] keySize Key size of the one to modify.
-+ @Param[in] p_Key A pointer to the requested key to modify.
-+ @Param[in] p_Mask A pointer to the mask if relevant,
-+ otherwise pointer to NULL
-+ @Param[in] p_NewKey A pointer to the new key
-+ @Param[in] p_NewMask A pointer to the new mask if relevant,
-+ otherwise pointer to NULL
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_MatchTableSet() was called for this
-+ node and the nodes that lead to it.
-+*//***************************************************************************/
-+t_Error FM_PCD_MatchTableFindNModifyKey(t_Handle h_CcNode,
-+ uint8_t keySize,
-+ uint8_t *p_Key,
-+ uint8_t *p_Mask,
-+ uint8_t *p_NewKey,
-+ uint8_t *p_NewMask);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_MatchTableGetKeyCounter
-+
-+ @Description This routine may be used to get a counter of specific key in a CC
-+ Node; This counter reflects how many frames passed that were matched
-+ this key.
-+
-+ @Param[in] h_CcNode A handle to the node
-+ @Param[in] keyIndex Key index for adding
-+
-+ @Return The specific key counter.
-+
-+ @Cautions Allowed only following FM_PCD_MatchTableSet().
-+*//***************************************************************************/
-+uint32_t FM_PCD_MatchTableGetKeyCounter(t_Handle h_CcNode, uint16_t keyIndex);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_MatchTableGetKeyStatistics
-+
-+ @Description This routine may be used to get statistics counters of specific key
-+ in a CC Node.
-+
-+ If 'e_FM_PCD_CC_STATS_MODE_FRAME' and
-+ 'e_FM_PCD_CC_STATS_MODE_BYTE_AND_FRAME' were set for this node,
-+ these counters reflect how many frames passed that were matched
-+ this key; The total frames count will be returned in the counter
-+ of the first range (as only one frame length range was defined).
-+ If 'e_FM_PCD_CC_STATS_MODE_RMON' was set for this node, the total
-+ frame count will be separated to frame length counters, based on
-+ provided frame length ranges.
-+
-+ @Param[in] h_CcNode A handle to the node
-+ @Param[in] keyIndex Key index for adding
-+ @Param[out] p_KeyStatistics Key statistics counters
-+
-+ @Return The specific key statistics.
-+
-+ @Cautions Allowed only following FM_PCD_MatchTableSet().
-+*//***************************************************************************/
-+t_Error FM_PCD_MatchTableGetKeyStatistics(t_Handle h_CcNode,
-+ uint16_t keyIndex,
-+ t_FmPcdCcKeyStatistics *p_KeyStatistics);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_MatchTableGetMissStatistics
-+
-+ @Description This routine may be used to get statistics counters of miss entry
-+ in a CC Node.
-+
-+ If 'e_FM_PCD_CC_STATS_MODE_FRAME' and
-+ 'e_FM_PCD_CC_STATS_MODE_BYTE_AND_FRAME' were set for this node,
-+ these counters reflect how many frames were not matched to any
-+ existing key and therefore passed through the miss entry; The
-+ total frames count will be returned in the counter of the
-+ first range (as only one frame length range was defined).
-+
-+ @Param[in] h_CcNode A handle to the node
-+ @Param[out] p_MissStatistics Statistics counters for 'miss'
-+
-+ @Return The statistics for 'miss'.
-+
-+ @Cautions Allowed only following FM_PCD_MatchTableSet().
-+*//***************************************************************************/
-+t_Error FM_PCD_MatchTableGetMissStatistics(t_Handle h_CcNode,
-+ t_FmPcdCcKeyStatistics *p_MissStatistics);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_MatchTableFindNGetKeyStatistics
-+
-+ @Description This routine may be used to get statistics counters of specific key
-+ in a CC Node.
-+
-+ If 'e_FM_PCD_CC_STATS_MODE_FRAME' and
-+ 'e_FM_PCD_CC_STATS_MODE_BYTE_AND_FRAME' were set for this node,
-+ these counters reflect how many frames passed that were matched
-+ this key; The total frames count will be returned in the counter
-+ of the first range (as only one frame length range was defined).
-+ If 'e_FM_PCD_CC_STATS_MODE_RMON' was set for this node, the total
-+ frame count will be separated to frame length counters, based on
-+ provided frame length ranges.
-+ Note that this routine will search the node to locate the index
-+ of the required key based on received key parameters.
-+
-+ @Param[in] h_CcNode A handle to the node
-+ @Param[in] keySize Size of the requested key
-+ @Param[in] p_Key A pointer to the requested key
-+ @Param[in] p_Mask A pointer to the mask if relevant,
-+ otherwise pointer to NULL
-+ @Param[out] p_KeyStatistics Key statistics counters
-+
-+ @Return The specific key statistics.
-+
-+ @Cautions Allowed only following FM_PCD_MatchTableSet().
-+*//***************************************************************************/
-+t_Error FM_PCD_MatchTableFindNGetKeyStatistics(t_Handle h_CcNode,
-+ uint8_t keySize,
-+ uint8_t *p_Key,
-+ uint8_t *p_Mask,
-+ t_FmPcdCcKeyStatistics *p_KeyStatistics);
-+
-+/**************************************************************************//*
-+ @Function FM_PCD_MatchTableGetNextEngine
-+
-+ @Description Gets NextEngine of the relevant keyIndex.
-+
-+ @Param[in] h_CcNode A handle to the node.
-+ @Param[in] keyIndex keyIndex in the relevant node.
-+ @Param[out] p_FmPcdCcNextEngineParams here updated nextEngine parameters for
-+ the relevant keyIndex of the CC Node
-+ received as parameter to this function
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_Init().
-+*//***************************************************************************/
-+t_Error FM_PCD_MatchTableGetNextEngine(t_Handle h_CcNode,
-+ uint16_t keyIndex,
-+ t_FmPcdCcNextEngineParams *p_FmPcdCcNextEngineParams);
-+
-+/**************************************************************************//*
-+ @Function FM_PCD_MatchTableGetIndexedHashBucket
-+
-+ @Description This routine simulates KeyGen operation on the provided key and
-+ calculates to which hash bucket it will be mapped.
-+
-+ @Param[in] h_CcNode A handle to the node.
-+ @Param[in] kgKeySize Key size as it was configured in the KG
-+ scheme that leads to this hash.
-+ @Param[in] p_KgKey Pointer to the key; must be like the key
-+ that the KG is generated, i.e. the same
-+ extraction and with mask if exist.
-+ @Param[in] kgHashShift Hash-shift as it was configured in the KG
-+ scheme that leads to this hash.
-+ @Param[out] p_CcNodeBucketHandle Pointer to the bucket of the provided key.
-+ @Param[out] p_BucketIndex Index to the bucket of the provided key
-+ @Param[out] p_LastIndex Pointer to last index in the bucket of the
-+ provided key.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_HashTableSet()
-+*//***************************************************************************/
-+t_Error FM_PCD_MatchTableGetIndexedHashBucket(t_Handle h_CcNode,
-+ uint8_t kgKeySize,
-+ uint8_t *p_KgKey,
-+ uint8_t kgHashShift,
-+ t_Handle *p_CcNodeBucketHandle,
-+ uint8_t *p_BucketIndex,
-+ uint16_t *p_LastIndex);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_HashTableSet
-+
-+ @Description This routine initializes a hash table structure.
-+ KeyGen hash result determines the hash bucket.
-+ Next, KeyGen key is compared against all keys of this
-+ bucket (exact match).
-+ Number of sets (number of buckets) of the hash equals to the
-+ number of 1-s in 'hashResMask' in the provided parameters.
-+ Number of hash table ways is then calculated by dividing
-+ 'maxNumOfKeys' equally between the hash sets. This is the maximal
-+ number of keys that a hash bucket may hold.
-+ The hash table is initialized empty and keys may be
-+ added to it following the initialization. Keys masks are not
-+ supported in current hash table implementation.
-+ The initialized hash table can be integrated as a node in a
-+ CC tree.
-+
-+ @Param[in] h_FmPcd FM PCD module descriptor.
-+ @Param[in] p_Param A structure of parameters defining the hash table
-+
-+ @Return A handle to the initialized object on success; NULL code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_Init().
-+*//***************************************************************************/
-+t_Handle FM_PCD_HashTableSet(t_Handle h_FmPcd, t_FmPcdHashTableParams *p_Param);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_HashTableDelete
-+
-+ @Description This routine deletes the provided hash table and released all
-+ its allocated resources.
-+
-+ @Param[in] h_HashTbl A handle to a hash table
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_HashTableSet().
-+*//***************************************************************************/
-+t_Error FM_PCD_HashTableDelete(t_Handle h_HashTbl);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_HashTableAddKey
-+
-+ @Description This routine adds the provided key (including next engine
-+ parameters of this key) to the hash table.
-+ The key is added as the last key of the bucket that it is
-+ mapped to.
-+
-+ @Param[in] h_HashTbl A handle to a hash table
-+ @Param[in] keySize Key size of added key
-+ @Param[in] p_KeyParams A pointer to the parameters includes
-+ new key with next engine parameters; The pointer
-+ to the key mask must be NULL, as masks are not
-+ supported in hash table implementation.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_HashTableSet().
-+*//***************************************************************************/
-+t_Error FM_PCD_HashTableAddKey(t_Handle h_HashTbl,
-+ uint8_t keySize,
-+ t_FmPcdCcKeyParams *p_KeyParams);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_HashTableRemoveKey
-+
-+ @Description This routine removes the requested key (including next engine
-+ parameters of this key) from the hash table.
-+
-+ @Param[in] h_HashTbl A handle to a hash table
-+ @Param[in] keySize Key size of the one to remove.
-+ @Param[in] p_Key A pointer to the requested key to remove.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_HashTableSet().
-+*//***************************************************************************/
-+t_Error FM_PCD_HashTableRemoveKey(t_Handle h_HashTbl,
-+ uint8_t keySize,
-+ uint8_t *p_Key);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_HashTableModifyNextEngine
-+
-+ @Description This routine modifies the next engine for the provided key. The
-+ key should be previously added to the hash table.
-+
-+ @Param[in] h_HashTbl A handle to a hash table
-+ @Param[in] keySize Key size of the key to modify.
-+ @Param[in] p_Key A pointer to the requested key to modify.
-+ @Param[in] p_FmPcdCcNextEngineParams A structure for defining new next engine
-+ parameters.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_HashTableSet().
-+ When configuring nextEngine = e_FM_PCD_CC, note that
-+ p_FmPcdCcNextEngineParams->ccParams.h_CcNode must be different
-+ from the currently changed table.
-+*//***************************************************************************/
-+t_Error FM_PCD_HashTableModifyNextEngine(t_Handle h_HashTbl,
-+ uint8_t keySize,
-+ uint8_t *p_Key,
-+ t_FmPcdCcNextEngineParams *p_FmPcdCcNextEngineParams);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_HashTableModifyMissNextEngine
-+
-+ @Description This routine modifies the next engine on key match miss.
-+
-+ @Param[in] h_HashTbl A handle to a hash table
-+ @Param[in] p_FmPcdCcNextEngineParams A structure for defining new next engine
-+ parameters.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_HashTableSet().
-+ When configuring nextEngine = e_FM_PCD_CC, note that
-+ p_FmPcdCcNextEngineParams->ccParams.h_CcNode must be different
-+ from the currently changed table.
-+*//***************************************************************************/
-+t_Error FM_PCD_HashTableModifyMissNextEngine(t_Handle h_HashTbl,
-+ t_FmPcdCcNextEngineParams *p_FmPcdCcNextEngineParams);
-+
-+/**************************************************************************//*
-+ @Function FM_PCD_HashTableGetMissNextEngine
-+
-+ @Description Gets NextEngine in case of key match miss.
-+
-+ @Param[in] h_HashTbl A handle to a hash table
-+ @Param[out] p_FmPcdCcNextEngineParams Next engine parameters for the specified
-+ hash table.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_HashTableSet().
-+*//***************************************************************************/
-+t_Error FM_PCD_HashTableGetMissNextEngine(t_Handle h_HashTbl,
-+ t_FmPcdCcNextEngineParams *p_FmPcdCcNextEngineParams);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_HashTableFindNGetKeyStatistics
-+
-+ @Description This routine may be used to get statistics counters of specific key
-+ in a hash table.
-+
-+ If 'e_FM_PCD_CC_STATS_MODE_FRAME' and
-+ 'e_FM_PCD_CC_STATS_MODE_BYTE_AND_FRAME' were set for this node,
-+ these counters reflect how many frames passed that were matched
-+ this key; The total frames count will be returned in the counter
-+ of the first range (as only one frame length range was defined).
-+ If 'e_FM_PCD_CC_STATS_MODE_RMON' was set for this node, the total
-+ frame count will be separated to frame length counters, based on
-+ provided frame length ranges.
-+ Note that this routine will identify the bucket of this key in
-+ the hash table and will search the bucket to locate the index
-+ of the required key based on received key parameters.
-+
-+ @Param[in] h_HashTbl A handle to a hash table
-+ @Param[in] keySize Size of the requested key
-+ @Param[in] p_Key A pointer to the requested key
-+ @Param[out] p_KeyStatistics Key statistics counters
-+
-+ @Return The specific key statistics.
-+
-+ @Cautions Allowed only following FM_PCD_HashTableSet().
-+*//***************************************************************************/
-+t_Error FM_PCD_HashTableFindNGetKeyStatistics(t_Handle h_HashTbl,
-+ uint8_t keySize,
-+ uint8_t *p_Key,
-+ t_FmPcdCcKeyStatistics *p_KeyStatistics);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_HashTableGetMissStatistics
-+
-+ @Description This routine may be used to get statistics counters of 'miss'
-+ entry of the a hash table.
-+
-+ If 'e_FM_PCD_CC_STATS_MODE_FRAME' and
-+ 'e_FM_PCD_CC_STATS_MODE_BYTE_AND_FRAME' were set for this node,
-+ these counters reflect how many frames were not matched to any
-+ existing key and therefore passed through the miss entry;
-+
-+ @Param[in] h_HashTbl A handle to a hash table
-+ @Param[out] p_MissStatistics Statistics counters for 'miss'
-+
-+ @Return The statistics for 'miss'.
-+
-+ @Cautions Allowed only following FM_PCD_HashTableSet().
-+*//***************************************************************************/
-+t_Error FM_PCD_HashTableGetMissStatistics(t_Handle h_HashTbl,
-+ t_FmPcdCcKeyStatistics *p_MissStatistics);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_ManipNodeSet
-+
-+ @Description This routine should be called for defining a manipulation
-+ node. A manipulation node must be defined before the CC node
-+ that precedes it.
-+
-+ @Param[in] h_FmPcd FM PCD module descriptor.
-+ @Param[in] p_FmPcdManipParams A structure of parameters defining the manipulation
-+
-+ @Return A handle to the initialized object on success; NULL code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_Init().
-+*//***************************************************************************/
-+t_Handle FM_PCD_ManipNodeSet(t_Handle h_FmPcd, t_FmPcdManipParams *p_FmPcdManipParams);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_ManipNodeDelete
-+
-+ @Description Delete an existing manipulation node.
-+
-+ @Param[in] h_ManipNode A handle to a manipulation node.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_ManipNodeSet().
-+*//***************************************************************************/
-+t_Error FM_PCD_ManipNodeDelete(t_Handle h_ManipNode);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_ManipGetStatistics
-+
-+ @Description Retrieve the manipulation statistics.
-+
-+ @Param[in] h_ManipNode A handle to a manipulation node.
-+ @Param[out] p_FmPcdManipStats A structure for retrieving the manipulation statistics
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_ManipNodeSet().
-+*//***************************************************************************/
-+t_Error FM_PCD_ManipGetStatistics(t_Handle h_ManipNode, t_FmPcdManipStats *p_FmPcdManipStats);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_ManipNodeReplace
-+
-+ @Description Change existing manipulation node to be according to new requirement.
-+
-+ @Param[in] h_ManipNode A handle to a manipulation node.
-+ @Param[out] p_ManipParams A structure of parameters defining the change requirement
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_ManipNodeSet().
-+*//***************************************************************************/
-+t_Error FM_PCD_ManipNodeReplace(t_Handle h_ManipNode, t_FmPcdManipParams *p_ManipParams);
-+
-+#if (DPAA_VERSION >= 11)
-+/**************************************************************************//**
-+ @Function FM_PCD_FrmReplicSetGroup
-+
-+ @Description Initialize a Frame Replicator group.
-+
-+ @Param[in] h_FmPcd FM PCD module descriptor.
-+ @Param[in] p_FrmReplicGroupParam A structure of parameters for the initialization of
-+ the frame replicator group.
-+
-+ @Return A handle to the initialized object on success; NULL code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_Init().
-+*//***************************************************************************/
-+t_Handle FM_PCD_FrmReplicSetGroup(t_Handle h_FmPcd, t_FmPcdFrmReplicGroupParams *p_FrmReplicGroupParam);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_FrmReplicDeleteGroup
-+
-+ @Description Delete a Frame Replicator group.
-+
-+ @Param[in] h_FrmReplicGroup A handle to the frame replicator group.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_FrmReplicSetGroup().
-+*//***************************************************************************/
-+t_Error FM_PCD_FrmReplicDeleteGroup(t_Handle h_FrmReplicGroup);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_FrmReplicAddMember
-+
-+ @Description Add the member in the index defined by the memberIndex.
-+
-+ @Param[in] h_FrmReplicGroup A handle to the frame replicator group.
-+ @Param[in] memberIndex member index for adding.
-+ @Param[in] p_MemberParams A pointer to the new member parameters.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_FrmReplicSetGroup() of this group.
-+*//***************************************************************************/
-+t_Error FM_PCD_FrmReplicAddMember(t_Handle h_FrmReplicGroup,
-+ uint16_t memberIndex,
-+ t_FmPcdCcNextEngineParams *p_MemberParams);
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_FrmReplicRemoveMember
-+
-+ @Description Remove the member defined by the index from the relevant group.
-+
-+ @Param[in] h_FrmReplicGroup A handle to the frame replicator group.
-+ @Param[in] memberIndex member index for removing.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_FrmReplicSetGroup() of this group.
-+*//***************************************************************************/
-+t_Error FM_PCD_FrmReplicRemoveMember(t_Handle h_FrmReplicGroup,
-+ uint16_t memberIndex);
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+#if ((DPAA_VERSION == 10) && defined(FM_CAPWAP_SUPPORT))
-+/**************************************************************************//**
-+ @Function FM_PCD_StatisticsSetNode
-+
-+ @Description This routine should be called for defining a statistics node.
-+
-+ @Param[in] h_FmPcd FM PCD module descriptor.
-+ @Param[in] p_FmPcdstatsParams A structure of parameters defining the statistics
-+
-+ @Return A handle to the initialized object on success; NULL code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_Init().
-+*//***************************************************************************/
-+t_Handle FM_PCD_StatisticsSetNode(t_Handle h_FmPcd, t_FmPcdStatsParams *p_FmPcdstatsParams);
-+#endif /* ((DPAA_VERSION == 10) && defined(FM_CAPWAP_SUPPORT)) */
-+
-+/** @} */ /* end of FM_PCD_Runtime_build_grp group */
-+/** @} */ /* end of FM_PCD_Runtime_grp group */
-+/** @} */ /* end of FM_PCD_grp group */
-+/** @} */ /* end of FM_grp group */
-+
-+
-+#ifdef NCSW_BACKWARD_COMPATIBLE_API
-+#define FM_PCD_MAX_NUM_OF_INTERCHANGABLE_HDRS FM_PCD_MAX_NUM_OF_INTERCHANGEABLE_HDRS
-+#define e_FM_PCD_MANIP_ONE_WAYS_HASH e_FM_PCD_MANIP_ONE_WAY_HASH
-+#define e_FM_PCD_MANIP_TOW_WAYS_HASH e_FM_PCD_MANIP_TWO_WAYS_HASH
-+
-+#define e_FM_PCD_MANIP_FRAGMENT_PACKECT e_FM_PCD_MANIP_FRAGMENT_PACKET /* Feb13 */
-+
-+#define FM_PCD_SetNetEnvCharacteristics(_pcd, _params) \
-+ FM_PCD_NetEnvCharacteristicsSet(_pcd, _params)
-+#define FM_PCD_KgSetScheme(_pcd, _params) FM_PCD_KgSchemeSet(_pcd, _params)
-+#define FM_PCD_CcBuildTree(_pcd, _params) FM_PCD_CcRootBuild(_pcd, _params)
-+#define FM_PCD_CcSetNode(_pcd, _params) FM_PCD_MatchTableSet(_pcd, _params)
-+#define FM_PCD_PlcrSetProfile(_pcd, _params) FM_PCD_PlcrProfileSet(_pcd, _params)
-+#define FM_PCD_ManipSetNode(_pcd, _params) FM_PCD_ManipNodeSet(_pcd, _params)
-+
-+#define FM_PCD_DeleteNetEnvCharacteristics(_pcd, ...) \
-+ FM_PCD_NetEnvCharacteristicsDelete(__VA_ARGS__)
-+#define FM_PCD_KgDeleteScheme(_pcd, ...) \
-+ FM_PCD_KgSchemeDelete(__VA_ARGS__)
-+#define FM_PCD_KgGetSchemeCounter(_pcd, ...) \
-+ FM_PCD_KgSchemeGetCounter(__VA_ARGS__)
-+#define FM_PCD_KgSetSchemeCounter(_pcd, ...) \
-+ FM_PCD_KgSchemeSetCounter(__VA_ARGS__)
-+#define FM_PCD_PlcrDeleteProfile(_pcd, ...) \
-+ FM_PCD_PlcrProfileDelete(__VA_ARGS__)
-+#define FM_PCD_PlcrGetProfileCounter(_pcd, ...) \
-+ FM_PCD_PlcrProfileGetCounter(__VA_ARGS__)
-+#define FM_PCD_PlcrSetProfileCounter(_pcd, ...) \
-+ FM_PCD_PlcrProfileSetCounter(__VA_ARGS__)
-+#define FM_PCD_CcDeleteTree(_pcd, ...) \
-+ FM_PCD_CcRootDelete(__VA_ARGS__)
-+#define FM_PCD_CcTreeModifyNextEngine(_pcd, ...) \
-+ FM_PCD_CcRootModifyNextEngine(__VA_ARGS__)
-+#define FM_PCD_CcDeleteNode(_pcd, ...) \
-+ FM_PCD_MatchTableDelete(__VA_ARGS__)
-+#define FM_PCD_CcNodeModifyMissNextEngine(_pcd, ...) \
-+ FM_PCD_MatchTableModifyMissNextEngine(__VA_ARGS__)
-+#define FM_PCD_CcNodeRemoveKey(_pcd, ...) \
-+ FM_PCD_MatchTableRemoveKey(__VA_ARGS__)
-+#define FM_PCD_CcNodeAddKey(_pcd, ...) \
-+ FM_PCD_MatchTableAddKey(__VA_ARGS__)
-+#define FM_PCD_CcNodeModifyNextEngine(_pcd, ...) \
-+ FM_PCD_MatchTableModifyNextEngine(__VA_ARGS__)
-+#define FM_PCD_CcNodeModifyKeyAndNextEngine(_pcd, ...) \
-+ FM_PCD_MatchTableModifyKeyAndNextEngine(__VA_ARGS__)
-+#define FM_PCD_CcNodeModifyKey(_pcd, ...) \
-+ FM_PCD_MatchTableModifyKey(__VA_ARGS__)
-+#define FM_PCD_CcNodeFindNRemoveKey(_pcd, ...) \
-+ FM_PCD_MatchTableFindNRemoveKey(__VA_ARGS__)
-+#define FM_PCD_CcNodeFindNModifyNextEngine(_pcd, ...) \
-+ FM_PCD_MatchTableFindNModifyNextEngine(__VA_ARGS__)
-+#define FM_PCD_CcNodeFindNModifyKeyAndNextEngine(_pcd, ...) \
-+ FM_PCD_MatchTableFindNModifyKeyAndNextEngine(__VA_ARGS__)
-+#define FM_PCD_CcNodeFindNModifyKey(_pcd, ...) \
-+ FM_PCD_MatchTableFindNModifyKey(__VA_ARGS__)
-+#define FM_PCD_CcIndexedHashNodeGetBucket(_pcd, ...) \
-+ FM_PCD_MatchTableGetIndexedHashBucket(__VA_ARGS__)
-+#define FM_PCD_CcNodeGetNextEngine(_pcd, ...) \
-+ FM_PCD_MatchTableGetNextEngine(__VA_ARGS__)
-+#define FM_PCD_CcNodeGetKeyCounter(_pcd, ...) \
-+ FM_PCD_MatchTableGetKeyCounter(__VA_ARGS__)
-+#define FM_PCD_ManipDeleteNode(_pcd, ...) \
-+ FM_PCD_ManipNodeDelete(__VA_ARGS__)
-+#endif /* NCSW_BACKWARD_COMPATIBLE_API */
-+
-+
-+#endif /* __FM_PCD_EXT */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/Peripherals/fm_port_ext.h
-@@ -0,0 +1,2608 @@
-+/* Copyright (c) 2008-2012 Freescale Semiconductor, Inc.
-+ * All rights reserved.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/**************************************************************************//**
-+ @File fm_port_ext.h
-+
-+ @Description FM-Port Application Programming Interface.
-+*//***************************************************************************/
-+#ifndef __FM_PORT_EXT
-+#define __FM_PORT_EXT
-+
-+#include "error_ext.h"
-+#include "std_ext.h"
-+#include "fm_pcd_ext.h"
-+#include "fm_ext.h"
-+#include "net_ext.h"
-+
-+
-+/**************************************************************************//**
-+
-+ @Group FM_grp Frame Manager API
-+
-+ @Description FM API functions, definitions and enums
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Group FM_PORT_grp FM Port
-+
-+ @Description FM Port API
-+
-+ The FM uses a general module called "port" to represent a Tx port
-+ (MAC), an Rx port (MAC) or Offline Parsing port.
-+ The number of ports in an FM varies between SOCs.
-+ The SW driver manages these ports as sub-modules of the FM, i.e.
-+ after an FM is initialized, its ports may be initialized and
-+ operated upon.
-+
-+ The port is initialized aware of its type, but other functions on
-+ a port may be indifferent to its type. When necessary, the driver
-+ verifies coherence and returns error if applicable.
-+
-+ On initialization, user specifies the port type and it's index
-+ (relative to the port's type) - always starting at 0.
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Description An enum for defining port PCD modes.
-+ This enum defines the superset of PCD engines support - i.e. not
-+ all engines have to be used, but all have to be enabled. The real
-+ flow of a specific frame depends on the PCD configuration and the
-+ frame headers and payload.
-+ Note: the first engine and the first engine after the parser (if
-+ exists) should be in order, the order is important as it will
-+ define the flow of the port. However, as for the rest engines
-+ (the ones that follows), the order is not important anymore as
-+ it is defined by the PCD graph itself.
-+*//***************************************************************************/
-+typedef enum e_FmPortPcdSupport {
-+ e_FM_PORT_PCD_SUPPORT_NONE = 0 /**< BMI to BMI, PCD is not used */
-+ , e_FM_PORT_PCD_SUPPORT_PRS_ONLY /**< Use only Parser */
-+ , e_FM_PORT_PCD_SUPPORT_PLCR_ONLY /**< Use only Policer */
-+ , e_FM_PORT_PCD_SUPPORT_PRS_AND_PLCR /**< Use Parser and Policer */
-+ , e_FM_PORT_PCD_SUPPORT_PRS_AND_KG /**< Use Parser and Keygen */
-+ , e_FM_PORT_PCD_SUPPORT_PRS_AND_KG_AND_CC /**< Use Parser, Keygen and Coarse Classification */
-+ , e_FM_PORT_PCD_SUPPORT_PRS_AND_KG_AND_CC_AND_PLCR
-+ /**< Use all PCD engines */
-+ , e_FM_PORT_PCD_SUPPORT_PRS_AND_KG_AND_PLCR /**< Use Parser, Keygen and Policer */
-+ , e_FM_PORT_PCD_SUPPORT_PRS_AND_CC /**< Use Parser and Coarse Classification */
-+ , e_FM_PORT_PCD_SUPPORT_PRS_AND_CC_AND_PLCR /**< Use Parser and Coarse Classification and Policer */
-+ , e_FM_PORT_PCD_SUPPORT_CC_ONLY /**< Use only Coarse Classification */
-+#ifdef FM_CAPWAP_SUPPORT
-+ , e_FM_PORT_PCD_SUPPORT_CC_AND_KG /**< Use Coarse Classification,and Keygen */
-+ , e_FM_PORT_PCD_SUPPORT_CC_AND_KG_AND_PLCR /**< Use Coarse Classification, Keygen and Policer */
-+#endif /* FM_CAPWAP_SUPPORT */
-+} e_FmPortPcdSupport;
-+
-+/**************************************************************************//**
-+ @Description Port interrupts
-+*//***************************************************************************/
-+typedef enum e_FmPortExceptions {
-+ e_FM_PORT_EXCEPTION_IM_BUSY /**< Independent-Mode Rx-BUSY */
-+} e_FmPortExceptions;
-+
-+
-+/**************************************************************************//**
-+ @Collection General FM Port defines
-+*//***************************************************************************/
-+#define FM_PORT_PRS_RESULT_NUM_OF_WORDS 8 /**< Number of 4 bytes words in parser result */
-+/* @} */
-+
-+/**************************************************************************//**
-+ @Collection FM Frame error
-+*//***************************************************************************/
-+typedef uint32_t fmPortFrameErrSelect_t; /**< typedef for defining Frame Descriptor errors */
-+
-+#define FM_PORT_FRM_ERR_UNSUPPORTED_FORMAT FM_FD_ERR_UNSUPPORTED_FORMAT /**< Not for Rx-Port! Unsupported Format */
-+#define FM_PORT_FRM_ERR_LENGTH FM_FD_ERR_LENGTH /**< Not for Rx-Port! Length Error */
-+#define FM_PORT_FRM_ERR_DMA FM_FD_ERR_DMA /**< DMA Data error */
-+#define FM_PORT_FRM_ERR_NON_FM FM_FD_RX_STATUS_ERR_NON_FM /**< non Frame-Manager error; probably come from SEC that
-+ was chained to FM */
-+
-+#define FM_PORT_FRM_ERR_IPRE (FM_FD_ERR_IPR & ~FM_FD_IPR) /**< IPR error */
-+#define FM_PORT_FRM_ERR_IPR_NCSP (FM_FD_ERR_IPR_NCSP & ~FM_FD_IPR) /**< IPR non-consistent-sp */
-+
-+#define FM_PORT_FRM_ERR_IPFE 0 /**< Obsolete; will be removed in the future */
-+
-+#ifdef FM_CAPWAP_SUPPORT
-+#define FM_PORT_FRM_ERR_CRE FM_FD_ERR_CRE
-+#define FM_PORT_FRM_ERR_CHE FM_FD_ERR_CHE
-+#endif /* FM_CAPWAP_SUPPORT */
-+
-+#define FM_PORT_FRM_ERR_PHYSICAL FM_FD_ERR_PHYSICAL /**< Rx FIFO overflow, FCS error, code error, running disparity
-+ error (SGMII and TBI modes), FIFO parity error. PHY
-+ Sequence error, PHY error control character detected. */
-+#define FM_PORT_FRM_ERR_SIZE FM_FD_ERR_SIZE /**< Frame too long OR Frame size exceeds max_length_frame */
-+#define FM_PORT_FRM_ERR_CLS_DISCARD FM_FD_ERR_CLS_DISCARD /**< indicates a classifier "drop" operation */
-+#define FM_PORT_FRM_ERR_EXTRACTION FM_FD_ERR_EXTRACTION /**< Extract Out of Frame */
-+#define FM_PORT_FRM_ERR_NO_SCHEME FM_FD_ERR_NO_SCHEME /**< No Scheme Selected */
-+#define FM_PORT_FRM_ERR_KEYSIZE_OVERFLOW FM_FD_ERR_KEYSIZE_OVERFLOW /**< Keysize Overflow */
-+#define FM_PORT_FRM_ERR_COLOR_RED FM_FD_ERR_COLOR_RED /**< Frame color is red */
-+#define FM_PORT_FRM_ERR_COLOR_YELLOW FM_FD_ERR_COLOR_YELLOW /**< Frame color is yellow */
-+#define FM_PORT_FRM_ERR_ILL_PLCR FM_FD_ERR_ILL_PLCR /**< Illegal Policer Profile selected */
-+#define FM_PORT_FRM_ERR_PLCR_FRAME_LEN FM_FD_ERR_PLCR_FRAME_LEN /**< Policer frame length error */
-+#define FM_PORT_FRM_ERR_PRS_TIMEOUT FM_FD_ERR_PRS_TIMEOUT /**< Parser Time out Exceed */
-+#define FM_PORT_FRM_ERR_PRS_ILL_INSTRUCT FM_FD_ERR_PRS_ILL_INSTRUCT /**< Invalid Soft Parser instruction */
-+#define FM_PORT_FRM_ERR_PRS_HDR_ERR FM_FD_ERR_PRS_HDR_ERR /**< Header error was identified during parsing */
-+#define FM_PORT_FRM_ERR_BLOCK_LIMIT_EXCEEDED FM_FD_ERR_BLOCK_LIMIT_EXCEEDED /**< Frame parsed beyind 256 first bytes */
-+#define FM_PORT_FRM_ERR_PROCESS_TIMEOUT 0x00000001 /**< FPM Frame Processing Timeout Exceeded */
-+/* @} */
-+
-+
-+
-+/**************************************************************************//**
-+ @Group FM_PORT_init_grp FM Port Initialization Unit
-+
-+ @Description FM Port Initialization Unit
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Description Exceptions user callback routine, will be called upon an
-+ exception passing the exception identification.
-+
-+ @Param[in] h_App - User's application descriptor.
-+ @Param[in] exception - The exception.
-+ *//***************************************************************************/
-+typedef void (t_FmPortExceptionCallback) (t_Handle h_App, e_FmPortExceptions exception);
-+
-+/**************************************************************************//**
-+ @Description User callback function called by driver with received data.
-+
-+ User provides this function. Driver invokes it.
-+
-+ @Param[in] h_App Application's handle originally specified to
-+ the API Config function
-+ @Param[in] p_Data A pointer to data received
-+ @Param[in] length length of received data
-+ @Param[in] status receive status and errors
-+ @Param[in] position position of buffer in frame
-+ @Param[in] h_BufContext A handle of the user acossiated with this buffer
-+
-+ @Retval e_RX_STORE_RESPONSE_CONTINUE - order the driver to continue Rx
-+ operation for all ready data.
-+ @Retval e_RX_STORE_RESPONSE_PAUSE - order the driver to stop Rx operation.
-+*//***************************************************************************/
-+typedef e_RxStoreResponse (t_FmPortImRxStoreCallback) (t_Handle h_App,
-+ uint8_t *p_Data,
-+ uint16_t length,
-+ uint16_t status,
-+ uint8_t position,
-+ t_Handle h_BufContext);
-+
-+/**************************************************************************//**
-+ @Description User callback function called by driver when transmit completed.
-+
-+ User provides this function. Driver invokes it.
-+
-+ @Param[in] h_App Application's handle originally specified to
-+ the API Config function
-+ @Param[in] p_Data A pointer to data received
-+ @Param[in] status transmit status and errors
-+ @Param[in] lastBuffer is last buffer in frame
-+ @Param[in] h_BufContext A handle of the user acossiated with this buffer
-+ *//***************************************************************************/
-+typedef void (t_FmPortImTxConfCallback) (t_Handle h_App,
-+ uint8_t *p_Data,
-+ uint16_t status,
-+ t_Handle h_BufContext);
-+
-+/**************************************************************************//**
-+ @Description A structure for additional Rx port parameters
-+*//***************************************************************************/
-+typedef struct t_FmPortRxParams {
-+ uint32_t errFqid; /**< Error Queue Id. */
-+ uint32_t dfltFqid; /**< Default Queue Id. */
-+ uint16_t liodnOffset; /**< Port's LIODN offset. */
-+ t_FmExtPools extBufPools; /**< Which external buffer pools are used
-+ (up to FM_PORT_MAX_NUM_OF_EXT_POOLS), and their sizes. */
-+} t_FmPortRxParams;
-+
-+/**************************************************************************//**
-+ @Description A structure for additional non-Rx port parameters
-+*//***************************************************************************/
-+typedef struct t_FmPortNonRxParams {
-+ uint32_t errFqid; /**< Error Queue Id. */
-+ uint32_t dfltFqid; /**< For Tx - Default Confirmation queue,
-+ 0 means no Tx confirmation for processed
-+ frames. For OP port - default Rx queue. */
-+ uint32_t qmChannel; /**< QM-channel dedicated to this port; will be used
-+ by the FM for dequeue. */
-+} t_FmPortNonRxParams;
-+
-+/**************************************************************************//**
-+ @Description A structure for additional Rx port parameters
-+*//***************************************************************************/
-+typedef struct t_FmPortImRxTxParams {
-+ t_Handle h_FmMuram; /**< A handle of the FM-MURAM partition */
-+ uint16_t liodnOffset; /**< For Rx ports only. Port's LIODN Offset. */
-+ uint8_t dataMemId; /**< Memory partition ID for data buffers */
-+ uint32_t dataMemAttributes; /**< Memory attributes for data buffers */
-+ t_BufferPoolInfo rxPoolParams; /**< For Rx ports only. */
-+ t_FmPortImRxStoreCallback *f_RxStore; /**< For Rx ports only. */
-+ t_FmPortImTxConfCallback *f_TxConf; /**< For Tx ports only. */
-+} t_FmPortImRxTxParams;
-+
-+/**************************************************************************//**
-+ @Description A union for additional parameters depending on port type
-+*//***************************************************************************/
-+typedef union u_FmPortSpecificParams {
-+ t_FmPortImRxTxParams imRxTxParams; /**< Rx/Tx Independent-Mode port parameter structure */
-+ t_FmPortRxParams rxParams; /**< Rx port parameters structure */
-+ t_FmPortNonRxParams nonRxParams; /**< Non-Rx port parameters structure */
-+} u_FmPortSpecificParams;
-+
-+/**************************************************************************//**
-+ @Description A structure representing FM initialization parameters
-+*//***************************************************************************/
-+typedef struct t_FmPortParams {
-+ uintptr_t baseAddr; /**< Virtual Address of memory mapped FM Port registers.*/
-+ t_Handle h_Fm; /**< A handle to the FM object this port related to */
-+ e_FmPortType portType; /**< Port type */
-+ uint8_t portId; /**< Port Id - relative to type;
-+ NOTE: When configuring Offline Parsing port for
-+ FMANv3 devices (DPAA_VERSION 11 and higher),
-+ it is highly recommended NOT to use portId=0 due to lack
-+ of HW resources on portId=0. */
-+ bool independentModeEnable;
-+ /**< This port is Independent-Mode - Used for Rx/Tx ports only! */
-+ uint16_t liodnBase; /**< Irrelevant for P4080 rev 1. LIODN base for this port, to be
-+ used together with LIODN offset. */
-+ u_FmPortSpecificParams specificParams; /**< Additional parameters depending on port
-+ type. */
-+
-+ t_FmPortExceptionCallback *f_Exception; /**< Relevant for IM only Callback routine to be called on BUSY exception */
-+ t_Handle h_App; /**< A handle to an application layer object; This handle will
-+ be passed by the driver upon calling the above callbacks */
-+} t_FmPortParams;
-+
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_Config
-+
-+ @Description Creates a descriptor for the FM PORT module.
-+
-+ The routine returns a handle (descriptor) to the FM PORT object.
-+ This descriptor must be passed as first parameter to all other
-+ FM PORT function calls.
-+
-+ No actual initialization or configuration of FM hardware is
-+ done by this routine.
-+
-+ @Param[in] p_FmPortParams - Pointer to data structure of parameters
-+
-+ @Retval Handle to FM object, or NULL for Failure.
-+*//***************************************************************************/
-+t_Handle FM_PORT_Config(t_FmPortParams *p_FmPortParams);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_Init
-+
-+ @Description Initializes the FM PORT module by defining the software structure
-+ and configuring the hardware registers.
-+
-+ @Param[in] h_FmPort - FM PORT module descriptor
-+
-+ @Return E_OK on success; Error code otherwise.
-+*//***************************************************************************/
-+t_Error FM_PORT_Init(t_Handle h_FmPort);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_Free
-+
-+ @Description Frees all resources that were assigned to FM PORT module.
-+
-+ Calling this routine invalidates the descriptor.
-+
-+ @Param[in] h_FmPort - FM PORT module descriptor
-+
-+ @Return E_OK on success; Error code otherwise.
-+*//***************************************************************************/
-+t_Error FM_PORT_Free(t_Handle h_FmPort);
-+
-+
-+/**************************************************************************//**
-+ @Group FM_PORT_advanced_init_grp FM Port Advanced Configuration Unit
-+
-+ @Description Configuration functions used to change default values.
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Description enum for defining QM frame dequeue
-+*//***************************************************************************/
-+typedef enum e_FmPortDeqType {
-+ e_FM_PORT_DEQ_TYPE1, /**< Dequeue from the SP channel - with priority precedence,
-+ and Intra-Class Scheduling respected. */
-+ e_FM_PORT_DEQ_TYPE2, /**< Dequeue from the SP channel - with active FQ precedence,
-+ and Intra-Class Scheduling respected. */
-+ e_FM_PORT_DEQ_TYPE3 /**< Dequeue from the SP channel - with active FQ precedence,
-+ and override Intra-Class Scheduling */
-+} e_FmPortDeqType;
-+
-+/**************************************************************************//**
-+ @Description enum for defining QM frame dequeue
-+*//***************************************************************************/
-+typedef enum e_FmPortDeqPrefetchOption {
-+ e_FM_PORT_DEQ_NO_PREFETCH, /**< QMI preforms a dequeue action for a single frame
-+ only when a dedicated portID Tnum is waiting. */
-+ e_FM_PORT_DEQ_PARTIAL_PREFETCH, /**< QMI preforms a dequeue action for 3 frames when
-+ one dedicated portId tnum is waiting. */
-+ e_FM_PORT_DEQ_FULL_PREFETCH /**< QMI preforms a dequeue action for 3 frames when
-+ no dedicated portId tnums are waiting. */
-+
-+} e_FmPortDeqPrefetchOption;
-+
-+/**************************************************************************//**
-+ @Description enum for defining port default color
-+*//***************************************************************************/
-+typedef enum e_FmPortColor {
-+ e_FM_PORT_COLOR_GREEN, /**< Default port color is green */
-+ e_FM_PORT_COLOR_YELLOW, /**< Default port color is yellow */
-+ e_FM_PORT_COLOR_RED, /**< Default port color is red */
-+ e_FM_PORT_COLOR_OVERRIDE /**< Ignore color */
-+} e_FmPortColor;
-+
-+/**************************************************************************//**
-+ @Description A structure for defining Dual Tx rate limiting scale
-+*//***************************************************************************/
-+typedef enum e_FmPortDualRateLimiterScaleDown {
-+ e_FM_PORT_DUAL_RATE_LIMITER_NONE = 0, /**< Use only single rate limiter */
-+ e_FM_PORT_DUAL_RATE_LIMITER_SCALE_DOWN_BY_2, /**< Divide high rate limiter by 2 */
-+ e_FM_PORT_DUAL_RATE_LIMITER_SCALE_DOWN_BY_4, /**< Divide high rate limiter by 4 */
-+ e_FM_PORT_DUAL_RATE_LIMITER_SCALE_DOWN_BY_8 /**< Divide high rate limiter by 8 */
-+} e_FmPortDualRateLimiterScaleDown;
-+
-+
-+/**************************************************************************//**
-+ @Description A structure for defining FM port resources
-+*//***************************************************************************/
-+typedef struct t_FmPortRsrc {
-+ uint32_t num; /**< Committed required resource */
-+ uint32_t extra; /**< Extra (not committed) required resource */
-+} t_FmPortRsrc;
-+
-+/**************************************************************************//**
-+ @Description A structure for defining observed pool depletion
-+*//***************************************************************************/
-+typedef struct t_FmPortObservedBufPoolDepletion {
-+ t_FmBufPoolDepletion poolDepletionParams;/**< parameters to define pool depletion */
-+ t_FmExtPools poolsParams; /**< Which external buffer pools are observed
-+ (up to FM_PORT_MAX_NUM_OF_OBSERVED_EXT_POOLS),
-+ and their sizes. */
-+} t_FmPortObservedBufPoolDepletion;
-+
-+/**************************************************************************//**
-+ @Description A structure for defining Tx rate limiting
-+*//***************************************************************************/
-+typedef struct t_FmPortRateLimit {
-+ uint16_t maxBurstSize; /**< in KBytes for Tx ports, in frames
-+ for OP ports. (note that
-+ for early chips burst size is
-+ rounded up to a multiply of 1000 frames).*/
-+ uint32_t rateLimit; /**< in Kb/sec for Tx ports, in frame/sec for
-+ OP ports. Rate limit refers to
-+ data rate (rather than line rate). */
-+ e_FmPortDualRateLimiterScaleDown rateLimitDivider; /**< For OP ports only. Not-valid
-+ for some earlier chip revisions */
-+} t_FmPortRateLimit;
-+
-+/**************************************************************************//**
-+ @Description A structure for defining the parameters of
-+ the Rx port performance counters
-+*//***************************************************************************/
-+typedef struct t_FmPortPerformanceCnt {
-+ uint8_t taskCompVal; /**< Task compare value */
-+ uint8_t queueCompVal; /**< Rx queue/Tx confirm queue compare
-+ value (unused for H/O) */
-+ uint8_t dmaCompVal; /**< Dma compare value */
-+ uint32_t fifoCompVal; /**< Fifo compare value (in bytes) */
-+} t_FmPortPerformanceCnt;
-+
-+
-+/**************************************************************************//**
-+ @Description A structure for defining the sizes of the Deep Sleep
-+ the Auto Response tables
-+*//***************************************************************************/
-+typedef struct t_FmPortDsarTablesSizes
-+{
-+ uint16_t maxNumOfArpEntries;
-+ uint16_t maxNumOfEchoIpv4Entries;
-+ uint16_t maxNumOfNdpEntries;
-+ uint16_t maxNumOfEchoIpv6Entries;
-+ uint16_t maxNumOfSnmpIPV4Entries;
-+ uint16_t maxNumOfSnmpIPV6Entries;
-+ uint16_t maxNumOfSnmpOidEntries;
-+ uint16_t maxNumOfSnmpOidChar; /* total amount of character needed for the snmp table */
-+
-+ uint16_t maxNumOfIpProtFiltering;
-+ uint16_t maxNumOfTcpPortFiltering;
-+ uint16_t maxNumOfUdpPortFiltering;
-+} t_FmPortDsarTablesSizes;
-+
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_ConfigDsarSupport
-+
-+ @Description This function will allocate the amount of MURAM needed for
-+ this max number of entries for Deep Sleep Auto Response.
-+ it will calculate all needed MURAM for autoresponse including
-+ necesary common stuff.
-+
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in] params A pointer to a structure containing the maximum
-+ sizes of the auto response tables
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Config() and before FM_PORT_Init().
-+*//***************************************************************************/
-+t_Error FM_PORT_ConfigDsarSupport(t_Handle h_FmPortRx, t_FmPortDsarTablesSizes *params);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_ConfigNumOfOpenDmas
-+
-+ @Description Calling this routine changes the max number of open DMA's
-+ available for this port. It changes this parameter in the
-+ internal driver data base from its default configuration
-+ [OP: 1]
-+ [1G-RX, 1G-TX: 1 (+1)]
-+ [10G-RX, 10G-TX: 8 (+8)]
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in] p_OpenDmas A pointer to a structure of parameters defining
-+ the open DMA allocation.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Config() and before FM_PORT_Init().
-+*//***************************************************************************/
-+t_Error FM_PORT_ConfigNumOfOpenDmas(t_Handle h_FmPort, t_FmPortRsrc *p_OpenDmas);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_ConfigNumOfTasks
-+
-+ @Description Calling this routine changes the max number of tasks
-+ available for this port. It changes this parameter in the
-+ internal driver data base from its default configuration
-+ [OP: 1]
-+ [1G-RX, 1G-TX: 3 (+2)]
-+ [10G-RX, 10G-TX: 16 (+8)]
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in] p_NumOfTasks A pointer to a structure of parameters defining
-+ the tasks allocation.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Config() and before FM_PORT_Init().
-+*//***************************************************************************/
-+t_Error FM_PORT_ConfigNumOfTasks(t_Handle h_FmPort, t_FmPortRsrc *p_NumOfTasks);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_ConfigSizeOfFifo
-+
-+ @Description Calling this routine changes the max FIFO size configured for this port.
-+
-+ This function changes the internal driver data base from its
-+ default configuration. Please refer to the driver's User Guide for
-+ information on default FIFO sizes in the various devices.
-+ [OP: 2KB]
-+ [1G-RX, 1G-TX: 11KB]
-+ [10G-RX, 10G-TX: 12KB]
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in] p_SizeOfFifo A pointer to a structure of parameters defining
-+ the FIFO allocation.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Config() and before FM_PORT_Init().
-+*//***************************************************************************/
-+t_Error FM_PORT_ConfigSizeOfFifo(t_Handle h_FmPort, t_FmPortRsrc *p_SizeOfFifo);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_ConfigDeqHighPriority
-+
-+ @Description Calling this routine changes the dequeue priority in the
-+ internal driver data base from its default configuration
-+ 1G: [DEFAULT_PORT_deqHighPriority_1G]
-+ 10G: [DEFAULT_PORT_deqHighPriority_10G]
-+
-+ May be used for Non-Rx ports only
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in] highPri TRUE to select high priority, FALSE for normal operation.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Config() and before FM_PORT_Init().
-+*//***************************************************************************/
-+t_Error FM_PORT_ConfigDeqHighPriority(t_Handle h_FmPort, bool highPri);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_ConfigDeqType
-+
-+ @Description Calling this routine changes the dequeue type parameter in the
-+ internal driver data base from its default configuration
-+ [DEFAULT_PORT_deqType].
-+
-+ May be used for Non-Rx ports only
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in] deqType According to QM definition.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Config() and before FM_PORT_Init().
-+*//***************************************************************************/
-+t_Error FM_PORT_ConfigDeqType(t_Handle h_FmPort, e_FmPortDeqType deqType);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_ConfigDeqPrefetchOption
-+
-+ @Description Calling this routine changes the dequeue prefetch option parameter in the
-+ internal driver data base from its default configuration
-+ [DEFAULT_PORT_deqPrefetchOption]
-+ Note: Available for some chips only
-+
-+ May be used for Non-Rx ports only
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in] deqPrefetchOption New option
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Config() and before FM_PORT_Init().
-+*//***************************************************************************/
-+t_Error FM_PORT_ConfigDeqPrefetchOption(t_Handle h_FmPort, e_FmPortDeqPrefetchOption deqPrefetchOption);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_ConfigDeqByteCnt
-+
-+ @Description Calling this routine changes the dequeue byte count parameter in
-+ the internal driver data base from its default configuration
-+ 1G:[DEFAULT_PORT_deqByteCnt_1G].
-+ 10G:[DEFAULT_PORT_deqByteCnt_10G].
-+
-+ May be used for Non-Rx ports only
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in] deqByteCnt New byte count
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Config() and before FM_PORT_Init().
-+*//***************************************************************************/
-+t_Error FM_PORT_ConfigDeqByteCnt(t_Handle h_FmPort, uint16_t deqByteCnt);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_ConfigBufferPrefixContent
-+
-+ @Description Defines the structure, size and content of the application buffer.
-+ The prefix will
-+ In Tx ports, if 'passPrsResult', the application
-+ should set a value to their offsets in the prefix of
-+ the FM will save the first 'privDataSize', than,
-+ depending on 'passPrsResult' and 'passTimeStamp', copy parse result
-+ and timeStamp, and the packet itself (in this order), to the
-+ application buffer, and to offset.
-+ Calling this routine changes the buffer margins definitions
-+ in the internal driver data base from its default
-+ configuration: Data size: [DEFAULT_PORT_bufferPrefixContent_privDataSize]
-+ Pass Parser result: [DEFAULT_PORT_bufferPrefixContent_passPrsResult].
-+ Pass timestamp: [DEFAULT_PORT_bufferPrefixContent_passTimeStamp].
-+
-+ May be used for all ports
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in,out] p_FmBufferPrefixContent A structure of parameters describing the
-+ structure of the buffer.
-+ Out parameter: Start margin - offset
-+ of data from start of external buffer.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Config() and before FM_PORT_Init().
-+*//***************************************************************************/
-+t_Error FM_PORT_ConfigBufferPrefixContent(t_Handle h_FmPort,
-+ t_FmBufferPrefixContent *p_FmBufferPrefixContent);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_ConfigCheksumLastBytesIgnore
-+
-+ @Description Calling this routine changes the number of checksum bytes to ignore
-+ parameter in the internal driver data base from its default configuration
-+ [DEFAULT_PORT_cheksumLastBytesIgnore]
-+
-+ May be used by Tx & Rx ports only
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in] cheksumLastBytesIgnore New value
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Config() and before FM_PORT_Init().
-+*//***************************************************************************/
-+t_Error FM_PORT_ConfigCheksumLastBytesIgnore(t_Handle h_FmPort, uint8_t cheksumLastBytesIgnore);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_ConfigCutBytesFromEnd
-+
-+ @Description Calling this routine changes the number of bytes to cut from a
-+ frame's end parameter in the internal driver data base
-+ from its default configuration [DEFAULT_PORT_cutBytesFromEnd]
-+ Note that if the result of (frame length before chop - cutBytesFromEnd) is
-+ less than 14 bytes, the chop operation is not executed.
-+
-+ May be used for Rx ports only
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in] cutBytesFromEnd New value
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Config() and before FM_PORT_Init().
-+*//***************************************************************************/
-+t_Error FM_PORT_ConfigCutBytesFromEnd(t_Handle h_FmPort, uint8_t cutBytesFromEnd);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_ConfigPoolDepletion
-+
-+ @Description Calling this routine enables pause frame generation depending on the
-+ depletion status of BM pools. It also defines the conditions to activate
-+ this functionality. By default, this functionality is disabled.
-+
-+ May be used for Rx ports only
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in] p_BufPoolDepletion A structure of pool depletion parameters
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Config() and before FM_PORT_Init().
-+*//***************************************************************************/
-+t_Error FM_PORT_ConfigPoolDepletion(t_Handle h_FmPort, t_FmBufPoolDepletion *p_BufPoolDepletion);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_ConfigObservedPoolDepletion
-+
-+ @Description Calling this routine enables a mechanism to stop port enqueue
-+ depending on the depletion status of selected BM pools.
-+ It also defines the conditions to activate
-+ this functionality. By default, this functionality is disabled.
-+
-+ Note: Available for some chips only
-+
-+ May be used for OP ports only
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in] p_FmPortObservedBufPoolDepletion A structure of parameters for pool depletion.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Config() and before FM_PORT_Init().
-+*//***************************************************************************/
-+t_Error FM_PORT_ConfigObservedPoolDepletion(t_Handle h_FmPort,
-+ t_FmPortObservedBufPoolDepletion *p_FmPortObservedBufPoolDepletion);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_ConfigExtBufPools
-+
-+ @Description This routine should be called for OP ports
-+ that internally use BM buffer pools. In such cases, e.g. for fragmentation and
-+ re-assembly, the FM needs new BM buffers. By calling this routine the user
-+ specifies the BM buffer pools that should be used.
-+
-+ Note: Available for some chips only
-+
-+ May be used for OP ports only
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in] p_FmExtPools A structure of parameters for the external pools.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Config() and before FM_PORT_Init().
-+*//***************************************************************************/
-+t_Error FM_PORT_ConfigExtBufPools(t_Handle h_FmPort, t_FmExtPools *p_FmExtPools);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_ConfigBackupPools
-+
-+ @Description Calling this routine allows the configuration of some of the BM pools
-+ defined for this port as backup pools.
-+ A pool configured to be a backup pool will be used only if all other
-+ enabled non-backup pools are depleted.
-+
-+ May be used for Rx ports only
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in] p_FmPortBackupBmPools An array of pool id's. All pools specified here will
-+ be defined as backup pools.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Config() and before FM_PORT_Init().
-+*//***************************************************************************/
-+t_Error FM_PORT_ConfigBackupPools(t_Handle h_FmPort, t_FmBackupBmPools *p_FmPortBackupBmPools);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_ConfigFrmDiscardOverride
-+
-+ @Description Calling this routine changes the error frames destination parameter
-+ in the internal driver data base from its default configuration:
-+ override = [DEFAULT_PORT_frmDiscardOverride]
-+
-+ May be used for Rx and OP ports only
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in] override TRUE to override discarding of error frames and
-+ enqueueing them to error queue.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Config() and before FM_PORT_Init().
-+*//***************************************************************************/
-+t_Error FM_PORT_ConfigFrmDiscardOverride(t_Handle h_FmPort, bool override);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_ConfigErrorsToDiscard
-+
-+ @Description Calling this routine changes the behaviour on error parameter
-+ in the internal driver data base from its default configuration:
-+ [DEFAULT_PORT_errorsToDiscard].
-+ If a requested error was previously defined as "ErrorsToEnqueue" it's
-+ definition will change and the frame will be discarded.
-+ Errors that were not defined either as "ErrorsToEnqueue" nor as
-+ "ErrorsToDiscard", will be forwarded to CPU.
-+
-+ May be used for Rx and OP ports only
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in] errs A list of errors to discard
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Config() and before FM_PORT_Init().
-+*//***************************************************************************/
-+t_Error FM_PORT_ConfigErrorsToDiscard(t_Handle h_FmPort, fmPortFrameErrSelect_t errs);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_ConfigDmaSwapData
-+
-+ @Description Calling this routine changes the DMA swap data aparameter
-+ in the internal driver data base from its default
-+ configuration [DEFAULT_PORT_dmaSwapData]
-+
-+ May be used for all port types
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in] swapData New selection
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Config() and before FM_PORT_Init().
-+*//***************************************************************************/
-+t_Error FM_PORT_ConfigDmaSwapData(t_Handle h_FmPort, e_FmDmaSwapOption swapData);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_ConfigDmaIcCacheAttr
-+
-+ @Description Calling this routine changes the internal context cache
-+ attribute parameter in the internal driver data base
-+ from its default configuration [DEFAULT_PORT_dmaIntContextCacheAttr]
-+
-+ May be used for all port types
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in] intContextCacheAttr New selection
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Config() and before FM_PORT_Init().
-+*//***************************************************************************/
-+t_Error FM_PORT_ConfigDmaIcCacheAttr(t_Handle h_FmPort, e_FmDmaCacheOption intContextCacheAttr);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_ConfigDmaHdrAttr
-+
-+ @Description Calling this routine changes the header cache
-+ attribute parameter in the internal driver data base
-+ from its default configuration [DEFAULT_PORT_dmaHeaderCacheAttr]
-+
-+ May be used for all port types
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in] headerCacheAttr New selection
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Config() and before FM_PORT_Init().
-+*//***************************************************************************/
-+t_Error FM_PORT_ConfigDmaHdrAttr(t_Handle h_FmPort, e_FmDmaCacheOption headerCacheAttr);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_ConfigDmaScatterGatherAttr
-+
-+ @Description Calling this routine changes the scatter gather cache
-+ attribute parameter in the internal driver data base
-+ from its default configuration [DEFAULT_PORT_dmaScatterGatherCacheAttr]
-+
-+ May be used for all port types
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in] scatterGatherCacheAttr New selection
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Config() and before FM_PORT_Init().
-+*//***************************************************************************/
-+t_Error FM_PORT_ConfigDmaScatterGatherAttr(t_Handle h_FmPort, e_FmDmaCacheOption scatterGatherCacheAttr);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_ConfigDmaWriteOptimize
-+
-+ @Description Calling this routine changes the write optimization
-+ parameter in the internal driver data base
-+ from its default configuration: By default optimize = [DEFAULT_PORT_dmaWriteOptimize].
-+ Note:
-+
-+ 1. For head optimization, data alignment must be >= 16 (supported by default).
-+
-+ 3. For tail optimization, note that the optimization is performed by extending the write transaction
-+ of the frame payload at the tail as needed to achieve optimal bus transfers, so that the last write
-+ is extended to be on 16/64 bytes aligned block (chip dependent).
-+
-+ Relevant for non-Tx port types
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in] optimize TRUE to enable optimization, FALSE for normal operation
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Config() and before FM_PORT_Init().
-+*//***************************************************************************/
-+t_Error FM_PORT_ConfigDmaWriteOptimize(t_Handle h_FmPort, bool optimize);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_ConfigNoScatherGather
-+
-+ @Description Calling this routine changes the noScatherGather parameter in internal driver data base
-+ from its default configuration.
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in] noScatherGather (TRUE - frame is discarded if can not be stored in single buffer,
-+ FALSE - frame can be stored in scatter gather (S/G) format).
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Config() and before FM_PORT_Init().
-+*//***************************************************************************/
-+t_Error FM_PORT_ConfigNoScatherGather(t_Handle h_FmPort, bool noScatherGather);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_ConfigDfltColor
-+
-+ @Description Calling this routine changes the internal default color parameter
-+ in the internal driver data base
-+ from its default configuration [DEFAULT_PORT_color]
-+
-+ May be used for all port types
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in] color New selection
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Config() and before FM_PORT_Init().
-+*//***************************************************************************/
-+t_Error FM_PORT_ConfigDfltColor(t_Handle h_FmPort, e_FmPortColor color);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_ConfigSyncReq
-+
-+ @Description Calling this routine changes the synchronization attribute parameter
-+ in the internal driver data base from its default configuration:
-+ syncReq = [DEFAULT_PORT_syncReq]
-+
-+ May be used for all port types
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in] syncReq TRUE to request synchronization, FALSE otherwize.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Config() and before FM_PORT_Init().
-+*//***************************************************************************/
-+t_Error FM_PORT_ConfigSyncReq(t_Handle h_FmPort, bool syncReq);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_ConfigForwardReuseIntContext
-+
-+ @Description This routine is relevant for Rx ports that are routed to OP port.
-+ It changes the internal context reuse option in the internal
-+ driver data base from its default configuration:
-+ reuse = [DEFAULT_PORT_forwardIntContextReuse]
-+
-+ May be used for Rx ports only
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in] reuse TRUE to reuse internal context on frames
-+ forwarded to OP port.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Config() and before FM_PORT_Init().
-+*//***************************************************************************/
-+t_Error FM_PORT_ConfigForwardReuseIntContext(t_Handle h_FmPort, bool reuse);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_ConfigDontReleaseTxBufToBM
-+
-+ @Description This routine should be called if no Tx confirmation
-+ is done, and yet buffers should not be released to the BM.
-+ Normally, buffers are returned using the Tx confirmation
-+ process. When Tx confirmation is not used (defFqid=0),
-+ buffers are typically released to the BM. This routine
-+ may be called to avoid this behavior and not release the
-+ buffers.
-+
-+ May be used for Tx ports only
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Config() and before FM_PORT_Init().
-+*//***************************************************************************/
-+t_Error FM_PORT_ConfigDontReleaseTxBufToBM(t_Handle h_FmPort);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_ConfigIMMaxRxBufLength
-+
-+ @Description Changes the maximum receive buffer length from its default
-+ configuration: Closest rounded down power of 2 value of the
-+ data buffer size.
-+
-+ The maximum receive buffer length directly affects the structure
-+ of received frames (single- or multi-buffered) and the performance
-+ of both the FM and the driver.
-+
-+ The selection between single- or multi-buffered frames should be
-+ done according to the characteristics of the specific application.
-+ The recommended mode is to use a single data buffer per packet,
-+ as this mode provides the best performance. However, the user can
-+ select to use multiple data buffers per packet.
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in] newVal Maximum receive buffer length (in bytes).
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Config() and before FM_PORT_Init().
-+ This routine is to be used only if Independent-Mode is enabled.
-+*//***************************************************************************/
-+t_Error FM_PORT_ConfigIMMaxRxBufLength(t_Handle h_FmPort, uint16_t newVal);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_ConfigIMRxBdRingLength
-+
-+ @Description Changes the receive BD ring length from its default
-+ configuration:[DEFAULT_PORT_rxBdRingLength]
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in] newVal The desired BD ring length.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Config() and before FM_PORT_Init().
-+ This routine is to be used only if Independent-Mode is enabled.
-+*//***************************************************************************/
-+t_Error FM_PORT_ConfigIMRxBdRingLength(t_Handle h_FmPort, uint16_t newVal);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_ConfigIMTxBdRingLength
-+
-+ @Description Changes the transmit BD ring length from its default
-+ configuration:[DEFAULT_PORT_txBdRingLength]
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in] newVal The desired BD ring length.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Config() and before FM_PORT_Init().
-+ This routine is to be used only if Independent-Mode is enabled.
-+*//***************************************************************************/
-+t_Error FM_PORT_ConfigIMTxBdRingLength(t_Handle h_FmPort, uint16_t newVal);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_ConfigIMFmanCtrlExternalStructsMemory
-+
-+ @Description Configures memory partition and attributes for FMan-Controller
-+ data structures (e.g. BD rings).
-+ Calling this routine changes the internal driver data base
-+ from its default configuration
-+ [DEFAULT_PORT_ImfwExtStructsMemId, DEFAULT_PORT_ImfwExtStructsMemAttr].
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in] memId Memory partition ID.
-+ @Param[in] memAttributes Memory attributes mask (a combination of MEMORY_ATTR_x flags).
-+
-+ @Return E_OK on success; Error code otherwise.
-+*//***************************************************************************/
-+t_Error FM_PORT_ConfigIMFmanCtrlExternalStructsMemory(t_Handle h_FmPort,
-+ uint8_t memId,
-+ uint32_t memAttributes);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_ConfigIMPolling
-+
-+ @Description Changes the Rx flow from interrupt driven (default) to polling.
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Config() and before FM_PORT_Init().
-+ This routine is to be used only if Independent-Mode is enabled.
-+*//***************************************************************************/
-+t_Error FM_PORT_ConfigIMPolling(t_Handle h_FmPort);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_ConfigMaxFrameLength
-+
-+ @Description Changes the definition of the max size of frame that should be
-+ transmitted/received on this port from its default value [DEFAULT_PORT_maxFrameLength].
-+ This parameter is used for confirmation of the minimum Fifo
-+ size calculations and only for Tx ports or ports working in
-+ independent mode. This should be larger than the maximum possible
-+ MTU that will be used for this port (i.e. its MAC).
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in] length Max size of frame
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Config() and before FM_PORT_Init().
-+ This routine is to be used only if Independent-Mode is enabled.
-+*//***************************************************************************/
-+t_Error FM_PORT_ConfigMaxFrameLength(t_Handle h_FmPort, uint16_t length);
-+
-+/**************************************************************************//*
-+ @Function FM_PORT_ConfigTxFifoMinFillLevel
-+
-+ @Description Calling this routine changes the fifo minimum
-+ fill level parameter in the internal driver data base
-+ from its default configuration [DEFAULT_PORT_txFifoMinFillLevel]
-+
-+ May be used for Tx ports only
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in] minFillLevel New value
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Config() and before FM_PORT_Init().
-+*//***************************************************************************/
-+t_Error FM_PORT_ConfigTxFifoMinFillLevel(t_Handle h_FmPort, uint32_t minFillLevel);
-+
-+/**************************************************************************//*
-+ @Function FM_PORT_ConfigFifoDeqPipelineDepth
-+
-+ @Description Calling this routine changes the fifo dequeue
-+ pipeline depth parameter in the internal driver data base
-+
-+ from its default configuration: 1G ports: [DEFAULT_PORT_fifoDeqPipelineDepth_1G],
-+ 10G port: [DEFAULT_PORT_fifoDeqPipelineDepth_10G],
-+ OP port: [DEFAULT_PORT_fifoDeqPipelineDepth_OH]
-+
-+ May be used for Tx/OP ports only
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in] deqPipelineDepth New value
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Config() and before FM_PORT_Init().
-+*//***************************************************************************/
-+t_Error FM_PORT_ConfigFifoDeqPipelineDepth(t_Handle h_FmPort, uint8_t deqPipelineDepth);
-+
-+/**************************************************************************//*
-+ @Function FM_PORT_ConfigTxFifoLowComfLevel
-+
-+ @Description Calling this routine changes the fifo low comfort level
-+ parameter in internal driver data base
-+ from its default configuration [DEFAULT_PORT_txFifoLowComfLevel]
-+
-+ May be used for Tx ports only
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in] fifoLowComfLevel New value
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Config() and before FM_PORT_Init().
-+*//***************************************************************************/
-+t_Error FM_PORT_ConfigTxFifoLowComfLevel(t_Handle h_FmPort, uint32_t fifoLowComfLevel);
-+
-+/**************************************************************************//*
-+ @Function FM_PORT_ConfigRxFifoThreshold
-+
-+ @Description Calling this routine changes the threshold of the FIFO
-+ fill level parameter in the internal driver data base
-+ from its default configuration [DEFAULT_PORT_rxFifoThreshold]
-+
-+ If the total number of buffers which are
-+ currently in use and associated with the
-+ specific RX port exceed this threshold, the
-+ BMI will signal the MAC to send a pause frame
-+ over the link.
-+
-+ May be used for Rx ports only
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in] fifoThreshold New value
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Config() and before FM_PORT_Init().
-+*//***************************************************************************/
-+t_Error FM_PORT_ConfigRxFifoThreshold(t_Handle h_FmPort, uint32_t fifoThreshold);
-+
-+/**************************************************************************//*
-+ @Function FM_PORT_ConfigRxFifoPriElevationLevel
-+
-+ @Description Calling this routine changes the priority elevation level
-+ parameter in the internal driver data base from its default
-+ configuration [DEFAULT_PORT_rxFifoPriElevationLevel]
-+
-+ If the total number of buffers which are currently in use and
-+ associated with the specific RX port exceed the amount specified
-+ in priElevationLevel, BMI will signal the main FM's DMA to
-+ elevate the FM priority on the system bus.
-+
-+ May be used for Rx ports only
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in] priElevationLevel New value
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Config() and before FM_PORT_Init().
-+*//***************************************************************************/
-+t_Error FM_PORT_ConfigRxFifoPriElevationLevel(t_Handle h_FmPort, uint32_t priElevationLevel);
-+
-+#ifdef FM_HEAVY_TRAFFIC_HANG_ERRATA_FMAN_A005669
-+/**************************************************************************//*
-+ @Function FM_PORT_ConfigBCBWorkaround
-+
-+ @Description Configures BCB errata workaround.
-+
-+ When BCB errata is applicable, the workaround is always
-+ performed by FM Controller. Thus, this functions doesn't
-+ actually enable errata workaround but rather allows driver
-+ to perform adjustments required due to errata workaround
-+ execution in FM controller.
-+
-+ Applying BCB workaround also configures FM_PORT_FRM_ERR_PHYSICAL
-+ errors to be discarded. Thus FM_PORT_FRM_ERR_PHYSICAL can't be
-+ set by FM_PORT_SetErrorsRoute() function.
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Config() and before FM_PORT_Init().
-+*//***************************************************************************/
-+t_Error FM_PORT_ConfigBCBWorkaround(t_Handle h_FmPort);
-+#endif /* FM_HEAVY_TRAFFIC_HANG_ERRATA_FMAN_A005669 */
-+
-+#if (DPAA_VERSION >= 11)
-+/**************************************************************************//*
-+ @Function FM_PORT_ConfigInternalBuffOffset
-+
-+ @Description Configures internal buffer offset.
-+
-+ May be used for Rx and OP ports only
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in] val New value
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Config() and before FM_PORT_Init().
-+*//***************************************************************************/
-+t_Error FM_PORT_ConfigInternalBuffOffset(t_Handle h_FmPort, uint8_t val);
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+/** @} */ /* end of FM_PORT_advanced_init_grp group */
-+/** @} */ /* end of FM_PORT_init_grp group */
-+
-+
-+/**************************************************************************//**
-+ @Group FM_PORT_runtime_control_grp FM Port Runtime Control Unit
-+
-+ @Description FM Port Runtime control unit API functions, definitions and enums.
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Description enum for defining FM Port counters
-+*//***************************************************************************/
-+typedef enum e_FmPortCounters {
-+ e_FM_PORT_COUNTERS_CYCLE, /**< BMI performance counter */
-+ e_FM_PORT_COUNTERS_TASK_UTIL, /**< BMI performance counter */
-+ e_FM_PORT_COUNTERS_QUEUE_UTIL, /**< BMI performance counter */
-+ e_FM_PORT_COUNTERS_DMA_UTIL, /**< BMI performance counter */
-+ e_FM_PORT_COUNTERS_FIFO_UTIL, /**< BMI performance counter */
-+ e_FM_PORT_COUNTERS_RX_PAUSE_ACTIVATION, /**< BMI Rx only performance counter */
-+ e_FM_PORT_COUNTERS_FRAME, /**< BMI statistics counter */
-+ e_FM_PORT_COUNTERS_DISCARD_FRAME, /**< BMI statistics counter */
-+ e_FM_PORT_COUNTERS_DEALLOC_BUF, /**< BMI deallocate buffer statistics counter */
-+ e_FM_PORT_COUNTERS_RX_BAD_FRAME, /**< BMI Rx only statistics counter */
-+ e_FM_PORT_COUNTERS_RX_LARGE_FRAME, /**< BMI Rx only statistics counter */
-+ e_FM_PORT_COUNTERS_RX_FILTER_FRAME, /**< BMI Rx & OP only statistics counter */
-+ e_FM_PORT_COUNTERS_RX_LIST_DMA_ERR, /**< BMI Rx, OP & HC only statistics counter */
-+ e_FM_PORT_COUNTERS_RX_OUT_OF_BUFFERS_DISCARD, /**< BMI Rx, OP & HC statistics counter */
-+ e_FM_PORT_COUNTERS_PREPARE_TO_ENQUEUE_COUNTER, /**< BMI Rx, OP & HC only statistics counter */
-+ e_FM_PORT_COUNTERS_WRED_DISCARD, /**< BMI OP & HC only statistics counter */
-+ e_FM_PORT_COUNTERS_LENGTH_ERR, /**< BMI non-Rx statistics counter */
-+ e_FM_PORT_COUNTERS_UNSUPPRTED_FORMAT, /**< BMI non-Rx statistics counter */
-+ e_FM_PORT_COUNTERS_DEQ_TOTAL, /**< QMI total QM dequeues counter */
-+ e_FM_PORT_COUNTERS_ENQ_TOTAL, /**< QMI total QM enqueues counter */
-+ e_FM_PORT_COUNTERS_DEQ_FROM_DEFAULT, /**< QMI counter */
-+ e_FM_PORT_COUNTERS_DEQ_CONFIRM /**< QMI counter */
-+} e_FmPortCounters;
-+
-+typedef struct t_FmPortBmiStats {
-+ uint32_t cntCycle;
-+ uint32_t cntTaskUtil;
-+ uint32_t cntQueueUtil;
-+ uint32_t cntDmaUtil;
-+ uint32_t cntFifoUtil;
-+ uint32_t cntRxPauseActivation;
-+ uint32_t cntFrame;
-+ uint32_t cntDiscardFrame;
-+ uint32_t cntDeallocBuf;
-+ uint32_t cntRxBadFrame;
-+ uint32_t cntRxLargeFrame;
-+ uint32_t cntRxFilterFrame;
-+ uint32_t cntRxListDmaErr;
-+ uint32_t cntRxOutOfBuffersDiscard;
-+ uint32_t cntWredDiscard;
-+ uint32_t cntLengthErr;
-+ uint32_t cntUnsupportedFormat;
-+} t_FmPortBmiStats;
-+
-+/**************************************************************************//**
-+ @Description Structure for Port id parameters.
-+ Fields commented 'IN' are passed by the port module to be used
-+ by the FM module.
-+ Fields commented 'OUT' will be filled by FM before returning to port.
-+*//***************************************************************************/
-+typedef struct t_FmPortCongestionGrps {
-+ uint16_t numOfCongestionGrpsToConsider; /**< The number of required CGs
-+ to define the size of the following array */
-+ uint8_t congestionGrpsToConsider[FM_PORT_NUM_OF_CONGESTION_GRPS];
-+ /**< An array of CG indexes;
-+ Note that the size of the array should be
-+ 'numOfCongestionGrpsToConsider'. */
-+#if (DPAA_VERSION >= 11)
-+ bool pfcPrioritiesEn[FM_PORT_NUM_OF_CONGESTION_GRPS][FM_MAX_NUM_OF_PFC_PRIORITIES];
-+ /**< a matrix that represents the map between the CG ids
-+ defined in 'congestionGrpsToConsider' to the priorties
-+ mapping array. */
-+#endif /* (DPAA_VERSION >= 11) */
-+} t_FmPortCongestionGrps;
-+
-+/**************************************************************************//**
-+ @Description Structure for Deep Sleep Auto Response ARP Entry
-+*//***************************************************************************/
-+typedef struct t_FmPortDsarArpEntry
-+{
-+ uint32_t ipAddress;
-+ uint8_t mac[6];
-+ bool isVlan;
-+ uint16_t vid;
-+} t_FmPortDsarArpEntry;
-+
-+/**************************************************************************//**
-+ @Description Structure for Deep Sleep Auto Response ARP info
-+*//***************************************************************************/
-+typedef struct t_FmPortDsarArpInfo
-+{
-+ uint8_t tableSize;
-+ t_FmPortDsarArpEntry *p_AutoResTable;
-+ bool enableConflictDetection; /* when TRUE Conflict Detection will be checked and wake the host if needed */
-+} t_FmPortDsarArpInfo;
-+
-+/**************************************************************************//**
-+ @Description Structure for Deep Sleep Auto Response NDP Entry
-+*//***************************************************************************/
-+typedef struct t_FmPortDsarNdpEntry
-+{
-+ uint32_t ipAddress[4];
-+ uint8_t mac[6];
-+ bool isVlan;
-+ uint16_t vid;
-+} t_FmPortDsarNdpEntry;
-+
-+/**************************************************************************//**
-+ @Description Structure for Deep Sleep Auto Response NDP info
-+*//***************************************************************************/
-+typedef struct t_FmPortDsarNdpInfo
-+{
-+ uint32_t multicastGroup;
-+
-+ uint8_t tableSizeAssigned;
-+ t_FmPortDsarNdpEntry *p_AutoResTableAssigned; /* This list refer to solicitation IP addresses.
-+ Note that all IP adresses must be from the same multicast group.
-+ This will be checked and if not operation will fail. */
-+ uint8_t tableSizeTmp;
-+ t_FmPortDsarNdpEntry *p_AutoResTableTmp; /* This list refer to temp IP addresses.
-+ Note that all temp IP adresses must be from the same multicast group.
-+ This will be checked and if not operation will fail. */
-+
-+ bool enableConflictDetection; /* when TRUE Conflict Detection will be checked and wake the host if needed */
-+
-+} t_FmPortDsarNdpInfo;
-+
-+/**************************************************************************//**
-+ @Description Structure for Deep Sleep Auto Response ICMPV4 info
-+*//***************************************************************************/
-+typedef struct t_FmPortDsarEchoIpv4Info
-+{
-+ uint8_t tableSize;
-+ t_FmPortDsarArpEntry *p_AutoResTable;
-+} t_FmPortDsarEchoIpv4Info;
-+
-+/**************************************************************************//**
-+ @Description Structure for Deep Sleep Auto Response ICMPV6 info
-+*//***************************************************************************/
-+typedef struct t_FmPortDsarEchoIpv6Info
-+{
-+ uint8_t tableSize;
-+ t_FmPortDsarNdpEntry *p_AutoResTable;
-+} t_FmPortDsarEchoIpv6Info;
-+
-+/**************************************************************************//**
-+@Description Deep Sleep Auto Response SNMP OIDs table entry
-+
-+*//***************************************************************************/
-+typedef struct {
-+ uint16_t oidSize;
-+ uint8_t *oidVal; /* only the oid string */
-+ uint16_t resSize;
-+ uint8_t *resVal; /* resVal will be the entire reply,
-+ i.e. "Type|Length|Value" */
-+} t_FmPortDsarOidsEntry;
-+
-+/**************************************************************************//**
-+ @Description Deep Sleep Auto Response SNMP IPv4 Addresses Table Entry
-+ Refer to the FMan Controller spec for more details.
-+*//***************************************************************************/
-+typedef struct
-+{
-+ uint32_t ipv4Addr; /*!< 32 bit IPv4 Address. */
-+ bool isVlan;
-+ uint16_t vid; /*!< 12 bits VLAN ID. The 4 left-most bits should be cleared */
-+ /*!< This field should be 0x0000 for an entry with no VLAN tag or a null VLAN ID. */
-+} t_FmPortDsarSnmpIpv4AddrTblEntry;
-+
-+/**************************************************************************//**
-+ @Description Deep Sleep Auto Response SNMP IPv6 Addresses Table Entry
-+ Refer to the FMan Controller spec for more details.
-+*//***************************************************************************/
-+typedef struct
-+{
-+ uint32_t ipv6Addr[4]; /*!< 4 * 32 bit IPv6 Address. */
-+ bool isVlan;
-+ uint16_t vid; /*!< 12 bits VLAN ID. The 4 left-most bits should be cleared */
-+ /*!< This field should be 0x0000 for an entry with no VLAN tag or a null VLAN ID. */
-+} t_FmPortDsarSnmpIpv6AddrTblEntry;
-+
-+/**************************************************************************//**
-+ @Description Deep Sleep Auto Response SNMP Descriptor
-+
-+*//***************************************************************************/
-+typedef struct
-+{
-+ uint16_t control; /**< Control bits [0-15]. */
-+ uint16_t maxSnmpMsgLength; /**< Maximal allowed SNMP message length. */
-+ uint16_t numOfIpv4Addresses; /**< Number of entries in IPv4 addresses table. */
-+ uint16_t numOfIpv6Addresses; /**< Number of entries in IPv6 addresses table. */
-+ t_FmPortDsarSnmpIpv4AddrTblEntry *p_Ipv4AddrTbl; /**< Pointer to IPv4 addresses table. */
-+ t_FmPortDsarSnmpIpv6AddrTblEntry *p_Ipv6AddrTbl; /**< Pointer to IPv6 addresses table. */
-+ uint8_t *p_RdOnlyCommunityStr; /**< Pointer to the Read Only Community String. */
-+ uint8_t *p_RdWrCommunityStr; /**< Pointer to the Read Write Community String. */
-+ t_FmPortDsarOidsEntry *p_OidsTbl; /**< Pointer to OIDs table. */
-+ uint32_t oidsTblSize; /**< Number of entries in OIDs table. */
-+} t_FmPortDsarSnmpInfo;
-+
-+/**************************************************************************//**
-+ @Description Structure for Deep Sleep Auto Response filtering Entry
-+*//***************************************************************************/
-+typedef struct t_FmPortDsarFilteringEntry
-+{
-+ uint16_t srcPort;
-+ uint16_t dstPort;
-+ uint16_t srcPortMask;
-+ uint16_t dstPortMask;
-+} t_FmPortDsarFilteringEntry;
-+
-+/**************************************************************************//**
-+ @Description Structure for Deep Sleep Auto Response filtering info
-+*//***************************************************************************/
-+typedef struct t_FmPortDsarFilteringInfo
-+{
-+ /* IP protocol filtering parameters */
-+ uint8_t ipProtTableSize;
-+ uint8_t *p_IpProtTablePtr;
-+ bool ipProtPassOnHit; /* when TRUE, miss in the table will cause the packet to be droped,
-+ hit will pass the packet to UDP/TCP filters if needed and if not
-+ to the classification tree. If the classification tree will pass
-+ the packet to a queue it will cause a wake interupt.
-+ When FALSE it the other way around. */
-+ /* UDP port filtering parameters */
-+ uint8_t udpPortsTableSize;
-+ t_FmPortDsarFilteringEntry *p_UdpPortsTablePtr;
-+ bool udpPortPassOnHit; /* when TRUE, miss in the table will cause the packet to be droped,
-+ hit will pass the packet to classification tree.
-+ If the classification tree will pass the packet to a queue it
-+ will cause a wake interupt.
-+ When FALSE it the other way around. */
-+ /* TCP port filtering parameters */
-+ uint16_t tcpFlagsMask;
-+ uint8_t tcpPortsTableSize;
-+ t_FmPortDsarFilteringEntry *p_TcpPortsTablePtr;
-+ bool tcpPortPassOnHit; /* when TRUE, miss in the table will cause the packet to be droped,
-+ hit will pass the packet to classification tree.
-+ If the classification tree will pass the packet to a queue it
-+ will cause a wake interupt.
-+ When FALSE it the other way around. */
-+} t_FmPortDsarFilteringInfo;
-+
-+/**************************************************************************//**
-+ @Description Structure for Deep Sleep Auto Response parameters
-+*//***************************************************************************/
-+typedef struct t_FmPortDsarParams
-+{
-+ t_Handle h_FmPortTx;
-+ t_FmPortDsarArpInfo *p_AutoResArpInfo;
-+ t_FmPortDsarEchoIpv4Info *p_AutoResEchoIpv4Info;
-+ t_FmPortDsarNdpInfo *p_AutoResNdpInfo;
-+ t_FmPortDsarEchoIpv6Info *p_AutoResEchoIpv6Info;
-+ t_FmPortDsarSnmpInfo *p_AutoResSnmpInfo;
-+ t_FmPortDsarFilteringInfo *p_AutoResFilteringInfo;
-+} t_FmPortDsarParams;
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_EnterDsar
-+
-+ @Description Enter Deep Sleep Auto Response mode.
-+ This function write the apropriate values to in the relevant
-+ tables in the MURAM.
-+
-+ @Param[in] h_FmPortRx - FM PORT module descriptor
-+ @Param[in] params - Auto Response parameters
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Init().
-+*//***************************************************************************/
-+t_Error FM_PORT_EnterDsar(t_Handle h_FmPortRx, t_FmPortDsarParams *params);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_EnterDsarFinal
-+
-+ @Description Enter Deep Sleep Auto Response mode.
-+ This function sets the Tx port in independent mode as needed
-+ and redirect the receive flow to go through the
-+ Dsar Fman-ctrl code
-+
-+ @Param[in] h_DsarRxPort - FM Rx PORT module descriptor
-+ @Param[in] h_DsarTxPort - FM Tx PORT module descriptor
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Init().
-+*//***************************************************************************/
-+t_Error FM_PORT_EnterDsarFinal(t_Handle h_DsarRxPort, t_Handle h_DsarTxPort);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_ExitDsar
-+
-+ @Description Exit Deep Sleep Auto Response mode.
-+ This function reverse the AR mode and put the ports back into
-+ their original wake mode
-+
-+ @Param[in] h_FmPortRx - FM PORT Rx module descriptor
-+ @Param[in] h_FmPortTx - FM PORT Tx module descriptor
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_EnterDsar().
-+*//***************************************************************************/
-+void FM_PORT_ExitDsar(t_Handle h_FmPortRx, t_Handle h_FmPortTx);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_IsInDsar
-+
-+ @Description This function returns TRUE if the port was set as Auto Response
-+ and FALSE if not. Once Exit AR mode it will return FALSE as well
-+ until re-enabled once more.
-+
-+ @Param[in] h_FmPort - FM PORT module descriptor
-+
-+ @Return E_OK on success; Error code otherwise.
-+*//***************************************************************************/
-+bool FM_PORT_IsInDsar(t_Handle h_FmPort);
-+
-+typedef struct t_FmPortDsarStats
-+{
-+ uint32_t arpArCnt;
-+ uint32_t echoIcmpv4ArCnt;
-+ uint32_t ndpArCnt;
-+ uint32_t echoIcmpv6ArCnt;
-+ uint32_t snmpGetCnt;
-+ uint32_t snmpGetNextCnt;
-+} t_FmPortDsarStats;
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_GetDsarStats
-+
-+ @Description Return statistics for Deep Sleep Auto Response
-+
-+ @Param[in] h_FmPortRx - FM PORT module descriptor
-+ @Param[out] stats - structure containing the statistics counters
-+
-+ @Return E_OK on success; Error code otherwise.
-+*//***************************************************************************/
-+t_Error FM_PORT_GetDsarStats(t_Handle h_FmPortRx, t_FmPortDsarStats *stats);
-+
-+#if (defined(DEBUG_ERRORS) && (DEBUG_ERRORS > 0))
-+/**************************************************************************//**
-+ @Function FM_PORT_DumpRegs
-+
-+ @Description Dump all regs.
-+
-+ Calling this routine invalidates the descriptor.
-+
-+ @Param[in] h_FmPort - FM PORT module descriptor
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Init().
-+*//***************************************************************************/
-+t_Error FM_PORT_DumpRegs(t_Handle h_FmPort);
-+#endif /* (defined(DEBUG_ERRORS) && ... */
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_GetBufferDataOffset
-+
-+ @Description Relevant for Rx ports.
-+ Returns the data offset from the beginning of the data buffer
-+
-+ @Param[in] h_FmPort - FM PORT module descriptor
-+
-+ @Return data offset.
-+
-+ @Cautions Allowed only following FM_PORT_Init().
-+*//***************************************************************************/
-+uint32_t FM_PORT_GetBufferDataOffset(t_Handle h_FmPort);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_GetBufferICInfo
-+
-+ @Description Returns the Internal Context offset from the beginning of the data buffer
-+
-+ @Param[in] h_FmPort - FM PORT module descriptor
-+ @Param[in] p_Data - A pointer to the data buffer.
-+
-+ @Return Internal context info pointer on success, NULL if 'allOtherInfo' was not
-+ configured for this port.
-+
-+ @Cautions Allowed only following FM_PORT_Init().
-+*//***************************************************************************/
-+uint8_t * FM_PORT_GetBufferICInfo(t_Handle h_FmPort, char *p_Data);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_GetBufferPrsResult
-+
-+ @Description Returns the pointer to the parse result in the data buffer.
-+ In Rx ports this is relevant after reception, if parse
-+ result is configured to be part of the data passed to the
-+ application. For non Rx ports it may be used to get the pointer
-+ of the area in the buffer where parse result should be
-+ initialized - if so configured.
-+ See FM_PORT_ConfigBufferPrefixContent for data buffer prefix
-+ configuration.
-+
-+ @Param[in] h_FmPort - FM PORT module descriptor
-+ @Param[in] p_Data - A pointer to the data buffer.
-+
-+ @Return Parse result pointer on success, NULL if parse result was not
-+ configured for this port.
-+
-+ @Cautions Allowed only following FM_PORT_Init().
-+*//***************************************************************************/
-+t_FmPrsResult * FM_PORT_GetBufferPrsResult(t_Handle h_FmPort, char *p_Data);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_GetBufferTimeStamp
-+
-+ @Description Returns the time stamp in the data buffer.
-+ Relevant for Rx ports for getting the buffer time stamp.
-+ See FM_PORT_ConfigBufferPrefixContent for data buffer prefix
-+ configuration.
-+
-+ @Param[in] h_FmPort - FM PORT module descriptor
-+ @Param[in] p_Data - A pointer to the data buffer.
-+
-+ @Return A pointer to the hash result on success, NULL otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Init().
-+*//***************************************************************************/
-+uint64_t * FM_PORT_GetBufferTimeStamp(t_Handle h_FmPort, char *p_Data);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_GetBufferHashResult
-+
-+ @Description Given a data buffer, on the condition that hash result was defined
-+ as a part of the buffer content (see FM_PORT_ConfigBufferPrefixContent)
-+ this routine will return the pointer to the hash result location in the
-+ buffer prefix.
-+
-+ @Param[in] h_FmPort - FM PORT module descriptor
-+ @Param[in] p_Data - A pointer to the data buffer.
-+
-+ @Return A pointer to the hash result on success, NULL otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Init().
-+*//***************************************************************************/
-+uint8_t * FM_PORT_GetBufferHashResult(t_Handle h_FmPort, char *p_Data);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_Disable
-+
-+ @Description Gracefully disable an FM port. The port will not start new tasks after all
-+ tasks associated with the port are terminated.
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Init().
-+ This is a blocking routine, it returns after port is
-+ gracefully stopped, i.e. the port will not except new frames,
-+ but it will finish all frames or tasks which were already began
-+*//***************************************************************************/
-+t_Error FM_PORT_Disable(t_Handle h_FmPort);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_Enable
-+
-+ @Description A runtime routine provided to allow disable/enable of port.
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Init().
-+*//***************************************************************************/
-+t_Error FM_PORT_Enable(t_Handle h_FmPort);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_SetRateLimit
-+
-+ @Description Calling this routine enables rate limit algorithm.
-+ By default, this functionality is disabled.
-+ Note that rate-limit mechanism uses the FM time stamp.
-+ The selected rate limit specified here would be
-+ rounded DOWN to the nearest 16M.
-+
-+ May be used for Tx and OP ports only
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in] p_RateLimit A structure of rate limit parameters
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Init().
-+ If rate limit is set on a port that need to send PFC frames,
-+ it might violate the stop transmit timing.
-+*//***************************************************************************/
-+t_Error FM_PORT_SetRateLimit(t_Handle h_FmPort, t_FmPortRateLimit *p_RateLimit);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_DeleteRateLimit
-+
-+ @Description Calling this routine disables and clears rate limit
-+ initialization.
-+
-+ May be used for Tx and OP ports only
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Init().
-+*//***************************************************************************/
-+t_Error FM_PORT_DeleteRateLimit(t_Handle h_FmPort);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_SetPfcPrioritiesMappingToQmanWQ
-+
-+ @Description Calling this routine maps each PFC received priority to the transmit WQ.
-+ This WQ will be blocked upon receiving a PFC frame with this priority.
-+
-+ May be used for Tx ports only.
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in] prio PFC priority (0-7).
-+ @Param[in] wq Work Queue (0-7).
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Init().
-+*//***************************************************************************/
-+t_Error FM_PORT_SetPfcPrioritiesMappingToQmanWQ(t_Handle h_FmPort, uint8_t prio, uint8_t wq);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_SetStatisticsCounters
-+
-+ @Description Calling this routine enables/disables port's statistics counters.
-+ By default, counters are enabled.
-+
-+ May be used for all port types
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in] enable TRUE to enable, FALSE to disable.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Init().
-+*//***************************************************************************/
-+t_Error FM_PORT_SetStatisticsCounters(t_Handle h_FmPort, bool enable);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_SetFrameQueueCounters
-+
-+ @Description Calling this routine enables/disables port's enqueue/dequeue counters.
-+ By default, counters are enabled.
-+
-+ May be used for all ports
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in] enable TRUE to enable, FALSE to disable.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Init().
-+*//***************************************************************************/
-+t_Error FM_PORT_SetFrameQueueCounters(t_Handle h_FmPort, bool enable);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_AnalyzePerformanceParams
-+
-+ @Description User may call this routine to so the driver will analyze if the
-+ basic performance parameters are correct and also the driver may
-+ suggest of improvements; The basic parameters are FIFO sizes, number
-+ of DMAs and number of TNUMs for the port.
-+
-+ May be used for all port types
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Init().
-+*//***************************************************************************/
-+t_Error FM_PORT_AnalyzePerformanceParams(t_Handle h_FmPort);
-+
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_SetAllocBufCounter
-+
-+ @Description Calling this routine enables/disables BM pool allocate
-+ buffer counters.
-+ By default, counters are enabled.
-+
-+ May be used for Rx ports only
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in] poolId BM pool id.
-+ @Param[in] enable TRUE to enable, FALSE to disable.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Init().
-+*//***************************************************************************/
-+t_Error FM_PORT_SetAllocBufCounter(t_Handle h_FmPort, uint8_t poolId, bool enable);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_GetBmiCounters
-+
-+ @Description Read port's BMI stat counters and place them into
-+ a designated structure of counters.
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[out] p_BmiStats counters structure
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Init().
-+*//***************************************************************************/
-+t_Error FM_PORT_GetBmiCounters(t_Handle h_FmPort, t_FmPortBmiStats *p_BmiStats);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_GetCounter
-+
-+ @Description Reads one of the FM PORT counters.
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in] fmPortCounter The requested counter.
-+
-+ @Return Counter's current value.
-+
-+ @Cautions Allowed only following FM_PORT_Init().
-+ Note that it is user's responsibility to call this routine only
-+ for enabled counters, and there will be no indication if a
-+ disabled counter is accessed.
-+*//***************************************************************************/
-+uint32_t FM_PORT_GetCounter(t_Handle h_FmPort, e_FmPortCounters fmPortCounter);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_ModifyCounter
-+
-+ @Description Sets a value to an enabled counter. Use "0" to reset the counter.
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in] fmPortCounter The requested counter.
-+ @Param[in] value The requested value to be written into the counter.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Init().
-+*//***************************************************************************/
-+t_Error FM_PORT_ModifyCounter(t_Handle h_FmPort, e_FmPortCounters fmPortCounter, uint32_t value);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_GetAllocBufCounter
-+
-+ @Description Reads one of the FM PORT buffer counters.
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in] poolId The requested pool.
-+
-+ @Return Counter's current value.
-+
-+ @Cautions Allowed only following FM_PORT_Init().
-+ Note that it is user's responsibility to call this routine only
-+ for enabled counters, and there will be no indication if a
-+ disabled counter is accessed.
-+*//***************************************************************************/
-+uint32_t FM_PORT_GetAllocBufCounter(t_Handle h_FmPort, uint8_t poolId);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_ModifyAllocBufCounter
-+
-+ @Description Sets a value to an enabled counter. Use "0" to reset the counter.
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in] poolId The requested pool.
-+ @Param[in] value The requested value to be written into the counter.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Init().
-+*//***************************************************************************/
-+t_Error FM_PORT_ModifyAllocBufCounter(t_Handle h_FmPort, uint8_t poolId, uint32_t value);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_AddCongestionGrps
-+
-+ @Description This routine effects the corresponding Tx port.
-+ It should be called in order to enable pause
-+ frame transmission in case of congestion in one or more
-+ of the congestion groups relevant to this port.
-+ Each call to this routine may add one or more congestion
-+ groups to be considered relevant to this port.
-+
-+ May be used for Rx, or RX+OP ports only (depending on chip)
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in] p_CongestionGrps A pointer to an array of congestion groups
-+ id's to consider.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Init().
-+*//***************************************************************************/
-+t_Error FM_PORT_AddCongestionGrps(t_Handle h_FmPort, t_FmPortCongestionGrps *p_CongestionGrps);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_RemoveCongestionGrps
-+
-+ @Description This routine effects the corresponding Tx port. It should be
-+ called when congestion groups were
-+ defined for this port and are no longer relevant, or pause
-+ frames transmitting is not required on their behalf.
-+ Each call to this routine may remove one or more congestion
-+ groups to be considered relevant to this port.
-+
-+ May be used for Rx, or RX+OP ports only (depending on chip)
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in] p_CongestionGrps A pointer to an array of congestion groups
-+ id's to consider.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Init().
-+*//***************************************************************************/
-+t_Error FM_PORT_RemoveCongestionGrps(t_Handle h_FmPort, t_FmPortCongestionGrps *p_CongestionGrps);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_IsStalled
-+
-+ @Description A routine for checking whether the specified port is stalled.
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+
-+ @Return TRUE if port is stalled, FALSE otherwize
-+
-+ @Cautions Allowed only following FM_PORT_Init().
-+*//***************************************************************************/
-+bool FM_PORT_IsStalled(t_Handle h_FmPort);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_ReleaseStalled
-+
-+ @Description This routine may be called in case the port was stalled and may
-+ now be released.
-+ Note that this routine is available only on older FMan revisions
-+ (FMan v2, DPAA v1.0 only).
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Init().
-+*//***************************************************************************/
-+t_Error FM_PORT_ReleaseStalled(t_Handle h_FmPort);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_SetRxL4ChecksumVerify
-+
-+ @Description This routine is relevant for Rx ports (1G and 10G). The routine
-+ set/clear the L3/L4 checksum verification (on RX side).
-+ Note that this takes affect only if hw-parser is enabled!
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in] l4Checksum boolean indicates whether to do L3/L4 checksum
-+ on frames or not.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Init().
-+*//***************************************************************************/
-+t_Error FM_PORT_SetRxL4ChecksumVerify(t_Handle h_FmPort, bool l4Checksum);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_SetErrorsRoute
-+
-+ @Description Errors selected for this routine will cause a frame with that error
-+ to be enqueued to error queue.
-+ Errors not selected for this routine will cause a frame with that error
-+ to be enqueued to the one of the other port queues.
-+ By default all errors are defined to be enqueued to error queue.
-+ Errors that were configured to be discarded (at initialization)
-+ may not be selected here.
-+
-+ May be used for Rx and OP ports only
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in] errs A list of errors to enqueue to error queue
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Config() and before FM_PORT_Init().
-+*//***************************************************************************/
-+t_Error FM_PORT_SetErrorsRoute(t_Handle h_FmPort, fmPortFrameErrSelect_t errs);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_SetIMExceptions
-+
-+ @Description Calling this routine enables/disables FM PORT interrupts.
-+
-+ @Param[in] h_FmPort FM PORT module descriptor.
-+ @Param[in] exception The exception to be selected.
-+ @Param[in] enable TRUE to enable interrupt, FALSE to mask it.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Init().
-+ This routine should NOT be called from guest-partition
-+ (i.e. guestId != NCSW_MASTER_ID)
-+*//***************************************************************************/
-+t_Error FM_PORT_SetIMExceptions(t_Handle h_FmPort, e_FmPortExceptions exception, bool enable);
-+
-+/**************************************************************************//*
-+ @Function FM_PORT_SetPerformanceCounters
-+
-+ @Description Calling this routine enables/disables port's performance counters.
-+ By default, counters are enabled.
-+
-+ May be used for all port types
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in] enable TRUE to enable, FALSE to disable.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Init().
-+*//***************************************************************************/
-+t_Error FM_PORT_SetPerformanceCounters(t_Handle h_FmPort, bool enable);
-+
-+/**************************************************************************//*
-+ @Function FM_PORT_SetPerformanceCountersParams
-+
-+ @Description Calling this routine defines port's performance
-+ counters parameters.
-+
-+ May be used for all port types
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in] p_FmPortPerformanceCnt A pointer to a structure of performance
-+ counters parameters.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Init().
-+*//***************************************************************************/
-+t_Error FM_PORT_SetPerformanceCountersParams(t_Handle h_FmPort, t_FmPortPerformanceCnt *p_FmPortPerformanceCnt);
-+
-+/**************************************************************************//**
-+ @Group FM_PORT_pcd_runtime_control_grp FM Port PCD Runtime Control Unit
-+
-+ @Description FM Port PCD Runtime control unit API functions, definitions and enums.
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Description A structure defining the KG scheme after the parser.
-+ This is relevant only to change scheme selection mode - from
-+ direct to indirect and vice versa, or when the scheme is selected directly,
-+ to select the scheme id.
-+
-+*//***************************************************************************/
-+typedef struct t_FmPcdKgSchemeSelect {
-+ bool direct; /**< TRUE to use 'h_Scheme' directly, FALSE to use LCV. */
-+ t_Handle h_DirectScheme; /**< Scheme handle, selects the scheme after parser;
-+ Relevant only when 'direct' is TRUE. */
-+} t_FmPcdKgSchemeSelect;
-+
-+/**************************************************************************//**
-+ @Description A structure of scheme parameters
-+*//***************************************************************************/
-+typedef struct t_FmPcdPortSchemesParams {
-+ uint8_t numOfSchemes; /**< Number of schemes for port to be bound to. */
-+ t_Handle h_Schemes[FM_PCD_KG_NUM_OF_SCHEMES]; /**< Array of 'numOfSchemes' schemes for the
-+ port to be bound to */
-+} t_FmPcdPortSchemesParams;
-+
-+/**************************************************************************//**
-+ @Description Union for defining port protocol parameters for parser
-+*//***************************************************************************/
-+typedef union u_FmPcdHdrPrsOpts {
-+ /* MPLS */
-+ struct {
-+ bool labelInterpretationEnable; /**< When this bit is set, the last MPLS label will be
-+ interpreted as described in HW spec table. When the bit
-+ is cleared, the parser will advance to MPLS next parse */
-+ e_NetHeaderType nextParse; /**< must be equal or higher than IPv4 */
-+ } mplsPrsOptions;
-+ /* VLAN */
-+ struct {
-+ uint16_t tagProtocolId1; /**< User defined Tag Protocol Identifier, to be recognized
-+ on VLAN TAG on top of 0x8100 and 0x88A8 */
-+ uint16_t tagProtocolId2; /**< User defined Tag Protocol Identifier, to be recognized
-+ on VLAN TAG on top of 0x8100 and 0x88A8 */
-+ } vlanPrsOptions;
-+ /* PPP */
-+ struct{
-+ bool enableMTUCheck; /**< Check validity of MTU according to RFC2516 */
-+ } pppoePrsOptions;
-+
-+ /* IPV6 */
-+ struct{
-+ bool routingHdrEnable; /**< TRUE to enable routing header, otherwise ignore */
-+ } ipv6PrsOptions;
-+
-+ /* UDP */
-+ struct{
-+ bool padIgnoreChecksum; /**< TRUE to ignore pad in checksum */
-+ } udpPrsOptions;
-+
-+ /* TCP */
-+ struct {
-+ bool padIgnoreChecksum; /**< TRUE to ignore pad in checksum */
-+ } tcpPrsOptions;
-+} u_FmPcdHdrPrsOpts;
-+
-+/**************************************************************************//**
-+ @Description A structure for defining each header for the parser
-+*//***************************************************************************/
-+typedef struct t_FmPcdPrsAdditionalHdrParams {
-+ e_NetHeaderType hdr; /**< Selected header; use HEADER_TYPE_NONE
-+ to indicate that sw parser is to run first
-+ (before HW parser, and independent of the
-+ existence of any protocol), in this case,
-+ swPrsEnable must be set, and all other
-+ parameters are irrelevant. */
-+ bool errDisable; /**< TRUE to disable error indication */
-+ bool swPrsEnable; /**< Enable jump to SW parser when this
-+ header is recognized by the HW parser. */
-+ uint8_t indexPerHdr; /**< Normally 0, if more than one sw parser
-+ attachments exists for the same header,
-+ (in the main sw parser code) use this
-+ index to distinguish between them. */
-+ bool usePrsOpts; /**< TRUE to use parser options. */
-+ u_FmPcdHdrPrsOpts prsOpts; /**< A union according to header type,
-+ defining the parser options selected.*/
-+} t_FmPcdPrsAdditionalHdrParams;
-+
-+/**************************************************************************//**
-+ @Description struct for defining port PCD parameters
-+*//***************************************************************************/
-+typedef struct t_FmPortPcdPrsParams {
-+ uint8_t prsResultPrivateInfo; /**< The private info provides a method of inserting
-+ port information into the parser result. This information
-+ may be extracted by Keygen and be used for frames
-+ distribution when a per-port distinction is required,
-+ it may also be used as a port logical id for analyzing
-+ incoming frames. */
-+ uint8_t parsingOffset; /**< Number of bytes from beginning of packet to start parsing */
-+ e_NetHeaderType firstPrsHdr; /**< The type of the first header expected at 'parsingOffset' */
-+ bool includeInPrsStatistics; /**< TRUE to include this port in the parser statistics;
-+ NOTE: this field is not valid when the FM is in "guest" mode
-+ and IPC is not available. */
-+ uint8_t numOfHdrsWithAdditionalParams; /**< Normally 0, some headers may get
-+ special parameters */
-+ t_FmPcdPrsAdditionalHdrParams additionalParams[FM_PCD_PRS_NUM_OF_HDRS];
-+ /**< 'numOfHdrsWithAdditionalParams' structures
-+ of additional parameters
-+ for each header that requires them */
-+ bool setVlanTpid1; /**< TRUE to configure user selection of Ethertype to
-+ indicate a VLAN tag (in addition to the TPID values
-+ 0x8100 and 0x88A8). */
-+ uint16_t vlanTpid1; /**< extra tag to use if setVlanTpid1=TRUE. */
-+ bool setVlanTpid2; /**< TRUE to configure user selection of Ethertype to
-+ indicate a VLAN tag (in addition to the TPID values
-+ 0x8100 and 0x88A8). */
-+ uint16_t vlanTpid2; /**< extra tag to use if setVlanTpid1=TRUE. */
-+} t_FmPortPcdPrsParams;
-+
-+/**************************************************************************//**
-+ @Description struct for defining coarse alassification parameters
-+*//***************************************************************************/
-+typedef struct t_FmPortPcdCcParams {
-+ t_Handle h_CcTree; /**< A handle to a CC tree */
-+} t_FmPortPcdCcParams;
-+
-+/**************************************************************************//**
-+ @Description struct for defining keygen parameters
-+*//***************************************************************************/
-+typedef struct t_FmPortPcdKgParams {
-+ uint8_t numOfSchemes; /**< Number of schemes for port to be bound to. */
-+ t_Handle h_Schemes[FM_PCD_KG_NUM_OF_SCHEMES];
-+ /**< Array of 'numOfSchemes' schemes handles for the
-+ port to be bound to */
-+ bool directScheme; /**< TRUE for going from parser to a specific scheme,
-+ regardless of parser result */
-+ t_Handle h_DirectScheme; /**< relevant only if direct == TRUE, Scheme handle,
-+ as returned by FM_PCD_KgSetScheme */
-+} t_FmPortPcdKgParams;
-+
-+/**************************************************************************//**
-+ @Description struct for defining policer parameters
-+*//***************************************************************************/
-+typedef struct t_FmPortPcdPlcrParams {
-+ t_Handle h_Profile; /**< Selected profile handle */
-+} t_FmPortPcdPlcrParams;
-+
-+/**************************************************************************//**
-+ @Description struct for defining port PCD parameters
-+*//***************************************************************************/
-+typedef struct t_FmPortPcdParams {
-+ e_FmPortPcdSupport pcdSupport; /**< Relevant for Rx and offline ports only.
-+ Describes the active PCD engines for this port. */
-+ t_Handle h_NetEnv; /**< HL Unused in PLCR only mode */
-+ t_FmPortPcdPrsParams *p_PrsParams; /**< Parser parameters for this port */
-+ t_FmPortPcdCcParams *p_CcParams; /**< Coarse classification parameters for this port */
-+ t_FmPortPcdKgParams *p_KgParams; /**< Keygen parameters for this port */
-+ t_FmPortPcdPlcrParams *p_PlcrParams; /**< Policer parameters for this port; Relevant for one of
-+ following cases:
-+ e_FM_PORT_PCD_SUPPORT_PLCR_ONLY or
-+ e_FM_PORT_PCD_SUPPORT_PRS_AND_PLCR were selected,
-+ or if any flow uses a KG scheme were policer
-+ profile is not generated
-+ ('bypassPlcrProfileGeneration selected'). */
-+ t_Handle h_IpReassemblyManip; /**< IP Reassembly manipulation */
-+#if (DPAA_VERSION >= 11)
-+ t_Handle h_CapwapReassemblyManip;/**< CAPWAP Reassembly manipulation */
-+#endif /* (DPAA_VERSION >= 11) */
-+} t_FmPortPcdParams;
-+
-+/**************************************************************************//**
-+ @Description A structure for defining the Parser starting point
-+*//***************************************************************************/
-+typedef struct t_FmPcdPrsStart {
-+ uint8_t parsingOffset; /**< Number of bytes from beginning of packet to
-+ start parsing */
-+ e_NetHeaderType firstPrsHdr; /**< The type of the first header axpected at
-+ 'parsingOffset' */
-+} t_FmPcdPrsStart;
-+
-+#if (DPAA_VERSION >= 11)
-+/**************************************************************************//**
-+ @Description struct for defining external buffer margins
-+*//***************************************************************************/
-+typedef struct t_FmPortVSPAllocParams {
-+ uint8_t numOfProfiles; /**< Number of Virtual Storage Profiles; must be a power of 2 */
-+ uint8_t dfltRelativeId; /**< The default Virtual-Storage-Profile-id dedicated to Rx/OP port
-+ The same default Virtual-Storage-Profile-id will be for coupled Tx port
-+ if relevant function called for Rx port */
-+ t_Handle h_FmTxPort; /**< Handle to coupled Tx Port; not relevant for OP port. */
-+} t_FmPortVSPAllocParams;
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_SetPCD
-+
-+ @Description Calling this routine defines the port's PCD configuration.
-+ It changes it from its default configuration which is PCD
-+ disabled (BMI to BMI) and configures it according to the passed
-+ parameters.
-+
-+ May be used for Rx and OP ports only
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in] p_FmPortPcd A Structure of parameters defining the port's PCD
-+ configuration.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Init().
-+*//***************************************************************************/
-+t_Error FM_PORT_SetPCD(t_Handle h_FmPort, t_FmPortPcdParams *p_FmPortPcd);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_DeletePCD
-+
-+ @Description Calling this routine releases the port's PCD configuration.
-+ The port returns to its default configuration which is PCD
-+ disabled (BMI to BMI) and all PCD configuration is removed.
-+
-+ May be used for Rx and OP ports which are
-+ in PCD mode only
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Init().
-+*//***************************************************************************/
-+t_Error FM_PORT_DeletePCD(t_Handle h_FmPort);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_AttachPCD
-+
-+ @Description This routine may be called after FM_PORT_DetachPCD was called,
-+ to return to the originally configured PCD support flow.
-+ The couple of routines are used to allow PCD configuration changes
-+ that demand that PCD will not be used while changes take place.
-+
-+ May be used for Rx and OP ports which are
-+ in PCD mode only
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Init().
-+*//***************************************************************************/
-+t_Error FM_PORT_AttachPCD(t_Handle h_FmPort);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_DetachPCD
-+
-+ @Description Calling this routine detaches the port from its PCD functionality.
-+ The port returns to its default flow which is BMI to BMI.
-+
-+ May be used for Rx and OP ports which are
-+ in PCD mode only
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_AttachPCD().
-+*//***************************************************************************/
-+t_Error FM_PORT_DetachPCD(t_Handle h_FmPort);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_PcdPlcrAllocProfiles
-+
-+ @Description This routine may be called only for ports that use the Policer in
-+ order to allocate private policer profiles.
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in] numOfProfiles The number of required policer profiles
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Init() and FM_PCD_Init(),
-+ and before FM_PORT_SetPCD().
-+*//***************************************************************************/
-+t_Error FM_PORT_PcdPlcrAllocProfiles(t_Handle h_FmPort, uint16_t numOfProfiles);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_PcdPlcrFreeProfiles
-+
-+ @Description This routine should be called for freeing private policer profiles.
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Init() and FM_PCD_Init(),
-+ and before FM_PORT_SetPCD().
-+*//***************************************************************************/
-+t_Error FM_PORT_PcdPlcrFreeProfiles(t_Handle h_FmPort);
-+
-+#if (DPAA_VERSION >= 11)
-+/**************************************************************************//**
-+ @Function FM_PORT_VSPAlloc
-+
-+ @Description This routine allocated VSPs per port and forces the port to work
-+ in VSP mode. Note that the port is initialized by default with the
-+ physical-storage-profile only.
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in] p_Params A structure of parameters for allocation VSP's per port
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Init(), and before FM_PORT_SetPCD()
-+ and also before FM_PORT_Enable(); i.e. the port should be disabled.
-+*//***************************************************************************/
-+t_Error FM_PORT_VSPAlloc(t_Handle h_FmPort, t_FmPortVSPAllocParams *p_Params);
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_PcdKgModifyInitialScheme
-+
-+ @Description This routine may be called only for ports that use the keygen in
-+ order to change the initial scheme frame should be routed to.
-+ The change may be of a scheme id (in case of direct mode),
-+ from direct to indirect, or from indirect to direct - specifying the scheme id.
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in] p_FmPcdKgScheme A structure of parameters for defining whether
-+ a scheme is direct/indirect, and if direct - scheme id.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Init() and FM_PORT_SetPCD().
-+*//***************************************************************************/
-+t_Error FM_PORT_PcdKgModifyInitialScheme (t_Handle h_FmPort, t_FmPcdKgSchemeSelect *p_FmPcdKgScheme);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_PcdPlcrModifyInitialProfile
-+
-+ @Description This routine may be called for ports with flows
-+ e_FM_PORT_PCD_SUPPORT_PLCR_ONLY or e_FM_PORT_PCD_SUPPORT_PRS_AND_PLCR
-+ only, to change the initial Policer profile frame should be
-+ routed to. The change may be of a profile and/or absolute/direct
-+ mode selection.
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in] h_Profile Policer profile handle
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Init() and FM_PORT_SetPCD().
-+*//***************************************************************************/
-+t_Error FM_PORT_PcdPlcrModifyInitialProfile (t_Handle h_FmPort, t_Handle h_Profile);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_PcdCcModifyTree
-+
-+ @Description This routine may be called for ports that use coarse classification tree
-+ if the user wishes to replace the tree. The routine may not be called while port
-+ receives packets using the PCD functionalities, therefor port must be first detached
-+ from the PCD, only than the routine may be called, and than port be attached to PCD again.
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in] h_CcTree A CC tree that was already built. The tree id as returned from
-+ the BuildTree routine.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Init(), FM_PORT_SetPCD() and FM_PORT_DetachPCD()
-+*//***************************************************************************/
-+t_Error FM_PORT_PcdCcModifyTree (t_Handle h_FmPort, t_Handle h_CcTree);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_PcdKgBindSchemes
-+
-+ @Description These routines may be called for adding more schemes for the
-+ port to be bound to. The selected schemes are not added,
-+ just this specific port starts using them.
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in] p_PortScheme A structure defining the list of schemes to be added.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Init() and FM_PORT_SetPCD().
-+*//***************************************************************************/
-+t_Error FM_PORT_PcdKgBindSchemes (t_Handle h_FmPort, t_FmPcdPortSchemesParams *p_PortScheme);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_PcdKgUnbindSchemes
-+
-+ @Description These routines may be called for adding more schemes for the
-+ port to be bound to. The selected schemes are not removed or invalidated,
-+ just this specific port stops using them.
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in] p_PortScheme A structure defining the list of schemes to be added.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Init() and FM_PORT_SetPCD().
-+*//***************************************************************************/
-+t_Error FM_PORT_PcdKgUnbindSchemes (t_Handle h_FmPort, t_FmPcdPortSchemesParams *p_PortScheme);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_GetIPv4OptionsCount
-+
-+ @Description TODO
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[out] p_Ipv4OptionsCount will hold the counter value
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Init()
-+*//***************************************************************************/
-+t_Error FM_PORT_GetIPv4OptionsCount(t_Handle h_FmPort, uint32_t *p_Ipv4OptionsCount);
-+
-+/** @} */ /* end of FM_PORT_pcd_runtime_control_grp group */
-+/** @} */ /* end of FM_PORT_runtime_control_grp group */
-+
-+
-+/**************************************************************************//**
-+ @Group FM_PORT_runtime_data_grp FM Port Runtime Data-path Unit
-+
-+ @Description FM Port Runtime data unit API functions, definitions and enums.
-+ This API is valid only if working in Independent-Mode.
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_ImTx
-+
-+ @Description Tx function, called to transmit a data buffer on the port.
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in] p_Data A pointer to an LCP data buffer.
-+ @Param[in] length Size of data for transmission.
-+ @Param[in] lastBuffer Buffer position - TRUE for the last buffer
-+ of a frame, including a single buffer frame
-+ @Param[in] h_BufContext A handle of the user acossiated with this buffer
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Init().
-+ NOTE - This routine can be used only when working in
-+ Independent-Mode mode.
-+*//***************************************************************************/
-+t_Error FM_PORT_ImTx( t_Handle h_FmPort,
-+ uint8_t *p_Data,
-+ uint16_t length,
-+ bool lastBuffer,
-+ t_Handle h_BufContext);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_ImTxConf
-+
-+ @Description Tx port confirmation routine, optional, may be called to verify
-+ transmission of all frames. The procedure performed by this
-+ routine will be performed automatically on next buffer transmission,
-+ but if desired, calling this routine will invoke this action on
-+ demand.
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+
-+ @Cautions Allowed only following FM_PORT_Init().
-+ NOTE - This routine can be used only when working in
-+ Independent-Mode mode.
-+*//***************************************************************************/
-+void FM_PORT_ImTxConf(t_Handle h_FmPort);
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_ImRx
-+
-+ @Description Rx function, may be called to poll for received buffers.
-+ Normally, Rx process is invoked by the driver on Rx interrupt.
-+ Alternatively, this routine may be called on demand.
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Init().
-+ NOTE - This routine can be used only when working in
-+ Independent-Mode mode.
-+*//***************************************************************************/
-+t_Error FM_PORT_ImRx(t_Handle h_FmPort);
-+
-+/** @} */ /* end of FM_PORT_runtime_data_grp group */
-+/** @} */ /* end of FM_PORT_grp group */
-+/** @} */ /* end of FM_grp group */
-+
-+
-+
-+#ifdef NCSW_BACKWARD_COMPATIBLE_API
-+#define FM_PORT_ConfigTxFifoDeqPipelineDepth FM_PORT_ConfigFifoDeqPipelineDepth
-+#endif /* NCSW_BACKWARD_COMPATIBLE_API */
-+
-+
-+#endif /* __FM_PORT_EXT */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/Peripherals/fm_rtc_ext.h
-@@ -0,0 +1,619 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/**************************************************************************//**
-+ @File fm_rtc_ext.h
-+
-+ @Description External definitions and API for FM RTC IEEE1588 Timer Module.
-+
-+ @Cautions None.
-+*//***************************************************************************/
-+
-+#ifndef __FM_RTC_EXT_H__
-+#define __FM_RTC_EXT_H__
-+
-+
-+#include "error_ext.h"
-+#include "std_ext.h"
-+#include "fsl_fman_rtc.h"
-+
-+/**************************************************************************//**
-+
-+ @Group FM_grp Frame Manager API
-+
-+ @Description FM API functions, definitions and enums
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Group fm_rtc_grp FM RTC
-+
-+ @Description FM RTC functions, definitions and enums.
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Group fm_rtc_init_grp FM RTC Initialization Unit
-+
-+ @Description FM RTC initialization API.
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Description FM RTC Alarm Polarity Options.
-+*//***************************************************************************/
-+typedef enum e_FmRtcAlarmPolarity
-+{
-+ e_FM_RTC_ALARM_POLARITY_ACTIVE_HIGH = E_FMAN_RTC_ALARM_POLARITY_ACTIVE_HIGH, /**< Active-high output polarity */
-+ e_FM_RTC_ALARM_POLARITY_ACTIVE_LOW = E_FMAN_RTC_ALARM_POLARITY_ACTIVE_LOW /**< Active-low output polarity */
-+} e_FmRtcAlarmPolarity;
-+
-+/**************************************************************************//**
-+ @Description FM RTC Trigger Polarity Options.
-+*//***************************************************************************/
-+typedef enum e_FmRtcTriggerPolarity
-+{
-+ e_FM_RTC_TRIGGER_ON_RISING_EDGE = E_FMAN_RTC_TRIGGER_ON_RISING_EDGE, /**< Trigger on rising edge */
-+ e_FM_RTC_TRIGGER_ON_FALLING_EDGE = E_FMAN_RTC_TRIGGER_ON_FALLING_EDGE /**< Trigger on falling edge */
-+} e_FmRtcTriggerPolarity;
-+
-+/**************************************************************************//**
-+ @Description IEEE1588 Timer Module FM RTC Optional Clock Sources.
-+*//***************************************************************************/
-+typedef enum e_FmSrcClock
-+{
-+ e_FM_RTC_SOURCE_CLOCK_EXTERNAL = E_FMAN_RTC_SOURCE_CLOCK_EXTERNAL, /**< external high precision timer reference clock */
-+ e_FM_RTC_SOURCE_CLOCK_SYSTEM = E_FMAN_RTC_SOURCE_CLOCK_SYSTEM, /**< MAC system clock */
-+ e_FM_RTC_SOURCE_CLOCK_OSCILATOR = E_FMAN_RTC_SOURCE_CLOCK_OSCILATOR /**< RTC clock oscilator */
-+}e_FmSrcClk;
-+
-+/**************************************************************************//**
-+ @Description FM RTC configuration parameters structure.
-+
-+ This structure should be passed to FM_RTC_Config().
-+*//***************************************************************************/
-+typedef struct t_FmRtcParams
-+{
-+ t_Handle h_Fm; /**< FM Handle*/
-+ uintptr_t baseAddress; /**< Base address of FM RTC registers */
-+ t_Handle h_App; /**< A handle to an application layer object; This handle will
-+ be passed by the driver upon calling the above callbacks */
-+} t_FmRtcParams;
-+
-+
-+/**************************************************************************//**
-+ @Function FM_RTC_Config
-+
-+ @Description Configures the FM RTC module according to user's parameters.
-+
-+ The driver assigns default values to some FM RTC parameters.
-+ These parameters can be overwritten using the advanced
-+ configuration routines.
-+
-+ @Param[in] p_FmRtcParam - FM RTC configuration parameters.
-+
-+ @Return Handle to the new FM RTC object; NULL pointer on failure.
-+
-+ @Cautions None
-+*//***************************************************************************/
-+t_Handle FM_RTC_Config(t_FmRtcParams *p_FmRtcParam);
-+
-+/**************************************************************************//**
-+ @Function FM_RTC_Init
-+
-+ @Description Initializes the FM RTC driver and hardware.
-+
-+ @Param[in] h_FmRtc - Handle to FM RTC object.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions h_FmRtc must have been previously created using FM_RTC_Config().
-+*//***************************************************************************/
-+t_Error FM_RTC_Init(t_Handle h_FmRtc);
-+
-+/**************************************************************************//**
-+ @Function FM_RTC_Free
-+
-+ @Description Frees the FM RTC object and all allocated resources.
-+
-+ @Param[in] h_FmRtc - Handle to FM RTC object.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions h_FmRtc must have been previously created using FM_RTC_Config().
-+*//***************************************************************************/
-+t_Error FM_RTC_Free(t_Handle h_FmRtc);
-+
-+
-+/**************************************************************************//**
-+ @Group fm_rtc_adv_config_grp FM RTC Advanced Configuration Unit
-+
-+ @Description FM RTC advanced configuration functions.
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Function FM_RTC_ConfigPeriod
-+
-+ @Description Configures the period of the timestamp if different than
-+ default [DEFAULT_clockPeriod].
-+
-+ @Param[in] h_FmRtc - Handle to FM RTC object.
-+ @Param[in] period - Period in nano-seconds.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions h_FmRtc must have been previously created using FM_RTC_Config().
-+*//***************************************************************************/
-+t_Error FM_RTC_ConfigPeriod(t_Handle h_FmRtc, uint32_t period);
-+
-+/**************************************************************************//**
-+ @Function FM_RTC_ConfigSourceClock
-+
-+ @Description Configures the source clock of the RTC.
-+
-+ @Param[in] h_FmRtc - Handle to FM RTC object.
-+ @Param[in] srcClk - Source clock selection.
-+ @Param[in] freqInMhz - the source-clock frequency (in MHz).
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions h_FmRtc must have been previously created using FM_RTC_Config().
-+*//***************************************************************************/
-+t_Error FM_RTC_ConfigSourceClock(t_Handle h_FmRtc,
-+ e_FmSrcClk srcClk,
-+ uint32_t freqInMhz);
-+
-+/**************************************************************************//**
-+ @Function FM_RTC_ConfigPulseRealignment
-+
-+ @Description Configures the RTC to automatic FIPER pulse realignment in
-+ response to timer adjustments [DEFAULT_pulseRealign]
-+
-+ In this mode, the RTC clock is identical to the source clock.
-+ This feature can be useful when the system contains an external
-+ RTC with inherent frequency compensation.
-+
-+ @Param[in] h_FmRtc - Handle to FM RTC object.
-+ @Param[in] enable - TRUE to enable automatic realignment.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions h_FmRtc must have been previously created using FM_RTC_Config().
-+*//***************************************************************************/
-+t_Error FM_RTC_ConfigPulseRealignment(t_Handle h_FmRtc, bool enable);
-+
-+/**************************************************************************//**
-+ @Function FM_RTC_ConfigFrequencyBypass
-+
-+ @Description Configures the RTC to bypass the frequency compensation
-+ mechanism. [DEFAULT_bypass]
-+
-+ In this mode, the RTC clock is identical to the source clock.
-+ This feature can be useful when the system contains an external
-+ RTC with inherent frequency compensation.
-+
-+ @Param[in] h_FmRtc - Handle to FM RTC object.
-+ @Param[in] enabled - TRUE to bypass frequency compensation;
-+ FALSE otherwise.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions h_FmRtc must have been previously created using FM_RTC_Config().
-+*//***************************************************************************/
-+t_Error FM_RTC_ConfigFrequencyBypass(t_Handle h_FmRtc, bool enabled);
-+
-+/**************************************************************************//**
-+ @Function FM_RTC_ConfigInvertedInputClockPhase
-+
-+ @Description Configures the RTC to invert the source clock phase on input.
-+ [DEFAULT_invertInputClkPhase]
-+
-+ @Param[in] h_FmRtc - Handle to FM RTC object.
-+ @Param[in] inverted - TRUE to invert the source clock phase on input.
-+ FALSE otherwise.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions h_FmRtc must have been previously created using FM_RTC_Config().
-+*//***************************************************************************/
-+t_Error FM_RTC_ConfigInvertedInputClockPhase(t_Handle h_FmRtc, bool inverted);
-+
-+/**************************************************************************//**
-+ @Function FM_RTC_ConfigInvertedOutputClockPhase
-+
-+ @Description Configures the RTC to invert the output clock phase.
-+ [DEFAULT_invertOutputClkPhase]
-+
-+ @Param[in] h_FmRtc - Handle to FM RTC object.
-+ @Param[in] inverted - TRUE to invert the output clock phase.
-+ FALSE otherwise.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions h_FmRtc must have been previously created using FM_RTC_Config().
-+*//***************************************************************************/
-+t_Error FM_RTC_ConfigInvertedOutputClockPhase(t_Handle h_FmRtc, bool inverted);
-+
-+/**************************************************************************//**
-+ @Function FM_RTC_ConfigOutputClockDivisor
-+
-+ @Description Configures the divisor for generating the output clock from
-+ the RTC clock. [DEFAULT_outputClockDivisor]
-+
-+ @Param[in] h_FmRtc - Handle to FM RTC object.
-+ @Param[in] divisor - Divisor for generation of the output clock.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions h_FmRtc must have been previously created using FM_RTC_Config().
-+*//***************************************************************************/
-+t_Error FM_RTC_ConfigOutputClockDivisor(t_Handle h_FmRtc, uint16_t divisor);
-+
-+/**************************************************************************//**
-+ @Function FM_RTC_ConfigAlarmPolarity
-+
-+ @Description Configures the polarity (active-high/active-low) of a specific
-+ alarm signal. [DEFAULT_alarmPolarity]
-+
-+ @Param[in] h_FmRtc - Handle to FM RTC object.
-+ @Param[in] alarmId - Alarm ID.
-+ @Param[in] alarmPolarity - Alarm polarity.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions h_FmRtc must have been previously created using FM_RTC_Config().
-+*//***************************************************************************/
-+t_Error FM_RTC_ConfigAlarmPolarity(t_Handle h_FmRtc,
-+ uint8_t alarmId,
-+ e_FmRtcAlarmPolarity alarmPolarity);
-+
-+/**************************************************************************//**
-+ @Function FM_RTC_ConfigExternalTriggerPolarity
-+
-+ @Description Configures the polarity (rising/falling edge) of a specific
-+ external trigger signal. [DEFAULT_triggerPolarity]
-+
-+ @Param[in] h_FmRtc - Handle to FM RTC object.
-+ @Param[in] triggerId - Trigger ID.
-+ @Param[in] triggerPolarity - Trigger polarity.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions h_FmRtc must have been previously created using FM_RTC_Config().
-+*//***************************************************************************/
-+t_Error FM_RTC_ConfigExternalTriggerPolarity(t_Handle h_FmRtc,
-+ uint8_t triggerId,
-+ e_FmRtcTriggerPolarity triggerPolarity);
-+
-+/** @} */ /* end of fm_rtc_adv_config_grp */
-+/** @} */ /* end of fm_rtc_init_grp */
-+
-+
-+/**************************************************************************//**
-+ @Group fm_rtc_control_grp FM RTC Control Unit
-+
-+ @Description FM RTC runtime control API.
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Function t_FmRtcExceptionsCallback
-+
-+ @Description Exceptions user callback routine, used for RTC different mechanisms.
-+
-+ @Param[in] h_App - User's application descriptor.
-+ @Param[in] id - source id.
-+*//***************************************************************************/
-+typedef void (t_FmRtcExceptionsCallback) ( t_Handle h_App, uint8_t id);
-+
-+/**************************************************************************//**
-+ @Description FM RTC alarm parameters.
-+*//***************************************************************************/
-+typedef struct t_FmRtcAlarmParams {
-+ uint8_t alarmId; /**< 0 or 1 */
-+ uint64_t alarmTime; /**< In nanoseconds, the time when the alarm
-+ should go off - must be a multiple of
-+ the RTC period */
-+ t_FmRtcExceptionsCallback *f_AlarmCallback; /**< This routine will be called when RTC
-+ reaches alarmTime */
-+ bool clearOnExpiration; /**< TRUE to turn off the alarm once expired. */
-+} t_FmRtcAlarmParams;
-+
-+/**************************************************************************//**
-+ @Description FM RTC Periodic Pulse parameters.
-+*//***************************************************************************/
-+typedef struct t_FmRtcPeriodicPulseParams {
-+ uint8_t periodicPulseId; /**< 0 or 1 */
-+ uint64_t periodicPulsePeriod; /**< In Nanoseconds. Must be
-+ a multiple of the RTC period */
-+ t_FmRtcExceptionsCallback *f_PeriodicPulseCallback; /**< This routine will be called every
-+ periodicPulsePeriod. */
-+} t_FmRtcPeriodicPulseParams;
-+
-+/**************************************************************************//**
-+ @Description FM RTC Periodic Pulse parameters.
-+*//***************************************************************************/
-+typedef struct t_FmRtcExternalTriggerParams {
-+ uint8_t externalTriggerId; /**< 0 or 1 */
-+ bool usePulseAsInput; /**< Use the pulse interrupt instead of
-+ an external signal */
-+ t_FmRtcExceptionsCallback *f_ExternalTriggerCallback; /**< This routine will be called every
-+ periodicPulsePeriod. */
-+} t_FmRtcExternalTriggerParams;
-+
-+
-+/**************************************************************************//**
-+ @Function FM_RTC_Enable
-+
-+ @Description Enable the RTC (time count is started).
-+
-+ The user can select to resume the time count from previous
-+ point, or to restart the time count.
-+
-+ @Param[in] h_FmRtc - Handle to FM RTC object.
-+ @Param[in] resetClock - Restart the time count from zero.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions h_FmRtc must have been previously initialized using FM_RTC_Init().
-+*//***************************************************************************/
-+t_Error FM_RTC_Enable(t_Handle h_FmRtc, bool resetClock);
-+
-+/**************************************************************************//**
-+ @Function FM_RTC_Disable
-+
-+ @Description Disables the RTC (time count is stopped).
-+
-+ @Param[in] h_FmRtc - Handle to FM RTC object.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions h_FmRtc must have been previously initialized using FM_RTC_Init().
-+*//***************************************************************************/
-+t_Error FM_RTC_Disable(t_Handle h_FmRtc);
-+
-+/**************************************************************************//**
-+ @Function FM_RTC_SetClockOffset
-+
-+ @Description Sets the clock offset (usually relative to another clock).
-+
-+ The user can pass a negative offset value.
-+
-+ @Param[in] h_FmRtc - Handle to FM RTC object.
-+ @Param[in] offset - New clock offset (in nanoseconds).
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions h_FmRtc must have been previously initialized using FM_RTC_Init().
-+*//***************************************************************************/
-+t_Error FM_RTC_SetClockOffset(t_Handle h_FmRtc, int64_t offset);
-+
-+/**************************************************************************//**
-+ @Function FM_RTC_SetAlarm
-+
-+ @Description Schedules an alarm event to a given RTC time.
-+
-+ @Param[in] h_FmRtc - Handle to FM RTC object.
-+ @Param[in] p_FmRtcAlarmParams - Alarm parameters.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions h_FmRtc must have been previously initialized using FM_RTC_Init().
-+ Must be called only prior to FM_RTC_Enable().
-+*//***************************************************************************/
-+t_Error FM_RTC_SetAlarm(t_Handle h_FmRtc, t_FmRtcAlarmParams *p_FmRtcAlarmParams);
-+
-+/**************************************************************************//**
-+ @Function FM_RTC_SetPeriodicPulse
-+
-+ @Description Sets a periodic pulse.
-+
-+ @Param[in] h_FmRtc - Handle to FM RTC object.
-+ @Param[in] p_FmRtcPeriodicPulseParams - Periodic pulse parameters.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions h_FmRtc must have been previously initialized using FM_RTC_Init().
-+ Must be called only prior to FM_RTC_Enable().
-+*//***************************************************************************/
-+t_Error FM_RTC_SetPeriodicPulse(t_Handle h_FmRtc, t_FmRtcPeriodicPulseParams *p_FmRtcPeriodicPulseParams);
-+
-+/**************************************************************************//**
-+ @Function FM_RTC_ClearPeriodicPulse
-+
-+ @Description Clears a periodic pulse.
-+
-+ @Param[in] h_FmRtc - Handle to FM RTC object.
-+ @Param[in] periodicPulseId - Periodic pulse id.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions h_FmRtc must have been previously initialized using FM_RTC_Init().
-+*//***************************************************************************/
-+t_Error FM_RTC_ClearPeriodicPulse(t_Handle h_FmRtc, uint8_t periodicPulseId);
-+
-+/**************************************************************************//**
-+ @Function FM_RTC_SetExternalTrigger
-+
-+ @Description Sets an external trigger indication and define a callback
-+ routine to be called on such event.
-+
-+ @Param[in] h_FmRtc - Handle to FM RTC object.
-+ @Param[in] p_FmRtcExternalTriggerParams - External Trigger parameters.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions h_FmRtc must have been previously initialized using FM_RTC_Init().
-+*//***************************************************************************/
-+t_Error FM_RTC_SetExternalTrigger(t_Handle h_FmRtc, t_FmRtcExternalTriggerParams *p_FmRtcExternalTriggerParams);
-+
-+/**************************************************************************//**
-+ @Function FM_RTC_ClearExternalTrigger
-+
-+ @Description Clears external trigger indication.
-+
-+ @Param[in] h_FmRtc - Handle to FM RTC object.
-+ @Param[in] id - External Trigger id.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions h_FmRtc must have been previously initialized using FM_RTC_Init().
-+*//***************************************************************************/
-+t_Error FM_RTC_ClearExternalTrigger(t_Handle h_FmRtc, uint8_t id);
-+
-+/**************************************************************************//**
-+ @Function FM_RTC_GetExternalTriggerTimeStamp
-+
-+ @Description Reads the External Trigger TimeStamp.
-+
-+ @Param[in] h_FmRtc - Handle to FM RTC object.
-+ @Param[in] triggerId - External Trigger id.
-+ @Param[out] p_TimeStamp - External Trigger timestamp (in nanoseconds).
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions h_FmRtc must have been previously initialized using FM_RTC_Init().
-+*//***************************************************************************/
-+t_Error FM_RTC_GetExternalTriggerTimeStamp(t_Handle h_FmRtc,
-+ uint8_t triggerId,
-+ uint64_t *p_TimeStamp);
-+
-+/**************************************************************************//**
-+ @Function FM_RTC_GetCurrentTime
-+
-+ @Description Returns the current RTC time.
-+
-+ @Param[in] h_FmRtc - Handle to FM RTC object.
-+ @Param[out] p_Ts - returned time stamp (in nanoseconds).
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions h_FmRtc must have been previously initialized using FM_RTC_Init().
-+*//***************************************************************************/
-+t_Error FM_RTC_GetCurrentTime(t_Handle h_FmRtc, uint64_t *p_Ts);
-+
-+/**************************************************************************//**
-+ @Function FM_RTC_SetCurrentTime
-+
-+ @Description Sets the current RTC time.
-+
-+ @Param[in] h_FmRtc - Handle to FM RTC object.
-+ @Param[in] ts - The new time stamp (in nanoseconds).
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions h_FmRtc must have been previously initialized using FM_RTC_Init().
-+*//***************************************************************************/
-+t_Error FM_RTC_SetCurrentTime(t_Handle h_FmRtc, uint64_t ts);
-+
-+/**************************************************************************//**
-+ @Function FM_RTC_GetFreqCompensation
-+
-+ @Description Retrieves the frequency compensation value
-+
-+ @Param[in] h_FmRtc - Handle to FM RTC object.
-+ @Param[out] p_Compensation - A pointer to the returned value of compensation.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions h_FmRtc must have been previously initialized using FM_RTC_Init().
-+*//***************************************************************************/
-+t_Error FM_RTC_GetFreqCompensation(t_Handle h_FmRtc, uint32_t *p_Compensation);
-+
-+/**************************************************************************//**
-+ @Function FM_RTC_SetFreqCompensation
-+
-+ @Description Sets a new frequency compensation value.
-+
-+ @Param[in] h_FmRtc - Handle to FM RTC object.
-+ @Param[in] freqCompensation - The new frequency compensation value to set.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions h_FmRtc must have been previously initialized using FM_RTC_Init().
-+*//***************************************************************************/
-+t_Error FM_RTC_SetFreqCompensation(t_Handle h_FmRtc, uint32_t freqCompensation);
-+
-+#ifdef CONFIG_PTP_1588_CLOCK_DPAA
-+/**************************************************************************//**
-+*@Function FM_RTC_EnableInterrupt
-+*
-+*@Description Enable interrupt of FM RTC.
-+*
-+*@Param[in] h_FmRtc - Handle to FM RTC object.
-+*@Param[in] events - Interrupt events.
-+*
-+*@Return E_OK on success; Error code otherwise.
-+*//***************************************************************************/
-+t_Error FM_RTC_EnableInterrupt(t_Handle h_FmRtc, uint32_t events);
-+
-+/**************************************************************************//**
-+*@Function FM_RTC_DisableInterrupt
-+*
-+*@Description Disable interrupt of FM RTC.
-+*
-+*@Param[in] h_FmRtc - Handle to FM RTC object.
-+*@Param[in] events - Interrupt events.
-+*
-+*@Return E_OK on success; Error code otherwise.
-+*//***************************************************************************/
-+t_Error FM_RTC_DisableInterrupt(t_Handle h_FmRtc, uint32_t events);
-+#endif
-+
-+#if (defined(DEBUG_ERRORS) && (DEBUG_ERRORS > 0))
-+/**************************************************************************//**
-+ @Function FM_RTC_DumpRegs
-+
-+ @Description Dumps all FM registers
-+
-+ @Param[in] h_FmRtc A handle to an FM RTC Module.
-+
-+ @Return E_OK on success;
-+
-+ @Cautions Allowed only FM_Init().
-+*//***************************************************************************/
-+t_Error FM_RTC_DumpRegs(t_Handle h_FmRtc);
-+#endif /* (defined(DEBUG_ERRORS) && ... */
-+
-+/** @} */ /* end of fm_rtc_control_grp */
-+/** @} */ /* end of fm_rtc_grp */
-+/** @} */ /* end of FM_grp group */
-+
-+
-+#endif /* __FM_RTC_EXT_H__ */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/Peripherals/fm_vsp_ext.h
-@@ -0,0 +1,411 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/**************************************************************************//**
-+ @File fm_vsp_ext.h
-+
-+ @Description FM Virtual Storage-Profile ...
-+*//***************************************************************************/
-+#ifndef __FM_VSP_EXT_H
-+#define __FM_VSP_EXT_H
-+
-+#include "std_ext.h"
-+#include "error_ext.h"
-+#include "string_ext.h"
-+#include "debug_ext.h"
-+
-+#include "fm_ext.h"
-+
-+
-+/**************************************************************************//**
-+
-+ @Group FM_grp Frame Manager API
-+
-+ @Description FM API functions, definitions and enums
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Group FM_VSP_grp FM Virtual-Storage-Profile
-+
-+ @Description FM Virtual-Storage-Profile API
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Group FM_VSP_init_grp FM VSP Initialization Unit
-+
-+ @Description FM VSP initialization API.
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Description Virtual Storage Profile
-+*//***************************************************************************/
-+typedef struct t_FmVspParams {
-+ t_Handle h_Fm; /**< A handle to the FM object this VSP related to */
-+ t_FmExtPools extBufPools; /**< Which external buffer pools are used
-+ (up to FM_PORT_MAX_NUM_OF_EXT_POOLS), and their sizes.
-+ parameter associated with Rx / OP port */
-+ uint16_t liodnOffset; /**< VSP's LIODN offset */
-+ struct {
-+ e_FmPortType portType; /**< Port type */
-+ uint8_t portId; /**< Port Id - relative to type */
-+ } portParams;
-+ uint8_t relativeProfileId; /**< VSP Id - relative to VSP's range
-+ defined in relevant FM object */
-+} t_FmVspParams;
-+
-+
-+/**************************************************************************//**
-+ @Function FM_VSP_Config
-+
-+ @Description Creates descriptor for the FM VSP module.
-+
-+ The routine returns a handle (descriptor) to the FM VSP object.
-+ This descriptor must be passed as first parameter to all other
-+ FM VSP function calls.
-+
-+ No actual initialization or configuration of FM hardware is
-+ done by this routine.
-+
-+@Param[in] p_FmVspParams Pointer to data structure of parameters
-+
-+ @Retval Handle to FM VSP object, or NULL for Failure.
-+*//***************************************************************************/
-+t_Handle FM_VSP_Config(t_FmVspParams *p_FmVspParams);
-+
-+/**************************************************************************//**
-+ @Function FM_VSP_Init
-+
-+ @Description Initializes the FM VSP module
-+
-+ @Param[in] h_FmVsp - FM VSP module descriptor
-+
-+ @Return E_OK on success; Error code otherwise.
-+*//***************************************************************************/
-+t_Error FM_VSP_Init(t_Handle h_FmVsp);
-+
-+/**************************************************************************//**
-+ @Function FM_VSP_Free
-+
-+ @Description Frees all resources that were assigned to FM VSP module.
-+
-+ Calling this routine invalidates the descriptor.
-+
-+ @Param[in] h_FmVsp - FM VSP module descriptor
-+
-+ @Return E_OK on success; Error code otherwise.
-+*//***************************************************************************/
-+t_Error FM_VSP_Free(t_Handle h_FmVsp);
-+
-+
-+/**************************************************************************//**
-+ @Group FM_VSP_adv_config_grp FM VSP Advanced Configuration Unit
-+
-+ @Description FM VSP advanced configuration functions.
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Function FM_VSP_ConfigBufferPrefixContent
-+
-+ @Description Defines the structure, size and content of the application buffer.
-+
-+ The prefix will
-+ In VSPs defined for Tx ports, if 'passPrsResult', the application
-+ should set a value to their offsets in the prefix of
-+ the FM will save the first 'privDataSize', than,
-+ depending on 'passPrsResult' and 'passTimeStamp', copy parse result
-+ and timeStamp, and the packet itself (in this order), to the
-+ application buffer, and to offset.
-+
-+ Calling this routine changes the buffer margins definitions
-+ in the internal driver data base from its default
-+ configuration: Data size: [DEFAULT_FM_SP_bufferPrefixContent_privDataSize]
-+ Pass Parser result: [DEFAULT_FM_SP_bufferPrefixContent_passPrsResult].
-+ Pass timestamp: [DEFAULT_FM_SP_bufferPrefixContent_passTimeStamp].
-+
-+ @Param[in] h_FmVsp A handle to a FM VSP module.
-+ @Param[in,out] p_FmBufferPrefixContent A structure of parameters describing the
-+ structure of the buffer.
-+ Out parameter: Start margin - offset
-+ of data from start of external buffer.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_VSP_Config() and before FM_VSP_Init().
-+*//***************************************************************************/
-+t_Error FM_VSP_ConfigBufferPrefixContent(t_Handle h_FmVsp,
-+ t_FmBufferPrefixContent *p_FmBufferPrefixContent);
-+
-+/**************************************************************************//**
-+ @Function FM_VSP_ConfigDmaSwapData
-+
-+ @Description Calling this routine changes the DMA swap data parameter
-+ in the internal driver data base from its default
-+ configuration [DEFAULT_FM_SP_dmaSwapData]
-+
-+ @Param[in] h_FmVsp A handle to a FM VSP module.
-+ @Param[in] swapData New selection
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_VSP_Config() and before FM_VSP_Init().
-+*//***************************************************************************/
-+t_Error FM_VSP_ConfigDmaSwapData(t_Handle h_FmVsp, e_FmDmaSwapOption swapData);
-+
-+/**************************************************************************//**
-+ @Function FM_VSP_ConfigDmaIcCacheAttr
-+
-+ @Description Calling this routine changes the internal context cache
-+ attribute parameter in the internal driver data base
-+ from its default configuration [DEFAULT_FM_SP_dmaIntContextCacheAttr]
-+
-+ @Param[in] h_FmVsp A handle to a FM VSP module.
-+ @Param[in] intContextCacheAttr New selection
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_VSP_Config() and before FM_VSP_Init().
-+*//***************************************************************************/
-+t_Error FM_VSP_ConfigDmaIcCacheAttr(t_Handle h_FmVsp,
-+ e_FmDmaCacheOption intContextCacheAttr);
-+
-+/**************************************************************************//**
-+ @Function FM_VSP_ConfigDmaHdrAttr
-+
-+ @Description Calling this routine changes the header cache
-+ attribute parameter in the internal driver data base
-+ from its default configuration [DEFAULT_FM_SP_dmaHeaderCacheAttr]
-+
-+ @Param[in] h_FmVsp A handle to a FM VSP module.
-+ @Param[in] headerCacheAttr New selection
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_VSP_Config() and before FM_VSP_Init().
-+*//***************************************************************************/
-+t_Error FM_VSP_ConfigDmaHdrAttr(t_Handle h_FmVsp, e_FmDmaCacheOption headerCacheAttr);
-+
-+/**************************************************************************//**
-+ @Function FM_VSP_ConfigDmaScatterGatherAttr
-+
-+ @Description Calling this routine changes the scatter gather cache
-+ attribute parameter in the internal driver data base
-+ from its default configuration [DEFAULT_FM_SP_dmaScatterGatherCacheAttr]
-+
-+ @Param[in] h_FmVsp A handle to a FM VSP module.
-+ @Param[in] scatterGatherCacheAttr New selection
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_VSP_Config() and before FM_VSP_Init().
-+*//***************************************************************************/
-+t_Error FM_VSP_ConfigDmaScatterGatherAttr(t_Handle h_FmVsp,
-+ e_FmDmaCacheOption scatterGatherCacheAttr);
-+
-+/**************************************************************************//**
-+ @Function FM_VSP_ConfigDmaWriteOptimize
-+
-+ @Description Calling this routine changes the write optimization
-+ parameter in the internal driver data base
-+ from its default configuration: optimize = [DEFAULT_FM_SP_dmaWriteOptimize]
-+
-+ @Param[in] h_FmVsp A handle to a FM VSP module.
-+ @Param[in] optimize TRUE to enable optimization, FALSE for normal operation
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_VSP_Config() and before FM_VSP_Init().
-+*//***************************************************************************/
-+t_Error FM_VSP_ConfigDmaWriteOptimize(t_Handle h_FmVsp, bool optimize);
-+
-+/**************************************************************************//**
-+ @Function FM_VSP_ConfigNoScatherGather
-+
-+ @Description Calling this routine changes the possibility to receive S/G frame
-+ in the internal driver data base
-+ from its default configuration: optimize = [DEFAULT_FM_SP_noScatherGather]
-+
-+ @Param[in] h_FmVsp A handle to a FM VSP module.
-+ @Param[in] noScatherGather TRUE to operate without scatter/gather capability.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_VSP_Config() and before FM_VSP_Init().
-+*//***************************************************************************/
-+t_Error FM_VSP_ConfigNoScatherGather(t_Handle h_FmVsp, bool noScatherGather);
-+
-+/**************************************************************************//**
-+ @Function FM_VSP_ConfigPoolDepletion
-+
-+ @Description Calling this routine enables pause frame generation depending on the
-+ depletion status of BM pools. It also defines the conditions to activate
-+ this functionality. By default, this functionality is disabled.
-+
-+ @Param[in] h_FmVsp A handle to a FM VSP module.
-+ @Param[in] p_BufPoolDepletion A structure of pool depletion parameters
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_VSP_Config() and before FM_VSP_Init().
-+*//***************************************************************************/
-+t_Error FM_VSP_ConfigPoolDepletion(t_Handle h_FmVsp, t_FmBufPoolDepletion *p_BufPoolDepletion);
-+
-+/**************************************************************************//**
-+ @Function FM_VSP_ConfigBackupPools
-+
-+ @Description Calling this routine allows the configuration of some of the BM pools
-+ defined for this port as backup pools.
-+ A pool configured to be a backup pool will be used only if all other
-+ enabled non-backup pools are depleted.
-+
-+ @Param[in] h_FmVsp A handle to a FM VSP module.
-+ @Param[in] p_BackupBmPools An array of pool id's. All pools specified here will
-+ be defined as backup pools.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_VSP_Config() and before FM_VSP_Init().
-+*//***************************************************************************/
-+t_Error FM_VSP_ConfigBackupPools(t_Handle h_FmVsp, t_FmBackupBmPools *p_BackupBmPools);
-+
-+/** @} */ /* end of FM_VSP_adv_config_grp group */
-+/** @} */ /* end of FM_VSP_init_grp group */
-+
-+
-+/**************************************************************************//**
-+ @Group FM_VSP_control_grp FM VSP Control Unit
-+
-+ @Description FM VSP runtime control API.
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Function FM_VSP_GetBufferDataOffset
-+
-+ @Description Relevant for Rx ports.
-+ Returns the data offset from the beginning of the data buffer
-+
-+ @Param[in] h_FmVsp - FM PORT module descriptor
-+
-+ @Return data offset.
-+
-+ @Cautions Allowed only following FM_VSP_Init().
-+*//***************************************************************************/
-+uint32_t FM_VSP_GetBufferDataOffset(t_Handle h_FmVsp);
-+
-+/**************************************************************************//**
-+ @Function FM_VSP_GetBufferICInfo
-+
-+ @Description Returns the Internal Context offset from the beginning of the data buffer
-+
-+ @Param[in] h_FmVsp - FM PORT module descriptor
-+ @Param[in] p_Data - A pointer to the data buffer.
-+
-+ @Return Internal context info pointer on success, NULL if 'allOtherInfo' was not
-+ configured for this port.
-+
-+ @Cautions Allowed only following FM_VSP_Init().
-+*//***************************************************************************/
-+uint8_t * FM_VSP_GetBufferICInfo(t_Handle h_FmVsp, char *p_Data);
-+
-+/**************************************************************************//**
-+ @Function FM_VSP_GetBufferPrsResult
-+
-+ @Description Returns the pointer to the parse result in the data buffer.
-+ In Rx ports this is relevant after reception, if parse
-+ result is configured to be part of the data passed to the
-+ application. For non Rx ports it may be used to get the pointer
-+ of the area in the buffer where parse result should be
-+ initialized - if so configured.
-+ See FM_VSP_ConfigBufferPrefixContent for data buffer prefix
-+ configuration.
-+
-+ @Param[in] h_FmVsp - FM PORT module descriptor
-+ @Param[in] p_Data - A pointer to the data buffer.
-+
-+ @Return Parse result pointer on success, NULL if parse result was not
-+ configured for this port.
-+
-+ @Cautions Allowed only following FM_VSP_Init().
-+*//***************************************************************************/
-+t_FmPrsResult * FM_VSP_GetBufferPrsResult(t_Handle h_FmVsp, char *p_Data);
-+
-+/**************************************************************************//**
-+ @Function FM_VSP_GetBufferTimeStamp
-+
-+ @Description Returns the time stamp in the data buffer.
-+ Relevant for Rx ports for getting the buffer time stamp.
-+ See FM_VSP_ConfigBufferPrefixContent for data buffer prefix
-+ configuration.
-+
-+ @Param[in] h_FmVsp - FM PORT module descriptor
-+ @Param[in] p_Data - A pointer to the data buffer.
-+
-+ @Return A pointer to the hash result on success, NULL otherwise.
-+
-+ @Cautions Allowed only following FM_VSP_Init().
-+*//***************************************************************************/
-+uint64_t * FM_VSP_GetBufferTimeStamp(t_Handle h_FmVsp, char *p_Data);
-+
-+/**************************************************************************//**
-+ @Function FM_VSP_GetBufferHashResult
-+
-+ @Description Given a data buffer, on the condition that hash result was defined
-+ as a part of the buffer content (see FM_VSP_ConfigBufferPrefixContent)
-+ this routine will return the pointer to the hash result location in the
-+ buffer prefix.
-+
-+ @Param[in] h_FmVsp - FM PORT module descriptor
-+ @Param[in] p_Data - A pointer to the data buffer.
-+
-+ @Return A pointer to the hash result on success, NULL otherwise.
-+
-+ @Cautions Allowed only following FM_VSP_Init().
-+*//***************************************************************************/
-+uint8_t * FM_VSP_GetBufferHashResult(t_Handle h_FmVsp, char *p_Data);
-+
-+
-+/** @} */ /* end of FM_VSP_control_grp group */
-+/** @} */ /* end of FM_VSP_grp group */
-+/** @} */ /* end of FM_grp group */
-+
-+
-+#endif /* __FM_VSP_EXT_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/Peripherals/mii_acc_ext.h
-@@ -0,0 +1,76 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+
-+#ifndef __MII_ACC_EXT_H
-+#define __MII_ACC_EXT_H
-+
-+
-+/**************************************************************************//**
-+ @Function MII_ReadPhyReg
-+
-+ @Description This routine is called to read a specified PHY
-+ register value.
-+
-+ @Param[in] h_MiiAccess - Handle to MII configuration access registers
-+ @Param[in] phyAddr - PHY address (0-31).
-+ @Param[in] reg - PHY register to read
-+ @Param[out] p_Data - Gets the register value.
-+
-+ @Return Always zero (success).
-+*//***************************************************************************/
-+int MII_ReadPhyReg(t_Handle h_MiiAccess,
-+ uint8_t phyAddr,
-+ uint8_t reg,
-+ uint16_t *p_Data);
-+
-+/**************************************************************************//**
-+ @Function MII_WritePhyReg
-+
-+ @Description This routine is called to write data to a specified PHY
-+ register.
-+
-+ @Param[in] h_MiiAccess - Handle to MII configuration access registers
-+ @Param[in] phyAddr - PHY address (0-31).
-+ @Param[in] reg - PHY register to write
-+ @Param[in] data - Data to write in register.
-+
-+ @Return Always zero (success).
-+*//***************************************************************************/
-+int MII_WritePhyReg(t_Handle h_MiiAccess,
-+ uint8_t phyAddr,
-+ uint8_t reg,
-+ uint16_t data);
-+
-+
-+#endif /* __MII_ACC_EXT_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/core_ext.h
-@@ -0,0 +1,90 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/**************************************************************************//**
-+ @File core_ext.h
-+
-+ @Description Generic interface to basic core operations.
-+
-+ The system integrator must ensure that this interface is
-+ mapped to a specific core implementation, by including the
-+ appropriate header file.
-+*//***************************************************************************/
-+#ifndef __CORE_EXT_H
-+#define __CORE_EXT_H
-+
-+#ifdef CONFIG_FMAN_ARM
-+#include "arm_ext.h"
-+#include <linux/smp.h>
-+#else
-+#ifdef NCSW_PPC_CORE
-+#include "ppc_ext.h"
-+#elif defined(NCSW_VXWORKS)
-+#include "core_vxw_ext.h"
-+#else
-+#error "Core is not defined!"
-+#endif /* NCSW_CORE */
-+
-+#if (!defined(CORE_IS_LITTLE_ENDIAN) && !defined(CORE_IS_BIG_ENDIAN))
-+#error "Must define core as little-endian or big-endian!"
-+#endif /* (!defined(CORE_IS_LITTLE_ENDIAN) && ... */
-+
-+#ifndef CORE_CACHELINE_SIZE
-+#error "Must define the core cache-line size!"
-+#endif /* !CORE_CACHELINE_SIZE */
-+
-+#endif /* CONFIG_FMAN_ARM */
-+
-+
-+/**************************************************************************//**
-+ @Function CORE_GetId
-+
-+ @Description Returns the core ID in the system.
-+
-+ @Return Core ID.
-+*//***************************************************************************/
-+uint32_t CORE_GetId(void);
-+
-+/**************************************************************************//**
-+ @Function CORE_MemoryBarrier
-+
-+ @Description This routine will cause the core to stop executing any commands
-+ until all previous memory read/write commands are completely out
-+ of the core's pipeline.
-+
-+ @Return None.
-+*//***************************************************************************/
-+void CORE_MemoryBarrier(void);
-+#define fsl_mem_core_barrier() CORE_MemoryBarrier()
-+
-+#endif /* __CORE_EXT_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/cores/arm_ext.h
-@@ -0,0 +1,55 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/**************************************************************************//**
-+ @File arm_ext.h
-+
-+ @Description Core API for ARM cores
-+
-+ These routines must be implemented by each specific PowerPC
-+ core driver.
-+*//***************************************************************************/
-+#ifndef __ARM_EXT_H
-+#define __ARM_EXT_H
-+
-+#include "part_ext.h"
-+
-+
-+#define CORE_IS_LITTLE_ENDIAN
-+
-+static __inline__ void CORE_MemoryBarrier(void)
-+{
-+ mb();
-+}
-+
-+#endif /* __PPC_EXT_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/cores/e500v2_ext.h
-@@ -0,0 +1,476 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/**************************************************************************//**
-+ @File e500v2_ext.h
-+
-+ @Description E500 external definitions prototypes
-+ This file is not included by the E500
-+ source file as it is an assembly file. It is used
-+ only for prototypes exposure, for inclusion
-+ by user and other modules.
-+*//***************************************************************************/
-+
-+#ifndef __E500V2_EXT_H
-+#define __E500V2_EXT_H
-+
-+#include "std_ext.h"
-+
-+
-+/* Layer 1 Cache Manipulations
-+ *==============================
-+ * Should not be called directly by the user.
-+ */
-+void L1DCache_Invalidate (void);
-+void L1ICache_Invalidate(void);
-+void L1DCache_Enable(void);
-+void L1ICache_Enable(void);
-+void L1DCache_Disable(void);
-+void L1ICache_Disable(void);
-+void L1DCache_Flush(void);
-+void L1ICache_Flush(void);
-+uint32_t L1ICache_IsEnabled(void);
-+uint32_t L1DCache_IsEnabled(void);
-+/*
-+ *
-+ */
-+uint32_t L1DCache_LineLock(uint32_t addr);
-+uint32_t L1ICache_LineLock(uint32_t addr);
-+void L1Cache_BroadCastEnable(void);
-+void L1Cache_BroadCastDisable(void);
-+
-+
-+#define CORE_DCacheEnable E500_DCacheEnable
-+#define CORE_ICacheEnable E500_ICacheEnable
-+#define CORE_DCacheDisable E500_DCacheDisable
-+#define CORE_ICacheDisable E500_ICacheDisable
-+#define CORE_GetId E500_GetId
-+#define CORE_TestAndSet E500_TestAndSet
-+#define CORE_MemoryBarrier E500_MemoryBarrier
-+#define CORE_InstructionSync E500_InstructionSync
-+
-+#define CORE_SetDozeMode E500_SetDozeMode
-+#define CORE_SetNapMode E500_SetNapMode
-+#define CORE_SetSleepMode E500_SetSleepMode
-+#define CORE_SetJogMode E500_SetJogMode
-+#define CORE_SetDeepSleepMode E500_SetDeepSleepMode
-+
-+#define CORE_RecoverDozeMode E500_RecoverDozeMode
-+#define CORE_RecoverNapMode E500_RecoverNapMode
-+#define CORE_RecoverSleepMode E500_RecoverSleepMode
-+#define CORE_RecoverJogMode E500_RecoverJogMode
-+
-+void E500_SetDozeMode(void);
-+void E500_SetNapMode(void);
-+void E500_SetSleepMode(void);
-+void E500_SetJogMode(void);
-+t_Error E500_SetDeepSleepMode(uint32_t bptrAddress);
-+
-+void E500_RecoverDozeMode(void);
-+void E500_RecoverNapMode(void);
-+void E500_RecoverSleepMode(void);
-+void E500_RecoverJogMode(void);
-+
-+
-+/**************************************************************************//**
-+ @Group E500_id E500 Application Programming Interface
-+
-+ @Description E500 API functions, definitions and enums
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Group E500_init_grp E500 Initialization Unit
-+
-+ @Description E500 initialization unit API functions, definitions and enums
-+
-+ @{
-+*//***************************************************************************/
-+
-+
-+/**************************************************************************//**
-+ @Function E500_DCacheEnable
-+
-+ @Description Enables the data cache for memory pages that are
-+ not cache inhibited.
-+
-+ @Return None.
-+*//***************************************************************************/
-+void E500_DCacheEnable(void);
-+
-+/**************************************************************************//**
-+ @Function E500_ICacheEnable
-+
-+ @Description Enables the instruction cache for memory pages that are
-+ not cache inhibited.
-+
-+ @Return None.
-+*//***************************************************************************/
-+void E500_ICacheEnable(void);
-+
-+/**************************************************************************//**
-+ @Function E500_DCacheDisable
-+
-+ @Description Disables the data cache.
-+
-+ @Return None.
-+*//***************************************************************************/
-+void E500_DCacheDisable(void);
-+
-+/**************************************************************************//**
-+ @Function E500_ICacheDisable
-+
-+ @Description Disables the instruction cache.
-+
-+ @Return None.
-+*//***************************************************************************/
-+void E500_ICacheDisable(void);
-+
-+/**************************************************************************//**
-+ @Function E500_DCacheFlush
-+
-+ @Description Flushes the data cache
-+
-+ @Return None.
-+*//***************************************************************************/
-+void E500_DCacheFlush(void);
-+
-+/**************************************************************************//**
-+ @Function E500_ICacheFlush
-+
-+ @Description Flushes the instruction cache.
-+
-+ @Return None.
-+*//***************************************************************************/
-+void E500_ICacheFlush(void);
-+
-+/**************************************************************************//**
-+ @Function E500_DCacheSetStashId
-+
-+ @Description Set Stash Id for data cache
-+
-+ @Param[in] stashId the stash id to be set.
-+
-+ @Return None.
-+*//***************************************************************************/
-+void E500_DCacheSetStashId(uint8_t stashId);
-+
-+/**************************************************************************//**
-+ @Description E500mc L2 Cache Operation Mode
-+*//***************************************************************************/
-+typedef enum e_E500mcL2CacheMode
-+{
-+ e_L2_CACHE_MODE_DATA_ONLY = 0x00000001, /**< Cache data only */
-+ e_L2_CACHE_MODE_INST_ONLY = 0x00000002, /**< Cache instructions only */
-+ e_L2_CACHE_MODE_DATA_AND_INST = 0x00000003 /**< Cache data and instructions */
-+} e_E500mcL2CacheMode;
-+
-+#if defined(CORE_E500MC) || defined(CORE_E5500)
-+/**************************************************************************//**
-+ @Function E500_L2CacheEnable
-+
-+ @Description Enables the cache for memory pages that are not cache inhibited.
-+
-+ @param[in] mode - L2 cache mode: data only, instruction only or instruction and data.
-+
-+ @Return None.
-+
-+ @Cautions This routine must be call only ONCE for both caches. I.e. it is
-+ not possible to call this routine for i-cache and than to call
-+ again for d-cache; The second call will override the first one.
-+*//***************************************************************************/
-+void E500_L2CacheEnable(e_E500mcL2CacheMode mode);
-+
-+/**************************************************************************//**
-+ @Function E500_L2CacheDisable
-+
-+ @Description Disables the cache (data instruction or both).
-+
-+ @Return None.
-+
-+*//***************************************************************************/
-+void E500_L2CacheDisable(void);
-+
-+/**************************************************************************//**
-+ @Function E500_L2CacheFlush
-+
-+ @Description Flushes the cache.
-+
-+ @Return None.
-+*//***************************************************************************/
-+void E500_L2CacheFlush(void);
-+
-+/**************************************************************************//**
-+ @Function E500_L2SetStashId
-+
-+ @Description Set Stash Id
-+
-+ @Param[in] stashId the stash id to be set.
-+
-+ @Return None.
-+*//***************************************************************************/
-+void E500_L2SetStashId(uint8_t stashId);
-+#endif /* defined(CORE_E500MC) || defined(CORE_E5500) */
-+
-+#ifdef CORE_E6500
-+/**************************************************************************//**
-+ @Function E6500_L2CacheEnable
-+
-+ @Description Enables the cache for memory pages that are not cache inhibited.
-+
-+ @param[in] mode - L2 cache mode: support data & instruction only.
-+
-+ @Return None.
-+
-+ @Cautions This routine must be call only ONCE for both caches. I.e. it is
-+ not possible to call this routine for i-cache and than to call
-+ again for d-cache; The second call will override the first one.
-+*//***************************************************************************/
-+void E6500_L2CacheEnable(uintptr_t clusterBase);
-+
-+/**************************************************************************//**
-+ @Function E6500_L2CacheDisable
-+
-+ @Description Disables the cache (data instruction or both).
-+
-+ @Return None.
-+
-+*//***************************************************************************/
-+void E6500_L2CacheDisable(uintptr_t clusterBase);
-+
-+/**************************************************************************//**
-+ @Function E6500_L2CacheFlush
-+
-+ @Description Flushes the cache.
-+
-+ @Return None.
-+*//***************************************************************************/
-+void E6500_L2CacheFlush(uintptr_t clusterBase);
-+
-+/**************************************************************************//**
-+ @Function E6500_L2SetStashId
-+
-+ @Description Set Stash Id
-+
-+ @Param[in] stashId the stash id to be set.
-+
-+ @Return None.
-+*//***************************************************************************/
-+void E6500_L2SetStashId(uintptr_t clusterBase, uint8_t stashId);
-+
-+/**************************************************************************//**
-+ @Function E6500_GetCcsrBase
-+
-+ @Description Obtain SoC CCSR base address
-+
-+ @Param[in] None.
-+
-+ @Return Physical CCSR base address.
-+*//***************************************************************************/
-+physAddress_t E6500_GetCcsrBase(void);
-+#endif /* CORE_E6500 */
-+
-+/**************************************************************************//**
-+ @Function E500_AddressBusStreamingEnable
-+
-+ @Description Enables address bus streaming on the CCB.
-+
-+ This setting, along with the ECM streaming configuration
-+ parameters, enables address bus streaming on the CCB.
-+
-+ @Return None.
-+*//***************************************************************************/
-+void E500_AddressBusStreamingEnable(void);
-+
-+/**************************************************************************//**
-+ @Function E500_AddressBusStreamingDisable
-+
-+ @Description Disables address bus streaming on the CCB.
-+
-+ @Return None.
-+*//***************************************************************************/
-+void E500_AddressBusStreamingDisable(void);
-+
-+/**************************************************************************//**
-+ @Function E500_AddressBroadcastEnable
-+
-+ @Description Enables address broadcast.
-+
-+ The e500 broadcasts cache management instructions (dcbst, dcblc
-+ (CT = 1), icblc (CT = 1), dcbf, dcbi, mbar, msync, tlbsync, icbi)
-+ based on ABE. ABE must be set to allow management of external
-+ L2 caches.
-+
-+ @Return None.
-+*//***************************************************************************/
-+void E500_AddressBroadcastEnable(void);
-+
-+/**************************************************************************//**
-+ @Function E500_AddressBroadcastDisable
-+
-+ @Description Disables address broadcast.
-+
-+ The e500 broadcasts cache management instructions (dcbst, dcblc
-+ (CT = 1), icblc (CT = 1), dcbf, dcbi, mbar, msync, tlbsync, icbi)
-+ based on ABE. ABE must be set to allow management of external
-+ L2 caches.
-+
-+ @Return None.
-+*//***************************************************************************/
-+void E500_AddressBroadcastDisable(void);
-+
-+/**************************************************************************//**
-+ @Function E500_IsTaskletSupported
-+
-+ @Description Checks if tasklets are supported by the e500 interrupt handler.
-+
-+ @Retval TRUE - Tasklets are supported.
-+ @Retval FALSE - Tasklets are not supported.
-+*//***************************************************************************/
-+bool E500_IsTaskletSupported(void);
-+
-+void E500_EnableTimeBase(void);
-+void E500_DisableTimeBase(void);
-+
-+uint64_t E500_GetTimeBaseTime(void);
-+
-+void E500_GenericIntrInit(void);
-+
-+t_Error E500_SetIntr(int ppcIntrSrc,
-+ void (* Isr)(t_Handle handle),
-+ t_Handle handle);
-+
-+t_Error E500_ClearIntr(int ppcIntrSrc);
-+
-+/**************************************************************************//**
-+ @Function E500_GenericIntrHandler
-+
-+ @Description This is the general e500 interrupt handler.
-+
-+ It is called by the main assembly interrupt handler
-+ when an exception occurs and no other function has been
-+ assigned to this exception.
-+
-+ @Param intrEntry - (In) The exception interrupt vector entry.
-+*//***************************************************************************/
-+void E500_GenericIntrHandler(uint32_t intrEntry);
-+
-+/**************************************************************************//**
-+ @Function CriticalIntr
-+
-+ @Description This is the specific critical e500 interrupt handler.
-+
-+ It is called by the main assembly interrupt handler
-+ when an critical interrupt.
-+
-+ @Param intrEntry - (In) The exception interrupt vector entry.
-+*//***************************************************************************/
-+void CriticalIntr(uint32_t intrEntry);
-+
-+
-+/**************************************************************************//**
-+ @Function E500_GetId
-+
-+ @Description Returns the core ID in the system.
-+
-+ @Return Core ID.
-+*//***************************************************************************/
-+uint32_t E500_GetId(void);
-+
-+/**************************************************************************//**
-+ @Function E500_TestAndSet
-+
-+ @Description This routine tries to atomically test-and-set an integer
-+ in memory to a non-zero value.
-+
-+ The memory will be set only if it is tested as zero, in which
-+ case the routine returns the new non-zero value; otherwise the
-+ routine returns zero.
-+
-+ @Param[in] p - pointer to a volatile int in memory, on which test-and-set
-+ operation should be made.
-+
-+ @Retval Zero - Operation failed - memory was already set.
-+ @Retval Non-zero - Operation succeeded - memory has been set.
-+*//***************************************************************************/
-+int E500_TestAndSet(volatile int *p);
-+
-+/**************************************************************************//**
-+ @Function E500_MemoryBarrier
-+
-+ @Description This routine will cause the core to stop executing any commands
-+ until all previous memory read/write commands are completely out
-+ of the core's pipeline.
-+
-+ @Return None.
-+*//***************************************************************************/
-+static __inline__ void E500_MemoryBarrier(void)
-+{
-+#ifndef CORE_E500V2
-+ __asm__ ("mbar 1");
-+#else /* CORE_E500V2 */
-+ /**** ERRATA WORK AROUND START ****/
-+ /* ERRATA num: CPU1 */
-+ /* Description: "mbar MO = 1" instruction fails to order caching-inhibited
-+ guarded loads and stores. */
-+
-+ /* "msync" instruction is used instead */
-+
-+ __asm__ ("msync");
-+
-+ /**** ERRATA WORK AROUND END ****/
-+#endif /* CORE_E500V2 */
-+}
-+
-+/**************************************************************************//**
-+ @Function E500_InstructionSync
-+
-+ @Description This routine will cause the core to wait for previous instructions
-+ (including any interrupts they generate) to complete before the
-+ synchronization command executes, which purges all instructions
-+ from the processor's pipeline and refetches the next instruction.
-+
-+ @Return None.
-+*//***************************************************************************/
-+static __inline__ void E500_InstructionSync(void)
-+{
-+ __asm__ ("isync");
-+}
-+
-+
-+/** @} */ /* end of E500_init_grp group */
-+/** @} */ /* end of E500_grp group */
-+
-+
-+#endif /* __E500V2_EXT_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/cores/ppc_ext.h
-@@ -0,0 +1,141 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/**************************************************************************//**
-+ @File ppc_ext.h
-+
-+ @Description Core API for PowerPC cores
-+
-+ These routines must be implemented by each specific PowerPC
-+ core driver.
-+*//***************************************************************************/
-+#ifndef __PPC_EXT_H
-+#define __PPC_EXT_H
-+
-+#include "part_ext.h"
-+
-+
-+#define CORE_IS_BIG_ENDIAN
-+
-+#if defined(CORE_E300) || defined(CORE_E500V2)
-+#define CORE_CACHELINE_SIZE 32
-+#elif defined(CORE_E500MC) || defined(CORE_E5500) || defined(CORE_E6500)
-+#define CORE_CACHELINE_SIZE 64
-+#else
-+#error "Core not defined!"
-+#endif /* defined(CORE_E300) || ... */
-+
-+
-+/**************************************************************************//**
-+ @Function CORE_TestAndSet
-+
-+ @Description This routine tries to atomically test-and-set an integer
-+ in memory to a non-zero value.
-+
-+ The memory will be set only if it is tested as zero, in which
-+ case the routine returns the new non-zero value; otherwise the
-+ routine returns zero.
-+
-+ @Param[in] p - pointer to a volatile int in memory, on which test-and-set
-+ operation should be made.
-+
-+ @Retval Zero - Operation failed - memory was already set.
-+ @Retval Non-zero - Operation succeeded - memory has been set.
-+*//***************************************************************************/
-+int CORE_TestAndSet(volatile int *p);
-+
-+/**************************************************************************//**
-+ @Function CORE_InstructionSync
-+
-+ @Description This routine will cause the core to wait for previous instructions
-+ (including any interrupts they generate) to complete before the
-+ synchronization command executes, which purges all instructions
-+ from the processor's pipeline and refetches the next instruction.
-+
-+ @Return None.
-+*//***************************************************************************/
-+void CORE_InstructionSync(void);
-+
-+/**************************************************************************//**
-+ @Function CORE_DCacheEnable
-+
-+ @Description Enables the data cache for memory pages that are
-+ not cache inhibited.
-+
-+ @Return None.
-+*//***************************************************************************/
-+void CORE_DCacheEnable(void);
-+
-+/**************************************************************************//**
-+ @Function CORE_ICacheEnable
-+
-+ @Description Enables the instruction cache for memory pages that are
-+ not cache inhibited.
-+
-+ @Return None.
-+*//***************************************************************************/
-+void CORE_ICacheEnable(void);
-+
-+/**************************************************************************//**
-+ @Function CORE_DCacheDisable
-+
-+ @Description Disables the data cache.
-+
-+ @Return None.
-+*//***************************************************************************/
-+void CORE_DCacheDisable(void);
-+
-+/**************************************************************************//**
-+ @Function CORE_ICacheDisable
-+
-+ @Description Disables the instruction cache.
-+
-+ @Return None.
-+*//***************************************************************************/
-+void CORE_ICacheDisable(void);
-+
-+
-+
-+#if defined(CORE_E300)
-+#include "e300_ext.h"
-+#elif defined(CORE_E500V2) || defined(CORE_E500MC) || defined(CORE_E5500) || defined(CORE_E6500)
-+#include "e500v2_ext.h"
-+#if !defined(NCSW_LINUX)
-+#include "e500v2_asm_ext.h"
-+#endif
-+#else
-+#error "Core not defined!"
-+#endif
-+
-+
-+#endif /* __PPC_EXT_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/ddr_std_ext.h
-@@ -0,0 +1,77 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#ifndef __DDR_SDT_EXT_H
-+#define __DDR_SDT_EXT_H
-+
-+
-+/**************************************************************************//**
-+ @Group ddr_Generic_Resources
-+
-+ @Description ddr generic functions, definitions and enums.
-+
-+ @{
-+*//***************************************************************************/
-+
-+
-+/**************************************************************************//**
-+ @Description SPD maximum size
-+*//***************************************************************************/
-+#define SPD_MAX_SIZE 256
-+
-+/**************************************************************************//**
-+ @Description DDR types select
-+*//***************************************************************************/
-+typedef enum e_DdrType
-+{
-+ e_DDR_DDR1,
-+ e_DDR_DDR2,
-+ e_DDR_DDR3,
-+ e_DDR_DDR3L,
-+ e_DDR_DDR4
-+} e_DdrType;
-+
-+/**************************************************************************//**
-+ @Description DDR Mode.
-+*//***************************************************************************/
-+typedef enum e_DdrMode
-+{
-+ e_DDR_BUS_WIDTH_32BIT,
-+ e_DDR_BUS_WIDTH_64BIT
-+} e_DdrMode;
-+
-+/** @} */ /* end of ddr_Generic_Resources group */
-+
-+
-+
-+#endif /* __DDR_SDT_EXT_H */
-+
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/debug_ext.h
-@@ -0,0 +1,233 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/**************************************************************************//**
-+ @File debug_ext.h
-+
-+ @Description Debug mode definitions.
-+*//***************************************************************************/
-+
-+#ifndef __DEBUG_EXT_H
-+#define __DEBUG_EXT_H
-+
-+#include "std_ext.h"
-+#include "xx_ext.h"
-+#include "memcpy_ext.h"
-+#if (DEBUG_ERRORS > 0)
-+#include "sprint_ext.h"
-+#include "string_ext.h"
-+#endif /* DEBUG_ERRORS > 0 */
-+
-+
-+#if (DEBUG_ERRORS > 0)
-+
-+/* Internally used macros */
-+
-+#define DUMP_Print XX_Print
-+#define DUMP_MAX_LEVELS 6
-+#define DUMP_IDX_LEN 6
-+#define DUMP_MAX_STR 64
-+
-+
-+#define _CREATE_DUMP_SUBSTR(phrase) \
-+ dumpTmpLevel = 0; dumpSubStr[0] = '\0'; \
-+ snprintf(dumpTmpStr, DUMP_MAX_STR, "%s", #phrase); \
-+ p_DumpToken = strtok(dumpTmpStr, (dumpIsArr[0] ? "[" : ".")); \
-+ while ((p_DumpToken != NULL) && (dumpTmpLevel < DUMP_MAX_LEVELS)) \
-+ { \
-+ strlcat(dumpSubStr, p_DumpToken, DUMP_MAX_STR); \
-+ if (dumpIsArr[dumpTmpLevel]) \
-+ { \
-+ strlcat(dumpSubStr, dumpIdxStr[dumpTmpLevel], DUMP_MAX_STR); \
-+ p_DumpToken = strtok(NULL, "."); \
-+ } \
-+ if ((p_DumpToken != NULL) && \
-+ ((p_DumpToken = strtok(NULL, (dumpIsArr[++dumpTmpLevel] ? "[" : "."))) != NULL)) \
-+ strlcat(dumpSubStr, ".", DUMP_MAX_STR); \
-+ }
-+
-+
-+/**************************************************************************//**
-+ @Group gen_id General Drivers Utilities
-+
-+ @Description External routines.
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Group dump_id Memory and Registers Dump Mechanism
-+
-+ @Description Macros for dumping memory mapped structures.
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Description Declaration of dump mechanism variables.
-+
-+ This macro must be declared at the beginning of each routine
-+ which uses the dump mechanism macros, before the routine's code
-+ starts.
-+*//***************************************************************************/
-+#define DECLARE_DUMP \
-+ char dumpIdxStr[DUMP_MAX_LEVELS + 1][DUMP_IDX_LEN] = { "", }; \
-+ char dumpSubStr[DUMP_MAX_STR] = ""; \
-+ char dumpTmpStr[DUMP_MAX_STR] = ""; \
-+ char *p_DumpToken = NULL; \
-+ int dumpArrIdx = 0, dumpArrSize = 0, dumpLevel = 0, dumpTmpLevel = 0; \
-+ uint8_t dumpIsArr[DUMP_MAX_LEVELS + 1] = { 0 }; \
-+ /* Prevent warnings if not all used */ \
-+ UNUSED(dumpIdxStr[0][0]); \
-+ UNUSED(dumpSubStr[0]); \
-+ UNUSED(dumpTmpStr[0]); \
-+ UNUSED(p_DumpToken); \
-+ UNUSED(dumpArrIdx); \
-+ UNUSED(dumpArrSize); \
-+ UNUSED(dumpLevel); \
-+ UNUSED(dumpTmpLevel); \
-+ UNUSED(dumpIsArr[0]);
-+
-+
-+/**************************************************************************//**
-+ @Description Prints a title for a subsequent dumped structure or memory.
-+
-+ The inputs for this macro are the structure/memory title and
-+ its base addresses.
-+*//***************************************************************************/
-+#define DUMP_TITLE(addr, msg) \
-+ DUMP_Print("\r\n"); DUMP_Print msg; \
-+ if (addr) \
-+ DUMP_Print(" (%p)", (addr)); \
-+ DUMP_Print("\r\n---------------------------------------------------------\r\n");
-+
-+/**************************************************************************//**
-+ @Description Prints a subtitle for a subsequent dumped sub-structure (optional).
-+
-+ The inputs for this macro are the sub-structure subtitle.
-+ A separating line with this subtitle will be printed.
-+*//***************************************************************************/
-+#define DUMP_SUBTITLE(subtitle) \
-+ DUMP_Print("----------- "); DUMP_Print subtitle; DUMP_Print("\r\n")
-+
-+
-+/**************************************************************************//**
-+ @Description Dumps a memory region in 4-bytes aligned format.
-+
-+ The inputs for this macro are the base addresses and size
-+ (in bytes) of the memory region.
-+*//***************************************************************************/
-+#define DUMP_MEMORY(addr, size) \
-+ MemDisp((uint8_t *)(addr), (int)(size))
-+
-+
-+/**************************************************************************//**
-+ @Description Declares a dump loop, for dumping a sub-structure array.
-+
-+ The inputs for this macro are:
-+ - idx: an index variable, for indexing the sub-structure items
-+ inside the loop. This variable must be declared separately
-+ in the beginning of the routine.
-+ - cnt: the number of times to repeat the loop. This number should
-+ equal the number of items in the sub-structures array.
-+
-+ Note, that the body of the loop must be written inside brackets.
-+*//***************************************************************************/
-+#define DUMP_SUBSTRUCT_ARRAY(idx, cnt) \
-+ for (idx=0, dumpIsArr[dumpLevel++] = 1; \
-+ (idx < cnt) && (dumpLevel > 0) && snprintf(dumpIdxStr[dumpLevel-1], DUMP_IDX_LEN, "[%d]", idx); \
-+ idx++, ((idx < cnt) || (dumpIsArr[--dumpLevel] = 0)))
-+
-+
-+/**************************************************************************//**
-+ @Description Dumps a structure's member variable.
-+
-+ The input for this macro is the full reference for the member
-+ variable, where the structure is referenced using a pointer.
-+
-+ Note, that a members array must be dumped using DUMP_ARR macro,
-+ rather than using this macro.
-+
-+ If the member variable is part of a sub-structure hierarchy,
-+ the full hierarchy (including array indexing) must be specified.
-+
-+ Examples: p_Struct->member
-+ p_Struct->sub.member
-+ p_Struct->sub[i].member
-+*//***************************************************************************/
-+#define DUMP_VAR(st, phrase) \
-+ do { \
-+ void *addr = (void *)&((st)->phrase); \
-+ physAddress_t physAddr = XX_VirtToPhys(addr); \
-+ _CREATE_DUMP_SUBSTR(phrase); \
-+ DUMP_Print("0x%010llX: 0x%08x%8s\t%s\r\n", \
-+ physAddr, GET_UINT32(*(uint32_t*)addr), "", dumpSubStr); \
-+ } while (0)
-+
-+
-+/**************************************************************************//**
-+ @Description Dumps a structure's members array.
-+
-+ The input for this macro is the full reference for the members
-+ array, where the structure is referenced using a pointer.
-+
-+ If the members array is part of a sub-structure hierarchy,
-+ the full hierarchy (including array indexing) must be specified.
-+
-+ Examples: p_Struct->array
-+ p_Struct->sub.array
-+ p_Struct->sub[i].array
-+*//***************************************************************************/
-+#define DUMP_ARR(st, phrase) \
-+ do { \
-+ physAddress_t physAddr; \
-+ _CREATE_DUMP_SUBSTR(phrase); \
-+ dumpArrSize = ARRAY_SIZE((st)->phrase); \
-+ for (dumpArrIdx=0; dumpArrIdx < dumpArrSize; dumpArrIdx++) { \
-+ physAddr = XX_VirtToPhys((void *)&((st)->phrase[dumpArrIdx])); \
-+ DUMP_Print("0x%010llX: 0x%08x%8s\t%s[%d]\r\n", \
-+ physAddr, GET_UINT32((st)->phrase[dumpArrIdx]), "", dumpSubStr, dumpArrIdx); \
-+ } \
-+ } while (0)
-+
-+
-+
-+#endif /* DEBUG_ERRORS > 0 */
-+
-+
-+/** @} */ /* end of dump_id group */
-+/** @} */ /* end of gen_id group */
-+
-+
-+#endif /* __DEBUG_EXT_H */
-+
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/endian_ext.h
-@@ -0,0 +1,447 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/**************************************************************************//**
-+
-+ @File endian_ext.h
-+
-+ @Description Big/little endian swapping routines.
-+*//***************************************************************************/
-+
-+#ifndef __ENDIAN_EXT_H
-+#define __ENDIAN_EXT_H
-+
-+#include "std_ext.h"
-+
-+
-+/**************************************************************************//**
-+ @Group gen_id General Drivers Utilities
-+
-+ @Description General usage API. This API is intended for usage by both the
-+ internal modules and the user's application.
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Group endian_id Big/Little-Endian Conversion
-+
-+ @Description Routines and macros for Big/Little-Endian conversion and
-+ general byte swapping.
-+
-+ All routines and macros are expecting unsigned values as
-+ parameters, but will generate the correct result also for
-+ signed values. Therefore, signed/unsigned casting is allowed.
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Collection Byte-Swap Macros
-+
-+ Macros for swapping byte order.
-+
-+ @Cautions The parameters of these macros are evaluated multiple times.
-+ For calculated expressions or expressions that contain function
-+ calls it is recommended to use the byte-swap routines.
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Description Swaps the byte order of a given 16-bit value.
-+
-+ @Param[in] val - The 16-bit value to swap.
-+
-+ @Return The byte-swapped value..
-+
-+ @Cautions The given value is evaluated multiple times by this macro.
-+ For calculated expressions or expressions that contain function
-+ calls it is recommended to use the SwapUint16() routine.
-+
-+ @hideinitializer
-+*//***************************************************************************/
-+#define SWAP_UINT16(val) \
-+ ((uint16_t)((((val) & 0x00FF) << 8) | (((val) & 0xFF00) >> 8)))
-+
-+/**************************************************************************//**
-+ @Description Swaps the byte order of a given 32-bit value.
-+
-+ @Param[in] val - The 32-bit value to swap.
-+
-+ @Return The byte-swapped value..
-+
-+ @Cautions The given value is evaluated multiple times by this macro.
-+ For calculated expressions or expressions that contain function
-+ calls it is recommended to use the SwapUint32() routine.
-+
-+ @hideinitializer
-+*//***************************************************************************/
-+#define SWAP_UINT32(val) \
-+ ((uint32_t)((((val) & 0x000000FF) << 24) | \
-+ (((val) & 0x0000FF00) << 8) | \
-+ (((val) & 0x00FF0000) >> 8) | \
-+ (((val) & 0xFF000000) >> 24)))
-+
-+/**************************************************************************//**
-+ @Description Swaps the byte order of a given 64-bit value.
-+
-+ @Param[in] val - The 64-bit value to swap.
-+
-+ @Return The byte-swapped value..
-+
-+ @Cautions The given value is evaluated multiple times by this macro.
-+ For calculated expressions or expressions that contain function
-+ calls it is recommended to use the SwapUint64() routine.
-+
-+ @hideinitializer
-+*//***************************************************************************/
-+#define SWAP_UINT64(val) \
-+ ((uint64_t)((((val) & 0x00000000000000FFULL) << 56) | \
-+ (((val) & 0x000000000000FF00ULL) << 40) | \
-+ (((val) & 0x0000000000FF0000ULL) << 24) | \
-+ (((val) & 0x00000000FF000000ULL) << 8) | \
-+ (((val) & 0x000000FF00000000ULL) >> 8) | \
-+ (((val) & 0x0000FF0000000000ULL) >> 24) | \
-+ (((val) & 0x00FF000000000000ULL) >> 40) | \
-+ (((val) & 0xFF00000000000000ULL) >> 56)))
-+
-+/* @} */
-+
-+/**************************************************************************//**
-+ @Collection Byte-Swap Routines
-+
-+ Routines for swapping the byte order of a given parameter and
-+ returning the swapped value.
-+
-+ These inline routines are safer than the byte-swap macros,
-+ because they evaluate the parameter expression only once.
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Function SwapUint16
-+
-+ @Description Returns the byte-swapped value of a given 16-bit value.
-+
-+ @Param[in] val - The 16-bit value.
-+
-+ @Return The byte-swapped value of the parameter.
-+*//***************************************************************************/
-+static __inline__ uint16_t SwapUint16(uint16_t val)
-+{
-+ return (uint16_t)(((val & 0x00FF) << 8) |
-+ ((val & 0xFF00) >> 8));
-+}
-+
-+/**************************************************************************//**
-+ @Function SwapUint32
-+
-+ @Description Returns the byte-swapped value of a given 32-bit value.
-+
-+ @Param[in] val - The 32-bit value.
-+
-+ @Return The byte-swapped value of the parameter.
-+*//***************************************************************************/
-+static __inline__ uint32_t SwapUint32(uint32_t val)
-+{
-+ return (uint32_t)(((val & 0x000000FF) << 24) |
-+ ((val & 0x0000FF00) << 8) |
-+ ((val & 0x00FF0000) >> 8) |
-+ ((val & 0xFF000000) >> 24));
-+}
-+
-+/**************************************************************************//**
-+ @Function SwapUint64
-+
-+ @Description Returns the byte-swapped value of a given 64-bit value.
-+
-+ @Param[in] val - The 64-bit value.
-+
-+ @Return The byte-swapped value of the parameter.
-+*//***************************************************************************/
-+static __inline__ uint64_t SwapUint64(uint64_t val)
-+{
-+ return (uint64_t)(((val & 0x00000000000000FFULL) << 56) |
-+ ((val & 0x000000000000FF00ULL) << 40) |
-+ ((val & 0x0000000000FF0000ULL) << 24) |
-+ ((val & 0x00000000FF000000ULL) << 8) |
-+ ((val & 0x000000FF00000000ULL) >> 8) |
-+ ((val & 0x0000FF0000000000ULL) >> 24) |
-+ ((val & 0x00FF000000000000ULL) >> 40) |
-+ ((val & 0xFF00000000000000ULL) >> 56));
-+}
-+
-+/* @} */
-+
-+/**************************************************************************//**
-+ @Collection In-place Byte-Swap-And-Set Routines
-+
-+ Routines for swapping the byte order of a given variable and
-+ setting the swapped value back to the same variable.
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Function SwapUint16P
-+
-+ @Description Swaps the byte order of a given 16-bit variable.
-+
-+ @Param[in] p_Val - Pointer to the 16-bit variable.
-+
-+ @Return None.
-+*//***************************************************************************/
-+static __inline__ void SwapUint16P(uint16_t *p_Val)
-+{
-+ *p_Val = SwapUint16(*p_Val);
-+}
-+
-+/**************************************************************************//**
-+ @Function SwapUint32P
-+
-+ @Description Swaps the byte order of a given 32-bit variable.
-+
-+ @Param[in] p_Val - Pointer to the 32-bit variable.
-+
-+ @Return None.
-+*//***************************************************************************/
-+static __inline__ void SwapUint32P(uint32_t *p_Val)
-+{
-+ *p_Val = SwapUint32(*p_Val);
-+}
-+
-+/**************************************************************************//**
-+ @Function SwapUint64P
-+
-+ @Description Swaps the byte order of a given 64-bit variable.
-+
-+ @Param[in] p_Val - Pointer to the 64-bit variable.
-+
-+ @Return None.
-+*//***************************************************************************/
-+static __inline__ void SwapUint64P(uint64_t *p_Val)
-+{
-+ *p_Val = SwapUint64(*p_Val);
-+}
-+
-+/* @} */
-+
-+
-+/**************************************************************************//**
-+ @Collection Little-Endian Conversion Macros
-+
-+ These macros convert given parameters to or from Little-Endian
-+ format. Use these macros when you want to read or write a specific
-+ Little-Endian value in memory, without a-priori knowing the CPU
-+ byte order.
-+
-+ These macros use the byte-swap routines. For conversion of
-+ constants in initialization structures, you may use the CONST
-+ versions of these macros (see below), which are using the
-+ byte-swap macros instead.
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Description Converts a given 16-bit value from CPU byte order to
-+ Little-Endian byte order.
-+
-+ @Param[in] val - The 16-bit value to convert.
-+
-+ @Return The converted value.
-+
-+ @hideinitializer
-+*//***************************************************************************/
-+#define CPU_TO_LE16(val) SwapUint16(val)
-+
-+/**************************************************************************//**
-+ @Description Converts a given 32-bit value from CPU byte order to
-+ Little-Endian byte order.
-+
-+ @Param[in] val - The 32-bit value to convert.
-+
-+ @Return The converted value.
-+
-+ @hideinitializer
-+*//***************************************************************************/
-+#define CPU_TO_LE32(val) SwapUint32(val)
-+
-+/**************************************************************************//**
-+ @Description Converts a given 64-bit value from CPU byte order to
-+ Little-Endian byte order.
-+
-+ @Param[in] val - The 64-bit value to convert.
-+
-+ @Return The converted value.
-+
-+ @hideinitializer
-+*//***************************************************************************/
-+#define CPU_TO_LE64(val) SwapUint64(val)
-+
-+
-+/**************************************************************************//**
-+ @Description Converts a given 16-bit value from Little-Endian byte order to
-+ CPU byte order.
-+
-+ @Param[in] val - The 16-bit value to convert.
-+
-+ @Return The converted value.
-+
-+ @hideinitializer
-+*//***************************************************************************/
-+#define LE16_TO_CPU(val) CPU_TO_LE16(val)
-+
-+/**************************************************************************//**
-+ @Description Converts a given 32-bit value from Little-Endian byte order to
-+ CPU byte order.
-+
-+ @Param[in] val - The 32-bit value to convert.
-+
-+ @Return The converted value.
-+
-+ @hideinitializer
-+*//***************************************************************************/
-+#define LE32_TO_CPU(val) CPU_TO_LE32(val)
-+
-+/**************************************************************************//**
-+ @Description Converts a given 64-bit value from Little-Endian byte order to
-+ CPU byte order.
-+
-+ @Param[in] val - The 64-bit value to convert.
-+
-+ @Return The converted value.
-+
-+ @hideinitializer
-+*//***************************************************************************/
-+#define LE64_TO_CPU(val) CPU_TO_LE64(val)
-+
-+/* @} */
-+
-+/**************************************************************************//**
-+ @Collection Little-Endian Constant Conversion Macros
-+
-+ These macros convert given constants to or from Little-Endian
-+ format. Use these macros when you want to read or write a specific
-+ Little-Endian constant in memory, without a-priori knowing the
-+ CPU byte order.
-+
-+ These macros use the byte-swap macros, therefore can be used for
-+ conversion of constants in initialization structures.
-+
-+ @Cautions The parameters of these macros are evaluated multiple times.
-+ For non-constant expressions, use the non-CONST macro versions.
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Description Converts a given 16-bit constant from CPU byte order to
-+ Little-Endian byte order.
-+
-+ @Param[in] val - The 16-bit value to convert.
-+
-+ @Return The converted value.
-+
-+ @hideinitializer
-+*//***************************************************************************/
-+#define CONST_CPU_TO_LE16(val) SWAP_UINT16(val)
-+
-+/**************************************************************************//**
-+ @Description Converts a given 32-bit constant from CPU byte order to
-+ Little-Endian byte order.
-+
-+ @Param[in] val - The 32-bit value to convert.
-+
-+ @Return The converted value.
-+
-+ @hideinitializer
-+*//***************************************************************************/
-+#define CONST_CPU_TO_LE32(val) SWAP_UINT32(val)
-+
-+/**************************************************************************//**
-+ @Description Converts a given 64-bit constant from CPU byte order to
-+ Little-Endian byte order.
-+
-+ @Param[in] val - The 64-bit value to convert.
-+
-+ @Return The converted value.
-+
-+ @hideinitializer
-+*//***************************************************************************/
-+#define CONST_CPU_TO_LE64(val) SWAP_UINT64(val)
-+
-+
-+/**************************************************************************//**
-+ @Description Converts a given 16-bit constant from Little-Endian byte order
-+ to CPU byte order.
-+
-+ @Param[in] val - The 16-bit value to convert.
-+
-+ @Return The converted value.
-+
-+ @hideinitializer
-+*//***************************************************************************/
-+#define CONST_LE16_TO_CPU(val) CONST_CPU_TO_LE16(val)
-+
-+/**************************************************************************//**
-+ @Description Converts a given 32-bit constant from Little-Endian byte order
-+ to CPU byte order.
-+
-+ @Param[in] val - The 32-bit value to convert.
-+
-+ @Return The converted value.
-+
-+ @hideinitializer
-+*//***************************************************************************/
-+#define CONST_LE32_TO_CPU(val) CONST_CPU_TO_LE32(val)
-+
-+/**************************************************************************//**
-+ @Description Converts a given 64-bit constant from Little-Endian byte order
-+ to CPU byte order.
-+
-+ @Param[in] val - The 64-bit value to convert.
-+
-+ @Return The converted value.
-+
-+ @hideinitializer
-+*//***************************************************************************/
-+#define CONST_LE64_TO_CPU(val) CONST_CPU_TO_LE64(val)
-+
-+/* @} */
-+
-+
-+/** @} */ /* end of endian_id group */
-+/** @} */ /* end of gen_id group */
-+
-+
-+#endif /* __ENDIAN_EXT_H */
-+
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/enet_ext.h
-@@ -0,0 +1,205 @@
-+/* Copyright (c) 2008-2012 Freescale Semiconductor, Inc
-+ * All rights reserved.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/**************************************************************************//**
-+ @File enet_ext.h
-+
-+ @Description Ethernet generic definitions and enums.
-+*//***************************************************************************/
-+
-+#ifndef __ENET_EXT_H
-+#define __ENET_EXT_H
-+
-+#include "fsl_enet.h"
-+
-+#define ENET_NUM_OCTETS_PER_ADDRESS 6 /**< Number of octets (8-bit bytes) in an ethernet address */
-+#define ENET_GROUP_ADDR 0x01 /**< Group address mask for ethernet addresses */
-+
-+
-+/**************************************************************************//**
-+ @Description Ethernet Address
-+*//***************************************************************************/
-+typedef uint8_t t_EnetAddr[ENET_NUM_OCTETS_PER_ADDRESS];
-+
-+/**************************************************************************//**
-+ @Description Ethernet Address Type.
-+*//***************************************************************************/
-+typedef enum e_EnetAddrType
-+{
-+ e_ENET_ADDR_TYPE_INDIVIDUAL, /**< Individual (unicast) address */
-+ e_ENET_ADDR_TYPE_GROUP, /**< Group (multicast) address */
-+ e_ENET_ADDR_TYPE_BROADCAST /**< Broadcast address */
-+} e_EnetAddrType;
-+
-+/**************************************************************************//**
-+ @Description Ethernet MAC-PHY Interface
-+*//***************************************************************************/
-+typedef enum e_EnetInterface
-+{
-+ e_ENET_IF_MII = E_ENET_IF_MII, /**< MII interface */
-+ e_ENET_IF_RMII = E_ENET_IF_RMII, /**< RMII interface */
-+ e_ENET_IF_SMII = E_ENET_IF_SMII, /**< SMII interface */
-+ e_ENET_IF_GMII = E_ENET_IF_GMII, /**< GMII interface */
-+ e_ENET_IF_RGMII = E_ENET_IF_RGMII, /**< RGMII interface */
-+ e_ENET_IF_TBI = E_ENET_IF_TBI, /**< TBI interface */
-+ e_ENET_IF_RTBI = E_ENET_IF_RTBI, /**< RTBI interface */
-+ e_ENET_IF_SGMII = E_ENET_IF_SGMII, /**< SGMII interface */
-+ e_ENET_IF_XGMII = E_ENET_IF_XGMII, /**< XGMII interface */
-+ e_ENET_IF_QSGMII= E_ENET_IF_QSGMII, /**< QSGMII interface */
-+ e_ENET_IF_XFI = E_ENET_IF_XFI /**< XFI interface */
-+} e_EnetInterface;
-+
-+#define ENET_IF_SGMII_BASEX 0x80000000 /**< SGMII/QSGII interface with 1000BaseX
-+ auto-negotiation between MAC and phy
-+ or backplane;
-+ Note: 1000BaseX auto-negotiation relates
-+ only to interface between MAC and phy/backplane,
-+ SGMII phy can still synchronize with far-end phy
-+ at 10Mbps, 100Mbps or 1000Mbps */
-+
-+/**************************************************************************//**
-+ @Description Ethernet Duplex Mode
-+*//***************************************************************************/
-+typedef enum e_EnetDuplexMode
-+{
-+ e_ENET_HALF_DUPLEX, /**< Half-Duplex mode */
-+ e_ENET_FULL_DUPLEX /**< Full-Duplex mode */
-+} e_EnetDuplexMode;
-+
-+/**************************************************************************//**
-+ @Description Ethernet Speed (nominal data rate)
-+*//***************************************************************************/
-+typedef enum e_EnetSpeed
-+{
-+ e_ENET_SPEED_10 = E_ENET_SPEED_10, /**< 10 Mbps */
-+ e_ENET_SPEED_100 = E_ENET_SPEED_100, /**< 100 Mbps */
-+ e_ENET_SPEED_1000 = E_ENET_SPEED_1000, /**< 1000 Mbps = 1 Gbps */
-+ e_ENET_SPEED_2500 = E_ENET_SPEED_2500, /**< 2500 Mbps = 2.5 Gbps */
-+ e_ENET_SPEED_10000 = E_ENET_SPEED_10000 /**< 10000 Mbps = 10 Gbps */
-+} e_EnetSpeed;
-+
-+/**************************************************************************//**
-+ @Description Ethernet mode (combination of MAC-PHY interface and speed)
-+*//***************************************************************************/
-+typedef enum e_EnetMode
-+{
-+ e_ENET_MODE_INVALID = 0, /**< Invalid Ethernet mode */
-+ e_ENET_MODE_MII_10 = (e_ENET_IF_MII | e_ENET_SPEED_10), /**< 10 Mbps MII */
-+ e_ENET_MODE_MII_100 = (e_ENET_IF_MII | e_ENET_SPEED_100), /**< 100 Mbps MII */
-+ e_ENET_MODE_RMII_10 = (e_ENET_IF_RMII | e_ENET_SPEED_10), /**< 10 Mbps RMII */
-+ e_ENET_MODE_RMII_100 = (e_ENET_IF_RMII | e_ENET_SPEED_100), /**< 100 Mbps RMII */
-+ e_ENET_MODE_SMII_10 = (e_ENET_IF_SMII | e_ENET_SPEED_10), /**< 10 Mbps SMII */
-+ e_ENET_MODE_SMII_100 = (e_ENET_IF_SMII | e_ENET_SPEED_100), /**< 100 Mbps SMII */
-+ e_ENET_MODE_GMII_1000 = (e_ENET_IF_GMII | e_ENET_SPEED_1000), /**< 1000 Mbps GMII */
-+ e_ENET_MODE_RGMII_10 = (e_ENET_IF_RGMII | e_ENET_SPEED_10), /**< 10 Mbps RGMII */
-+ e_ENET_MODE_RGMII_100 = (e_ENET_IF_RGMII | e_ENET_SPEED_100), /**< 100 Mbps RGMII */
-+ e_ENET_MODE_RGMII_1000 = (e_ENET_IF_RGMII | e_ENET_SPEED_1000), /**< 1000 Mbps RGMII */
-+ e_ENET_MODE_TBI_1000 = (e_ENET_IF_TBI | e_ENET_SPEED_1000), /**< 1000 Mbps TBI */
-+ e_ENET_MODE_RTBI_1000 = (e_ENET_IF_RTBI | e_ENET_SPEED_1000), /**< 1000 Mbps RTBI */
-+ e_ENET_MODE_SGMII_10 = (e_ENET_IF_SGMII | e_ENET_SPEED_10),
-+ /**< 10 Mbps SGMII with auto-negotiation between MAC and
-+ SGMII phy according to Cisco SGMII specification */
-+ e_ENET_MODE_SGMII_100 = (e_ENET_IF_SGMII | e_ENET_SPEED_100),
-+ /**< 100 Mbps SGMII with auto-negotiation between MAC and
-+ SGMII phy according to Cisco SGMII specification */
-+ e_ENET_MODE_SGMII_1000 = (e_ENET_IF_SGMII | e_ENET_SPEED_1000),
-+ /**< 1000 Mbps SGMII with auto-negotiation between MAC and
-+ SGMII phy according to Cisco SGMII specification */
-+ e_ENET_MODE_SGMII_2500 = (e_ENET_IF_SGMII | e_ENET_SPEED_2500),
-+ e_ENET_MODE_SGMII_BASEX_10 = (ENET_IF_SGMII_BASEX | e_ENET_IF_SGMII | e_ENET_SPEED_10),
-+ /**< 10 Mbps SGMII with 1000BaseX auto-negotiation between
-+ MAC and SGMII phy or backplane */
-+ e_ENET_MODE_SGMII_BASEX_100 = (ENET_IF_SGMII_BASEX | e_ENET_IF_SGMII | e_ENET_SPEED_100),
-+ /**< 100 Mbps SGMII with 1000BaseX auto-negotiation between
-+ MAC and SGMII phy or backplane */
-+ e_ENET_MODE_SGMII_BASEX_1000 = (ENET_IF_SGMII_BASEX | e_ENET_IF_SGMII | e_ENET_SPEED_1000),
-+ /**< 1000 Mbps SGMII with 1000BaseX auto-negotiation between
-+ MAC and SGMII phy or backplane */
-+ e_ENET_MODE_QSGMII_1000 = (e_ENET_IF_QSGMII| e_ENET_SPEED_1000),
-+ /**< 1000 Mbps QSGMII with auto-negotiation between MAC and
-+ QSGMII phy according to Cisco QSGMII specification */
-+ e_ENET_MODE_QSGMII_BASEX_1000 = (ENET_IF_SGMII_BASEX | e_ENET_IF_QSGMII| e_ENET_SPEED_1000),
-+ /**< 1000 Mbps QSGMII with 1000BaseX auto-negotiation between
-+ MAC and QSGMII phy or backplane */
-+ e_ENET_MODE_XGMII_10000 = (e_ENET_IF_XGMII | e_ENET_SPEED_10000), /**< 10000 Mbps XGMII */
-+ e_ENET_MODE_XFI_10000 = (e_ENET_IF_XFI | e_ENET_SPEED_10000) /**< 10000 Mbps XFI */
-+} e_EnetMode;
-+
-+
-+#define IS_ENET_MODE_VALID(mode) \
-+ (((mode) == e_ENET_MODE_MII_10 ) || \
-+ ((mode) == e_ENET_MODE_MII_100 ) || \
-+ ((mode) == e_ENET_MODE_RMII_10 ) || \
-+ ((mode) == e_ENET_MODE_RMII_100 ) || \
-+ ((mode) == e_ENET_MODE_SMII_10 ) || \
-+ ((mode) == e_ENET_MODE_SMII_100 ) || \
-+ ((mode) == e_ENET_MODE_GMII_1000 ) || \
-+ ((mode) == e_ENET_MODE_RGMII_10 ) || \
-+ ((mode) == e_ENET_MODE_RGMII_100 ) || \
-+ ((mode) == e_ENET_MODE_RGMII_1000 ) || \
-+ ((mode) == e_ENET_MODE_TBI_1000 ) || \
-+ ((mode) == e_ENET_MODE_RTBI_1000 ) || \
-+ ((mode) == e_ENET_MODE_SGMII_10 ) || \
-+ ((mode) == e_ENET_MODE_SGMII_100 ) || \
-+ ((mode) == e_ENET_MODE_SGMII_1000 ) || \
-+ ((mode) == e_ENET_MODE_SGMII_BASEX_10 ) || \
-+ ((mode) == e_ENET_MODE_SGMII_BASEX_100 ) || \
-+ ((mode) == e_ENET_MODE_SGMII_BASEX_1000 ) || \
-+ ((mode) == e_ENET_MODE_XGMII_10000) || \
-+ ((mode) == e_ENET_MODE_QSGMII_1000) || \
-+ ((mode) == e_ENET_MODE_QSGMII_BASEX_1000) || \
-+ ((mode) == e_ENET_MODE_XFI_10000))
-+
-+
-+#define MAKE_ENET_MODE(_interface, _speed) (e_EnetMode)((_interface) | (_speed))
-+
-+#define ENET_INTERFACE_FROM_MODE(mode) (e_EnetInterface)((mode) & 0x0FFF0000)
-+#define ENET_SPEED_FROM_MODE(mode) (e_EnetSpeed)((mode) & 0x0000FFFF)
-+
-+#define ENET_ADDR_TO_UINT64(_enetAddr) \
-+ (uint64_t)(((uint64_t)(_enetAddr)[0] << 40) | \
-+ ((uint64_t)(_enetAddr)[1] << 32) | \
-+ ((uint64_t)(_enetAddr)[2] << 24) | \
-+ ((uint64_t)(_enetAddr)[3] << 16) | \
-+ ((uint64_t)(_enetAddr)[4] << 8) | \
-+ ((uint64_t)(_enetAddr)[5]))
-+
-+#define MAKE_ENET_ADDR_FROM_UINT64(_addr64, _enetAddr) \
-+ do { \
-+ int i; \
-+ for (i=0; i < ENET_NUM_OCTETS_PER_ADDRESS; i++) \
-+ (_enetAddr)[i] = (uint8_t)((_addr64) >> ((5-i)*8)); \
-+ } while (0)
-+
-+
-+#endif /* __ENET_EXT_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/error_ext.h
-@@ -0,0 +1,529 @@
-+/* Copyright (c) 2008-2012 Freescale Semiconductor, Inc
-+ * All rights reserved.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/**************************************************************************//**
-+ @File error_ext.h
-+
-+ @Description Error definitions.
-+*//***************************************************************************/
-+
-+#ifndef __ERROR_EXT_H
-+#define __ERROR_EXT_H
-+
-+#if !defined(NCSW_LINUX)
-+#include <errno.h>
-+#endif
-+
-+#include "std_ext.h"
-+#include "xx_ext.h"
-+#include "core_ext.h"
-+
-+
-+
-+
-+/**************************************************************************//**
-+ @Group gen_id General Drivers Utilities
-+
-+ @Description External routines.
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Group gen_error_id Errors, Events and Debug
-+
-+ @Description External routines.
-+
-+ @{
-+*//***************************************************************************/
-+
-+/******************************************************************************
-+The scheme below provides the bits description for error codes:
-+
-+ 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15
-+| Reserved (should be zero) | Module ID |
-+
-+ 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31
-+| Error Type |
-+******************************************************************************/
-+
-+#define ERROR_CODE(_err) ((((uint32_t)_err) & 0x0000FFFF) | __ERR_MODULE__)
-+
-+#define GET_ERROR_TYPE(_errcode) ((_errcode) & 0x0000FFFF)
-+ /**< Extract module code from error code (#t_Error) */
-+
-+#define GET_ERROR_MODULE(_errcode) ((_errcode) & 0x00FF0000)
-+ /**< Extract error type (#e_ErrorType) from
-+ error code (#t_Error) */
-+
-+
-+/**************************************************************************//**
-+ @Description Error Type Enumeration
-+*//***************************************************************************/
-+typedef enum e_ErrorType /* Comments / Associated Message Strings */
-+{ /* ------------------------------------------------------------ */
-+ E_OK = 0 /* Never use "RETURN_ERROR" with E_OK; Use "return E_OK;" */
-+ ,E_WRITE_FAILED = EIO /**< Write access failed on memory/device. */
-+ /* String: none, or device name. */
-+ ,E_NO_DEVICE = ENXIO /**< The associated device is not initialized. */
-+ /* String: none. */
-+ ,E_NOT_AVAILABLE = EAGAIN
-+ /**< Resource is unavailable. */
-+ /* String: none, unless the operation is not the main goal
-+ of the function (in this case add resource description). */
-+ ,E_NO_MEMORY = ENOMEM /**< External memory allocation failed. */
-+ /* String: description of item for which allocation failed. */
-+ ,E_INVALID_ADDRESS = EFAULT
-+ /**< Invalid address. */
-+ /* String: description of the specific violation. */
-+ ,E_BUSY = EBUSY /**< Resource or module is busy. */
-+ /* String: none, unless the operation is not the main goal
-+ of the function (in this case add resource description). */
-+ ,E_ALREADY_EXISTS = EEXIST
-+ /**< Requested resource or item already exists. */
-+ /* Use when resource duplication or sharing are not allowed.
-+ String: none, unless the operation is not the main goal
-+ of the function (in this case add item description). */
-+ ,E_INVALID_OPERATION = ENODEV
-+ /**< The operation/command is invalid (unrecognized). */
-+ /* String: none. */
-+ ,E_INVALID_VALUE = EDOM /**< Invalid value. */
-+ /* Use for non-enumeration parameters, and
-+ only when other error types are not suitable.
-+ String: parameter description + "(should be <attribute>)",
-+ e.g: "Maximum Rx buffer length (should be divisible by 8)",
-+ "Channel number (should be even)". */
-+ ,E_NOT_IN_RANGE = ERANGE/**< Parameter value is out of range. */
-+ /* Don't use this error for enumeration parameters.
-+ String: parameter description + "(should be %d-%d)",
-+ e.g: "Number of pad characters (should be 0-15)". */
-+ ,E_NOT_SUPPORTED = ENOSYS
-+ /**< The function is not supported or not implemented. */
-+ /* String: none. */
-+ ,E_INVALID_STATE /**< The operation is not allowed in current module state. */
-+ /* String: none. */
-+ ,E_INVALID_HANDLE /**< Invalid handle of module or object. */
-+ /* String: none, unless the function takes in more than one
-+ handle (in this case add the handle description) */
-+ ,E_INVALID_ID /**< Invalid module ID (usually enumeration or index). */
-+ /* String: none, unless the function takes in more than one
-+ ID (in this case add the ID description) */
-+ ,E_NULL_POINTER /**< Unexpected NULL pointer. */
-+ /* String: pointer description. */
-+ ,E_INVALID_SELECTION /**< Invalid selection or mode. */
-+ /* Use for enumeration values, only when other error types
-+ are not suitable.
-+ String: parameter description. */
-+ ,E_INVALID_COMM_MODE /**< Invalid communication mode. */
-+ /* String: none, unless the function takes in more than one
-+ communication mode indications (in this case add
-+ parameter description). */
-+ ,E_INVALID_MEMORY_TYPE /**< Invalid memory type. */
-+ /* String: none, unless the function takes in more than one
-+ memory types (in this case add memory description,
-+ e.g: "Data memory", "Buffer descriptors memory"). */
-+ ,E_INVALID_CLOCK /**< Invalid clock. */
-+ /* String: none, unless the function takes in more than one
-+ clocks (in this case add clock description,
-+ e.g: "Rx clock", "Tx clock"). */
-+ ,E_CONFLICT /**< Some setting conflicts with another setting. */
-+ /* String: description of the conflicting settings. */
-+ ,E_NOT_ALIGNED /**< Non-aligned address. */
-+ /* String: parameter description + "(should be %d-bytes aligned)",
-+ e.g: "Rx data buffer (should be 32-bytes aligned)". */
-+ ,E_NOT_FOUND /**< Requested resource or item was not found. */
-+ /* Use only when the resource/item is uniquely identified.
-+ String: none, unless the operation is not the main goal
-+ of the function (in this case add item description). */
-+ ,E_FULL /**< Resource is full. */
-+ /* String: none, unless the operation is not the main goal
-+ of the function (in this case add resource description). */
-+ ,E_EMPTY /**< Resource is empty. */
-+ /* String: none, unless the operation is not the main goal
-+ of the function (in this case add resource description). */
-+ ,E_ALREADY_FREE /**< Specified resource or item is already free or deleted. */
-+ /* String: none, unless the operation is not the main goal
-+ of the function (in this case add item description). */
-+ ,E_READ_FAILED /**< Read access failed on memory/device. */
-+ /* String: none, or device name. */
-+ ,E_INVALID_FRAME /**< Invalid frame object (NULL handle or missing buffers). */
-+ /* String: none. */
-+ ,E_SEND_FAILED /**< Send operation failed on device. */
-+ /* String: none, or device name. */
-+ ,E_RECEIVE_FAILED /**< Receive operation failed on device. */
-+ /* String: none, or device name. */
-+ ,E_TIMEOUT/* = ETIMEDOUT*/ /**< The operation timed out. */
-+ /* String: none. */
-+
-+ ,E_DUMMY_LAST /* NEVER USED */
-+
-+} e_ErrorType;
-+
-+/**************************************************************************//**
-+ @Description Event Type Enumeration
-+*//***************************************************************************/
-+typedef enum e_Event /* Comments / Associated Flags and Message Strings */
-+{ /* ------------------------------------------------------------ */
-+ EV_NO_EVENT = 0 /**< No event; Never used. */
-+
-+ ,EV_RX_DISCARD /**< Received packet discarded (by the driver, and only for
-+ complete packets);
-+ Flags: error flags in case of error, zero otherwise. */
-+ /* String: reason for discard, e.g: "Error in frame",
-+ "Disordered frame", "Incomplete frame", "No frame object". */
-+ ,EV_RX_ERROR /**< Receive error (by hardware/firmware);
-+ Flags: usually status flags from the buffer descriptor. */
-+ /* String: none. */
-+ ,EV_TX_ERROR /**< Transmit error (by hardware/firmware);
-+ Flags: usually status flags from the buffer descriptor. */
-+ /* String: none. */
-+ ,EV_NO_BUFFERS /**< System ran out of buffer objects;
-+ Flags: zero. */
-+ /* String: none. */
-+ ,EV_NO_MB_FRAMES /**< System ran out of multi-buffer frame objects;
-+ Flags: zero. */
-+ /* String: none. */
-+ ,EV_NO_SB_FRAMES /**< System ran out of single-buffer frame objects;
-+ Flags: zero. */
-+ /* String: none. */
-+ ,EV_TX_QUEUE_FULL /**< Transmit queue is full;
-+ Flags: zero. */
-+ /* String: none. */
-+ ,EV_RX_QUEUE_FULL /**< Receive queue is full;
-+ Flags: zero. */
-+ /* String: none. */
-+ ,EV_INTR_QUEUE_FULL /**< Interrupt queue overflow;
-+ Flags: zero. */
-+ /* String: none. */
-+ ,EV_NO_DATA_BUFFER /**< Data buffer allocation (from higher layer) failed;
-+ Flags: zero. */
-+ /* String: none. */
-+ ,EV_OBJ_POOL_EMPTY /**< Objects pool is empty;
-+ Flags: zero. */
-+ /* String: object description (name). */
-+ ,EV_BUS_ERROR /**< Illegal access on bus;
-+ Flags: the address (if available) or bus identifier */
-+ /* String: bus/address/module description. */
-+ ,EV_PTP_TXTS_QUEUE_FULL /**< PTP Tx timestamps queue is full;
-+ Flags: zero. */
-+ /* String: none. */
-+ ,EV_PTP_RXTS_QUEUE_FULL /**< PTP Rx timestamps queue is full;
-+ Flags: zero. */
-+ /* String: none. */
-+ ,EV_DUMMY_LAST
-+
-+} e_Event;
-+
-+
-+/**************************************************************************//**
-+ @Collection Debug Levels for Errors and Events
-+
-+ The level description refers to errors only.
-+ For events, classification is done by the user.
-+
-+ The TRACE, INFO and WARNING levels are allowed only when using
-+ the DBG macro, and are not allowed when using the error macros
-+ (RETURN_ERROR or REPORT_ERROR).
-+ @{
-+*//***************************************************************************/
-+#define REPORT_LEVEL_CRITICAL 1 /**< Crasher: Incorrect flow, NULL pointers/handles. */
-+#define REPORT_LEVEL_MAJOR 2 /**< Cannot proceed: Invalid operation, parameters or
-+ configuration. */
-+#define REPORT_LEVEL_MINOR 3 /**< Recoverable problem: a repeating call with the same
-+ parameters may be successful. */
-+#define REPORT_LEVEL_WARNING 4 /**< Something is not exactly right, yet it is not an error. */
-+#define REPORT_LEVEL_INFO 5 /**< Messages which may be of interest to user/programmer. */
-+#define REPORT_LEVEL_TRACE 6 /**< Program flow messages. */
-+
-+#define EVENT_DISABLED 0xFF /**< Disabled event (not reported at all) */
-+
-+/* @} */
-+
-+
-+
-+#define NO_MSG ("")
-+
-+#ifndef DEBUG_GLOBAL_LEVEL
-+#define DEBUG_GLOBAL_LEVEL REPORT_LEVEL_WARNING
-+#endif /* DEBUG_GLOBAL_LEVEL */
-+
-+#ifndef ERROR_GLOBAL_LEVEL
-+#define ERROR_GLOBAL_LEVEL DEBUG_GLOBAL_LEVEL
-+#endif /* ERROR_GLOBAL_LEVEL */
-+
-+#ifndef EVENT_GLOBAL_LEVEL
-+#define EVENT_GLOBAL_LEVEL REPORT_LEVEL_MINOR
-+#endif /* EVENT_GLOBAL_LEVEL */
-+
-+#ifdef EVENT_LOCAL_LEVEL
-+#define EVENT_DYNAMIC_LEVEL EVENT_LOCAL_LEVEL
-+#else
-+#define EVENT_DYNAMIC_LEVEL EVENT_GLOBAL_LEVEL
-+#endif /* EVENT_LOCAL_LEVEL */
-+
-+
-+#ifndef DEBUG_DYNAMIC_LEVEL
-+#define DEBUG_USING_STATIC_LEVEL
-+
-+#ifdef DEBUG_STATIC_LEVEL
-+#define DEBUG_DYNAMIC_LEVEL DEBUG_STATIC_LEVEL
-+#else
-+#define DEBUG_DYNAMIC_LEVEL DEBUG_GLOBAL_LEVEL
-+#endif /* DEBUG_STATIC_LEVEL */
-+
-+#else /* DEBUG_DYNAMIC_LEVEL */
-+#ifdef DEBUG_STATIC_LEVEL
-+#error "Please use either DEBUG_STATIC_LEVEL or DEBUG_DYNAMIC_LEVEL (not both)"
-+#else
-+int DEBUG_DYNAMIC_LEVEL = DEBUG_GLOBAL_LEVEL;
-+#endif /* DEBUG_STATIC_LEVEL */
-+#endif /* !DEBUG_DYNAMIC_LEVEL */
-+
-+
-+#ifndef ERROR_DYNAMIC_LEVEL
-+
-+#ifdef ERROR_STATIC_LEVEL
-+#define ERROR_DYNAMIC_LEVEL ERROR_STATIC_LEVEL
-+#else
-+#define ERROR_DYNAMIC_LEVEL ERROR_GLOBAL_LEVEL
-+#endif /* ERROR_STATIC_LEVEL */
-+
-+#else /* ERROR_DYNAMIC_LEVEL */
-+#ifdef ERROR_STATIC_LEVEL
-+#error "Please use either ERROR_STATIC_LEVEL or ERROR_DYNAMIC_LEVEL (not both)"
-+#else
-+int ERROR_DYNAMIC_LEVEL = ERROR_GLOBAL_LEVEL;
-+#endif /* ERROR_STATIC_LEVEL */
-+#endif /* !ERROR_DYNAMIC_LEVEL */
-+
-+#define PRINT_FORMAT "[CPU%02d, %s:%d %s]"
-+#define PRINT_FMT_PARAMS raw_smp_processor_id(), __FILE__, __LINE__, __FUNCTION__
-+
-+#if (!(defined(DEBUG_ERRORS)) || (DEBUG_ERRORS == 0))
-+/* No debug/error/event messages at all */
-+#define DBG(_level, _vmsg)
-+
-+#define REPORT_ERROR(_level, _err, _vmsg)
-+
-+#define RETURN_ERROR(_level, _err, _vmsg) \
-+ return ERROR_CODE(_err)
-+
-+#if (REPORT_EVENTS > 0)
-+
-+#define REPORT_EVENT(_ev, _appId, _flg, _vmsg) \
-+ do { \
-+ if (_ev##_LEVEL <= EVENT_DYNAMIC_LEVEL) { \
-+ XX_EventById((uint32_t)(_ev), (t_Handle)(_appId), (uint16_t)(_flg), NO_MSG); \
-+ } \
-+ } while (0)
-+
-+#else
-+
-+#define REPORT_EVENT(_ev, _appId, _flg, _vmsg)
-+
-+#endif /* (REPORT_EVENTS > 0) */
-+
-+
-+#else /* DEBUG_ERRORS > 0 */
-+
-+extern const char *dbgLevelStrings[];
-+extern const char *moduleStrings[];
-+#if (REPORT_EVENTS > 0)
-+extern const char *eventStrings[];
-+#endif /* (REPORT_EVENTS > 0) */
-+
-+char * ErrTypeStrings (e_ErrorType err);
-+
-+
-+#if ((defined(DEBUG_USING_STATIC_LEVEL)) && (DEBUG_DYNAMIC_LEVEL < REPORT_LEVEL_WARNING))
-+/* No need for DBG macro - debug level is higher anyway */
-+#define DBG(_level, _vmsg)
-+#else
-+#define DBG(_level, _vmsg) \
-+ do { \
-+ if (REPORT_LEVEL_##_level <= DEBUG_DYNAMIC_LEVEL) { \
-+ XX_Print("> %s (%s) " PRINT_FORMAT ": ", \
-+ dbgLevelStrings[REPORT_LEVEL_##_level - 1], \
-+ moduleStrings[__ERR_MODULE__ >> 16], \
-+ PRINT_FMT_PARAMS); \
-+ XX_Print _vmsg; \
-+ XX_Print("\r\n"); \
-+ } \
-+ } while (0)
-+#endif /* (defined(DEBUG_USING_STATIC_LEVEL) && (DEBUG_DYNAMIC_LEVEL < WARNING)) */
-+
-+
-+#define REPORT_ERROR(_level, _err, _vmsg) \
-+ do { \
-+ if (REPORT_LEVEL_##_level <= ERROR_DYNAMIC_LEVEL) { \
-+ XX_Print("! %s %s Error " PRINT_FORMAT ": %s; ", \
-+ dbgLevelStrings[REPORT_LEVEL_##_level - 1], \
-+ moduleStrings[__ERR_MODULE__ >> 16], \
-+ PRINT_FMT_PARAMS, \
-+ ErrTypeStrings((e_ErrorType)GET_ERROR_TYPE(_err))); \
-+ XX_Print _vmsg; \
-+ XX_Print("\r\n"); \
-+ } \
-+ } while (0)
-+
-+
-+#define RETURN_ERROR(_level, _err, _vmsg) \
-+ do { \
-+ REPORT_ERROR(_level, (_err), _vmsg); \
-+ return ERROR_CODE(_err); \
-+ } while (0)
-+
-+
-+#if (REPORT_EVENTS > 0)
-+
-+#define REPORT_EVENT(_ev, _appId, _flg, _vmsg) \
-+ do { \
-+ if (_ev##_LEVEL <= EVENT_DYNAMIC_LEVEL) { \
-+ XX_Print("~ %s %s Event " PRINT_FORMAT ": %s (flags: 0x%04x); ", \
-+ dbgLevelStrings[_ev##_LEVEL - 1], \
-+ moduleStrings[__ERR_MODULE__ >> 16], \
-+ PRINT_FMT_PARAMS, \
-+ eventStrings[((_ev) - EV_NO_EVENT - 1)], \
-+ (uint16_t)(_flg)); \
-+ XX_Print _vmsg; \
-+ XX_Print("\r\n"); \
-+ XX_EventById((uint32_t)(_ev), (t_Handle)(_appId), (uint16_t)(_flg), NO_MSG); \
-+ } \
-+ } while (0)
-+
-+#else /* not REPORT_EVENTS */
-+
-+#define REPORT_EVENT(_ev, _appId, _flg, _vmsg)
-+
-+#endif /* (REPORT_EVENTS > 0) */
-+
-+#endif /* (DEBUG_ERRORS > 0) */
-+
-+
-+/**************************************************************************//**
-+ @Function ASSERT_COND
-+
-+ @Description Assertion macro.
-+
-+ @Param[in] _cond - The condition being checked, in positive form;
-+ Failure of the condition triggers the assert.
-+*//***************************************************************************/
-+#ifdef DISABLE_ASSERTIONS
-+#define ASSERT_COND(_cond)
-+#else
-+#define ASSERT_COND(_cond) \
-+ do { \
-+ if (!(_cond)) { \
-+ XX_Print("*** ASSERT_COND failed " PRINT_FORMAT "\r\n", \
-+ PRINT_FMT_PARAMS); \
-+ XX_Exit(1); \
-+ } \
-+ } while (0)
-+#endif /* DISABLE_ASSERTIONS */
-+
-+
-+#ifdef DISABLE_INIT_PARAMETERS_CHECK
-+
-+#define CHECK_INIT_PARAMETERS(handle, f_check)
-+#define CHECK_INIT_PARAMETERS_RETURN_VALUE(handle, f_check, retval)
-+
-+#else
-+
-+#define CHECK_INIT_PARAMETERS(handle, f_check) \
-+ do { \
-+ t_Error err = f_check(handle); \
-+ if (err != E_OK) { \
-+ RETURN_ERROR(MAJOR, err, NO_MSG); \
-+ } \
-+ } while (0)
-+
-+#define CHECK_INIT_PARAMETERS_RETURN_VALUE(handle, f_check, retval) \
-+ do { \
-+ t_Error err = f_check(handle); \
-+ if (err != E_OK) { \
-+ REPORT_ERROR(MAJOR, err, NO_MSG); \
-+ return (retval); \
-+ } \
-+ } while (0)
-+
-+#endif /* DISABLE_INIT_PARAMETERS_CHECK */
-+
-+#ifdef DISABLE_SANITY_CHECKS
-+
-+#define SANITY_CHECK_RETURN_ERROR(_cond, _err)
-+#define SANITY_CHECK_RETURN_VALUE(_cond, _err, retval)
-+#define SANITY_CHECK_RETURN(_cond, _err)
-+#define SANITY_CHECK_EXIT(_cond, _err)
-+
-+#else /* DISABLE_SANITY_CHECKS */
-+
-+#define SANITY_CHECK_RETURN_ERROR(_cond, _err) \
-+ do { \
-+ if (!(_cond)) { \
-+ RETURN_ERROR(CRITICAL, (_err), NO_MSG); \
-+ } \
-+ } while (0)
-+
-+#define SANITY_CHECK_RETURN_VALUE(_cond, _err, retval) \
-+ do { \
-+ if (!(_cond)) { \
-+ REPORT_ERROR(CRITICAL, (_err), NO_MSG); \
-+ return (retval); \
-+ } \
-+ } while (0)
-+
-+#define SANITY_CHECK_RETURN(_cond, _err) \
-+ do { \
-+ if (!(_cond)) { \
-+ REPORT_ERROR(CRITICAL, (_err), NO_MSG); \
-+ return; \
-+ } \
-+ } while (0)
-+
-+#define SANITY_CHECK_EXIT(_cond, _err) \
-+ do { \
-+ if (!(_cond)) { \
-+ REPORT_ERROR(CRITICAL, (_err), NO_MSG); \
-+ XX_Exit(1); \
-+ } \
-+ } while (0)
-+
-+#endif /* DISABLE_SANITY_CHECKS */
-+
-+/** @} */ /* end of Debug/error Utils group */
-+
-+/** @} */ /* end of General Utils group */
-+
-+#endif /* __ERROR_EXT_H */
-+
-+
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/etc/list_ext.h
-@@ -0,0 +1,358 @@
-+/* Copyright (c) 2008-2012 Freescale Semiconductor, Inc
-+ * All rights reserved.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/**************************************************************************//**
-+
-+ @File list_ext.h
-+
-+ @Description External prototypes for list.c
-+*//***************************************************************************/
-+
-+#ifndef __LIST_EXT_H
-+#define __LIST_EXT_H
-+
-+
-+#include "std_ext.h"
-+
-+
-+/**************************************************************************//**
-+ @Group etc_id Utility Library Application Programming Interface
-+
-+ @Description External routines.
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Group list_id List
-+
-+ @Description List module functions,definitions and enums.
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Description List structure.
-+*//***************************************************************************/
-+typedef struct List
-+{
-+ struct List *p_Next; /**< A pointer to the next list object */
-+ struct List *p_Prev; /**< A pointer to the previous list object */
-+} t_List;
-+
-+
-+/**************************************************************************//**
-+ @Function LIST_FIRST/LIST_LAST/LIST_NEXT/LIST_PREV
-+
-+ @Description Macro to get first/last/next/previous entry in a list.
-+
-+ @Param[in] p_List - A pointer to a list.
-+*//***************************************************************************/
-+#define LIST_FIRST(p_List) (p_List)->p_Next
-+#define LIST_LAST(p_List) (p_List)->p_Prev
-+#define LIST_NEXT LIST_FIRST
-+#define LIST_PREV LIST_LAST
-+
-+
-+/**************************************************************************//**
-+ @Function LIST_INIT
-+
-+ @Description Macro for initialization of a list struct.
-+
-+ @Param[in] lst - The t_List object to initialize.
-+*//***************************************************************************/
-+#define LIST_INIT(lst) {&(lst), &(lst)}
-+
-+
-+/**************************************************************************//**
-+ @Function LIST
-+
-+ @Description Macro to declare of a list.
-+
-+ @Param[in] listName - The list object name.
-+*//***************************************************************************/
-+#define LIST(listName) t_List listName = LIST_INIT(listName)
-+
-+
-+/**************************************************************************//**
-+ @Function INIT_LIST
-+
-+ @Description Macro to initialize a list pointer.
-+
-+ @Param[in] p_List - The list pointer.
-+*//***************************************************************************/
-+#define INIT_LIST(p_List) LIST_FIRST(p_List) = LIST_LAST(p_List) = (p_List)
-+
-+
-+/**************************************************************************//**
-+ @Function LIST_OBJECT
-+
-+ @Description Macro to get the struct (object) for this entry.
-+
-+ @Param[in] type - The type of the struct (object) this list is embedded in.
-+ @Param[in] member - The name of the t_List object within the struct.
-+
-+ @Return The structure pointer for this entry.
-+*//***************************************************************************/
-+#define MEMBER_OFFSET(type, member) (PTR_TO_UINT(&((type *)0)->member))
-+#define LIST_OBJECT(p_List, type, member) \
-+ ((type *)((char *)(p_List)-MEMBER_OFFSET(type, member)))
-+
-+
-+/**************************************************************************//**
-+ @Function LIST_FOR_EACH
-+
-+ @Description Macro to iterate over a list.
-+
-+ @Param[in] p_Pos - A pointer to a list to use as a loop counter.
-+ @Param[in] p_Head - A pointer to the head for your list pointer.
-+
-+ @Cautions You can't delete items with this routine.
-+ For deletion use LIST_FOR_EACH_SAFE().
-+*//***************************************************************************/
-+#define LIST_FOR_EACH(p_Pos, p_Head) \
-+ for (p_Pos = LIST_FIRST(p_Head); p_Pos != (p_Head); p_Pos = LIST_NEXT(p_Pos))
-+
-+
-+/**************************************************************************//**
-+ @Function LIST_FOR_EACH_SAFE
-+
-+ @Description Macro to iterate over a list safe against removal of list entry.
-+
-+ @Param[in] p_Pos - A pointer to a list to use as a loop counter.
-+ @Param[in] p_Tmp - Another pointer to a list to use as temporary storage.
-+ @Param[in] p_Head - A pointer to the head for your list pointer.
-+*//***************************************************************************/
-+#define LIST_FOR_EACH_SAFE(p_Pos, p_Tmp, p_Head) \
-+ for (p_Pos = LIST_FIRST(p_Head), p_Tmp = LIST_FIRST(p_Pos); \
-+ p_Pos != (p_Head); \
-+ p_Pos = p_Tmp, p_Tmp = LIST_NEXT(p_Pos))
-+
-+
-+/**************************************************************************//**
-+ @Function LIST_FOR_EACH_OBJECT_SAFE
-+
-+ @Description Macro to iterate over list of given type safely.
-+
-+ @Param[in] p_Pos - A pointer to a list to use as a loop counter.
-+ @Param[in] p_Tmp - Another pointer to a list to use as temporary storage.
-+ @Param[in] type - The type of the struct this is embedded in.
-+ @Param[in] p_Head - A pointer to the head for your list pointer.
-+ @Param[in] member - The name of the list_struct within the struct.
-+
-+ @Cautions You can't delete items with this routine.
-+ For deletion use LIST_FOR_EACH_SAFE().
-+*//***************************************************************************/
-+#define LIST_FOR_EACH_OBJECT_SAFE(p_Pos, p_Tmp, p_Head, type, member) \
-+ for (p_Pos = LIST_OBJECT(LIST_FIRST(p_Head), type, member), \
-+ p_Tmp = LIST_OBJECT(LIST_FIRST(&p_Pos->member), type, member); \
-+ &p_Pos->member != (p_Head); \
-+ p_Pos = p_Tmp, \
-+ p_Tmp = LIST_OBJECT(LIST_FIRST(&p_Pos->member), type, member))
-+
-+/**************************************************************************//**
-+ @Function LIST_FOR_EACH_OBJECT
-+
-+ @Description Macro to iterate over list of given type.
-+
-+ @Param[in] p_Pos - A pointer to a list to use as a loop counter.
-+ @Param[in] type - The type of the struct this is embedded in.
-+ @Param[in] p_Head - A pointer to the head for your list pointer.
-+ @Param[in] member - The name of the list_struct within the struct.
-+
-+ @Cautions You can't delete items with this routine.
-+ For deletion use LIST_FOR_EACH_SAFE().
-+*//***************************************************************************/
-+#define LIST_FOR_EACH_OBJECT(p_Pos, type, p_Head, member) \
-+ for (p_Pos = LIST_OBJECT(LIST_FIRST(p_Head), type, member); \
-+ &p_Pos->member != (p_Head); \
-+ p_Pos = LIST_OBJECT(LIST_FIRST(&(p_Pos->member)), type, member))
-+
-+
-+/**************************************************************************//**
-+ @Function LIST_Add
-+
-+ @Description Add a new entry to a list.
-+
-+ Insert a new entry after the specified head.
-+ This is good for implementing stacks.
-+
-+ @Param[in] p_New - A pointer to a new list entry to be added.
-+ @Param[in] p_Head - A pointer to a list head to add it after.
-+
-+ @Return none.
-+*//***************************************************************************/
-+static __inline__ void LIST_Add(t_List *p_New, t_List *p_Head)
-+{
-+ LIST_PREV(LIST_NEXT(p_Head)) = p_New;
-+ LIST_NEXT(p_New) = LIST_NEXT(p_Head);
-+ LIST_PREV(p_New) = p_Head;
-+ LIST_NEXT(p_Head) = p_New;
-+}
-+
-+
-+/**************************************************************************//**
-+ @Function LIST_AddToTail
-+
-+ @Description Add a new entry to a list.
-+
-+ Insert a new entry before the specified head.
-+ This is useful for implementing queues.
-+
-+ @Param[in] p_New - A pointer to a new list entry to be added.
-+ @Param[in] p_Head - A pointer to a list head to add it before.
-+
-+ @Return none.
-+*//***************************************************************************/
-+static __inline__ void LIST_AddToTail(t_List *p_New, t_List *p_Head)
-+{
-+ LIST_NEXT(LIST_PREV(p_Head)) = p_New;
-+ LIST_PREV(p_New) = LIST_PREV(p_Head);
-+ LIST_NEXT(p_New) = p_Head;
-+ LIST_PREV(p_Head) = p_New;
-+}
-+
-+
-+/**************************************************************************//**
-+ @Function LIST_Del
-+
-+ @Description Deletes entry from a list.
-+
-+ @Param[in] p_Entry - A pointer to the element to delete from the list.
-+
-+ @Return none.
-+
-+ @Cautions LIST_IsEmpty() on entry does not return true after this,
-+ the entry is in an undefined state.
-+*//***************************************************************************/
-+static __inline__ void LIST_Del(t_List *p_Entry)
-+{
-+ LIST_PREV(LIST_NEXT(p_Entry)) = LIST_PREV(p_Entry);
-+ LIST_NEXT(LIST_PREV(p_Entry)) = LIST_NEXT(p_Entry);
-+}
-+
-+
-+/**************************************************************************//**
-+ @Function LIST_DelAndInit
-+
-+ @Description Deletes entry from list and reinitialize it.
-+
-+ @Param[in] p_Entry - A pointer to the element to delete from the list.
-+
-+ @Return none.
-+*//***************************************************************************/
-+static __inline__ void LIST_DelAndInit(t_List *p_Entry)
-+{
-+ LIST_Del(p_Entry);
-+ INIT_LIST(p_Entry);
-+}
-+
-+
-+/**************************************************************************//**
-+ @Function LIST_Move
-+
-+ @Description Delete from one list and add as another's head.
-+
-+ @Param[in] p_Entry - A pointer to the list entry to move.
-+ @Param[in] p_Head - A pointer to the list head that will precede our entry.
-+
-+ @Return none.
-+*//***************************************************************************/
-+static __inline__ void LIST_Move(t_List *p_Entry, t_List *p_Head)
-+{
-+ LIST_Del(p_Entry);
-+ LIST_Add(p_Entry, p_Head);
-+}
-+
-+
-+/**************************************************************************//**
-+ @Function LIST_MoveToTail
-+
-+ @Description Delete from one list and add as another's tail.
-+
-+ @Param[in] p_Entry - A pointer to the entry to move.
-+ @Param[in] p_Head - A pointer to the list head that will follow our entry.
-+
-+ @Return none.
-+*//***************************************************************************/
-+static __inline__ void LIST_MoveToTail(t_List *p_Entry, t_List *p_Head)
-+{
-+ LIST_Del(p_Entry);
-+ LIST_AddToTail(p_Entry, p_Head);
-+}
-+
-+
-+/**************************************************************************//**
-+ @Function LIST_IsEmpty
-+
-+ @Description Tests whether a list is empty.
-+
-+ @Param[in] p_List - A pointer to the list to test.
-+
-+ @Return 1 if the list is empty, 0 otherwise.
-+*//***************************************************************************/
-+static __inline__ int LIST_IsEmpty(t_List *p_List)
-+{
-+ return (LIST_FIRST(p_List) == p_List);
-+}
-+
-+
-+/**************************************************************************//**
-+ @Function LIST_Append
-+
-+ @Description Join two lists.
-+
-+ @Param[in] p_NewList - A pointer to the new list to add.
-+ @Param[in] p_Head - A pointer to the place to add it in the first list.
-+
-+ @Return none.
-+*//***************************************************************************/
-+void LIST_Append(t_List *p_NewList, t_List *p_Head);
-+
-+
-+/**************************************************************************//**
-+ @Function LIST_NumOfObjs
-+
-+ @Description Counts number of objects in the list
-+
-+ @Param[in] p_List - A pointer to the list which objects are to be counted.
-+
-+ @Return Number of objects in the list.
-+*//***************************************************************************/
-+int LIST_NumOfObjs(t_List *p_List);
-+
-+/** @} */ /* end of list_id group */
-+/** @} */ /* end of etc_id group */
-+
-+
-+#endif /* __LIST_EXT_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/etc/mem_ext.h
-@@ -0,0 +1,318 @@
-+/* Copyright (c) 2008-2012 Freescale Semiconductor, Inc
-+ * All rights reserved.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/**************************************************************************//**
-+
-+ @File mem_ext.h
-+
-+ @Description External prototypes for the memory manager object
-+*//***************************************************************************/
-+
-+#ifndef __MEM_EXT_H
-+#define __MEM_EXT_H
-+
-+#include "std_ext.h"
-+#include "part_ext.h"
-+
-+
-+/**************************************************************************//**
-+ @Group etc_id Utility Library Application Programming Interface
-+
-+ @Description External routines.
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Group mem_id Slab Memory Manager
-+
-+ @Description Slab Memory Manager module functions, definitions and enums.
-+
-+ @{
-+*//***************************************************************************/
-+
-+/* Each block is of the following structure:
-+ *
-+ *
-+ * +-----------+----------+---------------------------+-----------+-----------+
-+ * | Alignment | Prefix | Data | Postfix | Alignment |
-+ * | field | field | field | field | Padding |
-+ * | | | | | |
-+ * +-----------+----------+---------------------------+-----------+-----------+
-+ * and at the beginning of all bytes, an additional optional padding might reside
-+ * to ensure that the first blocks data field is aligned as requested.
-+ */
-+
-+
-+#define MEM_MAX_NAME_LENGTH 8
-+
-+/**************************************************************************//*
-+ @Description Memory Segment structure
-+*//***************************************************************************/
-+
-+typedef struct
-+{
-+ char name[MEM_MAX_NAME_LENGTH];
-+ /* The segment's name */
-+ uint8_t **p_Bases; /* Base addresses of the segments */
-+ uint8_t **p_BlocksStack; /* Array of pointers to blocks */
-+ t_Handle h_Spinlock;
-+ uint16_t dataSize; /* Size of each data block */
-+ uint16_t prefixSize; /* How many bytes to reserve before the data */
-+ uint16_t postfixSize; /* How many bytes to reserve after the data */
-+ uint16_t alignment; /* Requested alignment for the data field */
-+ int allocOwner; /* Memory allocation owner */
-+ uint32_t getFailures; /* Number of times get failed */
-+ uint32_t num; /* Number of blocks in segment */
-+ uint32_t current; /* Current block */
-+ bool consecutiveMem; /* Allocate consecutive data blocks memory */
-+#ifdef DEBUG_MEM_LEAKS
-+ void *p_MemDbg; /* MEM debug database (MEM leaks detection) */
-+ uint32_t blockOffset;
-+ uint32_t blockSize;
-+#endif /* DEBUG_MEM_LEAKS */
-+} t_MemorySegment;
-+
-+
-+
-+/**************************************************************************//**
-+ @Function MEM_Init
-+
-+ @Description Create a new memory segment.
-+
-+ @Param[in] name - Name of memory partition.
-+ @Param[in] p_Handle - Handle to new segment is returned through here.
-+ @Param[in] num - Number of blocks in new segment.
-+ @Param[in] dataSize - Size of blocks in segment.
-+ @Param[in] prefixSize - How many bytes to allocate before the data.
-+ @Param[in] postfixSize - How many bytes to allocate after the data.
-+ @Param[in] alignment - Requested alignment for data field (in bytes).
-+
-+ @Return E_OK - success, E_NO_MEMORY - out of memory.
-+*//***************************************************************************/
-+t_Error MEM_Init(char name[],
-+ t_Handle *p_Handle,
-+ uint32_t num,
-+ uint16_t dataSize,
-+ uint16_t prefixSize,
-+ uint16_t postfixSize,
-+ uint16_t alignment);
-+
-+/**************************************************************************//**
-+ @Function MEM_InitSmart
-+
-+ @Description Create a new memory segment.
-+
-+ @Param[in] name - Name of memory partition.
-+ @Param[in] p_Handle - Handle to new segment is returned through here.
-+ @Param[in] num - Number of blocks in new segment.
-+ @Param[in] dataSize - Size of blocks in segment.
-+ @Param[in] prefixSize - How many bytes to allocate before the data.
-+ @Param[in] postfixSize - How many bytes to allocate after the data.
-+ @Param[in] alignment - Requested alignment for data field (in bytes).
-+ @Param[in] memPartitionId - Memory partition ID for allocation.
-+ @Param[in] consecutiveMem - Whether to allocate the memory blocks
-+ continuously or not.
-+
-+ @Return E_OK - success, E_NO_MEMORY - out of memory.
-+*//***************************************************************************/
-+t_Error MEM_InitSmart(char name[],
-+ t_Handle *p_Handle,
-+ uint32_t num,
-+ uint16_t dataSize,
-+ uint16_t prefixSize,
-+ uint16_t postfixSize,
-+ uint16_t alignment,
-+ uint8_t memPartitionId,
-+ bool consecutiveMem);
-+
-+/**************************************************************************//**
-+ @Function MEM_InitByAddress
-+
-+ @Description Create a new memory segment with a specified base address.
-+
-+ @Param[in] name - Name of memory partition.
-+ @Param[in] p_Handle - Handle to new segment is returned through here.
-+ @Param[in] num - Number of blocks in new segment.
-+ @Param[in] dataSize - Size of blocks in segment.
-+ @Param[in] prefixSize - How many bytes to allocate before the data.
-+ @Param[in] postfixSize - How many bytes to allocate after the data.
-+ @Param[in] alignment - Requested alignment for data field (in bytes).
-+ @Param[in] address - The required base address.
-+
-+ @Return E_OK - success, E_NO_MEMORY - out of memory.
-+ *//***************************************************************************/
-+t_Error MEM_InitByAddress(char name[],
-+ t_Handle *p_Handle,
-+ uint32_t num,
-+ uint16_t dataSize,
-+ uint16_t prefixSize,
-+ uint16_t postfixSize,
-+ uint16_t alignment,
-+ uint8_t *address);
-+
-+/**************************************************************************//**
-+ @Function MEM_Free
-+
-+ @Description Free a specific memory segment.
-+
-+ @Param[in] h_Mem - Handle to memory segment.
-+
-+ @Return None.
-+*//***************************************************************************/
-+void MEM_Free(t_Handle h_Mem);
-+
-+/**************************************************************************//**
-+ @Function MEM_Get
-+
-+ @Description Get a block of memory from a segment.
-+
-+ @Param[in] h_Mem - Handle to memory segment.
-+
-+ @Return Pointer to new memory block on success,0 otherwise.
-+*//***************************************************************************/
-+void * MEM_Get(t_Handle h_Mem);
-+
-+/**************************************************************************//**
-+ @Function MEM_GetN
-+
-+ @Description Get up to N blocks of memory from a segment.
-+
-+ The blocks are assumed to be of a fixed size (one size per segment).
-+
-+ @Param[in] h_Mem - Handle to memory segment.
-+ @Param[in] num - Number of blocks to allocate.
-+ @Param[out] array - Array of at least num pointers to which the addresses
-+ of the allocated blocks are written.
-+
-+ @Return The number of blocks actually allocated.
-+
-+ @Cautions Interrupts are disabled for all of the allocation loop.
-+ Although this loop is very short for each block (several machine
-+ instructions), you should not allocate a very large number
-+ of blocks via this routine.
-+*//***************************************************************************/
-+uint16_t MEM_GetN(t_Handle h_Mem, uint32_t num, void *array[]);
-+
-+/**************************************************************************//**
-+ @Function MEM_Put
-+
-+ @Description Put a block of memory back to a segment.
-+
-+ @Param[in] h_Mem - Handle to memory segment.
-+ @Param[in] p_Block - The block to return.
-+
-+ @Return Pointer to new memory block on success,0 otherwise.
-+*//***************************************************************************/
-+t_Error MEM_Put(t_Handle h_Mem, void *p_Block);
-+
-+/**************************************************************************//**
-+ @Function MEM_ComputePartitionSize
-+
-+ @Description calculate a tight upper boundary of the size of a partition with
-+ given attributes.
-+
-+ The returned value is suitable if one wants to use MEM_InitByAddress().
-+
-+ @Param[in] num - The number of blocks in the segment.
-+ @Param[in] dataSize - Size of block to get.
-+ @Param[in] prefixSize - The prefix size
-+ @Param postfixSize - The postfix size
-+ @Param[in] alignment - The requested alignment value (in bytes)
-+
-+ @Return The memory block size a segment with the given attributes needs.
-+*//***************************************************************************/
-+uint32_t MEM_ComputePartitionSize(uint32_t num,
-+ uint16_t dataSize,
-+ uint16_t prefixSize,
-+ uint16_t postfixSize,
-+ uint16_t alignment);
-+
-+#ifdef DEBUG_MEM_LEAKS
-+#if !((defined(__MWERKS__) || defined(__GNUC__)) && (__dest_os == __ppc_eabi))
-+#error "Memory-Leaks-Debug option is supported only for freescale CodeWarrior"
-+#endif /* !(defined(__MWERKS__) && ... */
-+
-+/**************************************************************************//**
-+ @Function MEM_CheckLeaks
-+
-+ @Description Report MEM object leaks.
-+
-+ This routine is automatically called by the MEM_Free() routine,
-+ but it can also be invoked while the MEM object is alive.
-+
-+ @Param[in] h_Mem - Handle to memory segment.
-+
-+ @Return None.
-+*//***************************************************************************/
-+void MEM_CheckLeaks(t_Handle h_Mem);
-+
-+#else /* not DEBUG_MEM_LEAKS */
-+#define MEM_CheckLeaks(h_Mem)
-+#endif /* not DEBUG_MEM_LEAKS */
-+
-+/**************************************************************************//**
-+ @Description Get base of MEM
-+*//***************************************************************************/
-+#define MEM_GetBase(h_Mem) ((t_MemorySegment *)(h_Mem))->p_Bases[0]
-+
-+/**************************************************************************//**
-+ @Description Get size of MEM block
-+*//***************************************************************************/
-+#define MEM_GetSize(h_Mem) ((t_MemorySegment *)(h_Mem))->dataSize
-+
-+/**************************************************************************//**
-+ @Description Get prefix size of MEM block
-+*//***************************************************************************/
-+#define MEM_GetPrefixSize(h_Mem) ((t_MemorySegment *)(h_Mem))->prefixSize
-+
-+/**************************************************************************//**
-+ @Description Get postfix size of MEM block
-+*//***************************************************************************/
-+#define MEM_GetPostfixSize(h_Mem) ((t_MemorySegment *)(h_Mem))->postfixSize
-+
-+/**************************************************************************//**
-+ @Description Get alignment of MEM block (in bytes)
-+*//***************************************************************************/
-+#define MEM_GetAlignment(h_Mem) ((t_MemorySegment *)(h_Mem))->alignment
-+
-+/**************************************************************************//**
-+ @Description Get the number of blocks in the segment
-+*//***************************************************************************/
-+#define MEM_GetNumOfBlocks(h_Mem) ((t_MemorySegment *)(h_Mem))->num
-+
-+/** @} */ /* end of MEM group */
-+/** @} */ /* end of etc_id group */
-+
-+
-+#endif /* __MEM_EXT_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/etc/memcpy_ext.h
-@@ -0,0 +1,208 @@
-+/* Copyright (c) 2008-2012 Freescale Semiconductor, Inc
-+ * All rights reserved.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/**************************************************************************//**
-+
-+ @File memcpy_ext.h
-+
-+ @Description Efficient functions for copying and setting blocks of memory.
-+*//***************************************************************************/
-+
-+#ifndef __MEMCPY_EXT_H
-+#define __MEMCPY_EXT_H
-+
-+#include "std_ext.h"
-+
-+
-+/**************************************************************************//**
-+ @Group etc_id Utility Library Application Programming Interface
-+
-+ @Description External routines.
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Group mem_cpy Memory Copy
-+
-+ @Description Memory Copy module functions,definitions and enums.
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Function MemCpy32
-+
-+ @Description Copies one memory buffer into another one in 4-byte chunks!
-+ Which should be more efficient than byte by byte.
-+
-+ For large buffers (over 60 bytes) this function is about 4 times
-+ more efficient than the trivial memory copy. For short buffers
-+ it is reduced to the trivial copy and may be a bit worse.
-+
-+ @Param[in] pDst - The address of the destination buffer.
-+ @Param[in] pSrc - The address of the source buffer.
-+ @Param[in] size - The number of bytes that will be copied from pSrc to pDst.
-+
-+ @Return pDst (the address of the destination buffer).
-+
-+ @Cautions There is no parameter or boundary checking! It is up to the user
-+ to supply non-null parameters as source & destination and size
-+ that actually fits into the destination buffer.
-+*//***************************************************************************/
-+void * MemCpy32(void* pDst,void* pSrc, uint32_t size);
-+void * IO2IOCpy32(void* pDst,void* pSrc, uint32_t size);
-+void * IO2MemCpy32(void* pDst,void* pSrc, uint32_t size);
-+void * Mem2IOCpy32(void* pDst,void* pSrc, uint32_t size);
-+
-+/**************************************************************************//**
-+ @Function MemCpy64
-+
-+ @Description Copies one memory buffer into another one in 8-byte chunks!
-+ Which should be more efficient than byte by byte.
-+
-+ For large buffers (over 60 bytes) this function is about 8 times
-+ more efficient than the trivial memory copy. For short buffers
-+ it is reduced to the trivial copy and may be a bit worse.
-+
-+ Some testing suggests that MemCpy32() preforms better than
-+ MemCpy64() over small buffers. On average they break even at
-+ 100 byte buffers. For buffers larger than that MemCpy64 is
-+ superior.
-+
-+ @Param[in] pDst - The address of the destination buffer.
-+ @Param[in] pSrc - The address of the source buffer.
-+ @Param[in] size - The number of bytes that will be copied from pSrc to pDst.
-+
-+ @Return pDst (the address of the destination buffer).
-+
-+ @Cautions There is no parameter or boundary checking! It is up to the user
-+ to supply non null parameters as source & destination and size
-+ that actually fits into their buffer.
-+
-+ Do not use under Linux.
-+*//***************************************************************************/
-+void * MemCpy64(void* pDst,void* pSrc, uint32_t size);
-+
-+/**************************************************************************//**
-+ @Function MemSet32
-+
-+ @Description Sets all bytes of a memory buffer to a specific value, in
-+ 4-byte chunks.
-+
-+ @Param[in] pDst - The address of the destination buffer.
-+ @Param[in] val - Value to set destination bytes to.
-+ @Param[in] size - The number of bytes that will be set to val.
-+
-+ @Return pDst (the address of the destination buffer).
-+
-+ @Cautions There is no parameter or boundary checking! It is up to the user
-+ to supply non null parameter as destination and size
-+ that actually fits into the destination buffer.
-+*//***************************************************************************/
-+void * MemSet32(void* pDst, uint8_t val, uint32_t size);
-+void * IOMemSet32(void* pDst, uint8_t val, uint32_t size);
-+
-+/**************************************************************************//**
-+ @Function MemSet64
-+
-+ @Description Sets all bytes of a memory buffer to a specific value, in
-+ 8-byte chunks.
-+
-+ @Param[in] pDst - The address of the destination buffer.
-+ @Param[in] val - Value to set destination bytes to.
-+ @Param[in] size - The number of bytes that will be set to val.
-+
-+ @Return pDst (the address of the destination buffer).
-+
-+ @Cautions There is no parameter or boundary checking! It is up to the user
-+ to supply non null parameter as destination and size
-+ that actually fits into the destination buffer.
-+*//***************************************************************************/
-+void * MemSet64(void* pDst, uint8_t val, uint32_t size);
-+
-+/**************************************************************************//**
-+ @Function MemDisp
-+
-+ @Description Displays a block of memory in chunks of 32 bits.
-+
-+ @Param[in] addr - The address of the memory to display.
-+ @Param[in] size - The number of bytes that will be displayed.
-+
-+ @Return None.
-+
-+ @Cautions There is no parameter or boundary checking! It is up to the user
-+ to supply non null parameter as destination and size
-+ that actually fits into the destination buffer.
-+*//***************************************************************************/
-+void MemDisp(uint8_t *addr, int size);
-+
-+/**************************************************************************//**
-+ @Function MemCpy8
-+
-+ @Description Trivial copy one memory buffer into another byte by byte
-+
-+ @Param[in] pDst - The address of the destination buffer.
-+ @Param[in] pSrc - The address of the source buffer.
-+ @Param[in] size - The number of bytes that will be copied from pSrc to pDst.
-+
-+ @Return pDst (the address of the destination buffer).
-+
-+ @Cautions There is no parameter or boundary checking! It is up to the user
-+ to supply non-null parameters as source & destination and size
-+ that actually fits into the destination buffer.
-+*//***************************************************************************/
-+void * MemCpy8(void* pDst,void* pSrc, uint32_t size);
-+
-+/**************************************************************************//**
-+ @Function MemSet8
-+
-+ @Description Sets all bytes of a memory buffer to a specific value byte by byte.
-+
-+ @Param[in] pDst - The address of the destination buffer.
-+ @Param[in] c - Value to set destination bytes to.
-+ @Param[in] size - The number of bytes that will be set to val.
-+
-+ @Return pDst (the address of the destination buffer).
-+
-+ @Cautions There is no parameter or boundary checking! It is up to the user
-+ to supply non null parameter as destination and size
-+ that actually fits into the destination buffer.
-+*//***************************************************************************/
-+void * MemSet8(void* pDst, int c, uint32_t size);
-+
-+/** @} */ /* end of mem_cpy group */
-+/** @} */ /* end of etc_id group */
-+
-+
-+#endif /* __MEMCPY_EXT_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/etc/mm_ext.h
-@@ -0,0 +1,310 @@
-+/* Copyright (c) 2008-2012 Freescale Semiconductor, Inc
-+ * All rights reserved.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/**************************************************************************//**
-+ @File mm_ext.h
-+
-+ @Description Memory Manager Application Programming Interface
-+*//***************************************************************************/
-+#ifndef __MM_EXT
-+#define __MM_EXT
-+
-+#include "std_ext.h"
-+
-+#define MM_MAX_ALIGNMENT 20 /* Alignments from 2 to 128 are available
-+ where maximum alignment defined as
-+ MM_MAX_ALIGNMENT power of 2 */
-+
-+#define MM_MAX_NAME_LEN 32
-+
-+/**************************************************************************//**
-+ @Group etc_id Utility Library Application Programming Interface
-+
-+ @Description External routines.
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Group mm_grp Flexible Memory Manager
-+
-+ @Description Flexible Memory Manager module functions,definitions and enums.
-+ (All of the following functions,definitions and enums can be found in mm_ext.h)
-+
-+ @{
-+*//***************************************************************************/
-+
-+
-+/**************************************************************************//**
-+ @Function MM_Init
-+
-+ @Description Initializes a new MM object.
-+
-+ It initializes a new memory block consisting of base address
-+ and size of the available memory by calling to MemBlock_Init
-+ routine. It is also initializes a new free block for each
-+ by calling FreeBlock_Init routine, which is pointed to
-+ the almost all memory started from the required alignment
-+ from the base address and to the end of the memory.
-+ The handle to the new MM object is returned via "MM"
-+ argument (passed by reference).
-+
-+ @Param[in] h_MM - Handle to the MM object.
-+ @Param[in] base - Base address of the MM.
-+ @Param[in] size - Size of the MM.
-+
-+ @Return E_OK is returned on success. E_NOMEMORY is returned if the new MM object or a new free block can not be initialized.
-+*//***************************************************************************/
-+t_Error MM_Init(t_Handle *h_MM, uint64_t base, uint64_t size);
-+
-+/**************************************************************************//**
-+ @Function MM_Get
-+
-+ @Description Allocates a block of memory according to the given size and the alignment.
-+
-+ The Alignment argument tells from which
-+ free list allocate a block of memory. 2^alignment indicates
-+ the alignment that the base address of the allocated block
-+ should have. So, the only values 1, 2, 4, 8, 16, 32 and 64
-+ are available for the alignment argument.
-+ The routine passes through the specific free list of free
-+ blocks and seeks for a first block that have anough memory
-+ that is required (best fit).
-+ After the block is found and data is allocated, it calls
-+ the internal MM_CutFree routine to update all free lists
-+ do not include a just allocated block. Of course, each
-+ free list contains a free blocks with the same alignment.
-+ It is also creates a busy block that holds
-+ information about an allocated block.
-+
-+ @Param[in] h_MM - Handle to the MM object.
-+ @Param[in] size - Size of the MM.
-+ @Param[in] alignment - Index as a power of two defines a required
-+ alignment (in bytes); Should be 1, 2, 4, 8, 16, 32 or 64
-+ @Param[in] name - The name that specifies an allocated block.
-+
-+ @Return base address of an allocated block ILLEGAL_BASE if can't allocate a block
-+*//***************************************************************************/
-+uint64_t MM_Get(t_Handle h_MM, uint64_t size, uint64_t alignment, char *name);
-+
-+/**************************************************************************//**
-+ @Function MM_GetBase
-+
-+ @Description Gets the base address of the required MM objects.
-+
-+ @Param[in] h_MM - Handle to the MM object.
-+
-+ @Return base address of the block.
-+*//***************************************************************************/
-+uint64_t MM_GetBase(t_Handle h_MM);
-+
-+/**************************************************************************//**
-+ @Function MM_GetForce
-+
-+ @Description Force memory allocation.
-+
-+ It means to allocate a block of memory of the given
-+ size from the given base address.
-+ The routine checks if the required block can be allocated
-+ (that is it is free) and then, calls the internal MM_CutFree
-+ routine to update all free lists do not include that block.
-+
-+ @Param[in] h_MM - Handle to the MM object.
-+ @Param[in] base - Base address of the MM.
-+ @Param[in] size - Size of the MM.
-+ @Param[in] name - Name that specifies an allocated block.
-+
-+ @Return base address of an allocated block, ILLEGAL_BASE if can't allocate a block.
-+*//***************************************************************************/
-+uint64_t MM_GetForce(t_Handle h_MM, uint64_t base, uint64_t size, char *name);
-+
-+/**************************************************************************//**
-+ @Function MM_GetForceMin
-+
-+ @Description Allocates a block of memory according to the given size, the alignment and minimum base address.
-+
-+ The Alignment argument tells from which
-+ free list allocate a block of memory. 2^alignment indicates
-+ the alignment that the base address of the allocated block
-+ should have. So, the only values 1, 2, 4, 8, 16, 32 and 64
-+ are available for the alignment argument.
-+ The minimum baser address forces the location of the block
-+ to be from a given address onward.
-+ The routine passes through the specific free list of free
-+ blocks and seeks for the first base address equal or smaller
-+ than the required minimum address and end address larger than
-+ than the required base + its size - i.e. that may contain
-+ the required block.
-+ After the block is found and data is allocated, it calls
-+ the internal MM_CutFree routine to update all free lists
-+ do not include a just allocated block. Of course, each
-+ free list contains a free blocks with the same alignment.
-+ It is also creates a busy block that holds
-+ information about an allocated block.
-+
-+ @Param[in] h_MM - Handle to the MM object.
-+ @Param[in] size - Size of the MM.
-+ @Param[in] alignment - Index as a power of two defines a required
-+ alignment (in bytes); Should be 1, 2, 4, 8, 16, 32 or 64
-+ @Param[in] min - The minimum base address of the block.
-+ @Param[in] name - Name that specifies an allocated block.
-+
-+ @Return base address of an allocated block,ILLEGAL_BASE if can't allocate a block.
-+*//***************************************************************************/
-+uint64_t MM_GetForceMin(t_Handle h_MM,
-+ uint64_t size,
-+ uint64_t alignment,
-+ uint64_t min,
-+ char *name);
-+
-+/**************************************************************************//**
-+ @Function MM_Put
-+
-+ @Description Puts a block of memory of the given base address back to the memory.
-+
-+ It checks if there is a busy block with the
-+ given base address. If not, it returns 0, that
-+ means can't free a block. Otherwise, it gets parameters of
-+ the busy block and after it updates lists of free blocks,
-+ removes that busy block from the list by calling to MM_CutBusy
-+ routine.
-+ After that it calls to MM_AddFree routine to add a new free
-+ block to the free lists.
-+
-+ @Param[in] h_MM - Handle to the MM object.
-+ @Param[in] base - Base address of the MM.
-+
-+ @Return The size of bytes released, 0 if failed.
-+*//***************************************************************************/
-+uint64_t MM_Put(t_Handle h_MM, uint64_t base);
-+
-+/**************************************************************************//**
-+ @Function MM_PutForce
-+
-+ @Description Releases a block of memory of the required size from the required base address.
-+
-+ First, it calls to MM_CutBusy routine
-+ to cut a free block from the busy list. And then, calls to
-+ MM_AddFree routine to add the free block to the free lists.
-+
-+ @Param[in] h_MM - Handle to the MM object.
-+ @Param[in] base - Base address of of a block to free.
-+ @Param[in] size - Size of a block to free.
-+
-+ @Return The number of bytes released, 0 on failure.
-+*//***************************************************************************/
-+uint64_t MM_PutForce(t_Handle h_MM, uint64_t base, uint64_t size);
-+
-+/**************************************************************************//**
-+ @Function MM_Add
-+
-+ @Description Adds a new memory block for memory allocation.
-+
-+ When a new memory block is initialized and added to the
-+ memory list, it calls to MM_AddFree routine to add the
-+ new free block to the free lists.
-+
-+ @Param[in] h_MM - Handle to the MM object.
-+ @Param[in] base - Base address of the memory block.
-+ @Param[in] size - Size of the memory block.
-+
-+ @Return E_OK on success, otherwise returns an error code.
-+*//***************************************************************************/
-+t_Error MM_Add(t_Handle h_MM, uint64_t base, uint64_t size);
-+
-+/**************************************************************************//**
-+ @Function MM_Dump
-+
-+ @Description Prints results of free and busy lists.
-+
-+ @Param[in] h_MM - Handle to the MM object.
-+*//***************************************************************************/
-+void MM_Dump(t_Handle h_MM);
-+
-+/**************************************************************************//**
-+ @Function MM_Free
-+
-+ @Description Releases memory allocated for MM object.
-+
-+ @Param[in] h_MM - Handle of the MM object.
-+*//***************************************************************************/
-+void MM_Free(t_Handle h_MM);
-+
-+/**************************************************************************//**
-+ @Function MM_GetMemBlock
-+
-+ @Description Returns base address of the memory block specified by the index.
-+
-+ If index is 0, returns base address
-+ of the first memory block, 1 - returns base address
-+ of the second memory block, etc.
-+ Note, those memory blocks are allocated by the
-+ application before MM_Init or MM_Add and have to
-+ be released by the application before or after invoking
-+ the MM_Free routine.
-+
-+ @Param[in] h_MM - Handle to the MM object.
-+ @Param[in] index - Index of the memory block.
-+
-+ @Return valid base address or ILLEGAL_BASE if no memory block specified by the index.
-+*//***************************************************************************/
-+uint64_t MM_GetMemBlock(t_Handle h_MM, int index);
-+
-+/**************************************************************************//**
-+ @Function MM_InRange
-+
-+ @Description Checks if a specific address is in the memory range of the passed MM object.
-+
-+ @Param[in] h_MM - Handle to the MM object.
-+ @Param[in] addr - The address to be checked.
-+
-+ @Return TRUE if the address is in the address range of the block, FALSE otherwise.
-+*//***************************************************************************/
-+bool MM_InRange(t_Handle h_MM, uint64_t addr);
-+
-+/**************************************************************************//**
-+ @Function MM_GetFreeMemSize
-+
-+ @Description Returns the size (in bytes) of free memory.
-+
-+ @Param[in] h_MM - Handle to the MM object.
-+
-+ @Return Free memory size in bytes.
-+*//***************************************************************************/
-+uint64_t MM_GetFreeMemSize(t_Handle h_MM);
-+
-+
-+/** @} */ /* end of mm_grp group */
-+/** @} */ /* end of etc_id group */
-+
-+#endif /* __MM_EXT_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/etc/sprint_ext.h
-@@ -0,0 +1,118 @@
-+/* Copyright (c) 2008-2012 Freescale Semiconductor, Inc
-+ * All rights reserved.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/**************************************************************************//**
-+ @File sprint_ext.h
-+
-+ @Description Debug routines (externals).
-+
-+*//***************************************************************************/
-+
-+#ifndef __SPRINT_EXT_H
-+#define __SPRINT_EXT_H
-+
-+
-+#if defined(NCSW_LINUX) && defined(__KERNEL__)
-+#include <linux/kernel.h>
-+
-+#elif defined(NCSW_VXWORKS)
-+#include "private/stdioP.h"
-+
-+#else
-+#include <stdio.h>
-+#endif /* defined(NCSW_LINUX) && defined(__KERNEL__) */
-+
-+#include "std_ext.h"
-+
-+
-+/**************************************************************************//**
-+ @Group etc_id Utility Library Application Programming Interface
-+
-+ @Description External routines.
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Group sprint_id Sprint
-+
-+ @Description Sprint & Sscan module functions,definitions and enums.
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Function Sprint
-+
-+ @Description Format a string and place it in a buffer.
-+
-+ @Param[in] buff - The buffer to place the result into.
-+ @Param[in] str - The format string to use.
-+ @Param[in] ... - Arguments for the format string.
-+
-+ @Return Number of bytes formatted.
-+*//***************************************************************************/
-+int Sprint(char *buff, const char *str, ...);
-+
-+/**************************************************************************//**
-+ @Function Snprint
-+
-+ @Description Format a string and place it in a buffer.
-+
-+ @Param[in] buf - The buffer to place the result into.
-+ @Param[in] size - The size of the buffer, including the trailing null space.
-+ @Param[in] fmt - The format string to use.
-+ @Param[in] ... - Arguments for the format string.
-+
-+ @Return Number of bytes formatted.
-+*//***************************************************************************/
-+int Snprint(char * buf, uint32_t size, const char *fmt, ...);
-+
-+/**************************************************************************//**
-+ @Function Sscan
-+
-+ @Description Unformat a buffer into a list of arguments.
-+
-+ @Param[in] buf - input buffer.
-+ @Param[in] fmt - formatting of buffer.
-+ @Param[out] ... - resulting arguments.
-+
-+ @Return Number of bytes unformatted.
-+*//***************************************************************************/
-+int Sscan(const char * buf, const char * fmt, ...);
-+
-+/** @} */ /* end of sprint_id group */
-+/** @} */ /* end of etc_id group */
-+
-+
-+#endif /* __SPRINT_EXT_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/flib/common/arch/ppc_access.h
-@@ -0,0 +1,37 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#ifndef FL_E500_MACROS_H
-+#define FL_E500_MACROS_H
-+
-+#endif /* FL_E500_MACROS_H */
-+
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/flib/common/general.h
-@@ -0,0 +1,52 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#ifndef __GENERAL_H
-+#define __GENERAL_H
-+
-+#include "std_ext.h"
-+#if !defined(NCSW_LINUX)
-+#include "errno.h"
-+#endif
-+
-+
-+extern uint32_t get_mac_addr_crc(uint64_t _addr);
-+
-+#ifndef CONFIG_FMAN_ARM
-+#define iowrite32be(val, addr) WRITE_UINT32(*addr, val)
-+#define ioread32be(addr) GET_UINT32(*addr)
-+#endif
-+
-+#define ether_crc(len, addr) get_mac_addr_crc(*(uint64_t *)(addr)>>16)
-+
-+
-+#endif /* __GENERAL_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/flib/fman_common.h
-@@ -0,0 +1,78 @@
-+/*
-+ * Copyright 2008-2013 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+#ifndef __FMAN_COMMON_H
-+#define __FMAN_COMMON_H
-+
-+/**************************************************************************//**
-+ @Description NIA Description
-+*//***************************************************************************/
-+#define NIA_ORDER_RESTOR 0x00800000
-+#define NIA_ENG_FM_CTL 0x00000000
-+#define NIA_ENG_PRS 0x00440000
-+#define NIA_ENG_KG 0x00480000
-+#define NIA_ENG_PLCR 0x004C0000
-+#define NIA_ENG_BMI 0x00500000
-+#define NIA_ENG_QMI_ENQ 0x00540000
-+#define NIA_ENG_QMI_DEQ 0x00580000
-+#define NIA_ENG_MASK 0x007C0000
-+
-+#define NIA_FM_CTL_AC_CC 0x00000006
-+#define NIA_FM_CTL_AC_HC 0x0000000C
-+#define NIA_FM_CTL_AC_IND_MODE_TX 0x00000008
-+#define NIA_FM_CTL_AC_IND_MODE_RX 0x0000000A
-+#define NIA_FM_CTL_AC_FRAG 0x0000000e
-+#define NIA_FM_CTL_AC_PRE_FETCH 0x00000010
-+#define NIA_FM_CTL_AC_POST_FETCH_PCD 0x00000012
-+#define NIA_FM_CTL_AC_POST_FETCH_PCD_UDP_LEN 0x00000018
-+#define NIA_FM_CTL_AC_POST_FETCH_NO_PCD 0x00000012
-+#define NIA_FM_CTL_AC_FRAG_CHECK 0x00000014
-+#define NIA_FM_CTL_AC_PRE_CC 0x00000020
-+
-+
-+#define NIA_BMI_AC_ENQ_FRAME 0x00000002
-+#define NIA_BMI_AC_TX_RELEASE 0x000002C0
-+#define NIA_BMI_AC_RELEASE 0x000000C0
-+#define NIA_BMI_AC_DISCARD 0x000000C1
-+#define NIA_BMI_AC_TX 0x00000274
-+#define NIA_BMI_AC_FETCH 0x00000208
-+#define NIA_BMI_AC_MASK 0x000003FF
-+
-+#define NIA_KG_DIRECT 0x00000100
-+#define NIA_KG_CC_EN 0x00000200
-+#define NIA_PLCR_ABSOLUTE 0x00008000
-+
-+#define NIA_BMI_AC_ENQ_FRAME_WITHOUT_DMA 0x00000202
-+#define NIA_BMI_AC_FETCH_ALL_FRAME 0x0000020c
-+
-+#endif /* __FMAN_COMMON_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/flib/fsl_enet.h
-@@ -0,0 +1,273 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#ifndef __FSL_ENET_H
-+#define __FSL_ENET_H
-+
-+/**
-+ @Description Ethernet MAC-PHY Interface
-+*/
-+
-+enum enet_interface {
-+ E_ENET_IF_MII = 0x00010000, /**< MII interface */
-+ E_ENET_IF_RMII = 0x00020000, /**< RMII interface */
-+ E_ENET_IF_SMII = 0x00030000, /**< SMII interface */
-+ E_ENET_IF_GMII = 0x00040000, /**< GMII interface */
-+ E_ENET_IF_RGMII = 0x00050000, /**< RGMII interface */
-+ E_ENET_IF_TBI = 0x00060000, /**< TBI interface */
-+ E_ENET_IF_RTBI = 0x00070000, /**< RTBI interface */
-+ E_ENET_IF_SGMII = 0x00080000, /**< SGMII interface */
-+ E_ENET_IF_XGMII = 0x00090000, /**< XGMII interface */
-+ E_ENET_IF_QSGMII = 0x000a0000, /**< QSGMII interface */
-+ E_ENET_IF_XFI = 0x000b0000 /**< XFI interface */
-+};
-+
-+/**
-+ @Description Ethernet Speed (nominal data rate)
-+*/
-+enum enet_speed {
-+ E_ENET_SPEED_10 = 10, /**< 10 Mbps */
-+ E_ENET_SPEED_100 = 100, /**< 100 Mbps */
-+ E_ENET_SPEED_1000 = 1000, /**< 1000 Mbps = 1 Gbps */
-+ E_ENET_SPEED_2500 = 2500, /**< 2500 Mbps = 2.5 Gbps */
-+ E_ENET_SPEED_10000 = 10000 /**< 10000 Mbps = 10 Gbps */
-+};
-+
-+enum mac_type {
-+ E_MAC_DTSEC,
-+ E_MAC_TGEC,
-+ E_MAC_MEMAC
-+};
-+
-+/**************************************************************************//**
-+ @Description Enum for inter-module interrupts registration
-+*//***************************************************************************/
-+enum fman_event_modules {
-+ E_FMAN_MOD_PRS, /**< Parser event */
-+ E_FMAN_MOD_KG, /**< Keygen event */
-+ E_FMAN_MOD_PLCR, /**< Policer event */
-+ E_FMAN_MOD_10G_MAC, /**< 10G MAC event */
-+ E_FMAN_MOD_1G_MAC, /**< 1G MAC event */
-+ E_FMAN_MOD_TMR, /**< Timer event */
-+ E_FMAN_MOD_FMAN_CTRL, /**< FMAN Controller Timer event */
-+ E_FMAN_MOD_MACSEC,
-+ E_FMAN_MOD_DUMMY_LAST
-+};
-+
-+/**************************************************************************//**
-+ @Description Enum for interrupts types
-+*//***************************************************************************/
-+enum fman_intr_type {
-+ E_FMAN_INTR_TYPE_ERR,
-+ E_FMAN_INTR_TYPE_NORMAL
-+};
-+
-+/**************************************************************************//**
-+ @Description enum for defining MAC types
-+*//***************************************************************************/
-+enum fman_mac_type {
-+ E_FMAN_MAC_10G = 0, /**< 10G MAC */
-+ E_FMAN_MAC_1G /**< 1G MAC */
-+};
-+
-+enum fman_mac_exceptions {
-+ E_FMAN_MAC_EX_10G_MDIO_SCAN_EVENTMDIO = 0,
-+ /**< 10GEC MDIO scan event interrupt */
-+ E_FMAN_MAC_EX_10G_MDIO_CMD_CMPL,
-+ /**< 10GEC MDIO command completion interrupt */
-+ E_FMAN_MAC_EX_10G_REM_FAULT,
-+ /**< 10GEC, mEMAC Remote fault interrupt */
-+ E_FMAN_MAC_EX_10G_LOC_FAULT,
-+ /**< 10GEC, mEMAC Local fault interrupt */
-+ E_FMAN_MAC_EX_10G_1TX_ECC_ER,
-+ /**< 10GEC, mEMAC Transmit frame ECC error interrupt */
-+ E_FMAN_MAC_EX_10G_TX_FIFO_UNFL,
-+ /**< 10GEC, mEMAC Transmit FIFO underflow interrupt */
-+ E_FMAN_MAC_EX_10G_TX_FIFO_OVFL,
-+ /**< 10GEC, mEMAC Transmit FIFO overflow interrupt */
-+ E_FMAN_MAC_EX_10G_TX_ER,
-+ /**< 10GEC Transmit frame error interrupt */
-+ E_FMAN_MAC_EX_10G_RX_FIFO_OVFL,
-+ /**< 10GEC, mEMAC Receive FIFO overflow interrupt */
-+ E_FMAN_MAC_EX_10G_RX_ECC_ER,
-+ /**< 10GEC, mEMAC Receive frame ECC error interrupt */
-+ E_FMAN_MAC_EX_10G_RX_JAB_FRM,
-+ /**< 10GEC Receive jabber frame interrupt */
-+ E_FMAN_MAC_EX_10G_RX_OVRSZ_FRM,
-+ /**< 10GEC Receive oversized frame interrupt */
-+ E_FMAN_MAC_EX_10G_RX_RUNT_FRM,
-+ /**< 10GEC Receive runt frame interrupt */
-+ E_FMAN_MAC_EX_10G_RX_FRAG_FRM,
-+ /**< 10GEC Receive fragment frame interrupt */
-+ E_FMAN_MAC_EX_10G_RX_LEN_ER,
-+ /**< 10GEC Receive payload length error interrupt */
-+ E_FMAN_MAC_EX_10G_RX_CRC_ER,
-+ /**< 10GEC Receive CRC error interrupt */
-+ E_FMAN_MAC_EX_10G_RX_ALIGN_ER,
-+ /**< 10GEC Receive alignment error interrupt */
-+ E_FMAN_MAC_EX_1G_BAB_RX,
-+ /**< dTSEC Babbling receive error */
-+ E_FMAN_MAC_EX_1G_RX_CTL,
-+ /**< dTSEC Receive control (pause frame) interrupt */
-+ E_FMAN_MAC_EX_1G_GRATEFUL_TX_STP_COMPLET,
-+ /**< dTSEC Graceful transmit stop complete */
-+ E_FMAN_MAC_EX_1G_BAB_TX,
-+ /**< dTSEC Babbling transmit error */
-+ E_FMAN_MAC_EX_1G_TX_CTL,
-+ /**< dTSEC Transmit control (pause frame) interrupt */
-+ E_FMAN_MAC_EX_1G_TX_ERR,
-+ /**< dTSEC Transmit error */
-+ E_FMAN_MAC_EX_1G_LATE_COL,
-+ /**< dTSEC Late collision */
-+ E_FMAN_MAC_EX_1G_COL_RET_LMT,
-+ /**< dTSEC Collision retry limit */
-+ E_FMAN_MAC_EX_1G_TX_FIFO_UNDRN,
-+ /**< dTSEC Transmit FIFO underrun */
-+ E_FMAN_MAC_EX_1G_MAG_PCKT,
-+ /**< dTSEC Magic Packet detection */
-+ E_FMAN_MAC_EX_1G_MII_MNG_RD_COMPLET,
-+ /**< dTSEC MII management read completion */
-+ E_FMAN_MAC_EX_1G_MII_MNG_WR_COMPLET,
-+ /**< dTSEC MII management write completion */
-+ E_FMAN_MAC_EX_1G_GRATEFUL_RX_STP_COMPLET,
-+ /**< dTSEC Graceful receive stop complete */
-+ E_FMAN_MAC_EX_1G_TX_DATA_ERR,
-+ /**< dTSEC Internal data error on transmit */
-+ E_FMAN_MAC_EX_1G_RX_DATA_ERR,
-+ /**< dTSEC Internal data error on receive */
-+ E_FMAN_MAC_EX_1G_1588_TS_RX_ERR,
-+ /**< dTSEC Time-Stamp Receive Error */
-+ E_FMAN_MAC_EX_1G_RX_MIB_CNT_OVFL,
-+ /**< dTSEC MIB counter overflow */
-+ E_FMAN_MAC_EX_TS_FIFO_ECC_ERR,
-+ /**< mEMAC Time-stamp FIFO ECC error interrupt;
-+ not supported on T4240/B4860 rev1 chips */
-+};
-+
-+#define ENET_IF_SGMII_BASEX 0x80000000
-+ /**< SGMII/QSGII interface with 1000BaseX auto-negotiation between MAC
-+ and phy or backplane;
-+ Note: 1000BaseX auto-negotiation relates only to interface between MAC
-+ and phy/backplane, SGMII phy can still synchronize with far-end phy at
-+ 10Mbps, 100Mbps or 1000Mbps */
-+
-+enum enet_mode {
-+ E_ENET_MODE_INVALID = 0,
-+ /**< Invalid Ethernet mode */
-+ E_ENET_MODE_MII_10 = (E_ENET_IF_MII | E_ENET_SPEED_10),
-+ /**< 10 Mbps MII */
-+ E_ENET_MODE_MII_100 = (E_ENET_IF_MII | E_ENET_SPEED_100),
-+ /**< 100 Mbps MII */
-+ E_ENET_MODE_RMII_10 = (E_ENET_IF_RMII | E_ENET_SPEED_10),
-+ /**< 10 Mbps RMII */
-+ E_ENET_MODE_RMII_100 = (E_ENET_IF_RMII | E_ENET_SPEED_100),
-+ /**< 100 Mbps RMII */
-+ E_ENET_MODE_SMII_10 = (E_ENET_IF_SMII | E_ENET_SPEED_10),
-+ /**< 10 Mbps SMII */
-+ E_ENET_MODE_SMII_100 = (E_ENET_IF_SMII | E_ENET_SPEED_100),
-+ /**< 100 Mbps SMII */
-+ E_ENET_MODE_GMII_1000 = (E_ENET_IF_GMII | E_ENET_SPEED_1000),
-+ /**< 1000 Mbps GMII */
-+ E_ENET_MODE_RGMII_10 = (E_ENET_IF_RGMII | E_ENET_SPEED_10),
-+ /**< 10 Mbps RGMII */
-+ E_ENET_MODE_RGMII_100 = (E_ENET_IF_RGMII | E_ENET_SPEED_100),
-+ /**< 100 Mbps RGMII */
-+ E_ENET_MODE_RGMII_1000 = (E_ENET_IF_RGMII | E_ENET_SPEED_1000),
-+ /**< 1000 Mbps RGMII */
-+ E_ENET_MODE_TBI_1000 = (E_ENET_IF_TBI | E_ENET_SPEED_1000),
-+ /**< 1000 Mbps TBI */
-+ E_ENET_MODE_RTBI_1000 = (E_ENET_IF_RTBI | E_ENET_SPEED_1000),
-+ /**< 1000 Mbps RTBI */
-+ E_ENET_MODE_SGMII_10 = (E_ENET_IF_SGMII | E_ENET_SPEED_10),
-+ /**< 10 Mbps SGMII with auto-negotiation between MAC and
-+ SGMII phy according to Cisco SGMII specification */
-+ E_ENET_MODE_SGMII_100 = (E_ENET_IF_SGMII | E_ENET_SPEED_100),
-+ /**< 100 Mbps SGMII with auto-negotiation between MAC and
-+ SGMII phy according to Cisco SGMII specification */
-+ E_ENET_MODE_SGMII_1000 = (E_ENET_IF_SGMII | E_ENET_SPEED_1000),
-+ /**< 1000 Mbps SGMII with auto-negotiation between MAC and
-+ SGMII phy according to Cisco SGMII specification */
-+ E_ENET_MODE_SGMII_BASEX_10 = (ENET_IF_SGMII_BASEX | E_ENET_IF_SGMII
-+ | E_ENET_SPEED_10),
-+ /**< 10 Mbps SGMII with 1000BaseX auto-negotiation between
-+ MAC and SGMII phy or backplane */
-+ E_ENET_MODE_SGMII_BASEX_100 = (ENET_IF_SGMII_BASEX | E_ENET_IF_SGMII
-+ | E_ENET_SPEED_100),
-+ /**< 100 Mbps SGMII with 1000BaseX auto-negotiation between
-+ MAC and SGMII phy or backplane */
-+ E_ENET_MODE_SGMII_BASEX_1000 = (ENET_IF_SGMII_BASEX | E_ENET_IF_SGMII
-+ | E_ENET_SPEED_1000),
-+ /**< 1000 Mbps SGMII with 1000BaseX auto-negotiation between
-+ MAC and SGMII phy or backplane */
-+ E_ENET_MODE_QSGMII_1000 = (E_ENET_IF_QSGMII | E_ENET_SPEED_1000),
-+ /**< 1000 Mbps QSGMII with auto-negotiation between MAC and
-+ QSGMII phy according to Cisco QSGMII specification */
-+ E_ENET_MODE_QSGMII_BASEX_1000 = (ENET_IF_SGMII_BASEX | E_ENET_IF_QSGMII
-+ | E_ENET_SPEED_1000),
-+ /**< 1000 Mbps QSGMII with 1000BaseX auto-negotiation between
-+ MAC and QSGMII phy or backplane */
-+ E_ENET_MODE_XGMII_10000 = (E_ENET_IF_XGMII | E_ENET_SPEED_10000),
-+ /**< 10000 Mbps XGMII */
-+ E_ENET_MODE_XFI_10000 = (E_ENET_IF_XFI | E_ENET_SPEED_10000)
-+ /**< 10000 Mbps XFI */
-+};
-+
-+enum fmam_mac_statistics_level {
-+ E_FMAN_MAC_NONE_STATISTICS, /**< No statistics */
-+ E_FMAN_MAC_PARTIAL_STATISTICS, /**< Only error counters are available;
-+ Optimized for performance */
-+ E_FMAN_MAC_FULL_STATISTICS /**< All counters available; Not
-+ optimized for performance */
-+};
-+
-+#define _MAKE_ENET_MODE(_interface, _speed) (enum enet_mode)((_interface) \
-+ | (_speed))
-+
-+#define _ENET_INTERFACE_FROM_MODE(mode) (enum enet_interface) \
-+ ((mode) & 0x0FFF0000)
-+#define _ENET_SPEED_FROM_MODE(mode) (enum enet_speed)((mode) & 0x0000FFFF)
-+#define _ENET_ADDR_TO_UINT64(_enet_addr) \
-+ (uint64_t)(((uint64_t)(_enet_addr)[0] << 40) | \
-+ ((uint64_t)(_enet_addr)[1] << 32) | \
-+ ((uint64_t)(_enet_addr)[2] << 24) | \
-+ ((uint64_t)(_enet_addr)[3] << 16) | \
-+ ((uint64_t)(_enet_addr)[4] << 8) | \
-+ ((uint64_t)(_enet_addr)[5]))
-+
-+#define _MAKE_ENET_ADDR_FROM_UINT64(_addr64, _enet_addr) \
-+ do { \
-+ int i; \
-+ for (i = 0; i < ENET_NUM_OCTETS_PER_ADDRESS; i++) \
-+ (_enet_addr)[i] = (uint8_t)((_addr64) >> ((5-i)*8));\
-+ } while (0)
-+
-+#endif /* __FSL_ENET_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/flib/fsl_fman.h
-@@ -0,0 +1,825 @@
-+/*
-+ * Copyright 2013 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#ifndef __FSL_FMAN_H
-+#define __FSL_FMAN_H
-+
-+#include "common/general.h"
-+
-+struct fman_ext_pool_params {
-+ uint8_t id; /**< External buffer pool id */
-+ uint16_t size; /**< External buffer pool buffer size */
-+};
-+
-+struct fman_ext_pools {
-+ uint8_t num_pools_used; /**< Number of pools use by this port */
-+ struct fman_ext_pool_params *ext_buf_pool;
-+ /**< Parameters for each port */
-+};
-+
-+struct fman_backup_bm_pools {
-+ uint8_t num_backup_pools; /**< Number of BM backup pools -
-+ must be smaller than the total number
-+ of pools defined for the specified
-+ port.*/
-+ uint8_t *pool_ids; /**< numOfBackupPools pool id's,
-+ specifying which pools should be used
-+ only as backup. Pool id's specified
-+ here must be a subset of the pools
-+ used by the specified port.*/
-+};
-+
-+/**************************************************************************//**
-+ @Description A structure for defining BM pool depletion criteria
-+*//***************************************************************************/
-+struct fman_buf_pool_depletion {
-+ bool buf_pool_depletion_enabled;
-+ bool pools_grp_mode_enable; /**< select mode in which pause frames
-+ will be sent after a number of pools
-+ (all together!) are depleted */
-+ uint8_t num_pools; /**< the number of depleted pools that
-+ will invoke pause frames transmission.
-+ */
-+ bool *pools_to_consider; /**< For each pool, TRUE if it should be
-+ considered for depletion (Note - this
-+ pool must be used by this port!). */
-+ bool single_pool_mode_enable; /**< select mode in which pause frames
-+ will be sent after a single-pool
-+ is depleted; */
-+ bool *pools_to_consider_for_single_mode;
-+ /**< For each pool, TRUE if it should be
-+ considered for depletion (Note - this
-+ pool must be used by this port!) */
-+ bool has_pfc_priorities;
-+ bool *pfc_priorities_en; /**< This field is used by the MAC as
-+ the Priority Enable Vector in the PFC
-+ frame which is transmitted */
-+};
-+
-+/**************************************************************************//**
-+ @Description Enum for defining port DMA swap mode
-+*//***************************************************************************/
-+enum fman_dma_swap_option {
-+ FMAN_DMA_NO_SWP, /**< No swap, transfer data as is.*/
-+ FMAN_DMA_SWP_PPC_LE, /**< The transferred data should be swapped
-+ in PowerPc Little Endian mode. */
-+ FMAN_DMA_SWP_BE /**< The transferred data should be swapped
-+ in Big Endian mode */
-+};
-+
-+/**************************************************************************//**
-+ @Description Enum for defining port DMA cache attributes
-+*//***************************************************************************/
-+enum fman_dma_cache_option {
-+ FMAN_DMA_NO_STASH = 0, /**< Cacheable, no Allocate (No Stashing) */
-+ FMAN_DMA_STASH = 1 /**< Cacheable and Allocate (Stashing on) */
-+};
-+
-+typedef struct t_FmPrsResult fm_prs_result_t;
-+typedef enum e_EnetMode enet_mode_t;
-+typedef t_Handle handle_t;
-+
-+struct fman_revision_info {
-+ uint8_t majorRev; /**< Major revision */
-+ uint8_t minorRev; /**< Minor revision */
-+};
-+
-+/* sizes */
-+#define CAPWAP_FRAG_EXTRA_SPACE 32
-+#define OFFSET_UNITS 16
-+#define MAX_INT_OFFSET 240
-+#define MAX_IC_SIZE 256
-+#define MAX_EXT_OFFSET 496
-+#define MAX_EXT_BUFFER_OFFSET 511
-+
-+/**************************************************************************
-+ @Description Memory Mapped Registers
-+***************************************************************************/
-+#define FMAN_LIODN_TBL 64 /* size of LIODN table */
-+
-+struct fman_fpm_regs {
-+ uint32_t fmfp_tnc; /**< FPM TNUM Control 0x00 */
-+ uint32_t fmfp_prc; /**< FPM Port_ID FmCtl Association 0x04 */
-+ uint32_t fmfp_brkc; /**< FPM Breakpoint Control 0x08 */
-+ uint32_t fmfp_mxd; /**< FPM Flush Control 0x0c */
-+ uint32_t fmfp_dist1; /**< FPM Dispatch Thresholds1 0x10 */
-+ uint32_t fmfp_dist2; /**< FPM Dispatch Thresholds2 0x14 */
-+ uint32_t fm_epi; /**< FM Error Pending Interrupts 0x18 */
-+ uint32_t fm_rie; /**< FM Error Interrupt Enable 0x1c */
-+ uint32_t fmfp_fcev[4]; /**< FPM FMan-Controller Event 1-4 0x20-0x2f */
-+ uint32_t res0030[4]; /**< res 0x30 - 0x3f */
-+ uint32_t fmfp_cee[4]; /**< PM FMan-Controller Event 1-4 0x40-0x4f */
-+ uint32_t res0050[4]; /**< res 0x50-0x5f */
-+ uint32_t fmfp_tsc1; /**< FPM TimeStamp Control1 0x60 */
-+ uint32_t fmfp_tsc2; /**< FPM TimeStamp Control2 0x64 */
-+ uint32_t fmfp_tsp; /**< FPM Time Stamp 0x68 */
-+ uint32_t fmfp_tsf; /**< FPM Time Stamp Fraction 0x6c */
-+ uint32_t fm_rcr; /**< FM Rams Control 0x70 */
-+ uint32_t fmfp_extc; /**< FPM External Requests Control 0x74 */
-+ uint32_t fmfp_ext1; /**< FPM External Requests Config1 0x78 */
-+ uint32_t fmfp_ext2; /**< FPM External Requests Config2 0x7c */
-+ uint32_t fmfp_drd[16]; /**< FPM Data_Ram Data 0-15 0x80 - 0xbf */
-+ uint32_t fmfp_dra; /**< FPM Data Ram Access 0xc0 */
-+ uint32_t fm_ip_rev_1; /**< FM IP Block Revision 1 0xc4 */
-+ uint32_t fm_ip_rev_2; /**< FM IP Block Revision 2 0xc8 */
-+ uint32_t fm_rstc; /**< FM Reset Command 0xcc */
-+ uint32_t fm_cld; /**< FM Classifier Debug 0xd0 */
-+ uint32_t fm_npi; /**< FM Normal Pending Interrupts 0xd4 */
-+ uint32_t fmfp_exte; /**< FPM External Requests Enable 0xd8 */
-+ uint32_t fmfp_ee; /**< FPM Event & Mask 0xdc */
-+ uint32_t fmfp_cev[4]; /**< FPM CPU Event 1-4 0xe0-0xef */
-+ uint32_t res00f0[4]; /**< res 0xf0-0xff */
-+ uint32_t fmfp_ps[64]; /**< FPM Port Status 0x100-0x1ff */
-+ uint32_t fmfp_clfabc; /**< FPM CLFABC 0x200 */
-+ uint32_t fmfp_clfcc; /**< FPM CLFCC 0x204 */
-+ uint32_t fmfp_clfaval; /**< FPM CLFAVAL 0x208 */
-+ uint32_t fmfp_clfbval; /**< FPM CLFBVAL 0x20c */
-+ uint32_t fmfp_clfcval; /**< FPM CLFCVAL 0x210 */
-+ uint32_t fmfp_clfamsk; /**< FPM CLFAMSK 0x214 */
-+ uint32_t fmfp_clfbmsk; /**< FPM CLFBMSK 0x218 */
-+ uint32_t fmfp_clfcmsk; /**< FPM CLFCMSK 0x21c */
-+ uint32_t fmfp_clfamc; /**< FPM CLFAMC 0x220 */
-+ uint32_t fmfp_clfbmc; /**< FPM CLFBMC 0x224 */
-+ uint32_t fmfp_clfcmc; /**< FPM CLFCMC 0x228 */
-+ uint32_t fmfp_decceh; /**< FPM DECCEH 0x22c */
-+ uint32_t res0230[116]; /**< res 0x230 - 0x3ff */
-+ uint32_t fmfp_ts[128]; /**< 0x400: FPM Task Status 0x400 - 0x5ff */
-+ uint32_t res0600[0x400 - 384];
-+};
-+
-+struct fman_bmi_regs {
-+ uint32_t fmbm_init; /**< BMI Initialization 0x00 */
-+ uint32_t fmbm_cfg1; /**< BMI Configuration 1 0x04 */
-+ uint32_t fmbm_cfg2; /**< BMI Configuration 2 0x08 */
-+ uint32_t res000c[5]; /**< 0x0c - 0x1f */
-+ uint32_t fmbm_ievr; /**< Interrupt Event Register 0x20 */
-+ uint32_t fmbm_ier; /**< Interrupt Enable Register 0x24 */
-+ uint32_t fmbm_ifr; /**< Interrupt Force Register 0x28 */
-+ uint32_t res002c[5]; /**< 0x2c - 0x3f */
-+ uint32_t fmbm_arb[8]; /**< BMI Arbitration 0x40 - 0x5f */
-+ uint32_t res0060[12]; /**<0x60 - 0x8f */
-+ uint32_t fmbm_dtc[3]; /**< Debug Trap Counter 0x90 - 0x9b */
-+ uint32_t res009c; /**< 0x9c */
-+ uint32_t fmbm_dcv[3][4]; /**< Debug Compare val 0xa0-0xcf */
-+ uint32_t fmbm_dcm[3][4]; /**< Debug Compare Mask 0xd0-0xff */
-+ uint32_t fmbm_gde; /**< BMI Global Debug Enable 0x100 */
-+ uint32_t fmbm_pp[63]; /**< BMI Port Parameters 0x104 - 0x1ff */
-+ uint32_t res0200; /**< 0x200 */
-+ uint32_t fmbm_pfs[63]; /**< BMI Port FIFO Size 0x204 - 0x2ff */
-+ uint32_t res0300; /**< 0x300 */
-+ uint32_t fmbm_spliodn[63]; /**< Port Partition ID 0x304 - 0x3ff */
-+};
-+
-+struct fman_qmi_regs {
-+ uint32_t fmqm_gc; /**< General Configuration Register 0x00 */
-+ uint32_t res0004; /**< 0x04 */
-+ uint32_t fmqm_eie; /**< Error Interrupt Event Register 0x08 */
-+ uint32_t fmqm_eien; /**< Error Interrupt Enable Register 0x0c */
-+ uint32_t fmqm_eif; /**< Error Interrupt Force Register 0x10 */
-+ uint32_t fmqm_ie; /**< Interrupt Event Register 0x14 */
-+ uint32_t fmqm_ien; /**< Interrupt Enable Register 0x18 */
-+ uint32_t fmqm_if; /**< Interrupt Force Register 0x1c */
-+ uint32_t fmqm_gs; /**< Global Status Register 0x20 */
-+ uint32_t fmqm_ts; /**< Task Status Register 0x24 */
-+ uint32_t fmqm_etfc; /**< Enqueue Total Frame Counter 0x28 */
-+ uint32_t fmqm_dtfc; /**< Dequeue Total Frame Counter 0x2c */
-+ uint32_t fmqm_dc0; /**< Dequeue Counter 0 0x30 */
-+ uint32_t fmqm_dc1; /**< Dequeue Counter 1 0x34 */
-+ uint32_t fmqm_dc2; /**< Dequeue Counter 2 0x38 */
-+ uint32_t fmqm_dc3; /**< Dequeue Counter 3 0x3c */
-+ uint32_t fmqm_dfdc; /**< Dequeue FQID from Default Counter 0x40 */
-+ uint32_t fmqm_dfcc; /**< Dequeue FQID from Context Counter 0x44 */
-+ uint32_t fmqm_dffc; /**< Dequeue FQID from FD Counter 0x48 */
-+ uint32_t fmqm_dcc; /**< Dequeue Confirm Counter 0x4c */
-+ uint32_t res0050[7]; /**< 0x50 - 0x6b */
-+ uint32_t fmqm_tapc; /**< Tnum Aging Period Control 0x6c */
-+ uint32_t fmqm_dmcvc; /**< Dequeue MAC Command Valid Counter 0x70 */
-+ uint32_t fmqm_difdcc; /**< Dequeue Invalid FD Command Counter 0x74 */
-+ uint32_t fmqm_da1v; /**< Dequeue A1 Valid Counter 0x78 */
-+ uint32_t res007c; /**< 0x7c */
-+ uint32_t fmqm_dtc; /**< 0x80 Debug Trap Counter 0x80 */
-+ uint32_t fmqm_efddd; /**< 0x84 Enqueue Frame desc Dynamic dbg 0x84 */
-+ uint32_t res0088[2]; /**< 0x88 - 0x8f */
-+ struct {
-+ uint32_t fmqm_dtcfg1; /**< 0x90 dbg trap cfg 1 Register 0x00 */
-+ uint32_t fmqm_dtval1; /**< Debug Trap Value 1 Register 0x04 */
-+ uint32_t fmqm_dtm1; /**< Debug Trap Mask 1 Register 0x08 */
-+ uint32_t fmqm_dtc1; /**< Debug Trap Counter 1 Register 0x0c */
-+ uint32_t fmqm_dtcfg2; /**< dbg Trap cfg 2 Register 0x10 */
-+ uint32_t fmqm_dtval2; /**< Debug Trap Value 2 Register 0x14 */
-+ uint32_t fmqm_dtm2; /**< Debug Trap Mask 2 Register 0x18 */
-+ uint32_t res001c; /**< 0x1c */
-+ } dbg_traps[3]; /**< 0x90 - 0xef */
-+ uint8_t res00f0[0x400 - 0xf0]; /**< 0xf0 - 0x3ff */
-+};
-+
-+struct fman_dma_regs {
-+ uint32_t fmdmsr; /**< FM DMA status register 0x00 */
-+ uint32_t fmdmmr; /**< FM DMA mode register 0x04 */
-+ uint32_t fmdmtr; /**< FM DMA bus threshold register 0x08 */
-+ uint32_t fmdmhy; /**< FM DMA bus hysteresis register 0x0c */
-+ uint32_t fmdmsetr; /**< FM DMA SOS emergency Threshold Register 0x10 */
-+ uint32_t fmdmtah; /**< FM DMA transfer bus address high reg 0x14 */
-+ uint32_t fmdmtal; /**< FM DMA transfer bus address low reg 0x18 */
-+ uint32_t fmdmtcid; /**< FM DMA transfer bus communication ID reg 0x1c */
-+ uint32_t fmdmra; /**< FM DMA bus internal ram address register 0x20 */
-+ uint32_t fmdmrd; /**< FM DMA bus internal ram data register 0x24 */
-+ uint32_t fmdmwcr; /**< FM DMA CAM watchdog counter value 0x28 */
-+ uint32_t fmdmebcr; /**< FM DMA CAM base in MURAM register 0x2c */
-+ uint32_t fmdmccqdr; /**< FM DMA CAM and CMD Queue Debug reg 0x30 */
-+ uint32_t fmdmccqvr1; /**< FM DMA CAM and CMD Queue Value reg #1 0x34 */
-+ uint32_t fmdmccqvr2; /**< FM DMA CAM and CMD Queue Value reg #2 0x38 */
-+ uint32_t fmdmcqvr3; /**< FM DMA CMD Queue Value register #3 0x3c */
-+ uint32_t fmdmcqvr4; /**< FM DMA CMD Queue Value register #4 0x40 */
-+ uint32_t fmdmcqvr5; /**< FM DMA CMD Queue Value register #5 0x44 */
-+ uint32_t fmdmsefrc; /**< FM DMA Semaphore Entry Full Reject Cntr 0x48 */
-+ uint32_t fmdmsqfrc; /**< FM DMA Semaphore Queue Full Reject Cntr 0x4c */
-+ uint32_t fmdmssrc; /**< FM DMA Semaphore SYNC Reject Counter 0x50 */
-+ uint32_t fmdmdcr; /**< FM DMA Debug Counter 0x54 */
-+ uint32_t fmdmemsr; /**< FM DMA Emergency Smoother Register 0x58 */
-+ uint32_t res005c; /**< 0x5c */
-+ uint32_t fmdmplr[FMAN_LIODN_TBL / 2]; /**< DMA LIODN regs 0x60-0xdf */
-+ uint32_t res00e0[0x400 - 56];
-+};
-+
-+struct fman_rg {
-+ struct fman_fpm_regs *fpm_rg;
-+ struct fman_dma_regs *dma_rg;
-+ struct fman_bmi_regs *bmi_rg;
-+ struct fman_qmi_regs *qmi_rg;
-+};
-+
-+enum fman_dma_cache_override {
-+ E_FMAN_DMA_NO_CACHE_OR = 0, /**< No override of the Cache field */
-+ E_FMAN_DMA_NO_STASH_DATA, /**< No data stashing in system level cache */
-+ E_FMAN_DMA_MAY_STASH_DATA, /**< Stashing allowed in sys level cache */
-+ E_FMAN_DMA_STASH_DATA /**< Stashing performed in system level cache */
-+};
-+
-+enum fman_dma_aid_mode {
-+ E_FMAN_DMA_AID_OUT_PORT_ID = 0, /**< 4 LSB of PORT_ID */
-+ E_FMAN_DMA_AID_OUT_TNUM /**< 4 LSB of TNUM */
-+};
-+
-+enum fman_dma_dbg_cnt_mode {
-+ E_FMAN_DMA_DBG_NO_CNT = 0, /**< No counting */
-+ E_FMAN_DMA_DBG_CNT_DONE, /**< Count DONE commands */
-+ E_FMAN_DMA_DBG_CNT_COMM_Q_EM, /**< command Q emergency signal */
-+ E_FMAN_DMA_DBG_CNT_INT_READ_EM, /**< Read buf emergency signal */
-+ E_FMAN_DMA_DBG_CNT_INT_WRITE_EM, /**< Write buf emergency signal */
-+ E_FMAN_DMA_DBG_CNT_FPM_WAIT, /**< FPM WAIT signal */
-+ E_FMAN_DMA_DBG_CNT_SIGLE_BIT_ECC, /**< Single bit ECC errors */
-+ E_FMAN_DMA_DBG_CNT_RAW_WAR_PROT /**< RAW & WAR protection counter */
-+};
-+
-+enum fman_dma_emergency_level {
-+ E_FMAN_DMA_EM_EBS = 0, /**< EBS emergency */
-+ E_FMAN_DMA_EM_SOS /**< SOS emergency */
-+};
-+
-+enum fman_catastrophic_err {
-+ E_FMAN_CATAST_ERR_STALL_PORT = 0, /**< Port_ID stalled reset required */
-+ E_FMAN_CATAST_ERR_STALL_TASK /**< Only erroneous task is stalled */
-+};
-+
-+enum fman_dma_err {
-+ E_FMAN_DMA_ERR_CATASTROPHIC = 0, /**< Catastrophic DMA error */
-+ E_FMAN_DMA_ERR_REPORT /**< Reported DMA error */
-+};
-+
-+struct fman_cfg {
-+ uint16_t liodn_bs_pr_port[FMAN_LIODN_TBL];/* base per port */
-+ bool en_counters;
-+ uint8_t disp_limit_tsh;
-+ uint8_t prs_disp_tsh;
-+ uint8_t plcr_disp_tsh;
-+ uint8_t kg_disp_tsh;
-+ uint8_t bmi_disp_tsh;
-+ uint8_t qmi_enq_disp_tsh;
-+ uint8_t qmi_deq_disp_tsh;
-+ uint8_t fm_ctl1_disp_tsh;
-+ uint8_t fm_ctl2_disp_tsh;
-+ enum fman_dma_cache_override dma_cache_override;
-+ enum fman_dma_aid_mode dma_aid_mode;
-+ bool dma_aid_override;
-+ uint8_t dma_axi_dbg_num_of_beats;
-+ uint8_t dma_cam_num_of_entries;
-+ uint32_t dma_watchdog;
-+ uint8_t dma_comm_qtsh_asrt_emer;
-+ uint8_t dma_write_buf_tsh_asrt_emer;
-+ uint8_t dma_read_buf_tsh_asrt_emer;
-+ uint8_t dma_comm_qtsh_clr_emer;
-+ uint8_t dma_write_buf_tsh_clr_emer;
-+ uint8_t dma_read_buf_tsh_clr_emer;
-+ uint32_t dma_sos_emergency;
-+ enum fman_dma_dbg_cnt_mode dma_dbg_cnt_mode;
-+ bool dma_stop_on_bus_error;
-+ bool dma_en_emergency;
-+ uint32_t dma_emergency_bus_select;
-+ enum fman_dma_emergency_level dma_emergency_level;
-+ bool dma_en_emergency_smoother;
-+ uint32_t dma_emergency_switch_counter;
-+ bool halt_on_external_activ;
-+ bool halt_on_unrecov_ecc_err;
-+ enum fman_catastrophic_err catastrophic_err;
-+ enum fman_dma_err dma_err;
-+ bool en_muram_test_mode;
-+ bool en_iram_test_mode;
-+ bool external_ecc_rams_enable;
-+ uint16_t tnum_aging_period;
-+ uint32_t exceptions;
-+ uint16_t clk_freq;
-+ bool pedantic_dma;
-+ uint32_t cam_base_addr;
-+ uint32_t fifo_base_addr;
-+ uint32_t total_fifo_size;
-+ uint8_t total_num_of_tasks;
-+ bool qmi_deq_option_support;
-+ uint32_t qmi_def_tnums_thresh;
-+ bool fman_partition_array;
-+ uint8_t num_of_fman_ctrl_evnt_regs;
-+};
-+
-+/**************************************************************************//**
-+ @Description Exceptions
-+*//***************************************************************************/
-+#define FMAN_EX_DMA_BUS_ERROR 0x80000000
-+#define FMAN_EX_DMA_READ_ECC 0x40000000
-+#define FMAN_EX_DMA_SYSTEM_WRITE_ECC 0x20000000
-+#define FMAN_EX_DMA_FM_WRITE_ECC 0x10000000
-+#define FMAN_EX_FPM_STALL_ON_TASKS 0x08000000
-+#define FMAN_EX_FPM_SINGLE_ECC 0x04000000
-+#define FMAN_EX_FPM_DOUBLE_ECC 0x02000000
-+#define FMAN_EX_QMI_SINGLE_ECC 0x01000000
-+#define FMAN_EX_QMI_DEQ_FROM_UNKNOWN_PORTID 0x00800000
-+#define FMAN_EX_QMI_DOUBLE_ECC 0x00400000
-+#define FMAN_EX_BMI_LIST_RAM_ECC 0x00200000
-+#define FMAN_EX_BMI_PIPELINE_ECC 0x00100000
-+#define FMAN_EX_BMI_STATISTICS_RAM_ECC 0x00080000
-+#define FMAN_EX_IRAM_ECC 0x00040000
-+#define FMAN_EX_NURAM_ECC 0x00020000
-+#define FMAN_EX_BMI_DISPATCH_RAM_ECC 0x00010000
-+
-+enum fman_exceptions {
-+ E_FMAN_EX_DMA_BUS_ERROR = 0, /**< DMA bus error. */
-+ E_FMAN_EX_DMA_READ_ECC, /**< Read Buffer ECC error */
-+ E_FMAN_EX_DMA_SYSTEM_WRITE_ECC, /**< Write Buffer ECC err on sys side */
-+ E_FMAN_EX_DMA_FM_WRITE_ECC, /**< Write Buffer ECC error on FM side */
-+ E_FMAN_EX_FPM_STALL_ON_TASKS, /**< Stall of tasks on FPM */
-+ E_FMAN_EX_FPM_SINGLE_ECC, /**< Single ECC on FPM. */
-+ E_FMAN_EX_FPM_DOUBLE_ECC, /**< Double ECC error on FPM ram access */
-+ E_FMAN_EX_QMI_SINGLE_ECC, /**< Single ECC on QMI. */
-+ E_FMAN_EX_QMI_DOUBLE_ECC, /**< Double bit ECC occurred on QMI */
-+ E_FMAN_EX_QMI_DEQ_FROM_UNKNOWN_PORTID,/**< DeQ from unknown port id */
-+ E_FMAN_EX_BMI_LIST_RAM_ECC, /**< Linked List RAM ECC error */
-+ E_FMAN_EX_BMI_STORAGE_PROFILE_ECC, /**< storage profile */
-+ E_FMAN_EX_BMI_STATISTICS_RAM_ECC, /**< Statistics RAM ECC Err Enable */
-+ E_FMAN_EX_BMI_DISPATCH_RAM_ECC, /**< Dispatch RAM ECC Error Enable */
-+ E_FMAN_EX_IRAM_ECC, /**< Double bit ECC occurred on IRAM*/
-+ E_FMAN_EX_MURAM_ECC /**< Double bit ECC occurred on MURAM*/
-+};
-+
-+enum fman_counters {
-+ E_FMAN_COUNTERS_ENQ_TOTAL_FRAME = 0, /**< QMI tot enQ frames counter */
-+ E_FMAN_COUNTERS_DEQ_TOTAL_FRAME, /**< QMI tot deQ frames counter */
-+ E_FMAN_COUNTERS_DEQ_0, /**< QMI 0 frames from QMan counter */
-+ E_FMAN_COUNTERS_DEQ_1, /**< QMI 1 frames from QMan counter */
-+ E_FMAN_COUNTERS_DEQ_2, /**< QMI 2 frames from QMan counter */
-+ E_FMAN_COUNTERS_DEQ_3, /**< QMI 3 frames from QMan counter */
-+ E_FMAN_COUNTERS_DEQ_FROM_DEFAULT, /**< QMI deQ from dflt queue cntr */
-+ E_FMAN_COUNTERS_DEQ_FROM_CONTEXT, /**< QMI deQ from FQ context cntr */
-+ E_FMAN_COUNTERS_DEQ_FROM_FD, /**< QMI deQ from FD command field cntr */
-+ E_FMAN_COUNTERS_DEQ_CONFIRM, /**< QMI dequeue confirm counter */
-+ E_FMAN_COUNTERS_SEMAPHOR_ENTRY_FULL_REJECT, /**< DMA full entry cntr */
-+ E_FMAN_COUNTERS_SEMAPHOR_QUEUE_FULL_REJECT, /**< DMA full CAM Q cntr */
-+ E_FMAN_COUNTERS_SEMAPHOR_SYNC_REJECT /**< DMA sync counter */
-+};
-+
-+#define FPM_PRT_FM_CTL1 0x00000001
-+#define FPM_PRT_FM_CTL2 0x00000002
-+
-+/**************************************************************************//**
-+ @Description DMA definitions
-+*//***************************************************************************/
-+
-+/* masks */
-+#define DMA_MODE_AID_OR 0x20000000
-+#define DMA_MODE_SBER 0x10000000
-+#define DMA_MODE_BER 0x00200000
-+#define DMA_MODE_EB 0x00100000
-+#define DMA_MODE_ECC 0x00000020
-+#define DMA_MODE_PRIVILEGE_PROT 0x00001000
-+#define DMA_MODE_SECURE_PROT 0x00000800
-+#define DMA_MODE_EMER_READ 0x00080000
-+#define DMA_MODE_EMER_WRITE 0x00040000
-+#define DMA_MODE_CACHE_OR_MASK 0xC0000000
-+#define DMA_MODE_CEN_MASK 0x0000E000
-+#define DMA_MODE_DBG_MASK 0x00000380
-+#define DMA_MODE_AXI_DBG_MASK 0x0F000000
-+
-+#define DMA_EMSR_EMSTR_MASK 0x0000FFFF
-+
-+#define DMA_TRANSFER_PORTID_MASK 0xFF000000
-+#define DMA_TRANSFER_TNUM_MASK 0x00FF0000
-+#define DMA_TRANSFER_LIODN_MASK 0x00000FFF
-+
-+#define DMA_HIGH_LIODN_MASK 0x0FFF0000
-+#define DMA_LOW_LIODN_MASK 0x00000FFF
-+
-+#define DMA_STATUS_CMD_QUEUE_NOT_EMPTY 0x10000000
-+#define DMA_STATUS_BUS_ERR 0x08000000
-+#define DMA_STATUS_READ_ECC 0x04000000
-+#define DMA_STATUS_SYSTEM_WRITE_ECC 0x02000000
-+#define DMA_STATUS_FM_WRITE_ECC 0x01000000
-+#define DMA_STATUS_SYSTEM_DPEXT_ECC 0x00800000
-+#define DMA_STATUS_FM_DPEXT_ECC 0x00400000
-+#define DMA_STATUS_SYSTEM_DPDAT_ECC 0x00200000
-+#define DMA_STATUS_FM_DPDAT_ECC 0x00100000
-+#define DMA_STATUS_FM_SPDAT_ECC 0x00080000
-+
-+#define FM_LIODN_BASE_MASK 0x00000FFF
-+
-+/* shifts */
-+#define DMA_MODE_CACHE_OR_SHIFT 30
-+#define DMA_MODE_BUS_PRI_SHIFT 16
-+#define DMA_MODE_AXI_DBG_SHIFT 24
-+#define DMA_MODE_CEN_SHIFT 13
-+#define DMA_MODE_BUS_PROT_SHIFT 10
-+#define DMA_MODE_DBG_SHIFT 7
-+#define DMA_MODE_EMER_LVL_SHIFT 6
-+#define DMA_MODE_AID_MODE_SHIFT 4
-+#define DMA_MODE_MAX_AXI_DBG_NUM_OF_BEATS 16
-+#define DMA_MODE_MAX_CAM_NUM_OF_ENTRIES 32
-+
-+#define DMA_THRESH_COMMQ_SHIFT 24
-+#define DMA_THRESH_READ_INT_BUF_SHIFT 16
-+
-+#define DMA_LIODN_SHIFT 16
-+
-+#define DMA_TRANSFER_PORTID_SHIFT 24
-+#define DMA_TRANSFER_TNUM_SHIFT 16
-+
-+/* sizes */
-+#define DMA_MAX_WATCHDOG 0xffffffff
-+
-+/* others */
-+#define DMA_CAM_SIZEOF_ENTRY 0x40
-+#define DMA_CAM_ALIGN 0x1000
-+#define DMA_CAM_UNITS 8
-+
-+/**************************************************************************//**
-+ @Description General defines
-+*//***************************************************************************/
-+
-+#define FM_DEBUG_STATUS_REGISTER_OFFSET 0x000d1084UL
-+#define FM_UCODE_DEBUG_INSTRUCTION 0x6ffff805UL
-+
-+/**************************************************************************//**
-+ @Description FPM defines
-+*//***************************************************************************/
-+
-+/* masks */
-+#define FPM_EV_MASK_DOUBLE_ECC 0x80000000
-+#define FPM_EV_MASK_STALL 0x40000000
-+#define FPM_EV_MASK_SINGLE_ECC 0x20000000
-+#define FPM_EV_MASK_RELEASE_FM 0x00010000
-+#define FPM_EV_MASK_DOUBLE_ECC_EN 0x00008000
-+#define FPM_EV_MASK_STALL_EN 0x00004000
-+#define FPM_EV_MASK_SINGLE_ECC_EN 0x00002000
-+#define FPM_EV_MASK_EXTERNAL_HALT 0x00000008
-+#define FPM_EV_MASK_ECC_ERR_HALT 0x00000004
-+
-+#define FPM_RAM_RAMS_ECC_EN 0x80000000
-+#define FPM_RAM_IRAM_ECC_EN 0x40000000
-+#define FPM_RAM_MURAM_ECC 0x00008000
-+#define FPM_RAM_IRAM_ECC 0x00004000
-+#define FPM_RAM_MURAM_TEST_ECC 0x20000000
-+#define FPM_RAM_IRAM_TEST_ECC 0x10000000
-+#define FPM_RAM_RAMS_ECC_EN_SRC_SEL 0x08000000
-+
-+#define FPM_IRAM_ECC_ERR_EX_EN 0x00020000
-+#define FPM_MURAM_ECC_ERR_EX_EN 0x00040000
-+
-+#define FPM_REV1_MAJOR_MASK 0x0000FF00
-+#define FPM_REV1_MINOR_MASK 0x000000FF
-+
-+#define FPM_REV2_INTEG_MASK 0x00FF0000
-+#define FPM_REV2_ERR_MASK 0x0000FF00
-+#define FPM_REV2_CFG_MASK 0x000000FF
-+
-+#define FPM_TS_FRACTION_MASK 0x0000FFFF
-+#define FPM_TS_CTL_EN 0x80000000
-+
-+#define FPM_PRC_REALSE_STALLED 0x00800000
-+
-+#define FPM_PS_STALLED 0x00800000
-+#define FPM_PS_FM_CTL1_SEL 0x80000000
-+#define FPM_PS_FM_CTL2_SEL 0x40000000
-+#define FPM_PS_FM_CTL_SEL_MASK (FPM_PS_FM_CTL1_SEL | FPM_PS_FM_CTL2_SEL)
-+
-+#define FPM_RSTC_FM_RESET 0x80000000
-+#define FPM_RSTC_10G0_RESET 0x04000000
-+#define FPM_RSTC_1G0_RESET 0x40000000
-+#define FPM_RSTC_1G1_RESET 0x20000000
-+#define FPM_RSTC_1G2_RESET 0x10000000
-+#define FPM_RSTC_1G3_RESET 0x08000000
-+#define FPM_RSTC_1G4_RESET 0x02000000
-+
-+
-+#define FPM_DISP_LIMIT_MASK 0x1F000000
-+#define FPM_THR1_PRS_MASK 0xFF000000
-+#define FPM_THR1_KG_MASK 0x00FF0000
-+#define FPM_THR1_PLCR_MASK 0x0000FF00
-+#define FPM_THR1_BMI_MASK 0x000000FF
-+
-+#define FPM_THR2_QMI_ENQ_MASK 0xFF000000
-+#define FPM_THR2_QMI_DEQ_MASK 0x000000FF
-+#define FPM_THR2_FM_CTL1_MASK 0x00FF0000
-+#define FPM_THR2_FM_CTL2_MASK 0x0000FF00
-+
-+/* shifts */
-+#define FPM_DISP_LIMIT_SHIFT 24
-+
-+#define FPM_THR1_PRS_SHIFT 24
-+#define FPM_THR1_KG_SHIFT 16
-+#define FPM_THR1_PLCR_SHIFT 8
-+#define FPM_THR1_BMI_SHIFT 0
-+
-+#define FPM_THR2_QMI_ENQ_SHIFT 24
-+#define FPM_THR2_QMI_DEQ_SHIFT 0
-+#define FPM_THR2_FM_CTL1_SHIFT 16
-+#define FPM_THR2_FM_CTL2_SHIFT 8
-+
-+#define FPM_EV_MASK_CAT_ERR_SHIFT 1
-+#define FPM_EV_MASK_DMA_ERR_SHIFT 0
-+
-+#define FPM_REV1_MAJOR_SHIFT 8
-+#define FPM_REV1_MINOR_SHIFT 0
-+
-+#define FPM_REV2_INTEG_SHIFT 16
-+#define FPM_REV2_ERR_SHIFT 8
-+#define FPM_REV2_CFG_SHIFT 0
-+
-+#define FPM_TS_INT_SHIFT 16
-+
-+#define FPM_PORT_FM_CTL_PORTID_SHIFT 24
-+
-+#define FPM_PS_FM_CTL_SEL_SHIFT 30
-+#define FPM_PRC_ORA_FM_CTL_SEL_SHIFT 16
-+
-+#define FPM_DISP_LIMIT_SHIFT 24
-+
-+/* Interrupts defines */
-+#define FPM_EVENT_FM_CTL_0 0x00008000
-+#define FPM_EVENT_FM_CTL 0x0000FF00
-+#define FPM_EVENT_FM_CTL_BRK 0x00000080
-+
-+/* others */
-+#define FPM_MAX_DISP_LIMIT 31
-+#define FPM_RSTC_FM_RESET 0x80000000
-+#define FPM_RSTC_1G0_RESET 0x40000000
-+#define FPM_RSTC_1G1_RESET 0x20000000
-+#define FPM_RSTC_1G2_RESET 0x10000000
-+#define FPM_RSTC_1G3_RESET 0x08000000
-+#define FPM_RSTC_10G0_RESET 0x04000000
-+#define FPM_RSTC_1G4_RESET 0x02000000
-+#define FPM_RSTC_1G5_RESET 0x01000000
-+#define FPM_RSTC_1G6_RESET 0x00800000
-+#define FPM_RSTC_1G7_RESET 0x00400000
-+#define FPM_RSTC_10G1_RESET 0x00200000
-+/**************************************************************************//**
-+ @Description BMI defines
-+*//***************************************************************************/
-+/* masks */
-+#define BMI_INIT_START 0x80000000
-+#define BMI_ERR_INTR_EN_STORAGE_PROFILE_ECC 0x80000000
-+#define BMI_ERR_INTR_EN_LIST_RAM_ECC 0x40000000
-+#define BMI_ERR_INTR_EN_STATISTICS_RAM_ECC 0x20000000
-+#define BMI_ERR_INTR_EN_DISPATCH_RAM_ECC 0x10000000
-+#define BMI_NUM_OF_TASKS_MASK 0x3F000000
-+#define BMI_NUM_OF_EXTRA_TASKS_MASK 0x000F0000
-+#define BMI_NUM_OF_DMAS_MASK 0x00000F00
-+#define BMI_NUM_OF_EXTRA_DMAS_MASK 0x0000000F
-+#define BMI_FIFO_SIZE_MASK 0x000003FF
-+#define BMI_EXTRA_FIFO_SIZE_MASK 0x03FF0000
-+#define BMI_CFG2_DMAS_MASK 0x0000003F
-+#define BMI_TOTAL_FIFO_SIZE_MASK 0x07FF0000
-+#define BMI_TOTAL_NUM_OF_TASKS_MASK 0x007F0000
-+
-+/* shifts */
-+#define BMI_CFG2_TASKS_SHIFT 16
-+#define BMI_CFG2_DMAS_SHIFT 0
-+#define BMI_CFG1_FIFO_SIZE_SHIFT 16
-+#define BMI_FIFO_SIZE_SHIFT 0
-+#define BMI_EXTRA_FIFO_SIZE_SHIFT 16
-+#define BMI_NUM_OF_TASKS_SHIFT 24
-+#define BMI_EXTRA_NUM_OF_TASKS_SHIFT 16
-+#define BMI_NUM_OF_DMAS_SHIFT 8
-+#define BMI_EXTRA_NUM_OF_DMAS_SHIFT 0
-+
-+/* others */
-+#define BMI_FIFO_ALIGN 0x100
-+#define FMAN_BMI_FIFO_UNITS 0x100
-+
-+
-+/**************************************************************************//**
-+ @Description QMI defines
-+*//***************************************************************************/
-+/* masks */
-+#define QMI_CFG_ENQ_EN 0x80000000
-+#define QMI_CFG_DEQ_EN 0x40000000
-+#define QMI_CFG_EN_COUNTERS 0x10000000
-+#define QMI_CFG_SOFT_RESET 0x01000000
-+#define QMI_CFG_DEQ_MASK 0x0000003F
-+#define QMI_CFG_ENQ_MASK 0x00003F00
-+
-+#define QMI_ERR_INTR_EN_DOUBLE_ECC 0x80000000
-+#define QMI_ERR_INTR_EN_DEQ_FROM_DEF 0x40000000
-+#define QMI_INTR_EN_SINGLE_ECC 0x80000000
-+
-+/* shifts */
-+#define QMI_CFG_ENQ_SHIFT 8
-+#define QMI_TAPC_TAP 22
-+
-+#define QMI_GS_HALT_NOT_BUSY 0x00000002
-+
-+/**************************************************************************//**
-+ @Description IRAM defines
-+*//***************************************************************************/
-+/* masks */
-+#define IRAM_IADD_AIE 0x80000000
-+#define IRAM_READY 0x80000000
-+
-+uint32_t fman_get_bmi_err_event(struct fman_bmi_regs *bmi_rg);
-+uint32_t fman_get_qmi_err_event(struct fman_qmi_regs *qmi_rg);
-+uint32_t fman_get_dma_com_id(struct fman_dma_regs *dma_rg);
-+uint64_t fman_get_dma_addr(struct fman_dma_regs *dma_rg);
-+uint32_t fman_get_dma_err_event(struct fman_dma_regs *dma_rg);
-+uint32_t fman_get_fpm_err_event(struct fman_fpm_regs *fpm_rg);
-+uint32_t fman_get_muram_err_event(struct fman_fpm_regs *fpm_rg);
-+uint32_t fman_get_iram_err_event(struct fman_fpm_regs *fpm_rg);
-+uint32_t fman_get_qmi_event(struct fman_qmi_regs *qmi_rg);
-+uint32_t fman_get_fpm_error_interrupts(struct fman_fpm_regs *fpm_rg);
-+uint32_t fman_get_ctrl_intr(struct fman_fpm_regs *fpm_rg,
-+ uint8_t event_reg_id);
-+uint8_t fman_get_qmi_deq_th(struct fman_qmi_regs *qmi_rg);
-+uint8_t fman_get_qmi_enq_th(struct fman_qmi_regs *qmi_rg);
-+uint16_t fman_get_size_of_fifo(struct fman_bmi_regs *bmi_rg, uint8_t port_id);
-+uint32_t fman_get_total_fifo_size(struct fman_bmi_regs *bmi_rg);
-+uint16_t fman_get_size_of_extra_fifo(struct fman_bmi_regs *bmi_rg,
-+ uint8_t port_id);
-+uint8_t fman_get_num_of_tasks(struct fman_bmi_regs *bmi_rg, uint8_t port_id);
-+uint8_t fman_get_num_extra_tasks(struct fman_bmi_regs *bmi_rg,
-+ uint8_t port_id);
-+uint8_t fman_get_num_of_dmas(struct fman_bmi_regs *bmi_rg, uint8_t port_id);
-+uint8_t fman_get_num_extra_dmas(struct fman_bmi_regs *bmi_rg,
-+ uint8_t port_id);
-+uint32_t fman_get_normal_pending(struct fman_fpm_regs *fpm_rg);
-+uint32_t fman_get_controller_event(struct fman_fpm_regs *fpm_rg,
-+ uint8_t reg_id);
-+uint32_t fman_get_error_pending(struct fman_fpm_regs *fpm_rg);
-+void fman_get_revision(struct fman_fpm_regs *fpm_rg, uint8_t *major,
-+ uint8_t *minor);
-+uint32_t fman_get_counter(struct fman_rg *fman_rg,
-+ enum fman_counters reg_name);
-+uint32_t fman_get_dma_status(struct fman_dma_regs *dma_rg);
-+
-+
-+int fman_set_erratum_10gmac_a004_wa(struct fman_fpm_regs *fpm_rg);
-+void fman_set_ctrl_intr(struct fman_fpm_regs *fpm_rg, uint8_t event_reg_id,
-+ uint32_t enable_events);
-+void fman_set_num_of_riscs_per_port(struct fman_fpm_regs *fpm_rg,
-+ uint8_t port_id,
-+ uint8_t num_fman_ctrls,
-+ uint32_t or_fman_ctrl);
-+void fman_set_order_restoration_per_port(struct fman_fpm_regs *fpm_rg,
-+ uint8_t port_id,
-+ bool independent_mode,
-+ bool is_rx_port);
-+void fman_set_qmi_enq_th(struct fman_qmi_regs *qmi_rg, uint8_t val);
-+void fman_set_qmi_deq_th(struct fman_qmi_regs *qmi_rg, uint8_t val);
-+void fman_set_liodn_per_port(struct fman_rg *fman_rg,
-+ uint8_t port_id,
-+ uint16_t liodn_base,
-+ uint16_t liodn_offset);
-+void fman_set_size_of_fifo(struct fman_bmi_regs *bmi_rg,
-+ uint8_t port_id,
-+ uint32_t size_of_fifo,
-+ uint32_t extra_size_of_fifo);
-+void fman_set_num_of_tasks(struct fman_bmi_regs *bmi_rg,
-+ uint8_t port_id,
-+ uint8_t num_of_tasks,
-+ uint8_t num_of_extra_tasks);
-+void fman_set_num_of_open_dmas(struct fman_bmi_regs *bmi_rg,
-+ uint8_t port_id,
-+ uint8_t num_of_open_dmas,
-+ uint8_t num_of_extra_open_dmas,
-+ uint8_t total_num_of_dmas);
-+void fman_set_ports_bandwidth(struct fman_bmi_regs *bmi_rg, uint8_t *weights);
-+int fman_set_exception(struct fman_rg *fman_rg,
-+ enum fman_exceptions exception,
-+ bool enable);
-+void fman_set_dma_emergency(struct fman_dma_regs *dma_rg, bool is_write,
-+ bool enable);
-+void fman_set_dma_ext_bus_pri(struct fman_dma_regs *dma_rg, uint32_t pri);
-+void fman_set_congestion_group_pfc_priority(uint32_t *cpg_rg,
-+ uint32_t congestion_group_id,
-+ uint8_t piority_bit_map,
-+ uint32_t reg_num);
-+
-+
-+void fman_defconfig(struct fman_cfg *cfg, bool is_master);
-+void fman_regconfig(struct fman_rg *fman_rg, struct fman_cfg *cfg);
-+int fman_fpm_init(struct fman_fpm_regs *fpm_rg, struct fman_cfg *cfg);
-+int fman_bmi_init(struct fman_bmi_regs *bmi_rg, struct fman_cfg *cfg);
-+int fman_qmi_init(struct fman_qmi_regs *qmi_rg, struct fman_cfg *cfg);
-+int fman_dma_init(struct fman_dma_regs *dma_rg, struct fman_cfg *cfg);
-+void fman_free_resources(struct fman_rg *fman_rg);
-+int fman_enable(struct fman_rg *fman_rg, struct fman_cfg *cfg);
-+void fman_reset(struct fman_fpm_regs *fpm_rg);
-+void fman_resume(struct fman_fpm_regs *fpm_rg);
-+
-+
-+void fman_enable_time_stamp(struct fman_fpm_regs *fpm_rg,
-+ uint8_t count1ubit,
-+ uint16_t fm_clk_freq);
-+void fman_enable_rams_ecc(struct fman_fpm_regs *fpm_rg);
-+void fman_qmi_disable_dispatch_limit(struct fman_fpm_regs *fpm_rg);
-+void fman_disable_rams_ecc(struct fman_fpm_regs *fpm_rg);
-+void fman_resume_stalled_port(struct fman_fpm_regs *fpm_rg, uint8_t port_id);
-+int fman_reset_mac(struct fman_fpm_regs *fpm_rg, uint8_t macId, bool is_10g);
-+bool fman_is_port_stalled(struct fman_fpm_regs *fpm_rg, uint8_t port_id);
-+bool fman_rams_ecc_is_external_ctl(struct fman_fpm_regs *fpm_rg);
-+bool fman_is_qmi_halt_not_busy_state(struct fman_qmi_regs *qmi_rg);
-+int fman_modify_counter(struct fman_rg *fman_rg,
-+ enum fman_counters reg_name,
-+ uint32_t val);
-+void fman_force_intr(struct fman_rg *fman_rg,
-+ enum fman_exceptions exception);
-+void fman_set_vsp_window(struct fman_bmi_regs *bmi_rg,
-+ uint8_t port_id,
-+ uint8_t base_storage_profile,
-+ uint8_t log2_num_of_profiles);
-+
-+/**************************************************************************//**
-+ @Description default values
-+*//***************************************************************************/
-+#define DEFAULT_CATASTROPHIC_ERR E_FMAN_CATAST_ERR_STALL_PORT
-+#define DEFAULT_DMA_ERR E_FMAN_DMA_ERR_CATASTROPHIC
-+#define DEFAULT_HALT_ON_EXTERNAL_ACTIVATION FALSE /* do not change! if changed, must be disabled for rev1 ! */
-+#define DEFAULT_HALT_ON_UNRECOVERABLE_ECC_ERROR FALSE /* do not change! if changed, must be disabled for rev1 ! */
-+#define DEFAULT_EXTERNAL_ECC_RAMS_ENABLE FALSE
-+#define DEFAULT_AID_OVERRIDE FALSE
-+#define DEFAULT_AID_MODE E_FMAN_DMA_AID_OUT_TNUM
-+#define DEFAULT_DMA_COMM_Q_LOW 0x2A
-+#define DEFAULT_DMA_COMM_Q_HIGH 0x3F
-+#define DEFAULT_CACHE_OVERRIDE E_FMAN_DMA_NO_CACHE_OR
-+#define DEFAULT_DMA_CAM_NUM_OF_ENTRIES 64
-+#define DEFAULT_DMA_DBG_CNT_MODE E_FMAN_DMA_DBG_NO_CNT
-+#define DEFAULT_DMA_EN_EMERGENCY FALSE
-+#define DEFAULT_DMA_SOS_EMERGENCY 0
-+#define DEFAULT_DMA_WATCHDOG 0 /* disabled */
-+#define DEFAULT_DMA_EN_EMERGENCY_SMOOTHER FALSE
-+#define DEFAULT_DMA_EMERGENCY_SWITCH_COUNTER 0
-+#define DEFAULT_DISP_LIMIT 0
-+#define DEFAULT_PRS_DISP_TH 16
-+#define DEFAULT_PLCR_DISP_TH 16
-+#define DEFAULT_KG_DISP_TH 16
-+#define DEFAULT_BMI_DISP_TH 16
-+#define DEFAULT_QMI_ENQ_DISP_TH 16
-+#define DEFAULT_QMI_DEQ_DISP_TH 16
-+#define DEFAULT_FM_CTL1_DISP_TH 16
-+#define DEFAULT_FM_CTL2_DISP_TH 16
-+#define DEFAULT_TNUM_AGING_PERIOD 4
-+
-+
-+#endif /* __FSL_FMAN_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/flib/fsl_fman_dtsec.h
-@@ -0,0 +1,1096 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#ifndef __FSL_FMAN_DTSEC_H
-+#define __FSL_FMAN_DTSEC_H
-+
-+#include "common/general.h"
-+#include "fsl_enet.h"
-+
-+/**
-+ * DOC: dTSEC Init sequence
-+ *
-+ * To prepare dTSEC block for transfer use the following call sequence:
-+ *
-+ * - fman_dtsec_defconfig() - This step is optional and yet recommended. Its
-+ * use is to obtain the default dTSEC configuration parameters.
-+ *
-+ * - Change dtsec configuration in &dtsec_cfg. This structure will be used
-+ * to customize the dTSEC behavior.
-+ *
-+ * - fman_dtsec_init() - Applies the configuration on dTSEC hardware. Note that
-+ * dTSEC is initialized while both Tx and Rx are disabled.
-+ *
-+ * - fman_dtsec_set_mac_address() - Set the station address (mac address).
-+ * This is used by dTSEC to match against received packets.
-+ *
-+ * - fman_dtsec_adjust_link() - Set the link speed and duplex parameters
-+ * after the PHY establishes the link.
-+ *
-+ * - dtsec_enable_tx() and dtsec_enable_rx() to enable transmission and
-+ * reception.
-+ */
-+
-+/**
-+ * DOC: dTSEC Graceful stop
-+ *
-+ * To temporary stop dTSEC activity use fman_dtsec_stop_tx() and
-+ * fman_dtsec_stop_rx(). Note that these functions request dTSEC graceful stop
-+ * but return before this stop is complete. To query for graceful stop
-+ * completion use fman_dtsec_get_event() and check DTSEC_IEVENT_GTSC and
-+ * DTSEC_IEVENT_GRSC bits. Alternatively the dTSEC interrupt mask can be set to
-+ * enable graceful stop interrupts.
-+ *
-+ * To resume operation after graceful stop use fman_dtsec_start_tx() and
-+ * fman_dtsec_start_rx().
-+ */
-+
-+/**
-+ * DOC: dTSEC interrupt handling
-+ *
-+ * This code does not provide an interrupt handler for dTSEC. Instead this
-+ * handler should be implemented and registered to the operating system by the
-+ * caller. Some primitives for accessing the event status and mask registers
-+ * are provided.
-+ *
-+ * See "dTSEC Events" section for a list of events that dTSEC can generate.
-+ */
-+
-+/**
-+ * DOC: dTSEC Events
-+ *
-+ * Interrupt events cause dTSEC event bits to be set. Software may poll the
-+ * event register at any time to check for pending interrupts. If an event
-+ * occurs and its corresponding enable bit is set in the interrupt mask
-+ * register, the event also causes a hardware interrupt at the PIC.
-+ *
-+ * To poll for event status use the fman_dtsec_get_event() function.
-+ * To configure the interrupt mask use fman_dtsec_enable_interrupt() and
-+ * fman_dtsec_disable_interrupt() functions.
-+ * After servicing a dTSEC interrupt use fman_dtsec_ack_event to reset the
-+ * serviced event bit.
-+ *
-+ * The following events may be signaled by dTSEC hardware:
-+ *
-+ * %DTSEC_IEVENT_BABR - Babbling receive error. This bit indicates that
-+ * a frame was received with length in excess of the MAC's maximum frame length
-+ * register.
-+ *
-+ * %DTSEC_IEVENT_RXC - Receive control (pause frame) interrupt. A pause
-+ * control frame was received while Rx pause frame handling is enabled.
-+ * Also see fman_dtsec_handle_rx_pause().
-+ *
-+ * %DTSEC_IEVENT_MSRO - MIB counter overflow. The count for one of the MIB
-+ * counters has exceeded the size of its register.
-+ *
-+ * %DTSEC_IEVENT_GTSC - Graceful transmit stop complete. Graceful stop is now
-+ * complete. The transmitter is in a stopped state, in which only pause frames
-+ * can be transmitted.
-+ * Also see fman_dtsec_stop_tx().
-+ *
-+ * %DTSEC_IEVENT_BABT - Babbling transmit error. The transmitted frame length
-+ * has exceeded the value in the MAC's Maximum Frame Length register.
-+ *
-+ * %DTSEC_IEVENT_TXC - Transmit control (pause frame) interrupt. his bit
-+ * indicates that a control frame was transmitted.
-+ *
-+ * %DTSEC_IEVENT_TXE - Transmit error. This bit indicates that an error
-+ * occurred on the transmitted channel. This bit is set whenever any transmit
-+ * error occurs which causes the dTSEC to discard all or part of a frame
-+ * (LC, CRL, XFUN).
-+ *
-+ * %DTSEC_IEVENT_LC - Late collision. This bit indicates that a collision
-+ * occurred beyond the collision window (slot time) in half-duplex mode.
-+ * The frame is truncated with a bad CRC and the remainder of the frame
-+ * is discarded.
-+ *
-+ * %DTSEC_IEVENT_CRL - Collision retry limit. is bit indicates that the number
-+ * of successive transmission collisions has exceeded the MAC's half-duplex
-+ * register's retransmission maximum count. The frame is discarded without
-+ * being transmitted and transmission of the next frame commences. This only
-+ * occurs while in half-duplex mode.
-+ * The number of retransmit attempts can be set in
-+ * &dtsec_halfdup_cfg.@retransmit before calling fman_dtsec_init().
-+ *
-+ * %DTSEC_IEVENT_XFUN - Transmit FIFO underrun. This bit indicates that the
-+ * transmit FIFO became empty before the complete frame was transmitted.
-+ * The frame is truncated with a bad CRC and the remainder of the frame is
-+ * discarded.
-+ *
-+ * %DTSEC_IEVENT_MAG - TBD
-+ *
-+ * %DTSEC_IEVENT_MMRD - MII management read completion.
-+ *
-+ * %DTSEC_IEVENT_MMWR - MII management write completion.
-+ *
-+ * %DTSEC_IEVENT_GRSC - Graceful receive stop complete. It allows the user to
-+ * know if the system has completed the stop and it is safe to write to receive
-+ * registers (status, control or configuration registers) that are used by the
-+ * system during normal operation.
-+ *
-+ * %DTSEC_IEVENT_TDPE - Internal data error on transmit. This bit indicates
-+ * that the dTSEC has detected a parity error on its stored transmit data, which
-+ * is likely to compromise the validity of recently transferred frames.
-+ *
-+ * %DTSEC_IEVENT_RDPE - Internal data error on receive. This bit indicates that
-+ * the dTSEC has detected a parity error on its stored receive data, which is
-+ * likely to compromise the validity of recently transferred frames.
-+ */
-+/* Interrupt Mask Register (IMASK) */
-+#define DTSEC_IMASK_BREN 0x80000000
-+#define DTSEC_IMASK_RXCEN 0x40000000
-+#define DTSEC_IMASK_MSROEN 0x04000000
-+#define DTSEC_IMASK_GTSCEN 0x02000000
-+#define DTSEC_IMASK_BTEN 0x01000000
-+#define DTSEC_IMASK_TXCEN 0x00800000
-+#define DTSEC_IMASK_TXEEN 0x00400000
-+#define DTSEC_IMASK_LCEN 0x00040000
-+#define DTSEC_IMASK_CRLEN 0x00020000
-+#define DTSEC_IMASK_XFUNEN 0x00010000
-+#define DTSEC_IMASK_ABRTEN 0x00008000
-+#define DTSEC_IMASK_IFERREN 0x00004000
-+#define DTSEC_IMASK_MAGEN 0x00000800
-+#define DTSEC_IMASK_MMRDEN 0x00000400
-+#define DTSEC_IMASK_MMWREN 0x00000200
-+#define DTSEC_IMASK_GRSCEN 0x00000100
-+#define DTSEC_IMASK_TDPEEN 0x00000002
-+#define DTSEC_IMASK_RDPEEN 0x00000001
-+
-+#define DTSEC_EVENTS_MASK \
-+ ((uint32_t)(DTSEC_IMASK_BREN | \
-+ DTSEC_IMASK_RXCEN | \
-+ DTSEC_IMASK_BTEN | \
-+ DTSEC_IMASK_TXCEN | \
-+ DTSEC_IMASK_TXEEN | \
-+ DTSEC_IMASK_ABRTEN | \
-+ DTSEC_IMASK_LCEN | \
-+ DTSEC_IMASK_CRLEN | \
-+ DTSEC_IMASK_XFUNEN | \
-+ DTSEC_IMASK_IFERREN | \
-+ DTSEC_IMASK_MAGEN | \
-+ DTSEC_IMASK_TDPEEN | \
-+ DTSEC_IMASK_RDPEEN))
-+
-+/* dtsec timestamp event bits */
-+#define TMR_PEMASK_TSREEN 0x00010000
-+#define TMR_PEVENT_TSRE 0x00010000
-+
-+/* Group address bit indication */
-+#define MAC_GROUP_ADDRESS 0x0000010000000000ULL
-+/* size in bytes of L2 address */
-+#define MAC_ADDRLEN 6
-+
-+#define DEFAULT_HALFDUP_ON FALSE
-+#define DEFAULT_HALFDUP_RETRANSMIT 0xf
-+#define DEFAULT_HALFDUP_COLL_WINDOW 0x37
-+#define DEFAULT_HALFDUP_EXCESS_DEFER TRUE
-+#define DEFAULT_HALFDUP_NO_BACKOFF FALSE
-+#define DEFAULT_HALFDUP_BP_NO_BACKOFF FALSE
-+#define DEFAULT_HALFDUP_ALT_BACKOFF_VAL 0x0A
-+#define DEFAULT_HALFDUP_ALT_BACKOFF_EN FALSE
-+#define DEFAULT_RX_DROP_BCAST FALSE
-+#define DEFAULT_RX_SHORT_FRM TRUE
-+#define DEFAULT_RX_LEN_CHECK FALSE
-+#define DEFAULT_TX_PAD_CRC TRUE
-+#define DEFAULT_TX_CRC FALSE
-+#define DEFAULT_RX_CTRL_ACC FALSE
-+#define DEFAULT_TX_PAUSE_TIME 0xf000
-+#define DEFAULT_TBIPA 5
-+#define DEFAULT_RX_PREPEND 0
-+#define DEFAULT_PTP_TSU_EN TRUE
-+#define DEFAULT_PTP_EXCEPTION_EN TRUE
-+#define DEFAULT_PREAMBLE_LEN 7
-+#define DEFAULT_RX_PREAMBLE FALSE
-+#define DEFAULT_TX_PREAMBLE FALSE
-+#define DEFAULT_LOOPBACK FALSE
-+#define DEFAULT_RX_TIME_STAMP_EN FALSE
-+#define DEFAULT_TX_TIME_STAMP_EN FALSE
-+#define DEFAULT_RX_FLOW TRUE
-+#define DEFAULT_TX_FLOW TRUE
-+#define DEFAULT_RX_GROUP_HASH_EXD FALSE
-+#define DEFAULT_TX_PAUSE_TIME_EXTD 0
-+#define DEFAULT_RX_PROMISC FALSE
-+#define DEFAULT_NON_BACK_TO_BACK_IPG1 0x40
-+#define DEFAULT_NON_BACK_TO_BACK_IPG2 0x60
-+#define DEFAULT_MIN_IFG_ENFORCEMENT 0x50
-+#define DEFAULT_BACK_TO_BACK_IPG 0x60
-+#define DEFAULT_MAXIMUM_FRAME 0x600
-+#define DEFAULT_TBI_PHY_ADDR 5
-+#define DEFAULT_WAKE_ON_LAN FALSE
-+
-+/* register related defines (bits, field offsets..) */
-+#define DTSEC_ID1_ID 0xffff0000
-+#define DTSEC_ID1_REV_MJ 0x0000FF00
-+#define DTSEC_ID1_REV_MN 0x000000ff
-+
-+#define DTSEC_ID2_INT_REDUCED_OFF 0x00010000
-+#define DTSEC_ID2_INT_NORMAL_OFF 0x00020000
-+
-+#define DTSEC_ECNTRL_CLRCNT 0x00004000
-+#define DTSEC_ECNTRL_AUTOZ 0x00002000
-+#define DTSEC_ECNTRL_STEN 0x00001000
-+#define DTSEC_ECNTRL_CFG_RO 0x80000000
-+#define DTSEC_ECNTRL_GMIIM 0x00000040
-+#define DTSEC_ECNTRL_TBIM 0x00000020
-+#define DTSEC_ECNTRL_SGMIIM 0x00000002
-+#define DTSEC_ECNTRL_RPM 0x00000010
-+#define DTSEC_ECNTRL_R100M 0x00000008
-+#define DTSEC_ECNTRL_RMM 0x00000004
-+#define DTSEC_ECNTRL_QSGMIIM 0x00000001
-+
-+#define DTSEC_TCTRL_THDF 0x00000800
-+#define DTSEC_TCTRL_TTSE 0x00000040
-+#define DTSEC_TCTRL_GTS 0x00000020
-+#define DTSEC_TCTRL_TFC_PAUSE 0x00000010
-+
-+/* PTV offsets */
-+#define PTV_PTE_OFST 16
-+
-+#define RCTRL_CFA 0x00008000
-+#define RCTRL_GHTX 0x00000400
-+#define RCTRL_RTSE 0x00000040
-+#define RCTRL_GRS 0x00000020
-+#define RCTRL_BC_REJ 0x00000010
-+#define RCTRL_MPROM 0x00000008
-+#define RCTRL_RSF 0x00000004
-+#define RCTRL_UPROM 0x00000001
-+#define RCTRL_PROM (RCTRL_UPROM | RCTRL_MPROM)
-+
-+#define TMR_CTL_ESFDP 0x00000800
-+#define TMR_CTL_ESFDE 0x00000400
-+
-+#define MACCFG1_SOFT_RESET 0x80000000
-+#define MACCFG1_LOOPBACK 0x00000100
-+#define MACCFG1_RX_FLOW 0x00000020
-+#define MACCFG1_TX_FLOW 0x00000010
-+#define MACCFG1_TX_EN 0x00000001
-+#define MACCFG1_RX_EN 0x00000004
-+#define MACCFG1_RESET_RxMC 0x00080000
-+#define MACCFG1_RESET_TxMC 0x00040000
-+#define MACCFG1_RESET_RxFUN 0x00020000
-+#define MACCFG1_RESET_TxFUN 0x00010000
-+
-+#define MACCFG2_NIBBLE_MODE 0x00000100
-+#define MACCFG2_BYTE_MODE 0x00000200
-+#define MACCFG2_PRE_AM_Rx_EN 0x00000080
-+#define MACCFG2_PRE_AM_Tx_EN 0x00000040
-+#define MACCFG2_LENGTH_CHECK 0x00000010
-+#define MACCFG2_MAGIC_PACKET_EN 0x00000008
-+#define MACCFG2_PAD_CRC_EN 0x00000004
-+#define MACCFG2_CRC_EN 0x00000002
-+#define MACCFG2_FULL_DUPLEX 0x00000001
-+
-+#define PREAMBLE_LENGTH_SHIFT 12
-+
-+#define IPGIFG_NON_BACK_TO_BACK_IPG_1_SHIFT 24
-+#define IPGIFG_NON_BACK_TO_BACK_IPG_2_SHIFT 16
-+#define IPGIFG_MIN_IFG_ENFORCEMENT_SHIFT 8
-+
-+#define IPGIFG_NON_BACK_TO_BACK_IPG_1 0x7F000000
-+#define IPGIFG_NON_BACK_TO_BACK_IPG_2 0x007F0000
-+#define IPGIFG_MIN_IFG_ENFORCEMENT 0x0000FF00
-+#define IPGIFG_BACK_TO_BACK_IPG 0x0000007F
-+
-+#define HAFDUP_ALT_BEB 0x00080000
-+#define HAFDUP_BP_NO_BACKOFF 0x00040000
-+#define HAFDUP_NO_BACKOFF 0x00020000
-+#define HAFDUP_EXCESS_DEFER 0x00010000
-+#define HAFDUP_COLLISION_WINDOW 0x000003ff
-+
-+#define HAFDUP_ALTERNATE_BEB_TRUNCATION_SHIFT 20
-+#define HAFDUP_RETRANSMISSION_MAX_SHIFT 12
-+#define HAFDUP_RETRANSMISSION_MAX 0x0000f000
-+
-+#define NUM_OF_HASH_REGS 8 /* Number of hash table registers */
-+
-+/* CAR1/2 bits */
-+#define DTSEC_CAR1_TR64 0x80000000
-+#define DTSEC_CAR1_TR127 0x40000000
-+#define DTSEC_CAR1_TR255 0x20000000
-+#define DTSEC_CAR1_TR511 0x10000000
-+#define DTSEC_CAR1_TRK1 0x08000000
-+#define DTSEC_CAR1_TRMAX 0x04000000
-+#define DTSEC_CAR1_TRMGV 0x02000000
-+
-+#define DTSEC_CAR1_RBYT 0x00010000
-+#define DTSEC_CAR1_RPKT 0x00008000
-+#define DTSEC_CAR1_RFCS 0x00004000
-+#define DTSEC_CAR1_RMCA 0x00002000
-+#define DTSEC_CAR1_RBCA 0x00001000
-+#define DTSEC_CAR1_RXCF 0x00000800
-+#define DTSEC_CAR1_RXPF 0x00000400
-+#define DTSEC_CAR1_RXUO 0x00000200
-+#define DTSEC_CAR1_RALN 0x00000100
-+#define DTSEC_CAR1_RFLR 0x00000080
-+#define DTSEC_CAR1_RCDE 0x00000040
-+#define DTSEC_CAR1_RCSE 0x00000020
-+#define DTSEC_CAR1_RUND 0x00000010
-+#define DTSEC_CAR1_ROVR 0x00000008
-+#define DTSEC_CAR1_RFRG 0x00000004
-+#define DTSEC_CAR1_RJBR 0x00000002
-+#define DTSEC_CAR1_RDRP 0x00000001
-+
-+#define DTSEC_CAR2_TJBR 0x00080000
-+#define DTSEC_CAR2_TFCS 0x00040000
-+#define DTSEC_CAR2_TXCF 0x00020000
-+#define DTSEC_CAR2_TOVR 0x00010000
-+#define DTSEC_CAR2_TUND 0x00008000
-+#define DTSEC_CAR2_TFRG 0x00004000
-+#define DTSEC_CAR2_TBYT 0x00002000
-+#define DTSEC_CAR2_TPKT 0x00001000
-+#define DTSEC_CAR2_TMCA 0x00000800
-+#define DTSEC_CAR2_TBCA 0x00000400
-+#define DTSEC_CAR2_TXPF 0x00000200
-+#define DTSEC_CAR2_TDFR 0x00000100
-+#define DTSEC_CAR2_TEDF 0x00000080
-+#define DTSEC_CAR2_TSCL 0x00000040
-+#define DTSEC_CAR2_TMCL 0x00000020
-+#define DTSEC_CAR2_TLCL 0x00000010
-+#define DTSEC_CAR2_TXCL 0x00000008
-+#define DTSEC_CAR2_TNCL 0x00000004
-+#define DTSEC_CAR2_TDRP 0x00000001
-+
-+#define CAM1_ERRORS_ONLY \
-+ (DTSEC_CAR1_RXPF | DTSEC_CAR1_RALN | DTSEC_CAR1_RFLR \
-+ | DTSEC_CAR1_RCDE | DTSEC_CAR1_RCSE | DTSEC_CAR1_RUND \
-+ | DTSEC_CAR1_ROVR | DTSEC_CAR1_RFRG | DTSEC_CAR1_RJBR \
-+ | DTSEC_CAR1_RDRP)
-+
-+#define CAM2_ERRORS_ONLY (DTSEC_CAR2_TFCS | DTSEC_CAR2_TXPF | DTSEC_CAR2_TDRP)
-+
-+/*
-+ * Group of dTSEC specific counters relating to the standard RMON MIB Group 1
-+ * (or Ethernet) statistics.
-+ */
-+#define CAM1_MIB_GRP_1 \
-+ (DTSEC_CAR1_RDRP | DTSEC_CAR1_RBYT | DTSEC_CAR1_RPKT | DTSEC_CAR1_RMCA\
-+ | DTSEC_CAR1_RBCA | DTSEC_CAR1_RALN | DTSEC_CAR1_RUND | DTSEC_CAR1_ROVR\
-+ | DTSEC_CAR1_RFRG | DTSEC_CAR1_RJBR \
-+ | DTSEC_CAR1_TR64 | DTSEC_CAR1_TR127 | DTSEC_CAR1_TR255 \
-+ | DTSEC_CAR1_TR511 | DTSEC_CAR1_TRMAX)
-+
-+#define CAM2_MIB_GRP_1 (DTSEC_CAR2_TNCL | DTSEC_CAR2_TDRP)
-+
-+/* memory map */
-+
-+struct dtsec_regs {
-+ /* dTSEC General Control and Status Registers */
-+ uint32_t tsec_id; /* 0x000 ETSEC_ID register */
-+ uint32_t tsec_id2; /* 0x004 ETSEC_ID2 register */
-+ uint32_t ievent; /* 0x008 Interrupt event register */
-+ uint32_t imask; /* 0x00C Interrupt mask register */
-+ uint32_t reserved0010[1];
-+ uint32_t ecntrl; /* 0x014 E control register */
-+ uint32_t ptv; /* 0x018 Pause time value register */
-+ uint32_t tbipa; /* 0x01C TBI PHY address register */
-+ uint32_t tmr_ctrl; /* 0x020 Time-stamp Control register */
-+ uint32_t tmr_pevent; /* 0x024 Time-stamp event register */
-+ uint32_t tmr_pemask; /* 0x028 Timer event mask register */
-+ uint32_t reserved002c[5];
-+ uint32_t tctrl; /* 0x040 Transmit control register */
-+ uint32_t reserved0044[3];
-+ uint32_t rctrl; /* 0x050 Receive control register */
-+ uint32_t reserved0054[11];
-+ uint32_t igaddr[8]; /* 0x080-0x09C Individual/group address */
-+ uint32_t gaddr[8]; /* 0x0A0-0x0BC Group address registers 0-7 */
-+ uint32_t reserved00c0[16];
-+ uint32_t maccfg1; /* 0x100 MAC configuration #1 */
-+ uint32_t maccfg2; /* 0x104 MAC configuration #2 */
-+ uint32_t ipgifg; /* 0x108 IPG/IFG */
-+ uint32_t hafdup; /* 0x10C Half-duplex */
-+ uint32_t maxfrm; /* 0x110 Maximum frame */
-+ uint32_t reserved0114[10];
-+ uint32_t ifstat; /* 0x13C Interface status */
-+ uint32_t macstnaddr1; /* 0x140 Station Address,part 1 */
-+ uint32_t macstnaddr2; /* 0x144 Station Address,part 2 */
-+ struct {
-+ uint32_t exact_match1; /* octets 1-4 */
-+ uint32_t exact_match2; /* octets 5-6 */
-+ } macaddr[15]; /* 0x148-0x1BC mac exact match addresses 1-15 */
-+ uint32_t reserved01c0[16];
-+ uint32_t tr64; /* 0x200 transmit and receive 64 byte frame counter */
-+ uint32_t tr127; /* 0x204 transmit and receive 65 to 127 byte frame
-+ * counter */
-+ uint32_t tr255; /* 0x208 transmit and receive 128 to 255 byte frame
-+ * counter */
-+ uint32_t tr511; /* 0x20C transmit and receive 256 to 511 byte frame
-+ * counter */
-+ uint32_t tr1k; /* 0x210 transmit and receive 512 to 1023 byte frame
-+ * counter */
-+ uint32_t trmax; /* 0x214 transmit and receive 1024 to 1518 byte frame
-+ * counter */
-+ uint32_t trmgv; /* 0x218 transmit and receive 1519 to 1522 byte good
-+ * VLAN frame count */
-+ uint32_t rbyt; /* 0x21C receive byte counter */
-+ uint32_t rpkt; /* 0x220 receive packet counter */
-+ uint32_t rfcs; /* 0x224 receive FCS error counter */
-+ uint32_t rmca; /* 0x228 RMCA receive multicast packet counter */
-+ uint32_t rbca; /* 0x22C receive broadcast packet counter */
-+ uint32_t rxcf; /* 0x230 receive control frame packet counter */
-+ uint32_t rxpf; /* 0x234 receive pause frame packet counter */
-+ uint32_t rxuo; /* 0x238 receive unknown OP code counter */
-+ uint32_t raln; /* 0x23C receive alignment error counter */
-+ uint32_t rflr; /* 0x240 receive frame length error counter */
-+ uint32_t rcde; /* 0x244 receive code error counter */
-+ uint32_t rcse; /* 0x248 receive carrier sense error counter */
-+ uint32_t rund; /* 0x24C receive undersize packet counter */
-+ uint32_t rovr; /* 0x250 receive oversize packet counter */
-+ uint32_t rfrg; /* 0x254 receive fragments counter */
-+ uint32_t rjbr; /* 0x258 receive jabber counter */
-+ uint32_t rdrp; /* 0x25C receive drop */
-+ uint32_t tbyt; /* 0x260 transmit byte counter */
-+ uint32_t tpkt; /* 0x264 transmit packet counter */
-+ uint32_t tmca; /* 0x268 transmit multicast packet counter */
-+ uint32_t tbca; /* 0x26C transmit broadcast packet counter */
-+ uint32_t txpf; /* 0x270 transmit pause control frame counter */
-+ uint32_t tdfr; /* 0x274 transmit deferral packet counter */
-+ uint32_t tedf; /* 0x278 transmit excessive deferral packet counter */
-+ uint32_t tscl; /* 0x27C transmit single collision packet counter */
-+ uint32_t tmcl; /* 0x280 transmit multiple collision packet counter */
-+ uint32_t tlcl; /* 0x284 transmit late collision packet counter */
-+ uint32_t txcl; /* 0x288 transmit excessive collision packet counter */
-+ uint32_t tncl; /* 0x28C transmit total collision counter */
-+ uint32_t reserved0290[1];
-+ uint32_t tdrp; /* 0x294 transmit drop frame counter */
-+ uint32_t tjbr; /* 0x298 transmit jabber frame counter */
-+ uint32_t tfcs; /* 0x29C transmit FCS error counter */
-+ uint32_t txcf; /* 0x2A0 transmit control frame counter */
-+ uint32_t tovr; /* 0x2A4 transmit oversize frame counter */
-+ uint32_t tund; /* 0x2A8 transmit undersize frame counter */
-+ uint32_t tfrg; /* 0x2AC transmit fragments frame counter */
-+ uint32_t car1; /* 0x2B0 carry register one register* */
-+ uint32_t car2; /* 0x2B4 carry register two register* */
-+ uint32_t cam1; /* 0x2B8 carry register one mask register */
-+ uint32_t cam2; /* 0x2BC carry register two mask register */
-+ uint32_t reserved02c0[848];
-+};
-+
-+/**
-+ * struct dtsec_mib_grp_1_counters - MIB counter overflows
-+ *
-+ * @tr64: Transmit and Receive 64 byte frame count. Increment for each
-+ * good or bad frame, of any type, transmitted or received, which
-+ * is 64 bytes in length.
-+ * @tr127: Transmit and Receive 65 to 127 byte frame count. Increments for
-+ * each good or bad frame of any type, transmitted or received,
-+ * which is 65-127 bytes in length.
-+ * @tr255: Transmit and Receive 128 to 255 byte frame count. Increments
-+ * for each good or bad frame, of any type, transmitted or
-+ * received, which is 128-255 bytes in length.
-+ * @tr511: Transmit and Receive 256 to 511 byte frame count. Increments
-+ * for each good or bad frame, of any type, transmitted or
-+ * received, which is 256-511 bytes in length.
-+ * @tr1k: Transmit and Receive 512 to 1023 byte frame count. Increments
-+ * for each good or bad frame, of any type, transmitted or
-+ * received, which is 512-1023 bytes in length.
-+ * @trmax: Transmit and Receive 1024 to 1518 byte frame count. Increments
-+ * for each good or bad frame, of any type, transmitted or
-+ * received, which is 1024-1518 bytes in length.
-+ * @rfrg: Receive fragments count. Increments for each received frame
-+ * which is less than 64 bytes in length and contains an invalid
-+ * FCS. This includes integral and non-integral lengths.
-+ * @rjbr: Receive jabber count. Increments for received frames which
-+ * exceed 1518 (non VLAN) or 1522 (VLAN) bytes and contain an
-+ * invalid FCS. This includes alignment errors.
-+ * @rdrp: Receive dropped packets count. Increments for received frames
-+ * which are streamed to system but are later dropped due to lack
-+ * of system resources. Does not increment for frames rejected due
-+ * to address filtering.
-+ * @raln: Receive alignment error count. Increments for each received
-+ * frame from 64 to 1518 (non VLAN) or 1522 (VLAN) which contains
-+ * an invalid FCS and is not an integral number of bytes.
-+ * @rund: Receive undersize packet count. Increments each time a frame is
-+ * received which is less than 64 bytes in length and contains a
-+ * valid FCS and is otherwise well formed. This count does not
-+ * include range length errors.
-+ * @rovr: Receive oversize packet count. Increments each time a frame is
-+ * received which exceeded 1518 (non VLAN) or 1522 (VLAN) and
-+ * contains a valid FCS and is otherwise well formed.
-+ * @rbyt: Receive byte count. Increments by the byte count of frames
-+ * received, including those in bad packets, excluding preamble and
-+ * SFD but including FCS bytes.
-+ * @rpkt: Receive packet count. Increments for each received frame
-+ * (including bad packets, all unicast, broadcast, and multicast
-+ * packets).
-+ * @rmca: Receive multicast packet count. Increments for each multicast
-+ * frame with valid CRC and of lengths 64 to 1518 (non VLAN) or
-+ * 1522 (VLAN), excluding broadcast frames. This count does not
-+ * include range/length errors.
-+ * @rbca: Receive broadcast packet count. Increments for each broadcast
-+ * frame with valid CRC and of lengths 64 to 1518 (non VLAN) or
-+ * 1522 (VLAN), excluding multicast frames. Does not include
-+ * range/length errors.
-+ * @tdrp: Transmit drop frame count. Increments each time a memory error
-+ * or an underrun has occurred.
-+ * @tncl: Transmit total collision counter. Increments by the number of
-+ * collisions experienced during the transmission of a frame. Does
-+ * not increment for aborted frames.
-+ *
-+ * The structure contains a group of dTSEC HW specific counters relating to the
-+ * standard RMON MIB Group 1 (or Ethernet statistics) counters. This structure
-+ * is counting only the carry events of the corresponding HW counters.
-+ *
-+ * tr64 to trmax notes: Frame sizes specified are considered excluding preamble
-+ * and SFD but including FCS bytes.
-+ */
-+struct dtsec_mib_grp_1_counters {
-+ uint64_t rdrp;
-+ uint64_t tdrp;
-+ uint64_t rbyt;
-+ uint64_t rpkt;
-+ uint64_t rbca;
-+ uint64_t rmca;
-+ uint64_t raln;
-+ uint64_t rund;
-+ uint64_t rovr;
-+ uint64_t rfrg;
-+ uint64_t rjbr;
-+ uint64_t tncl;
-+ uint64_t tr64;
-+ uint64_t tr127;
-+ uint64_t tr255;
-+ uint64_t tr511;
-+ uint64_t tr1k;
-+ uint64_t trmax;
-+};
-+
-+enum dtsec_stat_counters {
-+ E_DTSEC_STAT_TR64,
-+ E_DTSEC_STAT_TR127,
-+ E_DTSEC_STAT_TR255,
-+ E_DTSEC_STAT_TR511,
-+ E_DTSEC_STAT_TR1K,
-+ E_DTSEC_STAT_TRMAX,
-+ E_DTSEC_STAT_TRMGV,
-+ E_DTSEC_STAT_RBYT,
-+ E_DTSEC_STAT_RPKT,
-+ E_DTSEC_STAT_RMCA,
-+ E_DTSEC_STAT_RBCA,
-+ E_DTSEC_STAT_RXPF,
-+ E_DTSEC_STAT_RALN,
-+ E_DTSEC_STAT_RFLR,
-+ E_DTSEC_STAT_RCDE,
-+ E_DTSEC_STAT_RCSE,
-+ E_DTSEC_STAT_RUND,
-+ E_DTSEC_STAT_ROVR,
-+ E_DTSEC_STAT_RFRG,
-+ E_DTSEC_STAT_RJBR,
-+ E_DTSEC_STAT_RDRP,
-+ E_DTSEC_STAT_TFCS,
-+ E_DTSEC_STAT_TBYT,
-+ E_DTSEC_STAT_TPKT,
-+ E_DTSEC_STAT_TMCA,
-+ E_DTSEC_STAT_TBCA,
-+ E_DTSEC_STAT_TXPF,
-+ E_DTSEC_STAT_TNCL,
-+ E_DTSEC_STAT_TDRP
-+};
-+
-+enum dtsec_stat_level {
-+ /* No statistics */
-+ E_MAC_STAT_NONE = 0,
-+ /* Only RMON MIB group 1 (ether stats). Optimized for performance */
-+ E_MAC_STAT_MIB_GRP1,
-+ /* Only error counters are available. Optimized for performance */
-+ E_MAC_STAT_PARTIAL,
-+ /* All counters available. Not optimized for performance */
-+ E_MAC_STAT_FULL
-+};
-+
-+
-+/**
-+ * struct dtsec_cfg - dTSEC configuration
-+ *
-+ * @halfdup_on: Transmit half-duplex flow control, under software
-+ * control for 10/100-Mbps half-duplex media. If set,
-+ * back pressure is applied to media by raising carrier.
-+ * @halfdup_retransmit: Number of retransmission attempts following a collision.
-+ * If this is exceeded dTSEC aborts transmission due to
-+ * excessive collisions. The standard specifies the
-+ * attempt limit to be 15.
-+ * @halfdup_coll_window:The number of bytes of the frame during which
-+ * collisions may occur. The default value of 55
-+ * corresponds to the frame byte at the end of the
-+ * standard 512-bit slot time window. If collisions are
-+ * detected after this byte, the late collision event is
-+ * asserted and transmission of current frame is aborted.
-+ * @rx_drop_bcast: Discard broadcast frames. If set, all broadcast frames
-+ * will be discarded by dTSEC.
-+ * @rx_short_frm: Accept short frames. If set, dTSEC will accept frames
-+ * of length 14..63 bytes.
-+ * @rx_len_check: Length check for received frames. If set, the MAC
-+ * checks the frame's length field on receive to ensure it
-+ * matches the actual data field length. This only works
-+ * for received frames with length field less than 1500.
-+ * No check is performed for larger frames.
-+ * @tx_pad_crc: Pad and append CRC. If set, the MAC pads all
-+ * transmitted short frames and appends a CRC to every
-+ * frame regardless of padding requirement.
-+ * @tx_crc: Transmission CRC enable. If set, the MAC appends a CRC
-+ * to all frames. If frames presented to the MAC have a
-+ * valid length and contain a valid CRC, @tx_crc should be
-+ * reset.
-+ * This field is ignored if @tx_pad_crc is set.
-+ * @rx_ctrl_acc: Control frame accept. If set, this overrides 802.3
-+ * standard control frame behavior, and all Ethernet frames
-+ * that have an ethertype of 0x8808 are treated as normal
-+ * Ethernet frames and passed up to the packet interface on
-+ * a DA match. Received pause control frames are passed to
-+ * the packet interface only if Rx flow control is also
-+ * disabled. See fman_dtsec_handle_rx_pause() function.
-+ * @tx_pause_time: Transmit pause time value. This pause value is used as
-+ * part of the pause frame to be sent when a transmit pause
-+ * frame is initiated. If set to 0 this disables
-+ * transmission of pause frames.
-+ * @rx_preamble: Receive preamble enable. If set, the MAC recovers the
-+ * received Ethernet 7-byte preamble and passes it to the
-+ * packet interface at the start of each received frame.
-+ * This field should be reset for internal MAC loop-back
-+ * mode.
-+ * @tx_preamble: User defined preamble enable for transmitted frames.
-+ * If set, a user-defined preamble must passed to the MAC
-+ * and it is transmitted instead of the standard preamble.
-+ * @preamble_len: Length, in bytes, of the preamble field preceding each
-+ * Ethernet start-of-frame delimiter byte. The default
-+ * value of 0x7 should be used in order to guarantee
-+ * reliable operation with IEEE 802.3 compliant hardware.
-+ * @rx_prepend: Packet alignment padding length. The specified number
-+ * of bytes (1-31) of zero padding are inserted before the
-+ * start of each received frame. For Ethernet, where
-+ * optional preamble extraction is enabled, the padding
-+ * appears before the preamble, otherwise the padding
-+ * precedes the layer 2 header.
-+ *
-+ * This structure contains basic dTSEC configuration and must be passed to
-+ * fman_dtsec_init() function. A default set of configuration values can be
-+ * obtained by calling fman_dtsec_defconfig().
-+ */
-+struct dtsec_cfg {
-+ bool halfdup_on;
-+ bool halfdup_alt_backoff_en;
-+ bool halfdup_excess_defer;
-+ bool halfdup_no_backoff;
-+ bool halfdup_bp_no_backoff;
-+ uint8_t halfdup_alt_backoff_val;
-+ uint16_t halfdup_retransmit;
-+ uint16_t halfdup_coll_window;
-+ bool rx_drop_bcast;
-+ bool rx_short_frm;
-+ bool rx_len_check;
-+ bool tx_pad_crc;
-+ bool tx_crc;
-+ bool rx_ctrl_acc;
-+ unsigned short tx_pause_time;
-+ unsigned short tbipa;
-+ bool ptp_tsu_en;
-+ bool ptp_exception_en;
-+ bool rx_preamble;
-+ bool tx_preamble;
-+ unsigned char preamble_len;
-+ unsigned char rx_prepend;
-+ bool loopback;
-+ bool rx_time_stamp_en;
-+ bool tx_time_stamp_en;
-+ bool rx_flow;
-+ bool tx_flow;
-+ bool rx_group_hash_exd;
-+ bool rx_promisc;
-+ uint8_t tbi_phy_addr;
-+ uint16_t tx_pause_time_extd;
-+ uint16_t maximum_frame;
-+ uint32_t non_back_to_back_ipg1;
-+ uint32_t non_back_to_back_ipg2;
-+ uint32_t min_ifg_enforcement;
-+ uint32_t back_to_back_ipg;
-+ bool wake_on_lan;
-+};
-+
-+
-+/**
-+ * fman_dtsec_defconfig() - Get default dTSEC configuration
-+ * @cfg: pointer to configuration structure.
-+ *
-+ * Call this function to obtain a default set of configuration values for
-+ * initializing dTSEC. The user can overwrite any of the values before calling
-+ * fman_dtsec_init(), if specific configuration needs to be applied.
-+ */
-+void fman_dtsec_defconfig(struct dtsec_cfg *cfg);
-+
-+/**
-+ * fman_dtsec_init() - Init dTSEC hardware block
-+ * @regs: Pointer to dTSEC register block
-+ * @cfg: dTSEC configuration data
-+ * @iface_mode: dTSEC interface mode, the type of MAC - PHY interface.
-+ * @iface_speed: 1G or 10G
-+ * @macaddr: MAC station address to be assigned to the device
-+ * @fm_rev_maj: major rev number
-+ * @fm_rev_min: minor rev number
-+ * @exceptions_mask: initial exceptions mask
-+ *
-+ * This function initializes dTSEC and applies basic configuration.
-+ *
-+ * dTSEC initialization sequence:
-+ * Before enabling Rx/Tx call dtsec_set_address() to set MAC address,
-+ * fman_dtsec_adjust_link() to configure interface speed and duplex and finally
-+ * dtsec_enable_tx()/dtsec_enable_rx() to start transmission and reception.
-+ *
-+ * Returns: 0 if successful, an error code otherwise.
-+ */
-+int fman_dtsec_init(struct dtsec_regs *regs, struct dtsec_cfg *cfg,
-+ enum enet_interface iface_mode,
-+ enum enet_speed iface_speed,
-+ uint8_t *macaddr, uint8_t fm_rev_maj,
-+ uint8_t fm_rev_min,
-+ uint32_t exception_mask);
-+
-+/**
-+ * fman_dtsec_enable() - Enable dTSEC Tx and Tx
-+ * @regs: Pointer to dTSEC register block
-+ * @apply_rx: enable rx side
-+ * @apply_tx: enable tx side
-+ *
-+ * This function resets Tx and Rx graceful stop bit and enables dTSEC Tx and Rx.
-+ */
-+void fman_dtsec_enable(struct dtsec_regs *regs, bool apply_rx, bool apply_tx);
-+
-+/**
-+ * fman_dtsec_disable() - Disable dTSEC Tx and Rx
-+ * @regs: Pointer to dTSEC register block
-+ * @apply_rx: disable rx side
-+ * @apply_tx: disable tx side
-+ *
-+ * This function disables Tx and Rx in dTSEC.
-+ */
-+void fman_dtsec_disable(struct dtsec_regs *regs, bool apply_rx, bool apply_tx);
-+
-+/**
-+ * fman_dtsec_get_revision() - Get dTSEC hardware revision
-+ * @regs: Pointer to dTSEC register block
-+ *
-+ * Returns dtsec_id content
-+ *
-+ * Call this function to obtain the dTSEC hardware version.
-+ */
-+uint32_t fman_dtsec_get_revision(struct dtsec_regs *regs);
-+
-+/**
-+ * fman_dtsec_set_mac_address() - Set MAC station address
-+ * @regs: Pointer to dTSEC register block
-+ * @macaddr: MAC address array
-+ *
-+ * This function sets MAC station address. To enable unicast reception call
-+ * this after fman_dtsec_init(). While promiscuous mode is disabled dTSEC will
-+ * match the destination address of received unicast frames against this
-+ * address.
-+ */
-+void fman_dtsec_set_mac_address(struct dtsec_regs *regs, uint8_t *macaddr);
-+
-+/**
-+ * fman_dtsec_get_mac_address() - Query MAC station address
-+ * @regs: Pointer to dTSEC register block
-+ * @macaddr: MAC address array
-+ */
-+void fman_dtsec_get_mac_address(struct dtsec_regs *regs, uint8_t *macaddr);
-+
-+/**
-+ * fman_dtsec_set_uc_promisc() - Sets unicast promiscuous mode
-+ * @regs: Pointer to dTSEC register block
-+ * @enable: Enable unicast promiscuous mode
-+ *
-+ * Use this function to enable/disable dTSEC L2 address filtering. If the
-+ * address filtering is disabled all unicast packets are accepted.
-+ * To set dTSEC in promiscuous mode call both fman_dtsec_set_uc_promisc() and
-+ * fman_dtsec_set_mc_promisc() to disable filtering for both unicast and
-+ * multicast addresses.
-+ */
-+void fman_dtsec_set_uc_promisc(struct dtsec_regs *regs, bool enable);
-+
-+/**
-+ * fman_dtsec_set_wol() - Enable/Disable wake on lan
-+ * (magic packet support)
-+ * @regs: Pointer to dTSEC register block
-+ * @en: Enable Wake On Lan support in dTSEC
-+ *
-+ */
-+void fman_dtsec_set_wol(struct dtsec_regs *regs, bool en);
-+
-+/**
-+ * fman_dtsec_adjust_link() - Adjust dTSEC speed/duplex settings
-+ * @regs: Pointer to dTSEC register block
-+ * @iface_mode: dTSEC interface mode
-+ * @speed: Link speed
-+ * @full_dx: True for full-duplex, false for half-duplex.
-+ *
-+ * This function configures the MAC to function and the desired rates. Use it
-+ * to configure dTSEC after fman_dtsec_init() and whenever the link speed
-+ * changes (for instance following PHY auto-negociation).
-+ *
-+ * Returns: 0 if successful, an error code otherwise.
-+ */
-+int fman_dtsec_adjust_link(struct dtsec_regs *regs,
-+ enum enet_interface iface_mode,
-+ enum enet_speed speed, bool full_dx);
-+
-+/**
-+ * fman_dtsec_set_tbi_phy_addr() - Updates TBI address field
-+ * @regs: Pointer to dTSEC register block
-+ * @address: Valid PHY address in the range of 1 to 31. 0 is reserved.
-+ *
-+ * In SGMII mode, the dTSEC's TBIPA field must contain a valid TBI PHY address
-+ * so that the associated TBI PHY (i.e. the link) may be initialized.
-+ *
-+ * Returns: 0 if successful, an error code otherwise.
-+ */
-+int fman_dtsec_set_tbi_phy_addr(struct dtsec_regs *regs,
-+ uint8_t addr);
-+
-+/**
-+ * fman_dtsec_set_max_frame_len() - Set max frame length
-+ * @regs: Pointer to dTSEC register block
-+ * @length: Max frame length.
-+ *
-+ * Sets maximum frame length for received and transmitted frames. Frames that
-+ * exceeds this length are truncated.
-+ */
-+void fman_dtsec_set_max_frame_len(struct dtsec_regs *regs, uint16_t length);
-+
-+/**
-+ * fman_dtsec_get_max_frame_len() - Query max frame length
-+ * @regs: Pointer to dTSEC register block
-+ *
-+ * Returns: the current value of the maximum frame length.
-+ */
-+uint16_t fman_dtsec_get_max_frame_len(struct dtsec_regs *regs);
-+
-+/**
-+ * fman_dtsec_handle_rx_pause() - Configure pause frame handling
-+ * @regs: Pointer to dTSEC register block
-+ * @en: Enable pause frame handling in dTSEC
-+ *
-+ * If enabled, dTSEC will handle pause frames internally. This must be disabled
-+ * if dTSEC is set in half-duplex mode.
-+ * If pause frame handling is disabled and &dtsec_cfg.rx_ctrl_acc is set, pause
-+ * frames will be transferred to the packet interface just like regular Ethernet
-+ * frames.
-+ */
-+void fman_dtsec_handle_rx_pause(struct dtsec_regs *regs, bool en);
-+
-+/**
-+ * fman_dtsec_set_tx_pause_frames() - Configure Tx pause time
-+ * @regs: Pointer to dTSEC register block
-+ * @time: Time value included in pause frames
-+ *
-+ * Call this function to set the time value used in transmitted pause frames.
-+ * If time is 0, transmission of pause frames is disabled
-+ */
-+void fman_dtsec_set_tx_pause_frames(struct dtsec_regs *regs, uint16_t time);
-+
-+/**
-+ * fman_dtsec_ack_event() - Acknowledge handled events
-+ * @regs: Pointer to dTSEC register block
-+ * @ev_mask: Events to acknowledge
-+ *
-+ * After handling events signaled by dTSEC in either polling or interrupt mode,
-+ * call this function to reset the associated status bits in dTSEC event
-+ * register.
-+ */
-+void fman_dtsec_ack_event(struct dtsec_regs *regs, uint32_t ev_mask);
-+
-+/**
-+ * fman_dtsec_get_event() - Returns currently asserted events
-+ * @regs: Pointer to dTSEC register block
-+ * @ev_mask: Mask of relevant events
-+ *
-+ * Call this function to obtain a bit-mask of events that are currently asserted
-+ * in dTSEC, taken from IEVENT register.
-+ *
-+ * Returns: a bit-mask of events asserted in dTSEC.
-+ */
-+uint32_t fman_dtsec_get_event(struct dtsec_regs *regs, uint32_t ev_mask);
-+
-+/**
-+ * fman_dtsec_get_interrupt_mask() - Returns a bit-mask of enabled interrupts
-+ * @regs: Pointer to dTSEC register block
-+ *
-+ * Call this function to obtain a bit-mask of enabled interrupts
-+ * in dTSEC, taken from IMASK register.
-+ *
-+ * Returns: a bit-mask of enabled interrupts in dTSEC.
-+ */
-+uint32_t fman_dtsec_get_interrupt_mask(struct dtsec_regs *regs);
-+
-+void fman_dtsec_clear_addr_in_paddr(struct dtsec_regs *regs,
-+ uint8_t paddr_num);
-+
-+void fman_dtsec_add_addr_in_paddr(struct dtsec_regs *regs,
-+ uint64_t addr,
-+ uint8_t paddr_num);
-+
-+void fman_dtsec_enable_tmr_interrupt (struct dtsec_regs *regs);
-+
-+void fman_dtsec_disable_tmr_interrupt(struct dtsec_regs *regs);
-+
-+/**
-+ * fman_dtsec_disable_interrupt() - Disables interrupts for the specified events
-+ * @regs: Pointer to dTSEC register block
-+ * @ev_mask: Mask of relevant events
-+ *
-+ * Call this function to disable interrupts in dTSEC for the specified events.
-+ * To enable interrupts use fman_dtsec_enable_interrupt().
-+ */
-+void fman_dtsec_disable_interrupt(struct dtsec_regs *regs, uint32_t ev_mask);
-+
-+/**
-+ * fman_dtsec_enable_interrupt() - Enable interrupts for the specified events
-+ * @regs: Pointer to dTSEC register block
-+ * @ev_mask: Mask of relevant events
-+ *
-+ * Call this function to enable interrupts in dTSEC for the specified events.
-+ * To disable interrupts use fman_dtsec_disable_interrupt().
-+ */
-+void fman_dtsec_enable_interrupt(struct dtsec_regs *regs, uint32_t ev_mask);
-+
-+/**
-+ * fman_dtsec_set_ts() - Enables dTSEC timestamps
-+ * @regs: Pointer to dTSEC register block
-+ * @en: true to enable timestamps, false to disable them
-+ *
-+ * Call this function to enable/disable dTSEC timestamps. This affects both
-+ * Tx and Rx.
-+ */
-+void fman_dtsec_set_ts(struct dtsec_regs *regs, bool en);
-+
-+/**
-+ * fman_dtsec_set_bucket() - Enables/disables a filter bucket
-+ * @regs: Pointer to dTSEC register block
-+ * @bucket: Bucket index
-+ * @enable: true/false to enable/disable this bucket
-+ *
-+ * This function enables or disables the specified bucket. Enabling a bucket
-+ * associated with an address configures dTSEC to accept received packets
-+ * with that destination address.
-+ * Multiple addresses may be associated with the same bucket. Disabling a
-+ * bucket will affect all addresses associated with that bucket. A bucket that
-+ * is enabled requires further filtering and verification in the upper layers
-+ *
-+ */
-+void fman_dtsec_set_bucket(struct dtsec_regs *regs, int bucket, bool enable);
-+
-+/**
-+ * dtsec_set_hash_table() - insert a crc code into thr filter table
-+ * @regs: Pointer to dTSEC register block
-+ * @crc: crc to insert
-+ * @mcast: true is this is a multicast address
-+ * @ghtx: true if we are in ghtx mode
-+ *
-+ * This function inserts a crc code into the filter table.
-+ */
-+void fman_dtsec_set_hash_table(struct dtsec_regs *regs, uint32_t crc,
-+ bool mcast, bool ghtx);
-+
-+/**
-+ * fman_dtsec_reset_filter_table() - Resets the address filtering table
-+ * @regs: Pointer to dTSEC register block
-+ * @mcast: Reset multicast entries
-+ * @ucast: Reset unicast entries
-+ *
-+ * Resets all entries in L2 address filter table. After calling this function
-+ * all buckets enabled using fman_dtsec_set_bucket() will be disabled.
-+ * If dtsec_init_filter_table() was called with @unicast_hash set to false,
-+ * @ucast argument is ignored.
-+ * This does not affect the primary nor the 15 additional addresses configured
-+ * using dtsec_set_address() or dtsec_set_match_address().
-+ */
-+void fman_dtsec_reset_filter_table(struct dtsec_regs *regs, bool mcast,
-+ bool ucast);
-+
-+/**
-+ * fman_dtsec_set_mc_promisc() - Set multicast promiscuous mode
-+ * @regs: Pointer to dTSEC register block
-+ * @enable: Enable multicast promiscuous mode
-+ *
-+ * Call this to enable/disable L2 address filtering for multicast packets.
-+ */
-+void fman_dtsec_set_mc_promisc(struct dtsec_regs *regs, bool enable);
-+
-+/* statistics APIs */
-+
-+/**
-+ * fman_dtsec_set_stat_level() - Enable a group of MIB statistics counters
-+ * @regs: Pointer to dTSEC register block
-+ * @level: Specifies a certain group of dTSEC MIB HW counters or _all_,
-+ * to specify all the existing counters.
-+ * If set to _none_, it disables all the counters.
-+ *
-+ * Enables the MIB statistics hw counters and sets up the carry interrupt
-+ * masks for the counters corresponding to the @level input parameter.
-+ *
-+ * Returns: error if invalid @level value given.
-+ */
-+int fman_dtsec_set_stat_level(struct dtsec_regs *regs,
-+ enum dtsec_stat_level level);
-+
-+/**
-+ * fman_dtsec_reset_stat() - Completely resets all dTSEC HW counters
-+ * @regs: Pointer to dTSEC register block
-+ */
-+void fman_dtsec_reset_stat(struct dtsec_regs *regs);
-+
-+/**
-+ * fman_dtsec_get_clear_carry_regs() - Read and clear carry bits (CAR1-2 registers)
-+ * @regs: Pointer to dTSEC register block
-+ * @car1: car1 register value
-+ * @car2: car2 register value
-+ *
-+ * When set, the carry bits signal that an overflow occurred on the
-+ * corresponding counters.
-+ * Note that the carry bits (CAR1-2 registers) will assert the
-+ * %DTSEC_IEVENT_MSRO interrupt if unmasked (via CAM1-2 regs).
-+ *
-+ * Returns: true if overflow occurred, otherwise - false
-+ */
-+bool fman_dtsec_get_clear_carry_regs(struct dtsec_regs *regs,
-+ uint32_t *car1, uint32_t *car2);
-+
-+uint32_t fman_dtsec_check_and_clear_tmr_event(struct dtsec_regs *regs);
-+
-+uint32_t fman_dtsec_get_stat_counter(struct dtsec_regs *regs,
-+ enum dtsec_stat_counters reg_name);
-+
-+void fman_dtsec_start_tx(struct dtsec_regs *regs);
-+void fman_dtsec_start_rx(struct dtsec_regs *regs);
-+void fman_dtsec_stop_tx(struct dtsec_regs *regs);
-+void fman_dtsec_stop_rx(struct dtsec_regs *regs);
-+uint32_t fman_dtsec_get_rctrl(struct dtsec_regs *regs);
-+
-+
-+#endif /* __FSL_FMAN_DTSEC_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/flib/fsl_fman_dtsec_mii_acc.h
-@@ -0,0 +1,107 @@
-+/*
-+ * Copyright 2008-2013 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#ifndef __FSL_FMAN_DTSEC_MII_ACC_H
-+#define __FSL_FMAN_DTSEC_MII_ACC_H
-+
-+#include "common/general.h"
-+
-+
-+/* MII Management Configuration Register */
-+#define MIIMCFG_RESET_MGMT 0x80000000
-+#define MIIMCFG_MGNTCLK_MASK 0x00000007
-+#define MIIMCFG_MGNTCLK_SHIFT 0
-+
-+/* MII Management Command Register */
-+#define MIIMCOM_SCAN_CYCLE 0x00000002
-+#define MIIMCOM_READ_CYCLE 0x00000001
-+
-+/* MII Management Address Register */
-+#define MIIMADD_PHY_ADDR_SHIFT 8
-+#define MIIMADD_PHY_ADDR_MASK 0x00001f00
-+
-+#define MIIMADD_REG_ADDR_SHIFT 0
-+#define MIIMADD_REG_ADDR_MASK 0x0000001f
-+
-+/* MII Management Indicator Register */
-+#define MIIMIND_BUSY 0x00000001
-+
-+
-+/* PHY Control Register */
-+#define PHY_CR_PHY_RESET 0x8000
-+#define PHY_CR_LOOPBACK 0x4000
-+#define PHY_CR_SPEED0 0x2000
-+#define PHY_CR_ANE 0x1000
-+#define PHY_CR_RESET_AN 0x0200
-+#define PHY_CR_FULLDUPLEX 0x0100
-+#define PHY_CR_SPEED1 0x0040
-+
-+#define PHY_TBICON_SRESET 0x8000
-+#define PHY_TBICON_SPEED2 0x0020
-+#define PHY_TBICON_CLK_SEL 0x0020
-+#define PHY_TBIANA_SGMII 0x4001
-+#define PHY_TBIANA_1000X 0x01a0
-+/* register map */
-+
-+/* MII Configuration Control Memory Map Registers */
-+struct dtsec_mii_reg {
-+ uint32_t reserved1[72];
-+ uint32_t miimcfg; /* MII Mgmt:configuration */
-+ uint32_t miimcom; /* MII Mgmt:command */
-+ uint32_t miimadd; /* MII Mgmt:address */
-+ uint32_t miimcon; /* MII Mgmt:control 3 */
-+ uint32_t miimstat; /* MII Mgmt:status */
-+ uint32_t miimind; /* MII Mgmt:indicators */
-+};
-+
-+/* dTSEC MII API */
-+
-+/* functions to access the mii registers for phy configuration.
-+ * this functionality may not be available for all dtsecs in the system.
-+ * consult the reference manual for details */
-+void fman_dtsec_mii_reset(struct dtsec_mii_reg *regs);
-+/* frequency is in MHz.
-+ * note that dtsec clock is 1/2 of fman clock */
-+void fman_dtsec_mii_init(struct dtsec_mii_reg *regs, uint16_t dtsec_freq);
-+int fman_dtsec_mii_write_reg(struct dtsec_mii_reg *regs,
-+ uint8_t addr,
-+ uint8_t reg,
-+ uint16_t data,
-+ uint16_t dtsec_freq);
-+
-+int fman_dtsec_mii_read_reg(struct dtsec_mii_reg *regs,
-+ uint8_t addr,
-+ uint8_t reg,
-+ uint16_t *data,
-+ uint16_t dtsec_freq);
-+
-+#endif /* __FSL_FMAN_DTSEC_MII_ACC_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/flib/fsl_fman_kg.h
-@@ -0,0 +1,514 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#ifndef __FSL_FMAN_KG_H
-+#define __FSL_FMAN_KG_H
-+
-+#include "common/general.h"
-+
-+#define FM_KG_NUM_OF_GENERIC_REGS 8 /**< Num of generic KeyGen regs */
-+#define FMAN_MAX_NUM_OF_HW_PORTS 64
-+/**< Total num of masks allowed on KG extractions */
-+#define FM_KG_EXTRACT_MASKS_NUM 4
-+#define FM_KG_NUM_CLS_PLAN_ENTR 8 /**< Num of class. plan regs */
-+#define FM_KG_CLS_PLAN_GRPS_NUM 32 /**< Max num of class. groups */
-+
-+struct fman_kg_regs {
-+ uint32_t fmkg_gcr;
-+ uint32_t res004;
-+ uint32_t res008;
-+ uint32_t fmkg_eer;
-+ uint32_t fmkg_eeer;
-+ uint32_t res014;
-+ uint32_t res018;
-+ uint32_t fmkg_seer;
-+ uint32_t fmkg_seeer;
-+ uint32_t fmkg_gsr;
-+ uint32_t fmkg_tpc;
-+ uint32_t fmkg_serc;
-+ uint32_t res030[4];
-+ uint32_t fmkg_fdor;
-+ uint32_t fmkg_gdv0r;
-+ uint32_t fmkg_gdv1r;
-+ uint32_t res04c[6];
-+ uint32_t fmkg_feer;
-+ uint32_t res068[38];
-+ uint32_t fmkg_indirect[63];
-+ uint32_t fmkg_ar;
-+};
-+
-+struct fman_kg_scheme_regs {
-+ uint32_t kgse_mode; /**< MODE */
-+ uint32_t kgse_ekfc; /**< Extract Known Fields Command */
-+ uint32_t kgse_ekdv; /**< Extract Known Default Value */
-+ uint32_t kgse_bmch; /**< Bit Mask Command High */
-+ uint32_t kgse_bmcl; /**< Bit Mask Command Low */
-+ uint32_t kgse_fqb; /**< Frame Queue Base */
-+ uint32_t kgse_hc; /**< Hash Command */
-+ uint32_t kgse_ppc; /**< Policer Profile Command */
-+ uint32_t kgse_gec[FM_KG_NUM_OF_GENERIC_REGS];
-+ /**< Generic Extract Command */
-+ uint32_t kgse_spc; /**< KeyGen Scheme Entry Statistic Packet Counter */
-+ uint32_t kgse_dv0; /**< KeyGen Scheme Entry Default Value 0 */
-+ uint32_t kgse_dv1; /**< KeyGen Scheme Entry Default Value 1 */
-+ uint32_t kgse_ccbs; /**< KeyGen Scheme Entry Coarse Classification Bit*/
-+ uint32_t kgse_mv; /**< KeyGen Scheme Entry Match vector */
-+ uint32_t kgse_om; /**< KeyGen Scheme Entry Operation Mode bits */
-+ uint32_t kgse_vsp; /**< KeyGen Scheme Entry Virtual Storage Profile */
-+};
-+
-+struct fman_kg_pe_regs{
-+ uint32_t fmkg_pe_sp;
-+ uint32_t fmkg_pe_cpp;
-+};
-+
-+struct fman_kg_cp_regs {
-+ uint32_t kgcpe[FM_KG_NUM_CLS_PLAN_ENTR];
-+};
-+
-+
-+#define FM_KG_KGAR_GO 0x80000000
-+#define FM_KG_KGAR_READ 0x40000000
-+#define FM_KG_KGAR_WRITE 0x00000000
-+#define FM_KG_KGAR_SEL_SCHEME_ENTRY 0x00000000
-+#define FM_KG_KGAR_SCM_WSEL_UPDATE_CNT 0x00008000
-+
-+#define KG_SCH_PP_SHIFT_HIGH 0x80000000
-+#define KG_SCH_PP_NO_GEN 0x10000000
-+#define KG_SCH_PP_SHIFT_LOW 0x0000F000
-+#define KG_SCH_MODE_NIA_PLCR 0x40000000
-+#define KG_SCH_GEN_EXTRACT_TYPE 0x00008000
-+#define KG_SCH_BITMASK_MASK 0x000000FF
-+#define KG_SCH_GEN_VALID 0x80000000
-+#define KG_SCH_GEN_MASK 0x00FF0000
-+#define FM_PCD_KG_KGAR_ERR 0x20000000
-+#define FM_PCD_KG_KGAR_SEL_CLS_PLAN_ENTRY 0x01000000
-+#define FM_PCD_KG_KGAR_SEL_PORT_ENTRY 0x02000000
-+#define FM_PCD_KG_KGAR_SEL_PORT_WSEL_SP 0x00008000
-+#define FM_PCD_KG_KGAR_SEL_PORT_WSEL_CPP 0x00004000
-+#define FM_PCD_KG_KGAR_WSEL_MASK 0x0000FF00
-+#define KG_SCH_HASH_CONFIG_NO_FQID 0x80000000
-+#define KG_SCH_HASH_CONFIG_SYM 0x40000000
-+
-+#define FM_EX_KG_DOUBLE_ECC 0x80000000
-+#define FM_EX_KG_KEYSIZE_OVERFLOW 0x40000000
-+
-+/* ECC capture register */
-+#define KG_FMKG_SERC_CAP 0x80000000
-+#define KG_FMKG_SERC_CET 0x40000000
-+#define KG_FMKG_SERC_CNT_MSK 0x00FF0000
-+#define KG_FMKG_SERC_CNT_SHIFT 16
-+#define KG_FMKG_SERC_ADDR_MSK 0x000003FF
-+
-+/* Masks */
-+#define FM_KG_KGGCR_EN 0x80000000
-+#define KG_SCH_GEN_VALID 0x80000000
-+#define KG_SCH_GEN_EXTRACT_TYPE 0x00008000
-+#define KG_ERR_TYPE_DOUBLE 0x40000000
-+#define KG_ERR_ADDR_MASK 0x00000FFF
-+#define KG_SCH_MODE_EN 0x80000000
-+
-+/* shifts */
-+#define FM_KG_KGAR_NUM_SHIFT 16
-+#define FM_KG_PE_CPP_MASK_SHIFT 16
-+#define FM_KG_KGAR_WSEL_SHIFT 8
-+
-+#define FM_KG_SCH_GEN_HT_INVALID 0
-+
-+#define FM_KG_MASK_SEL_GEN_BASE 0x20
-+
-+#define KG_GET_MASK_SEL_SHIFT(shift, i) \
-+switch (i) \
-+{ \
-+ case 0: (shift) = 26; break; \
-+ case 1: (shift) = 20; break; \
-+ case 2: (shift) = 10; break; \
-+ case 3: (shift) = 4; break; \
-+ default: (shift) = 0; \
-+}
-+
-+#define KG_GET_MASK_OFFSET_SHIFT(shift, i) \
-+switch (i) \
-+{ \
-+ case 0: (shift) = 16; break; \
-+ case 1: (shift) = 0; break; \
-+ case 2: (shift) = 28; break; \
-+ case 3: (shift) = 24; break; \
-+ default: (shift) = 0; \
-+}
-+
-+#define KG_GET_MASK_SHIFT(shift, i) \
-+switch (i) \
-+{ \
-+ case 0: shift = 24; break; \
-+ case 1: shift = 16; break; \
-+ case 2: shift = 8; break; \
-+ case 3: shift = 0; break; \
-+ default: shift = 0; \
-+}
-+
-+/* Port entry CPP register */
-+#define FMAN_KG_PE_CPP_MASK_SHIFT 16
-+
-+/* Scheme registers */
-+#define FMAN_KG_SCH_MODE_EN 0x80000000
-+#define FMAN_KG_SCH_MODE_NIA_PLCR 0x40000000
-+#define FMAN_KG_SCH_MODE_CCOBASE_SHIFT 24
-+
-+#define FMAN_KG_SCH_DEF_MAC_ADDR_SHIFT 30
-+#define FMAN_KG_SCH_DEF_VLAN_TCI_SHIFT 28
-+#define FMAN_KG_SCH_DEF_ETYPE_SHIFT 26
-+#define FMAN_KG_SCH_DEF_PPP_SID_SHIFT 24
-+#define FMAN_KG_SCH_DEF_PPP_PID_SHIFT 22
-+#define FMAN_KG_SCH_DEF_MPLS_SHIFT 20
-+#define FMAN_KG_SCH_DEF_IP_ADDR_SHIFT 18
-+#define FMAN_KG_SCH_DEF_PTYPE_SHIFT 16
-+#define FMAN_KG_SCH_DEF_IP_TOS_TC_SHIFT 14
-+#define FMAN_KG_SCH_DEF_IPv6_FL_SHIFT 12
-+#define FMAN_KG_SCH_DEF_IPSEC_SPI_SHIFT 10
-+#define FMAN_KG_SCH_DEF_L4_PORT_SHIFT 8
-+#define FMAN_KG_SCH_DEF_TCP_FLG_SHIFT 6
-+
-+#define FMAN_KG_SCH_GEN_VALID 0x80000000
-+#define FMAN_KG_SCH_GEN_SIZE_MAX 16
-+#define FMAN_KG_SCH_GEN_OR 0x00008000
-+
-+#define FMAN_KG_SCH_GEN_DEF_SHIFT 29
-+#define FMAN_KG_SCH_GEN_SIZE_SHIFT 24
-+#define FMAN_KG_SCH_GEN_MASK_SHIFT 16
-+#define FMAN_KG_SCH_GEN_HT_SHIFT 8
-+
-+#define FMAN_KG_SCH_HASH_HSHIFT_SHIFT 24
-+#define FMAN_KG_SCH_HASH_HSHIFT_MAX 0x28
-+#define FMAN_KG_SCH_HASH_SYM 0x40000000
-+#define FMAN_KG_SCH_HASH_NO_FQID_GEN 0x80000000
-+
-+#define FMAN_KG_SCH_PP_SH_SHIFT 27
-+#define FMAN_KG_SCH_PP_SL_SHIFT 12
-+#define FMAN_KG_SCH_PP_SH_MASK 0x80000000
-+#define FMAN_KG_SCH_PP_SL_MASK 0x0000F000
-+#define FMAN_KG_SCH_PP_SHIFT_MAX 0x17
-+#define FMAN_KG_SCH_PP_MASK_SHIFT 16
-+#define FMAN_KG_SCH_PP_NO_GEN 0x10000000
-+
-+enum fman_kg_gen_extract_src {
-+ E_FMAN_KG_GEN_EXTRACT_ETH,
-+ E_FMAN_KG_GEN_EXTRACT_ETYPE,
-+ E_FMAN_KG_GEN_EXTRACT_SNAP,
-+ E_FMAN_KG_GEN_EXTRACT_VLAN_TCI_1,
-+ E_FMAN_KG_GEN_EXTRACT_VLAN_TCI_N,
-+ E_FMAN_KG_GEN_EXTRACT_PPPoE,
-+ E_FMAN_KG_GEN_EXTRACT_MPLS_1,
-+ E_FMAN_KG_GEN_EXTRACT_MPLS_2,
-+ E_FMAN_KG_GEN_EXTRACT_MPLS_3,
-+ E_FMAN_KG_GEN_EXTRACT_MPLS_N,
-+ E_FMAN_KG_GEN_EXTRACT_IPv4_1,
-+ E_FMAN_KG_GEN_EXTRACT_IPv6_1,
-+ E_FMAN_KG_GEN_EXTRACT_IPv4_2,
-+ E_FMAN_KG_GEN_EXTRACT_IPv6_2,
-+ E_FMAN_KG_GEN_EXTRACT_MINENCAP,
-+ E_FMAN_KG_GEN_EXTRACT_IP_PID,
-+ E_FMAN_KG_GEN_EXTRACT_GRE,
-+ E_FMAN_KG_GEN_EXTRACT_TCP,
-+ E_FMAN_KG_GEN_EXTRACT_UDP,
-+ E_FMAN_KG_GEN_EXTRACT_SCTP,
-+ E_FMAN_KG_GEN_EXTRACT_DCCP,
-+ E_FMAN_KG_GEN_EXTRACT_IPSEC_AH,
-+ E_FMAN_KG_GEN_EXTRACT_IPSEC_ESP,
-+ E_FMAN_KG_GEN_EXTRACT_SHIM_1,
-+ E_FMAN_KG_GEN_EXTRACT_SHIM_2,
-+ E_FMAN_KG_GEN_EXTRACT_FROM_DFLT,
-+ E_FMAN_KG_GEN_EXTRACT_FROM_FRAME_START,
-+ E_FMAN_KG_GEN_EXTRACT_FROM_PARSE_RESULT,
-+ E_FMAN_KG_GEN_EXTRACT_FROM_END_OF_PARSE,
-+ E_FMAN_KG_GEN_EXTRACT_FROM_FQID
-+};
-+
-+struct fman_kg_ex_ecc_attr
-+{
-+ bool valid;
-+ bool double_ecc;
-+ uint16_t addr;
-+ uint8_t single_ecc_count;
-+};
-+
-+enum fman_kg_def_select
-+{
-+ E_FMAN_KG_DEF_GLOBAL_0,
-+ E_FMAN_KG_DEF_GLOBAL_1,
-+ E_FMAN_KG_DEF_SCHEME_0,
-+ E_FMAN_KG_DEF_SCHEME_1
-+};
-+
-+struct fman_kg_extract_def
-+{
-+ enum fman_kg_def_select mac_addr;
-+ enum fman_kg_def_select vlan_tci;
-+ enum fman_kg_def_select etype;
-+ enum fman_kg_def_select ppp_sid;
-+ enum fman_kg_def_select ppp_pid;
-+ enum fman_kg_def_select mpls;
-+ enum fman_kg_def_select ip_addr;
-+ enum fman_kg_def_select ptype;
-+ enum fman_kg_def_select ip_tos_tc;
-+ enum fman_kg_def_select ipv6_fl;
-+ enum fman_kg_def_select ipsec_spi;
-+ enum fman_kg_def_select l4_port;
-+ enum fman_kg_def_select tcp_flg;
-+};
-+
-+enum fman_kg_gen_extract_type
-+{
-+ E_FMAN_KG_HASH_EXTRACT,
-+ E_FMAN_KG_OR_EXTRACT
-+};
-+
-+struct fman_kg_gen_extract_params
-+{
-+ /* Hash or Or-ed extract */
-+ enum fman_kg_gen_extract_type type;
-+ enum fman_kg_gen_extract_src src;
-+ bool no_validation;
-+ /* Extraction offset from the header location specified above */
-+ uint8_t offset;
-+ /* Size of extraction for FMAN_KG_HASH_EXTRACT,
-+ * hash result shift for FMAN_KG_OR_EXTRACT */
-+ uint8_t extract;
-+ uint8_t mask;
-+ /* Default value to use when header specified
-+ * by fman_kg_gen_extract_src doesn't present */
-+ enum fman_kg_def_select def_val;
-+};
-+
-+struct fman_kg_extract_mask
-+{
-+ /**< Indication if mask is on known field extraction or
-+ * on general extraction; TRUE for known field */
-+ bool is_known;
-+ /**< One of FMAN_KG_EXTRACT_xxx defines for known fields mask and
-+ * generic register index for generic extracts mask */
-+ uint32_t field_or_gen_idx;
-+ /**< Byte offset from start of the extracted data specified
-+ * by field_or_gen_idx */
-+ uint8_t offset;
-+ /**< Byte mask (selected bits will be used) */
-+ uint8_t mask;
-+};
-+
-+struct fman_kg_extract_params
-+{
-+ /* Or-ed mask of FMAN_KG_EXTRACT_xxx defines */
-+ uint32_t known_fields;
-+ struct fman_kg_extract_def known_fields_def;
-+ /* Number of entries in gen_extract */
-+ uint8_t gen_extract_num;
-+ struct fman_kg_gen_extract_params gen_extract[FM_KG_NUM_OF_GENERIC_REGS];
-+ /* Number of entries in masks */
-+ uint8_t masks_num;
-+ struct fman_kg_extract_mask masks[FM_KG_EXTRACT_MASKS_NUM];
-+ uint32_t def_scheme_0;
-+ uint32_t def_scheme_1;
-+};
-+
-+struct fman_kg_hash_params
-+{
-+ bool use_hash;
-+ uint8_t shift_r;
-+ uint32_t mask; /**< 24-bit mask */
-+ bool sym; /**< Symmetric hash for src and dest pairs */
-+};
-+
-+struct fman_kg_pp_params
-+{
-+ uint8_t base;
-+ uint8_t shift;
-+ uint8_t mask;
-+ bool bypass_pp_gen;
-+};
-+
-+struct fman_kg_cc_params
-+{
-+ uint8_t base_offset;
-+ uint32_t qlcv_bits_sel;
-+};
-+
-+enum fman_pcd_engine
-+{
-+ E_FMAN_PCD_INVALID = 0, /**< Invalid PCD engine indicated*/
-+ E_FMAN_PCD_DONE, /**< No PCD Engine indicated */
-+ E_FMAN_PCD_KG, /**< Keygen indicated */
-+ E_FMAN_PCD_CC, /**< Coarse classification indicated */
-+ E_FMAN_PCD_PLCR, /**< Policer indicated */
-+ E_FMAN_PCD_PRS /**< Parser indicated */
-+};
-+
-+struct fman_kg_cls_plan_params
-+{
-+ uint8_t entries_mask;
-+ uint32_t mask_vector[FM_KG_NUM_CLS_PLAN_ENTR];
-+};
-+
-+struct fman_kg_scheme_params
-+{
-+ uint32_t match_vector;
-+ struct fman_kg_extract_params extract_params;
-+ struct fman_kg_hash_params hash_params;
-+ uint32_t base_fqid;
-+ /* What we do w/features supported per FM version ?? */
-+ bool bypass_fqid_gen;
-+ struct fman_kg_pp_params policer_params;
-+ struct fman_kg_cc_params cc_params;
-+ bool update_counter;
-+ /**< counter_value: Set scheme counter to the specified value;
-+ * relevant only when update_counter = TRUE. */
-+ uint32_t counter_value;
-+ enum fman_pcd_engine next_engine;
-+ /**< Next engine action code */
-+ uint32_t next_engine_action;
-+};
-+
-+
-+
-+int fman_kg_write_ar_wait(struct fman_kg_regs *regs, uint32_t fmkg_ar);
-+void fman_kg_write_sp(struct fman_kg_regs *regs, uint32_t sp, bool add);
-+void fman_kg_write_cpp(struct fman_kg_regs *regs, uint32_t cpp);
-+void fman_kg_get_event(struct fman_kg_regs *regs,
-+ uint32_t *event,
-+ uint32_t *scheme_idx);
-+void fman_kg_init(struct fman_kg_regs *regs,
-+ uint32_t exceptions,
-+ uint32_t dflt_nia);
-+void fman_kg_enable_scheme_interrupts(struct fman_kg_regs *regs);
-+void fman_kg_enable(struct fman_kg_regs *regs);
-+void fman_kg_disable(struct fman_kg_regs *regs);
-+int fman_kg_write_bind_cls_plans(struct fman_kg_regs *regs,
-+ uint8_t hwport_id,
-+ uint32_t bind_cls_plans);
-+int fman_kg_build_bind_cls_plans(uint8_t grp_base,
-+ uint8_t grp_mask,
-+ uint32_t *bind_cls_plans);
-+int fman_kg_write_bind_schemes(struct fman_kg_regs *regs,
-+ uint8_t hwport_id,
-+ uint32_t schemes);
-+int fman_kg_write_cls_plan(struct fman_kg_regs *regs,
-+ uint8_t grp_id,
-+ uint8_t entries_mask,
-+ uint8_t hwport_id,
-+ struct fman_kg_cp_regs *cls_plan_regs);
-+int fman_kg_build_cls_plan(struct fman_kg_cls_plan_params *params,
-+ struct fman_kg_cp_regs *cls_plan_regs);
-+uint32_t fman_kg_get_schemes_total_counter(struct fman_kg_regs *regs);
-+int fman_kg_set_scheme_counter(struct fman_kg_regs *regs,
-+ uint8_t scheme_id,
-+ uint8_t hwport_id,
-+ uint32_t counter);
-+int fman_kg_get_scheme_counter(struct fman_kg_regs *regs,
-+ uint8_t scheme_id,
-+ uint8_t hwport_id,
-+ uint32_t *counter);
-+int fman_kg_delete_scheme(struct fman_kg_regs *regs,
-+ uint8_t scheme_id,
-+ uint8_t hwport_id);
-+int fman_kg_write_scheme(struct fman_kg_regs *regs,
-+ uint8_t scheme_id,
-+ uint8_t hwport_id,
-+ struct fman_kg_scheme_regs *scheme_regs,
-+ bool update_counter);
-+int fman_kg_build_scheme(struct fman_kg_scheme_params *params,
-+ struct fman_kg_scheme_regs *scheme_regs);
-+void fman_kg_get_capture(struct fman_kg_regs *regs,
-+ struct fman_kg_ex_ecc_attr *ecc_attr,
-+ bool clear);
-+void fman_kg_get_exception(struct fman_kg_regs *regs,
-+ uint32_t *events,
-+ uint32_t *scheme_ids,
-+ bool clear);
-+void fman_kg_set_exception(struct fman_kg_regs *regs,
-+ uint32_t exception,
-+ bool enable);
-+void fman_kg_set_dflt_val(struct fman_kg_regs *regs,
-+ uint8_t def_id,
-+ uint32_t val);
-+void fman_kg_set_data_after_prs(struct fman_kg_regs *regs, uint8_t offset);
-+
-+
-+
-+/**************************************************************************//**
-+ @Description NIA Description
-+*//***************************************************************************/
-+#define KG_NIA_ORDER_RESTOR 0x00800000
-+#define KG_NIA_ENG_FM_CTL 0x00000000
-+#define KG_NIA_ENG_PRS 0x00440000
-+#define KG_NIA_ENG_KG 0x00480000
-+#define KG_NIA_ENG_PLCR 0x004C0000
-+#define KG_NIA_ENG_BMI 0x00500000
-+#define KG_NIA_ENG_QMI_ENQ 0x00540000
-+#define KG_NIA_ENG_QMI_DEQ 0x00580000
-+#define KG_NIA_ENG_MASK 0x007C0000
-+
-+#define KG_NIA_AC_MASK 0x0003FFFF
-+
-+#define KG_NIA_INVALID 0xFFFFFFFF
-+
-+static __inline__ uint32_t fm_kg_build_nia(enum fman_pcd_engine next_engine,
-+ uint32_t next_engine_action)
-+{
-+ uint32_t nia;
-+
-+ if (next_engine_action & ~KG_NIA_AC_MASK)
-+ return KG_NIA_INVALID;
-+
-+ switch (next_engine) {
-+ case E_FMAN_PCD_DONE:
-+ nia = KG_NIA_ENG_BMI | next_engine_action;
-+ break;
-+
-+ case E_FMAN_PCD_KG:
-+ nia = KG_NIA_ENG_KG | next_engine_action;
-+ break;
-+
-+ case E_FMAN_PCD_CC:
-+ nia = KG_NIA_ENG_FM_CTL | next_engine_action;
-+ break;
-+
-+ case E_FMAN_PCD_PLCR:
-+ nia = KG_NIA_ENG_PLCR | next_engine_action;
-+ break;
-+
-+ default:
-+ nia = KG_NIA_INVALID;
-+ }
-+
-+ return nia;
-+}
-+
-+#endif /* __FSL_FMAN_KG_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/flib/fsl_fman_memac.h
-@@ -0,0 +1,434 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+#ifndef __FSL_FMAN_MEMAC_H
-+#define __FSL_FMAN_MEMAC_H
-+
-+#include "common/general.h"
-+#include "fsl_enet.h"
-+
-+
-+#define MEMAC_NUM_OF_PADDRS 7 /* Num of additional exact match MAC adr regs */
-+
-+/* Control and Configuration Register (COMMAND_CONFIG) */
-+#define CMD_CFG_MG 0x80000000 /* 00 Magic Packet detection */
-+#define CMD_CFG_REG_LOWP_RXETY 0x01000000 /* 07 Rx low power indication */
-+#define CMD_CFG_TX_LOWP_ENA 0x00800000 /* 08 Tx Low Power Idle Enable */
-+#define CMD_CFG_SFD_ANY 0x00200000 /* 10 Disable SFD check */
-+#define CMD_CFG_PFC_MODE 0x00080000 /* 12 Enable PFC */
-+#define CMD_CFG_NO_LEN_CHK 0x00020000 /* 14 Payload length check disable */
-+#define CMD_CFG_SEND_IDLE 0x00010000 /* 15 Force idle generation */
-+#define CMD_CFG_CNT_FRM_EN 0x00002000 /* 18 Control frame rx enable */
-+#define CMD_CFG_SW_RESET 0x00001000 /* 19 S/W Reset, self clearing bit */
-+#define CMD_CFG_TX_PAD_EN 0x00000800 /* 20 Enable Tx padding of frames */
-+#define CMD_CFG_LOOPBACK_EN 0x00000400 /* 21 XGMII/GMII loopback enable */
-+#define CMD_CFG_TX_ADDR_INS 0x00000200 /* 22 Tx source MAC addr insertion */
-+#define CMD_CFG_PAUSE_IGNORE 0x00000100 /* 23 Ignore Pause frame quanta */
-+#define CMD_CFG_PAUSE_FWD 0x00000080 /* 24 Terminate/frwd Pause frames */
-+#define CMD_CFG_CRC_FWD 0x00000040 /* 25 Terminate/frwd CRC of frames */
-+#define CMD_CFG_PAD_EN 0x00000020 /* 26 Frame padding removal */
-+#define CMD_CFG_PROMIS_EN 0x00000010 /* 27 Promiscuous operation enable */
-+#define CMD_CFG_WAN_MODE 0x00000008 /* 28 WAN mode enable */
-+#define CMD_CFG_RX_EN 0x00000002 /* 30 MAC receive path enable */
-+#define CMD_CFG_TX_EN 0x00000001 /* 31 MAC transmit path enable */
-+
-+/* Transmit FIFO Sections Register (TX_FIFO_SECTIONS) */
-+#define TX_FIFO_SECTIONS_TX_EMPTY_MASK 0xFFFF0000
-+#define TX_FIFO_SECTIONS_TX_AVAIL_MASK 0x0000FFFF
-+#define TX_FIFO_SECTIONS_TX_EMPTY_DEFAULT_10G 0x00400000
-+#define TX_FIFO_SECTIONS_TX_EMPTY_DEFAULT_1G 0x00100000
-+#define TX_FIFO_SECTIONS_TX_EMPTY_PFC_10G 0x00360000
-+#define TX_FIFO_SECTIONS_TX_EMPTY_PFC_1G 0x00040000
-+#define TX_FIFO_SECTIONS_TX_AVAIL_10G 0x00000019
-+#define TX_FIFO_SECTIONS_TX_AVAIL_1G 0x00000020
-+#define TX_FIFO_SECTIONS_TX_AVAIL_SLOW_10G 0x00000060
-+
-+#define GET_TX_EMPTY_DEFAULT_VALUE(_val) \
-+_val &= ~TX_FIFO_SECTIONS_TX_EMPTY_MASK; \
-+((_val == TX_FIFO_SECTIONS_TX_AVAIL_10G) ? \
-+ (_val |= TX_FIFO_SECTIONS_TX_EMPTY_DEFAULT_10G) : \
-+ (_val |= TX_FIFO_SECTIONS_TX_EMPTY_DEFAULT_1G));
-+
-+#define GET_TX_EMPTY_PFC_VALUE(_val) \
-+_val &= ~TX_FIFO_SECTIONS_TX_EMPTY_MASK; \
-+((_val == TX_FIFO_SECTIONS_TX_AVAIL_10G) ? \
-+ (_val |= TX_FIFO_SECTIONS_TX_EMPTY_PFC_10G) : \
-+ (_val |= TX_FIFO_SECTIONS_TX_EMPTY_PFC_1G));
-+
-+/* Interface Mode Register (IF_MODE) */
-+#define IF_MODE_MASK 0x00000003 /* 30-31 Mask on i/f mode bits */
-+#define IF_MODE_XGMII 0x00000000 /* 30-31 XGMII (10G) interface */
-+#define IF_MODE_GMII 0x00000002 /* 30-31 GMII (1G) interface */
-+#define IF_MODE_RGMII 0x00000004
-+#define IF_MODE_RGMII_AUTO 0x00008000
-+#define IF_MODE_RGMII_1000 0x00004000 /* 10 - 1000Mbps RGMII */
-+#define IF_MODE_RGMII_100 0x00000000 /* 00 - 100Mbps RGMII */
-+#define IF_MODE_RGMII_10 0x00002000 /* 01 - 10Mbps RGMII */
-+#define IF_MODE_RGMII_SP_MASK 0x00006000 /* Setsp mask bits */
-+#define IF_MODE_RGMII_FD 0x00001000 /* Full duplex RGMII */
-+#define IF_MODE_HD 0x00000040 /* Half duplex operation */
-+
-+/* Hash table Control Register (HASHTABLE_CTRL) */
-+#define HASH_CTRL_MCAST_SHIFT 26
-+#define HASH_CTRL_MCAST_EN 0x00000100 /* 23 Mcast frame rx for hash */
-+#define HASH_CTRL_ADDR_MASK 0x0000003F /* 26-31 Hash table address code */
-+
-+#define GROUP_ADDRESS 0x0000010000000000LL /* MAC mcast indication */
-+#define HASH_TABLE_SIZE 64 /* Hash tbl size */
-+
-+/* Transmit Inter-Packet Gap Length Register (TX_IPG_LENGTH) */
-+#define MEMAC_TX_IPG_LENGTH_MASK 0x0000003F
-+
-+/* Statistics Configuration Register (STATN_CONFIG) */
-+#define STATS_CFG_CLR 0x00000004 /* 29 Reset all counters */
-+#define STATS_CFG_CLR_ON_RD 0x00000002 /* 30 Clear on read */
-+#define STATS_CFG_SATURATE 0x00000001 /* 31 Saturate at the maximum val */
-+
-+/* Interrupt Mask Register (IMASK) */
-+#define MEMAC_IMASK_MGI 0x40000000 /* 1 Magic pkt detect indication */
-+#define MEMAC_IMASK_TSECC_ER 0x20000000 /* 2 Timestamp FIFO ECC error evnt */
-+#define MEMAC_IMASK_TECC_ER 0x02000000 /* 6 Transmit frame ECC error evnt */
-+#define MEMAC_IMASK_RECC_ER 0x01000000 /* 7 Receive frame ECC error evnt */
-+
-+#define MEMAC_ALL_ERRS_IMASK \
-+ ((uint32_t)(MEMAC_IMASK_TSECC_ER | \
-+ MEMAC_IMASK_TECC_ER | \
-+ MEMAC_IMASK_RECC_ER | \
-+ MEMAC_IMASK_MGI))
-+
-+#define MEMAC_IEVNT_PCS 0x80000000 /* PCS (XG). Link sync (G) */
-+#define MEMAC_IEVNT_AN 0x40000000 /* Auto-negotiation */
-+#define MEMAC_IEVNT_LT 0x20000000 /* Link Training/New page */
-+#define MEMAC_IEVNT_MGI 0x00004000 /* Magic pkt detection */
-+#define MEMAC_IEVNT_TS_ECC_ER 0x00002000 /* Timestamp FIFO ECC error */
-+#define MEMAC_IEVNT_RX_FIFO_OVFL 0x00001000 /* Rx FIFO overflow */
-+#define MEMAC_IEVNT_TX_FIFO_UNFL 0x00000800 /* Tx FIFO underflow */
-+#define MEMAC_IEVNT_TX_FIFO_OVFL 0x00000400 /* Tx FIFO overflow */
-+#define MEMAC_IEVNT_TX_ECC_ER 0x00000200 /* Tx frame ECC error */
-+#define MEMAC_IEVNT_RX_ECC_ER 0x00000100 /* Rx frame ECC error */
-+#define MEMAC_IEVNT_LI_FAULT 0x00000080 /* Link Interruption flt */
-+#define MEMAC_IEVNT_RX_EMPTY 0x00000040 /* Rx FIFO empty */
-+#define MEMAC_IEVNT_TX_EMPTY 0x00000020 /* Tx FIFO empty */
-+#define MEMAC_IEVNT_RX_LOWP 0x00000010 /* Low Power Idle */
-+#define MEMAC_IEVNT_PHY_LOS 0x00000004 /* Phy loss of signal */
-+#define MEMAC_IEVNT_REM_FAULT 0x00000002 /* Remote fault (XGMII) */
-+#define MEMAC_IEVNT_LOC_FAULT 0x00000001 /* Local fault (XGMII) */
-+
-+enum memac_counters {
-+ E_MEMAC_COUNTER_R64,
-+ E_MEMAC_COUNTER_T64,
-+ E_MEMAC_COUNTER_R127,
-+ E_MEMAC_COUNTER_T127,
-+ E_MEMAC_COUNTER_R255,
-+ E_MEMAC_COUNTER_T255,
-+ E_MEMAC_COUNTER_R511,
-+ E_MEMAC_COUNTER_T511,
-+ E_MEMAC_COUNTER_R1023,
-+ E_MEMAC_COUNTER_T1023,
-+ E_MEMAC_COUNTER_R1518,
-+ E_MEMAC_COUNTER_T1518,
-+ E_MEMAC_COUNTER_R1519X,
-+ E_MEMAC_COUNTER_T1519X,
-+ E_MEMAC_COUNTER_RFRG,
-+ E_MEMAC_COUNTER_RJBR,
-+ E_MEMAC_COUNTER_RDRP,
-+ E_MEMAC_COUNTER_RALN,
-+ E_MEMAC_COUNTER_TUND,
-+ E_MEMAC_COUNTER_ROVR,
-+ E_MEMAC_COUNTER_RXPF,
-+ E_MEMAC_COUNTER_TXPF,
-+ E_MEMAC_COUNTER_ROCT,
-+ E_MEMAC_COUNTER_RMCA,
-+ E_MEMAC_COUNTER_RBCA,
-+ E_MEMAC_COUNTER_RPKT,
-+ E_MEMAC_COUNTER_RUCA,
-+ E_MEMAC_COUNTER_RERR,
-+ E_MEMAC_COUNTER_TOCT,
-+ E_MEMAC_COUNTER_TMCA,
-+ E_MEMAC_COUNTER_TBCA,
-+ E_MEMAC_COUNTER_TUCA,
-+ E_MEMAC_COUNTER_TERR
-+};
-+
-+#define DEFAULT_PAUSE_QUANTA 0xf000
-+#define DEFAULT_FRAME_LENGTH 0x600
-+#define DEFAULT_TX_IPG_LENGTH 12
-+
-+/*
-+ * memory map
-+ */
-+
-+struct mac_addr {
-+ uint32_t mac_addr_l; /* Lower 32 bits of 48-bit MAC address */
-+ uint32_t mac_addr_u; /* Upper 16 bits of 48-bit MAC address */
-+};
-+
-+struct memac_regs {
-+ /* General Control and Status */
-+ uint32_t res0000[2];
-+ uint32_t command_config; /* 0x008 Ctrl and cfg */
-+ struct mac_addr mac_addr0; /* 0x00C-0x010 MAC_ADDR_0...1 */
-+ uint32_t maxfrm; /* 0x014 Max frame length */
-+ uint32_t res0018[1];
-+ uint32_t rx_fifo_sections; /* Receive FIFO configuration reg */
-+ uint32_t tx_fifo_sections; /* Transmit FIFO configuration reg */
-+ uint32_t res0024[2];
-+ uint32_t hashtable_ctrl; /* 0x02C Hash table control */
-+ uint32_t res0030[4];
-+ uint32_t ievent; /* 0x040 Interrupt event */
-+ uint32_t tx_ipg_length; /* 0x044 Transmitter inter-packet-gap */
-+ uint32_t res0048;
-+ uint32_t imask; /* 0x04C Interrupt mask */
-+ uint32_t res0050;
-+ uint32_t pause_quanta[4]; /* 0x054 Pause quanta */
-+ uint32_t pause_thresh[4]; /* 0x064 Pause quanta threshold */
-+ uint32_t rx_pause_status; /* 0x074 Receive pause status */
-+ uint32_t res0078[2];
-+ struct mac_addr mac_addr[MEMAC_NUM_OF_PADDRS]; /* 0x80-0x0B4 mac padr */
-+ uint32_t lpwake_timer; /* 0x0B8 Low Power Wakeup Timer */
-+ uint32_t sleep_timer; /* 0x0BC Transmit EEE Low Power Timer */
-+ uint32_t res00c0[8];
-+ uint32_t statn_config; /* 0x0E0 Statistics configuration */
-+ uint32_t res00e4[7];
-+ /* Rx Statistics Counter */
-+ uint32_t reoct_l;
-+ uint32_t reoct_u;
-+ uint32_t roct_l;
-+ uint32_t roct_u;
-+ uint32_t raln_l;
-+ uint32_t raln_u;
-+ uint32_t rxpf_l;
-+ uint32_t rxpf_u;
-+ uint32_t rfrm_l;
-+ uint32_t rfrm_u;
-+ uint32_t rfcs_l;
-+ uint32_t rfcs_u;
-+ uint32_t rvlan_l;
-+ uint32_t rvlan_u;
-+ uint32_t rerr_l;
-+ uint32_t rerr_u;
-+ uint32_t ruca_l;
-+ uint32_t ruca_u;
-+ uint32_t rmca_l;
-+ uint32_t rmca_u;
-+ uint32_t rbca_l;
-+ uint32_t rbca_u;
-+ uint32_t rdrp_l;
-+ uint32_t rdrp_u;
-+ uint32_t rpkt_l;
-+ uint32_t rpkt_u;
-+ uint32_t rund_l;
-+ uint32_t rund_u;
-+ uint32_t r64_l;
-+ uint32_t r64_u;
-+ uint32_t r127_l;
-+ uint32_t r127_u;
-+ uint32_t r255_l;
-+ uint32_t r255_u;
-+ uint32_t r511_l;
-+ uint32_t r511_u;
-+ uint32_t r1023_l;
-+ uint32_t r1023_u;
-+ uint32_t r1518_l;
-+ uint32_t r1518_u;
-+ uint32_t r1519x_l;
-+ uint32_t r1519x_u;
-+ uint32_t rovr_l;
-+ uint32_t rovr_u;
-+ uint32_t rjbr_l;
-+ uint32_t rjbr_u;
-+ uint32_t rfrg_l;
-+ uint32_t rfrg_u;
-+ uint32_t rcnp_l;
-+ uint32_t rcnp_u;
-+ uint32_t rdrntp_l;
-+ uint32_t rdrntp_u;
-+ uint32_t res01d0[12];
-+ /* Tx Statistics Counter */
-+ uint32_t teoct_l;
-+ uint32_t teoct_u;
-+ uint32_t toct_l;
-+ uint32_t toct_u;
-+ uint32_t res0210[2];
-+ uint32_t txpf_l;
-+ uint32_t txpf_u;
-+ uint32_t tfrm_l;
-+ uint32_t tfrm_u;
-+ uint32_t tfcs_l;
-+ uint32_t tfcs_u;
-+ uint32_t tvlan_l;
-+ uint32_t tvlan_u;
-+ uint32_t terr_l;
-+ uint32_t terr_u;
-+ uint32_t tuca_l;
-+ uint32_t tuca_u;
-+ uint32_t tmca_l;
-+ uint32_t tmca_u;
-+ uint32_t tbca_l;
-+ uint32_t tbca_u;
-+ uint32_t res0258[2];
-+ uint32_t tpkt_l;
-+ uint32_t tpkt_u;
-+ uint32_t tund_l;
-+ uint32_t tund_u;
-+ uint32_t t64_l;
-+ uint32_t t64_u;
-+ uint32_t t127_l;
-+ uint32_t t127_u;
-+ uint32_t t255_l;
-+ uint32_t t255_u;
-+ uint32_t t511_l;
-+ uint32_t t511_u;
-+ uint32_t t1023_l;
-+ uint32_t t1023_u;
-+ uint32_t t1518_l;
-+ uint32_t t1518_u;
-+ uint32_t t1519x_l;
-+ uint32_t t1519x_u;
-+ uint32_t res02a8[6];
-+ uint32_t tcnp_l;
-+ uint32_t tcnp_u;
-+ uint32_t res02c8[14];
-+ /* Line Interface Control */
-+ uint32_t if_mode; /* 0x300 Interface Mode Control */
-+ uint32_t if_status; /* 0x304 Interface Status */
-+ uint32_t res0308[14];
-+ /* HiGig/2 */
-+ uint32_t hg_config; /* 0x340 Control and cfg */
-+ uint32_t res0344[3];
-+ uint32_t hg_pause_quanta; /* 0x350 Pause quanta */
-+ uint32_t res0354[3];
-+ uint32_t hg_pause_thresh; /* 0x360 Pause quanta threshold */
-+ uint32_t res0364[3];
-+ uint32_t hgrx_pause_status; /* 0x370 Receive pause status */
-+ uint32_t hg_fifos_status; /* 0x374 fifos status */
-+ uint32_t rhm; /* 0x378 rx messages counter */
-+ uint32_t thm; /* 0x37C tx messages counter */
-+};
-+
-+struct memac_cfg {
-+ bool reset_on_init;
-+ bool rx_error_discard;
-+ bool pause_ignore;
-+ bool pause_forward_enable;
-+ bool no_length_check_enable;
-+ bool cmd_frame_enable;
-+ bool send_idle_enable;
-+ bool wan_mode_enable;
-+ bool promiscuous_mode_enable;
-+ bool tx_addr_ins_enable;
-+ bool loopback_enable;
-+ bool lgth_check_nostdr;
-+ bool time_stamp_enable;
-+ bool pad_enable;
-+ bool phy_tx_ena_on;
-+ bool rx_sfd_any;
-+ bool rx_pbl_fwd;
-+ bool tx_pbl_fwd;
-+ bool debug_mode;
-+ bool wake_on_lan;
-+ uint16_t max_frame_length;
-+ uint16_t pause_quanta;
-+ uint32_t tx_ipg_length;
-+};
-+
-+
-+/**
-+ * fman_memac_defconfig() - Get default MEMAC configuration
-+ * @cfg: pointer to configuration structure.
-+ *
-+ * Call this function to obtain a default set of configuration values for
-+ * initializing MEMAC. The user can overwrite any of the values before calling
-+ * fman_memac_init(), if specific configuration needs to be applied.
-+ */
-+void fman_memac_defconfig(struct memac_cfg *cfg);
-+
-+int fman_memac_init(struct memac_regs *regs,
-+ struct memac_cfg *cfg,
-+ enum enet_interface enet_interface,
-+ enum enet_speed enet_speed,
-+ bool slow_10g_if,
-+ uint32_t exceptions);
-+
-+void fman_memac_enable(struct memac_regs *regs, bool apply_rx, bool apply_tx);
-+
-+void fman_memac_disable(struct memac_regs *regs, bool apply_rx, bool apply_tx);
-+
-+void fman_memac_set_promiscuous(struct memac_regs *regs, bool val);
-+
-+void fman_memac_add_addr_in_paddr(struct memac_regs *regs,
-+ uint8_t *adr,
-+ uint8_t paddr_num);
-+
-+void fman_memac_clear_addr_in_paddr(struct memac_regs *regs,
-+ uint8_t paddr_num);
-+
-+uint64_t fman_memac_get_counter(struct memac_regs *regs,
-+ enum memac_counters reg_name);
-+
-+void fman_memac_set_tx_pause_frames(struct memac_regs *regs,
-+ uint8_t priority, uint16_t pauseTime, uint16_t threshTime);
-+
-+uint16_t fman_memac_get_max_frame_len(struct memac_regs *regs);
-+
-+void fman_memac_set_exception(struct memac_regs *regs, uint32_t val,
-+ bool enable);
-+
-+void fman_memac_reset_stat(struct memac_regs *regs);
-+
-+void fman_memac_reset(struct memac_regs *regs);
-+
-+void fman_memac_reset_filter_table(struct memac_regs *regs);
-+
-+void fman_memac_set_hash_table_entry(struct memac_regs *regs, uint32_t crc);
-+
-+void fman_memac_set_hash_table(struct memac_regs *regs, uint32_t val);
-+
-+void fman_memac_set_rx_ignore_pause_frames(struct memac_regs *regs,
-+ bool enable);
-+
-+void fman_memac_set_wol(struct memac_regs *regs, bool enable);
-+
-+uint32_t fman_memac_get_event(struct memac_regs *regs, uint32_t ev_mask);
-+
-+void fman_memac_ack_event(struct memac_regs *regs, uint32_t ev_mask);
-+
-+uint32_t fman_memac_get_interrupt_mask(struct memac_regs *regs);
-+
-+void fman_memac_adjust_link(struct memac_regs *regs,
-+ enum enet_interface iface_mode,
-+ enum enet_speed speed, bool full_dx);
-+
-+
-+
-+#endif /*__FSL_FMAN_MEMAC_H*/
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/flib/fsl_fman_memac_mii_acc.h
-@@ -0,0 +1,78 @@
-+/*
-+ * Copyright 2008-2013 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#ifndef __FSL_FMAN_MEMAC_MII_ACC_H
-+#define __FSL_FMAN_MEMAC_MII_ACC_H
-+
-+#include "common/general.h"
-+#include "fsl_enet.h"
-+/* MII Management Registers */
-+#define MDIO_CFG_CLK_DIV_MASK 0x0080ff80
-+#define MDIO_CFG_CLK_DIV_SHIFT 7
-+#define MDIO_CFG_HOLD_MASK 0x0000001c
-+#define MDIO_CFG_ENC45 0x00000040
-+#define MDIO_CFG_READ_ERR 0x00000002
-+#define MDIO_CFG_BSY 0x00000001
-+
-+#define MDIO_CTL_PHY_ADDR_SHIFT 5
-+#define MDIO_CTL_READ 0x00008000
-+
-+#define MDIO_DATA_BSY 0x80000000
-+
-+/*MEMAC Internal PHY Registers - SGMII */
-+#define PHY_SGMII_CR_PHY_RESET 0x8000
-+#define PHY_SGMII_CR_RESET_AN 0x0200
-+#define PHY_SGMII_CR_DEF_VAL 0x1140
-+#define PHY_SGMII_DEV_ABILITY_SGMII 0x4001
-+#define PHY_SGMII_DEV_ABILITY_1000X 0x01A0
-+#define PHY_SGMII_IF_MODE_AN 0x0002
-+#define PHY_SGMII_IF_MODE_SGMII 0x0001
-+#define PHY_SGMII_IF_MODE_1000X 0x0000
-+
-+/*----------------------------------------------------*/
-+/* MII Configuration Control Memory Map Registers */
-+/*----------------------------------------------------*/
-+struct memac_mii_access_mem_map {
-+ uint32_t mdio_cfg; /* 0x030 */
-+ uint32_t mdio_ctrl; /* 0x034 */
-+ uint32_t mdio_data; /* 0x038 */
-+ uint32_t mdio_addr; /* 0x03c */
-+};
-+
-+int fman_memac_mii_read_phy_reg(struct memac_mii_access_mem_map *mii_regs,
-+ uint8_t phy_addr, uint8_t reg, uint16_t *data,
-+ enum enet_speed enet_speed);
-+int fman_memac_mii_write_phy_reg(struct memac_mii_access_mem_map *mii_regs,
-+ uint8_t phy_addr, uint8_t reg, uint16_t data,
-+ enum enet_speed enet_speed);
-+
-+#endif /* __MAC_API_MEMAC_MII_ACC_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/flib/fsl_fman_port.h
-@@ -0,0 +1,593 @@
-+/*
-+ * Copyright 2008-2013 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#ifndef __FSL_FMAN_PORT_H
-+#define __FSL_FMAN_PORT_H
-+
-+#include "fsl_fman_sp.h"
-+
-+/** @Collection Registers bit fields */
-+
-+/** @Description BMI defines */
-+#define BMI_EBD_EN 0x80000000
-+
-+#define BMI_PORT_CFG_EN 0x80000000
-+#define BMI_PORT_CFG_FDOVR 0x02000000
-+#define BMI_PORT_CFG_IM 0x01000000
-+
-+#define BMI_PORT_STATUS_BSY 0x80000000
-+
-+#define BMI_DMA_ATTR_SWP_SHIFT FMAN_SP_DMA_ATTR_SWP_SHIFT
-+#define BMI_DMA_ATTR_IC_STASH_ON 0x10000000
-+#define BMI_DMA_ATTR_HDR_STASH_ON 0x04000000
-+#define BMI_DMA_ATTR_SG_STASH_ON 0x01000000
-+#define BMI_DMA_ATTR_WRITE_OPTIMIZE FMAN_SP_DMA_ATTR_WRITE_OPTIMIZE
-+
-+#define BMI_RX_FIFO_PRI_ELEVATION_SHIFT 16
-+#define BMI_RX_FIFO_THRESHOLD_ETHE 0x80000000
-+
-+#define BMI_TX_FRAME_END_CS_IGNORE_SHIFT 24
-+#define BMI_RX_FRAME_END_CS_IGNORE_SHIFT 24
-+#define BMI_RX_FRAME_END_CUT_SHIFT 16
-+
-+#define BMI_IC_TO_EXT_SHIFT FMAN_SP_IC_TO_EXT_SHIFT
-+#define BMI_IC_FROM_INT_SHIFT FMAN_SP_IC_FROM_INT_SHIFT
-+
-+#define BMI_INT_BUF_MARG_SHIFT 28
-+#define BMI_EXT_BUF_MARG_START_SHIFT FMAN_SP_EXT_BUF_MARG_START_SHIFT
-+
-+#define BMI_CMD_MR_LEAC 0x00200000
-+#define BMI_CMD_MR_SLEAC 0x00100000
-+#define BMI_CMD_MR_MA 0x00080000
-+#define BMI_CMD_MR_DEAS 0x00040000
-+#define BMI_CMD_RX_MR_DEF (BMI_CMD_MR_LEAC | \
-+ BMI_CMD_MR_SLEAC | \
-+ BMI_CMD_MR_MA | \
-+ BMI_CMD_MR_DEAS)
-+#define BMI_CMD_TX_MR_DEF 0
-+#define BMI_CMD_OP_MR_DEF (BMI_CMD_MR_DEAS | \
-+ BMI_CMD_MR_MA)
-+
-+#define BMI_CMD_ATTR_ORDER 0x80000000
-+#define BMI_CMD_ATTR_SYNC 0x02000000
-+#define BMI_CMD_ATTR_COLOR_SHIFT 26
-+
-+#define BMI_FIFO_PIPELINE_DEPTH_SHIFT 12
-+#define BMI_NEXT_ENG_FD_BITS_SHIFT 24
-+#define BMI_FRAME_END_CS_IGNORE_SHIFT 24
-+
-+#define BMI_COUNTERS_EN 0x80000000
-+
-+#define BMI_EXT_BUF_POOL_VALID FMAN_SP_EXT_BUF_POOL_VALID
-+#define BMI_EXT_BUF_POOL_EN_COUNTER FMAN_SP_EXT_BUF_POOL_EN_COUNTER
-+#define BMI_EXT_BUF_POOL_BACKUP FMAN_SP_EXT_BUF_POOL_BACKUP
-+#define BMI_EXT_BUF_POOL_ID_SHIFT 16
-+#define BMI_EXT_BUF_POOL_ID_MASK 0x003F0000
-+#define BMI_POOL_DEP_NUM_OF_POOLS_SHIFT 16
-+
-+#define BMI_TX_FIFO_MIN_FILL_SHIFT 16
-+#define BMI_TX_FIFO_PIPELINE_DEPTH_SHIFT 12
-+
-+#define MAX_PERFORMANCE_TASK_COMP 64
-+#define MAX_PERFORMANCE_RX_QUEUE_COMP 64
-+#define MAX_PERFORMANCE_TX_QUEUE_COMP 8
-+#define MAX_PERFORMANCE_DMA_COMP 16
-+#define MAX_PERFORMANCE_FIFO_COMP 1024
-+
-+#define BMI_PERFORMANCE_TASK_COMP_SHIFT 24
-+#define BMI_PERFORMANCE_QUEUE_COMP_SHIFT 16
-+#define BMI_PERFORMANCE_DMA_COMP_SHIFT 12
-+
-+#define BMI_RATE_LIMIT_GRAN_TX 16000 /* In Kbps */
-+#define BMI_RATE_LIMIT_GRAN_OP 10000 /* In frames */
-+#define BMI_RATE_LIMIT_MAX_RATE_IN_GRAN_UNITS 1024
-+#define BMI_RATE_LIMIT_MAX_BURST_SIZE 1024 /* In KBytes */
-+#define BMI_RATE_LIMIT_MAX_BURST_SHIFT 16
-+#define BMI_RATE_LIMIT_HIGH_BURST_SIZE_GRAN 0x80000000
-+#define BMI_RATE_LIMIT_SCALE_TSBS_SHIFT 16
-+#define BMI_RATE_LIMIT_SCALE_EN 0x80000000
-+#define BMI_SG_DISABLE FMAN_SP_SG_DISABLE
-+
-+/** @Description QMI defines */
-+#define QMI_PORT_CFG_EN 0x80000000
-+#define QMI_PORT_CFG_EN_COUNTERS 0x10000000
-+
-+#define QMI_PORT_STATUS_DEQ_TNUM_BSY 0x80000000
-+#define QMI_PORT_STATUS_DEQ_FD_BSY 0x20000000
-+
-+#define QMI_DEQ_CFG_PRI 0x80000000
-+#define QMI_DEQ_CFG_TYPE1 0x10000000
-+#define QMI_DEQ_CFG_TYPE2 0x20000000
-+#define QMI_DEQ_CFG_TYPE3 0x30000000
-+#define QMI_DEQ_CFG_PREFETCH_PARTIAL 0x01000000
-+#define QMI_DEQ_CFG_PREFETCH_FULL 0x03000000
-+#define QMI_DEQ_CFG_SP_MASK 0xf
-+#define QMI_DEQ_CFG_SP_SHIFT 20
-+
-+
-+/** @Description General port defines */
-+#define FMAN_PORT_EXT_POOLS_NUM(fm_rev_maj) \
-+ (((fm_rev_maj) == 4) ? 4 : 8)
-+#define FMAN_PORT_MAX_EXT_POOLS_NUM 8
-+#define FMAN_PORT_OBS_EXT_POOLS_NUM 2
-+#define FMAN_PORT_CG_MAP_NUM 8
-+#define FMAN_PORT_PRS_RESULT_WORDS_NUM 8
-+#define FMAN_PORT_BMI_FIFO_UNITS 0x100
-+#define FMAN_PORT_IC_OFFSET_UNITS 0x10
-+
-+
-+/** @Collection FM Port Register Map */
-+
-+/** @Description BMI Rx port register map */
-+struct fman_port_rx_bmi_regs {
-+ uint32_t fmbm_rcfg; /**< Rx Configuration */
-+ uint32_t fmbm_rst; /**< Rx Status */
-+ uint32_t fmbm_rda; /**< Rx DMA attributes*/
-+ uint32_t fmbm_rfp; /**< Rx FIFO Parameters*/
-+ uint32_t fmbm_rfed; /**< Rx Frame End Data*/
-+ uint32_t fmbm_ricp; /**< Rx Internal Context Parameters*/
-+ uint32_t fmbm_rim; /**< Rx Internal Buffer Margins*/
-+ uint32_t fmbm_rebm; /**< Rx External Buffer Margins*/
-+ uint32_t fmbm_rfne; /**< Rx Frame Next Engine*/
-+ uint32_t fmbm_rfca; /**< Rx Frame Command Attributes.*/
-+ uint32_t fmbm_rfpne; /**< Rx Frame Parser Next Engine*/
-+ uint32_t fmbm_rpso; /**< Rx Parse Start Offset*/
-+ uint32_t fmbm_rpp; /**< Rx Policer Profile */
-+ uint32_t fmbm_rccb; /**< Rx Coarse Classification Base */
-+ uint32_t fmbm_reth; /**< Rx Excessive Threshold */
-+ uint32_t reserved003c[1]; /**< (0x03C 0x03F) */
-+ uint32_t fmbm_rprai[FMAN_PORT_PRS_RESULT_WORDS_NUM];
-+ /**< Rx Parse Results Array Init*/
-+ uint32_t fmbm_rfqid; /**< Rx Frame Queue ID*/
-+ uint32_t fmbm_refqid; /**< Rx Error Frame Queue ID*/
-+ uint32_t fmbm_rfsdm; /**< Rx Frame Status Discard Mask*/
-+ uint32_t fmbm_rfsem; /**< Rx Frame Status Error Mask*/
-+ uint32_t fmbm_rfene; /**< Rx Frame Enqueue Next Engine */
-+ uint32_t reserved0074[0x2]; /**< (0x074-0x07C) */
-+ uint32_t fmbm_rcmne; /**< Rx Frame Continuous Mode Next Engine */
-+ uint32_t reserved0080[0x20];/**< (0x080 0x0FF) */
-+ uint32_t fmbm_ebmpi[FMAN_PORT_MAX_EXT_POOLS_NUM];
-+ /**< Buffer Manager pool Information-*/
-+ uint32_t fmbm_acnt[FMAN_PORT_MAX_EXT_POOLS_NUM];
-+ /**< Allocate Counter-*/
-+ uint32_t reserved0130[8];
-+ /**< 0x130/0x140 - 0x15F reserved -*/
-+ uint32_t fmbm_rcgm[FMAN_PORT_CG_MAP_NUM];
-+ /**< Congestion Group Map*/
-+ uint32_t fmbm_mpd; /**< BM Pool Depletion */
-+ uint32_t reserved0184[0x1F]; /**< (0x184 0x1FF) */
-+ uint32_t fmbm_rstc; /**< Rx Statistics Counters*/
-+ uint32_t fmbm_rfrc; /**< Rx Frame Counter*/
-+ uint32_t fmbm_rfbc; /**< Rx Bad Frames Counter*/
-+ uint32_t fmbm_rlfc; /**< Rx Large Frames Counter*/
-+ uint32_t fmbm_rffc; /**< Rx Filter Frames Counter*/
-+ uint32_t fmbm_rfdc; /**< Rx Frame Discard Counter*/
-+ uint32_t fmbm_rfldec; /**< Rx Frames List DMA Error Counter*/
-+ uint32_t fmbm_rodc; /**< Rx Out of Buffers Discard nntr*/
-+ uint32_t fmbm_rbdc; /**< Rx Buffers Deallocate Counter*/
-+ uint32_t reserved0224[0x17]; /**< (0x224 0x27F) */
-+ uint32_t fmbm_rpc; /**< Rx Performance Counters*/
-+ uint32_t fmbm_rpcp; /**< Rx Performance Count Parameters*/
-+ uint32_t fmbm_rccn; /**< Rx Cycle Counter*/
-+ uint32_t fmbm_rtuc; /**< Rx Tasks Utilization Counter*/
-+ uint32_t fmbm_rrquc; /**< Rx Receive Queue Utilization cntr*/
-+ uint32_t fmbm_rduc; /**< Rx DMA Utilization Counter*/
-+ uint32_t fmbm_rfuc; /**< Rx FIFO Utilization Counter*/
-+ uint32_t fmbm_rpac; /**< Rx Pause Activation Counter*/
-+ uint32_t reserved02a0[0x18]; /**< (0x2A0 0x2FF) */
-+ uint32_t fmbm_rdbg; /**< Rx Debug-*/
-+};
-+
-+/** @Description BMI Tx port register map */
-+struct fman_port_tx_bmi_regs {
-+ uint32_t fmbm_tcfg; /**< Tx Configuration */
-+ uint32_t fmbm_tst; /**< Tx Status */
-+ uint32_t fmbm_tda; /**< Tx DMA attributes */
-+ uint32_t fmbm_tfp; /**< Tx FIFO Parameters */
-+ uint32_t fmbm_tfed; /**< Tx Frame End Data */
-+ uint32_t fmbm_ticp; /**< Tx Internal Context Parameters */
-+ uint32_t fmbm_tfdne; /**< Tx Frame Dequeue Next Engine. */
-+ uint32_t fmbm_tfca; /**< Tx Frame Command attribute. */
-+ uint32_t fmbm_tcfqid; /**< Tx Confirmation Frame Queue ID. */
-+ uint32_t fmbm_tefqid; /**< Tx Frame Error Queue ID */
-+ uint32_t fmbm_tfene; /**< Tx Frame Enqueue Next Engine */
-+ uint32_t fmbm_trlmts; /**< Tx Rate Limiter Scale */
-+ uint32_t fmbm_trlmt; /**< Tx Rate Limiter */
-+ uint32_t reserved0034[0x0e]; /**< (0x034-0x6c) */
-+ uint32_t fmbm_tccb; /**< Tx Coarse Classification base */
-+ uint32_t fmbm_tfne; /**< Tx Frame Next Engine */
-+ uint32_t fmbm_tpfcm[0x02]; /**< Tx Priority based Flow Control (PFC) Mapping */
-+ uint32_t fmbm_tcmne; /**< Tx Frame Continuous Mode Next Engine */
-+ uint32_t reserved0080[0x60]; /**< (0x080-0x200) */
-+ uint32_t fmbm_tstc; /**< Tx Statistics Counters */
-+ uint32_t fmbm_tfrc; /**< Tx Frame Counter */
-+ uint32_t fmbm_tfdc; /**< Tx Frames Discard Counter */
-+ uint32_t fmbm_tfledc; /**< Tx Frame len error discard cntr */
-+ uint32_t fmbm_tfufdc; /**< Tx Frame unsprt frmt discard cntr*/
-+ uint32_t fmbm_tbdc; /**< Tx Buffers Deallocate Counter */
-+ uint32_t reserved0218[0x1A]; /**< (0x218-0x280) */
-+ uint32_t fmbm_tpc; /**< Tx Performance Counters*/
-+ uint32_t fmbm_tpcp; /**< Tx Performance Count Parameters*/
-+ uint32_t fmbm_tccn; /**< Tx Cycle Counter*/
-+ uint32_t fmbm_ttuc; /**< Tx Tasks Utilization Counter*/
-+ uint32_t fmbm_ttcquc; /**< Tx Transmit conf Q util Counter*/
-+ uint32_t fmbm_tduc; /**< Tx DMA Utilization Counter*/
-+ uint32_t fmbm_tfuc; /**< Tx FIFO Utilization Counter*/
-+};
-+
-+/** @Description BMI O/H port register map */
-+struct fman_port_oh_bmi_regs {
-+ uint32_t fmbm_ocfg; /**< O/H Configuration */
-+ uint32_t fmbm_ost; /**< O/H Status */
-+ uint32_t fmbm_oda; /**< O/H DMA attributes */
-+ uint32_t fmbm_oicp; /**< O/H Internal Context Parameters */
-+ uint32_t fmbm_ofdne; /**< O/H Frame Dequeue Next Engine */
-+ uint32_t fmbm_ofne; /**< O/H Frame Next Engine */
-+ uint32_t fmbm_ofca; /**< O/H Frame Command Attributes. */
-+ uint32_t fmbm_ofpne; /**< O/H Frame Parser Next Engine */
-+ uint32_t fmbm_opso; /**< O/H Parse Start Offset */
-+ uint32_t fmbm_opp; /**< O/H Policer Profile */
-+ uint32_t fmbm_occb; /**< O/H Coarse Classification base */
-+ uint32_t fmbm_oim; /**< O/H Internal margins*/
-+ uint32_t fmbm_ofp; /**< O/H Fifo Parameters*/
-+ uint32_t fmbm_ofed; /**< O/H Frame End Data*/
-+ uint32_t reserved0030[2]; /**< (0x038 - 0x03F) */
-+ uint32_t fmbm_oprai[FMAN_PORT_PRS_RESULT_WORDS_NUM];
-+ /**< O/H Parse Results Array Initialization */
-+ uint32_t fmbm_ofqid; /**< O/H Frame Queue ID */
-+ uint32_t fmbm_oefqid; /**< O/H Error Frame Queue ID */
-+ uint32_t fmbm_ofsdm; /**< O/H Frame Status Discard Mask */
-+ uint32_t fmbm_ofsem; /**< O/H Frame Status Error Mask */
-+ uint32_t fmbm_ofene; /**< O/H Frame Enqueue Next Engine */
-+ uint32_t fmbm_orlmts; /**< O/H Rate Limiter Scale */
-+ uint32_t fmbm_orlmt; /**< O/H Rate Limiter */
-+ uint32_t fmbm_ocmne; /**< O/H Continuous Mode Next Engine */
-+ uint32_t reserved0080[0x20]; /**< 0x080 - 0x0FF Reserved */
-+ uint32_t fmbm_oebmpi[2]; /**< Buf Mngr Observed Pool Info */
-+ uint32_t reserved0108[0x16]; /**< 0x108 - 0x15F Reserved */
-+ uint32_t fmbm_ocgm[FMAN_PORT_CG_MAP_NUM]; /**< Observed Congestion Group Map */
-+ uint32_t fmbm_ompd; /**< Observed BMan Pool Depletion */
-+ uint32_t reserved0184[0x1F]; /**< 0x184 - 0x1FF Reserved */
-+ uint32_t fmbm_ostc; /**< O/H Statistics Counters */
-+ uint32_t fmbm_ofrc; /**< O/H Frame Counter */
-+ uint32_t fmbm_ofdc; /**< O/H Frames Discard Counter */
-+ uint32_t fmbm_ofledc; /**< O/H Frames Len Err Discard Cntr */
-+ uint32_t fmbm_ofufdc; /**< O/H Frames Unsprtd Discard Cutr */
-+ uint32_t fmbm_offc; /**< O/H Filter Frames Counter */
-+ uint32_t fmbm_ofwdc; /**< Rx Frames WRED Discard Counter */
-+ uint32_t fmbm_ofldec; /**< O/H Frames List DMA Error Cntr */
-+ uint32_t fmbm_obdc; /**< O/H Buffers Deallocate Counter */
-+ uint32_t reserved0218[0x17]; /**< (0x218 - 0x27F) */
-+ uint32_t fmbm_opc; /**< O/H Performance Counters */
-+ uint32_t fmbm_opcp; /**< O/H Performance Count Parameters */
-+ uint32_t fmbm_occn; /**< O/H Cycle Counter */
-+ uint32_t fmbm_otuc; /**< O/H Tasks Utilization Counter */
-+ uint32_t fmbm_oduc; /**< O/H DMA Utilization Counter */
-+ uint32_t fmbm_ofuc; /**< O/H FIFO Utilization Counter */
-+};
-+
-+/** @Description BMI port register map */
-+union fman_port_bmi_regs {
-+ struct fman_port_rx_bmi_regs rx;
-+ struct fman_port_tx_bmi_regs tx;
-+ struct fman_port_oh_bmi_regs oh;
-+};
-+
-+/** @Description QMI port register map */
-+struct fman_port_qmi_regs {
-+ uint32_t fmqm_pnc; /**< PortID n Configuration Register */
-+ uint32_t fmqm_pns; /**< PortID n Status Register */
-+ uint32_t fmqm_pnts; /**< PortID n Task Status Register */
-+ uint32_t reserved00c[4]; /**< 0xn00C - 0xn01B */
-+ uint32_t fmqm_pnen; /**< PortID n Enqueue NIA Register */
-+ uint32_t fmqm_pnetfc; /**< PortID n Enq Total Frame Counter */
-+ uint32_t reserved024[2]; /**< 0xn024 - 0x02B */
-+ uint32_t fmqm_pndn; /**< PortID n Dequeue NIA Register */
-+ uint32_t fmqm_pndc; /**< PortID n Dequeue Config Register */
-+ uint32_t fmqm_pndtfc; /**< PortID n Dequeue tot Frame cntr */
-+ uint32_t fmqm_pndfdc; /**< PortID n Dequeue FQID Dflt Cntr */
-+ uint32_t fmqm_pndcc; /**< PortID n Dequeue Confirm Counter */
-+};
-+
-+
-+enum fman_port_dma_swap {
-+ E_FMAN_PORT_DMA_NO_SWAP, /**< No swap, transfer data as is */
-+ E_FMAN_PORT_DMA_SWAP_LE,
-+ /**< The transferred data should be swapped in PPC Little Endian mode */
-+ E_FMAN_PORT_DMA_SWAP_BE
-+ /**< The transferred data should be swapped in Big Endian mode */
-+};
-+
-+/* Default port color */
-+enum fman_port_color {
-+ E_FMAN_PORT_COLOR_GREEN, /**< Default port color is green */
-+ E_FMAN_PORT_COLOR_YELLOW, /**< Default port color is yellow */
-+ E_FMAN_PORT_COLOR_RED, /**< Default port color is red */
-+ E_FMAN_PORT_COLOR_OVERRIDE /**< Ignore color */
-+};
-+
-+/* QMI dequeue from the SP channel - types */
-+enum fman_port_deq_type {
-+ E_FMAN_PORT_DEQ_BY_PRI,
-+ /**< Priority precedence and Intra-Class scheduling */
-+ E_FMAN_PORT_DEQ_ACTIVE_FQ,
-+ /**< Active FQ precedence and Intra-Class scheduling */
-+ E_FMAN_PORT_DEQ_ACTIVE_FQ_NO_ICS
-+ /**< Active FQ precedence and override Intra-Class scheduling */
-+};
-+
-+/* QMI dequeue prefetch modes */
-+enum fman_port_deq_prefetch {
-+ E_FMAN_PORT_DEQ_NO_PREFETCH, /**< No prefetch mode */
-+ E_FMAN_PORT_DEQ_PART_PREFETCH, /**< Partial prefetch mode */
-+ E_FMAN_PORT_DEQ_FULL_PREFETCH /**< Full prefetch mode */
-+};
-+
-+/* Parameters for defining performance counters behavior */
-+struct fman_port_perf_cnt_params {
-+ uint8_t task_val; /**< Task compare value */
-+ uint8_t queue_val;
-+ /**< Rx or Tx conf queue compare value (unused for O/H ports) */
-+ uint8_t dma_val; /**< Dma compare value */
-+ uint32_t fifo_val; /**< Fifo compare value (in bytes) */
-+};
-+
-+/** @Description FM Port configuration structure, used at init */
-+struct fman_port_cfg {
-+ struct fman_port_perf_cnt_params perf_cnt_params;
-+ /* BMI parameters */
-+ enum fman_port_dma_swap dma_swap_data;
-+ bool dma_ic_stash_on;
-+ bool dma_header_stash_on;
-+ bool dma_sg_stash_on;
-+ bool dma_write_optimize;
-+ uint16_t ic_ext_offset;
-+ uint8_t ic_int_offset;
-+ uint16_t ic_size;
-+ enum fman_port_color color;
-+ bool sync_req;
-+ bool discard_override;
-+ uint8_t checksum_bytes_ignore;
-+ uint8_t rx_cut_end_bytes;
-+ uint32_t rx_pri_elevation;
-+ uint32_t rx_fifo_thr;
-+ uint8_t rx_fd_bits;
-+ uint8_t int_buf_start_margin;
-+ uint16_t ext_buf_start_margin;
-+ uint16_t ext_buf_end_margin;
-+ uint32_t tx_fifo_min_level;
-+ uint32_t tx_fifo_low_comf_level;
-+ uint8_t tx_fifo_deq_pipeline_depth;
-+ bool stats_counters_enable;
-+ bool perf_counters_enable;
-+ /* QMI parameters */
-+ bool deq_high_pri;
-+ enum fman_port_deq_type deq_type;
-+ enum fman_port_deq_prefetch deq_prefetch_opt;
-+ uint16_t deq_byte_cnt;
-+ bool queue_counters_enable;
-+ bool no_scatter_gather;
-+ int errata_A006675;
-+ int errata_A006320;
-+ int excessive_threshold_register;
-+ int fmbm_rebm_has_sgd;
-+ int fmbm_tfne_has_features;
-+ int qmi_deq_options_support;
-+};
-+
-+enum fman_port_type {
-+ E_FMAN_PORT_TYPE_OP = 0,
-+ /**< Offline parsing port, shares id-s with
-+ * host command, so must have exclusive id-s */
-+ E_FMAN_PORT_TYPE_RX, /**< 1G Rx port */
-+ E_FMAN_PORT_TYPE_RX_10G, /**< 10G Rx port */
-+ E_FMAN_PORT_TYPE_TX, /**< 1G Tx port */
-+ E_FMAN_PORT_TYPE_TX_10G, /**< 10G Tx port */
-+ E_FMAN_PORT_TYPE_DUMMY,
-+ E_FMAN_PORT_TYPE_HC = E_FMAN_PORT_TYPE_DUMMY
-+ /**< Host command port, shares id-s with
-+ * offline parsing ports, so must have exclusive id-s */
-+};
-+
-+struct fman_port_params {
-+ uint32_t discard_mask;
-+ uint32_t err_mask;
-+ uint32_t dflt_fqid;
-+ uint32_t err_fqid;
-+ uint8_t deq_sp;
-+ bool dont_release_buf;
-+};
-+
-+/* Port context - used by most API functions */
-+struct fman_port {
-+ enum fman_port_type type;
-+ uint8_t fm_rev_maj;
-+ uint8_t fm_rev_min;
-+ union fman_port_bmi_regs *bmi_regs;
-+ struct fman_port_qmi_regs *qmi_regs;
-+ bool im_en;
-+ uint8_t ext_pools_num;
-+};
-+
-+/** @Description External buffer pools configuration */
-+struct fman_port_bpools {
-+ uint8_t count; /**< Num of pools to set up */
-+ bool counters_enable; /**< Enable allocate counters */
-+ uint8_t grp_bp_depleted_num;
-+ /**< Number of depleted pools - if reached the BMI indicates
-+ * the MAC to send a pause frame */
-+ struct {
-+ uint8_t bpid; /**< BM pool ID */
-+ uint16_t size;
-+ /**< Pool's size - must be in ascending order */
-+ bool is_backup;
-+ /**< If this is a backup pool */
-+ bool grp_bp_depleted;
-+ /**< Consider this buffer in multiple pools depletion criteria*/
-+ bool single_bp_depleted;
-+ /**< Consider this buffer in single pool depletion criteria */
-+ bool pfc_priorities_en;
-+ } bpool[FMAN_PORT_MAX_EXT_POOLS_NUM];
-+};
-+
-+enum fman_port_rate_limiter_scale_down {
-+ E_FMAN_PORT_RATE_DOWN_NONE,
-+ E_FMAN_PORT_RATE_DOWN_BY_2,
-+ E_FMAN_PORT_RATE_DOWN_BY_4,
-+ E_FMAN_PORT_RATE_DOWN_BY_8
-+};
-+
-+/* Rate limiter configuration */
-+struct fman_port_rate_limiter {
-+ uint8_t count_1micro_bit;
-+ bool high_burst_size_gran;
-+ /**< Defines burst_size granularity for OP ports; when TRUE,
-+ * burst_size below counts in frames, otherwise in 10^3 frames */
-+ uint16_t burst_size;
-+ /**< Max burst size, in KBytes for Tx port, according to
-+ * high_burst_size_gran definition for OP port */
-+ uint32_t rate;
-+ /**< In Kbps for Tx port, in frames/sec for OP port */
-+ enum fman_port_rate_limiter_scale_down rate_factor;
-+};
-+
-+/* BMI statistics counters */
-+enum fman_port_stats_counters {
-+ E_FMAN_PORT_STATS_CNT_FRAME,
-+ /**< Number of processed frames; valid for all ports */
-+ E_FMAN_PORT_STATS_CNT_DISCARD,
-+ /**< For Rx ports - frames discarded by QMAN, for Tx or O/H ports -
-+ * frames discarded due to DMA error; valid for all ports */
-+ E_FMAN_PORT_STATS_CNT_DEALLOC_BUF,
-+ /**< Number of buffer deallocate operations; valid for all ports */
-+ E_FMAN_PORT_STATS_CNT_RX_BAD_FRAME,
-+ /**< Number of bad Rx frames, like CRC error, Rx FIFO overflow etc;
-+ * valid for Rx ports only */
-+ E_FMAN_PORT_STATS_CNT_RX_LARGE_FRAME,
-+ /**< Number of Rx oversized frames, that is frames exceeding max frame
-+ * size configured for the corresponding ETH controller;
-+ * valid for Rx ports only */
-+ E_FMAN_PORT_STATS_CNT_RX_OUT_OF_BUF,
-+ /**< Frames discarded due to lack of external buffers; valid for
-+ * Rx ports only */
-+ E_FMAN_PORT_STATS_CNT_LEN_ERR,
-+ /**< Frames discarded due to frame length error; valid for Tx and
-+ * O/H ports only */
-+ E_FMAN_PORT_STATS_CNT_UNSUPPORTED_FORMAT,
-+ /**< Frames discarded due to unsupported FD format; valid for Tx
-+ * and O/H ports only */
-+ E_FMAN_PORT_STATS_CNT_FILTERED_FRAME,
-+ /**< Number of frames filtered out by PCD module; valid for
-+ * Rx and OP ports only */
-+ E_FMAN_PORT_STATS_CNT_DMA_ERR,
-+ /**< Frames rejected by QMAN that were not able to release their
-+ * buffers due to DMA error; valid for Rx and O/H ports only */
-+ E_FMAN_PORT_STATS_CNT_WRED_DISCARD
-+ /**< Frames going through O/H port that were not able to to enter the
-+ * return queue due to WRED algorithm; valid for O/H ports only */
-+};
-+
-+/* BMI performance counters */
-+enum fman_port_perf_counters {
-+ E_FMAN_PORT_PERF_CNT_CYCLE, /**< Cycle counter */
-+ E_FMAN_PORT_PERF_CNT_TASK_UTIL, /**< Tasks utilization counter */
-+ E_FMAN_PORT_PERF_CNT_QUEUE_UTIL,
-+ /**< For Rx ports - Rx queue utilization, for Tx ports - Tx conf queue
-+ * utilization; not valid for O/H ports */
-+ E_FMAN_PORT_PERF_CNT_DMA_UTIL, /**< DMA utilization counter */
-+ E_FMAN_PORT_PERF_CNT_FIFO_UTIL, /**< FIFO utilization counter */
-+ E_FMAN_PORT_PERF_CNT_RX_PAUSE
-+ /**< Number of cycles in which Rx pause activation control is on;
-+ * valid for Rx ports only */
-+};
-+
-+/* QMI counters */
-+enum fman_port_qmi_counters {
-+ E_FMAN_PORT_ENQ_TOTAL, /**< EnQ tot frame cntr */
-+ E_FMAN_PORT_DEQ_TOTAL, /**< DeQ tot frame cntr; invalid for Rx ports */
-+ E_FMAN_PORT_DEQ_FROM_DFLT,
-+ /**< Dequeue from default FQID counter not valid for Rx ports */
-+ E_FMAN_PORT_DEQ_CONFIRM /**< DeQ confirm cntr invalid for Rx ports */
-+};
-+
-+
-+/** @Collection FM Port API */
-+void fman_port_defconfig(struct fman_port_cfg *cfg, enum fman_port_type type);
-+int fman_port_init(struct fman_port *port,
-+ struct fman_port_cfg *cfg,
-+ struct fman_port_params *params);
-+int fman_port_enable(struct fman_port *port);
-+int fman_port_disable(const struct fman_port *port);
-+int fman_port_set_bpools(const struct fman_port *port,
-+ const struct fman_port_bpools *bp);
-+int fman_port_set_rate_limiter(struct fman_port *port,
-+ struct fman_port_rate_limiter *rate_limiter);
-+int fman_port_delete_rate_limiter(struct fman_port *port);
-+int fman_port_set_err_mask(struct fman_port *port, uint32_t err_mask);
-+int fman_port_set_discard_mask(struct fman_port *port, uint32_t discard_mask);
-+int fman_port_modify_rx_fd_bits(struct fman_port *port,
-+ uint8_t rx_fd_bits,
-+ bool add);
-+int fman_port_set_perf_cnt_params(struct fman_port *port,
-+ struct fman_port_perf_cnt_params *params);
-+int fman_port_set_stats_cnt_mode(struct fman_port *port, bool enable);
-+int fman_port_set_perf_cnt_mode(struct fman_port *port, bool enable);
-+int fman_port_set_queue_cnt_mode(struct fman_port *port, bool enable);
-+int fman_port_set_bpool_cnt_mode(struct fman_port *port,
-+ uint8_t bpid,
-+ bool enable);
-+uint32_t fman_port_get_stats_counter(struct fman_port *port,
-+ enum fman_port_stats_counters counter);
-+void fman_port_set_stats_counter(struct fman_port *port,
-+ enum fman_port_stats_counters counter,
-+ uint32_t value);
-+uint32_t fman_port_get_perf_counter(struct fman_port *port,
-+ enum fman_port_perf_counters counter);
-+void fman_port_set_perf_counter(struct fman_port *port,
-+ enum fman_port_perf_counters counter,
-+ uint32_t value);
-+uint32_t fman_port_get_qmi_counter(struct fman_port *port,
-+ enum fman_port_qmi_counters counter);
-+void fman_port_set_qmi_counter(struct fman_port *port,
-+ enum fman_port_qmi_counters counter,
-+ uint32_t value);
-+uint32_t fman_port_get_bpool_counter(struct fman_port *port, uint8_t bpid);
-+void fman_port_set_bpool_counter(struct fman_port *port,
-+ uint8_t bpid,
-+ uint32_t value);
-+int fman_port_add_congestion_grps(struct fman_port *port,
-+ uint32_t grps_map[FMAN_PORT_CG_MAP_NUM]);
-+int fman_port_remove_congestion_grps(struct fman_port *port,
-+ uint32_t grps_map[FMAN_PORT_CG_MAP_NUM]);
-+
-+
-+#endif /* __FSL_FMAN_PORT_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/flib/fsl_fman_prs.h
-@@ -0,0 +1,102 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#ifndef __FSL_FMAN_PRS_H
-+#define __FSL_FMAN_PRS_H
-+
-+#include "common/general.h"
-+
-+#define FM_PCD_EX_PRS_DOUBLE_ECC 0x02000000
-+#define FM_PCD_EX_PRS_SINGLE_ECC 0x01000000
-+
-+#define FM_PCD_PRS_PPSC_ALL_PORTS 0xffff0000
-+#define FM_PCD_PRS_RPIMAC_EN 0x00000001
-+#define FM_PCD_PRS_PORT_IDLE_STS 0xffff0000
-+#define FM_PCD_PRS_SINGLE_ECC 0x00004000
-+#define FM_PCD_PRS_DOUBLE_ECC 0x00004000
-+#define PRS_MAX_CYCLE_LIMIT 8191
-+
-+#define DEFAULT_MAX_PRS_CYC_LIM 0
-+
-+struct fman_prs_regs {
-+ uint32_t fmpr_rpclim;
-+ uint32_t fmpr_rpimac;
-+ uint32_t pmeec;
-+ uint32_t res00c[5];
-+ uint32_t fmpr_pevr;
-+ uint32_t fmpr_pever;
-+ uint32_t res028;
-+ uint32_t fmpr_perr;
-+ uint32_t fmpr_perer;
-+ uint32_t res034;
-+ uint32_t res038[10];
-+ uint32_t fmpr_ppsc;
-+ uint32_t res064;
-+ uint32_t fmpr_pds;
-+ uint32_t fmpr_l2rrs;
-+ uint32_t fmpr_l3rrs;
-+ uint32_t fmpr_l4rrs;
-+ uint32_t fmpr_srrs;
-+ uint32_t fmpr_l2rres;
-+ uint32_t fmpr_l3rres;
-+ uint32_t fmpr_l4rres;
-+ uint32_t fmpr_srres;
-+ uint32_t fmpr_spcs;
-+ uint32_t fmpr_spscs;
-+ uint32_t fmpr_hxscs;
-+ uint32_t fmpr_mrcs;
-+ uint32_t fmpr_mwcs;
-+ uint32_t fmpr_mrscs;
-+ uint32_t fmpr_mwscs;
-+ uint32_t fmpr_fcscs;
-+};
-+
-+struct fman_prs_cfg {
-+ uint32_t port_id_stat;
-+ uint16_t max_prs_cyc_lim;
-+ uint32_t prs_exceptions;
-+};
-+
-+uint32_t fman_prs_get_err_event(struct fman_prs_regs *regs, uint32_t ev_mask);
-+uint32_t fman_prs_get_err_ev_mask(struct fman_prs_regs *regs);
-+void fman_prs_ack_err_event(struct fman_prs_regs *regs, uint32_t event);
-+uint32_t fman_prs_get_expt_event(struct fman_prs_regs *regs, uint32_t ev_mask);
-+uint32_t fman_prs_get_expt_ev_mask(struct fman_prs_regs *regs);
-+void fman_prs_ack_expt_event(struct fman_prs_regs *regs, uint32_t event);
-+void fman_prs_defconfig(struct fman_prs_cfg *cfg);
-+int fman_prs_init(struct fman_prs_regs *regs, struct fman_prs_cfg *cfg);
-+void fman_prs_enable(struct fman_prs_regs *regs);
-+void fman_prs_disable(struct fman_prs_regs *regs);
-+int fman_prs_is_enabled(struct fman_prs_regs *regs);
-+void fman_prs_set_stst_port_msk(struct fman_prs_regs *regs, uint32_t pid_msk);
-+void fman_prs_set_stst(struct fman_prs_regs *regs, bool enable);
-+#endif /* __FSL_FMAN_PRS_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/flib/fsl_fman_rtc.h
-@@ -0,0 +1,449 @@
-+/*
-+ * Copyright 2013 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#ifndef __FSL_FMAN_RTC_H
-+#define __FSL_FMAN_RTC_H
-+
-+#include "common/general.h"
-+
-+/* FM RTC Registers definitions */
-+#define FMAN_RTC_TMR_CTRL_ALMP1 0x80000000
-+#define FMAN_RTC_TMR_CTRL_ALMP2 0x40000000
-+#define FMAN_RTC_TMR_CTRL_FS 0x10000000
-+#define FMAN_RTC_TMR_CTRL_PP1L 0x08000000
-+#define FMAN_RTC_TMR_CTRL_PP2L 0x04000000
-+#define FMAN_RTC_TMR_CTRL_TCLK_PERIOD_MASK 0x03FF0000
-+#define FMAN_RTC_TMR_CTRL_FRD 0x00004000
-+#define FMAN_RTC_TMR_CTRL_SLV 0x00002000
-+#define FMAN_RTC_TMR_CTRL_ETEP1 0x00000100
-+#define FMAN_RTC_TMR_CTRL_COPH 0x00000080
-+#define FMAN_RTC_TMR_CTRL_CIPH 0x00000040
-+#define FMAN_RTC_TMR_CTRL_TMSR 0x00000020
-+#define FMAN_RTC_TMR_CTRL_DBG 0x00000010
-+#define FMAN_RTC_TMR_CTRL_BYP 0x00000008
-+#define FMAN_RTC_TMR_CTRL_TE 0x00000004
-+#define FMAN_RTC_TMR_CTRL_CKSEL_OSC_CLK 0x00000003
-+#define FMAN_RTC_TMR_CTRL_CKSEL_MAC_CLK 0x00000001
-+#define FMAN_RTC_TMR_CTRL_CKSEL_EXT_CLK 0x00000000
-+#define FMAN_RTC_TMR_CTRL_TCLK_PERIOD_SHIFT 16
-+
-+#define FMAN_RTC_TMR_TEVENT_ETS2 0x02000000
-+#define FMAN_RTC_TMR_TEVENT_ETS1 0x01000000
-+#define FMAN_RTC_TMR_TEVENT_ALM2 0x00020000
-+#define FMAN_RTC_TMR_TEVENT_ALM1 0x00010000
-+#define FMAN_RTC_TMR_TEVENT_PP1 0x00000080
-+#define FMAN_RTC_TMR_TEVENT_PP2 0x00000040
-+#define FMAN_RTC_TMR_TEVENT_PP3 0x00000020
-+#define FMAN_RTC_TMR_TEVENT_ALL (FMAN_RTC_TMR_TEVENT_ETS2 |\
-+ FMAN_RTC_TMR_TEVENT_ETS1 |\
-+ FMAN_RTC_TMR_TEVENT_ALM2 |\
-+ FMAN_RTC_TMR_TEVENT_ALM1 |\
-+ FMAN_RTC_TMR_TEVENT_PP1 |\
-+ FMAN_RTC_TMR_TEVENT_PP2 |\
-+ FMAN_RTC_TMR_TEVENT_PP3)
-+
-+#define FMAN_RTC_TMR_PRSC_OCK_MASK 0x0000FFFF
-+
-+/**************************************************************************//**
-+ @Description FM RTC Alarm Polarity Options.
-+*//***************************************************************************/
-+enum fman_rtc_alarm_polarity {
-+ E_FMAN_RTC_ALARM_POLARITY_ACTIVE_HIGH, /**< Active-high output polarity */
-+ E_FMAN_RTC_ALARM_POLARITY_ACTIVE_LOW /**< Active-low output polarity */
-+};
-+
-+/**************************************************************************//**
-+ @Description FM RTC Trigger Polarity Options.
-+*//***************************************************************************/
-+enum fman_rtc_trigger_polarity {
-+ E_FMAN_RTC_TRIGGER_ON_RISING_EDGE, /**< Trigger on rising edge */
-+ E_FMAN_RTC_TRIGGER_ON_FALLING_EDGE /**< Trigger on falling edge */
-+};
-+
-+/**************************************************************************//**
-+ @Description IEEE1588 Timer Module FM RTC Optional Clock Sources.
-+*//***************************************************************************/
-+enum fman_src_clock {
-+ E_FMAN_RTC_SOURCE_CLOCK_EXTERNAL, /**< external high precision timer
-+ reference clock */
-+ E_FMAN_RTC_SOURCE_CLOCK_SYSTEM, /**< MAC system clock */
-+ E_FMAN_RTC_SOURCE_CLOCK_OSCILATOR /**< RTC clock oscilator */
-+};
-+
-+/* RTC default values */
-+#define DEFAULT_SRC_CLOCK E_FMAN_RTC_SOURCE_CLOCK_SYSTEM
-+#define DEFAULT_INVERT_INPUT_CLK_PHASE FALSE
-+#define DEFAULT_INVERT_OUTPUT_CLK_PHASE FALSE
-+#define DEFAULT_ALARM_POLARITY E_FMAN_RTC_ALARM_POLARITY_ACTIVE_HIGH
-+#define DEFAULT_TRIGGER_POLARITY E_FMAN_RTC_TRIGGER_ON_FALLING_EDGE
-+#define DEFAULT_PULSE_REALIGN FALSE
-+
-+#define FMAN_RTC_MAX_NUM_OF_ALARMS 3
-+#define FMAN_RTC_MAX_NUM_OF_PERIODIC_PULSES 4
-+#define FMAN_RTC_MAX_NUM_OF_EXT_TRIGGERS 3
-+
-+/**************************************************************************//**
-+ @Description FM RTC timer alarm
-+*//***************************************************************************/
-+struct t_tmr_alarm{
-+ uint32_t tmr_alarm_h; /**< */
-+ uint32_t tmr_alarm_l; /**< */
-+};
-+
-+/**************************************************************************//**
-+ @Description FM RTC timer Ex trigger
-+*//***************************************************************************/
-+struct t_tmr_ext_trigger{
-+ uint32_t tmr_etts_h; /**< */
-+ uint32_t tmr_etts_l; /**< */
-+};
-+
-+struct rtc_regs {
-+ uint32_t tmr_id; /* 0x000 Module ID register */
-+ uint32_t tmr_id2; /* 0x004 Controller ID register */
-+ uint32_t reserved0008[30];
-+ uint32_t tmr_ctrl; /* 0x0080 timer control register */
-+ uint32_t tmr_tevent; /* 0x0084 timer event register */
-+ uint32_t tmr_temask; /* 0x0088 timer event mask register */
-+ uint32_t reserved008c[3];
-+ uint32_t tmr_cnt_h; /* 0x0098 timer counter high register */
-+ uint32_t tmr_cnt_l; /* 0x009c timer counter low register */
-+ uint32_t tmr_add; /* 0x00a0 timer drift compensation addend register */
-+ uint32_t tmr_acc; /* 0x00a4 timer accumulator register */
-+ uint32_t tmr_prsc; /* 0x00a8 timer prescale */
-+ uint32_t reserved00ac;
-+ uint32_t tmr_off_h; /* 0x00b0 timer offset high */
-+ uint32_t tmr_off_l; /* 0x00b4 timer offset low */
-+ struct t_tmr_alarm tmr_alarm[FMAN_RTC_MAX_NUM_OF_ALARMS]; /* 0x00b8 timer
-+ alarm */
-+ uint32_t tmr_fiper[FMAN_RTC_MAX_NUM_OF_PERIODIC_PULSES]; /* 0x00d0 timer
-+ fixed period interval */
-+ struct t_tmr_ext_trigger tmr_etts[FMAN_RTC_MAX_NUM_OF_EXT_TRIGGERS];
-+ /* 0x00e0 time stamp general purpose external */
-+ uint32_t reserved00f0[4];
-+};
-+
-+struct rtc_cfg {
-+ enum fman_src_clock src_clk;
-+ uint32_t ext_src_clk_freq;
-+ uint32_t rtc_freq_hz;
-+ bool timer_slave_mode;
-+ bool invert_input_clk_phase;
-+ bool invert_output_clk_phase;
-+ uint32_t events_mask;
-+ bool bypass; /**< Indicates if frequency compensation
-+ is bypassed */
-+ bool pulse_realign;
-+ enum fman_rtc_alarm_polarity alarm_polarity[FMAN_RTC_MAX_NUM_OF_ALARMS];
-+ enum fman_rtc_trigger_polarity trigger_polarity
-+ [FMAN_RTC_MAX_NUM_OF_EXT_TRIGGERS];
-+};
-+
-+/**
-+ * fman_rtc_defconfig() - Get default RTC configuration
-+ * @cfg: pointer to configuration structure.
-+ *
-+ * Call this function to obtain a default set of configuration values for
-+ * initializing RTC. The user can overwrite any of the values before calling
-+ * fman_rtc_init(), if specific configuration needs to be applied.
-+ */
-+void fman_rtc_defconfig(struct rtc_cfg *cfg);
-+
-+/**
-+ * fman_rtc_get_events() - Get the events
-+ * @regs: Pointer to RTC register block
-+ *
-+ * Returns: The events
-+ */
-+uint32_t fman_rtc_get_events(struct rtc_regs *regs);
-+
-+/**
-+ * fman_rtc_get_interrupt_mask() - Get the events mask
-+ * @regs: Pointer to RTC register block
-+ *
-+ * Returns: The events mask
-+ */
-+uint32_t fman_rtc_get_interrupt_mask(struct rtc_regs *regs);
-+
-+
-+/**
-+ * fman_rtc_set_interrupt_mask() - Set the events mask
-+ * @regs: Pointer to RTC register block
-+ * @mask: The mask to set
-+ */
-+void fman_rtc_set_interrupt_mask(struct rtc_regs *regs, uint32_t mask);
-+
-+/**
-+ * fman_rtc_get_event() - Check if specific events occurred
-+ * @regs: Pointer to RTC register block
-+ * @ev_mask: a mask of the events to check
-+ *
-+ * Returns: 0 if the events did not occur. Non zero if one of the events occurred
-+ */
-+uint32_t fman_rtc_get_event(struct rtc_regs *regs, uint32_t ev_mask);
-+
-+/**
-+ * fman_rtc_check_and_clear_event() - Clear events which are on
-+ * @regs: Pointer to RTC register block
-+ *
-+ * Returns: A mask of the events which were cleared
-+ */
-+uint32_t fman_rtc_check_and_clear_event(struct rtc_regs *regs);
-+
-+/**
-+ * fman_rtc_ack_event() - Clear events
-+ * @regs: Pointer to RTC register block
-+ * @events: The events to disable
-+ */
-+void fman_rtc_ack_event(struct rtc_regs *regs, uint32_t events);
-+
-+/**
-+ * fman_rtc_enable_interupt() - Enable events interrupts
-+ * @regs: Pointer to RTC register block
-+ * @mask: The events to disable
-+ */
-+void fman_rtc_enable_interupt(struct rtc_regs *regs, uint32_t mask);
-+
-+/**
-+ * fman_rtc_disable_interupt() - Disable events interrupts
-+ * @regs: Pointer to RTC register block
-+ * @mask: The events to disable
-+ */
-+void fman_rtc_disable_interupt(struct rtc_regs *regs, uint32_t mask);
-+
-+/**
-+ * fman_rtc_get_timer_ctrl() - Get the control register
-+ * @regs: Pointer to RTC register block
-+ *
-+ * Returns: The control register value
-+ */
-+uint32_t fman_rtc_get_timer_ctrl(struct rtc_regs *regs);
-+
-+/**
-+ * fman_rtc_set_timer_ctrl() - Set timer control register
-+ * @regs: Pointer to RTC register block
-+ * @val: The value to set
-+ */
-+void fman_rtc_set_timer_ctrl(struct rtc_regs *regs, uint32_t val);
-+
-+/**
-+ * fman_rtc_get_frequency_compensation() - Get the frequency compensation
-+ * @regs: Pointer to RTC register block
-+ *
-+ * Returns: The timer counter
-+ */
-+uint32_t fman_rtc_get_frequency_compensation(struct rtc_regs *regs);
-+
-+/**
-+ * fman_rtc_set_frequency_compensation() - Set frequency compensation
-+ * @regs: Pointer to RTC register block
-+ * @val: The value to set
-+ */
-+void fman_rtc_set_frequency_compensation(struct rtc_regs *regs, uint32_t val);
-+
-+/**
-+ * fman_rtc_get_trigger_stamp() - Get a trigger stamp
-+ * @regs: Pointer to RTC register block
-+ * @id: The id of the trigger stamp
-+ *
-+ * Returns: The time stamp
-+ */
-+uint64_t fman_rtc_get_trigger_stamp(struct rtc_regs *regs, int id);
-+
-+/**
-+ * fman_rtc_set_timer_alarm_l() - Set timer alarm low register
-+ * @regs: Pointer to RTC register block
-+ * @index: The index of alarm to set
-+ * @val: The value to set
-+ */
-+void fman_rtc_set_timer_alarm_l(struct rtc_regs *regs, int index,
-+ uint32_t val);
-+
-+/**
-+ * fman_rtc_set_timer_alarm() - Set timer alarm
-+ * @regs: Pointer to RTC register block
-+ * @index: The index of alarm to set
-+ * @val: The value to set
-+ */
-+void fman_rtc_set_timer_alarm(struct rtc_regs *regs, int index, int64_t val);
-+
-+/**
-+ * fman_rtc_set_timer_fiper() - Set timer fiper
-+ * @regs: Pointer to RTC register block
-+ * @index: The index of fiper to set
-+ * @val: The value to set
-+ */
-+void fman_rtc_set_timer_fiper(struct rtc_regs *regs, int index, uint32_t val);
-+
-+/**
-+ * fman_rtc_set_timer_offset() - Set timer offset
-+ * @regs: Pointer to RTC register block
-+ * @val: The value to set
-+ */
-+void fman_rtc_set_timer_offset(struct rtc_regs *regs, int64_t val);
-+
-+/**
-+ * fman_rtc_get_timer() - Get the timer counter
-+ * @regs: Pointer to RTC register block
-+ *
-+ * Returns: The timer counter
-+ */
-+static inline uint64_t fman_rtc_get_timer(struct rtc_regs *regs)
-+{
-+ uint64_t time;
-+ /* TMR_CNT_L must be read first to get an accurate value */
-+ time = (uint64_t)ioread32be(&regs->tmr_cnt_l);
-+ time |= ((uint64_t)ioread32be(&regs->tmr_cnt_h) << 32);
-+
-+ return time;
-+}
-+
-+/**
-+ * fman_rtc_set_timer() - Set timer counter
-+ * @regs: Pointer to RTC register block
-+ * @val: The value to set
-+ */
-+static inline void fman_rtc_set_timer(struct rtc_regs *regs, int64_t val)
-+{
-+ iowrite32be((uint32_t)val, &regs->tmr_cnt_l);
-+ iowrite32be((uint32_t)(val >> 32), &regs->tmr_cnt_h);
-+}
-+
-+/**
-+ * fman_rtc_timers_soft_reset() - Soft reset
-+ * @regs: Pointer to RTC register block
-+ *
-+ * Resets all the timer registers and state machines for the 1588 IP and
-+ * the attached client 1588
-+ */
-+void fman_rtc_timers_soft_reset(struct rtc_regs *regs);
-+
-+/**
-+ * fman_rtc_clear_external_trigger() - Clear an external trigger
-+ * @regs: Pointer to RTC register block
-+ * @id: The id of the trigger to clear
-+ */
-+void fman_rtc_clear_external_trigger(struct rtc_regs *regs, int id);
-+
-+/**
-+ * fman_rtc_clear_periodic_pulse() - Clear periodic pulse
-+ * @regs: Pointer to RTC register block
-+ * @id: The id of the fiper to clear
-+ */
-+void fman_rtc_clear_periodic_pulse(struct rtc_regs *regs, int id);
-+
-+/**
-+ * fman_rtc_enable() - Enable RTC hardware block
-+ * @regs: Pointer to RTC register block
-+ */
-+void fman_rtc_enable(struct rtc_regs *regs, bool reset_clock);
-+
-+/**
-+ * fman_rtc_is_enabled() - Is RTC hardware block enabled
-+ * @regs: Pointer to RTC register block
-+ *
-+ * Return: TRUE if enabled
-+ */
-+bool fman_rtc_is_enabled(struct rtc_regs *regs);
-+
-+/**
-+ * fman_rtc_disable() - Disable RTC hardware block
-+ * @regs: Pointer to RTC register block
-+ */
-+void fman_rtc_disable(struct rtc_regs *regs);
-+
-+/**
-+ * fman_rtc_init() - Init RTC hardware block
-+ * @cfg: RTC configuration data
-+ * @regs: Pointer to RTC register block
-+ * @num_alarms: Number of alarms in RTC
-+ * @num_fipers: Number of fipers in RTC
-+ * @num_ext_triggers: Number of external triggers in RTC
-+ * @freq_compensation: Frequency compensation
-+ * @output_clock_divisor: Output clock divisor
-+ *
-+ * This function initializes RTC and applies basic configuration.
-+ */
-+void fman_rtc_init(struct rtc_cfg *cfg, struct rtc_regs *regs, int num_alarms,
-+ int num_fipers, int num_ext_triggers, bool init_freq_comp,
-+ uint32_t freq_compensation, uint32_t output_clock_divisor);
-+
-+/**
-+ * fman_rtc_set_alarm() - Set an alarm
-+ * @regs: Pointer to RTC register block
-+ * @id: id of alarm
-+ * @val: value to write
-+ * @enable: should interrupt be enabled
-+ */
-+void fman_rtc_set_alarm(struct rtc_regs *regs, int id, uint32_t val, bool enable);
-+
-+/**
-+ * fman_rtc_set_periodic_pulse() - Set an alarm
-+ * @regs: Pointer to RTC register block
-+ * @id: id of fiper
-+ * @val: value to write
-+ * @enable: should interrupt be enabled
-+ */
-+void fman_rtc_set_periodic_pulse(struct rtc_regs *regs, int id, uint32_t val,
-+ bool enable);
-+
-+/**
-+ * fman_rtc_set_ext_trigger() - Set an external trigger
-+ * @regs: Pointer to RTC register block
-+ * @id: id of trigger
-+ * @enable: should interrupt be enabled
-+ * @use_pulse_as_input: use the pulse as input
-+ */
-+void fman_rtc_set_ext_trigger(struct rtc_regs *regs, int id, bool enable,
-+ bool use_pulse_as_input);
-+
-+struct fm_rtc_alarm_params {
-+ uint8_t alarm_id; /**< 0 or 1 */
-+ uint64_t alarm_time; /**< In nanoseconds, the time when the
-+ alarm should go off - must be a
-+ multiple of the RTC period */
-+ void (*f_alarm_callback)(void* app, uint8_t id); /**< This routine will
-+ be called when RTC reaches alarmTime */
-+ bool clear_on_expiration; /**< TRUE to turn off the alarm once
-+ expired.*/
-+};
-+
-+struct fm_rtc_periodic_pulse_params {
-+ uint8_t periodic_pulse_id; /**< 0 or 1 */
-+ uint64_t periodic_pulse_period; /**< In Nanoseconds. Must be a multiple
-+ of the RTC period */
-+ void (*f_periodic_pulse_callback)(void* app, uint8_t id); /**< This
-+ routine will be called every
-+ periodicPulsePeriod. */
-+};
-+
-+#endif /* __FSL_FMAN_RTC_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/flib/fsl_fman_sp.h
-@@ -0,0 +1,138 @@
-+/*
-+ * Copyright 2013 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#ifndef __FSL_FMAN_SP_H
-+#define __FSL_FMAN_SP_H
-+
-+#include "common/general.h"
-+#include "fsl_fman.h"
-+
-+
-+struct fm_pcd_storage_profile_regs{
-+ uint32_t fm_sp_ebmpi[8];
-+ /*offset 0 - 0xc*/
-+ /**< Buffer Manager pool Information */
-+
-+ uint32_t fm_sp_acnt; /*offset 0x20*/
-+ uint32_t fm_sp_ebm; /*offset 0x24*/
-+ uint32_t fm_sp_da; /*offset 0x28*/
-+ uint32_t fm_sp_icp; /*offset 0x2c*/
-+ uint32_t fm_sp_mpd; /*offset 0x30*/
-+ uint32_t res1[2]; /*offset 0x34 - 0x38*/
-+ uint32_t fm_sp_spliodn; /*offset 0x3c*/
-+};
-+
-+/**************************************************************************//**
-+ @Description structure for defining internal context copying
-+*//***************************************************************************/
-+struct fman_sp_int_context_data_copy{
-+ uint16_t ext_buf_offset; /**< Offset in External buffer to which
-+ internal context is copied to (Rx)
-+ or taken from (Tx, Op). */
-+ uint8_t int_context_offset; /**< Offset within internal context to copy
-+ from (Rx) or to copy to (Tx, Op).*/
-+ uint16_t size; /**< Internal offset size to be copied */
-+};
-+
-+/**************************************************************************//**
-+ @Description struct for defining external buffer margins
-+*//***************************************************************************/
-+struct fman_sp_buf_margins{
-+ uint16_t start_margins; /**< Number of bytes to be left at the
-+ beginning of the external buffer (must be
-+ divisible by 16) */
-+ uint16_t end_margins; /**< number of bytes to be left at the end of
-+ the external buffer(must be divisible by 16)*/
-+};
-+
-+struct fm_storage_profile_params {
-+ struct fman_ext_pools fm_ext_pools;
-+ struct fman_backup_bm_pools backup_pools;
-+ struct fman_sp_int_context_data_copy *int_context;
-+ struct fman_sp_buf_margins *buf_margins;
-+ enum fman_dma_swap_option dma_swap_data;
-+ enum fman_dma_cache_option int_context_cache_attr;
-+ enum fman_dma_cache_option header_cache_attr;
-+ enum fman_dma_cache_option scatter_gather_cache_attr;
-+ bool dma_write_optimize;
-+ uint16_t liodn_offset;
-+ bool no_scather_gather;
-+ struct fman_buf_pool_depletion buf_pool_depletion;
-+};
-+
-+/**************************************************************************//**
-+ @Description Registers bit fields
-+*//***************************************************************************/
-+#define FMAN_SP_EXT_BUF_POOL_EN_COUNTER 0x40000000
-+#define FMAN_SP_EXT_BUF_POOL_VALID 0x80000000
-+#define FMAN_SP_EXT_BUF_POOL_BACKUP 0x20000000
-+#define FMAN_SP_DMA_ATTR_WRITE_OPTIMIZE 0x00100000
-+#define FMAN_SP_SG_DISABLE 0x80000000
-+
-+/* shifts */
-+#define FMAN_SP_EXT_BUF_POOL_ID_SHIFT 16
-+#define FMAN_SP_POOL_DEP_NUM_OF_POOLS_SHIFT 16
-+#define FMAN_SP_EXT_BUF_MARG_START_SHIFT 16
-+#define FMAN_SP_EXT_BUF_MARG_END_SHIFT 0
-+#define FMAN_SP_DMA_ATTR_SWP_SHIFT 30
-+#define FMAN_SP_DMA_ATTR_IC_CACHE_SHIFT 28
-+#define FMAN_SP_DMA_ATTR_HDR_CACHE_SHIFT 26
-+#define FMAN_SP_DMA_ATTR_SG_CACHE_SHIFT 24
-+#define FMAN_SP_IC_TO_EXT_SHIFT 16
-+#define FMAN_SP_IC_FROM_INT_SHIFT 8
-+#define FMAN_SP_IC_SIZE_SHIFT 0
-+
-+/**************************************************************************//**
-+ @Description defaults
-+*//***************************************************************************/
-+#define DEFAULT_FMAN_SP_DMA_SWAP_DATA FMAN_DMA_NO_SWP
-+#define DEFAULT_FMAN_SP_DMA_INT_CONTEXT_CACHE_ATTR FMAN_DMA_NO_STASH
-+#define DEFAULT_FMAN_SP_DMA_HEADER_CACHE_ATTR FMAN_DMA_NO_STASH
-+#define DEFAULT_FMAN_SP_DMA_SCATTER_GATHER_CACHE_ATTR FMAN_DMA_NO_STASH
-+#define DEFAULT_FMAN_SP_DMA_WRITE_OPTIMIZE TRUE
-+#define DEFAULT_FMAN_SP_NO_SCATTER_GATHER FALSE
-+
-+void fman_vsp_defconfig(struct fm_storage_profile_params *cfg);
-+
-+void fman_vsp_init(struct fm_pcd_storage_profile_regs *regs,
-+ uint16_t index, struct fm_storage_profile_params *fm_vsp_params,
-+ int port_max_num_of_ext_pools, int bm_max_num_of_pools,
-+ int max_num_of_pfc_priorities);
-+
-+uint32_t fman_vsp_get_statistics(struct fm_pcd_storage_profile_regs *regs,
-+ uint16_t index);
-+
-+void fman_vsp_set_statistics(struct fm_pcd_storage_profile_regs *regs,
-+ uint16_t index, uint32_t value);
-+
-+
-+#endif /* __FSL_FMAN_SP_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/flib/fsl_fman_tgec.h
-@@ -0,0 +1,479 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#ifndef __FSL_FMAN_TGEC_H
-+#define __FSL_FMAN_TGEC_H
-+
-+#include "common/general.h"
-+#include "fsl_enet.h"
-+
-+
-+/* Transmit Inter-Packet Gap Length Register (TX_IPG_LENGTH) */
-+#define TGEC_TX_IPG_LENGTH_MASK 0x000003ff
-+
-+enum tgec_counters {
-+ E_TGEC_COUNTER_R64,
-+ E_TGEC_COUNTER_R127,
-+ E_TGEC_COUNTER_R255,
-+ E_TGEC_COUNTER_R511,
-+ E_TGEC_COUNTER_R1023,
-+ E_TGEC_COUNTER_R1518,
-+ E_TGEC_COUNTER_R1519X,
-+ E_TGEC_COUNTER_TRFRG,
-+ E_TGEC_COUNTER_TRJBR,
-+ E_TGEC_COUNTER_RDRP,
-+ E_TGEC_COUNTER_RALN,
-+ E_TGEC_COUNTER_TRUND,
-+ E_TGEC_COUNTER_TROVR,
-+ E_TGEC_COUNTER_RXPF,
-+ E_TGEC_COUNTER_TXPF,
-+ E_TGEC_COUNTER_ROCT,
-+ E_TGEC_COUNTER_RMCA,
-+ E_TGEC_COUNTER_RBCA,
-+ E_TGEC_COUNTER_RPKT,
-+ E_TGEC_COUNTER_RUCA,
-+ E_TGEC_COUNTER_RERR,
-+ E_TGEC_COUNTER_TOCT,
-+ E_TGEC_COUNTER_TMCA,
-+ E_TGEC_COUNTER_TBCA,
-+ E_TGEC_COUNTER_TUCA,
-+ E_TGEC_COUNTER_TERR
-+};
-+
-+/* Command and Configuration Register (COMMAND_CONFIG) */
-+#define CMD_CFG_EN_TIMESTAMP 0x00100000
-+#define CMD_CFG_TX_ADDR_INS_SEL 0x00080000
-+#define CMD_CFG_NO_LEN_CHK 0x00020000
-+#define CMD_CFG_SEND_IDLE 0x00010000
-+#define CMD_CFG_RX_ER_DISC 0x00004000
-+#define CMD_CFG_CMD_FRM_EN 0x00002000
-+#define CMD_CFG_STAT_CLR 0x00001000
-+#define CMD_CFG_LOOPBACK_EN 0x00000400
-+#define CMD_CFG_TX_ADDR_INS 0x00000200
-+#define CMD_CFG_PAUSE_IGNORE 0x00000100
-+#define CMD_CFG_PAUSE_FWD 0x00000080
-+#define CMD_CFG_PROMIS_EN 0x00000010
-+#define CMD_CFG_WAN_MODE 0x00000008
-+#define CMD_CFG_RX_EN 0x00000002
-+#define CMD_CFG_TX_EN 0x00000001
-+
-+/* Interrupt Mask Register (IMASK) */
-+#define TGEC_IMASK_MDIO_SCAN_EVENT 0x00010000
-+#define TGEC_IMASK_MDIO_CMD_CMPL 0x00008000
-+#define TGEC_IMASK_REM_FAULT 0x00004000
-+#define TGEC_IMASK_LOC_FAULT 0x00002000
-+#define TGEC_IMASK_TX_ECC_ER 0x00001000
-+#define TGEC_IMASK_TX_FIFO_UNFL 0x00000800
-+#define TGEC_IMASK_TX_FIFO_OVFL 0x00000400
-+#define TGEC_IMASK_TX_ER 0x00000200
-+#define TGEC_IMASK_RX_FIFO_OVFL 0x00000100
-+#define TGEC_IMASK_RX_ECC_ER 0x00000080
-+#define TGEC_IMASK_RX_JAB_FRM 0x00000040
-+#define TGEC_IMASK_RX_OVRSZ_FRM 0x00000020
-+#define TGEC_IMASK_RX_RUNT_FRM 0x00000010
-+#define TGEC_IMASK_RX_FRAG_FRM 0x00000008
-+#define TGEC_IMASK_RX_LEN_ER 0x00000004
-+#define TGEC_IMASK_RX_CRC_ER 0x00000002
-+#define TGEC_IMASK_RX_ALIGN_ER 0x00000001
-+
-+#define TGEC_EVENTS_MASK \
-+ ((uint32_t)(TGEC_IMASK_MDIO_SCAN_EVENT | \
-+ TGEC_IMASK_MDIO_CMD_CMPL | \
-+ TGEC_IMASK_REM_FAULT | \
-+ TGEC_IMASK_LOC_FAULT | \
-+ TGEC_IMASK_TX_ECC_ER | \
-+ TGEC_IMASK_TX_FIFO_UNFL | \
-+ TGEC_IMASK_TX_FIFO_OVFL | \
-+ TGEC_IMASK_TX_ER | \
-+ TGEC_IMASK_RX_FIFO_OVFL | \
-+ TGEC_IMASK_RX_ECC_ER | \
-+ TGEC_IMASK_RX_JAB_FRM | \
-+ TGEC_IMASK_RX_OVRSZ_FRM | \
-+ TGEC_IMASK_RX_RUNT_FRM | \
-+ TGEC_IMASK_RX_FRAG_FRM | \
-+ TGEC_IMASK_RX_LEN_ER | \
-+ TGEC_IMASK_RX_CRC_ER | \
-+ TGEC_IMASK_RX_ALIGN_ER))
-+
-+/* Hashtable Control Register (HASHTABLE_CTRL) */
-+#define TGEC_HASH_MCAST_SHIFT 23
-+#define TGEC_HASH_MCAST_EN 0x00000200
-+#define TGEC_HASH_ADR_MSK 0x000001ff
-+
-+#define DEFAULT_WAN_MODE_ENABLE FALSE
-+#define DEFAULT_PROMISCUOUS_MODE_ENABLE FALSE
-+#define DEFAULT_PAUSE_FORWARD_ENABLE FALSE
-+#define DEFAULT_PAUSE_IGNORE FALSE
-+#define DEFAULT_TX_ADDR_INS_ENABLE FALSE
-+#define DEFAULT_LOOPBACK_ENABLE FALSE
-+#define DEFAULT_CMD_FRAME_ENABLE FALSE
-+#define DEFAULT_RX_ERROR_DISCARD FALSE
-+#define DEFAULT_SEND_IDLE_ENABLE FALSE
-+#define DEFAULT_NO_LENGTH_CHECK_ENABLE TRUE
-+#define DEFAULT_LGTH_CHECK_NOSTDR FALSE
-+#define DEFAULT_TIME_STAMP_ENABLE FALSE
-+#define DEFAULT_TX_IPG_LENGTH 12
-+#define DEFAULT_MAX_FRAME_LENGTH 0x600
-+#define DEFAULT_PAUSE_QUANT 0xf000
-+
-+/*
-+ * 10G memory map
-+ */
-+struct tgec_regs {
-+ uint32_t tgec_id; /* 0x000 Controller ID */
-+ uint32_t reserved001[1]; /* 0x004 */
-+ uint32_t command_config; /* 0x008 Control and configuration */
-+ uint32_t mac_addr_0; /* 0x00c Lower 32 bits of the MAC adr */
-+ uint32_t mac_addr_1; /* 0x010 Upper 16 bits of the MAC adr */
-+ uint32_t maxfrm; /* 0x014 Maximum frame length */
-+ uint32_t pause_quant; /* 0x018 Pause quanta */
-+ uint32_t rx_fifo_sections; /* 0x01c */
-+ uint32_t tx_fifo_sections; /* 0x020 */
-+ uint32_t rx_fifo_almost_f_e; /* 0x024 */
-+ uint32_t tx_fifo_almost_f_e; /* 0x028 */
-+ uint32_t hashtable_ctrl; /* 0x02c Hash table control*/
-+ uint32_t mdio_cfg_status; /* 0x030 */
-+ uint32_t mdio_command; /* 0x034 */
-+ uint32_t mdio_data; /* 0x038 */
-+ uint32_t mdio_regaddr; /* 0x03c */
-+ uint32_t status; /* 0x040 */
-+ uint32_t tx_ipg_len; /* 0x044 Transmitter inter-packet-gap */
-+ uint32_t mac_addr_2; /* 0x048 Lower 32 bits of 2nd MAC adr */
-+ uint32_t mac_addr_3; /* 0x04c Upper 16 bits of 2nd MAC adr */
-+ uint32_t rx_fifo_ptr_rd; /* 0x050 */
-+ uint32_t rx_fifo_ptr_wr; /* 0x054 */
-+ uint32_t tx_fifo_ptr_rd; /* 0x058 */
-+ uint32_t tx_fifo_ptr_wr; /* 0x05c */
-+ uint32_t imask; /* 0x060 Interrupt mask */
-+ uint32_t ievent; /* 0x064 Interrupt event */
-+ uint32_t udp_port; /* 0x068 Defines a UDP Port number */
-+ uint32_t type_1588v2; /* 0x06c Type field for 1588v2 */
-+ uint32_t reserved070[4]; /* 0x070 */
-+ /*10Ge Statistics Counter */
-+ uint32_t tfrm_u; /* 80 aFramesTransmittedOK */
-+ uint32_t tfrm_l; /* 84 aFramesTransmittedOK */
-+ uint32_t rfrm_u; /* 88 aFramesReceivedOK */
-+ uint32_t rfrm_l; /* 8c aFramesReceivedOK */
-+ uint32_t rfcs_u; /* 90 aFrameCheckSequenceErrors */
-+ uint32_t rfcs_l; /* 94 aFrameCheckSequenceErrors */
-+ uint32_t raln_u; /* 98 aAlignmentErrors */
-+ uint32_t raln_l; /* 9c aAlignmentErrors */
-+ uint32_t txpf_u; /* A0 aPAUSEMACCtrlFramesTransmitted */
-+ uint32_t txpf_l; /* A4 aPAUSEMACCtrlFramesTransmitted */
-+ uint32_t rxpf_u; /* A8 aPAUSEMACCtrlFramesReceived */
-+ uint32_t rxpf_l; /* Ac aPAUSEMACCtrlFramesReceived */
-+ uint32_t rlong_u; /* B0 aFrameTooLongErrors */
-+ uint32_t rlong_l; /* B4 aFrameTooLongErrors */
-+ uint32_t rflr_u; /* B8 aInRangeLengthErrors */
-+ uint32_t rflr_l; /* Bc aInRangeLengthErrors */
-+ uint32_t tvlan_u; /* C0 VLANTransmittedOK */
-+ uint32_t tvlan_l; /* C4 VLANTransmittedOK */
-+ uint32_t rvlan_u; /* C8 VLANReceivedOK */
-+ uint32_t rvlan_l; /* Cc VLANReceivedOK */
-+ uint32_t toct_u; /* D0 ifOutOctets */
-+ uint32_t toct_l; /* D4 ifOutOctets */
-+ uint32_t roct_u; /* D8 ifInOctets */
-+ uint32_t roct_l; /* Dc ifInOctets */
-+ uint32_t ruca_u; /* E0 ifInUcastPkts */
-+ uint32_t ruca_l; /* E4 ifInUcastPkts */
-+ uint32_t rmca_u; /* E8 ifInMulticastPkts */
-+ uint32_t rmca_l; /* Ec ifInMulticastPkts */
-+ uint32_t rbca_u; /* F0 ifInBroadcastPkts */
-+ uint32_t rbca_l; /* F4 ifInBroadcastPkts */
-+ uint32_t terr_u; /* F8 ifOutErrors */
-+ uint32_t terr_l; /* Fc ifOutErrors */
-+ uint32_t reserved100[2]; /* 100-108*/
-+ uint32_t tuca_u; /* 108 ifOutUcastPkts */
-+ uint32_t tuca_l; /* 10c ifOutUcastPkts */
-+ uint32_t tmca_u; /* 110 ifOutMulticastPkts */
-+ uint32_t tmca_l; /* 114 ifOutMulticastPkts */
-+ uint32_t tbca_u; /* 118 ifOutBroadcastPkts */
-+ uint32_t tbca_l; /* 11c ifOutBroadcastPkts */
-+ uint32_t rdrp_u; /* 120 etherStatsDropEvents */
-+ uint32_t rdrp_l; /* 124 etherStatsDropEvents */
-+ uint32_t reoct_u; /* 128 etherStatsOctets */
-+ uint32_t reoct_l; /* 12c etherStatsOctets */
-+ uint32_t rpkt_u; /* 130 etherStatsPkts */
-+ uint32_t rpkt_l; /* 134 etherStatsPkts */
-+ uint32_t trund_u; /* 138 etherStatsUndersizePkts */
-+ uint32_t trund_l; /* 13c etherStatsUndersizePkts */
-+ uint32_t r64_u; /* 140 etherStatsPkts64Octets */
-+ uint32_t r64_l; /* 144 etherStatsPkts64Octets */
-+ uint32_t r127_u; /* 148 etherStatsPkts65to127Octets */
-+ uint32_t r127_l; /* 14c etherStatsPkts65to127Octets */
-+ uint32_t r255_u; /* 150 etherStatsPkts128to255Octets */
-+ uint32_t r255_l; /* 154 etherStatsPkts128to255Octets */
-+ uint32_t r511_u; /* 158 etherStatsPkts256to511Octets */
-+ uint32_t r511_l; /* 15c etherStatsPkts256to511Octets */
-+ uint32_t r1023_u; /* 160 etherStatsPkts512to1023Octets */
-+ uint32_t r1023_l; /* 164 etherStatsPkts512to1023Octets */
-+ uint32_t r1518_u; /* 168 etherStatsPkts1024to1518Octets */
-+ uint32_t r1518_l; /* 16c etherStatsPkts1024to1518Octets */
-+ uint32_t r1519x_u; /* 170 etherStatsPkts1519toX */
-+ uint32_t r1519x_l; /* 174 etherStatsPkts1519toX */
-+ uint32_t trovr_u; /* 178 etherStatsOversizePkts */
-+ uint32_t trovr_l; /* 17c etherStatsOversizePkts */
-+ uint32_t trjbr_u; /* 180 etherStatsJabbers */
-+ uint32_t trjbr_l; /* 184 etherStatsJabbers */
-+ uint32_t trfrg_u; /* 188 etherStatsFragments */
-+ uint32_t trfrg_l; /* 18C etherStatsFragments */
-+ uint32_t rerr_u; /* 190 ifInErrors */
-+ uint32_t rerr_l; /* 194 ifInErrors */
-+};
-+
-+/**
-+ * struct tgec_cfg - TGEC configuration
-+ *
-+ * @rx_error_discard: Receive Erroneous Frame Discard Enable. When set to 1
-+ * any frame received with an error is discarded in the
-+ * Core and not forwarded to the Client interface.
-+ * When set to 0 (Reset value), erroneous Frames are
-+ * forwarded to the Client interface with ff_rx_err
-+ * asserted.
-+ * @pause_ignore: Ignore Pause Frame Quanta. If set to 1 received pause
-+ * frames are ignored by the MAC. When set to 0
-+ * (Reset value) the transmit process is stopped for the
-+ * amount of time specified in the pause quanta received
-+ * within a pause frame.
-+ * @pause_forward_enable:
-+ * Terminate / Forward Pause Frames. If set to 1 pause
-+ * frames are forwarded to the user application. When set
-+ * to 0 (Reset value) pause frames are terminated and
-+ * discarded within the MAC.
-+ * @no_length_check_enable:
-+ * Payload Length Check Disable. When set to 0
-+ * (Reset value), the Core checks the frame's payload
-+ * length with the Frame Length/Type field, when set to 1
-+ * the payload length check is disabled.
-+ * @cmd_frame_enable: Enables reception of all command frames. When set to 1
-+ * all Command Frames are accepted, when set to 0
-+ * (Reset Value) only Pause Frames are accepted and all
-+ * other Command Frames are rejected.
-+ * @send_idle_enable: Force Idle Generation. When set to 1, the MAC
-+ * permanently sends XGMII Idle sequences even when faults
-+ * are received.
-+ * @wan_mode_enable: WAN Mode Enable. Sets WAN mode (1) or LAN mode
-+ * (0, default) of operation.
-+ * @promiscuous_mode_enable:
-+ * Enables MAC promiscuous operation. When set to 1, all
-+ * frames are received without any MAC address filtering,
-+ * when set to 0 (Reset value) Unicast Frames with a
-+ * destination address not matching the Core MAC Address
-+ * (MAC Address programmed in Registers MAC_ADDR_0 and
-+ * MAC_ADDR_1 or the MAC address programmed in Registers
-+ * MAC_ADDR_2 and MAC_ADDR_3) are rejected.
-+ * @tx_addr_ins_enable: Set Source MAC Address on Transmit. If set to 1 the
-+ * MAC overwrites the source MAC address received from the
-+ * Client Interface with one of the MAC addresses. If set
-+ * to 0 (Reset value), the source MAC address from the
-+ * Client Interface is transmitted unmodified to the line.
-+ * @loopback_enable: PHY Interface Loopback. When set to 1, the signal
-+ * loop_ena is set to '1', when set to 0 (Reset value)
-+ * the signal loop_ena is set to 0.
-+ * @lgth_check_nostdr: The Core interprets the Length/Type field differently
-+ * depending on the value of this Bit
-+ * @time_stamp_enable: This bit selects between enabling and disabling the
-+ * IEEE 1588 functionality. 1: IEEE 1588 is enabled
-+ * 0: IEEE 1588 is disabled
-+ * @max_frame_length: Maximum supported received frame length.
-+ * The 10GEC MAC supports reception of any frame size up
-+ * to 16,352 bytes (0x3FE0). Typical settings are
-+ * 0x05EE (1,518 bytes) for standard frames.
-+ * Default setting is 0x0600 (1,536 bytes).
-+ * Received frames that exceed this stated maximum
-+ * are truncated.
-+ * @pause_quant: Pause quanta value used with transmitted pause frames.
-+ * Each quanta represents a 512 bit-times.
-+ * @tx_ipg_length: Transmit Inter-Packet-Gap (IPG) value. A 6-bit value:
-+ * Depending on LAN or WAN mode of operation the value has
-+ * the following meaning: - LAN Mode: Number of octets in
-+ * steps of 4. Valid values are 8, 12, 16, ... 100. DIC is
-+ * fully supported (see 10.6.1 page 49) for any setting. A
-+ * default of 12 (reset value) must be set to conform to
-+ * IEEE802.3ae. Warning: When set to 8, PCS layers may not
-+ * be able to perform clock rate compensation. - WAN Mode:
-+ * Stretch factor. Valid values are 4..15. The stretch
-+ * factor is calculated as (value+1)*8. A default of 12
-+ * (reset value) must be set to conform to IEEE 802.3ae
-+ * (i.e. 13*8=104). A larger value shrinks the IPG
-+ * (increasing bandwidth).
-+ *
-+ * This structure contains basic TGEC configuration and must be passed to
-+ * fman_tgec_init() function. A default set of configuration values can be
-+ * obtained by calling fman_tgec_defconfig().
-+ */
-+struct tgec_cfg {
-+ bool rx_error_discard;
-+ bool pause_ignore;
-+ bool pause_forward_enable;
-+ bool no_length_check_enable;
-+ bool cmd_frame_enable;
-+ bool send_idle_enable;
-+ bool wan_mode_enable;
-+ bool promiscuous_mode_enable;
-+ bool tx_addr_ins_enable;
-+ bool loopback_enable;
-+ bool lgth_check_nostdr;
-+ bool time_stamp_enable;
-+ uint16_t max_frame_length;
-+ uint16_t pause_quant;
-+ uint32_t tx_ipg_length;
-+ bool skip_fman11_workaround;
-+};
-+
-+
-+void fman_tgec_defconfig(struct tgec_cfg *cfg);
-+
-+/**
-+ * fman_tgec_init() - Init tgec hardware block
-+ * @regs: Pointer to tgec register block
-+ * @cfg: tgec configuration data
-+ * @exceptions_mask: initial exceptions mask
-+ *
-+ * This function initializes the tgec controller and applies its
-+ * basic configuration.
-+ *
-+ * Returns: 0 if successful, an error code otherwise.
-+ */
-+
-+int fman_tgec_init(struct tgec_regs *regs, struct tgec_cfg *cfg,
-+ uint32_t exception_mask);
-+
-+void fman_tgec_enable(struct tgec_regs *regs, bool apply_rx, bool apply_tx);
-+
-+void fman_tgec_disable(struct tgec_regs *regs, bool apply_rx, bool apply_tx);
-+
-+uint32_t fman_tgec_get_revision(struct tgec_regs *regs);
-+
-+void fman_tgec_set_mac_address(struct tgec_regs *regs, uint8_t *macaddr);
-+
-+void fman_tgec_set_promiscuous(struct tgec_regs *regs, bool val);
-+
-+/**
-+ * fman_tgec_reset_stat() - Completely resets all TGEC HW counters
-+ * @regs: Pointer to TGEC register block
-+ */
-+void fman_tgec_reset_stat(struct tgec_regs *regs);
-+
-+/**
-+ * fman_tgec_get_counter() - Reads TGEC HW counters
-+ * @regs: Pointer to TGEC register block
-+ * @reg_name: Counter name according to the appropriate enum
-+ *
-+ * Returns: Required counter value
-+ */
-+uint64_t fman_tgec_get_counter(struct tgec_regs *regs,
-+ enum tgec_counters reg_name);
-+
-+/**
-+ * fman_tgec_set_hash_table() - Sets the Hashtable Control Register
-+ * @regs: Pointer to TGEC register block
-+ * @value: Value to be written in Hashtable Control Register
-+ */
-+void fman_tgec_set_hash_table(struct tgec_regs *regs, uint32_t value);
-+
-+/**
-+ * fman_tgec_set_tx_pause_frames() - Sets the Pause Quanta Register
-+ * @regs: Pointer to TGEC register block
-+ * @pause_time: Pause quanta value used with transmitted pause frames.
-+ * Each quanta represents a 512 bit-times
-+ */
-+void fman_tgec_set_tx_pause_frames(struct tgec_regs *regs, uint16_t pause_time);
-+
-+/**
-+ * fman_tgec_set_rx_ignore_pause_frames() - Changes the policy WRT pause frames
-+ * @regs: Pointer to TGEC register block
-+ * @en: Ignore/Respond to pause frame quanta
-+ *
-+ * Sets the value of PAUSE_IGNORE field in the COMMAND_CONFIG Register
-+ * 0 - MAC stops transmit process for the duration specified
-+ * in the Pause frame quanta of a received Pause frame.
-+ * 1 - MAC ignores received Pause frames.
-+ */
-+void fman_tgec_set_rx_ignore_pause_frames(struct tgec_regs *regs, bool en);
-+
-+/**
-+ * fman_tgec_enable_1588_time_stamp() - change timestamp functionality
-+ * @regs: Pointer to TGEC register block
-+ * @en: enable/disable timestamp functionality
-+ *
-+ * Sets the value of EN_TIMESTAMP field in the COMMAND_CONFIG Register
-+ * IEEE 1588 timestamp functionality control:
-+ * 0 disabled, 1 enabled
-+ */
-+
-+void fman_tgec_enable_1588_time_stamp(struct tgec_regs *regs, bool en);
-+
-+uint32_t fman_tgec_get_event(struct tgec_regs *regs, uint32_t ev_mask);
-+
-+void fman_tgec_ack_event(struct tgec_regs *regs, uint32_t ev_mask);
-+
-+uint32_t fman_tgec_get_interrupt_mask(struct tgec_regs *regs);
-+
-+/**
-+ * fman_tgec_add_addr_in_paddr() - Sets additional exact match MAC address
-+ * @regs: Pointer to TGEC register block
-+ * @addr_ptr: Pointer to 6-byte array containing the MAC address
-+ *
-+ * Sets the additional station MAC address
-+ */
-+void fman_tgec_add_addr_in_paddr(struct tgec_regs *regs, uint8_t *addr_ptr);
-+
-+void fman_tgec_clear_addr_in_paddr(struct tgec_regs *regs);
-+
-+void fman_tgec_enable_interrupt(struct tgec_regs *regs, uint32_t ev_mask);
-+
-+void fman_tgec_disable_interrupt(struct tgec_regs *regs, uint32_t ev_mask);
-+
-+void fman_tgec_reset_filter_table(struct tgec_regs *regs);
-+
-+void fman_tgec_set_hash_table_entry(struct tgec_regs *regs, uint32_t crc);
-+
-+
-+/**
-+ * fman_tgec_get_max_frame_len() - Returns the maximum frame length value
-+ * @regs: Pointer to TGEC register block
-+ */
-+uint16_t fman_tgec_get_max_frame_len(struct tgec_regs *regs);
-+
-+/**
-+ * fman_tgec_set_erratum_tx_fifo_corruption_10gmac_a007() - Initialize the
-+ * main tgec configuration parameters
-+ * @regs: Pointer to TGEC register block
-+ *
-+ * TODO
-+ */
-+void fman_tgec_set_erratum_tx_fifo_corruption_10gmac_a007(struct tgec_regs
-+ *regs);
-+
-+
-+#endif /* __FSL_FMAN_TGEC_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/integrations/FMANV3H/dpaa_integration_ext.h
-@@ -0,0 +1,291 @@
-+/*
-+ * Copyright 2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+/**
-+
-+ @File dpaa_integration_ext.h
-+
-+ @Description T4240 FM external definitions and structures.
-+*//***************************************************************************/
-+#ifndef __DPAA_INTEGRATION_EXT_H
-+#define __DPAA_INTEGRATION_EXT_H
-+
-+#include "std_ext.h"
-+
-+
-+#define DPAA_VERSION 11
-+
-+/**************************************************************************//**
-+ @Description DPAA SW Portals Enumeration.
-+*//***************************************************************************/
-+typedef enum
-+{
-+ e_DPAA_SWPORTAL0 = 0,
-+ e_DPAA_SWPORTAL1,
-+ e_DPAA_SWPORTAL2,
-+ e_DPAA_SWPORTAL3,
-+ e_DPAA_SWPORTAL4,
-+ e_DPAA_SWPORTAL5,
-+ e_DPAA_SWPORTAL6,
-+ e_DPAA_SWPORTAL7,
-+ e_DPAA_SWPORTAL8,
-+ e_DPAA_SWPORTAL9,
-+ e_DPAA_SWPORTAL10,
-+ e_DPAA_SWPORTAL11,
-+ e_DPAA_SWPORTAL12,
-+ e_DPAA_SWPORTAL13,
-+ e_DPAA_SWPORTAL14,
-+ e_DPAA_SWPORTAL15,
-+ e_DPAA_SWPORTAL16,
-+ e_DPAA_SWPORTAL17,
-+ e_DPAA_SWPORTAL18,
-+ e_DPAA_SWPORTAL19,
-+ e_DPAA_SWPORTAL20,
-+ e_DPAA_SWPORTAL21,
-+ e_DPAA_SWPORTAL22,
-+ e_DPAA_SWPORTAL23,
-+ e_DPAA_SWPORTAL24,
-+ e_DPAA_SWPORTAL_DUMMY_LAST
-+} e_DpaaSwPortal;
-+
-+/**************************************************************************//**
-+ @Description DPAA Direct Connect Portals Enumeration.
-+*//***************************************************************************/
-+typedef enum
-+{
-+ e_DPAA_DCPORTAL0 = 0,
-+ e_DPAA_DCPORTAL1,
-+ e_DPAA_DCPORTAL2,
-+ e_DPAA_DCPORTAL_DUMMY_LAST
-+} e_DpaaDcPortal;
-+
-+#define DPAA_MAX_NUM_OF_SW_PORTALS e_DPAA_SWPORTAL_DUMMY_LAST
-+#define DPAA_MAX_NUM_OF_DC_PORTALS e_DPAA_DCPORTAL_DUMMY_LAST
-+
-+/*****************************************************************************
-+ QMan INTEGRATION-SPECIFIC DEFINITIONS
-+******************************************************************************/
-+#define QM_MAX_NUM_OF_POOL_CHANNELS 15 /**< Total number of channels, dedicated and pool */
-+#define QM_MAX_NUM_OF_WQ 8 /**< Number of work queues per channel */
-+#define QM_MAX_NUM_OF_CGS 256 /**< Congestion groups number */
-+#define QM_MAX_NUM_OF_FQIDS (16 * MEGABYTE)
-+ /**< FQIDs range - 24 bits */
-+
-+/**************************************************************************//**
-+ @Description Work Queue Channel assignments in QMan.
-+*//***************************************************************************/
-+typedef enum
-+{
-+ e_QM_FQ_CHANNEL_SWPORTAL0 = 0x0, /**< Dedicated channels serviced by software portals 0 to 24 */
-+ e_QM_FQ_CHANNEL_SWPORTAL1,
-+ e_QM_FQ_CHANNEL_SWPORTAL2,
-+ e_QM_FQ_CHANNEL_SWPORTAL3,
-+ e_QM_FQ_CHANNEL_SWPORTAL4,
-+ e_QM_FQ_CHANNEL_SWPORTAL5,
-+ e_QM_FQ_CHANNEL_SWPORTAL6,
-+ e_QM_FQ_CHANNEL_SWPORTAL7,
-+ e_QM_FQ_CHANNEL_SWPORTAL8,
-+ e_QM_FQ_CHANNEL_SWPORTAL9,
-+ e_QM_FQ_CHANNEL_SWPORTAL10,
-+ e_QM_FQ_CHANNEL_SWPORTAL11,
-+ e_QM_FQ_CHANNEL_SWPORTAL12,
-+ e_QM_FQ_CHANNEL_SWPORTAL13,
-+ e_QM_FQ_CHANNEL_SWPORTAL14,
-+ e_QM_FQ_CHANNEL_SWPORTAL15,
-+ e_QM_FQ_CHANNEL_SWPORTAL16,
-+ e_QM_FQ_CHANNEL_SWPORTAL17,
-+ e_QM_FQ_CHANNEL_SWPORTAL18,
-+ e_QM_FQ_CHANNEL_SWPORTAL19,
-+ e_QM_FQ_CHANNEL_SWPORTAL20,
-+ e_QM_FQ_CHANNEL_SWPORTAL21,
-+ e_QM_FQ_CHANNEL_SWPORTAL22,
-+ e_QM_FQ_CHANNEL_SWPORTAL23,
-+ e_QM_FQ_CHANNEL_SWPORTAL24,
-+
-+ e_QM_FQ_CHANNEL_POOL1 = 0x401, /**< Pool channels that can be serviced by any of the software portals */
-+ e_QM_FQ_CHANNEL_POOL2,
-+ e_QM_FQ_CHANNEL_POOL3,
-+ e_QM_FQ_CHANNEL_POOL4,
-+ e_QM_FQ_CHANNEL_POOL5,
-+ e_QM_FQ_CHANNEL_POOL6,
-+ e_QM_FQ_CHANNEL_POOL7,
-+ e_QM_FQ_CHANNEL_POOL8,
-+ e_QM_FQ_CHANNEL_POOL9,
-+ e_QM_FQ_CHANNEL_POOL10,
-+ e_QM_FQ_CHANNEL_POOL11,
-+ e_QM_FQ_CHANNEL_POOL12,
-+ e_QM_FQ_CHANNEL_POOL13,
-+ e_QM_FQ_CHANNEL_POOL14,
-+ e_QM_FQ_CHANNEL_POOL15,
-+
-+ e_QM_FQ_CHANNEL_FMAN0_SP0 = 0x800, /**< Dedicated channels serviced by Direct Connect Portal 0:
-+ connected to FMan 0; assigned in incrementing order to
-+ each sub-portal (SP) in the portal */
-+ e_QM_FQ_CHANNEL_FMAN0_SP1,
-+ e_QM_FQ_CHANNEL_FMAN0_SP2,
-+ e_QM_FQ_CHANNEL_FMAN0_SP3,
-+ e_QM_FQ_CHANNEL_FMAN0_SP4,
-+ e_QM_FQ_CHANNEL_FMAN0_SP5,
-+ e_QM_FQ_CHANNEL_FMAN0_SP6,
-+ e_QM_FQ_CHANNEL_FMAN0_SP7,
-+ e_QM_FQ_CHANNEL_FMAN0_SP8,
-+ e_QM_FQ_CHANNEL_FMAN0_SP9,
-+ e_QM_FQ_CHANNEL_FMAN0_SP10,
-+ e_QM_FQ_CHANNEL_FMAN0_SP11,
-+ e_QM_FQ_CHANNEL_FMAN0_SP12,
-+ e_QM_FQ_CHANNEL_FMAN0_SP13,
-+ e_QM_FQ_CHANNEL_FMAN0_SP14,
-+ e_QM_FQ_CHANNEL_FMAN0_SP15,
-+
-+ e_QM_FQ_CHANNEL_RMAN_SP0 = 0x820, /**< Dedicated channels serviced by Direct Connect Portal 1: connected to RMan */
-+ e_QM_FQ_CHANNEL_RMAN_SP1,
-+
-+ e_QM_FQ_CHANNEL_CAAM = 0x840 /**< Dedicated channel serviced by Direct Connect Portal 2:
-+ connected to SEC */
-+} e_QmFQChannel;
-+
-+/*****************************************************************************
-+ BMan INTEGRATION-SPECIFIC DEFINITIONS
-+******************************************************************************/
-+#define BM_MAX_NUM_OF_POOLS 64 /**< Number of buffers pools */
-+
-+/*****************************************************************************
-+ SEC INTEGRATION-SPECIFIC DEFINITIONS
-+******************************************************************************/
-+#define SEC_NUM_OF_DECOS 3
-+#define SEC_ALL_DECOS_MASK 0x00000003
-+
-+
-+/*****************************************************************************
-+ FM INTEGRATION-SPECIFIC DEFINITIONS
-+******************************************************************************/
-+#define INTG_MAX_NUM_OF_FM 2
-+/* Ports defines */
-+#define FM_MAX_NUM_OF_1G_MACS 6
-+#define FM_MAX_NUM_OF_10G_MACS 2
-+#define FM_MAX_NUM_OF_MACS (FM_MAX_NUM_OF_1G_MACS + FM_MAX_NUM_OF_10G_MACS)
-+#define FM_MAX_NUM_OF_OH_PORTS 6
-+
-+#define FM_MAX_NUM_OF_1G_RX_PORTS FM_MAX_NUM_OF_1G_MACS
-+#define FM_MAX_NUM_OF_10G_RX_PORTS FM_MAX_NUM_OF_10G_MACS
-+#define FM_MAX_NUM_OF_RX_PORTS (FM_MAX_NUM_OF_10G_RX_PORTS + FM_MAX_NUM_OF_1G_RX_PORTS)
-+
-+#define FM_MAX_NUM_OF_1G_TX_PORTS FM_MAX_NUM_OF_1G_MACS
-+#define FM_MAX_NUM_OF_10G_TX_PORTS FM_MAX_NUM_OF_10G_MACS
-+#define FM_MAX_NUM_OF_TX_PORTS (FM_MAX_NUM_OF_10G_TX_PORTS + FM_MAX_NUM_OF_1G_TX_PORTS)
-+
-+#define FM_PORT_MAX_NUM_OF_EXT_POOLS 4 /**< Number of external BM pools per Rx port */
-+#define FM_PORT_NUM_OF_CONGESTION_GRPS 256 /**< Total number of congestion groups in QM */
-+#define FM_MAX_NUM_OF_SUB_PORTALS 16
-+#define FM_PORT_MAX_NUM_OF_OBSERVED_EXT_POOLS 0
-+
-+#define FM_VSP_MAX_NUM_OF_ENTRIES 64
-+#define FM_MAX_NUM_OF_PFC_PRIORITIES 8
-+
-+/* RAMs defines */
-+#define FM_MURAM_SIZE (384 * KILOBYTE)
-+#define FM_IRAM_SIZE(major, minor) (64 * KILOBYTE)
-+#define FM_NUM_OF_CTRL 4
-+
-+/* PCD defines */
-+#define FM_PCD_PLCR_NUM_ENTRIES 256 /**< Total number of policer profiles */
-+#define FM_PCD_KG_NUM_OF_SCHEMES 32 /**< Total number of KG schemes */
-+#define FM_PCD_MAX_NUM_OF_CLS_PLANS 256 /**< Number of classification plan entries. */
-+#define FM_PCD_PRS_SW_PATCHES_SIZE 0x00000600 /**< Number of bytes saved for patches */
-+#define FM_PCD_SW_PRS_SIZE 0x00000800 /**< Total size of SW parser area */
-+
-+/* RTC defines */
-+#define FM_RTC_NUM_OF_ALARMS 2 /**< RTC number of alarms */
-+#define FM_RTC_NUM_OF_PERIODIC_PULSES 3 /**< RTC number of periodic pulses */
-+#define FM_RTC_NUM_OF_EXT_TRIGGERS 2 /**< RTC number of external triggers */
-+
-+/* QMI defines */
-+#define QMI_MAX_NUM_OF_TNUMS 64
-+#define QMI_DEF_TNUMS_THRESH 32
-+/* FPM defines */
-+#define FM_NUM_OF_FMAN_CTRL_EVENT_REGS 4
-+
-+/* DMA defines */
-+#define DMA_THRESH_MAX_COMMQ 83
-+#define DMA_THRESH_MAX_BUF 127
-+
-+/* BMI defines */
-+#define BMI_MAX_NUM_OF_TASKS 128
-+#define BMI_MAX_NUM_OF_DMAS 84
-+
-+#define BMI_MAX_FIFO_SIZE (FM_MURAM_SIZE)
-+#define PORT_MAX_WEIGHT 16
-+
-+#define FM_CHECK_PORT_RESTRICTIONS(__validPorts, __newPortIndx) TRUE
-+
-+/* Unique T4240 */
-+#define FM_OP_OPEN_DMA_MIN_LIMIT
-+#define FM_NO_RESTRICT_ON_ACCESS_RSRC
-+#define FM_NO_OP_OBSERVED_POOLS
-+#define FM_FRAME_END_PARAMS_FOR_OP
-+#define FM_DEQ_PIPELINE_PARAMS_FOR_OP
-+#define FM_QMI_NO_SINGLE_ECC_EXCEPTION
-+
-+#define FM_NO_GUARANTEED_RESET_VALUES
-+
-+/* FM errata */
-+#define FM_HEAVY_TRAFFIC_HANG_ERRATA_FMAN_A005669
-+#define FM_WRONG_RESET_VALUES_ERRATA_FMAN_A005127
-+#define FM_RX_FIFO_CORRUPT_ERRATA_10GMAC_A006320
-+#define FM_OP_NO_VSP_NO_RELEASE_ERRATA_FMAN_A006675
-+#define FM_HEAVY_TRAFFIC_SEQUENCER_HANG_ERRATA_FMAN_A006981
-+#define FM_HANG_AT_RESET_MAC_CLK_DISABLED_ERRATA_FMAN_A007273
-+
-+#define FM_BCB_ERRATA_BMI_SW001
-+#define FM_LEN_CHECK_ERRATA_FMAN_SW002
-+#define FM_AID_MODE_NO_TNUM_SW005 /* refer to pdm TKT068794 - only support of port_id on aid */
-+#define FM_ERROR_VSP_NO_MATCH_SW006 /* refer to pdm TKT174304 - no match between errorQ and VSP */
-+
-+/*****************************************************************************
-+ RMan INTEGRATION-SPECIFIC DEFINITIONS
-+******************************************************************************/
-+#define RM_MAX_NUM_OF_IB 4 /**< Number of inbound blocks */
-+#define RM_NUM_OF_IBCU 8 /**< NUmber of classification units in an inbound block */
-+
-+/* RMan erratas */
-+#define RM_ERRONEOUS_ACK_ERRATA_RMAN_A006756
-+
-+/*****************************************************************************
-+ FM MACSEC INTEGRATION-SPECIFIC DEFINITIONS
-+******************************************************************************/
-+#define NUM_OF_RX_SC 16
-+#define NUM_OF_TX_SC 16
-+
-+#define NUM_OF_SA_PER_RX_SC 2
-+#define NUM_OF_SA_PER_TX_SC 2
-+
-+#endif /* __DPAA_INTEGRATION_EXT_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/integrations/FMANV3H/part_ext.h
-@@ -0,0 +1,71 @@
-+/*
-+ * Copyright 2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+/**************************************************************************//**
-+
-+ @File part_ext.h
-+
-+ @Description Definitions for the part (integration) module.
-+*//***************************************************************************/
-+
-+#ifndef __PART_EXT_H
-+#define __PART_EXT_H
-+
-+#include "std_ext.h"
-+#include "part_integration_ext.h"
-+
-+#if !(defined(P1023) || \
-+ defined(P2041) || \
-+ defined(P3041) || \
-+ defined(P4080) || \
-+ defined(P5020) || \
-+ defined(P5040) || \
-+ defined(B4860) || \
-+ defined(T4240))
-+#error "unable to proceed without chip-definition"
-+#endif
-+
-+
-+/**************************************************************************//*
-+ @Description Part data structure - must be contained in any integration
-+ data structure.
-+*//***************************************************************************/
-+typedef struct t_Part
-+{
-+ uintptr_t (* f_GetModuleBase)(t_Handle h_Part, e_ModuleId moduleId);
-+ /**< Returns the address of the module's memory map base. */
-+ e_ModuleId (* f_GetModuleIdByBase)(t_Handle h_Part, uintptr_t baseAddress);
-+ /**< Returns the module's ID according to its memory map base. */
-+} t_Part;
-+
-+
-+#endif /* __PART_EXT_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/integrations/FMANV3H/part_integration_ext.h
-@@ -0,0 +1,304 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+/**
-+
-+ @File part_integration_ext.h
-+
-+ @Description T4240 external definitions and structures.
-+*//***************************************************************************/
-+#ifndef __PART_INTEGRATION_EXT_H
-+#define __PART_INTEGRATION_EXT_H
-+
-+#include "std_ext.h"
-+#include "ddr_std_ext.h"
-+#include "enet_ext.h"
-+#include "dpaa_integration_ext.h"
-+
-+
-+/**************************************************************************//**
-+ @Group T4240_chip_id T4240 Application Programming Interface
-+
-+ @Description T4240 Chip functions,definitions and enums.
-+
-+ @{
-+*//***************************************************************************/
-+
-+#define CORE_E6500
-+
-+#define INTG_MAX_NUM_OF_CORES 24
-+
-+
-+/**************************************************************************//**
-+ @Description Module types.
-+*//***************************************************************************/
-+typedef enum e_ModuleId
-+{
-+ e_MODULE_ID_DUART_1 = 0,
-+ e_MODULE_ID_DUART_2,
-+ e_MODULE_ID_DUART_3,
-+ e_MODULE_ID_DUART_4,
-+ e_MODULE_ID_LAW,
-+ e_MODULE_ID_IFC,
-+ e_MODULE_ID_PAMU,
-+ e_MODULE_ID_QM, /**< Queue manager module */
-+ e_MODULE_ID_BM, /**< Buffer manager module */
-+ e_MODULE_ID_QM_CE_PORTAL_0,
-+ e_MODULE_ID_QM_CI_PORTAL_0,
-+ e_MODULE_ID_QM_CE_PORTAL_1,
-+ e_MODULE_ID_QM_CI_PORTAL_1,
-+ e_MODULE_ID_QM_CE_PORTAL_2,
-+ e_MODULE_ID_QM_CI_PORTAL_2,
-+ e_MODULE_ID_QM_CE_PORTAL_3,
-+ e_MODULE_ID_QM_CI_PORTAL_3,
-+ e_MODULE_ID_QM_CE_PORTAL_4,
-+ e_MODULE_ID_QM_CI_PORTAL_4,
-+ e_MODULE_ID_QM_CE_PORTAL_5,
-+ e_MODULE_ID_QM_CI_PORTAL_5,
-+ e_MODULE_ID_QM_CE_PORTAL_6,
-+ e_MODULE_ID_QM_CI_PORTAL_6,
-+ e_MODULE_ID_QM_CE_PORTAL_7,
-+ e_MODULE_ID_QM_CI_PORTAL_7,
-+ e_MODULE_ID_QM_CE_PORTAL_8,
-+ e_MODULE_ID_QM_CI_PORTAL_8,
-+ e_MODULE_ID_QM_CE_PORTAL_9,
-+ e_MODULE_ID_QM_CI_PORTAL_9,
-+ e_MODULE_ID_BM_CE_PORTAL_0,
-+ e_MODULE_ID_BM_CI_PORTAL_0,
-+ e_MODULE_ID_BM_CE_PORTAL_1,
-+ e_MODULE_ID_BM_CI_PORTAL_1,
-+ e_MODULE_ID_BM_CE_PORTAL_2,
-+ e_MODULE_ID_BM_CI_PORTAL_2,
-+ e_MODULE_ID_BM_CE_PORTAL_3,
-+ e_MODULE_ID_BM_CI_PORTAL_3,
-+ e_MODULE_ID_BM_CE_PORTAL_4,
-+ e_MODULE_ID_BM_CI_PORTAL_4,
-+ e_MODULE_ID_BM_CE_PORTAL_5,
-+ e_MODULE_ID_BM_CI_PORTAL_5,
-+ e_MODULE_ID_BM_CE_PORTAL_6,
-+ e_MODULE_ID_BM_CI_PORTAL_6,
-+ e_MODULE_ID_BM_CE_PORTAL_7,
-+ e_MODULE_ID_BM_CI_PORTAL_7,
-+ e_MODULE_ID_BM_CE_PORTAL_8,
-+ e_MODULE_ID_BM_CI_PORTAL_8,
-+ e_MODULE_ID_BM_CE_PORTAL_9,
-+ e_MODULE_ID_BM_CI_PORTAL_9,
-+ e_MODULE_ID_FM, /**< Frame manager module */
-+ e_MODULE_ID_FM_RTC, /**< FM Real-Time-Clock */
-+ e_MODULE_ID_FM_MURAM, /**< FM Multi-User-RAM */
-+ e_MODULE_ID_FM_BMI, /**< FM BMI block */
-+ e_MODULE_ID_FM_QMI, /**< FM QMI block */
-+ e_MODULE_ID_FM_PARSER, /**< FM parser block */
-+ e_MODULE_ID_FM_PORT_HO1, /**< FM Host-command/offline-parsing port block */
-+ e_MODULE_ID_FM_PORT_HO2, /**< FM Host-command/offline-parsing port block */
-+ e_MODULE_ID_FM_PORT_HO3, /**< FM Host-command/offline-parsing port block */
-+ e_MODULE_ID_FM_PORT_HO4, /**< FM Host-command/offline-parsing port block */
-+ e_MODULE_ID_FM_PORT_HO5, /**< FM Host-command/offline-parsing port block */
-+ e_MODULE_ID_FM_PORT_HO6, /**< FM Host-command/offline-parsing port block */
-+ e_MODULE_ID_FM_PORT_HO7, /**< FM Host-command/offline-parsing port block */
-+ e_MODULE_ID_FM_PORT_1GRx1, /**< FM Rx 1G MAC port block */
-+ e_MODULE_ID_FM_PORT_1GRx2, /**< FM Rx 1G MAC port block */
-+ e_MODULE_ID_FM_PORT_1GRx3, /**< FM Rx 1G MAC port block */
-+ e_MODULE_ID_FM_PORT_1GRx4, /**< FM Rx 1G MAC port block */
-+ e_MODULE_ID_FM_PORT_1GRx5, /**< FM Rx 1G MAC port block */
-+ e_MODULE_ID_FM_PORT_1GRx6, /**< FM Rx 1G MAC port block */
-+ e_MODULE_ID_FM_PORT_10GRx1, /**< FM Rx 10G MAC port block */
-+ e_MODULE_ID_FM_PORT_10GRx2, /**< FM Rx 10G MAC port block */
-+ e_MODULE_ID_FM_PORT_1GTx1, /**< FM Tx 1G MAC port block */
-+ e_MODULE_ID_FM_PORT_1GTx2, /**< FM Tx 1G MAC port block */
-+ e_MODULE_ID_FM_PORT_1GTx3, /**< FM Tx 1G MAC port block */
-+ e_MODULE_ID_FM_PORT_1GTx4, /**< FM Tx 1G MAC port block */
-+ e_MODULE_ID_FM_PORT_1GTx5, /**< FM Tx 1G MAC port block */
-+ e_MODULE_ID_FM_PORT_1GTx6, /**< FM Tx 1G MAC port block */
-+ e_MODULE_ID_FM_PORT_10GTx1, /**< FM Tx 10G MAC port block */
-+ e_MODULE_ID_FM_PORT_10GTx2, /**< FM Tx 10G MAC port block */
-+ e_MODULE_ID_FM_PLCR, /**< FM Policer */
-+ e_MODULE_ID_FM_KG, /**< FM Keygen */
-+ e_MODULE_ID_FM_DMA, /**< FM DMA */
-+ e_MODULE_ID_FM_FPM, /**< FM FPM */
-+ e_MODULE_ID_FM_IRAM, /**< FM Instruction-RAM */
-+ e_MODULE_ID_FM_1GMDIO, /**< FM 1G MDIO MAC */
-+ e_MODULE_ID_FM_10GMDIO, /**< FM 10G MDIO */
-+ e_MODULE_ID_FM_PRS_IRAM, /**< FM SW-parser Instruction-RAM */
-+ e_MODULE_ID_FM_1GMAC1, /**< FM 1G MAC #1 */
-+ e_MODULE_ID_FM_1GMAC2, /**< FM 1G MAC #2 */
-+ e_MODULE_ID_FM_1GMAC3, /**< FM 1G MAC #3 */
-+ e_MODULE_ID_FM_1GMAC4, /**< FM 1G MAC #4 */
-+ e_MODULE_ID_FM_1GMAC5, /**< FM 1G MAC #5 */
-+ e_MODULE_ID_FM_1GMAC6, /**< FM 1G MAC #6 */
-+ e_MODULE_ID_FM_10GMAC1, /**< FM 10G MAC */
-+ e_MODULE_ID_FM_10GMAC2, /**< FM 10G MAC */
-+
-+ e_MODULE_ID_SEC_GEN, /**< SEC 4.0 General registers */
-+ e_MODULE_ID_SEC_QI, /**< SEC 4.0 QI registers */
-+ e_MODULE_ID_SEC_JQ0, /**< SEC 4.0 JQ-0 registers */
-+ e_MODULE_ID_SEC_JQ1, /**< SEC 4.0 JQ-1 registers */
-+ e_MODULE_ID_SEC_JQ2, /**< SEC 4.0 JQ-2 registers */
-+ e_MODULE_ID_SEC_JQ3, /**< SEC 4.0 JQ-3 registers */
-+ e_MODULE_ID_SEC_RTIC, /**< SEC 4.0 RTIC registers */
-+ e_MODULE_ID_SEC_DECO0_CCB0, /**< SEC 4.0 DECO-0/CCB-0 registers */
-+ e_MODULE_ID_SEC_DECO1_CCB1, /**< SEC 4.0 DECO-1/CCB-1 registers */
-+ e_MODULE_ID_SEC_DECO2_CCB2, /**< SEC 4.0 DECO-2/CCB-2 registers */
-+ e_MODULE_ID_SEC_DECO3_CCB3, /**< SEC 4.0 DECO-3/CCB-3 registers */
-+ e_MODULE_ID_SEC_DECO4_CCB4, /**< SEC 4.0 DECO-4/CCB-4 registers */
-+
-+ e_MODULE_ID_PIC, /**< PIC */
-+ e_MODULE_ID_GPIO, /**< GPIO */
-+ e_MODULE_ID_SERDES, /**< SERDES */
-+ e_MODULE_ID_CPC_1, /**< CoreNet-Platform-Cache 1 */
-+ e_MODULE_ID_CPC_2, /**< CoreNet-Platform-Cache 2 */
-+
-+ e_MODULE_ID_SRIO_PORTS, /**< RapidIO controller */
-+
-+ e_MODULE_ID_DUMMY_LAST
-+} e_ModuleId;
-+
-+#define NUM_OF_MODULES e_MODULE_ID_DUMMY_LAST
-+
-+#if 0 /* using unified values */
-+/*****************************************************************************
-+ INTEGRATION-SPECIFIC MODULE CODES
-+******************************************************************************/
-+#define MODULE_UNKNOWN 0x00000000
-+#define MODULE_MEM 0x00010000
-+#define MODULE_MM 0x00020000
-+#define MODULE_CORE 0x00030000
-+#define MODULE_T4240 0x00040000
-+#define MODULE_T4240_PLATFORM 0x00050000
-+#define MODULE_PM 0x00060000
-+#define MODULE_MMU 0x00070000
-+#define MODULE_PIC 0x00080000
-+#define MODULE_CPC 0x00090000
-+#define MODULE_DUART 0x000a0000
-+#define MODULE_SERDES 0x000b0000
-+#define MODULE_PIO 0x000c0000
-+#define MODULE_QM 0x000d0000
-+#define MODULE_BM 0x000e0000
-+#define MODULE_SEC 0x000f0000
-+#define MODULE_LAW 0x00100000
-+#define MODULE_LBC 0x00110000
-+#define MODULE_PAMU 0x00120000
-+#define MODULE_FM 0x00130000
-+#define MODULE_FM_MURAM 0x00140000
-+#define MODULE_FM_PCD 0x00150000
-+#define MODULE_FM_RTC 0x00160000
-+#define MODULE_FM_MAC 0x00170000
-+#define MODULE_FM_PORT 0x00180000
-+#define MODULE_FM_SP 0x00190000
-+#define MODULE_DPA_PORT 0x001a0000
-+#define MODULE_MII 0x001b0000
-+#define MODULE_I2C 0x001c0000
-+#define MODULE_DMA 0x001d0000
-+#define MODULE_DDR 0x001e0000
-+#define MODULE_ESPI 0x001f0000
-+#define MODULE_DPAA_IPSEC 0x00200000
-+#endif /* using unified values */
-+
-+/*****************************************************************************
-+ PAMU INTEGRATION-SPECIFIC DEFINITIONS
-+******************************************************************************/
-+#define PAMU_NUM_OF_PARTITIONS 4
-+
-+/*****************************************************************************
-+ LAW INTEGRATION-SPECIFIC DEFINITIONS
-+******************************************************************************/
-+#define LAW_NUM_OF_WINDOWS 32
-+#define LAW_MIN_WINDOW_SIZE 0x0000000000001000LL /**< 4 Kbytes */
-+#define LAW_MAX_WINDOW_SIZE 0x0000010000000000LL /**< 1 Tbytes for 40-bit address space */
-+
-+
-+/*****************************************************************************
-+ LBC INTEGRATION-SPECIFIC DEFINITIONS
-+******************************************************************************/
-+/**************************************************************************//**
-+ @Group lbc_exception_grp LBC Exception Unit
-+
-+ @Description LBC Exception unit API functions, definitions and enums
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Anchor lbc_exbm
-+
-+ @Collection LBC Errors Bit Mask
-+
-+ These errors are reported through the exceptions callback..
-+ The values can be or'ed in any combination in the errors mask
-+ parameter of the errors report structure.
-+
-+ These errors can also be passed as a bit-mask to
-+ LBC_EnableErrorChecking() or LBC_DisableErrorChecking(),
-+ for enabling or disabling error checking.
-+ @{
-+*//***************************************************************************/
-+#define LBC_ERR_BUS_MONITOR 0x80000000 /**< Bus monitor error */
-+#define LBC_ERR_PARITY_ECC 0x20000000 /**< Parity error for GPCM/UPM */
-+#define LBC_ERR_WRITE_PROTECT 0x04000000 /**< Write protection error */
-+#define LBC_ERR_CHIP_SELECT 0x00080000 /**< Unrecognized chip select */
-+
-+#define LBC_ERR_ALL (LBC_ERR_BUS_MONITOR | LBC_ERR_PARITY_ECC | \
-+ LBC_ERR_WRITE_PROTECT | LBC_ERR_CHIP_SELECT)
-+ /**< All possible errors */
-+/* @} */
-+/** @} */ /* end of lbc_exception_grp group */
-+
-+#define LBC_INCORRECT_ERROR_REPORT_ERRATA
-+
-+#define LBC_NUM_OF_BANKS 8
-+#define LBC_MAX_CS_SIZE 0x0000000100000000LL /* Up to 4G memory block size */
-+#define LBC_PARITY_SUPPORT
-+#define LBC_ADDRESS_HOLD_TIME_CTRL
-+#define LBC_HIGH_CLK_DIVIDERS
-+#define LBC_FCM_AVAILABLE
-+
-+/*****************************************************************************
-+ GPIO INTEGRATION-SPECIFIC DEFINITIONS
-+******************************************************************************/
-+#define GPIO_PORT_OFFSET_0x1000
-+
-+#define GPIO_NUM_OF_PORTS 3 /**< Number of ports in GPIO module;
-+ Each port contains up to 32 I/O pins. */
-+
-+#define GPIO_VALID_PIN_MASKS \
-+ { /* Port A */ 0xFFFFFFFF, \
-+ /* Port B */ 0xFFFFFFFF, \
-+ /* Port C */ 0xFFFFFFFF }
-+
-+#define GPIO_VALID_INTR_MASKS \
-+ { /* Port A */ 0xFFFFFFFF, \
-+ /* Port B */ 0xFFFFFFFF, \
-+ /* Port C */ 0xFFFFFFFF }
-+
-+
-+
-+#endif /* __PART_INTEGRATION_EXT_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/integrations/FMANV3L/dpaa_integration_ext.h
-@@ -0,0 +1,293 @@
-+/*
-+ * Copyright 2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+/**
-+
-+ @File dpaa_integration_ext.h
-+
-+ @Description T4240 FM external definitions and structures.
-+*//***************************************************************************/
-+#ifndef __DPAA_INTEGRATION_EXT_H
-+#define __DPAA_INTEGRATION_EXT_H
-+
-+#include "std_ext.h"
-+
-+
-+#define DPAA_VERSION 11
-+
-+/**************************************************************************//**
-+ @Description DPAA SW Portals Enumeration.
-+*//***************************************************************************/
-+typedef enum
-+{
-+ e_DPAA_SWPORTAL0 = 0,
-+ e_DPAA_SWPORTAL1,
-+ e_DPAA_SWPORTAL2,
-+ e_DPAA_SWPORTAL3,
-+ e_DPAA_SWPORTAL4,
-+ e_DPAA_SWPORTAL5,
-+ e_DPAA_SWPORTAL6,
-+ e_DPAA_SWPORTAL7,
-+ e_DPAA_SWPORTAL8,
-+ e_DPAA_SWPORTAL9,
-+ e_DPAA_SWPORTAL10,
-+ e_DPAA_SWPORTAL11,
-+ e_DPAA_SWPORTAL12,
-+ e_DPAA_SWPORTAL13,
-+ e_DPAA_SWPORTAL14,
-+ e_DPAA_SWPORTAL15,
-+ e_DPAA_SWPORTAL16,
-+ e_DPAA_SWPORTAL17,
-+ e_DPAA_SWPORTAL18,
-+ e_DPAA_SWPORTAL19,
-+ e_DPAA_SWPORTAL20,
-+ e_DPAA_SWPORTAL21,
-+ e_DPAA_SWPORTAL22,
-+ e_DPAA_SWPORTAL23,
-+ e_DPAA_SWPORTAL24,
-+ e_DPAA_SWPORTAL_DUMMY_LAST
-+} e_DpaaSwPortal;
-+
-+/**************************************************************************//**
-+ @Description DPAA Direct Connect Portals Enumeration.
-+*//***************************************************************************/
-+typedef enum
-+{
-+ e_DPAA_DCPORTAL0 = 0,
-+ e_DPAA_DCPORTAL1,
-+ e_DPAA_DCPORTAL2,
-+ e_DPAA_DCPORTAL_DUMMY_LAST
-+} e_DpaaDcPortal;
-+
-+#define DPAA_MAX_NUM_OF_SW_PORTALS e_DPAA_SWPORTAL_DUMMY_LAST
-+#define DPAA_MAX_NUM_OF_DC_PORTALS e_DPAA_DCPORTAL_DUMMY_LAST
-+
-+/*****************************************************************************
-+ QMan INTEGRATION-SPECIFIC DEFINITIONS
-+******************************************************************************/
-+#define QM_MAX_NUM_OF_POOL_CHANNELS 15 /**< Total number of channels, dedicated and pool */
-+#define QM_MAX_NUM_OF_WQ 8 /**< Number of work queues per channel */
-+#define QM_MAX_NUM_OF_CGS 256 /**< Congestion groups number */
-+#define QM_MAX_NUM_OF_FQIDS (16 * MEGABYTE)
-+ /**< FQIDs range - 24 bits */
-+
-+/**************************************************************************//**
-+ @Description Work Queue Channel assignments in QMan.
-+*//***************************************************************************/
-+typedef enum
-+{
-+ e_QM_FQ_CHANNEL_SWPORTAL0 = 0x0, /**< Dedicated channels serviced by software portals 0 to 24 */
-+ e_QM_FQ_CHANNEL_SWPORTAL1,
-+ e_QM_FQ_CHANNEL_SWPORTAL2,
-+ e_QM_FQ_CHANNEL_SWPORTAL3,
-+ e_QM_FQ_CHANNEL_SWPORTAL4,
-+ e_QM_FQ_CHANNEL_SWPORTAL5,
-+ e_QM_FQ_CHANNEL_SWPORTAL6,
-+ e_QM_FQ_CHANNEL_SWPORTAL7,
-+ e_QM_FQ_CHANNEL_SWPORTAL8,
-+ e_QM_FQ_CHANNEL_SWPORTAL9,
-+ e_QM_FQ_CHANNEL_SWPORTAL10,
-+ e_QM_FQ_CHANNEL_SWPORTAL11,
-+ e_QM_FQ_CHANNEL_SWPORTAL12,
-+ e_QM_FQ_CHANNEL_SWPORTAL13,
-+ e_QM_FQ_CHANNEL_SWPORTAL14,
-+ e_QM_FQ_CHANNEL_SWPORTAL15,
-+ e_QM_FQ_CHANNEL_SWPORTAL16,
-+ e_QM_FQ_CHANNEL_SWPORTAL17,
-+ e_QM_FQ_CHANNEL_SWPORTAL18,
-+ e_QM_FQ_CHANNEL_SWPORTAL19,
-+ e_QM_FQ_CHANNEL_SWPORTAL20,
-+ e_QM_FQ_CHANNEL_SWPORTAL21,
-+ e_QM_FQ_CHANNEL_SWPORTAL22,
-+ e_QM_FQ_CHANNEL_SWPORTAL23,
-+ e_QM_FQ_CHANNEL_SWPORTAL24,
-+
-+ e_QM_FQ_CHANNEL_POOL1 = 0x401, /**< Pool channels that can be serviced by any of the software portals */
-+ e_QM_FQ_CHANNEL_POOL2,
-+ e_QM_FQ_CHANNEL_POOL3,
-+ e_QM_FQ_CHANNEL_POOL4,
-+ e_QM_FQ_CHANNEL_POOL5,
-+ e_QM_FQ_CHANNEL_POOL6,
-+ e_QM_FQ_CHANNEL_POOL7,
-+ e_QM_FQ_CHANNEL_POOL8,
-+ e_QM_FQ_CHANNEL_POOL9,
-+ e_QM_FQ_CHANNEL_POOL10,
-+ e_QM_FQ_CHANNEL_POOL11,
-+ e_QM_FQ_CHANNEL_POOL12,
-+ e_QM_FQ_CHANNEL_POOL13,
-+ e_QM_FQ_CHANNEL_POOL14,
-+ e_QM_FQ_CHANNEL_POOL15,
-+
-+ e_QM_FQ_CHANNEL_FMAN0_SP0 = 0x800, /**< Dedicated channels serviced by Direct Connect Portal 0:
-+ connected to FMan 0; assigned in incrementing order to
-+ each sub-portal (SP) in the portal */
-+ e_QM_FQ_CHANNEL_FMAN0_SP1,
-+ e_QM_FQ_CHANNEL_FMAN0_SP2,
-+ e_QM_FQ_CHANNEL_FMAN0_SP3,
-+ e_QM_FQ_CHANNEL_FMAN0_SP4,
-+ e_QM_FQ_CHANNEL_FMAN0_SP5,
-+ e_QM_FQ_CHANNEL_FMAN0_SP6,
-+ e_QM_FQ_CHANNEL_FMAN0_SP7,
-+ e_QM_FQ_CHANNEL_FMAN0_SP8,
-+ e_QM_FQ_CHANNEL_FMAN0_SP9,
-+ e_QM_FQ_CHANNEL_FMAN0_SP10,
-+ e_QM_FQ_CHANNEL_FMAN0_SP11,
-+ e_QM_FQ_CHANNEL_FMAN0_SP12,
-+ e_QM_FQ_CHANNEL_FMAN0_SP13,
-+ e_QM_FQ_CHANNEL_FMAN0_SP14,
-+ e_QM_FQ_CHANNEL_FMAN0_SP15,
-+
-+ e_QM_FQ_CHANNEL_RMAN_SP0 = 0x820, /**< Dedicated channels serviced by Direct Connect Portal 1: connected to RMan */
-+ e_QM_FQ_CHANNEL_RMAN_SP1,
-+
-+ e_QM_FQ_CHANNEL_CAAM = 0x840 /**< Dedicated channel serviced by Direct Connect Portal 2:
-+ connected to SEC */
-+} e_QmFQChannel;
-+
-+/*****************************************************************************
-+ BMan INTEGRATION-SPECIFIC DEFINITIONS
-+******************************************************************************/
-+#define BM_MAX_NUM_OF_POOLS 64 /**< Number of buffers pools */
-+
-+/*****************************************************************************
-+ SEC INTEGRATION-SPECIFIC DEFINITIONS
-+******************************************************************************/
-+#define SEC_NUM_OF_DECOS 3
-+#define SEC_ALL_DECOS_MASK 0x00000003
-+
-+
-+/*****************************************************************************
-+ FM INTEGRATION-SPECIFIC DEFINITIONS
-+******************************************************************************/
-+#define INTG_MAX_NUM_OF_FM 1
-+/* Ports defines */
-+#define FM_MAX_NUM_OF_1G_MACS 5
-+#define FM_MAX_NUM_OF_10G_MACS 1
-+#define FM_MAX_NUM_OF_MACS (FM_MAX_NUM_OF_1G_MACS + FM_MAX_NUM_OF_10G_MACS)
-+#define FM_MAX_NUM_OF_OH_PORTS 4
-+
-+#define FM_MAX_NUM_OF_1G_RX_PORTS FM_MAX_NUM_OF_1G_MACS
-+#define FM_MAX_NUM_OF_10G_RX_PORTS FM_MAX_NUM_OF_10G_MACS
-+#define FM_MAX_NUM_OF_RX_PORTS (FM_MAX_NUM_OF_10G_RX_PORTS + FM_MAX_NUM_OF_1G_RX_PORTS)
-+
-+#define FM_MAX_NUM_OF_1G_TX_PORTS FM_MAX_NUM_OF_1G_MACS
-+#define FM_MAX_NUM_OF_10G_TX_PORTS FM_MAX_NUM_OF_10G_MACS
-+#define FM_MAX_NUM_OF_TX_PORTS (FM_MAX_NUM_OF_10G_TX_PORTS + FM_MAX_NUM_OF_1G_TX_PORTS)
-+
-+#define FM_MAX_NUM_OF_MACSECS 1 /* Should be updated */
-+
-+#define FM_PORT_MAX_NUM_OF_EXT_POOLS 4 /**< Number of external BM pools per Rx port */
-+#define FM_PORT_NUM_OF_CONGESTION_GRPS 256 /**< Total number of congestion groups in QM */
-+#define FM_MAX_NUM_OF_SUB_PORTALS 16
-+#define FM_PORT_MAX_NUM_OF_OBSERVED_EXT_POOLS 0
-+
-+#define FM_VSP_MAX_NUM_OF_ENTRIES 32
-+#define FM_MAX_NUM_OF_PFC_PRIORITIES 8
-+
-+/* RAMs defines */
-+#define FM_MURAM_SIZE (192 * KILOBYTE)
-+#define FM_IRAM_SIZE(major, minor) \
-+ (((major == 6) && ((minor == 4) )) ? (64 * KILOBYTE) : (32 * KILOBYTE))
-+#define FM_NUM_OF_CTRL 2
-+
-+/* PCD defines */
-+#define FM_PCD_PLCR_NUM_ENTRIES 256 /**< Total number of policer profiles */
-+#define FM_PCD_KG_NUM_OF_SCHEMES 32 /**< Total number of KG schemes */
-+#define FM_PCD_MAX_NUM_OF_CLS_PLANS 256 /**< Number of classification plan entries. */
-+#define FM_PCD_PRS_SW_PATCHES_SIZE 0x00000600 /**< Number of bytes saved for patches */
-+#define FM_PCD_SW_PRS_SIZE 0x00000800 /**< Total size of SW parser area */
-+
-+/* RTC defines */
-+#define FM_RTC_NUM_OF_ALARMS 2 /**< RTC number of alarms */
-+#define FM_RTC_NUM_OF_PERIODIC_PULSES 3 /**< RTC number of periodic pulses */
-+#define FM_RTC_NUM_OF_EXT_TRIGGERS 2 /**< RTC number of external triggers */
-+
-+/* QMI defines */
-+#define QMI_MAX_NUM_OF_TNUMS 64
-+#define QMI_DEF_TNUMS_THRESH 32
-+/* FPM defines */
-+#define FM_NUM_OF_FMAN_CTRL_EVENT_REGS 4
-+
-+/* DMA defines */
-+#define DMA_THRESH_MAX_COMMQ 83
-+#define DMA_THRESH_MAX_BUF 127
-+
-+/* BMI defines */
-+#define BMI_MAX_NUM_OF_TASKS 64
-+#define BMI_MAX_NUM_OF_DMAS 32
-+
-+#define BMI_MAX_FIFO_SIZE (FM_MURAM_SIZE)
-+#define PORT_MAX_WEIGHT 16
-+
-+#define FM_CHECK_PORT_RESTRICTIONS(__validPorts, __newPortIndx) TRUE
-+
-+/* Unique T4240 */
-+#define FM_OP_OPEN_DMA_MIN_LIMIT
-+#define FM_NO_RESTRICT_ON_ACCESS_RSRC
-+#define FM_NO_OP_OBSERVED_POOLS
-+#define FM_FRAME_END_PARAMS_FOR_OP
-+#define FM_DEQ_PIPELINE_PARAMS_FOR_OP
-+#define FM_QMI_NO_SINGLE_ECC_EXCEPTION
-+
-+#define FM_NO_GUARANTEED_RESET_VALUES
-+
-+/* FM errata */
-+#define FM_HEAVY_TRAFFIC_HANG_ERRATA_FMAN_A005669
-+#define FM_RX_FIFO_CORRUPT_ERRATA_10GMAC_A006320
-+#define FM_OP_NO_VSP_NO_RELEASE_ERRATA_FMAN_A006675
-+#define FM_HEAVY_TRAFFIC_SEQUENCER_HANG_ERRATA_FMAN_A006981
-+#define FM_HANG_AT_RESET_MAC_CLK_DISABLED_ERRATA_FMAN_A007273
-+
-+#define FM_BCB_ERRATA_BMI_SW001
-+#define FM_LEN_CHECK_ERRATA_FMAN_SW002
-+#define FM_AID_MODE_NO_TNUM_SW005 /* refer to pdm TKT068794 - only support of port_id on aid */
-+#define FM_ERROR_VSP_NO_MATCH_SW006 /* refer to pdm TKT174304 - no match between errorQ and VSP */
-+
-+/*****************************************************************************
-+ RMan INTEGRATION-SPECIFIC DEFINITIONS
-+******************************************************************************/
-+#define RM_MAX_NUM_OF_IB 4 /**< Number of inbound blocks */
-+#define RM_NUM_OF_IBCU 8 /**< NUmber of classification units in an inbound block */
-+
-+/* RMan erratas */
-+#define RM_ERRONEOUS_ACK_ERRATA_RMAN_A006756
-+
-+/*****************************************************************************
-+ FM MACSEC INTEGRATION-SPECIFIC DEFINITIONS
-+******************************************************************************/
-+#define NUM_OF_RX_SC 16
-+#define NUM_OF_TX_SC 16
-+
-+#define NUM_OF_SA_PER_RX_SC 2
-+#define NUM_OF_SA_PER_TX_SC 2
-+
-+#endif /* __DPAA_INTEGRATION_EXT_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/integrations/FMANV3L/part_ext.h
-@@ -0,0 +1,59 @@
-+/*
-+ * Copyright 2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+/**************************************************************************//**
-+
-+ @File part_ext.h
-+
-+ @Description Definitions for the part (integration) module.
-+*//***************************************************************************/
-+
-+#ifndef __PART_EXT_H
-+#define __PART_EXT_H
-+
-+#include "std_ext.h"
-+#include "part_integration_ext.h"
-+
-+/**************************************************************************//*
-+ @Description Part data structure - must be contained in any integration
-+ data structure.
-+*//***************************************************************************/
-+typedef struct t_Part
-+{
-+ uintptr_t (* f_GetModuleBase)(t_Handle h_Part, e_ModuleId moduleId);
-+ /**< Returns the address of the module's memory map base. */
-+ e_ModuleId (* f_GetModuleIdByBase)(t_Handle h_Part, uintptr_t baseAddress);
-+ /**< Returns the module's ID according to its memory map base. */
-+} t_Part;
-+
-+
-+#endif /* __PART_EXT_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/integrations/FMANV3L/part_integration_ext.h
-@@ -0,0 +1,304 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+/**
-+
-+ @File part_integration_ext.h
-+
-+ @Description T4240 external definitions and structures.
-+*//***************************************************************************/
-+#ifndef __PART_INTEGRATION_EXT_H
-+#define __PART_INTEGRATION_EXT_H
-+
-+#include "std_ext.h"
-+#include "ddr_std_ext.h"
-+#include "enet_ext.h"
-+#include "dpaa_integration_ext.h"
-+
-+
-+/**************************************************************************//**
-+ @Group T4240_chip_id T4240 Application Programming Interface
-+
-+ @Description T4240 Chip functions,definitions and enums.
-+
-+ @{
-+*//***************************************************************************/
-+
-+#define CORE_E6500
-+
-+#define INTG_MAX_NUM_OF_CORES 24
-+
-+
-+/**************************************************************************//**
-+ @Description Module types.
-+*//***************************************************************************/
-+typedef enum e_ModuleId
-+{
-+ e_MODULE_ID_DUART_1 = 0,
-+ e_MODULE_ID_DUART_2,
-+ e_MODULE_ID_DUART_3,
-+ e_MODULE_ID_DUART_4,
-+ e_MODULE_ID_LAW,
-+ e_MODULE_ID_IFC,
-+ e_MODULE_ID_PAMU,
-+ e_MODULE_ID_QM, /**< Queue manager module */
-+ e_MODULE_ID_BM, /**< Buffer manager module */
-+ e_MODULE_ID_QM_CE_PORTAL_0,
-+ e_MODULE_ID_QM_CI_PORTAL_0,
-+ e_MODULE_ID_QM_CE_PORTAL_1,
-+ e_MODULE_ID_QM_CI_PORTAL_1,
-+ e_MODULE_ID_QM_CE_PORTAL_2,
-+ e_MODULE_ID_QM_CI_PORTAL_2,
-+ e_MODULE_ID_QM_CE_PORTAL_3,
-+ e_MODULE_ID_QM_CI_PORTAL_3,
-+ e_MODULE_ID_QM_CE_PORTAL_4,
-+ e_MODULE_ID_QM_CI_PORTAL_4,
-+ e_MODULE_ID_QM_CE_PORTAL_5,
-+ e_MODULE_ID_QM_CI_PORTAL_5,
-+ e_MODULE_ID_QM_CE_PORTAL_6,
-+ e_MODULE_ID_QM_CI_PORTAL_6,
-+ e_MODULE_ID_QM_CE_PORTAL_7,
-+ e_MODULE_ID_QM_CI_PORTAL_7,
-+ e_MODULE_ID_QM_CE_PORTAL_8,
-+ e_MODULE_ID_QM_CI_PORTAL_8,
-+ e_MODULE_ID_QM_CE_PORTAL_9,
-+ e_MODULE_ID_QM_CI_PORTAL_9,
-+ e_MODULE_ID_BM_CE_PORTAL_0,
-+ e_MODULE_ID_BM_CI_PORTAL_0,
-+ e_MODULE_ID_BM_CE_PORTAL_1,
-+ e_MODULE_ID_BM_CI_PORTAL_1,
-+ e_MODULE_ID_BM_CE_PORTAL_2,
-+ e_MODULE_ID_BM_CI_PORTAL_2,
-+ e_MODULE_ID_BM_CE_PORTAL_3,
-+ e_MODULE_ID_BM_CI_PORTAL_3,
-+ e_MODULE_ID_BM_CE_PORTAL_4,
-+ e_MODULE_ID_BM_CI_PORTAL_4,
-+ e_MODULE_ID_BM_CE_PORTAL_5,
-+ e_MODULE_ID_BM_CI_PORTAL_5,
-+ e_MODULE_ID_BM_CE_PORTAL_6,
-+ e_MODULE_ID_BM_CI_PORTAL_6,
-+ e_MODULE_ID_BM_CE_PORTAL_7,
-+ e_MODULE_ID_BM_CI_PORTAL_7,
-+ e_MODULE_ID_BM_CE_PORTAL_8,
-+ e_MODULE_ID_BM_CI_PORTAL_8,
-+ e_MODULE_ID_BM_CE_PORTAL_9,
-+ e_MODULE_ID_BM_CI_PORTAL_9,
-+ e_MODULE_ID_FM, /**< Frame manager module */
-+ e_MODULE_ID_FM_RTC, /**< FM Real-Time-Clock */
-+ e_MODULE_ID_FM_MURAM, /**< FM Multi-User-RAM */
-+ e_MODULE_ID_FM_BMI, /**< FM BMI block */
-+ e_MODULE_ID_FM_QMI, /**< FM QMI block */
-+ e_MODULE_ID_FM_PARSER, /**< FM parser block */
-+ e_MODULE_ID_FM_PORT_HO1, /**< FM Host-command/offline-parsing port block */
-+ e_MODULE_ID_FM_PORT_HO2, /**< FM Host-command/offline-parsing port block */
-+ e_MODULE_ID_FM_PORT_HO3, /**< FM Host-command/offline-parsing port block */
-+ e_MODULE_ID_FM_PORT_HO4, /**< FM Host-command/offline-parsing port block */
-+ e_MODULE_ID_FM_PORT_HO5, /**< FM Host-command/offline-parsing port block */
-+ e_MODULE_ID_FM_PORT_HO6, /**< FM Host-command/offline-parsing port block */
-+ e_MODULE_ID_FM_PORT_HO7, /**< FM Host-command/offline-parsing port block */
-+ e_MODULE_ID_FM_PORT_1GRx1, /**< FM Rx 1G MAC port block */
-+ e_MODULE_ID_FM_PORT_1GRx2, /**< FM Rx 1G MAC port block */
-+ e_MODULE_ID_FM_PORT_1GRx3, /**< FM Rx 1G MAC port block */
-+ e_MODULE_ID_FM_PORT_1GRx4, /**< FM Rx 1G MAC port block */
-+ e_MODULE_ID_FM_PORT_1GRx5, /**< FM Rx 1G MAC port block */
-+ e_MODULE_ID_FM_PORT_1GRx6, /**< FM Rx 1G MAC port block */
-+ e_MODULE_ID_FM_PORT_10GRx1, /**< FM Rx 10G MAC port block */
-+ e_MODULE_ID_FM_PORT_10GRx2, /**< FM Rx 10G MAC port block */
-+ e_MODULE_ID_FM_PORT_1GTx1, /**< FM Tx 1G MAC port block */
-+ e_MODULE_ID_FM_PORT_1GTx2, /**< FM Tx 1G MAC port block */
-+ e_MODULE_ID_FM_PORT_1GTx3, /**< FM Tx 1G MAC port block */
-+ e_MODULE_ID_FM_PORT_1GTx4, /**< FM Tx 1G MAC port block */
-+ e_MODULE_ID_FM_PORT_1GTx5, /**< FM Tx 1G MAC port block */
-+ e_MODULE_ID_FM_PORT_1GTx6, /**< FM Tx 1G MAC port block */
-+ e_MODULE_ID_FM_PORT_10GTx1, /**< FM Tx 10G MAC port block */
-+ e_MODULE_ID_FM_PORT_10GTx2, /**< FM Tx 10G MAC port block */
-+ e_MODULE_ID_FM_PLCR, /**< FM Policer */
-+ e_MODULE_ID_FM_KG, /**< FM Keygen */
-+ e_MODULE_ID_FM_DMA, /**< FM DMA */
-+ e_MODULE_ID_FM_FPM, /**< FM FPM */
-+ e_MODULE_ID_FM_IRAM, /**< FM Instruction-RAM */
-+ e_MODULE_ID_FM_1GMDIO, /**< FM 1G MDIO MAC */
-+ e_MODULE_ID_FM_10GMDIO, /**< FM 10G MDIO */
-+ e_MODULE_ID_FM_PRS_IRAM, /**< FM SW-parser Instruction-RAM */
-+ e_MODULE_ID_FM_1GMAC1, /**< FM 1G MAC #1 */
-+ e_MODULE_ID_FM_1GMAC2, /**< FM 1G MAC #2 */
-+ e_MODULE_ID_FM_1GMAC3, /**< FM 1G MAC #3 */
-+ e_MODULE_ID_FM_1GMAC4, /**< FM 1G MAC #4 */
-+ e_MODULE_ID_FM_1GMAC5, /**< FM 1G MAC #5 */
-+ e_MODULE_ID_FM_1GMAC6, /**< FM 1G MAC #6 */
-+ e_MODULE_ID_FM_10GMAC1, /**< FM 10G MAC */
-+ e_MODULE_ID_FM_10GMAC2, /**< FM 10G MAC */
-+
-+ e_MODULE_ID_SEC_GEN, /**< SEC 4.0 General registers */
-+ e_MODULE_ID_SEC_QI, /**< SEC 4.0 QI registers */
-+ e_MODULE_ID_SEC_JQ0, /**< SEC 4.0 JQ-0 registers */
-+ e_MODULE_ID_SEC_JQ1, /**< SEC 4.0 JQ-1 registers */
-+ e_MODULE_ID_SEC_JQ2, /**< SEC 4.0 JQ-2 registers */
-+ e_MODULE_ID_SEC_JQ3, /**< SEC 4.0 JQ-3 registers */
-+ e_MODULE_ID_SEC_RTIC, /**< SEC 4.0 RTIC registers */
-+ e_MODULE_ID_SEC_DECO0_CCB0, /**< SEC 4.0 DECO-0/CCB-0 registers */
-+ e_MODULE_ID_SEC_DECO1_CCB1, /**< SEC 4.0 DECO-1/CCB-1 registers */
-+ e_MODULE_ID_SEC_DECO2_CCB2, /**< SEC 4.0 DECO-2/CCB-2 registers */
-+ e_MODULE_ID_SEC_DECO3_CCB3, /**< SEC 4.0 DECO-3/CCB-3 registers */
-+ e_MODULE_ID_SEC_DECO4_CCB4, /**< SEC 4.0 DECO-4/CCB-4 registers */
-+
-+ e_MODULE_ID_PIC, /**< PIC */
-+ e_MODULE_ID_GPIO, /**< GPIO */
-+ e_MODULE_ID_SERDES, /**< SERDES */
-+ e_MODULE_ID_CPC_1, /**< CoreNet-Platform-Cache 1 */
-+ e_MODULE_ID_CPC_2, /**< CoreNet-Platform-Cache 2 */
-+
-+ e_MODULE_ID_SRIO_PORTS, /**< RapidIO controller */
-+
-+ e_MODULE_ID_DUMMY_LAST
-+} e_ModuleId;
-+
-+#define NUM_OF_MODULES e_MODULE_ID_DUMMY_LAST
-+
-+#if 0 /* using unified values */
-+/*****************************************************************************
-+ INTEGRATION-SPECIFIC MODULE CODES
-+******************************************************************************/
-+#define MODULE_UNKNOWN 0x00000000
-+#define MODULE_MEM 0x00010000
-+#define MODULE_MM 0x00020000
-+#define MODULE_CORE 0x00030000
-+#define MODULE_T4240 0x00040000
-+#define MODULE_T4240_PLATFORM 0x00050000
-+#define MODULE_PM 0x00060000
-+#define MODULE_MMU 0x00070000
-+#define MODULE_PIC 0x00080000
-+#define MODULE_CPC 0x00090000
-+#define MODULE_DUART 0x000a0000
-+#define MODULE_SERDES 0x000b0000
-+#define MODULE_PIO 0x000c0000
-+#define MODULE_QM 0x000d0000
-+#define MODULE_BM 0x000e0000
-+#define MODULE_SEC 0x000f0000
-+#define MODULE_LAW 0x00100000
-+#define MODULE_LBC 0x00110000
-+#define MODULE_PAMU 0x00120000
-+#define MODULE_FM 0x00130000
-+#define MODULE_FM_MURAM 0x00140000
-+#define MODULE_FM_PCD 0x00150000
-+#define MODULE_FM_RTC 0x00160000
-+#define MODULE_FM_MAC 0x00170000
-+#define MODULE_FM_PORT 0x00180000
-+#define MODULE_FM_SP 0x00190000
-+#define MODULE_DPA_PORT 0x001a0000
-+#define MODULE_MII 0x001b0000
-+#define MODULE_I2C 0x001c0000
-+#define MODULE_DMA 0x001d0000
-+#define MODULE_DDR 0x001e0000
-+#define MODULE_ESPI 0x001f0000
-+#define MODULE_DPAA_IPSEC 0x00200000
-+#endif /* using unified values */
-+
-+/*****************************************************************************
-+ PAMU INTEGRATION-SPECIFIC DEFINITIONS
-+******************************************************************************/
-+#define PAMU_NUM_OF_PARTITIONS 4
-+
-+/*****************************************************************************
-+ LAW INTEGRATION-SPECIFIC DEFINITIONS
-+******************************************************************************/
-+#define LAW_NUM_OF_WINDOWS 32
-+#define LAW_MIN_WINDOW_SIZE 0x0000000000001000LL /**< 4 Kbytes */
-+#define LAW_MAX_WINDOW_SIZE 0x0000010000000000LL /**< 1 Tbytes for 40-bit address space */
-+
-+
-+/*****************************************************************************
-+ LBC INTEGRATION-SPECIFIC DEFINITIONS
-+******************************************************************************/
-+/**************************************************************************//**
-+ @Group lbc_exception_grp LBC Exception Unit
-+
-+ @Description LBC Exception unit API functions, definitions and enums
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Anchor lbc_exbm
-+
-+ @Collection LBC Errors Bit Mask
-+
-+ These errors are reported through the exceptions callback..
-+ The values can be or'ed in any combination in the errors mask
-+ parameter of the errors report structure.
-+
-+ These errors can also be passed as a bit-mask to
-+ LBC_EnableErrorChecking() or LBC_DisableErrorChecking(),
-+ for enabling or disabling error checking.
-+ @{
-+*//***************************************************************************/
-+#define LBC_ERR_BUS_MONITOR 0x80000000 /**< Bus monitor error */
-+#define LBC_ERR_PARITY_ECC 0x20000000 /**< Parity error for GPCM/UPM */
-+#define LBC_ERR_WRITE_PROTECT 0x04000000 /**< Write protection error */
-+#define LBC_ERR_CHIP_SELECT 0x00080000 /**< Unrecognized chip select */
-+
-+#define LBC_ERR_ALL (LBC_ERR_BUS_MONITOR | LBC_ERR_PARITY_ECC | \
-+ LBC_ERR_WRITE_PROTECT | LBC_ERR_CHIP_SELECT)
-+ /**< All possible errors */
-+/* @} */
-+/** @} */ /* end of lbc_exception_grp group */
-+
-+#define LBC_INCORRECT_ERROR_REPORT_ERRATA
-+
-+#define LBC_NUM_OF_BANKS 8
-+#define LBC_MAX_CS_SIZE 0x0000000100000000LL /* Up to 4G memory block size */
-+#define LBC_PARITY_SUPPORT
-+#define LBC_ADDRESS_HOLD_TIME_CTRL
-+#define LBC_HIGH_CLK_DIVIDERS
-+#define LBC_FCM_AVAILABLE
-+
-+/*****************************************************************************
-+ GPIO INTEGRATION-SPECIFIC DEFINITIONS
-+******************************************************************************/
-+#define GPIO_PORT_OFFSET_0x1000
-+
-+#define GPIO_NUM_OF_PORTS 3 /**< Number of ports in GPIO module;
-+ Each port contains up to 32 I/O pins. */
-+
-+#define GPIO_VALID_PIN_MASKS \
-+ { /* Port A */ 0xFFFFFFFF, \
-+ /* Port B */ 0xFFFFFFFF, \
-+ /* Port C */ 0xFFFFFFFF }
-+
-+#define GPIO_VALID_INTR_MASKS \
-+ { /* Port A */ 0xFFFFFFFF, \
-+ /* Port B */ 0xFFFFFFFF, \
-+ /* Port C */ 0xFFFFFFFF }
-+
-+
-+
-+#endif /* __PART_INTEGRATION_EXT_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/integrations/LS1043/dpaa_integration_ext.h
-@@ -0,0 +1,291 @@
-+/*
-+ * Copyright 2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+/**
-+
-+ @File dpaa_integration_ext.h
-+
-+ @Description T4240 FM external definitions and structures.
-+*//***************************************************************************/
-+#ifndef __DPAA_INTEGRATION_EXT_H
-+#define __DPAA_INTEGRATION_EXT_H
-+
-+#include "std_ext.h"
-+
-+
-+#define DPAA_VERSION 11
-+
-+/**************************************************************************//**
-+ @Description DPAA SW Portals Enumeration.
-+*//***************************************************************************/
-+typedef enum
-+{
-+ e_DPAA_SWPORTAL0 = 0,
-+ e_DPAA_SWPORTAL1,
-+ e_DPAA_SWPORTAL2,
-+ e_DPAA_SWPORTAL3,
-+ e_DPAA_SWPORTAL4,
-+ e_DPAA_SWPORTAL5,
-+ e_DPAA_SWPORTAL6,
-+ e_DPAA_SWPORTAL7,
-+ e_DPAA_SWPORTAL8,
-+ e_DPAA_SWPORTAL9,
-+ e_DPAA_SWPORTAL10,
-+ e_DPAA_SWPORTAL11,
-+ e_DPAA_SWPORTAL12,
-+ e_DPAA_SWPORTAL13,
-+ e_DPAA_SWPORTAL14,
-+ e_DPAA_SWPORTAL15,
-+ e_DPAA_SWPORTAL16,
-+ e_DPAA_SWPORTAL17,
-+ e_DPAA_SWPORTAL18,
-+ e_DPAA_SWPORTAL19,
-+ e_DPAA_SWPORTAL20,
-+ e_DPAA_SWPORTAL21,
-+ e_DPAA_SWPORTAL22,
-+ e_DPAA_SWPORTAL23,
-+ e_DPAA_SWPORTAL24,
-+ e_DPAA_SWPORTAL_DUMMY_LAST
-+} e_DpaaSwPortal;
-+
-+/**************************************************************************//**
-+ @Description DPAA Direct Connect Portals Enumeration.
-+*//***************************************************************************/
-+typedef enum
-+{
-+ e_DPAA_DCPORTAL0 = 0,
-+ e_DPAA_DCPORTAL1,
-+ e_DPAA_DCPORTAL2,
-+ e_DPAA_DCPORTAL_DUMMY_LAST
-+} e_DpaaDcPortal;
-+
-+#define DPAA_MAX_NUM_OF_SW_PORTALS e_DPAA_SWPORTAL_DUMMY_LAST
-+#define DPAA_MAX_NUM_OF_DC_PORTALS e_DPAA_DCPORTAL_DUMMY_LAST
-+
-+/*****************************************************************************
-+ QMan INTEGRATION-SPECIFIC DEFINITIONS
-+******************************************************************************/
-+#define QM_MAX_NUM_OF_POOL_CHANNELS 15 /**< Total number of channels, dedicated and pool */
-+#define QM_MAX_NUM_OF_WQ 8 /**< Number of work queues per channel */
-+#define QM_MAX_NUM_OF_CGS 256 /**< Congestion groups number */
-+#define QM_MAX_NUM_OF_FQIDS (16 * MEGABYTE)
-+ /**< FQIDs range - 24 bits */
-+
-+/**************************************************************************//**
-+ @Description Work Queue Channel assignments in QMan.
-+*//***************************************************************************/
-+typedef enum
-+{
-+ e_QM_FQ_CHANNEL_SWPORTAL0 = 0x0, /**< Dedicated channels serviced by software portals 0 to 24 */
-+ e_QM_FQ_CHANNEL_SWPORTAL1,
-+ e_QM_FQ_CHANNEL_SWPORTAL2,
-+ e_QM_FQ_CHANNEL_SWPORTAL3,
-+ e_QM_FQ_CHANNEL_SWPORTAL4,
-+ e_QM_FQ_CHANNEL_SWPORTAL5,
-+ e_QM_FQ_CHANNEL_SWPORTAL6,
-+ e_QM_FQ_CHANNEL_SWPORTAL7,
-+ e_QM_FQ_CHANNEL_SWPORTAL8,
-+ e_QM_FQ_CHANNEL_SWPORTAL9,
-+ e_QM_FQ_CHANNEL_SWPORTAL10,
-+ e_QM_FQ_CHANNEL_SWPORTAL11,
-+ e_QM_FQ_CHANNEL_SWPORTAL12,
-+ e_QM_FQ_CHANNEL_SWPORTAL13,
-+ e_QM_FQ_CHANNEL_SWPORTAL14,
-+ e_QM_FQ_CHANNEL_SWPORTAL15,
-+ e_QM_FQ_CHANNEL_SWPORTAL16,
-+ e_QM_FQ_CHANNEL_SWPORTAL17,
-+ e_QM_FQ_CHANNEL_SWPORTAL18,
-+ e_QM_FQ_CHANNEL_SWPORTAL19,
-+ e_QM_FQ_CHANNEL_SWPORTAL20,
-+ e_QM_FQ_CHANNEL_SWPORTAL21,
-+ e_QM_FQ_CHANNEL_SWPORTAL22,
-+ e_QM_FQ_CHANNEL_SWPORTAL23,
-+ e_QM_FQ_CHANNEL_SWPORTAL24,
-+
-+ e_QM_FQ_CHANNEL_POOL1 = 0x401, /**< Pool channels that can be serviced by any of the software portals */
-+ e_QM_FQ_CHANNEL_POOL2,
-+ e_QM_FQ_CHANNEL_POOL3,
-+ e_QM_FQ_CHANNEL_POOL4,
-+ e_QM_FQ_CHANNEL_POOL5,
-+ e_QM_FQ_CHANNEL_POOL6,
-+ e_QM_FQ_CHANNEL_POOL7,
-+ e_QM_FQ_CHANNEL_POOL8,
-+ e_QM_FQ_CHANNEL_POOL9,
-+ e_QM_FQ_CHANNEL_POOL10,
-+ e_QM_FQ_CHANNEL_POOL11,
-+ e_QM_FQ_CHANNEL_POOL12,
-+ e_QM_FQ_CHANNEL_POOL13,
-+ e_QM_FQ_CHANNEL_POOL14,
-+ e_QM_FQ_CHANNEL_POOL15,
-+
-+ e_QM_FQ_CHANNEL_FMAN0_SP0 = 0x800, /**< Dedicated channels serviced by Direct Connect Portal 0:
-+ connected to FMan 0; assigned in incrementing order to
-+ each sub-portal (SP) in the portal */
-+ e_QM_FQ_CHANNEL_FMAN0_SP1,
-+ e_QM_FQ_CHANNEL_FMAN0_SP2,
-+ e_QM_FQ_CHANNEL_FMAN0_SP3,
-+ e_QM_FQ_CHANNEL_FMAN0_SP4,
-+ e_QM_FQ_CHANNEL_FMAN0_SP5,
-+ e_QM_FQ_CHANNEL_FMAN0_SP6,
-+ e_QM_FQ_CHANNEL_FMAN0_SP7,
-+ e_QM_FQ_CHANNEL_FMAN0_SP8,
-+ e_QM_FQ_CHANNEL_FMAN0_SP9,
-+ e_QM_FQ_CHANNEL_FMAN0_SP10,
-+ e_QM_FQ_CHANNEL_FMAN0_SP11,
-+ e_QM_FQ_CHANNEL_FMAN0_SP12,
-+ e_QM_FQ_CHANNEL_FMAN0_SP13,
-+ e_QM_FQ_CHANNEL_FMAN0_SP14,
-+ e_QM_FQ_CHANNEL_FMAN0_SP15,
-+
-+ e_QM_FQ_CHANNEL_RMAN_SP0 = 0x820, /**< Dedicated channels serviced by Direct Connect Portal 1: connected to RMan */
-+ e_QM_FQ_CHANNEL_RMAN_SP1,
-+
-+ e_QM_FQ_CHANNEL_CAAM = 0x840 /**< Dedicated channel serviced by Direct Connect Portal 2:
-+ connected to SEC */
-+} e_QmFQChannel;
-+
-+/*****************************************************************************
-+ BMan INTEGRATION-SPECIFIC DEFINITIONS
-+******************************************************************************/
-+#define BM_MAX_NUM_OF_POOLS 64 /**< Number of buffers pools */
-+
-+/*****************************************************************************
-+ SEC INTEGRATION-SPECIFIC DEFINITIONS
-+******************************************************************************/
-+#define SEC_NUM_OF_DECOS 3
-+#define SEC_ALL_DECOS_MASK 0x00000003
-+
-+
-+/*****************************************************************************
-+ FM INTEGRATION-SPECIFIC DEFINITIONS
-+******************************************************************************/
-+#define INTG_MAX_NUM_OF_FM 2
-+
-+/* Ports defines */
-+#define FM_MAX_NUM_OF_1G_MACS 6
-+#define FM_MAX_NUM_OF_10G_MACS 2
-+#define FM_MAX_NUM_OF_MACS (FM_MAX_NUM_OF_1G_MACS + FM_MAX_NUM_OF_10G_MACS)
-+#define FM_MAX_NUM_OF_OH_PORTS 6
-+
-+#define FM_MAX_NUM_OF_1G_RX_PORTS FM_MAX_NUM_OF_1G_MACS
-+#define FM_MAX_NUM_OF_10G_RX_PORTS FM_MAX_NUM_OF_10G_MACS
-+#define FM_MAX_NUM_OF_RX_PORTS (FM_MAX_NUM_OF_10G_RX_PORTS + FM_MAX_NUM_OF_1G_RX_PORTS)
-+
-+#define FM_MAX_NUM_OF_1G_TX_PORTS FM_MAX_NUM_OF_1G_MACS
-+#define FM_MAX_NUM_OF_10G_TX_PORTS FM_MAX_NUM_OF_10G_MACS
-+#define FM_MAX_NUM_OF_TX_PORTS (FM_MAX_NUM_OF_10G_TX_PORTS + FM_MAX_NUM_OF_1G_TX_PORTS)
-+
-+#define FM_PORT_MAX_NUM_OF_EXT_POOLS 4 /**< Number of external BM pools per Rx port */
-+#define FM_PORT_NUM_OF_CONGESTION_GRPS 256 /**< Total number of congestion groups in QM */
-+#define FM_MAX_NUM_OF_SUB_PORTALS 16
-+#define FM_PORT_MAX_NUM_OF_OBSERVED_EXT_POOLS 0
-+
-+#define FM_VSP_MAX_NUM_OF_ENTRIES 64
-+#define FM_MAX_NUM_OF_PFC_PRIORITIES 8
-+
-+/* RAMs defines */
-+#define FM_MURAM_SIZE (384 * KILOBYTE)
-+#define FM_IRAM_SIZE(major, minor) (64 * KILOBYTE)
-+#define FM_NUM_OF_CTRL 4
-+
-+/* PCD defines */
-+#define FM_PCD_PLCR_NUM_ENTRIES 256 /**< Total number of policer profiles */
-+#define FM_PCD_KG_NUM_OF_SCHEMES 32 /**< Total number of KG schemes */
-+#define FM_PCD_MAX_NUM_OF_CLS_PLANS 256 /**< Number of classification plan entries. */
-+#define FM_PCD_PRS_SW_PATCHES_SIZE 0x00000600 /**< Number of bytes saved for patches */
-+#define FM_PCD_SW_PRS_SIZE 0x00000800 /**< Total size of SW parser area */
-+
-+/* RTC defines */
-+#define FM_RTC_NUM_OF_ALARMS 2 /**< RTC number of alarms */
-+#define FM_RTC_NUM_OF_PERIODIC_PULSES 3 /**< RTC number of periodic pulses */
-+#define FM_RTC_NUM_OF_EXT_TRIGGERS 2 /**< RTC number of external triggers */
-+
-+/* QMI defines */
-+#define QMI_MAX_NUM_OF_TNUMS 64
-+#define QMI_DEF_TNUMS_THRESH 32
-+/* FPM defines */
-+#define FM_NUM_OF_FMAN_CTRL_EVENT_REGS 4
-+
-+/* DMA defines */
-+#define DMA_THRESH_MAX_COMMQ 83
-+#define DMA_THRESH_MAX_BUF 127
-+
-+/* BMI defines */
-+#define BMI_MAX_NUM_OF_TASKS 128
-+#define BMI_MAX_NUM_OF_DMAS 84
-+
-+#define BMI_MAX_FIFO_SIZE (FM_MURAM_SIZE)
-+#define PORT_MAX_WEIGHT 16
-+
-+#define FM_CHECK_PORT_RESTRICTIONS(__validPorts, __newPortIndx) TRUE
-+
-+/* Unique T4240 */
-+#define FM_OP_OPEN_DMA_MIN_LIMIT
-+#define FM_NO_RESTRICT_ON_ACCESS_RSRC
-+#define FM_NO_OP_OBSERVED_POOLS
-+#define FM_FRAME_END_PARAMS_FOR_OP
-+#define FM_DEQ_PIPELINE_PARAMS_FOR_OP
-+#define FM_QMI_NO_SINGLE_ECC_EXCEPTION
-+
-+#define FM_NO_GUARANTEED_RESET_VALUES
-+
-+/* FM errata */
-+#define FM_HEAVY_TRAFFIC_HANG_ERRATA_FMAN_A005669
-+#define FM_WRONG_RESET_VALUES_ERRATA_FMAN_A005127
-+#define FM_RX_FIFO_CORRUPT_ERRATA_10GMAC_A006320
-+#define FM_OP_NO_VSP_NO_RELEASE_ERRATA_FMAN_A006675
-+#define FM_HEAVY_TRAFFIC_SEQUENCER_HANG_ERRATA_FMAN_A006981
-+
-+#define FM_BCB_ERRATA_BMI_SW001
-+#define FM_LEN_CHECK_ERRATA_FMAN_SW002
-+#define FM_AID_MODE_NO_TNUM_SW005 /* refer to pdm TKT068794 - only support of port_id on aid */
-+#define FM_ERROR_VSP_NO_MATCH_SW006 /* refer to pdm TKT174304 - no match between errorQ and VSP */
-+
-+/*****************************************************************************
-+ RMan INTEGRATION-SPECIFIC DEFINITIONS
-+******************************************************************************/
-+#define RM_MAX_NUM_OF_IB 4 /**< Number of inbound blocks */
-+#define RM_NUM_OF_IBCU 8 /**< NUmber of classification units in an inbound block */
-+
-+/* RMan erratas */
-+#define RM_ERRONEOUS_ACK_ERRATA_RMAN_A006756
-+
-+/*****************************************************************************
-+ FM MACSEC INTEGRATION-SPECIFIC DEFINITIONS
-+******************************************************************************/
-+#define NUM_OF_RX_SC 16
-+#define NUM_OF_TX_SC 16
-+
-+#define NUM_OF_SA_PER_RX_SC 2
-+#define NUM_OF_SA_PER_TX_SC 2
-+
-+#endif /* __DPAA_INTEGRATION_EXT_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/integrations/LS1043/part_ext.h
-@@ -0,0 +1,64 @@
-+/*
-+ * Copyright 2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+/**************************************************************************//**
-+
-+ @File part_ext.h
-+
-+ @Description Definitions for the part (integration) module.
-+*//***************************************************************************/
-+
-+#ifndef __PART_EXT_H
-+#define __PART_EXT_H
-+
-+#include "std_ext.h"
-+#include "part_integration_ext.h"
-+
-+#if !(defined(LS1043))
-+#error "unable to proceed without chip-definition"
-+#endif
-+
-+
-+/**************************************************************************//*
-+ @Description Part data structure - must be contained in any integration
-+ data structure.
-+*//***************************************************************************/
-+typedef struct t_Part
-+{
-+ uintptr_t (* f_GetModuleBase)(t_Handle h_Part, e_ModuleId moduleId);
-+ /**< Returns the address of the module's memory map base. */
-+ e_ModuleId (* f_GetModuleIdByBase)(t_Handle h_Part, uintptr_t baseAddress);
-+ /**< Returns the module's ID according to its memory map base. */
-+} t_Part;
-+
-+
-+#endif /* __PART_EXT_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/integrations/LS1043/part_integration_ext.h
-@@ -0,0 +1,185 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+/**
-+
-+ @File part_integration_ext.h
-+
-+ @Description T4240 external definitions and structures.
-+*//***************************************************************************/
-+#ifndef __PART_INTEGRATION_EXT_H
-+#define __PART_INTEGRATION_EXT_H
-+
-+#include "std_ext.h"
-+#include "ddr_std_ext.h"
-+#include "enet_ext.h"
-+#include "dpaa_integration_ext.h"
-+
-+
-+/**************************************************************************//**
-+ @Group T4240_chip_id T4240 Application Programming Interface
-+
-+ @Description T4240 Chip functions,definitions and enums.
-+
-+ @{
-+*//***************************************************************************/
-+
-+#define INTG_MAX_NUM_OF_CORES 4
-+
-+/**************************************************************************//**
-+ @Description Module types.
-+*//***************************************************************************/
-+typedef enum e_ModuleId
-+{
-+ e_MODULE_ID_DUART_1 = 0,
-+ e_MODULE_ID_DUART_2,
-+ e_MODULE_ID_DUART_3,
-+ e_MODULE_ID_DUART_4,
-+ e_MODULE_ID_LAW,
-+ e_MODULE_ID_IFC,
-+ e_MODULE_ID_PAMU,
-+ e_MODULE_ID_QM, /**< Queue manager module */
-+ e_MODULE_ID_BM, /**< Buffer manager module */
-+ e_MODULE_ID_QM_CE_PORTAL_0,
-+ e_MODULE_ID_QM_CI_PORTAL_0,
-+ e_MODULE_ID_QM_CE_PORTAL_1,
-+ e_MODULE_ID_QM_CI_PORTAL_1,
-+ e_MODULE_ID_QM_CE_PORTAL_2,
-+ e_MODULE_ID_QM_CI_PORTAL_2,
-+ e_MODULE_ID_QM_CE_PORTAL_3,
-+ e_MODULE_ID_QM_CI_PORTAL_3,
-+ e_MODULE_ID_QM_CE_PORTAL_4,
-+ e_MODULE_ID_QM_CI_PORTAL_4,
-+ e_MODULE_ID_QM_CE_PORTAL_5,
-+ e_MODULE_ID_QM_CI_PORTAL_5,
-+ e_MODULE_ID_QM_CE_PORTAL_6,
-+ e_MODULE_ID_QM_CI_PORTAL_6,
-+ e_MODULE_ID_QM_CE_PORTAL_7,
-+ e_MODULE_ID_QM_CI_PORTAL_7,
-+ e_MODULE_ID_QM_CE_PORTAL_8,
-+ e_MODULE_ID_QM_CI_PORTAL_8,
-+ e_MODULE_ID_QM_CE_PORTAL_9,
-+ e_MODULE_ID_QM_CI_PORTAL_9,
-+ e_MODULE_ID_BM_CE_PORTAL_0,
-+ e_MODULE_ID_BM_CI_PORTAL_0,
-+ e_MODULE_ID_BM_CE_PORTAL_1,
-+ e_MODULE_ID_BM_CI_PORTAL_1,
-+ e_MODULE_ID_BM_CE_PORTAL_2,
-+ e_MODULE_ID_BM_CI_PORTAL_2,
-+ e_MODULE_ID_BM_CE_PORTAL_3,
-+ e_MODULE_ID_BM_CI_PORTAL_3,
-+ e_MODULE_ID_BM_CE_PORTAL_4,
-+ e_MODULE_ID_BM_CI_PORTAL_4,
-+ e_MODULE_ID_BM_CE_PORTAL_5,
-+ e_MODULE_ID_BM_CI_PORTAL_5,
-+ e_MODULE_ID_BM_CE_PORTAL_6,
-+ e_MODULE_ID_BM_CI_PORTAL_6,
-+ e_MODULE_ID_BM_CE_PORTAL_7,
-+ e_MODULE_ID_BM_CI_PORTAL_7,
-+ e_MODULE_ID_BM_CE_PORTAL_8,
-+ e_MODULE_ID_BM_CI_PORTAL_8,
-+ e_MODULE_ID_BM_CE_PORTAL_9,
-+ e_MODULE_ID_BM_CI_PORTAL_9,
-+ e_MODULE_ID_FM, /**< Frame manager module */
-+ e_MODULE_ID_FM_RTC, /**< FM Real-Time-Clock */
-+ e_MODULE_ID_FM_MURAM, /**< FM Multi-User-RAM */
-+ e_MODULE_ID_FM_BMI, /**< FM BMI block */
-+ e_MODULE_ID_FM_QMI, /**< FM QMI block */
-+ e_MODULE_ID_FM_PARSER, /**< FM parser block */
-+ e_MODULE_ID_FM_PORT_HO1, /**< FM Host-command/offline-parsing port block */
-+ e_MODULE_ID_FM_PORT_HO2, /**< FM Host-command/offline-parsing port block */
-+ e_MODULE_ID_FM_PORT_HO3, /**< FM Host-command/offline-parsing port block */
-+ e_MODULE_ID_FM_PORT_HO4, /**< FM Host-command/offline-parsing port block */
-+ e_MODULE_ID_FM_PORT_HO5, /**< FM Host-command/offline-parsing port block */
-+ e_MODULE_ID_FM_PORT_HO6, /**< FM Host-command/offline-parsing port block */
-+ e_MODULE_ID_FM_PORT_HO7, /**< FM Host-command/offline-parsing port block */
-+ e_MODULE_ID_FM_PORT_1GRx1, /**< FM Rx 1G MAC port block */
-+ e_MODULE_ID_FM_PORT_1GRx2, /**< FM Rx 1G MAC port block */
-+ e_MODULE_ID_FM_PORT_1GRx3, /**< FM Rx 1G MAC port block */
-+ e_MODULE_ID_FM_PORT_1GRx4, /**< FM Rx 1G MAC port block */
-+ e_MODULE_ID_FM_PORT_1GRx5, /**< FM Rx 1G MAC port block */
-+ e_MODULE_ID_FM_PORT_1GRx6, /**< FM Rx 1G MAC port block */
-+ e_MODULE_ID_FM_PORT_10GRx1, /**< FM Rx 10G MAC port block */
-+ e_MODULE_ID_FM_PORT_10GRx2, /**< FM Rx 10G MAC port block */
-+ e_MODULE_ID_FM_PORT_1GTx1, /**< FM Tx 1G MAC port block */
-+ e_MODULE_ID_FM_PORT_1GTx2, /**< FM Tx 1G MAC port block */
-+ e_MODULE_ID_FM_PORT_1GTx3, /**< FM Tx 1G MAC port block */
-+ e_MODULE_ID_FM_PORT_1GTx4, /**< FM Tx 1G MAC port block */
-+ e_MODULE_ID_FM_PORT_1GTx5, /**< FM Tx 1G MAC port block */
-+ e_MODULE_ID_FM_PORT_1GTx6, /**< FM Tx 1G MAC port block */
-+ e_MODULE_ID_FM_PORT_10GTx1, /**< FM Tx 10G MAC port block */
-+ e_MODULE_ID_FM_PORT_10GTx2, /**< FM Tx 10G MAC port block */
-+ e_MODULE_ID_FM_PLCR, /**< FM Policer */
-+ e_MODULE_ID_FM_KG, /**< FM Keygen */
-+ e_MODULE_ID_FM_DMA, /**< FM DMA */
-+ e_MODULE_ID_FM_FPM, /**< FM FPM */
-+ e_MODULE_ID_FM_IRAM, /**< FM Instruction-RAM */
-+ e_MODULE_ID_FM_1GMDIO, /**< FM 1G MDIO MAC */
-+ e_MODULE_ID_FM_10GMDIO, /**< FM 10G MDIO */
-+ e_MODULE_ID_FM_PRS_IRAM, /**< FM SW-parser Instruction-RAM */
-+ e_MODULE_ID_FM_1GMAC1, /**< FM 1G MAC #1 */
-+ e_MODULE_ID_FM_1GMAC2, /**< FM 1G MAC #2 */
-+ e_MODULE_ID_FM_1GMAC3, /**< FM 1G MAC #3 */
-+ e_MODULE_ID_FM_1GMAC4, /**< FM 1G MAC #4 */
-+ e_MODULE_ID_FM_1GMAC5, /**< FM 1G MAC #5 */
-+ e_MODULE_ID_FM_1GMAC6, /**< FM 1G MAC #6 */
-+ e_MODULE_ID_FM_10GMAC1, /**< FM 10G MAC */
-+ e_MODULE_ID_FM_10GMAC2, /**< FM 10G MAC */
-+
-+ e_MODULE_ID_SEC_GEN, /**< SEC 4.0 General registers */
-+ e_MODULE_ID_SEC_QI, /**< SEC 4.0 QI registers */
-+ e_MODULE_ID_SEC_JQ0, /**< SEC 4.0 JQ-0 registers */
-+ e_MODULE_ID_SEC_JQ1, /**< SEC 4.0 JQ-1 registers */
-+ e_MODULE_ID_SEC_JQ2, /**< SEC 4.0 JQ-2 registers */
-+ e_MODULE_ID_SEC_JQ3, /**< SEC 4.0 JQ-3 registers */
-+ e_MODULE_ID_SEC_RTIC, /**< SEC 4.0 RTIC registers */
-+ e_MODULE_ID_SEC_DECO0_CCB0, /**< SEC 4.0 DECO-0/CCB-0 registers */
-+ e_MODULE_ID_SEC_DECO1_CCB1, /**< SEC 4.0 DECO-1/CCB-1 registers */
-+ e_MODULE_ID_SEC_DECO2_CCB2, /**< SEC 4.0 DECO-2/CCB-2 registers */
-+ e_MODULE_ID_SEC_DECO3_CCB3, /**< SEC 4.0 DECO-3/CCB-3 registers */
-+ e_MODULE_ID_SEC_DECO4_CCB4, /**< SEC 4.0 DECO-4/CCB-4 registers */
-+
-+ e_MODULE_ID_PIC, /**< PIC */
-+ e_MODULE_ID_GPIO, /**< GPIO */
-+ e_MODULE_ID_SERDES, /**< SERDES */
-+ e_MODULE_ID_CPC_1, /**< CoreNet-Platform-Cache 1 */
-+ e_MODULE_ID_CPC_2, /**< CoreNet-Platform-Cache 2 */
-+
-+ e_MODULE_ID_SRIO_PORTS, /**< RapidIO controller */
-+
-+ e_MODULE_ID_DUMMY_LAST
-+} e_ModuleId;
-+
-+#define NUM_OF_MODULES e_MODULE_ID_DUMMY_LAST
-+
-+
-+#endif /* __PART_INTEGRATION_EXT_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/integrations/P1023/dpaa_integration_ext.h
-@@ -0,0 +1,213 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/**
-+
-+ @File dpaa_integration_ext.h
-+
-+ @Description P1023 FM external definitions and structures.
-+*//***************************************************************************/
-+#ifndef __DPAA_INTEGRATION_EXT_H
-+#define __DPAA_INTEGRATION_EXT_H
-+
-+#include "std_ext.h"
-+
-+
-+#define DPAA_VERSION 10
-+
-+typedef enum e_DpaaSwPortal {
-+ e_DPAA_SWPORTAL0 = 0,
-+ e_DPAA_SWPORTAL1,
-+ e_DPAA_SWPORTAL2,
-+ e_DPAA_SWPORTAL_DUMMY_LAST
-+} e_DpaaSwPortal;
-+
-+typedef enum {
-+ e_DPAA_DCPORTAL0 = 0,
-+ e_DPAA_DCPORTAL2,
-+ e_DPAA_DCPORTAL_DUMMY_LAST
-+} e_DpaaDcPortal;
-+
-+#define DPAA_MAX_NUM_OF_SW_PORTALS e_DPAA_SWPORTAL_DUMMY_LAST
-+#define DPAA_MAX_NUM_OF_DC_PORTALS e_DPAA_DCPORTAL_DUMMY_LAST
-+
-+/*****************************************************************************
-+ QMAN INTEGRATION-SPECIFIC DEFINITIONS
-+******************************************************************************/
-+#define QM_MAX_NUM_OF_POOL_CHANNELS 3
-+#define QM_MAX_NUM_OF_WQ 8
-+#define QM_MAX_NUM_OF_SWP_AS 2
-+#define QM_MAX_NUM_OF_CGS 64
-+#define QM_MAX_NUM_OF_FQIDS (16*MEGABYTE)
-+
-+typedef enum {
-+ e_QM_FQ_CHANNEL_SWPORTAL0 = 0,
-+ e_QM_FQ_CHANNEL_SWPORTAL1,
-+ e_QM_FQ_CHANNEL_SWPORTAL2,
-+
-+ e_QM_FQ_CHANNEL_POOL1 = 0x21,
-+ e_QM_FQ_CHANNEL_POOL2,
-+ e_QM_FQ_CHANNEL_POOL3,
-+
-+ e_QM_FQ_CHANNEL_FMAN0_SP0 = 0x40,
-+ e_QM_FQ_CHANNEL_FMAN0_SP1,
-+ e_QM_FQ_CHANNEL_FMAN0_SP2,
-+ e_QM_FQ_CHANNEL_FMAN0_SP3,
-+ e_QM_FQ_CHANNEL_FMAN0_SP4,
-+ e_QM_FQ_CHANNEL_FMAN0_SP5,
-+ e_QM_FQ_CHANNEL_FMAN0_SP6,
-+
-+
-+ e_QM_FQ_CHANNEL_CAAM = 0x80
-+} e_QmFQChannel;
-+
-+/*****************************************************************************
-+ BMAN INTEGRATION-SPECIFIC DEFINITIONS
-+******************************************************************************/
-+#define BM_MAX_NUM_OF_POOLS 8
-+
-+/*****************************************************************************
-+ SEC INTEGRATION-SPECIFIC DEFINITIONS
-+******************************************************************************/
-+#define SEC_NUM_OF_DECOS 2
-+#define SEC_ALL_DECOS_MASK 0x00000003
-+#define SEC_RNGB
-+#define SEC_NO_ESP_TRAILER_REMOVAL
-+
-+/*****************************************************************************
-+ FM INTEGRATION-SPECIFIC DEFINITIONS
-+******************************************************************************/
-+#define INTG_MAX_NUM_OF_FM 1
-+
-+/* Ports defines */
-+#define FM_MAX_NUM_OF_1G_MACS 2
-+#define FM_MAX_NUM_OF_10G_MACS 0
-+#define FM_MAX_NUM_OF_MACS (FM_MAX_NUM_OF_1G_MACS + FM_MAX_NUM_OF_10G_MACS)
-+#define FM_MAX_NUM_OF_OH_PORTS 5
-+
-+#define FM_MAX_NUM_OF_1G_RX_PORTS FM_MAX_NUM_OF_1G_MACS
-+#define FM_MAX_NUM_OF_10G_RX_PORTS FM_MAX_NUM_OF_10G_MACS
-+#define FM_MAX_NUM_OF_RX_PORTS (FM_MAX_NUM_OF_10G_RX_PORTS + FM_MAX_NUM_OF_1G_RX_PORTS)
-+
-+#define FM_MAX_NUM_OF_1G_TX_PORTS FM_MAX_NUM_OF_1G_MACS
-+#define FM_MAX_NUM_OF_10G_TX_PORTS FM_MAX_NUM_OF_10G_MACS
-+#define FM_MAX_NUM_OF_TX_PORTS (FM_MAX_NUM_OF_10G_TX_PORTS + FM_MAX_NUM_OF_1G_TX_PORTS)
-+
-+#define FM_MAX_NUM_OF_MACSECS 1
-+
-+#define FM_MACSEC_SUPPORT
-+
-+#define FM_LOW_END_RESTRICTION /* prevents the use of TX port 1 with OP port 0 */
-+
-+#define FM_PORT_MAX_NUM_OF_EXT_POOLS 4 /**< Number of external BM pools per Rx port */
-+#define FM_PORT_MAX_NUM_OF_OBSERVED_EXT_POOLS 2 /**< Number of Offline parsing port external BM pools per Rx port */
-+#define FM_PORT_NUM_OF_CONGESTION_GRPS 32 /**< Total number of congestion groups in QM */
-+#define FM_MAX_NUM_OF_SUB_PORTALS 7
-+
-+/* Rams defines */
-+#define FM_MURAM_SIZE (64*KILOBYTE)
-+#define FM_IRAM_SIZE(major, minor) (32 * KILOBYTE)
-+#define FM_NUM_OF_CTRL 2
-+
-+/* PCD defines */
-+#define FM_PCD_PLCR_NUM_ENTRIES 32 /**< Total number of policer profiles */
-+#define FM_PCD_KG_NUM_OF_SCHEMES 16 /**< Total number of KG schemes */
-+#define FM_PCD_MAX_NUM_OF_CLS_PLANS 128 /**< Number of classification plan entries. */
-+#define FM_PCD_PRS_SW_PATCHES_SIZE 0x00000240 /**< Number of bytes saved for patches */
-+#define FM_PCD_SW_PRS_SIZE 0x00000800 /**< Total size of SW parser area */
-+
-+/* RTC defines */
-+#define FM_RTC_NUM_OF_ALARMS 2
-+#define FM_RTC_NUM_OF_PERIODIC_PULSES 2
-+#define FM_RTC_NUM_OF_EXT_TRIGGERS 2
-+
-+/* QMI defines */
-+#define QMI_MAX_NUM_OF_TNUMS 15
-+
-+/* FPM defines */
-+#define FM_NUM_OF_FMAN_CTRL_EVENT_REGS 4
-+
-+/* DMA defines */
-+#define DMA_THRESH_MAX_COMMQ 15
-+#define DMA_THRESH_MAX_BUF 7
-+
-+/* BMI defines */
-+#define BMI_MAX_NUM_OF_TASKS 64
-+#define BMI_MAX_NUM_OF_DMAS 16
-+#define BMI_MAX_FIFO_SIZE (FM_MURAM_SIZE)
-+#define PORT_MAX_WEIGHT 4
-+
-+/*****************************************************************************
-+ FM MACSEC INTEGRATION-SPECIFIC DEFINITIONS
-+******************************************************************************/
-+#define NUM_OF_RX_SC 16
-+#define NUM_OF_TX_SC 16
-+
-+#define NUM_OF_SA_PER_RX_SC 2
-+#define NUM_OF_SA_PER_TX_SC 2
-+
-+/**************************************************************************//**
-+ @Description Enum for inter-module interrupts registration
-+*//***************************************************************************/
-+
-+/* 1023 unique features */
-+#define FM_QMI_NO_ECC_EXCEPTIONS
-+#define FM_CSI_CFED_LIMIT
-+#define FM_PEDANTIC_DMA
-+#define FM_QMI_NO_DEQ_OPTIONS_SUPPORT
-+#define FM_FIFO_ALLOCATION_ALG
-+#define FM_DEQ_PIPELINE_PARAMS_FOR_OP
-+#define FM_HAS_TOTAL_DMAS
-+#define FM_KG_NO_IPPID_SUPPORT
-+#define FM_NO_GUARANTEED_RESET_VALUES
-+#define FM_MAC_RESET
-+
-+/* FM erratas */
-+#define FM_RX_PREAM_4_ERRATA_DTSEC_A001
-+#define FM_MAGIC_PACKET_UNRECOGNIZED_ERRATA_DTSEC2 /* No implementation, Out of LLD scope */
-+
-+#define FM_DEBUG_TRACE_FMAN_A004 /* No implementation, Out of LLD scope */
-+#define FM_INT_BUF_LEAK_FMAN_A005 /* No implementation, Out of LLD scope. App must avoid S/G */
-+
-+#define FM_GTS_AFTER_DROPPED_FRAME_ERRATA_DTSEC_A004839
-+
-+/* #define FM_UCODE_NOT_RESET_ERRATA_BUGZILLA6173 */
-+
-+/*
-+TKT056919 - axi12axi0 can hang if read request follows the single byte write on the very next cycle
-+TKT038900 - FM dma lockup occur due to AXI slave protocol violation
-+*/
-+#define FM_LOCKUP_ALIGNMENT_ERRATA_FMAN_SW004
-+
-+
-+#endif /* __DPAA_INTEGRATION_EXT_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/integrations/P1023/part_ext.h
-@@ -0,0 +1,82 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/**************************************************************************//**
-+
-+ @File part_ext.h
-+
-+ @Description Definitions for the part (integration) module.
-+*//***************************************************************************/
-+
-+#ifndef __PART_EXT_H
-+#define __PART_EXT_H
-+
-+#include "std_ext.h"
-+#include "part_integration_ext.h"
-+
-+
-+#if !(defined(MPC8306) || \
-+ defined(MPC8309) || \
-+ defined(MPC834x) || \
-+ defined(MPC836x) || \
-+ defined(MPC832x) || \
-+ defined(MPC837x) || \
-+ defined(MPC8568) || \
-+ defined(MPC8569) || \
-+ defined(P1020) || \
-+ defined(P1021) || \
-+ defined(P1022) || \
-+ defined(P1023) || \
-+ defined(P2020) || \
-+ defined(P3041) || \
-+ defined(P4080) || \
-+ defined(P5020) || \
-+ defined(MSC814x))
-+#error "unable to proceed without chip-definition"
-+#endif
-+
-+
-+/**************************************************************************//*
-+ @Description Part data structure - must be contained in any integration
-+ data structure.
-+*//***************************************************************************/
-+typedef struct t_Part
-+{
-+ uint64_t (* f_GetModuleBase)(t_Handle h_Part, e_ModuleId moduleId);
-+ /**< Returns the address of the module's memory map base. */
-+ e_ModuleId (* f_GetModuleIdByBase)(t_Handle h_Part, uint64_t baseAddress);
-+ /**< Returns the module's ID according to its memory map base. */
-+} t_Part;
-+
-+
-+#endif /* __PART_EXT_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/integrations/P1023/part_integration_ext.h
-@@ -0,0 +1,635 @@
-+/* Copyright (c) 2008-2012 Freescale Semiconductor, Inc
-+ * All rights reserved.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+/**************************************************************************//**
-+ @File part_integration_ext.h
-+
-+ @Description P1023 external definitions and structures.
-+*//***************************************************************************/
-+#ifndef __PART_INTEGRATION_EXT_H
-+#define __PART_INTEGRATION_EXT_H
-+
-+#include "std_ext.h"
-+#include "dpaa_integration_ext.h"
-+
-+
-+/**************************************************************************//**
-+ @Group 1023_chip_id P1023 Application Programming Interface
-+
-+ @Description P1023 Chip functions,definitions and enums.
-+
-+ @{
-+*//***************************************************************************/
-+
-+#define INTG_MAX_NUM_OF_CORES 2
-+
-+
-+/**************************************************************************//**
-+ @Description Module types.
-+*//***************************************************************************/
-+typedef enum e_ModuleId
-+{
-+ e_MODULE_ID_LAW, /**< Local Access module */
-+ e_MODULE_ID_ECM, /**< e500 Coherency Module */
-+ e_MODULE_ID_DDR, /**< DDR memory controller */
-+ e_MODULE_ID_I2C_1, /**< I2C 1 */
-+ e_MODULE_ID_I2C_2, /**< I2C 1 */
-+ e_MODULE_ID_DUART_1, /**< DUART module 1 */
-+ e_MODULE_ID_DUART_2, /**< DUART module 2 */
-+ e_MODULE_ID_LBC, /**< Local bus memory controller module */
-+ e_MODULE_ID_PCIE_1, /**< PCI Express 1 controller module */
-+ e_MODULE_ID_PCIE_ATMU_1, /**< PCI 1 ATMU Window */
-+ e_MODULE_ID_PCIE_2, /**< PCI Express 2 controller module */
-+ e_MODULE_ID_PCIE_ATMU_2, /**< PCI 2 ATMU Window */
-+ e_MODULE_ID_PCIE_3, /**< PCI Express 3 controller module */
-+ e_MODULE_ID_PCIE_ATMU_3, /**< PCI 3 ATMU Window */
-+ e_MODULE_ID_MSI, /**< MSI registers */
-+ e_MODULE_ID_L2_SRAM, /**< L2/SRAM Memory-Mapped controller module */
-+ e_MODULE_ID_DMA_1, /**< DMA controller 1 */
-+ e_MODULE_ID_DMA_2, /**< DMA controller 2 */
-+ e_MODULE_ID_EPIC, /**< Programmable interrupt controller */
-+ e_MODULE_ID_ESPI, /**< ESPI module */
-+ e_MODULE_ID_GPIO, /**< General Purpose I/O */
-+ e_MODULE_ID_SEC_GEN, /**< SEC 4.0 General registers */
-+ e_MODULE_ID_SEC_QI, /**< SEC 4.0 QI registers */
-+ e_MODULE_ID_SEC_JQ0, /**< SEC 4.0 JQ-0 registers */
-+ e_MODULE_ID_SEC_JQ1, /**< SEC 4.0 JQ-1 registers */
-+ e_MODULE_ID_SEC_JQ2, /**< SEC 4.0 JQ-2 registers */
-+ e_MODULE_ID_SEC_JQ3, /**< SEC 4.0 JQ-3 registers */
-+ e_MODULE_ID_SEC_RTIC, /**< SEC 4.0 RTIC registers */
-+ e_MODULE_ID_SEC_DECO0_CCB0, /**< SEC 4.0 DECO-0/CCB-0 registers */
-+ e_MODULE_ID_SEC_DECO1_CCB1, /**< SEC 4.0 DECO-1/CCB-1 registers */
-+ e_MODULE_ID_SEC_DECO2_CCB2, /**< SEC 4.0 DECO-2/CCB-2 registers */
-+ e_MODULE_ID_SEC_DECO3_CCB3, /**< SEC 4.0 DECO-3/CCB-3 registers */
-+ e_MODULE_ID_SEC_DECO4_CCB4, /**< SEC 4.0 DECO-4/CCB-4 registers */
-+ e_MODULE_ID_USB_DR_1, /**< USB 2.0 module 1 */
-+ e_MODULE_ID_USB_DR_2, /**< USB 2.0 module 2 */
-+ e_MODULE_ID_ETSEC_MII_MNG, /**< MII MNG registers */
-+ e_MODULE_ID_ETSEC_1, /**< ETSEC module 1 */
-+ e_MODULE_ID_ETSEC_2, /**< ETSEC module 2 */
-+ e_MODULE_ID_GUTS, /**< Serial DMA */
-+ e_MODULE_ID_PM, /**< Performance Monitor module */
-+ e_MODULE_ID_QM, /**< Queue manager module */
-+ e_MODULE_ID_BM, /**< Buffer manager module */
-+ e_MODULE_ID_QM_CE_PORTAL,
-+ e_MODULE_ID_QM_CI_PORTAL,
-+ e_MODULE_ID_BM_CE_PORTAL,
-+ e_MODULE_ID_BM_CI_PORTAL,
-+ e_MODULE_ID_FM, /**< Frame manager #1 module */
-+ e_MODULE_ID_FM_RTC, /**< FM Real-Time-Clock */
-+ e_MODULE_ID_FM_MURAM, /**< FM Multi-User-RAM */
-+ e_MODULE_ID_FM_BMI, /**< FM BMI block */
-+ e_MODULE_ID_FM_QMI, /**< FM QMI block */
-+ e_MODULE_ID_FM_PRS, /**< FM parser block */
-+ e_MODULE_ID_FM_PORT_HO0, /**< FM Host-command/offline-parsing port block */
-+ e_MODULE_ID_FM_PORT_HO1, /**< FM Host-command/offline-parsing port block */
-+ e_MODULE_ID_FM_PORT_HO2, /**< FM Host-command/offline-parsing port block */
-+ e_MODULE_ID_FM_PORT_HO3, /**< FM Host-command/offline-parsing port block */
-+ e_MODULE_ID_FM_PORT_HO4, /**< FM Host-command/offline-parsing port block */
-+ e_MODULE_ID_FM_PORT_1GRx0, /**< FM Rx 1G MAC port block */
-+ e_MODULE_ID_FM_PORT_1GRx1, /**< FM Rx 1G MAC port block */
-+ e_MODULE_ID_FM_PORT_1GTx0, /**< FM Tx 1G MAC port block */
-+ e_MODULE_ID_FM_PORT_1GTx1, /**< FM Tx 1G MAC port block */
-+ e_MODULE_ID_FM_PLCR, /**< FM Policer */
-+ e_MODULE_ID_FM_KG, /**< FM Keygen */
-+ e_MODULE_ID_FM_DMA, /**< FM DMA */
-+ e_MODULE_ID_FM_FPM, /**< FM FPM */
-+ e_MODULE_ID_FM_IRAM, /**< FM Instruction-RAM */
-+ e_MODULE_ID_FM_1GMDIO0, /**< FM 1G MDIO MAC 0*/
-+ e_MODULE_ID_FM_1GMDIO1, /**< FM 1G MDIO MAC 1*/
-+ e_MODULE_ID_FM_PRS_IRAM, /**< FM SW-parser Instruction-RAM */
-+ e_MODULE_ID_FM_RISC0, /**< FM risc #0 */
-+ e_MODULE_ID_FM_RISC1, /**< FM risc #1 */
-+ e_MODULE_ID_FM_1GMAC0, /**< FM 1G MAC #0 */
-+ e_MODULE_ID_FM_1GMAC1, /**< FM 1G MAC #1 */
-+ e_MODULE_ID_FM_MACSEC, /**< FM MACSEC */
-+
-+ e_MODULE_ID_DUMMY_LAST
-+} e_ModuleId;
-+
-+#define NUM_OF_MODULES e_MODULE_ID_DUMMY_LAST
-+
-+
-+#define P1023_OFFSET_LAW 0x00000C08
-+#define P1023_OFFSET_ECM 0x00001000
-+#define P1023_OFFSET_DDR 0x00002000
-+#define P1023_OFFSET_I2C1 0x00003000
-+#define P1023_OFFSET_I2C2 0x00003100
-+#define P1023_OFFSET_DUART1 0x00004500
-+#define P1023_OFFSET_DUART2 0x00004600
-+#define P1023_OFFSET_LBC 0x00005000
-+#define P1023_OFFSET_ESPI 0x00007000
-+#define P1023_OFFSET_PCIE2 0x00009000
-+#define P1023_OFFSET_PCIE2_ATMU 0x00009C00
-+#define P1023_OFFSET_PCIE1 0x0000A000
-+#define P1023_OFFSET_PCIE1_ATMU 0x0000AC00
-+#define P1023_OFFSET_PCIE3 0x0000B000
-+#define P1023_OFFSET_PCIE3_ATMU 0x0000BC00
-+#define P1023_OFFSET_DMA2 0x0000C100
-+#define P1023_OFFSET_GPIO 0x0000F000
-+#define P1023_OFFSET_L2_SRAM 0x00020000
-+#define P1023_OFFSET_DMA1 0x00021100
-+#define P1023_OFFSET_USB1 0x00022000
-+#define P1023_OFFSET_SEC_GEN 0x00030000
-+#define P1023_OFFSET_SEC_JQ0 0x00031000
-+#define P1023_OFFSET_SEC_JQ1 0x00032000
-+#define P1023_OFFSET_SEC_JQ2 0x00033000
-+#define P1023_OFFSET_SEC_JQ3 0x00034000
-+#define P1023_OFFSET_SEC_RTIC 0x00036000
-+#define P1023_OFFSET_SEC_QI 0x00037000
-+#define P1023_OFFSET_SEC_DECO0_CCB0 0x00038000
-+#define P1023_OFFSET_SEC_DECO1_CCB1 0x00039000
-+#define P1023_OFFSET_SEC_DECO2_CCB2 0x0003a000
-+#define P1023_OFFSET_SEC_DECO3_CCB3 0x0003b000
-+#define P1023_OFFSET_SEC_DECO4_CCB4 0x0003c000
-+#define P1023_OFFSET_PIC 0x00040000
-+#define P1023_OFFSET_MSI 0x00041600
-+#define P1023_OFFSET_AXI 0x00081000
-+#define P1023_OFFSET_QM 0x00088000
-+#define P1023_OFFSET_BM 0x0008A000
-+#define P1022_OFFSET_PM 0x000E1000
-+
-+#define P1023_OFFSET_GUTIL 0x000E0000
-+#define P1023_OFFSET_PM 0x000E1000
-+#define P1023_OFFSET_DEBUG 0x000E2000
-+#define P1023_OFFSET_SERDES 0x000E3000
-+#define P1023_OFFSET_ROM 0x000F0000
-+#define P1023_OFFSET_FM 0x00100000
-+
-+#define P1023_OFFSET_FM_MURAM (P1023_OFFSET_FM + 0x00000000)
-+#define P1023_OFFSET_FM_BMI (P1023_OFFSET_FM + 0x00080000)
-+#define P1023_OFFSET_FM_QMI (P1023_OFFSET_FM + 0x00080400)
-+#define P1023_OFFSET_FM_PRS (P1023_OFFSET_FM + 0x00080800)
-+#define P1023_OFFSET_FM_PORT_HO0 (P1023_OFFSET_FM + 0x00081000)
-+#define P1023_OFFSET_FM_PORT_HO1 (P1023_OFFSET_FM + 0x00082000)
-+#define P1023_OFFSET_FM_PORT_HO2 (P1023_OFFSET_FM + 0x00083000)
-+#define P1023_OFFSET_FM_PORT_HO3 (P1023_OFFSET_FM + 0x00084000)
-+#define P1023_OFFSET_FM_PORT_HO4 (P1023_OFFSET_FM + 0x00085000)
-+#define P1023_OFFSET_FM_PORT_1GRX0 (P1023_OFFSET_FM + 0x00088000)
-+#define P1023_OFFSET_FM_PORT_1GRX1 (P1023_OFFSET_FM + 0x00089000)
-+#define P1023_OFFSET_FM_PORT_1GTX0 (P1023_OFFSET_FM + 0x000A8000)
-+#define P1023_OFFSET_FM_PORT_1GTX1 (P1023_OFFSET_FM + 0x000A9000)
-+#define P1023_OFFSET_FM_PLCR (P1023_OFFSET_FM + 0x000C0000)
-+#define P1023_OFFSET_FM_KG (P1023_OFFSET_FM + 0x000C1000)
-+#define P1023_OFFSET_FM_DMA (P1023_OFFSET_FM + 0x000C2000)
-+#define P1023_OFFSET_FM_FPM (P1023_OFFSET_FM + 0x000C3000)
-+#define P1023_OFFSET_FM_IRAM (P1023_OFFSET_FM + 0x000C4000)
-+#define P1023_OFFSET_FM_PRS_IRAM (P1023_OFFSET_FM + 0x000C7000)
-+#define P1023_OFFSET_FM_RISC0 (P1023_OFFSET_FM + 0x000D0000)
-+#define P1023_OFFSET_FM_RISC1 (P1023_OFFSET_FM + 0x000D0400)
-+#define P1023_OFFSET_FM_MACSEC (P1023_OFFSET_FM + 0x000D8000)
-+#define P1023_OFFSET_FM_1GMAC0 (P1023_OFFSET_FM + 0x000E0000)
-+#define P1023_OFFSET_FM_1GMDIO0 (P1023_OFFSET_FM + 0x000E1120)
-+#define P1023_OFFSET_FM_1GMAC1 (P1023_OFFSET_FM + 0x000E2000)
-+#define P1023_OFFSET_FM_1GMDIO1 (P1023_OFFSET_FM + 0x000E3000)
-+#define P1023_OFFSET_FM_RTC (P1023_OFFSET_FM + 0x000FE000)
-+
-+/* Offsets relative to QM or BM portals base */
-+#define P1023_OFFSET_PORTALS_CE_AREA 0x00000000 /* cache enabled area */
-+#define P1023_OFFSET_PORTALS_CI_AREA 0x00100000 /* cache inhibited area */
-+
-+#define P1023_OFFSET_PORTALS_CE(portal) (P1023_OFFSET_PORTALS_CE_AREA + 0x4000 * (portal))
-+#define P1023_OFFSET_PORTALS_CI(portal) (P1023_OFFSET_PORTALS_CI_AREA + 0x1000 * (portal))
-+
-+/**************************************************************************//**
-+ @Description Transaction source ID (for memory controllers error reporting).
-+*//***************************************************************************/
-+typedef enum e_TransSrc
-+{
-+ e_TRANS_SRC_PCIE_2 = 0x01, /**< PCIe port 2 */
-+ e_TRANS_SRC_PCIE_1 = 0x02, /**< PCIe port 1 */
-+ e_TRANS_SRC_PCIE_3 = 0x03, /**< PCIe port 3 */
-+ e_TRANS_SRC_LBC = 0x04, /**< Enhanced local bus */
-+ e_TRANS_SRC_DPAA_SW_PORTALS = 0x0E, /**< DPAA software portals or SRAM */
-+ e_TRANS_SRC_DDR = 0x0F, /**< DDR controller */
-+ e_TRANS_SRC_CORE_INS_FETCH = 0x10, /**< Processor (instruction) */
-+ e_TRANS_SRC_CORE_DATA = 0x11, /**< Processor (data) */
-+ e_TRANS_SRC_DMA = 0x15 /**< DMA */
-+} e_TransSrc;
-+
-+/**************************************************************************//**
-+ @Description Local Access Window Target interface ID
-+*//***************************************************************************/
-+typedef enum e_P1023LawTargetId
-+{
-+ e_P1023_LAW_TARGET_PCIE_2 = 0x01, /**< PCI Express 2 target interface */
-+ e_P1023_LAW_TARGET_PCIE_1 = 0x02, /**< PCI Express 1 target interface */
-+ e_P1023_LAW_TARGET_PCIE_3 = 0x03, /**< PCI Express 3 target interface */
-+ e_P1023_LAW_TARGET_LBC = 0x04, /**< Local bus target interface */
-+ e_P1023_LAW_TARGET_QM_PORTALS = 0x0E, /**< Queue Manager Portals */
-+ e_P1023_LAW_TARGET_BM_PORTALS = 0x0E, /**< Buffer Manager Portals */
-+ e_P1023_LAW_TARGET_SRAM = 0x0E, /**< SRAM scratchpad */
-+ e_P1023_LAW_TARGET_DDR = 0x0F, /**< DDR target interface */
-+ e_P1023_LAW_TARGET_NONE = 0xFF /**< Invalid target interface */
-+} e_P1023LawTargetId;
-+
-+
-+/**************************************************************************//**
-+ @Group 1023_init_grp P1023 Initialization Unit
-+
-+ @Description P1023 initialization unit API functions, definitions and enums
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Description Part ID and revision number
-+*//***************************************************************************/
-+typedef enum e_P1023DeviceName
-+{
-+ e_P1023_REV_INVALID = 0x00000000, /**< Invalid revision */
-+ e_SC1023_REV_1_0 = (int)0x80FC0010, /**< SC1023 rev 1.0 */
-+ e_SC1023_REV_1_1 = (int)0x80FC0011, /**< SC1023 rev 1.1 */
-+ e_P1023_REV_1_0 = (int)0x80FE0010, /**< P1023 rev 1.0 with security */
-+ e_P1023_REV_1_1 = (int)0x80FE0011, /**< P1023 rev 1.1 with security */
-+ e_P1017_REV_1_1 = (int)0x80FF0011, /**< P1017 rev 1.1 with security */
-+ e_P1023_REV_1_0_NO_SEC = (int)0x80F60010, /**< P1023 rev 1.0 without security */
-+ e_P1023_REV_1_1_NO_SEC = (int)0x80F60011, /**< P1023 rev 1.1 without security */
-+ e_P1017_REV_1_1_NO_SEC = (int)0x80F70011 /**< P1017 rev 1.1 without security */
-+} e_P1023DeviceName;
-+
-+/**************************************************************************//**
-+ @Description structure representing P1023 initialization parameters
-+*//***************************************************************************/
-+typedef struct t_P1023Params
-+{
-+ uintptr_t ccsrBaseAddress; /**< CCSR base address (virtual) */
-+ uintptr_t bmPortalsBaseAddress; /**< Portals base address (virtual) */
-+ uintptr_t qmPortalsBaseAddress; /**< Portals base address (virtual) */
-+} t_P1023Params;
-+
-+/**************************************************************************//**
-+ @Function P1023_ConfigAndInit
-+
-+ @Description General initiation of the chip registers.
-+
-+ @Param[in] p_P1023Params - A pointer to data structure of parameters
-+
-+ @Return A handle to the P1023 data structure.
-+*//***************************************************************************/
-+t_Handle P1023_ConfigAndInit(t_P1023Params *p_P1023Params);
-+
-+/**************************************************************************//**
-+ @Function P1023_Free
-+
-+ @Description Free all resources.
-+
-+ @Param h_P1023 - (In) The handle of the initialized P1023 object.
-+
-+ @Return E_OK on success; Other value otherwise.
-+*//***************************************************************************/
-+t_Error P1023_Free(t_Handle h_P1023);
-+
-+/**************************************************************************//**
-+ @Function P1023_GetRevInfo
-+
-+ @Description This routine enables access to chip and revision information.
-+
-+ @Param[in] gutilBase - Base address of P1023 GUTIL registers.
-+
-+ @Return Part ID and revision.
-+*//***************************************************************************/
-+e_P1023DeviceName P1023_GetRevInfo(uintptr_t gutilBase);
-+
-+/**************************************************************************//**
-+ @Function P1023_GetE500Factor
-+
-+ @Description Returns E500 core clock multiplication factor.
-+
-+ @Param[in] gutilBase - Base address of P1023 GUTIL registers.
-+ @Param[in] coreId - Id of the requested core.
-+ @Param[out] p_E500MulFactor - Returns E500 to CCB multification factor.
-+ @Param[out] p_E500DivFactor - Returns E500 to CCB division factor.
-+
-+ @Return E_OK on success; Other value otherwise.
-+*
-+*//***************************************************************************/
-+t_Error P1023_GetE500Factor(uintptr_t gutilBase,
-+ uint32_t coreId,
-+ uint32_t *p_E500MulFactor,
-+ uint32_t *p_E500DivFactor);
-+
-+/**************************************************************************//**
-+ @Function P1023_GetFmFactor
-+
-+ @Description returns FM multiplication factors. (This value is returned using
-+ two parameters to avoid using float parameter).
-+
-+ @Param[in] gutilBase - Base address of P1023 GUTIL registers.
-+ @Param[out] p_FmMulFactor - returns E500 to CCB multification factor.
-+ @Param[out] p_FmDivFactor - returns E500 to CCB division factor.
-+
-+ @Return E_OK on success; Other value otherwise.
-+*//***************************************************************************/
-+t_Error P1023_GetFmFactor(uintptr_t gutilBase, uint32_t *p_FmMulFactor, uint32_t *p_FmDivFactor);
-+
-+/**************************************************************************//**
-+ @Function P1023_GetCcbFactor
-+
-+ @Description returns system multiplication factor.
-+
-+ @Param[in] gutilBase - Base address of P1023 GUTIL registers.
-+
-+ @Return System multiplication factor.
-+*//***************************************************************************/
-+uint32_t P1023_GetCcbFactor(uintptr_t gutilBase);
-+
-+#if 0
-+/**************************************************************************//**
-+ @Function P1023_GetDdrFactor
-+
-+ @Description returns the multiplication factor of the clock in for the DDR clock .
-+ Note: assumes the ddr_in_clk is identical to the sys_in_clk
-+
-+ @Param[in] gutilBase - Base address of P1023 GUTIL registers.
-+ @Param p_DdrMulFactor - returns DDR in clk multification factor.
-+ @Param p_DdrDivFactor - returns DDR division factor.
-+
-+ @Return E_OK on success; Other value otherwise..
-+*//***************************************************************************/
-+t_Error P1023_GetDdrFactor( uintptr_t gutilBase,
-+ uint32_t *p_DdrMulFactor,
-+ uint32_t *p_DdrDivFactor);
-+
-+/**************************************************************************//**
-+ @Function P1023_GetDdrType
-+
-+ @Description returns the multiplication factor of the clock in for the DDR clock .
-+
-+ @Param[in] gutilBase - Base address of P1023 GUTIL registers.
-+ @Param p_DdrType - (Out) returns DDR type DDR1/DDR2/DDR3.
-+
-+ @Return E_OK on success; Other value otherwise.
-+*//***************************************************************************/
-+t_Error P1023_GetDdrType(uintptr_t gutilBase, e_DdrType *p_DdrType );
-+#endif
-+
-+/** @} */ /* end of 1023_init_grp group */
-+/** @} */ /* end of 1023_grp group */
-+
-+#define CORE_E500V2
-+
-+#if 0 /* using unified values */
-+/*****************************************************************************
-+ INTEGRATION-SPECIFIC MODULE CODES
-+******************************************************************************/
-+#define MODULE_UNKNOWN 0x00000000
-+#define MODULE_MEM 0x00010000
-+#define MODULE_MM 0x00020000
-+#define MODULE_CORE 0x00030000
-+#define MODULE_P1023 0x00040000
-+#define MODULE_MII 0x00050000
-+#define MODULE_PM 0x00060000
-+#define MODULE_MMU 0x00070000
-+#define MODULE_PIC 0x00080000
-+#define MODULE_L2_CACHE 0x00090000
-+#define MODULE_DUART 0x000a0000
-+#define MODULE_SERDES 0x000b0000
-+#define MODULE_PIO 0x000c0000
-+#define MODULE_QM 0x000d0000
-+#define MODULE_BM 0x000e0000
-+#define MODULE_SEC 0x000f0000
-+#define MODULE_FM 0x00100000
-+#define MODULE_FM_MURAM 0x00110000
-+#define MODULE_FM_PCD 0x00120000
-+#define MODULE_FM_RTC 0x00130000
-+#define MODULE_FM_MAC 0x00140000
-+#define MODULE_FM_PORT 0x00150000
-+#define MODULE_FM_MACSEC 0x00160000
-+#define MODULE_FM_MACSEC_SECY 0x00170000
-+#define MODULE_FM_SP 0x00280000
-+#define MODULE_ECM 0x00190000
-+#define MODULE_DMA 0x001a0000
-+#define MODULE_DDR 0x001b0000
-+#define MODULE_LAW 0x001c0000
-+#define MODULE_LBC 0x001d0000
-+#define MODULE_I2C 0x001e0000
-+#define MODULE_ESPI 0x001f0000
-+#define MODULE_PCI 0x00200000
-+#define MODULE_DPA_PORT 0x00210000
-+#define MODULE_USB 0x00220000
-+#endif /* using unified values */
-+
-+/*****************************************************************************
-+ LBC INTEGRATION-SPECIFIC DEFINITIONS
-+******************************************************************************/
-+/**************************************************************************//**
-+ @Group lbc_exception_grp LBC Exception Unit
-+
-+ @Description LBC Exception unit API functions, definitions and enums
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Anchor lbc_exbm
-+
-+ @Collection LBC Errors Bit Mask
-+
-+ These errors are reported through the exceptions callback..
-+ The values can be or'ed in any combination in the errors mask
-+ parameter of the errors report structure.
-+
-+ These errors can also be passed as a bit-mask to
-+ LBC_EnableErrorChecking() or LBC_DisableErrorChecking(),
-+ for enabling or disabling error checking.
-+ @{
-+*//***************************************************************************/
-+#define LBC_ERR_BUS_MONITOR 0x80000000 /**< Bus monitor error */
-+#define LBC_ERR_PARITY_ECC 0x20000000 /**< Parity error for GPCM/UPM */
-+#define LBC_ERR_WRITE_PROTECT 0x04000000 /**< Write protection error */
-+#define LBC_ERR_CHIP_SELECT 0x00080000 /**< Unrecognized chip select */
-+
-+#define LBC_ERR_ALL (LBC_ERR_BUS_MONITOR | LBC_ERR_PARITY_ECC | \
-+ LBC_ERR_WRITE_PROTECT | LBC_ERR_CHIP_SELECT)
-+ /**< All possible errors */
-+/* @} */
-+/** @} */ /* end of lbc_exception_grp group */
-+
-+#define LBC_NUM_OF_BANKS 2
-+#define LBC_MAX_CS_SIZE 0x0000000100000000LL
-+#define LBC_ATOMIC_OPERATION_SUPPORT
-+#define LBC_PARITY_SUPPORT
-+#define LBC_ADDRESS_SHIFT_SUPPORT
-+#define LBC_ADDRESS_HOLD_TIME_CTRL
-+#define LBC_HIGH_CLK_DIVIDERS
-+#define LBC_FCM_AVAILABLE
-+
-+
-+/*****************************************************************************
-+ LAW INTEGRATION-SPECIFIC DEFINITIONS
-+******************************************************************************/
-+#define LAW_ARCH_CCB
-+#define LAW_NUM_OF_WINDOWS 12
-+#define LAW_MIN_WINDOW_SIZE 0x0000000000001000LL /**< 4KB */
-+#define LAW_MAX_WINDOW_SIZE 0x0000001000000000LL /**< 32GB */
-+
-+
-+/*****************************************************************************
-+ SPI INTEGRATION-SPECIFIC DEFINITIONS
-+******************************************************************************/
-+#define SPI_NUM_OF_CONTROLLERS 1
-+
-+/*****************************************************************************
-+ PCI/PCIe INTEGRATION-SPECIFIC DEFINITIONS
-+******************************************************************************/
-+
-+#define PCI_MAX_INBOUND_WINDOWS_NUM 4
-+#define PCI_MAX_OUTBOUND_WINDOWS_NUM 5
-+
-+/**************************************************************************//**
-+ @Description Target interface of an inbound window
-+*//***************************************************************************/
-+typedef enum e_PciTargetInterface
-+{
-+ e_PCI_TARGET_PCIE_2 = 0x1, /**< PCI Express target interface 2 */
-+ e_PCI_TARGET_PCIE_1 = 0x2, /**< PCI Express target interface 1 */
-+ e_PCI_TARGET_PCIE_3 = 0x3, /**< PCI Express target interface 3 */
-+ e_PCI_TARGET_LOCAL_MEMORY = 0xF /**< Local Memory (DDR SDRAM, Local Bus, SRAM) target interface */
-+
-+} e_PciTargetInterface;
-+
-+/*****************************************************************************
-+ DDR INTEGRATION-SPECIFIC DEFINITIONS
-+******************************************************************************/
-+#define DDR_NUM_OF_VALID_CS 2
-+
-+/*****************************************************************************
-+ SEC INTEGRATION-SPECIFIC DEFINITIONS
-+******************************************************************************/
-+#define SEC_ERRATA_STAT_REGS_UNUSABLE
-+
-+/*****************************************************************************
-+ DMA INTEGRATION-SPECIFIC DEFINITIONS
-+******************************************************************************/
-+#define DMA_NUM_OF_CONTROLLERS 2
-+
-+
-+
-+
-+/*****************************************************************************
-+ 1588 INTEGRATION-SPECIFIC DEFINITIONS
-+******************************************************************************/
-+#define PTP_V2
-+
-+/**************************************************************************//**
-+ @Function P1023_GetMuxControlReg
-+
-+ @Description Returns the value of PMUXCR (Alternate Function Signal Multiplex
-+ Control Register)
-+
-+ @Param[in] gutilBase - Base address of P1023 GUTIL registers.
-+
-+ @Return Value of PMUXCR
-+*//***************************************************************************/
-+uint32_t P1023_GetMuxControlReg(uintptr_t gutilBase);
-+
-+/**************************************************************************//**
-+ @Function P1023_SetMuxControlReg
-+
-+ @Description Sets the value of PMUXCR (Alternate Function Signal Multiplex
-+ Control Register)
-+
-+ @Param[in] gutilBase - Base address of P1023 GUTIL registers.
-+ @Param[in] val - the new value for PMUXCR.
-+
-+ @Return None
-+*//***************************************************************************/
-+void P1023_SetMuxControlReg(uintptr_t gutilBase, uint32_t val);
-+
-+/**************************************************************************//**
-+ @Function P1023_GetDeviceDisableStatusRegister
-+
-+ @Description Returns the value of DEVDISR (Device Disable Register)
-+
-+ @Param[in] gutilBase - Base address of P1023 GUTIL registers.
-+
-+ @Return Value of DEVDISR
-+*//***************************************************************************/
-+uint32_t P1023_GetDeviceDisableStatusRegister(uintptr_t gutilBase);
-+
-+/**************************************************************************//**
-+ @Function P1023_GetPorDeviceStatusRegister
-+
-+ @Description Returns the value of POR Device Status Register
-+
-+ @Param[in] gutilBase - Base address of P1023 GUTIL registers.
-+
-+ @Return POR Device Status Register
-+*//***************************************************************************/
-+uint32_t P1023_GetPorDeviceStatusRegister(uintptr_t gutilBase);
-+
-+/**************************************************************************//**
-+ @Function P1023_GetPorBootModeStatusRegister
-+
-+ @Description Returns the value of POR Boot Mode Status Register
-+
-+ @Param[in] gutilBase - Base address of P1023 GUTIL registers.
-+
-+ @Return POR Boot Mode Status Register value
-+*//***************************************************************************/
-+uint32_t P1023_GetPorBootModeStatusRegister(uintptr_t gutilBase);
-+
-+
-+#define PORDEVSR_SGMII1_DIS 0x10000000
-+#define PORDEVSR_SGMII2_DIS 0x08000000
-+#define PORDEVSR_ECP1 0x02000000
-+#define PORDEVSR_IO_SEL 0x00780000
-+#define PORDEVSR_IO_SEL_SHIFT 19
-+#define PORBMSR_HA 0x00070000
-+#define PORBMSR_HA_SHIFT 16
-+
-+#define DEVDISR_QM_BM 0x80000000
-+#define DEVDISR_FM 0x40000000
-+#define DEVDISR_PCIE1 0x20000000
-+#define DEVDISR_MAC_SEC 0x10000000
-+#define DEVDISR_ELBC 0x08000000
-+#define DEVDISR_PCIE2 0x04000000
-+#define DEVDISR_PCIE3 0x02000000
-+#define DEVDISR_CAAM 0x01000000
-+#define DEVDISR_USB0 0x00800000
-+#define DEVDISR_1588 0x00020000
-+#define DEVDISR_CORE0 0x00008000
-+#define DEVDISR_TB0 0x00004000
-+#define DEVDISR_CORE1 0x00002000
-+#define DEVDISR_TB1 0x00001000
-+#define DEVDISR_DMA1 0x00000400
-+#define DEVDISR_DMA2 0x00000200
-+#define DEVDISR_DDR 0x00000010
-+#define DEVDISR_TSEC1 0x00000080
-+#define DEVDISR_TSEC2 0x00000040
-+#define DEVDISR_SPI 0x00000008
-+#define DEVDISR_I2C 0x00000004
-+#define DEVDISR_DUART 0x00000002
-+
-+
-+#endif /* __PART_INTEGRATION_EXT_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/integrations/P3040_P4080_P5020/dpaa_integration_ext.h
-@@ -0,0 +1,276 @@
-+/* Copyright (c) 2009-2012 Freescale Semiconductor, Inc
-+ * All rights reserved.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+/**************************************************************************//**
-+ @File dpaa_integration_ext.h
-+
-+ @Description P3040/P4080/P5020 FM external definitions and structures.
-+*//***************************************************************************/
-+#ifndef __DPAA_INTEGRATION_EXT_H
-+#define __DPAA_INTEGRATION_EXT_H
-+
-+#include "std_ext.h"
-+
-+
-+#define DPAA_VERSION 10
-+
-+typedef enum {
-+ e_DPAA_SWPORTAL0 = 0,
-+ e_DPAA_SWPORTAL1,
-+ e_DPAA_SWPORTAL2,
-+ e_DPAA_SWPORTAL3,
-+ e_DPAA_SWPORTAL4,
-+ e_DPAA_SWPORTAL5,
-+ e_DPAA_SWPORTAL6,
-+ e_DPAA_SWPORTAL7,
-+ e_DPAA_SWPORTAL8,
-+ e_DPAA_SWPORTAL9,
-+ e_DPAA_SWPORTAL_DUMMY_LAST
-+} e_DpaaSwPortal;
-+
-+typedef enum {
-+ e_DPAA_DCPORTAL0 = 0,
-+ e_DPAA_DCPORTAL1,
-+ e_DPAA_DCPORTAL2,
-+ e_DPAA_DCPORTAL3,
-+ e_DPAA_DCPORTAL4,
-+ e_DPAA_DCPORTAL_DUMMY_LAST
-+} e_DpaaDcPortal;
-+
-+#define DPAA_MAX_NUM_OF_SW_PORTALS e_DPAA_SWPORTAL_DUMMY_LAST
-+#define DPAA_MAX_NUM_OF_DC_PORTALS e_DPAA_DCPORTAL_DUMMY_LAST
-+
-+/*****************************************************************************
-+ QMan INTEGRATION-SPECIFIC DEFINITIONS
-+******************************************************************************/
-+#define QM_MAX_NUM_OF_POOL_CHANNELS 15 /**< Total number of channels, dedicated and pool */
-+#define QM_MAX_NUM_OF_WQ 8 /**< Number of work queues per channel */
-+#define QM_MAX_NUM_OF_SWP_AS 4
-+#define QM_MAX_NUM_OF_CGS 256 /**< Number of congestion groups */
-+#define QM_MAX_NUM_OF_FQIDS (16 * MEGABYTE) /**< FQIDs range - 24 bits */
-+
-+/**************************************************************************//**
-+ @Description Work Queue Channel assignments in QMan.
-+*//***************************************************************************/
-+typedef enum
-+{
-+ e_QM_FQ_CHANNEL_SWPORTAL0 = 0, /**< Dedicated channels serviced by software portals 0 to 9 */
-+ e_QM_FQ_CHANNEL_SWPORTAL1,
-+ e_QM_FQ_CHANNEL_SWPORTAL2,
-+ e_QM_FQ_CHANNEL_SWPORTAL3,
-+ e_QM_FQ_CHANNEL_SWPORTAL4,
-+ e_QM_FQ_CHANNEL_SWPORTAL5,
-+ e_QM_FQ_CHANNEL_SWPORTAL6,
-+ e_QM_FQ_CHANNEL_SWPORTAL7,
-+ e_QM_FQ_CHANNEL_SWPORTAL8,
-+ e_QM_FQ_CHANNEL_SWPORTAL9,
-+
-+ e_QM_FQ_CHANNEL_POOL1 = 0x21, /**< Pool channels that can be serviced by any of the software portals */
-+ e_QM_FQ_CHANNEL_POOL2,
-+ e_QM_FQ_CHANNEL_POOL3,
-+ e_QM_FQ_CHANNEL_POOL4,
-+ e_QM_FQ_CHANNEL_POOL5,
-+ e_QM_FQ_CHANNEL_POOL6,
-+ e_QM_FQ_CHANNEL_POOL7,
-+ e_QM_FQ_CHANNEL_POOL8,
-+ e_QM_FQ_CHANNEL_POOL9,
-+ e_QM_FQ_CHANNEL_POOL10,
-+ e_QM_FQ_CHANNEL_POOL11,
-+ e_QM_FQ_CHANNEL_POOL12,
-+ e_QM_FQ_CHANNEL_POOL13,
-+ e_QM_FQ_CHANNEL_POOL14,
-+ e_QM_FQ_CHANNEL_POOL15,
-+
-+ e_QM_FQ_CHANNEL_FMAN0_SP0 = 0x40, /**< Dedicated channels serviced by Direct Connect Portal 0:
-+ connected to FMan 0; assigned in incrementing order to
-+ each sub-portal (SP) in the portal */
-+ e_QM_FQ_CHANNEL_FMAN0_SP1,
-+ e_QM_FQ_CHANNEL_FMAN0_SP2,
-+ e_QM_FQ_CHANNEL_FMAN0_SP3,
-+ e_QM_FQ_CHANNEL_FMAN0_SP4,
-+ e_QM_FQ_CHANNEL_FMAN0_SP5,
-+ e_QM_FQ_CHANNEL_FMAN0_SP6,
-+ e_QM_FQ_CHANNEL_FMAN0_SP7,
-+ e_QM_FQ_CHANNEL_FMAN0_SP8,
-+ e_QM_FQ_CHANNEL_FMAN0_SP9,
-+ e_QM_FQ_CHANNEL_FMAN0_SP10,
-+ e_QM_FQ_CHANNEL_FMAN0_SP11,
-+/* difference between 5020 and 4080 :) */
-+ e_QM_FQ_CHANNEL_FMAN1_SP0 = 0x60,
-+ e_QM_FQ_CHANNEL_FMAN1_SP1,
-+ e_QM_FQ_CHANNEL_FMAN1_SP2,
-+ e_QM_FQ_CHANNEL_FMAN1_SP3,
-+ e_QM_FQ_CHANNEL_FMAN1_SP4,
-+ e_QM_FQ_CHANNEL_FMAN1_SP5,
-+ e_QM_FQ_CHANNEL_FMAN1_SP6,
-+ e_QM_FQ_CHANNEL_FMAN1_SP7,
-+ e_QM_FQ_CHANNEL_FMAN1_SP8,
-+ e_QM_FQ_CHANNEL_FMAN1_SP9,
-+ e_QM_FQ_CHANNEL_FMAN1_SP10,
-+ e_QM_FQ_CHANNEL_FMAN1_SP11,
-+
-+ e_QM_FQ_CHANNEL_CAAM = 0x80, /**< Dedicated channel serviced by Direct Connect Portal 2:
-+ connected to SEC 4.x */
-+
-+ e_QM_FQ_CHANNEL_PME = 0xA0, /**< Dedicated channel serviced by Direct Connect Portal 3:
-+ connected to PME */
-+ e_QM_FQ_CHANNEL_RAID = 0xC0 /**< Dedicated channel serviced by Direct Connect Portal 4:
-+ connected to RAID */
-+} e_QmFQChannel;
-+
-+/*****************************************************************************
-+ BMan INTEGRATION-SPECIFIC DEFINITIONS
-+******************************************************************************/
-+#define BM_MAX_NUM_OF_POOLS 64 /**< Number of buffers pools */
-+
-+
-+/*****************************************************************************
-+ FM INTEGRATION-SPECIFIC DEFINITIONS
-+******************************************************************************/
-+#define INTG_MAX_NUM_OF_FM 2
-+
-+/* Ports defines */
-+#define FM_MAX_NUM_OF_1G_MACS 5
-+#define FM_MAX_NUM_OF_10G_MACS 1
-+#define FM_MAX_NUM_OF_MACS (FM_MAX_NUM_OF_1G_MACS + FM_MAX_NUM_OF_10G_MACS)
-+#define FM_MAX_NUM_OF_OH_PORTS 7
-+
-+#define FM_MAX_NUM_OF_1G_RX_PORTS FM_MAX_NUM_OF_1G_MACS
-+#define FM_MAX_NUM_OF_10G_RX_PORTS FM_MAX_NUM_OF_10G_MACS
-+#define FM_MAX_NUM_OF_RX_PORTS (FM_MAX_NUM_OF_10G_RX_PORTS + FM_MAX_NUM_OF_1G_RX_PORTS)
-+
-+#define FM_MAX_NUM_OF_1G_TX_PORTS FM_MAX_NUM_OF_1G_MACS
-+#define FM_MAX_NUM_OF_10G_TX_PORTS FM_MAX_NUM_OF_10G_MACS
-+#define FM_MAX_NUM_OF_TX_PORTS (FM_MAX_NUM_OF_10G_TX_PORTS + FM_MAX_NUM_OF_1G_TX_PORTS)
-+
-+#define FM_PORT_MAX_NUM_OF_EXT_POOLS 8 /**< Number of external BM pools per Rx port */
-+#define FM_PORT_NUM_OF_CONGESTION_GRPS 256 /**< Total number of congestion groups in QM */
-+#define FM_MAX_NUM_OF_SUB_PORTALS 12
-+#define FM_PORT_MAX_NUM_OF_OBSERVED_EXT_POOLS 0
-+
-+/* Rams defines */
-+#define FM_MURAM_SIZE (160*KILOBYTE)
-+#define FM_IRAM_SIZE(major, minor) (64 * KILOBYTE)
-+#define FM_NUM_OF_CTRL 2
-+
-+/* PCD defines */
-+#define FM_PCD_PLCR_NUM_ENTRIES 256 /**< Total number of policer profiles */
-+#define FM_PCD_KG_NUM_OF_SCHEMES 32 /**< Total number of KG schemes */
-+#define FM_PCD_MAX_NUM_OF_CLS_PLANS 256 /**< Number of classification plan entries. */
-+#define FM_PCD_PRS_SW_PATCHES_SIZE 0x00000200 /**< Number of bytes saved for patches */
-+#define FM_PCD_SW_PRS_SIZE 0x00000800 /**< Total size of SW parser area */
-+
-+/* RTC defines */
-+#define FM_RTC_NUM_OF_ALARMS 2 /**< RTC number of alarms */
-+#define FM_RTC_NUM_OF_PERIODIC_PULSES 2 /**< RTC number of periodic pulses */
-+#define FM_RTC_NUM_OF_EXT_TRIGGERS 2 /**< RTC number of external triggers */
-+
-+/* QMI defines */
-+#define QMI_MAX_NUM_OF_TNUMS 64
-+#define QMI_DEF_TNUMS_THRESH 48
-+
-+/* FPM defines */
-+#define FM_NUM_OF_FMAN_CTRL_EVENT_REGS 4
-+
-+/* DMA defines */
-+#define DMA_THRESH_MAX_COMMQ 31
-+#define DMA_THRESH_MAX_BUF 127
-+
-+/* BMI defines */
-+#define BMI_MAX_NUM_OF_TASKS 128
-+#define BMI_MAX_NUM_OF_DMAS 32
-+#define BMI_MAX_FIFO_SIZE (FM_MURAM_SIZE)
-+#define PORT_MAX_WEIGHT 16
-+
-+
-+#define FM_CHECK_PORT_RESTRICTIONS(__validPorts, __newPortIndx) TRUE
-+
-+/* p4080-rev1 unique features */
-+#define QM_CGS_NO_FRAME_MODE
-+
-+/* p4080 unique features */
-+#define FM_NO_DISPATCH_RAM_ECC
-+#define FM_NO_WATCHDOG
-+#define FM_NO_TNUM_AGING
-+#define FM_KG_NO_BYPASS_FQID_GEN
-+#define FM_KG_NO_BYPASS_PLCR_PROFILE_GEN
-+#define FM_NO_BACKUP_POOLS
-+#define FM_NO_OP_OBSERVED_POOLS
-+#define FM_NO_ADVANCED_RATE_LIMITER
-+#define FM_NO_OP_OBSERVED_CGS
-+#define FM_HAS_TOTAL_DMAS
-+#define FM_KG_NO_IPPID_SUPPORT
-+#define FM_NO_GUARANTEED_RESET_VALUES
-+#define FM_MAC_RESET
-+
-+/* FM erratas */
-+#define FM_TX_ECC_FRMS_ERRATA_10GMAC_A004
-+#define FM_TX_SHORT_FRAME_BAD_TS_ERRATA_10GMAC_A006 /* No implementation, Out of LLD scope */
-+#define FM_TX_FIFO_CORRUPTION_ERRATA_10GMAC_A007
-+#define FM_ECC_HALT_NO_SYNC_ERRATA_10GMAC_A008
-+#define FM_TX_INVALID_ECC_ERRATA_10GMAC_A009 /* Out of LLD scope, user may disable ECC exceptions using FM_DisableRamsEcc */
-+#define FM_BAD_VLAN_DETECT_ERRATA_10GMAC_A010
-+
-+#define FM_RX_PREAM_4_ERRATA_DTSEC_A001
-+#define FM_GRS_ERRATA_DTSEC_A002
-+#define FM_BAD_TX_TS_IN_B_2_B_ERRATA_DTSEC_A003
-+#define FM_GTS_ERRATA_DTSEC_A004
-+#define FM_GTS_AFTER_MAC_ABORTED_FRAME_ERRATA_DTSEC_A0012
-+#define FM_GTS_UNDERRUN_ERRATA_DTSEC_A0014
-+#define FM_GTS_AFTER_DROPPED_FRAME_ERRATA_DTSEC_A004839
-+
-+#define FM_MAGIC_PACKET_UNRECOGNIZED_ERRATA_DTSEC2 /* No implementation, Out of LLD scope */
-+#define FM_TX_LOCKUP_ERRATA_DTSEC6
-+
-+#define FM_HC_DEF_FQID_ONLY_ERRATA_FMAN_A003 /* Implemented by ucode */
-+#define FM_DEBUG_TRACE_FMAN_A004 /* No implementation, Out of LLD scope */
-+
-+#define FM_UCODE_NOT_RESET_ERRATA_BUGZILLA6173
-+
-+#define FM_10G_REM_N_LCL_FLT_EX_10GMAC_ERRATA_SW005
-+
-+#define FM_LEN_CHECK_ERRATA_FMAN_SW002
-+
-+#define FM_NO_CTXA_COPY_ERRATA_FMAN_SW001
-+#define FM_KG_ERASE_FLOW_ID_ERRATA_FMAN_SW004
-+
-+/*****************************************************************************
-+ FM MACSEC INTEGRATION-SPECIFIC DEFINITIONS
-+******************************************************************************/
-+#define NUM_OF_RX_SC 16
-+#define NUM_OF_TX_SC 16
-+
-+#define NUM_OF_SA_PER_RX_SC 2
-+#define NUM_OF_SA_PER_TX_SC 2
-+
-+
-+#endif /* __DPAA_INTEGRATION_EXT_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/integrations/P3040_P4080_P5020/part_ext.h
-@@ -0,0 +1,83 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+/**************************************************************************//**
-+
-+ @File part_ext.h
-+
-+ @Description Definitions for the part (integration) module.
-+*//***************************************************************************/
-+
-+#ifndef __PART_EXT_H
-+#define __PART_EXT_H
-+
-+#include "std_ext.h"
-+#include "part_integration_ext.h"
-+
-+
-+#if !(defined(MPC8306) || \
-+ defined(MPC8309) || \
-+ defined(MPC834x) || \
-+ defined(MPC836x) || \
-+ defined(MPC832x) || \
-+ defined(MPC837x) || \
-+ defined(MPC8568) || \
-+ defined(MPC8569) || \
-+ defined(P1020) || \
-+ defined(P1021) || \
-+ defined(P1022) || \
-+ defined(P1023) || \
-+ defined(P2020) || \
-+ defined(P2040) || \
-+ defined(P3041) || \
-+ defined(P4080) || \
-+ defined(SC4080) || \
-+ defined(P5020) || \
-+ defined(MSC814x))
-+#error "unable to proceed without chip-definition"
-+#endif /* !(defined(MPC834x) || ... */
-+
-+
-+/**************************************************************************//*
-+ @Description Part data structure - must be contained in any integration
-+ data structure.
-+*//***************************************************************************/
-+typedef struct t_Part
-+{
-+ uintptr_t (* f_GetModuleBase)(t_Handle h_Part, e_ModuleId moduleId);
-+ /**< Returns the address of the module's memory map base. */
-+ e_ModuleId (* f_GetModuleIdByBase)(t_Handle h_Part, uintptr_t baseAddress);
-+ /**< Returns the module's ID according to its memory map base. */
-+} t_Part;
-+
-+
-+#endif /* __PART_EXT_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/integrations/P3040_P4080_P5020/part_integration_ext.h
-@@ -0,0 +1,336 @@
-+/* Copyright (c) 2008-2012 Freescale Semiconductor, Inc
-+ * All rights reserved.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+/**************************************************************************//**
-+ @File part_integration_ext.h
-+
-+ @Description P3040/P4080/P5020 external definitions and structures.
-+*//***************************************************************************/
-+#ifndef __PART_INTEGRATION_EXT_H
-+#define __PART_INTEGRATION_EXT_H
-+
-+#include "std_ext.h"
-+#include "dpaa_integration_ext.h"
-+
-+
-+/**************************************************************************//**
-+ @Group P3040/P4080/P5020_chip_id P5020 Application Programming Interface
-+
-+ @Description P3040/P4080/P5020 Chip functions,definitions and enums.
-+
-+ @{
-+*//***************************************************************************/
-+
-+#define CORE_E500MC
-+
-+#define INTG_MAX_NUM_OF_CORES 1
-+
-+
-+/**************************************************************************//**
-+ @Description Module types.
-+*//***************************************************************************/
-+typedef enum e_ModuleId
-+{
-+ e_MODULE_ID_DUART_1 = 0,
-+ e_MODULE_ID_DUART_2,
-+ e_MODULE_ID_DUART_3,
-+ e_MODULE_ID_DUART_4,
-+ e_MODULE_ID_LAW,
-+ e_MODULE_ID_LBC,
-+ e_MODULE_ID_PAMU,
-+ e_MODULE_ID_QM, /**< Queue manager module */
-+ e_MODULE_ID_BM, /**< Buffer manager module */
-+ e_MODULE_ID_QM_CE_PORTAL_0,
-+ e_MODULE_ID_QM_CI_PORTAL_0,
-+ e_MODULE_ID_QM_CE_PORTAL_1,
-+ e_MODULE_ID_QM_CI_PORTAL_1,
-+ e_MODULE_ID_QM_CE_PORTAL_2,
-+ e_MODULE_ID_QM_CI_PORTAL_2,
-+ e_MODULE_ID_QM_CE_PORTAL_3,
-+ e_MODULE_ID_QM_CI_PORTAL_3,
-+ e_MODULE_ID_QM_CE_PORTAL_4,
-+ e_MODULE_ID_QM_CI_PORTAL_4,
-+ e_MODULE_ID_QM_CE_PORTAL_5,
-+ e_MODULE_ID_QM_CI_PORTAL_5,
-+ e_MODULE_ID_QM_CE_PORTAL_6,
-+ e_MODULE_ID_QM_CI_PORTAL_6,
-+ e_MODULE_ID_QM_CE_PORTAL_7,
-+ e_MODULE_ID_QM_CI_PORTAL_7,
-+ e_MODULE_ID_QM_CE_PORTAL_8,
-+ e_MODULE_ID_QM_CI_PORTAL_8,
-+ e_MODULE_ID_QM_CE_PORTAL_9,
-+ e_MODULE_ID_QM_CI_PORTAL_9,
-+ e_MODULE_ID_BM_CE_PORTAL_0,
-+ e_MODULE_ID_BM_CI_PORTAL_0,
-+ e_MODULE_ID_BM_CE_PORTAL_1,
-+ e_MODULE_ID_BM_CI_PORTAL_1,
-+ e_MODULE_ID_BM_CE_PORTAL_2,
-+ e_MODULE_ID_BM_CI_PORTAL_2,
-+ e_MODULE_ID_BM_CE_PORTAL_3,
-+ e_MODULE_ID_BM_CI_PORTAL_3,
-+ e_MODULE_ID_BM_CE_PORTAL_4,
-+ e_MODULE_ID_BM_CI_PORTAL_4,
-+ e_MODULE_ID_BM_CE_PORTAL_5,
-+ e_MODULE_ID_BM_CI_PORTAL_5,
-+ e_MODULE_ID_BM_CE_PORTAL_6,
-+ e_MODULE_ID_BM_CI_PORTAL_6,
-+ e_MODULE_ID_BM_CE_PORTAL_7,
-+ e_MODULE_ID_BM_CI_PORTAL_7,
-+ e_MODULE_ID_BM_CE_PORTAL_8,
-+ e_MODULE_ID_BM_CI_PORTAL_8,
-+ e_MODULE_ID_BM_CE_PORTAL_9,
-+ e_MODULE_ID_BM_CI_PORTAL_9,
-+ e_MODULE_ID_FM1, /**< Frame manager #1 module */
-+ e_MODULE_ID_FM1_RTC, /**< FM Real-Time-Clock */
-+ e_MODULE_ID_FM1_MURAM, /**< FM Multi-User-RAM */
-+ e_MODULE_ID_FM1_BMI, /**< FM BMI block */
-+ e_MODULE_ID_FM1_QMI, /**< FM QMI block */
-+ e_MODULE_ID_FM1_PRS, /**< FM parser block */
-+ e_MODULE_ID_FM1_PORT_HO0, /**< FM Host-command/offline-parsing port block */
-+ e_MODULE_ID_FM1_PORT_HO1, /**< FM Host-command/offline-parsing port block */
-+ e_MODULE_ID_FM1_PORT_HO2, /**< FM Host-command/offline-parsing port block */
-+ e_MODULE_ID_FM1_PORT_HO3, /**< FM Host-command/offline-parsing port block */
-+ e_MODULE_ID_FM1_PORT_HO4, /**< FM Host-command/offline-parsing port block */
-+ e_MODULE_ID_FM1_PORT_HO5, /**< FM Host-command/offline-parsing port block */
-+ e_MODULE_ID_FM1_PORT_HO6, /**< FM Host-command/offline-parsing port block */
-+ e_MODULE_ID_FM1_PORT_1GRx0, /**< FM Rx 1G MAC port block */
-+ e_MODULE_ID_FM1_PORT_1GRx1, /**< FM Rx 1G MAC port block */
-+ e_MODULE_ID_FM1_PORT_1GRx2, /**< FM Rx 1G MAC port block */
-+ e_MODULE_ID_FM1_PORT_1GRx3, /**< FM Rx 1G MAC port block */
-+ e_MODULE_ID_FM1_PORT_1GRx4, /**< FM Rx 1G MAC port block */
-+ e_MODULE_ID_FM1_PORT_10GRx0, /**< FM Rx 10G MAC port block */
-+ e_MODULE_ID_FM1_PORT_1GTx0, /**< FM Tx 1G MAC port block */
-+ e_MODULE_ID_FM1_PORT_1GTx1, /**< FM Tx 1G MAC port block */
-+ e_MODULE_ID_FM1_PORT_1GTx2, /**< FM Tx 1G MAC port block */
-+ e_MODULE_ID_FM1_PORT_1GTx3, /**< FM Tx 1G MAC port block */
-+ e_MODULE_ID_FM1_PORT_1GTx4, /**< FM Tx 1G MAC port block */
-+ e_MODULE_ID_FM1_PORT_10GTx0, /**< FM Tx 10G MAC port block */
-+ e_MODULE_ID_FM1_PLCR, /**< FM Policer */
-+ e_MODULE_ID_FM1_KG, /**< FM Keygen */
-+ e_MODULE_ID_FM1_DMA, /**< FM DMA */
-+ e_MODULE_ID_FM1_FPM, /**< FM FPM */
-+ e_MODULE_ID_FM1_IRAM, /**< FM Instruction-RAM */
-+ e_MODULE_ID_FM1_1GMDIO0, /**< FM 1G MDIO MAC 0*/
-+ e_MODULE_ID_FM1_1GMDIO1, /**< FM 1G MDIO MAC 1*/
-+ e_MODULE_ID_FM1_1GMDIO2, /**< FM 1G MDIO MAC 2*/
-+ e_MODULE_ID_FM1_1GMDIO3, /**< FM 1G MDIO MAC 3*/
-+ e_MODULE_ID_FM1_10GMDIO, /**< FM 10G MDIO */
-+ e_MODULE_ID_FM1_PRS_IRAM, /**< FM SW-parser Instruction-RAM */
-+ e_MODULE_ID_FM1_1GMAC0, /**< FM 1G MAC #0 */
-+ e_MODULE_ID_FM1_1GMAC1, /**< FM 1G MAC #1 */
-+ e_MODULE_ID_FM1_1GMAC2, /**< FM 1G MAC #2 */
-+ e_MODULE_ID_FM1_1GMAC3, /**< FM 1G MAC #3 */
-+ e_MODULE_ID_FM1_10GMAC0, /**< FM 10G MAC #0 */
-+
-+ e_MODULE_ID_FM2, /**< Frame manager #2 module */
-+ e_MODULE_ID_FM2_RTC, /**< FM Real-Time-Clock */
-+ e_MODULE_ID_FM2_MURAM, /**< FM Multi-User-RAM */
-+ e_MODULE_ID_FM2_BMI, /**< FM BMI block */
-+ e_MODULE_ID_FM2_QMI, /**< FM QMI block */
-+ e_MODULE_ID_FM2_PRS, /**< FM parser block */
-+ e_MODULE_ID_FM2_PORT_HO0, /**< FM Host-command/offline-parsing port block */
-+ e_MODULE_ID_FM2_PORT_HO1, /**< FM Host-command/offline-parsing port block */
-+ e_MODULE_ID_FM2_PORT_HO2, /**< FM Host-command/offline-parsing port block */
-+ e_MODULE_ID_FM2_PORT_HO3, /**< FM Host-command/offline-parsing port block */
-+ e_MODULE_ID_FM2_PORT_HO4, /**< FM Host-command/offline-parsing port block */
-+ e_MODULE_ID_FM2_PORT_HO5, /**< FM Host-command/offline-parsing port block */
-+ e_MODULE_ID_FM2_PORT_HO6, /**< FM Host-command/offline-parsing port block */
-+ e_MODULE_ID_FM2_PORT_1GRx0, /**< FM Rx 1G MAC port block */
-+ e_MODULE_ID_FM2_PORT_1GRx1, /**< FM Rx 1G MAC port block */
-+ e_MODULE_ID_FM2_PORT_1GRx2, /**< FM Rx 1G MAC port block */
-+ e_MODULE_ID_FM2_PORT_1GRx3, /**< FM Rx 1G MAC port block */
-+ e_MODULE_ID_FM2_PORT_10GRx0, /**< FM Rx 10G MAC port block */
-+ e_MODULE_ID_FM2_PORT_1GTx0, /**< FM Tx 1G MAC port block */
-+ e_MODULE_ID_FM2_PORT_1GTx1, /**< FM Tx 1G MAC port block */
-+ e_MODULE_ID_FM2_PORT_1GTx2, /**< FM Tx 1G MAC port block */
-+ e_MODULE_ID_FM2_PORT_1GTx3, /**< FM Tx 1G MAC port block */
-+ e_MODULE_ID_FM2_PORT_10GTx0, /**< FM Tx 10G MAC port block */
-+ e_MODULE_ID_FM2_PLCR, /**< FM Policer */
-+ e_MODULE_ID_FM2_KG, /**< FM Keygen */
-+ e_MODULE_ID_FM2_DMA, /**< FM DMA */
-+ e_MODULE_ID_FM2_FPM, /**< FM FPM */
-+ e_MODULE_ID_FM2_IRAM, /**< FM Instruction-RAM */
-+ e_MODULE_ID_FM2_1GMDIO0, /**< FM 1G MDIO MAC 0*/
-+ e_MODULE_ID_FM2_1GMDIO1, /**< FM 1G MDIO MAC 1*/
-+ e_MODULE_ID_FM2_1GMDIO2, /**< FM 1G MDIO MAC 2*/
-+ e_MODULE_ID_FM2_1GMDIO3, /**< FM 1G MDIO MAC 3*/
-+ e_MODULE_ID_FM2_10GMDIO, /**< FM 10G MDIO */
-+ e_MODULE_ID_FM2_PRS_IRAM, /**< FM SW-parser Instruction-RAM */
-+ e_MODULE_ID_FM2_1GMAC0, /**< FM 1G MAC #0 */
-+ e_MODULE_ID_FM2_1GMAC1, /**< FM 1G MAC #1 */
-+ e_MODULE_ID_FM2_1GMAC2, /**< FM 1G MAC #2 */
-+ e_MODULE_ID_FM2_1GMAC3, /**< FM 1G MAC #3 */
-+ e_MODULE_ID_FM2_10GMAC0, /**< FM 10G MAC #0 */
-+
-+ e_MODULE_ID_SEC_GEN, /**< SEC 4.0 General registers */
-+ e_MODULE_ID_SEC_QI, /**< SEC 4.0 QI registers */
-+ e_MODULE_ID_SEC_JQ0, /**< SEC 4.0 JQ-0 registers */
-+ e_MODULE_ID_SEC_JQ1, /**< SEC 4.0 JQ-1 registers */
-+ e_MODULE_ID_SEC_JQ2, /**< SEC 4.0 JQ-2 registers */
-+ e_MODULE_ID_SEC_JQ3, /**< SEC 4.0 JQ-3 registers */
-+ e_MODULE_ID_SEC_RTIC, /**< SEC 4.0 RTIC registers */
-+ e_MODULE_ID_SEC_DECO0_CCB0, /**< SEC 4.0 DECO-0/CCB-0 registers */
-+ e_MODULE_ID_SEC_DECO1_CCB1, /**< SEC 4.0 DECO-1/CCB-1 registers */
-+ e_MODULE_ID_SEC_DECO2_CCB2, /**< SEC 4.0 DECO-2/CCB-2 registers */
-+ e_MODULE_ID_SEC_DECO3_CCB3, /**< SEC 4.0 DECO-3/CCB-3 registers */
-+ e_MODULE_ID_SEC_DECO4_CCB4, /**< SEC 4.0 DECO-4/CCB-4 registers */
-+
-+ e_MODULE_ID_MPIC, /**< MPIC */
-+ e_MODULE_ID_GPIO, /**< GPIO */
-+ e_MODULE_ID_SERDES, /**< SERDES */
-+ e_MODULE_ID_CPC_1, /**< CoreNet-Platform-Cache 1 */
-+ e_MODULE_ID_CPC_2, /**< CoreNet-Platform-Cache 2 */
-+
-+ e_MODULE_ID_SRIO_PORTS, /**< RapidIO controller */
-+ e_MODULE_ID_SRIO_MU, /**< RapidIO messaging unit module */
-+
-+ e_MODULE_ID_DUMMY_LAST
-+} e_ModuleId;
-+
-+#define NUM_OF_MODULES e_MODULE_ID_DUMMY_LAST
-+
-+#if 0 /* using unified values */
-+/*****************************************************************************
-+ INTEGRATION-SPECIFIC MODULE CODES
-+******************************************************************************/
-+#define MODULE_UNKNOWN 0x00000000
-+#define MODULE_MEM 0x00010000
-+#define MODULE_MM 0x00020000
-+#define MODULE_CORE 0x00030000
-+#define MODULE_CHIP 0x00040000
-+#define MODULE_PLTFRM 0x00050000
-+#define MODULE_PM 0x00060000
-+#define MODULE_MMU 0x00070000
-+#define MODULE_PIC 0x00080000
-+#define MODULE_CPC 0x00090000
-+#define MODULE_DUART 0x000a0000
-+#define MODULE_SERDES 0x000b0000
-+#define MODULE_PIO 0x000c0000
-+#define MODULE_QM 0x000d0000
-+#define MODULE_BM 0x000e0000
-+#define MODULE_SEC 0x000f0000
-+#define MODULE_LAW 0x00100000
-+#define MODULE_LBC 0x00110000
-+#define MODULE_PAMU 0x00120000
-+#define MODULE_FM 0x00130000
-+#define MODULE_FM_MURAM 0x00140000
-+#define MODULE_FM_PCD 0x00150000
-+#define MODULE_FM_RTC 0x00160000
-+#define MODULE_FM_MAC 0x00170000
-+#define MODULE_FM_PORT 0x00180000
-+#define MODULE_FM_SP 0x00190000
-+#define MODULE_DPA_PORT 0x001a0000
-+#define MODULE_MII 0x001b0000
-+#define MODULE_I2C 0x001c0000
-+#define MODULE_DMA 0x001d0000
-+#define MODULE_DDR 0x001e0000
-+#define MODULE_ESPI 0x001f0000
-+#define MODULE_DPAA_IPSEC 0x00200000
-+#endif /* using unified values */
-+
-+/*****************************************************************************
-+ PAMU INTEGRATION-SPECIFIC DEFINITIONS
-+******************************************************************************/
-+#define PAMU_NUM_OF_PARTITIONS 5
-+
-+#define PAMU_PICS_AVICS_ERRATA_PAMU3
-+
-+/*****************************************************************************
-+ LAW INTEGRATION-SPECIFIC DEFINITIONS
-+******************************************************************************/
-+#define LAW_NUM_OF_WINDOWS 32
-+#define LAW_MIN_WINDOW_SIZE 0x0000000000001000LL /**< 4KB */
-+#define LAW_MAX_WINDOW_SIZE 0x0000002000000000LL /**< 64GB */
-+
-+
-+/*****************************************************************************
-+ LBC INTEGRATION-SPECIFIC DEFINITIONS
-+******************************************************************************/
-+/**************************************************************************//**
-+ @Group lbc_exception_grp LBC Exception Unit
-+
-+ @Description LBC Exception unit API functions, definitions and enums
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Anchor lbc_exbm
-+
-+ @Collection LBC Errors Bit Mask
-+
-+ These errors are reported through the exceptions callback..
-+ The values can be or'ed in any combination in the errors mask
-+ parameter of the errors report structure.
-+
-+ These errors can also be passed as a bit-mask to
-+ LBC_EnableErrorChecking() or LBC_DisableErrorChecking(),
-+ for enabling or disabling error checking.
-+ @{
-+*//***************************************************************************/
-+#define LBC_ERR_BUS_MONITOR 0x80000000 /**< Bus monitor error */
-+#define LBC_ERR_PARITY_ECC 0x20000000 /**< Parity error for GPCM/UPM */
-+#define LBC_ERR_WRITE_PROTECT 0x04000000 /**< Write protection error */
-+#define LBC_ERR_ATOMIC_WRITE 0x00800000 /**< Atomic write error */
-+#define LBC_ERR_ATOMIC_READ 0x00400000 /**< Atomic read error */
-+#define LBC_ERR_CHIP_SELECT 0x00080000 /**< Unrecognized chip select */
-+
-+#define LBC_ERR_ALL (LBC_ERR_BUS_MONITOR | LBC_ERR_PARITY_ECC | \
-+ LBC_ERR_WRITE_PROTECT | LBC_ERR_ATOMIC_WRITE | \
-+ LBC_ERR_ATOMIC_READ | LBC_ERR_CHIP_SELECT)
-+ /**< All possible errors */
-+/* @} */
-+/** @} */ /* end of lbc_exception_grp group */
-+
-+#define LBC_INCORRECT_ERROR_REPORT_ERRATA
-+
-+#define LBC_NUM_OF_BANKS 8
-+#define LBC_MAX_CS_SIZE 0x0000000100000000LL
-+#define LBC_ATOMIC_OPERATION_SUPPORT
-+#define LBC_PARITY_SUPPORT
-+#define LBC_ADDRESS_HOLD_TIME_CTRL
-+#define LBC_HIGH_CLK_DIVIDERS
-+#define LBC_FCM_AVAILABLE
-+
-+/*****************************************************************************
-+ GPIO INTEGRATION-SPECIFIC DEFINITIONS
-+******************************************************************************/
-+#define GPIO_NUM_OF_PORTS 1 /**< Number of ports in GPIO module;
-+ Each port contains up to 32 i/O pins. */
-+
-+#define GPIO_VALID_PIN_MASKS \
-+ { /* Port A */ 0xFFFFFFFF }
-+
-+#define GPIO_VALID_INTR_MASKS \
-+ { /* Port A */ 0xFFFFFFFF }
-+
-+#endif /* __PART_INTEGRATION_EXT_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/math_ext.h
-@@ -0,0 +1,100 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+#ifndef __MATH_EXT_H
-+#define __MATH_EXT_H
-+
-+
-+#if defined(NCSW_LINUX) && defined(__KERNEL__)
-+#include <linux/math.h>
-+#include <linux/math64.h>
-+
-+#elif defined(__MWERKS__)
-+#define LOW(x) ( sizeof(x)==8 ? *(1+(int32_t*)&x) : (*(int32_t*)&x))
-+#define HIGH(x) (*(int32_t*)&x)
-+#define ULOW(x) ( sizeof(x)==8 ? *(1+(uint32_t*)&x) : (*(uint32_t*)&x))
-+#define UHIGH(x) (*(uint32_t*)&x)
-+
-+static const double big = 1.0e300;
-+
-+/* Macro for checking if a number is a power of 2 */
-+static __inline__ double ceil(double x)
-+{
-+ int32_t i0,i1,j0; /*- cc 020130 -*/
-+ uint32_t i,j; /*- cc 020130 -*/
-+ i0 = HIGH(x);
-+ i1 = LOW(x);
-+ j0 = ((i0>>20)&0x7ff)-0x3ff;
-+ if(j0<20) {
-+ if(j0<0) { /* raise inexact if x != 0 */
-+ if(big+x>0.0) {/* return 0*sign(x) if |x|<1 */
-+ if(i0<0) {i0=0x80000000;i1=0;}
-+ else if((i0|i1)!=0) { i0=0x3ff00000;i1=0;}
-+ }
-+ } else {
-+ i = (uint32_t)(0x000fffff)>>j0;
-+ if(((i0&i)|i1)==0) return x; /* x is integral */
-+ if(big+x>0.0) { /* raise inexact flag */
-+ if(i0>0) i0 += (0x00100000)>>j0;
-+ i0 &= (~i); i1=0;
-+ }
-+ }
-+ } else if (j0>51) {
-+ if(j0==0x400) return x+x; /* inf or NaN */
-+ else return x; /* x is integral */
-+ } else {
-+ i = ((uint32_t)(0xffffffff))>>(j0-20); /*- cc 020130 -*/
-+ if((i1&i)==0) return x; /* x is integral */
-+ if(big+x>0.0) { /* raise inexact flag */
-+ if(i0>0) {
-+ if(j0==20) i0+=1;
-+ else {
-+ j = (uint32_t)(i1 + (1<<(52-j0)));
-+ if(j<i1) i0+=1; /* got a carry */
-+ i1 = (int32_t)j;
-+ }
-+ }
-+ i1 &= (~i);
-+ }
-+ }
-+ HIGH(x) = i0;
-+ LOW(x) = i1;
-+ return x;
-+}
-+
-+#else
-+#include <math.h>
-+#endif /* defined(NCSW_LINUX) && defined(__KERNEL__) */
-+
-+
-+#endif /* __MATH_EXT_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/ncsw_ext.h
-@@ -0,0 +1,435 @@
-+/* Copyright (c) 2008-2012 Freescale Semiconductor, Inc
-+ * All rights reserved.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/**************************************************************************//**
-+ @File ncsw_ext.h
-+
-+ @Description General NetCommSw Standard Definitions
-+*//***************************************************************************/
-+
-+#ifndef __NCSW_EXT_H
-+#define __NCSW_EXT_H
-+
-+
-+#include "memcpy_ext.h"
-+
-+#define WRITE_BLOCK IOMemSet32 /* include memcpy_ext.h */
-+#define COPY_BLOCK Mem2IOCpy32 /* include memcpy_ext.h */
-+
-+#define PTR_TO_UINT(_ptr) ((uintptr_t)(_ptr))
-+#define UINT_TO_PTR(_val) ((void*)(uintptr_t)(_val))
-+
-+#define PTR_MOVE(_ptr, _offset) (void*)((uint8_t*)(_ptr) + (_offset))
-+
-+
-+#define WRITE_UINT8_UINT24(arg, data08, data24) \
-+ WRITE_UINT32(arg,((uint32_t)(data08)<<24)|((uint32_t)(data24)&0x00FFFFFF))
-+#define WRITE_UINT24_UINT8(arg, data24, data08) \
-+ WRITE_UINT32(arg,((uint32_t)(data24)<< 8)|((uint32_t)(data08)&0x000000FF))
-+
-+/* Little-Endian access macros */
-+
-+#define WRITE_UINT16_LE(arg, data) \
-+ WRITE_UINT16((arg), SwapUint16(data))
-+
-+#define WRITE_UINT32_LE(arg, data) \
-+ WRITE_UINT32((arg), SwapUint32(data))
-+
-+#define WRITE_UINT64_LE(arg, data) \
-+ WRITE_UINT64((arg), SwapUint64(data))
-+
-+#define GET_UINT16_LE(arg) \
-+ SwapUint16(GET_UINT16(arg))
-+
-+#define GET_UINT32_LE(arg) \
-+ SwapUint32(GET_UINT32(arg))
-+
-+#define GET_UINT64_LE(arg) \
-+ SwapUint64(GET_UINT64(arg))
-+
-+/* Write and Read again macros */
-+#define WRITE_UINT_SYNC(size, arg, data) \
-+ do { \
-+ WRITE_UINT##size((arg), (data)); \
-+ CORE_MemoryBarrier(); \
-+ } while (0)
-+
-+#define WRITE_UINT8_SYNC(arg, data) WRITE_UINT_SYNC(8, (arg), (data))
-+
-+#define WRITE_UINT16_SYNC(arg, data) WRITE_UINT_SYNC(16, (arg), (data))
-+#define WRITE_UINT32_SYNC(arg, data) WRITE_UINT_SYNC(32, (arg), (data))
-+
-+#define MAKE_UINT64(high32, low32) (((uint64_t)high32 << 32) | (low32))
-+
-+
-+/*----------------------*/
-+/* Miscellaneous macros */
-+/*----------------------*/
-+
-+#define UNUSED(_x) ((void)(_x))
-+
-+#define KILOBYTE 0x400UL /* 1024 */
-+#define MEGABYTE (KILOBYTE * KILOBYTE) /* 1024*1024 */
-+#define GIGABYTE ((uint64_t)(KILOBYTE * MEGABYTE)) /* 1024*1024*1024 */
-+#define TERABYTE ((uint64_t)(KILOBYTE * GIGABYTE)) /* 1024*1024*1024*1024 */
-+
-+#ifndef NO_IRQ
-+#define NO_IRQ (0)
-+#endif
-+#define NCSW_MASTER_ID (0)
-+
-+/* Macro for checking if a number is a power of 2 */
-+#define POWER_OF_2(n) (!((n) & ((n)-1)))
-+
-+/* Macro for calculating log of base 2 */
-+#define LOG2(num, log2Num) \
-+ do \
-+ { \
-+ uint64_t tmp = (num); \
-+ log2Num = 0; \
-+ while (tmp > 1) \
-+ { \
-+ log2Num++; \
-+ tmp >>= 1; \
-+ } \
-+ } while (0)
-+
-+#define NEXT_POWER_OF_2(_num, _nextPow) \
-+do \
-+{ \
-+ if (POWER_OF_2(_num)) \
-+ _nextPow = (_num); \
-+ else \
-+ { \
-+ uint64_t tmp = (_num); \
-+ _nextPow = 1; \
-+ while (tmp) \
-+ { \
-+ _nextPow <<= 1; \
-+ tmp >>= 1; \
-+ } \
-+ } \
-+} while (0)
-+
-+/* Ceiling division - not the fastest way, but safer in terms of overflow */
-+#define DIV_CEIL(x,y) (div64_u64((x),(y)) + (((div64_u64((x),(y))*(y)) == (x)) ? 0 : 1))
-+
-+/* Round up a number to be a multiple of a second number */
-+#define ROUND_UP(x,y) ((((x) + (y) - 1) / (y)) * (y))
-+
-+/* Timing macro for converting usec units to number of ticks. */
-+/* (number of usec * clock_Hz) / 1,000,000) - since */
-+/* clk is in MHz units, no division needed. */
-+#define USEC_TO_CLK(usec,clk) ((usec) * (clk))
-+#define CYCLES_TO_USEC(cycles,clk) ((cycles) / (clk))
-+
-+/* Timing macros for converting between nsec units and number of clocks. */
-+#define NSEC_TO_CLK(nsec,clk) DIV_CEIL(((nsec) * (clk)), 1000)
-+#define CYCLES_TO_NSEC(cycles,clk) (((cycles) * 1000) / (clk))
-+
-+/* Timing macros for converting between psec units and number of clocks. */
-+#define PSEC_TO_CLK(psec,clk) DIV_CEIL(((psec) * (clk)), 1000000)
-+#define CYCLES_TO_PSEC(cycles,clk) (((cycles) * 1000000) / (clk))
-+
-+/* Min, Max macros */
-+#define MIN(a,b) ((a) < (b) ? (a) : (b))
-+#define MAX(a,b) ((a) > (b) ? (a) : (b))
-+#define IN_RANGE(min,val,max) ((min)<=(val) && (val)<=(max))
-+
-+#define ABS(a) ((a<0)?(a*-1):a)
-+
-+#if !(defined(ARRAY_SIZE))
-+#define ARRAY_SIZE(arr) (sizeof(arr) / sizeof((arr)[0]))
-+#endif /* !defined(ARRAY_SIZE) */
-+
-+
-+/* possible alignments */
-+#define HALF_WORD_ALIGNMENT 2
-+#define WORD_ALIGNMENT 4
-+#define DOUBLE_WORD_ALIGNMENT 8
-+#define BURST_ALIGNMENT 32
-+
-+#define HALF_WORD_ALIGNED 0x00000001
-+#define WORD_ALIGNED 0x00000003
-+#define DOUBLE_WORD_ALIGNED 0x00000007
-+#define BURST_ALIGNED 0x0000001f
-+#ifndef IS_ALIGNED
-+#define IS_ALIGNED(n,align) (!((uint32_t)(n) & (align - 1)))
-+#endif /* IS_ALIGNED */
-+
-+
-+#define LAST_BUF 1
-+#define FIRST_BUF 2
-+#define SINGLE_BUF (LAST_BUF | FIRST_BUF)
-+#define MIDDLE_BUF 4
-+
-+#define ARRAY_END -1
-+
-+#define ILLEGAL_BASE (~0)
-+
-+#define BUF_POSITION(first, last) state[(!!(last))<<1 | !!(first)]
-+#define DECLARE_POSITION static uint8_t state[4] = { (uint8_t)MIDDLE_BUF, (uint8_t)FIRST_BUF, (uint8_t)LAST_BUF, (uint8_t)SINGLE_BUF };
-+
-+
-+/**************************************************************************//**
-+ @Description Timers operation mode
-+*//***************************************************************************/
-+typedef enum e_TimerMode
-+{
-+ e_TIMER_MODE_INVALID = 0,
-+ e_TIMER_MODE_FREE_RUN, /**< Free run - counter continues to increase
-+ after reaching the reference value. */
-+ e_TIMER_MODE_PERIODIC, /**< Periodic - counter restarts counting from 0
-+ after reaching the reference value. */
-+ e_TIMER_MODE_SINGLE /**< Single (one-shot) - counter stops counting
-+ after reaching the reference value. */
-+} e_TimerMode;
-+
-+
-+/**************************************************************************//**
-+ @Description Enumeration (bit flags) of communication modes (Transmit,
-+ receive or both).
-+*//***************************************************************************/
-+typedef enum e_CommMode
-+{
-+ e_COMM_MODE_NONE = 0, /**< No transmit/receive communication */
-+ e_COMM_MODE_RX = 1, /**< Only receive communication */
-+ e_COMM_MODE_TX = 2, /**< Only transmit communication */
-+ e_COMM_MODE_RX_AND_TX = 3 /**< Both transmit and receive communication */
-+} e_CommMode;
-+
-+/**************************************************************************//**
-+ @Description General Diagnostic Mode
-+*//***************************************************************************/
-+typedef enum e_DiagMode
-+{
-+ e_DIAG_MODE_NONE = 0, /**< Normal operation; no diagnostic mode */
-+ e_DIAG_MODE_CTRL_LOOPBACK, /**< Loopback in the controller */
-+ e_DIAG_MODE_CHIP_LOOPBACK, /**< Loopback in the chip but not in the
-+ controller; e.g. IO-pins, SerDes, etc. */
-+ e_DIAG_MODE_PHY_LOOPBACK, /**< Loopback in the external PHY */
-+ e_DIAG_MODE_EXT_LOOPBACK, /**< Loopback in the external line (beyond the PHY) */
-+ e_DIAG_MODE_CTRL_ECHO, /**< Echo incoming data by the controller */
-+ e_DIAG_MODE_PHY_ECHO /**< Echo incoming data by the PHY */
-+} e_DiagMode;
-+
-+/**************************************************************************//**
-+ @Description Possible RxStore callback responses.
-+*//***************************************************************************/
-+typedef enum e_RxStoreResponse
-+{
-+ e_RX_STORE_RESPONSE_PAUSE /**< Pause invoking callback with received data;
-+ in polling mode, start again invoking callback
-+ only next time user invokes the receive routine;
-+ in interrupt mode, start again invoking callback
-+ only next time a receive event triggers an interrupt;
-+ in all cases, received data that are pending are not
-+ lost, rather, their processing is temporarily deferred;
-+ in all cases, received data are processed in the order
-+ in which they were received. */
-+ , e_RX_STORE_RESPONSE_CONTINUE /**< Continue invoking callback with received data. */
-+} e_RxStoreResponse;
-+
-+
-+/**************************************************************************//**
-+ @Description General Handle
-+*//***************************************************************************/
-+typedef void * t_Handle; /**< handle, used as object's descriptor */
-+
-+/**************************************************************************//**
-+ @Description MUTEX type
-+*//***************************************************************************/
-+typedef uint32_t t_Mutex;
-+
-+/**************************************************************************//**
-+ @Description Error Code.
-+
-+ The high word of the error code is the code of the software
-+ module (driver). The low word is the error type (e_ErrorType).
-+ To get the values from the error code, use GET_ERROR_TYPE()
-+ and GET_ERROR_MODULE().
-+*//***************************************************************************/
-+typedef uint32_t t_Error;
-+
-+/**************************************************************************//**
-+ @Description General prototype of interrupt service routine (ISR).
-+
-+ @Param[in] handle - Optional handle of the module handling the interrupt.
-+
-+ @Return None
-+ *//***************************************************************************/
-+typedef void (t_Isr)(t_Handle handle);
-+
-+/**************************************************************************//**
-+ @Anchor mem_attr
-+
-+ @Collection Memory Attributes
-+
-+ Various attributes of memory partitions. These values may be
-+ or'ed together to create a mask of all memory attributes.
-+ @{
-+*//***************************************************************************/
-+#define MEMORY_ATTR_CACHEABLE 0x00000001
-+ /**< Memory is cacheable */
-+#define MEMORY_ATTR_QE_2ND_BUS_ACCESS 0x00000002
-+ /**< Memory can be accessed by QUICC Engine
-+ through its secondary bus interface */
-+
-+/* @} */
-+
-+
-+/**************************************************************************//**
-+ @Function t_GetBufFunction
-+
-+ @Description User callback function called by driver to get data buffer.
-+
-+ User provides this function. Driver invokes it.
-+
-+ @Param[in] h_BufferPool - A handle to buffer pool manager
-+ @Param[out] p_BufContextHandle - Returns the user's private context that
-+ should be associated with the buffer
-+
-+ @Return Pointer to data buffer, NULL if error
-+ *//***************************************************************************/
-+typedef uint8_t * (t_GetBufFunction)(t_Handle h_BufferPool,
-+ t_Handle *p_BufContextHandle);
-+
-+/**************************************************************************//**
-+ @Function t_PutBufFunction
-+
-+ @Description User callback function called by driver to return data buffer.
-+
-+ User provides this function. Driver invokes it.
-+
-+ @Param[in] h_BufferPool - A handle to buffer pool manager
-+ @Param[in] p_Buffer - A pointer to buffer to return
-+ @Param[in] h_BufContext - The user's private context associated with
-+ the returned buffer
-+
-+ @Return E_OK on success; Error code otherwise
-+ *//***************************************************************************/
-+typedef t_Error (t_PutBufFunction)(t_Handle h_BufferPool,
-+ uint8_t *p_Buffer,
-+ t_Handle h_BufContext);
-+
-+/**************************************************************************//**
-+ @Function t_PhysToVirt
-+
-+ @Description Translates a physical address to the matching virtual address.
-+
-+ @Param[in] addr - The physical address to translate.
-+
-+ @Return Virtual address.
-+*//***************************************************************************/
-+typedef void * t_PhysToVirt(physAddress_t addr);
-+
-+/**************************************************************************//**
-+ @Function t_VirtToPhys
-+
-+ @Description Translates a virtual address to the matching physical address.
-+
-+ @Param[in] addr - The virtual address to translate.
-+
-+ @Return Physical address.
-+*//***************************************************************************/
-+typedef physAddress_t t_VirtToPhys(void *addr);
-+
-+/**************************************************************************//**
-+ @Description Buffer Pool Information Structure.
-+*//***************************************************************************/
-+typedef struct t_BufferPoolInfo
-+{
-+ t_Handle h_BufferPool; /**< A handle to the buffer pool manager */
-+ t_GetBufFunction *f_GetBuf; /**< User callback to get a free buffer */
-+ t_PutBufFunction *f_PutBuf; /**< User callback to return a buffer */
-+ uint16_t bufferSize; /**< Buffer size (in bytes) */
-+
-+ t_PhysToVirt *f_PhysToVirt; /**< User callback to translate pool buffers
-+ physical addresses to virtual addresses */
-+ t_VirtToPhys *f_VirtToPhys; /**< User callback to translate pool buffers
-+ virtual addresses to physical addresses */
-+} t_BufferPoolInfo;
-+
-+
-+/**************************************************************************//**
-+ @Description User callback function called by driver when transmit completed.
-+
-+ User provides this function. Driver invokes it.
-+
-+ @Param[in] h_App - Application's handle, as was provided to the
-+ driver by the user
-+ @Param[in] queueId - Transmit queue ID
-+ @Param[in] p_Data - Pointer to the data buffer
-+ @Param[in] h_BufContext - The user's private context associated with
-+ the given data buffer
-+ @Param[in] status - Transmit status and errors
-+ @Param[in] flags - Driver-dependent information
-+ *//***************************************************************************/
-+typedef void (t_TxConfFunction)(t_Handle h_App,
-+ uint32_t queueId,
-+ uint8_t *p_Data,
-+ t_Handle h_BufContext,
-+ uint16_t status,
-+ uint32_t flags);
-+
-+/**************************************************************************//**
-+ @Description User callback function called by driver with receive data.
-+
-+ User provides this function. Driver invokes it.
-+
-+ @Param[in] h_App - Application's handle, as was provided to the
-+ driver by the user
-+ @Param[in] queueId - Receive queue ID
-+ @Param[in] p_Data - Pointer to the buffer with received data
-+ @Param[in] h_BufContext - The user's private context associated with
-+ the given data buffer
-+ @Param[in] length - Length of received data
-+ @Param[in] status - Receive status and errors
-+ @Param[in] position - Position of buffer in frame
-+ @Param[in] flags - Driver-dependent information
-+
-+ @Retval e_RX_STORE_RESPONSE_CONTINUE - order the driver to continue Rx
-+ operation for all ready data.
-+ @Retval e_RX_STORE_RESPONSE_PAUSE - order the driver to stop Rx operation.
-+ *//***************************************************************************/
-+typedef e_RxStoreResponse (t_RxStoreFunction)(t_Handle h_App,
-+ uint32_t queueId,
-+ uint8_t *p_Data,
-+ t_Handle h_BufContext,
-+ uint32_t length,
-+ uint16_t status,
-+ uint8_t position,
-+ uint32_t flags);
-+
-+
-+#endif /* __NCSW_EXT_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/net_ext.h
-@@ -0,0 +1,430 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/**************************************************************************//**
-+ @File net_ext.h
-+
-+ @Description This file contains common and general netcomm headers definitions.
-+*//***************************************************************************/
-+#ifndef __NET_EXT_H
-+#define __NET_EXT_H
-+
-+#include "std_ext.h"
-+
-+
-+typedef uint8_t headerFieldPpp_t;
-+
-+#define NET_HEADER_FIELD_PPP_PID (1)
-+#define NET_HEADER_FIELD_PPP_COMPRESSED (NET_HEADER_FIELD_PPP_PID << 1)
-+#define NET_HEADER_FIELD_PPP_ALL_FIELDS ((NET_HEADER_FIELD_PPP_PID << 2) - 1)
-+
-+
-+typedef uint8_t headerFieldPppoe_t;
-+
-+#define NET_HEADER_FIELD_PPPoE_VER (1)
-+#define NET_HEADER_FIELD_PPPoE_TYPE (NET_HEADER_FIELD_PPPoE_VER << 1)
-+#define NET_HEADER_FIELD_PPPoE_CODE (NET_HEADER_FIELD_PPPoE_VER << 2)
-+#define NET_HEADER_FIELD_PPPoE_SID (NET_HEADER_FIELD_PPPoE_VER << 3)
-+#define NET_HEADER_FIELD_PPPoE_LEN (NET_HEADER_FIELD_PPPoE_VER << 4)
-+#define NET_HEADER_FIELD_PPPoE_SESSION (NET_HEADER_FIELD_PPPoE_VER << 5)
-+#define NET_HEADER_FIELD_PPPoE_PID (NET_HEADER_FIELD_PPPoE_VER << 6)
-+#define NET_HEADER_FIELD_PPPoE_ALL_FIELDS ((NET_HEADER_FIELD_PPPoE_VER << 7) - 1)
-+
-+#define NET_HEADER_FIELD_PPPMUX_PID (1)
-+#define NET_HEADER_FIELD_PPPMUX_CKSUM (NET_HEADER_FIELD_PPPMUX_PID << 1)
-+#define NET_HEADER_FIELD_PPPMUX_COMPRESSED (NET_HEADER_FIELD_PPPMUX_PID << 2)
-+#define NET_HEADER_FIELD_PPPMUX_ALL_FIELDS ((NET_HEADER_FIELD_PPPMUX_PID << 3) - 1)
-+
-+#define NET_HEADER_FIELD_PPPMUX_SUBFRAME_PFF (1)
-+#define NET_HEADER_FIELD_PPPMUX_SUBFRAME_LXT (NET_HEADER_FIELD_PPPMUX_SUBFRAME_PFF << 1)
-+#define NET_HEADER_FIELD_PPPMUX_SUBFRAME_LEN (NET_HEADER_FIELD_PPPMUX_SUBFRAME_PFF << 2)
-+#define NET_HEADER_FIELD_PPPMUX_SUBFRAME_PID (NET_HEADER_FIELD_PPPMUX_SUBFRAME_PFF << 3)
-+#define NET_HEADER_FIELD_PPPMUX_SUBFRAME_USE_PID (NET_HEADER_FIELD_PPPMUX_SUBFRAME_PFF << 4)
-+#define NET_HEADER_FIELD_PPPMUX_SUBFRAME_ALL_FIELDS ((NET_HEADER_FIELD_PPPMUX_SUBFRAME_PFF << 5) - 1)
-+
-+
-+typedef uint8_t headerFieldEth_t;
-+
-+#define NET_HEADER_FIELD_ETH_DA (1)
-+#define NET_HEADER_FIELD_ETH_SA (NET_HEADER_FIELD_ETH_DA << 1)
-+#define NET_HEADER_FIELD_ETH_LENGTH (NET_HEADER_FIELD_ETH_DA << 2)
-+#define NET_HEADER_FIELD_ETH_TYPE (NET_HEADER_FIELD_ETH_DA << 3)
-+#define NET_HEADER_FIELD_ETH_FINAL_CKSUM (NET_HEADER_FIELD_ETH_DA << 4)
-+#define NET_HEADER_FIELD_ETH_PADDING (NET_HEADER_FIELD_ETH_DA << 5)
-+#define NET_HEADER_FIELD_ETH_ALL_FIELDS ((NET_HEADER_FIELD_ETH_DA << 6) - 1)
-+
-+#define NET_HEADER_FIELD_ETH_ADDR_SIZE 6
-+
-+typedef uint16_t headerFieldIp_t;
-+
-+#define NET_HEADER_FIELD_IP_VER (1)
-+#define NET_HEADER_FIELD_IP_DSCP (NET_HEADER_FIELD_IP_VER << 2)
-+#define NET_HEADER_FIELD_IP_ECN (NET_HEADER_FIELD_IP_VER << 3)
-+#define NET_HEADER_FIELD_IP_PROTO (NET_HEADER_FIELD_IP_VER << 4)
-+
-+#define NET_HEADER_FIELD_IP_PROTO_SIZE 1
-+
-+typedef uint16_t headerFieldIpv4_t;
-+
-+#define NET_HEADER_FIELD_IPv4_VER (1)
-+#define NET_HEADER_FIELD_IPv4_HDR_LEN (NET_HEADER_FIELD_IPv4_VER << 1)
-+#define NET_HEADER_FIELD_IPv4_TOS (NET_HEADER_FIELD_IPv4_VER << 2)
-+#define NET_HEADER_FIELD_IPv4_TOTAL_LEN (NET_HEADER_FIELD_IPv4_VER << 3)
-+#define NET_HEADER_FIELD_IPv4_ID (NET_HEADER_FIELD_IPv4_VER << 4)
-+#define NET_HEADER_FIELD_IPv4_FLAG_D (NET_HEADER_FIELD_IPv4_VER << 5)
-+#define NET_HEADER_FIELD_IPv4_FLAG_M (NET_HEADER_FIELD_IPv4_VER << 6)
-+#define NET_HEADER_FIELD_IPv4_OFFSET (NET_HEADER_FIELD_IPv4_VER << 7)
-+#define NET_HEADER_FIELD_IPv4_TTL (NET_HEADER_FIELD_IPv4_VER << 8)
-+#define NET_HEADER_FIELD_IPv4_PROTO (NET_HEADER_FIELD_IPv4_VER << 9)
-+#define NET_HEADER_FIELD_IPv4_CKSUM (NET_HEADER_FIELD_IPv4_VER << 10)
-+#define NET_HEADER_FIELD_IPv4_SRC_IP (NET_HEADER_FIELD_IPv4_VER << 11)
-+#define NET_HEADER_FIELD_IPv4_DST_IP (NET_HEADER_FIELD_IPv4_VER << 12)
-+#define NET_HEADER_FIELD_IPv4_OPTS (NET_HEADER_FIELD_IPv4_VER << 13)
-+#define NET_HEADER_FIELD_IPv4_OPTS_COUNT (NET_HEADER_FIELD_IPv4_VER << 14)
-+#define NET_HEADER_FIELD_IPv4_ALL_FIELDS ((NET_HEADER_FIELD_IPv4_VER << 15) - 1)
-+
-+#define NET_HEADER_FIELD_IPv4_ADDR_SIZE 4
-+#define NET_HEADER_FIELD_IPv4_PROTO_SIZE 1
-+
-+
-+typedef uint8_t headerFieldIpv6_t;
-+
-+#define NET_HEADER_FIELD_IPv6_VER (1)
-+#define NET_HEADER_FIELD_IPv6_TC (NET_HEADER_FIELD_IPv6_VER << 1)
-+#define NET_HEADER_FIELD_IPv6_SRC_IP (NET_HEADER_FIELD_IPv6_VER << 2)
-+#define NET_HEADER_FIELD_IPv6_DST_IP (NET_HEADER_FIELD_IPv6_VER << 3)
-+#define NET_HEADER_FIELD_IPv6_NEXT_HDR (NET_HEADER_FIELD_IPv6_VER << 4)
-+#define NET_HEADER_FIELD_IPv6_FL (NET_HEADER_FIELD_IPv6_VER << 5)
-+#define NET_HEADER_FIELD_IPv6_HOP_LIMIT (NET_HEADER_FIELD_IPv6_VER << 6)
-+#define NET_HEADER_FIELD_IPv6_ALL_FIELDS ((NET_HEADER_FIELD_IPv6_VER << 7) - 1)
-+
-+#define NET_HEADER_FIELD_IPv6_ADDR_SIZE 16
-+#define NET_HEADER_FIELD_IPv6_NEXT_HDR_SIZE 1
-+
-+#define NET_HEADER_FIELD_ICMP_TYPE (1)
-+#define NET_HEADER_FIELD_ICMP_CODE (NET_HEADER_FIELD_ICMP_TYPE << 1)
-+#define NET_HEADER_FIELD_ICMP_CKSUM (NET_HEADER_FIELD_ICMP_TYPE << 2)
-+#define NET_HEADER_FIELD_ICMP_ID (NET_HEADER_FIELD_ICMP_TYPE << 3)
-+#define NET_HEADER_FIELD_ICMP_SQ_NUM (NET_HEADER_FIELD_ICMP_TYPE << 4)
-+#define NET_HEADER_FIELD_ICMP_ALL_FIELDS ((NET_HEADER_FIELD_ICMP_TYPE << 5) - 1)
-+
-+#define NET_HEADER_FIELD_ICMP_CODE_SIZE 1
-+#define NET_HEADER_FIELD_ICMP_TYPE_SIZE 1
-+
-+#define NET_HEADER_FIELD_IGMP_VERSION (1)
-+#define NET_HEADER_FIELD_IGMP_TYPE (NET_HEADER_FIELD_IGMP_VERSION << 1)
-+#define NET_HEADER_FIELD_IGMP_CKSUM (NET_HEADER_FIELD_IGMP_VERSION << 2)
-+#define NET_HEADER_FIELD_IGMP_DATA (NET_HEADER_FIELD_IGMP_VERSION << 3)
-+#define NET_HEADER_FIELD_IGMP_ALL_FIELDS ((NET_HEADER_FIELD_IGMP_VERSION << 4) - 1)
-+
-+
-+typedef uint16_t headerFieldTcp_t;
-+
-+#define NET_HEADER_FIELD_TCP_PORT_SRC (1)
-+#define NET_HEADER_FIELD_TCP_PORT_DST (NET_HEADER_FIELD_TCP_PORT_SRC << 1)
-+#define NET_HEADER_FIELD_TCP_SEQ (NET_HEADER_FIELD_TCP_PORT_SRC << 2)
-+#define NET_HEADER_FIELD_TCP_ACK (NET_HEADER_FIELD_TCP_PORT_SRC << 3)
-+#define NET_HEADER_FIELD_TCP_OFFSET (NET_HEADER_FIELD_TCP_PORT_SRC << 4)
-+#define NET_HEADER_FIELD_TCP_FLAGS (NET_HEADER_FIELD_TCP_PORT_SRC << 5)
-+#define NET_HEADER_FIELD_TCP_WINDOW (NET_HEADER_FIELD_TCP_PORT_SRC << 6)
-+#define NET_HEADER_FIELD_TCP_CKSUM (NET_HEADER_FIELD_TCP_PORT_SRC << 7)
-+#define NET_HEADER_FIELD_TCP_URGPTR (NET_HEADER_FIELD_TCP_PORT_SRC << 8)
-+#define NET_HEADER_FIELD_TCP_OPTS (NET_HEADER_FIELD_TCP_PORT_SRC << 9)
-+#define NET_HEADER_FIELD_TCP_OPTS_COUNT (NET_HEADER_FIELD_TCP_PORT_SRC << 10)
-+#define NET_HEADER_FIELD_TCP_ALL_FIELDS ((NET_HEADER_FIELD_TCP_PORT_SRC << 11) - 1)
-+
-+#define NET_HEADER_FIELD_TCP_PORT_SIZE 2
-+
-+
-+typedef uint8_t headerFieldSctp_t;
-+
-+#define NET_HEADER_FIELD_SCTP_PORT_SRC (1)
-+#define NET_HEADER_FIELD_SCTP_PORT_DST (NET_HEADER_FIELD_SCTP_PORT_SRC << 1)
-+#define NET_HEADER_FIELD_SCTP_VER_TAG (NET_HEADER_FIELD_SCTP_PORT_SRC << 2)
-+#define NET_HEADER_FIELD_SCTP_CKSUM (NET_HEADER_FIELD_SCTP_PORT_SRC << 3)
-+#define NET_HEADER_FIELD_SCTP_ALL_FIELDS ((NET_HEADER_FIELD_SCTP_PORT_SRC << 4) - 1)
-+
-+#define NET_HEADER_FIELD_SCTP_PORT_SIZE 2
-+
-+typedef uint8_t headerFieldDccp_t;
-+
-+#define NET_HEADER_FIELD_DCCP_PORT_SRC (1)
-+#define NET_HEADER_FIELD_DCCP_PORT_DST (NET_HEADER_FIELD_DCCP_PORT_SRC << 1)
-+#define NET_HEADER_FIELD_DCCP_ALL_FIELDS ((NET_HEADER_FIELD_DCCP_PORT_SRC << 2) - 1)
-+
-+#define NET_HEADER_FIELD_DCCP_PORT_SIZE 2
-+
-+
-+typedef uint8_t headerFieldUdp_t;
-+
-+#define NET_HEADER_FIELD_UDP_PORT_SRC (1)
-+#define NET_HEADER_FIELD_UDP_PORT_DST (NET_HEADER_FIELD_UDP_PORT_SRC << 1)
-+#define NET_HEADER_FIELD_UDP_LEN (NET_HEADER_FIELD_UDP_PORT_SRC << 2)
-+#define NET_HEADER_FIELD_UDP_CKSUM (NET_HEADER_FIELD_UDP_PORT_SRC << 3)
-+#define NET_HEADER_FIELD_UDP_ALL_FIELDS ((NET_HEADER_FIELD_UDP_PORT_SRC << 4) - 1)
-+
-+#define NET_HEADER_FIELD_UDP_PORT_SIZE 2
-+
-+typedef uint8_t headerFieldUdpLite_t;
-+
-+#define NET_HEADER_FIELD_UDP_LITE_PORT_SRC (1)
-+#define NET_HEADER_FIELD_UDP_LITE_PORT_DST (NET_HEADER_FIELD_UDP_LITE_PORT_SRC << 1)
-+#define NET_HEADER_FIELD_UDP_LITE_ALL_FIELDS ((NET_HEADER_FIELD_UDP_LITE_PORT_SRC << 2) - 1)
-+
-+#define NET_HEADER_FIELD_UDP_LITE_PORT_SIZE 2
-+
-+typedef uint8_t headerFieldUdpEncapEsp_t;
-+
-+#define NET_HEADER_FIELD_UDP_ENCAP_ESP_PORT_SRC (1)
-+#define NET_HEADER_FIELD_UDP_ENCAP_ESP_PORT_DST (NET_HEADER_FIELD_UDP_ENCAP_ESP_PORT_SRC << 1)
-+#define NET_HEADER_FIELD_UDP_ENCAP_ESP_LEN (NET_HEADER_FIELD_UDP_ENCAP_ESP_PORT_SRC << 2)
-+#define NET_HEADER_FIELD_UDP_ENCAP_ESP_CKSUM (NET_HEADER_FIELD_UDP_ENCAP_ESP_PORT_SRC << 3)
-+#define NET_HEADER_FIELD_UDP_ENCAP_ESP_SPI (NET_HEADER_FIELD_UDP_ENCAP_ESP_PORT_SRC << 4)
-+#define NET_HEADER_FIELD_UDP_ENCAP_ESP_SEQUENCE_NUM (NET_HEADER_FIELD_UDP_ENCAP_ESP_PORT_SRC << 5)
-+#define NET_HEADER_FIELD_UDP_ENCAP_ESP_ALL_FIELDS ((NET_HEADER_FIELD_UDP_ENCAP_ESP_PORT_SRC << 6) - 1)
-+
-+#define NET_HEADER_FIELD_UDP_ENCAP_ESP_PORT_SIZE 2
-+#define NET_HEADER_FIELD_UDP_ENCAP_ESP_SPI_SIZE 4
-+
-+#define NET_HEADER_FIELD_IPHC_CID (1)
-+#define NET_HEADER_FIELD_IPHC_CID_TYPE (NET_HEADER_FIELD_IPHC_CID << 1)
-+#define NET_HEADER_FIELD_IPHC_HCINDEX (NET_HEADER_FIELD_IPHC_CID << 2)
-+#define NET_HEADER_FIELD_IPHC_GEN (NET_HEADER_FIELD_IPHC_CID << 3)
-+#define NET_HEADER_FIELD_IPHC_D_BIT (NET_HEADER_FIELD_IPHC_CID << 4)
-+#define NET_HEADER_FIELD_IPHC_ALL_FIELDS ((NET_HEADER_FIELD_IPHC_CID << 5) - 1)
-+
-+#define NET_HEADER_FIELD_SCTP_CHUNK_DATA_TYPE (1)
-+#define NET_HEADER_FIELD_SCTP_CHUNK_DATA_FLAGS (NET_HEADER_FIELD_SCTP_CHUNK_DATA_TYPE << 1)
-+#define NET_HEADER_FIELD_SCTP_CHUNK_DATA_LENGTH (NET_HEADER_FIELD_SCTP_CHUNK_DATA_TYPE << 2)
-+#define NET_HEADER_FIELD_SCTP_CHUNK_DATA_TSN (NET_HEADER_FIELD_SCTP_CHUNK_DATA_TYPE << 3)
-+#define NET_HEADER_FIELD_SCTP_CHUNK_DATA_STREAM_ID (NET_HEADER_FIELD_SCTP_CHUNK_DATA_TYPE << 4)
-+#define NET_HEADER_FIELD_SCTP_CHUNK_DATA_STREAM_SQN (NET_HEADER_FIELD_SCTP_CHUNK_DATA_TYPE << 5)
-+#define NET_HEADER_FIELD_SCTP_CHUNK_DATA_PAYLOAD_PID (NET_HEADER_FIELD_SCTP_CHUNK_DATA_TYPE << 6)
-+#define NET_HEADER_FIELD_SCTP_CHUNK_DATA_UNORDERED (NET_HEADER_FIELD_SCTP_CHUNK_DATA_TYPE << 7)
-+#define NET_HEADER_FIELD_SCTP_CHUNK_DATA_BEGGINING (NET_HEADER_FIELD_SCTP_CHUNK_DATA_TYPE << 8)
-+#define NET_HEADER_FIELD_SCTP_CHUNK_DATA_END (NET_HEADER_FIELD_SCTP_CHUNK_DATA_TYPE << 9)
-+#define NET_HEADER_FIELD_SCTP_CHUNK_DATA_ALL_FIELDS ((NET_HEADER_FIELD_SCTP_CHUNK_DATA_TYPE << 10) - 1)
-+
-+#define NET_HEADER_FIELD_L2TPv2_TYPE_BIT (1)
-+#define NET_HEADER_FIELD_L2TPv2_LENGTH_BIT (NET_HEADER_FIELD_L2TPv2_TYPE_BIT << 1)
-+#define NET_HEADER_FIELD_L2TPv2_SEQUENCE_BIT (NET_HEADER_FIELD_L2TPv2_TYPE_BIT << 2)
-+#define NET_HEADER_FIELD_L2TPv2_OFFSET_BIT (NET_HEADER_FIELD_L2TPv2_TYPE_BIT << 3)
-+#define NET_HEADER_FIELD_L2TPv2_PRIORITY_BIT (NET_HEADER_FIELD_L2TPv2_TYPE_BIT << 4)
-+#define NET_HEADER_FIELD_L2TPv2_VERSION (NET_HEADER_FIELD_L2TPv2_TYPE_BIT << 5)
-+#define NET_HEADER_FIELD_L2TPv2_LEN (NET_HEADER_FIELD_L2TPv2_TYPE_BIT << 6)
-+#define NET_HEADER_FIELD_L2TPv2_TUNNEL_ID (NET_HEADER_FIELD_L2TPv2_TYPE_BIT << 7)
-+#define NET_HEADER_FIELD_L2TPv2_SESSION_ID (NET_HEADER_FIELD_L2TPv2_TYPE_BIT << 8)
-+#define NET_HEADER_FIELD_L2TPv2_NS (NET_HEADER_FIELD_L2TPv2_TYPE_BIT << 9)
-+#define NET_HEADER_FIELD_L2TPv2_NR (NET_HEADER_FIELD_L2TPv2_TYPE_BIT << 10)
-+#define NET_HEADER_FIELD_L2TPv2_OFFSET_SIZE (NET_HEADER_FIELD_L2TPv2_TYPE_BIT << 11)
-+#define NET_HEADER_FIELD_L2TPv2_FIRST_BYTE (NET_HEADER_FIELD_L2TPv2_TYPE_BIT << 12)
-+#define NET_HEADER_FIELD_L2TPv2_ALL_FIELDS ((NET_HEADER_FIELD_L2TPv2_TYPE_BIT << 13) - 1)
-+
-+#define NET_HEADER_FIELD_L2TPv3_CTRL_TYPE_BIT (1)
-+#define NET_HEADER_FIELD_L2TPv3_CTRL_LENGTH_BIT (NET_HEADER_FIELD_L2TPv3_CTRL_TYPE_BIT << 1)
-+#define NET_HEADER_FIELD_L2TPv3_CTRL_SEQUENCE_BIT (NET_HEADER_FIELD_L2TPv3_CTRL_TYPE_BIT << 2)
-+#define NET_HEADER_FIELD_L2TPv3_CTRL_VERSION (NET_HEADER_FIELD_L2TPv3_CTRL_TYPE_BIT << 3)
-+#define NET_HEADER_FIELD_L2TPv3_CTRL_LENGTH (NET_HEADER_FIELD_L2TPv3_CTRL_TYPE_BIT << 4)
-+#define NET_HEADER_FIELD_L2TPv3_CTRL_CONTROL (NET_HEADER_FIELD_L2TPv3_CTRL_TYPE_BIT << 5)
-+#define NET_HEADER_FIELD_L2TPv3_CTRL_SENT (NET_HEADER_FIELD_L2TPv3_CTRL_TYPE_BIT << 6)
-+#define NET_HEADER_FIELD_L2TPv3_CTRL_RECV (NET_HEADER_FIELD_L2TPv3_CTRL_TYPE_BIT << 7)
-+#define NET_HEADER_FIELD_L2TPv3_CTRL_FIRST_BYTE (NET_HEADER_FIELD_L2TPv3_CTRL_TYPE_BIT << 8)
-+#define NET_HEADER_FIELD_L2TPv3_CTRL_ALL_FIELDS ((NET_HEADER_FIELD_L2TPv3_CTRL_TYPE_BIT << 9) - 1)
-+
-+#define NET_HEADER_FIELD_L2TPv3_SESS_TYPE_BIT (1)
-+#define NET_HEADER_FIELD_L2TPv3_SESS_VERSION (NET_HEADER_FIELD_L2TPv3_SESS_TYPE_BIT << 1)
-+#define NET_HEADER_FIELD_L2TPv3_SESS_ID (NET_HEADER_FIELD_L2TPv3_SESS_TYPE_BIT << 2)
-+#define NET_HEADER_FIELD_L2TPv3_SESS_COOKIE (NET_HEADER_FIELD_L2TPv3_SESS_TYPE_BIT << 3)
-+#define NET_HEADER_FIELD_L2TPv3_SESS_ALL_FIELDS ((NET_HEADER_FIELD_L2TPv3_SESS_TYPE_BIT << 4) - 1)
-+
-+
-+typedef uint8_t headerFieldVlan_t;
-+
-+#define NET_HEADER_FIELD_VLAN_VPRI (1)
-+#define NET_HEADER_FIELD_VLAN_CFI (NET_HEADER_FIELD_VLAN_VPRI << 1)
-+#define NET_HEADER_FIELD_VLAN_VID (NET_HEADER_FIELD_VLAN_VPRI << 2)
-+#define NET_HEADER_FIELD_VLAN_LENGTH (NET_HEADER_FIELD_VLAN_VPRI << 3)
-+#define NET_HEADER_FIELD_VLAN_TYPE (NET_HEADER_FIELD_VLAN_VPRI << 4)
-+#define NET_HEADER_FIELD_VLAN_ALL_FIELDS ((NET_HEADER_FIELD_VLAN_VPRI << 5) - 1)
-+
-+#define NET_HEADER_FIELD_VLAN_TCI (NET_HEADER_FIELD_VLAN_VPRI | \
-+ NET_HEADER_FIELD_VLAN_CFI | \
-+ NET_HEADER_FIELD_VLAN_VID)
-+
-+
-+typedef uint8_t headerFieldLlc_t;
-+
-+#define NET_HEADER_FIELD_LLC_DSAP (1)
-+#define NET_HEADER_FIELD_LLC_SSAP (NET_HEADER_FIELD_LLC_DSAP << 1)
-+#define NET_HEADER_FIELD_LLC_CTRL (NET_HEADER_FIELD_LLC_DSAP << 2)
-+#define NET_HEADER_FIELD_LLC_ALL_FIELDS ((NET_HEADER_FIELD_LLC_DSAP << 3) - 1)
-+
-+#define NET_HEADER_FIELD_NLPID_NLPID (1)
-+#define NET_HEADER_FIELD_NLPID_ALL_FIELDS ((NET_HEADER_FIELD_NLPID_NLPID << 1) - 1)
-+
-+
-+typedef uint8_t headerFieldSnap_t;
-+
-+#define NET_HEADER_FIELD_SNAP_OUI (1)
-+#define NET_HEADER_FIELD_SNAP_PID (NET_HEADER_FIELD_SNAP_OUI << 1)
-+#define NET_HEADER_FIELD_SNAP_ALL_FIELDS ((NET_HEADER_FIELD_SNAP_OUI << 2) - 1)
-+
-+
-+typedef uint8_t headerFieldLlcSnap_t;
-+
-+#define NET_HEADER_FIELD_LLC_SNAP_TYPE (1)
-+#define NET_HEADER_FIELD_LLC_SNAP_ALL_FIELDS ((NET_HEADER_FIELD_LLC_SNAP_TYPE << 1) - 1)
-+
-+#define NET_HEADER_FIELD_ARP_HTYPE (1)
-+#define NET_HEADER_FIELD_ARP_PTYPE (NET_HEADER_FIELD_ARP_HTYPE << 1)
-+#define NET_HEADER_FIELD_ARP_HLEN (NET_HEADER_FIELD_ARP_HTYPE << 2)
-+#define NET_HEADER_FIELD_ARP_PLEN (NET_HEADER_FIELD_ARP_HTYPE << 3)
-+#define NET_HEADER_FIELD_ARP_OPER (NET_HEADER_FIELD_ARP_HTYPE << 4)
-+#define NET_HEADER_FIELD_ARP_SHA (NET_HEADER_FIELD_ARP_HTYPE << 5)
-+#define NET_HEADER_FIELD_ARP_SPA (NET_HEADER_FIELD_ARP_HTYPE << 6)
-+#define NET_HEADER_FIELD_ARP_THA (NET_HEADER_FIELD_ARP_HTYPE << 7)
-+#define NET_HEADER_FIELD_ARP_TPA (NET_HEADER_FIELD_ARP_HTYPE << 8)
-+#define NET_HEADER_FIELD_ARP_ALL_FIELDS ((NET_HEADER_FIELD_ARP_HTYPE << 9) - 1)
-+
-+#define NET_HEADER_FIELD_RFC2684_LLC (1)
-+#define NET_HEADER_FIELD_RFC2684_NLPID (NET_HEADER_FIELD_RFC2684_LLC << 1)
-+#define NET_HEADER_FIELD_RFC2684_OUI (NET_HEADER_FIELD_RFC2684_LLC << 2)
-+#define NET_HEADER_FIELD_RFC2684_PID (NET_HEADER_FIELD_RFC2684_LLC << 3)
-+#define NET_HEADER_FIELD_RFC2684_VPN_OUI (NET_HEADER_FIELD_RFC2684_LLC << 4)
-+#define NET_HEADER_FIELD_RFC2684_VPN_IDX (NET_HEADER_FIELD_RFC2684_LLC << 5)
-+#define NET_HEADER_FIELD_RFC2684_ALL_FIELDS ((NET_HEADER_FIELD_RFC2684_LLC << 6) - 1)
-+
-+#define NET_HEADER_FIELD_USER_DEFINED_SRCPORT (1)
-+#define NET_HEADER_FIELD_USER_DEFINED_PCDID (NET_HEADER_FIELD_USER_DEFINED_SRCPORT << 1)
-+#define NET_HEADER_FIELD_USER_DEFINED_ALL_FIELDS ((NET_HEADER_FIELD_USER_DEFINED_SRCPORT << 2) - 1)
-+
-+#define NET_HEADER_FIELD_PAYLOAD_BUFFER (1)
-+#define NET_HEADER_FIELD_PAYLOAD_SIZE (NET_HEADER_FIELD_PAYLOAD_BUFFER << 1)
-+#define NET_HEADER_FIELD_MAX_FRM_SIZE (NET_HEADER_FIELD_PAYLOAD_BUFFER << 2)
-+#define NET_HEADER_FIELD_MIN_FRM_SIZE (NET_HEADER_FIELD_PAYLOAD_BUFFER << 3)
-+#define NET_HEADER_FIELD_PAYLOAD_TYPE (NET_HEADER_FIELD_PAYLOAD_BUFFER << 4)
-+#define NET_HEADER_FIELD_FRAME_SIZE (NET_HEADER_FIELD_PAYLOAD_BUFFER << 5)
-+#define NET_HEADER_FIELD_PAYLOAD_ALL_FIELDS ((NET_HEADER_FIELD_PAYLOAD_BUFFER << 6) - 1)
-+
-+
-+typedef uint8_t headerFieldGre_t;
-+
-+#define NET_HEADER_FIELD_GRE_TYPE (1)
-+#define NET_HEADER_FIELD_GRE_ALL_FIELDS ((NET_HEADER_FIELD_GRE_TYPE << 1) - 1)
-+
-+
-+typedef uint8_t headerFieldMinencap_t;
-+
-+#define NET_HEADER_FIELD_MINENCAP_SRC_IP (1)
-+#define NET_HEADER_FIELD_MINENCAP_DST_IP (NET_HEADER_FIELD_MINENCAP_SRC_IP << 1)
-+#define NET_HEADER_FIELD_MINENCAP_TYPE (NET_HEADER_FIELD_MINENCAP_SRC_IP << 2)
-+#define NET_HEADER_FIELD_MINENCAP_ALL_FIELDS ((NET_HEADER_FIELD_MINENCAP_SRC_IP << 3) - 1)
-+
-+
-+typedef uint8_t headerFieldIpsecAh_t;
-+
-+#define NET_HEADER_FIELD_IPSEC_AH_SPI (1)
-+#define NET_HEADER_FIELD_IPSEC_AH_NH (NET_HEADER_FIELD_IPSEC_AH_SPI << 1)
-+#define NET_HEADER_FIELD_IPSEC_AH_ALL_FIELDS ((NET_HEADER_FIELD_IPSEC_AH_SPI << 2) - 1)
-+
-+
-+typedef uint8_t headerFieldIpsecEsp_t;
-+
-+#define NET_HEADER_FIELD_IPSEC_ESP_SPI (1)
-+#define NET_HEADER_FIELD_IPSEC_ESP_SEQUENCE_NUM (NET_HEADER_FIELD_IPSEC_ESP_SPI << 1)
-+#define NET_HEADER_FIELD_IPSEC_ESP_ALL_FIELDS ((NET_HEADER_FIELD_IPSEC_ESP_SPI << 2) - 1)
-+
-+#define NET_HEADER_FIELD_IPSEC_ESP_SPI_SIZE 4
-+
-+
-+typedef uint8_t headerFieldMpls_t;
-+
-+#define NET_HEADER_FIELD_MPLS_LABEL_STACK (1)
-+#define NET_HEADER_FIELD_MPLS_LABEL_STACK_ALL_FIELDS ((NET_HEADER_FIELD_MPLS_LABEL_STACK << 1) - 1)
-+
-+
-+typedef uint8_t headerFieldMacsec_t;
-+
-+#define NET_HEADER_FIELD_MACSEC_SECTAG (1)
-+#define NET_HEADER_FIELD_MACSEC_ALL_FIELDS ((NET_HEADER_FIELD_MACSEC_SECTAG << 1) - 1)
-+
-+
-+typedef enum {
-+ HEADER_TYPE_NONE = 0,
-+ HEADER_TYPE_PAYLOAD,
-+ HEADER_TYPE_ETH,
-+ HEADER_TYPE_VLAN,
-+ HEADER_TYPE_IPv4,
-+ HEADER_TYPE_IPv6,
-+ HEADER_TYPE_IP,
-+ HEADER_TYPE_TCP,
-+ HEADER_TYPE_UDP,
-+ HEADER_TYPE_UDP_LITE,
-+ HEADER_TYPE_IPHC,
-+ HEADER_TYPE_SCTP,
-+ HEADER_TYPE_SCTP_CHUNK_DATA,
-+ HEADER_TYPE_PPPoE,
-+ HEADER_TYPE_PPP,
-+ HEADER_TYPE_PPPMUX,
-+ HEADER_TYPE_PPPMUX_SUBFRAME,
-+ HEADER_TYPE_L2TPv2,
-+ HEADER_TYPE_L2TPv3_CTRL,
-+ HEADER_TYPE_L2TPv3_SESS,
-+ HEADER_TYPE_LLC,
-+ HEADER_TYPE_LLC_SNAP,
-+ HEADER_TYPE_NLPID,
-+ HEADER_TYPE_SNAP,
-+ HEADER_TYPE_MPLS,
-+ HEADER_TYPE_IPSEC_AH,
-+ HEADER_TYPE_IPSEC_ESP,
-+ HEADER_TYPE_UDP_ENCAP_ESP, /* RFC 3948 */
-+ HEADER_TYPE_MACSEC,
-+ HEADER_TYPE_GRE,
-+ HEADER_TYPE_MINENCAP,
-+ HEADER_TYPE_DCCP,
-+ HEADER_TYPE_ICMP,
-+ HEADER_TYPE_IGMP,
-+ HEADER_TYPE_ARP,
-+ HEADER_TYPE_CAPWAP,
-+ HEADER_TYPE_CAPWAP_DTLS,
-+ HEADER_TYPE_RFC2684,
-+ HEADER_TYPE_USER_DEFINED_L2,
-+ HEADER_TYPE_USER_DEFINED_L3,
-+ HEADER_TYPE_USER_DEFINED_L4,
-+ HEADER_TYPE_USER_DEFINED_SHIM1,
-+ HEADER_TYPE_USER_DEFINED_SHIM2,
-+ MAX_HEADER_TYPE_COUNT
-+} e_NetHeaderType;
-+
-+
-+#endif /* __NET_EXT_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/std_ext.h
-@@ -0,0 +1,48 @@
-+/* Copyright (c) 2008-2012 Freescale Semiconductor, Inc
-+ * All rights reserved.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/**************************************************************************//**
-+ @File std_ext.h
-+
-+ @Description General Standard Definitions
-+*//***************************************************************************/
-+
-+#ifndef __STD_EXT_H
-+#define __STD_EXT_H
-+
-+
-+#include "types_ext.h"
-+#include "ncsw_ext.h"
-+
-+
-+#endif /* __STD_EXT_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/stdarg_ext.h
-@@ -0,0 +1,49 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+#ifndef __STDARG_EXT_H
-+#define __STDARG_EXT_H
-+
-+
-+#if defined(NCSW_LINUX) && defined(__KERNEL__)
-+#include <stdarg.h>
-+
-+#else
-+#include <stdarg.h>
-+
-+#endif /* defined(NCSW_LINUX) && defined(__KERNEL__) */
-+
-+#include "std_ext.h"
-+
-+
-+#endif /* __STDARG_EXT_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/stdlib_ext.h
-@@ -0,0 +1,162 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+
-+#ifndef __STDLIB_EXT_H
-+#define __STDLIB_EXT_H
-+
-+
-+#if (defined(NCSW_LINUX)) && defined(__KERNEL__)
-+#include "stdarg_ext.h"
-+#include "std_ext.h"
-+
-+
-+/**
-+ * strtoul - convert a string to an uint32_t
-+ * @cp: The start of the string
-+ * @endp: A pointer to the end of the parsed string will be placed here
-+ * @base: The number base to use
-+ */
-+uint32_t strtoul(const char *cp,char **endp,uint32_t base);
-+
-+/**
-+ * strtol - convert a string to a int32_t
-+ * @cp: The start of the string
-+ * @endp: A pointer to the end of the parsed string will be placed here
-+ * @base: The number base to use
-+ */
-+long strtol(const char *cp,char **endp,uint32_t base);
-+
-+/**
-+ * strtoull - convert a string to an uint64_t
-+ * @cp: The start of the string
-+ * @endp: A pointer to the end of the parsed string will be placed here
-+ * @base: The number base to use
-+ */
-+uint64_t strtoull(const char *cp,char **endp,uint32_t base);
-+
-+/**
-+ * strtoll - convert a string to a int64 long
-+ * @cp: The start of the string
-+ * @endp: A pointer to the end of the parsed string will be placed here
-+ * @base: The number base to use
-+ */
-+long long strtoll(const char *cp,char **endp,uint32_t base);
-+
-+/**
-+ * atoi - convert a character to a int
-+ * @s: The start of the string
-+ */
-+int atoi(const char *s);
-+
-+/**
-+ * strnlen - Find the length of a length-limited string
-+ * @s: The string to be sized
-+ * @count: The maximum number of bytes to search
-+ */
-+size_t strnlen(const char * s, size_t count);
-+
-+/**
-+ * strlen - Find the length of a string
-+ * @s: The string to be sized
-+ */
-+size_t strlen(const char * s);
-+
-+/**
-+ * strtok - Split a string into tokens
-+ * @s: The string to be searched
-+ * @ct: The characters to search for
-+ *
-+ * WARNING: strtok is deprecated, use strsep instead.
-+ */
-+char * strtok(char * s,const char * ct);
-+
-+/**
-+ * strncpy - Copy a length-limited, %NUL-terminated string
-+ * @dest: Where to copy the string to
-+ * @src: Where to copy the string from
-+ * @count: The maximum number of bytes to copy
-+ *
-+ * Note that unlike userspace strncpy, this does not %NUL-pad the buffer.
-+ * However, the result is not %NUL-terminated if the source exceeds
-+ * @count bytes.
-+ */
-+char * strncpy(char * dest,const char *src,size_t count);
-+
-+/**
-+ * strcpy - Copy a %NUL terminated string
-+ * @dest: Where to copy the string to
-+ * @src: Where to copy the string from
-+ */
-+char * strcpy(char * dest,const char *src);
-+
-+/**
-+ * vsscanf - Unformat a buffer into a list of arguments
-+ * @buf: input buffer
-+ * @fmt: format of buffer
-+ * @args: arguments
-+ */
-+int vsscanf(const char * buf, const char * fmt, va_list args);
-+
-+/**
-+ * vsnprintf - Format a string and place it in a buffer
-+ * @buf: The buffer to place the result into
-+ * @size: The size of the buffer, including the trailing null space
-+ * @fmt: The format string to use
-+ * @args: Arguments for the format string
-+ *
-+ * Call this function if you are already dealing with a va_list.
-+ * You probably want snprintf instead.
-+ */
-+int vsnprintf(char *buf, size_t size, const char *fmt, va_list args);
-+
-+/**
-+ * vsprintf - Format a string and place it in a buffer
-+ * @buf: The buffer to place the result into
-+ * @fmt: The format string to use
-+ * @args: Arguments for the format string
-+ *
-+ * Call this function if you are already dealing with a va_list.
-+ * You probably want sprintf instead.
-+ */
-+int vsprintf(char *buf, const char *fmt, va_list args);
-+
-+#else
-+#include <stdlib.h>
-+#include <stdio.h>
-+#endif /* defined(NCSW_LINUX) && defined(__KERNEL__) */
-+
-+#include "std_ext.h"
-+
-+
-+#endif /* __STDLIB_EXT_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/string_ext.h
-@@ -0,0 +1,56 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+#ifndef __STRING_EXT_H
-+#define __STRING_EXT_H
-+
-+
-+#if defined(NCSW_LINUX) && defined(__KERNEL__)
-+#include <linux/kernel.h>
-+#include <linux/string.h>
-+extern char * strtok ( char * str, const char * delimiters );
-+
-+#elif defined(__KERNEL__)
-+#include "linux/types.h"
-+#include "linux/posix_types.h"
-+#include "linux/string.h"
-+
-+#else
-+#include <string.h>
-+
-+#endif /* defined(NCSW_LINUX) && defined(__KERNEL__) */
-+
-+#include "std_ext.h"
-+
-+
-+#endif /* __STRING_EXT_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/types_ext.h
-@@ -0,0 +1,62 @@
-+/* Copyright (c) 2008-2012 Freescale Semiconductor, Inc
-+ * All rights reserved.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/**************************************************************************//**
-+ @File types_ext.h
-+
-+ @Description General types Standard Definitions
-+*//***************************************************************************/
-+
-+#ifndef __TYPES_EXT_H
-+#define __TYPES_EXT_H
-+
-+#if defined(NCSW_LINUX)
-+#include "types_linux.h"
-+
-+#elif defined(NCSW_VXWORKS)
-+#include "types_vxworks.h"
-+
-+#elif defined(__GNUC__) && defined(__cplusplus)
-+#include "types_bb_gpp.h"
-+
-+#elif defined(__GNUC__)
-+#include "types_bb_gcc.h"
-+
-+#elif defined(__ghs__)
-+#include "types_ghs.h"
-+
-+#else
-+#include "types_dflt.h"
-+#endif /* defined (__ROCOO__) */
-+
-+#endif /* __TYPES_EXT_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/xx_common.h
-@@ -0,0 +1,56 @@
-+/*
-+ * Copyright 2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/**************************************************************************//**
-+ @File debug_ext.h
-+
-+ @Description Debug mode definitions.
-+*//***************************************************************************/
-+
-+#ifndef __XX_COMMON_H
-+#define __XX_COMMON_H
-+
-+/*****************************************************************************
-+ * UNIFIED MODULE CODES
-+ *****************************************************************************/
-+#define MODULE_UNKNOWN 0x00000000
-+#define MODULE_FM 0x00010000
-+#define MODULE_FM_MURAM 0x00020000
-+#define MODULE_FM_PCD 0x00030000
-+#define MODULE_FM_RTC 0x00040000
-+#define MODULE_FM_MAC 0x00050000
-+#define MODULE_FM_PORT 0x00060000
-+#define MODULE_MM 0x00070000
-+#define MODULE_FM_SP 0x00080000
-+#define MODULE_FM_MACSEC 0x00090000
-+#endif /* __XX_COMMON_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/inc/xx_ext.h
-@@ -0,0 +1,791 @@
-+/* Copyright (c) 2008-2012 Freescale Semiconductor, Inc
-+ * All rights reserved.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/**************************************************************************//**
-+ @File xx_ext.h
-+
-+ @Description Prototypes, externals and typedefs for system-supplied
-+ (external) routines
-+*//***************************************************************************/
-+
-+#ifndef __XX_EXT_H
-+#define __XX_EXT_H
-+
-+#include "std_ext.h"
-+#include "xx_common.h"
-+#include "part_ext.h"
-+
-+
-+
-+/**************************************************************************//**
-+ @Group xx_id XX Interface (System call hooks)
-+
-+ @Description Prototypes, externals and typedefs for system-supplied
-+ (external) routines
-+
-+ @{
-+*//***************************************************************************/
-+
-+#ifdef DEBUG_XX_MALLOC
-+void * XX_MallocDebug(uint32_t size, char *fname, int line);
-+
-+void * XX_MallocSmartDebug(uint32_t size,
-+ int memPartitionId,
-+ uint32_t alignment,
-+ char *fname,
-+ int line);
-+
-+#define XX_Malloc(sz) \
-+ XX_MallocDebug((sz), __FILE__, __LINE__)
-+
-+#define XX_MallocSmart(sz, memt, al) \
-+ XX_MallocSmartDebug((sz), (memt), (al), __FILE__, __LINE__)
-+
-+#else /* not DEBUG_XX_MALLOC */
-+/**************************************************************************//**
-+ @Function XX_Malloc
-+
-+ @Description allocates contiguous block of memory.
-+
-+ @Param[in] size - Number of bytes to allocate.
-+
-+ @Return The address of the newly allocated block on success, NULL on failure.
-+*//***************************************************************************/
-+void * XX_Malloc(uint32_t size);
-+
-+/**************************************************************************//**
-+ @Function XX_MallocSmart
-+
-+ @Description Allocates contiguous block of memory in a specified
-+ alignment and from the specified segment.
-+
-+ @Param[in] size - Number of bytes to allocate.
-+ @Param[in] memPartitionId - Memory partition ID; The value zero must
-+ be mapped to the default heap partition.
-+ @Param[in] alignment - Required memory alignment (in bytes).
-+
-+ @Return The address of the newly allocated block on success, NULL on failure.
-+*//***************************************************************************/
-+void * XX_MallocSmart(uint32_t size, int memPartitionId, uint32_t alignment);
-+#endif /* not DEBUG_XX_MALLOC */
-+
-+/**************************************************************************//**
-+ @Function XX_FreeSmart
-+
-+ @Description Frees the memory block pointed to by "p".
-+ Only for memory allocated by XX_MallocSmart
-+
-+ @Param[in] p_Memory - pointer to the memory block.
-+
-+ @Return None.
-+*//***************************************************************************/
-+void XX_FreeSmart(void *p_Memory);
-+
-+/**************************************************************************//**
-+ @Function XX_Free
-+
-+ @Description frees the memory block pointed to by "p".
-+
-+ @Param[in] p_Memory - pointer to the memory block.
-+
-+ @Return None.
-+*//***************************************************************************/
-+void XX_Free(void *p_Memory);
-+
-+/**************************************************************************//**
-+ @Function XX_Print
-+
-+ @Description print a string.
-+
-+ @Param[in] str - string to print.
-+
-+ @Return None.
-+*//***************************************************************************/
-+void XX_Print(char *str, ...);
-+
-+/**************************************************************************//**
-+ @Function XX_SetIntr
-+
-+ @Description Set an interrupt service routine for a specific interrupt source.
-+
-+ @Param[in] irq - Interrupt ID (system-specific number).
-+ @Param[in] f_Isr - Callback routine that will be called when the interrupt occurs.
-+ @Param[in] handle - The argument for the user callback routine.
-+
-+ @Return E_OK on success; error code otherwise..
-+*//***************************************************************************/
-+t_Error XX_SetIntr(int irq, t_Isr *f_Isr, t_Handle handle);
-+
-+/**************************************************************************//**
-+ @Function XX_FreeIntr
-+
-+ @Description Free a specific interrupt and a specific callback routine.
-+
-+ @Param[in] irq - Interrupt ID (system-specific number).
-+
-+ @Return E_OK on success; error code otherwise..
-+*//***************************************************************************/
-+t_Error XX_FreeIntr(int irq);
-+
-+/**************************************************************************//**
-+ @Function XX_EnableIntr
-+
-+ @Description Enable a specific interrupt.
-+
-+ @Param[in] irq - Interrupt ID (system-specific number).
-+
-+ @Return E_OK on success; error code otherwise..
-+*//***************************************************************************/
-+t_Error XX_EnableIntr(int irq);
-+
-+/**************************************************************************//**
-+ @Function XX_DisableIntr
-+
-+ @Description Disable a specific interrupt.
-+
-+ @Param[in] irq - Interrupt ID (system-specific number).
-+
-+ @Return E_OK on success; error code otherwise..
-+*//***************************************************************************/
-+t_Error XX_DisableIntr(int irq);
-+
-+/**************************************************************************//**
-+ @Function XX_DisableAllIntr
-+
-+ @Description Disable all interrupts by masking them at the CPU.
-+
-+ @Return A value that represents the interrupts state before the
-+ operation, and should be passed to the matching
-+ XX_RestoreAllIntr() call.
-+*//***************************************************************************/
-+uint32_t XX_DisableAllIntr(void);
-+
-+/**************************************************************************//**
-+ @Function XX_RestoreAllIntr
-+
-+ @Description Restore previous state of interrupts level at the CPU.
-+
-+ @Param[in] flags - A value that represents the interrupts state to restore,
-+ as returned by the matching call for XX_DisableAllIntr().
-+
-+ @Return None.
-+*//***************************************************************************/
-+void XX_RestoreAllIntr(uint32_t flags);
-+
-+
-+/**************************************************************************//**
-+ @Function XX_Exit
-+
-+ @Description Stop execution and report status (where it is applicable)
-+
-+ @Param[in] status - exit status
-+*//***************************************************************************/
-+void XX_Exit(int status);
-+
-+
-+/*****************************************************************************/
-+/* Tasklet Service Routines */
-+/*****************************************************************************/
-+typedef t_Handle t_TaskletHandle;
-+
-+/**************************************************************************//**
-+ @Function XX_InitTasklet
-+
-+ @Description Create and initialize a tasklet object.
-+
-+ @Param[in] routine - A routine to be ran as a tasklet.
-+ @Param[in] data - An argument to pass to the tasklet.
-+
-+ @Return Tasklet handle is returned on success. NULL is returned otherwise.
-+*//***************************************************************************/
-+t_TaskletHandle XX_InitTasklet (void (*routine)(void *), void *data);
-+
-+/**************************************************************************//**
-+ @Function XX_FreeTasklet
-+
-+ @Description Free a tasklet object.
-+
-+ @Param[in] h_Tasklet - A handle to a tasklet to be free.
-+
-+ @Return None.
-+*//***************************************************************************/
-+void XX_FreeTasklet (t_TaskletHandle h_Tasklet);
-+
-+/**************************************************************************//**
-+ @Function XX_ScheduleTask
-+
-+ @Description Schedule a tasklet object.
-+
-+ @Param[in] h_Tasklet - A handle to a tasklet to be scheduled.
-+ @Param[in] immediate - Indicate whether to schedule this tasklet on
-+ the immediate queue or on the delayed one.
-+
-+ @Return 0 - on success. Error code - otherwise.
-+*//***************************************************************************/
-+int XX_ScheduleTask(t_TaskletHandle h_Tasklet, int immediate);
-+
-+/**************************************************************************//**
-+ @Function XX_FlushScheduledTasks
-+
-+ @Description Flush all tasks there are in the scheduled tasks queue.
-+
-+ @Return None.
-+*//***************************************************************************/
-+void XX_FlushScheduledTasks(void);
-+
-+/**************************************************************************//**
-+ @Function XX_TaskletIsQueued
-+
-+ @Description Check if task is queued.
-+
-+ @Param[in] h_Tasklet - A handle to a tasklet to be scheduled.
-+
-+ @Return 1 - task is queued. 0 - otherwise.
-+*//***************************************************************************/
-+int XX_TaskletIsQueued(t_TaskletHandle h_Tasklet);
-+
-+/**************************************************************************//**
-+ @Function XX_SetTaskletData
-+
-+ @Description Set data to a scheduled task. Used to change data of already
-+ scheduled task.
-+
-+ @Param[in] h_Tasklet - A handle to a tasklet to be scheduled.
-+ @Param[in] data - Data to be set.
-+*//***************************************************************************/
-+void XX_SetTaskletData(t_TaskletHandle h_Tasklet, t_Handle data);
-+
-+/**************************************************************************//**
-+ @Function XX_GetTaskletData
-+
-+ @Description Get the data of scheduled task.
-+
-+ @Param[in] h_Tasklet - A handle to a tasklet to be scheduled.
-+
-+ @Return handle to the data of the task.
-+*//***************************************************************************/
-+t_Handle XX_GetTaskletData(t_TaskletHandle h_Tasklet);
-+
-+/**************************************************************************//**
-+ @Function XX_BottomHalf
-+
-+ @Description Bottom half implementation, invoked by the interrupt handler.
-+
-+ This routine handles all bottom-half tasklets with interrupts
-+ enabled.
-+
-+ @Return None.
-+*//***************************************************************************/
-+void XX_BottomHalf(void);
-+
-+
-+/*****************************************************************************/
-+/* Spinlock Service Routines */
-+/*****************************************************************************/
-+
-+/**************************************************************************//**
-+ @Function XX_InitSpinlock
-+
-+ @Description Creates a spinlock.
-+
-+ @Return Spinlock handle is returned on success; NULL otherwise.
-+*//***************************************************************************/
-+t_Handle XX_InitSpinlock(void);
-+
-+/**************************************************************************//**
-+ @Function XX_FreeSpinlock
-+
-+ @Description Frees the memory allocated for the spinlock creation.
-+
-+ @Param[in] h_Spinlock - A handle to a spinlock.
-+
-+ @Return None.
-+*//***************************************************************************/
-+void XX_FreeSpinlock(t_Handle h_Spinlock);
-+
-+/**************************************************************************//**
-+ @Function XX_LockSpinlock
-+
-+ @Description Locks a spinlock.
-+
-+ @Param[in] h_Spinlock - A handle to a spinlock.
-+
-+ @Return None.
-+*//***************************************************************************/
-+void XX_LockSpinlock(t_Handle h_Spinlock);
-+
-+/**************************************************************************//**
-+ @Function XX_UnlockSpinlock
-+
-+ @Description Unlocks a spinlock.
-+
-+ @Param[in] h_Spinlock - A handle to a spinlock.
-+
-+ @Return None.
-+*//***************************************************************************/
-+void XX_UnlockSpinlock(t_Handle h_Spinlock);
-+
-+/**************************************************************************//**
-+ @Function XX_LockIntrSpinlock
-+
-+ @Description Locks a spinlock (interrupt safe).
-+
-+ @Param[in] h_Spinlock - A handle to a spinlock.
-+
-+ @Return A value that represents the interrupts state before the
-+ operation, and should be passed to the matching
-+ XX_UnlockIntrSpinlock() call.
-+*//***************************************************************************/
-+uint32_t XX_LockIntrSpinlock(t_Handle h_Spinlock);
-+
-+/**************************************************************************//**
-+ @Function XX_UnlockIntrSpinlock
-+
-+ @Description Unlocks a spinlock (interrupt safe).
-+
-+ @Param[in] h_Spinlock - A handle to a spinlock.
-+ @Param[in] intrFlags - A value that represents the interrupts state to
-+ restore, as returned by the matching call for
-+ XX_LockIntrSpinlock().
-+
-+ @Return None.
-+*//***************************************************************************/
-+void XX_UnlockIntrSpinlock(t_Handle h_Spinlock, uint32_t intrFlags);
-+
-+
-+/*****************************************************************************/
-+/* Timers Service Routines */
-+/*****************************************************************************/
-+
-+/**************************************************************************//**
-+ @Function XX_CurrentTime
-+
-+ @Description Returns current system time.
-+
-+ @Return Current system time (in milliseconds).
-+*//***************************************************************************/
-+uint32_t XX_CurrentTime(void);
-+
-+/**************************************************************************//**
-+ @Function XX_CreateTimer
-+
-+ @Description Creates a timer.
-+
-+ @Return Timer handle is returned on success; NULL otherwise.
-+*//***************************************************************************/
-+t_Handle XX_CreateTimer(void);
-+
-+/**************************************************************************//**
-+ @Function XX_FreeTimer
-+
-+ @Description Frees the memory allocated for the timer creation.
-+
-+ @Param[in] h_Timer - A handle to a timer.
-+
-+ @Return None.
-+*//***************************************************************************/
-+void XX_FreeTimer(t_Handle h_Timer);
-+
-+/**************************************************************************//**
-+ @Function XX_StartTimer
-+
-+ @Description Starts a timer.
-+
-+ The user can select to start the timer as periodic timer or as
-+ one-shot timer. The user should provide a callback routine that
-+ will be called when the timer expires.
-+
-+ @Param[in] h_Timer - A handle to a timer.
-+ @Param[in] msecs - Timer expiration period (in milliseconds).
-+ @Param[in] periodic - TRUE for a periodic timer;
-+ FALSE for a one-shot timer..
-+ @Param[in] f_TimerExpired - A callback routine to be called when the
-+ timer expires.
-+ @Param[in] h_Arg - The argument to pass in the timer-expired
-+ callback routine.
-+
-+ @Return None.
-+*//***************************************************************************/
-+void XX_StartTimer(t_Handle h_Timer,
-+ uint32_t msecs,
-+ bool periodic,
-+ void (*f_TimerExpired)(t_Handle h_Arg),
-+ t_Handle h_Arg);
-+
-+/**************************************************************************//**
-+ @Function XX_StopTimer
-+
-+ @Description Frees the memory allocated for the timer creation.
-+
-+ @Param[in] h_Timer - A handle to a timer.
-+
-+ @Return None.
-+*//***************************************************************************/
-+void XX_StopTimer(t_Handle h_Timer);
-+
-+/**************************************************************************//**
-+ @Function XX_ModTimer
-+
-+ @Description Updates the expiration time of a timer.
-+
-+ This routine adds the given time to the current system time,
-+ and sets this value as the new expiration time of the timer.
-+
-+ @Param[in] h_Timer - A handle to a timer.
-+ @Param[in] msecs - The new interval until timer expiration
-+ (in milliseconds).
-+
-+ @Return None.
-+*//***************************************************************************/
-+void XX_ModTimer(t_Handle h_Timer, uint32_t msecs);
-+
-+/**************************************************************************//**
-+ @Function XX_Sleep
-+
-+ @Description Non-busy wait until the desired time (in milliseconds) has passed.
-+
-+ @Param[in] msecs - The requested sleep time (in milliseconds).
-+
-+ @Return Zero if the requested time has elapsed; Otherwise, the value
-+ returned will be the unslept amount) in milliseconds.
-+
-+ @Cautions This routine enables interrupts during its wait time.
-+*//***************************************************************************/
-+uint32_t XX_Sleep(uint32_t msecs);
-+
-+/**************************************************************************//**
-+ @Function XX_UDelay
-+
-+ @Description Busy-wait until the desired time (in microseconds) has passed.
-+
-+ @Param[in] usecs - The requested delay time (in microseconds).
-+
-+ @Return None.
-+
-+ @Cautions It is highly unrecommended to call this routine during interrupt
-+ time, because the system time may not be updated properly during
-+ the delay loop. The behavior of this routine during interrupt
-+ time is unexpected.
-+*//***************************************************************************/
-+void XX_UDelay(uint32_t usecs);
-+
-+
-+/*****************************************************************************/
-+/* Other Service Routines */
-+/*****************************************************************************/
-+
-+/**************************************************************************//**
-+ @Function XX_PhysToVirt
-+
-+ @Description Translates a physical address to the matching virtual address.
-+
-+ @Param[in] addr - The physical address to translate.
-+
-+ @Return Virtual address.
-+*//***************************************************************************/
-+void * XX_PhysToVirt(physAddress_t addr);
-+
-+/**************************************************************************//**
-+ @Function XX_VirtToPhys
-+
-+ @Description Translates a virtual address to the matching physical address.
-+
-+ @Param[in] addr - The virtual address to translate.
-+
-+ @Return Physical address.
-+*//***************************************************************************/
-+physAddress_t XX_VirtToPhys(void *addr);
-+
-+
-+/**************************************************************************//**
-+ @Group xx_ipc XX Inter-Partition-Communication API
-+
-+ @Description The following API is to be used when working with multiple
-+ partitions configuration.
-+
-+ @{
-+*//***************************************************************************/
-+
-+#define XX_IPC_MAX_ADDR_NAME_LENGTH 16 /**< Maximum length of an endpoint name string;
-+ The IPC service can use this constant to limit
-+ the storage space for IPC endpoint names. */
-+
-+
-+/**************************************************************************//**
-+ @Function t_IpcMsgCompletion
-+
-+ @Description Callback function used upon IPC non-blocking transaction completion
-+ to return message buffer to the caller and to forward reply if available.
-+
-+ This callback function may be attached by the source endpoint to any outgoing
-+ IPC message to indicate a non-blocking send (see also XX_IpcSendMessage() routine).
-+ Upon completion of an IPC transaction (consisting of a message and an optional reply),
-+ the IPC service invokes this callback routine to return the message buffer to the sender
-+ and to provide the received reply, if requested.
-+
-+ User provides this function. Driver invokes it.
-+
-+ @Param[in] h_Module - Abstract handle to the sending module - the same handle as was passed
-+ in the XX_IpcSendMessage() function; This handle is typically used to point
-+ to the internal data structure of the source endpoint.
-+ @Param[in] p_Msg - Pointer to original (sent) message buffer;
-+ The source endpoint can free (or reuse) this buffer when message
-+ completion callback is called.
-+ @Param[in] p_Reply - Pointer to (received) reply buffer;
-+ This pointer is the same as was provided by the source endpoint in
-+ XX_IpcSendMessage().
-+ @Param[in] replyLength - Length (in bytes) of actual data in the reply buffer.
-+ @Param[in] status - Completion status - E_OK or failure indication, e.g. IPC transaction completion
-+ timeout.
-+
-+ @Return None
-+ *//***************************************************************************/
-+typedef void (t_IpcMsgCompletion)(t_Handle h_Module,
-+ uint8_t *p_Msg,
-+ uint8_t *p_Reply,
-+ uint32_t replyLength,
-+ t_Error status);
-+
-+/**************************************************************************//**
-+ @Function t_IpcMsgHandler
-+
-+ @Description Callback function used as IPC message handler.
-+
-+ The IPC service invokes message handlers for each IPC message received.
-+ The actual function pointer should be registered by each destination endpoint
-+ via the XX_IpcRegisterMsgHandler() routine.
-+
-+ User provides this function. Driver invokes it.
-+
-+ @Param[in] h_Module - Abstract handle to the message handling module - the same handle as
-+ was passed in the XX_IpcRegisterMsgHandler() function; this handle is
-+ typically used to point to the internal data structure of the destination
-+ endpoint.
-+ @Param[in] p_Msg - Pointer to message buffer with data received from peer.
-+ @Param[in] msgLength - Length (in bytes) of message data.
-+ @Param[in] p_Reply - Pointer to reply buffer, to be filled by the message handler and then sent
-+ by the IPC service;
-+ The reply buffer is allocated by the IPC service with size equals to the
-+ replyLength parameter provided in message handler registration (see
-+ XX_IpcRegisterMsgHandler() function);
-+ If replyLength was initially specified as zero during message handler registration,
-+ the IPC service may set this pointer to NULL and assume that a reply is not needed;
-+ The IPC service is also responsible for freeing the reply buffer after the
-+ reply has been sent or dismissed.
-+ @Param[in,out] p_ReplyLength - Pointer to reply length, which has a dual role in this function:
-+ [In] equals the replyLength parameter provided in message handler
-+ registration (see XX_IpcRegisterMsgHandler() function), and
-+ [Out] should be updated by message handler to the actual reply length; if
-+ this value is set to zero, the IPC service must assume that a reply should
-+ not be sent;
-+ Note: If p_Reply is not NULL, p_ReplyLength must not be NULL as well.
-+
-+ @Return E_OK on success; Error code otherwise.
-+ *//***************************************************************************/
-+typedef t_Error (t_IpcMsgHandler)(t_Handle h_Module,
-+ uint8_t *p_Msg,
-+ uint32_t msgLength,
-+ uint8_t *p_Reply,
-+ uint32_t *p_ReplyLength);
-+
-+/**************************************************************************//**
-+ @Function XX_IpcRegisterMsgHandler
-+
-+ @Description IPC mailbox registration.
-+
-+ This function is used for registering an IPC message handler in the IPC service.
-+ This function is called by each destination endpoint to indicate that it is ready
-+ to handle incoming messages. The IPC service invokes the message handler upon receiving
-+ a message addressed to the specified destination endpoint.
-+
-+ @Param[in] addr - The address name string associated with the destination endpoint;
-+ This address must be unique across the IPC service domain to ensure
-+ correct message routing.
-+ @Param[in] f_MsgHandler - Pointer to the message handler callback for processing incoming
-+ message; invoked by the IPC service upon receiving a message
-+ addressed to the destination endpoint specified by the addr
-+ parameter.
-+ @Param[in] h_Module - Abstract handle to the message handling module, passed unchanged
-+ to f_MsgHandler callback function.
-+ @Param[in] replyLength - The maximal data length (in bytes) of any reply that the specified message handler
-+ may generate; the IPC service provides the message handler with buffer
-+ for reply according to the length specified here (refer also to the description
-+ of #t_IpcMsgHandler callback function type);
-+ This size shall be zero if the message handler never generates replies.
-+
-+ @Return E_OK on success; Error code otherwise.
-+*//***************************************************************************/
-+t_Error XX_IpcRegisterMsgHandler(char addr[XX_IPC_MAX_ADDR_NAME_LENGTH],
-+ t_IpcMsgHandler *f_MsgHandler,
-+ t_Handle h_Module,
-+ uint32_t replyLength);
-+
-+/**************************************************************************//**
-+ @Function XX_IpcUnregisterMsgHandler
-+
-+ @Description Release IPC mailbox routine.
-+
-+ This function is used for unregistering an IPC message handler from the IPC service.
-+ This function is called by each destination endpoint to indicate that it is no longer
-+ capable of handling incoming messages.
-+
-+ @Param[in] addr - The address name string associated with the destination endpoint;
-+ This address is the same as was used when the message handler was
-+ registered via XX_IpcRegisterMsgHandler().
-+
-+ @Return E_OK on success; Error code otherwise.
-+*//***************************************************************************/
-+t_Error XX_IpcUnregisterMsgHandler(char addr[XX_IPC_MAX_ADDR_NAME_LENGTH]);
-+
-+/**************************************************************************//**
-+ @Function XX_IpcInitSession
-+
-+ @Description This function is used for creating an IPC session between the source endpoint
-+ and the destination endpoint.
-+
-+ The actual implementation and representation of a session is left for the IPC service.
-+ The function returns an abstract handle to the created session. This handle shall be used
-+ by the source endpoint in subsequent calls to XX_IpcSendMessage().
-+ The IPC service assumes that before this function is called, no messages are sent from
-+ the specified source endpoint to the specified destination endpoint.
-+
-+ The IPC service may use a connection-oriented approach or a connectionless approach (or both)
-+ as described below.
-+
-+ @par Connection-Oriented Approach
-+
-+ The IPC service may implement a session in a connection-oriented approach - when this function is called,
-+ the IPC service should take the necessary steps to bring up a source-to-destination channel for messages
-+ and a destination-to-source channel for replies. The returned handle should represent the internal
-+ representation of these channels.
-+
-+ @par Connectionless Approach
-+
-+ The IPC service may implement a session in a connectionless approach - when this function is called, the
-+ IPC service should not perform any particular steps, but it must store the pair of source and destination
-+ addresses in some session representation and return it as a handle. When XX_IpcSendMessage() shall be
-+ called, the IPC service may use this handle to provide the necessary identifiers for routing the messages
-+ through the connectionless medium.
-+
-+ @Param[in] destAddr - The address name string associated with the destination endpoint.
-+ @Param[in] srcAddr - The address name string associated with the source endpoint.
-+
-+ @Return Abstract handle to the initialized session, or NULL on error.
-+*//***************************************************************************/
-+t_Handle XX_IpcInitSession(char destAddr[XX_IPC_MAX_ADDR_NAME_LENGTH],
-+ char srcAddr[XX_IPC_MAX_ADDR_NAME_LENGTH]);
-+
-+/**************************************************************************//**
-+ @Function XX_IpcFreeSession
-+
-+ @Description This function is used for terminating an existing IPC session between a source endpoint
-+ and a destination endpoint.
-+
-+ The IPC service assumes that after this function is called, no messages shall be sent from
-+ the associated source endpoint to the associated destination endpoint.
-+
-+ @Param[in] h_Session - Abstract handle to the IPC session - the same handle as was originally
-+ returned by the XX_IpcInitSession() function.
-+
-+ @Return E_OK on success; Error code otherwise.
-+*//***************************************************************************/
-+t_Error XX_IpcFreeSession(t_Handle h_Session);
-+
-+/**************************************************************************//**
-+ @Function XX_IpcSendMessage
-+
-+ @Description IPC message send routine.
-+
-+ This function may be used by a source endpoint to send an IPC message to a destination
-+ endpoint. The source endpoint cannot send a message to the destination endpoint without
-+ first initiating a session with that destination endpoint via XX_IpcInitSession() routine.
-+
-+ The source endpoint must provide the buffer pointer and length of the outgoing message.
-+ Optionally, it may also provide a buffer for an expected reply. In the latter case, the
-+ transaction is not considered complete by the IPC service until the reply has been received.
-+ If the source endpoint does not provide a reply buffer, the transaction is considered
-+ complete after the message has been sent. The source endpoint must keep the message (and
-+ optional reply) buffers valid until the transaction is complete.
-+
-+ @par Non-blocking mode
-+
-+ The source endpoint may request a non-blocking send by providing a non-NULL pointer to a message
-+ completion callback function (f_Completion). Upon completion of the IPC transaction (consisting of a
-+ message and an optional reply), the IPC service invokes this callback routine to return the message
-+ buffer to the sender and to provide the received reply, if requested.
-+
-+ @par Blocking mode
-+
-+ The source endpoint may request a blocking send by setting f_Completion to NULL. The function is
-+ expected to block until the IPC transaction is complete - either the reply has been received or (if no reply
-+ was requested) the message has been sent.
-+
-+ @Param[in] h_Session - Abstract handle to the IPC session - the same handle as was originally
-+ returned by the XX_IpcInitSession() function.
-+ @Param[in] p_Msg - Pointer to message buffer to send.
-+ @Param[in] msgLength - Length (in bytes) of actual data in the message buffer.
-+ @Param[in] p_Reply - Pointer to reply buffer - if this buffer is not NULL, the IPC service
-+ fills this buffer with the received reply data;
-+ In blocking mode, the reply data must be valid when the function returns;
-+ In non-blocking mode, the reply data is valid when f_Completion is called;
-+ If this pointer is NULL, no reply is expected.
-+ @Param[in,out] p_ReplyLength - Pointer to reply length, which has a dual role in this function:
-+ [In] specifies the maximal length (in bytes) of the reply buffer pointed by
-+ p_Reply, and
-+ [Out] in non-blocking mode this value is updated by the IPC service to the
-+ actual reply length (in bytes).
-+ @Param[in] f_Completion - Pointer to a completion callback to be used in non-blocking send mode;
-+ The completion callback is invoked by the IPC service upon
-+ completion of the IPC transaction (consisting of a message and an optional
-+ reply);
-+ If this pointer is NULL, the function is expected to block until the IPC
-+ transaction is complete.
-+ @Param[in] h_Arg - Abstract handle to the sending module; passed unchanged to the f_Completion
-+ callback function as the first argument.
-+
-+ @Return E_OK on success; Error code otherwise.
-+*//***************************************************************************/
-+t_Error XX_IpcSendMessage(t_Handle h_Session,
-+ uint8_t *p_Msg,
-+ uint32_t msgLength,
-+ uint8_t *p_Reply,
-+ uint32_t *p_ReplyLength,
-+ t_IpcMsgCompletion *f_Completion,
-+ t_Handle h_Arg);
-+
-+
-+/** @} */ /* end of xx_ipc group */
-+/** @} */ /* end of xx_id group */
-+
-+
-+#endif /* __XX_EXT_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/ls1043_dflags.h
-@@ -0,0 +1,56 @@
-+/*
-+ * Copyright 2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#ifndef __dflags_h
-+#define __dflags_h
-+
-+
-+#define NCSW_LINUX
-+
-+#define LS1043
-+
-+#define DEBUG_ERRORS 1
-+
-+#if defined(DEBUG)
-+#define DEBUG_GLOBAL_LEVEL REPORT_LEVEL_INFO
-+
-+#define DEBUG_XX_MALLOC
-+#define DEBUG_MEM_LEAKS
-+
-+#else
-+#define DEBUG_GLOBAL_LEVEL REPORT_LEVEL_WARNING
-+#endif /* (DEBUG) */
-+
-+#define REPORT_EVENTS 1
-+#define EVENT_GLOBAL_LEVEL REPORT_LEVEL_MINOR
-+
-+#endif /* __dflags_h */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/ncsw_config.mk
-@@ -0,0 +1,53 @@
-+#
-+# Makefile config for the Freescale NetcommSW
-+#
-+NET_DPA = $(srctree)/drivers/net
-+DRV_DPA = $(srctree)/drivers/net/ethernet/freescale/sdk_dpaa
-+FMAN = $(srctree)/drivers/net/ethernet/freescale/sdk_fman
-+
-+ifeq ("$(CONFIG_FMAN_P3040_P4080_P5020)", "y")
-+ccflags-y +=-include $(FMAN)/p3040_4080_5020_dflags.h
-+endif
-+ifeq ("$(CONFIG_FMAN_P1023)", "y")
-+ccflags-y +=-include $(FMAN)/p1023_dflags.h
-+endif
-+ifdef CONFIG_FMAN_V3H
-+ccflags-y +=-include $(FMAN)/fmanv3h_dflags.h
-+endif
-+ifdef CONFIG_FMAN_V3L
-+ccflags-y +=-include $(FMAN)/fmanv3l_dflags.h
-+endif
-+ifdef CONFIG_FMAN_ARM
-+ccflags-y +=-include $(FMAN)/ls1043_dflags.h
-+endif
-+
-+ccflags-y += -I$(DRV_DPA)/
-+ccflags-y += -I$(FMAN)/inc
-+ccflags-y += -I$(FMAN)/inc/cores
-+ccflags-y += -I$(FMAN)/inc/etc
-+ccflags-y += -I$(FMAN)/inc/Peripherals
-+ccflags-y += -I$(FMAN)/inc/flib
-+
-+ifeq ("$(CONFIG_FMAN_P3040_P4080_P5020)", "y")
-+ccflags-y += -I$(FMAN)/inc/integrations/P3040_P4080_P5020
-+endif
-+ifeq ("$(CONFIG_FMAN_P1023)", "y")
-+ccflags-y += -I$(FMAN)/inc/integrations/P1023
-+endif
-+ifdef CONFIG_FMAN_V3H
-+ccflags-y += -I$(FMAN)/inc/integrations/FMANV3H
-+endif
-+ifdef CONFIG_FMAN_V3L
-+ccflags-y += -I$(FMAN)/inc/integrations/FMANV3L
-+endif
-+ifdef CONFIG_FMAN_ARM
-+ccflags-y += -I$(FMAN)/inc/integrations/LS1043
-+endif
-+
-+ccflags-y += -I$(FMAN)/src/inc
-+ccflags-y += -I$(FMAN)/src/inc/system
-+ccflags-y += -I$(FMAN)/src/inc/wrapper
-+ccflags-y += -I$(FMAN)/src/inc/xx
-+ccflags-y += -I$(srctree)/include/uapi/linux/fmd
-+ccflags-y += -I$(srctree)/include/uapi/linux/fmd/Peripherals
-+ccflags-y += -I$(srctree)/include/uapi/linux/fmd/integrations
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/p1023_dflags.h
-@@ -0,0 +1,65 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#ifndef __dflags_h
-+#define __dflags_h
-+
-+
-+#define NCSW_LINUX
-+#if 0
-+#define DEBUG
-+#endif
-+
-+#define P1023
-+#define NCSW_PPC_CORE
-+
-+#define DEBUG_ERRORS 1
-+
-+#if defined(DEBUG)
-+#define DEBUG_GLOBAL_LEVEL REPORT_LEVEL_INFO
-+
-+#define DEBUG_XX_MALLOC
-+#define DEBUG_MEM_LEAKS
-+
-+#else
-+#define DEBUG_GLOBAL_LEVEL REPORT_LEVEL_WARNING
-+#endif /* (DEBUG) */
-+
-+#define REPORT_EVENTS 1
-+#define EVENT_GLOBAL_LEVEL REPORT_LEVEL_MINOR
-+
-+#ifdef CONFIG_P4080_SIM
-+#error "Do not define CONFIG_P4080_SIM..."
-+#endif
-+
-+
-+#endif /* __dflags_h */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/p3040_4080_5020_dflags.h
-@@ -0,0 +1,62 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#ifndef __dflags_h
-+#define __dflags_h
-+
-+
-+#define NCSW_LINUX
-+
-+#define P4080
-+#define NCSW_PPC_CORE
-+
-+#define DEBUG_ERRORS 1
-+
-+#if defined(DEBUG)
-+#define DEBUG_GLOBAL_LEVEL REPORT_LEVEL_INFO
-+
-+#define DEBUG_XX_MALLOC
-+#define DEBUG_MEM_LEAKS
-+
-+#else
-+#define DEBUG_GLOBAL_LEVEL REPORT_LEVEL_WARNING
-+#endif /* (DEBUG) */
-+
-+#define REPORT_EVENTS 1
-+#define EVENT_GLOBAL_LEVEL REPORT_LEVEL_MINOR
-+
-+#ifdef CONFIG_P4080_SIM
-+#define SIMULATOR
-+#endif /* CONFIG_P4080_SIM */
-+
-+
-+#endif /* __dflags_h */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/src/Makefile
-@@ -0,0 +1,11 @@
-+#
-+# Makefile for the Freescale Ethernet controllers
-+#
-+ccflags-y += -DVERSION=\"\"
-+#
-+#Include netcomm SW specific definitions
-+include $(srctree)/drivers/net/ethernet/freescale/sdk_fman/ncsw_config.mk
-+#
-+obj-y += system/
-+obj-y += wrapper/
-+obj-y += xx/
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/src/inc/system/sys_ext.h
-@@ -0,0 +1,118 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#ifndef __SYS_EXT_H
-+#define __SYS_EXT_H
-+
-+#include "std_ext.h"
-+
-+
-+/**************************************************************************//**
-+ @Group sys_grp System Interfaces
-+
-+ @Description Linux system programming interfaces.
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Group sys_gen_grp System General Interface
-+
-+ @Description General definitions, structures and routines of the linux
-+ system programming interface.
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Collection Macros for Advanced Configuration Requests
-+ @{
-+*//***************************************************************************/
-+#define SYS_MAX_ADV_CONFIG_ARGS 4
-+ /**< Maximum number of arguments in
-+ an advanced configuration entry */
-+/* @} */
-+
-+/**************************************************************************//**
-+ @Description System Object Advanced Configuration Entry
-+
-+ This structure represents a single request for an advanced
-+ configuration call on the initialized object. An array of such
-+ requests may be contained in the settings structure of the
-+ corresponding object.
-+
-+ The maximum number of arguments is limited to #SYS_MAX_ADV_CONFIG_ARGS.
-+*//***************************************************************************/
-+typedef struct t_SysObjectAdvConfigEntry
-+{
-+ void *p_Function; /**< Pointer to advanced configuration routine */
-+
-+ uintptr_t args[SYS_MAX_ADV_CONFIG_ARGS];
-+ /**< Array of arguments for the specified routine;
-+ All arguments should be casted to uint32_t. */
-+} t_SysObjectAdvConfigEntry;
-+
-+
-+/** @} */ /* end of sys_gen_grp */
-+/** @} */ /* end of sys_grp */
-+
-+#define NCSW_PARAMS(_num, _params) ADV_CONFIG_PARAMS_##_num _params
-+
-+#define ADV_CONFIG_PARAMS_1(_type) \
-+ , (_type)p_Entry->args[0]
-+
-+#define SET_ADV_CONFIG_ARGS_1(_arg0) \
-+ p_Entry->args[0] = (uintptr_t )(_arg0); \
-+
-+#define ARGS(_num, _params) SET_ADV_CONFIG_ARGS_##_num _params
-+
-+#define ADD_ADV_CONFIG_START(_p_Entries, _maxEntries) \
-+ { \
-+ t_SysObjectAdvConfigEntry *p_Entry; \
-+ t_SysObjectAdvConfigEntry *p_Entrys = (_p_Entries); \
-+ int i=0, max = (_maxEntries); \
-+
-+#define ADD_ADV_CONFIG_END \
-+ }
-+
-+#define ADV_CONFIG_CHECK_START(_p_Entry) \
-+ { \
-+ t_SysObjectAdvConfigEntry *p_Entry = _p_Entry; \
-+ t_Error errCode; \
-+
-+#define ADV_CONFIG_CHECK(_handle, _func, _params) \
-+ if (p_Entry->p_Function == _func) \
-+ { \
-+ errCode = _func(_handle _params); \
-+ } else
-+
-+#endif /* __SYS_EXT_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/src/inc/system/sys_io_ext.h
-@@ -0,0 +1,46 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#ifndef __SYS_IO_EXT_H
-+#define __SYS_IO_EXT_H
-+
-+#include "std_ext.h"
-+#include "error_ext.h"
-+
-+
-+t_Error SYS_RegisterIoMap (uint64_t virtAddr, uint64_t physAddr, uint32_t size);
-+t_Error SYS_UnregisterIoMap (uint64_t virtAddr);
-+uint64_t SYS_PhysToVirt (uint64_t addr);
-+uint64_t SYS_VirtToPhys (uint64_t addr);
-+
-+
-+#endif /* __SYS_IO_EXT_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/src/inc/types_linux.h
-@@ -0,0 +1,208 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#ifndef __TYPES_LINUX_H__
-+#define __TYPES_LINUX_H__
-+
-+#include <linux/version.h>
-+
-+#if defined(CONFIG_MODVERSIONS) && !defined(MODVERSIONS)
-+#define MODVERSIONS
-+#endif
-+#ifdef MODVERSIONS
-+#include <config/modversions.h>
-+#endif /* MODVERSIONS */
-+
-+#include <linux/kernel.h>
-+#include <linux/types.h>
-+#include <asm/io.h>
-+#include <linux/delay.h>
-+
-+#if LINUX_VERSION_CODE < KERNEL_VERSION(2,6,11)
-+ #error "This kernel is probably not supported!!!"
-+#elif (!((LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,19)) || \
-+ (LINUX_VERSION_CODE <= KERNEL_VERSION(2,6,27)) || \
-+ (LINUX_VERSION_CODE <= KERNEL_VERSION(2,6,30))))
-+ #warning "This kernel is probably not supported!!! You may need to add some fixes."
-+#endif /* LINUX_VERSION_CODE */
-+
-+
-+typedef float float_t; /* Single precision floating point */
-+typedef double double_t; /* Double precision floating point */
-+
-+
-+#define _Packed
-+#define _PackedType __attribute__ ((packed))
-+
-+typedef phys_addr_t physAddress_t;
-+
-+#define UINT8_MAX 0xFF
-+#define UINT8_MIN 0
-+#define UINT16_MAX 0xFFFF
-+#define UINT16_MIN 0
-+#define UINT32_MAX 0xFFFFFFFF
-+#define UINT32_MIN 0
-+#define UINT64_MAX 0xFFFFFFFFFFFFFFFFLL
-+#define UINT64_MIN 0
-+#define INT8_MAX 0x7F
-+#define INT8_MIN 0x80
-+#define INT16_MAX 0x7FFF
-+#define INT16_MIN 0x8000
-+#define INT32_MAX 0x7FFFFFFF
-+#define INT32_MIN 0x80000000
-+#define INT64_MAX 0x7FFFFFFFFFFFFFFFLL
-+#define INT64_MIN 0x8000000000000000LL
-+
-+#define ON 1
-+#define OFF 0
-+
-+#define FALSE false
-+#define TRUE true
-+
-+
-+/************************/
-+/* memory access macros */
-+/************************/
-+#ifdef CONFIG_FMAN_ARM
-+#define in_be16(a) __be16_to_cpu(__raw_readw(a))
-+#define in_be32(a) __be32_to_cpu(__raw_readl(a))
-+#define out_be16(a, v) __raw_writew(__cpu_to_be16(v), a)
-+#define out_be32(a, v) __raw_writel(__cpu_to_be32(v), a)
-+#endif
-+
-+#define GET_UINT8(arg) *(volatile uint8_t *)(&(arg))
-+#define GET_UINT16(arg) in_be16(&(arg))//*(volatile uint16_t*)(&(arg))
-+#define GET_UINT32(arg) in_be32(&(arg))//*(volatile uint32_t*)(&(arg))
-+#define GET_UINT64(arg) *(volatile uint64_t*)(&(arg))
-+
-+#ifdef VERBOSE_WRITE
-+void XX_Print(char *str, ...);
-+#define WRITE_UINT8(arg, data) \
-+ do { XX_Print("ADDR: 0x%08x, VAL: 0x%02x\r\n", (uint32_t)&(arg), (data)); *(volatile uint8_t *)(&(arg)) = (data); } while (0)
-+#define WRITE_UINT16(arg, data) \
-+ do { XX_Print("ADDR: 0x%08x, VAL: 0x%04x\r\n", (uint32_t)&(arg), (data)); out_be16(&(arg), data); /* *(volatile uint16_t*)(&(arg)) = (data);*/ } while (0)
-+#define WRITE_UINT32(arg, data) \
-+ do { XX_Print("ADDR: 0x%08x, VAL: 0x%08x\r\n", (uint32_t)&(arg), (data)); out_be32(&(arg), data); /* *(volatile uint32_t*)(&(arg)) = (data);*/ } while (0)
-+#define WRITE_UINT64(arg, data) \
-+ do { XX_Print("ADDR: 0x%08x, VAL: 0x%016llx\r\n", (uint32_t)&(arg), (data)); *(volatile uint64_t*)(&(arg)) = (data); } while (0)
-+
-+#else /* not VERBOSE_WRITE */
-+#define WRITE_UINT8(arg, data) *(volatile uint8_t *)(&(arg)) = (data)
-+#define WRITE_UINT16(arg, data) out_be16(&(arg), data)//*(volatile uint16_t*)(&(arg)) = (data)
-+#define WRITE_UINT32(arg, data) out_be32(&(arg), data)//*(volatile unsigned int *)(&(arg)) = (data)
-+#define WRITE_UINT64(arg, data) *(volatile uint64_t*)(&(arg)) = (data)
-+#endif /* not VERBOSE_WRITE */
-+
-+
-+/*****************************************************************************/
-+/* General stuff */
-+/*****************************************************************************/
-+#ifdef ARRAY_SIZE
-+#undef ARRAY_SIZE
-+#endif /* ARRAY_SIZE */
-+
-+#ifdef MAJOR
-+#undef MAJOR
-+#endif /* MAJOR */
-+
-+#ifdef MINOR
-+#undef MINOR
-+#endif /* MINOR */
-+
-+#ifdef QE_SIZEOF_BD
-+#undef QE_SIZEOF_BD
-+#endif /* QE_SIZEOF_BD */
-+
-+#ifdef BD_BUFFER_CLEAR
-+#undef BD_BUFFER_CLEAR
-+#endif /* BD_BUFFER_CLEAR */
-+
-+#ifdef BD_BUFFER
-+#undef BD_BUFFER
-+#endif /* BD_BUFFER */
-+
-+#ifdef BD_STATUS_AND_LENGTH_SET
-+#undef BD_STATUS_AND_LENGTH_SET
-+#endif /* BD_STATUS_AND_LENGTH_SET */
-+
-+#ifdef BD_STATUS_AND_LENGTH
-+#undef BD_STATUS_AND_LENGTH
-+#endif /* BD_STATUS_AND_LENGTH */
-+
-+#ifdef BD_BUFFER_ARG
-+#undef BD_BUFFER_ARG
-+#endif /* BD_BUFFER_ARG */
-+
-+#ifdef BD_GET_NEXT
-+#undef BD_GET_NEXT
-+#endif /* BD_GET_NEXT */
-+
-+#ifdef QE_SDEBCR_BA_MASK
-+#undef QE_SDEBCR_BA_MASK
-+#endif /* QE_SDEBCR_BA_MASK */
-+
-+#ifdef BD_BUFFER_SET
-+#undef BD_BUFFER_SET
-+#endif /* BD_BUFFER_SET */
-+
-+#ifdef UPGCR_PROTOCOL
-+#undef UPGCR_PROTOCOL
-+#endif /* UPGCR_PROTOCOL */
-+
-+#ifdef UPGCR_TMS
-+#undef UPGCR_TMS
-+#endif /* UPGCR_TMS */
-+
-+#ifdef UPGCR_RMS
-+#undef UPGCR_RMS
-+#endif /* UPGCR_RMS */
-+
-+#ifdef UPGCR_ADDR
-+#undef UPGCR_ADDR
-+#endif /* UPGCR_ADDR */
-+
-+#ifdef UPGCR_DIAG
-+#undef UPGCR_DIAG
-+#endif /* UPGCR_DIAG */
-+
-+#ifdef NCSW_PARAMS
-+#undef NCSW_PARAMS
-+#endif /* NCSW_PARAMS */
-+
-+#ifdef NO_IRQ
-+#undef NO_IRQ
-+#endif /* NO_IRQ */
-+
-+#define PRINT_LINE XX_Print("%s:\n %s [%d]\n",__FILE__,__FUNCTION__,__LINE__);
-+
-+
-+#endif /* __TYPES_LINUX_H__ */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/src/inc/wrapper/fsl_fman_test.h
-@@ -0,0 +1,84 @@
-+/* Copyright (c) 2008-2011 Freescale Semiconductor, Inc.
-+ * All rights reserved.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+/******************************************************************************
-+ @File fsl_fman_test.h
-+
-+ @Description
-+*//***************************************************************************/
-+
-+#ifndef __FSL_FMAN_TEST_H
-+#define __FSL_FMAN_TEST_H
-+
-+#include <linux/types.h>
-+#include <linux/smp.h> /* raw_smp_processor_id() */
-+
-+//#define FMT_K_DBG
-+//#define FMT_K_DBG_RUNTIME
-+
-+#define _fmt_prk(stage, format, arg...) \
-+ printk(stage "fmt (cpu:%u): " format, raw_smp_processor_id(), ##arg)
-+
-+#define _fmt_inf(format, arg...) _fmt_prk(KERN_INFO, format, ##arg)
-+#define _fmt_wrn(format, arg...) _fmt_prk(KERN_WARNING, format, ##arg)
-+#define _fmt_err(format, arg...) _fmt_prk(KERN_ERR, format, ##arg)
-+
-+/* there are two macros for debugging: for runtime and generic.
-+ * Helps when the runtime functions are not targeted for debugging,
-+ * thus all the unnecessary information will be skipped.
-+ */
-+/* used for generic debugging */
-+#if defined(FMT_K_DBG)
-+ #define _fmt_dbg(format, arg...) \
-+ printk("fmt [%s:%u](cpu:%u) - " format, \
-+ __func__, __LINE__, raw_smp_processor_id(), ##arg)
-+#else
-+# define _fmt_dbg(arg...)
-+#endif
-+
-+/* used for debugging runtime functions */
-+#if defined(FMT_K_DBG_RUNTIME)
-+ #define _fmt_dbgr(format, arg...) \
-+ printk("fmt [%s:%u](cpu:%u) - " format, \
-+ __func__, __LINE__, raw_smp_processor_id(), ##arg)
-+#else
-+# define _fmt_dbgr(arg...)
-+#endif
-+
-+#define FMT_RX_ERR_Q 0xffffffff
-+#define FMT_RX_DFLT_Q 0xfffffffe
-+#define FMT_TX_ERR_Q 0xfffffffd
-+#define FMT_TX_CONF_Q 0xfffffffc
-+
-+#define FMAN_TEST_MAX_TX_FQS 8
-+
-+#endif /* __FSL_FMAN_TEST_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/src/inc/wrapper/lnxwrp_exp_sym.h
-@@ -0,0 +1,130 @@
-+/* Copyright (c) 2008-2012 Freescale Semiconductor, Inc.
-+ * All rights reserved.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+/*
-+ @File lnxwrp_exp_sym.h
-+ @Description FMan exported routines
-+*/
-+
-+#ifndef __LNXWRP_EXP_SYM_H
-+#define __LNXWRP_EXP_SYM_H
-+
-+#include "fm_port_ext.h"
-+#include "fm_pcd_ext.h"
-+#include "fm_mac_ext.h"
-+
-+
-+/* FMAN Port exported routines */
-+EXPORT_SYMBOL(FM_PORT_Disable);
-+EXPORT_SYMBOL(FM_PORT_Enable);
-+EXPORT_SYMBOL(FM_PORT_SetPCD);
-+EXPORT_SYMBOL(FM_PORT_DeletePCD);
-+
-+/* Runtime PCD exported routines */
-+EXPORT_SYMBOL(FM_PCD_Enable);
-+EXPORT_SYMBOL(FM_PCD_Disable);
-+EXPORT_SYMBOL(FM_PCD_GetCounter);
-+EXPORT_SYMBOL(FM_PCD_PrsLoadSw);
-+EXPORT_SYMBOL(FM_PCD_KgSetDfltValue);
-+EXPORT_SYMBOL(FM_PCD_KgSetAdditionalDataAfterParsing);
-+EXPORT_SYMBOL(FM_PCD_SetException);
-+EXPORT_SYMBOL(FM_PCD_ModifyCounter);
-+EXPORT_SYMBOL(FM_PCD_SetPlcrStatistics);
-+EXPORT_SYMBOL(FM_PCD_SetPrsStatistics);
-+EXPORT_SYMBOL(FM_PCD_ForceIntr);
-+EXPORT_SYMBOL(FM_PCD_HcTxConf);
-+
-+EXPORT_SYMBOL(FM_PCD_NetEnvCharacteristicsSet);
-+EXPORT_SYMBOL(FM_PCD_NetEnvCharacteristicsDelete);
-+EXPORT_SYMBOL(FM_PCD_KgSchemeSet);
-+EXPORT_SYMBOL(FM_PCD_KgSchemeDelete);
-+EXPORT_SYMBOL(FM_PCD_KgSchemeGetCounter);
-+EXPORT_SYMBOL(FM_PCD_KgSchemeSetCounter);
-+EXPORT_SYMBOL(FM_PCD_CcRootBuild);
-+EXPORT_SYMBOL(FM_PCD_CcRootDelete);
-+EXPORT_SYMBOL(FM_PCD_MatchTableSet);
-+EXPORT_SYMBOL(FM_PCD_MatchTableDelete);
-+EXPORT_SYMBOL(FM_PCD_CcRootModifyNextEngine);
-+EXPORT_SYMBOL(FM_PCD_MatchTableModifyNextEngine);
-+EXPORT_SYMBOL(FM_PCD_MatchTableFindNModifyNextEngine);
-+EXPORT_SYMBOL(FM_PCD_MatchTableModifyMissNextEngine);
-+EXPORT_SYMBOL(FM_PCD_MatchTableRemoveKey);
-+EXPORT_SYMBOL(FM_PCD_MatchTableFindNRemoveKey);
-+EXPORT_SYMBOL(FM_PCD_MatchTableAddKey);
-+EXPORT_SYMBOL(FM_PCD_MatchTableModifyKeyAndNextEngine);
-+EXPORT_SYMBOL(FM_PCD_MatchTableFindNModifyKeyAndNextEngine);
-+EXPORT_SYMBOL(FM_PCD_MatchTableModifyKey);
-+EXPORT_SYMBOL(FM_PCD_MatchTableFindNModifyKey);
-+EXPORT_SYMBOL(FM_PCD_MatchTableGetIndexedHashBucket);
-+EXPORT_SYMBOL(FM_PCD_MatchTableGetNextEngine);
-+EXPORT_SYMBOL(FM_PCD_MatchTableGetKeyCounter);
-+EXPORT_SYMBOL(FM_PCD_MatchTableGetKeyStatistics);
-+EXPORT_SYMBOL(FM_PCD_MatchTableFindNGetKeyStatistics);
-+EXPORT_SYMBOL(FM_PCD_MatchTableGetMissStatistics);
-+EXPORT_SYMBOL(FM_PCD_HashTableGetMissStatistics);
-+EXPORT_SYMBOL(FM_PCD_HashTableSet);
-+EXPORT_SYMBOL(FM_PCD_HashTableDelete);
-+EXPORT_SYMBOL(FM_PCD_HashTableAddKey);
-+EXPORT_SYMBOL(FM_PCD_HashTableRemoveKey);
-+EXPORT_SYMBOL(FM_PCD_HashTableModifyNextEngine);
-+EXPORT_SYMBOL(FM_PCD_HashTableModifyMissNextEngine);
-+EXPORT_SYMBOL(FM_PCD_HashTableGetMissNextEngine);
-+EXPORT_SYMBOL(FM_PCD_HashTableFindNGetKeyStatistics);
-+EXPORT_SYMBOL(FM_PCD_PlcrProfileSet);
-+EXPORT_SYMBOL(FM_PCD_PlcrProfileDelete);
-+EXPORT_SYMBOL(FM_PCD_PlcrProfileGetCounter);
-+EXPORT_SYMBOL(FM_PCD_PlcrProfileSetCounter);
-+EXPORT_SYMBOL(FM_PCD_ManipNodeSet);
-+EXPORT_SYMBOL(FM_PCD_ManipNodeDelete);
-+EXPORT_SYMBOL(FM_PCD_ManipGetStatistics);
-+EXPORT_SYMBOL(FM_PCD_ManipNodeReplace);
-+#if (DPAA_VERSION >= 11)
-+EXPORT_SYMBOL(FM_PCD_FrmReplicSetGroup);
-+EXPORT_SYMBOL(FM_PCD_FrmReplicDeleteGroup);
-+EXPORT_SYMBOL(FM_PCD_FrmReplicAddMember);
-+EXPORT_SYMBOL(FM_PCD_FrmReplicRemoveMember);
-+#endif /* DPAA_VERSION >= 11 */
-+
-+#ifdef FM_CAPWAP_SUPPORT
-+EXPORT_SYMBOL(FM_PCD_StatisticsSetNode);
-+#endif /* FM_CAPWAP_SUPPORT */
-+
-+EXPORT_SYMBOL(FM_PCD_SetAdvancedOffloadSupport);
-+
-+/* FMAN MAC exported routines */
-+EXPORT_SYMBOL(FM_MAC_GetStatistics);
-+
-+EXPORT_SYMBOL(FM_MAC_GetFrameSizeCounters);
-+
-+EXPORT_SYMBOL(FM_GetSpecialOperationCoding);
-+
-+#endif /* __LNXWRP_EXP_SYM_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/src/inc/wrapper/lnxwrp_fm_ext.h
-@@ -0,0 +1,163 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+/******************************************************************************
-+ @File lnxwrp_fm_ext.h
-+
-+ @Description TODO
-+*//***************************************************************************/
-+
-+#ifndef __LNXWRP_FM_EXT_H
-+#define __LNXWRP_FM_EXT_H
-+
-+#include "std_ext.h"
-+#include "sys_ext.h"
-+#include "fm_ext.h"
-+#include "fm_muram_ext.h"
-+#include "fm_pcd_ext.h"
-+#include "fm_port_ext.h"
-+#include "fm_mac_ext.h"
-+#include "fm_rtc_ext.h"
-+
-+
-+/**************************************************************************//**
-+ @Group FM_LnxKern_grp Frame Manager Linux wrapper API
-+
-+ @Description FM API functions, definitions and enums.
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Group FM_LnxKern_init_grp Initialization Unit
-+
-+ @Description Initialization Unit
-+
-+ Initialization Flow:
-+ Initialization of the FM Module will be carried out by the Linux
-+ kernel according to the following sequence:
-+ a. Calling the initialization routine with no parameters.
-+ b. The driver will register to the Device-Tree.
-+ c. The Linux Device-Tree will initiate a call to the driver for
-+ initialization.
-+ d. The driver will read the appropriate information from the Device-Tree
-+ e. [Optional] Calling the advance initialization routines to change
-+ driver's defaults.
-+ f. Initialization of the device will be automatically upon using it.
-+
-+ @{
-+*//***************************************************************************/
-+
-+typedef struct t_WrpFmDevSettings
-+{
-+ t_FmParams param;
-+ t_SysObjectAdvConfigEntry *advConfig;
-+} t_WrpFmDevSettings;
-+
-+typedef struct t_WrpFmPcdDevSettings
-+{
-+ t_FmPcdParams param;
-+ t_SysObjectAdvConfigEntry *advConfig;
-+} t_WrpFmPcdDevSettings;
-+
-+typedef struct t_WrpFmPortDevSettings
-+{
-+ bool frag_enabled;
-+ t_FmPortParams param;
-+ t_SysObjectAdvConfigEntry *advConfig;
-+} t_WrpFmPortDevSettings;
-+
-+typedef struct t_WrpFmMacDevSettings
-+{
-+ t_FmMacParams param;
-+ t_SysObjectAdvConfigEntry *advConfig;
-+} t_WrpFmMacDevSettings;
-+
-+
-+/**************************************************************************//**
-+ @Function LNXWRP_FM_Init
-+
-+ @Description Initialize the FM linux wrapper.
-+
-+ @Return A handle (descriptor) of the newly created FM Linux wrapper
-+ structure.
-+*//***************************************************************************/
-+t_Handle LNXWRP_FM_Init(void);
-+
-+/**************************************************************************//**
-+ @Function LNXWRP_FM_Free
-+
-+ @Description Free the FM linux wrapper.
-+
-+ @Param[in] h_LnxWrpFm - A handle to the FM linux wrapper.
-+
-+ @Return E_OK on success; Error code otherwise.
-+*//***************************************************************************/
-+t_Error LNXWRP_FM_Free(t_Handle h_LnxWrpFm);
-+
-+/**************************************************************************//**
-+ @Function LNXWRP_FM_GetMacHandle
-+
-+ @Description Get the FM-MAC LLD handle from the FM linux wrapper.
-+
-+ @Param[in] h_LnxWrpFm - A handle to the FM linux wrapper.
-+ @Param[in] fmId - Index of the FM device to get the MAC handle from.
-+ @Param[in] macId - Index of the mac handle.
-+
-+ @Return A handle of the LLD compressor.
-+*//***************************************************************************/
-+t_Handle LNXWRP_FM_GetMacHandle(t_Handle h_LnxWrpFm, uint8_t fmId, uint8_t macId);
-+
-+#ifdef CONFIG_FSL_SDK_FMAN_TEST
-+t_Handle LNXWRP_FM_TEST_Init(void);
-+t_Error LNXWRP_FM_TEST_Free(t_Handle h_FmTestLnxWrp);
-+#endif /* CONFIG_FSL_SDK_FMAN_TEST */
-+
-+/** @} */ /* end of FM_LnxKern_init_grp group */
-+
-+
-+/**************************************************************************//**
-+ @Group FM_LnxKern_ctrl_grp Control Unit
-+
-+ @Description Control Unit
-+
-+ TODO
-+ @{
-+*//***************************************************************************/
-+
-+#include "lnxwrp_fsl_fman.h"
-+
-+/** @} */ /* end of FM_LnxKern_ctrl_grp group */
-+/** @} */ /* end of FM_LnxKern_grp group */
-+
-+
-+#endif /* __LNXWRP_FM_EXT_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/src/inc/wrapper/lnxwrp_fsl_fman.h
-@@ -0,0 +1,921 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+/******************************************************************************
-+ @File lnxwrp_fsl_fman.h
-+
-+ @Description Linux internal kernel API
-+*//***************************************************************************/
-+
-+#ifndef __LNXWRP_FSL_FMAN_H
-+#define __LNXWRP_FSL_FMAN_H
-+
-+#include <linux/types.h>
-+#include <linux/device.h> /* struct device */
-+#include <linux/fsl_qman.h> /* struct qman_fq */
-+#include "dpaa_integration_ext.h"
-+#include "fm_port_ext.h"
-+#include "fm_mac_ext.h"
-+#include "fm_macsec_ext.h"
-+#include "fm_rtc_ext.h"
-+
-+/**************************************************************************//**
-+ @Group FM_LnxKern_grp Frame Manager Linux wrapper API
-+
-+ @Description FM API functions, definitions and enums.
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Group FM_LnxKern_ctrl_grp Control Unit
-+
-+ @Description Control Unit
-+
-+ Internal Kernel Control Unit API
-+ @{
-+*//***************************************************************************/
-+
-+/*****************************************************************************/
-+/* Internal Linux kernel routines */
-+/*****************************************************************************/
-+
-+/**************************************************************************//**
-+ @Description MACSEC Exceptions wrapper
-+*//***************************************************************************/
-+typedef enum fm_macsec_exception {
-+ SINGLE_BIT_ECC = e_FM_MACSEC_EX_SINGLE_BIT_ECC,
-+ MULTI_BIT_ECC = e_FM_MACSEC_EX_MULTI_BIT_ECC
-+} fm_macsec_exception;
-+
-+/**************************************************************************//**
-+ @Description Unknown sci frame treatment wrapper
-+*//***************************************************************************/
-+typedef enum fm_macsec_unknown_sci_frame_treatment {
-+ SCI_DISCARD_BOTH = e_FM_MACSEC_UNKNOWN_SCI_FRAME_TREATMENT_DISCARD_BOTH,
-+ SCI_DISCARD_UNCTRL_DELIVER_DISCARD_CTRL = \
-+ e_FM_MACSEC_UNKNOWN_SCI_FRAME_TREATMENT_DISCARD_UNCONTROLLED_DELIVER_OR_DISCARD_CONTROLLED,
-+ SCI_DELIVER_UNCTRL_DISCARD_CTRL = \
-+ e_FM_MACSEC_UNKNOWN_SCI_FRAME_TREATMENT_DELIVER_UNCONTROLLED_DISCARD_CONTROLLED,
-+ SCI_DELIVER_DISCARD_UNCTRL_DELIVER_DISCARD_CTRL = \
-+ e_FM_MACSEC_UNKNOWN_SCI_FRAME_TREATMENT_DELIVER_OR_DISCARD_UNCONTROLLED_DELIVER_OR_DISCARD_CONTROLLED
-+} fm_macsec_unknown_sci_frame_treatment;
-+
-+/**************************************************************************//**
-+ @Description Untag frame treatment wrapper
-+*//***************************************************************************/
-+typedef enum fm_macsec_untag_frame_treatment {
-+ UNTAG_DELIVER_UNCTRL_DISCARD_CTRL = \
-+ e_FM_MACSEC_UNTAG_FRAME_TREATMENT_DELIVER_UNCONTROLLED_DISCARD_CONTROLLED,
-+ UNTAG_DISCARD_BOTH = e_FM_MACSEC_UNTAG_FRAME_TREATMENT_DISCARD_BOTH,
-+ UNTAG_DISCARD_UNCTRL_DELIVER_CTRL_UNMODIFIED = \
-+ e_FM_MACSEC_UNTAG_FRAME_TREATMENT_DISCARD_UNCONTROLLED_DELIVER_CONTROLLED_UNMODIFIED
-+} fm_macsec_untag_frame_treatment;
-+
-+/**************************************************************************//**
-+@Description MACSEC SECY Cipher Suite wrapper
-+*//***************************************************************************/
-+typedef enum fm_macsec_secy_cipher_suite {
-+ SECY_GCM_AES_128 = e_FM_MACSEC_SECY_GCM_AES_128, /**< GCM-AES-128 */
-+#if (DPAA_VERSION >= 11)
-+ SECY_GCM_AES_256 = e_FM_MACSEC_SECY_GCM_AES_256 /**< GCM-AES-256 */
-+#endif /* (DPAA_VERSION >= 11) */
-+} fm_macsec_secy_cipher_suite;
-+
-+/**************************************************************************//**
-+ @Description MACSEC SECY Exceptions wrapper
-+*//***************************************************************************/
-+typedef enum fm_macsec_secy_exception {
-+ SECY_EX_FRAME_DISCARDED = e_FM_MACSEC_SECY_EX_FRAME_DISCARDED
-+} fm_macsec_secy_exception;
-+
-+/**************************************************************************//**
-+ @Description MACSEC SECY Events wrapper
-+*//***************************************************************************/
-+typedef enum fm_macsec_secy_event {
-+ SECY_EV_NEXT_PN = e_FM_MACSEC_SECY_EV_NEXT_PN
-+} fm_macsec_secy_event;
-+
-+/**************************************************************************//**
-+ @Description Valid frame behaviors wrapper
-+*//***************************************************************************/
-+typedef enum fm_macsec_valid_frame_behavior {
-+ VALID_FRAME_BEHAVIOR_DISABLE = e_FM_MACSEC_VALID_FRAME_BEHAVIOR_DISABLE,
-+ VALID_FRAME_BEHAVIOR_CHECK = e_FM_MACSEC_VALID_FRAME_BEHAVIOR_CHECK,
-+ VALID_FRAME_BEHAVIOR_STRICT = e_FM_MACSEC_VALID_FRAME_BEHAVIOR_STRICT
-+} fm_macsec_valid_frame_behavior;
-+
-+/**************************************************************************//**
-+ @Description SCI insertion modes wrapper
-+*//***************************************************************************/
-+typedef enum fm_macsec_sci_insertion_mode {
-+ SCI_INSERTION_MODE_EXPLICIT_SECTAG = \
-+ e_FM_MACSEC_SCI_INSERTION_MODE_EXPLICIT_SECTAG,
-+ SCI_INSERTION_MODE_EXPLICIT_MAC_SA = \
-+ e_FM_MACSEC_SCI_INSERTION_MODE_EXPLICIT_MAC_SA,
-+ SCI_INSERTION_MODE_IMPLICT_PTP = e_FM_MACSEC_SCI_INSERTION_MODE_IMPLICT_PTP
-+} fm_macsec_sci_insertion_mode;
-+
-+typedef macsecSAKey_t macsec_sa_key_t;
-+typedef macsecSCI_t macsec_sci_t;
-+typedef macsecAN_t macsec_an_t;
-+typedef t_Handle handle_t;
-+
-+/**************************************************************************//**
-+ @Function fm_macsec_secy_exception_callback wrapper
-+ @Description Exceptions user callback routine, will be called upon an
-+ exception passing the exception identification.
-+ @Param[in] app_h A handle to an application layer object; This handle
-+ will be passed by the driver upon calling this callback.
-+ @Param[in] exception The exception.
-+*//***************************************************************************/
-+typedef void (fm_macsec_secy_exception_callback) (handle_t app_h,
-+ fm_macsec_secy_exception exception);
-+
-+/**************************************************************************//**
-+ @Function fm_macsec_secy_event_callback wrapper
-+ @Description Events user callback routine, will be called upon an
-+ event passing the event identification.
-+ @Param[in] app_h A handle to an application layer object; This handle
-+ will be passed by the driver upon calling this callback.
-+ @Param[in] event The event.
-+*//***************************************************************************/
-+typedef void (fm_macsec_secy_event_callback) (handle_t app_h,
-+ fm_macsec_secy_event event);
-+
-+/**************************************************************************//**
-+ @Function fm_macsec_exception_callback wrapper
-+ @Description Exceptions user callback routine, will be called upon an
-+ exception passing the exception identification.
-+ @Param[in] app_h A handle to an application layer object; This handle
-+ will be passed by the driver upon calling this callback.
-+ @Param[in] exception The exception.
-+*//***************************************************************************/
-+typedef void (fm_macsec_exception_callback) (handle_t app_h,
-+ fm_macsec_exception exception);
-+
-+/**************************************************************************//**
-+ @Description MACSEC SecY SC Params wrapper
-+*//***************************************************************************/
-+struct fm_macsec_secy_sc_params {
-+ macsec_sci_t sci;
-+ fm_macsec_secy_cipher_suite cipher_suite;
-+};
-+
-+/**************************************************************************//**
-+ @Description FM MACSEC SecY config input wrapper
-+*//***************************************************************************/
-+struct fm_macsec_secy_params {
-+ handle_t fm_macsec_h;
-+ struct fm_macsec_secy_sc_params tx_sc_params;
-+ uint32_t num_receive_channels;
-+ fm_macsec_secy_exception_callback *exception_f;
-+ fm_macsec_secy_event_callback *event_f;
-+ handle_t app_h;
-+};
-+
-+/**************************************************************************//**
-+ @Description FM MACSEC config input wrapper
-+*//***************************************************************************/
-+struct fm_macsec_params {
-+ handle_t fm_h;
-+ bool guest_mode;
-+
-+ union {
-+ struct {
-+ uint8_t fm_mac_id;
-+ } guest_params;
-+
-+ struct {
-+ uintptr_t base_addr;
-+ handle_t fm_mac_h;
-+ fm_macsec_exception_callback *exception_f;
-+ handle_t app_h;
-+ } non_guest_params;
-+ };
-+
-+};
-+
-+/**************************************************************************//**
-+ @Description FM device opaque structure used for type checking
-+*//***************************************************************************/
-+struct fm;
-+
-+/**************************************************************************//**
-+ @Description FM MAC device opaque structure used for type checking
-+*//***************************************************************************/
-+struct fm_mac_dev;
-+
-+/**************************************************************************//**
-+ @Description FM MACSEC device opaque structure used for type checking
-+*//***************************************************************************/
-+struct fm_macsec_dev;
-+struct fm_macsec_secy_dev;
-+
-+/**************************************************************************//**
-+ @Description A structure ..,
-+*//***************************************************************************/
-+struct fm_port;
-+
-+typedef int (*alloc_pcd_fqids)(struct device *dev, uint32_t num,
-+ uint8_t alignment, uint32_t *base_fqid);
-+
-+typedef int (*free_pcd_fqids)(struct device *dev, uint32_t base_fqid);
-+
-+struct fm_port_pcd_param {
-+ alloc_pcd_fqids cba;
-+ free_pcd_fqids cbf;
-+ struct device *dev;
-+};
-+
-+/**************************************************************************//**
-+ @Description A structure of information about each of the external
-+ buffer pools used by the port,
-+*//***************************************************************************/
-+struct fm_port_pool_param {
-+ uint8_t id; /**< External buffer pool id */
-+ uint16_t size; /**< External buffer pool buffer size */
-+};
-+
-+/**************************************************************************//**
-+ @Description structure for additional port parameters
-+*//***************************************************************************/
-+struct fm_port_params {
-+ uint32_t errq; /**< Error Queue Id. */
-+ uint32_t defq; /**< For Tx and HC - Default Confirmation queue,
-+ 0 means no Tx conf for processed frames.
-+ For Rx and OP - default Rx queue. */
-+ uint8_t num_pools; /**< Number of pools use by this port */
-+ struct fm_port_pool_param pool_param[FM_PORT_MAX_NUM_OF_EXT_POOLS];
-+ /**< Parameters for each pool */
-+ uint16_t priv_data_size; /**< Area that user may save for his own
-+ need (E.g. save the SKB) */
-+ bool parse_results; /**< Put the parser-results in the Rx/Tx buffer */
-+ bool hash_results; /**< Put the hash-results in the Rx/Tx buffer */
-+ bool time_stamp; /**< Put the time-stamp in the Rx/Tx buffer */
-+ bool frag_enable; /**< Fragmentation support, for OP only */
-+ uint16_t data_align; /**< value for selecting a data alignment (must be a power of 2);
-+ if write optimization is used, must be >= 16. */
-+ uint8_t manip_extra_space; /**< Maximum extra size needed (insertion-size minus removal-size);
-+ Note that this field impacts the size of the buffer-prefix
-+ (i.e. it pushes the data offset); */
-+};
-+
-+/**************************************************************************//**
-+ @Function fm_bind
-+
-+ @Description Bind to a specific FM device.
-+
-+ @Param[in] fm_dev - the OF handle of the FM device.
-+
-+ @Return A handle of the FM device.
-+
-+ @Cautions Allowed only after the port was created.
-+*//***************************************************************************/
-+struct fm *fm_bind(struct device *fm_dev);
-+
-+/**************************************************************************//**
-+ @Function fm_unbind
-+
-+ @Description Un-bind from a specific FM device.
-+
-+ @Param[in] fm - A handle of the FM device.
-+
-+ @Cautions Allowed only after the port was created.
-+*//***************************************************************************/
-+void fm_unbind(struct fm *fm);
-+
-+void *fm_get_handle(struct fm *fm);
-+void *fm_get_rtc_handle(struct fm *fm);
-+struct resource *fm_get_mem_region(struct fm *fm);
-+
-+/**************************************************************************//**
-+ @Function fm_port_bind
-+
-+ @Description Bind to a specific FM-port device (may be Rx or Tx port).
-+
-+ @Param[in] fm_port_dev - the OF handle of the FM port device.
-+
-+ @Return A handle of the FM port device.
-+
-+ @Cautions Allowed only after the port was created.
-+*//***************************************************************************/
-+struct fm_port *fm_port_bind(struct device *fm_port_dev);
-+
-+/**************************************************************************//**
-+ @Function fm_port_unbind
-+
-+ @Description Un-bind from a specific FM-port device (may be Rx or Tx port).
-+
-+ @Param[in] port - A handle of the FM port device.
-+
-+ @Cautions Allowed only after the port was created.
-+*//***************************************************************************/
-+void fm_port_unbind(struct fm_port *port);
-+
-+/**************************************************************************//**
-+ @Function fm_set_rx_port_params
-+
-+ @Description Configure parameters for a specific Rx FM-port device.
-+
-+ @Param[in] port - A handle of the FM port device.
-+ @Param[in] params - Rx port parameters
-+
-+ @Cautions Allowed only after the port is binded.
-+*//***************************************************************************/
-+void fm_set_rx_port_params(struct fm_port *port,
-+ struct fm_port_params *params);
-+
-+/**************************************************************************//**
-+ @Function fm_port_pcd_bind
-+
-+ @Description Bind as a listener on a port PCD.
-+
-+ @Param[in] port - A handle of the FM port device.
-+ @Param[in] params - PCD port parameters
-+
-+ @Cautions Allowed only after the port is binded.
-+*//***************************************************************************/
-+void fm_port_pcd_bind (struct fm_port *port, struct fm_port_pcd_param *params);
-+
-+/**************************************************************************//**
-+ @Function fm_port_get_buff_layout_ext_params
-+
-+ @Description Get data_align and manip_extra_space from the device tree
-+ chosen node if applied.
-+ This function will only update these two parameters.
-+ When this port has no such parameters in the device tree
-+ values will be set to 0.
-+
-+ @Param[in] port - A handle of the FM port device.
-+ @Param[in] params - PCD port parameters
-+
-+ @Cautions Allowed only after the port is binded.
-+*//***************************************************************************/
-+void fm_port_get_buff_layout_ext_params(struct fm_port *port, struct fm_port_params *params);
-+
-+/**************************************************************************//**
-+ @Function fm_get_tx_port_channel
-+
-+ @Description Get qman-channel number for this Tx port.
-+
-+ @Param[in] port - A handle of the FM port device.
-+
-+ @Return qman-channel number for this Tx port.
-+
-+ @Cautions Allowed only after the port is binded.
-+*//***************************************************************************/
-+uint16_t fm_get_tx_port_channel(struct fm_port *port);
-+
-+/**************************************************************************//**
-+ @Function fm_set_tx_port_params
-+
-+ @Description Configure parameters for a specific Tx FM-port device
-+
-+ @Param[in] port - A handle of the FM port device.
-+ @Param[in] params - Tx port parameters
-+
-+ @Cautions Allowed only after the port is binded.
-+*//***************************************************************************/
-+void fm_set_tx_port_params(struct fm_port *port, struct fm_port_params *params);
-+
-+
-+/**************************************************************************//**
-+ @Function fm_mac_set_handle
-+
-+ @Description Set mac handle
-+
-+ @Param[in] h_lnx_wrp_fm_dev - A handle of the LnxWrp FM device.
-+ @Param[in] h_fm_mac - A handle of the LnxWrp FM MAC device.
-+ @Param[in] mac_id - MAC id.
-+*//***************************************************************************/
-+void fm_mac_set_handle(t_Handle h_lnx_wrp_fm_dev, t_Handle h_fm_mac,
-+ int mac_id);
-+
-+/**************************************************************************//**
-+ @Function fm_port_enable
-+
-+ @Description Enable specific FM-port device (may be Rx or Tx port).
-+
-+ @Param[in] port - A handle of the FM port device.
-+
-+ @Cautions Allowed only after the port is initialized.
-+*//***************************************************************************/
-+int fm_port_enable(struct fm_port *port);
-+
-+/**************************************************************************//**
-+ @Function fm_port_disable
-+
-+ @Description Disable specific FM-port device (may be Rx or Tx port).
-+
-+ @Param[in] port - A handle of the FM port device.
-+
-+ @Cautions Allowed only after the port is initialized.
-+*//***************************************************************************/
-+int fm_port_disable(struct fm_port *port);
-+
-+void *fm_port_get_handle(const struct fm_port *port);
-+
-+u64 *fm_port_get_buffer_time_stamp(const struct fm_port *port,
-+ const void *data);
-+
-+/**************************************************************************//**
-+ @Function fm_port_get_base_address
-+
-+ @Description Get base address of this port. Useful for accessing
-+ port-specific registers (i.e., not common ones).
-+
-+ @Param[in] port - A handle of the FM port device.
-+
-+ @Param[out] base_addr - The port's base addr (virtual address).
-+*//***************************************************************************/
-+void fm_port_get_base_addr(const struct fm_port *port, uint64_t *base_addr);
-+
-+/**************************************************************************//**
-+ @Function fm_mutex_lock
-+
-+ @Description Lock function required before any FMD/LLD call.
-+*//***************************************************************************/
-+void fm_mutex_lock(void);
-+
-+/**************************************************************************//**
-+ @Function fm_mutex_unlock
-+
-+ @Description Unlock function required after any FMD/LLD call.
-+*//***************************************************************************/
-+void fm_mutex_unlock(void);
-+
-+/**************************************************************************//**
-+ @Function fm_get_max_frm
-+
-+ @Description Get the maximum frame size
-+*//***************************************************************************/
-+int fm_get_max_frm(void);
-+
-+/**************************************************************************//**
-+ @Function fm_get_rx_extra_headroom
-+
-+ @Description Get the extra headroom size
-+*//***************************************************************************/
-+int fm_get_rx_extra_headroom(void);
-+
-+/**************************************************************************//**
-+@Function fm_port_set_rate_limit
-+
-+@Description Configure Shaper parameter on FM-port device (Tx port).
-+
-+@Param[in] port - A handle of the FM port device.
-+@Param[in] max_burst_size - Value of maximum burst size allowed.
-+@Param[in] rate_limit - The required rate value.
-+
-+@Cautions Allowed only after the port is initialized.
-+*//***************************************************************************/
-+int fm_port_set_rate_limit(struct fm_port *port,
-+ uint16_t max_burst_size,
-+ uint32_t rate_limit);
-+/**************************************************************************//**
-+@Function fm_port_set_rate_limit
-+
-+@Description Delete Shaper configuration on FM-port device (Tx port).
-+
-+@Param[in] port - A handle of the FM port device.
-+
-+@Cautions Allowed only after the port is initialized.
-+*//***************************************************************************/
-+int fm_port_del_rate_limit(struct fm_port *port);
-+
-+struct auto_res_tables_sizes
-+{
-+ uint16_t max_num_of_arp_entries;
-+ uint16_t max_num_of_echo_ipv4_entries;
-+ uint16_t max_num_of_ndp_entries;
-+ uint16_t max_num_of_echo_ipv6_entries;
-+ uint16_t max_num_of_snmp_ipv4_entries;
-+ uint16_t max_num_of_snmp_ipv6_entries;
-+ uint16_t max_num_of_snmp_oid_entries;
-+ uint16_t max_num_of_snmp_char; /* total amount of character needed
-+ for the snmp table */
-+ uint16_t max_num_of_ip_prot_filtering;
-+ uint16_t max_num_of_tcp_port_filtering;
-+ uint16_t max_num_of_udp_port_filtering;
-+};
-+/* ARP */
-+struct auto_res_arp_entry
-+{
-+ uint32_t ip_address;
-+ uint8_t mac[6];
-+ bool is_vlan;
-+ uint16_t vid;
-+};
-+struct auto_res_arp_info
-+{
-+ uint8_t table_size;
-+ struct auto_res_arp_entry *auto_res_table;
-+ bool enable_conflict_detection; /* when TRUE
-+ Conflict Detection will be checked and wake the host if
-+ needed */
-+};
-+
-+/* NDP */
-+struct auto_res_ndp_entry
-+{
-+ uint32_t ip_address[4];
-+ uint8_t mac[6];
-+ bool is_vlan;
-+ uint16_t vid;
-+};
-+struct auto_res_ndp_info
-+{
-+ uint32_t multicast_group;
-+ uint8_t table_size_assigned;
-+ struct auto_res_ndp_entry *auto_res_table_assigned; /* This list
-+ refer to solicitation IP addresses. Note that all IP adresses
-+ must be from the same multicast group. This will be checked and
-+ if not operation will fail. */
-+ uint8_t table_size_tmp;
-+ struct auto_res_ndp_entry *auto_res_table_tmp; /* This list
-+ refer to temp IP addresses. Note that all temp IP adresses must
-+ be from the same multicast group. This will be checked and if
-+ not operation will fail. */
-+
-+ bool enable_conflict_detection; /* when TRUE
-+ Conflict Detection will be checked and wake the host if
-+ needed */
-+};
-+
-+/* ICMP ECHO */
-+struct auto_res_echo_ipv4_info
-+{
-+ uint8_t table_size;
-+ struct auto_res_arp_entry *auto_res_table;
-+};
-+
-+struct auto_res_echo_ipv6_info
-+{
-+ uint8_t table_size;
-+ struct auto_res_ndp_entry *auto_res_table;
-+};
-+
-+/* SNMP */
-+struct auto_res_snmp_entry
-+{
-+ uint16_t oidSize;
-+ uint8_t *oidVal; /* only the oid string */
-+ uint16_t resSize;
-+ uint8_t *resVal; /* resVal will be the entire reply,
-+ i.e. "Type|Length|Value" */
-+};
-+
-+/**************************************************************************//**
-+ @Description Deep Sleep Auto Response SNMP IPv4 Addresses Table Entry
-+ Refer to the FMan Controller spec for more details.
-+*//***************************************************************************/
-+struct auto_res_snmp_ipv4addr_tbl_entry
-+{
-+ uint32_t ipv4addr; /*!< 32 bit IPv4 Address. */
-+ bool is_vlan;
-+ uint16_t vid; /*!< 12 bits VLAN ID. The 4 left-most bits should be cleared */
-+ /*!< This field should be 0x0000 for an entry with no VLAN tag or a null VLAN ID. */
-+};
-+
-+/**************************************************************************//**
-+ @Description Deep Sleep Auto Response SNMP IPv6 Addresses Table Entry
-+ Refer to the FMan Controller spec for more details.
-+*//***************************************************************************/
-+struct auto_res_snmp_ipv6addr_tbl_entry
-+{
-+ uint32_t ipv6Addr[4]; /*!< 4 * 32 bit IPv6 Address. */
-+ bool isVlan;
-+ uint16_t vid; /*!< 12 bits VLAN ID. The 4 left-most bits should be cleared */
-+ /*!< This field should be 0x0000 for an entry with no VLAN tag or a null VLAN ID. */
-+};
-+
-+struct auto_res_snmp_info
-+{
-+ uint16_t control; /**< Control bits [0-15]. */
-+ uint16_t max_snmp_msg_length; /**< Maximal allowed SNMP message length. */
-+ uint16_t num_ipv4_addresses; /**< Number of entries in IPv4 addresses table. */
-+ uint16_t num_ipv6_addresses; /**< Number of entries in IPv6 addresses table. */
-+ struct auto_res_snmp_ipv4addr_tbl_entry *ipv4addr_tbl; /**< Pointer to IPv4 addresses table. */
-+ struct auto_res_snmp_ipv6addr_tbl_entry *ipv6addr_tbl; /**< Pointer to IPv6 addresses table. */
-+ char *community_read_write_string;
-+ char *community_read_only_string;
-+ struct auto_res_snmp_entry *oid_table;
-+ uint32_t oid_table_size;
-+ uint32_t *statistics;
-+};
-+
-+/* Filtering */
-+struct auto_res_port_filtering_entry
-+{
-+ uint16_t src_port;
-+ uint16_t dst_port;
-+ uint16_t src_port_mask;
-+ uint16_t dst_port_mask;
-+};
-+struct auto_res_filtering_info
-+{
-+ /* IP protocol filtering parameters */
-+ uint8_t ip_prot_table_size;
-+ uint8_t *ip_prot_table_ptr;
-+ bool ip_prot_pass_on_hit; /* when TRUE, miss in the table will
-+ cause the packet to be droped, hit will pass the packet to
-+ UDP/TCP filters if needed and if not to the classification
-+ tree. If the classification tree will pass the packet to a
-+ queue it will cause a wake interupt. When FALSE it the other
-+ way around. */
-+ /* UDP port filtering parameters */
-+ uint8_t udp_ports_table_size;
-+ struct auto_res_port_filtering_entry *udp_ports_table_ptr;
-+ bool udp_port_pass_on_hit; /* when TRUE, miss in the table will
-+ cause the packet to be droped, hit will pass the packet to
-+ classification tree. If the classification tree will pass the
-+ packet to a queue it will cause a wake interupt. When FALSE it
-+ the other way around. */
-+ /* TCP port filtering parameters */
-+ uint16_t tcp_flags_mask;
-+ uint8_t tcp_ports_table_size;
-+ struct auto_res_port_filtering_entry *tcp_ports_table_ptr;
-+ bool tcp_port_pass_on_hit; /* when TRUE, miss in the table will
-+ cause the packet to be droped, hit will pass the packet to
-+ classification tree. If the classification tree will pass the
-+ packet to a queue it will cause a wake interupt. When FALSE it
-+ the other way around. */
-+};
-+
-+struct auto_res_port_params
-+{
-+ t_Handle h_FmPortTx;
-+ struct auto_res_arp_info *p_auto_res_arp_info;
-+ struct auto_res_echo_ipv4_info *p_auto_res_echo_ipv4_info;
-+ struct auto_res_ndp_info *p_auto_res_ndp_info;
-+ struct auto_res_echo_ipv6_info *p_auto_res_echo_ipv6_info;
-+ struct auto_res_snmp_info *p_auto_res_snmp_info;
-+ struct auto_res_filtering_info *p_auto_res_filtering_info;
-+};
-+
-+struct auto_res_port_stats
-+{
-+ uint32_t arp_ar_cnt;
-+ uint32_t echo_icmpv4_ar_cnt;
-+ uint32_t ndp_ar_cnt;
-+ uint32_t echo_icmpv6_ar_cnt;
-+};
-+
-+int fm_port_config_autores_for_deepsleep_support(struct fm_port *port,
-+ struct auto_res_tables_sizes *params);
-+
-+int fm_port_enter_autores_for_deepsleep(struct fm_port *port,
-+ struct auto_res_port_params *params);
-+
-+void fm_port_exit_auto_res_for_deep_sleep(struct fm_port *port_rx,
-+ struct fm_port *port_tx);
-+
-+bool fm_port_is_in_auto_res_mode(struct fm_port *port);
-+
-+struct auto_res_tables_sizes *fm_port_get_autores_maxsize(
-+ struct fm_port *port);
-+
-+int fm_port_get_autores_stats(struct fm_port *port, struct auto_res_port_stats
-+ *stats);
-+
-+int fm_port_resume(struct fm_port *port);
-+
-+int fm_port_suspend(struct fm_port *port);
-+
-+#ifdef CONFIG_FMAN_PFC
-+/**************************************************************************//**
-+@Function fm_port_set_pfc_priorities_mapping_to_qman_wq
-+
-+@Description Associate a QMan Work Queue with a PFC priority on this
-+ FM-port device (Tx port).
-+
-+@Param[in] port - A handle of the FM port device.
-+
-+@Param[in] prio - The PFC priority.
-+
-+@Param[in] wq - The Work Queue associated with the PFC priority.
-+
-+@Cautions Allowed only after the port is initialized.
-+*//***************************************************************************/
-+int fm_port_set_pfc_priorities_mapping_to_qman_wq(struct fm_port *port,
-+ uint8_t prio, uint8_t wq);
-+#endif
-+
-+/**************************************************************************//**
-+@Function fm_mac_set_exception
-+
-+@Description Set MAC exception state.
-+
-+@Param[in] fm_mac_dev - A handle of the FM MAC device.
-+@Param[in] exception - FM MAC exception type.
-+@Param[in] enable - new state.
-+
-+*//***************************************************************************/
-+int fm_mac_set_exception(struct fm_mac_dev *fm_mac_dev,
-+ e_FmMacExceptions exception, bool enable);
-+
-+int fm_mac_free(struct fm_mac_dev *fm_mac_dev);
-+
-+struct fm_mac_dev *fm_mac_config(t_FmMacParams *params);
-+
-+int fm_mac_config_max_frame_length(struct fm_mac_dev *fm_mac_dev,
-+ int len);
-+
-+int fm_mac_config_pad_and_crc(struct fm_mac_dev *fm_mac_dev, bool enable);
-+
-+int fm_mac_config_half_duplex(struct fm_mac_dev *fm_mac_dev, bool enable);
-+
-+int fm_mac_config_reset_on_init(struct fm_mac_dev *fm_mac_dev, bool enable);
-+
-+int fm_mac_init(struct fm_mac_dev *fm_mac_dev);
-+
-+int fm_mac_get_version(struct fm_mac_dev *fm_mac_dev, uint32_t *version);
-+
-+int fm_mac_enable(struct fm_mac_dev *fm_mac_dev);
-+
-+int fm_mac_disable(struct fm_mac_dev *fm_mac_dev);
-+
-+int fm_mac_resume(struct fm_mac_dev *fm_mac_dev);
-+
-+int fm_mac_set_promiscuous(struct fm_mac_dev *fm_mac_dev,
-+ bool enable);
-+
-+int fm_mac_remove_hash_mac_addr(struct fm_mac_dev *fm_mac_dev,
-+ t_EnetAddr *mac_addr);
-+
-+int fm_mac_add_hash_mac_addr(struct fm_mac_dev *fm_mac_dev,
-+ t_EnetAddr *mac_addr);
-+
-+int fm_mac_modify_mac_addr(struct fm_mac_dev *fm_mac_dev,
-+ uint8_t *addr);
-+
-+int fm_mac_adjust_link(struct fm_mac_dev *fm_mac_dev,
-+ bool link, int speed, bool duplex);
-+
-+int fm_mac_enable_1588_time_stamp(struct fm_mac_dev *fm_mac_dev);
-+
-+int fm_mac_disable_1588_time_stamp(struct fm_mac_dev *fm_mac_dev);
-+
-+int fm_mac_set_rx_pause_frames(
-+ struct fm_mac_dev *fm_mac_dev, bool en);
-+
-+int fm_mac_set_tx_pause_frames(struct fm_mac_dev *fm_mac_dev,
-+ bool en);
-+
-+int fm_rtc_enable(struct fm *fm_dev);
-+
-+int fm_rtc_disable(struct fm *fm_dev);
-+
-+int fm_rtc_get_cnt(struct fm *fm_dev, uint64_t *ts);
-+
-+int fm_rtc_set_cnt(struct fm *fm_dev, uint64_t ts);
-+
-+int fm_rtc_get_drift(struct fm *fm_dev, uint32_t *drift);
-+
-+int fm_rtc_set_drift(struct fm *fm_dev, uint32_t drift);
-+
-+int fm_rtc_set_alarm(struct fm *fm_dev, uint32_t id,
-+ uint64_t time);
-+
-+int fm_rtc_set_fiper(struct fm *fm_dev, uint32_t id,
-+ uint64_t fiper);
-+
-+int fm_mac_set_wol(struct fm_port *port, struct fm_mac_dev *fm_mac_dev,
-+ bool en);
-+
-+/**************************************************************************//**
-+@Function fm_macsec_set_exception
-+
-+@Description Set MACSEC exception state.
-+
-+@Param[in] fm_macsec_dev - A handle of the FM MACSEC device.
-+@Param[in] exception - FM MACSEC exception type.
-+@Param[in] enable - new state.
-+
-+*//***************************************************************************/
-+
-+int fm_macsec_set_exception(struct fm_macsec_dev *fm_macsec_dev,
-+ fm_macsec_exception exception, bool enable);
-+int fm_macsec_free(struct fm_macsec_dev *fm_macsec_dev);
-+struct fm_macsec_dev *fm_macsec_config(struct fm_macsec_params *fm_params);
-+int fm_macsec_init(struct fm_macsec_dev *fm_macsec_dev);
-+int fm_macsec_config_unknown_sci_frame_treatment(struct fm_macsec_dev
-+ *fm_macsec_dev,
-+ fm_macsec_unknown_sci_frame_treatment treat_mode);
-+int fm_macsec_config_invalid_tags_frame_treatment(struct fm_macsec_dev *fm_macsec_dev,
-+ bool deliver_uncontrolled);
-+int fm_macsec_config_kay_frame_treatment(struct fm_macsec_dev *fm_macsec_dev,
-+ bool discard_uncontrolled);
-+int fm_macsec_config_untag_frame_treatment(struct fm_macsec_dev *fm_macsec_dev,
-+ fm_macsec_untag_frame_treatment treat_mode);
-+int fm_macsec_config_pn_exhaustion_threshold(struct fm_macsec_dev *fm_macsec_dev,
-+ uint32_t pnExhThr);
-+int fm_macsec_config_keys_unreadable(struct fm_macsec_dev *fm_macsec_dev);
-+int fm_macsec_config_sectag_without_sci(struct fm_macsec_dev *fm_macsec_dev);
-+int fm_macsec_config_exception(struct fm_macsec_dev *fm_macsec_dev,
-+ fm_macsec_exception exception, bool enable);
-+int fm_macsec_get_revision(struct fm_macsec_dev *fm_macsec_dev,
-+ int *macsec_revision);
-+int fm_macsec_enable(struct fm_macsec_dev *fm_macsec_dev);
-+int fm_macsec_disable(struct fm_macsec_dev *fm_macsec_dev);
-+
-+
-+int fm_macsec_secy_config_exception(struct fm_macsec_secy_dev *fm_macsec_secy_dev,
-+ fm_macsec_secy_exception exception,
-+ bool enable);
-+int fm_macsec_secy_free(struct fm_macsec_secy_dev *fm_macsec_secy_dev);
-+struct fm_macsec_secy_dev *fm_macsec_secy_config(struct fm_macsec_secy_params *secy_params);
-+int fm_macsec_secy_init(struct fm_macsec_secy_dev *fm_macsec_secy_dev);
-+int fm_macsec_secy_config_sci_insertion_mode(struct fm_macsec_secy_dev *fm_macsec_secy_dev,
-+ fm_macsec_sci_insertion_mode sci_insertion_mode);
-+int fm_macsec_secy_config_protect_frames(struct fm_macsec_secy_dev *fm_macsec_secy_dev,
-+ bool protect_frames);
-+int fm_macsec_secy_config_replay_window(struct fm_macsec_secy_dev *fm_macsec_secy_dev,
-+ bool replay_protect, uint32_t replay_window);
-+int fm_macsec_secy_config_validation_mode(struct fm_macsec_secy_dev *fm_macsec_secy_dev,
-+ fm_macsec_valid_frame_behavior validate_frames);
-+int fm_macsec_secy_config_confidentiality(struct fm_macsec_secy_dev *fm_macsec_secy_dev,
-+ bool confidentiality_enable,
-+ uint32_t confidentiality_offset);
-+int fm_macsec_secy_config_point_to_point(struct fm_macsec_secy_dev *fm_macsec_secy_dev);
-+int fm_macsec_secy_config_event(struct fm_macsec_secy_dev *fm_macsec_secy_dev,
-+ fm_macsec_secy_event event,
-+ bool enable);
-+struct rx_sc_dev *fm_macsec_secy_create_rxsc(struct fm_macsec_secy_dev *fm_macsec_secy_dev,
-+ struct fm_macsec_secy_sc_params *params);
-+int fm_macsec_secy_delete_rxsc(struct fm_macsec_secy_dev *fm_macsec_secy_dev,
-+ struct rx_sc_dev *sc);
-+int fm_macsec_secy_create_rx_sa(struct fm_macsec_secy_dev *fm_macsec_secy_dev,
-+ struct rx_sc_dev *sc, macsec_an_t an,
-+ uint32_t lowest_pn, macsec_sa_key_t key);
-+int fm_macsec_secy_delete_rx_sa(struct fm_macsec_secy_dev *fm_macsec_secy_dev,
-+ struct rx_sc_dev *sc, macsec_an_t an);
-+int fm_macsec_secy_rxsa_enable_receive(struct fm_macsec_secy_dev *fm_macsec_secy_dev,
-+ struct rx_sc_dev *sc,
-+ macsec_an_t an);
-+int fm_macsec_secy_rxsa_disable_receive(struct fm_macsec_secy_dev *fm_macsec_secy_dev,
-+ struct rx_sc_dev *sc,
-+ macsec_an_t an);
-+int fm_macsec_secy_rxsa_update_next_pn(struct fm_macsec_secy_dev *fm_macsec_secy_dev,
-+ struct rx_sc_dev *sc,
-+ macsec_an_t an, uint32_t updt_next_pn);
-+int fm_macsec_secy_rxsa_update_lowest_pn(struct fm_macsec_secy_dev *fm_macsec_secy_dev,
-+ struct rx_sc_dev *sc,
-+ macsec_an_t an, uint32_t updt_lowest_pn);
-+int fm_macsec_secy_rxsa_modify_key(struct fm_macsec_secy_dev *fm_macsec_secy_dev,
-+ struct rx_sc_dev *sc,
-+ macsec_an_t an, macsec_sa_key_t key);
-+int fm_macsec_secy_create_tx_sa(struct fm_macsec_secy_dev *fm_macsec_secy_dev,
-+ macsec_an_t an, macsec_sa_key_t key);
-+int fm_macsec_secy_delete_tx_sa(struct fm_macsec_secy_dev *fm_macsec_secy_dev,
-+ macsec_an_t an);
-+int fm_macsec_secy_txsa_modify_key(struct fm_macsec_secy_dev *fm_macsec_secy_dev,
-+ macsec_an_t next_active_an,
-+ macsec_sa_key_t key);
-+int fm_macsec_secy_txsa_set_active(struct fm_macsec_secy_dev *fm_macsec_secy_dev,
-+ macsec_an_t an);
-+int fm_macsec_secy_txsa_get_active(struct fm_macsec_secy_dev *fm_macsec_secy_dev,
-+ macsec_an_t *p_an);
-+int fm_macsec_secy_get_rxsc_phys_id(struct fm_macsec_secy_dev *fm_macsec_secy_dev,
-+ struct rx_sc_dev *sc, uint32_t *sc_phys_id);
-+int fm_macsec_secy_get_txsc_phys_id(struct fm_macsec_secy_dev *fm_macsec_secy_dev,
-+ uint32_t *sc_phys_id);
-+
-+/** @} */ /* end of FM_LnxKern_ctrl_grp group */
-+/** @} */ /* end of FM_LnxKern_grp group */
-+
-+/* default values for initializing PTP 1588 timer clock */
-+#define DPA_PTP_NOMINAL_FREQ_PERIOD_SHIFT 2 /* power of 2 for better performance */
-+#define DPA_PTP_NOMINAL_FREQ_PERIOD_NS (1 << DPA_PTP_NOMINAL_FREQ_PERIOD_SHIFT) /* 4ns,250MHz */
-+
-+#endif /* __LNXWRP_FSL_FMAN_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/src/inc/xx/xx.h
-@@ -0,0 +1,50 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#ifndef __XX_H
-+#define __XX_H
-+
-+#include "xx_ext.h"
-+
-+void * xx_Malloc(uint32_t n);
-+void xx_Free(void *p);
-+
-+void *xx_MallocSmart(uint32_t size, int memPartitionId, uint32_t align);
-+void xx_FreeSmart(void *p);
-+
-+/* never used: */
-+#define GetDeviceName(irq) ((char *)NULL)
-+
-+int GetDeviceIrqNum(int irq);
-+
-+
-+#endif /* __XX_H */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/src/system/Makefile
-@@ -0,0 +1,10 @@
-+#
-+# Makefile for the Freescale Ethernet controllers
-+#
-+ccflags-y += -DVERSION=\"\"
-+#
-+#Include netcomm SW specific definitions
-+include $(srctree)/drivers/net/ethernet/freescale/sdk_fman/ncsw_config.mk
-+#
-+
-+obj-y += sys_io.o
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/src/system/sys_io.c
-@@ -0,0 +1,171 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#include <linux/version.h>
-+
-+#if defined(CONFIG_MODVERSIONS) && !defined(MODVERSIONS)
-+#define MODVERSIONS
-+#endif
-+#ifdef MODVERSIONS
-+#if LINUX_VERSION_CODE < KERNEL_VERSION(2,6,0)
-+#include <linux/modversions.h>
-+#else
-+#include <config/modversions.h>
-+#endif /* LINUX_VERSION_CODE */
-+#endif /* MODVERSIONS */
-+
-+#include <linux/module.h>
-+#include <linux/kernel.h>
-+
-+#include <asm/io.h>
-+
-+#include "std_ext.h"
-+#include "error_ext.h"
-+#include "string_ext.h"
-+#include "list_ext.h"
-+#include "sys_io_ext.h"
-+
-+
-+#define __ERR_MODULE__ MODULE_UNKNOWN
-+
-+
-+typedef struct {
-+ uint64_t virtAddr;
-+ uint64_t physAddr;
-+ uint32_t size;
-+ t_List node;
-+} t_IoMap;
-+#define IOMAP_OBJECT(ptr) LIST_OBJECT(ptr, t_IoMap, node)
-+
-+LIST(mapsList);
-+
-+
-+static void EnqueueIoMap(t_IoMap *p_IoMap)
-+{
-+ uint32_t intFlags;
-+
-+ intFlags = XX_DisableAllIntr();
-+ LIST_AddToTail(&p_IoMap->node, &mapsList);
-+ XX_RestoreAllIntr(intFlags);
-+}
-+
-+static t_IoMap * FindIoMapByVirtAddr(uint64_t addr)
-+{
-+ t_IoMap *p_IoMap;
-+ t_List *p_Pos;
-+
-+ LIST_FOR_EACH(p_Pos, &mapsList)
-+ {
-+ p_IoMap = IOMAP_OBJECT(p_Pos);
-+ if ((addr >= p_IoMap->virtAddr) && (addr < p_IoMap->virtAddr+p_IoMap->size))
-+ return p_IoMap;
-+ }
-+
-+ return NULL;
-+}
-+
-+static t_IoMap * FindIoMapByPhysAddr(uint64_t addr)
-+{
-+ t_IoMap *p_IoMap;
-+ t_List *p_Pos;
-+
-+ LIST_FOR_EACH(p_Pos, &mapsList)
-+ {
-+ p_IoMap = IOMAP_OBJECT(p_Pos);
-+ if ((addr >= p_IoMap->physAddr) && (addr < p_IoMap->physAddr+p_IoMap->size))
-+ return p_IoMap;
-+ }
-+
-+ return NULL;
-+}
-+
-+t_Error SYS_RegisterIoMap (uint64_t virtAddr, uint64_t physAddr, uint32_t size)
-+{
-+ t_IoMap *p_IoMap;
-+
-+ p_IoMap = (t_IoMap*)XX_Malloc(sizeof(t_IoMap));
-+ if (!p_IoMap)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("message handler object!!!"));
-+ memset(p_IoMap, 0, sizeof(t_IoMap));
-+
-+ p_IoMap->virtAddr = virtAddr;
-+ p_IoMap->physAddr = physAddr;
-+ p_IoMap->size = size;
-+
-+ INIT_LIST(&p_IoMap->node);
-+ EnqueueIoMap(p_IoMap);
-+
-+ return E_OK;
-+}
-+
-+t_Error SYS_UnregisterIoMap (uint64_t virtAddr)
-+{
-+ t_IoMap *p_IoMap = FindIoMapByVirtAddr(virtAddr);
-+ if (!p_IoMap)
-+ RETURN_ERROR(MINOR, E_NO_DEVICE, ("message handler not found in list!!!"));
-+
-+ LIST_Del(&p_IoMap->node);
-+ XX_Free(p_IoMap);
-+
-+ return E_OK;
-+}
-+
-+uint64_t SYS_PhysToVirt(uint64_t addr)
-+{
-+ t_IoMap *p_IoMap = FindIoMapByPhysAddr(addr);
-+ if (p_IoMap)
-+ {
-+ /* This is optimization - put the latest in the list-head - like a cache */
-+ if (mapsList.p_Next != &p_IoMap->node)
-+ {
-+ uint32_t intFlags = XX_DisableAllIntr();
-+ LIST_DelAndInit(&p_IoMap->node);
-+ LIST_Add(&p_IoMap->node, &mapsList);
-+ XX_RestoreAllIntr(intFlags);
-+ }
-+ return (uint64_t)(addr - p_IoMap->physAddr + p_IoMap->virtAddr);
-+ }
-+ return PTR_TO_UINT(phys_to_virt((unsigned long)addr));
-+}
-+
-+uint64_t SYS_VirtToPhys(uint64_t addr)
-+{
-+ t_IoMap *p_IoMap;
-+
-+ if (addr == 0)
-+ return 0;
-+
-+ p_IoMap = FindIoMapByVirtAddr(addr);
-+ if (p_IoMap)
-+ return (uint64_t)(addr - p_IoMap->virtAddr + p_IoMap->physAddr);
-+ return (uint64_t)virt_to_phys(UINT_TO_PTR(addr));
-+}
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/src/wrapper/Makefile
-@@ -0,0 +1,19 @@
-+#
-+# Makefile for the Freescale Ethernet controllers
-+#
-+ccflags-y += -DVERSION=\"\"
-+#
-+#Include netcomm SW specific definitions
-+include $(srctree)/drivers/net/ethernet/freescale/sdk_fman/ncsw_config.mk
-+
-+NCSW_FM_INC = $(srctree)/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/inc
-+
-+ccflags-y += -I$(NCSW_FM_INC)
-+ccflags-y += -I$(NET_DPA)
-+
-+obj-y += fsl-ncsw-PFM.o
-+obj-$(CONFIG_FSL_SDK_FMAN_TEST) += fman_test.o
-+
-+fsl-ncsw-PFM-objs := lnxwrp_fm.o lnxwrp_fm_port.o lnxwrp_ioctls_fm.o \
-+ lnxwrp_sysfs.o lnxwrp_sysfs_fm.o lnxwrp_sysfs_fm_port.o
-+obj-$(CONFIG_COMPAT) += lnxwrp_ioctls_fm_compat.o
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/src/wrapper/fman_test.c
-@@ -0,0 +1,1665 @@
-+/* Copyright (c) 2008-2011 Freescale Semiconductor, Inc.
-+ * All rights reserved.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+/*
-+ @File fman_test.c
-+ @Authors Pistirica Sorin Andrei
-+ @Description FM Linux test environment
-+*/
-+
-+#include <linux/kernel.h>
-+#include <linux/module.h>
-+#include <linux/fs.h>
-+#include <linux/cdev.h>
-+#include <linux/device.h>
-+#include <linux/io.h>
-+#include <linux/ioport.h>
-+#include <linux/of_platform.h>
-+#include <linux/ip.h>
-+#include <linux/compat.h>
-+#include <linux/uaccess.h>
-+#include <linux/errno.h>
-+#include <linux/netdevice.h>
-+#include <linux/spinlock.h>
-+#include <linux/types.h>
-+#include <linux/fsl_qman.h>
-+#include <linux/fsl_bman.h>
-+
-+/* private headers */
-+#include "fm_ext.h"
-+#include "lnxwrp_fsl_fman.h"
-+#include "fm_port_ext.h"
-+#if (DPAA_VERSION == 11)
-+#include "../../Peripherals/FM/MAC/memac.h"
-+#endif
-+#include "fm_test_ioctls.h"
-+#include "fsl_fman_test.h"
-+
-+#include "dpaa_eth.h"
-+#include "dpaa_eth_common.h"
-+
-+#define FMT_FRM_WATERMARK 0xdeadbeefdeadbeeaLL
-+
-+struct fmt_frame_s {
-+ ioc_fmt_buff_desc_t buff;
-+ struct list_head list;
-+};
-+
-+struct fmt_fqs_s {
-+ struct qman_fq fq_base;
-+ bool init;
-+ struct fmt_port_s *fmt_port_priv;
-+};
-+
-+struct fmt_port_pcd_s {
-+ int num_queues;
-+ struct fmt_fqs_s *fmt_pcd_fqs;
-+ uint32_t fqid_base;
-+};
-+
-+/* char dev structure: fm test port */
-+struct fmt_port_s {
-+ bool valid;
-+ uint8_t id;
-+ ioc_fmt_port_type port_type;
-+ ioc_diag_mode diag;
-+ bool compat_test_type;
-+
-+ /* fm ports */
-+ /* ! for oh ports p_tx_fm_port_dev == p_rx_fm_port_dev &&
-+ * p_tx_port == p_rx_port */
-+ /* t_LnxWrpFmPortDev */
-+ struct fm_port *p_tx_port;
-+ /* t_LnxWrpFmPortDev->h_Dev: t_FmPort */
-+ void *p_tx_fm_port_dev;
-+ /* t_LnxWrpFmPortDev */
-+ struct fm_port *p_rx_port;
-+ /* t_LnxWrpFmPortDev->h_Dev: t_FmPort */
-+ void *p_rx_fm_port_dev;
-+
-+ void *p_mac_dev;
-+ uint64_t fm_phys_base_addr;
-+
-+ /* read/write queue manipulation */
-+ spinlock_t rx_q_lock;
-+ struct list_head rx_q;
-+
-+ /* tx queuee for injecting traffic */
-+ int num_of_tx_fqs;
-+ struct fmt_fqs_s p_tx_fqs[FMAN_TEST_MAX_TX_FQS];
-+
-+ /* pcd private queues manipulation */
-+ struct fmt_port_pcd_s fmt_port_pcd;
-+
-+ /* debugging stuff */
-+
-+#if defined(FMT_K_DBG) || defined(FMT_K_DBG_RUNTIME)
-+ atomic_t enqueue_to_qman_frm;
-+ atomic_t enqueue_to_rxq;
-+ atomic_t dequeue_from_rxq;
-+ atomic_t not_enqueue_to_rxq_wrong_frm;
-+#endif
-+
-+};
-+
-+/* The devices. */
-+struct fmt_s {
-+ int major;
-+ struct fmt_port_s ports[IOC_FMT_MAX_NUM_OF_PORTS];
-+ struct class *fmt_class;
-+};
-+
-+/* fm test structure */
-+static struct fmt_s fm_test;
-+
-+#if (DPAA_VERSION == 11)
-+struct mac_priv_s {
-+ t_Handle mac;
-+};
-+#endif
-+
-+#define DTSEC_BASE_ADDR 0x000e0000
-+#define DTSEC_MEM_RANGE 0x00002000
-+#define MAC_1G_MACCFG1 0x00000100
-+#define MAC_1G_LOOP_MASK 0x00000100
-+static int set_1gmac_loopback(
-+ struct fmt_port_s *fmt_port,
-+ bool en)
-+{
-+#if (DPAA_VERSION <= 10)
-+ uint32_t dtsec_idx = fmt_port->id; /* dtsec for which port */
-+ uint32_t dtsec_idx_off = dtsec_idx * DTSEC_MEM_RANGE;
-+ phys_addr_t maccfg1_hw;
-+ void *maccfg1_map;
-+ uint32_t maccfg1_val;
-+
-+ /* compute the maccfg1 register address */
-+ maccfg1_hw = fmt_port->fm_phys_base_addr +
-+ (phys_addr_t)(DTSEC_BASE_ADDR +
-+ dtsec_idx_off +
-+ MAC_1G_MACCFG1);
-+
-+ /* map register */
-+ maccfg1_map = ioremap(maccfg1_hw, sizeof(u32));
-+
-+ /* set register */
-+ maccfg1_val = in_be32(maccfg1_map);
-+ if (en)
-+ maccfg1_val |= MAC_1G_LOOP_MASK;
-+ else
-+ maccfg1_val &= ~MAC_1G_LOOP_MASK;
-+ out_be32(maccfg1_map, maccfg1_val);
-+
-+ /* unmap register */
-+ iounmap(maccfg1_map);
-+#else
-+ struct mac_device *mac_dev;
-+ struct mac_priv_s *priv;
-+ t_Memac *p_memac;
-+
-+ if (!fmt_port)
-+ return -EINVAL;
-+
-+ mac_dev = (struct mac_device *)fmt_port->p_mac_dev;
-+
-+ if (!mac_dev)
-+ return -EINVAL;
-+
-+ priv = macdev_priv(mac_dev);
-+
-+ if (!priv)
-+ return -EINVAL;
-+
-+ p_memac = priv->mac;
-+
-+ if (!p_memac)
-+ return -EINVAL;
-+
-+ memac_set_loopback(p_memac->p_MemMap, en);
-+#endif
-+ return 0;
-+}
-+
-+/* TODO: re-write this function */
-+static int set_10gmac_int_loopback(
-+ struct fmt_port_s *fmt_port,
-+ bool en)
-+{
-+#ifndef FM_10G_MAC_NO_CTRL_LOOPBACK
-+#define FM_10GMAC0_OFFSET 0x000f0000
-+#define FM_10GMAC_CMD_CONF_CTRL_OFFSET 0x8
-+#define CMD_CFG_LOOPBACK_EN 0x00000400
-+
-+ uint64_t base_addr, reg_addr;
-+ uint32_t tmp_val;
-+
-+ base_addr = fmt_port->fm_phys_base_addr + (FM_10GMAC0_OFFSET +
-+ ((fmt_port->id-FM_MAX_NUM_OF_1G_RX_PORTS)*0x2000));
-+
-+ base_addr = PTR_TO_UINT(ioremap(base_addr, 0x1000));
-+
-+ reg_addr = base_addr + FM_10GMAC_CMD_CONF_CTRL_OFFSET;
-+ tmp_val = GET_UINT32(*((uint32_t *)UINT_TO_PTR(reg_addr)));
-+ if (en)
-+ tmp_val |= CMD_CFG_LOOPBACK_EN;
-+ else
-+ tmp_val &= ~CMD_CFG_LOOPBACK_EN;
-+ WRITE_UINT32(*((uint32_t *)UINT_TO_PTR(reg_addr)), tmp_val);
-+
-+ iounmap(UINT_TO_PTR(base_addr));
-+
-+ return 0;
-+#else
-+ _fmt_err("TGEC don't have internal-loopback.\n");
-+ return -EPERM;
-+#endif
-+}
-+
-+static int set_mac_int_loopback(struct fmt_port_s *fmt_port, bool en)
-+{
-+ int _err = 0;
-+
-+ switch (fmt_port->port_type) {
-+
-+ case e_IOC_FMT_PORT_T_RXTX:
-+ /* 1G port */
-+ if (fmt_port->id < FM_MAX_NUM_OF_1G_RX_PORTS)
-+ _err = set_1gmac_loopback(fmt_port, en);
-+ /* 10g port */
-+ else if ((fmt_port->id >= FM_MAX_NUM_OF_1G_RX_PORTS) &&
-+ (fmt_port->id < FM_MAX_NUM_OF_1G_RX_PORTS +
-+ FM_MAX_NUM_OF_10G_RX_PORTS)) {
-+
-+ _err = set_10gmac_int_loopback(fmt_port, en);
-+ } else
-+ _err = -EINVAL;
-+ break;
-+ /* op port does not have MAC (loopback mode) */
-+ case e_IOC_FMT_PORT_T_OP:
-+
-+ _err = 0;
-+ break;
-+ default:
-+
-+ _err = -EPERM;
-+ break;
-+ }
-+
-+ return _err;
-+}
-+
-+static void enqueue_fmt_frame(
-+ struct fmt_port_s *fmt_port,
-+ struct fmt_frame_s *p_fmt_frame)
-+{
-+ spinlock_t *rx_q_lock = NULL;
-+
-+ rx_q_lock = &fmt_port->rx_q_lock;
-+
-+ spin_lock(rx_q_lock);
-+ list_add_tail(&p_fmt_frame->list, &fmt_port->rx_q);
-+ spin_unlock(rx_q_lock);
-+
-+#if defined(FMT_K_DBG) || defined(FMT_K_DBG_RUNTIME)
-+ atomic_inc(&fmt_port->enqueue_to_rxq);
-+#endif
-+}
-+
-+static struct fmt_frame_s *dequeue_fmt_frame(
-+ struct fmt_port_s *fmt_port)
-+{
-+ struct fmt_frame_s *p_fmt_frame = NULL;
-+ spinlock_t *rx_q_lock = NULL;
-+
-+ rx_q_lock = &fmt_port->rx_q_lock;
-+
-+ spin_lock(rx_q_lock);
-+
-+#define list_last_entry(ptr, type, member) list_entry((ptr)->prev, type, member)
-+
-+ if (!list_empty(&fmt_port->rx_q)) {
-+ p_fmt_frame = list_last_entry(&fmt_port->rx_q,
-+ struct fmt_frame_s,
-+ list);
-+ list_del(&p_fmt_frame->list);
-+
-+#if defined(FMT_K_DBG) || defined(FMT_K_DBG_RUNTIME)
-+ atomic_inc(&fmt_port->dequeue_from_rxq);
-+#endif
-+ }
-+
-+ spin_unlock(rx_q_lock);
-+
-+ return p_fmt_frame;
-+}
-+
-+/* eth-dev -to- fmt port association */
-+struct fmt_port_s *match_dpa_to_fmt_port(
-+ struct dpa_priv_s *dpa_priv) {
-+ struct mac_device *mac_dev = dpa_priv->mac_dev;
-+ struct fm_port *fm_port = (struct fm_port *) mac_dev;
-+ struct fmt_port_s *fmt_port = NULL;
-+ int i;
-+
-+ _fmt_dbgr("calling...\n");
-+
-+ /* find the FM-test-port object */
-+ for (i = 0; i < IOC_FMT_MAX_NUM_OF_PORTS; i++)
-+ if ((fm_test.ports[i].p_mac_dev &&
-+ mac_dev == fm_test.ports[i].p_mac_dev) ||
-+ fm_port == fm_test.ports[i].p_tx_port) {
-+
-+ fmt_port = &fm_test.ports[i];
-+ break;
-+ }
-+
-+ _fmt_dbgr("called\n");
-+ return fmt_port;
-+}
-+
-+void dump_frame(
-+ uint8_t *buffer,
-+ uint32_t size)
-+{
-+#if defined(FMT_K_DBG) || defined(FMT_K_DBG_RUNTIME)
-+ unsigned int i;
-+
-+ for (i = 0; i < size; i++) {
-+ if (i%16 == 0)
-+ printk(KERN_DEBUG "\n");
-+ printk(KERN_DEBUG "%2x ", *(buffer+i));
-+ }
-+#endif
-+ return;
-+}
-+
-+bool test_and_steal_frame(struct fmt_port_s *fmt_port,
-+ uint32_t fqid,
-+ uint8_t *buffer,
-+ uint32_t size)
-+{
-+ struct fmt_frame_s *p_fmt_frame = NULL;
-+ bool test_and_steal_frame_frame;
-+ uint32_t data_offset;
-+ uint32_t i;
-+
-+ _fmt_dbgr("calling...\n");
-+
-+ if (!fmt_port || !fmt_port->p_rx_fm_port_dev)
-+ return false;
-+
-+ /* check watermark */
-+ test_and_steal_frame_frame = false;
-+ for (i = 0; i < size; i++) {
-+ uint64_t temp = *((uint64_t *)(buffer + i));
-+
-+ if (temp == (uint64_t) FMT_FRM_WATERMARK) {
-+ _fmt_dbgr("watermark found!\n");
-+ test_and_steal_frame_frame = true;
-+ break;
-+ }
-+ }
-+
-+ if (!test_and_steal_frame_frame) {
-+#if defined(FMT_K_DBG) || defined(FMT_K_DBG_RUNTIME)
-+ atomic_inc(&fmt_port->not_enqueue_to_rxq_wrong_frm);
-+#endif
-+ _fmt_dbgr("NOT watermark found!\n");
-+ return false;
-+ }
-+
-+ /* do not enqueue the tx conf/err frames */
-+ if ((fqid == FMT_TX_CONF_Q) || (fqid == FMT_TX_ERR_Q))
-+ goto _test_and_steal_frame_return_true;
-+
-+ _fmt_dbgr("on port %d got FMUC frame\n", fmt_port->id);
-+ data_offset = FM_PORT_GetBufferDataOffset(
-+ fmt_port->p_rx_fm_port_dev);
-+
-+ p_fmt_frame = kmalloc(sizeof(struct fmt_frame_s), GFP_KERNEL);
-+
-+ /* dump frame... no more space left on device */
-+ if (p_fmt_frame == NULL) {
-+ _fmt_err("no space left on device!\n");
-+ goto _test_and_steal_frame_return_true;
-+ }
-+
-+ memset(p_fmt_frame, 0, sizeof(struct fmt_frame_s));
-+ p_fmt_frame->buff.p_data = kmalloc(size * sizeof(uint8_t), GFP_KERNEL);
-+
-+ /* No more space left on device*/
-+ if (p_fmt_frame->buff.p_data == NULL) {
-+ _fmt_err("no space left on device!\n");
-+ kfree(p_fmt_frame);
-+ goto _test_and_steal_frame_return_true;
-+ }
-+
-+ p_fmt_frame->buff.size = size-data_offset;
-+ p_fmt_frame->buff.qid = fqid;
-+
-+ memcpy(p_fmt_frame->buff.p_data,
-+ (uint8_t *)PTR_MOVE(buffer, data_offset),
-+ p_fmt_frame->buff.size);
-+
-+ memcpy(p_fmt_frame->buff.buff_context.fm_prs_res,
-+ FM_PORT_GetBufferPrsResult(fmt_port->p_rx_fm_port_dev,
-+ (char *)buffer),
-+ 32);
-+
-+ /* enqueue frame - this frame will go to us */
-+ enqueue_fmt_frame(fmt_port, p_fmt_frame);
-+
-+_test_and_steal_frame_return_true:
-+ return true;
-+}
-+
-+static int fmt_fq_release(const struct qm_fd *fd)
-+{
-+ struct dpa_bp *_dpa_bp;
-+ struct bm_buffer _bmb;
-+
-+ if (fd->format == qm_fd_contig) {
-+ _dpa_bp = dpa_bpid2pool(fd->bpid);
-+ BUG_ON(IS_ERR(_dpa_bp));
-+
-+ _bmb.hi = fd->addr_hi;
-+ _bmb.lo = fd->addr_lo;
-+
-+ while (bman_release(_dpa_bp->pool, &_bmb, 1, 0))
-+ cpu_relax();
-+
-+ } else {
-+ _fmt_err("frame not supported !\n");
-+ return -1;
-+ }
-+
-+ return 0;
-+}
-+
-+/* sync it w/ dpaa_eth.c: DPA_BP_HEAD */
-+#define DPA_BP_HEADROOM (DPA_TX_PRIV_DATA_SIZE + \
-+ fm_get_rx_extra_headroom() + \
-+ DPA_PARSE_RESULTS_SIZE + \
-+ DPA_HASH_RESULTS_SIZE)
-+#define MAC_HEADER_LENGTH 14
-+#define L2_AND_HEADROOM_OFF ((DPA_BP_HEADROOM) + (MAC_HEADER_LENGTH))
-+
-+/* dpa ingress hooks definition */
-+enum dpaa_eth_hook_result fmt_rx_default_hook(
-+ struct sk_buff *skb,
-+ struct net_device *net_dev,
-+ u32 fqid)
-+{
-+ struct dpa_priv_s *dpa_priv = NULL;
-+ struct fmt_port_s *fmt_port = NULL;
-+ uint8_t *buffer;
-+ uint32_t buffer_len;
-+
-+ _fmt_dbgr("calling...\n");
-+
-+ dpa_priv = netdev_priv(net_dev);
-+ fmt_port = match_dpa_to_fmt_port(dpa_priv);
-+
-+ /* conversion from skb to fd:
-+ * skb cames processed for L3, so we need to go back for
-+ * layer 2 offset */
-+ buffer = (uint8_t *)(skb->data - ((int)L2_AND_HEADROOM_OFF));
-+ buffer_len = skb->len + ((int)L2_AND_HEADROOM_OFF);
-+
-+ /* if is not out frame let dpa to handle it */
-+ if (test_and_steal_frame(fmt_port,
-+ FMT_RX_DFLT_Q,
-+ buffer,
-+ buffer_len))
-+ goto _fmt_rx_default_hook_stolen;
-+
-+ _fmt_dbgr("called:DPAA_ETH_CONTINUE.\n");
-+ return DPAA_ETH_CONTINUE;
-+
-+_fmt_rx_default_hook_stolen:
-+ dev_kfree_skb(skb);
-+
-+ _fmt_dbgr("called:DPAA_ETH_STOLEN.\n");
-+ return DPAA_ETH_STOLEN;
-+}
-+
-+enum dpaa_eth_hook_result fmt_rx_error_hook(
-+ struct net_device *net_dev,
-+ const struct qm_fd *fd,
-+ u32 fqid)
-+{
-+ struct dpa_priv_s *dpa_priv = NULL;
-+ struct dpa_bp *dpa_bp = NULL;
-+ struct fmt_port_s *fmt_port = NULL;
-+ void *fd_virt_addr = NULL;
-+ dma_addr_t addr = qm_fd_addr(fd);
-+
-+ _fmt_dbgr("calling...\n");
-+
-+ dpa_priv = netdev_priv(net_dev);
-+ fmt_port = match_dpa_to_fmt_port(dpa_priv);
-+
-+ /* dpaa doesn't do this... we have to do it here */
-+ dpa_bp = dpa_bpid2pool(fd->bpid);
-+ dma_unmap_single(dpa_bp->dev, addr, dpa_bp->size, DMA_BIDIRECTIONAL);
-+
-+ fd_virt_addr = phys_to_virt(addr);
-+ /* if is not out frame let dpa to handle it */
-+ if (test_and_steal_frame(fmt_port,
-+ FMT_RX_ERR_Q,
-+ fd_virt_addr,
-+ fd->length20 + fd->offset)) {
-+ goto _fmt_rx_error_hook_stolen;
-+ }
-+
-+ _fmt_dbgr("called:DPAA_ETH_CONTINUE.\n");
-+ return DPAA_ETH_CONTINUE;
-+
-+_fmt_rx_error_hook_stolen:
-+ /* the frame data doesn't matter,
-+ * so, no mapping is needed */
-+ fmt_fq_release(fd);
-+
-+ _fmt_dbgr("called:DPAA_ETH_STOLEN.\n");
-+ return DPAA_ETH_STOLEN;
-+}
-+
-+enum dpaa_eth_hook_result fmt_tx_confirm_hook(
-+ struct net_device *net_dev,
-+ const struct qm_fd *fd,
-+ u32 fqid)
-+{
-+ struct dpa_priv_s *dpa_priv = NULL;
-+ struct fmt_port_s *fmt_port = NULL;
-+ dma_addr_t addr = qm_fd_addr(fd);
-+ void *fd_virt_addr = NULL;
-+ uint32_t fd_len = 0;
-+
-+ _fmt_dbgr("calling...\n");
-+
-+ dpa_priv = netdev_priv(net_dev);
-+ fmt_port = match_dpa_to_fmt_port(dpa_priv);
-+
-+ fd_virt_addr = phys_to_virt(addr);
-+ fd_len = fd->length20 + fd->offset;
-+
-+ if (fd_len > fm_get_max_frm()) {
-+ _fmt_err("tx confirm bad frame size: %u!\n", fd_len);
-+ goto _fmt_tx_confirm_hook_continue;
-+ }
-+
-+ if (test_and_steal_frame(fmt_port,
-+ FMT_TX_CONF_Q,
-+ fd_virt_addr,
-+ fd_len))
-+ goto _fmt_tx_confirm_hook_stolen;
-+
-+_fmt_tx_confirm_hook_continue:
-+ _fmt_dbgr("called:DPAA_ETH_CONTINUE.\n");
-+ return DPAA_ETH_CONTINUE;
-+
-+_fmt_tx_confirm_hook_stolen:
-+ kfree(fd_virt_addr);
-+
-+ _fmt_dbgr("called:DPAA_ETH_STOLEN.\n");
-+ return DPAA_ETH_STOLEN;
-+}
-+
-+enum dpaa_eth_hook_result fmt_tx_confirm_error_hook(
-+ struct net_device *net_dev,
-+ const struct qm_fd *fd,
-+ u32 fqid)
-+{
-+ struct dpa_priv_s *dpa_priv = NULL;
-+ struct fmt_port_s *fmt_port = NULL;
-+ dma_addr_t addr = qm_fd_addr(fd);
-+ void *fd_virt_addr = NULL;
-+ uint32_t fd_len = 0;
-+
-+ _fmt_dbgr("calling...\n");
-+
-+ dpa_priv = netdev_priv(net_dev);
-+ fmt_port = match_dpa_to_fmt_port(dpa_priv);
-+
-+ fd_virt_addr = phys_to_virt(addr);
-+ fd_len = fd->length20 + fd->offset;
-+
-+ if (fd_len > fm_get_max_frm()) {
-+ _fmt_err("tx confirm err bad frame size: %u !\n", fd_len);
-+ goto _priv_ingress_tx_err_continue;
-+ }
-+
-+ if (test_and_steal_frame(fmt_port, FMT_TX_ERR_Q, fd_virt_addr, fd_len))
-+ goto _priv_ingress_tx_err_stolen;
-+
-+_priv_ingress_tx_err_continue:
-+ _fmt_dbgr("called:DPAA_ETH_CONTINUE.\n");
-+ return DPAA_ETH_CONTINUE;
-+
-+_priv_ingress_tx_err_stolen:
-+ kfree(fd_virt_addr);
-+
-+ _fmt_dbgr("called:DPAA_ETH_STOLEN.\n");
-+ return DPAA_ETH_STOLEN;
-+}
-+
-+/* egress callbacks definition */
-+enum qman_cb_dqrr_result fmt_egress_dqrr(
-+ struct qman_portal *portal,
-+ struct qman_fq *fq,
-+ const struct qm_dqrr_entry *dqrr)
-+{
-+ /* this callback should never be called */
-+ BUG();
-+ return qman_cb_dqrr_consume;
-+}
-+
-+static void fmt_egress_error_dqrr(
-+ struct qman_portal *p,
-+ struct qman_fq *fq,
-+ const struct qm_mr_entry *msg)
-+{
-+ uint8_t *fd_virt_addr = NULL;
-+
-+ /* tx failure, on the ern callback - release buffer */
-+ fd_virt_addr = (uint8_t *)phys_to_virt(qm_fd_addr(&msg->ern.fd));
-+ kfree(fd_virt_addr);
-+
-+ return;
-+}
-+
-+static const struct qman_fq fmt_egress_fq = {
-+ .cb = { .dqrr = fmt_egress_dqrr,
-+ .ern = fmt_egress_error_dqrr,
-+ .fqs = NULL}
-+};
-+
-+int fmt_fq_alloc(
-+ struct fmt_fqs_s *fmt_fqs,
-+ const struct qman_fq *qman_fq,
-+ uint32_t fqid, uint32_t flags,
-+ uint16_t channel, uint8_t wq)
-+{
-+ int _errno = 0;
-+
-+ _fmt_dbg("calling...\n");
-+
-+ fmt_fqs->fq_base = *qman_fq;
-+
-+ if (fqid == 0) {
-+ flags |= QMAN_FQ_FLAG_DYNAMIC_FQID;
-+ flags &= ~QMAN_FQ_FLAG_NO_MODIFY;
-+ } else
-+ flags &= ~QMAN_FQ_FLAG_DYNAMIC_FQID;
-+
-+ fmt_fqs->init = !(flags & QMAN_FQ_FLAG_NO_MODIFY);
-+
-+ _errno = qman_create_fq(fqid, flags, &fmt_fqs->fq_base);
-+ if (_errno < 0) {
-+ _fmt_err("frame queues create failed.\n");
-+ return -EINVAL;
-+ }
-+
-+ if (fmt_fqs->init) {
-+ struct qm_mcc_initfq initfq;
-+
-+ initfq.we_mask = QM_INITFQ_WE_DESTWQ;
-+ initfq.fqd.dest.channel = channel;
-+ initfq.fqd.dest.wq = wq;
-+
-+ _errno = qman_init_fq(&fmt_fqs->fq_base,
-+ QMAN_INITFQ_FLAG_SCHED,
-+ &initfq);
-+ if (_errno < 0) {
-+ _fmt_err("frame queues init erorr.\n");
-+ qman_destroy_fq(&fmt_fqs->fq_base, 0);
-+ return -EINVAL;
-+ }
-+ }
-+
-+ _fmt_dbg("called.\n");
-+ return 0;
-+}
-+
-+static int fmt_fq_free(struct fmt_fqs_s *fmt_fq)
-+{
-+ int _err = 0;
-+
-+ _fmt_dbg("calling...\n");
-+
-+ if (fmt_fq->init) {
-+ _err = qman_retire_fq(&fmt_fq->fq_base, NULL);
-+ if (unlikely(_err < 0))
-+ _fmt_err("qman_retire_fq(%u) = %d\n",
-+ qman_fq_fqid(&fmt_fq->fq_base), _err);
-+
-+ _err = qman_oos_fq(&fmt_fq->fq_base);
-+ if (unlikely(_err < 0))
-+ _fmt_err("qman_oos_fq(%u) = %d\n",
-+ qman_fq_fqid(&fmt_fq->fq_base), _err);
-+ }
-+
-+ qman_destroy_fq(&fmt_fq->fq_base, 0);
-+
-+ _fmt_dbg("called.\n");
-+ return _err;
-+}
-+
-+/* private pcd dqrr calbacks */
-+static enum qman_cb_dqrr_result fmt_pcd_dqrr(
-+ struct qman_portal *portal,
-+ struct qman_fq *fq,
-+ const struct qm_dqrr_entry *dq)
-+{
-+ struct dpa_bp *dpa_bp = NULL;
-+ dma_addr_t addr = qm_fd_addr(&dq->fd);
-+ uint8_t *fd_virt_addr = NULL;
-+ struct fmt_port_s *fmt_port;
-+ struct fmt_port_pcd_s *fmt_port_pcd;
-+ uint32_t relative_fqid = 0;
-+ uint32_t fd_len = 0;
-+
-+ _fmt_dbgr("calling...\n");
-+
-+ /* upcast - from pcd_alloc_fq */
-+ fmt_port = ((struct fmt_fqs_s *)fq)->fmt_port_priv;
-+ if (!fmt_port) {
-+ _fmt_err(" wrong fmt port -to- fq match.\n");
-+ goto _fmt_pcd_dqrr_return;
-+ }
-+ fmt_port_pcd = &fmt_port->fmt_port_pcd;
-+
-+ relative_fqid = dq->fqid - fmt_port_pcd->fqid_base;
-+ _fmt_dbgr("pcd dqrr got frame on relative fq:%u@base:%u\n",
-+ relative_fqid, fmt_port_pcd->fqid_base);
-+
-+ fd_len = dq->fd.length20 + dq->fd.offset;
-+
-+ if (fd_len > fm_get_max_frm()) {
-+ _fmt_err("pcd dqrr wrong frame size: %u (%u:%u)!\n",
-+ fd_len, dq->fd.length20, dq->fd.offset);
-+ goto _fmt_pcd_dqrr_return;
-+ }
-+
-+ dpa_bp = dpa_bpid2pool(dq->fd.bpid);
-+ dma_unmap_single(dpa_bp->dev, addr, dpa_bp->size, DMA_BIDIRECTIONAL);
-+
-+ fd_virt_addr = phys_to_virt(addr);
-+ if (!test_and_steal_frame(fmt_port, relative_fqid, fd_virt_addr,
-+ fd_len)) {
-+
-+#if defined(FMT_K_DBG) || defined(FMT_K_DBG_RUNTIME)
-+ atomic_inc(&fmt_port->not_enqueue_to_rxq_wrong_frm);
-+#endif
-+ _fmt_wrn("pcd dqrr unrecognized frame@fqid: %u,"
-+ " frame len: %u (dropped).\n",
-+ dq->fqid, dq->fd.length20);
-+ dump_frame(fd_virt_addr, fd_len);
-+ }
-+
-+_fmt_pcd_dqrr_return:
-+ /* no need to map again here */
-+ fmt_fq_release(&dq->fd);
-+
-+ _fmt_dbgr("calle.\n");
-+ return qman_cb_dqrr_consume;
-+}
-+
-+static void fmt_pcd_err_dqrr(
-+ struct qman_portal *qm,
-+ struct qman_fq *fq,
-+ const struct qm_mr_entry *msg)
-+{
-+ _fmt_err("this callback should never be called.\n");
-+ BUG();
-+ return;
-+}
-+
-+static void fmt_pcd_fqs_dqrr(
-+ struct qman_portal *qm,
-+ struct qman_fq *fq,
-+ const struct qm_mr_entry *msg)
-+{
-+ _fmt_dbg(" fq state(0x%x)@fqid(%u.\n", msg->fq.fqs, msg->fq.fqid);
-+ return;
-+}
-+
-+/* private pcd queue template */
-+static const struct qman_fq pcd_fq = {
-+ .cb = { .dqrr = fmt_pcd_dqrr,
-+ .ern = fmt_pcd_err_dqrr,
-+ .fqs = fmt_pcd_fqs_dqrr}
-+};
-+
-+/* defined as weak in dpaa driver. */
-+/* ! parameters come from IOCTL call - US */
-+int dpa_alloc_pcd_fqids(
-+ struct device *dev,
-+ uint32_t num, uint8_t alignment,
-+ uint32_t *base_fqid)
-+{
-+ int _err = 0, i;
-+ struct net_device *net_dev = NULL;
-+ struct dpa_priv_s *dpa_priv = NULL;
-+ struct fmt_port_pcd_s *fmt_port_pcd = NULL;
-+ struct fmt_fqs_s *fmt_fqs = NULL;
-+ struct fmt_port_s *fmt_port = NULL;
-+ int num_allocated = 0;
-+
-+ _fmt_dbg("calling...\n");
-+
-+ net_dev = (typeof(net_dev))dev_get_drvdata(dev);
-+ dpa_priv = (typeof(dpa_priv))netdev_priv(net_dev);
-+
-+ if (!netif_msg_probe(dpa_priv)) {
-+ _fmt_err("dpa not probe.\n");
-+ _err = -ENODEV;
-+ goto _pcd_alloc_fqs_err;
-+ }
-+
-+ fmt_port = match_dpa_to_fmt_port(dpa_priv);
-+ if (!fmt_port) {
-+ _fmt_err("fmt port not found.");
-+ _err = -EINVAL;
-+ goto _pcd_alloc_fqs_err;
-+ }
-+
-+ fmt_port_pcd = &fmt_port->fmt_port_pcd;
-+
-+ num_allocated = qman_alloc_fqid_range(base_fqid, num, alignment, 0);
-+
-+ if ((num_allocated <= 0) ||
-+ (num_allocated < num) ||
-+ (alignment && (*base_fqid) % alignment)) {
-+ *base_fqid = 0;
-+ _fmt_err("Failed to alloc pcd fqs rang.\n");
-+ _err = -EINVAL;
-+ goto _pcd_alloc_fqs_err;
-+ }
-+
-+ _fmt_dbg("wanted %d fqs(align %d), got %d fqids@%u.\n",
-+ num, alignment, num_allocated, *base_fqid);
-+
-+ /* alloc pcd queues */
-+ fmt_port_pcd->fmt_pcd_fqs = kmalloc(num_allocated *
-+ sizeof(struct fmt_fqs_s),
-+ GFP_KERNEL);
-+ fmt_port_pcd->num_queues = num_allocated;
-+ fmt_port_pcd->fqid_base = *base_fqid;
-+ fmt_fqs = fmt_port_pcd->fmt_pcd_fqs;
-+
-+ /* alloc the pcd queues */
-+ for (i = 0; i < num_allocated; i++, fmt_fqs++) {
-+ _err = fmt_fq_alloc(
-+ fmt_fqs,
-+ &pcd_fq,
-+ (*base_fqid) + i, QMAN_FQ_FLAG_NO_ENQUEUE,
-+ dpa_priv->channel, 7);
-+
-+ if (_err < 0)
-+ goto _pcd_alloc_fqs_err;
-+
-+ /* upcast to identify from where the frames came from */
-+ fmt_fqs->fmt_port_priv = fmt_port;
-+ }
-+
-+ _fmt_dbg("called.\n");
-+ return _err;
-+_pcd_alloc_fqs_err:
-+ if (num_allocated > 0)
-+ qman_release_fqid_range(*base_fqid, num_allocated);
-+ /*TODO: free fmt_pcd_fqs if are any */
-+
-+ _fmt_dbg("called(_err:%d).\n", _err);
-+ return _err;
-+}
-+
-+/* defined as weak in dpaa driver. */
-+int dpa_free_pcd_fqids(
-+ struct device *dev,
-+ uint32_t base_fqid)
-+{
-+
-+ int _err = 0, i;
-+ struct net_device *net_dev = NULL;
-+ struct dpa_priv_s *dpa_priv = NULL;
-+ struct fmt_port_pcd_s *fmt_port_pcd = NULL;
-+ struct fmt_fqs_s *fmt_fqs = NULL;
-+ struct fmt_port_s *fmt_port = NULL;
-+ int num_allocated = 0;
-+
-+ _fmt_dbg("calling...\n");
-+
-+ net_dev = (typeof(net_dev))dev_get_drvdata(dev);
-+ dpa_priv = (typeof(dpa_priv))netdev_priv(net_dev);
-+
-+ if (!netif_msg_probe(dpa_priv)) {
-+ _fmt_err("dpa not probe.\n");
-+ _err = -ENODEV;
-+ goto _pcd_free_fqs_err;
-+ }
-+
-+ fmt_port = match_dpa_to_fmt_port(dpa_priv);
-+ if (!fmt_port) {
-+ _fmt_err("fmt port not found.");
-+ _err = -EINVAL;
-+ goto _pcd_free_fqs_err;
-+ }
-+
-+ fmt_port_pcd = &fmt_port->fmt_port_pcd;
-+ num_allocated = fmt_port_pcd->num_queues;
-+ fmt_fqs = fmt_port_pcd->fmt_pcd_fqs;
-+
-+ for (i = 0; i < num_allocated; i++, fmt_fqs++)
-+ fmt_fq_free(fmt_fqs);
-+
-+ qman_release_fqid_range(base_fqid,num_allocated);
-+
-+ kfree(fmt_port_pcd->fmt_pcd_fqs);
-+ memset(fmt_port_pcd, 0, sizeof(*fmt_port_pcd));
-+
-+ /* debugging stuff */
-+#if defined(FMT_K_DBG) || defined(FMT_K_DBG_RUNTIME)
-+ _fmt_dbg(" portid: %u.\n", fmt_port->id);
-+ _fmt_dbg(" frames enqueue to qman: %u.\n",
-+ atomic_read(&fmt_port->enqueue_to_qman_frm));
-+ _fmt_dbg(" frames enqueue to rxq: %u.\n",
-+ atomic_read(&fmt_port->enqueue_to_rxq));
-+ _fmt_dbg(" frames dequeue from rxq: %u.\n",
-+ atomic_read(&fmt_port->dequeue_from_rxq));
-+ _fmt_dbg(" frames not enqueue to rxq - wrong frm: %u.\n",
-+ atomic_read(&fmt_port->not_enqueue_to_rxq_wrong_frm));
-+ atomic_set(&fmt_port->enqueue_to_qman_frm, 0);
-+ atomic_set(&fmt_port->enqueue_to_rxq, 0);
-+ atomic_set(&fmt_port->dequeue_from_rxq, 0);
-+ atomic_set(&fmt_port->not_enqueue_to_rxq_wrong_frm, 0);
-+#endif
-+ return 0;
-+
-+_pcd_free_fqs_err:
-+ return _err;
-+}
-+
-+static int fmt_port_init(
-+ struct fmt_port_s *fmt_port,
-+ ioc_fmt_port_param_t *p_Params)
-+{
-+ struct device_node *fm_node, *fm_port_node;
-+ const uint32_t *uint32_prop;
-+ int _errno = 0, lenp = 0, i;
-+ static struct of_device_id fm_node_of_match[] = {
-+ { .compatible = "fsl,fman", },
-+ { /* end of list */ },
-+ };
-+
-+ _fmt_dbg("calling...\n");
-+
-+ /* init send/receive tu US list */
-+ INIT_LIST_HEAD(&fmt_port->rx_q);
-+
-+ /* check parameters */
-+ if (p_Params->num_tx_queues > FMAN_TEST_MAX_TX_FQS ||
-+ p_Params->fm_port_id > IOC_FMT_MAX_NUM_OF_PORTS) {
-+ _fmt_dbg("wrong test parameters.\n");
-+ return -EINVAL;
-+ }
-+
-+ /* set port parameters */
-+ fmt_port->num_of_tx_fqs = p_Params->num_tx_queues;
-+ fmt_port->id = p_Params->fm_port_id;
-+ fmt_port->port_type = p_Params->fm_port_type;
-+ fmt_port->diag = e_IOC_DIAG_MODE_NONE;
-+
-+ /* init debugging stuff */
-+#if defined(FMT_K_DBG) || defined(FMT_K_DBG_RUNTIME)
-+ atomic_set(&fmt_port->enqueue_to_qman_frm, 0);
-+ atomic_set(&fmt_port->enqueue_to_rxq, 0);
-+ atomic_set(&fmt_port->dequeue_from_rxq, 0);
-+ atomic_set(&fmt_port->not_enqueue_to_rxq_wrong_frm, 0);
-+#endif
-+
-+ /* TODO: This should be done at probe time not at runtime
-+ * very ugly function */
-+ /* fill fmt port properties from dts */
-+ for_each_matching_node(fm_node, fm_node_of_match) {
-+
-+ uint32_prop = (uint32_t *)of_get_property(fm_node,
-+ "cell-index", &lenp);
-+ if (unlikely(uint32_prop == NULL)) {
-+ _fmt_wrn("of_get_property(%s, cell-index) invalid",
-+ fm_node->full_name);
-+ return -EINVAL;
-+ }
-+ if (WARN_ON(lenp != sizeof(uint32_t))) {
-+ _fmt_wrn("of_get_property(%s, cell-index) invalid",
-+ fm_node->full_name);
-+ return -EINVAL;
-+ }
-+
-+ if (*uint32_prop == p_Params->fm_id) {
-+ struct resource res;
-+
-+ /* Get the FM address */
-+ _errno = of_address_to_resource(fm_node, 0, &res);
-+ if (unlikely(_errno < 0)) {
-+ _fmt_wrn("of_address_to_resource() = %u.\n", _errno);
-+ return -EINVAL;
-+ }
-+
-+ fmt_port->fm_phys_base_addr = res.start;
-+
-+ for_each_child_of_node(fm_node, fm_port_node) {
-+ struct platform_device *of_dev;
-+
-+ if (!of_device_is_available(fm_port_node))
-+ continue;
-+
-+ uint32_prop = (uint32_t *)of_get_property(
-+ fm_port_node,
-+ "cell-index",
-+ &lenp);
-+ if (uint32_prop == NULL)
-+ continue;
-+
-+ if (of_device_is_compatible(fm_port_node,
-+ "fsl,fman-port-oh") &&
-+ (fmt_port->port_type == e_IOC_FMT_PORT_T_OP)) {
-+
-+ if (*uint32_prop == fmt_port->id) {
-+ of_dev = of_find_device_by_node(fm_port_node);
-+ if (unlikely(of_dev == NULL)) {
-+ _fmt_wrn("fm id invalid\n");
-+ return -EINVAL;
-+ }
-+
-+ fmt_port->p_tx_port =
-+ fm_port_bind(&of_dev->dev);
-+ fmt_port->p_tx_fm_port_dev =
-+ (void *)fm_port_get_handle(
-+ fmt_port->p_tx_port);
-+ fmt_port->p_rx_port =
-+ fmt_port->p_tx_port;
-+ fmt_port->p_rx_fm_port_dev =
-+ fmt_port->p_tx_fm_port_dev;
-+ fmt_port->p_mac_dev = NULL;
-+ break;
-+ }
-+ } else if ((*uint32_prop == fmt_port->id) &&
-+ fmt_port->port_type == e_IOC_FMT_PORT_T_RXTX) {
-+
-+ of_dev = of_find_device_by_node(fm_port_node);
-+ if (unlikely(of_dev == NULL)) {
-+ _fmt_wrn("dtb fm id invalid value");
-+ return -EINVAL;
-+ }
-+
-+ if (of_device_is_compatible(fm_port_node,
-+ "fsl,fman-port-1g-tx")) {
-+ fmt_port->p_tx_port =
-+ fm_port_bind(&of_dev->dev);
-+ fmt_port->p_tx_fm_port_dev = (void *)
-+ fm_port_get_handle(
-+ fmt_port->p_tx_port);
-+ } else if (of_device_is_compatible(fm_port_node,
-+ "fsl,fman-port-1g-rx")) {
-+ fmt_port->p_rx_port =
-+ fm_port_bind(&of_dev->dev);
-+ fmt_port->p_rx_fm_port_dev = (void *)
-+ fm_port_get_handle(
-+ fmt_port->p_rx_port);
-+ } else if (of_device_is_compatible(fm_port_node,
-+ "fsl,fman-1g-mac") ||
-+ of_device_is_compatible(fm_port_node,
-+ "fsl,fman-memac"))
-+ fmt_port->p_mac_dev =
-+ (typeof(fmt_port->p_mac_dev))
-+ dev_get_drvdata(&of_dev->dev);
-+ else
-+ continue;
-+
-+ if (fmt_port->p_tx_fm_port_dev &&
-+ fmt_port->p_rx_fm_port_dev && fmt_port->p_mac_dev)
-+ break;
-+ } else if (((*uint32_prop + FM_MAX_NUM_OF_1G_RX_PORTS) ==
-+ fmt_port->id) &&
-+ fmt_port->port_type == e_IOC_FMT_PORT_T_RXTX) {
-+
-+ of_dev = of_find_device_by_node(fm_port_node);
-+ if (unlikely(of_dev == NULL)) {
-+ _fmt_wrn("dtb fm id invalid value\n");
-+ return -EINVAL;
-+ }
-+
-+ if (of_device_is_compatible(fm_port_node,
-+ "fsl,fman-port-10g-tx")) {
-+ fmt_port->p_tx_port =
-+ fm_port_bind(&of_dev->dev);
-+ fmt_port->p_tx_fm_port_dev = (void *)
-+ fm_port_get_handle(
-+ fmt_port->p_tx_port);
-+ } else if (of_device_is_compatible(fm_port_node,
-+ "fsl,fman-port-10g-rx")) {
-+ fmt_port->p_rx_port =
-+ fm_port_bind(&of_dev->dev);
-+ fmt_port->p_rx_fm_port_dev = (void *)
-+ fm_port_get_handle(
-+ fmt_port->p_rx_port);
-+ } else if (of_device_is_compatible(fm_port_node,
-+ "fsl,fman-10g-mac") ||
-+ of_device_is_compatible(fm_port_node,
-+ "fsl,fman-memac"))
-+ fmt_port->p_mac_dev =
-+ (typeof(fmt_port->p_mac_dev))
-+ dev_get_drvdata(&of_dev->dev);
-+ else
-+ continue;
-+
-+ if (fmt_port->p_tx_fm_port_dev &&
-+ fmt_port->p_rx_fm_port_dev && fmt_port->p_mac_dev)
-+ break;
-+ }
-+ } /* for_each_child */
-+ }
-+ } /* for each matching node */
-+
-+ if (fmt_port->p_tx_fm_port_dev == 0 ||
-+ fmt_port->p_rx_fm_port_dev == 0) {
-+
-+ _fmt_err("bad fm port pointers.\n");
-+ return -EINVAL;
-+ }
-+
-+ _fmt_dbg("alloc %u tx queues.\n", fmt_port->num_of_tx_fqs);
-+
-+ /* init fman test egress dynamic frame queues */
-+ for (i = 0; i < fmt_port->num_of_tx_fqs; i++) {
-+ int _errno;
-+ _errno = fmt_fq_alloc(
-+ &fmt_port->p_tx_fqs[i],
-+ &fmt_egress_fq,
-+ 0,
-+ QMAN_FQ_FLAG_TO_DCPORTAL,
-+ fm_get_tx_port_channel(fmt_port->p_tx_port),
-+ i);
-+
-+ if (_errno < 0) {
-+ _fmt_err("tx queues allocation failed.\n");
-+ /* TODO: memory leak here if 1 queue is allocated and
-+ * next queues are failing ... */
-+ return -EINVAL;
-+ }
-+ }
-+
-+ /* port is valid and ready to use. */
-+ fmt_port->valid = TRUE;
-+
-+ _fmt_dbg("called.\n");
-+ return 0;
-+}
-+
-+/* fm test chardev functions */
-+static int fmt_open(struct inode *inode, struct file *file)
-+{
-+ unsigned int minor = iminor(inode);
-+
-+ _fmt_dbg("calling...\n");
-+
-+ if (file->private_data != NULL)
-+ return 0;
-+
-+ /* The minor represent the port number.
-+ * Set the port structure accordingly, thus all the operations
-+ * will be done on this port. */
-+ if ((minor >= DEV_FM_TEST_PORTS_MINOR_BASE) &&
-+ (minor < DEV_FM_TEST_MAX_MINORS))
-+ file->private_data = &fm_test.ports[minor];
-+ else
-+ return -ENXIO;
-+
-+ _fmt_dbg("called.\n");
-+ return 0;
-+}
-+
-+static int fmt_close(struct inode *inode, struct file *file)
-+{
-+ struct fmt_port_s *fmt_port = NULL;
-+ struct fmt_frame_s *fmt_frame = NULL;
-+
-+ int err = 0;
-+
-+ _fmt_dbg("calling...\n");
-+
-+ fmt_port = file->private_data;
-+ if (!fmt_port)
-+ return -ENODEV;
-+
-+ /* Close the current test port by invalidating it. */
-+ fmt_port->valid = FALSE;
-+
-+ /* clean the fmt port queue */
-+ while ((fmt_frame = dequeue_fmt_frame(fmt_port)) != NULL) {
-+ if (fmt_frame && fmt_frame->buff.p_data){
-+ kfree(fmt_frame->buff.p_data);
-+ kfree(fmt_frame);
-+ }
-+ }
-+
-+ /* !!! the qman queues are cleaning from fm_ioctl...
-+ * - very ugly */
-+
-+ _fmt_dbg("called.\n");
-+ return err;
-+}
-+
-+static int fmt_ioctls(unsigned int minor,
-+ struct file *file,
-+ unsigned int cmd,
-+ unsigned long arg,
-+ bool compat)
-+{
-+ struct fmt_port_s *fmt_port = NULL;
-+
-+ _fmt_dbg("IOCTL minor:%u "
-+ " arg:0x%08lx ioctl cmd (0x%08x):(0x%02x:0x%02x.\n",
-+ minor, arg, cmd, _IOC_TYPE(cmd), _IOC_NR(cmd));
-+
-+ fmt_port = file->private_data;
-+ if (!fmt_port) {
-+ _fmt_err("invalid fmt port.\n");
-+ return -ENODEV;
-+ }
-+
-+ /* set test type properly */
-+ if (compat)
-+ fmt_port->compat_test_type = true;
-+ else
-+ fmt_port->compat_test_type = false;
-+
-+ switch (cmd) {
-+ case FMT_PORT_IOC_INIT:
-+ {
-+ ioc_fmt_port_param_t param;
-+
-+ if (fmt_port->valid) {
-+ _fmt_wrn("port is already initialized.\n");
-+ return -EFAULT;
-+ }
-+#if defined(CONFIG_COMPAT)
-+ if (compat) {
-+ if (copy_from_user(&param,
-+ (ioc_fmt_port_param_t *)compat_ptr(arg),
-+ sizeof(ioc_fmt_port_param_t)))
-+
-+ return -EFAULT;
-+ } else
-+#endif
-+ {
-+ if (copy_from_user(&param,
-+ (ioc_fmt_port_param_t *) arg,
-+ sizeof(ioc_fmt_port_param_t)))
-+
-+ return -EFAULT;
-+ }
-+
-+ return fmt_port_init(fmt_port, &param);
-+ }
-+
-+ case FMT_PORT_IOC_SET_DIAG_MODE:
-+ if (get_user(fmt_port->diag, (ioc_diag_mode *)arg))
-+ return -EFAULT;
-+
-+ if (fmt_port->diag == e_IOC_DIAG_MODE_CTRL_LOOPBACK)
-+ return set_mac_int_loopback(fmt_port, TRUE);
-+ else
-+ return set_mac_int_loopback(fmt_port, FALSE);
-+ break;
-+
-+ case FMT_PORT_IOC_SET_DPAECHO_MODE:
-+ case FMT_PORT_IOC_SET_IP_HEADER_MANIP:
-+ default:
-+ _fmt_wrn("ioctl unimplemented minor:%u@ioctl"
-+ " cmd:0x%08x(type:0x%02x, nr:0x%02x.\n",
-+ minor, cmd, _IOC_TYPE(cmd), _IOC_NR(cmd));
-+ return -EFAULT;
-+ }
-+
-+ return 0;
-+}
-+
-+#ifdef CONFIG_COMPAT
-+static long fmt_compat_ioctl(
-+ struct file *file,
-+ unsigned int cmd,
-+ unsigned long arg)
-+{
-+ unsigned int minor = iminor(file->f_path.dentry->d_inode);
-+
-+ _fmt_dbg("calling...\n");
-+ return fmt_ioctls(minor, file, cmd, arg, true);
-+}
-+#endif
-+
-+static long fmt_ioctl(
-+ struct file *file,
-+ unsigned int cmd,
-+ unsigned long arg)
-+{
-+ unsigned int minor = iminor(file->f_path.dentry->d_inode);
-+ unsigned int res;
-+
-+ _fmt_dbg("calling...\n");
-+
-+ fm_mutex_lock();
-+ res = fmt_ioctls(minor, file, cmd, arg, false);
-+ fm_mutex_unlock();
-+
-+ _fmt_dbg("called.\n");
-+
-+ return res;
-+}
-+
-+#ifdef CONFIG_COMPAT
-+void copy_compat_test_frame_buffer(
-+ ioc_fmt_buff_desc_t *buff,
-+ ioc_fmt_compat_buff_desc_t *compat_buff)
-+{
-+ compat_buff->qid = buff->qid;
-+ compat_buff->p_data = ptr_to_compat(buff->p_data);
-+ compat_buff->size = buff->size;
-+ compat_buff->status = buff->status;
-+
-+ compat_buff->buff_context.p_user_priv =
-+ ptr_to_compat(buff->buff_context.p_user_priv);
-+ memcpy(compat_buff->buff_context.fm_prs_res,
-+ buff->buff_context.fm_prs_res,
-+ FM_PRS_MAX * sizeof(uint8_t));
-+ memcpy(compat_buff->buff_context.fm_time_stamp,
-+ buff->buff_context.fm_time_stamp,
-+ FM_TIME_STAMP_MAX * sizeof(uint8_t));
-+}
-+#endif
-+
-+ssize_t fmt_read(
-+ struct file *file,
-+ char __user *buf,
-+ size_t size,
-+ loff_t *ppos)
-+{
-+ struct fmt_port_s *fmt_port = NULL;
-+ struct fmt_frame_s *p_fmt_frame = NULL;
-+ ssize_t cnt = 0;
-+
-+ fmt_port = file->private_data;
-+ if (!fmt_port || !fmt_port->valid) {
-+ _fmt_err("fmt port not valid!\n");
-+ return -ENODEV;
-+ }
-+
-+ p_fmt_frame = dequeue_fmt_frame(fmt_port);
-+ if (p_fmt_frame == NULL)
-+ return 0;
-+
-+ _fmt_dbgr("calling...\n");
-+
-+#ifdef CONFIG_COMPAT
-+ if (fmt_port->compat_test_type){
-+ cnt = sizeof(ioc_fmt_compat_buff_desc_t);
-+ }
-+ else
-+#endif
-+ {
-+ cnt = sizeof(ioc_fmt_buff_desc_t);
-+ }
-+
-+ if (size < cnt) {
-+ _fmt_err("illegal buffer-size!\n");
-+ cnt = 0;
-+ goto _fmt_read_return;
-+ }
-+
-+ /* Copy structure */
-+#ifdef CONFIG_COMPAT
-+ if (fmt_port->compat_test_type) {
-+ {
-+ ioc_fmt_compat_buff_desc_t compat_buff;
-+ copy_compat_test_frame_buffer(&p_fmt_frame->buff,
-+ &compat_buff);
-+
-+ if (copy_to_user(buf, &compat_buff, cnt)) {
-+ _fmt_err("copy_to_user failed!\n");
-+ goto _fmt_read_return;
-+ }
-+ }
-+
-+ ((ioc_fmt_compat_buff_desc_t *)buf)->p_data =
-+ ptr_to_compat(buf+sizeof(ioc_fmt_compat_buff_desc_t));
-+ cnt += MIN(p_fmt_frame->buff.size, size-cnt);
-+ } else
-+#endif
-+ {
-+ if (copy_to_user(buf, &p_fmt_frame->buff, cnt)) {
-+ _fmt_err("copy_to_user failed!\n");
-+ goto _fmt_read_return;
-+ }
-+
-+ ((ioc_fmt_buff_desc_t *)buf)->p_data =
-+ buf + sizeof(ioc_fmt_buff_desc_t);
-+ cnt += MIN(p_fmt_frame->buff.size, size-cnt);
-+ }
-+
-+ if (size < cnt) {
-+ _fmt_err("illegal buffer-size!\n");
-+ goto _fmt_read_return;
-+ }
-+
-+ /* copy frame */
-+#ifdef CONFIG_COMPAT
-+ if (fmt_port->compat_test_type) {
-+ if (copy_to_user(buf+sizeof(ioc_fmt_compat_buff_desc_t),
-+ p_fmt_frame->buff.p_data, cnt)) {
-+ _fmt_err("copy_to_user failed!\n");
-+ goto _fmt_read_return;
-+ }
-+ } else
-+#endif
-+ {
-+ if (copy_to_user(buf+sizeof(ioc_fmt_buff_desc_t),
-+ p_fmt_frame->buff.p_data, cnt)) {
-+ _fmt_err("copy_to_user failed!\n");
-+ goto _fmt_read_return;
-+ }
-+ }
-+
-+_fmt_read_return:
-+ kfree(p_fmt_frame->buff.p_data);
-+ kfree(p_fmt_frame);
-+
-+ _fmt_dbgr("called.\n");
-+ return cnt;
-+}
-+
-+ssize_t fmt_write(
-+ struct file *file,
-+ const char __user *buf,
-+ size_t size,
-+ loff_t *ppos)
-+{
-+ struct fmt_port_s *fmt_port = NULL;
-+ ioc_fmt_buff_desc_t buff_desc;
-+#ifdef CONFIG_COMPAT
-+ ioc_fmt_compat_buff_desc_t buff_desc_compat;
-+#endif
-+ uint8_t *p_data = NULL;
-+ uint32_t data_offset;
-+ int _errno;
-+ t_DpaaFD fd;
-+
-+ _fmt_dbgr("calling...\n");
-+
-+ fmt_port = file->private_data;
-+ if (!fmt_port || !fmt_port->valid) {
-+ _fmt_err("fmt port not valid.\n");
-+ return -EINVAL;
-+ }
-+
-+ /* If Compat (32B UserSpace - 64B KernelSpace) */
-+#ifdef CONFIG_COMPAT
-+ if (fmt_port->compat_test_type) {
-+ if (size < sizeof(ioc_fmt_compat_buff_desc_t)) {
-+ _fmt_err("invalid buff_desc size.\n");
-+ return -EFAULT;
-+ }
-+
-+ if (copy_from_user(&buff_desc_compat, buf,
-+ sizeof(ioc_fmt_compat_buff_desc_t)))
-+ return -EFAULT;
-+
-+ buff_desc.qid = buff_desc_compat.qid;
-+ buff_desc.p_data = compat_ptr(buff_desc_compat.p_data);
-+ buff_desc.size = buff_desc_compat.size;
-+ buff_desc.status = buff_desc_compat.status;
-+
-+ buff_desc.buff_context.p_user_priv =
-+ compat_ptr(buff_desc_compat.buff_context.p_user_priv);
-+ memcpy(buff_desc.buff_context.fm_prs_res,
-+ buff_desc_compat.buff_context.fm_prs_res,
-+ FM_PRS_MAX * sizeof(uint8_t));
-+ memcpy(buff_desc.buff_context.fm_time_stamp,
-+ buff_desc_compat.buff_context.fm_time_stamp,
-+ FM_TIME_STAMP_MAX * sizeof(uint8_t));
-+ } else
-+#endif
-+ {
-+ if (size < sizeof(ioc_fmt_buff_desc_t)) {
-+ _fmt_err("invalid buff_desc size.\n");
-+ return -EFAULT;
-+ }
-+
-+ if (copy_from_user(&buff_desc, (ioc_fmt_buff_desc_t *)buf,
-+ sizeof(ioc_fmt_buff_desc_t)))
-+ return -EFAULT;
-+ }
-+
-+ data_offset = FM_PORT_GetBufferDataOffset(fmt_port->p_tx_fm_port_dev);
-+ p_data = kmalloc(buff_desc.size+data_offset, GFP_KERNEL);
-+ if (!p_data)
-+ return -ENOMEM;
-+
-+ /* If Compat (32UserSpace - 64KernelSpace) the buff_desc.p_data is ok */
-+ if (copy_from_user((uint8_t *)PTR_MOVE(p_data, data_offset),
-+ buff_desc.p_data,
-+ buff_desc.size)) {
-+ kfree(p_data);
-+ return -EFAULT;
-+ }
-+
-+ /* TODO: dma_map_single here (cannot access the bpool struct) */
-+
-+ /* prepare fd */
-+ memset(&fd, 0, sizeof(fd));
-+ DPAA_FD_SET_ADDR(&fd, p_data);
-+ DPAA_FD_SET_OFFSET(&fd, data_offset);
-+ DPAA_FD_SET_LENGTH(&fd, buff_desc.size);
-+
-+ _errno = qman_enqueue(&fmt_port->p_tx_fqs[buff_desc.qid].fq_base,
-+ (struct qm_fd *)&fd, 0);
-+ if (_errno) {
-+ buff_desc.status = (uint32_t)_errno;
-+ if (copy_to_user((ioc_fmt_buff_desc_t *)buf, &buff_desc,
-+ sizeof(ioc_fmt_buff_desc_t))) {
-+ kfree(p_data);
-+ return -EFAULT;
-+ }
-+ }
-+
-+ /* for debugging */
-+#if defined(FMT_K_DBG) || defined(FMT_K_DBG_RUNTIME)
-+ atomic_inc(&fmt_port->enqueue_to_qman_frm);
-+#endif
-+ _fmt_dbgr("called.\n");
-+ return buff_desc.size;
-+}
-+
-+/* fm test character device definition */
-+static const struct file_operations fmt_fops =
-+{
-+ .owner = THIS_MODULE,
-+#ifdef CONFIG_COMPAT
-+ .compat_ioctl = fmt_compat_ioctl,
-+#endif
-+ .unlocked_ioctl = fmt_ioctl,
-+ .open = fmt_open,
-+ .release = fmt_close,
-+ .read = fmt_read,
-+ .write = fmt_write,
-+};
-+
-+static int fmt_init(void)
-+{
-+ int id;
-+
-+ _fmt_dbg("calling...\n");
-+
-+ /* Register to the /dev for IOCTL API */
-+ /* Register dynamically a new major number for the character device: */
-+ fm_test.major = register_chrdev(0, DEV_FM_TEST_NAME, &fmt_fops);
-+ if (fm_test.major <= 0) {
-+ _fmt_wrn("Failed to allocate major number for device %s.\n",
-+ DEV_FM_TEST_NAME);
-+ return -ENODEV;
-+ }
-+
-+ /* Creating class for FMan_test */
-+ fm_test.fmt_class = class_create(THIS_MODULE, DEV_FM_TEST_NAME);
-+ if (IS_ERR(fm_test.fmt_class)) {
-+ unregister_chrdev(fm_test.major, DEV_FM_TEST_NAME);
-+ _fmt_wrn("Error creating %s class.\n", DEV_FM_TEST_NAME);
-+ return -ENODEV;
-+ }
-+
-+ for (id = 0; id < IOC_FMT_MAX_NUM_OF_PORTS; id++)
-+ if (NULL == device_create(fm_test.fmt_class, NULL,
-+ MKDEV(fm_test.major,
-+ DEV_FM_TEST_PORTS_MINOR_BASE + id), NULL,
-+ DEV_FM_TEST_NAME "%d", id)) {
-+
-+ _fmt_err("Error creating %s device.\n",
-+ DEV_FM_TEST_NAME);
-+ return -ENODEV;
-+ }
-+
-+ return 0;
-+}
-+
-+static void fmt_free(void)
-+{
-+ int id;
-+
-+ for (id = 0; id < IOC_FMT_MAX_NUM_OF_PORTS; id++)
-+ device_destroy(fm_test.fmt_class, MKDEV(fm_test.major,
-+ DEV_FM_TEST_PORTS_MINOR_BASE + id));
-+ class_destroy(fm_test.fmt_class);
-+}
-+
-+static int __init __cold fmt_load(void)
-+{
-+ struct dpaa_eth_hooks_s priv_dpaa_eth_hooks;
-+
-+ /* set dpaa hooks for default queues */
-+ memset(&priv_dpaa_eth_hooks, 0, sizeof(priv_dpaa_eth_hooks));
-+ priv_dpaa_eth_hooks.rx_default = fmt_rx_default_hook;
-+ priv_dpaa_eth_hooks.rx_error = fmt_rx_error_hook;
-+ priv_dpaa_eth_hooks.tx_confirm = fmt_tx_confirm_hook;
-+ priv_dpaa_eth_hooks.tx_error = fmt_tx_confirm_error_hook;
-+
-+ fsl_dpaa_eth_set_hooks(&priv_dpaa_eth_hooks);
-+
-+ /* initialize the fman test environment */
-+ if (fmt_init() < 0) {
-+ _fmt_err("Failed to init FM-test modul.\n");
-+ fmt_free();
-+ return -ENODEV;
-+ }
-+
-+ _fmt_inf("FSL FM test module loaded.\n");
-+
-+ return 0;
-+}
-+
-+static void __exit __cold fmt_unload(void)
-+{
-+ fmt_free();
-+ _fmt_inf("FSL FM test module unloaded.\n");
-+}
-+
-+module_init(fmt_load);
-+module_exit(fmt_unload);
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/src/wrapper/lnxwrp_fm.c
-@@ -0,0 +1,2908 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+/*
-+ @File lnxwrp_fm.c
-+ @Author Shlomi Gridish
-+ @Description FM Linux wrapper functions.
-+*/
-+
-+#include <linux/version.h>
-+#include <linux/slab.h>
-+#if defined(CONFIG_MODVERSIONS) && !defined(MODVERSIONS)
-+#define MODVERSIONS
-+#endif
-+#ifdef MODVERSIONS
-+#include <config/modversions.h>
-+#endif /* MODVERSIONS */
-+#include <linux/kernel.h>
-+#include <linux/module.h>
-+#include <linux/fs.h>
-+#include <linux/cdev.h>
-+#include <linux/device.h>
-+#include <linux/irq.h>
-+#include <linux/interrupt.h>
-+#include <linux/io.h>
-+#include <linux/ioport.h>
-+#include <linux/of_platform.h>
-+#include <linux/of_address.h>
-+#include <linux/of_irq.h>
-+#include <linux/clk.h>
-+#include <asm/uaccess.h>
-+#include <asm/errno.h>
-+#ifndef CONFIG_FMAN_ARM
-+#include <sysdev/fsl_soc.h>
-+#include <linux/fsl/guts.h>
-+#include <linux/fsl/svr.h>
-+#endif
-+#include <linux/stat.h> /* For file access mask */
-+#include <linux/skbuff.h>
-+#include <linux/proc_fs.h>
-+
-+/* NetCommSw Headers --------------- */
-+#include "std_ext.h"
-+#include "error_ext.h"
-+#include "sprint_ext.h"
-+#include "debug_ext.h"
-+#include "sys_io_ext.h"
-+
-+#include "fm_ioctls.h"
-+
-+#include "lnxwrp_fm.h"
-+#include "lnxwrp_resources.h"
-+#include "lnxwrp_sysfs_fm.h"
-+#include "lnxwrp_sysfs_fm_port.h"
-+#include "lnxwrp_exp_sym.h"
-+#include "fm_common.h"
-+#include "../../sdk_fman/Peripherals/FM/fm.h"
-+#define __ERR_MODULE__ MODULE_FM
-+
-+extern struct device_node *GetFmPortAdvArgsDevTreeNode (struct device_node *fm_node,
-+ e_FmPortType portType,
-+ uint8_t portId);
-+
-+#define PROC_PRINT(args...) offset += sprintf(buf+offset,args)
-+
-+#define ADD_ADV_CONFIG_NO_RET(_func, _param) \
-+ do { \
-+ if (i<max){ \
-+ p_Entry = &p_Entrys[i]; \
-+ p_Entry->p_Function = _func; \
-+ _param \
-+ i++; \
-+ } \
-+ else \
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE,\
-+ ("Number of advanced-configuration entries exceeded"));\
-+ } while (0)
-+
-+/* Bootarg used to override the Kconfig FSL_FM_MAX_FRAME_SIZE value */
-+#define FSL_FM_MAX_FRM_BOOTARG "fsl_fm_max_frm"
-+
-+/* Bootarg used to override FSL_FM_RX_EXTRA_HEADROOM Kconfig value */
-+#define FSL_FM_RX_EXTRA_HEADROOM_BOOTARG "fsl_fm_rx_extra_headroom"
-+
-+/* Minimum and maximum value for the fsl_fm_rx_extra_headroom bootarg */
-+#define FSL_FM_RX_EXTRA_HEADROOM_MIN 16
-+#define FSL_FM_RX_EXTRA_HEADROOM_MAX 384
-+
-+#define FSL_FM_PAUSE_TIME_ENABLE 0xf000
-+#define FSL_FM_PAUSE_TIME_DISABLE 0
-+#define FSL_FM_PAUSE_THRESH_DEFAULT 0
-+
-+/*
-+ * Max frame size, across all interfaces.
-+ * Configurable from Kconfig or bootargs, to avoid allocating
-+ * oversized (socket) buffers when not using jumbo frames.
-+ * Must be large enough to accommodate the network MTU, but small enough
-+ * to avoid wasting skb memory.
-+ *
-+ * Could be overridden once, at boot-time, via the
-+ * fm_set_max_frm() callback.
-+ */
-+int fsl_fm_max_frm = CONFIG_FSL_FM_MAX_FRAME_SIZE;
-+
-+/*
-+ * Extra headroom for Rx buffers.
-+ * FMan is instructed to allocate, on the Rx path, this amount of
-+ * space at the beginning of a data buffer, beside the DPA private
-+ * data area and the IC fields.
-+ * Does not impact Tx buffer layout.
-+ *
-+ * Configurable from Kconfig or bootargs. Zero by default, it's needed
-+ * on particular forwarding scenarios that add extra headers to the
-+ * forwarded frame.
-+ */
-+int fsl_fm_rx_extra_headroom = CONFIG_FSL_FM_RX_EXTRA_HEADROOM;
-+
-+#ifdef CONFIG_FMAN_PFC
-+static int fsl_fm_pfc_quanta[] = {
-+ CONFIG_FMAN_PFC_QUANTA_0,
-+ CONFIG_FMAN_PFC_QUANTA_1,
-+ CONFIG_FMAN_PFC_QUANTA_2,
-+ CONFIG_FMAN_PFC_QUANTA_3
-+};
-+#endif
-+
-+static t_LnxWrpFm lnxWrpFm;
-+
-+int fm_get_max_frm()
-+{
-+ return fsl_fm_max_frm;
-+}
-+EXPORT_SYMBOL(fm_get_max_frm);
-+
-+int fm_get_rx_extra_headroom()
-+{
-+ return ALIGN(fsl_fm_rx_extra_headroom, 16);
-+}
-+EXPORT_SYMBOL(fm_get_rx_extra_headroom);
-+
-+static int __init fm_set_max_frm(char *str)
-+{
-+ int ret = 0;
-+
-+ ret = get_option(&str, &fsl_fm_max_frm);
-+ if (ret != 1) {
-+ /*
-+ * This will only work if CONFIG_EARLY_PRINTK is compiled in,
-+ * and something like "earlyprintk=serial,uart0,115200" is
-+ * specified in the bootargs
-+ */
-+ printk(KERN_WARNING "No suitable %s=<int> prop in bootargs; "
-+ "will use the default FSL_FM_MAX_FRAME_SIZE (%d) "
-+ "from Kconfig.\n", FSL_FM_MAX_FRM_BOOTARG,
-+ CONFIG_FSL_FM_MAX_FRAME_SIZE);
-+
-+ fsl_fm_max_frm = CONFIG_FSL_FM_MAX_FRAME_SIZE;
-+ return 1;
-+ }
-+
-+ /* Don't allow invalid bootargs; fallback to the Kconfig value */
-+ if (fsl_fm_max_frm < 64 || fsl_fm_max_frm > 9600) {
-+ printk(KERN_WARNING "Invalid %s=%d in bootargs, valid range is "
-+ "64-9600. Falling back to the FSL_FM_MAX_FRAME_SIZE (%d) "
-+ "from Kconfig.\n",
-+ FSL_FM_MAX_FRM_BOOTARG, fsl_fm_max_frm,
-+ CONFIG_FSL_FM_MAX_FRAME_SIZE);
-+
-+ fsl_fm_max_frm = CONFIG_FSL_FM_MAX_FRAME_SIZE;
-+ return 1;
-+ }
-+
-+ printk(KERN_INFO "Using fsl_fm_max_frm=%d from bootargs\n",
-+ fsl_fm_max_frm);
-+ return 0;
-+}
-+early_param(FSL_FM_MAX_FRM_BOOTARG, fm_set_max_frm);
-+
-+static int __init fm_set_rx_extra_headroom(char *str)
-+{
-+ int ret;
-+
-+ ret = get_option(&str, &fsl_fm_rx_extra_headroom);
-+
-+ if (ret != 1) {
-+ printk(KERN_WARNING "No suitable %s=<int> prop in bootargs; "
-+ "will use the default FSL_FM_RX_EXTRA_HEADROOM (%d) "
-+ "from Kconfig.\n", FSL_FM_RX_EXTRA_HEADROOM_BOOTARG,
-+ CONFIG_FSL_FM_RX_EXTRA_HEADROOM);
-+ fsl_fm_rx_extra_headroom = CONFIG_FSL_FM_RX_EXTRA_HEADROOM;
-+
-+ return 1;
-+ }
-+
-+ if (fsl_fm_rx_extra_headroom < FSL_FM_RX_EXTRA_HEADROOM_MIN ||
-+ fsl_fm_rx_extra_headroom > FSL_FM_RX_EXTRA_HEADROOM_MAX) {
-+ printk(KERN_WARNING "Invalid value for %s=%d prop in "
-+ "bootargs; will use the default "
-+ "FSL_FM_RX_EXTRA_HEADROOM (%d) from Kconfig.\n",
-+ FSL_FM_RX_EXTRA_HEADROOM_BOOTARG,
-+ fsl_fm_rx_extra_headroom,
-+ CONFIG_FSL_FM_RX_EXTRA_HEADROOM);
-+ fsl_fm_rx_extra_headroom = CONFIG_FSL_FM_RX_EXTRA_HEADROOM;
-+ }
-+
-+ printk(KERN_INFO "Using fsl_fm_rx_extra_headroom=%d from bootargs\n",
-+ fsl_fm_rx_extra_headroom);
-+
-+ return 0;
-+}
-+early_param(FSL_FM_RX_EXTRA_HEADROOM_BOOTARG, fm_set_rx_extra_headroom);
-+
-+static irqreturn_t fm_irq(int irq, void *_dev)
-+{
-+ t_LnxWrpFmDev *p_LnxWrpFmDev = (t_LnxWrpFmDev *)_dev;
-+#ifdef CONFIG_PM_SLEEP
-+ t_Fm *p_Fm = (t_Fm*)p_LnxWrpFmDev->h_Dev;
-+#endif
-+ if (!p_LnxWrpFmDev || !p_LnxWrpFmDev->h_Dev)
-+ return IRQ_NONE;
-+
-+#ifdef CONFIG_PM_SLEEP
-+ if (fman_get_normal_pending(p_Fm->p_FmFpmRegs) & INTR_EN_WAKEUP)
-+ {
-+ pm_wakeup_event(p_LnxWrpFmDev->dev, 200);
-+ }
-+#endif
-+ FM_EventIsr(p_LnxWrpFmDev->h_Dev);
-+ return IRQ_HANDLED;
-+}
-+
-+static irqreturn_t fm_err_irq(int irq, void *_dev)
-+{
-+ t_LnxWrpFmDev *p_LnxWrpFmDev = (t_LnxWrpFmDev *)_dev;
-+
-+ if (!p_LnxWrpFmDev || !p_LnxWrpFmDev->h_Dev)
-+ return IRQ_NONE;
-+
-+ if (FM_ErrorIsr(p_LnxWrpFmDev->h_Dev) == E_OK)
-+ return IRQ_HANDLED;
-+
-+ return IRQ_NONE;
-+}
-+
-+/* used to protect FMD/LLD from concurrent calls in functions fm_mutex_lock / fm_mutex_unlock */
-+static struct mutex lnxwrp_mutex;
-+
-+static t_LnxWrpFmDev * CreateFmDev(uint8_t id)
-+{
-+ t_LnxWrpFmDev *p_LnxWrpFmDev;
-+ int j;
-+
-+ p_LnxWrpFmDev = (t_LnxWrpFmDev *)XX_Malloc(sizeof(t_LnxWrpFmDev));
-+ if (!p_LnxWrpFmDev)
-+ {
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, NO_MSG);
-+ return NULL;
-+ }
-+
-+ memset(p_LnxWrpFmDev, 0, sizeof(t_LnxWrpFmDev));
-+ p_LnxWrpFmDev->fmDevSettings.advConfig = (t_SysObjectAdvConfigEntry*)XX_Malloc(FM_MAX_NUM_OF_ADV_SETTINGS*sizeof(t_SysObjectAdvConfigEntry));
-+ memset(p_LnxWrpFmDev->fmDevSettings.advConfig, 0, (FM_MAX_NUM_OF_ADV_SETTINGS*sizeof(t_SysObjectAdvConfigEntry)));
-+ p_LnxWrpFmDev->fmPcdDevSettings.advConfig = (t_SysObjectAdvConfigEntry*)XX_Malloc(FM_MAX_NUM_OF_ADV_SETTINGS*sizeof(t_SysObjectAdvConfigEntry));
-+ memset(p_LnxWrpFmDev->fmPcdDevSettings.advConfig, 0, (FM_MAX_NUM_OF_ADV_SETTINGS*sizeof(t_SysObjectAdvConfigEntry)));
-+ p_LnxWrpFmDev->hcPort.settings.advConfig = (t_SysObjectAdvConfigEntry*)XX_Malloc(FM_MAX_NUM_OF_ADV_SETTINGS*sizeof(t_SysObjectAdvConfigEntry));
-+ memset(p_LnxWrpFmDev->hcPort.settings.advConfig, 0, (FM_MAX_NUM_OF_ADV_SETTINGS*sizeof(t_SysObjectAdvConfigEntry)));
-+ for (j=0; j<FM_MAX_NUM_OF_RX_PORTS; j++)
-+ {
-+ p_LnxWrpFmDev->rxPorts[j].settings.advConfig = (t_SysObjectAdvConfigEntry*)XX_Malloc(FM_MAX_NUM_OF_ADV_SETTINGS*sizeof(t_SysObjectAdvConfigEntry));
-+ memset(p_LnxWrpFmDev->rxPorts[j].settings.advConfig, 0, (FM_MAX_NUM_OF_ADV_SETTINGS*sizeof(t_SysObjectAdvConfigEntry)));
-+ }
-+ for (j=0; j<FM_MAX_NUM_OF_TX_PORTS; j++)
-+ {
-+ p_LnxWrpFmDev->txPorts[j].settings.advConfig = (t_SysObjectAdvConfigEntry*)XX_Malloc(FM_MAX_NUM_OF_ADV_SETTINGS*sizeof(t_SysObjectAdvConfigEntry));
-+ memset(p_LnxWrpFmDev->txPorts[j].settings.advConfig, 0, (FM_MAX_NUM_OF_ADV_SETTINGS*sizeof(t_SysObjectAdvConfigEntry)));
-+ }
-+ for (j=0; j<FM_MAX_NUM_OF_OH_PORTS-1; j++)
-+ {
-+ p_LnxWrpFmDev->opPorts[j].settings.advConfig = (t_SysObjectAdvConfigEntry*)XX_Malloc(FM_MAX_NUM_OF_ADV_SETTINGS*sizeof(t_SysObjectAdvConfigEntry));
-+ memset(p_LnxWrpFmDev->opPorts[j].settings.advConfig, 0, (FM_MAX_NUM_OF_ADV_SETTINGS*sizeof(t_SysObjectAdvConfigEntry)));
-+ }
-+
-+ return p_LnxWrpFmDev;
-+}
-+
-+static void DestroyFmDev(t_LnxWrpFmDev *p_LnxWrpFmDev)
-+{
-+ int j;
-+
-+ for (j=0; j<FM_MAX_NUM_OF_OH_PORTS-1; j++)
-+ if (p_LnxWrpFmDev->opPorts[j].settings.advConfig)
-+ XX_Free(p_LnxWrpFmDev->opPorts[j].settings.advConfig);
-+ for (j=0; j<FM_MAX_NUM_OF_TX_PORTS; j++)
-+ if (p_LnxWrpFmDev->txPorts[j].settings.advConfig)
-+ XX_Free(p_LnxWrpFmDev->txPorts[j].settings.advConfig);
-+ for (j=0; j<FM_MAX_NUM_OF_RX_PORTS; j++)
-+ if (p_LnxWrpFmDev->rxPorts[j].settings.advConfig)
-+ XX_Free(p_LnxWrpFmDev->rxPorts[j].settings.advConfig);
-+ if (p_LnxWrpFmDev->hcPort.settings.advConfig)
-+ XX_Free(p_LnxWrpFmDev->hcPort.settings.advConfig);
-+ if (p_LnxWrpFmDev->fmPcdDevSettings.advConfig)
-+ XX_Free(p_LnxWrpFmDev->fmPcdDevSettings.advConfig);
-+ if (p_LnxWrpFmDev->fmDevSettings.advConfig)
-+ XX_Free(p_LnxWrpFmDev->fmDevSettings.advConfig);
-+
-+ XX_Free(p_LnxWrpFmDev);
-+}
-+
-+static t_Error FillRestFmInfo(t_LnxWrpFmDev *p_LnxWrpFmDev)
-+{
-+#define FM_BMI_PPIDS_OFFSET 0x00080304
-+#define FM_DMA_PLR_OFFSET 0x000c2060
-+#define FM_FPM_IP_REV_1_OFFSET 0x000c30c4
-+#define DMA_HIGH_LIODN_MASK 0x0FFF0000
-+#define DMA_LOW_LIODN_MASK 0x00000FFF
-+#define DMA_LIODN_SHIFT 16
-+
-+typedef _Packed struct {
-+ uint32_t plr[32];
-+} _PackedType t_Plr;
-+
-+typedef _Packed struct {
-+ volatile uint32_t fmbm_ppid[63];
-+} _PackedType t_Ppids;
-+
-+ t_Plr *p_Plr;
-+ t_Ppids *p_Ppids;
-+ int i,j;
-+ uint32_t fmRev;
-+
-+ static const uint8_t phys1GRxPortId[] = {0x8,0x9,0xa,0xb,0xc,0xd,0xe,0xf};
-+ static const uint8_t phys10GRxPortId[] = {0x10,0x11};
-+#if (DPAA_VERSION >= 11)
-+ static const uint8_t physOhPortId[] = {/* 0x1, */0x2,0x3,0x4,0x5,0x6,0x7};
-+#else
-+ static const uint8_t physOhPortId[] = {0x1,0x2,0x3,0x4,0x5,0x6,0x7};
-+#endif
-+ static const uint8_t phys1GTxPortId[] = {0x28,0x29,0x2a,0x2b,0x2c,0x2d,0x2e,0x2f};
-+ static const uint8_t phys10GTxPortId[] = {0x30,0x31};
-+
-+ fmRev = (uint32_t)(*((volatile uint32_t *)UINT_TO_PTR(p_LnxWrpFmDev->fmBaseAddr+FM_FPM_IP_REV_1_OFFSET)));
-+ fmRev &= 0xffff;
-+
-+ p_Plr = (t_Plr *)UINT_TO_PTR(p_LnxWrpFmDev->fmBaseAddr+FM_DMA_PLR_OFFSET);
-+#ifdef MODULE
-+ for (i=0;i<FM_MAX_NUM_OF_PARTITIONS/2;i++)
-+ p_Plr->plr[i] = 0;
-+#endif /* MODULE */
-+
-+ for (i=0; i<FM_MAX_NUM_OF_PARTITIONS; i++)
-+ {
-+ uint16_t liodnBase = (uint16_t)((i%2) ?
-+ (p_Plr->plr[i/2] & DMA_LOW_LIODN_MASK) :
-+ ((p_Plr->plr[i/2] & DMA_HIGH_LIODN_MASK) >> DMA_LIODN_SHIFT));
-+#ifdef FM_PARTITION_ARRAY
-+ /* TODO: this was .liodnPerPartition[i] = liodnBase; is the index meaning the same? */
-+ p_LnxWrpFmDev->fmDevSettings.param.liodnBasePerPort[i] = liodnBase;
-+#endif /* FM_PARTITION_ARRAY */
-+
-+ if ((i >= phys1GRxPortId[0]) &&
-+ (i <= phys1GRxPortId[FM_MAX_NUM_OF_1G_RX_PORTS-1]))
-+ {
-+ for (j=0; j<ARRAY_SIZE(phys1GRxPortId); j++)
-+ if (phys1GRxPortId[j] == i)
-+ break;
-+ ASSERT_COND(j<ARRAY_SIZE(phys1GRxPortId));
-+ p_LnxWrpFmDev->rxPorts[j].settings.param.liodnBase = liodnBase;
-+ }
-+ else if (FM_MAX_NUM_OF_10G_RX_PORTS &&
-+ (i >= phys10GRxPortId[0]) &&
-+ (i <= phys10GRxPortId[FM_MAX_NUM_OF_10G_RX_PORTS-1]))
-+ {
-+ for (j=0; j<ARRAY_SIZE(phys10GRxPortId); j++)
-+ if (phys10GRxPortId[j] == i)
-+ break;
-+ ASSERT_COND(j<ARRAY_SIZE(phys10GRxPortId));
-+ p_LnxWrpFmDev->rxPorts[FM_MAX_NUM_OF_1G_RX_PORTS+j].settings.param.liodnBase = liodnBase;
-+ }
-+ else if ((i >= physOhPortId[0]) &&
-+ (i <= physOhPortId[FM_MAX_NUM_OF_OH_PORTS-1]))
-+ {
-+ for (j=0; j<ARRAY_SIZE(physOhPortId); j++)
-+ if (physOhPortId[j] == i)
-+ break;
-+ ASSERT_COND(j<ARRAY_SIZE(physOhPortId));
-+ if (j == 0)
-+ p_LnxWrpFmDev->hcPort.settings.param.liodnBase = liodnBase;
-+ else
-+ p_LnxWrpFmDev->opPorts[j - 1].settings.param.liodnBase = liodnBase;
-+ }
-+ else if ((i >= phys1GTxPortId[0]) &&
-+ (i <= phys1GTxPortId[FM_MAX_NUM_OF_1G_TX_PORTS-1]))
-+ {
-+ for (j=0; j<ARRAY_SIZE(phys1GTxPortId); j++)
-+ if (phys1GTxPortId[j] == i)
-+ break;
-+ ASSERT_COND(j<ARRAY_SIZE(phys1GTxPortId));
-+ p_LnxWrpFmDev->txPorts[j].settings.param.liodnBase = liodnBase;
-+ }
-+ else if (FM_MAX_NUM_OF_10G_TX_PORTS &&
-+ (i >= phys10GTxPortId[0]) &&
-+ (i <= phys10GTxPortId[FM_MAX_NUM_OF_10G_TX_PORTS-1]))
-+ {
-+ for (j=0; j<ARRAY_SIZE(phys10GTxPortId); j++)
-+ if (phys10GTxPortId[j] == i)
-+ break;
-+ ASSERT_COND(j<ARRAY_SIZE(phys10GTxPortId));
-+ p_LnxWrpFmDev->txPorts[FM_MAX_NUM_OF_1G_TX_PORTS+j].settings.param.liodnBase = liodnBase;
-+ }
-+ }
-+
-+ p_Ppids = (t_Ppids *)UINT_TO_PTR(p_LnxWrpFmDev->fmBaseAddr+FM_BMI_PPIDS_OFFSET);
-+
-+ for (i=0; i<FM_MAX_NUM_OF_1G_RX_PORTS; i++)
-+ p_LnxWrpFmDev->rxPorts[i].settings.param.specificParams.rxParams.liodnOffset =
-+ p_Ppids->fmbm_ppid[phys1GRxPortId[i]-1];
-+
-+ for (i=0; i<FM_MAX_NUM_OF_10G_RX_PORTS; i++)
-+ p_LnxWrpFmDev->rxPorts[FM_MAX_NUM_OF_1G_RX_PORTS+i].settings.param.specificParams.rxParams.liodnOffset =
-+ p_Ppids->fmbm_ppid[phys10GRxPortId[i]-1];
-+
-+ return E_OK;
-+}
-+
-+/* Structure that defines QE firmware binary files.
-+ *
-+ * See Documentation/powerpc/qe_firmware.txt for a description of these
-+ * fields.
-+ */
-+struct qe_firmware {
-+ struct qe_header {
-+ __be32 length; /* Length of the entire structure, in bytes */
-+ u8 magic[3]; /* Set to { 'Q', 'E', 'F' } */
-+ u8 version; /* Version of this layout. First ver is '1' */
-+ } header;
-+ u8 id[62]; /* Null-terminated identifier string */
-+ u8 split; /* 0 = shared I-RAM, 1 = split I-RAM */
-+ u8 count; /* Number of microcode[] structures */
-+ struct {
-+ __be16 model; /* The SOC model */
-+ u8 major; /* The SOC revision major */
-+ u8 minor; /* The SOC revision minor */
-+ } __attribute__ ((packed)) soc;
-+ u8 padding[4]; /* Reserved, for alignment */
-+ __be64 extended_modes; /* Extended modes */
-+ __be32 vtraps[8]; /* Virtual trap addresses */
-+ u8 reserved[4]; /* Reserved, for future expansion */
-+ struct qe_microcode {
-+ u8 id[32]; /* Null-terminated identifier */
-+ __be32 traps[16]; /* Trap addresses, 0 == ignore */
-+ __be32 eccr; /* The value for the ECCR register */
-+ __be32 iram_offset; /* Offset into I-RAM for the code */
-+ __be32 count; /* Number of 32-bit words of the code */
-+ __be32 code_offset; /* Offset of the actual microcode */
-+ u8 major; /* The microcode version major */
-+ u8 minor; /* The microcode version minor */
-+ u8 revision; /* The microcode version revision */
-+ u8 padding; /* Reserved, for alignment */
-+ u8 reserved[4]; /* Reserved, for future expansion */
-+ } __attribute__ ((packed)) microcode[1];
-+ /* All microcode binaries should be located here */
-+ /* CRC32 should be located here, after the microcode binaries */
-+} __attribute__ ((packed));
-+
-+
-+/**
-+ * FindFmanMicrocode - find the Fman microcode
-+ *
-+ * This function returns a pointer to the QE Firmware blob that holds
-+ * the Fman microcode. We use the QE Firmware structure because Fman microcode
-+ * is similar to QE microcode, so there's no point in defining a new layout.
-+ *
-+ * Current versions of U-Boot embed the Fman firmware into the device tree,
-+ * so we check for that first. Each Fman node in the device tree contains a
-+ * node or a pointer to node that holds the firmware. Technically, we should
-+ * be fetching the firmware node for the current Fman, but we don't have that
-+ * information any more, so we assume that there is only one firmware node in
-+ * the device tree, and that all Fmen use the same firmware.
-+ */
-+static const struct qe_firmware *FindFmanMicrocode(void)
-+{
-+ static const struct qe_firmware *P4080_UCPatch;
-+ struct device_node *np;
-+
-+ if (P4080_UCPatch)
-+ return P4080_UCPatch;
-+
-+ /* The firmware should be inside the device tree. */
-+ np = of_find_compatible_node(NULL, NULL, "fsl,fman-firmware");
-+ if (np) {
-+ P4080_UCPatch = of_get_property(np, "fsl,firmware", NULL);
-+ of_node_put(np);
-+ if (P4080_UCPatch)
-+ return P4080_UCPatch;
-+ else
-+ REPORT_ERROR(WARNING, E_NOT_FOUND, ("firmware node is incomplete"));
-+ }
-+
-+ /* Returning NULL here forces the reuse of the IRAM content */
-+ return NULL;
-+}
-+#define SVR_SECURITY_MASK 0x00080000
-+#define SVR_PERSONALITY_MASK 0x0000FF00
-+#define SVR_VER_IGNORE_MASK (SVR_SECURITY_MASK | SVR_PERSONALITY_MASK)
-+#define SVR_B4860_REV1_VALUE 0x86800010
-+#define SVR_B4860_REV2_VALUE 0x86800020
-+#define SVR_T4240_VALUE 0x82400000
-+#define SVR_T4120_VALUE 0x82400100
-+#define SVR_T4160_VALUE 0x82410000
-+#define SVR_T4080_VALUE 0x82410200
-+#define SVR_T4_DEVICE_ID 0x82400000
-+#define SVR_DEVICE_ID_MASK 0xFFF00000
-+
-+#define OF_DEV_ID_NUM 2 /* one used, another one zeroed */
-+
-+/* searches for a subnode with the given name/compatible */
-+static bool HasFmPcdOfNode(struct device_node *fm_node,
-+ struct of_device_id *ids,
-+ const char *name,
-+ const char *compatible)
-+{
-+ struct device_node *dev_node;
-+ bool ret = false;
-+
-+ memset(ids, 0, OF_DEV_ID_NUM*sizeof(struct of_device_id));
-+ if (WARN_ON(strlen(name) >= sizeof(ids[0].name)))
-+ return false;
-+ strcpy(ids[0].name, name);
-+ if (WARN_ON(strlen(compatible) >= sizeof(ids[0].compatible)))
-+ return false;
-+ strcpy(ids[0].compatible, compatible);
-+ for_each_child_of_node(fm_node, dev_node)
-+ if (of_match_node(ids, dev_node) != NULL)
-+ ret = true;
-+ return ret;
-+}
-+
-+static t_LnxWrpFmDev * ReadFmDevTreeNode (struct platform_device *of_dev)
-+{
-+ t_LnxWrpFmDev *p_LnxWrpFmDev;
-+ struct device_node *fm_node, *dev_node;
-+ struct of_device_id ids[OF_DEV_ID_NUM];
-+ struct resource res;
-+ struct clk *clk;
-+ u32 clk_rate;
-+ const uint32_t *uint32_prop;
-+ int _errno=0, lenp;
-+ uint32_t tmp_prop;
-+
-+ fm_node = of_node_get(of_dev->dev.of_node);
-+
-+ uint32_prop = (uint32_t *)of_get_property(fm_node, "cell-index", &lenp);
-+ if (unlikely(uint32_prop == NULL)) {
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE, ("of_get_property(%s, cell-index) failed", fm_node->full_name));
-+ return NULL;
-+ }
-+ tmp_prop = be32_to_cpu(*uint32_prop);
-+
-+ if (WARN_ON(lenp != sizeof(uint32_t)))
-+ return NULL;
-+
-+ if (tmp_prop > INTG_MAX_NUM_OF_FM) {
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE, ("fm id!"));
-+ return NULL;
-+ }
-+ p_LnxWrpFmDev = CreateFmDev(tmp_prop);
-+ if (!p_LnxWrpFmDev) {
-+ REPORT_ERROR(MAJOR, E_NULL_POINTER, NO_MSG);
-+ return NULL;
-+ }
-+ p_LnxWrpFmDev->dev = &of_dev->dev;
-+ p_LnxWrpFmDev->id = tmp_prop;
-+
-+ /* Get the FM interrupt */
-+ p_LnxWrpFmDev->irq = of_irq_to_resource(fm_node, 0, NULL);
-+ if (unlikely(p_LnxWrpFmDev->irq == /*NO_IRQ*/0)) {
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE, ("of_irq_to_resource() = %d", NO_IRQ));
-+ DestroyFmDev(p_LnxWrpFmDev);
-+ return NULL;
-+ }
-+
-+ /* Get the FM error interrupt */
-+ p_LnxWrpFmDev->err_irq = of_irq_to_resource(fm_node, 1, NULL);
-+
-+ if (unlikely(p_LnxWrpFmDev->err_irq == /*NO_IRQ*/0)) {
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE, ("of_irq_to_resource() = %d", NO_IRQ));
-+ DestroyFmDev(p_LnxWrpFmDev);
-+ return NULL;
-+ }
-+
-+ /* Get the FM address */
-+ _errno = of_address_to_resource(fm_node, 0, &res);
-+ if (unlikely(_errno < 0)) {
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE, ("of_address_to_resource() = %d", _errno));
-+ DestroyFmDev(p_LnxWrpFmDev);
-+ return NULL;
-+ }
-+
-+
-+ p_LnxWrpFmDev->fmBaseAddr = 0;
-+ p_LnxWrpFmDev->fmPhysBaseAddr = res.start;
-+ p_LnxWrpFmDev->fmMemSize = res.end + 1 - res.start;
-+
-+ clk = of_clk_get(fm_node, 0);
-+ if (IS_ERR(clk)) {
-+ dev_err(&of_dev->dev, "%s: Failed to get FM clock structure\n",
-+ __func__);
-+ of_node_put(fm_node);
-+ DestroyFmDev(p_LnxWrpFmDev);
-+ return NULL;
-+ }
-+
-+ clk_rate = clk_get_rate(clk);
-+ if (!clk_rate) {
-+ dev_err(&of_dev->dev, "%s: Failed to determine FM clock rate\n",
-+ __func__);
-+ of_node_put(fm_node);
-+ DestroyFmDev(p_LnxWrpFmDev);
-+ return NULL;
-+ }
-+
-+ p_LnxWrpFmDev->fmDevSettings.param.fmClkFreq = DIV_ROUND_UP(clk_rate, 1000000); /* In MHz, rounded */
-+ /* Get the MURAM base address and size */
-+ memset(ids, 0, sizeof(ids));
-+ if (WARN_ON(strlen("muram") >= sizeof(ids[0].name)))
-+ return NULL;
-+ strcpy(ids[0].name, "muram");
-+ if (WARN_ON(strlen("fsl,fman-muram") >= sizeof(ids[0].compatible)))
-+ return NULL;
-+ strcpy(ids[0].compatible, "fsl,fman-muram");
-+ for_each_child_of_node(fm_node, dev_node) {
-+ if (likely(of_match_node(ids, dev_node) != NULL)) {
-+ _errno = of_address_to_resource(dev_node, 0, &res);
-+ if (unlikely(_errno < 0)) {
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE, ("of_address_to_resource() = %d", _errno));
-+ DestroyFmDev(p_LnxWrpFmDev);
-+ return NULL;
-+ }
-+
-+ p_LnxWrpFmDev->fmMuramBaseAddr = 0;
-+ p_LnxWrpFmDev->fmMuramPhysBaseAddr = res.start;
-+ p_LnxWrpFmDev->fmMuramMemSize = res.end + 1 - res.start;
-+
-+#ifndef CONFIG_FMAN_ARM
-+ {
-+ uint32_t svr;
-+ svr = mfspr(SPRN_SVR);
-+
-+ if ((svr & ~SVR_VER_IGNORE_MASK) >= SVR_B4860_REV2_VALUE)
-+ p_LnxWrpFmDev->fmMuramMemSize = 0x80000;
-+ }
-+#endif
-+ }
-+ }
-+
-+ /* Get the RTC base address and size */
-+ memset(ids, 0, sizeof(ids));
-+ if (WARN_ON(strlen("ptp-timer") >= sizeof(ids[0].name)))
-+ return NULL;
-+ strcpy(ids[0].name, "ptp-timer");
-+ if (WARN_ON(strlen("fsl,fman-rtc") >= sizeof(ids[0].compatible)))
-+ return NULL;
-+ strcpy(ids[0].compatible, "fsl,fman-rtc");
-+ for_each_child_of_node(fm_node, dev_node) {
-+ if (likely(of_match_node(ids, dev_node) != NULL)) {
-+ _errno = of_address_to_resource(dev_node, 0, &res);
-+ if (unlikely(_errno < 0)) {
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE, ("of_address_to_resource() = %d", _errno));
-+ DestroyFmDev(p_LnxWrpFmDev);
-+ return NULL;
-+ }
-+
-+ p_LnxWrpFmDev->fmRtcBaseAddr = 0;
-+ p_LnxWrpFmDev->fmRtcPhysBaseAddr = res.start;
-+ p_LnxWrpFmDev->fmRtcMemSize = res.end + 1 - res.start;
-+ }
-+ }
-+
-+#if (DPAA_VERSION >= 11)
-+ /* Get the VSP base address */
-+ for_each_child_of_node(fm_node, dev_node) {
-+ if (of_device_is_compatible(dev_node, "fsl,fman-vsps")) {
-+ _errno = of_address_to_resource(dev_node, 0, &res);
-+ if (unlikely(_errno < 0)) {
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE, ("of_address_to_resource() = %d", _errno));
-+ DestroyFmDev(p_LnxWrpFmDev);
-+ return NULL;
-+ }
-+ p_LnxWrpFmDev->fmVspBaseAddr = 0;
-+ p_LnxWrpFmDev->fmVspPhysBaseAddr = res.start;
-+ p_LnxWrpFmDev->fmVspMemSize = res.end + 1 - res.start;
-+ }
-+ }
-+#endif
-+
-+ /* Get all PCD nodes */
-+ p_LnxWrpFmDev->prsActive = HasFmPcdOfNode(fm_node, ids, "parser", "fsl,fman-parser");
-+ p_LnxWrpFmDev->kgActive = HasFmPcdOfNode(fm_node, ids, "keygen", "fsl,fman-keygen");
-+ p_LnxWrpFmDev->ccActive = HasFmPcdOfNode(fm_node, ids, "cc", "fsl,fman-cc");
-+ p_LnxWrpFmDev->plcrActive = HasFmPcdOfNode(fm_node, ids, "policer", "fsl,fman-policer");
-+
-+ if (p_LnxWrpFmDev->prsActive || p_LnxWrpFmDev->kgActive ||
-+ p_LnxWrpFmDev->ccActive || p_LnxWrpFmDev->plcrActive)
-+ p_LnxWrpFmDev->pcdActive = TRUE;
-+
-+ if (p_LnxWrpFmDev->pcdActive)
-+ {
-+ const char *str_prop = (char *)of_get_property(fm_node, "fsl,default-pcd", &lenp);
-+ if (str_prop) {
-+ if (strncmp(str_prop, "3-tuple", strlen("3-tuple")) == 0)
-+ p_LnxWrpFmDev->defPcd = e_FM_PCD_3_TUPLE;
-+ }
-+ else
-+ p_LnxWrpFmDev->defPcd = e_NO_PCD;
-+ }
-+
-+ of_node_put(fm_node);
-+
-+ p_LnxWrpFmDev->hcCh =
-+ qman_affine_channel(cpumask_first(qman_affine_cpus()));
-+
-+ p_LnxWrpFmDev->active = TRUE;
-+
-+ return p_LnxWrpFmDev;
-+}
-+
-+struct device_node *GetFmAdvArgsDevTreeNode (uint8_t fmIndx)
-+{
-+ struct device_node *dev_node;
-+ const uint32_t *uint32_prop;
-+ int lenp;
-+ uint32_t tmp_prop;
-+
-+ for_each_compatible_node(dev_node, NULL, "fsl,fman-extended-args") {
-+ uint32_prop = (uint32_t *)of_get_property(dev_node, "cell-index", &lenp);
-+ if (unlikely(uint32_prop == NULL)) {
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE,
-+ ("of_get_property(%s, cell-index) failed",
-+ dev_node->full_name));
-+ return NULL;
-+ }
-+ tmp_prop = be32_to_cpu(*uint32_prop);
-+ if (WARN_ON(lenp != sizeof(uint32_t)))
-+ return NULL;
-+ if (tmp_prop > INTG_MAX_NUM_OF_FM) {
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE, ("fm id!"));
-+ return NULL;
-+ }
-+ if (fmIndx == tmp_prop)
-+ return dev_node;
-+ }
-+
-+ return NULL;
-+}
-+
-+static t_Error CheckNConfigFmAdvArgs (t_LnxWrpFmDev *p_LnxWrpFmDev)
-+{
-+ struct device_node *dev_node;
-+ t_Error err = E_INVALID_VALUE;
-+ const uint32_t *uint32_prop;
-+ const char *str_prop;
-+ int lenp;
-+ uint32_t tmp_prop;
-+
-+ dev_node = GetFmAdvArgsDevTreeNode(p_LnxWrpFmDev->id);
-+ if (!dev_node) /* no advance parameters for FMan */
-+ return E_OK;
-+
-+ str_prop = (char *)of_get_property(dev_node, "dma-aid-mode", &lenp);
-+ if (str_prop) {
-+ if (strcmp(str_prop, "port") == 0)
-+ err = FM_ConfigDmaAidMode(p_LnxWrpFmDev->h_Dev, e_FM_DMA_AID_OUT_PORT_ID);
-+ else if (strcmp(str_prop, "tnum") == 0)
-+ err = FM_ConfigDmaAidMode(p_LnxWrpFmDev->h_Dev, e_FM_DMA_AID_OUT_TNUM);
-+
-+ if (err != E_OK)
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+ }
-+
-+ uint32_prop = (uint32_t *)of_get_property(dev_node,
-+ "total-fifo-size", &lenp);
-+ if (uint32_prop) {
-+ tmp_prop = be32_to_cpu(*uint32_prop);
-+ if (WARN_ON(lenp != sizeof(uint32_t)))
-+ RETURN_ERROR(MINOR, E_INVALID_VALUE, NO_MSG);
-+
-+ if (FM_ConfigTotalFifoSize(p_LnxWrpFmDev->h_Dev,
-+ tmp_prop) != E_OK)
-+ RETURN_ERROR(MINOR, E_INVALID_VALUE, NO_MSG);
-+ }
-+
-+ uint32_prop = (uint32_t *)of_get_property(dev_node, "tnum-aging-period",
-+ &lenp);
-+ if (uint32_prop) {
-+ tmp_prop = be32_to_cpu(*uint32_prop);
-+ if (WARN_ON(lenp != sizeof(uint32_t)))
-+ RETURN_ERROR(MINOR, E_INVALID_VALUE, NO_MSG);
-+
-+ err = FM_ConfigTnumAgingPeriod(p_LnxWrpFmDev->h_Dev,
-+ (uint16_t)tmp_prop/*tnumAgingPeriod*/);
-+
-+ if (err != E_OK)
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+ }
-+
-+ of_node_put(dev_node);
-+
-+ return E_OK;
-+}
-+
-+static void LnxwrpFmDevExceptionsCb(t_Handle h_App, e_FmExceptions exception)
-+{
-+ t_LnxWrpFmDev *p_LnxWrpFmDev = (t_LnxWrpFmDev *)h_App;
-+
-+ ASSERT_COND(p_LnxWrpFmDev);
-+
-+ DBG(INFO, ("got fm exception %d", exception));
-+
-+ /* do nothing */
-+ UNUSED(exception);
-+}
-+
-+static void LnxwrpFmDevBusErrorCb(t_Handle h_App,
-+ e_FmPortType portType,
-+ uint8_t portId,
-+ uint64_t addr,
-+ uint8_t tnum,
-+ uint16_t liodn)
-+{
-+ t_LnxWrpFmDev *p_LnxWrpFmDev = (t_LnxWrpFmDev *)h_App;
-+
-+ ASSERT_COND(p_LnxWrpFmDev);
-+
-+ /* do nothing */
-+ UNUSED(portType);UNUSED(portId);UNUSED(addr);UNUSED(tnum);UNUSED(liodn);
-+}
-+
-+static t_Error ConfigureFmDev(t_LnxWrpFmDev *p_LnxWrpFmDev)
-+{
-+ struct resource *dev_res;
-+ int _errno;
-+
-+ if (!p_LnxWrpFmDev->active)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("FM not configured!!!"));
-+
-+#ifndef MODULE
-+ _errno = can_request_irq(p_LnxWrpFmDev->irq, 0);
-+ if (unlikely(_errno < 0))
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("can_request_irq() = %d", _errno));
-+#endif
-+ _errno = devm_request_irq(p_LnxWrpFmDev->dev, p_LnxWrpFmDev->irq, fm_irq, 0, "fman", p_LnxWrpFmDev);
-+ if (unlikely(_errno < 0))
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("request_irq(%d) = %d", p_LnxWrpFmDev->irq, _errno));
-+
-+ enable_irq_wake(p_LnxWrpFmDev->irq);
-+
-+ if (p_LnxWrpFmDev->err_irq != 0) {
-+#ifndef MODULE
-+ _errno = can_request_irq(p_LnxWrpFmDev->err_irq, 0);
-+ if (unlikely(_errno < 0))
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("can_request_irq() = %d", _errno));
-+#endif
-+ _errno = devm_request_irq(p_LnxWrpFmDev->dev, p_LnxWrpFmDev->err_irq, fm_err_irq, IRQF_SHARED, "fman-err", p_LnxWrpFmDev);
-+ if (unlikely(_errno < 0))
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("request_irq(%d) = %d", p_LnxWrpFmDev->err_irq, _errno));
-+
-+ enable_irq_wake(p_LnxWrpFmDev->err_irq);
-+ }
-+
-+ p_LnxWrpFmDev->res = devm_request_mem_region(p_LnxWrpFmDev->dev, p_LnxWrpFmDev->fmPhysBaseAddr, p_LnxWrpFmDev->fmMemSize, "fman");
-+ if (unlikely(p_LnxWrpFmDev->res == NULL))
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("request_mem_region() failed"));
-+
-+ p_LnxWrpFmDev->fmBaseAddr = PTR_TO_UINT(devm_ioremap(p_LnxWrpFmDev->dev, p_LnxWrpFmDev->fmPhysBaseAddr, p_LnxWrpFmDev->fmMemSize));
-+ if (unlikely(p_LnxWrpFmDev->fmBaseAddr == 0))
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("devm_ioremap() failed"));
-+
-+ if (SYS_RegisterIoMap((uint64_t)p_LnxWrpFmDev->fmBaseAddr, (uint64_t)p_LnxWrpFmDev->fmPhysBaseAddr, p_LnxWrpFmDev->fmMemSize) != E_OK)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("FM memory map"));
-+
-+ dev_res = __devm_request_region(p_LnxWrpFmDev->dev, p_LnxWrpFmDev->res, p_LnxWrpFmDev->fmMuramPhysBaseAddr, p_LnxWrpFmDev->fmMuramMemSize, "fman-muram");
-+ if (unlikely(dev_res == NULL))
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("__devm_request_region() failed"));
-+
-+ p_LnxWrpFmDev->fmMuramBaseAddr = PTR_TO_UINT(devm_ioremap(p_LnxWrpFmDev->dev, p_LnxWrpFmDev->fmMuramPhysBaseAddr, p_LnxWrpFmDev->fmMuramMemSize));
-+ if (unlikely(p_LnxWrpFmDev->fmMuramBaseAddr == 0))
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("devm_ioremap() failed"));
-+
-+ if (SYS_RegisterIoMap((uint64_t)p_LnxWrpFmDev->fmMuramBaseAddr, (uint64_t)p_LnxWrpFmDev->fmMuramPhysBaseAddr, p_LnxWrpFmDev->fmMuramMemSize) != E_OK)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("FM MURAM memory map"));
-+
-+ if (p_LnxWrpFmDev->fmRtcPhysBaseAddr)
-+ {
-+ dev_res = __devm_request_region(p_LnxWrpFmDev->dev, p_LnxWrpFmDev->res, p_LnxWrpFmDev->fmRtcPhysBaseAddr, p_LnxWrpFmDev->fmRtcMemSize, "fman-rtc");
-+ if (unlikely(dev_res == NULL))
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("__devm_request_region() failed"));
-+
-+ p_LnxWrpFmDev->fmRtcBaseAddr = PTR_TO_UINT(devm_ioremap(p_LnxWrpFmDev->dev, p_LnxWrpFmDev->fmRtcPhysBaseAddr, p_LnxWrpFmDev->fmRtcMemSize));
-+ if (unlikely(p_LnxWrpFmDev->fmRtcBaseAddr == 0))
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("devm_ioremap() failed"));
-+
-+ if (SYS_RegisterIoMap((uint64_t)p_LnxWrpFmDev->fmRtcBaseAddr, (uint64_t)p_LnxWrpFmDev->fmRtcPhysBaseAddr, p_LnxWrpFmDev->fmRtcMemSize) != E_OK)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("FM-RTC memory map"));
-+ }
-+
-+#if (DPAA_VERSION >= 11)
-+ if (p_LnxWrpFmDev->fmVspPhysBaseAddr) {
-+ dev_res = __devm_request_region(p_LnxWrpFmDev->dev, p_LnxWrpFmDev->res, p_LnxWrpFmDev->fmVspPhysBaseAddr, p_LnxWrpFmDev->fmVspMemSize, "fman-vsp");
-+ if (unlikely(dev_res == NULL))
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("__devm_request_region() failed"));
-+
-+ p_LnxWrpFmDev->fmVspBaseAddr = PTR_TO_UINT(devm_ioremap(p_LnxWrpFmDev->dev, p_LnxWrpFmDev->fmVspPhysBaseAddr, p_LnxWrpFmDev->fmVspMemSize));
-+ if (unlikely(p_LnxWrpFmDev->fmVspBaseAddr == 0))
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("devm_ioremap() failed"));
-+ }
-+#endif
-+
-+ p_LnxWrpFmDev->fmDevSettings.param.baseAddr = p_LnxWrpFmDev->fmBaseAddr;
-+ p_LnxWrpFmDev->fmDevSettings.param.fmId = p_LnxWrpFmDev->id;
-+ p_LnxWrpFmDev->fmDevSettings.param.irq = NO_IRQ;
-+ p_LnxWrpFmDev->fmDevSettings.param.errIrq = NO_IRQ;
-+ p_LnxWrpFmDev->fmDevSettings.param.f_Exception = LnxwrpFmDevExceptionsCb;
-+ p_LnxWrpFmDev->fmDevSettings.param.f_BusError = LnxwrpFmDevBusErrorCb;
-+ p_LnxWrpFmDev->fmDevSettings.param.h_App = p_LnxWrpFmDev;
-+
-+ return FillRestFmInfo(p_LnxWrpFmDev);
-+}
-+
-+#ifndef CONFIG_FMAN_ARM
-+/*
-+ * Table for matching compatible strings, for device tree
-+ * guts node, for QorIQ SOCs.
-+ * "fsl,qoriq-device-config-2.0" corresponds to T4 & B4
-+ * SOCs. For the older SOCs "fsl,qoriq-device-config-1.0"
-+ * string would be used.
-+*/
-+static const struct of_device_id guts_device_ids[] = {
-+ { .compatible = "fsl,qoriq-device-config-1.0", },
-+ { .compatible = "fsl,qoriq-device-config-2.0", },
-+ {}
-+};
-+
-+static unsigned int get_rcwsr(int regnum)
-+{
-+ struct ccsr_guts __iomem *guts_regs = NULL;
-+ struct device_node *guts_node;
-+
-+ guts_node = of_find_matching_node(NULL, guts_device_ids);
-+ if (!guts_node) {
-+ pr_err("could not find GUTS node\n");
-+ return 0;
-+ }
-+ guts_regs = of_iomap(guts_node, 0);
-+ of_node_put(guts_node);
-+ if (!guts_regs) {
-+ pr_err("ioremap of GUTS node failed\n");
-+ return 0;
-+ }
-+
-+ return ioread32be(&guts_regs->rcwsr[regnum]);
-+}
-+
-+#define FMAN1_ALL_MACS_MASK 0xFCC00000
-+#define FMAN2_ALL_MACS_MASK 0x000FCC00
-+
-+/**
-+ * @Function ResetOnInitErrata_A007273
-+ *
-+ * @Description Workaround for Errata A-007273
-+ * This workaround is required to avoid a FMan hang during reset on initialization.
-+ * Enable all MACs in guts.devdisr2 register,
-+ * then perform a regular FMan reset and then restore MACs to their original state.
-+ *
-+ * @Param[in] h_Fm - FM module descriptor
-+ *
-+ * @Return None.
-+ */
-+void ResetOnInitErrata_A007273(t_Handle h_Fm)
-+{
-+ struct ccsr_guts __iomem *guts_regs = NULL;
-+ struct device_node *guts_node;
-+ u32 devdisr2, enableMacs;
-+
-+ /* Get guts registers */
-+ guts_node = of_find_matching_node(NULL, guts_device_ids);
-+ if (!guts_node) {
-+ pr_err("could not find GUTS node\n");
-+ return;
-+ }
-+ guts_regs = of_iomap(guts_node, 0);
-+ of_node_put(guts_node);
-+ if (!guts_regs) {
-+ pr_err("ioremap of GUTS node failed\n");
-+ return;
-+ }
-+
-+ /* Read current state */
-+ devdisr2 = ioread32be(&guts_regs->devdisr2);
-+
-+ if (FmGetId(h_Fm) == 0)
-+ enableMacs = devdisr2 & ~FMAN1_ALL_MACS_MASK;
-+ else
-+ enableMacs = devdisr2 & ~FMAN2_ALL_MACS_MASK;
-+
-+ /* Enable all MACs */
-+ iowrite32be(enableMacs, &guts_regs->devdisr2);
-+
-+ /* Perform standard FMan reset */
-+ FmReset(h_Fm);
-+
-+ /* Restore devdisr2 value */
-+ iowrite32be(devdisr2, &guts_regs->devdisr2);
-+
-+ iounmap(guts_regs);
-+}
-+#endif
-+
-+static t_Error InitFmDev(t_LnxWrpFmDev *p_LnxWrpFmDev)
-+{
-+ const struct qe_firmware *fw;
-+
-+ if (!p_LnxWrpFmDev->active)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("FM not configured!!!"));
-+
-+ if ((p_LnxWrpFmDev->h_MuramDev = FM_MURAM_ConfigAndInit(p_LnxWrpFmDev->fmMuramBaseAddr, p_LnxWrpFmDev->fmMuramMemSize)) == NULL)
-+ RETURN_ERROR(MAJOR, E_INVALID_HANDLE, ("FM-MURAM!"));
-+
-+ /* Loading the fman-controller code */
-+ fw = FindFmanMicrocode();
-+
-+ if (!fw) {
-+ /* this forces the reuse of the current IRAM content */
-+ p_LnxWrpFmDev->fmDevSettings.param.firmware.size = 0;
-+ p_LnxWrpFmDev->fmDevSettings.param.firmware.p_Code = NULL;
-+ } else {
-+ p_LnxWrpFmDev->fmDevSettings.param.firmware.p_Code =
-+ (void *) fw + be32_to_cpu(fw->microcode[0].code_offset);
-+ p_LnxWrpFmDev->fmDevSettings.param.firmware.size =
-+ sizeof(u32) * be32_to_cpu(fw->microcode[0].count);
-+ DBG(INFO, ("Loading fman-controller code version %d.%d.%d",
-+ fw->microcode[0].major,
-+ fw->microcode[0].minor,
-+ fw->microcode[0].revision));
-+ }
-+
-+#ifdef CONFIG_FMAN_ARM
-+ { /* endianness adjustments: byteswap the ucode retrieved from the f/w blob */
-+ int i;
-+ int usz = p_LnxWrpFmDev->fmDevSettings.param.firmware.size;
-+ void * p_Code = p_LnxWrpFmDev->fmDevSettings.param.firmware.p_Code;
-+ u32 *dest = kzalloc(usz, GFP_KERNEL);
-+
-+ if (p_Code && dest)
-+ for(i=0; i < usz / 4; ++i)
-+ dest[i] = be32_to_cpu(((u32 *)p_Code)[i]);
-+
-+ p_LnxWrpFmDev->fmDevSettings.param.firmware.p_Code = dest;
-+ }
-+#endif
-+
-+ p_LnxWrpFmDev->fmDevSettings.param.h_FmMuram = p_LnxWrpFmDev->h_MuramDev;
-+
-+#if (DPAA_VERSION >= 11)
-+ if (p_LnxWrpFmDev->fmVspBaseAddr) {
-+ p_LnxWrpFmDev->fmDevSettings.param.vspBaseAddr = p_LnxWrpFmDev->fmVspBaseAddr;
-+ p_LnxWrpFmDev->fmDevSettings.param.partVSPBase = 0;
-+ p_LnxWrpFmDev->fmDevSettings.param.partNumOfVSPs = FM_VSP_MAX_NUM_OF_ENTRIES;
-+ }
-+#endif
-+
-+#ifdef CONFIG_FMAN_ARM
-+ p_LnxWrpFmDev->fmDevSettings.param.fmMacClkRatio = 1;
-+#else
-+ if(p_LnxWrpFmDev->fmDevSettings.param.fmId == 0)
-+ p_LnxWrpFmDev->fmDevSettings.param.fmMacClkRatio =
-+ !!(get_rcwsr(4) & 0x2); /* RCW[FM_MAC_RAT0] */
-+ else
-+ p_LnxWrpFmDev->fmDevSettings.param.fmMacClkRatio =
-+ !!(get_rcwsr(4) & 0x1); /* RCW[FM_MAC_RAT1] */
-+
-+ {
-+ /* T4 Devices ClkRatio is always 1 regardless of RCW[FM_MAC_RAT1] */
-+ uint32_t svr;
-+ svr = mfspr(SPRN_SVR);
-+
-+ if ((svr & SVR_DEVICE_ID_MASK) == SVR_T4_DEVICE_ID)
-+ p_LnxWrpFmDev->fmDevSettings.param.fmMacClkRatio = 1;
-+ }
-+#endif /* CONFIG_FMAN_ARM */
-+
-+ if ((p_LnxWrpFmDev->h_Dev = FM_Config(&p_LnxWrpFmDev->fmDevSettings.param)) == NULL)
-+ RETURN_ERROR(MAJOR, E_INVALID_HANDLE, ("FM"));
-+
-+
-+ if (FM_ConfigResetOnInit(p_LnxWrpFmDev->h_Dev, TRUE) != E_OK)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("FM"));
-+
-+#ifndef CONFIG_FMAN_ARM
-+#ifdef FM_HANG_AT_RESET_MAC_CLK_DISABLED_ERRATA_FMAN_A007273
-+ if (FM_ConfigResetOnInitOverrideCallback(p_LnxWrpFmDev->h_Dev, ResetOnInitErrata_A007273) != E_OK)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("FM"));
-+#endif /* FM_HANG_AT_RESET_MAC_CLK_DISABLED_ERRATA_FMAN_A007273 */
-+#endif /* CONFIG_FMAN_ARM */
-+
-+#ifdef CONFIG_FMAN_P1023
-+ if (FM_ConfigDmaAidOverride(p_LnxWrpFmDev->h_Dev, TRUE) != E_OK)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("FM"));
-+#endif
-+
-+
-+ CheckNConfigFmAdvArgs(p_LnxWrpFmDev);
-+
-+ if (FM_Init(p_LnxWrpFmDev->h_Dev) != E_OK)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("FM"));
-+
-+ /* TODO: Why we mask these interrupts? */
-+ if (p_LnxWrpFmDev->err_irq == 0) {
-+ FM_SetException(p_LnxWrpFmDev->h_Dev, e_FM_EX_DMA_BUS_ERROR,FALSE);
-+ FM_SetException(p_LnxWrpFmDev->h_Dev,e_FM_EX_DMA_READ_ECC,FALSE);
-+ FM_SetException(p_LnxWrpFmDev->h_Dev,e_FM_EX_DMA_SYSTEM_WRITE_ECC,FALSE);
-+ FM_SetException(p_LnxWrpFmDev->h_Dev,e_FM_EX_DMA_FM_WRITE_ECC,FALSE);
-+ FM_SetException(p_LnxWrpFmDev->h_Dev,e_FM_EX_DMA_SINGLE_PORT_ECC, FALSE);
-+ FM_SetException(p_LnxWrpFmDev->h_Dev,e_FM_EX_FPM_STALL_ON_TASKS , FALSE);
-+ FM_SetException(p_LnxWrpFmDev->h_Dev,e_FM_EX_FPM_SINGLE_ECC, FALSE);
-+ FM_SetException(p_LnxWrpFmDev->h_Dev,e_FM_EX_FPM_DOUBLE_ECC,FALSE);
-+ FM_SetException(p_LnxWrpFmDev->h_Dev,e_FM_EX_QMI_SINGLE_ECC, FALSE);
-+ FM_SetException(p_LnxWrpFmDev->h_Dev,e_FM_EX_QMI_DOUBLE_ECC,FALSE);
-+ FM_SetException(p_LnxWrpFmDev->h_Dev,e_FM_EX_QMI_DEQ_FROM_UNKNOWN_PORTID,FALSE);
-+ FM_SetException(p_LnxWrpFmDev->h_Dev,e_FM_EX_BMI_LIST_RAM_ECC,FALSE);
-+ FM_SetException(p_LnxWrpFmDev->h_Dev,e_FM_EX_BMI_STORAGE_PROFILE_ECC, FALSE);
-+ FM_SetException(p_LnxWrpFmDev->h_Dev,e_FM_EX_BMI_STATISTICS_RAM_ECC, FALSE);
-+ FM_SetException(p_LnxWrpFmDev->h_Dev,e_FM_EX_BMI_DISPATCH_RAM_ECC, FALSE);
-+ FM_SetException(p_LnxWrpFmDev->h_Dev,e_FM_EX_IRAM_ECC,FALSE);
-+ /* TODO: FmDisableRamsEcc assert for ramsEccOwners.
-+ * FM_SetException(p_LnxWrpFmDev->h_Dev,e_FM_EX_MURAM_ECC,FALSE);*/
-+ }
-+
-+ if (p_LnxWrpFmDev->fmRtcBaseAddr)
-+ {
-+ t_FmRtcParams fmRtcParam;
-+
-+ memset(&fmRtcParam, 0, sizeof(fmRtcParam));
-+ fmRtcParam.h_App = p_LnxWrpFmDev;
-+ fmRtcParam.h_Fm = p_LnxWrpFmDev->h_Dev;
-+ fmRtcParam.baseAddress = p_LnxWrpFmDev->fmRtcBaseAddr;
-+
-+ if(!(p_LnxWrpFmDev->h_RtcDev = FM_RTC_Config(&fmRtcParam)))
-+ RETURN_ERROR(MAJOR, E_INVALID_HANDLE, ("FM-RTC"));
-+
-+ if (FM_RTC_ConfigPeriod(p_LnxWrpFmDev->h_RtcDev, DPA_PTP_NOMINAL_FREQ_PERIOD_NS) != E_OK)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("FM-RTC"));
-+
-+ if (FM_RTC_Init(p_LnxWrpFmDev->h_RtcDev) != E_OK)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, ("FM-RTC"));
-+ }
-+
-+ return E_OK;
-+}
-+
-+/* TODO: to be moved back here */
-+extern void FreeFmPcdDev(t_LnxWrpFmDev *p_LnxWrpFmDev);
-+
-+static void FreeFmDev(t_LnxWrpFmDev *p_LnxWrpFmDev)
-+{
-+ if (!p_LnxWrpFmDev->active)
-+ return;
-+
-+ FreeFmPcdDev(p_LnxWrpFmDev);
-+
-+ if (p_LnxWrpFmDev->h_RtcDev)
-+ FM_RTC_Free(p_LnxWrpFmDev->h_RtcDev);
-+
-+ if (p_LnxWrpFmDev->h_Dev)
-+ FM_Free(p_LnxWrpFmDev->h_Dev);
-+
-+ if (p_LnxWrpFmDev->h_MuramDev)
-+ FM_MURAM_Free(p_LnxWrpFmDev->h_MuramDev);
-+
-+ if (p_LnxWrpFmDev->fmRtcBaseAddr)
-+ {
-+ SYS_UnregisterIoMap(p_LnxWrpFmDev->fmRtcBaseAddr);
-+ devm_iounmap(p_LnxWrpFmDev->dev, UINT_TO_PTR(p_LnxWrpFmDev->fmRtcBaseAddr));
-+ __devm_release_region(p_LnxWrpFmDev->dev, p_LnxWrpFmDev->res, p_LnxWrpFmDev->fmRtcPhysBaseAddr, p_LnxWrpFmDev->fmRtcMemSize);
-+ }
-+ SYS_UnregisterIoMap(p_LnxWrpFmDev->fmMuramBaseAddr);
-+ devm_iounmap(p_LnxWrpFmDev->dev, UINT_TO_PTR(p_LnxWrpFmDev->fmMuramBaseAddr));
-+ __devm_release_region(p_LnxWrpFmDev->dev, p_LnxWrpFmDev->res, p_LnxWrpFmDev->fmMuramPhysBaseAddr, p_LnxWrpFmDev->fmMuramMemSize);
-+ SYS_UnregisterIoMap(p_LnxWrpFmDev->fmBaseAddr);
-+ devm_iounmap(p_LnxWrpFmDev->dev, UINT_TO_PTR(p_LnxWrpFmDev->fmBaseAddr));
-+ devm_release_mem_region(p_LnxWrpFmDev->dev, p_LnxWrpFmDev->fmPhysBaseAddr, p_LnxWrpFmDev->fmMemSize);
-+ if (p_LnxWrpFmDev->err_irq != 0) {
-+ devm_free_irq(p_LnxWrpFmDev->dev, p_LnxWrpFmDev->err_irq, p_LnxWrpFmDev);
-+ }
-+
-+ devm_free_irq(p_LnxWrpFmDev->dev, p_LnxWrpFmDev->irq, p_LnxWrpFmDev);
-+}
-+
-+/* FMan character device file operations */
-+extern struct file_operations fm_fops;
-+
-+static int /*__devinit*/ fm_probe(struct platform_device *of_dev)
-+{
-+ t_LnxWrpFmDev *p_LnxWrpFmDev;
-+
-+ if ((p_LnxWrpFmDev = ReadFmDevTreeNode(of_dev)) == NULL)
-+ return -EIO;
-+ if (ConfigureFmDev(p_LnxWrpFmDev) != E_OK)
-+ return -EIO;
-+ if (InitFmDev(p_LnxWrpFmDev) != E_OK)
-+ return -EIO;
-+
-+ /* IOCTL ABI checking */
-+ LnxWrpPCDIOCTLEnumChecking();
-+ LnxWrpPCDIOCTLTypeChecking();
-+
-+ Sprint (p_LnxWrpFmDev->name, "%s%d", DEV_FM_NAME, p_LnxWrpFmDev->id);
-+
-+ /* Register to the /dev for IOCTL API */
-+ /* Register dynamically a new major number for the character device: */
-+ if ((p_LnxWrpFmDev->major = register_chrdev(0, p_LnxWrpFmDev->name, &fm_fops)) <= 0) {
-+ REPORT_ERROR(MAJOR, E_INVALID_STATE, ("Failed to allocate a major number for device \"%s\"", p_LnxWrpFmDev->name));
-+ return -EIO;
-+ }
-+
-+ /* Creating classes for FM */
-+ DBG(TRACE ,("class_create fm_class"));
-+ p_LnxWrpFmDev->fm_class = class_create(THIS_MODULE, p_LnxWrpFmDev->name);
-+ if (IS_ERR(p_LnxWrpFmDev->fm_class)) {
-+ unregister_chrdev(p_LnxWrpFmDev->major, p_LnxWrpFmDev->name);
-+ REPORT_ERROR(MAJOR, E_INVALID_STATE, ("class_create error fm_class"));
-+ return -EIO;
-+ }
-+
-+ device_create(p_LnxWrpFmDev->fm_class, NULL, MKDEV(p_LnxWrpFmDev->major, DEV_FM_MINOR_BASE), NULL,
-+ "fm%d", p_LnxWrpFmDev->id);
-+ device_create(p_LnxWrpFmDev->fm_class, NULL, MKDEV(p_LnxWrpFmDev->major, DEV_FM_PCD_MINOR_BASE), NULL,
-+ "fm%d-pcd", p_LnxWrpFmDev->id);
-+ dev_set_drvdata(p_LnxWrpFmDev->dev, p_LnxWrpFmDev);
-+
-+ /* create sysfs entries for stats and regs */
-+ if ( fm_sysfs_create(p_LnxWrpFmDev->dev) !=0 )
-+ {
-+ FreeFmDev(p_LnxWrpFmDev);
-+ REPORT_ERROR(MAJOR, E_INVALID_STATE, ("Unable to create sysfs entry - fm!!!"));
-+ return -EIO;
-+ }
-+
-+#ifdef CONFIG_PM
-+ device_set_wakeup_capable(p_LnxWrpFmDev->dev, true);
-+#endif
-+
-+ DBG(TRACE, ("FM%d probed", p_LnxWrpFmDev->id));
-+
-+ return 0;
-+}
-+
-+static int fm_remove(struct platform_device *of_dev)
-+{
-+ t_LnxWrpFmDev *p_LnxWrpFmDev;
-+ struct device *dev;
-+
-+ dev = &of_dev->dev;
-+ p_LnxWrpFmDev = dev_get_drvdata(dev);
-+
-+ fm_sysfs_destroy(dev);
-+
-+ DBG(TRACE, ("destroy fm_class"));
-+ device_destroy(p_LnxWrpFmDev->fm_class, MKDEV(p_LnxWrpFmDev->major, DEV_FM_MINOR_BASE));
-+ device_destroy(p_LnxWrpFmDev->fm_class, MKDEV(p_LnxWrpFmDev->major, DEV_FM_PCD_MINOR_BASE));
-+ class_destroy(p_LnxWrpFmDev->fm_class);
-+
-+ /* Destroy chardev */
-+ unregister_chrdev(p_LnxWrpFmDev->major, p_LnxWrpFmDev->name);
-+
-+ FreeFmDev(p_LnxWrpFmDev);
-+
-+ DestroyFmDev(p_LnxWrpFmDev);
-+
-+ dev_set_drvdata(dev, NULL);
-+
-+ return 0;
-+}
-+
-+static const struct of_device_id fm_match[] = {
-+ {
-+ .compatible = "fsl,fman"
-+ },
-+ {}
-+};
-+#ifndef MODULE
-+MODULE_DEVICE_TABLE(of, fm_match);
-+#endif /* !MODULE */
-+
-+#ifdef CONFIG_PM
-+
-+#define SCFG_FMCLKDPSLPCR_ADDR 0xFFE0FC00C
-+#define SCFG_FMCLKDPSLPCR_DS_VAL 0x48402000
-+#define SCFG_FMCLKDPSLPCR_NORMAL_VAL 0x00402000
-+
-+struct device *g_fm_dev;
-+
-+static int fm_soc_suspend(struct device *dev)
-+{
-+ int err = 0;
-+ uint32_t *fmclk;
-+ t_LnxWrpFmDev *p_LnxWrpFmDev = dev_get_drvdata(get_device(dev));
-+ g_fm_dev = dev;
-+ fmclk = ioremap(SCFG_FMCLKDPSLPCR_ADDR, 4);
-+ WRITE_UINT32(*fmclk, SCFG_FMCLKDPSLPCR_DS_VAL);
-+ if (p_LnxWrpFmDev->h_DsarRxPort)
-+ {
-+#ifdef CONFIG_FSL_QORIQ_PM
-+ device_set_wakeup_enable(p_LnxWrpFmDev->dev, 1);
-+#endif
-+ err = FM_PORT_EnterDsarFinal(p_LnxWrpFmDev->h_DsarRxPort,
-+ p_LnxWrpFmDev->h_DsarTxPort);
-+ }
-+ return err;
-+}
-+
-+static int fm_soc_resume(struct device *dev)
-+{
-+ t_LnxWrpFmDev *p_LnxWrpFmDev = dev_get_drvdata(get_device(dev));
-+ uint32_t *fmclk;
-+ fmclk = ioremap(SCFG_FMCLKDPSLPCR_ADDR, 4);
-+ WRITE_UINT32(*fmclk, SCFG_FMCLKDPSLPCR_NORMAL_VAL);
-+ if (p_LnxWrpFmDev->h_DsarRxPort)
-+ {
-+#ifdef CONFIG_FSL_QORIQ_PM
-+ device_set_wakeup_enable(p_LnxWrpFmDev->dev, 0);
-+#endif
-+ FM_PORT_ExitDsar(p_LnxWrpFmDev->h_DsarRxPort,
-+ p_LnxWrpFmDev->h_DsarTxPort);
-+ p_LnxWrpFmDev->h_DsarRxPort = 0;
-+ p_LnxWrpFmDev->h_DsarTxPort = 0;
-+ }
-+ return 0;
-+}
-+
-+static const struct dev_pm_ops fm_pm_ops = {
-+ .suspend = fm_soc_suspend,
-+ .resume = fm_soc_resume,
-+};
-+
-+#define FM_PM_OPS (&fm_pm_ops)
-+
-+#else /* CONFIG_PM */
-+
-+#define FM_PM_OPS NULL
-+
-+#endif /* CONFIG_PM */
-+
-+static struct platform_driver fm_driver = {
-+ .driver = {
-+ .name = "fsl-fman",
-+ .of_match_table = fm_match,
-+ .owner = THIS_MODULE,
-+ .pm = FM_PM_OPS,
-+ },
-+ .probe = fm_probe,
-+ .remove = fm_remove
-+};
-+
-+t_Handle LNXWRP_FM_Init(void)
-+{
-+ memset(&lnxWrpFm, 0, sizeof(lnxWrpFm));
-+ mutex_init(&lnxwrp_mutex);
-+
-+ /* Register to the DTB for basic FM API */
-+ platform_driver_register(&fm_driver);
-+
-+ return &lnxWrpFm;
-+}
-+
-+t_Error LNXWRP_FM_Free(t_Handle h_LnxWrpFm)
-+{
-+ platform_driver_unregister(&fm_driver);
-+ mutex_destroy(&lnxwrp_mutex);
-+
-+ return E_OK;
-+}
-+
-+
-+struct fm * fm_bind(struct device *fm_dev)
-+{
-+ return (struct fm *)(dev_get_drvdata(get_device(fm_dev)));
-+}
-+EXPORT_SYMBOL(fm_bind);
-+
-+void fm_unbind(struct fm *fm)
-+{
-+ t_LnxWrpFmDev *p_LnxWrpFmDev = (t_LnxWrpFmDev*)fm;
-+
-+ put_device(p_LnxWrpFmDev->dev);
-+}
-+EXPORT_SYMBOL(fm_unbind);
-+
-+struct resource * fm_get_mem_region(struct fm *fm)
-+{
-+ t_LnxWrpFmDev *p_LnxWrpFmDev = (t_LnxWrpFmDev*)fm;
-+
-+ return p_LnxWrpFmDev->res;
-+}
-+EXPORT_SYMBOL(fm_get_mem_region);
-+
-+void * fm_get_handle(struct fm *fm)
-+{
-+ t_LnxWrpFmDev *p_LnxWrpFmDev = (t_LnxWrpFmDev*)fm;
-+
-+ return (void *)p_LnxWrpFmDev->h_Dev;
-+}
-+EXPORT_SYMBOL(fm_get_handle);
-+
-+void * fm_get_rtc_handle(struct fm *fm)
-+{
-+ t_LnxWrpFmDev *p_LnxWrpFmDev = (t_LnxWrpFmDev*)fm;
-+
-+ return (void *)p_LnxWrpFmDev->h_RtcDev;
-+}
-+EXPORT_SYMBOL(fm_get_rtc_handle);
-+
-+struct fm_port * fm_port_bind (struct device *fm_port_dev)
-+{
-+ return (struct fm_port *)(dev_get_drvdata(get_device(fm_port_dev)));
-+}
-+EXPORT_SYMBOL(fm_port_bind);
-+
-+void fm_port_unbind(struct fm_port *port)
-+{
-+ t_LnxWrpFmPortDev *p_LnxWrpFmPortDev = (t_LnxWrpFmPortDev*)port;
-+
-+ put_device(p_LnxWrpFmPortDev->dev);
-+}
-+EXPORT_SYMBOL(fm_port_unbind);
-+
-+void *fm_port_get_handle(const struct fm_port *port)
-+{
-+ t_LnxWrpFmPortDev *p_LnxWrpFmPortDev = (t_LnxWrpFmPortDev*)port;
-+
-+ return (void *)p_LnxWrpFmPortDev->h_Dev;
-+}
-+EXPORT_SYMBOL(fm_port_get_handle);
-+
-+u64 *fm_port_get_buffer_time_stamp(const struct fm_port *port,
-+ const void *data)
-+{
-+ return FM_PORT_GetBufferTimeStamp(fm_port_get_handle(port),
-+ (void *)data);
-+}
-+EXPORT_SYMBOL(fm_port_get_buffer_time_stamp);
-+
-+void fm_port_get_base_addr(const struct fm_port *port, uint64_t *base_addr)
-+{
-+ t_LnxWrpFmPortDev *p_LnxWrpFmPortDev = (t_LnxWrpFmPortDev *)port;
-+
-+ *base_addr = p_LnxWrpFmPortDev->settings.param.baseAddr;
-+}
-+EXPORT_SYMBOL(fm_port_get_base_addr);
-+
-+void fm_port_pcd_bind (struct fm_port *port, struct fm_port_pcd_param *params)
-+{
-+ t_LnxWrpFmPortDev *p_LnxWrpFmPortDev = (t_LnxWrpFmPortDev*)port;
-+
-+ p_LnxWrpFmPortDev->pcd_owner_params.cba = params->cba;
-+ p_LnxWrpFmPortDev->pcd_owner_params.cbf = params->cbf;
-+ p_LnxWrpFmPortDev->pcd_owner_params.dev = params->dev;
-+}
-+EXPORT_SYMBOL(fm_port_pcd_bind);
-+
-+void fm_port_get_buff_layout_ext_params(struct fm_port *port, struct fm_port_params *params)
-+{
-+ t_LnxWrpFmPortDev *p_LnxWrpFmPortDev = (t_LnxWrpFmPortDev *)port;
-+ struct device_node *fm_node, *port_node;
-+ const uint32_t *uint32_prop;
-+ int lenp;
-+
-+ params->data_align = 0;
-+ params->manip_extra_space = 0;
-+
-+ fm_node = GetFmAdvArgsDevTreeNode(((t_LnxWrpFmDev *) p_LnxWrpFmPortDev->h_LnxWrpFmDev)->id);
-+ if (!fm_node) /* no advance parameters for FMan */
-+ return;
-+
-+ port_node = GetFmPortAdvArgsDevTreeNode(fm_node,
-+ p_LnxWrpFmPortDev->settings.param.portType,
-+ p_LnxWrpFmPortDev->settings.param.portId);
-+ if (!port_node) /* no advance parameters for FMan-Port */
-+ return;
-+
-+ uint32_prop = (uint32_t *)of_get_property(port_node, "buffer-layout", &lenp);
-+ if (uint32_prop) {
-+ if (WARN_ON(lenp != sizeof(uint32_t)*2))
-+ return;
-+
-+ params->manip_extra_space = (uint8_t)be32_to_cpu(uint32_prop[0]);
-+ params->data_align = (uint16_t)be32_to_cpu(uint32_prop[1]);
-+ }
-+
-+ of_node_put(port_node);
-+ of_node_put(fm_node);
-+}
-+EXPORT_SYMBOL(fm_port_get_buff_layout_ext_params);
-+
-+uint16_t fm_get_tx_port_channel(struct fm_port *port)
-+{
-+ t_LnxWrpFmPortDev *p_LnxWrpFmPortDev = (t_LnxWrpFmPortDev*)port;
-+
-+ return p_LnxWrpFmPortDev->txCh;
-+}
-+EXPORT_SYMBOL(fm_get_tx_port_channel);
-+
-+int fm_port_enable (struct fm_port *port)
-+{
-+ t_LnxWrpFmPortDev *p_LnxWrpFmPortDev = (t_LnxWrpFmPortDev*)port;
-+ t_Error err = FM_PORT_Enable(p_LnxWrpFmPortDev->h_Dev);
-+
-+ return GET_ERROR_TYPE(err);
-+}
-+EXPORT_SYMBOL(fm_port_enable);
-+
-+int fm_port_disable(struct fm_port *port)
-+{
-+ t_LnxWrpFmPortDev *p_LnxWrpFmPortDev = (t_LnxWrpFmPortDev*)port;
-+ t_Error err = FM_PORT_Disable(p_LnxWrpFmPortDev->h_Dev);
-+
-+ return GET_ERROR_TYPE(err);
-+}
-+EXPORT_SYMBOL(fm_port_disable);
-+
-+int fm_port_set_rate_limit(struct fm_port *port,
-+ uint16_t max_burst_size,
-+ uint32_t rate_limit)
-+{
-+ t_FmPortRateLimit param;
-+ t_LnxWrpFmPortDev *p_LnxWrpFmPortDev = (t_LnxWrpFmPortDev *)port;
-+ int err = 0;
-+
-+ param.maxBurstSize = max_burst_size;
-+ param.rateLimit = rate_limit;
-+ param.rateLimitDivider = 0;
-+
-+ err = FM_PORT_SetRateLimit(p_LnxWrpFmPortDev->h_Dev, &param);
-+ return err;
-+}
-+EXPORT_SYMBOL(fm_port_set_rate_limit);
-+
-+int fm_port_del_rate_limit(struct fm_port *port)
-+{
-+ t_LnxWrpFmPortDev *p_LnxWrpFmPortDev = (t_LnxWrpFmPortDev *)port;
-+
-+ FM_PORT_DeleteRateLimit(p_LnxWrpFmPortDev->h_Dev);
-+ return 0;
-+}
-+EXPORT_SYMBOL(fm_port_del_rate_limit);
-+
-+void FM_PORT_Dsar_DumpRegs(void);
-+int ar_showmem(struct file *file, const char __user *buffer,
-+ unsigned long count, void *data)
-+{
-+ FM_PORT_Dsar_DumpRegs();
-+ return 2;
-+}
-+
-+struct auto_res_tables_sizes *fm_port_get_autores_maxsize(
-+ struct fm_port *port)
-+{
-+ t_LnxWrpFmPortDev *p_LnxWrpFmPortDev = (t_LnxWrpFmPortDev *)port;
-+ return &p_LnxWrpFmPortDev->dsar_table_sizes;
-+}
-+EXPORT_SYMBOL(fm_port_get_autores_maxsize);
-+
-+int fm_port_enter_autores_for_deepsleep(struct fm_port *port,
-+ struct auto_res_port_params *params)
-+{
-+ t_LnxWrpFmPortDev *p_LnxWrpFmPortDev = (t_LnxWrpFmPortDev *)port;
-+ t_LnxWrpFmDev* p_LnxWrpFmDev = (t_LnxWrpFmDev*)p_LnxWrpFmPortDev->h_LnxWrpFmDev;
-+ p_LnxWrpFmDev->h_DsarRxPort = p_LnxWrpFmPortDev->h_Dev;
-+ p_LnxWrpFmDev->h_DsarTxPort = params->h_FmPortTx;
-+
-+ /*Register other under /proc/autoresponse */
-+ if (WARN_ON(sizeof(t_FmPortDsarParams) != sizeof(struct auto_res_port_params)))
-+ return -EFAULT;
-+
-+ FM_PORT_EnterDsar(p_LnxWrpFmPortDev->h_Dev, (t_FmPortDsarParams*)params);
-+ return 0;
-+}
-+EXPORT_SYMBOL(fm_port_enter_autores_for_deepsleep);
-+
-+void fm_port_exit_auto_res_for_deep_sleep(struct fm_port *port_rx,
-+ struct fm_port *port_tx)
-+{
-+}
-+EXPORT_SYMBOL(fm_port_exit_auto_res_for_deep_sleep);
-+
-+int fm_port_get_autores_stats(struct fm_port *port,
-+ struct auto_res_port_stats *stats)
-+{
-+ t_LnxWrpFmPortDev *p_LnxWrpFmPortDev = (t_LnxWrpFmPortDev *)port;
-+ if (WARN_ON(sizeof(t_FmPortDsarStats) != sizeof(struct auto_res_port_stats)))
-+ return -EFAULT;
-+ return FM_PORT_GetDsarStats(p_LnxWrpFmPortDev->h_Dev, (t_FmPortDsarStats*)stats);
-+}
-+EXPORT_SYMBOL(fm_port_get_autores_stats);
-+
-+int fm_port_suspend(struct fm_port *port)
-+{
-+ t_LnxWrpFmPortDev *p_LnxWrpFmPortDev = (t_LnxWrpFmPortDev *)port;
-+ if (!FM_PORT_IsInDsar(p_LnxWrpFmPortDev->h_Dev))
-+ return FM_PORT_Disable(p_LnxWrpFmPortDev->h_Dev);
-+ else
-+ return 0;
-+}
-+EXPORT_SYMBOL(fm_port_suspend);
-+
-+int fm_port_resume(struct fm_port *port)
-+{
-+ t_LnxWrpFmPortDev *p_LnxWrpFmPortDev = (t_LnxWrpFmPortDev *)port;
-+ if (!FM_PORT_IsInDsar(p_LnxWrpFmPortDev->h_Dev))
-+ return FM_PORT_Enable(p_LnxWrpFmPortDev->h_Dev);
-+ else
-+ return 0;
-+}
-+EXPORT_SYMBOL(fm_port_resume);
-+
-+bool fm_port_is_in_auto_res_mode(struct fm_port *port)
-+{
-+ return FM_PORT_IsInDsar(port);
-+}
-+EXPORT_SYMBOL(fm_port_is_in_auto_res_mode);
-+
-+#ifdef CONFIG_FMAN_PFC
-+int fm_port_set_pfc_priorities_mapping_to_qman_wq(struct fm_port *port,
-+ uint8_t prio, uint8_t wq)
-+{
-+ t_LnxWrpFmPortDev *p_LnxWrpFmPortDev = (t_LnxWrpFmPortDev *)port;
-+ int err;
-+ int _errno;
-+
-+ err = FM_PORT_SetPfcPrioritiesMappingToQmanWQ(p_LnxWrpFmPortDev->h_Dev,
-+ prio, wq);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_PORT_SetPfcPrioritiesMappingToQmanWQ() = 0x%08x\n", err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_port_set_pfc_priorities_mapping_to_qman_wq);
-+#endif
-+
-+int fm_mac_set_exception(struct fm_mac_dev *fm_mac_dev,
-+ e_FmMacExceptions exception, bool enable)
-+{
-+ int err;
-+ int _errno;
-+
-+ err = FM_MAC_SetException(fm_mac_dev, exception, enable);
-+
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_MAC_SetException() = 0x%08x\n", err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_mac_set_exception);
-+
-+int fm_mac_free(struct fm_mac_dev *fm_mac_dev)
-+{
-+ int err;
-+ int _error;
-+
-+ err = FM_MAC_Free(fm_mac_dev);
-+ _error = -GET_ERROR_TYPE(err);
-+
-+ if (unlikely(_error < 0))
-+ pr_err("FM_MAC_Free() = 0x%08x\n", err);
-+
-+ return _error;
-+}
-+EXPORT_SYMBOL(fm_mac_free);
-+
-+struct fm_mac_dev *fm_mac_config(t_FmMacParams *params)
-+{
-+ struct fm_mac_dev *fm_mac_dev;
-+
-+ fm_mac_dev = FM_MAC_Config(params);
-+ if (unlikely(fm_mac_dev == NULL))
-+ pr_err("FM_MAC_Config() failed\n");
-+
-+ return fm_mac_dev;
-+}
-+EXPORT_SYMBOL(fm_mac_config);
-+
-+int fm_mac_config_max_frame_length(struct fm_mac_dev *fm_mac_dev,
-+ int len)
-+{
-+ int err;
-+ int _errno;
-+
-+ err = FM_MAC_ConfigMaxFrameLength(fm_mac_dev, len);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_MAC_ConfigMaxFrameLength() = 0x%08x\n", err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_mac_config_max_frame_length);
-+
-+int fm_mac_config_pad_and_crc(struct fm_mac_dev *fm_mac_dev, bool enable)
-+{
-+ int err;
-+ int _errno;
-+
-+ err = FM_MAC_ConfigPadAndCrc(fm_mac_dev, enable);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_MAC_ConfigPadAndCrc() = 0x%08x\n", err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_mac_config_pad_and_crc);
-+
-+int fm_mac_config_half_duplex(struct fm_mac_dev *fm_mac_dev, bool enable)
-+{
-+ int err;
-+ int _errno;
-+
-+ err = FM_MAC_ConfigHalfDuplex(fm_mac_dev, enable);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_MAC_ConfigHalfDuplex() = 0x%08x\n", err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_mac_config_half_duplex);
-+
-+int fm_mac_config_reset_on_init(struct fm_mac_dev *fm_mac_dev, bool enable)
-+{
-+ int err;
-+ int _errno;
-+
-+ err = FM_MAC_ConfigResetOnInit(fm_mac_dev, enable);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_MAC_ConfigResetOnInit() = 0x%08x\n", err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_mac_config_reset_on_init);
-+
-+int fm_mac_init(struct fm_mac_dev *fm_mac_dev)
-+{
-+ int err;
-+ int _errno;
-+
-+ err = FM_MAC_Init(fm_mac_dev);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_MAC_Init() = 0x%08x\n", err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_mac_init);
-+
-+int fm_mac_get_version(struct fm_mac_dev *fm_mac_dev, uint32_t *version)
-+{
-+ int err;
-+ int _errno;
-+
-+ err = FM_MAC_GetVesrion(fm_mac_dev, version);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_MAC_GetVesrion() = 0x%08x\n", err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_mac_get_version);
-+
-+int fm_mac_enable(struct fm_mac_dev *fm_mac_dev)
-+{
-+ int _errno;
-+ t_Error err;
-+
-+ err = FM_MAC_Enable(fm_mac_dev, e_COMM_MODE_RX_AND_TX);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_MAC_Enable() = 0x%08x\n", err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_mac_enable);
-+
-+int fm_mac_disable(struct fm_mac_dev *fm_mac_dev)
-+{
-+ int _errno;
-+ t_Error err;
-+
-+ err = FM_MAC_Disable(fm_mac_dev, e_COMM_MODE_RX_AND_TX);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_MAC_Disable() = 0x%08x\n", err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_mac_disable);
-+
-+int fm_mac_resume(struct fm_mac_dev *fm_mac_dev)
-+{
-+ int _errno;
-+ t_Error err;
-+
-+ err = FM_MAC_Resume(fm_mac_dev);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_MAC_Resume() = 0x%08x\n", err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_mac_resume);
-+
-+int fm_mac_set_promiscuous(struct fm_mac_dev *fm_mac_dev,
-+ bool enable)
-+{
-+ int _errno;
-+ t_Error err;
-+
-+ err = FM_MAC_SetPromiscuous(fm_mac_dev, enable);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_MAC_SetPromiscuous() = 0x%08x\n", err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_mac_set_promiscuous);
-+
-+int fm_mac_remove_hash_mac_addr(struct fm_mac_dev *fm_mac_dev,
-+ t_EnetAddr *mac_addr)
-+{
-+ int _errno;
-+ t_Error err;
-+
-+ err = FM_MAC_RemoveHashMacAddr(fm_mac_dev, mac_addr);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (_errno < 0) {
-+ pr_err("FM_MAC_RemoveHashMacAddr() = 0x%08x\n", err);
-+ return _errno;
-+ }
-+
-+ return 0;
-+}
-+EXPORT_SYMBOL(fm_mac_remove_hash_mac_addr);
-+
-+int fm_mac_add_hash_mac_addr(struct fm_mac_dev *fm_mac_dev,
-+ t_EnetAddr *mac_addr)
-+{
-+ int _errno;
-+ t_Error err;
-+
-+ err = FM_MAC_AddHashMacAddr(fm_mac_dev, mac_addr);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (_errno < 0) {
-+ pr_err("FM_MAC_AddHashMacAddr() = 0x%08x\n", err);
-+ return _errno;
-+ }
-+
-+ return 0;
-+}
-+EXPORT_SYMBOL(fm_mac_add_hash_mac_addr);
-+
-+int fm_mac_modify_mac_addr(struct fm_mac_dev *fm_mac_dev,
-+ uint8_t *addr)
-+{
-+ int _errno;
-+ t_Error err;
-+
-+ err = FM_MAC_ModifyMacAddr(fm_mac_dev, (t_EnetAddr *)addr);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (_errno < 0)
-+ pr_err("FM_MAC_ModifyMacAddr() = 0x%08x\n", err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_mac_modify_mac_addr);
-+
-+int fm_mac_adjust_link(struct fm_mac_dev *fm_mac_dev,
-+ bool link, int speed, bool duplex)
-+{
-+ int _errno;
-+ t_Error err;
-+
-+ if (!link) {
-+#if (DPAA_VERSION < 11)
-+ FM_MAC_RestartAutoneg(fm_mac_dev);
-+#endif
-+ return 0;
-+ }
-+
-+ err = FM_MAC_AdjustLink(fm_mac_dev, speed, duplex);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_MAC_AdjustLink() = 0x%08x\n", err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_mac_adjust_link);
-+
-+int fm_mac_enable_1588_time_stamp(struct fm_mac_dev *fm_mac_dev)
-+{
-+ int _errno;
-+ t_Error err;
-+
-+ err = FM_MAC_Enable1588TimeStamp(fm_mac_dev);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_MAC_Enable1588TimeStamp() = 0x%08x\n", err);
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_mac_enable_1588_time_stamp);
-+
-+int fm_mac_disable_1588_time_stamp(struct fm_mac_dev *fm_mac_dev)
-+{
-+ int _errno;
-+ t_Error err;
-+
-+ err = FM_MAC_Disable1588TimeStamp(fm_mac_dev);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_MAC_Disable1588TimeStamp() = 0x%08x\n", err);
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_mac_disable_1588_time_stamp);
-+
-+int fm_mac_set_rx_pause_frames(
-+ struct fm_mac_dev *fm_mac_dev, bool en)
-+{
-+ int _errno;
-+ t_Error err;
-+
-+ /* if rx pause is enabled, do NOT ignore pause frames */
-+ err = FM_MAC_SetRxIgnorePauseFrames(fm_mac_dev, !en);
-+
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (_errno < 0)
-+ pr_err("FM_MAC_SetRxIgnorePauseFrames() = 0x%08x\n", err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_mac_set_rx_pause_frames);
-+
-+#ifdef CONFIG_FMAN_PFC
-+int fm_mac_set_tx_pause_frames(struct fm_mac_dev *fm_mac_dev,
-+ bool en)
-+{
-+ int _errno, i;
-+ t_Error err;
-+
-+ if (en)
-+ for (i = 0; i < CONFIG_FMAN_PFC_COS_COUNT; i++) {
-+ err = FM_MAC_SetTxPauseFrames(fm_mac_dev,
-+ i, fsl_fm_pfc_quanta[i],
-+ FSL_FM_PAUSE_THRESH_DEFAULT);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (_errno < 0) {
-+ pr_err("FM_MAC_SetTxPauseFrames() = 0x%08x\n", err);
-+ return _errno;
-+ }
-+ }
-+ else
-+ for (i = 0; i < CONFIG_FMAN_PFC_COS_COUNT; i++) {
-+ err = FM_MAC_SetTxPauseFrames(fm_mac_dev,
-+ i, FSL_FM_PAUSE_TIME_DISABLE,
-+ FSL_FM_PAUSE_THRESH_DEFAULT);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (_errno < 0) {
-+ pr_err("FM_MAC_SetTxPauseFrames() = 0x%08x\n", err);
-+ return _errno;
-+ }
-+ }
-+
-+ return _errno;
-+}
-+#else
-+int fm_mac_set_tx_pause_frames(struct fm_mac_dev *fm_mac_dev,
-+ bool en)
-+{
-+ int _errno;
-+ t_Error err;
-+
-+ if (en)
-+ err = FM_MAC_SetTxAutoPauseFrames(fm_mac_dev,
-+ FSL_FM_PAUSE_TIME_ENABLE);
-+ else
-+ err = FM_MAC_SetTxAutoPauseFrames(fm_mac_dev,
-+ FSL_FM_PAUSE_TIME_DISABLE);
-+
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (_errno < 0)
-+ pr_err("FM_MAC_SetTxAutoPauseFrames() = 0x%08x\n", err);
-+
-+ return _errno;
-+}
-+#endif
-+EXPORT_SYMBOL(fm_mac_set_tx_pause_frames);
-+
-+int fm_rtc_enable(struct fm *fm_dev)
-+{
-+ int _errno;
-+ t_Error err;
-+
-+ err = FM_RTC_Enable(fm_get_rtc_handle(fm_dev), 0);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_RTC_Enable = 0x%08x\n", err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_rtc_enable);
-+
-+int fm_rtc_disable(struct fm *fm_dev)
-+{
-+ int _errno;
-+ t_Error err;
-+
-+ err = FM_RTC_Disable(fm_get_rtc_handle(fm_dev));
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_RTC_Disable = 0x%08x\n", err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_rtc_disable);
-+
-+int fm_rtc_get_cnt(struct fm *fm_dev, uint64_t *ts)
-+{
-+ int _errno;
-+ t_Error err;
-+
-+ err = FM_RTC_GetCurrentTime(fm_get_rtc_handle(fm_dev), ts);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_RTC_GetCurrentTime = 0x%08x\n", err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_rtc_get_cnt);
-+
-+int fm_rtc_set_cnt(struct fm *fm_dev, uint64_t ts)
-+{
-+ int _errno;
-+ t_Error err;
-+
-+ err = FM_RTC_SetCurrentTime(fm_get_rtc_handle(fm_dev), ts);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_RTC_SetCurrentTime = 0x%08x\n", err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_rtc_set_cnt);
-+
-+int fm_rtc_get_drift(struct fm *fm_dev, uint32_t *drift)
-+{
-+ int _errno;
-+ t_Error err;
-+
-+ err = FM_RTC_GetFreqCompensation(fm_get_rtc_handle(fm_dev),
-+ drift);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_RTC_GetFreqCompensation = 0x%08x\n", err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_rtc_get_drift);
-+
-+int fm_rtc_set_drift(struct fm *fm_dev, uint32_t drift)
-+{
-+ int _errno;
-+ t_Error err;
-+
-+ err = FM_RTC_SetFreqCompensation(fm_get_rtc_handle(fm_dev),
-+ drift);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_RTC_SetFreqCompensation = 0x%08x\n", err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_rtc_set_drift);
-+
-+int fm_rtc_set_alarm(struct fm *fm_dev, uint32_t id,
-+ uint64_t time)
-+{
-+ t_FmRtcAlarmParams alarm;
-+ int _errno;
-+ t_Error err;
-+
-+ alarm.alarmId = id;
-+ alarm.alarmTime = time;
-+ alarm.f_AlarmCallback = NULL;
-+ err = FM_RTC_SetAlarm(fm_get_rtc_handle(fm_dev),
-+ &alarm);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_RTC_SetAlarm = 0x%08x\n", err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_rtc_set_alarm);
-+
-+int fm_rtc_set_fiper(struct fm *fm_dev, uint32_t id,
-+ uint64_t fiper)
-+{
-+ t_FmRtcPeriodicPulseParams pp;
-+ int _errno;
-+ t_Error err;
-+
-+ pp.periodicPulseId = id;
-+ pp.periodicPulsePeriod = fiper;
-+ pp.f_PeriodicPulseCallback = NULL;
-+ err = FM_RTC_SetPeriodicPulse(fm_get_rtc_handle(fm_dev), &pp);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_RTC_SetPeriodicPulse = 0x%08x\n", err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_rtc_set_fiper);
-+
-+#ifdef CONFIG_PTP_1588_CLOCK_DPAA
-+int fm_rtc_enable_interrupt(struct fm *fm_dev, uint32_t events)
-+{
-+ int _errno;
-+ t_Error err;
-+
-+ err = FM_RTC_EnableInterrupt(fm_get_rtc_handle(fm_dev),
-+ events);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_RTC_EnableInterrupt = 0x%08x\n", err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_rtc_enable_interrupt);
-+
-+int fm_rtc_disable_interrupt(struct fm *fm_dev, uint32_t events)
-+{
-+ int _errno;
-+ t_Error err;
-+
-+ err = FM_RTC_DisableInterrupt(fm_get_rtc_handle(fm_dev),
-+ events);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_RTC_DisableInterrupt = 0x%08x\n", err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_rtc_disable_interrupt);
-+#endif
-+
-+int fm_mac_set_wol(struct fm_port *port, struct fm_mac_dev *fm_mac_dev, bool en)
-+{
-+ int _errno;
-+ t_Error err;
-+ t_LnxWrpFmPortDev *p_LnxWrpFmPortDev = (t_LnxWrpFmPortDev *)port;
-+
-+ /* Do not set WoL on AR ports */
-+ if (FM_PORT_IsInDsar(p_LnxWrpFmPortDev->h_Dev)) {
-+ printk(KERN_WARNING "Port is AutoResponse enabled! WoL will not be set on this port!\n");
-+ return 0;
-+ }
-+
-+ err = FM_MAC_SetWakeOnLan(fm_mac_dev, en);
-+
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (_errno < 0)
-+ pr_err("FM_MAC_SetWakeOnLan() = 0x%08x\n", err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_mac_set_wol);
-+
-+void fm_mutex_lock(void)
-+{
-+ mutex_lock(&lnxwrp_mutex);
-+}
-+EXPORT_SYMBOL(fm_mutex_lock);
-+
-+void fm_mutex_unlock(void)
-+{
-+ mutex_unlock(&lnxwrp_mutex);
-+}
-+EXPORT_SYMBOL(fm_mutex_unlock);
-+
-+/*Macsec wrapper functions*/
-+struct fm_macsec_dev *fm_macsec_config(struct fm_macsec_params *fm_params)
-+{
-+ struct fm_macsec_dev *fm_macsec_dev;
-+
-+ fm_macsec_dev = FM_MACSEC_Config((t_FmMacsecParams *)fm_params);
-+ if (unlikely(fm_macsec_dev == NULL))
-+ pr_err("FM_MACSEC_Config() failed\n");
-+
-+ return fm_macsec_dev;
-+}
-+EXPORT_SYMBOL(fm_macsec_config);
-+
-+int fm_macsec_init(struct fm_macsec_dev *fm_macsec_dev)
-+{
-+ int err;
-+ int _errno;
-+
-+ err = FM_MACSEC_Init(fm_macsec_dev);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_MACSEC_Init() = 0x%08x\n", err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_macsec_init);
-+
-+int fm_macsec_free(struct fm_macsec_dev *fm_macsec_dev)
-+{
-+ int err;
-+ int _error;
-+
-+ err = FM_MACSEC_Free(fm_macsec_dev);
-+ _error = -GET_ERROR_TYPE(err);
-+
-+ if (unlikely(_error < 0))
-+ pr_err("FM_MACSEC_Free() = 0x%08x\n", err);
-+
-+ return _error;
-+}
-+EXPORT_SYMBOL(fm_macsec_free);
-+
-+int fm_macsec_config_unknown_sci_frame_treatment(struct fm_macsec_dev
-+ *fm_macsec_dev,
-+ fm_macsec_unknown_sci_frame_treatment treat_mode)
-+{
-+ int err;
-+ int _errno;
-+
-+ err = FM_MACSEC_ConfigUnknownSciFrameTreatment(fm_macsec_dev,
-+ treat_mode);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_MACSEC_ConfigUnknownSciFrameTreatmen() = 0x%08x\n", err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_macsec_config_unknown_sci_frame_treatment);
-+
-+int fm_macsec_config_invalid_tags_frame_treatment(struct fm_macsec_dev *fm_macsec_dev,
-+ bool deliver_uncontrolled)
-+{
-+ int err;
-+ int _errno;
-+
-+ err = FM_MACSEC_ConfigInvalidTagsFrameTreatment(fm_macsec_dev,
-+ deliver_uncontrolled);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_MAC_ConfigMaxFrameLength() = 0x%08x\n", err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_macsec_config_invalid_tags_frame_treatment);
-+
-+int fm_macsec_config_kay_frame_treatment(struct fm_macsec_dev *fm_macsec_dev,
-+ bool discard_uncontrolled)
-+{
-+ int err;
-+ int _errno;
-+
-+ err = FM_MACSEC_ConfigEncryptWithNoChangedTextFrameTreatment(fm_macsec_dev,
-+ discard_uncontrolled);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_MACSEC_ConfigEncryptWithNoChangedTextFrameTreatmen() = 0x%08x\n", err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_macsec_config_kay_frame_treatment);
-+
-+int fm_macsec_config_untag_frame_treatment(struct fm_macsec_dev *fm_macsec_dev,
-+ fm_macsec_untag_frame_treatment treat_mode)
-+{
-+ int err;
-+ int _errno;
-+
-+ err = FM_MACSEC_ConfigUntagFrameTreatment(fm_macsec_dev, treat_mode);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_MACSEC_ConfigUntagFrameTreatment() = 0x%08x\n", err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_macsec_config_untag_frame_treatment);
-+
-+int fm_macsec_config_pn_exhaustion_threshold(struct fm_macsec_dev *fm_macsec_dev,
-+ uint32_t pn_exh_thr)
-+{
-+ int err;
-+ int _errno;
-+
-+ err = FM_MACSEC_ConfigPnExhaustionThreshold(fm_macsec_dev, pn_exh_thr);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_MACSEC_ConfigPnExhaustionThreshold() = 0x%08x\n", err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_macsec_config_pn_exhaustion_threshold);
-+
-+int fm_macsec_config_keys_unreadable(struct fm_macsec_dev *fm_macsec_dev)
-+{
-+ int err;
-+ int _errno;
-+
-+ err = FM_MACSEC_ConfigKeysUnreadable(fm_macsec_dev);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_MACSEC_ConfigKeysUnreadable() = 0x%08x\n", err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_macsec_config_keys_unreadable);
-+
-+int fm_macsec_config_sectag_without_sci(struct fm_macsec_dev *fm_macsec_dev)
-+{
-+ int err;
-+ int _errno;
-+
-+ err = FM_MACSEC_ConfigSectagWithoutSCI(fm_macsec_dev);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_MACSEC_ConfigSectagWithoutSCI() = 0x%08x\n", err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_macsec_config_sectag_without_sci);
-+
-+int fm_macsec_config_exception(struct fm_macsec_dev *fm_macsec_dev,
-+ fm_macsec_exception exception, bool enable)
-+{
-+ int err;
-+ int _errno;
-+
-+ err = FM_MACSEC_ConfigException(fm_macsec_dev, exception, enable);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_MACSEC_ConfigException() = 0x%08x\n", err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_macsec_config_exception);
-+
-+int fm_macsec_get_revision(struct fm_macsec_dev *fm_macsec_dev,
-+ int *macsec_revision)
-+{
-+ int err;
-+ int _errno;
-+
-+ err = FM_MACSEC_GetRevision(fm_macsec_dev, macsec_revision);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_MACSEC_GetRevision() = 0x%08x\n", err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_macsec_get_revision);
-+
-+int fm_macsec_enable(struct fm_macsec_dev *fm_macsec_dev)
-+{
-+ int err;
-+ int _errno;
-+
-+ err = FM_MACSEC_Enable(fm_macsec_dev);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_MACSEC_Enable() = 0x%08x\n", err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_macsec_enable);
-+
-+int fm_macsec_disable(struct fm_macsec_dev *fm_macsec_dev)
-+{
-+ int err;
-+ int _errno;
-+
-+ err = FM_MACSEC_Disable(fm_macsec_dev);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_MACSEC_Disable() = 0x%08x\n", err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_macsec_disable);
-+
-+int fm_macsec_set_exception(struct fm_macsec_dev *fm_macsec_dev,
-+ fm_macsec_exception exception, bool enable)
-+{
-+ int err;
-+ int _errno;
-+
-+ err = FM_MACSEC_SetException(fm_macsec_dev, exception, enable);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_MACSEC_SetException() = 0x%08x\n", err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_macsec_set_exception);
-+
-+/* Macsec SECY wrapper API */
-+struct fm_macsec_secy_dev *fm_macsec_secy_config(struct fm_macsec_secy_params *secy_params)
-+{
-+ struct fm_macsec_secy_dev *fm_macsec_secy;
-+
-+ fm_macsec_secy = FM_MACSEC_SECY_Config((t_FmMacsecSecYParams *)secy_params);
-+ if (unlikely(fm_macsec_secy < 0))
-+ pr_err("FM_MACSEC_SECY_Config() failed\n");
-+
-+ return fm_macsec_secy;
-+}
-+EXPORT_SYMBOL(fm_macsec_secy_config);
-+
-+int fm_macsec_secy_init(struct fm_macsec_secy_dev *fm_macsec_secy_dev)
-+{
-+ int err;
-+ int _errno;
-+
-+ err = FM_MACSEC_SECY_Init(fm_macsec_secy_dev);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_MACSEC_SECY_Init() = 0x%08x\n", err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_macsec_secy_init);
-+
-+int fm_macsec_secy_free(struct fm_macsec_secy_dev *fm_macsec_secy_dev)
-+{
-+ int err;
-+ int _errno;
-+
-+ err = FM_MACSEC_SECY_Free(fm_macsec_secy_dev);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_MACSEC_SECY_Free() = 0x%08x\n", err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_macsec_secy_free);
-+
-+int fm_macsec_secy_config_sci_insertion_mode(struct fm_macsec_secy_dev *fm_macsec_secy_dev,
-+ fm_macsec_sci_insertion_mode sci_insertion_mode)
-+{
-+ int err;
-+ int _errno;
-+
-+ err = FM_MACSEC_SECY_ConfigSciInsertionMode(fm_macsec_secy_dev,
-+ sci_insertion_mode);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_MACSEC_SECY_ConfigSciInsertionMode() = 0x%08x\n", err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_macsec_secy_config_sci_insertion_mode);
-+
-+int fm_macsec_secy_config_protect_frames(struct fm_macsec_secy_dev *fm_macsec_secy_dev,
-+ bool protect_frames)
-+{
-+ int err;
-+ int _errno;
-+
-+ err = FM_MACSEC_SECY_ConfigProtectFrames(fm_macsec_secy_dev,
-+ protect_frames);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_MACSEC_SECY_ConfigProtectFrames() = 0x%08x\n", err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_macsec_secy_config_protect_frames);
-+
-+int fm_macsec_secy_config_replay_window(struct fm_macsec_secy_dev *fm_macsec_secy_dev,
-+ bool replay_protect, uint32_t replay_window)
-+{
-+ int err;
-+ int _errno;
-+
-+ err = FM_MACSEC_SECY_ConfigReplayWindow(fm_macsec_secy_dev,
-+ replay_protect, replay_window);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_MACSEC_SECY_ConfigReplayWindow() = 0x%08x\n", err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_macsec_secy_config_replay_window);
-+
-+int fm_macsec_secy_config_validation_mode(struct fm_macsec_secy_dev *fm_macsec_secy_dev,
-+ fm_macsec_valid_frame_behavior validate_frames)
-+{
-+ int err;
-+ int _errno;
-+
-+ err = FM_MACSEC_SECY_ConfigValidationMode(fm_macsec_secy_dev,
-+ validate_frames);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_MACSEC_SECY_ConfigValidationMode() = 0x%08x\n", err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_macsec_secy_config_validation_mode);
-+
-+int fm_macsec_secy_config_confidentiality(struct fm_macsec_secy_dev *fm_macsec_secy_dev,
-+ bool confidentiality_enable,
-+ uint32_t confidentiality_offset)
-+{
-+ int err;
-+ int _errno;
-+
-+ err = FM_MACSEC_SECY_ConfigConfidentiality(fm_macsec_secy_dev,
-+ confidentiality_enable,
-+ confidentiality_offset);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_MACSEC_SECY_ConfigConfidentiality() = 0x%08x\n",
-+ err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_macsec_secy_config_confidentiality);
-+
-+int fm_macsec_secy_config_point_to_point(struct fm_macsec_secy_dev *fm_macsec_secy_dev)
-+{
-+ int err;
-+ int _errno;
-+
-+ err = FM_MACSEC_SECY_ConfigPointToPoint(fm_macsec_secy_dev);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_MACSEC_SECY_ConfigPointToPoint() = 0x%08x\n",
-+ err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_macsec_secy_config_point_to_point);
-+
-+int fm_macsec_secy_config_exception(struct fm_macsec_secy_dev *fm_macsec_secy_dev,
-+ fm_macsec_secy_exception exception,
-+ bool enable)
-+{
-+ int err;
-+ int _errno;
-+
-+ err = FM_MACSEC_SECY_ConfigException(fm_macsec_secy_dev, exception,
-+ enable);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_MACSEC_SECY_ConfigException() = 0x%08x\n",
-+ err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_macsec_secy_config_exception);
-+
-+int fm_macsec_secy_config_event(struct fm_macsec_secy_dev *fm_macsec_secy_dev,
-+ fm_macsec_secy_event event,
-+ bool enable)
-+{
-+ int err;
-+ int _errno;
-+
-+ err = FM_MACSEC_SECY_ConfigEvent(fm_macsec_secy_dev, event, enable);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_MACSEC_SECY_ConfigEvent() = 0x%08x\n",
-+ err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_macsec_secy_config_event);
-+
-+struct rx_sc_dev *fm_macsec_secy_create_rxsc(struct fm_macsec_secy_dev *fm_macsec_secy_dev,
-+ struct fm_macsec_secy_sc_params *params)
-+{
-+ struct rx_sc_dev *rx_sc_dev;
-+
-+ rx_sc_dev = FM_MACSEC_SECY_CreateRxSc(fm_macsec_secy_dev, (t_FmMacsecSecYSCParams *)params);
-+ if (unlikely(rx_sc_dev == NULL))
-+ pr_err("FM_MACSEC_SECY_CreateRxSc() failed\n");
-+
-+ return rx_sc_dev;
-+}
-+EXPORT_SYMBOL(fm_macsec_secy_create_rxsc);
-+
-+int fm_macsec_secy_delete_rxsc(struct fm_macsec_secy_dev *fm_macsec_secy_dev,
-+ struct rx_sc_dev *sc)
-+{
-+ int err;
-+ int _errno;
-+
-+ err = FM_MACSEC_SECY_DeleteRxSc(fm_macsec_secy_dev, sc);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_MACSEC_SECY_DeleteRxSc() = 0x%08x\n",
-+ err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_macsec_secy_delete_rxsc);
-+
-+int fm_macsec_secy_create_rx_sa(struct fm_macsec_secy_dev *fm_macsec_secy_dev,
-+ struct rx_sc_dev *sc, macsec_an_t an,
-+ uint32_t lowest_pn, macsec_sa_key_t key)
-+{
-+ int err;
-+ int _errno;
-+
-+ err = FM_MACSEC_SECY_CreateRxSa(fm_macsec_secy_dev, sc, an,
-+ lowest_pn, key);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_MACSEC_SECY_CreateRxSa() = 0x%08x\n",
-+ err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_macsec_secy_create_rx_sa);
-+
-+int fm_macsec_secy_delete_rx_sa(struct fm_macsec_secy_dev *fm_macsec_secy_dev,
-+ struct rx_sc_dev *sc, macsec_an_t an)
-+{
-+ int err;
-+ int _errno;
-+
-+ err = FM_MACSEC_SECY_DeleteRxSa(fm_macsec_secy_dev, sc, an);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_MACSEC_SECY_DeleteRxSa() = 0x%08x\n",
-+ err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_macsec_secy_delete_rx_sa);
-+
-+int fm_macsec_secy_rxsa_enable_receive(struct fm_macsec_secy_dev *fm_macsec_secy_dev,
-+ struct rx_sc_dev *sc,
-+ macsec_an_t an)
-+{
-+ int err;
-+ int _errno;
-+
-+ err = FM_MACSEC_SECY_RxSaEnableReceive(fm_macsec_secy_dev, sc, an);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_MACSEC_SECY_RxSaEnableReceive() = 0x%08x\n",
-+ err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_macsec_secy_rxsa_enable_receive);
-+
-+int fm_macsec_secy_rxsa_disable_receive(struct fm_macsec_secy_dev *fm_macsec_secy_dev,
-+ struct rx_sc_dev *sc,
-+ macsec_an_t an)
-+{
-+ int err;
-+ int _errno;
-+
-+ err = FM_MACSEC_SECY_RxSaDisableReceive(fm_macsec_secy_dev, sc, an);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_MACSEC_SECY_RxSaDisableReceive() = 0x%08x\n",
-+ err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_macsec_secy_rxsa_disable_receive);
-+
-+int fm_macsec_secy_rxsa_update_next_pn(struct fm_macsec_secy_dev *fm_macsec_secy_dev,
-+ struct rx_sc_dev *sc,
-+ macsec_an_t an, uint32_t updt_next_pn)
-+{
-+ int err;
-+ int _errno;
-+
-+ err = FM_MACSEC_SECY_RxSaUpdateNextPn(fm_macsec_secy_dev, sc, an,
-+ updt_next_pn);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_MACSEC_SECY_RxSaUpdateNextPn() = 0x%08x\n", err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_macsec_secy_rxsa_update_next_pn);
-+
-+int fm_macsec_secy_rxsa_update_lowest_pn(struct fm_macsec_secy_dev *fm_macsec_secy_dev,
-+ struct rx_sc_dev *sc,
-+ macsec_an_t an, uint32_t updt_lowest_pn)
-+{
-+ int err;
-+ int _errno;
-+
-+ err = FM_MACSEC_SECY_RxSaUpdateLowestPn(fm_macsec_secy_dev, sc, an,
-+ updt_lowest_pn);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_MACSEC_SECY_RxSaUpdateLowestPn() = 0x%08x\n",
-+ err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_macsec_secy_rxsa_update_lowest_pn);
-+
-+int fm_macsec_secy_rxsa_modify_key(struct fm_macsec_secy_dev *fm_macsec_secy_dev,
-+ struct rx_sc_dev *sc,
-+ macsec_an_t an, macsec_sa_key_t key)
-+{
-+ int err;
-+ int _errno;
-+
-+ err = FM_MACSEC_SECY_RxSaModifyKey(fm_macsec_secy_dev, sc, an, key);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_MACSEC_SECY_RxSaModifyKey() = 0x%08x\n",
-+ err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_macsec_secy_rxsa_modify_key);
-+
-+int fm_macsec_secy_create_tx_sa(struct fm_macsec_secy_dev *fm_macsec_secy_dev,
-+ macsec_an_t an, macsec_sa_key_t key)
-+{
-+ int err;
-+ int _errno;
-+
-+ err = FM_MACSEC_SECY_CreateTxSa(fm_macsec_secy_dev, an, key);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_MACSEC_SECY_CreateTxSa() = 0x%08x\n",
-+ err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_macsec_secy_create_tx_sa);
-+
-+int fm_macsec_secy_delete_tx_sa(struct fm_macsec_secy_dev *fm_macsec_secy_dev,
-+ macsec_an_t an)
-+{
-+ int err;
-+ int _errno;
-+
-+ err = FM_MACSEC_SECY_DeleteTxSa(fm_macsec_secy_dev, an);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_MACSEC_SECY_DeleteTxSa() = 0x%08x\n",
-+ err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_macsec_secy_delete_tx_sa);
-+
-+int fm_macsec_secy_txsa_modify_key(struct fm_macsec_secy_dev *fm_macsec_secy_dev,
-+ macsec_an_t next_active_an,
-+ macsec_sa_key_t key)
-+{
-+ int err;
-+ int _errno;
-+
-+ err = FM_MACSEC_SECY_TxSaModifyKey(fm_macsec_secy_dev, next_active_an,
-+ key);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_MACSEC_SECY_TxSaModifyKey() = 0x%08x\n",
-+ err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_macsec_secy_txsa_modify_key);
-+
-+int fm_macsec_secy_txsa_set_active(struct fm_macsec_secy_dev *fm_macsec_secy_dev,
-+ macsec_an_t an)
-+{
-+ int err;
-+ int _errno;
-+
-+ err = FM_MACSEC_SECY_TxSaSetActive(fm_macsec_secy_dev, an);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_MACSEC_SECY_TxSaSetActive() = 0x%08x\n",
-+ err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_macsec_secy_txsa_set_active);
-+
-+int fm_macsec_secy_txsa_get_active(struct fm_macsec_secy_dev *fm_macsec_secy_dev,
-+ macsec_an_t *p_an)
-+{
-+ int err;
-+ int _errno;
-+
-+ err = FM_MACSEC_SECY_TxSaGetActive(fm_macsec_secy_dev, p_an);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_MACSEC_SECY_TxSaGetActive() = 0x%08x\n",
-+ err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_macsec_secy_txsa_get_active);
-+
-+int fm_macsec_secy_get_rxsc_phys_id(struct fm_macsec_secy_dev *fm_macsec_secy_dev,
-+ struct rx_sc_dev *sc, uint32_t *sc_phys_id)
-+{
-+ int err;
-+ int _errno;
-+
-+ err = FM_MACSEC_SECY_GetRxScPhysId(fm_macsec_secy_dev, sc, sc_phys_id);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_MACSEC_SECY_GetRxScPhysId() = 0x%08x\n",
-+ err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_macsec_secy_get_rxsc_phys_id);
-+
-+int fm_macsec_secy_get_txsc_phys_id(struct fm_macsec_secy_dev *fm_macsec_secy_dev,
-+ uint32_t *sc_phys_id)
-+{
-+ int err;
-+ int _errno;
-+
-+ err = FM_MACSEC_SECY_GetTxScPhysId(fm_macsec_secy_dev, sc_phys_id);
-+ _errno = -GET_ERROR_TYPE(err);
-+ if (unlikely(_errno < 0))
-+ pr_err("FM_MACSEC_SECY_GetTxScPhysId() = 0x%08x\n",
-+ err);
-+
-+ return _errno;
-+}
-+EXPORT_SYMBOL(fm_macsec_secy_get_txsc_phys_id);
-+
-+static t_Handle h_FmLnxWrp;
-+
-+static int __init __cold fm_load (void)
-+{
-+ if ((h_FmLnxWrp = LNXWRP_FM_Init()) == NULL)
-+ {
-+ printk("Failed to init FM wrapper!\n");
-+ return -ENODEV;
-+ }
-+
-+ printk(KERN_CRIT "Freescale FM module," \
-+ " FMD API version %d.%d.%d\n",
-+ FMD_API_VERSION_MAJOR,
-+ FMD_API_VERSION_MINOR,
-+ FMD_API_VERSION_RESPIN);
-+ return 0;
-+}
-+
-+static void __exit __cold fm_unload (void)
-+{
-+ if (h_FmLnxWrp)
-+ LNXWRP_FM_Free(h_FmLnxWrp);
-+}
-+
-+module_init (fm_load);
-+module_exit (fm_unload);
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/src/wrapper/lnxwrp_fm.h
-@@ -0,0 +1,294 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+/*
-+ @File lnxwrp_fm.h
-+
-+ @Author Shlomi Gridish
-+
-+ @Description FM Linux wrapper functions.
-+
-+*/
-+
-+#ifndef __LNXWRP_FM_H__
-+#define __LNXWRP_FM_H__
-+
-+#include <linux/fsl_qman.h> /* struct qman_fq */
-+
-+#include "std_ext.h"
-+#include "error_ext.h"
-+#include "list_ext.h"
-+
-+#include "lnxwrp_fm_ext.h"
-+
-+#define FM_MAX_NUM_OF_ADV_SETTINGS 10
-+
-+#define LNXWRP_FM_NUM_OF_SHARED_PROFILES 16
-+
-+#if defined(CONFIG_FMAN_DISABLE_OH_TO_REUSE_RESOURCES)
-+#define FM_10G_OPENDMA_MIN_TRESHOLD 8 /* 10g minimum treshold if only HC is enabled and no OH port enabled */
-+#define FM_OPENDMA_RX_TX_RAPORT 2 /* RX = 2*TX */
-+#else
-+#define FM_10G_OPENDMA_MIN_TRESHOLD 7 /* 10g minimum treshold if 7 OH ports are enabled */
-+#define FM_OPENDMA_RX_TX_RAPORT 1 /* RX = TX */
-+#endif
-+#define FM_DEFAULT_TX10G_OPENDMA 8 /* default TX 10g open dmas */
-+#define FM_DEFAULT_RX10G_OPENDMA 8 /* default RX 10g open dmas */
-+
-+#define FRAG_MANIP_SPACE 128
-+#define FRAG_DATA_ALIGN 64
-+
-+#ifndef CONFIG_FSL_FM_MAX_FRAME_SIZE
-+#define CONFIG_FSL_FM_MAX_FRAME_SIZE 0
-+#endif
-+
-+#ifndef CONFIG_FSL_FM_RX_EXTRA_HEADROOM
-+#define CONFIG_FSL_FM_RX_EXTRA_HEADROOM 16
-+#endif
-+
-+typedef enum {
-+ e_NO_PCD = 0,
-+ e_FM_PCD_3_TUPLE
-+} e_LnxWrpFmPortPcdDefUseCase;
-+
-+
-+typedef struct t_FmTestFq {
-+ struct qman_fq fq_base;
-+ t_Handle h_Arg;
-+} t_FmTestFq;
-+
-+typedef struct {
-+ uint8_t id; /* sw port id, see SW_PORT_ID_TO_HW_PORT_ID() in fm_common.h */
-+ int minor;
-+ char name[20];
-+ bool active;
-+ uint64_t phys_baseAddr;
-+ uint64_t baseAddr; /* Port's *virtual* address */
-+ uint32_t memSize;
-+ t_WrpFmPortDevSettings settings;
-+ t_FmExtPools opExtPools;
-+ uint8_t totalNumOfSchemes;
-+ uint8_t schemesBase;
-+ uint8_t numOfSchemesUsed;
-+ uint32_t pcdBaseQ;
-+ uint16_t pcdNumOfQs;
-+ struct fm_port_pcd_param pcd_owner_params;
-+ e_LnxWrpFmPortPcdDefUseCase defPcd;
-+ t_Handle h_DefNetEnv;
-+ t_Handle h_Schemes[FM_PCD_KG_NUM_OF_SCHEMES];
-+ t_FmBufferPrefixContent buffPrefixContent;
-+ t_Handle h_Dev;
-+ t_Handle h_DfltVsp;
-+ t_Handle h_LnxWrpFmDev;
-+ uint16_t txCh;
-+ struct device *dev;
-+ struct device_attribute *dev_attr_stats;
-+ struct device_attribute *dev_attr_regs;
-+ struct device_attribute *dev_attr_bmi_regs;
-+ struct device_attribute *dev_attr_qmi_regs;
-+#if (DPAA_VERSION >= 11)
-+ struct device_attribute *dev_attr_ipv4_opt;
-+#endif
-+ struct device_attribute *dev_attr_dsar_regs;
-+ struct device_attribute *dev_attr_dsar_mem;
-+ struct auto_res_tables_sizes dsar_table_sizes;
-+} t_LnxWrpFmPortDev;
-+
-+typedef struct {
-+ uint8_t id;
-+ bool active;
-+ uint64_t baseAddr;
-+ uint32_t memSize;
-+ t_WrpFmMacDevSettings settings;
-+ t_Handle h_Dev;
-+ t_Handle h_LnxWrpFmDev;
-+} t_LnxWrpFmMacDev;
-+
-+/* information about all active ports for an FMan.
-+ * !Some ports may be disabled by u-boot, thus will not be available */
-+struct fm_active_ports {
-+ uint32_t num_oh_ports;
-+ uint32_t num_tx_ports;
-+ uint32_t num_rx_ports;
-+ uint32_t num_tx25_ports;
-+ uint32_t num_rx25_ports;
-+ uint32_t num_tx10_ports;
-+ uint32_t num_rx10_ports;
-+};
-+
-+/* FMan resources precalculated at fm probe based
-+ * on available FMan port. */
-+struct fm_resource_settings {
-+ /* buffers - fifo sizes */
-+ uint32_t tx1g_num_buffers;
-+ uint32_t rx1g_num_buffers;
-+ uint32_t tx2g5_num_buffers; /* Not supported yet by LLD */
-+ uint32_t rx2g5_num_buffers; /* Not supported yet by LLD */
-+ uint32_t tx10g_num_buffers;
-+ uint32_t rx10g_num_buffers;
-+ uint32_t oh_num_buffers;
-+ uint32_t shared_ext_buffers;
-+
-+ /* open DMAs */
-+ uint32_t tx_1g_dmas;
-+ uint32_t rx_1g_dmas;
-+ uint32_t tx_2g5_dmas; /* Not supported yet by LLD */
-+ uint32_t rx_2g5_dmas; /* Not supported yet by LLD */
-+ uint32_t tx_10g_dmas;
-+ uint32_t rx_10g_dmas;
-+ uint32_t oh_dmas;
-+ uint32_t shared_ext_open_dma;
-+
-+ /* Tnums */
-+ uint32_t tx_1g_tnums;
-+ uint32_t rx_1g_tnums;
-+ uint32_t tx_2g5_tnums; /* Not supported yet by LLD */
-+ uint32_t rx_2g5_tnums; /* Not supported yet by LLD */
-+ uint32_t tx_10g_tnums;
-+ uint32_t rx_10g_tnums;
-+ uint32_t oh_tnums;
-+ uint32_t shared_ext_tnums;
-+};
-+
-+typedef struct {
-+ uint8_t id;
-+ char name[10];
-+ bool active;
-+ bool pcdActive;
-+ bool prsActive;
-+ bool kgActive;
-+ bool ccActive;
-+ bool plcrActive;
-+ e_LnxWrpFmPortPcdDefUseCase defPcd;
-+ uint32_t usedSchemes;
-+ uint8_t totalNumOfSharedSchemes;
-+ uint8_t sharedSchemesBase;
-+ uint8_t numOfSchemesUsed;
-+ uint8_t defNetEnvId;
-+ uint64_t fmPhysBaseAddr;
-+ uint64_t fmBaseAddr;
-+ uint32_t fmMemSize;
-+ uint64_t fmMuramPhysBaseAddr;
-+ uint64_t fmMuramBaseAddr;
-+ uint32_t fmMuramMemSize;
-+ uint64_t fmRtcPhysBaseAddr;
-+ uint64_t fmRtcBaseAddr;
-+ uint32_t fmRtcMemSize;
-+ uint64_t fmVspPhysBaseAddr;
-+ uint64_t fmVspBaseAddr;
-+ uint32_t fmVspMemSize;
-+ int irq;
-+ int err_irq;
-+ t_WrpFmDevSettings fmDevSettings;
-+ t_WrpFmPcdDevSettings fmPcdDevSettings;
-+ t_Handle h_Dev;
-+ uint16_t hcCh;
-+
-+ t_Handle h_MuramDev;
-+ t_Handle h_PcdDev;
-+ t_Handle h_RtcDev;
-+
-+ t_Handle h_DsarRxPort;
-+ t_Handle h_DsarTxPort;
-+
-+ t_LnxWrpFmPortDev hcPort;
-+ t_LnxWrpFmPortDev opPorts[FM_MAX_NUM_OF_OH_PORTS-1];
-+ t_LnxWrpFmPortDev rxPorts[FM_MAX_NUM_OF_RX_PORTS];
-+ t_LnxWrpFmPortDev txPorts[FM_MAX_NUM_OF_TX_PORTS];
-+ t_LnxWrpFmMacDev macs[FM_MAX_NUM_OF_MACS];
-+ struct fm_active_ports fm_active_ports_info;
-+ struct fm_resource_settings fm_resource_settings_info;
-+
-+ struct device *dev;
-+ struct resource *res;
-+ int major;
-+ struct class *fm_class;
-+ struct device_attribute *dev_attr_stats;
-+ struct device_attribute *dev_attr_regs;
-+ struct device_attribute *dev_attr_risc_load;
-+
-+ struct device_attribute *dev_pcd_attr_stats;
-+ struct device_attribute *dev_plcr_attr_regs;
-+ struct device_attribute *dev_prs_attr_regs;
-+ struct device_attribute *dev_fm_fpm_attr_regs;
-+ struct device_attribute *dev_fm_kg_attr_regs;
-+ struct device_attribute *dev_fm_kg_pe_attr_regs;
-+ struct device_attribute *dev_attr_muram_free_size;
-+ struct device_attribute *dev_attr_fm_ctrl_code_ver;
-+
-+
-+ struct qman_fq *hc_tx_conf_fq, *hc_tx_err_fq, *hc_tx_fq;
-+} t_LnxWrpFmDev;
-+
-+typedef struct {
-+ t_LnxWrpFmDev *p_FmDevs[INTG_MAX_NUM_OF_FM];
-+} t_LnxWrpFm;
-+#define LNXWRP_FM_OBJECT(ptr) LIST_OBJECT(ptr, t_LnxWrpFm, fms[((t_LnxWrpFmDev *)ptr)->id])
-+
-+
-+t_Error LnxwrpFmIOCTL(t_LnxWrpFmDev *p_LnxWrpFmDev, unsigned int cmd, unsigned long arg, bool compat);
-+t_Error LnxwrpFmPortIOCTL(t_LnxWrpFmPortDev *p_LnxWrpFmPortDev, unsigned int cmd, unsigned long arg, bool compat);
-+
-+
-+#if 0
-+static __inline__ t_Error AllocSchemesForPort(t_LnxWrpFmDev *p_LnxWrpFmDev, uint8_t numSchemes, uint8_t *p_BaseSchemeNum)
-+{
-+ uint32_t schemeMask;
-+ uint8_t i;
-+
-+ if (!numSchemes)
-+ RETURN_ERROR(MINOR, E_INVALID_VALUE, NO_MSG);
-+
-+ schemeMask = 0x80000000;
-+ *p_BaseSchemeNum = 0xff;
-+
-+ for (i=0; schemeMask && numSchemes; schemeMask>>=1, i++)
-+ if ((p_LnxWrpFmDev->usedSchemes & schemeMask) == 0)
-+ {
-+ p_LnxWrpFmDev->usedSchemes |= schemeMask;
-+ numSchemes--;
-+ if (*p_BaseSchemeNum==0xff)
-+ *p_BaseSchemeNum = i;
-+ }
-+ else if (*p_BaseSchemeNum!=0xff)
-+ RETURN_ERROR(MINOR, E_INVALID_STATE, ("Fragmentation on schemes array!!!"));
-+
-+ if (numSchemes)
-+ RETURN_ERROR(MINOR, E_FULL, ("schemes!!!"));
-+ return E_OK;
-+}
-+#endif
-+
-+void LnxWrpPCDIOCTLTypeChecking(void);
-+void LnxWrpPCDIOCTLEnumChecking(void);
-+
-+#endif /* __LNXWRP_FM_H__ */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/src/wrapper/lnxwrp_fm_port.c
-@@ -0,0 +1,1512 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+/*
-+ @File lnxwrp_fm_port.c
-+
-+ @Description FMD wrapper - FMan port functions.
-+
-+*/
-+
-+#include <linux/version.h>
-+#if defined(CONFIG_MODVERSIONS) && !defined(MODVERSIONS)
-+#define MODVERSIONS
-+#endif
-+#ifdef MODVERSIONS
-+#include <config/modversions.h>
-+#endif /* MODVERSIONS */
-+#include <linux/kernel.h>
-+#include <linux/module.h>
-+#include <linux/of_platform.h>
-+#include <linux/of_address.h>
-+#include <linux/cdev.h>
-+#include <linux/slab.h>
-+#include <linux/spinlock.h>
-+#ifndef CONFIG_FMAN_ARM
-+#include <linux/fsl/svr.h>
-+#endif
-+#include <linux/io.h>
-+
-+#include "sprint_ext.h"
-+#include "fm_common.h"
-+#include "lnxwrp_fsl_fman.h"
-+#include "fm_port_ext.h"
-+#if (DPAA_VERSION >= 11)
-+#include "fm_vsp_ext.h"
-+#endif /* DPAA_VERSION >= 11 */
-+#include "fm_ioctls.h"
-+#include "lnxwrp_resources.h"
-+#include "lnxwrp_sysfs_fm_port.h"
-+
-+#define __ERR_MODULE__ MODULE_FM
-+
-+extern struct device_node *GetFmAdvArgsDevTreeNode (uint8_t fmIndx);
-+
-+/* TODO: duplicated, see lnxwrp_fm.c */
-+#define ADD_ADV_CONFIG_NO_RET(_func, _param)\
-+do {\
-+ if (i < max) {\
-+ p_Entry = &p_Entrys[i];\
-+ p_Entry->p_Function = _func;\
-+ _param\
-+ i++;\
-+ } else {\
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE,\
-+ ("Number of advanced-configuration entries exceeded"));\
-+ } \
-+} while (0)
-+
-+#ifndef CONFIG_FMAN_ARM
-+#define IS_T1023_T1024 (SVR_SOC_VER(mfspr(SPRN_SVR)) == SVR_T1024 || \
-+ SVR_SOC_VER(mfspr(SPRN_SVR)) == SVR_T1023)
-+#endif
-+
-+static volatile int hcFrmRcv/* = 0 */;
-+static spinlock_t lock;
-+
-+static enum qman_cb_dqrr_result qm_tx_conf_dqrr_cb(struct qman_portal *portal,
-+ struct qman_fq *fq,
-+ const struct qm_dqrr_entry
-+ *dq)
-+{
-+ t_LnxWrpFmDev *p_LnxWrpFmDev = ((t_FmTestFq *) fq)->h_Arg;
-+ unsigned long flags;
-+
-+#if __BYTE_ORDER__ == __ORDER_LITTLE_ENDIAN__
-+{
-+ /* extract the HC frame address */
-+ uint32_t *hcf_va = XX_PhysToVirt(qm_fd_addr((struct qm_fd *)&dq->fd));
-+ int hcf_l = ((struct qm_fd *)&dq->fd)->length20;
-+ int i;
-+
-+ /* 32b byteswap of all data in the HC Frame */
-+ for(i = 0; i < hcf_l / 4; ++i)
-+ hcf_va[i] =
-+ ___constant_swab32(hcf_va[i]);
-+}
-+#endif
-+ FM_PCD_HcTxConf(p_LnxWrpFmDev->h_PcdDev, (t_DpaaFD *)&dq->fd);
-+ spin_lock_irqsave(&lock, flags);
-+ hcFrmRcv--;
-+ spin_unlock_irqrestore(&lock, flags);
-+
-+ return qman_cb_dqrr_consume;
-+}
-+
-+static enum qman_cb_dqrr_result qm_tx_dqrr_cb(struct qman_portal *portal,
-+ struct qman_fq *fq,
-+ const struct qm_dqrr_entry *dq)
-+{
-+ WARN(1, "FMD: failure at %s:%d/%s()!\n", __FILE__, __LINE__,
-+ __func__);
-+ return qman_cb_dqrr_consume;
-+}
-+
-+static void qm_err_cb(struct qman_portal *portal,
-+ struct qman_fq *fq, const struct qm_mr_entry *msg)
-+{
-+ WARN(1, "FMD: failure at %s:%d/%s()!\n", __FILE__, __LINE__,
-+ __func__);
-+}
-+
-+static struct qman_fq *FqAlloc(t_LnxWrpFmDev * p_LnxWrpFmDev,
-+ uint32_t fqid,
-+ uint32_t flags, uint16_t channel, uint8_t wq)
-+{
-+ int _errno;
-+ struct qman_fq *fq = NULL;
-+ t_FmTestFq *p_FmtFq;
-+ struct qm_mcc_initfq initfq;
-+
-+ p_FmtFq = (t_FmTestFq *) XX_Malloc(sizeof(t_FmTestFq));
-+ if (!p_FmtFq) {
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("FQ obj!!!"));
-+ return NULL;
-+ }
-+
-+ p_FmtFq->fq_base.cb.dqrr = ((flags & QMAN_FQ_FLAG_NO_ENQUEUE)
-+ ? qm_tx_conf_dqrr_cb
-+ : qm_tx_dqrr_cb);
-+ p_FmtFq->fq_base.cb.ern = qm_err_cb;
-+ /* p_FmtFq->fq_base.cb.fqs = qm_err_cb; */
-+ /* qm_err_cb wrongly called when the FQ is parked */
-+ p_FmtFq->fq_base.cb.fqs = NULL;
-+ p_FmtFq->h_Arg = (t_Handle) p_LnxWrpFmDev;
-+ if (fqid == 0) {
-+ flags |= QMAN_FQ_FLAG_DYNAMIC_FQID;
-+ flags &= ~QMAN_FQ_FLAG_NO_MODIFY;
-+ } else {
-+ flags &= ~QMAN_FQ_FLAG_DYNAMIC_FQID;
-+ }
-+
-+ if (qman_create_fq(fqid, flags, &p_FmtFq->fq_base)) {
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY, ("FQ obj - qman_new_fq!!!"));
-+ XX_Free(p_FmtFq);
-+ return NULL;
-+ }
-+ fq = &p_FmtFq->fq_base;
-+
-+ if (!(flags & QMAN_FQ_FLAG_NO_MODIFY)) {
-+ initfq.we_mask = QM_INITFQ_WE_DESTWQ;
-+ initfq.fqd.dest.channel = channel;
-+ initfq.fqd.dest.wq = wq;
-+
-+ _errno = qman_init_fq(fq, QMAN_INITFQ_FLAG_SCHED, &initfq);
-+ if (unlikely(_errno < 0)) {
-+ REPORT_ERROR(MAJOR, E_NO_MEMORY,
-+ ("FQ obj - qman_init_fq!!!"));
-+ qman_destroy_fq(fq, 0);
-+ XX_Free(p_FmtFq);
-+ return NULL;
-+ }
-+ }
-+
-+ DBG(TRACE,
-+ ("fqid %d, flags 0x%08x, channel %d, wq %d", qman_fq_fqid(fq),
-+ flags, channel, wq));
-+
-+ return fq;
-+}
-+
-+static void FqFree(struct qman_fq *fq)
-+{
-+ int _errno;
-+
-+ _errno = qman_retire_fq(fq, NULL);
-+ if (unlikely(_errno < 0))
-+ printk(KERN_WARNING "qman_retire_fq(%u) = %d\n", qman_fq_fqid(fq), _errno);
-+
-+ _errno = qman_oos_fq(fq);
-+ if (unlikely(_errno < 0))
-+ printk(KERN_WARNING "qman_oos_fq(%u) = %d\n", qman_fq_fqid(fq), _errno);
-+
-+ qman_destroy_fq(fq, 0);
-+ XX_Free((t_FmTestFq *) fq);
-+}
-+
-+static t_Error QmEnqueueCB(t_Handle h_Arg, void *p_Fd)
-+{
-+ t_LnxWrpFmDev *p_LnxWrpFmDev = (t_LnxWrpFmDev *) h_Arg;
-+ int _errno, timeout = 1000000;
-+ unsigned long flags;
-+
-+ ASSERT_COND(p_LnxWrpFmDev);
-+
-+ spin_lock_irqsave(&lock, flags);
-+ hcFrmRcv++;
-+ spin_unlock_irqrestore(&lock, flags);
-+
-+#if __BYTE_ORDER__ == __ORDER_LITTLE_ENDIAN__
-+{
-+ /* extract the HC frame address */
-+ uint32_t *hcf_va = XX_PhysToVirt(qm_fd_addr((struct qm_fd *) p_Fd));
-+ int hcf_l = ((struct qm_fd *)p_Fd)->length20;
-+ int i;
-+
-+ /* 32b byteswap of all data in the HC Frame */
-+ for(i = 0; i < hcf_l / 4; ++i)
-+ hcf_va[i] =
-+ ___constant_swab32(hcf_va[i]);
-+}
-+#endif
-+
-+ _errno = qman_enqueue(p_LnxWrpFmDev->hc_tx_fq, (struct qm_fd *) p_Fd,
-+ 0);
-+ if (_errno)
-+ RETURN_ERROR(MINOR, E_INVALID_STATE,
-+ ("qman_enqueue() failed"));
-+
-+ while (hcFrmRcv && --timeout) {
-+ udelay(1);
-+ cpu_relax();
-+ }
-+ if (timeout == 0) {
-+ dump_stack();
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED,
-+ ("timeout waiting for Tx confirmation"));
-+ return E_WRITE_FAILED;
-+ }
-+
-+ return E_OK;
-+}
-+
-+static t_LnxWrpFmPortDev *ReadFmPortDevTreeNode(struct platform_device
-+ *of_dev)
-+{
-+ t_LnxWrpFmDev *p_LnxWrpFmDev;
-+ t_LnxWrpFmPortDev *p_LnxWrpFmPortDev;
-+ struct device_node *fm_node, *port_node;
-+ struct resource res;
-+ const uint32_t *uint32_prop;
-+ int _errno = 0, lenp;
-+ uint32_t tmp_prop;
-+
-+#ifdef CONFIG_FMAN_P1023
-+ static unsigned char have_oh_port/* = 0 */;
-+#endif
-+
-+ port_node = of_node_get(of_dev->dev.of_node);
-+
-+ /* Get the FM node */
-+ fm_node = of_get_parent(port_node);
-+ if (unlikely(fm_node == NULL)) {
-+ REPORT_ERROR(MAJOR, E_NO_DEVICE,
-+ ("of_get_parent() = %d", _errno));
-+ return NULL;
-+ }
-+
-+ p_LnxWrpFmDev =
-+ dev_get_drvdata(&of_find_device_by_node(fm_node)->dev);
-+ of_node_put(fm_node);
-+
-+ /* if fm_probe() failed, no point in going further with port probing */
-+ if (p_LnxWrpFmDev == NULL)
-+ return NULL;
-+
-+ uint32_prop =
-+ (uint32_t *) of_get_property(port_node, "cell-index", &lenp);
-+ if (unlikely(uint32_prop == NULL)) {
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE,
-+ ("of_get_property(%s, cell-index) failed",
-+ port_node->full_name));
-+ return NULL;
-+ }
-+ tmp_prop = be32_to_cpu(*uint32_prop);
-+ if (WARN_ON(lenp != sizeof(uint32_t)))
-+ return NULL;
-+ if (of_device_is_compatible(port_node, "fsl,fman-port-oh") ||
-+ of_device_is_compatible(port_node, "fsl,fman-v2-port-oh") ||
-+ of_device_is_compatible(port_node, "fsl,fman-v3-port-oh")) {
-+#ifndef CONFIG_FMAN_ARM
-+#ifdef CONFIG_FMAN_P3040_P4080_P5020
-+ /* On PPC FMan v2, OH ports start from cell-index 0x1 */
-+ tmp_prop -= 0x1;
-+#else
-+ /* On PPC FMan v3 (Low and High), OH ports start from
-+ * cell-index 0x2
-+ */
-+ tmp_prop -= 0x2;
-+#endif // CONFIG_FMAN_P3040_P4080_P5020
-+#endif // CONFIG_FMAN_ARM
-+
-+ if (unlikely(tmp_prop >= FM_MAX_NUM_OF_OH_PORTS)) {
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE,
-+ ("of_get_property(%s, cell-index) failed",
-+ port_node->full_name));
-+ return NULL;
-+ }
-+
-+#ifdef CONFIG_FMAN_P1023
-+ /* Beware, this can be done when there is only
-+ one FMan to be initialized */
-+ if (!have_oh_port) {
-+ have_oh_port = 1; /* first OP/HC port
-+ is used for host command */
-+#else
-+ /* Here it is hardcoded the use of the OH port 1
-+ (with cell-index 0) */
-+ if (tmp_prop == 0) {
-+#endif
-+ p_LnxWrpFmPortDev = &p_LnxWrpFmDev->hcPort;
-+ p_LnxWrpFmPortDev->id = 0;
-+ /*
-+ p_LnxWrpFmPortDev->id = *uint32_prop-1;
-+ p_LnxWrpFmPortDev->id = *uint32_prop;
-+ */
-+ p_LnxWrpFmPortDev->settings.param.portType =
-+ e_FM_PORT_TYPE_OH_HOST_COMMAND;
-+ } else {
-+ p_LnxWrpFmPortDev =
-+ &p_LnxWrpFmDev->opPorts[tmp_prop - 1];
-+ p_LnxWrpFmPortDev->id = tmp_prop- 1;
-+ p_LnxWrpFmPortDev->settings.param.portType =
-+ e_FM_PORT_TYPE_OH_OFFLINE_PARSING;
-+ }
-+ p_LnxWrpFmPortDev->settings.param.portId = tmp_prop;
-+
-+ uint32_prop =
-+ (uint32_t *) of_get_property(port_node,
-+ "fsl,qman-channel-id",
-+ &lenp);
-+ if (uint32_prop == NULL) {
-+ /*
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE, ("missing fsl,qman-channel-id"));
-+ */
-+ XX_Print("FM warning: missing fsl,qman-channel-id"
-+ " for OH port.\n");
-+ return NULL;
-+ }
-+ tmp_prop = be32_to_cpu(*uint32_prop);
-+ if (WARN_ON(lenp != sizeof(uint32_t)))
-+ return NULL;
-+ p_LnxWrpFmPortDev->txCh = tmp_prop;
-+
-+ p_LnxWrpFmPortDev->settings.param.specificParams.nonRxParams.
-+ qmChannel = p_LnxWrpFmPortDev->txCh;
-+ } else if (of_device_is_compatible(port_node, "fsl,fman-port-1g-tx")) {
-+ tmp_prop -= 0x28;
-+ if (unlikely(tmp_prop >= FM_MAX_NUM_OF_1G_TX_PORTS)) {
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE,
-+ ("of_get_property(%s, cell-index) failed",
-+ port_node->full_name));
-+ return NULL;
-+ }
-+ p_LnxWrpFmPortDev = &p_LnxWrpFmDev->txPorts[tmp_prop];
-+
-+ p_LnxWrpFmPortDev->id = tmp_prop;
-+ p_LnxWrpFmPortDev->settings.param.portId =
-+ p_LnxWrpFmPortDev->id;
-+ p_LnxWrpFmPortDev->settings.param.portType = e_FM_PORT_TYPE_TX;
-+
-+ uint32_prop = (uint32_t *) of_get_property(port_node,
-+ "fsl,qman-channel-id", &lenp);
-+ if (uint32_prop == NULL) {
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE,
-+ ("missing fsl,qman-channel-id"));
-+ return NULL;
-+ }
-+ tmp_prop = be32_to_cpu(*uint32_prop);
-+ if (WARN_ON(lenp != sizeof(uint32_t)))
-+ return NULL;
-+ p_LnxWrpFmPortDev->txCh = tmp_prop;
-+ p_LnxWrpFmPortDev->
-+ settings.param.specificParams.nonRxParams.qmChannel =
-+ p_LnxWrpFmPortDev->txCh;
-+ } else if (of_device_is_compatible(port_node, "fsl,fman-port-10g-tx")) {
-+#ifndef CONFIG_FMAN_ARM
-+ /* On T102x, the 10G TX port IDs start from 0x28 */
-+ if (IS_T1023_T1024)
-+ tmp_prop -= 0x28;
-+ else
-+#endif
-+ tmp_prop -= 0x30;
-+
-+ if (unlikely(tmp_prop>= FM_MAX_NUM_OF_10G_TX_PORTS)) {
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE,
-+ ("of_get_property(%s, cell-index) failed",
-+ port_node->full_name));
-+ return NULL;
-+ }
-+ p_LnxWrpFmPortDev = &p_LnxWrpFmDev->txPorts[tmp_prop +
-+ FM_MAX_NUM_OF_1G_TX_PORTS];
-+#ifndef CONFIG_FMAN_ARM
-+ if (IS_T1023_T1024)
-+ p_LnxWrpFmPortDev = &p_LnxWrpFmDev->txPorts[tmp_prop];
-+#endif
-+
-+ p_LnxWrpFmPortDev->id = tmp_prop;
-+ p_LnxWrpFmPortDev->settings.param.portId =
-+ p_LnxWrpFmPortDev->id;
-+ p_LnxWrpFmPortDev->settings.param.portType =
-+ e_FM_PORT_TYPE_TX_10G;
-+ uint32_prop = (uint32_t *) of_get_property(port_node,
-+ "fsl,qman-channel-id", &lenp);
-+ if (uint32_prop == NULL) {
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE,
-+ ("missing fsl,qman-channel-id"));
-+ return NULL;
-+ }
-+ tmp_prop = be32_to_cpu(*uint32_prop);
-+ if (WARN_ON(lenp != sizeof(uint32_t)))
-+ return NULL;
-+ p_LnxWrpFmPortDev->txCh = tmp_prop;
-+ p_LnxWrpFmPortDev->settings.param.specificParams.nonRxParams.
-+ qmChannel = p_LnxWrpFmPortDev->txCh;
-+ } else if (of_device_is_compatible(port_node, "fsl,fman-port-1g-rx")) {
-+ tmp_prop -= 0x08;
-+ if (unlikely(tmp_prop >= FM_MAX_NUM_OF_1G_RX_PORTS)) {
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE,
-+ ("of_get_property(%s, cell-index) failed",
-+ port_node->full_name));
-+ return NULL;
-+ }
-+ p_LnxWrpFmPortDev = &p_LnxWrpFmDev->rxPorts[tmp_prop];
-+
-+ p_LnxWrpFmPortDev->id = tmp_prop;
-+ p_LnxWrpFmPortDev->settings.param.portId =
-+ p_LnxWrpFmPortDev->id;
-+ p_LnxWrpFmPortDev->settings.param.portType = e_FM_PORT_TYPE_RX;
-+ if (p_LnxWrpFmDev->pcdActive)
-+ p_LnxWrpFmPortDev->defPcd = p_LnxWrpFmDev->defPcd;
-+ } else if (of_device_is_compatible(port_node, "fsl,fman-port-10g-rx")) {
-+#ifndef CONFIG_FMAN_ARM
-+ /* On T102x, the 10G RX port IDs start from 0x08 */
-+ if (IS_T1023_T1024)
-+ tmp_prop -= 0x8;
-+ else
-+#endif
-+ tmp_prop -= 0x10;
-+
-+ if (unlikely(tmp_prop >= FM_MAX_NUM_OF_10G_RX_PORTS)) {
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE,
-+ ("of_get_property(%s, cell-index) failed",
-+ port_node->full_name));
-+ return NULL;
-+ }
-+ p_LnxWrpFmPortDev = &p_LnxWrpFmDev->rxPorts[tmp_prop +
-+ FM_MAX_NUM_OF_1G_RX_PORTS];
-+
-+#ifndef CONFIG_FMAN_ARM
-+ if (IS_T1023_T1024)
-+ p_LnxWrpFmPortDev = &p_LnxWrpFmDev->rxPorts[tmp_prop];
-+#endif
-+
-+ p_LnxWrpFmPortDev->id = tmp_prop;
-+ p_LnxWrpFmPortDev->settings.param.portId =
-+ p_LnxWrpFmPortDev->id;
-+ p_LnxWrpFmPortDev->settings.param.portType =
-+ e_FM_PORT_TYPE_RX_10G;
-+ if (p_LnxWrpFmDev->pcdActive)
-+ p_LnxWrpFmPortDev->defPcd = p_LnxWrpFmDev->defPcd;
-+ } else {
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE, ("Illegal port type"));
-+ return NULL;
-+ }
-+
-+ _errno = of_address_to_resource(port_node, 0, &res);
-+ if (unlikely(_errno < 0)) {
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE,
-+ ("of_address_to_resource() = %d", _errno));
-+ return NULL;
-+ }
-+
-+ p_LnxWrpFmPortDev->dev = &of_dev->dev;
-+ p_LnxWrpFmPortDev->baseAddr = 0;
-+ p_LnxWrpFmPortDev->phys_baseAddr = res.start;
-+ p_LnxWrpFmPortDev->memSize = res.end + 1 - res.start;
-+ p_LnxWrpFmPortDev->settings.param.h_Fm = p_LnxWrpFmDev->h_Dev;
-+ p_LnxWrpFmPortDev->h_LnxWrpFmDev = (t_Handle) p_LnxWrpFmDev;
-+
-+ of_node_put(port_node);
-+
-+ p_LnxWrpFmPortDev->active = TRUE;
-+
-+#if defined(CONFIG_FMAN_DISABLE_OH_TO_REUSE_RESOURCES)
-+ /* for performance mode no OH port available. */
-+ if (p_LnxWrpFmPortDev->settings.param.portType ==
-+ e_FM_PORT_TYPE_OH_OFFLINE_PARSING)
-+ p_LnxWrpFmPortDev->active = FALSE;
-+#endif
-+
-+ return p_LnxWrpFmPortDev;
-+}
-+
-+struct device_node * GetFmPortAdvArgsDevTreeNode (struct device_node *fm_node,
-+ e_FmPortType portType,
-+ uint8_t portId)
-+{
-+ struct device_node *port_node;
-+ const uint32_t *uint32_prop;
-+ int lenp;
-+ char *portTypeString;
-+ uint32_t tmp_prop;
-+
-+ switch(portType) {
-+ case e_FM_PORT_TYPE_OH_OFFLINE_PARSING:
-+ portTypeString = "fsl,fman-port-op-extended-args";
-+ break;
-+ case e_FM_PORT_TYPE_TX:
-+ portTypeString = "fsl,fman-port-1g-tx-extended-args";
-+ break;
-+ case e_FM_PORT_TYPE_TX_10G:
-+ portTypeString = "fsl,fman-port-10g-tx-extended-args";
-+ break;
-+ case e_FM_PORT_TYPE_RX:
-+ portTypeString = "fsl,fman-port-1g-rx-extended-args";
-+ break;
-+ case e_FM_PORT_TYPE_RX_10G:
-+ portTypeString = "fsl,fman-port-10g-rx-extended-args";
-+ break;
-+ default:
-+ return NULL;
-+ }
-+
-+ for_each_child_of_node(fm_node, port_node) {
-+ uint32_prop = (uint32_t *)of_get_property(port_node, "cell-index", &lenp);
-+ if (unlikely(uint32_prop == NULL)) {
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE,
-+ ("of_get_property(%s, cell-index) failed",
-+ port_node->full_name));
-+ return NULL;
-+ }
-+ tmp_prop = be32_to_cpu(*uint32_prop);
-+ if (WARN_ON(lenp != sizeof(uint32_t)))
-+ return NULL;
-+ if ((portId == tmp_prop) &&
-+ (of_device_is_compatible(port_node, portTypeString))) {
-+ return port_node;
-+ }
-+ }
-+
-+ return NULL;
-+}
-+
-+static t_Error CheckNConfigFmPortAdvArgs (t_LnxWrpFmPortDev *p_LnxWrpFmPortDev)
-+{
-+ struct device_node *fm_node, *port_node;
-+ t_Error err;
-+ t_FmPortRsrc portRsrc;
-+ const uint32_t *uint32_prop;
-+ /*const char *str_prop;*/
-+ int lenp;
-+#ifdef CONFIG_FMAN_PFC
-+ uint8_t i, id, num_pools;
-+ t_FmBufPoolDepletion poolDepletion;
-+
-+ if (p_LnxWrpFmPortDev->settings.param.portType == e_FM_PORT_TYPE_RX ||
-+ p_LnxWrpFmPortDev->settings.param.portType == e_FM_PORT_TYPE_RX_10G) {
-+ memset(&poolDepletion, 0, sizeof(t_FmBufPoolDepletion));
-+ poolDepletion.singlePoolModeEnable = true;
-+ num_pools = p_LnxWrpFmPortDev->settings.param.specificParams.rxParams.
-+ extBufPools.numOfPoolsUsed;
-+ for (i = 0; i < num_pools; i++) {
-+ id = p_LnxWrpFmPortDev->settings.param.specificParams.rxParams.
-+ extBufPools.extBufPool[i].id;
-+ poolDepletion.poolsToConsiderForSingleMode[id] = true;
-+ }
-+
-+ for (i = 0; i < CONFIG_FMAN_PFC_COS_COUNT; i++)
-+ poolDepletion.pfcPrioritiesEn[i] = true;
-+
-+ err = FM_PORT_ConfigPoolDepletion(p_LnxWrpFmPortDev->h_Dev,
-+ &poolDepletion);
-+ if (err != E_OK)
-+ RETURN_ERROR(MAJOR, err, ("FM_PORT_ConfigPoolDepletion() failed"));
-+ }
-+#endif
-+
-+ fm_node = GetFmAdvArgsDevTreeNode(((t_LnxWrpFmDev *) p_LnxWrpFmPortDev->h_LnxWrpFmDev)->id);
-+ if (!fm_node) /* no advance parameters for FMan */
-+ return E_OK;
-+
-+ port_node = GetFmPortAdvArgsDevTreeNode(fm_node,
-+ p_LnxWrpFmPortDev->settings.param.portType,
-+ p_LnxWrpFmPortDev->settings.param.portId);
-+ if (!port_node) /* no advance parameters for FMan-Port */
-+ return E_OK;
-+
-+ uint32_prop = (uint32_t *)of_get_property(port_node, "num-tnums", &lenp);
-+ if (uint32_prop) {
-+ if (WARN_ON(lenp != sizeof(uint32_t)*2))
-+ RETURN_ERROR(MINOR, E_INVALID_VALUE, NO_MSG);
-+
-+ portRsrc.num = be32_to_cpu(uint32_prop[0]);
-+ portRsrc.extra = be32_to_cpu(uint32_prop[1]);
-+
-+ if ((err = FM_PORT_ConfigNumOfTasks(p_LnxWrpFmPortDev->h_Dev,
-+ &portRsrc)) != E_OK)
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+ }
-+
-+ uint32_prop = (uint32_t *)of_get_property(port_node, "num-dmas", &lenp);
-+ if (uint32_prop) {
-+ if (WARN_ON(lenp != sizeof(uint32_t)*2))
-+ RETURN_ERROR(MINOR, E_INVALID_VALUE, NO_MSG);
-+
-+ portRsrc.num = be32_to_cpu(uint32_prop[0]);
-+ portRsrc.extra = be32_to_cpu(uint32_prop[1]);
-+
-+ if ((err = FM_PORT_ConfigNumOfOpenDmas(p_LnxWrpFmPortDev->h_Dev,
-+ &portRsrc)) != E_OK)
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+ }
-+
-+ uint32_prop = (uint32_t *)of_get_property(port_node, "fifo-size", &lenp);
-+ if (uint32_prop) {
-+ if (WARN_ON(lenp != sizeof(uint32_t)*2))
-+ RETURN_ERROR(MINOR, E_INVALID_VALUE, NO_MSG);
-+
-+ portRsrc.num = be32_to_cpu(uint32_prop[0]);
-+ portRsrc.extra = be32_to_cpu(uint32_prop[1]);
-+
-+ if ((err = FM_PORT_ConfigSizeOfFifo(p_LnxWrpFmPortDev->h_Dev,
-+ &portRsrc)) != E_OK)
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+ }
-+
-+ uint32_prop = (uint32_t *)of_get_property(port_node, "errors-to-discard", &lenp);
-+ if (uint32_prop) {
-+ if (WARN_ON(lenp != sizeof(uint32_t)))
-+ RETURN_ERROR(MINOR, E_INVALID_VALUE, NO_MSG);
-+ if ((err = FM_PORT_ConfigErrorsToDiscard(p_LnxWrpFmPortDev->h_Dev,
-+ be32_to_cpu(uint32_prop[0]))) != E_OK)
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+ }
-+
-+ uint32_prop = (uint32_t *)of_get_property(port_node, "ar-tables-sizes",
-+ &lenp);
-+ if (uint32_prop) {
-+
-+ if (WARN_ON(lenp != sizeof(uint32_t)*8))
-+ RETURN_ERROR(MINOR, E_INVALID_VALUE, NO_MSG);
-+ if (WARN_ON(p_LnxWrpFmPortDev->settings.param.portType !=
-+ e_FM_PORT_TYPE_RX) &&
-+ (p_LnxWrpFmPortDev->settings.param.portType !=
-+ e_FM_PORT_TYPE_RX_10G))
-+ RETURN_ERROR(MINOR, E_INVALID_VALUE,
-+ ("Auto Response is an Rx port atribute."));
-+
-+ memset(&p_LnxWrpFmPortDev->dsar_table_sizes, 0, sizeof(struct auto_res_tables_sizes));
-+
-+ p_LnxWrpFmPortDev->dsar_table_sizes.max_num_of_arp_entries =
-+ (uint16_t)be32_to_cpu(uint32_prop[0]);
-+ p_LnxWrpFmPortDev->dsar_table_sizes.max_num_of_echo_ipv4_entries =
-+ (uint16_t)be32_to_cpu(uint32_prop[1]);
-+ p_LnxWrpFmPortDev->dsar_table_sizes.max_num_of_ndp_entries =
-+ (uint16_t)be32_to_cpu(uint32_prop[2]);
-+ p_LnxWrpFmPortDev->dsar_table_sizes.max_num_of_echo_ipv6_entries =
-+ (uint16_t)be32_to_cpu(uint32_prop[3]);
-+ p_LnxWrpFmPortDev->dsar_table_sizes.max_num_of_snmp_ipv4_entries =
-+ (uint16_t)be32_to_cpu(uint32_prop[4]);
-+ p_LnxWrpFmPortDev->dsar_table_sizes.max_num_of_snmp_ipv6_entries =
-+ (uint16_t)be32_to_cpu(uint32_prop[5]);
-+ p_LnxWrpFmPortDev->dsar_table_sizes.max_num_of_snmp_oid_entries =
-+ (uint16_t)be32_to_cpu(uint32_prop[6]);
-+ p_LnxWrpFmPortDev->dsar_table_sizes.max_num_of_snmp_char =
-+ (uint16_t)be32_to_cpu(uint32_prop[7]);
-+
-+ uint32_prop = (uint32_t *)of_get_property(port_node,
-+ "ar-filters-sizes", &lenp);
-+ if (uint32_prop) {
-+ if (WARN_ON(lenp != sizeof(uint32_t)*3))
-+ RETURN_ERROR(MINOR, E_INVALID_VALUE, NO_MSG);
-+
-+ p_LnxWrpFmPortDev->dsar_table_sizes.max_num_of_ip_prot_filtering =
-+ (uint16_t)be32_to_cpu(uint32_prop[0]);
-+ p_LnxWrpFmPortDev->dsar_table_sizes.max_num_of_tcp_port_filtering =
-+ (uint16_t)be32_to_cpu(uint32_prop[1]);
-+ p_LnxWrpFmPortDev->dsar_table_sizes.max_num_of_udp_port_filtering =
-+ (uint16_t)be32_to_cpu(uint32_prop[2]);
-+ }
-+
-+ if ((err = FM_PORT_ConfigDsarSupport(p_LnxWrpFmPortDev->h_Dev,
-+ (t_FmPortDsarTablesSizes*)&p_LnxWrpFmPortDev->dsar_table_sizes)) != E_OK)
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+ }
-+
-+ of_node_put(port_node);
-+ of_node_put(fm_node);
-+
-+ return E_OK;
-+}
-+
-+static t_Error CheckNSetFmPortAdvArgs (t_LnxWrpFmPortDev *p_LnxWrpFmPortDev)
-+{
-+ struct device_node *fm_node, *port_node;
-+ t_Error err;
-+ const uint32_t *uint32_prop;
-+ /*const char *str_prop;*/
-+ int lenp;
-+
-+ fm_node = GetFmAdvArgsDevTreeNode(((t_LnxWrpFmDev *) p_LnxWrpFmPortDev->h_LnxWrpFmDev)->id);
-+ if (!fm_node) /* no advance parameters for FMan */
-+ return E_OK;
-+
-+ port_node = GetFmPortAdvArgsDevTreeNode(fm_node,
-+ p_LnxWrpFmPortDev->settings.param.portType,
-+ p_LnxWrpFmPortDev->settings.param.portId);
-+ if (!port_node) /* no advance parameters for FMan-Port */
-+ return E_OK;
-+
-+#if (DPAA_VERSION >= 11)
-+ uint32_prop = (uint32_t *)of_get_property(port_node, "vsp-window", &lenp);
-+ if (uint32_prop) {
-+ t_FmPortVSPAllocParams portVSPAllocParams;
-+ t_FmVspParams fmVspParams;
-+ t_LnxWrpFmDev *p_LnxWrpFmDev;
-+ uint8_t portId;
-+
-+ p_LnxWrpFmDev = ((t_LnxWrpFmDev *)p_LnxWrpFmPortDev->h_LnxWrpFmDev);
-+
-+ if (WARN_ON(lenp != sizeof(uint32_t)*2))
-+ RETURN_ERROR(MINOR, E_INVALID_VALUE, NO_MSG);
-+
-+ if ((p_LnxWrpFmPortDev->settings.param.portType == e_FM_PORT_TYPE_TX) ||
-+ (p_LnxWrpFmPortDev->settings.param.portType == e_FM_PORT_TYPE_TX_10G) ||
-+ ((p_LnxWrpFmPortDev->settings.param.portType == e_FM_PORT_TYPE_OH_OFFLINE_PARSING) &&
-+ p_LnxWrpFmPortDev->settings.frag_enabled))
-+ return E_OK;
-+
-+ memset(&portVSPAllocParams, 0, sizeof(portVSPAllocParams));
-+ memset(&fmVspParams, 0, sizeof(fmVspParams));
-+
-+ portVSPAllocParams.numOfProfiles = (uint8_t)be32_to_cpu(uint32_prop[0]);
-+ portVSPAllocParams.dfltRelativeId = (uint8_t)be32_to_cpu(uint32_prop[1]);
-+ fmVspParams.h_Fm = p_LnxWrpFmDev->h_Dev;
-+
-+ fmVspParams.portParams.portType = p_LnxWrpFmPortDev->settings.param.portType;
-+ fmVspParams.portParams.portId = p_LnxWrpFmPortDev->settings.param.portId;
-+ fmVspParams.relativeProfileId = portVSPAllocParams.dfltRelativeId;
-+
-+ if (p_LnxWrpFmPortDev->settings.param.portType != e_FM_PORT_TYPE_OH_OFFLINE_PARSING)
-+ {
-+ portId = fmVspParams.portParams.portId;
-+ if (p_LnxWrpFmPortDev->settings.param.portType == e_FM_PORT_TYPE_RX_10G){
-+#ifndef CONFIG_FMAN_ARM
-+ if (!(IS_T1023_T1024))
-+#endif
-+ portId += FM_MAX_NUM_OF_1G_RX_PORTS;
-+ }
-+ portVSPAllocParams.h_FmTxPort =
-+ p_LnxWrpFmDev->txPorts[portId].h_Dev;
-+ fmVspParams.liodnOffset =
-+ p_LnxWrpFmDev->rxPorts[portId].settings.param.specificParams.rxParams.liodnOffset;
-+ memcpy(&fmVspParams.extBufPools,
-+ &p_LnxWrpFmPortDev->settings.param.specificParams.rxParams.extBufPools,
-+ sizeof(t_FmExtPools));
-+ }
-+ else
-+ {
-+ memcpy(&fmVspParams.extBufPools,
-+ &p_LnxWrpFmPortDev->opExtPools,
-+ sizeof(t_FmExtPools));
-+ }
-+
-+ if ((err = FM_PORT_VSPAlloc(p_LnxWrpFmPortDev->h_Dev,
-+ &portVSPAllocParams)) != E_OK)
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+
-+ /* We're initializing only the default VSP that are being used by the Linux-Ethernet-driver */
-+ if ((p_LnxWrpFmPortDev->settings.param.portType == e_FM_PORT_TYPE_OH_OFFLINE_PARSING) &&
-+ !p_LnxWrpFmPortDev->opExtPools.numOfPoolsUsed)
-+ return E_OK;
-+
-+ p_LnxWrpFmPortDev->h_DfltVsp = FM_VSP_Config(&fmVspParams);
-+ if (!p_LnxWrpFmPortDev->h_DfltVsp)
-+ RETURN_ERROR(MAJOR, E_INVALID_HANDLE, ("default-VSP for port!"));
-+
-+ if ((err = FM_VSP_ConfigBufferPrefixContent(p_LnxWrpFmPortDev->h_DfltVsp,
-+ &p_LnxWrpFmPortDev->buffPrefixContent)) != E_OK)
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+
-+ if ((err = FM_VSP_Init(p_LnxWrpFmPortDev->h_DfltVsp)) != E_OK)
-+ RETURN_ERROR(MINOR, err, NO_MSG);
-+ }
-+#else
-+UNUSED(err); UNUSED(uint32_prop); UNUSED(lenp);
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+ of_node_put(port_node);
-+ of_node_put(fm_node);
-+
-+ return E_OK;
-+}
-+
-+static t_Error ConfigureFmPortDev(t_LnxWrpFmPortDev *p_LnxWrpFmPortDev)
-+{
-+ t_LnxWrpFmDev *p_LnxWrpFmDev =
-+ (t_LnxWrpFmDev *) p_LnxWrpFmPortDev->h_LnxWrpFmDev;
-+ struct resource *dev_res;
-+
-+ if (!p_LnxWrpFmPortDev->active)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE,
-+ ("FM port not configured!!!"));
-+
-+ dev_res =
-+ __devm_request_region(p_LnxWrpFmDev->dev, p_LnxWrpFmDev->res,
-+ p_LnxWrpFmPortDev->phys_baseAddr,
-+ p_LnxWrpFmPortDev->memSize,
-+ "fman-port-hc");
-+ if (unlikely(dev_res == NULL))
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE,
-+ ("__devm_request_region() failed"));
-+ p_LnxWrpFmPortDev->baseAddr =
-+ PTR_TO_UINT(devm_ioremap
-+ (p_LnxWrpFmDev->dev,
-+ p_LnxWrpFmPortDev->phys_baseAddr,
-+ p_LnxWrpFmPortDev->memSize));
-+ if (unlikely(p_LnxWrpFmPortDev->baseAddr == 0))
-+ REPORT_ERROR(MAJOR, E_INVALID_STATE,
-+ ("devm_ioremap() failed"));
-+
-+ p_LnxWrpFmPortDev->settings.param.baseAddr =
-+ p_LnxWrpFmPortDev->baseAddr;
-+
-+ return E_OK;
-+}
-+
-+static t_Error InitFmPortDev(t_LnxWrpFmPortDev *p_LnxWrpFmPortDev)
-+{
-+#define MY_ADV_CONFIG_CHECK_END \
-+ RETURN_ERROR(MAJOR, E_INVALID_SELECTION,\
-+ ("Advanced configuration routine"));\
-+ if (errCode != E_OK)\
-+ RETURN_ERROR(MAJOR, errCode, NO_MSG);\
-+ }
-+
-+ int i = 0;
-+
-+ if (!p_LnxWrpFmPortDev->active || p_LnxWrpFmPortDev->h_Dev)
-+ return E_INVALID_STATE;
-+
-+ p_LnxWrpFmPortDev->h_Dev =
-+ FM_PORT_Config(&p_LnxWrpFmPortDev->settings.param);
-+ if (p_LnxWrpFmPortDev->h_Dev == NULL)
-+ RETURN_ERROR(MAJOR, E_INVALID_HANDLE, ("FM-port"));
-+
-+#ifndef FM_QMI_NO_DEQ_OPTIONS_SUPPORT
-+ if ((p_LnxWrpFmPortDev->settings.param.portType ==
-+ e_FM_PORT_TYPE_TX_10G)
-+ || (p_LnxWrpFmPortDev->settings.param.portType ==
-+ e_FM_PORT_TYPE_TX)) {
-+ t_Error errCode = E_OK;
-+ errCode =
-+ FM_PORT_ConfigDeqHighPriority(p_LnxWrpFmPortDev->h_Dev,
-+ TRUE);
-+ if (errCode != E_OK)
-+ RETURN_ERROR(MAJOR, errCode, NO_MSG);
-+ errCode =
-+ FM_PORT_ConfigDeqPrefetchOption(p_LnxWrpFmPortDev->h_Dev,
-+ e_FM_PORT_DEQ_FULL_PREFETCH);
-+ if (errCode
-+ != E_OK)
-+ RETURN_ERROR(MAJOR, errCode, NO_MSG);
-+ }
-+#endif /* !FM_QMI_NO_DEQ_OPTIONS_SUPPORT */
-+
-+#ifndef CONFIG_FMAN_ARM
-+#ifdef FM_BCB_ERRATA_BMI_SW001
-+/* Configure BCB workaround on Rx ports, only for B4860 rev1 */
-+#define SVR_SECURITY_MASK 0x00080000
-+#define SVR_PERSONALITY_MASK 0x0000FF00
-+#define SVR_VER_IGNORE_MASK (SVR_SECURITY_MASK | SVR_PERSONALITY_MASK)
-+#define SVR_B4860_REV1_VALUE 0x86800010
-+
-+ if ((p_LnxWrpFmPortDev->settings.param.portType ==
-+ e_FM_PORT_TYPE_RX_10G) ||
-+ (p_LnxWrpFmPortDev->settings.param.portType ==
-+ e_FM_PORT_TYPE_RX)) {
-+ unsigned int svr;
-+
-+ svr = mfspr(SPRN_SVR);
-+
-+ if ((svr & ~SVR_VER_IGNORE_MASK) == SVR_B4860_REV1_VALUE)
-+ FM_PORT_ConfigBCBWorkaround(p_LnxWrpFmPortDev->h_Dev);
-+ }
-+#endif /* FM_BCB_ERRATA_BMI_SW001 */
-+#endif /* CONFIG_FMAN_ARM */
-+/* Call the driver's advanced configuration routines, if requested:
-+ Compare the function pointer of each entry to the available routines,
-+ and invoke the matching routine with proper casting of arguments. */
-+ while (p_LnxWrpFmPortDev->settings.advConfig[i].p_Function
-+ && (i < FM_MAX_NUM_OF_ADV_SETTINGS)) {
-+
-+/* TODO: Change this MACRO */
-+ ADV_CONFIG_CHECK_START(
-+ &(p_LnxWrpFmPortDev->settings.advConfig[i]))
-+
-+ ADV_CONFIG_CHECK(p_LnxWrpFmPortDev->h_Dev,
-+ FM_PORT_ConfigBufferPrefixContent,
-+ NCSW_PARAMS(1,
-+ (t_FmBufferPrefixContent *)))
-+
-+ if ((p_LnxWrpFmPortDev->settings.param.portType ==
-+ e_FM_PORT_TYPE_OH_OFFLINE_PARSING) &&
-+ (p_LnxWrpFmPortDev->settings.frag_enabled == TRUE)) {
-+
-+ ADV_CONFIG_CHECK(p_LnxWrpFmPortDev->h_Dev,
-+ FM_PORT_ConfigExtBufPools,
-+ NCSW_PARAMS(1, (t_FmExtPools *)))
-+
-+ /* this define contains an else */
-+ MY_ADV_CONFIG_CHECK_END
-+ }
-+
-+ /* Advance to next advanced configuration entry */
-+ i++;
-+ }
-+
-+
-+ if ((p_LnxWrpFmPortDev->settings.param.portType != e_FM_PORT_TYPE_TX) &&
-+ (p_LnxWrpFmPortDev->settings.param.portType != e_FM_PORT_TYPE_TX_10G)) {
-+ if (FM_PORT_ConfigErrorsToDiscard(p_LnxWrpFmPortDev->h_Dev, (FM_PORT_FRM_ERR_IPRE |
-+ FM_PORT_FRM_ERR_IPR_NCSP |
-+ FM_PORT_FRM_ERR_CLS_DISCARD)) !=E_OK)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, NO_MSG);
-+ }
-+
-+ if (CheckNConfigFmPortAdvArgs(p_LnxWrpFmPortDev) != E_OK)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, NO_MSG);
-+
-+ if (FM_PORT_Init(p_LnxWrpFmPortDev->h_Dev) != E_OK)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, NO_MSG);
-+
-+ if (CheckNSetFmPortAdvArgs(p_LnxWrpFmPortDev) != E_OK)
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, NO_MSG);
-+
-+/* FMan Fifo sizes behind the scene":
-+ * Using the following formulae (*), under a set of simplifying assumptions (.):
-+ * . all ports are configured in Normal Mode (rather than Independent Mode)
-+ * . the DPAA Eth driver allocates buffers of size:
-+ * . MAXFRM + NET_IP_ALIGN + DPA_PRIV_DATA_SIZE + DPA_PARSE_RESULTS_SIZE
-+ * + DPA_HASH_RESULTS_SIZE, i.e.:
-+ * MAXFRM + 2 + 16 + sizeof(t_FmPrsResult) + 16, i.e.:
-+ * MAXFRM + 66
-+ * . excessive buffer pools not accounted for
-+ *
-+ * * for Rx ports on P4080:
-+ * . IFSZ = ceil(max(FMBM_EBMPI[PBS]) / 256) * 256 + 7 * 256
-+ * . no internal frame offset (FMBM_RIM[FOF] == 0) - otherwise,
-+ * add up to 256 to the above
-+ *
-+ * * for Rx ports on P1023:
-+ * . IFSZ = ceil(second_largest(FMBM_EBMPI[PBS] / 256)) * 256 + 7 * 256,
-+ * if at least 2 bpools are configured
-+ * . IFSZ = 8 * 256, if only a single bpool is configured
-+ *
-+ * * for Tx ports:
-+ * . IFSZ = ceil(frame_size / 256) * 256 + 3 * 256
-+ * + FMBM_TFP[DPDE] * 256, i.e.:
-+ * IFSZ = ceil(MAXFRM / 256) * 256 + 3 x 256 + FMBM_TFP[DPDE] * 256
-+ *
-+ * * for OH ports on P4080:
-+ * . IFSZ = ceil(frame_size / 256) * 256 + 1 * 256 + FMBM_PP[MXT] * 256
-+ * * for OH ports on P1023:
-+ * . IFSZ = ceil(frame_size / 256) * 256 + 3 * 256 + FMBM_TFP[DPDE] * 256
-+ * * for both P4080 and P1023:
-+ * . (conservative decisions, assuming that BMI must bring the entire
-+ * frame, not only the frame header)
-+ * . no internal frame offset (FMBM_OIM[FOF] == 0) - otherwise,
-+ * add up to 256 to the above
-+ *
-+ * . for P4080/P5020/P3041/P2040, DPDE is:
-+ * > 0 or 1, for 1Gb ports, HW default: 0
-+ * > 2..7 (recommended: 3..7) for 10Gb ports, HW default: 3
-+ * . for P1023, DPDE should be 1
-+ *
-+ * . for P1023, MXT is in range (0..31)
-+ * . for P4080, MXT is in range (0..63)
-+ *
-+ */
-+#if 0
-+ if ((p_LnxWrpFmPortDev->defPcd != e_NO_PCD) &&
-+ (InitFmPort3TupleDefPcd(p_LnxWrpFmPortDev) != E_OK))
-+ RETURN_ERROR(MAJOR, E_INVALID_STATE, NO_MSG);
-+#endif
-+ return E_OK;
-+}
-+
-+void fm_set_rx_port_params(struct fm_port *port,
-+ struct fm_port_params *params)
-+{
-+ t_LnxWrpFmPortDev *p_LnxWrpFmPortDev = (t_LnxWrpFmPortDev *) port;
-+ int i;
-+
-+ p_LnxWrpFmPortDev->settings.param.specificParams.rxParams.errFqid =
-+ params->errq;
-+ p_LnxWrpFmPortDev->settings.param.specificParams.rxParams.dfltFqid =
-+ params->defq;
-+ p_LnxWrpFmPortDev->settings.param.specificParams.rxParams.extBufPools.
-+ numOfPoolsUsed = params->num_pools;
-+ for (i = 0; i < params->num_pools; i++) {
-+ p_LnxWrpFmPortDev->settings.param.specificParams.rxParams.
-+ extBufPools.extBufPool[i].id =
-+ params->pool_param[i].id;
-+ p_LnxWrpFmPortDev->settings.param.specificParams.rxParams.
-+ extBufPools.extBufPool[i].size =
-+ params->pool_param[i].size;
-+ }
-+
-+ p_LnxWrpFmPortDev->buffPrefixContent.privDataSize =
-+ params->priv_data_size;
-+ p_LnxWrpFmPortDev->buffPrefixContent.passPrsResult =
-+ params->parse_results;
-+ p_LnxWrpFmPortDev->buffPrefixContent.passHashResult =
-+ params->hash_results;
-+ p_LnxWrpFmPortDev->buffPrefixContent.passTimeStamp =
-+ params->time_stamp;
-+ p_LnxWrpFmPortDev->buffPrefixContent.dataAlign =
-+ params->data_align;
-+ p_LnxWrpFmPortDev->buffPrefixContent.manipExtraSpace =
-+ params->manip_extra_space;
-+
-+ ADD_ADV_CONFIG_START(p_LnxWrpFmPortDev->settings.advConfig,
-+ FM_MAX_NUM_OF_ADV_SETTINGS)
-+
-+ ADD_ADV_CONFIG_NO_RET(FM_PORT_ConfigBufferPrefixContent,
-+ ARGS(1,
-+ (&p_LnxWrpFmPortDev->
-+ buffPrefixContent)));
-+
-+ ADD_ADV_CONFIG_END InitFmPortDev(p_LnxWrpFmPortDev);
-+}
-+EXPORT_SYMBOL(fm_set_rx_port_params);
-+
-+/* this function is called from oh_probe as well, thus it contains oh port
-+ * specific parameters (make sure everything is checked) */
-+void fm_set_tx_port_params(struct fm_port *port,
-+ struct fm_port_params *params)
-+{
-+ t_LnxWrpFmPortDev *p_LnxWrpFmPortDev = (t_LnxWrpFmPortDev *) port;
-+
-+ p_LnxWrpFmPortDev->settings.param.specificParams.nonRxParams.errFqid =
-+ params->errq;
-+ p_LnxWrpFmPortDev->settings.param.specificParams.nonRxParams.
-+ dfltFqid = params->defq;
-+
-+ p_LnxWrpFmPortDev->buffPrefixContent.privDataSize =
-+ params->priv_data_size;
-+ p_LnxWrpFmPortDev->buffPrefixContent.passPrsResult =
-+ params->parse_results;
-+ p_LnxWrpFmPortDev->buffPrefixContent.passHashResult =
-+ params->hash_results;
-+ p_LnxWrpFmPortDev->buffPrefixContent.passTimeStamp =
-+ params->time_stamp;
-+ p_LnxWrpFmPortDev->settings.frag_enabled =
-+ params->frag_enable;
-+ p_LnxWrpFmPortDev->buffPrefixContent.dataAlign =
-+ params->data_align;
-+ p_LnxWrpFmPortDev->buffPrefixContent.manipExtraSpace =
-+ params->manip_extra_space;
-+
-+ ADD_ADV_CONFIG_START(p_LnxWrpFmPortDev->settings.advConfig,
-+ FM_MAX_NUM_OF_ADV_SETTINGS)
-+
-+ ADD_ADV_CONFIG_NO_RET(FM_PORT_ConfigBufferPrefixContent,
-+ ARGS(1,
-+ (&p_LnxWrpFmPortDev->
-+ buffPrefixContent)));
-+
-+ /* oh port specific parameter (for fragmentation only) */
-+ if ((p_LnxWrpFmPortDev->settings.param.portType ==
-+ e_FM_PORT_TYPE_OH_OFFLINE_PARSING) &&
-+ params->num_pools) {
-+ int i;
-+
-+ p_LnxWrpFmPortDev->opExtPools.numOfPoolsUsed = params->num_pools;
-+ for (i = 0; i < params->num_pools; i++) {
-+ p_LnxWrpFmPortDev->opExtPools.extBufPool[i].id = params->pool_param[i].id;
-+ p_LnxWrpFmPortDev->opExtPools.extBufPool[i].size = params->pool_param[i].size;
-+ }
-+
-+ if (p_LnxWrpFmPortDev->settings.frag_enabled)
-+ ADD_ADV_CONFIG_NO_RET(FM_PORT_ConfigExtBufPools,
-+ ARGS(1, (&p_LnxWrpFmPortDev->opExtPools)));
-+ }
-+
-+ ADD_ADV_CONFIG_END InitFmPortDev(p_LnxWrpFmPortDev);
-+}
-+EXPORT_SYMBOL(fm_set_tx_port_params);
-+
-+void fm_mac_set_handle(t_Handle h_lnx_wrp_fm_dev,
-+ t_Handle h_fm_mac,
-+ int mac_id)
-+{
-+ t_LnxWrpFmDev *p_lnx_wrp_fm_dev = (t_LnxWrpFmDev *)h_lnx_wrp_fm_dev;
-+
-+ p_lnx_wrp_fm_dev->macs[mac_id].h_Dev = h_fm_mac;
-+ p_lnx_wrp_fm_dev->macs[mac_id].h_LnxWrpFmDev = h_lnx_wrp_fm_dev;
-+}
-+EXPORT_SYMBOL(fm_mac_set_handle);
-+
-+static void LnxwrpFmPcdDevExceptionsCb(t_Handle h_App,
-+ e_FmPcdExceptions exception)
-+{
-+ t_LnxWrpFmDev *p_LnxWrpFmDev = (t_LnxWrpFmDev *) h_App;
-+
-+ ASSERT_COND(p_LnxWrpFmDev);
-+
-+ DBG(INFO, ("got fm-pcd exception %d", exception));
-+
-+ /* do nothing */
-+ UNUSED(exception);
-+}
-+
-+static void LnxwrpFmPcdDevIndexedExceptionsCb(t_Handle h_App,
-+ e_FmPcdExceptions exception,
-+ uint16_t index)
-+{
-+ t_LnxWrpFmDev *p_LnxWrpFmDev = (t_LnxWrpFmDev *) h_App;
-+
-+ ASSERT_COND(p_LnxWrpFmDev);
-+
-+ DBG(INFO,
-+ ("got fm-pcd-indexed exception %d, indx %d", exception, index));
-+
-+ /* do nothing */
-+ UNUSED(exception);
-+ UNUSED(index);
-+}
-+
-+static t_Error InitFmPcdDev(t_LnxWrpFmDev *p_LnxWrpFmDev)
-+{
-+ spin_lock_init(&lock);
-+
-+ if (p_LnxWrpFmDev->pcdActive) {
-+ t_LnxWrpFmPortDev *p_LnxWrpFmPortDev = &p_LnxWrpFmDev->hcPort;
-+ t_FmPcdParams fmPcdParams;
-+ t_Error err;
-+
-+ memset(&fmPcdParams, 0, sizeof(fmPcdParams));
-+ fmPcdParams.h_Fm = p_LnxWrpFmDev->h_Dev;
-+ fmPcdParams.prsSupport = p_LnxWrpFmDev->prsActive;
-+ fmPcdParams.kgSupport = p_LnxWrpFmDev->kgActive;
-+ fmPcdParams.plcrSupport = p_LnxWrpFmDev->plcrActive;
-+ fmPcdParams.ccSupport = p_LnxWrpFmDev->ccActive;
-+ fmPcdParams.numOfSchemes = FM_PCD_KG_NUM_OF_SCHEMES;
-+
-+#ifndef CONFIG_GUEST_PARTITION
-+ fmPcdParams.f_Exception = LnxwrpFmPcdDevExceptionsCb;
-+ if (fmPcdParams.kgSupport)
-+ fmPcdParams.f_ExceptionId =
-+ LnxwrpFmPcdDevIndexedExceptionsCb;
-+ fmPcdParams.h_App = p_LnxWrpFmDev;
-+#endif /* !CONFIG_GUEST_PARTITION */
-+
-+#ifdef CONFIG_MULTI_PARTITION_SUPPORT
-+ fmPcdParams.numOfSchemes = 0;
-+ fmPcdParams.numOfClsPlanEntries = 0;
-+ fmPcdParams.partitionId = 0;
-+#endif /* CONFIG_MULTI_PARTITION_SUPPORT */
-+ fmPcdParams.useHostCommand = TRUE;
-+
-+ p_LnxWrpFmDev->hc_tx_fq =
-+ FqAlloc(p_LnxWrpFmDev,
-+ 0,
-+ QMAN_FQ_FLAG_TO_DCPORTAL,
-+ p_LnxWrpFmPortDev->txCh, 0);
-+ if (!p_LnxWrpFmDev->hc_tx_fq)
-+ RETURN_ERROR(MAJOR, E_NULL_POINTER,
-+ ("Frame queue allocation failed..."));
-+
-+ p_LnxWrpFmDev->hc_tx_conf_fq =
-+ FqAlloc(p_LnxWrpFmDev,
-+ 0,
-+ QMAN_FQ_FLAG_NO_ENQUEUE,
-+ p_LnxWrpFmDev->hcCh, 1);
-+ if (!p_LnxWrpFmDev->hc_tx_conf_fq)
-+ RETURN_ERROR(MAJOR, E_NULL_POINTER,
-+ ("Frame queue allocation failed..."));
-+
-+ p_LnxWrpFmDev->hc_tx_err_fq =
-+ FqAlloc(p_LnxWrpFmDev,
-+ 0,
-+ QMAN_FQ_FLAG_NO_ENQUEUE,
-+ p_LnxWrpFmDev->hcCh, 2);
-+ if (!p_LnxWrpFmDev->hc_tx_err_fq)
-+ RETURN_ERROR(MAJOR, E_NULL_POINTER,
-+ ("Frame queue allocation failed..."));
-+
-+ fmPcdParams.hc.portBaseAddr = p_LnxWrpFmPortDev->baseAddr;
-+ fmPcdParams.hc.portId =
-+ p_LnxWrpFmPortDev->settings.param.portId;
-+ fmPcdParams.hc.liodnBase =
-+ p_LnxWrpFmPortDev->settings.param.liodnBase;
-+ fmPcdParams.hc.errFqid =
-+ qman_fq_fqid(p_LnxWrpFmDev->hc_tx_err_fq);
-+ fmPcdParams.hc.confFqid =
-+ qman_fq_fqid(p_LnxWrpFmDev->hc_tx_conf_fq);
-+ fmPcdParams.hc.qmChannel = p_LnxWrpFmPortDev->txCh;
-+ fmPcdParams.hc.f_QmEnqueue = QmEnqueueCB;
-+ fmPcdParams.hc.h_QmArg = (t_Handle) p_LnxWrpFmDev;
-+
-+ p_LnxWrpFmDev->h_PcdDev = FM_PCD_Config(&fmPcdParams);
-+ if (!p_LnxWrpFmDev->h_PcdDev)
-+ RETURN_ERROR(MAJOR, E_INVALID_HANDLE, ("FM PCD!"));
-+
-+ err =
-+ FM_PCD_ConfigPlcrNumOfSharedProfiles(p_LnxWrpFmDev->h_PcdDev,
-+ LNXWRP_FM_NUM_OF_SHARED_PROFILES);
-+ if (err != E_OK)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+
-+ err = FM_PCD_Init(p_LnxWrpFmDev->h_PcdDev);
-+ if (err != E_OK)
-+ RETURN_ERROR(MAJOR, err, NO_MSG);
-+
-+ if (p_LnxWrpFmDev->err_irq == 0) {
-+ FM_PCD_SetException(p_LnxWrpFmDev->h_PcdDev,
-+ e_FM_PCD_KG_EXCEPTION_DOUBLE_ECC,
-+ FALSE);
-+ FM_PCD_SetException(p_LnxWrpFmDev->h_PcdDev,
-+ e_FM_PCD_KG_EXCEPTION_KEYSIZE_OVERFLOW,
-+ FALSE);
-+ FM_PCD_SetException(p_LnxWrpFmDev->h_PcdDev,
-+ e_FM_PCD_PLCR_EXCEPTION_INIT_ENTRY_ERROR,
-+ FALSE);
-+ FM_PCD_SetException(p_LnxWrpFmDev->h_PcdDev,
-+ e_FM_PCD_PLCR_EXCEPTION_DOUBLE_ECC,
-+ FALSE);
-+ FM_PCD_SetException(p_LnxWrpFmDev->h_PcdDev,
-+ e_FM_PCD_PRS_EXCEPTION_DOUBLE_ECC,
-+ FALSE);
-+ FM_PCD_SetException(p_LnxWrpFmDev->h_PcdDev,
-+ e_FM_PCD_PLCR_EXCEPTION_PRAM_SELF_INIT_COMPLETE,
-+ FALSE);
-+ FM_PCD_SetException(p_LnxWrpFmDev->h_PcdDev,
-+ e_FM_PCD_PLCR_EXCEPTION_ATOMIC_ACTION_COMPLETE,
-+ FALSE);
-+ FM_PCD_SetException(p_LnxWrpFmDev->h_PcdDev,
-+ e_FM_PCD_PRS_EXCEPTION_SINGLE_ECC,
-+ FALSE);
-+ }
-+ }
-+
-+ return E_OK;
-+}
-+
-+void FreeFmPcdDev(t_LnxWrpFmDev *p_LnxWrpFmDev)
-+{
-+
-+ if (p_LnxWrpFmDev->h_PcdDev)
-+ FM_PCD_Free(p_LnxWrpFmDev->h_PcdDev);
-+
-+ if (p_LnxWrpFmDev->hc_tx_err_fq)
-+ FqFree(p_LnxWrpFmDev->hc_tx_err_fq);
-+
-+ if (p_LnxWrpFmDev->hc_tx_conf_fq)
-+ FqFree(p_LnxWrpFmDev->hc_tx_conf_fq);
-+
-+ if (p_LnxWrpFmDev->hc_tx_fq)
-+ FqFree(p_LnxWrpFmDev->hc_tx_fq);
-+}
-+
-+static void FreeFmPortDev(t_LnxWrpFmPortDev *p_LnxWrpFmPortDev)
-+{
-+ t_LnxWrpFmDev *p_LnxWrpFmDev =
-+ (t_LnxWrpFmDev *) p_LnxWrpFmPortDev->h_LnxWrpFmDev;
-+
-+ if (!p_LnxWrpFmPortDev->active)
-+ return;
-+
-+ if (p_LnxWrpFmPortDev->h_Dev)
-+ FM_PORT_Free(p_LnxWrpFmPortDev->h_Dev);
-+
-+ devm_iounmap(p_LnxWrpFmDev->dev,
-+ UINT_TO_PTR(p_LnxWrpFmPortDev->baseAddr));
-+ __devm_release_region(p_LnxWrpFmDev->dev, p_LnxWrpFmDev->res,
-+ p_LnxWrpFmPortDev->phys_baseAddr,
-+ p_LnxWrpFmPortDev->memSize);
-+}
-+
-+static int /*__devinit*/ fm_port_probe(struct platform_device *of_dev)
-+{
-+ t_LnxWrpFmPortDev *p_LnxWrpFmPortDev;
-+ t_LnxWrpFmDev *p_LnxWrpFmDev;
-+ struct device *dev;
-+
-+ dev = &of_dev->dev;
-+
-+ p_LnxWrpFmPortDev = ReadFmPortDevTreeNode(of_dev);
-+ if (p_LnxWrpFmPortDev == NULL)
-+ return -EIO;
-+ /* Port can be inactive, thus will not be probed:
-+ - in performance mode, OH ports are disabled
-+ ...
-+ */
-+ if (!p_LnxWrpFmPortDev->active)
-+ return 0;
-+
-+ if (ConfigureFmPortDev(p_LnxWrpFmPortDev) != E_OK)
-+ return -EIO;
-+
-+ dev_set_drvdata(dev, p_LnxWrpFmPortDev);
-+
-+ if (p_LnxWrpFmPortDev->settings.param.portType ==
-+ e_FM_PORT_TYPE_OH_HOST_COMMAND)
-+ InitFmPcdDev((t_LnxWrpFmDev *) p_LnxWrpFmPortDev->h_LnxWrpFmDev);
-+
-+ p_LnxWrpFmDev = (t_LnxWrpFmDev *) p_LnxWrpFmPortDev->h_LnxWrpFmDev;
-+
-+ if (p_LnxWrpFmPortDev->settings.param.portType == e_FM_PORT_TYPE_RX) {
-+ Sprint(p_LnxWrpFmPortDev->name, "%s-port-rx%d",
-+ p_LnxWrpFmDev->name, p_LnxWrpFmPortDev->id);
-+ p_LnxWrpFmPortDev->minor =
-+ p_LnxWrpFmPortDev->id + DEV_FM_RX_PORTS_MINOR_BASE;
-+ } else if (p_LnxWrpFmPortDev->settings.param.portType ==
-+ e_FM_PORT_TYPE_RX_10G) {
-+ Sprint(p_LnxWrpFmPortDev->name, "%s-port-rx%d",
-+ p_LnxWrpFmDev->name,
-+ p_LnxWrpFmPortDev->id + FM_MAX_NUM_OF_1G_RX_PORTS);
-+ p_LnxWrpFmPortDev->minor =
-+ p_LnxWrpFmPortDev->id + FM_MAX_NUM_OF_1G_RX_PORTS +
-+ DEV_FM_RX_PORTS_MINOR_BASE;
-+#ifndef CONFIG_FMAN_ARM
-+ if (IS_T1023_T1024) {
-+ Sprint(p_LnxWrpFmPortDev->name, "%s-port-rx%d",
-+ p_LnxWrpFmDev->name,
-+ p_LnxWrpFmPortDev->id);
-+ p_LnxWrpFmPortDev->minor =
-+ p_LnxWrpFmPortDev->id +
-+ DEV_FM_RX_PORTS_MINOR_BASE;
-+ }
-+#endif
-+ } else if (p_LnxWrpFmPortDev->settings.param.portType ==
-+ e_FM_PORT_TYPE_TX) {
-+ Sprint(p_LnxWrpFmPortDev->name, "%s-port-tx%d",
-+ p_LnxWrpFmDev->name, p_LnxWrpFmPortDev->id);
-+ p_LnxWrpFmPortDev->minor =
-+ p_LnxWrpFmPortDev->id + DEV_FM_TX_PORTS_MINOR_BASE;
-+ } else if (p_LnxWrpFmPortDev->settings.param.portType ==
-+ e_FM_PORT_TYPE_TX_10G) {
-+ Sprint(p_LnxWrpFmPortDev->name, "%s-port-tx%d",
-+ p_LnxWrpFmDev->name,
-+ p_LnxWrpFmPortDev->id + FM_MAX_NUM_OF_1G_TX_PORTS);
-+ p_LnxWrpFmPortDev->minor =
-+ p_LnxWrpFmPortDev->id + FM_MAX_NUM_OF_1G_TX_PORTS +
-+ DEV_FM_TX_PORTS_MINOR_BASE;
-+#ifndef CONFIG_FMAN_ARM
-+ if (IS_T1023_T1024) {
-+ Sprint(p_LnxWrpFmPortDev->name, "%s-port-tx%d",
-+ p_LnxWrpFmDev->name,
-+ p_LnxWrpFmPortDev->id);
-+ p_LnxWrpFmPortDev->minor =
-+ p_LnxWrpFmPortDev->id +
-+ DEV_FM_TX_PORTS_MINOR_BASE;
-+ }
-+#endif
-+ } else if (p_LnxWrpFmPortDev->settings.param.portType ==
-+ e_FM_PORT_TYPE_OH_HOST_COMMAND) {
-+ Sprint(p_LnxWrpFmPortDev->name, "%s-port-oh%d",
-+ p_LnxWrpFmDev->name, p_LnxWrpFmPortDev->id);
-+ p_LnxWrpFmPortDev->minor =
-+ p_LnxWrpFmPortDev->id + DEV_FM_OH_PORTS_MINOR_BASE;
-+ } else if (p_LnxWrpFmPortDev->settings.param.portType ==
-+ e_FM_PORT_TYPE_OH_OFFLINE_PARSING) {
-+ Sprint(p_LnxWrpFmPortDev->name, "%s-port-oh%d",
-+ p_LnxWrpFmDev->name, p_LnxWrpFmPortDev->id + 1);
-+ p_LnxWrpFmPortDev->minor =
-+ p_LnxWrpFmPortDev->id + 1 +
-+ DEV_FM_OH_PORTS_MINOR_BASE;
-+ }
-+
-+ device_create(p_LnxWrpFmDev->fm_class, NULL,
-+ MKDEV(p_LnxWrpFmDev->major, p_LnxWrpFmPortDev->minor),
-+ NULL, p_LnxWrpFmPortDev->name);
-+
-+ /* create sysfs entries for stats and regs */
-+
-+ if (fm_port_sysfs_create(dev) != 0) {
-+ FreeFmPortDev(p_LnxWrpFmPortDev);
-+ REPORT_ERROR(MAJOR, E_INVALID_STATE,
-+ ("Unable to create sys entry - fm port!!!"));
-+ return -EIO;
-+ }
-+
-+#ifdef FM_TX_INVALID_ECC_ERRATA_10GMAC_A009
-+ FM_DisableRamsEcc(p_LnxWrpFmDev->h_Dev);
-+#endif /* FM_TX_INVALID_ECC_ERRATA_10GMAC_A009 */
-+
-+ DBG(TRACE, ("%s probed", p_LnxWrpFmPortDev->name));
-+
-+ return 0;
-+}
-+
-+static int fm_port_remove(struct platform_device *of_dev)
-+{
-+ t_LnxWrpFmPortDev *p_LnxWrpFmPortDev;
-+ t_LnxWrpFmDev *p_LnxWrpFmDev;
-+ struct device *dev;
-+
-+ dev = &of_dev->dev;
-+ p_LnxWrpFmPortDev = dev_get_drvdata(dev);
-+
-+ fm_port_sysfs_destroy(dev);
-+
-+ p_LnxWrpFmDev = (t_LnxWrpFmDev *) p_LnxWrpFmPortDev->h_LnxWrpFmDev;
-+ device_destroy(p_LnxWrpFmDev->fm_class,
-+ MKDEV(p_LnxWrpFmDev->major, p_LnxWrpFmPortDev->minor));
-+
-+ FreeFmPortDev(p_LnxWrpFmPortDev);
-+
-+ dev_set_drvdata(dev, NULL);
-+
-+ return 0;
-+}
-+
-+static const struct of_device_id fm_port_match[] = {
-+ {
-+ .compatible = "fsl,fman-port-oh"},
-+ {
-+ .compatible = "fsl,fman-v2-port-oh"},
-+ {
-+ .compatible = "fsl,fman-v3-port-oh"},
-+ {
-+ .compatible = "fsl,fman-port-1g-rx"},
-+ {
-+ .compatible = "fsl,fman-port-10g-rx"},
-+ {
-+ .compatible = "fsl,fman-port-1g-tx"},
-+ {
-+ .compatible = "fsl,fman-port-10g-tx"},
-+ {}
-+};
-+
-+#ifndef MODULE
-+MODULE_DEVICE_TABLE(of, fm_port_match);
-+#endif /* !MODULE */
-+
-+static struct platform_driver fm_port_driver = {
-+
-+ .driver = {
-+ .name = "fsl-fman-port",
-+ .of_match_table = fm_port_match,
-+ .owner = THIS_MODULE,
-+ },
-+ .probe = fm_port_probe,
-+ .remove = fm_port_remove
-+};
-+
-+
-+t_Error LNXWRP_FM_Port_Init(void)
-+{
-+ /* Register to the DTB for basic FM port API */
-+ if (platform_driver_register(&fm_port_driver))
-+ return E_NO_DEVICE;
-+
-+ return E_OK;
-+}
-+
-+void LNXWRP_FM_Port_Free(void)
-+{
-+ platform_driver_unregister(&fm_port_driver);
-+}
-+
-+static int __init __cold fm_port_load(void)
-+{
-+ if (LNXWRP_FM_Port_Init() != E_OK) {
-+ printk(KERN_CRIT "Failed to init FM Ports wrapper!\n");
-+ return -ENODEV;
-+ }
-+
-+ printk(KERN_CRIT "Freescale FM Ports module\n");
-+
-+ return 0;
-+}
-+
-+static void __exit __cold fm_port_unload(void)
-+{
-+ LNXWRP_FM_Port_Free();
-+}
-+
-+module_init(fm_port_load);
-+module_exit(fm_port_unload);
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/src/wrapper/lnxwrp_ioctls_fm.c
-@@ -0,0 +1,4854 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+/*
-+ @File lnxwrp_ioctls_fm.c
-+ @Author Shlomi Gridish
-+ @Description FM Linux wrapper functions.
-+*/
-+
-+/* Linux Headers ------------------- */
-+#include <linux/version.h>
-+
-+#if defined(CONFIG_MODVERSIONS) && !defined(MODVERSIONS)
-+#define MODVERSIONS
-+#endif
-+#ifdef MODVERSIONS
-+#include <config/modversions.h>
-+#endif /* MODVERSIONS */
-+
-+#include <linux/kernel.h>
-+#include <linux/module.h>
-+#include <linux/slab.h>
-+#include <linux/fs.h>
-+#include <linux/cdev.h>
-+#include <linux/device.h>
-+#include <linux/irq.h>
-+#include <linux/interrupt.h>
-+#include <linux/io.h>
-+#include <linux/ioport.h>
-+#include <linux/of_platform.h>
-+#include <linux/uaccess.h>
-+#include <asm/errno.h>
-+#ifndef CONFIG_FMAN_ARM
-+#include <sysdev/fsl_soc.h>
-+#include <linux/fsl/svr.h>
-+#endif
-+
-+#if defined(CONFIG_COMPAT)
-+#include <linux/compat.h>
-+#endif
-+
-+#include "part_ext.h"
-+#include "fm_ioctls.h"
-+#include "fm_pcd_ioctls.h"
-+#include "fm_port_ioctls.h"
-+#include "fm_vsp_ext.h"
-+
-+#ifndef CONFIG_FMAN_ARM
-+#define IS_T1023_T1024 (SVR_SOC_VER(mfspr(SPRN_SVR)) == SVR_T1024 || \
-+ SVR_SOC_VER(mfspr(SPRN_SVR)) == SVR_T1023)
-+#endif
-+
-+#define __ERR_MODULE__ MODULE_FM
-+
-+#if defined(CONFIG_COMPAT)
-+#include "lnxwrp_ioctls_fm_compat.h"
-+#endif
-+
-+#include "lnxwrp_fm.h"
-+
-+#define CMP_IOC_DEFINE(def) (IOC_##def != def)
-+
-+/* fm_pcd_ioctls.h === fm_pcd_ext.h assertions */
-+#if CMP_IOC_DEFINE(FM_PCD_MAX_NUM_OF_PRIVATE_HDRS)
-+#error Error: please synchronize IOC_ defines!
-+#endif
-+
-+#if CMP_IOC_DEFINE(FM_PCD_PRS_NUM_OF_HDRS)
-+#error Error: please synchronize IOC_ defines!
-+#endif
-+
-+#if CMP_IOC_DEFINE(FM_PCD_MAX_NUM_OF_DISTINCTION_UNITS)
-+#error Error: please synchronize IOC_ defines!
-+#endif
-+
-+#if CMP_IOC_DEFINE(FM_PCD_MAX_NUM_OF_INTERCHANGEABLE_HDRS)
-+#error Error: please synchronize IOC_ defines!
-+#endif
-+
-+#if CMP_IOC_DEFINE(FM_PCD_KG_NUM_OF_GENERIC_REGS)
-+#error Error: please synchronize IOC_ defines!
-+#endif
-+
-+#if CMP_IOC_DEFINE(FM_PCD_KG_MAX_NUM_OF_EXTRACTS_PER_KEY)
-+#error Error: please synchronize IOC_ defines!
-+#endif
-+
-+#if CMP_IOC_DEFINE(FM_PCD_KG_NUM_OF_EXTRACT_MASKS)
-+#error Error: please synchronize IOC_ defines!
-+#endif
-+
-+#if CMP_IOC_DEFINE(FM_PCD_KG_NUM_OF_DEFAULT_GROUPS)
-+#error Error: please synchronize IOC_ defines!
-+#endif
-+
-+#if CMP_IOC_DEFINE(FM_PCD_PRS_NUM_OF_LABELS)
-+#error Error: please synchronize IOC_ defines!
-+#endif
-+
-+#if CMP_IOC_DEFINE(FM_PCD_SW_PRS_SIZE)
-+#error Error: please synchronize IOC_ defines!
-+#endif
-+
-+#if CMP_IOC_DEFINE(FM_PCD_MAX_MANIP_INSRT_TEMPLATE_SIZE)
-+#error Error: please synchronize IOC_ defines!
-+#endif
-+
-+#if DPAA_VERSION >= 11
-+#if CMP_IOC_DEFINE(FM_PCD_FRM_REPLIC_MAX_NUM_OF_ENTRIES)
-+#error Error: please synchronize IOC_ defines!
-+#endif
-+#endif
-+
-+#if CMP_IOC_DEFINE(FM_PCD_MAX_NUM_OF_CC_TREES)
-+#error Error: please synchronize IOC_ defines!
-+#endif
-+
-+#if CMP_IOC_DEFINE(FM_PCD_MAX_NUM_OF_CC_GROUPS)
-+#error Error: please synchronize IOC_ defines!
-+#endif
-+
-+#if CMP_IOC_DEFINE(FM_PCD_MAX_NUM_OF_CC_UNITS)
-+#error Error: please synchronize IOC_ defines!
-+#endif
-+
-+#if CMP_IOC_DEFINE(FM_PCD_MAX_NUM_OF_KEYS)
-+#error Error: please synchronize IOC_ defines!
-+#endif
-+
-+#if CMP_IOC_DEFINE(FM_PCD_MAX_SIZE_OF_KEY)
-+#error Error: please synchronize IOC_ defines!
-+#endif
-+
-+#if CMP_IOC_DEFINE(FM_PCD_MAX_NUM_OF_CC_ENTRIES_IN_GRP)
-+#error Error: please synchronize IOC_ defines!
-+#endif
-+
-+#if CMP_IOC_DEFINE(FM_PCD_LAST_KEY_INDEX)
-+#error Error: please synchronize IOC_ defines!
-+#endif
-+
-+/* net_ioctls.h === net_ext.h assertions */
-+#if CMP_IOC_DEFINE(NET_HEADER_FIELD_PPP_PID)
-+#error Error: please synchronize IOC_ defines!
-+#endif
-+
-+#if CMP_IOC_DEFINE(NET_HEADER_FIELD_PPP_COMPRESSED)
-+#error Error: please synchronize IOC_ defines!
-+#endif
-+
-+#if CMP_IOC_DEFINE(NET_HEADER_FIELD_PPP_ALL_FIELDS)
-+#error Error: please synchronize IOC_ defines!
-+#endif
-+
-+#if CMP_IOC_DEFINE(NET_HEADER_FIELD_PPPoE_ALL_FIELDS)
-+#error Error: please synchronize IOC_ defines!
-+#endif
-+
-+#if CMP_IOC_DEFINE(NET_HEADER_FIELD_PPPMUX_ALL_FIELDS)
-+#error Error: please synchronize IOC_ defines!
-+#endif
-+
-+#if CMP_IOC_DEFINE(NET_HEADER_FIELD_PPPMUX_SUBFRAME_ALL_FIELDS)
-+#error Error: please synchronize IOC_ defines!
-+#endif
-+
-+#if CMP_IOC_DEFINE(NET_HEADER_FIELD_ETH_ALL_FIELDS)
-+#error Error: please synchronize IOC_ defines!
-+#endif
-+
-+#if CMP_IOC_DEFINE(NET_HEADER_FIELD_IPv4_ALL_FIELDS)
-+#error Error: please synchronize IOC_ defines!
-+#endif
-+
-+#if CMP_IOC_DEFINE(NET_HEADER_FIELD_IPv6_ALL_FIELDS)
-+#error Error: please synchronize IOC_ defines!
-+#endif
-+
-+#if CMP_IOC_DEFINE(NET_HEADER_FIELD_ICMP_ALL_FIELDS)
-+#error Error: please synchronize IOC_ defines!
-+#endif
-+
-+#if CMP_IOC_DEFINE(NET_HEADER_FIELD_IGMP_ALL_FIELDS)
-+#error Error: please synchronize IOC_ defines!
-+#endif
-+
-+#if CMP_IOC_DEFINE(NET_HEADER_FIELD_TCP_ALL_FIELDS)
-+#error Error: please synchronize IOC_ defines!
-+#endif
-+
-+#if CMP_IOC_DEFINE(NET_HEADER_FIELD_SCTP_ALL_FIELDS)
-+#error Error: please synchronize IOC_ defines!
-+#endif
-+
-+#if CMP_IOC_DEFINE(NET_HEADER_FIELD_DCCP_ALL_FIELDS)
-+#error Error: please synchronize IOC_ defines!
-+#endif
-+
-+#if CMP_IOC_DEFINE(NET_HEADER_FIELD_UDP_ALL_FIELDS)
-+#error Error: please synchronize IOC_ defines!
-+#endif
-+
-+#if CMP_IOC_DEFINE(NET_HEADER_FIELD_UDP_ENCAP_ESP_ALL_FIELDS)
-+#error Error: please synchronize IOC_ defines!
-+#endif
-+
-+#if CMP_IOC_DEFINE(NET_HEADER_FIELD_IPHC_ALL_FIELDS)
-+#error Error: please synchronize IOC_ defines!
-+#endif
-+
-+#if CMP_IOC_DEFINE(NET_HEADER_FIELD_SCTP_CHUNK_DATA_ALL_FIELDS)
-+#error Error: please synchronize IOC_ defines!
-+#endif
-+
-+#if CMP_IOC_DEFINE(NET_HEADER_FIELD_L2TPv2_ALL_FIELDS)
-+#error Error: please synchronize IOC_ defines!
-+#endif
-+
-+#if CMP_IOC_DEFINE(NET_HEADER_FIELD_L2TPv3_CTRL_ALL_FIELDS)
-+#error Error: please synchronize IOC_ defines!
-+#endif
-+
-+#if CMP_IOC_DEFINE(NET_HEADER_FIELD_L2TPv3_SESS_ALL_FIELDS)
-+#error Error: please synchronize IOC_ defines!
-+#endif
-+
-+#if CMP_IOC_DEFINE(NET_HEADER_FIELD_VLAN_ALL_FIELDS)
-+#error Error: please synchronize IOC_ defines!
-+#endif
-+
-+#if CMP_IOC_DEFINE(NET_HEADER_FIELD_LLC_ALL_FIELDS)
-+#error Error: please synchronize IOC_ defines!
-+#endif
-+
-+#if CMP_IOC_DEFINE(NET_HEADER_FIELD_NLPID_ALL_FIELDS)
-+#error Error: please synchronize IOC_ defines!
-+#endif
-+
-+#if CMP_IOC_DEFINE(NET_HEADER_FIELD_SNAP_ALL_FIELDS)
-+#error Error: please synchronize IOC_ defines!
-+#endif
-+
-+#if CMP_IOC_DEFINE(NET_HEADER_FIELD_LLC_SNAP_ALL_FIELDS)
-+#warning Error: please synchronize IOC_ defines!
-+#endif
-+
-+#if CMP_IOC_DEFINE(NET_HEADER_FIELD_ARP_ALL_FIELDS)
-+#error Error: please synchronize IOC_ defines!
-+#endif
-+
-+#if CMP_IOC_DEFINE(NET_HEADER_FIELD_RFC2684_ALL_FIELDS)
-+#error Error: please synchronize IOC_ defines!
-+#endif
-+
-+#if CMP_IOC_DEFINE(NET_HEADER_FIELD_USER_DEFINED_ALL_FIELDS)
-+#error Error: please synchronize IOC_ defines!
-+#endif
-+
-+#if CMP_IOC_DEFINE(NET_HEADER_FIELD_PAYLOAD_ALL_FIELDS)
-+#error Error: please synchronize IOC_ defines!
-+#endif
-+
-+#if CMP_IOC_DEFINE(NET_HEADER_FIELD_GRE_ALL_FIELDS)
-+#error Error: please synchronize IOC_ defines!
-+#endif
-+
-+#if CMP_IOC_DEFINE(NET_HEADER_FIELD_MINENCAP_ALL_FIELDS)
-+#error Error: please synchronize IOC_ defines!
-+#endif
-+
-+#if CMP_IOC_DEFINE(NET_HEADER_FIELD_IPSEC_AH_ALL_FIELDS)
-+#error Error: please synchronize IOC_ defines!
-+#endif
-+
-+#if CMP_IOC_DEFINE(NET_HEADER_FIELD_IPSEC_ESP_ALL_FIELDS)
-+#error Error: please synchronize IOC_ defines!
-+#endif
-+
-+#if CMP_IOC_DEFINE(NET_HEADER_FIELD_MPLS_LABEL_STACK_ALL_FIELDS)
-+#error Error: please synchronize IOC_ defines!
-+#endif
-+
-+#if CMP_IOC_DEFINE(NET_HEADER_FIELD_MACSEC_ALL_FIELDS)
-+#error Error: please synchronize IOC_ defines!
-+#endif
-+
-+/* fm_ioctls.h === fm_ext.h assertions */
-+#if CMP_IOC_DEFINE(FM_MAX_NUM_OF_VALID_PORTS)
-+#error Error: please synchronize IOC_ defines!
-+#endif
-+
-+void LnxWrpPCDIOCTLTypeChecking(void)
-+{
-+ /* fm_ext.h == fm_ioctls.h */
-+ ASSERT_COND(sizeof(ioc_fm_port_bandwidth_params) == sizeof(t_FmPortsBandwidthParams));
-+ ASSERT_COND(sizeof(ioc_fm_revision_info_t) == sizeof(t_FmRevisionInfo));
-+
-+ /* fm_pcd_ext.h == fm_pcd_ioctls.h */
-+ /*ioc_fm_pcd_counters_params_t : NOT USED */
-+ /*ioc_fm_pcd_exception_params_t : private */
-+#if (DPAA_VERSION >= 11)
-+ ASSERT_COND(sizeof(ioc_fm_pcd_manip_frag_capwap_params_t) == sizeof(t_FmPcdManipFragCapwapParams));
-+ ASSERT_COND(sizeof(ioc_fm_pcd_manip_reassem_capwap_params_t) == sizeof(t_FmPcdManipReassemCapwapParams));
-+ ASSERT_COND(sizeof(ioc_fm_pcd_manip_hdr_insrt_by_hdr_params_t) == sizeof(t_FmPcdManipHdrInsrtByHdrParams));
-+ ASSERT_COND(sizeof(ioc_fm_pcd_manip_hdr_insrt_ip_params_t) == sizeof(t_FmPcdManipHdrInsrtIpParams));
-+ ASSERT_COND(sizeof(ioc_fm_pcd_manip_hdr_insrt_t) == sizeof(t_FmPcdManipHdrInsrt));
-+ ASSERT_COND(sizeof(ioc_fm_manip_hdr_info_t) == sizeof(t_FmManipHdrInfo));
-+ ASSERT_COND(sizeof(ioc_fm_pcd_manip_hdr_rmv_by_hdr_params_t) == sizeof(t_FmPcdManipHdrRmvByHdrParams));
-+ ASSERT_COND(sizeof(ioc_fm_pcd_manip_special_offload_capwap_params_t) == sizeof(t_FmPcdManipSpecialOffloadCapwapParams));
-+ ASSERT_COND(sizeof(ioc_fm_pcd_manip_frag_capwap_stats_t) == sizeof(t_FmPcdManipFragCapwapStats));
-+ ASSERT_COND(sizeof(ioc_fm_pcd_manip_reassem_capwap_stats_t) == sizeof(t_FmPcdManipReassemCapwapStats));
-+ ASSERT_COND(sizeof(ioc_fm_pcd_manip_frag_params_t) == sizeof(t_FmPcdManipFragParams));
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+ ASSERT_COND(sizeof(ioc_fm_pcd_prs_label_params_t) == sizeof(t_FmPcdPrsLabelParams));
-+ ASSERT_COND(sizeof(ioc_fm_pcd_prs_sw_params_t) == sizeof(t_FmPcdPrsSwParams));
-+ /*ioc_fm_pcd_kg_dflt_value_params_t : private */
-+ ASSERT_COND(sizeof(ioc_fm_pcd_hdr_protocol_opt_u) == sizeof(u_FmPcdHdrProtocolOpt));
-+ ASSERT_COND(sizeof(ioc_fm_pcd_fields_u) == sizeof(t_FmPcdFields));
-+ ASSERT_COND(sizeof(ioc_fm_pcd_from_hdr_t) == sizeof(t_FmPcdFromHdr));
-+ ASSERT_COND(sizeof(ioc_fm_pcd_from_field_t) == sizeof(t_FmPcdFromField));
-+ ASSERT_COND(sizeof(ioc_fm_pcd_distinction_unit_t) == sizeof(t_FmPcdDistinctionUnit));
-+
-+#if defined(CONFIG_ARM64)
-+ /* different alignment */
-+ ASSERT_COND(sizeof(ioc_fm_pcd_net_env_params_t) == sizeof(t_FmPcdNetEnvParams) + sizeof(void *) + 4);
-+#else
-+#if !defined(CONFIG_COMPAT)
-+ /* different alignment */
-+ ASSERT_COND(sizeof(ioc_fm_pcd_net_env_params_t) == sizeof(t_FmPcdNetEnvParams) + sizeof(void *));
-+#endif
-+#endif
-+ ASSERT_COND(sizeof(ioc_fm_pcd_extract_entry_t) == sizeof(t_FmPcdExtractEntry));
-+ ASSERT_COND(sizeof(ioc_fm_pcd_kg_extract_mask_t) == sizeof(t_FmPcdKgExtractMask));
-+ ASSERT_COND(sizeof(ioc_fm_pcd_kg_extract_dflt_t) == sizeof(t_FmPcdKgExtractDflt));
-+ ASSERT_COND(sizeof(ioc_fm_pcd_kg_key_extract_and_hash_params_t) == sizeof(t_FmPcdKgKeyExtractAndHashParams));
-+ ASSERT_COND(sizeof(ioc_fm_pcd_kg_extracted_or_params_t) == sizeof(t_FmPcdKgExtractedOrParams));
-+ ASSERT_COND(sizeof(ioc_fm_pcd_kg_scheme_counter_t) == sizeof(t_FmPcdKgSchemeCounter));
-+ ASSERT_COND(sizeof(ioc_fm_pcd_kg_plcr_profile_t) == sizeof(t_FmPcdKgPlcrProfile));
-+#if (DPAA_VERSION >= 11)
-+ ASSERT_COND(sizeof(ioc_fm_pcd_kg_storage_profile_t) == sizeof(t_FmPcdKgStorageProfile));
-+#endif
-+ ASSERT_COND(sizeof(ioc_fm_pcd_kg_cc_t) == sizeof(t_FmPcdKgCc));
-+#if !defined(CONFIG_COMPAT)
-+ /* different alignment */
-+ ASSERT_COND(sizeof(ioc_fm_pcd_kg_scheme_params_t) == sizeof(t_FmPcdKgSchemeParams) + sizeof(void *));
-+#endif
-+ ASSERT_COND(sizeof(ioc_fm_pcd_cc_next_cc_params_t) == sizeof(t_FmPcdCcNextCcParams));
-+ ASSERT_COND(sizeof(ioc_fm_pcd_cc_next_plcr_params_t) == sizeof(t_FmPcdCcNextPlcrParams));
-+ ASSERT_COND(sizeof(ioc_fm_pcd_cc_next_enqueue_params_t) == sizeof(t_FmPcdCcNextEnqueueParams));
-+ ASSERT_COND(sizeof(ioc_fm_pcd_cc_next_kg_params_t) == sizeof(t_FmPcdCcNextKgParams));
-+ ASSERT_COND(sizeof(ioc_fm_pcd_cc_next_engine_params_t) == sizeof(t_FmPcdCcNextEngineParams));
-+ ASSERT_COND(sizeof(ioc_fm_pcd_cc_key_params_t) == sizeof(t_FmPcdCcKeyParams));
-+ ASSERT_COND(sizeof(ioc_keys_params_t) == sizeof(t_KeysParams));
-+#if !defined(CONFIG_COMPAT)
-+ /* different alignment */
-+ ASSERT_COND(sizeof(ioc_fm_pcd_cc_node_params_t) == sizeof(t_FmPcdCcNodeParams) + sizeof(void *));
-+ ASSERT_COND(sizeof(ioc_fm_pcd_hash_table_params_t) == sizeof(t_FmPcdHashTableParams) + sizeof(void *));
-+#endif
-+ ASSERT_COND(sizeof(ioc_fm_pcd_cc_grp_params_t) == sizeof(t_FmPcdCcGrpParams));
-+#if !defined(CONFIG_COMPAT)
-+ /* different alignment */
-+ ASSERT_COND(sizeof(ioc_fm_pcd_cc_tree_params_t) == sizeof(t_FmPcdCcTreeParams) + sizeof(void *));
-+#endif
-+ ASSERT_COND(sizeof(ioc_fm_pcd_plcr_byte_rate_mode_param_t) == sizeof(t_FmPcdPlcrByteRateModeParams));
-+ ASSERT_COND(sizeof(ioc_fm_pcd_plcr_non_passthrough_alg_param_t) == sizeof(t_FmPcdPlcrNonPassthroughAlgParams));
-+ ASSERT_COND(sizeof(ioc_fm_pcd_plcr_next_engine_params_u) == sizeof(u_FmPcdPlcrNextEngineParams));
-+ /*ioc_fm_pcd_port_params_t : private */
-+ ASSERT_COND(sizeof(ioc_fm_pcd_plcr_profile_params_t) == sizeof(t_FmPcdPlcrProfileParams) + sizeof(void *));
-+ /*ioc_fm_pcd_cc_tree_modify_next_engine_params_t : private */
-+
-+#ifdef FM_CAPWAP_SUPPORT
-+#error TODO: unsupported feature
-+/*
-+ ASSERT_COND(sizeof(TODO) == sizeof(t_FmPcdManipHdrInsrtByTemplateParams));
-+ ASSERT_COND(sizeof(TODO) == sizeof(t_CapwapFragmentationParams));
-+ ASSERT_COND(sizeof(TODO) == sizeof(t_CapwapReassemblyParams));
-+*/
-+#endif
-+
-+ /*ioc_fm_pcd_cc_node_modify_next_engine_params_t : private */
-+ /*ioc_fm_pcd_cc_node_remove_key_params_t : private */
-+ /*ioc_fm_pcd_cc_node_modify_key_and_next_engine_params_t : private */
-+ /*ioc_fm_pcd_cc_node_modify_key_params_t : private */
-+ /*ioc_fm_manip_hdr_info_t : private */
-+ /*ioc_fm_pcd_hash_table_set_t : private */
-+
-+ ASSERT_COND(sizeof(ioc_fm_pcd_manip_frag_ip_params_t) == sizeof(t_FmPcdManipFragIpParams));
-+ ASSERT_COND(sizeof(ioc_fm_pcd_manip_reassem_ip_params_t) == sizeof(t_FmPcdManipReassemIpParams));
-+ ASSERT_COND(sizeof(ioc_fm_pcd_manip_special_offload_ipsec_params_t) == sizeof(t_FmPcdManipSpecialOffloadIPSecParams));
-+ ASSERT_COND(sizeof(ioc_fm_pcd_manip_special_offload_params_t) == sizeof(t_FmPcdManipSpecialOffloadParams));
-+ ASSERT_COND(sizeof(ioc_fm_pcd_manip_hdr_rmv_generic_params_t) == sizeof(t_FmPcdManipHdrRmvGenericParams));
-+ ASSERT_COND(sizeof(ioc_fm_pcd_manip_hdr_insrt_generic_params_t) == sizeof(t_FmPcdManipHdrInsrtGenericParams));
-+ ASSERT_COND(sizeof(ioc_fm_pcd_manip_hdr_insrt_params_t) == sizeof(t_FmPcdManipHdrInsrtParams));
-+ ASSERT_COND(sizeof(ioc_fm_pcd_manip_hdr_rmv_params_t) == sizeof(t_FmPcdManipHdrRmvParams));
-+ ASSERT_COND(sizeof(ioc_fm_pcd_manip_hdr_params_t) == sizeof(t_FmPcdManipHdrParams));
-+ ASSERT_COND(sizeof(ioc_fm_pcd_manip_frag_params_t) == sizeof(t_FmPcdManipFragParams));
-+ ASSERT_COND(sizeof(ioc_fm_pcd_manip_reassem_params_t) == sizeof(t_FmPcdManipReassemParams));
-+#if !defined(CONFIG_COMPAT)
-+ /* different alignment */
-+ ASSERT_COND(sizeof(ioc_fm_pcd_manip_params_t) == sizeof(t_FmPcdManipParams) + sizeof(void *));
-+#endif
-+ ASSERT_COND(sizeof(ioc_fm_pcd_manip_reassem_ip_stats_t) == sizeof(t_FmPcdManipReassemIpStats));
-+ ASSERT_COND(sizeof(ioc_fm_pcd_manip_frag_ip_stats_t) == sizeof(t_FmPcdManipFragIpStats));
-+ ASSERT_COND(sizeof(ioc_fm_pcd_manip_reassem_stats_t) == sizeof(t_FmPcdManipReassemStats));
-+ ASSERT_COND(sizeof(ioc_fm_pcd_manip_frag_stats_t) == sizeof(t_FmPcdManipFragStats));
-+ ASSERT_COND(sizeof(ioc_fm_pcd_manip_stats_t) == sizeof(t_FmPcdManipStats));
-+#if DPAA_VERSION >= 11
-+ ASSERT_COND(sizeof(ioc_fm_pcd_frm_replic_group_params_t) == sizeof(t_FmPcdFrmReplicGroupParams) + sizeof(void *));
-+#endif
-+
-+ /* fm_port_ext.h == fm_port_ioctls.h */
-+ ASSERT_COND(sizeof(ioc_fm_port_rate_limit_t) == sizeof(t_FmPortRateLimit));
-+ ASSERT_COND(sizeof(ioc_fm_port_pcd_params_t) == sizeof(t_FmPortPcdParams));
-+ ASSERT_COND(sizeof(ioc_fm_pcd_kg_scheme_select_t) == sizeof(t_FmPcdKgSchemeSelect));
-+ ASSERT_COND(sizeof(ioc_fm_pcd_port_schemes_params_t) == sizeof(t_FmPcdPortSchemesParams));
-+ ASSERT_COND(sizeof(ioc_fm_pcd_prs_start_t) == sizeof(t_FmPcdPrsStart));
-+
-+ return;
-+}
-+
-+#define ASSERT_IOC_NET_ENUM(def) ASSERT_COND((unsigned long)e_IOC_NET_##def == (unsigned long)def)
-+
-+void LnxWrpPCDIOCTLEnumChecking(void)
-+{
-+ /* net_ext.h == net_ioctls.h : sampling checks */
-+ ASSERT_IOC_NET_ENUM(HEADER_TYPE_MACSEC);
-+ ASSERT_IOC_NET_ENUM(HEADER_TYPE_PPP);
-+ ASSERT_IOC_NET_ENUM(MAX_HEADER_TYPE_COUNT);
-+
-+ /* fm_ext.h == fm_ioctls.h */
-+ ASSERT_COND((unsigned long)e_IOC_FM_PORT_TYPE_DUMMY == (unsigned long)e_FM_PORT_TYPE_DUMMY);
-+ ASSERT_COND((unsigned long)e_IOC_EX_MURAM_ECC == (unsigned long)e_FM_EX_MURAM_ECC);
-+ ASSERT_COND((unsigned long)e_IOC_FM_COUNTERS_DEQ_CONFIRM == (unsigned long)e_FM_COUNTERS_DEQ_CONFIRM);
-+
-+ /* fm_pcd_ext.h == fm_pcd_ioctls.h */
-+ ASSERT_COND((unsigned long)e_IOC_FM_PCD_PRS_COUNTERS_FPM_COMMAND_STALL_CYCLES == (unsigned long)e_FM_PCD_PRS_COUNTERS_FPM_COMMAND_STALL_CYCLES);
-+ ASSERT_COND((unsigned long)e_IOC_FM_PCD_PRS_EXCEPTION_SINGLE_ECC == (unsigned long)e_FM_PCD_PRS_EXCEPTION_SINGLE_ECC);
-+ ASSERT_COND((unsigned long)e_IOC_FM_PCD_PRS == (unsigned long)e_FM_PCD_PRS);
-+ ASSERT_COND((unsigned long)e_IOC_FM_PCD_EXTRACT_FULL_FIELD == (unsigned long)e_FM_PCD_EXTRACT_FULL_FIELD);
-+ ASSERT_COND((unsigned long)e_IOC_FM_PCD_EXTRACT_FROM_FLOW_ID == (unsigned long)e_FM_PCD_EXTRACT_FROM_FLOW_ID);
-+ ASSERT_COND((unsigned long)e_IOC_FM_PCD_KG_EXTRACT_PORT_PRIVATE_INFO == (unsigned long)e_FM_PCD_KG_EXTRACT_PORT_PRIVATE_INFO);
-+ ASSERT_COND((unsigned long)e_IOC_FM_PCD_KG_DFLT_ILLEGAL == (unsigned long)e_FM_PCD_KG_DFLT_ILLEGAL);
-+ ASSERT_COND((unsigned long)e_IOC_FM_PCD_KG_GENERIC_NOT_FROM_DATA == (unsigned long)e_FM_PCD_KG_GENERIC_NOT_FROM_DATA);
-+ ASSERT_COND((unsigned long)e_IOC_FM_PCD_HDR_INDEX_LAST == (unsigned long)e_FM_PCD_HDR_INDEX_LAST);
-+ ASSERT_COND((unsigned long)e_IOC_FM_PCD_PLCR_SHARED == (unsigned long)e_FM_PCD_PLCR_SHARED);
-+ ASSERT_COND((unsigned long)e_IOC_FM_PCD_PLCR_RFC_4115 == (unsigned long)e_FM_PCD_PLCR_RFC_4115);
-+ ASSERT_COND((unsigned long)e_IOC_FM_PCD_PLCR_COLOR_AWARE == (unsigned long)e_FM_PCD_PLCR_COLOR_AWARE);
-+ ASSERT_COND((unsigned long)e_IOC_FM_PCD_PLCR_OVERRIDE == (unsigned long)e_FM_PCD_PLCR_OVERRIDE);
-+ ASSERT_COND((unsigned long)e_IOC_FM_PCD_PLCR_FULL_FRM_LEN == (unsigned long)e_FM_PCD_PLCR_FULL_FRM_LEN);
-+ ASSERT_COND((unsigned long)e_IOC_FM_PCD_PLCR_ROLLBACK_FULL_FRM_LEN == (unsigned long)e_FM_PCD_PLCR_ROLLBACK_FULL_FRM_LEN);
-+ ASSERT_COND((unsigned long)e_IOC_FM_PCD_PLCR_PACKET_MODE == (unsigned long)e_FM_PCD_PLCR_PACKET_MODE);
-+ ASSERT_COND((unsigned long)e_IOC_FM_PCD_DROP_FRAME == (unsigned long)e_FM_PCD_DROP_FRAME);
-+ ASSERT_COND((unsigned long)e_IOC_FM_PCD_PLCR_PROFILE_RECOLOURED_RED_PACKET_TOTAL_COUNTER == (unsigned long)e_FM_PCD_PLCR_PROFILE_RECOLOURED_RED_PACKET_TOTAL_COUNTER);
-+ ASSERT_COND((unsigned long)e_IOC_FM_PCD_ACTION_INDEXED_LOOKUP == (unsigned long)e_FM_PCD_ACTION_INDEXED_LOOKUP);
-+ ASSERT_COND((unsigned long)e_IOC_FM_PORT_PCD_SUPPORT_PRS_AND_KG_AND_PLCR == (unsigned long)e_FM_PORT_PCD_SUPPORT_PRS_AND_KG_AND_PLCR);
-+#if !defined(FM_CAPWAP_SUPPORT)
-+ ASSERT_COND((unsigned long)e_IOC_FM_PCD_MANIP_INSRT_GENERIC == (unsigned long)e_FM_PCD_MANIP_INSRT_GENERIC);
-+ ASSERT_COND((unsigned long)e_IOC_FM_PCD_MANIP_RMV_GENERIC == (unsigned long)e_FM_PCD_MANIP_RMV_GENERIC);
-+#else
-+ ASSERT_COND((unsigned long)e_IOC_FM_PCD_MANIP_INSRT_BY_TEMPLATE == (unsigned long)e_FM_PCD_MANIP_INSRT_BY_TEMPLATE);
-+ ASSERT_COND((unsigned long)e_IOC_FM_PCD_MANIP_RMV_BY_HDR == (unsigned long)e_FM_PCD_MANIP_RMV_BY_HDR);
-+ ASSERT_COND((unsigned long)e_IOC_FM_PCD_MANIP_RMV_BY_HDR_FROM_START == (unsigned long)e_FM_PCD_MANIP_RMV_BY_HDR_FROM_START);
-+#endif
-+ ASSERT_COND((unsigned long)e_IOC_FM_PCD_MANIP_TIME_OUT_BETWEEN_FRAG == (unsigned long)e_FM_PCD_MANIP_TIME_OUT_BETWEEN_FRAG);
-+ ASSERT_COND((unsigned long)e_IOC_FM_PCD_MANIP_EIGHT_WAYS_HASH == (unsigned long)e_FM_PCD_MANIP_EIGHT_WAYS_HASH);
-+
-+#ifdef FM_CAPWAP_SUPPORT
-+ ASSERT_COND((unsigned long)e_IOC_FM_PCD_STATS_PER_FLOWID == (unsigned long)e_FM_PCD_STATS_PER_FLOWID);
-+#endif
-+ ASSERT_COND((unsigned long)e_IOC_FM_PCD_MANIP_SPECIAL_OFFLOAD == (unsigned long)e_FM_PCD_MANIP_SPECIAL_OFFLOAD);
-+ ASSERT_COND((unsigned long)e_IOC_FM_PCD_CC_STATS_MODE_FRAME == (unsigned long)e_FM_PCD_CC_STATS_MODE_FRAME);
-+ ASSERT_COND((unsigned long)e_IOC_FM_PCD_MANIP_CONTINUE_WITHOUT_FRAG == (unsigned long)e_FM_PCD_MANIP_CONTINUE_WITHOUT_FRAG);
-+ ASSERT_COND((unsigned long)e_IOC_FM_PCD_MANIP_SPECIAL_OFFLOAD_IPSEC == (unsigned long)e_FM_PCD_MANIP_SPECIAL_OFFLOAD_IPSEC);
-+
-+ /* fm_port_ext.h == fm_port_ioctls.h */
-+#if !defined(FM_CAPWAP_SUPPORT)
-+ ASSERT_COND((unsigned long)e_IOC_FM_PORT_PCD_SUPPORT_PRS_AND_KG_AND_PLCR == (unsigned long)e_FM_PORT_PCD_SUPPORT_PRS_AND_KG_AND_PLCR);
-+#else
-+ ASSERT_COND((unsigned long)e_IOC_FM_PORT_PCD_SUPPORT_CC_AND_KG_AND_PLCR == (unsigned long)e_FM_PORT_PCD_SUPPORT_CC_AND_KG_AND_PLCR);
-+#endif
-+ ASSERT_COND((unsigned long)e_IOC_FM_PORT_COUNTERS_DEQ_CONFIRM == (unsigned long)e_FM_PORT_COUNTERS_DEQ_CONFIRM);
-+ ASSERT_COND((unsigned long)e_IOC_FM_PORT_DUAL_RATE_LIMITER_SCALE_DOWN_BY_8 == (unsigned long)e_FM_PORT_DUAL_RATE_LIMITER_SCALE_DOWN_BY_8);
-+
-+ return;
-+}
-+
-+static t_Error LnxwrpFmPcdIOCTL(t_LnxWrpFmDev *p_LnxWrpFmDev, unsigned int cmd, unsigned long arg, bool compat)
-+{
-+ t_Error err = E_OK;
-+
-+/*
-+Status: PCD API to fmlib (file: drivers/net/dpa/NetCommSw/inc/Peripherals/fm_pcd_ext.h):
-+
-+ FM_PCD_PrsLoadSw
-+ FM_PCD_SetAdvancedOffloadSupport
-+ FM_PCD_Enable
-+ FM_PCD_Disable
-+ FM_PCD_ForceIntr
-+ FM_PCD_SetException
-+ FM_PCD_KgSetAdditionalDataAfterParsing
-+ FM_PCD_KgSetDfltValue
-+ FM_PCD_NetEnvCharacteristicsSet
-+ FM_PCD_NetEnvCharacteristicsDelete
-+ FM_PCD_KgSchemeSet
-+ FM_PCD_KgSchemeDelete
-+ FM_PCD_MatchTableSet
-+ FM_PCD_MatchTableDelete
-+ FM_PCD_CcRootBuild
-+ FM_PCD_CcRootDelete
-+ FM_PCD_PlcrProfileSet
-+ FM_PCD_PlcrProfileDelete
-+ FM_PCD_CcRootModifyNextEngine
-+ FM_PCD_MatchTableModifyNextEngine
-+ FM_PCD_MatchTableModifyMissNextEngine
-+ FM_PCD_MatchTableRemoveKey
-+ FM_PCD_MatchTableAddKey
-+ FM_PCD_MatchTableModifyKeyAndNextEngine
-+ FM_PCD_HashTableSet
-+ FM_PCD_HashTableDelete
-+ FM_PCD_HashTableAddKey
-+ FM_PCD_HashTableRemoveKey
-+ FM_PCD_MatchTableModifyKey
-+ FM_PCD_ManipNodeReplace
-+ FM_PCD_ManipNodeSet
-+ FM_PCD_ManipNodeDelete
-+
-+Status: not exported, should be thru sysfs
-+ FM_PCD_KgSchemeGetCounter
-+ FM_PCD_KgSchemeSetCounter
-+ FM_PCD_PlcrProfileGetCounter
-+ FM_PCD_PlcrProfileSetCounter
-+
-+Status: not exported
-+ FM_PCD_MatchTableFindNRemoveKey
-+ FM_PCD_MatchTableFindNModifyNextEngine
-+ FM_PCD_MatchTableFindNModifyKeyAndNextEngine
-+ FM_PCD_MatchTableFindNModifyKey
-+ FM_PCD_MatchTableGetIndexedHashBucket
-+ FM_PCD_MatchTableGetNextEngine
-+ FM_PCD_MatchTableGetKeyCounter
-+
-+Status: not exported, would be nice to have
-+ FM_PCD_HashTableModifyNextEngine
-+ FM_PCD_HashTableModifyMissNextEngine
-+ FM_PCD_HashTableGetMissNextEngine
-+ FM_PCD_ManipGetStatistics
-+
-+Status: not exported
-+#if DPAA_VERSION >= 11
-+
-+ FM_VSP_GetStatistics -- it's not available yet
-+#endif
-+
-+Status: feature not supported
-+#ifdef FM_CAPWAP_SUPPORT
-+#error unsupported feature
-+ FM_PCD_StatisticsSetNode
-+#endif
-+
-+ */
-+ _fm_ioctl_dbg("cmd:0x%08x(type:0x%02x, nr:%u).\n",
-+ cmd, _IOC_TYPE(cmd), _IOC_NR(cmd) - 20);
-+
-+ switch (cmd)
-+ {
-+#if defined(CONFIG_COMPAT)
-+ case FM_PCD_IOC_PRS_LOAD_SW_COMPAT:
-+#endif
-+ case FM_PCD_IOC_PRS_LOAD_SW:
-+ {
-+ ioc_fm_pcd_prs_sw_params_t *param;
-+ uint8_t *p_code;
-+
-+ param = (ioc_fm_pcd_prs_sw_params_t *) XX_Malloc(sizeof(ioc_fm_pcd_prs_sw_params_t));
-+ if (!param)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+
-+ memset(param, 0, sizeof(ioc_fm_pcd_prs_sw_params_t));
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_pcd_prs_sw_params_t *compat_param;
-+
-+ compat_param = (ioc_compat_fm_pcd_prs_sw_params_t *) XX_Malloc(
-+ sizeof(ioc_compat_fm_pcd_prs_sw_params_t));
-+ if (!compat_param)
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+ }
-+
-+ memset(compat_param, 0, sizeof(ioc_compat_fm_pcd_prs_sw_params_t));
-+ if (copy_from_user(compat_param,
-+ (ioc_compat_fm_pcd_prs_sw_params_t *) compat_ptr(arg),
-+ sizeof(ioc_compat_fm_pcd_prs_sw_params_t)))
-+ {
-+ XX_Free(compat_param);
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+
-+ compat_fm_pcd_prs_sw(compat_param, param, COMPAT_US_TO_K);
-+
-+ XX_Free(compat_param);
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_from_user(param, (ioc_fm_pcd_prs_sw_params_t *)arg,
-+ sizeof(ioc_fm_pcd_prs_sw_params_t)))
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+ }
-+
-+ if (!param->p_code || !param->size)
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+
-+ p_code = (uint8_t *) XX_Malloc(param->size);
-+ if (!p_code)
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+ }
-+
-+ memset(p_code, 0, param->size);
-+ if (copy_from_user(p_code, param->p_code, param->size))
-+ {
-+ XX_Free(p_code);
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+
-+ param->p_code = p_code;
-+
-+ err = FM_PCD_PrsLoadSw(p_LnxWrpFmDev->h_PcdDev, (t_FmPcdPrsSwParams*)param);
-+
-+ XX_Free(p_code);
-+ XX_Free(param);
-+ break;
-+ }
-+
-+ case FM_PCD_IOC_SET_ADVANCED_OFFLOAD_SUPPORT:
-+ err = FM_PCD_SetAdvancedOffloadSupport(p_LnxWrpFmDev->h_PcdDev);
-+ break;
-+
-+ case FM_PCD_IOC_ENABLE:
-+ err = FM_PCD_Enable(p_LnxWrpFmDev->h_PcdDev);
-+ break;
-+
-+ case FM_PCD_IOC_DISABLE:
-+ err = FM_PCD_Disable(p_LnxWrpFmDev->h_PcdDev);
-+ break;
-+
-+ case FM_PCD_IOC_FORCE_INTR:
-+ {
-+ int exception;
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ if (get_user(exception, (int *) compat_ptr(arg)))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+ else
-+#endif
-+ {
-+ if (get_user(exception, (int *)arg))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+
-+ err = FM_PCD_ForceIntr(p_LnxWrpFmDev->h_PcdDev, (e_FmPcdExceptions)exception);
-+ break;
-+ }
-+
-+ case FM_PCD_IOC_SET_EXCEPTION:
-+ {
-+ ioc_fm_pcd_exception_params_t *param;
-+
-+ param = (ioc_fm_pcd_exception_params_t *) XX_Malloc(
-+ sizeof(ioc_fm_pcd_exception_params_t));
-+ if (!param)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+
-+ memset(param, 0, sizeof(ioc_fm_pcd_exception_params_t));
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ if (copy_from_user(param, (ioc_fm_pcd_exception_params_t *)compat_ptr(arg),
-+ sizeof(ioc_fm_pcd_exception_params_t)))
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_from_user(param, (ioc_fm_pcd_exception_params_t *)arg,
-+ sizeof(ioc_fm_pcd_exception_params_t)))
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+ }
-+
-+ err = FM_PCD_SetException(p_LnxWrpFmDev->h_PcdDev, param->exception, param->enable);
-+
-+ XX_Free(param);
-+ break;
-+ }
-+
-+ case FM_PCD_IOC_KG_SET_ADDITIONAL_DATA_AFTER_PARSING:
-+ {
-+ uint8_t payloadOffset;
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ if (get_user(payloadOffset, (uint8_t*) compat_ptr(arg)))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+ else
-+#endif
-+ {
-+ if (get_user(payloadOffset, (uint8_t*) arg))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+
-+ err = FM_PCD_KgSetAdditionalDataAfterParsing(p_LnxWrpFmDev->h_PcdDev, payloadOffset);
-+ break;
-+ }
-+
-+ case FM_PCD_IOC_KG_SET_DFLT_VALUE:
-+ {
-+ ioc_fm_pcd_kg_dflt_value_params_t *param;
-+
-+ param = (ioc_fm_pcd_kg_dflt_value_params_t *) XX_Malloc(
-+ sizeof(ioc_fm_pcd_kg_dflt_value_params_t));
-+ if (!param)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+
-+ memset(param, 0, sizeof(ioc_fm_pcd_kg_dflt_value_params_t));
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ if (copy_from_user(param, (ioc_fm_pcd_kg_dflt_value_params_t *)compat_ptr(arg),
-+ sizeof(ioc_fm_pcd_kg_dflt_value_params_t)))
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_from_user(param, (ioc_fm_pcd_kg_dflt_value_params_t *)arg,
-+ sizeof(ioc_fm_pcd_kg_dflt_value_params_t)))
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+ }
-+
-+ err = FM_PCD_KgSetDfltValue(p_LnxWrpFmDev->h_PcdDev, param->valueId, param->value);
-+
-+ XX_Free(param);
-+ break;
-+ }
-+
-+#if defined(CONFIG_COMPAT)
-+ case FM_PCD_IOC_NET_ENV_CHARACTERISTICS_SET_COMPAT:
-+#endif
-+ case FM_PCD_IOC_NET_ENV_CHARACTERISTICS_SET:
-+ {
-+ ioc_fm_pcd_net_env_params_t *param;
-+
-+ param = (ioc_fm_pcd_net_env_params_t *) XX_Malloc(sizeof(ioc_fm_pcd_net_env_params_t));
-+ if (!param)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+
-+ memset(param, 0, sizeof(ioc_fm_pcd_net_env_params_t));
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_pcd_net_env_params_t *compat_param;
-+
-+ compat_param = (ioc_compat_fm_pcd_net_env_params_t *) XX_Malloc(
-+ sizeof(ioc_compat_fm_pcd_net_env_params_t));
-+ if (!compat_param)
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+ }
-+
-+ memset(compat_param, 0, sizeof(ioc_compat_fm_pcd_net_env_params_t));
-+ if (copy_from_user(compat_param, (ioc_compat_fm_pcd_net_env_params_t *) compat_ptr(arg),
-+ sizeof(ioc_compat_fm_pcd_net_env_params_t)))
-+ {
-+ XX_Free(compat_param);
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+
-+ compat_copy_fm_pcd_net_env(compat_param, param, COMPAT_US_TO_K);
-+ XX_Free(compat_param);
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_from_user(param, (ioc_fm_pcd_net_env_params_t *) arg,
-+ sizeof(ioc_fm_pcd_net_env_params_t)))
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+ }
-+
-+ param->id = FM_PCD_NetEnvCharacteristicsSet(p_LnxWrpFmDev->h_PcdDev, (t_FmPcdNetEnvParams*)param);
-+
-+ if (!param->id)
-+ {
-+ XX_Free(param);
-+ err = E_INVALID_VALUE;
-+ /* Since the LLD has no errno-style error reporting,
-+ we're left here with no other option than to report
-+ a generic E_INVALID_VALUE */
-+ break;
-+ }
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_pcd_net_env_params_t *compat_param;
-+
-+ compat_param = (ioc_compat_fm_pcd_net_env_params_t *) XX_Malloc(
-+ sizeof(ioc_compat_fm_pcd_net_env_params_t));
-+ if (!compat_param)
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+ }
-+
-+ memset(compat_param, 0, sizeof(ioc_compat_fm_pcd_net_env_params_t));
-+ compat_copy_fm_pcd_net_env(compat_param, param, COMPAT_K_TO_US);
-+
-+ if (copy_to_user((ioc_compat_fm_pcd_net_env_params_t *) compat_ptr(arg),
-+ compat_param,
-+ sizeof(ioc_compat_fm_pcd_net_env_params_t)))
-+ err = E_READ_FAILED;
-+
-+ XX_Free(compat_param);
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_to_user((ioc_fm_pcd_net_env_params_t *)arg,
-+ param,
-+ sizeof(ioc_fm_pcd_net_env_params_t)))
-+ err = E_READ_FAILED;
-+ }
-+
-+ XX_Free(param);
-+ break;
-+ }
-+
-+#if defined(CONFIG_COMPAT)
-+ case FM_PCD_IOC_NET_ENV_CHARACTERISTICS_DELETE_COMPAT:
-+#endif
-+ case FM_PCD_IOC_NET_ENV_CHARACTERISTICS_DELETE:
-+ {
-+ ioc_fm_obj_t id;
-+
-+ memset(&id, 0 , sizeof(ioc_fm_obj_t));
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_obj_t compat_id;
-+
-+ if (copy_from_user(&compat_id, (ioc_compat_fm_obj_t *) compat_ptr(arg), sizeof(ioc_compat_fm_obj_t)))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+
-+ compat_obj_delete(&compat_id, &id);
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_from_user(&id, (ioc_fm_obj_t *) arg, sizeof(ioc_fm_obj_t)))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+
-+ err = FM_PCD_NetEnvCharacteristicsDelete(id.obj);
-+ break;
-+ }
-+
-+#if defined(CONFIG_COMPAT)
-+ case FM_PCD_IOC_KG_SCHEME_SET_COMPAT:
-+#endif
-+ case FM_PCD_IOC_KG_SCHEME_SET:
-+ {
-+ ioc_fm_pcd_kg_scheme_params_t *param;
-+
-+ param = (ioc_fm_pcd_kg_scheme_params_t *) XX_Malloc(sizeof(ioc_fm_pcd_kg_scheme_params_t));
-+ if (!param)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+
-+ memset(param, 0, sizeof(ioc_fm_pcd_kg_scheme_params_t));
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_pcd_kg_scheme_params_t *compat_param = NULL;
-+
-+ compat_param = (ioc_compat_fm_pcd_kg_scheme_params_t *) XX_Malloc(
-+ sizeof(ioc_compat_fm_pcd_kg_scheme_params_t));
-+ if (!compat_param)
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+ }
-+
-+ memset(compat_param, 0, sizeof(ioc_compat_fm_pcd_kg_scheme_params_t));
-+
-+ if (copy_from_user(compat_param, (ioc_compat_fm_pcd_kg_scheme_params_t *) compat_ptr(arg),
-+ sizeof(ioc_compat_fm_pcd_kg_scheme_params_t)))
-+ {
-+ XX_Free(compat_param);
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+
-+ compat_copy_fm_pcd_kg_scheme(compat_param, param, COMPAT_US_TO_K);
-+
-+ XX_Free(compat_param);
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_from_user(param, (ioc_fm_pcd_kg_scheme_params_t *)arg,
-+ sizeof(ioc_fm_pcd_kg_scheme_params_t)))
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+ }
-+
-+ param->id = FM_PCD_KgSchemeSet(p_LnxWrpFmDev->h_PcdDev, (t_FmPcdKgSchemeParams*)param);
-+
-+ if (!param->id)
-+ {
-+ XX_Free(param);
-+ err = E_INVALID_VALUE;
-+ /* Since the LLD has no errno-style error reporting,
-+ we're left here with no other option than to report
-+ a generic E_INVALID_VALUE */
-+ break;
-+ }
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_pcd_kg_scheme_params_t *compat_param;
-+
-+ compat_param = (ioc_compat_fm_pcd_kg_scheme_params_t *) XX_Malloc(
-+ sizeof(ioc_compat_fm_pcd_kg_scheme_params_t));
-+ if (!compat_param)
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+ }
-+
-+ memset(compat_param, 0, sizeof(ioc_compat_fm_pcd_kg_scheme_params_t));
-+ compat_copy_fm_pcd_kg_scheme(compat_param, param, COMPAT_K_TO_US);
-+ if (copy_to_user((ioc_compat_fm_pcd_kg_scheme_params_t *)compat_ptr(arg),
-+ compat_param,
-+ sizeof(ioc_compat_fm_pcd_kg_scheme_params_t)))
-+ err = E_READ_FAILED;
-+
-+ XX_Free(compat_param);
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_to_user((ioc_fm_pcd_kg_scheme_params_t *)arg,
-+ param,
-+ sizeof(ioc_fm_pcd_kg_scheme_params_t)))
-+ err = E_READ_FAILED;
-+ }
-+
-+ XX_Free(param);
-+ break;
-+ }
-+
-+#if defined(CONFIG_COMPAT)
-+ case FM_PCD_IOC_KG_SCHEME_GET_CNTR_COMPAT:
-+#endif
-+ case FM_PCD_IOC_KG_SCHEME_GET_CNTR:
-+ {
-+ ioc_fm_pcd_kg_scheme_spc_t *param;
-+
-+ param = (ioc_fm_pcd_kg_scheme_spc_t *) XX_Malloc(sizeof(ioc_fm_pcd_kg_scheme_spc_t));
-+ if (!param)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+
-+ memset(param, 0, sizeof(ioc_fm_pcd_kg_scheme_spc_t));
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_pcd_kg_scheme_spc_t *compat_param = NULL;
-+
-+ compat_param = (ioc_compat_fm_pcd_kg_scheme_spc_t *) XX_Malloc(
-+ sizeof(ioc_compat_fm_pcd_kg_scheme_spc_t));
-+ if (!compat_param)
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+ }
-+
-+ memset(compat_param, 0, sizeof(ioc_compat_fm_pcd_kg_scheme_spc_t));
-+
-+ if (copy_from_user(compat_param, (ioc_compat_fm_pcd_kg_scheme_spc_t *) compat_ptr(arg),
-+ sizeof(ioc_compat_fm_pcd_kg_scheme_spc_t)))
-+ {
-+ XX_Free(compat_param);
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+
-+ compat_copy_fm_pcd_kg_scheme_spc(compat_param, param, COMPAT_US_TO_K);
-+
-+ XX_Free(compat_param);
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_from_user(param, (ioc_fm_pcd_kg_scheme_spc_t *)arg,
-+ sizeof(ioc_fm_pcd_kg_scheme_spc_t)))
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+ }
-+
-+ param->val = FM_PCD_KgSchemeGetCounter((t_Handle)param->id);
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_pcd_kg_scheme_spc_t *compat_param;
-+
-+ compat_param = (ioc_compat_fm_pcd_kg_scheme_spc_t *) XX_Malloc(
-+ sizeof(ioc_compat_fm_pcd_kg_scheme_spc_t));
-+ if (!compat_param)
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+ }
-+
-+ memset(compat_param, 0, sizeof(ioc_compat_fm_pcd_kg_scheme_spc_t));
-+ compat_copy_fm_pcd_kg_scheme_spc(compat_param, param, COMPAT_K_TO_US);
-+ if (copy_to_user((ioc_compat_fm_pcd_kg_scheme_spc_t *)compat_ptr(arg),
-+ compat_param,
-+ sizeof(ioc_compat_fm_pcd_kg_scheme_spc_t)))
-+ err = E_READ_FAILED;
-+
-+ XX_Free(compat_param);
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_to_user((ioc_fm_pcd_kg_scheme_spc_t *)arg,
-+ param,
-+ sizeof(ioc_fm_pcd_kg_scheme_spc_t)))
-+ err = E_READ_FAILED;
-+ }
-+
-+ XX_Free(param);
-+ break;
-+ }
-+
-+#if defined(CONFIG_COMPAT)
-+ case FM_PCD_IOC_KG_SCHEME_DELETE_COMPAT:
-+#endif
-+ case FM_PCD_IOC_KG_SCHEME_DELETE:
-+ {
-+ ioc_fm_obj_t id;
-+
-+ memset(&id, 0 , sizeof(ioc_fm_obj_t));
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_obj_t compat_id;
-+
-+ if (copy_from_user(&compat_id, (ioc_compat_fm_obj_t *) compat_ptr(arg), sizeof(ioc_compat_fm_obj_t)))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+
-+ compat_obj_delete(&compat_id, &id);
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_from_user(&id, (ioc_fm_obj_t *) arg, sizeof(ioc_fm_obj_t)))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+
-+ err = FM_PCD_KgSchemeDelete(id.obj);
-+ break;
-+ }
-+
-+#if defined(CONFIG_COMPAT)
-+ case FM_PCD_IOC_MATCH_TABLE_SET_COMPAT:
-+#endif
-+ case FM_PCD_IOC_MATCH_TABLE_SET:
-+ {
-+ ioc_fm_pcd_cc_node_params_t *param;
-+ uint8_t *keys;
-+ uint8_t *masks;
-+ int i,k;
-+
-+ param = (ioc_fm_pcd_cc_node_params_t *) XX_Malloc(
-+ sizeof(ioc_fm_pcd_cc_node_params_t) +
-+ 2 * IOC_FM_PCD_MAX_NUM_OF_KEYS * IOC_FM_PCD_MAX_SIZE_OF_KEY);
-+ if (!param)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+
-+ memset(param, 0, sizeof(ioc_fm_pcd_cc_node_params_t) +
-+ 2 * IOC_FM_PCD_MAX_NUM_OF_KEYS * IOC_FM_PCD_MAX_SIZE_OF_KEY);
-+
-+ keys = (uint8_t *) (param + 1);
-+ masks = keys + IOC_FM_PCD_MAX_NUM_OF_KEYS * IOC_FM_PCD_MAX_SIZE_OF_KEY;
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_pcd_cc_node_params_t *compat_param;
-+
-+ compat_param = (ioc_compat_fm_pcd_cc_node_params_t *) XX_Malloc(
-+ sizeof(ioc_compat_fm_pcd_cc_node_params_t) +
-+ 2 * IOC_FM_PCD_MAX_NUM_OF_KEYS * IOC_FM_PCD_MAX_SIZE_OF_KEY);
-+ if (!compat_param)
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+ }
-+
-+ memset(compat_param, 0, sizeof(ioc_compat_fm_pcd_cc_node_params_t) +
-+ 2 * IOC_FM_PCD_MAX_NUM_OF_KEYS * IOC_FM_PCD_MAX_SIZE_OF_KEY);
-+
-+ if (copy_from_user(compat_param,
-+ (ioc_compat_fm_pcd_cc_node_params_t *)compat_ptr(arg),
-+ sizeof(ioc_compat_fm_pcd_cc_node_params_t)))
-+ {
-+ XX_Free(compat_param);
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+
-+ compat_copy_fm_pcd_cc_node(compat_param, param, COMPAT_US_TO_K);
-+
-+ XX_Free(compat_param);
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_from_user(param, (ioc_fm_pcd_cc_node_params_t *)arg, sizeof(ioc_fm_pcd_cc_node_params_t)))
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+ }
-+
-+ ASSERT_COND(param->keys_params.num_of_keys <= IOC_FM_PCD_MAX_NUM_OF_KEYS);
-+ ASSERT_COND(param->keys_params.key_size <= IOC_FM_PCD_MAX_SIZE_OF_KEY);
-+
-+ /* support for indexed lookup */
-+ if( !(param->extract_cc_params.type == e_IOC_FM_PCD_EXTRACT_NON_HDR &&
-+ param->extract_cc_params.extract_params.extract_non_hdr.src == e_IOC_FM_PCD_EXTRACT_FROM_HASH &&
-+ param->extract_cc_params.extract_params.extract_non_hdr.action == e_IOC_FM_PCD_ACTION_INDEXED_LOOKUP))
-+ {
-+ for (i=0, k=0;
-+ i < param->keys_params.num_of_keys;
-+ i++, k += IOC_FM_PCD_MAX_SIZE_OF_KEY)
-+ {
-+ if (param->keys_params.key_params[i].p_key &&
-+ param->keys_params.key_size)
-+ {
-+ if (copy_from_user(&keys[k],
-+ param->keys_params.key_params[i].p_key,
-+ param->keys_params.key_size))
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+
-+ param->keys_params.key_params[i].p_key = &keys[k];
-+ }
-+
-+ if (param->keys_params.key_params[i].p_mask)
-+ {
-+ if (copy_from_user(&masks[k],
-+ param->keys_params.key_params[i].p_mask,
-+ param->keys_params.key_size))
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+
-+ param->keys_params.key_params[i].p_mask = &masks[k];
-+ }
-+ }
-+ }
-+
-+ param->id = FM_PCD_MatchTableSet(p_LnxWrpFmDev->h_PcdDev, (t_FmPcdCcNodeParams*)param);
-+
-+ if (!param->id) {
-+ XX_Free(param);
-+ err = E_INVALID_VALUE;
-+ /* Since the LLD has no errno-style error reporting,
-+ we're left here with no other option than to report
-+ a generic E_INVALID_VALUE */
-+ break;
-+ }
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_pcd_cc_node_params_t *compat_param;
-+ compat_param = (ioc_compat_fm_pcd_cc_node_params_t *) XX_Malloc(
-+ sizeof(ioc_compat_fm_pcd_cc_node_params_t) +
-+ 2 * IOC_FM_PCD_MAX_NUM_OF_KEYS * IOC_FM_PCD_MAX_SIZE_OF_KEY);
-+ if (!compat_param)
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+ }
-+
-+ memset(compat_param, 0, sizeof(ioc_compat_fm_pcd_cc_node_params_t) +
-+ 2 * IOC_FM_PCD_MAX_NUM_OF_KEYS * IOC_FM_PCD_MAX_SIZE_OF_KEY);
-+ compat_copy_fm_pcd_cc_node(compat_param, param, COMPAT_K_TO_US);
-+
-+ if (copy_to_user((ioc_compat_fm_pcd_cc_node_params_t *)compat_ptr(arg),
-+ compat_param,
-+ sizeof(ioc_compat_fm_pcd_cc_node_params_t)))
-+ err = E_READ_FAILED;
-+
-+ XX_Free(compat_param);
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_to_user((ioc_fm_pcd_cc_node_params_t *)arg,
-+ param,
-+ sizeof(ioc_fm_pcd_cc_node_params_t)))
-+ err = E_READ_FAILED;
-+ }
-+
-+ XX_Free(param);
-+ break;
-+ }
-+
-+#if defined(CONFIG_COMPAT)
-+ case FM_PCD_IOC_MATCH_TABLE_DELETE_COMPAT:
-+#endif
-+ case FM_PCD_IOC_MATCH_TABLE_DELETE:
-+ {
-+ ioc_fm_obj_t id;
-+
-+ memset(&id, 0 , sizeof(ioc_fm_obj_t));
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_obj_t compat_id;
-+
-+ if (copy_from_user(&compat_id, (ioc_compat_fm_obj_t *) compat_ptr(arg), sizeof(ioc_compat_fm_obj_t)))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+
-+ compat_obj_delete(&compat_id, &id);
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_from_user(&id, (ioc_fm_obj_t *) arg, sizeof(ioc_fm_obj_t)))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+
-+ err = FM_PCD_MatchTableDelete(id.obj);
-+ break;
-+ }
-+
-+#if defined(CONFIG_COMPAT)
-+ case FM_PCD_IOC_CC_ROOT_BUILD_COMPAT:
-+#endif
-+ case FM_PCD_IOC_CC_ROOT_BUILD:
-+ {
-+ ioc_fm_pcd_cc_tree_params_t *param;
-+
-+ param = (ioc_fm_pcd_cc_tree_params_t *) XX_Malloc(sizeof(ioc_fm_pcd_cc_tree_params_t));
-+ if (!param)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+
-+ memset(param, 0, sizeof(ioc_fm_pcd_cc_tree_params_t));
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_pcd_cc_tree_params_t *compat_param;
-+
-+ compat_param = (ioc_compat_fm_pcd_cc_tree_params_t *) XX_Malloc(
-+ sizeof(ioc_compat_fm_pcd_cc_tree_params_t));
-+ if (!compat_param)
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+ }
-+
-+ memset(compat_param, 0, sizeof(ioc_compat_fm_pcd_cc_tree_params_t));
-+ if (copy_from_user(compat_param,
-+ (ioc_compat_fm_pcd_cc_tree_params_t *)compat_ptr(arg),
-+ sizeof(ioc_compat_fm_pcd_cc_tree_params_t)))
-+ {
-+ XX_Free(compat_param);
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+
-+ compat_copy_fm_pcd_cc_tree(compat_param, param, COMPAT_US_TO_K);
-+
-+ XX_Free(compat_param);
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_from_user(param, (ioc_fm_pcd_cc_tree_params_t *)arg,
-+ sizeof(ioc_fm_pcd_cc_tree_params_t)))
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+ }
-+
-+ param->id = FM_PCD_CcRootBuild(p_LnxWrpFmDev->h_PcdDev, (t_FmPcdCcTreeParams*)param);
-+
-+ if (!param->id) {
-+ XX_Free(param);
-+ err = E_INVALID_VALUE;
-+ /* Since the LLD has no errno-style error reporting,
-+ we're left here with no other option than to report
-+ a generic E_INVALID_VALUE */
-+ break;
-+ }
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_pcd_cc_tree_params_t *compat_param;
-+
-+ compat_param = (ioc_compat_fm_pcd_cc_tree_params_t *) XX_Malloc(sizeof(ioc_compat_fm_pcd_cc_tree_params_t));
-+ if (!compat_param)
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+ }
-+
-+ memset(compat_param, 0, sizeof(ioc_compat_fm_pcd_cc_tree_params_t));
-+
-+ compat_copy_fm_pcd_cc_tree(compat_param, param, COMPAT_K_TO_US);
-+
-+ if (copy_to_user((ioc_compat_fm_pcd_cc_tree_params_t *)compat_ptr(arg),
-+ compat_param,
-+ sizeof(ioc_compat_fm_pcd_cc_tree_params_t)))
-+ err = E_READ_FAILED;
-+
-+ XX_Free(compat_param);
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_to_user((ioc_fm_pcd_cc_tree_params_t *)arg,
-+ param,
-+ sizeof(ioc_fm_pcd_cc_tree_params_t)))
-+ err = E_READ_FAILED;
-+ }
-+
-+ XX_Free(param);
-+ break;
-+ }
-+
-+#if defined(CONFIG_COMPAT)
-+ case FM_PCD_IOC_CC_ROOT_DELETE_COMPAT:
-+#endif
-+ case FM_PCD_IOC_CC_ROOT_DELETE:
-+ {
-+ ioc_fm_obj_t id;
-+
-+ memset(&id, 0 , sizeof(ioc_fm_obj_t));
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_obj_t compat_id;
-+
-+ if (copy_from_user(&compat_id, (ioc_compat_fm_obj_t *) compat_ptr(arg), sizeof(ioc_compat_fm_obj_t)))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+
-+ compat_obj_delete(&compat_id, &id);
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_from_user(&id, (ioc_fm_obj_t *) arg, sizeof(ioc_fm_obj_t)))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+
-+ err = FM_PCD_CcRootDelete(id.obj);
-+ break;
-+ }
-+
-+#if defined(CONFIG_COMPAT)
-+ case FM_PCD_IOC_PLCR_PROFILE_SET_COMPAT:
-+#endif
-+ case FM_PCD_IOC_PLCR_PROFILE_SET:
-+ {
-+ ioc_fm_pcd_plcr_profile_params_t *param;
-+
-+ param = (ioc_fm_pcd_plcr_profile_params_t *) XX_Malloc(
-+ sizeof(ioc_fm_pcd_plcr_profile_params_t));
-+ if (!param)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+
-+ memset(param, 0, sizeof(ioc_fm_pcd_plcr_profile_params_t));
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_pcd_plcr_profile_params_t *compat_param;
-+
-+ compat_param = (ioc_compat_fm_pcd_plcr_profile_params_t *) XX_Malloc(
-+ sizeof(ioc_compat_fm_pcd_plcr_profile_params_t));
-+ if (!compat_param)
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+ }
-+
-+ memset(compat_param, 0, sizeof(ioc_compat_fm_pcd_plcr_profile_params_t));
-+ if (copy_from_user(compat_param, (
-+ ioc_compat_fm_pcd_plcr_profile_params_t *)compat_ptr(arg),
-+ sizeof(ioc_compat_fm_pcd_plcr_profile_params_t)))
-+ {
-+ XX_Free(compat_param);
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+
-+ compat_copy_fm_pcd_plcr_profile(compat_param, param, COMPAT_US_TO_K);
-+
-+ XX_Free(compat_param);
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_from_user(param, (ioc_fm_pcd_plcr_profile_params_t *)arg,
-+ sizeof(ioc_fm_pcd_plcr_profile_params_t)))
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+ }
-+
-+ if (!param->modify &&
-+ (((t_FmPcdPlcrProfileParams*)param)->id.newParams.profileType != e_FM_PCD_PLCR_SHARED))
-+ {
-+ t_Handle h_Port;
-+ ioc_fm_pcd_port_params_t *port_params;
-+
-+ port_params = (ioc_fm_pcd_port_params_t*) XX_Malloc(sizeof(ioc_fm_pcd_port_params_t));
-+ if (!port_params)
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+ }
-+
-+ memset(port_params, 0, sizeof(ioc_fm_pcd_port_params_t));
-+ if (copy_from_user(port_params, (ioc_fm_pcd_port_params_t*)((t_FmPcdPlcrProfileParams*)param)->id.newParams.h_FmPort,
-+ sizeof(ioc_fm_pcd_port_params_t)))
-+ {
-+ XX_Free(port_params);
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+
-+ switch(port_params->port_type)
-+ {
-+ case (e_IOC_FM_PORT_TYPE_RX):
-+ if (port_params->port_id < FM_MAX_NUM_OF_1G_RX_PORTS) {
-+ h_Port = p_LnxWrpFmDev->rxPorts[port_params->port_id].h_Dev;
-+ break;
-+ }
-+ goto invalid_port_id;
-+
-+ case (e_IOC_FM_PORT_TYPE_RX_10G):
-+ if (port_params->port_id < FM_MAX_NUM_OF_10G_RX_PORTS) {
-+#ifndef CONFIG_FMAN_ARM
-+ if (IS_T1023_T1024) {
-+ h_Port = p_LnxWrpFmDev->rxPorts[port_params->port_id].h_Dev;
-+ } else {
-+#else
-+ {
-+#endif
-+ h_Port = p_LnxWrpFmDev->rxPorts[port_params->port_id + FM_MAX_NUM_OF_1G_RX_PORTS].h_Dev;
-+ }
-+ break;
-+ }
-+ goto invalid_port_id;
-+
-+ case (e_IOC_FM_PORT_TYPE_OH_OFFLINE_PARSING):
-+ if (port_params->port_id && port_params->port_id < FM_MAX_NUM_OF_OH_PORTS) {
-+ h_Port = p_LnxWrpFmDev->opPorts[port_params->port_id - 1].h_Dev;
-+ break;
-+ }
-+ goto invalid_port_id;
-+
-+ default:
-+invalid_port_id:
-+ XX_Free(port_params);
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_INVALID_SELECTION, NO_MSG);
-+ }
-+
-+ ((t_FmPcdPlcrProfileParams*)param)->id.newParams.h_FmPort = h_Port;
-+ XX_Free(port_params);
-+ }
-+
-+ param->id = FM_PCD_PlcrProfileSet(p_LnxWrpFmDev->h_PcdDev, (t_FmPcdPlcrProfileParams*)param);
-+
-+ if (!param->id) {
-+ XX_Free(param);
-+ err = E_INVALID_VALUE;
-+ /* Since the LLD has no errno-style error reporting,
-+ we're left here with no other option than to report
-+ a generic E_INVALID_VALUE */
-+ break;
-+ }
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_pcd_plcr_profile_params_t *compat_param;
-+
-+ compat_param = (ioc_compat_fm_pcd_plcr_profile_params_t *) XX_Malloc(
-+ sizeof(ioc_compat_fm_pcd_plcr_profile_params_t));
-+ if (!compat_param)
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+ }
-+
-+ memset(compat_param, 0, sizeof(ioc_compat_fm_pcd_plcr_profile_params_t));
-+ compat_copy_fm_pcd_plcr_profile(compat_param, param, COMPAT_K_TO_US);
-+ if (copy_to_user((ioc_compat_fm_pcd_plcr_profile_params_t *) compat_ptr(arg),
-+ compat_param,
-+ sizeof(ioc_compat_fm_pcd_plcr_profile_params_t)))
-+ err = E_READ_FAILED;
-+
-+ XX_Free(compat_param);
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_to_user((ioc_fm_pcd_plcr_profile_params_t *)arg,
-+ param,
-+ sizeof(ioc_fm_pcd_plcr_profile_params_t)))
-+ err = E_READ_FAILED;
-+ }
-+
-+ XX_Free(param);
-+ break;
-+ }
-+
-+#if defined(CONFIG_COMPAT)
-+ case FM_PCD_IOC_PLCR_PROFILE_DELETE_COMPAT:
-+#endif
-+ case FM_PCD_IOC_PLCR_PROFILE_DELETE:
-+ {
-+ ioc_fm_obj_t id;
-+
-+ memset(&id, 0 , sizeof(ioc_fm_obj_t));
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_obj_t compat_id;
-+
-+ if (copy_from_user(&compat_id, (ioc_compat_fm_obj_t *) compat_ptr(arg), sizeof(ioc_compat_fm_obj_t)))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+
-+ compat_obj_delete(&compat_id, &id);
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_from_user(&id, (ioc_fm_obj_t *) arg, sizeof(ioc_fm_obj_t)))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+
-+ err = FM_PCD_PlcrProfileDelete(id.obj);
-+ break;
-+ }
-+
-+#if defined(CONFIG_COMPAT)
-+ case FM_PCD_IOC_CC_ROOT_MODIFY_NEXT_ENGINE_COMPAT:
-+#endif
-+ case FM_PCD_IOC_CC_ROOT_MODIFY_NEXT_ENGINE:
-+ {
-+ ioc_fm_pcd_cc_tree_modify_next_engine_params_t *param;
-+
-+ param = (ioc_fm_pcd_cc_tree_modify_next_engine_params_t *) XX_Malloc(
-+ sizeof(ioc_fm_pcd_cc_tree_modify_next_engine_params_t));
-+ if (!param)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+
-+ memset(param, 0, sizeof(ioc_fm_pcd_cc_tree_modify_next_engine_params_t));
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_pcd_cc_tree_modify_next_engine_params_t *compat_param;
-+
-+ compat_param = (ioc_compat_fm_pcd_cc_tree_modify_next_engine_params_t *) XX_Malloc(
-+ sizeof(ioc_compat_fm_pcd_cc_tree_modify_next_engine_params_t));
-+ if (!compat_param)
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+ }
-+
-+ memset(compat_param, 0, sizeof(ioc_compat_fm_pcd_cc_tree_modify_next_engine_params_t));
-+ if (copy_from_user(compat_param, (ioc_compat_fm_pcd_cc_tree_modify_next_engine_params_t *) compat_ptr(arg),
-+ sizeof(ioc_compat_fm_pcd_cc_tree_modify_next_engine_params_t)))
-+ {
-+ XX_Free(compat_param);
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+
-+ compat_fm_pcd_cc_tree_modify_next_engine(compat_param, param, COMPAT_US_TO_K);
-+
-+ XX_Free(compat_param);
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_from_user(param, (ioc_fm_pcd_cc_tree_modify_next_engine_params_t *)arg,
-+ sizeof(ioc_fm_pcd_cc_tree_modify_next_engine_params_t)))
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+ }
-+
-+ err = FM_PCD_CcRootModifyNextEngine(param->id,
-+ param->grp_indx,
-+ param->indx,
-+ (t_FmPcdCcNextEngineParams*)(&param->cc_next_engine_params));
-+
-+ XX_Free(param);
-+ break;
-+ }
-+
-+#if defined(CONFIG_COMPAT)
-+ case FM_PCD_IOC_MATCH_TABLE_MODIFY_NEXT_ENGINE_COMPAT:
-+#endif
-+ case FM_PCD_IOC_MATCH_TABLE_MODIFY_NEXT_ENGINE:
-+ {
-+ ioc_fm_pcd_cc_node_modify_next_engine_params_t *param;
-+
-+ param = (ioc_fm_pcd_cc_node_modify_next_engine_params_t *) XX_Malloc(
-+ sizeof(ioc_fm_pcd_cc_node_modify_next_engine_params_t));
-+ if (!param)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+
-+ memset(param, 0, sizeof(ioc_fm_pcd_cc_node_modify_next_engine_params_t));
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_pcd_cc_node_modify_next_engine_params_t *compat_param;
-+
-+ compat_param = (ioc_compat_fm_pcd_cc_node_modify_next_engine_params_t *) XX_Malloc(
-+ sizeof(ioc_compat_fm_pcd_cc_node_modify_next_engine_params_t));
-+ if (!compat_param)
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+ }
-+
-+ memset(compat_param, 0, sizeof(ioc_compat_fm_pcd_cc_node_modify_next_engine_params_t));
-+ if (copy_from_user(compat_param, (ioc_compat_fm_pcd_cc_node_modify_next_engine_params_t *) compat_ptr(arg),
-+ sizeof(ioc_compat_fm_pcd_cc_node_modify_next_engine_params_t)))
-+ {
-+ XX_Free(compat_param);
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+
-+ compat_copy_fm_pcd_cc_node_modify_next_engine(compat_param, param, COMPAT_US_TO_K);
-+
-+ XX_Free(compat_param);
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_from_user(param, (ioc_fm_pcd_cc_node_modify_next_engine_params_t *)arg,
-+ sizeof(ioc_fm_pcd_cc_node_modify_next_engine_params_t)))
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+ }
-+
-+ err = FM_PCD_MatchTableModifyNextEngine(param->id,
-+ param->key_indx,
-+ (t_FmPcdCcNextEngineParams*)(&param->cc_next_engine_params));
-+
-+ XX_Free(param);
-+ break;
-+ }
-+
-+#if defined(CONFIG_COMPAT)
-+ case FM_PCD_IOC_MATCH_TABLE_MODIFY_MISS_NEXT_ENGINE_COMPAT:
-+#endif
-+ case FM_PCD_IOC_MATCH_TABLE_MODIFY_MISS_NEXT_ENGINE:
-+ {
-+ ioc_fm_pcd_cc_node_modify_next_engine_params_t *param;
-+
-+ param = (ioc_fm_pcd_cc_node_modify_next_engine_params_t *) XX_Malloc(
-+ sizeof(ioc_fm_pcd_cc_node_modify_next_engine_params_t));
-+ if (!param)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+
-+ memset(param, 0, sizeof(ioc_fm_pcd_cc_node_modify_next_engine_params_t));
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_pcd_cc_node_modify_next_engine_params_t *compat_param;
-+
-+ compat_param = (ioc_compat_fm_pcd_cc_node_modify_next_engine_params_t *) XX_Malloc(
-+ sizeof(ioc_compat_fm_pcd_cc_node_modify_next_engine_params_t));
-+ if (!compat_param)
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+ }
-+
-+ memset(compat_param, 0, sizeof(ioc_compat_fm_pcd_cc_node_modify_next_engine_params_t));
-+ if (copy_from_user(compat_param, (ioc_compat_fm_pcd_cc_node_modify_next_engine_params_t *) compat_ptr(arg),
-+ sizeof(ioc_compat_fm_pcd_cc_node_modify_next_engine_params_t)))
-+ {
-+ XX_Free(compat_param);
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+
-+ compat_copy_fm_pcd_cc_node_modify_next_engine(compat_param, param, COMPAT_US_TO_K);
-+
-+ XX_Free(compat_param);
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_from_user(param, (ioc_fm_pcd_cc_node_modify_next_engine_params_t *) arg,
-+ sizeof(ioc_fm_pcd_cc_node_modify_next_engine_params_t)))
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+ }
-+
-+ err = FM_PCD_MatchTableModifyMissNextEngine(param->id,
-+ (t_FmPcdCcNextEngineParams*)(&param->cc_next_engine_params));
-+
-+ XX_Free(param);
-+ break;
-+ }
-+
-+#if defined(CONFIG_COMPAT)
-+ case FM_PCD_IOC_MATCH_TABLE_REMOVE_KEY_COMPAT:
-+#endif
-+ case FM_PCD_IOC_MATCH_TABLE_REMOVE_KEY:
-+ {
-+ ioc_fm_pcd_cc_node_remove_key_params_t *param;
-+
-+ param = (ioc_fm_pcd_cc_node_remove_key_params_t *) XX_Malloc(
-+ sizeof(ioc_fm_pcd_cc_node_remove_key_params_t));
-+ if (!param)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+
-+ memset(param, 0, sizeof(ioc_fm_pcd_cc_node_remove_key_params_t));
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_pcd_cc_node_remove_key_params_t *compat_param;
-+
-+ compat_param = (ioc_compat_fm_pcd_cc_node_remove_key_params_t *) XX_Malloc(
-+ sizeof(ioc_compat_fm_pcd_cc_node_remove_key_params_t));
-+ if (!compat_param)
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+ }
-+
-+ memset(compat_param, 0, sizeof(ioc_compat_fm_pcd_cc_node_remove_key_params_t));
-+ if (copy_from_user(compat_param,
-+ (ioc_compat_fm_pcd_cc_node_remove_key_params_t *)compat_ptr(arg),
-+ sizeof(ioc_compat_fm_pcd_cc_node_remove_key_params_t)))
-+ {
-+ XX_Free(compat_param);
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+
-+ param->id = compat_ptr(compat_param->id);
-+ param->key_indx = compat_param->key_indx;
-+
-+ XX_Free(compat_param);
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_from_user(param, (ioc_fm_pcd_cc_node_remove_key_params_t *) arg,
-+ sizeof(ioc_fm_pcd_cc_node_remove_key_params_t)))
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+ }
-+
-+ err = FM_PCD_MatchTableRemoveKey(param->id, param->key_indx);
-+
-+ XX_Free(param);
-+ break;
-+ }
-+#if defined(CONFIG_COMPAT)
-+ case FM_PCD_IOC_MATCH_TABLE_ADD_KEY_COMPAT:
-+#endif
-+ case FM_PCD_IOC_MATCH_TABLE_ADD_KEY:
-+ {
-+ ioc_fm_pcd_cc_node_modify_key_and_next_engine_params_t *param;
-+
-+ param = (ioc_fm_pcd_cc_node_modify_key_and_next_engine_params_t *) XX_Malloc(
-+ sizeof(ioc_fm_pcd_cc_node_modify_key_and_next_engine_params_t));
-+ if (!param)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+
-+ memset(param, 0, sizeof(ioc_fm_pcd_cc_node_modify_key_and_next_engine_params_t));
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_pcd_cc_node_modify_key_and_next_engine_params_t *compat_param;
-+
-+ compat_param = (ioc_compat_fm_pcd_cc_node_modify_key_and_next_engine_params_t *) XX_Malloc(
-+ sizeof(ioc_compat_fm_pcd_cc_node_modify_key_and_next_engine_params_t));
-+ if (!compat_param)
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+ }
-+
-+ memset(compat_param, 0, sizeof(ioc_compat_fm_pcd_cc_node_modify_key_and_next_engine_params_t));
-+ if (copy_from_user(compat_param,
-+ (ioc_compat_fm_pcd_cc_node_modify_key_and_next_engine_params_t *)compat_ptr(arg),
-+ sizeof(ioc_compat_fm_pcd_cc_node_modify_key_and_next_engine_params_t)))
-+ {
-+ XX_Free(compat_param);
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+
-+ compat_copy_fm_pcd_cc_node_modify_key_and_next_engine(compat_param, param, COMPAT_US_TO_K);
-+
-+ XX_Free(compat_param);
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_from_user(param, (ioc_fm_pcd_cc_node_modify_key_and_next_engine_params_t *)arg,
-+ sizeof(ioc_fm_pcd_cc_node_modify_key_and_next_engine_params_t)))
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+ }
-+
-+ if (param->key_size)
-+ {
-+ int size = 0;
-+
-+ if (param->key_params.p_key) size += param->key_size;
-+ if (param->key_params.p_mask) size += param->key_size;
-+
-+ if (size)
-+ {
-+ uint8_t *p_tmp;
-+
-+ p_tmp = (uint8_t*) XX_Malloc(size);
-+ if (!p_tmp)
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD key/mask"));
-+ }
-+
-+ if (param->key_params.p_key)
-+ {
-+ if (copy_from_user(p_tmp, param->key_params.p_key, param->key_size))
-+ {
-+ XX_Free(p_tmp);
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+
-+ param->key_params.p_key = p_tmp;
-+ }
-+
-+ if (param->key_params.p_mask)
-+ {
-+ p_tmp += param->key_size;
-+ if (copy_from_user(p_tmp, param->key_params.p_mask, param->key_size))
-+ {
-+ XX_Free(p_tmp - param->key_size);
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+
-+ param->key_params.p_mask = p_tmp;
-+ }
-+ }
-+ }
-+
-+ err = FM_PCD_MatchTableAddKey(
-+ param->id,
-+ param->key_indx,
-+ param->key_size,
-+ (t_FmPcdCcKeyParams*)&param->key_params);
-+
-+ if (param->key_params.p_key)
-+ XX_Free(param->key_params.p_key);
-+ XX_Free(param);
-+ break;
-+ }
-+
-+#if defined(CONFIG_COMPAT)
-+ case FM_PCD_IOC_MATCH_TABLE_MODIFY_KEY_AND_NEXT_ENGINE_COMPAT:
-+#endif
-+ case FM_PCD_IOC_MATCH_TABLE_MODIFY_KEY_AND_NEXT_ENGINE:
-+ {
-+ ioc_fm_pcd_cc_node_modify_key_and_next_engine_params_t *param;
-+
-+ param = (ioc_fm_pcd_cc_node_modify_key_and_next_engine_params_t *) XX_Malloc(
-+ sizeof(ioc_fm_pcd_cc_node_modify_key_and_next_engine_params_t));
-+ if (!param)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+
-+ memset(param, 0, sizeof(ioc_fm_pcd_cc_node_modify_key_and_next_engine_params_t));
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_pcd_cc_node_modify_key_and_next_engine_params_t *compat_param;
-+
-+ compat_param = (ioc_compat_fm_pcd_cc_node_modify_key_and_next_engine_params_t *) XX_Malloc(
-+ sizeof(ioc_compat_fm_pcd_cc_node_modify_key_and_next_engine_params_t));
-+ if (!compat_param)
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+ }
-+
-+ memset(compat_param, 0, sizeof(ioc_compat_fm_pcd_cc_node_modify_key_and_next_engine_params_t));
-+ if (copy_from_user(compat_param,
-+ (ioc_compat_fm_pcd_cc_node_modify_key_and_next_engine_params_t *)compat_ptr(arg),
-+ sizeof(ioc_compat_fm_pcd_cc_node_modify_key_and_next_engine_params_t)))
-+ {
-+ XX_Free(compat_param);
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+
-+ compat_copy_fm_pcd_cc_node_modify_key_and_next_engine(compat_param, param, COMPAT_US_TO_K);
-+
-+ XX_Free(compat_param);
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_from_user(param, (ioc_fm_pcd_cc_node_modify_key_and_next_engine_params_t *)arg,
-+ sizeof(ioc_fm_pcd_cc_node_modify_key_and_next_engine_params_t)))
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+ }
-+
-+ err = FM_PCD_MatchTableModifyKeyAndNextEngine(param->id,
-+ param->key_indx,
-+ param->key_size,
-+ (t_FmPcdCcKeyParams*)(&param->key_params));
-+
-+ XX_Free(param);
-+ break;
-+ }
-+
-+
-+#if defined(CONFIG_COMPAT)
-+ case FM_PCD_IOC_MATCH_TABLE_GET_KEY_STAT_COMPAT:
-+#endif
-+ case FM_PCD_IOC_MATCH_TABLE_GET_KEY_STAT:
-+ {
-+ ioc_fm_pcd_cc_tbl_get_stats_t param;
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_pcd_cc_tbl_get_stats_t *compat_param;
-+
-+ compat_param = (ioc_compat_fm_pcd_cc_tbl_get_stats_t *) XX_Malloc(
-+ sizeof(ioc_compat_fm_pcd_cc_tbl_get_stats_t));
-+ if (!compat_param)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+
-+ memset(compat_param, 0, sizeof(ioc_compat_fm_pcd_cc_tbl_get_stats_t));
-+ if (copy_from_user(compat_param,
-+ (ioc_compat_fm_pcd_cc_tbl_get_stats_t *)compat_ptr(arg),
-+ sizeof(ioc_compat_fm_pcd_cc_tbl_get_stats_t)))
-+ {
-+ XX_Free(compat_param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+
-+ compat_copy_fm_pcd_cc_tbl_get_stats(compat_param, &param, COMPAT_US_TO_K);
-+
-+ XX_Free(compat_param);
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_from_user(&param, (ioc_fm_pcd_cc_tbl_get_stats_t *)arg,
-+ sizeof(ioc_fm_pcd_cc_tbl_get_stats_t)))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+
-+
-+ err = FM_PCD_MatchTableGetKeyStatistics((t_Handle) param.id,
-+ param.key_index,
-+ (t_FmPcdCcKeyStatistics *) &param.statistics);
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_pcd_cc_tbl_get_stats_t *compat_param;
-+
-+ compat_param = (ioc_compat_fm_pcd_cc_tbl_get_stats_t*) XX_Malloc(
-+ sizeof(ioc_compat_fm_pcd_cc_tbl_get_stats_t));
-+ if (!compat_param)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+
-+ memset(compat_param, 0, sizeof(ioc_compat_fm_pcd_cc_tbl_get_stats_t));
-+ compat_copy_fm_pcd_cc_tbl_get_stats(compat_param, &param, COMPAT_K_TO_US);
-+ if (copy_to_user((ioc_compat_fm_pcd_cc_tbl_get_stats_t*) compat_ptr(arg),
-+ compat_param,
-+ sizeof(ioc_compat_fm_pcd_cc_tbl_get_stats_t))){
-+ XX_Free(compat_param);
-+ RETURN_ERROR(MINOR, E_READ_FAILED, NO_MSG);
-+ }
-+ XX_Free(compat_param);
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_to_user((ioc_fm_pcd_cc_tbl_get_stats_t *)arg,
-+ &param,
-+ sizeof(ioc_fm_pcd_cc_tbl_get_stats_t)))
-+ RETURN_ERROR(MINOR, E_READ_FAILED, NO_MSG);
-+ }
-+
-+ break;
-+ }
-+
-+
-+#if defined(CONFIG_COMPAT)
-+ case FM_PCD_IOC_MATCH_TABLE_GET_MISS_STAT_COMPAT:
-+#endif
-+ case FM_PCD_IOC_MATCH_TABLE_GET_MISS_STAT:
-+ {
-+ ioc_fm_pcd_cc_tbl_get_stats_t param;
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_pcd_cc_tbl_get_stats_t *compat_param;
-+
-+ compat_param = (ioc_compat_fm_pcd_cc_tbl_get_stats_t *) XX_Malloc(
-+ sizeof(ioc_compat_fm_pcd_cc_tbl_get_stats_t));
-+ if (!compat_param)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+
-+ memset(compat_param, 0, sizeof(ioc_compat_fm_pcd_cc_tbl_get_stats_t));
-+ if (copy_from_user(compat_param,
-+ (ioc_compat_fm_pcd_cc_tbl_get_stats_t *)compat_ptr(arg),
-+ sizeof(ioc_compat_fm_pcd_cc_tbl_get_stats_t)))
-+ {
-+ XX_Free(compat_param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+
-+ compat_copy_fm_pcd_cc_tbl_get_stats(compat_param, &param, COMPAT_US_TO_K);
-+
-+ XX_Free(compat_param);
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_from_user(&param, (ioc_fm_pcd_cc_tbl_get_stats_t *)arg,
-+ sizeof(ioc_fm_pcd_cc_tbl_get_stats_t)))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+
-+
-+ err = FM_PCD_MatchTableGetMissStatistics((t_Handle) param.id,
-+ (t_FmPcdCcKeyStatistics *) &param.statistics);
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_pcd_cc_tbl_get_stats_t *compat_param;
-+
-+ compat_param = (ioc_compat_fm_pcd_cc_tbl_get_stats_t*) XX_Malloc(
-+ sizeof(ioc_compat_fm_pcd_cc_tbl_get_stats_t));
-+ if (!compat_param)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+
-+ memset(compat_param, 0, sizeof(ioc_compat_fm_pcd_cc_tbl_get_stats_t));
-+ compat_copy_fm_pcd_cc_tbl_get_stats(compat_param, &param, COMPAT_K_TO_US);
-+ if (copy_to_user((ioc_compat_fm_pcd_cc_tbl_get_stats_t*) compat_ptr(arg),
-+ compat_param,
-+ sizeof(ioc_compat_fm_pcd_cc_tbl_get_stats_t))){
-+ XX_Free(compat_param);
-+ RETURN_ERROR(MINOR, E_READ_FAILED, NO_MSG);
-+ }
-+ XX_Free(compat_param);
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_to_user((ioc_fm_pcd_cc_tbl_get_stats_t *)arg,
-+ &param,
-+ sizeof(ioc_fm_pcd_cc_tbl_get_stats_t)))
-+ RETURN_ERROR(MINOR, E_READ_FAILED, NO_MSG);
-+ }
-+
-+ break;
-+ }
-+
-+
-+#if defined(CONFIG_COMPAT)
-+ case FM_PCD_IOC_HASH_TABLE_GET_MISS_STAT_COMPAT:
-+#endif
-+ case FM_PCD_IOC_HASH_TABLE_GET_MISS_STAT:
-+ {
-+ ioc_fm_pcd_cc_tbl_get_stats_t param;
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_pcd_cc_tbl_get_stats_t *compat_param;
-+
-+ compat_param = (ioc_compat_fm_pcd_cc_tbl_get_stats_t *) XX_Malloc(
-+ sizeof(ioc_compat_fm_pcd_cc_tbl_get_stats_t));
-+ if (!compat_param)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+
-+ memset(compat_param, 0, sizeof(ioc_compat_fm_pcd_cc_tbl_get_stats_t));
-+ if (copy_from_user(compat_param,
-+ (ioc_compat_fm_pcd_cc_tbl_get_stats_t *)compat_ptr(arg),
-+ sizeof(ioc_compat_fm_pcd_cc_tbl_get_stats_t)))
-+ {
-+ XX_Free(compat_param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+
-+ compat_copy_fm_pcd_cc_tbl_get_stats(compat_param, &param, COMPAT_US_TO_K);
-+
-+ XX_Free(compat_param);
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_from_user(&param, (ioc_fm_pcd_cc_tbl_get_stats_t *)arg,
-+ sizeof(ioc_fm_pcd_cc_tbl_get_stats_t)))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+
-+
-+ err = FM_PCD_HashTableGetMissStatistics((t_Handle) param.id,
-+ (t_FmPcdCcKeyStatistics *) &param.statistics);
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_pcd_cc_tbl_get_stats_t *compat_param;
-+
-+ compat_param = (ioc_compat_fm_pcd_cc_tbl_get_stats_t*) XX_Malloc(
-+ sizeof(ioc_compat_fm_pcd_cc_tbl_get_stats_t));
-+ if (!compat_param)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+
-+ memset(compat_param, 0, sizeof(ioc_compat_fm_pcd_cc_tbl_get_stats_t));
-+ compat_copy_fm_pcd_cc_tbl_get_stats(compat_param, &param, COMPAT_K_TO_US);
-+ if (copy_to_user((ioc_compat_fm_pcd_cc_tbl_get_stats_t*) compat_ptr(arg),
-+ compat_param,
-+ sizeof(ioc_compat_fm_pcd_cc_tbl_get_stats_t))){
-+ XX_Free(compat_param);
-+ RETURN_ERROR(MINOR, E_READ_FAILED, NO_MSG);
-+ }
-+ XX_Free(compat_param);
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_to_user((ioc_fm_pcd_cc_tbl_get_stats_t *)arg,
-+ &param,
-+ sizeof(ioc_fm_pcd_cc_tbl_get_stats_t)))
-+ RETURN_ERROR(MINOR, E_READ_FAILED, NO_MSG);
-+ }
-+
-+ break;
-+ }
-+
-+#if defined(CONFIG_COMPAT)
-+ case FM_PCD_IOC_HASH_TABLE_SET_COMPAT:
-+#endif
-+ case FM_PCD_IOC_HASH_TABLE_SET:
-+ {
-+ ioc_fm_pcd_hash_table_params_t *param;
-+
-+ param = (ioc_fm_pcd_hash_table_params_t*) XX_Malloc(
-+ sizeof(ioc_fm_pcd_hash_table_params_t));
-+ if (!param)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+
-+ memset(param, 0, sizeof(ioc_fm_pcd_hash_table_params_t));
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_pcd_hash_table_params_t *compat_param;
-+
-+ compat_param = (ioc_compat_fm_pcd_hash_table_params_t*) XX_Malloc(
-+ sizeof(ioc_compat_fm_pcd_hash_table_params_t));
-+ if (!compat_param)
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+ }
-+
-+ memset(compat_param, 0, sizeof(ioc_compat_fm_pcd_hash_table_params_t));
-+ if (copy_from_user(compat_param,
-+ (ioc_compat_fm_pcd_hash_table_params_t*)compat_ptr(arg),
-+ sizeof(ioc_compat_fm_pcd_hash_table_params_t)))
-+ {
-+ XX_Free(compat_param);
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+
-+ compat_copy_fm_pcd_hash_table(compat_param, param, COMPAT_US_TO_K);
-+
-+ XX_Free(compat_param);
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_from_user(param, (ioc_fm_pcd_hash_table_params_t *)arg,
-+ sizeof(ioc_fm_pcd_hash_table_params_t)))
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+ }
-+
-+ param->id = FM_PCD_HashTableSet(p_LnxWrpFmDev->h_PcdDev, (t_FmPcdHashTableParams *) param);
-+
-+ if (!param->id)
-+ {
-+ XX_Free(param);
-+ err = E_INVALID_VALUE;
-+ /* Since the LLD has no errno-style error reporting,
-+ we're left here with no other option than to report
-+ a generic E_INVALID_VALUE */
-+ break;
-+ }
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_pcd_hash_table_params_t *compat_param;
-+
-+ compat_param = (ioc_compat_fm_pcd_hash_table_params_t*) XX_Malloc(
-+ sizeof(ioc_compat_fm_pcd_hash_table_params_t));
-+ if (!compat_param)
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+ }
-+
-+ memset(compat_param, 0, sizeof(ioc_compat_fm_pcd_hash_table_params_t));
-+ compat_copy_fm_pcd_hash_table(compat_param, param, COMPAT_K_TO_US);
-+ if (copy_to_user((ioc_compat_fm_pcd_hash_table_params_t*) compat_ptr(arg),
-+ compat_param,
-+ sizeof(ioc_compat_fm_pcd_hash_table_params_t)))
-+ err = E_READ_FAILED;
-+
-+ XX_Free(compat_param);
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_to_user((ioc_fm_pcd_hash_table_params_t *)arg,
-+ param,
-+ sizeof(ioc_fm_pcd_hash_table_params_t)))
-+ err = E_READ_FAILED;
-+ }
-+
-+ XX_Free(param);
-+ break;
-+ }
-+
-+#if defined(CONFIG_COMPAT)
-+ case FM_PCD_IOC_HASH_TABLE_DELETE_COMPAT:
-+#endif
-+ case FM_PCD_IOC_HASH_TABLE_DELETE:
-+ {
-+ ioc_fm_obj_t id;
-+
-+ memset(&id, 0, sizeof(ioc_fm_obj_t));
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_obj_t compat_id;
-+
-+ if (copy_from_user(&compat_id, (ioc_compat_fm_obj_t *) compat_ptr(arg), sizeof(ioc_compat_fm_obj_t)))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+
-+ id.obj = compat_pcd_id2ptr(compat_id.obj);
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_from_user(&id, (ioc_fm_obj_t *) arg, sizeof(ioc_fm_obj_t)))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+
-+ err = FM_PCD_HashTableDelete(id.obj);
-+ break;
-+ }
-+
-+#if defined(CONFIG_COMPAT)
-+ case FM_PCD_IOC_HASH_TABLE_ADD_KEY_COMPAT:
-+#endif
-+ case FM_PCD_IOC_HASH_TABLE_ADD_KEY:
-+ {
-+ ioc_fm_pcd_hash_table_add_key_params_t *param = NULL;
-+
-+ param = (ioc_fm_pcd_hash_table_add_key_params_t*) XX_Malloc(
-+ sizeof(ioc_fm_pcd_hash_table_add_key_params_t));
-+ if (!param)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+
-+ memset(param, 0, sizeof(ioc_fm_pcd_hash_table_add_key_params_t));
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_pcd_hash_table_add_key_params_t *compat_param;
-+
-+ compat_param = (ioc_compat_fm_pcd_hash_table_add_key_params_t*) XX_Malloc(
-+ sizeof(ioc_compat_fm_pcd_hash_table_add_key_params_t));
-+ if (!compat_param)
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+ }
-+
-+ memset(compat_param, 0, sizeof(ioc_compat_fm_pcd_hash_table_add_key_params_t));
-+ if (copy_from_user(compat_param,
-+ (ioc_compat_fm_pcd_hash_table_add_key_params_t*) compat_ptr(arg),
-+ sizeof(ioc_compat_fm_pcd_hash_table_add_key_params_t)))
-+ {
-+ XX_Free(compat_param);
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+
-+ if (compat_param->key_size)
-+ {
-+ param->p_hash_tbl = compat_pcd_id2ptr(compat_param->p_hash_tbl);
-+ param->key_size = compat_param->key_size;
-+
-+ compat_copy_fm_pcd_cc_key(&compat_param->key_params, &param->key_params, COMPAT_US_TO_K);
-+ }
-+ else
-+ {
-+ XX_Free(compat_param);
-+ XX_Free(param);
-+ err = E_INVALID_VALUE;
-+ break;
-+ }
-+
-+ XX_Free(compat_param);
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_from_user(param, (ioc_fm_pcd_hash_table_add_key_params_t*) arg,
-+ sizeof(ioc_fm_pcd_hash_table_add_key_params_t)))
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+ }
-+
-+ if (param->key_size)
-+ {
-+ int size = 0;
-+
-+ if (param->key_params.p_key) size += param->key_size;
-+ if (param->key_params.p_mask) size += param->key_size;
-+
-+ if (size)
-+ {
-+ uint8_t *p_tmp;
-+
-+ p_tmp = (uint8_t*) XX_Malloc(size);
-+ if (!p_tmp)
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD key/mask"));
-+ }
-+
-+ if (param->key_params.p_key)
-+ {
-+ if (copy_from_user(p_tmp, param->key_params.p_key, param->key_size))
-+ {
-+ XX_Free(p_tmp);
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+
-+ param->key_params.p_key = p_tmp;
-+ }
-+
-+ if (param->key_params.p_mask)
-+ {
-+ p_tmp += param->key_size;
-+ if (copy_from_user(p_tmp, param->key_params.p_mask, param->key_size))
-+ {
-+ XX_Free(p_tmp - param->key_size);
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+
-+ param->key_params.p_mask = p_tmp;
-+ }
-+ }
-+ }
-+
-+ err = FM_PCD_HashTableAddKey(
-+ param->p_hash_tbl,
-+ param->key_size,
-+ (t_FmPcdCcKeyParams*)&param->key_params);
-+
-+ if (param->key_params.p_key)
-+ XX_Free(param->key_params.p_key);
-+ XX_Free(param);
-+ break;
-+ }
-+
-+#if defined(CONFIG_COMPAT)
-+ case FM_PCD_IOC_HASH_TABLE_REMOVE_KEY_COMPAT:
-+#endif
-+ case FM_PCD_IOC_HASH_TABLE_REMOVE_KEY:
-+ {
-+ ioc_fm_pcd_hash_table_remove_key_params_t *param = NULL;
-+
-+ param = (ioc_fm_pcd_hash_table_remove_key_params_t*) XX_Malloc(
-+ sizeof(ioc_fm_pcd_hash_table_remove_key_params_t));
-+ if (!param)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+
-+ memset(param, 0, sizeof(ioc_fm_pcd_hash_table_remove_key_params_t));
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_pcd_hash_table_remove_key_params_t *compat_param;
-+
-+ compat_param = (ioc_compat_fm_pcd_hash_table_remove_key_params_t*) XX_Malloc(
-+ sizeof(ioc_compat_fm_pcd_hash_table_remove_key_params_t));
-+ if (!compat_param)
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+ }
-+
-+ memset(compat_param, 0, sizeof(ioc_compat_fm_pcd_hash_table_remove_key_params_t));
-+ if (copy_from_user(compat_param,
-+ (ioc_compat_fm_pcd_hash_table_remove_key_params_t*) compat_ptr(arg),
-+ sizeof(ioc_compat_fm_pcd_hash_table_remove_key_params_t)))
-+ {
-+ XX_Free(compat_param);
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+
-+ param->p_hash_tbl = compat_pcd_id2ptr(compat_param->p_hash_tbl);
-+ param->key_size = compat_param->key_size;
-+
-+ XX_Free(compat_param);
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_from_user(param, (ioc_fm_pcd_hash_table_remove_key_params_t*)arg,
-+ sizeof(ioc_fm_pcd_hash_table_remove_key_params_t)))
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+ }
-+
-+ if (param->key_size)
-+ {
-+ uint8_t *p_key;
-+
-+ p_key = (uint8_t*) XX_Malloc(param->key_size);
-+ if (!p_key)
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+ }
-+
-+ if (param->p_key && copy_from_user(p_key, param->p_key, param->key_size))
-+ {
-+ XX_Free(p_key);
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+ param->p_key = p_key;
-+ }
-+
-+ err = FM_PCD_HashTableRemoveKey(
-+ param->p_hash_tbl,
-+ param->key_size,
-+ param->p_key);
-+
-+ if (param->p_key)
-+ XX_Free(param->p_key);
-+ XX_Free(param);
-+ break;
-+ }
-+
-+#if defined(CONFIG_COMPAT)
-+ case FM_PCD_IOC_MATCH_TABLE_MODIFY_KEY_COMPAT:
-+#endif
-+ case FM_PCD_IOC_MATCH_TABLE_MODIFY_KEY:
-+ {
-+ ioc_fm_pcd_cc_node_modify_key_params_t *param;
-+
-+ param = (ioc_fm_pcd_cc_node_modify_key_params_t *) XX_Malloc(
-+ sizeof(ioc_fm_pcd_cc_node_modify_key_params_t));
-+ if (!param)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+
-+ memset(param, 0, sizeof(ioc_fm_pcd_cc_node_modify_key_params_t));
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_pcd_cc_node_modify_key_params_t *compat_param;
-+
-+ compat_param = (ioc_compat_fm_pcd_cc_node_modify_key_params_t *) XX_Malloc(
-+ sizeof(ioc_compat_fm_pcd_cc_node_modify_key_params_t));
-+ if (!compat_param)
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+ }
-+
-+ memset(compat_param, 0, sizeof(ioc_compat_fm_pcd_cc_node_modify_key_params_t));
-+ if (copy_from_user(compat_param, (ioc_compat_fm_pcd_cc_node_modify_key_params_t *)compat_ptr(arg),
-+ sizeof(ioc_compat_fm_pcd_cc_node_modify_key_params_t)))
-+ {
-+ XX_Free(compat_param);
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+
-+ compat_copy_fm_pcd_cc_node_modify_key(compat_param, param, COMPAT_US_TO_K);
-+
-+ XX_Free(compat_param);
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_from_user(param, (ioc_fm_pcd_cc_node_modify_key_params_t *)arg,
-+ sizeof(ioc_fm_pcd_cc_node_modify_key_params_t)))
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+ }
-+
-+ if (param->key_size)
-+ {
-+ int size = 0;
-+
-+ if (param->p_key) size += param->key_size;
-+ if (param->p_mask) size += param->key_size;
-+
-+ if (size)
-+ {
-+ uint8_t *p_tmp;
-+
-+ p_tmp = (uint8_t*) XX_Malloc(size);
-+ if (!p_tmp)
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD key/mask"));
-+ }
-+
-+ if (param->p_key)
-+ {
-+ if (copy_from_user(p_tmp, param->p_key, param->key_size))
-+ {
-+ XX_Free(p_tmp);
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+
-+ param->p_key = p_tmp;
-+ }
-+
-+ if (param->p_mask)
-+ {
-+ p_tmp += param->key_size;
-+ if (copy_from_user(p_tmp, param->p_mask, param->key_size))
-+ {
-+ XX_Free(p_tmp - param->key_size);
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+
-+ param->p_mask = p_tmp;
-+ }
-+ }
-+ }
-+
-+ err = FM_PCD_MatchTableModifyKey(param->id,
-+ param->key_indx,
-+ param->key_size,
-+ param->p_key,
-+ param->p_mask);
-+
-+ if (param->p_key)
-+ XX_Free(param->p_key);
-+ else if (param->p_mask)
-+ XX_Free(param->p_mask);
-+ XX_Free(param);
-+ break;
-+ }
-+
-+#if defined(CONFIG_COMPAT)
-+ case FM_PCD_IOC_MANIP_NODE_SET_COMPAT:
-+#endif
-+ case FM_PCD_IOC_MANIP_NODE_SET:
-+ {
-+ ioc_fm_pcd_manip_params_t *param;
-+ uint8_t *p_data = NULL;
-+ uint8_t size;
-+
-+ param = (ioc_fm_pcd_manip_params_t *) XX_Malloc(
-+ sizeof(ioc_fm_pcd_manip_params_t));
-+
-+ if (!param)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+
-+ memset(param, 0, sizeof(ioc_fm_pcd_manip_params_t));
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_pcd_manip_params_t *compat_param;
-+
-+ compat_param = (ioc_compat_fm_pcd_manip_params_t *) XX_Malloc(
-+ sizeof(ioc_compat_fm_pcd_manip_params_t));
-+ if (!compat_param)
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+ }
-+
-+ memset(compat_param, 0, sizeof(ioc_compat_fm_pcd_manip_params_t));
-+ if (copy_from_user(compat_param,
-+ (ioc_compat_fm_pcd_manip_params_t *) compat_ptr(arg),
-+ sizeof(ioc_compat_fm_pcd_manip_params_t)))
-+ {
-+ XX_Free(compat_param);
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+
-+ compat_fm_pcd_manip_set_node(compat_param, param, COMPAT_US_TO_K);
-+
-+ XX_Free(compat_param);
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_from_user(param, (ioc_fm_pcd_manip_params_t *)arg,
-+ sizeof(ioc_fm_pcd_manip_params_t)))
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+ }
-+
-+ if (param->type == e_IOC_FM_PCD_MANIP_HDR)
-+ {
-+ size = param->u.hdr.insrt_params.u.generic.size;
-+ p_data = (uint8_t *) XX_Malloc(size);
-+ if (!p_data )
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, NO_MSG);
-+ }
-+
-+ if (param->u.hdr.insrt_params.u.generic.p_data &&
-+ copy_from_user(p_data,
-+ param->u.hdr.insrt_params.u.generic.p_data, size))
-+ {
-+ XX_Free(p_data);
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+
-+ param->u.hdr.insrt_params.u.generic.p_data = p_data;
-+ }
-+
-+ if (param->id)
-+ {
-+ /* Security Hole: the user can pass any piece of garbage
-+ in 'param->id', and that will go straight through to the LLD,
-+ no checks being done by the wrapper! */
-+ err = FM_PCD_ManipNodeReplace(
-+ (t_Handle) param->id,
-+ (t_FmPcdManipParams*) param);
-+ if (err)
-+ {
-+ if (p_data)
-+ XX_Free(p_data);
-+ XX_Free(param);
-+ break;
-+ }
-+ }
-+ else
-+ {
-+ param->id = FM_PCD_ManipNodeSet(
-+ p_LnxWrpFmDev->h_PcdDev,
-+ (t_FmPcdManipParams*) param);
-+ if (!param->id)
-+ {
-+ if (p_data)
-+ XX_Free(p_data);
-+ XX_Free(param);
-+ err = E_INVALID_VALUE;
-+ /* Since the LLD has no errno-style error reporting,
-+ we're left here with no other option than to report
-+ a generic E_INVALID_VALUE */
-+ break;
-+ }
-+ }
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_pcd_manip_params_t *compat_param;
-+
-+ compat_param = (ioc_compat_fm_pcd_manip_params_t *) XX_Malloc(
-+ sizeof(ioc_compat_fm_pcd_manip_params_t));
-+ if (!compat_param)
-+ {
-+ if (p_data)
-+ XX_Free(p_data);
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+ }
-+
-+ memset(compat_param, 0, sizeof(ioc_compat_fm_pcd_manip_params_t));
-+
-+ compat_fm_pcd_manip_set_node(compat_param, param, COMPAT_K_TO_US);
-+
-+ if (copy_to_user((ioc_compat_fm_pcd_manip_params_t *) compat_ptr(arg),
-+ compat_param,
-+ sizeof(ioc_compat_fm_pcd_manip_params_t)))
-+ err = E_READ_FAILED;
-+
-+ XX_Free(compat_param);
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_to_user((ioc_fm_pcd_manip_params_t *)arg,
-+ param, sizeof(ioc_fm_pcd_manip_params_t)))
-+ err = E_READ_FAILED;
-+ }
-+
-+ if (p_data)
-+ XX_Free(p_data);
-+ XX_Free(param);
-+ break;
-+ }
-+
-+#if defined(CONFIG_COMPAT)
-+ case FM_PCD_IOC_MANIP_NODE_DELETE_COMPAT:
-+#endif
-+ case FM_PCD_IOC_MANIP_NODE_DELETE:
-+ {
-+ ioc_fm_obj_t id;
-+
-+ memset(&id, 0, sizeof(ioc_fm_obj_t));
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_obj_t compat_id;
-+
-+ if (copy_from_user(&compat_id, (ioc_compat_fm_obj_t *) compat_ptr(arg), sizeof(ioc_compat_fm_obj_t)))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+
-+ compat_obj_delete(&compat_id, &id);
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_from_user(&id, (ioc_fm_obj_t *) arg, sizeof(ioc_fm_obj_t)))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+
-+ err = FM_PCD_ManipNodeDelete(id.obj);
-+ break;
-+ }
-+
-+#if defined(CONFIG_COMPAT)
-+ case FM_PCD_IOC_MANIP_GET_STATS_COMPAT:
-+#endif
-+ case FM_PCD_IOC_MANIP_GET_STATS:
-+ {
-+ ioc_fm_pcd_manip_get_stats_t param;
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_pcd_manip_get_stats_t *compat_param;
-+
-+ compat_param = (ioc_compat_fm_pcd_manip_get_stats_t *) XX_Malloc(
-+ sizeof(ioc_compat_fm_pcd_manip_get_stats_t));
-+ if (!compat_param)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+
-+ memset(compat_param, 0, sizeof(ioc_compat_fm_pcd_manip_get_stats_t));
-+ if (copy_from_user(compat_param,
-+ (ioc_compat_fm_pcd_manip_get_stats_t *)compat_ptr(arg),
-+ sizeof(ioc_compat_fm_pcd_manip_get_stats_t)))
-+ {
-+ XX_Free(compat_param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+
-+ compat_copy_fm_pcd_manip_get_stats(compat_param, &param, COMPAT_US_TO_K);
-+
-+ XX_Free(compat_param);
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_from_user(&param, (ioc_fm_pcd_manip_get_stats_t *)arg,
-+ sizeof(ioc_fm_pcd_manip_get_stats_t)))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+
-+ err = FM_PCD_ManipGetStatistics((t_Handle) param.id,
-+ (t_FmPcdManipStats*) &param.stats);
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_pcd_manip_get_stats_t *compat_param;
-+
-+ compat_param = (ioc_compat_fm_pcd_manip_get_stats_t*) XX_Malloc(
-+ sizeof(ioc_compat_fm_pcd_manip_get_stats_t));
-+ if (!compat_param)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+
-+ memset(compat_param, 0, sizeof(ioc_compat_fm_pcd_manip_get_stats_t));
-+ compat_copy_fm_pcd_manip_get_stats(compat_param, &param, COMPAT_K_TO_US);
-+ if (copy_to_user((ioc_compat_fm_pcd_manip_get_stats_t*) compat_ptr(arg),
-+ compat_param,
-+ sizeof(ioc_compat_fm_pcd_manip_get_stats_t))){
-+ XX_Free(compat_param);
-+ RETURN_ERROR(MINOR, E_READ_FAILED, NO_MSG);
-+ }
-+ XX_Free(compat_param);
-+ }
-+ else
-+#endif
-+ if (copy_to_user((ioc_fm_pcd_manip_get_stats_t *)arg,
-+ &param,
-+ sizeof(ioc_fm_pcd_manip_get_stats_t)))
-+ RETURN_ERROR(MINOR, E_READ_FAILED, NO_MSG);
-+
-+ break;
-+ }
-+
-+#if (DPAA_VERSION >= 11)
-+#if defined(CONFIG_COMPAT)
-+ case FM_PCD_IOC_FRM_REPLIC_GROUP_SET_COMPAT:
-+#endif
-+ case FM_PCD_IOC_FRM_REPLIC_GROUP_SET:
-+ {
-+ ioc_fm_pcd_frm_replic_group_params_t *param;
-+
-+ param = (ioc_fm_pcd_frm_replic_group_params_t *) XX_Malloc(
-+ sizeof(ioc_fm_pcd_frm_replic_group_params_t));
-+ if (!param)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+
-+ memset(param, 0, sizeof(ioc_fm_pcd_frm_replic_group_params_t));
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_pcd_frm_replic_group_params_t
-+ *compat_param;
-+
-+ compat_param =
-+ (ioc_compat_fm_pcd_frm_replic_group_params_t *)
-+ XX_Malloc(sizeof(ioc_compat_fm_pcd_frm_replic_group_params_t));
-+ if (!compat_param)
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_NO_MEMORY,
-+ ("IOCTL FM PCD"));
-+ }
-+
-+ memset(compat_param, 0, sizeof(ioc_compat_fm_pcd_frm_replic_group_params_t));
-+ if (copy_from_user(compat_param,
-+ (ioc_compat_fm_pcd_frm_replic_group_params_t *)
-+ compat_ptr(arg),
-+ sizeof(ioc_compat_fm_pcd_frm_replic_group_params_t))) {
-+ XX_Free(compat_param);
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_READ_FAILED, NO_MSG);
-+ }
-+
-+ compat_copy_fm_pcd_frm_replic_group_params(compat_param,
-+ param, COMPAT_US_TO_K);
-+
-+ XX_Free(compat_param);
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_from_user(param,
-+ (ioc_fm_pcd_frm_replic_group_params_t *)arg,
-+ sizeof(ioc_fm_pcd_frm_replic_group_params_t)))
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_READ_FAILED, NO_MSG);
-+ }
-+ }
-+
-+ param->id = FM_PCD_FrmReplicSetGroup(p_LnxWrpFmDev->h_PcdDev,
-+ (t_FmPcdFrmReplicGroupParams*)param);
-+
-+ if (!param->id) {
-+ XX_Free(param);
-+ err = E_INVALID_VALUE;
-+ /*
-+ * Since the LLD has no errno-style error reporting,
-+ * we're left here with no other option than to report
-+ * a generic E_INVALID_VALUE
-+ */
-+ break;
-+ }
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_pcd_frm_replic_group_params_t
-+ *compat_param;
-+
-+ compat_param =
-+ (ioc_compat_fm_pcd_frm_replic_group_params_t *)
-+ XX_Malloc(sizeof(ioc_compat_fm_pcd_frm_replic_group_params_t));
-+ if (!compat_param)
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_NO_MEMORY,
-+ ("IOCTL FM PCD"));
-+ }
-+
-+ memset(compat_param, 0, sizeof(ioc_compat_fm_pcd_frm_replic_group_params_t));
-+ compat_copy_fm_pcd_frm_replic_group_params(compat_param,
-+ param, COMPAT_K_TO_US);
-+ if (copy_to_user(
-+ (ioc_compat_fm_pcd_frm_replic_group_params_t *)
-+ compat_ptr(arg),
-+ compat_param,
-+ sizeof(ioc_compat_fm_pcd_frm_replic_group_params_t)))
-+ err = E_WRITE_FAILED;
-+
-+ XX_Free(compat_param);
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_to_user(
-+ (ioc_fm_pcd_frm_replic_group_params_t *)arg,
-+ param,
-+ sizeof(ioc_fm_pcd_frm_replic_group_params_t)))
-+ err = E_WRITE_FAILED;
-+ }
-+
-+ XX_Free(param);
-+ break;
-+ }
-+ break;
-+
-+#if defined(CONFIG_COMPAT)
-+ case FM_PCD_IOC_FRM_REPLIC_GROUP_DELETE_COMPAT:
-+#endif
-+ case FM_PCD_IOC_FRM_REPLIC_GROUP_DELETE:
-+ {
-+ ioc_fm_obj_t id;
-+
-+ memset(&id, 0, sizeof(ioc_fm_obj_t));
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_obj_t compat_id;
-+
-+ if (copy_from_user(&compat_id,
-+ (ioc_compat_fm_obj_t *) compat_ptr(arg),
-+ sizeof(ioc_compat_fm_obj_t)))
-+ break;
-+ compat_obj_delete(&compat_id, &id);
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_from_user(&id, (ioc_fm_obj_t *) arg,
-+ sizeof(ioc_fm_obj_t)))
-+ break;
-+ }
-+
-+ return FM_PCD_FrmReplicDeleteGroup(id.obj);
-+ }
-+ break;
-+
-+#if defined(CONFIG_COMPAT)
-+ case FM_PCD_IOC_FRM_REPLIC_MEMBER_ADD_COMPAT:
-+#endif
-+ case FM_PCD_IOC_FRM_REPLIC_MEMBER_ADD:
-+ {
-+ ioc_fm_pcd_frm_replic_member_params_t param;
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_pcd_frm_replic_member_params_t compat_param;
-+
-+ if (copy_from_user(&compat_param, compat_ptr(arg), sizeof(compat_param)))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+
-+ compat_copy_fm_pcd_frm_replic_member_params(&compat_param, &param, COMPAT_US_TO_K);
-+ }
-+ else
-+#endif
-+ if (copy_from_user(&param, (void *)arg, sizeof(param)))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+
-+ return FM_PCD_FrmReplicAddMember(param.member.h_replic_group,
-+ param.member.member_index,
-+ (t_FmPcdCcNextEngineParams*)&param.next_engine_params);
-+ }
-+ break;
-+
-+#if defined(CONFIG_COMPAT)
-+ case FM_PCD_IOC_FRM_REPLIC_MEMBER_REMOVE_COMPAT:
-+#endif
-+ case FM_PCD_IOC_FRM_REPLIC_MEMBER_REMOVE:
-+ {
-+ ioc_fm_pcd_frm_replic_member_t param;
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_pcd_frm_replic_member_t compat_param;
-+
-+ if (copy_from_user(&compat_param, compat_ptr(arg), sizeof(compat_param)))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+
-+ compat_copy_fm_pcd_frm_replic_member(&compat_param, &param, COMPAT_US_TO_K);
-+ }
-+ else
-+#endif
-+ if (copy_from_user(&param, (void *)arg, sizeof(param)))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+
-+ return FM_PCD_FrmReplicRemoveMember(param.h_replic_group, param.member_index);
-+ }
-+ break;
-+
-+#if defined(CONFIG_COMPAT)
-+ case FM_IOC_VSP_CONFIG_COMPAT:
-+#endif
-+ case FM_IOC_VSP_CONFIG:
-+ {
-+ ioc_fm_vsp_params_t param;
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_vsp_params_t compat_param;
-+
-+ if (copy_from_user(&compat_param, compat_ptr(arg), sizeof(compat_param)))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+
-+ compat_copy_fm_vsp_params(&compat_param, &param, COMPAT_US_TO_K);
-+ }
-+ else
-+#endif
-+ if (copy_from_user(&param, (void *)arg, sizeof(param)))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ {
-+ uint8_t portId = param.port_params.port_id;
-+ param.liodn_offset =
-+ p_LnxWrpFmDev->rxPorts[portId].settings.param.specificParams.rxParams.liodnOffset;
-+ }
-+ param.p_fm = p_LnxWrpFmDev->h_Dev;
-+ param.id = FM_VSP_Config((t_FmVspParams *)&param);
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_vsp_params_t compat_param;
-+
-+ memset(&compat_param, 0, sizeof(compat_param));
-+ compat_copy_fm_vsp_params(&compat_param, &param, COMPAT_K_TO_US);
-+
-+ if (copy_to_user(compat_ptr(arg), &compat_param, sizeof(compat_param)))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+ else
-+#endif
-+ if (copy_to_user((void *)arg, &param, sizeof(param)))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ break;
-+ }
-+
-+#if defined(CONFIG_COMPAT)
-+ case FM_IOC_VSP_INIT_COMPAT:
-+#endif
-+ case FM_IOC_VSP_INIT:
-+ {
-+ ioc_fm_obj_t id;
-+
-+ memset(&id, 0, sizeof(ioc_fm_obj_t));
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_obj_t compat_id;
-+
-+ if (copy_from_user(&compat_id,
-+ (ioc_compat_fm_obj_t *) compat_ptr(arg),
-+ sizeof(ioc_compat_fm_obj_t)))
-+ break;
-+ id.obj = compat_pcd_id2ptr(compat_id.obj);
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_from_user(&id, (ioc_fm_obj_t *) arg,
-+ sizeof(ioc_fm_obj_t)))
-+ break;
-+ }
-+
-+ return FM_VSP_Init(id.obj);
-+ }
-+
-+#if defined(CONFIG_COMPAT)
-+ case FM_IOC_VSP_FREE_COMPAT:
-+#endif
-+ case FM_IOC_VSP_FREE:
-+ {
-+ ioc_fm_obj_t id;
-+
-+ memset(&id, 0, sizeof(ioc_fm_obj_t));
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_obj_t compat_id;
-+
-+ if (copy_from_user(&compat_id,
-+ (ioc_compat_fm_obj_t *) compat_ptr(arg),
-+ sizeof(ioc_compat_fm_obj_t)))
-+ break;
-+ compat_obj_delete(&compat_id, &id);
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_from_user(&id, (ioc_fm_obj_t *) arg,
-+ sizeof(ioc_fm_obj_t)))
-+ break;
-+ }
-+
-+ return FM_VSP_Free(id.obj);
-+ }
-+
-+#if defined(CONFIG_COMPAT)
-+ case FM_IOC_VSP_CONFIG_POOL_DEPLETION_COMPAT:
-+#endif
-+ case FM_IOC_VSP_CONFIG_POOL_DEPLETION:
-+ {
-+ ioc_fm_buf_pool_depletion_params_t param;
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_buf_pool_depletion_params_t compat_param;
-+
-+ if (copy_from_user(&compat_param, compat_ptr(arg), sizeof(compat_param)))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+
-+ compat_copy_fm_buf_pool_depletion_params(&compat_param, &param, COMPAT_US_TO_K);
-+ }
-+ else
-+#endif
-+ if (copy_from_user(&param, (void *)arg, sizeof(param)))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+
-+ if (FM_VSP_ConfigPoolDepletion(param.p_fm_vsp,
-+ (t_FmBufPoolDepletion *)&param.fm_buf_pool_depletion))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+
-+ break;
-+ }
-+
-+
-+#if defined(CONFIG_COMPAT)
-+ case FM_IOC_VSP_CONFIG_BUFFER_PREFIX_CONTENT_COMPAT:
-+#endif
-+ case FM_IOC_VSP_CONFIG_BUFFER_PREFIX_CONTENT:
-+ {
-+ ioc_fm_buffer_prefix_content_params_t param;
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_buffer_prefix_content_params_t compat_param;
-+
-+ if (copy_from_user(&compat_param, compat_ptr(arg), sizeof(compat_param)))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+
-+ compat_copy_fm_buffer_prefix_content_params(&compat_param, &param, COMPAT_US_TO_K);
-+ }
-+ else
-+#endif
-+ if (copy_from_user(&param, (void *)arg, sizeof(param)))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+
-+ if (FM_VSP_ConfigBufferPrefixContent(param.p_fm_vsp,
-+ (t_FmBufferPrefixContent *)&param.fm_buffer_prefix_content))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+
-+ break;
-+ }
-+
-+#if defined(CONFIG_COMPAT)
-+ case FM_IOC_VSP_CONFIG_NO_SG_COMPAT:
-+#endif
-+ case FM_IOC_VSP_CONFIG_NO_SG:
-+ {
-+ ioc_fm_vsp_config_no_sg_params_t param;
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_vsp_config_no_sg_params_t compat_param;
-+
-+ if (copy_from_user(&compat_param, compat_ptr(arg), sizeof(compat_param)))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+
-+ compat_copy_fm_vsp_config_no_sg_params(&compat_param, &param, COMPAT_US_TO_K);
-+ }
-+ else
-+#endif
-+ if (copy_from_user(&param, (void *)arg, sizeof(param)))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+
-+ if (FM_VSP_ConfigNoScatherGather(param.p_fm_vsp, param.no_sg))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+
-+ break;
-+ }
-+
-+#if defined(CONFIG_COMPAT)
-+ case FM_IOC_VSP_GET_BUFFER_PRS_RESULT_COMPAT:
-+#endif
-+ case FM_IOC_VSP_GET_BUFFER_PRS_RESULT:
-+ {
-+ ioc_fm_vsp_prs_result_params_t param;
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_vsp_prs_result_params_t compat_param;
-+
-+ if (copy_from_user(&compat_param, compat_ptr(arg), sizeof(compat_param)))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+
-+ compat_copy_fm_vsp_prs_result_params(&compat_param, &param, COMPAT_US_TO_K);
-+ }
-+ else
-+#endif
-+ if (copy_from_user(&param, (void *)arg, sizeof(param)))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+
-+ /* this call just adds the parse results offset to p_data */
-+ param.p_data = FM_VSP_GetBufferPrsResult(param.p_fm_vsp, param.p_data);
-+
-+ if (!param.p_data)
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_vsp_prs_result_params_t compat_param;
-+
-+ memset(&compat_param, 0, sizeof(compat_param));
-+ compat_copy_fm_vsp_prs_result_params(&compat_param, &param, COMPAT_K_TO_US);
-+
-+ if (copy_to_user(compat_ptr(arg), &compat_param, sizeof(compat_param)))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+ else
-+#endif
-+ if (copy_to_user((void *)arg, &param, sizeof(param)))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+
-+ break;
-+ }
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+#ifdef FM_CAPWAP_SUPPORT
-+#warning "feature not supported!"
-+#if defined(CONFIG_COMPAT)
-+ case FM_PCD_IOC_STATISTICS_SET_NODE_COMPAT:
-+#endif
-+ case FM_PCD_IOC_STATISTICS_SET_NODE:
-+ {
-+/* ioc_fm_pcd_stats_params_t param;
-+ ...
-+ param->id = FM_PCD_StatisticsSetNode(p_LnxWrpFmDev->h_PcdDev,
-+ (t_FmPcdStatsParams *)&param);
-+*/
-+ err = E_NOT_SUPPORTED;
-+ break;
-+ }
-+#endif /* FM_CAPWAP_SUPPORT */
-+
-+ default:
-+ RETURN_ERROR(MINOR, E_INVALID_SELECTION,
-+ ("invalid ioctl: cmd:0x%08x(type:0x%02x, nr: %d.\n",
-+ cmd, _IOC_TYPE(cmd), _IOC_NR(cmd)));
-+ }
-+
-+ if (err)
-+ RETURN_ERROR(MINOR, err, ("IOCTL FM PCD"));
-+
-+ return E_OK;
-+}
-+
-+void FM_Get_Api_Version(ioc_fm_api_version_t *p_version)
-+{
-+ p_version->version.major = FMD_API_VERSION_MAJOR;
-+ p_version->version.minor = FMD_API_VERSION_MINOR;
-+ p_version->version.respin = FMD_API_VERSION_RESPIN;
-+ p_version->version.reserved = 0;
-+}
-+
-+t_Error LnxwrpFmIOCTL(t_LnxWrpFmDev *p_LnxWrpFmDev, unsigned int cmd, unsigned long arg, bool compat)
-+{
-+ t_Error err = E_OK;
-+
-+ switch (cmd)
-+ {
-+ case FM_IOC_SET_PORTS_BANDWIDTH:
-+ {
-+ ioc_fm_port_bandwidth_params *param;
-+
-+ param = (ioc_fm_port_bandwidth_params*) XX_Malloc(sizeof(ioc_fm_port_bandwidth_params));
-+ if (!param)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+
-+ memset(param, 0, sizeof(ioc_fm_port_bandwidth_params));
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ if (copy_from_user(param, (ioc_fm_port_bandwidth_params*)compat_ptr(arg), sizeof(ioc_fm_port_bandwidth_params)))
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_from_user(param, (ioc_fm_port_bandwidth_params*)arg, sizeof(ioc_fm_port_bandwidth_params)))
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+ }
-+
-+ err = FM_SetPortsBandwidth(p_LnxWrpFmDev->h_Dev, (t_FmPortsBandwidthParams*) param);
-+
-+ XX_Free(param);
-+ break;
-+ }
-+
-+ case FM_IOC_GET_REVISION:
-+ {
-+ ioc_fm_revision_info_t *param;
-+
-+ param = (ioc_fm_revision_info_t *) XX_Malloc(sizeof(ioc_fm_revision_info_t));
-+ if (!param)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+
-+ FM_GetRevision(p_LnxWrpFmDev->h_Dev, (t_FmRevisionInfo*)param);
-+ /* This one never returns anything other than E_OK */
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ if (copy_to_user((ioc_fm_revision_info_t *)compat_ptr(arg),
-+ param,
-+ sizeof(ioc_fm_revision_info_t))){
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_READ_FAILED, NO_MSG);
-+ }
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_to_user((ioc_fm_revision_info_t *)arg,
-+ param,
-+ sizeof(ioc_fm_revision_info_t))){
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_READ_FAILED, NO_MSG);
-+ }
-+ }
-+ XX_Free(param);
-+ break;
-+ }
-+
-+ case FM_IOC_SET_COUNTER:
-+ {
-+ ioc_fm_counters_params_t *param;
-+
-+ param = (ioc_fm_counters_params_t *) XX_Malloc(sizeof(ioc_fm_counters_params_t));
-+ if (!param)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+
-+ memset(param, 0, sizeof(ioc_fm_counters_params_t));
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ if (copy_from_user(param, (ioc_fm_counters_params_t *)compat_ptr(arg), sizeof(ioc_fm_counters_params_t)))
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_from_user(param, (ioc_fm_counters_params_t *)arg, sizeof(ioc_fm_counters_params_t)))
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+ }
-+
-+ err = FM_ModifyCounter(p_LnxWrpFmDev->h_Dev, param->cnt, param->val);
-+
-+ XX_Free(param);
-+ break;
-+ }
-+
-+ case FM_IOC_GET_COUNTER:
-+ {
-+ ioc_fm_counters_params_t *param;
-+
-+ param = (ioc_fm_counters_params_t *) XX_Malloc(sizeof(ioc_fm_counters_params_t));
-+ if (!param)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PCD"));
-+
-+ memset(param, 0, sizeof(ioc_fm_counters_params_t));
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ if (copy_from_user(param, (ioc_fm_counters_params_t *)compat_ptr(arg), sizeof(ioc_fm_counters_params_t)))
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_from_user(param, (ioc_fm_counters_params_t *)arg, sizeof(ioc_fm_counters_params_t)))
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+ }
-+
-+ param->val = FM_GetCounter(p_LnxWrpFmDev->h_Dev, param->cnt);
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ if (copy_to_user((ioc_fm_counters_params_t *)compat_ptr(arg), param, sizeof(ioc_fm_counters_params_t)))
-+ err = E_READ_FAILED;
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_to_user((ioc_fm_counters_params_t *)arg, param, sizeof(ioc_fm_counters_params_t)))
-+ err = E_READ_FAILED;
-+ }
-+
-+ XX_Free(param);
-+ break;
-+ }
-+
-+ case FM_IOC_FORCE_INTR:
-+ {
-+ ioc_fm_exceptions param;
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ if (get_user(param, (ioc_fm_exceptions*) compat_ptr(arg)))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+ else
-+#endif
-+ {
-+ if (get_user(param, (ioc_fm_exceptions*)arg))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+
-+ err = FM_ForceIntr(p_LnxWrpFmDev->h_Dev, (e_FmExceptions)param);
-+ break;
-+ }
-+
-+ case FM_IOC_GET_API_VERSION:
-+ {
-+ ioc_fm_api_version_t version;
-+
-+ FM_Get_Api_Version(&version);
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ if (copy_to_user(
-+ (ioc_fm_api_version_t *)compat_ptr(arg),
-+ &version, sizeof(version)))
-+ err = E_READ_FAILED;
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_to_user((ioc_fm_api_version_t *)arg,
-+ &version, sizeof(version)))
-+ err = E_READ_FAILED;
-+ }
-+ }
-+ break;
-+
-+ case FM_IOC_CTRL_MON_START:
-+ {
-+ FM_CtrlMonStart(p_LnxWrpFmDev->h_Dev);
-+ }
-+ break;
-+
-+ case FM_IOC_CTRL_MON_STOP:
-+ {
-+ FM_CtrlMonStop(p_LnxWrpFmDev->h_Dev);
-+ }
-+ break;
-+
-+#if defined(CONFIG_COMPAT)
-+ case FM_IOC_CTRL_MON_GET_COUNTERS_COMPAT:
-+#endif
-+ case FM_IOC_CTRL_MON_GET_COUNTERS:
-+ {
-+ ioc_fm_ctrl_mon_counters_params_t param;
-+ t_FmCtrlMon mon;
-+
-+#if defined(CONFIG_COMPAT)
-+ ioc_compat_fm_ctrl_mon_counters_params_t compat_param;
-+
-+ if (compat)
-+ {
-+ if (copy_from_user(&compat_param, (void *)compat_ptr(arg),
-+ sizeof(compat_param)))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+
-+ param.fm_ctrl_index = compat_param.fm_ctrl_index;
-+ param.p_mon = (fm_ctrl_mon_t *)compat_ptr(compat_param.p_mon);
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_from_user(&param, (void *)arg, sizeof(ioc_fm_ctrl_mon_counters_params_t)))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+
-+ if (FM_CtrlMonGetCounters(p_LnxWrpFmDev->h_Dev, param.fm_ctrl_index, &mon))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+
-+ if (copy_to_user(param.p_mon, &mon, sizeof(t_FmCtrlMon)))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+ break;
-+
-+ default:
-+ return LnxwrpFmPcdIOCTL(p_LnxWrpFmDev, cmd, arg, compat);
-+ }
-+
-+ if (err)
-+ RETURN_ERROR(MINOR, E_INVALID_OPERATION, ("IOCTL FM"));
-+
-+ return E_OK;
-+}
-+
-+t_Error LnxwrpFmPortIOCTL(t_LnxWrpFmPortDev *p_LnxWrpFmPortDev, unsigned int cmd, unsigned long arg, bool compat)
-+{
-+ t_Error err = E_OK;
-+
-+ _fm_ioctl_dbg("cmd:0x%08x(type:0x%02x, nr:%u).\n",
-+ cmd, _IOC_TYPE(cmd), _IOC_NR(cmd) - 70);
-+
-+ switch (cmd)
-+ {
-+ case FM_PORT_IOC_DISABLE:
-+ FM_PORT_Disable(p_LnxWrpFmPortDev->h_Dev);
-+ /* deliberately ignoring error codes here */
-+ return E_OK;
-+
-+ case FM_PORT_IOC_ENABLE:
-+ FM_PORT_Enable(p_LnxWrpFmPortDev->h_Dev);
-+ /* deliberately ignoring error codes here */
-+ return E_OK;
-+
-+ case FM_PORT_IOC_SET_ERRORS_ROUTE:
-+ {
-+ ioc_fm_port_frame_err_select_t errs;
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ if (get_user(errs, (ioc_fm_port_frame_err_select_t*)compat_ptr(arg)))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+ else
-+#endif
-+ {
-+ if (get_user(errs, (ioc_fm_port_frame_err_select_t*)arg))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+
-+ err = FM_PORT_SetErrorsRoute(p_LnxWrpFmPortDev->h_Dev, (fmPortFrameErrSelect_t)errs);
-+ break;
-+ }
-+
-+ case FM_PORT_IOC_SET_RATE_LIMIT:
-+ {
-+ ioc_fm_port_rate_limit_t *param;
-+
-+ param = (ioc_fm_port_rate_limit_t *) XX_Malloc(sizeof(ioc_fm_port_rate_limit_t));
-+ if (!param)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PORT"));
-+
-+ memset(param, 0, sizeof(ioc_fm_port_rate_limit_t));
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ if (copy_from_user(param, (ioc_fm_port_rate_limit_t *)compat_ptr(arg), sizeof(ioc_fm_port_rate_limit_t)))
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_from_user(param, (ioc_fm_port_rate_limit_t *)arg, sizeof(ioc_fm_port_rate_limit_t)))
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+ }
-+
-+ err = FM_PORT_SetRateLimit(p_LnxWrpFmPortDev->h_Dev, (t_FmPortRateLimit *)param);
-+
-+ XX_Free(param);
-+ break;
-+ }
-+
-+ case FM_PORT_IOC_REMOVE_RATE_LIMIT:
-+ FM_PORT_DeleteRateLimit(p_LnxWrpFmPortDev->h_Dev);
-+ /* deliberately ignoring error codes here */
-+ return E_OK;
-+
-+ case FM_PORT_IOC_ALLOC_PCD_FQIDS:
-+ {
-+ ioc_fm_port_pcd_fqids_params_t *param;
-+
-+ if (!p_LnxWrpFmPortDev->pcd_owner_params.cba)
-+ RETURN_ERROR(MINOR, E_INVALID_STATE, ("No one to listen on this PCD!!!"));
-+
-+ param = (ioc_fm_port_pcd_fqids_params_t *) XX_Malloc(sizeof(ioc_fm_port_pcd_fqids_params_t));
-+ if (!param)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PORT"));
-+
-+ memset(param, 0, sizeof(ioc_fm_port_pcd_fqids_params_t));
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ if (copy_from_user(param, (ioc_fm_port_pcd_fqids_params_t *)compat_ptr(arg),
-+ sizeof(ioc_fm_port_pcd_fqids_params_t)))
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_from_user(param, (ioc_fm_port_pcd_fqids_params_t *)arg,
-+ sizeof(ioc_fm_port_pcd_fqids_params_t)))
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+ }
-+
-+ if (p_LnxWrpFmPortDev->pcd_owner_params.cba(p_LnxWrpFmPortDev->pcd_owner_params.dev,
-+ param->num_fqids,
-+ param->alignment,
-+ &param->base_fqid))
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_INVALID_STATE, ("can't allocate fqids for PCD!!!"));
-+ }
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ if (copy_to_user((ioc_fm_port_pcd_fqids_params_t *)compat_ptr(arg),
-+ param, sizeof(ioc_fm_port_pcd_fqids_params_t)))
-+ err = E_READ_FAILED;
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_to_user((ioc_fm_port_pcd_fqids_params_t *)arg,
-+ param, sizeof(ioc_fm_port_pcd_fqids_params_t)))
-+ err = E_READ_FAILED;
-+ }
-+
-+ XX_Free(param);
-+ break;
-+ }
-+
-+ case FM_PORT_IOC_FREE_PCD_FQIDS:
-+ {
-+ uint32_t base_fqid;
-+
-+ if (!p_LnxWrpFmPortDev->pcd_owner_params.cbf)
-+ RETURN_ERROR(MINOR, E_INVALID_STATE, ("No one to listen on this PCD!!!"));
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ if (get_user(base_fqid, (uint32_t*) compat_ptr(arg)))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+ else
-+#endif
-+ {
-+ if (get_user(base_fqid, (uint32_t*)arg))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+
-+ if (p_LnxWrpFmPortDev->pcd_owner_params.cbf(p_LnxWrpFmPortDev->pcd_owner_params.dev, base_fqid))
-+ err = E_WRITE_FAILED;
-+
-+ break;
-+ }
-+
-+#if defined(CONFIG_COMPAT)
-+ case FM_PORT_IOC_SET_PCD_COMPAT:
-+#endif
-+ case FM_PORT_IOC_SET_PCD:
-+ {
-+ ioc_fm_port_pcd_params_t *port_pcd_params;
-+ ioc_fm_port_pcd_prs_params_t *port_pcd_prs_params;
-+ ioc_fm_port_pcd_cc_params_t *port_pcd_cc_params;
-+ ioc_fm_port_pcd_kg_params_t *port_pcd_kg_params;
-+ ioc_fm_port_pcd_plcr_params_t *port_pcd_plcr_params;
-+
-+ port_pcd_params = (ioc_fm_port_pcd_params_t *) XX_Malloc(
-+ sizeof(ioc_fm_port_pcd_params_t) +
-+ sizeof(ioc_fm_port_pcd_prs_params_t) +
-+ sizeof(ioc_fm_port_pcd_cc_params_t) +
-+ sizeof(ioc_fm_port_pcd_kg_params_t) +
-+ sizeof(ioc_fm_port_pcd_plcr_params_t));
-+ if (!port_pcd_params)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PORT"));
-+
-+ memset(port_pcd_params, 0,
-+ sizeof(ioc_fm_port_pcd_params_t) +
-+ sizeof(ioc_fm_port_pcd_prs_params_t) +
-+ sizeof(ioc_fm_port_pcd_cc_params_t) +
-+ sizeof(ioc_fm_port_pcd_kg_params_t) +
-+ sizeof(ioc_fm_port_pcd_plcr_params_t));
-+
-+ port_pcd_prs_params = (ioc_fm_port_pcd_prs_params_t *) (port_pcd_params + 1);
-+ port_pcd_cc_params = (ioc_fm_port_pcd_cc_params_t *) (port_pcd_prs_params + 1);
-+ port_pcd_kg_params = (ioc_fm_port_pcd_kg_params_t *) (port_pcd_cc_params + 1);
-+ port_pcd_plcr_params = (ioc_fm_port_pcd_plcr_params_t *) (port_pcd_kg_params + 1);
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_port_pcd_params_t *compat_port_pcd_params;
-+ ioc_fm_port_pcd_prs_params_t *same_port_pcd_prs_params;
-+ ioc_compat_fm_port_pcd_cc_params_t *compat_port_pcd_cc_params;
-+ ioc_compat_fm_port_pcd_kg_params_t *compat_port_pcd_kg_params;
-+ ioc_compat_fm_port_pcd_plcr_params_t *compat_port_pcd_plcr_params;
-+
-+ compat_port_pcd_params = (ioc_compat_fm_port_pcd_params_t *) XX_Malloc(
-+ sizeof(ioc_compat_fm_port_pcd_params_t) +
-+ sizeof(ioc_fm_port_pcd_prs_params_t) +
-+ sizeof(ioc_compat_fm_port_pcd_cc_params_t) +
-+ sizeof(ioc_compat_fm_port_pcd_kg_params_t) +
-+ sizeof(ioc_compat_fm_port_pcd_plcr_params_t));
-+ if (!compat_port_pcd_params)
-+ {
-+ XX_Free(port_pcd_params);
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PORT"));
-+ }
-+
-+ memset(compat_port_pcd_params, 0,
-+ sizeof(ioc_compat_fm_port_pcd_params_t) +
-+ sizeof(ioc_fm_port_pcd_prs_params_t) +
-+ sizeof(ioc_compat_fm_port_pcd_cc_params_t) +
-+ sizeof(ioc_compat_fm_port_pcd_kg_params_t) +
-+ sizeof(ioc_compat_fm_port_pcd_plcr_params_t));
-+ same_port_pcd_prs_params = (ioc_fm_port_pcd_prs_params_t *) (compat_port_pcd_params + 1);
-+ compat_port_pcd_cc_params = (ioc_compat_fm_port_pcd_cc_params_t *) (same_port_pcd_prs_params + 1);
-+ compat_port_pcd_kg_params = (ioc_compat_fm_port_pcd_kg_params_t *) (compat_port_pcd_cc_params + 1);
-+ compat_port_pcd_plcr_params = (ioc_compat_fm_port_pcd_plcr_params_t *) (compat_port_pcd_kg_params + 1);
-+
-+ if (copy_from_user(compat_port_pcd_params,
-+ (ioc_compat_fm_port_pcd_params_t*) compat_ptr(arg),
-+ sizeof(ioc_compat_fm_port_pcd_params_t)))
-+ err = E_WRITE_FAILED;
-+
-+ while (!err) /* pseudo-while */
-+ {
-+ /* set pointers from where to copy from: */
-+ port_pcd_params->p_prs_params = compat_ptr(compat_port_pcd_params->p_prs_params); /* same structure */
-+ port_pcd_params->p_cc_params = compat_ptr(compat_port_pcd_params->p_cc_params);
-+ port_pcd_params->p_kg_params = compat_ptr(compat_port_pcd_params->p_kg_params);
-+ port_pcd_params->p_plcr_params = compat_ptr(compat_port_pcd_params->p_plcr_params);
-+ port_pcd_params->p_ip_reassembly_manip = compat_ptr(compat_port_pcd_params->p_ip_reassembly_manip);
-+#if (DPAA_VERSION >= 11)
-+ port_pcd_params->p_capwap_reassembly_manip = compat_ptr(compat_port_pcd_params->p_capwap_reassembly_manip);
-+#endif
-+ /* the prs member is the same, no compat structure...memcpy only */
-+ if (port_pcd_params->p_prs_params)
-+ {
-+ if (copy_from_user(same_port_pcd_prs_params,
-+ port_pcd_params->p_prs_params,
-+ sizeof(ioc_fm_port_pcd_prs_params_t)))
-+ {
-+ err = E_WRITE_FAILED;
-+ break; /* from pseudo-while */
-+ }
-+
-+ memcpy(port_pcd_prs_params, same_port_pcd_prs_params, sizeof(ioc_fm_port_pcd_prs_params_t));
-+ port_pcd_params->p_prs_params = port_pcd_prs_params;
-+ }
-+
-+ if (port_pcd_params->p_cc_params)
-+ {
-+ if (copy_from_user(compat_port_pcd_cc_params,
-+ port_pcd_params->p_cc_params,
-+ sizeof(ioc_compat_fm_port_pcd_cc_params_t)))
-+ {
-+ err = E_WRITE_FAILED;
-+ break; /* from pseudo-while */
-+ }
-+
-+ port_pcd_params->p_cc_params = port_pcd_cc_params;
-+ }
-+
-+ if (port_pcd_params->p_kg_params)
-+ {
-+ if (copy_from_user(compat_port_pcd_kg_params,
-+ port_pcd_params->p_kg_params,
-+ sizeof(ioc_compat_fm_port_pcd_kg_params_t)))
-+ {
-+ err = E_WRITE_FAILED;
-+ break; /* from pseudo-while */
-+ }
-+
-+ port_pcd_params->p_kg_params = port_pcd_kg_params;
-+ }
-+
-+ if (port_pcd_params->p_plcr_params)
-+ {
-+ if (copy_from_user(compat_port_pcd_plcr_params,
-+ port_pcd_params->p_plcr_params,
-+ sizeof(ioc_compat_fm_port_pcd_plcr_params_t)))
-+ {
-+ err = E_WRITE_FAILED;
-+ break; /* from pseudo-while */
-+ }
-+
-+ port_pcd_params->p_plcr_params = port_pcd_plcr_params;
-+ }
-+
-+ break; /* pseudo-while: always run once! */
-+ }
-+
-+ if (!err)
-+ compat_copy_fm_port_pcd(compat_port_pcd_params, port_pcd_params, COMPAT_US_TO_K);
-+
-+ XX_Free(compat_port_pcd_params);
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_from_user(port_pcd_params,
-+ (ioc_fm_port_pcd_params_t*) arg,
-+ sizeof(ioc_fm_port_pcd_params_t)))
-+ err = E_WRITE_FAILED;
-+
-+ while (!err) /* pseudo-while */
-+ {
-+ if (port_pcd_params->p_prs_params)
-+ {
-+ if (copy_from_user(port_pcd_prs_params,
-+ port_pcd_params->p_prs_params,
-+ sizeof(ioc_fm_port_pcd_prs_params_t)))
-+ {
-+ err = E_WRITE_FAILED;
-+ break; /* from pseudo-while */
-+ }
-+
-+ port_pcd_params->p_prs_params = port_pcd_prs_params;
-+ }
-+
-+ if (port_pcd_params->p_cc_params)
-+ {
-+ if (copy_from_user(port_pcd_cc_params,
-+ port_pcd_params->p_cc_params,
-+ sizeof(ioc_fm_port_pcd_cc_params_t)))
-+ {
-+ err = E_WRITE_FAILED;
-+ break; /* from pseudo-while */
-+ }
-+
-+ port_pcd_params->p_cc_params = port_pcd_cc_params;
-+ }
-+
-+ if (port_pcd_params->p_kg_params)
-+ {
-+ if (copy_from_user(port_pcd_kg_params,
-+ port_pcd_params->p_kg_params,
-+ sizeof(ioc_fm_port_pcd_kg_params_t)))
-+ {
-+ err = E_WRITE_FAILED;
-+ break; /* from pseudo-while */
-+ }
-+
-+ port_pcd_params->p_kg_params = port_pcd_kg_params;
-+ }
-+
-+ if (port_pcd_params->p_plcr_params)
-+ {
-+ if (copy_from_user(port_pcd_plcr_params,
-+ port_pcd_params->p_plcr_params,
-+ sizeof(ioc_fm_port_pcd_plcr_params_t)))
-+ {
-+ err = E_WRITE_FAILED;
-+ break; /* from pseudo-while */
-+ }
-+
-+ port_pcd_params->p_plcr_params = port_pcd_plcr_params;
-+ }
-+
-+ break; /* pseudo-while: always run once! */
-+ }
-+ }
-+
-+ if (!err)
-+ err = FM_PORT_SetPCD(p_LnxWrpFmPortDev->h_Dev, (t_FmPortPcdParams*) port_pcd_params);
-+
-+ XX_Free(port_pcd_params);
-+ break;
-+ }
-+
-+ case FM_PORT_IOC_DELETE_PCD:
-+ err = FM_PORT_DeletePCD(p_LnxWrpFmPortDev->h_Dev);
-+ break;
-+
-+#if defined(CONFIG_COMPAT)
-+ case FM_PORT_IOC_PCD_KG_MODIFY_INITIAL_SCHEME_COMPAT:
-+#endif
-+ case FM_PORT_IOC_PCD_KG_MODIFY_INITIAL_SCHEME:
-+ {
-+ ioc_fm_pcd_kg_scheme_select_t *param;
-+
-+ param = (ioc_fm_pcd_kg_scheme_select_t *) XX_Malloc(
-+ sizeof(ioc_fm_pcd_kg_scheme_select_t));
-+ if (!param)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PORT"));
-+
-+ memset(param, 0, sizeof(ioc_fm_pcd_kg_scheme_select_t));
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_pcd_kg_scheme_select_t *compat_param;
-+
-+ compat_param = (ioc_compat_fm_pcd_kg_scheme_select_t *) XX_Malloc(
-+ sizeof(ioc_compat_fm_pcd_kg_scheme_select_t));
-+ if (!compat_param)
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PORT"));
-+ }
-+
-+ memset(compat_param, 0, sizeof(ioc_compat_fm_pcd_kg_scheme_select_t));
-+ if (copy_from_user(compat_param,
-+ (ioc_compat_fm_pcd_kg_scheme_select_t *) compat_ptr(arg),
-+ sizeof(ioc_compat_fm_pcd_kg_scheme_select_t)))
-+ {
-+ XX_Free(compat_param);
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+
-+ compat_copy_fm_pcd_kg_scheme_select(compat_param, param, COMPAT_US_TO_K);
-+
-+ XX_Free(compat_param);
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_from_user(param, (ioc_fm_pcd_kg_scheme_select_t *)arg,
-+ sizeof(ioc_fm_pcd_kg_scheme_select_t)))
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+ }
-+
-+ err = FM_PORT_PcdKgModifyInitialScheme(p_LnxWrpFmPortDev->h_Dev, (t_FmPcdKgSchemeSelect *)param);
-+
-+ XX_Free(param);
-+ break;
-+ }
-+
-+#if defined(CONFIG_COMPAT)
-+ case FM_PORT_IOC_PCD_PLCR_MODIFY_INITIAL_PROFILE_COMPAT:
-+#endif
-+ case FM_PORT_IOC_PCD_PLCR_MODIFY_INITIAL_PROFILE:
-+ {
-+ ioc_fm_obj_t id;
-+
-+ memset(&id, 0 , sizeof(ioc_fm_obj_t));
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_obj_t compat_id;
-+
-+ if (copy_from_user(&compat_id, (ioc_compat_fm_obj_t *) compat_ptr(arg), sizeof(ioc_compat_fm_obj_t)))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+
-+ id.obj = compat_ptr(compat_id.obj);
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_from_user(&id, (ioc_fm_obj_t *) arg, sizeof(ioc_fm_obj_t)))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+
-+ err = FM_PORT_PcdPlcrModifyInitialProfile(p_LnxWrpFmPortDev->h_Dev, id.obj);
-+ break;
-+ }
-+
-+#if defined(CONFIG_COMPAT)
-+ case FM_PORT_IOC_PCD_KG_BIND_SCHEMES_COMPAT:
-+#endif
-+ case FM_PORT_IOC_PCD_KG_BIND_SCHEMES:
-+ {
-+ ioc_fm_pcd_port_schemes_params_t *param;
-+
-+ param = (ioc_fm_pcd_port_schemes_params_t *) XX_Malloc(
-+ sizeof(ioc_fm_pcd_port_schemes_params_t));
-+ if (!param)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PORT"));
-+
-+ memset(param, 0 , sizeof(ioc_fm_pcd_port_schemes_params_t));
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_pcd_port_schemes_params_t compat_param;
-+
-+ if (copy_from_user(&compat_param,
-+ (ioc_compat_fm_pcd_port_schemes_params_t *) compat_ptr(arg),
-+ sizeof(ioc_compat_fm_pcd_port_schemes_params_t)))
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+
-+ compat_copy_fm_pcd_kg_schemes_params(&compat_param, param, COMPAT_US_TO_K);
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_from_user(param, (ioc_fm_pcd_port_schemes_params_t *) arg,
-+ sizeof(ioc_fm_pcd_port_schemes_params_t)))
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+ }
-+
-+ err = FM_PORT_PcdKgBindSchemes(p_LnxWrpFmPortDev->h_Dev, (t_FmPcdPortSchemesParams *)param);
-+
-+ XX_Free(param);
-+ break;
-+ }
-+
-+#if defined(CONFIG_COMPAT)
-+ case FM_PORT_IOC_PCD_KG_UNBIND_SCHEMES_COMPAT:
-+#endif
-+ case FM_PORT_IOC_PCD_KG_UNBIND_SCHEMES:
-+ {
-+ ioc_fm_pcd_port_schemes_params_t *param;
-+
-+ param = (ioc_fm_pcd_port_schemes_params_t *) XX_Malloc(
-+ sizeof(ioc_fm_pcd_port_schemes_params_t));
-+ if (!param)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PORT"));
-+
-+ memset(param, 0 , sizeof(ioc_fm_pcd_port_schemes_params_t));
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_pcd_port_schemes_params_t compat_param;
-+
-+ if (copy_from_user(&compat_param,
-+ (ioc_compat_fm_pcd_port_schemes_params_t *) compat_ptr(arg),
-+ sizeof(ioc_compat_fm_pcd_port_schemes_params_t)))
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+
-+ compat_copy_fm_pcd_kg_schemes_params(&compat_param, param, COMPAT_US_TO_K);
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_from_user(param, (ioc_fm_pcd_port_schemes_params_t *) arg,
-+ sizeof(ioc_fm_pcd_port_schemes_params_t)))
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+ }
-+
-+ err = FM_PORT_PcdKgUnbindSchemes(p_LnxWrpFmPortDev->h_Dev, (t_FmPcdPortSchemesParams *)param);
-+
-+ XX_Free(param);
-+ break;
-+ }
-+
-+ case FM_PORT_IOC_PCD_PLCR_ALLOC_PROFILES:
-+ {
-+ uint16_t num;
-+ if (get_user(num, (uint16_t*) arg))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+
-+ err = FM_PORT_PcdPlcrAllocProfiles(p_LnxWrpFmPortDev->h_Dev, num);
-+ break;
-+ }
-+
-+ case FM_PORT_IOC_PCD_PLCR_FREE_PROFILES:
-+ err = FM_PORT_PcdPlcrFreeProfiles(p_LnxWrpFmPortDev->h_Dev);
-+ break;
-+
-+ case FM_PORT_IOC_DETACH_PCD:
-+ err = FM_PORT_DetachPCD(p_LnxWrpFmPortDev->h_Dev);
-+ break;
-+
-+ case FM_PORT_IOC_ATTACH_PCD:
-+ err = FM_PORT_AttachPCD(p_LnxWrpFmPortDev->h_Dev);
-+ break;
-+
-+#if defined(CONFIG_COMPAT)
-+ case FM_PORT_IOC_PCD_CC_MODIFY_TREE_COMPAT:
-+#endif
-+ case FM_PORT_IOC_PCD_CC_MODIFY_TREE:
-+ {
-+ ioc_fm_obj_t id;
-+
-+ memset(&id, 0 , sizeof(ioc_fm_obj_t));
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_obj_t compat_id;
-+
-+ if (copy_from_user(&compat_id, (ioc_compat_fm_obj_t *) compat_ptr(arg), sizeof(ioc_compat_fm_obj_t)))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+
-+ compat_copy_fm_port_pcd_modify_tree(&compat_id, &id, COMPAT_US_TO_K);
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_from_user(&id, (ioc_fm_obj_t *) arg, sizeof(ioc_fm_obj_t)))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+
-+ err = FM_PORT_PcdCcModifyTree(p_LnxWrpFmPortDev->h_Dev, id.obj);
-+ break;
-+ }
-+
-+ case FM_PORT_IOC_ADD_CONGESTION_GRPS:
-+ case FM_PORT_IOC_REMOVE_CONGESTION_GRPS:
-+ {
-+ ioc_fm_port_congestion_groups_t *param;
-+
-+ param = (ioc_fm_port_congestion_groups_t*) XX_Malloc(sizeof(ioc_fm_port_congestion_groups_t));
-+ if (!param)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PORT"));
-+
-+ memset(param, 0, sizeof(ioc_fm_port_congestion_groups_t));
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ if (copy_from_user(param, (t_FmPortCongestionGrps*) compat_ptr(arg),
-+ sizeof(t_FmPortCongestionGrps)))
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+ }
-+ else
-+#endif /* CONFIG_COMPAT */
-+ {
-+ if (copy_from_user(param, (t_FmPortCongestionGrps*) arg,
-+ sizeof(t_FmPortCongestionGrps)))
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+ }
-+
-+ err = (cmd == FM_PORT_IOC_ADD_CONGESTION_GRPS)
-+ ? FM_PORT_AddCongestionGrps(p_LnxWrpFmPortDev->h_Dev, (t_FmPortCongestionGrps*) param)
-+ : FM_PORT_RemoveCongestionGrps(p_LnxWrpFmPortDev->h_Dev, (t_FmPortCongestionGrps*) param)
-+ ;
-+
-+ XX_Free(param);
-+ break;
-+ }
-+
-+ case FM_PORT_IOC_ADD_RX_HASH_MAC_ADDR:
-+ case FM_PORT_IOC_REMOVE_RX_HASH_MAC_ADDR:
-+ {
-+ ioc_fm_port_mac_addr_params_t *param;
-+
-+ param = (ioc_fm_port_mac_addr_params_t*) XX_Malloc(
-+ sizeof(ioc_fm_port_mac_addr_params_t));
-+ if (!param)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PORT"));
-+
-+ memset(param, 0, sizeof(ioc_fm_port_mac_addr_params_t));
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ if (copy_from_user(param, (ioc_fm_port_mac_addr_params_t*) compat_ptr(arg),
-+ sizeof(ioc_fm_port_mac_addr_params_t)))
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+ }
-+ else
-+#endif /* CONFIG_COMPAT */
-+ {
-+ if (copy_from_user(param, (ioc_fm_port_mac_addr_params_t*) arg,
-+ sizeof(ioc_fm_port_mac_addr_params_t)))
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+ }
-+
-+ if (p_LnxWrpFmPortDev->pcd_owner_params.dev)
-+ {
-+ int id = -1;
-+
-+ switch(p_LnxWrpFmPortDev->settings.param.portType)
-+ {
-+ case e_FM_PORT_TYPE_RX:
-+ case e_FM_PORT_TYPE_TX:
-+ id = p_LnxWrpFmPortDev->id;
-+ break;
-+ case e_FM_PORT_TYPE_RX_10G:
-+ case e_FM_PORT_TYPE_TX_10G:
-+ id = p_LnxWrpFmPortDev->id + FM_MAX_NUM_OF_1G_MACS;
-+ break;
-+ default:
-+ err = E_NOT_AVAILABLE;
-+ REPORT_ERROR(MINOR, err, ("Attempt to add/remove hash MAC addr. to/from MAC-less port!"));
-+ }
-+ if (id >= 0)
-+ {
-+ t_LnxWrpFmDev *fm = (t_LnxWrpFmDev *)p_LnxWrpFmPortDev->h_LnxWrpFmDev;
-+ t_Handle mac_handle = fm->macs[id].h_Dev;
-+
-+ err = (cmd == FM_PORT_IOC_ADD_RX_HASH_MAC_ADDR)
-+ ? FM_MAC_AddHashMacAddr(mac_handle, (t_EnetAddr*) param)
-+ : FM_MAC_RemoveHashMacAddr(mac_handle, (t_EnetAddr*) param);
-+ }
-+ }
-+ else
-+ {
-+ err = E_NOT_AVAILABLE;
-+ REPORT_ERROR(MINOR, err, ("Port not initialized or other error!?!?"));
-+ }
-+
-+ XX_Free(param);
-+ break;
-+ }
-+
-+ case FM_PORT_IOC_SET_TX_PAUSE_FRAMES:
-+ {
-+ t_LnxWrpFmDev *p_LnxWrpFmDev =
-+ (t_LnxWrpFmDev *)p_LnxWrpFmPortDev->h_LnxWrpFmDev;
-+ ioc_fm_port_tx_pause_frames_params_t param;
-+ int mac_id = p_LnxWrpFmPortDev->id;
-+
-+ if(&p_LnxWrpFmDev->txPorts[mac_id] != p_LnxWrpFmPortDev)
-+ mac_id += FM_MAX_NUM_OF_1G_MACS; /* 10G port */
-+
-+ if (copy_from_user(&param, (ioc_fm_port_tx_pause_frames_params_t *)arg,
-+ sizeof(ioc_fm_port_tx_pause_frames_params_t)))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+
-+ if (p_LnxWrpFmDev && p_LnxWrpFmDev->macs[mac_id].h_Dev)
-+ {
-+ FM_MAC_SetTxPauseFrames(p_LnxWrpFmDev->macs[mac_id].h_Dev,
-+ param.priority,
-+ param.pause_time,
-+ param.thresh_time);
-+ }
-+ else
-+ {
-+ err = E_NOT_AVAILABLE;
-+ REPORT_ERROR(MINOR, err, ("Port not initialized or other error!"));
-+ }
-+
-+ break;
-+ }
-+
-+ case FM_PORT_IOC_CONFIG_BUFFER_PREFIX_CONTENT:
-+ {
-+ ioc_fm_buffer_prefix_content_t *param;
-+
-+ param = (ioc_fm_buffer_prefix_content_t*) XX_Malloc(sizeof(ioc_fm_buffer_prefix_content_t));
-+ if (!param)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PORT"));
-+
-+ memset(param, 0, sizeof(ioc_fm_buffer_prefix_content_t));
-+
-+ if (copy_from_user(param, (ioc_fm_buffer_prefix_content_t*) arg,
-+ sizeof(ioc_fm_buffer_prefix_content_t)))
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+
-+ if (FM_PORT_ConfigBufferPrefixContent(p_LnxWrpFmPortDev->h_Dev,
-+ (t_FmBufferPrefixContent *)param))
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+
-+ XX_Free(param);
-+ break;
-+ }
-+
-+#if (DPAA_VERSION >= 11)
-+#if defined(CONFIG_COMPAT)
-+ case FM_PORT_IOC_VSP_ALLOC_COMPAT:
-+#endif
-+ case FM_PORT_IOC_VSP_ALLOC:
-+ {
-+ ioc_fm_port_vsp_alloc_params_t *param;
-+ t_LnxWrpFmDev *p_LnxWrpFmDev;
-+ t_LnxWrpFmPortDev *p_LnxWrpFmTxPortDev;
-+
-+ param = (ioc_fm_port_vsp_alloc_params_t *) XX_Malloc(
-+ sizeof(ioc_fm_port_vsp_alloc_params_t));
-+ if (!param)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PORT"));
-+
-+ memset(param, 0, sizeof(ioc_fm_port_vsp_alloc_params_t));
-+
-+#if defined(CONFIG_COMPAT)
-+ if (compat)
-+ {
-+ ioc_compat_fm_port_vsp_alloc_params_t *compat_param;
-+
-+ compat_param = (ioc_compat_fm_port_vsp_alloc_params_t *) XX_Malloc(
-+ sizeof(ioc_compat_fm_port_vsp_alloc_params_t));
-+ if (!compat_param)
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("IOCTL FM PORT"));
-+ }
-+
-+ memset(compat_param, 0, sizeof(ioc_compat_fm_port_vsp_alloc_params_t));
-+ if (copy_from_user(compat_param,
-+ (ioc_compat_fm_port_vsp_alloc_params_t *) compat_ptr(arg),
-+ sizeof(ioc_compat_fm_port_vsp_alloc_params_t)))
-+ {
-+ XX_Free(compat_param);
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+
-+ compat_copy_fm_port_vsp_alloc_params(compat_param, param, COMPAT_US_TO_K);
-+
-+ XX_Free(compat_param);
-+ }
-+ else
-+#endif
-+ {
-+ if (copy_from_user(param, (ioc_fm_port_vsp_alloc_params_t *)arg,
-+ sizeof(ioc_fm_port_vsp_alloc_params_t)))
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+ }
-+
-+ /* Userspace may not have the Tx port t_handle when issuing the IOCTL */
-+ if (p_LnxWrpFmPortDev->settings.param.portType == e_FM_PORT_TYPE_RX ||
-+ p_LnxWrpFmPortDev->settings.param.portType == e_FM_PORT_TYPE_RX_10G)
-+ {
-+ /* Determine the Tx port t_Handle from the Rx port id */
-+ p_LnxWrpFmDev = p_LnxWrpFmPortDev->h_LnxWrpFmDev;
-+ p_LnxWrpFmTxPortDev = &p_LnxWrpFmDev->txPorts[p_LnxWrpFmPortDev->id];
-+ param->p_fm_tx_port = p_LnxWrpFmTxPortDev->h_Dev;
-+ }
-+
-+ if (FM_PORT_VSPAlloc(p_LnxWrpFmPortDev->h_Dev, (t_FmPortVSPAllocParams *)param))
-+ {
-+ XX_Free(param);
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+ }
-+
-+ XX_Free(param);
-+ break;
-+ }
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+ case FM_PORT_IOC_GET_MAC_STATISTICS:
-+ {
-+ t_LnxWrpFmDev *p_LnxWrpFmDev =
-+ (t_LnxWrpFmDev *)p_LnxWrpFmPortDev->h_LnxWrpFmDev;
-+ ioc_fm_port_mac_statistics_t param;
-+ int mac_id = p_LnxWrpFmPortDev->id;
-+
-+ if (!p_LnxWrpFmDev)
-+ RETURN_ERROR(MINOR, E_NOT_AVAILABLE, ("Port not initialized or other error!"));
-+
-+ if (&p_LnxWrpFmDev->txPorts[mac_id] != p_LnxWrpFmPortDev &&
-+ &p_LnxWrpFmDev->rxPorts[mac_id] != p_LnxWrpFmPortDev)
-+ mac_id += FM_MAX_NUM_OF_1G_MACS; /* 10G port */
-+
-+ if (!p_LnxWrpFmDev->macs[mac_id].h_Dev)
-+ RETURN_ERROR(MINOR, E_NOT_AVAILABLE, ("Port not initialized or other error!"));
-+
-+ if (FM_MAC_GetStatistics(p_LnxWrpFmDev->macs[mac_id].h_Dev,
-+ (t_FmMacStatistics *)&param))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+
-+ if (copy_to_user((ioc_fm_port_mac_statistics_t *)arg, &param,
-+ sizeof(ioc_fm_port_mac_statistics_t)))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+
-+ break;
-+ }
-+
-+ case FM_PORT_IOC_GET_MAC_FRAME_SIZE_COUNTERS:
-+ {
-+ t_LnxWrpFmDev *p_LnxWrpFmDev =
-+ (t_LnxWrpFmDev *)p_LnxWrpFmPortDev->h_LnxWrpFmDev;
-+ ioc_fm_port_mac_frame_size_counters_t param;
-+ t_FmMacFrameSizeCounters frameSizeCounters;
-+ int mac_id = p_LnxWrpFmPortDev->id;
-+
-+ if (!p_LnxWrpFmDev)
-+ RETURN_ERROR(MINOR, E_NOT_AVAILABLE, ("Port not initialized or other error!"));
-+
-+ if (&p_LnxWrpFmDev->txPorts[mac_id] != p_LnxWrpFmPortDev &&
-+ &p_LnxWrpFmDev->rxPorts[mac_id] != p_LnxWrpFmPortDev)
-+ mac_id += FM_MAX_NUM_OF_1G_MACS; /* 10G port */
-+
-+ if (!p_LnxWrpFmDev->macs[mac_id].h_Dev)
-+ RETURN_ERROR(MINOR, E_NOT_AVAILABLE, ("Port not initialized or other error!"));
-+
-+ if (copy_from_user(&param, (ioc_fm_port_mac_frame_size_counters_t *)arg,
-+ sizeof(ioc_fm_port_mac_frame_size_counters_t)))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+
-+ if (FM_MAC_GetFrameSizeCounters(p_LnxWrpFmDev->macs[mac_id].h_Dev,
-+ &frameSizeCounters, param.type))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+
-+ param.count_pkts_64 = frameSizeCounters.count_pkts_64;
-+ param.count_pkts_65_to_127 = frameSizeCounters.count_pkts_65_to_127;
-+ param.count_pkts_128_to_255 = frameSizeCounters.count_pkts_128_to_255;
-+ param.count_pkts_256_to_511 = frameSizeCounters.count_pkts_256_to_511;
-+ param.count_pkts_512_to_1023 = frameSizeCounters.count_pkts_512_to_1023;
-+ param.count_pkts_1024_to_1518 = frameSizeCounters.count_pkts_1024_to_1518;
-+ param.count_pkts_1519_to_1522 = frameSizeCounters.count_pkts_1519_to_1522;
-+
-+ if (copy_to_user((ioc_fm_port_mac_frame_size_counters_t *)arg, &param,
-+ sizeof(ioc_fm_port_mac_frame_size_counters_t)))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+
-+ break;
-+ }
-+
-+ case FM_PORT_IOC_GET_BMI_COUNTERS:
-+ {
-+ t_LnxWrpFmDev *p_LnxWrpFmDev =
-+ (t_LnxWrpFmDev *)p_LnxWrpFmPortDev->h_LnxWrpFmDev;
-+ ioc_fm_port_bmi_stats_t param;
-+
-+ if (!p_LnxWrpFmDev)
-+ RETURN_ERROR(MINOR, E_NOT_AVAILABLE, ("Port not initialized or other error!"));
-+
-+ if (FM_PORT_GetBmiCounters(p_LnxWrpFmPortDev->h_Dev,
-+ (t_FmPortBmiStats *)&param))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+
-+ if (copy_to_user((ioc_fm_port_bmi_stats_t *)arg, &param,
-+ sizeof(ioc_fm_port_bmi_stats_t)))
-+ RETURN_ERROR(MINOR, E_WRITE_FAILED, NO_MSG);
-+
-+ break;
-+ }
-+
-+ default:
-+ RETURN_ERROR(MINOR, E_INVALID_SELECTION,
-+ ("invalid ioctl: cmd:0x%08x(type:0x%02x, nr:0x%02x.\n",
-+ cmd, _IOC_TYPE(cmd), _IOC_NR(cmd)));
-+ }
-+
-+ if (err)
-+ RETURN_ERROR(MINOR, E_INVALID_OPERATION, ("IOCTL FM PORT"));
-+
-+ return E_OK;
-+}
-+
-+/*****************************************************************************/
-+/* API routines for the FM Linux Device */
-+/*****************************************************************************/
-+
-+static int fm_open(struct inode *inode, struct file *file)
-+{
-+ t_LnxWrpFmDev *p_LnxWrpFmDev = NULL;
-+ t_LnxWrpFmPortDev *p_LnxWrpFmPortDev = NULL;
-+ unsigned int major = imajor(inode);
-+ unsigned int minor = iminor(inode);
-+ struct device_node *fm_node;
-+ static struct of_device_id fm_node_of_match[] = {
-+ { .compatible = "fsl,fman", },
-+ { /* end of list */ },
-+ };
-+
-+ DBG(TRACE, ("Opening minor - %d - ", minor));
-+
-+ if (file->private_data != NULL)
-+ return 0;
-+
-+ /* Get all the FM nodes */
-+ for_each_matching_node(fm_node, fm_node_of_match) {
-+ struct platform_device *of_dev;
-+
-+ of_dev = of_find_device_by_node(fm_node);
-+ if (unlikely(of_dev == NULL)) {
-+ REPORT_ERROR(MAJOR, E_INVALID_VALUE, ("fm id!"));
-+ return -ENXIO;
-+ }
-+
-+ p_LnxWrpFmDev = (t_LnxWrpFmDev *)fm_bind(&of_dev->dev);
-+ if (p_LnxWrpFmDev->major == major)
-+ break;
-+ fm_unbind((struct fm *)p_LnxWrpFmDev);
-+ p_LnxWrpFmDev = NULL;
-+ }
-+
-+ if (!p_LnxWrpFmDev)
-+ return -ENODEV;
-+
-+ if (minor == DEV_FM_MINOR_BASE)
-+ file->private_data = p_LnxWrpFmDev;
-+ else if (minor == DEV_FM_PCD_MINOR_BASE)
-+ file->private_data = p_LnxWrpFmDev;
-+ else {
-+ if (minor == DEV_FM_OH_PORTS_MINOR_BASE)
-+ p_LnxWrpFmPortDev = &p_LnxWrpFmDev->hcPort;
-+ else if ((minor > DEV_FM_OH_PORTS_MINOR_BASE) && (minor < DEV_FM_RX_PORTS_MINOR_BASE))
-+ p_LnxWrpFmPortDev = &p_LnxWrpFmDev->opPorts[minor-DEV_FM_OH_PORTS_MINOR_BASE-1];
-+ else if ((minor >= DEV_FM_RX_PORTS_MINOR_BASE) && (minor < DEV_FM_TX_PORTS_MINOR_BASE))
-+ p_LnxWrpFmPortDev = &p_LnxWrpFmDev->rxPorts[minor-DEV_FM_RX_PORTS_MINOR_BASE];
-+ else if ((minor >= DEV_FM_TX_PORTS_MINOR_BASE) && (minor < DEV_FM_MAX_MINORS))
-+ p_LnxWrpFmPortDev = &p_LnxWrpFmDev->txPorts[minor-DEV_FM_TX_PORTS_MINOR_BASE];
-+ else
-+ return -EINVAL;
-+
-+ /* if trying to open port, check if it initialized */
-+ if (!p_LnxWrpFmPortDev->h_Dev)
-+ return -ENODEV;
-+
-+ p_LnxWrpFmPortDev = (t_LnxWrpFmPortDev *)fm_port_bind(p_LnxWrpFmPortDev->dev);
-+ file->private_data = p_LnxWrpFmPortDev;
-+ fm_unbind((struct fm *)p_LnxWrpFmDev);
-+ }
-+
-+ if (file->private_data == NULL)
-+ return -ENXIO;
-+
-+ return 0;
-+}
-+
-+static int fm_close(struct inode *inode, struct file *file)
-+{
-+ t_LnxWrpFmDev *p_LnxWrpFmDev;
-+ t_LnxWrpFmPortDev *p_LnxWrpFmPortDev;
-+ unsigned int minor = iminor(inode);
-+ int err = 0;
-+
-+ DBG(TRACE, ("Closing minor - %d - ", minor));
-+
-+ if ((minor == DEV_FM_MINOR_BASE) ||
-+ (minor == DEV_FM_PCD_MINOR_BASE))
-+ {
-+ p_LnxWrpFmDev = (t_LnxWrpFmDev*)file->private_data;
-+ if (!p_LnxWrpFmDev)
-+ return -ENODEV;
-+ fm_unbind((struct fm *)p_LnxWrpFmDev);
-+ }
-+ else if (((minor >= DEV_FM_OH_PORTS_MINOR_BASE) && (minor < DEV_FM_RX_PORTS_MINOR_BASE)) ||
-+ ((minor >= DEV_FM_RX_PORTS_MINOR_BASE) && (minor < DEV_FM_TX_PORTS_MINOR_BASE)) ||
-+ ((minor >= DEV_FM_TX_PORTS_MINOR_BASE) && (minor < DEV_FM_MAX_MINORS)))
-+ {
-+ p_LnxWrpFmPortDev = (t_LnxWrpFmPortDev*)file->private_data;
-+ if (!p_LnxWrpFmPortDev)
-+ return -ENODEV;
-+ fm_port_unbind((struct fm_port *)p_LnxWrpFmPortDev);
-+ }
-+
-+ return err;
-+}
-+
-+static int fm_ioctls(unsigned int minor, struct file *file, unsigned int cmd, unsigned long arg, bool compat)
-+{
-+ DBG(TRACE, ("IOCTL minor - %u, cmd - 0x%08x, arg - 0x%08lx \n", minor, cmd, arg));
-+
-+ if ((minor == DEV_FM_MINOR_BASE) ||
-+ (minor == DEV_FM_PCD_MINOR_BASE))
-+ {
-+ t_LnxWrpFmDev *p_LnxWrpFmDev = ((t_LnxWrpFmDev*)file->private_data);
-+ if (!p_LnxWrpFmDev)
-+ return -ENODEV;
-+ if (LnxwrpFmIOCTL(p_LnxWrpFmDev, cmd, arg, compat))
-+ return -EFAULT;
-+ }
-+ else if (((minor >= DEV_FM_OH_PORTS_MINOR_BASE) && (minor < DEV_FM_RX_PORTS_MINOR_BASE)) ||
-+ ((minor >= DEV_FM_RX_PORTS_MINOR_BASE) && (minor < DEV_FM_TX_PORTS_MINOR_BASE)) ||
-+ ((minor >= DEV_FM_TX_PORTS_MINOR_BASE) && (minor < DEV_FM_MAX_MINORS)))
-+ {
-+ t_LnxWrpFmPortDev *p_LnxWrpFmPortDev = ((t_LnxWrpFmPortDev*)file->private_data);
-+ if (!p_LnxWrpFmPortDev)
-+ return -ENODEV;
-+ if (LnxwrpFmPortIOCTL(p_LnxWrpFmPortDev, cmd, arg, compat))
-+ return -EFAULT;
-+ }
-+ else
-+ {
-+ REPORT_ERROR(MINOR, E_INVALID_VALUE, ("minor"));
-+ return -ENODEV;
-+ }
-+
-+ return 0;
-+}
-+
-+#ifdef CONFIG_COMPAT
-+static long fm_compat_ioctl(struct file *file, unsigned int cmd, unsigned long arg)
-+{
-+ unsigned int minor = iminor(file->f_path.dentry->d_inode);
-+ long res;
-+
-+ fm_mutex_lock();
-+ res = fm_ioctls(minor, file, cmd, arg, true);
-+ fm_mutex_unlock();
-+
-+ return res;
-+}
-+#endif
-+
-+static long fm_ioctl(struct file *file, unsigned int cmd, unsigned long arg)
-+{
-+ unsigned int minor = iminor(file->f_path.dentry->d_inode);
-+ long res;
-+
-+ fm_mutex_lock();
-+ res = fm_ioctls(minor, file, cmd, arg, false);
-+ fm_mutex_unlock();
-+
-+ return res;
-+}
-+
-+/* Globals for FM character device */
-+struct file_operations fm_fops =
-+{
-+ .owner = THIS_MODULE,
-+ .unlocked_ioctl = fm_ioctl,
-+#ifdef CONFIG_COMPAT
-+ .compat_ioctl = fm_compat_ioctl,
-+#endif
-+ .open = fm_open,
-+ .release = fm_close,
-+};
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/src/wrapper/lnxwrp_ioctls_fm_compat.c
-@@ -0,0 +1,1297 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+/*
-+ @File lnxwrp_fm_compat_ioctls.c
-+
-+ @Description FM PCD compat functions
-+
-+*/
-+
-+#if !defined(CONFIG_COMPAT)
-+#error "missing COMPAT layer..."
-+#endif
-+
-+
-+#include <linux/kernel.h>
-+#include <linux/module.h>
-+#include <linux/fs.h>
-+#include <linux/cdev.h>
-+#include <linux/device.h>
-+#include <linux/irq.h>
-+#include <linux/interrupt.h>
-+#include <linux/io.h>
-+#include <linux/ioport.h>
-+#include <asm/uaccess.h>
-+#include <asm/errno.h>
-+#ifndef CONFIG_FMAN_ARM
-+#include <sysdev/fsl_soc.h>
-+#endif
-+
-+#include "part_ext.h"
-+#include "fm_ioctls.h"
-+#include "fm_pcd_ioctls.h"
-+#include "fm_port_ioctls.h"
-+#include "lnxwrp_ioctls_fm_compat.h"
-+
-+#if defined(FM_COMPAT_DBG)
-+static void hex_dump(void * p_addr, unsigned int size)
-+{
-+ int i;
-+
-+ for(i=0; i<size; i+=16)
-+ {
-+ printk("%p: 0x%08x 0x%08x 0x%08x 0x%08x\n", p_addr + i,
-+ *(unsigned int *)(p_addr + i),
-+ *(unsigned int *)(p_addr + i + 4),
-+ *(unsigned int *)(p_addr + i + 8),
-+ *(unsigned int *)(p_addr + i +12)
-+ );
-+ }
-+}
-+#endif
-+
-+/* maping kernel pointers w/ UserSpace id's { */
-+struct map_node {
-+ void *ptr;
-+ u8 node_type;
-+};
-+
-+static struct map_node compat_ptr2id_array[COMPAT_PTR2ID_ARRAY_MAX] = {{NULL},{FM_MAP_TYPE_UNSPEC}};
-+
-+void compat_del_ptr2id(void *p, enum fm_map_node_type node_type)
-+{
-+ compat_uptr_t k;
-+
-+ _fm_cpt_dbg(COMPAT_GENERIC, "delete (%p)\n", p);
-+
-+ for(k=1; k < COMPAT_PTR2ID_ARRAY_MAX; k++)
-+ if(compat_ptr2id_array[k].ptr == p){
-+ compat_ptr2id_array[k].ptr = NULL;
-+ compat_ptr2id_array[k].node_type = FM_MAP_TYPE_UNSPEC;
-+ }
-+}
-+EXPORT_SYMBOL(compat_del_ptr2id);
-+
-+compat_uptr_t compat_add_ptr2id(void *p, enum fm_map_node_type node_type)
-+{
-+ compat_uptr_t k;
-+
-+ _fm_cpt_dbg(COMPAT_GENERIC, " (%p) do ->\n", p);
-+
-+ if(!p)
-+ return 0;
-+
-+ for(k=1; k < COMPAT_PTR2ID_ARRAY_MAX; k++)
-+ if(compat_ptr2id_array[k].ptr == NULL)
-+ {
-+ compat_ptr2id_array[k].ptr = p;
-+ compat_ptr2id_array[k].node_type = node_type;
-+ _fm_cpt_dbg(COMPAT_GENERIC, "0x%08x \n", k | COMPAT_PTR2ID_WATERMARK);
-+ return k | COMPAT_PTR2ID_WATERMARK;
-+ }
-+
-+ printk(KERN_WARNING "FMan map list full! No more PCD space on kernel!\n");
-+ return 0;
-+}
-+EXPORT_SYMBOL(compat_add_ptr2id);
-+
-+compat_uptr_t compat_get_ptr2id(void *p, enum fm_map_node_type node_type)
-+{
-+ compat_uptr_t k;
-+
-+ _fm_cpt_dbg(COMPAT_GENERIC, " (%p) get -> \n", p);
-+
-+ for(k=1; k < COMPAT_PTR2ID_ARRAY_MAX; k++)
-+ if(compat_ptr2id_array[k].ptr == p &&
-+ compat_ptr2id_array[k].node_type == node_type) {
-+
-+ _fm_cpt_dbg(COMPAT_GENERIC, "0x%08x\n", k | COMPAT_PTR2ID_WATERMARK);
-+ return k | COMPAT_PTR2ID_WATERMARK;
-+ }
-+
-+ return 0;
-+}
-+EXPORT_SYMBOL(compat_get_ptr2id);
-+
-+void *compat_get_id2ptr(compat_uptr_t comp, enum fm_map_node_type node_type)
-+{
-+
-+ _fm_cpt_dbg(COMPAT_GENERIC, " (0x%08x) get -> \n", comp);
-+
-+ if((COMPAT_PTR2ID_WM_MASK & comp) != COMPAT_PTR2ID_WATERMARK) {
-+ _fm_cpt_dbg(COMPAT_GENERIC, "Error, invalid watermark (0x%08x)!\n\n", comp);
-+ dump_stack();
-+ return compat_ptr(comp);
-+ }
-+
-+ comp &= ~COMPAT_PTR2ID_WM_MASK;
-+
-+ if(((0 < comp) && (comp < COMPAT_PTR2ID_ARRAY_MAX) && (compat_ptr2id_array[comp].ptr != NULL)
-+ && compat_ptr2id_array[comp].node_type == node_type)) {
-+ _fm_cpt_dbg(COMPAT_GENERIC, "%p\n", compat_ptr2id_array[comp].ptr);
-+ return compat_ptr2id_array[comp].ptr;
-+ }
-+ return NULL;
-+}
-+EXPORT_SYMBOL(compat_get_id2ptr);
-+/* } maping kernel pointers w/ UserSpace id's */
-+
-+void compat_obj_delete(
-+ ioc_compat_fm_obj_t *compat_id,
-+ ioc_fm_obj_t *id)
-+{
-+ id->obj = compat_pcd_id2ptr(compat_id->obj);
-+ compat_del_ptr2id(id->obj, FM_MAP_TYPE_PCD_NODE);
-+}
-+
-+static inline void compat_copy_fm_pcd_plcr_next_engine(
-+ ioc_compat_fm_pcd_plcr_next_engine_params_u *compat_param,
-+ ioc_fm_pcd_plcr_next_engine_params_u *param,
-+ ioc_fm_pcd_engine next_engine,
-+ uint8_t compat)
-+{
-+ _fm_cpt_dbg (compat, " {->...\n");
-+
-+ switch (next_engine)
-+ {
-+ case e_IOC_FM_PCD_PLCR:
-+ if (compat == COMPAT_US_TO_K)
-+ param->p_profile = compat_pcd_id2ptr(compat_param->p_profile);
-+ else
-+ compat_param->p_profile = compat_pcd_ptr2id(param->p_profile);
-+ break;
-+ case e_IOC_FM_PCD_KG:
-+ if (compat == COMPAT_US_TO_K)
-+ param->p_direct_scheme = compat_pcd_id2ptr(compat_param->p_direct_scheme);
-+ else
-+ compat_param->p_direct_scheme = compat_pcd_ptr2id(param->p_direct_scheme);
-+ break;
-+ default:
-+ if (compat == COMPAT_US_TO_K)
-+ param->action = compat_param->action;
-+ else
-+ compat_param->action = param->action;
-+ break;
-+ }
-+
-+ _fm_cpt_dbg (compat, " ...->}\n");
-+}
-+
-+void compat_copy_fm_pcd_plcr_profile(
-+ ioc_compat_fm_pcd_plcr_profile_params_t *compat_param,
-+ ioc_fm_pcd_plcr_profile_params_t *param,
-+ uint8_t compat)
-+{
-+ _fm_cpt_dbg (compat, " {->...\n");
-+
-+ if (compat == COMPAT_US_TO_K)
-+ {
-+ param->modify = compat_param->modify;
-+
-+ /* profile_select */
-+ if (!compat_param->modify)
-+ {
-+ param->profile_select.new_params.profile_type =
-+ compat_param->profile_select.new_params.profile_type;
-+ param->profile_select.new_params.p_fm_port =
-+ compat_ptr(compat_param->profile_select.new_params.p_fm_port);
-+ param->profile_select.new_params.relative_profile_id =
-+ compat_param->profile_select.new_params.relative_profile_id;
-+ }
-+ else
-+ param->profile_select.p_profile =
-+ compat_pcd_id2ptr(compat_param->profile_select.p_profile);
-+
-+ param->alg_selection = compat_param->alg_selection;
-+ param->color_mode = compat_param->color_mode;
-+
-+ /* both parameters in the union has the same size, so memcpy works */
-+ memcpy(&param->color, &compat_param->color, sizeof(param->color));
-+
-+ memcpy(&param->non_passthrough_alg_param,
-+ &compat_param->non_passthrough_alg_param,
-+ sizeof(ioc_fm_pcd_plcr_non_passthrough_alg_param_t));
-+
-+ param->next_engine_on_green = compat_param->next_engine_on_green;
-+ param->next_engine_on_yellow = compat_param->next_engine_on_yellow;
-+ param->next_engine_on_red = compat_param->next_engine_on_red;
-+
-+ param->trap_profile_on_flow_A = compat_param->trap_profile_on_flow_A;
-+ param->trap_profile_on_flow_B = compat_param->trap_profile_on_flow_B;
-+ param->trap_profile_on_flow_C = compat_param->trap_profile_on_flow_C;
-+ }
-+ else
-+ {
-+ compat_param->modify = param->modify;
-+
-+ /* profile_select */
-+ if (!param->modify)
-+ {
-+ compat_param->profile_select.new_params.profile_type =
-+ param->profile_select.new_params.profile_type;
-+ compat_param->profile_select.new_params.p_fm_port =
-+ ptr_to_compat(param->profile_select.new_params.p_fm_port);
-+ compat_param->profile_select.new_params.relative_profile_id =
-+ param->profile_select.new_params.relative_profile_id;
-+ }
-+ else
-+ compat_param->profile_select.p_profile =
-+ compat_pcd_ptr2id(param->profile_select.p_profile);
-+
-+ compat_param->alg_selection = param->alg_selection;
-+ compat_param->color_mode = param->color_mode;
-+
-+ /* both parameters in the union has the same size, so memcpy works */
-+ memcpy(&compat_param->color, &param->color, sizeof(compat_param->color));
-+
-+ memcpy(&compat_param->non_passthrough_alg_param,
-+ &param->non_passthrough_alg_param,
-+ sizeof(ioc_fm_pcd_plcr_non_passthrough_alg_param_t));
-+
-+ compat_param->next_engine_on_green = param->next_engine_on_green;
-+ compat_param->next_engine_on_yellow = param->next_engine_on_yellow;
-+ compat_param->next_engine_on_red = param->next_engine_on_red;
-+
-+ compat_param->trap_profile_on_flow_A = param->trap_profile_on_flow_A;
-+ compat_param->trap_profile_on_flow_B = param->trap_profile_on_flow_B;
-+ compat_param->trap_profile_on_flow_C = param->trap_profile_on_flow_C;
-+
-+ compat_param->id = compat_add_ptr2id(param->id, FM_MAP_TYPE_PCD_NODE);
-+ }
-+
-+ compat_copy_fm_pcd_plcr_next_engine(&compat_param->params_on_green,
-+ &param->params_on_green, param->next_engine_on_green, compat);
-+
-+ compat_copy_fm_pcd_plcr_next_engine(&compat_param->params_on_yellow,
-+ &param->params_on_yellow, param->next_engine_on_yellow, compat);
-+
-+ compat_copy_fm_pcd_plcr_next_engine(&compat_param->params_on_red,
-+ &param->params_on_red, param->next_engine_on_red, compat);
-+
-+ _fm_cpt_dbg (compat, " ...->}\n");
-+}
-+
-+static inline void compat_copy_fm_pcd_cc_next_kg(
-+ ioc_compat_fm_pcd_cc_next_kg_params_t *compat_param,
-+ ioc_fm_pcd_cc_next_kg_params_t *param,
-+ uint8_t compat)
-+{
-+ _fm_cpt_dbg (compat, " {->...\n");
-+
-+ if (compat == COMPAT_US_TO_K)
-+ {
-+ param->new_fqid = compat_param->new_fqid;
-+ param->override_fqid = compat_param->override_fqid;
-+#if DPAA_VERSION >= 11
-+ param->new_relative_storage_profile_id = compat_param->new_relative_storage_profile_id;
-+#endif
-+ param->p_direct_scheme = compat_pcd_id2ptr(compat_param->p_direct_scheme);
-+ }
-+ else
-+ {
-+ compat_param->new_fqid = param->new_fqid;
-+ compat_param->override_fqid = param->override_fqid;
-+#if DPAA_VERSION >= 11
-+ compat_param->new_relative_storage_profile_id = param->new_relative_storage_profile_id;
-+#endif
-+ compat_param->p_direct_scheme = compat_pcd_ptr2id(param->p_direct_scheme);
-+ }
-+
-+ _fm_cpt_dbg (compat, " ...->}\n");
-+}
-+
-+static inline void compat_copy_fm_pcd_cc_next_cc(
-+ ioc_compat_fm_pcd_cc_next_cc_params_t *compat_param,
-+ ioc_fm_pcd_cc_next_cc_params_t *param,
-+ uint8_t compat)
-+{
-+ _fm_cpt_dbg (compat, " {->...\n");
-+
-+ if (compat == COMPAT_US_TO_K)
-+ param->cc_node_id = compat_pcd_id2ptr(compat_param->cc_node_id);
-+ else
-+ compat_param->cc_node_id = compat_pcd_ptr2id(param->cc_node_id);
-+
-+ _fm_cpt_dbg (compat, " ...->}\n");
-+}
-+
-+static inline void compat_copy_fm_pcd_cc_next_engine(
-+ ioc_compat_fm_pcd_cc_next_engine_params_t *compat_param,
-+ ioc_fm_pcd_cc_next_engine_params_t *param,
-+ uint8_t compat)
-+{
-+ _fm_cpt_dbg (compat, " {->...\n");
-+
-+ if (compat == COMPAT_US_TO_K)
-+ {
-+ param->next_engine = compat_param->next_engine;
-+ if (param->next_engine != e_IOC_FM_PCD_INVALID )
-+ _fm_cpt_dbg(compat, " param->next_engine = %i \n", param->next_engine);
-+
-+ switch (param->next_engine)
-+ {
-+#if DPAA_VERSION >= 11
-+ case e_IOC_FM_PCD_FR:
-+ param->params.fr_params.frm_replic_id = compat_pcd_id2ptr(compat_param->params.fr_params.frm_replic_id);
-+ break;
-+#endif /* DPAA_VERSION >= 11 */
-+ case e_IOC_FM_PCD_CC:
-+ param->manip_id = compat_pcd_id2ptr(compat_param->manip_id);
-+ compat_copy_fm_pcd_cc_next_cc(&compat_param->params.cc_params, &param->params.cc_params, compat);
-+ break;
-+ case e_IOC_FM_PCD_KG:
-+ param->manip_id = compat_pcd_id2ptr(compat_param->manip_id);
-+ compat_copy_fm_pcd_cc_next_kg(&compat_param->params.kg_params, &param->params.kg_params, compat);
-+ break;
-+ case e_IOC_FM_PCD_DONE:
-+ case e_IOC_FM_PCD_PLCR:
-+ param->manip_id = compat_pcd_id2ptr(compat_param->manip_id);
-+ default:
-+ memcpy(&param->params, &compat_param->params, sizeof(param->params));
-+ }
-+ param->statistics_en = compat_param->statistics_en;
-+ }
-+ else
-+ {
-+ compat_param->next_engine = param->next_engine;
-+
-+ switch (compat_param->next_engine)
-+ {
-+#if DPAA_VERSION >= 11
-+ case e_IOC_FM_PCD_FR:
-+ compat_param->params.fr_params.frm_replic_id = compat_pcd_ptr2id(param->params.fr_params.frm_replic_id);
-+ break;
-+#endif /* DPAA_VERSION >= 11 */
-+ case e_IOC_FM_PCD_CC:
-+ compat_param->manip_id = compat_pcd_ptr2id(param->manip_id);
-+ compat_copy_fm_pcd_cc_next_cc(&compat_param->params.cc_params, &param->params.cc_params, compat);
-+ break;
-+ case e_IOC_FM_PCD_KG:
-+ compat_param->manip_id = compat_pcd_ptr2id(param->manip_id);
-+ compat_copy_fm_pcd_cc_next_kg(&compat_param->params.kg_params, &param->params.kg_params, compat);
-+ break;
-+ case e_IOC_FM_PCD_DONE:
-+ case e_IOC_FM_PCD_PLCR:
-+ compat_param->manip_id = compat_pcd_ptr2id(param->manip_id);
-+ default:
-+ memcpy(&compat_param->params, &param->params, sizeof(compat_param->params));
-+ }
-+ compat_param->statistics_en = param->statistics_en;
-+ }
-+
-+ _fm_cpt_dbg (compat, " ...->}\n");
-+}
-+
-+void compat_copy_fm_pcd_cc_key(
-+ ioc_compat_fm_pcd_cc_key_params_t *compat_param,
-+ ioc_fm_pcd_cc_key_params_t *param,
-+ uint8_t compat)
-+{
-+ if (compat == COMPAT_US_TO_K)
-+ {
-+ param->p_key = compat_ptr(compat_param->p_key);
-+ param->p_mask = compat_ptr(compat_param->p_mask);
-+ }
-+ else
-+ {
-+ compat_param->p_key = ptr_to_compat(param->p_key);
-+ compat_param->p_mask = ptr_to_compat(param->p_mask);
-+ }
-+
-+ compat_copy_fm_pcd_cc_next_engine(
-+ &compat_param->cc_next_engine_params,
-+ &param->cc_next_engine_params,
-+ compat);
-+}
-+
-+void compat_copy_fm_pcd_cc_node_modify_key_and_next_engine(
-+ ioc_compat_fm_pcd_cc_node_modify_key_and_next_engine_params_t *compat_param,
-+ ioc_fm_pcd_cc_node_modify_key_and_next_engine_params_t *param,
-+ uint8_t compat)
-+{
-+ if (compat == COMPAT_US_TO_K)
-+ {
-+ param->id = compat_pcd_id2ptr(compat_param->id);
-+ param->key_indx = compat_param->key_indx;
-+ param->key_size = compat_param->key_size;
-+ compat_copy_fm_pcd_cc_key(
-+ &compat_param->key_params,
-+ &param->key_params,
-+ compat);
-+ }
-+ else
-+ {
-+ compat_param->id = compat_pcd_ptr2id(param->id);
-+ compat_param->key_indx = param->key_indx;
-+ compat_param->key_size = param->key_size;
-+ compat_copy_fm_pcd_cc_key(
-+ &compat_param->key_params,
-+ &param->key_params,
-+ compat);
-+ }
-+}
-+
-+void compat_copy_fm_pcd_cc_node_modify_next_engine(
-+ ioc_compat_fm_pcd_cc_node_modify_next_engine_params_t *compat_param,
-+ ioc_fm_pcd_cc_node_modify_next_engine_params_t *param,
-+ uint8_t compat)
-+{
-+ if (compat == COMPAT_US_TO_K)
-+ {
-+ param->id = compat_pcd_id2ptr(compat_param->id);
-+ param->key_indx = compat_param->key_indx;
-+ param->key_size = compat_param->key_size;
-+ }
-+ else
-+ {
-+ compat_param->id = compat_pcd_ptr2id(param->id);
-+ compat_param->key_indx = param->key_indx;
-+ compat_param->key_size = param->key_size;
-+ }
-+
-+ compat_copy_fm_pcd_cc_next_engine(
-+ &compat_param->cc_next_engine_params,
-+ &param->cc_next_engine_params,
-+ compat);
-+}
-+
-+void compat_fm_pcd_cc_tree_modify_next_engine(
-+ ioc_compat_fm_pcd_cc_tree_modify_next_engine_params_t *compat_param,
-+ ioc_fm_pcd_cc_tree_modify_next_engine_params_t *param,
-+ uint8_t compat)
-+{
-+ if (compat == COMPAT_US_TO_K)
-+ {
-+ param->id = compat_pcd_id2ptr(compat_param->id);
-+ param->grp_indx = compat_param->grp_indx;
-+ param->indx = compat_param->indx;
-+ }
-+ else
-+ {
-+ compat_param->id = compat_pcd_ptr2id(param->id);
-+ compat_param->grp_indx = param->grp_indx;
-+ compat_param->indx = param->indx;
-+ }
-+
-+ compat_copy_fm_pcd_cc_next_engine(
-+ &compat_param->cc_next_engine_params,
-+ &param->cc_next_engine_params,
-+ compat);
-+}
-+
-+void compat_copy_fm_pcd_hash_table(
-+ ioc_compat_fm_pcd_hash_table_params_t *compat_param,
-+ ioc_fm_pcd_hash_table_params_t *param,
-+ uint8_t compat)
-+{
-+ if (compat == COMPAT_US_TO_K)
-+ {
-+ param->max_num_of_keys = compat_param->max_num_of_keys;
-+ param->statistics_mode = compat_param->statistics_mode;
-+ param->kg_hash_shift = compat_param->kg_hash_shift;
-+ param->hash_res_mask = compat_param->hash_res_mask;
-+ param->hash_shift = compat_param->hash_shift;
-+ param->match_key_size = compat_param->match_key_size;
-+ param->id = compat_pcd_id2ptr(compat_param->id);
-+ }
-+ else
-+ {
-+ compat_param->max_num_of_keys = param->max_num_of_keys;
-+ compat_param->statistics_mode = param->statistics_mode;
-+ compat_param->kg_hash_shift = param->kg_hash_shift;
-+ compat_param->hash_res_mask = param->hash_res_mask;
-+ compat_param->hash_shift = param->hash_shift;
-+ compat_param->match_key_size = param->match_key_size;
-+
-+ compat_param->id = compat_add_ptr2id(param->id, FM_MAP_TYPE_PCD_NODE);
-+ }
-+
-+ compat_copy_fm_pcd_cc_next_engine(
-+ &compat_param->cc_next_engine_params_for_miss,
-+ &param->cc_next_engine_params_for_miss,
-+ compat);
-+}
-+
-+void compat_copy_fm_pcd_cc_grp(
-+ ioc_compat_fm_pcd_cc_grp_params_t *compat_param,
-+ ioc_fm_pcd_cc_grp_params_t *param,
-+ uint8_t compat)
-+{
-+ int k;
-+
-+ _fm_cpt_dbg (compat, " {->...\n");
-+
-+ if (compat == COMPAT_US_TO_K)
-+ {
-+ param->num_of_distinction_units = compat_param->num_of_distinction_units;
-+ memcpy(param->unit_ids, compat_param->unit_ids, IOC_FM_PCD_MAX_NUM_OF_CC_UNITS);
-+ }
-+ else
-+ {
-+ compat_param->num_of_distinction_units = param->num_of_distinction_units;
-+ memcpy(compat_param->unit_ids, param->unit_ids, IOC_FM_PCD_MAX_NUM_OF_CC_UNITS);
-+ }
-+
-+ for (k=0; k < IOC_FM_PCD_MAX_NUM_OF_CC_ENTRIES_IN_GRP; k++)
-+ compat_copy_fm_pcd_cc_next_engine(
-+ &compat_param->next_engine_per_entries_in_grp[k],
-+ &param->next_engine_per_entries_in_grp[k],
-+ compat);
-+
-+ _fm_cpt_dbg (compat, " ...->}\n");
-+}
-+
-+void compat_copy_fm_pcd_cc_tree(
-+ ioc_compat_fm_pcd_cc_tree_params_t *compat_param,
-+ ioc_fm_pcd_cc_tree_params_t *param,
-+ uint8_t compat)
-+{
-+ int k;
-+ _fm_cpt_dbg (compat, " {->...\n");
-+
-+ if (compat == COMPAT_US_TO_K)
-+ {
-+ param->net_env_id = compat_pcd_id2ptr(compat_param->net_env_id);
-+ param->num_of_groups = compat_param->num_of_groups;
-+ }
-+ else
-+ {
-+ compat_param->net_env_id = compat_pcd_ptr2id(param->net_env_id);
-+ compat_param->num_of_groups = param->num_of_groups;
-+
-+ compat_param->id = compat_add_ptr2id(param->id, FM_MAP_TYPE_PCD_NODE);
-+ }
-+
-+ for (k=0; k < IOC_FM_PCD_MAX_NUM_OF_CC_GROUPS; k++)
-+ compat_copy_fm_pcd_cc_grp(
-+ &compat_param->fm_pcd_cc_group_params[k],
-+ &param->fm_pcd_cc_group_params[k],
-+ compat);
-+
-+ _fm_cpt_dbg (compat, " ...->}\n");
-+}
-+
-+void compat_fm_pcd_prs_sw(
-+ ioc_compat_fm_pcd_prs_sw_params_t *compat_param,
-+ ioc_fm_pcd_prs_sw_params_t *param,
-+ uint8_t compat)
-+{
-+ if (compat == COMPAT_US_TO_K)
-+ {
-+ param->override = compat_param->override;
-+ param->size = compat_param->size;
-+ param->base = compat_param->base;
-+ param->p_code = compat_ptr(compat_param->p_code);
-+ memcpy(param->sw_prs_data_params,compat_param->sw_prs_data_params,IOC_FM_PCD_PRS_NUM_OF_HDRS*sizeof(uint32_t));
-+ param->num_of_labels = compat_param->num_of_labels;
-+ memcpy(param->labels_table,compat_param->labels_table,IOC_FM_PCD_PRS_NUM_OF_LABELS*sizeof(ioc_fm_pcd_prs_label_params_t));
-+ }
-+}
-+
-+void compat_copy_fm_pcd_kg_scheme(
-+ ioc_compat_fm_pcd_kg_scheme_params_t *compat_param,
-+ ioc_fm_pcd_kg_scheme_params_t *param,
-+ uint8_t compat)
-+{
-+ _fm_cpt_dbg(compat," {->...\n");
-+
-+ if (compat == COMPAT_US_TO_K)
-+ {
-+ param->modify = compat_param->modify;
-+
-+ /* scm_id */
-+ if (compat_param->modify)
-+ {
-+ param->scm_id.scheme_id = compat_pcd_id2ptr(compat_param->scm_id.scheme_id);
-+ _fm_cpt_dbg(compat," param->scm_id.scheme_id = %p \n", param->scm_id.scheme_id);
-+ }
-+ else
-+ param->scm_id.relative_scheme_id = compat_param->scm_id.relative_scheme_id;
-+
-+ param->always_direct = compat_param->always_direct;
-+ /* net_env_params */
-+ param->net_env_params.net_env_id = compat_pcd_id2ptr(compat_param->net_env_params.net_env_id);
-+ param->net_env_params.num_of_distinction_units = compat_param->net_env_params.num_of_distinction_units;
-+ memcpy(param->net_env_params.unit_ids,
-+ compat_param->net_env_params.unit_ids,
-+ IOC_FM_PCD_MAX_NUM_OF_DISTINCTION_UNITS);
-+
-+ param->use_hash = compat_param->use_hash;
-+ memcpy(&param->key_extract_and_hash_params,
-+ &compat_param->key_extract_and_hash_params,
-+ sizeof(ioc_fm_pcd_kg_key_extract_and_hash_params_t));
-+ param->bypass_fqid_generation = compat_param->bypass_fqid_generation;
-+ param->base_fqid = compat_param->base_fqid;
-+#if DPAA_VERSION >= 11
-+ param->override_storage_profile =
-+ compat_param->override_storage_profile;
-+ param->storage_profile = compat_param->storage_profile;
-+#endif
-+ param->num_of_used_extracted_ors = compat_param->num_of_used_extracted_ors;
-+ memcpy(param->extracted_ors,
-+ compat_param->extracted_ors,
-+ IOC_FM_PCD_KG_NUM_OF_GENERIC_REGS * sizeof(ioc_fm_pcd_kg_extracted_or_params_t));
-+ param->next_engine = compat_param->next_engine;
-+
-+ /* kg_next_engine_params */
-+ if (param->next_engine == e_IOC_FM_PCD_CC)
-+ {
-+ param->kg_next_engine_params.cc.tree_id = compat_pcd_id2ptr(compat_param->kg_next_engine_params.cc.tree_id);
-+ param->kg_next_engine_params.cc.grp_id = compat_param->kg_next_engine_params.cc.grp_id;
-+ param->kg_next_engine_params.cc.plcr_next = compat_param->kg_next_engine_params.cc.plcr_next;
-+ param->kg_next_engine_params.cc.bypass_plcr_profile_generation
-+ = compat_param->kg_next_engine_params.cc.bypass_plcr_profile_generation;
-+ memcpy(&param->kg_next_engine_params.cc.plcr_profile,
-+ &compat_param->kg_next_engine_params.cc.plcr_profile,
-+ sizeof(ioc_fm_pcd_kg_plcr_profile_t));
-+ }
-+ else
-+ memcpy(&param->kg_next_engine_params,
-+ &compat_param->kg_next_engine_params,
-+ sizeof(param->kg_next_engine_params));
-+
-+ memcpy(&param->scheme_counter,
-+ &compat_param->scheme_counter,
-+ sizeof(ioc_fm_pcd_kg_scheme_counter_t));
-+ }
-+ else
-+ {
-+ compat_param->modify = param->modify;
-+
-+ /* scm_id */
-+ if (param->modify)
-+ compat_param->scm_id.scheme_id = compat_pcd_ptr2id(param->scm_id.scheme_id);
-+ else
-+ compat_param->scm_id.relative_scheme_id = param->scm_id.relative_scheme_id;
-+
-+ compat_param->always_direct = param->always_direct;
-+
-+ /* net_env_params */
-+ compat_param->net_env_params.net_env_id = compat_pcd_ptr2id(param->net_env_params.net_env_id);
-+ compat_param->net_env_params.num_of_distinction_units = param->net_env_params.num_of_distinction_units;
-+ memcpy(compat_param->net_env_params.unit_ids, param->net_env_params.unit_ids, IOC_FM_PCD_MAX_NUM_OF_DISTINCTION_UNITS);
-+
-+ compat_param->use_hash = param->use_hash;
-+ memcpy(&compat_param->key_extract_and_hash_params, &param->key_extract_and_hash_params, sizeof(ioc_fm_pcd_kg_key_extract_and_hash_params_t));
-+ compat_param->bypass_fqid_generation = param->bypass_fqid_generation;
-+ compat_param->base_fqid = param->base_fqid;
-+#if DPAA_VERSION >= 11
-+ compat_param->override_storage_profile =
-+ param->override_storage_profile;
-+ compat_param->storage_profile = param->storage_profile;
-+#endif
-+ compat_param->num_of_used_extracted_ors = param->num_of_used_extracted_ors;
-+ memcpy(compat_param->extracted_ors, param->extracted_ors, IOC_FM_PCD_KG_NUM_OF_GENERIC_REGS * sizeof(ioc_fm_pcd_kg_extracted_or_params_t));
-+ compat_param->next_engine = param->next_engine;
-+
-+ /* kg_next_engine_params */
-+ if (compat_param->next_engine == e_IOC_FM_PCD_CC)
-+ {
-+ compat_param->kg_next_engine_params.cc.tree_id = compat_pcd_ptr2id(param->kg_next_engine_params.cc.tree_id);
-+ compat_param->kg_next_engine_params.cc.grp_id = param->kg_next_engine_params.cc.grp_id;
-+ compat_param->kg_next_engine_params.cc.plcr_next = param->kg_next_engine_params.cc.plcr_next;
-+ compat_param->kg_next_engine_params.cc.bypass_plcr_profile_generation
-+ = param->kg_next_engine_params.cc.bypass_plcr_profile_generation;
-+ memcpy(&compat_param->kg_next_engine_params.cc.plcr_profile,
-+ &param->kg_next_engine_params.cc.plcr_profile,
-+ sizeof(ioc_fm_pcd_kg_plcr_profile_t));
-+ }
-+ else
-+ memcpy(&param->kg_next_engine_params, &compat_param->kg_next_engine_params, sizeof(compat_param->kg_next_engine_params));
-+
-+ memcpy(&compat_param->scheme_counter, &param->scheme_counter, sizeof(ioc_fm_pcd_kg_scheme_counter_t));
-+
-+ compat_param->id = compat_add_ptr2id(param->id, FM_MAP_TYPE_PCD_NODE);
-+ }
-+
-+ _fm_cpt_dbg(compat," ...->}\n");
-+}
-+
-+void compat_copy_fm_pcd_kg_scheme_spc(
-+ ioc_compat_fm_pcd_kg_scheme_spc_t *compat_param,
-+ ioc_fm_pcd_kg_scheme_spc_t *param,
-+ uint8_t compat)
-+{
-+ if (compat == COMPAT_US_TO_K)
-+ {
-+ param->id = compat_pcd_id2ptr(compat_param->id);
-+ param->val = compat_param->val;
-+ } else {
-+ compat_param->id = compat_pcd_ptr2id(param->id);
-+ compat_param->val = param->val;
-+ }
-+}
-+
-+
-+void compat_copy_fm_pcd_kg_scheme_select(
-+ ioc_compat_fm_pcd_kg_scheme_select_t *compat_param,
-+ ioc_fm_pcd_kg_scheme_select_t *param,
-+ uint8_t compat)
-+{
-+ if (compat == COMPAT_US_TO_K)
-+ {
-+ param->direct = compat_param->direct;
-+ if (param->direct)
-+ param->scheme_id = compat_pcd_id2ptr(compat_param->scheme_id);
-+ }
-+}
-+
-+void compat_copy_fm_pcd_kg_schemes_params(
-+ ioc_compat_fm_pcd_port_schemes_params_t *compat_param,
-+ ioc_fm_pcd_port_schemes_params_t *param,
-+ uint8_t compat)
-+{
-+ int k;
-+
-+ if (compat == COMPAT_US_TO_K) {
-+ param->num_of_schemes = compat_param->num_of_schemes;
-+ for(k=0; k < compat_param->num_of_schemes; k++)
-+ param->scheme_ids[k] = compat_pcd_id2ptr(compat_param->scheme_ids[k]);
-+ }
-+}
-+
-+void compat_copy_fm_port_pcd_cc(
-+ ioc_compat_fm_port_pcd_cc_params_t *compat_cc_params ,
-+ ioc_fm_port_pcd_cc_params_t *p_cc_params,
-+ uint8_t compat)
-+{
-+ if (compat == COMPAT_US_TO_K){
-+ p_cc_params->cc_tree_id = compat_pcd_id2ptr(compat_cc_params->cc_tree_id);
-+ }
-+}
-+
-+void compat_copy_fm_port_pcd_kg(
-+ ioc_compat_fm_port_pcd_kg_params_t *compat_param,
-+ ioc_fm_port_pcd_kg_params_t *param,
-+ uint8_t compat)
-+{
-+ if (compat == COMPAT_US_TO_K){
-+ uint8_t k;
-+
-+ param->num_of_schemes = compat_param->num_of_schemes;
-+ for(k=0; k<compat_param->num_of_schemes; k++)
-+ param->scheme_ids[k] = compat_pcd_id2ptr(compat_param->scheme_ids[k]);
-+
-+ param->direct_scheme = compat_param->direct_scheme;
-+ if (param->direct_scheme)
-+ param->direct_scheme_id = compat_pcd_id2ptr(compat_param->direct_scheme_id);
-+ }
-+}
-+
-+void compat_copy_fm_port_pcd(
-+ ioc_compat_fm_port_pcd_params_t *compat_param,
-+ ioc_fm_port_pcd_params_t *param,
-+ uint8_t compat)
-+{
-+ if (compat == COMPAT_US_TO_K)
-+ {
-+ ioc_fm_port_pcd_prs_params_t *same_port_pcd_prs_params;
-+ ioc_compat_fm_port_pcd_cc_params_t *compat_port_pcd_cc_params;
-+ ioc_compat_fm_port_pcd_kg_params_t *compat_port_pcd_kg_params;
-+ ioc_compat_fm_port_pcd_plcr_params_t *compat_port_pcd_plcr_params;
-+
-+ same_port_pcd_prs_params = (ioc_fm_port_pcd_prs_params_t *) (compat_param + 1);
-+ compat_port_pcd_cc_params = (ioc_compat_fm_port_pcd_cc_params_t *) (same_port_pcd_prs_params + 1);
-+ compat_port_pcd_kg_params = (ioc_compat_fm_port_pcd_kg_params_t *) (compat_port_pcd_cc_params + 1);
-+ compat_port_pcd_plcr_params = (ioc_compat_fm_port_pcd_plcr_params_t *) (compat_port_pcd_kg_params + 1);
-+
-+ _fm_cpt_dbg(compat,"\n param->p_prs_params=%p \n", param->p_prs_params);
-+ _fm_cpt_dbg(compat," param->p_cc_params=%p \n", param->p_cc_params);
-+ _fm_cpt_dbg(compat," param->p_kg_params=%p \n", param->p_kg_params);
-+ _fm_cpt_dbg(compat," param->p_plcr_params=%p \n", param->p_plcr_params);
-+ _fm_cpt_dbg(compat," param->p_ip_reassembly_manip=%p \n", param->p_ip_reassembly_manip);
-+#if (DPAA_VERSION >= 11)
-+ _fm_cpt_dbg(compat," param->p_capwap_reassembly_manip=%p \n", param->p_capwap_reassembly_manip);
-+#endif
-+ param->pcd_support = compat_param->pcd_support;
-+ param->net_env_id = compat_pcd_id2ptr(compat_param->net_env_id);
-+
-+ if (param->p_cc_params)
-+ compat_copy_fm_port_pcd_cc(compat_port_pcd_cc_params, param->p_cc_params, COMPAT_US_TO_K);
-+ if (param->p_kg_params)
-+ compat_copy_fm_port_pcd_kg(compat_port_pcd_kg_params, param->p_kg_params, COMPAT_US_TO_K);
-+ if (param->p_plcr_params)
-+ param->p_plcr_params->plcr_profile_id = compat_pcd_id2ptr(compat_port_pcd_plcr_params->plcr_profile_id);
-+ param->p_ip_reassembly_manip = compat_pcd_id2ptr(compat_param->p_ip_reassembly_manip);
-+#if (DPAA_VERSION >= 11)
-+ param->p_capwap_reassembly_manip = compat_pcd_id2ptr(compat_param->p_capwap_reassembly_manip);
-+#endif
-+ }
-+}
-+
-+void compat_copy_fm_port_pcd_modify_tree(
-+ ioc_compat_fm_obj_t *compat_id,
-+ ioc_fm_obj_t *id,
-+ uint8_t compat)
-+{
-+ if (compat == COMPAT_US_TO_K)
-+ id->obj = compat_pcd_id2ptr(compat_id->obj);
-+}
-+
-+#if (DPAA_VERSION >= 11)
-+void compat_copy_fm_port_vsp_alloc_params(
-+ ioc_compat_fm_port_vsp_alloc_params_t *compat_param,
-+ ioc_fm_port_vsp_alloc_params_t *param,
-+ uint8_t compat)
-+{
-+ if (compat == COMPAT_US_TO_K)
-+ {
-+ _fm_cpt_dbg(compat," param->p_fm_tx_port=%p \n", param->p_fm_tx_port);
-+
-+ param->dflt_relative_id = compat_param->dflt_relative_id;
-+ param->num_of_profiles = compat_param->num_of_profiles;
-+ param->p_fm_tx_port = compat_pcd_id2ptr(compat_param->p_fm_tx_port);
-+ }
-+}
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+void compat_copy_fm_pcd_cc_tbl_get_stats(
-+ ioc_compat_fm_pcd_cc_tbl_get_stats_t *compat_param,
-+ ioc_fm_pcd_cc_tbl_get_stats_t *param,
-+ uint8_t compat)
-+{
-+ if (compat == COMPAT_US_TO_K)
-+ {
-+ param->id = compat_pcd_id2ptr(compat_param->id);
-+ param->key_index = compat_param->key_index;
-+ memcpy(&param->statistics, &compat_param->statistics, sizeof(ioc_fm_pcd_cc_key_statistics_t));
-+ } else {
-+ compat_param->id = compat_add_ptr2id(param->id, FM_MAP_TYPE_PCD_NODE);
-+ compat_param->key_index = param->key_index;
-+ memcpy(&compat_param->statistics, &param->statistics, sizeof(ioc_fm_pcd_cc_key_statistics_t));
-+ }
-+}
-+
-+
-+void compat_copy_fm_pcd_net_env(
-+ ioc_compat_fm_pcd_net_env_params_t *compat_param,
-+ ioc_fm_pcd_net_env_params_t *param,
-+ uint8_t compat)
-+{
-+ if (compat == COMPAT_US_TO_K)
-+ {
-+ param->num_of_distinction_units = compat_param->num_of_distinction_units;
-+ memcpy(param->units, compat_param->units, sizeof(ioc_fm_pcd_distinction_unit_t)*IOC_FM_PCD_MAX_NUM_OF_DISTINCTION_UNITS);
-+ param->id = NULL; /* to avoid passing garbage to the kernel */
-+ }
-+ else
-+ {
-+ compat_param->num_of_distinction_units = param->num_of_distinction_units;
-+ memcpy(compat_param->units, param->units, sizeof(ioc_fm_pcd_distinction_unit_t)*IOC_FM_PCD_MAX_NUM_OF_DISTINCTION_UNITS);
-+
-+ compat_param->id = compat_add_ptr2id(param->id, FM_MAP_TYPE_PCD_NODE);
-+ }
-+}
-+
-+void compat_copy_fm_pcd_cc_node_modify_key(
-+ ioc_compat_fm_pcd_cc_node_modify_key_params_t *compat_param,
-+ ioc_fm_pcd_cc_node_modify_key_params_t *param,
-+ uint8_t compat)
-+{
-+ if (compat == COMPAT_US_TO_K)
-+ {
-+ param->key_indx = compat_param->key_indx;
-+ param->key_size = compat_param->key_size;
-+ param->p_key = (uint8_t *)compat_ptr(compat_param->p_key);
-+ _fm_cpt_dbg(compat," param->p_key = %p \n", param->p_key);
-+ param->p_mask = (uint8_t *)compat_ptr(compat_param->p_mask);
-+ _fm_cpt_dbg(compat," param->p_mask = %p\n", param->p_mask);
-+ param->id = compat_pcd_id2ptr(compat_param->id);
-+ _fm_cpt_dbg(compat," param->id = %p \n", param->id);
-+ }
-+ else
-+ {
-+ compat_param->key_indx = param->key_indx;
-+ compat_param->key_size = param->key_size;
-+ compat_param->p_key = ptr_to_compat((void *)param->p_key);
-+ compat_param->p_mask = ptr_to_compat((void *)param->p_mask);
-+
-+ compat_param->id = compat_add_ptr2id(param->id, FM_MAP_TYPE_PCD_NODE);
-+ }
-+}
-+
-+void compat_copy_keys(
-+ ioc_compat_keys_params_t *compat_param,
-+ ioc_keys_params_t *param,
-+ uint8_t compat)
-+{
-+ int k = 0;
-+
-+ _fm_cpt_dbg(compat," {->...\n");
-+
-+ if (compat == COMPAT_US_TO_K) {
-+ param->max_num_of_keys = compat_param->max_num_of_keys;
-+ param->mask_support = compat_param->mask_support;
-+ param->statistics_mode = compat_param->statistics_mode;
-+ param->num_of_keys = compat_param->num_of_keys;
-+ param->key_size = compat_param->key_size;
-+#if (DPAA_VERSION >= 11)
-+ memcpy(&param->frame_length_ranges,
-+ &compat_param->frame_length_ranges,
-+ sizeof(param->frame_length_ranges[0]) *
-+ IOC_FM_PCD_CC_STATS_MAX_NUM_OF_FLR);
-+#endif /* (DPAA_VERSION >= 11) */
-+ }
-+ else {
-+ compat_param->max_num_of_keys = param->max_num_of_keys;
-+ compat_param->mask_support = param->mask_support;
-+ compat_param->statistics_mode = param->statistics_mode;
-+ compat_param->num_of_keys = param->num_of_keys;
-+ compat_param->key_size = param->key_size;
-+#if (DPAA_VERSION >= 11)
-+ memcpy(&compat_param->frame_length_ranges,
-+ &param->frame_length_ranges,
-+ sizeof(compat_param->frame_length_ranges[0]) *
-+ IOC_FM_PCD_CC_STATS_MAX_NUM_OF_FLR);
-+#endif /* (DPAA_VERSION >= 11) */
-+ }
-+
-+ for (k=0; k < IOC_FM_PCD_MAX_NUM_OF_KEYS; k++)
-+ compat_copy_fm_pcd_cc_key(
-+ &compat_param->key_params[k],
-+ &param->key_params[k],
-+ compat);
-+
-+ compat_copy_fm_pcd_cc_next_engine(
-+ &compat_param->cc_next_engine_params_for_miss,
-+ &param->cc_next_engine_params_for_miss,
-+ compat);
-+
-+ _fm_cpt_dbg(compat," ...->}\n");
-+}
-+
-+void compat_copy_fm_pcd_cc_node(
-+ ioc_compat_fm_pcd_cc_node_params_t *compat_param,
-+ ioc_fm_pcd_cc_node_params_t *param,
-+ uint8_t compat)
-+{
-+ _fm_cpt_dbg(compat," {->...\n");
-+
-+ if (compat == COMPAT_US_TO_K)
-+ memcpy(&param->extract_cc_params, &compat_param->extract_cc_params, sizeof(ioc_fm_pcd_extract_entry_t));
-+
-+ else
-+ {
-+ compat_copy_keys(&compat_param->keys_params, &param->keys_params, compat);
-+
-+ compat_param->id = compat_add_ptr2id(param->id, FM_MAP_TYPE_PCD_NODE);
-+ _fm_cpt_dbg(compat," param->id = %p \n", param->id);
-+ }
-+
-+ compat_copy_keys(&compat_param->keys_params, &param->keys_params, compat);
-+
-+ _fm_cpt_dbg(compat," ...->}\n");
-+}
-+
-+void compat_fm_pcd_manip_set_node(
-+ ioc_compat_fm_pcd_manip_params_t *compat_param,
-+ ioc_fm_pcd_manip_params_t *param,
-+ uint8_t compat)
-+{
-+ if (compat == COMPAT_US_TO_K) {
-+ param->type = compat_param->type;
-+ switch (param->type) {
-+ case e_IOC_FM_PCD_MANIP_HDR:
-+ param->u.hdr.rmv = compat_param->u.hdr.rmv;
-+ memcpy(&param->u.hdr.rmv_params,
-+ &compat_param->u.hdr.rmv_params,
-+ sizeof(param->u.hdr.rmv_params));
-+
-+ param->u.hdr.insrt = compat_param->u.hdr.insrt;
-+ param->u.hdr.insrt_params.type =
-+ compat_param->u.hdr.insrt_params.type;
-+ switch (compat_param->u.hdr.insrt_params.type)
-+ {
-+ case e_IOC_FM_PCD_MANIP_INSRT_GENERIC:
-+ param->u.hdr.insrt_params.u.generic.offset =
-+ compat_param->u.hdr.insrt_params.u.generic.offset;
-+ param->u.hdr.insrt_params.u.generic.size =
-+ compat_param->u.hdr.insrt_params.u.generic.size;
-+ param->u.hdr.insrt_params.u.generic.replace =
-+ compat_param->u.hdr.insrt_params.u.generic.replace;
-+ param->u.hdr.insrt_params.u.generic.p_data =
-+ compat_ptr(compat_param->u.hdr.insrt_params.u.generic.p_data);
-+ break;
-+ case e_IOC_FM_PCD_MANIP_INSRT_BY_HDR:
-+ param->u.hdr.insrt_params.u.by_hdr.type =
-+ compat_param->u.hdr.insrt_params.u.by_hdr.type;
-+ param->u.hdr.insrt_params.u.by_hdr.u.specific_l2_params.specific_l2 =
-+ compat_param->u.hdr.insrt_params.u.by_hdr.u.specific_l2_params.specific_l2;
-+ param->u.hdr.insrt_params.u.by_hdr.u.specific_l2_params.update =
-+ compat_param->u.hdr.insrt_params.u.by_hdr.u.specific_l2_params.update;
-+ param->u.hdr.insrt_params.u.by_hdr.u.specific_l2_params.size =
-+ compat_param->u.hdr.insrt_params.u.by_hdr.u.specific_l2_params.size;
-+ param->u.hdr.insrt_params.u.by_hdr.u.specific_l2_params.p_data =
-+ compat_ptr(compat_param->u.hdr.insrt_params.u.by_hdr.u.specific_l2_params.p_data);
-+ break;
-+ default:
-+ _fm_cpt_err("Unsupported type: %d", compat_param->u.hdr.insrt_params.type);
-+ }
-+
-+ param->u.hdr.field_update = compat_param->u.hdr.field_update;
-+ memcpy(&param->u.hdr.field_update_params,
-+ &compat_param->u.hdr.field_update_params,
-+ sizeof(param->u.hdr.field_update_params));
-+
-+ param->u.hdr.custom = compat_param->u.hdr.custom;
-+ memcpy(&param->u.hdr.custom_params,
-+ &compat_param->u.hdr.custom_params,
-+ sizeof(param->u.hdr.custom_params));
-+
-+ param->u.hdr.dont_parse_after_manip =
-+ compat_param->u.hdr.dont_parse_after_manip;
-+ break;
-+ case e_IOC_FM_PCD_MANIP_REASSEM:
-+ memcpy(&param->u.reassem, &compat_param->u.reassem, sizeof(param->u.reassem));
-+ break;
-+ case e_IOC_FM_PCD_MANIP_FRAG:
-+ memcpy(&param->u.frag, &compat_param->u.frag, sizeof(param->u.frag));
-+ break;
-+ case e_IOC_FM_PCD_MANIP_SPECIAL_OFFLOAD:
-+ memcpy(&param->u.special_offload,
-+ &compat_param->u.special_offload,
-+ sizeof(param->u.special_offload));
-+ break;
-+ }
-+
-+ param->p_next_manip = compat_pcd_id2ptr(compat_param->p_next_manip);
-+ param->id = compat_pcd_id2ptr(compat_param->id);
-+ }
-+ else {
-+ compat_param->type = param->type;
-+ memcpy(&compat_param->u, &param->u, sizeof(compat_param->u));
-+
-+ if (param->type == e_IOC_FM_PCD_MANIP_HDR &&
-+ param->u.hdr.insrt_params.type == e_IOC_FM_PCD_MANIP_INSRT_GENERIC)
-+ compat_param->u.hdr.insrt_params.u.generic.p_data =
-+ ptr_to_compat(param->u.hdr.insrt_params.u.generic.p_data);
-+
-+ compat_param->p_next_manip = compat_pcd_ptr2id(param->id);
-+ /* ... should be one that was added previously by the very call to
-+ compat_add_ptr2id() below: */
-+ compat_param->id = compat_add_ptr2id(param->id, FM_MAP_TYPE_PCD_NODE);
-+ }
-+}
-+
-+void compat_copy_fm_pcd_manip_get_stats(
-+ ioc_compat_fm_pcd_manip_get_stats_t *compat_param,
-+ ioc_fm_pcd_manip_get_stats_t *param,
-+ uint8_t compat)
-+{
-+ _fm_cpt_dbg (compat, " {->...\n");
-+
-+ if (compat == COMPAT_US_TO_K)
-+ {
-+ param->id = compat_pcd_id2ptr(compat_param->id);
-+ memcpy(&param->stats, &compat_param->stats,
-+ sizeof(ioc_fm_pcd_manip_stats_t));
-+ }
-+ else
-+ {
-+ compat_param->id = compat_add_ptr2id(param->id,
-+ FM_MAP_TYPE_PCD_NODE);
-+ memcpy(&compat_param->stats, &param->stats,
-+ sizeof(ioc_fm_pcd_manip_stats_t));
-+ }
-+
-+ _fm_cpt_dbg (compat, " ...->}\n");
-+}
-+
-+#if (DPAA_VERSION >= 11)
-+void compat_copy_fm_pcd_frm_replic_group_params(
-+ ioc_compat_fm_pcd_frm_replic_group_params_t *compat_param,
-+ ioc_fm_pcd_frm_replic_group_params_t *param,
-+ uint8_t compat)
-+{
-+ int k;
-+
-+ _fm_cpt_dbg (compat, " {->...\n");
-+
-+ if (compat == COMPAT_US_TO_K)
-+ {
-+ param->max_num_of_entries = compat_param->max_num_of_entries;
-+ param->num_of_entries = compat_param->num_of_entries;
-+ param->id = compat_pcd_id2ptr(compat_param->id);
-+ }
-+ else
-+ {
-+ compat_param->max_num_of_entries = param->max_num_of_entries;
-+ compat_param->num_of_entries = param->num_of_entries;
-+ compat_param->id = compat_add_ptr2id(param->id,
-+ FM_MAP_TYPE_PCD_NODE);
-+ }
-+
-+ for (k=0; k < IOC_FM_PCD_FRM_REPLIC_MAX_NUM_OF_ENTRIES; k++)
-+ compat_copy_fm_pcd_cc_next_engine(
-+ &compat_param->next_engine_params[k],
-+ &param->next_engine_params[k],
-+ compat);
-+
-+ _fm_cpt_dbg (compat, " ...->}\n");
-+}
-+
-+void compat_copy_fm_pcd_frm_replic_member(
-+ ioc_compat_fm_pcd_frm_replic_member_t *compat_param,
-+ ioc_fm_pcd_frm_replic_member_t *param,
-+ uint8_t compat)
-+{
-+ _fm_cpt_dbg (compat, " {->...\n");
-+
-+ if (compat == COMPAT_US_TO_K)
-+ {
-+ param->h_replic_group = compat_pcd_id2ptr(compat_param->h_replic_group);
-+ param->member_index = compat_param->member_index;
-+ }
-+
-+ _fm_cpt_dbg (compat, " ...->}\n");
-+}
-+
-+void compat_copy_fm_pcd_frm_replic_member_params(
-+ ioc_compat_fm_pcd_frm_replic_member_params_t *compat_param,
-+ ioc_fm_pcd_frm_replic_member_params_t *param,
-+ uint8_t compat)
-+{
-+ _fm_cpt_dbg (compat, " {->...\n");
-+
-+ compat_copy_fm_pcd_frm_replic_member(&compat_param->member,
-+ &param->member, compat);
-+
-+ compat_copy_fm_pcd_cc_next_engine(&compat_param->next_engine_params,
-+ &param->next_engine_params, compat);
-+
-+ _fm_cpt_dbg (compat, " ...->}\n");
-+}
-+
-+void compat_copy_fm_vsp_params(
-+ ioc_compat_fm_vsp_params_t *compat_param,
-+ ioc_fm_vsp_params_t *param,
-+ uint8_t compat)
-+{
-+ _fm_cpt_dbg (compat, " {->...\n");
-+
-+ if (compat == COMPAT_US_TO_K)
-+ {
-+ memcpy(&param->ext_buf_pools, &compat_param->ext_buf_pools, sizeof(ioc_fm_ext_pools));
-+ param->liodn_offset = compat_param->liodn_offset;
-+ param->port_params.port_id = compat_param->port_params.port_id;
-+ param->port_params.port_type = compat_param->port_params.port_type;
-+ param->relative_profile_id = compat_param->relative_profile_id;
-+ }
-+ else
-+ {
-+ memcpy(&compat_param->ext_buf_pools, &param->ext_buf_pools, sizeof(ioc_fm_ext_pools));
-+ compat_param->liodn_offset = param->liodn_offset;
-+ compat_param->port_params.port_id = param->port_params.port_id;
-+ compat_param->port_params.port_type = param->port_params.port_type;
-+ compat_param->relative_profile_id = param->relative_profile_id;
-+ compat_param->id = compat_add_ptr2id(param->id, FM_MAP_TYPE_PCD_NODE);
-+ }
-+
-+ _fm_cpt_dbg (compat, " ...->}\n");
-+}
-+
-+void compat_copy_fm_buf_pool_depletion_params(
-+ ioc_compat_fm_buf_pool_depletion_params_t *compat_param,
-+ ioc_fm_buf_pool_depletion_params_t *param,
-+ uint8_t compat)
-+{
-+ _fm_cpt_dbg (compat, " {->...\n");
-+
-+ if (compat == COMPAT_US_TO_K)
-+ {
-+ param->p_fm_vsp = compat_pcd_id2ptr(compat_param->p_fm_vsp);
-+ memcpy(&param->fm_buf_pool_depletion,
-+ &compat_param->fm_buf_pool_depletion,
-+ sizeof(ioc_fm_buf_pool_depletion_t));
-+ }
-+
-+ _fm_cpt_dbg (compat, " ...->}\n");
-+}
-+
-+void compat_copy_fm_buffer_prefix_content_params(
-+ ioc_compat_fm_buffer_prefix_content_params_t *compat_param,
-+ ioc_fm_buffer_prefix_content_params_t *param,
-+ uint8_t compat)
-+{
-+ _fm_cpt_dbg (compat, " {->...\n");
-+
-+ if (compat == COMPAT_US_TO_K)
-+ {
-+ param->p_fm_vsp = compat_pcd_id2ptr(compat_param->p_fm_vsp);
-+ memcpy(&param->fm_buffer_prefix_content,
-+ &compat_param->fm_buffer_prefix_content,
-+ sizeof(ioc_fm_buffer_prefix_content_t));
-+ }
-+
-+ _fm_cpt_dbg (compat, " ...->}\n");
-+}
-+
-+void compat_copy_fm_vsp_config_no_sg_params(
-+ ioc_compat_fm_vsp_config_no_sg_params_t *compat_param,
-+ ioc_fm_vsp_config_no_sg_params_t *param,
-+ uint8_t compat)
-+{
-+ _fm_cpt_dbg (compat, " {->...\n");
-+
-+ if (compat == COMPAT_US_TO_K)
-+ {
-+ param->p_fm_vsp = compat_pcd_id2ptr(compat_param->p_fm_vsp);
-+ param->no_sg = compat_param->no_sg;
-+ }
-+
-+ _fm_cpt_dbg (compat, " ...->}\n");
-+}
-+
-+void compat_copy_fm_vsp_prs_result_params(
-+ ioc_compat_fm_vsp_prs_result_params_t *compat_param,
-+ ioc_fm_vsp_prs_result_params_t *param,
-+ uint8_t compat)
-+{
-+ _fm_cpt_dbg (compat, " {->...\n");
-+
-+ if (compat == COMPAT_US_TO_K)
-+ {
-+ param->p_fm_vsp = compat_pcd_id2ptr(compat_param->p_fm_vsp);
-+ /* p_data is an user-space pointer that needs to remain unmodified */
-+ param->p_data = (void *)(unsigned long long)compat_param->p_data;
-+ }
-+ else
-+ {
-+ compat_param->p_fm_vsp = compat_pcd_ptr2id(param->p_fm_vsp);
-+ /* p_data is an user-space pointer that needs to remain unmodified */
-+ compat_param->p_data = (compat_uptr_t)((unsigned long long)param->p_data & 0xFFFFFFFF);
-+ }
-+
-+ _fm_cpt_dbg (compat, " ...->}\n");
-+}
-+#endif /* (DPAA_VERSION >= 11) */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/src/wrapper/lnxwrp_ioctls_fm_compat.h
-@@ -0,0 +1,755 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+/*
-+ @File lnxwrp_ioctls_fm_compat.h
-+
-+ @Description FM PCD compat structures definition.
-+
-+*/
-+
-+#ifndef __FM_COMPAT_IOCTLS_H
-+#define __FM_COMPAT_IOCTLS_H
-+
-+#include <linux/compat.h>
-+
-+#define COMPAT_K_TO_US 0 /* copy from Kernel to User */
-+#define COMPAT_US_TO_K 1 /* copy from User to Kernel */
-+#define COMPAT_GENERIC 2
-+
-+#define COMPAT_COPY_K2US(dest, src, type) compat_copy_##type(src, dest, 0)
-+#define COMPAT_COPY_US2K(dest, src, type) compat_copy_##type(dest, src, 1)
-+
-+/* mapping kernel pointers w/ UserSpace id's { */
-+/* Because compat_ptr(ptr_to_compat(X)) != X, this way we cannot exchange pointers
-+ back and forth (US - KS). compat_ptr is a cast and pointers are broken. */
-+#define COMPAT_PTR2ID_ARRAY_MAX (512+1) /* first location is not used */
-+#define COMPAT_PTR2ID_WATERMARK 0xface0000
-+#define COMPAT_PTR2ID_WM_MASK 0xffff0000
-+
-+/* define it for debug trace */
-+/*#define FM_COMPAT_DBG*/
-+
-+#define _fm_cpt_prk(stage, format, arg...) \
-+ printk(stage "fm_cpt (cpu:%u): " format, raw_smp_processor_id(), ##arg)
-+
-+#define _fm_cpt_inf(format, arg...) _fm_cpt_prk(KERN_INFO, format, ##arg)
-+#define _fm_cpt_wrn(format, arg...) _fm_cpt_prk(KERN_WARNING, format, ##arg)
-+#define _fm_cpt_err(format, arg...) _fm_cpt_prk(KERN_ERR, format, ##arg)
-+
-+/* used for compat IOCTL debugging */
-+#if defined(FM_COMPAT_DBG)
-+ #define _fm_cpt_dbg(from, format, arg...) \
-+ do{ \
-+ if (from == COMPAT_US_TO_K) \
-+ printk("fm_cpt to KS [%s:%u](cpu:%u) - " format, \
-+ __func__, __LINE__, raw_smp_processor_id(), ##arg); \
-+ else if (from == COMPAT_K_TO_US) \
-+ printk("fm_cpt to US [%s:%u](cpu:%u) - " format, \
-+ __func__, __LINE__, raw_smp_processor_id(), ##arg); \
-+ else \
-+ printk("fm_cpt [%s:%u](cpu:%u) - " format, \
-+ __func__, __LINE__, raw_smp_processor_id(), ##arg); \
-+ }while(0)
-+#else
-+# define _fm_cpt_dbg(arg...)
-+#endif
-+
-+/*TODO: per FMan module:
-+ *
-+ * Parser: FM_MAP_TYPE_PARSER_NODE,
-+ * Kg: FM_MAP_TYPE_KG_NODE,
-+ * Policer: FM_MAP_TYPE_POLICER_NODE
-+ * Manip: FM_MAP_TYPE_MANIP_NODE
-+ **/
-+enum fm_map_node_type {
-+ FM_MAP_TYPE_UNSPEC = 0,
-+ FM_MAP_TYPE_PCD_NODE,
-+
-+ /* add types here, update the policy */
-+
-+ __FM_MAP_TYPE_AFTER_LAST,
-+ FM_MAP_TYPE_MAX = __FM_MAP_TYPE_AFTER_LAST - 1
-+};
-+
-+void compat_del_ptr2id(void *p, enum fm_map_node_type);
-+compat_uptr_t compat_add_ptr2id(void *p, enum fm_map_node_type);
-+compat_uptr_t compat_get_ptr2id(void *p, enum fm_map_node_type);
-+void *compat_get_id2ptr(compat_uptr_t comp, enum fm_map_node_type);
-+
-+static inline compat_uptr_t compat_pcd_ptr2id(void *ptr) {
-+ return (ptr)? compat_get_ptr2id(ptr, FM_MAP_TYPE_PCD_NODE)
-+ : (compat_uptr_t) 0;
-+}
-+
-+static inline void *compat_pcd_id2ptr(compat_uptr_t id) {
-+ return (id) ? compat_get_id2ptr(id, FM_MAP_TYPE_PCD_NODE)
-+ : NULL;
-+}
-+
-+/* other similar inlines may be added as new nodes are added
-+ to enum fm_map_node_type above... */
-+/* } mapping kernel pointers w/ UserSpace id's */
-+
-+/* pcd compat structures { */
-+typedef struct ioc_compat_fm_pcd_cc_node_remove_key_params_t {
-+ compat_uptr_t id;
-+ uint16_t key_indx;
-+} ioc_compat_fm_pcd_cc_node_remove_key_params_t;
-+
-+typedef union ioc_compat_fm_pcd_plcr_next_engine_params_u {
-+ ioc_fm_pcd_done_action action;
-+ compat_uptr_t p_profile;
-+ compat_uptr_t p_direct_scheme;
-+} ioc_compat_fm_pcd_plcr_next_engine_params_u;
-+
-+typedef struct ioc_compat_fm_pcd_plcr_profile_params_t {
-+ bool modify;
-+ union {
-+ struct {
-+ ioc_fm_pcd_profile_type_selection profile_type;
-+ compat_uptr_t p_fm_port;
-+ uint16_t relative_profile_id;
-+ } new_params;
-+ compat_uptr_t p_profile;
-+ } profile_select;
-+ ioc_fm_pcd_plcr_algorithm_selection alg_selection;
-+ ioc_fm_pcd_plcr_color_mode color_mode;
-+
-+ union {
-+ ioc_fm_pcd_plcr_color dflt_color;
-+ ioc_fm_pcd_plcr_color override;
-+ } color;
-+
-+ ioc_fm_pcd_plcr_non_passthrough_alg_param_t non_passthrough_alg_param;
-+
-+ ioc_fm_pcd_engine next_engine_on_green;
-+ ioc_compat_fm_pcd_plcr_next_engine_params_u params_on_green;
-+
-+ ioc_fm_pcd_engine next_engine_on_yellow;
-+ ioc_compat_fm_pcd_plcr_next_engine_params_u params_on_yellow;
-+
-+ ioc_fm_pcd_engine next_engine_on_red;
-+ ioc_compat_fm_pcd_plcr_next_engine_params_u params_on_red;
-+
-+ bool trap_profile_on_flow_A;
-+ bool trap_profile_on_flow_B;
-+ bool trap_profile_on_flow_C;
-+ compat_uptr_t id;
-+} ioc_compat_fm_pcd_plcr_profile_params_t;
-+
-+typedef struct ioc_compat_fm_obj_t {
-+ compat_uptr_t obj;
-+} ioc_compat_fm_obj_t;
-+
-+typedef struct ioc_compat_fm_pcd_kg_scheme_select_t {
-+ bool direct;
-+ compat_uptr_t scheme_id;
-+} ioc_compat_fm_pcd_kg_scheme_select_t;
-+
-+typedef struct ioc_compat_fm_pcd_port_schemes_params_t {
-+ uint8_t num_of_schemes;
-+ compat_uptr_t scheme_ids[FM_PCD_KG_NUM_OF_SCHEMES];
-+} ioc_compat_fm_pcd_port_schemes_params_t;
-+
-+#if (DPAA_VERSION >= 11)
-+typedef struct ioc_compat_fm_port_vsp_alloc_params_t {
-+ uint8_t num_of_profiles; /**< Number of Virtual Storage Profiles */
-+ uint8_t dflt_relative_id; /**< The default Virtual-Storage-Profile-id dedicated to Rx/OP port
-+ The same default Virtual-Storage-Profile-id will be for coupled Tx port
-+ if relevant function called for Rx port */
-+ compat_uptr_t p_fm_tx_port; /**< Handle to coupled Tx Port; not relevant for OP port. */
-+}ioc_compat_fm_port_vsp_alloc_params_t;
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+typedef struct ioc_compat_fm_pcd_net_env_params_t {
-+ uint8_t num_of_distinction_units;
-+ ioc_fm_pcd_distinction_unit_t units[IOC_FM_PCD_MAX_NUM_OF_DISTINCTION_UNITS]; /* same structure*/
-+ compat_uptr_t id;
-+} ioc_compat_fm_pcd_net_env_params_t;
-+
-+typedef struct ioc_compat_fm_pcd_prs_sw_params_t {
-+ bool override;
-+ uint32_t size;
-+ uint16_t base;
-+ compat_uptr_t p_code;
-+ uint32_t sw_prs_data_params[IOC_FM_PCD_PRS_NUM_OF_HDRS];
-+ uint8_t num_of_labels;
-+ ioc_fm_pcd_prs_label_params_t labels_table[IOC_FM_PCD_PRS_NUM_OF_LABELS];
-+} ioc_compat_fm_pcd_prs_sw_params_t;
-+
-+typedef struct ioc_compat_fm_pcd_cc_next_kg_params_t {
-+ bool override_fqid;
-+ uint32_t new_fqid;
-+#if DPAA_VERSION >= 11
-+ uint8_t new_relative_storage_profile_id;
-+#endif
-+ compat_uptr_t p_direct_scheme;
-+} ioc_compat_fm_pcd_cc_next_kg_params_t;
-+
-+typedef struct ioc_compat_fm_pcd_cc_next_cc_params_t {
-+ compat_uptr_t cc_node_id;
-+} ioc_compat_fm_pcd_cc_next_cc_params_t;
-+
-+#if DPAA_VERSION >= 11
-+typedef struct ioc_compat_fm_pcd_cc_next_fr_params_t {
-+ compat_uptr_t frm_replic_id;
-+} ioc_compat_fm_pcd_cc_next_fr_params_t;
-+#endif /* DPAA_VERSION >= 11 */
-+
-+typedef struct ioc_compat_fm_pcd_cc_next_engine_params_t {
-+ ioc_fm_pcd_engine next_engine;
-+ union {
-+ ioc_compat_fm_pcd_cc_next_cc_params_t cc_params; /**< compat structure*/
-+ ioc_fm_pcd_cc_next_plcr_params_t plcr_params; /**< same structure*/
-+ ioc_fm_pcd_cc_next_enqueue_params_t enqueue_params; /**< same structure*/
-+ ioc_compat_fm_pcd_cc_next_kg_params_t kg_params; /**< compat structure*/
-+#if DPAA_VERSION >= 11
-+ ioc_compat_fm_pcd_cc_next_fr_params_t fr_params; /**< compat structure*/
-+#endif /* DPAA_VERSION >= 11 */
-+ } params;
-+ compat_uptr_t manip_id;
-+ bool statistics_en;
-+} ioc_compat_fm_pcd_cc_next_engine_params_t;
-+
-+typedef struct ioc_compat_fm_pcd_cc_grp_params_t {
-+ uint8_t num_of_distinction_units;
-+ uint8_t unit_ids [IOC_FM_PCD_MAX_NUM_OF_CC_UNITS];
-+ ioc_compat_fm_pcd_cc_next_engine_params_t next_engine_per_entries_in_grp[IOC_FM_PCD_MAX_NUM_OF_CC_ENTRIES_IN_GRP];
-+} ioc_compat_fm_pcd_cc_grp_params_t;
-+
-+typedef struct ioc_compat_fm_pcd_cc_tree_params_t {
-+ compat_uptr_t net_env_id;
-+ uint8_t num_of_groups;
-+ ioc_compat_fm_pcd_cc_grp_params_t fm_pcd_cc_group_params [IOC_FM_PCD_MAX_NUM_OF_CC_GROUPS];
-+ compat_uptr_t id;
-+} ioc_compat_fm_pcd_cc_tree_params_t;
-+
-+typedef struct ioc_compat_fm_pcd_cc_tree_modify_next_engine_params_t {
-+ compat_uptr_t id;
-+ uint8_t grp_indx;
-+ uint8_t indx;
-+ ioc_compat_fm_pcd_cc_next_engine_params_t cc_next_engine_params;
-+} ioc_compat_fm_pcd_cc_tree_modify_next_engine_params_t;
-+
-+typedef struct ioc_compat_fm_pcd_cc_key_params_t {
-+ compat_uptr_t p_key;
-+ compat_uptr_t p_mask;
-+ ioc_compat_fm_pcd_cc_next_engine_params_t cc_next_engine_params; /**< compat structure*/
-+} ioc_compat_fm_pcd_cc_key_params_t;
-+
-+typedef struct ioc_compat_keys_params_t {
-+ uint16_t max_num_of_keys;
-+ bool mask_support;
-+ ioc_fm_pcd_cc_stats_mode statistics_mode;
-+#if (DPAA_VERSION >= 11)
-+ uint16_t frame_length_ranges[IOC_FM_PCD_CC_STATS_MAX_NUM_OF_FLR];
-+#endif /* (DPAA_VERSION >= 11) */
-+ uint16_t num_of_keys;
-+ uint8_t key_size;
-+ ioc_compat_fm_pcd_cc_key_params_t key_params[IOC_FM_PCD_MAX_NUM_OF_KEYS]; /**< compat structure*/
-+ ioc_compat_fm_pcd_cc_next_engine_params_t cc_next_engine_params_for_miss; /**< compat structure*/
-+} ioc_compat_keys_params_t;
-+
-+typedef struct ioc_compat_fm_pcd_cc_node_params_t {
-+ ioc_fm_pcd_extract_entry_t extract_cc_params; /**< same structure*/
-+ ioc_compat_keys_params_t keys_params; /**< compat structure*/
-+ compat_uptr_t id;
-+} ioc_compat_fm_pcd_cc_node_params_t;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining a hash table
-+*//***************************************************************************/
-+typedef struct ioc_compat_fm_pcd_hash_table_params_t {
-+ uint16_t max_num_of_keys;
-+ ioc_fm_pcd_cc_stats_mode statistics_mode;
-+ uint8_t kg_hash_shift;
-+ uint16_t hash_res_mask;
-+ uint8_t hash_shift;
-+ uint8_t match_key_size;
-+ ioc_compat_fm_pcd_cc_next_engine_params_t cc_next_engine_params_for_miss;
-+ compat_uptr_t id;
-+} ioc_compat_fm_pcd_hash_table_params_t;
-+
-+typedef struct ioc_compat_fm_pcd_hash_table_add_key_params_t {
-+ compat_uptr_t p_hash_tbl;
-+ uint8_t key_size;
-+ ioc_compat_fm_pcd_cc_key_params_t key_params;
-+} ioc_compat_fm_pcd_hash_table_add_key_params_t;
-+
-+typedef struct ioc_compat_fm_pcd_cc_node_modify_key_params_t {
-+ compat_uptr_t id;
-+ uint16_t key_indx;
-+ uint8_t key_size;
-+ compat_uptr_t p_key;
-+ compat_uptr_t p_mask;
-+} ioc_compat_fm_pcd_cc_node_modify_key_params_t;
-+
-+typedef struct ioc_compat_fm_pcd_hash_table_remove_key_params_t {
-+ compat_uptr_t p_hash_tbl;
-+ uint8_t key_size;
-+ compat_uptr_t p_key;
-+} ioc_compat_fm_pcd_hash_table_remove_key_params_t;
-+
-+typedef struct ioc_compat_fm_pcd_cc_node_modify_key_and_next_engine_params_t {
-+ compat_uptr_t id;
-+ uint16_t key_indx;
-+ uint8_t key_size;
-+ ioc_compat_fm_pcd_cc_key_params_t key_params;
-+} ioc_compat_fm_pcd_cc_node_modify_key_and_next_engine_params_t;
-+
-+typedef struct ioc_compat_fm_port_pcd_plcr_params_t {
-+ compat_uptr_t plcr_profile_id;
-+} ioc_compat_fm_port_pcd_plcr_params_t;
-+
-+typedef struct ioc_compat_fm_port_pcd_cc_params_t {
-+ compat_uptr_t cc_tree_id;
-+} ioc_compat_fm_port_pcd_cc_params_t;
-+
-+typedef struct ioc_compat_fm_port_pcd_kg_params_t {
-+ uint8_t num_of_schemes;
-+ compat_uptr_t scheme_ids[FM_PCD_KG_NUM_OF_SCHEMES];
-+ bool direct_scheme;
-+ compat_uptr_t direct_scheme_id;
-+} ioc_compat_fm_port_pcd_kg_params_t;
-+
-+typedef struct ioc_compat_fm_port_pcd_params_t {
-+ ioc_fm_port_pcd_support pcd_support;
-+ compat_uptr_t net_env_id;
-+ compat_uptr_t p_prs_params;
-+ compat_uptr_t p_cc_params;
-+ compat_uptr_t p_kg_params;
-+ compat_uptr_t p_plcr_params;
-+ compat_uptr_t p_ip_reassembly_manip;
-+#if DPAA_VERSION >= 11
-+ compat_uptr_t p_capwap_reassembly_manip;
-+#endif
-+} ioc_compat_fm_port_pcd_params_t;
-+
-+typedef struct ioc_compat_fm_pcd_kg_cc_t {
-+ compat_uptr_t tree_id;
-+ uint8_t grp_id;
-+ bool plcr_next;
-+ bool bypass_plcr_profile_generation;
-+ ioc_fm_pcd_kg_plcr_profile_t plcr_profile;
-+} ioc_compat_fm_pcd_kg_cc_t;
-+
-+typedef struct ioc_compat_fm_pcd_kg_scheme_params_t {
-+ bool modify;
-+ union {
-+ uint8_t relative_scheme_id;
-+ compat_uptr_t scheme_id;
-+ } scm_id;
-+ bool always_direct;
-+ struct {
-+ compat_uptr_t net_env_id;
-+ uint8_t num_of_distinction_units;
-+ uint8_t unit_ids[IOC_FM_PCD_MAX_NUM_OF_DISTINCTION_UNITS];
-+ } net_env_params;
-+ bool use_hash;
-+ ioc_fm_pcd_kg_key_extract_and_hash_params_t key_extract_and_hash_params;
-+ bool bypass_fqid_generation;
-+ uint32_t base_fqid;
-+ uint8_t num_of_used_extracted_ors;
-+ ioc_fm_pcd_kg_extracted_or_params_t extracted_ors[IOC_FM_PCD_KG_NUM_OF_GENERIC_REGS];
-+#if DPAA_VERSION >= 11
-+ bool override_storage_profile;
-+ ioc_fm_pcd_kg_storage_profile_t storage_profile;
-+#endif /* DPAA_VERSION >= 11 */
-+ ioc_fm_pcd_engine next_engine;
-+ union{
-+ ioc_fm_pcd_done_action done_action;
-+ ioc_fm_pcd_kg_plcr_profile_t plcr_profile;
-+ ioc_compat_fm_pcd_kg_cc_t cc;
-+ } kg_next_engine_params;
-+ ioc_fm_pcd_kg_scheme_counter_t scheme_counter;
-+ compat_uptr_t id;
-+} ioc_compat_fm_pcd_kg_scheme_params_t;
-+
-+typedef struct ioc_compat_fm_pcd_cc_node_modify_next_engine_params_t {
-+ compat_uptr_t id;
-+ uint16_t key_indx;
-+ uint8_t key_size;
-+ ioc_compat_fm_pcd_cc_next_engine_params_t cc_next_engine_params;
-+} ioc_compat_fm_pcd_cc_node_modify_next_engine_params_t;
-+
-+typedef struct ioc_compat_fm_pcd_manip_hdr_insrt_generic_params_t {
-+ uint8_t offset;
-+ uint8_t size;
-+ bool replace;
-+ compat_uptr_t p_data;
-+} ioc_compat_fm_pcd_manip_hdr_insrt_generic_params_t;
-+
-+typedef struct ioc_compat_fm_pcd_manip_hdr_insrt_specific_l2_params_t {
-+ ioc_fm_pcd_manip_hdr_insrt_specific_l2 specific_l2;
-+ bool update;
-+ uint8_t size;
-+ compat_uptr_t p_data;
-+} ioc_compat_fm_pcd_manip_hdr_insrt_specific_l2_params_t;
-+
-+typedef struct ioc_compat_fm_pcd_manip_hdr_insrt_t {
-+ uint8_t size; /**< size of inserted section */
-+ compat_uptr_t p_data; /**< data to be inserted */
-+} ioc_compat_fm_pcd_manip_hdr_insrt_t;
-+
-+#if (DPAA_VERSION >= 11)
-+typedef struct ioc_compat_fm_pcd_manip_hdr_insrt_ip_params_t {
-+ bool calc_l4_checksum; /**< Calculate L4 checksum. */
-+ ioc_fm_pcd_manip_hdr_qos_mapping_mode mapping_mode; /**< TODO */
-+ uint8_t last_pid_offset; /**< the offset of the last Protocol within
-+ the inserted header */
-+ uint16_t id; /**< 16 bit New IP ID */
-+ bool dont_frag_overwrite;
-+ /**< IPv4 only. DF is overwritten with the hash-result next-to-last byte.
-+ * This byte is configured to be overwritten when RPD is set. */
-+ uint8_t last_dst_offset;
-+ /**< IPv6 only. if routing extension exist, user should set the offset of the destination address
-+ * in order to calculate UDP checksum pseudo header;
-+ * Otherwise set it to '0'. */
-+ ioc_compat_fm_pcd_manip_hdr_insrt_t insrt; /**< size and data to be inserted. */
-+} ioc_compat_fm_pcd_manip_hdr_insrt_ip_params_t;
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+typedef struct ioc_compat_fm_pcd_manip_hdr_insrt_by_hdr_params_t {
-+ ioc_fm_pcd_manip_hdr_insrt_by_hdr_type type;
-+ union {
-+ ioc_compat_fm_pcd_manip_hdr_insrt_specific_l2_params_t specific_l2_params;
-+#if (DPAA_VERSION >= 11)
-+ ioc_compat_fm_pcd_manip_hdr_insrt_ip_params_t ip_params;
-+ ioc_compat_fm_pcd_manip_hdr_insrt_t insrt;
-+#endif /* (DPAA_VERSION >= 11) */
-+ } u;
-+} ioc_compat_fm_pcd_manip_hdr_insrt_by_hdr_params_t;
-+
-+typedef struct ioc_compat_fm_pcd_manip_hdr_insrt_params_t {
-+ ioc_fm_pcd_manip_hdr_insrt_type type;
-+ union {
-+ ioc_compat_fm_pcd_manip_hdr_insrt_by_hdr_params_t by_hdr;
-+ ioc_compat_fm_pcd_manip_hdr_insrt_generic_params_t generic;
-+#if (defined(FM_CAPWAP_SUPPORT) && (DPAA_VERSION == 10))
-+#error "FM_CAPWAP_SUPPORT feature not supported!"
-+ ioc_fm_pcd_manip_hdr_insrt_by_template_params_t by_template;
-+#endif /* FM_CAPWAP_SUPPORT */
-+ } u;
-+} ioc_compat_fm_pcd_manip_hdr_insrt_params_t;
-+
-+typedef struct ioc_compat_fm_pcd_manip_hdr_params_t {
-+ bool rmv;
-+ ioc_fm_pcd_manip_hdr_rmv_params_t rmv_params;
-+ bool insrt;
-+ ioc_compat_fm_pcd_manip_hdr_insrt_params_t insrt_params;
-+ bool field_update;
-+ ioc_fm_pcd_manip_hdr_field_update_params_t field_update_params;
-+ bool custom;
-+ ioc_fm_pcd_manip_hdr_custom_params_t custom_params;
-+ bool dont_parse_after_manip;
-+} ioc_compat_fm_pcd_manip_hdr_params_t;
-+
-+typedef struct ioc_compat_fm_pcd_manip_special_offload_params_t {
-+ bool decryption;
-+ bool ecn_copy;
-+ bool dscp_copy;
-+ bool variable_ip_hdr_len;
-+ bool variable_ip_version;
-+ uint8_t outer_ip_hdr_len;
-+ uint16_t arw_size;
-+ compat_uptr_t arw_addr;
-+} ioc_compat_fm_pcd_manip_special_offload_params_t;
-+
-+typedef struct ioc_compat_fm_pcd_manip_params_t {
-+ ioc_fm_pcd_manip_type type;
-+ union {
-+ ioc_compat_fm_pcd_manip_hdr_params_t hdr;
-+ ioc_fm_pcd_manip_reassem_params_t reassem;
-+ ioc_fm_pcd_manip_frag_params_t frag;
-+ ioc_compat_fm_pcd_manip_special_offload_params_t special_offload;
-+ } u;
-+ compat_uptr_t p_next_manip;
-+#if (defined(FM_CAPWAP_SUPPORT) && (DPAA_VERSION == 10))
-+#error "FM_CAPWAP_SUPPORT feature not supported!"
-+ bool frag_or_reasm;
-+ ioc_fm_pcd_manip_frag_or_reasm_params_t frag_or_reasm_params;
-+#endif /* FM_CAPWAP_SUPPORT */
-+ compat_uptr_t id;
-+} ioc_compat_fm_pcd_manip_params_t;
-+
-+typedef struct ioc_compat_fm_pcd_manip_get_stats_t {
-+ compat_uptr_t id;
-+ ioc_fm_pcd_manip_stats_t stats;
-+} ioc_compat_fm_pcd_manip_get_stats_t;
-+
-+#if (DPAA_VERSION >= 11)
-+typedef struct ioc_compat_fm_pcd_frm_replic_group_params_t {
-+ uint8_t max_num_of_entries;
-+ uint8_t num_of_entries;
-+ ioc_compat_fm_pcd_cc_next_engine_params_t
-+ next_engine_params[IOC_FM_PCD_FRM_REPLIC_MAX_NUM_OF_ENTRIES];
-+ compat_uptr_t id;
-+} ioc_compat_fm_pcd_frm_replic_group_params_t;
-+
-+typedef struct ioc_compat_fm_pcd_frm_replic_member_t {
-+ compat_uptr_t h_replic_group;
-+ uint16_t member_index;
-+} ioc_compat_fm_pcd_frm_replic_member_t;
-+
-+typedef struct ioc_compat_fm_pcd_frm_replic_member_params_t {
-+ ioc_compat_fm_pcd_frm_replic_member_t member;
-+ ioc_compat_fm_pcd_cc_next_engine_params_t next_engine_params;
-+} ioc_compat_fm_pcd_frm_replic_member_params_t;
-+
-+typedef struct ioc_compat_fm_vsp_params_t {
-+ compat_uptr_t p_fm; /**< A handle to the FM object this VSP related to */
-+ ioc_fm_ext_pools ext_buf_pools; /**< Which external buffer pools are used
-+ (up to FM_PORT_MAX_NUM_OF_EXT_POOLS), and their sizes.
-+ parameter associated with Rx / OP port */
-+ uint16_t liodn_offset; /**< VSP's LIODN offset */
-+ struct {
-+ ioc_fm_port_type port_type; /**< Port type */
-+ uint8_t port_id; /**< Port Id - relative to type */
-+ } port_params;
-+ uint8_t relative_profile_id; /**< VSP Id - relative to VSP's range
-+ defined in relevant FM object */
-+ compat_uptr_t id; /**< return value */
-+} ioc_compat_fm_vsp_params_t;
-+
-+typedef struct ioc_compat_fm_buf_pool_depletion_params_t {
-+ compat_uptr_t p_fm_vsp;
-+ ioc_fm_buf_pool_depletion_t fm_buf_pool_depletion;
-+} ioc_compat_fm_buf_pool_depletion_params_t;
-+
-+typedef struct ioc_compat_fm_buffer_prefix_content_params_t {
-+ compat_uptr_t p_fm_vsp;
-+ ioc_fm_buffer_prefix_content_t fm_buffer_prefix_content;
-+} ioc_compat_fm_buffer_prefix_content_params_t;
-+
-+typedef struct ioc_compat_fm_vsp_config_no_sg_params_t {
-+ compat_uptr_t p_fm_vsp;
-+ bool no_sg;
-+} ioc_compat_fm_vsp_config_no_sg_params_t;
-+
-+typedef struct ioc_compat_fm_vsp_prs_result_params_t {
-+ compat_uptr_t p_fm_vsp;
-+ compat_uptr_t p_data;
-+} ioc_compat_fm_vsp_prs_result_params_t;
-+
-+#endif /* (DPAA_VERSION >= 11) */
-+typedef struct ioc_compat_fm_pcd_kg_scheme_spc_t {
-+ uint32_t val;
-+ compat_uptr_t id;
-+} ioc_compat_fm_pcd_kg_scheme_spc_t;
-+
-+typedef struct ioc_compat_fm_ctrl_mon_counters_params_t {
-+ uint8_t fm_ctrl_index;
-+ compat_uptr_t p_mon;
-+} ioc_compat_fm_ctrl_mon_counters_params_t;
-+
-+typedef struct ioc_compat_fm_pcd_cc_tbl_get_stats_t {
-+ compat_uptr_t id;
-+ uint16_t key_index;
-+ ioc_fm_pcd_cc_key_statistics_t statistics;
-+} ioc_compat_fm_pcd_cc_tbl_get_stats_t;
-+
-+
-+/* } pcd compat structures */
-+
-+void compat_obj_delete(
-+ ioc_compat_fm_obj_t *compat_id,
-+ ioc_fm_obj_t *id);
-+
-+/* pcd compat functions { */
-+void compat_copy_fm_pcd_plcr_profile(
-+ ioc_compat_fm_pcd_plcr_profile_params_t *compat_param,
-+ ioc_fm_pcd_plcr_profile_params_t *param,
-+ uint8_t compat);
-+
-+void compat_copy_fm_pcd_cc_key(
-+ ioc_compat_fm_pcd_cc_key_params_t *compat_param,
-+ ioc_fm_pcd_cc_key_params_t *param,
-+ uint8_t compat);
-+
-+void compat_copy_fm_pcd_cc_node_modify_key_and_next_engine(
-+ ioc_compat_fm_pcd_cc_node_modify_key_and_next_engine_params_t *compat_param,
-+ ioc_fm_pcd_cc_node_modify_key_and_next_engine_params_t *param,
-+ uint8_t compat);
-+
-+void compat_copy_fm_pcd_cc_node_modify_next_engine(
-+ ioc_compat_fm_pcd_cc_node_modify_next_engine_params_t *compat_param,
-+ ioc_fm_pcd_cc_node_modify_next_engine_params_t *param,
-+ uint8_t compat);
-+
-+void compat_fm_pcd_cc_tree_modify_next_engine(
-+ ioc_compat_fm_pcd_cc_tree_modify_next_engine_params_t *compat_param,
-+ ioc_fm_pcd_cc_tree_modify_next_engine_params_t *param,
-+ uint8_t compat);
-+
-+void compat_copy_fm_pcd_hash_table(
-+ ioc_compat_fm_pcd_hash_table_params_t *compat_param,
-+ ioc_fm_pcd_hash_table_params_t *param,
-+ uint8_t compat);
-+
-+void compat_copy_fm_pcd_cc_grp(
-+ ioc_compat_fm_pcd_cc_grp_params_t *compat_param,
-+ ioc_fm_pcd_cc_grp_params_t *param,
-+ uint8_t compat);
-+
-+void compat_copy_fm_pcd_cc_tree(
-+ ioc_compat_fm_pcd_cc_tree_params_t *compat_param,
-+ ioc_fm_pcd_cc_tree_params_t *param,
-+ uint8_t compat);
-+
-+void compat_copy_fm_pcd_cc_tbl_get_stats(
-+ ioc_compat_fm_pcd_cc_tbl_get_stats_t *compat_param,
-+ ioc_fm_pcd_cc_tbl_get_stats_t *param,
-+ uint8_t compat);
-+
-+void compat_fm_pcd_prs_sw(
-+ ioc_compat_fm_pcd_prs_sw_params_t *compat_param,
-+ ioc_fm_pcd_prs_sw_params_t *param,
-+ uint8_t compat);
-+
-+void compat_copy_fm_pcd_kg_scheme(
-+ ioc_compat_fm_pcd_kg_scheme_params_t *compat_param,
-+ ioc_fm_pcd_kg_scheme_params_t *param,
-+ uint8_t compat);
-+
-+void compat_copy_fm_pcd_kg_scheme_select(
-+ ioc_compat_fm_pcd_kg_scheme_select_t *compat_param,
-+ ioc_fm_pcd_kg_scheme_select_t *param,
-+ uint8_t compat);
-+
-+void compat_copy_fm_pcd_kg_schemes_params(
-+ ioc_compat_fm_pcd_port_schemes_params_t *compat_param,
-+ ioc_fm_pcd_port_schemes_params_t *param,
-+ uint8_t compat);
-+
-+void compat_copy_fm_port_pcd_kg(
-+ ioc_compat_fm_port_pcd_kg_params_t *compat_param,
-+ ioc_fm_port_pcd_kg_params_t *param,
-+ uint8_t compat);
-+
-+void compat_copy_fm_port_pcd(
-+ ioc_compat_fm_port_pcd_params_t *compat_param,
-+ ioc_fm_port_pcd_params_t *param,
-+ uint8_t compat);
-+
-+#if (DPAA_VERSION >= 11)
-+void compat_copy_fm_port_vsp_alloc_params(
-+ ioc_compat_fm_port_vsp_alloc_params_t *compat_param,
-+ ioc_fm_port_vsp_alloc_params_t *param,
-+ uint8_t compat);
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+void compat_copy_fm_pcd_net_env(
-+ ioc_compat_fm_pcd_net_env_params_t *compat_param,
-+ ioc_fm_pcd_net_env_params_t *param,
-+ uint8_t compat);
-+
-+void compat_copy_fm_pcd_cc_node_modify_key(
-+ ioc_compat_fm_pcd_cc_node_modify_key_params_t *compat_param,
-+ ioc_fm_pcd_cc_node_modify_key_params_t *param,
-+ uint8_t compat);
-+
-+void compat_copy_keys(
-+ ioc_compat_keys_params_t *compat_param,
-+ ioc_keys_params_t *param,
-+ uint8_t compat);
-+
-+void compat_copy_fm_pcd_cc_node(
-+ ioc_compat_fm_pcd_cc_node_params_t *compat_param,
-+ ioc_fm_pcd_cc_node_params_t *param,
-+ uint8_t compat);
-+
-+void compat_fm_pcd_manip_set_node(
-+ ioc_compat_fm_pcd_manip_params_t *compat_param,
-+ ioc_fm_pcd_manip_params_t *param,
-+ uint8_t compat);
-+
-+void compat_copy_fm_pcd_manip_get_stats(
-+ ioc_compat_fm_pcd_manip_get_stats_t *compat_param,
-+ ioc_fm_pcd_manip_get_stats_t *param,
-+ uint8_t compat);
-+
-+void compat_copy_fm_port_pcd_modify_tree(
-+ ioc_compat_fm_obj_t *compat_id,
-+ ioc_fm_obj_t *id,
-+ uint8_t compat);
-+
-+#if (DPAA_VERSION >= 11)
-+void compat_copy_fm_pcd_frm_replic_group_params(
-+ ioc_compat_fm_pcd_frm_replic_group_params_t *compat_param,
-+ ioc_fm_pcd_frm_replic_group_params_t *param,
-+ uint8_t compat);
-+
-+void compat_copy_fm_pcd_frm_replic_member(
-+ ioc_compat_fm_pcd_frm_replic_member_t *compat_param,
-+ ioc_fm_pcd_frm_replic_member_t *param,
-+ uint8_t compat);
-+
-+void compat_copy_fm_pcd_frm_replic_member_params(
-+ ioc_compat_fm_pcd_frm_replic_member_params_t *compat_param,
-+ ioc_fm_pcd_frm_replic_member_params_t *param,
-+ uint8_t compat);
-+
-+void compat_copy_fm_vsp_params(
-+ ioc_compat_fm_vsp_params_t *compat_param,
-+ ioc_fm_vsp_params_t *param,
-+ uint8_t compat);
-+
-+void compat_copy_fm_buf_pool_depletion_params(
-+ ioc_compat_fm_buf_pool_depletion_params_t *compat_param,
-+ ioc_fm_buf_pool_depletion_params_t *param,
-+ uint8_t compat);
-+
-+void compat_copy_fm_buffer_prefix_content_params(
-+ ioc_compat_fm_buffer_prefix_content_params_t *compat_param,
-+ ioc_fm_buffer_prefix_content_params_t *param,
-+ uint8_t compat);
-+
-+void compat_copy_fm_vsp_config_no_sg_params(
-+ ioc_compat_fm_vsp_config_no_sg_params_t *compat_param,
-+ ioc_fm_vsp_config_no_sg_params_t *param,
-+ uint8_t compat);
-+
-+void compat_copy_fm_vsp_prs_result_params(
-+ ioc_compat_fm_vsp_prs_result_params_t *compat_param,
-+ ioc_fm_vsp_prs_result_params_t *param,
-+ uint8_t compat);
-+
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+void compat_copy_fm_pcd_kg_scheme_spc(
-+ ioc_compat_fm_pcd_kg_scheme_spc_t *compat_param,
-+ ioc_fm_pcd_kg_scheme_spc_t *param,
-+ uint8_t compat);
-+
-+/* } pcd compat functions */
-+#endif
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/src/wrapper/lnxwrp_resources.h
-@@ -0,0 +1,121 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+/*
-+ @File lnxwrp_resources.h
-+
-+ @Description FMD wrapper resource allocation functions.
-+
-+*/
-+
-+#ifndef LNXWRP_RESOURCES_H_
-+#define LNXWRP_RESOURCES_H_
-+
-+#if !defined(FMAN_RESOURCES_UNIT_TEST)
-+#include "lnxwrp_fm.h"
-+#else
-+#include "lnxwrp_resources_ut.h"
-+#endif
-+
-+#define ROUND(X) ((2*(X)+1)/2)
-+#define CEIL(X) ((X)+1)
-+/* #define ROUND_DIV(X, Y) (((X)+(Y)/2)/(Y)) */
-+#define ROUND_DIV(X, Y) ((2*(X)+(Y))/(2*(Y)))
-+#define CEIL_DIV(X, Y) (((X)+(Y)-1)/(Y))
-+
-+/* used for resource calculus */
-+#define DPDE_1G 2 /* DQDP 1g - from LLD:
-+ DEFAULT_PORT_txFifoDeqPipelineDepth_1G */
-+#define DPDE_10G 8 /* DQDP 10g - from LLD:
-+ DEFAULT_PORT_txFifoDeqPipelineDepth_10G */
-+
-+int fm_set_active_fman_ports(struct platform_device *of_dev,
-+ t_LnxWrpFmDev *p_LnxWrpFmDev);
-+
-+/* Calculate the fifosize based on MURAM allocation, number of ports, dpde
-+ * value and s/g software support (! Kernel does not suport s/g).
-+ *
-+ * Algorithm summary:
-+ * - Calculate the the minimum fifosize required for every type of port
-+ * (TX,RX for 1G, 2.5G and 10G).
-+ * - Set TX the minimum fifosize required.
-+ * - Distribute the remaining buffers (after all TX were set) to RX ports
-+ * based on:
-+ * 1G RX = Remaining_buffers * 1/(1+2.5+10)
-+ * 2.5G RX = Remaining_buffers * 2.5/(1+2.5+10)
-+ * 10G RX = Remaining_buffers * 10/(1+2.5+10)
-+ * - if the RX is smaller than the minimum required, then set the minimum
-+ * required
-+ * - In the end distribuite the leftovers if there are any (due to
-+ * unprecise calculus) or if over allocation cat some buffers from all RX
-+ * ports w/o pass over minimum required treshold, but if there must be
-+ * pass the treshold in order to cat the over allocation ,then this
-+ * configuration can not be set - KERN_ALERT.
-+*/
-+int fm_precalculate_fifosizes(t_LnxWrpFmDev *p_LnxWrpFmDev,
-+ int muram_fifo_size);
-+
-+#if !defined(FMAN_RESOURCES_UNIT_TEST)
-+int fm_config_precalculate_fifosize(t_LnxWrpFmPortDev *p_LnxWrpFmPortDev);
-+#endif
-+
-+/* Compute FMan open DMA based on total number of open DMAs and
-+ * number of available fman ports.
-+ *
-+ * By default 10g ports are set to input parameters. The other ports
-+ * tries to keep the proportion rx=2tx open dmas or tresholds.
-+ *
-+ * If leftovers, then those will be set as shared.
-+ *
-+ * If after computing overflow appears, then it decrements open dma
-+ * for all ports w/o cross the tresholds. If the tresholds are meet
-+ * and is still overflow, then it returns error.
-+*/
-+int fm_precalculate_open_dma(t_LnxWrpFmDev *p_LnxWrpFmDev,
-+ int max_fm_open_dma,
-+ int default_tx_10g_dmas,
-+ int default_rx_10g_dmas,
-+ int min_tx_10g_treshold, int min_rx_10g_treshold);
-+
-+#if !defined(FMAN_RESOURCES_UNIT_TEST)
-+int fm_config_precalculate_open_dma(t_LnxWrpFmPortDev *p_LnxWrpFmPortDev);
-+#endif
-+
-+/* Compute FMan tnums based on available tnums and number of ports.
-+ * Set defaults (minim tresholds) and then distribute leftovers.*/
-+int fm_precalculate_tnums(t_LnxWrpFmDev *p_LnxWrpFmDev, int max_fm_tnums);
-+
-+#if !defined(FMAN_RESOURCES_UNIT_TEST)
-+int fm_config_precalculate_tnums(t_LnxWrpFmPortDev *p_LnxWrpFmPortDev);
-+#endif
-+
-+#endif /* LNXWRP_RESOURCES_H_ */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/src/wrapper/lnxwrp_resources_ut.c
-@@ -0,0 +1,191 @@
-+/* Copyright (c) 2012 Freescale Semiconductor, Inc.
-+ * All rights reserved.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#include "lnxwrp_resources.h"
-+#include "lnxwrp_resources_ut.h"
-+
-+#define KILOBYTE 0x400 /* 1024 */
-+
-+typedef enum e_board_type {
-+ e_p3041,
-+ e_p4080,
-+ e_p5020,
-+ e_p1023
-+} e_board_type;
-+
-+uint8_t board_type;
-+uint32_t muram_size = 0;
-+uint32_t dmas_num = 0;
-+uint32_t task_num = 0;
-+uint32_t frame_size = 0;
-+uint32_t oh_num = 0;
-+uint32_t num_ports_1g = 0;
-+uint32_t num_ports_10g = 0;
-+uint32_t num_ports_2g5 = 0;
-+uint32_t fsl_fman_phy_maxfrm = 0;
-+uint32_t dpa_rx_extra_headroom = 0;
-+
-+void show_help(void){
-+ printf(" help: \n");
-+ printf(" -b <board_type> -f <max_fram_size(mtu)> -o <num_oh_ports> -g1"
-+ " <num_1g_ports> -g10 <num_10g_ports> -g25 <num_2g5_ports>\n");
-+ printf(" Maxim num of DMAS availbale: P3/P4/P5:32 , P1023:16 \n");
-+ printf(" Maxim num of TNUMs availbale: P3/P4/P5:128, P1023:32 \n");
-+ printf(" Muram size: P3/P4/P5:160K, P1023:64K \n");
-+ printf(" Number of ports:\n");
-+ printf(" P3/P5: 5p 1g, 1p 10g, 7p oh \n");
-+ printf(" P4 : 4p 1g, 1p 10g, 7p oh \n");
-+ printf(" P1 : 2p 1g, 0p 10g, 4p oh \n");
-+ printf(" MTU: Default:1522, Jumbo:9600 \n");
-+}
-+
-+int fm_set_param(t_LnxWrpFmDev *p_LnxWrpFmDev) {
-+ struct fm_active_ports *fm_active_ports_info = NULL;
-+ fm_active_ports_info = &p_LnxWrpFmDev->fm_active_ports_info;
-+
-+ switch(board_type){
-+ case e_p3041:
-+ case e_p5020:
-+ muram_size = 160*KILOBYTE;
-+ dmas_num = 32;
-+ task_num = 128;
-+ if ((num_ports_1g+num_ports_2g5) > 5 || num_ports_10g > 1 || oh_num > 7)
-+ goto err_fm_set_param;
-+ break;
-+ case e_p4080:
-+ muram_size = 160*KILOBYTE;
-+ dmas_num = 32;
-+ task_num = 128;
-+ if ((num_ports_1g+num_ports_2g5) > 4 || num_ports_10g > 1 || oh_num > 7)
-+ goto err_fm_set_param;
-+ break;
-+ case e_p1023:
-+ muram_size = 64*KILOBYTE;
-+ dmas_num = 16;
-+ task_num = 128;
-+ if ((num_ports_1g+num_ports_2g5) > 2 || oh_num > 4)
-+ goto err_fm_set_param;
-+ break;
-+ default:
-+ goto err_fm_set_param;
-+ break;
-+ }
-+
-+ p_LnxWrpFmDev->id = 0;
-+ fsl_fman_phy_maxfrm = frame_size;
-+ dpa_rx_extra_headroom = 0; /* ATTENTION: can be != 0 */
-+ fm_active_ports_info->num_oh_ports = oh_num;
-+ fm_active_ports_info->num_tx_ports = num_ports_1g;
-+ fm_active_ports_info->num_rx_ports = num_ports_1g;
-+ fm_active_ports_info->num_tx25_ports = num_ports_2g5;
-+ fm_active_ports_info->num_rx25_ports = num_ports_2g5;
-+ fm_active_ports_info->num_tx10_ports = num_ports_10g;
-+ fm_active_ports_info->num_rx10_ports = num_ports_10g;
-+
-+ return 0;
-+
-+err_fm_set_param:
-+ printf(" ERR: To many ports!!! \n");
-+ return -1;
-+}
-+
-+int main (int argc, char *argv[]){
-+ t_LnxWrpFmDev LnxWrpFmDev;
-+ t_LnxWrpFmDev *p_LnxWrpFmDev = &LnxWrpFmDev;
-+ int tokens_cnt = 1;
-+
-+ char *token = NULL;
-+
-+ while(tokens_cnt < argc)
-+ {
-+ token = argv[tokens_cnt++];
-+ if (strcmp(token, "-b") == 0){
-+ if(strcmp(argv[tokens_cnt],"p3") == 0)
-+ board_type = e_p3041;
-+ else if(strcmp(argv[tokens_cnt],"p4") == 0)
-+ board_type = e_p4080;
-+ else if(strcmp(argv[tokens_cnt],"p5") == 0)
-+ board_type = e_p5020;
-+ else if(strcmp(argv[tokens_cnt],"p1") == 0)
-+ board_type = e_p1023;
-+ else
-+ show_help();
-+ tokens_cnt++;
-+ }
-+ else if(strcmp(token, "-d") == 0){
-+ dmas_num = atoi(argv[tokens_cnt++]);
-+ }
-+ else if(strcmp(token, "-t") == 0)
-+ task_num = atoi(argv[tokens_cnt++]);
-+ else if(strcmp(token, "-f") == 0)
-+ frame_size = atoi(argv[tokens_cnt++]);
-+ else if(strcmp(token, "-o") == 0)
-+ oh_num = atoi(argv[tokens_cnt++]);
-+ else if(strcmp(token, "-g1") == 0)
-+ num_ports_1g = atoi(argv[tokens_cnt++]);
-+ else if(strcmp(token, "-g10") == 0)
-+ num_ports_10g = atoi(argv[tokens_cnt++]);
-+ else if(strcmp(token, "-g25") == 0)
-+ num_ports_2g5 = atoi(argv[tokens_cnt++]);
-+ else {
-+ show_help();
-+ return -1;
-+ }
-+ }
-+
-+ if(fm_set_param(p_LnxWrpFmDev) < 0){
-+ show_help();
-+ return -1;
-+ }
-+
-+ if(fm_precalculate_fifosizes(
-+ p_LnxWrpFmDev,
-+ 128*KILOBYTE)
-+ != 0)
-+ return -1;
-+ if(fm_precalculate_open_dma(
-+ p_LnxWrpFmDev,
-+ dmas_num, /* max open dmas:dpaa_integration_ext.h */
-+ FM_DEFAULT_TX10G_OPENDMA, /* default TX 10g open dmas */
-+ FM_DEFAULT_RX10G_OPENDMA, /* default RX 10g open dmas */
-+ FM_10G_OPENDMA_MIN_TRESHOLD,/* TX 10g minimum treshold */
-+ FM_10G_OPENDMA_MIN_TRESHOLD)/* RX 10g minimum treshold */
-+ != 0)
-+ return -1;
-+ if(fm_precalculate_tnums(
-+ p_LnxWrpFmDev,
-+ task_num) /* max TNUMS: dpa integration file. */
-+ != 0)
-+ return -1;
-+
-+ return 0;
-+}
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/src/wrapper/lnxwrp_resources_ut.h
-@@ -0,0 +1,144 @@
-+/* Copyright (c) 2012 Freescale Semiconductor, Inc
-+ * All rights reserved.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#ifndef FM_RESS_TEST_H_
-+#define FM_RESS_TEST_H_
-+
-+#include <stdint.h>
-+#include <stdbool.h>
-+#include <stdio.h>
-+#include <assert.h>
-+#include <string.h>
-+#include <stdlib.h>
-+
-+#define _Packed
-+#define _PackedType __attribute__ ((packed))
-+#define MAX(x, y) (((x) > (y)) ? (x) : (y))
-+#define MIN(x, y) (((x) < (y)) ? (x) : (y))
-+#define KERN_ALERT ""
-+#define KERN_INFO ""
-+#define ASSERT_COND assert
-+#define printk printf
-+#define NET_IP_ALIGN 0
-+#define FM_FIFO_ALLOCATION_OLD_ALG
-+
-+#if defined(CONFIG_FMAN_DISABLE_OH_AND_DISTRIBUTE_RESOURCES)
-+#define FM_10G_OPENDMA_MIN_TRESHOLD 8 /* 10g minimum treshold if only HC is enabled and no OH port enabled */
-+#define FM_OPENDMA_RX_TX_RAPORT 2 /* RX = 2*TX */
-+#else
-+#define FM_10G_OPENDMA_MIN_TRESHOLD 7 /* 10g minimum treshold if 7 OH ports are enabled */
-+#define FM_OPENDMA_RX_TX_RAPORT 1 /* RX = TX */
-+#endif
-+#define FM_DEFAULT_TX10G_OPENDMA 8 /* default TX 10g open dmas */
-+#define FM_DEFAULT_RX10G_OPENDMA 8 /* default RX 10g open dmas */
-+
-+/* information about all active ports for an FMan.
-+ * !Some ports may be disabled by u-boot, thus will not be available */
-+struct fm_active_ports {
-+ uint32_t num_oh_ports;
-+ uint32_t num_tx_ports;
-+ uint32_t num_rx_ports;
-+ uint32_t num_tx25_ports;
-+ uint32_t num_rx25_ports;
-+ uint32_t num_tx10_ports;
-+ uint32_t num_rx10_ports;
-+};
-+
-+/* FMan resources precalculated at fm probe based
-+ * on available FMan port. */
-+struct fm_resource_settings {
-+ /* buffers - fifo sizes */
-+ uint32_t tx1g_num_buffers;
-+ uint32_t rx1g_num_buffers;
-+ uint32_t tx2g5_num_buffers; /* Not supported yet by LLD */
-+ uint32_t rx2g5_num_buffers; /* Not supported yet by LLD */
-+ uint32_t tx10g_num_buffers;
-+ uint32_t rx10g_num_buffers;
-+ uint32_t oh_num_buffers;
-+ uint32_t shared_ext_buffers;
-+
-+
-+ /* open DMAs */
-+ uint32_t tx_1g_dmas;
-+ uint32_t rx_1g_dmas;
-+ uint32_t tx_2g5_dmas; /* Not supported yet by LLD */
-+ uint32_t rx_2g5_dmas; /* Not supported yet by LLD */
-+ uint32_t tx_10g_dmas;
-+ uint32_t rx_10g_dmas;
-+ uint32_t oh_dmas;
-+ uint32_t shared_ext_open_dma;
-+
-+ /* Tnums */
-+ uint32_t tx_1g_tnums;
-+ uint32_t rx_1g_tnums;
-+ uint32_t tx_2g5_tnums; /* Not supported yet by LLD */
-+ uint32_t rx_2g5_tnums; /* Not supported yet by LLD */
-+ uint32_t tx_10g_tnums;
-+ uint32_t rx_10g_tnums;
-+ uint32_t oh_tnums;
-+ uint32_t shared_ext_tnums;
-+};
-+
-+typedef struct {
-+ uint8_t id;
-+ struct fm_active_ports fm_active_ports_info;
-+ struct fm_resource_settings fm_resource_settings_info;
-+} t_LnxWrpFmDev;
-+
-+typedef struct {
-+ uint8_t id;
-+} t_LnxWrpFmPortDev;
-+
-+typedef _Packed struct t_FmPrsResult {
-+ volatile uint8_t lpid; /**< Logical port id */
-+ volatile uint8_t shimr; /**< Shim header result */
-+ volatile uint16_t l2r; /**< Layer 2 result */
-+ volatile uint16_t l3r; /**< Layer 3 result */
-+ volatile uint8_t l4r; /**< Layer 4 result */
-+ volatile uint8_t cplan; /**< Classification plan id */
-+ volatile uint16_t nxthdr; /**< Next Header */
-+ volatile uint16_t cksum; /**< Checksum */
-+ volatile uint32_t lcv; /**< LCV */
-+ volatile uint8_t shim_off[3]; /**< Shim offset */
-+ volatile uint8_t eth_off; /**< ETH offset */
-+ volatile uint8_t llc_snap_off; /**< LLC_SNAP offset */
-+ volatile uint8_t vlan_off[2]; /**< VLAN offset */
-+ volatile uint8_t etype_off; /**< ETYPE offset */
-+ volatile uint8_t pppoe_off; /**< PPP offset */
-+ volatile uint8_t mpls_off[2]; /**< MPLS offset */
-+ volatile uint8_t ip_off[2]; /**< IP offset */
-+ volatile uint8_t gre_off; /**< GRE offset */
-+ volatile uint8_t l4_off; /**< Layer 4 offset */
-+ volatile uint8_t nxthdr_off; /**< Parser end point */
-+} _PackedType t_FmPrsResult;
-+
-+#endif
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/src/wrapper/lnxwrp_resources_ut.make
-@@ -0,0 +1,28 @@
-+CC=gcc
-+
-+LNXWRP_RESS_UT=lnxwrp_resources_ut
-+OBJ=lnxwrp_resources
-+
-+INC_PATH=
-+LIB_PATH=
-+
-+INC=$(addprefix -I,$(INC_PATH))
-+LIB=$(addprefix -L,$(LIB_PATH))
-+
-+CFLAGS= -gdwarf-2 -g -O0 -Wall
-+XFLAGS= -DFMAN_RESOURCES_UNIT_TEST
-+
-+all: $(LNXWRP_RESS_UT)
-+
-+$(LNXWRP_RESS_UT):$(addsuffix .o,$(OBJ)) $(LNXWRP_RESS_UT).o
-+ $(CC) -o $(LNXWRP_RESS_UT) $(LNXWRP_RESS_UT).o $(addsuffix .o,$(OBJ))
-+
-+%.o: %.c
-+ @(echo " (CC) $@")
-+ @($(CC) $(INC) $(CFLAGS) $(XFLAGS) -o $(@) -c $<)
-+
-+.PHONY: clean
-+
-+clean:
-+ rm -f *.o
-+ rm -f $(LNXWRP_RESS_UT)
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/src/wrapper/lnxwrp_sysfs.c
-@@ -0,0 +1,60 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+/*
-+ @File lnxwrp_sysfs.c
-+
-+ @Description FM wrapper sysfs related functions.
-+
-+*/
-+
-+#include <linux/types.h>
-+#include "lnxwrp_sysfs.h"
-+
-+uint8_t fm_find_statistic_counter_by_name(const char *attr_name,
-+ const struct sysfs_stats_t *sysfs_stats,
-+ uint8_t *offset)
-+{
-+ int i = 0;
-+
-+ while (sysfs_stats[i].stat_name != NULL) {
-+ if (strcmp(sysfs_stats[i].stat_name, attr_name) == 0) {
-+ if (offset != NULL)
-+ *offset = i;
-+ return sysfs_stats[i].stat_counter;
-+ }
-+
-+ i++;
-+ }
-+ WARN(1, "FMD: Should never get here!");
-+ return 0;
-+}
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/src/wrapper/lnxwrp_sysfs.h
-@@ -0,0 +1,60 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#ifndef LNXWRP_SYSFS_H_
-+#define LNXWRP_SYSFS_H_
-+
-+/* Linux Headers ------------------- */
-+#include <linux/version.h>
-+
-+#if defined(CONFIG_MODVERSIONS) && !defined(MODVERSIONS)
-+#define MODVERSIONS
-+#endif
-+#ifdef MODVERSIONS
-+#include <config/modversions.h>
-+#endif /* MODVERSIONS */
-+
-+#include <linux/kernel.h>
-+#include <linux/module.h>
-+#include <linux/device.h>
-+#include <linux/sysfs.h>
-+
-+struct sysfs_stats_t {
-+ const char *stat_name;
-+ uint8_t stat_counter;
-+};
-+
-+uint8_t fm_find_statistic_counter_by_name(const char *attr_name,
-+ const struct sysfs_stats_t *sysfs_stats,
-+ uint8_t *offset);
-+
-+#endif /* LNXWRP_SYSFS_H_ */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/src/wrapper/lnxwrp_sysfs_fm.c
-@@ -0,0 +1,1855 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#include "lnxwrp_sysfs.h"
-+#include "lnxwrp_sysfs_fm.h"
-+#include "lnxwrp_fm.h"
-+
-+#include "../../sdk_fman/Peripherals/FM/inc/fm_common.h"
-+#include "../../sdk_fman/Peripherals/FM/Pcd/fm_pcd.h"
-+#include "../../sdk_fman/Peripherals/FM/Pcd/fm_kg.h"
-+#include "../../sdk_fman/Peripherals/FM/Pcd/fm_plcr.h"
-+
-+#if defined(__ERR_MODULE__)
-+#undef __ERR_MODULE__
-+#endif
-+
-+#include "../../sdk_fman/Peripherals/FM/fm.h"
-+#include <linux/delay.h>
-+
-+
-+static int fm_get_counter(void *h_fm, e_FmCounters cnt_e, uint32_t *cnt_val);
-+
-+enum fm_dma_match_stats {
-+ FM_DMA_COUNTERS_CMQ_NOT_EMPTY,
-+ FM_DMA_COUNTERS_BUS_ERROR,
-+ FM_DMA_COUNTERS_READ_BUF_ECC_ERROR,
-+ FM_DMA_COUNTERS_WRITE_BUF_ECC_SYS_ERROR,
-+ FM_DMA_COUNTERS_WRITE_BUF_ECC_FM_ERROR
-+};
-+
-+static const struct sysfs_stats_t fm_sysfs_stats[] = {
-+ /* FM statistics */
-+ {
-+ .stat_name = "enq_total_frame",
-+ .stat_counter = e_FM_COUNTERS_ENQ_TOTAL_FRAME,
-+ },
-+ {
-+ .stat_name = "deq_total_frame",
-+ .stat_counter = e_FM_COUNTERS_DEQ_TOTAL_FRAME,
-+ },
-+ {
-+ .stat_name = "deq_0",
-+ .stat_counter = e_FM_COUNTERS_DEQ_0,
-+ },
-+ {
-+ .stat_name = "deq_1",
-+ .stat_counter = e_FM_COUNTERS_DEQ_1,
-+ },
-+ {
-+ .stat_name = "deq_2",
-+ .stat_counter = e_FM_COUNTERS_DEQ_2,
-+ },
-+ {
-+ .stat_name = "deq_3",
-+ .stat_counter = e_FM_COUNTERS_DEQ_3,
-+ },
-+ {
-+ .stat_name = "deq_from_default",
-+ .stat_counter = e_FM_COUNTERS_DEQ_FROM_DEFAULT,
-+ },
-+ {
-+ .stat_name = "deq_from_context",
-+ .stat_counter = e_FM_COUNTERS_DEQ_FROM_CONTEXT,
-+ },
-+ {
-+ .stat_name = "deq_from_fd",
-+ .stat_counter = e_FM_COUNTERS_DEQ_FROM_FD,
-+ },
-+ {
-+ .stat_name = "deq_confirm",
-+ .stat_counter = e_FM_COUNTERS_DEQ_CONFIRM,
-+ },
-+ /* FM:DMA statistics */
-+ {
-+ .stat_name = "cmq_not_empty",
-+ .stat_counter = FM_DMA_COUNTERS_CMQ_NOT_EMPTY,
-+ },
-+ {
-+ .stat_name = "bus_error",
-+ .stat_counter = FM_DMA_COUNTERS_BUS_ERROR,
-+ },
-+ {
-+ .stat_name = "read_buf_ecc_error",
-+ .stat_counter = FM_DMA_COUNTERS_READ_BUF_ECC_ERROR,
-+ },
-+ {
-+ .stat_name = "write_buf_ecc_sys_error",
-+ .stat_counter = FM_DMA_COUNTERS_WRITE_BUF_ECC_SYS_ERROR,
-+ },
-+ {
-+ .stat_name = "write_buf_ecc_fm_error",
-+ .stat_counter = FM_DMA_COUNTERS_WRITE_BUF_ECC_FM_ERROR,
-+ },
-+ /* FM:PCD statistics */
-+ {
-+ .stat_name = "pcd_kg_total",
-+ .stat_counter = e_FM_PCD_KG_COUNTERS_TOTAL,
-+ },
-+ {
-+ .stat_name = "pcd_plcr_yellow",
-+ .stat_counter = e_FM_PCD_PLCR_COUNTERS_YELLOW,
-+ },
-+ {
-+ .stat_name = "pcd_plcr_red",
-+ .stat_counter = e_FM_PCD_PLCR_COUNTERS_RED,
-+ },
-+ {
-+ .stat_name = "pcd_plcr_recolored_to_red",
-+ .stat_counter = e_FM_PCD_PLCR_COUNTERS_RECOLORED_TO_RED,
-+ },
-+ {
-+ .stat_name = "pcd_plcr_recolored_to_yellow",
-+ .stat_counter = e_FM_PCD_PLCR_COUNTERS_RECOLORED_TO_YELLOW,
-+ },
-+ {
-+ .stat_name = "pcd_plcr_total",
-+ .stat_counter = e_FM_PCD_PLCR_COUNTERS_TOTAL,
-+ },
-+ {
-+ .stat_name = "pcd_plcr_length_mismatch",
-+ .stat_counter = e_FM_PCD_PLCR_COUNTERS_LENGTH_MISMATCH,
-+ },
-+ {
-+ .stat_name = "pcd_prs_parse_dispatch",
-+ .stat_counter = e_FM_PCD_PRS_COUNTERS_PARSE_DISPATCH,
-+ },
-+ {
-+ .stat_name = "pcd_prs_l2_parse_result_returned",
-+ .stat_counter = e_FM_PCD_PRS_COUNTERS_L2_PARSE_RESULT_RETURNED,
-+ },
-+ {
-+ .stat_name = "pcd_prs_l3_parse_result_returned",
-+ .stat_counter = e_FM_PCD_PRS_COUNTERS_L3_PARSE_RESULT_RETURNED,
-+ },
-+ {
-+ .stat_name = "pcd_prs_l4_parse_result_returned",
-+ .stat_counter = e_FM_PCD_PRS_COUNTERS_L4_PARSE_RESULT_RETURNED,
-+ },
-+ {
-+ .stat_name = "pcd_prs_shim_parse_result_returned",
-+ .stat_counter = e_FM_PCD_PRS_COUNTERS_SHIM_PARSE_RESULT_RETURNED,
-+ },
-+ {
-+ .stat_name = "pcd_prs_l2_parse_result_returned_with_err",
-+ .stat_counter =
-+ e_FM_PCD_PRS_COUNTERS_L2_PARSE_RESULT_RETURNED_WITH_ERR,
-+ },
-+ {
-+ .stat_name = "pcd_prs_l3_parse_result_returned_with_err",
-+ .stat_counter =
-+ e_FM_PCD_PRS_COUNTERS_L3_PARSE_RESULT_RETURNED_WITH_ERR,
-+ },
-+ {
-+ .stat_name = "pcd_prs_l4_parse_result_returned_with_err",
-+ .stat_counter =
-+ e_FM_PCD_PRS_COUNTERS_L4_PARSE_RESULT_RETURNED_WITH_ERR,
-+ },
-+ {
-+ .stat_name = "pcd_prs_shim_parse_result_returned_with_err",
-+ .stat_counter =
-+ e_FM_PCD_PRS_COUNTERS_SHIM_PARSE_RESULT_RETURNED_WITH_ERR,
-+ },
-+ {
-+ .stat_name = "pcd_prs_soft_prs_cycles",
-+ .stat_counter = e_FM_PCD_PRS_COUNTERS_SOFT_PRS_CYCLES,
-+ },
-+ {
-+ .stat_name = "pcd_prs_soft_prs_stall_cycles",
-+ .stat_counter = e_FM_PCD_PRS_COUNTERS_SOFT_PRS_STALL_CYCLES,
-+ },
-+ {
-+ .stat_name = "pcd_prs_hard_prs_cycle_incl_stall_cycles",
-+ .stat_counter =
-+ e_FM_PCD_PRS_COUNTERS_HARD_PRS_CYCLE_INCL_STALL_CYCLES,
-+ },
-+ {
-+ .stat_name = "pcd_prs_muram_read_cycles",
-+ .stat_counter = e_FM_PCD_PRS_COUNTERS_MURAM_READ_CYCLES,
-+ },
-+ {
-+ .stat_name = "pcd_prs_muram_read_stall_cycles",
-+ .stat_counter = e_FM_PCD_PRS_COUNTERS_MURAM_READ_STALL_CYCLES,
-+ },
-+ {
-+ .stat_name = "pcd_prs_muram_write_cycles",
-+ .stat_counter = e_FM_PCD_PRS_COUNTERS_MURAM_WRITE_CYCLES,
-+ },
-+ {
-+ .stat_name = "pcd_prs_muram_write_stall_cycles",
-+ .stat_counter = e_FM_PCD_PRS_COUNTERS_MURAM_WRITE_STALL_CYCLES,
-+ },
-+ {
-+ .stat_name = "pcd_prs_fpm_command_stall_cycles",
-+ .stat_counter = e_FM_PCD_PRS_COUNTERS_FPM_COMMAND_STALL_CYCLES,
-+ },
-+ {}
-+};
-+
-+
-+static ssize_t show_fm_risc_load(struct device *dev,
-+ struct device_attribute *attr, char *buf)
-+{
-+ t_LnxWrpFmDev *p_wrp_fm_dev = NULL;
-+ unsigned long flags;
-+ int m =0;
-+ int err =0;
-+ unsigned n = 0;
-+ t_FmCtrlMon util;
-+ uint8_t i =0 ;
-+
-+ if (attr == NULL || buf == NULL || dev == NULL)
-+ return -EINVAL;
-+
-+ p_wrp_fm_dev = (t_LnxWrpFmDev *) dev_get_drvdata(dev);
-+ if (WARN_ON(p_wrp_fm_dev == NULL))
-+ return -EINVAL;
-+
-+ if (!p_wrp_fm_dev->active || !p_wrp_fm_dev->h_Dev)
-+ return -EIO;
-+
-+ local_irq_save(flags);
-+
-+ /* Calculate risc load */
-+ FM_CtrlMonStart(p_wrp_fm_dev->h_Dev);
-+ msleep(1000);
-+ FM_CtrlMonStop(p_wrp_fm_dev->h_Dev);
-+
-+ for (i = 0; i < FM_NUM_OF_CTRL; i++) {
-+ err |= FM_CtrlMonGetCounters(p_wrp_fm_dev->h_Dev, i, &util);
-+ m = snprintf(&buf[n],PAGE_SIZE,"\tRisc%u: util-%u%%, efficiency-%u%%\n",
-+ i, util.percentCnt[0], util.percentCnt[1]);
-+ n=m+n;
-+ }
-+
-+ local_irq_restore(flags);
-+
-+ return n;
-+}
-+
-+/* Fm stats and regs dumps via sysfs */
-+static ssize_t show_fm_dma_stats(struct device *dev,
-+ struct device_attribute *attr, char *buf)
-+{
-+ t_LnxWrpFmDev *p_wrp_fm_dev = NULL;
-+ t_FmDmaStatus dma_status;
-+ unsigned long flags = 0;
-+ unsigned n = 0;
-+ uint8_t counter_value = 0, counter = 0;
-+
-+ if (attr == NULL || buf == NULL || dev == NULL)
-+ return -EINVAL;
-+
-+ p_wrp_fm_dev = (t_LnxWrpFmDev *) dev_get_drvdata(dev);
-+ if (WARN_ON(p_wrp_fm_dev == NULL))
-+ return -EINVAL;
-+
-+ if (!p_wrp_fm_dev->active || !p_wrp_fm_dev->h_Dev)
-+ return -EIO;
-+
-+ counter = fm_find_statistic_counter_by_name(
-+ attr->attr.name,
-+ fm_sysfs_stats, NULL);
-+
-+ local_irq_save(flags);
-+
-+ memset(&dma_status, 0, sizeof(dma_status));
-+ FM_GetDmaStatus(p_wrp_fm_dev->h_Dev, &dma_status);
-+
-+ switch (counter) {
-+ case FM_DMA_COUNTERS_CMQ_NOT_EMPTY:
-+ counter_value = dma_status.cmqNotEmpty;
-+ break;
-+ case FM_DMA_COUNTERS_BUS_ERROR:
-+ counter_value = dma_status.busError;
-+ break;
-+ case FM_DMA_COUNTERS_READ_BUF_ECC_ERROR:
-+ counter_value = dma_status.readBufEccError;
-+ break;
-+ case FM_DMA_COUNTERS_WRITE_BUF_ECC_SYS_ERROR:
-+ counter_value = dma_status.writeBufEccSysError;
-+ break;
-+ case FM_DMA_COUNTERS_WRITE_BUF_ECC_FM_ERROR:
-+ counter_value = dma_status.writeBufEccFmError;
-+ break;
-+ default:
-+ WARN(1, "FMD: failure at %s:%d/%s()!\n", __FILE__, __LINE__,
-+ __func__);
-+ break;
-+ };
-+
-+ n = snprintf(buf, PAGE_SIZE, "\tFM %u counter: %c\n",
-+ p_wrp_fm_dev->id, counter_value ? 'T' : 'F');
-+
-+ local_irq_restore(flags);
-+
-+ return n;
-+}
-+
-+static ssize_t show_fm_stats(struct device *dev,
-+ struct device_attribute *attr, char *buf)
-+{
-+ t_LnxWrpFmDev *p_wrp_fm_dev = NULL;
-+ unsigned long flags = 0;
-+ unsigned n = 0, cnt_e = 0;
-+ uint32_t cnt_val;
-+ int err;
-+
-+ if (attr == NULL || buf == NULL || dev == NULL)
-+ return -EINVAL;
-+
-+ p_wrp_fm_dev = (t_LnxWrpFmDev *) dev_get_drvdata(dev);
-+ if (WARN_ON(p_wrp_fm_dev == NULL))
-+ return -EINVAL;
-+
-+ if (!p_wrp_fm_dev->active || !p_wrp_fm_dev->h_Dev)
-+ return -EIO;
-+
-+ cnt_e = fm_find_statistic_counter_by_name(
-+ attr->attr.name,
-+ fm_sysfs_stats, NULL);
-+
-+ err = fm_get_counter(p_wrp_fm_dev->h_Dev,
-+ (e_FmCounters) cnt_e, &cnt_val);
-+
-+ if (err)
-+ return err;
-+
-+ local_irq_save(flags);
-+
-+ n = snprintf(buf, PAGE_SIZE, "\tFM %d counter: %d\n",
-+ p_wrp_fm_dev->id, cnt_val);
-+
-+ local_irq_restore(flags);
-+
-+ return n;
-+}
-+
-+static ssize_t show_fm_muram_free_sz(struct device *dev,
-+ struct device_attribute *attr, char *buf)
-+{
-+ t_LnxWrpFmDev *p_wrp_fm_dev = NULL;
-+ unsigned long flags = 0;
-+ unsigned n = 0;
-+ uint64_t muram_free_size = 0;
-+
-+ if (attr == NULL || buf == NULL || dev == NULL)
-+ return -EINVAL;
-+
-+ p_wrp_fm_dev = (t_LnxWrpFmDev *) dev_get_drvdata(dev);
-+ if (WARN_ON(p_wrp_fm_dev == NULL))
-+ return -EINVAL;
-+
-+ if (!p_wrp_fm_dev->active || !p_wrp_fm_dev->h_Dev)
-+ return -EIO;
-+
-+ muram_free_size = FM_MURAM_GetFreeMemSize(p_wrp_fm_dev->h_MuramDev);
-+
-+ local_irq_save(flags);
-+
-+ n = snprintf(buf, PAGE_SIZE, "\tFM %d muram_free_size: %lld\n",
-+ p_wrp_fm_dev->id, muram_free_size);
-+
-+ local_irq_restore(flags);
-+
-+ return n;
-+}
-+
-+static ssize_t show_fm_ctrl_code_ver(struct device *dev,
-+ struct device_attribute *attr, char *buf)
-+{
-+ t_LnxWrpFmDev *p_wrp_fm_dev = NULL;
-+ unsigned long flags = 0;
-+ unsigned n = 0;
-+ t_FmCtrlCodeRevisionInfo rv_info;
-+
-+ if (attr == NULL || buf == NULL || dev == NULL)
-+ return -EINVAL;
-+
-+ p_wrp_fm_dev = (t_LnxWrpFmDev *) dev_get_drvdata(dev);
-+ if (WARN_ON(p_wrp_fm_dev == NULL))
-+ return -EINVAL;
-+
-+ if (!p_wrp_fm_dev->active || !p_wrp_fm_dev->h_Dev)
-+ return -EIO;
-+
-+ FM_GetFmanCtrlCodeRevision((t_Fm *)p_wrp_fm_dev->h_Dev, &rv_info);
-+
-+ local_irq_save(flags);
-+
-+ FM_DMP_LN(buf, n, "- FM %d ctrl code pkg info:\n", p_wrp_fm_dev->id);
-+ FM_DMP_LN(buf, n, "Package rev: %d\n", rv_info.packageRev);
-+ FM_DMP_LN(buf, n, "major rev: %d\n", rv_info.majorRev);
-+ FM_DMP_LN(buf, n, "minor rev: %d\n", rv_info.minorRev);
-+
-+ local_irq_restore(flags);
-+
-+ return n;
-+}
-+
-+static ssize_t show_fm_pcd_stats(struct device *dev,
-+ struct device_attribute *attr, char *buf)
-+{
-+ t_LnxWrpFmDev *p_wrp_fm_dev = NULL;
-+ unsigned long flags = 0;
-+ unsigned n = 0, counter = 0;
-+
-+ if (attr == NULL || buf == NULL || dev == NULL)
-+ return -EINVAL;
-+
-+ p_wrp_fm_dev = (t_LnxWrpFmDev *) dev_get_drvdata(dev);
-+ if (WARN_ON(p_wrp_fm_dev == NULL))
-+ return -EINVAL;
-+
-+ if (!p_wrp_fm_dev->active || !p_wrp_fm_dev->h_Dev ||
-+ !p_wrp_fm_dev->h_PcdDev)
-+ return -EIO;
-+
-+ counter = fm_find_statistic_counter_by_name(
-+ attr->attr.name,
-+ fm_sysfs_stats, NULL);
-+
-+ local_irq_save(flags);
-+
-+ n = snprintf(buf, PAGE_SIZE, "\tFM %d counter: %d\n",
-+ p_wrp_fm_dev->id,
-+ FM_PCD_GetCounter(p_wrp_fm_dev->h_PcdDev,
-+ (e_FmPcdCounters) counter));
-+
-+ local_irq_restore(flags);
-+
-+ return n;
-+}
-+
-+static ssize_t show_fm_tnum_dbg(struct device *dev,
-+ struct device_attribute *attr,
-+ char *buf)
-+{
-+ unsigned long flags;
-+ unsigned n = 0;
-+#if (defined(DEBUG_ERRORS) && (DEBUG_ERRORS > 0))
-+ t_LnxWrpFmDev *p_wrp_fm_dev = NULL;
-+#endif
-+
-+ if (attr == NULL || buf == NULL || dev == NULL)
-+ return -EINVAL;
-+
-+#if (defined(DEBUG_ERRORS) && (DEBUG_ERRORS > 0))
-+
-+ p_wrp_fm_dev = (t_LnxWrpFmDev *) dev_get_drvdata(dev);
-+ if (WARN_ON(p_wrp_fm_dev == NULL))
-+ return -EINVAL;
-+
-+ local_irq_save(flags);
-+
-+ if (!p_wrp_fm_dev->active)
-+ return -EIO;
-+ else {
-+ int tn_s;
-+
-+ if (!sscanf(attr->attr.name, "tnum_dbg_%d", &tn_s))
-+ return -EINVAL;
-+
-+ n = fm_dump_tnum_dbg(p_wrp_fm_dev->h_Dev,
-+ tn_s, tn_s + 15, buf, n);
-+ }
-+ local_irq_restore(flags);
-+#else
-+
-+ local_irq_save(flags);
-+ n = snprintf(buf, PAGE_SIZE,
-+ "Debug level is too low to dump registers!!!\n");
-+ local_irq_restore(flags);
-+#endif /* (defined(DEBUG_ERRORS) && ... */
-+
-+ return n;
-+}
-+
-+static ssize_t show_fm_cls_plan(struct device *dev,
-+ struct device_attribute *attr,
-+ char *buf)
-+{
-+ unsigned long flags;
-+ unsigned n = 0;
-+#if (defined(DEBUG_ERRORS) && (DEBUG_ERRORS > 0))
-+ t_LnxWrpFmDev *p_wrp_fm_dev = NULL;
-+#endif
-+
-+ if (attr == NULL || buf == NULL || dev == NULL)
-+ return -EINVAL;
-+
-+#if (defined(DEBUG_ERRORS) && (DEBUG_ERRORS > 0))
-+ p_wrp_fm_dev = (t_LnxWrpFmDev *) dev_get_drvdata(dev);
-+ if (WARN_ON(p_wrp_fm_dev == NULL))
-+ return -EINVAL;
-+
-+ local_irq_save(flags);
-+
-+ n = snprintf(buf, PAGE_SIZE, "\n FM-KG classification plan dump.\n");
-+
-+ if (!p_wrp_fm_dev->active || !p_wrp_fm_dev->h_PcdDev)
-+ return -EIO;
-+ else {
-+ int cpn;
-+
-+ if (!sscanf(attr->attr.name, "cls_plan_%d", &cpn))
-+ return -EINVAL;
-+
-+ n = fm_dump_cls_plan(p_wrp_fm_dev->h_PcdDev, cpn, buf, n);
-+ }
-+ local_irq_restore(flags);
-+#else
-+ local_irq_save(flags);
-+ n = snprintf(buf, PAGE_SIZE,
-+ "Debug level is too low to dump registers!!!\n");
-+ local_irq_restore(flags);
-+#endif /* (defined(DEBUG_ERRORS) && ... */
-+
-+ return n;
-+}
-+
-+static ssize_t show_fm_profiles(struct device *dev,
-+ struct device_attribute *attr,
-+ char *buf)
-+{
-+ unsigned long flags;
-+ unsigned n = 0;
-+#if (defined(DEBUG_ERRORS) && (DEBUG_ERRORS > 0))
-+ t_LnxWrpFmDev *p_wrp_fm_dev = NULL;
-+#endif
-+
-+ if (attr == NULL || buf == NULL || dev == NULL)
-+ return -EINVAL;
-+
-+#if (defined(DEBUG_ERRORS) && (DEBUG_ERRORS > 0))
-+
-+ p_wrp_fm_dev = (t_LnxWrpFmDev *) dev_get_drvdata(dev);
-+ if (WARN_ON(p_wrp_fm_dev == NULL))
-+ return -EINVAL;
-+
-+ local_irq_save(flags);
-+
-+ n = snprintf(buf, PAGE_SIZE, "FM policer profile dump.\n");
-+
-+ if (!p_wrp_fm_dev->active || !p_wrp_fm_dev->h_PcdDev)
-+ return -EIO;
-+ else {
-+ int pn;
-+
-+ if (!sscanf(attr->attr.name, "profile_%d", &pn))
-+ return -EINVAL;
-+
-+ n = fm_profile_dump_regs(p_wrp_fm_dev->h_PcdDev, pn, buf, n);
-+ }
-+ local_irq_restore(flags);
-+#else
-+ local_irq_save(flags);
-+ n = snprintf(buf, PAGE_SIZE,
-+ "Debug level is too low to dump registers!!!\n");
-+ local_irq_restore(flags);
-+#endif /* (defined(DEBUG_ERRORS) && ... */
-+
-+ return n;
-+}
-+
-+static ssize_t show_fm_schemes(struct device *dev,
-+ struct device_attribute *attr,
-+ char *buf)
-+{
-+ unsigned long flags;
-+ unsigned n = 0;
-+#if (defined(DEBUG_ERRORS) && (DEBUG_ERRORS > 0))
-+ t_LnxWrpFmDev *p_wrp_fm_dev = NULL;
-+#endif
-+
-+ if (attr == NULL || buf == NULL || dev == NULL)
-+ return -EINVAL;
-+
-+#if (defined(DEBUG_ERRORS) && (DEBUG_ERRORS > 0))
-+
-+ p_wrp_fm_dev = (t_LnxWrpFmDev *) dev_get_drvdata(dev);
-+ if (WARN_ON(p_wrp_fm_dev == NULL))
-+ return -EINVAL;
-+
-+ local_irq_save(flags);
-+
-+ n = snprintf(buf, PAGE_SIZE, "FM-KG driver schemes dump.\n");
-+
-+ if (!p_wrp_fm_dev->active || !p_wrp_fm_dev->h_PcdDev)
-+ return -EIO;
-+ else {
-+ int sn;
-+
-+ if (!sscanf(attr->attr.name, "scheme_%d", &sn))
-+ return -EINVAL;
-+
-+ n = fm_dump_scheme(p_wrp_fm_dev->h_PcdDev, sn, buf, n);
-+ }
-+ local_irq_restore(flags);
-+#else
-+
-+ local_irq_save(flags);
-+ n = snprintf(buf, PAGE_SIZE,
-+ "Debug level is too low to dump registers!!!\n");
-+ local_irq_restore(flags);
-+#endif /* (defined(DEBUG_ERRORS) && ... */
-+
-+ return n;
-+}
-+
-+/* FM */
-+static DEVICE_ATTR(enq_total_frame, S_IRUGO, show_fm_stats, NULL);
-+static DEVICE_ATTR(deq_total_frame, S_IRUGO, show_fm_stats, NULL);
-+static DEVICE_ATTR(fm_risc_load_val, S_IRUGO, show_fm_risc_load, NULL);
-+static DEVICE_ATTR(deq_0, S_IRUGO, show_fm_stats, NULL);
-+static DEVICE_ATTR(deq_1, S_IRUGO, show_fm_stats, NULL);
-+static DEVICE_ATTR(deq_2, S_IRUGO, show_fm_stats, NULL);
-+static DEVICE_ATTR(deq_3, S_IRUGO, show_fm_stats, NULL);
-+static DEVICE_ATTR(deq_from_default, S_IRUGO, show_fm_stats, NULL);
-+static DEVICE_ATTR(deq_from_context, S_IRUGO, show_fm_stats, NULL);
-+static DEVICE_ATTR(deq_from_fd, S_IRUGO, show_fm_stats, NULL);
-+static DEVICE_ATTR(deq_confirm, S_IRUGO, show_fm_stats, NULL);
-+/* FM:DMA */
-+static DEVICE_ATTR(cmq_not_empty, S_IRUGO, show_fm_dma_stats, NULL);
-+static DEVICE_ATTR(bus_error, S_IRUGO, show_fm_dma_stats, NULL);
-+static DEVICE_ATTR(read_buf_ecc_error, S_IRUGO, show_fm_dma_stats, NULL);
-+static DEVICE_ATTR(write_buf_ecc_sys_error, S_IRUGO, show_fm_dma_stats, NULL);
-+static DEVICE_ATTR(write_buf_ecc_fm_error, S_IRUGO, show_fm_dma_stats, NULL);
-+/* FM:PCD */
-+static DEVICE_ATTR(pcd_kg_total, S_IRUGO, show_fm_pcd_stats, NULL);
-+static DEVICE_ATTR(pcd_plcr_yellow, S_IRUGO, show_fm_pcd_stats, NULL);
-+static DEVICE_ATTR(pcd_plcr_red, S_IRUGO, show_fm_pcd_stats, NULL);
-+static DEVICE_ATTR(pcd_plcr_recolored_to_red, S_IRUGO, show_fm_pcd_stats,
-+ NULL);
-+static DEVICE_ATTR(pcd_plcr_recolored_to_yellow, S_IRUGO, show_fm_pcd_stats,
-+ NULL);
-+static DEVICE_ATTR(pcd_plcr_total, S_IRUGO, show_fm_pcd_stats, NULL);
-+static DEVICE_ATTR(pcd_plcr_length_mismatch, S_IRUGO, show_fm_pcd_stats,
-+ NULL);
-+static DEVICE_ATTR(pcd_prs_parse_dispatch, S_IRUGO, show_fm_pcd_stats, NULL);
-+static DEVICE_ATTR(pcd_prs_l2_parse_result_returned, S_IRUGO,
-+ show_fm_pcd_stats, NULL);
-+static DEVICE_ATTR(pcd_prs_l3_parse_result_returned, S_IRUGO,
-+ show_fm_pcd_stats, NULL);
-+static DEVICE_ATTR(pcd_prs_l4_parse_result_returned, S_IRUGO,
-+ show_fm_pcd_stats, NULL);
-+static DEVICE_ATTR(pcd_prs_shim_parse_result_returned, S_IRUGO,
-+ show_fm_pcd_stats, NULL);
-+static DEVICE_ATTR(pcd_prs_l2_parse_result_returned_with_err, S_IRUGO,
-+ show_fm_pcd_stats, NULL);
-+static DEVICE_ATTR(pcd_prs_l3_parse_result_returned_with_err, S_IRUGO,
-+ show_fm_pcd_stats, NULL);
-+static DEVICE_ATTR(pcd_prs_l4_parse_result_returned_with_err, S_IRUGO,
-+ show_fm_pcd_stats, NULL);
-+static DEVICE_ATTR(pcd_prs_shim_parse_result_returned_with_err, S_IRUGO,
-+ show_fm_pcd_stats, NULL);
-+static DEVICE_ATTR(pcd_prs_soft_prs_cycles, S_IRUGO, show_fm_pcd_stats, NULL);
-+static DEVICE_ATTR(pcd_prs_soft_prs_stall_cycles, S_IRUGO, show_fm_pcd_stats,
-+ NULL);
-+static DEVICE_ATTR(pcd_prs_hard_prs_cycle_incl_stall_cycles, S_IRUGO,
-+ show_fm_pcd_stats, NULL);
-+static DEVICE_ATTR(pcd_prs_muram_read_cycles, S_IRUGO, show_fm_pcd_stats,
-+ NULL);
-+static DEVICE_ATTR(pcd_prs_muram_read_stall_cycles, S_IRUGO,
-+ show_fm_pcd_stats, NULL);
-+static DEVICE_ATTR(pcd_prs_muram_write_cycles, S_IRUGO, show_fm_pcd_stats,
-+ NULL);
-+static DEVICE_ATTR(pcd_prs_muram_write_stall_cycles, S_IRUGO,
-+ show_fm_pcd_stats, NULL);
-+static DEVICE_ATTR(pcd_prs_fpm_command_stall_cycles, S_IRUGO,
-+ show_fm_pcd_stats, NULL);
-+
-+static DEVICE_ATTR(tnum_dbg_0, S_IRUGO, show_fm_tnum_dbg, NULL);
-+static DEVICE_ATTR(tnum_dbg_16, S_IRUGO, show_fm_tnum_dbg, NULL);
-+static DEVICE_ATTR(tnum_dbg_32, S_IRUGO, show_fm_tnum_dbg, NULL);
-+static DEVICE_ATTR(tnum_dbg_48, S_IRUGO, show_fm_tnum_dbg, NULL);
-+static DEVICE_ATTR(tnum_dbg_64, S_IRUGO, show_fm_tnum_dbg, NULL);
-+static DEVICE_ATTR(tnum_dbg_80, S_IRUGO, show_fm_tnum_dbg, NULL);
-+static DEVICE_ATTR(tnum_dbg_96, S_IRUGO, show_fm_tnum_dbg, NULL);
-+static DEVICE_ATTR(tnum_dbg_112, S_IRUGO, show_fm_tnum_dbg, NULL);
-+
-+static DEVICE_ATTR(cls_plan_0, S_IRUGO, show_fm_cls_plan, NULL);
-+static DEVICE_ATTR(cls_plan_1, S_IRUGO, show_fm_cls_plan, NULL);
-+static DEVICE_ATTR(cls_plan_2, S_IRUGO, show_fm_cls_plan, NULL);
-+static DEVICE_ATTR(cls_plan_3, S_IRUGO, show_fm_cls_plan, NULL);
-+static DEVICE_ATTR(cls_plan_4, S_IRUGO, show_fm_cls_plan, NULL);
-+static DEVICE_ATTR(cls_plan_5, S_IRUGO, show_fm_cls_plan, NULL);
-+static DEVICE_ATTR(cls_plan_6, S_IRUGO, show_fm_cls_plan, NULL);
-+static DEVICE_ATTR(cls_plan_7, S_IRUGO, show_fm_cls_plan, NULL);
-+static DEVICE_ATTR(cls_plan_8, S_IRUGO, show_fm_cls_plan, NULL);
-+static DEVICE_ATTR(cls_plan_9, S_IRUGO, show_fm_cls_plan, NULL);
-+static DEVICE_ATTR(cls_plan_10, S_IRUGO, show_fm_cls_plan, NULL);
-+static DEVICE_ATTR(cls_plan_11, S_IRUGO, show_fm_cls_plan, NULL);
-+static DEVICE_ATTR(cls_plan_12, S_IRUGO, show_fm_cls_plan, NULL);
-+static DEVICE_ATTR(cls_plan_13, S_IRUGO, show_fm_cls_plan, NULL);
-+static DEVICE_ATTR(cls_plan_14, S_IRUGO, show_fm_cls_plan, NULL);
-+static DEVICE_ATTR(cls_plan_15, S_IRUGO, show_fm_cls_plan, NULL);
-+static DEVICE_ATTR(cls_plan_16, S_IRUGO, show_fm_cls_plan, NULL);
-+static DEVICE_ATTR(cls_plan_17, S_IRUGO, show_fm_cls_plan, NULL);
-+static DEVICE_ATTR(cls_plan_18, S_IRUGO, show_fm_cls_plan, NULL);
-+static DEVICE_ATTR(cls_plan_19, S_IRUGO, show_fm_cls_plan, NULL);
-+static DEVICE_ATTR(cls_plan_20, S_IRUGO, show_fm_cls_plan, NULL);
-+static DEVICE_ATTR(cls_plan_21, S_IRUGO, show_fm_cls_plan, NULL);
-+static DEVICE_ATTR(cls_plan_22, S_IRUGO, show_fm_cls_plan, NULL);
-+static DEVICE_ATTR(cls_plan_23, S_IRUGO, show_fm_cls_plan, NULL);
-+static DEVICE_ATTR(cls_plan_24, S_IRUGO, show_fm_cls_plan, NULL);
-+static DEVICE_ATTR(cls_plan_25, S_IRUGO, show_fm_cls_plan, NULL);
-+static DEVICE_ATTR(cls_plan_26, S_IRUGO, show_fm_cls_plan, NULL);
-+static DEVICE_ATTR(cls_plan_27, S_IRUGO, show_fm_cls_plan, NULL);
-+static DEVICE_ATTR(cls_plan_28, S_IRUGO, show_fm_cls_plan, NULL);
-+static DEVICE_ATTR(cls_plan_29, S_IRUGO, show_fm_cls_plan, NULL);
-+static DEVICE_ATTR(cls_plan_30, S_IRUGO, show_fm_cls_plan, NULL);
-+static DEVICE_ATTR(cls_plan_31, S_IRUGO, show_fm_cls_plan, NULL);
-+
-+static DEVICE_ATTR(profile_0, S_IRUGO, show_fm_profiles, NULL);
-+static DEVICE_ATTR(profile_1, S_IRUGO, show_fm_profiles, NULL);
-+static DEVICE_ATTR(profile_2, S_IRUGO, show_fm_profiles, NULL);
-+static DEVICE_ATTR(profile_3, S_IRUGO, show_fm_profiles, NULL);
-+static DEVICE_ATTR(profile_4, S_IRUGO, show_fm_profiles, NULL);
-+static DEVICE_ATTR(profile_5, S_IRUGO, show_fm_profiles, NULL);
-+static DEVICE_ATTR(profile_6, S_IRUGO, show_fm_profiles, NULL);
-+static DEVICE_ATTR(profile_7, S_IRUGO, show_fm_profiles, NULL);
-+static DEVICE_ATTR(profile_8, S_IRUGO, show_fm_profiles, NULL);
-+static DEVICE_ATTR(profile_9, S_IRUGO, show_fm_profiles, NULL);
-+static DEVICE_ATTR(profile_10, S_IRUGO, show_fm_profiles, NULL);
-+static DEVICE_ATTR(profile_11, S_IRUGO, show_fm_profiles, NULL);
-+static DEVICE_ATTR(profile_12, S_IRUGO, show_fm_profiles, NULL);
-+static DEVICE_ATTR(profile_13, S_IRUGO, show_fm_profiles, NULL);
-+static DEVICE_ATTR(profile_14, S_IRUGO, show_fm_profiles, NULL);
-+static DEVICE_ATTR(profile_15, S_IRUGO, show_fm_profiles, NULL);
-+static DEVICE_ATTR(profile_16, S_IRUGO, show_fm_profiles, NULL);
-+static DEVICE_ATTR(profile_17, S_IRUGO, show_fm_profiles, NULL);
-+static DEVICE_ATTR(profile_18, S_IRUGO, show_fm_profiles, NULL);
-+static DEVICE_ATTR(profile_19, S_IRUGO, show_fm_profiles, NULL);
-+static DEVICE_ATTR(profile_20, S_IRUGO, show_fm_profiles, NULL);
-+static DEVICE_ATTR(profile_21, S_IRUGO, show_fm_profiles, NULL);
-+static DEVICE_ATTR(profile_22, S_IRUGO, show_fm_profiles, NULL);
-+static DEVICE_ATTR(profile_23, S_IRUGO, show_fm_profiles, NULL);
-+static DEVICE_ATTR(profile_24, S_IRUGO, show_fm_profiles, NULL);
-+static DEVICE_ATTR(profile_25, S_IRUGO, show_fm_profiles, NULL);
-+static DEVICE_ATTR(profile_26, S_IRUGO, show_fm_profiles, NULL);
-+static DEVICE_ATTR(profile_27, S_IRUGO, show_fm_profiles, NULL);
-+static DEVICE_ATTR(profile_28, S_IRUGO, show_fm_profiles, NULL);
-+static DEVICE_ATTR(profile_29, S_IRUGO, show_fm_profiles, NULL);
-+static DEVICE_ATTR(profile_30, S_IRUGO, show_fm_profiles, NULL);
-+static DEVICE_ATTR(profile_31, S_IRUGO, show_fm_profiles, NULL);
-+
-+static DEVICE_ATTR(scheme_0, S_IRUGO, show_fm_schemes, NULL);
-+static DEVICE_ATTR(scheme_1, S_IRUGO, show_fm_schemes, NULL);
-+static DEVICE_ATTR(scheme_2, S_IRUGO, show_fm_schemes, NULL);
-+static DEVICE_ATTR(scheme_3, S_IRUGO, show_fm_schemes, NULL);
-+static DEVICE_ATTR(scheme_4, S_IRUGO, show_fm_schemes, NULL);
-+static DEVICE_ATTR(scheme_5, S_IRUGO, show_fm_schemes, NULL);
-+static DEVICE_ATTR(scheme_6, S_IRUGO, show_fm_schemes, NULL);
-+static DEVICE_ATTR(scheme_7, S_IRUGO, show_fm_schemes, NULL);
-+static DEVICE_ATTR(scheme_8, S_IRUGO, show_fm_schemes, NULL);
-+static DEVICE_ATTR(scheme_9, S_IRUGO, show_fm_schemes, NULL);
-+static DEVICE_ATTR(scheme_10, S_IRUGO, show_fm_schemes, NULL);
-+static DEVICE_ATTR(scheme_11, S_IRUGO, show_fm_schemes, NULL);
-+static DEVICE_ATTR(scheme_12, S_IRUGO, show_fm_schemes, NULL);
-+static DEVICE_ATTR(scheme_13, S_IRUGO, show_fm_schemes, NULL);
-+static DEVICE_ATTR(scheme_14, S_IRUGO, show_fm_schemes, NULL);
-+static DEVICE_ATTR(scheme_15, S_IRUGO, show_fm_schemes, NULL);
-+static DEVICE_ATTR(scheme_16, S_IRUGO, show_fm_schemes, NULL);
-+static DEVICE_ATTR(scheme_17, S_IRUGO, show_fm_schemes, NULL);
-+static DEVICE_ATTR(scheme_18, S_IRUGO, show_fm_schemes, NULL);
-+static DEVICE_ATTR(scheme_19, S_IRUGO, show_fm_schemes, NULL);
-+static DEVICE_ATTR(scheme_20, S_IRUGO, show_fm_schemes, NULL);
-+static DEVICE_ATTR(scheme_21, S_IRUGO, show_fm_schemes, NULL);
-+static DEVICE_ATTR(scheme_22, S_IRUGO, show_fm_schemes, NULL);
-+static DEVICE_ATTR(scheme_23, S_IRUGO, show_fm_schemes, NULL);
-+static DEVICE_ATTR(scheme_24, S_IRUGO, show_fm_schemes, NULL);
-+static DEVICE_ATTR(scheme_25, S_IRUGO, show_fm_schemes, NULL);
-+static DEVICE_ATTR(scheme_26, S_IRUGO, show_fm_schemes, NULL);
-+static DEVICE_ATTR(scheme_27, S_IRUGO, show_fm_schemes, NULL);
-+static DEVICE_ATTR(scheme_28, S_IRUGO, show_fm_schemes, NULL);
-+static DEVICE_ATTR(scheme_29, S_IRUGO, show_fm_schemes, NULL);
-+static DEVICE_ATTR(scheme_30, S_IRUGO, show_fm_schemes, NULL);
-+static DEVICE_ATTR(scheme_31, S_IRUGO, show_fm_schemes, NULL);
-+
-+
-+static struct attribute *fm_dev_stats_attributes[] = {
-+ &dev_attr_enq_total_frame.attr,
-+ &dev_attr_deq_total_frame.attr,
-+ &dev_attr_deq_0.attr,
-+ &dev_attr_deq_1.attr,
-+ &dev_attr_deq_2.attr,
-+ &dev_attr_deq_3.attr,
-+ &dev_attr_deq_from_default.attr,
-+ &dev_attr_deq_from_context.attr,
-+ &dev_attr_deq_from_fd.attr,
-+ &dev_attr_deq_confirm.attr,
-+ &dev_attr_cmq_not_empty.attr,
-+ &dev_attr_bus_error.attr,
-+ &dev_attr_read_buf_ecc_error.attr,
-+ &dev_attr_write_buf_ecc_sys_error.attr,
-+ &dev_attr_write_buf_ecc_fm_error.attr,
-+ &dev_attr_pcd_kg_total.attr,
-+ &dev_attr_pcd_plcr_yellow.attr,
-+ &dev_attr_pcd_plcr_red.attr,
-+ &dev_attr_pcd_plcr_recolored_to_red.attr,
-+ &dev_attr_pcd_plcr_recolored_to_yellow.attr,
-+ &dev_attr_pcd_plcr_total.attr,
-+ &dev_attr_pcd_plcr_length_mismatch.attr,
-+ &dev_attr_pcd_prs_parse_dispatch.attr,
-+ &dev_attr_pcd_prs_l2_parse_result_returned.attr,
-+ &dev_attr_pcd_prs_l3_parse_result_returned.attr,
-+ &dev_attr_pcd_prs_l4_parse_result_returned.attr,
-+ &dev_attr_pcd_prs_shim_parse_result_returned.attr,
-+ &dev_attr_pcd_prs_l2_parse_result_returned_with_err.attr,
-+ &dev_attr_pcd_prs_l3_parse_result_returned_with_err.attr,
-+ &dev_attr_pcd_prs_l4_parse_result_returned_with_err.attr,
-+ &dev_attr_pcd_prs_shim_parse_result_returned_with_err.attr,
-+ &dev_attr_pcd_prs_soft_prs_cycles.attr,
-+ &dev_attr_pcd_prs_soft_prs_stall_cycles.attr,
-+ &dev_attr_pcd_prs_hard_prs_cycle_incl_stall_cycles.attr,
-+ &dev_attr_pcd_prs_muram_read_cycles.attr,
-+ &dev_attr_pcd_prs_muram_read_stall_cycles.attr,
-+ &dev_attr_pcd_prs_muram_write_cycles.attr,
-+ &dev_attr_pcd_prs_muram_write_stall_cycles.attr,
-+ &dev_attr_pcd_prs_fpm_command_stall_cycles.attr,
-+ NULL
-+};
-+
-+static struct attribute *fm_dev_tnums_dbg_attributes[] = {
-+ &dev_attr_tnum_dbg_0.attr,
-+ &dev_attr_tnum_dbg_16.attr,
-+ &dev_attr_tnum_dbg_32.attr,
-+ &dev_attr_tnum_dbg_48.attr,
-+ &dev_attr_tnum_dbg_64.attr,
-+ &dev_attr_tnum_dbg_80.attr,
-+ &dev_attr_tnum_dbg_96.attr,
-+ &dev_attr_tnum_dbg_112.attr,
-+ NULL
-+};
-+
-+static struct attribute *fm_dev_cls_plans_attributes[] = {
-+ &dev_attr_cls_plan_0.attr,
-+ &dev_attr_cls_plan_1.attr,
-+ &dev_attr_cls_plan_2.attr,
-+ &dev_attr_cls_plan_3.attr,
-+ &dev_attr_cls_plan_4.attr,
-+ &dev_attr_cls_plan_5.attr,
-+ &dev_attr_cls_plan_6.attr,
-+ &dev_attr_cls_plan_7.attr,
-+ &dev_attr_cls_plan_8.attr,
-+ &dev_attr_cls_plan_9.attr,
-+ &dev_attr_cls_plan_10.attr,
-+ &dev_attr_cls_plan_11.attr,
-+ &dev_attr_cls_plan_12.attr,
-+ &dev_attr_cls_plan_13.attr,
-+ &dev_attr_cls_plan_14.attr,
-+ &dev_attr_cls_plan_15.attr,
-+ &dev_attr_cls_plan_16.attr,
-+ &dev_attr_cls_plan_17.attr,
-+ &dev_attr_cls_plan_18.attr,
-+ &dev_attr_cls_plan_19.attr,
-+ &dev_attr_cls_plan_20.attr,
-+ &dev_attr_cls_plan_21.attr,
-+ &dev_attr_cls_plan_22.attr,
-+ &dev_attr_cls_plan_23.attr,
-+ &dev_attr_cls_plan_24.attr,
-+ &dev_attr_cls_plan_25.attr,
-+ &dev_attr_cls_plan_26.attr,
-+ &dev_attr_cls_plan_27.attr,
-+ &dev_attr_cls_plan_28.attr,
-+ &dev_attr_cls_plan_29.attr,
-+ &dev_attr_cls_plan_30.attr,
-+ &dev_attr_cls_plan_31.attr,
-+ NULL
-+};
-+
-+static struct attribute *fm_dev_profiles_attributes[] = {
-+ &dev_attr_profile_0.attr,
-+ &dev_attr_profile_1.attr,
-+ &dev_attr_profile_2.attr,
-+ &dev_attr_profile_3.attr,
-+ &dev_attr_profile_4.attr,
-+ &dev_attr_profile_5.attr,
-+ &dev_attr_profile_6.attr,
-+ &dev_attr_profile_7.attr,
-+ &dev_attr_profile_8.attr,
-+ &dev_attr_profile_9.attr,
-+ &dev_attr_profile_10.attr,
-+ &dev_attr_profile_11.attr,
-+ &dev_attr_profile_12.attr,
-+ &dev_attr_profile_13.attr,
-+ &dev_attr_profile_14.attr,
-+ &dev_attr_profile_15.attr,
-+ &dev_attr_profile_16.attr,
-+ &dev_attr_profile_17.attr,
-+ &dev_attr_profile_18.attr,
-+ &dev_attr_profile_19.attr,
-+ &dev_attr_profile_20.attr,
-+ &dev_attr_profile_21.attr,
-+ &dev_attr_profile_22.attr,
-+ &dev_attr_profile_23.attr,
-+ &dev_attr_profile_24.attr,
-+ &dev_attr_profile_25.attr,
-+ &dev_attr_profile_26.attr,
-+ &dev_attr_profile_27.attr,
-+ &dev_attr_profile_28.attr,
-+ &dev_attr_profile_29.attr,
-+ &dev_attr_profile_30.attr,
-+ &dev_attr_profile_31.attr,
-+ NULL
-+};
-+
-+static struct attribute *fm_dev_schemes_attributes[] = {
-+ &dev_attr_scheme_0.attr,
-+ &dev_attr_scheme_1.attr,
-+ &dev_attr_scheme_2.attr,
-+ &dev_attr_scheme_3.attr,
-+ &dev_attr_scheme_4.attr,
-+ &dev_attr_scheme_5.attr,
-+ &dev_attr_scheme_6.attr,
-+ &dev_attr_scheme_7.attr,
-+ &dev_attr_scheme_8.attr,
-+ &dev_attr_scheme_9.attr,
-+ &dev_attr_scheme_10.attr,
-+ &dev_attr_scheme_11.attr,
-+ &dev_attr_scheme_12.attr,
-+ &dev_attr_scheme_13.attr,
-+ &dev_attr_scheme_14.attr,
-+ &dev_attr_scheme_15.attr,
-+ &dev_attr_scheme_16.attr,
-+ &dev_attr_scheme_17.attr,
-+ &dev_attr_scheme_18.attr,
-+ &dev_attr_scheme_19.attr,
-+ &dev_attr_scheme_20.attr,
-+ &dev_attr_scheme_21.attr,
-+ &dev_attr_scheme_22.attr,
-+ &dev_attr_scheme_23.attr,
-+ &dev_attr_scheme_24.attr,
-+ &dev_attr_scheme_25.attr,
-+ &dev_attr_scheme_26.attr,
-+ &dev_attr_scheme_27.attr,
-+ &dev_attr_scheme_28.attr,
-+ &dev_attr_scheme_29.attr,
-+ &dev_attr_scheme_30.attr,
-+ &dev_attr_scheme_31.attr,
-+ NULL
-+};
-+
-+static const struct attribute_group fm_dev_stats_attr_grp = {
-+ .name = "statistics",
-+ .attrs = fm_dev_stats_attributes
-+};
-+
-+static const struct attribute_group fm_dev_tnums_dbg_attr_grp = {
-+ .name = "tnums_dbg",
-+ .attrs = fm_dev_tnums_dbg_attributes
-+};
-+
-+static const struct attribute_group fm_dev_cls_plans_attr_grp = {
-+ .name = "cls_plans",
-+ .attrs = fm_dev_cls_plans_attributes
-+};
-+
-+static const struct attribute_group fm_dev_schemes_attr_grp = {
-+ .name = "schemes",
-+ .attrs = fm_dev_schemes_attributes
-+};
-+
-+static const struct attribute_group fm_dev_profiles_attr_grp = {
-+ .name = "profiles",
-+ .attrs = fm_dev_profiles_attributes
-+};
-+
-+static ssize_t show_fm_regs(struct device *dev,
-+ struct device_attribute *attr,
-+ char *buf)
-+{
-+ unsigned long flags;
-+ unsigned n = 0;
-+#if (defined(DEBUG_ERRORS) && (DEBUG_ERRORS > 0))
-+ t_LnxWrpFmDev *p_wrp_fm_dev = NULL;
-+#endif
-+ if (attr == NULL || buf == NULL || dev == NULL)
-+ return -EINVAL;
-+
-+#if (defined(DEBUG_ERRORS) && (DEBUG_ERRORS > 0))
-+
-+ p_wrp_fm_dev = (t_LnxWrpFmDev *) dev_get_drvdata(dev);
-+ if (WARN_ON(p_wrp_fm_dev == NULL))
-+ return -EINVAL;
-+
-+ local_irq_save(flags);
-+
-+ n = snprintf(buf, PAGE_SIZE, "FM driver registers dump.\n");
-+
-+ if (!p_wrp_fm_dev->active || !p_wrp_fm_dev->h_Dev)
-+ return -EIO;
-+ else
-+ n = fm_dump_regs(p_wrp_fm_dev->h_Dev, buf, n);
-+
-+ local_irq_restore(flags);
-+#else
-+
-+ local_irq_save(flags);
-+ n = snprintf(buf, PAGE_SIZE,
-+ "Debug level is too low to dump registers!!!\n");
-+ local_irq_restore(flags);
-+#endif /* (defined(DEBUG_ERRORS) && ... */
-+
-+ return n;
-+}
-+
-+static ssize_t show_fm_kg_pe_regs(struct device *dev,
-+ struct device_attribute *attr,
-+ char *buf)
-+{
-+ unsigned long flags;
-+ unsigned n = 0;
-+#if (defined(DEBUG_ERRORS) && (DEBUG_ERRORS > 0))
-+ t_LnxWrpFmDev *p_wrp_fm_dev = NULL;
-+#endif
-+
-+ if (attr == NULL || buf == NULL || dev == NULL)
-+ return -EINVAL;
-+
-+#if (defined(DEBUG_ERRORS) && (DEBUG_ERRORS > 0))
-+
-+ p_wrp_fm_dev = (t_LnxWrpFmDev *) dev_get_drvdata(dev);
-+ if (WARN_ON(p_wrp_fm_dev == NULL))
-+ return -EINVAL;
-+
-+ local_irq_save(flags);
-+
-+ n = snprintf(buf, PAGE_SIZE,
-+ "\n FM-KG Port Partition Config registers dump.\n");
-+
-+ if (!p_wrp_fm_dev->active || !p_wrp_fm_dev->h_PcdDev)
-+ return -EIO;
-+ else
-+ n = fm_kg_pe_dump_regs(p_wrp_fm_dev->h_PcdDev, buf, n);
-+
-+ local_irq_restore(flags);
-+#else
-+
-+ local_irq_save(flags);
-+ n = snprintf(buf, PAGE_SIZE,
-+ "Debug level is too low to dump registers!!!\n");
-+ local_irq_restore(flags);
-+#endif /* (defined(DEBUG_ERRORS) && ... */
-+
-+ return n;
-+}
-+
-+static ssize_t show_fm_kg_regs(struct device *dev,
-+ struct device_attribute *attr,
-+ char *buf)
-+{
-+ unsigned long flags;
-+ unsigned n = 0;
-+#if (defined(DEBUG_ERRORS) && (DEBUG_ERRORS > 0))
-+ t_LnxWrpFmDev *p_wrp_fm_dev = NULL;
-+#endif
-+
-+ if (attr == NULL || buf == NULL || dev == NULL)
-+ return -EINVAL;
-+
-+#if (defined(DEBUG_ERRORS) && (DEBUG_ERRORS > 0))
-+
-+ p_wrp_fm_dev = (t_LnxWrpFmDev *) dev_get_drvdata(dev);
-+ if (WARN_ON(p_wrp_fm_dev == NULL))
-+ return -EINVAL;
-+
-+ local_irq_save(flags);
-+
-+ n = snprintf(buf, PAGE_SIZE, "FM-KG registers dump.\n");
-+
-+ if (!p_wrp_fm_dev->active || !p_wrp_fm_dev->h_PcdDev)
-+ return -EIO;
-+ else
-+ n = fm_kg_dump_regs(p_wrp_fm_dev->h_PcdDev, buf, n);
-+
-+ local_irq_restore(flags);
-+#else
-+
-+ local_irq_save(flags);
-+ n = snprintf(buf, PAGE_SIZE,
-+ "Debug level is too low to dump registers!!!\n");
-+ local_irq_restore(flags);
-+#endif /* (defined(DEBUG_ERRORS) && ... */
-+
-+ return n;
-+}
-+
-+
-+static ssize_t show_fm_fpm_regs(struct device *dev,
-+ struct device_attribute *attr,
-+ char *buf)
-+{
-+ unsigned long flags;
-+ unsigned n = 0;
-+#if (defined(DEBUG_ERRORS) && (DEBUG_ERRORS > 0))
-+ t_LnxWrpFmDev *p_wrp_fm_dev = NULL;
-+#endif
-+
-+ if (attr == NULL || buf == NULL || dev == NULL)
-+ return -EINVAL;
-+
-+#if (defined(DEBUG_ERRORS) && (DEBUG_ERRORS > 0))
-+
-+ p_wrp_fm_dev = (t_LnxWrpFmDev *) dev_get_drvdata(dev);
-+ if (WARN_ON(p_wrp_fm_dev == NULL))
-+ return -EINVAL;
-+
-+ local_irq_save(flags);
-+
-+ n = snprintf(buf, PAGE_SIZE, "FM-FPM registers dump.\n");
-+
-+ if (!p_wrp_fm_dev->active || !p_wrp_fm_dev->h_Dev)
-+ return -EIO;
-+ else
-+ n = fm_fpm_dump_regs(p_wrp_fm_dev->h_Dev, buf, n);
-+
-+ local_irq_restore(flags);
-+#else
-+
-+ local_irq_save(flags);
-+ n = snprintf(buf, PAGE_SIZE,
-+ "Debug level is too low to dump registers!!!\n");
-+ local_irq_restore(flags);
-+#endif /* (defined(DEBUG_ERRORS) && ... */
-+
-+ return n;
-+}
-+
-+static ssize_t show_prs_regs(struct device *dev,
-+ struct device_attribute *attr, char *buf)
-+{
-+ unsigned long flags;
-+ unsigned n = 0;
-+#if (defined(DEBUG_ERRORS) && (DEBUG_ERRORS > 0))
-+ t_LnxWrpFmDev *p_wrp_fm_dev = NULL;
-+#endif
-+
-+ if (attr == NULL || buf == NULL || dev == NULL)
-+ return -EINVAL;
-+
-+#if (defined(DEBUG_ERRORS) && (DEBUG_ERRORS > 0))
-+ p_wrp_fm_dev = (t_LnxWrpFmDev *) dev_get_drvdata(dev);
-+ if (WARN_ON(p_wrp_fm_dev == NULL))
-+ return -EINVAL;
-+
-+ local_irq_save(flags);
-+ n = snprintf(buf, PAGE_SIZE, "FM Policer registers dump.\n");
-+
-+ if (!p_wrp_fm_dev->active || !p_wrp_fm_dev->h_PcdDev)
-+ return -EIO;
-+ else
-+ n = fm_prs_dump_regs(p_wrp_fm_dev->h_PcdDev, buf, n);
-+
-+ local_irq_restore(flags);
-+#else
-+
-+ local_irq_save(flags);
-+ n = snprintf(buf, PAGE_SIZE,
-+ "Debug level is too low to dump registers!!!\n");
-+ local_irq_restore(flags);
-+
-+#endif /* (defined(DEBUG_ERRORS) && ... */
-+
-+ return n;
-+}
-+
-+static ssize_t show_plcr_regs(struct device *dev,
-+ struct device_attribute *attr,
-+ char *buf)
-+{
-+ unsigned long flags;
-+ unsigned n = 0;
-+#if (defined(DEBUG_ERRORS) && (DEBUG_ERRORS > 0))
-+ t_LnxWrpFmDev *p_wrp_fm_dev = NULL;
-+#endif
-+
-+ if (attr == NULL || buf == NULL || dev == NULL)
-+ return -EINVAL;
-+
-+#if (defined(DEBUG_ERRORS) && (DEBUG_ERRORS > 0))
-+ p_wrp_fm_dev = (t_LnxWrpFmDev *) dev_get_drvdata(dev);
-+ if (WARN_ON(p_wrp_fm_dev == NULL))
-+ return -EINVAL;
-+
-+ local_irq_save(flags);
-+ n = snprintf(buf, PAGE_SIZE, "FM Policer registers dump.\n");
-+
-+ if (!p_wrp_fm_dev->active || !p_wrp_fm_dev->h_PcdDev)
-+ return -EIO;
-+ else
-+ n = fm_plcr_dump_regs(p_wrp_fm_dev->h_PcdDev, buf, n);
-+
-+ local_irq_restore(flags);
-+#else
-+
-+ local_irq_save(flags);
-+ n = snprintf(buf, PAGE_SIZE,
-+ "Debug level is too low to dump registers!!!\n");
-+ local_irq_restore(flags);
-+
-+#endif /* (defined(DEBUG_ERRORS) && ... */
-+
-+ return n;
-+}
-+
-+static DEVICE_ATTR(fm_regs, S_IRUGO, show_fm_regs, NULL);
-+static DEVICE_ATTR(fm_fpm_regs, S_IRUGO, show_fm_fpm_regs, NULL);
-+static DEVICE_ATTR(fm_kg_regs, S_IRUGO, show_fm_kg_regs, NULL);
-+static DEVICE_ATTR(fm_kg_pe_regs, S_IRUGO, show_fm_kg_pe_regs, NULL);
-+static DEVICE_ATTR(fm_plcr_regs, S_IRUGO, show_plcr_regs, NULL);
-+static DEVICE_ATTR(fm_prs_regs, S_IRUGO, show_prs_regs, NULL);
-+static DEVICE_ATTR(fm_muram_free_size, S_IRUGO, show_fm_muram_free_sz, NULL);
-+static DEVICE_ATTR(fm_ctrl_code_ver, S_IRUGO, show_fm_ctrl_code_ver, NULL);
-+
-+int fm_sysfs_create(struct device *dev)
-+{
-+ t_LnxWrpFmDev *p_wrp_fm_dev = NULL;
-+
-+ if (dev == NULL)
-+ return -EIO;
-+
-+ p_wrp_fm_dev = (t_LnxWrpFmDev *) dev_get_drvdata(dev);
-+
-+ /* store to remove them when module is disabled */
-+ p_wrp_fm_dev->dev_attr_regs = &dev_attr_fm_regs;
-+ p_wrp_fm_dev->dev_attr_risc_load = &dev_attr_fm_risc_load_val;
-+ p_wrp_fm_dev->dev_fm_fpm_attr_regs = &dev_attr_fm_fpm_regs;
-+ p_wrp_fm_dev->dev_fm_kg_attr_regs = &dev_attr_fm_kg_regs;
-+ p_wrp_fm_dev->dev_fm_kg_pe_attr_regs = &dev_attr_fm_kg_pe_regs;
-+ p_wrp_fm_dev->dev_plcr_attr_regs = &dev_attr_fm_plcr_regs;
-+ p_wrp_fm_dev->dev_prs_attr_regs = &dev_attr_fm_prs_regs;
-+ p_wrp_fm_dev->dev_attr_muram_free_size = &dev_attr_fm_muram_free_size;
-+ p_wrp_fm_dev->dev_attr_fm_ctrl_code_ver = &dev_attr_fm_ctrl_code_ver;
-+
-+ /* Create sysfs statistics group for FM module */
-+ if (sysfs_create_group(&dev->kobj, &fm_dev_stats_attr_grp) != 0)
-+ return -EIO;
-+
-+ if (sysfs_create_group(&dev->kobj, &fm_dev_schemes_attr_grp) != 0)
-+ return -EIO;
-+
-+ if (sysfs_create_group(&dev->kobj, &fm_dev_profiles_attr_grp) != 0)
-+ return -EIO;
-+
-+ if (sysfs_create_group(&dev->kobj, &fm_dev_tnums_dbg_attr_grp) != 0)
-+ return -EIO;
-+
-+ if (sysfs_create_group(&dev->kobj, &fm_dev_cls_plans_attr_grp) != 0)
-+ return -EIO;
-+
-+ /* Registers dump entry - in future will be moved to debugfs */
-+ if (device_create_file(dev, &dev_attr_fm_regs) != 0)
-+ return -EIO;
-+
-+ if (device_create_file(dev, &dev_attr_fm_risc_load_val) != 0)
-+ return -EIO;
-+
-+ if (device_create_file(dev, &dev_attr_fm_fpm_regs) != 0)
-+ return -EIO;
-+
-+ if (device_create_file(dev, &dev_attr_fm_kg_regs) != 0)
-+ return -EIO;
-+
-+ if (device_create_file(dev, &dev_attr_fm_kg_pe_regs) != 0)
-+ return -EIO;
-+
-+ if (device_create_file(dev, &dev_attr_fm_plcr_regs) != 0)
-+ return -EIO;
-+
-+ if (device_create_file(dev, &dev_attr_fm_prs_regs) != 0)
-+ return -EIO;
-+
-+ /* muram free size */
-+ if (device_create_file(dev, &dev_attr_fm_muram_free_size) != 0)
-+ return -EIO;
-+
-+ /* fm ctrl code version */
-+ if (device_create_file(dev, &dev_attr_fm_ctrl_code_ver) != 0)
-+ return -EIO;
-+
-+ return 0;
-+}
-+
-+void fm_sysfs_destroy(struct device *dev)
-+{
-+ t_LnxWrpFmDev *p_wrp_fm_dev = NULL;
-+
-+ if (WARN_ON(dev == NULL))
-+ return;
-+
-+ p_wrp_fm_dev = (t_LnxWrpFmDev *) dev_get_drvdata(dev);
-+ if (WARN_ON(p_wrp_fm_dev == NULL))
-+ return;
-+
-+ sysfs_remove_group(&dev->kobj, &fm_dev_stats_attr_grp);
-+ sysfs_remove_group(&dev->kobj, &fm_dev_schemes_attr_grp);
-+ sysfs_remove_group(&dev->kobj, &fm_dev_profiles_attr_grp);
-+ sysfs_remove_group(&dev->kobj, &fm_dev_cls_plans_attr_grp);
-+ sysfs_remove_group(&dev->kobj, &fm_dev_tnums_dbg_attr_grp);
-+ device_remove_file(dev, p_wrp_fm_dev->dev_attr_regs);
-+ device_remove_file(dev, p_wrp_fm_dev->dev_fm_fpm_attr_regs);
-+ device_remove_file(dev, p_wrp_fm_dev->dev_fm_kg_attr_regs);
-+ device_remove_file(dev, p_wrp_fm_dev->dev_fm_kg_pe_attr_regs);
-+ device_remove_file(dev, p_wrp_fm_dev->dev_plcr_attr_regs);
-+ device_remove_file(dev, p_wrp_fm_dev->dev_prs_attr_regs);
-+ device_remove_file(dev, p_wrp_fm_dev->dev_attr_muram_free_size);
-+ device_remove_file(dev, p_wrp_fm_dev->dev_attr_fm_ctrl_code_ver);
-+}
-+
-+int fm_dump_regs(void *h_fm, char *buf, int nn)
-+{
-+ t_Fm *p_Fm = (t_Fm *)h_fm;
-+ uint8_t i = 0;
-+ int n = nn;
-+
-+ FM_DMP_SUBTITLE(buf, n, "\n");
-+
-+ FM_DMP_TITLE(buf, n, p_Fm->p_FmDmaRegs, "FM-DMA Regs");
-+
-+ FM_DMP_V32(buf, n, p_Fm->p_FmDmaRegs, fmdmsr);
-+ FM_DMP_V32(buf, n, p_Fm->p_FmDmaRegs, fmdmemsr);
-+ FM_DMP_V32(buf, n, p_Fm->p_FmDmaRegs, fmdmmr);
-+ FM_DMP_V32(buf, n, p_Fm->p_FmDmaRegs, fmdmtr);
-+ FM_DMP_V32(buf, n, p_Fm->p_FmDmaRegs, fmdmhy);
-+ FM_DMP_V32(buf, n, p_Fm->p_FmDmaRegs, fmdmsetr);
-+ FM_DMP_V32(buf, n, p_Fm->p_FmDmaRegs, fmdmtah);
-+ FM_DMP_V32(buf, n, p_Fm->p_FmDmaRegs, fmdmtal);
-+ FM_DMP_V32(buf, n, p_Fm->p_FmDmaRegs, fmdmtcid);
-+ FM_DMP_V32(buf, n, p_Fm->p_FmDmaRegs, fmdmra);
-+ FM_DMP_V32(buf, n, p_Fm->p_FmDmaRegs, fmdmrd);
-+ FM_DMP_V32(buf, n, p_Fm->p_FmDmaRegs, fmdmwcr);
-+ FM_DMP_V32(buf, n, p_Fm->p_FmDmaRegs, fmdmebcr);
-+ FM_DMP_V32(buf, n, p_Fm->p_FmDmaRegs, fmdmdcr);
-+
-+ FM_DMP_TITLE(buf, n, &p_Fm->p_FmDmaRegs->fmdmplr, "fmdmplr");
-+
-+ for (i = 0; i < FM_MAX_NUM_OF_HW_PORT_IDS / 2 ; ++i)
-+ FM_DMP_MEM_32(buf, n, &p_Fm->p_FmDmaRegs->fmdmplr[i]);
-+
-+ FM_DMP_TITLE(buf, n, p_Fm->p_FmBmiRegs, "FM-BMI COMMON Regs");
-+ FM_DMP_V32(buf, n, p_Fm->p_FmBmiRegs, fmbm_init);
-+ FM_DMP_V32(buf, n, p_Fm->p_FmBmiRegs, fmbm_cfg1);
-+ FM_DMP_V32(buf, n, p_Fm->p_FmBmiRegs, fmbm_cfg2);
-+ FM_DMP_V32(buf, n, p_Fm->p_FmBmiRegs, fmbm_ievr);
-+ FM_DMP_V32(buf, n, p_Fm->p_FmBmiRegs, fmbm_ier);
-+
-+ FM_DMP_TITLE(buf, n, &p_Fm->p_FmBmiRegs->fmbm_arb, "fmbm_arb");
-+ for (i = 0; i < 8 ; ++i)
-+ FM_DMP_MEM_32(buf, n, &p_Fm->p_FmBmiRegs->fmbm_arb[i]);
-+
-+ FM_DMP_TITLE(buf, n, p_Fm->p_FmQmiRegs, "FM-QMI COMMON Regs");
-+ FM_DMP_V32(buf, n, p_Fm->p_FmQmiRegs, fmqm_gc);
-+ FM_DMP_V32(buf, n, p_Fm->p_FmQmiRegs, fmqm_eie);
-+ FM_DMP_V32(buf, n, p_Fm->p_FmQmiRegs, fmqm_eien);
-+ FM_DMP_V32(buf, n, p_Fm->p_FmQmiRegs, fmqm_eif);
-+ FM_DMP_V32(buf, n, p_Fm->p_FmQmiRegs, fmqm_ie);
-+ FM_DMP_V32(buf, n, p_Fm->p_FmQmiRegs, fmqm_ien);
-+ FM_DMP_V32(buf, n, p_Fm->p_FmQmiRegs, fmqm_if);
-+ FM_DMP_V32(buf, n, p_Fm->p_FmQmiRegs, fmqm_gs);
-+ FM_DMP_V32(buf, n, p_Fm->p_FmQmiRegs, fmqm_etfc);
-+
-+ return n;
-+}
-+
-+int fm_dump_tnum_dbg(void *h_fm, int tn_s, int tn_e, char *buf, int nn)
-+{
-+ t_Fm *p_Fm = (t_Fm *)h_fm;
-+ uint8_t i, j = 0;
-+ int n = nn;
-+
-+ FM_DMP_TITLE(buf, n, NULL, "Tnums and Tnum dbg regs %d - %d",
-+ tn_s, tn_e);
-+
-+ iowrite32be(tn_s << 24, &p_Fm->p_FmFpmRegs->fmfp_dra);
-+
-+ mb();
-+
-+ for (j = tn_s; j <= tn_e; j++) {
-+ FM_DMP_LN(buf, n, "> fmfp_ts[%d]\n", j);
-+ FM_DMP_MEM_32(buf, n, &p_Fm->p_FmFpmRegs->fmfp_ts[j]);
-+ FM_DMP_V32(buf, n, p_Fm->p_FmFpmRegs, fmfp_dra);
-+ FM_DMP_LN(buf, n, "> fmfp_drd[0-3]\n");
-+
-+ for (i = 0; i < 4 ; ++i)
-+ FM_DMP_MEM_32(buf, n, &p_Fm->p_FmFpmRegs->fmfp_drd[i]);
-+
-+ FM_DMP_LN(buf, n, "\n");
-+
-+ }
-+
-+ return n;
-+}
-+
-+int fm_dump_cls_plan(void *h_fm_pcd, int cpn, char *buf, int nn)
-+{
-+ t_FmPcd *p_pcd = (t_FmPcd *)h_fm_pcd;
-+ int i = 0;
-+ uint32_t tmp;
-+ unsigned long i_flg;
-+ int n = nn;
-+ u_FmPcdKgIndirectAccessRegs *idac;
-+ spinlock_t *p_lk;
-+
-+ p_lk = (spinlock_t *)p_pcd->p_FmPcdKg->h_HwSpinlock;
-+ idac = p_pcd->p_FmPcdKg->p_IndirectAccessRegs;
-+
-+ spin_lock_irqsave(p_lk, i_flg);
-+
-+ /* Read ClsPlan Block Action Regs */
-+ tmp = (uint32_t)(FM_KG_KGAR_GO |
-+ FM_KG_KGAR_READ |
-+ FM_PCD_KG_KGAR_SEL_CLS_PLAN_ENTRY |
-+ DUMMY_PORT_ID |
-+ ((uint32_t)cpn << FM_PCD_KG_KGAR_NUM_SHIFT) |
-+ FM_PCD_KG_KGAR_WSEL_MASK);
-+
-+ if (fman_kg_write_ar_wait(p_pcd->p_FmPcdKg->p_FmPcdKgRegs, tmp)) {
-+ FM_DMP_LN(buf, nn, "Keygen scheme access violation");
-+ spin_unlock_irqrestore(p_lk, i_flg);
-+ return nn;
-+ }
-+ FM_DMP_TITLE(buf, n, &idac->clsPlanRegs,
-+ "ClsPlan %d Indirect Access Regs", cpn);
-+
-+ for (i = 0; i < 8; i++)
-+ FM_DMP_MEM_32(buf, n, &idac->clsPlanRegs.kgcpe[i]);
-+
-+ spin_unlock_irqrestore(p_lk, i_flg);
-+
-+ return n;
-+}
-+
-+int fm_profile_dump_regs(void *h_fm_pcd, int ppn, char *buf, int nn)
-+{
-+ t_FmPcd *p_pcd = (t_FmPcd *)h_fm_pcd;
-+ t_FmPcdPlcrProfileRegs *p_prof_regs;
-+ t_FmPcdPlcrRegs *p_plcr_regs;
-+ t_FmPcdPlcr *p_plcr;
-+ uint32_t tmp;
-+ unsigned long i_flg;
-+ int n = nn;
-+ int toc = 10;
-+ spinlock_t *p_lk;
-+
-+ p_plcr = p_pcd->p_FmPcdPlcr;
-+ p_prof_regs = &p_pcd->p_FmPcdPlcr->p_FmPcdPlcrRegs->profileRegs;
-+ p_plcr_regs = p_pcd->p_FmPcdPlcr->p_FmPcdPlcrRegs;
-+
-+ p_lk = (spinlock_t *)((t_FmPcdPlcr *)p_plcr)->h_HwSpinlock;
-+
-+ FM_DMP_SUBTITLE(buf, n, "\n");
-+ FM_DMP_TITLE(buf, n, p_plcr_regs, "FM-PCD policer-profile regs");
-+
-+ tmp = (uint32_t)(FM_PCD_PLCR_PAR_GO |
-+ FM_PCD_PLCR_PAR_R |
-+ ((uint32_t)ppn << FM_PCD_PLCR_PAR_PNUM_SHIFT) |
-+ FM_PCD_PLCR_PAR_PWSEL_MASK);
-+
-+ spin_lock_irqsave(p_lk, i_flg);
-+
-+ iowrite32be(tmp, &p_plcr_regs->fmpl_par);
-+
-+ mb();
-+
-+ /* wait for the porfile regs to be present */
-+ do {
-+ --toc;
-+ udelay(10);
-+ if (!toc) {
-+ /* looks like PLCR_PAR_GO refuses to clear */
-+ spin_unlock_irqrestore(p_lk, i_flg);
-+ FM_DMP_LN(buf, n, "Profile regs not accessible -");
-+ FM_DMP_LN(buf, n, " check profile init process\n");
-+ return n;
-+ }
-+ } while ((ioread32be(&p_plcr_regs->fmpl_par) & FM_PCD_PLCR_PAR_GO));
-+
-+ FM_DMP_TITLE(buf, n, p_prof_regs, "Profile %d regs", ppn);
-+
-+ FM_DMP_V32(buf, n, p_prof_regs, fmpl_pemode);
-+ FM_DMP_V32(buf, n, p_prof_regs, fmpl_pegnia);
-+ FM_DMP_V32(buf, n, p_prof_regs, fmpl_peynia);
-+ FM_DMP_V32(buf, n, p_prof_regs, fmpl_pernia);
-+ FM_DMP_V32(buf, n, p_prof_regs, fmpl_pecir);
-+ FM_DMP_V32(buf, n, p_prof_regs, fmpl_pecbs);
-+ FM_DMP_V32(buf, n, p_prof_regs, fmpl_pepepir_eir);
-+ FM_DMP_V32(buf, n, p_prof_regs, fmpl_pepbs_ebs);
-+ FM_DMP_V32(buf, n, p_prof_regs, fmpl_pelts);
-+ FM_DMP_V32(buf, n, p_prof_regs, fmpl_pects);
-+ FM_DMP_V32(buf, n, p_prof_regs, fmpl_pepts_ets);
-+ FM_DMP_V32(buf, n, p_prof_regs, fmpl_pegpc);
-+ FM_DMP_V32(buf, n, p_prof_regs, fmpl_peypc);
-+ FM_DMP_V32(buf, n, p_prof_regs, fmpl_perpc);
-+ FM_DMP_V32(buf, n, p_prof_regs, fmpl_perypc);
-+ FM_DMP_V32(buf, n, p_prof_regs, fmpl_perrpc);
-+
-+ spin_unlock_irqrestore(p_lk, i_flg);
-+
-+ return n;
-+}
-+
-+int fm_dump_scheme(void *h_fm_pcd, int scnum, char *buf, int nn)
-+{
-+ t_FmPcd *p_pcd = (t_FmPcd *)h_fm_pcd;
-+ uint32_t tmp_ar;
-+ unsigned long i_flg;
-+ int i, n = nn;
-+ spinlock_t *p_lk;
-+ u_FmPcdKgIndirectAccessRegs *idac;
-+
-+ idac = p_pcd->p_FmPcdKg->p_IndirectAccessRegs;
-+ p_lk = (spinlock_t *)p_pcd->p_FmPcdKg->h_HwSpinlock;
-+
-+ spin_lock_irqsave(p_lk, i_flg);
-+
-+ tmp_ar = FmPcdKgBuildReadSchemeActionReg((uint8_t)scnum);
-+ if (fman_kg_write_ar_wait(p_pcd->p_FmPcdKg->p_FmPcdKgRegs, tmp_ar)) {
-+ FM_DMP_LN(buf, nn,
-+ "Keygen scheme access violation or no such scheme");
-+ spin_unlock_irqrestore(p_lk, i_flg);
-+ return nn;
-+ }
-+
-+ FM_DMP_TITLE(buf, n, &idac->schemeRegs,
-+ "Scheme %d Indirect Access Regs", scnum);
-+
-+ FM_DMP_V32(buf, n, &idac->schemeRegs, kgse_mode);
-+ FM_DMP_V32(buf, n, &idac->schemeRegs, kgse_ekfc);
-+ FM_DMP_V32(buf, n, &idac->schemeRegs, kgse_ekdv);
-+ FM_DMP_V32(buf, n, &idac->schemeRegs, kgse_bmch);
-+ FM_DMP_V32(buf, n, &idac->schemeRegs, kgse_bmcl);
-+ FM_DMP_V32(buf, n, &idac->schemeRegs, kgse_fqb);
-+ FM_DMP_V32(buf, n, &idac->schemeRegs, kgse_hc);
-+ FM_DMP_V32(buf, n, &idac->schemeRegs, kgse_ppc);
-+
-+ FM_DMP_TITLE(buf, n, &idac->schemeRegs.kgse_gec, "kgse_gec");
-+
-+ for (i = 0; i < FM_KG_NUM_OF_GENERIC_REGS; i++)
-+ FM_DMP_MEM_32(buf, n, &idac->schemeRegs.kgse_gec[i]);
-+
-+ FM_DMP_V32(buf, n, &idac->schemeRegs, kgse_spc);
-+ FM_DMP_V32(buf, n, &idac->schemeRegs, kgse_dv0);
-+ FM_DMP_V32(buf, n, &idac->schemeRegs, kgse_dv1);
-+ FM_DMP_V32(buf, n, &idac->schemeRegs, kgse_ccbs);
-+ FM_DMP_V32(buf, n, &idac->schemeRegs, kgse_mv);
-+
-+ FM_DMP_SUBTITLE(buf, n, "\n");
-+
-+ spin_unlock_irqrestore(p_lk, i_flg);
-+
-+ return n;
-+}
-+
-+int fm_kg_pe_dump_regs(void *h_fm_pcd, char *buf, int nn)
-+{
-+ t_FmPcd *p_pcd = (t_FmPcd *)h_fm_pcd;
-+ int i = 0;
-+ uint8_t prt_id = 0;
-+ uint32_t tmp_ar;
-+ unsigned long i_flg;
-+ int n = nn;
-+ u_FmPcdKgIndirectAccessRegs *idac;
-+ t_FmPcdKg *p_kg;
-+ spinlock_t *p_lk;
-+
-+ p_kg = p_pcd->p_FmPcdKg;
-+ idac = p_pcd->p_FmPcdKg->p_IndirectAccessRegs;
-+ p_lk = (spinlock_t *)p_kg->h_HwSpinlock;
-+
-+ spin_lock_irqsave(p_lk, i_flg);
-+
-+ FM_DMP_SUBTITLE(buf, n, "\n");
-+
-+ for (i = 0; i < FM_MAX_NUM_OF_PORTS; i++) {
-+ SW_PORT_INDX_TO_HW_PORT_ID(prt_id, i);
-+
-+ tmp_ar = FmPcdKgBuildReadPortSchemeBindActionReg(prt_id);
-+
-+ if (fman_kg_write_ar_wait(p_kg->p_FmPcdKgRegs, tmp_ar)) {
-+ FM_DMP_LN(buf, nn, "Keygen scheme access violation");
-+ spin_unlock_irqrestore(p_lk, i_flg);
-+ return nn;
-+ }
-+ FM_DMP_TITLE(buf, n, &idac->portRegs, "Port %d regs", prt_id);
-+ FM_DMP_V32(buf, n, &idac->portRegs, fmkg_pe_sp);
-+ FM_DMP_V32(buf, n, &idac->portRegs, fmkg_pe_cpp);
-+ }
-+
-+ FM_DMP_SUBTITLE(buf, n, "\n");
-+
-+ spin_unlock_irqrestore(p_lk, i_flg);
-+
-+ return n;
-+}
-+
-+int fm_kg_dump_regs(void *h_fm_pcd, char *buf, int nn)
-+{
-+ t_FmPcd *p_pcd = (t_FmPcd *)h_fm_pcd;
-+ int n = nn;
-+
-+ FM_DMP_SUBTITLE(buf, n, "\n");
-+ FM_DMP_TITLE(buf, n, p_pcd->p_FmPcdKg->p_FmPcdKgRegs,
-+ "FmPcdKgRegs Regs");
-+
-+ FM_DMP_V32(buf, n, p_pcd->p_FmPcdKg->p_FmPcdKgRegs, fmkg_gcr);
-+ FM_DMP_V32(buf, n, p_pcd->p_FmPcdKg->p_FmPcdKgRegs, fmkg_eer);
-+ FM_DMP_V32(buf, n, p_pcd->p_FmPcdKg->p_FmPcdKgRegs, fmkg_eeer);
-+ FM_DMP_V32(buf, n, p_pcd->p_FmPcdKg->p_FmPcdKgRegs, fmkg_seer);
-+ FM_DMP_V32(buf, n, p_pcd->p_FmPcdKg->p_FmPcdKgRegs, fmkg_seeer);
-+ FM_DMP_V32(buf, n, p_pcd->p_FmPcdKg->p_FmPcdKgRegs, fmkg_gsr);
-+ FM_DMP_V32(buf, n, p_pcd->p_FmPcdKg->p_FmPcdKgRegs, fmkg_tpc);
-+ FM_DMP_V32(buf, n, p_pcd->p_FmPcdKg->p_FmPcdKgRegs, fmkg_serc);
-+ FM_DMP_V32(buf, n, p_pcd->p_FmPcdKg->p_FmPcdKgRegs, fmkg_fdor);
-+ FM_DMP_V32(buf, n, p_pcd->p_FmPcdKg->p_FmPcdKgRegs, fmkg_gdv0r);
-+ FM_DMP_V32(buf, n, p_pcd->p_FmPcdKg->p_FmPcdKgRegs, fmkg_gdv1r);
-+ FM_DMP_V32(buf, n, p_pcd->p_FmPcdKg->p_FmPcdKgRegs, fmkg_feer);
-+ FM_DMP_V32(buf, n, p_pcd->p_FmPcdKg->p_FmPcdKgRegs, fmkg_ar);
-+
-+ FM_DMP_SUBTITLE(buf, n, "\n");
-+
-+ return n;
-+}
-+
-+
-+int fm_fpm_dump_regs(void *h_fm, char *buf, int nn)
-+{
-+ t_Fm *p_fm = (t_Fm *)h_fm;
-+ uint8_t i;
-+ int n = nn;
-+
-+ FM_DMP_SUBTITLE(buf, n, "\n");
-+
-+ FM_DMP_TITLE(buf, n, p_fm->p_FmFpmRegs, "FM-FPM Regs");
-+
-+ FM_DMP_V32(buf, n, p_fm->p_FmFpmRegs, fmfp_tnc);
-+ FM_DMP_V32(buf, n, p_fm->p_FmFpmRegs, fmfp_prc);
-+ FM_DMP_V32(buf, n, p_fm->p_FmFpmRegs, fmfp_brkc);
-+ FM_DMP_V32(buf, n, p_fm->p_FmFpmRegs, fmfp_mxd);
-+ FM_DMP_V32(buf, n, p_fm->p_FmFpmRegs, fmfp_dist1);
-+ FM_DMP_V32(buf, n, p_fm->p_FmFpmRegs, fmfp_dist2);
-+ FM_DMP_V32(buf, n, p_fm->p_FmFpmRegs, fm_epi);
-+ FM_DMP_V32(buf, n, p_fm->p_FmFpmRegs, fm_rie);
-+
-+ FM_DMP_TITLE(buf, n, &p_fm->p_FmFpmRegs->fmfp_fcev, "fmfp_fcev");
-+ for (i = 0; i < 4; ++i)
-+ FM_DMP_MEM_32(buf, n, &p_fm->p_FmFpmRegs->fmfp_fcev[i]);
-+
-+ FM_DMP_TITLE(buf, n, &p_fm->p_FmFpmRegs->fmfp_cee, "fmfp_cee");
-+ for (i = 0; i < 4; ++i)
-+ FM_DMP_MEM_32(buf, n, &p_fm->p_FmFpmRegs->fmfp_cee[i]);
-+
-+ FM_DMP_SUBTITLE(buf, n, "\n");
-+ FM_DMP_V32(buf, n, p_fm->p_FmFpmRegs, fmfp_tsc1);
-+ FM_DMP_V32(buf, n, p_fm->p_FmFpmRegs, fmfp_tsc2);
-+ FM_DMP_V32(buf, n, p_fm->p_FmFpmRegs, fmfp_tsp);
-+ FM_DMP_V32(buf, n, p_fm->p_FmFpmRegs, fmfp_tsf);
-+ FM_DMP_V32(buf, n, p_fm->p_FmFpmRegs, fm_rcr);
-+ FM_DMP_V32(buf, n, p_fm->p_FmFpmRegs, fmfp_extc);
-+ FM_DMP_V32(buf, n, p_fm->p_FmFpmRegs, fmfp_ext1);
-+ FM_DMP_V32(buf, n, p_fm->p_FmFpmRegs, fmfp_ext2);
-+
-+ FM_DMP_SUBTITLE(buf, n, "\n");
-+ FM_DMP_V32(buf, n, p_fm->p_FmFpmRegs, fm_ip_rev_1);
-+ FM_DMP_V32(buf, n, p_fm->p_FmFpmRegs, fm_ip_rev_2);
-+ FM_DMP_V32(buf, n, p_fm->p_FmFpmRegs, fm_rstc);
-+ FM_DMP_V32(buf, n, p_fm->p_FmFpmRegs, fm_cld);
-+ FM_DMP_V32(buf, n, p_fm->p_FmFpmRegs, fm_npi);
-+ FM_DMP_V32(buf, n, p_fm->p_FmFpmRegs, fmfp_ee);
-+
-+ FM_DMP_TITLE(buf, n, &p_fm->p_FmFpmRegs->fmfp_cev, "fmfp_cev");
-+ for (i = 0; i < 4; ++i)
-+ FM_DMP_MEM_32(buf, n, &p_fm->p_FmFpmRegs->fmfp_cev[i]);
-+
-+ FM_DMP_TITLE(buf, n, &p_fm->p_FmFpmRegs->fmfp_ps, "fmfp_ps");
-+ for (i = 0; i < 64; ++i)
-+ FM_DMP_MEM_32(buf, n, &p_fm->p_FmFpmRegs->fmfp_ps[i]);
-+
-+ return n;
-+}
-+
-+int fm_prs_dump_regs(void *h_fm_pcd, char *buf, int nn)
-+{
-+ t_FmPcd *p_pcd = (t_FmPcd *)h_fm_pcd;
-+ int n = nn;
-+
-+ FM_DMP_SUBTITLE(buf, n, "\n");
-+
-+ FM_DMP_TITLE(buf, n, p_pcd->p_FmPcdPrs->p_FmPcdPrsRegs,
-+ "FM-PCD parser regs");
-+
-+ FM_DMP_V32(buf, n, p_pcd->p_FmPcdPrs->p_FmPcdPrsRegs, fmpr_rpclim);
-+ FM_DMP_V32(buf, n, p_pcd->p_FmPcdPrs->p_FmPcdPrsRegs, fmpr_rpimac);
-+ FM_DMP_V32(buf, n, p_pcd->p_FmPcdPrs->p_FmPcdPrsRegs, pmeec);
-+ FM_DMP_V32(buf, n, p_pcd->p_FmPcdPrs->p_FmPcdPrsRegs, fmpr_pevr);
-+ FM_DMP_V32(buf, n, p_pcd->p_FmPcdPrs->p_FmPcdPrsRegs, fmpr_pever);
-+ FM_DMP_V32(buf, n, p_pcd->p_FmPcdPrs->p_FmPcdPrsRegs, fmpr_perr);
-+ FM_DMP_V32(buf, n, p_pcd->p_FmPcdPrs->p_FmPcdPrsRegs, fmpr_perer);
-+ FM_DMP_V32(buf, n, p_pcd->p_FmPcdPrs->p_FmPcdPrsRegs, fmpr_ppsc);
-+ FM_DMP_V32(buf, n, p_pcd->p_FmPcdPrs->p_FmPcdPrsRegs, fmpr_pds);
-+ FM_DMP_V32(buf, n, p_pcd->p_FmPcdPrs->p_FmPcdPrsRegs, fmpr_l2rrs);
-+ FM_DMP_V32(buf, n, p_pcd->p_FmPcdPrs->p_FmPcdPrsRegs, fmpr_l3rrs);
-+ FM_DMP_V32(buf, n, p_pcd->p_FmPcdPrs->p_FmPcdPrsRegs, fmpr_l4rrs);
-+ FM_DMP_V32(buf, n, p_pcd->p_FmPcdPrs->p_FmPcdPrsRegs, fmpr_srrs);
-+ FM_DMP_V32(buf, n, p_pcd->p_FmPcdPrs->p_FmPcdPrsRegs, fmpr_l2rres);
-+ FM_DMP_V32(buf, n, p_pcd->p_FmPcdPrs->p_FmPcdPrsRegs, fmpr_l3rres);
-+ FM_DMP_V32(buf, n, p_pcd->p_FmPcdPrs->p_FmPcdPrsRegs, fmpr_l4rres);
-+ FM_DMP_V32(buf, n, p_pcd->p_FmPcdPrs->p_FmPcdPrsRegs, fmpr_srres);
-+ FM_DMP_V32(buf, n, p_pcd->p_FmPcdPrs->p_FmPcdPrsRegs, fmpr_spcs);
-+ FM_DMP_V32(buf, n, p_pcd->p_FmPcdPrs->p_FmPcdPrsRegs, fmpr_spscs);
-+ FM_DMP_V32(buf, n, p_pcd->p_FmPcdPrs->p_FmPcdPrsRegs, fmpr_hxscs);
-+ FM_DMP_V32(buf, n, p_pcd->p_FmPcdPrs->p_FmPcdPrsRegs, fmpr_mrcs);
-+ FM_DMP_V32(buf, n, p_pcd->p_FmPcdPrs->p_FmPcdPrsRegs, fmpr_mwcs);
-+ FM_DMP_V32(buf, n, p_pcd->p_FmPcdPrs->p_FmPcdPrsRegs, fmpr_mrscs);
-+ FM_DMP_V32(buf, n, p_pcd->p_FmPcdPrs->p_FmPcdPrsRegs, fmpr_mwscs);
-+ FM_DMP_V32(buf, n, p_pcd->p_FmPcdPrs->p_FmPcdPrsRegs, fmpr_fcscs);
-+
-+ return n;
-+}
-+
-+int fm_plcr_dump_regs(void *h_fm_pcd, char *buf, int nn)
-+{
-+ t_FmPcd *p_pcd = (t_FmPcd *)h_fm_pcd;
-+ int i = 0;
-+ int n = nn;
-+
-+ FM_DMP_SUBTITLE(buf, n, "\n");
-+
-+ FM_DMP_TITLE(buf, n,
-+ p_pcd->p_FmPcdPlcr->p_FmPcdPlcrRegs,
-+ "FM policer regs");
-+
-+ FM_DMP_V32(buf, n, p_pcd->p_FmPcdPlcr->p_FmPcdPlcrRegs, fmpl_gcr);
-+ FM_DMP_V32(buf, n, p_pcd->p_FmPcdPlcr->p_FmPcdPlcrRegs, fmpl_gsr);
-+ FM_DMP_V32(buf, n, p_pcd->p_FmPcdPlcr->p_FmPcdPlcrRegs, fmpl_evr);
-+ FM_DMP_V32(buf, n, p_pcd->p_FmPcdPlcr->p_FmPcdPlcrRegs, fmpl_ier);
-+ FM_DMP_V32(buf, n, p_pcd->p_FmPcdPlcr->p_FmPcdPlcrRegs, fmpl_ifr);
-+ FM_DMP_V32(buf, n, p_pcd->p_FmPcdPlcr->p_FmPcdPlcrRegs, fmpl_eevr);
-+ FM_DMP_V32(buf, n, p_pcd->p_FmPcdPlcr->p_FmPcdPlcrRegs, fmpl_eier);
-+ FM_DMP_V32(buf, n, p_pcd->p_FmPcdPlcr->p_FmPcdPlcrRegs, fmpl_eifr);
-+ FM_DMP_V32(buf, n, p_pcd->p_FmPcdPlcr->p_FmPcdPlcrRegs, fmpl_rpcnt);
-+ FM_DMP_V32(buf, n, p_pcd->p_FmPcdPlcr->p_FmPcdPlcrRegs, fmpl_ypcnt);
-+ FM_DMP_V32(buf, n, p_pcd->p_FmPcdPlcr->p_FmPcdPlcrRegs, fmpl_rrpcnt);
-+ FM_DMP_V32(buf, n, p_pcd->p_FmPcdPlcr->p_FmPcdPlcrRegs, fmpl_rypcnt);
-+ FM_DMP_V32(buf, n, p_pcd->p_FmPcdPlcr->p_FmPcdPlcrRegs, fmpl_tpcnt);
-+ FM_DMP_V32(buf, n, p_pcd->p_FmPcdPlcr->p_FmPcdPlcrRegs, fmpl_flmcnt);
-+
-+ FM_DMP_V32(buf, n, p_pcd->p_FmPcdPlcr->p_FmPcdPlcrRegs, fmpl_serc);
-+ FM_DMP_V32(buf, n, p_pcd->p_FmPcdPlcr->p_FmPcdPlcrRegs, fmpl_upcr);
-+ FM_DMP_V32(buf, n, p_pcd->p_FmPcdPlcr->p_FmPcdPlcrRegs, fmpl_dpmr);
-+
-+ FM_DMP_TITLE(buf, n,
-+ &p_pcd->p_FmPcdPlcr->p_FmPcdPlcrRegs->fmpl_pmr,
-+ "fmpl_pmr");
-+
-+ for (i = 0; i < 63; ++i)
-+ FM_DMP_MEM_32(buf, n,
-+ &p_pcd->p_FmPcdPlcr->p_FmPcdPlcrRegs->fmpl_pmr[i]);
-+
-+ return n;
-+}
-+
-+int fm_get_counter(void *h_fm, e_FmCounters cnt_e, uint32_t *cnt_val)
-+{
-+ t_Fm *p_fm = (t_Fm *)h_fm;
-+
-+ /* When applicable (when there is an "enable counters" bit),
-+ check that counters are enabled */
-+
-+ switch (cnt_e) {
-+ case (e_FM_COUNTERS_DEQ_1):
-+ case (e_FM_COUNTERS_DEQ_2):
-+ case (e_FM_COUNTERS_DEQ_3):
-+ if (p_fm->p_FmStateStruct->revInfo.majorRev >= 6)
-+ return -EINVAL; /* counter not available */
-+
-+ case (e_FM_COUNTERS_ENQ_TOTAL_FRAME):
-+ case (e_FM_COUNTERS_DEQ_TOTAL_FRAME):
-+ case (e_FM_COUNTERS_DEQ_0):
-+ case (e_FM_COUNTERS_DEQ_FROM_DEFAULT):
-+ case (e_FM_COUNTERS_DEQ_FROM_CONTEXT):
-+ case (e_FM_COUNTERS_DEQ_FROM_FD):
-+ case (e_FM_COUNTERS_DEQ_CONFIRM):
-+ if (!(ioread32be(&p_fm->p_FmQmiRegs->fmqm_gc) &
-+ QMI_CFG_EN_COUNTERS))
-+ return -EINVAL; /* Requested counter not available */
-+ break;
-+ default:
-+ break;
-+ }
-+
-+ switch (cnt_e) {
-+ case (e_FM_COUNTERS_ENQ_TOTAL_FRAME):
-+ *cnt_val = ioread32be(&p_fm->p_FmQmiRegs->fmqm_etfc);
-+ return 0;
-+ case (e_FM_COUNTERS_DEQ_TOTAL_FRAME):
-+ *cnt_val = ioread32be(&p_fm->p_FmQmiRegs->fmqm_dtfc);
-+ return 0;
-+ case (e_FM_COUNTERS_DEQ_0):
-+ *cnt_val = ioread32be(&p_fm->p_FmQmiRegs->fmqm_dc0);
-+ return 0;
-+ case (e_FM_COUNTERS_DEQ_1):
-+ *cnt_val = ioread32be(&p_fm->p_FmQmiRegs->fmqm_dc1);
-+ return 0;
-+ case (e_FM_COUNTERS_DEQ_2):
-+ *cnt_val = ioread32be(&p_fm->p_FmQmiRegs->fmqm_dc2);
-+ return 0;
-+ case (e_FM_COUNTERS_DEQ_3):
-+ *cnt_val = ioread32be(&p_fm->p_FmQmiRegs->fmqm_dc3);
-+ return 0;
-+ case (e_FM_COUNTERS_DEQ_FROM_DEFAULT):
-+ *cnt_val = ioread32be(&p_fm->p_FmQmiRegs->fmqm_dfdc);
-+ return 0;
-+ case (e_FM_COUNTERS_DEQ_FROM_CONTEXT):
-+ *cnt_val = ioread32be(&p_fm->p_FmQmiRegs->fmqm_dfcc);
-+ return 0;
-+ case (e_FM_COUNTERS_DEQ_FROM_FD):
-+ *cnt_val = ioread32be(&p_fm->p_FmQmiRegs->fmqm_dffc);
-+ return 0;
-+ case (e_FM_COUNTERS_DEQ_CONFIRM):
-+ *cnt_val = ioread32be(&p_fm->p_FmQmiRegs->fmqm_dcc);
-+ return 0;
-+ }
-+ /* should never get here */
-+ return -EINVAL; /* counter not available */
-+}
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/src/wrapper/lnxwrp_sysfs_fm.h
-@@ -0,0 +1,136 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+#ifndef LNXWRP_SYSFS_FM_H_
-+#define LNXWRP_SYSFS_FM_H_
-+
-+#include "lnxwrp_sysfs.h"
-+
-+int fm_sysfs_create(struct device *dev);
-+void fm_sysfs_destroy(struct device *dev);
-+int fm_dump_regs(void *h_dev, char *buf, int nn);
-+int fm_fpm_dump_regs(void *h_dev, char *buf, int nn);
-+int fm_kg_dump_regs(void *h_pcd, char *buf, int nn);
-+int fm_kg_pe_dump_regs(void *h_pcd, char *buf, int nn);
-+int fm_dump_scheme(void *h_pcd, int scnum, char *buf, int nn);
-+int fm_dump_tnum_dbg(void *h_fm, int tn_s, int tn_e, char *buf, int nn);
-+int fm_dump_cls_plan(void *h_pcd, int cpn, char *buf, int nn);
-+int fm_plcr_dump_regs(void *h_pcd, char *buf, int nn);
-+int fm_prs_dump_regs(void *h_pcd, char *buf, int nn);
-+int fm_profile_dump_regs(void *h_pcd, int ppnum, char *buf, int nn);
-+
-+#define FM_DMP_PGSZ_ERR { \
-+ snprintf(&buf[PAGE_SIZE - 80], 70, \
-+ "\n Err: current sysfs buffer reached PAGE_SIZE\n");\
-+ n = PAGE_SIZE - 2; \
-+ }
-+
-+#define FM_DMP_LN(buf, n, ...) \
-+ do { \
-+ int k, m = n; \
-+ m += k = snprintf(&buf[m], PAGE_SIZE - m, __VA_ARGS__); \
-+ if (k < 0 || m > PAGE_SIZE - 90) \
-+ FM_DMP_PGSZ_ERR \
-+ n = m; \
-+ } while (0)
-+
-+#define FM_DMP_TITLE(buf, n, addr, ...) \
-+ do { \
-+ int k, m = n; \
-+ m += k = snprintf(&buf[m], PAGE_SIZE - m, "\n"); \
-+ if (k < 0 || m > PAGE_SIZE - 90) \
-+ FM_DMP_PGSZ_ERR \
-+ m += k = snprintf(&buf[m], PAGE_SIZE - m, __VA_ARGS__); \
-+ if (k < 0 || m > PAGE_SIZE - 90) \
-+ FM_DMP_PGSZ_ERR \
-+ if (addr) { \
-+ phys_addr_t pa; \
-+ pa = virt_to_phys(addr); \
-+ m += k = \
-+ snprintf(&buf[m], PAGE_SIZE - m, " (0x%lX)", \
-+ (long unsigned int)(pa)); \
-+ if (k < 0 || m > PAGE_SIZE - 90) \
-+ FM_DMP_PGSZ_ERR \
-+ } \
-+ m += k = snprintf(&buf[m], PAGE_SIZE - m, \
-+ "\n----------------------------------------\n\n"); \
-+ if (k < 0 || m > PAGE_SIZE - 90) \
-+ FM_DMP_PGSZ_ERR \
-+ n = m; \
-+ } while (0)
-+
-+#define FM_DMP_SUBTITLE(buf, n, ...) \
-+ do { \
-+ int k, m = n; \
-+ m += k = snprintf(&buf[m], PAGE_SIZE - m, "------- "); \
-+ if (k < 0 || m > PAGE_SIZE - 90) \
-+ FM_DMP_PGSZ_ERR \
-+ m += k = snprintf(&buf[m], PAGE_SIZE - m, __VA_ARGS__); \
-+ if (k < 0 || m > PAGE_SIZE - 90) \
-+ FM_DMP_PGSZ_ERR \
-+ m += k = snprintf(&buf[m], PAGE_SIZE - m, "\n"); \
-+ if (k < 0 || m > PAGE_SIZE - 90) \
-+ FM_DMP_PGSZ_ERR \
-+ n = m; \
-+ } while (0)
-+
-+#define FM_DMP_MEM_32(buf, n, addr) \
-+ { \
-+ uint32_t val; \
-+ phys_addr_t pa; \
-+ int k, m = n; \
-+ pa = virt_to_phys(addr); \
-+ val = ioread32be((addr)); \
-+ do { \
-+ m += k = snprintf(&buf[m], \
-+ PAGE_SIZE - m, "0x%010llX: 0x%08x\n", \
-+ pa, val); \
-+ if (k < 0 || m > PAGE_SIZE - 90) \
-+ FM_DMP_PGSZ_ERR \
-+ n += k; \
-+ } while (0) ;\
-+ }
-+
-+#define FM_DMP_V32(buf, n, st, phrase) \
-+ do { \
-+ int k, m = n; \
-+ phys_addr_t pa = virt_to_phys(&((st)->phrase)); \
-+ k = snprintf(&buf[m], PAGE_SIZE - m, \
-+ "0x%010llX: 0x%08x%8s\t%s\n", (unsigned long long) pa, \
-+ ioread32be((uint32_t *)&((st)->phrase)), "", #phrase); \
-+ if (k < 0 || m > PAGE_SIZE - 90) \
-+ FM_DMP_PGSZ_ERR \
-+ n += k; \
-+ } while (0)
-+
-+#endif /* LNXWRP_SYSFS_FM_H_ */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/src/wrapper/lnxwrp_sysfs_fm_port.c
-@@ -0,0 +1,1268 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#include "lnxwrp_sysfs.h"
-+#include "lnxwrp_fm.h"
-+#include "debug_ext.h"
-+#include "lnxwrp_sysfs_fm_port.h"
-+#include "lnxwrp_sysfs_fm.h"
-+
-+#include "../../sdk_fman/Peripherals/FM/Port/fm_port.h"
-+#include "../../sdk_fman/Peripherals/FM/Port/fm_port_dsar.h"
-+
-+#if defined(__ERR_MODULE__)
-+#undef __ERR_MODULE__
-+#endif
-+
-+#include "../../sdk_fman/Peripherals/FM/fm.h"
-+
-+static const struct sysfs_stats_t portSysfsStats[] = {
-+ /* RX/TX/OH common statistics */
-+ {
-+ .stat_name = "port_frame",
-+ .stat_counter = e_FM_PORT_COUNTERS_FRAME,
-+ },
-+ {
-+ .stat_name = "port_discard_frame",
-+ .stat_counter = e_FM_PORT_COUNTERS_DISCARD_FRAME,
-+ },
-+ {
-+ .stat_name = "port_dealloc_buf",
-+ .stat_counter = e_FM_PORT_COUNTERS_DEALLOC_BUF,
-+ },
-+ {
-+ .stat_name = "port_enq_total",
-+ .stat_counter = e_FM_PORT_COUNTERS_ENQ_TOTAL,
-+ },
-+ /* TX/OH */
-+ {
-+ .stat_name = "port_length_err",
-+ .stat_counter = e_FM_PORT_COUNTERS_LENGTH_ERR,
-+ },
-+ {
-+ .stat_name = "port_unsupprted_format",
-+ .stat_counter = e_FM_PORT_COUNTERS_UNSUPPRTED_FORMAT,
-+ },
-+ {
-+ .stat_name = "port_deq_total",
-+ .stat_counter = e_FM_PORT_COUNTERS_DEQ_TOTAL,
-+ },
-+ {
-+ .stat_name = "port_deq_from_default",
-+ .stat_counter = e_FM_PORT_COUNTERS_DEQ_FROM_DEFAULT,
-+ },
-+ {
-+ .stat_name = "port_deq_confirm",
-+ .stat_counter = e_FM_PORT_COUNTERS_DEQ_CONFIRM,
-+ },
-+ /* RX/OH */
-+ {
-+ .stat_name = "port_rx_bad_frame",
-+ .stat_counter = e_FM_PORT_COUNTERS_RX_BAD_FRAME,
-+ },
-+ {
-+ .stat_name = "port_rx_large_frame",
-+ .stat_counter = e_FM_PORT_COUNTERS_RX_LARGE_FRAME,
-+ },
-+ {
-+ .stat_name = "port_rx_out_of_buffers_discard",
-+ .stat_counter = e_FM_PORT_COUNTERS_RX_OUT_OF_BUFFERS_DISCARD,
-+ },
-+ {
-+ .stat_name = "port_rx_filter_frame",
-+ .stat_counter = e_FM_PORT_COUNTERS_RX_FILTER_FRAME,
-+ },
-+ /* TODO: Particular statistics for OH ports */
-+ {}
-+};
-+
-+static ssize_t show_fm_port_stats(struct device *dev,
-+ struct device_attribute *attr, char *buf)
-+{
-+ t_LnxWrpFmPortDev *p_LnxWrpFmPortDev;
-+ t_LnxWrpFmDev *p_LnxWrpFmDev;
-+ unsigned long flags;
-+ int n = 0;
-+ uint8_t counter = 0;
-+
-+ if (attr == NULL || buf == NULL || dev == NULL)
-+ return -EINVAL;
-+
-+ p_LnxWrpFmPortDev = (t_LnxWrpFmPortDev *) dev_get_drvdata(dev);
-+ if (WARN_ON(p_LnxWrpFmPortDev == NULL))
-+ return -EINVAL;
-+
-+ p_LnxWrpFmDev = (t_LnxWrpFmDev *) p_LnxWrpFmPortDev->h_LnxWrpFmDev;
-+ if (WARN_ON(p_LnxWrpFmDev == NULL))
-+ return -EINVAL;
-+
-+ if (!p_LnxWrpFmDev->active || !p_LnxWrpFmDev->h_Dev)
-+ return -EIO;
-+
-+ if (!p_LnxWrpFmPortDev->h_Dev) {
-+ n = snprintf(buf, PAGE_SIZE, "\tFM Port not configured...\n");
-+ return n;
-+ }
-+
-+ counter = fm_find_statistic_counter_by_name(
-+ attr->attr.name,
-+ portSysfsStats, NULL);
-+
-+ if (counter == e_FM_PORT_COUNTERS_RX_LIST_DMA_ERR) {
-+ uint32_t fmRev = 0;
-+ fmRev = 0xffff &
-+ ioread32(UINT_TO_PTR(p_LnxWrpFmDev->fmBaseAddr +
-+ 0x000c30c4));
-+
-+ if (fmRev == 0x0100) {
-+ local_irq_save(flags);
-+ n = snprintf(buf, PAGE_SIZE,
-+ "counter not available for revision 1\n");
-+ local_irq_restore(flags);
-+ }
-+ return n;
-+ }
-+
-+ local_irq_save(flags);
-+ n = snprintf(buf, PAGE_SIZE, "\t%s counter: %u\n",
-+ p_LnxWrpFmPortDev->name,
-+ FM_PORT_GetCounter(p_LnxWrpFmPortDev->h_Dev,
-+ (e_FmPortCounters) counter));
-+ local_irq_restore(flags);
-+
-+ return n;
-+}
-+
-+/* FM PORT RX/TX/OH statistics */
-+static DEVICE_ATTR(port_frame, S_IRUGO, show_fm_port_stats, NULL);
-+static DEVICE_ATTR(port_discard_frame, S_IRUGO, show_fm_port_stats, NULL);
-+static DEVICE_ATTR(port_dealloc_buf, S_IRUGO, show_fm_port_stats, NULL);
-+static DEVICE_ATTR(port_enq_total, S_IRUGO, show_fm_port_stats, NULL);
-+/* FM PORT TX/OH statistics */
-+static DEVICE_ATTR(port_length_err, S_IRUGO, show_fm_port_stats, NULL);
-+static DEVICE_ATTR(port_unsupprted_format, S_IRUGO, show_fm_port_stats, NULL);
-+static DEVICE_ATTR(port_deq_total, S_IRUGO, show_fm_port_stats, NULL);
-+static DEVICE_ATTR(port_deq_from_default, S_IRUGO, show_fm_port_stats, NULL);
-+static DEVICE_ATTR(port_deq_confirm, S_IRUGO, show_fm_port_stats, NULL);
-+/* FM PORT RX/OH statistics */
-+static DEVICE_ATTR(port_rx_bad_frame, S_IRUGO, show_fm_port_stats, NULL);
-+static DEVICE_ATTR(port_rx_large_frame, S_IRUGO, show_fm_port_stats, NULL);
-+static DEVICE_ATTR(port_rx_out_of_buffers_discard, S_IRUGO,
-+ show_fm_port_stats, NULL);
-+static DEVICE_ATTR(port_rx_filter_frame, S_IRUGO, show_fm_port_stats, NULL);
-+
-+/* FM PORT TX statistics */
-+static struct attribute *fm_tx_port_dev_stats_attributes[] = {
-+ &dev_attr_port_frame.attr,
-+ &dev_attr_port_discard_frame.attr,
-+ &dev_attr_port_dealloc_buf.attr,
-+ &dev_attr_port_enq_total.attr,
-+ &dev_attr_port_length_err.attr,
-+ &dev_attr_port_unsupprted_format.attr,
-+ &dev_attr_port_deq_total.attr,
-+ &dev_attr_port_deq_from_default.attr,
-+ &dev_attr_port_deq_confirm.attr,
-+ NULL
-+};
-+
-+static const struct attribute_group fm_tx_port_dev_stats_attr_grp = {
-+ .name = "statistics",
-+ .attrs = fm_tx_port_dev_stats_attributes
-+};
-+
-+/* FM PORT RX statistics */
-+static struct attribute *fm_rx_port_dev_stats_attributes[] = {
-+ &dev_attr_port_frame.attr,
-+ &dev_attr_port_discard_frame.attr,
-+ &dev_attr_port_dealloc_buf.attr,
-+ &dev_attr_port_enq_total.attr,
-+ &dev_attr_port_rx_bad_frame.attr,
-+ &dev_attr_port_rx_large_frame.attr,
-+ &dev_attr_port_rx_out_of_buffers_discard.attr,
-+ &dev_attr_port_rx_filter_frame.attr,
-+ NULL
-+};
-+
-+static const struct attribute_group fm_rx_port_dev_stats_attr_grp = {
-+ .name = "statistics",
-+ .attrs = fm_rx_port_dev_stats_attributes
-+};
-+
-+/* TODO: add particular OH ports statistics */
-+static struct attribute *fm_oh_port_dev_stats_attributes[] = {
-+ &dev_attr_port_frame.attr,
-+ &dev_attr_port_discard_frame.attr,
-+ &dev_attr_port_dealloc_buf.attr,
-+ &dev_attr_port_enq_total.attr,
-+ /*TX*/ &dev_attr_port_length_err.attr,
-+ &dev_attr_port_unsupprted_format.attr,
-+ &dev_attr_port_deq_total.attr,
-+ &dev_attr_port_deq_from_default.attr,
-+ &dev_attr_port_deq_confirm.attr,
-+ /* &dev_attr_port_rx_bad_frame.attr, */
-+ /* &dev_attr_port_rx_large_frame.attr, */
-+ &dev_attr_port_rx_out_of_buffers_discard.attr,
-+ /*&dev_attr_port_rx_filter_frame.attr, */
-+ NULL
-+};
-+
-+static const struct attribute_group fm_oh_port_dev_stats_attr_grp = {
-+ .name = "statistics",
-+ .attrs = fm_oh_port_dev_stats_attributes
-+};
-+
-+static ssize_t show_fm_port_regs(struct device *dev,
-+ struct device_attribute *attr, char *buf)
-+{
-+ unsigned long flags;
-+ unsigned n = 0;
-+#if (defined(DEBUG_ERRORS) && (DEBUG_ERRORS > 0))
-+ t_LnxWrpFmPortDev *p_LnxWrpFmPortDev;
-+#endif
-+ if (attr == NULL || buf == NULL || dev == NULL)
-+ return -EINVAL;
-+
-+#if (defined(DEBUG_ERRORS) && (DEBUG_ERRORS > 0))
-+ p_LnxWrpFmPortDev =
-+ (t_LnxWrpFmPortDev *) dev_get_drvdata(dev);
-+
-+
-+ local_irq_save(flags);
-+
-+ if (!p_LnxWrpFmPortDev->h_Dev) {
-+ n = snprintf(buf, PAGE_SIZE, "\tFM Port not configured...\n");
-+ return n;
-+ } else {
-+ n = snprintf(buf, PAGE_SIZE,
-+ "FM port driver registers dump.\n");
-+ n = fm_port_dump_regs(p_LnxWrpFmPortDev->h_Dev, buf, n);
-+ }
-+
-+ local_irq_restore(flags);
-+
-+ return n;
-+#else
-+
-+ local_irq_save(flags);
-+ n = snprintf(buf, PAGE_SIZE,
-+ "Debug level is too low to dump registers!!!\n");
-+ local_irq_restore(flags);
-+
-+ return n;
-+#endif
-+}
-+static int fm_port_dsar_dump_mem(void *h_dev, char *buf, int nn)
-+{
-+ t_FmPort *p_FmPort;
-+ t_Fm *p_Fm;
-+ uint8_t hardwarePortId;
-+ uint32_t *param_page;
-+ t_ArCommonDesc *ArCommonDescPtr;
-+ uint32_t *mem;
-+ int i, n = nn;
-+
-+ p_FmPort = (t_FmPort *)h_dev;
-+ hardwarePortId = p_FmPort->hardwarePortId;
-+ p_Fm = (t_Fm *)p_FmPort->h_Fm;
-+
-+ if (!FM_PORT_IsInDsar(p_FmPort))
-+ {
-+ FM_DMP_LN(buf, n, "port %u is not a DSAR port\n",
-+ hardwarePortId);
-+ return n;
-+ }
-+ FM_DMP_LN(buf, n, "port %u DSAR mem\n", hardwarePortId);
-+ FM_DMP_LN(buf, n, "========================\n");
-+
-+ /* do I need request_mem_region here? */
-+ param_page = ioremap(p_FmPort->fmMuramPhysBaseAddr + ioread32be(&p_FmPort->p_FmPortBmiRegs->rxPortBmiRegs.fmbm_rgpr), 4);
-+ ArCommonDescPtr = (t_ArCommonDesc*)(ioremap(p_FmPort->fmMuramPhysBaseAddr + ioread32be(param_page), 300*4)); /* this should be changed*/
-+ mem = (uint32_t*)ArCommonDescPtr;
-+ for (i = 0; i < 300; i+=4)
-+ FM_DMP_LN(buf, n, "%08x: %08x %08x %08x %08x\n", i*4, mem[i], mem[i + 1], mem[i + 2], mem[i + 3]);
-+ iounmap(ArCommonDescPtr);
-+ iounmap(param_page);
-+ return n;
-+}
-+
-+static int fm_port_dsar_dump_regs(void *h_dev, char *buf, int nn)
-+{
-+ t_FmPort *p_FmPort;
-+ t_Fm *p_Fm;
-+ uint8_t hardwarePortId;
-+ uint32_t *param_page;
-+ t_ArCommonDesc *ArCommonDescPtr;
-+ int i, n = nn;
-+
-+ p_FmPort = (t_FmPort *)h_dev;
-+ hardwarePortId = p_FmPort->hardwarePortId;
-+ p_Fm = (t_Fm *)p_FmPort->h_Fm;
-+
-+ if (!FM_PORT_IsInDsar(p_FmPort))
-+ {
-+ FM_DMP_LN(buf, n, "port %u is not a DSAR port\n",
-+ hardwarePortId);
-+ return n;
-+ }
-+ FM_DMP_LN(buf, n, "port %u DSAR information\n", hardwarePortId);
-+ FM_DMP_LN(buf, n, "========================\n");
-+
-+ /* do I need request_mem_region here? */
-+ param_page = ioremap(p_FmPort->fmMuramPhysBaseAddr + ioread32be(&p_FmPort->p_FmPortBmiRegs->rxPortBmiRegs.fmbm_rgpr), 4);
-+ ArCommonDescPtr = (t_ArCommonDesc*)(ioremap(p_FmPort->fmMuramPhysBaseAddr + ioread32be(param_page), sizeof(t_ArCommonDesc))); /* this should be changed*/
-+ FM_DMP_LN(buf, n, "Tx port: 0x%x\n", ArCommonDescPtr->arTxPort);
-+ FM_DMP_LN(buf, n, "Active HPNIA: 0x%08x\n", ArCommonDescPtr->activeHPNIA);
-+ FM_DMP_LN(buf, n, "Snmp port: 0x%x\n", ArCommonDescPtr->snmpPort);
-+ FM_DMP_LN(buf, n, "MAC address: %02x:%02x:%02x:%02x:%02x:%02x\n", ArCommonDescPtr->macStationAddr[0],
-+ ArCommonDescPtr->macStationAddr[1], ArCommonDescPtr->macStationAddr[2],
-+ ArCommonDescPtr->macStationAddr[3], ArCommonDescPtr->macStationAddr[4],
-+ ArCommonDescPtr->macStationAddr[5]);
-+ FM_DMP_LN(buf, n, "filterControl: 0x%02x\n", ArCommonDescPtr->filterControl);
-+ FM_DMP_LN(buf, n, "tcpControlPass: 0x%04x\n", ArCommonDescPtr->tcpControlPass);
-+ FM_DMP_LN(buf, n, "ipProtocolTblSize: 0x%x\n", ArCommonDescPtr->ipProtocolTblSize);
-+ FM_DMP_LN(buf, n, "udpPortTblSize: 0x%x\n", ArCommonDescPtr->udpPortTblSize);
-+ FM_DMP_LN(buf, n, "tcpPortTblSize: 0x%x\n", ArCommonDescPtr->tcpPortTblSize);
-+ if (ArCommonDescPtr->p_ArStats)
-+ {
-+ t_ArStatistics *arStatistics = (t_ArStatistics*)
-+ ioremap(ioread32be(&ArCommonDescPtr->p_ArStats) +
-+ p_FmPort->fmMuramPhysBaseAddr,
-+ sizeof (t_ArStatistics));
-+ FM_DMP_LN(buf, n, "\nDSAR statistics\n");
-+ FM_DMP_LN(buf, n, "DSAR_Discarded: 0x%x\n", arStatistics->dsarDiscarded);
-+ FM_DMP_LN(buf, n, "DSAR_Err_Discarded: 0x%x\n", arStatistics->dsarErrDiscarded);
-+ FM_DMP_LN(buf, n, "DSAR_Frag_Discarded: 0x%x\n", arStatistics->dsarFragDiscarded);
-+ FM_DMP_LN(buf, n, "DSAR_Tunnel_Discarded: 0x%x\n", arStatistics->dsarTunnelDiscarded);
-+ FM_DMP_LN(buf, n, "DSAR_ARP_Discarded: 0x%x\n", arStatistics->dsarArpDiscarded);
-+ FM_DMP_LN(buf, n, "DSAR_IP_Discarded: 0x%x\n", arStatistics->dsarIpDiscarded);
-+ FM_DMP_LN(buf, n, "DSAR_TCP_Discarded: 0x%x\n", arStatistics->dsarTcpDiscarded);
-+ FM_DMP_LN(buf, n, "DSAR_UDP_Discarded: 0x%x\n", arStatistics->dsarUdpDiscarded);
-+ FM_DMP_LN(buf, n, "DSAR_ICMPv6_Checksum_Err: 0x%x\n", arStatistics->dsarIcmpV6ChecksumErr);
-+ FM_DMP_LN(buf, n, "DSAR_ICMPv6_Other_Type: 0x%x\n", arStatistics->dsarIcmpV6OtherType);
-+ FM_DMP_LN(buf, n, "DSAR_ICMPv4_Other_Type: 0x%x\n", arStatistics->dsarIcmpV4OtherType);
-+
-+ iounmap(arStatistics);
-+ }
-+ if (ArCommonDescPtr->p_ArpDescriptor)
-+ {
-+ t_DsarArpDescriptor* ArpDescriptor = (t_DsarArpDescriptor*)
-+ ioremap(ioread32be(&ArCommonDescPtr->p_ArpDescriptor) +
-+ p_FmPort->fmMuramPhysBaseAddr,
-+ sizeof (t_DsarArpDescriptor));
-+ FM_DMP_LN(buf, n, "\nARP\n");
-+ FM_DMP_LN(buf, n, "===\n");
-+ FM_DMP_LN(buf, n, "control bits 0x%04x\n", ArpDescriptor->control);
-+ if (ArpDescriptor->numOfBindings)
-+ {
-+ char ip_str[100];
-+ t_DsarArpBindingEntry* bindings = ioremap(
-+ ioread32be(&ArpDescriptor->p_Bindings) +
-+ p_FmPort->fmMuramPhysBaseAddr,
-+ ArpDescriptor->numOfBindings *
-+ sizeof(t_DsarArpBindingEntry));
-+ uint8_t* ip_addr = (uint8_t*)&bindings->ipv4Addr;
-+ FM_DMP_LN(buf, n, " ip vlan id\n");
-+ for (i = 0; i < ArpDescriptor->numOfBindings; i++)
-+ {
-+ n += snprintf(ip_str, 100, "%d.%d.%d.%d",
-+ ip_addr[0], ip_addr[1],
-+ ip_addr[2], ip_addr[3]);
-+ FM_DMP_LN(buf, n, "%-15s 0x%x\n",
-+ ip_str, bindings->vlanId);
-+ }
-+ iounmap(bindings);
-+ }
-+ if (ArpDescriptor->p_Statistics)
-+ {
-+ t_DsarArpStatistics* arpStats = ioremap(
-+ ioread32be(&ArpDescriptor->p_Statistics) +
-+ p_FmPort->fmMuramPhysBaseAddr,
-+ sizeof(t_DsarArpStatistics));
-+ FM_DMP_LN(buf, n, "statistics\n");
-+ FM_DMP_LN(buf, n, "INVAL_CNT: 0x%x\n", arpStats->invalCnt);
-+ FM_DMP_LN(buf, n, "ECHO_CNT: 0x%x\n", arpStats->echoCnt);
-+ FM_DMP_LN(buf, n, "CD_CNT: 0x%x\n", arpStats->cdCnt);
-+ FM_DMP_LN(buf, n, "AR_CNT: 0x%x\n", arpStats->arCnt);
-+ FM_DMP_LN(buf, n, "RATM_CNT: 0x%x\n", arpStats->ratmCnt);
-+ FM_DMP_LN(buf, n, "UKOP_CNT: 0x%x\n", arpStats->ukopCnt);
-+ FM_DMP_LN(buf, n, "NMTP_CNT: 0x%x\n", arpStats->nmtpCnt);
-+ FM_DMP_LN(buf, n, "NMVLAN_CNT: 0x%x\n", arpStats->nmVlanCnt);
-+ iounmap(arpStats);
-+ }
-+
-+ iounmap(ArpDescriptor);
-+ }
-+ if (ArCommonDescPtr->p_IcmpV4Descriptor)
-+ {
-+ t_DsarIcmpV4Descriptor* ICMPV4Descriptor =
-+ (t_DsarIcmpV4Descriptor*)ioremap(ioread32be(
-+ &ArCommonDescPtr->p_IcmpV4Descriptor) +
-+ p_FmPort->fmMuramPhysBaseAddr,
-+ sizeof (t_DsarIcmpV4Descriptor));
-+ FM_DMP_LN(buf, n, "\nEcho ICMPv4\n");
-+ FM_DMP_LN(buf, n, "===========\n");
-+ FM_DMP_LN(buf, n, "control bits 0x%04x\n", ICMPV4Descriptor->control);
-+ if (ICMPV4Descriptor->numOfBindings)
-+ {
-+ char ip_str[100];
-+ t_DsarArpBindingEntry* bindings = ioremap(
-+ ioread32be(&ICMPV4Descriptor->p_Bindings) +
-+ p_FmPort->fmMuramPhysBaseAddr,
-+ ICMPV4Descriptor->numOfBindings *
-+ sizeof(t_DsarArpBindingEntry));
-+ uint8_t* ip_addr = (uint8_t*)&bindings->ipv4Addr;
-+ FM_DMP_LN(buf, n, " ip vlan id\n");
-+ for (i = 0; i < ICMPV4Descriptor->numOfBindings; i++)
-+ {
-+ n += snprintf(ip_str, 100, "%d.%d.%d.%d",
-+ ip_addr[0], ip_addr[1],
-+ ip_addr[2], ip_addr[3]);
-+ FM_DMP_LN(buf, n, "%-15s 0x%x\n",
-+ ip_str, bindings->vlanId);
-+ }
-+ iounmap(bindings);
-+ }
-+ if (ICMPV4Descriptor->p_Statistics)
-+ {
-+ t_DsarIcmpV4Statistics* icmpv4Stats = ioremap(
-+ ioread32be(&ICMPV4Descriptor->p_Statistics) +
-+ p_FmPort->fmMuramPhysBaseAddr,
-+ sizeof(t_DsarIcmpV4Statistics));
-+ FM_DMP_LN(buf, n, "statistics\n");
-+ FM_DMP_LN(buf, n, "INVAL_CNT: 0x%x\n", icmpv4Stats->invalCnt);
-+ FM_DMP_LN(buf, n, "NMVLAN_CNT: 0x%x\n", icmpv4Stats->nmVlanCnt);
-+ FM_DMP_LN(buf, n, "NMIP_CNT: 0x%x\n", icmpv4Stats->nmIpCnt);
-+ FM_DMP_LN(buf, n, "AR_CNT: 0x%x\n", icmpv4Stats->arCnt);
-+ FM_DMP_LN(buf, n, "CSERR_CNT: 0x%x\n", icmpv4Stats->cserrCnt);
-+ iounmap(icmpv4Stats);
-+ }
-+ iounmap(ICMPV4Descriptor);
-+ }
-+ if (ArCommonDescPtr->p_NdDescriptor)
-+ {
-+ t_DsarNdDescriptor *NDDescriptor =
-+ (t_DsarNdDescriptor*)ioremap(ioread32be(
-+ &ArCommonDescPtr->p_NdDescriptor) + p_FmPort->
-+ fmMuramPhysBaseAddr, sizeof (t_DsarNdDescriptor));
-+ FM_DMP_LN(buf, n, "\nNDP\n");
-+ FM_DMP_LN(buf, n, "===\n");
-+ FM_DMP_LN(buf, n, "control bits 0x%04x\n", NDDescriptor->control);
-+ FM_DMP_LN(buf, n, "solicited address 0x%08x\n", NDDescriptor->solicitedAddr);
-+ if (NDDescriptor->numOfBindings)
-+ {
-+ char ip_str[100];
-+ t_DsarIcmpV6BindingEntry* bindings = ioremap(
-+ ioread32be(&NDDescriptor->p_Bindings) +
-+ p_FmPort->fmMuramPhysBaseAddr,
-+ NDDescriptor->numOfBindings *
-+ sizeof(t_DsarIcmpV6BindingEntry));
-+ uint16_t* ip_addr = (uint16_t*)&bindings->ipv6Addr;
-+ FM_DMP_LN(buf, n, " ip vlan id\n");
-+ for (i = 0; i < NDDescriptor->numOfBindings; i++)
-+ {
-+ n += snprintf(ip_str, 100,
-+ "%04x:%04x:%04x:%04x:%04x:%04x:%04x:%04x",
-+ ip_addr[0], ip_addr[1], ip_addr[2], ip_addr[3],
-+ ip_addr[4], ip_addr[5], ip_addr[6], ip_addr[7]);
-+ FM_DMP_LN(buf, n, "%s 0x%x\n", ip_str, bindings->vlanId);
-+ }
-+ iounmap(bindings);
-+ }
-+ if (NDDescriptor->p_Statistics)
-+ {
-+ t_NdStatistics* ndStats = ioremap(
-+ ioread32be(&NDDescriptor->p_Statistics) +
-+ p_FmPort->fmMuramPhysBaseAddr,
-+ sizeof(t_NdStatistics));
-+ FM_DMP_LN(buf, n, "statistics\n");
-+ FM_DMP_LN(buf, n, "INVAL_CNT: 0x%x\n", ndStats->invalCnt);
-+ FM_DMP_LN(buf, n, "NMVLAN_CNT: 0x%x\n", ndStats->nmVlanCnt);
-+ FM_DMP_LN(buf, n, "NMIP_CNT: 0x%x\n", ndStats->nmIpCnt);
-+ FM_DMP_LN(buf, n, "AR_CNT: 0x%x\n", ndStats->arCnt);
-+ FM_DMP_LN(buf, n, "USADVERT_CNT: 0x%x\n", ndStats->usadvertCnt);
-+ FM_DMP_LN(buf, n, "NMMCAST_CNT: 0x%x\n", ndStats->nmmcastCnt);
-+ FM_DMP_LN(buf, n, "NSLLA_CNT: 0x%x\n", ndStats->nsllaCnt);
-+ iounmap(ndStats);
-+ }
-+ iounmap(NDDescriptor);
-+ }
-+ if (ArCommonDescPtr->p_IcmpV6Descriptor)
-+ {
-+ t_DsarIcmpV6Descriptor *ICMPV6Descriptor =
-+ (t_DsarIcmpV6Descriptor*)ioremap(ioread32be(
-+ &ArCommonDescPtr->p_IcmpV6Descriptor) + p_FmPort->
-+ fmMuramPhysBaseAddr, sizeof (t_DsarIcmpV6Descriptor));
-+ FM_DMP_LN(buf, n, "\nEcho ICMPv6\n");
-+ FM_DMP_LN(buf, n, "===========\n");
-+ FM_DMP_LN(buf, n, "control bits 0x%04x\n", ICMPV6Descriptor->control);
-+ if (ICMPV6Descriptor->numOfBindings)
-+ {
-+ char ip_str[100];
-+ t_DsarIcmpV6BindingEntry* bindings = ioremap(
-+ ioread32be(&ICMPV6Descriptor->p_Bindings) +
-+ p_FmPort->fmMuramPhysBaseAddr,
-+ ICMPV6Descriptor->numOfBindings *
-+ sizeof(t_DsarIcmpV6BindingEntry));
-+ uint16_t* ip_addr = (uint16_t*)&bindings->ipv6Addr;
-+ FM_DMP_LN(buf, n, " ip vlan id\n");
-+ for (i = 0; i < ICMPV6Descriptor->numOfBindings; i++)
-+ {
-+ n += snprintf(ip_str, 100,
-+ "%04x:%04x:%04x:%04x:%04x:%04x:%04x:%04x",
-+ ip_addr[0], ip_addr[1], ip_addr[2], ip_addr[3],
-+ ip_addr[4], ip_addr[5], ip_addr[6], ip_addr[7]);
-+ FM_DMP_LN(buf, n, "%s 0x%x\n", ip_str, bindings->vlanId);
-+ }
-+ iounmap(bindings);
-+ }
-+ if (ICMPV6Descriptor->p_Statistics)
-+ {
-+ t_DsarIcmpV6Statistics* icmpv6Stats = ioremap(
-+ ioread32be(&ICMPV6Descriptor->p_Statistics) +
-+ p_FmPort->fmMuramPhysBaseAddr,
-+ sizeof(t_DsarIcmpV6Statistics));
-+ FM_DMP_LN(buf, n, "statistics\n");
-+ FM_DMP_LN(buf, n, "INVAL_CNT: 0x%x\n", icmpv6Stats->invalCnt);
-+ FM_DMP_LN(buf, n, "NMVLAN_CNT: 0x%x\n", icmpv6Stats->nmVlanCnt);
-+ FM_DMP_LN(buf, n, "NMIP_CNT: 0x%x\n", icmpv6Stats->nmIpCnt);
-+ FM_DMP_LN(buf, n, "AR_CNT: 0x%x\n", icmpv6Stats->arCnt);
-+ iounmap(icmpv6Stats);
-+ }
-+ iounmap(ICMPV6Descriptor);
-+ }
-+ if (ArCommonDescPtr->p_SnmpDescriptor)
-+ {
-+ t_DsarSnmpDescriptor *SnmpDescriptor =
-+ (t_DsarSnmpDescriptor*)ioremap(ioread32be(
-+ &ArCommonDescPtr->p_SnmpDescriptor) + p_FmPort->
-+ fmMuramPhysBaseAddr, sizeof (t_DsarSnmpDescriptor));
-+ FM_DMP_LN(buf, n, "\nSNMP\n");
-+ FM_DMP_LN(buf, n, "===========\n");
-+ FM_DMP_LN(buf, n, "control bits 0x%04x\n", SnmpDescriptor->control);
-+ FM_DMP_LN(buf, n, "max message length 0x%04x\n", SnmpDescriptor->maxSnmpMsgLength);
-+ if (SnmpDescriptor->numOfIpv4Addresses)
-+ {
-+ char ip_str[100];
-+ t_DsarSnmpIpv4AddrTblEntry* addrs = ioremap(
-+ ioread32be(&SnmpDescriptor->p_Ipv4AddrTbl) +
-+ p_FmPort->fmMuramPhysBaseAddr,
-+ SnmpDescriptor->numOfIpv4Addresses *
-+ sizeof(t_DsarSnmpIpv4AddrTblEntry));
-+ uint8_t* ip_addr = (uint8_t*)&addrs->ipv4Addr;
-+ FM_DMP_LN(buf, n, " ip vlan id\n");
-+ for (i = 0; i < SnmpDescriptor->numOfIpv4Addresses; i++)
-+ {
-+ n += snprintf(ip_str, 100, "%d.%d.%d.%d",
-+ ip_addr[0], ip_addr[1],
-+ ip_addr[2], ip_addr[3]);
-+ FM_DMP_LN(buf, n, "%-15s 0x%x\n", ip_str, addrs->vlanId);
-+ }
-+ iounmap(addrs);
-+ }
-+ if (SnmpDescriptor->p_Statistics)
-+ {
-+ t_DsarSnmpStatistics* snmpStats = ioremap(
-+ ioread32be(&SnmpDescriptor->p_Statistics) +
-+ p_FmPort->fmMuramPhysBaseAddr,
-+ sizeof(t_DsarSnmpStatistics));
-+ FM_DMP_LN(buf, n, "statistics\n");
-+ FM_DMP_LN(buf, n, "snmpErrCnt: 0x%x\n", snmpStats->snmpErrCnt);
-+ FM_DMP_LN(buf, n, "snmpCommunityErrCnt: 0x%x\n", snmpStats->snmpCommunityErrCnt);
-+ FM_DMP_LN(buf, n, "snmpTotalDiscardCnt: 0x%x\n", snmpStats->snmpTotalDiscardCnt);
-+ FM_DMP_LN(buf, n, "snmpGetReqCnt: 0x%x\n", snmpStats->snmpGetReqCnt);
-+ FM_DMP_LN(buf, n, "snmpGetNextReqCnt: 0x%x\n", snmpStats->snmpGetNextReqCnt);
-+ iounmap(snmpStats);
-+ }
-+ iounmap(SnmpDescriptor);
-+ }
-+ iounmap(ArCommonDescPtr);
-+ iounmap(param_page);
-+ return n;
-+}
-+
-+static ssize_t show_fm_port_dsar_mem(struct device *dev,
-+ struct device_attribute *attr, char *buf)
-+{
-+ unsigned long flags;
-+ unsigned n = 0;
-+#if (defined(DEBUG_ERRORS) && (DEBUG_ERRORS > 0))
-+ t_LnxWrpFmPortDev *p_LnxWrpFmPortDev;
-+#endif
-+ if (attr == NULL || buf == NULL || dev == NULL)
-+ return -EINVAL;
-+
-+#if (defined(DEBUG_ERRORS) && (DEBUG_ERRORS > 0))
-+ p_LnxWrpFmPortDev =
-+ (t_LnxWrpFmPortDev *) dev_get_drvdata(dev);
-+
-+ local_irq_save(flags);
-+
-+ if (!p_LnxWrpFmPortDev->h_Dev) {
-+ n = snprintf(buf, PAGE_SIZE, "\tFM Port not configured...\n");
-+ return n;
-+ } else {
-+ n = snprintf(buf, PAGE_SIZE,
-+ "FM port driver registers dump.\n");
-+ n = fm_port_dsar_dump_mem(p_LnxWrpFmPortDev->h_Dev, buf, n);
-+ }
-+
-+ local_irq_restore(flags);
-+
-+ return n;
-+#else
-+
-+ local_irq_save(flags);
-+ n = snprintf(buf, PAGE_SIZE,
-+ "Debug level is too low to dump registers!!!\n");
-+ local_irq_restore(flags);
-+
-+ return n;
-+#endif
-+}
-+
-+static ssize_t show_fm_port_dsar_regs(struct device *dev,
-+ struct device_attribute *attr, char *buf)
-+{
-+ unsigned long flags;
-+ unsigned n = 0;
-+#if (defined(DEBUG_ERRORS) && (DEBUG_ERRORS > 0))
-+ t_LnxWrpFmPortDev *p_LnxWrpFmPortDev;
-+#endif
-+ if (attr == NULL || buf == NULL || dev == NULL)
-+ return -EINVAL;
-+
-+#if (defined(DEBUG_ERRORS) && (DEBUG_ERRORS > 0))
-+ p_LnxWrpFmPortDev =
-+ (t_LnxWrpFmPortDev *) dev_get_drvdata(dev);
-+
-+ local_irq_save(flags);
-+
-+ if (!p_LnxWrpFmPortDev->h_Dev) {
-+ n = snprintf(buf, PAGE_SIZE, "\tFM Port not configured...\n");
-+ return n;
-+ } else {
-+ n = snprintf(buf, PAGE_SIZE,
-+ "FM port driver registers dump.\n");
-+ n = fm_port_dsar_dump_regs(p_LnxWrpFmPortDev->h_Dev, buf, n);
-+ }
-+
-+ local_irq_restore(flags);
-+
-+ return n;
-+#else
-+
-+ local_irq_save(flags);
-+ n = snprintf(buf, PAGE_SIZE,
-+ "Debug level is too low to dump registers!!!\n");
-+ local_irq_restore(flags);
-+
-+ return n;
-+#endif
-+}
-+
-+#if (DPAA_VERSION >= 11)
-+static ssize_t show_fm_port_ipv4_options(struct device *dev,
-+ struct device_attribute *attr, char *buf)
-+{
-+ unsigned long flags;
-+ unsigned n = 0;
-+#if (defined(DEBUG_ERRORS) && (DEBUG_ERRORS > 0))
-+ t_LnxWrpFmPortDev *p_LnxWrpFmPortDev;
-+#endif
-+
-+ if (attr == NULL || buf == NULL || dev == NULL)
-+ return -EINVAL;
-+
-+#if (defined(DEBUG_ERRORS) && (DEBUG_ERRORS > 0))
-+ p_LnxWrpFmPortDev =
-+ (t_LnxWrpFmPortDev *) dev_get_drvdata(dev);
-+
-+ local_irq_save(flags);
-+
-+ if (!p_LnxWrpFmPortDev->h_Dev) {
-+ n = snprintf(buf, PAGE_SIZE, "\tFM Port not configured...\n");
-+ return n;
-+ } else if (((t_FmPort *)p_LnxWrpFmPortDev->h_Dev)->p_ParamsPage
-+ == NULL) {
-+ n = snprintf(buf, PAGE_SIZE,
-+ "\tPort: FMan-controller params page not set\n");
-+ return n;
-+ } else {
-+ n = snprintf(buf, PAGE_SIZE,
-+ "Counter for fragmented pkt with IP header options\n");
-+ n = fm_port_dump_ipv4_opt(p_LnxWrpFmPortDev->h_Dev, buf, n);
-+ }
-+
-+ local_irq_restore(flags);
-+
-+ return n;
-+#else
-+
-+ local_irq_save(flags);
-+ n = snprintf(buf, PAGE_SIZE,
-+ "Debug level is too low to dump registers!!!\n");
-+ local_irq_restore(flags);
-+
-+ return n;
-+#endif
-+}
-+
-+#endif
-+
-+static ssize_t show_fm_port_bmi_regs(struct device *dev,
-+ struct device_attribute *attr, char *buf)
-+{
-+ unsigned long flags;
-+ unsigned n = 0;
-+#if (defined(DEBUG_ERRORS) && (DEBUG_ERRORS > 0))
-+ t_LnxWrpFmPortDev *p_LnxWrpFmPortDev;
-+#endif
-+
-+ if (attr == NULL || buf == NULL || dev == NULL)
-+ return -EINVAL;
-+
-+#if (defined(DEBUG_ERRORS) && (DEBUG_ERRORS > 0))
-+ p_LnxWrpFmPortDev =
-+ (t_LnxWrpFmPortDev *) dev_get_drvdata(dev);
-+
-+ local_irq_save(flags);
-+
-+ if (!p_LnxWrpFmPortDev->h_Dev) {
-+ n = snprintf(buf, PAGE_SIZE, "\tFM Port not configured...\n");
-+ return n;
-+ } else {
-+ n = snprintf(buf, PAGE_SIZE,
-+ "FM port driver registers dump.\n");
-+ n = fm_port_dump_regs_bmi(p_LnxWrpFmPortDev->h_Dev, buf, n);
-+ }
-+
-+ local_irq_restore(flags);
-+
-+ return n;
-+#else
-+
-+ local_irq_save(flags);
-+ n = snprintf(buf, PAGE_SIZE,
-+ "Debug level is too low to dump registers!!!\n");
-+ local_irq_restore(flags);
-+
-+ return n;
-+#endif
-+}
-+
-+static ssize_t show_fm_port_qmi_regs(struct device *dev,
-+ struct device_attribute *attr, char *buf)
-+{
-+ unsigned long flags;
-+ unsigned n = 0;
-+#if (defined(DEBUG_ERRORS) && (DEBUG_ERRORS > 0))
-+ t_LnxWrpFmPortDev *p_LnxWrpFmPortDev;
-+#endif
-+
-+ if (attr == NULL || buf == NULL || dev == NULL)
-+ return -EINVAL;
-+
-+#if (defined(DEBUG_ERRORS) && (DEBUG_ERRORS > 0))
-+ p_LnxWrpFmPortDev =
-+ (t_LnxWrpFmPortDev *) dev_get_drvdata(dev);
-+
-+ local_irq_save(flags);
-+
-+ if (!p_LnxWrpFmPortDev->h_Dev) {
-+ n = snprintf(buf, PAGE_SIZE, "\tFM Port not configured...\n");
-+ return n;
-+ } else {
-+ n = snprintf(buf, PAGE_SIZE,
-+ "FM port driver registers dump.\n");
-+ n = fm_port_dump_regs_qmi(p_LnxWrpFmPortDev->h_Dev, buf, n);
-+ }
-+
-+ local_irq_restore(flags);
-+
-+ return n;
-+#else
-+
-+ local_irq_save(flags);
-+ n = snprintf(buf, PAGE_SIZE,
-+ "Debug level is too low to dump registers!!!\n");
-+ local_irq_restore(flags);
-+
-+ return n;
-+#endif
-+}
-+
-+static DEVICE_ATTR(fm_port_regs, S_IRUGO | S_IRUSR, show_fm_port_regs, NULL);
-+static DEVICE_ATTR(fm_port_qmi_regs, S_IRUGO | S_IRUSR, show_fm_port_qmi_regs, NULL);
-+static DEVICE_ATTR(fm_port_bmi_regs, S_IRUGO | S_IRUSR, show_fm_port_bmi_regs, NULL);
-+#if (DPAA_VERSION >= 11)
-+static DEVICE_ATTR(fm_port_ipv4_opt, S_IRUGO | S_IRUSR, show_fm_port_ipv4_options, NULL);
-+#endif
-+static DEVICE_ATTR(fm_port_dsar_regs, S_IRUGO | S_IRUSR, show_fm_port_dsar_regs, NULL);
-+static DEVICE_ATTR(fm_port_dsar_mem, S_IRUGO | S_IRUSR, show_fm_port_dsar_mem, NULL);
-+
-+int fm_port_sysfs_create(struct device *dev)
-+{
-+ t_LnxWrpFmPortDev *p_LnxWrpFmPortDev;
-+
-+ if (dev == NULL)
-+ return -EINVAL;
-+
-+ p_LnxWrpFmPortDev = (t_LnxWrpFmPortDev *) dev_get_drvdata(dev);
-+ if (WARN_ON(p_LnxWrpFmPortDev == NULL))
-+ return -EINVAL;
-+
-+ /* store to remove them when module is disabled */
-+ p_LnxWrpFmPortDev->dev_attr_regs = &dev_attr_fm_port_regs;
-+ p_LnxWrpFmPortDev->dev_attr_qmi_regs = &dev_attr_fm_port_qmi_regs;
-+ p_LnxWrpFmPortDev->dev_attr_bmi_regs = &dev_attr_fm_port_bmi_regs;
-+#if (DPAA_VERSION >= 11)
-+ p_LnxWrpFmPortDev->dev_attr_ipv4_opt = &dev_attr_fm_port_ipv4_opt;
-+#endif
-+ p_LnxWrpFmPortDev->dev_attr_dsar_regs = &dev_attr_fm_port_dsar_regs;
-+ p_LnxWrpFmPortDev->dev_attr_dsar_mem = &dev_attr_fm_port_dsar_mem;
-+ /* Registers dump entry - in future will be moved to debugfs */
-+ if (device_create_file(dev, &dev_attr_fm_port_regs) != 0)
-+ return -EIO;
-+ if (device_create_file(dev, &dev_attr_fm_port_qmi_regs) != 0)
-+ return -EIO;
-+ if (device_create_file(dev, &dev_attr_fm_port_bmi_regs) != 0)
-+ return -EIO;
-+#if (DPAA_VERSION >= 11)
-+ if (device_create_file(dev, &dev_attr_fm_port_ipv4_opt) != 0)
-+ return -EIO;
-+#endif
-+ if (device_create_file(dev, &dev_attr_fm_port_dsar_regs) != 0)
-+ return -EIO;
-+ if (device_create_file(dev, &dev_attr_fm_port_dsar_mem) != 0)
-+ return -EIO;
-+
-+ /* FM Ports statistics */
-+ switch (p_LnxWrpFmPortDev->settings.param.portType) {
-+ case e_FM_PORT_TYPE_TX:
-+ case e_FM_PORT_TYPE_TX_10G:
-+ if (sysfs_create_group
-+ (&dev->kobj, &fm_tx_port_dev_stats_attr_grp) != 0)
-+ return -EIO;
-+ break;
-+ case e_FM_PORT_TYPE_RX:
-+ case e_FM_PORT_TYPE_RX_10G:
-+ if (sysfs_create_group
-+ (&dev->kobj, &fm_rx_port_dev_stats_attr_grp) != 0)
-+ return -EIO;
-+ break;
-+ case e_FM_PORT_TYPE_DUMMY:
-+ case e_FM_PORT_TYPE_OH_OFFLINE_PARSING:
-+ if (sysfs_create_group
-+ (&dev->kobj, &fm_oh_port_dev_stats_attr_grp) != 0)
-+ return -EIO;
-+ break;
-+ default:
-+ WARN(1, "FMD: failure at %s:%d/%s()!\n", __FILE__, __LINE__,
-+ __func__);
-+ return -EINVAL;
-+ break;
-+ };
-+
-+ return 0;
-+}
-+
-+void fm_port_sysfs_destroy(struct device *dev)
-+{
-+ t_LnxWrpFmPortDev *p_LnxWrpFmPortDev = NULL;
-+
-+ /* this function has never been tested !!! */
-+
-+ if (WARN_ON(dev == NULL))
-+ return;
-+
-+ p_LnxWrpFmPortDev = (t_LnxWrpFmPortDev *) dev_get_drvdata(dev);
-+ if (WARN_ON(p_LnxWrpFmPortDev == NULL))
-+ return;
-+
-+ /* The name attribute will be freed also by these 2 functions? */
-+ switch (p_LnxWrpFmPortDev->settings.param.portType) {
-+ case e_FM_PORT_TYPE_TX:
-+ case e_FM_PORT_TYPE_TX_10G:
-+ sysfs_remove_group(&dev->kobj, &fm_tx_port_dev_stats_attr_grp);
-+ break;
-+ case e_FM_PORT_TYPE_RX:
-+ case e_FM_PORT_TYPE_RX_10G:
-+ sysfs_remove_group(&dev->kobj, &fm_rx_port_dev_stats_attr_grp);
-+ break;
-+ case e_FM_PORT_TYPE_DUMMY:
-+ case e_FM_PORT_TYPE_OH_OFFLINE_PARSING:
-+ sysfs_remove_group(&dev->kobj, &fm_oh_port_dev_stats_attr_grp);
-+ break;
-+ default:
-+ WARN(1, "FMD: failure at %s:%d/%s()!\n", __FILE__, __LINE__,
-+ __func__);
-+ break;
-+ };
-+
-+ device_remove_file(dev, p_LnxWrpFmPortDev->dev_attr_regs);
-+ device_remove_file(dev, p_LnxWrpFmPortDev->dev_attr_qmi_regs);
-+ device_remove_file(dev, p_LnxWrpFmPortDev->dev_attr_bmi_regs);
-+#if (DPAA_VERSION >= 11)
-+ device_remove_file(dev, p_LnxWrpFmPortDev->dev_attr_ipv4_opt);
-+#endif
-+ device_remove_file(dev, p_LnxWrpFmPortDev->dev_attr_dsar_regs);
-+ device_remove_file(dev, p_LnxWrpFmPortDev->dev_attr_dsar_mem);
-+}
-+
-+
-+int fm_port_dump_regs(void *h_dev, char *buf, int nn)
-+{
-+ t_FmPort *p_FmPort;
-+ t_Fm *p_Fm;
-+ uint8_t hardwarePortId;
-+ int n = nn;
-+
-+ p_FmPort = (t_FmPort *)h_dev;
-+ hardwarePortId = p_FmPort->hardwarePortId;
-+ p_Fm = (t_Fm *)p_FmPort->h_Fm;
-+
-+ FM_DMP_TITLE(buf, n, &p_Fm->p_FmBmiRegs->fmbm_pp[hardwarePortId - 1],
-+ "fmbm_pp for port %u", hardwarePortId);
-+ FM_DMP_MEM_32(buf, n,
-+ &p_Fm->p_FmBmiRegs->fmbm_pp[hardwarePortId - 1]);
-+
-+ FM_DMP_TITLE(buf, n, &p_Fm->p_FmBmiRegs->fmbm_pfs[hardwarePortId - 1],
-+ "fmbm_pfs for port %u", hardwarePortId);
-+ FM_DMP_MEM_32(buf, n,
-+ &p_Fm->p_FmBmiRegs->fmbm_pfs[hardwarePortId - 1]);
-+
-+ FM_DMP_TITLE(buf, n,
-+ &p_Fm->p_FmBmiRegs->fmbm_spliodn[hardwarePortId - 1],
-+ "fmbm_spliodn for port %u", hardwarePortId);
-+ FM_DMP_MEM_32(buf, n,
-+ &p_Fm->p_FmBmiRegs->fmbm_spliodn[hardwarePortId - 1]);
-+
-+ FM_DMP_TITLE(buf, n, &p_Fm->p_FmFpmRegs->fmfp_ps[hardwarePortId],
-+ "fmfp_psfor port %u", hardwarePortId);
-+ FM_DMP_MEM_32(buf, n, &p_Fm->p_FmFpmRegs->fmfp_ps[hardwarePortId]);
-+
-+ FM_DMP_TITLE(buf, n, &p_Fm->p_FmDmaRegs->fmdmplr[hardwarePortId / 2],
-+ "fmdmplrfor port %u", hardwarePortId);
-+ FM_DMP_MEM_32(buf, n,
-+ &p_Fm->p_FmDmaRegs->fmdmplr[hardwarePortId / 2]);
-+ return n;
-+}
-+
-+#if (DPAA_VERSION >= 11)
-+
-+int fm_port_dump_ipv4_opt(void *h_dev, char *buf, int nn)
-+{
-+ t_FmPort *p_FmPort;
-+ int n = nn;
-+
-+ p_FmPort = (t_FmPort *)h_dev;
-+
-+ FM_DMP_V32(buf, n, p_FmPort->p_ParamsPage, ipfOptionsCounter);
-+
-+ FM_DMP_SUBTITLE(buf, n, "\n");
-+
-+ return n;
-+}
-+#endif
-+
-+int fm_port_dump_regs_bmi(void *h_dev, char *buf, int nn)
-+{
-+ t_FmPort *p_FmPort;
-+ u_FmPortBmiRegs *p_bmi;
-+
-+ char arr[20];
-+ uint8_t flag;
-+ int i = 0;
-+ int n = nn;
-+
-+ p_FmPort = (t_FmPort *)h_dev;
-+ p_bmi = p_FmPort->p_FmPortBmiRegs;
-+
-+ memset(arr, 0, sizeof(arr));
-+ switch (p_FmPort->portType) {
-+ case (e_FM_PORT_TYPE_OH_OFFLINE_PARSING):
-+ strcpy(arr, "OFFLINE-PARSING");
-+ flag = 0;
-+ break;
-+ case (e_FM_PORT_TYPE_OH_HOST_COMMAND):
-+ strcpy(arr, "HOST-COMMAND");
-+ flag = 0;
-+ break;
-+ case (e_FM_PORT_TYPE_RX):
-+ strcpy(arr, "RX");
-+ flag = 1;
-+ break;
-+ case (e_FM_PORT_TYPE_RX_10G):
-+ strcpy(arr, "RX-10G");
-+ flag = 1;
-+ break;
-+ case (e_FM_PORT_TYPE_TX):
-+ strcpy(arr, "TX");
-+ flag = 2;
-+ break;
-+ case (e_FM_PORT_TYPE_TX_10G):
-+ strcpy(arr, "TX-10G");
-+ flag = 2;
-+ break;
-+ default:
-+ return -EINVAL;
-+ }
-+
-+ FM_DMP_TITLE(buf, n, NULL,
-+ "FMan-Port (%s #%d) registers:",
-+ arr, p_FmPort->portId);
-+
-+ FM_DMP_TITLE(buf, n, p_bmi, "Bmi Port Regs");
-+
-+ switch (flag) {
-+ case (0):
-+ FM_DMP_SUBTITLE(buf, n, "\n");
-+ FM_DMP_V32(buf, n, &p_bmi->ohPortBmiRegs, fmbm_ocfg);
-+ FM_DMP_V32(buf, n, &p_bmi->ohPortBmiRegs, fmbm_ost);
-+ FM_DMP_V32(buf, n, &p_bmi->ohPortBmiRegs, fmbm_oda);
-+ FM_DMP_V32(buf, n, &p_bmi->ohPortBmiRegs, fmbm_oicp);
-+ FM_DMP_V32(buf, n, &p_bmi->ohPortBmiRegs, fmbm_ofdne);
-+ FM_DMP_V32(buf, n, &p_bmi->ohPortBmiRegs, fmbm_ofne);
-+ FM_DMP_V32(buf, n, &p_bmi->ohPortBmiRegs, fmbm_ofca);
-+ FM_DMP_V32(buf, n, &p_bmi->ohPortBmiRegs, fmbm_ofpne);
-+ FM_DMP_V32(buf, n, &p_bmi->ohPortBmiRegs, fmbm_opso);
-+ FM_DMP_V32(buf, n, &p_bmi->ohPortBmiRegs, fmbm_opp);
-+ FM_DMP_V32(buf, n, &p_bmi->ohPortBmiRegs, fmbm_occb);
-+ FM_DMP_V32(buf, n, &p_bmi->ohPortBmiRegs, fmbm_oim);
-+ FM_DMP_V32(buf, n, &p_bmi->ohPortBmiRegs, fmbm_ofp);
-+ FM_DMP_V32(buf, n, &p_bmi->ohPortBmiRegs, fmbm_ofed);
-+
-+ FM_DMP_TITLE(buf, n,
-+ &(p_bmi->ohPortBmiRegs.fmbm_oprai), "fmbm_oprai");
-+ for (i = 0; i < FM_PORT_PRS_RESULT_NUM_OF_WORDS; ++i) {
-+ FM_DMP_MEM_32(buf, n,
-+ &(p_bmi->ohPortBmiRegs.fmbm_oprai[i]));
-+ }
-+ FM_DMP_SUBTITLE(buf, n, "\n");
-+ FM_DMP_V32(buf, n, &p_bmi->ohPortBmiRegs, fmbm_ofqid);
-+ FM_DMP_V32(buf, n, &p_bmi->ohPortBmiRegs, fmbm_oefqid);
-+ FM_DMP_V32(buf, n, &p_bmi->ohPortBmiRegs, fmbm_ofsdm);
-+ FM_DMP_V32(buf, n, &p_bmi->ohPortBmiRegs, fmbm_ofsem);
-+ FM_DMP_V32(buf, n, &p_bmi->ohPortBmiRegs, fmbm_ofene);
-+ FM_DMP_V32(buf, n, &p_bmi->ohPortBmiRegs, fmbm_orlmts);
-+ FM_DMP_V32(buf, n, &p_bmi->ohPortBmiRegs, fmbm_orlmt);
-+ FM_DMP_V32(buf, n, &p_bmi->ohPortBmiRegs, fmbm_ocmne);
-+ {
-+#ifndef FM_NO_OP_OBSERVED_POOLS
-+ if (p_FmPort->fmRevInfo.majorRev == 4) {
-+ FM_DMP_TITLE(buf, n,
-+ &p_bmi->ohPortBmiRegs.fmbm_oebmpi,
-+ "fmbm_oebmpi");
-+
-+ for (i = 0; i < FM_PORT_MAX_NUM_OF_OBSERVED_EXT_POOLS; ++i) {
-+ FM_DMP_MEM_32(buf, n,
-+ &(p_bmi->ohPortBmiRegs.fmbm_oebmpi[i]));
-+ }
-+ FM_DMP_V32(buf, n, &p_bmi->ohPortBmiRegs, fmbm_ocgm);
-+ }
-+#endif /* !FM_NO_OP_OBSERVED_POOLS */
-+ }
-+ FM_DMP_V32(buf, n, &p_bmi->ohPortBmiRegs, fmbm_ostc);
-+ FM_DMP_V32(buf, n, &p_bmi->ohPortBmiRegs, fmbm_ofrc);
-+ FM_DMP_V32(buf, n, &p_bmi->ohPortBmiRegs, fmbm_ofdc);
-+ FM_DMP_V32(buf, n, &p_bmi->ohPortBmiRegs, fmbm_ofledc);
-+ FM_DMP_V32(buf, n, &p_bmi->ohPortBmiRegs, fmbm_ofufdc);
-+ FM_DMP_V32(buf, n, &p_bmi->ohPortBmiRegs, fmbm_offc);
-+ FM_DMP_V32(buf, n, &p_bmi->ohPortBmiRegs, fmbm_ofwdc);
-+ FM_DMP_V32(buf, n, &p_bmi->ohPortBmiRegs, fmbm_ofldec);
-+ FM_DMP_V32(buf, n, &p_bmi->ohPortBmiRegs, fmbm_opc);
-+ FM_DMP_V32(buf, n, &p_bmi->ohPortBmiRegs, fmbm_opcp);
-+ FM_DMP_V32(buf, n, &p_bmi->ohPortBmiRegs, fmbm_occn);
-+ FM_DMP_V32(buf, n, &p_bmi->ohPortBmiRegs, fmbm_otuc);
-+ FM_DMP_V32(buf, n, &p_bmi->ohPortBmiRegs, fmbm_oduc);
-+ FM_DMP_V32(buf, n, &p_bmi->ohPortBmiRegs, fmbm_ofuc);
-+ FM_DMP_TITLE(buf, n, &(p_bmi->ohPortBmiRegs.fmbm_odcfg),
-+ "fmbm_odcfg");
-+ for (i = 0; i < 3; ++i) {
-+ FM_DMP_MEM_32(buf, n,
-+ &(p_bmi->ohPortBmiRegs.fmbm_odcfg[i]));
-+ }
-+ FM_DMP_SUBTITLE(buf, n, "\n");
-+
-+ FM_DMP_V32(buf, n, &p_bmi->ohPortBmiRegs, fmbm_ogpr);
-+ break;
-+ case (1):
-+ FM_DMP_SUBTITLE(buf, n, "\n");
-+ FM_DMP_V32(buf, n, &p_bmi->rxPortBmiRegs, fmbm_rcfg);
-+ FM_DMP_V32(buf, n, &p_bmi->rxPortBmiRegs, fmbm_rst);
-+ FM_DMP_V32(buf, n, &p_bmi->rxPortBmiRegs, fmbm_rda);
-+ FM_DMP_V32(buf, n, &p_bmi->rxPortBmiRegs, fmbm_rfp);
-+ FM_DMP_V32(buf, n, &p_bmi->rxPortBmiRegs, fmbm_reth);
-+ FM_DMP_V32(buf, n, &p_bmi->rxPortBmiRegs, fmbm_rfed);
-+ FM_DMP_V32(buf, n, &p_bmi->rxPortBmiRegs, fmbm_ricp);
-+ FM_DMP_V32(buf, n, &p_bmi->rxPortBmiRegs, fmbm_rebm);
-+ FM_DMP_V32(buf, n, &p_bmi->rxPortBmiRegs, fmbm_rfne);
-+ FM_DMP_V32(buf, n, &p_bmi->rxPortBmiRegs, fmbm_rfca);
-+ FM_DMP_V32(buf, n, &p_bmi->rxPortBmiRegs, fmbm_rfpne);
-+ FM_DMP_V32(buf, n, &p_bmi->rxPortBmiRegs, fmbm_rpso);
-+ FM_DMP_V32(buf, n, &p_bmi->rxPortBmiRegs, fmbm_rpp);
-+ FM_DMP_TITLE(buf, n, &(p_bmi->rxPortBmiRegs.fmbm_rprai),
-+ "fmbm_rprai");
-+ for (i = 0; i < FM_PORT_PRS_RESULT_NUM_OF_WORDS; ++i) {
-+ FM_DMP_MEM_32(buf, n,
-+ &(p_bmi->rxPortBmiRegs.fmbm_rprai[i]));
-+ }
-+ FM_DMP_SUBTITLE(buf, n, "\n");
-+ FM_DMP_V32(buf, n, &p_bmi->rxPortBmiRegs, fmbm_rfqid);
-+ FM_DMP_V32(buf, n, &p_bmi->rxPortBmiRegs, fmbm_refqid);
-+ FM_DMP_V32(buf, n, &p_bmi->rxPortBmiRegs, fmbm_rfsdm);
-+ FM_DMP_V32(buf, n, &p_bmi->rxPortBmiRegs, fmbm_rfsem);
-+ FM_DMP_V32(buf, n, &p_bmi->rxPortBmiRegs, fmbm_rfene);
-+ FM_DMP_V32(buf, n, &p_bmi->rxPortBmiRegs, fmbm_rcmne);
-+ FM_DMP_TITLE(buf, n, &p_bmi->rxPortBmiRegs.fmbm_ebmpi,
-+ "fmbm_ebmpi");
-+ for (i = 0; i < FM_PORT_MAX_NUM_OF_EXT_POOLS; ++i) {
-+ FM_DMP_MEM_32(buf, n,
-+ &(p_bmi->rxPortBmiRegs.fmbm_ebmpi[i]));
-+ }
-+ FM_DMP_TITLE(buf, n, &p_bmi->rxPortBmiRegs.fmbm_acnt,
-+ "fmbm_acnt");
-+ for (i = 0; i < FM_PORT_MAX_NUM_OF_EXT_POOLS; ++i) {
-+ FM_DMP_MEM_32(buf, n,
-+ &(p_bmi->rxPortBmiRegs.fmbm_acnt[i]));
-+ }
-+ FM_DMP_TITLE(buf, n, &p_bmi->rxPortBmiRegs.fmbm_rcgm,
-+ "fmbm_rcgm");
-+ for (i = 0; i < FM_PORT_NUM_OF_CONGESTION_GRPS / 32; ++i) {
-+ FM_DMP_MEM_32(buf, n,
-+ &(p_bmi->rxPortBmiRegs.fmbm_rcgm[i]));
-+ }
-+
-+ FM_DMP_SUBTITLE(buf, n, "\n");
-+ FM_DMP_V32(buf, n, &p_bmi->rxPortBmiRegs, fmbm_rmpd);
-+ FM_DMP_V32(buf, n, &p_bmi->rxPortBmiRegs, fmbm_rstc);
-+ FM_DMP_V32(buf, n, &p_bmi->rxPortBmiRegs, fmbm_rfrc);
-+ FM_DMP_V32(buf, n, &p_bmi->rxPortBmiRegs, fmbm_rfbc);
-+ FM_DMP_V32(buf, n, &p_bmi->rxPortBmiRegs, fmbm_rlfc);
-+ FM_DMP_V32(buf, n, &p_bmi->rxPortBmiRegs, fmbm_rffc);
-+ FM_DMP_V32(buf, n, &p_bmi->rxPortBmiRegs, fmbm_rfcd);
-+ FM_DMP_V32(buf, n, &p_bmi->rxPortBmiRegs, fmbm_rfldec);
-+ FM_DMP_V32(buf, n, &p_bmi->rxPortBmiRegs, fmbm_rodc);
-+ FM_DMP_V32(buf, n, &p_bmi->rxPortBmiRegs, fmbm_rpc);
-+ FM_DMP_V32(buf, n, &p_bmi->rxPortBmiRegs, fmbm_rpcp);
-+ FM_DMP_V32(buf, n, &p_bmi->rxPortBmiRegs, fmbm_rccn);
-+ FM_DMP_V32(buf, n, &p_bmi->rxPortBmiRegs, fmbm_rtuc);
-+ FM_DMP_V32(buf, n, &p_bmi->rxPortBmiRegs, fmbm_rrquc);
-+ FM_DMP_V32(buf, n, &p_bmi->rxPortBmiRegs, fmbm_rduc);
-+ FM_DMP_V32(buf, n, &p_bmi->rxPortBmiRegs, fmbm_rfuc);
-+ FM_DMP_V32(buf, n, &p_bmi->rxPortBmiRegs, fmbm_rpac);
-+ FM_DMP_TITLE(buf, n, &(p_bmi->rxPortBmiRegs.fmbm_rdcfg),
-+ "fmbm_rdcfg");
-+ for (i = 0; i < 3; ++i) {
-+ FM_DMP_MEM_32(buf, n,
-+ &(p_bmi->rxPortBmiRegs.fmbm_rdcfg[i]));
-+ }
-+ FM_DMP_SUBTITLE(buf, n, "\n");
-+ FM_DMP_V32(buf, n, &p_bmi->rxPortBmiRegs, fmbm_rgpr);
-+ break;
-+ case (2):
-+ FM_DMP_SUBTITLE(buf, n, "\n");
-+ FM_DMP_V32(buf, n, &p_bmi->txPortBmiRegs, fmbm_tcfg);
-+ FM_DMP_V32(buf, n, &p_bmi->txPortBmiRegs, fmbm_tst);
-+ FM_DMP_V32(buf, n, &p_bmi->txPortBmiRegs, fmbm_tda);
-+ FM_DMP_V32(buf, n, &p_bmi->txPortBmiRegs, fmbm_tfp);
-+ FM_DMP_V32(buf, n, &p_bmi->txPortBmiRegs, fmbm_tfed);
-+ FM_DMP_V32(buf, n, &p_bmi->txPortBmiRegs, fmbm_ticp);
-+ FM_DMP_V32(buf, n, &p_bmi->txPortBmiRegs, fmbm_tfdne);
-+ FM_DMP_V32(buf, n, &p_bmi->txPortBmiRegs, fmbm_tfca);
-+ FM_DMP_V32(buf, n, &p_bmi->txPortBmiRegs, fmbm_tcfqid);
-+ FM_DMP_V32(buf, n, &p_bmi->txPortBmiRegs, fmbm_tfeqid);
-+ FM_DMP_V32(buf, n, &p_bmi->txPortBmiRegs, fmbm_tfene);
-+#if (DPAA_VERSION >= 11)
-+ FM_DMP_V32(buf, n, &p_bmi->txPortBmiRegs, fmbm_tfne);
-+ FM_DMP_V32(buf, n, &p_bmi->txPortBmiRegs, fmbm_tcmne);
-+#endif /* (DPAA_VERSION >= 11) */
-+ FM_DMP_V32(buf, n, &p_bmi->txPortBmiRegs, fmbm_trlmts);
-+ FM_DMP_V32(buf, n, &p_bmi->txPortBmiRegs, fmbm_trlmt);
-+ FM_DMP_V32(buf, n, &p_bmi->txPortBmiRegs, fmbm_tstc);
-+ FM_DMP_V32(buf, n, &p_bmi->txPortBmiRegs, fmbm_tfrc);
-+ FM_DMP_V32(buf, n, &p_bmi->txPortBmiRegs, fmbm_tfdc);
-+ FM_DMP_V32(buf, n, &p_bmi->txPortBmiRegs, fmbm_tfledc);
-+ FM_DMP_V32(buf, n, &p_bmi->txPortBmiRegs, fmbm_tfufdc);
-+ FM_DMP_V32(buf, n, &p_bmi->txPortBmiRegs, fmbm_tpc);
-+ FM_DMP_V32(buf, n, &p_bmi->txPortBmiRegs, fmbm_tpcp);
-+ FM_DMP_V32(buf, n, &p_bmi->txPortBmiRegs, fmbm_tccn);
-+ FM_DMP_V32(buf, n, &p_bmi->txPortBmiRegs, fmbm_ttuc);
-+ FM_DMP_V32(buf, n, &p_bmi->txPortBmiRegs, fmbm_ttcquc);
-+ FM_DMP_V32(buf, n, &p_bmi->txPortBmiRegs, fmbm_tduc);
-+ FM_DMP_V32(buf, n, &p_bmi->txPortBmiRegs, fmbm_tfuc);
-+ FM_DMP_TITLE(buf, n, &(p_bmi->txPortBmiRegs.fmbm_tdcfg),
-+ "fmbm_tdcfg");
-+ for (i = 0; i < 3 ; ++i) {
-+ FM_DMP_MEM_32(buf, n,
-+ &(p_bmi->txPortBmiRegs.fmbm_tdcfg[i]));
-+ }
-+ FM_DMP_SUBTITLE(buf, n, "\n");
-+ FM_DMP_V32(buf, n, &p_bmi->txPortBmiRegs, fmbm_tgpr);
-+ break;
-+ }
-+
-+ FM_DMP_SUBTITLE(buf, n, "\n");
-+
-+ return n;
-+}
-+
-+int fm_port_dump_regs_qmi(void *h_dev, char *buf, int nn)
-+{
-+ t_FmPort *p_FmPort;
-+ int n = nn;
-+
-+ p_FmPort = (t_FmPort *)h_dev;
-+
-+ FM_DMP_TITLE(buf, n, p_FmPort->p_FmPortQmiRegs, "Qmi Port Regs");
-+
-+ FM_DMP_V32(buf, n, p_FmPort->p_FmPortQmiRegs, fmqm_pnc);
-+ FM_DMP_V32(buf, n, p_FmPort->p_FmPortQmiRegs, fmqm_pns);
-+ FM_DMP_V32(buf, n, p_FmPort->p_FmPortQmiRegs, fmqm_pnts);
-+ FM_DMP_V32(buf, n, p_FmPort->p_FmPortQmiRegs, fmqm_pnen);
-+ FM_DMP_V32(buf, n, p_FmPort->p_FmPortQmiRegs, fmqm_pnetfc);
-+ FM_DMP_V32(buf, n,
-+ &p_FmPort->p_FmPortQmiRegs->nonRxQmiRegs, fmqm_pndn);
-+ FM_DMP_V32(buf, n,
-+ &p_FmPort->p_FmPortQmiRegs->nonRxQmiRegs, fmqm_pndc);
-+ FM_DMP_V32(buf, n,
-+ &p_FmPort->p_FmPortQmiRegs->nonRxQmiRegs, fmqm_pndtfc);
-+ FM_DMP_V32(buf, n,
-+ &p_FmPort->p_FmPortQmiRegs->nonRxQmiRegs, fmqm_pndfdc);
-+ FM_DMP_V32(buf, n,
-+ &p_FmPort->p_FmPortQmiRegs->nonRxQmiRegs, fmqm_pndcc);
-+
-+ FM_DMP_SUBTITLE(buf, n, "\n");
-+
-+ return n;
-+}
-+
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/src/wrapper/lnxwrp_sysfs_fm_port.h
-@@ -0,0 +1,56 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+/*
-+ @File lnxwrp_sysfs_fm_port.h
-+
-+ @Description FM port sysfs functions.
-+
-+*/
-+
-+#ifndef LNXWRP_SYSFS_FM_PORT_H_
-+#define LNXWRP_SYSFS_FM_PORT_H_
-+
-+#include "lnxwrp_sysfs.h"
-+
-+int fm_port_sysfs_create(struct device *dev);
-+void fm_port_sysfs_destroy(struct device *dev);
-+
-+int fm_port_dump_regs(void *h_dev, char *buf, int n);
-+int fm_port_dump_regs_bmi(void *h_dev, char *buf, int n);
-+int fm_port_dump_regs_qmi(void *h_dev, char *buf, int n);
-+
-+#if (DPAA_VERSION >= 11)
-+int fm_port_dump_ipv4_opt(void *h_dev, char *buf, int n);
-+#endif
-+
-+#endif /* LNXWRP_SYSFS_FM_PORT_H_ */
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/src/xx/Makefile
-@@ -0,0 +1,18 @@
-+#
-+# Makefile for the Freescale Ethernet controllers
-+#
-+ccflags-y += -DVERSION=\"\"
-+#
-+#Include netcomm SW specific definitions
-+include $(srctree)/drivers/net/ethernet/freescale/sdk_fman/ncsw_config.mk
-+
-+obj-y += fsl-ncsw-xx.o
-+
-+ifneq ($(CONFIG_FMAN_ARM),y)
-+fsl-ncsw-xx-objs := xx_linux.o \
-+ module_strings.o
-+else
-+fsl-ncsw-xx-objs := xx_arm_linux.o \
-+ module_strings.o
-+endif
-+
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/src/xx/module_strings.c
-@@ -0,0 +1,46 @@
-+/*
-+ * Copyright 2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+/* Module names for debug messages */
-+const char *moduleStrings[] =
-+{
-+ "", /* MODULE_UNKNOWN */
-+ "FM", /* MODULE_FM */
-+ "FM-MURAM", /* MODULE_FM_MURAM */
-+ "FM-PCD", /* MODULE_FM_PCD */
-+ "FM-RTC", /* MODULE_FM_RTC */
-+ "FM-MAC", /* MODULE_FM_MAC */
-+ "FM-Port", /* MODULE_FM_PORT */
-+ "MM", /* MODULE_MM */
-+ "FM-SP", /* MODULE_FM_SP */
-+ "FM-MACSEC" /* MODULE_FM_MACSEC */
-+};
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/src/xx/xx_arm_linux.c
-@@ -0,0 +1,905 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+/**************************************************************************//**
-+ @File xx_arm_linux.c
-+
-+ @Description XX routines implementation for Linux.
-+*//***************************************************************************/
-+#include <linux/version.h>
-+
-+#if defined(CONFIG_MODVERSIONS) && !defined(MODVERSIONS)
-+#define MODVERSIONS
-+#endif
-+#ifdef MODVERSIONS
-+#include <config/modversions.h>
-+#endif /* MODVERSIONS */
-+
-+#include <linux/module.h>
-+#include <linux/kernel.h>
-+#include <linux/sched.h>
-+#include <linux/string.h>
-+#include <linux/ptrace.h>
-+#include <linux/errno.h>
-+#include <linux/ioport.h>
-+#include <linux/slab.h>
-+#include <linux/interrupt.h>
-+#include <linux/fs.h>
-+#include <linux/vmalloc.h>
-+#include <linux/init.h>
-+#include <linux/timer.h>
-+#include <linux/spinlock.h>
-+#include <linux/delay.h>
-+#include <linux/proc_fs.h>
-+#include <linux/smp.h>
-+#include <linux/of.h>
-+#include <linux/irqdomain.h>
-+
-+#include <linux/workqueue.h>
-+
-+#ifdef BIGPHYSAREA_ENABLE
-+#include <linux/bigphysarea.h>
-+#endif /* BIGPHYSAREA_ENABLE */
-+
-+//#include <sysdev/fsl_soc.h>
-+#include <asm/pgtable.h>
-+#include <asm/irq.h>
-+#include <asm/bitops.h>
-+#include <asm/uaccess.h>
-+#include <asm/io.h>
-+#include <asm/atomic.h>
-+#include <asm/string.h>
-+#include <asm/byteorder.h>
-+#include <asm/page.h>
-+
-+#include "error_ext.h"
-+#include "std_ext.h"
-+#include "list_ext.h"
-+#include "mm_ext.h"
-+#include "sys_io_ext.h"
-+#include "xx.h"
-+
-+
-+#define __ERR_MODULE__ MODULE_UNKNOWN
-+
-+#ifdef BIGPHYSAREA_ENABLE
-+#define MAX_ALLOCATION_SIZE 128 * 1024 /* Maximum size allocated with kmalloc is 128K */
-+
-+
-+/* TODO: large allocations => use big phys area */
-+/******************************************************************************
-+ * routine: get_nr_pages
-+ *
-+ * description:
-+ * calculates the number of memory pages for a given size (in bytes)
-+ *
-+ * arguments:
-+ * size - the number of bytes
-+ *
-+ * return code:
-+ * The number of pages
-+ *
-+ *****************************************************************************/
-+static __inline__ uint32_t get_nr_pages (uint32_t size)
-+{
-+ return (uint32_t)((size >> PAGE_SHIFT) + (size & PAGE_SHIFT ? 1 : 0));
-+}
-+
-+static bool in_big_phys_area (uint32_t addr)
-+{
-+ uint32_t base, size;
-+
-+ bigphysarea_get_details (&base, &size);
-+ return ((addr >= base) && (addr < base + size));
-+}
-+#endif /* BIGPHYSAREA_ENABLE */
-+
-+void * xx_Malloc(uint32_t n)
-+{
-+ void *a;
-+ uint32_t flags;
-+
-+ flags = XX_DisableAllIntr();
-+#ifdef BIGPHYSAREA_ENABLE
-+ if (n >= MAX_ALLOCATION_SIZE)
-+ a = (void*)bigphysarea_alloc_pages(get_nr_pages(n), 0, GFP_ATOMIC);
-+ else
-+#endif /* BIGPHYSAREA_ENABLE */
-+ a = (void *)kmalloc((uint32_t)n, GFP_ATOMIC);
-+ if (!a)
-+ XX_Print("No memory for XX_Malloc\n");
-+ XX_RestoreAllIntr(flags);
-+
-+ return a;
-+}
-+
-+void xx_Free(void *p)
-+{
-+#ifdef BIGPHYSAREA_ENABLE
-+ if (in_big_phys_area ((uint32_t)p))
-+ bigphysarea_free_pages(p);
-+ else
-+#endif /* BIGPHYSAREA_ENABLE */
-+ kfree(p);
-+}
-+
-+void XX_Exit(int status)
-+{
-+ WARN(1, "\n\nFMD: fatal error, driver can't go on!!!\n\n");
-+}
-+
-+#define BUF_SIZE 512
-+void XX_Print(char *str, ...)
-+{
-+ va_list args;
-+#ifdef CONFIG_SMP
-+ char buf[BUF_SIZE];
-+#endif /* CONFIG_SMP */
-+
-+ va_start(args, str);
-+#ifdef CONFIG_SMP
-+ if (vsnprintf (buf, BUF_SIZE, str, args) >= BUF_SIZE)
-+ printk(KERN_WARNING "Illegal string to print!\n more than %d characters.\n\tString was not printed completelly.\n", BUF_SIZE);
-+ printk(KERN_CRIT "cpu %d: %s", raw_smp_processor_id(), buf);
-+#else
-+ vprintk(str, args);
-+#endif /* CONFIG_SMP */
-+ va_end(args);
-+}
-+
-+void XX_Fprint(void *file, char *str, ...)
-+{
-+ va_list args;
-+#ifdef CONFIG_SMP
-+ char buf[BUF_SIZE];
-+#endif /* CONFIG_SMP */
-+
-+ va_start(args, str);
-+#ifdef CONFIG_SMP
-+ if (vsnprintf (buf, BUF_SIZE, str, args) >= BUF_SIZE)
-+ printk(KERN_WARNING "Illegal string to print!\n more than %d characters.\n\tString was not printed completelly.\n", BUF_SIZE);
-+ printk (KERN_CRIT "cpu %d: %s", smp_processor_id(), buf);
-+
-+#else
-+ vprintk(str, args);
-+#endif /* CONFIG_SMP */
-+ va_end(args);
-+}
-+
-+#ifdef DEBUG_XX_MALLOC
-+typedef void (*t_ffn)(void *);
-+typedef struct {
-+ t_ffn f_free;
-+ void *mem;
-+ char *fname;
-+ int fline;
-+ uint32_t size;
-+ t_List node;
-+} t_MemDebug;
-+#define MEMDBG_OBJECT(p_List) LIST_OBJECT(p_List, t_MemDebug, node)
-+
-+LIST(memDbgLst);
-+
-+
-+void * XX_MallocDebug(uint32_t size, char *fname, int line)
-+{
-+ void *mem;
-+ t_MemDebug *p_MemDbg;
-+
-+ p_MemDbg = (t_MemDebug *)xx_Malloc(sizeof(t_MemDebug));
-+ if (p_MemDbg == NULL)
-+ return NULL;
-+
-+ mem = xx_Malloc(size);
-+ if (mem == NULL)
-+ {
-+ XX_Free(p_MemDbg);
-+ return NULL;
-+ }
-+
-+ INIT_LIST(&p_MemDbg->node);
-+ p_MemDbg->f_free = xx_Free;
-+ p_MemDbg->mem = mem;
-+ p_MemDbg->fname = fname;
-+ p_MemDbg->fline = line;
-+ p_MemDbg->size = size+sizeof(t_MemDebug);
-+ LIST_AddToTail(&p_MemDbg->node, &memDbgLst);
-+
-+ return mem;
-+}
-+
-+void * XX_MallocSmartDebug(uint32_t size,
-+ int memPartitionId,
-+ uint32_t align,
-+ char *fname,
-+ int line)
-+{
-+ void *mem;
-+ t_MemDebug *p_MemDbg;
-+
-+ p_MemDbg = (t_MemDebug *)XX_Malloc(sizeof(t_MemDebug));
-+ if (p_MemDbg == NULL)
-+ return NULL;
-+
-+ mem = xx_MallocSmart((uint32_t)size, memPartitionId, align);
-+ if (mem == NULL)
-+ {
-+ XX_Free(p_MemDbg);
-+ return NULL;
-+ }
-+
-+ INIT_LIST(&p_MemDbg->node);
-+ p_MemDbg->f_free = xx_FreeSmart;
-+ p_MemDbg->mem = mem;
-+ p_MemDbg->fname = fname;
-+ p_MemDbg->fline = line;
-+ p_MemDbg->size = size+sizeof(t_MemDebug);
-+ LIST_AddToTail(&p_MemDbg->node, &memDbgLst);
-+
-+ return mem;
-+}
-+
-+static void debug_free(void *mem)
-+{
-+ t_List *p_MemDbgLh = NULL;
-+ t_MemDebug *p_MemDbg;
-+ bool found = FALSE;
-+
-+ if (LIST_IsEmpty(&memDbgLst))
-+ {
-+ REPORT_ERROR(MAJOR, E_ALREADY_FREE, ("Unbalanced free (0x%08x)", mem));
-+ return;
-+ }
-+
-+ LIST_FOR_EACH(p_MemDbgLh, &memDbgLst)
-+ {
-+ p_MemDbg = MEMDBG_OBJECT(p_MemDbgLh);
-+ if (p_MemDbg->mem == mem)
-+ {
-+ found = TRUE;
-+ break;
-+ }
-+ }
-+
-+ if (!found)
-+ {
-+ REPORT_ERROR(MAJOR, E_NOT_FOUND,
-+ ("Attempt to free unallocated address (0x%08x)",mem));
-+ dump_stack();
-+ return;
-+ }
-+
-+ LIST_Del(p_MemDbgLh);
-+ p_MemDbg->f_free(mem);
-+ p_MemDbg->f_free(p_MemDbg);
-+}
-+
-+void XX_FreeSmart(void *p)
-+{
-+ debug_free(p);
-+}
-+
-+
-+void XX_Free(void *p)
-+{
-+ debug_free(p);
-+}
-+
-+#else /* not DEBUG_XX_MALLOC */
-+void * XX_Malloc(uint32_t size)
-+{
-+ return xx_Malloc(size);
-+}
-+
-+void * XX_MallocSmart(uint32_t size, int memPartitionId, uint32_t alignment)
-+{
-+ return xx_MallocSmart(size,memPartitionId, alignment);
-+}
-+
-+void XX_FreeSmart(void *p)
-+{
-+ xx_FreeSmart(p);
-+}
-+
-+
-+void XX_Free(void *p)
-+{
-+ xx_Free(p);
-+}
-+#endif /* not DEBUG_XX_MALLOC */
-+
-+
-+#if (defined(REPORT_EVENTS) && (REPORT_EVENTS > 0))
-+void XX_EventById(uint32_t event, t_Handle appId, uint16_t flags, char *msg)
-+{
-+ e_Event eventCode = (e_Event)event;
-+
-+ UNUSED(eventCode);
-+ UNUSED(appId);
-+ UNUSED(flags);
-+ UNUSED(msg);
-+}
-+#endif /* (defined(REPORT_EVENTS) && ... */
-+
-+
-+uint32_t XX_DisableAllIntr(void)
-+{
-+ unsigned long flags;
-+
-+#ifdef local_irq_save_nort
-+ local_irq_save_nort(flags);
-+#else
-+ local_irq_save(flags);
-+#endif
-+
-+ return (uint32_t)flags;
-+}
-+
-+void XX_RestoreAllIntr(uint32_t flags)
-+{
-+#ifdef local_irq_restore_nort
-+ local_irq_restore_nort((unsigned long)flags);
-+#else
-+ local_irq_restore((unsigned long)flags);
-+#endif
-+}
-+
-+t_Error XX_Call( uint32_t qid, t_Error (* f)(t_Handle), t_Handle id, t_Handle appId, uint16_t flags )
-+{
-+ UNUSED(qid);
-+ UNUSED(appId);
-+ UNUSED(flags);
-+
-+ return f(id);
-+}
-+
-+int XX_IsICacheEnable(void)
-+{
-+ return TRUE;
-+}
-+
-+int XX_IsDCacheEnable(void)
-+{
-+ return TRUE;
-+}
-+
-+
-+typedef struct {
-+ t_Isr *f_Isr;
-+ t_Handle handle;
-+} t_InterruptHandler;
-+
-+
-+t_Handle interruptHandlers[0x00010000];
-+
-+static irqreturn_t LinuxInterruptHandler (int irq, void *dev_id)
-+{
-+ t_InterruptHandler *p_IntrHndl = (t_InterruptHandler *)dev_id;
-+ p_IntrHndl->f_Isr(p_IntrHndl->handle);
-+ return IRQ_HANDLED;
-+}
-+
-+t_Error XX_SetIntr(int irq, t_Isr *f_Isr, t_Handle handle)
-+{
-+ const char *device;
-+ t_InterruptHandler *p_IntrHndl;
-+
-+ device = GetDeviceName(irq);
-+ if (device == NULL)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("Interrupt source - %d", irq));
-+
-+ p_IntrHndl = (t_InterruptHandler *)XX_Malloc(sizeof(t_InterruptHandler));
-+ if (p_IntrHndl == NULL)
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, NO_MSG);
-+ p_IntrHndl->f_Isr = f_Isr;
-+ p_IntrHndl->handle = handle;
-+ interruptHandlers[irq] = p_IntrHndl;
-+
-+ if (request_irq(GetDeviceIrqNum(irq), LinuxInterruptHandler, 0, device, p_IntrHndl) < 0)
-+ RETURN_ERROR(MAJOR, E_BUSY, ("Can't get IRQ %s\n", device));
-+ disable_irq(GetDeviceIrqNum(irq));
-+
-+ return E_OK;
-+}
-+
-+t_Error XX_FreeIntr(int irq)
-+{
-+ t_InterruptHandler *p_IntrHndl = interruptHandlers[irq];
-+ free_irq(GetDeviceIrqNum(irq), p_IntrHndl);
-+ XX_Free(p_IntrHndl);
-+ interruptHandlers[irq] = 0;
-+ return E_OK;
-+}
-+
-+t_Error XX_EnableIntr(int irq)
-+{
-+ enable_irq(GetDeviceIrqNum(irq));
-+ return E_OK;
-+}
-+
-+t_Error XX_DisableIntr(int irq)
-+{
-+ disable_irq(GetDeviceIrqNum(irq));
-+ return E_OK;
-+}
-+
-+
-+/*****************************************************************************/
-+/* Tasklet Service Routines */
-+/*****************************************************************************/
-+#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,20)
-+typedef struct
-+{
-+ t_Handle h_Data;
-+ void (*f_Callback) (void *);
-+ struct delayed_work dwork;
-+} t_Tasklet;
-+
-+static void GenericTaskletCallback(struct work_struct *p_Work)
-+{
-+ t_Tasklet *p_Task = container_of(p_Work, t_Tasklet, dwork.work);
-+
-+ p_Task->f_Callback(p_Task->h_Data);
-+}
-+#endif /* LINUX_VERSION_CODE */
-+
-+
-+t_TaskletHandle XX_InitTasklet (void (*routine)(void *), void *data)
-+{
-+#if LINUX_VERSION_CODE < KERNEL_VERSION(2,6,20)
-+ struct work_struct *p_Task;
-+ p_Task = (struct work_struct *)XX_Malloc(sizeof(struct work_struct));
-+ INIT_WORK(p_Task, routine, data);
-+#else
-+ t_Tasklet *p_Task = (t_Tasklet *)XX_Malloc(sizeof(t_Tasklet));
-+ p_Task->h_Data = data;
-+ p_Task->f_Callback = routine;
-+ INIT_DELAYED_WORK(&p_Task->dwork, GenericTaskletCallback);
-+#endif /* LINUX_VERSION_CODE */
-+
-+ return (t_TaskletHandle)p_Task;
-+}
-+
-+
-+void XX_FreeTasklet (t_TaskletHandle h_Tasklet)
-+{
-+ if (h_Tasklet)
-+ XX_Free(h_Tasklet);
-+}
-+
-+int XX_ScheduleTask(t_TaskletHandle h_Tasklet, int immediate)
-+{
-+ int ans;
-+
-+#if LINUX_VERSION_CODE < KERNEL_VERSION(2,6,20)
-+ if (immediate)
-+ ans = schedule_work(h_Tasklet);
-+ else
-+ ans = schedule_delayed_work(h_Tasklet, 1);
-+#else
-+ if (immediate)
-+ ans = schedule_delayed_work(&((t_Tasklet *)h_Tasklet)->dwork, 0);
-+ else
-+ ans = schedule_delayed_work(&((t_Tasklet *)h_Tasklet)->dwork, HZ);
-+#endif /* LINUX_VERSION_CODE */
-+
-+ return ans;
-+}
-+
-+void XX_FlushScheduledTasks(void)
-+{
-+#if LINUX_VERSION_CODE < KERNEL_VERSION(2,6,0)
-+ flush_scheduled_tasks();
-+#else
-+ flush_scheduled_work();
-+#endif /* LINUX_VERSION_CODE */
-+}
-+
-+int XX_TaskletIsQueued(t_TaskletHandle h_Tasklet)
-+{
-+#if LINUX_VERSION_CODE < KERNEL_VERSION(2,6,20)
-+ return (int)(((struct work_struct *)h_Tasklet)->pending);
-+#else
-+ return (int)delayed_work_pending(&((t_Tasklet *)h_Tasklet)->dwork);
-+#endif /* LINUX_VERSION_CODE */
-+}
-+
-+void XX_SetTaskletData(t_TaskletHandle h_Tasklet, t_Handle data)
-+{
-+#if LINUX_VERSION_CODE < KERNEL_VERSION(2,6,0)
-+ ((struct tq_struct *)h_Tasklet)->data = data;
-+#elif LINUX_VERSION_CODE < KERNEL_VERSION(2,6,20)
-+ ((struct work_struct *)h_Tasklet)->data = data;
-+#else
-+ ((t_Tasklet *)h_Tasklet)->h_Data = data;
-+#endif /* LINUX_VERSION_CODE */
-+}
-+
-+t_Handle XX_GetTaskletData(t_TaskletHandle h_Tasklet)
-+{
-+#if LINUX_VERSION_CODE < KERNEL_VERSION(2,6,20)
-+ return (t_Handle)(((struct work_struct *)h_Tasklet)->data);
-+#else
-+ return ((t_Tasklet *)h_Tasklet)->h_Data;
-+#endif /* LINUX_VERSION_CODE */
-+}
-+
-+
-+/*****************************************************************************/
-+/* Spinlock Service Routines */
-+/*****************************************************************************/
-+
-+t_Handle XX_InitSpinlock(void)
-+{
-+ spinlock_t *p_Spinlock = (spinlock_t *)XX_Malloc(sizeof(spinlock_t));
-+ if (!p_Spinlock)
-+ return NULL;
-+
-+ spin_lock_init(p_Spinlock);
-+
-+ return (t_Handle)p_Spinlock;
-+}
-+
-+void XX_FreeSpinlock(t_Handle h_Spinlock)
-+{
-+ if (h_Spinlock)
-+ XX_Free(h_Spinlock);
-+}
-+
-+void XX_LockSpinlock(t_Handle h_Spinlock)
-+{
-+ spin_lock((spinlock_t *)h_Spinlock);
-+}
-+
-+void XX_UnlockSpinlock(t_Handle h_Spinlock)
-+{
-+ spin_unlock((spinlock_t *)h_Spinlock);
-+}
-+
-+uint32_t XX_LockIntrSpinlock(t_Handle h_Spinlock)
-+{
-+ unsigned long intrFlags;
-+ spin_lock_irqsave((spinlock_t *)h_Spinlock, intrFlags);
-+ return intrFlags;
-+}
-+
-+void XX_UnlockIntrSpinlock(t_Handle h_Spinlock, uint32_t intrFlags)
-+{
-+ spin_unlock_irqrestore((spinlock_t *)h_Spinlock, (unsigned long)intrFlags);
-+}
-+
-+
-+/*****************************************************************************/
-+/* Timers Service Routines */
-+/*****************************************************************************/
-+/* The time now is in mili sec. resolution */
-+uint32_t XX_CurrentTime(void)
-+{
-+ return (jiffies*1000)/HZ;
-+}
-+
-+
-+t_Handle XX_CreateTimer(void)
-+{
-+ struct timer_list *p_Timer = (struct timer_list *)XX_Malloc(sizeof(struct timer_list));
-+ if (p_Timer)
-+ {
-+ memset(p_Timer, 0, sizeof(struct timer_list));
-+ init_timer(p_Timer);
-+ }
-+ return (t_Handle)p_Timer;
-+}
-+
-+void XX_FreeTimer(t_Handle h_Timer)
-+{
-+ if (h_Timer)
-+ XX_Free(h_Timer);
-+}
-+
-+void XX_StartTimer(t_Handle h_Timer,
-+ uint32_t msecs,
-+ bool periodic,
-+ void (*f_TimerExpired)(t_Handle),
-+ t_Handle h_Arg)
-+{
-+ int tmp_jiffies = (msecs*HZ)/1000;
-+ struct timer_list *p_Timer = (struct timer_list *)h_Timer;
-+
-+ SANITY_CHECK_RETURN((periodic == FALSE), E_NOT_SUPPORTED);
-+
-+ p_Timer->function = (void (*)(unsigned long))f_TimerExpired;
-+ p_Timer->data = (unsigned long)h_Arg;
-+ if ((msecs*HZ)%1000)
-+ tmp_jiffies++;
-+ p_Timer->expires = (jiffies + tmp_jiffies);
-+
-+ add_timer((struct timer_list *)h_Timer);
-+}
-+
-+void XX_SetTimerData(t_Handle h_Timer, t_Handle data)
-+{
-+ struct timer_list *p_Timer = (struct timer_list *)h_Timer;
-+
-+ p_Timer->data = (unsigned long)data;
-+}
-+
-+t_Handle XX_GetTimerData(t_Handle h_Timer)
-+{
-+ struct timer_list *p_Timer = (struct timer_list *)h_Timer;
-+
-+ return (t_Handle)p_Timer->data;
-+}
-+
-+uint32_t XX_GetExpirationTime(t_Handle h_Timer)
-+{
-+ struct timer_list *p_Timer = (struct timer_list *)h_Timer;
-+
-+ return (uint32_t)p_Timer->expires;
-+}
-+
-+void XX_StopTimer(t_Handle h_Timer)
-+{
-+ del_timer((struct timer_list *)h_Timer);
-+}
-+
-+void XX_ModTimer(t_Handle h_Timer, uint32_t msecs)
-+{
-+ int tmp_jiffies = (msecs*HZ)/1000;
-+
-+ if ((msecs*HZ)%1000)
-+ tmp_jiffies++;
-+ mod_timer((struct timer_list *)h_Timer, jiffies + tmp_jiffies);
-+}
-+
-+int XX_TimerIsActive(t_Handle h_Timer)
-+{
-+ return timer_pending((struct timer_list *)h_Timer);
-+}
-+
-+uint32_t XX_Sleep(uint32_t msecs)
-+{
-+ int tmp_jiffies = (msecs*HZ)/1000;
-+
-+ if ((msecs*HZ)%1000)
-+ tmp_jiffies++;
-+ return schedule_timeout(tmp_jiffies);
-+}
-+
-+/*BEWARE!!!!! UDelay routine is BUSY WAITTING!!!!!*/
-+void XX_UDelay(uint32_t usecs)
-+{
-+ udelay(usecs);
-+}
-+
-+/* TODO: verify that these are correct */
-+#define MSG_BODY_SIZE 512
-+typedef t_Error (t_MsgHandler) (t_Handle h_Mod, uint32_t msgId, uint8_t msgBody[MSG_BODY_SIZE]);
-+typedef void (t_MsgCompletionCB) (t_Handle h_Arg, uint8_t msgBody[MSG_BODY_SIZE]);
-+t_Error XX_SendMessage(char *p_DestAddr,
-+ uint32_t msgId,
-+ uint8_t msgBody[MSG_BODY_SIZE],
-+ t_MsgCompletionCB *f_CompletionCB,
-+ t_Handle h_CBArg);
-+
-+typedef struct {
-+ char *p_Addr;
-+ t_MsgHandler *f_MsgHandlerCB;
-+ t_Handle h_Mod;
-+ t_List node;
-+} t_MsgHndlr;
-+#define MSG_HNDLR_OBJECT(ptr) LIST_OBJECT(ptr, t_MsgHndlr, node)
-+
-+LIST(msgHndlrList);
-+
-+static void EnqueueMsgHndlr(t_MsgHndlr *p_MsgHndlr)
-+{
-+ uint32_t intFlags;
-+
-+ intFlags = XX_DisableAllIntr();
-+ LIST_AddToTail(&p_MsgHndlr->node, &msgHndlrList);
-+ XX_RestoreAllIntr(intFlags);
-+}
-+/* TODO: add this for multi-platform support
-+static t_MsgHndlr * DequeueMsgHndlr(void)
-+{
-+ t_MsgHndlr *p_MsgHndlr = NULL;
-+ uint32_t intFlags;
-+
-+ intFlags = XX_DisableAllIntr();
-+ if (!LIST_IsEmpty(&msgHndlrList))
-+ {
-+ p_MsgHndlr = MSG_HNDLR_OBJECT(msgHndlrList.p_Next);
-+ LIST_DelAndInit(&p_MsgHndlr->node);
-+ }
-+ XX_RestoreAllIntr(intFlags);
-+
-+ return p_MsgHndlr;
-+}
-+*/
-+static t_MsgHndlr * FindMsgHndlr(char *p_Addr)
-+{
-+ t_MsgHndlr *p_MsgHndlr;
-+ t_List *p_Pos;
-+
-+ LIST_FOR_EACH(p_Pos, &msgHndlrList)
-+ {
-+ p_MsgHndlr = MSG_HNDLR_OBJECT(p_Pos);
-+ if (strstr(p_MsgHndlr->p_Addr, p_Addr))
-+ return p_MsgHndlr;
-+ }
-+
-+ return NULL;
-+}
-+
-+t_Error XX_RegisterMessageHandler (char *p_Addr, t_MsgHandler *f_MsgHandlerCB, t_Handle h_Mod)
-+{
-+ t_MsgHndlr *p_MsgHndlr;
-+ uint32_t len;
-+
-+ p_MsgHndlr = (t_MsgHndlr*)XX_Malloc(sizeof(t_MsgHndlr));
-+ if (!p_MsgHndlr)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("message handler object!!!"));
-+ memset(p_MsgHndlr, 0, sizeof(t_MsgHndlr));
-+
-+ len = strlen(p_Addr);
-+ p_MsgHndlr->p_Addr = (char*)XX_Malloc(len+1);
-+ strncpy(p_MsgHndlr->p_Addr,p_Addr, (uint32_t)(len+1));
-+
-+ p_MsgHndlr->f_MsgHandlerCB = f_MsgHandlerCB;
-+ p_MsgHndlr->h_Mod = h_Mod;
-+ INIT_LIST(&p_MsgHndlr->node);
-+ EnqueueMsgHndlr(p_MsgHndlr);
-+
-+ return E_OK;
-+}
-+
-+t_Error XX_UnregisterMessageHandler (char *p_Addr)
-+{
-+ t_MsgHndlr *p_MsgHndlr = FindMsgHndlr(p_Addr);
-+ if (!p_MsgHndlr)
-+ RETURN_ERROR(MINOR, E_NO_DEVICE, ("message handler not found in list!!!"));
-+
-+ LIST_Del(&p_MsgHndlr->node);
-+ XX_Free(p_MsgHndlr->p_Addr);
-+ XX_Free(p_MsgHndlr);
-+
-+ return E_OK;
-+}
-+
-+t_Error XX_SendMessage(char *p_DestAddr,
-+ uint32_t msgId,
-+ uint8_t msgBody[MSG_BODY_SIZE],
-+ t_MsgCompletionCB *f_CompletionCB,
-+ t_Handle h_CBArg)
-+{
-+ t_Error ans;
-+ t_MsgHndlr *p_MsgHndlr = FindMsgHndlr(p_DestAddr);
-+ if (!p_MsgHndlr)
-+ RETURN_ERROR(MINOR, E_NO_DEVICE, ("message handler not found in list!!!"));
-+
-+ ans = p_MsgHndlr->f_MsgHandlerCB(p_MsgHndlr->h_Mod, msgId, msgBody);
-+
-+ if (f_CompletionCB)
-+ f_CompletionCB(h_CBArg, msgBody);
-+
-+ return ans;
-+}
-+
-+t_Error XX_IpcRegisterMsgHandler(char addr[XX_IPC_MAX_ADDR_NAME_LENGTH],
-+ t_IpcMsgHandler *f_MsgHandler,
-+ t_Handle h_Module,
-+ uint32_t replyLength)
-+{
-+ UNUSED(addr);UNUSED(f_MsgHandler);UNUSED(h_Module);UNUSED(replyLength);
-+ return E_OK;
-+}
-+
-+t_Error XX_IpcUnregisterMsgHandler(char addr[XX_IPC_MAX_ADDR_NAME_LENGTH])
-+{
-+ UNUSED(addr);
-+ return E_OK;
-+}
-+
-+
-+t_Error XX_IpcSendMessage(t_Handle h_Session,
-+ uint8_t *p_Msg,
-+ uint32_t msgLength,
-+ uint8_t *p_Reply,
-+ uint32_t *p_ReplyLength,
-+ t_IpcMsgCompletion *f_Completion,
-+ t_Handle h_Arg)
-+{
-+ UNUSED(h_Session); UNUSED(p_Msg); UNUSED(msgLength); UNUSED(p_Reply);
-+ UNUSED(p_ReplyLength); UNUSED(f_Completion); UNUSED(h_Arg);
-+ return E_OK;
-+}
-+
-+t_Handle XX_IpcInitSession(char destAddr[XX_IPC_MAX_ADDR_NAME_LENGTH],
-+ char srcAddr[XX_IPC_MAX_ADDR_NAME_LENGTH])
-+{
-+ UNUSED(destAddr); UNUSED(srcAddr);
-+ return E_OK;
-+}
-+
-+#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,22)
-+int GetDeviceIrqNum(int irq)
-+{
-+ struct device_node *iPar;
-+ struct irq_domain *irqHost;
-+ uint32_t hwIrq;
-+
-+ /* Get the interrupt controller */
-+ iPar = of_find_node_by_name(NULL, "mpic");
-+ hwIrq = 0;
-+
-+ ASSERT_COND(iPar != NULL);
-+ /* Get the irq host */
-+ irqHost = irq_find_host(iPar);
-+ of_node_put(iPar);
-+
-+ /* Create irq mapping */
-+ return irq_create_mapping(irqHost, hwIrq);
-+}
-+#else
-+#error "kernel not supported!!!"
-+#endif /* LINUX_VERSION_CODE */
-+
-+void * XX_PhysToVirt(physAddress_t addr)
-+{
-+ return UINT_TO_PTR(SYS_PhysToVirt((uint64_t)addr));
-+}
-+
-+physAddress_t XX_VirtToPhys(void * addr)
-+{
-+ return (physAddress_t)SYS_VirtToPhys(PTR_TO_UINT(addr));
-+}
-+
-+void * xx_MallocSmart(uint32_t size, int memPartitionId, uint32_t alignment)
-+{
-+ uintptr_t *returnCode, tmp;
-+
-+ if (alignment < sizeof(uintptr_t))
-+ alignment = sizeof(uintptr_t);
-+ size += alignment + sizeof(returnCode);
-+ tmp = (uintptr_t)xx_Malloc(size);
-+ if (tmp == 0)
-+ return NULL;
-+ returnCode = (uintptr_t*)((tmp + alignment + sizeof(returnCode)) & ~((uintptr_t)alignment - 1));
-+ *(returnCode - 1) = tmp;
-+
-+ return (void*)returnCode;
-+}
-+
-+void xx_FreeSmart(void *p)
-+{
-+ xx_Free((void*)(*((uintptr_t *)(p) - 1)));
-+}
---- /dev/null
-+++ b/drivers/net/ethernet/freescale/sdk_fman/src/xx/xx_linux.c
-@@ -0,0 +1,918 @@
-+/*
-+ * Copyright 2008-2012 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+/**************************************************************************//**
-+ @File xx_linux.c
-+
-+ @Description XX routines implementation for Linux.
-+*//***************************************************************************/
-+#include <linux/version.h>
-+
-+#if defined(CONFIG_MODVERSIONS) && !defined(MODVERSIONS)
-+#define MODVERSIONS
-+#endif
-+#ifdef MODVERSIONS
-+#include <config/modversions.h>
-+#endif /* MODVERSIONS */
-+
-+#include <linux/module.h>
-+#include <linux/kernel.h>
-+#include <linux/sched.h>
-+#include <linux/string.h>
-+#include <linux/ptrace.h>
-+#include <linux/errno.h>
-+#include <linux/ioport.h>
-+#include <linux/slab.h>
-+#include <linux/interrupt.h>
-+#include <linux/fs.h>
-+#include <linux/vmalloc.h>
-+#include <linux/init.h>
-+#include <linux/timer.h>
-+#include <linux/spinlock.h>
-+#include <linux/delay.h>
-+#include <linux/proc_fs.h>
-+#include <linux/smp.h>
-+#include <linux/of.h>
-+#ifdef CONFIG_FMAN_ARM
-+#include <linux/irqdomain.h>
-+#endif
-+
-+#include <linux/workqueue.h>
-+
-+#ifdef BIGPHYSAREA_ENABLE
-+#include <linux/bigphysarea.h>
-+#endif /* BIGPHYSAREA_ENABLE */
-+
-+#ifndef CONFIG_FMAN_ARM
-+#include <sysdev/fsl_soc.h>
-+#endif
-+#include <asm/pgtable.h>
-+#include <asm/irq.h>
-+#include <asm/bitops.h>
-+#include <asm/uaccess.h>
-+#include <asm/io.h>
-+#include <asm/atomic.h>
-+#include <asm/string.h>
-+#include <asm/byteorder.h>
-+#include <asm/page.h>
-+
-+#include "error_ext.h"
-+#include "std_ext.h"
-+#include "list_ext.h"
-+#include "mm_ext.h"
-+#include "sys_io_ext.h"
-+#include "xx.h"
-+
-+
-+#define __ERR_MODULE__ MODULE_UNKNOWN
-+
-+#ifdef BIGPHYSAREA_ENABLE
-+#define MAX_ALLOCATION_SIZE 128 * 1024 /* Maximum size allocated with kmalloc is 128K */
-+
-+
-+/* TODO: large allocations => use big phys area */
-+/******************************************************************************
-+ * routine: get_nr_pages
-+ *
-+ * description:
-+ * calculates the number of memory pages for a given size (in bytes)
-+ *
-+ * arguments:
-+ * size - the number of bytes
-+ *
-+ * return code:
-+ * The number of pages
-+ *
-+ *****************************************************************************/
-+static __inline__ uint32_t get_nr_pages (uint32_t size)
-+{
-+ return (uint32_t)((size >> PAGE_SHIFT) + (size & PAGE_SHIFT ? 1 : 0));
-+}
-+
-+static bool in_big_phys_area (uint32_t addr)
-+{
-+ uint32_t base, size;
-+
-+ bigphysarea_get_details (&base, &size);
-+ return ((addr >= base) && (addr < base + size));
-+}
-+#endif /* BIGPHYSAREA_ENABLE */
-+
-+void * xx_Malloc(uint32_t n)
-+{
-+ void *a;
-+ uint32_t flags;
-+
-+ flags = XX_DisableAllIntr();
-+#ifdef BIGPHYSAREA_ENABLE
-+ if (n >= MAX_ALLOCATION_SIZE)
-+ a = (void*)bigphysarea_alloc_pages(get_nr_pages(n), 0, GFP_ATOMIC);
-+ else
-+#endif /* BIGPHYSAREA_ENABLE */
-+ a = (void *)kmalloc((uint32_t)n, GFP_ATOMIC);
-+ if (!a)
-+ XX_Print("No memory for XX_Malloc\n");
-+ XX_RestoreAllIntr(flags);
-+
-+ return a;
-+}
-+
-+void xx_Free(void *p)
-+{
-+#ifdef BIGPHYSAREA_ENABLE
-+ if (in_big_phys_area ((uint32_t)p))
-+ bigphysarea_free_pages(p);
-+ else
-+#endif /* BIGPHYSAREA_ENABLE */
-+ kfree(p);
-+}
-+
-+void XX_Exit(int status)
-+{
-+ WARN(1, "\n\nFMD: fatal error, driver can't go on!!!\n\n");
-+}
-+
-+#define BUF_SIZE 512
-+void XX_Print(char *str, ...)
-+{
-+ va_list args;
-+#ifdef CONFIG_SMP
-+ char buf[BUF_SIZE];
-+#endif /* CONFIG_SMP */
-+
-+ va_start(args, str);
-+#ifdef CONFIG_SMP
-+ if (vsnprintf (buf, BUF_SIZE, str, args) >= BUF_SIZE)
-+ printk(KERN_WARNING "Illegal string to print!\n more than %d characters.\n\tString was not printed completelly.\n", BUF_SIZE);
-+ printk(KERN_CRIT "cpu%d/%d: %s", raw_smp_processor_id(), NR_CPUS, buf);
-+#else
-+ vprintk(str, args);
-+#endif /* CONFIG_SMP */
-+ va_end(args);
-+}
-+
-+void XX_Fprint(void *file, char *str, ...)
-+{
-+ va_list args;
-+#ifdef CONFIG_SMP
-+ char buf[BUF_SIZE];
-+#endif /* CONFIG_SMP */
-+
-+ va_start(args, str);
-+#ifdef CONFIG_SMP
-+ if (vsnprintf (buf, BUF_SIZE, str, args) >= BUF_SIZE)
-+ printk(KERN_WARNING "Illegal string to print!\n more than %d characters.\n\tString was not printed completelly.\n", BUF_SIZE);
-+ printk (KERN_CRIT "cpu%d/%d: %s", raw_smp_processor_id(), NR_CPUS, buf);
-+
-+#else
-+ vprintk(str, args);
-+#endif /* CONFIG_SMP */
-+ va_end(args);
-+}
-+
-+#ifdef DEBUG_XX_MALLOC
-+typedef void (*t_ffn)(void *);
-+typedef struct {
-+ t_ffn f_free;
-+ void *mem;
-+ char *fname;
-+ int fline;
-+ uint32_t size;
-+ t_List node;
-+} t_MemDebug;
-+#define MEMDBG_OBJECT(p_List) LIST_OBJECT(p_List, t_MemDebug, node)
-+
-+LIST(memDbgLst);
-+
-+
-+void * XX_MallocDebug(uint32_t size, char *fname, int line)
-+{
-+ void *mem;
-+ t_MemDebug *p_MemDbg;
-+
-+ p_MemDbg = (t_MemDebug *)xx_Malloc(sizeof(t_MemDebug));
-+ if (p_MemDbg == NULL)
-+ return NULL;
-+
-+ mem = xx_Malloc(size);
-+ if (mem == NULL)
-+ {
-+ XX_Free(p_MemDbg);
-+ return NULL;
-+ }
-+
-+ INIT_LIST(&p_MemDbg->node);
-+ p_MemDbg->f_free = xx_Free;
-+ p_MemDbg->mem = mem;
-+ p_MemDbg->fname = fname;
-+ p_MemDbg->fline = line;
-+ p_MemDbg->size = size+sizeof(t_MemDebug);
-+ LIST_AddToTail(&p_MemDbg->node, &memDbgLst);
-+
-+ return mem;
-+}
-+
-+void * XX_MallocSmartDebug(uint32_t size,
-+ int memPartitionId,
-+ uint32_t align,
-+ char *fname,
-+ int line)
-+{
-+ void *mem;
-+ t_MemDebug *p_MemDbg;
-+
-+ p_MemDbg = (t_MemDebug *)XX_Malloc(sizeof(t_MemDebug));
-+ if (p_MemDbg == NULL)
-+ return NULL;
-+
-+ mem = xx_MallocSmart((uint32_t)size, memPartitionId, align);
-+ if (mem == NULL)
-+ {
-+ XX_Free(p_MemDbg);
-+ return NULL;
-+ }
-+
-+ INIT_LIST(&p_MemDbg->node);
-+ p_MemDbg->f_free = xx_FreeSmart;
-+ p_MemDbg->mem = mem;
-+ p_MemDbg->fname = fname;
-+ p_MemDbg->fline = line;
-+ p_MemDbg->size = size+sizeof(t_MemDebug);
-+ LIST_AddToTail(&p_MemDbg->node, &memDbgLst);
-+
-+ return mem;
-+}
-+
-+static void debug_free(void *mem)
-+{
-+ t_List *p_MemDbgLh = NULL;
-+ t_MemDebug *p_MemDbg;
-+ bool found = FALSE;
-+
-+ if (LIST_IsEmpty(&memDbgLst))
-+ {
-+ REPORT_ERROR(MAJOR, E_ALREADY_FREE, ("Unbalanced free (0x%08x)", mem));
-+ return;
-+ }
-+
-+ LIST_FOR_EACH(p_MemDbgLh, &memDbgLst)
-+ {
-+ p_MemDbg = MEMDBG_OBJECT(p_MemDbgLh);
-+ if (p_MemDbg->mem == mem)
-+ {
-+ found = TRUE;
-+ break;
-+ }
-+ }
-+
-+ if (!found)
-+ {
-+ REPORT_ERROR(MAJOR, E_NOT_FOUND,
-+ ("Attempt to free unallocated address (0x%08x)",mem));
-+ dump_stack();
-+ return;
-+ }
-+
-+ LIST_Del(p_MemDbgLh);
-+ p_MemDbg->f_free(mem);
-+ p_MemDbg->f_free(p_MemDbg);
-+}
-+
-+void XX_FreeSmart(void *p)
-+{
-+ debug_free(p);
-+}
-+
-+
-+void XX_Free(void *p)
-+{
-+ debug_free(p);
-+}
-+
-+#else /* not DEBUG_XX_MALLOC */
-+void * XX_Malloc(uint32_t size)
-+{
-+ return xx_Malloc(size);
-+}
-+
-+void * XX_MallocSmart(uint32_t size, int memPartitionId, uint32_t alignment)
-+{
-+ return xx_MallocSmart(size,memPartitionId, alignment);
-+}
-+
-+void XX_FreeSmart(void *p)
-+{
-+ xx_FreeSmart(p);
-+}
-+
-+
-+void XX_Free(void *p)
-+{
-+ xx_Free(p);
-+}
-+#endif /* not DEBUG_XX_MALLOC */
-+
-+
-+#if (defined(REPORT_EVENTS) && (REPORT_EVENTS > 0))
-+void XX_EventById(uint32_t event, t_Handle appId, uint16_t flags, char *msg)
-+{
-+ e_Event eventCode = (e_Event)event;
-+
-+ UNUSED(eventCode);
-+ UNUSED(appId);
-+ UNUSED(flags);
-+ UNUSED(msg);
-+}
-+#endif /* (defined(REPORT_EVENTS) && ... */
-+
-+
-+uint32_t XX_DisableAllIntr(void)
-+{
-+ unsigned long flags;
-+
-+#ifdef local_irq_save_nort
-+ local_irq_save_nort(flags);
-+#else
-+ local_irq_save(flags);
-+#endif
-+
-+ return (uint32_t)flags;
-+}
-+
-+void XX_RestoreAllIntr(uint32_t flags)
-+{
-+#ifdef local_irq_restore_nort
-+ local_irq_restore_nort((unsigned long)flags);
-+#else
-+ local_irq_restore((unsigned long)flags);
-+#endif
-+}
-+
-+t_Error XX_Call( uint32_t qid, t_Error (* f)(t_Handle), t_Handle id, t_Handle appId, uint16_t flags )
-+{
-+ UNUSED(qid);
-+ UNUSED(appId);
-+ UNUSED(flags);
-+
-+ return f(id);
-+}
-+
-+int XX_IsICacheEnable(void)
-+{
-+ return TRUE;
-+}
-+
-+int XX_IsDCacheEnable(void)
-+{
-+ return TRUE;
-+}
-+
-+
-+typedef struct {
-+ t_Isr *f_Isr;
-+ t_Handle handle;
-+} t_InterruptHandler;
-+
-+
-+t_Handle interruptHandlers[0x00010000];
-+
-+#ifdef CONFIG_FMAN_ARM
-+static irqreturn_t LinuxInterruptHandler (int irq, void *dev_id)
-+{
-+ t_InterruptHandler *p_IntrHndl = (t_InterruptHandler *)dev_id;
-+ p_IntrHndl->f_Isr(p_IntrHndl->handle);
-+ return IRQ_HANDLED;
-+}
-+#endif
-+
-+t_Error XX_SetIntr(int irq, t_Isr *f_Isr, t_Handle handle)
-+{
-+#ifdef CONFIG_FMAN_ARM
-+ const char *device;
-+ t_InterruptHandler *p_IntrHndl;
-+
-+ device = GetDeviceName(irq);
-+ if (device == NULL)
-+ RETURN_ERROR(MAJOR, E_INVALID_VALUE, ("Interrupt source - %d", irq));
-+
-+ p_IntrHndl = (t_InterruptHandler *)XX_Malloc(sizeof(t_InterruptHandler));
-+ if (p_IntrHndl == NULL)
-+ RETURN_ERROR(MAJOR, E_NO_MEMORY, NO_MSG);
-+ p_IntrHndl->f_Isr = f_Isr;
-+ p_IntrHndl->handle = handle;
-+ interruptHandlers[irq] = p_IntrHndl;
-+
-+ if (request_irq(GetDeviceIrqNum(irq), LinuxInterruptHandler, 0, device, p_IntrHndl) < 0)
-+ RETURN_ERROR(MAJOR, E_BUSY, ("Can't get IRQ %s\n", device));
-+ disable_irq(GetDeviceIrqNum(irq));
-+#endif
-+ return E_OK;
-+}
-+
-+t_Error XX_FreeIntr(int irq)
-+{
-+ t_InterruptHandler *p_IntrHndl = interruptHandlers[irq];
-+ free_irq(GetDeviceIrqNum(irq), p_IntrHndl);
-+ XX_Free(p_IntrHndl);
-+ interruptHandlers[irq] = 0;
-+ return E_OK;
-+}
-+
-+t_Error XX_EnableIntr(int irq)
-+{
-+ enable_irq(GetDeviceIrqNum(irq));
-+ return E_OK;
-+}
-+
-+t_Error XX_DisableIntr(int irq)
-+{
-+ disable_irq(GetDeviceIrqNum(irq));
-+ return E_OK;
-+}
-+
-+
-+/*****************************************************************************/
-+/* Tasklet Service Routines */
-+/*****************************************************************************/
-+#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,20)
-+typedef struct
-+{
-+ t_Handle h_Data;
-+ void (*f_Callback) (void *);
-+ struct delayed_work dwork;
-+} t_Tasklet;
-+
-+static void GenericTaskletCallback(struct work_struct *p_Work)
-+{
-+ t_Tasklet *p_Task = container_of(p_Work, t_Tasklet, dwork.work);
-+
-+ p_Task->f_Callback(p_Task->h_Data);
-+}
-+#endif /* LINUX_VERSION_CODE */
-+
-+
-+t_TaskletHandle XX_InitTasklet (void (*routine)(void *), void *data)
-+{
-+#if LINUX_VERSION_CODE < KERNEL_VERSION(2,6,20)
-+ struct work_struct *p_Task;
-+ p_Task = (struct work_struct *)XX_Malloc(sizeof(struct work_struct));
-+ INIT_WORK(p_Task, routine, data);
-+#else
-+ t_Tasklet *p_Task = (t_Tasklet *)XX_Malloc(sizeof(t_Tasklet));
-+ p_Task->h_Data = data;
-+ p_Task->f_Callback = routine;
-+ INIT_DELAYED_WORK(&p_Task->dwork, GenericTaskletCallback);
-+#endif /* LINUX_VERSION_CODE */
-+
-+ return (t_TaskletHandle)p_Task;
-+}
-+
-+
-+void XX_FreeTasklet (t_TaskletHandle h_Tasklet)
-+{
-+ if (h_Tasklet)
-+ XX_Free(h_Tasklet);
-+}
-+
-+int XX_ScheduleTask(t_TaskletHandle h_Tasklet, int immediate)
-+{
-+ int ans;
-+
-+#if LINUX_VERSION_CODE < KERNEL_VERSION(2,6,20)
-+ if (immediate)
-+ ans = schedule_work(h_Tasklet);
-+ else
-+ ans = schedule_delayed_work(h_Tasklet, 1);
-+#else
-+ if (immediate)
-+ ans = schedule_delayed_work(&((t_Tasklet *)h_Tasklet)->dwork, 0);
-+ else
-+ ans = schedule_delayed_work(&((t_Tasklet *)h_Tasklet)->dwork, HZ);
-+#endif /* LINUX_VERSION_CODE */
-+
-+ return ans;
-+}
-+
-+void XX_FlushScheduledTasks(void)
-+{
-+#if LINUX_VERSION_CODE < KERNEL_VERSION(2,6,0)
-+ flush_scheduled_tasks();
-+#else
-+ flush_scheduled_work();
-+#endif /* LINUX_VERSION_CODE */
-+}
-+
-+int XX_TaskletIsQueued(t_TaskletHandle h_Tasklet)
-+{
-+#if LINUX_VERSION_CODE < KERNEL_VERSION(2,6,20)
-+ return (int)(((struct work_struct *)h_Tasklet)->pending);
-+#else
-+ return (int)delayed_work_pending(&((t_Tasklet *)h_Tasklet)->dwork);
-+#endif /* LINUX_VERSION_CODE */
-+}
-+
-+void XX_SetTaskletData(t_TaskletHandle h_Tasklet, t_Handle data)
-+{
-+#if LINUX_VERSION_CODE < KERNEL_VERSION(2,6,0)
-+ ((struct tq_struct *)h_Tasklet)->data = data;
-+#elif LINUX_VERSION_CODE < KERNEL_VERSION(2,6,20)
-+ ((struct work_struct *)h_Tasklet)->data = data;
-+#else
-+ ((t_Tasklet *)h_Tasklet)->h_Data = data;
-+#endif /* LINUX_VERSION_CODE */
-+}
-+
-+t_Handle XX_GetTaskletData(t_TaskletHandle h_Tasklet)
-+{
-+#if LINUX_VERSION_CODE < KERNEL_VERSION(2,6,20)
-+ return (t_Handle)(((struct work_struct *)h_Tasklet)->data);
-+#else
-+ return ((t_Tasklet *)h_Tasklet)->h_Data;
-+#endif /* LINUX_VERSION_CODE */
-+}
-+
-+
-+/*****************************************************************************/
-+/* Spinlock Service Routines */
-+/*****************************************************************************/
-+
-+t_Handle XX_InitSpinlock(void)
-+{
-+ spinlock_t *p_Spinlock = (spinlock_t *)XX_Malloc(sizeof(spinlock_t));
-+ if (!p_Spinlock)
-+ return NULL;
-+
-+ spin_lock_init(p_Spinlock);
-+
-+ return (t_Handle)p_Spinlock;
-+}
-+
-+void XX_FreeSpinlock(t_Handle h_Spinlock)
-+{
-+ if (h_Spinlock)
-+ XX_Free(h_Spinlock);
-+}
-+
-+void XX_LockSpinlock(t_Handle h_Spinlock)
-+{
-+ spin_lock((spinlock_t *)h_Spinlock);
-+}
-+
-+void XX_UnlockSpinlock(t_Handle h_Spinlock)
-+{
-+ spin_unlock((spinlock_t *)h_Spinlock);
-+}
-+
-+uint32_t XX_LockIntrSpinlock(t_Handle h_Spinlock)
-+{
-+ unsigned long intrFlags;
-+ spin_lock_irqsave((spinlock_t *)h_Spinlock, intrFlags);
-+ return intrFlags;
-+}
-+
-+void XX_UnlockIntrSpinlock(t_Handle h_Spinlock, uint32_t intrFlags)
-+{
-+ spin_unlock_irqrestore((spinlock_t *)h_Spinlock, (unsigned long)intrFlags);
-+}
-+
-+
-+/*****************************************************************************/
-+/* Timers Service Routines */
-+/*****************************************************************************/
-+/* The time now is in mili sec. resolution */
-+uint32_t XX_CurrentTime(void)
-+{
-+ return (jiffies*1000)/HZ;
-+}
-+
-+
-+t_Handle XX_CreateTimer(void)
-+{
-+ struct timer_list *p_Timer = (struct timer_list *)XX_Malloc(sizeof(struct timer_list));
-+ if (p_Timer)
-+ {
-+ memset(p_Timer, 0, sizeof(struct timer_list));
-+ init_timer(p_Timer);
-+ }
-+ return (t_Handle)p_Timer;
-+}
-+
-+void XX_FreeTimer(t_Handle h_Timer)
-+{
-+ if (h_Timer)
-+ XX_Free(h_Timer);
-+}
-+
-+void XX_StartTimer(t_Handle h_Timer,
-+ uint32_t msecs,
-+ bool periodic,
-+ void (*f_TimerExpired)(t_Handle),
-+ t_Handle h_Arg)
-+{
-+ int tmp_jiffies = (msecs*HZ)/1000;
-+ struct timer_list *p_Timer = (struct timer_list *)h_Timer;
-+
-+ SANITY_CHECK_RETURN((periodic == FALSE), E_NOT_SUPPORTED);
-+
-+ p_Timer->function = (void (*)(unsigned long))f_TimerExpired;
-+ p_Timer->data = (unsigned long)h_Arg;
-+ if ((msecs*HZ)%1000)
-+ tmp_jiffies++;
-+ p_Timer->expires = (jiffies + tmp_jiffies);
-+
-+ add_timer((struct timer_list *)h_Timer);
-+}
-+
-+void XX_SetTimerData(t_Handle h_Timer, t_Handle data)
-+{
-+ struct timer_list *p_Timer = (struct timer_list *)h_Timer;
-+
-+ p_Timer->data = (unsigned long)data;
-+}
-+
-+t_Handle XX_GetTimerData(t_Handle h_Timer)
-+{
-+ struct timer_list *p_Timer = (struct timer_list *)h_Timer;
-+
-+ return (t_Handle)p_Timer->data;
-+}
-+
-+uint32_t XX_GetExpirationTime(t_Handle h_Timer)
-+{
-+ struct timer_list *p_Timer = (struct timer_list *)h_Timer;
-+
-+ return (uint32_t)p_Timer->expires;
-+}
-+
-+void XX_StopTimer(t_Handle h_Timer)
-+{
-+ del_timer((struct timer_list *)h_Timer);
-+}
-+
-+void XX_ModTimer(t_Handle h_Timer, uint32_t msecs)
-+{
-+ int tmp_jiffies = (msecs*HZ)/1000;
-+
-+ if ((msecs*HZ)%1000)
-+ tmp_jiffies++;
-+ mod_timer((struct timer_list *)h_Timer, jiffies + tmp_jiffies);
-+}
-+
-+int XX_TimerIsActive(t_Handle h_Timer)
-+{
-+ return timer_pending((struct timer_list *)h_Timer);
-+}
-+
-+uint32_t XX_Sleep(uint32_t msecs)
-+{
-+ int tmp_jiffies = (msecs*HZ)/1000;
-+
-+ if ((msecs*HZ)%1000)
-+ tmp_jiffies++;
-+ return schedule_timeout(tmp_jiffies);
-+}
-+
-+/*BEWARE!!!!! UDelay routine is BUSY WAITTING!!!!!*/
-+void XX_UDelay(uint32_t usecs)
-+{
-+ udelay(usecs);
-+}
-+
-+/* TODO: verify that these are correct */
-+#define MSG_BODY_SIZE 512
-+typedef t_Error (t_MsgHandler) (t_Handle h_Mod, uint32_t msgId, uint8_t msgBody[MSG_BODY_SIZE]);
-+typedef void (t_MsgCompletionCB) (t_Handle h_Arg, uint8_t msgBody[MSG_BODY_SIZE]);
-+t_Error XX_SendMessage(char *p_DestAddr,
-+ uint32_t msgId,
-+ uint8_t msgBody[MSG_BODY_SIZE],
-+ t_MsgCompletionCB *f_CompletionCB,
-+ t_Handle h_CBArg);
-+
-+typedef struct {
-+ char *p_Addr;
-+ t_MsgHandler *f_MsgHandlerCB;
-+ t_Handle h_Mod;
-+ t_List node;
-+} t_MsgHndlr;
-+#define MSG_HNDLR_OBJECT(ptr) LIST_OBJECT(ptr, t_MsgHndlr, node)
-+
-+LIST(msgHndlrList);
-+
-+static void EnqueueMsgHndlr(t_MsgHndlr *p_MsgHndlr)
-+{
-+ uint32_t intFlags;
-+
-+ intFlags = XX_DisableAllIntr();
-+ LIST_AddToTail(&p_MsgHndlr->node, &msgHndlrList);
-+ XX_RestoreAllIntr(intFlags);
-+}
-+/* TODO: add this for multi-platform support
-+static t_MsgHndlr * DequeueMsgHndlr(void)
-+{
-+ t_MsgHndlr *p_MsgHndlr = NULL;
-+ uint32_t intFlags;
-+
-+ intFlags = XX_DisableAllIntr();
-+ if (!LIST_IsEmpty(&msgHndlrList))
-+ {
-+ p_MsgHndlr = MSG_HNDLR_OBJECT(msgHndlrList.p_Next);
-+ LIST_DelAndInit(&p_MsgHndlr->node);
-+ }
-+ XX_RestoreAllIntr(intFlags);
-+
-+ return p_MsgHndlr;
-+}
-+*/
-+static t_MsgHndlr * FindMsgHndlr(char *p_Addr)
-+{
-+ t_MsgHndlr *p_MsgHndlr;
-+ t_List *p_Pos;
-+
-+ LIST_FOR_EACH(p_Pos, &msgHndlrList)
-+ {
-+ p_MsgHndlr = MSG_HNDLR_OBJECT(p_Pos);
-+ if (strstr(p_MsgHndlr->p_Addr, p_Addr))
-+ return p_MsgHndlr;
-+ }
-+
-+ return NULL;
-+}
-+
-+t_Error XX_RegisterMessageHandler (char *p_Addr, t_MsgHandler *f_MsgHandlerCB, t_Handle h_Mod)
-+{
-+ t_MsgHndlr *p_MsgHndlr;
-+ uint32_t len;
-+
-+ p_MsgHndlr = (t_MsgHndlr*)XX_Malloc(sizeof(t_MsgHndlr));
-+ if (!p_MsgHndlr)
-+ RETURN_ERROR(MINOR, E_NO_MEMORY, ("message handler object!!!"));
-+ memset(p_MsgHndlr, 0, sizeof(t_MsgHndlr));
-+
-+ len = strlen(p_Addr);
-+ p_MsgHndlr->p_Addr = (char*)XX_Malloc(len+1);
-+ strncpy(p_MsgHndlr->p_Addr,p_Addr, (uint32_t)(len+1));
-+
-+ p_MsgHndlr->f_MsgHandlerCB = f_MsgHandlerCB;
-+ p_MsgHndlr->h_Mod = h_Mod;
-+ INIT_LIST(&p_MsgHndlr->node);
-+ EnqueueMsgHndlr(p_MsgHndlr);
-+
-+ return E_OK;
-+}
-+
-+t_Error XX_UnregisterMessageHandler (char *p_Addr)
-+{
-+ t_MsgHndlr *p_MsgHndlr = FindMsgHndlr(p_Addr);
-+ if (!p_MsgHndlr)
-+ RETURN_ERROR(MINOR, E_NO_DEVICE, ("message handler not found in list!!!"));
-+
-+ LIST_Del(&p_MsgHndlr->node);
-+ XX_Free(p_MsgHndlr->p_Addr);
-+ XX_Free(p_MsgHndlr);
-+
-+ return E_OK;
-+}
-+
-+t_Error XX_SendMessage(char *p_DestAddr,
-+ uint32_t msgId,
-+ uint8_t msgBody[MSG_BODY_SIZE],
-+ t_MsgCompletionCB *f_CompletionCB,
-+ t_Handle h_CBArg)
-+{
-+ t_Error ans;
-+ t_MsgHndlr *p_MsgHndlr = FindMsgHndlr(p_DestAddr);
-+ if (!p_MsgHndlr)
-+ RETURN_ERROR(MINOR, E_NO_DEVICE, ("message handler not found in list!!!"));
-+
-+ ans = p_MsgHndlr->f_MsgHandlerCB(p_MsgHndlr->h_Mod, msgId, msgBody);
-+
-+ if (f_CompletionCB)
-+ f_CompletionCB(h_CBArg, msgBody);
-+
-+ return ans;
-+}
-+
-+t_Error XX_IpcRegisterMsgHandler(char addr[XX_IPC_MAX_ADDR_NAME_LENGTH],
-+ t_IpcMsgHandler *f_MsgHandler,
-+ t_Handle h_Module,
-+ uint32_t replyLength)
-+{
-+ UNUSED(addr);UNUSED(f_MsgHandler);UNUSED(h_Module);UNUSED(replyLength);
-+ return E_OK;
-+}
-+
-+t_Error XX_IpcUnregisterMsgHandler(char addr[XX_IPC_MAX_ADDR_NAME_LENGTH])
-+{
-+ UNUSED(addr);
-+ return E_OK;
-+}
-+
-+
-+t_Error XX_IpcSendMessage(t_Handle h_Session,
-+ uint8_t *p_Msg,
-+ uint32_t msgLength,
-+ uint8_t *p_Reply,
-+ uint32_t *p_ReplyLength,
-+ t_IpcMsgCompletion *f_Completion,
-+ t_Handle h_Arg)
-+{
-+ UNUSED(h_Session); UNUSED(p_Msg); UNUSED(msgLength); UNUSED(p_Reply);
-+ UNUSED(p_ReplyLength); UNUSED(f_Completion); UNUSED(h_Arg);
-+ return E_OK;
-+}
-+
-+t_Handle XX_IpcInitSession(char destAddr[XX_IPC_MAX_ADDR_NAME_LENGTH],
-+ char srcAddr[XX_IPC_MAX_ADDR_NAME_LENGTH])
-+{
-+ UNUSED(destAddr); UNUSED(srcAddr);
-+ return E_OK;
-+}
-+
-+/*Forced to introduce due to PRINT_FMT_PARAMS define*/
-+uint32_t E500_GetId(void)
-+{
-+ return raw_smp_processor_id();
-+}
-+
-+#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,22)
-+int GetDeviceIrqNum(int irq)
-+{
-+ struct device_node *iPar;
-+ struct irq_domain *irqHost;
-+ uint32_t hwIrq;
-+
-+ /* Get the interrupt controller */
-+ iPar = of_find_node_by_name(NULL, "mpic");
-+ hwIrq = 0;
-+
-+ ASSERT_COND(iPar != NULL);
-+ /* Get the irq host */
-+ irqHost = irq_find_host(iPar);
-+ of_node_put(iPar);
-+
-+ /* Create irq mapping */
-+ return irq_create_mapping(irqHost, hwIrq);
-+}
-+#else
-+#error "kernel not supported!!!"
-+#endif /* LINUX_VERSION_CODE */
-+
-+void * XX_PhysToVirt(physAddress_t addr)
-+{
-+ return UINT_TO_PTR(SYS_PhysToVirt((uint64_t)addr));
-+}
-+
-+physAddress_t XX_VirtToPhys(void * addr)
-+{
-+ return (physAddress_t)SYS_VirtToPhys(PTR_TO_UINT(addr));
-+}
-+
-+void * xx_MallocSmart(uint32_t size, int memPartitionId, uint32_t alignment)
-+{
-+ uintptr_t *returnCode, tmp;
-+
-+ if (alignment < sizeof(uintptr_t))
-+ alignment = sizeof(uintptr_t);
-+ size += alignment + sizeof(returnCode);
-+ tmp = (uintptr_t)xx_Malloc(size);
-+ if (tmp == 0)
-+ return NULL;
-+ returnCode = (uintptr_t*)((tmp + alignment + sizeof(returnCode)) & ~((uintptr_t)alignment - 1));
-+ *(returnCode - 1) = tmp;
-+
-+ return (void*)returnCode;
-+}
-+
-+void xx_FreeSmart(void *p)
-+{
-+ xx_Free((void*)(*((uintptr_t *)(p) - 1)));
-+}
---- /dev/null
-+++ b/drivers/staging/fsl_qbman/Kconfig
-@@ -0,0 +1,228 @@
-+config FSL_SDK_DPA
-+ bool "Freescale Datapath Queue and Buffer management"
-+ depends on !FSL_DPAA
-+ select FSL_QMAN_FQ_LOOKUP if PPC64
-+ select FSL_QMAN_FQ_LOOKUP if ARM64
-+
-+
-+menu "Freescale Datapath QMan/BMan options"
-+ depends on FSL_SDK_DPA
-+
-+config FSL_DPA_CHECKING
-+ bool "additional driver checking"
-+ default n
-+ ---help---
-+ Compiles in additional checks to sanity-check the drivers and any
-+ use of it by other code. Not recommended for performance.
-+
-+config FSL_DPA_CAN_WAIT
-+ bool
-+ default y
-+
-+config FSL_DPA_CAN_WAIT_SYNC
-+ bool
-+ default y
-+
-+config FSL_DPA_PIRQ_FAST
-+ bool
-+ default y
-+
-+config FSL_DPA_PIRQ_SLOW
-+ bool
-+ default y
-+
-+config FSL_DPA_PORTAL_SHARE
-+ bool
-+ default y
-+
-+config FSL_SDK_BMAN
-+ bool "Freescale Buffer Manager (BMan) support"
-+ default y
-+
-+if FSL_SDK_BMAN
-+
-+config FSL_BMAN_CONFIG
-+ bool "BMan device management"
-+ default y
-+ ---help---
-+ If this linux image is running natively, you need this option. If this
-+ linux image is running as a guest OS under the hypervisor, only one
-+ guest OS ("the control plane") needs this option.
-+
-+config FSL_BMAN_TEST
-+ tristate "BMan self-tests"
-+ default n
-+ ---help---
-+ This option compiles self-test code for BMan.
-+
-+config FSL_BMAN_TEST_HIGH
-+ bool "BMan high-level self-test"
-+ depends on FSL_BMAN_TEST
-+ default y
-+ ---help---
-+ This requires the presence of cpu-affine portals, and performs
-+ high-level API testing with them (whichever portal(s) are affine to
-+ the cpu(s) the test executes on).
-+
-+config FSL_BMAN_TEST_THRESH
-+ bool "BMan threshold test"
-+ depends on FSL_BMAN_TEST
-+ default y
-+ ---help---
-+ Multi-threaded (SMP) test of BMan pool depletion. A pool is seeded
-+ before multiple threads (one per cpu) create pool objects to track
-+ depletion state changes. The pool is then drained to empty by a
-+ "drainer" thread, and the other threads that they observe exactly
-+ the depletion state changes that are expected.
-+
-+config FSL_BMAN_DEBUGFS
-+ tristate "BMan debugfs interface"
-+ depends on DEBUG_FS
-+ default y
-+ ---help---
-+ This option compiles debugfs code for BMan.
-+
-+endif # FSL_SDK_BMAN
-+
-+config FSL_SDK_QMAN
-+ bool "Freescale Queue Manager (QMan) support"
-+ default y
-+
-+if FSL_SDK_QMAN
-+
-+config FSL_QMAN_POLL_LIMIT
-+ int
-+ default 32
-+
-+config FSL_QMAN_CONFIG
-+ bool "QMan device management"
-+ default y
-+ ---help---
-+ If this linux image is running natively, you need this option. If this
-+ linux image is running as a guest OS under the hypervisor, only one
-+ guest OS ("the control plane") needs this option.
-+
-+config FSL_QMAN_TEST
-+ tristate "QMan self-tests"
-+ default n
-+ ---help---
-+ This option compiles self-test code for QMan.
-+
-+config FSL_QMAN_TEST_STASH_POTATO
-+ bool "QMan 'hot potato' data-stashing self-test"
-+ depends on FSL_QMAN_TEST
-+ default y
-+ ---help---
-+ This performs a "hot potato" style test enqueuing/dequeuing a frame
-+ across a series of FQs scheduled to different portals (and cpus), with
-+ DQRR, data and context stashing always on.
-+
-+config FSL_QMAN_TEST_HIGH
-+ bool "QMan high-level self-test"
-+ depends on FSL_QMAN_TEST
-+ default y
-+ ---help---
-+ This requires the presence of cpu-affine portals, and performs
-+ high-level API testing with them (whichever portal(s) are affine to
-+ the cpu(s) the test executes on).
-+
-+config FSL_QMAN_DEBUGFS
-+ tristate "QMan debugfs interface"
-+ depends on DEBUG_FS
-+ default y
-+ ---help---
-+ This option compiles debugfs code for QMan.
-+
-+# H/w settings that can be hard-coded for now.
-+config FSL_QMAN_FQD_SZ
-+ int "size of Frame Queue Descriptor region"
-+ default 10
-+ ---help---
-+ This is the size of the FQD region defined as: PAGE_SIZE * (2^value)
-+ ex: 10 => PAGE_SIZE * (2^10)
-+ Note: Default device-trees now require minimum Kconfig setting of 10.
-+
-+config FSL_QMAN_PFDR_SZ
-+ int "size of the PFDR pool"
-+ default 13
-+ ---help---
-+ This is the size of the PFDR pool defined as: PAGE_SIZE * (2^value)
-+ ex: 13 => PAGE_SIZE * (2^13)
-+
-+# Corenet initiator settings. Stash request queues are 4-deep to match cores'
-+# ability to snart. Stash priority is 3, other priorities are 2.
-+config FSL_QMAN_CI_SCHED_CFG_SRCCIV
-+ int
-+ depends on FSL_QMAN_CONFIG
-+ default 4
-+config FSL_QMAN_CI_SCHED_CFG_SRQ_W
-+ int
-+ depends on FSL_QMAN_CONFIG
-+ default 3
-+config FSL_QMAN_CI_SCHED_CFG_RW_W
-+ int
-+ depends on FSL_QMAN_CONFIG
-+ default 2
-+config FSL_QMAN_CI_SCHED_CFG_BMAN_W
-+ int
-+ depends on FSL_QMAN_CONFIG
-+ default 2
-+
-+# portal interrupt settings
-+config FSL_QMAN_PIRQ_DQRR_ITHRESH
-+ int
-+ default 12
-+config FSL_QMAN_PIRQ_MR_ITHRESH
-+ int
-+ default 4
-+config FSL_QMAN_PIRQ_IPERIOD
-+ int
-+ default 100
-+
-+# 64 bit kernel support
-+config FSL_QMAN_FQ_LOOKUP
-+ bool
-+ default n
-+
-+config QMAN_CEETM_UPDATE_PERIOD
-+ int "Token update period for shaping, in nanoseconds"
-+ default 1000
-+ ---help---
-+ Traffic shaping works by performing token calculations (using
-+ credits) on shaper instances periodically. This update period
-+ sets the granularity for how often those token rate credit
-+ updates are performed, and thus determines the accuracy and
-+ range of traffic rates that can be configured by users. The
-+ reference manual recommends a 1 microsecond period as providing
-+ a good balance between granularity and range.
-+
-+ Unless you know what you are doing, leave this value at its default.
-+
-+config FSL_QMAN_INIT_TIMEOUT
-+ int "timeout for qman init stage, in seconds"
-+ default 10
-+ ---help---
-+ The timeout setting to quit the initialization loop for non-control
-+ partition in case the control partition fails to boot-up.
-+
-+endif # FSL_SDK_QMAN
-+
-+config FSL_USDPAA
-+ bool "Freescale USDPAA process driver"
-+ depends on FSL_SDK_DPA
-+ default y
-+ ---help---
-+ This driver provides user-space access to kernel-managed
-+ resource interfaces for USDPAA applications, on the assumption
-+ that each process will open this device once. Specifically, this
-+ device exposes functionality that would be awkward if exposed
-+ via the portal devices - ie. this device exposes functionality
-+ that is inherently process-wide rather than portal-specific.
-+ This device is necessary for obtaining access to DMA memory and
-+ for allocation of Qman and Bman resources. In short, if you wish
-+ to use USDPAA applications, you need this.
-+
-+ If unsure, say Y.
-+
-+
-+endmenu
---- /dev/null
-+++ b/drivers/staging/fsl_qbman/Makefile
-@@ -0,0 +1,28 @@
-+subdir-ccflags-y := -Werror
-+
-+# Common
-+obj-$(CONFIG_FSL_SDK_DPA) += dpa_alloc.o
-+obj-$(CONFIG_FSL_SDK_DPA) += qbman_driver.o
-+
-+# Bman
-+obj-$(CONFIG_FSL_SDK_BMAN) += bman_high.o
-+obj-$(CONFIG_FSL_BMAN_CONFIG) += bman_config.o bman_driver.o
-+obj-$(CONFIG_FSL_BMAN_TEST) += bman_tester.o
-+obj-$(CONFIG_FSL_BMAN_DEBUGFS) += bman_debugfs_interface.o
-+bman_tester-y = bman_test.o
-+bman_tester-$(CONFIG_FSL_BMAN_TEST_HIGH) += bman_test_high.o
-+bman_tester-$(CONFIG_FSL_BMAN_TEST_THRESH) += bman_test_thresh.o
-+bman_debugfs_interface-y = bman_debugfs.o
-+
-+# Qman
-+obj-$(CONFIG_FSL_SDK_QMAN) += qman_high.o qman_utility.o
-+obj-$(CONFIG_FSL_QMAN_CONFIG) += qman_config.o qman_driver.o
-+obj-$(CONFIG_FSL_QMAN_TEST) += qman_tester.o
-+qman_tester-y = qman_test.o
-+qman_tester-$(CONFIG_FSL_QMAN_TEST_STASH_POTATO) += qman_test_hotpotato.o
-+qman_tester-$(CONFIG_FSL_QMAN_TEST_HIGH) += qman_test_high.o
-+obj-$(CONFIG_FSL_QMAN_DEBUGFS) += qman_debugfs_interface.o
-+qman_debugfs_interface-y = qman_debugfs.o
-+
-+# USDPAA
-+obj-$(CONFIG_FSL_USDPAA) += fsl_usdpaa.o fsl_usdpaa_irq.o
---- /dev/null
-+++ b/drivers/staging/fsl_qbman/bman_config.c
-@@ -0,0 +1,720 @@
-+/* Copyright (c) 2009-2012 Freescale Semiconductor, Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#include <asm/cacheflush.h>
-+#include "bman_private.h"
-+#include <linux/of_reserved_mem.h>
-+
-+/* Last updated for v00.79 of the BG */
-+
-+struct bman;
-+
-+/* Register offsets */
-+#define REG_POOL_SWDET(n) (0x0000 + ((n) * 0x04))
-+#define REG_POOL_HWDET(n) (0x0100 + ((n) * 0x04))
-+#define REG_POOL_SWDXT(n) (0x0200 + ((n) * 0x04))
-+#define REG_POOL_HWDXT(n) (0x0300 + ((n) * 0x04))
-+#define REG_POOL_CONTENT(n) (0x0600 + ((n) * 0x04))
-+#define REG_FBPR_FPC 0x0800
-+#define REG_STATE_IDLE 0x960
-+#define REG_STATE_STOP 0x964
-+#define REG_ECSR 0x0a00
-+#define REG_ECIR 0x0a04
-+#define REG_EADR 0x0a08
-+#define REG_EDATA(n) (0x0a10 + ((n) * 0x04))
-+#define REG_SBEC(n) (0x0a80 + ((n) * 0x04))
-+#define REG_IP_REV_1 0x0bf8
-+#define REG_IP_REV_2 0x0bfc
-+#define REG_FBPR_BARE 0x0c00
-+#define REG_FBPR_BAR 0x0c04
-+#define REG_FBPR_AR 0x0c10
-+#define REG_SRCIDR 0x0d04
-+#define REG_LIODNR 0x0d08
-+#define REG_ERR_ISR 0x0e00 /* + "enum bm_isr_reg" */
-+
-+/* Used by all error interrupt registers except 'inhibit' */
-+#define BM_EIRQ_IVCI 0x00000010 /* Invalid Command Verb */
-+#define BM_EIRQ_FLWI 0x00000008 /* FBPR Low Watermark */
-+#define BM_EIRQ_MBEI 0x00000004 /* Multi-bit ECC Error */
-+#define BM_EIRQ_SBEI 0x00000002 /* Single-bit ECC Error */
-+#define BM_EIRQ_BSCN 0x00000001 /* pool State Change Notification */
-+
-+/* BMAN_ECIR valid error bit */
-+#define PORTAL_ECSR_ERR (BM_EIRQ_IVCI)
-+
-+union bman_ecir {
-+ u32 ecir_raw;
-+ struct {
-+#if __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
-+ u32 __reserved1:4;
-+ u32 portal_num:4;
-+ u32 __reserved2:12;
-+ u32 numb:4;
-+ u32 __reserved3:2;
-+ u32 pid:6;
-+#else
-+ u32 pid:6;
-+ u32 __reserved3:2;
-+ u32 numb:4;
-+ u32 __reserved2:12;
-+ u32 portal_num:4;
-+ u32 __reserved1:4;
-+#endif
-+ } __packed info;
-+};
-+
-+union bman_eadr {
-+ u32 eadr_raw;
-+ struct {
-+#if __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
-+ u32 __reserved1:5;
-+ u32 memid:3;
-+ u32 __reserved2:14;
-+ u32 eadr:10;
-+#else
-+ u32 eadr:10;
-+ u32 __reserved2:14;
-+ u32 memid:3;
-+ u32 __reserved1:5;
-+#endif
-+ } __packed info;
-+};
-+
-+struct bman_hwerr_txt {
-+ u32 mask;
-+ const char *txt;
-+};
-+
-+#define BMAN_HWE_TXT(a, b) { .mask = BM_EIRQ_##a, .txt = b }
-+
-+static const struct bman_hwerr_txt bman_hwerr_txts[] = {
-+ BMAN_HWE_TXT(IVCI, "Invalid Command Verb"),
-+ BMAN_HWE_TXT(FLWI, "FBPR Low Watermark"),
-+ BMAN_HWE_TXT(MBEI, "Multi-bit ECC Error"),
-+ BMAN_HWE_TXT(SBEI, "Single-bit ECC Error"),
-+ BMAN_HWE_TXT(BSCN, "Pool State Change Notification"),
-+};
-+#define BMAN_HWE_COUNT (sizeof(bman_hwerr_txts)/sizeof(struct bman_hwerr_txt))
-+
-+struct bman_error_info_mdata {
-+ u16 addr_mask;
-+ u16 bits;
-+ const char *txt;
-+};
-+
-+#define BMAN_ERR_MDATA(a, b, c) { .addr_mask = a, .bits = b, .txt = c}
-+static const struct bman_error_info_mdata error_mdata[] = {
-+ BMAN_ERR_MDATA(0x03FF, 192, "Stockpile memory"),
-+ BMAN_ERR_MDATA(0x00FF, 256, "SW portal ring memory port 1"),
-+ BMAN_ERR_MDATA(0x00FF, 256, "SW portal ring memory port 2"),
-+};
-+#define BMAN_ERR_MDATA_COUNT \
-+ (sizeof(error_mdata)/sizeof(struct bman_error_info_mdata))
-+
-+/* Add this in Kconfig */
-+#define BMAN_ERRS_TO_UNENABLE (BM_EIRQ_FLWI)
-+
-+/**
-+ * bm_err_isr_<reg>_<verb> - Manipulate global interrupt registers
-+ * @v: for accessors that write values, this is the 32-bit value
-+ *
-+ * Manipulates BMAN_ERR_ISR, BMAN_ERR_IER, BMAN_ERR_ISDR, BMAN_ERR_IIR. All
-+ * manipulations except bm_err_isr_[un]inhibit() use 32-bit masks composed of
-+ * the BM_EIRQ_*** definitions. Note that "bm_err_isr_enable_write" means
-+ * "write the enable register" rather than "enable the write register"!
-+ */
-+#define bm_err_isr_status_read(bm) \
-+ __bm_err_isr_read(bm, bm_isr_status)
-+#define bm_err_isr_status_clear(bm, m) \
-+ __bm_err_isr_write(bm, bm_isr_status, m)
-+#define bm_err_isr_enable_read(bm) \
-+ __bm_err_isr_read(bm, bm_isr_enable)
-+#define bm_err_isr_enable_write(bm, v) \
-+ __bm_err_isr_write(bm, bm_isr_enable, v)
-+#define bm_err_isr_disable_read(bm) \
-+ __bm_err_isr_read(bm, bm_isr_disable)
-+#define bm_err_isr_disable_write(bm, v) \
-+ __bm_err_isr_write(bm, bm_isr_disable, v)
-+#define bm_err_isr_inhibit(bm) \
-+ __bm_err_isr_write(bm, bm_isr_inhibit, 1)
-+#define bm_err_isr_uninhibit(bm) \
-+ __bm_err_isr_write(bm, bm_isr_inhibit, 0)
-+
-+/*
-+ * TODO: unimplemented registers
-+ *
-+ * BMAN_POOLk_SDCNT, BMAN_POOLk_HDCNT, BMAN_FULT,
-+ * BMAN_VLDPL, BMAN_EECC, BMAN_SBET, BMAN_EINJ
-+ */
-+
-+/* Encapsulate "struct bman *" as a cast of the register space address. */
-+
-+static struct bman *bm_create(void *regs)
-+{
-+ return (struct bman *)regs;
-+}
-+
-+static inline u32 __bm_in(struct bman *bm, u32 offset)
-+{
-+ return in_be32((void *)bm + offset);
-+}
-+static inline void __bm_out(struct bman *bm, u32 offset, u32 val)
-+{
-+ out_be32((void *)bm + offset, val);
-+}
-+#define bm_in(reg) __bm_in(bm, REG_##reg)
-+#define bm_out(reg, val) __bm_out(bm, REG_##reg, val)
-+
-+static u32 __bm_err_isr_read(struct bman *bm, enum bm_isr_reg n)
-+{
-+ return __bm_in(bm, REG_ERR_ISR + (n << 2));
-+}
-+
-+static void __bm_err_isr_write(struct bman *bm, enum bm_isr_reg n, u32 val)
-+{
-+ __bm_out(bm, REG_ERR_ISR + (n << 2), val);
-+}
-+
-+static void bm_get_version(struct bman *bm, u16 *id, u8 *major, u8 *minor)
-+{
-+ u32 v = bm_in(IP_REV_1);
-+ *id = (v >> 16);
-+ *major = (v >> 8) & 0xff;
-+ *minor = v & 0xff;
-+}
-+
-+static u32 __generate_thresh(u32 val, int roundup)
-+{
-+ u32 e = 0; /* co-efficient, exponent */
-+ int oddbit = 0;
-+ while (val > 0xff) {
-+ oddbit = val & 1;
-+ val >>= 1;
-+ e++;
-+ if (roundup && oddbit)
-+ val++;
-+ }
-+ DPA_ASSERT(e < 0x10);
-+ return val | (e << 8);
-+}
-+
-+static void bm_set_pool(struct bman *bm, u8 pool, u32 swdet, u32 swdxt,
-+ u32 hwdet, u32 hwdxt)
-+{
-+ DPA_ASSERT(pool < bman_pool_max);
-+ bm_out(POOL_SWDET(pool), __generate_thresh(swdet, 0));
-+ bm_out(POOL_SWDXT(pool), __generate_thresh(swdxt, 1));
-+ bm_out(POOL_HWDET(pool), __generate_thresh(hwdet, 0));
-+ bm_out(POOL_HWDXT(pool), __generate_thresh(hwdxt, 1));
-+}
-+
-+static void bm_set_memory(struct bman *bm, u64 ba, int prio, u32 size)
-+{
-+ u32 exp = ilog2(size);
-+ /* choke if size isn't within range */
-+ DPA_ASSERT((size >= 4096) && (size <= 1073741824) &&
-+ is_power_of_2(size));
-+ /* choke if '[e]ba' has lower-alignment than 'size' */
-+ DPA_ASSERT(!(ba & (size - 1)));
-+ bm_out(FBPR_BARE, upper_32_bits(ba));
-+ bm_out(FBPR_BAR, lower_32_bits(ba));
-+ bm_out(FBPR_AR, (prio ? 0x40000000 : 0) | (exp - 1));
-+}
-+
-+/*****************/
-+/* Config driver */
-+/*****************/
-+
-+/* TODO: Kconfig these? */
-+#define DEFAULT_FBPR_SZ (PAGE_SIZE << 12)
-+
-+/* We support only one of these. */
-+static struct bman *bm;
-+static struct device_node *bm_node;
-+
-+/* And this state belongs to 'bm'. It is set during fsl_bman_init(), but used
-+ * during bman_init_ccsr(). */
-+static dma_addr_t fbpr_a;
-+static size_t fbpr_sz = DEFAULT_FBPR_SZ;
-+
-+static int bman_fbpr(struct reserved_mem *rmem)
-+{
-+ fbpr_a = rmem->base;
-+ fbpr_sz = rmem->size;
-+
-+ WARN_ON(!(fbpr_a && fbpr_sz));
-+
-+ return 0;
-+}
-+RESERVEDMEM_OF_DECLARE(bman_fbpr, "fsl,bman-fbpr", bman_fbpr);
-+
-+static int __init fsl_bman_init(struct device_node *node)
-+{
-+ struct resource res;
-+ u32 __iomem *regs;
-+ const char *s;
-+ int ret, standby = 0;
-+ u16 id;
-+ u8 major, minor;
-+
-+ ret = of_address_to_resource(node, 0, &res);
-+ if (ret) {
-+ pr_err("Can't get %s property 'reg'\n",
-+ node->full_name);
-+ return ret;
-+ }
-+ s = of_get_property(node, "fsl,hv-claimable", &ret);
-+ if (s && !strcmp(s, "standby"))
-+ standby = 1;
-+ /* Global configuration */
-+ regs = ioremap(res.start, res.end - res.start + 1);
-+ bm = bm_create(regs);
-+ BUG_ON(!bm);
-+ bm_node = node;
-+ bm_get_version(bm, &id, &major, &minor);
-+ pr_info("Bman ver:%04x,%02x,%02x\n", id, major, minor);
-+ if ((major == 1) && (minor == 0)) {
-+ bman_ip_rev = BMAN_REV10;
-+ bman_pool_max = 64;
-+ } else if ((major == 2) && (minor == 0)) {
-+ bman_ip_rev = BMAN_REV20;
-+ bman_pool_max = 8;
-+ } else if ((major == 2) && (minor == 1)) {
-+ bman_ip_rev = BMAN_REV21;
-+ bman_pool_max = 64;
-+ } else {
-+ pr_warn("unknown Bman version, default to rev1.0\n");
-+ }
-+
-+ if (standby) {
-+ pr_info(" -> in standby mode\n");
-+ return 0;
-+ }
-+ return 0;
-+}
-+
-+int bman_have_ccsr(void)
-+{
-+ return bm ? 1 : 0;
-+}
-+
-+int bm_pool_set(u32 bpid, const u32 *thresholds)
-+{
-+ if (!bm)
-+ return -ENODEV;
-+ bm_set_pool(bm, bpid, thresholds[0],
-+ thresholds[1], thresholds[2],
-+ thresholds[3]);
-+ return 0;
-+}
-+EXPORT_SYMBOL(bm_pool_set);
-+
-+__init int bman_init_early(void)
-+{
-+ struct device_node *dn;
-+ int ret;
-+
-+ for_each_compatible_node(dn, NULL, "fsl,bman") {
-+ if (bm)
-+ pr_err("%s: only one 'fsl,bman' allowed\n",
-+ dn->full_name);
-+ else {
-+ if (!of_device_is_available(dn))
-+ continue;
-+
-+ ret = fsl_bman_init(dn);
-+ BUG_ON(ret);
-+ }
-+ }
-+ return 0;
-+}
-+postcore_initcall_sync(bman_init_early);
-+
-+
-+static void log_edata_bits(u32 bit_count)
-+{
-+ u32 i, j, mask = 0xffffffff;
-+
-+ pr_warn("Bman ErrInt, EDATA:\n");
-+ i = bit_count/32;
-+ if (bit_count%32) {
-+ i++;
-+ mask = ~(mask << bit_count%32);
-+ }
-+ j = 16-i;
-+ pr_warn(" 0x%08x\n", bm_in(EDATA(j)) & mask);
-+ j++;
-+ for (; j < 16; j++)
-+ pr_warn(" 0x%08x\n", bm_in(EDATA(j)));
-+}
-+
-+static void log_additional_error_info(u32 isr_val, u32 ecsr_val)
-+{
-+ union bman_ecir ecir_val;
-+ union bman_eadr eadr_val;
-+
-+ ecir_val.ecir_raw = bm_in(ECIR);
-+ /* Is portal info valid */
-+ if (ecsr_val & PORTAL_ECSR_ERR) {
-+ pr_warn("Bman ErrInt: SWP id %d, numb %d, pid %d\n",
-+ ecir_val.info.portal_num, ecir_val.info.numb,
-+ ecir_val.info.pid);
-+ }
-+ if (ecsr_val & (BM_EIRQ_SBEI|BM_EIRQ_MBEI)) {
-+ eadr_val.eadr_raw = bm_in(EADR);
-+ pr_warn("Bman ErrInt: EADR Memory: %s, 0x%x\n",
-+ error_mdata[eadr_val.info.memid].txt,
-+ error_mdata[eadr_val.info.memid].addr_mask
-+ & eadr_val.info.eadr);
-+ log_edata_bits(error_mdata[eadr_val.info.memid].bits);
-+ }
-+}
-+
-+/* Bman interrupt handler */
-+static irqreturn_t bman_isr(int irq, void *ptr)
-+{
-+ u32 isr_val, ier_val, ecsr_val, isr_mask, i;
-+
-+ ier_val = bm_err_isr_enable_read(bm);
-+ isr_val = bm_err_isr_status_read(bm);
-+ ecsr_val = bm_in(ECSR);
-+ isr_mask = isr_val & ier_val;
-+
-+ if (!isr_mask)
-+ return IRQ_NONE;
-+ for (i = 0; i < BMAN_HWE_COUNT; i++) {
-+ if (bman_hwerr_txts[i].mask & isr_mask) {
-+ pr_warn("Bman ErrInt: %s\n", bman_hwerr_txts[i].txt);
-+ if (bman_hwerr_txts[i].mask & ecsr_val) {
-+ log_additional_error_info(isr_mask, ecsr_val);
-+ /* Re-arm error capture registers */
-+ bm_out(ECSR, ecsr_val);
-+ }
-+ if (bman_hwerr_txts[i].mask & BMAN_ERRS_TO_UNENABLE) {
-+ pr_devel("Bman un-enabling error 0x%x\n",
-+ bman_hwerr_txts[i].mask);
-+ ier_val &= ~bman_hwerr_txts[i].mask;
-+ bm_err_isr_enable_write(bm, ier_val);
-+ }
-+ }
-+ }
-+ bm_err_isr_status_clear(bm, isr_val);
-+ return IRQ_HANDLED;
-+}
-+
-+static int __bind_irq(void)
-+{
-+ int ret, err_irq;
-+
-+ err_irq = of_irq_to_resource(bm_node, 0, NULL);
-+ if (err_irq == 0) {
-+ pr_info("Can't get %s property '%s'\n", bm_node->full_name,
-+ "interrupts");
-+ return -ENODEV;
-+ }
-+ ret = request_irq(err_irq, bman_isr, IRQF_SHARED, "bman-err", bm_node);
-+ if (ret) {
-+ pr_err("request_irq() failed %d for '%s'\n", ret,
-+ bm_node->full_name);
-+ return -ENODEV;
-+ }
-+ /* Disable Buffer Pool State Change */
-+ bm_err_isr_disable_write(bm, BM_EIRQ_BSCN);
-+ /* Write-to-clear any stale bits, (eg. starvation being asserted prior
-+ * to resource allocation during driver init). */
-+ bm_err_isr_status_clear(bm, 0xffffffff);
-+ /* Enable Error Interrupts */
-+ bm_err_isr_enable_write(bm, 0xffffffff);
-+ return 0;
-+}
-+
-+int bman_init_ccsr(struct device_node *node)
-+{
-+ int ret;
-+ if (!bman_have_ccsr())
-+ return 0;
-+ if (node != bm_node)
-+ return -EINVAL;
-+ /* FBPR memory */
-+ bm_set_memory(bm, fbpr_a, 0, fbpr_sz);
-+ pr_info("bman-fbpr addr %pad size 0x%zx\n", &fbpr_a, fbpr_sz);
-+
-+ ret = __bind_irq();
-+ if (ret)
-+ return ret;
-+ return 0;
-+}
-+
-+u32 bm_pool_free_buffers(u32 bpid)
-+{
-+ return bm_in(POOL_CONTENT(bpid));
-+}
-+
-+#ifdef CONFIG_SYSFS
-+
-+#define DRV_NAME "fsl-bman"
-+#define SBEC_MAX_ID 1
-+#define SBEC_MIN_ID 0
-+
-+static ssize_t show_fbpr_fpc(struct device *dev,
-+ struct device_attribute *dev_attr, char *buf)
-+{
-+ return snprintf(buf, PAGE_SIZE, "%u\n", bm_in(FBPR_FPC));
-+};
-+
-+static ssize_t show_pool_count(struct device *dev,
-+ struct device_attribute *dev_attr, char *buf)
-+{
-+ u32 data;
-+ int i;
-+
-+ if (!sscanf(dev_attr->attr.name, "%d", &i) || (i >= bman_pool_max))
-+ return -EINVAL;
-+ data = bm_in(POOL_CONTENT(i));
-+ return snprintf(buf, PAGE_SIZE, "%d\n", data);
-+};
-+
-+static ssize_t show_err_isr(struct device *dev,
-+ struct device_attribute *dev_attr, char *buf)
-+{
-+ return snprintf(buf, PAGE_SIZE, "0x%08x\n", bm_in(ERR_ISR));
-+};
-+
-+static ssize_t show_sbec(struct device *dev,
-+ struct device_attribute *dev_attr, char *buf)
-+{
-+ int i;
-+
-+ if (!sscanf(dev_attr->attr.name, "sbec_%d", &i))
-+ return -EINVAL;
-+ if (i < SBEC_MIN_ID || i > SBEC_MAX_ID)
-+ return -EINVAL;
-+ return snprintf(buf, PAGE_SIZE, "%u\n", bm_in(SBEC(i)));
-+};
-+
-+static DEVICE_ATTR(err_isr, S_IRUSR, show_err_isr, NULL);
-+static DEVICE_ATTR(fbpr_fpc, S_IRUSR, show_fbpr_fpc, NULL);
-+
-+/* Didn't use DEVICE_ATTR as 64 of this would be required.
-+ * Initialize them when needed. */
-+static char *name_attrs_pool_count; /* "xx" + null-terminator */
-+static struct device_attribute *dev_attr_buffer_pool_count;
-+
-+static DEVICE_ATTR(sbec_0, S_IRUSR, show_sbec, NULL);
-+static DEVICE_ATTR(sbec_1, S_IRUSR, show_sbec, NULL);
-+
-+static struct attribute *bman_dev_attributes[] = {
-+ &dev_attr_fbpr_fpc.attr,
-+ &dev_attr_err_isr.attr,
-+ NULL
-+};
-+
-+static struct attribute *bman_dev_ecr_attributes[] = {
-+ &dev_attr_sbec_0.attr,
-+ &dev_attr_sbec_1.attr,
-+ NULL
-+};
-+
-+static struct attribute **bman_dev_pool_count_attributes;
-+
-+
-+/* root level */
-+static const struct attribute_group bman_dev_attr_grp = {
-+ .name = NULL,
-+ .attrs = bman_dev_attributes
-+};
-+static const struct attribute_group bman_dev_ecr_grp = {
-+ .name = "error_capture",
-+ .attrs = bman_dev_ecr_attributes
-+};
-+static struct attribute_group bman_dev_pool_countent_grp = {
-+ .name = "pool_count",
-+};
-+
-+static int of_fsl_bman_remove(struct platform_device *ofdev)
-+{
-+ sysfs_remove_group(&ofdev->dev.kobj, &bman_dev_attr_grp);
-+ return 0;
-+};
-+
-+static int of_fsl_bman_probe(struct platform_device *ofdev)
-+{
-+ int ret, i;
-+
-+ ret = sysfs_create_group(&ofdev->dev.kobj, &bman_dev_attr_grp);
-+ if (ret)
-+ goto done;
-+ ret = sysfs_create_group(&ofdev->dev.kobj, &bman_dev_ecr_grp);
-+ if (ret)
-+ goto del_group_0;
-+
-+ name_attrs_pool_count = kmalloc(sizeof(char) * bman_pool_max * 3,
-+ GFP_KERNEL);
-+ if (!name_attrs_pool_count) {
-+ pr_err("Can't alloc name_attrs_pool_count\n");
-+ goto del_group_1;
-+ }
-+
-+ dev_attr_buffer_pool_count = kmalloc(sizeof(struct device_attribute) *
-+ bman_pool_max, GFP_KERNEL);
-+ if (!dev_attr_buffer_pool_count) {
-+ pr_err("Can't alloc dev_attr-buffer_pool_count\n");
-+ goto del_group_2;
-+ }
-+
-+ bman_dev_pool_count_attributes = kmalloc(sizeof(struct attribute *) *
-+ (bman_pool_max + 1), GFP_KERNEL);
-+ if (!bman_dev_pool_count_attributes) {
-+ pr_err("can't alloc bman_dev_pool_count_attributes\n");
-+ goto del_group_3;
-+ }
-+
-+ for (i = 0; i < bman_pool_max; i++) {
-+ ret = scnprintf((name_attrs_pool_count + i * 3), 3, "%d", i);
-+ if (!ret)
-+ goto del_group_4;
-+ dev_attr_buffer_pool_count[i].attr.name =
-+ (name_attrs_pool_count + i * 3);
-+ dev_attr_buffer_pool_count[i].attr.mode = S_IRUSR;
-+ dev_attr_buffer_pool_count[i].show = show_pool_count;
-+ bman_dev_pool_count_attributes[i] =
-+ &dev_attr_buffer_pool_count[i].attr;
-+ sysfs_attr_init(bman_dev_pool_count_attributes[i]);
-+ }
-+ bman_dev_pool_count_attributes[bman_pool_max] = NULL;
-+
-+ bman_dev_pool_countent_grp.attrs = bman_dev_pool_count_attributes;
-+
-+ ret = sysfs_create_group(&ofdev->dev.kobj, &bman_dev_pool_countent_grp);
-+ if (ret)
-+ goto del_group_4;
-+
-+ goto done;
-+
-+del_group_4:
-+ kfree(bman_dev_pool_count_attributes);
-+del_group_3:
-+ kfree(dev_attr_buffer_pool_count);
-+del_group_2:
-+ kfree(name_attrs_pool_count);
-+del_group_1:
-+ sysfs_remove_group(&ofdev->dev.kobj, &bman_dev_ecr_grp);
-+del_group_0:
-+ sysfs_remove_group(&ofdev->dev.kobj, &bman_dev_attr_grp);
-+done:
-+ if (ret)
-+ dev_err(&ofdev->dev,
-+ "Cannot create dev attributes ret=%d\n", ret);
-+ return ret;
-+};
-+
-+static struct of_device_id of_fsl_bman_ids[] = {
-+ {
-+ .compatible = "fsl,bman",
-+ },
-+ {}
-+};
-+MODULE_DEVICE_TABLE(of, of_fsl_bman_ids);
-+
-+#ifdef CONFIG_SUSPEND
-+static u32 saved_isdr;
-+
-+static int bman_pm_suspend_noirq(struct device *dev)
-+{
-+ uint32_t idle_state;
-+
-+ suspend_unused_bportal();
-+ /* save isdr, disable all, clear isr */
-+ saved_isdr = bm_err_isr_disable_read(bm);
-+ bm_err_isr_disable_write(bm, 0xffffffff);
-+ bm_err_isr_status_clear(bm, 0xffffffff);
-+
-+ if (bman_ip_rev < BMAN_REV21) {
-+#ifdef CONFIG_PM_DEBUG
-+ pr_info("Bman version doesn't have STATE_IDLE\n");
-+#endif
-+ return 0;
-+ }
-+ idle_state = bm_in(STATE_IDLE);
-+ if (!(idle_state & 0x1)) {
-+ pr_err("Bman not idle 0x%x aborting\n", idle_state);
-+ bm_err_isr_disable_write(bm, saved_isdr);
-+ resume_unused_bportal();
-+ return -EBUSY;
-+ }
-+#ifdef CONFIG_PM_DEBUG
-+ pr_info("Bman suspend code, IDLE_STAT = 0x%x\n", idle_state);
-+#endif
-+ return 0;
-+}
-+
-+static int bman_pm_resume_noirq(struct device *dev)
-+{
-+ /* restore isdr */
-+ bm_err_isr_disable_write(bm, saved_isdr);
-+ resume_unused_bportal();
-+ return 0;
-+}
-+#else
-+#define bman_pm_suspend_noirq NULL
-+#define bman_pm_resume_noirq NULL
-+#endif
-+
-+static const struct dev_pm_ops bman_pm_ops = {
-+ .suspend_noirq = bman_pm_suspend_noirq,
-+ .resume_noirq = bman_pm_resume_noirq,
-+};
-+
-+static struct platform_driver of_fsl_bman_driver = {
-+ .driver = {
-+ .owner = THIS_MODULE,
-+ .name = DRV_NAME,
-+ .of_match_table = of_fsl_bman_ids,
-+ .pm = &bman_pm_ops,
-+ },
-+ .probe = of_fsl_bman_probe,
-+ .remove = of_fsl_bman_remove,
-+};
-+
-+static int bman_ctrl_init(void)
-+{
-+ return platform_driver_register(&of_fsl_bman_driver);
-+}
-+
-+static void bman_ctrl_exit(void)
-+{
-+ platform_driver_unregister(&of_fsl_bman_driver);
-+}
-+
-+module_init(bman_ctrl_init);
-+module_exit(bman_ctrl_exit);
-+
-+#endif /* CONFIG_SYSFS */
---- /dev/null
-+++ b/drivers/staging/fsl_qbman/bman_debugfs.c
-@@ -0,0 +1,119 @@
-+/* Copyright 2010-2011 Freescale Semiconductor, Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+#include <linux/module.h>
-+#include <linux/fsl_bman.h>
-+#include <linux/debugfs.h>
-+#include <linux/seq_file.h>
-+#include <linux/uaccess.h>
-+
-+static struct dentry *dfs_root; /* debugfs root directory */
-+
-+/*******************************************************************************
-+ * Query Buffer Pool State
-+ ******************************************************************************/
-+static int query_bp_state_show(struct seq_file *file, void *offset)
-+{
-+ int ret;
-+ struct bm_pool_state state;
-+ int i, j;
-+ u32 mask;
-+
-+ memset(&state, 0, sizeof(struct bm_pool_state));
-+ ret = bman_query_pools(&state);
-+ if (ret) {
-+ seq_printf(file, "Error %d\n", ret);
-+ return 0;
-+ }
-+ seq_puts(file, "bp_id free_buffers_avail bp_depleted\n");
-+ for (i = 0; i < 2; i++) {
-+ mask = 0x80000000;
-+ for (j = 0; j < 32; j++) {
-+ seq_printf(file,
-+ " %-2u %-3s %-3s\n",
-+ (i*32)+j,
-+ (state.as.state.__state[i] & mask) ? "no" : "yes",
-+ (state.ds.state.__state[i] & mask) ? "yes" : "no");
-+ mask >>= 1;
-+ }
-+ }
-+ return 0;
-+}
-+
-+static int query_bp_state_open(struct inode *inode, struct file *file)
-+{
-+ return single_open(file, query_bp_state_show, NULL);
-+}
-+
-+static const struct file_operations query_bp_state_fops = {
-+ .owner = THIS_MODULE,
-+ .open = query_bp_state_open,
-+ .read = seq_read,
-+ .release = single_release,
-+};
-+
-+static int __init bman_debugfs_module_init(void)
-+{
-+ int ret = 0;
-+ struct dentry *d;
-+
-+ dfs_root = debugfs_create_dir("bman", NULL);
-+
-+ if (dfs_root == NULL) {
-+ ret = -ENOMEM;
-+ pr_err("Cannot create bman debugfs dir\n");
-+ goto _return;
-+ }
-+ d = debugfs_create_file("query_bp_state",
-+ S_IRUGO,
-+ dfs_root,
-+ NULL,
-+ &query_bp_state_fops);
-+ if (d == NULL) {
-+ ret = -ENOMEM;
-+ pr_err("Cannot create query_bp_state\n");
-+ goto _return;
-+ }
-+ return 0;
-+
-+_return:
-+ debugfs_remove_recursive(dfs_root);
-+ return ret;
-+}
-+
-+static void __exit bman_debugfs_module_exit(void)
-+{
-+ debugfs_remove_recursive(dfs_root);
-+}
-+
-+
-+module_init(bman_debugfs_module_init);
-+module_exit(bman_debugfs_module_exit);
-+MODULE_LICENSE("Dual BSD/GPL");
---- /dev/null
-+++ b/drivers/staging/fsl_qbman/bman_driver.c
-@@ -0,0 +1,575 @@
-+/* Copyright 2008-2012 Freescale Semiconductor, Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+#include "bman_low.h"
-+#ifdef CONFIG_HOTPLUG_CPU
-+#include <linux/cpu.h>
-+#endif
-+/*
-+ * Global variables of the max portal/pool number this bman version supported
-+ */
-+u16 bman_ip_rev;
-+EXPORT_SYMBOL(bman_ip_rev);
-+u16 bman_pool_max;
-+EXPORT_SYMBOL(bman_pool_max);
-+static u16 bman_portal_max;
-+
-+/* After initialising cpus that own shared portal configs, we cache the
-+ * resulting portals (ie. not just the configs) in this array. Then we
-+ * initialise slave cpus that don't have their own portals, redirecting them to
-+ * portals from this cache in a round-robin assignment. */
-+static struct bman_portal *shared_portals[NR_CPUS];
-+static int num_shared_portals;
-+static int shared_portals_idx;
-+static LIST_HEAD(unused_pcfgs);
-+static DEFINE_SPINLOCK(unused_pcfgs_lock);
-+static void *affine_bportals[NR_CPUS];
-+
-+static int __init fsl_bpool_init(struct device_node *node)
-+{
-+ int ret;
-+ u32 *thresh, *bpid = (u32 *)of_get_property(node, "fsl,bpid", &ret);
-+ if (!bpid || (ret != 4)) {
-+ pr_err("Can't get %s property 'fsl,bpid'\n", node->full_name);
-+ return -ENODEV;
-+ }
-+ thresh = (u32 *)of_get_property(node, "fsl,bpool-thresholds", &ret);
-+ if (thresh) {
-+ if (ret != 16) {
-+ pr_err("Invalid %s property '%s'\n",
-+ node->full_name, "fsl,bpool-thresholds");
-+ return -ENODEV;
-+ }
-+ }
-+ if (thresh) {
-+#ifdef CONFIG_FSL_BMAN_CONFIG
-+ ret = bm_pool_set(be32_to_cpu(*bpid), thresh);
-+ if (ret)
-+ pr_err("No CCSR node for %s property '%s'\n",
-+ node->full_name, "fsl,bpool-thresholds");
-+ return ret;
-+#else
-+ pr_err("Ignoring %s property '%s', no CCSR support\n",
-+ node->full_name, "fsl,bpool-thresholds");
-+#endif
-+ }
-+ return 0;
-+}
-+
-+static int __init fsl_bpid_range_init(struct device_node *node)
-+{
-+ int ret;
-+ u32 *range = (u32 *)of_get_property(node, "fsl,bpid-range", &ret);
-+ if (!range) {
-+ pr_err("No 'fsl,bpid-range' property in node %s\n",
-+ node->full_name);
-+ return -EINVAL;
-+ }
-+ if (ret != 8) {
-+ pr_err("'fsl,bpid-range' is not a 2-cell range in node %s\n",
-+ node->full_name);
-+ return -EINVAL;
-+ }
-+ bman_seed_bpid_range(be32_to_cpu(range[0]), be32_to_cpu(range[1]));
-+ pr_info("Bman: BPID allocator includes range %d:%d\n",
-+ be32_to_cpu(range[0]), be32_to_cpu(range[1]));
-+ return 0;
-+}
-+
-+static struct bm_portal_config * __init parse_pcfg(struct device_node *node)
-+{
-+ struct bm_portal_config *pcfg;
-+ const u32 *index;
-+ int irq, ret;
-+ resource_size_t len;
-+
-+ pcfg = kmalloc(sizeof(*pcfg), GFP_KERNEL);
-+ if (!pcfg) {
-+ pr_err("can't allocate portal config");
-+ return NULL;
-+ }
-+
-+ if (of_device_is_compatible(node, "fsl,bman-portal-1.0") ||
-+ of_device_is_compatible(node, "fsl,bman-portal-1.0.0")) {
-+ bman_ip_rev = BMAN_REV10;
-+ bman_pool_max = 64;
-+ bman_portal_max = 10;
-+ } else if (of_device_is_compatible(node, "fsl,bman-portal-2.0") ||
-+ of_device_is_compatible(node, "fsl,bman-portal-2.0.8")) {
-+ bman_ip_rev = BMAN_REV20;
-+ bman_pool_max = 8;
-+ bman_portal_max = 3;
-+ } else if (of_device_is_compatible(node, "fsl,bman-portal-2.1.0")) {
-+ bman_ip_rev = BMAN_REV21;
-+ bman_pool_max = 64;
-+ bman_portal_max = 50;
-+ } else if (of_device_is_compatible(node, "fsl,bman-portal-2.1.1")) {
-+ bman_ip_rev = BMAN_REV21;
-+ bman_pool_max = 64;
-+ bman_portal_max = 25;
-+ } else if (of_device_is_compatible(node, "fsl,bman-portal-2.1.2")) {
-+ bman_ip_rev = BMAN_REV21;
-+ bman_pool_max = 64;
-+ bman_portal_max = 18;
-+ } else if (of_device_is_compatible(node, "fsl,bman-portal-2.1.3")) {
-+ bman_ip_rev = BMAN_REV21;
-+ bman_pool_max = 64;
-+ bman_portal_max = 10;
-+ } else {
-+ pr_warn("unknown BMan version in portal node,"
-+ "default to rev1.0\n");
-+ bman_ip_rev = BMAN_REV10;
-+ bman_pool_max = 64;
-+ bman_portal_max = 10;
-+ }
-+
-+ ret = of_address_to_resource(node, DPA_PORTAL_CE,
-+ &pcfg->addr_phys[DPA_PORTAL_CE]);
-+ if (ret) {
-+ pr_err("Can't get %s property 'reg::CE'\n", node->full_name);
-+ goto err;
-+ }
-+ ret = of_address_to_resource(node, DPA_PORTAL_CI,
-+ &pcfg->addr_phys[DPA_PORTAL_CI]);
-+ if (ret) {
-+ pr_err("Can't get %s property 'reg::CI'\n", node->full_name);
-+ goto err;
-+ }
-+
-+ index = of_get_property(node, "cell-index", &ret);
-+ if (!index || (ret != 4)) {
-+ pr_err("Can't get %s property '%s'\n", node->full_name,
-+ "cell-index");
-+ goto err;
-+ }
-+ if (be32_to_cpu(*index) >= bman_portal_max) {
-+ pr_err("BMan portal cell index %d out of range, max %d\n",
-+ be32_to_cpu(*index), bman_portal_max);
-+ goto err;
-+ }
-+
-+ pcfg->public_cfg.cpu = -1;
-+
-+ irq = irq_of_parse_and_map(node, 0);
-+ if (irq == 0) {
-+ pr_err("Can't get %s property 'interrupts'\n", node->full_name);
-+ goto err;
-+ }
-+ pcfg->public_cfg.irq = irq;
-+ pcfg->public_cfg.index = be32_to_cpu(*index);
-+ bman_depletion_fill(&pcfg->public_cfg.mask);
-+
-+ len = resource_size(&pcfg->addr_phys[DPA_PORTAL_CE]);
-+ if (len != (unsigned long)len)
-+ goto err;
-+
-+#if defined(CONFIG_ARM) || defined(CONFIG_ARM64)
-+ pcfg->addr_virt[DPA_PORTAL_CE] = ioremap_cache_ns(
-+ pcfg->addr_phys[DPA_PORTAL_CE].start,
-+ resource_size(&pcfg->addr_phys[DPA_PORTAL_CE]));
-+ pcfg->addr_virt[DPA_PORTAL_CI] = ioremap(
-+ pcfg->addr_phys[DPA_PORTAL_CI].start,
-+ resource_size(&pcfg->addr_phys[DPA_PORTAL_CI]));
-+
-+#else
-+ pcfg->addr_virt[DPA_PORTAL_CE] = ioremap_prot(
-+ pcfg->addr_phys[DPA_PORTAL_CE].start,
-+ (unsigned long)len,
-+ 0);
-+ pcfg->addr_virt[DPA_PORTAL_CI] = ioremap_prot(
-+ pcfg->addr_phys[DPA_PORTAL_CI].start,
-+ resource_size(&pcfg->addr_phys[DPA_PORTAL_CI]),
-+ _PAGE_GUARDED | _PAGE_NO_CACHE);
-+#endif
-+ /* disable bp depletion */
-+ __raw_writel(0x0, pcfg->addr_virt[DPA_PORTAL_CI] + BM_REG_SCN(0));
-+ __raw_writel(0x0, pcfg->addr_virt[DPA_PORTAL_CI] + BM_REG_SCN(1));
-+ return pcfg;
-+err:
-+ kfree(pcfg);
-+ return NULL;
-+}
-+
-+static struct bm_portal_config *get_pcfg(struct list_head *list)
-+{
-+ struct bm_portal_config *pcfg;
-+ if (list_empty(list))
-+ return NULL;
-+ pcfg = list_entry(list->prev, struct bm_portal_config, list);
-+ list_del(&pcfg->list);
-+ return pcfg;
-+}
-+
-+static struct bm_portal_config *get_pcfg_idx(struct list_head *list,
-+ uint32_t idx)
-+{
-+ struct bm_portal_config *pcfg;
-+ if (list_empty(list))
-+ return NULL;
-+ list_for_each_entry(pcfg, list, list) {
-+ if (pcfg->public_cfg.index == idx) {
-+ list_del(&pcfg->list);
-+ return pcfg;
-+ }
-+ }
-+ return NULL;
-+}
-+
-+struct bm_portal_config *bm_get_unused_portal(void)
-+{
-+ return bm_get_unused_portal_idx(QBMAN_ANY_PORTAL_IDX);
-+}
-+
-+struct bm_portal_config *bm_get_unused_portal_idx(uint32_t idx)
-+{
-+ struct bm_portal_config *ret;
-+ spin_lock(&unused_pcfgs_lock);
-+ if (idx == QBMAN_ANY_PORTAL_IDX)
-+ ret = get_pcfg(&unused_pcfgs);
-+ else
-+ ret = get_pcfg_idx(&unused_pcfgs, idx);
-+ spin_unlock(&unused_pcfgs_lock);
-+ return ret;
-+}
-+
-+void bm_put_unused_portal(struct bm_portal_config *pcfg)
-+{
-+ spin_lock(&unused_pcfgs_lock);
-+ list_add(&pcfg->list, &unused_pcfgs);
-+ spin_unlock(&unused_pcfgs_lock);
-+}
-+
-+static struct bman_portal *init_pcfg(struct bm_portal_config *pcfg)
-+{
-+ struct bman_portal *p;
-+ p = bman_create_affine_portal(pcfg);
-+ if (p) {
-+#ifdef CONFIG_FSL_DPA_PIRQ_SLOW
-+ bman_p_irqsource_add(p, BM_PIRQ_RCRI | BM_PIRQ_BSCN);
-+#endif
-+ pr_info("Bman portal %sinitialised, cpu %d\n",
-+ pcfg->public_cfg.is_shared ? "(shared) " : "",
-+ pcfg->public_cfg.cpu);
-+ affine_bportals[pcfg->public_cfg.cpu] = p;
-+ } else
-+ pr_crit("Bman portal failure on cpu %d\n",
-+ pcfg->public_cfg.cpu);
-+ return p;
-+}
-+
-+static void init_slave(int cpu)
-+{
-+ struct bman_portal *p;
-+ p = bman_create_affine_slave(shared_portals[shared_portals_idx++], cpu);
-+ if (!p)
-+ pr_err("Bman slave portal failure on cpu %d\n", cpu);
-+ else
-+ pr_info("Bman portal %sinitialised, cpu %d\n", "(slave) ", cpu);
-+ if (shared_portals_idx >= num_shared_portals)
-+ shared_portals_idx = 0;
-+ affine_bportals[cpu] = p;
-+}
-+
-+/* Bootarg "bportals=[...]" has the same syntax as "qportals=", and so the
-+ * parsing is in dpa_sys.h. The syntax is a comma-separated list of indexes
-+ * and/or ranges of indexes, with each being optionally prefixed by "s" to
-+ * explicitly mark it or them for sharing.
-+ * Eg;
-+ * bportals=s0,1-3,s4
-+ * means that cpus 1,2,3 get "unshared" portals, cpus 0 and 4 get "shared"
-+ * portals, and any remaining cpus share the portals that are assigned to cpus 0
-+ * or 4, selected in a round-robin fashion. (In this example, cpu 5 would share
-+ * cpu 0's portal, cpu 6 would share cpu4's portal, and cpu 7 would share cpu
-+ * 0's portal.) */
-+static struct cpumask want_unshared __initdata; /* cpus requested without "s" */
-+static struct cpumask want_shared __initdata; /* cpus requested with "s" */
-+
-+static int __init parse_bportals(char *str)
-+{
-+ return parse_portals_bootarg(str, &want_shared, &want_unshared,
-+ "bportals");
-+}
-+__setup("bportals=", parse_bportals);
-+
-+static int bman_offline_cpu(unsigned int cpu)
-+{
-+ struct bman_portal *p;
-+ const struct bm_portal_config *pcfg;
-+ p = (struct bman_portal *)affine_bportals[cpu];
-+ if (p) {
-+ pcfg = bman_get_bm_portal_config(p);
-+ if (pcfg)
-+ irq_set_affinity(pcfg->public_cfg.irq, cpumask_of(0));
-+ }
-+ return 0;
-+}
-+
-+#ifdef CONFIG_HOTPLUG_CPU
-+static int bman_online_cpu(unsigned int cpu)
-+{
-+ struct bman_portal *p;
-+ const struct bm_portal_config *pcfg;
-+ p = (struct bman_portal *)affine_bportals[cpu];
-+ if (p) {
-+ pcfg = bman_get_bm_portal_config(p);
-+ if (pcfg)
-+ irq_set_affinity(pcfg->public_cfg.irq, cpumask_of(cpu));
-+ }
-+ return 0;
-+}
-+static int bman_hotplug_cpu_callback(struct notifier_block *nfb,
-+ unsigned long action, void *hcpu)
-+{
-+ unsigned int cpu = (unsigned long)hcpu;
-+
-+ switch (action) {
-+ case CPU_ONLINE:
-+ case CPU_ONLINE_FROZEN:
-+ bman_online_cpu(cpu);
-+ break;
-+ case CPU_DOWN_PREPARE:
-+ case CPU_DOWN_PREPARE_FROZEN:
-+ bman_offline_cpu(cpu);
-+ default:
-+ break;
-+ }
-+ return NOTIFY_OK;
-+}
-+
-+static struct notifier_block bman_hotplug_cpu_notifier = {
-+ .notifier_call = bman_hotplug_cpu_callback,
-+};
-+#endif /* CONFIG_HOTPLUG_CPU */
-+
-+/* Initialise the Bman driver. The meat of this function deals with portals. The
-+ * following describes the flow of portal-handling, the code "steps" refer to
-+ * this description;
-+ * 1. Portal configs are parsed from the device-tree into 'unused_pcfgs', with
-+ * ::cpu==-1. Regions and interrupts are mapped (but interrupts are not
-+ * bound).
-+ * 2. The "want_shared" and "want_unshared" lists (as filled by the
-+ * "bportals=[...]" bootarg) are processed, allocating portals and assigning
-+ * them to cpus, placing them in the relevant list and setting ::cpu as
-+ * appropriate. If no "bportals" bootarg was present, the defaut is to try to
-+ * assign portals to all online cpus at the time of driver initialisation.
-+ * Any failure to allocate portals (when parsing the "want" lists or when
-+ * using default behaviour) will be silently tolerated (the "fixup" logic in
-+ * step 3 will determine what happens in this case).
-+ * 3. Do fixups relative to cpu_online_mask(). If no portals are marked for
-+ * sharing and sharing is required (because not all cpus have been assigned
-+ * portals), then one portal will marked for sharing. Conversely if no
-+ * sharing is required, any portals marked for sharing will not be shared. It
-+ * may be that sharing occurs when it wasn't expected, if portal allocation
-+ * failed to honour all the requested assignments (including the default
-+ * assignments if no bootarg is present).
-+ * 4. Unshared portals are initialised on their respective cpus.
-+ * 5. Shared portals are initialised on their respective cpus.
-+ * 6. Each remaining cpu is initialised to slave to one of the shared portals,
-+ * which are selected in a round-robin fashion.
-+ * Any portal configs left unused are available for USDPAA allocation.
-+ */
-+__init int bman_init(void)
-+{
-+ struct cpumask slave_cpus;
-+ struct cpumask unshared_cpus = *cpu_none_mask;
-+ struct cpumask shared_cpus = *cpu_none_mask;
-+ LIST_HEAD(unshared_pcfgs);
-+ LIST_HEAD(shared_pcfgs);
-+ struct device_node *dn;
-+ struct bm_portal_config *pcfg;
-+ struct bman_portal *p;
-+ int cpu, ret;
-+ struct cpumask offline_cpus;
-+
-+ /* Initialise the Bman (CCSR) device */
-+ for_each_compatible_node(dn, NULL, "fsl,bman") {
-+ if (!bman_init_ccsr(dn))
-+ pr_info("Bman err interrupt handler present\n");
-+ else
-+ pr_err("Bman CCSR setup failed\n");
-+ }
-+ /* Initialise any declared buffer pools */
-+ for_each_compatible_node(dn, NULL, "fsl,bpool") {
-+ ret = fsl_bpool_init(dn);
-+ if (ret)
-+ return ret;
-+ }
-+ /* Step 1. See comments at the beginning of the file. */
-+ for_each_compatible_node(dn, NULL, "fsl,bman-portal") {
-+ if (!of_device_is_available(dn))
-+ continue;
-+ pcfg = parse_pcfg(dn);
-+ if (pcfg)
-+ list_add_tail(&pcfg->list, &unused_pcfgs);
-+ }
-+ /* Step 2. */
-+ for_each_possible_cpu(cpu) {
-+ if (cpumask_test_cpu(cpu, &want_shared)) {
-+ pcfg = get_pcfg(&unused_pcfgs);
-+ if (!pcfg)
-+ break;
-+ pcfg->public_cfg.cpu = cpu;
-+ list_add_tail(&pcfg->list, &shared_pcfgs);
-+ cpumask_set_cpu(cpu, &shared_cpus);
-+ }
-+ if (cpumask_test_cpu(cpu, &want_unshared)) {
-+ if (cpumask_test_cpu(cpu, &shared_cpus))
-+ continue;
-+ pcfg = get_pcfg(&unused_pcfgs);
-+ if (!pcfg)
-+ break;
-+ pcfg->public_cfg.cpu = cpu;
-+ list_add_tail(&pcfg->list, &unshared_pcfgs);
-+ cpumask_set_cpu(cpu, &unshared_cpus);
-+ }
-+ }
-+ if (list_empty(&shared_pcfgs) && list_empty(&unshared_pcfgs)) {
-+ /* Default, give an unshared portal to each online cpu */
-+ for_each_online_cpu(cpu) {
-+ pcfg = get_pcfg(&unused_pcfgs);
-+ if (!pcfg)
-+ break;
-+ pcfg->public_cfg.cpu = cpu;
-+ list_add_tail(&pcfg->list, &unshared_pcfgs);
-+ cpumask_set_cpu(cpu, &unshared_cpus);
-+ }
-+ }
-+ /* Step 3. */
-+ cpumask_andnot(&slave_cpus, cpu_possible_mask, &shared_cpus);
-+ cpumask_andnot(&slave_cpus, &slave_cpus, &unshared_cpus);
-+ if (cpumask_empty(&slave_cpus)) {
-+ /* No sharing required */
-+ if (!list_empty(&shared_pcfgs)) {
-+ /* Migrate "shared" to "unshared" */
-+ cpumask_or(&unshared_cpus, &unshared_cpus,
-+ &shared_cpus);
-+ cpumask_clear(&shared_cpus);
-+ list_splice_tail(&shared_pcfgs, &unshared_pcfgs);
-+ INIT_LIST_HEAD(&shared_pcfgs);
-+ }
-+ } else {
-+ /* Sharing required */
-+ if (list_empty(&shared_pcfgs)) {
-+ /* Migrate one "unshared" to "shared" */
-+ pcfg = get_pcfg(&unshared_pcfgs);
-+ if (!pcfg) {
-+ pr_crit("No BMan portals available!\n");
-+ return 0;
-+ }
-+ cpumask_clear_cpu(pcfg->public_cfg.cpu, &unshared_cpus);
-+ cpumask_set_cpu(pcfg->public_cfg.cpu, &shared_cpus);
-+ list_add_tail(&pcfg->list, &shared_pcfgs);
-+ }
-+ }
-+ /* Step 4. */
-+ list_for_each_entry(pcfg, &unshared_pcfgs, list) {
-+ pcfg->public_cfg.is_shared = 0;
-+ p = init_pcfg(pcfg);
-+ if (!p) {
-+ pr_crit("Unable to initialize bman portal\n");
-+ return 0;
-+ }
-+ }
-+ /* Step 5. */
-+ list_for_each_entry(pcfg, &shared_pcfgs, list) {
-+ pcfg->public_cfg.is_shared = 1;
-+ p = init_pcfg(pcfg);
-+ if (p)
-+ shared_portals[num_shared_portals++] = p;
-+ }
-+ /* Step 6. */
-+ if (!cpumask_empty(&slave_cpus))
-+ for_each_cpu(cpu, &slave_cpus)
-+ init_slave(cpu);
-+ pr_info("Bman portals initialised\n");
-+ cpumask_andnot(&offline_cpus, cpu_possible_mask, cpu_online_mask);
-+ for_each_cpu(cpu, &offline_cpus)
-+ bman_offline_cpu(cpu);
-+#ifdef CONFIG_HOTPLUG_CPU
-+ register_hotcpu_notifier(&bman_hotplug_cpu_notifier);
-+#endif
-+ return 0;
-+}
-+
-+__init int bman_resource_init(void)
-+{
-+ struct device_node *dn;
-+ int ret;
-+
-+ /* Initialise BPID allocation ranges */
-+ for_each_compatible_node(dn, NULL, "fsl,bpid-range") {
-+ ret = fsl_bpid_range_init(dn);
-+ if (ret)
-+ return ret;
-+ }
-+ return 0;
-+}
-+
-+#ifdef CONFIG_SUSPEND
-+void suspend_unused_bportal(void)
-+{
-+ struct bm_portal_config *pcfg;
-+
-+ if (list_empty(&unused_pcfgs))
-+ return;
-+
-+ list_for_each_entry(pcfg, &unused_pcfgs, list) {
-+#ifdef CONFIG_PM_DEBUG
-+ pr_info("Need to save bportal %d\n", pcfg->public_cfg.index);
-+#endif
-+ /* save isdr, disable all via isdr, clear isr */
-+ pcfg->saved_isdr =
-+ __raw_readl(pcfg->addr_virt[DPA_PORTAL_CI] + 0xe08);
-+ __raw_writel(0xffffffff, pcfg->addr_virt[DPA_PORTAL_CI] +
-+ 0xe08);
-+ __raw_writel(0xffffffff, pcfg->addr_virt[DPA_PORTAL_CI] +
-+ 0xe00);
-+ }
-+ return;
-+}
-+
-+void resume_unused_bportal(void)
-+{
-+ struct bm_portal_config *pcfg;
-+
-+ if (list_empty(&unused_pcfgs))
-+ return;
-+
-+ list_for_each_entry(pcfg, &unused_pcfgs, list) {
-+#ifdef CONFIG_PM_DEBUG
-+ pr_info("Need to resume bportal %d\n", pcfg->public_cfg.index);
-+#endif
-+ /* restore isdr */
-+ __raw_writel(pcfg->saved_isdr,
-+ pcfg->addr_virt[DPA_PORTAL_CI] + 0xe08);
-+ }
-+ return;
-+}
-+#endif
---- /dev/null
-+++ b/drivers/staging/fsl_qbman/bman_high.c
-@@ -0,0 +1,1145 @@
-+/* Copyright 2008-2012 Freescale Semiconductor, Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#include "bman_low.h"
-+
-+/* Compilation constants */
-+#define RCR_THRESH 2 /* reread h/w CI when running out of space */
-+#define IRQNAME "BMan portal %d"
-+#define MAX_IRQNAME 16 /* big enough for "BMan portal %d" */
-+
-+struct bman_portal {
-+ struct bm_portal p;
-+ /* 2-element array. pools[0] is mask, pools[1] is snapshot. */
-+ struct bman_depletion *pools;
-+ int thresh_set;
-+ unsigned long irq_sources;
-+ u32 slowpoll; /* only used when interrupts are off */
-+#ifdef CONFIG_FSL_DPA_CAN_WAIT_SYNC
-+ struct bman_pool *rcri_owned; /* only 1 release WAIT_SYNC at a time */
-+#endif
-+#ifdef CONFIG_FSL_DPA_PORTAL_SHARE
-+ raw_spinlock_t sharing_lock; /* only used if is_shared */
-+ int is_shared;
-+ struct bman_portal *sharing_redirect;
-+#endif
-+ /* When the cpu-affine portal is activated, this is non-NULL */
-+ const struct bm_portal_config *config;
-+ /* This is needed for power management */
-+ struct platform_device *pdev;
-+ /* 64-entry hash-table of pool objects that are tracking depletion
-+ * entry/exit (ie. BMAN_POOL_FLAG_DEPLETION). This isn't fast-path, so
-+ * we're not fussy about cache-misses and so forth - whereas the above
-+ * members should all fit in one cacheline.
-+ * BTW, with 64 entries in the hash table and 64 buffer pools to track,
-+ * you'll never guess the hash-function ... */
-+ struct bman_pool *cb[64];
-+ char irqname[MAX_IRQNAME];
-+ /* Track if the portal was alloced by the driver */
-+ u8 alloced;
-+ /* power management data */
-+ u32 save_isdr;
-+};
-+
-+/* For an explanation of the locking, redirection, or affine-portal logic,
-+ * please consult the Qman driver for details. This is the same, only simpler
-+ * (no fiddly Qman-specific bits.) */
-+#ifdef CONFIG_FSL_DPA_PORTAL_SHARE
-+#define PORTAL_IRQ_LOCK(p, irqflags) \
-+ do { \
-+ if ((p)->is_shared) \
-+ raw_spin_lock_irqsave(&(p)->sharing_lock, irqflags); \
-+ else \
-+ local_irq_save(irqflags); \
-+ } while (0)
-+#define PORTAL_IRQ_UNLOCK(p, irqflags) \
-+ do { \
-+ if ((p)->is_shared) \
-+ raw_spin_unlock_irqrestore(&(p)->sharing_lock, \
-+ irqflags); \
-+ else \
-+ local_irq_restore(irqflags); \
-+ } while (0)
-+#else
-+#define PORTAL_IRQ_LOCK(p, irqflags) local_irq_save(irqflags)
-+#define PORTAL_IRQ_UNLOCK(p, irqflags) local_irq_restore(irqflags)
-+#endif
-+
-+static cpumask_t affine_mask;
-+static DEFINE_SPINLOCK(affine_mask_lock);
-+static DEFINE_PER_CPU(struct bman_portal, bman_affine_portal);
-+static inline struct bman_portal *get_raw_affine_portal(void)
-+{
-+ return &get_cpu_var(bman_affine_portal);
-+}
-+#ifdef CONFIG_FSL_DPA_PORTAL_SHARE
-+static inline struct bman_portal *get_affine_portal(void)
-+{
-+ struct bman_portal *p = get_raw_affine_portal();
-+ if (p->sharing_redirect)
-+ return p->sharing_redirect;
-+ return p;
-+}
-+#else
-+#define get_affine_portal() get_raw_affine_portal()
-+#endif
-+static inline void put_affine_portal(void)
-+{
-+ put_cpu_var(bman_affine_portal);
-+}
-+static inline struct bman_portal *get_poll_portal(void)
-+{
-+ return &get_cpu_var(bman_affine_portal);
-+}
-+#define put_poll_portal()
-+
-+/* GOTCHA: this object type refers to a pool, it isn't *the* pool. There may be
-+ * more than one such object per Bman buffer pool, eg. if different users of the
-+ * pool are operating via different portals. */
-+struct bman_pool {
-+ struct bman_pool_params params;
-+ /* Used for hash-table admin when using depletion notifications. */
-+ struct bman_portal *portal;
-+ struct bman_pool *next;
-+ /* stockpile state - NULL unless BMAN_POOL_FLAG_STOCKPILE is set */
-+ struct bm_buffer *sp;
-+ unsigned int sp_fill;
-+#ifdef CONFIG_FSL_DPA_CHECKING
-+ atomic_t in_use;
-+#endif
-+};
-+
-+/* (De)Registration of depletion notification callbacks */
-+static void depletion_link(struct bman_portal *portal, struct bman_pool *pool)
-+{
-+ __maybe_unused unsigned long irqflags;
-+ pool->portal = portal;
-+ PORTAL_IRQ_LOCK(portal, irqflags);
-+ pool->next = portal->cb[pool->params.bpid];
-+ portal->cb[pool->params.bpid] = pool;
-+ if (!pool->next)
-+ /* First object for that bpid on this portal, enable the BSCN
-+ * mask bit. */
-+ bm_isr_bscn_mask(&portal->p, pool->params.bpid, 1);
-+ PORTAL_IRQ_UNLOCK(portal, irqflags);
-+}
-+static void depletion_unlink(struct bman_pool *pool)
-+{
-+ struct bman_pool *it, *last = NULL;
-+ struct bman_pool **base = &pool->portal->cb[pool->params.bpid];
-+ __maybe_unused unsigned long irqflags;
-+ PORTAL_IRQ_LOCK(pool->portal, irqflags);
-+ it = *base; /* <-- gotcha, don't do this prior to the irq_save */
-+ while (it != pool) {
-+ last = it;
-+ it = it->next;
-+ }
-+ if (!last)
-+ *base = pool->next;
-+ else
-+ last->next = pool->next;
-+ if (!last && !pool->next) {
-+ /* Last object for that bpid on this portal, disable the BSCN
-+ * mask bit. */
-+ bm_isr_bscn_mask(&pool->portal->p, pool->params.bpid, 0);
-+ /* And "forget" that we last saw this pool as depleted */
-+ bman_depletion_unset(&pool->portal->pools[1],
-+ pool->params.bpid);
-+ }
-+ PORTAL_IRQ_UNLOCK(pool->portal, irqflags);
-+}
-+
-+/* In the case that the application's core loop calls qman_poll() and
-+ * bman_poll(), we ought to balance how often we incur the overheads of the
-+ * slow-path poll. We'll use two decrementer sources. The idle decrementer
-+ * constant is used when the last slow-poll detected no work to do, and the busy
-+ * decrementer constant when the last slow-poll had work to do. */
-+#define SLOW_POLL_IDLE 1000
-+#define SLOW_POLL_BUSY 10
-+static u32 __poll_portal_slow(struct bman_portal *p, u32 is);
-+
-+/* Portal interrupt handler */
-+static irqreturn_t portal_isr(__always_unused int irq, void *ptr)
-+{
-+ struct bman_portal *p = ptr;
-+ u32 clear = p->irq_sources;
-+ u32 is = bm_isr_status_read(&p->p) & p->irq_sources;
-+ clear |= __poll_portal_slow(p, is);
-+ bm_isr_status_clear(&p->p, clear);
-+ return IRQ_HANDLED;
-+}
-+
-+#ifdef CONFIG_SUSPEND
-+static int _bman_portal_suspend_noirq(struct device *dev)
-+{
-+ struct bman_portal *p = (struct bman_portal *)dev->platform_data;
-+#ifdef CONFIG_PM_DEBUG
-+ struct platform_device *pdev = to_platform_device(dev);
-+#endif
-+ p->save_isdr = bm_isr_disable_read(&p->p);
-+ bm_isr_disable_write(&p->p, 0xffffffff);
-+ bm_isr_status_clear(&p->p, 0xffffffff);
-+#ifdef CONFIG_PM_DEBUG
-+ pr_info("Suspend for %s\n", pdev->name);
-+#endif
-+ return 0;
-+}
-+
-+static int _bman_portal_resume_noirq(struct device *dev)
-+{
-+ struct bman_portal *p = (struct bman_portal *)dev->platform_data;
-+
-+ /* restore isdr */
-+ bm_isr_disable_write(&p->p, p->save_isdr);
-+ return 0;
-+}
-+#else
-+#define _bman_portal_suspend_noirq NULL
-+#define _bman_portal_resume_noirq NULL
-+#endif
-+
-+struct dev_pm_domain bman_portal_device_pm_domain = {
-+ .ops = {
-+ USE_PLATFORM_PM_SLEEP_OPS
-+ .suspend_noirq = _bman_portal_suspend_noirq,
-+ .resume_noirq = _bman_portal_resume_noirq,
-+ }
-+};
-+
-+struct bman_portal *bman_create_portal(
-+ struct bman_portal *portal,
-+ const struct bm_portal_config *config)
-+{
-+ struct bm_portal *__p;
-+ const struct bman_depletion *pools = &config->public_cfg.mask;
-+ int ret;
-+ u8 bpid = 0;
-+ char buf[16];
-+
-+ if (!portal) {
-+ portal = kmalloc(sizeof(*portal), GFP_KERNEL);
-+ if (!portal)
-+ return portal;
-+ portal->alloced = 1;
-+ } else
-+ portal->alloced = 0;
-+
-+ __p = &portal->p;
-+
-+ /* prep the low-level portal struct with the mapped addresses from the
-+ * config, everything that follows depends on it and "config" is more
-+ * for (de)reference... */
-+ __p->addr.addr_ce = config->addr_virt[DPA_PORTAL_CE];
-+ __p->addr.addr_ci = config->addr_virt[DPA_PORTAL_CI];
-+ if (bm_rcr_init(__p, bm_rcr_pvb, bm_rcr_cce)) {
-+ pr_err("Bman RCR initialisation failed\n");
-+ goto fail_rcr;
-+ }
-+ if (bm_mc_init(__p)) {
-+ pr_err("Bman MC initialisation failed\n");
-+ goto fail_mc;
-+ }
-+ if (bm_isr_init(__p)) {
-+ pr_err("Bman ISR initialisation failed\n");
-+ goto fail_isr;
-+ }
-+ portal->pools = kmalloc(2 * sizeof(*pools), GFP_KERNEL);
-+ if (!portal->pools)
-+ goto fail_pools;
-+ portal->pools[0] = *pools;
-+ bman_depletion_init(portal->pools + 1);
-+ while (bpid < bman_pool_max) {
-+ /* Default to all BPIDs disabled, we enable as required at
-+ * run-time. */
-+ bm_isr_bscn_mask(__p, bpid, 0);
-+ bpid++;
-+ }
-+ portal->slowpoll = 0;
-+#ifdef CONFIG_FSL_DPA_CAN_WAIT_SYNC
-+ portal->rcri_owned = NULL;
-+#endif
-+#ifdef CONFIG_FSL_DPA_PORTAL_SHARE
-+ raw_spin_lock_init(&portal->sharing_lock);
-+ portal->is_shared = config->public_cfg.is_shared;
-+ portal->sharing_redirect = NULL;
-+#endif
-+ sprintf(buf, "bportal-%u", config->public_cfg.index);
-+ portal->pdev = platform_device_alloc(buf, -1);
-+ if (!portal->pdev)
-+ goto fail_devalloc;
-+ portal->pdev->dev.pm_domain = &bman_portal_device_pm_domain;
-+ portal->pdev->dev.platform_data = portal;
-+ ret = platform_device_add(portal->pdev);
-+ if (ret)
-+ goto fail_devadd;
-+ memset(&portal->cb, 0, sizeof(portal->cb));
-+ /* Write-to-clear any stale interrupt status bits */
-+ bm_isr_disable_write(__p, 0xffffffff);
-+ portal->irq_sources = 0;
-+ bm_isr_enable_write(__p, portal->irq_sources);
-+ bm_isr_status_clear(__p, 0xffffffff);
-+ snprintf(portal->irqname, MAX_IRQNAME, IRQNAME, config->public_cfg.cpu);
-+ if (request_irq(config->public_cfg.irq, portal_isr, 0, portal->irqname,
-+ portal)) {
-+ pr_err("request_irq() failed\n");
-+ goto fail_irq;
-+ }
-+ if ((config->public_cfg.cpu != -1) &&
-+ irq_can_set_affinity(config->public_cfg.irq) &&
-+ irq_set_affinity(config->public_cfg.irq,
-+ cpumask_of(config->public_cfg.cpu))) {
-+ pr_err("irq_set_affinity() failed %s\n", portal->irqname);
-+ goto fail_affinity;
-+ }
-+
-+ /* Need RCR to be empty before continuing */
-+ ret = bm_rcr_get_fill(__p);
-+ if (ret) {
-+ pr_err("Bman RCR unclean\n");
-+ goto fail_rcr_empty;
-+ }
-+ /* Success */
-+ portal->config = config;
-+
-+ bm_isr_disable_write(__p, 0);
-+ bm_isr_uninhibit(__p);
-+ return portal;
-+fail_rcr_empty:
-+fail_affinity:
-+ free_irq(config->public_cfg.irq, portal);
-+fail_irq:
-+ platform_device_del(portal->pdev);
-+fail_devadd:
-+ platform_device_put(portal->pdev);
-+fail_devalloc:
-+ kfree(portal->pools);
-+fail_pools:
-+ bm_isr_finish(__p);
-+fail_isr:
-+ bm_mc_finish(__p);
-+fail_mc:
-+ bm_rcr_finish(__p);
-+fail_rcr:
-+ if (portal->alloced)
-+ kfree(portal);
-+ return NULL;
-+}
-+
-+struct bman_portal *bman_create_affine_portal(
-+ const struct bm_portal_config *config)
-+{
-+ struct bman_portal *portal;
-+
-+ portal = &per_cpu(bman_affine_portal, config->public_cfg.cpu);
-+ portal = bman_create_portal(portal, config);
-+ if (portal) {
-+ spin_lock(&affine_mask_lock);
-+ cpumask_set_cpu(config->public_cfg.cpu, &affine_mask);
-+ spin_unlock(&affine_mask_lock);
-+ }
-+ return portal;
-+}
-+
-+
-+struct bman_portal *bman_create_affine_slave(struct bman_portal *redirect,
-+ int cpu)
-+{
-+#ifdef CONFIG_FSL_DPA_PORTAL_SHARE
-+ struct bman_portal *p;
-+ p = &per_cpu(bman_affine_portal, cpu);
-+ BUG_ON(p->config);
-+ BUG_ON(p->is_shared);
-+ BUG_ON(!redirect->config->public_cfg.is_shared);
-+ p->irq_sources = 0;
-+ p->sharing_redirect = redirect;
-+ return p;
-+#else
-+ BUG();
-+ return NULL;
-+#endif
-+}
-+
-+void bman_destroy_portal(struct bman_portal *bm)
-+{
-+ const struct bm_portal_config *pcfg;
-+ pcfg = bm->config;
-+ bm_rcr_cce_update(&bm->p);
-+ bm_rcr_cce_update(&bm->p);
-+
-+ free_irq(pcfg->public_cfg.irq, bm);
-+
-+ kfree(bm->pools);
-+ bm_isr_finish(&bm->p);
-+ bm_mc_finish(&bm->p);
-+ bm_rcr_finish(&bm->p);
-+ bm->config = NULL;
-+ if (bm->alloced)
-+ kfree(bm);
-+}
-+
-+const struct bm_portal_config *bman_destroy_affine_portal(void)
-+{
-+ struct bman_portal *bm = get_raw_affine_portal();
-+ const struct bm_portal_config *pcfg;
-+#ifdef CONFIG_FSL_DPA_PORTAL_SHARE
-+ if (bm->sharing_redirect) {
-+ bm->sharing_redirect = NULL;
-+ put_affine_portal();
-+ return NULL;
-+ }
-+ bm->is_shared = 0;
-+#endif
-+ pcfg = bm->config;
-+ bman_destroy_portal(bm);
-+ spin_lock(&affine_mask_lock);
-+ cpumask_clear_cpu(pcfg->public_cfg.cpu, &affine_mask);
-+ spin_unlock(&affine_mask_lock);
-+ put_affine_portal();
-+ return pcfg;
-+}
-+
-+/* When release logic waits on available RCR space, we need a global waitqueue
-+ * in the case of "affine" use (as the waits wake on different cpus which means
-+ * different portals - so we can't wait on any per-portal waitqueue). */
-+static DECLARE_WAIT_QUEUE_HEAD(affine_queue);
-+
-+static u32 __poll_portal_slow(struct bman_portal *p, u32 is)
-+{
-+ struct bman_depletion tmp;
-+ u32 ret = is;
-+
-+ /* There is a gotcha to be aware of. If we do the query before clearing
-+ * the status register, we may miss state changes that occur between the
-+ * two. If we write to clear the status register before the query, the
-+ * cache-enabled query command may overtake the status register write
-+ * unless we use a heavyweight sync (which we don't want). Instead, we
-+ * write-to-clear the status register then *read it back* before doing
-+ * the query, hence the odd while loop with the 'is' accumulation. */
-+ if (is & BM_PIRQ_BSCN) {
-+ struct bm_mc_result *mcr;
-+ __maybe_unused unsigned long irqflags;
-+ unsigned int i, j;
-+ u32 __is;
-+ bm_isr_status_clear(&p->p, BM_PIRQ_BSCN);
-+ while ((__is = bm_isr_status_read(&p->p)) & BM_PIRQ_BSCN) {
-+ is |= __is;
-+ bm_isr_status_clear(&p->p, BM_PIRQ_BSCN);
-+ }
-+ is &= ~BM_PIRQ_BSCN;
-+ PORTAL_IRQ_LOCK(p, irqflags);
-+ bm_mc_start(&p->p);
-+ bm_mc_commit(&p->p, BM_MCC_VERB_CMD_QUERY);
-+ while (!(mcr = bm_mc_result(&p->p)))
-+ cpu_relax();
-+ tmp = mcr->query.ds.state;
-+ tmp.__state[0] = be32_to_cpu(tmp.__state[0]);
-+ tmp.__state[1] = be32_to_cpu(tmp.__state[1]);
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+ for (i = 0; i < 2; i++) {
-+ int idx = i * 32;
-+ /* tmp is a mask of currently-depleted pools.
-+ * pools[0] is mask of those we care about.
-+ * pools[1] is our previous view (we only want to
-+ * be told about changes). */
-+ tmp.__state[i] &= p->pools[0].__state[i];
-+ if (tmp.__state[i] == p->pools[1].__state[i])
-+ /* fast-path, nothing to see, move along */
-+ continue;
-+ for (j = 0; j <= 31; j++, idx++) {
-+ struct bman_pool *pool = p->cb[idx];
-+ int b4 = bman_depletion_get(&p->pools[1], idx);
-+ int af = bman_depletion_get(&tmp, idx);
-+ if (b4 == af)
-+ continue;
-+ while (pool) {
-+ pool->params.cb(p, pool,
-+ pool->params.cb_ctx, af);
-+ pool = pool->next;
-+ }
-+ }
-+ }
-+ p->pools[1] = tmp;
-+ }
-+
-+ if (is & BM_PIRQ_RCRI) {
-+ __maybe_unused unsigned long irqflags;
-+ PORTAL_IRQ_LOCK(p, irqflags);
-+ bm_rcr_cce_update(&p->p);
-+#ifdef CONFIG_FSL_DPA_CAN_WAIT_SYNC
-+ /* If waiting for sync, we only cancel the interrupt threshold
-+ * when the ring utilisation hits zero. */
-+ if (p->rcri_owned) {
-+ if (!bm_rcr_get_fill(&p->p)) {
-+ p->rcri_owned = NULL;
-+ bm_rcr_set_ithresh(&p->p, 0);
-+ }
-+ } else
-+#endif
-+ bm_rcr_set_ithresh(&p->p, 0);
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+ wake_up(&affine_queue);
-+ bm_isr_status_clear(&p->p, BM_PIRQ_RCRI);
-+ is &= ~BM_PIRQ_RCRI;
-+ }
-+
-+ /* There should be no status register bits left undefined */
-+ DPA_ASSERT(!is);
-+ return ret;
-+}
-+
-+const struct bman_portal_config *bman_get_portal_config(void)
-+{
-+ struct bman_portal *p = get_affine_portal();
-+ const struct bman_portal_config *ret = &p->config->public_cfg;
-+ put_affine_portal();
-+ return ret;
-+}
-+EXPORT_SYMBOL(bman_get_portal_config);
-+
-+u32 bman_irqsource_get(void)
-+{
-+ struct bman_portal *p = get_raw_affine_portal();
-+ u32 ret = p->irq_sources & BM_PIRQ_VISIBLE;
-+ put_affine_portal();
-+ return ret;
-+}
-+EXPORT_SYMBOL(bman_irqsource_get);
-+
-+int bman_p_irqsource_add(struct bman_portal *p, __maybe_unused u32 bits)
-+{
-+#ifdef CONFIG_FSL_DPA_PORTAL_SHARE
-+ if (p->sharing_redirect)
-+ return -EINVAL;
-+ else
-+#endif
-+ {
-+ __maybe_unused unsigned long irqflags;
-+ PORTAL_IRQ_LOCK(p, irqflags);
-+ set_bits(bits & BM_PIRQ_VISIBLE, &p->irq_sources);
-+ bm_isr_enable_write(&p->p, p->irq_sources);
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+ }
-+ return 0;
-+}
-+EXPORT_SYMBOL(bman_p_irqsource_add);
-+
-+int bman_irqsource_add(__maybe_unused u32 bits)
-+{
-+ struct bman_portal *p = get_raw_affine_portal();
-+ int ret = 0;
-+ ret = bman_p_irqsource_add(p, bits);
-+ put_affine_portal();
-+ return ret;
-+}
-+EXPORT_SYMBOL(bman_irqsource_add);
-+
-+int bman_irqsource_remove(u32 bits)
-+{
-+ struct bman_portal *p = get_raw_affine_portal();
-+ __maybe_unused unsigned long irqflags;
-+ u32 ier;
-+#ifdef CONFIG_FSL_DPA_PORTAL_SHARE
-+ if (p->sharing_redirect) {
-+ put_affine_portal();
-+ return -EINVAL;
-+ }
-+#endif
-+ /* Our interrupt handler only processes+clears status register bits that
-+ * are in p->irq_sources. As we're trimming that mask, if one of them
-+ * were to assert in the status register just before we remove it from
-+ * the enable register, there would be an interrupt-storm when we
-+ * release the IRQ lock. So we wait for the enable register update to
-+ * take effect in h/w (by reading it back) and then clear all other bits
-+ * in the status register. Ie. we clear them from ISR once it's certain
-+ * IER won't allow them to reassert. */
-+ PORTAL_IRQ_LOCK(p, irqflags);
-+ bits &= BM_PIRQ_VISIBLE;
-+ clear_bits(bits, &p->irq_sources);
-+ bm_isr_enable_write(&p->p, p->irq_sources);
-+ ier = bm_isr_enable_read(&p->p);
-+ /* Using "~ier" (rather than "bits" or "~p->irq_sources") creates a
-+ * data-dependency, ie. to protect against re-ordering. */
-+ bm_isr_status_clear(&p->p, ~ier);
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+ put_affine_portal();
-+ return 0;
-+}
-+EXPORT_SYMBOL(bman_irqsource_remove);
-+
-+const cpumask_t *bman_affine_cpus(void)
-+{
-+ return &affine_mask;
-+}
-+EXPORT_SYMBOL(bman_affine_cpus);
-+
-+u32 bman_poll_slow(void)
-+{
-+ struct bman_portal *p = get_poll_portal();
-+ u32 ret;
-+#ifdef CONFIG_FSL_DPA_PORTAL_SHARE
-+ if (unlikely(p->sharing_redirect))
-+ ret = (u32)-1;
-+ else
-+#endif
-+ {
-+ u32 is = bm_isr_status_read(&p->p) & ~p->irq_sources;
-+ ret = __poll_portal_slow(p, is);
-+ bm_isr_status_clear(&p->p, ret);
-+ }
-+ put_poll_portal();
-+ return ret;
-+}
-+EXPORT_SYMBOL(bman_poll_slow);
-+
-+/* Legacy wrapper */
-+void bman_poll(void)
-+{
-+ struct bman_portal *p = get_poll_portal();
-+#ifdef CONFIG_FSL_DPA_PORTAL_SHARE
-+ if (unlikely(p->sharing_redirect))
-+ goto done;
-+#endif
-+ if (!(p->slowpoll--)) {
-+ u32 is = bm_isr_status_read(&p->p) & ~p->irq_sources;
-+ u32 active = __poll_portal_slow(p, is);
-+ if (active)
-+ p->slowpoll = SLOW_POLL_BUSY;
-+ else
-+ p->slowpoll = SLOW_POLL_IDLE;
-+ }
-+#ifdef CONFIG_FSL_DPA_PORTAL_SHARE
-+done:
-+#endif
-+ put_poll_portal();
-+}
-+EXPORT_SYMBOL(bman_poll);
-+
-+static const u32 zero_thresholds[4] = {0, 0, 0, 0};
-+
-+struct bman_pool *bman_new_pool(const struct bman_pool_params *params)
-+{
-+ struct bman_pool *pool = NULL;
-+ u32 bpid;
-+
-+ if (params->flags & BMAN_POOL_FLAG_DYNAMIC_BPID) {
-+ int ret = bman_alloc_bpid(&bpid);
-+ if (ret)
-+ return NULL;
-+ } else {
-+ if (params->bpid >= bman_pool_max)
-+ return NULL;
-+ bpid = params->bpid;
-+ }
-+#ifdef CONFIG_FSL_BMAN_CONFIG
-+ if (params->flags & BMAN_POOL_FLAG_THRESH) {
-+ int ret = bm_pool_set(bpid, params->thresholds);
-+ if (ret)
-+ goto err;
-+ }
-+#else
-+ if (params->flags & BMAN_POOL_FLAG_THRESH)
-+ goto err;
-+#endif
-+ pool = kmalloc(sizeof(*pool), GFP_KERNEL);
-+ if (!pool)
-+ goto err;
-+ pool->sp = NULL;
-+ pool->sp_fill = 0;
-+ pool->params = *params;
-+#ifdef CONFIG_FSL_DPA_CHECKING
-+ atomic_set(&pool->in_use, 1);
-+#endif
-+ if (params->flags & BMAN_POOL_FLAG_DYNAMIC_BPID)
-+ pool->params.bpid = bpid;
-+ if (params->flags & BMAN_POOL_FLAG_STOCKPILE) {
-+ pool->sp = kmalloc(sizeof(struct bm_buffer) * BMAN_STOCKPILE_SZ,
-+ GFP_KERNEL);
-+ if (!pool->sp)
-+ goto err;
-+ }
-+ if (pool->params.flags & BMAN_POOL_FLAG_DEPLETION) {
-+ struct bman_portal *p = get_affine_portal();
-+ if (!p->pools || !bman_depletion_get(&p->pools[0], bpid)) {
-+ pr_err("Depletion events disabled for bpid %d\n", bpid);
-+ goto err;
-+ }
-+ depletion_link(p, pool);
-+ put_affine_portal();
-+ }
-+ return pool;
-+err:
-+#ifdef CONFIG_FSL_BMAN_CONFIG
-+ if (params->flags & BMAN_POOL_FLAG_THRESH)
-+ bm_pool_set(bpid, zero_thresholds);
-+#endif
-+ if (params->flags & BMAN_POOL_FLAG_DYNAMIC_BPID)
-+ bman_release_bpid(bpid);
-+ if (pool) {
-+ kfree(pool->sp);
-+ kfree(pool);
-+ }
-+ return NULL;
-+}
-+EXPORT_SYMBOL(bman_new_pool);
-+
-+void bman_free_pool(struct bman_pool *pool)
-+{
-+#ifdef CONFIG_FSL_BMAN_CONFIG
-+ if (pool->params.flags & BMAN_POOL_FLAG_THRESH)
-+ bm_pool_set(pool->params.bpid, zero_thresholds);
-+#endif
-+ if (pool->params.flags & BMAN_POOL_FLAG_DEPLETION)
-+ depletion_unlink(pool);
-+ if (pool->params.flags & BMAN_POOL_FLAG_STOCKPILE) {
-+ if (pool->sp_fill)
-+ pr_err("Stockpile not flushed, has %u in bpid %u.\n",
-+ pool->sp_fill, pool->params.bpid);
-+ kfree(pool->sp);
-+ pool->sp = NULL;
-+ pool->params.flags ^= BMAN_POOL_FLAG_STOCKPILE;
-+ }
-+ if (pool->params.flags & BMAN_POOL_FLAG_DYNAMIC_BPID)
-+ bman_release_bpid(pool->params.bpid);
-+ kfree(pool);
-+}
-+EXPORT_SYMBOL(bman_free_pool);
-+
-+const struct bman_pool_params *bman_get_params(const struct bman_pool *pool)
-+{
-+ return &pool->params;
-+}
-+EXPORT_SYMBOL(bman_get_params);
-+
-+static noinline void update_rcr_ci(struct bman_portal *p, u8 avail)
-+{
-+ if (avail)
-+ bm_rcr_cce_prefetch(&p->p);
-+ else
-+ bm_rcr_cce_update(&p->p);
-+}
-+
-+int bman_rcr_is_empty(void)
-+{
-+ __maybe_unused unsigned long irqflags;
-+ struct bman_portal *p = get_affine_portal();
-+ u8 avail;
-+
-+ PORTAL_IRQ_LOCK(p, irqflags);
-+ update_rcr_ci(p, 0);
-+ avail = bm_rcr_get_fill(&p->p);
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+ put_affine_portal();
-+ return avail == 0;
-+}
-+EXPORT_SYMBOL(bman_rcr_is_empty);
-+
-+static inline struct bm_rcr_entry *try_rel_start(struct bman_portal **p,
-+#ifdef CONFIG_FSL_DPA_CAN_WAIT
-+ __maybe_unused struct bman_pool *pool,
-+#endif
-+ __maybe_unused unsigned long *irqflags,
-+ __maybe_unused u32 flags)
-+{
-+ struct bm_rcr_entry *r;
-+ u8 avail;
-+
-+ *p = get_affine_portal();
-+ PORTAL_IRQ_LOCK(*p, (*irqflags));
-+#ifdef CONFIG_FSL_DPA_CAN_WAIT_SYNC
-+ if (unlikely((flags & BMAN_RELEASE_FLAG_WAIT) &&
-+ (flags & BMAN_RELEASE_FLAG_WAIT_SYNC))) {
-+ if ((*p)->rcri_owned) {
-+ PORTAL_IRQ_UNLOCK(*p, (*irqflags));
-+ put_affine_portal();
-+ return NULL;
-+ }
-+ (*p)->rcri_owned = pool;
-+ }
-+#endif
-+ avail = bm_rcr_get_avail(&(*p)->p);
-+ if (avail < 2)
-+ update_rcr_ci(*p, avail);
-+ r = bm_rcr_start(&(*p)->p);
-+ if (unlikely(!r)) {
-+#ifdef CONFIG_FSL_DPA_CAN_WAIT_SYNC
-+ if (unlikely((flags & BMAN_RELEASE_FLAG_WAIT) &&
-+ (flags & BMAN_RELEASE_FLAG_WAIT_SYNC)))
-+ (*p)->rcri_owned = NULL;
-+#endif
-+ PORTAL_IRQ_UNLOCK(*p, (*irqflags));
-+ put_affine_portal();
-+ }
-+ return r;
-+}
-+
-+#ifdef CONFIG_FSL_DPA_CAN_WAIT
-+static noinline struct bm_rcr_entry *__wait_rel_start(struct bman_portal **p,
-+ struct bman_pool *pool,
-+ __maybe_unused unsigned long *irqflags,
-+ u32 flags)
-+{
-+ struct bm_rcr_entry *rcr = try_rel_start(p, pool, irqflags, flags);
-+ if (!rcr)
-+ bm_rcr_set_ithresh(&(*p)->p, 1);
-+ return rcr;
-+}
-+
-+static noinline struct bm_rcr_entry *wait_rel_start(struct bman_portal **p,
-+ struct bman_pool *pool,
-+ __maybe_unused unsigned long *irqflags,
-+ u32 flags)
-+{
-+ struct bm_rcr_entry *rcr;
-+#ifndef CONFIG_FSL_DPA_CAN_WAIT_SYNC
-+ pool = NULL;
-+#endif
-+ if (flags & BMAN_RELEASE_FLAG_WAIT_INT)
-+ /* NB: return NULL if signal occurs before completion. Signal
-+ * can occur during return. Caller must check for signal */
-+ wait_event_interruptible(affine_queue,
-+ (rcr = __wait_rel_start(p, pool, irqflags, flags)));
-+ else
-+ wait_event(affine_queue,
-+ (rcr = __wait_rel_start(p, pool, irqflags, flags)));
-+ return rcr;
-+}
-+#endif
-+
-+static inline int __bman_release(struct bman_pool *pool,
-+ const struct bm_buffer *bufs, u8 num, u32 flags)
-+{
-+ struct bman_portal *p;
-+ struct bm_rcr_entry *r;
-+ __maybe_unused unsigned long irqflags;
-+ u32 i = num - 1;
-+
-+#ifdef CONFIG_FSL_DPA_CAN_WAIT
-+ if (flags & BMAN_RELEASE_FLAG_WAIT)
-+ r = wait_rel_start(&p, pool, &irqflags, flags);
-+ else
-+ r = try_rel_start(&p, pool, &irqflags, flags);
-+#else
-+ r = try_rel_start(&p, &irqflags, flags);
-+#endif
-+ if (!r)
-+ return -EBUSY;
-+ /* We can copy all but the first entry, as this can trigger badness
-+ * with the valid-bit. Use the overlay to mask the verb byte. */
-+ r->bufs[0].opaque =
-+ ((cpu_to_be64((bufs[0].opaque |
-+ ((u64)pool->params.bpid<<48))
-+ & 0x00ffffffffffffff)));
-+ if (i) {
-+ for (i = 1; i < num; i++)
-+ r->bufs[i].opaque =
-+ cpu_to_be64(bufs[i].opaque);
-+ }
-+
-+ bm_rcr_pvb_commit(&p->p, BM_RCR_VERB_CMD_BPID_SINGLE |
-+ (num & BM_RCR_VERB_BUFCOUNT_MASK));
-+#ifdef CONFIG_FSL_DPA_CAN_WAIT_SYNC
-+ /* if we wish to sync we need to set the threshold after h/w sees the
-+ * new ring entry. As we're mixing cache-enabled and cache-inhibited
-+ * accesses, this requires a heavy-weight sync. */
-+ if (unlikely((flags & BMAN_RELEASE_FLAG_WAIT) &&
-+ (flags & BMAN_RELEASE_FLAG_WAIT_SYNC))) {
-+ hwsync();
-+ bm_rcr_set_ithresh(&p->p, 1);
-+ }
-+#endif
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+ put_affine_portal();
-+#ifdef CONFIG_FSL_DPA_CAN_WAIT_SYNC
-+ if (unlikely((flags & BMAN_RELEASE_FLAG_WAIT) &&
-+ (flags & BMAN_RELEASE_FLAG_WAIT_SYNC))) {
-+ if (flags & BMAN_RELEASE_FLAG_WAIT_INT)
-+ /* NB: return success even if signal occurs before
-+ * condition is true. pvb_commit guarantees success */
-+ wait_event_interruptible(affine_queue,
-+ (p->rcri_owned != pool));
-+ else
-+ wait_event(affine_queue, (p->rcri_owned != pool));
-+ }
-+#endif
-+ return 0;
-+}
-+
-+int bman_release(struct bman_pool *pool, const struct bm_buffer *bufs, u8 num,
-+ u32 flags)
-+{
-+ int ret;
-+#ifdef CONFIG_FSL_DPA_CHECKING
-+ if (!num || (num > 8))
-+ return -EINVAL;
-+ if (pool->params.flags & BMAN_POOL_FLAG_NO_RELEASE)
-+ return -EINVAL;
-+#endif
-+ /* Without stockpile, this API is a pass-through to the h/w operation */
-+ if (!(pool->params.flags & BMAN_POOL_FLAG_STOCKPILE))
-+ return __bman_release(pool, bufs, num, flags);
-+#ifdef CONFIG_FSL_DPA_CHECKING
-+ if (!atomic_dec_and_test(&pool->in_use)) {
-+ pr_crit("Parallel attempts to enter bman_released() detected.");
-+ panic("only one instance of bman_released/acquired allowed");
-+ }
-+#endif
-+ /* Two movements of buffers are possible, and can occur in either order.
-+ * A: moving buffers from the caller to the stockpile.
-+ * B: moving buffers from the stockpile to hardware.
-+ * Order 1: if there is already enough space in the stockpile for A
-+ * then we want to do A first, and only do B if we trigger the
-+ * stockpile-high threshold.
-+ * Order 2: if there is not enough space in the stockpile for A, then
-+ * we want to do B first, then do A if B had succeeded. However in this
-+ * case B is dependent on how many buffers the user needs to release,
-+ * not the stockpile-high threshold.
-+ * Due to the different handling of B between the two cases, putting A
-+ * and B in a while() loop would require quite obscure logic, so handle
-+ * the different sequences explicitly. */
-+ if ((pool->sp_fill + num) <= BMAN_STOCKPILE_SZ) {
-+ /* Order 1: do A */
-+ copy_words(pool->sp + pool->sp_fill, bufs,
-+ sizeof(struct bm_buffer) * num);
-+ pool->sp_fill += num;
-+ /* do B relative to STOCKPILE_HIGH */
-+ while (pool->sp_fill >= BMAN_STOCKPILE_HIGH) {
-+ ret = __bman_release(pool,
-+ pool->sp + (pool->sp_fill - 8), 8,
-+ flags);
-+ if (ret >= 0)
-+ pool->sp_fill -= 8;
-+ }
-+ } else {
-+ /* Order 2: do B relative to 'num' */
-+ do {
-+ ret = __bman_release(pool,
-+ pool->sp + (pool->sp_fill - 8), 8,
-+ flags);
-+ if (ret < 0)
-+ /* failure */
-+ goto release_done;
-+ pool->sp_fill -= 8;
-+ } while ((pool->sp_fill + num) > BMAN_STOCKPILE_SZ);
-+ /* do A */
-+ copy_words(pool->sp + pool->sp_fill, bufs,
-+ sizeof(struct bm_buffer) * num);
-+ pool->sp_fill += num;
-+ }
-+ /* success */
-+ ret = 0;
-+release_done:
-+#ifdef CONFIG_FSL_DPA_CHECKING
-+ atomic_inc(&pool->in_use);
-+#endif
-+ return ret;
-+}
-+EXPORT_SYMBOL(bman_release);
-+
-+static inline int __bman_acquire(struct bman_pool *pool, struct bm_buffer *bufs,
-+ u8 num)
-+{
-+ struct bman_portal *p = get_affine_portal();
-+ struct bm_mc_command *mcc;
-+ struct bm_mc_result *mcr;
-+ __maybe_unused unsigned long irqflags;
-+ int ret, i;
-+
-+ PORTAL_IRQ_LOCK(p, irqflags);
-+ mcc = bm_mc_start(&p->p);
-+ mcc->acquire.bpid = pool->params.bpid;
-+ bm_mc_commit(&p->p, BM_MCC_VERB_CMD_ACQUIRE |
-+ (num & BM_MCC_VERB_ACQUIRE_BUFCOUNT));
-+ while (!(mcr = bm_mc_result(&p->p)))
-+ cpu_relax();
-+ ret = mcr->verb & BM_MCR_VERB_ACQUIRE_BUFCOUNT;
-+ if (bufs) {
-+ for (i = 0; i < num; i++)
-+ bufs[i].opaque =
-+ be64_to_cpu(mcr->acquire.bufs[i].opaque);
-+ }
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+ put_affine_portal();
-+ if (ret != num)
-+ ret = -ENOMEM;
-+ return ret;
-+}
-+
-+int bman_acquire(struct bman_pool *pool, struct bm_buffer *bufs, u8 num,
-+ u32 flags)
-+{
-+ int ret;
-+#ifdef CONFIG_FSL_DPA_CHECKING
-+ if (!num || (num > 8))
-+ return -EINVAL;
-+ if (pool->params.flags & BMAN_POOL_FLAG_ONLY_RELEASE)
-+ return -EINVAL;
-+#endif
-+ /* Without stockpile, this API is a pass-through to the h/w operation */
-+ if (!(pool->params.flags & BMAN_POOL_FLAG_STOCKPILE))
-+ return __bman_acquire(pool, bufs, num);
-+#ifdef CONFIG_FSL_DPA_CHECKING
-+ if (!atomic_dec_and_test(&pool->in_use)) {
-+ pr_crit("Parallel attempts to enter bman_acquire() detected.");
-+ panic("only one instance of bman_released/acquired allowed");
-+ }
-+#endif
-+ /* Two movements of buffers are possible, and can occur in either order.
-+ * A: moving buffers from stockpile to the caller.
-+ * B: moving buffers from hardware to the stockpile.
-+ * Order 1: if there are already enough buffers in the stockpile for A
-+ * then we want to do A first, and only do B if we trigger the
-+ * stockpile-low threshold.
-+ * Order 2: if there are not enough buffers in the stockpile for A,
-+ * then we want to do B first, then do A if B had succeeded. However in
-+ * this case B is dependent on how many buffers the user needs, not the
-+ * stockpile-low threshold.
-+ * Due to the different handling of B between the two cases, putting A
-+ * and B in a while() loop would require quite obscure logic, so handle
-+ * the different sequences explicitly. */
-+ if (num <= pool->sp_fill) {
-+ /* Order 1: do A */
-+ copy_words(bufs, pool->sp + (pool->sp_fill - num),
-+ sizeof(struct bm_buffer) * num);
-+ pool->sp_fill -= num;
-+ /* do B relative to STOCKPILE_LOW */
-+ while (pool->sp_fill <= BMAN_STOCKPILE_LOW) {
-+ ret = __bman_acquire(pool, pool->sp + pool->sp_fill, 8);
-+ if (ret < 0)
-+ ret = __bman_acquire(pool,
-+ pool->sp + pool->sp_fill, 1);
-+ if (ret < 0)
-+ break;
-+ pool->sp_fill += ret;
-+ }
-+ } else {
-+ /* Order 2: do B relative to 'num' */
-+ do {
-+ ret = __bman_acquire(pool, pool->sp + pool->sp_fill, 8);
-+ if (ret < 0)
-+ ret = __bman_acquire(pool,
-+ pool->sp + pool->sp_fill, 1);
-+ if (ret < 0)
-+ /* failure */
-+ goto acquire_done;
-+ pool->sp_fill += ret;
-+ } while (pool->sp_fill < num);
-+ /* do A */
-+ copy_words(bufs, pool->sp + (pool->sp_fill - num),
-+ sizeof(struct bm_buffer) * num);
-+ pool->sp_fill -= num;
-+ }
-+ /* success */
-+ ret = num;
-+acquire_done:
-+#ifdef CONFIG_FSL_DPA_CHECKING
-+ atomic_inc(&pool->in_use);
-+#endif
-+ return ret;
-+}
-+EXPORT_SYMBOL(bman_acquire);
-+
-+int bman_flush_stockpile(struct bman_pool *pool, u32 flags)
-+{
-+ u8 num;
-+ int ret;
-+
-+ while (pool->sp_fill) {
-+ num = ((pool->sp_fill > 8) ? 8 : pool->sp_fill);
-+ ret = __bman_release(pool, pool->sp + (pool->sp_fill - num),
-+ num, flags);
-+ if (ret)
-+ return ret;
-+ pool->sp_fill -= num;
-+ }
-+ return 0;
-+}
-+EXPORT_SYMBOL(bman_flush_stockpile);
-+
-+int bman_query_pools(struct bm_pool_state *state)
-+{
-+ struct bman_portal *p = get_affine_portal();
-+ struct bm_mc_result *mcr;
-+ __maybe_unused unsigned long irqflags;
-+
-+ PORTAL_IRQ_LOCK(p, irqflags);
-+ bm_mc_start(&p->p);
-+ bm_mc_commit(&p->p, BM_MCC_VERB_CMD_QUERY);
-+ while (!(mcr = bm_mc_result(&p->p)))
-+ cpu_relax();
-+ DPA_ASSERT((mcr->verb & BM_MCR_VERB_CMD_MASK) == BM_MCR_VERB_CMD_QUERY);
-+ *state = mcr->query;
-+ state->as.state.__state[0] = be32_to_cpu(state->as.state.__state[0]);
-+ state->as.state.__state[1] = be32_to_cpu(state->as.state.__state[1]);
-+ state->ds.state.__state[0] = be32_to_cpu(state->ds.state.__state[0]);
-+ state->ds.state.__state[1] = be32_to_cpu(state->ds.state.__state[1]);
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+ put_affine_portal();
-+ return 0;
-+}
-+EXPORT_SYMBOL(bman_query_pools);
-+
-+#ifdef CONFIG_FSL_BMAN_CONFIG
-+u32 bman_query_free_buffers(struct bman_pool *pool)
-+{
-+ return bm_pool_free_buffers(pool->params.bpid);
-+}
-+EXPORT_SYMBOL(bman_query_free_buffers);
-+
-+int bman_update_pool_thresholds(struct bman_pool *pool, const u32 *thresholds)
-+{
-+ u32 bpid;
-+
-+ bpid = bman_get_params(pool)->bpid;
-+
-+ return bm_pool_set(bpid, thresholds);
-+}
-+EXPORT_SYMBOL(bman_update_pool_thresholds);
-+#endif
-+
-+int bman_shutdown_pool(u32 bpid)
-+{
-+ struct bman_portal *p = get_affine_portal();
-+ __maybe_unused unsigned long irqflags;
-+ int ret;
-+
-+ PORTAL_IRQ_LOCK(p, irqflags);
-+ ret = bm_shutdown_pool(&p->p, bpid);
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+ put_affine_portal();
-+ return ret;
-+}
-+EXPORT_SYMBOL(bman_shutdown_pool);
-+
-+const struct bm_portal_config *bman_get_bm_portal_config(
-+ struct bman_portal *portal)
-+{
-+ return portal->sharing_redirect ? NULL : portal->config;
-+}
---- /dev/null
-+++ b/drivers/staging/fsl_qbman/bman_low.h
-@@ -0,0 +1,565 @@
-+/* Copyright 2008-2011 Freescale Semiconductor, Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#include "bman_private.h"
-+
-+/***************************/
-+/* Portal register assists */
-+/***************************/
-+
-+#if defined(CONFIG_PPC32) || defined(CONFIG_PPC64)
-+
-+/* Cache-inhibited register offsets */
-+#define BM_REG_RCR_PI_CINH 0x0000
-+#define BM_REG_RCR_CI_CINH 0x0004
-+#define BM_REG_RCR_ITR 0x0008
-+#define BM_REG_CFG 0x0100
-+#define BM_REG_SCN(n) (0x0200 + ((n) << 2))
-+#define BM_REG_ISR 0x0e00
-+#define BM_REG_IIR 0x0e0c
-+
-+/* Cache-enabled register offsets */
-+#define BM_CL_CR 0x0000
-+#define BM_CL_RR0 0x0100
-+#define BM_CL_RR1 0x0140
-+#define BM_CL_RCR 0x1000
-+#define BM_CL_RCR_PI_CENA 0x3000
-+#define BM_CL_RCR_CI_CENA 0x3100
-+
-+#endif
-+
-+#if defined(CONFIG_ARM) || defined(CONFIG_ARM64)
-+
-+/* Cache-inhibited register offsets */
-+#define BM_REG_RCR_PI_CINH 0x3000
-+#define BM_REG_RCR_CI_CINH 0x3100
-+#define BM_REG_RCR_ITR 0x3200
-+#define BM_REG_CFG 0x3300
-+#define BM_REG_SCN(n) (0x3400 + ((n) << 6))
-+#define BM_REG_ISR 0x3e00
-+#define BM_REG_IIR 0x3ec0
-+
-+/* Cache-enabled register offsets */
-+#define BM_CL_CR 0x0000
-+#define BM_CL_RR0 0x0100
-+#define BM_CL_RR1 0x0140
-+#define BM_CL_RCR 0x1000
-+#define BM_CL_RCR_PI_CENA 0x3000
-+#define BM_CL_RCR_CI_CENA 0x3100
-+
-+#endif
-+
-+/* BTW, the drivers (and h/w programming model) already obtain the required
-+ * synchronisation for portal accesses via lwsync(), hwsync(), and
-+ * data-dependencies. Use of barrier()s or other order-preserving primitives
-+ * simply degrade performance. Hence the use of the __raw_*() interfaces, which
-+ * simply ensure that the compiler treats the portal registers as volatile (ie.
-+ * non-coherent). */
-+
-+/* Cache-inhibited register access. */
-+#define __bm_in(bm, o) be32_to_cpu(__raw_readl((bm)->addr_ci + (o)))
-+#define __bm_out(bm, o, val) __raw_writel(cpu_to_be32(val), \
-+ (bm)->addr_ci + (o));
-+#define bm_in(reg) __bm_in(&portal->addr, BM_REG_##reg)
-+#define bm_out(reg, val) __bm_out(&portal->addr, BM_REG_##reg, val)
-+
-+/* Cache-enabled (index) register access */
-+#define __bm_cl_touch_ro(bm, o) dcbt_ro((bm)->addr_ce + (o))
-+#define __bm_cl_touch_rw(bm, o) dcbt_rw((bm)->addr_ce + (o))
-+#define __bm_cl_in(bm, o) be32_to_cpu(__raw_readl((bm)->addr_ce + (o)))
-+#define __bm_cl_out(bm, o, val) \
-+ do { \
-+ u32 *__tmpclout = (bm)->addr_ce + (o); \
-+ __raw_writel(cpu_to_be32(val), __tmpclout); \
-+ dcbf(__tmpclout); \
-+ } while (0)
-+#define __bm_cl_invalidate(bm, o) dcbi((bm)->addr_ce + (o))
-+#define bm_cl_touch_ro(reg) __bm_cl_touch_ro(&portal->addr, BM_CL_##reg##_CENA)
-+#define bm_cl_touch_rw(reg) __bm_cl_touch_rw(&portal->addr, BM_CL_##reg##_CENA)
-+#define bm_cl_in(reg) __bm_cl_in(&portal->addr, BM_CL_##reg##_CENA)
-+#define bm_cl_out(reg, val) __bm_cl_out(&portal->addr, BM_CL_##reg##_CENA, val)
-+#define bm_cl_invalidate(reg)\
-+ __bm_cl_invalidate(&portal->addr, BM_CL_##reg##_CENA)
-+
-+/* Cyclic helper for rings. FIXME: once we are able to do fine-grain perf
-+ * analysis, look at using the "extra" bit in the ring index registers to avoid
-+ * cyclic issues. */
-+static inline u8 bm_cyc_diff(u8 ringsize, u8 first, u8 last)
-+{
-+ /* 'first' is included, 'last' is excluded */
-+ if (first <= last)
-+ return last - first;
-+ return ringsize + last - first;
-+}
-+
-+/* Portal modes.
-+ * Enum types;
-+ * pmode == production mode
-+ * cmode == consumption mode,
-+ * Enum values use 3 letter codes. First letter matches the portal mode,
-+ * remaining two letters indicate;
-+ * ci == cache-inhibited portal register
-+ * ce == cache-enabled portal register
-+ * vb == in-band valid-bit (cache-enabled)
-+ */
-+enum bm_rcr_pmode { /* matches BCSP_CFG::RPM */
-+ bm_rcr_pci = 0, /* PI index, cache-inhibited */
-+ bm_rcr_pce = 1, /* PI index, cache-enabled */
-+ bm_rcr_pvb = 2 /* valid-bit */
-+};
-+enum bm_rcr_cmode { /* s/w-only */
-+ bm_rcr_cci, /* CI index, cache-inhibited */
-+ bm_rcr_cce /* CI index, cache-enabled */
-+};
-+
-+
-+/* ------------------------- */
-+/* --- Portal structures --- */
-+
-+#define BM_RCR_SIZE 8
-+
-+struct bm_rcr {
-+ struct bm_rcr_entry *ring, *cursor;
-+ u8 ci, available, ithresh, vbit;
-+#ifdef CONFIG_FSL_DPA_CHECKING
-+ u32 busy;
-+ enum bm_rcr_pmode pmode;
-+ enum bm_rcr_cmode cmode;
-+#endif
-+};
-+
-+struct bm_mc {
-+ struct bm_mc_command *cr;
-+ struct bm_mc_result *rr;
-+ u8 rridx, vbit;
-+#ifdef CONFIG_FSL_DPA_CHECKING
-+ enum {
-+ /* Can only be _mc_start()ed */
-+ mc_idle,
-+ /* Can only be _mc_commit()ed or _mc_abort()ed */
-+ mc_user,
-+ /* Can only be _mc_retry()ed */
-+ mc_hw
-+ } state;
-+#endif
-+};
-+
-+struct bm_addr {
-+ void __iomem *addr_ce; /* cache-enabled */
-+ void __iomem *addr_ci; /* cache-inhibited */
-+};
-+
-+struct bm_portal {
-+ struct bm_addr addr;
-+ struct bm_rcr rcr;
-+ struct bm_mc mc;
-+ struct bm_portal_config config;
-+} ____cacheline_aligned;
-+
-+
-+/* --------------- */
-+/* --- RCR API --- */
-+
-+/* Bit-wise logic to wrap a ring pointer by clearing the "carry bit" */
-+#define RCR_CARRYCLEAR(p) \
-+ (void *)((unsigned long)(p) & (~(unsigned long)(BM_RCR_SIZE << 6)))
-+
-+/* Bit-wise logic to convert a ring pointer to a ring index */
-+static inline u8 RCR_PTR2IDX(struct bm_rcr_entry *e)
-+{
-+ return ((uintptr_t)e >> 6) & (BM_RCR_SIZE - 1);
-+}
-+
-+/* Increment the 'cursor' ring pointer, taking 'vbit' into account */
-+static inline void RCR_INC(struct bm_rcr *rcr)
-+{
-+ /* NB: this is odd-looking, but experiments show that it generates
-+ * fast code with essentially no branching overheads. We increment to
-+ * the next RCR pointer and handle overflow and 'vbit'. */
-+ struct bm_rcr_entry *partial = rcr->cursor + 1;
-+ rcr->cursor = RCR_CARRYCLEAR(partial);
-+ if (partial != rcr->cursor)
-+ rcr->vbit ^= BM_RCR_VERB_VBIT;
-+}
-+
-+static inline int bm_rcr_init(struct bm_portal *portal, enum bm_rcr_pmode pmode,
-+ __maybe_unused enum bm_rcr_cmode cmode)
-+{
-+ /* This use of 'register', as well as all other occurrences, is because
-+ * it has been observed to generate much faster code with gcc than is
-+ * otherwise the case. */
-+ register struct bm_rcr *rcr = &portal->rcr;
-+ u32 cfg;
-+ u8 pi;
-+
-+ rcr->ring = portal->addr.addr_ce + BM_CL_RCR;
-+ rcr->ci = bm_in(RCR_CI_CINH) & (BM_RCR_SIZE - 1);
-+
-+ pi = bm_in(RCR_PI_CINH) & (BM_RCR_SIZE - 1);
-+ rcr->cursor = rcr->ring + pi;
-+ rcr->vbit = (bm_in(RCR_PI_CINH) & BM_RCR_SIZE) ? BM_RCR_VERB_VBIT : 0;
-+ rcr->available = BM_RCR_SIZE - 1
-+ - bm_cyc_diff(BM_RCR_SIZE, rcr->ci, pi);
-+ rcr->ithresh = bm_in(RCR_ITR);
-+#ifdef CONFIG_FSL_DPA_CHECKING
-+ rcr->busy = 0;
-+ rcr->pmode = pmode;
-+ rcr->cmode = cmode;
-+#endif
-+ cfg = (bm_in(CFG) & 0xffffffe0) | (pmode & 0x3); /* BCSP_CFG::RPM */
-+ bm_out(CFG, cfg);
-+ return 0;
-+}
-+
-+static inline void bm_rcr_finish(struct bm_portal *portal)
-+{
-+ register struct bm_rcr *rcr = &portal->rcr;
-+ u8 pi = bm_in(RCR_PI_CINH) & (BM_RCR_SIZE - 1);
-+ u8 ci = bm_in(RCR_CI_CINH) & (BM_RCR_SIZE - 1);
-+ DPA_ASSERT(!rcr->busy);
-+ if (pi != RCR_PTR2IDX(rcr->cursor))
-+ pr_crit("losing uncommited RCR entries\n");
-+ if (ci != rcr->ci)
-+ pr_crit("missing existing RCR completions\n");
-+ if (rcr->ci != RCR_PTR2IDX(rcr->cursor))
-+ pr_crit("RCR destroyed unquiesced\n");
-+}
-+
-+static inline struct bm_rcr_entry *bm_rcr_start(struct bm_portal *portal)
-+{
-+ register struct bm_rcr *rcr = &portal->rcr;
-+ DPA_ASSERT(!rcr->busy);
-+ if (!rcr->available)
-+ return NULL;
-+#ifdef CONFIG_FSL_DPA_CHECKING
-+ rcr->busy = 1;
-+#endif
-+#if defined(CONFIG_PPC32) || defined(CONFIG_PPC64)
-+ dcbz_64(rcr->cursor);
-+#endif
-+ return rcr->cursor;
-+}
-+
-+static inline void bm_rcr_abort(struct bm_portal *portal)
-+{
-+ __maybe_unused register struct bm_rcr *rcr = &portal->rcr;
-+ DPA_ASSERT(rcr->busy);
-+#ifdef CONFIG_FSL_DPA_CHECKING
-+ rcr->busy = 0;
-+#endif
-+}
-+
-+static inline struct bm_rcr_entry *bm_rcr_pend_and_next(
-+ struct bm_portal *portal, u8 myverb)
-+{
-+ register struct bm_rcr *rcr = &portal->rcr;
-+ DPA_ASSERT(rcr->busy);
-+ DPA_ASSERT(rcr->pmode != bm_rcr_pvb);
-+ if (rcr->available == 1)
-+ return NULL;
-+ rcr->cursor->__dont_write_directly__verb = myverb | rcr->vbit;
-+ dcbf_64(rcr->cursor);
-+ RCR_INC(rcr);
-+ rcr->available--;
-+#if defined(CONFIG_PPC32) || defined(CONFIG_PPC64)
-+ dcbz_64(rcr->cursor);
-+#endif
-+ return rcr->cursor;
-+}
-+
-+static inline void bm_rcr_pci_commit(struct bm_portal *portal, u8 myverb)
-+{
-+ register struct bm_rcr *rcr = &portal->rcr;
-+ DPA_ASSERT(rcr->busy);
-+ DPA_ASSERT(rcr->pmode == bm_rcr_pci);
-+ rcr->cursor->__dont_write_directly__verb = myverb | rcr->vbit;
-+ RCR_INC(rcr);
-+ rcr->available--;
-+ hwsync();
-+ bm_out(RCR_PI_CINH, RCR_PTR2IDX(rcr->cursor));
-+#ifdef CONFIG_FSL_DPA_CHECKING
-+ rcr->busy = 0;
-+#endif
-+}
-+
-+static inline void bm_rcr_pce_prefetch(struct bm_portal *portal)
-+{
-+ __maybe_unused register struct bm_rcr *rcr = &portal->rcr;
-+ DPA_ASSERT(rcr->pmode == bm_rcr_pce);
-+ bm_cl_invalidate(RCR_PI);
-+ bm_cl_touch_rw(RCR_PI);
-+}
-+
-+static inline void bm_rcr_pce_commit(struct bm_portal *portal, u8 myverb)
-+{
-+ register struct bm_rcr *rcr = &portal->rcr;
-+ DPA_ASSERT(rcr->busy);
-+ DPA_ASSERT(rcr->pmode == bm_rcr_pce);
-+ rcr->cursor->__dont_write_directly__verb = myverb | rcr->vbit;
-+ RCR_INC(rcr);
-+ rcr->available--;
-+ lwsync();
-+ bm_cl_out(RCR_PI, RCR_PTR2IDX(rcr->cursor));
-+#ifdef CONFIG_FSL_DPA_CHECKING
-+ rcr->busy = 0;
-+#endif
-+}
-+
-+static inline void bm_rcr_pvb_commit(struct bm_portal *portal, u8 myverb)
-+{
-+ register struct bm_rcr *rcr = &portal->rcr;
-+ struct bm_rcr_entry *rcursor;
-+ DPA_ASSERT(rcr->busy);
-+ DPA_ASSERT(rcr->pmode == bm_rcr_pvb);
-+ lwsync();
-+ rcursor = rcr->cursor;
-+ rcursor->__dont_write_directly__verb = myverb | rcr->vbit;
-+ dcbf_64(rcursor);
-+ RCR_INC(rcr);
-+ rcr->available--;
-+#ifdef CONFIG_FSL_DPA_CHECKING
-+ rcr->busy = 0;
-+#endif
-+}
-+
-+static inline u8 bm_rcr_cci_update(struct bm_portal *portal)
-+{
-+ register struct bm_rcr *rcr = &portal->rcr;
-+ u8 diff, old_ci = rcr->ci;
-+ DPA_ASSERT(rcr->cmode == bm_rcr_cci);
-+ rcr->ci = bm_in(RCR_CI_CINH) & (BM_RCR_SIZE - 1);
-+ diff = bm_cyc_diff(BM_RCR_SIZE, old_ci, rcr->ci);
-+ rcr->available += diff;
-+ return diff;
-+}
-+
-+static inline void bm_rcr_cce_prefetch(struct bm_portal *portal)
-+{
-+ __maybe_unused register struct bm_rcr *rcr = &portal->rcr;
-+ DPA_ASSERT(rcr->cmode == bm_rcr_cce);
-+ bm_cl_touch_ro(RCR_CI);
-+}
-+
-+static inline u8 bm_rcr_cce_update(struct bm_portal *portal)
-+{
-+ register struct bm_rcr *rcr = &portal->rcr;
-+ u8 diff, old_ci = rcr->ci;
-+ DPA_ASSERT(rcr->cmode == bm_rcr_cce);
-+ rcr->ci = bm_cl_in(RCR_CI) & (BM_RCR_SIZE - 1);
-+ bm_cl_invalidate(RCR_CI);
-+ diff = bm_cyc_diff(BM_RCR_SIZE, old_ci, rcr->ci);
-+ rcr->available += diff;
-+ return diff;
-+}
-+
-+static inline u8 bm_rcr_get_ithresh(struct bm_portal *portal)
-+{
-+ register struct bm_rcr *rcr = &portal->rcr;
-+ return rcr->ithresh;
-+}
-+
-+static inline void bm_rcr_set_ithresh(struct bm_portal *portal, u8 ithresh)
-+{
-+ register struct bm_rcr *rcr = &portal->rcr;
-+ rcr->ithresh = ithresh;
-+ bm_out(RCR_ITR, ithresh);
-+}
-+
-+static inline u8 bm_rcr_get_avail(struct bm_portal *portal)
-+{
-+ register struct bm_rcr *rcr = &portal->rcr;
-+ return rcr->available;
-+}
-+
-+static inline u8 bm_rcr_get_fill(struct bm_portal *portal)
-+{
-+ register struct bm_rcr *rcr = &portal->rcr;
-+ return BM_RCR_SIZE - 1 - rcr->available;
-+}
-+
-+
-+/* ------------------------------ */
-+/* --- Management command API --- */
-+
-+static inline int bm_mc_init(struct bm_portal *portal)
-+{
-+ register struct bm_mc *mc = &portal->mc;
-+ mc->cr = portal->addr.addr_ce + BM_CL_CR;
-+ mc->rr = portal->addr.addr_ce + BM_CL_RR0;
-+ mc->rridx = (__raw_readb(&mc->cr->__dont_write_directly__verb) &
-+ BM_MCC_VERB_VBIT) ? 0 : 1;
-+ mc->vbit = mc->rridx ? BM_MCC_VERB_VBIT : 0;
-+#ifdef CONFIG_FSL_DPA_CHECKING
-+ mc->state = mc_idle;
-+#endif
-+ return 0;
-+}
-+
-+static inline void bm_mc_finish(struct bm_portal *portal)
-+{
-+ __maybe_unused register struct bm_mc *mc = &portal->mc;
-+ DPA_ASSERT(mc->state == mc_idle);
-+#ifdef CONFIG_FSL_DPA_CHECKING
-+ if (mc->state != mc_idle)
-+ pr_crit("Losing incomplete MC command\n");
-+#endif
-+}
-+
-+static inline struct bm_mc_command *bm_mc_start(struct bm_portal *portal)
-+{
-+ register struct bm_mc *mc = &portal->mc;
-+ DPA_ASSERT(mc->state == mc_idle);
-+#ifdef CONFIG_FSL_DPA_CHECKING
-+ mc->state = mc_user;
-+#endif
-+#if defined(CONFIG_PPC32) || defined(CONFIG_PPC64)
-+ dcbz_64(mc->cr);
-+#endif
-+ return mc->cr;
-+}
-+
-+static inline void bm_mc_abort(struct bm_portal *portal)
-+{
-+ __maybe_unused register struct bm_mc *mc = &portal->mc;
-+ DPA_ASSERT(mc->state == mc_user);
-+#ifdef CONFIG_FSL_DPA_CHECKING
-+ mc->state = mc_idle;
-+#endif
-+}
-+
-+static inline void bm_mc_commit(struct bm_portal *portal, u8 myverb)
-+{
-+ register struct bm_mc *mc = &portal->mc;
-+ struct bm_mc_result *rr = mc->rr + mc->rridx;
-+ DPA_ASSERT(mc->state == mc_user);
-+ lwsync();
-+ mc->cr->__dont_write_directly__verb = myverb | mc->vbit;
-+ dcbf(mc->cr);
-+ dcbit_ro(rr);
-+#ifdef CONFIG_FSL_DPA_CHECKING
-+ mc->state = mc_hw;
-+#endif
-+}
-+
-+static inline struct bm_mc_result *bm_mc_result(struct bm_portal *portal)
-+{
-+ register struct bm_mc *mc = &portal->mc;
-+ struct bm_mc_result *rr = mc->rr + mc->rridx;
-+ DPA_ASSERT(mc->state == mc_hw);
-+ /* The inactive response register's verb byte always returns zero until
-+ * its command is submitted and completed. This includes the valid-bit,
-+ * in case you were wondering... */
-+ if (!__raw_readb(&rr->verb)) {
-+ dcbit_ro(rr);
-+ return NULL;
-+ }
-+ mc->rridx ^= 1;
-+ mc->vbit ^= BM_MCC_VERB_VBIT;
-+#ifdef CONFIG_FSL_DPA_CHECKING
-+ mc->state = mc_idle;
-+#endif
-+ return rr;
-+}
-+
-+
-+/* ------------------------------------- */
-+/* --- Portal interrupt register API --- */
-+
-+static inline int bm_isr_init(__always_unused struct bm_portal *portal)
-+{
-+ return 0;
-+}
-+
-+static inline void bm_isr_finish(__always_unused struct bm_portal *portal)
-+{
-+}
-+
-+#define SCN_REG(bpid) BM_REG_SCN((bpid) / 32)
-+#define SCN_BIT(bpid) (0x80000000 >> (bpid & 31))
-+static inline void bm_isr_bscn_mask(struct bm_portal *portal, u8 bpid,
-+ int enable)
-+{
-+ u32 val;
-+ DPA_ASSERT(bpid < bman_pool_max);
-+ /* REG_SCN for bpid=0..31, REG_SCN+4 for bpid=32..63 */
-+ val = __bm_in(&portal->addr, SCN_REG(bpid));
-+ if (enable)
-+ val |= SCN_BIT(bpid);
-+ else
-+ val &= ~SCN_BIT(bpid);
-+ __bm_out(&portal->addr, SCN_REG(bpid), val);
-+}
-+
-+static inline u32 __bm_isr_read(struct bm_portal *portal, enum bm_isr_reg n)
-+{
-+#if defined(CONFIG_ARM) || defined(CONFIG_ARM64)
-+ return __bm_in(&portal->addr, BM_REG_ISR + (n << 6));
-+#else
-+ return __bm_in(&portal->addr, BM_REG_ISR + (n << 2));
-+#endif
-+}
-+
-+static inline void __bm_isr_write(struct bm_portal *portal, enum bm_isr_reg n,
-+ u32 val)
-+{
-+#if defined(CONFIG_ARM) || defined(CONFIG_ARM64)
-+ __bm_out(&portal->addr, BM_REG_ISR + (n << 6), val);
-+#else
-+ __bm_out(&portal->addr, BM_REG_ISR + (n << 2), val);
-+#endif
-+}
-+
-+/* Buffer Pool Cleanup */
-+static inline int bm_shutdown_pool(struct bm_portal *p, u32 bpid)
-+{
-+ struct bm_mc_command *bm_cmd;
-+ struct bm_mc_result *bm_res;
-+
-+ int aq_count = 0;
-+ bool stop = false;
-+ while (!stop) {
-+ /* Acquire buffers until empty */
-+ bm_cmd = bm_mc_start(p);
-+ bm_cmd->acquire.bpid = bpid;
-+ bm_mc_commit(p, BM_MCC_VERB_CMD_ACQUIRE | 1);
-+ while (!(bm_res = bm_mc_result(p)))
-+ cpu_relax();
-+ if (!(bm_res->verb & BM_MCR_VERB_ACQUIRE_BUFCOUNT)) {
-+ /* Pool is empty */
-+ /* TBD : Should we do a few extra iterations in
-+ case some other some blocks keep buffers 'on deck',
-+ which may also be problematic */
-+ stop = true;
-+ } else
-+ ++aq_count;
-+ }
-+ return 0;
-+}
---- /dev/null
-+++ b/drivers/staging/fsl_qbman/bman_private.h
-@@ -0,0 +1,166 @@
-+/* Copyright 2008-2012 Freescale Semiconductor, Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#include "dpa_sys.h"
-+#include <linux/fsl_bman.h>
-+
-+/* Revision info (for errata and feature handling) */
-+#define BMAN_REV10 0x0100
-+#define BMAN_REV20 0x0200
-+#define BMAN_REV21 0x0201
-+#define QBMAN_ANY_PORTAL_IDX 0xffffffff
-+extern u16 bman_ip_rev; /* 0 if uninitialised, otherwise QMAN_REVx */
-+
-+/*
-+ * Global variables of the max portal/pool number this bman version supported
-+ */
-+extern u16 bman_pool_max;
-+
-+/* used by CCSR and portal interrupt code */
-+enum bm_isr_reg {
-+ bm_isr_status = 0,
-+ bm_isr_enable = 1,
-+ bm_isr_disable = 2,
-+ bm_isr_inhibit = 3
-+};
-+
-+struct bm_portal_config {
-+ /* Corenet portal addresses;
-+ * [0]==cache-enabled, [1]==cache-inhibited. */
-+ __iomem void *addr_virt[2];
-+ struct resource addr_phys[2];
-+ /* Allow these to be joined in lists */
-+ struct list_head list;
-+ /* User-visible portal configuration settings */
-+ struct bman_portal_config public_cfg;
-+ /* power management saved data */
-+ u32 saved_isdr;
-+};
-+
-+#ifdef CONFIG_FSL_BMAN_CONFIG
-+/* Hooks from bman_driver.c to bman_config.c */
-+int bman_init_ccsr(struct device_node *node);
-+#endif
-+
-+/* Hooks from bman_driver.c in to bman_high.c */
-+struct bman_portal *bman_create_portal(
-+ struct bman_portal *portal,
-+ const struct bm_portal_config *config);
-+struct bman_portal *bman_create_affine_portal(
-+ const struct bm_portal_config *config);
-+struct bman_portal *bman_create_affine_slave(struct bman_portal *redirect,
-+ int cpu);
-+void bman_destroy_portal(struct bman_portal *bm);
-+
-+const struct bm_portal_config *bman_destroy_affine_portal(void);
-+
-+/* Hooks from fsl_usdpaa.c to bman_driver.c */
-+struct bm_portal_config *bm_get_unused_portal(void);
-+struct bm_portal_config *bm_get_unused_portal_idx(uint32_t idx);
-+void bm_put_unused_portal(struct bm_portal_config *pcfg);
-+void bm_set_liodns(struct bm_portal_config *pcfg);
-+
-+/* Pool logic in the portal driver, during initialisation, needs to know if
-+ * there's access to CCSR or not (if not, it'll cripple the pool allocator). */
-+#ifdef CONFIG_FSL_BMAN_CONFIG
-+int bman_have_ccsr(void);
-+#else
-+#define bman_have_ccsr() 0
-+#endif
-+
-+/* Stockpile build constants. The _LOW value: when bman_acquire() is called and
-+ * the stockpile fill-level is <= _LOW, an acquire is attempted from h/w but it
-+ * might fail (if the buffer pool is depleted). So this value provides some
-+ * "stagger" in that the bman_acquire() function will only fail if lots of bufs
-+ * are requested at once or if h/w has been tested a couple of times without
-+ * luck. The _HIGH value: when bman_release() is called and the stockpile
-+ * fill-level is >= _HIGH, a release is attempted to h/w but it might fail (if
-+ * the release ring is full). So this value provides some "stagger" so that
-+ * ring-access is retried a couple of times prior to the API returning a
-+ * failure. The following *must* be true;
-+ * BMAN_STOCKPILE_HIGH-BMAN_STOCKPILE_LOW > 8
-+ * (to avoid thrashing)
-+ * BMAN_STOCKPILE_SZ >= 16
-+ * (as the release logic expects to either send 8 buffers to hw prior to
-+ * adding the given buffers to the stockpile or add the buffers to the
-+ * stockpile before sending 8 to hw, as the API must be an all-or-nothing
-+ * success/fail.)
-+ */
-+#define BMAN_STOCKPILE_SZ 16u /* number of bufs in per-pool cache */
-+#define BMAN_STOCKPILE_LOW 2u /* when fill is <= this, acquire from hw */
-+#define BMAN_STOCKPILE_HIGH 14u /* when fill is >= this, release to hw */
-+
-+/*************************************************/
-+/* BMan s/w corenet portal, low-level i/face */
-+/*************************************************/
-+
-+/* Used by all portal interrupt registers except 'inhibit'
-+ * This mask contains all the "irqsource" bits visible to API users
-+ */
-+#define BM_PIRQ_VISIBLE (BM_PIRQ_RCRI | BM_PIRQ_BSCN)
-+
-+/* These are bm_<reg>_<verb>(). So for example, bm_disable_write() means "write
-+ * the disable register" rather than "disable the ability to write". */
-+#define bm_isr_status_read(bm) __bm_isr_read(bm, bm_isr_status)
-+#define bm_isr_status_clear(bm, m) __bm_isr_write(bm, bm_isr_status, m)
-+#define bm_isr_enable_read(bm) __bm_isr_read(bm, bm_isr_enable)
-+#define bm_isr_enable_write(bm, v) __bm_isr_write(bm, bm_isr_enable, v)
-+#define bm_isr_disable_read(bm) __bm_isr_read(bm, bm_isr_disable)
-+#define bm_isr_disable_write(bm, v) __bm_isr_write(bm, bm_isr_disable, v)
-+#define bm_isr_inhibit(bm) __bm_isr_write(bm, bm_isr_inhibit, 1)
-+#define bm_isr_uninhibit(bm) __bm_isr_write(bm, bm_isr_inhibit, 0)
-+
-+#ifdef CONFIG_FSL_BMAN_CONFIG
-+/* Set depletion thresholds associated with a buffer pool. Requires that the
-+ * operating system have access to Bman CCSR (ie. compiled in support and
-+ * run-time access courtesy of the device-tree). */
-+int bm_pool_set(u32 bpid, const u32 *thresholds);
-+#define BM_POOL_THRESH_SW_ENTER 0
-+#define BM_POOL_THRESH_SW_EXIT 1
-+#define BM_POOL_THRESH_HW_ENTER 2
-+#define BM_POOL_THRESH_HW_EXIT 3
-+
-+/* Read the free buffer count for a given buffer */
-+u32 bm_pool_free_buffers(u32 bpid);
-+
-+__init int bman_init(void);
-+__init int bman_resource_init(void);
-+
-+const struct bm_portal_config *bman_get_bm_portal_config(
-+ struct bman_portal *portal);
-+
-+/* power management */
-+#ifdef CONFIG_SUSPEND
-+void suspend_unused_bportal(void);
-+void resume_unused_bportal(void);
-+#endif
-+
-+#endif /* CONFIG_FSL_BMAN_CONFIG */
---- /dev/null
-+++ b/drivers/staging/fsl_qbman/bman_test.c
-@@ -0,0 +1,56 @@
-+/* Copyright 2008-2011 Freescale Semiconductor, Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#include "bman_test.h"
-+
-+MODULE_AUTHOR("Geoff Thorpe");
-+MODULE_LICENSE("Dual BSD/GPL");
-+MODULE_DESCRIPTION("Bman testing");
-+
-+static int test_init(void)
-+{
-+#ifdef CONFIG_FSL_BMAN_TEST_HIGH
-+ int loop = 1;
-+ while (loop--)
-+ bman_test_high();
-+#endif
-+#ifdef CONFIG_FSL_BMAN_TEST_THRESH
-+ bman_test_thresh();
-+#endif
-+ return 0;
-+}
-+
-+static void test_exit(void)
-+{
-+}
-+
-+module_init(test_init);
-+module_exit(test_exit);
---- /dev/null
-+++ b/drivers/staging/fsl_qbman/bman_test.h
-@@ -0,0 +1,44 @@
-+/* Copyright 2008-2011 Freescale Semiconductor, Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#include <linux/kernel.h>
-+#include <linux/errno.h>
-+#include <linux/io.h>
-+#include <linux/slab.h>
-+#include <linux/module.h>
-+#include <linux/interrupt.h>
-+#include <linux/delay.h>
-+#include <linux/kthread.h>
-+
-+#include <linux/fsl_bman.h>
-+
-+void bman_test_high(void);
-+void bman_test_thresh(void);
---- /dev/null
-+++ b/drivers/staging/fsl_qbman/bman_test_high.c
-@@ -0,0 +1,183 @@
-+/* Copyright 2008-2011 Freescale Semiconductor, Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#include "bman_test.h"
-+#include "bman_private.h"
-+
-+/*************/
-+/* constants */
-+/*************/
-+
-+#define PORTAL_OPAQUE ((void *)0xf00dbeef)
-+#define POOL_OPAQUE ((void *)0xdeadabba)
-+#define NUM_BUFS 93
-+#define LOOPS 3
-+#define BMAN_TOKEN_MASK 0x00FFFFFFFFFFLLU
-+
-+/***************/
-+/* global vars */
-+/***************/
-+
-+static struct bman_pool *pool;
-+static int depleted;
-+static struct bm_buffer bufs_in[NUM_BUFS] ____cacheline_aligned;
-+static struct bm_buffer bufs_out[NUM_BUFS] ____cacheline_aligned;
-+static int bufs_received;
-+
-+/* Predeclare the callback so we can instantiate pool parameters */
-+static void depletion_cb(struct bman_portal *, struct bman_pool *, void *, int);
-+
-+/**********************/
-+/* internal functions */
-+/**********************/
-+
-+static void bufs_init(void)
-+{
-+ int i;
-+ for (i = 0; i < NUM_BUFS; i++)
-+ bm_buffer_set64(&bufs_in[i], 0xfedc01234567LLU * i);
-+ bufs_received = 0;
-+}
-+
-+static inline int bufs_cmp(const struct bm_buffer *a, const struct bm_buffer *b)
-+{
-+ if ((bman_ip_rev == BMAN_REV20) || (bman_ip_rev == BMAN_REV21)) {
-+
-+ /* On SoCs with Bman revison 2.0, Bman only respects the 40
-+ * LS-bits of buffer addresses, masking off the upper 8-bits on
-+ * release commands. The API provides for 48-bit addresses
-+ * because some SoCs support all 48-bits. When generating
-+ * garbage addresses for testing, we either need to zero the
-+ * upper 8-bits when releasing to Bman (otherwise we'll be
-+ * disappointed when the buffers we acquire back from Bman
-+ * don't match), or we need to mask the upper 8-bits off when
-+ * comparing. We do the latter.
-+ */
-+ if ((bm_buffer_get64(a) & BMAN_TOKEN_MASK)
-+ < (bm_buffer_get64(b) & BMAN_TOKEN_MASK))
-+ return -1;
-+ if ((bm_buffer_get64(a) & BMAN_TOKEN_MASK)
-+ > (bm_buffer_get64(b) & BMAN_TOKEN_MASK))
-+ return 1;
-+ } else {
-+ if (bm_buffer_get64(a) < bm_buffer_get64(b))
-+ return -1;
-+ if (bm_buffer_get64(a) > bm_buffer_get64(b))
-+ return 1;
-+ }
-+
-+ return 0;
-+}
-+
-+static void bufs_confirm(void)
-+{
-+ int i, j;
-+ for (i = 0; i < NUM_BUFS; i++) {
-+ int matches = 0;
-+ for (j = 0; j < NUM_BUFS; j++)
-+ if (!bufs_cmp(&bufs_in[i], &bufs_out[j]))
-+ matches++;
-+ BUG_ON(matches != 1);
-+ }
-+}
-+
-+/********/
-+/* test */
-+/********/
-+
-+static void depletion_cb(struct bman_portal *__portal, struct bman_pool *__pool,
-+ void *pool_ctx, int __depleted)
-+{
-+ BUG_ON(__pool != pool);
-+ BUG_ON(pool_ctx != POOL_OPAQUE);
-+ depleted = __depleted;
-+}
-+
-+void bman_test_high(void)
-+{
-+ struct bman_pool_params pparams = {
-+ .flags = BMAN_POOL_FLAG_DEPLETION | BMAN_POOL_FLAG_DYNAMIC_BPID,
-+ .cb = depletion_cb,
-+ .cb_ctx = POOL_OPAQUE,
-+ };
-+ int i, loops = LOOPS;
-+ struct bm_buffer tmp_buf;
-+
-+ bufs_init();
-+
-+ pr_info("BMAN: --- starting high-level test ---\n");
-+
-+ pool = bman_new_pool(&pparams);
-+ BUG_ON(!pool);
-+
-+ /*******************/
-+ /* Release buffers */
-+ /*******************/
-+do_loop:
-+ i = 0;
-+ while (i < NUM_BUFS) {
-+ u32 flags = BMAN_RELEASE_FLAG_WAIT;
-+ int num = 8;
-+ if ((i + num) > NUM_BUFS)
-+ num = NUM_BUFS - i;
-+ if ((i + num) == NUM_BUFS)
-+ flags |= BMAN_RELEASE_FLAG_WAIT_SYNC;
-+ if (bman_release(pool, bufs_in + i, num, flags))
-+ panic("bman_release() failed\n");
-+ i += num;
-+ }
-+
-+ /*******************/
-+ /* Acquire buffers */
-+ /*******************/
-+ while (i > 0) {
-+ int tmp, num = 8;
-+ if (num > i)
-+ num = i;
-+ tmp = bman_acquire(pool, bufs_out + i - num, num, 0);
-+ BUG_ON(tmp != num);
-+ i -= num;
-+ }
-+
-+ i = bman_acquire(pool, &tmp_buf, 1, 0);
-+ BUG_ON(i > 0);
-+
-+ bufs_confirm();
-+
-+ if (--loops)
-+ goto do_loop;
-+
-+ /************/
-+ /* Clean up */
-+ /************/
-+ bman_free_pool(pool);
-+ pr_info("BMAN: --- finished high-level test ---\n");
-+}
---- /dev/null
-+++ b/drivers/staging/fsl_qbman/bman_test_thresh.c
-@@ -0,0 +1,196 @@
-+/* Copyright 2010-2011 Freescale Semiconductor, Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#include "bman_test.h"
-+
-+/* Test constants */
-+#define TEST_NUMBUFS 129728
-+#define TEST_EXIT 129536
-+#define TEST_ENTRY 129024
-+
-+struct affine_test_data {
-+ struct task_struct *t;
-+ int cpu;
-+ int expect_affinity;
-+ int drain;
-+ int num_enter;
-+ int num_exit;
-+ struct list_head node;
-+ struct completion wakethread;
-+ struct completion wakeparent;
-+};
-+
-+static void cb_depletion(struct bman_portal *portal,
-+ struct bman_pool *pool,
-+ void *opaque,
-+ int depleted)
-+{
-+ struct affine_test_data *data = opaque;
-+ int c = smp_processor_id();
-+ pr_info("cb_depletion: bpid=%d, depleted=%d, cpu=%d, original=%d\n",
-+ bman_get_params(pool)->bpid, !!depleted, c, data->cpu);
-+ /* We should be executing on the CPU of the thread that owns the pool if
-+ * and that CPU has an affine portal (ie. it isn't slaved). */
-+ BUG_ON((c != data->cpu) && data->expect_affinity);
-+ BUG_ON((c == data->cpu) && !data->expect_affinity);
-+ if (depleted)
-+ data->num_enter++;
-+ else
-+ data->num_exit++;
-+}
-+
-+/* Params used to set up a pool, this also dynamically allocates a BPID */
-+static const struct bman_pool_params params_nocb = {
-+ .flags = BMAN_POOL_FLAG_DYNAMIC_BPID | BMAN_POOL_FLAG_THRESH,
-+ .thresholds = { TEST_ENTRY, TEST_EXIT, 0, 0 }
-+};
-+
-+/* Params used to set up each cpu's pool with callbacks enabled */
-+static struct bman_pool_params params_cb = {
-+ .bpid = 0, /* will be replaced to match pool_nocb */
-+ .flags = BMAN_POOL_FLAG_DEPLETION,
-+ .cb = cb_depletion
-+};
-+
-+static struct bman_pool *pool_nocb;
-+static LIST_HEAD(threads);
-+
-+static int affine_test(void *__data)
-+{
-+ struct bman_pool *pool;
-+ struct affine_test_data *data = __data;
-+ struct bman_pool_params my_params = params_cb;
-+
-+ pr_info("thread %d: starting\n", data->cpu);
-+ /* create the pool */
-+ my_params.cb_ctx = data;
-+ pool = bman_new_pool(&my_params);
-+ BUG_ON(!pool);
-+ complete(&data->wakeparent);
-+ wait_for_completion(&data->wakethread);
-+ init_completion(&data->wakethread);
-+
-+ /* if we're the drainer, we get signalled for that */
-+ if (data->drain) {
-+ struct bm_buffer buf;
-+ int ret;
-+ pr_info("thread %d: draining...\n", data->cpu);
-+ do {
-+ ret = bman_acquire(pool, &buf, 1, 0);
-+ } while (ret > 0);
-+ pr_info("thread %d: draining done.\n", data->cpu);
-+ complete(&data->wakeparent);
-+ wait_for_completion(&data->wakethread);
-+ init_completion(&data->wakethread);
-+ }
-+
-+ /* cleanup */
-+ bman_free_pool(pool);
-+ while (!kthread_should_stop())
-+ cpu_relax();
-+ pr_info("thread %d: exiting\n", data->cpu);
-+ return 0;
-+}
-+
-+static struct affine_test_data *start_affine_test(int cpu, int drain)
-+{
-+ struct affine_test_data *data = kmalloc(sizeof(*data), GFP_KERNEL);
-+
-+ if (!data)
-+ return NULL;
-+ data->cpu = cpu;
-+ data->expect_affinity = cpumask_test_cpu(cpu, bman_affine_cpus());
-+ data->drain = drain;
-+ data->num_enter = 0;
-+ data->num_exit = 0;
-+ init_completion(&data->wakethread);
-+ init_completion(&data->wakeparent);
-+ list_add_tail(&data->node, &threads);
-+ data->t = kthread_create(affine_test, data, "threshtest%d", cpu);
-+ BUG_ON(IS_ERR(data->t));
-+ kthread_bind(data->t, cpu);
-+ wake_up_process(data->t);
-+ return data;
-+}
-+
-+void bman_test_thresh(void)
-+{
-+ int loop = TEST_NUMBUFS;
-+ int ret, num_cpus = 0;
-+ struct affine_test_data *data, *drainer = NULL;
-+
-+ pr_info("bman_test_thresh: start\n");
-+
-+ /* allocate a BPID and seed it */
-+ pool_nocb = bman_new_pool(&params_nocb);
-+ BUG_ON(!pool_nocb);
-+ while (loop--) {
-+ struct bm_buffer buf;
-+ bm_buffer_set64(&buf, 0x0badbeef + loop);
-+ ret = bman_release(pool_nocb, &buf, 1,
-+ BMAN_RELEASE_FLAG_WAIT);
-+ BUG_ON(ret);
-+ }
-+ while (!bman_rcr_is_empty())
-+ cpu_relax();
-+ pr_info("bman_test_thresh: buffers are in\n");
-+
-+ /* create threads and wait for them to create pools */
-+ params_cb.bpid = bman_get_params(pool_nocb)->bpid;
-+ for_each_cpu(loop, cpu_online_mask) {
-+ data = start_affine_test(loop, drainer ? 0 : 1);
-+ BUG_ON(!data);
-+ if (!drainer)
-+ drainer = data;
-+ num_cpus++;
-+ wait_for_completion(&data->wakeparent);
-+ }
-+
-+ /* signal the drainer to start draining */
-+ complete(&drainer->wakethread);
-+ wait_for_completion(&drainer->wakeparent);
-+ init_completion(&drainer->wakeparent);
-+
-+ /* tear down */
-+ list_for_each_entry_safe(data, drainer, &threads, node) {
-+ complete(&data->wakethread);
-+ ret = kthread_stop(data->t);
-+ BUG_ON(ret);
-+ list_del(&data->node);
-+ /* check that we get the expected callbacks (and no others) */
-+ BUG_ON(data->num_enter != 1);
-+ BUG_ON(data->num_exit != 0);
-+ kfree(data);
-+ }
-+ bman_free_pool(pool_nocb);
-+
-+ pr_info("bman_test_thresh: done\n");
-+}
---- /dev/null
-+++ b/drivers/staging/fsl_qbman/dpa_alloc.c
-@@ -0,0 +1,706 @@
-+/* Copyright 2009-2012 Freescale Semiconductor, Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#include "dpa_sys.h"
-+#include <linux/fsl_qman.h>
-+#include <linux/fsl_bman.h>
-+
-+/* Qman and Bman APIs are front-ends to the common code; */
-+
-+static DECLARE_DPA_ALLOC(bpalloc); /* BPID allocator */
-+static DECLARE_DPA_ALLOC(fqalloc); /* FQID allocator */
-+static DECLARE_DPA_ALLOC(qpalloc); /* pool-channel allocator */
-+static DECLARE_DPA_ALLOC(cgralloc); /* CGR ID allocator */
-+static DECLARE_DPA_ALLOC(ceetm0_challoc); /* CEETM Channel ID allocator */
-+static DECLARE_DPA_ALLOC(ceetm0_lfqidalloc); /* CEETM LFQID allocator */
-+static DECLARE_DPA_ALLOC(ceetm1_challoc); /* CEETM Channel ID allocator */
-+static DECLARE_DPA_ALLOC(ceetm1_lfqidalloc); /* CEETM LFQID allocator */
-+
-+/* This is a sort-of-conditional dpa_alloc_free() routine. Eg. when releasing
-+ * FQIDs (probably from user-space), it can filter out those that aren't in the
-+ * OOS state (better to leak a h/w resource than to crash). This function
-+ * returns the number of invalid IDs that were not released. */
-+static u32 release_id_range(struct dpa_alloc *alloc, u32 id, u32 count,
-+ int (*is_valid)(u32 id))
-+{
-+ int valid_mode = 0;
-+ u32 loop = id, total_invalid = 0;
-+ while (loop < (id + count)) {
-+ int isvalid = is_valid ? is_valid(loop) : 1;
-+ if (!valid_mode) {
-+ /* We're looking for a valid ID to terminate an invalid
-+ * range */
-+ if (isvalid) {
-+ /* We finished a range of invalid IDs, a valid
-+ * range is now underway */
-+ valid_mode = 1;
-+ count -= (loop - id);
-+ id = loop;
-+ } else
-+ total_invalid++;
-+ } else {
-+ /* We're looking for an invalid ID to terminate a
-+ * valid range */
-+ if (!isvalid) {
-+ /* Release the range of valid IDs, an unvalid
-+ * range is now underway */
-+ if (loop > id)
-+ dpa_alloc_free(alloc, id, loop - id);
-+ valid_mode = 0;
-+ }
-+ }
-+ loop++;
-+ }
-+ /* Release any unterminated range of valid IDs */
-+ if (valid_mode && count)
-+ dpa_alloc_free(alloc, id, count);
-+ return total_invalid;
-+}
-+
-+/* BPID allocator front-end */
-+
-+int bman_alloc_bpid_range(u32 *result, u32 count, u32 align, int partial)
-+{
-+ return dpa_alloc_new(&bpalloc, result, count, align, partial);
-+}
-+EXPORT_SYMBOL(bman_alloc_bpid_range);
-+
-+static int bp_cleanup(u32 bpid)
-+{
-+ return bman_shutdown_pool(bpid) == 0;
-+}
-+void bman_release_bpid_range(u32 bpid, u32 count)
-+{
-+ u32 total_invalid = release_id_range(&bpalloc, bpid, count, bp_cleanup);
-+ if (total_invalid)
-+ pr_err("BPID range [%d..%d] (%d) had %d leaks\n",
-+ bpid, bpid + count - 1, count, total_invalid);
-+}
-+EXPORT_SYMBOL(bman_release_bpid_range);
-+
-+void bman_seed_bpid_range(u32 bpid, u32 count)
-+{
-+ dpa_alloc_seed(&bpalloc, bpid, count);
-+}
-+EXPORT_SYMBOL(bman_seed_bpid_range);
-+
-+int bman_reserve_bpid_range(u32 bpid, u32 count)
-+{
-+ return dpa_alloc_reserve(&bpalloc, bpid, count);
-+}
-+EXPORT_SYMBOL(bman_reserve_bpid_range);
-+
-+
-+/* FQID allocator front-end */
-+
-+int qman_alloc_fqid_range(u32 *result, u32 count, u32 align, int partial)
-+{
-+ return dpa_alloc_new(&fqalloc, result, count, align, partial);
-+}
-+EXPORT_SYMBOL(qman_alloc_fqid_range);
-+
-+static int fq_cleanup(u32 fqid)
-+{
-+ return qman_shutdown_fq(fqid) == 0;
-+}
-+void qman_release_fqid_range(u32 fqid, u32 count)
-+{
-+ u32 total_invalid = release_id_range(&fqalloc, fqid, count, fq_cleanup);
-+ if (total_invalid)
-+ pr_err("FQID range [%d..%d] (%d) had %d leaks\n",
-+ fqid, fqid + count - 1, count, total_invalid);
-+}
-+EXPORT_SYMBOL(qman_release_fqid_range);
-+
-+int qman_reserve_fqid_range(u32 fqid, u32 count)
-+{
-+ return dpa_alloc_reserve(&fqalloc, fqid, count);
-+}
-+EXPORT_SYMBOL(qman_reserve_fqid_range);
-+
-+void qman_seed_fqid_range(u32 fqid, u32 count)
-+{
-+ dpa_alloc_seed(&fqalloc, fqid, count);
-+}
-+EXPORT_SYMBOL(qman_seed_fqid_range);
-+
-+/* Pool-channel allocator front-end */
-+
-+int qman_alloc_pool_range(u32 *result, u32 count, u32 align, int partial)
-+{
-+ return dpa_alloc_new(&qpalloc, result, count, align, partial);
-+}
-+EXPORT_SYMBOL(qman_alloc_pool_range);
-+
-+static int qpool_cleanup(u32 qp)
-+{
-+ /* We query all FQDs starting from
-+ * FQID 1 until we get an "invalid FQID" error, looking for non-OOS FQDs
-+ * whose destination channel is the pool-channel being released.
-+ * When a non-OOS FQD is found we attempt to clean it up */
-+ struct qman_fq fq = {
-+ .fqid = 1
-+ };
-+ int err;
-+ do {
-+ struct qm_mcr_queryfq_np np;
-+ err = qman_query_fq_np(&fq, &np);
-+ if (err)
-+ /* FQID range exceeded, found no problems */
-+ return 1;
-+ if ((np.state & QM_MCR_NP_STATE_MASK) != QM_MCR_NP_STATE_OOS) {
-+ struct qm_fqd fqd;
-+ err = qman_query_fq(&fq, &fqd);
-+ BUG_ON(err);
-+ if (fqd.dest.channel == qp) {
-+ /* The channel is the FQ's target, clean it */
-+ if (qman_shutdown_fq(fq.fqid) != 0)
-+ /* Couldn't shut down the FQ
-+ so the pool must be leaked */
-+ return 0;
-+ }
-+ }
-+ /* Move to the next FQID */
-+ fq.fqid++;
-+ } while (1);
-+}
-+void qman_release_pool_range(u32 qp, u32 count)
-+{
-+ u32 total_invalid = release_id_range(&qpalloc, qp,
-+ count, qpool_cleanup);
-+ if (total_invalid) {
-+ /* Pool channels are almost always used individually */
-+ if (count == 1)
-+ pr_err("Pool channel 0x%x had %d leaks\n",
-+ qp, total_invalid);
-+ else
-+ pr_err("Pool channels [%d..%d] (%d) had %d leaks\n",
-+ qp, qp + count - 1, count, total_invalid);
-+ }
-+}
-+EXPORT_SYMBOL(qman_release_pool_range);
-+
-+
-+void qman_seed_pool_range(u32 poolid, u32 count)
-+{
-+ dpa_alloc_seed(&qpalloc, poolid, count);
-+
-+}
-+EXPORT_SYMBOL(qman_seed_pool_range);
-+
-+int qman_reserve_pool_range(u32 poolid, u32 count)
-+{
-+ return dpa_alloc_reserve(&qpalloc, poolid, count);
-+}
-+EXPORT_SYMBOL(qman_reserve_pool_range);
-+
-+
-+/* CGR ID allocator front-end */
-+
-+int qman_alloc_cgrid_range(u32 *result, u32 count, u32 align, int partial)
-+{
-+ return dpa_alloc_new(&cgralloc, result, count, align, partial);
-+}
-+EXPORT_SYMBOL(qman_alloc_cgrid_range);
-+
-+static int cqr_cleanup(u32 cgrid)
-+{
-+ /* We query all FQDs starting from
-+ * FQID 1 until we get an "invalid FQID" error, looking for non-OOS FQDs
-+ * whose CGR is the CGR being released.
-+ */
-+ struct qman_fq fq = {
-+ .fqid = 1
-+ };
-+ int err;
-+ do {
-+ struct qm_mcr_queryfq_np np;
-+ err = qman_query_fq_np(&fq, &np);
-+ if (err)
-+ /* FQID range exceeded, found no problems */
-+ return 1;
-+ if ((np.state & QM_MCR_NP_STATE_MASK) != QM_MCR_NP_STATE_OOS) {
-+ struct qm_fqd fqd;
-+ err = qman_query_fq(&fq, &fqd);
-+ BUG_ON(err);
-+ if ((fqd.fq_ctrl & QM_FQCTRL_CGE) &&
-+ (fqd.cgid == cgrid)) {
-+ pr_err("CRGID 0x%x is being used by FQID 0x%x,"
-+ " CGR will be leaked\n",
-+ cgrid, fq.fqid);
-+ return 1;
-+ }
-+ }
-+ /* Move to the next FQID */
-+ fq.fqid++;
-+ } while (1);
-+}
-+
-+void qman_release_cgrid_range(u32 cgrid, u32 count)
-+{
-+ u32 total_invalid = release_id_range(&cgralloc, cgrid,
-+ count, cqr_cleanup);
-+ if (total_invalid)
-+ pr_err("CGRID range [%d..%d] (%d) had %d leaks\n",
-+ cgrid, cgrid + count - 1, count, total_invalid);
-+}
-+EXPORT_SYMBOL(qman_release_cgrid_range);
-+
-+void qman_seed_cgrid_range(u32 cgrid, u32 count)
-+{
-+ dpa_alloc_seed(&cgralloc, cgrid, count);
-+
-+}
-+EXPORT_SYMBOL(qman_seed_cgrid_range);
-+
-+/* CEETM CHANNEL ID allocator front-end */
-+int qman_alloc_ceetm0_channel_range(u32 *result, u32 count, u32 align,
-+ int partial)
-+{
-+ return dpa_alloc_new(&ceetm0_challoc, result, count, align, partial);
-+}
-+EXPORT_SYMBOL(qman_alloc_ceetm0_channel_range);
-+
-+int qman_alloc_ceetm1_channel_range(u32 *result, u32 count, u32 align,
-+ int partial)
-+{
-+ return dpa_alloc_new(&ceetm1_challoc, result, count, align, partial);
-+}
-+EXPORT_SYMBOL(qman_alloc_ceetm1_channel_range);
-+
-+void qman_release_ceetm0_channel_range(u32 channelid, u32 count)
-+{
-+ u32 total_invalid;
-+
-+ total_invalid = release_id_range(&ceetm0_challoc, channelid, count,
-+ NULL);
-+ if (total_invalid)
-+ pr_err("CEETM channel range [%d..%d] (%d) had %d leaks\n",
-+ channelid, channelid + count - 1, count, total_invalid);
-+}
-+EXPORT_SYMBOL(qman_release_ceetm0_channel_range);
-+
-+void qman_seed_ceetm0_channel_range(u32 channelid, u32 count)
-+{
-+ dpa_alloc_seed(&ceetm0_challoc, channelid, count);
-+
-+}
-+EXPORT_SYMBOL(qman_seed_ceetm0_channel_range);
-+
-+void qman_release_ceetm1_channel_range(u32 channelid, u32 count)
-+{
-+ u32 total_invalid;
-+ total_invalid = release_id_range(&ceetm1_challoc, channelid, count,
-+ NULL);
-+ if (total_invalid)
-+ pr_err("CEETM channel range [%d..%d] (%d) had %d leaks\n",
-+ channelid, channelid + count - 1, count, total_invalid);
-+}
-+EXPORT_SYMBOL(qman_release_ceetm1_channel_range);
-+
-+void qman_seed_ceetm1_channel_range(u32 channelid, u32 count)
-+{
-+ dpa_alloc_seed(&ceetm1_challoc, channelid, count);
-+
-+}
-+EXPORT_SYMBOL(qman_seed_ceetm1_channel_range);
-+
-+/* CEETM LFQID allocator front-end */
-+int qman_alloc_ceetm0_lfqid_range(u32 *result, u32 count, u32 align,
-+ int partial)
-+{
-+ return dpa_alloc_new(&ceetm0_lfqidalloc, result, count, align, partial);
-+}
-+EXPORT_SYMBOL(qman_alloc_ceetm0_lfqid_range);
-+
-+int qman_alloc_ceetm1_lfqid_range(u32 *result, u32 count, u32 align,
-+ int partial)
-+{
-+ return dpa_alloc_new(&ceetm1_lfqidalloc, result, count, align, partial);
-+}
-+EXPORT_SYMBOL(qman_alloc_ceetm1_lfqid_range);
-+
-+void qman_release_ceetm0_lfqid_range(u32 lfqid, u32 count)
-+{
-+ u32 total_invalid;
-+
-+ total_invalid = release_id_range(&ceetm0_lfqidalloc, lfqid, count,
-+ NULL);
-+ if (total_invalid)
-+ pr_err("CEETM LFQID range [0x%x..0x%x] (%d) had %d leaks\n",
-+ lfqid, lfqid + count - 1, count, total_invalid);
-+}
-+EXPORT_SYMBOL(qman_release_ceetm0_lfqid_range);
-+
-+void qman_seed_ceetm0_lfqid_range(u32 lfqid, u32 count)
-+{
-+ dpa_alloc_seed(&ceetm0_lfqidalloc, lfqid, count);
-+
-+}
-+EXPORT_SYMBOL(qman_seed_ceetm0_lfqid_range);
-+
-+void qman_release_ceetm1_lfqid_range(u32 lfqid, u32 count)
-+{
-+ u32 total_invalid;
-+
-+ total_invalid = release_id_range(&ceetm1_lfqidalloc, lfqid, count,
-+ NULL);
-+ if (total_invalid)
-+ pr_err("CEETM LFQID range [0x%x..0x%x] (%d) had %d leaks\n",
-+ lfqid, lfqid + count - 1, count, total_invalid);
-+}
-+EXPORT_SYMBOL(qman_release_ceetm1_lfqid_range);
-+
-+void qman_seed_ceetm1_lfqid_range(u32 lfqid, u32 count)
-+{
-+ dpa_alloc_seed(&ceetm1_lfqidalloc, lfqid, count);
-+
-+}
-+EXPORT_SYMBOL(qman_seed_ceetm1_lfqid_range);
-+
-+
-+/* Everything else is the common backend to all the allocators */
-+
-+/* The allocator is a (possibly-empty) list of these; */
-+struct alloc_node {
-+ struct list_head list;
-+ u32 base;
-+ u32 num;
-+ /* refcount and is_alloced are only set
-+ when the node is in the used list */
-+ unsigned int refcount;
-+ int is_alloced;
-+};
-+
-+/* #define DPA_ALLOC_DEBUG */
-+
-+#ifdef DPA_ALLOC_DEBUG
-+#define DPRINT pr_info
-+static void DUMP(struct dpa_alloc *alloc)
-+{
-+ int off = 0;
-+ char buf[256];
-+ struct alloc_node *p;
-+ pr_info("Free Nodes\n");
-+ list_for_each_entry(p, &alloc->free, list) {
-+ if (off < 255)
-+ off += snprintf(buf + off, 255-off, "{%d,%d}",
-+ p->base, p->base + p->num - 1);
-+ }
-+ pr_info("%s\n", buf);
-+
-+ off = 0;
-+ pr_info("Used Nodes\n");
-+ list_for_each_entry(p, &alloc->used, list) {
-+ if (off < 255)
-+ off += snprintf(buf + off, 255-off, "{%d,%d}",
-+ p->base, p->base + p->num - 1);
-+ }
-+ pr_info("%s\n", buf);
-+
-+
-+
-+}
-+#else
-+#define DPRINT(x...)
-+#define DUMP(a)
-+#endif
-+
-+int dpa_alloc_new(struct dpa_alloc *alloc, u32 *result, u32 count, u32 align,
-+ int partial)
-+{
-+ struct alloc_node *i = NULL, *next_best = NULL, *used_node = NULL;
-+ u32 base, next_best_base = 0, num = 0, next_best_num = 0;
-+ struct alloc_node *margin_left, *margin_right;
-+
-+ *result = (u32)-1;
-+ DPRINT("alloc_range(%d,%d,%d)\n", count, align, partial);
-+ DUMP(alloc);
-+ /* If 'align' is 0, it should behave as though it was 1 */
-+ if (!align)
-+ align = 1;
-+ margin_left = kmalloc(sizeof(*margin_left), GFP_KERNEL);
-+ if (!margin_left)
-+ goto err;
-+ margin_right = kmalloc(sizeof(*margin_right), GFP_KERNEL);
-+ if (!margin_right) {
-+ kfree(margin_left);
-+ goto err;
-+ }
-+ spin_lock_irq(&alloc->lock);
-+ list_for_each_entry(i, &alloc->free, list) {
-+ base = (i->base + align - 1) / align;
-+ base *= align;
-+ if ((base - i->base) >= i->num)
-+ /* alignment is impossible, regardless of count */
-+ continue;
-+ num = i->num - (base - i->base);
-+ if (num >= count) {
-+ /* this one will do nicely */
-+ num = count;
-+ goto done;
-+ }
-+ if (num > next_best_num) {
-+ next_best = i;
-+ next_best_base = base;
-+ next_best_num = num;
-+ }
-+ }
-+ if (partial && next_best) {
-+ i = next_best;
-+ base = next_best_base;
-+ num = next_best_num;
-+ } else
-+ i = NULL;
-+done:
-+ if (i) {
-+ if (base != i->base) {
-+ margin_left->base = i->base;
-+ margin_left->num = base - i->base;
-+ list_add_tail(&margin_left->list, &i->list);
-+ } else
-+ kfree(margin_left);
-+ if ((base + num) < (i->base + i->num)) {
-+ margin_right->base = base + num;
-+ margin_right->num = (i->base + i->num) -
-+ (base + num);
-+ list_add(&margin_right->list, &i->list);
-+ } else
-+ kfree(margin_right);
-+ list_del(&i->list);
-+ kfree(i);
-+ *result = base;
-+ } else {
-+ spin_unlock_irq(&alloc->lock);
-+ kfree(margin_left);
-+ kfree(margin_right);
-+ }
-+
-+err:
-+ DPRINT("returning %d\n", i ? num : -ENOMEM);
-+ DUMP(alloc);
-+ if (!i)
-+ return -ENOMEM;
-+
-+ /* Add the allocation to the used list with a refcount of 1 */
-+ used_node = kmalloc(sizeof(*used_node), GFP_KERNEL);
-+ if (!used_node) {
-+ spin_unlock_irq(&alloc->lock);
-+ return -ENOMEM;
-+ }
-+ used_node->base = *result;
-+ used_node->num = num;
-+ used_node->refcount = 1;
-+ used_node->is_alloced = 1;
-+ list_add_tail(&used_node->list, &alloc->used);
-+ spin_unlock_irq(&alloc->lock);
-+ return (int)num;
-+}
-+
-+/* Allocate the list node using GFP_ATOMIC, because we *really* want to avoid
-+ * forcing error-handling on to users in the deallocation path. */
-+static void _dpa_alloc_free(struct dpa_alloc *alloc, u32 base_id, u32 count)
-+{
-+ struct alloc_node *i, *node = kmalloc(sizeof(*node), GFP_ATOMIC);
-+ BUG_ON(!node);
-+ DPRINT("release_range(%d,%d)\n", base_id, count);
-+ DUMP(alloc);
-+ BUG_ON(!count);
-+ spin_lock_irq(&alloc->lock);
-+
-+
-+ node->base = base_id;
-+ node->num = count;
-+ list_for_each_entry(i, &alloc->free, list) {
-+ if (i->base >= node->base) {
-+ /* BUG_ON(any overlapping) */
-+ BUG_ON(i->base < (node->base + node->num));
-+ list_add_tail(&node->list, &i->list);
-+ goto done;
-+ }
-+ }
-+ list_add_tail(&node->list, &alloc->free);
-+done:
-+ /* Merge to the left */
-+ i = list_entry(node->list.prev, struct alloc_node, list);
-+ if (node->list.prev != &alloc->free) {
-+ BUG_ON((i->base + i->num) > node->base);
-+ if ((i->base + i->num) == node->base) {
-+ node->base = i->base;
-+ node->num += i->num;
-+ list_del(&i->list);
-+ kfree(i);
-+ }
-+ }
-+ /* Merge to the right */
-+ i = list_entry(node->list.next, struct alloc_node, list);
-+ if (node->list.next != &alloc->free) {
-+ BUG_ON((node->base + node->num) > i->base);
-+ if ((node->base + node->num) == i->base) {
-+ node->num += i->num;
-+ list_del(&i->list);
-+ kfree(i);
-+ }
-+ }
-+ spin_unlock_irq(&alloc->lock);
-+ DUMP(alloc);
-+}
-+
-+
-+void dpa_alloc_free(struct dpa_alloc *alloc, u32 base_id, u32 count)
-+{
-+ struct alloc_node *i = NULL;
-+ spin_lock_irq(&alloc->lock);
-+
-+ /* First find the node in the used list and decrement its ref count */
-+ list_for_each_entry(i, &alloc->used, list) {
-+ if (i->base == base_id && i->num == count) {
-+ --i->refcount;
-+ if (i->refcount == 0) {
-+ list_del(&i->list);
-+ spin_unlock_irq(&alloc->lock);
-+ if (i->is_alloced)
-+ _dpa_alloc_free(alloc, base_id, count);
-+ kfree(i);
-+ return;
-+ }
-+ spin_unlock_irq(&alloc->lock);
-+ return;
-+ }
-+ }
-+ /* Couldn't find the allocation */
-+ pr_err("Attempt to free ID 0x%x COUNT %d that wasn't alloc'd or reserved\n",
-+ base_id, count);
-+ spin_unlock_irq(&alloc->lock);
-+}
-+
-+void dpa_alloc_seed(struct dpa_alloc *alloc, u32 base_id, u32 count)
-+{
-+ /* Same as free but no previous allocation checking is needed */
-+ _dpa_alloc_free(alloc, base_id, count);
-+}
-+
-+
-+int dpa_alloc_reserve(struct dpa_alloc *alloc, u32 base, u32 num)
-+{
-+ struct alloc_node *i = NULL, *used_node;
-+
-+ DPRINT("alloc_reserve(%d,%d)\n", base, num);
-+ DUMP(alloc);
-+
-+ spin_lock_irq(&alloc->lock);
-+
-+ /* Check for the node in the used list.
-+ If found, increase it's refcount */
-+ list_for_each_entry(i, &alloc->used, list) {
-+ if ((i->base == base) && (i->num == num)) {
-+ ++i->refcount;
-+ spin_unlock_irq(&alloc->lock);
-+ return 0;
-+ }
-+ if ((base >= i->base) && (base < (i->base + i->num))) {
-+ /* This is an attempt to reserve a region that was
-+ already reserved or alloced with a different
-+ base or num */
-+ pr_err("Cannot reserve %d - %d, it overlaps with"
-+ " existing reservation from %d - %d\n",
-+ base, base + num - 1, i->base,
-+ i->base + i->num - 1);
-+ spin_unlock_irq(&alloc->lock);
-+ return -1;
-+ }
-+ }
-+ /* Check to make sure this ID isn't in the free list */
-+ list_for_each_entry(i, &alloc->free, list) {
-+ if ((base >= i->base) && (base < (i->base + i->num))) {
-+ /* yep, the reservation is within this node */
-+ pr_err("Cannot reserve %d - %d, it overlaps with"
-+ " free range %d - %d and must be alloced\n",
-+ base, base + num - 1,
-+ i->base, i->base + i->num - 1);
-+ spin_unlock_irq(&alloc->lock);
-+ return -1;
-+ }
-+ }
-+ /* Add the allocation to the used list with a refcount of 1 */
-+ used_node = kmalloc(sizeof(*used_node), GFP_KERNEL);
-+ if (!used_node) {
-+ spin_unlock_irq(&alloc->lock);
-+ return -ENOMEM;
-+
-+ }
-+ used_node->base = base;
-+ used_node->num = num;
-+ used_node->refcount = 1;
-+ used_node->is_alloced = 0;
-+ list_add_tail(&used_node->list, &alloc->used);
-+ spin_unlock_irq(&alloc->lock);
-+ return 0;
-+}
-+
-+
-+int dpa_alloc_pop(struct dpa_alloc *alloc, u32 *result, u32 *count)
-+{
-+ struct alloc_node *i = NULL;
-+ DPRINT("alloc_pop()\n");
-+ DUMP(alloc);
-+ spin_lock_irq(&alloc->lock);
-+ if (!list_empty(&alloc->free)) {
-+ i = list_entry(alloc->free.next, struct alloc_node, list);
-+ list_del(&i->list);
-+ }
-+ spin_unlock_irq(&alloc->lock);
-+ DPRINT("returning %d\n", i ? 0 : -ENOMEM);
-+ DUMP(alloc);
-+ if (!i)
-+ return -ENOMEM;
-+ *result = i->base;
-+ *count = i->num;
-+ kfree(i);
-+ return 0;
-+}
-+
-+int dpa_alloc_check(struct dpa_alloc *list_head, u32 item)
-+{
-+ struct alloc_node *i = NULL;
-+ int res = 0;
-+ DPRINT("alloc_check()\n");
-+ spin_lock_irq(&list_head->lock);
-+
-+ list_for_each_entry(i, &list_head->free, list) {
-+ if ((item >= i->base) && (item < (i->base + i->num))) {
-+ res = 1;
-+ break;
-+ }
-+ }
-+ spin_unlock_irq(&list_head->lock);
-+ return res;
-+}
---- /dev/null
-+++ b/drivers/staging/fsl_qbman/dpa_sys.h
-@@ -0,0 +1,259 @@
-+/* Copyright 2008-2012 Freescale Semiconductor, Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#ifndef DPA_SYS_H
-+#define DPA_SYS_H
-+
-+#include <linux/kernel.h>
-+#include <linux/errno.h>
-+#include <linux/io.h>
-+#include <linux/dma-mapping.h>
-+#include <linux/bootmem.h>
-+#include <linux/slab.h>
-+#include <linux/module.h>
-+#include <linux/init.h>
-+#include <linux/interrupt.h>
-+#include <linux/delay.h>
-+#include <linux/of_platform.h>
-+#include <linux/of_address.h>
-+#include <linux/of_irq.h>
-+#include <linux/kthread.h>
-+#include <linux/memblock.h>
-+#include <linux/completion.h>
-+#include <linux/log2.h>
-+#include <linux/types.h>
-+#include <linux/ioctl.h>
-+#include <linux/miscdevice.h>
-+#include <linux/uaccess.h>
-+#include <linux/debugfs.h>
-+#include <linux/seq_file.h>
-+#include <linux/device.h>
-+#include <linux/uio_driver.h>
-+#include <linux/smp.h>
-+#include <linux/fsl_hypervisor.h>
-+#include <linux/vmalloc.h>
-+#include <linux/ctype.h>
-+#include <linux/math64.h>
-+#include <linux/bitops.h>
-+
-+#include <linux/fsl_usdpaa.h>
-+
-+/* When copying aligned words or shorts, try to avoid memcpy() */
-+#define CONFIG_TRY_BETTER_MEMCPY
-+
-+/* For 2-element tables related to cache-inhibited and cache-enabled mappings */
-+#define DPA_PORTAL_CE 0
-+#define DPA_PORTAL_CI 1
-+
-+/***********************/
-+/* Misc inline assists */
-+/***********************/
-+
-+#if defined CONFIG_PPC32
-+#include "dpa_sys_ppc32.h"
-+#elif defined CONFIG_PPC64
-+#include "dpa_sys_ppc64.h"
-+#elif defined CONFIG_ARM
-+#include "dpa_sys_arm.h"
-+#elif defined CONFIG_ARM64
-+#include "dpa_sys_arm64.h"
-+#endif
-+
-+
-+#ifdef CONFIG_FSL_DPA_CHECKING
-+#define DPA_ASSERT(x) \
-+ do { \
-+ if (!(x)) { \
-+ pr_crit("ASSERT: (%s:%d) %s\n", __FILE__, __LINE__, \
-+ __stringify_1(x)); \
-+ dump_stack(); \
-+ panic("assertion failure"); \
-+ } \
-+ } while (0)
-+#else
-+#define DPA_ASSERT(x)
-+#endif
-+
-+/* memcpy() stuff - when you know alignments in advance */
-+#ifdef CONFIG_TRY_BETTER_MEMCPY
-+static inline void copy_words(void *dest, const void *src, size_t sz)
-+{
-+ u32 *__dest = dest;
-+ const u32 *__src = src;
-+ size_t __sz = sz >> 2;
-+ BUG_ON((unsigned long)dest & 0x3);
-+ BUG_ON((unsigned long)src & 0x3);
-+ BUG_ON(sz & 0x3);
-+ while (__sz--)
-+ *(__dest++) = *(__src++);
-+}
-+static inline void copy_shorts(void *dest, const void *src, size_t sz)
-+{
-+ u16 *__dest = dest;
-+ const u16 *__src = src;
-+ size_t __sz = sz >> 1;
-+ BUG_ON((unsigned long)dest & 0x1);
-+ BUG_ON((unsigned long)src & 0x1);
-+ BUG_ON(sz & 0x1);
-+ while (__sz--)
-+ *(__dest++) = *(__src++);
-+}
-+static inline void copy_bytes(void *dest, const void *src, size_t sz)
-+{
-+ u8 *__dest = dest;
-+ const u8 *__src = src;
-+ while (sz--)
-+ *(__dest++) = *(__src++);
-+}
-+#else
-+#define copy_words memcpy
-+#define copy_shorts memcpy
-+#define copy_bytes memcpy
-+#endif
-+
-+/************/
-+/* RB-trees */
-+/************/
-+
-+/* We encapsulate RB-trees so that its easier to use non-linux forms in
-+ * non-linux systems. This also encapsulates the extra plumbing that linux code
-+ * usually provides when using RB-trees. This encapsulation assumes that the
-+ * data type held by the tree is u32. */
-+
-+struct dpa_rbtree {
-+ struct rb_root root;
-+};
-+#define DPA_RBTREE { .root = RB_ROOT }
-+
-+static inline void dpa_rbtree_init(struct dpa_rbtree *tree)
-+{
-+ tree->root = RB_ROOT;
-+}
-+
-+#define IMPLEMENT_DPA_RBTREE(name, type, node_field, val_field) \
-+static inline int name##_push(struct dpa_rbtree *tree, type *obj) \
-+{ \
-+ struct rb_node *parent = NULL, **p = &tree->root.rb_node; \
-+ while (*p) { \
-+ u32 item; \
-+ parent = *p; \
-+ item = rb_entry(parent, type, node_field)->val_field; \
-+ if (obj->val_field < item) \
-+ p = &parent->rb_left; \
-+ else if (obj->val_field > item) \
-+ p = &parent->rb_right; \
-+ else \
-+ return -EBUSY; \
-+ } \
-+ rb_link_node(&obj->node_field, parent, p); \
-+ rb_insert_color(&obj->node_field, &tree->root); \
-+ return 0; \
-+} \
-+static inline void name##_del(struct dpa_rbtree *tree, type *obj) \
-+{ \
-+ rb_erase(&obj->node_field, &tree->root); \
-+} \
-+static inline type *name##_find(struct dpa_rbtree *tree, u32 val) \
-+{ \
-+ type *ret; \
-+ struct rb_node *p = tree->root.rb_node; \
-+ while (p) { \
-+ ret = rb_entry(p, type, node_field); \
-+ if (val < ret->val_field) \
-+ p = p->rb_left; \
-+ else if (val > ret->val_field) \
-+ p = p->rb_right; \
-+ else \
-+ return ret; \
-+ } \
-+ return NULL; \
-+}
-+
-+/************/
-+/* Bootargs */
-+/************/
-+
-+/* Qman has "qportals=" and Bman has "bportals=", they use the same syntax
-+ * though; a comma-separated list of items, each item being a cpu index and/or a
-+ * range of cpu indices, and each item optionally be prefixed by "s" to indicate
-+ * that the portal associated with that cpu should be shared. See bman_driver.c
-+ * for more specifics. */
-+static int __parse_portals_cpu(const char **s, unsigned int *cpu)
-+{
-+ *cpu = 0;
-+ if (!isdigit(**s))
-+ return -EINVAL;
-+ while (isdigit(**s))
-+ *cpu = *cpu * 10 + (*((*s)++) - '0');
-+ return 0;
-+}
-+static inline int parse_portals_bootarg(char *str, struct cpumask *want_shared,
-+ struct cpumask *want_unshared,
-+ const char *argname)
-+{
-+ const char *s = str;
-+ unsigned int shared, cpu1, cpu2, loop;
-+
-+keep_going:
-+ if (*s == 's') {
-+ shared = 1;
-+ s++;
-+ } else
-+ shared = 0;
-+ if (__parse_portals_cpu(&s, &cpu1))
-+ goto err;
-+ if (*s == '-') {
-+ s++;
-+ if (__parse_portals_cpu(&s, &cpu2))
-+ goto err;
-+ if (cpu2 < cpu1)
-+ goto err;
-+ } else
-+ cpu2 = cpu1;
-+ for (loop = cpu1; loop <= cpu2; loop++)
-+ cpumask_set_cpu(loop, shared ? want_shared : want_unshared);
-+ if (*s == ',') {
-+ s++;
-+ goto keep_going;
-+ } else if ((*s == '\0') || isspace(*s))
-+ return 0;
-+err:
-+ pr_crit("Malformed %s argument: %s, offset: %lu\n", argname, str,
-+ (unsigned long)s - (unsigned long)str);
-+ return -EINVAL;
-+}
-+#ifdef CONFIG_FSL_USDPAA
-+/* Hooks from fsl_usdpaa_irq.c to fsl_usdpaa.c */
-+int usdpaa_get_portal_config(struct file *filp, void *cinh,
-+ enum usdpaa_portal_type ptype, unsigned int *irq,
-+ void **iir_reg);
-+#endif
-+#endif /* DPA_SYS_H */
---- /dev/null
-+++ b/drivers/staging/fsl_qbman/dpa_sys_arm.h
-@@ -0,0 +1,95 @@
-+/* Copyright 2016 Freescale Semiconductor, Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#ifndef DPA_SYS_ARM_H
-+#define DPA_SYS_ARM_H
-+
-+#include <asm/cacheflush.h>
-+#include <asm/barrier.h>
-+
-+/* Implementation of ARM specific routines */
-+
-+/* TODO: NB, we currently assume that hwsync() and lwsync() imply compiler
-+ * barriers and that dcb*() won't fall victim to compiler or execution
-+ * reordering with respect to other code/instructions that manipulate the same
-+ * cacheline. */
-+#define hwsync() { asm volatile("dmb st" : : : "memory"); }
-+#define lwsync() { asm volatile("dmb st" : : : "memory"); }
-+#define dcbf(p) { asm volatile("mcr p15, 0, %0, c7, c10, 1" : : "r" (p) : "memory"); }
-+#define dcbt_ro(p) { asm volatile("pld [%0, #64];": : "r" (p)); }
-+#define dcbt_rw(p) { asm volatile("pldw [%0, #64];": : "r" (p)); }
-+#define dcbi(p) { asm volatile("mcr p15, 0, %0, c7, c6, 1" : : "r" (p) : "memory"); }
-+
-+#define dcbz_64(p) { memset(p, 0, sizeof(*p)); }
-+
-+#define dcbf_64(p) \
-+ do { \
-+ dcbf((u32)p); \
-+ } while (0)
-+/* Commonly used combo */
-+#define dcbit_ro(p) \
-+ do { \
-+ dcbi((u32)p); \
-+ dcbt_ro((u32)p); \
-+ } while (0)
-+
-+static inline u64 mfatb(void)
-+{
-+ return get_cycles();
-+}
-+
-+static inline u32 in_be32(volatile void *addr)
-+{
-+ return be32_to_cpu(*((volatile u32 *) addr));
-+}
-+
-+static inline void out_be32(void *addr, u32 val)
-+{
-+ *((u32 *) addr) = cpu_to_be32(val);
-+}
-+
-+
-+static inline void set_bits(unsigned long mask, volatile unsigned long *p)
-+{
-+ *p |= mask;
-+}
-+static inline void clear_bits(unsigned long mask, volatile unsigned long *p)
-+{
-+ *p &= ~mask;
-+}
-+
-+static inline void flush_dcache_range(unsigned long start, unsigned long stop)
-+{
-+ __cpuc_flush_dcache_area((void *) start, stop - start);
-+}
-+
-+#define hard_smp_processor_id() raw_smp_processor_id()
-+#endif
---- /dev/null
-+++ b/drivers/staging/fsl_qbman/dpa_sys_arm64.h
-@@ -0,0 +1,102 @@
-+/* Copyright 2014 Freescale Semiconductor, Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#ifndef DPA_SYS_ARM64_H
-+#define DPA_SYS_ARM64_H
-+
-+#include <asm/cacheflush.h>
-+#include <asm/barrier.h>
-+
-+/* Implementation of ARM 64 bit specific routines */
-+
-+/* TODO: NB, we currently assume that hwsync() and lwsync() imply compiler
-+ * barriers and that dcb*() won't fall victim to compiler or execution
-+ * reordering with respect to other code/instructions that manipulate the same
-+ * cacheline. */
-+#define hwsync() { asm volatile("dmb st" : : : "memory"); }
-+#define lwsync() { asm volatile("dmb st" : : : "memory"); }
-+#define dcbf(p) { asm volatile("dc cvac, %0;" : : "r" (p) : "memory"); }
-+#define dcbt_ro(p) { asm volatile("prfm pldl1keep, [%0, #0]" : : "r" (p)); }
-+#define dcbt_rw(p) { asm volatile("prfm pstl1keep, [%0, #0]" : : "r" (p)); }
-+#define dcbi(p) { asm volatile("dc ivac, %0" : : "r"(p) : "memory"); }
-+#define dcbz(p) { asm volatile("dc zva, %0" : : "r" (p) : "memory"); }
-+
-+#define dcbz_64(p) \
-+ do { \
-+ dcbz(p); \
-+ } while (0)
-+
-+#define dcbf_64(p) \
-+ do { \
-+ dcbf(p); \
-+ } while (0)
-+/* Commonly used combo */
-+#define dcbit_ro(p) \
-+ do { \
-+ dcbi(p); \
-+ dcbt_ro(p); \
-+ } while (0)
-+
-+static inline u64 mfatb(void)
-+{
-+ return get_cycles();
-+}
-+
-+static inline u32 in_be32(volatile void *addr)
-+{
-+ return be32_to_cpu(*((volatile u32 *) addr));
-+}
-+
-+static inline void out_be32(void *addr, u32 val)
-+{
-+ *((u32 *) addr) = cpu_to_be32(val);
-+}
-+
-+
-+static inline void set_bits(unsigned long mask, volatile unsigned long *p)
-+{
-+ *p |= mask;
-+}
-+static inline void clear_bits(unsigned long mask, volatile unsigned long *p)
-+{
-+ *p &= ~mask;
-+}
-+
-+static inline void flush_dcache_range(unsigned long start, unsigned long stop)
-+{
-+ __flush_dcache_area((void *) start, stop - start);
-+}
-+
-+#define hard_smp_processor_id() raw_smp_processor_id()
-+
-+
-+
-+#endif
---- /dev/null
-+++ b/drivers/staging/fsl_qbman/dpa_sys_ppc32.h
-@@ -0,0 +1,70 @@
-+/* Copyright 2014 Freescale Semiconductor, Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#ifndef DPA_SYS_PPC32_H
-+#define DPA_SYS_PPC32_H
-+
-+/* Implementation of PowerPC 32 bit specific routines */
-+
-+/* TODO: NB, we currently assume that hwsync() and lwsync() imply compiler
-+ * barriers and that dcb*() won't fall victim to compiler or execution
-+ * reordering with respect to other code/instructions that manipulate the same
-+ * cacheline. */
-+#define hwsync() __asm__ __volatile__ ("sync" : : : "memory")
-+#define lwsync() __asm__ __volatile__ (stringify_in_c(LWSYNC) : : : "memory")
-+#define dcbf(p) __asm__ __volatile__ ("dcbf 0,%0" : : "r" (p) : "memory")
-+#define dcbt_ro(p) __asm__ __volatile__ ("dcbt 0,%0" : : "r" (p))
-+#define dcbt_rw(p) __asm__ __volatile__ ("dcbtst 0,%0" : : "r" (p))
-+#define dcbi(p) dcbf(p)
-+
-+#define dcbzl(p) __asm__ __volatile__ ("dcbzl 0,%0" : : "r" (p))
-+#define dcbz_64(p) dcbzl(p)
-+#define dcbf_64(p) dcbf(p)
-+
-+/* Commonly used combo */
-+#define dcbit_ro(p) \
-+ do { \
-+ dcbi(p); \
-+ dcbt_ro(p); \
-+ } while (0)
-+
-+static inline u64 mfatb(void)
-+{
-+ u32 hi, lo, chk;
-+ do {
-+ hi = mfspr(SPRN_ATBU);
-+ lo = mfspr(SPRN_ATBL);
-+ chk = mfspr(SPRN_ATBU);
-+ } while (unlikely(hi != chk));
-+ return ((u64)hi << 32) | (u64)lo;
-+}
-+
-+#endif
---- /dev/null
-+++ b/drivers/staging/fsl_qbman/dpa_sys_ppc64.h
-@@ -0,0 +1,79 @@
-+/* Copyright 2014 Freescale Semiconductor, Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#ifndef DPA_SYS_PPC64_H
-+#define DPA_SYS_PPC64_H
-+
-+/* Implementation of PowerPC 64 bit specific routines */
-+
-+/* TODO: NB, we currently assume that hwsync() and lwsync() imply compiler
-+ * barriers and that dcb*() won't fall victim to compiler or execution
-+ * reordering with respect to other code/instructions that manipulate the same
-+ * cacheline. */
-+#define hwsync() __asm__ __volatile__ ("sync" : : : "memory")
-+#define lwsync() __asm__ __volatile__ (stringify_in_c(LWSYNC) : : : "memory")
-+#define dcbf(p) __asm__ __volatile__ ("dcbf 0,%0" : : "r" (p) : "memory")
-+#define dcbt_ro(p) __asm__ __volatile__ ("dcbt 0,%0" : : "r" (p))
-+#define dcbt_rw(p) __asm__ __volatile__ ("dcbtst 0,%0" : : "r" (p))
-+#define dcbi(p) dcbf(p)
-+
-+#define dcbz(p) __asm__ __volatile__ ("dcbz 0,%0" : : "r" (p))
-+#define dcbz_64(p) \
-+ do { \
-+ dcbz((void*)p + 32); \
-+ dcbz(p); \
-+ } while (0)
-+#define dcbf_64(p) \
-+ do { \
-+ dcbf((void*)p + 32); \
-+ dcbf(p); \
-+ } while (0)
-+/* Commonly used combo */
-+#define dcbit_ro(p) \
-+ do { \
-+ dcbi(p); \
-+ dcbi((void*)p + 32); \
-+ dcbt_ro(p); \
-+ dcbt_ro((void*)p + 32); \
-+ } while (0)
-+
-+static inline u64 mfatb(void)
-+{
-+ u32 hi, lo, chk;
-+ do {
-+ hi = mfspr(SPRN_ATBU);
-+ lo = mfspr(SPRN_ATBL);
-+ chk = mfspr(SPRN_ATBU);
-+ } while (unlikely(hi != chk));
-+ return ((u64)hi << 32) | (u64)lo;
-+}
-+
-+#endif
---- /dev/null
-+++ b/drivers/staging/fsl_qbman/fsl_usdpaa.c
-@@ -0,0 +1,2007 @@
-+/* Copyright (C) 2008-2012 Freescale Semiconductor, Inc.
-+ * Authors: Andy Fleming <afleming@freescale.com>
-+ * Timur Tabi <timur@freescale.com>
-+ * Geoff Thorpe <Geoff.Thorpe@freescale.com>
-+ *
-+ * This file is licensed under the terms of the GNU General Public License
-+ * version 2. This program is licensed "as is" without any warranty of any
-+ * kind, whether express or implied.
-+ */
-+
-+
-+#include <linux/miscdevice.h>
-+#include <linux/fs.h>
-+#include <linux/cdev.h>
-+#include <linux/mm.h>
-+#include <linux/of.h>
-+#include <linux/memblock.h>
-+#include <linux/slab.h>
-+#include <linux/mman.h>
-+#include <linux/of_reserved_mem.h>
-+
-+#if !(defined(CONFIG_ARM) || defined(CONFIG_ARM64))
-+#include <mm/mmu_decl.h>
-+#endif
-+
-+#include "dpa_sys.h"
-+#include <linux/fsl_usdpaa.h>
-+#include "bman_low.h"
-+#include "qman_low.h"
-+
-+/* Physical address range of the memory reservation, exported for mm/mem.c */
-+static u64 phys_start;
-+static u64 phys_size;
-+static u64 arg_phys_size;
-+
-+/* PFN versions of the above */
-+static unsigned long pfn_start;
-+static unsigned long pfn_size;
-+
-+/* Memory reservations are manipulated under this spinlock (which is why 'refs'
-+ * isn't atomic_t). */
-+static DEFINE_SPINLOCK(mem_lock);
-+
-+/* The range of TLB1 indices */
-+static unsigned int first_tlb;
-+static unsigned int num_tlb = 1;
-+static unsigned int current_tlb; /* loops around for fault handling */
-+
-+/* Memory reservation is represented as a list of 'mem_fragment's, some of which
-+ * may be mapped. Unmapped fragments are always merged where possible. */
-+static LIST_HEAD(mem_list);
-+
-+struct mem_mapping;
-+
-+/* Memory fragments are in 'mem_list'. */
-+struct mem_fragment {
-+ u64 base;
-+ u64 len;
-+ unsigned long pfn_base; /* PFN version of 'base' */
-+ unsigned long pfn_len; /* PFN version of 'len' */
-+ unsigned int refs; /* zero if unmapped */
-+ u64 root_len; /* Size of the orignal fragment */
-+ unsigned long root_pfn; /* PFN of the orignal fragment */
-+ struct list_head list;
-+ /* if mapped, flags+name captured at creation time */
-+ u32 flags;
-+ char name[USDPAA_DMA_NAME_MAX];
-+ u64 map_len;
-+ /* support multi-process locks per-memory-fragment. */
-+ int has_locking;
-+ wait_queue_head_t wq;
-+ struct mem_mapping *owner;
-+};
-+
-+/* Mappings of memory fragments in 'struct ctx'. These are created from
-+ * ioctl(USDPAA_IOCTL_DMA_MAP), though the actual mapping then happens via a
-+ * mmap(). */
-+struct mem_mapping {
-+ struct mem_fragment *root_frag;
-+ u32 frag_count;
-+ u64 total_size;
-+ struct list_head list;
-+ int refs;
-+ void *virt_addr;
-+};
-+
-+struct portal_mapping {
-+ struct usdpaa_ioctl_portal_map user;
-+ union {
-+ struct qm_portal_config *qportal;
-+ struct bm_portal_config *bportal;
-+ };
-+ /* Declare space for the portals in case the process
-+ exits unexpectedly and needs to be cleaned by the kernel */
-+ union {
-+ struct qm_portal qman_portal_low;
-+ struct bm_portal bman_portal_low;
-+ };
-+ struct list_head list;
-+ struct resource *phys;
-+ struct iommu_domain *iommu_domain;
-+};
-+
-+/* Track the DPAA resources the process is using */
-+struct active_resource {
-+ struct list_head list;
-+ u32 id;
-+ u32 num;
-+ unsigned int refcount;
-+};
-+
-+/* Per-FD state (which should also be per-process but we don't enforce that) */
-+struct ctx {
-+ /* Lock to protect the context */
-+ spinlock_t lock;
-+ /* Allocated resources get put here for accounting */
-+ struct list_head resources[usdpaa_id_max];
-+ /* list of DMA maps */
-+ struct list_head maps;
-+ /* list of portal maps */
-+ struct list_head portals;
-+};
-+
-+/* Different resource classes */
-+static const struct alloc_backend {
-+ enum usdpaa_id_type id_type;
-+ int (*alloc)(u32 *, u32, u32, int);
-+ void (*release)(u32 base, unsigned int count);
-+ int (*reserve)(u32 base, unsigned int count);
-+ const char *acronym;
-+} alloc_backends[] = {
-+ {
-+ .id_type = usdpaa_id_fqid,
-+ .alloc = qman_alloc_fqid_range,
-+ .release = qman_release_fqid_range,
-+ .reserve = qman_reserve_fqid_range,
-+ .acronym = "FQID"
-+ },
-+ {
-+ .id_type = usdpaa_id_bpid,
-+ .alloc = bman_alloc_bpid_range,
-+ .release = bman_release_bpid_range,
-+ .reserve = bman_reserve_bpid_range,
-+ .acronym = "BPID"
-+ },
-+ {
-+ .id_type = usdpaa_id_qpool,
-+ .alloc = qman_alloc_pool_range,
-+ .release = qman_release_pool_range,
-+ .reserve = qman_reserve_pool_range,
-+ .acronym = "QPOOL"
-+ },
-+ {
-+ .id_type = usdpaa_id_cgrid,
-+ .alloc = qman_alloc_cgrid_range,
-+ .release = qman_release_cgrid_range,
-+ .acronym = "CGRID"
-+ },
-+ {
-+ .id_type = usdpaa_id_ceetm0_lfqid,
-+ .alloc = qman_alloc_ceetm0_lfqid_range,
-+ .release = qman_release_ceetm0_lfqid_range,
-+ .acronym = "CEETM0_LFQID"
-+ },
-+ {
-+ .id_type = usdpaa_id_ceetm0_channelid,
-+ .alloc = qman_alloc_ceetm0_channel_range,
-+ .release = qman_release_ceetm0_channel_range,
-+ .acronym = "CEETM0_LFQID"
-+ },
-+ {
-+ .id_type = usdpaa_id_ceetm1_lfqid,
-+ .alloc = qman_alloc_ceetm1_lfqid_range,
-+ .release = qman_release_ceetm1_lfqid_range,
-+ .acronym = "CEETM1_LFQID"
-+ },
-+ {
-+ .id_type = usdpaa_id_ceetm1_channelid,
-+ .alloc = qman_alloc_ceetm1_channel_range,
-+ .release = qman_release_ceetm1_channel_range,
-+ .acronym = "CEETM1_LFQID"
-+ },
-+ {
-+ /* This terminates the array */
-+ .id_type = usdpaa_id_max
-+ }
-+};
-+
-+/* Determines the largest acceptable page size for a given size
-+ The sizes are determined by what the TLB1 acceptable page sizes are */
-+static u32 largest_page_size(u32 size)
-+{
-+ int shift = 30; /* Start at 1G size */
-+ if (size < 4096)
-+ return 0;
-+ do {
-+ if (size >= (1<<shift))
-+ return 1<<shift;
-+ shift -= 2;
-+ } while (shift >= 12); /* Up to 4k */
-+ return 0;
-+}
-+
-+/* Determine if value is power of 4 */
-+static inline bool is_power_of_4(u64 x)
-+{
-+ if (x == 0 || ((x & (x - 1)) != 0))
-+ return false;
-+ return !!(x & 0x5555555555555555ull);
-+}
-+
-+/* Helper for ioctl_dma_map() when we have a larger fragment than we need. This
-+ * splits the fragment into 4 and returns the upper-most. (The caller can loop
-+ * until it has a suitable fragment size.) */
-+static struct mem_fragment *split_frag(struct mem_fragment *frag)
-+{
-+ struct mem_fragment *x[3];
-+
-+ x[0] = kmalloc(sizeof(struct mem_fragment), GFP_ATOMIC);
-+ x[1] = kmalloc(sizeof(struct mem_fragment), GFP_ATOMIC);
-+ x[2] = kmalloc(sizeof(struct mem_fragment), GFP_ATOMIC);
-+ if (!x[0] || !x[1] || !x[2]) {
-+ kfree(x[0]);
-+ kfree(x[1]);
-+ kfree(x[2]);
-+ return NULL;
-+ }
-+ BUG_ON(frag->refs);
-+ frag->len >>= 2;
-+ frag->pfn_len >>= 2;
-+ x[0]->base = frag->base + frag->len;
-+ x[1]->base = x[0]->base + frag->len;
-+ x[2]->base = x[1]->base + frag->len;
-+ x[0]->len = x[1]->len = x[2]->len = frag->len;
-+ x[0]->pfn_base = frag->pfn_base + frag->pfn_len;
-+ x[1]->pfn_base = x[0]->pfn_base + frag->pfn_len;
-+ x[2]->pfn_base = x[1]->pfn_base + frag->pfn_len;
-+ x[0]->pfn_len = x[1]->pfn_len = x[2]->pfn_len = frag->pfn_len;
-+ x[0]->refs = x[1]->refs = x[2]->refs = 0;
-+ x[0]->root_len = x[1]->root_len = x[2]->root_len = frag->root_len;
-+ x[0]->root_pfn = x[1]->root_pfn = x[2]->root_pfn = frag->root_pfn;
-+ x[0]->name[0] = x[1]->name[0] = x[2]->name[0] = 0;
-+ list_add_tail(&x[0]->list, &frag->list);
-+ list_add_tail(&x[1]->list, &x[0]->list);
-+ list_add_tail(&x[2]->list, &x[1]->list);
-+ return x[2];
-+}
-+
-+static __maybe_unused void dump_frags(void)
-+{
-+ struct mem_fragment *frag;
-+ int i = 0;
-+ list_for_each_entry(frag, &mem_list, list) {
-+ pr_info("FRAG %d: base 0x%llx pfn_base 0x%lx len 0x%llx root_len 0x%llx root_pfn 0x%lx refs %d name %s\n",
-+ i, frag->base, frag->pfn_base,
-+ frag->len, frag->root_len, frag->root_pfn,
-+ frag->refs, frag->name);
-+ ++i;
-+ }
-+}
-+
-+/* Walk the list of fragments and adjoin neighbouring segments if possible */
-+static void compress_frags(void)
-+{
-+ /* Walk the fragment list and combine fragments */
-+ struct mem_fragment *frag, *nxtfrag;
-+ u64 len = 0;
-+
-+ int i, numfrags;
-+
-+
-+ frag = list_entry(mem_list.next, struct mem_fragment, list);
-+
-+ while (&frag->list != &mem_list) {
-+ /* Must combine consecutive fragemenst with
-+ same root_pfn such that they are power of 4 */
-+ if (frag->refs != 0) {
-+ frag = list_entry(frag->list.next,
-+ struct mem_fragment, list);
-+ continue; /* Not this window */
-+ }
-+ len = frag->len;
-+ numfrags = 0;
-+ nxtfrag = list_entry(frag->list.next,
-+ struct mem_fragment, list);
-+ while (true) {
-+ if (&nxtfrag->list == &mem_list) {
-+ numfrags = 0;
-+ break; /* End of list */
-+ }
-+ if (nxtfrag->refs) {
-+ numfrags = 0;
-+ break; /* In use still */
-+ }
-+ if (nxtfrag->root_pfn != frag->root_pfn) {
-+ numfrags = 0;
-+ break; /* Crosses root fragment boundary */
-+ }
-+ len += nxtfrag->len;
-+ numfrags++;
-+ if (is_power_of_4(len)) {
-+ /* These fragments can be combined */
-+ break;
-+ }
-+ nxtfrag = list_entry(nxtfrag->list.next,
-+ struct mem_fragment, list);
-+ }
-+ if (numfrags == 0) {
-+ frag = list_entry(frag->list.next,
-+ struct mem_fragment, list);
-+ continue; /* try the next window */
-+ }
-+ for (i = 0; i < numfrags; i++) {
-+ struct mem_fragment *todel =
-+ list_entry(nxtfrag->list.prev,
-+ struct mem_fragment, list);
-+ nxtfrag->len += todel->len;
-+ nxtfrag->pfn_len += todel->pfn_len;
-+ list_del(&todel->list);
-+ }
-+ /* Re evaluate the list, things may merge now */
-+ frag = list_entry(mem_list.next, struct mem_fragment, list);
-+ }
-+}
-+
-+/* Hook from arch/powerpc/mm/mem.c */
-+int usdpaa_test_fault(unsigned long pfn, u64 *phys_addr, u64 *size)
-+{
-+ struct mem_fragment *frag;
-+ int idx = -1;
-+ if ((pfn < pfn_start) || (pfn >= (pfn_start + pfn_size)))
-+ return -1;
-+ /* It's in-range, we need to find the fragment */
-+ spin_lock(&mem_lock);
-+ list_for_each_entry(frag, &mem_list, list) {
-+ if ((pfn >= frag->pfn_base) && (pfn < (frag->pfn_base +
-+ frag->pfn_len))) {
-+ *phys_addr = frag->base;
-+ *size = frag->len;
-+ idx = current_tlb++;
-+ if (current_tlb >= (first_tlb + num_tlb))
-+ current_tlb = first_tlb;
-+ break;
-+ }
-+ }
-+ spin_unlock(&mem_lock);
-+ return idx;
-+}
-+
-+static int usdpaa_open(struct inode *inode, struct file *filp)
-+{
-+ const struct alloc_backend *backend = &alloc_backends[0];
-+ struct ctx *ctx = kmalloc(sizeof(struct ctx), GFP_KERNEL);
-+ if (!ctx)
-+ return -ENOMEM;
-+ filp->private_data = ctx;
-+
-+ while (backend->id_type != usdpaa_id_max) {
-+ INIT_LIST_HEAD(&ctx->resources[backend->id_type]);
-+ backend++;
-+ }
-+
-+ INIT_LIST_HEAD(&ctx->maps);
-+ INIT_LIST_HEAD(&ctx->portals);
-+ spin_lock_init(&ctx->lock);
-+
-+ //filp->f_mapping->backing_dev_info = &directly_mappable_cdev_bdi;
-+
-+ return 0;
-+}
-+
-+#define DQRR_MAXFILL 15
-+
-+
-+/* Invalidate a portal */
-+void dbci_portal(void *addr)
-+{
-+ int i;
-+
-+ for (i = 0; i < 0x4000; i += 64)
-+ dcbi(addr + i);
-+}
-+
-+/* Reset a QMan portal to its default state */
-+static int init_qm_portal(struct qm_portal_config *config,
-+ struct qm_portal *portal)
-+{
-+ const struct qm_dqrr_entry *dqrr = NULL;
-+ int i;
-+
-+ portal->addr.addr_ce = config->addr_virt[DPA_PORTAL_CE];
-+ portal->addr.addr_ci = config->addr_virt[DPA_PORTAL_CI];
-+
-+ /* Make sure interrupts are inhibited */
-+ qm_out(IIR, 1);
-+
-+ /*
-+ * Invalidate the entire CE portal are to ensure no stale
-+ * cachelines are present. This should be done on all
-+ * cores as the portal is mapped as M=0 (non-coherent).
-+ */
-+ on_each_cpu(dbci_portal, portal->addr.addr_ce, 1);
-+
-+ /* Initialize the DQRR. This will stop any dequeue
-+ commands that are in progress */
-+ if (qm_dqrr_init(portal, config, qm_dqrr_dpush, qm_dqrr_pvb,
-+ qm_dqrr_cdc, DQRR_MAXFILL)) {
-+ pr_err("qm_dqrr_init() failed when trying to"
-+ " recover portal, portal will be leaked\n");
-+ return 1;
-+ }
-+
-+ /* Discard any entries on the DQRR */
-+ /* If we consume the ring twice something is wrong */
-+ for (i = 0; i < DQRR_MAXFILL * 2; i++) {
-+ qm_dqrr_pvb_update(portal);
-+ dqrr = qm_dqrr_current(portal);
-+ if (!dqrr)
-+ break;
-+ qm_dqrr_cdc_consume_1ptr(portal, dqrr, 0);
-+ qm_dqrr_pvb_update(portal);
-+ qm_dqrr_next(portal);
-+ }
-+ /* Initialize the EQCR */
-+ if (qm_eqcr_init(portal, qm_eqcr_pvb,
-+ qm_eqcr_get_ci_stashing(portal), 1)) {
-+ pr_err("Qman EQCR initialisation failed\n");
-+ return 1;
-+ }
-+ /* initialize the MR */
-+ if (qm_mr_init(portal, qm_mr_pvb, qm_mr_cci)) {
-+ pr_err("Qman MR initialisation failed\n");
-+ return 1;
-+ }
-+ qm_mr_pvb_update(portal);
-+ while (qm_mr_current(portal)) {
-+ qm_mr_next(portal);
-+ qm_mr_cci_consume_to_current(portal);
-+ qm_mr_pvb_update(portal);
-+ }
-+
-+ if (qm_mc_init(portal)) {
-+ pr_err("Qman MC initialisation failed\n");
-+ return 1;
-+ }
-+ return 0;
-+}
-+
-+static int init_bm_portal(struct bm_portal_config *config,
-+ struct bm_portal *portal)
-+{
-+ portal->addr.addr_ce = config->addr_virt[DPA_PORTAL_CE];
-+ portal->addr.addr_ci = config->addr_virt[DPA_PORTAL_CI];
-+
-+ /*
-+ * Invalidate the entire CE portal are to ensure no stale
-+ * cachelines are present. This should be done on all
-+ * cores as the portal is mapped as M=0 (non-coherent).
-+ */
-+ on_each_cpu(dbci_portal, portal->addr.addr_ce, 1);
-+
-+ if (bm_rcr_init(portal, bm_rcr_pvb, bm_rcr_cce)) {
-+ pr_err("Bman RCR initialisation failed\n");
-+ return 1;
-+ }
-+ if (bm_mc_init(portal)) {
-+ pr_err("Bman MC initialisation failed\n");
-+ return 1;
-+ }
-+ return 0;
-+}
-+
-+/* Function that will scan all FQ's in the system. For each FQ that is not
-+ OOS it will call the check_channel helper to determine if the FQ should
-+ be torn down. If the check_channel helper returns true the FQ will be
-+ transitioned to the OOS state */
-+static int qm_check_and_destroy_fqs(struct qm_portal *portal, void *ctx,
-+ bool (*check_channel)(void*, u32))
-+{
-+ u32 fq_id = 0;
-+ while (1) {
-+ struct qm_mc_command *mcc;
-+ struct qm_mc_result *mcr;
-+ u8 state;
-+ u32 channel;
-+
-+ /* Determine the channel for the FQID */
-+ mcc = qm_mc_start(portal);
-+ mcc->queryfq.fqid = fq_id;
-+ qm_mc_commit(portal, QM_MCC_VERB_QUERYFQ);
-+ while (!(mcr = qm_mc_result(portal)))
-+ cpu_relax();
-+ DPA_ASSERT((mcr->verb & QM_MCR_VERB_MASK)
-+ == QM_MCR_VERB_QUERYFQ);
-+ if (mcr->result != QM_MCR_RESULT_OK)
-+ break; /* End of valid FQIDs */
-+
-+ channel = mcr->queryfq.fqd.dest.channel;
-+ /* Determine the state of the FQID */
-+ mcc = qm_mc_start(portal);
-+ mcc->queryfq_np.fqid = fq_id;
-+ qm_mc_commit(portal, QM_MCC_VERB_QUERYFQ_NP);
-+ while (!(mcr = qm_mc_result(portal)))
-+ cpu_relax();
-+ DPA_ASSERT((mcr->verb & QM_MCR_VERB_MASK)
-+ == QM_MCR_VERB_QUERYFQ_NP);
-+ state = mcr->queryfq_np.state & QM_MCR_NP_STATE_MASK;
-+ if (state == QM_MCR_NP_STATE_OOS)
-+ /* Already OOS, no need to do anymore checks */
-+ goto next;
-+
-+ if (check_channel(ctx, channel))
-+ qm_shutdown_fq(&portal, 1, fq_id);
-+ next:
-+ ++fq_id;
-+ }
-+ return 0;
-+}
-+
-+static bool check_channel_device(void *_ctx, u32 channel)
-+{
-+ struct ctx *ctx = _ctx;
-+ struct portal_mapping *portal, *tmpportal;
-+ struct active_resource *res;
-+
-+ /* See if the FQ is destined for one of the portals we're cleaning up */
-+ list_for_each_entry_safe(portal, tmpportal, &ctx->portals, list) {
-+ if (portal->user.type == usdpaa_portal_qman) {
-+ if (portal->qportal->public_cfg.channel == channel) {
-+ /* This FQs destination is a portal
-+ we're cleaning, send a retire */
-+ return true;
-+ }
-+ }
-+ }
-+
-+ /* Check the pool channels that will be released as well */
-+ list_for_each_entry(res, &ctx->resources[usdpaa_id_qpool], list) {
-+ if ((res->id >= channel) &&
-+ ((res->id + res->num - 1) <= channel))
-+ return true;
-+ }
-+ return false;
-+}
-+
-+static bool check_portal_channel(void *ctx, u32 channel)
-+{
-+ u32 portal_channel = *(u32 *)ctx;
-+ if (portal_channel == channel) {
-+ /* This FQs destination is a portal
-+ we're cleaning, send a retire */
-+ return true;
-+ }
-+ return false;
-+}
-+
-+
-+
-+
-+static int usdpaa_release(struct inode *inode, struct file *filp)
-+{
-+ struct ctx *ctx = filp->private_data;
-+ struct mem_mapping *map, *tmpmap;
-+ struct portal_mapping *portal, *tmpportal;
-+ const struct alloc_backend *backend = &alloc_backends[0];
-+ struct active_resource *res;
-+ struct qm_portal *qm_cleanup_portal = NULL;
-+ struct bm_portal *bm_cleanup_portal = NULL;
-+ struct qm_portal_config *qm_alloced_portal = NULL;
-+ struct bm_portal_config *bm_alloced_portal = NULL;
-+
-+ struct qm_portal *portal_array[qman_portal_max];
-+ int portal_count = 0;
-+
-+ /* Ensure the release operation cannot be migrated to another
-+ CPU as CPU specific variables may be needed during cleanup */
-+#ifdef CONFIG_PREEMPT_RT_FULL
-+ migrate_disable();
-+#endif
-+ /* The following logic is used to recover resources that were not
-+ correctly released by the process that is closing the FD.
-+ Step 1: syncronize the HW with the qm_portal/bm_portal structures
-+ in the kernel
-+ */
-+
-+ list_for_each_entry_safe(portal, tmpportal, &ctx->portals, list) {
-+ /* Try to recover any portals that weren't shut down */
-+ if (portal->user.type == usdpaa_portal_qman) {
-+ portal_array[portal_count] = &portal->qman_portal_low;
-+ ++portal_count;
-+ init_qm_portal(portal->qportal,
-+ &portal->qman_portal_low);
-+ if (!qm_cleanup_portal) {
-+ qm_cleanup_portal = &portal->qman_portal_low;
-+ } else {
-+ /* Clean FQs on the dedicated channel */
-+ u32 chan = portal->qportal->public_cfg.channel;
-+ qm_check_and_destroy_fqs(
-+ &portal->qman_portal_low, &chan,
-+ check_portal_channel);
-+ }
-+ } else {
-+ /* BMAN */
-+ init_bm_portal(portal->bportal,
-+ &portal->bman_portal_low);
-+ if (!bm_cleanup_portal)
-+ bm_cleanup_portal = &portal->bman_portal_low;
-+ }
-+ }
-+ /* If no portal was found, allocate one for cleanup */
-+ if (!qm_cleanup_portal) {
-+ qm_alloced_portal = qm_get_unused_portal();
-+ if (!qm_alloced_portal) {
-+ pr_crit("No QMan portal avalaible for cleanup\n");
-+#ifdef CONFIG_PREEMPT_RT_FULL
-+ migrate_enable();
-+#endif
-+ return -1;
-+ }
-+ qm_cleanup_portal = kmalloc(sizeof(struct qm_portal),
-+ GFP_KERNEL);
-+ if (!qm_cleanup_portal) {
-+#ifdef CONFIG_PREEMPT_RT_FULL
-+ migrate_enable();
-+#endif
-+ return -ENOMEM;
-+ }
-+ init_qm_portal(qm_alloced_portal, qm_cleanup_portal);
-+ portal_array[portal_count] = qm_cleanup_portal;
-+ ++portal_count;
-+ }
-+ if (!bm_cleanup_portal) {
-+ bm_alloced_portal = bm_get_unused_portal();
-+ if (!bm_alloced_portal) {
-+ pr_crit("No BMan portal avalaible for cleanup\n");
-+#ifdef CONFIG_PREEMPT_RT_FULL
-+ migrate_enable();
-+#endif
-+ return -1;
-+ }
-+ bm_cleanup_portal = kmalloc(sizeof(struct bm_portal),
-+ GFP_KERNEL);
-+ if (!bm_cleanup_portal) {
-+#ifdef CONFIG_PREEMPT_RT_FULL
-+ migrate_enable();
-+#endif
-+ return -ENOMEM;
-+ }
-+ init_bm_portal(bm_alloced_portal, bm_cleanup_portal);
-+ }
-+
-+ /* OOS the FQs associated with this process */
-+ qm_check_and_destroy_fqs(qm_cleanup_portal, ctx, check_channel_device);
-+
-+ while (backend->id_type != usdpaa_id_max) {
-+ int leaks = 0;
-+ list_for_each_entry(res, &ctx->resources[backend->id_type],
-+ list) {
-+ if (backend->id_type == usdpaa_id_fqid) {
-+ int i = 0;
-+ for (; i < res->num; i++) {
-+ /* Clean FQs with the cleanup portal */
-+ qm_shutdown_fq(portal_array,
-+ portal_count,
-+ res->id + i);
-+ }
-+ }
-+ leaks += res->num;
-+ backend->release(res->id, res->num);
-+ }
-+ if (leaks)
-+ pr_crit("USDPAA process leaking %d %s%s\n", leaks,
-+ backend->acronym, (leaks > 1) ? "s" : "");
-+ backend++;
-+ }
-+ /* Release any DMA regions */
-+ spin_lock(&mem_lock);
-+ list_for_each_entry_safe(map, tmpmap, &ctx->maps, list) {
-+ struct mem_fragment *current_frag = map->root_frag;
-+ int i;
-+ if (map->root_frag->has_locking &&
-+ (map->root_frag->owner == map)) {
-+ map->root_frag->owner = NULL;
-+ wake_up(&map->root_frag->wq);
-+ }
-+ /* Check each fragment and merge if the ref count is 0 */
-+ for (i = 0; i < map->frag_count; i++) {
-+ --current_frag->refs;
-+ current_frag = list_entry(current_frag->list.prev,
-+ struct mem_fragment, list);
-+ }
-+
-+ compress_frags();
-+ list_del(&map->list);
-+ kfree(map);
-+ }
-+ spin_unlock(&mem_lock);
-+
-+ /* Return portals */
-+ list_for_each_entry_safe(portal, tmpportal, &ctx->portals, list) {
-+ if (portal->user.type == usdpaa_portal_qman) {
-+ /* Give the portal back to the allocator */
-+ init_qm_portal(portal->qportal,
-+ &portal->qman_portal_low);
-+ qm_put_unused_portal(portal->qportal);
-+ } else {
-+ init_bm_portal(portal->bportal,
-+ &portal->bman_portal_low);
-+ bm_put_unused_portal(portal->bportal);
-+ }
-+ list_del(&portal->list);
-+ kfree(portal);
-+ }
-+ if (qm_alloced_portal) {
-+ qm_put_unused_portal(qm_alloced_portal);
-+ kfree(qm_cleanup_portal);
-+ }
-+ if (bm_alloced_portal) {
-+ bm_put_unused_portal(bm_alloced_portal);
-+ kfree(bm_cleanup_portal);
-+ }
-+
-+ kfree(ctx);
-+#ifdef CONFIG_PREEMPT_RT_FULL
-+ migrate_enable();
-+#endif
-+ return 0;
-+}
-+
-+static int check_mmap_dma(struct ctx *ctx, struct vm_area_struct *vma,
-+ int *match, unsigned long *pfn)
-+{
-+ struct mem_mapping *map;
-+
-+ list_for_each_entry(map, &ctx->maps, list) {
-+ int i;
-+ struct mem_fragment *frag = map->root_frag;
-+
-+ for (i = 0; i < map->frag_count; i++) {
-+ if (frag->pfn_base == vma->vm_pgoff) {
-+ *match = 1;
-+ *pfn = frag->pfn_base;
-+ return 0;
-+ }
-+ frag = list_entry(frag->list.next, struct mem_fragment,
-+ list);
-+ }
-+ }
-+ *match = 0;
-+ return 0;
-+}
-+
-+static int check_mmap_resource(struct resource *res, struct vm_area_struct *vma,
-+ int *match, unsigned long *pfn)
-+{
-+ *pfn = res->start >> PAGE_SHIFT;
-+ if (*pfn == vma->vm_pgoff) {
-+ *match = 1;
-+ if ((vma->vm_end - vma->vm_start) != resource_size(res))
-+ return -EINVAL;
-+ } else
-+ *match = 0;
-+ return 0;
-+}
-+
-+static int check_mmap_portal(struct ctx *ctx, struct vm_area_struct *vma,
-+ int *match, unsigned long *pfn)
-+{
-+ struct portal_mapping *portal;
-+ int ret;
-+
-+ list_for_each_entry(portal, &ctx->portals, list) {
-+ ret = check_mmap_resource(&portal->phys[DPA_PORTAL_CE], vma,
-+ match, pfn);
-+ if (*match) {
-+ vma->vm_page_prot =
-+#if defined(CONFIG_ARM) || defined(CONFIG_ARM64)
-+ pgprot_cached_ns(vma->vm_page_prot);
-+#else
-+ pgprot_cached_noncoherent(vma->vm_page_prot);
-+#endif
-+ return ret;
-+ }
-+ ret = check_mmap_resource(&portal->phys[DPA_PORTAL_CI], vma,
-+ match, pfn);
-+ if (*match) {
-+ vma->vm_page_prot = pgprot_noncached(vma->vm_page_prot);
-+ return ret;
-+ }
-+ }
-+ *match = 0;
-+ return 0;
-+}
-+
-+static int usdpaa_mmap(struct file *filp, struct vm_area_struct *vma)
-+{
-+ struct ctx *ctx = filp->private_data;
-+ unsigned long pfn = 0;
-+ int match, ret;
-+
-+ spin_lock(&mem_lock);
-+ ret = check_mmap_dma(ctx, vma, &match, &pfn);
-+ if (!match)
-+ ret = check_mmap_portal(ctx, vma, &match, &pfn);
-+ spin_unlock(&mem_lock);
-+ if (!match)
-+ return -EINVAL;
-+ if (!ret)
-+ ret = remap_pfn_range(vma, vma->vm_start, pfn,
-+ vma->vm_end - vma->vm_start,
-+ vma->vm_page_prot);
-+ return ret;
-+}
-+
-+/* Return the nearest rounded-up address >= 'addr' that is 'sz'-aligned. 'sz'
-+ * must be a power of 2, but both 'addr' and 'sz' can be expressions. */
-+#define USDPAA_MEM_ROUNDUP(addr, sz) \
-+ ({ \
-+ unsigned long foo_align = (sz) - 1; \
-+ ((addr) + foo_align) & ~foo_align; \
-+ })
-+/* Searching for a size-aligned virtual address range starting from 'addr' */
-+static unsigned long usdpaa_get_unmapped_area(struct file *file,
-+ unsigned long addr,
-+ unsigned long len,
-+ unsigned long pgoff,
-+ unsigned long flags)
-+{
-+ struct vm_area_struct *vma;
-+
-+ if (len % PAGE_SIZE)
-+ return -EINVAL;
-+ if (!len)
-+ return -EINVAL;
-+
-+ /* Need to align the address to the largest pagesize of the mapping
-+ * because the MMU requires the virtual address to have the same
-+ * alignment as the physical address */
-+ addr = USDPAA_MEM_ROUNDUP(addr, largest_page_size(len));
-+ vma = find_vma(current->mm, addr);
-+ /* Keep searching until we reach the end of currently-used virtual
-+ * address-space or we find a big enough gap. */
-+ while (vma) {
-+ if ((addr + len) < vma->vm_start)
-+ return addr;
-+
-+ addr = USDPAA_MEM_ROUNDUP(vma->vm_end, largest_page_size(len));
-+ vma = vma->vm_next;
-+ }
-+ if ((TASK_SIZE - len) < addr)
-+ return -ENOMEM;
-+ return addr;
-+}
-+
-+static long ioctl_id_alloc(struct ctx *ctx, void __user *arg)
-+{
-+ struct usdpaa_ioctl_id_alloc i;
-+ const struct alloc_backend *backend;
-+ struct active_resource *res;
-+ int ret = copy_from_user(&i, arg, sizeof(i));
-+ if (ret)
-+ return ret;
-+ if ((i.id_type >= usdpaa_id_max) || !i.num)
-+ return -EINVAL;
-+ backend = &alloc_backends[i.id_type];
-+ /* Allocate the required resource type */
-+ ret = backend->alloc(&i.base, i.num, i.align, i.partial);
-+ if (ret < 0)
-+ return ret;
-+ i.num = ret;
-+ /* Copy the result to user-space */
-+ ret = copy_to_user(arg, &i, sizeof(i));
-+ if (ret) {
-+ backend->release(i.base, i.num);
-+ return ret;
-+ }
-+ /* Assign the allocated range to the FD accounting */
-+ res = kmalloc(sizeof(*res), GFP_KERNEL);
-+ if (!res) {
-+ backend->release(i.base, i.num);
-+ return -ENOMEM;
-+ }
-+ spin_lock(&ctx->lock);
-+ res->id = i.base;
-+ res->num = i.num;
-+ res->refcount = 1;
-+ list_add(&res->list, &ctx->resources[i.id_type]);
-+ spin_unlock(&ctx->lock);
-+ return 0;
-+}
-+
-+static long ioctl_id_release(struct ctx *ctx, void __user *arg)
-+{
-+ struct usdpaa_ioctl_id_release i;
-+ const struct alloc_backend *backend;
-+ struct active_resource *tmp, *pos;
-+
-+ int ret = copy_from_user(&i, arg, sizeof(i));
-+ if (ret)
-+ return ret;
-+ if ((i.id_type >= usdpaa_id_max) || !i.num)
-+ return -EINVAL;
-+ backend = &alloc_backends[i.id_type];
-+ /* Pull the range out of the FD accounting - the range is valid iff this
-+ * succeeds. */
-+ spin_lock(&ctx->lock);
-+ list_for_each_entry_safe(pos, tmp, &ctx->resources[i.id_type], list) {
-+ if (pos->id == i.base && pos->num == i.num) {
-+ pos->refcount--;
-+ if (pos->refcount) {
-+ spin_unlock(&ctx->lock);
-+ return 0; /* Still being used */
-+ }
-+ list_del(&pos->list);
-+ kfree(pos);
-+ spin_unlock(&ctx->lock);
-+ goto found;
-+ }
-+ }
-+ /* Failed to find the resource */
-+ spin_unlock(&ctx->lock);
-+ pr_err("Couldn't find resource type %d base 0x%x num %d\n",
-+ i.id_type, i.base, i.num);
-+ return -EINVAL;
-+found:
-+ /* Release the resource to the backend */
-+ backend->release(i.base, i.num);
-+ return 0;
-+}
-+
-+static long ioctl_id_reserve(struct ctx *ctx, void __user *arg)
-+{
-+ struct usdpaa_ioctl_id_reserve i;
-+ const struct alloc_backend *backend;
-+ struct active_resource *tmp, *pos;
-+
-+ int ret = copy_from_user(&i, arg, sizeof(i));
-+ if (ret)
-+ return ret;
-+ if ((i.id_type >= usdpaa_id_max) || !i.num)
-+ return -EINVAL;
-+ backend = &alloc_backends[i.id_type];
-+ if (!backend->reserve)
-+ return -EINVAL;
-+ /* Pull the range out of the FD accounting - the range is valid iff this
-+ * succeeds. */
-+ spin_lock(&ctx->lock);
-+ list_for_each_entry_safe(pos, tmp, &ctx->resources[i.id_type], list) {
-+ if (pos->id == i.base && pos->num == i.num) {
-+ pos->refcount++;
-+ spin_unlock(&ctx->lock);
-+ return 0;
-+ }
-+ }
-+
-+ /* Failed to find the resource */
-+ spin_unlock(&ctx->lock);
-+
-+ /* Reserve the resource in the backend */
-+ ret = backend->reserve(i.base, i.num);
-+ if (ret)
-+ return ret;
-+ /* Assign the reserved range to the FD accounting */
-+ pos = kmalloc(sizeof(*pos), GFP_KERNEL);
-+ if (!pos) {
-+ backend->release(i.base, i.num);
-+ return -ENOMEM;
-+ }
-+ spin_lock(&ctx->lock);
-+ pos->id = i.base;
-+ pos->num = i.num;
-+ pos->refcount = 1;
-+ list_add(&pos->list, &ctx->resources[i.id_type]);
-+ spin_unlock(&ctx->lock);
-+ return 0;
-+}
-+
-+static long ioctl_dma_map(struct file *fp, struct ctx *ctx,
-+ struct usdpaa_ioctl_dma_map *i)
-+{
-+ struct mem_fragment *frag, *start_frag, *next_frag;
-+ struct mem_mapping *map, *tmp;
-+ int ret = 0;
-+ u32 largest_page, so_far = 0;
-+ int frag_count = 0;
-+ unsigned long next_addr = PAGE_SIZE, populate;
-+
-+ /* error checking to ensure values copied from user space are valid */
-+ if (i->len % PAGE_SIZE)
-+ return -EINVAL;
-+
-+ map = kmalloc(sizeof(*map), GFP_KERNEL);
-+ if (!map)
-+ return -ENOMEM;
-+
-+ spin_lock(&mem_lock);
-+ if (i->flags & USDPAA_DMA_FLAG_SHARE) {
-+ list_for_each_entry(frag, &mem_list, list) {
-+ if (frag->refs && (frag->flags &
-+ USDPAA_DMA_FLAG_SHARE) &&
-+ !strncmp(i->name, frag->name,
-+ USDPAA_DMA_NAME_MAX)) {
-+ /* Matching entry */
-+ if ((i->flags & USDPAA_DMA_FLAG_CREATE) &&
-+ !(i->flags & USDPAA_DMA_FLAG_LAZY)) {
-+ ret = -EBUSY;
-+ goto out;
-+ }
-+
-+ /* Check to ensure size matches record */
-+ if (i->len != frag->map_len && i->len) {
-+ pr_err("ioctl_dma_map() Size requested does not match %s and is none zero\n",
-+ frag->name);
-+ return -EINVAL;
-+ }
-+
-+ /* Check if this has already been mapped
-+ to this process */
-+ list_for_each_entry(tmp, &ctx->maps, list)
-+ if (tmp->root_frag == frag) {
-+ /* Already mapped, just need to
-+ inc ref count */
-+ tmp->refs++;
-+ kfree(map);
-+ i->did_create = 0;
-+ i->len = tmp->total_size;
-+ i->phys_addr = frag->base;
-+ i->ptr = tmp->virt_addr;
-+ spin_unlock(&mem_lock);
-+ return 0;
-+ }
-+ /* Matching entry - just need to map */
-+ i->has_locking = frag->has_locking;
-+ i->did_create = 0;
-+ i->len = frag->map_len;
-+ start_frag = frag;
-+ goto do_map;
-+ }
-+ }
-+ /* No matching entry */
-+ if (!(i->flags & USDPAA_DMA_FLAG_CREATE)) {
-+ pr_err("ioctl_dma_map() No matching entry\n");
-+ ret = -ENOMEM;
-+ goto out;
-+ }
-+ }
-+ /* New fragment required, size must be provided. */
-+ if (!i->len) {
-+ ret = -EINVAL;
-+ goto out;
-+ }
-+
-+ /* Find one of more contiguous fragments that satisfy the total length
-+ trying to minimize the number of fragments
-+ compute the largest page size that the allocation could use */
-+ largest_page = largest_page_size(i->len);
-+ start_frag = NULL;
-+ while (largest_page &&
-+ largest_page <= largest_page_size(phys_size) &&
-+ start_frag == NULL) {
-+ /* Search the list for a frag of that size */
-+ list_for_each_entry(frag, &mem_list, list) {
-+ if (!frag->refs && (frag->len == largest_page)) {
-+ /* See if the next x fragments are free
-+ and can accomidate the size */
-+ u32 found_size = largest_page;
-+ next_frag = list_entry(frag->list.prev,
-+ struct mem_fragment,
-+ list);
-+ /* If the fragement is too small check
-+ if the neighbours cab support it */
-+ while (found_size < i->len) {
-+ if (&mem_list == &next_frag->list)
-+ break; /* End of list */
-+ if (next_frag->refs != 0 ||
-+ next_frag->len == 0)
-+ break; /* not enough space */
-+ found_size += next_frag->len;
-+ next_frag = list_entry(
-+ next_frag->list.prev,
-+ struct mem_fragment,
-+ list);
-+ }
-+ if (found_size >= i->len) {
-+ /* Success! there is enough contigous
-+ free space */
-+ start_frag = frag;
-+ break;
-+ }
-+ }
-+ } /* next frag loop */
-+ /* Couldn't statisfy the request with this
-+ largest page size, try a smaller one */
-+ largest_page <<= 2;
-+ }
-+ if (start_frag == NULL) {
-+ /* Couldn't find proper amount of space */
-+ ret = -ENOMEM;
-+ goto out;
-+ }
-+ i->did_create = 1;
-+do_map:
-+ /* Verify there is sufficient space to do the mapping */
-+ down_write(&current->mm->mmap_sem);
-+ next_addr = usdpaa_get_unmapped_area(fp, next_addr, i->len, 0, 0);
-+ up_write(&current->mm->mmap_sem);
-+
-+ if (next_addr & ~PAGE_MASK) {
-+ ret = -ENOMEM;
-+ goto out;
-+ }
-+
-+ /* We may need to divide the final fragment to accomidate the mapping */
-+ next_frag = start_frag;
-+ while (so_far != i->len) {
-+ BUG_ON(next_frag->len == 0);
-+ while ((next_frag->len + so_far) > i->len) {
-+ /* Split frag until they match */
-+ split_frag(next_frag);
-+ }
-+ so_far += next_frag->len;
-+ next_frag->refs++;
-+ ++frag_count;
-+ next_frag = list_entry(next_frag->list.prev,
-+ struct mem_fragment, list);
-+ }
-+ if (i->did_create) {
-+ size_t name_len = 0;
-+ start_frag->flags = i->flags;
-+ strncpy(start_frag->name, i->name, USDPAA_DMA_NAME_MAX);
-+ name_len = strnlen(start_frag->name, USDPAA_DMA_NAME_MAX);
-+ if (name_len >= USDPAA_DMA_NAME_MAX) {
-+ ret = -EFAULT;
-+ goto out;
-+ }
-+ start_frag->map_len = i->len;
-+ start_frag->has_locking = i->has_locking;
-+ init_waitqueue_head(&start_frag->wq);
-+ start_frag->owner = NULL;
-+ }
-+
-+ /* Setup the map entry */
-+ map->root_frag = start_frag;
-+ map->total_size = i->len;
-+ map->frag_count = frag_count;
-+ map->refs = 1;
-+ list_add(&map->list, &ctx->maps);
-+ i->phys_addr = start_frag->base;
-+out:
-+ spin_unlock(&mem_lock);
-+
-+ if (!ret) {
-+ unsigned long longret;
-+ down_write(&current->mm->mmap_sem);
-+ longret = do_mmap_pgoff(fp, next_addr, map->total_size,
-+ PROT_READ |
-+ (i->flags &
-+ USDPAA_DMA_FLAG_RDONLY ? 0
-+ : PROT_WRITE),
-+ MAP_SHARED,
-+ start_frag->pfn_base,
-+ &populate);
-+ up_write(&current->mm->mmap_sem);
-+ if (longret & ~PAGE_MASK) {
-+ ret = (int)longret;
-+ } else {
-+ i->ptr = (void *)longret;
-+ map->virt_addr = i->ptr;
-+ }
-+ } else
-+ kfree(map);
-+ return ret;
-+}
-+
-+static long ioctl_dma_unmap(struct ctx *ctx, void __user *arg)
-+{
-+ struct mem_mapping *map;
-+ struct vm_area_struct *vma;
-+ int ret, i;
-+ struct mem_fragment *current_frag;
-+ size_t sz;
-+ unsigned long base;
-+ unsigned long vaddr;
-+
-+ down_write(&current->mm->mmap_sem);
-+ vma = find_vma(current->mm, (unsigned long)arg);
-+ if (!vma || (vma->vm_start > (unsigned long)arg)) {
-+ up_write(&current->mm->mmap_sem);
-+ return -EFAULT;
-+ }
-+ spin_lock(&mem_lock);
-+ list_for_each_entry(map, &ctx->maps, list) {
-+ if (map->root_frag->pfn_base == vma->vm_pgoff) {
-+ /* Drop the map lock if we hold it */
-+ if (map->root_frag->has_locking &&
-+ (map->root_frag->owner == map)) {
-+ map->root_frag->owner = NULL;
-+ wake_up(&map->root_frag->wq);
-+ }
-+ goto map_match;
-+ }
-+ }
-+ /* Failed to find a matching mapping for this process */
-+ ret = -EFAULT;
-+ spin_unlock(&mem_lock);
-+ goto out;
-+map_match:
-+ map->refs--;
-+ if (map->refs != 0) {
-+ /* Another call the dma_map is referencing this */
-+ ret = 0;
-+ spin_unlock(&mem_lock);
-+ goto out;
-+ }
-+
-+ current_frag = map->root_frag;
-+ vaddr = (unsigned long) map->virt_addr;
-+ for (i = 0; i < map->frag_count; i++) {
-+ DPA_ASSERT(current_frag->refs > 0);
-+ --current_frag->refs;
-+#if !(defined(CONFIG_ARM) || defined(CONFIG_ARM64))
-+ /*
-+ * Make sure we invalidate the TLB entry for
-+ * this fragment, otherwise a remap of a different
-+ * page to this vaddr would give acces to an
-+ * incorrect piece of memory
-+ */
-+ cleartlbcam(vaddr, mfspr(SPRN_PID));
-+#endif
-+ vaddr += current_frag->len;
-+ current_frag = list_entry(current_frag->list.prev,
-+ struct mem_fragment, list);
-+ }
-+ map->root_frag->name[0] = 0;
-+ list_del(&map->list);
-+ compress_frags();
-+ spin_unlock(&mem_lock);
-+
-+ base = vma->vm_start;
-+ sz = vma->vm_end - vma->vm_start;
-+ do_munmap(current->mm, base, sz);
-+ ret = 0;
-+ out:
-+ up_write(&current->mm->mmap_sem);
-+ return ret;
-+}
-+
-+static long ioctl_dma_stats(struct ctx *ctx, void __user *arg)
-+{
-+ struct mem_fragment *frag;
-+ struct usdpaa_ioctl_dma_used result;
-+
-+ result.free_bytes = 0;
-+ result.total_bytes = phys_size;
-+
-+ list_for_each_entry(frag, &mem_list, list) {
-+ if (frag->refs == 0)
-+ result.free_bytes += frag->len;
-+ }
-+
-+ return copy_to_user(arg, &result, sizeof(result)); }
-+
-+static int test_lock(struct mem_mapping *map)
-+{
-+ int ret = 0;
-+ spin_lock(&mem_lock);
-+ if (!map->root_frag->owner) {
-+ map->root_frag->owner = map;
-+ ret = 1;
-+ }
-+ spin_unlock(&mem_lock);
-+ return ret;
-+}
-+
-+static long ioctl_dma_lock(struct ctx *ctx, void __user *arg)
-+{
-+ struct mem_mapping *map;
-+ struct vm_area_struct *vma;
-+
-+ down_read(&current->mm->mmap_sem);
-+ vma = find_vma(current->mm, (unsigned long)arg);
-+ if (!vma || (vma->vm_start > (unsigned long)arg)) {
-+ up_read(&current->mm->mmap_sem);
-+ return -EFAULT;
-+ }
-+ spin_lock(&mem_lock);
-+ list_for_each_entry(map, &ctx->maps, list) {
-+ if (map->root_frag->pfn_base == vma->vm_pgoff)
-+ goto map_match;
-+ }
-+ map = NULL;
-+map_match:
-+ spin_unlock(&mem_lock);
-+ up_read(&current->mm->mmap_sem);
-+
-+ if (!map)
-+ return -EFAULT;
-+ if (!map->root_frag->has_locking)
-+ return -ENODEV;
-+ return wait_event_interruptible(map->root_frag->wq, test_lock(map));
-+}
-+
-+static long ioctl_dma_unlock(struct ctx *ctx, void __user *arg)
-+{
-+ struct mem_mapping *map;
-+ struct vm_area_struct *vma;
-+ int ret;
-+
-+ down_read(&current->mm->mmap_sem);
-+ vma = find_vma(current->mm, (unsigned long)arg);
-+ if (!vma || (vma->vm_start > (unsigned long)arg))
-+ ret = -EFAULT;
-+ else {
-+ spin_lock(&mem_lock);
-+ list_for_each_entry(map, &ctx->maps, list) {
-+ if (map->root_frag->pfn_base == vma->vm_pgoff) {
-+ if (!map->root_frag->has_locking)
-+ ret = -ENODEV;
-+ else if (map->root_frag->owner == map) {
-+ map->root_frag->owner = NULL;
-+ wake_up(&map->root_frag->wq);
-+ ret = 0;
-+ } else
-+ ret = -EBUSY;
-+ goto map_match;
-+ }
-+ }
-+ ret = -EINVAL;
-+map_match:
-+ spin_unlock(&mem_lock);
-+ }
-+ up_read(&current->mm->mmap_sem);
-+ return ret;
-+}
-+
-+static int portal_mmap(struct file *fp, struct resource *res, void **ptr)
-+{
-+ unsigned long longret = 0, populate;
-+ resource_size_t len;
-+
-+ down_write(&current->mm->mmap_sem);
-+ len = resource_size(res);
-+ if (len != (unsigned long)len)
-+ return -EINVAL;
-+ longret = do_mmap_pgoff(fp, PAGE_SIZE, (unsigned long)len,
-+ PROT_READ | PROT_WRITE, MAP_SHARED,
-+ res->start >> PAGE_SHIFT, &populate);
-+ up_write(&current->mm->mmap_sem);
-+
-+ if (longret & ~PAGE_MASK)
-+ return (int)longret;
-+
-+ *ptr = (void *) longret;
-+ return 0;
-+}
-+
-+static void portal_munmap(struct resource *res, void *ptr)
-+{
-+ down_write(&current->mm->mmap_sem);
-+ do_munmap(current->mm, (unsigned long)ptr, resource_size(res));
-+ up_write(&current->mm->mmap_sem);
-+}
-+
-+static long ioctl_portal_map(struct file *fp, struct ctx *ctx,
-+ struct usdpaa_ioctl_portal_map *arg)
-+{
-+ struct portal_mapping *mapping = kmalloc(sizeof(*mapping), GFP_KERNEL);
-+ int ret;
-+
-+ if (!mapping)
-+ return -ENOMEM;
-+
-+ mapping->user = *arg;
-+ mapping->iommu_domain = NULL;
-+
-+ if (mapping->user.type == usdpaa_portal_qman) {
-+ mapping->qportal =
-+ qm_get_unused_portal_idx(mapping->user.index);
-+ if (!mapping->qportal) {
-+ ret = -ENODEV;
-+ goto err_get_portal;
-+ }
-+ mapping->phys = &mapping->qportal->addr_phys[0];
-+ mapping->user.channel = mapping->qportal->public_cfg.channel;
-+ mapping->user.pools = mapping->qportal->public_cfg.pools;
-+ mapping->user.index = mapping->qportal->public_cfg.index;
-+ } else if (mapping->user.type == usdpaa_portal_bman) {
-+ mapping->bportal =
-+ bm_get_unused_portal_idx(mapping->user.index);
-+ if (!mapping->bportal) {
-+ ret = -ENODEV;
-+ goto err_get_portal;
-+ }
-+ mapping->phys = &mapping->bportal->addr_phys[0];
-+ mapping->user.index = mapping->bportal->public_cfg.index;
-+ } else {
-+ ret = -EINVAL;
-+ goto err_copy_from_user;
-+ }
-+ /* Need to put pcfg in ctx's list before the mmaps because the mmap
-+ * handlers look it up. */
-+ spin_lock(&mem_lock);
-+ list_add(&mapping->list, &ctx->portals);
-+ spin_unlock(&mem_lock);
-+ ret = portal_mmap(fp, &mapping->phys[DPA_PORTAL_CE],
-+ &mapping->user.addr.cena);
-+ if (ret)
-+ goto err_mmap_cena;
-+ ret = portal_mmap(fp, &mapping->phys[DPA_PORTAL_CI],
-+ &mapping->user.addr.cinh);
-+ if (ret)
-+ goto err_mmap_cinh;
-+ *arg = mapping->user;
-+ return ret;
-+
-+err_mmap_cinh:
-+ portal_munmap(&mapping->phys[DPA_PORTAL_CE], mapping->user.addr.cena);
-+err_mmap_cena:
-+ if ((mapping->user.type == usdpaa_portal_qman) && mapping->qportal)
-+ qm_put_unused_portal(mapping->qportal);
-+ else if ((mapping->user.type == usdpaa_portal_bman) && mapping->bportal)
-+ bm_put_unused_portal(mapping->bportal);
-+ spin_lock(&mem_lock);
-+ list_del(&mapping->list);
-+ spin_unlock(&mem_lock);
-+err_get_portal:
-+err_copy_from_user:
-+ kfree(mapping);
-+ return ret;
-+}
-+
-+static long ioctl_portal_unmap(struct ctx *ctx, struct usdpaa_portal_map *i)
-+{
-+ struct portal_mapping *mapping;
-+ struct vm_area_struct *vma;
-+ unsigned long pfn;
-+ u32 channel;
-+
-+ /* Get the PFN corresponding to one of the virt addresses */
-+ down_read(&current->mm->mmap_sem);
-+ vma = find_vma(current->mm, (unsigned long)i->cinh);
-+ if (!vma || (vma->vm_start > (unsigned long)i->cinh)) {
-+ up_read(&current->mm->mmap_sem);
-+ return -EFAULT;
-+ }
-+ pfn = vma->vm_pgoff;
-+ up_read(&current->mm->mmap_sem);
-+
-+ /* Find the corresponding portal */
-+ spin_lock(&mem_lock);
-+ list_for_each_entry(mapping, &ctx->portals, list) {
-+ if (pfn == (mapping->phys[DPA_PORTAL_CI].start >> PAGE_SHIFT))
-+ goto found;
-+ }
-+ mapping = NULL;
-+found:
-+ if (mapping)
-+ list_del(&mapping->list);
-+ spin_unlock(&mem_lock);
-+ if (!mapping)
-+ return -ENODEV;
-+ portal_munmap(&mapping->phys[DPA_PORTAL_CI], mapping->user.addr.cinh);
-+ portal_munmap(&mapping->phys[DPA_PORTAL_CE], mapping->user.addr.cena);
-+ if (mapping->user.type == usdpaa_portal_qman) {
-+ init_qm_portal(mapping->qportal,
-+ &mapping->qman_portal_low);
-+
-+ /* Tear down any FQs this portal is referencing */
-+ channel = mapping->qportal->public_cfg.channel;
-+ qm_check_and_destroy_fqs(&mapping->qman_portal_low,
-+ &channel,
-+ check_portal_channel);
-+ qm_put_unused_portal(mapping->qportal);
-+ } else if (mapping->user.type == usdpaa_portal_bman) {
-+ init_bm_portal(mapping->bportal,
-+ &mapping->bman_portal_low);
-+ bm_put_unused_portal(mapping->bportal);
-+ }
-+ kfree(mapping);
-+ return 0;
-+}
-+
-+static void portal_config_pamu(struct qm_portal_config *pcfg, uint8_t sdest,
-+ uint32_t cpu, uint32_t cache, uint32_t window)
-+{
-+#ifdef CONFIG_FSL_PAMU
-+ int ret;
-+ int window_count = 1;
-+ struct iommu_domain_geometry geom_attr;
-+ struct pamu_stash_attribute stash_attr;
-+
-+ pcfg->iommu_domain = iommu_domain_alloc(&platform_bus_type);
-+ if (!pcfg->iommu_domain) {
-+ pr_err(KBUILD_MODNAME ":%s(): iommu_domain_alloc() failed",
-+ __func__);
-+ goto _no_iommu;
-+ }
-+ geom_attr.aperture_start = 0;
-+ geom_attr.aperture_end =
-+ ((dma_addr_t)1 << min(8 * sizeof(dma_addr_t), (size_t)36)) - 1;
-+ geom_attr.force_aperture = true;
-+ ret = iommu_domain_set_attr(pcfg->iommu_domain, DOMAIN_ATTR_GEOMETRY,
-+ &geom_attr);
-+ if (ret < 0) {
-+ pr_err(KBUILD_MODNAME ":%s(): iommu_domain_set_attr() = %d",
-+ __func__, ret);
-+ goto _iommu_domain_free;
-+ }
-+ ret = iommu_domain_set_attr(pcfg->iommu_domain, DOMAIN_ATTR_WINDOWS,
-+ &window_count);
-+ if (ret < 0) {
-+ pr_err(KBUILD_MODNAME ":%s(): iommu_domain_set_attr() = %d",
-+ __func__, ret);
-+ goto _iommu_domain_free;
-+ }
-+ stash_attr.cpu = cpu;
-+ stash_attr.cache = cache;
-+ /* set stash information for the window */
-+ stash_attr.window = 0;
-+
-+ ret = iommu_domain_set_attr(pcfg->iommu_domain,
-+ DOMAIN_ATTR_FSL_PAMU_STASH,
-+ &stash_attr);
-+ if (ret < 0) {
-+ pr_err(KBUILD_MODNAME ":%s(): iommu_domain_set_attr() = %d",
-+ __func__, ret);
-+ goto _iommu_domain_free;
-+ }
-+ ret = iommu_domain_window_enable(pcfg->iommu_domain, 0, 0, 1ULL << 36,
-+ IOMMU_READ | IOMMU_WRITE);
-+ if (ret < 0) {
-+ pr_err(KBUILD_MODNAME ":%s(): iommu_domain_window_enable() = %d",
-+ __func__, ret);
-+ goto _iommu_domain_free;
-+ }
-+ ret = iommu_attach_device(pcfg->iommu_domain, &pcfg->dev);
-+ if (ret < 0) {
-+ pr_err(KBUILD_MODNAME ":%s(): iommu_device_attach() = %d",
-+ __func__, ret);
-+ goto _iommu_domain_free;
-+ }
-+ ret = iommu_domain_set_attr(pcfg->iommu_domain,
-+ DOMAIN_ATTR_FSL_PAMU_ENABLE,
-+ &window_count);
-+ if (ret < 0) {
-+ pr_err(KBUILD_MODNAME ":%s(): iommu_domain_set_attr() = %d",
-+ __func__, ret);
-+ goto _iommu_detach_device;
-+ }
-+_no_iommu:
-+#endif
-+
-+#ifdef CONFIG_FSL_QMAN_CONFIG
-+ if (qman_set_sdest(pcfg->public_cfg.channel, sdest))
-+#endif
-+ pr_warn("Failed to set QMan portal's stash request queue\n");
-+
-+ return;
-+
-+#ifdef CONFIG_FSL_PAMU
-+_iommu_detach_device:
-+ iommu_detach_device(pcfg->iommu_domain, NULL);
-+_iommu_domain_free:
-+ iommu_domain_free(pcfg->iommu_domain);
-+#endif
-+}
-+
-+static long ioctl_allocate_raw_portal(struct file *fp, struct ctx *ctx,
-+ struct usdpaa_ioctl_raw_portal *arg)
-+{
-+ struct portal_mapping *mapping = kmalloc(sizeof(*mapping), GFP_KERNEL);
-+ int ret;
-+
-+ if (!mapping)
-+ return -ENOMEM;
-+
-+ mapping->user.type = arg->type;
-+ mapping->iommu_domain = NULL;
-+ if (arg->type == usdpaa_portal_qman) {
-+ mapping->qportal = qm_get_unused_portal_idx(arg->index);
-+ if (!mapping->qportal) {
-+ ret = -ENODEV;
-+ goto err;
-+ }
-+ mapping->phys = &mapping->qportal->addr_phys[0];
-+ arg->index = mapping->qportal->public_cfg.index;
-+ arg->cinh = mapping->qportal->addr_phys[DPA_PORTAL_CI].start;
-+ arg->cena = mapping->qportal->addr_phys[DPA_PORTAL_CE].start;
-+ if (arg->enable_stash) {
-+ /* Setup the PAMU with the supplied parameters */
-+ portal_config_pamu(mapping->qportal, arg->sdest,
-+ arg->cpu, arg->cache, arg->window);
-+ }
-+ } else if (mapping->user.type == usdpaa_portal_bman) {
-+ mapping->bportal =
-+ bm_get_unused_portal_idx(arg->index);
-+ if (!mapping->bportal) {
-+ ret = -ENODEV;
-+ goto err;
-+ }
-+ mapping->phys = &mapping->bportal->addr_phys[0];
-+ arg->index = mapping->bportal->public_cfg.index;
-+ arg->cinh = mapping->bportal->addr_phys[DPA_PORTAL_CI].start;
-+ arg->cena = mapping->bportal->addr_phys[DPA_PORTAL_CE].start;
-+ } else {
-+ ret = -EINVAL;
-+ goto err;
-+ }
-+ /* Need to put pcfg in ctx's list before the mmaps because the mmap
-+ * handlers look it up. */
-+ spin_lock(&mem_lock);
-+ list_add(&mapping->list, &ctx->portals);
-+ spin_unlock(&mem_lock);
-+ return 0;
-+err:
-+ kfree(mapping);
-+ return ret;
-+}
-+
-+static long ioctl_free_raw_portal(struct file *fp, struct ctx *ctx,
-+ struct usdpaa_ioctl_raw_portal *arg)
-+{
-+ struct portal_mapping *mapping;
-+ u32 channel;
-+
-+ /* Find the corresponding portal */
-+ spin_lock(&mem_lock);
-+ list_for_each_entry(mapping, &ctx->portals, list) {
-+ if (mapping->phys[DPA_PORTAL_CI].start == arg->cinh)
-+ goto found;
-+ }
-+ mapping = NULL;
-+found:
-+ if (mapping)
-+ list_del(&mapping->list);
-+ spin_unlock(&mem_lock);
-+ if (!mapping)
-+ return -ENODEV;
-+ if (mapping->user.type == usdpaa_portal_qman) {
-+ init_qm_portal(mapping->qportal,
-+ &mapping->qman_portal_low);
-+
-+ /* Tear down any FQs this portal is referencing */
-+ channel = mapping->qportal->public_cfg.channel;
-+ qm_check_and_destroy_fqs(&mapping->qman_portal_low,
-+ &channel,
-+ check_portal_channel);
-+ qm_put_unused_portal(mapping->qportal);
-+ } else if (mapping->user.type == usdpaa_portal_bman) {
-+ init_bm_portal(mapping->bportal,
-+ &mapping->bman_portal_low);
-+ bm_put_unused_portal(mapping->bportal);
-+ }
-+ kfree(mapping);
-+ return 0;
-+}
-+
-+static long usdpaa_ioctl(struct file *fp, unsigned int cmd, unsigned long arg)
-+{
-+ struct ctx *ctx = fp->private_data;
-+ void __user *a = (void __user *)arg;
-+ switch (cmd) {
-+ case USDPAA_IOCTL_ID_ALLOC:
-+ return ioctl_id_alloc(ctx, a);
-+ case USDPAA_IOCTL_ID_RELEASE:
-+ return ioctl_id_release(ctx, a);
-+ case USDPAA_IOCTL_ID_RESERVE:
-+ return ioctl_id_reserve(ctx, a);
-+ case USDPAA_IOCTL_DMA_MAP:
-+ {
-+ struct usdpaa_ioctl_dma_map input;
-+ int ret;
-+ if (copy_from_user(&input, a, sizeof(input)))
-+ return -EFAULT;
-+ ret = ioctl_dma_map(fp, ctx, &input);
-+ if (copy_to_user(a, &input, sizeof(input)))
-+ return -EFAULT;
-+ return ret;
-+ }
-+ case USDPAA_IOCTL_DMA_UNMAP:
-+ return ioctl_dma_unmap(ctx, a);
-+ case USDPAA_IOCTL_DMA_LOCK:
-+ return ioctl_dma_lock(ctx, a);
-+ case USDPAA_IOCTL_DMA_UNLOCK:
-+ return ioctl_dma_unlock(ctx, a);
-+ case USDPAA_IOCTL_PORTAL_MAP:
-+ {
-+ struct usdpaa_ioctl_portal_map input;
-+ int ret;
-+ if (copy_from_user(&input, a, sizeof(input)))
-+ return -EFAULT;
-+ ret = ioctl_portal_map(fp, ctx, &input);
-+ if (copy_to_user(a, &input, sizeof(input)))
-+ return -EFAULT;
-+ return ret;
-+ }
-+ case USDPAA_IOCTL_PORTAL_UNMAP:
-+ {
-+ struct usdpaa_portal_map input;
-+ if (copy_from_user(&input, a, sizeof(input)))
-+ return -EFAULT;
-+ return ioctl_portal_unmap(ctx, &input);
-+ }
-+ case USDPAA_IOCTL_DMA_USED:
-+ return ioctl_dma_stats(ctx, a);
-+ case USDPAA_IOCTL_ALLOC_RAW_PORTAL:
-+ {
-+ struct usdpaa_ioctl_raw_portal input;
-+ int ret;
-+ if (copy_from_user(&input, a, sizeof(input)))
-+ return -EFAULT;
-+ ret = ioctl_allocate_raw_portal(fp, ctx, &input);
-+ if (copy_to_user(a, &input, sizeof(input)))
-+ return -EFAULT;
-+ return ret;
-+ }
-+ case USDPAA_IOCTL_FREE_RAW_PORTAL:
-+ {
-+ struct usdpaa_ioctl_raw_portal input;
-+ if (copy_from_user(&input, a, sizeof(input)))
-+ return -EFAULT;
-+ return ioctl_free_raw_portal(fp, ctx, &input);
-+ }
-+ }
-+ return -EINVAL;
-+}
-+
-+static long usdpaa_ioctl_compat(struct file *fp, unsigned int cmd,
-+ unsigned long arg)
-+{
-+#ifdef CONFIG_COMPAT
-+ struct ctx *ctx = fp->private_data;
-+ void __user *a = (void __user *)arg;
-+#endif
-+ switch (cmd) {
-+#ifdef CONFIG_COMPAT
-+ case USDPAA_IOCTL_DMA_MAP_COMPAT:
-+ {
-+ int ret;
-+ struct usdpaa_ioctl_dma_map_compat input;
-+ struct usdpaa_ioctl_dma_map converted;
-+
-+ if (copy_from_user(&input, a, sizeof(input)))
-+ return -EFAULT;
-+
-+ converted.ptr = compat_ptr(input.ptr);
-+ converted.phys_addr = input.phys_addr;
-+ converted.len = input.len;
-+ converted.flags = input.flags;
-+ strncpy(converted.name, input.name, USDPAA_DMA_NAME_MAX);
-+ converted.has_locking = input.has_locking;
-+ converted.did_create = input.did_create;
-+
-+ ret = ioctl_dma_map(fp, ctx, &converted);
-+ input.ptr = ptr_to_compat(converted.ptr);
-+ input.phys_addr = converted.phys_addr;
-+ input.len = converted.len;
-+ input.flags = converted.flags;
-+ strncpy(input.name, converted.name, USDPAA_DMA_NAME_MAX);
-+ input.has_locking = converted.has_locking;
-+ input.did_create = converted.did_create;
-+ if (copy_to_user(a, &input, sizeof(input)))
-+ return -EFAULT;
-+ return ret;
-+ }
-+ case USDPAA_IOCTL_PORTAL_MAP_COMPAT:
-+ {
-+ int ret;
-+ struct compat_usdpaa_ioctl_portal_map input;
-+ struct usdpaa_ioctl_portal_map converted;
-+ if (copy_from_user(&input, a, sizeof(input)))
-+ return -EFAULT;
-+ converted.type = input.type;
-+ converted.index = input.index;
-+ ret = ioctl_portal_map(fp, ctx, &converted);
-+ input.addr.cinh = ptr_to_compat(converted.addr.cinh);
-+ input.addr.cena = ptr_to_compat(converted.addr.cena);
-+ input.channel = converted.channel;
-+ input.pools = converted.pools;
-+ input.index = converted.index;
-+ if (copy_to_user(a, &input, sizeof(input)))
-+ return -EFAULT;
-+ return ret;
-+ }
-+ case USDPAA_IOCTL_PORTAL_UNMAP_COMPAT:
-+ {
-+ struct usdpaa_portal_map_compat input;
-+ struct usdpaa_portal_map converted;
-+
-+ if (copy_from_user(&input, a, sizeof(input)))
-+ return -EFAULT;
-+ converted.cinh = compat_ptr(input.cinh);
-+ converted.cena = compat_ptr(input.cena);
-+ return ioctl_portal_unmap(ctx, &converted);
-+ }
-+ case USDPAA_IOCTL_ALLOC_RAW_PORTAL_COMPAT:
-+ {
-+ int ret;
-+ struct usdpaa_ioctl_raw_portal converted;
-+ struct compat_ioctl_raw_portal input;
-+ if (copy_from_user(&input, a, sizeof(input)))
-+ return -EFAULT;
-+ converted.type = input.type;
-+ converted.index = input.index;
-+ converted.enable_stash = input.enable_stash;
-+ converted.cpu = input.cpu;
-+ converted.cache = input.cache;
-+ converted.window = input.window;
-+ converted.sdest = input.sdest;
-+ ret = ioctl_allocate_raw_portal(fp, ctx, &converted);
-+
-+ input.cinh = converted.cinh;
-+ input.cena = converted.cena;
-+ input.index = converted.index;
-+
-+ if (copy_to_user(a, &input, sizeof(input)))
-+ return -EFAULT;
-+ return ret;
-+ }
-+ case USDPAA_IOCTL_FREE_RAW_PORTAL_COMPAT:
-+ {
-+ struct usdpaa_ioctl_raw_portal converted;
-+ struct compat_ioctl_raw_portal input;
-+ if (copy_from_user(&input, a, sizeof(input)))
-+ return -EFAULT;
-+ converted.type = input.type;
-+ converted.index = input.index;
-+ converted.cinh = input.cinh;
-+ converted.cena = input.cena;
-+ return ioctl_free_raw_portal(fp, ctx, &converted);
-+ }
-+#endif
-+ default:
-+ return usdpaa_ioctl(fp, cmd, arg);
-+ }
-+ return -EINVAL;
-+}
-+
-+int usdpaa_get_portal_config(struct file *filp, void *cinh,
-+ enum usdpaa_portal_type ptype, unsigned int *irq,
-+ void **iir_reg)
-+{
-+ /* Walk the list of portals for filp and return the config
-+ for the portal that matches the hint */
-+ struct ctx *context;
-+ struct portal_mapping *portal;
-+
-+ /* First sanitize the filp */
-+ if (filp->f_op->open != usdpaa_open)
-+ return -ENODEV;
-+ context = filp->private_data;
-+ spin_lock(&context->lock);
-+ list_for_each_entry(portal, &context->portals, list) {
-+ if (portal->user.type == ptype &&
-+ portal->user.addr.cinh == cinh) {
-+ if (ptype == usdpaa_portal_qman) {
-+ *irq = portal->qportal->public_cfg.irq;
-+ *iir_reg = portal->qportal->addr_virt[1] +
-+ QM_REG_IIR;
-+ } else {
-+ *irq = portal->bportal->public_cfg.irq;
-+ *iir_reg = portal->bportal->addr_virt[1] +
-+ BM_REG_IIR;
-+ }
-+ spin_unlock(&context->lock);
-+ return 0;
-+ }
-+ }
-+ spin_unlock(&context->lock);
-+ return -EINVAL;
-+}
-+
-+static const struct file_operations usdpaa_fops = {
-+ .open = usdpaa_open,
-+ .release = usdpaa_release,
-+ .mmap = usdpaa_mmap,
-+ .get_unmapped_area = usdpaa_get_unmapped_area,
-+ .unlocked_ioctl = usdpaa_ioctl,
-+ .compat_ioctl = usdpaa_ioctl_compat
-+};
-+
-+static struct miscdevice usdpaa_miscdev = {
-+ .name = "fsl-usdpaa",
-+ .fops = &usdpaa_fops,
-+ .minor = MISC_DYNAMIC_MINOR,
-+};
-+
-+/* Early-boot memory allocation. The boot-arg "usdpaa_mem=<x>" is used to
-+ * indicate how much memory (if any) to allocate during early boot. If the
-+ * format "usdpaa_mem=<x>,<y>" is used, then <y> will be interpreted as the
-+ * number of TLB1 entries to reserve (default is 1). If there are more mappings
-+ * than there are TLB1 entries, fault-handling will occur. */
-+
-+static __init int usdpaa_mem(char *arg)
-+{
-+ pr_warn("uspdaa_mem argument is depracated\n");
-+ arg_phys_size = memparse(arg, &arg);
-+ num_tlb = 1;
-+ if (*arg == ',') {
-+ unsigned long ul;
-+ int err = kstrtoul(arg + 1, 0, &ul);
-+ if (err < 0) {
-+ num_tlb = 1;
-+ pr_warn("ERROR, usdpaa_mem arg is invalid\n");
-+ } else
-+ num_tlb = (unsigned int)ul;
-+ }
-+ return 0;
-+}
-+early_param("usdpaa_mem", usdpaa_mem);
-+
-+static int usdpaa_mem_init(struct reserved_mem *rmem)
-+{
-+ phys_start = rmem->base;
-+ phys_size = rmem->size;
-+
-+ WARN_ON(!(phys_start && phys_size));
-+
-+ return 0;
-+}
-+RESERVEDMEM_OF_DECLARE(usdpaa_mem_init, "fsl,usdpaa-mem", usdpaa_mem_init);
-+
-+__init int fsl_usdpaa_init_early(void)
-+{
-+ if (!phys_size || !phys_start) {
-+ pr_info("No USDPAA memory, no 'fsl,usdpaa-mem' in device-tree\n");
-+ return 0;
-+ }
-+ if (phys_size % PAGE_SIZE) {
-+ pr_err("'fsl,usdpaa-mem' size must be a multiple of page size\n");
-+ phys_size = 0;
-+ return 0;
-+ }
-+ if (arg_phys_size && phys_size != arg_phys_size) {
-+ pr_err("'usdpaa_mem argument size (0x%llx) does not match device tree size (0x%llx)\n",
-+ arg_phys_size, phys_size);
-+ phys_size = 0;
-+ return 0;
-+ }
-+ pfn_start = phys_start >> PAGE_SHIFT;
-+ pfn_size = phys_size >> PAGE_SHIFT;
-+#ifdef CONFIG_PPC
-+ first_tlb = current_tlb = tlbcam_index;
-+ tlbcam_index += num_tlb;
-+#endif
-+ pr_info("USDPAA region at %llx:%llx(%lx:%lx), %d TLB1 entries)\n",
-+ phys_start, phys_size, pfn_start, pfn_size, num_tlb);
-+ return 0;
-+}
-+subsys_initcall(fsl_usdpaa_init_early);
-+
-+
-+static int __init usdpaa_init(void)
-+{
-+ struct mem_fragment *frag;
-+ int ret;
-+ u64 tmp_size = phys_size;
-+ u64 tmp_start = phys_start;
-+ u64 tmp_pfn_size = pfn_size;
-+ u64 tmp_pfn_start = pfn_start;
-+
-+ pr_info("Freescale USDPAA process driver\n");
-+ if (!phys_start) {
-+ pr_warn("fsl-usdpaa: no region found\n");
-+ return 0;
-+ }
-+
-+ while (tmp_size != 0) {
-+ u32 frag_size = largest_page_size(tmp_size);
-+ frag = kmalloc(sizeof(*frag), GFP_KERNEL);
-+ if (!frag) {
-+ pr_err("Failed to setup USDPAA memory accounting\n");
-+ return -ENOMEM;
-+ }
-+ frag->base = tmp_start;
-+ frag->len = frag->root_len = frag_size;
-+ frag->root_pfn = tmp_pfn_start;
-+ frag->pfn_base = tmp_pfn_start;
-+ frag->pfn_len = frag_size / PAGE_SIZE;
-+ frag->refs = 0;
-+ init_waitqueue_head(&frag->wq);
-+ frag->owner = NULL;
-+ list_add(&frag->list, &mem_list);
-+
-+ /* Adjust for this frag */
-+ tmp_start += frag_size;
-+ tmp_size -= frag_size;
-+ tmp_pfn_start += frag_size / PAGE_SIZE;
-+ tmp_pfn_size -= frag_size / PAGE_SIZE;
-+ }
-+ ret = misc_register(&usdpaa_miscdev);
-+ if (ret)
-+ pr_err("fsl-usdpaa: failed to register misc device\n");
-+ return ret;
-+}
-+
-+static void __exit usdpaa_exit(void)
-+{
-+ misc_deregister(&usdpaa_miscdev);
-+}
-+
-+module_init(usdpaa_init);
-+module_exit(usdpaa_exit);
-+
-+MODULE_LICENSE("GPL");
-+MODULE_AUTHOR("Freescale Semiconductor");
-+MODULE_DESCRIPTION("Freescale USDPAA process driver");
---- /dev/null
-+++ b/drivers/staging/fsl_qbman/fsl_usdpaa_irq.c
-@@ -0,0 +1,289 @@
-+/* Copyright (c) 2013 Freescale Semiconductor, Inc.
-+ * All rights reserved.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+/* define a device that allows USPDAA processes to open a file
-+ descriptor and specify which IRQ it wants to montior using an ioctl()
-+ When an IRQ is received, the device becomes readable so that a process
-+ can use read() or select() type calls to monitor for IRQs */
-+
-+#include <linux/miscdevice.h>
-+#include <linux/fs.h>
-+#include <linux/cdev.h>
-+#include <linux/slab.h>
-+#include <linux/interrupt.h>
-+#include <linux/poll.h>
-+#include <linux/uaccess.h>
-+#include <linux/fsl_usdpaa.h>
-+#include <linux/module.h>
-+#include <linux/fdtable.h>
-+#include <linux/file.h>
-+
-+#include "qman_low.h"
-+#include "bman_low.h"
-+
-+struct usdpaa_irq_ctx {
-+ int irq_set; /* Set to true once the irq is set via ioctl */
-+ unsigned int irq_num;
-+ u32 last_irq_count; /* Last value returned from read */
-+ u32 irq_count; /* Number of irqs since last read */
-+ wait_queue_head_t wait_queue; /* Waiting processes */
-+ spinlock_t lock;
-+ void *inhibit_addr; /* inhibit register address */
-+ struct file *usdpaa_filp;
-+ char irq_name[128];
-+};
-+
-+static int usdpaa_irq_open(struct inode *inode, struct file *filp)
-+{
-+ struct usdpaa_irq_ctx *ctx = kmalloc(sizeof(*ctx), GFP_KERNEL);
-+ if (!ctx)
-+ return -ENOMEM;
-+ ctx->irq_set = 0;
-+ ctx->irq_count = 0;
-+ ctx->last_irq_count = 0;
-+ init_waitqueue_head(&ctx->wait_queue);
-+ spin_lock_init(&ctx->lock);
-+ filp->private_data = ctx;
-+ return 0;
-+}
-+
-+static int usdpaa_irq_release(struct inode *inode, struct file *filp)
-+{
-+ struct usdpaa_irq_ctx *ctx = filp->private_data;
-+ if (ctx->irq_set) {
-+ /* Inhibit the IRQ */
-+ out_be32(ctx->inhibit_addr, 0x1);
-+ irq_set_affinity_hint(ctx->irq_num, NULL);
-+ free_irq(ctx->irq_num, ctx);
-+ ctx->irq_set = 0;
-+ fput(ctx->usdpaa_filp);
-+ }
-+ kfree(filp->private_data);
-+ return 0;
-+}
-+
-+static irqreturn_t usdpaa_irq_handler(int irq, void *_ctx)
-+{
-+ unsigned long flags;
-+ struct usdpaa_irq_ctx *ctx = _ctx;
-+ spin_lock_irqsave(&ctx->lock, flags);
-+ ++ctx->irq_count;
-+ spin_unlock_irqrestore(&ctx->lock, flags);
-+ wake_up_all(&ctx->wait_queue);
-+ /* Set the inhibit register. This will be reenabled
-+ once the USDPAA code handles the IRQ */
-+ out_be32(ctx->inhibit_addr, 0x1);
-+ pr_info("Inhibit at %p count %d", ctx->inhibit_addr, ctx->irq_count);
-+ return IRQ_HANDLED;
-+}
-+
-+static int map_irq(struct file *fp, struct usdpaa_ioctl_irq_map *irq_map)
-+{
-+ struct usdpaa_irq_ctx *ctx = fp->private_data;
-+ int ret;
-+
-+ if (ctx->irq_set) {
-+ pr_debug("Setting USDPAA IRQ when it was already set!\n");
-+ return -EBUSY;
-+ }
-+
-+ ctx->usdpaa_filp = fget(irq_map->fd);
-+ if (!ctx->usdpaa_filp) {
-+ pr_debug("USDPAA fget(%d) returned NULL\n", irq_map->fd);
-+ return -EINVAL;
-+ }
-+
-+ ret = usdpaa_get_portal_config(ctx->usdpaa_filp, irq_map->portal_cinh,
-+ irq_map->type, &ctx->irq_num,
-+ &ctx->inhibit_addr);
-+ if (ret) {
-+ pr_debug("USDPAA IRQ couldn't identify portal\n");
-+ fput(ctx->usdpaa_filp);
-+ return ret;
-+ }
-+
-+ ctx->irq_set = 1;
-+
-+ snprintf(ctx->irq_name, sizeof(ctx->irq_name),
-+ "usdpaa_irq %d", ctx->irq_num);
-+
-+ ret = request_irq(ctx->irq_num, usdpaa_irq_handler, 0,
-+ ctx->irq_name, ctx);
-+ if (ret) {
-+ pr_err("USDPAA request_irq(%d) failed, ret= %d\n",
-+ ctx->irq_num, ret);
-+ ctx->irq_set = 0;
-+ fput(ctx->usdpaa_filp);
-+ return ret;
-+ }
-+ ret = irq_set_affinity(ctx->irq_num, &current->cpus_allowed);
-+ if (ret)
-+ pr_err("USDPAA irq_set_affinity() failed, ret= %d\n", ret);
-+
-+ ret = irq_set_affinity_hint(ctx->irq_num, &current->cpus_allowed);
-+ if (ret)
-+ pr_err("USDPAA irq_set_affinity_hint() failed, ret= %d\n", ret);
-+
-+ return 0;
-+}
-+
-+static long usdpaa_irq_ioctl(struct file *fp, unsigned int cmd,
-+ unsigned long arg)
-+{
-+ int ret;
-+ struct usdpaa_ioctl_irq_map irq_map;
-+
-+ if (cmd != USDPAA_IOCTL_PORTAL_IRQ_MAP) {
-+ pr_debug("USDPAA IRQ unknown command 0x%x\n", cmd);
-+ return -EINVAL;
-+ }
-+
-+ ret = copy_from_user(&irq_map, (void __user *)arg,
-+ sizeof(irq_map));
-+ if (ret)
-+ return ret;
-+ return map_irq(fp, &irq_map);
-+}
-+
-+static ssize_t usdpaa_irq_read(struct file *filp, char __user *buff,
-+ size_t count, loff_t *offp)
-+{
-+ struct usdpaa_irq_ctx *ctx = filp->private_data;
-+ int ret;
-+
-+ if (!ctx->irq_set) {
-+ pr_debug("Reading USDPAA IRQ before it was set\n");
-+ return -EINVAL;
-+ }
-+
-+ if (count < sizeof(ctx->irq_count)) {
-+ pr_debug("USDPAA IRQ Read too small\n");
-+ return -EINVAL;
-+ }
-+ if (ctx->irq_count == ctx->last_irq_count) {
-+ if (filp->f_flags & O_NONBLOCK)
-+ return -EAGAIN;
-+
-+ ret = wait_event_interruptible(ctx->wait_queue,
-+ ctx->irq_count != ctx->last_irq_count);
-+ if (ret == -ERESTARTSYS)
-+ return ret;
-+ }
-+
-+ ctx->last_irq_count = ctx->irq_count;
-+
-+ if (copy_to_user(buff, &ctx->last_irq_count,
-+ sizeof(ctx->last_irq_count)))
-+ return -EFAULT;
-+ return sizeof(ctx->irq_count);
-+}
-+
-+static unsigned int usdpaa_irq_poll(struct file *filp, poll_table *wait)
-+{
-+ struct usdpaa_irq_ctx *ctx = filp->private_data;
-+ unsigned int ret = 0;
-+ unsigned long flags;
-+
-+ if (!ctx->irq_set)
-+ return POLLHUP;
-+
-+ poll_wait(filp, &ctx->wait_queue, wait);
-+
-+ spin_lock_irqsave(&ctx->lock, flags);
-+ if (ctx->irq_count != ctx->last_irq_count)
-+ ret |= POLLIN | POLLRDNORM;
-+ spin_unlock_irqrestore(&ctx->lock, flags);
-+ return ret;
-+}
-+
-+static long usdpaa_irq_ioctl_compat(struct file *fp, unsigned int cmd,
-+ unsigned long arg)
-+{
-+#ifdef CONFIG_COMPAT
-+ void __user *a = (void __user *)arg;
-+#endif
-+ switch (cmd) {
-+#ifdef CONFIG_COMPAT
-+ case USDPAA_IOCTL_PORTAL_IRQ_MAP_COMPAT:
-+ {
-+ struct compat_ioctl_irq_map input;
-+ struct usdpaa_ioctl_irq_map converted;
-+ if (copy_from_user(&input, a, sizeof(input)))
-+ return -EFAULT;
-+ converted.type = input.type;
-+ converted.fd = input.fd;
-+ converted.portal_cinh = compat_ptr(input.portal_cinh);
-+ return map_irq(fp, &converted);
-+ }
-+#endif
-+ default:
-+ return usdpaa_irq_ioctl(fp, cmd, arg);
-+ }
-+}
-+
-+static const struct file_operations usdpaa_irq_fops = {
-+ .open = usdpaa_irq_open,
-+ .release = usdpaa_irq_release,
-+ .unlocked_ioctl = usdpaa_irq_ioctl,
-+ .compat_ioctl = usdpaa_irq_ioctl_compat,
-+ .read = usdpaa_irq_read,
-+ .poll = usdpaa_irq_poll
-+};
-+
-+static struct miscdevice usdpaa_miscdev = {
-+ .name = "fsl-usdpaa-irq",
-+ .fops = &usdpaa_irq_fops,
-+ .minor = MISC_DYNAMIC_MINOR,
-+};
-+
-+static int __init usdpaa_irq_init(void)
-+{
-+ int ret;
-+
-+ pr_info("Freescale USDPAA process IRQ driver\n");
-+ ret = misc_register(&usdpaa_miscdev);
-+ if (ret)
-+ pr_err("fsl-usdpaa-irq: failed to register misc device\n");
-+ return ret;
-+}
-+
-+static void __exit usdpaa_irq_exit(void)
-+{
-+ misc_deregister(&usdpaa_miscdev);
-+}
-+
-+module_init(usdpaa_irq_init);
-+module_exit(usdpaa_irq_exit);
-+
-+MODULE_LICENSE("GPL");
-+MODULE_AUTHOR("Freescale Semiconductor");
-+MODULE_DESCRIPTION("Freescale USDPAA process IRQ driver");
---- /dev/null
-+++ b/drivers/staging/fsl_qbman/qbman_driver.c
-@@ -0,0 +1,88 @@
-+/* Copyright 2013 Freescale Semiconductor, Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#include <linux/time.h>
-+#include "qman_private.h"
-+#include "bman_private.h"
-+__init void qman_init_early(void);
-+__init void bman_init_early(void);
-+
-+static __init int qbman_init(void)
-+{
-+ struct device_node *dn;
-+ u32 is_portal_available;
-+
-+ bman_init();
-+ qman_init();
-+
-+ is_portal_available = 0;
-+ for_each_compatible_node(dn, NULL, "fsl,qman-portal") {
-+ if (!of_device_is_available(dn))
-+ continue;
-+ else
-+ is_portal_available = 1;
-+ }
-+
-+ if (!qman_have_ccsr() && is_portal_available) {
-+ struct qman_fq fq = {
-+ .fqid = 1
-+ };
-+ struct qm_mcr_queryfq_np np;
-+ int err, retry = CONFIG_FSL_QMAN_INIT_TIMEOUT;
-+ struct timespec nowts, diffts, startts = current_kernel_time();
-+ /* Loop while querying given fqid succeeds or time out */
-+ while (1) {
-+ err = qman_query_fq_np(&fq, &np);
-+ if (!err) {
-+ /* success, control-plane has configured QMan */
-+ break;
-+ } else if (err != -ERANGE) {
-+ pr_err("QMan: I/O error, continuing anyway\n");
-+ break;
-+ }
-+ nowts = current_kernel_time();
-+ diffts = timespec_sub(nowts, startts);
-+ if (diffts.tv_sec > 0) {
-+ if (!retry--) {
-+ pr_err("QMan: time out, control-plane"
-+ " dead?\n");
-+ break;
-+ }
-+ pr_warn("QMan: polling for the control-plane"
-+ " (%d)\n", retry);
-+ }
-+ }
-+ }
-+ bman_resource_init();
-+ qman_resource_init();
-+ return 0;
-+}
-+subsys_initcall(qbman_init);
---- /dev/null
-+++ b/drivers/staging/fsl_qbman/qman_config.c
-@@ -0,0 +1,1224 @@
-+/* Copyright 2008-2012 Freescale Semiconductor, Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#include <asm/cacheflush.h>
-+#include "qman_private.h"
-+#include <linux/highmem.h>
-+#include <linux/of_reserved_mem.h>
-+
-+/* Last updated for v00.800 of the BG */
-+
-+/* Register offsets */
-+#define REG_QCSP_LIO_CFG(n) (0x0000 + ((n) * 0x10))
-+#define REG_QCSP_IO_CFG(n) (0x0004 + ((n) * 0x10))
-+#define REG_QCSP_DD_CFG(n) (0x000c + ((n) * 0x10))
-+#define REG_DD_CFG 0x0200
-+#define REG_DCP_CFG(n) (0x0300 + ((n) * 0x10))
-+#define REG_DCP_DD_CFG(n) (0x0304 + ((n) * 0x10))
-+#define REG_DCP_DLM_AVG(n) (0x030c + ((n) * 0x10))
-+#define REG_PFDR_FPC 0x0400
-+#define REG_PFDR_FP_HEAD 0x0404
-+#define REG_PFDR_FP_TAIL 0x0408
-+#define REG_PFDR_FP_LWIT 0x0410
-+#define REG_PFDR_CFG 0x0414
-+#define REG_SFDR_CFG 0x0500
-+#define REG_SFDR_IN_USE 0x0504
-+#define REG_WQ_CS_CFG(n) (0x0600 + ((n) * 0x04))
-+#define REG_WQ_DEF_ENC_WQID 0x0630
-+#define REG_WQ_SC_DD_CFG(n) (0x640 + ((n) * 0x04))
-+#define REG_WQ_PC_DD_CFG(n) (0x680 + ((n) * 0x04))
-+#define REG_WQ_DC0_DD_CFG(n) (0x6c0 + ((n) * 0x04))
-+#define REG_WQ_DC1_DD_CFG(n) (0x700 + ((n) * 0x04))
-+#define REG_WQ_DCn_DD_CFG(n) (0x6c0 + ((n) * 0x40)) /* n=2,3 */
-+#define REG_CM_CFG 0x0800
-+#define REG_ECSR 0x0a00
-+#define REG_ECIR 0x0a04
-+#define REG_EADR 0x0a08
-+#define REG_ECIR2 0x0a0c
-+#define REG_EDATA(n) (0x0a10 + ((n) * 0x04))
-+#define REG_SBEC(n) (0x0a80 + ((n) * 0x04))
-+#define REG_MCR 0x0b00
-+#define REG_MCP(n) (0x0b04 + ((n) * 0x04))
-+#define REG_MISC_CFG 0x0be0
-+#define REG_HID_CFG 0x0bf0
-+#define REG_IDLE_STAT 0x0bf4
-+#define REG_IP_REV_1 0x0bf8
-+#define REG_IP_REV_2 0x0bfc
-+#define REG_FQD_BARE 0x0c00
-+#define REG_PFDR_BARE 0x0c20
-+#define REG_offset_BAR 0x0004 /* relative to REG_[FQD|PFDR]_BARE */
-+#define REG_offset_AR 0x0010 /* relative to REG_[FQD|PFDR]_BARE */
-+#define REG_QCSP_BARE 0x0c80
-+#define REG_QCSP_BAR 0x0c84
-+#define REG_CI_SCHED_CFG 0x0d00
-+#define REG_SRCIDR 0x0d04
-+#define REG_LIODNR 0x0d08
-+#define REG_CI_RLM_AVG 0x0d14
-+#define REG_ERR_ISR 0x0e00 /* + "enum qm_isr_reg" */
-+#define REG_REV3_QCSP_LIO_CFG(n) (0x1000 + ((n) * 0x10))
-+#define REG_REV3_QCSP_IO_CFG(n) (0x1004 + ((n) * 0x10))
-+#define REG_REV3_QCSP_DD_CFG(n) (0x100c + ((n) * 0x10))
-+#define REG_CEETM_CFG_IDX 0x900
-+#define REG_CEETM_CFG_PRES 0x904
-+#define REG_CEETM_XSFDR_IN_USE 0x908
-+
-+/* Assists for QMAN_MCR */
-+#define MCR_INIT_PFDR 0x01000000
-+#define MCR_get_rslt(v) (u8)((v) >> 24)
-+#define MCR_rslt_idle(r) (!rslt || (rslt >= 0xf0))
-+#define MCR_rslt_ok(r) (rslt == 0xf0)
-+#define MCR_rslt_eaccess(r) (rslt == 0xf8)
-+#define MCR_rslt_inval(r) (rslt == 0xff)
-+
-+struct qman;
-+
-+/* Follows WQ_CS_CFG0-5 */
-+enum qm_wq_class {
-+ qm_wq_portal = 0,
-+ qm_wq_pool = 1,
-+ qm_wq_fman0 = 2,
-+ qm_wq_fman1 = 3,
-+ qm_wq_caam = 4,
-+ qm_wq_pme = 5,
-+ qm_wq_first = qm_wq_portal,
-+ qm_wq_last = qm_wq_pme
-+};
-+
-+/* Follows FQD_[BARE|BAR|AR] and PFDR_[BARE|BAR|AR] */
-+enum qm_memory {
-+ qm_memory_fqd,
-+ qm_memory_pfdr
-+};
-+
-+/* Used by all error interrupt registers except 'inhibit' */
-+#define QM_EIRQ_CIDE 0x20000000 /* Corenet Initiator Data Error */
-+#define QM_EIRQ_CTDE 0x10000000 /* Corenet Target Data Error */
-+#define QM_EIRQ_CITT 0x08000000 /* Corenet Invalid Target Transaction */
-+#define QM_EIRQ_PLWI 0x04000000 /* PFDR Low Watermark */
-+#define QM_EIRQ_MBEI 0x02000000 /* Multi-bit ECC Error */
-+#define QM_EIRQ_SBEI 0x01000000 /* Single-bit ECC Error */
-+#define QM_EIRQ_PEBI 0x00800000 /* PFDR Enqueues Blocked Interrupt */
-+#define QM_EIRQ_IFSI 0x00020000 /* Invalid FQ Flow Control State */
-+#define QM_EIRQ_ICVI 0x00010000 /* Invalid Command Verb */
-+#define QM_EIRQ_IDDI 0x00000800 /* Invalid Dequeue (Direct-connect) */
-+#define QM_EIRQ_IDFI 0x00000400 /* Invalid Dequeue FQ */
-+#define QM_EIRQ_IDSI 0x00000200 /* Invalid Dequeue Source */
-+#define QM_EIRQ_IDQI 0x00000100 /* Invalid Dequeue Queue */
-+#define QM_EIRQ_IECE 0x00000010 /* Invalid Enqueue Configuration */
-+#define QM_EIRQ_IEOI 0x00000008 /* Invalid Enqueue Overflow */
-+#define QM_EIRQ_IESI 0x00000004 /* Invalid Enqueue State */
-+#define QM_EIRQ_IECI 0x00000002 /* Invalid Enqueue Channel */
-+#define QM_EIRQ_IEQI 0x00000001 /* Invalid Enqueue Queue */
-+
-+/* QMAN_ECIR valid error bit */
-+#define PORTAL_ECSR_ERR (QM_EIRQ_IEQI | QM_EIRQ_IESI | QM_EIRQ_IEOI | \
-+ QM_EIRQ_IDQI | QM_EIRQ_IDSI | QM_EIRQ_IDFI | \
-+ QM_EIRQ_IDDI | QM_EIRQ_ICVI | QM_EIRQ_IFSI)
-+#define FQID_ECSR_ERR (QM_EIRQ_IEQI | QM_EIRQ_IECI | QM_EIRQ_IESI | \
-+ QM_EIRQ_IEOI | QM_EIRQ_IDQI | QM_EIRQ_IDFI | \
-+ QM_EIRQ_IFSI)
-+
-+union qman_ecir {
-+ u32 ecir_raw;
-+ struct {
-+#if __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
-+ u32 __reserved:2;
-+ u32 portal_type:1;
-+ u32 portal_num:5;
-+ u32 fqid:24;
-+#else
-+ u32 fqid:24;
-+ u32 portal_num:5;
-+ u32 portal_type:1;
-+ u32 __reserved:2;
-+#endif
-+ } __packed info;
-+};
-+
-+union qman_ecir2 {
-+ u32 ecir2_raw;
-+ struct {
-+#if __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
-+ u32 portal_type:1;
-+ u32 __reserved:21;
-+ u32 portal_num:10;
-+#else
-+ u32 portal_num:10;
-+ u32 __reserved:21;
-+ u32 portal_type:1;
-+#endif
-+ } __packed info;
-+};
-+
-+union qman_eadr {
-+ u32 eadr_raw;
-+ struct {
-+#if __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
-+ u32 __reserved1:4;
-+ u32 memid:4;
-+ u32 __reserved2:12;
-+ u32 eadr:12;
-+#else
-+ u32 eadr:12;
-+ u32 __reserved2:12;
-+ u32 memid:4;
-+ u32 __reserved1:4;
-+#endif
-+ } __packed info;
-+ struct {
-+#if __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
-+ u32 __reserved1:3;
-+ u32 memid:5;
-+ u32 __reserved:8;
-+ u32 eadr:16;
-+#else
-+ u32 eadr:16;
-+ u32 __reserved:8;
-+ u32 memid:5;
-+ u32 __reserved1:3;
-+#endif
-+ } __packed info_rev3;
-+};
-+
-+struct qman_hwerr_txt {
-+ u32 mask;
-+ const char *txt;
-+};
-+
-+#define QMAN_HWE_TXT(a, b) { .mask = QM_EIRQ_##a, .txt = b }
-+
-+static const struct qman_hwerr_txt qman_hwerr_txts[] = {
-+ QMAN_HWE_TXT(CIDE, "Corenet Initiator Data Error"),
-+ QMAN_HWE_TXT(CTDE, "Corenet Target Data Error"),
-+ QMAN_HWE_TXT(CITT, "Corenet Invalid Target Transaction"),
-+ QMAN_HWE_TXT(PLWI, "PFDR Low Watermark"),
-+ QMAN_HWE_TXT(MBEI, "Multi-bit ECC Error"),
-+ QMAN_HWE_TXT(SBEI, "Single-bit ECC Error"),
-+ QMAN_HWE_TXT(PEBI, "PFDR Enqueues Blocked Interrupt"),
-+ QMAN_HWE_TXT(ICVI, "Invalid Command Verb"),
-+ QMAN_HWE_TXT(IFSI, "Invalid Flow Control State"),
-+ QMAN_HWE_TXT(IDDI, "Invalid Dequeue (Direct-connect)"),
-+ QMAN_HWE_TXT(IDFI, "Invalid Dequeue FQ"),
-+ QMAN_HWE_TXT(IDSI, "Invalid Dequeue Source"),
-+ QMAN_HWE_TXT(IDQI, "Invalid Dequeue Queue"),
-+ QMAN_HWE_TXT(IECE, "Invalid Enqueue Configuration"),
-+ QMAN_HWE_TXT(IEOI, "Invalid Enqueue Overflow"),
-+ QMAN_HWE_TXT(IESI, "Invalid Enqueue State"),
-+ QMAN_HWE_TXT(IECI, "Invalid Enqueue Channel"),
-+ QMAN_HWE_TXT(IEQI, "Invalid Enqueue Queue")
-+};
-+#define QMAN_HWE_COUNT (sizeof(qman_hwerr_txts)/sizeof(struct qman_hwerr_txt))
-+
-+struct qman_error_info_mdata {
-+ u16 addr_mask;
-+ u16 bits;
-+ const char *txt;
-+};
-+
-+#define QMAN_ERR_MDATA(a, b, c) { .addr_mask = a, .bits = b, .txt = c}
-+static const struct qman_error_info_mdata error_mdata[] = {
-+ QMAN_ERR_MDATA(0x01FF, 24, "FQD cache tag memory 0"),
-+ QMAN_ERR_MDATA(0x01FF, 24, "FQD cache tag memory 1"),
-+ QMAN_ERR_MDATA(0x01FF, 24, "FQD cache tag memory 2"),
-+ QMAN_ERR_MDATA(0x01FF, 24, "FQD cache tag memory 3"),
-+ QMAN_ERR_MDATA(0x0FFF, 512, "FQD cache memory"),
-+ QMAN_ERR_MDATA(0x07FF, 128, "SFDR memory"),
-+ QMAN_ERR_MDATA(0x01FF, 72, "WQ context memory"),
-+ QMAN_ERR_MDATA(0x00FF, 240, "CGR memory"),
-+ QMAN_ERR_MDATA(0x00FF, 302, "Internal Order Restoration List memory"),
-+ QMAN_ERR_MDATA(0x01FF, 256, "SW portal ring memory"),
-+ QMAN_ERR_MDATA(0x07FF, 181, "CEETM class queue descriptor memory"),
-+ QMAN_ERR_MDATA(0x0FFF, 140, "CEETM extended SFDR memory"),
-+ QMAN_ERR_MDATA(0x0FFF, 25, "CEETM logical FQ mapping memory"),
-+ QMAN_ERR_MDATA(0x0FFF, 96, "CEETM dequeue context memory"),
-+ QMAN_ERR_MDATA(0x07FF, 396, "CEETM ccgr memory"),
-+ QMAN_ERR_MDATA(0x00FF, 146, "CEETM CQ channel shaping memory"),
-+ QMAN_ERR_MDATA(0x007F, 256, "CEETM CQ channel scheduling memory"),
-+ QMAN_ERR_MDATA(0x01FF, 88, "CEETM dequeue statistics memory"),
-+};
-+#define QMAN_ERR_MDATA_COUNT \
-+ (sizeof(error_mdata)/sizeof(struct qman_error_info_mdata))
-+
-+/* Add this in Kconfig */
-+#define QMAN_ERRS_TO_UNENABLE (QM_EIRQ_PLWI | QM_EIRQ_PEBI)
-+
-+/**
-+ * qm_err_isr_<reg>_<verb> - Manipulate global interrupt registers
-+ * @v: for accessors that write values, this is the 32-bit value
-+ *
-+ * Manipulates QMAN_ERR_ISR, QMAN_ERR_IER, QMAN_ERR_ISDR, QMAN_ERR_IIR. All
-+ * manipulations except qm_err_isr_[un]inhibit() use 32-bit masks composed of
-+ * the QM_EIRQ_*** definitions. Note that "qm_err_isr_enable_write" means
-+ * "write the enable register" rather than "enable the write register"!
-+ */
-+#define qm_err_isr_status_read(qm) \
-+ __qm_err_isr_read(qm, qm_isr_status)
-+#define qm_err_isr_status_clear(qm, m) \
-+ __qm_err_isr_write(qm, qm_isr_status, m)
-+#define qm_err_isr_enable_read(qm) \
-+ __qm_err_isr_read(qm, qm_isr_enable)
-+#define qm_err_isr_enable_write(qm, v) \
-+ __qm_err_isr_write(qm, qm_isr_enable, v)
-+#define qm_err_isr_disable_read(qm) \
-+ __qm_err_isr_read(qm, qm_isr_disable)
-+#define qm_err_isr_disable_write(qm, v) \
-+ __qm_err_isr_write(qm, qm_isr_disable, v)
-+#define qm_err_isr_inhibit(qm) \
-+ __qm_err_isr_write(qm, qm_isr_inhibit, 1)
-+#define qm_err_isr_uninhibit(qm) \
-+ __qm_err_isr_write(qm, qm_isr_inhibit, 0)
-+
-+/*
-+ * TODO: unimplemented registers
-+ *
-+ * Keeping a list here of Qman registers I have not yet covered;
-+ * QCSP_DD_IHRSR, QCSP_DD_IHRFR, QCSP_DD_HASR,
-+ * DCP_DD_IHRSR, DCP_DD_IHRFR, DCP_DD_HASR, CM_CFG,
-+ * QMAN_EECC, QMAN_SBET, QMAN_EINJ, QMAN_SBEC0-12
-+ */
-+
-+/* Encapsulate "struct qman *" as a cast of the register space address. */
-+
-+static struct qman *qm_create(void *regs)
-+{
-+ return (struct qman *)regs;
-+}
-+
-+static inline u32 __qm_in(struct qman *qm, u32 offset)
-+{
-+ return in_be32((void *)qm + offset);
-+}
-+static inline void __qm_out(struct qman *qm, u32 offset, u32 val)
-+{
-+ out_be32((void *)qm + offset, val);
-+}
-+#define qm_in(reg) __qm_in(qm, REG_##reg)
-+#define qm_out(reg, val) __qm_out(qm, REG_##reg, val)
-+
-+static u32 __qm_err_isr_read(struct qman *qm, enum qm_isr_reg n)
-+{
-+ return __qm_in(qm, REG_ERR_ISR + (n << 2));
-+}
-+
-+static void __qm_err_isr_write(struct qman *qm, enum qm_isr_reg n, u32 val)
-+{
-+ __qm_out(qm, REG_ERR_ISR + (n << 2), val);
-+}
-+
-+static void qm_set_dc(struct qman *qm, enum qm_dc_portal portal,
-+ int ed, u8 sernd)
-+{
-+ DPA_ASSERT(!ed || (portal == qm_dc_portal_fman0) ||
-+ (portal == qm_dc_portal_fman1));
-+ if ((qman_ip_rev & 0xFF00) >= QMAN_REV30)
-+ qm_out(DCP_CFG(portal), (ed ? 0x1000 : 0) | (sernd & 0x3ff));
-+ else
-+ qm_out(DCP_CFG(portal), (ed ? 0x100 : 0) | (sernd & 0x1f));
-+}
-+
-+static void qm_set_wq_scheduling(struct qman *qm, enum qm_wq_class wq_class,
-+ u8 cs_elev, u8 csw2, u8 csw3, u8 csw4, u8 csw5,
-+ u8 csw6, u8 csw7)
-+{
-+ qm_out(WQ_CS_CFG(wq_class), ((cs_elev & 0xff) << 24) |
-+ ((csw2 & 0x7) << 20) | ((csw3 & 0x7) << 16) |
-+ ((csw4 & 0x7) << 12) | ((csw5 & 0x7) << 8) |
-+ ((csw6 & 0x7) << 4) | (csw7 & 0x7));
-+}
-+
-+static void qm_set_hid(struct qman *qm)
-+{
-+ qm_out(HID_CFG, 0);
-+}
-+
-+static void qm_set_corenet_initiator(struct qman *qm)
-+{
-+ qm_out(CI_SCHED_CFG,
-+ 0x80000000 | /* write srcciv enable */
-+ (CONFIG_FSL_QMAN_CI_SCHED_CFG_SRCCIV << 24) |
-+ (CONFIG_FSL_QMAN_CI_SCHED_CFG_SRQ_W << 8) |
-+ (CONFIG_FSL_QMAN_CI_SCHED_CFG_RW_W << 4) |
-+ CONFIG_FSL_QMAN_CI_SCHED_CFG_BMAN_W);
-+}
-+
-+static void qm_get_version(struct qman *qm, u16 *id, u8 *major, u8 *minor,
-+ u8 *cfg)
-+{
-+ u32 v = qm_in(IP_REV_1);
-+ u32 v2 = qm_in(IP_REV_2);
-+ *id = (v >> 16);
-+ *major = (v >> 8) & 0xff;
-+ *minor = v & 0xff;
-+ *cfg = v2 & 0xff;
-+}
-+
-+static void qm_set_memory(struct qman *qm, enum qm_memory memory, u64 ba,
-+ int enable, int prio, int stash, u32 size)
-+{
-+ u32 offset = (memory == qm_memory_fqd) ? REG_FQD_BARE : REG_PFDR_BARE;
-+ u32 exp = ilog2(size);
-+ /* choke if size isn't within range */
-+ DPA_ASSERT((size >= 4096) && (size <= 1073741824) &&
-+ is_power_of_2(size));
-+ /* choke if 'ba' has lower-alignment than 'size' */
-+ DPA_ASSERT(!(ba & (size - 1)));
-+ __qm_out(qm, offset, upper_32_bits(ba));
-+ __qm_out(qm, offset + REG_offset_BAR, lower_32_bits(ba));
-+ __qm_out(qm, offset + REG_offset_AR,
-+ (enable ? 0x80000000 : 0) |
-+ (prio ? 0x40000000 : 0) |
-+ (stash ? 0x20000000 : 0) |
-+ (exp - 1));
-+}
-+
-+static void qm_set_pfdr_threshold(struct qman *qm, u32 th, u8 k)
-+{
-+ qm_out(PFDR_FP_LWIT, th & 0xffffff);
-+ qm_out(PFDR_CFG, k);
-+}
-+
-+static void qm_set_sfdr_threshold(struct qman *qm, u16 th)
-+{
-+ qm_out(SFDR_CFG, th & 0x3ff);
-+}
-+
-+static int qm_init_pfdr(struct qman *qm, u32 pfdr_start, u32 num)
-+{
-+ u8 rslt = MCR_get_rslt(qm_in(MCR));
-+
-+ DPA_ASSERT(pfdr_start && !(pfdr_start & 7) && !(num & 7) && num);
-+ /* Make sure the command interface is 'idle' */
-+ if (!MCR_rslt_idle(rslt))
-+ panic("QMAN_MCR isn't idle");
-+
-+ /* Write the MCR command params then the verb */
-+ qm_out(MCP(0), pfdr_start);
-+ /* TODO: remove this - it's a workaround for a model bug that is
-+ * corrected in more recent versions. We use the workaround until
-+ * everyone has upgraded. */
-+ qm_out(MCP(1), (pfdr_start + num - 16));
-+ lwsync();
-+ qm_out(MCR, MCR_INIT_PFDR);
-+ /* Poll for the result */
-+ do {
-+ rslt = MCR_get_rslt(qm_in(MCR));
-+ } while (!MCR_rslt_idle(rslt));
-+ if (MCR_rslt_ok(rslt))
-+ return 0;
-+ if (MCR_rslt_eaccess(rslt))
-+ return -EACCES;
-+ if (MCR_rslt_inval(rslt))
-+ return -EINVAL;
-+ pr_crit("Unexpected result from MCR_INIT_PFDR: %02x\n", rslt);
-+ return -ENOSYS;
-+}
-+
-+/*****************/
-+/* Config driver */
-+/*****************/
-+
-+#define DEFAULT_FQD_SZ (PAGE_SIZE << CONFIG_FSL_QMAN_FQD_SZ)
-+#define DEFAULT_PFDR_SZ (PAGE_SIZE << CONFIG_FSL_QMAN_PFDR_SZ)
-+
-+/* We support only one of these */
-+static struct qman *qm;
-+static struct device_node *qm_node;
-+
-+/* And this state belongs to 'qm'. It is set during fsl_qman_init(), but used
-+ * during qman_init_ccsr(). */
-+static dma_addr_t fqd_a, pfdr_a;
-+static size_t fqd_sz = DEFAULT_FQD_SZ, pfdr_sz = DEFAULT_PFDR_SZ;
-+
-+static int qman_fqd(struct reserved_mem *rmem)
-+{
-+ fqd_a = rmem->base;
-+ fqd_sz = rmem->size;
-+
-+ WARN_ON(!(fqd_a && fqd_sz));
-+
-+ return 0;
-+}
-+RESERVEDMEM_OF_DECLARE(qman_fqd, "fsl,qman-fqd", qman_fqd);
-+
-+static int qman_pfdr(struct reserved_mem *rmem)
-+{
-+ pfdr_a = rmem->base;
-+ pfdr_sz = rmem->size;
-+
-+ WARN_ON(!(pfdr_a && pfdr_sz));
-+
-+ return 0;
-+}
-+RESERVEDMEM_OF_DECLARE(qman_fbpr, "fsl,qman-pfdr", qman_pfdr);
-+
-+size_t get_qman_fqd_size()
-+{
-+ return fqd_sz;
-+}
-+
-+/* Parse the <name> property to extract the memory location and size and
-+ * memblock_reserve() it. If it isn't supplied, memblock_alloc() the default
-+ * size. Also flush this memory range from data cache so that QMAN originated
-+ * transactions for this memory region could be marked non-coherent.
-+ */
-+static __init int parse_mem_property(struct device_node *node, const char *name,
-+ dma_addr_t *addr, size_t *sz, int zero)
-+{
-+ int ret;
-+
-+ /* If using a "zero-pma", don't try to zero it, even if you asked */
-+ if (zero && of_find_property(node, "zero-pma", &ret)) {
-+ pr_info(" it's a 'zero-pma', not zeroing from s/w\n");
-+ zero = 0;
-+ }
-+
-+ if (zero) {
-+ /* map as cacheable, non-guarded */
-+#if defined(CONFIG_ARM) || defined(CONFIG_ARM64)
-+ void __iomem *tmpp = ioremap_cache(*addr, *sz);
-+#else
-+ void __iomem *tmpp = ioremap(*addr, *sz);
-+#endif
-+
-+ if (!tmpp)
-+ return -ENOMEM;
-+ memset_io(tmpp, 0, *sz);
-+ flush_dcache_range((unsigned long)tmpp,
-+ (unsigned long)tmpp + *sz);
-+ iounmap(tmpp);
-+ }
-+
-+ return 0;
-+}
-+
-+/* TODO:
-+ * - there is obviously no handling of errors,
-+ * - the calls to qm_set_memory() hard-code the priority and CPC-stashing for
-+ * both memory resources to zero.
-+ */
-+static int __init fsl_qman_init(struct device_node *node)
-+{
-+ struct resource res;
-+ resource_size_t len;
-+ u32 __iomem *regs;
-+ const char *s;
-+ int ret, standby = 0;
-+ u16 id;
-+ u8 major, minor, cfg;
-+ ret = of_address_to_resource(node, 0, &res);
-+ if (ret) {
-+ pr_err("Can't get %s property '%s'\n", node->full_name, "reg");
-+ return ret;
-+ }
-+ s = of_get_property(node, "fsl,hv-claimable", &ret);
-+ if (s && !strcmp(s, "standby"))
-+ standby = 1;
-+ if (!standby) {
-+ ret = parse_mem_property(node, "fsl,qman-fqd",
-+ &fqd_a, &fqd_sz, 1);
-+ pr_info("qman-fqd addr %pad size 0x%zx\n", &fqd_a, fqd_sz);
-+ BUG_ON(ret);
-+ ret = parse_mem_property(node, "fsl,qman-pfdr",
-+ &pfdr_a, &pfdr_sz, 0);
-+ pr_info("qman-pfdr addr %pad size 0x%zx\n", &pfdr_a, pfdr_sz);
-+ BUG_ON(ret);
-+ }
-+ /* Global configuration */
-+ len = resource_size(&res);
-+ if (len != (unsigned long)len)
-+ return -EINVAL;
-+ regs = ioremap(res.start, (unsigned long)len);
-+ qm = qm_create(regs);
-+ qm_node = node;
-+ qm_get_version(qm, &id, &major, &minor, &cfg);
-+ pr_info("Qman ver:%04x,%02x,%02x,%02x\n", id, major, minor, cfg);
-+ if (!qman_ip_rev) {
-+ if ((major == 1) && (minor == 0)) {
-+ pr_err("QMAN rev1.0 on P4080 rev1 is not supported!\n");
-+ iounmap(regs);
-+ return -ENODEV;
-+ } else if ((major == 1) && (minor == 1))
-+ qman_ip_rev = QMAN_REV11;
-+ else if ((major == 1) && (minor == 2))
-+ qman_ip_rev = QMAN_REV12;
-+ else if ((major == 2) && (minor == 0))
-+ qman_ip_rev = QMAN_REV20;
-+ else if ((major == 3) && (minor == 0))
-+ qman_ip_rev = QMAN_REV30;
-+ else if ((major == 3) && (minor == 1))
-+ qman_ip_rev = QMAN_REV31;
-+ else if ((major == 3) && (minor == 2))
-+ qman_ip_rev = QMAN_REV32;
-+ else {
-+ pr_warn("unknown Qman version, default to rev1.1\n");
-+ qman_ip_rev = QMAN_REV11;
-+ }
-+ qman_ip_cfg = cfg;
-+ }
-+
-+ if (standby) {
-+ pr_info(" -> in standby mode\n");
-+ return 0;
-+ }
-+ return 0;
-+}
-+
-+int qman_have_ccsr(void)
-+{
-+ return qm ? 1 : 0;
-+}
-+
-+__init int qman_init_early(void)
-+{
-+ struct device_node *dn;
-+ int ret;
-+
-+ for_each_compatible_node(dn, NULL, "fsl,qman") {
-+ if (qm)
-+ pr_err("%s: only one 'fsl,qman' allowed\n",
-+ dn->full_name);
-+ else {
-+ if (!of_device_is_available(dn))
-+ continue;
-+
-+ ret = fsl_qman_init(dn);
-+ BUG_ON(ret);
-+ }
-+ }
-+ return 0;
-+}
-+postcore_initcall_sync(qman_init_early);
-+
-+static void log_edata_bits(u32 bit_count)
-+{
-+ u32 i, j, mask = 0xffffffff;
-+
-+ pr_warn("Qman ErrInt, EDATA:\n");
-+ i = bit_count/32;
-+ if (bit_count%32) {
-+ i++;
-+ mask = ~(mask << bit_count%32);
-+ }
-+ j = 16-i;
-+ pr_warn(" 0x%08x\n", qm_in(EDATA(j)) & mask);
-+ j++;
-+ for (; j < 16; j++)
-+ pr_warn(" 0x%08x\n", qm_in(EDATA(j)));
-+}
-+
-+static void log_additional_error_info(u32 isr_val, u32 ecsr_val)
-+{
-+ union qman_ecir ecir_val;
-+ union qman_eadr eadr_val;
-+
-+ ecir_val.ecir_raw = qm_in(ECIR);
-+ /* Is portal info valid */
-+ if ((qman_ip_rev & 0xFF00) >= QMAN_REV30) {
-+ union qman_ecir2 ecir2_val;
-+ ecir2_val.ecir2_raw = qm_in(ECIR2);
-+ if (ecsr_val & PORTAL_ECSR_ERR) {
-+ pr_warn("Qman ErrInt: %s id %d\n",
-+ (ecir2_val.info.portal_type) ?
-+ "DCP" : "SWP", ecir2_val.info.portal_num);
-+ }
-+ if (ecsr_val & (FQID_ECSR_ERR | QM_EIRQ_IECE)) {
-+ pr_warn("Qman ErrInt: ecir.fqid 0x%x\n",
-+ ecir_val.info.fqid);
-+ }
-+ if (ecsr_val & (QM_EIRQ_SBEI|QM_EIRQ_MBEI)) {
-+ eadr_val.eadr_raw = qm_in(EADR);
-+ pr_warn("Qman ErrInt: EADR Memory: %s, 0x%x\n",
-+ error_mdata[eadr_val.info_rev3.memid].txt,
-+ error_mdata[eadr_val.info_rev3.memid].addr_mask
-+ & eadr_val.info_rev3.eadr);
-+ log_edata_bits(
-+ error_mdata[eadr_val.info_rev3.memid].bits);
-+ }
-+ } else {
-+ if (ecsr_val & PORTAL_ECSR_ERR) {
-+ pr_warn("Qman ErrInt: %s id %d\n",
-+ (ecir_val.info.portal_type) ?
-+ "DCP" : "SWP", ecir_val.info.portal_num);
-+ }
-+ if (ecsr_val & FQID_ECSR_ERR) {
-+ pr_warn("Qman ErrInt: ecir.fqid 0x%x\n",
-+ ecir_val.info.fqid);
-+ }
-+ if (ecsr_val & (QM_EIRQ_SBEI|QM_EIRQ_MBEI)) {
-+ eadr_val.eadr_raw = qm_in(EADR);
-+ pr_warn("Qman ErrInt: EADR Memory: %s, 0x%x\n",
-+ error_mdata[eadr_val.info.memid].txt,
-+ error_mdata[eadr_val.info.memid].addr_mask
-+ & eadr_val.info.eadr);
-+ log_edata_bits(error_mdata[eadr_val.info.memid].bits);
-+ }
-+ }
-+}
-+
-+/* Qman interrupt handler */
-+static irqreturn_t qman_isr(int irq, void *ptr)
-+{
-+ u32 isr_val, ier_val, ecsr_val, isr_mask, i;
-+
-+ ier_val = qm_err_isr_enable_read(qm);
-+ isr_val = qm_err_isr_status_read(qm);
-+ ecsr_val = qm_in(ECSR);
-+ isr_mask = isr_val & ier_val;
-+
-+ if (!isr_mask)
-+ return IRQ_NONE;
-+ for (i = 0; i < QMAN_HWE_COUNT; i++) {
-+ if (qman_hwerr_txts[i].mask & isr_mask) {
-+ pr_warn("Qman ErrInt: %s\n", qman_hwerr_txts[i].txt);
-+ if (qman_hwerr_txts[i].mask & ecsr_val) {
-+ log_additional_error_info(isr_mask, ecsr_val);
-+ /* Re-arm error capture registers */
-+ qm_out(ECSR, ecsr_val);
-+ }
-+ if (qman_hwerr_txts[i].mask & QMAN_ERRS_TO_UNENABLE) {
-+ pr_devel("Qman un-enabling error 0x%x\n",
-+ qman_hwerr_txts[i].mask);
-+ ier_val &= ~qman_hwerr_txts[i].mask;
-+ qm_err_isr_enable_write(qm, ier_val);
-+ }
-+ }
-+ }
-+ qm_err_isr_status_clear(qm, isr_val);
-+ return IRQ_HANDLED;
-+}
-+
-+static int __bind_irq(void)
-+{
-+ int ret, err_irq;
-+
-+ err_irq = of_irq_to_resource(qm_node, 0, NULL);
-+ if (err_irq == 0) {
-+ pr_info("Can't get %s property '%s'\n", qm_node->full_name,
-+ "interrupts");
-+ return -ENODEV;
-+ }
-+ ret = request_irq(err_irq, qman_isr, IRQF_SHARED, "qman-err", qm_node);
-+ if (ret) {
-+ pr_err("request_irq() failed %d for '%s'\n", ret,
-+ qm_node->full_name);
-+ return -ENODEV;
-+ }
-+ /* Write-to-clear any stale bits, (eg. starvation being asserted prior
-+ * to resource allocation during driver init). */
-+ qm_err_isr_status_clear(qm, 0xffffffff);
-+ /* Enable Error Interrupts */
-+ qm_err_isr_enable_write(qm, 0xffffffff);
-+ return 0;
-+}
-+
-+int qman_init_ccsr(struct device_node *node)
-+{
-+ int ret;
-+ if (!qman_have_ccsr())
-+ return 0;
-+ if (node != qm_node)
-+ return -EINVAL;
-+#if defined(CONFIG_ARM) || defined(CONFIG_ARM64)
-+ /* TEMP for LS1043 : should be done in uboot */
-+ qm_out(QCSP_BARE, 0x5);
-+ qm_out(QCSP_BAR, 0x0);
-+#endif
-+ /* FQD memory */
-+ qm_set_memory(qm, qm_memory_fqd, fqd_a, 1, 0, 0, fqd_sz);
-+ /* PFDR memory */
-+ qm_set_memory(qm, qm_memory_pfdr, pfdr_a, 1, 0, 0, pfdr_sz);
-+ qm_init_pfdr(qm, 8, pfdr_sz / 64 - 8);
-+ /* thresholds */
-+ qm_set_pfdr_threshold(qm, 512, 64);
-+ qm_set_sfdr_threshold(qm, 128);
-+ /* clear stale PEBI bit from interrupt status register */
-+ qm_err_isr_status_clear(qm, QM_EIRQ_PEBI);
-+ /* corenet initiator settings */
-+ qm_set_corenet_initiator(qm);
-+ /* HID settings */
-+ qm_set_hid(qm);
-+ /* Set scheduling weights to defaults */
-+ for (ret = qm_wq_first; ret <= qm_wq_last; ret++)
-+ qm_set_wq_scheduling(qm, ret, 0, 0, 0, 0, 0, 0, 0);
-+ /* We are not prepared to accept ERNs for hardware enqueues */
-+ qm_set_dc(qm, qm_dc_portal_fman0, 1, 0);
-+ qm_set_dc(qm, qm_dc_portal_fman1, 1, 0);
-+ /* Initialise Error Interrupt Handler */
-+ ret = __bind_irq();
-+ if (ret)
-+ return ret;
-+ return 0;
-+}
-+
-+#define LIO_CFG_LIODN_MASK 0x0fff0000
-+void qman_liodn_fixup(u16 channel)
-+{
-+ static int done;
-+ static u32 liodn_offset;
-+ u32 before, after;
-+ int idx = channel - QM_CHANNEL_SWPORTAL0;
-+
-+ if (!qman_have_ccsr())
-+ return;
-+ if ((qman_ip_rev & 0xFF00) >= QMAN_REV30)
-+ before = qm_in(REV3_QCSP_LIO_CFG(idx));
-+ else
-+ before = qm_in(QCSP_LIO_CFG(idx));
-+ if (!done) {
-+ liodn_offset = before & LIO_CFG_LIODN_MASK;
-+ done = 1;
-+ return;
-+ }
-+ after = (before & (~LIO_CFG_LIODN_MASK)) | liodn_offset;
-+ if ((qman_ip_rev & 0xFF00) >= QMAN_REV30)
-+ qm_out(REV3_QCSP_LIO_CFG(idx), after);
-+ else
-+ qm_out(QCSP_LIO_CFG(idx), after);
-+}
-+
-+#define IO_CFG_SDEST_MASK 0x00ff0000
-+int qman_set_sdest(u16 channel, unsigned int cpu_idx)
-+{
-+ int idx = channel - QM_CHANNEL_SWPORTAL0;
-+ u32 before, after;
-+
-+ if (!qman_have_ccsr())
-+ return -ENODEV;
-+ if ((qman_ip_rev & 0xFF00) == QMAN_REV31) {
-+ /* LS1043A - only one L2 cache */
-+ cpu_idx = 0;
-+ }
-+
-+ if ((qman_ip_rev & 0xFF00) >= QMAN_REV30) {
-+ before = qm_in(REV3_QCSP_IO_CFG(idx));
-+ /* Each pair of vcpu share the same SRQ(SDEST) */
-+ cpu_idx /= 2;
-+ after = (before & (~IO_CFG_SDEST_MASK)) | (cpu_idx << 16);
-+ qm_out(REV3_QCSP_IO_CFG(idx), after);
-+ } else {
-+ before = qm_in(QCSP_IO_CFG(idx));
-+ after = (before & (~IO_CFG_SDEST_MASK)) | (cpu_idx << 16);
-+ qm_out(QCSP_IO_CFG(idx), after);
-+ }
-+ return 0;
-+}
-+
-+#define MISC_CFG_WPM_MASK 0x00000002
-+int qm_set_wpm(int wpm)
-+{
-+ u32 before;
-+ u32 after;
-+
-+ if (!qman_have_ccsr())
-+ return -ENODEV;
-+
-+ before = qm_in(MISC_CFG);
-+ after = (before & (~MISC_CFG_WPM_MASK)) | (wpm << 1);
-+ qm_out(MISC_CFG, after);
-+ return 0;
-+}
-+
-+int qm_get_wpm(int *wpm)
-+{
-+ u32 before;
-+
-+ if (!qman_have_ccsr())
-+ return -ENODEV;
-+
-+ before = qm_in(MISC_CFG);
-+ *wpm = (before & MISC_CFG_WPM_MASK) >> 1;
-+ return 0;
-+}
-+
-+/* CEETM_CFG_PRES register has PRES field which is calculated by:
-+ * PRES = (2^22 / credit update reference period) * QMan clock period
-+ * = (2^22 * 10^9)/ CONFIG_QMAN_CEETM_UPDATE_PERIOD) / qman_clk
-+ */
-+
-+int qman_ceetm_set_prescaler(enum qm_dc_portal portal)
-+{
-+ u64 temp;
-+ u16 pres;
-+
-+ if (!qman_have_ccsr())
-+ return -ENODEV;
-+
-+ temp = 0x400000 * 100;
-+ do_div(temp, CONFIG_QMAN_CEETM_UPDATE_PERIOD);
-+ temp *= 10000000;
-+ do_div(temp, qman_clk);
-+ pres = (u16) temp;
-+ qm_out(CEETM_CFG_IDX, portal);
-+ qm_out(CEETM_CFG_PRES, pres);
-+ return 0;
-+}
-+
-+int qman_ceetm_get_prescaler(u16 *pres)
-+{
-+ if (!qman_have_ccsr())
-+ return -ENODEV;
-+ *pres = (u16)qm_in(CEETM_CFG_PRES);
-+ return 0;
-+}
-+
-+#define DCP_CFG_CEETME_MASK 0xFFFF0000
-+#define QM_SP_ENABLE_CEETM(n) (0x80000000 >> (n))
-+int qman_sp_enable_ceetm_mode(enum qm_dc_portal portal, u16 sub_portal)
-+{
-+ u32 dcp_cfg;
-+
-+ if (!qman_have_ccsr())
-+ return -ENODEV;
-+
-+ dcp_cfg = qm_in(DCP_CFG(portal));
-+ dcp_cfg |= QM_SP_ENABLE_CEETM(sub_portal);
-+ qm_out(DCP_CFG(portal), dcp_cfg);
-+ return 0;
-+}
-+
-+int qman_sp_disable_ceetm_mode(enum qm_dc_portal portal, u16 sub_portal)
-+{
-+ u32 dcp_cfg;
-+
-+ if (!qman_have_ccsr())
-+ return -ENODEV;
-+ dcp_cfg = qm_in(DCP_CFG(portal));
-+ dcp_cfg &= ~(QM_SP_ENABLE_CEETM(sub_portal));
-+ qm_out(DCP_CFG(portal), dcp_cfg);
-+ return 0;
-+}
-+
-+int qman_ceetm_get_xsfdr(enum qm_dc_portal portal, unsigned int *num)
-+{
-+ if (!qman_have_ccsr())
-+ return -ENODEV;
-+ *num = qm_in(CEETM_XSFDR_IN_USE);
-+ return 0;
-+}
-+EXPORT_SYMBOL(qman_ceetm_get_xsfdr);
-+
-+#ifdef CONFIG_SYSFS
-+
-+#define DRV_NAME "fsl-qman"
-+#define DCP_MAX_ID 3
-+#define DCP_MIN_ID 0
-+
-+static ssize_t show_pfdr_fpc(struct device *dev,
-+ struct device_attribute *dev_attr, char *buf)
-+{
-+ return snprintf(buf, PAGE_SIZE, "%u\n", qm_in(PFDR_FPC));
-+};
-+
-+static ssize_t show_dlm_avg(struct device *dev,
-+ struct device_attribute *dev_attr, char *buf)
-+{
-+ u32 data;
-+ int i;
-+
-+ if (!sscanf(dev_attr->attr.name, "dcp%d_dlm_avg", &i))
-+ return -EINVAL;
-+ if (i < DCP_MIN_ID || i > DCP_MAX_ID)
-+ return -EINVAL;
-+ data = qm_in(DCP_DLM_AVG(i));
-+ return snprintf(buf, PAGE_SIZE, "%d.%08d\n", data>>8,
-+ (data & 0x000000ff)*390625);
-+};
-+
-+static ssize_t set_dlm_avg(struct device *dev,
-+ struct device_attribute *dev_attr, const char *buf, size_t count)
-+{
-+ unsigned long val;
-+ int i;
-+
-+ if (!sscanf(dev_attr->attr.name, "dcp%d_dlm_avg", &i))
-+ return -EINVAL;
-+ if (i < DCP_MIN_ID || i > DCP_MAX_ID)
-+ return -EINVAL;
-+ if (kstrtoul(buf, 0, &val)) {
-+ dev_dbg(dev, "invalid input %s\n", buf);
-+ return -EINVAL;
-+ }
-+ qm_out(DCP_DLM_AVG(i), val);
-+ return count;
-+};
-+
-+static ssize_t show_pfdr_cfg(struct device *dev,
-+ struct device_attribute *dev_attr, char *buf)
-+{
-+ return snprintf(buf, PAGE_SIZE, "%u\n", qm_in(PFDR_CFG));
-+};
-+
-+static ssize_t set_pfdr_cfg(struct device *dev,
-+ struct device_attribute *dev_attr, const char *buf, size_t count)
-+{
-+ unsigned long val;
-+
-+ if (kstrtoul(buf, 0, &val)) {
-+ dev_dbg(dev, "invalid input %s\n", buf);
-+ return -EINVAL;
-+ }
-+ qm_out(PFDR_CFG, val);
-+ return count;
-+};
-+
-+static ssize_t show_sfdr_in_use(struct device *dev,
-+ struct device_attribute *dev_attr, char *buf)
-+{
-+ return snprintf(buf, PAGE_SIZE, "%u\n", qm_in(SFDR_IN_USE));
-+};
-+
-+static ssize_t show_idle_stat(struct device *dev,
-+ struct device_attribute *dev_attr, char *buf)
-+{
-+ return snprintf(buf, PAGE_SIZE, "%u\n", qm_in(IDLE_STAT));
-+};
-+
-+static ssize_t show_ci_rlm_avg(struct device *dev,
-+ struct device_attribute *dev_attr, char *buf)
-+{
-+ u32 data = qm_in(CI_RLM_AVG);
-+ return snprintf(buf, PAGE_SIZE, "%d.%08d\n", data>>8,
-+ (data & 0x000000ff)*390625);
-+};
-+
-+static ssize_t set_ci_rlm_avg(struct device *dev,
-+ struct device_attribute *dev_attr, const char *buf, size_t count)
-+{
-+ unsigned long val;
-+
-+ if (kstrtoul(buf, 0, &val)) {
-+ dev_dbg(dev, "invalid input %s\n", buf);
-+ return -EINVAL;
-+ }
-+ qm_out(CI_RLM_AVG, val);
-+ return count;
-+};
-+
-+static ssize_t show_err_isr(struct device *dev,
-+ struct device_attribute *dev_attr, char *buf)
-+{
-+ return snprintf(buf, PAGE_SIZE, "0x%08x\n", qm_in(ERR_ISR));
-+};
-+
-+#define SBEC_MAX_ID 14
-+#define SBEC_MIN_ID 0
-+
-+static ssize_t show_sbec(struct device *dev,
-+ struct device_attribute *dev_attr, char *buf)
-+{
-+ int i;
-+
-+ if (!sscanf(dev_attr->attr.name, "sbec_%d", &i))
-+ return -EINVAL;
-+ if (i < SBEC_MIN_ID || i > SBEC_MAX_ID)
-+ return -EINVAL;
-+ return snprintf(buf, PAGE_SIZE, "%u\n", qm_in(SBEC(i)));
-+};
-+
-+static DEVICE_ATTR(pfdr_fpc, S_IRUSR, show_pfdr_fpc, NULL);
-+static DEVICE_ATTR(pfdr_cfg, S_IRUSR, show_pfdr_cfg, set_pfdr_cfg);
-+static DEVICE_ATTR(idle_stat, S_IRUSR, show_idle_stat, NULL);
-+static DEVICE_ATTR(ci_rlm_avg, (S_IRUSR|S_IWUSR),
-+ show_ci_rlm_avg, set_ci_rlm_avg);
-+static DEVICE_ATTR(err_isr, S_IRUSR, show_err_isr, NULL);
-+static DEVICE_ATTR(sfdr_in_use, S_IRUSR, show_sfdr_in_use, NULL);
-+
-+static DEVICE_ATTR(dcp0_dlm_avg, (S_IRUSR|S_IWUSR), show_dlm_avg, set_dlm_avg);
-+static DEVICE_ATTR(dcp1_dlm_avg, (S_IRUSR|S_IWUSR), show_dlm_avg, set_dlm_avg);
-+static DEVICE_ATTR(dcp2_dlm_avg, (S_IRUSR|S_IWUSR), show_dlm_avg, set_dlm_avg);
-+static DEVICE_ATTR(dcp3_dlm_avg, (S_IRUSR|S_IWUSR), show_dlm_avg, set_dlm_avg);
-+
-+static DEVICE_ATTR(sbec_0, S_IRUSR, show_sbec, NULL);
-+static DEVICE_ATTR(sbec_1, S_IRUSR, show_sbec, NULL);
-+static DEVICE_ATTR(sbec_2, S_IRUSR, show_sbec, NULL);
-+static DEVICE_ATTR(sbec_3, S_IRUSR, show_sbec, NULL);
-+static DEVICE_ATTR(sbec_4, S_IRUSR, show_sbec, NULL);
-+static DEVICE_ATTR(sbec_5, S_IRUSR, show_sbec, NULL);
-+static DEVICE_ATTR(sbec_6, S_IRUSR, show_sbec, NULL);
-+static DEVICE_ATTR(sbec_7, S_IRUSR, show_sbec, NULL);
-+static DEVICE_ATTR(sbec_8, S_IRUSR, show_sbec, NULL);
-+static DEVICE_ATTR(sbec_9, S_IRUSR, show_sbec, NULL);
-+static DEVICE_ATTR(sbec_10, S_IRUSR, show_sbec, NULL);
-+static DEVICE_ATTR(sbec_11, S_IRUSR, show_sbec, NULL);
-+static DEVICE_ATTR(sbec_12, S_IRUSR, show_sbec, NULL);
-+static DEVICE_ATTR(sbec_13, S_IRUSR, show_sbec, NULL);
-+static DEVICE_ATTR(sbec_14, S_IRUSR, show_sbec, NULL);
-+
-+static struct attribute *qman_dev_attributes[] = {
-+ &dev_attr_pfdr_fpc.attr,
-+ &dev_attr_pfdr_cfg.attr,
-+ &dev_attr_idle_stat.attr,
-+ &dev_attr_ci_rlm_avg.attr,
-+ &dev_attr_err_isr.attr,
-+ &dev_attr_dcp0_dlm_avg.attr,
-+ &dev_attr_dcp1_dlm_avg.attr,
-+ &dev_attr_dcp2_dlm_avg.attr,
-+ &dev_attr_dcp3_dlm_avg.attr,
-+ /* sfdr_in_use will be added if necessary */
-+ NULL
-+};
-+
-+static struct attribute *qman_dev_ecr_attributes[] = {
-+ &dev_attr_sbec_0.attr,
-+ &dev_attr_sbec_1.attr,
-+ &dev_attr_sbec_2.attr,
-+ &dev_attr_sbec_3.attr,
-+ &dev_attr_sbec_4.attr,
-+ &dev_attr_sbec_5.attr,
-+ &dev_attr_sbec_6.attr,
-+ &dev_attr_sbec_7.attr,
-+ &dev_attr_sbec_8.attr,
-+ &dev_attr_sbec_9.attr,
-+ &dev_attr_sbec_10.attr,
-+ &dev_attr_sbec_11.attr,
-+ &dev_attr_sbec_12.attr,
-+ &dev_attr_sbec_13.attr,
-+ &dev_attr_sbec_14.attr,
-+ NULL
-+};
-+
-+/* root level */
-+static const struct attribute_group qman_dev_attr_grp = {
-+ .name = NULL,
-+ .attrs = qman_dev_attributes
-+};
-+static const struct attribute_group qman_dev_ecr_grp = {
-+ .name = "error_capture",
-+ .attrs = qman_dev_ecr_attributes
-+};
-+
-+static int of_fsl_qman_remove(struct platform_device *ofdev)
-+{
-+ sysfs_remove_group(&ofdev->dev.kobj, &qman_dev_attr_grp);
-+ return 0;
-+};
-+
-+static int of_fsl_qman_probe(struct platform_device *ofdev)
-+{
-+ int ret;
-+
-+ ret = sysfs_create_group(&ofdev->dev.kobj, &qman_dev_attr_grp);
-+ if (ret)
-+ goto done;
-+ ret = sysfs_add_file_to_group(&ofdev->dev.kobj,
-+ &dev_attr_sfdr_in_use.attr, qman_dev_attr_grp.name);
-+ if (ret)
-+ goto del_group_0;
-+ ret = sysfs_create_group(&ofdev->dev.kobj, &qman_dev_ecr_grp);
-+ if (ret)
-+ goto del_group_0;
-+
-+ goto done;
-+
-+del_group_0:
-+ sysfs_remove_group(&ofdev->dev.kobj, &qman_dev_attr_grp);
-+done:
-+ if (ret)
-+ dev_err(&ofdev->dev,
-+ "Cannot create dev attributes ret=%d\n", ret);
-+ return ret;
-+};
-+
-+static struct of_device_id of_fsl_qman_ids[] = {
-+ {
-+ .compatible = "fsl,qman",
-+ },
-+ {}
-+};
-+MODULE_DEVICE_TABLE(of, of_fsl_qman_ids);
-+
-+#ifdef CONFIG_SUSPEND
-+
-+static u32 saved_isdr;
-+static int qman_pm_suspend_noirq(struct device *dev)
-+{
-+ uint32_t idle_state;
-+
-+ suspend_unused_qportal();
-+ /* save isdr, disable all, clear isr */
-+ saved_isdr = qm_err_isr_disable_read(qm);
-+ qm_err_isr_disable_write(qm, 0xffffffff);
-+ qm_err_isr_status_clear(qm, 0xffffffff);
-+ idle_state = qm_in(IDLE_STAT);
-+ if (!(idle_state & 0x1)) {
-+ pr_err("Qman not idle 0x%x aborting\n", idle_state);
-+ qm_err_isr_disable_write(qm, saved_isdr);
-+ resume_unused_qportal();
-+ return -EBUSY;
-+ }
-+#ifdef CONFIG_PM_DEBUG
-+ pr_info("Qman suspend code, IDLE_STAT = 0x%x\n", idle_state);
-+#endif
-+ return 0;
-+}
-+
-+static int qman_pm_resume_noirq(struct device *dev)
-+{
-+ /* restore isdr */
-+ qm_err_isr_disable_write(qm, saved_isdr);
-+ resume_unused_qportal();
-+ return 0;
-+}
-+#else
-+#define qman_pm_suspend_noirq NULL
-+#define qman_pm_resume_noirq NULL
-+#endif
-+
-+static const struct dev_pm_ops qman_pm_ops = {
-+ .suspend_noirq = qman_pm_suspend_noirq,
-+ .resume_noirq = qman_pm_resume_noirq,
-+};
-+
-+static struct platform_driver of_fsl_qman_driver = {
-+ .driver = {
-+ .owner = THIS_MODULE,
-+ .name = DRV_NAME,
-+ .of_match_table = of_fsl_qman_ids,
-+ .pm = &qman_pm_ops,
-+ },
-+ .probe = of_fsl_qman_probe,
-+ .remove = of_fsl_qman_remove,
-+};
-+
-+static int qman_ctrl_init(void)
-+{
-+ return platform_driver_register(&of_fsl_qman_driver);
-+}
-+
-+static void qman_ctrl_exit(void)
-+{
-+ platform_driver_unregister(&of_fsl_qman_driver);
-+}
-+
-+module_init(qman_ctrl_init);
-+module_exit(qman_ctrl_exit);
-+
-+#endif /* CONFIG_SYSFS */
---- /dev/null
-+++ b/drivers/staging/fsl_qbman/qman_debugfs.c
-@@ -0,0 +1,1594 @@
-+/* Copyright 2010-2011 Freescale Semiconductor, Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+#include "qman_private.h"
-+
-+#define MAX_FQID (0x00ffffff)
-+#define QM_FQD_BLOCK_SIZE 64
-+#define QM_FQD_AR (0xC10)
-+
-+static u32 fqid_max;
-+static u64 qman_ccsr_start;
-+static u64 qman_ccsr_size;
-+
-+static const char * const state_txt[] = {
-+ "Out of Service",
-+ "Retired",
-+ "Tentatively Scheduled",
-+ "Truly Scheduled",
-+ "Parked",
-+ "Active, Active Held or Held Suspended",
-+ "Unknown State 6",
-+ "Unknown State 7",
-+ NULL,
-+};
-+
-+static const u8 fqd_states[] = {
-+ QM_MCR_NP_STATE_OOS, QM_MCR_NP_STATE_RETIRED, QM_MCR_NP_STATE_TEN_SCHED,
-+ QM_MCR_NP_STATE_TRU_SCHED, QM_MCR_NP_STATE_PARKED,
-+ QM_MCR_NP_STATE_ACTIVE};
-+
-+struct mask_to_text {
-+ u16 mask;
-+ const char *txt;
-+};
-+
-+struct mask_filter_s {
-+ u16 mask;
-+ u8 filter;
-+};
-+
-+static const struct mask_filter_s mask_filter[] = {
-+ {QM_FQCTRL_PREFERINCACHE, 0},
-+ {QM_FQCTRL_PREFERINCACHE, 1},
-+ {QM_FQCTRL_HOLDACTIVE, 0},
-+ {QM_FQCTRL_HOLDACTIVE, 1},
-+ {QM_FQCTRL_AVOIDBLOCK, 0},
-+ {QM_FQCTRL_AVOIDBLOCK, 1},
-+ {QM_FQCTRL_FORCESFDR, 0},
-+ {QM_FQCTRL_FORCESFDR, 1},
-+ {QM_FQCTRL_CPCSTASH, 0},
-+ {QM_FQCTRL_CPCSTASH, 1},
-+ {QM_FQCTRL_CTXASTASHING, 0},
-+ {QM_FQCTRL_CTXASTASHING, 1},
-+ {QM_FQCTRL_ORP, 0},
-+ {QM_FQCTRL_ORP, 1},
-+ {QM_FQCTRL_TDE, 0},
-+ {QM_FQCTRL_TDE, 1},
-+ {QM_FQCTRL_CGE, 0},
-+ {QM_FQCTRL_CGE, 1}
-+};
-+
-+static const struct mask_to_text fq_ctrl_text_list[] = {
-+ {
-+ .mask = QM_FQCTRL_PREFERINCACHE,
-+ .txt = "Prefer in cache",
-+ },
-+ {
-+ .mask = QM_FQCTRL_HOLDACTIVE,
-+ .txt = "Hold active in portal",
-+ },
-+ {
-+ .mask = QM_FQCTRL_AVOIDBLOCK,
-+ .txt = "Avoid Blocking",
-+ },
-+ {
-+ .mask = QM_FQCTRL_FORCESFDR,
-+ .txt = "High-priority SFDRs",
-+ },
-+ {
-+ .mask = QM_FQCTRL_CPCSTASH,
-+ .txt = "CPC Stash Enable",
-+ },
-+ {
-+ .mask = QM_FQCTRL_CTXASTASHING,
-+ .txt = "Context-A stashing",
-+ },
-+ {
-+ .mask = QM_FQCTRL_ORP,
-+ .txt = "ORP Enable",
-+ },
-+ {
-+ .mask = QM_FQCTRL_TDE,
-+ .txt = "Tail-Drop Enable",
-+ },
-+ {
-+ .mask = QM_FQCTRL_CGE,
-+ .txt = "Congestion Group Enable",
-+ },
-+ {
-+ .mask = 0,
-+ .txt = NULL,
-+ }
-+};
-+
-+static const char *get_fqd_ctrl_text(u16 mask)
-+{
-+ int i = 0;
-+
-+ while (fq_ctrl_text_list[i].txt != NULL) {
-+ if (fq_ctrl_text_list[i].mask == mask)
-+ return fq_ctrl_text_list[i].txt;
-+ i++;
-+ }
-+ return NULL;
-+}
-+
-+static const struct mask_to_text stashing_text_list[] = {
-+ {
-+ .mask = QM_STASHING_EXCL_CTX,
-+ .txt = "FQ Ctx Stash"
-+ },
-+ {
-+ .mask = QM_STASHING_EXCL_DATA,
-+ .txt = "Frame Data Stash",
-+ },
-+ {
-+ .mask = QM_STASHING_EXCL_ANNOTATION,
-+ .txt = "Frame Annotation Stash",
-+ },
-+ {
-+ .mask = 0,
-+ .txt = NULL,
-+ },
-+};
-+
-+static int user_input_convert(const char __user *user_buf, size_t count,
-+ unsigned long *val)
-+{
-+ char buf[12];
-+
-+ if (count > sizeof(buf) - 1)
-+ return -EINVAL;
-+ if (copy_from_user(buf, user_buf, count))
-+ return -EFAULT;
-+ buf[count] = '\0';
-+ if (kstrtoul(buf, 0, val))
-+ return -EINVAL;
-+ return 0;
-+}
-+
-+struct line_buffer_fq {
-+ u32 buf[8];
-+ u32 buf_cnt;
-+ int line_cnt;
-+};
-+
-+static void add_to_line_buffer(struct line_buffer_fq *line_buf, u32 fqid,
-+ struct seq_file *file)
-+{
-+ line_buf->buf[line_buf->buf_cnt] = fqid;
-+ line_buf->buf_cnt++;
-+ if (line_buf->buf_cnt == 8) {
-+ /* Buffer is full, flush it */
-+ if (line_buf->line_cnt != 0)
-+ seq_puts(file, ",\n");
-+ seq_printf(file, "0x%06x,0x%06x,0x%06x,0x%06x,0x%06x,"
-+ "0x%06x,0x%06x,0x%06x",
-+ line_buf->buf[0], line_buf->buf[1], line_buf->buf[2],
-+ line_buf->buf[3], line_buf->buf[4], line_buf->buf[5],
-+ line_buf->buf[6], line_buf->buf[7]);
-+ line_buf->buf_cnt = 0;
-+ line_buf->line_cnt++;
-+ }
-+}
-+
-+static void flush_line_buffer(struct line_buffer_fq *line_buf,
-+ struct seq_file *file)
-+{
-+ if (line_buf->buf_cnt) {
-+ int y = 0;
-+ if (line_buf->line_cnt != 0)
-+ seq_puts(file, ",\n");
-+ while (y != line_buf->buf_cnt) {
-+ if (y+1 == line_buf->buf_cnt)
-+ seq_printf(file, "0x%06x", line_buf->buf[y]);
-+ else
-+ seq_printf(file, "0x%06x,", line_buf->buf[y]);
-+ y++;
-+ }
-+ line_buf->line_cnt++;
-+ }
-+ if (line_buf->line_cnt)
-+ seq_putc(file, '\n');
-+}
-+
-+static struct dentry *dfs_root; /* debugfs root directory */
-+
-+/*******************************************************************************
-+ * Query Frame Queue Non Programmable Fields
-+ ******************************************************************************/
-+struct query_fq_np_fields_data_s {
-+ u32 fqid;
-+};
-+static struct query_fq_np_fields_data_s query_fq_np_fields_data = {
-+ .fqid = 1,
-+};
-+
-+static int query_fq_np_fields_show(struct seq_file *file, void *offset)
-+{
-+ int ret;
-+ struct qm_mcr_queryfq_np np;
-+ struct qman_fq fq;
-+
-+ fq.fqid = query_fq_np_fields_data.fqid;
-+ ret = qman_query_fq_np(&fq, &np);
-+ if (ret)
-+ return ret;
-+ /* Print state */
-+ seq_printf(file, "Query FQ Non Programmable Fields Result fqid 0x%x\n",
-+ fq.fqid);
-+ seq_printf(file, " force eligible pending: %s\n",
-+ (np.state & QM_MCR_NP_STATE_FE) ? "yes" : "no");
-+ seq_printf(file, " retirement pending: %s\n",
-+ (np.state & QM_MCR_NP_STATE_R) ? "yes" : "no");
-+ seq_printf(file, " state: %s\n",
-+ state_txt[np.state & QM_MCR_NP_STATE_MASK]);
-+ seq_printf(file, " fq_link: 0x%x\n", np.fqd_link);
-+ seq_printf(file, " odp_seq: %u\n", np.odp_seq);
-+ seq_printf(file, " orp_nesn: %u\n", np.orp_nesn);
-+ seq_printf(file, " orp_ea_hseq: %u\n", np.orp_ea_hseq);
-+ seq_printf(file, " orp_ea_tseq: %u\n", np.orp_ea_tseq);
-+ seq_printf(file, " orp_ea_hptr: 0x%x\n", np.orp_ea_hptr);
-+ seq_printf(file, " orp_ea_tptr: 0x%x\n", np.orp_ea_tptr);
-+ seq_printf(file, " pfdr_hptr: 0x%x\n", np.pfdr_hptr);
-+ seq_printf(file, " pfdr_tptr: 0x%x\n", np.pfdr_tptr);
-+ seq_printf(file, " is: ics_surp contains a %s\n",
-+ (np.is) ? "deficit" : "surplus");
-+ seq_printf(file, " ics_surp: %u\n", np.ics_surp);
-+ seq_printf(file, " byte_cnt: %u\n", np.byte_cnt);
-+ seq_printf(file, " frm_cnt: %u\n", np.frm_cnt);
-+ seq_printf(file, " ra1_sfdr: 0x%x\n", np.ra1_sfdr);
-+ seq_printf(file, " ra2_sfdr: 0x%x\n", np.ra2_sfdr);
-+ seq_printf(file, " od1_sfdr: 0x%x\n", np.od1_sfdr);
-+ seq_printf(file, " od2_sfdr: 0x%x\n", np.od2_sfdr);
-+ seq_printf(file, " od3_sfdr: 0x%x\n", np.od3_sfdr);
-+ return 0;
-+}
-+
-+static int query_fq_np_fields_open(struct inode *inode,
-+ struct file *file)
-+{
-+ return single_open(file, query_fq_np_fields_show, NULL);
-+}
-+
-+static ssize_t query_fq_np_fields_write(struct file *f,
-+ const char __user *buf, size_t count, loff_t *off)
-+{
-+ int ret;
-+ unsigned long val;
-+
-+ ret = user_input_convert(buf, count, &val);
-+ if (ret)
-+ return ret;
-+ if (val > MAX_FQID)
-+ return -EINVAL;
-+ query_fq_np_fields_data.fqid = (u32)val;
-+ return count;
-+}
-+
-+static const struct file_operations query_fq_np_fields_fops = {
-+ .owner = THIS_MODULE,
-+ .open = query_fq_np_fields_open,
-+ .read = seq_read,
-+ .write = query_fq_np_fields_write,
-+ .release = single_release,
-+};
-+
-+/*******************************************************************************
-+ * Frame Queue Programmable Fields
-+ ******************************************************************************/
-+struct query_fq_fields_data_s {
-+ u32 fqid;
-+};
-+
-+static struct query_fq_fields_data_s query_fq_fields_data = {
-+ .fqid = 1,
-+};
-+
-+static int query_fq_fields_show(struct seq_file *file, void *offset)
-+{
-+ int ret;
-+ struct qm_fqd fqd;
-+ struct qman_fq fq;
-+ int i = 0;
-+
-+ memset(&fqd, 0, sizeof(struct qm_fqd));
-+ fq.fqid = query_fq_fields_data.fqid;
-+ ret = qman_query_fq(&fq, &fqd);
-+ if (ret)
-+ return ret;
-+ seq_printf(file, "Query FQ Programmable Fields Result fqid 0x%x\n",
-+ fq.fqid);
-+ seq_printf(file, " orprws: %u\n", fqd.orprws);
-+ seq_printf(file, " oa: %u\n", fqd.oa);
-+ seq_printf(file, " olws: %u\n", fqd.olws);
-+
-+ seq_printf(file, " cgid: %u\n", fqd.cgid);
-+
-+ if ((fqd.fq_ctrl & QM_FQCTRL_MASK) == 0)
-+ seq_puts(file, " fq_ctrl: None\n");
-+ else {
-+ i = 0;
-+ seq_puts(file, " fq_ctrl:\n");
-+ while (fq_ctrl_text_list[i].txt != NULL) {
-+ if ((fqd.fq_ctrl & QM_FQCTRL_MASK) &
-+ fq_ctrl_text_list[i].mask)
-+ seq_printf(file, " %s\n",
-+ fq_ctrl_text_list[i].txt);
-+ i++;
-+ }
-+ }
-+ seq_printf(file, " dest_channel: %u\n", fqd.dest.channel);
-+ seq_printf(file, " dest_wq: %u\n", fqd.dest.wq);
-+ seq_printf(file, " ics_cred: %u\n", fqd.ics_cred);
-+ seq_printf(file, " td_mant: %u\n", fqd.td.mant);
-+ seq_printf(file, " td_exp: %u\n", fqd.td.exp);
-+
-+ seq_printf(file, " ctx_b: 0x%x\n", fqd.context_b);
-+
-+ seq_printf(file, " ctx_a: 0x%llx\n", qm_fqd_stashing_get64(&fqd));
-+ /* Any stashing configured */
-+ if ((fqd.context_a.stashing.exclusive & 0x7) == 0)
-+ seq_puts(file, " ctx_a_stash_exclusive: None\n");
-+ else {
-+ seq_puts(file, " ctx_a_stash_exclusive:\n");
-+ i = 0;
-+ while (stashing_text_list[i].txt != NULL) {
-+ if ((fqd.fq_ctrl & 0x7) & stashing_text_list[i].mask)
-+ seq_printf(file, " %s\n",
-+ stashing_text_list[i].txt);
-+ i++;
-+ }
-+ }
-+ seq_printf(file, " ctx_a_stash_annotation_cl: %u\n",
-+ fqd.context_a.stashing.annotation_cl);
-+ seq_printf(file, " ctx_a_stash_data_cl: %u\n",
-+ fqd.context_a.stashing.data_cl);
-+ seq_printf(file, " ctx_a_stash_context_cl: %u\n",
-+ fqd.context_a.stashing.context_cl);
-+ return 0;
-+}
-+
-+static int query_fq_fields_open(struct inode *inode,
-+ struct file *file)
-+{
-+ return single_open(file, query_fq_fields_show, NULL);
-+}
-+
-+static ssize_t query_fq_fields_write(struct file *f,
-+ const char __user *buf, size_t count, loff_t *off)
-+{
-+ int ret;
-+ unsigned long val;
-+
-+ ret = user_input_convert(buf, count, &val);
-+ if (ret)
-+ return ret;
-+ if (val > MAX_FQID)
-+ return -EINVAL;
-+ query_fq_fields_data.fqid = (u32)val;
-+ return count;
-+}
-+
-+static const struct file_operations query_fq_fields_fops = {
-+ .owner = THIS_MODULE,
-+ .open = query_fq_fields_open,
-+ .read = seq_read,
-+ .write = query_fq_fields_write,
-+ .release = single_release,
-+};
-+
-+/*******************************************************************************
-+ * Query WQ lengths
-+ ******************************************************************************/
-+struct query_wq_lengths_data_s {
-+ union {
-+ u16 channel_wq; /* ignores wq (3 lsbits) */
-+ struct {
-+ u16 id:13; /* qm_channel */
-+ u16 __reserved:3;
-+ } __packed channel;
-+ };
-+};
-+static struct query_wq_lengths_data_s query_wq_lengths_data;
-+static int query_wq_lengths_show(struct seq_file *file, void *offset)
-+{
-+ int ret;
-+ struct qm_mcr_querywq wq;
-+ int i;
-+
-+ memset(&wq, 0, sizeof(struct qm_mcr_querywq));
-+ wq.channel.id = query_wq_lengths_data.channel.id;
-+ ret = qman_query_wq(0, &wq);
-+ if (ret)
-+ return ret;
-+ seq_printf(file, "Query Result For Channel: 0x%x\n", wq.channel.id);
-+ for (i = 0; i < 8; i++)
-+ /* mask out upper 4 bits since they are not part of length */
-+ seq_printf(file, " wq%d_len : %u\n", i, wq.wq_len[i] & 0x0fff);
-+ return 0;
-+}
-+
-+static int query_wq_lengths_open(struct inode *inode,
-+ struct file *file)
-+{
-+ return single_open(file, query_wq_lengths_show, NULL);
-+}
-+
-+static ssize_t query_wq_lengths_write(struct file *f,
-+ const char __user *buf, size_t count, loff_t *off)
-+{
-+ int ret;
-+ unsigned long val;
-+
-+ ret = user_input_convert(buf, count, &val);
-+ if (ret)
-+ return ret;
-+ if (val > 0xfff8)
-+ return -EINVAL;
-+ query_wq_lengths_data.channel.id = (u16)val;
-+ return count;
-+}
-+
-+static const struct file_operations query_wq_lengths_fops = {
-+ .owner = THIS_MODULE,
-+ .open = query_wq_lengths_open,
-+ .read = seq_read,
-+ .write = query_wq_lengths_write,
-+ .release = single_release,
-+};
-+
-+/*******************************************************************************
-+ * Query CGR
-+ ******************************************************************************/
-+struct query_cgr_s {
-+ u8 cgid;
-+};
-+static struct query_cgr_s query_cgr_data;
-+
-+static int query_cgr_show(struct seq_file *file, void *offset)
-+{
-+ int ret;
-+ struct qm_mcr_querycgr cgrd;
-+ struct qman_cgr cgr;
-+ int i, j;
-+ u32 mask;
-+
-+ memset(&cgr, 0, sizeof(cgr));
-+ memset(&cgrd, 0, sizeof(cgrd));
-+ cgr.cgrid = query_cgr_data.cgid;
-+ ret = qman_query_cgr(&cgr, &cgrd);
-+ if (ret)
-+ return ret;
-+ seq_printf(file, "Query CGR id 0x%x\n", cgr.cgrid);
-+ seq_printf(file, " wr_parm_g MA: %u, Mn: %u, SA: %u, Sn: %u, Pn: %u\n",
-+ cgrd.cgr.wr_parm_g.MA, cgrd.cgr.wr_parm_g.Mn,
-+ cgrd.cgr.wr_parm_g.SA, cgrd.cgr.wr_parm_g.Sn,
-+ cgrd.cgr.wr_parm_g.Pn);
-+
-+ seq_printf(file, " wr_parm_y MA: %u, Mn: %u, SA: %u, Sn: %u, Pn: %u\n",
-+ cgrd.cgr.wr_parm_y.MA, cgrd.cgr.wr_parm_y.Mn,
-+ cgrd.cgr.wr_parm_y.SA, cgrd.cgr.wr_parm_y.Sn,
-+ cgrd.cgr.wr_parm_y.Pn);
-+
-+ seq_printf(file, " wr_parm_r MA: %u, Mn: %u, SA: %u, Sn: %u, Pn: %u\n",
-+ cgrd.cgr.wr_parm_r.MA, cgrd.cgr.wr_parm_r.Mn,
-+ cgrd.cgr.wr_parm_r.SA, cgrd.cgr.wr_parm_r.Sn,
-+ cgrd.cgr.wr_parm_r.Pn);
-+
-+ seq_printf(file, " wr_en_g: %u, wr_en_y: %u, we_en_r: %u\n",
-+ cgrd.cgr.wr_en_g, cgrd.cgr.wr_en_y, cgrd.cgr.wr_en_r);
-+
-+ seq_printf(file, " cscn_en: %u\n", cgrd.cgr.cscn_en);
-+ if ((qman_ip_rev & 0xFF00) >= QMAN_REV30) {
-+ seq_puts(file, " cscn_targ_dcp:\n");
-+ mask = 0x80000000;
-+ for (i = 0; i < 32; i++) {
-+ if (cgrd.cgr.cscn_targ & mask)
-+ seq_printf(file, " send CSCN to dcp %u\n",
-+ (31 - i));
-+ mask >>= 1;
-+ }
-+
-+ seq_puts(file, " cscn_targ_swp:\n");
-+ for (i = 0; i < 4; i++) {
-+ mask = 0x80000000;
-+ for (j = 0; j < 32; j++) {
-+ if (cgrd.cscn_targ_swp[i] & mask)
-+ seq_printf(file, " send CSCN to swp"
-+ " %u\n", (127 - (i * 32) - j));
-+ mask >>= 1;
-+ }
-+ }
-+ } else {
-+ seq_printf(file, " cscn_targ: %u\n", cgrd.cgr.cscn_targ);
-+ }
-+ seq_printf(file, " cstd_en: %u\n", cgrd.cgr.cstd_en);
-+ seq_printf(file, " cs: %u\n", cgrd.cgr.cs);
-+
-+ seq_printf(file, " cs_thresh_TA: %u, cs_thresh_Tn: %u\n",
-+ cgrd.cgr.cs_thres.TA, cgrd.cgr.cs_thres.Tn);
-+
-+ seq_printf(file, " mode: %s\n",
-+ (cgrd.cgr.mode & QMAN_CGR_MODE_FRAME) ?
-+ "frame count" : "byte count");
-+ seq_printf(file, " i_bcnt: %llu\n", qm_mcr_querycgr_i_get64(&cgrd));
-+ seq_printf(file, " a_bcnt: %llu\n", qm_mcr_querycgr_a_get64(&cgrd));
-+
-+ return 0;
-+}
-+
-+static int query_cgr_open(struct inode *inode, struct file *file)
-+{
-+ return single_open(file, query_cgr_show, NULL);
-+}
-+
-+static ssize_t query_cgr_write(struct file *f, const char __user *buf,
-+ size_t count, loff_t *off)
-+{
-+ int ret;
-+ unsigned long val;
-+
-+ ret = user_input_convert(buf, count, &val);
-+ if (ret)
-+ return ret;
-+ if (val > 0xff)
-+ return -EINVAL;
-+ query_cgr_data.cgid = (u8)val;
-+ return count;
-+}
-+
-+static const struct file_operations query_cgr_fops = {
-+ .owner = THIS_MODULE,
-+ .open = query_cgr_open,
-+ .read = seq_read,
-+ .write = query_cgr_write,
-+ .release = single_release,
-+};
-+
-+/*******************************************************************************
-+ * Test Write CGR
-+ ******************************************************************************/
-+struct test_write_cgr_s {
-+ u64 i_bcnt;
-+ u8 cgid;
-+};
-+static struct test_write_cgr_s test_write_cgr_data;
-+
-+static int testwrite_cgr_show(struct seq_file *file, void *offset)
-+{
-+ int ret;
-+ struct qm_mcr_cgrtestwrite result;
-+ struct qman_cgr cgr;
-+ u64 i_bcnt;
-+
-+ memset(&cgr, 0, sizeof(struct qman_cgr));
-+ memset(&result, 0, sizeof(struct qm_mcr_cgrtestwrite));
-+ cgr.cgrid = test_write_cgr_data.cgid;
-+ i_bcnt = test_write_cgr_data.i_bcnt;
-+ ret = qman_testwrite_cgr(&cgr, i_bcnt, &result);
-+ if (ret)
-+ return ret;
-+ seq_printf(file, "CGR Test Write CGR id 0x%x\n", cgr.cgrid);
-+ seq_printf(file, " wr_parm_g MA: %u, Mn: %u, SA: %u, Sn: %u, Pn: %u\n",
-+ result.cgr.wr_parm_g.MA, result.cgr.wr_parm_g.Mn,
-+ result.cgr.wr_parm_g.SA, result.cgr.wr_parm_g.Sn,
-+ result.cgr.wr_parm_g.Pn);
-+ seq_printf(file, " wr_parm_y MA: %u, Mn: %u, SA: %u, Sn: %u, Pn: %u\n",
-+ result.cgr.wr_parm_y.MA, result.cgr.wr_parm_y.Mn,
-+ result.cgr.wr_parm_y.SA, result.cgr.wr_parm_y.Sn,
-+ result.cgr.wr_parm_y.Pn);
-+ seq_printf(file, " wr_parm_r MA: %u, Mn: %u, SA: %u, Sn: %u, Pn: %u\n",
-+ result.cgr.wr_parm_r.MA, result.cgr.wr_parm_r.Mn,
-+ result.cgr.wr_parm_r.SA, result.cgr.wr_parm_r.Sn,
-+ result.cgr.wr_parm_r.Pn);
-+ seq_printf(file, " wr_en_g: %u, wr_en_y: %u, we_en_r: %u\n",
-+ result.cgr.wr_en_g, result.cgr.wr_en_y, result.cgr.wr_en_r);
-+ seq_printf(file, " cscn_en: %u\n", result.cgr.cscn_en);
-+ seq_printf(file, " cscn_targ: %u\n", result.cgr.cscn_targ);
-+ seq_printf(file, " cstd_en: %u\n", result.cgr.cstd_en);
-+ seq_printf(file, " cs: %u\n", result.cgr.cs);
-+ seq_printf(file, " cs_thresh_TA: %u, cs_thresh_Tn: %u\n",
-+ result.cgr.cs_thres.TA, result.cgr.cs_thres.Tn);
-+
-+ /* Add Mode for Si 2 */
-+ seq_printf(file, " mode: %s\n",
-+ (result.cgr.mode & QMAN_CGR_MODE_FRAME) ?
-+ "frame count" : "byte count");
-+
-+ seq_printf(file, " i_bcnt: %llu\n",
-+ qm_mcr_cgrtestwrite_i_get64(&result));
-+ seq_printf(file, " a_bcnt: %llu\n",
-+ qm_mcr_cgrtestwrite_a_get64(&result));
-+ seq_printf(file, " wr_prob_g: %u\n", result.wr_prob_g);
-+ seq_printf(file, " wr_prob_y: %u\n", result.wr_prob_y);
-+ seq_printf(file, " wr_prob_r: %u\n", result.wr_prob_r);
-+ return 0;
-+}
-+
-+static int testwrite_cgr_open(struct inode *inode, struct file *file)
-+{
-+ return single_open(file, testwrite_cgr_show, NULL);
-+}
-+
-+static const struct file_operations testwrite_cgr_fops = {
-+ .owner = THIS_MODULE,
-+ .open = testwrite_cgr_open,
-+ .read = seq_read,
-+ .release = single_release,
-+};
-+
-+
-+static int testwrite_cgr_ibcnt_show(struct seq_file *file, void *offset)
-+{
-+ seq_printf(file, "i_bcnt: %llu\n", test_write_cgr_data.i_bcnt);
-+ return 0;
-+}
-+static int testwrite_cgr_ibcnt_open(struct inode *inode, struct file *file)
-+{
-+ return single_open(file, testwrite_cgr_ibcnt_show, NULL);
-+}
-+
-+static ssize_t testwrite_cgr_ibcnt_write(struct file *f, const char __user *buf,
-+ size_t count, loff_t *off)
-+{
-+ int ret;
-+ unsigned long val;
-+
-+ ret = user_input_convert(buf, count, &val);
-+ if (ret)
-+ return ret;
-+ test_write_cgr_data.i_bcnt = val;
-+ return count;
-+}
-+
-+static const struct file_operations teswrite_cgr_ibcnt_fops = {
-+ .owner = THIS_MODULE,
-+ .open = testwrite_cgr_ibcnt_open,
-+ .read = seq_read,
-+ .write = testwrite_cgr_ibcnt_write,
-+ .release = single_release,
-+};
-+
-+static int testwrite_cgr_cgrid_show(struct seq_file *file, void *offset)
-+{
-+ seq_printf(file, "cgrid: %u\n", (u32)test_write_cgr_data.cgid);
-+ return 0;
-+}
-+static int testwrite_cgr_cgrid_open(struct inode *inode, struct file *file)
-+{
-+ return single_open(file, testwrite_cgr_cgrid_show, NULL);
-+}
-+
-+static ssize_t testwrite_cgr_cgrid_write(struct file *f, const char __user *buf,
-+ size_t count, loff_t *off)
-+{
-+ int ret;
-+ unsigned long val;
-+
-+ ret = user_input_convert(buf, count, &val);
-+ if (ret)
-+ return ret;
-+ if (val > 0xff)
-+ return -EINVAL;
-+ test_write_cgr_data.cgid = (u8)val;
-+ return count;
-+}
-+
-+static const struct file_operations teswrite_cgr_cgrid_fops = {
-+ .owner = THIS_MODULE,
-+ .open = testwrite_cgr_cgrid_open,
-+ .read = seq_read,
-+ .write = testwrite_cgr_cgrid_write,
-+ .release = single_release,
-+};
-+
-+/*******************************************************************************
-+ * Query Congestion State
-+ ******************************************************************************/
-+static int query_congestion_show(struct seq_file *file, void *offset)
-+{
-+ int ret;
-+ struct qm_mcr_querycongestion cs;
-+ int i, j, in_cong = 0;
-+ u32 mask;
-+
-+ memset(&cs, 0, sizeof(struct qm_mcr_querycongestion));
-+ ret = qman_query_congestion(&cs);
-+ if (ret)
-+ return ret;
-+ seq_puts(file, "Query Congestion Result\n");
-+ for (i = 0; i < 8; i++) {
-+ mask = 0x80000000;
-+ for (j = 0; j < 32; j++) {
-+ if (cs.state.__state[i] & mask) {
-+ in_cong = 1;
-+ seq_printf(file, " cg %u: %s\n", (i*32)+j,
-+ "in congestion");
-+ }
-+ mask >>= 1;
-+ }
-+ }
-+ if (!in_cong)
-+ seq_puts(file, " All congestion groups not congested.\n");
-+ return 0;
-+}
-+
-+static int query_congestion_open(struct inode *inode, struct file *file)
-+{
-+ return single_open(file, query_congestion_show, NULL);
-+}
-+
-+static const struct file_operations query_congestion_fops = {
-+ .owner = THIS_MODULE,
-+ .open = query_congestion_open,
-+ .read = seq_read,
-+ .release = single_release,
-+};
-+
-+/*******************************************************************************
-+ * Query CCGR
-+ ******************************************************************************/
-+struct query_ccgr_s {
-+ u32 ccgid;
-+};
-+static struct query_ccgr_s query_ccgr_data;
-+
-+static int query_ccgr_show(struct seq_file *file, void *offset)
-+{
-+ int ret;
-+ struct qm_mcr_ceetm_ccgr_query ccgr_query;
-+ struct qm_mcc_ceetm_ccgr_query query_opts;
-+ int i, j;
-+ u32 mask;
-+
-+ memset(&ccgr_query, 0, sizeof(struct qm_mcr_ceetm_ccgr_query));
-+ memset(&query_opts, 0, sizeof(struct qm_mcc_ceetm_ccgr_query));
-+
-+ if ((qman_ip_rev & 0xFF00) < QMAN_REV30)
-+ return -EINVAL;
-+
-+ seq_printf(file, "Query CCGID %x\n", query_ccgr_data.ccgid);
-+ query_opts.dcpid = ((query_ccgr_data.ccgid & 0xFF000000) >> 24);
-+ query_opts.ccgrid = query_ccgr_data.ccgid & 0x000001FF;
-+ ret = qman_ceetm_query_ccgr(&query_opts, &ccgr_query);
-+ if (ret)
-+ return ret;
-+ seq_printf(file, "Query CCGR id %x in DCP %d\n", query_opts.ccgrid,
-+ query_opts.dcpid);
-+ seq_printf(file, " wr_parm_g MA: %u, Mn: %u, SA: %u, Sn: %u, Pn: %u\n",
-+ ccgr_query.cm_query.wr_parm_g.MA,
-+ ccgr_query.cm_query.wr_parm_g.Mn,
-+ ccgr_query.cm_query.wr_parm_g.SA,
-+ ccgr_query.cm_query.wr_parm_g.Sn,
-+ ccgr_query.cm_query.wr_parm_g.Pn);
-+
-+ seq_printf(file, " wr_parm_y MA: %u, Mn: %u, SA: %u, Sn: %u, Pn: %u\n",
-+ ccgr_query.cm_query.wr_parm_y.MA,
-+ ccgr_query.cm_query.wr_parm_y.Mn,
-+ ccgr_query.cm_query.wr_parm_y.SA,
-+ ccgr_query.cm_query.wr_parm_y.Sn,
-+ ccgr_query.cm_query.wr_parm_y.Pn);
-+
-+ seq_printf(file, " wr_parm_r MA: %u, Mn: %u, SA: %u, Sn: %u, Pn: %u\n",
-+ ccgr_query.cm_query.wr_parm_r.MA,
-+ ccgr_query.cm_query.wr_parm_r.Mn,
-+ ccgr_query.cm_query.wr_parm_r.SA,
-+ ccgr_query.cm_query.wr_parm_r.Sn,
-+ ccgr_query.cm_query.wr_parm_r.Pn);
-+
-+ seq_printf(file, " wr_en_g: %u, wr_en_y: %u, we_en_r: %u\n",
-+ ccgr_query.cm_query.ctl_wr_en_g,
-+ ccgr_query.cm_query.ctl_wr_en_y,
-+ ccgr_query.cm_query.ctl_wr_en_r);
-+
-+ seq_printf(file, " cscn_en: %u\n", ccgr_query.cm_query.ctl_cscn_en);
-+ seq_puts(file, " cscn_targ_dcp:\n");
-+ mask = 0x80000000;
-+ for (i = 0; i < 32; i++) {
-+ if (ccgr_query.cm_query.cscn_targ_dcp & mask)
-+ seq_printf(file, " send CSCN to dcp %u\n", (31 - i));
-+ mask >>= 1;
-+ }
-+
-+ seq_puts(file, " cscn_targ_swp:\n");
-+ for (i = 0; i < 4; i++) {
-+ mask = 0x80000000;
-+ for (j = 0; j < 32; j++) {
-+ if (ccgr_query.cm_query.cscn_targ_swp[i] & mask)
-+ seq_printf(file, " send CSCN to swp"
-+ "%u\n", (127 - (i * 32) - j));
-+ mask >>= 1;
-+ }
-+ }
-+
-+ seq_printf(file, " td_en: %u\n", ccgr_query.cm_query.ctl_td_en);
-+
-+ seq_printf(file, " cs_thresh_in_TA: %u, cs_thresh_in_Tn: %u\n",
-+ ccgr_query.cm_query.cs_thres.TA,
-+ ccgr_query.cm_query.cs_thres.Tn);
-+
-+ seq_printf(file, " cs_thresh_out_TA: %u, cs_thresh_out_Tn: %u\n",
-+ ccgr_query.cm_query.cs_thres_x.TA,
-+ ccgr_query.cm_query.cs_thres_x.Tn);
-+
-+ seq_printf(file, " td_thresh_TA: %u, td_thresh_Tn: %u\n",
-+ ccgr_query.cm_query.td_thres.TA,
-+ ccgr_query.cm_query.td_thres.Tn);
-+
-+ seq_printf(file, " mode: %s\n",
-+ (ccgr_query.cm_query.ctl_mode &
-+ QMAN_CGR_MODE_FRAME) ?
-+ "frame count" : "byte count");
-+ seq_printf(file, " i_cnt: %llu\n", (u64)ccgr_query.cm_query.i_cnt);
-+ seq_printf(file, " a_cnt: %llu\n", (u64)ccgr_query.cm_query.a_cnt);
-+
-+ return 0;
-+}
-+
-+static int query_ccgr_open(struct inode *inode, struct file *file)
-+{
-+ return single_open(file, query_ccgr_show, NULL);
-+}
-+
-+static ssize_t query_ccgr_write(struct file *f, const char __user *buf,
-+ size_t count, loff_t *off)
-+{
-+ int ret;
-+ unsigned long val;
-+
-+ ret = user_input_convert(buf, count, &val);
-+ if (ret)
-+ return ret;
-+ query_ccgr_data.ccgid = val;
-+ return count;
-+}
-+
-+static const struct file_operations query_ccgr_fops = {
-+ .owner = THIS_MODULE,
-+ .open = query_ccgr_open,
-+ .read = seq_read,
-+ .write = query_ccgr_write,
-+ .release = single_release,
-+};
-+/*******************************************************************************
-+ * QMan register
-+ ******************************************************************************/
-+struct qman_register_s {
-+ u32 val;
-+};
-+static struct qman_register_s qman_register_data;
-+
-+static void init_ccsrmempeek(void)
-+{
-+ struct device_node *dn;
-+ const u32 *regaddr_p;
-+
-+ dn = of_find_compatible_node(NULL, NULL, "fsl,qman");
-+ if (!dn) {
-+ pr_info("No fsl,qman node\n");
-+ return;
-+ }
-+ regaddr_p = of_get_address(dn, 0, &qman_ccsr_size, NULL);
-+ if (!regaddr_p) {
-+ of_node_put(dn);
-+ return;
-+ }
-+ qman_ccsr_start = of_translate_address(dn, regaddr_p);
-+ of_node_put(dn);
-+}
-+/* This function provides access to QMan ccsr memory map */
-+static int qman_ccsrmempeek(u32 *val, u32 offset)
-+{
-+ void __iomem *addr;
-+ u64 phys_addr;
-+
-+ if (!qman_ccsr_start)
-+ return -EINVAL;
-+
-+ if (offset > (qman_ccsr_size - sizeof(u32)))
-+ return -EINVAL;
-+
-+ phys_addr = qman_ccsr_start + offset;
-+ addr = ioremap(phys_addr, sizeof(u32));
-+ if (!addr) {
-+ pr_err("ccsrmempeek, ioremap failed\n");
-+ return -EINVAL;
-+ }
-+ *val = in_be32(addr);
-+ iounmap(addr);
-+ return 0;
-+}
-+
-+static int qman_ccsrmempeek_show(struct seq_file *file, void *offset)
-+{
-+ u32 b;
-+
-+ qman_ccsrmempeek(&b, qman_register_data.val);
-+ seq_printf(file, "QMan register offset = 0x%x\n",
-+ qman_register_data.val);
-+ seq_printf(file, "value = 0x%08x\n", b);
-+
-+ return 0;
-+}
-+
-+static int qman_ccsrmempeek_open(struct inode *inode, struct file *file)
-+{
-+ return single_open(file, qman_ccsrmempeek_show, NULL);
-+}
-+
-+static ssize_t qman_ccsrmempeek_write(struct file *f, const char __user *buf,
-+ size_t count, loff_t *off)
-+{
-+ int ret;
-+ unsigned long val;
-+
-+ ret = user_input_convert(buf, count, &val);
-+ if (ret)
-+ return ret;
-+ /* multiple of 4 */
-+ if (val > (qman_ccsr_size - sizeof(u32))) {
-+ pr_info("Input 0x%lx > 0x%llx\n",
-+ val, (qman_ccsr_size - sizeof(u32)));
-+ return -EINVAL;
-+ }
-+ if (val & 0x3) {
-+ pr_info("Input 0x%lx not multiple of 4\n", val);
-+ return -EINVAL;
-+ }
-+ qman_register_data.val = val;
-+ return count;
-+}
-+
-+static const struct file_operations qman_ccsrmempeek_fops = {
-+ .owner = THIS_MODULE,
-+ .open = qman_ccsrmempeek_open,
-+ .read = seq_read,
-+ .write = qman_ccsrmempeek_write,
-+};
-+
-+/*******************************************************************************
-+ * QMan state
-+ ******************************************************************************/
-+static int qman_fqd_state_show(struct seq_file *file, void *offset)
-+{
-+ struct qm_mcr_queryfq_np np;
-+ struct qman_fq fq;
-+ struct line_buffer_fq line_buf;
-+ int ret, i;
-+ u8 *state = file->private;
-+ u32 qm_fq_state_cnt[ARRAY_SIZE(fqd_states)];
-+
-+ memset(qm_fq_state_cnt, 0, sizeof(qm_fq_state_cnt));
-+ memset(&line_buf, 0, sizeof(line_buf));
-+
-+ seq_printf(file, "List of fq ids in state: %s\n", state_txt[*state]);
-+
-+ for (i = 1; i < fqid_max; i++) {
-+ fq.fqid = i;
-+ ret = qman_query_fq_np(&fq, &np);
-+ if (ret)
-+ return ret;
-+ if (*state == (np.state & QM_MCR_NP_STATE_MASK))
-+ add_to_line_buffer(&line_buf, fq.fqid, file);
-+ /* Keep a summary count of all states */
-+ if ((np.state & QM_MCR_NP_STATE_MASK) < ARRAY_SIZE(fqd_states))
-+ qm_fq_state_cnt[(np.state & QM_MCR_NP_STATE_MASK)]++;
-+ }
-+ flush_line_buffer(&line_buf, file);
-+
-+ for (i = 0; i < ARRAY_SIZE(fqd_states); i++) {
-+ seq_printf(file, "%s count = %u\n", state_txt[i],
-+ qm_fq_state_cnt[i]);
-+ }
-+ return 0;
-+}
-+
-+static int qman_fqd_state_open(struct inode *inode, struct file *file)
-+{
-+ return single_open(file, qman_fqd_state_show, inode->i_private);
-+}
-+
-+static const struct file_operations qman_fqd_state_fops = {
-+ .owner = THIS_MODULE,
-+ .open = qman_fqd_state_open,
-+ .read = seq_read,
-+};
-+
-+static int qman_fqd_ctrl_show(struct seq_file *file, void *offset)
-+{
-+ struct qm_fqd fqd;
-+ struct qman_fq fq;
-+ u32 fq_en_cnt = 0, fq_di_cnt = 0;
-+ int ret, i;
-+ struct mask_filter_s *data = file->private;
-+ const char *ctrl_txt = get_fqd_ctrl_text(data->mask);
-+ struct line_buffer_fq line_buf;
-+
-+ memset(&line_buf, 0, sizeof(line_buf));
-+ seq_printf(file, "List of fq ids with: %s :%s\n",
-+ ctrl_txt, (data->filter) ? "enabled" : "disabled");
-+ for (i = 1; i < fqid_max; i++) {
-+ fq.fqid = i;
-+ memset(&fqd, 0, sizeof(struct qm_fqd));
-+ ret = qman_query_fq(&fq, &fqd);
-+ if (ret)
-+ return ret;
-+ if (data->filter) {
-+ if (fqd.fq_ctrl & data->mask)
-+ add_to_line_buffer(&line_buf, fq.fqid, file);
-+ } else {
-+ if (!(fqd.fq_ctrl & data->mask))
-+ add_to_line_buffer(&line_buf, fq.fqid, file);
-+ }
-+ if (fqd.fq_ctrl & data->mask)
-+ fq_en_cnt++;
-+ else
-+ fq_di_cnt++;
-+ }
-+ flush_line_buffer(&line_buf, file);
-+
-+ seq_printf(file, "Total FQD with: %s : enabled = %u\n",
-+ ctrl_txt, fq_en_cnt);
-+ seq_printf(file, "Total FQD with: %s : disabled = %u\n",
-+ ctrl_txt, fq_di_cnt);
-+ return 0;
-+}
-+
-+/*******************************************************************************
-+ * QMan ctrl CGE, TDE, ORP, CTX, CPC, SFDR, BLOCK, HOLD, CACHE
-+ ******************************************************************************/
-+static int qman_fqd_ctrl_open(struct inode *inode, struct file *file)
-+{
-+ return single_open(file, qman_fqd_ctrl_show, inode->i_private);
-+}
-+
-+static const struct file_operations qman_fqd_ctrl_fops = {
-+ .owner = THIS_MODULE,
-+ .open = qman_fqd_ctrl_open,
-+ .read = seq_read,
-+};
-+
-+/*******************************************************************************
-+ * QMan ctrl summary
-+ ******************************************************************************/
-+/*******************************************************************************
-+ * QMan summary state
-+ ******************************************************************************/
-+static int qman_fqd_non_prog_summary_show(struct seq_file *file, void *offset)
-+{
-+ struct qm_mcr_queryfq_np np;
-+ struct qman_fq fq;
-+ int ret, i;
-+ u32 qm_fq_state_cnt[ARRAY_SIZE(fqd_states)];
-+
-+ memset(qm_fq_state_cnt, 0, sizeof(qm_fq_state_cnt));
-+
-+ for (i = 1; i < fqid_max; i++) {
-+ fq.fqid = i;
-+ ret = qman_query_fq_np(&fq, &np);
-+ if (ret)
-+ return ret;
-+ /* Keep a summary count of all states */
-+ if ((np.state & QM_MCR_NP_STATE_MASK) < ARRAY_SIZE(fqd_states))
-+ qm_fq_state_cnt[(np.state & QM_MCR_NP_STATE_MASK)]++;
-+ }
-+
-+ for (i = 0; i < ARRAY_SIZE(fqd_states); i++) {
-+ seq_printf(file, "%s count = %u\n", state_txt[i],
-+ qm_fq_state_cnt[i]);
-+ }
-+ return 0;
-+}
-+
-+static int qman_fqd_prog_summary_show(struct seq_file *file, void *offset)
-+{
-+ struct qm_fqd fqd;
-+ struct qman_fq fq;
-+ int ret, i , j;
-+ u32 qm_prog_cnt[ARRAY_SIZE(mask_filter)/2];
-+
-+ memset(qm_prog_cnt, 0, sizeof(qm_prog_cnt));
-+
-+ for (i = 1; i < fqid_max; i++) {
-+ memset(&fqd, 0, sizeof(struct qm_fqd));
-+ fq.fqid = i;
-+ ret = qman_query_fq(&fq, &fqd);
-+ if (ret)
-+ return ret;
-+ /* Keep a summary count of all states */
-+ for (j = 0; j < ARRAY_SIZE(mask_filter); j += 2)
-+ if ((fqd.fq_ctrl & QM_FQCTRL_MASK) &
-+ mask_filter[j].mask)
-+ qm_prog_cnt[j/2]++;
-+ }
-+ for (i = 0; i < ARRAY_SIZE(mask_filter) / 2; i++) {
-+ seq_printf(file, "%s count = %u\n",
-+ get_fqd_ctrl_text(mask_filter[i*2].mask),
-+ qm_prog_cnt[i]);
-+ }
-+ return 0;
-+}
-+
-+static int qman_fqd_summary_show(struct seq_file *file, void *offset)
-+{
-+ int ret;
-+
-+ /* Display summary of non programmable fields */
-+ ret = qman_fqd_non_prog_summary_show(file, offset);
-+ if (ret)
-+ return ret;
-+ seq_puts(file, "-----------------------------------------\n");
-+ /* Display programmable fields */
-+ ret = qman_fqd_prog_summary_show(file, offset);
-+ if (ret)
-+ return ret;
-+ return 0;
-+}
-+
-+static int qman_fqd_summary_open(struct inode *inode, struct file *file)
-+{
-+ return single_open(file, qman_fqd_summary_show, NULL);
-+}
-+
-+static const struct file_operations qman_fqd_summary_fops = {
-+ .owner = THIS_MODULE,
-+ .open = qman_fqd_summary_open,
-+ .read = seq_read,
-+};
-+
-+/*******************************************************************************
-+ * QMan destination work queue
-+ ******************************************************************************/
-+struct qman_dest_wq_s {
-+ u16 wq_id;
-+};
-+static struct qman_dest_wq_s qman_dest_wq_data = {
-+ .wq_id = 0,
-+};
-+
-+static int qman_fqd_dest_wq_show(struct seq_file *file, void *offset)
-+{
-+ struct qm_fqd fqd;
-+ struct qman_fq fq;
-+ int ret, i;
-+ u16 *wq, wq_id = qman_dest_wq_data.wq_id;
-+ struct line_buffer_fq line_buf;
-+
-+ memset(&line_buf, 0, sizeof(line_buf));
-+ /* use vmalloc : need to allocate large memory region and don't
-+ * require the memory to be physically contiguous. */
-+ wq = vzalloc(sizeof(u16) * (0xFFFF+1));
-+ if (!wq)
-+ return -ENOMEM;
-+
-+ seq_printf(file, "List of fq ids with destination work queue id"
-+ " = 0x%x\n", wq_id);
-+
-+ for (i = 1; i < fqid_max; i++) {
-+ fq.fqid = i;
-+ memset(&fqd, 0, sizeof(struct qm_fqd));
-+ ret = qman_query_fq(&fq, &fqd);
-+ if (ret) {
-+ vfree(wq);
-+ return ret;
-+ }
-+ if (wq_id == fqd.dest_wq)
-+ add_to_line_buffer(&line_buf, fq.fqid, file);
-+ wq[fqd.dest_wq]++;
-+ }
-+ flush_line_buffer(&line_buf, file);
-+
-+ seq_puts(file, "Summary of all FQD destination work queue values\n");
-+ for (i = 0; i < 0xFFFF; i++) {
-+ if (wq[i])
-+ seq_printf(file, "Channel: 0x%x WQ: 0x%x WQ_ID: 0x%x, "
-+ "count = %u\n", i >> 3, i & 0x3, i, wq[i]);
-+ }
-+ vfree(wq);
-+ return 0;
-+}
-+
-+static ssize_t qman_fqd_dest_wq_write(struct file *f, const char __user *buf,
-+ size_t count, loff_t *off)
-+{
-+ int ret;
-+ unsigned long val;
-+
-+ ret = user_input_convert(buf, count, &val);
-+ if (ret)
-+ return ret;
-+ if (val > 0xFFFF)
-+ return -EINVAL;
-+ qman_dest_wq_data.wq_id = val;
-+ return count;
-+}
-+
-+static int qman_fqd_dest_wq_open(struct inode *inode, struct file *file)
-+{
-+ return single_open(file, qman_fqd_dest_wq_show, NULL);
-+}
-+
-+static const struct file_operations qman_fqd_dest_wq_fops = {
-+ .owner = THIS_MODULE,
-+ .open = qman_fqd_dest_wq_open,
-+ .read = seq_read,
-+ .write = qman_fqd_dest_wq_write,
-+};
-+
-+/*******************************************************************************
-+ * QMan Intra-Class Scheduling Credit
-+ ******************************************************************************/
-+static int qman_fqd_cred_show(struct seq_file *file, void *offset)
-+{
-+ struct qm_fqd fqd;
-+ struct qman_fq fq;
-+ int ret, i;
-+ u32 fq_cnt = 0;
-+ struct line_buffer_fq line_buf;
-+
-+ memset(&line_buf, 0, sizeof(line_buf));
-+ seq_puts(file, "List of fq ids with Intra-Class Scheduling Credit > 0"
-+ "\n");
-+
-+ for (i = 1; i < fqid_max; i++) {
-+ fq.fqid = i;
-+ memset(&fqd, 0, sizeof(struct qm_fqd));
-+ ret = qman_query_fq(&fq, &fqd);
-+ if (ret)
-+ return ret;
-+ if (fqd.ics_cred > 0) {
-+ add_to_line_buffer(&line_buf, fq.fqid, file);
-+ fq_cnt++;
-+ }
-+ }
-+ flush_line_buffer(&line_buf, file);
-+
-+ seq_printf(file, "Total FQD with ics_cred > 0 = %d\n", fq_cnt);
-+ return 0;
-+}
-+
-+static int qman_fqd_cred_open(struct inode *inode, struct file *file)
-+{
-+ return single_open(file, qman_fqd_cred_show, NULL);
-+}
-+
-+static const struct file_operations qman_fqd_cred_fops = {
-+ .owner = THIS_MODULE,
-+ .open = qman_fqd_cred_open,
-+ .read = seq_read,
-+};
-+
-+/*******************************************************************************
-+ * Class Queue Fields
-+ ******************************************************************************/
-+struct query_cq_fields_data_s {
-+ u32 cqid;
-+};
-+
-+static struct query_cq_fields_data_s query_cq_fields_data = {
-+ .cqid = 1,
-+};
-+
-+static int query_cq_fields_show(struct seq_file *file, void *offset)
-+{
-+ int ret;
-+ struct qm_mcr_ceetm_cq_query query_result;
-+ unsigned int cqid;
-+ unsigned int portal;
-+
-+ if ((qman_ip_rev & 0xFF00) < QMAN_REV30)
-+ return -EINVAL;
-+
-+ cqid = query_cq_fields_data.cqid & 0x00FFFFFF;
-+ portal = query_cq_fields_data.cqid >> 24;
-+ if (portal > qm_dc_portal_fman1)
-+ return -EINVAL;
-+
-+ ret = qman_ceetm_query_cq(cqid, portal, &query_result);
-+ if (ret)
-+ return ret;
-+ seq_printf(file, "Query CQ Fields Result cqid 0x%x on DCP %d\n",
-+ cqid, portal);
-+ seq_printf(file, " ccgid: %u\n", query_result.ccgid);
-+ seq_printf(file, " state: %u\n", query_result.state);
-+ seq_printf(file, " pfdr_hptr: %u\n", query_result.pfdr_hptr);
-+ seq_printf(file, " pfdr_tptr: %u\n", query_result.pfdr_tptr);
-+ seq_printf(file, " od1_xsfdr: %u\n", query_result.od1_xsfdr);
-+ seq_printf(file, " od2_xsfdr: %u\n", query_result.od2_xsfdr);
-+ seq_printf(file, " od3_xsfdr: %u\n", query_result.od3_xsfdr);
-+ seq_printf(file, " od4_xsfdr: %u\n", query_result.od4_xsfdr);
-+ seq_printf(file, " od5_xsfdr: %u\n", query_result.od5_xsfdr);
-+ seq_printf(file, " od6_xsfdr: %u\n", query_result.od6_xsfdr);
-+ seq_printf(file, " ra1_xsfdr: %u\n", query_result.ra1_xsfdr);
-+ seq_printf(file, " ra2_xsfdr: %u\n", query_result.ra2_xsfdr);
-+ seq_printf(file, " frame_count: %u\n", query_result.frm_cnt);
-+
-+ return 0;
-+}
-+
-+static int query_cq_fields_open(struct inode *inode,
-+ struct file *file)
-+{
-+ return single_open(file, query_cq_fields_show, NULL);
-+}
-+
-+static ssize_t query_cq_fields_write(struct file *f,
-+ const char __user *buf, size_t count, loff_t *off)
-+{
-+ int ret;
-+ unsigned long val;
-+
-+ ret = user_input_convert(buf, count, &val);
-+ if (ret)
-+ return ret;
-+ query_cq_fields_data.cqid = (u32)val;
-+ return count;
-+}
-+
-+static const struct file_operations query_cq_fields_fops = {
-+ .owner = THIS_MODULE,
-+ .open = query_cq_fields_open,
-+ .read = seq_read,
-+ .write = query_cq_fields_write,
-+ .release = single_release,
-+};
-+
-+/*******************************************************************************
-+ * READ CEETM_XSFDR_IN_USE
-+ ******************************************************************************/
-+struct query_ceetm_xsfdr_data_s {
-+ enum qm_dc_portal dcp_portal;
-+};
-+
-+static struct query_ceetm_xsfdr_data_s query_ceetm_xsfdr_data;
-+
-+static int query_ceetm_xsfdr_show(struct seq_file *file, void *offset)
-+{
-+ int ret;
-+ unsigned int xsfdr_in_use;
-+ enum qm_dc_portal portal;
-+
-+
-+ if (qman_ip_rev < QMAN_REV31)
-+ return -EINVAL;
-+
-+ portal = query_ceetm_xsfdr_data.dcp_portal;
-+ ret = qman_ceetm_get_xsfdr(portal, &xsfdr_in_use);
-+ if (ret) {
-+ seq_printf(file, "Read CEETM_XSFDR_IN_USE on DCP %d failed\n",
-+ portal);
-+ return ret;
-+ }
-+
-+ seq_printf(file, "DCP%d: CEETM_XSFDR_IN_USE number is %u\n", portal,
-+ (xsfdr_in_use & 0x1FFF));
-+ return 0;
-+}
-+
-+static int query_ceetm_xsfdr_open(struct inode *inode,
-+ struct file *file)
-+{
-+ return single_open(file, query_ceetm_xsfdr_show, NULL);
-+}
-+
-+static ssize_t query_ceetm_xsfdr_write(struct file *f,
-+ const char __user *buf, size_t count, loff_t *off)
-+{
-+ int ret;
-+ unsigned long val;
-+
-+ ret = user_input_convert(buf, count, &val);
-+ if (ret)
-+ return ret;
-+ if (val > qm_dc_portal_fman1)
-+ return -EINVAL;
-+ query_ceetm_xsfdr_data.dcp_portal = (u32)val;
-+ return count;
-+}
-+
-+static const struct file_operations query_ceetm_xsfdr_fops = {
-+ .owner = THIS_MODULE,
-+ .open = query_ceetm_xsfdr_open,
-+ .read = seq_read,
-+ .write = query_ceetm_xsfdr_write,
-+ .release = single_release,
-+};
-+
-+/* helper macros used in qman_debugfs_module_init */
-+#define QMAN_DBGFS_ENTRY(name, mode, parent, data, fops) \
-+ do { \
-+ d = debugfs_create_file(name, \
-+ mode, parent, \
-+ data, \
-+ fops); \
-+ if (d == NULL) { \
-+ ret = -ENOMEM; \
-+ goto _return; \
-+ } \
-+ } while (0)
-+
-+/* dfs_root as parent */
-+#define QMAN_DBGFS_ENTRY_ROOT(name, mode, data, fops) \
-+ QMAN_DBGFS_ENTRY(name, mode, dfs_root, data, fops)
-+
-+/* fqd_root as parent */
-+#define QMAN_DBGFS_ENTRY_FQDROOT(name, mode, data, fops) \
-+ QMAN_DBGFS_ENTRY(name, mode, fqd_root, data, fops)
-+
-+/* fqd state */
-+#define QMAN_DBGFS_ENTRY_FQDSTATE(name, index) \
-+ QMAN_DBGFS_ENTRY_FQDROOT(name, S_IRUGO, \
-+ (void *)&mask_filter[index], &qman_fqd_ctrl_fops)
-+
-+static int __init qman_debugfs_module_init(void)
-+{
-+ int ret = 0;
-+ struct dentry *d, *fqd_root;
-+ u32 reg;
-+
-+ fqid_max = 0;
-+ init_ccsrmempeek();
-+ if (qman_ccsr_start) {
-+ if (!qman_ccsrmempeek(&reg, QM_FQD_AR)) {
-+ /* extract the size of the FQD window */
-+ reg = reg & 0x3f;
-+ /* calculate valid frame queue descriptor range */
-+ fqid_max = (1 << (reg + 1)) / QM_FQD_BLOCK_SIZE;
-+ }
-+ }
-+ dfs_root = debugfs_create_dir("qman", NULL);
-+ fqd_root = debugfs_create_dir("fqd", dfs_root);
-+ if (dfs_root == NULL || fqd_root == NULL) {
-+ ret = -ENOMEM;
-+ pr_err("Cannot create qman/fqd debugfs dir\n");
-+ goto _return;
-+ }
-+ if (fqid_max) {
-+ QMAN_DBGFS_ENTRY_ROOT("ccsrmempeek", S_IRUGO | S_IWUGO,
-+ NULL, &qman_ccsrmempeek_fops);
-+ }
-+ QMAN_DBGFS_ENTRY_ROOT("query_fq_np_fields", S_IRUGO | S_IWUGO,
-+ &query_fq_np_fields_data, &query_fq_np_fields_fops);
-+
-+ QMAN_DBGFS_ENTRY_ROOT("query_fq_fields", S_IRUGO | S_IWUGO,
-+ &query_fq_fields_data, &query_fq_fields_fops);
-+
-+ QMAN_DBGFS_ENTRY_ROOT("query_wq_lengths", S_IRUGO | S_IWUGO,
-+ &query_wq_lengths_data, &query_wq_lengths_fops);
-+
-+ QMAN_DBGFS_ENTRY_ROOT("query_cgr", S_IRUGO | S_IWUGO,
-+ &query_cgr_data, &query_cgr_fops);
-+
-+ QMAN_DBGFS_ENTRY_ROOT("query_congestion", S_IRUGO,
-+ NULL, &query_congestion_fops);
-+
-+ QMAN_DBGFS_ENTRY_ROOT("testwrite_cgr", S_IRUGO,
-+ NULL, &testwrite_cgr_fops);
-+
-+ QMAN_DBGFS_ENTRY_ROOT("testwrite_cgr_cgrid", S_IRUGO | S_IWUGO,
-+ NULL, &teswrite_cgr_cgrid_fops);
-+
-+ QMAN_DBGFS_ENTRY_ROOT("testwrite_cgr_ibcnt", S_IRUGO | S_IWUGO,
-+ NULL, &teswrite_cgr_ibcnt_fops);
-+
-+ QMAN_DBGFS_ENTRY_ROOT("query_ceetm_ccgr", S_IRUGO | S_IWUGO,
-+ &query_ccgr_data, &query_ccgr_fops);
-+ /* Create files with fqd_root as parent */
-+
-+ QMAN_DBGFS_ENTRY_FQDROOT("stateoos", S_IRUGO,
-+ (void *)&fqd_states[QM_MCR_NP_STATE_OOS], &qman_fqd_state_fops);
-+
-+ QMAN_DBGFS_ENTRY_FQDROOT("state_retired", S_IRUGO,
-+ (void *)&fqd_states[QM_MCR_NP_STATE_RETIRED],
-+ &qman_fqd_state_fops);
-+
-+ QMAN_DBGFS_ENTRY_FQDROOT("state_tentatively_sched", S_IRUGO,
-+ (void *)&fqd_states[QM_MCR_NP_STATE_TEN_SCHED],
-+ &qman_fqd_state_fops);
-+
-+ QMAN_DBGFS_ENTRY_FQDROOT("state_truly_sched", S_IRUGO,
-+ (void *)&fqd_states[QM_MCR_NP_STATE_TRU_SCHED],
-+ &qman_fqd_state_fops);
-+
-+ QMAN_DBGFS_ENTRY_FQDROOT("state_parked", S_IRUGO,
-+ (void *)&fqd_states[QM_MCR_NP_STATE_PARKED],
-+ &qman_fqd_state_fops);
-+
-+ QMAN_DBGFS_ENTRY_FQDROOT("state_active", S_IRUGO,
-+ (void *)&fqd_states[QM_MCR_NP_STATE_ACTIVE],
-+ &qman_fqd_state_fops);
-+ QMAN_DBGFS_ENTRY_ROOT("query_cq_fields", S_IRUGO | S_IWUGO,
-+ &query_cq_fields_data, &query_cq_fields_fops);
-+ QMAN_DBGFS_ENTRY_ROOT("query_ceetm_xsfdr_in_use", S_IRUGO | S_IWUGO,
-+ &query_ceetm_xsfdr_data, &query_ceetm_xsfdr_fops);
-+
-+
-+ QMAN_DBGFS_ENTRY_FQDSTATE("cge_enable", 17);
-+
-+ QMAN_DBGFS_ENTRY_FQDSTATE("cge_disable", 16);
-+
-+ QMAN_DBGFS_ENTRY_FQDSTATE("tde_enable", 15);
-+
-+ QMAN_DBGFS_ENTRY_FQDSTATE("tde_disable", 14);
-+
-+ QMAN_DBGFS_ENTRY_FQDSTATE("orp_enable", 13);
-+
-+ QMAN_DBGFS_ENTRY_FQDSTATE("orp_disable", 12);
-+
-+ QMAN_DBGFS_ENTRY_FQDSTATE("ctx_a_stashing_enable", 11);
-+
-+ QMAN_DBGFS_ENTRY_FQDSTATE("ctx_a_stashing_disable", 10);
-+
-+ QMAN_DBGFS_ENTRY_FQDSTATE("cpc_enable", 9);
-+
-+ QMAN_DBGFS_ENTRY_FQDSTATE("cpc_disable", 8);
-+
-+ QMAN_DBGFS_ENTRY_FQDSTATE("sfdr_enable", 7);
-+
-+ QMAN_DBGFS_ENTRY_FQDSTATE("sfdr_disable", 6);
-+
-+ QMAN_DBGFS_ENTRY_FQDSTATE("avoid_blocking_enable", 5);
-+
-+ QMAN_DBGFS_ENTRY_FQDSTATE("avoid_blocking_disable", 4);
-+
-+ QMAN_DBGFS_ENTRY_FQDSTATE("hold_active_enable", 3);
-+
-+ QMAN_DBGFS_ENTRY_FQDSTATE("hold_active_disable", 2);
-+
-+ QMAN_DBGFS_ENTRY_FQDSTATE("prefer_in_cache_enable", 1);
-+
-+ QMAN_DBGFS_ENTRY_FQDSTATE("prefer_in_cache_disable", 0);
-+
-+ QMAN_DBGFS_ENTRY_FQDROOT("summary", S_IRUGO,
-+ NULL, &qman_fqd_summary_fops);
-+
-+ QMAN_DBGFS_ENTRY_FQDROOT("wq", S_IRUGO | S_IWUGO,
-+ NULL, &qman_fqd_dest_wq_fops);
-+
-+ QMAN_DBGFS_ENTRY_FQDROOT("cred", S_IRUGO,
-+ NULL, &qman_fqd_cred_fops);
-+
-+ return 0;
-+
-+_return:
-+ debugfs_remove_recursive(dfs_root);
-+ return ret;
-+}
-+
-+static void __exit qman_debugfs_module_exit(void)
-+{
-+ debugfs_remove_recursive(dfs_root);
-+}
-+
-+module_init(qman_debugfs_module_init);
-+module_exit(qman_debugfs_module_exit);
-+MODULE_LICENSE("Dual BSD/GPL");
---- /dev/null
-+++ b/drivers/staging/fsl_qbman/qman_driver.c
-@@ -0,0 +1,977 @@
-+/* Copyright 2008-2012 Freescale Semiconductor, Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#include "qman_private.h"
-+
-+#include <asm/smp.h> /* hard_smp_processor_id() if !CONFIG_SMP */
-+#ifdef CONFIG_HOTPLUG_CPU
-+#include <linux/cpu.h>
-+#endif
-+
-+/* Global variable containing revision id (even on non-control plane systems
-+ * where CCSR isn't available) */
-+u16 qman_ip_rev;
-+EXPORT_SYMBOL(qman_ip_rev);
-+u8 qman_ip_cfg;
-+EXPORT_SYMBOL(qman_ip_cfg);
-+u16 qm_channel_pool1 = QMAN_CHANNEL_POOL1;
-+EXPORT_SYMBOL(qm_channel_pool1);
-+u16 qm_channel_caam = QMAN_CHANNEL_CAAM;
-+EXPORT_SYMBOL(qm_channel_caam);
-+u16 qm_channel_pme = QMAN_CHANNEL_PME;
-+EXPORT_SYMBOL(qm_channel_pme);
-+u16 qm_channel_dce = QMAN_CHANNEL_DCE;
-+EXPORT_SYMBOL(qm_channel_dce);
-+u16 qman_portal_max;
-+EXPORT_SYMBOL(qman_portal_max);
-+
-+u32 qman_clk;
-+struct qm_ceetm qman_ceetms[QMAN_CEETM_MAX];
-+/* the qman ceetm instances on the given SoC */
-+u8 num_ceetms;
-+
-+/* For these variables, and the portal-initialisation logic, the
-+ * comments in bman_driver.c apply here so won't be repeated. */
-+static struct qman_portal *shared_portals[NR_CPUS];
-+static int num_shared_portals;
-+static int shared_portals_idx;
-+static LIST_HEAD(unused_pcfgs);
-+static DEFINE_SPINLOCK(unused_pcfgs_lock);
-+
-+/* A SDQCR mask comprising all the available/visible pool channels */
-+static u32 pools_sdqcr;
-+
-+#define STR_ERR_NOPROP "No '%s' property in node %s\n"
-+#define STR_ERR_CELL "'%s' is not a %d-cell range in node %s\n"
-+#define STR_FQID_RANGE "fsl,fqid-range"
-+#define STR_POOL_CHAN_RANGE "fsl,pool-channel-range"
-+#define STR_CGRID_RANGE "fsl,cgrid-range"
-+
-+/* A "fsl,fqid-range" node; release the given range to the allocator */
-+static __init int fsl_fqid_range_init(struct device_node *node)
-+{
-+ int ret;
-+ const u32 *range = of_get_property(node, STR_FQID_RANGE, &ret);
-+ if (!range) {
-+ pr_err(STR_ERR_NOPROP, STR_FQID_RANGE, node->full_name);
-+ return -EINVAL;
-+ }
-+ if (ret != 8) {
-+ pr_err(STR_ERR_CELL, STR_FQID_RANGE, 2, node->full_name);
-+ return -EINVAL;
-+ }
-+ qman_seed_fqid_range(be32_to_cpu(range[0]), be32_to_cpu(range[1]));
-+ pr_info("Qman: FQID allocator includes range %d:%d\n",
-+ be32_to_cpu(range[0]), be32_to_cpu(range[1]));
-+ return 0;
-+}
-+
-+/* A "fsl,pool-channel-range" node; add to the SDQCR mask only */
-+static __init int fsl_pool_channel_range_sdqcr(struct device_node *node)
-+{
-+ int ret;
-+ const u32 *chanid = of_get_property(node, STR_POOL_CHAN_RANGE, &ret);
-+ if (!chanid) {
-+ pr_err(STR_ERR_NOPROP, STR_POOL_CHAN_RANGE, node->full_name);
-+ return -EINVAL;
-+ }
-+ if (ret != 8) {
-+ pr_err(STR_ERR_CELL, STR_POOL_CHAN_RANGE, 1, node->full_name);
-+ return -EINVAL;
-+ }
-+ for (ret = 0; ret < be32_to_cpu(chanid[1]); ret++)
-+ pools_sdqcr |= QM_SDQCR_CHANNELS_POOL_CONV(be32_to_cpu(chanid[0]) + ret);
-+ return 0;
-+}
-+
-+/* A "fsl,pool-channel-range" node; release the given range to the allocator */
-+static __init int fsl_pool_channel_range_init(struct device_node *node)
-+{
-+ int ret;
-+ const u32 *chanid = of_get_property(node, STR_POOL_CHAN_RANGE, &ret);
-+ if (!chanid) {
-+ pr_err(STR_ERR_NOPROP, STR_POOL_CHAN_RANGE, node->full_name);
-+ return -EINVAL;
-+ }
-+ if (ret != 8) {
-+ pr_err(STR_ERR_CELL, STR_POOL_CHAN_RANGE, 1, node->full_name);
-+ return -EINVAL;
-+ }
-+ qman_seed_pool_range(be32_to_cpu(chanid[0]), be32_to_cpu(chanid[1]));
-+ pr_info("Qman: pool channel allocator includes range %d:%d\n",
-+ be32_to_cpu(chanid[0]), be32_to_cpu(chanid[1]));
-+ return 0;
-+}
-+
-+/* A "fsl,cgrid-range" node; release the given range to the allocator */
-+static __init int fsl_cgrid_range_init(struct device_node *node)
-+{
-+ struct qman_cgr cgr;
-+ int ret, errors = 0;
-+ const u32 *range = of_get_property(node, STR_CGRID_RANGE, &ret);
-+ if (!range) {
-+ pr_err(STR_ERR_NOPROP, STR_CGRID_RANGE, node->full_name);
-+ return -EINVAL;
-+ }
-+ if (ret != 8) {
-+ pr_err(STR_ERR_CELL, STR_CGRID_RANGE, 2, node->full_name);
-+ return -EINVAL;
-+ }
-+ qman_seed_cgrid_range(be32_to_cpu(range[0]), be32_to_cpu(range[1]));
-+ pr_info("Qman: CGRID allocator includes range %d:%d\n",
-+ be32_to_cpu(range[0]), be32_to_cpu(range[1]));
-+ for (cgr.cgrid = 0; cgr.cgrid < __CGR_NUM; cgr.cgrid++) {
-+ ret = qman_modify_cgr(&cgr, QMAN_CGR_FLAG_USE_INIT, NULL);
-+ if (ret)
-+ errors++;
-+ }
-+ if (errors)
-+ pr_err("Warning: %d error%s while initialising CGRs %d:%d\n",
-+ errors, (errors > 1) ? "s" : "", range[0], range[1]);
-+ return 0;
-+}
-+
-+static __init int fsl_ceetm_init(struct device_node *node)
-+{
-+ enum qm_dc_portal dcp_portal;
-+ struct qm_ceetm_sp *sp;
-+ struct qm_ceetm_lni *lni;
-+ int ret, i;
-+ const u32 *range;
-+
-+ /* Find LFQID range */
-+ range = of_get_property(node, "fsl,ceetm-lfqid-range", &ret);
-+ if (!range) {
-+ pr_err("No fsl,ceetm-lfqid-range in node %s\n",
-+ node->full_name);
-+ return -EINVAL;
-+ }
-+ if (ret != 8) {
-+ pr_err("fsl,ceetm-lfqid-range is not a 2-cell range in node"
-+ " %s\n", node->full_name);
-+ return -EINVAL;
-+ }
-+
-+ dcp_portal = (be32_to_cpu(range[0]) & 0x0F0000) >> 16;
-+ if (dcp_portal > qm_dc_portal_fman1) {
-+ pr_err("The DCP portal %d doesn't support CEETM\n", dcp_portal);
-+ return -EINVAL;
-+ }
-+
-+ if (dcp_portal == qm_dc_portal_fman0)
-+ qman_seed_ceetm0_lfqid_range(be32_to_cpu(range[0]), be32_to_cpu(range[1]));
-+ if (dcp_portal == qm_dc_portal_fman1)
-+ qman_seed_ceetm1_lfqid_range(be32_to_cpu(range[0]), be32_to_cpu(range[1]));
-+ pr_debug("Qman: The lfqid allocator of CEETM %d includes range"
-+ " 0x%x:0x%x\n", dcp_portal, be32_to_cpu(range[0]), be32_to_cpu(range[1]));
-+
-+ qman_ceetms[dcp_portal].idx = dcp_portal;
-+ INIT_LIST_HEAD(&qman_ceetms[dcp_portal].sub_portals);
-+ INIT_LIST_HEAD(&qman_ceetms[dcp_portal].lnis);
-+
-+ /* Find Sub-portal range */
-+ range = of_get_property(node, "fsl,ceetm-sp-range", &ret);
-+ if (!range) {
-+ pr_err("No fsl,ceetm-sp-range in node %s\n", node->full_name);
-+ return -EINVAL;
-+ }
-+ if (ret != 8) {
-+ pr_err("fsl,ceetm-sp-range is not a 2-cell range in node %s\n",
-+ node->full_name);
-+ return -EINVAL;
-+ }
-+
-+ for (i = 0; i < be32_to_cpu(range[1]); i++) {
-+ sp = kzalloc(sizeof(*sp), GFP_KERNEL);
-+ if (!sp) {
-+ pr_err("Can't alloc memory for sub-portal %d\n",
-+ range[0] + i);
-+ return -ENOMEM;
-+ }
-+ sp->idx = be32_to_cpu(range[0]) + i;
-+ sp->dcp_idx = dcp_portal;
-+ sp->is_claimed = 0;
-+ list_add_tail(&sp->node, &qman_ceetms[dcp_portal].sub_portals);
-+ sp++;
-+ }
-+ pr_debug("Qman: Reserve sub-portal %d:%d for CEETM %d\n",
-+ be32_to_cpu(range[0]), be32_to_cpu(range[1]), dcp_portal);
-+ qman_ceetms[dcp_portal].sp_range[0] = be32_to_cpu(range[0]);
-+ qman_ceetms[dcp_portal].sp_range[1] = be32_to_cpu(range[1]);
-+
-+ /* Find LNI range */
-+ range = of_get_property(node, "fsl,ceetm-lni-range", &ret);
-+ if (!range) {
-+ pr_err("No fsl,ceetm-lni-range in node %s\n", node->full_name);
-+ return -EINVAL;
-+ }
-+ if (ret != 8) {
-+ pr_err("fsl,ceetm-lni-range is not a 2-cell range in node %s\n",
-+ node->full_name);
-+ return -EINVAL;
-+ }
-+
-+ for (i = 0; i < be32_to_cpu(range[1]); i++) {
-+ lni = kzalloc(sizeof(*lni), GFP_KERNEL);
-+ if (!lni) {
-+ pr_err("Can't alloc memory for LNI %d\n",
-+ range[0] + i);
-+ return -ENOMEM;
-+ }
-+ lni->idx = be32_to_cpu(range[0]) + i;
-+ lni->dcp_idx = dcp_portal;
-+ lni->is_claimed = 0;
-+ INIT_LIST_HEAD(&lni->channels);
-+ list_add_tail(&lni->node, &qman_ceetms[dcp_portal].lnis);
-+ lni++;
-+ }
-+ pr_debug("Qman: Reserve LNI %d:%d for CEETM %d\n",
-+ be32_to_cpu(range[0]), be32_to_cpu(range[1]), dcp_portal);
-+ qman_ceetms[dcp_portal].lni_range[0] = be32_to_cpu(range[0]);
-+ qman_ceetms[dcp_portal].lni_range[1] = be32_to_cpu(range[1]);
-+
-+ /* Find CEETM channel range */
-+ range = of_get_property(node, "fsl,ceetm-channel-range", &ret);
-+ if (!range) {
-+ pr_err("No fsl,ceetm-channel-range in node %s\n",
-+ node->full_name);
-+ return -EINVAL;
-+ }
-+ if (ret != 8) {
-+ pr_err("fsl,ceetm-channel-range is not a 2-cell range in node"
-+ "%s\n", node->full_name);
-+ return -EINVAL;
-+ }
-+
-+ if (dcp_portal == qm_dc_portal_fman0)
-+ qman_seed_ceetm0_channel_range(be32_to_cpu(range[0]), be32_to_cpu(range[1]));
-+ if (dcp_portal == qm_dc_portal_fman1)
-+ qman_seed_ceetm1_channel_range(be32_to_cpu(range[0]), be32_to_cpu(range[1]));
-+ pr_debug("Qman: The channel allocator of CEETM %d includes"
-+ " range %d:%d\n", dcp_portal, be32_to_cpu(range[0]), be32_to_cpu(range[1]));
-+
-+ /* Set CEETM PRES register */
-+ ret = qman_ceetm_set_prescaler(dcp_portal);
-+ if (ret)
-+ return ret;
-+ return 0;
-+}
-+
-+static void qman_get_ip_revision(struct device_node *dn)
-+{
-+ u16 ip_rev = 0;
-+ u8 ip_cfg = QMAN_REV_CFG_0;
-+ for_each_compatible_node(dn, NULL, "fsl,qman-portal") {
-+ if (!of_device_is_available(dn))
-+ continue;
-+ if (of_device_is_compatible(dn, "fsl,qman-portal-1.0") ||
-+ of_device_is_compatible(dn, "fsl,qman-portal-1.0.0")) {
-+ pr_err("QMAN rev1.0 on P4080 rev1 is not supported!\n");
-+ BUG_ON(1);
-+ } else if (of_device_is_compatible(dn, "fsl,qman-portal-1.1") ||
-+ of_device_is_compatible(dn, "fsl,qman-portal-1.1.0")) {
-+ ip_rev = QMAN_REV11;
-+ qman_portal_max = 10;
-+ } else if (of_device_is_compatible(dn, "fsl,qman-portal-1.2") ||
-+ of_device_is_compatible(dn, "fsl,qman-portal-1.2.0")) {
-+ ip_rev = QMAN_REV12;
-+ qman_portal_max = 10;
-+ } else if (of_device_is_compatible(dn, "fsl,qman-portal-2.0") ||
-+ of_device_is_compatible(dn, "fsl,qman-portal-2.0.0")) {
-+ ip_rev = QMAN_REV20;
-+ qman_portal_max = 3;
-+ } else if (of_device_is_compatible(dn,
-+ "fsl,qman-portal-3.0.0")) {
-+ ip_rev = QMAN_REV30;
-+ qman_portal_max = 50;
-+ } else if (of_device_is_compatible(dn,
-+ "fsl,qman-portal-3.0.1")) {
-+ ip_rev = QMAN_REV30;
-+ qman_portal_max = 25;
-+ ip_cfg = QMAN_REV_CFG_1;
-+ } else if (of_device_is_compatible(dn,
-+ "fsl,qman-portal-3.1.0")) {
-+ ip_rev = QMAN_REV31;
-+ qman_portal_max = 50;
-+ } else if (of_device_is_compatible(dn,
-+ "fsl,qman-portal-3.1.1")) {
-+ ip_rev = QMAN_REV31;
-+ qman_portal_max = 25;
-+ ip_cfg = QMAN_REV_CFG_1;
-+ } else if (of_device_is_compatible(dn,
-+ "fsl,qman-portal-3.1.2")) {
-+ ip_rev = QMAN_REV31;
-+ qman_portal_max = 18;
-+ ip_cfg = QMAN_REV_CFG_2;
-+ } else if (of_device_is_compatible(dn,
-+ "fsl,qman-portal-3.1.3")) {
-+ ip_rev = QMAN_REV31;
-+ qman_portal_max = 10;
-+ ip_cfg = QMAN_REV_CFG_3;
-+ } else if (of_device_is_compatible(dn,
-+ "fsl,qman-portal-3.2.0")) {
-+ ip_rev = QMAN_REV32;
-+ qman_portal_max = 10;
-+ ip_cfg = QMAN_REV_CFG_3; // TODO: Verify for ls1043
-+ } else if (of_device_is_compatible(dn,
-+ "fsl,qman-portal-3.2.1")) {
-+ ip_rev = QMAN_REV32;
-+ qman_portal_max = 10;
-+ ip_cfg = QMAN_REV_CFG_3;
-+ } else {
-+ pr_warn("unknown QMan version in portal node,"
-+ "default to rev1.1\n");
-+ ip_rev = QMAN_REV11;
-+ qman_portal_max = 10;
-+ }
-+
-+ if (!qman_ip_rev) {
-+ if (ip_rev) {
-+ qman_ip_rev = ip_rev;
-+ qman_ip_cfg = ip_cfg;
-+ } else {
-+ pr_warn("unknown Qman version,"
-+ " default to rev1.1\n");
-+ qman_ip_rev = QMAN_REV11;
-+ qman_ip_cfg = QMAN_REV_CFG_0;
-+ }
-+ } else if (ip_rev && (qman_ip_rev != ip_rev))
-+ pr_warn("Revision=0x%04x, but portal '%s' has"
-+ " 0x%04x\n",
-+ qman_ip_rev, dn->full_name, ip_rev);
-+ if (qman_ip_rev == ip_rev)
-+ break;
-+ }
-+}
-+
-+/* Parse a portal node, perform generic mapping duties and return the config. It
-+ * is not known at this stage for what purpose (or even if) the portal will be
-+ * used. */
-+static struct qm_portal_config * __init parse_pcfg(struct device_node *node)
-+{
-+ struct qm_portal_config *pcfg;
-+ const u32 *index_p;
-+ u32 index, channel;
-+ int irq, ret;
-+ resource_size_t len;
-+
-+ pcfg = kmalloc(sizeof(*pcfg), GFP_KERNEL);
-+ if (!pcfg) {
-+ pr_err("can't allocate portal config");
-+ return NULL;
-+ }
-+
-+ /*
-+ * This is a *horrible hack*, but the IOMMU/PAMU driver needs a
-+ * 'struct device' in order to get the PAMU stashing setup and the QMan
-+ * portal [driver] won't function at all without ring stashing
-+ *
-+ * Making the QMan portal driver nice and proper is part of the
-+ * upstreaming effort
-+ */
-+ pcfg->dev.bus = &platform_bus_type;
-+ pcfg->dev.of_node = node;
-+#ifdef CONFIG_FSL_PAMU
-+ pcfg->dev.archdata.iommu_domain = NULL;
-+#endif
-+
-+ ret = of_address_to_resource(node, DPA_PORTAL_CE,
-+ &pcfg->addr_phys[DPA_PORTAL_CE]);
-+ if (ret) {
-+ pr_err("Can't get %s property '%s'\n", node->full_name,
-+ "reg::CE");
-+ goto err;
-+ }
-+ ret = of_address_to_resource(node, DPA_PORTAL_CI,
-+ &pcfg->addr_phys[DPA_PORTAL_CI]);
-+ if (ret) {
-+ pr_err("Can't get %s property '%s'\n", node->full_name,
-+ "reg::CI");
-+ goto err;
-+ }
-+ index_p = of_get_property(node, "cell-index", &ret);
-+ if (!index_p || (ret != 4)) {
-+ pr_err("Can't get %s property '%s'\n", node->full_name,
-+ "cell-index");
-+ goto err;
-+ }
-+ index = be32_to_cpu(*index_p);
-+ if (index >= qman_portal_max) {
-+ pr_err("QMan portal index %d is beyond max (%d)\n",
-+ index, qman_portal_max);
-+ goto err;
-+ }
-+
-+ channel = index + QM_CHANNEL_SWPORTAL0;
-+ pcfg->public_cfg.channel = channel;
-+ pcfg->public_cfg.cpu = -1;
-+ irq = irq_of_parse_and_map(node, 0);
-+ if (irq == 0) {
-+ pr_err("Can't get %s property '%s'\n", node->full_name,
-+ "interrupts");
-+ goto err;
-+ }
-+ pcfg->public_cfg.irq = irq;
-+ pcfg->public_cfg.index = index;
-+#ifdef CONFIG_FSL_QMAN_CONFIG
-+ /* We need the same LIODN offset for all portals */
-+ qman_liodn_fixup(pcfg->public_cfg.channel);
-+#endif
-+
-+ len = resource_size(&pcfg->addr_phys[DPA_PORTAL_CE]);
-+ if (len != (unsigned long)len)
-+ goto err;
-+
-+#if defined(CONFIG_ARM) || defined(CONFIG_ARM64)
-+ pcfg->addr_virt[DPA_PORTAL_CE] = ioremap_cache_ns(
-+ pcfg->addr_phys[DPA_PORTAL_CE].start,
-+ resource_size(&pcfg->addr_phys[DPA_PORTAL_CE]));
-+
-+ pcfg->addr_virt[DPA_PORTAL_CI] = ioremap(
-+ pcfg->addr_phys[DPA_PORTAL_CI].start,
-+ resource_size(&pcfg->addr_phys[DPA_PORTAL_CI]));
-+#else
-+ pcfg->addr_virt[DPA_PORTAL_CE] = ioremap_prot(
-+ pcfg->addr_phys[DPA_PORTAL_CE].start,
-+ (unsigned long)len,
-+ 0);
-+ pcfg->addr_virt[DPA_PORTAL_CI] = ioremap_prot(
-+ pcfg->addr_phys[DPA_PORTAL_CI].start,
-+ resource_size(&pcfg->addr_phys[DPA_PORTAL_CI]),
-+ _PAGE_GUARDED | _PAGE_NO_CACHE);
-+#endif
-+ return pcfg;
-+err:
-+ kfree(pcfg);
-+ return NULL;
-+}
-+
-+static struct qm_portal_config *get_pcfg(struct list_head *list)
-+{
-+ struct qm_portal_config *pcfg;
-+ if (list_empty(list))
-+ return NULL;
-+ pcfg = list_entry(list->prev, struct qm_portal_config, list);
-+ list_del(&pcfg->list);
-+ return pcfg;
-+}
-+
-+static struct qm_portal_config *get_pcfg_idx(struct list_head *list, u32 idx)
-+{
-+ struct qm_portal_config *pcfg;
-+ if (list_empty(list))
-+ return NULL;
-+ list_for_each_entry(pcfg, list, list) {
-+ if (pcfg->public_cfg.index == idx) {
-+ list_del(&pcfg->list);
-+ return pcfg;
-+ }
-+ }
-+ return NULL;
-+}
-+
-+static void portal_set_cpu(struct qm_portal_config *pcfg, int cpu)
-+{
-+#ifdef CONFIG_FSL_PAMU
-+ int ret;
-+ int window_count = 1;
-+ struct iommu_domain_geometry geom_attr;
-+ struct pamu_stash_attribute stash_attr;
-+
-+ pcfg->iommu_domain = iommu_domain_alloc(&platform_bus_type);
-+ if (!pcfg->iommu_domain) {
-+ pr_err(KBUILD_MODNAME ":%s(): iommu_domain_alloc() failed",
-+ __func__);
-+ goto _no_iommu;
-+ }
-+ geom_attr.aperture_start = 0;
-+ geom_attr.aperture_end =
-+ ((dma_addr_t)1 << min(8 * sizeof(dma_addr_t), (size_t)36)) - 1;
-+ geom_attr.force_aperture = true;
-+ ret = iommu_domain_set_attr(pcfg->iommu_domain, DOMAIN_ATTR_GEOMETRY,
-+ &geom_attr);
-+ if (ret < 0) {
-+ pr_err(KBUILD_MODNAME ":%s(): iommu_domain_set_attr() = %d",
-+ __func__, ret);
-+ goto _iommu_domain_free;
-+ }
-+ ret = iommu_domain_set_attr(pcfg->iommu_domain, DOMAIN_ATTR_WINDOWS,
-+ &window_count);
-+ if (ret < 0) {
-+ pr_err(KBUILD_MODNAME ":%s(): iommu_domain_set_attr() = %d",
-+ __func__, ret);
-+ goto _iommu_domain_free;
-+ }
-+ stash_attr.cpu = cpu;
-+ stash_attr.cache = PAMU_ATTR_CACHE_L1;
-+ /* set stash information for the window */
-+ stash_attr.window = 0;
-+ ret = iommu_domain_set_attr(pcfg->iommu_domain,
-+ DOMAIN_ATTR_FSL_PAMU_STASH,
-+ &stash_attr);
-+ if (ret < 0) {
-+ pr_err(KBUILD_MODNAME ":%s(): iommu_domain_set_attr() = %d",
-+ __func__, ret);
-+ goto _iommu_domain_free;
-+ }
-+ ret = iommu_domain_window_enable(pcfg->iommu_domain, 0, 0, 1ULL << 36,
-+ IOMMU_READ | IOMMU_WRITE);
-+ if (ret < 0) {
-+ pr_err(KBUILD_MODNAME ":%s(): iommu_domain_window_enable() = %d",
-+ __func__, ret);
-+ goto _iommu_domain_free;
-+ }
-+ ret = iommu_attach_device(pcfg->iommu_domain, &pcfg->dev);
-+ if (ret < 0) {
-+ pr_err(KBUILD_MODNAME ":%s(): iommu_device_attach() = %d",
-+ __func__, ret);
-+ goto _iommu_domain_free;
-+ }
-+ ret = iommu_domain_set_attr(pcfg->iommu_domain,
-+ DOMAIN_ATTR_FSL_PAMU_ENABLE,
-+ &window_count);
-+ if (ret < 0) {
-+ pr_err(KBUILD_MODNAME ":%s(): iommu_domain_set_attr() = %d",
-+ __func__, ret);
-+ goto _iommu_detach_device;
-+ }
-+
-+_no_iommu:
-+#endif
-+#ifdef CONFIG_FSL_QMAN_CONFIG
-+ if (qman_set_sdest(pcfg->public_cfg.channel, cpu))
-+#endif
-+ pr_warn("Failed to set QMan portal's stash request queue\n");
-+
-+ return;
-+
-+#ifdef CONFIG_FSL_PAMU
-+_iommu_detach_device:
-+ iommu_detach_device(pcfg->iommu_domain, NULL);
-+_iommu_domain_free:
-+ iommu_domain_free(pcfg->iommu_domain);
-+#endif
-+}
-+
-+struct qm_portal_config *qm_get_unused_portal_idx(u32 idx)
-+{
-+ struct qm_portal_config *ret;
-+ spin_lock(&unused_pcfgs_lock);
-+ if (idx == QBMAN_ANY_PORTAL_IDX)
-+ ret = get_pcfg(&unused_pcfgs);
-+ else
-+ ret = get_pcfg_idx(&unused_pcfgs, idx);
-+ spin_unlock(&unused_pcfgs_lock);
-+ /* Bind stashing LIODNs to the CPU we are currently executing on, and
-+ * set the portal to use the stashing request queue corresonding to the
-+ * cpu as well. The user-space driver assumption is that the pthread has
-+ * to already be affine to one cpu only before opening a portal. If that
-+ * check is circumvented, the only risk is a performance degradation -
-+ * stashing will go to whatever cpu they happened to be running on when
-+ * opening the device file, and if that isn't the cpu they subsequently
-+ * bind to and do their polling on, tough. */
-+ if (ret)
-+ portal_set_cpu(ret, hard_smp_processor_id());
-+ return ret;
-+}
-+
-+struct qm_portal_config *qm_get_unused_portal(void)
-+{
-+ return qm_get_unused_portal_idx(QBMAN_ANY_PORTAL_IDX);
-+}
-+
-+void qm_put_unused_portal(struct qm_portal_config *pcfg)
-+{
-+ spin_lock(&unused_pcfgs_lock);
-+ list_add(&pcfg->list, &unused_pcfgs);
-+ spin_unlock(&unused_pcfgs_lock);
-+}
-+
-+static struct qman_portal *init_pcfg(struct qm_portal_config *pcfg)
-+{
-+ struct qman_portal *p;
-+
-+ pcfg->iommu_domain = NULL;
-+ portal_set_cpu(pcfg, pcfg->public_cfg.cpu);
-+ p = qman_create_affine_portal(pcfg, NULL);
-+ if (p) {
-+ u32 irq_sources = 0;
-+ /* Determine what should be interrupt-vs-poll driven */
-+#ifdef CONFIG_FSL_DPA_PIRQ_SLOW
-+ irq_sources |= QM_PIRQ_EQCI | QM_PIRQ_EQRI | QM_PIRQ_MRI |
-+ QM_PIRQ_CSCI | QM_PIRQ_CCSCI;
-+#endif
-+#ifdef CONFIG_FSL_DPA_PIRQ_FAST
-+ irq_sources |= QM_PIRQ_DQRI;
-+#endif
-+ qman_p_irqsource_add(p, irq_sources);
-+ pr_info("Qman portal %sinitialised, cpu %d\n",
-+ pcfg->public_cfg.is_shared ? "(shared) " : "",
-+ pcfg->public_cfg.cpu);
-+ } else
-+ pr_crit("Qman portal failure on cpu %d\n",
-+ pcfg->public_cfg.cpu);
-+ return p;
-+}
-+
-+static void init_slave(int cpu)
-+{
-+ struct qman_portal *p;
-+ struct cpumask oldmask = current->cpus_allowed;
-+ set_cpus_allowed_ptr(current, get_cpu_mask(cpu));
-+ p = qman_create_affine_slave(shared_portals[shared_portals_idx++], cpu);
-+ if (!p)
-+ pr_err("Qman slave portal failure on cpu %d\n", cpu);
-+ else
-+ pr_info("Qman portal %sinitialised, cpu %d\n", "(slave) ", cpu);
-+ set_cpus_allowed_ptr(current, &oldmask);
-+ if (shared_portals_idx >= num_shared_portals)
-+ shared_portals_idx = 0;
-+}
-+
-+static struct cpumask want_unshared __initdata;
-+static struct cpumask want_shared __initdata;
-+
-+static int __init parse_qportals(char *str)
-+{
-+ return parse_portals_bootarg(str, &want_shared, &want_unshared,
-+ "qportals");
-+}
-+__setup("qportals=", parse_qportals);
-+
-+static void qman_portal_update_sdest(const struct qm_portal_config *pcfg,
-+ unsigned int cpu)
-+{
-+#ifdef CONFIG_FSL_PAMU
-+ struct pamu_stash_attribute stash_attr;
-+ int ret;
-+
-+ if (pcfg->iommu_domain) {
-+ stash_attr.cpu = cpu;
-+ stash_attr.cache = PAMU_ATTR_CACHE_L1;
-+ /* set stash information for the window */
-+ stash_attr.window = 0;
-+ ret = iommu_domain_set_attr(pcfg->iommu_domain,
-+ DOMAIN_ATTR_FSL_PAMU_STASH, &stash_attr);
-+ if (ret < 0) {
-+ pr_err("Failed to update pamu stash setting\n");
-+ return;
-+ }
-+ }
-+#endif
-+#ifdef CONFIG_FSL_QMAN_CONFIG
-+ if (qman_set_sdest(pcfg->public_cfg.channel, cpu))
-+ pr_warn("Failed to update portal's stash request queue\n");
-+#endif
-+}
-+
-+static int qman_offline_cpu(unsigned int cpu)
-+{
-+ struct qman_portal *p;
-+ const struct qm_portal_config *pcfg;
-+ p = (struct qman_portal *)affine_portals[cpu];
-+ if (p) {
-+ pcfg = qman_get_qm_portal_config(p);
-+ if (pcfg) {
-+ irq_set_affinity(pcfg->public_cfg.irq, cpumask_of(0));
-+ qman_portal_update_sdest(pcfg, 0);
-+ }
-+ }
-+ return 0;
-+}
-+
-+#ifdef CONFIG_HOTPLUG_CPU
-+static int qman_online_cpu(unsigned int cpu)
-+{
-+ struct qman_portal *p;
-+ const struct qm_portal_config *pcfg;
-+ p = (struct qman_portal *)affine_portals[cpu];
-+ if (p) {
-+ pcfg = qman_get_qm_portal_config(p);
-+ if (pcfg) {
-+ irq_set_affinity(pcfg->public_cfg.irq, cpumask_of(cpu));
-+ qman_portal_update_sdest(pcfg, cpu);
-+ }
-+ }
-+ return 0;
-+}
-+
-+static int qman_hotplug_cpu_callback(struct notifier_block *nfb,
-+ unsigned long action, void *hcpu)
-+{
-+ unsigned int cpu = (unsigned long)hcpu;
-+
-+ switch (action) {
-+ case CPU_ONLINE:
-+ case CPU_ONLINE_FROZEN:
-+ qman_online_cpu(cpu);
-+ break;
-+ case CPU_DOWN_PREPARE:
-+ case CPU_DOWN_PREPARE_FROZEN:
-+ qman_offline_cpu(cpu);
-+ default:
-+ break;
-+ }
-+ return NOTIFY_OK;
-+}
-+
-+static struct notifier_block qman_hotplug_cpu_notifier = {
-+ .notifier_call = qman_hotplug_cpu_callback,
-+};
-+#endif /* CONFIG_HOTPLUG_CPU */
-+
-+__init int qman_init(void)
-+{
-+ struct cpumask slave_cpus;
-+ struct cpumask unshared_cpus = *cpu_none_mask;
-+ struct cpumask shared_cpus = *cpu_none_mask;
-+ LIST_HEAD(unshared_pcfgs);
-+ LIST_HEAD(shared_pcfgs);
-+ struct device_node *dn;
-+ struct qm_portal_config *pcfg;
-+ struct qman_portal *p;
-+ int cpu, ret;
-+ const u32 *clk;
-+ struct cpumask offline_cpus;
-+
-+ /* Initialise the Qman (CCSR) device */
-+ for_each_compatible_node(dn, NULL, "fsl,qman") {
-+ if (!qman_init_ccsr(dn))
-+ pr_info("Qman err interrupt handler present\n");
-+ else
-+ pr_err("Qman CCSR setup failed\n");
-+
-+ clk = of_get_property(dn, "clock-frequency", NULL);
-+ if (!clk)
-+ pr_warn("Can't find Qman clock frequency\n");
-+ else
-+ qman_clk = be32_to_cpu(*clk);
-+ }
-+#ifdef CONFIG_FSL_QMAN_FQ_LOOKUP
-+ /* Setup lookup table for FQ demux */
-+ ret = qman_setup_fq_lookup_table(get_qman_fqd_size()/64);
-+ if (ret)
-+ return ret;
-+#endif
-+
-+ /* Get qman ip revision */
-+ qman_get_ip_revision(dn);
-+ if ((qman_ip_rev & 0xff00) >= QMAN_REV30) {
-+ qm_channel_pool1 = QMAN_CHANNEL_POOL1_REV3;
-+ qm_channel_caam = QMAN_CHANNEL_CAAM_REV3;
-+ qm_channel_pme = QMAN_CHANNEL_PME_REV3;
-+ }
-+
-+ if ((qman_ip_rev == QMAN_REV31) && (qman_ip_cfg == QMAN_REV_CFG_2))
-+ qm_channel_dce = QMAN_CHANNEL_DCE_QMANREV312;
-+
-+ /*
-+ * Parse the ceetm node to get how many ceetm instances are supported
-+ * on the current silicon. num_ceetms must be confirmed before portals
-+ * are intiailized.
-+ */
-+ num_ceetms = 0;
-+ for_each_compatible_node(dn, NULL, "fsl,qman-ceetm")
-+ num_ceetms++;
-+
-+ /* Parse pool channels into the SDQCR mask. (Must happen before portals
-+ * are initialised.) */
-+ for_each_compatible_node(dn, NULL, "fsl,pool-channel-range") {
-+ ret = fsl_pool_channel_range_sdqcr(dn);
-+ if (ret)
-+ return ret;
-+ }
-+
-+ memset(affine_portals, 0, sizeof(void *) * num_possible_cpus());
-+ /* Initialise portals. See bman_driver.c for comments */
-+ for_each_compatible_node(dn, NULL, "fsl,qman-portal") {
-+ if (!of_device_is_available(dn))
-+ continue;
-+ pcfg = parse_pcfg(dn);
-+ if (pcfg) {
-+ pcfg->public_cfg.pools = pools_sdqcr;
-+ list_add_tail(&pcfg->list, &unused_pcfgs);
-+ }
-+ }
-+ for_each_possible_cpu(cpu) {
-+ if (cpumask_test_cpu(cpu, &want_shared)) {
-+ pcfg = get_pcfg(&unused_pcfgs);
-+ if (!pcfg)
-+ break;
-+ pcfg->public_cfg.cpu = cpu;
-+ list_add_tail(&pcfg->list, &shared_pcfgs);
-+ cpumask_set_cpu(cpu, &shared_cpus);
-+ }
-+ if (cpumask_test_cpu(cpu, &want_unshared)) {
-+ if (cpumask_test_cpu(cpu, &shared_cpus))
-+ continue;
-+ pcfg = get_pcfg(&unused_pcfgs);
-+ if (!pcfg)
-+ break;
-+ pcfg->public_cfg.cpu = cpu;
-+ list_add_tail(&pcfg->list, &unshared_pcfgs);
-+ cpumask_set_cpu(cpu, &unshared_cpus);
-+ }
-+ }
-+ if (list_empty(&shared_pcfgs) && list_empty(&unshared_pcfgs)) {
-+ for_each_online_cpu(cpu) {
-+ pcfg = get_pcfg(&unused_pcfgs);
-+ if (!pcfg)
-+ break;
-+ pcfg->public_cfg.cpu = cpu;
-+ list_add_tail(&pcfg->list, &unshared_pcfgs);
-+ cpumask_set_cpu(cpu, &unshared_cpus);
-+ }
-+ }
-+ cpumask_andnot(&slave_cpus, cpu_possible_mask, &shared_cpus);
-+ cpumask_andnot(&slave_cpus, &slave_cpus, &unshared_cpus);
-+ if (cpumask_empty(&slave_cpus)) {
-+ if (!list_empty(&shared_pcfgs)) {
-+ cpumask_or(&unshared_cpus, &unshared_cpus,
-+ &shared_cpus);
-+ cpumask_clear(&shared_cpus);
-+ list_splice_tail(&shared_pcfgs, &unshared_pcfgs);
-+ INIT_LIST_HEAD(&shared_pcfgs);
-+ }
-+ } else {
-+ if (list_empty(&shared_pcfgs)) {
-+ pcfg = get_pcfg(&unshared_pcfgs);
-+ if (!pcfg) {
-+ pr_crit("No QMan portals available!\n");
-+ return 0;
-+ }
-+ cpumask_clear_cpu(pcfg->public_cfg.cpu, &unshared_cpus);
-+ cpumask_set_cpu(pcfg->public_cfg.cpu, &shared_cpus);
-+ list_add_tail(&pcfg->list, &shared_pcfgs);
-+ }
-+ }
-+ list_for_each_entry(pcfg, &unshared_pcfgs, list) {
-+ pcfg->public_cfg.is_shared = 0;
-+ p = init_pcfg(pcfg);
-+ if (!p) {
-+ pr_crit("Unable to configure portals\n");
-+ return 0;
-+ }
-+ }
-+ list_for_each_entry(pcfg, &shared_pcfgs, list) {
-+ pcfg->public_cfg.is_shared = 1;
-+ p = init_pcfg(pcfg);
-+ if (p)
-+ shared_portals[num_shared_portals++] = p;
-+ }
-+ if (!cpumask_empty(&slave_cpus))
-+ for_each_cpu(cpu, &slave_cpus)
-+ init_slave(cpu);
-+ pr_info("Qman portals initialised\n");
-+ cpumask_andnot(&offline_cpus, cpu_possible_mask, cpu_online_mask);
-+ for_each_cpu(cpu, &offline_cpus)
-+ qman_offline_cpu(cpu);
-+#ifdef CONFIG_HOTPLUG_CPU
-+ register_hotcpu_notifier(&qman_hotplug_cpu_notifier);
-+#endif
-+ return 0;
-+}
-+
-+__init int qman_resource_init(void)
-+{
-+ struct device_node *dn;
-+ int ret;
-+
-+ /* Initialise FQID allocation ranges */
-+ for_each_compatible_node(dn, NULL, "fsl,fqid-range") {
-+ ret = fsl_fqid_range_init(dn);
-+ if (ret)
-+ return ret;
-+ }
-+ /* Initialise CGRID allocation ranges */
-+ for_each_compatible_node(dn, NULL, "fsl,cgrid-range") {
-+ ret = fsl_cgrid_range_init(dn);
-+ if (ret)
-+ return ret;
-+ }
-+ /* Parse pool channels into the allocator. (Must happen after portals
-+ * are initialised.) */
-+ for_each_compatible_node(dn, NULL, "fsl,pool-channel-range") {
-+ ret = fsl_pool_channel_range_init(dn);
-+ if (ret)
-+ return ret;
-+ }
-+
-+ /* Parse CEETM */
-+ for_each_compatible_node(dn, NULL, "fsl,qman-ceetm") {
-+ ret = fsl_ceetm_init(dn);
-+ if (ret)
-+ return ret;
-+ }
-+ return 0;
-+}
-+
-+#ifdef CONFIG_SUSPEND
-+void suspend_unused_qportal(void)
-+{
-+ struct qm_portal_config *pcfg;
-+
-+ if (list_empty(&unused_pcfgs))
-+ return;
-+
-+ list_for_each_entry(pcfg, &unused_pcfgs, list) {
-+#ifdef CONFIG_PM_DEBUG
-+ pr_info("Need to save qportal %d\n", pcfg->public_cfg.index);
-+#endif
-+ /* save isdr, disable all via isdr, clear isr */
-+ pcfg->saved_isdr =
-+ __raw_readl(pcfg->addr_virt[DPA_PORTAL_CI] + 0xe08);
-+ __raw_writel(0xffffffff, pcfg->addr_virt[DPA_PORTAL_CI] +
-+ 0xe08);
-+ __raw_writel(0xffffffff, pcfg->addr_virt[DPA_PORTAL_CI] +
-+ 0xe00);
-+ }
-+ return;
-+}
-+
-+void resume_unused_qportal(void)
-+{
-+ struct qm_portal_config *pcfg;
-+
-+ if (list_empty(&unused_pcfgs))
-+ return;
-+
-+ list_for_each_entry(pcfg, &unused_pcfgs, list) {
-+#ifdef CONFIG_PM_DEBUG
-+ pr_info("Need to resume qportal %d\n", pcfg->public_cfg.index);
-+#endif
-+ /* restore isdr */
-+ __raw_writel(pcfg->saved_isdr,
-+ pcfg->addr_virt[DPA_PORTAL_CI] + 0xe08);
-+ }
-+ return;
-+}
-+#endif
---- /dev/null
-+++ b/drivers/staging/fsl_qbman/qman_high.c
-@@ -0,0 +1,5652 @@
-+/* Copyright 2008-2012 Freescale Semiconductor, Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#include "qman_low.h"
-+
-+/* Compilation constants */
-+#define DQRR_MAXFILL 15
-+#define EQCR_ITHRESH 4 /* if EQCR congests, interrupt threshold */
-+#define IRQNAME "QMan portal %d"
-+#define MAX_IRQNAME 16 /* big enough for "QMan portal %d" */
-+
-+/* Divide 'n' by 'd', rounding down if 'r' is negative, rounding up if it's
-+ * positive, and rounding to the closest value if it's zero. NB, this macro
-+ * implicitly upgrades parameters to unsigned 64-bit, so feed it with types
-+ * that are compatible with this. NB, these arguments should not be expressions
-+ * unless it is safe for them to be evaluated multiple times. Eg. do not pass
-+ * in "some_value++" as a parameter to the macro! */
-+#define ROUNDING(n, d, r) \
-+ (((r) < 0) ? div64_u64((n), (d)) : \
-+ (((r) > 0) ? div64_u64(((n) + (d) - 1), (d)) : \
-+ div64_u64(((n) + ((d) / 2)), (d))))
-+
-+/* Lock/unlock frame queues, subject to the "LOCKED" flag. This is about
-+ * inter-processor locking only. Note, FQLOCK() is always called either under a
-+ * local_irq_save() or from interrupt context - hence there's no need for irq
-+ * protection (and indeed, attempting to nest irq-protection doesn't work, as
-+ * the "irq en/disable" machinery isn't recursive...). */
-+#define FQLOCK(fq) \
-+ do { \
-+ struct qman_fq *__fq478 = (fq); \
-+ if (fq_isset(__fq478, QMAN_FQ_FLAG_LOCKED)) \
-+ spin_lock(&__fq478->fqlock); \
-+ } while (0)
-+#define FQUNLOCK(fq) \
-+ do { \
-+ struct qman_fq *__fq478 = (fq); \
-+ if (fq_isset(__fq478, QMAN_FQ_FLAG_LOCKED)) \
-+ spin_unlock(&__fq478->fqlock); \
-+ } while (0)
-+
-+static inline void fq_set(struct qman_fq *fq, u32 mask)
-+{
-+ set_bits(mask, &fq->flags);
-+}
-+static inline void fq_clear(struct qman_fq *fq, u32 mask)
-+{
-+ clear_bits(mask, &fq->flags);
-+}
-+static inline int fq_isset(struct qman_fq *fq, u32 mask)
-+{
-+ return fq->flags & mask;
-+}
-+static inline int fq_isclear(struct qman_fq *fq, u32 mask)
-+{
-+ return !(fq->flags & mask);
-+}
-+
-+struct qman_portal {
-+ struct qm_portal p;
-+ unsigned long bits; /* PORTAL_BITS_*** - dynamic, strictly internal */
-+ unsigned long irq_sources;
-+ u32 use_eqcr_ci_stashing;
-+ u32 slowpoll; /* only used when interrupts are off */
-+ struct qman_fq *vdqcr_owned; /* only 1 volatile dequeue at a time */
-+#ifdef CONFIG_FSL_DPA_CAN_WAIT_SYNC
-+ struct qman_fq *eqci_owned; /* only 1 enqueue WAIT_SYNC at a time */
-+#endif
-+#ifdef CONFIG_FSL_DPA_PORTAL_SHARE
-+ raw_spinlock_t sharing_lock; /* only used if is_shared */
-+ int is_shared;
-+ struct qman_portal *sharing_redirect;
-+#endif
-+ u32 sdqcr;
-+ int dqrr_disable_ref;
-+ /* A portal-specific handler for DCP ERNs. If this is NULL, the global
-+ * handler is called instead. */
-+ qman_cb_dc_ern cb_dc_ern;
-+ /* When the cpu-affine portal is activated, this is non-NULL */
-+ const struct qm_portal_config *config;
-+ /* This is needed for providing a non-NULL device to dma_map_***() */
-+ struct platform_device *pdev;
-+ struct dpa_rbtree retire_table;
-+ char irqname[MAX_IRQNAME];
-+ /* 2-element array. cgrs[0] is mask, cgrs[1] is snapshot. */
-+ struct qman_cgrs *cgrs;
-+ /* linked-list of CSCN handlers. */
-+ struct list_head cgr_cbs;
-+ /* list lock */
-+ spinlock_t cgr_lock;
-+ /* 2-element array. ccgrs[0] is mask, ccgrs[1] is snapshot. */
-+ struct qman_ccgrs *ccgrs[QMAN_CEETM_MAX];
-+ /* 256-element array, each is a linked-list of CCSCN handlers. */
-+ struct list_head ccgr_cbs[QMAN_CEETM_MAX];
-+ /* list lock */
-+ spinlock_t ccgr_lock;
-+ /* track if memory was allocated by the driver */
-+ u8 alloced;
-+ /* power management data */
-+ u32 save_isdr;
-+#if __BYTE_ORDER__ == __ORDER_LITTLE_ENDIAN__
-+ /* Keep a shadow copy of the DQRR on LE systems as the SW needs to
-+ * do byte swaps of DQRR read only memory. First entry must be aligned
-+ * to 2 ** 10 to ensure DQRR index calculations based shadow copy
-+ * address (6 bits for address shift + 4 bits for the DQRR size).
-+ */
-+ struct qm_dqrr_entry shadow_dqrr[QM_DQRR_SIZE] __aligned(1024);
-+#endif
-+};
-+
-+#ifdef CONFIG_FSL_DPA_PORTAL_SHARE
-+#define PORTAL_IRQ_LOCK(p, irqflags) \
-+ do { \
-+ if ((p)->is_shared) \
-+ raw_spin_lock_irqsave(&(p)->sharing_lock, irqflags); \
-+ else \
-+ local_irq_save(irqflags); \
-+ } while (0)
-+#define PORTAL_IRQ_UNLOCK(p, irqflags) \
-+ do { \
-+ if ((p)->is_shared) \
-+ raw_spin_unlock_irqrestore(&(p)->sharing_lock, \
-+ irqflags); \
-+ else \
-+ local_irq_restore(irqflags); \
-+ } while (0)
-+#else
-+#define PORTAL_IRQ_LOCK(p, irqflags) local_irq_save(irqflags)
-+#define PORTAL_IRQ_UNLOCK(p, irqflags) local_irq_restore(irqflags)
-+#endif
-+
-+/* Global handler for DCP ERNs. Used when the portal receiving the message does
-+ * not have a portal-specific handler. */
-+static qman_cb_dc_ern cb_dc_ern;
-+
-+static cpumask_t affine_mask;
-+static DEFINE_SPINLOCK(affine_mask_lock);
-+static u16 affine_channels[NR_CPUS];
-+static DEFINE_PER_CPU(struct qman_portal, qman_affine_portal);
-+void *affine_portals[NR_CPUS];
-+
-+/* "raw" gets the cpu-local struct whether it's a redirect or not. */
-+static inline struct qman_portal *get_raw_affine_portal(void)
-+{
-+ return &get_cpu_var(qman_affine_portal);
-+}
-+/* For ops that can redirect, this obtains the portal to use */
-+#ifdef CONFIG_FSL_DPA_PORTAL_SHARE
-+static inline struct qman_portal *get_affine_portal(void)
-+{
-+ struct qman_portal *p = get_raw_affine_portal();
-+ if (p->sharing_redirect)
-+ return p->sharing_redirect;
-+ return p;
-+}
-+#else
-+#define get_affine_portal() get_raw_affine_portal()
-+#endif
-+/* For every "get", there must be a "put" */
-+static inline void put_affine_portal(void)
-+{
-+ put_cpu_var(qman_affine_portal);
-+}
-+/* Exception: poll functions assume the caller is cpu-affine and in no risk of
-+ * re-entrance, which are the two reasons we usually use the get/put_cpu_var()
-+ * semantic - ie. to disable pre-emption. Some use-cases expect the execution
-+ * context to remain as non-atomic during poll-triggered callbacks as it was
-+ * when the poll API was first called (eg. NAPI), so we go out of our way in
-+ * this case to not disable pre-emption. */
-+static inline struct qman_portal *get_poll_portal(void)
-+{
-+ return &get_cpu_var(qman_affine_portal);
-+}
-+#define put_poll_portal()
-+
-+/* This gives a FQID->FQ lookup to cover the fact that we can't directly demux
-+ * retirement notifications (the fact they are sometimes h/w-consumed means that
-+ * contextB isn't always a s/w demux - and as we can't know which case it is
-+ * when looking at the notification, we have to use the slow lookup for all of
-+ * them). NB, it's possible to have multiple FQ objects refer to the same FQID
-+ * (though at most one of them should be the consumer), so this table isn't for
-+ * all FQs - FQs are added when retirement commands are issued, and removed when
-+ * they complete, which also massively reduces the size of this table. */
-+IMPLEMENT_DPA_RBTREE(fqtree, struct qman_fq, node, fqid);
-+
-+/* This is what everything can wait on, even if it migrates to a different cpu
-+ * to the one whose affine portal it is waiting on. */
-+static DECLARE_WAIT_QUEUE_HEAD(affine_queue);
-+
-+static inline int table_push_fq(struct qman_portal *p, struct qman_fq *fq)
-+{
-+ int ret = fqtree_push(&p->retire_table, fq);
-+ if (ret)
-+ pr_err("ERROR: double FQ-retirement %d\n", fq->fqid);
-+ return ret;
-+}
-+
-+static inline void table_del_fq(struct qman_portal *p, struct qman_fq *fq)
-+{
-+ fqtree_del(&p->retire_table, fq);
-+}
-+
-+static inline struct qman_fq *table_find_fq(struct qman_portal *p, u32 fqid)
-+{
-+ return fqtree_find(&p->retire_table, fqid);
-+}
-+
-+#ifdef CONFIG_FSL_QMAN_FQ_LOOKUP
-+static void **qman_fq_lookup_table;
-+static size_t qman_fq_lookup_table_size;
-+
-+int qman_setup_fq_lookup_table(size_t num_entries)
-+{
-+ num_entries++;
-+ /* Allocate 1 more entry since the first entry is not used */
-+ qman_fq_lookup_table = vzalloc((num_entries * sizeof(void *)));
-+ if (!qman_fq_lookup_table) {
-+ pr_err("QMan: Could not allocate fq lookup table\n");
-+ return -ENOMEM;
-+ }
-+ qman_fq_lookup_table_size = num_entries;
-+ pr_info("QMan: Allocated lookup table at %p, entry count %lu\n",
-+ qman_fq_lookup_table,
-+ (unsigned long)qman_fq_lookup_table_size);
-+ return 0;
-+}
-+
-+/* global structure that maintains fq object mapping */
-+static DEFINE_SPINLOCK(fq_hash_table_lock);
-+
-+static int find_empty_fq_table_entry(u32 *entry, struct qman_fq *fq)
-+{
-+ u32 i;
-+
-+ spin_lock(&fq_hash_table_lock);
-+ /* Can't use index zero because this has special meaning
-+ * in context_b field. */
-+ for (i = 1; i < qman_fq_lookup_table_size; i++) {
-+ if (qman_fq_lookup_table[i] == NULL) {
-+ *entry = i;
-+ qman_fq_lookup_table[i] = fq;
-+ spin_unlock(&fq_hash_table_lock);
-+ return 0;
-+ }
-+ }
-+ spin_unlock(&fq_hash_table_lock);
-+ return -ENOMEM;
-+}
-+
-+static void clear_fq_table_entry(u32 entry)
-+{
-+ spin_lock(&fq_hash_table_lock);
-+ BUG_ON(entry >= qman_fq_lookup_table_size);
-+ qman_fq_lookup_table[entry] = NULL;
-+ spin_unlock(&fq_hash_table_lock);
-+}
-+
-+static inline struct qman_fq *get_fq_table_entry(u32 entry)
-+{
-+ BUG_ON(entry >= qman_fq_lookup_table_size);
-+ return qman_fq_lookup_table[entry];
-+}
-+#endif
-+
-+static inline void cpu_to_hw_fqd(struct qm_fqd *fqd)
-+{
-+ /* Byteswap the FQD to HW format */
-+ fqd->fq_ctrl = cpu_to_be16(fqd->fq_ctrl);
-+ fqd->dest_wq = cpu_to_be16(fqd->dest_wq);
-+ fqd->ics_cred = cpu_to_be16(fqd->ics_cred);
-+ fqd->context_b = cpu_to_be32(fqd->context_b);
-+ fqd->context_a.opaque = cpu_to_be64(fqd->context_a.opaque);
-+}
-+
-+static inline void hw_fqd_to_cpu(struct qm_fqd *fqd)
-+{
-+ /* Byteswap the FQD to CPU format */
-+ fqd->fq_ctrl = be16_to_cpu(fqd->fq_ctrl);
-+ fqd->dest_wq = be16_to_cpu(fqd->dest_wq);
-+ fqd->ics_cred = be16_to_cpu(fqd->ics_cred);
-+ fqd->context_b = be32_to_cpu(fqd->context_b);
-+ fqd->context_a.opaque = be64_to_cpu(fqd->context_a.opaque);
-+}
-+
-+/* Swap a 40 bit address */
-+static inline u64 cpu_to_be40(u64 in)
-+{
-+#if __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
-+ return in;
-+#else
-+ u64 out = 0;
-+ u8 *p = (u8 *) &out;
-+ p[0] = in >> 32;
-+ p[1] = in >> 24;
-+ p[2] = in >> 16;
-+ p[3] = in >> 8;
-+ p[4] = in >> 0;
-+ return out;
-+#endif
-+}
-+static inline u64 be40_to_cpu(u64 in)
-+{
-+#if __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
-+ return in;
-+#else
-+ u64 out = 0;
-+ u8 *pout = (u8 *) &out;
-+ u8 *pin = (u8 *) &in;
-+ pout[0] = pin[4];
-+ pout[1] = pin[3];
-+ pout[2] = pin[2];
-+ pout[3] = pin[1];
-+ pout[4] = pin[0];
-+ return out;
-+#endif
-+}
-+
-+/* Swap a 24 bit value */
-+static inline u32 cpu_to_be24(u32 in)
-+{
-+#if __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
-+ return in;
-+#else
-+ u32 out = 0;
-+ u8 *p = (u8 *) &out;
-+ p[0] = in >> 16;
-+ p[1] = in >> 8;
-+ p[2] = in >> 0;
-+ return out;
-+#endif
-+}
-+
-+static inline u32 be24_to_cpu(u32 in)
-+{
-+#if __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
-+ return in;
-+#else
-+ u32 out = 0;
-+ u8 *pout = (u8 *) &out;
-+ u8 *pin = (u8 *) &in;
-+ pout[0] = pin[2];
-+ pout[1] = pin[1];
-+ pout[2] = pin[0];
-+ return out;
-+#endif
-+}
-+
-+static inline u64 be48_to_cpu(u64 in)
-+{
-+#if __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
-+ return in;
-+#else
-+ u64 out = 0;
-+ u8 *pout = (u8 *) &out;
-+ u8 *pin = (u8 *) &in;
-+
-+ pout[0] = pin[5];
-+ pout[1] = pin[4];
-+ pout[2] = pin[3];
-+ pout[3] = pin[2];
-+ pout[4] = pin[1];
-+ pout[5] = pin[0];
-+ return out;
-+#endif
-+}
-+static inline void cpu_to_hw_fd(struct qm_fd *fd)
-+{
-+ fd->opaque_addr = cpu_to_be64(fd->opaque_addr);
-+ fd->status = cpu_to_be32(fd->status);
-+ fd->opaque = cpu_to_be32(fd->opaque);
-+}
-+
-+static inline void hw_fd_to_cpu(struct qm_fd *fd)
-+{
-+ fd->opaque_addr = be64_to_cpu(fd->opaque_addr);
-+ fd->status = be32_to_cpu(fd->status);
-+ fd->opaque = be32_to_cpu(fd->opaque);
-+}
-+
-+static inline void hw_cq_query_to_cpu(struct qm_mcr_ceetm_cq_query *cq_query)
-+{
-+ cq_query->ccgid = be16_to_cpu(cq_query->ccgid);
-+ cq_query->state = be16_to_cpu(cq_query->state);
-+ cq_query->pfdr_hptr = be24_to_cpu(cq_query->pfdr_hptr);
-+ cq_query->pfdr_tptr = be24_to_cpu(cq_query->pfdr_tptr);
-+ cq_query->od1_xsfdr = be16_to_cpu(cq_query->od1_xsfdr);
-+ cq_query->od2_xsfdr = be16_to_cpu(cq_query->od2_xsfdr);
-+ cq_query->od3_xsfdr = be16_to_cpu(cq_query->od3_xsfdr);
-+ cq_query->od4_xsfdr = be16_to_cpu(cq_query->od4_xsfdr);
-+ cq_query->od5_xsfdr = be16_to_cpu(cq_query->od5_xsfdr);
-+ cq_query->od6_xsfdr = be16_to_cpu(cq_query->od6_xsfdr);
-+ cq_query->ra1_xsfdr = be16_to_cpu(cq_query->ra1_xsfdr);
-+ cq_query->ra2_xsfdr = be16_to_cpu(cq_query->ra2_xsfdr);
-+ cq_query->frm_cnt = be24_to_cpu(cq_query->frm_cnt);
-+}
-+
-+static inline void hw_ccgr_query_to_cpu(struct qm_mcr_ceetm_ccgr_query *ccgr_q)
-+{
-+ int i;
-+
-+ ccgr_q->cm_query.cs_thres.hword =
-+ be16_to_cpu(ccgr_q->cm_query.cs_thres.hword);
-+ ccgr_q->cm_query.cs_thres_x.hword =
-+ be16_to_cpu(ccgr_q->cm_query.cs_thres_x.hword);
-+ ccgr_q->cm_query.td_thres.hword =
-+ be16_to_cpu(ccgr_q->cm_query.td_thres.hword);
-+ ccgr_q->cm_query.wr_parm_g.word =
-+ be32_to_cpu(ccgr_q->cm_query.wr_parm_g.word);
-+ ccgr_q->cm_query.wr_parm_y.word =
-+ be32_to_cpu(ccgr_q->cm_query.wr_parm_y.word);
-+ ccgr_q->cm_query.wr_parm_r.word =
-+ be32_to_cpu(ccgr_q->cm_query.wr_parm_r.word);
-+ ccgr_q->cm_query.cscn_targ_dcp =
-+ be16_to_cpu(ccgr_q->cm_query.cscn_targ_dcp);
-+ ccgr_q->cm_query.i_cnt = be40_to_cpu(ccgr_q->cm_query.i_cnt);
-+ ccgr_q->cm_query.a_cnt = be40_to_cpu(ccgr_q->cm_query.a_cnt);
-+ for (i = 0; i < ARRAY_SIZE(ccgr_q->cm_query.cscn_targ_swp); i++)
-+ ccgr_q->cm_query.cscn_targ_swp[i] =
-+ be32_to_cpu(ccgr_q->cm_query.cscn_targ_swp[i]);
-+}
-+
-+/* In the case that slow- and fast-path handling are both done by qman_poll()
-+ * (ie. because there is no interrupt handling), we ought to balance how often
-+ * we do the fast-path poll versus the slow-path poll. We'll use two decrementer
-+ * sources, so we call the fast poll 'n' times before calling the slow poll
-+ * once. The idle decrementer constant is used when the last slow-poll detected
-+ * no work to do, and the busy decrementer constant when the last slow-poll had
-+ * work to do. */
-+#define SLOW_POLL_IDLE 1000
-+#define SLOW_POLL_BUSY 10
-+static u32 __poll_portal_slow(struct qman_portal *p, u32 is);
-+static inline unsigned int __poll_portal_fast(struct qman_portal *p,
-+ unsigned int poll_limit);
-+
-+/* Portal interrupt handler */
-+static irqreturn_t portal_isr(__always_unused int irq, void *ptr)
-+{
-+ struct qman_portal *p = ptr;
-+ /*
-+ * The CSCI/CCSCI source is cleared inside __poll_portal_slow(), because
-+ * it could race against a Query Congestion State command also given
-+ * as part of the handling of this interrupt source. We mustn't
-+ * clear it a second time in this top-level function.
-+ */
-+ u32 clear = QM_DQAVAIL_MASK | (p->irq_sources &
-+ ~(QM_PIRQ_CSCI | QM_PIRQ_CCSCI));
-+ u32 is = qm_isr_status_read(&p->p) & p->irq_sources;
-+ /* DQRR-handling if it's interrupt-driven */
-+ if (is & QM_PIRQ_DQRI)
-+ __poll_portal_fast(p, CONFIG_FSL_QMAN_POLL_LIMIT);
-+ /* Handling of anything else that's interrupt-driven */
-+ clear |= __poll_portal_slow(p, is);
-+ qm_isr_status_clear(&p->p, clear);
-+ return IRQ_HANDLED;
-+}
-+
-+/* This inner version is used privately by qman_create_affine_portal(), as well
-+ * as by the exported qman_stop_dequeues(). */
-+static inline void qman_stop_dequeues_ex(struct qman_portal *p)
-+{
-+ unsigned long irqflags __maybe_unused;
-+ PORTAL_IRQ_LOCK(p, irqflags);
-+ if (!(p->dqrr_disable_ref++))
-+ qm_dqrr_set_maxfill(&p->p, 0);
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+}
-+
-+static int drain_mr_fqrni(struct qm_portal *p)
-+{
-+ const struct qm_mr_entry *msg;
-+loop:
-+ msg = qm_mr_current(p);
-+ if (!msg) {
-+ /* if MR was full and h/w had other FQRNI entries to produce, we
-+ * need to allow it time to produce those entries once the
-+ * existing entries are consumed. A worst-case situation
-+ * (fully-loaded system) means h/w sequencers may have to do 3-4
-+ * other things before servicing the portal's MR pump, each of
-+ * which (if slow) may take ~50 qman cycles (which is ~200
-+ * processor cycles). So rounding up and then multiplying this
-+ * worst-case estimate by a factor of 10, just to be
-+ * ultra-paranoid, goes as high as 10,000 cycles. NB, we consume
-+ * one entry at a time, so h/w has an opportunity to produce new
-+ * entries well before the ring has been fully consumed, so
-+ * we're being *really* paranoid here. */
-+ u64 now, then = mfatb();
-+ do {
-+ now = mfatb();
-+ } while ((then + 10000) > now);
-+ msg = qm_mr_current(p);
-+ if (!msg)
-+ return 0;
-+ }
-+ if ((msg->verb & QM_MR_VERB_TYPE_MASK) != QM_MR_VERB_FQRNI) {
-+ /* We aren't draining anything but FQRNIs */
-+ pr_err("QMan found verb 0x%x in MR\n", msg->verb);
-+ return -1;
-+ }
-+ qm_mr_next(p);
-+ qm_mr_cci_consume(p, 1);
-+ goto loop;
-+}
-+
-+#ifdef CONFIG_SUSPEND
-+static int _qman_portal_suspend_noirq(struct device *dev)
-+{
-+ struct qman_portal *p = (struct qman_portal *)dev->platform_data;
-+#ifdef CONFIG_PM_DEBUG
-+ struct platform_device *pdev = to_platform_device(dev);
-+#endif
-+
-+ p->save_isdr = qm_isr_disable_read(&p->p);
-+ qm_isr_disable_write(&p->p, 0xffffffff);
-+ qm_isr_status_clear(&p->p, 0xffffffff);
-+#ifdef CONFIG_PM_DEBUG
-+ pr_info("Suspend for %s\n", pdev->name);
-+#endif
-+ return 0;
-+}
-+
-+static int _qman_portal_resume_noirq(struct device *dev)
-+{
-+ struct qman_portal *p = (struct qman_portal *)dev->platform_data;
-+
-+ /* restore isdr */
-+ qm_isr_disable_write(&p->p, p->save_isdr);
-+ return 0;
-+}
-+#else
-+#define _qman_portal_suspend_noirq NULL
-+#define _qman_portal_resume_noirq NULL
-+#endif
-+
-+struct dev_pm_domain qman_portal_device_pm_domain = {
-+ .ops = {
-+ USE_PLATFORM_PM_SLEEP_OPS
-+ .suspend_noirq = _qman_portal_suspend_noirq,
-+ .resume_noirq = _qman_portal_resume_noirq,
-+ }
-+};
-+
-+struct qman_portal *qman_create_portal(
-+ struct qman_portal *portal,
-+ const struct qm_portal_config *config,
-+ const struct qman_cgrs *cgrs)
-+{
-+ struct qm_portal *__p;
-+ char buf[16];
-+ int ret;
-+ u32 isdr;
-+
-+ if (!portal) {
-+ portal = kmalloc(sizeof(*portal), GFP_KERNEL);
-+ if (!portal)
-+ return portal;
-+ portal->alloced = 1;
-+ } else
-+ portal->alloced = 0;
-+
-+ __p = &portal->p;
-+
-+#if (defined CONFIG_PPC || defined CONFIG_PPC64) && defined CONFIG_FSL_PAMU
-+ /* PAMU is required for stashing */
-+ portal->use_eqcr_ci_stashing = ((qman_ip_rev >= QMAN_REV30) ?
-+ 1 : 0);
-+#elif defined(CONFIG_ARM) || defined(CONFIG_ARM64)
-+ portal->use_eqcr_ci_stashing = 1;
-+#else
-+ portal->use_eqcr_ci_stashing = 0;
-+#endif
-+
-+ /* prep the low-level portal struct with the mapped addresses from the
-+ * config, everything that follows depends on it and "config" is more
-+ * for (de)reference... */
-+ __p->addr.addr_ce = config->addr_virt[DPA_PORTAL_CE];
-+ __p->addr.addr_ci = config->addr_virt[DPA_PORTAL_CI];
-+ /*
-+ * If CI-stashing is used, the current defaults use a threshold of 3,
-+ * and stash with high-than-DQRR priority.
-+ */
-+ if (qm_eqcr_init(__p, qm_eqcr_pvb,
-+ portal->use_eqcr_ci_stashing ? 3 : 0, 1)) {
-+ pr_err("Qman EQCR initialisation failed\n");
-+ goto fail_eqcr;
-+ }
-+ if (qm_dqrr_init(__p, config, qm_dqrr_dpush, qm_dqrr_pvb,
-+ qm_dqrr_cdc, DQRR_MAXFILL)) {
-+ pr_err("Qman DQRR initialisation failed\n");
-+ goto fail_dqrr;
-+ }
-+ if (qm_mr_init(__p, qm_mr_pvb, qm_mr_cci)) {
-+ pr_err("Qman MR initialisation failed\n");
-+ goto fail_mr;
-+ }
-+ if (qm_mc_init(__p)) {
-+ pr_err("Qman MC initialisation failed\n");
-+ goto fail_mc;
-+ }
-+ if (qm_isr_init(__p)) {
-+ pr_err("Qman ISR initialisation failed\n");
-+ goto fail_isr;
-+ }
-+ /* static interrupt-gating controls */
-+ qm_dqrr_set_ithresh(__p, CONFIG_FSL_QMAN_PIRQ_DQRR_ITHRESH);
-+ qm_mr_set_ithresh(__p, CONFIG_FSL_QMAN_PIRQ_MR_ITHRESH);
-+ qm_isr_set_iperiod(__p, CONFIG_FSL_QMAN_PIRQ_IPERIOD);
-+ portal->cgrs = kmalloc(2 * sizeof(*cgrs), GFP_KERNEL);
-+ if (!portal->cgrs)
-+ goto fail_cgrs;
-+ /* initial snapshot is no-depletion */
-+ qman_cgrs_init(&portal->cgrs[1]);
-+ if (cgrs)
-+ portal->cgrs[0] = *cgrs;
-+ else
-+ /* if the given mask is NULL, assume all CGRs can be seen */
-+ qman_cgrs_fill(&portal->cgrs[0]);
-+ INIT_LIST_HEAD(&portal->cgr_cbs);
-+ spin_lock_init(&portal->cgr_lock);
-+ if (num_ceetms) {
-+ for (ret = 0; ret < num_ceetms; ret++) {
-+ portal->ccgrs[ret] = kmalloc(2 *
-+ sizeof(struct qman_ccgrs), GFP_KERNEL);
-+ if (!portal->ccgrs[ret])
-+ goto fail_ccgrs;
-+ qman_ccgrs_init(&portal->ccgrs[ret][1]);
-+ qman_ccgrs_fill(&portal->ccgrs[ret][0]);
-+ INIT_LIST_HEAD(&portal->ccgr_cbs[ret]);
-+ }
-+ }
-+ spin_lock_init(&portal->ccgr_lock);
-+ portal->bits = 0;
-+ portal->slowpoll = 0;
-+#ifdef CONFIG_FSL_DPA_CAN_WAIT_SYNC
-+ portal->eqci_owned = NULL;
-+#endif
-+#ifdef CONFIG_FSL_DPA_PORTAL_SHARE
-+ raw_spin_lock_init(&portal->sharing_lock);
-+ portal->is_shared = config->public_cfg.is_shared;
-+ portal->sharing_redirect = NULL;
-+#endif
-+ portal->sdqcr = QM_SDQCR_SOURCE_CHANNELS | QM_SDQCR_COUNT_UPTO3 |
-+ QM_SDQCR_DEDICATED_PRECEDENCE | QM_SDQCR_TYPE_PRIO_QOS |
-+ QM_SDQCR_TOKEN_SET(0xab) | QM_SDQCR_CHANNELS_DEDICATED;
-+ portal->dqrr_disable_ref = 0;
-+ portal->cb_dc_ern = NULL;
-+ sprintf(buf, "qportal-%d", config->public_cfg.channel);
-+ portal->pdev = platform_device_alloc(buf, -1);
-+ if (!portal->pdev) {
-+ pr_err("qman_portal - platform_device_alloc() failed\n");
-+ goto fail_devalloc;
-+ }
-+#if defined(CONFIG_ARM) || defined(CONFIG_ARM64)
-+ portal->pdev->dev.coherent_dma_mask = DMA_BIT_MASK(40);
-+ portal->pdev->dev.dma_mask = &portal->pdev->dev.coherent_dma_mask;
-+#else
-+ if (dma_set_mask(&portal->pdev->dev, DMA_BIT_MASK(40))) {
-+ pr_err("qman_portal - dma_set_mask() failed\n");
-+ goto fail_devadd;
-+ }
-+#endif
-+ portal->pdev->dev.pm_domain = &qman_portal_device_pm_domain;
-+ portal->pdev->dev.platform_data = portal;
-+ ret = platform_device_add(portal->pdev);
-+ if (ret) {
-+ pr_err("qman_portal - platform_device_add() failed\n");
-+ goto fail_devadd;
-+ }
-+ dpa_rbtree_init(&portal->retire_table);
-+ isdr = 0xffffffff;
-+ qm_isr_disable_write(__p, isdr);
-+ portal->irq_sources = 0;
-+ qm_isr_enable_write(__p, portal->irq_sources);
-+ qm_isr_status_clear(__p, 0xffffffff);
-+ snprintf(portal->irqname, MAX_IRQNAME, IRQNAME, config->public_cfg.cpu);
-+ if (request_irq(config->public_cfg.irq, portal_isr, 0, portal->irqname,
-+ portal)) {
-+ pr_err("request_irq() failed\n");
-+ goto fail_irq;
-+ }
-+ if ((config->public_cfg.cpu != -1) &&
-+ irq_can_set_affinity(config->public_cfg.irq) &&
-+ irq_set_affinity(config->public_cfg.irq,
-+ cpumask_of(config->public_cfg.cpu))) {
-+ pr_err("irq_set_affinity() failed\n");
-+ goto fail_affinity;
-+ }
-+
-+ /* Need EQCR to be empty before continuing */
-+ isdr ^= QM_PIRQ_EQCI;
-+ qm_isr_disable_write(__p, isdr);
-+ ret = qm_eqcr_get_fill(__p);
-+ if (ret) {
-+ pr_err("Qman EQCR unclean\n");
-+ goto fail_eqcr_empty;
-+ }
-+ isdr ^= (QM_PIRQ_DQRI | QM_PIRQ_MRI);
-+ qm_isr_disable_write(__p, isdr);
-+ if (qm_dqrr_current(__p) != NULL) {
-+ pr_err("Qman DQRR unclean\n");
-+ qm_dqrr_cdc_consume_n(__p, 0xffff);
-+ }
-+ if (qm_mr_current(__p) != NULL) {
-+ /* special handling, drain just in case it's a few FQRNIs */
-+ if (drain_mr_fqrni(__p)) {
-+ const struct qm_mr_entry *e = qm_mr_current(__p);
-+ /*
-+ * Message ring cannot be empty no need to check
-+ * qm_mr_current returned successfully
-+ */
-+ pr_err("Qman MR unclean, MR VERB 0x%x, rc 0x%x\n, addr 0x%x",
-+ e->verb, e->ern.rc, e->ern.fd.addr_lo);
-+ goto fail_dqrr_mr_empty;
-+ }
-+ }
-+ /* Success */
-+ portal->config = config;
-+ qm_isr_disable_write(__p, 0);
-+ qm_isr_uninhibit(__p);
-+ /* Write a sane SDQCR */
-+ qm_dqrr_sdqcr_set(__p, portal->sdqcr);
-+ return portal;
-+fail_dqrr_mr_empty:
-+fail_eqcr_empty:
-+fail_affinity:
-+ free_irq(config->public_cfg.irq, portal);
-+fail_irq:
-+ platform_device_del(portal->pdev);
-+fail_devadd:
-+ platform_device_put(portal->pdev);
-+fail_devalloc:
-+ if (num_ceetms)
-+ for (ret = 0; ret < num_ceetms; ret++)
-+ kfree(portal->ccgrs[ret]);
-+fail_ccgrs:
-+ kfree(portal->cgrs);
-+fail_cgrs:
-+ qm_isr_finish(__p);
-+fail_isr:
-+ qm_mc_finish(__p);
-+fail_mc:
-+ qm_mr_finish(__p);
-+fail_mr:
-+ qm_dqrr_finish(__p);
-+fail_dqrr:
-+ qm_eqcr_finish(__p);
-+fail_eqcr:
-+ if (portal->alloced)
-+ kfree(portal);
-+ return NULL;
-+}
-+
-+struct qman_portal *qman_create_affine_portal(
-+ const struct qm_portal_config *config,
-+ const struct qman_cgrs *cgrs)
-+{
-+ struct qman_portal *res;
-+ struct qman_portal *portal;
-+
-+ portal = &per_cpu(qman_affine_portal, config->public_cfg.cpu);
-+ res = qman_create_portal(portal, config, cgrs);
-+ if (res) {
-+ spin_lock(&affine_mask_lock);
-+ cpumask_set_cpu(config->public_cfg.cpu, &affine_mask);
-+ affine_channels[config->public_cfg.cpu] =
-+ config->public_cfg.channel;
-+ affine_portals[config->public_cfg.cpu] = portal;
-+ spin_unlock(&affine_mask_lock);
-+ }
-+ return res;
-+}
-+
-+/* These checks are BUG_ON()s because the driver is already supposed to avoid
-+ * these cases. */
-+struct qman_portal *qman_create_affine_slave(struct qman_portal *redirect,
-+ int cpu)
-+{
-+#ifdef CONFIG_FSL_DPA_PORTAL_SHARE
-+ struct qman_portal *p;
-+ p = &per_cpu(qman_affine_portal, cpu);
-+ /* Check that we don't already have our own portal */
-+ BUG_ON(p->config);
-+ /* Check that we aren't already slaving to another portal */
-+ BUG_ON(p->is_shared);
-+ /* Check that 'redirect' is prepared to have us */
-+ BUG_ON(!redirect->config->public_cfg.is_shared);
-+ /* These are the only elements to initialise when redirecting */
-+ p->irq_sources = 0;
-+ p->sharing_redirect = redirect;
-+ affine_portals[cpu] = p;
-+ return p;
-+#else
-+ BUG();
-+ return NULL;
-+#endif
-+}
-+
-+void qman_destroy_portal(struct qman_portal *qm)
-+{
-+ const struct qm_portal_config *pcfg;
-+ int i;
-+
-+ /* Stop dequeues on the portal */
-+ qm_dqrr_sdqcr_set(&qm->p, 0);
-+
-+ /* NB we do this to "quiesce" EQCR. If we add enqueue-completions or
-+ * something related to QM_PIRQ_EQCI, this may need fixing.
-+ * Also, due to the prefetching model used for CI updates in the enqueue
-+ * path, this update will only invalidate the CI cacheline *after*
-+ * working on it, so we need to call this twice to ensure a full update
-+ * irrespective of where the enqueue processing was at when the teardown
-+ * began. */
-+ qm_eqcr_cce_update(&qm->p);
-+ qm_eqcr_cce_update(&qm->p);
-+ pcfg = qm->config;
-+
-+ free_irq(pcfg->public_cfg.irq, qm);
-+
-+ kfree(qm->cgrs);
-+ if (num_ceetms)
-+ for (i = 0; i < num_ceetms; i++)
-+ kfree(qm->ccgrs[i]);
-+ qm_isr_finish(&qm->p);
-+ qm_mc_finish(&qm->p);
-+ qm_mr_finish(&qm->p);
-+ qm_dqrr_finish(&qm->p);
-+ qm_eqcr_finish(&qm->p);
-+
-+ platform_device_del(qm->pdev);
-+ platform_device_put(qm->pdev);
-+
-+ qm->config = NULL;
-+ if (qm->alloced)
-+ kfree(qm);
-+}
-+
-+const struct qm_portal_config *qman_destroy_affine_portal(void)
-+{
-+ /* We don't want to redirect if we're a slave, use "raw" */
-+ struct qman_portal *qm = get_raw_affine_portal();
-+ const struct qm_portal_config *pcfg;
-+ int cpu;
-+#ifdef CONFIG_FSL_DPA_PORTAL_SHARE
-+ if (qm->sharing_redirect) {
-+ qm->sharing_redirect = NULL;
-+ put_affine_portal();
-+ return NULL;
-+ }
-+ qm->is_shared = 0;
-+#endif
-+ pcfg = qm->config;
-+ cpu = pcfg->public_cfg.cpu;
-+
-+ qman_destroy_portal(qm);
-+
-+ spin_lock(&affine_mask_lock);
-+ cpumask_clear_cpu(cpu, &affine_mask);
-+ spin_unlock(&affine_mask_lock);
-+ put_affine_portal();
-+ return pcfg;
-+}
-+
-+const struct qman_portal_config *qman_p_get_portal_config(struct qman_portal *p)
-+{
-+ return &p->config->public_cfg;
-+}
-+EXPORT_SYMBOL(qman_p_get_portal_config);
-+
-+const struct qman_portal_config *qman_get_portal_config(void)
-+{
-+ struct qman_portal *p = get_affine_portal();
-+ const struct qman_portal_config *ret = qman_p_get_portal_config(p);
-+ put_affine_portal();
-+ return ret;
-+}
-+EXPORT_SYMBOL(qman_get_portal_config);
-+
-+/* Inline helper to reduce nesting in __poll_portal_slow() */
-+static inline void fq_state_change(struct qman_portal *p, struct qman_fq *fq,
-+ const struct qm_mr_entry *msg, u8 verb)
-+{
-+ FQLOCK(fq);
-+ switch (verb) {
-+ case QM_MR_VERB_FQRL:
-+ DPA_ASSERT(fq_isset(fq, QMAN_FQ_STATE_ORL));
-+ fq_clear(fq, QMAN_FQ_STATE_ORL);
-+ table_del_fq(p, fq);
-+ break;
-+ case QM_MR_VERB_FQRN:
-+ DPA_ASSERT((fq->state == qman_fq_state_parked) ||
-+ (fq->state == qman_fq_state_sched));
-+ DPA_ASSERT(fq_isset(fq, QMAN_FQ_STATE_CHANGING));
-+ fq_clear(fq, QMAN_FQ_STATE_CHANGING);
-+ if (msg->fq.fqs & QM_MR_FQS_NOTEMPTY)
-+ fq_set(fq, QMAN_FQ_STATE_NE);
-+ if (msg->fq.fqs & QM_MR_FQS_ORLPRESENT)
-+ fq_set(fq, QMAN_FQ_STATE_ORL);
-+ else
-+ table_del_fq(p, fq);
-+ fq->state = qman_fq_state_retired;
-+ break;
-+ case QM_MR_VERB_FQPN:
-+ DPA_ASSERT(fq->state == qman_fq_state_sched);
-+ DPA_ASSERT(fq_isclear(fq, QMAN_FQ_STATE_CHANGING));
-+ fq->state = qman_fq_state_parked;
-+ }
-+ FQUNLOCK(fq);
-+}
-+
-+static u32 __poll_portal_slow(struct qman_portal *p, u32 is)
-+{
-+ const struct qm_mr_entry *msg;
-+ struct qm_mr_entry swapped_msg;
-+ int k;
-+
-+ if (is & QM_PIRQ_CSCI) {
-+ struct qman_cgrs rr, c;
-+ struct qm_mc_result *mcr;
-+ struct qman_cgr *cgr;
-+ unsigned long irqflags __maybe_unused;
-+
-+ spin_lock_irqsave(&p->cgr_lock, irqflags);
-+ /*
-+ * The CSCI bit must be cleared _before_ issuing the
-+ * Query Congestion State command, to ensure that a long
-+ * CGR State Change callback cannot miss an intervening
-+ * state change.
-+ */
-+ qm_isr_status_clear(&p->p, QM_PIRQ_CSCI);
-+ qm_mc_start(&p->p);
-+ qm_mc_commit(&p->p, QM_MCC_VERB_QUERYCONGESTION);
-+ while (!(mcr = qm_mc_result(&p->p)))
-+ cpu_relax();
-+ for (k = 0; k < 8; k++)
-+ mcr->querycongestion.state.__state[k] = be32_to_cpu(
-+ mcr->querycongestion.state.__state[k]);
-+ /* mask out the ones I'm not interested in */
-+ qman_cgrs_and(&rr, (const struct qman_cgrs *)
-+ &mcr->querycongestion.state, &p->cgrs[0]);
-+ /* check previous snapshot for delta, enter/exit congestion */
-+ qman_cgrs_xor(&c, &rr, &p->cgrs[1]);
-+ /* update snapshot */
-+ qman_cgrs_cp(&p->cgrs[1], &rr);
-+ /* Invoke callback */
-+ list_for_each_entry(cgr, &p->cgr_cbs, node)
-+ if (cgr->cb && qman_cgrs_get(&c, cgr->cgrid))
-+ cgr->cb(p, cgr, qman_cgrs_get(&rr, cgr->cgrid));
-+ spin_unlock_irqrestore(&p->cgr_lock, irqflags);
-+ }
-+ if (is & QM_PIRQ_CCSCI) {
-+ struct qman_ccgrs rr, c, congestion_result;
-+ struct qm_mc_result *mcr;
-+ struct qm_mc_command *mcc;
-+ struct qm_ceetm_ccg *ccg;
-+ unsigned long irqflags __maybe_unused;
-+ int i, j;
-+
-+ spin_lock_irqsave(&p->ccgr_lock, irqflags);
-+ /*
-+ * The CCSCI bit must be cleared _before_ issuing the
-+ * Query Congestion State command, to ensure that a long
-+ * CCGR State Change callback cannot miss an intervening
-+ * state change.
-+ */
-+ qm_isr_status_clear(&p->p, QM_PIRQ_CCSCI);
-+
-+ for (i = 0; i < num_ceetms; i++) {
-+ for (j = 0; j < 2; j++) {
-+ mcc = qm_mc_start(&p->p);
-+ mcc->ccgr_query.ccgrid = cpu_to_be16(
-+ CEETM_QUERY_CONGESTION_STATE | j);
-+ mcc->ccgr_query.dcpid = i;
-+ qm_mc_commit(&p->p, QM_CEETM_VERB_CCGR_QUERY);
-+ while (!(mcr = qm_mc_result(&p->p)))
-+ cpu_relax();
-+ for (k = 0; k < 8; k++)
-+ mcr->ccgr_query.congestion_state.state.
-+ __state[k] = be32_to_cpu(
-+ mcr->ccgr_query.
-+ congestion_state.state.
-+ __state[k]);
-+ congestion_result.q[j] =
-+ mcr->ccgr_query.congestion_state.state;
-+ }
-+ /* mask out the ones I'm not interested in */
-+ qman_ccgrs_and(&rr, &congestion_result,
-+ &p->ccgrs[i][0]);
-+ /*
-+ * check previous snapshot for delta, enter/exit
-+ * congestion.
-+ */
-+ qman_ccgrs_xor(&c, &rr, &p->ccgrs[i][1]);
-+ /* update snapshot */
-+ qman_ccgrs_cp(&p->ccgrs[i][1], &rr);
-+ /* Invoke callback */
-+ list_for_each_entry(ccg, &p->ccgr_cbs[i], cb_node)
-+ if (ccg->cb && qman_ccgrs_get(&c,
-+ (ccg->parent->idx << 4) | ccg->idx))
-+ ccg->cb(ccg, ccg->cb_ctx,
-+ qman_ccgrs_get(&rr,
-+ (ccg->parent->idx << 4)
-+ | ccg->idx));
-+ }
-+ spin_unlock_irqrestore(&p->ccgr_lock, irqflags);
-+ }
-+
-+#ifdef CONFIG_FSL_DPA_CAN_WAIT_SYNC
-+ if (is & QM_PIRQ_EQCI) {
-+ unsigned long irqflags;
-+ PORTAL_IRQ_LOCK(p, irqflags);
-+ p->eqci_owned = NULL;
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+ wake_up(&affine_queue);
-+ }
-+#endif
-+
-+ if (is & QM_PIRQ_EQRI) {
-+ unsigned long irqflags __maybe_unused;
-+ PORTAL_IRQ_LOCK(p, irqflags);
-+ qm_eqcr_cce_update(&p->p);
-+ qm_eqcr_set_ithresh(&p->p, 0);
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+ wake_up(&affine_queue);
-+ }
-+
-+ if (is & QM_PIRQ_MRI) {
-+ struct qman_fq *fq;
-+ u8 verb, num = 0;
-+mr_loop:
-+ qm_mr_pvb_update(&p->p);
-+ msg = qm_mr_current(&p->p);
-+ if (!msg)
-+ goto mr_done;
-+ swapped_msg = *msg;
-+ hw_fd_to_cpu(&swapped_msg.ern.fd);
-+ verb = msg->verb & QM_MR_VERB_TYPE_MASK;
-+ /* The message is a software ERN iff the 0x20 bit is set */
-+ if (verb & 0x20) {
-+ switch (verb) {
-+ case QM_MR_VERB_FQRNI:
-+ /* nada, we drop FQRNIs on the floor */
-+ break;
-+ case QM_MR_VERB_FQRN:
-+ case QM_MR_VERB_FQRL:
-+ /* Lookup in the retirement table */
-+ fq = table_find_fq(p, be32_to_cpu(msg->fq.fqid));
-+ BUG_ON(!fq);
-+ fq_state_change(p, fq, &swapped_msg, verb);
-+ if (fq->cb.fqs)
-+ fq->cb.fqs(p, fq, &swapped_msg);
-+ break;
-+ case QM_MR_VERB_FQPN:
-+ /* Parked */
-+#ifdef CONFIG_FSL_QMAN_FQ_LOOKUP
-+ fq = get_fq_table_entry(
-+ be32_to_cpu(msg->fq.contextB));
-+#else
-+ fq = (void *)(uintptr_t)
-+ be32_to_cpu(msg->fq.contextB);
-+#endif
-+ fq_state_change(p, fq, msg, verb);
-+ if (fq->cb.fqs)
-+ fq->cb.fqs(p, fq, &swapped_msg);
-+ break;
-+ case QM_MR_VERB_DC_ERN:
-+ /* DCP ERN */
-+ if (p->cb_dc_ern)
-+ p->cb_dc_ern(p, msg);
-+ else if (cb_dc_ern)
-+ cb_dc_ern(p, msg);
-+ else {
-+ static int warn_once;
-+ if (!warn_once) {
-+ pr_crit("Leaking DCP ERNs!\n");
-+ warn_once = 1;
-+ }
-+ }
-+ break;
-+ default:
-+ pr_crit("Invalid MR verb 0x%02x\n", verb);
-+ }
-+ } else {
-+ /* Its a software ERN */
-+#ifdef CONFIG_FSL_QMAN_FQ_LOOKUP
-+ fq = get_fq_table_entry(be32_to_cpu(msg->ern.tag));
-+#else
-+ fq = (void *)(uintptr_t)be32_to_cpu(msg->ern.tag);
-+#endif
-+ fq->cb.ern(p, fq, &swapped_msg);
-+ }
-+ num++;
-+ qm_mr_next(&p->p);
-+ goto mr_loop;
-+mr_done:
-+ qm_mr_cci_consume(&p->p, num);
-+ }
-+ /*
-+ * QM_PIRQ_CSCI/CCSCI has already been cleared, as part of its specific
-+ * processing. If that interrupt source has meanwhile been re-asserted,
-+ * we mustn't clear it here (or in the top-level interrupt handler).
-+ */
-+ return is & (QM_PIRQ_EQCI | QM_PIRQ_EQRI | QM_PIRQ_MRI);
-+}
-+
-+/* remove some slowish-path stuff from the "fast path" and make sure it isn't
-+ * inlined. */
-+static noinline void clear_vdqcr(struct qman_portal *p, struct qman_fq *fq)
-+{
-+ p->vdqcr_owned = NULL;
-+ FQLOCK(fq);
-+ fq_clear(fq, QMAN_FQ_STATE_VDQCR);
-+ FQUNLOCK(fq);
-+ wake_up(&affine_queue);
-+}
-+
-+/* Copy a DQRR entry ensuring reads reach QBMan in order */
-+static inline void safe_copy_dqrr(struct qm_dqrr_entry *dst,
-+ const struct qm_dqrr_entry *src)
-+{
-+ int i = 0;
-+ const u64 *s64 = (u64*)src;
-+ u64 *d64 = (u64*)dst;
-+
-+ /* DQRR only has 32 bytes of valid data so only need to
-+ * copy 4 - 64 bit values */
-+ *d64 = *s64;
-+#if defined(CONFIG_ARM) || defined(CONFIG_ARM64)
-+ {
-+ u32 res, zero = 0;
-+ /* Create a dependancy after copying first bytes ensures no wrap
-+ transaction generated to QBMan */
-+ /* Logical AND the value pointed to by s64 with 0x0 and
-+ store the result in res */
-+ asm volatile("and %[result], %[in1], %[in2]"
-+ : [result] "=r" (res)
-+ : [in1] "r" (zero), [in2] "r" (*s64)
-+ : "memory");
-+ /* Add res to s64 - this creates a dependancy on the result of
-+ reading the value of s64 before the next read. The side
-+ effect of this is that the core must stall until the first
-+ aligned read is complete therefore preventing a WRAP
-+ transaction to be seen by the QBMan */
-+ asm volatile("add %[result], %[in1], %[in2]"
-+ : [result] "=r" (s64)
-+ : [in1] "r" (res), [in2] "r" (s64)
-+ : "memory");
-+ }
-+#endif
-+ /* Copy the last 3 64 bit parts */
-+ d64++; s64++;
-+ for (;i<3; i++)
-+ *d64++ = *s64++;
-+}
-+
-+/* Look: no locks, no irq_save()s, no preempt_disable()s! :-) The only states
-+ * that would conflict with other things if they ran at the same time on the
-+ * same cpu are;
-+ *
-+ * (i) setting/clearing vdqcr_owned, and
-+ * (ii) clearing the NE (Not Empty) flag.
-+ *
-+ * Both are safe. Because;
-+ *
-+ * (i) this clearing can only occur after qman_volatile_dequeue() has set the
-+ * vdqcr_owned field (which it does before setting VDQCR), and
-+ * qman_volatile_dequeue() blocks interrupts and preemption while this is
-+ * done so that we can't interfere.
-+ * (ii) the NE flag is only cleared after qman_retire_fq() has set it, and as
-+ * with (i) that API prevents us from interfering until it's safe.
-+ *
-+ * The good thing is that qman_volatile_dequeue() and qman_retire_fq() run far
-+ * less frequently (ie. per-FQ) than __poll_portal_fast() does, so the nett
-+ * advantage comes from this function not having to "lock" anything at all.
-+ *
-+ * Note also that the callbacks are invoked at points which are safe against the
-+ * above potential conflicts, but that this function itself is not re-entrant
-+ * (this is because the function tracks one end of each FIFO in the portal and
-+ * we do *not* want to lock that). So the consequence is that it is safe for
-+ * user callbacks to call into any Qman API *except* qman_poll() (as that's the
-+ * sole API that could be invoking the callback through this function).
-+ */
-+static inline unsigned int __poll_portal_fast(struct qman_portal *p,
-+ unsigned int poll_limit)
-+{
-+ const struct qm_dqrr_entry *dq;
-+ struct qman_fq *fq;
-+ enum qman_cb_dqrr_result res;
-+ unsigned int limit = 0;
-+#if __BYTE_ORDER__ == __ORDER_LITTLE_ENDIAN__
-+ struct qm_dqrr_entry *shadow;
-+ const struct qm_dqrr_entry *orig_dq;
-+#endif
-+loop:
-+ qm_dqrr_pvb_update(&p->p);
-+ dq = qm_dqrr_current(&p->p);
-+ if (!dq)
-+ goto done;
-+#if __BYTE_ORDER__ == __ORDER_LITTLE_ENDIAN__
-+ /* If running on an LE system the fields of the
-+ dequeue entry must be swapped. Because the
-+ QMan HW will ignore writes the DQRR entry is
-+ copied and the index stored within the copy */
-+ shadow = &p->shadow_dqrr[DQRR_PTR2IDX(dq)];
-+ /* Use safe copy here to avoid WRAP transaction */
-+ safe_copy_dqrr(shadow, dq);
-+ orig_dq = dq;
-+ dq = shadow;
-+ shadow->fqid = be32_to_cpu(shadow->fqid);
-+ shadow->contextB = be32_to_cpu(shadow->contextB);
-+ shadow->seqnum = be16_to_cpu(shadow->seqnum);
-+ hw_fd_to_cpu(&shadow->fd);
-+#endif
-+ if (dq->stat & QM_DQRR_STAT_UNSCHEDULED) {
-+ /* VDQCR: don't trust contextB as the FQ may have been
-+ * configured for h/w consumption and we're draining it
-+ * post-retirement. */
-+ fq = p->vdqcr_owned;
-+ /* We only set QMAN_FQ_STATE_NE when retiring, so we only need
-+ * to check for clearing it when doing volatile dequeues. It's
-+ * one less thing to check in the critical path (SDQCR). */
-+ if (dq->stat & QM_DQRR_STAT_FQ_EMPTY)
-+ fq_clear(fq, QMAN_FQ_STATE_NE);
-+ /* this is duplicated from the SDQCR code, but we have stuff to
-+ * do before *and* after this callback, and we don't want
-+ * multiple if()s in the critical path (SDQCR). */
-+ res = fq->cb.dqrr(p, fq, dq);
-+ if (res == qman_cb_dqrr_stop)
-+ goto done;
-+ /* Check for VDQCR completion */
-+ if (dq->stat & QM_DQRR_STAT_DQCR_EXPIRED)
-+ clear_vdqcr(p, fq);
-+ } else {
-+ /* SDQCR: contextB points to the FQ */
-+#ifdef CONFIG_FSL_QMAN_FQ_LOOKUP
-+ fq = get_fq_table_entry(dq->contextB);
-+#else
-+ fq = (void *)(uintptr_t)dq->contextB;
-+#endif
-+ /* Now let the callback do its stuff */
-+ res = fq->cb.dqrr(p, fq, dq);
-+
-+ /* The callback can request that we exit without consuming this
-+ * entry nor advancing; */
-+ if (res == qman_cb_dqrr_stop)
-+ goto done;
-+ }
-+ /* Interpret 'dq' from a driver perspective. */
-+ /* Parking isn't possible unless HELDACTIVE was set. NB,
-+ * FORCEELIGIBLE implies HELDACTIVE, so we only need to
-+ * check for HELDACTIVE to cover both. */
-+ DPA_ASSERT((dq->stat & QM_DQRR_STAT_FQ_HELDACTIVE) ||
-+ (res != qman_cb_dqrr_park));
-+#if __BYTE_ORDER__ == __ORDER_LITTLE_ENDIAN__
-+ if (res != qman_cb_dqrr_defer)
-+ qm_dqrr_cdc_consume_1ptr(&p->p, orig_dq,
-+ (res == qman_cb_dqrr_park));
-+#else
-+ /* Defer just means "skip it, I'll consume it myself later on" */
-+ if (res != qman_cb_dqrr_defer)
-+ qm_dqrr_cdc_consume_1ptr(&p->p, dq, (res == qman_cb_dqrr_park));
-+#endif
-+ /* Move forward */
-+ qm_dqrr_next(&p->p);
-+ /* Entry processed and consumed, increment our counter. The callback can
-+ * request that we exit after consuming the entry, and we also exit if
-+ * we reach our processing limit, so loop back only if neither of these
-+ * conditions is met. */
-+ if ((++limit < poll_limit) && (res != qman_cb_dqrr_consume_stop))
-+ goto loop;
-+done:
-+ return limit;
-+}
-+
-+u32 qman_irqsource_get(void)
-+{
-+ /* "irqsource" and "poll" APIs mustn't redirect when sharing, they
-+ * should shut the user out if they are not the primary CPU hosting the
-+ * portal. That's why we use the "raw" interface. */
-+ struct qman_portal *p = get_raw_affine_portal();
-+ u32 ret = p->irq_sources & QM_PIRQ_VISIBLE;
-+ put_affine_portal();
-+ return ret;
-+}
-+EXPORT_SYMBOL(qman_irqsource_get);
-+
-+int qman_p_irqsource_add(struct qman_portal *p, u32 bits __maybe_unused)
-+{
-+ __maybe_unused unsigned long irqflags;
-+#ifdef CONFIG_FSL_DPA_PORTAL_SHARE
-+ if (p->sharing_redirect)
-+ return -EINVAL;
-+ else
-+#endif
-+ {
-+ bits = bits & QM_PIRQ_VISIBLE;
-+ PORTAL_IRQ_LOCK(p, irqflags);
-+
-+ /* Clear any previously remaining interrupt conditions in
-+ * QCSP_ISR. This prevents raising a false interrupt when
-+ * interrupt conditions are enabled in QCSP_IER.
-+ */
-+ qm_isr_status_clear(&p->p, bits);
-+ set_bits(bits, &p->irq_sources);
-+ qm_isr_enable_write(&p->p, p->irq_sources);
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+ }
-+ return 0;
-+}
-+EXPORT_SYMBOL(qman_p_irqsource_add);
-+
-+int qman_irqsource_add(u32 bits __maybe_unused)
-+{
-+ struct qman_portal *p = get_raw_affine_portal();
-+ int ret;
-+ ret = qman_p_irqsource_add(p, bits);
-+ put_affine_portal();
-+ return ret;
-+}
-+EXPORT_SYMBOL(qman_irqsource_add);
-+
-+int qman_p_irqsource_remove(struct qman_portal *p, u32 bits)
-+{
-+ __maybe_unused unsigned long irqflags;
-+ u32 ier;
-+#ifdef CONFIG_FSL_DPA_PORTAL_SHARE
-+ if (p->sharing_redirect) {
-+ put_affine_portal();
-+ return -EINVAL;
-+ }
-+#endif
-+ /* Our interrupt handler only processes+clears status register bits that
-+ * are in p->irq_sources. As we're trimming that mask, if one of them
-+ * were to assert in the status register just before we remove it from
-+ * the enable register, there would be an interrupt-storm when we
-+ * release the IRQ lock. So we wait for the enable register update to
-+ * take effect in h/w (by reading it back) and then clear all other bits
-+ * in the status register. Ie. we clear them from ISR once it's certain
-+ * IER won't allow them to reassert. */
-+ PORTAL_IRQ_LOCK(p, irqflags);
-+ bits &= QM_PIRQ_VISIBLE;
-+ clear_bits(bits, &p->irq_sources);
-+ qm_isr_enable_write(&p->p, p->irq_sources);
-+
-+ ier = qm_isr_enable_read(&p->p);
-+ /* Using "~ier" (rather than "bits" or "~p->irq_sources") creates a
-+ * data-dependency, ie. to protect against re-ordering. */
-+ qm_isr_status_clear(&p->p, ~ier);
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+ return 0;
-+}
-+EXPORT_SYMBOL(qman_p_irqsource_remove);
-+
-+int qman_irqsource_remove(u32 bits)
-+{
-+ struct qman_portal *p = get_raw_affine_portal();
-+ int ret;
-+ ret = qman_p_irqsource_remove(p, bits);
-+ put_affine_portal();
-+ return ret;
-+}
-+EXPORT_SYMBOL(qman_irqsource_remove);
-+
-+const cpumask_t *qman_affine_cpus(void)
-+{
-+ return &affine_mask;
-+}
-+EXPORT_SYMBOL(qman_affine_cpus);
-+
-+u16 qman_affine_channel(int cpu)
-+{
-+ if (cpu < 0) {
-+ struct qman_portal *portal = get_raw_affine_portal();
-+#ifdef CONFIG_FSL_DPA_PORTAL_SHARE
-+ BUG_ON(portal->sharing_redirect);
-+#endif
-+ cpu = portal->config->public_cfg.cpu;
-+ put_affine_portal();
-+ }
-+ BUG_ON(!cpumask_test_cpu(cpu, &affine_mask));
-+ return affine_channels[cpu];
-+}
-+EXPORT_SYMBOL(qman_affine_channel);
-+
-+void *qman_get_affine_portal(int cpu)
-+{
-+ return affine_portals[cpu];
-+}
-+EXPORT_SYMBOL(qman_get_affine_portal);
-+
-+int qman_p_poll_dqrr(struct qman_portal *p, unsigned int limit)
-+{
-+ int ret;
-+
-+#ifdef CONFIG_FSL_DPA_PORTAL_SHARE
-+ if (unlikely(p->sharing_redirect))
-+ ret = -EINVAL;
-+ else
-+#endif
-+ {
-+ BUG_ON(p->irq_sources & QM_PIRQ_DQRI);
-+ ret = __poll_portal_fast(p, limit);
-+ }
-+ return ret;
-+}
-+EXPORT_SYMBOL(qman_p_poll_dqrr);
-+
-+int qman_poll_dqrr(unsigned int limit)
-+{
-+ struct qman_portal *p = get_poll_portal();
-+ int ret;
-+ ret = qman_p_poll_dqrr(p, limit);
-+ put_poll_portal();
-+ return ret;
-+}
-+EXPORT_SYMBOL(qman_poll_dqrr);
-+
-+u32 qman_p_poll_slow(struct qman_portal *p)
-+{
-+ u32 ret;
-+#ifdef CONFIG_FSL_DPA_PORTAL_SHARE
-+ if (unlikely(p->sharing_redirect))
-+ ret = (u32)-1;
-+ else
-+#endif
-+ {
-+ u32 is = qm_isr_status_read(&p->p) & ~p->irq_sources;
-+ ret = __poll_portal_slow(p, is);
-+ qm_isr_status_clear(&p->p, ret);
-+ }
-+ return ret;
-+}
-+EXPORT_SYMBOL(qman_p_poll_slow);
-+
-+u32 qman_poll_slow(void)
-+{
-+ struct qman_portal *p = get_poll_portal();
-+ u32 ret;
-+ ret = qman_p_poll_slow(p);
-+ put_poll_portal();
-+ return ret;
-+}
-+EXPORT_SYMBOL(qman_poll_slow);
-+
-+/* Legacy wrapper */
-+void qman_p_poll(struct qman_portal *p)
-+{
-+#ifdef CONFIG_FSL_DPA_PORTAL_SHARE
-+ if (unlikely(p->sharing_redirect))
-+ return;
-+#endif
-+ if ((~p->irq_sources) & QM_PIRQ_SLOW) {
-+ if (!(p->slowpoll--)) {
-+ u32 is = qm_isr_status_read(&p->p) & ~p->irq_sources;
-+ u32 active = __poll_portal_slow(p, is);
-+ if (active) {
-+ qm_isr_status_clear(&p->p, active);
-+ p->slowpoll = SLOW_POLL_BUSY;
-+ } else
-+ p->slowpoll = SLOW_POLL_IDLE;
-+ }
-+ }
-+ if ((~p->irq_sources) & QM_PIRQ_DQRI)
-+ __poll_portal_fast(p, CONFIG_FSL_QMAN_POLL_LIMIT);
-+}
-+EXPORT_SYMBOL(qman_p_poll);
-+
-+void qman_poll(void)
-+{
-+ struct qman_portal *p = get_poll_portal();
-+ qman_p_poll(p);
-+ put_poll_portal();
-+}
-+EXPORT_SYMBOL(qman_poll);
-+
-+void qman_p_stop_dequeues(struct qman_portal *p)
-+{
-+ qman_stop_dequeues_ex(p);
-+}
-+EXPORT_SYMBOL(qman_p_stop_dequeues);
-+
-+void qman_stop_dequeues(void)
-+{
-+ struct qman_portal *p = get_affine_portal();
-+ qman_p_stop_dequeues(p);
-+ put_affine_portal();
-+}
-+EXPORT_SYMBOL(qman_stop_dequeues);
-+
-+void qman_p_start_dequeues(struct qman_portal *p)
-+{
-+ unsigned long irqflags __maybe_unused;
-+ PORTAL_IRQ_LOCK(p, irqflags);
-+ DPA_ASSERT(p->dqrr_disable_ref > 0);
-+ if (!(--p->dqrr_disable_ref))
-+ qm_dqrr_set_maxfill(&p->p, DQRR_MAXFILL);
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+}
-+EXPORT_SYMBOL(qman_p_start_dequeues);
-+
-+void qman_start_dequeues(void)
-+{
-+ struct qman_portal *p = get_affine_portal();
-+ qman_p_start_dequeues(p);
-+ put_affine_portal();
-+}
-+EXPORT_SYMBOL(qman_start_dequeues);
-+
-+void qman_p_static_dequeue_add(struct qman_portal *p, u32 pools)
-+{
-+ unsigned long irqflags __maybe_unused;
-+ PORTAL_IRQ_LOCK(p, irqflags);
-+ pools &= p->config->public_cfg.pools;
-+ p->sdqcr |= pools;
-+ qm_dqrr_sdqcr_set(&p->p, p->sdqcr);
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+}
-+EXPORT_SYMBOL(qman_p_static_dequeue_add);
-+
-+void qman_static_dequeue_add(u32 pools)
-+{
-+ struct qman_portal *p = get_affine_portal();
-+ qman_p_static_dequeue_add(p, pools);
-+ put_affine_portal();
-+}
-+EXPORT_SYMBOL(qman_static_dequeue_add);
-+
-+void qman_p_static_dequeue_del(struct qman_portal *p, u32 pools)
-+{
-+ unsigned long irqflags __maybe_unused;
-+ PORTAL_IRQ_LOCK(p, irqflags);
-+ pools &= p->config->public_cfg.pools;
-+ p->sdqcr &= ~pools;
-+ qm_dqrr_sdqcr_set(&p->p, p->sdqcr);
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+}
-+EXPORT_SYMBOL(qman_p_static_dequeue_del);
-+
-+void qman_static_dequeue_del(u32 pools)
-+{
-+ struct qman_portal *p = get_affine_portal();
-+ qman_p_static_dequeue_del(p, pools);
-+ put_affine_portal();
-+}
-+EXPORT_SYMBOL(qman_static_dequeue_del);
-+
-+u32 qman_p_static_dequeue_get(struct qman_portal *p)
-+{
-+ return p->sdqcr;
-+}
-+EXPORT_SYMBOL(qman_p_static_dequeue_get);
-+
-+u32 qman_static_dequeue_get(void)
-+{
-+ struct qman_portal *p = get_affine_portal();
-+ u32 ret = qman_p_static_dequeue_get(p);
-+ put_affine_portal();
-+ return ret;
-+}
-+EXPORT_SYMBOL(qman_static_dequeue_get);
-+
-+void qman_p_dca(struct qman_portal *p, struct qm_dqrr_entry *dq,
-+ int park_request)
-+{
-+ qm_dqrr_cdc_consume_1ptr(&p->p, dq, park_request);
-+}
-+EXPORT_SYMBOL(qman_p_dca);
-+
-+void qman_dca(struct qm_dqrr_entry *dq, int park_request)
-+{
-+ struct qman_portal *p = get_affine_portal();
-+ qman_p_dca(p, dq, park_request);
-+ put_affine_portal();
-+}
-+EXPORT_SYMBOL(qman_dca);
-+
-+/*******************/
-+/* Frame queue API */
-+/*******************/
-+
-+static const char *mcr_result_str(u8 result)
-+{
-+ switch (result) {
-+ case QM_MCR_RESULT_NULL:
-+ return "QM_MCR_RESULT_NULL";
-+ case QM_MCR_RESULT_OK:
-+ return "QM_MCR_RESULT_OK";
-+ case QM_MCR_RESULT_ERR_FQID:
-+ return "QM_MCR_RESULT_ERR_FQID";
-+ case QM_MCR_RESULT_ERR_FQSTATE:
-+ return "QM_MCR_RESULT_ERR_FQSTATE";
-+ case QM_MCR_RESULT_ERR_NOTEMPTY:
-+ return "QM_MCR_RESULT_ERR_NOTEMPTY";
-+ case QM_MCR_RESULT_PENDING:
-+ return "QM_MCR_RESULT_PENDING";
-+ case QM_MCR_RESULT_ERR_BADCOMMAND:
-+ return "QM_MCR_RESULT_ERR_BADCOMMAND";
-+ }
-+ return "<unknown MCR result>";
-+}
-+
-+int qman_create_fq(u32 fqid, u32 flags, struct qman_fq *fq)
-+{
-+ struct qm_fqd fqd;
-+ struct qm_mcr_queryfq_np np;
-+ struct qm_mc_command *mcc;
-+ struct qm_mc_result *mcr;
-+ struct qman_portal *p;
-+ unsigned long irqflags __maybe_unused;
-+
-+ if (flags & QMAN_FQ_FLAG_DYNAMIC_FQID) {
-+ int ret = qman_alloc_fqid(&fqid);
-+ if (ret)
-+ return ret;
-+ }
-+ spin_lock_init(&fq->fqlock);
-+ fq->fqid = fqid;
-+ fq->flags = flags;
-+ fq->state = qman_fq_state_oos;
-+ fq->cgr_groupid = 0;
-+#ifdef CONFIG_FSL_QMAN_FQ_LOOKUP
-+ if (unlikely(find_empty_fq_table_entry(&fq->key, fq)))
-+ return -ENOMEM;
-+#endif
-+ if (!(flags & QMAN_FQ_FLAG_AS_IS) || (flags & QMAN_FQ_FLAG_NO_MODIFY))
-+ return 0;
-+ /* Everything else is AS_IS support */
-+ p = get_affine_portal();
-+ PORTAL_IRQ_LOCK(p, irqflags);
-+ mcc = qm_mc_start(&p->p);
-+ mcc->queryfq.fqid = cpu_to_be32(fqid);
-+ qm_mc_commit(&p->p, QM_MCC_VERB_QUERYFQ);
-+ while (!(mcr = qm_mc_result(&p->p)))
-+ cpu_relax();
-+ DPA_ASSERT((mcr->verb & QM_MCR_VERB_MASK) == QM_MCC_VERB_QUERYFQ);
-+ if (mcr->result != QM_MCR_RESULT_OK) {
-+ pr_err("QUERYFQ failed: %s\n", mcr_result_str(mcr->result));
-+ goto err;
-+ }
-+ fqd = mcr->queryfq.fqd;
-+ hw_fqd_to_cpu(&fqd);
-+ mcc = qm_mc_start(&p->p);
-+ mcc->queryfq_np.fqid = cpu_to_be32(fqid);
-+ qm_mc_commit(&p->p, QM_MCC_VERB_QUERYFQ_NP);
-+ while (!(mcr = qm_mc_result(&p->p)))
-+ cpu_relax();
-+ DPA_ASSERT((mcr->verb & QM_MCR_VERB_MASK) == QM_MCC_VERB_QUERYFQ_NP);
-+ if (mcr->result != QM_MCR_RESULT_OK) {
-+ pr_err("QUERYFQ_NP failed: %s\n", mcr_result_str(mcr->result));
-+ goto err;
-+ }
-+ np = mcr->queryfq_np;
-+ /* Phew, have queryfq and queryfq_np results, stitch together
-+ * the FQ object from those. */
-+ fq->cgr_groupid = fqd.cgid;
-+ switch (np.state & QM_MCR_NP_STATE_MASK) {
-+ case QM_MCR_NP_STATE_OOS:
-+ break;
-+ case QM_MCR_NP_STATE_RETIRED:
-+ fq->state = qman_fq_state_retired;
-+ if (np.frm_cnt)
-+ fq_set(fq, QMAN_FQ_STATE_NE);
-+ break;
-+ case QM_MCR_NP_STATE_TEN_SCHED:
-+ case QM_MCR_NP_STATE_TRU_SCHED:
-+ case QM_MCR_NP_STATE_ACTIVE:
-+ fq->state = qman_fq_state_sched;
-+ if (np.state & QM_MCR_NP_STATE_R)
-+ fq_set(fq, QMAN_FQ_STATE_CHANGING);
-+ break;
-+ case QM_MCR_NP_STATE_PARKED:
-+ fq->state = qman_fq_state_parked;
-+ break;
-+ default:
-+ DPA_ASSERT(NULL == "invalid FQ state");
-+ }
-+ if (fqd.fq_ctrl & QM_FQCTRL_CGE)
-+ fq->state |= QMAN_FQ_STATE_CGR_EN;
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+ put_affine_portal();
-+ return 0;
-+err:
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+ put_affine_portal();
-+ if (flags & QMAN_FQ_FLAG_DYNAMIC_FQID)
-+ qman_release_fqid(fqid);
-+ return -EIO;
-+}
-+EXPORT_SYMBOL(qman_create_fq);
-+
-+void qman_destroy_fq(struct qman_fq *fq, u32 flags __maybe_unused)
-+{
-+
-+ /* We don't need to lock the FQ as it is a pre-condition that the FQ be
-+ * quiesced. Instead, run some checks. */
-+ switch (fq->state) {
-+ case qman_fq_state_parked:
-+ DPA_ASSERT(flags & QMAN_FQ_DESTROY_PARKED);
-+ case qman_fq_state_oos:
-+ if (fq_isset(fq, QMAN_FQ_FLAG_DYNAMIC_FQID))
-+ qman_release_fqid(fq->fqid);
-+#ifdef CONFIG_FSL_QMAN_FQ_LOOKUP
-+ clear_fq_table_entry(fq->key);
-+#endif
-+ return;
-+ default:
-+ break;
-+ }
-+ DPA_ASSERT(NULL == "qman_free_fq() on unquiesced FQ!");
-+}
-+EXPORT_SYMBOL(qman_destroy_fq);
-+
-+u32 qman_fq_fqid(struct qman_fq *fq)
-+{
-+ return fq->fqid;
-+}
-+EXPORT_SYMBOL(qman_fq_fqid);
-+
-+void qman_fq_state(struct qman_fq *fq, enum qman_fq_state *state, u32 *flags)
-+{
-+ if (state)
-+ *state = fq->state;
-+ if (flags)
-+ *flags = fq->flags;
-+}
-+EXPORT_SYMBOL(qman_fq_state);
-+
-+int qman_init_fq(struct qman_fq *fq, u32 flags, struct qm_mcc_initfq *opts)
-+{
-+ struct qm_mc_command *mcc;
-+ struct qm_mc_result *mcr;
-+ struct qman_portal *p;
-+ unsigned long irqflags __maybe_unused;
-+ u8 res, myverb = (flags & QMAN_INITFQ_FLAG_SCHED) ?
-+ QM_MCC_VERB_INITFQ_SCHED : QM_MCC_VERB_INITFQ_PARKED;
-+
-+ if ((fq->state != qman_fq_state_oos) &&
-+ (fq->state != qman_fq_state_parked))
-+ return -EINVAL;
-+#ifdef CONFIG_FSL_DPA_CHECKING
-+ if (unlikely(fq_isset(fq, QMAN_FQ_FLAG_NO_MODIFY)))
-+ return -EINVAL;
-+#endif
-+ if (opts && (opts->we_mask & QM_INITFQ_WE_OAC)) {
-+ /* And can't be set at the same time as TDTHRESH */
-+ if (opts->we_mask & QM_INITFQ_WE_TDTHRESH)
-+ return -EINVAL;
-+ }
-+ /* Issue an INITFQ_[PARKED|SCHED] management command */
-+ p = get_affine_portal();
-+ PORTAL_IRQ_LOCK(p, irqflags);
-+ FQLOCK(fq);
-+ if (unlikely((fq_isset(fq, QMAN_FQ_STATE_CHANGING)) ||
-+ ((fq->state != qman_fq_state_oos) &&
-+ (fq->state != qman_fq_state_parked)))) {
-+ FQUNLOCK(fq);
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+ put_affine_portal();
-+ return -EBUSY;
-+ }
-+ mcc = qm_mc_start(&p->p);
-+ if (opts)
-+ mcc->initfq = *opts;
-+ mcc->initfq.fqid = cpu_to_be32(fq->fqid);
-+ mcc->initfq.count = 0;
-+
-+ /* If the FQ does *not* have the TO_DCPORTAL flag, contextB is set as a
-+ * demux pointer. Otherwise, the caller-provided value is allowed to
-+ * stand, don't overwrite it. */
-+ if (fq_isclear(fq, QMAN_FQ_FLAG_TO_DCPORTAL)) {
-+ dma_addr_t phys_fq;
-+ mcc->initfq.we_mask |= QM_INITFQ_WE_CONTEXTB;
-+#ifdef CONFIG_FSL_QMAN_FQ_LOOKUP
-+ mcc->initfq.fqd.context_b = fq->key;
-+#else
-+ mcc->initfq.fqd.context_b = (u32)(uintptr_t)fq;
-+#endif
-+ /* and the physical address - NB, if the user wasn't trying to
-+ * set CONTEXTA, clear the stashing settings. */
-+ if (!(mcc->initfq.we_mask & QM_INITFQ_WE_CONTEXTA)) {
-+ mcc->initfq.we_mask |= QM_INITFQ_WE_CONTEXTA;
-+ memset(&mcc->initfq.fqd.context_a, 0,
-+ sizeof(mcc->initfq.fqd.context_a));
-+ } else {
-+ phys_fq = dma_map_single(&p->pdev->dev, fq, sizeof(*fq),
-+ DMA_TO_DEVICE);
-+ qm_fqd_stashing_set64(&mcc->initfq.fqd, phys_fq);
-+ }
-+ }
-+ if (flags & QMAN_INITFQ_FLAG_LOCAL) {
-+ mcc->initfq.fqd.dest.channel = p->config->public_cfg.channel;
-+ if (!(mcc->initfq.we_mask & QM_INITFQ_WE_DESTWQ)) {
-+ mcc->initfq.we_mask |= QM_INITFQ_WE_DESTWQ;
-+ mcc->initfq.fqd.dest.wq = 4;
-+ }
-+ }
-+ mcc->initfq.we_mask = cpu_to_be16(mcc->initfq.we_mask);
-+ cpu_to_hw_fqd(&mcc->initfq.fqd);
-+ qm_mc_commit(&p->p, myverb);
-+ while (!(mcr = qm_mc_result(&p->p)))
-+ cpu_relax();
-+ DPA_ASSERT((mcr->verb & QM_MCR_VERB_MASK) == myverb);
-+ res = mcr->result;
-+ if (res != QM_MCR_RESULT_OK) {
-+ FQUNLOCK(fq);
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+ put_affine_portal();
-+ return -EIO;
-+ }
-+ if (opts) {
-+ if (opts->we_mask & QM_INITFQ_WE_FQCTRL) {
-+ if (opts->fqd.fq_ctrl & QM_FQCTRL_CGE)
-+ fq_set(fq, QMAN_FQ_STATE_CGR_EN);
-+ else
-+ fq_clear(fq, QMAN_FQ_STATE_CGR_EN);
-+ }
-+ if (opts->we_mask & QM_INITFQ_WE_CGID)
-+ fq->cgr_groupid = opts->fqd.cgid;
-+ }
-+ fq->state = (flags & QMAN_INITFQ_FLAG_SCHED) ?
-+ qman_fq_state_sched : qman_fq_state_parked;
-+ FQUNLOCK(fq);
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+ put_affine_portal();
-+ return 0;
-+}
-+EXPORT_SYMBOL(qman_init_fq);
-+
-+int qman_schedule_fq(struct qman_fq *fq)
-+{
-+ struct qm_mc_command *mcc;
-+ struct qm_mc_result *mcr;
-+ struct qman_portal *p;
-+ unsigned long irqflags __maybe_unused;
-+ int ret = 0;
-+ u8 res;
-+
-+ if (fq->state != qman_fq_state_parked)
-+ return -EINVAL;
-+#ifdef CONFIG_FSL_DPA_CHECKING
-+ if (unlikely(fq_isset(fq, QMAN_FQ_FLAG_NO_MODIFY)))
-+ return -EINVAL;
-+#endif
-+ /* Issue a ALTERFQ_SCHED management command */
-+ p = get_affine_portal();
-+ PORTAL_IRQ_LOCK(p, irqflags);
-+ FQLOCK(fq);
-+ if (unlikely((fq_isset(fq, QMAN_FQ_STATE_CHANGING)) ||
-+ (fq->state != qman_fq_state_parked))) {
-+ ret = -EBUSY;
-+ goto out;
-+ }
-+ mcc = qm_mc_start(&p->p);
-+ mcc->alterfq.fqid = cpu_to_be32(fq->fqid);
-+ qm_mc_commit(&p->p, QM_MCC_VERB_ALTER_SCHED);
-+ while (!(mcr = qm_mc_result(&p->p)))
-+ cpu_relax();
-+ DPA_ASSERT((mcr->verb & QM_MCR_VERB_MASK) == QM_MCR_VERB_ALTER_SCHED);
-+ res = mcr->result;
-+ if (res != QM_MCR_RESULT_OK) {
-+ ret = -EIO;
-+ goto out;
-+ }
-+ fq->state = qman_fq_state_sched;
-+out:
-+ FQUNLOCK(fq);
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+ put_affine_portal();
-+ return ret;
-+}
-+EXPORT_SYMBOL(qman_schedule_fq);
-+
-+int qman_retire_fq(struct qman_fq *fq, u32 *flags)
-+{
-+ struct qm_mc_command *mcc;
-+ struct qm_mc_result *mcr;
-+ struct qman_portal *p;
-+ unsigned long irqflags __maybe_unused;
-+ int rval;
-+ u8 res;
-+
-+ if ((fq->state != qman_fq_state_parked) &&
-+ (fq->state != qman_fq_state_sched))
-+ return -EINVAL;
-+#ifdef CONFIG_FSL_DPA_CHECKING
-+ if (unlikely(fq_isset(fq, QMAN_FQ_FLAG_NO_MODIFY)))
-+ return -EINVAL;
-+#endif
-+ p = get_affine_portal();
-+ PORTAL_IRQ_LOCK(p, irqflags);
-+ FQLOCK(fq);
-+ if (unlikely((fq_isset(fq, QMAN_FQ_STATE_CHANGING)) ||
-+ (fq->state == qman_fq_state_retired) ||
-+ (fq->state == qman_fq_state_oos))) {
-+ rval = -EBUSY;
-+ goto out;
-+ }
-+ rval = table_push_fq(p, fq);
-+ if (rval)
-+ goto out;
-+ mcc = qm_mc_start(&p->p);
-+ mcc->alterfq.fqid = cpu_to_be32(fq->fqid);
-+ qm_mc_commit(&p->p, QM_MCC_VERB_ALTER_RETIRE);
-+ while (!(mcr = qm_mc_result(&p->p)))
-+ cpu_relax();
-+ DPA_ASSERT((mcr->verb & QM_MCR_VERB_MASK) == QM_MCR_VERB_ALTER_RETIRE);
-+ res = mcr->result;
-+ /* "Elegant" would be to treat OK/PENDING the same way; set CHANGING,
-+ * and defer the flags until FQRNI or FQRN (respectively) show up. But
-+ * "Friendly" is to process OK immediately, and not set CHANGING. We do
-+ * friendly, otherwise the caller doesn't necessarily have a fully
-+ * "retired" FQ on return even if the retirement was immediate. However
-+ * this does mean some code duplication between here and
-+ * fq_state_change(). */
-+ if (likely(res == QM_MCR_RESULT_OK)) {
-+ rval = 0;
-+ /* Process 'fq' right away, we'll ignore FQRNI */
-+ if (mcr->alterfq.fqs & QM_MCR_FQS_NOTEMPTY)
-+ fq_set(fq, QMAN_FQ_STATE_NE);
-+ if (mcr->alterfq.fqs & QM_MCR_FQS_ORLPRESENT)
-+ fq_set(fq, QMAN_FQ_STATE_ORL);
-+ else
-+ table_del_fq(p, fq);
-+ if (flags)
-+ *flags = fq->flags;
-+ fq->state = qman_fq_state_retired;
-+ if (fq->cb.fqs) {
-+ /* Another issue with supporting "immediate" retirement
-+ * is that we're forced to drop FQRNIs, because by the
-+ * time they're seen it may already be "too late" (the
-+ * fq may have been OOS'd and free()'d already). But if
-+ * the upper layer wants a callback whether it's
-+ * immediate or not, we have to fake a "MR" entry to
-+ * look like an FQRNI... */
-+ struct qm_mr_entry msg;
-+ msg.verb = QM_MR_VERB_FQRNI;
-+ msg.fq.fqs = mcr->alterfq.fqs;
-+ msg.fq.fqid = fq->fqid;
-+#ifdef CONFIG_FSL_QMAN_FQ_LOOKUP
-+ msg.fq.contextB = fq->key;
-+#else
-+ msg.fq.contextB = (u32)(uintptr_t)fq;
-+#endif
-+ fq->cb.fqs(p, fq, &msg);
-+ }
-+ } else if (res == QM_MCR_RESULT_PENDING) {
-+ rval = 1;
-+ fq_set(fq, QMAN_FQ_STATE_CHANGING);
-+ } else {
-+ rval = -EIO;
-+ table_del_fq(p, fq);
-+ }
-+out:
-+ FQUNLOCK(fq);
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+ put_affine_portal();
-+ return rval;
-+}
-+EXPORT_SYMBOL(qman_retire_fq);
-+
-+int qman_oos_fq(struct qman_fq *fq)
-+{
-+ struct qm_mc_command *mcc;
-+ struct qm_mc_result *mcr;
-+ struct qman_portal *p;
-+ unsigned long irqflags __maybe_unused;
-+ int ret = 0;
-+ u8 res;
-+
-+ if (fq->state != qman_fq_state_retired)
-+ return -EINVAL;
-+#ifdef CONFIG_FSL_DPA_CHECKING
-+ if (unlikely(fq_isset(fq, QMAN_FQ_FLAG_NO_MODIFY)))
-+ return -EINVAL;
-+#endif
-+ p = get_affine_portal();
-+ PORTAL_IRQ_LOCK(p, irqflags);
-+ FQLOCK(fq);
-+ if (unlikely((fq_isset(fq, QMAN_FQ_STATE_BLOCKOOS)) ||
-+ (fq->state != qman_fq_state_retired))) {
-+ ret = -EBUSY;
-+ goto out;
-+ }
-+ mcc = qm_mc_start(&p->p);
-+ mcc->alterfq.fqid = cpu_to_be32(fq->fqid);
-+ qm_mc_commit(&p->p, QM_MCC_VERB_ALTER_OOS);
-+ while (!(mcr = qm_mc_result(&p->p)))
-+ cpu_relax();
-+ DPA_ASSERT((mcr->verb & QM_MCR_VERB_MASK) == QM_MCR_VERB_ALTER_OOS);
-+ res = mcr->result;
-+ if (res != QM_MCR_RESULT_OK) {
-+ ret = -EIO;
-+ goto out;
-+ }
-+ fq->state = qman_fq_state_oos;
-+out:
-+ FQUNLOCK(fq);
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+ put_affine_portal();
-+ return ret;
-+}
-+EXPORT_SYMBOL(qman_oos_fq);
-+
-+int qman_fq_flow_control(struct qman_fq *fq, int xon)
-+{
-+ struct qm_mc_command *mcc;
-+ struct qm_mc_result *mcr;
-+ struct qman_portal *p;
-+ unsigned long irqflags __maybe_unused;
-+ int ret = 0;
-+ u8 res;
-+ u8 myverb;
-+
-+ if ((fq->state == qman_fq_state_oos) ||
-+ (fq->state == qman_fq_state_retired) ||
-+ (fq->state == qman_fq_state_parked))
-+ return -EINVAL;
-+
-+#ifdef CONFIG_FSL_DPA_CHECKING
-+ if (unlikely(fq_isset(fq, QMAN_FQ_FLAG_NO_MODIFY)))
-+ return -EINVAL;
-+#endif
-+ /* Issue a ALTER_FQXON or ALTER_FQXOFF management command */
-+ p = get_affine_portal();
-+ PORTAL_IRQ_LOCK(p, irqflags);
-+ FQLOCK(fq);
-+ if (unlikely((fq_isset(fq, QMAN_FQ_STATE_CHANGING)) ||
-+ (fq->state == qman_fq_state_parked) ||
-+ (fq->state == qman_fq_state_oos) ||
-+ (fq->state == qman_fq_state_retired))) {
-+ ret = -EBUSY;
-+ goto out;
-+ }
-+ mcc = qm_mc_start(&p->p);
-+ mcc->alterfq.fqid = fq->fqid;
-+ mcc->alterfq.count = 0;
-+ myverb = xon ? QM_MCC_VERB_ALTER_FQXON : QM_MCC_VERB_ALTER_FQXOFF;
-+
-+ qm_mc_commit(&p->p, myverb);
-+ while (!(mcr = qm_mc_result(&p->p)))
-+ cpu_relax();
-+ DPA_ASSERT((mcr->verb & QM_MCR_VERB_MASK) == myverb);
-+
-+ res = mcr->result;
-+ if (res != QM_MCR_RESULT_OK) {
-+ ret = -EIO;
-+ goto out;
-+ }
-+out:
-+ FQUNLOCK(fq);
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+ put_affine_portal();
-+ return ret;
-+}
-+EXPORT_SYMBOL(qman_fq_flow_control);
-+
-+int qman_query_fq(struct qman_fq *fq, struct qm_fqd *fqd)
-+{
-+ struct qm_mc_command *mcc;
-+ struct qm_mc_result *mcr;
-+ struct qman_portal *p = get_affine_portal();
-+ unsigned long irqflags __maybe_unused;
-+ u8 res;
-+
-+ PORTAL_IRQ_LOCK(p, irqflags);
-+ mcc = qm_mc_start(&p->p);
-+ mcc->queryfq.fqid = cpu_to_be32(fq->fqid);
-+ qm_mc_commit(&p->p, QM_MCC_VERB_QUERYFQ);
-+ while (!(mcr = qm_mc_result(&p->p)))
-+ cpu_relax();
-+ DPA_ASSERT((mcr->verb & QM_MCR_VERB_MASK) == QM_MCR_VERB_QUERYFQ);
-+ res = mcr->result;
-+ if (res == QM_MCR_RESULT_OK)
-+ *fqd = mcr->queryfq.fqd;
-+ hw_fqd_to_cpu(fqd);
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+ put_affine_portal();
-+ if (res != QM_MCR_RESULT_OK)
-+ return -EIO;
-+ return 0;
-+}
-+EXPORT_SYMBOL(qman_query_fq);
-+
-+int qman_query_fq_np(struct qman_fq *fq, struct qm_mcr_queryfq_np *np)
-+{
-+ struct qm_mc_command *mcc;
-+ struct qm_mc_result *mcr;
-+ struct qman_portal *p = get_affine_portal();
-+ unsigned long irqflags __maybe_unused;
-+ u8 res;
-+
-+ PORTAL_IRQ_LOCK(p, irqflags);
-+ mcc = qm_mc_start(&p->p);
-+ mcc->queryfq.fqid = cpu_to_be32(fq->fqid);
-+ qm_mc_commit(&p->p, QM_MCC_VERB_QUERYFQ_NP);
-+ while (!(mcr = qm_mc_result(&p->p)))
-+ cpu_relax();
-+ DPA_ASSERT((mcr->verb & QM_MCR_VERB_MASK) == QM_MCR_VERB_QUERYFQ_NP);
-+ res = mcr->result;
-+ if (res == QM_MCR_RESULT_OK) {
-+ *np = mcr->queryfq_np;
-+ np->fqd_link = be24_to_cpu(np->fqd_link);
-+ np->odp_seq = be16_to_cpu(np->odp_seq);
-+ np->orp_nesn = be16_to_cpu(np->orp_nesn);
-+ np->orp_ea_hseq = be16_to_cpu(np->orp_ea_hseq);
-+ np->orp_ea_tseq = be16_to_cpu(np->orp_ea_tseq);
-+ np->orp_ea_hptr = be24_to_cpu(np->orp_ea_hptr);
-+ np->orp_ea_tptr = be24_to_cpu(np->orp_ea_tptr);
-+ np->pfdr_hptr = be24_to_cpu(np->pfdr_hptr);
-+ np->pfdr_tptr = be24_to_cpu(np->pfdr_tptr);
-+ np->ics_surp = be16_to_cpu(np->ics_surp);
-+ np->byte_cnt = be32_to_cpu(np->byte_cnt);
-+ np->frm_cnt = be24_to_cpu(np->frm_cnt);
-+ np->ra1_sfdr = be16_to_cpu(np->ra1_sfdr);
-+ np->ra2_sfdr = be16_to_cpu(np->ra2_sfdr);
-+ np->od1_sfdr = be16_to_cpu(np->od1_sfdr);
-+ np->od2_sfdr = be16_to_cpu(np->od2_sfdr);
-+ np->od3_sfdr = be16_to_cpu(np->od3_sfdr);
-+ }
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+ put_affine_portal();
-+ if (res == QM_MCR_RESULT_ERR_FQID)
-+ return -ERANGE;
-+ else if (res != QM_MCR_RESULT_OK)
-+ return -EIO;
-+ return 0;
-+}
-+EXPORT_SYMBOL(qman_query_fq_np);
-+
-+int qman_query_wq(u8 query_dedicated, struct qm_mcr_querywq *wq)
-+{
-+ struct qm_mc_command *mcc;
-+ struct qm_mc_result *mcr;
-+ struct qman_portal *p = get_affine_portal();
-+ unsigned long irqflags __maybe_unused;
-+ u8 res, myverb;
-+
-+ PORTAL_IRQ_LOCK(p, irqflags);
-+ myverb = (query_dedicated) ? QM_MCR_VERB_QUERYWQ_DEDICATED :
-+ QM_MCR_VERB_QUERYWQ;
-+ mcc = qm_mc_start(&p->p);
-+ mcc->querywq.channel.id = cpu_to_be16(wq->channel.id);
-+ qm_mc_commit(&p->p, myverb);
-+ while (!(mcr = qm_mc_result(&p->p)))
-+ cpu_relax();
-+ DPA_ASSERT((mcr->verb & QM_MCR_VERB_MASK) == myverb);
-+ res = mcr->result;
-+ if (res == QM_MCR_RESULT_OK) {
-+ int i, array_len;
-+ wq->channel.id = be16_to_cpu(mcr->querywq.channel.id);
-+ array_len = ARRAY_SIZE(mcr->querywq.wq_len);
-+ for (i = 0; i < array_len; i++)
-+ wq->wq_len[i] = be32_to_cpu(mcr->querywq.wq_len[i]);
-+ }
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+ put_affine_portal();
-+ if (res != QM_MCR_RESULT_OK) {
-+ pr_err("QUERYWQ failed: %s\n", mcr_result_str(res));
-+ return -EIO;
-+ }
-+ return 0;
-+}
-+EXPORT_SYMBOL(qman_query_wq);
-+
-+int qman_testwrite_cgr(struct qman_cgr *cgr, u64 i_bcnt,
-+ struct qm_mcr_cgrtestwrite *result)
-+{
-+ struct qm_mc_command *mcc;
-+ struct qm_mc_result *mcr;
-+ struct qman_portal *p = get_affine_portal();
-+ unsigned long irqflags __maybe_unused;
-+ u8 res;
-+
-+ PORTAL_IRQ_LOCK(p, irqflags);
-+ mcc = qm_mc_start(&p->p);
-+ mcc->cgrtestwrite.cgid = cgr->cgrid;
-+ mcc->cgrtestwrite.i_bcnt_hi = (u8)(i_bcnt >> 32);
-+ mcc->cgrtestwrite.i_bcnt_lo = (u32)i_bcnt;
-+ qm_mc_commit(&p->p, QM_MCC_VERB_CGRTESTWRITE);
-+ while (!(mcr = qm_mc_result(&p->p)))
-+ cpu_relax();
-+ DPA_ASSERT((mcr->verb & QM_MCR_VERB_MASK) == QM_MCC_VERB_CGRTESTWRITE);
-+ res = mcr->result;
-+ if (res == QM_MCR_RESULT_OK)
-+ *result = mcr->cgrtestwrite;
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+ put_affine_portal();
-+ if (res != QM_MCR_RESULT_OK) {
-+ pr_err("CGR TEST WRITE failed: %s\n", mcr_result_str(res));
-+ return -EIO;
-+ }
-+ return 0;
-+}
-+EXPORT_SYMBOL(qman_testwrite_cgr);
-+
-+int qman_query_cgr(struct qman_cgr *cgr, struct qm_mcr_querycgr *cgrd)
-+{
-+ struct qm_mc_command *mcc;
-+ struct qm_mc_result *mcr;
-+ struct qman_portal *p = get_affine_portal();
-+ unsigned long irqflags __maybe_unused;
-+ u8 res;
-+ int i;
-+
-+ PORTAL_IRQ_LOCK(p, irqflags);
-+ mcc = qm_mc_start(&p->p);
-+ mcc->querycgr.cgid = cgr->cgrid;
-+ qm_mc_commit(&p->p, QM_MCC_VERB_QUERYCGR);
-+ while (!(mcr = qm_mc_result(&p->p)))
-+ cpu_relax();
-+ DPA_ASSERT((mcr->verb & QM_MCR_VERB_MASK) == QM_MCC_VERB_QUERYCGR);
-+ res = mcr->result;
-+ if (res == QM_MCR_RESULT_OK)
-+ *cgrd = mcr->querycgr;
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+ put_affine_portal();
-+ if (res != QM_MCR_RESULT_OK) {
-+ pr_err("QUERY_CGR failed: %s\n", mcr_result_str(res));
-+ return -EIO;
-+ }
-+ cgrd->cgr.wr_parm_g.word =
-+ be32_to_cpu(cgrd->cgr.wr_parm_g.word);
-+ cgrd->cgr.wr_parm_y.word =
-+ be32_to_cpu(cgrd->cgr.wr_parm_y.word);
-+ cgrd->cgr.wr_parm_r.word =
-+ be32_to_cpu(cgrd->cgr.wr_parm_r.word);
-+ cgrd->cgr.cscn_targ = be32_to_cpu(cgrd->cgr.cscn_targ);
-+ cgrd->cgr.__cs_thres = be16_to_cpu(cgrd->cgr.__cs_thres);
-+ for (i = 0; i < ARRAY_SIZE(cgrd->cscn_targ_swp); i++)
-+ be32_to_cpus(&cgrd->cscn_targ_swp[i]);
-+ return 0;
-+}
-+EXPORT_SYMBOL(qman_query_cgr);
-+
-+int qman_query_congestion(struct qm_mcr_querycongestion *congestion)
-+{
-+ struct qm_mc_result *mcr;
-+ struct qman_portal *p = get_affine_portal();
-+ unsigned long irqflags __maybe_unused;
-+ u8 res;
-+ int i;
-+
-+ PORTAL_IRQ_LOCK(p, irqflags);
-+ qm_mc_start(&p->p);
-+ qm_mc_commit(&p->p, QM_MCC_VERB_QUERYCONGESTION);
-+ while (!(mcr = qm_mc_result(&p->p)))
-+ cpu_relax();
-+ DPA_ASSERT((mcr->verb & QM_MCR_VERB_MASK) ==
-+ QM_MCC_VERB_QUERYCONGESTION);
-+ res = mcr->result;
-+ if (res == QM_MCR_RESULT_OK)
-+ memcpy_fromio(congestion, &mcr->querycongestion,
-+ sizeof(*congestion));
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+ put_affine_portal();
-+ if (res != QM_MCR_RESULT_OK) {
-+ pr_err("QUERY_CONGESTION failed: %s\n", mcr_result_str(res));
-+ return -EIO;
-+ }
-+
-+ for (i = 0; i < ARRAY_SIZE(congestion->state.__state); i++)
-+ be32_to_cpus(&congestion->state.__state[i]);
-+ return 0;
-+}
-+EXPORT_SYMBOL(qman_query_congestion);
-+
-+/* internal function used as a wait_event() expression */
-+static int set_p_vdqcr(struct qman_portal *p, struct qman_fq *fq, u32 vdqcr)
-+{
-+ unsigned long irqflags __maybe_unused;
-+ int ret = -EBUSY;
-+ PORTAL_IRQ_LOCK(p, irqflags);
-+ if (!p->vdqcr_owned) {
-+ FQLOCK(fq);
-+ if (fq_isset(fq, QMAN_FQ_STATE_VDQCR))
-+ goto escape;
-+ fq_set(fq, QMAN_FQ_STATE_VDQCR);
-+ FQUNLOCK(fq);
-+ p->vdqcr_owned = fq;
-+ ret = 0;
-+ }
-+escape:
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+ if (!ret)
-+ qm_dqrr_vdqcr_set(&p->p, vdqcr);
-+ return ret;
-+}
-+
-+static int set_vdqcr(struct qman_portal **p, struct qman_fq *fq, u32 vdqcr)
-+{
-+ int ret;
-+ *p = get_affine_portal();
-+ ret = set_p_vdqcr(*p, fq, vdqcr);
-+ put_affine_portal();
-+ return ret;
-+}
-+
-+#ifdef CONFIG_FSL_DPA_CAN_WAIT
-+static int wait_p_vdqcr_start(struct qman_portal *p, struct qman_fq *fq,
-+ u32 vdqcr, u32 flags)
-+{
-+ int ret = 0;
-+ if (flags & QMAN_VOLATILE_FLAG_WAIT_INT)
-+ ret = wait_event_interruptible(affine_queue,
-+ !(ret = set_p_vdqcr(p, fq, vdqcr)));
-+ else
-+ wait_event(affine_queue, !(ret = set_p_vdqcr(p, fq, vdqcr)));
-+ return ret;
-+}
-+
-+static int wait_vdqcr_start(struct qman_portal **p, struct qman_fq *fq,
-+ u32 vdqcr, u32 flags)
-+{
-+ int ret = 0;
-+ if (flags & QMAN_VOLATILE_FLAG_WAIT_INT)
-+ ret = wait_event_interruptible(affine_queue,
-+ !(ret = set_vdqcr(p, fq, vdqcr)));
-+ else
-+ wait_event(affine_queue, !(ret = set_vdqcr(p, fq, vdqcr)));
-+ return ret;
-+}
-+#endif
-+
-+int qman_p_volatile_dequeue(struct qman_portal *p, struct qman_fq *fq,
-+ u32 flags __maybe_unused, u32 vdqcr)
-+{
-+ int ret;
-+
-+ if ((fq->state != qman_fq_state_parked) &&
-+ (fq->state != qman_fq_state_retired))
-+ return -EINVAL;
-+ if (vdqcr & QM_VDQCR_FQID_MASK)
-+ return -EINVAL;
-+ if (fq_isset(fq, QMAN_FQ_STATE_VDQCR))
-+ return -EBUSY;
-+ vdqcr = (vdqcr & ~QM_VDQCR_FQID_MASK) | fq->fqid;
-+#ifdef CONFIG_FSL_DPA_CAN_WAIT
-+ if (flags & QMAN_VOLATILE_FLAG_WAIT)
-+ ret = wait_p_vdqcr_start(p, fq, vdqcr, flags);
-+ else
-+#endif
-+ ret = set_p_vdqcr(p, fq, vdqcr);
-+ if (ret)
-+ return ret;
-+ /* VDQCR is set */
-+#ifdef CONFIG_FSL_DPA_CAN_WAIT
-+ if (flags & QMAN_VOLATILE_FLAG_FINISH) {
-+ if (flags & QMAN_VOLATILE_FLAG_WAIT_INT)
-+ /* NB: don't propagate any error - the caller wouldn't
-+ * know whether the VDQCR was issued or not. A signal
-+ * could arrive after returning anyway, so the caller
-+ * can check signal_pending() if that's an issue. */
-+ wait_event_interruptible(affine_queue,
-+ !fq_isset(fq, QMAN_FQ_STATE_VDQCR));
-+ else
-+ wait_event(affine_queue,
-+ !fq_isset(fq, QMAN_FQ_STATE_VDQCR));
-+ }
-+#endif
-+ return 0;
-+}
-+EXPORT_SYMBOL(qman_p_volatile_dequeue);
-+
-+int qman_volatile_dequeue(struct qman_fq *fq, u32 flags __maybe_unused,
-+ u32 vdqcr)
-+{
-+ struct qman_portal *p;
-+ int ret;
-+
-+ if ((fq->state != qman_fq_state_parked) &&
-+ (fq->state != qman_fq_state_retired))
-+ return -EINVAL;
-+ if (vdqcr & QM_VDQCR_FQID_MASK)
-+ return -EINVAL;
-+ if (fq_isset(fq, QMAN_FQ_STATE_VDQCR))
-+ return -EBUSY;
-+ vdqcr = (vdqcr & ~QM_VDQCR_FQID_MASK) | fq->fqid;
-+#ifdef CONFIG_FSL_DPA_CAN_WAIT
-+ if (flags & QMAN_VOLATILE_FLAG_WAIT)
-+ ret = wait_vdqcr_start(&p, fq, vdqcr, flags);
-+ else
-+#endif
-+ ret = set_vdqcr(&p, fq, vdqcr);
-+ if (ret)
-+ return ret;
-+ /* VDQCR is set */
-+#ifdef CONFIG_FSL_DPA_CAN_WAIT
-+ if (flags & QMAN_VOLATILE_FLAG_FINISH) {
-+ if (flags & QMAN_VOLATILE_FLAG_WAIT_INT)
-+ /* NB: don't propagate any error - the caller wouldn't
-+ * know whether the VDQCR was issued or not. A signal
-+ * could arrive after returning anyway, so the caller
-+ * can check signal_pending() if that's an issue. */
-+ wait_event_interruptible(affine_queue,
-+ !fq_isset(fq, QMAN_FQ_STATE_VDQCR));
-+ else
-+ wait_event(affine_queue,
-+ !fq_isset(fq, QMAN_FQ_STATE_VDQCR));
-+ }
-+#endif
-+ return 0;
-+}
-+EXPORT_SYMBOL(qman_volatile_dequeue);
-+
-+static noinline void update_eqcr_ci(struct qman_portal *p, u8 avail)
-+{
-+ if (avail)
-+ qm_eqcr_cce_prefetch(&p->p);
-+ else
-+ qm_eqcr_cce_update(&p->p);
-+}
-+
-+int qman_eqcr_is_empty(void)
-+{
-+ unsigned long irqflags __maybe_unused;
-+ struct qman_portal *p = get_affine_portal();
-+ u8 avail;
-+
-+ PORTAL_IRQ_LOCK(p, irqflags);
-+ update_eqcr_ci(p, 0);
-+ avail = qm_eqcr_get_fill(&p->p);
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+ put_affine_portal();
-+ return avail == 0;
-+}
-+EXPORT_SYMBOL(qman_eqcr_is_empty);
-+
-+void qman_set_dc_ern(qman_cb_dc_ern handler, int affine)
-+{
-+ if (affine) {
-+ unsigned long irqflags __maybe_unused;
-+ struct qman_portal *p = get_affine_portal();
-+ PORTAL_IRQ_LOCK(p, irqflags);
-+ p->cb_dc_ern = handler;
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+ put_affine_portal();
-+ } else
-+ cb_dc_ern = handler;
-+}
-+EXPORT_SYMBOL(qman_set_dc_ern);
-+
-+static inline struct qm_eqcr_entry *try_p_eq_start(struct qman_portal *p,
-+ unsigned long *irqflags __maybe_unused,
-+ struct qman_fq *fq,
-+ const struct qm_fd *fd,
-+ u32 flags)
-+{
-+ struct qm_eqcr_entry *eq;
-+ u8 avail;
-+ PORTAL_IRQ_LOCK(p, (*irqflags));
-+#ifdef CONFIG_FSL_DPA_CAN_WAIT_SYNC
-+ if (unlikely((flags & QMAN_ENQUEUE_FLAG_WAIT) &&
-+ (flags & QMAN_ENQUEUE_FLAG_WAIT_SYNC))) {
-+ if (p->eqci_owned) {
-+ PORTAL_IRQ_UNLOCK(p, (*irqflags));
-+ return NULL;
-+ }
-+ p->eqci_owned = fq;
-+ }
-+#endif
-+ if (p->use_eqcr_ci_stashing) {
-+ /*
-+ * The stashing case is easy, only update if we need to in
-+ * order to try and liberate ring entries.
-+ */
-+ eq = qm_eqcr_start_stash(&p->p);
-+ } else {
-+ /*
-+ * The non-stashing case is harder, need to prefetch ahead of
-+ * time.
-+ */
-+ avail = qm_eqcr_get_avail(&p->p);
-+ if (avail < 2)
-+ update_eqcr_ci(p, avail);
-+ eq = qm_eqcr_start_no_stash(&p->p);
-+ }
-+
-+ if (unlikely(!eq)) {
-+#ifdef CONFIG_FSL_DPA_CAN_WAIT_SYNC
-+ if (unlikely((flags & QMAN_ENQUEUE_FLAG_WAIT) &&
-+ (flags & QMAN_ENQUEUE_FLAG_WAIT_SYNC)))
-+ p->eqci_owned = NULL;
-+#endif
-+ PORTAL_IRQ_UNLOCK(p, (*irqflags));
-+ return NULL;
-+ }
-+ if (flags & QMAN_ENQUEUE_FLAG_DCA)
-+ eq->dca = QM_EQCR_DCA_ENABLE |
-+ ((flags & QMAN_ENQUEUE_FLAG_DCA_PARK) ?
-+ QM_EQCR_DCA_PARK : 0) |
-+ ((flags >> 8) & QM_EQCR_DCA_IDXMASK);
-+ eq->fqid = cpu_to_be32(fq->fqid);
-+#ifdef CONFIG_FSL_QMAN_FQ_LOOKUP
-+ eq->tag = cpu_to_be32(fq->key);
-+#else
-+ eq->tag = cpu_to_be32((u32)(uintptr_t)fq);
-+#endif
-+ eq->fd = *fd;
-+ cpu_to_hw_fd(&eq->fd);
-+ return eq;
-+}
-+
-+static inline struct qm_eqcr_entry *try_eq_start(struct qman_portal **p,
-+ unsigned long *irqflags __maybe_unused,
-+ struct qman_fq *fq,
-+ const struct qm_fd *fd,
-+ u32 flags)
-+{
-+ struct qm_eqcr_entry *eq;
-+ *p = get_affine_portal();
-+ eq = try_p_eq_start(*p, irqflags, fq, fd, flags);
-+ if (!eq)
-+ put_affine_portal();
-+ return eq;
-+}
-+
-+#ifdef CONFIG_FSL_DPA_CAN_WAIT
-+static noinline struct qm_eqcr_entry *__wait_eq_start(struct qman_portal **p,
-+ unsigned long *irqflags __maybe_unused,
-+ struct qman_fq *fq,
-+ const struct qm_fd *fd,
-+ u32 flags)
-+{
-+ struct qm_eqcr_entry *eq = try_eq_start(p, irqflags, fq, fd, flags);
-+ if (!eq)
-+ qm_eqcr_set_ithresh(&(*p)->p, EQCR_ITHRESH);
-+ return eq;
-+}
-+static noinline struct qm_eqcr_entry *wait_eq_start(struct qman_portal **p,
-+ unsigned long *irqflags __maybe_unused,
-+ struct qman_fq *fq,
-+ const struct qm_fd *fd,
-+ u32 flags)
-+{
-+ struct qm_eqcr_entry *eq;
-+ if (flags & QMAN_ENQUEUE_FLAG_WAIT_INT)
-+ /* NB: return NULL if signal occurs before completion. Signal
-+ * can occur during return. Caller must check for signal */
-+ wait_event_interruptible(affine_queue,
-+ (eq = __wait_eq_start(p, irqflags, fq, fd, flags)));
-+ else
-+ wait_event(affine_queue,
-+ (eq = __wait_eq_start(p, irqflags, fq, fd, flags)));
-+ return eq;
-+}
-+static noinline struct qm_eqcr_entry *__wait_p_eq_start(struct qman_portal *p,
-+ unsigned long *irqflags __maybe_unused,
-+ struct qman_fq *fq,
-+ const struct qm_fd *fd,
-+ u32 flags)
-+{
-+ struct qm_eqcr_entry *eq = try_p_eq_start(p, irqflags, fq, fd, flags);
-+ if (!eq)
-+ qm_eqcr_set_ithresh(&p->p, EQCR_ITHRESH);
-+ return eq;
-+}
-+static noinline struct qm_eqcr_entry *wait_p_eq_start(struct qman_portal *p,
-+ unsigned long *irqflags __maybe_unused,
-+ struct qman_fq *fq,
-+ const struct qm_fd *fd,
-+ u32 flags)
-+{
-+ struct qm_eqcr_entry *eq;
-+ if (flags & QMAN_ENQUEUE_FLAG_WAIT_INT)
-+ /* NB: return NULL if signal occurs before completion. Signal
-+ * can occur during return. Caller must check for signal */
-+ wait_event_interruptible(affine_queue,
-+ (eq = __wait_p_eq_start(p, irqflags, fq, fd, flags)));
-+ else
-+ wait_event(affine_queue,
-+ (eq = __wait_p_eq_start(p, irqflags, fq, fd, flags)));
-+ return eq;
-+}
-+#endif
-+
-+int qman_p_enqueue(struct qman_portal *p, struct qman_fq *fq,
-+ const struct qm_fd *fd, u32 flags)
-+{
-+ struct qm_eqcr_entry *eq;
-+ unsigned long irqflags __maybe_unused;
-+
-+#ifdef CONFIG_FSL_DPA_CAN_WAIT
-+ if (flags & QMAN_ENQUEUE_FLAG_WAIT)
-+ eq = wait_p_eq_start(p, &irqflags, fq, fd, flags);
-+ else
-+#endif
-+ eq = try_p_eq_start(p, &irqflags, fq, fd, flags);
-+ if (!eq)
-+ return -EBUSY;
-+ /* Note: QM_EQCR_VERB_INTERRUPT == QMAN_ENQUEUE_FLAG_WAIT_SYNC */
-+ qm_eqcr_pvb_commit(&p->p, QM_EQCR_VERB_CMD_ENQUEUE |
-+ (flags & (QM_EQCR_VERB_COLOUR_MASK | QM_EQCR_VERB_INTERRUPT)));
-+ /* Factor the below out, it's used from qman_enqueue_orp() too */
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+#ifdef CONFIG_FSL_DPA_CAN_WAIT_SYNC
-+ if (unlikely((flags & QMAN_ENQUEUE_FLAG_WAIT) &&
-+ (flags & QMAN_ENQUEUE_FLAG_WAIT_SYNC))) {
-+ if (flags & QMAN_ENQUEUE_FLAG_WAIT_INT)
-+ /* NB: return success even if signal occurs before
-+ * condition is true. pvb_commit guarantees success */
-+ wait_event_interruptible(affine_queue,
-+ (p->eqci_owned != fq));
-+ else
-+ wait_event(affine_queue, (p->eqci_owned != fq));
-+ }
-+#endif
-+ return 0;
-+}
-+EXPORT_SYMBOL(qman_p_enqueue);
-+
-+int qman_enqueue(struct qman_fq *fq, const struct qm_fd *fd, u32 flags)
-+{
-+ struct qman_portal *p;
-+ struct qm_eqcr_entry *eq;
-+ unsigned long irqflags __maybe_unused;
-+
-+#ifdef CONFIG_FSL_DPA_CAN_WAIT
-+ if (flags & QMAN_ENQUEUE_FLAG_WAIT)
-+ eq = wait_eq_start(&p, &irqflags, fq, fd, flags);
-+ else
-+#endif
-+ eq = try_eq_start(&p, &irqflags, fq, fd, flags);
-+ if (!eq)
-+ return -EBUSY;
-+ /* Note: QM_EQCR_VERB_INTERRUPT == QMAN_ENQUEUE_FLAG_WAIT_SYNC */
-+ qm_eqcr_pvb_commit(&p->p, QM_EQCR_VERB_CMD_ENQUEUE |
-+ (flags & (QM_EQCR_VERB_COLOUR_MASK | QM_EQCR_VERB_INTERRUPT)));
-+ /* Factor the below out, it's used from qman_enqueue_orp() too */
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+ put_affine_portal();
-+#ifdef CONFIG_FSL_DPA_CAN_WAIT_SYNC
-+ if (unlikely((flags & QMAN_ENQUEUE_FLAG_WAIT) &&
-+ (flags & QMAN_ENQUEUE_FLAG_WAIT_SYNC))) {
-+ if (flags & QMAN_ENQUEUE_FLAG_WAIT_INT)
-+ /* NB: return success even if signal occurs before
-+ * condition is true. pvb_commit guarantees success */
-+ wait_event_interruptible(affine_queue,
-+ (p->eqci_owned != fq));
-+ else
-+ wait_event(affine_queue, (p->eqci_owned != fq));
-+ }
-+#endif
-+ return 0;
-+}
-+EXPORT_SYMBOL(qman_enqueue);
-+
-+int qman_p_enqueue_orp(struct qman_portal *p, struct qman_fq *fq,
-+ const struct qm_fd *fd, u32 flags,
-+ struct qman_fq *orp, u16 orp_seqnum)
-+{
-+ struct qm_eqcr_entry *eq;
-+ unsigned long irqflags __maybe_unused;
-+
-+#ifdef CONFIG_FSL_DPA_CAN_WAIT
-+ if (flags & QMAN_ENQUEUE_FLAG_WAIT)
-+ eq = wait_p_eq_start(p, &irqflags, fq, fd, flags);
-+ else
-+#endif
-+ eq = try_p_eq_start(p, &irqflags, fq, fd, flags);
-+ if (!eq)
-+ return -EBUSY;
-+ /* Process ORP-specifics here */
-+ if (flags & QMAN_ENQUEUE_FLAG_NLIS)
-+ orp_seqnum |= QM_EQCR_SEQNUM_NLIS;
-+ else {
-+ orp_seqnum &= ~QM_EQCR_SEQNUM_NLIS;
-+ if (flags & QMAN_ENQUEUE_FLAG_NESN)
-+ orp_seqnum |= QM_EQCR_SEQNUM_NESN;
-+ else
-+ /* No need to check 4 QMAN_ENQUEUE_FLAG_HOLE */
-+ orp_seqnum &= ~QM_EQCR_SEQNUM_NESN;
-+ }
-+ eq->seqnum = cpu_to_be16(orp_seqnum);
-+ eq->orp = cpu_to_be32(orp->fqid);
-+ /* Note: QM_EQCR_VERB_INTERRUPT == QMAN_ENQUEUE_FLAG_WAIT_SYNC */
-+ qm_eqcr_pvb_commit(&p->p, QM_EQCR_VERB_ORP |
-+ ((flags & (QMAN_ENQUEUE_FLAG_HOLE | QMAN_ENQUEUE_FLAG_NESN)) ?
-+ 0 : QM_EQCR_VERB_CMD_ENQUEUE) |
-+ (flags & (QM_EQCR_VERB_COLOUR_MASK | QM_EQCR_VERB_INTERRUPT)));
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+#ifdef CONFIG_FSL_DPA_CAN_WAIT_SYNC
-+ if (unlikely((flags & QMAN_ENQUEUE_FLAG_WAIT) &&
-+ (flags & QMAN_ENQUEUE_FLAG_WAIT_SYNC))) {
-+ if (flags & QMAN_ENQUEUE_FLAG_WAIT_INT)
-+ /* NB: return success even if signal occurs before
-+ * condition is true. pvb_commit guarantees success */
-+ wait_event_interruptible(affine_queue,
-+ (p->eqci_owned != fq));
-+ else
-+ wait_event(affine_queue, (p->eqci_owned != fq));
-+ }
-+#endif
-+ return 0;
-+}
-+EXPORT_SYMBOL(qman_p_enqueue_orp);
-+
-+int qman_enqueue_orp(struct qman_fq *fq, const struct qm_fd *fd, u32 flags,
-+ struct qman_fq *orp, u16 orp_seqnum)
-+{
-+ struct qman_portal *p;
-+ struct qm_eqcr_entry *eq;
-+ unsigned long irqflags __maybe_unused;
-+
-+#ifdef CONFIG_FSL_DPA_CAN_WAIT
-+ if (flags & QMAN_ENQUEUE_FLAG_WAIT)
-+ eq = wait_eq_start(&p, &irqflags, fq, fd, flags);
-+ else
-+#endif
-+ eq = try_eq_start(&p, &irqflags, fq, fd, flags);
-+ if (!eq)
-+ return -EBUSY;
-+ /* Process ORP-specifics here */
-+ if (flags & QMAN_ENQUEUE_FLAG_NLIS)
-+ orp_seqnum |= QM_EQCR_SEQNUM_NLIS;
-+ else {
-+ orp_seqnum &= ~QM_EQCR_SEQNUM_NLIS;
-+ if (flags & QMAN_ENQUEUE_FLAG_NESN)
-+ orp_seqnum |= QM_EQCR_SEQNUM_NESN;
-+ else
-+ /* No need to check 4 QMAN_ENQUEUE_FLAG_HOLE */
-+ orp_seqnum &= ~QM_EQCR_SEQNUM_NESN;
-+ }
-+ eq->seqnum = cpu_to_be16(orp_seqnum);
-+ eq->orp = cpu_to_be32(orp->fqid);
-+ /* Note: QM_EQCR_VERB_INTERRUPT == QMAN_ENQUEUE_FLAG_WAIT_SYNC */
-+ qm_eqcr_pvb_commit(&p->p, QM_EQCR_VERB_ORP |
-+ ((flags & (QMAN_ENQUEUE_FLAG_HOLE | QMAN_ENQUEUE_FLAG_NESN)) ?
-+ 0 : QM_EQCR_VERB_CMD_ENQUEUE) |
-+ (flags & (QM_EQCR_VERB_COLOUR_MASK | QM_EQCR_VERB_INTERRUPT)));
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+ put_affine_portal();
-+#ifdef CONFIG_FSL_DPA_CAN_WAIT_SYNC
-+ if (unlikely((flags & QMAN_ENQUEUE_FLAG_WAIT) &&
-+ (flags & QMAN_ENQUEUE_FLAG_WAIT_SYNC))) {
-+ if (flags & QMAN_ENQUEUE_FLAG_WAIT_INT)
-+ /* NB: return success even if signal occurs before
-+ * condition is true. pvb_commit guarantees success */
-+ wait_event_interruptible(affine_queue,
-+ (p->eqci_owned != fq));
-+ else
-+ wait_event(affine_queue, (p->eqci_owned != fq));
-+ }
-+#endif
-+ return 0;
-+}
-+EXPORT_SYMBOL(qman_enqueue_orp);
-+
-+int qman_p_enqueue_precommit(struct qman_portal *p, struct qman_fq *fq,
-+ const struct qm_fd *fd, u32 flags,
-+ qman_cb_precommit cb, void *cb_arg)
-+{
-+ struct qm_eqcr_entry *eq;
-+ unsigned long irqflags __maybe_unused;
-+
-+#ifdef CONFIG_FSL_DPA_CAN_WAIT
-+ if (flags & QMAN_ENQUEUE_FLAG_WAIT)
-+ eq = wait_p_eq_start(p, &irqflags, fq, fd, flags);
-+ else
-+#endif
-+ eq = try_p_eq_start(p, &irqflags, fq, fd, flags);
-+ if (!eq)
-+ return -EBUSY;
-+ /* invoke user supplied callback function before writing commit verb */
-+ if (cb(cb_arg)) {
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+ return -EINVAL;
-+ }
-+ /* Note: QM_EQCR_VERB_INTERRUPT == QMAN_ENQUEUE_FLAG_WAIT_SYNC */
-+ qm_eqcr_pvb_commit(&p->p, QM_EQCR_VERB_CMD_ENQUEUE |
-+ (flags & (QM_EQCR_VERB_COLOUR_MASK | QM_EQCR_VERB_INTERRUPT)));
-+ /* Factor the below out, it's used from qman_enqueue_orp() too */
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+#ifdef CONFIG_FSL_DPA_CAN_WAIT_SYNC
-+ if (unlikely((flags & QMAN_ENQUEUE_FLAG_WAIT) &&
-+ (flags & QMAN_ENQUEUE_FLAG_WAIT_SYNC))) {
-+ if (flags & QMAN_ENQUEUE_FLAG_WAIT_INT)
-+ /* NB: return success even if signal occurs before
-+ * condition is true. pvb_commit guarantees success */
-+ wait_event_interruptible(affine_queue,
-+ (p->eqci_owned != fq));
-+ else
-+ wait_event(affine_queue, (p->eqci_owned != fq));
-+ }
-+#endif
-+ return 0;
-+}
-+EXPORT_SYMBOL(qman_p_enqueue_precommit);
-+
-+int qman_enqueue_precommit(struct qman_fq *fq, const struct qm_fd *fd,
-+ u32 flags, qman_cb_precommit cb, void *cb_arg)
-+{
-+ struct qman_portal *p;
-+ struct qm_eqcr_entry *eq;
-+ unsigned long irqflags __maybe_unused;
-+
-+#ifdef CONFIG_FSL_DPA_CAN_WAIT
-+ if (flags & QMAN_ENQUEUE_FLAG_WAIT)
-+ eq = wait_eq_start(&p, &irqflags, fq, fd, flags);
-+ else
-+#endif
-+ eq = try_eq_start(&p, &irqflags, fq, fd, flags);
-+ if (!eq)
-+ return -EBUSY;
-+ /* invoke user supplied callback function before writing commit verb */
-+ if (cb(cb_arg)) {
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+ put_affine_portal();
-+ return -EINVAL;
-+ }
-+ /* Note: QM_EQCR_VERB_INTERRUPT == QMAN_ENQUEUE_FLAG_WAIT_SYNC */
-+ qm_eqcr_pvb_commit(&p->p, QM_EQCR_VERB_CMD_ENQUEUE |
-+ (flags & (QM_EQCR_VERB_COLOUR_MASK | QM_EQCR_VERB_INTERRUPT)));
-+ /* Factor the below out, it's used from qman_enqueue_orp() too */
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+ put_affine_portal();
-+#ifdef CONFIG_FSL_DPA_CAN_WAIT_SYNC
-+ if (unlikely((flags & QMAN_ENQUEUE_FLAG_WAIT) &&
-+ (flags & QMAN_ENQUEUE_FLAG_WAIT_SYNC))) {
-+ if (flags & QMAN_ENQUEUE_FLAG_WAIT_INT)
-+ /* NB: return success even if signal occurs before
-+ * condition is true. pvb_commit guarantees success */
-+ wait_event_interruptible(affine_queue,
-+ (p->eqci_owned != fq));
-+ else
-+ wait_event(affine_queue, (p->eqci_owned != fq));
-+ }
-+#endif
-+ return 0;
-+}
-+EXPORT_SYMBOL(qman_enqueue_precommit);
-+
-+int qman_modify_cgr(struct qman_cgr *cgr, u32 flags,
-+ struct qm_mcc_initcgr *opts)
-+{
-+ struct qm_mc_command *mcc;
-+ struct qm_mc_result *mcr;
-+ struct qman_portal *p = get_affine_portal();
-+ unsigned long irqflags __maybe_unused;
-+ u8 res;
-+ u8 verb = QM_MCC_VERB_MODIFYCGR;
-+
-+ PORTAL_IRQ_LOCK(p, irqflags);
-+ mcc = qm_mc_start(&p->p);
-+ if (opts)
-+ mcc->initcgr = *opts;
-+ mcc->initcgr.we_mask = cpu_to_be16(mcc->initcgr.we_mask);
-+ mcc->initcgr.cgr.wr_parm_g.word =
-+ cpu_to_be32(mcc->initcgr.cgr.wr_parm_g.word);
-+ mcc->initcgr.cgr.wr_parm_y.word =
-+ cpu_to_be32(mcc->initcgr.cgr.wr_parm_y.word);
-+ mcc->initcgr.cgr.wr_parm_r.word =
-+ cpu_to_be32(mcc->initcgr.cgr.wr_parm_r.word);
-+ mcc->initcgr.cgr.cscn_targ = cpu_to_be32(mcc->initcgr.cgr.cscn_targ);
-+ mcc->initcgr.cgr.__cs_thres = cpu_to_be16(mcc->initcgr.cgr.__cs_thres);
-+
-+ mcc->initcgr.cgid = cgr->cgrid;
-+ if (flags & QMAN_CGR_FLAG_USE_INIT)
-+ verb = QM_MCC_VERB_INITCGR;
-+ qm_mc_commit(&p->p, verb);
-+ while (!(mcr = qm_mc_result(&p->p)))
-+ cpu_relax();
-+ DPA_ASSERT((mcr->verb & QM_MCR_VERB_MASK) == verb);
-+ res = mcr->result;
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+ put_affine_portal();
-+ return (res == QM_MCR_RESULT_OK) ? 0 : -EIO;
-+}
-+EXPORT_SYMBOL(qman_modify_cgr);
-+
-+#define TARG_MASK(n) (0x80000000 >> (n->config->public_cfg.channel - \
-+ QM_CHANNEL_SWPORTAL0))
-+#define TARG_DCP_MASK(n) (0x80000000 >> (10 + n))
-+#define PORTAL_IDX(n) (n->config->public_cfg.channel - QM_CHANNEL_SWPORTAL0)
-+
-+static u8 qman_cgr_cpus[__CGR_NUM];
-+
-+int qman_create_cgr(struct qman_cgr *cgr, u32 flags,
-+ struct qm_mcc_initcgr *opts)
-+{
-+ unsigned long irqflags __maybe_unused;
-+ struct qm_mcr_querycgr cgr_state;
-+ struct qm_mcc_initcgr local_opts;
-+ int ret;
-+ struct qman_portal *p;
-+
-+ /* We have to check that the provided CGRID is within the limits of the
-+ * data-structures, for obvious reasons. However we'll let h/w take
-+ * care of determining whether it's within the limits of what exists on
-+ * the SoC. */
-+ if (cgr->cgrid >= __CGR_NUM)
-+ return -EINVAL;
-+
-+ preempt_disable();
-+ p = get_affine_portal();
-+ qman_cgr_cpus[cgr->cgrid] = smp_processor_id();
-+ preempt_enable();
-+
-+ memset(&local_opts, 0, sizeof(struct qm_mcc_initcgr));
-+ cgr->chan = p->config->public_cfg.channel;
-+ spin_lock_irqsave(&p->cgr_lock, irqflags);
-+
-+ /* if no opts specified, just add it to the list */
-+ if (!opts)
-+ goto add_list;
-+
-+ ret = qman_query_cgr(cgr, &cgr_state);
-+ if (ret)
-+ goto release_lock;
-+ if (opts)
-+ local_opts = *opts;
-+ if ((qman_ip_rev & 0xFF00) >= QMAN_REV30)
-+ local_opts.cgr.cscn_targ_upd_ctrl =
-+ QM_CGR_TARG_UDP_CTRL_WRITE_BIT | PORTAL_IDX(p);
-+ else
-+ /* Overwrite TARG */
-+ local_opts.cgr.cscn_targ = cgr_state.cgr.cscn_targ |
-+ TARG_MASK(p);
-+ local_opts.we_mask |= QM_CGR_WE_CSCN_TARG;
-+
-+ /* send init if flags indicate so */
-+ if (opts && (flags & QMAN_CGR_FLAG_USE_INIT))
-+ ret = qman_modify_cgr(cgr, QMAN_CGR_FLAG_USE_INIT, &local_opts);
-+ else
-+ ret = qman_modify_cgr(cgr, 0, &local_opts);
-+ if (ret)
-+ goto release_lock;
-+add_list:
-+ list_add(&cgr->node, &p->cgr_cbs);
-+
-+ /* Determine if newly added object requires its callback to be called */
-+ ret = qman_query_cgr(cgr, &cgr_state);
-+ if (ret) {
-+ /* we can't go back, so proceed and return success, but screen
-+ * and wail to the log file */
-+ pr_crit("CGR HW state partially modified\n");
-+ ret = 0;
-+ goto release_lock;
-+ }
-+ if (cgr->cb && cgr_state.cgr.cscn_en && qman_cgrs_get(&p->cgrs[1],
-+ cgr->cgrid))
-+ cgr->cb(p, cgr, 1);
-+release_lock:
-+ spin_unlock_irqrestore(&p->cgr_lock, irqflags);
-+ put_affine_portal();
-+ return ret;
-+}
-+EXPORT_SYMBOL(qman_create_cgr);
-+
-+int qman_create_cgr_to_dcp(struct qman_cgr *cgr, u32 flags, u16 dcp_portal,
-+ struct qm_mcc_initcgr *opts)
-+{
-+ unsigned long irqflags __maybe_unused;
-+ struct qm_mcc_initcgr local_opts;
-+ struct qm_mcr_querycgr cgr_state;
-+ int ret;
-+
-+ if ((qman_ip_rev & 0xFF00) < QMAN_REV30) {
-+ pr_warn("This QMan version doesn't support to send CSCN to DCP portal\n");
-+ return -EINVAL;
-+ }
-+ /* We have to check that the provided CGRID is within the limits of the
-+ * data-structures, for obvious reasons. However we'll let h/w take
-+ * care of determining whether it's within the limits of what exists on
-+ * the SoC.
-+ */
-+ if (cgr->cgrid >= __CGR_NUM)
-+ return -EINVAL;
-+
-+ ret = qman_query_cgr(cgr, &cgr_state);
-+ if (ret)
-+ return ret;
-+
-+ memset(&local_opts, 0, sizeof(struct qm_mcc_initcgr));
-+ if (opts)
-+ local_opts = *opts;
-+
-+ if ((qman_ip_rev & 0xFF00) >= QMAN_REV30)
-+ local_opts.cgr.cscn_targ_upd_ctrl =
-+ QM_CGR_TARG_UDP_CTRL_WRITE_BIT |
-+ QM_CGR_TARG_UDP_CTRL_DCP | dcp_portal;
-+ else
-+ local_opts.cgr.cscn_targ = cgr_state.cgr.cscn_targ |
-+ TARG_DCP_MASK(dcp_portal);
-+ local_opts.we_mask |= QM_CGR_WE_CSCN_TARG;
-+
-+ /* send init if flags indicate so */
-+ if (opts && (flags & QMAN_CGR_FLAG_USE_INIT))
-+ ret = qman_modify_cgr(cgr, QMAN_CGR_FLAG_USE_INIT,
-+ &local_opts);
-+ else
-+ ret = qman_modify_cgr(cgr, 0, &local_opts);
-+
-+ return ret;
-+}
-+EXPORT_SYMBOL(qman_create_cgr_to_dcp);
-+
-+int qman_delete_cgr(struct qman_cgr *cgr)
-+{
-+ unsigned long irqflags __maybe_unused;
-+ struct qm_mcr_querycgr cgr_state;
-+ struct qm_mcc_initcgr local_opts;
-+ int ret = 0;
-+ struct qman_cgr *i;
-+ struct qman_portal *p = get_affine_portal();
-+
-+ if (cgr->chan != p->config->public_cfg.channel) {
-+ pr_crit("Attempting to delete cgr from different portal "
-+ "than it was create: create 0x%x, delete 0x%x\n",
-+ cgr->chan, p->config->public_cfg.channel);
-+ ret = -EINVAL;
-+ goto put_portal;
-+ }
-+ memset(&local_opts, 0, sizeof(struct qm_mcc_initcgr));
-+ spin_lock_irqsave(&p->cgr_lock, irqflags);
-+ list_del(&cgr->node);
-+ /*
-+ * If there are no other CGR objects for this CGRID in the list, update
-+ * CSCN_TARG accordingly
-+ */
-+ list_for_each_entry(i, &p->cgr_cbs, node)
-+ if ((i->cgrid == cgr->cgrid) && i->cb)
-+ goto release_lock;
-+ ret = qman_query_cgr(cgr, &cgr_state);
-+ if (ret) {
-+ /* add back to the list */
-+ list_add(&cgr->node, &p->cgr_cbs);
-+ goto release_lock;
-+ }
-+ /* Overwrite TARG */
-+ local_opts.we_mask = QM_CGR_WE_CSCN_TARG;
-+ if ((qman_ip_rev & 0xFF00) >= QMAN_REV30)
-+ local_opts.cgr.cscn_targ_upd_ctrl = PORTAL_IDX(p);
-+ else
-+ local_opts.cgr.cscn_targ = cgr_state.cgr.cscn_targ &
-+ ~(TARG_MASK(p));
-+ ret = qman_modify_cgr(cgr, 0, &local_opts);
-+ if (ret)
-+ /* add back to the list */
-+ list_add(&cgr->node, &p->cgr_cbs);
-+release_lock:
-+ spin_unlock_irqrestore(&p->cgr_lock, irqflags);
-+put_portal:
-+ put_affine_portal();
-+ return ret;
-+}
-+EXPORT_SYMBOL(qman_delete_cgr);
-+
-+struct cgr_comp {
-+ struct qman_cgr *cgr;
-+ struct completion completion;
-+};
-+
-+static void qman_delete_cgr_smp_call(void *p)
-+{
-+ qman_delete_cgr((struct qman_cgr *)p);
-+}
-+
-+void qman_delete_cgr_safe(struct qman_cgr *cgr)
-+{
-+ preempt_disable();
-+ if (qman_cgr_cpus[cgr->cgrid] != smp_processor_id()) {
-+ smp_call_function_single(qman_cgr_cpus[cgr->cgrid],
-+ qman_delete_cgr_smp_call, cgr, true);
-+ preempt_enable();
-+ return;
-+ }
-+ qman_delete_cgr(cgr);
-+ preempt_enable();
-+}
-+EXPORT_SYMBOL(qman_delete_cgr_safe);
-+
-+int qm_get_clock(u64 *clock_hz)
-+{
-+ if (!qman_clk) {
-+ pr_warn("Qman clock speed is unknown\n");
-+ return -EINVAL;
-+ }
-+ *clock_hz = (u64)qman_clk;
-+ return 0;
-+}
-+EXPORT_SYMBOL(qm_get_clock);
-+
-+int qm_set_clock(u64 clock_hz)
-+{
-+ if (qman_clk)
-+ return -1;
-+ qman_clk = (u32)clock_hz;
-+ return 0;
-+}
-+EXPORT_SYMBOL(qm_set_clock);
-+
-+/* CEETM management command */
-+static int qman_ceetm_configure_lfqmt(struct qm_mcc_ceetm_lfqmt_config *opts)
-+{
-+ struct qm_mc_command *mcc;
-+ struct qm_mc_result *mcr;
-+ struct qman_portal *p;
-+ unsigned long irqflags __maybe_unused;
-+ u8 res;
-+
-+ p = get_affine_portal();
-+ PORTAL_IRQ_LOCK(p, irqflags);
-+
-+ mcc = qm_mc_start(&p->p);
-+ mcc->lfqmt_config = *opts;
-+ qm_mc_commit(&p->p, QM_CEETM_VERB_LFQMT_CONFIG);
-+ while (!(mcr = qm_mc_result(&p->p)))
-+ cpu_relax();
-+ DPA_ASSERT((mcr->verb & QM_MCR_VERB_MASK) ==
-+ QM_CEETM_VERB_LFQMT_CONFIG);
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+ put_affine_portal();
-+
-+ res = mcr->result;
-+ if (res != QM_MCR_RESULT_OK) {
-+ pr_err("CEETM: CONFIGURE LFQMT failed\n");
-+ return -EIO;
-+ }
-+ return 0;
-+}
-+
-+int qman_ceetm_query_lfqmt(int lfqid,
-+ struct qm_mcr_ceetm_lfqmt_query *lfqmt_query)
-+{
-+ struct qm_mc_command *mcc;
-+ struct qm_mc_result *mcr;
-+ struct qman_portal *p;
-+ unsigned long irqflags __maybe_unused;
-+ u8 res;
-+
-+ p = get_affine_portal();
-+ PORTAL_IRQ_LOCK(p, irqflags);
-+
-+ mcc = qm_mc_start(&p->p);
-+ mcc->lfqmt_query.lfqid = lfqid;
-+ qm_mc_commit(&p->p, QM_CEETM_VERB_LFQMT_QUERY);
-+ while (!(mcr = qm_mc_result(&p->p)))
-+ cpu_relax();
-+ DPA_ASSERT((mcr->verb & QM_MCR_VERB_MASK) == QM_CEETM_VERB_LFQMT_QUERY);
-+ res = mcr->result;
-+ if (res == QM_MCR_RESULT_OK)
-+ *lfqmt_query = mcr->lfqmt_query;
-+
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+ put_affine_portal();
-+ if (res != QM_MCR_RESULT_OK) {
-+ pr_err("CEETM: QUERY LFQMT failed\n");
-+ return -EIO;
-+ }
-+ return 0;
-+}
-+EXPORT_SYMBOL(qman_ceetm_query_lfqmt);
-+
-+static int qman_ceetm_configure_cq(struct qm_mcc_ceetm_cq_config *opts)
-+{
-+ struct qm_mc_command *mcc;
-+ struct qm_mc_result *mcr;
-+ struct qman_portal *p;
-+ unsigned long irqflags __maybe_unused;
-+ u8 res;
-+
-+ p = get_affine_portal();
-+ PORTAL_IRQ_LOCK(p, irqflags);
-+
-+ mcc = qm_mc_start(&p->p);
-+ mcc->cq_config = *opts;
-+ qm_mc_commit(&p->p, QM_CEETM_VERB_CQ_CONFIG);
-+ while (!(mcr = qm_mc_result(&p->p)))
-+ cpu_relax();
-+ res = mcr->result;
-+ DPA_ASSERT((mcr->verb & QM_MCR_VERB_MASK) == QM_CEETM_VERB_CQ_CONFIG);
-+
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+ put_affine_portal();
-+
-+ if (res != QM_MCR_RESULT_OK) {
-+ pr_err("CEETM: CONFIGURE CQ failed\n");
-+ return -EIO;
-+ }
-+ return 0;
-+}
-+
-+int qman_ceetm_query_cq(unsigned int cqid, unsigned int dcpid,
-+ struct qm_mcr_ceetm_cq_query *cq_query)
-+{
-+ struct qm_mc_command *mcc;
-+ struct qm_mc_result *mcr;
-+ struct qman_portal *p;
-+ unsigned long irqflags __maybe_unused;
-+ u8 res;
-+
-+ p = get_affine_portal();
-+ PORTAL_IRQ_LOCK(p, irqflags);
-+
-+ mcc = qm_mc_start(&p->p);
-+ mcc->cq_query.cqid = cpu_to_be16(cqid);
-+ mcc->cq_query.dcpid = dcpid;
-+ qm_mc_commit(&p->p, QM_CEETM_VERB_CQ_QUERY);
-+ while (!(mcr = qm_mc_result(&p->p)))
-+ cpu_relax();
-+ DPA_ASSERT((mcr->verb & QM_MCR_VERB_MASK) == QM_CEETM_VERB_CQ_QUERY);
-+ res = mcr->result;
-+ if (res == QM_MCR_RESULT_OK) {
-+ *cq_query = mcr->cq_query;
-+ hw_cq_query_to_cpu(cq_query);
-+ }
-+
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+ put_affine_portal();
-+
-+ if (res != QM_MCR_RESULT_OK) {
-+ pr_err("CEETM: QUERY CQ failed\n");
-+ return -EIO;
-+ }
-+
-+ return 0;
-+}
-+EXPORT_SYMBOL(qman_ceetm_query_cq);
-+
-+static int qman_ceetm_configure_dct(struct qm_mcc_ceetm_dct_config *opts)
-+{
-+ struct qm_mc_command *mcc;
-+ struct qm_mc_result *mcr;
-+ struct qman_portal *p;
-+ unsigned long irqflags __maybe_unused;
-+ u8 res;
-+
-+ p = get_affine_portal();
-+ PORTAL_IRQ_LOCK(p, irqflags);
-+
-+ mcc = qm_mc_start(&p->p);
-+ mcc->dct_config = *opts;
-+ qm_mc_commit(&p->p, QM_CEETM_VERB_DCT_CONFIG);
-+ while (!(mcr = qm_mc_result(&p->p)))
-+ cpu_relax();
-+ DPA_ASSERT((mcr->verb & QM_MCR_VERB_MASK) == QM_CEETM_VERB_DCT_CONFIG);
-+ res = mcr->result;
-+
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+ put_affine_portal();
-+
-+ if (res != QM_MCR_RESULT_OK) {
-+ pr_err("CEETM: CONFIGURE DCT failed\n");
-+ return -EIO;
-+ }
-+ return 0;
-+}
-+
-+static int qman_ceetm_query_dct(struct qm_mcc_ceetm_dct_query *opts,
-+ struct qm_mcr_ceetm_dct_query *dct_query)
-+{
-+ struct qm_mc_command *mcc;
-+ struct qm_mc_result *mcr;
-+ struct qman_portal *p = get_affine_portal();
-+ unsigned long irqflags __maybe_unused;
-+ u8 res;
-+
-+ PORTAL_IRQ_LOCK(p, irqflags);
-+
-+ mcc = qm_mc_start(&p->p);
-+ mcc->dct_query = *opts;
-+ qm_mc_commit(&p->p, QM_CEETM_VERB_DCT_QUERY);
-+ while (!(mcr = qm_mc_result(&p->p)))
-+ cpu_relax();
-+ DPA_ASSERT((mcr->verb & QM_MCR_VERB_MASK) == QM_CEETM_VERB_DCT_QUERY);
-+ res = mcr->result;
-+
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+ put_affine_portal();
-+
-+ if (res != QM_MCR_RESULT_OK) {
-+ pr_err("CEETM: QUERY DCT failed\n");
-+ return -EIO;
-+ }
-+
-+ *dct_query = mcr->dct_query;
-+ return 0;
-+}
-+
-+static int qman_ceetm_configure_class_scheduler(
-+ struct qm_mcc_ceetm_class_scheduler_config *opts)
-+{
-+ struct qm_mc_command *mcc;
-+ struct qm_mc_result *mcr;
-+ struct qman_portal *p;
-+ unsigned long irqflags __maybe_unused;
-+ u8 res;
-+
-+ p = get_affine_portal();
-+ PORTAL_IRQ_LOCK(p, irqflags);
-+
-+ mcc = qm_mc_start(&p->p);
-+ mcc->csch_config = *opts;
-+ qm_mc_commit(&p->p, QM_CEETM_VERB_CLASS_SCHEDULER_CONFIG);
-+ while (!(mcr = qm_mc_result(&p->p)))
-+ cpu_relax();
-+ DPA_ASSERT((mcr->verb & QM_MCR_VERB_MASK) ==
-+ QM_CEETM_VERB_CLASS_SCHEDULER_CONFIG);
-+ res = mcr->result;
-+
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+ put_affine_portal();
-+
-+ if (res != QM_MCR_RESULT_OK) {
-+ pr_err("CEETM: CONFIGURE CLASS SCHEDULER failed\n");
-+ return -EIO;
-+ }
-+ return 0;
-+}
-+
-+static int qman_ceetm_query_class_scheduler(struct qm_ceetm_channel *channel,
-+ struct qm_mcr_ceetm_class_scheduler_query *query)
-+{
-+ struct qm_mc_command *mcc;
-+ struct qm_mc_result *mcr;
-+ struct qman_portal *p;
-+ unsigned long irqflags __maybe_unused;
-+ u8 res;
-+
-+ p = get_affine_portal();
-+ PORTAL_IRQ_LOCK(p, irqflags);
-+
-+ mcc = qm_mc_start(&p->p);
-+ mcc->csch_query.cqcid = cpu_to_be16(channel->idx);
-+ mcc->csch_query.dcpid = channel->dcp_idx;
-+ qm_mc_commit(&p->p, QM_CEETM_VERB_CLASS_SCHEDULER_QUERY);
-+ while (!(mcr = qm_mc_result(&p->p)))
-+ cpu_relax();
-+ DPA_ASSERT((mcr->verb & QM_MCR_VERB_MASK) ==
-+ QM_CEETM_VERB_CLASS_SCHEDULER_QUERY);
-+ res = mcr->result;
-+
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+ put_affine_portal();
-+
-+ if (res != QM_MCR_RESULT_OK) {
-+ pr_err("CEETM: QUERY CLASS SCHEDULER failed\n");
-+ return -EIO;
-+ }
-+ *query = mcr->csch_query;
-+ return 0;
-+}
-+
-+static int qman_ceetm_configure_mapping_shaper_tcfc(
-+ struct qm_mcc_ceetm_mapping_shaper_tcfc_config *opts)
-+{
-+ struct qm_mc_command *mcc;
-+ struct qm_mc_result *mcr;
-+ struct qman_portal *p;
-+ unsigned long irqflags __maybe_unused;
-+ u8 res;
-+
-+ p = get_affine_portal();
-+ PORTAL_IRQ_LOCK(p, irqflags);
-+
-+ mcc = qm_mc_start(&p->p);
-+ mcc->mst_config = *opts;
-+ qm_mc_commit(&p->p, QM_CEETM_VERB_MAPPING_SHAPER_TCFC_CONFIG);
-+ while (!(mcr = qm_mc_result(&p->p)))
-+ cpu_relax();
-+ DPA_ASSERT((mcr->verb & QM_MCR_VERB_MASK) ==
-+ QM_CEETM_VERB_MAPPING_SHAPER_TCFC_CONFIG);
-+ res = mcr->result;
-+
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+ put_affine_portal();
-+
-+ if (res != QM_MCR_RESULT_OK) {
-+ pr_err("CEETM: CONFIGURE CHANNEL MAPPING failed\n");
-+ return -EIO;
-+ }
-+ return 0;
-+}
-+
-+static int qman_ceetm_query_mapping_shaper_tcfc(
-+ struct qm_mcc_ceetm_mapping_shaper_tcfc_query *opts,
-+ struct qm_mcr_ceetm_mapping_shaper_tcfc_query *response)
-+{
-+ struct qm_mc_command *mcc;
-+ struct qm_mc_result *mcr;
-+ struct qman_portal *p;
-+ unsigned long irqflags __maybe_unused;
-+ u8 res;
-+
-+ p = get_affine_portal();
-+ PORTAL_IRQ_LOCK(p, irqflags);
-+
-+ mcc = qm_mc_start(&p->p);
-+ mcc->mst_query = *opts;
-+ qm_mc_commit(&p->p, QM_CEETM_VERB_MAPPING_SHAPER_TCFC_QUERY);
-+ while (!(mcr = qm_mc_result(&p->p)))
-+ cpu_relax();
-+ DPA_ASSERT((mcr->verb & QM_MCR_VERB_MASK) ==
-+ QM_CEETM_VERB_MAPPING_SHAPER_TCFC_QUERY);
-+ res = mcr->result;
-+
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+ put_affine_portal();
-+
-+ if (res != QM_MCR_RESULT_OK) {
-+ pr_err("CEETM: QUERY CHANNEL MAPPING failed\n");
-+ return -EIO;
-+ }
-+
-+ *response = mcr->mst_query;
-+ return 0;
-+}
-+
-+static int qman_ceetm_configure_ccgr(struct qm_mcc_ceetm_ccgr_config *opts)
-+{
-+ struct qm_mc_command *mcc;
-+ struct qm_mc_result *mcr;
-+ struct qman_portal *p;
-+ unsigned long irqflags __maybe_unused;
-+ u8 res;
-+
-+ p = get_affine_portal();
-+ PORTAL_IRQ_LOCK(p, irqflags);
-+
-+ mcc = qm_mc_start(&p->p);
-+ mcc->ccgr_config = *opts;
-+
-+ qm_mc_commit(&p->p, QM_CEETM_VERB_CCGR_CONFIG);
-+ while (!(mcr = qm_mc_result(&p->p)))
-+ cpu_relax();
-+ DPA_ASSERT((mcr->verb & QM_MCR_VERB_MASK) == QM_CEETM_VERB_CCGR_CONFIG);
-+
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+ put_affine_portal();
-+
-+ res = mcr->result;
-+ if (res != QM_MCR_RESULT_OK) {
-+ pr_err("CEETM: CONFIGURE CCGR failed\n");
-+ return -EIO;
-+ }
-+ return 0;
-+}
-+
-+int qman_ceetm_query_ccgr(struct qm_mcc_ceetm_ccgr_query *ccgr_query,
-+ struct qm_mcr_ceetm_ccgr_query *response)
-+{
-+ struct qm_mc_command *mcc;
-+ struct qm_mc_result *mcr;
-+ struct qman_portal *p;
-+ unsigned long irqflags __maybe_unused;
-+ u8 res;
-+
-+ p = get_affine_portal();
-+ PORTAL_IRQ_LOCK(p, irqflags);
-+
-+ mcc = qm_mc_start(&p->p);
-+ mcc->ccgr_query.ccgrid = cpu_to_be16(ccgr_query->ccgrid);
-+ mcc->ccgr_query.dcpid = ccgr_query->dcpid;
-+ qm_mc_commit(&p->p, QM_CEETM_VERB_CCGR_QUERY);
-+
-+ while (!(mcr = qm_mc_result(&p->p)))
-+ cpu_relax();
-+ DPA_ASSERT((mcr->verb & QM_MCR_VERB_MASK) == QM_CEETM_VERB_CCGR_QUERY);
-+ res = mcr->result;
-+ if (res == QM_MCR_RESULT_OK) {
-+ *response = mcr->ccgr_query;
-+ hw_ccgr_query_to_cpu(response);
-+ }
-+
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+ put_affine_portal();
-+ if (res != QM_MCR_RESULT_OK) {
-+ pr_err("CEETM: QUERY CCGR failed\n");
-+ return -EIO;
-+ }
-+ return 0;
-+}
-+EXPORT_SYMBOL(qman_ceetm_query_ccgr);
-+
-+static int qman_ceetm_cq_peek_pop_xsfdrread(struct qm_ceetm_cq *cq,
-+ u8 command_type, u16 xsfdr,
-+ struct qm_mcr_ceetm_cq_peek_pop_xsfdrread *cq_ppxr)
-+{
-+ struct qm_mc_command *mcc;
-+ struct qm_mc_result *mcr;
-+ struct qman_portal *p;
-+ unsigned long irqflags __maybe_unused;
-+ u8 res;
-+
-+ p = get_affine_portal();
-+ PORTAL_IRQ_LOCK(p, irqflags);
-+
-+ mcc = qm_mc_start(&p->p);
-+ switch (command_type) {
-+ case 0:
-+ case 1:
-+ mcc->cq_ppxr.cqid = (cq->parent->idx << 4) | cq->idx;
-+ break;
-+ case 2:
-+ mcc->cq_ppxr.xsfdr = xsfdr;
-+ break;
-+ default:
-+ break;
-+ }
-+ mcc->cq_ppxr.ct = command_type;
-+ mcc->cq_ppxr.dcpid = cq->parent->dcp_idx;
-+ qm_mc_commit(&p->p, QM_CEETM_VERB_CQ_PEEK_POP_XFDRREAD);
-+ while (!(mcr = qm_mc_result(&p->p)))
-+ cpu_relax();
-+ DPA_ASSERT((mcr->verb & QM_MCR_VERB_MASK) ==
-+ QM_CEETM_VERB_CQ_PEEK_POP_XFDRREAD);
-+
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+ put_affine_portal();
-+
-+ res = mcr->result;
-+ if (res != QM_MCR_RESULT_OK) {
-+ pr_err("CEETM: CQ PEEK/POP/XSFDR READ failed\n");
-+ return -EIO;
-+ }
-+ *cq_ppxr = mcr->cq_ppxr;
-+ return 0;
-+}
-+
-+static int qman_ceetm_query_statistics(u16 cid,
-+ enum qm_dc_portal dcp_idx,
-+ u16 command_type,
-+ struct qm_mcr_ceetm_statistics_query *query_result)
-+{
-+ struct qm_mc_command *mcc;
-+ struct qm_mc_result *mcr;
-+ struct qman_portal *p;
-+ unsigned long irqflags __maybe_unused;
-+ u8 res;
-+
-+ p = get_affine_portal();
-+ PORTAL_IRQ_LOCK(p, irqflags);
-+
-+ mcc = qm_mc_start(&p->p);
-+ mcc->stats_query_write.cid = cid;
-+ mcc->stats_query_write.dcpid = dcp_idx;
-+ mcc->stats_query_write.ct = command_type;
-+ qm_mc_commit(&p->p, QM_CEETM_VERB_STATISTICS_QUERY_WRITE);
-+
-+ while (!(mcr = qm_mc_result(&p->p)))
-+ cpu_relax();
-+ DPA_ASSERT((mcr->verb & QM_MCR_VERB_MASK) ==
-+ QM_CEETM_VERB_STATISTICS_QUERY_WRITE);
-+
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+ put_affine_portal();
-+
-+ res = mcr->result;
-+ if (res != QM_MCR_RESULT_OK) {
-+ pr_err("CEETM: STATISTICS QUERY failed\n");
-+ return -EIO;
-+ }
-+ *query_result = mcr->stats_query;
-+ return 0;
-+}
-+
-+int qman_ceetm_query_write_statistics(u16 cid, enum qm_dc_portal dcp_idx,
-+ u16 command_type, u64 frame_count,
-+ u64 byte_count)
-+{
-+ struct qm_mc_command *mcc;
-+ struct qm_mc_result *mcr;
-+ struct qman_portal *p;
-+ unsigned long irqflags __maybe_unused;
-+ u8 res;
-+
-+ p = get_affine_portal();
-+ PORTAL_IRQ_LOCK(p, irqflags);
-+
-+ mcc = qm_mc_start(&p->p);
-+ mcc->stats_query_write.cid = cid;
-+ mcc->stats_query_write.dcpid = dcp_idx;
-+ mcc->stats_query_write.ct = command_type;
-+ mcc->stats_query_write.frm_cnt = frame_count;
-+ mcc->stats_query_write.byte_cnt = byte_count;
-+ qm_mc_commit(&p->p, QM_CEETM_VERB_STATISTICS_QUERY_WRITE);
-+
-+ while (!(mcr = qm_mc_result(&p->p)))
-+ cpu_relax();
-+ DPA_ASSERT((mcr->verb & QM_MCR_VERB_MASK) ==
-+ QM_CEETM_VERB_STATISTICS_QUERY_WRITE);
-+
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+ put_affine_portal();
-+
-+ res = mcr->result;
-+ if (res != QM_MCR_RESULT_OK) {
-+ pr_err("CEETM: STATISTICS WRITE failed\n");
-+ return -EIO;
-+ }
-+ return 0;
-+}
-+EXPORT_SYMBOL(qman_ceetm_query_write_statistics);
-+
-+int qman_ceetm_bps2tokenrate(u64 bps, struct qm_ceetm_rate *token_rate,
-+ int rounding)
-+{
-+ u16 pres;
-+ u64 temp;
-+ u64 qman_freq;
-+ int ret;
-+
-+ /* Read PRES from CEET_CFG_PRES register */
-+ ret = qman_ceetm_get_prescaler(&pres);
-+ if (ret)
-+ return -EINVAL;
-+
-+ ret = qm_get_clock(&qman_freq);
-+ if (ret)
-+ return -EINVAL;
-+
-+ /* token-rate = bytes-per-second * update-reference-period
-+ *
-+ * Where token-rate is N/8192 for a integer N, and
-+ * update-reference-period is (2^22)/(PRES*QHz), where PRES
-+ * is the prescalar value and QHz is the QMan clock frequency.
-+ * So:
-+ *
-+ * token-rate = (byte-per-second*2^22)/PRES*QHZ)
-+ *
-+ * Converting to bits-per-second gives;
-+ *
-+ * token-rate = (bps*2^19) / (PRES*QHZ)
-+ * N = (bps*2^32) / (PRES*QHz)
-+ *
-+ * And to avoid 64-bit overflow if 'bps' is larger than 4Gbps
-+ * (yet minimise rounding error if 'bps' is small), we reorganise
-+ * the formula to use two 16-bit shifts rather than 1 32-bit shift.
-+ * N = (((bps*2^16)/PRES)*2^16)/QHz
-+ */
-+ temp = ROUNDING((bps << 16), pres, rounding);
-+ temp = ROUNDING((temp << 16), qman_freq, rounding);
-+ token_rate->whole = temp >> 13;
-+ token_rate->fraction = temp & (((u64)1 << 13) - 1);
-+ return 0;
-+}
-+EXPORT_SYMBOL(qman_ceetm_bps2tokenrate);
-+
-+int qman_ceetm_tokenrate2bps(const struct qm_ceetm_rate *token_rate, u64 *bps,
-+ int rounding)
-+{
-+ u16 pres;
-+ u64 temp;
-+ u64 qman_freq;
-+ int ret;
-+
-+ /* Read PRES from CEET_CFG_PRES register */
-+ ret = qman_ceetm_get_prescaler(&pres);
-+ if (ret)
-+ return -EINVAL;
-+
-+ ret = qm_get_clock(&qman_freq);
-+ if (ret)
-+ return -EINVAL;
-+
-+ /* bytes-per-second = token-rate / update-reference-period
-+ *
-+ * where "token-rate" is N/8192 for an integer N, and
-+ * "update-reference-period" is (2^22)/(PRES*QHz), where PRES is
-+ * the prescalar value and QHz is the QMan clock frequency. So;
-+ *
-+ * bytes-per-second = (N/8192) / (4194304/PRES*QHz)
-+ * = N*PRES*QHz / (4194304*8192)
-+ * = N*PRES*QHz / (2^35)
-+ *
-+ * Converting to bits-per-second gives;
-+ *
-+ * bps = N*PRES*QHZ / (2^32)
-+ *
-+ * Note, the numerator has a maximum width of 72 bits! So to
-+ * avoid 64-bit overflow errors, we calculate PRES*QHZ (maximum
-+ * width 48 bits) divided by 2^9 (reducing to maximum 39 bits), before
-+ * multiplying by N (goes to maximum of 63 bits).
-+ *
-+ * temp = PRES*QHZ / (2^16)
-+ * kbps = temp*N / (2^16)
-+ */
-+ temp = ROUNDING(qman_freq * pres, (u64)1 << 16 , rounding);
-+ temp *= ((token_rate->whole << 13) + token_rate->fraction);
-+ *bps = ROUNDING(temp, (u64)(1) << 16, rounding);
-+ return 0;
-+}
-+EXPORT_SYMBOL(qman_ceetm_tokenrate2bps);
-+
-+int qman_ceetm_sp_claim(struct qm_ceetm_sp **sp, enum qm_dc_portal dcp_idx,
-+ unsigned int sp_idx)
-+{
-+ struct qm_ceetm_sp *p;
-+
-+ DPA_ASSERT((dcp_idx == qm_dc_portal_fman0) ||
-+ (dcp_idx == qm_dc_portal_fman1));
-+
-+ if ((sp_idx < qman_ceetms[dcp_idx].sp_range[0]) ||
-+ (sp_idx >= (qman_ceetms[dcp_idx].sp_range[0] +
-+ qman_ceetms[dcp_idx].sp_range[1]))) {
-+ pr_err("Sub-portal index doesn't exist\n");
-+ return -EINVAL;
-+ }
-+
-+ list_for_each_entry(p, &qman_ceetms[dcp_idx].sub_portals, node) {
-+ if ((p->idx == sp_idx) && (p->is_claimed == 0)) {
-+ p->is_claimed = 1;
-+ *sp = p;
-+ return 0;
-+ }
-+ }
-+ pr_err("The sub-portal#%d is not available!\n", sp_idx);
-+ return -ENODEV;
-+}
-+EXPORT_SYMBOL(qman_ceetm_sp_claim);
-+
-+int qman_ceetm_sp_release(struct qm_ceetm_sp *sp)
-+{
-+ struct qm_ceetm_sp *p;
-+
-+ if (sp->lni && sp->lni->is_claimed == 1) {
-+ pr_err("The dependency of sub-portal has not been released!\n");
-+ return -EBUSY;
-+ }
-+
-+ list_for_each_entry(p, &qman_ceetms[sp->dcp_idx].sub_portals, node) {
-+ if (p->idx == sp->idx) {
-+ p->is_claimed = 0;
-+ p->lni = NULL;
-+ }
-+ }
-+ /* Disable CEETM mode of this sub-portal */
-+ qman_sp_disable_ceetm_mode(sp->dcp_idx, sp->idx);
-+
-+ return 0;
-+}
-+EXPORT_SYMBOL(qman_ceetm_sp_release);
-+
-+int qman_ceetm_lni_claim(struct qm_ceetm_lni **lni, enum qm_dc_portal dcp_idx,
-+ unsigned int lni_idx)
-+{
-+ struct qm_ceetm_lni *p;
-+
-+ if ((lni_idx < qman_ceetms[dcp_idx].lni_range[0]) ||
-+ (lni_idx >= (qman_ceetms[dcp_idx].lni_range[0] +
-+ qman_ceetms[dcp_idx].lni_range[1]))) {
-+ pr_err("The lni index is out of range\n");
-+ return -EINVAL;
-+ }
-+
-+ list_for_each_entry(p, &qman_ceetms[dcp_idx].lnis, node) {
-+ if ((p->idx == lni_idx) && (p->is_claimed == 0)) {
-+ *lni = p;
-+ p->is_claimed = 1;
-+ return 0;
-+ }
-+ }
-+
-+ pr_err("The LNI#%d is not available!\n", lni_idx);
-+ return -EINVAL;
-+}
-+EXPORT_SYMBOL(qman_ceetm_lni_claim);
-+
-+int qman_ceetm_lni_release(struct qm_ceetm_lni *lni)
-+{
-+ struct qm_ceetm_lni *p;
-+ struct qm_mcc_ceetm_mapping_shaper_tcfc_config config_opts;
-+
-+ if (!list_empty(&lni->channels)) {
-+ pr_err("The LNI dependencies are not released!\n");
-+ return -EBUSY;
-+ }
-+
-+ list_for_each_entry(p, &qman_ceetms[lni->dcp_idx].lnis, node) {
-+ if (p->idx == lni->idx) {
-+ p->shaper_enable = 0;
-+ p->shaper_couple = 0;
-+ p->cr_token_rate.whole = 0;
-+ p->cr_token_rate.fraction = 0;
-+ p->er_token_rate.whole = 0;
-+ p->er_token_rate.fraction = 0;
-+ p->cr_token_bucket_limit = 0;
-+ p->er_token_bucket_limit = 0;
-+ p->is_claimed = 0;
-+ }
-+ }
-+ config_opts.cid = cpu_to_be16(CEETM_COMMAND_LNI_SHAPER | lni->idx);
-+ config_opts.dcpid = lni->dcp_idx;
-+ memset(&config_opts.shaper_config, 0,
-+ sizeof(config_opts.shaper_config));
-+ return qman_ceetm_configure_mapping_shaper_tcfc(&config_opts);
-+}
-+EXPORT_SYMBOL(qman_ceetm_lni_release);
-+
-+int qman_ceetm_sp_set_lni(struct qm_ceetm_sp *sp, struct qm_ceetm_lni *lni)
-+{
-+ struct qm_mcc_ceetm_mapping_shaper_tcfc_config config_opts;
-+
-+ config_opts.cid = cpu_to_be16(CEETM_COMMAND_SP_MAPPING | sp->idx);
-+ config_opts.dcpid = sp->dcp_idx;
-+ config_opts.sp_mapping.map_lni_id = lni->idx;
-+ sp->lni = lni;
-+
-+ if (qman_ceetm_configure_mapping_shaper_tcfc(&config_opts))
-+ return -EINVAL;
-+
-+ /* Enable CEETM mode for this sub-portal */
-+ return qman_sp_enable_ceetm_mode(sp->dcp_idx, sp->idx);
-+}
-+EXPORT_SYMBOL(qman_ceetm_sp_set_lni);
-+
-+int qman_ceetm_sp_get_lni(struct qm_ceetm_sp *sp, unsigned int *lni_idx)
-+{
-+ struct qm_mcc_ceetm_mapping_shaper_tcfc_query query_opts;
-+ struct qm_mcr_ceetm_mapping_shaper_tcfc_query query_result;
-+
-+ query_opts.cid = cpu_to_be16(CEETM_COMMAND_SP_MAPPING | sp->idx);
-+ query_opts.dcpid = sp->dcp_idx;
-+ if (qman_ceetm_query_mapping_shaper_tcfc(&query_opts, &query_result)) {
-+ pr_err("Can't get SP <-> LNI mapping\n");
-+ return -EINVAL;
-+ }
-+ *lni_idx = query_result.sp_mapping_query.map_lni_id;
-+ sp->lni->idx = query_result.sp_mapping_query.map_lni_id;
-+ return 0;
-+}
-+EXPORT_SYMBOL(qman_ceetm_sp_get_lni);
-+
-+int qman_ceetm_lni_enable_shaper(struct qm_ceetm_lni *lni, int coupled,
-+ int oal)
-+{
-+ struct qm_mcc_ceetm_mapping_shaper_tcfc_config config_opts;
-+
-+ if (lni->shaper_enable) {
-+ pr_err("The shaper has already been enabled\n");
-+ return -EINVAL;
-+ }
-+ lni->shaper_enable = 1;
-+ lni->shaper_couple = coupled;
-+ lni->oal = oal;
-+
-+ config_opts.cid = cpu_to_be16(CEETM_COMMAND_LNI_SHAPER | lni->idx);
-+ config_opts.dcpid = lni->dcp_idx;
-+ config_opts.shaper_config.cpl = coupled;
-+ config_opts.shaper_config.oal = oal;
-+ config_opts.shaper_config.crtcr = cpu_to_be24((lni->cr_token_rate.whole
-+ << 13) | lni->cr_token_rate.fraction);
-+ config_opts.shaper_config.ertcr = cpu_to_be24((lni->er_token_rate.whole
-+ << 13) | lni->er_token_rate.fraction);
-+ config_opts.shaper_config.crtbl =
-+ cpu_to_be16(lni->cr_token_bucket_limit);
-+ config_opts.shaper_config.ertbl =
-+ cpu_to_be16(lni->er_token_bucket_limit);
-+ config_opts.shaper_config.mps = 60;
-+
-+ return qman_ceetm_configure_mapping_shaper_tcfc(&config_opts);
-+}
-+EXPORT_SYMBOL(qman_ceetm_lni_enable_shaper);
-+
-+int qman_ceetm_lni_disable_shaper(struct qm_ceetm_lni *lni)
-+{
-+ struct qm_mcc_ceetm_mapping_shaper_tcfc_config config_opts;
-+
-+ if (!lni->shaper_enable) {
-+ pr_err("The shaper has been disabled\n");
-+ return -EINVAL;
-+ }
-+
-+ config_opts.cid = cpu_to_be16(CEETM_COMMAND_LNI_SHAPER | lni->idx);
-+ config_opts.dcpid = lni->dcp_idx;
-+ config_opts.shaper_config.cpl = lni->shaper_couple;
-+ config_opts.shaper_config.oal = lni->oal;
-+ config_opts.shaper_config.crtbl =
-+ cpu_to_be16(lni->cr_token_bucket_limit);
-+ config_opts.shaper_config.ertbl =
-+ cpu_to_be16(lni->er_token_bucket_limit);
-+ /* Set CR/ER rate with all 1's to configure an infinite rate, thus
-+ * disable the shaping.
-+ */
-+ config_opts.shaper_config.crtcr = 0xFFFFFF;
-+ config_opts.shaper_config.ertcr = 0xFFFFFF;
-+ config_opts.shaper_config.mps = 60;
-+ lni->shaper_enable = 0;
-+ return qman_ceetm_configure_mapping_shaper_tcfc(&config_opts);
-+}
-+EXPORT_SYMBOL(qman_ceetm_lni_disable_shaper);
-+
-+int qman_ceetm_lni_is_shaper_enabled(struct qm_ceetm_lni *lni)
-+{
-+ return lni->shaper_enable;
-+}
-+EXPORT_SYMBOL(qman_ceetm_lni_is_shaper_enabled);
-+
-+int qman_ceetm_lni_set_commit_rate(struct qm_ceetm_lni *lni,
-+ const struct qm_ceetm_rate *token_rate,
-+ u16 token_limit)
-+{
-+ struct qm_mcc_ceetm_mapping_shaper_tcfc_config config_opts;
-+ struct qm_mcc_ceetm_mapping_shaper_tcfc_query query_opts;
-+ struct qm_mcr_ceetm_mapping_shaper_tcfc_query query_result;
-+ int ret;
-+
-+ lni->cr_token_rate.whole = token_rate->whole;
-+ lni->cr_token_rate.fraction = token_rate->fraction;
-+ lni->cr_token_bucket_limit = token_limit;
-+ if (!lni->shaper_enable)
-+ return 0;
-+ query_opts.cid = cpu_to_be16(CEETM_COMMAND_LNI_SHAPER | lni->idx);
-+ query_opts.dcpid = lni->dcp_idx;
-+ ret = qman_ceetm_query_mapping_shaper_tcfc(&query_opts,
-+ &query_result);
-+ if (ret) {
-+ pr_err("Fail to get current LNI shaper setting\n");
-+ return -EINVAL;
-+ }
-+
-+ config_opts.cid = cpu_to_be16(CEETM_COMMAND_LNI_SHAPER | lni->idx);
-+ config_opts.dcpid = lni->dcp_idx;
-+ config_opts.shaper_config.crtcr = cpu_to_be24((token_rate->whole << 13)
-+ | (token_rate->fraction));
-+ config_opts.shaper_config.crtbl = cpu_to_be16(token_limit);
-+ config_opts.shaper_config.cpl = query_result.shaper_query.cpl;
-+ config_opts.shaper_config.oal = query_result.shaper_query.oal;
-+ config_opts.shaper_config.ertcr = query_result.shaper_query.ertcr;
-+ config_opts.shaper_config.ertbl = query_result.shaper_query.ertbl;
-+ config_opts.shaper_config.mps = query_result.shaper_query.mps;
-+ return qman_ceetm_configure_mapping_shaper_tcfc(&config_opts);
-+}
-+EXPORT_SYMBOL(qman_ceetm_lni_set_commit_rate);
-+
-+int qman_ceetm_lni_set_commit_rate_bps(struct qm_ceetm_lni *lni,
-+ u64 bps,
-+ u16 token_limit)
-+{
-+ struct qm_ceetm_rate token_rate;
-+ int ret;
-+
-+ ret = qman_ceetm_bps2tokenrate(bps, &token_rate, 0);
-+ if (ret) {
-+ pr_err("Can not convert bps to token rate\n");
-+ return -EINVAL;
-+ }
-+
-+ return qman_ceetm_lni_set_commit_rate(lni, &token_rate, token_limit);
-+}
-+EXPORT_SYMBOL(qman_ceetm_lni_set_commit_rate_bps);
-+
-+int qman_ceetm_lni_get_commit_rate(struct qm_ceetm_lni *lni,
-+ struct qm_ceetm_rate *token_rate,
-+ u16 *token_limit)
-+{
-+ struct qm_mcc_ceetm_mapping_shaper_tcfc_query query_opts;
-+ struct qm_mcr_ceetm_mapping_shaper_tcfc_query query_result;
-+ int ret;
-+
-+ query_opts.cid = cpu_to_be16(CEETM_COMMAND_LNI_SHAPER | lni->idx);
-+ query_opts.dcpid = lni->dcp_idx;
-+
-+ ret = qman_ceetm_query_mapping_shaper_tcfc(&query_opts, &query_result);
-+ if (ret) {
-+ pr_err("The LNI CR rate or limit is not set\n");
-+ return -EINVAL;
-+ }
-+ token_rate->whole = be24_to_cpu(query_result.shaper_query.crtcr) >> 13;
-+ token_rate->fraction = be24_to_cpu(query_result.shaper_query.crtcr) &
-+ 0x1FFF;
-+ *token_limit = be16_to_cpu(query_result.shaper_query.crtbl);
-+ return 0;
-+}
-+EXPORT_SYMBOL(qman_ceetm_lni_get_commit_rate);
-+
-+int qman_ceetm_lni_get_commit_rate_bps(struct qm_ceetm_lni *lni,
-+ u64 *bps, u16 *token_limit)
-+{
-+ struct qm_ceetm_rate token_rate;
-+ int ret;
-+
-+ ret = qman_ceetm_lni_get_commit_rate(lni, &token_rate, token_limit);
-+ if (ret) {
-+ pr_err("The LNI CR rate or limit is not available\n");
-+ return -EINVAL;
-+ }
-+
-+ return qman_ceetm_tokenrate2bps(&token_rate, bps, 0);
-+}
-+EXPORT_SYMBOL(qman_ceetm_lni_get_commit_rate_bps);
-+
-+int qman_ceetm_lni_set_excess_rate(struct qm_ceetm_lni *lni,
-+ const struct qm_ceetm_rate *token_rate,
-+ u16 token_limit)
-+{
-+ struct qm_mcc_ceetm_mapping_shaper_tcfc_config config_opts;
-+ struct qm_mcc_ceetm_mapping_shaper_tcfc_query query_opts;
-+ struct qm_mcr_ceetm_mapping_shaper_tcfc_query query_result;
-+ int ret;
-+
-+ lni->er_token_rate.whole = token_rate->whole;
-+ lni->er_token_rate.fraction = token_rate->fraction;
-+ lni->er_token_bucket_limit = token_limit;
-+ if (!lni->shaper_enable)
-+ return 0;
-+
-+ query_opts.cid = cpu_to_be16(CEETM_COMMAND_LNI_SHAPER | lni->idx);
-+ query_opts.dcpid = lni->dcp_idx;
-+ ret = qman_ceetm_query_mapping_shaper_tcfc(&query_opts,
-+ &query_result);
-+ if (ret) {
-+ pr_err("Fail to get current LNI shaper setting\n");
-+ return -EINVAL;
-+ }
-+
-+ config_opts.cid = cpu_to_be16(CEETM_COMMAND_LNI_SHAPER | lni->idx);
-+ config_opts.dcpid = lni->dcp_idx;
-+ config_opts.shaper_config.ertcr = cpu_to_be24(
-+ (token_rate->whole << 13) | (token_rate->fraction));
-+ config_opts.shaper_config.ertbl = cpu_to_be16(token_limit);
-+ config_opts.shaper_config.cpl = query_result.shaper_query.cpl;
-+ config_opts.shaper_config.oal = query_result.shaper_query.oal;
-+ config_opts.shaper_config.crtcr = query_result.shaper_query.crtcr;
-+ config_opts.shaper_config.crtbl = query_result.shaper_query.crtbl;
-+ config_opts.shaper_config.mps = query_result.shaper_query.mps;
-+ return qman_ceetm_configure_mapping_shaper_tcfc(&config_opts);
-+}
-+EXPORT_SYMBOL(qman_ceetm_lni_set_excess_rate);
-+
-+int qman_ceetm_lni_set_excess_rate_bps(struct qm_ceetm_lni *lni,
-+ u64 bps,
-+ u16 token_limit)
-+{
-+ struct qm_ceetm_rate token_rate;
-+ int ret;
-+
-+ ret = qman_ceetm_bps2tokenrate(bps, &token_rate, 0);
-+ if (ret) {
-+ pr_err("Can not convert bps to token rate\n");
-+ return -EINVAL;
-+ }
-+ return qman_ceetm_lni_set_excess_rate(lni, &token_rate, token_limit);
-+}
-+EXPORT_SYMBOL(qman_ceetm_lni_set_excess_rate_bps);
-+
-+int qman_ceetm_lni_get_excess_rate(struct qm_ceetm_lni *lni,
-+ struct qm_ceetm_rate *token_rate,
-+ u16 *token_limit)
-+{
-+ struct qm_mcc_ceetm_mapping_shaper_tcfc_query query_opts;
-+ struct qm_mcr_ceetm_mapping_shaper_tcfc_query query_result;
-+ int ret;
-+
-+ query_opts.cid = cpu_to_be16(CEETM_COMMAND_LNI_SHAPER | lni->idx);
-+ query_opts.dcpid = lni->dcp_idx;
-+ ret = qman_ceetm_query_mapping_shaper_tcfc(&query_opts, &query_result);
-+ if (ret) {
-+ pr_err("The LNI ER rate or limit is not set\n");
-+ return -EINVAL;
-+ }
-+ token_rate->whole = be24_to_cpu(query_result.shaper_query.ertcr) >> 13;
-+ token_rate->fraction = be24_to_cpu(query_result.shaper_query.ertcr) &
-+ 0x1FFF;
-+ *token_limit = be16_to_cpu(query_result.shaper_query.ertbl);
-+ return 0;
-+}
-+EXPORT_SYMBOL(qman_ceetm_lni_get_excess_rate);
-+
-+int qman_ceetm_lni_get_excess_rate_bps(struct qm_ceetm_lni *lni,
-+ u64 *bps, u16 *token_limit)
-+{
-+ struct qm_ceetm_rate token_rate;
-+ int ret;
-+
-+ ret = qman_ceetm_lni_get_excess_rate(lni, &token_rate, token_limit);
-+ if (ret) {
-+ pr_err("The LNI ER rate or limit is not available\n");
-+ return -EINVAL;
-+ }
-+
-+ return qman_ceetm_tokenrate2bps(&token_rate, bps, 0);
-+}
-+EXPORT_SYMBOL(qman_ceetm_lni_get_excess_rate_bps);
-+
-+#define QMAN_CEETM_LNITCFCC_CQ_LEVEL_SHIFT(n) ((15 - n) * 4)
-+#define QMAN_CEETM_LNITCFCC_ENABLE 0x8
-+int qman_ceetm_lni_set_tcfcc(struct qm_ceetm_lni *lni,
-+ unsigned int cq_level,
-+ int traffic_class)
-+{
-+ struct qm_mcc_ceetm_mapping_shaper_tcfc_config config_opts;
-+ struct qm_mcc_ceetm_mapping_shaper_tcfc_query query_opts;
-+ struct qm_mcr_ceetm_mapping_shaper_tcfc_query query_result;
-+ u64 lnitcfcc;
-+
-+ if ((cq_level > 15) | (traffic_class > 7)) {
-+ pr_err("The CQ or traffic class id is out of range\n");
-+ return -EINVAL;
-+ }
-+
-+ query_opts.cid = cpu_to_be16(CEETM_COMMAND_TCFC | lni->idx);
-+ query_opts.dcpid = lni->dcp_idx;
-+ if (qman_ceetm_query_mapping_shaper_tcfc(&query_opts, &query_result)) {
-+ pr_err("Fail to query tcfcc\n");
-+ return -EINVAL;
-+ }
-+
-+ lnitcfcc = be64_to_cpu(query_result.tcfc_query.lnitcfcc);
-+ if (traffic_class == -1) {
-+ /* disable tcfc for this CQ */
-+ lnitcfcc &= ~((u64)QMAN_CEETM_LNITCFCC_ENABLE <<
-+ QMAN_CEETM_LNITCFCC_CQ_LEVEL_SHIFT(cq_level));
-+ } else {
-+ lnitcfcc &= ~((u64)0xF <<
-+ QMAN_CEETM_LNITCFCC_CQ_LEVEL_SHIFT(cq_level));
-+ lnitcfcc |= ((u64)(QMAN_CEETM_LNITCFCC_ENABLE |
-+ traffic_class)) <<
-+ QMAN_CEETM_LNITCFCC_CQ_LEVEL_SHIFT(cq_level);
-+ }
-+ config_opts.tcfc_config.lnitcfcc = cpu_to_be64(lnitcfcc);
-+ config_opts.cid = cpu_to_be16(CEETM_COMMAND_TCFC | lni->idx);
-+ config_opts.dcpid = lni->dcp_idx;
-+ return qman_ceetm_configure_mapping_shaper_tcfc(&config_opts);
-+}
-+EXPORT_SYMBOL(qman_ceetm_lni_set_tcfcc);
-+
-+#define QMAN_CEETM_LNITCFCC_TC_MASK 0x7
-+int qman_ceetm_lni_get_tcfcc(struct qm_ceetm_lni *lni, unsigned int cq_level,
-+ int *traffic_class)
-+{
-+ struct qm_mcc_ceetm_mapping_shaper_tcfc_query query_opts;
-+ struct qm_mcr_ceetm_mapping_shaper_tcfc_query query_result;
-+ int ret;
-+ u8 lnitcfcc;
-+
-+ if (cq_level > 15) {
-+ pr_err("the CQ level is out of range\n");
-+ return -EINVAL;
-+ }
-+
-+ query_opts.cid = cpu_to_be16(CEETM_COMMAND_TCFC | lni->idx);
-+ query_opts.dcpid = lni->dcp_idx;
-+ ret = qman_ceetm_query_mapping_shaper_tcfc(&query_opts, &query_result);
-+ if (ret)
-+ return ret;
-+ lnitcfcc = (u8)be64_to_cpu((query_result.tcfc_query.lnitcfcc) >>
-+ QMAN_CEETM_LNITCFCC_CQ_LEVEL_SHIFT(cq_level));
-+ if (lnitcfcc & QMAN_CEETM_LNITCFCC_ENABLE)
-+ *traffic_class = lnitcfcc & QMAN_CEETM_LNITCFCC_TC_MASK;
-+ else
-+ *traffic_class = -1;
-+ return 0;
-+}
-+EXPORT_SYMBOL(qman_ceetm_lni_get_tcfcc);
-+
-+int qman_ceetm_channel_claim(struct qm_ceetm_channel **channel,
-+ struct qm_ceetm_lni *lni)
-+{
-+ struct qm_ceetm_channel *p;
-+ u32 channel_idx;
-+ int ret = 0;
-+ struct qm_mcc_ceetm_mapping_shaper_tcfc_config config_opts;
-+
-+ if (lni->dcp_idx == qm_dc_portal_fman0) {
-+ ret = qman_alloc_ceetm0_channel(&channel_idx);
-+ } else if (lni->dcp_idx == qm_dc_portal_fman1) {
-+ ret = qman_alloc_ceetm1_channel(&channel_idx);
-+ } else {
-+ pr_err("dcp_idx %u does not correspond to a known fman in this driver\n",
-+ lni->dcp_idx);
-+ return -EINVAL;
-+ }
-+
-+ if (ret) {
-+ pr_err("The is no channel available for LNI#%d\n", lni->idx);
-+ return -ENODEV;
-+ }
-+
-+ p = kzalloc(sizeof(*p), GFP_KERNEL);
-+ if (!p)
-+ return -ENOMEM;
-+ p->idx = channel_idx;
-+ p->dcp_idx = lni->dcp_idx;
-+ p->lni_idx = lni->idx;
-+ list_add_tail(&p->node, &lni->channels);
-+ INIT_LIST_HEAD(&p->class_queues);
-+ INIT_LIST_HEAD(&p->ccgs);
-+ config_opts.cid = cpu_to_be16(CEETM_COMMAND_CHANNEL_MAPPING |
-+ channel_idx);
-+ config_opts.dcpid = lni->dcp_idx;
-+ config_opts.channel_mapping.map_lni_id = lni->idx;
-+ config_opts.channel_mapping.map_shaped = 0;
-+ if (qman_ceetm_configure_mapping_shaper_tcfc(&config_opts)) {
-+ pr_err("Can't map channel#%d for LNI#%d\n",
-+ channel_idx, lni->idx);
-+ return -EINVAL;
-+ }
-+ *channel = p;
-+ return 0;
-+}
-+EXPORT_SYMBOL(qman_ceetm_channel_claim);
-+
-+int qman_ceetm_channel_release(struct qm_ceetm_channel *channel)
-+{
-+ struct qm_mcc_ceetm_mapping_shaper_tcfc_config config_opts;
-+ if (!list_empty(&channel->class_queues)) {
-+ pr_err("CEETM channel#%d has class queue unreleased!\n",
-+ channel->idx);
-+ return -EBUSY;
-+ }
-+ if (!list_empty(&channel->ccgs)) {
-+ pr_err("CEETM channel#%d has ccg unreleased!\n",
-+ channel->idx);
-+ return -EBUSY;
-+ }
-+
-+ /* channel->dcp_idx corresponds to known fman validation */
-+ if ((channel->dcp_idx != qm_dc_portal_fman0) &&
-+ (channel->dcp_idx != qm_dc_portal_fman1)) {
-+ pr_err("dcp_idx %u does not correspond to a known fman in this driver\n",
-+ channel->dcp_idx);
-+ return -EINVAL;
-+ }
-+
-+ config_opts.cid = cpu_to_be16(CEETM_COMMAND_CHANNEL_SHAPER |
-+ channel->idx);
-+ config_opts.dcpid = channel->dcp_idx;
-+ memset(&config_opts.shaper_config, 0,
-+ sizeof(config_opts.shaper_config));
-+ if (qman_ceetm_configure_mapping_shaper_tcfc(&config_opts)) {
-+ pr_err("Can't reset channel shapping parameters\n");
-+ return -EINVAL;
-+ }
-+
-+ if (channel->dcp_idx == qm_dc_portal_fman0) {
-+ qman_release_ceetm0_channelid(channel->idx);
-+ } else if (channel->dcp_idx == qm_dc_portal_fman1) {
-+ qman_release_ceetm1_channelid(channel->idx);
-+ } else {
-+ pr_err("dcp_idx %u does not correspond to a known fman in this driver\n",
-+ channel->dcp_idx);
-+ return -EINVAL;
-+ }
-+ list_del(&channel->node);
-+ kfree(channel);
-+
-+ return 0;
-+}
-+EXPORT_SYMBOL(qman_ceetm_channel_release);
-+
-+int qman_ceetm_channel_enable_shaper(struct qm_ceetm_channel *channel,
-+ int coupled)
-+{
-+ struct qm_mcc_ceetm_mapping_shaper_tcfc_query query_opts;
-+ struct qm_mcr_ceetm_mapping_shaper_tcfc_query query_result;
-+ struct qm_mcc_ceetm_mapping_shaper_tcfc_config config_opts;
-+
-+ if (channel->shaper_enable == 1) {
-+ pr_err("This channel shaper has been enabled!\n");
-+ return -EINVAL;
-+ }
-+
-+ channel->shaper_enable = 1;
-+ channel->shaper_couple = coupled;
-+
-+ query_opts.cid = cpu_to_be16(CEETM_COMMAND_CHANNEL_MAPPING |
-+ channel->idx);
-+ query_opts.dcpid = channel->dcp_idx;
-+
-+ if (qman_ceetm_query_mapping_shaper_tcfc(&query_opts, &query_result)) {
-+ pr_err("Can't query channel mapping\n");
-+ return -EINVAL;
-+ }
-+
-+ config_opts.cid = cpu_to_be16(CEETM_COMMAND_CHANNEL_MAPPING |
-+ channel->idx);
-+ config_opts.dcpid = channel->dcp_idx;
-+ config_opts.channel_mapping.map_lni_id =
-+ query_result.channel_mapping_query.map_lni_id;
-+ config_opts.channel_mapping.map_shaped = 1;
-+ if (qman_ceetm_configure_mapping_shaper_tcfc(&config_opts)) {
-+ pr_err("Can't enable shaper for channel #%d\n", channel->idx);
-+ return -EINVAL;
-+ }
-+
-+ config_opts.cid = cpu_to_be16(CEETM_COMMAND_CHANNEL_SHAPER |
-+ channel->idx);
-+ config_opts.shaper_config.cpl = coupled;
-+ config_opts.shaper_config.crtcr =
-+ cpu_to_be24((channel->cr_token_rate.whole
-+ << 13) |
-+ channel->cr_token_rate.fraction);
-+ config_opts.shaper_config.ertcr =
-+ cpu_to_be24(channel->er_token_rate.whole
-+ << 13 |
-+ channel->er_token_rate.fraction);
-+ config_opts.shaper_config.crtbl =
-+ cpu_to_be16(channel->cr_token_bucket_limit);
-+ config_opts.shaper_config.ertbl =
-+ cpu_to_be16(channel->er_token_bucket_limit);
-+
-+ return qman_ceetm_configure_mapping_shaper_tcfc(&config_opts);
-+}
-+EXPORT_SYMBOL(qman_ceetm_channel_enable_shaper);
-+
-+int qman_ceetm_channel_disable_shaper(struct qm_ceetm_channel *channel)
-+{
-+ struct qm_mcc_ceetm_mapping_shaper_tcfc_query query_opts;
-+ struct qm_mcr_ceetm_mapping_shaper_tcfc_query query_result;
-+ struct qm_mcc_ceetm_mapping_shaper_tcfc_config config_opts;
-+
-+
-+ query_opts.cid = cpu_to_be16(CEETM_COMMAND_CHANNEL_MAPPING |
-+ channel->idx);
-+ query_opts.dcpid = channel->dcp_idx;
-+
-+ if (qman_ceetm_query_mapping_shaper_tcfc(&query_opts, &query_result)) {
-+ pr_err("Can't query channel mapping\n");
-+ return -EINVAL;
-+ }
-+
-+ config_opts.cid = cpu_to_be16(CEETM_COMMAND_CHANNEL_MAPPING |
-+ channel->idx);
-+ config_opts.dcpid = channel->dcp_idx;
-+ config_opts.channel_mapping.map_shaped = 0;
-+ config_opts.channel_mapping.map_lni_id =
-+ query_result.channel_mapping_query.map_lni_id;
-+
-+ return qman_ceetm_configure_mapping_shaper_tcfc(&config_opts);
-+}
-+EXPORT_SYMBOL(qman_ceetm_channel_disable_shaper);
-+
-+int qman_ceetm_channel_is_shaper_enabled(struct qm_ceetm_channel *channel)
-+{
-+ struct qm_mcc_ceetm_mapping_shaper_tcfc_query query_opts;
-+ struct qm_mcr_ceetm_mapping_shaper_tcfc_query query_result;
-+
-+ query_opts.cid = cpu_to_be16(CEETM_COMMAND_CHANNEL_MAPPING |
-+ channel->idx);
-+ query_opts.dcpid = channel->dcp_idx;
-+
-+ if (qman_ceetm_query_mapping_shaper_tcfc(&query_opts, &query_result)) {
-+ pr_err("Can't query channel mapping\n");
-+ return -EINVAL;
-+ }
-+
-+ return query_result.channel_mapping_query.map_shaped;
-+}
-+EXPORT_SYMBOL(qman_ceetm_channel_is_shaper_enabled);
-+
-+int qman_ceetm_channel_set_commit_rate(struct qm_ceetm_channel *channel,
-+ const struct qm_ceetm_rate *token_rate,
-+ u16 token_limit)
-+{
-+ struct qm_mcc_ceetm_mapping_shaper_tcfc_config config_opts;
-+ struct qm_mcc_ceetm_mapping_shaper_tcfc_query query_opts;
-+ struct qm_mcr_ceetm_mapping_shaper_tcfc_query query_result;
-+ int ret;
-+
-+ query_opts.cid = cpu_to_be16(CEETM_COMMAND_CHANNEL_SHAPER |
-+ channel->idx);
-+ query_opts.dcpid = channel->dcp_idx;
-+
-+ ret = qman_ceetm_query_mapping_shaper_tcfc(&query_opts, &query_result);
-+ if (ret) {
-+ pr_err("Fail to get the current channel shaper setting\n");
-+ return -EINVAL;
-+ }
-+
-+ channel->cr_token_rate.whole = token_rate->whole;
-+ channel->cr_token_rate.fraction = token_rate->fraction;
-+ channel->cr_token_bucket_limit = token_limit;
-+ config_opts.cid = cpu_to_be16(CEETM_COMMAND_CHANNEL_SHAPER |
-+ channel->idx);
-+ config_opts.dcpid = channel->dcp_idx;
-+ config_opts.shaper_config.crtcr = cpu_to_be24((token_rate->whole
-+ << 13) | (token_rate->fraction));
-+ config_opts.shaper_config.crtbl = cpu_to_be16(token_limit);
-+ config_opts.shaper_config.cpl = query_result.shaper_query.cpl;
-+ config_opts.shaper_config.ertcr = query_result.shaper_query.ertcr;
-+ config_opts.shaper_config.ertbl = query_result.shaper_query.ertbl;
-+ return qman_ceetm_configure_mapping_shaper_tcfc(&config_opts);
-+}
-+EXPORT_SYMBOL(qman_ceetm_channel_set_commit_rate);
-+
-+int qman_ceetm_channel_set_commit_rate_bps(struct qm_ceetm_channel *channel,
-+ u64 bps, u16 token_limit)
-+{
-+ struct qm_ceetm_rate token_rate;
-+ int ret;
-+
-+ ret = qman_ceetm_bps2tokenrate(bps, &token_rate, 0);
-+ if (ret) {
-+ pr_err("Can not convert bps to token rate\n");
-+ return -EINVAL;
-+ }
-+ return qman_ceetm_channel_set_commit_rate(channel, &token_rate,
-+ token_limit);
-+}
-+EXPORT_SYMBOL(qman_ceetm_channel_set_commit_rate_bps);
-+
-+int qman_ceetm_channel_get_commit_rate(struct qm_ceetm_channel *channel,
-+ struct qm_ceetm_rate *token_rate,
-+ u16 *token_limit)
-+{
-+ struct qm_mcc_ceetm_mapping_shaper_tcfc_query query_opts;
-+ struct qm_mcr_ceetm_mapping_shaper_tcfc_query query_result;
-+ int ret;
-+
-+ query_opts.cid = cpu_to_be16(CEETM_COMMAND_CHANNEL_SHAPER |
-+ channel->idx);
-+ query_opts.dcpid = channel->dcp_idx;
-+
-+ ret = qman_ceetm_query_mapping_shaper_tcfc(&query_opts, &query_result);
-+ if (ret | !query_result.shaper_query.crtcr |
-+ !query_result.shaper_query.crtbl) {
-+ pr_err("The channel commit rate or limit is not set\n");
-+ return -EINVAL;
-+ }
-+ token_rate->whole = be24_to_cpu(query_result.shaper_query.crtcr) >> 13;
-+ token_rate->fraction = be24_to_cpu(query_result.shaper_query.crtcr) &
-+ 0x1FFF;
-+ *token_limit = be16_to_cpu(query_result.shaper_query.crtbl);
-+ return 0;
-+}
-+EXPORT_SYMBOL(qman_ceetm_channel_get_commit_rate);
-+
-+int qman_ceetm_channel_get_commit_rate_bps(struct qm_ceetm_channel *channel,
-+ u64 *bps, u16 *token_limit)
-+{
-+ struct qm_ceetm_rate token_rate;
-+ int ret;
-+
-+ ret = qman_ceetm_channel_get_commit_rate(channel, &token_rate,
-+ token_limit);
-+ if (ret) {
-+ pr_err("The channel CR rate or limit is not available\n");
-+ return -EINVAL;
-+ }
-+
-+ return qman_ceetm_tokenrate2bps(&token_rate, bps, 0);
-+}
-+EXPORT_SYMBOL(qman_ceetm_channel_get_commit_rate_bps);
-+
-+int qman_ceetm_channel_set_excess_rate(struct qm_ceetm_channel *channel,
-+ const struct qm_ceetm_rate *token_rate,
-+ u16 token_limit)
-+{
-+ struct qm_mcc_ceetm_mapping_shaper_tcfc_config config_opts;
-+ struct qm_mcc_ceetm_mapping_shaper_tcfc_query query_opts;
-+ struct qm_mcr_ceetm_mapping_shaper_tcfc_query query_result;
-+ int ret;
-+
-+ query_opts.cid = cpu_to_be16(CEETM_COMMAND_CHANNEL_SHAPER |
-+ channel->idx);
-+ query_opts.dcpid = channel->dcp_idx;
-+ ret = qman_ceetm_query_mapping_shaper_tcfc(&query_opts, &query_result);
-+ if (ret) {
-+ pr_err("Fail to get the current channel shaper setting\n");
-+ return -EINVAL;
-+ }
-+
-+ channel->er_token_rate.whole = token_rate->whole;
-+ channel->er_token_rate.fraction = token_rate->fraction;
-+ channel->er_token_bucket_limit = token_limit;
-+ config_opts.cid = cpu_to_be16(CEETM_COMMAND_CHANNEL_SHAPER |
-+ channel->idx);
-+ config_opts.dcpid = channel->dcp_idx;
-+ config_opts.shaper_config.ertcr = cpu_to_be24(
-+ (token_rate->whole << 13) | (token_rate->fraction));
-+ config_opts.shaper_config.ertbl = cpu_to_be16(token_limit);
-+ config_opts.shaper_config.cpl = query_result.shaper_query.cpl;
-+ config_opts.shaper_config.crtcr = query_result.shaper_query.crtcr;
-+ config_opts.shaper_config.crtbl = query_result.shaper_query.crtbl;
-+ return qman_ceetm_configure_mapping_shaper_tcfc(&config_opts);
-+}
-+EXPORT_SYMBOL(qman_ceetm_channel_set_excess_rate);
-+
-+int qman_ceetm_channel_set_excess_rate_bps(struct qm_ceetm_channel *channel,
-+ u64 bps, u16 token_limit)
-+{
-+ struct qm_ceetm_rate token_rate;
-+ int ret;
-+
-+ ret = qman_ceetm_bps2tokenrate(bps, &token_rate, 0);
-+ if (ret) {
-+ pr_err("Can not convert bps to token rate\n");
-+ return -EINVAL;
-+ }
-+ return qman_ceetm_channel_set_excess_rate(channel, &token_rate,
-+ token_limit);
-+}
-+EXPORT_SYMBOL(qman_ceetm_channel_set_excess_rate_bps);
-+
-+int qman_ceetm_channel_get_excess_rate(struct qm_ceetm_channel *channel,
-+ struct qm_ceetm_rate *token_rate,
-+ u16 *token_limit)
-+{
-+ struct qm_mcc_ceetm_mapping_shaper_tcfc_query query_opts;
-+ struct qm_mcr_ceetm_mapping_shaper_tcfc_query query_result;
-+ int ret;
-+
-+ query_opts.cid = cpu_to_be16(CEETM_COMMAND_CHANNEL_SHAPER |
-+ channel->idx);
-+ query_opts.dcpid = channel->dcp_idx;
-+ ret = qman_ceetm_query_mapping_shaper_tcfc(&query_opts, &query_result);
-+ if (ret | !query_result.shaper_query.ertcr |
-+ !query_result.shaper_query.ertbl) {
-+ pr_err("The channel excess rate or limit is not set\n");
-+ return -EINVAL;
-+ }
-+ token_rate->whole = be24_to_cpu(query_result.shaper_query.ertcr) >> 13;
-+ token_rate->fraction = be24_to_cpu(query_result.shaper_query.ertcr) &
-+ 0x1FFF;
-+ *token_limit = be16_to_cpu(query_result.shaper_query.ertbl);
-+ return 0;
-+}
-+EXPORT_SYMBOL(qman_ceetm_channel_get_excess_rate);
-+
-+int qman_ceetm_channel_get_excess_rate_bps(struct qm_ceetm_channel *channel,
-+ u64 *bps, u16 *token_limit)
-+{
-+ struct qm_ceetm_rate token_rate;
-+ int ret;
-+
-+ ret = qman_ceetm_channel_get_excess_rate(channel, &token_rate,
-+ token_limit);
-+ if (ret) {
-+ pr_err("The channel ER rate or limit is not available\n");
-+ return -EINVAL;
-+ }
-+
-+ return qman_ceetm_tokenrate2bps(&token_rate, bps, 0);
-+}
-+EXPORT_SYMBOL(qman_ceetm_channel_get_excess_rate_bps);
-+
-+int qman_ceetm_channel_set_weight(struct qm_ceetm_channel *channel,
-+ u16 token_limit)
-+{
-+ struct qm_mcc_ceetm_mapping_shaper_tcfc_config config_opts;
-+
-+ if (channel->shaper_enable) {
-+ pr_err("This channel is a shaped one\n");
-+ return -EINVAL;
-+ }
-+
-+ channel->cr_token_bucket_limit = token_limit;
-+ config_opts.cid = cpu_to_be16(CEETM_COMMAND_CHANNEL_SHAPER |
-+ channel->idx);
-+ config_opts.dcpid = channel->dcp_idx;
-+ config_opts.shaper_config.crtbl = cpu_to_be16(token_limit);
-+ return qman_ceetm_configure_mapping_shaper_tcfc(&config_opts);
-+}
-+EXPORT_SYMBOL(qman_ceetm_channel_set_weight);
-+
-+int qman_ceetm_channel_get_weight(struct qm_ceetm_channel *channel,
-+ u16 *token_limit)
-+{
-+ struct qm_mcc_ceetm_mapping_shaper_tcfc_query query_opts;
-+ struct qm_mcr_ceetm_mapping_shaper_tcfc_query query_result;
-+ int ret;
-+
-+ query_opts.cid = cpu_to_be16(CEETM_COMMAND_CHANNEL_SHAPER |
-+ channel->idx);
-+ query_opts.dcpid = channel->dcp_idx;
-+ ret = qman_ceetm_query_mapping_shaper_tcfc(&query_opts, &query_result);
-+ if (ret | !query_result.shaper_query.crtbl) {
-+ pr_err("This unshaped channel's uFQ wight is unavailable\n");
-+ return -EINVAL;
-+ }
-+ *token_limit = be16_to_cpu(query_result.shaper_query.crtbl);
-+ return 0;
-+}
-+EXPORT_SYMBOL(qman_ceetm_channel_get_weight);
-+
-+int qman_ceetm_channel_set_group(struct qm_ceetm_channel *channel, int group_b,
-+ unsigned int prio_a, unsigned int prio_b)
-+{
-+ struct qm_mcc_ceetm_class_scheduler_config config_opts;
-+ struct qm_mcr_ceetm_class_scheduler_query query_result;
-+ int i;
-+
-+ if (prio_a > 7) {
-+ pr_err("The priority of group A is out of range\n");
-+ return -EINVAL;
-+ }
-+ if (group_b && (prio_b > 7)) {
-+ pr_err("The priority of group B is out of range\n");
-+ return -EINVAL;
-+ }
-+
-+ if (qman_ceetm_query_class_scheduler(channel, &query_result)) {
-+ pr_err("Can't query channel#%d's scheduler!\n", channel->idx);
-+ return -EINVAL;
-+ }
-+
-+ config_opts.cqcid = cpu_to_be16(channel->idx);
-+ config_opts.dcpid = channel->dcp_idx;
-+ config_opts.gpc_combine_flag = !group_b;
-+ config_opts.gpc_prio_a = prio_a;
-+ config_opts.gpc_prio_b = prio_b;
-+
-+ for (i = 0; i < 8; i++)
-+ config_opts.w[i] = query_result.w[i];
-+ config_opts.crem = query_result.crem;
-+ config_opts.erem = query_result.erem;
-+
-+ return qman_ceetm_configure_class_scheduler(&config_opts);
-+}
-+EXPORT_SYMBOL(qman_ceetm_channel_set_group);
-+
-+int qman_ceetm_channel_get_group(struct qm_ceetm_channel *channel, int *group_b,
-+ unsigned int *prio_a, unsigned int *prio_b)
-+{
-+ struct qm_mcr_ceetm_class_scheduler_query query_result;
-+
-+ if (qman_ceetm_query_class_scheduler(channel, &query_result)) {
-+ pr_err("Can't query channel#%d's scheduler!\n", channel->idx);
-+ return -EINVAL;
-+ }
-+ *group_b = !query_result.gpc_combine_flag;
-+ *prio_a = query_result.gpc_prio_a;
-+ *prio_b = query_result.gpc_prio_b;
-+
-+ return 0;
-+}
-+EXPORT_SYMBOL(qman_ceetm_channel_get_group);
-+
-+#define GROUP_A_ELIGIBILITY_SET (1 << 8)
-+#define GROUP_B_ELIGIBILITY_SET (1 << 9)
-+#define CQ_ELIGIBILITY_SET(n) (1 << (7 - n))
-+int qman_ceetm_channel_set_group_cr_eligibility(struct qm_ceetm_channel
-+ *channel, int group_b, int cre)
-+{
-+ struct qm_mcc_ceetm_class_scheduler_config csch_config;
-+ struct qm_mcr_ceetm_class_scheduler_query csch_query;
-+ int i;
-+
-+ if (qman_ceetm_query_class_scheduler(channel, &csch_query)) {
-+ pr_err("Cannot get the channel %d scheduler setting.\n",
-+ channel->idx);
-+ return -EINVAL;
-+ }
-+ csch_config.cqcid = cpu_to_be16(channel->idx);
-+ csch_config.dcpid = channel->dcp_idx;
-+ csch_config.gpc_combine_flag = csch_query.gpc_combine_flag;
-+ csch_config.gpc_prio_a = csch_query.gpc_prio_a;
-+ csch_config.gpc_prio_b = csch_query.gpc_prio_b;
-+
-+ for (i = 0; i < 8; i++)
-+ csch_config.w[i] = csch_query.w[i];
-+ csch_config.erem = csch_query.erem;
-+ if (group_b)
-+ csch_config.crem = (be16_to_cpu(csch_query.crem)
-+ & ~GROUP_B_ELIGIBILITY_SET)
-+ | (cre ? GROUP_B_ELIGIBILITY_SET : 0);
-+ else
-+ csch_config.crem = (be16_to_cpu(csch_query.crem)
-+ & ~GROUP_A_ELIGIBILITY_SET)
-+ | (cre ? GROUP_A_ELIGIBILITY_SET : 0);
-+
-+ csch_config.crem = cpu_to_be16(csch_config.crem);
-+
-+ if (qman_ceetm_configure_class_scheduler(&csch_config)) {
-+ pr_err("Cannot config channel %d's scheduler with "
-+ "group_%c's cr eligibility\n", channel->idx,
-+ group_b ? 'b' : 'a');
-+ return -EINVAL;
-+ }
-+
-+ return 0;
-+}
-+EXPORT_SYMBOL(qman_ceetm_channel_set_group_cr_eligibility);
-+
-+int qman_ceetm_channel_set_group_er_eligibility(struct qm_ceetm_channel
-+ *channel, int group_b, int ere)
-+{
-+ struct qm_mcc_ceetm_class_scheduler_config csch_config;
-+ struct qm_mcr_ceetm_class_scheduler_query csch_query;
-+ int i;
-+
-+ if (qman_ceetm_query_class_scheduler(channel, &csch_query)) {
-+ pr_err("Cannot get the channel %d scheduler setting.\n",
-+ channel->idx);
-+ return -EINVAL;
-+ }
-+ csch_config.cqcid = cpu_to_be16(channel->idx);
-+ csch_config.dcpid = channel->dcp_idx;
-+ csch_config.gpc_combine_flag = csch_query.gpc_combine_flag;
-+ csch_config.gpc_prio_a = csch_query.gpc_prio_a;
-+ csch_config.gpc_prio_b = csch_query.gpc_prio_b;
-+
-+ for (i = 0; i < 8; i++)
-+ csch_config.w[i] = csch_query.w[i];
-+ csch_config.crem = csch_query.crem;
-+ if (group_b)
-+ csch_config.erem = (be16_to_cpu(csch_query.erem)
-+ & ~GROUP_B_ELIGIBILITY_SET)
-+ | (ere ? GROUP_B_ELIGIBILITY_SET : 0);
-+ else
-+ csch_config.erem = (be16_to_cpu(csch_query.erem)
-+ & ~GROUP_A_ELIGIBILITY_SET)
-+ | (ere ? GROUP_A_ELIGIBILITY_SET : 0);
-+
-+ csch_config.erem = cpu_to_be16(csch_config.erem);
-+
-+ if (qman_ceetm_configure_class_scheduler(&csch_config)) {
-+ pr_err("Cannot config channel %d's scheduler with "
-+ "group_%c's er eligibility\n", channel->idx,
-+ group_b ? 'b' : 'a');
-+ return -EINVAL;
-+ }
-+
-+ return 0;
-+}
-+EXPORT_SYMBOL(qman_ceetm_channel_set_group_er_eligibility);
-+
-+int qman_ceetm_channel_set_cq_cr_eligibility(struct qm_ceetm_channel *channel,
-+ unsigned int idx, int cre)
-+{
-+ struct qm_mcc_ceetm_class_scheduler_config csch_config;
-+ struct qm_mcr_ceetm_class_scheduler_query csch_query;
-+ int i;
-+
-+ if (idx > 7) {
-+ pr_err("CQ index is out of range\n");
-+ return -EINVAL;
-+ }
-+ if (qman_ceetm_query_class_scheduler(channel, &csch_query)) {
-+ pr_err("Cannot get the channel %d scheduler setting.\n",
-+ channel->idx);
-+ return -EINVAL;
-+ }
-+ csch_config.cqcid = cpu_to_be16(channel->idx);
-+ csch_config.dcpid = channel->dcp_idx;
-+ csch_config.gpc_combine_flag = csch_query.gpc_combine_flag;
-+ csch_config.gpc_prio_a = csch_query.gpc_prio_a;
-+ csch_config.gpc_prio_b = csch_query.gpc_prio_b;
-+ for (i = 0; i < 8; i++)
-+ csch_config.w[i] = csch_query.w[i];
-+ csch_config.erem = csch_query.erem;
-+ csch_config.crem = (be16_to_cpu(csch_query.crem)
-+ & ~CQ_ELIGIBILITY_SET(idx)) |
-+ (cre ? CQ_ELIGIBILITY_SET(idx) : 0);
-+ csch_config.crem = cpu_to_be16(csch_config.crem);
-+ if (qman_ceetm_configure_class_scheduler(&csch_config)) {
-+ pr_err("Cannot config channel scheduler to set "
-+ "cr eligibility mask for CQ#%d\n", idx);
-+ return -EINVAL;
-+ }
-+
-+ return 0;
-+}
-+EXPORT_SYMBOL(qman_ceetm_channel_set_cq_cr_eligibility);
-+
-+int qman_ceetm_channel_set_cq_er_eligibility(struct qm_ceetm_channel *channel,
-+ unsigned int idx, int ere)
-+{
-+ struct qm_mcc_ceetm_class_scheduler_config csch_config;
-+ struct qm_mcr_ceetm_class_scheduler_query csch_query;
-+ int i;
-+
-+ if (idx > 7) {
-+ pr_err("CQ index is out of range\n");
-+ return -EINVAL;
-+ }
-+ if (qman_ceetm_query_class_scheduler(channel, &csch_query)) {
-+ pr_err("Cannot get the channel %d scheduler setting.\n",
-+ channel->idx);
-+ return -EINVAL;
-+ }
-+ csch_config.cqcid = cpu_to_be16(channel->idx);
-+ csch_config.dcpid = channel->dcp_idx;
-+ csch_config.gpc_combine_flag = csch_query.gpc_combine_flag;
-+ csch_config.gpc_prio_a = csch_query.gpc_prio_a;
-+ csch_config.gpc_prio_b = csch_query.gpc_prio_b;
-+ for (i = 0; i < 8; i++)
-+ csch_config.w[i] = csch_query.w[i];
-+ csch_config.crem = csch_query.crem;
-+ csch_config.erem = (be16_to_cpu(csch_query.erem)
-+ & ~CQ_ELIGIBILITY_SET(idx)) |
-+ (ere ? CQ_ELIGIBILITY_SET(idx) : 0);
-+ csch_config.erem = cpu_to_be16(csch_config.erem);
-+ if (qman_ceetm_configure_class_scheduler(&csch_config)) {
-+ pr_err("Cannot config channel scheduler to set "
-+ "er eligibility mask for CQ#%d\n", idx);
-+ return -EINVAL;
-+ }
-+ return 0;
-+}
-+EXPORT_SYMBOL(qman_ceetm_channel_set_cq_er_eligibility);
-+
-+int qman_ceetm_cq_claim(struct qm_ceetm_cq **cq,
-+ struct qm_ceetm_channel *channel, unsigned int idx,
-+ struct qm_ceetm_ccg *ccg)
-+{
-+ struct qm_ceetm_cq *p;
-+ struct qm_mcc_ceetm_cq_config cq_config;
-+
-+ if (idx > 7) {
-+ pr_err("The independent class queue id is out of range\n");
-+ return -EINVAL;
-+ }
-+
-+ list_for_each_entry(p, &channel->class_queues, node) {
-+ if (p->idx == idx) {
-+ pr_err("The CQ#%d has been claimed!\n", idx);
-+ return -EINVAL;
-+ }
-+ }
-+
-+ p = kmalloc(sizeof(*p), GFP_KERNEL);
-+ if (!p) {
-+ pr_err("Can't allocate memory for CQ#%d!\n", idx);
-+ return -ENOMEM;
-+ }
-+
-+ list_add_tail(&p->node, &channel->class_queues);
-+ p->idx = idx;
-+ p->is_claimed = 1;
-+ p->parent = channel;
-+ INIT_LIST_HEAD(&p->bound_lfqids);
-+
-+ if (ccg) {
-+ cq_config.cqid = cpu_to_be16((channel->idx << 4) | idx);
-+ cq_config.dcpid = channel->dcp_idx;
-+ cq_config.ccgid = cpu_to_be16(ccg->idx);
-+ if (qman_ceetm_configure_cq(&cq_config)) {
-+ pr_err("Can't configure the CQ#%d with CCGRID#%d\n",
-+ idx, ccg->idx);
-+ list_del(&p->node);
-+ kfree(p);
-+ return -EINVAL;
-+ }
-+ }
-+
-+ *cq = p;
-+ return 0;
-+}
-+EXPORT_SYMBOL(qman_ceetm_cq_claim);
-+
-+int qman_ceetm_cq_claim_A(struct qm_ceetm_cq **cq,
-+ struct qm_ceetm_channel *channel, unsigned int idx,
-+ struct qm_ceetm_ccg *ccg)
-+{
-+ struct qm_ceetm_cq *p;
-+ struct qm_mcc_ceetm_cq_config cq_config;
-+
-+ if ((idx < 8) || (idx > 15)) {
-+ pr_err("This grouped class queue id is out of range\n");
-+ return -EINVAL;
-+ }
-+
-+ list_for_each_entry(p, &channel->class_queues, node) {
-+ if (p->idx == idx) {
-+ pr_err("The CQ#%d has been claimed!\n", idx);
-+ return -EINVAL;
-+ }
-+ }
-+
-+ p = kmalloc(sizeof(*p), GFP_KERNEL);
-+ if (!p) {
-+ pr_err("Can't allocate memory for CQ#%d!\n", idx);
-+ return -ENOMEM;
-+ }
-+
-+ list_add_tail(&p->node, &channel->class_queues);
-+ p->idx = idx;
-+ p->is_claimed = 1;
-+ p->parent = channel;
-+ INIT_LIST_HEAD(&p->bound_lfqids);
-+
-+ if (ccg) {
-+ cq_config.cqid = cpu_to_be16((channel->idx << 4) | idx);
-+ cq_config.dcpid = channel->dcp_idx;
-+ cq_config.ccgid = cpu_to_be16(ccg->idx);
-+ if (qman_ceetm_configure_cq(&cq_config)) {
-+ pr_err("Can't configure the CQ#%d with CCGRID#%d\n",
-+ idx, ccg->idx);
-+ list_del(&p->node);
-+ kfree(p);
-+ return -EINVAL;
-+ }
-+ }
-+ *cq = p;
-+ return 0;
-+}
-+EXPORT_SYMBOL(qman_ceetm_cq_claim_A);
-+
-+int qman_ceetm_cq_claim_B(struct qm_ceetm_cq **cq,
-+ struct qm_ceetm_channel *channel, unsigned int idx,
-+ struct qm_ceetm_ccg *ccg)
-+{
-+ struct qm_ceetm_cq *p;
-+ struct qm_mcc_ceetm_cq_config cq_config;
-+
-+ if ((idx < 12) || (idx > 15)) {
-+ pr_err("This grouped class queue id is out of range\n");
-+ return -EINVAL;
-+ }
-+
-+ list_for_each_entry(p, &channel->class_queues, node) {
-+ if (p->idx == idx) {
-+ pr_err("The CQ#%d has been claimed!\n", idx);
-+ return -EINVAL;
-+ }
-+ }
-+
-+ p = kmalloc(sizeof(*p), GFP_KERNEL);
-+ if (!p) {
-+ pr_err("Can't allocate memory for CQ#%d!\n", idx);
-+ return -ENOMEM;
-+ }
-+
-+ list_add_tail(&p->node, &channel->class_queues);
-+ p->idx = idx;
-+ p->is_claimed = 1;
-+ p->parent = channel;
-+ INIT_LIST_HEAD(&p->bound_lfqids);
-+
-+ if (ccg) {
-+ cq_config.cqid = cpu_to_be16((channel->idx << 4) | idx);
-+ cq_config.dcpid = channel->dcp_idx;
-+ cq_config.ccgid = cpu_to_be16(ccg->idx);
-+ if (qman_ceetm_configure_cq(&cq_config)) {
-+ pr_err("Can't configure the CQ#%d with CCGRID#%d\n",
-+ idx, ccg->idx);
-+ list_del(&p->node);
-+ kfree(p);
-+ return -EINVAL;
-+ }
-+ }
-+ *cq = p;
-+ return 0;
-+}
-+EXPORT_SYMBOL(qman_ceetm_cq_claim_B);
-+
-+int qman_ceetm_cq_release(struct qm_ceetm_cq *cq)
-+{
-+ if (!list_empty(&cq->bound_lfqids)) {
-+ pr_err("The CQ#%d has unreleased LFQID\n", cq->idx);
-+ return -EBUSY;
-+ }
-+ list_del(&cq->node);
-+ qman_ceetm_drain_cq(cq);
-+ kfree(cq);
-+ return 0;
-+}
-+EXPORT_SYMBOL(qman_ceetm_cq_release);
-+
-+int qman_ceetm_set_queue_weight(struct qm_ceetm_cq *cq,
-+ struct qm_ceetm_weight_code *weight_code)
-+{
-+ struct qm_mcc_ceetm_class_scheduler_config config_opts;
-+ struct qm_mcr_ceetm_class_scheduler_query query_result;
-+ int i;
-+
-+ if (cq->idx < 8) {
-+ pr_err("Can not set weight for ungrouped class queue\n");
-+ return -EINVAL;
-+ }
-+
-+ if (qman_ceetm_query_class_scheduler(cq->parent, &query_result)) {
-+ pr_err("Can't query channel#%d's scheduler!\n",
-+ cq->parent->idx);
-+ return -EINVAL;
-+ }
-+
-+ config_opts.cqcid = cpu_to_be16(cq->parent->idx);
-+ config_opts.dcpid = cq->parent->dcp_idx;
-+ config_opts.crem = query_result.crem;
-+ config_opts.erem = query_result.erem;
-+ config_opts.gpc_combine_flag = query_result.gpc_combine_flag;
-+ config_opts.gpc_prio_a = query_result.gpc_prio_a;
-+ config_opts.gpc_prio_b = query_result.gpc_prio_b;
-+
-+ for (i = 0; i < 8; i++)
-+ config_opts.w[i] = query_result.w[i];
-+ config_opts.w[cq->idx - 8] = ((weight_code->y << 3) |
-+ (weight_code->x & 0x7));
-+ return qman_ceetm_configure_class_scheduler(&config_opts);
-+}
-+EXPORT_SYMBOL(qman_ceetm_set_queue_weight);
-+
-+int qman_ceetm_get_queue_weight(struct qm_ceetm_cq *cq,
-+ struct qm_ceetm_weight_code *weight_code)
-+{
-+ struct qm_mcr_ceetm_class_scheduler_query query_result;
-+
-+ if (cq->idx < 8) {
-+ pr_err("Can not get weight for ungrouped class queue\n");
-+ return -EINVAL;
-+ }
-+
-+ if (qman_ceetm_query_class_scheduler(cq->parent,
-+ &query_result)) {
-+ pr_err("Can't get the weight code for CQ#%d!\n", cq->idx);
-+ return -EINVAL;
-+ }
-+ weight_code->y = query_result.w[cq->idx - 8] >> 3;
-+ weight_code->x = query_result.w[cq->idx - 8] & 0x7;
-+
-+ return 0;
-+}
-+EXPORT_SYMBOL(qman_ceetm_get_queue_weight);
-+
-+/* The WBFS code is represent as {x,y}, the effect wieght can be calculated as:
-+ * effective weight = 2^x / (1 - (y/64))
-+ * = 2^(x+6) / (64 - y)
-+ */
-+static void reduce_fraction(u32 *n, u32 *d)
-+{
-+ u32 factor = 2;
-+ u32 lesser = (*n < *d) ? *n : *d;
-+ /* If factor exceeds the square-root of the lesser of *n and *d,
-+ * then there's no point continuing. Proof: if there was a factor
-+ * bigger than the square root, that would imply there exists
-+ * another factor smaller than the square-root with which it
-+ * multiplies to give 'lesser' - but that's a contradiction
-+ * because the other factor would have already been found and
-+ * divided out.
-+ */
-+ while ((factor * factor) <= lesser) {
-+ /* If 'factor' is a factor of *n and *d, divide them both
-+ * by 'factor' as many times as possible.
-+ */
-+ while (!(*n % factor) && !(*d % factor)) {
-+ *n /= factor;
-+ *d /= factor;
-+ lesser /= factor;
-+ }
-+ if (factor == 2)
-+ factor = 3;
-+ else
-+ factor += 2;
-+ }
-+}
-+
-+int qman_ceetm_wbfs2ratio(struct qm_ceetm_weight_code *weight_code,
-+ u32 *numerator,
-+ u32 *denominator)
-+{
-+ *numerator = (u32) 1 << (weight_code->x + 6);
-+ *denominator = 64 - weight_code->y;
-+ reduce_fraction(numerator, denominator);
-+ return 0;
-+}
-+EXPORT_SYMBOL(qman_ceetm_wbfs2ratio);
-+
-+/* For a given x, the weight is between 2^x (inclusive) and 2^(x+1) (exclusive).
-+ * So find 'x' by range, and then estimate 'y' using:
-+ * 64 - y = 2^(x + 6) / weight
-+ * = 2^(x + 6) / (n/d)
-+ * = d * 2^(x+6) / n
-+ * y = 64 - (d * 2^(x+6) / n)
-+ */
-+int qman_ceetm_ratio2wbfs(u32 numerator,
-+ u32 denominator,
-+ struct qm_ceetm_weight_code *weight_code,
-+ int rounding)
-+{
-+ unsigned int y, x = 0;
-+ /* search incrementing 'x' until:
-+ * weight < 2^(x+1)
-+ * n/d < 2^(x+1)
-+ * n < d * 2^(x+1)
-+ */
-+ while ((x < 8) && (numerator >= (denominator << (x + 1))))
-+ x++;
-+ if (x >= 8)
-+ return -ERANGE;
-+ /* because of the subtraction, use '-rounding' */
-+ y = 64 - ROUNDING(denominator << (x + 6), numerator, -rounding);
-+ if (y >= 32)
-+ return -ERANGE;
-+ weight_code->x = x;
-+ weight_code->y = y;
-+ return 0;
-+}
-+EXPORT_SYMBOL(qman_ceetm_ratio2wbfs);
-+
-+int qman_ceetm_set_queue_weight_in_ratio(struct qm_ceetm_cq *cq, u32 ratio)
-+{
-+ struct qm_ceetm_weight_code weight_code;
-+
-+ if (qman_ceetm_ratio2wbfs(ratio, 100, &weight_code, 0)) {
-+ pr_err("Cannot get wbfs code for cq %x\n", cq->idx);
-+ return -EINVAL;
-+ }
-+ return qman_ceetm_set_queue_weight(cq, &weight_code);
-+}
-+EXPORT_SYMBOL(qman_ceetm_set_queue_weight_in_ratio);
-+
-+int qman_ceetm_get_queue_weight_in_ratio(struct qm_ceetm_cq *cq, u32 *ratio)
-+{
-+ struct qm_ceetm_weight_code weight_code;
-+ u32 n, d;
-+
-+ if (qman_ceetm_get_queue_weight(cq, &weight_code)) {
-+ pr_err("Cannot query the weight code for cq%x\n", cq->idx);
-+ return -EINVAL;
-+ }
-+
-+ if (qman_ceetm_wbfs2ratio(&weight_code, &n, &d)) {
-+ pr_err("Cannot get the ratio with wbfs code\n");
-+ return -EINVAL;
-+ }
-+
-+ *ratio = (n * 100) / d;
-+ return 0;
-+}
-+EXPORT_SYMBOL(qman_ceetm_get_queue_weight_in_ratio);
-+
-+int qman_ceetm_cq_get_dequeue_statistics(struct qm_ceetm_cq *cq, u32 flags,
-+ u64 *frame_count, u64 *byte_count)
-+{
-+ struct qm_mcr_ceetm_statistics_query result;
-+ u16 cid, command_type;
-+ enum qm_dc_portal dcp_idx;
-+ int ret;
-+
-+ cid = cpu_to_be16((cq->parent->idx << 4) | cq->idx);
-+ dcp_idx = cq->parent->dcp_idx;
-+ if (flags == QMAN_CEETM_FLAG_CLEAR_STATISTICS_COUNTER)
-+ command_type = CEETM_QUERY_DEQUEUE_CLEAR_STATISTICS;
-+ else
-+ command_type = CEETM_QUERY_DEQUEUE_STATISTICS;
-+
-+ ret = qman_ceetm_query_statistics(cid, dcp_idx, command_type, &result);
-+ if (ret) {
-+ pr_err("Can't query the statistics of CQ#%d!\n", cq->idx);
-+ return -EINVAL;
-+ }
-+
-+ *frame_count = be40_to_cpu(result.frm_cnt);
-+ *byte_count = be48_to_cpu(result.byte_cnt);
-+ return 0;
-+}
-+EXPORT_SYMBOL(qman_ceetm_cq_get_dequeue_statistics);
-+
-+int qman_ceetm_drain_cq(struct qm_ceetm_cq *cq)
-+{
-+ struct qm_mcr_ceetm_cq_peek_pop_xsfdrread ppxr;
-+ int ret;
-+
-+ do {
-+ ret = qman_ceetm_cq_peek_pop_xsfdrread(cq, 1, 0, &ppxr);
-+ if (ret) {
-+ pr_err("Failed to pop frame from CQ\n");
-+ return -EINVAL;
-+ }
-+ } while (!(ppxr.stat & 0x2));
-+
-+ return 0;
-+}
-+EXPORT_SYMBOL(qman_ceetm_drain_cq);
-+
-+#define CEETM_LFQMT_LFQID_MSB 0xF00000
-+#define CEETM_LFQMT_LFQID_LSB 0x000FFF
-+int qman_ceetm_lfq_claim(struct qm_ceetm_lfq **lfq,
-+ struct qm_ceetm_cq *cq)
-+{
-+ struct qm_ceetm_lfq *p;
-+ u32 lfqid;
-+ int ret = 0;
-+ struct qm_mcc_ceetm_lfqmt_config lfqmt_config;
-+
-+ if (cq->parent->dcp_idx == qm_dc_portal_fman0) {
-+ ret = qman_alloc_ceetm0_lfqid(&lfqid);
-+ } else if (cq->parent->dcp_idx == qm_dc_portal_fman1) {
-+ ret = qman_alloc_ceetm1_lfqid(&lfqid);
-+ } else {
-+ pr_err("dcp_idx %u does not correspond to a known fman in this driver\n",
-+ cq->parent->dcp_idx);
-+ return -EINVAL;
-+ }
-+
-+ if (ret) {
-+ pr_err("There is no lfqid avalaible for CQ#%d!\n", cq->idx);
-+ return -ENODEV;
-+ }
-+ p = kmalloc(sizeof(*p), GFP_KERNEL);
-+ if (!p)
-+ return -ENOMEM;
-+ p->idx = lfqid;
-+ p->dctidx = (u16)(lfqid & CEETM_LFQMT_LFQID_LSB);
-+ p->parent = cq->parent;
-+ list_add_tail(&p->node, &cq->bound_lfqids);
-+
-+ lfqmt_config.lfqid = cpu_to_be24(CEETM_LFQMT_LFQID_MSB |
-+ (cq->parent->dcp_idx << 16) |
-+ (lfqid & CEETM_LFQMT_LFQID_LSB));
-+ lfqmt_config.cqid = cpu_to_be16((cq->parent->idx << 4) | (cq->idx));
-+ lfqmt_config.dctidx = cpu_to_be16(p->dctidx);
-+ if (qman_ceetm_configure_lfqmt(&lfqmt_config)) {
-+ pr_err("Can't configure LFQMT for LFQID#%d @ CQ#%d\n",
-+ lfqid, cq->idx);
-+ list_del(&p->node);
-+ kfree(p);
-+ return -EINVAL;
-+ }
-+ *lfq = p;
-+ return 0;
-+}
-+EXPORT_SYMBOL(qman_ceetm_lfq_claim);
-+
-+int qman_ceetm_lfq_release(struct qm_ceetm_lfq *lfq)
-+{
-+ if (lfq->parent->dcp_idx == qm_dc_portal_fman0) {
-+ qman_release_ceetm0_lfqid(lfq->idx);
-+ } else if (lfq->parent->dcp_idx == qm_dc_portal_fman1) {
-+ qman_release_ceetm1_lfqid(lfq->idx);
-+ } else {
-+ pr_err("dcp_idx %u does not correspond to a known fman in this driver\n",
-+ lfq->parent->dcp_idx);
-+ return -EINVAL;
-+ }
-+ list_del(&lfq->node);
-+ kfree(lfq);
-+ return 0;
-+}
-+EXPORT_SYMBOL(qman_ceetm_lfq_release);
-+
-+int qman_ceetm_lfq_set_context(struct qm_ceetm_lfq *lfq, u64 context_a,
-+ u32 context_b)
-+{
-+ struct qm_mcc_ceetm_dct_config dct_config;
-+ lfq->context_a = context_a;
-+ lfq->context_b = context_b;
-+ dct_config.dctidx = cpu_to_be16((u16)lfq->dctidx);
-+ dct_config.dcpid = lfq->parent->dcp_idx;
-+ dct_config.context_b = cpu_to_be32(context_b);
-+ dct_config.context_a = cpu_to_be64(context_a);
-+
-+ return qman_ceetm_configure_dct(&dct_config);
-+}
-+EXPORT_SYMBOL(qman_ceetm_lfq_set_context);
-+
-+int qman_ceetm_lfq_get_context(struct qm_ceetm_lfq *lfq, u64 *context_a,
-+ u32 *context_b)
-+{
-+ struct qm_mcc_ceetm_dct_query dct_query;
-+ struct qm_mcr_ceetm_dct_query query_result;
-+
-+ dct_query.dctidx = cpu_to_be16(lfq->dctidx);
-+ dct_query.dcpid = lfq->parent->dcp_idx;
-+ if (qman_ceetm_query_dct(&dct_query, &query_result)) {
-+ pr_err("Can't query LFQID#%d's context!\n", lfq->idx);
-+ return -EINVAL;
-+ }
-+ *context_a = be64_to_cpu(query_result.context_a);
-+ *context_b = be32_to_cpu(query_result.context_b);
-+ return 0;
-+}
-+EXPORT_SYMBOL(qman_ceetm_lfq_get_context);
-+
-+int qman_ceetm_create_fq(struct qm_ceetm_lfq *lfq, struct qman_fq *fq)
-+{
-+ spin_lock_init(&fq->fqlock);
-+ fq->fqid = lfq->idx;
-+ fq->flags = QMAN_FQ_FLAG_NO_MODIFY;
-+ if (lfq->ern)
-+ fq->cb.ern = lfq->ern;
-+#ifdef CONFIG_FSL_QMAN_FQ_LOOKUP
-+ if (unlikely(find_empty_fq_table_entry(&fq->key, fq)))
-+ return -ENOMEM;
-+#endif
-+ return 0;
-+}
-+EXPORT_SYMBOL(qman_ceetm_create_fq);
-+
-+#define MAX_CCG_IDX 0x000F
-+int qman_ceetm_ccg_claim(struct qm_ceetm_ccg **ccg,
-+ struct qm_ceetm_channel *channel,
-+ unsigned int idx,
-+ void (*cscn)(struct qm_ceetm_ccg *,
-+ void *cb_ctx,
-+ int congested),
-+ void *cb_ctx)
-+{
-+ struct qm_ceetm_ccg *p;
-+
-+ if (idx > MAX_CCG_IDX) {
-+ pr_err("The given ccg index is out of range\n");
-+ return -EINVAL;
-+ }
-+
-+ list_for_each_entry(p, &channel->ccgs, node) {
-+ if (p->idx == idx) {
-+ pr_err("The CCG#%d has been claimed\n", idx);
-+ return -EINVAL;
-+ }
-+ }
-+
-+ p = kmalloc(sizeof(*p), GFP_KERNEL);
-+ if (!p) {
-+ pr_err("Can't allocate memory for CCG#%d!\n", idx);
-+ return -ENOMEM;
-+ }
-+
-+ list_add_tail(&p->node, &channel->ccgs);
-+
-+ p->idx = idx;
-+ p->parent = channel;
-+ p->cb = cscn;
-+ p->cb_ctx = cb_ctx;
-+ INIT_LIST_HEAD(&p->cb_node);
-+
-+ *ccg = p;
-+ return 0;
-+}
-+EXPORT_SYMBOL(qman_ceetm_ccg_claim);
-+
-+int qman_ceetm_ccg_release(struct qm_ceetm_ccg *ccg)
-+{
-+ unsigned long irqflags __maybe_unused;
-+ struct qm_mcc_ceetm_ccgr_config config_opts;
-+ int ret = 0;
-+ struct qman_portal *p = get_affine_portal();
-+
-+ memset(&config_opts, 0, sizeof(struct qm_mcc_ceetm_ccgr_config));
-+ spin_lock_irqsave(&p->ccgr_lock, irqflags);
-+ if (!list_empty(&ccg->cb_node))
-+ list_del(&ccg->cb_node);
-+ config_opts.ccgrid = cpu_to_be16(CEETM_CCGR_CM_CONFIGURE |
-+ (ccg->parent->idx << 4) | ccg->idx);
-+ config_opts.dcpid = ccg->parent->dcp_idx;
-+ config_opts.we_mask = cpu_to_be16(QM_CCGR_WE_CSCN_TUPD);
-+ config_opts.cm_config.cscn_tupd = cpu_to_be16(PORTAL_IDX(p));
-+ ret = qman_ceetm_configure_ccgr(&config_opts);
-+ spin_unlock_irqrestore(&p->ccgr_lock, irqflags);
-+ put_affine_portal();
-+
-+ list_del(&ccg->node);
-+ kfree(ccg);
-+ return ret;
-+}
-+EXPORT_SYMBOL(qman_ceetm_ccg_release);
-+
-+int qman_ceetm_ccg_set(struct qm_ceetm_ccg *ccg, u16 we_mask,
-+ const struct qm_ceetm_ccg_params *params)
-+{
-+ struct qm_mcc_ceetm_ccgr_config config_opts;
-+ unsigned long irqflags __maybe_unused;
-+ int ret;
-+ struct qman_portal *p;
-+
-+ if (((ccg->parent->idx << 4) | ccg->idx) >= (2 * __CGR_NUM))
-+ return -EINVAL;
-+
-+ p = get_affine_portal();
-+
-+ memset(&config_opts, 0, sizeof(struct qm_mcc_ceetm_ccgr_config));
-+ spin_lock_irqsave(&p->ccgr_lock, irqflags);
-+
-+ config_opts.ccgrid = cpu_to_be16(CEETM_CCGR_CM_CONFIGURE |
-+ (ccg->parent->idx << 4) | ccg->idx);
-+ config_opts.dcpid = ccg->parent->dcp_idx;
-+ config_opts.we_mask = we_mask;
-+ if (we_mask & QM_CCGR_WE_CSCN_EN) {
-+ config_opts.we_mask |= QM_CCGR_WE_CSCN_TUPD;
-+ config_opts.cm_config.cscn_tupd = cpu_to_be16(
-+ QM_CGR_TARG_UDP_CTRL_WRITE_BIT | PORTAL_IDX(p));
-+ }
-+ config_opts.we_mask = cpu_to_be16(config_opts.we_mask);
-+ config_opts.cm_config.ctl_wr_en_g = params->wr_en_g;
-+ config_opts.cm_config.ctl_wr_en_y = params->wr_en_y;
-+ config_opts.cm_config.ctl_wr_en_r = params->wr_en_r;
-+ config_opts.cm_config.ctl_td_en = params->td_en;
-+ config_opts.cm_config.ctl_td_mode = params->td_mode;
-+ config_opts.cm_config.ctl_cscn_en = params->cscn_en;
-+ config_opts.cm_config.ctl_mode = params->mode;
-+ config_opts.cm_config.oal = params->oal;
-+ config_opts.cm_config.cs_thres.hword =
-+ cpu_to_be16(params->cs_thres_in.hword);
-+ config_opts.cm_config.cs_thres_x.hword =
-+ cpu_to_be16(params->cs_thres_out.hword);
-+ config_opts.cm_config.td_thres.hword =
-+ cpu_to_be16(params->td_thres.hword);
-+ config_opts.cm_config.wr_parm_g.word =
-+ cpu_to_be32(params->wr_parm_g.word);
-+ config_opts.cm_config.wr_parm_y.word =
-+ cpu_to_be32(params->wr_parm_y.word);
-+ config_opts.cm_config.wr_parm_r.word =
-+ cpu_to_be32(params->wr_parm_r.word);
-+ ret = qman_ceetm_configure_ccgr(&config_opts);
-+ if (ret) {
-+ pr_err("Configure CCGR CM failed!\n");
-+ goto release_lock;
-+ }
-+
-+ if (we_mask & QM_CCGR_WE_CSCN_EN)
-+ if (list_empty(&ccg->cb_node))
-+ list_add(&ccg->cb_node,
-+ &p->ccgr_cbs[ccg->parent->dcp_idx]);
-+release_lock:
-+ spin_unlock_irqrestore(&p->ccgr_lock, irqflags);
-+ put_affine_portal();
-+ return ret;
-+}
-+EXPORT_SYMBOL(qman_ceetm_ccg_set);
-+
-+#define CEETM_CCGR_CTL_MASK 0x01
-+int qman_ceetm_ccg_get(struct qm_ceetm_ccg *ccg,
-+ struct qm_ceetm_ccg_params *params)
-+{
-+ struct qm_mcc_ceetm_ccgr_query query_opts;
-+ struct qm_mcr_ceetm_ccgr_query query_result;
-+
-+ query_opts.ccgrid = cpu_to_be16(CEETM_CCGR_CM_QUERY |
-+ (ccg->parent->idx << 4) | ccg->idx);
-+ query_opts.dcpid = ccg->parent->dcp_idx;
-+
-+ if (qman_ceetm_query_ccgr(&query_opts, &query_result)) {
-+ pr_err("Can't query CCGR#%d\n", ccg->idx);
-+ return -EINVAL;
-+ }
-+
-+ params->wr_parm_r.word = query_result.cm_query.wr_parm_r.word;
-+ params->wr_parm_y.word = query_result.cm_query.wr_parm_y.word;
-+ params->wr_parm_g.word = query_result.cm_query.wr_parm_g.word;
-+ params->td_thres.hword = query_result.cm_query.td_thres.hword;
-+ params->cs_thres_out.hword = query_result.cm_query.cs_thres_x.hword;
-+ params->cs_thres_in.hword = query_result.cm_query.cs_thres.hword;
-+ params->oal = query_result.cm_query.oal;
-+ params->wr_en_g = query_result.cm_query.ctl_wr_en_g;
-+ params->wr_en_y = query_result.cm_query.ctl_wr_en_y;
-+ params->wr_en_r = query_result.cm_query.ctl_wr_en_r;
-+ params->td_en = query_result.cm_query.ctl_td_en;
-+ params->td_mode = query_result.cm_query.ctl_td_mode;
-+ params->cscn_en = query_result.cm_query.ctl_cscn_en;
-+ params->mode = query_result.cm_query.ctl_mode;
-+
-+ return 0;
-+}
-+EXPORT_SYMBOL(qman_ceetm_ccg_get);
-+
-+int qman_ceetm_ccg_get_reject_statistics(struct qm_ceetm_ccg *ccg, u32 flags,
-+ u64 *frame_count, u64 *byte_count)
-+{
-+ struct qm_mcr_ceetm_statistics_query result;
-+ u16 cid, command_type;
-+ enum qm_dc_portal dcp_idx;
-+ int ret;
-+
-+ cid = cpu_to_be16((ccg->parent->idx << 4) | ccg->idx);
-+ dcp_idx = ccg->parent->dcp_idx;
-+ if (flags == QMAN_CEETM_FLAG_CLEAR_STATISTICS_COUNTER)
-+ command_type = CEETM_QUERY_REJECT_CLEAR_STATISTICS;
-+ else
-+ command_type = CEETM_QUERY_REJECT_STATISTICS;
-+
-+ ret = qman_ceetm_query_statistics(cid, dcp_idx, command_type, &result);
-+ if (ret) {
-+ pr_err("Can't query the statistics of CCG#%d!\n", ccg->idx);
-+ return -EINVAL;
-+ }
-+
-+ *frame_count = be40_to_cpu(result.frm_cnt);
-+ *byte_count = be48_to_cpu(result.byte_cnt);
-+ return 0;
-+}
-+EXPORT_SYMBOL(qman_ceetm_ccg_get_reject_statistics);
-+
-+int qman_ceetm_cscn_swp_get(struct qm_ceetm_ccg *ccg,
-+ u16 swp_idx,
-+ unsigned int *cscn_enabled)
-+{
-+ struct qm_mcc_ceetm_ccgr_query query_opts;
-+ struct qm_mcr_ceetm_ccgr_query query_result;
-+ int i;
-+
-+ DPA_ASSERT(swp_idx < 127);
-+ query_opts.ccgrid = cpu_to_be16(CEETM_CCGR_CM_QUERY |
-+ (ccg->parent->idx << 4) | ccg->idx);
-+ query_opts.dcpid = ccg->parent->dcp_idx;
-+
-+ if (qman_ceetm_query_ccgr(&query_opts, &query_result)) {
-+ pr_err("Can't query CCGR#%d\n", ccg->idx);
-+ return -EINVAL;
-+ }
-+
-+ i = swp_idx / 32;
-+ i = 3 - i;
-+ *cscn_enabled = query_result.cm_query.cscn_targ_swp[i] >>
-+ (31 - swp_idx % 32);
-+
-+ return 0;
-+}
-+EXPORT_SYMBOL(qman_ceetm_cscn_swp_get);
-+
-+int qman_ceetm_cscn_dcp_set(struct qm_ceetm_ccg *ccg,
-+ u16 dcp_idx,
-+ u8 vcgid,
-+ unsigned int cscn_enabled,
-+ u16 we_mask,
-+ const struct qm_ceetm_ccg_params *params)
-+{
-+ struct qm_mcc_ceetm_ccgr_config config_opts;
-+ int ret;
-+
-+ config_opts.ccgrid = cpu_to_be16(CEETM_CCGR_CM_CONFIGURE |
-+ (ccg->parent->idx << 4) | ccg->idx);
-+ config_opts.dcpid = ccg->parent->dcp_idx;
-+ config_opts.we_mask = cpu_to_be16(we_mask | QM_CCGR_WE_CSCN_TUPD |
-+ QM_CCGR_WE_CDV);
-+ config_opts.cm_config.cdv = vcgid;
-+ config_opts.cm_config.cscn_tupd = cpu_to_be16((cscn_enabled << 15) |
-+ QM_CGR_TARG_UDP_CTRL_DCP | dcp_idx);
-+ config_opts.cm_config.ctl_wr_en_g = params->wr_en_g;
-+ config_opts.cm_config.ctl_wr_en_y = params->wr_en_y;
-+ config_opts.cm_config.ctl_wr_en_r = params->wr_en_r;
-+ config_opts.cm_config.ctl_td_en = params->td_en;
-+ config_opts.cm_config.ctl_td_mode = params->td_mode;
-+ config_opts.cm_config.ctl_cscn_en = params->cscn_en;
-+ config_opts.cm_config.ctl_mode = params->mode;
-+ config_opts.cm_config.cs_thres.hword =
-+ cpu_to_be16(params->cs_thres_in.hword);
-+ config_opts.cm_config.cs_thres_x.hword =
-+ cpu_to_be16(params->cs_thres_out.hword);
-+ config_opts.cm_config.td_thres.hword =
-+ cpu_to_be16(params->td_thres.hword);
-+ config_opts.cm_config.wr_parm_g.word =
-+ cpu_to_be32(params->wr_parm_g.word);
-+ config_opts.cm_config.wr_parm_y.word =
-+ cpu_to_be32(params->wr_parm_y.word);
-+ config_opts.cm_config.wr_parm_r.word =
-+ cpu_to_be32(params->wr_parm_r.word);
-+
-+ ret = qman_ceetm_configure_ccgr(&config_opts);
-+ if (ret) {
-+ pr_err("Configure CSCN_TARG_DCP failed!\n");
-+ return -EINVAL;
-+ }
-+ return 0;
-+}
-+EXPORT_SYMBOL(qman_ceetm_cscn_dcp_set);
-+
-+int qman_ceetm_cscn_dcp_get(struct qm_ceetm_ccg *ccg,
-+ u16 dcp_idx,
-+ u8 *vcgid,
-+ unsigned int *cscn_enabled)
-+{
-+ struct qm_mcc_ceetm_ccgr_query query_opts;
-+ struct qm_mcr_ceetm_ccgr_query query_result;
-+
-+ query_opts.ccgrid = cpu_to_be16(CEETM_CCGR_CM_QUERY |
-+ (ccg->parent->idx << 4) | ccg->idx);
-+ query_opts.dcpid = ccg->parent->dcp_idx;
-+
-+ if (qman_ceetm_query_ccgr(&query_opts, &query_result)) {
-+ pr_err("Can't query CCGR#%d\n", ccg->idx);
-+ return -EINVAL;
-+ }
-+
-+ *vcgid = query_result.cm_query.cdv;
-+ *cscn_enabled = (query_result.cm_query.cscn_targ_dcp >> dcp_idx) & 0x1;
-+ return 0;
-+}
-+EXPORT_SYMBOL(qman_ceetm_cscn_dcp_get);
-+
-+int qman_ceetm_querycongestion(struct __qm_mcr_querycongestion *ccg_state,
-+ unsigned int dcp_idx)
-+{
-+ struct qm_mc_command *mcc;
-+ struct qm_mc_result *mcr;
-+ struct qman_portal *p;
-+ unsigned long irqflags __maybe_unused;
-+ u8 res;
-+ int i, j;
-+
-+ p = get_affine_portal();
-+ PORTAL_IRQ_LOCK(p, irqflags);
-+
-+ mcc = qm_mc_start(&p->p);
-+ for (i = 0; i < 2; i++) {
-+ mcc->ccgr_query.ccgrid =
-+ cpu_to_be16(CEETM_QUERY_CONGESTION_STATE | i);
-+ mcc->ccgr_query.dcpid = dcp_idx;
-+ qm_mc_commit(&p->p, QM_CEETM_VERB_CCGR_QUERY);
-+
-+ while (!(mcr = qm_mc_result(&p->p)))
-+ cpu_relax();
-+ DPA_ASSERT((mcr->verb & QM_MCR_VERB_MASK) ==
-+ QM_CEETM_VERB_CCGR_QUERY);
-+ res = mcr->result;
-+ if (res == QM_MCR_RESULT_OK) {
-+ for (j = 0; j < 8; j++)
-+ mcr->ccgr_query.congestion_state.state.
-+ __state[j] = be32_to_cpu(mcr->ccgr_query.
-+ congestion_state.state.__state[j]);
-+ *(ccg_state + i) =
-+ mcr->ccgr_query.congestion_state.state;
-+ } else {
-+ pr_err("QUERY CEETM CONGESTION STATE failed\n");
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+ return -EIO;
-+ }
-+ }
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+ put_affine_portal();
-+ return 0;
-+}
-+
-+int qman_set_wpm(int wpm_enable)
-+{
-+ return qm_set_wpm(wpm_enable);
-+}
-+EXPORT_SYMBOL(qman_set_wpm);
-+
-+int qman_get_wpm(int *wpm_enable)
-+{
-+ return qm_get_wpm(wpm_enable);
-+}
-+EXPORT_SYMBOL(qman_get_wpm);
-+
-+int qman_shutdown_fq(u32 fqid)
-+{
-+ struct qman_portal *p;
-+ unsigned long irqflags __maybe_unused;
-+ int ret;
-+ struct qm_portal *low_p;
-+ p = get_affine_portal();
-+ PORTAL_IRQ_LOCK(p, irqflags);
-+ low_p = &p->p;
-+ ret = qm_shutdown_fq(&low_p, 1, fqid);
-+ PORTAL_IRQ_UNLOCK(p, irqflags);
-+ put_affine_portal();
-+ return ret;
-+}
-+
-+const struct qm_portal_config *qman_get_qm_portal_config(
-+ struct qman_portal *portal)
-+{
-+ return portal->sharing_redirect ? NULL : portal->config;
-+}
---- /dev/null
-+++ b/drivers/staging/fsl_qbman/qman_low.h
-@@ -0,0 +1,1442 @@
-+/* Copyright 2008-2011 Freescale Semiconductor, Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#include "qman_private.h"
-+
-+/***************************/
-+/* Portal register assists */
-+/***************************/
-+
-+/* Cache-inhibited register offsets */
-+#if defined(CONFIG_PPC32) || defined(CONFIG_PPC64)
-+
-+#define QM_REG_EQCR_PI_CINH 0x0000
-+#define QM_REG_EQCR_CI_CINH 0x0004
-+#define QM_REG_EQCR_ITR 0x0008
-+#define QM_REG_DQRR_PI_CINH 0x0040
-+#define QM_REG_DQRR_CI_CINH 0x0044
-+#define QM_REG_DQRR_ITR 0x0048
-+#define QM_REG_DQRR_DCAP 0x0050
-+#define QM_REG_DQRR_SDQCR 0x0054
-+#define QM_REG_DQRR_VDQCR 0x0058
-+#define QM_REG_DQRR_PDQCR 0x005c
-+#define QM_REG_MR_PI_CINH 0x0080
-+#define QM_REG_MR_CI_CINH 0x0084
-+#define QM_REG_MR_ITR 0x0088
-+#define QM_REG_CFG 0x0100
-+#define QM_REG_ISR 0x0e00
-+#define QM_REG_IIR 0x0e0c
-+#define QM_REG_ITPR 0x0e14
-+
-+/* Cache-enabled register offsets */
-+#define QM_CL_EQCR 0x0000
-+#define QM_CL_DQRR 0x1000
-+#define QM_CL_MR 0x2000
-+#define QM_CL_EQCR_PI_CENA 0x3000
-+#define QM_CL_EQCR_CI_CENA 0x3100
-+#define QM_CL_DQRR_PI_CENA 0x3200
-+#define QM_CL_DQRR_CI_CENA 0x3300
-+#define QM_CL_MR_PI_CENA 0x3400
-+#define QM_CL_MR_CI_CENA 0x3500
-+#define QM_CL_CR 0x3800
-+#define QM_CL_RR0 0x3900
-+#define QM_CL_RR1 0x3940
-+
-+#endif
-+
-+#if defined(CONFIG_ARM) || defined(CONFIG_ARM64)
-+
-+#define QM_REG_EQCR_PI_CINH 0x3000
-+#define QM_REG_EQCR_CI_CINH 0x3040
-+#define QM_REG_EQCR_ITR 0x3080
-+#define QM_REG_DQRR_PI_CINH 0x3100
-+#define QM_REG_DQRR_CI_CINH 0x3140
-+#define QM_REG_DQRR_ITR 0x3180
-+#define QM_REG_DQRR_DCAP 0x31C0
-+#define QM_REG_DQRR_SDQCR 0x3200
-+#define QM_REG_DQRR_VDQCR 0x3240
-+#define QM_REG_DQRR_PDQCR 0x3280
-+#define QM_REG_MR_PI_CINH 0x3300
-+#define QM_REG_MR_CI_CINH 0x3340
-+#define QM_REG_MR_ITR 0x3380
-+#define QM_REG_CFG 0x3500
-+#define QM_REG_ISR 0x3600
-+#define QM_REG_IIR 0x36C0
-+#define QM_REG_ITPR 0x3740
-+
-+/* Cache-enabled register offsets */
-+#define QM_CL_EQCR 0x0000
-+#define QM_CL_DQRR 0x1000
-+#define QM_CL_MR 0x2000
-+#define QM_CL_EQCR_PI_CENA 0x3000
-+#define QM_CL_EQCR_CI_CENA 0x3040
-+#define QM_CL_DQRR_PI_CENA 0x3100
-+#define QM_CL_DQRR_CI_CENA 0x3140
-+#define QM_CL_MR_PI_CENA 0x3300
-+#define QM_CL_MR_CI_CENA 0x3340
-+#define QM_CL_CR 0x3800
-+#define QM_CL_RR0 0x3900
-+#define QM_CL_RR1 0x3940
-+
-+#endif
-+
-+
-+/* BTW, the drivers (and h/w programming model) already obtain the required
-+ * synchronisation for portal accesses via lwsync(), hwsync(), and
-+ * data-dependencies. Use of barrier()s or other order-preserving primitives
-+ * simply degrade performance. Hence the use of the __raw_*() interfaces, which
-+ * simply ensure that the compiler treats the portal registers as volatile (ie.
-+ * non-coherent). */
-+
-+/* Cache-inhibited register access. */
-+#define __qm_in(qm, o) be32_to_cpu(__raw_readl((qm)->addr_ci + (o)))
-+#define __qm_out(qm, o, val) __raw_writel((cpu_to_be32(val)), \
-+ (qm)->addr_ci + (o));
-+#define qm_in(reg) __qm_in(&portal->addr, QM_REG_##reg)
-+#define qm_out(reg, val) __qm_out(&portal->addr, QM_REG_##reg, val)
-+
-+/* Cache-enabled (index) register access */
-+#define __qm_cl_touch_ro(qm, o) dcbt_ro((qm)->addr_ce + (o))
-+#define __qm_cl_touch_rw(qm, o) dcbt_rw((qm)->addr_ce + (o))
-+#define __qm_cl_in(qm, o) be32_to_cpu(__raw_readl((qm)->addr_ce + (o)))
-+#define __qm_cl_out(qm, o, val) \
-+ do { \
-+ u32 *__tmpclout = (qm)->addr_ce + (o); \
-+ __raw_writel(cpu_to_be32(val), __tmpclout); \
-+ dcbf(__tmpclout); \
-+ } while (0)
-+#define __qm_cl_invalidate(qm, o) dcbi((qm)->addr_ce + (o))
-+#define qm_cl_touch_ro(reg) __qm_cl_touch_ro(&portal->addr, QM_CL_##reg##_CENA)
-+#define qm_cl_touch_rw(reg) __qm_cl_touch_rw(&portal->addr, QM_CL_##reg##_CENA)
-+#define qm_cl_in(reg) __qm_cl_in(&portal->addr, QM_CL_##reg##_CENA)
-+#define qm_cl_out(reg, val) __qm_cl_out(&portal->addr, QM_CL_##reg##_CENA, val)
-+#define qm_cl_invalidate(reg)\
-+ __qm_cl_invalidate(&portal->addr, QM_CL_##reg##_CENA)
-+
-+/* Cache-enabled ring access */
-+#define qm_cl(base, idx) ((void *)base + ((idx) << 6))
-+
-+/* Cyclic helper for rings. FIXME: once we are able to do fine-grain perf
-+ * analysis, look at using the "extra" bit in the ring index registers to avoid
-+ * cyclic issues. */
-+static inline u8 qm_cyc_diff(u8 ringsize, u8 first, u8 last)
-+{
-+ /* 'first' is included, 'last' is excluded */
-+ if (first <= last)
-+ return last - first;
-+ return ringsize + last - first;
-+}
-+
-+/* Portal modes.
-+ * Enum types;
-+ * pmode == production mode
-+ * cmode == consumption mode,
-+ * dmode == h/w dequeue mode.
-+ * Enum values use 3 letter codes. First letter matches the portal mode,
-+ * remaining two letters indicate;
-+ * ci == cache-inhibited portal register
-+ * ce == cache-enabled portal register
-+ * vb == in-band valid-bit (cache-enabled)
-+ * dc == DCA (Discrete Consumption Acknowledgement), DQRR-only
-+ * As for "enum qm_dqrr_dmode", it should be self-explanatory.
-+ */
-+enum qm_eqcr_pmode { /* matches QCSP_CFG::EPM */
-+ qm_eqcr_pci = 0, /* PI index, cache-inhibited */
-+ qm_eqcr_pce = 1, /* PI index, cache-enabled */
-+ qm_eqcr_pvb = 2 /* valid-bit */
-+};
-+enum qm_dqrr_dmode { /* matches QCSP_CFG::DP */
-+ qm_dqrr_dpush = 0, /* SDQCR + VDQCR */
-+ qm_dqrr_dpull = 1 /* PDQCR */
-+};
-+enum qm_dqrr_pmode { /* s/w-only */
-+ qm_dqrr_pci, /* reads DQRR_PI_CINH */
-+ qm_dqrr_pce, /* reads DQRR_PI_CENA */
-+ qm_dqrr_pvb /* reads valid-bit */
-+};
-+enum qm_dqrr_cmode { /* matches QCSP_CFG::DCM */
-+ qm_dqrr_cci = 0, /* CI index, cache-inhibited */
-+ qm_dqrr_cce = 1, /* CI index, cache-enabled */
-+ qm_dqrr_cdc = 2 /* Discrete Consumption Acknowledgement */
-+};
-+enum qm_mr_pmode { /* s/w-only */
-+ qm_mr_pci, /* reads MR_PI_CINH */
-+ qm_mr_pce, /* reads MR_PI_CENA */
-+ qm_mr_pvb /* reads valid-bit */
-+};
-+enum qm_mr_cmode { /* matches QCSP_CFG::MM */
-+ qm_mr_cci = 0, /* CI index, cache-inhibited */
-+ qm_mr_cce = 1 /* CI index, cache-enabled */
-+};
-+
-+
-+/* ------------------------- */
-+/* --- Portal structures --- */
-+
-+#define QM_EQCR_SIZE 8
-+#define QM_DQRR_SIZE 16
-+#define QM_MR_SIZE 8
-+
-+struct qm_eqcr {
-+ struct qm_eqcr_entry *ring, *cursor;
-+ u8 ci, available, ithresh, vbit;
-+#ifdef CONFIG_FSL_DPA_CHECKING
-+ u32 busy;
-+ enum qm_eqcr_pmode pmode;
-+#endif
-+};
-+
-+struct qm_dqrr {
-+ const struct qm_dqrr_entry *ring, *cursor;
-+ u8 pi, ci, fill, ithresh, vbit;
-+#ifdef CONFIG_FSL_DPA_CHECKING
-+ enum qm_dqrr_dmode dmode;
-+ enum qm_dqrr_pmode pmode;
-+ enum qm_dqrr_cmode cmode;
-+#endif
-+};
-+
-+struct qm_mr {
-+ const struct qm_mr_entry *ring, *cursor;
-+ u8 pi, ci, fill, ithresh, vbit;
-+#ifdef CONFIG_FSL_DPA_CHECKING
-+ enum qm_mr_pmode pmode;
-+ enum qm_mr_cmode cmode;
-+#endif
-+};
-+
-+struct qm_mc {
-+ struct qm_mc_command *cr;
-+ struct qm_mc_result *rr;
-+ u8 rridx, vbit;
-+#ifdef CONFIG_FSL_DPA_CHECKING
-+ enum {
-+ /* Can be _mc_start()ed */
-+ qman_mc_idle,
-+ /* Can be _mc_commit()ed or _mc_abort()ed */
-+ qman_mc_user,
-+ /* Can only be _mc_retry()ed */
-+ qman_mc_hw
-+ } state;
-+#endif
-+};
-+
-+#define QM_PORTAL_ALIGNMENT ____cacheline_aligned
-+
-+struct qm_addr {
-+ void __iomem *addr_ce; /* cache-enabled */
-+ void __iomem *addr_ci; /* cache-inhibited */
-+};
-+
-+struct qm_portal {
-+ /* In the non-CONFIG_FSL_DPA_CHECKING case, the following stuff up to
-+ * and including 'mc' fits within a cacheline (yay!). The 'config' part
-+ * is setup-only, so isn't a cause for a concern. In other words, don't
-+ * rearrange this structure on a whim, there be dragons ... */
-+ struct qm_addr addr;
-+ struct qm_eqcr eqcr;
-+ struct qm_dqrr dqrr;
-+ struct qm_mr mr;
-+ struct qm_mc mc;
-+} QM_PORTAL_ALIGNMENT;
-+
-+
-+/* ---------------- */
-+/* --- EQCR API --- */
-+
-+/* Bit-wise logic to wrap a ring pointer by clearing the "carry bit" */
-+#define EQCR_CARRYCLEAR(p) \
-+ (void *)((unsigned long)(p) & (~(unsigned long)(QM_EQCR_SIZE << 6)))
-+
-+/* Bit-wise logic to convert a ring pointer to a ring index */
-+static inline u8 EQCR_PTR2IDX(struct qm_eqcr_entry *e)
-+{
-+ return ((uintptr_t)e >> 6) & (QM_EQCR_SIZE - 1);
-+}
-+
-+/* Increment the 'cursor' ring pointer, taking 'vbit' into account */
-+static inline void EQCR_INC(struct qm_eqcr *eqcr)
-+{
-+ /* NB: this is odd-looking, but experiments show that it generates fast
-+ * code with essentially no branching overheads. We increment to the
-+ * next EQCR pointer and handle overflow and 'vbit'. */
-+ struct qm_eqcr_entry *partial = eqcr->cursor + 1;
-+ eqcr->cursor = EQCR_CARRYCLEAR(partial);
-+ if (partial != eqcr->cursor)
-+ eqcr->vbit ^= QM_EQCR_VERB_VBIT;
-+}
-+
-+static inline int qm_eqcr_init(struct qm_portal *portal,
-+ enum qm_eqcr_pmode pmode,
-+ unsigned int eq_stash_thresh,
-+ int eq_stash_prio)
-+{
-+ /* This use of 'register', as well as all other occurrences, is because
-+ * it has been observed to generate much faster code with gcc than is
-+ * otherwise the case. */
-+ register struct qm_eqcr *eqcr = &portal->eqcr;
-+ u32 cfg;
-+ u8 pi;
-+
-+ eqcr->ring = portal->addr.addr_ce + QM_CL_EQCR;
-+ eqcr->ci = qm_in(EQCR_CI_CINH) & (QM_EQCR_SIZE - 1);
-+ qm_cl_invalidate(EQCR_CI);
-+ pi = qm_in(EQCR_PI_CINH) & (QM_EQCR_SIZE - 1);
-+ eqcr->cursor = eqcr->ring + pi;
-+ eqcr->vbit = (qm_in(EQCR_PI_CINH) & QM_EQCR_SIZE) ?
-+ QM_EQCR_VERB_VBIT : 0;
-+ eqcr->available = QM_EQCR_SIZE - 1 -
-+ qm_cyc_diff(QM_EQCR_SIZE, eqcr->ci, pi);
-+ eqcr->ithresh = qm_in(EQCR_ITR);
-+#ifdef CONFIG_FSL_DPA_CHECKING
-+ eqcr->busy = 0;
-+ eqcr->pmode = pmode;
-+#endif
-+ cfg = (qm_in(CFG) & 0x00ffffff) |
-+ (eq_stash_thresh << 28) | /* QCSP_CFG: EST */
-+ (eq_stash_prio << 26) | /* QCSP_CFG: EP */
-+ ((pmode & 0x3) << 24); /* QCSP_CFG::EPM */
-+ qm_out(CFG, cfg);
-+ return 0;
-+}
-+
-+static inline unsigned int qm_eqcr_get_ci_stashing(struct qm_portal *portal)
-+{
-+ return (qm_in(CFG) >> 28) & 0x7;
-+}
-+
-+static inline void qm_eqcr_finish(struct qm_portal *portal)
-+{
-+ register struct qm_eqcr *eqcr = &portal->eqcr;
-+ u8 pi, ci;
-+ u32 cfg;
-+
-+ /*
-+ * Disable EQCI stashing because the QMan only
-+ * presents the value it previously stashed to
-+ * maintain coherency. Setting the stash threshold
-+ * to 1 then 0 ensures that QMan has resyncronized
-+ * its internal copy so that the portal is clean
-+ * when it is reinitialized in the future
-+ */
-+ cfg = (qm_in(CFG) & 0x0fffffff) |
-+ (1 << 28); /* QCSP_CFG: EST */
-+ qm_out(CFG, cfg);
-+ cfg &= 0x0fffffff; /* stash threshold = 0 */
-+ qm_out(CFG, cfg);
-+
-+ pi = qm_in(EQCR_PI_CINH) & (QM_EQCR_SIZE - 1);
-+ ci = qm_in(EQCR_CI_CINH) & (QM_EQCR_SIZE - 1);
-+
-+ /* Refresh EQCR CI cache value */
-+ qm_cl_invalidate(EQCR_CI);
-+ eqcr->ci = qm_cl_in(EQCR_CI) & (QM_EQCR_SIZE - 1);
-+
-+ DPA_ASSERT(!eqcr->busy);
-+ if (pi != EQCR_PTR2IDX(eqcr->cursor))
-+ pr_crit("losing uncommited EQCR entries\n");
-+ if (ci != eqcr->ci)
-+ pr_crit("missing existing EQCR completions\n");
-+ if (eqcr->ci != EQCR_PTR2IDX(eqcr->cursor))
-+ pr_crit("EQCR destroyed unquiesced\n");
-+}
-+
-+static inline struct qm_eqcr_entry *qm_eqcr_start_no_stash(struct qm_portal
-+ *portal)
-+{
-+ register struct qm_eqcr *eqcr = &portal->eqcr;
-+ DPA_ASSERT(!eqcr->busy);
-+ if (!eqcr->available)
-+ return NULL;
-+
-+
-+#ifdef CONFIG_FSL_DPA_CHECKING
-+ eqcr->busy = 1;
-+#endif
-+#if defined(CONFIG_PPC32) || defined(CONFIG_PPC64)
-+ dcbz_64(eqcr->cursor);
-+#endif
-+ return eqcr->cursor;
-+}
-+
-+static inline struct qm_eqcr_entry *qm_eqcr_start_stash(struct qm_portal
-+ *portal)
-+{
-+ register struct qm_eqcr *eqcr = &portal->eqcr;
-+ u8 diff, old_ci;
-+
-+ DPA_ASSERT(!eqcr->busy);
-+ if (!eqcr->available) {
-+ old_ci = eqcr->ci;
-+ eqcr->ci = qm_cl_in(EQCR_CI) & (QM_EQCR_SIZE - 1);
-+ diff = qm_cyc_diff(QM_EQCR_SIZE, old_ci, eqcr->ci);
-+ eqcr->available += diff;
-+ if (!diff)
-+ return NULL;
-+ }
-+#ifdef CONFIG_FSL_DPA_CHECKING
-+ eqcr->busy = 1;
-+#endif
-+#if defined(CONFIG_PPC32) || defined(CONFIG_PPC64)
-+ dcbz_64(eqcr->cursor);
-+#endif
-+ return eqcr->cursor;
-+}
-+
-+static inline void qm_eqcr_abort(struct qm_portal *portal)
-+{
-+ __maybe_unused register struct qm_eqcr *eqcr = &portal->eqcr;
-+ DPA_ASSERT(eqcr->busy);
-+#ifdef CONFIG_FSL_DPA_CHECKING
-+ eqcr->busy = 0;
-+#endif
-+}
-+
-+static inline struct qm_eqcr_entry *qm_eqcr_pend_and_next(
-+ struct qm_portal *portal, u8 myverb)
-+{
-+ register struct qm_eqcr *eqcr = &portal->eqcr;
-+ DPA_ASSERT(eqcr->busy);
-+ DPA_ASSERT(eqcr->pmode != qm_eqcr_pvb);
-+ if (eqcr->available == 1)
-+ return NULL;
-+ eqcr->cursor->__dont_write_directly__verb = myverb | eqcr->vbit;
-+ dcbf(eqcr->cursor);
-+ EQCR_INC(eqcr);
-+ eqcr->available--;
-+#if defined(CONFIG_PPC32) || defined(CONFIG_PPC64)
-+ dcbz_64(eqcr->cursor);
-+#endif
-+ return eqcr->cursor;
-+}
-+
-+#if __BYTE_ORDER__ == __ORDER_LITTLE_ENDIAN__
-+#define EQCR_COMMIT_CHECKS(eqcr) \
-+do { \
-+ DPA_ASSERT(eqcr->busy); \
-+ DPA_ASSERT(eqcr->cursor->orp == (eqcr->cursor->orp & 0xffffff00)); \
-+ DPA_ASSERT(eqcr->cursor->fqid == (eqcr->cursor->fqid & 0xffffff00)); \
-+} while (0)
-+#else
-+#define EQCR_COMMIT_CHECKS(eqcr) \
-+do { \
-+ DPA_ASSERT(eqcr->busy); \
-+ DPA_ASSERT(eqcr->cursor->orp == (eqcr->cursor->orp & \
-+ cpu_to_be32(0x00ffffff))); \
-+ DPA_ASSERT(eqcr->cursor->fqid == (eqcr->cursor->fqid & \
-+ cpu_to_be32(0x00ffffff))); \
-+} while (0)
-+#endif
-+
-+static inline void qm_eqcr_pci_commit(struct qm_portal *portal, u8 myverb)
-+{
-+ register struct qm_eqcr *eqcr = &portal->eqcr;
-+ EQCR_COMMIT_CHECKS(eqcr);
-+ DPA_ASSERT(eqcr->pmode == qm_eqcr_pci);
-+ eqcr->cursor->__dont_write_directly__verb = myverb | eqcr->vbit;
-+ EQCR_INC(eqcr);
-+ eqcr->available--;
-+ dcbf(eqcr->cursor);
-+ hwsync();
-+ qm_out(EQCR_PI_CINH, EQCR_PTR2IDX(eqcr->cursor));
-+#ifdef CONFIG_FSL_DPA_CHECKING
-+ eqcr->busy = 0;
-+#endif
-+}
-+
-+static inline void qm_eqcr_pce_prefetch(struct qm_portal *portal)
-+{
-+ __maybe_unused register struct qm_eqcr *eqcr = &portal->eqcr;
-+ DPA_ASSERT(eqcr->pmode == qm_eqcr_pce);
-+ qm_cl_invalidate(EQCR_PI);
-+ qm_cl_touch_rw(EQCR_PI);
-+}
-+
-+static inline void qm_eqcr_pce_commit(struct qm_portal *portal, u8 myverb)
-+{
-+ register struct qm_eqcr *eqcr = &portal->eqcr;
-+ EQCR_COMMIT_CHECKS(eqcr);
-+ DPA_ASSERT(eqcr->pmode == qm_eqcr_pce);
-+ eqcr->cursor->__dont_write_directly__verb = myverb | eqcr->vbit;
-+ EQCR_INC(eqcr);
-+ eqcr->available--;
-+ dcbf(eqcr->cursor);
-+ lwsync();
-+ qm_cl_out(EQCR_PI, EQCR_PTR2IDX(eqcr->cursor));
-+#ifdef CONFIG_FSL_DPA_CHECKING
-+ eqcr->busy = 0;
-+#endif
-+}
-+
-+static inline void qm_eqcr_pvb_commit(struct qm_portal *portal, u8 myverb)
-+{
-+ register struct qm_eqcr *eqcr = &portal->eqcr;
-+ struct qm_eqcr_entry *eqcursor;
-+ EQCR_COMMIT_CHECKS(eqcr);
-+ DPA_ASSERT(eqcr->pmode == qm_eqcr_pvb);
-+ lwsync();
-+ eqcursor = eqcr->cursor;
-+ eqcursor->__dont_write_directly__verb = myverb | eqcr->vbit;
-+ dcbf(eqcursor);
-+ EQCR_INC(eqcr);
-+ eqcr->available--;
-+#ifdef CONFIG_FSL_DPA_CHECKING
-+ eqcr->busy = 0;
-+#endif
-+}
-+
-+static inline u8 qm_eqcr_cci_update(struct qm_portal *portal)
-+{
-+ register struct qm_eqcr *eqcr = &portal->eqcr;
-+ u8 diff, old_ci = eqcr->ci;
-+ eqcr->ci = qm_in(EQCR_CI_CINH) & (QM_EQCR_SIZE - 1);
-+ diff = qm_cyc_diff(QM_EQCR_SIZE, old_ci, eqcr->ci);
-+ eqcr->available += diff;
-+ return diff;
-+}
-+
-+static inline void qm_eqcr_cce_prefetch(struct qm_portal *portal)
-+{
-+ __maybe_unused register struct qm_eqcr *eqcr = &portal->eqcr;
-+ qm_cl_touch_ro(EQCR_CI);
-+}
-+
-+static inline u8 qm_eqcr_cce_update(struct qm_portal *portal)
-+{
-+ register struct qm_eqcr *eqcr = &portal->eqcr;
-+ u8 diff, old_ci = eqcr->ci;
-+ eqcr->ci = qm_cl_in(EQCR_CI) & (QM_EQCR_SIZE - 1);
-+ qm_cl_invalidate(EQCR_CI);
-+ diff = qm_cyc_diff(QM_EQCR_SIZE, old_ci, eqcr->ci);
-+ eqcr->available += diff;
-+ return diff;
-+}
-+
-+static inline u8 qm_eqcr_get_ithresh(struct qm_portal *portal)
-+{
-+ register struct qm_eqcr *eqcr = &portal->eqcr;
-+ return eqcr->ithresh;
-+}
-+
-+static inline void qm_eqcr_set_ithresh(struct qm_portal *portal, u8 ithresh)
-+{
-+ register struct qm_eqcr *eqcr = &portal->eqcr;
-+ eqcr->ithresh = ithresh;
-+ qm_out(EQCR_ITR, ithresh);
-+}
-+
-+static inline u8 qm_eqcr_get_avail(struct qm_portal *portal)
-+{
-+ register struct qm_eqcr *eqcr = &portal->eqcr;
-+ return eqcr->available;
-+}
-+
-+static inline u8 qm_eqcr_get_fill(struct qm_portal *portal)
-+{
-+ register struct qm_eqcr *eqcr = &portal->eqcr;
-+ return QM_EQCR_SIZE - 1 - eqcr->available;
-+}
-+
-+
-+/* ---------------- */
-+/* --- DQRR API --- */
-+
-+/* FIXME: many possible improvements;
-+ * - look at changing the API to use pointer rather than index parameters now
-+ * that 'cursor' is a pointer,
-+ * - consider moving other parameters to pointer if it could help (ci)
-+ */
-+
-+#define DQRR_CARRYCLEAR(p) \
-+ (void *)((unsigned long)(p) & (~(unsigned long)(QM_DQRR_SIZE << 6)))
-+
-+static inline u8 DQRR_PTR2IDX(const struct qm_dqrr_entry *e)
-+{
-+ return ((uintptr_t)e >> 6) & (QM_DQRR_SIZE - 1);
-+}
-+
-+static inline const struct qm_dqrr_entry *DQRR_INC(
-+ const struct qm_dqrr_entry *e)
-+{
-+ return DQRR_CARRYCLEAR(e + 1);
-+}
-+
-+static inline void qm_dqrr_set_maxfill(struct qm_portal *portal, u8 mf)
-+{
-+ qm_out(CFG, (qm_in(CFG) & 0xff0fffff) |
-+ ((mf & (QM_DQRR_SIZE - 1)) << 20));
-+}
-+
-+static inline void qm_dqrr_cci_consume(struct qm_portal *portal, u8 num)
-+{
-+ register struct qm_dqrr *dqrr = &portal->dqrr;
-+ DPA_ASSERT(dqrr->cmode == qm_dqrr_cci);
-+ dqrr->ci = (dqrr->ci + num) & (QM_DQRR_SIZE - 1);
-+ qm_out(DQRR_CI_CINH, dqrr->ci);
-+}
-+
-+static inline void qm_dqrr_cce_consume(struct qm_portal *portal, u8 num)
-+{
-+ register struct qm_dqrr *dqrr = &portal->dqrr;
-+ DPA_ASSERT(dqrr->cmode == qm_dqrr_cce);
-+ dqrr->ci = (dqrr->ci + num) & (QM_DQRR_SIZE - 1);
-+ qm_cl_out(DQRR_CI, dqrr->ci);
-+}
-+
-+static inline void qm_dqrr_cdc_consume_n(struct qm_portal *portal, u16 bitmask)
-+{
-+ __maybe_unused register struct qm_dqrr *dqrr = &portal->dqrr;
-+ DPA_ASSERT(dqrr->cmode == qm_dqrr_cdc);
-+ qm_out(DQRR_DCAP, (1 << 8) | /* DQRR_DCAP::S */
-+ ((u32)bitmask << 16)); /* DQRR_DCAP::DCAP_CI */
-+ dqrr->ci = qm_in(DQRR_CI_CINH) & (QM_DQRR_SIZE - 1);
-+ dqrr->fill = qm_cyc_diff(QM_DQRR_SIZE, dqrr->ci, dqrr->pi);
-+}
-+
-+static inline int qm_dqrr_init(struct qm_portal *portal,
-+ const struct qm_portal_config *config,
-+ enum qm_dqrr_dmode dmode,
-+ __maybe_unused enum qm_dqrr_pmode pmode,
-+ enum qm_dqrr_cmode cmode, u8 max_fill)
-+{
-+ register struct qm_dqrr *dqrr = &portal->dqrr;
-+ u32 cfg;
-+
-+ /* Make sure the DQRR will be idle when we enable */
-+ qm_out(DQRR_SDQCR, 0);
-+ qm_out(DQRR_VDQCR, 0);
-+ qm_out(DQRR_PDQCR, 0);
-+ dqrr->ring = portal->addr.addr_ce + QM_CL_DQRR;
-+ dqrr->pi = qm_in(DQRR_PI_CINH) & (QM_DQRR_SIZE - 1);
-+ dqrr->ci = qm_in(DQRR_CI_CINH) & (QM_DQRR_SIZE - 1);
-+ dqrr->cursor = dqrr->ring + dqrr->ci;
-+ dqrr->fill = qm_cyc_diff(QM_DQRR_SIZE, dqrr->ci, dqrr->pi);
-+ dqrr->vbit = (qm_in(DQRR_PI_CINH) & QM_DQRR_SIZE) ?
-+ QM_DQRR_VERB_VBIT : 0;
-+ dqrr->ithresh = qm_in(DQRR_ITR);
-+
-+ /* Free up pending DQRR entries if any as per current DCM */
-+ if (dqrr->fill) {
-+ enum qm_dqrr_cmode dcm = (qm_in(CFG) >> 16) & 3;
-+
-+#ifdef CONFIG_FSL_DPA_CHECKING
-+ dqrr->cmode = dcm;
-+#endif
-+ switch (dcm) {
-+ case qm_dqrr_cci:
-+ qm_dqrr_cci_consume(portal, dqrr->fill);
-+ break;
-+ case qm_dqrr_cce:
-+ qm_dqrr_cce_consume(portal, dqrr->fill);
-+ break;
-+ case qm_dqrr_cdc:
-+ qm_dqrr_cdc_consume_n(portal, (QM_DQRR_SIZE - 1));
-+ break;
-+ default:
-+ DPA_ASSERT(0);
-+ }
-+ }
-+
-+#ifdef CONFIG_FSL_DPA_CHECKING
-+ dqrr->dmode = dmode;
-+ dqrr->pmode = pmode;
-+ dqrr->cmode = cmode;
-+#endif
-+ /* Invalidate every ring entry before beginning */
-+ for (cfg = 0; cfg < QM_DQRR_SIZE; cfg++)
-+ dcbi(qm_cl(dqrr->ring, cfg));
-+ cfg = (qm_in(CFG) & 0xff000f00) |
-+ ((max_fill & (QM_DQRR_SIZE - 1)) << 20) | /* DQRR_MF */
-+ ((dmode & 1) << 18) | /* DP */
-+ ((cmode & 3) << 16) | /* DCM */
-+ 0xa0 | /* RE+SE */
-+ (0 ? 0x40 : 0) | /* Ignore RP */
-+ (0 ? 0x10 : 0); /* Ignore SP */
-+ qm_out(CFG, cfg);
-+ qm_dqrr_set_maxfill(portal, max_fill);
-+ return 0;
-+}
-+
-+static inline void qm_dqrr_finish(struct qm_portal *portal)
-+{
-+ __maybe_unused register struct qm_dqrr *dqrr = &portal->dqrr;
-+#ifdef CONFIG_FSL_DPA_CHECKING
-+ if ((dqrr->cmode != qm_dqrr_cdc) &&
-+ (dqrr->ci != DQRR_PTR2IDX(dqrr->cursor)))
-+ pr_crit("Ignoring completed DQRR entries\n");
-+#endif
-+}
-+
-+static inline const struct qm_dqrr_entry *qm_dqrr_current(
-+ struct qm_portal *portal)
-+{
-+ register struct qm_dqrr *dqrr = &portal->dqrr;
-+ if (!dqrr->fill)
-+ return NULL;
-+ return dqrr->cursor;
-+}
-+
-+static inline u8 qm_dqrr_cursor(struct qm_portal *portal)
-+{
-+ register struct qm_dqrr *dqrr = &portal->dqrr;
-+ return DQRR_PTR2IDX(dqrr->cursor);
-+}
-+
-+static inline u8 qm_dqrr_next(struct qm_portal *portal)
-+{
-+ register struct qm_dqrr *dqrr = &portal->dqrr;
-+ DPA_ASSERT(dqrr->fill);
-+ dqrr->cursor = DQRR_INC(dqrr->cursor);
-+ return --dqrr->fill;
-+}
-+
-+static inline u8 qm_dqrr_pci_update(struct qm_portal *portal)
-+{
-+ register struct qm_dqrr *dqrr = &portal->dqrr;
-+ u8 diff, old_pi = dqrr->pi;
-+ DPA_ASSERT(dqrr->pmode == qm_dqrr_pci);
-+ dqrr->pi = qm_in(DQRR_PI_CINH) & (QM_DQRR_SIZE - 1);
-+ diff = qm_cyc_diff(QM_DQRR_SIZE, old_pi, dqrr->pi);
-+ dqrr->fill += diff;
-+ return diff;
-+}
-+
-+static inline void qm_dqrr_pce_prefetch(struct qm_portal *portal)
-+{
-+ __maybe_unused register struct qm_dqrr *dqrr = &portal->dqrr;
-+ DPA_ASSERT(dqrr->pmode == qm_dqrr_pce);
-+ qm_cl_invalidate(DQRR_PI);
-+ qm_cl_touch_ro(DQRR_PI);
-+}
-+
-+static inline u8 qm_dqrr_pce_update(struct qm_portal *portal)
-+{
-+ register struct qm_dqrr *dqrr = &portal->dqrr;
-+ u8 diff, old_pi = dqrr->pi;
-+ DPA_ASSERT(dqrr->pmode == qm_dqrr_pce);
-+ dqrr->pi = qm_cl_in(DQRR_PI) & (QM_DQRR_SIZE - 1);
-+ diff = qm_cyc_diff(QM_DQRR_SIZE, old_pi, dqrr->pi);
-+ dqrr->fill += diff;
-+ return diff;
-+}
-+
-+static inline void qm_dqrr_pvb_update(struct qm_portal *portal)
-+{
-+ register struct qm_dqrr *dqrr = &portal->dqrr;
-+ const struct qm_dqrr_entry *res = qm_cl(dqrr->ring, dqrr->pi);
-+ DPA_ASSERT(dqrr->pmode == qm_dqrr_pvb);
-+#if (defined CONFIG_PPC || defined CONFIG_PPC64) && !defined CONFIG_FSL_PAMU
-+ /*
-+ * On PowerPC platforms if PAMU is not available we need to
-+ * manually invalidate the cache. When PAMU is available the
-+ * cache is updated by stashing operations generated by QMan
-+ */
-+ dcbi(res);
-+ dcbt_ro(res);
-+#endif
-+
-+ /* when accessing 'verb', use __raw_readb() to ensure that compiler
-+ * inlining doesn't try to optimise out "excess reads". */
-+ if ((__raw_readb(&res->verb) & QM_DQRR_VERB_VBIT) == dqrr->vbit) {
-+ dqrr->pi = (dqrr->pi + 1) & (QM_DQRR_SIZE - 1);
-+ if (!dqrr->pi)
-+ dqrr->vbit ^= QM_DQRR_VERB_VBIT;
-+ dqrr->fill++;
-+ }
-+}
-+
-+
-+static inline void qm_dqrr_cci_consume_to_current(struct qm_portal *portal)
-+{
-+ register struct qm_dqrr *dqrr = &portal->dqrr;
-+ DPA_ASSERT(dqrr->cmode == qm_dqrr_cci);
-+ dqrr->ci = DQRR_PTR2IDX(dqrr->cursor);
-+ qm_out(DQRR_CI_CINH, dqrr->ci);
-+}
-+
-+static inline void qm_dqrr_cce_prefetch(struct qm_portal *portal)
-+{
-+ __maybe_unused register struct qm_dqrr *dqrr = &portal->dqrr;
-+ DPA_ASSERT(dqrr->cmode == qm_dqrr_cce);
-+ qm_cl_invalidate(DQRR_CI);
-+ qm_cl_touch_rw(DQRR_CI);
-+}
-+
-+static inline void qm_dqrr_cce_consume_to_current(struct qm_portal *portal)
-+{
-+ register struct qm_dqrr *dqrr = &portal->dqrr;
-+ DPA_ASSERT(dqrr->cmode == qm_dqrr_cce);
-+ dqrr->ci = DQRR_PTR2IDX(dqrr->cursor);
-+ qm_cl_out(DQRR_CI, dqrr->ci);
-+}
-+
-+static inline void qm_dqrr_cdc_consume_1(struct qm_portal *portal, u8 idx,
-+ int park)
-+{
-+ __maybe_unused register struct qm_dqrr *dqrr = &portal->dqrr;
-+ DPA_ASSERT(dqrr->cmode == qm_dqrr_cdc);
-+ DPA_ASSERT(idx < QM_DQRR_SIZE);
-+ qm_out(DQRR_DCAP, (0 << 8) | /* S */
-+ ((park ? 1 : 0) << 6) | /* PK */
-+ idx); /* DCAP_CI */
-+}
-+
-+static inline void qm_dqrr_cdc_consume_1ptr(struct qm_portal *portal,
-+ const struct qm_dqrr_entry *dq,
-+ int park)
-+{
-+ __maybe_unused register struct qm_dqrr *dqrr = &portal->dqrr;
-+ u8 idx = DQRR_PTR2IDX(dq);
-+ DPA_ASSERT(dqrr->cmode == qm_dqrr_cdc);
-+ DPA_ASSERT((dqrr->ring + idx) == dq);
-+ DPA_ASSERT(idx < QM_DQRR_SIZE);
-+ qm_out(DQRR_DCAP, (0 << 8) | /* DQRR_DCAP::S */
-+ ((park ? 1 : 0) << 6) | /* DQRR_DCAP::PK */
-+ idx); /* DQRR_DCAP::DCAP_CI */
-+}
-+
-+static inline u8 qm_dqrr_cdc_cci(struct qm_portal *portal)
-+{
-+ __maybe_unused register struct qm_dqrr *dqrr = &portal->dqrr;
-+ DPA_ASSERT(dqrr->cmode == qm_dqrr_cdc);
-+ return qm_in(DQRR_CI_CINH) & (QM_DQRR_SIZE - 1);
-+}
-+
-+static inline void qm_dqrr_cdc_cce_prefetch(struct qm_portal *portal)
-+{
-+ __maybe_unused register struct qm_dqrr *dqrr = &portal->dqrr;
-+ DPA_ASSERT(dqrr->cmode == qm_dqrr_cdc);
-+ qm_cl_invalidate(DQRR_CI);
-+ qm_cl_touch_ro(DQRR_CI);
-+}
-+
-+static inline u8 qm_dqrr_cdc_cce(struct qm_portal *portal)
-+{
-+ __maybe_unused register struct qm_dqrr *dqrr = &portal->dqrr;
-+ DPA_ASSERT(dqrr->cmode == qm_dqrr_cdc);
-+ return qm_cl_in(DQRR_CI) & (QM_DQRR_SIZE - 1);
-+}
-+
-+static inline u8 qm_dqrr_get_ci(struct qm_portal *portal)
-+{
-+ register struct qm_dqrr *dqrr = &portal->dqrr;
-+ DPA_ASSERT(dqrr->cmode != qm_dqrr_cdc);
-+ return dqrr->ci;
-+}
-+
-+static inline void qm_dqrr_park(struct qm_portal *portal, u8 idx)
-+{
-+ __maybe_unused register struct qm_dqrr *dqrr = &portal->dqrr;
-+ DPA_ASSERT(dqrr->cmode != qm_dqrr_cdc);
-+ qm_out(DQRR_DCAP, (0 << 8) | /* S */
-+ (1 << 6) | /* PK */
-+ (idx & (QM_DQRR_SIZE - 1))); /* DCAP_CI */
-+}
-+
-+static inline void qm_dqrr_park_current(struct qm_portal *portal)
-+{
-+ register struct qm_dqrr *dqrr = &portal->dqrr;
-+ DPA_ASSERT(dqrr->cmode != qm_dqrr_cdc);
-+ qm_out(DQRR_DCAP, (0 << 8) | /* S */
-+ (1 << 6) | /* PK */
-+ DQRR_PTR2IDX(dqrr->cursor)); /* DCAP_CI */
-+}
-+
-+static inline void qm_dqrr_sdqcr_set(struct qm_portal *portal, u32 sdqcr)
-+{
-+ qm_out(DQRR_SDQCR, sdqcr);
-+}
-+
-+static inline u32 qm_dqrr_sdqcr_get(struct qm_portal *portal)
-+{
-+ return qm_in(DQRR_SDQCR);
-+}
-+
-+static inline void qm_dqrr_vdqcr_set(struct qm_portal *portal, u32 vdqcr)
-+{
-+ qm_out(DQRR_VDQCR, vdqcr);
-+}
-+
-+static inline u32 qm_dqrr_vdqcr_get(struct qm_portal *portal)
-+{
-+ return qm_in(DQRR_VDQCR);
-+}
-+
-+static inline void qm_dqrr_pdqcr_set(struct qm_portal *portal, u32 pdqcr)
-+{
-+ qm_out(DQRR_PDQCR, pdqcr);
-+}
-+
-+static inline u32 qm_dqrr_pdqcr_get(struct qm_portal *portal)
-+{
-+ return qm_in(DQRR_PDQCR);
-+}
-+
-+static inline u8 qm_dqrr_get_ithresh(struct qm_portal *portal)
-+{
-+ register struct qm_dqrr *dqrr = &portal->dqrr;
-+ return dqrr->ithresh;
-+}
-+
-+static inline void qm_dqrr_set_ithresh(struct qm_portal *portal, u8 ithresh)
-+{
-+ qm_out(DQRR_ITR, ithresh);
-+}
-+
-+static inline u8 qm_dqrr_get_maxfill(struct qm_portal *portal)
-+{
-+ return (qm_in(CFG) & 0x00f00000) >> 20;
-+}
-+
-+
-+/* -------------- */
-+/* --- MR API --- */
-+
-+#define MR_CARRYCLEAR(p) \
-+ (void *)((unsigned long)(p) & (~(unsigned long)(QM_MR_SIZE << 6)))
-+
-+static inline u8 MR_PTR2IDX(const struct qm_mr_entry *e)
-+{
-+ return ((uintptr_t)e >> 6) & (QM_MR_SIZE - 1);
-+}
-+
-+static inline const struct qm_mr_entry *MR_INC(const struct qm_mr_entry *e)
-+{
-+ return MR_CARRYCLEAR(e + 1);
-+}
-+
-+static inline int qm_mr_init(struct qm_portal *portal, enum qm_mr_pmode pmode,
-+ enum qm_mr_cmode cmode)
-+{
-+ register struct qm_mr *mr = &portal->mr;
-+ u32 cfg;
-+
-+ mr->ring = portal->addr.addr_ce + QM_CL_MR;
-+ mr->pi = qm_in(MR_PI_CINH) & (QM_MR_SIZE - 1);
-+ mr->ci = qm_in(MR_CI_CINH) & (QM_MR_SIZE - 1);
-+ mr->cursor = mr->ring + mr->ci;
-+ mr->fill = qm_cyc_diff(QM_MR_SIZE, mr->ci, mr->pi);
-+ mr->vbit = (qm_in(MR_PI_CINH) & QM_MR_SIZE) ? QM_MR_VERB_VBIT : 0;
-+ mr->ithresh = qm_in(MR_ITR);
-+#ifdef CONFIG_FSL_DPA_CHECKING
-+ mr->pmode = pmode;
-+ mr->cmode = cmode;
-+#endif
-+ cfg = (qm_in(CFG) & 0xfffff0ff) |
-+ ((cmode & 1) << 8); /* QCSP_CFG:MM */
-+ qm_out(CFG, cfg);
-+ return 0;
-+}
-+
-+static inline void qm_mr_finish(struct qm_portal *portal)
-+{
-+ register struct qm_mr *mr = &portal->mr;
-+ if (mr->ci != MR_PTR2IDX(mr->cursor))
-+ pr_crit("Ignoring completed MR entries\n");
-+}
-+
-+static inline const struct qm_mr_entry *qm_mr_current(struct qm_portal *portal)
-+{
-+ register struct qm_mr *mr = &portal->mr;
-+ if (!mr->fill)
-+ return NULL;
-+ return mr->cursor;
-+}
-+
-+static inline u8 qm_mr_cursor(struct qm_portal *portal)
-+{
-+ register struct qm_mr *mr = &portal->mr;
-+ return MR_PTR2IDX(mr->cursor);
-+}
-+
-+static inline u8 qm_mr_next(struct qm_portal *portal)
-+{
-+ register struct qm_mr *mr = &portal->mr;
-+ DPA_ASSERT(mr->fill);
-+ mr->cursor = MR_INC(mr->cursor);
-+ return --mr->fill;
-+}
-+
-+static inline u8 qm_mr_pci_update(struct qm_portal *portal)
-+{
-+ register struct qm_mr *mr = &portal->mr;
-+ u8 diff, old_pi = mr->pi;
-+ DPA_ASSERT(mr->pmode == qm_mr_pci);
-+ mr->pi = qm_in(MR_PI_CINH);
-+ diff = qm_cyc_diff(QM_MR_SIZE, old_pi, mr->pi);
-+ mr->fill += diff;
-+ return diff;
-+}
-+
-+static inline void qm_mr_pce_prefetch(struct qm_portal *portal)
-+{
-+ __maybe_unused register struct qm_mr *mr = &portal->mr;
-+ DPA_ASSERT(mr->pmode == qm_mr_pce);
-+ qm_cl_invalidate(MR_PI);
-+ qm_cl_touch_ro(MR_PI);
-+}
-+
-+static inline u8 qm_mr_pce_update(struct qm_portal *portal)
-+{
-+ register struct qm_mr *mr = &portal->mr;
-+ u8 diff, old_pi = mr->pi;
-+ DPA_ASSERT(mr->pmode == qm_mr_pce);
-+ mr->pi = qm_cl_in(MR_PI) & (QM_MR_SIZE - 1);
-+ diff = qm_cyc_diff(QM_MR_SIZE, old_pi, mr->pi);
-+ mr->fill += diff;
-+ return diff;
-+}
-+
-+static inline void qm_mr_pvb_update(struct qm_portal *portal)
-+{
-+ register struct qm_mr *mr = &portal->mr;
-+ const struct qm_mr_entry *res = qm_cl(mr->ring, mr->pi);
-+ DPA_ASSERT(mr->pmode == qm_mr_pvb);
-+ /* when accessing 'verb', use __raw_readb() to ensure that compiler
-+ * inlining doesn't try to optimise out "excess reads". */
-+ if ((__raw_readb(&res->verb) & QM_MR_VERB_VBIT) == mr->vbit) {
-+ mr->pi = (mr->pi + 1) & (QM_MR_SIZE - 1);
-+ if (!mr->pi)
-+ mr->vbit ^= QM_MR_VERB_VBIT;
-+ mr->fill++;
-+ res = MR_INC(res);
-+ }
-+ dcbit_ro(res);
-+}
-+
-+static inline void qm_mr_cci_consume(struct qm_portal *portal, u8 num)
-+{
-+ register struct qm_mr *mr = &portal->mr;
-+ DPA_ASSERT(mr->cmode == qm_mr_cci);
-+ mr->ci = (mr->ci + num) & (QM_MR_SIZE - 1);
-+ qm_out(MR_CI_CINH, mr->ci);
-+}
-+
-+static inline void qm_mr_cci_consume_to_current(struct qm_portal *portal)
-+{
-+ register struct qm_mr *mr = &portal->mr;
-+ DPA_ASSERT(mr->cmode == qm_mr_cci);
-+ mr->ci = MR_PTR2IDX(mr->cursor);
-+ qm_out(MR_CI_CINH, mr->ci);
-+}
-+
-+static inline void qm_mr_cce_prefetch(struct qm_portal *portal)
-+{
-+ __maybe_unused register struct qm_mr *mr = &portal->mr;
-+ DPA_ASSERT(mr->cmode == qm_mr_cce);
-+ qm_cl_invalidate(MR_CI);
-+ qm_cl_touch_rw(MR_CI);
-+}
-+
-+static inline void qm_mr_cce_consume(struct qm_portal *portal, u8 num)
-+{
-+ register struct qm_mr *mr = &portal->mr;
-+ DPA_ASSERT(mr->cmode == qm_mr_cce);
-+ mr->ci = (mr->ci + num) & (QM_MR_SIZE - 1);
-+ qm_cl_out(MR_CI, mr->ci);
-+}
-+
-+static inline void qm_mr_cce_consume_to_current(struct qm_portal *portal)
-+{
-+ register struct qm_mr *mr = &portal->mr;
-+ DPA_ASSERT(mr->cmode == qm_mr_cce);
-+ mr->ci = MR_PTR2IDX(mr->cursor);
-+ qm_cl_out(MR_CI, mr->ci);
-+}
-+
-+static inline u8 qm_mr_get_ci(struct qm_portal *portal)
-+{
-+ register struct qm_mr *mr = &portal->mr;
-+ return mr->ci;
-+}
-+
-+static inline u8 qm_mr_get_ithresh(struct qm_portal *portal)
-+{
-+ register struct qm_mr *mr = &portal->mr;
-+ return mr->ithresh;
-+}
-+
-+static inline void qm_mr_set_ithresh(struct qm_portal *portal, u8 ithresh)
-+{
-+ qm_out(MR_ITR, ithresh);
-+}
-+
-+
-+/* ------------------------------ */
-+/* --- Management command API --- */
-+
-+static inline int qm_mc_init(struct qm_portal *portal)
-+{
-+ u8 rr0, rr1;
-+ register struct qm_mc *mc = &portal->mc;
-+
-+ mc->cr = portal->addr.addr_ce + QM_CL_CR;
-+ mc->rr = portal->addr.addr_ce + QM_CL_RR0;
-+
-+ /*
-+ * The expected valid bit polarity for the next CR command is 0
-+ * if RR1 contains a valid response, and is 1 if RR0 contains a
-+ * valid response. If both RR contain all 0, this indicates either
-+ * that no command has been executed since reset (in which case the
-+ * expected valid bit polarity is 1)
-+ */
-+ rr0 = __raw_readb(&mc->rr->verb);
-+ rr1 = __raw_readb(&(mc->rr+1)->verb);
-+ if ((rr0 == 0 && rr1 == 0) || rr0 != 0)
-+ mc->rridx = 1;
-+ else
-+ mc->rridx = 0;
-+
-+ mc->vbit = mc->rridx ? QM_MCC_VERB_VBIT : 0;
-+#ifdef CONFIG_FSL_DPA_CHECKING
-+ mc->state = qman_mc_idle;
-+#endif
-+ return 0;
-+}
-+
-+static inline void qm_mc_finish(struct qm_portal *portal)
-+{
-+ __maybe_unused register struct qm_mc *mc = &portal->mc;
-+ DPA_ASSERT(mc->state == qman_mc_idle);
-+#ifdef CONFIG_FSL_DPA_CHECKING
-+ if (mc->state != qman_mc_idle)
-+ pr_crit("Losing incomplete MC command\n");
-+#endif
-+}
-+
-+static inline struct qm_mc_command *qm_mc_start(struct qm_portal *portal)
-+{
-+ register struct qm_mc *mc = &portal->mc;
-+ DPA_ASSERT(mc->state == qman_mc_idle);
-+#ifdef CONFIG_FSL_DPA_CHECKING
-+ mc->state = qman_mc_user;
-+#endif
-+#if defined(CONFIG_PPC32) || defined(CONFIG_PPC64)
-+ dcbz_64(mc->cr);
-+#endif
-+ return mc->cr;
-+}
-+
-+static inline void qm_mc_abort(struct qm_portal *portal)
-+{
-+ __maybe_unused register struct qm_mc *mc = &portal->mc;
-+ DPA_ASSERT(mc->state == qman_mc_user);
-+#ifdef CONFIG_FSL_DPA_CHECKING
-+ mc->state = qman_mc_idle;
-+#endif
-+}
-+
-+static inline void qm_mc_commit(struct qm_portal *portal, u8 myverb)
-+{
-+ register struct qm_mc *mc = &portal->mc;
-+ struct qm_mc_result *rr = mc->rr + mc->rridx;
-+ DPA_ASSERT(mc->state == qman_mc_user);
-+ lwsync();
-+ mc->cr->__dont_write_directly__verb = myverb | mc->vbit;
-+ dcbf(mc->cr);
-+ dcbit_ro(rr);
-+#ifdef CONFIG_FSL_DPA_CHECKING
-+ mc->state = qman_mc_hw;
-+#endif
-+}
-+
-+static inline struct qm_mc_result *qm_mc_result(struct qm_portal *portal)
-+{
-+ register struct qm_mc *mc = &portal->mc;
-+ struct qm_mc_result *rr = mc->rr + mc->rridx;
-+ DPA_ASSERT(mc->state == qman_mc_hw);
-+ /* The inactive response register's verb byte always returns zero until
-+ * its command is submitted and completed. This includes the valid-bit,
-+ * in case you were wondering... */
-+ if (!__raw_readb(&rr->verb)) {
-+ dcbit_ro(rr);
-+ return NULL;
-+ }
-+ mc->rridx ^= 1;
-+ mc->vbit ^= QM_MCC_VERB_VBIT;
-+#ifdef CONFIG_FSL_DPA_CHECKING
-+ mc->state = qman_mc_idle;
-+#endif
-+ return rr;
-+}
-+
-+
-+/* ------------------------------------- */
-+/* --- Portal interrupt register API --- */
-+
-+static inline int qm_isr_init(__always_unused struct qm_portal *portal)
-+{
-+ return 0;
-+}
-+
-+static inline void qm_isr_finish(__always_unused struct qm_portal *portal)
-+{
-+}
-+
-+static inline void qm_isr_set_iperiod(struct qm_portal *portal, u16 iperiod)
-+{
-+ qm_out(ITPR, iperiod);
-+}
-+
-+static inline u32 __qm_isr_read(struct qm_portal *portal, enum qm_isr_reg n)
-+{
-+#if defined(CONFIG_ARM) || defined(CONFIG_ARM64)
-+ return __qm_in(&portal->addr, QM_REG_ISR + (n << 6));
-+#else
-+ return __qm_in(&portal->addr, QM_REG_ISR + (n << 2));
-+#endif
-+}
-+
-+static inline void __qm_isr_write(struct qm_portal *portal, enum qm_isr_reg n,
-+ u32 val)
-+{
-+#if defined(CONFIG_ARM) || defined(CONFIG_ARM64)
-+ __qm_out(&portal->addr, QM_REG_ISR + (n << 6), val);
-+#else
-+ __qm_out(&portal->addr, QM_REG_ISR + (n << 2), val);
-+#endif
-+}
-+
-+/* Cleanup FQs */
-+static inline int qm_shutdown_fq(struct qm_portal **portal, int portal_count,
-+ u32 fqid)
-+{
-+
-+ struct qm_mc_command *mcc;
-+ struct qm_mc_result *mcr;
-+ u8 state;
-+ int orl_empty, fq_empty, i, drain = 0;
-+ u32 result;
-+ u32 channel, wq;
-+ u16 dest_wq;
-+
-+ /* Determine the state of the FQID */
-+ mcc = qm_mc_start(portal[0]);
-+ mcc->queryfq_np.fqid = cpu_to_be32(fqid);
-+ qm_mc_commit(portal[0], QM_MCC_VERB_QUERYFQ_NP);
-+ while (!(mcr = qm_mc_result(portal[0])))
-+ cpu_relax();
-+ DPA_ASSERT((mcr->verb & QM_MCR_VERB_MASK) == QM_MCR_VERB_QUERYFQ_NP);
-+ state = mcr->queryfq_np.state & QM_MCR_NP_STATE_MASK;
-+ if (state == QM_MCR_NP_STATE_OOS)
-+ return 0; /* Already OOS, no need to do anymore checks */
-+
-+ /* Query which channel the FQ is using */
-+ mcc = qm_mc_start(portal[0]);
-+ mcc->queryfq.fqid = cpu_to_be32(fqid);
-+ qm_mc_commit(portal[0], QM_MCC_VERB_QUERYFQ);
-+ while (!(mcr = qm_mc_result(portal[0])))
-+ cpu_relax();
-+ DPA_ASSERT((mcr->verb & QM_MCR_VERB_MASK) == QM_MCR_VERB_QUERYFQ);
-+
-+ /* Need to store these since the MCR gets reused */
-+ dest_wq = be16_to_cpu(mcr->queryfq.fqd.dest_wq);
-+ wq = dest_wq & 0x7;
-+ channel = dest_wq>>3;
-+
-+ switch (state) {
-+ case QM_MCR_NP_STATE_TEN_SCHED:
-+ case QM_MCR_NP_STATE_TRU_SCHED:
-+ case QM_MCR_NP_STATE_ACTIVE:
-+ case QM_MCR_NP_STATE_PARKED:
-+ orl_empty = 0;
-+ mcc = qm_mc_start(portal[0]);
-+ mcc->alterfq.fqid = cpu_to_be32(fqid);
-+ qm_mc_commit(portal[0], QM_MCC_VERB_ALTER_RETIRE);
-+ while (!(mcr = qm_mc_result(portal[0])))
-+ cpu_relax();
-+ DPA_ASSERT((mcr->verb & QM_MCR_VERB_MASK) ==
-+ QM_MCR_VERB_ALTER_RETIRE);
-+ result = mcr->result; /* Make a copy as we reuse MCR below */
-+
-+ if (result == QM_MCR_RESULT_PENDING) {
-+ /* Need to wait for the FQRN in the message ring, which
-+ will only occur once the FQ has been drained. In
-+ order for the FQ to drain the portal needs to be set
-+ to dequeue from the channel the FQ is scheduled on */
-+ const struct qm_mr_entry *msg;
-+ const struct qm_dqrr_entry *dqrr = NULL;
-+ int found_fqrn = 0;
-+ u16 dequeue_wq = 0;
-+
-+ /* Flag that we need to drain FQ */
-+ drain = 1;
-+
-+ if (channel >= qm_channel_pool1 &&
-+ channel < (qm_channel_pool1 + 15)) {
-+ /* Pool channel, enable the bit in the portal */
-+ dequeue_wq = (channel -
-+ qm_channel_pool1 + 1)<<4 | wq;
-+ } else if (channel < qm_channel_pool1) {
-+ /* Dedicated channel */
-+ dequeue_wq = wq;
-+ } else {
-+ pr_info("Cannot recover FQ 0x%x, it is "
-+ "scheduled on channel 0x%x",
-+ fqid, channel);
-+ return -EBUSY;
-+ }
-+ /* Set the sdqcr to drain this channel */
-+ if (channel < qm_channel_pool1)
-+ for (i = 0; i < portal_count; i++)
-+ qm_dqrr_sdqcr_set(portal[i],
-+ QM_SDQCR_TYPE_ACTIVE |
-+ QM_SDQCR_CHANNELS_DEDICATED);
-+ else
-+ for (i = 0; i < portal_count; i++)
-+ qm_dqrr_sdqcr_set(
-+ portal[i],
-+ QM_SDQCR_TYPE_ACTIVE |
-+ QM_SDQCR_CHANNELS_POOL_CONV
-+ (channel));
-+ while (!found_fqrn) {
-+ /* Keep draining DQRR while checking the MR*/
-+ for (i = 0; i < portal_count; i++) {
-+ qm_dqrr_pvb_update(portal[i]);
-+ dqrr = qm_dqrr_current(portal[i]);
-+ while (dqrr) {
-+ qm_dqrr_cdc_consume_1ptr(
-+ portal[i], dqrr, 0);
-+ qm_dqrr_pvb_update(portal[i]);
-+ qm_dqrr_next(portal[i]);
-+ dqrr = qm_dqrr_current(
-+ portal[i]);
-+ }
-+ /* Process message ring too */
-+ qm_mr_pvb_update(portal[i]);
-+ msg = qm_mr_current(portal[i]);
-+ while (msg) {
-+ if ((msg->verb &
-+ QM_MR_VERB_TYPE_MASK)
-+ == QM_MR_VERB_FQRN)
-+ found_fqrn = 1;
-+ qm_mr_next(portal[i]);
-+ qm_mr_cci_consume_to_current(
-+ portal[i]);
-+ qm_mr_pvb_update(portal[i]);
-+ msg = qm_mr_current(portal[i]);
-+ }
-+ cpu_relax();
-+ }
-+ }
-+ }
-+ if (result != QM_MCR_RESULT_OK &&
-+ result != QM_MCR_RESULT_PENDING) {
-+ /* error */
-+ pr_err("qman_retire_fq failed on FQ 0x%x, result=0x%x\n",
-+ fqid, result);
-+ return -1;
-+ }
-+ if (!(mcr->alterfq.fqs & QM_MCR_FQS_ORLPRESENT)) {
-+ /* ORL had no entries, no need to wait until the
-+ ERNs come in */
-+ orl_empty = 1;
-+ }
-+ /* Retirement succeeded, check to see if FQ needs
-+ to be drained */
-+ if (drain || mcr->alterfq.fqs & QM_MCR_FQS_NOTEMPTY) {
-+ /* FQ is Not Empty, drain using volatile DQ commands */
-+ fq_empty = 0;
-+ do {
-+ const struct qm_dqrr_entry *dqrr = NULL;
-+ u32 vdqcr = fqid | QM_VDQCR_NUMFRAMES_SET(3);
-+ qm_dqrr_vdqcr_set(portal[0], vdqcr);
-+
-+ /* Wait for a dequeue to occur */
-+ while (dqrr == NULL) {
-+ qm_dqrr_pvb_update(portal[0]);
-+ dqrr = qm_dqrr_current(portal[0]);
-+ if (!dqrr)
-+ cpu_relax();
-+ }
-+ /* Process the dequeues, making sure to
-+ empty the ring completely */
-+ while (dqrr) {
-+ if (be32_to_cpu(dqrr->fqid) == fqid &&
-+ dqrr->stat & QM_DQRR_STAT_FQ_EMPTY)
-+ fq_empty = 1;
-+ qm_dqrr_cdc_consume_1ptr(portal[0],
-+ dqrr, 0);
-+ qm_dqrr_pvb_update(portal[0]);
-+ qm_dqrr_next(portal[0]);
-+ dqrr = qm_dqrr_current(portal[0]);
-+ }
-+ } while (fq_empty == 0);
-+ }
-+ for (i = 0; i < portal_count; i++)
-+ qm_dqrr_sdqcr_set(portal[i], 0);
-+
-+ /* Wait for the ORL to have been completely drained */
-+ while (orl_empty == 0) {
-+ const struct qm_mr_entry *msg;
-+ qm_mr_pvb_update(portal[0]);
-+ msg = qm_mr_current(portal[0]);
-+ while (msg) {
-+ if ((msg->verb & QM_MR_VERB_TYPE_MASK) ==
-+ QM_MR_VERB_FQRL)
-+ orl_empty = 1;
-+ qm_mr_next(portal[0]);
-+ qm_mr_cci_consume_to_current(portal[0]);
-+ qm_mr_pvb_update(portal[0]);
-+ msg = qm_mr_current(portal[0]);
-+ }
-+ cpu_relax();
-+ }
-+ mcc = qm_mc_start(portal[0]);
-+ mcc->alterfq.fqid = cpu_to_be32(fqid);
-+ qm_mc_commit(portal[0], QM_MCC_VERB_ALTER_OOS);
-+ while (!(mcr = qm_mc_result(portal[0])))
-+ cpu_relax();
-+ DPA_ASSERT((mcr->verb & QM_MCR_VERB_MASK) ==
-+ QM_MCR_VERB_ALTER_OOS);
-+ if (mcr->result != QM_MCR_RESULT_OK) {
-+ pr_err("OOS after drain Failed on FQID 0x%x, result 0x%x\n",
-+ fqid, mcr->result);
-+ return -1;
-+ }
-+ return 0;
-+ case QM_MCR_NP_STATE_RETIRED:
-+ /* Send OOS Command */
-+ mcc = qm_mc_start(portal[0]);
-+ mcc->alterfq.fqid = cpu_to_be32(fqid);
-+ qm_mc_commit(portal[0], QM_MCC_VERB_ALTER_OOS);
-+ while (!(mcr = qm_mc_result(portal[0])))
-+ cpu_relax();
-+ DPA_ASSERT((mcr->verb & QM_MCR_VERB_MASK) ==
-+ QM_MCR_VERB_ALTER_OOS);
-+ if (mcr->result) {
-+ pr_err("OOS Failed on FQID 0x%x\n", fqid);
-+ return -1;
-+ }
-+ return 0;
-+ }
-+ return -1;
-+}
---- /dev/null
-+++ b/drivers/staging/fsl_qbman/qman_private.h
-@@ -0,0 +1,398 @@
-+/* Copyright 2008-2012 Freescale Semiconductor, Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#include "dpa_sys.h"
-+#include <linux/fsl_qman.h>
-+#include <linux/iommu.h>
-+
-+#if defined(CONFIG_FSL_PAMU)
-+#include <asm/fsl_pamu_stash.h>
-+#endif
-+
-+#if !defined(CONFIG_FSL_QMAN_FQ_LOOKUP) && defined(CONFIG_PPC64)
-+#error "_PPC64 requires _FSL_QMAN_FQ_LOOKUP"
-+#endif
-+
-+#define QBMAN_ANY_PORTAL_IDX 0xffffffff
-+ /* ----------------- */
-+ /* Congestion Groups */
-+ /* ----------------- */
-+/* This wrapper represents a bit-array for the state of the 256 Qman congestion
-+ * groups. Is also used as a *mask* for congestion groups, eg. so we ignore
-+ * those that don't concern us. We harness the structure and accessor details
-+ * already used in the management command to query congestion groups. */
-+struct qman_cgrs {
-+ struct __qm_mcr_querycongestion q;
-+};
-+static inline void qman_cgrs_init(struct qman_cgrs *c)
-+{
-+ memset(c, 0, sizeof(*c));
-+}
-+static inline void qman_cgrs_fill(struct qman_cgrs *c)
-+{
-+ memset(c, 0xff, sizeof(*c));
-+}
-+static inline int qman_cgrs_get(struct qman_cgrs *c, int num)
-+{
-+ return QM_MCR_QUERYCONGESTION(&c->q, num);
-+}
-+static inline void qman_cgrs_set(struct qman_cgrs *c, int num)
-+{
-+ c->q.__state[__CGR_WORD(num)] |= (0x80000000 >> __CGR_SHIFT(num));
-+}
-+static inline void qman_cgrs_unset(struct qman_cgrs *c, int num)
-+{
-+ c->q.__state[__CGR_WORD(num)] &= ~(0x80000000 >> __CGR_SHIFT(num));
-+}
-+static inline int qman_cgrs_next(struct qman_cgrs *c, int num)
-+{
-+ while ((++num < __CGR_NUM) && !qman_cgrs_get(c, num))
-+ ;
-+ return num;
-+}
-+static inline void qman_cgrs_cp(struct qman_cgrs *dest,
-+ const struct qman_cgrs *src)
-+{
-+ *dest = *src;
-+}
-+static inline void qman_cgrs_and(struct qman_cgrs *dest,
-+ const struct qman_cgrs *a, const struct qman_cgrs *b)
-+{
-+ int ret;
-+ u32 *_d = dest->q.__state;
-+ const u32 *_a = a->q.__state;
-+ const u32 *_b = b->q.__state;
-+ for (ret = 0; ret < 8; ret++)
-+ *(_d++) = *(_a++) & *(_b++);
-+}
-+static inline void qman_cgrs_xor(struct qman_cgrs *dest,
-+ const struct qman_cgrs *a, const struct qman_cgrs *b)
-+{
-+ int ret;
-+ u32 *_d = dest->q.__state;
-+ const u32 *_a = a->q.__state;
-+ const u32 *_b = b->q.__state;
-+ for (ret = 0; ret < 8; ret++)
-+ *(_d++) = *(_a++) ^ *(_b++);
-+}
-+
-+ /* ----------------------- */
-+ /* CEETM Congestion Groups */
-+ /* ----------------------- */
-+/* This wrapper represents a bit-array for the state of the 512 Qman CEETM
-+ * congestion groups.
-+ */
-+struct qman_ccgrs {
-+ struct __qm_mcr_querycongestion q[2];
-+};
-+static inline void qman_ccgrs_init(struct qman_ccgrs *c)
-+{
-+ memset(c, 0, sizeof(*c));
-+}
-+static inline void qman_ccgrs_fill(struct qman_ccgrs *c)
-+{
-+ memset(c, 0xff, sizeof(*c));
-+}
-+static inline int qman_ccgrs_get(struct qman_ccgrs *c, int num)
-+{
-+ if (num < __CGR_NUM)
-+ return QM_MCR_QUERYCONGESTION(&c->q[0], num);
-+ else
-+ return QM_MCR_QUERYCONGESTION(&c->q[1], (num - __CGR_NUM));
-+}
-+static inline int qman_ccgrs_next(struct qman_ccgrs *c, int num)
-+{
-+ while ((++num < __CGR_NUM) && !qman_ccgrs_get(c, num))
-+ ;
-+ return num;
-+}
-+static inline void qman_ccgrs_cp(struct qman_ccgrs *dest,
-+ const struct qman_ccgrs *src)
-+{
-+ *dest = *src;
-+}
-+static inline void qman_ccgrs_and(struct qman_ccgrs *dest,
-+ const struct qman_ccgrs *a, const struct qman_ccgrs *b)
-+{
-+ int ret, i;
-+ u32 *_d;
-+ const u32 *_a, *_b;
-+ for (i = 0; i < 2; i++) {
-+ _d = dest->q[i].__state;
-+ _a = a->q[i].__state;
-+ _b = b->q[i].__state;
-+ for (ret = 0; ret < 8; ret++)
-+ *(_d++) = *(_a++) & *(_b++);
-+ }
-+}
-+static inline void qman_ccgrs_xor(struct qman_ccgrs *dest,
-+ const struct qman_ccgrs *a, const struct qman_ccgrs *b)
-+{
-+ int ret, i;
-+ u32 *_d;
-+ const u32 *_a, *_b;
-+ for (i = 0; i < 2; i++) {
-+ _d = dest->q[i].__state;
-+ _a = a->q[i].__state;
-+ _b = b->q[i].__state;
-+ for (ret = 0; ret < 8; ret++)
-+ *(_d++) = *(_a++) ^ *(_b++);
-+ }
-+}
-+
-+/* used by CCSR and portal interrupt code */
-+enum qm_isr_reg {
-+ qm_isr_status = 0,
-+ qm_isr_enable = 1,
-+ qm_isr_disable = 2,
-+ qm_isr_inhibit = 3
-+};
-+
-+struct qm_portal_config {
-+ /* Corenet portal addresses;
-+ * [0]==cache-enabled, [1]==cache-inhibited. */
-+ __iomem void *addr_virt[2];
-+ struct resource addr_phys[2];
-+ struct device dev;
-+ struct iommu_domain *iommu_domain;
-+ /* Allow these to be joined in lists */
-+ struct list_head list;
-+ /* User-visible portal configuration settings */
-+ struct qman_portal_config public_cfg;
-+ /* power management saved data */
-+ u32 saved_isdr;
-+};
-+
-+/* Revision info (for errata and feature handling) */
-+#define QMAN_REV11 0x0101
-+#define QMAN_REV12 0x0102
-+#define QMAN_REV20 0x0200
-+#define QMAN_REV30 0x0300
-+#define QMAN_REV31 0x0301
-+#define QMAN_REV32 0x0302
-+
-+/* QMan REV_2 register contains the Cfg option */
-+#define QMAN_REV_CFG_0 0x0
-+#define QMAN_REV_CFG_1 0x1
-+#define QMAN_REV_CFG_2 0x2
-+#define QMAN_REV_CFG_3 0x3
-+
-+extern u16 qman_ip_rev; /* 0 if uninitialised, otherwise QMAN_REVx */
-+extern u8 qman_ip_cfg;
-+extern u32 qman_clk;
-+extern u16 qman_portal_max;
-+
-+#ifdef CONFIG_FSL_QMAN_CONFIG
-+/* Hooks from qman_driver.c to qman_config.c */
-+int qman_init_ccsr(struct device_node *node);
-+void qman_liodn_fixup(u16 channel);
-+int qman_set_sdest(u16 channel, unsigned int cpu_idx);
-+size_t get_qman_fqd_size(void);
-+#else
-+static inline size_t get_qman_fqd_size(void)
-+{
-+ return (PAGE_SIZE << CONFIG_FSL_QMAN_FQD_SZ);
-+}
-+#endif
-+
-+int qm_set_wpm(int wpm);
-+int qm_get_wpm(int *wpm);
-+
-+/* Hooks from qman_driver.c in to qman_high.c */
-+struct qman_portal *qman_create_portal(
-+ struct qman_portal *portal,
-+ const struct qm_portal_config *config,
-+ const struct qman_cgrs *cgrs);
-+
-+struct qman_portal *qman_create_affine_portal(
-+ const struct qm_portal_config *config,
-+ const struct qman_cgrs *cgrs);
-+struct qman_portal *qman_create_affine_slave(struct qman_portal *redirect,
-+ int cpu);
-+const struct qm_portal_config *qman_destroy_affine_portal(void);
-+void qman_destroy_portal(struct qman_portal *qm);
-+
-+/* Hooks from fsl_usdpaa.c to qman_driver.c */
-+struct qm_portal_config *qm_get_unused_portal(void);
-+struct qm_portal_config *qm_get_unused_portal_idx(uint32_t idx);
-+
-+void qm_put_unused_portal(struct qm_portal_config *pcfg);
-+void qm_set_liodns(struct qm_portal_config *pcfg);
-+
-+/* This CGR feature is supported by h/w and required by unit-tests and the
-+ * debugfs hooks, so is implemented in the driver. However it allows an explicit
-+ * corruption of h/w fields by s/w that are usually incorruptible (because the
-+ * counters are usually maintained entirely within h/w). As such, we declare
-+ * this API internally. */
-+int qman_testwrite_cgr(struct qman_cgr *cgr, u64 i_bcnt,
-+ struct qm_mcr_cgrtestwrite *result);
-+
-+#ifdef CONFIG_FSL_QMAN_FQ_LOOKUP
-+/* If the fq object pointer is greater than the size of context_b field,
-+ * than a lookup table is required. */
-+int qman_setup_fq_lookup_table(size_t num_entries);
-+#endif
-+
-+
-+/*************************************************/
-+/* QMan s/w corenet portal, low-level i/face */
-+/*************************************************/
-+
-+/* Note: most functions are only used by the high-level interface, so are
-+ * inlined from qman_low.h. The stuff below is for use by other parts of the
-+ * driver. */
-+
-+/* For qm_dqrr_sdqcr_set(); Choose one SOURCE. Choose one COUNT. Choose one
-+ * dequeue TYPE. Choose TOKEN (8-bit).
-+ * If SOURCE == CHANNELS,
-+ * Choose CHANNELS_DEDICATED and/or CHANNELS_POOL(n).
-+ * You can choose DEDICATED_PRECEDENCE if the portal channel should have
-+ * priority.
-+ * If SOURCE == SPECIFICWQ,
-+ * Either select the work-queue ID with SPECIFICWQ_WQ(), or select the
-+ * channel (SPECIFICWQ_DEDICATED or SPECIFICWQ_POOL()) and specify the
-+ * work-queue priority (0-7) with SPECIFICWQ_WQ() - either way, you get the
-+ * same value.
-+ */
-+#define QM_SDQCR_SOURCE_CHANNELS 0x0
-+#define QM_SDQCR_SOURCE_SPECIFICWQ 0x40000000
-+#define QM_SDQCR_COUNT_EXACT1 0x0
-+#define QM_SDQCR_COUNT_UPTO3 0x20000000
-+#define QM_SDQCR_DEDICATED_PRECEDENCE 0x10000000
-+#define QM_SDQCR_TYPE_MASK 0x03000000
-+#define QM_SDQCR_TYPE_NULL 0x0
-+#define QM_SDQCR_TYPE_PRIO_QOS 0x01000000
-+#define QM_SDQCR_TYPE_ACTIVE_QOS 0x02000000
-+#define QM_SDQCR_TYPE_ACTIVE 0x03000000
-+#define QM_SDQCR_TOKEN_MASK 0x00ff0000
-+#define QM_SDQCR_TOKEN_SET(v) (((v) & 0xff) << 16)
-+#define QM_SDQCR_TOKEN_GET(v) (((v) >> 16) & 0xff)
-+#define QM_SDQCR_CHANNELS_DEDICATED 0x00008000
-+#define QM_SDQCR_SPECIFICWQ_MASK 0x000000f7
-+#define QM_SDQCR_SPECIFICWQ_DEDICATED 0x00000000
-+#define QM_SDQCR_SPECIFICWQ_POOL(n) ((n) << 4)
-+#define QM_SDQCR_SPECIFICWQ_WQ(n) (n)
-+
-+/* For qm_dqrr_vdqcr_set(): use FQID(n) to fill in the frame queue ID */
-+#define QM_VDQCR_FQID_MASK 0x00ffffff
-+#define QM_VDQCR_FQID(n) ((n) & QM_VDQCR_FQID_MASK)
-+
-+/* For qm_dqrr_pdqcr_set(); Choose one MODE. Choose one COUNT.
-+ * If MODE==SCHEDULED
-+ * Choose SCHEDULED_CHANNELS or SCHEDULED_SPECIFICWQ. Choose one dequeue TYPE.
-+ * If CHANNELS,
-+ * Choose CHANNELS_DEDICATED and/or CHANNELS_POOL() channels.
-+ * You can choose DEDICATED_PRECEDENCE if the portal channel should have
-+ * priority.
-+ * If SPECIFICWQ,
-+ * Either select the work-queue ID with SPECIFICWQ_WQ(), or select the
-+ * channel (SPECIFICWQ_DEDICATED or SPECIFICWQ_POOL()) and specify the
-+ * work-queue priority (0-7) with SPECIFICWQ_WQ() - either way, you get the
-+ * same value.
-+ * If MODE==UNSCHEDULED
-+ * Choose FQID().
-+ */
-+#define QM_PDQCR_MODE_SCHEDULED 0x0
-+#define QM_PDQCR_MODE_UNSCHEDULED 0x80000000
-+#define QM_PDQCR_SCHEDULED_CHANNELS 0x0
-+#define QM_PDQCR_SCHEDULED_SPECIFICWQ 0x40000000
-+#define QM_PDQCR_COUNT_EXACT1 0x0
-+#define QM_PDQCR_COUNT_UPTO3 0x20000000
-+#define QM_PDQCR_DEDICATED_PRECEDENCE 0x10000000
-+#define QM_PDQCR_TYPE_MASK 0x03000000
-+#define QM_PDQCR_TYPE_NULL 0x0
-+#define QM_PDQCR_TYPE_PRIO_QOS 0x01000000
-+#define QM_PDQCR_TYPE_ACTIVE_QOS 0x02000000
-+#define QM_PDQCR_TYPE_ACTIVE 0x03000000
-+#define QM_PDQCR_CHANNELS_DEDICATED 0x00008000
-+#define QM_PDQCR_CHANNELS_POOL(n) (0x00008000 >> (n))
-+#define QM_PDQCR_SPECIFICWQ_MASK 0x000000f7
-+#define QM_PDQCR_SPECIFICWQ_DEDICATED 0x00000000
-+#define QM_PDQCR_SPECIFICWQ_POOL(n) ((n) << 4)
-+#define QM_PDQCR_SPECIFICWQ_WQ(n) (n)
-+#define QM_PDQCR_FQID(n) ((n) & 0xffffff)
-+
-+/* Used by all portal interrupt registers except 'inhibit'
-+ * Channels with frame availability
-+ */
-+#define QM_PIRQ_DQAVAIL 0x0000ffff
-+
-+/* The DQAVAIL interrupt fields break down into these bits; */
-+#define QM_DQAVAIL_PORTAL 0x8000 /* Portal channel */
-+#define QM_DQAVAIL_POOL(n) (0x8000 >> (n)) /* Pool channel, n==[1..15] */
-+#define QM_DQAVAIL_MASK 0xffff
-+/* This mask contains all the "irqsource" bits visible to API users */
-+#define QM_PIRQ_VISIBLE (QM_PIRQ_SLOW | QM_PIRQ_DQRI)
-+
-+/* These are qm_<reg>_<verb>(). So for example, qm_disable_write() means "write
-+ * the disable register" rather than "disable the ability to write". */
-+#define qm_isr_status_read(qm) __qm_isr_read(qm, qm_isr_status)
-+#define qm_isr_status_clear(qm, m) __qm_isr_write(qm, qm_isr_status, m)
-+#define qm_isr_enable_read(qm) __qm_isr_read(qm, qm_isr_enable)
-+#define qm_isr_enable_write(qm, v) __qm_isr_write(qm, qm_isr_enable, v)
-+#define qm_isr_disable_read(qm) __qm_isr_read(qm, qm_isr_disable)
-+#define qm_isr_disable_write(qm, v) __qm_isr_write(qm, qm_isr_disable, v)
-+/* TODO: unfortunate name-clash here, reword? */
-+#define qm_isr_inhibit(qm) __qm_isr_write(qm, qm_isr_inhibit, 1)
-+#define qm_isr_uninhibit(qm) __qm_isr_write(qm, qm_isr_inhibit, 0)
-+
-+#ifdef CONFIG_FSL_QMAN_CONFIG
-+int qman_have_ccsr(void);
-+#else
-+#define qman_have_ccsr 0
-+#endif
-+
-+__init int qman_init(void);
-+__init int qman_resource_init(void);
-+
-+/* CEETM related */
-+#define QMAN_CEETM_MAX 2
-+extern u8 num_ceetms;
-+extern struct qm_ceetm qman_ceetms[QMAN_CEETM_MAX];
-+int qman_sp_enable_ceetm_mode(enum qm_dc_portal portal, u16 sub_portal);
-+int qman_sp_disable_ceetm_mode(enum qm_dc_portal portal, u16 sub_portal);
-+int qman_ceetm_set_prescaler(enum qm_dc_portal portal);
-+int qman_ceetm_get_prescaler(u16 *pres);
-+int qman_ceetm_query_cq(unsigned int cqid, unsigned int dcpid,
-+ struct qm_mcr_ceetm_cq_query *cq_query);
-+int qman_ceetm_query_ccgr(struct qm_mcc_ceetm_ccgr_query *ccgr_query,
-+ struct qm_mcr_ceetm_ccgr_query *response);
-+int qman_ceetm_get_xsfdr(enum qm_dc_portal portal, unsigned int *num);
-+
-+extern void *affine_portals[NR_CPUS];
-+const struct qm_portal_config *qman_get_qm_portal_config(
-+ struct qman_portal *portal);
-+
-+/* power management */
-+#ifdef CONFIG_SUSPEND
-+void suspend_unused_qportal(void);
-+void resume_unused_qportal(void);
-+#endif
---- /dev/null
-+++ b/drivers/staging/fsl_qbman/qman_test.c
-@@ -0,0 +1,57 @@
-+/* Copyright 2008-2011 Freescale Semiconductor, Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#include "qman_test.h"
-+
-+MODULE_AUTHOR("Geoff Thorpe");
-+MODULE_LICENSE("Dual BSD/GPL");
-+MODULE_DESCRIPTION("Qman testing");
-+
-+static int test_init(void)
-+{
-+ int loop = 1;
-+ while (loop--) {
-+#ifdef CONFIG_FSL_QMAN_TEST_STASH_POTATO
-+ qman_test_hotpotato();
-+#endif
-+#ifdef CONFIG_FSL_QMAN_TEST_HIGH
-+ qman_test_high();
-+#endif
-+ }
-+ return 0;
-+}
-+
-+static void test_exit(void)
-+{
-+}
-+
-+module_init(test_init);
-+module_exit(test_exit);
---- /dev/null
-+++ b/drivers/staging/fsl_qbman/qman_test.h
-@@ -0,0 +1,45 @@
-+/* Copyright 2008-2011 Freescale Semiconductor, Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#include <linux/kernel.h>
-+#include <linux/errno.h>
-+#include <linux/io.h>
-+#include <linux/slab.h>
-+#include <linux/module.h>
-+#include <linux/interrupt.h>
-+#include <linux/delay.h>
-+#include <linux/sched.h>
-+
-+#include <linux/fsl_qman.h>
-+
-+void qman_test_hotpotato(void);
-+void qman_test_high(void);
-+
---- /dev/null
-+++ b/drivers/staging/fsl_qbman/qman_test_high.c
-@@ -0,0 +1,216 @@
-+/* Copyright 2008-2011 Freescale Semiconductor, Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#include "qman_test.h"
-+
-+/*************/
-+/* constants */
-+/*************/
-+
-+#define CGR_ID 27
-+#define POOL_ID 2
-+#define FQ_FLAGS QMAN_FQ_FLAG_DYNAMIC_FQID
-+#define NUM_ENQUEUES 10
-+#define NUM_PARTIAL 4
-+#define PORTAL_SDQCR (QM_SDQCR_SOURCE_CHANNELS | \
-+ QM_SDQCR_TYPE_PRIO_QOS | \
-+ QM_SDQCR_TOKEN_SET(0x98) | \
-+ QM_SDQCR_CHANNELS_DEDICATED | \
-+ QM_SDQCR_CHANNELS_POOL(POOL_ID))
-+#define PORTAL_OPAQUE ((void *)0xf00dbeef)
-+#define VDQCR_FLAGS (QMAN_VOLATILE_FLAG_WAIT | QMAN_VOLATILE_FLAG_FINISH)
-+
-+/*************************************/
-+/* Predeclarations (eg. for fq_base) */
-+/*************************************/
-+
-+static enum qman_cb_dqrr_result cb_dqrr(struct qman_portal *,
-+ struct qman_fq *,
-+ const struct qm_dqrr_entry *);
-+static void cb_ern(struct qman_portal *, struct qman_fq *,
-+ const struct qm_mr_entry *);
-+static void cb_fqs(struct qman_portal *, struct qman_fq *,
-+ const struct qm_mr_entry *);
-+
-+/***************/
-+/* global vars */
-+/***************/
-+
-+static struct qm_fd fd, fd_dq;
-+static struct qman_fq fq_base = {
-+ .cb.dqrr = cb_dqrr,
-+ .cb.ern = cb_ern,
-+ .cb.fqs = cb_fqs
-+};
-+static DECLARE_WAIT_QUEUE_HEAD(waitqueue);
-+static int retire_complete, sdqcr_complete;
-+
-+/**********************/
-+/* internal functions */
-+/**********************/
-+
-+/* Helpers for initialising and "incrementing" a frame descriptor */
-+static void fd_init(struct qm_fd *__fd)
-+{
-+ qm_fd_addr_set64(__fd, 0xabdeadbeefLLU);
-+ __fd->format = qm_fd_contig_big;
-+ __fd->length29 = 0x0000ffff;
-+ __fd->cmd = 0xfeedf00d;
-+}
-+
-+static void fd_inc(struct qm_fd *__fd)
-+{
-+ u64 t = qm_fd_addr_get64(__fd);
-+ int z = t >> 40;
-+ t <<= 1;
-+ if (z)
-+ t |= 1;
-+ qm_fd_addr_set64(__fd, t);
-+ __fd->length29--;
-+ __fd->cmd++;
-+}
-+
-+/* The only part of the 'fd' we can't memcmp() is the ppid */
-+static int fd_cmp(const struct qm_fd *a, const struct qm_fd *b)
-+{
-+ int r = (qm_fd_addr_get64(a) == qm_fd_addr_get64(b)) ? 0 : -1;
-+ if (!r)
-+ r = a->format - b->format;
-+ if (!r)
-+ r = a->opaque - b->opaque;
-+ if (!r)
-+ r = a->cmd - b->cmd;
-+ return r;
-+}
-+
-+/********/
-+/* test */
-+/********/
-+
-+static void do_enqueues(struct qman_fq *fq)
-+{
-+ unsigned int loop;
-+ for (loop = 0; loop < NUM_ENQUEUES; loop++) {
-+ if (qman_enqueue(fq, &fd, QMAN_ENQUEUE_FLAG_WAIT |
-+ (((loop + 1) == NUM_ENQUEUES) ?
-+ QMAN_ENQUEUE_FLAG_WAIT_SYNC : 0)))
-+ panic("qman_enqueue() failed\n");
-+ fd_inc(&fd);
-+ }
-+}
-+
-+void qman_test_high(void)
-+{
-+ unsigned int flags;
-+ int res;
-+ struct qman_fq *fq = &fq_base;
-+
-+ pr_info("qman_test_high starting\n");
-+ fd_init(&fd);
-+ fd_init(&fd_dq);
-+
-+ /* Initialise (parked) FQ */
-+ if (qman_create_fq(0, FQ_FLAGS, fq))
-+ panic("qman_create_fq() failed\n");
-+ if (qman_init_fq(fq, QMAN_INITFQ_FLAG_LOCAL, NULL))
-+ panic("qman_init_fq() failed\n");
-+
-+ /* Do enqueues + VDQCR, twice. (Parked FQ) */
-+ do_enqueues(fq);
-+ pr_info("VDQCR (till-empty);\n");
-+ if (qman_volatile_dequeue(fq, VDQCR_FLAGS,
-+ QM_VDQCR_NUMFRAMES_TILLEMPTY))
-+ panic("qman_volatile_dequeue() failed\n");
-+ do_enqueues(fq);
-+ pr_info("VDQCR (%d of %d);\n", NUM_PARTIAL, NUM_ENQUEUES);
-+ if (qman_volatile_dequeue(fq, VDQCR_FLAGS,
-+ QM_VDQCR_NUMFRAMES_SET(NUM_PARTIAL)))
-+ panic("qman_volatile_dequeue() failed\n");
-+ pr_info("VDQCR (%d of %d);\n", NUM_ENQUEUES - NUM_PARTIAL,
-+ NUM_ENQUEUES);
-+ if (qman_volatile_dequeue(fq, VDQCR_FLAGS,
-+ QM_VDQCR_NUMFRAMES_SET(NUM_ENQUEUES - NUM_PARTIAL)))
-+ panic("qman_volatile_dequeue() failed\n");
-+
-+ do_enqueues(fq);
-+ pr_info("scheduled dequeue (till-empty)\n");
-+ if (qman_schedule_fq(fq))
-+ panic("qman_schedule_fq() failed\n");
-+ wait_event(waitqueue, sdqcr_complete);
-+
-+ /* Retire and OOS the FQ */
-+ res = qman_retire_fq(fq, &flags);
-+ if (res < 0)
-+ panic("qman_retire_fq() failed\n");
-+ wait_event(waitqueue, retire_complete);
-+ if (flags & QMAN_FQ_STATE_BLOCKOOS)
-+ panic("leaking frames\n");
-+ if (qman_oos_fq(fq))
-+ panic("qman_oos_fq() failed\n");
-+ qman_destroy_fq(fq, 0);
-+ pr_info("qman_test_high finished\n");
-+}
-+
-+static enum qman_cb_dqrr_result cb_dqrr(struct qman_portal *p,
-+ struct qman_fq *fq,
-+ const struct qm_dqrr_entry *dq)
-+{
-+ if (fd_cmp(&fd_dq, &dq->fd)) {
-+ pr_err("BADNESS: dequeued frame doesn't match;\n");
-+ pr_err("Expected 0x%llx, got 0x%llx\n",
-+ (unsigned long long)fd_dq.length29,
-+ (unsigned long long)dq->fd.length29);
-+ BUG();
-+ }
-+ fd_inc(&fd_dq);
-+ if (!(dq->stat & QM_DQRR_STAT_UNSCHEDULED) && !fd_cmp(&fd_dq, &fd)) {
-+ sdqcr_complete = 1;
-+ wake_up(&waitqueue);
-+ }
-+ return qman_cb_dqrr_consume;
-+}
-+
-+static void cb_ern(struct qman_portal *p, struct qman_fq *fq,
-+ const struct qm_mr_entry *msg)
-+{
-+ panic("cb_ern() unimplemented");
-+}
-+
-+static void cb_fqs(struct qman_portal *p, struct qman_fq *fq,
-+ const struct qm_mr_entry *msg)
-+{
-+ u8 verb = (msg->verb & QM_MR_VERB_TYPE_MASK);
-+ if ((verb != QM_MR_VERB_FQRN) && (verb != QM_MR_VERB_FQRNI))
-+ panic("unexpected FQS message");
-+ pr_info("Retirement message received\n");
-+ retire_complete = 1;
-+ wake_up(&waitqueue);
-+}
---- /dev/null
-+++ b/drivers/staging/fsl_qbman/qman_test_hotpotato.c
-@@ -0,0 +1,502 @@
-+/* Copyright 2009-2012 Freescale Semiconductor, Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#include <linux/kthread.h>
-+#include <linux/platform_device.h>
-+#include <linux/dma-mapping.h>
-+#include "qman_test.h"
-+
-+/* Algorithm:
-+ *
-+ * Each cpu will have HP_PER_CPU "handlers" set up, each of which incorporates
-+ * an rx/tx pair of FQ objects (both of which are stashed on dequeue). The
-+ * organisation of FQIDs is such that the HP_PER_CPU*NUM_CPUS handlers will
-+ * shuttle a "hot potato" frame around them such that every forwarding action
-+ * moves it from one cpu to another. (The use of more than one handler per cpu
-+ * is to allow enough handlers/FQs to truly test the significance of caching -
-+ * ie. when cache-expiries are occurring.)
-+ *
-+ * The "hot potato" frame content will be HP_NUM_WORDS*4 bytes in size, and the
-+ * first and last words of the frame data will undergo a transformation step on
-+ * each forwarding action. To achieve this, each handler will be assigned a
-+ * 32-bit "mixer", that is produced using a 32-bit LFSR. When a frame is
-+ * received by a handler, the mixer of the expected sender is XOR'd into all
-+ * words of the entire frame, which is then validated against the original
-+ * values. Then, before forwarding, the entire frame is XOR'd with the mixer of
-+ * the current handler. Apart from validating that the frame is taking the
-+ * expected path, this also provides some quasi-realistic overheads to each
-+ * forwarding action - dereferencing *all* the frame data, computation, and
-+ * conditional branching. There is a "special" handler designated to act as the
-+ * instigator of the test by creating an enqueuing the "hot potato" frame, and
-+ * to determine when the test has completed by counting HP_LOOPS iterations.
-+ *
-+ * Init phases:
-+ *
-+ * 1. prepare each cpu's 'hp_cpu' struct using on_each_cpu(,,1) and link them
-+ * into 'hp_cpu_list'. Specifically, set processor_id, allocate HP_PER_CPU
-+ * handlers and link-list them (but do no other handler setup).
-+ *
-+ * 2. scan over 'hp_cpu_list' HP_PER_CPU times, the first time sets each
-+ * hp_cpu's 'iterator' to point to its first handler. With each loop,
-+ * allocate rx/tx FQIDs and mixer values to the hp_cpu's iterator handler
-+ * and advance the iterator for the next loop. This includes a final fixup,
-+ * which connects the last handler to the first (and which is why phase 2
-+ * and 3 are separate).
-+ *
-+ * 3. scan over 'hp_cpu_list' HP_PER_CPU times, the first time sets each
-+ * hp_cpu's 'iterator' to point to its first handler. With each loop,
-+ * initialise FQ objects and advance the iterator for the next loop.
-+ * Moreover, do this initialisation on the cpu it applies to so that Rx FQ
-+ * initialisation targets the correct cpu.
-+ */
-+
-+/* helper to run something on all cpus (can't use on_each_cpu(), as that invokes
-+ * the fn from irq context, which is too restrictive). */
-+struct bstrap {
-+ void (*fn)(void);
-+ atomic_t started;
-+};
-+static int bstrap_fn(void *__bstrap)
-+{
-+ struct bstrap *bstrap = __bstrap;
-+ atomic_inc(&bstrap->started);
-+ bstrap->fn();
-+ while (!kthread_should_stop())
-+ msleep(1);
-+ return 0;
-+}
-+static int on_all_cpus(void (*fn)(void))
-+{
-+ int cpu;
-+ for_each_cpu(cpu, cpu_online_mask) {
-+ struct bstrap bstrap = {
-+ .fn = fn,
-+ .started = ATOMIC_INIT(0)
-+ };
-+ struct task_struct *k = kthread_create(bstrap_fn, &bstrap,
-+ "hotpotato%d", cpu);
-+ int ret;
-+ if (IS_ERR(k))
-+ return -ENOMEM;
-+ kthread_bind(k, cpu);
-+ wake_up_process(k);
-+ /* If we call kthread_stop() before the "wake up" has had an
-+ * effect, then the thread may exit with -EINTR without ever
-+ * running the function. So poll until it's started before
-+ * requesting it to stop. */
-+ while (!atomic_read(&bstrap.started))
-+ msleep(10);
-+ ret = kthread_stop(k);
-+ if (ret)
-+ return ret;
-+ }
-+ return 0;
-+}
-+
-+struct hp_handler {
-+
-+ /* The following data is stashed when 'rx' is dequeued; */
-+ /* -------------- */
-+ /* The Rx FQ, dequeues of which will stash the entire hp_handler */
-+ struct qman_fq rx;
-+ /* The Tx FQ we should forward to */
-+ struct qman_fq tx;
-+ /* The value we XOR post-dequeue, prior to validating */
-+ u32 rx_mixer;
-+ /* The value we XOR pre-enqueue, after validating */
-+ u32 tx_mixer;
-+ /* what the hotpotato address should be on dequeue */
-+ dma_addr_t addr;
-+ u32 *frame_ptr;
-+
-+ /* The following data isn't (necessarily) stashed on dequeue; */
-+ /* -------------- */
-+ u32 fqid_rx, fqid_tx;
-+ /* list node for linking us into 'hp_cpu' */
-+ struct list_head node;
-+ /* Just to check ... */
-+ unsigned int processor_id;
-+} ____cacheline_aligned;
-+
-+struct hp_cpu {
-+ /* identify the cpu we run on; */
-+ unsigned int processor_id;
-+ /* root node for the per-cpu list of handlers */
-+ struct list_head handlers;
-+ /* list node for linking us into 'hp_cpu_list' */
-+ struct list_head node;
-+ /* when repeatedly scanning 'hp_list', each time linking the n'th
-+ * handlers together, this is used as per-cpu iterator state */
-+ struct hp_handler *iterator;
-+};
-+
-+/* Each cpu has one of these */
-+static DEFINE_PER_CPU(struct hp_cpu, hp_cpus);
-+
-+/* links together the hp_cpu structs, in first-come first-serve order. */
-+static LIST_HEAD(hp_cpu_list);
-+static spinlock_t hp_lock = __SPIN_LOCK_UNLOCKED(hp_lock);
-+
-+static unsigned int hp_cpu_list_length;
-+
-+/* the "special" handler, that starts and terminates the test. */
-+static struct hp_handler *special_handler;
-+static int loop_counter;
-+
-+/* handlers are allocated out of this, so they're properly aligned. */
-+static struct kmem_cache *hp_handler_slab;
-+
-+/* this is the frame data */
-+static void *__frame_ptr;
-+static u32 *frame_ptr;
-+static dma_addr_t frame_dma;
-+
-+/* the main function waits on this */
-+static DECLARE_WAIT_QUEUE_HEAD(queue);
-+
-+#define HP_PER_CPU 2
-+#define HP_LOOPS 8
-+/* 80 bytes, like a small ethernet frame, and bleeds into a second cacheline */
-+#define HP_NUM_WORDS 80
-+/* First word of the LFSR-based frame data */
-+#define HP_FIRST_WORD 0xabbaf00d
-+
-+static inline u32 do_lfsr(u32 prev)
-+{
-+ return (prev >> 1) ^ (-(prev & 1u) & 0xd0000001u);
-+}
-+
-+static void allocate_frame_data(void)
-+{
-+ u32 lfsr = HP_FIRST_WORD;
-+ int loop;
-+ struct platform_device *pdev = platform_device_alloc("foobar", -1);
-+ if (!pdev)
-+ panic("platform_device_alloc() failed");
-+ if (platform_device_add(pdev))
-+ panic("platform_device_add() failed");
-+ __frame_ptr = kmalloc(4 * HP_NUM_WORDS, GFP_KERNEL);
-+ if (!__frame_ptr)
-+ panic("kmalloc() failed");
-+ frame_ptr = (void *)(((unsigned long)__frame_ptr + 63) &
-+ ~(unsigned long)63);
-+ for (loop = 0; loop < HP_NUM_WORDS; loop++) {
-+ frame_ptr[loop] = lfsr;
-+ lfsr = do_lfsr(lfsr);
-+ }
-+ frame_dma = dma_map_single(&pdev->dev, frame_ptr, 4 * HP_NUM_WORDS,
-+ DMA_BIDIRECTIONAL);
-+ platform_device_del(pdev);
-+ platform_device_put(pdev);
-+}
-+
-+static void deallocate_frame_data(void)
-+{
-+ kfree(__frame_ptr);
-+}
-+
-+static inline void process_frame_data(struct hp_handler *handler,
-+ const struct qm_fd *fd)
-+{
-+ u32 *p = handler->frame_ptr;
-+ u32 lfsr = HP_FIRST_WORD;
-+ int loop;
-+ if (qm_fd_addr_get64(fd) != (handler->addr & 0xffffffffff)) {
-+ pr_err("Got 0x%llx expected 0x%llx\n",
-+ qm_fd_addr_get64(fd), handler->addr);
-+ panic("bad frame address");
-+ }
-+ for (loop = 0; loop < HP_NUM_WORDS; loop++, p++) {
-+ *p ^= handler->rx_mixer;
-+ if (*p != lfsr)
-+ panic("corrupt frame data");
-+ *p ^= handler->tx_mixer;
-+ lfsr = do_lfsr(lfsr);
-+ }
-+}
-+
-+static enum qman_cb_dqrr_result normal_dqrr(struct qman_portal *portal,
-+ struct qman_fq *fq,
-+ const struct qm_dqrr_entry *dqrr)
-+{
-+ struct hp_handler *handler = (struct hp_handler *)fq;
-+
-+ process_frame_data(handler, &dqrr->fd);
-+ if (qman_enqueue(&handler->tx, &dqrr->fd, 0))
-+ panic("qman_enqueue() failed");
-+ return qman_cb_dqrr_consume;
-+}
-+
-+static enum qman_cb_dqrr_result special_dqrr(struct qman_portal *portal,
-+ struct qman_fq *fq,
-+ const struct qm_dqrr_entry *dqrr)
-+{
-+ struct hp_handler *handler = (struct hp_handler *)fq;
-+
-+ process_frame_data(handler, &dqrr->fd);
-+ if (++loop_counter < HP_LOOPS) {
-+ if (qman_enqueue(&handler->tx, &dqrr->fd, 0))
-+ panic("qman_enqueue() failed");
-+ } else {
-+ pr_info("Received final (%dth) frame\n", loop_counter);
-+ wake_up(&queue);
-+ }
-+ return qman_cb_dqrr_consume;
-+}
-+
-+static void create_per_cpu_handlers(void)
-+{
-+ struct hp_handler *handler;
-+ int loop;
-+ struct hp_cpu *hp_cpu = &get_cpu_var(hp_cpus);
-+
-+ hp_cpu->processor_id = smp_processor_id();
-+ spin_lock(&hp_lock);
-+ list_add_tail(&hp_cpu->node, &hp_cpu_list);
-+ hp_cpu_list_length++;
-+ spin_unlock(&hp_lock);
-+ INIT_LIST_HEAD(&hp_cpu->handlers);
-+ for (loop = 0; loop < HP_PER_CPU; loop++) {
-+ handler = kmem_cache_alloc(hp_handler_slab, GFP_KERNEL);
-+ if (!handler)
-+ panic("kmem_cache_alloc() failed");
-+ handler->processor_id = hp_cpu->processor_id;
-+ handler->addr = frame_dma;
-+ handler->frame_ptr = frame_ptr;
-+ list_add_tail(&handler->node, &hp_cpu->handlers);
-+ }
-+ put_cpu_var(hp_cpus);
-+}
-+
-+static void destroy_per_cpu_handlers(void)
-+{
-+ struct list_head *loop, *tmp;
-+ struct hp_cpu *hp_cpu = &get_cpu_var(hp_cpus);
-+
-+ spin_lock(&hp_lock);
-+ list_del(&hp_cpu->node);
-+ spin_unlock(&hp_lock);
-+ list_for_each_safe(loop, tmp, &hp_cpu->handlers) {
-+ u32 flags;
-+ struct hp_handler *handler = list_entry(loop, struct hp_handler,
-+ node);
-+ if (qman_retire_fq(&handler->rx, &flags))
-+ panic("qman_retire_fq(rx) failed");
-+ BUG_ON(flags & QMAN_FQ_STATE_BLOCKOOS);
-+ if (qman_oos_fq(&handler->rx))
-+ panic("qman_oos_fq(rx) failed");
-+ qman_destroy_fq(&handler->rx, 0);
-+ qman_destroy_fq(&handler->tx, 0);
-+ qman_release_fqid(handler->fqid_rx);
-+ list_del(&handler->node);
-+ kmem_cache_free(hp_handler_slab, handler);
-+ }
-+ put_cpu_var(hp_cpus);
-+}
-+
-+static inline u8 num_cachelines(u32 offset)
-+{
-+ u8 res = (offset + (L1_CACHE_BYTES - 1))
-+ / (L1_CACHE_BYTES);
-+ if (res > 3)
-+ return 3;
-+ return res;
-+}
-+#define STASH_DATA_CL \
-+ num_cachelines(HP_NUM_WORDS * 4)
-+#define STASH_CTX_CL \
-+ num_cachelines(offsetof(struct hp_handler, fqid_rx))
-+
-+static void init_handler(void *__handler)
-+{
-+ struct qm_mcc_initfq opts;
-+ struct hp_handler *handler = __handler;
-+ BUG_ON(handler->processor_id != smp_processor_id());
-+ /* Set up rx */
-+ memset(&handler->rx, 0, sizeof(handler->rx));
-+ if (handler == special_handler)
-+ handler->rx.cb.dqrr = special_dqrr;
-+ else
-+ handler->rx.cb.dqrr = normal_dqrr;
-+ if (qman_create_fq(handler->fqid_rx, 0, &handler->rx))
-+ panic("qman_create_fq(rx) failed");
-+ memset(&opts, 0, sizeof(opts));
-+ opts.we_mask = QM_INITFQ_WE_FQCTRL | QM_INITFQ_WE_CONTEXTA;
-+ opts.fqd.fq_ctrl = QM_FQCTRL_CTXASTASHING;
-+ opts.fqd.context_a.stashing.data_cl = STASH_DATA_CL;
-+ opts.fqd.context_a.stashing.context_cl = STASH_CTX_CL;
-+ if (qman_init_fq(&handler->rx, QMAN_INITFQ_FLAG_SCHED |
-+ QMAN_INITFQ_FLAG_LOCAL, &opts))
-+ panic("qman_init_fq(rx) failed");
-+ /* Set up tx */
-+ memset(&handler->tx, 0, sizeof(handler->tx));
-+ if (qman_create_fq(handler->fqid_tx, QMAN_FQ_FLAG_NO_MODIFY,
-+ &handler->tx))
-+ panic("qman_create_fq(tx) failed");
-+}
-+
-+static void init_phase2(void)
-+{
-+ int loop;
-+ u32 fqid = 0;
-+ u32 lfsr = 0xdeadbeef;
-+ struct hp_cpu *hp_cpu;
-+ struct hp_handler *handler;
-+
-+ for (loop = 0; loop < HP_PER_CPU; loop++) {
-+ list_for_each_entry(hp_cpu, &hp_cpu_list, node) {
-+ int ret;
-+ if (!loop)
-+ hp_cpu->iterator = list_first_entry(
-+ &hp_cpu->handlers,
-+ struct hp_handler, node);
-+ else
-+ hp_cpu->iterator = list_entry(
-+ hp_cpu->iterator->node.next,
-+ struct hp_handler, node);
-+ /* Rx FQID is the previous handler's Tx FQID */
-+ hp_cpu->iterator->fqid_rx = fqid;
-+ /* Allocate new FQID for Tx */
-+ ret = qman_alloc_fqid(&fqid);
-+ if (ret)
-+ panic("qman_alloc_fqid() failed");
-+ hp_cpu->iterator->fqid_tx = fqid;
-+ /* Rx mixer is the previous handler's Tx mixer */
-+ hp_cpu->iterator->rx_mixer = lfsr;
-+ /* Get new mixer for Tx */
-+ lfsr = do_lfsr(lfsr);
-+ hp_cpu->iterator->tx_mixer = lfsr;
-+ }
-+ }
-+ /* Fix up the first handler (fqid_rx==0, rx_mixer=0xdeadbeef) */
-+ hp_cpu = list_first_entry(&hp_cpu_list, struct hp_cpu, node);
-+ handler = list_first_entry(&hp_cpu->handlers, struct hp_handler, node);
-+ BUG_ON((handler->fqid_rx != 0) || (handler->rx_mixer != 0xdeadbeef));
-+ handler->fqid_rx = fqid;
-+ handler->rx_mixer = lfsr;
-+ /* and tag it as our "special" handler */
-+ special_handler = handler;
-+}
-+
-+static void init_phase3(void)
-+{
-+ int loop;
-+ struct hp_cpu *hp_cpu;
-+
-+ for (loop = 0; loop < HP_PER_CPU; loop++) {
-+ list_for_each_entry(hp_cpu, &hp_cpu_list, node) {
-+ if (!loop)
-+ hp_cpu->iterator = list_first_entry(
-+ &hp_cpu->handlers,
-+ struct hp_handler, node);
-+ else
-+ hp_cpu->iterator = list_entry(
-+ hp_cpu->iterator->node.next,
-+ struct hp_handler, node);
-+ preempt_disable();
-+ if (hp_cpu->processor_id == smp_processor_id())
-+ init_handler(hp_cpu->iterator);
-+ else
-+ smp_call_function_single(hp_cpu->processor_id,
-+ init_handler, hp_cpu->iterator, 1);
-+ preempt_enable();
-+ }
-+ }
-+}
-+
-+static void send_first_frame(void *ignore)
-+{
-+ u32 *p = special_handler->frame_ptr;
-+ u32 lfsr = HP_FIRST_WORD;
-+ int loop;
-+ struct qm_fd fd;
-+
-+ BUG_ON(special_handler->processor_id != smp_processor_id());
-+ memset(&fd, 0, sizeof(fd));
-+ qm_fd_addr_set64(&fd, special_handler->addr);
-+ fd.format = qm_fd_contig_big;
-+ fd.length29 = HP_NUM_WORDS * 4;
-+ for (loop = 0; loop < HP_NUM_WORDS; loop++, p++) {
-+ if (*p != lfsr)
-+ panic("corrupt frame data");
-+ *p ^= special_handler->tx_mixer;
-+ lfsr = do_lfsr(lfsr);
-+ }
-+ pr_info("Sending first frame\n");
-+ if (qman_enqueue(&special_handler->tx, &fd, 0))
-+ panic("qman_enqueue() failed");
-+}
-+
-+void qman_test_hotpotato(void)
-+{
-+ if (cpumask_weight(cpu_online_mask) < 2) {
-+ pr_info("qman_test_hotpotato, skip - only 1 CPU\n");
-+ return;
-+ }
-+
-+ pr_info("qman_test_hotpotato starting\n");
-+
-+ hp_cpu_list_length = 0;
-+ loop_counter = 0;
-+ hp_handler_slab = kmem_cache_create("hp_handler_slab",
-+ sizeof(struct hp_handler), L1_CACHE_BYTES,
-+ SLAB_HWCACHE_ALIGN, NULL);
-+ if (!hp_handler_slab)
-+ panic("kmem_cache_create() failed");
-+
-+ allocate_frame_data();
-+
-+ /* Init phase 1 */
-+ pr_info("Creating %d handlers per cpu...\n", HP_PER_CPU);
-+ if (on_all_cpus(create_per_cpu_handlers))
-+ panic("on_each_cpu() failed");
-+ pr_info("Number of cpus: %d, total of %d handlers\n",
-+ hp_cpu_list_length, hp_cpu_list_length * HP_PER_CPU);
-+
-+ init_phase2();
-+
-+ init_phase3();
-+
-+ preempt_disable();
-+ if (special_handler->processor_id == smp_processor_id())
-+ send_first_frame(NULL);
-+ else
-+ smp_call_function_single(special_handler->processor_id,
-+ send_first_frame, NULL, 1);
-+ preempt_enable();
-+
-+ wait_event(queue, loop_counter == HP_LOOPS);
-+ deallocate_frame_data();
-+ if (on_all_cpus(destroy_per_cpu_handlers))
-+ panic("on_each_cpu() failed");
-+ kmem_cache_destroy(hp_handler_slab);
-+ pr_info("qman_test_hotpotato finished\n");
-+}
---- /dev/null
-+++ b/drivers/staging/fsl_qbman/qman_utility.c
-@@ -0,0 +1,129 @@
-+/* Copyright 2008-2011 Freescale Semiconductor, Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#include "qman_private.h"
-+
-+/* ----------------- */
-+/* --- FQID Pool --- */
-+
-+struct qman_fqid_pool {
-+ /* Base and size of the FQID range */
-+ u32 fqid_base;
-+ u32 total;
-+ /* Number of FQIDs currently "allocated" */
-+ u32 used;
-+ /* Allocation optimisation. When 'used<total', it is the index of an
-+ * available FQID. Otherwise there are no available FQIDs, and this
-+ * will be set when the next deallocation occurs. */
-+ u32 next;
-+ /* A bit-field representation of the FQID range. */
-+ unsigned long *bits;
-+};
-+
-+#define QLONG_BYTES sizeof(unsigned long)
-+#define QLONG_BITS (QLONG_BYTES * 8)
-+/* Number of 'longs' required for the given number of bits */
-+#define QNUM_LONGS(b) (((b) + QLONG_BITS - 1) / QLONG_BITS)
-+/* Shorthand for the number of bytes of same (kmalloc, memset, etc) */
-+#define QNUM_BYTES(b) (QNUM_LONGS(b) * QLONG_BYTES)
-+/* And in bits */
-+#define QNUM_BITS(b) (QNUM_LONGS(b) * QLONG_BITS)
-+
-+struct qman_fqid_pool *qman_fqid_pool_create(u32 fqid_start, u32 num)
-+{
-+ struct qman_fqid_pool *pool = kmalloc(sizeof(*pool), GFP_KERNEL);
-+ unsigned int i;
-+
-+ BUG_ON(!num);
-+ if (!pool)
-+ return NULL;
-+ pool->fqid_base = fqid_start;
-+ pool->total = num;
-+ pool->used = 0;
-+ pool->next = 0;
-+ pool->bits = kzalloc(QNUM_BYTES(num), GFP_KERNEL);
-+ if (!pool->bits) {
-+ kfree(pool);
-+ return NULL;
-+ }
-+ /* If num is not an even multiple of QLONG_BITS (or even 8, for
-+ * byte-oriented searching) then we fill the trailing bits with 1, to
-+ * make them look allocated (permanently). */
-+ for (i = num + 1; i < QNUM_BITS(num); i++)
-+ set_bit(i, pool->bits);
-+ return pool;
-+}
-+EXPORT_SYMBOL(qman_fqid_pool_create);
-+
-+int qman_fqid_pool_destroy(struct qman_fqid_pool *pool)
-+{
-+ int ret = pool->used;
-+ kfree(pool->bits);
-+ kfree(pool);
-+ return ret;
-+}
-+EXPORT_SYMBOL(qman_fqid_pool_destroy);
-+
-+int qman_fqid_pool_alloc(struct qman_fqid_pool *pool, u32 *fqid)
-+{
-+ int ret;
-+ if (pool->used == pool->total)
-+ return -ENOMEM;
-+ *fqid = pool->fqid_base + pool->next;
-+ ret = test_and_set_bit(pool->next, pool->bits);
-+ BUG_ON(ret);
-+ if (++pool->used == pool->total)
-+ return 0;
-+ pool->next = find_next_zero_bit(pool->bits, pool->total, pool->next);
-+ if (pool->next >= pool->total)
-+ pool->next = find_first_zero_bit(pool->bits, pool->total);
-+ BUG_ON(pool->next >= pool->total);
-+ return 0;
-+}
-+EXPORT_SYMBOL(qman_fqid_pool_alloc);
-+
-+void qman_fqid_pool_free(struct qman_fqid_pool *pool, u32 fqid)
-+{
-+ int ret;
-+
-+ fqid -= pool->fqid_base;
-+ ret = test_and_clear_bit(fqid, pool->bits);
-+ BUG_ON(!ret);
-+ if (pool->used-- == pool->total)
-+ pool->next = fqid;
-+}
-+EXPORT_SYMBOL(qman_fqid_pool_free);
-+
-+u32 qman_fqid_pool_used(struct qman_fqid_pool *pool)
-+{
-+ return pool->used;
-+}
-+EXPORT_SYMBOL(qman_fqid_pool_used);
---- /dev/null
-+++ b/include/linux/fsl_bman.h
-@@ -0,0 +1,532 @@
-+/* Copyright 2008-2012 Freescale Semiconductor, Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#ifndef FSL_BMAN_H
-+#define FSL_BMAN_H
-+
-+#ifdef __cplusplus
-+extern "C" {
-+#endif
-+
-+/* Last updated for v00.79 of the BG */
-+
-+/* Portal processing (interrupt) sources */
-+#define BM_PIRQ_RCRI 0x00000002 /* RCR Ring (below threshold) */
-+#define BM_PIRQ_BSCN 0x00000001 /* Buffer depletion State Change */
-+
-+/* This wrapper represents a bit-array for the depletion state of the 64 Bman
-+ * buffer pools. */
-+struct bman_depletion {
-+ u32 __state[2];
-+};
-+#define BMAN_DEPLETION_EMPTY { { 0x00000000, 0x00000000 } }
-+#define BMAN_DEPLETION_FULL { { 0xffffffff, 0xffffffff } }
-+#define __bmdep_word(x) ((x) >> 5)
-+#define __bmdep_shift(x) ((x) & 0x1f)
-+#define __bmdep_bit(x) (0x80000000 >> __bmdep_shift(x))
-+static inline void bman_depletion_init(struct bman_depletion *c)
-+{
-+ c->__state[0] = c->__state[1] = 0;
-+}
-+static inline void bman_depletion_fill(struct bman_depletion *c)
-+{
-+ c->__state[0] = c->__state[1] = ~0;
-+}
-+static inline int bman_depletion_get(const struct bman_depletion *c, u8 bpid)
-+{
-+ return c->__state[__bmdep_word(bpid)] & __bmdep_bit(bpid);
-+}
-+static inline void bman_depletion_set(struct bman_depletion *c, u8 bpid)
-+{
-+ c->__state[__bmdep_word(bpid)] |= __bmdep_bit(bpid);
-+}
-+static inline void bman_depletion_unset(struct bman_depletion *c, u8 bpid)
-+{
-+ c->__state[__bmdep_word(bpid)] &= ~__bmdep_bit(bpid);
-+}
-+
-+/* ------------------------------------------------------- */
-+/* --- Bman data structures (and associated constants) --- */
-+
-+/* Represents s/w corenet portal mapped data structures */
-+struct bm_rcr_entry; /* RCR (Release Command Ring) entries */
-+struct bm_mc_command; /* MC (Management Command) command */
-+struct bm_mc_result; /* MC result */
-+
-+/* Code-reduction, define a wrapper for 48-bit buffers. In cases where a buffer
-+ * pool id specific to this buffer is needed (BM_RCR_VERB_CMD_BPID_MULTI,
-+ * BM_MCC_VERB_ACQUIRE), the 'bpid' field is used. */
-+struct bm_buffer {
-+ union {
-+ struct {
-+#if __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
-+ u8 __reserved1;
-+ u8 bpid;
-+ u16 hi; /* High 16-bits of 48-bit address */
-+ u32 lo; /* Low 32-bits of 48-bit address */
-+#else
-+ u32 lo;
-+ u16 hi;
-+ u8 bpid;
-+ u8 __reserved;
-+#endif
-+ };
-+ struct {
-+#if __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
-+ u64 __notaddress:16;
-+ u64 addr:48;
-+#else
-+ u64 addr:48;
-+ u64 __notaddress:16;
-+#endif
-+ };
-+ u64 opaque;
-+ };
-+} __aligned(8);
-+static inline u64 bm_buffer_get64(const struct bm_buffer *buf)
-+{
-+ return buf->addr;
-+}
-+static inline dma_addr_t bm_buf_addr(const struct bm_buffer *buf)
-+{
-+ return (dma_addr_t)buf->addr;
-+}
-+/* Macro, so we compile better if 'v' isn't always 64-bit */
-+#define bm_buffer_set64(buf, v) \
-+ do { \
-+ struct bm_buffer *__buf931 = (buf); \
-+ __buf931->hi = upper_32_bits(v); \
-+ __buf931->lo = lower_32_bits(v); \
-+ } while (0)
-+
-+/* See 1.5.3.5.4: "Release Command" */
-+struct bm_rcr_entry {
-+ union {
-+ struct {
-+ u8 __dont_write_directly__verb;
-+ u8 bpid; /* used with BM_RCR_VERB_CMD_BPID_SINGLE */
-+ u8 __reserved1[62];
-+ };
-+ struct bm_buffer bufs[8];
-+ };
-+} __packed;
-+#define BM_RCR_VERB_VBIT 0x80
-+#define BM_RCR_VERB_CMD_MASK 0x70 /* one of two values; */
-+#define BM_RCR_VERB_CMD_BPID_SINGLE 0x20
-+#define BM_RCR_VERB_CMD_BPID_MULTI 0x30
-+#define BM_RCR_VERB_BUFCOUNT_MASK 0x0f /* values 1..8 */
-+
-+/* See 1.5.3.1: "Acquire Command" */
-+/* See 1.5.3.2: "Query Command" */
-+struct bm_mcc_acquire {
-+ u8 bpid;
-+ u8 __reserved1[62];
-+} __packed;
-+struct bm_mcc_query {
-+ u8 __reserved2[63];
-+} __packed;
-+struct bm_mc_command {
-+ u8 __dont_write_directly__verb;
-+ union {
-+ struct bm_mcc_acquire acquire;
-+ struct bm_mcc_query query;
-+ };
-+} __packed;
-+#define BM_MCC_VERB_VBIT 0x80
-+#define BM_MCC_VERB_CMD_MASK 0x70 /* where the verb contains; */
-+#define BM_MCC_VERB_CMD_ACQUIRE 0x10
-+#define BM_MCC_VERB_CMD_QUERY 0x40
-+#define BM_MCC_VERB_ACQUIRE_BUFCOUNT 0x0f /* values 1..8 go here */
-+
-+/* See 1.5.3.3: "Acquire Response" */
-+/* See 1.5.3.4: "Query Response" */
-+struct bm_pool_state {
-+ u8 __reserved1[32];
-+ /* "availability state" and "depletion state" */
-+ struct {
-+ u8 __reserved1[8];
-+ /* Access using bman_depletion_***() */
-+ struct bman_depletion state;
-+ } as, ds;
-+};
-+struct bm_mc_result {
-+ union {
-+ struct {
-+ u8 verb;
-+ u8 __reserved1[63];
-+ };
-+ union {
-+ struct {
-+ u8 __reserved1;
-+ u8 bpid;
-+ u8 __reserved2[62];
-+ };
-+ struct bm_buffer bufs[8];
-+ } acquire;
-+ struct bm_pool_state query;
-+ };
-+} __packed;
-+#define BM_MCR_VERB_VBIT 0x80
-+#define BM_MCR_VERB_CMD_MASK BM_MCC_VERB_CMD_MASK
-+#define BM_MCR_VERB_CMD_ACQUIRE BM_MCC_VERB_CMD_ACQUIRE
-+#define BM_MCR_VERB_CMD_QUERY BM_MCC_VERB_CMD_QUERY
-+#define BM_MCR_VERB_CMD_ERR_INVALID 0x60
-+#define BM_MCR_VERB_CMD_ERR_ECC 0x70
-+#define BM_MCR_VERB_ACQUIRE_BUFCOUNT BM_MCC_VERB_ACQUIRE_BUFCOUNT /* 0..8 */
-+/* Determine the "availability state" of pool 'p' from a query result 'r' */
-+#define BM_MCR_QUERY_AVAILABILITY(r, p) \
-+ bman_depletion_get(&r->query.as.state, p)
-+/* Determine the "depletion state" of pool 'p' from a query result 'r' */
-+#define BM_MCR_QUERY_DEPLETION(r, p) \
-+ bman_depletion_get(&r->query.ds.state, p)
-+
-+/*******************************************************************/
-+/* Managed (aka "shared" or "mux/demux") portal, high-level i/face */
-+/*******************************************************************/
-+
-+ /* Portal and Buffer Pools */
-+ /* ----------------------- */
-+/* Represents a managed portal */
-+struct bman_portal;
-+
-+/* This object type represents Bman buffer pools. */
-+struct bman_pool;
-+
-+struct bman_portal_config {
-+ /* This is used for any "core-affine" portals, ie. default portals
-+ * associated to the corresponding cpu. -1 implies that there is no core
-+ * affinity configured. */
-+ int cpu;
-+ /* portal interrupt line */
-+ int irq;
-+ /* the unique index of this portal */
-+ u32 index;
-+ /* Is this portal shared? (If so, it has coarser locking and demuxes
-+ * processing on behalf of other CPUs.) */
-+ int is_shared;
-+ /* These are the buffer pool IDs that may be used via this portal. */
-+ struct bman_depletion mask;
-+};
-+
-+/* This callback type is used when handling pool depletion entry/exit. The
-+ * 'cb_ctx' value is the opaque value associated with the pool object in
-+ * bman_new_pool(). 'depleted' is non-zero on depletion-entry, and zero on
-+ * depletion-exit. */
-+typedef void (*bman_cb_depletion)(struct bman_portal *bm,
-+ struct bman_pool *pool, void *cb_ctx, int depleted);
-+
-+/* This struct specifies parameters for a bman_pool object. */
-+struct bman_pool_params {
-+ /* index of the buffer pool to encapsulate (0-63), ignored if
-+ * BMAN_POOL_FLAG_DYNAMIC_BPID is set. */
-+ u32 bpid;
-+ /* bit-mask of BMAN_POOL_FLAG_*** options */
-+ u32 flags;
-+ /* depletion-entry/exit callback, if BMAN_POOL_FLAG_DEPLETION is set */
-+ bman_cb_depletion cb;
-+ /* opaque user value passed as a parameter to 'cb' */
-+ void *cb_ctx;
-+ /* depletion-entry/exit thresholds, if BMAN_POOL_FLAG_THRESH is set. NB:
-+ * this is only allowed if BMAN_POOL_FLAG_DYNAMIC_BPID is used *and*
-+ * when run in the control plane (which controls Bman CCSR). This array
-+ * matches the definition of bm_pool_set(). */
-+ u32 thresholds[4];
-+};
-+
-+/* Flags to bman_new_pool() */
-+#define BMAN_POOL_FLAG_NO_RELEASE 0x00000001 /* can't release to pool */
-+#define BMAN_POOL_FLAG_ONLY_RELEASE 0x00000002 /* can only release to pool */
-+#define BMAN_POOL_FLAG_DEPLETION 0x00000004 /* track depletion entry/exit */
-+#define BMAN_POOL_FLAG_DYNAMIC_BPID 0x00000008 /* (de)allocate bpid */
-+#define BMAN_POOL_FLAG_THRESH 0x00000010 /* set depletion thresholds */
-+#define BMAN_POOL_FLAG_STOCKPILE 0x00000020 /* stockpile to reduce hw ops */
-+
-+/* Flags to bman_release() */
-+#ifdef CONFIG_FSL_DPA_CAN_WAIT
-+#define BMAN_RELEASE_FLAG_WAIT 0x00000001 /* wait if RCR is full */
-+#define BMAN_RELEASE_FLAG_WAIT_INT 0x00000002 /* if we wait, interruptible? */
-+#ifdef CONFIG_FSL_DPA_CAN_WAIT_SYNC
-+#define BMAN_RELEASE_FLAG_WAIT_SYNC 0x00000004 /* if wait, until consumed? */
-+#endif
-+#endif
-+#define BMAN_RELEASE_FLAG_NOW 0x00000008 /* issue immediate release */
-+
-+/* Flags to bman_acquire() */
-+#define BMAN_ACQUIRE_FLAG_STOCKPILE 0x00000001 /* no hw op, stockpile only */
-+
-+ /* Portal Management */
-+ /* ----------------- */
-+/**
-+ * bman_get_portal_config - get portal configuration settings
-+ *
-+ * This returns a read-only view of the current cpu's affine portal settings.
-+ */
-+const struct bman_portal_config *bman_get_portal_config(void);
-+
-+/**
-+ * bman_irqsource_get - return the portal work that is interrupt-driven
-+ *
-+ * Returns a bitmask of BM_PIRQ_**I processing sources that are currently
-+ * enabled for interrupt handling on the current cpu's affine portal. These
-+ * sources will trigger the portal interrupt and the interrupt handler (or a
-+ * tasklet/bottom-half it defers to) will perform the corresponding processing
-+ * work. The bman_poll_***() functions will only process sources that are not in
-+ * this bitmask. If the current CPU is sharing a portal hosted on another CPU,
-+ * this always returns zero.
-+ */
-+u32 bman_irqsource_get(void);
-+
-+/**
-+ * bman_irqsource_add - add processing sources to be interrupt-driven
-+ * @bits: bitmask of BM_PIRQ_**I processing sources
-+ *
-+ * Adds processing sources that should be interrupt-driven (rather than
-+ * processed via bman_poll_***() functions). Returns zero for success, or
-+ * -EINVAL if the current CPU is sharing a portal hosted on another CPU. */
-+int bman_irqsource_add(u32 bits);
-+
-+/**
-+ * bman_irqsource_remove - remove processing sources from being interrupt-driven
-+ * @bits: bitmask of BM_PIRQ_**I processing sources
-+ *
-+ * Removes processing sources from being interrupt-driven, so that they will
-+ * instead be processed via bman_poll_***() functions. Returns zero for success,
-+ * or -EINVAL if the current CPU is sharing a portal hosted on another CPU. */
-+int bman_irqsource_remove(u32 bits);
-+
-+/**
-+ * bman_affine_cpus - return a mask of cpus that have affine portals
-+ */
-+const cpumask_t *bman_affine_cpus(void);
-+
-+/**
-+ * bman_poll_slow - process anything that isn't interrupt-driven.
-+ *
-+ * This function does any portal processing that isn't interrupt-driven. If the
-+ * current CPU is sharing a portal hosted on another CPU, this function will
-+ * return -EINVAL, otherwise the return value is a bitmask of BM_PIRQ_* sources
-+ * indicating what interrupt sources were actually processed by the call.
-+ *
-+ * NB, unlike the legacy wrapper bman_poll(), this function will
-+ * deterministically check for the presence of portal processing work and do it,
-+ * which implies some latency even if there's nothing to do. The bman_poll()
-+ * wrapper on the other hand (like the qman_poll() wrapper) attenuates this by
-+ * checking for (and doing) portal processing infrequently. Ie. such that
-+ * qman_poll() and bman_poll() can be called from core-processing loops. Use
-+ * bman_poll_slow() when you yourself are deciding when to incur the overhead of
-+ * processing.
-+ */
-+u32 bman_poll_slow(void);
-+
-+/**
-+ * bman_poll - process anything that isn't interrupt-driven.
-+ *
-+ * Dispatcher logic on a cpu can use this to trigger any maintenance of the
-+ * affine portal. This function does whatever processing is not triggered by
-+ * interrupts. This is a legacy wrapper that can be used in core-processing
-+ * loops but mitigates the performance overhead of portal processing by
-+ * adaptively bypassing true portal processing most of the time. (Processing is
-+ * done once every 10 calls if the previous processing revealed that work needed
-+ * to be done, or once very 1000 calls if the previous processing revealed no
-+ * work needed doing.) If you wish to control this yourself, call
-+ * bman_poll_slow() instead, which always checks for portal processing work.
-+ */
-+void bman_poll(void);
-+
-+/**
-+ * bman_rcr_is_empty - Determine if portal's RCR is empty
-+ *
-+ * For use in situations where a cpu-affine caller needs to determine when all
-+ * releases for the local portal have been processed by Bman but can't use the
-+ * BMAN_RELEASE_FLAG_WAIT_SYNC flag to do this from the final bman_release().
-+ * The function forces tracking of RCR consumption (which normally doesn't
-+ * happen until release processing needs to find space to put new release
-+ * commands), and returns zero if the ring still has unprocessed entries,
-+ * non-zero if it is empty.
-+ */
-+int bman_rcr_is_empty(void);
-+
-+/**
-+ * bman_alloc_bpid_range - Allocate a contiguous range of BPIDs
-+ * @result: is set by the API to the base BPID of the allocated range
-+ * @count: the number of BPIDs required
-+ * @align: required alignment of the allocated range
-+ * @partial: non-zero if the API can return fewer than @count BPIDs
-+ *
-+ * Returns the number of buffer pools allocated, or a negative error code. If
-+ * @partial is non zero, the allocation request may return a smaller range of
-+ * BPs than requested (though alignment will be as requested). If @partial is
-+ * zero, the return value will either be 'count' or negative.
-+ */
-+int bman_alloc_bpid_range(u32 *result, u32 count, u32 align, int partial);
-+static inline int bman_alloc_bpid(u32 *result)
-+{
-+ int ret = bman_alloc_bpid_range(result, 1, 0, 0);
-+ return (ret > 0) ? 0 : ret;
-+}
-+
-+/**
-+ * bman_release_bpid_range - Release the specified range of buffer pool IDs
-+ * @bpid: the base BPID of the range to deallocate
-+ * @count: the number of BPIDs in the range
-+ *
-+ * This function can also be used to seed the allocator with ranges of BPIDs
-+ * that it can subsequently allocate from.
-+ */
-+void bman_release_bpid_range(u32 bpid, unsigned int count);
-+static inline void bman_release_bpid(u32 bpid)
-+{
-+ bman_release_bpid_range(bpid, 1);
-+}
-+
-+int bman_reserve_bpid_range(u32 bpid, unsigned int count);
-+static inline int bman_reserve_bpid(u32 bpid)
-+{
-+ return bman_reserve_bpid_range(bpid, 1);
-+}
-+
-+void bman_seed_bpid_range(u32 bpid, unsigned int count);
-+
-+
-+int bman_shutdown_pool(u32 bpid);
-+
-+ /* Pool management */
-+ /* --------------- */
-+/**
-+ * bman_new_pool - Allocates a Buffer Pool object
-+ * @params: parameters specifying the buffer pool ID and behaviour
-+ *
-+ * Creates a pool object for the given @params. A portal and the depletion
-+ * callback field of @params are only used if the BMAN_POOL_FLAG_DEPLETION flag
-+ * is set. NB, the fields from @params are copied into the new pool object, so
-+ * the structure provided by the caller can be released or reused after the
-+ * function returns.
-+ */
-+struct bman_pool *bman_new_pool(const struct bman_pool_params *params);
-+
-+/**
-+ * bman_free_pool - Deallocates a Buffer Pool object
-+ * @pool: the pool object to release
-+ *
-+ */
-+void bman_free_pool(struct bman_pool *pool);
-+
-+/**
-+ * bman_get_params - Returns a pool object's parameters.
-+ * @pool: the pool object
-+ *
-+ * The returned pointer refers to state within the pool object so must not be
-+ * modified and can no longer be read once the pool object is destroyed.
-+ */
-+const struct bman_pool_params *bman_get_params(const struct bman_pool *pool);
-+
-+/**
-+ * bman_release - Release buffer(s) to the buffer pool
-+ * @pool: the buffer pool object to release to
-+ * @bufs: an array of buffers to release
-+ * @num: the number of buffers in @bufs (1-8)
-+ * @flags: bit-mask of BMAN_RELEASE_FLAG_*** options
-+ *
-+ * Adds the given buffers to RCR entries. If the portal @p was created with the
-+ * "COMPACT" flag, then it will be using a compaction algorithm to improve
-+ * utilisation of RCR. As such, these buffers may join an existing ring entry
-+ * and/or it may not be issued right away so as to allow future releases to join
-+ * the same ring entry. Use the BMAN_RELEASE_FLAG_NOW flag to override this
-+ * behaviour by committing the RCR entry (or entries) right away. If the RCR
-+ * ring is full, the function will return -EBUSY unless BMAN_RELEASE_FLAG_WAIT
-+ * is selected, in which case it will sleep waiting for space to become
-+ * available in RCR. If the function receives a signal before such time (and
-+ * BMAN_RELEASE_FLAG_WAIT_INT is set), the function returns -EINTR. Otherwise,
-+ * it returns zero.
-+ */
-+int bman_release(struct bman_pool *pool, const struct bm_buffer *bufs, u8 num,
-+ u32 flags);
-+
-+/**
-+ * bman_acquire - Acquire buffer(s) from a buffer pool
-+ * @pool: the buffer pool object to acquire from
-+ * @bufs: array for storing the acquired buffers
-+ * @num: the number of buffers desired (@bufs is at least this big)
-+ *
-+ * Issues an "Acquire" command via the portal's management command interface.
-+ * The return value will be the number of buffers obtained from the pool, or a
-+ * negative error code if a h/w error or pool starvation was encountered. In
-+ * the latter case, the content of @bufs is undefined.
-+ */
-+int bman_acquire(struct bman_pool *pool, struct bm_buffer *bufs, u8 num,
-+ u32 flags);
-+
-+/**
-+ * bman_flush_stockpile - Flush stockpile buffer(s) to the buffer pool
-+ * @pool: the buffer pool object the stockpile belongs
-+ * @flags: bit-mask of BMAN_RELEASE_FLAG_*** options
-+ *
-+ * Adds stockpile buffers to RCR entries until the stockpile is empty.
-+ * The return value will be a negative error code if a h/w error occurred.
-+ * If BMAN_RELEASE_FLAG_NOW flag is passed and RCR ring is full,
-+ * -EAGAIN will be returned.
-+ */
-+int bman_flush_stockpile(struct bman_pool *pool, u32 flags);
-+
-+/**
-+ * bman_query_pools - Query all buffer pool states
-+ * @state: storage for the queried availability and depletion states
-+ */
-+int bman_query_pools(struct bm_pool_state *state);
-+
-+#ifdef CONFIG_FSL_BMAN_CONFIG
-+/**
-+ * bman_query_free_buffers - Query how many free buffers are in buffer pool
-+ * @pool: the buffer pool object to query
-+ *
-+ * Return the number of the free buffers
-+ */
-+u32 bman_query_free_buffers(struct bman_pool *pool);
-+
-+/**
-+ * bman_update_pool_thresholds - Change the buffer pool's depletion thresholds
-+ * @pool: the buffer pool object to which the thresholds will be set
-+ * @thresholds: the new thresholds
-+ */
-+int bman_update_pool_thresholds(struct bman_pool *pool, const u32 *thresholds);
-+#endif
-+
-+/**
-+ * The below bman_p_***() variant might be called in a situation that the cpu
-+ * which the portal affine to is not online yet.
-+ * @bman_portal specifies which portal the API will use.
-+*/
-+int bman_p_irqsource_add(struct bman_portal *p, __maybe_unused u32 bits);
-+#ifdef __cplusplus
-+}
-+#endif
-+
-+#endif /* FSL_BMAN_H */
---- /dev/null
-+++ b/include/linux/fsl_qman.h
-@@ -0,0 +1,3900 @@
-+/* Copyright 2008-2012 Freescale Semiconductor, Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#ifndef FSL_QMAN_H
-+#define FSL_QMAN_H
-+
-+#ifdef __cplusplus
-+extern "C" {
-+#endif
-+
-+/* Last updated for v00.800 of the BG */
-+
-+/* Hardware constants */
-+#define QM_CHANNEL_SWPORTAL0 0
-+#define QMAN_CHANNEL_POOL1 0x21
-+#define QMAN_CHANNEL_CAAM 0x80
-+#define QMAN_CHANNEL_PME 0xa0
-+#define QMAN_CHANNEL_POOL1_REV3 0x401
-+#define QMAN_CHANNEL_CAAM_REV3 0x840
-+#define QMAN_CHANNEL_PME_REV3 0x860
-+#define QMAN_CHANNEL_DCE 0x8a0
-+#define QMAN_CHANNEL_DCE_QMANREV312 0x880
-+extern u16 qm_channel_pool1;
-+extern u16 qm_channel_caam;
-+extern u16 qm_channel_pme;
-+extern u16 qm_channel_dce;
-+enum qm_dc_portal {
-+ qm_dc_portal_fman0 = 0,
-+ qm_dc_portal_fman1 = 1,
-+ qm_dc_portal_caam = 2,
-+ qm_dc_portal_pme = 3,
-+ qm_dc_portal_rman = 4,
-+ qm_dc_portal_dce = 5
-+};
-+
-+/* Portal processing (interrupt) sources */
-+#define QM_PIRQ_CCSCI 0x00200000 /* CEETM Congestion State Change */
-+#define QM_PIRQ_CSCI 0x00100000 /* Congestion State Change */
-+#define QM_PIRQ_EQCI 0x00080000 /* Enqueue Command Committed */
-+#define QM_PIRQ_EQRI 0x00040000 /* EQCR Ring (below threshold) */
-+#define QM_PIRQ_DQRI 0x00020000 /* DQRR Ring (non-empty) */
-+#define QM_PIRQ_MRI 0x00010000 /* MR Ring (non-empty) */
-+/* This mask contains all the interrupt sources that need handling except DQRI,
-+ * ie. that if present should trigger slow-path processing. */
-+#define QM_PIRQ_SLOW (QM_PIRQ_CSCI | QM_PIRQ_EQCI | QM_PIRQ_EQRI | \
-+ QM_PIRQ_MRI | QM_PIRQ_CCSCI)
-+
-+/* --- Clock speed --- */
-+/* A qman driver instance may or may not know the current qman clock speed.
-+ * However, certain CEETM calculations may not be possible if this is not known.
-+ * The 'set' function will only succeed (return zero) if the driver did not
-+ * already know the clock speed. Likewise, the 'get' function will only succeed
-+ * if the driver does know the clock speed (either because it knew when booting,
-+ * or was told via 'set'). In cases where software is running on a driver
-+ * instance that does not know the clock speed (eg. on a hypervised data-plane),
-+ * and the user can obtain the current qman clock speed by other means (eg. from
-+ * a message sent from the control-plane), then the 'set' function can be used
-+ * to enable rate-calculations in a driver where it would otherwise not be
-+ * possible. */
-+int qm_get_clock(u64 *clock_hz);
-+int qm_set_clock(u64 clock_hz);
-+
-+/* For qman_static_dequeue_*** APIs */
-+#define QM_SDQCR_CHANNELS_POOL_MASK 0x00007fff
-+/* for n in [1,15] */
-+#define QM_SDQCR_CHANNELS_POOL(n) (0x00008000 >> (n))
-+/* for conversion from n of qm_channel */
-+static inline u32 QM_SDQCR_CHANNELS_POOL_CONV(u16 channel)
-+{
-+ return QM_SDQCR_CHANNELS_POOL(channel + 1 - qm_channel_pool1);
-+}
-+
-+/* For qman_volatile_dequeue(); Choose one PRECEDENCE. EXACT is optional. Use
-+ * NUMFRAMES(n) (6-bit) or NUMFRAMES_TILLEMPTY to fill in the frame-count. Use
-+ * FQID(n) to fill in the frame queue ID. */
-+#define QM_VDQCR_PRECEDENCE_VDQCR 0x0
-+#define QM_VDQCR_PRECEDENCE_SDQCR 0x80000000
-+#define QM_VDQCR_EXACT 0x40000000
-+#define QM_VDQCR_NUMFRAMES_MASK 0x3f000000
-+#define QM_VDQCR_NUMFRAMES_SET(n) (((n) & 0x3f) << 24)
-+#define QM_VDQCR_NUMFRAMES_GET(n) (((n) >> 24) & 0x3f)
-+#define QM_VDQCR_NUMFRAMES_TILLEMPTY QM_VDQCR_NUMFRAMES_SET(0)
-+
-+
-+/* ------------------------------------------------------- */
-+/* --- Qman data structures (and associated constants) --- */
-+
-+/* Represents s/w corenet portal mapped data structures */
-+struct qm_eqcr_entry; /* EQCR (EnQueue Command Ring) entries */
-+struct qm_dqrr_entry; /* DQRR (DeQueue Response Ring) entries */
-+struct qm_mr_entry; /* MR (Message Ring) entries */
-+struct qm_mc_command; /* MC (Management Command) command */
-+struct qm_mc_result; /* MC result */
-+
-+/* See David Lapp's "Frame formats" document, "dpateam", Jan 07, 2008 */
-+#define QM_FD_FORMAT_SG 0x4
-+#define QM_FD_FORMAT_LONG 0x2
-+#define QM_FD_FORMAT_COMPOUND 0x1
-+enum qm_fd_format {
-+ /* 'contig' implies a contiguous buffer, whereas 'sg' implies a
-+ * scatter-gather table. 'big' implies a 29-bit length with no offset
-+ * field, otherwise length is 20-bit and offset is 9-bit. 'compound'
-+ * implies a s/g-like table, where each entry itself represents a frame
-+ * (contiguous or scatter-gather) and the 29-bit "length" is
-+ * interpreted purely for congestion calculations, ie. a "congestion
-+ * weight". */
-+ qm_fd_contig = 0,
-+ qm_fd_contig_big = QM_FD_FORMAT_LONG,
-+ qm_fd_sg = QM_FD_FORMAT_SG,
-+ qm_fd_sg_big = QM_FD_FORMAT_SG | QM_FD_FORMAT_LONG,
-+ qm_fd_compound = QM_FD_FORMAT_COMPOUND
-+};
-+
-+/* Capitalised versions are un-typed but can be used in static expressions */
-+#define QM_FD_CONTIG 0
-+#define QM_FD_CONTIG_BIG QM_FD_FORMAT_LONG
-+#define QM_FD_SG QM_FD_FORMAT_SG
-+#define QM_FD_SG_BIG (QM_FD_FORMAT_SG | QM_FD_FORMAT_LONG)
-+#define QM_FD_COMPOUND QM_FD_FORMAT_COMPOUND
-+
-+/* See 1.5.1.1: "Frame Descriptor (FD)" */
-+struct qm_fd {
-+ union {
-+ struct {
-+#if __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
-+ u8 dd:2; /* dynamic debug */
-+ u8 liodn_offset:6;
-+ u8 bpid:8; /* Buffer Pool ID */
-+ u8 eliodn_offset:4;
-+ u8 __reserved:4;
-+ u8 addr_hi; /* high 8-bits of 40-bit address */
-+ u32 addr_lo; /* low 32-bits of 40-bit address */
-+#else
-+ u32 addr_lo; /* low 32-bits of 40-bit address */
-+ u8 addr_hi; /* high 8-bits of 40-bit address */
-+ u8 __reserved:4;
-+ u8 eliodn_offset:4;
-+ u8 bpid:8; /* Buffer Pool ID */
-+ u8 liodn_offset:6;
-+ u8 dd:2; /* dynamic debug */
-+#endif
-+ };
-+ struct {
-+#if __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
-+ u64 __notaddress:24;
-+ u64 addr:40;
-+#else
-+ u64 addr:40;
-+ u64 __notaddress:24;
-+#endif
-+ };
-+ u64 opaque_addr;
-+ };
-+ /* The 'format' field indicates the interpretation of the remaining 29
-+ * bits of the 32-bit word. For packing reasons, it is duplicated in the
-+ * other union elements. Note, union'd structs are difficult to use with
-+ * static initialisation under gcc, in which case use the "opaque" form
-+ * with one of the macros. */
-+ union {
-+ /* For easier/faster copying of this part of the fd (eg. from a
-+ * DQRR entry to an EQCR entry) copy 'opaque' */
-+ u32 opaque;
-+ /* If 'format' is _contig or _sg, 20b length and 9b offset */
-+ struct {
-+#if __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
-+ enum qm_fd_format format:3;
-+ u16 offset:9;
-+ u32 length20:20;
-+#else
-+ u32 length20:20;
-+ u16 offset:9;
-+ enum qm_fd_format format:3;
-+#endif
-+ };
-+ /* If 'format' is _contig_big or _sg_big, 29b length */
-+ struct {
-+#if __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
-+ enum qm_fd_format _format1:3;
-+ u32 length29:29;
-+#else
-+ u32 length29:29;
-+ enum qm_fd_format _format1:3;
-+#endif
-+ };
-+ /* If 'format' is _compound, 29b "congestion weight" */
-+ struct {
-+#if __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
-+ enum qm_fd_format _format2:3;
-+ u32 cong_weight:29;
-+#else
-+ u32 cong_weight:29;
-+ enum qm_fd_format _format2:3;
-+#endif
-+ };
-+ };
-+ union {
-+ u32 cmd;
-+ u32 status;
-+ };
-+} __aligned(8);
-+#define QM_FD_DD_NULL 0x00
-+#define QM_FD_PID_MASK 0x3f
-+static inline u64 qm_fd_addr_get64(const struct qm_fd *fd)
-+{
-+ return fd->addr;
-+}
-+
-+static inline dma_addr_t qm_fd_addr(const struct qm_fd *fd)
-+{
-+ return (dma_addr_t)fd->addr;
-+}
-+/* Macro, so we compile better if 'v' isn't always 64-bit */
-+#define qm_fd_addr_set64(fd, v) \
-+ do { \
-+ struct qm_fd *__fd931 = (fd); \
-+ __fd931->addr = v; \
-+ } while (0)
-+
-+/* For static initialisation of FDs (which is complicated by the use of unions
-+ * in "struct qm_fd"), use the following macros. Note that;
-+ * - 'dd', 'pid' and 'bpid' are ignored because there's no static initialisation
-+ * use-case),
-+ * - use capitalised QM_FD_*** formats for static initialisation.
-+ */
-+#define QM_FD_FMT_20(cmd, addr_hi, addr_lo, fmt, off, len) \
-+ { 0, 0, 0, 0, 0, addr_hi, addr_lo, \
-+ { (((fmt)&0x7) << 29) | (((off)&0x1ff) << 20) | ((len)&0xfffff) }, \
-+ { cmd } }
-+#define QM_FD_FMT_29(cmd, addr_hi, addr_lo, fmt, len) \
-+ { 0, 0, 0, 0, 0, addr_hi, addr_lo, \
-+ { (((fmt)&0x7) << 29) | ((len)&0x1fffffff) }, \
-+ { cmd } }
-+
-+/* See 2.2.1.3 Multi-Core Datapath Acceleration Architecture */
-+#define QM_SG_OFFSET_MASK 0x1FFF
-+struct qm_sg_entry {
-+ union {
-+ struct {
-+#if __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
-+ u8 __reserved1[3];
-+ u8 addr_hi; /* high 8-bits of 40-bit address */
-+ u32 addr_lo; /* low 32-bits of 40-bit address */
-+#else
-+ u32 addr_lo; /* low 32-bits of 40-bit address */
-+ u8 addr_hi; /* high 8-bits of 40-bit address */
-+ u8 __reserved1[3];
-+#endif
-+ };
-+ struct {
-+#if __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
-+ u64 __notaddress:24;
-+ u64 addr:40;
-+#else
-+ u64 addr:40;
-+ u64 __notaddress:24;
-+#endif
-+ };
-+ u64 opaque;
-+ };
-+ union {
-+ struct {
-+#if __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
-+ u32 extension:1; /* Extension bit */
-+ u32 final:1; /* Final bit */
-+ u32 length:30;
-+#else
-+ u32 length:30;
-+ u32 final:1; /* Final bit */
-+ u32 extension:1; /* Extension bit */
-+#endif
-+ };
-+ u32 sgt_efl;
-+ };
-+ u8 __reserved2;
-+ u8 bpid;
-+ union {
-+ struct {
-+#if __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
-+ u16 __reserved3:3;
-+ u16 offset:13;
-+#else
-+ u16 offset:13;
-+ u16 __reserved3:3;
-+#endif
-+ };
-+ u16 opaque_offset;
-+ };
-+} __packed;
-+union qm_sg_efl {
-+ struct {
-+#if __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
-+ u32 extension:1; /* Extension bit */
-+ u32 final:1; /* Final bit */
-+ u32 length:30;
-+#else
-+ u32 length:30;
-+ u32 final:1; /* Final bit */
-+ u32 extension:1; /* Extension bit */
-+#endif
-+ };
-+ u32 efl;
-+};
-+static inline dma_addr_t qm_sg_addr(const struct qm_sg_entry *sg)
-+{
-+ return (dma_addr_t)be64_to_cpu(sg->opaque) & 0xffffffffffULL;
-+}
-+static inline u8 qm_sg_entry_get_ext(const struct qm_sg_entry *sg)
-+{
-+ union qm_sg_efl u;
-+
-+ u.efl = be32_to_cpu(sg->sgt_efl);
-+ return u.extension;
-+}
-+static inline u8 qm_sg_entry_get_final(const struct qm_sg_entry *sg)
-+{
-+ union qm_sg_efl u;
-+
-+ u.efl = be32_to_cpu(sg->sgt_efl);
-+ return u.final;
-+}
-+static inline u32 qm_sg_entry_get_len(const struct qm_sg_entry *sg)
-+{
-+ union qm_sg_efl u;
-+
-+ u.efl = be32_to_cpu(sg->sgt_efl);
-+ return u.length;
-+}
-+static inline u8 qm_sg_entry_get_bpid(const struct qm_sg_entry *sg)
-+{
-+ return sg->bpid;
-+}
-+static inline u16 qm_sg_entry_get_offset(const struct qm_sg_entry *sg)
-+{
-+ u32 opaque_offset = be16_to_cpu(sg->opaque_offset);
-+
-+ return opaque_offset & 0x1fff;
-+}
-+
-+/* Macro, so we compile better if 'v' isn't always 64-bit */
-+#define qm_sg_entry_set64(sg, v) \
-+ do { \
-+ struct qm_sg_entry *__sg931 = (sg); \
-+ __sg931->opaque = cpu_to_be64(v); \
-+ } while (0)
-+#define qm_sg_entry_set_ext(sg, v) \
-+ do { \
-+ union qm_sg_efl __u932; \
-+ __u932.efl = be32_to_cpu((sg)->sgt_efl); \
-+ __u932.extension = v; \
-+ (sg)->sgt_efl = cpu_to_be32(__u932.efl); \
-+ } while (0)
-+#define qm_sg_entry_set_final(sg, v) \
-+ do { \
-+ union qm_sg_efl __u933; \
-+ __u933.efl = be32_to_cpu((sg)->sgt_efl); \
-+ __u933.final = v; \
-+ (sg)->sgt_efl = cpu_to_be32(__u933.efl); \
-+ } while (0)
-+#define qm_sg_entry_set_len(sg, v) \
-+ do { \
-+ union qm_sg_efl __u934; \
-+ __u934.efl = be32_to_cpu((sg)->sgt_efl); \
-+ __u934.length = v; \
-+ (sg)->sgt_efl = cpu_to_be32(__u934.efl); \
-+ } while (0)
-+#define qm_sg_entry_set_bpid(sg, v) \
-+ do { \
-+ struct qm_sg_entry *__u935 = (sg); \
-+ __u935->bpid = v; \
-+ } while (0)
-+#define qm_sg_entry_set_offset(sg, v) \
-+ do { \
-+ struct qm_sg_entry *__u936 = (sg); \
-+ __u936->opaque_offset = cpu_to_be16(v); \
-+ } while (0)
-+
-+/* See 1.5.8.1: "Enqueue Command" */
-+struct qm_eqcr_entry {
-+ u8 __dont_write_directly__verb;
-+ u8 dca;
-+ u16 seqnum;
-+ u32 orp; /* 24-bit */
-+ u32 fqid; /* 24-bit */
-+ u32 tag;
-+ struct qm_fd fd;
-+ u8 __reserved3[32];
-+} __packed;
-+#define QM_EQCR_VERB_VBIT 0x80
-+#define QM_EQCR_VERB_CMD_MASK 0x61 /* but only one value; */
-+#define QM_EQCR_VERB_CMD_ENQUEUE 0x01
-+#define QM_EQCR_VERB_COLOUR_MASK 0x18 /* 4 possible values; */
-+#define QM_EQCR_VERB_COLOUR_GREEN 0x00
-+#define QM_EQCR_VERB_COLOUR_YELLOW 0x08
-+#define QM_EQCR_VERB_COLOUR_RED 0x10
-+#define QM_EQCR_VERB_COLOUR_OVERRIDE 0x18
-+#define QM_EQCR_VERB_INTERRUPT 0x04 /* on command consumption */
-+#define QM_EQCR_VERB_ORP 0x02 /* enable order restoration */
-+#define QM_EQCR_DCA_ENABLE 0x80
-+#define QM_EQCR_DCA_PARK 0x40
-+#define QM_EQCR_DCA_IDXMASK 0x0f /* "DQRR::idx" goes here */
-+#define QM_EQCR_SEQNUM_NESN 0x8000 /* Advance NESN */
-+#define QM_EQCR_SEQNUM_NLIS 0x4000 /* More fragments to come */
-+#define QM_EQCR_SEQNUM_SEQMASK 0x3fff /* sequence number goes here */
-+#define QM_EQCR_FQID_NULL 0 /* eg. for an ORP seqnum hole */
-+
-+/* See 1.5.8.2: "Frame Dequeue Response" */
-+struct qm_dqrr_entry {
-+ u8 verb;
-+ u8 stat;
-+ u16 seqnum; /* 15-bit */
-+ u8 tok;
-+ u8 __reserved2[3];
-+ u32 fqid; /* 24-bit */
-+ u32 contextB;
-+ struct qm_fd fd;
-+ u8 __reserved4[32];
-+};
-+#define QM_DQRR_VERB_VBIT 0x80
-+#define QM_DQRR_VERB_MASK 0x7f /* where the verb contains; */
-+#define QM_DQRR_VERB_FRAME_DEQUEUE 0x60 /* "this format" */
-+#define QM_DQRR_STAT_FQ_EMPTY 0x80 /* FQ empty */
-+#define QM_DQRR_STAT_FQ_HELDACTIVE 0x40 /* FQ held active */
-+#define QM_DQRR_STAT_FQ_FORCEELIGIBLE 0x20 /* FQ was force-eligible'd */
-+#define QM_DQRR_STAT_FD_VALID 0x10 /* has a non-NULL FD */
-+#define QM_DQRR_STAT_UNSCHEDULED 0x02 /* Unscheduled dequeue */
-+#define QM_DQRR_STAT_DQCR_EXPIRED 0x01 /* VDQCR or PDQCR expired*/
-+
-+/* See 1.5.8.3: "ERN Message Response" */
-+/* See 1.5.8.4: "FQ State Change Notification" */
-+struct qm_mr_entry {
-+ u8 verb;
-+ union {
-+ struct {
-+ u8 dca;
-+ u16 seqnum;
-+ u8 rc; /* Rejection Code */
-+ u32 orp:24;
-+ u32 fqid; /* 24-bit */
-+ u32 tag;
-+ struct qm_fd fd;
-+ } __packed ern;
-+ struct {
-+#if __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
-+ u8 colour:2; /* See QM_MR_DCERN_COLOUR_* */
-+ u8 __reserved1:3;
-+ enum qm_dc_portal portal:3;
-+#else
-+ enum qm_dc_portal portal:3;
-+ u8 __reserved1:3;
-+ u8 colour:2; /* See QM_MR_DCERN_COLOUR_* */
-+#endif
-+ u16 __reserved2;
-+ u8 rc; /* Rejection Code */
-+ u32 __reserved3:24;
-+ u32 fqid; /* 24-bit */
-+ u32 tag;
-+ struct qm_fd fd;
-+ } __packed dcern;
-+ struct {
-+ u8 fqs; /* Frame Queue Status */
-+ u8 __reserved1[6];
-+ u32 fqid; /* 24-bit */
-+ u32 contextB;
-+ u8 __reserved2[16];
-+ } __packed fq; /* FQRN/FQRNI/FQRL/FQPN */
-+ };
-+ u8 __reserved2[32];
-+} __packed;
-+#define QM_MR_VERB_VBIT 0x80
-+/* The "ern" VERB bits match QM_EQCR_VERB_*** so aren't reproduced here. ERNs
-+ * originating from direct-connect portals ("dcern") use 0x20 as a verb which
-+ * would be invalid as a s/w enqueue verb. A s/w ERN can be distinguished from
-+ * the other MR types by noting if the 0x20 bit is unset. */
-+#define QM_MR_VERB_TYPE_MASK 0x27
-+#define QM_MR_VERB_DC_ERN 0x20
-+#define QM_MR_VERB_FQRN 0x21
-+#define QM_MR_VERB_FQRNI 0x22
-+#define QM_MR_VERB_FQRL 0x23
-+#define QM_MR_VERB_FQPN 0x24
-+#define QM_MR_RC_MASK 0xf0 /* contains one of; */
-+#define QM_MR_RC_CGR_TAILDROP 0x00
-+#define QM_MR_RC_WRED 0x10
-+#define QM_MR_RC_ERROR 0x20
-+#define QM_MR_RC_ORPWINDOW_EARLY 0x30
-+#define QM_MR_RC_ORPWINDOW_LATE 0x40
-+#define QM_MR_RC_FQ_TAILDROP 0x50
-+#define QM_MR_RC_ORPWINDOW_RETIRED 0x60
-+#define QM_MR_RC_ORP_ZERO 0x70
-+#define QM_MR_FQS_ORLPRESENT 0x02 /* ORL fragments to come */
-+#define QM_MR_FQS_NOTEMPTY 0x01 /* FQ has enqueued frames */
-+#define QM_MR_DCERN_COLOUR_GREEN 0x00
-+#define QM_MR_DCERN_COLOUR_YELLOW 0x01
-+#define QM_MR_DCERN_COLOUR_RED 0x02
-+#define QM_MR_DCERN_COLOUR_OVERRIDE 0x03
-+
-+/* An identical structure of FQD fields is present in the "Init FQ" command and
-+ * the "Query FQ" result, it's suctioned out into the "struct qm_fqd" type.
-+ * Within that, the 'stashing' and 'taildrop' pieces are also factored out, the
-+ * latter has two inlines to assist with converting to/from the mant+exp
-+ * representation. */
-+struct qm_fqd_stashing {
-+ /* See QM_STASHING_EXCL_<...> */
-+#if __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
-+ u8 exclusive;
-+ u8 __reserved1:2;
-+ /* Numbers of cachelines */
-+ u8 annotation_cl:2;
-+ u8 data_cl:2;
-+ u8 context_cl:2;
-+#else
-+ u8 context_cl:2;
-+ u8 data_cl:2;
-+ u8 annotation_cl:2;
-+ u8 __reserved1:2;
-+ u8 exclusive;
-+#endif
-+} __packed;
-+struct qm_fqd_taildrop {
-+#if __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
-+ u16 __reserved1:3;
-+ u16 mant:8;
-+ u16 exp:5;
-+#else
-+ u16 exp:5;
-+ u16 mant:8;
-+ u16 __reserved1:3;
-+#endif
-+} __packed;
-+struct qm_fqd_oac {
-+ /* See QM_OAC_<...> */
-+#if __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
-+ u8 oac:2; /* "Overhead Accounting Control" */
-+ u8 __reserved1:6;
-+#else
-+ u8 __reserved1:6;
-+ u8 oac:2; /* "Overhead Accounting Control" */
-+#endif
-+ /* Two's-complement value (-128 to +127) */
-+ signed char oal; /* "Overhead Accounting Length" */
-+} __packed;
-+struct qm_fqd {
-+ union {
-+ u8 orpc;
-+ struct {
-+#if __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
-+ u8 __reserved1:2;
-+ u8 orprws:3;
-+ u8 oa:1;
-+ u8 olws:2;
-+#else
-+ u8 olws:2;
-+ u8 oa:1;
-+ u8 orprws:3;
-+ u8 __reserved1:2;
-+#endif
-+ } __packed;
-+ };
-+ u8 cgid;
-+ u16 fq_ctrl; /* See QM_FQCTRL_<...> */
-+ union {
-+ u16 dest_wq;
-+ struct {
-+#if __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
-+ u16 channel:13; /* qm_channel */
-+ u16 wq:3;
-+#else
-+ u16 wq:3;
-+ u16 channel:13; /* qm_channel */
-+#endif
-+ } __packed dest;
-+ };
-+#if __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
-+ u16 __reserved2:1;
-+ u16 ics_cred:15;
-+#else
-+ u16 __reserved2:1;
-+ u16 ics_cred:15;
-+#endif
-+ /* For "Initialize Frame Queue" commands, the write-enable mask
-+ * determines whether 'td' or 'oac_init' is observed. For query
-+ * commands, this field is always 'td', and 'oac_query' (below) reflects
-+ * the Overhead ACcounting values. */
-+ union {
-+ struct qm_fqd_taildrop td;
-+ struct qm_fqd_oac oac_init;
-+ };
-+ u32 context_b;
-+ union {
-+ /* Treat it as 64-bit opaque */
-+ u64 opaque;
-+ struct {
-+#if __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
-+ u32 hi;
-+ u32 lo;
-+#else
-+ u32 lo;
-+ u32 hi;
-+#endif
-+ };
-+ /* Treat it as s/w portal stashing config */
-+ /* See 1.5.6.7.1: "FQD Context_A field used for [...] */
-+ struct {
-+#if __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
-+ struct qm_fqd_stashing stashing;
-+ /* 48-bit address of FQ context to
-+ * stash, must be cacheline-aligned */
-+ u16 context_hi;
-+ u32 context_lo;
-+#else
-+ u32 context_lo;
-+ u16 context_hi;
-+ struct qm_fqd_stashing stashing;
-+#endif
-+ } __packed;
-+ } context_a;
-+ struct qm_fqd_oac oac_query;
-+} __packed;
-+/* 64-bit converters for context_hi/lo */
-+static inline u64 qm_fqd_stashing_get64(const struct qm_fqd *fqd)
-+{
-+ return ((u64)fqd->context_a.context_hi << 32) |
-+ (u64)fqd->context_a.context_lo;
-+}
-+static inline dma_addr_t qm_fqd_stashing_addr(const struct qm_fqd *fqd)
-+{
-+ return (dma_addr_t)qm_fqd_stashing_get64(fqd);
-+}
-+static inline u64 qm_fqd_context_a_get64(const struct qm_fqd *fqd)
-+{
-+ return ((u64)fqd->context_a.hi << 32) |
-+ (u64)fqd->context_a.lo;
-+}
-+/* Macro, so we compile better when 'v' isn't necessarily 64-bit */
-+#define qm_fqd_stashing_set64(fqd, v) \
-+ do { \
-+ struct qm_fqd *__fqd931 = (fqd); \
-+ __fqd931->context_a.context_hi = upper_32_bits(v); \
-+ __fqd931->context_a.context_lo = lower_32_bits(v); \
-+ } while (0)
-+#define qm_fqd_context_a_set64(fqd, v) \
-+ do { \
-+ struct qm_fqd *__fqd931 = (fqd); \
-+ __fqd931->context_a.hi = upper_32_bits(v); \
-+ __fqd931->context_a.lo = lower_32_bits(v); \
-+ } while (0)
-+/* convert a threshold value into mant+exp representation */
-+static inline int qm_fqd_taildrop_set(struct qm_fqd_taildrop *td, u32 val,
-+ int roundup)
-+{
-+ u32 e = 0;
-+ int oddbit = 0;
-+ if (val > 0xe0000000)
-+ return -ERANGE;
-+ while (val > 0xff) {
-+ oddbit = val & 1;
-+ val >>= 1;
-+ e++;
-+ if (roundup && oddbit)
-+ val++;
-+ }
-+ td->exp = e;
-+ td->mant = val;
-+ return 0;
-+}
-+/* and the other direction */
-+static inline u32 qm_fqd_taildrop_get(const struct qm_fqd_taildrop *td)
-+{
-+ return (u32)td->mant << td->exp;
-+}
-+
-+/* See 1.5.2.2: "Frame Queue Descriptor (FQD)" */
-+/* Frame Queue Descriptor (FQD) field 'fq_ctrl' uses these constants */
-+#define QM_FQCTRL_MASK 0x07ff /* 'fq_ctrl' flags; */
-+#define QM_FQCTRL_CGE 0x0400 /* Congestion Group Enable */
-+#define QM_FQCTRL_TDE 0x0200 /* Tail-Drop Enable */
-+#define QM_FQCTRL_ORP 0x0100 /* ORP Enable */
-+#define QM_FQCTRL_CTXASTASHING 0x0080 /* Context-A stashing */
-+#define QM_FQCTRL_CPCSTASH 0x0040 /* CPC Stash Enable */
-+#define QM_FQCTRL_FORCESFDR 0x0008 /* High-priority SFDRs */
-+#define QM_FQCTRL_AVOIDBLOCK 0x0004 /* Don't block active */
-+#define QM_FQCTRL_HOLDACTIVE 0x0002 /* Hold active in portal */
-+#define QM_FQCTRL_PREFERINCACHE 0x0001 /* Aggressively cache FQD */
-+#define QM_FQCTRL_LOCKINCACHE QM_FQCTRL_PREFERINCACHE /* older naming */
-+
-+/* See 1.5.6.7.1: "FQD Context_A field used for [...] */
-+/* Frame Queue Descriptor (FQD) field 'CONTEXT_A' uses these constants */
-+#define QM_STASHING_EXCL_ANNOTATION 0x04
-+#define QM_STASHING_EXCL_DATA 0x02
-+#define QM_STASHING_EXCL_CTX 0x01
-+
-+/* See 1.5.5.3: "Intra Class Scheduling" */
-+/* FQD field 'OAC' (Overhead ACcounting) uses these constants */
-+#define QM_OAC_ICS 0x2 /* Accounting for Intra-Class Scheduling */
-+#define QM_OAC_CG 0x1 /* Accounting for Congestion Groups */
-+
-+/* See 1.5.8.4: "FQ State Change Notification" */
-+/* This struct represents the 32-bit "WR_PARM_[GYR]" parameters in CGR fields
-+ * and associated commands/responses. The WRED parameters are calculated from
-+ * these fields as follows;
-+ * MaxTH = MA * (2 ^ Mn)
-+ * Slope = SA / (2 ^ Sn)
-+ * MaxP = 4 * (Pn + 1)
-+ */
-+struct qm_cgr_wr_parm {
-+ union {
-+ u32 word;
-+ struct {
-+#if __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
-+ u32 MA:8;
-+ u32 Mn:5;
-+ u32 SA:7; /* must be between 64-127 */
-+ u32 Sn:6;
-+ u32 Pn:6;
-+#else
-+ u32 Pn:6;
-+ u32 Sn:6;
-+ u32 SA:7; /* must be between 64-127 */
-+ u32 Mn:5;
-+ u32 MA:8;
-+#endif
-+ } __packed;
-+ };
-+} __packed;
-+/* This struct represents the 13-bit "CS_THRES" CGR field. In the corresponding
-+ * management commands, this is padded to a 16-bit structure field, so that's
-+ * how we represent it here. The congestion state threshold is calculated from
-+ * these fields as follows;
-+ * CS threshold = TA * (2 ^ Tn)
-+ */
-+struct qm_cgr_cs_thres {
-+ union {
-+ u16 hword;
-+ struct {
-+#if __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
-+ u16 __reserved:3;
-+ u16 TA:8;
-+ u16 Tn:5;
-+#else
-+ u16 Tn:5;
-+ u16 TA:8;
-+ u16 __reserved:3;
-+#endif
-+ } __packed;
-+ };
-+} __packed;
-+/* This identical structure of CGR fields is present in the "Init/Modify CGR"
-+ * commands and the "Query CGR" result. It's suctioned out here into its own
-+ * struct. */
-+struct __qm_mc_cgr {
-+ struct qm_cgr_wr_parm wr_parm_g;
-+ struct qm_cgr_wr_parm wr_parm_y;
-+ struct qm_cgr_wr_parm wr_parm_r;
-+ u8 wr_en_g; /* boolean, use QM_CGR_EN */
-+ u8 wr_en_y; /* boolean, use QM_CGR_EN */
-+ u8 wr_en_r; /* boolean, use QM_CGR_EN */
-+ u8 cscn_en; /* boolean, use QM_CGR_EN */
-+ union {
-+ struct {
-+#if __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
-+ u16 cscn_targ_upd_ctrl; /* use QM_CSCN_TARG_UDP_ */
-+ u16 cscn_targ_dcp_low; /* CSCN_TARG_DCP low-16bits */
-+#else
-+ u16 cscn_targ_dcp_low; /* CSCN_TARG_DCP low-16bits */
-+ u16 cscn_targ_upd_ctrl; /* use QM_CSCN_TARG_UDP_ */
-+#endif
-+ };
-+ u32 cscn_targ; /* use QM_CGR_TARG_* */
-+ };
-+ u8 cstd_en; /* boolean, use QM_CGR_EN */
-+ u8 cs; /* boolean, only used in query response */
-+ union {
-+ /* use qm_cgr_cs_thres_set64() */
-+ struct qm_cgr_cs_thres cs_thres;
-+ u16 __cs_thres;
-+ };
-+ u8 mode; /* QMAN_CGR_MODE_FRAME not supported in rev1.0 */
-+} __packed;
-+#define QM_CGR_EN 0x01 /* For wr_en_*, cscn_en, cstd_en */
-+#define QM_CGR_TARG_UDP_CTRL_WRITE_BIT 0x8000 /* value written to portal bit*/
-+#define QM_CGR_TARG_UDP_CTRL_DCP 0x4000 /* 0: SWP, 1: DCP */
-+#define QM_CGR_TARG_PORTAL(n) (0x80000000 >> (n)) /* s/w portal, 0-9 */
-+#define QM_CGR_TARG_FMAN0 0x00200000 /* direct-connect portal: fman0 */
-+#define QM_CGR_TARG_FMAN1 0x00100000 /* : fman1 */
-+/* Convert CGR thresholds to/from "cs_thres" format */
-+static inline u64 qm_cgr_cs_thres_get64(const struct qm_cgr_cs_thres *th)
-+{
-+ return (u64)th->TA << th->Tn;
-+}
-+static inline int qm_cgr_cs_thres_set64(struct qm_cgr_cs_thres *th, u64 val,
-+ int roundup)
-+{
-+ u32 e = 0;
-+ int oddbit = 0;
-+ while (val > 0xff) {
-+ oddbit = val & 1;
-+ val >>= 1;
-+ e++;
-+ if (roundup && oddbit)
-+ val++;
-+ }
-+ th->Tn = e;
-+ th->TA = val;
-+ return 0;
-+}
-+
-+/* See 1.5.8.5.1: "Initialize FQ" */
-+/* See 1.5.8.5.2: "Query FQ" */
-+/* See 1.5.8.5.3: "Query FQ Non-Programmable Fields" */
-+/* See 1.5.8.5.4: "Alter FQ State Commands " */
-+/* See 1.5.8.6.1: "Initialize/Modify CGR" */
-+/* See 1.5.8.6.2: "CGR Test Write" */
-+/* See 1.5.8.6.3: "Query CGR" */
-+/* See 1.5.8.6.4: "Query Congestion Group State" */
-+struct qm_mcc_initfq {
-+ u8 __reserved1;
-+ u16 we_mask; /* Write Enable Mask */
-+ u32 fqid; /* 24-bit */
-+ u16 count; /* Initialises 'count+1' FQDs */
-+ struct qm_fqd fqd; /* the FQD fields go here */
-+ u8 __reserved3[30];
-+} __packed;
-+struct qm_mcc_queryfq {
-+ u8 __reserved1[3];
-+ u32 fqid; /* 24-bit */
-+ u8 __reserved2[56];
-+} __packed;
-+struct qm_mcc_queryfq_np {
-+ u8 __reserved1[3];
-+ u32 fqid; /* 24-bit */
-+ u8 __reserved2[56];
-+} __packed;
-+struct qm_mcc_alterfq {
-+ u8 __reserved1[3];
-+ u32 fqid; /* 24-bit */
-+ u8 __reserved2;
-+ u8 count; /* number of consecutive FQID */
-+ u8 __reserved3[10];
-+ u32 context_b; /* frame queue context b */
-+ u8 __reserved4[40];
-+} __packed;
-+struct qm_mcc_initcgr {
-+ u8 __reserved1;
-+ u16 we_mask; /* Write Enable Mask */
-+ struct __qm_mc_cgr cgr; /* CGR fields */
-+ u8 __reserved2[2];
-+ u8 cgid;
-+ u8 __reserved4[32];
-+} __packed;
-+struct qm_mcc_cgrtestwrite {
-+ u8 __reserved1[2];
-+ u8 i_bcnt_hi:8;/* high 8-bits of 40-bit "Instant" */
-+ u32 i_bcnt_lo; /* low 32-bits of 40-bit */
-+ u8 __reserved2[23];
-+ u8 cgid;
-+ u8 __reserved3[32];
-+} __packed;
-+struct qm_mcc_querycgr {
-+ u8 __reserved1[30];
-+ u8 cgid;
-+ u8 __reserved2[32];
-+} __packed;
-+struct qm_mcc_querycongestion {
-+ u8 __reserved[63];
-+} __packed;
-+struct qm_mcc_querywq {
-+ u8 __reserved;
-+ /* select channel if verb != QUERYWQ_DEDICATED */
-+ union {
-+ u16 channel_wq; /* ignores wq (3 lsbits) */
-+ struct {
-+#if __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
-+ u16 id:13; /* qm_channel */
-+ u16 __reserved1:3;
-+#else
-+ u16 __reserved1:3;
-+ u16 id:13; /* qm_channel */
-+#endif
-+ } __packed channel;
-+ };
-+ u8 __reserved2[60];
-+} __packed;
-+
-+struct qm_mcc_ceetm_lfqmt_config {
-+ u8 __reserved1[4];
-+ u32 lfqid:24;
-+ u8 __reserved2[2];
-+ u16 cqid;
-+ u8 __reserved3[2];
-+ u16 dctidx;
-+ u8 __reserved4[48];
-+} __packed;
-+
-+struct qm_mcc_ceetm_lfqmt_query {
-+ u8 __reserved1[4];
-+ u32 lfqid:24;
-+ u8 __reserved2[56];
-+} __packed;
-+
-+struct qm_mcc_ceetm_cq_config {
-+ u8 __reserved1;
-+ u16 cqid;
-+ u8 dcpid;
-+ u8 __reserved2;
-+ u16 ccgid;
-+ u8 __reserved3[56];
-+} __packed;
-+
-+struct qm_mcc_ceetm_cq_query {
-+ u8 __reserved1;
-+ u16 cqid;
-+ u8 dcpid;
-+ u8 __reserved2[59];
-+} __packed;
-+
-+struct qm_mcc_ceetm_dct_config {
-+ u8 __reserved1;
-+ u16 dctidx;
-+ u8 dcpid;
-+ u8 __reserved2[15];
-+ u32 context_b;
-+ u64 context_a;
-+ u8 __reserved3[32];
-+} __packed;
-+
-+struct qm_mcc_ceetm_dct_query {
-+ u8 __reserved1;
-+ u16 dctidx;
-+ u8 dcpid;
-+ u8 __reserved2[59];
-+} __packed;
-+
-+struct qm_mcc_ceetm_class_scheduler_config {
-+ u8 __reserved1;
-+ u16 cqcid;
-+ u8 dcpid;
-+ u8 __reserved2[6];
-+#if __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
-+ u8 gpc_reserved:1;
-+ u8 gpc_combine_flag:1;
-+ u8 gpc_prio_b:3;
-+ u8 gpc_prio_a:3;
-+#else
-+ u8 gpc_prio_a:3;
-+ u8 gpc_prio_b:3;
-+ u8 gpc_combine_flag:1;
-+ u8 gpc_reserved:1;
-+#endif
-+ u16 crem;
-+ u16 erem;
-+ u8 w[8];
-+ u8 __reserved3[40];
-+} __packed;
-+
-+struct qm_mcc_ceetm_class_scheduler_query {
-+ u8 __reserved1;
-+ u16 cqcid;
-+ u8 dcpid;
-+ u8 __reserved2[59];
-+} __packed;
-+
-+#define CEETM_COMMAND_CHANNEL_MAPPING (0 << 12)
-+#define CEETM_COMMAND_SP_MAPPING (1 << 12)
-+#define CEETM_COMMAND_CHANNEL_SHAPER (2 << 12)
-+#define CEETM_COMMAND_LNI_SHAPER (3 << 12)
-+#define CEETM_COMMAND_TCFC (4 << 12)
-+
-+#define CEETM_CCGRID_MASK 0x01FF
-+#define CEETM_CCGR_CM_CONFIGURE (0 << 14)
-+#define CEETM_CCGR_DN_CONFIGURE (1 << 14)
-+#define CEETM_CCGR_TEST_WRITE (2 << 14)
-+#define CEETM_CCGR_CM_QUERY (0 << 14)
-+#define CEETM_CCGR_DN_QUERY (1 << 14)
-+#define CEETM_CCGR_DN_QUERY_FLUSH (2 << 14)
-+#define CEETM_QUERY_CONGESTION_STATE (3 << 14)
-+
-+struct qm_mcc_ceetm_mapping_shaper_tcfc_config {
-+ u8 __reserved1;
-+ u16 cid;
-+ u8 dcpid;
-+ union {
-+ struct {
-+#if __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
-+ u8 map_shaped:1;
-+ u8 map_reserved:4;
-+ u8 map_lni_id:3;
-+#else
-+ u8 map_lni_id:3;
-+ u8 map_reserved:4;
-+ u8 map_shaped:1;
-+#endif
-+ u8 __reserved2[58];
-+ } __packed channel_mapping;
-+ struct {
-+#if __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
-+ u8 map_reserved:5;
-+ u8 map_lni_id:3;
-+#else
-+ u8 map_lni_id:3;
-+ u8 map_reserved:5;
-+#endif
-+ u8 __reserved2[58];
-+ } __packed sp_mapping;
-+ struct {
-+#if __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
-+ u8 cpl:1;
-+ u8 cpl_reserved:2;
-+ u8 oal:5;
-+#else
-+ u8 oal:5;
-+ u8 cpl_reserved:2;
-+ u8 cpl:1;
-+#endif
-+ u32 crtcr:24;
-+ u32 ertcr:24;
-+ u16 crtbl;
-+ u16 ertbl;
-+ u8 mps; /* This will be hardcoded by driver with 60 */
-+ u8 __reserved2[47];
-+ } __packed shaper_config;
-+ struct {
-+ u8 __reserved2[11];
-+ u64 lnitcfcc;
-+ u8 __reserved3[40];
-+ } __packed tcfc_config;
-+ };
-+} __packed;
-+
-+struct qm_mcc_ceetm_mapping_shaper_tcfc_query {
-+ u8 __reserved1;
-+ u16 cid;
-+ u8 dcpid;
-+ u8 __reserved2[59];
-+} __packed;
-+
-+struct qm_mcc_ceetm_ccgr_config {
-+ u8 __reserved1;
-+ u16 ccgrid;
-+ u8 dcpid;
-+ u8 __reserved2;
-+ u16 we_mask;
-+ union {
-+ struct {
-+#if __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
-+ u8 ctl_reserved:1;
-+ u8 ctl_wr_en_g:1;
-+ u8 ctl_wr_en_y:1;
-+ u8 ctl_wr_en_r:1;
-+ u8 ctl_td_en:1;
-+ u8 ctl_td_mode:1;
-+ u8 ctl_cscn_en:1;
-+ u8 ctl_mode:1;
-+#else
-+ u8 ctl_mode:1;
-+ u8 ctl_cscn_en:1;
-+ u8 ctl_td_mode:1;
-+ u8 ctl_td_en:1;
-+ u8 ctl_wr_en_r:1;
-+ u8 ctl_wr_en_y:1;
-+ u8 ctl_wr_en_g:1;
-+ u8 ctl_reserved:1;
-+#endif
-+ u8 cdv;
-+ u16 cscn_tupd;
-+ u8 oal;
-+ u8 __reserved3;
-+ struct qm_cgr_cs_thres cs_thres;
-+ struct qm_cgr_cs_thres cs_thres_x;
-+ struct qm_cgr_cs_thres td_thres;
-+ struct qm_cgr_wr_parm wr_parm_g;
-+ struct qm_cgr_wr_parm wr_parm_y;
-+ struct qm_cgr_wr_parm wr_parm_r;
-+ } __packed cm_config;
-+ struct {
-+ u8 dnc;
-+ u8 dn0;
-+ u8 dn1;
-+ u64 dnba:40;
-+ u8 __reserved3[2];
-+ u16 dnth_0;
-+ u8 __reserved4[2];
-+ u16 dnth_1;
-+ u8 __reserved5[8];
-+ } __packed dn_config;
-+ struct {
-+ u8 __reserved3[3];
-+ u64 i_cnt:40;
-+ u8 __reserved4[16];
-+ } __packed test_write;
-+ };
-+ u8 __reserved5[32];
-+} __packed;
-+
-+struct qm_mcc_ceetm_ccgr_query {
-+ u8 __reserved1;
-+ u16 ccgrid;
-+ u8 dcpid;
-+ u8 __reserved2[59];
-+} __packed;
-+
-+struct qm_mcc_ceetm_cq_peek_pop_xsfdrread {
-+ u8 __reserved1;
-+ u16 cqid;
-+ u8 dcpid;
-+ u8 ct;
-+ u16 xsfdr;
-+ u8 __reserved2[56];
-+} __packed;
-+
-+#define CEETM_QUERY_DEQUEUE_STATISTICS 0x00
-+#define CEETM_QUERY_DEQUEUE_CLEAR_STATISTICS 0x01
-+#define CEETM_WRITE_DEQUEUE_STATISTICS 0x02
-+#define CEETM_QUERY_REJECT_STATISTICS 0x03
-+#define CEETM_QUERY_REJECT_CLEAR_STATISTICS 0x04
-+#define CEETM_WRITE_REJECT_STATISTICS 0x05
-+struct qm_mcc_ceetm_statistics_query_write {
-+ u8 __reserved1;
-+ u16 cid;
-+ u8 dcpid;
-+ u8 ct;
-+ u8 __reserved2[13];
-+ u64 frm_cnt:40;
-+ u8 __reserved3[2];
-+ u64 byte_cnt:48;
-+ u8 __reserved[32];
-+} __packed;
-+
-+struct qm_mc_command {
-+ u8 __dont_write_directly__verb;
-+ union {
-+ struct qm_mcc_initfq initfq;
-+ struct qm_mcc_queryfq queryfq;
-+ struct qm_mcc_queryfq_np queryfq_np;
-+ struct qm_mcc_alterfq alterfq;
-+ struct qm_mcc_initcgr initcgr;
-+ struct qm_mcc_cgrtestwrite cgrtestwrite;
-+ struct qm_mcc_querycgr querycgr;
-+ struct qm_mcc_querycongestion querycongestion;
-+ struct qm_mcc_querywq querywq;
-+ struct qm_mcc_ceetm_lfqmt_config lfqmt_config;
-+ struct qm_mcc_ceetm_lfqmt_query lfqmt_query;
-+ struct qm_mcc_ceetm_cq_config cq_config;
-+ struct qm_mcc_ceetm_cq_query cq_query;
-+ struct qm_mcc_ceetm_dct_config dct_config;
-+ struct qm_mcc_ceetm_dct_query dct_query;
-+ struct qm_mcc_ceetm_class_scheduler_config csch_config;
-+ struct qm_mcc_ceetm_class_scheduler_query csch_query;
-+ struct qm_mcc_ceetm_mapping_shaper_tcfc_config mst_config;
-+ struct qm_mcc_ceetm_mapping_shaper_tcfc_query mst_query;
-+ struct qm_mcc_ceetm_ccgr_config ccgr_config;
-+ struct qm_mcc_ceetm_ccgr_query ccgr_query;
-+ struct qm_mcc_ceetm_cq_peek_pop_xsfdrread cq_ppxr;
-+ struct qm_mcc_ceetm_statistics_query_write stats_query_write;
-+ };
-+} __packed;
-+#define QM_MCC_VERB_VBIT 0x80
-+#define QM_MCC_VERB_MASK 0x7f /* where the verb contains; */
-+#define QM_MCC_VERB_INITFQ_PARKED 0x40
-+#define QM_MCC_VERB_INITFQ_SCHED 0x41
-+#define QM_MCC_VERB_QUERYFQ 0x44
-+#define QM_MCC_VERB_QUERYFQ_NP 0x45 /* "non-programmable" fields */
-+#define QM_MCC_VERB_QUERYWQ 0x46
-+#define QM_MCC_VERB_QUERYWQ_DEDICATED 0x47
-+#define QM_MCC_VERB_ALTER_SCHED 0x48 /* Schedule FQ */
-+#define QM_MCC_VERB_ALTER_FE 0x49 /* Force Eligible FQ */
-+#define QM_MCC_VERB_ALTER_RETIRE 0x4a /* Retire FQ */
-+#define QM_MCC_VERB_ALTER_OOS 0x4b /* Take FQ out of service */
-+#define QM_MCC_VERB_ALTER_FQXON 0x4d /* FQ XON */
-+#define QM_MCC_VERB_ALTER_FQXOFF 0x4e /* FQ XOFF */
-+#define QM_MCC_VERB_INITCGR 0x50
-+#define QM_MCC_VERB_MODIFYCGR 0x51
-+#define QM_MCC_VERB_CGRTESTWRITE 0x52
-+#define QM_MCC_VERB_QUERYCGR 0x58
-+#define QM_MCC_VERB_QUERYCONGESTION 0x59
-+/* INITFQ-specific flags */
-+#define QM_INITFQ_WE_MASK 0x01ff /* 'Write Enable' flags; */
-+#define QM_INITFQ_WE_OAC 0x0100
-+#define QM_INITFQ_WE_ORPC 0x0080
-+#define QM_INITFQ_WE_CGID 0x0040
-+#define QM_INITFQ_WE_FQCTRL 0x0020
-+#define QM_INITFQ_WE_DESTWQ 0x0010
-+#define QM_INITFQ_WE_ICSCRED 0x0008
-+#define QM_INITFQ_WE_TDTHRESH 0x0004
-+#define QM_INITFQ_WE_CONTEXTB 0x0002
-+#define QM_INITFQ_WE_CONTEXTA 0x0001
-+/* INITCGR/MODIFYCGR-specific flags */
-+#define QM_CGR_WE_MASK 0x07ff /* 'Write Enable Mask'; */
-+#define QM_CGR_WE_WR_PARM_G 0x0400
-+#define QM_CGR_WE_WR_PARM_Y 0x0200
-+#define QM_CGR_WE_WR_PARM_R 0x0100
-+#define QM_CGR_WE_WR_EN_G 0x0080
-+#define QM_CGR_WE_WR_EN_Y 0x0040
-+#define QM_CGR_WE_WR_EN_R 0x0020
-+#define QM_CGR_WE_CSCN_EN 0x0010
-+#define QM_CGR_WE_CSCN_TARG 0x0008
-+#define QM_CGR_WE_CSTD_EN 0x0004
-+#define QM_CGR_WE_CS_THRES 0x0002
-+#define QM_CGR_WE_MODE 0x0001
-+
-+/* See 1.5.9.7 CEETM Management Commands */
-+#define QM_CEETM_VERB_LFQMT_CONFIG 0x70
-+#define QM_CEETM_VERB_LFQMT_QUERY 0x71
-+#define QM_CEETM_VERB_CQ_CONFIG 0x72
-+#define QM_CEETM_VERB_CQ_QUERY 0x73
-+#define QM_CEETM_VERB_DCT_CONFIG 0x74
-+#define QM_CEETM_VERB_DCT_QUERY 0x75
-+#define QM_CEETM_VERB_CLASS_SCHEDULER_CONFIG 0x76
-+#define QM_CEETM_VERB_CLASS_SCHEDULER_QUERY 0x77
-+#define QM_CEETM_VERB_MAPPING_SHAPER_TCFC_CONFIG 0x78
-+#define QM_CEETM_VERB_MAPPING_SHAPER_TCFC_QUERY 0x79
-+#define QM_CEETM_VERB_CCGR_CONFIG 0x7A
-+#define QM_CEETM_VERB_CCGR_QUERY 0x7B
-+#define QM_CEETM_VERB_CQ_PEEK_POP_XFDRREAD 0x7C
-+#define QM_CEETM_VERB_STATISTICS_QUERY_WRITE 0x7D
-+
-+/* See 1.5.8.5.1: "Initialize FQ" */
-+/* See 1.5.8.5.2: "Query FQ" */
-+/* See 1.5.8.5.3: "Query FQ Non-Programmable Fields" */
-+/* See 1.5.8.5.4: "Alter FQ State Commands " */
-+/* See 1.5.8.6.1: "Initialize/Modify CGR" */
-+/* See 1.5.8.6.2: "CGR Test Write" */
-+/* See 1.5.8.6.3: "Query CGR" */
-+/* See 1.5.8.6.4: "Query Congestion Group State" */
-+struct qm_mcr_initfq {
-+ u8 __reserved1[62];
-+} __packed;
-+struct qm_mcr_queryfq {
-+ u8 __reserved1[8];
-+ struct qm_fqd fqd; /* the FQD fields are here */
-+ u8 __reserved2[30];
-+} __packed;
-+struct qm_mcr_queryfq_np {
-+ u8 __reserved1;
-+ u8 state; /* QM_MCR_NP_STATE_*** */
-+#if __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
-+ u8 __reserved2;
-+ u32 fqd_link:24;
-+ u16 __reserved3:2;
-+ u16 odp_seq:14;
-+ u16 __reserved4:2;
-+ u16 orp_nesn:14;
-+ u16 __reserved5:1;
-+ u16 orp_ea_hseq:15;
-+ u16 __reserved6:1;
-+ u16 orp_ea_tseq:15;
-+ u8 __reserved7;
-+ u32 orp_ea_hptr:24;
-+ u8 __reserved8;
-+ u32 orp_ea_tptr:24;
-+ u8 __reserved9;
-+ u32 pfdr_hptr:24;
-+ u8 __reserved10;
-+ u32 pfdr_tptr:24;
-+ u8 __reserved11[5];
-+ u8 __reserved12:7;
-+ u8 is:1;
-+ u16 ics_surp;
-+ u32 byte_cnt;
-+ u8 __reserved13;
-+ u32 frm_cnt:24;
-+ u32 __reserved14;
-+ u16 ra1_sfdr; /* QM_MCR_NP_RA1_*** */
-+ u16 ra2_sfdr; /* QM_MCR_NP_RA2_*** */
-+ u16 __reserved15;
-+ u16 od1_sfdr; /* QM_MCR_NP_OD1_*** */
-+ u16 od2_sfdr; /* QM_MCR_NP_OD2_*** */
-+ u16 od3_sfdr; /* QM_MCR_NP_OD3_*** */
-+#else
-+ u8 __reserved2;
-+ u32 fqd_link:24;
-+
-+ u16 odp_seq:14;
-+ u16 __reserved3:2;
-+
-+ u16 orp_nesn:14;
-+ u16 __reserved4:2;
-+
-+ u16 orp_ea_hseq:15;
-+ u16 __reserved5:1;
-+
-+ u16 orp_ea_tseq:15;
-+ u16 __reserved6:1;
-+
-+ u8 __reserved7;
-+ u32 orp_ea_hptr:24;
-+
-+ u8 __reserved8;
-+ u32 orp_ea_tptr:24;
-+
-+ u8 __reserved9;
-+ u32 pfdr_hptr:24;
-+
-+ u8 __reserved10;
-+ u32 pfdr_tptr:24;
-+
-+ u8 __reserved11[5];
-+ u8 is:1;
-+ u8 __reserved12:7;
-+ u16 ics_surp;
-+ u32 byte_cnt;
-+ u8 __reserved13;
-+ u32 frm_cnt:24;
-+ u32 __reserved14;
-+ u16 ra1_sfdr; /* QM_MCR_NP_RA1_*** */
-+ u16 ra2_sfdr; /* QM_MCR_NP_RA2_*** */
-+ u16 __reserved15;
-+ u16 od1_sfdr; /* QM_MCR_NP_OD1_*** */
-+ u16 od2_sfdr; /* QM_MCR_NP_OD2_*** */
-+ u16 od3_sfdr; /* QM_MCR_NP_OD3_*** */
-+#endif
-+} __packed;
-+
-+
-+struct qm_mcr_alterfq {
-+ u8 fqs; /* Frame Queue Status */
-+ u8 __reserved1[61];
-+} __packed;
-+struct qm_mcr_initcgr {
-+ u8 __reserved1[62];
-+} __packed;
-+struct qm_mcr_cgrtestwrite {
-+ u16 __reserved1;
-+ struct __qm_mc_cgr cgr; /* CGR fields */
-+ u8 __reserved2[3];
-+ u32 __reserved3:24;
-+ u32 i_bcnt_hi:8;/* high 8-bits of 40-bit "Instant" */
-+ u32 i_bcnt_lo; /* low 32-bits of 40-bit */
-+ u32 __reserved4:24;
-+ u32 a_bcnt_hi:8;/* high 8-bits of 40-bit "Average" */
-+ u32 a_bcnt_lo; /* low 32-bits of 40-bit */
-+ u16 lgt; /* Last Group Tick */
-+ u16 wr_prob_g;
-+ u16 wr_prob_y;
-+ u16 wr_prob_r;
-+ u8 __reserved5[8];
-+} __packed;
-+struct qm_mcr_querycgr {
-+ u16 __reserved1;
-+ struct __qm_mc_cgr cgr; /* CGR fields */
-+ u8 __reserved2[3];
-+ union {
-+ struct {
-+#if __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
-+ u32 __reserved3:24;
-+ u32 i_bcnt_hi:8;/* high 8-bits of 40-bit "Instant" */
-+ u32 i_bcnt_lo; /* low 32-bits of 40-bit */
-+#else
-+ u32 i_bcnt_lo; /* low 32-bits of 40-bit */
-+ u32 i_bcnt_hi:8;/* high 8-bits of 40-bit "Instant" */
-+ u32 __reserved3:24;
-+#endif
-+ };
-+ u64 i_bcnt;
-+ };
-+ union {
-+ struct {
-+#if __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
-+ u32 __reserved4:24;
-+ u32 a_bcnt_hi:8;/* high 8-bits of 40-bit "Average" */
-+ u32 a_bcnt_lo; /* low 32-bits of 40-bit */
-+#else
-+ u32 a_bcnt_lo; /* low 32-bits of 40-bit */
-+ u32 a_bcnt_hi:8;/* high 8-bits of 40-bit "Average" */
-+ u32 __reserved4:24;
-+#endif
-+ };
-+ u64 a_bcnt;
-+ };
-+ union {
-+ u32 cscn_targ_swp[4];
-+ u8 __reserved5[16];
-+ };
-+} __packed;
-+static inline u64 qm_mcr_querycgr_i_get64(const struct qm_mcr_querycgr *q)
-+{
-+ return be64_to_cpu(q->i_bcnt);
-+}
-+static inline u64 qm_mcr_querycgr_a_get64(const struct qm_mcr_querycgr *q)
-+{
-+ return be64_to_cpu(q->a_bcnt);
-+}
-+static inline u64 qm_mcr_cgrtestwrite_i_get64(
-+ const struct qm_mcr_cgrtestwrite *q)
-+{
-+ return be64_to_cpu(((u64)q->i_bcnt_hi << 32) | (u64)q->i_bcnt_lo);
-+}
-+static inline u64 qm_mcr_cgrtestwrite_a_get64(
-+ const struct qm_mcr_cgrtestwrite *q)
-+{
-+ return be64_to_cpu(((u64)q->a_bcnt_hi << 32) | (u64)q->a_bcnt_lo);
-+}
-+/* Macro, so we compile better if 'v' isn't always 64-bit */
-+#define qm_mcr_querycgr_i_set64(q, v) \
-+ do { \
-+ struct qm_mcr_querycgr *__q931 = (fd); \
-+ __q931->i_bcnt_hi = upper_32_bits(v); \
-+ __q931->i_bcnt_lo = lower_32_bits(v); \
-+ } while (0)
-+#define qm_mcr_querycgr_a_set64(q, v) \
-+ do { \
-+ struct qm_mcr_querycgr *__q931 = (fd); \
-+ __q931->a_bcnt_hi = upper_32_bits(v); \
-+ __q931->a_bcnt_lo = lower_32_bits(v); \
-+ } while (0)
-+struct __qm_mcr_querycongestion {
-+ u32 __state[8];
-+};
-+struct qm_mcr_querycongestion {
-+ u8 __reserved[30];
-+ /* Access this struct using QM_MCR_QUERYCONGESTION() */
-+ struct __qm_mcr_querycongestion state;
-+} __packed;
-+struct qm_mcr_querywq {
-+ union {
-+ u16 channel_wq; /* ignores wq (3 lsbits) */
-+ struct {
-+#if __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
-+ u16 id:13; /* qm_channel */
-+ u16 __reserved:3;
-+#else
-+ u16 __reserved:3;
-+ u16 id:13; /* qm_channel */
-+#endif
-+ } __packed channel;
-+ };
-+ u8 __reserved[28];
-+ u32 wq_len[8];
-+} __packed;
-+
-+/* QMAN CEETM Management Command Response */
-+struct qm_mcr_ceetm_lfqmt_config {
-+ u8 __reserved1[62];
-+} __packed;
-+struct qm_mcr_ceetm_lfqmt_query {
-+ u8 __reserved1[8];
-+ u16 cqid;
-+ u8 __reserved2[2];
-+ u16 dctidx;
-+ u8 __reserved3[2];
-+ u16 ccgid;
-+ u8 __reserved4[44];
-+} __packed;
-+
-+struct qm_mcr_ceetm_cq_config {
-+ u8 __reserved1[62];
-+} __packed;
-+
-+struct qm_mcr_ceetm_cq_query {
-+ u8 __reserved1[4];
-+ u16 ccgid;
-+ u16 state;
-+ u32 pfdr_hptr:24;
-+ u32 pfdr_tptr:24;
-+ u16 od1_xsfdr;
-+ u16 od2_xsfdr;
-+ u16 od3_xsfdr;
-+ u16 od4_xsfdr;
-+ u16 od5_xsfdr;
-+ u16 od6_xsfdr;
-+ u16 ra1_xsfdr;
-+ u16 ra2_xsfdr;
-+ u8 __reserved2;
-+ u32 frm_cnt:24;
-+ u8 __reserved333[28];
-+} __packed;
-+
-+struct qm_mcr_ceetm_dct_config {
-+ u8 __reserved1[62];
-+} __packed;
-+
-+struct qm_mcr_ceetm_dct_query {
-+ u8 __reserved1[18];
-+ u32 context_b;
-+ u64 context_a;
-+ u8 __reserved2[32];
-+} __packed;
-+
-+struct qm_mcr_ceetm_class_scheduler_config {
-+ u8 __reserved1[62];
-+} __packed;
-+
-+struct qm_mcr_ceetm_class_scheduler_query {
-+ u8 __reserved1[9];
-+#if __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
-+ u8 gpc_reserved:1;
-+ u8 gpc_combine_flag:1;
-+ u8 gpc_prio_b:3;
-+ u8 gpc_prio_a:3;
-+#else
-+ u8 gpc_prio_a:3;
-+ u8 gpc_prio_b:3;
-+ u8 gpc_combine_flag:1;
-+ u8 gpc_reserved:1;
-+#endif
-+ u16 crem;
-+ u16 erem;
-+ u8 w[8];
-+ u8 __reserved2[5];
-+ u32 wbfslist:24;
-+ u32 d8;
-+ u32 d9;
-+ u32 d10;
-+ u32 d11;
-+ u32 d12;
-+ u32 d13;
-+ u32 d14;
-+ u32 d15;
-+} __packed;
-+
-+struct qm_mcr_ceetm_mapping_shaper_tcfc_config {
-+ u16 cid;
-+ u8 __reserved2[60];
-+} __packed;
-+
-+struct qm_mcr_ceetm_mapping_shaper_tcfc_query {
-+ u16 cid;
-+ u8 __reserved1;
-+ union {
-+ struct {
-+#if __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
-+ u8 map_shaped:1;
-+ u8 map_reserved:4;
-+ u8 map_lni_id:3;
-+#else
-+ u8 map_lni_id:3;
-+ u8 map_reserved:4;
-+ u8 map_shaped:1;
-+#endif
-+ u8 __reserved2[58];
-+ } __packed channel_mapping_query;
-+ struct {
-+#if __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
-+ u8 map_reserved:5;
-+ u8 map_lni_id:3;
-+#else
-+ u8 map_lni_id:3;
-+ u8 map_reserved:5;
-+#endif
-+ u8 __reserved2[58];
-+ } __packed sp_mapping_query;
-+ struct {
-+#if __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
-+ u8 cpl:1;
-+ u8 cpl_reserved:2;
-+ u8 oal:5;
-+#else
-+ u8 oal:5;
-+ u8 cpl_reserved:2;
-+ u8 cpl:1;
-+#endif
-+ u32 crtcr:24;
-+ u32 ertcr:24;
-+ u16 crtbl;
-+ u16 ertbl;
-+ u8 mps;
-+ u8 __reserved2[15];
-+ u32 crat;
-+ u32 erat;
-+ u8 __reserved3[24];
-+ } __packed shaper_query;
-+ struct {
-+ u8 __reserved1[11];
-+ u64 lnitcfcc;
-+ u8 __reserved3[40];
-+ } __packed tcfc_query;
-+ };
-+} __packed;
-+
-+struct qm_mcr_ceetm_ccgr_config {
-+ u8 __reserved1[46];
-+ union {
-+ u8 __reserved2[8];
-+ struct {
-+ u16 timestamp;
-+ u16 wr_porb_g;
-+ u16 wr_prob_y;
-+ u16 wr_prob_r;
-+ } __packed test_write;
-+ };
-+ u8 __reserved3[8];
-+} __packed;
-+
-+struct qm_mcr_ceetm_ccgr_query {
-+ u8 __reserved1[6];
-+ union {
-+ struct {
-+#if __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
-+ u8 ctl_reserved:1;
-+ u8 ctl_wr_en_g:1;
-+ u8 ctl_wr_en_y:1;
-+ u8 ctl_wr_en_r:1;
-+ u8 ctl_td_en:1;
-+ u8 ctl_td_mode:1;
-+ u8 ctl_cscn_en:1;
-+ u8 ctl_mode:1;
-+#else
-+ u8 ctl_mode:1;
-+ u8 ctl_cscn_en:1;
-+ u8 ctl_td_mode:1;
-+ u8 ctl_td_en:1;
-+ u8 ctl_wr_en_r:1;
-+ u8 ctl_wr_en_y:1;
-+ u8 ctl_wr_en_g:1;
-+ u8 ctl_reserved:1;
-+#endif
-+ u8 cdv;
-+ u8 __reserved2[2];
-+ u8 oal;
-+ u8 __reserved3;
-+ struct qm_cgr_cs_thres cs_thres;
-+ struct qm_cgr_cs_thres cs_thres_x;
-+ struct qm_cgr_cs_thres td_thres;
-+ struct qm_cgr_wr_parm wr_parm_g;
-+ struct qm_cgr_wr_parm wr_parm_y;
-+ struct qm_cgr_wr_parm wr_parm_r;
-+ u16 cscn_targ_dcp;
-+ u8 dcp_lsn;
-+ u64 i_cnt:40;
-+ u8 __reserved4[3];
-+ u64 a_cnt:40;
-+ u32 cscn_targ_swp[4];
-+ } __packed cm_query;
-+ struct {
-+ u8 dnc;
-+ u8 dn0;
-+ u8 dn1;
-+ u64 dnba:40;
-+ u8 __reserved2[2];
-+ u16 dnth_0;
-+ u8 __reserved3[2];
-+ u16 dnth_1;
-+ u8 __reserved4[10];
-+ u16 dnacc_0;
-+ u8 __reserved5[2];
-+ u16 dnacc_1;
-+ u8 __reserved6[24];
-+ } __packed dn_query;
-+ struct {
-+ u8 __reserved2[24];
-+ struct __qm_mcr_querycongestion state;
-+ } __packed congestion_state;
-+
-+ };
-+} __packed;
-+
-+struct qm_mcr_ceetm_cq_peek_pop_xsfdrread {
-+ u8 stat;
-+ u8 __reserved1[11];
-+ u16 dctidx;
-+ struct qm_fd fd;
-+ u8 __reserved2[32];
-+} __packed;
-+
-+struct qm_mcr_ceetm_statistics_query {
-+ u8 __reserved1[17];
-+ u64 frm_cnt:40;
-+ u8 __reserved2[2];
-+ u64 byte_cnt:48;
-+ u8 __reserved3[32];
-+} __packed;
-+
-+struct qm_mc_result {
-+ u8 verb;
-+ u8 result;
-+ union {
-+ struct qm_mcr_initfq initfq;
-+ struct qm_mcr_queryfq queryfq;
-+ struct qm_mcr_queryfq_np queryfq_np;
-+ struct qm_mcr_alterfq alterfq;
-+ struct qm_mcr_initcgr initcgr;
-+ struct qm_mcr_cgrtestwrite cgrtestwrite;
-+ struct qm_mcr_querycgr querycgr;
-+ struct qm_mcr_querycongestion querycongestion;
-+ struct qm_mcr_querywq querywq;
-+ struct qm_mcr_ceetm_lfqmt_config lfqmt_config;
-+ struct qm_mcr_ceetm_lfqmt_query lfqmt_query;
-+ struct qm_mcr_ceetm_cq_config cq_config;
-+ struct qm_mcr_ceetm_cq_query cq_query;
-+ struct qm_mcr_ceetm_dct_config dct_config;
-+ struct qm_mcr_ceetm_dct_query dct_query;
-+ struct qm_mcr_ceetm_class_scheduler_config csch_config;
-+ struct qm_mcr_ceetm_class_scheduler_query csch_query;
-+ struct qm_mcr_ceetm_mapping_shaper_tcfc_config mst_config;
-+ struct qm_mcr_ceetm_mapping_shaper_tcfc_query mst_query;
-+ struct qm_mcr_ceetm_ccgr_config ccgr_config;
-+ struct qm_mcr_ceetm_ccgr_query ccgr_query;
-+ struct qm_mcr_ceetm_cq_peek_pop_xsfdrread cq_ppxr;
-+ struct qm_mcr_ceetm_statistics_query stats_query;
-+ };
-+} __packed;
-+
-+#define QM_MCR_VERB_RRID 0x80
-+#define QM_MCR_VERB_MASK QM_MCC_VERB_MASK
-+#define QM_MCR_VERB_INITFQ_PARKED QM_MCC_VERB_INITFQ_PARKED
-+#define QM_MCR_VERB_INITFQ_SCHED QM_MCC_VERB_INITFQ_SCHED
-+#define QM_MCR_VERB_QUERYFQ QM_MCC_VERB_QUERYFQ
-+#define QM_MCR_VERB_QUERYFQ_NP QM_MCC_VERB_QUERYFQ_NP
-+#define QM_MCR_VERB_QUERYWQ QM_MCC_VERB_QUERYWQ
-+#define QM_MCR_VERB_QUERYWQ_DEDICATED QM_MCC_VERB_QUERYWQ_DEDICATED
-+#define QM_MCR_VERB_ALTER_SCHED QM_MCC_VERB_ALTER_SCHED
-+#define QM_MCR_VERB_ALTER_FE QM_MCC_VERB_ALTER_FE
-+#define QM_MCR_VERB_ALTER_RETIRE QM_MCC_VERB_ALTER_RETIRE
-+#define QM_MCR_VERB_ALTER_OOS QM_MCC_VERB_ALTER_OOS
-+#define QM_MCR_RESULT_NULL 0x00
-+#define QM_MCR_RESULT_OK 0xf0
-+#define QM_MCR_RESULT_ERR_FQID 0xf1
-+#define QM_MCR_RESULT_ERR_FQSTATE 0xf2
-+#define QM_MCR_RESULT_ERR_NOTEMPTY 0xf3 /* OOS fails if FQ is !empty */
-+#define QM_MCR_RESULT_ERR_BADCHANNEL 0xf4
-+#define QM_MCR_RESULT_PENDING 0xf8
-+#define QM_MCR_RESULT_ERR_BADCOMMAND 0xff
-+#define QM_MCR_NP_STATE_FE 0x10
-+#define QM_MCR_NP_STATE_R 0x08
-+#define QM_MCR_NP_STATE_MASK 0x07 /* Reads FQD::STATE; */
-+#define QM_MCR_NP_STATE_OOS 0x00
-+#define QM_MCR_NP_STATE_RETIRED 0x01
-+#define QM_MCR_NP_STATE_TEN_SCHED 0x02
-+#define QM_MCR_NP_STATE_TRU_SCHED 0x03
-+#define QM_MCR_NP_STATE_PARKED 0x04
-+#define QM_MCR_NP_STATE_ACTIVE 0x05
-+#define QM_MCR_NP_PTR_MASK 0x07ff /* for RA[12] & OD[123] */
-+#define QM_MCR_NP_RA1_NRA(v) (((v) >> 14) & 0x3) /* FQD::NRA */
-+#define QM_MCR_NP_RA2_IT(v) (((v) >> 14) & 0x1) /* FQD::IT */
-+#define QM_MCR_NP_OD1_NOD(v) (((v) >> 14) & 0x3) /* FQD::NOD */
-+#define QM_MCR_NP_OD3_NPC(v) (((v) >> 14) & 0x3) /* FQD::NPC */
-+#define QM_MCR_FQS_ORLPRESENT 0x02 /* ORL fragments to come */
-+#define QM_MCR_FQS_NOTEMPTY 0x01 /* FQ has enqueued frames */
-+/* This extracts the state for congestion group 'n' from a query response.
-+ * Eg.
-+ * u8 cgr = [...];
-+ * struct qm_mc_result *res = [...];
-+ * printf("congestion group %d congestion state: %d\n", cgr,
-+ * QM_MCR_QUERYCONGESTION(&res->querycongestion.state, cgr));
-+ */
-+#define __CGR_WORD(num) (num >> 5)
-+#define __CGR_SHIFT(num) (num & 0x1f)
-+#define __CGR_NUM (sizeof(struct __qm_mcr_querycongestion) << 3)
-+static inline int QM_MCR_QUERYCONGESTION(struct __qm_mcr_querycongestion *p,
-+ u8 cgr)
-+{
-+ return p->__state[__CGR_WORD(cgr)] & (0x80000000 >> __CGR_SHIFT(cgr));
-+}
-+
-+
-+/*********************/
-+/* Utility interface */
-+/*********************/
-+
-+/* Represents an allocator over a range of FQIDs. NB, accesses are not locked,
-+ * spinlock them yourself if needed. */
-+struct qman_fqid_pool;
-+
-+/* Create/destroy a FQID pool, num must be a multiple of 32. NB, _destroy()
-+ * always succeeds, but returns non-zero if there were "leaked" FQID
-+ * allocations. */
-+struct qman_fqid_pool *qman_fqid_pool_create(u32 fqid_start, u32 num);
-+int qman_fqid_pool_destroy(struct qman_fqid_pool *pool);
-+/* Alloc/free a FQID from the range. _alloc() returns zero for success. */
-+int qman_fqid_pool_alloc(struct qman_fqid_pool *pool, u32 *fqid);
-+void qman_fqid_pool_free(struct qman_fqid_pool *pool, u32 fqid);
-+u32 qman_fqid_pool_used(struct qman_fqid_pool *pool);
-+
-+/*******************************************************************/
-+/* Managed (aka "shared" or "mux/demux") portal, high-level i/face */
-+/*******************************************************************/
-+
-+ /* Portal and Frame Queues */
-+ /* ----------------------- */
-+/* Represents a managed portal */
-+struct qman_portal;
-+
-+/* This object type represents Qman frame queue descriptors (FQD), it is
-+ * cacheline-aligned, and initialised by qman_create_fq(). The structure is
-+ * defined further down. */
-+struct qman_fq;
-+
-+/* This object type represents a Qman congestion group, it is defined further
-+ * down. */
-+struct qman_cgr;
-+
-+struct qman_portal_config {
-+ /* If the caller enables DQRR stashing (and thus wishes to operate the
-+ * portal from only one cpu), this is the logical CPU that the portal
-+ * will stash to. Whether stashing is enabled or not, this setting is
-+ * also used for any "core-affine" portals, ie. default portals
-+ * associated to the corresponding cpu. -1 implies that there is no core
-+ * affinity configured. */
-+ int cpu;
-+ /* portal interrupt line */
-+ int irq;
-+ /* the unique index of this portal */
-+ u32 index;
-+ /* Is this portal shared? (If so, it has coarser locking and demuxes
-+ * processing on behalf of other CPUs.) */
-+ int is_shared;
-+ /* The portal's dedicated channel id, use this value for initialising
-+ * frame queues to target this portal when scheduled. */
-+ u16 channel;
-+ /* A mask of which pool channels this portal has dequeue access to
-+ * (using QM_SDQCR_CHANNELS_POOL(n) for the bitmask) */
-+ u32 pools;
-+};
-+
-+/* This enum, and the callback type that returns it, are used when handling
-+ * dequeued frames via DQRR. Note that for "null" callbacks registered with the
-+ * portal object (for handling dequeues that do not demux because contextB is
-+ * NULL), the return value *MUST* be qman_cb_dqrr_consume. */
-+enum qman_cb_dqrr_result {
-+ /* DQRR entry can be consumed */
-+ qman_cb_dqrr_consume,
-+ /* Like _consume, but requests parking - FQ must be held-active */
-+ qman_cb_dqrr_park,
-+ /* Does not consume, for DCA mode only. This allows out-of-order
-+ * consumes by explicit calls to qman_dca() and/or the use of implicit
-+ * DCA via EQCR entries. */
-+ qman_cb_dqrr_defer,
-+ /* Stop processing without consuming this ring entry. Exits the current
-+ * qman_poll_dqrr() or interrupt-handling, as appropriate. If within an
-+ * interrupt handler, the callback would typically call
-+ * qman_irqsource_remove(QM_PIRQ_DQRI) before returning this value,
-+ * otherwise the interrupt will reassert immediately. */
-+ qman_cb_dqrr_stop,
-+ /* Like qman_cb_dqrr_stop, but consumes the current entry. */
-+ qman_cb_dqrr_consume_stop
-+};
-+typedef enum qman_cb_dqrr_result (*qman_cb_dqrr)(struct qman_portal *qm,
-+ struct qman_fq *fq,
-+ const struct qm_dqrr_entry *dqrr);
-+
-+/* This callback type is used when handling ERNs, FQRNs and FQRLs via MR. They
-+ * are always consumed after the callback returns. */
-+typedef void (*qman_cb_mr)(struct qman_portal *qm, struct qman_fq *fq,
-+ const struct qm_mr_entry *msg);
-+
-+/* This callback type is used when handling DCP ERNs */
-+typedef void (*qman_cb_dc_ern)(struct qman_portal *qm,
-+ const struct qm_mr_entry *msg);
-+
-+/* s/w-visible states. Ie. tentatively scheduled + truly scheduled + active +
-+ * held-active + held-suspended are just "sched". Things like "retired" will not
-+ * be assumed until it is complete (ie. QMAN_FQ_STATE_CHANGING is set until
-+ * then, to indicate it's completing and to gate attempts to retry the retire
-+ * command). Note, park commands do not set QMAN_FQ_STATE_CHANGING because it's
-+ * technically impossible in the case of enqueue DCAs (which refer to DQRR ring
-+ * index rather than the FQ that ring entry corresponds to), so repeated park
-+ * commands are allowed (if you're silly enough to try) but won't change FQ
-+ * state, and the resulting park notifications move FQs from "sched" to
-+ * "parked". */
-+enum qman_fq_state {
-+ qman_fq_state_oos,
-+ qman_fq_state_parked,
-+ qman_fq_state_sched,
-+ qman_fq_state_retired
-+};
-+
-+/* Frame queue objects (struct qman_fq) are stored within memory passed to
-+ * qman_create_fq(), as this allows stashing of caller-provided demux callback
-+ * pointers at no extra cost to stashing of (driver-internal) FQ state. If the
-+ * caller wishes to add per-FQ state and have it benefit from dequeue-stashing,
-+ * they should;
-+ *
-+ * (a) extend the qman_fq structure with their state; eg.
-+ *
-+ * // myfq is allocated and driver_fq callbacks filled in;
-+ * struct my_fq {
-+ * struct qman_fq base;
-+ * int an_extra_field;
-+ * [ ... add other fields to be associated with each FQ ...]
-+ * } *myfq = some_my_fq_allocator();
-+ * struct qman_fq *fq = qman_create_fq(fqid, flags, &myfq->base);
-+ *
-+ * // in a dequeue callback, access extra fields from 'fq' via a cast;
-+ * struct my_fq *myfq = (struct my_fq *)fq;
-+ * do_something_with(myfq->an_extra_field);
-+ * [...]
-+ *
-+ * (b) when and if configuring the FQ for context stashing, specify how ever
-+ * many cachelines are required to stash 'struct my_fq', to accelerate not
-+ * only the Qman driver but the callback as well.
-+ */
-+
-+struct qman_fq_cb {
-+ qman_cb_dqrr dqrr; /* for dequeued frames */
-+ qman_cb_mr ern; /* for s/w ERNs */
-+ qman_cb_mr fqs; /* frame-queue state changes*/
-+};
-+
-+struct qman_fq {
-+ /* Caller of qman_create_fq() provides these demux callbacks */
-+ struct qman_fq_cb cb;
-+ /* These are internal to the driver, don't touch. In particular, they
-+ * may change, be removed, or extended (so you shouldn't rely on
-+ * sizeof(qman_fq) being a constant). */
-+ spinlock_t fqlock;
-+ u32 fqid;
-+ volatile unsigned long flags;
-+ enum qman_fq_state state;
-+ int cgr_groupid;
-+ struct rb_node node;
-+#ifdef CONFIG_FSL_QMAN_FQ_LOOKUP
-+ u32 key;
-+#endif
-+};
-+
-+/* This callback type is used when handling congestion group entry/exit.
-+ * 'congested' is non-zero on congestion-entry, and zero on congestion-exit. */
-+typedef void (*qman_cb_cgr)(struct qman_portal *qm,
-+ struct qman_cgr *cgr, int congested);
-+
-+struct qman_cgr {
-+ /* Set these prior to qman_create_cgr() */
-+ u32 cgrid; /* 0..255, but u32 to allow specials like -1, 256, etc.*/
-+ qman_cb_cgr cb;
-+ /* These are private to the driver */
-+ u16 chan; /* portal channel this object is created on */
-+ struct list_head node;
-+};
-+
-+/* Flags to qman_create_fq() */
-+#define QMAN_FQ_FLAG_NO_ENQUEUE 0x00000001 /* can't enqueue */
-+#define QMAN_FQ_FLAG_NO_MODIFY 0x00000002 /* can only enqueue */
-+#define QMAN_FQ_FLAG_TO_DCPORTAL 0x00000004 /* consumed by CAAM/PME/Fman */
-+#define QMAN_FQ_FLAG_LOCKED 0x00000008 /* multi-core locking */
-+#define QMAN_FQ_FLAG_AS_IS 0x00000010 /* query h/w state */
-+#define QMAN_FQ_FLAG_DYNAMIC_FQID 0x00000020 /* (de)allocate fqid */
-+
-+/* Flags to qman_destroy_fq() */
-+#define QMAN_FQ_DESTROY_PARKED 0x00000001 /* FQ can be parked or OOS */
-+
-+/* Flags from qman_fq_state() */
-+#define QMAN_FQ_STATE_CHANGING 0x80000000 /* 'state' is changing */
-+#define QMAN_FQ_STATE_NE 0x40000000 /* retired FQ isn't empty */
-+#define QMAN_FQ_STATE_ORL 0x20000000 /* retired FQ has ORL */
-+#define QMAN_FQ_STATE_BLOCKOOS 0xe0000000 /* if any are set, no OOS */
-+#define QMAN_FQ_STATE_CGR_EN 0x10000000 /* CGR enabled */
-+#define QMAN_FQ_STATE_VDQCR 0x08000000 /* being volatile dequeued */
-+
-+/* Flags to qman_init_fq() */
-+#define QMAN_INITFQ_FLAG_SCHED 0x00000001 /* schedule rather than park */
-+#define QMAN_INITFQ_FLAG_LOCAL 0x00000004 /* set dest portal */
-+
-+/* Flags to qman_volatile_dequeue() */
-+#ifdef CONFIG_FSL_DPA_CAN_WAIT
-+#define QMAN_VOLATILE_FLAG_WAIT 0x00000001 /* wait if VDQCR is in use */
-+#define QMAN_VOLATILE_FLAG_WAIT_INT 0x00000002 /* if wait, interruptible? */
-+#define QMAN_VOLATILE_FLAG_FINISH 0x00000004 /* wait till VDQCR completes */
-+#endif
-+
-+/* Flags to qman_enqueue(). NB, the strange numbering is to align with hardware,
-+ * bit-wise. (NB: the PME API is sensitive to these precise numberings too, so
-+ * any change here should be audited in PME.) */
-+#ifdef CONFIG_FSL_DPA_CAN_WAIT
-+#define QMAN_ENQUEUE_FLAG_WAIT 0x00010000 /* wait if EQCR is full */
-+#define QMAN_ENQUEUE_FLAG_WAIT_INT 0x00020000 /* if wait, interruptible? */
-+#ifdef CONFIG_FSL_DPA_CAN_WAIT_SYNC
-+#define QMAN_ENQUEUE_FLAG_WAIT_SYNC 0x00000004 /* if wait, until consumed? */
-+#endif
-+#endif
-+#define QMAN_ENQUEUE_FLAG_WATCH_CGR 0x00080000 /* watch congestion state */
-+#define QMAN_ENQUEUE_FLAG_DCA 0x00008000 /* perform enqueue-DCA */
-+#define QMAN_ENQUEUE_FLAG_DCA_PARK 0x00004000 /* If DCA, requests park */
-+#define QMAN_ENQUEUE_FLAG_DCA_PTR(p) /* If DCA, p is DQRR entry */ \
-+ (((u32)(p) << 2) & 0x00000f00)
-+#define QMAN_ENQUEUE_FLAG_C_GREEN 0x00000000 /* choose one C_*** flag */
-+#define QMAN_ENQUEUE_FLAG_C_YELLOW 0x00000008
-+#define QMAN_ENQUEUE_FLAG_C_RED 0x00000010
-+#define QMAN_ENQUEUE_FLAG_C_OVERRIDE 0x00000018
-+/* For the ORP-specific qman_enqueue_orp() variant;
-+ * - this flag indicates "Not Last In Sequence", ie. all but the final fragment
-+ * of a frame. */
-+#define QMAN_ENQUEUE_FLAG_NLIS 0x01000000
-+/* - this flag performs no enqueue but fills in an ORP sequence number that
-+ * would otherwise block it (eg. if a frame has been dropped). */
-+#define QMAN_ENQUEUE_FLAG_HOLE 0x02000000
-+/* - this flag performs no enqueue but advances NESN to the given sequence
-+ * number. */
-+#define QMAN_ENQUEUE_FLAG_NESN 0x04000000
-+
-+/* Flags to qman_modify_cgr() */
-+#define QMAN_CGR_FLAG_USE_INIT 0x00000001
-+#define QMAN_CGR_MODE_FRAME 0x00000001
-+
-+ /* Portal Management */
-+ /* ----------------- */
-+/**
-+ * qman_get_portal_config - get portal configuration settings
-+ *
-+ * This returns a read-only view of the current cpu's affine portal settings.
-+ */
-+const struct qman_portal_config *qman_get_portal_config(void);
-+
-+/**
-+ * qman_irqsource_get - return the portal work that is interrupt-driven
-+ *
-+ * Returns a bitmask of QM_PIRQ_**I processing sources that are currently
-+ * enabled for interrupt handling on the current cpu's affine portal. These
-+ * sources will trigger the portal interrupt and the interrupt handler (or a
-+ * tasklet/bottom-half it defers to) will perform the corresponding processing
-+ * work. The qman_poll_***() functions will only process sources that are not in
-+ * this bitmask. If the current CPU is sharing a portal hosted on another CPU,
-+ * this always returns zero.
-+ */
-+u32 qman_irqsource_get(void);
-+
-+/**
-+ * qman_irqsource_add - add processing sources to be interrupt-driven
-+ * @bits: bitmask of QM_PIRQ_**I processing sources
-+ *
-+ * Adds processing sources that should be interrupt-driven (rather than
-+ * processed via qman_poll_***() functions). Returns zero for success, or
-+ * -EINVAL if the current CPU is sharing a portal hosted on another CPU.
-+ */
-+int qman_irqsource_add(u32 bits);
-+
-+/**
-+ * qman_irqsource_remove - remove processing sources from being interrupt-driven
-+ * @bits: bitmask of QM_PIRQ_**I processing sources
-+ *
-+ * Removes processing sources from being interrupt-driven, so that they will
-+ * instead be processed via qman_poll_***() functions. Returns zero for success,
-+ * or -EINVAL if the current CPU is sharing a portal hosted on another CPU.
-+ */
-+int qman_irqsource_remove(u32 bits);
-+
-+/**
-+ * qman_affine_cpus - return a mask of cpus that have affine portals
-+ */
-+const cpumask_t *qman_affine_cpus(void);
-+
-+/**
-+ * qman_affine_channel - return the channel ID of an portal
-+ * @cpu: the cpu whose affine portal is the subject of the query
-+ *
-+ * If @cpu is -1, the affine portal for the current CPU will be used. It is a
-+ * bug to call this function for any value of @cpu (other than -1) that is not a
-+ * member of the mask returned from qman_affine_cpus().
-+ */
-+u16 qman_affine_channel(int cpu);
-+
-+/**
-+ * qman_get_affine_portal - return the portal pointer affine to cpu
-+ * @cpu: the cpu whose affine portal is the subject of the query
-+ *
-+ */
-+void *qman_get_affine_portal(int cpu);
-+
-+/**
-+ * qman_poll_dqrr - process DQRR (fast-path) entries
-+ * @limit: the maximum number of DQRR entries to process
-+ *
-+ * Use of this function requires that DQRR processing not be interrupt-driven.
-+ * Ie. the value returned by qman_irqsource_get() should not include
-+ * QM_PIRQ_DQRI. If the current CPU is sharing a portal hosted on another CPU,
-+ * this function will return -EINVAL, otherwise the return value is >=0 and
-+ * represents the number of DQRR entries processed.
-+ */
-+int qman_poll_dqrr(unsigned int limit);
-+
-+/**
-+ * qman_poll_slow - process anything (except DQRR) that isn't interrupt-driven.
-+ *
-+ * This function does any portal processing that isn't interrupt-driven. If the
-+ * current CPU is sharing a portal hosted on another CPU, this function will
-+ * return (u32)-1, otherwise the return value is a bitmask of QM_PIRQ_* sources
-+ * indicating what interrupt sources were actually processed by the call.
-+ */
-+u32 qman_poll_slow(void);
-+
-+/**
-+ * qman_poll - legacy wrapper for qman_poll_dqrr() and qman_poll_slow()
-+ *
-+ * Dispatcher logic on a cpu can use this to trigger any maintenance of the
-+ * affine portal. There are two classes of portal processing in question;
-+ * fast-path (which involves demuxing dequeue ring (DQRR) entries and tracking
-+ * enqueue ring (EQCR) consumption), and slow-path (which involves EQCR
-+ * thresholds, congestion state changes, etc). This function does whatever
-+ * processing is not triggered by interrupts.
-+ *
-+ * Note, if DQRR and some slow-path processing are poll-driven (rather than
-+ * interrupt-driven) then this function uses a heuristic to determine how often
-+ * to run slow-path processing - as slow-path processing introduces at least a
-+ * minimum latency each time it is run, whereas fast-path (DQRR) processing is
-+ * close to zero-cost if there is no work to be done. Applications can tune this
-+ * behaviour themselves by using qman_poll_dqrr() and qman_poll_slow() directly
-+ * rather than going via this wrapper.
-+ */
-+void qman_poll(void);
-+
-+/**
-+ * qman_stop_dequeues - Stop h/w dequeuing to the s/w portal
-+ *
-+ * Disables DQRR processing of the portal. This is reference-counted, so
-+ * qman_start_dequeues() must be called as many times as qman_stop_dequeues() to
-+ * truly re-enable dequeuing.
-+ */
-+void qman_stop_dequeues(void);
-+
-+/**
-+ * qman_start_dequeues - (Re)start h/w dequeuing to the s/w portal
-+ *
-+ * Enables DQRR processing of the portal. This is reference-counted, so
-+ * qman_start_dequeues() must be called as many times as qman_stop_dequeues() to
-+ * truly re-enable dequeuing.
-+ */
-+void qman_start_dequeues(void);
-+
-+/**
-+ * qman_static_dequeue_add - Add pool channels to the portal SDQCR
-+ * @pools: bit-mask of pool channels, using QM_SDQCR_CHANNELS_POOL(n)
-+ *
-+ * Adds a set of pool channels to the portal's static dequeue command register
-+ * (SDQCR). The requested pools are limited to those the portal has dequeue
-+ * access to.
-+ */
-+void qman_static_dequeue_add(u32 pools);
-+
-+/**
-+ * qman_static_dequeue_del - Remove pool channels from the portal SDQCR
-+ * @pools: bit-mask of pool channels, using QM_SDQCR_CHANNELS_POOL(n)
-+ *
-+ * Removes a set of pool channels from the portal's static dequeue command
-+ * register (SDQCR). The requested pools are limited to those the portal has
-+ * dequeue access to.
-+ */
-+void qman_static_dequeue_del(u32 pools);
-+
-+/**
-+ * qman_static_dequeue_get - return the portal's current SDQCR
-+ *
-+ * Returns the portal's current static dequeue command register (SDQCR). The
-+ * entire register is returned, so if only the currently-enabled pool channels
-+ * are desired, mask the return value with QM_SDQCR_CHANNELS_POOL_MASK.
-+ */
-+u32 qman_static_dequeue_get(void);
-+
-+/**
-+ * qman_dca - Perform a Discrete Consumption Acknowledgement
-+ * @dq: the DQRR entry to be consumed
-+ * @park_request: indicates whether the held-active @fq should be parked
-+ *
-+ * Only allowed in DCA-mode portals, for DQRR entries whose handler callback had
-+ * previously returned 'qman_cb_dqrr_defer'. NB, as with the other APIs, this
-+ * does not take a 'portal' argument but implies the core affine portal from the
-+ * cpu that is currently executing the function. For reasons of locking, this
-+ * function must be called from the same CPU as that which processed the DQRR
-+ * entry in the first place.
-+ */
-+void qman_dca(struct qm_dqrr_entry *dq, int park_request);
-+
-+/**
-+ * qman_eqcr_is_empty - Determine if portal's EQCR is empty
-+ *
-+ * For use in situations where a cpu-affine caller needs to determine when all
-+ * enqueues for the local portal have been processed by Qman but can't use the
-+ * QMAN_ENQUEUE_FLAG_WAIT_SYNC flag to do this from the final qman_enqueue().
-+ * The function forces tracking of EQCR consumption (which normally doesn't
-+ * happen until enqueue processing needs to find space to put new enqueue
-+ * commands), and returns zero if the ring still has unprocessed entries,
-+ * non-zero if it is empty.
-+ */
-+int qman_eqcr_is_empty(void);
-+
-+/**
-+ * qman_set_dc_ern - Set the handler for DCP enqueue rejection notifications
-+ * @handler: callback for processing DCP ERNs
-+ * @affine: whether this handler is specific to the locally affine portal
-+ *
-+ * If a hardware block's interface to Qman (ie. its direct-connect portal, or
-+ * DCP) is configured not to receive enqueue rejections, then any enqueues
-+ * through that DCP that are rejected will be sent to a given software portal.
-+ * If @affine is non-zero, then this handler will only be used for DCP ERNs
-+ * received on the portal affine to the current CPU. If multiple CPUs share a
-+ * portal and they all call this function, they will be setting the handler for
-+ * the same portal! If @affine is zero, then this handler will be global to all
-+ * portals handled by this instance of the driver. Only those portals that do
-+ * not have their own affine handler will use the global handler.
-+ */
-+void qman_set_dc_ern(qman_cb_dc_ern handler, int affine);
-+
-+ /* FQ management */
-+ /* ------------- */
-+/**
-+ * qman_create_fq - Allocates a FQ
-+ * @fqid: the index of the FQD to encapsulate, must be "Out of Service"
-+ * @flags: bit-mask of QMAN_FQ_FLAG_*** options
-+ * @fq: memory for storing the 'fq', with callbacks filled in
-+ *
-+ * Creates a frame queue object for the given @fqid, unless the
-+ * QMAN_FQ_FLAG_DYNAMIC_FQID flag is set in @flags, in which case a FQID is
-+ * dynamically allocated (or the function fails if none are available). Once
-+ * created, the caller should not touch the memory at 'fq' except as extended to
-+ * adjacent memory for user-defined fields (see the definition of "struct
-+ * qman_fq" for more info). NO_MODIFY is only intended for enqueuing to
-+ * pre-existing frame-queues that aren't to be otherwise interfered with, it
-+ * prevents all other modifications to the frame queue. The TO_DCPORTAL flag
-+ * causes the driver to honour any contextB modifications requested in the
-+ * qm_init_fq() API, as this indicates the frame queue will be consumed by a
-+ * direct-connect portal (PME, CAAM, or Fman). When frame queues are consumed by
-+ * software portals, the contextB field is controlled by the driver and can't be
-+ * modified by the caller. If the AS_IS flag is specified, management commands
-+ * will be used on portal @p to query state for frame queue @fqid and construct
-+ * a frame queue object based on that, rather than assuming/requiring that it be
-+ * Out of Service.
-+ */
-+int qman_create_fq(u32 fqid, u32 flags, struct qman_fq *fq);
-+
-+/**
-+ * qman_destroy_fq - Deallocates a FQ
-+ * @fq: the frame queue object to release
-+ * @flags: bit-mask of QMAN_FQ_FREE_*** options
-+ *
-+ * The memory for this frame queue object ('fq' provided in qman_create_fq()) is
-+ * not deallocated but the caller regains ownership, to do with as desired. The
-+ * FQ must be in the 'out-of-service' state unless the QMAN_FQ_FREE_PARKED flag
-+ * is specified, in which case it may also be in the 'parked' state.
-+ */
-+void qman_destroy_fq(struct qman_fq *fq, u32 flags);
-+
-+/**
-+ * qman_fq_fqid - Queries the frame queue ID of a FQ object
-+ * @fq: the frame queue object to query
-+ */
-+u32 qman_fq_fqid(struct qman_fq *fq);
-+
-+/**
-+ * qman_fq_state - Queries the state of a FQ object
-+ * @fq: the frame queue object to query
-+ * @state: pointer to state enum to return the FQ scheduling state
-+ * @flags: pointer to state flags to receive QMAN_FQ_STATE_*** bitmask
-+ *
-+ * Queries the state of the FQ object, without performing any h/w commands.
-+ * This captures the state, as seen by the driver, at the time the function
-+ * executes.
-+ */
-+void qman_fq_state(struct qman_fq *fq, enum qman_fq_state *state, u32 *flags);
-+
-+/**
-+ * qman_init_fq - Initialises FQ fields, leaves the FQ "parked" or "scheduled"
-+ * @fq: the frame queue object to modify, must be 'parked' or new.
-+ * @flags: bit-mask of QMAN_INITFQ_FLAG_*** options
-+ * @opts: the FQ-modification settings, as defined in the low-level API
-+ *
-+ * The @opts parameter comes from the low-level portal API. Select
-+ * QMAN_INITFQ_FLAG_SCHED in @flags to cause the frame queue to be scheduled
-+ * rather than parked. NB, @opts can be NULL.
-+ *
-+ * Note that some fields and options within @opts may be ignored or overwritten
-+ * by the driver;
-+ * 1. the 'count' and 'fqid' fields are always ignored (this operation only
-+ * affects one frame queue: @fq).
-+ * 2. the QM_INITFQ_WE_CONTEXTB option of the 'we_mask' field and the associated
-+ * 'fqd' structure's 'context_b' field are sometimes overwritten;
-+ * - if @fq was not created with QMAN_FQ_FLAG_TO_DCPORTAL, then context_b is
-+ * initialised to a value used by the driver for demux.
-+ * - if context_b is initialised for demux, so is context_a in case stashing
-+ * is requested (see item 4).
-+ * (So caller control of context_b is only possible for TO_DCPORTAL frame queue
-+ * objects.)
-+ * 3. if @flags contains QMAN_INITFQ_FLAG_LOCAL, the 'fqd' structure's
-+ * 'dest::channel' field will be overwritten to match the portal used to issue
-+ * the command. If the WE_DESTWQ write-enable bit had already been set by the
-+ * caller, the channel workqueue will be left as-is, otherwise the write-enable
-+ * bit is set and the workqueue is set to a default of 4. If the "LOCAL" flag
-+ * isn't set, the destination channel/workqueue fields and the write-enable bit
-+ * are left as-is.
-+ * 4. if the driver overwrites context_a/b for demux, then if
-+ * QM_INITFQ_WE_CONTEXTA is set, the driver will only overwrite
-+ * context_a.address fields and will leave the stashing fields provided by the
-+ * user alone, otherwise it will zero out the context_a.stashing fields.
-+ */
-+int qman_init_fq(struct qman_fq *fq, u32 flags, struct qm_mcc_initfq *opts);
-+
-+/**
-+ * qman_schedule_fq - Schedules a FQ
-+ * @fq: the frame queue object to schedule, must be 'parked'
-+ *
-+ * Schedules the frame queue, which must be Parked, which takes it to
-+ * Tentatively-Scheduled or Truly-Scheduled depending on its fill-level.
-+ */
-+int qman_schedule_fq(struct qman_fq *fq);
-+
-+/**
-+ * qman_retire_fq - Retires a FQ
-+ * @fq: the frame queue object to retire
-+ * @flags: FQ flags (as per qman_fq_state) if retirement completes immediately
-+ *
-+ * Retires the frame queue. This returns zero if it succeeds immediately, +1 if
-+ * the retirement was started asynchronously, otherwise it returns negative for
-+ * failure. When this function returns zero, @flags is set to indicate whether
-+ * the retired FQ is empty and/or whether it has any ORL fragments (to show up
-+ * as ERNs). Otherwise the corresponding flags will be known when a subsequent
-+ * FQRN message shows up on the portal's message ring.
-+ *
-+ * NB, if the retirement is asynchronous (the FQ was in the Truly Scheduled or
-+ * Active state), the completion will be via the message ring as a FQRN - but
-+ * the corresponding callback may occur before this function returns!! Ie. the
-+ * caller should be prepared to accept the callback as the function is called,
-+ * not only once it has returned.
-+ */
-+int qman_retire_fq(struct qman_fq *fq, u32 *flags);
-+
-+/**
-+ * qman_oos_fq - Puts a FQ "out of service"
-+ * @fq: the frame queue object to be put out-of-service, must be 'retired'
-+ *
-+ * The frame queue must be retired and empty, and if any order restoration list
-+ * was released as ERNs at the time of retirement, they must all be consumed.
-+ */
-+int qman_oos_fq(struct qman_fq *fq);
-+
-+/**
-+ * qman_fq_flow_control - Set the XON/XOFF state of a FQ
-+ * @fq: the frame queue object to be set to XON/XOFF state, must not be 'oos',
-+ * or 'retired' or 'parked' state
-+ * @xon: boolean to set fq in XON or XOFF state
-+ *
-+ * The frame should be in Tentatively Scheduled state or Truly Schedule sate,
-+ * otherwise the IFSI interrupt will be asserted.
-+ */
-+int qman_fq_flow_control(struct qman_fq *fq, int xon);
-+
-+/**
-+ * qman_query_fq - Queries FQD fields (via h/w query command)
-+ * @fq: the frame queue object to be queried
-+ * @fqd: storage for the queried FQD fields
-+ */
-+int qman_query_fq(struct qman_fq *fq, struct qm_fqd *fqd);
-+
-+/**
-+ * qman_query_fq_np - Queries non-programmable FQD fields
-+ * @fq: the frame queue object to be queried
-+ * @np: storage for the queried FQD fields
-+ */
-+int qman_query_fq_np(struct qman_fq *fq, struct qm_mcr_queryfq_np *np);
-+
-+/**
-+ * qman_query_wq - Queries work queue lengths
-+ * @query_dedicated: If non-zero, query length of WQs in the channel dedicated
-+ * to this software portal. Otherwise, query length of WQs in a
-+ * channel specified in wq.
-+ * @wq: storage for the queried WQs lengths. Also specified the channel to
-+ * to query if query_dedicated is zero.
-+ */
-+int qman_query_wq(u8 query_dedicated, struct qm_mcr_querywq *wq);
-+
-+/**
-+ * qman_volatile_dequeue - Issue a volatile dequeue command
-+ * @fq: the frame queue object to dequeue from
-+ * @flags: a bit-mask of QMAN_VOLATILE_FLAG_*** options
-+ * @vdqcr: bit mask of QM_VDQCR_*** options, as per qm_dqrr_vdqcr_set()
-+ *
-+ * Attempts to lock access to the portal's VDQCR volatile dequeue functionality.
-+ * The function will block and sleep if QMAN_VOLATILE_FLAG_WAIT is specified and
-+ * the VDQCR is already in use, otherwise returns non-zero for failure. If
-+ * QMAN_VOLATILE_FLAG_FINISH is specified, the function will only return once
-+ * the VDQCR command has finished executing (ie. once the callback for the last
-+ * DQRR entry resulting from the VDQCR command has been called). If not using
-+ * the FINISH flag, completion can be determined either by detecting the
-+ * presence of the QM_DQRR_STAT_UNSCHEDULED and QM_DQRR_STAT_DQCR_EXPIRED bits
-+ * in the "stat" field of the "struct qm_dqrr_entry" passed to the FQ's dequeue
-+ * callback, or by waiting for the QMAN_FQ_STATE_VDQCR bit to disappear from the
-+ * "flags" retrieved from qman_fq_state().
-+ */
-+int qman_volatile_dequeue(struct qman_fq *fq, u32 flags, u32 vdqcr);
-+
-+/**
-+ * qman_enqueue - Enqueue a frame to a frame queue
-+ * @fq: the frame queue object to enqueue to
-+ * @fd: a descriptor of the frame to be enqueued
-+ * @flags: bit-mask of QMAN_ENQUEUE_FLAG_*** options
-+ *
-+ * Fills an entry in the EQCR of portal @qm to enqueue the frame described by
-+ * @fd. The descriptor details are copied from @fd to the EQCR entry, the 'pid'
-+ * field is ignored. The return value is non-zero on error, such as ring full
-+ * (and FLAG_WAIT not specified), congestion avoidance (FLAG_WATCH_CGR
-+ * specified), etc. If the ring is full and FLAG_WAIT is specified, this
-+ * function will block. If FLAG_INTERRUPT is set, the EQCI bit of the portal
-+ * interrupt will assert when Qman consumes the EQCR entry (subject to "status
-+ * disable", "enable", and "inhibit" registers). If FLAG_DCA is set, Qman will
-+ * perform an implied "discrete consumption acknowledgement" on the dequeue
-+ * ring's (DQRR) entry, at the ring index specified by the FLAG_DCA_IDX(x)
-+ * macro. (As an alternative to issuing explicit DCA actions on DQRR entries,
-+ * this implicit DCA can delay the release of a "held active" frame queue
-+ * corresponding to a DQRR entry until Qman consumes the EQCR entry - providing
-+ * order-preservation semantics in packet-forwarding scenarios.) If FLAG_DCA is
-+ * set, then FLAG_DCA_PARK can also be set to imply that the DQRR consumption
-+ * acknowledgement should "park request" the "held active" frame queue. Ie.
-+ * when the portal eventually releases that frame queue, it will be left in the
-+ * Parked state rather than Tentatively Scheduled or Truly Scheduled. If the
-+ * portal is watching congestion groups, the QMAN_ENQUEUE_FLAG_WATCH_CGR flag
-+ * is requested, and the FQ is a member of a congestion group, then this
-+ * function returns -EAGAIN if the congestion group is currently congested.
-+ * Note, this does not eliminate ERNs, as the async interface means we can be
-+ * sending enqueue commands to an un-congested FQ that becomes congested before
-+ * the enqueue commands are processed, but it does minimise needless thrashing
-+ * of an already busy hardware resource by throttling many of the to-be-dropped
-+ * enqueues "at the source".
-+ */
-+int qman_enqueue(struct qman_fq *fq, const struct qm_fd *fd, u32 flags);
-+
-+typedef int (*qman_cb_precommit) (void *arg);
-+/**
-+ * qman_enqueue_precommit - Enqueue a frame to a frame queue and call cb
-+ * @fq: the frame queue object to enqueue to
-+ * @fd: a descriptor of the frame to be enqueued
-+ * @flags: bit-mask of QMAN_ENQUEUE_FLAG_*** options
-+ * @cb: user supplied callback function to invoke before writing commit verb.
-+ * @cb_arg: callback function argument
-+ *
-+ * This is similar to qman_enqueue except that it will invoke a user supplied
-+ * callback function just before writng the commit verb. This is useful
-+ * when the user want to do something *just before* enqueuing the request and
-+ * the enqueue can't fail.
-+ */
-+int qman_enqueue_precommit(struct qman_fq *fq, const struct qm_fd *fd,
-+ u32 flags, qman_cb_precommit cb, void *cb_arg);
-+
-+/**
-+ * qman_enqueue_orp - Enqueue a frame to a frame queue using an ORP
-+ * @fq: the frame queue object to enqueue to
-+ * @fd: a descriptor of the frame to be enqueued
-+ * @flags: bit-mask of QMAN_ENQUEUE_FLAG_*** options
-+ * @orp: the frame queue object used as an order restoration point.
-+ * @orp_seqnum: the sequence number of this frame in the order restoration path
-+ *
-+ * Similar to qman_enqueue(), but with the addition of an Order Restoration
-+ * Point (@orp) and corresponding sequence number (@orp_seqnum) for this
-+ * enqueue operation to employ order restoration. Each frame queue object acts
-+ * as an Order Definition Point (ODP) by providing each frame dequeued from it
-+ * with an incrementing sequence number, this value is generally ignored unless
-+ * that sequence of dequeued frames will need order restoration later. Each
-+ * frame queue object also encapsulates an Order Restoration Point (ORP), which
-+ * is a re-assembly context for re-ordering frames relative to their sequence
-+ * numbers as they are enqueued. The ORP does not have to be within the frame
-+ * queue that receives the enqueued frame, in fact it is usually the frame
-+ * queue from which the frames were originally dequeued. For the purposes of
-+ * order restoration, multiple frames (or "fragments") can be enqueued for a
-+ * single sequence number by setting the QMAN_ENQUEUE_FLAG_NLIS flag for all
-+ * enqueues except the final fragment of a given sequence number. Ordering
-+ * between sequence numbers is guaranteed, even if fragments of different
-+ * sequence numbers are interlaced with one another. Fragments of the same
-+ * sequence number will retain the order in which they are enqueued. If no
-+ * enqueue is to performed, QMAN_ENQUEUE_FLAG_HOLE indicates that the given
-+ * sequence number is to be "skipped" by the ORP logic (eg. if a frame has been
-+ * dropped from a sequence), or QMAN_ENQUEUE_FLAG_NESN indicates that the given
-+ * sequence number should become the ORP's "Next Expected Sequence Number".
-+ *
-+ * Side note: a frame queue object can be used purely as an ORP, without
-+ * carrying any frames at all. Care should be taken not to deallocate a frame
-+ * queue object that is being actively used as an ORP, as a future allocation
-+ * of the frame queue object may start using the internal ORP before the
-+ * previous use has finished.
-+ */
-+int qman_enqueue_orp(struct qman_fq *fq, const struct qm_fd *fd, u32 flags,
-+ struct qman_fq *orp, u16 orp_seqnum);
-+
-+/**
-+ * qman_alloc_fqid_range - Allocate a contiguous range of FQIDs
-+ * @result: is set by the API to the base FQID of the allocated range
-+ * @count: the number of FQIDs required
-+ * @align: required alignment of the allocated range
-+ * @partial: non-zero if the API can return fewer than @count FQIDs
-+ *
-+ * Returns the number of frame queues allocated, or a negative error code. If
-+ * @partial is non zero, the allocation request may return a smaller range of
-+ * FQs than requested (though alignment will be as requested). If @partial is
-+ * zero, the return value will either be 'count' or negative.
-+ */
-+int qman_alloc_fqid_range(u32 *result, u32 count, u32 align, int partial);
-+static inline int qman_alloc_fqid(u32 *result)
-+{
-+ int ret = qman_alloc_fqid_range(result, 1, 0, 0);
-+ return (ret > 0) ? 0 : ret;
-+}
-+
-+/**
-+ * qman_release_fqid_range - Release the specified range of frame queue IDs
-+ * @fqid: the base FQID of the range to deallocate
-+ * @count: the number of FQIDs in the range
-+ *
-+ * This function can also be used to seed the allocator with ranges of FQIDs
-+ * that it can subsequently allocate from.
-+ */
-+void qman_release_fqid_range(u32 fqid, unsigned int count);
-+static inline void qman_release_fqid(u32 fqid)
-+{
-+ qman_release_fqid_range(fqid, 1);
-+}
-+
-+void qman_seed_fqid_range(u32 fqid, unsigned int count);
-+
-+
-+int qman_shutdown_fq(u32 fqid);
-+
-+/**
-+ * qman_reserve_fqid_range - Reserve the specified range of frame queue IDs
-+ * @fqid: the base FQID of the range to deallocate
-+ * @count: the number of FQIDs in the range
-+ */
-+int qman_reserve_fqid_range(u32 fqid, unsigned int count);
-+static inline int qman_reserve_fqid(u32 fqid)
-+{
-+ return qman_reserve_fqid_range(fqid, 1);
-+}
-+
-+ /* Pool-channel management */
-+ /* ----------------------- */
-+/**
-+ * qman_alloc_pool_range - Allocate a contiguous range of pool-channel IDs
-+ * @result: is set by the API to the base pool-channel ID of the allocated range
-+ * @count: the number of pool-channel IDs required
-+ * @align: required alignment of the allocated range
-+ * @partial: non-zero if the API can return fewer than @count
-+ *
-+ * Returns the number of pool-channel IDs allocated, or a negative error code.
-+ * If @partial is non zero, the allocation request may return a smaller range of
-+ * than requested (though alignment will be as requested). If @partial is zero,
-+ * the return value will either be 'count' or negative.
-+ */
-+int qman_alloc_pool_range(u32 *result, u32 count, u32 align, int partial);
-+static inline int qman_alloc_pool(u32 *result)
-+{
-+ int ret = qman_alloc_pool_range(result, 1, 0, 0);
-+ return (ret > 0) ? 0 : ret;
-+}
-+
-+/**
-+ * qman_release_pool_range - Release the specified range of pool-channel IDs
-+ * @id: the base pool-channel ID of the range to deallocate
-+ * @count: the number of pool-channel IDs in the range
-+ */
-+void qman_release_pool_range(u32 id, unsigned int count);
-+static inline void qman_release_pool(u32 id)
-+{
-+ qman_release_pool_range(id, 1);
-+}
-+
-+/**
-+ * qman_reserve_pool_range - Reserve the specified range of pool-channel IDs
-+ * @id: the base pool-channel ID of the range to reserve
-+ * @count: the number of pool-channel IDs in the range
-+ */
-+int qman_reserve_pool_range(u32 id, unsigned int count);
-+static inline int qman_reserve_pool(u32 id)
-+{
-+ return qman_reserve_pool_range(id, 1);
-+}
-+
-+void qman_seed_pool_range(u32 id, unsigned int count);
-+
-+ /* CGR management */
-+ /* -------------- */
-+/**
-+ * qman_create_cgr - Register a congestion group object
-+ * @cgr: the 'cgr' object, with fields filled in
-+ * @flags: QMAN_CGR_FLAG_* values
-+ * @opts: optional state of CGR settings
-+ *
-+ * Registers this object to receiving congestion entry/exit callbacks on the
-+ * portal affine to the cpu portal on which this API is executed. If opts is
-+ * NULL then only the callback (cgr->cb) function is registered. If @flags
-+ * contains QMAN_CGR_FLAG_USE_INIT, then an init hw command (which will reset
-+ * any unspecified parameters) will be used rather than a modify hw hardware
-+ * (which only modifies the specified parameters).
-+ */
-+int qman_create_cgr(struct qman_cgr *cgr, u32 flags,
-+ struct qm_mcc_initcgr *opts);
-+
-+/**
-+ * qman_create_cgr_to_dcp - Register a congestion group object to DCP portal
-+ * @cgr: the 'cgr' object, with fields filled in
-+ * @flags: QMAN_CGR_FLAG_* values
-+ * @dcp_portal: the DCP portal to which the cgr object is registered.
-+ * @opts: optional state of CGR settings
-+ *
-+ */
-+int qman_create_cgr_to_dcp(struct qman_cgr *cgr, u32 flags, u16 dcp_portal,
-+ struct qm_mcc_initcgr *opts);
-+
-+/**
-+ * qman_delete_cgr - Deregisters a congestion group object
-+ * @cgr: the 'cgr' object to deregister
-+ *
-+ * "Unplugs" this CGR object from the portal affine to the cpu on which this API
-+ * is executed. This must be excuted on the same affine portal on which it was
-+ * created.
-+ */
-+int qman_delete_cgr(struct qman_cgr *cgr);
-+
-+/**
-+ * qman_delete_cgr_safe - Deregisters a congestion group object from any CPU
-+ * @cgr: the 'cgr' object to deregister
-+ *
-+ * This will select the proper CPU and run there qman_delete_cgr().
-+ */
-+void qman_delete_cgr_safe(struct qman_cgr *cgr);
-+
-+/**
-+ * qman_modify_cgr - Modify CGR fields
-+ * @cgr: the 'cgr' object to modify
-+ * @flags: QMAN_CGR_FLAG_* values
-+ * @opts: the CGR-modification settings
-+ *
-+ * The @opts parameter comes from the low-level portal API, and can be NULL.
-+ * Note that some fields and options within @opts may be ignored or overwritten
-+ * by the driver, in particular the 'cgrid' field is ignored (this operation
-+ * only affects the given CGR object). If @flags contains
-+ * QMAN_CGR_FLAG_USE_INIT, then an init hw command (which will reset any
-+ * unspecified parameters) will be used rather than a modify hw hardware (which
-+ * only modifies the specified parameters).
-+ */
-+int qman_modify_cgr(struct qman_cgr *cgr, u32 flags,
-+ struct qm_mcc_initcgr *opts);
-+
-+/**
-+* qman_query_cgr - Queries CGR fields
-+* @cgr: the 'cgr' object to query
-+* @result: storage for the queried congestion group record
-+*/
-+int qman_query_cgr(struct qman_cgr *cgr, struct qm_mcr_querycgr *result);
-+
-+/**
-+ * qman_query_congestion - Queries the state of all congestion groups
-+ * @congestion: storage for the queried state of all congestion groups
-+ */
-+int qman_query_congestion(struct qm_mcr_querycongestion *congestion);
-+
-+/**
-+ * qman_alloc_cgrid_range - Allocate a contiguous range of CGR IDs
-+ * @result: is set by the API to the base CGR ID of the allocated range
-+ * @count: the number of CGR IDs required
-+ * @align: required alignment of the allocated range
-+ * @partial: non-zero if the API can return fewer than @count
-+ *
-+ * Returns the number of CGR IDs allocated, or a negative error code.
-+ * If @partial is non zero, the allocation request may return a smaller range of
-+ * than requested (though alignment will be as requested). If @partial is zero,
-+ * the return value will either be 'count' or negative.
-+ */
-+int qman_alloc_cgrid_range(u32 *result, u32 count, u32 align, int partial);
-+static inline int qman_alloc_cgrid(u32 *result)
-+{
-+ int ret = qman_alloc_cgrid_range(result, 1, 0, 0);
-+ return (ret > 0) ? 0 : ret;
-+}
-+
-+/**
-+ * qman_release_cgrid_range - Release the specified range of CGR IDs
-+ * @id: the base CGR ID of the range to deallocate
-+ * @count: the number of CGR IDs in the range
-+ */
-+void qman_release_cgrid_range(u32 id, unsigned int count);
-+static inline void qman_release_cgrid(u32 id)
-+{
-+ qman_release_cgrid_range(id, 1);
-+}
-+
-+/**
-+ * qman_reserve_cgrid_range - Reserve the specified range of CGR ID
-+ * @id: the base CGR ID of the range to reserve
-+ * @count: the number of CGR IDs in the range
-+ */
-+int qman_reserve_cgrid_range(u32 id, unsigned int count);
-+static inline int qman_reserve_cgrid(u32 id)
-+{
-+ return qman_reserve_cgrid_range(id, 1);
-+}
-+
-+void qman_seed_cgrid_range(u32 id, unsigned int count);
-+
-+
-+ /* Helpers */
-+ /* ------- */
-+/**
-+ * qman_poll_fq_for_init - Check if an FQ has been initialised from OOS
-+ * @fqid: the FQID that will be initialised by other s/w
-+ *
-+ * In many situations, a FQID is provided for communication between s/w
-+ * entities, and whilst the consumer is responsible for initialising and
-+ * scheduling the FQ, the producer(s) generally create a wrapper FQ object using
-+ * and only call qman_enqueue() (no FQ initialisation, scheduling, etc). Ie;
-+ * qman_create_fq(..., QMAN_FQ_FLAG_NO_MODIFY, ...);
-+ * However, data can not be enqueued to the FQ until it is initialised out of
-+ * the OOS state - this function polls for that condition. It is particularly
-+ * useful for users of IPC functions - each endpoint's Rx FQ is the other
-+ * endpoint's Tx FQ, so each side can initialise and schedule their Rx FQ object
-+ * and then use this API on the (NO_MODIFY) Tx FQ object in order to
-+ * synchronise. The function returns zero for success, +1 if the FQ is still in
-+ * the OOS state, or negative if there was an error.
-+ */
-+static inline int qman_poll_fq_for_init(struct qman_fq *fq)
-+{
-+ struct qm_mcr_queryfq_np np;
-+ int err;
-+ err = qman_query_fq_np(fq, &np);
-+ if (err)
-+ return err;
-+ if ((np.state & QM_MCR_NP_STATE_MASK) == QM_MCR_NP_STATE_OOS)
-+ return 1;
-+ return 0;
-+}
-+
-+ /* -------------- */
-+ /* CEETM :: types */
-+ /* -------------- */
-+/**
-+ * Token Rate Structure
-+ * Shaping rates are based on a "credit" system and a pre-configured h/w
-+ * internal timer. The following type represents a shaper "rate" parameter as a
-+ * fractional number of "tokens". Here's how it works. This (fractional) number
-+ * of tokens is added to the shaper's "credit" every time the h/w timer elapses
-+ * (up to a limit which is set by another shaper parameter). Every time a frame
-+ * is enqueued through a shaper, the shaper deducts as many tokens as there are
-+ * bytes of data in the enqueued frame. A shaper will not allow itself to
-+ * enqueue any frames if its token count is negative. As such;
-+ *
-+ * The rate at which data is enqueued is limited by the
-+ * rate at which tokens are added.
-+ *
-+ * Therefore if the user knows the period between these h/w timer updates in
-+ * seconds, they can calculate the maximum traffic rate of the shaper (in
-+ * bytes-per-second) from the token rate. And vice versa, they can calculate
-+ * the token rate to use in order to achieve a given traffic rate.
-+ */
-+struct qm_ceetm_rate {
-+ /* The token rate is; whole + (fraction/8192) */
-+ u32 whole:11; /* 0..2047 */
-+ u32 fraction:13; /* 0..8191 */
-+};
-+
-+struct qm_ceetm_weight_code {
-+ /* The weight code is; 5 msbits + 3 lsbits */
-+ u8 y:5;
-+ u8 x:3;
-+};
-+
-+struct qm_ceetm {
-+ unsigned int idx;
-+ struct list_head sub_portals;
-+ struct list_head lnis;
-+ unsigned int sp_range[2];
-+ unsigned int lni_range[2];
-+};
-+
-+struct qm_ceetm_sp {
-+ struct list_head node;
-+ unsigned int idx;
-+ unsigned int dcp_idx;
-+ int is_claimed;
-+ struct qm_ceetm_lni *lni;
-+};
-+
-+/* Logical Network Interface */
-+struct qm_ceetm_lni {
-+ struct list_head node;
-+ unsigned int idx;
-+ unsigned int dcp_idx;
-+ int is_claimed;
-+ struct qm_ceetm_sp *sp;
-+ struct list_head channels;
-+ int shaper_enable;
-+ int shaper_couple;
-+ int oal;
-+ struct qm_ceetm_rate cr_token_rate;
-+ struct qm_ceetm_rate er_token_rate;
-+ u16 cr_token_bucket_limit;
-+ u16 er_token_bucket_limit;
-+};
-+
-+/* Class Queue Channel */
-+struct qm_ceetm_channel {
-+ struct list_head node;
-+ unsigned int idx;
-+ unsigned int lni_idx;
-+ unsigned int dcp_idx;
-+ struct list_head class_queues;
-+ struct list_head ccgs;
-+ u8 shaper_enable;
-+ u8 shaper_couple;
-+ struct qm_ceetm_rate cr_token_rate;
-+ struct qm_ceetm_rate er_token_rate;
-+ u16 cr_token_bucket_limit;
-+ u16 er_token_bucket_limit;
-+};
-+
-+struct qm_ceetm_ccg;
-+
-+/* This callback type is used when handling congestion entry/exit. The
-+ * 'cb_ctx' value is the opaque value associated with ccg object.
-+ * 'congested' is non-zero on congestion-entry, and zero on congestion-exit.
-+ */
-+typedef void (*qman_cb_ccgr)(struct qm_ceetm_ccg *ccg, void *cb_ctx,
-+ int congested);
-+
-+/* Class Congestion Group */
-+struct qm_ceetm_ccg {
-+ struct qm_ceetm_channel *parent;
-+ struct list_head node;
-+ struct list_head cb_node;
-+ qman_cb_ccgr cb;
-+ void *cb_ctx;
-+ unsigned int idx;
-+};
-+
-+/* Class Queue */
-+struct qm_ceetm_cq {
-+ struct qm_ceetm_channel *parent;
-+ struct qm_ceetm_ccg *ccg;
-+ struct list_head node;
-+ unsigned int idx;
-+ int is_claimed;
-+ struct list_head bound_lfqids;
-+ struct list_head binding_node;
-+};
-+
-+/* Logical Frame Queue */
-+struct qm_ceetm_lfq {
-+ struct qm_ceetm_channel *parent;
-+ struct list_head node;
-+ unsigned int idx;
-+ unsigned int dctidx;
-+ u64 context_a;
-+ u32 context_b;
-+ qman_cb_mr ern;
-+};
-+
-+/**
-+ * qman_ceetm_bps2tokenrate - Given a desired rate 'bps' measured in bps
-+ * (ie. bits-per-second), compute the 'token_rate' fraction that best
-+ * approximates that rate.
-+ * @bps: the desired shaper rate in bps.
-+ * @token_rate: the output token rate computed with the given kbps.
-+ * @rounding: dictates how to round if an exact conversion is not possible; if
-+ * it is negative then 'token_rate' will round down to the highest value that
-+ * does not exceed the desired rate, if it is positive then 'token_rate' will
-+ * round up to the lowest value that is greater than or equal to the desired
-+ * rate, and if it is zero then it will round to the nearest approximation,
-+ * whether that be up or down.
-+ *
-+ * Return 0 for success, or -EINVAL if prescaler or qman clock is not available.
-+ */
-+int qman_ceetm_bps2tokenrate(u64 bps,
-+ struct qm_ceetm_rate *token_rate,
-+ int rounding);
-+
-+/**
-+ * qman_ceetm_tokenrate2bps - Given a 'token_rate', compute the
-+ * corresponding number of 'bps'.
-+ * @token_rate: the input desired token_rate fraction.
-+ * @bps: the output shaper rate in bps computed with the give token rate.
-+ * @rounding: has the same semantics as the previous function.
-+ *
-+ * Return 0 for success, or -EINVAL if prescaler or qman clock is not available.
-+ */
-+int qman_ceetm_tokenrate2bps(const struct qm_ceetm_rate *token_rate,
-+ u64 *bps,
-+ int rounding);
-+
-+int qman_alloc_ceetm0_channel_range(u32 *result, u32 count, u32 align,
-+ int partial);
-+static inline int qman_alloc_ceetm0_channel(u32 *result)
-+{
-+ int ret = qman_alloc_ceetm0_channel_range(result, 1, 0, 0);
-+ return (ret > 0) ? 0 : ret;
-+}
-+void qman_release_ceetm0_channel_range(u32 channelid, u32 count);
-+static inline void qman_release_ceetm0_channelid(u32 channelid)
-+{
-+ qman_release_ceetm0_channel_range(channelid, 1);
-+}
-+
-+int qman_reserve_ceetm0_channel_range(u32 channelid, u32 count);
-+static inline int qman_reserve_ceetm0_channelid(u32 channelid)
-+{
-+ return qman_reserve_ceetm0_channel_range(channelid, 1);
-+}
-+
-+void qman_seed_ceetm0_channel_range(u32 channelid, u32 count);
-+
-+
-+int qman_alloc_ceetm1_channel_range(u32 *result, u32 count, u32 align,
-+ int partial);
-+static inline int qman_alloc_ceetm1_channel(u32 *result)
-+{
-+ int ret = qman_alloc_ceetm1_channel_range(result, 1, 0, 0);
-+ return (ret > 0) ? 0 : ret;
-+}
-+void qman_release_ceetm1_channel_range(u32 channelid, u32 count);
-+static inline void qman_release_ceetm1_channelid(u32 channelid)
-+{
-+ qman_release_ceetm1_channel_range(channelid, 1);
-+}
-+int qman_reserve_ceetm1_channel_range(u32 channelid, u32 count);
-+static inline int qman_reserve_ceetm1_channelid(u32 channelid)
-+{
-+ return qman_reserve_ceetm1_channel_range(channelid, 1);
-+}
-+
-+void qman_seed_ceetm1_channel_range(u32 channelid, u32 count);
-+
-+
-+int qman_alloc_ceetm0_lfqid_range(u32 *result, u32 count, u32 align,
-+ int partial);
-+static inline int qman_alloc_ceetm0_lfqid(u32 *result)
-+{
-+ int ret = qman_alloc_ceetm0_lfqid_range(result, 1, 0, 0);
-+ return (ret > 0) ? 0 : ret;
-+}
-+void qman_release_ceetm0_lfqid_range(u32 lfqid, u32 count);
-+static inline void qman_release_ceetm0_lfqid(u32 lfqid)
-+{
-+ qman_release_ceetm0_lfqid_range(lfqid, 1);
-+}
-+int qman_reserve_ceetm0_lfqid_range(u32 lfqid, u32 count);
-+static inline int qman_reserve_ceetm0_lfqid(u32 lfqid)
-+{
-+ return qman_reserve_ceetm0_lfqid_range(lfqid, 1);
-+}
-+
-+void qman_seed_ceetm0_lfqid_range(u32 lfqid, u32 count);
-+
-+
-+int qman_alloc_ceetm1_lfqid_range(u32 *result, u32 count, u32 align,
-+ int partial);
-+static inline int qman_alloc_ceetm1_lfqid(u32 *result)
-+{
-+ int ret = qman_alloc_ceetm1_lfqid_range(result, 1, 0, 0);
-+ return (ret > 0) ? 0 : ret;
-+}
-+void qman_release_ceetm1_lfqid_range(u32 lfqid, u32 count);
-+static inline void qman_release_ceetm1_lfqid(u32 lfqid)
-+{
-+ qman_release_ceetm1_lfqid_range(lfqid, 1);
-+}
-+int qman_reserve_ceetm1_lfqid_range(u32 lfqid, u32 count);
-+static inline int qman_reserve_ceetm1_lfqid(u32 lfqid)
-+{
-+ return qman_reserve_ceetm1_lfqid_range(lfqid, 1);
-+}
-+
-+void qman_seed_ceetm1_lfqid_range(u32 lfqid, u32 count);
-+
-+
-+ /* ----------------------------- */
-+ /* CEETM :: sub-portals */
-+ /* ----------------------------- */
-+
-+/**
-+ * qman_ceetm_sp_claim - Claims the given sub-portal, provided it is available
-+ * to us and configured for traffic-management.
-+ * @sp: the returned sub-portal object, if successful.
-+ * @dcp_id: specifies the desired Fman block (and thus the relevant CEETM
-+ * instance),
-+ * @sp_idx" is the desired sub-portal index from 0 to 15.
-+ *
-+ * Returns zero for success, or -ENODEV if the sub-portal is in use, or -EINVAL
-+ * if the sp_idx is out of range.
-+ *
-+ * Note that if there are multiple driver domains (eg. a linux kernel versus
-+ * user-space drivers in USDPAA, or multiple guests running under a hypervisor)
-+ * then a sub-portal may be accessible by more than one instance of a qman
-+ * driver and so it may be claimed multiple times. If this is the case, it is
-+ * up to the system architect to prevent conflicting configuration actions
-+ * coming from the different driver domains. The qman drivers do not have any
-+ * behind-the-scenes coordination to prevent this from happening.
-+ */
-+int qman_ceetm_sp_claim(struct qm_ceetm_sp **sp,
-+ enum qm_dc_portal dcp_idx,
-+ unsigned int sp_idx);
-+
-+/**
-+ * qman_ceetm_sp_release - Releases a previously claimed sub-portal.
-+ * @sp: the sub-portal to be released.
-+ *
-+ * Returns 0 for success, or -EBUSY for failure if the dependencies are not
-+ * released.
-+ */
-+int qman_ceetm_sp_release(struct qm_ceetm_sp *sp);
-+
-+ /* ----------------------------------- */
-+ /* CEETM :: logical network interfaces */
-+ /* ----------------------------------- */
-+
-+/**
-+ * qman_ceetm_lni_claim - Claims an unclaimed LNI.
-+ * @lni: the returned LNI object, if successful.
-+ * @dcp_id: specifies the desired Fman block (and thus the relevant CEETM
-+ * instance)
-+ * @lni_idx: is the desired LNI index.
-+ *
-+ * Returns zero for success, or -EINVAL on failure, which will happen if the LNI
-+ * is not available or has already been claimed (and not yet successfully
-+ * released), or lni_dix is out of range.
-+ *
-+ * Note that there may be multiple driver domains (or instances) that need to
-+ * transmit out the same LNI, so this claim is only guaranteeing exclusivity
-+ * within the domain of the driver being called. See qman_ceetm_sp_claim() and
-+ * qman_ceetm_sp_get_lni() for more information.
-+ */
-+int qman_ceetm_lni_claim(struct qm_ceetm_lni **lni,
-+ enum qm_dc_portal dcp_id,
-+ unsigned int lni_idx);
-+
-+/**
-+ * qman_ceetm_lni_releaes - Releases a previously claimed LNI.
-+ * @lni: the lni needs to be released.
-+ *
-+ * This will only succeed if all dependent objects have been released.
-+ * Returns zero for success, or -EBUSY if the dependencies are not released.
-+ */
-+int qman_ceetm_lni_release(struct qm_ceetm_lni *lni);
-+
-+/**
-+ * qman_ceetm_sp_set_lni
-+ * qman_ceetm_sp_get_lni - Set/get the LNI that the sub-portal is currently
-+ * mapped to.
-+ * @sp: the given sub-portal.
-+ * @lni(in "set"function): the LNI object which the sp will be mappaed to.
-+ * @lni_idx(in "get" function): the LNI index which the sp is mapped to.
-+ *
-+ * Returns zero for success, or -EINVAL for the "set" function when this sp-lni
-+ * mapping has been set, or configure mapping command returns error, and
-+ * -EINVAL for "get" function when this sp-lni mapping is not set or the query
-+ * mapping command returns error.
-+ *
-+ * This may be useful in situations where multiple driver domains have access
-+ * to the same sub-portals in order to all be able to transmit out the same
-+ * physical interface (perhaps they're on different IP addresses or VPNs, so
-+ * Fman is splitting Rx traffic and here we need to converge Tx traffic). In
-+ * that case, a control-plane is likely to use qman_ceetm_lni_claim() followed
-+ * by qman_ceetm_sp_set_lni() to configure the sub-portal, and other domains
-+ * are likely to use qman_ceetm_sp_get_lni() followed by qman_ceetm_lni_claim()
-+ * in order to determine the LNI that the control-plane had assigned. This is
-+ * why the "get" returns an index, whereas the "set" takes an (already claimed)
-+ * LNI object.
-+ */
-+int qman_ceetm_sp_set_lni(struct qm_ceetm_sp *sp,
-+ struct qm_ceetm_lni *lni);
-+int qman_ceetm_sp_get_lni(struct qm_ceetm_sp *sp,
-+ unsigned int *lni_idx);
-+
-+/**
-+ * qman_ceetm_lni_enable_shaper
-+ * qman_ceetm_lni_disable_shaper - Enables/disables shaping on the LNI.
-+ * @lni: the given LNI.
-+ * @coupled: indicates whether CR and ER are coupled.
-+ * @oal: the overhead accounting length which is added to the actual length of
-+ * each frame when performing shaper calculations.
-+ *
-+ * When the number of (unused) committed-rate tokens reach the committed-rate
-+ * token limit, 'coupled' indicates whether surplus tokens should be added to
-+ * the excess-rate token count (up to the excess-rate token limit).
-+ * When LNI is claimed, the shaper is disabled by default. The enable function
-+ * will turn on this shaper for this lni.
-+ * Whenever a claimed LNI is first enabled for shaping, its committed and
-+ * excess token rates and limits are zero, so will need to be changed to do
-+ * anything useful. The shaper can subsequently be enabled/disabled without
-+ * resetting the shaping parameters, but the shaping parameters will be reset
-+ * when the LNI is released.
-+ *
-+ * Returns zero for success, or errno for "enable" function in the cases as:
-+ * a) -EINVAL if the shaper is already enabled,
-+ * b) -EIO if the configure shaper command returns error.
-+ * For "disable" function, returns:
-+ * a) -EINVAL if the shaper is has already disabled.
-+ * b) -EIO if calling configure shaper command returns error.
-+ */
-+int qman_ceetm_lni_enable_shaper(struct qm_ceetm_lni *lni, int coupled,
-+ int oal);
-+int qman_ceetm_lni_disable_shaper(struct qm_ceetm_lni *lni);
-+
-+/**
-+ * qman_ceetm_lni_is_shaper_enabled - Check LNI shaper status
-+ * @lni: the give LNI
-+ */
-+int qman_ceetm_lni_is_shaper_enabled(struct qm_ceetm_lni *lni);
-+
-+/**
-+ * qman_ceetm_lni_set_commit_rate
-+ * qman_ceetm_lni_get_commit_rate
-+ * qman_ceetm_lni_set_excess_rate
-+ * qman_ceetm_lni_get_excess_rate - Set/get the shaper CR/ER token rate and
-+ * token limit for the given LNI.
-+ * @lni: the given LNI.
-+ * @token_rate: the desired token rate for "set" fuction, or the token rate of
-+ * the LNI queried by "get" function.
-+ * @token_limit: the desired token bucket limit for "set" function, or the token
-+ * limit of the given LNI queried by "get" function.
-+ *
-+ * Returns zero for success. The "set" function returns -EINVAL if the given
-+ * LNI is unshapped or -EIO if the configure shaper command returns error.
-+ * The "get" function returns -EINVAL if the token rate or the token limit is
-+ * not set or the query command returns error.
-+ */
-+int qman_ceetm_lni_set_commit_rate(struct qm_ceetm_lni *lni,
-+ const struct qm_ceetm_rate *token_rate,
-+ u16 token_limit);
-+int qman_ceetm_lni_get_commit_rate(struct qm_ceetm_lni *lni,
-+ struct qm_ceetm_rate *token_rate,
-+ u16 *token_limit);
-+int qman_ceetm_lni_set_excess_rate(struct qm_ceetm_lni *lni,
-+ const struct qm_ceetm_rate *token_rate,
-+ u16 token_limit);
-+int qman_ceetm_lni_get_excess_rate(struct qm_ceetm_lni *lni,
-+ struct qm_ceetm_rate *token_rate,
-+ u16 *token_limit);
-+/**
-+ * qman_ceetm_lni_set_commit_rate_bps
-+ * qman_ceetm_lni_get_commit_rate_bps
-+ * qman_ceetm_lni_set_excess_rate_bps
-+ * qman_ceetm_lni_get_excess_rate_bps - Set/get the shaper CR/ER rate
-+ * and token limit for the given LNI.
-+ * @lni: the given LNI.
-+ * @bps: the desired shaping rate in bps for "set" fuction, or the shaping rate
-+ * of the LNI queried by "get" function.
-+ * @token_limit: the desired token bucket limit for "set" function, or the token
-+ * limit of the given LNI queried by "get" function.
-+ *
-+ * Returns zero for success. The "set" function returns -EINVAL if the given
-+ * LNI is unshapped or -EIO if the configure shaper command returns error.
-+ * The "get" function returns -EINVAL if the token rate or the token limit is
-+ * not set or the query command returns error.
-+ */
-+int qman_ceetm_lni_set_commit_rate_bps(struct qm_ceetm_lni *lni,
-+ u64 bps,
-+ u16 token_limit);
-+int qman_ceetm_lni_get_commit_rate_bps(struct qm_ceetm_lni *lni,
-+ u64 *bps, u16 *token_limit);
-+int qman_ceetm_lni_set_excess_rate_bps(struct qm_ceetm_lni *lni,
-+ u64 bps,
-+ u16 token_limit);
-+int qman_ceetm_lni_get_excess_rate_bps(struct qm_ceetm_lni *lni,
-+ u64 *bps, u16 *token_limit);
-+
-+/**
-+ * qman_ceetm_lni_set_tcfcc
-+ * qman_ceetm_lni_get_tcfcc - Configure/query "Traffic Class Flow Control".
-+ * @lni: the given LNI.
-+ * @cq_level: is between 0 and 15, representing individual class queue levels
-+ * (CQ0 to CQ7 for every channel) and grouped class queue levels (CQ8 to CQ15
-+ * for every channel).
-+ * @traffic_class: is between 0 and 7 when associating a given class queue level
-+ * to a traffic class, or -1 when disabling traffic class flow control for this
-+ * class queue level.
-+ *
-+ * Return zero for success, or -EINVAL if the cq_level or traffic_class is out
-+ * of range as indicated above, or -EIO if the configure/query tcfcc command
-+ * returns error.
-+ *
-+ * Refer to the section of QMan CEETM traffic class flow control in the
-+ * Reference Manual.
-+ */
-+int qman_ceetm_lni_set_tcfcc(struct qm_ceetm_lni *lni,
-+ unsigned int cq_level,
-+ int traffic_class);
-+int qman_ceetm_lni_get_tcfcc(struct qm_ceetm_lni *lni,
-+ unsigned int cq_level,
-+ int *traffic_class);
-+
-+ /* ----------------------------- */
-+ /* CEETM :: class queue channels */
-+ /* ----------------------------- */
-+
-+/**
-+ * qman_ceetm_channel_claim - Claims an unclaimed CQ channel that is mapped to
-+ * the given LNI.
-+ * @channel: the returned class queue channel object, if successful.
-+ * @lni: the LNI that the channel belongs to.
-+ *
-+ * Channels are always initially "unshaped".
-+ *
-+ * Return zero for success, or -ENODEV if there is no channel available(all 32
-+ * channels are claimed) or -EINVAL if the channel mapping command returns
-+ * error.
-+ */
-+int qman_ceetm_channel_claim(struct qm_ceetm_channel **channel,
-+ struct qm_ceetm_lni *lni);
-+
-+/**
-+ * qman_ceetm_channel_release - Releases a previously claimed CQ channel.
-+ * @channel: the channel needs to be released.
-+ *
-+ * Returns zero for success, or -EBUSY if the dependencies are still in use.
-+ *
-+ * Note any shaping of the channel will be cleared to leave it in an unshaped
-+ * state.
-+ */
-+int qman_ceetm_channel_release(struct qm_ceetm_channel *channel);
-+
-+/**
-+ * qman_ceetm_channel_enable_shaper
-+ * qman_ceetm_channel_disable_shaper - Enables/disables shaping on the channel.
-+ * @channel: the given channel.
-+ * @coupled: indicates whether surplus CR tokens should be added to the
-+ * excess-rate token count (up to the excess-rate token limit) when the number
-+ * of (unused) committed-rate tokens reach the committed_rate token limit.
-+ *
-+ * Whenever a claimed channel is first enabled for shaping, its committed and
-+ * excess token rates and limits are zero, so will need to be changed to do
-+ * anything useful. The shaper can subsequently be enabled/disabled without
-+ * resetting the shaping parameters, but the shaping parameters will be reset
-+ * when the channel is released.
-+ *
-+ * Return 0 for success, or -EINVAL for failure, in the case that the channel
-+ * shaper has been enabled/disabled or the management command returns error.
-+ */
-+int qman_ceetm_channel_enable_shaper(struct qm_ceetm_channel *channel,
-+ int coupled);
-+int qman_ceetm_channel_disable_shaper(struct qm_ceetm_channel *channel);
-+
-+/**
-+ * qman_ceetm_channel_is_shaper_enabled - Check channel shaper status.
-+ * @channel: the give channel.
-+ */
-+int qman_ceetm_channel_is_shaper_enabled(struct qm_ceetm_channel *channel);
-+
-+/**
-+ * qman_ceetm_channel_set_commit_rate
-+ * qman_ceetm_channel_get_commit_rate
-+ * qman_ceetm_channel_set_excess_rate
-+ * qman_ceetm_channel_get_excess_rate - Set/get channel CR/ER shaper parameters.
-+ * @channel: the given channel.
-+ * @token_rate: the desired token rate for "set" function, or the queried token
-+ * rate for "get" function.
-+ * @token_limit: the desired token limit for "set" function, or the queried
-+ * token limit for "get" function.
-+ *
-+ * Return zero for success. The "set" function returns -EINVAL if the channel
-+ * is unshaped, or -EIO if the configure shapper command returns error. The
-+ * "get" function returns -EINVAL if token rate of token limit is not set, or
-+ * the query shaper command returns error.
-+ */
-+int qman_ceetm_channel_set_commit_rate(struct qm_ceetm_channel *channel,
-+ const struct qm_ceetm_rate *token_rate,
-+ u16 token_limit);
-+int qman_ceetm_channel_get_commit_rate(struct qm_ceetm_channel *channel,
-+ struct qm_ceetm_rate *token_rate,
-+ u16 *token_limit);
-+int qman_ceetm_channel_set_excess_rate(struct qm_ceetm_channel *channel,
-+ const struct qm_ceetm_rate *token_rate,
-+ u16 token_limit);
-+int qman_ceetm_channel_get_excess_rate(struct qm_ceetm_channel *channel,
-+ struct qm_ceetm_rate *token_rate,
-+ u16 *token_limit);
-+/**
-+ * qman_ceetm_channel_set_commit_rate_bps
-+ * qman_ceetm_channel_get_commit_rate_bps
-+ * qman_ceetm_channel_set_excess_rate_bps
-+ * qman_ceetm_channel_get_excess_rate_bps - Set/get channel CR/ER shaper
-+ * parameters.
-+ * @channel: the given channel.
-+ * @token_rate: the desired shaper rate in bps for "set" function, or the
-+ * shaper rate in bps for "get" function.
-+ * @token_limit: the desired token limit for "set" function, or the queried
-+ * token limit for "get" function.
-+ *
-+ * Return zero for success. The "set" function returns -EINVAL if the channel
-+ * is unshaped, or -EIO if the configure shapper command returns error. The
-+ * "get" function returns -EINVAL if token rate of token limit is not set, or
-+ * the query shaper command returns error.
-+ */
-+int qman_ceetm_channel_set_commit_rate_bps(struct qm_ceetm_channel *channel,
-+ u64 bps, u16 token_limit);
-+int qman_ceetm_channel_get_commit_rate_bps(struct qm_ceetm_channel *channel,
-+ u64 *bps, u16 *token_limit);
-+int qman_ceetm_channel_set_excess_rate_bps(struct qm_ceetm_channel *channel,
-+ u64 bps, u16 token_limit);
-+int qman_ceetm_channel_get_excess_rate_bps(struct qm_ceetm_channel *channel,
-+ u64 *bps, u16 *token_limit);
-+
-+/**
-+ * qman_ceetm_channel_set_weight
-+ * qman_ceetm_channel_get_weight - Set/get the weight for unshaped channel
-+ * @channel: the given channel.
-+ * @token_limit: the desired token limit as the weight of the unshaped channel
-+ * for "set" function, or the queried token limit for "get" function.
-+ *
-+ * The algorithm of unshaped fair queuing (uFQ) is used for unshaped channel.
-+ * It allows the unshaped channels to be included in the CR time eligible list,
-+ * and thus use the configured CR token limit value as their fair queuing
-+ * weight.
-+ *
-+ * Return zero for success, or -EINVAL if the channel is a shaped channel or
-+ * the management command returns error.
-+ */
-+int qman_ceetm_channel_set_weight(struct qm_ceetm_channel *channel,
-+ u16 token_limit);
-+int qman_ceetm_channel_get_weight(struct qm_ceetm_channel *channel,
-+ u16 *token_limit);
-+
-+/**
-+ * qman_ceetm_channel_set_group
-+ * qman_ceetm_channel_get_group - Set/get the grouping of the class scheduler.
-+ * @channel: the given channel.
-+ * @group_b: indicates whether there is group B in this channel.
-+ * @prio_a: the priority of group A.
-+ * @prio_b: the priority of group B.
-+ *
-+ * There are 8 individual class queues (CQ0-CQ7), and 8 grouped class queues
-+ * (CQ8-CQ15). If 'group_b' is zero, then all the grouped class queues are in
-+ * group A, otherwise they are split into group A (CQ8-11) and group B
-+ * (CQ12-C15). The individual class queues and the group(s) are in strict
-+ * priority order relative to each other. Within the group(s), the scheduling
-+ * is not strict priority order, but the result of scheduling within a group
-+ * is in strict priority order relative to the other class queues in the
-+ * channel. 'prio_a' and 'prio_b' control the priority order of the groups
-+ * relative to the individual class queues, and take values from 0-7. Eg. if
-+ * 'group_b' is non-zero, 'prio_a' is 2 and 'prio_b' is 6, then the strict
-+ * priority order would be;
-+ * CQ0, CQ1, CQ2, GROUPA, CQ3, CQ4, CQ5, CQ6, GROUPB, CQ7
-+ *
-+ * Return 0 for success. For "set" function, returns -EINVAL if prio_a or
-+ * prio_b are out of the range 0 - 7 (priority of group A or group B can not
-+ * be 0, CQ0 is always the highest class queue in this channel.), or -EIO if
-+ * the configure scheduler command returns error. For "get" function, return
-+ * -EINVAL if the query scheduler command returns error.
-+ */
-+int qman_ceetm_channel_set_group(struct qm_ceetm_channel *channel,
-+ int group_b,
-+ unsigned int prio_a,
-+ unsigned int prio_b);
-+int qman_ceetm_channel_get_group(struct qm_ceetm_channel *channel,
-+ int *group_b,
-+ unsigned int *prio_a,
-+ unsigned int *prio_b);
-+
-+/**
-+ * qman_ceetm_channel_set_group_cr_eligibility
-+ * qman_ceetm_channel_set_group_er_eligibility - Set channel group eligibility
-+ * @channel: the given channel object
-+ * @group_b: indicates whether there is group B in this channel.
-+ * @cre: the commit rate eligibility, 1 for enable, 0 for disable.
-+ *
-+ * Return zero for success, or -EINVAL if eligibility setting fails.
-+*/
-+int qman_ceetm_channel_set_group_cr_eligibility(struct qm_ceetm_channel
-+ *channel, int group_b, int cre);
-+int qman_ceetm_channel_set_group_er_eligibility(struct qm_ceetm_channel
-+ *channel, int group_b, int ere);
-+
-+/**
-+ * qman_ceetm_channel_set_cq_cr_eligibility
-+ * qman_ceetm_channel_set_cq_er_eligibility - Set channel cq eligibility
-+ * @channel: the given channel object
-+ * @idx: is from 0 to 7 (representing CQ0 to CQ7).
-+ * @cre: the commit rate eligibility, 1 for enable, 0 for disable.
-+ *
-+ * Return zero for success, or -EINVAL if eligibility setting fails.
-+*/
-+int qman_ceetm_channel_set_cq_cr_eligibility(struct qm_ceetm_channel *channel,
-+ unsigned int idx, int cre);
-+int qman_ceetm_channel_set_cq_er_eligibility(struct qm_ceetm_channel *channel,
-+ unsigned int idx, int ere);
-+
-+ /* --------------------- */
-+ /* CEETM :: class queues */
-+ /* --------------------- */
-+
-+/**
-+ * qman_ceetm_cq_claim - Claims an individual class queue.
-+ * @cq: the returned class queue object, if successful.
-+ * @channel: the class queue channel.
-+ * @idx: is from 0 to 7 (representing CQ0 to CQ7).
-+ * @ccg: represents the class congestion group that this class queue should be
-+ * subscribed to, or NULL if no congestion group membership is desired.
-+ *
-+ * Returns zero for success, or -EINVAL if @idx is out of range 0 - 7 or
-+ * if this class queue has been claimed, or configure class queue command
-+ * returns error, or returns -ENOMEM if allocating CQ memory fails.
-+ */
-+int qman_ceetm_cq_claim(struct qm_ceetm_cq **cq,
-+ struct qm_ceetm_channel *channel,
-+ unsigned int idx,
-+ struct qm_ceetm_ccg *ccg);
-+
-+/**
-+ * qman_ceetm_cq_claim_A - Claims a class queue group A.
-+ * @cq: the returned class queue object, if successful.
-+ * @channel: the class queue channel.
-+ * @idx: is from 8 to 15 if only group A exits, otherwise, it is from 8 to 11.
-+ * @ccg: represents the class congestion group that this class queue should be
-+ * subscribed to, or NULL if no congestion group membership is desired.
-+ *
-+ * Return zero for success, or -EINVAL if @idx is out the range or if
-+ * this class queue has been claimed or configure class queue command returns
-+ * error, or returns -ENOMEM if allocating CQ memory fails.
-+ */
-+int qman_ceetm_cq_claim_A(struct qm_ceetm_cq **cq,
-+ struct qm_ceetm_channel *channel,
-+ unsigned int idx,
-+ struct qm_ceetm_ccg *ccg);
-+
-+/**
-+ * qman_ceetm_cq_claim_B - Claims a class queue group B.
-+ * @cq: the returned class queue object, if successful.
-+ * @channel: the class queue channel.
-+ * @idx: is from 0 to 3 (CQ12 to CQ15).
-+ * @ccg: represents the class congestion group that this class queue should be
-+ * subscribed to, or NULL if no congestion group membership is desired.
-+ *
-+ * Return zero for success, or -EINVAL if @idx is out the range or if
-+ * this class queue has been claimed or configure class queue command returns
-+ * error, or returns -ENOMEM if allocating CQ memory fails.
-+ */
-+int qman_ceetm_cq_claim_B(struct qm_ceetm_cq **cq,
-+ struct qm_ceetm_channel *channel,
-+ unsigned int idx,
-+ struct qm_ceetm_ccg *ccg);
-+
-+/**
-+ * qman_ceetm_cq_release - Releases a previously claimed class queue.
-+ * @cq: The class queue to be released.
-+ *
-+ * Return zero for success, or -EBUSY if the dependent objects (eg. logical
-+ * FQIDs) have not been released.
-+ */
-+int qman_ceetm_cq_release(struct qm_ceetm_cq *cq);
-+
-+/**
-+ * qman_ceetm_set_queue_weight
-+ * qman_ceetm_get_queue_weight - Configure/query the weight of a grouped class
-+ * queue.
-+ * @cq: the given class queue.
-+ * @weight_code: the desired weight code to set for the given class queue for
-+ * "set" function or the queired weight code for "get" function.
-+ *
-+ * Grouped class queues have a default weight code of zero, which corresponds to
-+ * a scheduler weighting of 1. This function can be used to modify a grouped
-+ * class queue to another weight, (Use the helpers qman_ceetm_wbfs2ratio()
-+ * and qman_ceetm_ratio2wbfs() to convert between these 'weight_code' values
-+ * and the corresponding sharing weight.)
-+ *
-+ * Returns zero for success, or -EIO if the configure weight command returns
-+ * error for "set" function, or -EINVAL if the query command returns
-+ * error for "get" function.
-+ * See section "CEETM Weighted Scheduling among Grouped Classes" in Reference
-+ * Manual for weight and weight code.
-+ */
-+int qman_ceetm_set_queue_weight(struct qm_ceetm_cq *cq,
-+ struct qm_ceetm_weight_code *weight_code);
-+int qman_ceetm_get_queue_weight(struct qm_ceetm_cq *cq,
-+ struct qm_ceetm_weight_code *weight_code);
-+
-+/**
-+ * qman_ceetm_set_queue_weight_in_ratio
-+ * qman_ceetm_get_queue_weight_in_ratio - Configure/query the weight of a
-+ * grouped class queue.
-+ * @cq: the given class queue.
-+ * @ratio: the weight in ratio. It should be the real ratio number multiplied
-+ * by 100 to get rid of fraction.
-+ *
-+ * Returns zero for success, or -EIO if the configure weight command returns
-+ * error for "set" function, or -EINVAL if the query command returns
-+ * error for "get" function.
-+ */
-+int qman_ceetm_set_queue_weight_in_ratio(struct qm_ceetm_cq *cq, u32 ratio);
-+int qman_ceetm_get_queue_weight_in_ratio(struct qm_ceetm_cq *cq, u32 *ratio);
-+
-+/* Weights are encoded using a pseudo-exponential scheme. The weight codes 0,
-+ * 32, 64, [...] correspond to weights of 1, 2, 4, [...]. The weights
-+ * corresponding to intermediate weight codes are calculated using linear
-+ * interpolation on the inverted values. Or put another way, the inverse weights
-+ * for each 32nd weight code are 1, 1/2, 1/4, [...], and so the intervals
-+ * between these are divided linearly into 32 intermediate values, the inverses
-+ * of which form the remaining weight codes.
-+ *
-+ * The Weighted Bandwidth Fair Scheduling (WBFS) algorithm provides a form of
-+ * scheduling within a group of class queues (group A or B). Weights are used to
-+ * normalise the class queues to an underlying BFS algorithm where all class
-+ * queues are assumed to require "equal bandwidth". So the weights referred to
-+ * by the weight codes act as divisors on the size of frames being enqueued. Ie.
-+ * one class queue in a group is assigned a weight of 2 whilst the other class
-+ * queues in the group keep the default weight of 1, then the WBFS scheduler
-+ * will effectively treat all frames enqueued on the weight-2 class queue as
-+ * having half the number of bytes they really have. Ie. if all other things are
-+ * equal, that class queue would get twice as much bytes-per-second bandwidth as
-+ * the others. So weights should be chosen to provide bandwidth ratios between
-+ * members of the same class queue group. These weights have no bearing on
-+ * behaviour outside that group's WBFS mechanism though.
-+ */
-+
-+/**
-+ * qman_ceetm_wbfs2ratio - Given a weight code ('wbfs'), an accurate fractional
-+ * representation of the corresponding weight is given (in order to not lose
-+ * any precision).
-+ * @weight_code: The given weight code in WBFS.
-+ * @numerator: the numerator part of the weight computed by the weight code.
-+ * @denominator: the denominator part of the weight computed by the weight code
-+ *
-+ * Returns zero for success or -EINVAL if the given weight code is illegal.
-+ */
-+int qman_ceetm_wbfs2ratio(struct qm_ceetm_weight_code *weight_code,
-+ u32 *numerator,
-+ u32 *denominator);
-+/**
-+ * qman_ceetm_ratio2wbfs - Given a weight, find the nearest possible weight code
-+ * If the user needs to know how close this is, convert the resulting weight
-+ * code back to a weight and compare.
-+ * @numerator: numerator part of the given weight.
-+ * @denominator: denominator part of the given weight.
-+ * @weight_code: the weight code computed from the given weight.
-+ *
-+ * Returns zero for success, or -ERANGE if "numerator/denominator" is outside
-+ * the range of weights.
-+ */
-+int qman_ceetm_ratio2wbfs(u32 numerator,
-+ u32 denominator,
-+ struct qm_ceetm_weight_code *weight_code,
-+ int rounding);
-+
-+#define QMAN_CEETM_FLAG_CLEAR_STATISTICS_COUNTER 0x1
-+/**
-+ * qman_ceetm_cq_get_dequeue_statistics - Get the statistics provided by CEETM
-+ * CQ counters.
-+ * @cq: the given CQ object.
-+ * @flags: indicates whether the statistics counter will be cleared after query.
-+ * @frame_count: The number of the frames that have been counted since the
-+ * counter was cleared last time.
-+ * @byte_count: the number of bytes in all frames that have been counted.
-+ *
-+ * Return zero for success or -EINVAL if query statistics command returns error.
-+ *
-+ */
-+int qman_ceetm_cq_get_dequeue_statistics(struct qm_ceetm_cq *cq, u32 flags,
-+ u64 *frame_count, u64 *byte_count);
-+
-+/**
-+ * qman_ceetm_drain_cq - drain the CQ till it is empty.
-+ * @cq: the give CQ object.
-+ * Return 0 for success or -EINVAL for unsuccessful command to empty CQ.
-+ */
-+int qman_ceetm_drain_cq(struct qm_ceetm_cq *cq);
-+
-+ /* ---------------------- */
-+ /* CEETM :: logical FQIDs */
-+ /* ---------------------- */
-+/**
-+ * qman_ceetm_lfq_claim - Claims an unused logical FQID, associates it with
-+ * the given class queue.
-+ * @lfq: the returned lfq object, if successful.
-+ * @cq: the class queue which needs to claim a LFQID.
-+ *
-+ * Return zero for success, or -ENODEV if no LFQID is available or -ENOMEM if
-+ * allocating memory for lfq fails, or -EINVAL if configuring LFQMT fails.
-+ */
-+int qman_ceetm_lfq_claim(struct qm_ceetm_lfq **lfq,
-+ struct qm_ceetm_cq *cq);
-+
-+/**
-+ * qman_ceetm_lfq_release - Releases a previously claimed logical FQID.
-+ * @lfq: the lfq to be released.
-+ *
-+ * Return zero for success.
-+ */
-+int qman_ceetm_lfq_release(struct qm_ceetm_lfq *lfq);
-+
-+/**
-+ * qman_ceetm_lfq_set_context
-+ * qman_ceetm_lfq_get_context - Set/get the context_a/context_b pair to the
-+ * "dequeue context table" associated with the logical FQID.
-+ * @lfq: the given logical FQ object.
-+ * @context_a: contextA of the dequeue context.
-+ * @context_b: contextB of the dequeue context.
-+ *
-+ * Returns zero for success, or -EINVAL if there is error to set/get the
-+ * context pair.
-+ */
-+int qman_ceetm_lfq_set_context(struct qm_ceetm_lfq *lfq,
-+ u64 context_a,
-+ u32 context_b);
-+int qman_ceetm_lfq_get_context(struct qm_ceetm_lfq *lfq,
-+ u64 *context_a,
-+ u32 *context_b);
-+
-+/**
-+ * qman_ceetm_create_fq - Initialise a FQ object for the LFQ.
-+ * @lfq: the given logic fq.
-+ * @fq: the fq object created for the given logic fq.
-+ *
-+ * The FQ object can be used in qman_enqueue() and qman_enqueue_orp() APIs to
-+ * target a logical FQID (and the class queue it is associated with).
-+ * Note that this FQ object can only be used for enqueues, and
-+ * in the case of qman_enqueue_orp() it can not be used as the 'orp' parameter,
-+ * only as 'fq'. This FQ object can not (and shouldn't) be destroyed, it is only
-+ * valid as long as the underlying 'lfq' remains claimed. It is the user's
-+ * responsibility to ensure that the underlying 'lfq' is not released until any
-+ * enqueues to this FQ object have completed. The only field the user needs to
-+ * fill in is fq->cb.ern, as that enqueue rejection handler is the callback that
-+ * could conceivably be called on this FQ object. This API can be called
-+ * multiple times to create multiple FQ objects referring to the same logical
-+ * FQID, and any enqueue rejections will respect the callback of the object that
-+ * issued the enqueue (and will identify the object via the parameter passed to
-+ * the callback too). There is no 'flags' parameter to this API as there is for
-+ * qman_create_fq() - the created FQ object behaves as though qman_create_fq()
-+ * had been called with the single flag QMAN_FQ_FLAG_NO_MODIFY.
-+ *
-+ * Returns 0 for success.
-+ */
-+int qman_ceetm_create_fq(struct qm_ceetm_lfq *lfq, struct qman_fq *fq);
-+
-+ /* -------------------------------- */
-+ /* CEETM :: class congestion groups */
-+ /* -------------------------------- */
-+
-+/**
-+ * qman_ceetm_ccg_claim - Claims an unused CCG.
-+ * @ccg: the returned CCG object, if successful.
-+ * @channel: the given class queue channel
-+ * @cscn: the callback function of this CCG.
-+ * @cb_ctx: the corresponding context to be used used if state change
-+ * notifications are later enabled for this CCG.
-+ *
-+ * The congestion group is local to the given class queue channel, so only
-+ * class queues within the channel can be associated with that congestion group.
-+ * The association of class queues to congestion groups occurs when the class
-+ * queues are claimed, see qman_ceetm_cq_claim() and related functions.
-+ * Congestion groups are in a "zero" state when initially claimed, and they are
-+ * returned to that state when released.
-+ *
-+ * Return zero for success, or -EINVAL if no CCG in the channel is available.
-+ */
-+int qman_ceetm_ccg_claim(struct qm_ceetm_ccg **ccg,
-+ struct qm_ceetm_channel *channel,
-+ unsigned int idx,
-+ void (*cscn)(struct qm_ceetm_ccg *,
-+ void *cb_ctx,
-+ int congested),
-+ void *cb_ctx);
-+
-+/**
-+ * qman_ceetm_ccg_release - Releases a previously claimed CCG.
-+ * @ccg: the given ccg.
-+ *
-+ * Returns zero for success, or -EBUSY if the given ccg's dependent objects
-+ * (class queues that are associated with the CCG) have not been released.
-+ */
-+int qman_ceetm_ccg_release(struct qm_ceetm_ccg *ccg);
-+
-+/* This struct is used to specify attributes for a CCG. The 'we_mask' field
-+ * controls which CCG attributes are to be updated, and the remainder specify
-+ * the values for those attributes. A CCG counts either frames or the bytes
-+ * within those frames, but not both ('mode'). A CCG can optionally cause
-+ * enqueues to be rejected, due to tail-drop or WRED, or both (they are
-+ * independent options, 'td_en' and 'wr_en_g,wr_en_y,wr_en_r'). Tail-drop can be
-+ * level-triggered due to a single threshold ('td_thres') or edge-triggered due
-+ * to a "congestion state", but not both ('td_mode'). Congestion state has
-+ * distinct entry and exit thresholds ('cs_thres_in' and 'cs_thres_out'), and
-+ * notifications can be sent to software the CCG goes in to and out of this
-+ * congested state ('cscn_en'). */
-+struct qm_ceetm_ccg_params {
-+ /* Boolean fields together in a single bitfield struct */
-+ struct {
-+ /* Whether to count bytes or frames. 1==frames */
-+ u8 mode:1;
-+ /* En/disable tail-drop. 1==enable */
-+ u8 td_en:1;
-+ /* Tail-drop on congestion-state or threshold. 1=threshold */
-+ u8 td_mode:1;
-+ /* Generate congestion state change notifications. 1==enable */
-+ u8 cscn_en:1;
-+ /* Enable WRED rejections (per colour). 1==enable */
-+ u8 wr_en_g:1;
-+ u8 wr_en_y:1;
-+ u8 wr_en_r:1;
-+ } __packed;
-+ /* Tail-drop threshold. See qm_cgr_thres_[gs]et64(). */
-+ struct qm_cgr_cs_thres td_thres;
-+ /* Congestion state thresholds, for entry and exit. */
-+ struct qm_cgr_cs_thres cs_thres_in;
-+ struct qm_cgr_cs_thres cs_thres_out;
-+ /* Overhead accounting length. Per-packet "tax", from -128 to +127 */
-+ signed char oal;
-+ /* Congestion state change notification for DCP portal, virtual CCGID*/
-+ /* WRED parameters. */
-+ struct qm_cgr_wr_parm wr_parm_g;
-+ struct qm_cgr_wr_parm wr_parm_y;
-+ struct qm_cgr_wr_parm wr_parm_r;
-+};
-+/* Bits used in 'we_mask' to qman_ceetm_ccg_set(), controls which attributes of
-+ * the CCGR are to be updated. */
-+#define QM_CCGR_WE_MODE 0x0001 /* mode (bytes/frames) */
-+#define QM_CCGR_WE_CS_THRES_IN 0x0002 /* congestion state entry threshold */
-+#define QM_CCGR_WE_TD_EN 0x0004 /* congestion state tail-drop enable */
-+#define QM_CCGR_WE_CSCN_TUPD 0x0008 /* CSCN target update */
-+#define QM_CCGR_WE_CSCN_EN 0x0010 /* congestion notification enable */
-+#define QM_CCGR_WE_WR_EN_R 0x0020 /* WRED enable - red */
-+#define QM_CCGR_WE_WR_EN_Y 0x0040 /* WRED enable - yellow */
-+#define QM_CCGR_WE_WR_EN_G 0x0080 /* WRED enable - green */
-+#define QM_CCGR_WE_WR_PARM_R 0x0100 /* WRED parameters - red */
-+#define QM_CCGR_WE_WR_PARM_Y 0x0200 /* WRED parameters - yellow */
-+#define QM_CCGR_WE_WR_PARM_G 0x0400 /* WRED parameters - green */
-+#define QM_CCGR_WE_OAL 0x0800 /* overhead accounting length */
-+#define QM_CCGR_WE_CS_THRES_OUT 0x1000 /* congestion state exit threshold */
-+#define QM_CCGR_WE_TD_THRES 0x2000 /* tail-drop threshold */
-+#define QM_CCGR_WE_TD_MODE 0x4000 /* tail-drop mode (state/threshold) */
-+#define QM_CCGR_WE_CDV 0x8000 /* cdv */
-+
-+/**
-+ * qman_ceetm_ccg_set
-+ * qman_ceetm_ccg_get - Configure/query a subset of CCG attributes.
-+ * @ccg: the given CCG object.
-+ * @we_mask: the write enable mask.
-+ * @params: the parameters setting for this ccg
-+ *
-+ * Return 0 for success, or -EIO if configure ccg command returns error for
-+ * "set" function, or -EINVAL if query ccg command returns error for "get"
-+ * function.
-+ */
-+int qman_ceetm_ccg_set(struct qm_ceetm_ccg *ccg,
-+ u16 we_mask,
-+ const struct qm_ceetm_ccg_params *params);
-+int qman_ceetm_ccg_get(struct qm_ceetm_ccg *ccg,
-+ struct qm_ceetm_ccg_params *params);
-+
-+/** qman_ceetm_cscn_swp_set - Add or remove a software portal from the target
-+ * mask.
-+ * qman_ceetm_cscn_swp_get - Query whether a given software portal index is
-+ * in the cscn target mask.
-+ * @ccg: the give CCG object.
-+ * @swp_idx: the index of the software portal.
-+ * @cscn_enabled: 1: Set the swp to be cscn target. 0: remove the swp from
-+ * the target mask.
-+ * @we_mask: the write enable mask.
-+ * @params: the parameters setting for this ccg
-+ *
-+ * Return 0 for success, or -EINVAL if command in set/get function fails.
-+ */
-+int qman_ceetm_cscn_swp_set(struct qm_ceetm_ccg *ccg,
-+ u16 swp_idx,
-+ unsigned int cscn_enabled,
-+ u16 we_mask,
-+ const struct qm_ceetm_ccg_params *params);
-+int qman_ceetm_cscn_swp_get(struct qm_ceetm_ccg *ccg,
-+ u16 swp_idx,
-+ unsigned int *cscn_enabled);
-+
-+/** qman_ceetm_cscn_dcp_set - Add or remove a direct connect portal from the\
-+ * target mask.
-+ * qman_ceetm_cscn_dcp_get - Query whether a given direct connect portal index
-+ * is in the cscn target mask.
-+ * @ccg: the give CCG object.
-+ * @dcp_idx: the index of the direct connect portal.
-+ * @vcgid: congestion state change notification for dcp portal, virtual CGID.
-+ * @cscn_enabled: 1: Set the dcp to be cscn target. 0: remove the dcp from
-+ * the target mask.
-+ * @we_mask: the write enable mask.
-+ * @params: the parameters setting for this ccg
-+ *
-+ * Return 0 for success, or -EINVAL if command in set/get function fails.
-+ */
-+int qman_ceetm_cscn_dcp_set(struct qm_ceetm_ccg *ccg,
-+ u16 dcp_idx,
-+ u8 vcgid,
-+ unsigned int cscn_enabled,
-+ u16 we_mask,
-+ const struct qm_ceetm_ccg_params *params);
-+int qman_ceetm_cscn_dcp_get(struct qm_ceetm_ccg *ccg,
-+ u16 dcp_idx,
-+ u8 *vcgid,
-+ unsigned int *cscn_enabled);
-+
-+/**
-+ * qman_ceetm_ccg_get_reject_statistics - Get the statistics provided by
-+ * CEETM CCG counters.
-+ * @ccg: the given CCG object.
-+ * @flags: indicates whether the statistics counter will be cleared after query.
-+ * @frame_count: The number of the frames that have been counted since the
-+ * counter was cleared last time.
-+ * @byte_count: the number of bytes in all frames that have been counted.
-+ *
-+ * Return zero for success or -EINVAL if query statistics command returns error.
-+ *
-+ */
-+int qman_ceetm_ccg_get_reject_statistics(struct qm_ceetm_ccg *ccg, u32 flags,
-+ u64 *frame_count, u64 *byte_count);
-+
-+/**
-+ * qman_ceetm_query_lfqmt - Query the logical frame queue mapping table
-+ * @lfqid: Logical Frame Queue ID
-+ * @lfqmt_query: Results of the query command
-+ *
-+ * Returns zero for success or -EIO if the query command returns error.
-+ *
-+ */
-+int qman_ceetm_query_lfqmt(int lfqid,
-+ struct qm_mcr_ceetm_lfqmt_query *lfqmt_query);
-+
-+/**
-+ * qman_ceetm_query_cq - Queries a CEETM CQ
-+ * @cqid: the channel ID (first byte) followed by the CQ idx
-+ * @dcpid: CEETM portal ID
-+ * @cq_query: storage for the queried CQ fields
-+ *
-+ * Returns zero for success or -EIO if the query command returns error.
-+ *
-+*/
-+int qman_ceetm_query_cq(unsigned int cqid, unsigned int dcpid,
-+ struct qm_mcr_ceetm_cq_query *cq_query);
-+
-+/**
-+ * qman_ceetm_query_write_statistics - Query (and optionally write) statistics
-+ * @cid: Target ID (CQID or CCGRID)
-+ * @dcp_idx: CEETM portal ID
-+ * @command_type: One of the following:
-+ * 0 = Query dequeue statistics. CID carries the CQID to be queried.
-+ * 1 = Query and clear dequeue statistics. CID carries the CQID to be queried
-+ * 2 = Write dequeue statistics. CID carries the CQID to be written.
-+ * 3 = Query reject statistics. CID carries the CCGRID to be queried.
-+ * 4 = Query and clear reject statistics. CID carries the CCGRID to be queried
-+ * 5 = Write reject statistics. CID carries the CCGRID to be written
-+ * @frame_count: Frame count value to be written if this is a write command
-+ * @byte_count: Bytes count value to be written if this is a write command
-+ *
-+ * Returns zero for success or -EIO if the query command returns error.
-+ */
-+int qman_ceetm_query_write_statistics(u16 cid, enum qm_dc_portal dcp_idx,
-+ u16 command_type, u64 frame_count,
-+ u64 byte_count);
-+
-+/**
-+ * qman_set_wpm - Set waterfall power management
-+ *
-+ * @wpm_enable: boolean, 1 = enable wpm, 0 = disable wpm.
-+ *
-+ * Return 0 for success, return -ENODEV if QMan misc_cfg register is not
-+ * accessible.
-+ */
-+int qman_set_wpm(int wpm_enable);
-+
-+/**
-+ * qman_get_wpm - Query the waterfall power management setting
-+ *
-+ * @wpm_enable: boolean, 1 = enable wpm, 0 = disable wpm.
-+ *
-+ * Return 0 for success, return -ENODEV if QMan misc_cfg register is not
-+ * accessible.
-+ */
-+int qman_get_wpm(int *wpm_enable);
-+
-+/* The below qman_p_***() variants might be called in a migration situation
-+ * (e.g. cpu hotplug). They are used to continue accessing the portal that
-+ * execution was affine to prior to migration.
-+ * @qman_portal specifies which portal the APIs will use.
-+*/
-+const struct qman_portal_config *qman_p_get_portal_config(struct qman_portal
-+ *p);
-+int qman_p_irqsource_add(struct qman_portal *p, u32 bits);
-+int qman_p_irqsource_remove(struct qman_portal *p, u32 bits);
-+int qman_p_poll_dqrr(struct qman_portal *p, unsigned int limit);
-+u32 qman_p_poll_slow(struct qman_portal *p);
-+void qman_p_poll(struct qman_portal *p);
-+void qman_p_stop_dequeues(struct qman_portal *p);
-+void qman_p_start_dequeues(struct qman_portal *p);
-+void qman_p_static_dequeue_add(struct qman_portal *p, u32 pools);
-+void qman_p_static_dequeue_del(struct qman_portal *p, u32 pools);
-+u32 qman_p_static_dequeue_get(struct qman_portal *p);
-+void qman_p_dca(struct qman_portal *p, struct qm_dqrr_entry *dq,
-+ int park_request);
-+int qman_p_volatile_dequeue(struct qman_portal *p, struct qman_fq *fq,
-+ u32 flags __maybe_unused, u32 vdqcr);
-+int qman_p_enqueue(struct qman_portal *p, struct qman_fq *fq,
-+ const struct qm_fd *fd, u32 flags);
-+int qman_p_enqueue_orp(struct qman_portal *p, struct qman_fq *fq,
-+ const struct qm_fd *fd, u32 flags,
-+ struct qman_fq *orp, u16 orp_seqnum);
-+int qman_p_enqueue_precommit(struct qman_portal *p, struct qman_fq *fq,
-+ const struct qm_fd *fd, u32 flags,
-+ qman_cb_precommit cb, void *cb_arg);
-+#ifdef __cplusplus
-+}
-+#endif
-+
-+#endif /* FSL_QMAN_H */
---- /dev/null
-+++ b/include/linux/fsl_usdpaa.h
-@@ -0,0 +1,372 @@
-+/* Copyright 2011-2012 Freescale Semiconductor, Inc.
-+ *
-+ * This file is licensed under the terms of the GNU General Public License
-+ * version 2. This program is licensed "as is" without any warranty of any
-+ * kind, whether express or implied.
-+ */
-+
-+#ifndef FSL_USDPAA_H
-+#define FSL_USDPAA_H
-+
-+#ifdef __cplusplus
-+extern "C" {
-+#endif
-+
-+#include <linux/uaccess.h>
-+#include <linux/ioctl.h>
-+#include <linux/fsl_qman.h> /* For "enum qm_channel" */
-+#include <linux/compat.h>
-+
-+#ifdef CONFIG_FSL_USDPAA
-+
-+/******************************/
-+/* Allocation of resource IDs */
-+/******************************/
-+
-+/* This enum is used to distinguish between the type of underlying object being
-+ * manipulated. */
-+enum usdpaa_id_type {
-+ usdpaa_id_fqid,
-+ usdpaa_id_bpid,
-+ usdpaa_id_qpool,
-+ usdpaa_id_cgrid,
-+ usdpaa_id_ceetm0_lfqid,
-+ usdpaa_id_ceetm0_channelid,
-+ usdpaa_id_ceetm1_lfqid,
-+ usdpaa_id_ceetm1_channelid,
-+ usdpaa_id_max /* <-- not a valid type, represents the number of types */
-+};
-+#define USDPAA_IOCTL_MAGIC 'u'
-+struct usdpaa_ioctl_id_alloc {
-+ uint32_t base; /* Return value, the start of the allocated range */
-+ enum usdpaa_id_type id_type; /* what kind of resource(s) to allocate */
-+ uint32_t num; /* how many IDs to allocate (and return value) */
-+ uint32_t align; /* must be a power of 2, 0 is treated like 1 */
-+ int partial; /* whether to allow less than 'num' */
-+};
-+struct usdpaa_ioctl_id_release {
-+ /* Input; */
-+ enum usdpaa_id_type id_type;
-+ uint32_t base;
-+ uint32_t num;
-+};
-+struct usdpaa_ioctl_id_reserve {
-+ enum usdpaa_id_type id_type;
-+ uint32_t base;
-+ uint32_t num;
-+};
-+
-+
-+/* ioctl() commands */
-+#define USDPAA_IOCTL_ID_ALLOC \
-+ _IOWR(USDPAA_IOCTL_MAGIC, 0x01, struct usdpaa_ioctl_id_alloc)
-+#define USDPAA_IOCTL_ID_RELEASE \
-+ _IOW(USDPAA_IOCTL_MAGIC, 0x02, struct usdpaa_ioctl_id_release)
-+#define USDPAA_IOCTL_ID_RESERVE \
-+ _IOW(USDPAA_IOCTL_MAGIC, 0x0A, struct usdpaa_ioctl_id_reserve)
-+
-+/**********************/
-+/* Mapping DMA memory */
-+/**********************/
-+
-+/* Maximum length for a map name, including NULL-terminator */
-+#define USDPAA_DMA_NAME_MAX 16
-+/* Flags for requesting DMA maps. Maps are private+unnamed or sharable+named.
-+ * For a sharable and named map, specify _SHARED (whether creating one or
-+ * binding to an existing one). If _SHARED is specified and _CREATE is not, then
-+ * the mapping must already exist. If _SHARED and _CREATE are specified and the
-+ * mapping doesn't already exist, it will be created. If _SHARED and _CREATE are
-+ * specified and the mapping already exists, the mapping will fail unless _LAZY
-+ * is specified. When mapping to a pre-existing sharable map, the length must be
-+ * an exact match. Lengths must be a power-of-4 multiple of page size.
-+ *
-+ * Note that this does not actually map the memory to user-space, that is done
-+ * by a subsequent mmap() using the page offset returned from this ioctl(). The
-+ * ioctl() is what gives the process permission to do this, and a page-offset
-+ * with which to do so.
-+ */
-+#define USDPAA_DMA_FLAG_SHARE 0x01
-+#define USDPAA_DMA_FLAG_CREATE 0x02
-+#define USDPAA_DMA_FLAG_LAZY 0x04
-+#define USDPAA_DMA_FLAG_RDONLY 0x08
-+struct usdpaa_ioctl_dma_map {
-+ /* Output parameters - virtual and physical addresses */
-+ void *ptr;
-+ uint64_t phys_addr;
-+ /* Input parameter, the length of the region to be created (or if
-+ * mapping an existing region, this must match it). Must be a power-of-4
-+ * multiple of page size. */
-+ uint64_t len;
-+ /* Input parameter, the USDPAA_DMA_FLAG_* settings. */
-+ uint32_t flags;
-+ /* If _FLAG_SHARE is specified, the name of the region to be created (or
-+ * of the existing mapping to use). */
-+ char name[USDPAA_DMA_NAME_MAX];
-+ /* If this ioctl() creates the mapping, this is an input parameter
-+ * stating whether the region supports locking. If mapping an existing
-+ * region, this is a return value indicating the same thing. */
-+ int has_locking;
-+ /* In the case of a successful map with _CREATE and _LAZY, this return
-+ * value indicates whether we created the mapped region or whether it
-+ * already existed. */
-+ int did_create;
-+};
-+
-+#ifdef CONFIG_COMPAT
-+struct usdpaa_ioctl_dma_map_compat {
-+ /* Output parameters - virtual and physical addresses */
-+ compat_uptr_t ptr;
-+ uint64_t phys_addr;
-+ /* Input parameter, the length of the region to be created (or if
-+ * mapping an existing region, this must match it). Must be a power-of-4
-+ * multiple of page size. */
-+ uint64_t len;
-+ /* Input parameter, the USDPAA_DMA_FLAG_* settings. */
-+ uint32_t flags;
-+ /* If _FLAG_SHARE is specified, the name of the region to be created (or
-+ * of the existing mapping to use). */
-+ char name[USDPAA_DMA_NAME_MAX];
-+ /* If this ioctl() creates the mapping, this is an input parameter
-+ * stating whether the region supports locking. If mapping an existing
-+ * region, this is a return value indicating the same thing. */
-+ int has_locking;
-+ /* In the case of a successful map with _CREATE and _LAZY, this return
-+ * value indicates whether we created the mapped region or whether it
-+ * already existed. */
-+ int did_create;
-+};
-+
-+#define USDPAA_IOCTL_DMA_MAP_COMPAT \
-+ _IOWR(USDPAA_IOCTL_MAGIC, 0x03, struct usdpaa_ioctl_dma_map_compat)
-+#endif
-+
-+
-+#define USDPAA_IOCTL_DMA_MAP \
-+ _IOWR(USDPAA_IOCTL_MAGIC, 0x03, struct usdpaa_ioctl_dma_map)
-+/* munmap() does not remove the DMA map, just the user-space mapping to it.
-+ * This ioctl will do both (though you can munmap() before calling the ioctl
-+ * too). */
-+#define USDPAA_IOCTL_DMA_UNMAP \
-+ _IOW(USDPAA_IOCTL_MAGIC, 0x04, unsigned char)
-+/* We implement a cross-process locking scheme per DMA map. Call this ioctl()
-+ * with a mmap()'d address, and the process will (interruptible) sleep if the
-+ * lock is already held by another process. Process destruction will
-+ * automatically clean up any held locks. */
-+#define USDPAA_IOCTL_DMA_LOCK \
-+ _IOW(USDPAA_IOCTL_MAGIC, 0x05, unsigned char)
-+#define USDPAA_IOCTL_DMA_UNLOCK \
-+ _IOW(USDPAA_IOCTL_MAGIC, 0x06, unsigned char)
-+
-+/***************************************/
-+/* Mapping and using QMan/BMan portals */
-+/***************************************/
-+enum usdpaa_portal_type {
-+ usdpaa_portal_qman,
-+ usdpaa_portal_bman,
-+};
-+
-+#define QBMAN_ANY_PORTAL_IDX 0xffffffff
-+
-+struct usdpaa_ioctl_portal_map {
-+ /* Input parameter, is a qman or bman portal required. */
-+
-+ enum usdpaa_portal_type type;
-+ /* Specifes a specific portal index to map or QBMAN_ANY_PORTAL_IDX
-+ for don't care. The portal index will be populated by the
-+ driver when the ioctl() successfully completes */
-+ uint32_t index;
-+
-+ /* Return value if the map succeeds, this gives the mapped
-+ * cache-inhibited (cinh) and cache-enabled (cena) addresses. */
-+ struct usdpaa_portal_map {
-+ void *cinh;
-+ void *cena;
-+ } addr;
-+ /* Qman-specific return values */
-+ uint16_t channel;
-+ uint32_t pools;
-+};
-+
-+#ifdef CONFIG_COMPAT
-+struct compat_usdpaa_ioctl_portal_map {
-+ /* Input parameter, is a qman or bman portal required. */
-+ enum usdpaa_portal_type type;
-+ /* Specifes a specific portal index to map or QBMAN_ANY_PORTAL_IDX
-+ for don't care. The portal index will be populated by the
-+ driver when the ioctl() successfully completes */
-+ uint32_t index;
-+ /* Return value if the map succeeds, this gives the mapped
-+ * cache-inhibited (cinh) and cache-enabled (cena) addresses. */
-+ struct usdpaa_portal_map_compat {
-+ compat_uptr_t cinh;
-+ compat_uptr_t cena;
-+ } addr;
-+ /* Qman-specific return values */
-+ uint16_t channel;
-+ uint32_t pools;
-+};
-+#define USDPAA_IOCTL_PORTAL_MAP_COMPAT \
-+ _IOWR(USDPAA_IOCTL_MAGIC, 0x07, struct compat_usdpaa_ioctl_portal_map)
-+#define USDPAA_IOCTL_PORTAL_UNMAP_COMPAT \
-+ _IOW(USDPAA_IOCTL_MAGIC, 0x08, struct usdpaa_portal_map_compat)
-+#endif
-+
-+#define USDPAA_IOCTL_PORTAL_MAP \
-+ _IOWR(USDPAA_IOCTL_MAGIC, 0x07, struct usdpaa_ioctl_portal_map)
-+#define USDPAA_IOCTL_PORTAL_UNMAP \
-+ _IOW(USDPAA_IOCTL_MAGIC, 0x08, struct usdpaa_portal_map)
-+
-+struct usdpaa_ioctl_irq_map {
-+ enum usdpaa_portal_type type; /* Type of portal to map */
-+ int fd; /* File descriptor that contains the portal */
-+ void *portal_cinh; /* Cache inhibited area to identify the portal */
-+};
-+
-+#define USDPAA_IOCTL_PORTAL_IRQ_MAP \
-+ _IOW(USDPAA_IOCTL_MAGIC, 0x09, struct usdpaa_ioctl_irq_map)
-+
-+#ifdef CONFIG_COMPAT
-+
-+struct compat_ioctl_irq_map {
-+ enum usdpaa_portal_type type; /* Type of portal to map */
-+ compat_int_t fd; /* File descriptor that contains the portal */
-+ compat_uptr_t portal_cinh; /* Used identify the portal */};
-+
-+#define USDPAA_IOCTL_PORTAL_IRQ_MAP_COMPAT \
-+ _IOW(USDPAA_IOCTL_MAGIC, 0x09, struct compat_ioctl_irq_map)
-+#endif
-+
-+/* ioctl to query the amount of DMA memory used in the system */
-+struct usdpaa_ioctl_dma_used {
-+ uint64_t free_bytes;
-+ uint64_t total_bytes;
-+};
-+#define USDPAA_IOCTL_DMA_USED \
-+ _IOR(USDPAA_IOCTL_MAGIC, 0x0B, struct usdpaa_ioctl_dma_used)
-+
-+/* ioctl to allocate a raw portal */
-+struct usdpaa_ioctl_raw_portal {
-+ /* inputs */
-+ enum usdpaa_portal_type type; /* Type of portal to allocate */
-+
-+ /* set to non zero to turn on stashing */
-+ uint8_t enable_stash;
-+ /* Stashing attributes for the portal */
-+ uint32_t cpu;
-+ uint32_t cache;
-+ uint32_t window;
-+
-+ /* Specifies the stash request queue this portal should use */
-+ uint8_t sdest;
-+
-+ /* Specifes a specific portal index to map or QBMAN_ANY_PORTAL_IDX
-+ * for don't care. The portal index will be populated by the
-+ * driver when the ioctl() successfully completes */
-+ uint32_t index;
-+
-+ /* outputs */
-+ uint64_t cinh;
-+ uint64_t cena;
-+};
-+
-+#define USDPAA_IOCTL_ALLOC_RAW_PORTAL \
-+ _IOWR(USDPAA_IOCTL_MAGIC, 0x0C, struct usdpaa_ioctl_raw_portal)
-+
-+#define USDPAA_IOCTL_FREE_RAW_PORTAL \
-+ _IOR(USDPAA_IOCTL_MAGIC, 0x0D, struct usdpaa_ioctl_raw_portal)
-+
-+#ifdef CONFIG_COMPAT
-+
-+struct compat_ioctl_raw_portal {
-+ /* inputs */
-+ enum usdpaa_portal_type type; /* Type of portal to allocate */
-+
-+ /* set to non zero to turn on stashing */
-+ uint8_t enable_stash;
-+ /* Stashing attributes for the portal */
-+ uint32_t cpu;
-+ uint32_t cache;
-+ uint32_t window;
-+ /* Specifies the stash request queue this portal should use */
-+ uint8_t sdest;
-+
-+ /* Specifes a specific portal index to map or QBMAN_ANY_PORTAL_IDX
-+ * for don't care. The portal index will be populated by the
-+ * driver when the ioctl() successfully completes */
-+ uint32_t index;
-+
-+ /* outputs */
-+ uint64_t cinh;
-+ uint64_t cena;
-+};
-+
-+#define USDPAA_IOCTL_ALLOC_RAW_PORTAL_COMPAT \
-+ _IOWR(USDPAA_IOCTL_MAGIC, 0x0C, struct compat_ioctl_raw_portal)
-+
-+#define USDPAA_IOCTL_FREE_RAW_PORTAL_COMPAT \
-+ _IOR(USDPAA_IOCTL_MAGIC, 0x0D, struct compat_ioctl_raw_portal)
-+
-+#endif
-+
-+#ifdef __KERNEL__
-+
-+/* Early-boot hook */
-+int __init fsl_usdpaa_init_early(void);
-+
-+/* Fault-handling in arch/powerpc/mm/mem.c gives USDPAA an opportunity to detect
-+ * faults within its ranges via this hook. */
-+int usdpaa_test_fault(unsigned long pfn, u64 *phys_addr, u64 *size);
-+
-+#endif /* __KERNEL__ */
-+
-+#endif /* CONFIG_FSL_USDPAA */
-+
-+#ifdef __KERNEL__
-+/* This interface is needed in a few places and though it's not specific to
-+ * USDPAA as such, creating a new header for it doesn't make any sense. The
-+ * qbman kernel driver implements this interface and uses it as the backend for
-+ * both the FQID and BPID allocators. The fsl_usdpaa driver also uses this
-+ * interface for tracking per-process allocations handed out to user-space. */
-+struct dpa_alloc {
-+ struct list_head free;
-+ spinlock_t lock;
-+ struct list_head used;
-+};
-+#define DECLARE_DPA_ALLOC(name) \
-+ struct dpa_alloc name = { \
-+ .free = { \
-+ .prev = &name.free, \
-+ .next = &name.free \
-+ }, \
-+ .lock = __SPIN_LOCK_UNLOCKED(name.lock), \
-+ .used = { \
-+ .prev = &name.used, \
-+ .next = &name.used \
-+ } \
-+ }
-+static inline void dpa_alloc_init(struct dpa_alloc *alloc)
-+{
-+ INIT_LIST_HEAD(&alloc->free);
-+ INIT_LIST_HEAD(&alloc->used);
-+ spin_lock_init(&alloc->lock);
-+}
-+int dpa_alloc_new(struct dpa_alloc *alloc, u32 *result, u32 count, u32 align,
-+ int partial);
-+void dpa_alloc_free(struct dpa_alloc *alloc, u32 base_id, u32 count);
-+void dpa_alloc_seed(struct dpa_alloc *alloc, u32 fqid, u32 count);
-+
-+/* Like 'new' but specifies the desired range, returns -ENOMEM if the entire
-+ * desired range is not available, or 0 for success. */
-+int dpa_alloc_reserve(struct dpa_alloc *alloc, u32 base_id, u32 count);
-+/* Pops and returns contiguous ranges from the allocator. Returns -ENOMEM when
-+ * 'alloc' is empty. */
-+int dpa_alloc_pop(struct dpa_alloc *alloc, u32 *result, u32 *count);
-+/* Returns 1 if the specified id is alloced, 0 otherwise */
-+int dpa_alloc_check(struct dpa_alloc *list, u32 id);
-+#endif /* __KERNEL__ */
-+
-+#ifdef __cplusplus
-+}
-+#endif
-+
-+#endif /* FSL_USDPAA_H */
---- /dev/null
-+++ b/include/uapi/linux/fmd/Kbuild
-@@ -0,0 +1,5 @@
-+header-y += integrations/
-+header-y += Peripherals/
-+
-+header-y += ioctls.h
-+header-y += net_ioctls.h
---- /dev/null
-+++ b/include/uapi/linux/fmd/Peripherals/Kbuild
-@@ -0,0 +1,4 @@
-+header-y += fm_ioctls.h
-+header-y += fm_port_ioctls.h
-+header-y += fm_pcd_ioctls.h
-+header-y += fm_test_ioctls.h
---- /dev/null
-+++ b/include/uapi/linux/fmd/Peripherals/fm_ioctls.h
-@@ -0,0 +1,628 @@
-+/* Copyright (c) 2008-2012 Freescale Semiconductor, Inc.
-+ * All rights reserved.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+/**************************************************************************//**
-+ @File fm_ioctls.h
-+
-+ @Description FM Char device ioctls
-+*//***************************************************************************/
-+#ifndef __FM_IOCTLS_H
-+#define __FM_IOCTLS_H
-+
-+
-+/**************************************************************************//**
-+ @Group lnx_ioctl_FM_grp Frame Manager Linux IOCTL API
-+
-+ @Description FM Linux ioctls definitions and enums
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Collection FM IOCTL device ('/dev') definitions
-+*//***************************************************************************/
-+#define DEV_FM_NAME "fm" /**< Name of the FM chardev */
-+
-+#define DEV_FM_MINOR_BASE 0
-+#define DEV_FM_PCD_MINOR_BASE (DEV_FM_MINOR_BASE + 1) /*/dev/fmx-pcd */
-+#define DEV_FM_OH_PORTS_MINOR_BASE (DEV_FM_PCD_MINOR_BASE + 1) /*/dev/fmx-port-ohy */
-+#define DEV_FM_RX_PORTS_MINOR_BASE (DEV_FM_OH_PORTS_MINOR_BASE + FM_MAX_NUM_OF_OH_PORTS) /*/dev/fmx-port-rxy */
-+#define DEV_FM_TX_PORTS_MINOR_BASE (DEV_FM_RX_PORTS_MINOR_BASE + FM_MAX_NUM_OF_RX_PORTS) /*/dev/fmx-port-txy */
-+#define DEV_FM_MAX_MINORS (DEV_FM_TX_PORTS_MINOR_BASE + FM_MAX_NUM_OF_TX_PORTS)
-+
-+#define FM_IOC_NUM(n) (n)
-+#define FM_PCD_IOC_NUM(n) (n+20)
-+#define FM_PORT_IOC_NUM(n) (n+70)
-+/* @} */
-+
-+#define IOC_FM_MAX_NUM_OF_PORTS 64
-+
-+
-+/**************************************************************************//**
-+ @Description Enum for defining port types
-+ (must match enum e_FmPortType defined in fm_ext.h)
-+*//***************************************************************************/
-+typedef enum ioc_fm_port_type {
-+ e_IOC_FM_PORT_TYPE_OH_OFFLINE_PARSING = 0, /**< Offline parsing port */
-+ e_IOC_FM_PORT_TYPE_RX, /**< 1G Rx port */
-+ e_IOC_FM_PORT_TYPE_RX_10G, /**< 10G Rx port */
-+ e_IOC_FM_PORT_TYPE_TX, /**< 1G Tx port */
-+ e_IOC_FM_PORT_TYPE_TX_10G, /**< 10G Tx port */
-+ e_IOC_FM_PORT_TYPE_DUMMY
-+} ioc_fm_port_type;
-+
-+
-+/**************************************************************************//**
-+ @Group lnx_ioctl_FM_lib_grp FM library
-+
-+ @Description FM API functions, definitions and enums
-+ The FM module is the main driver module and is a mandatory module
-+ for FM driver users. Before any further module initialization,
-+ this module must be initialized.
-+ The FM is a "single-tone" module. It is responsible of the common
-+ HW modules: FPM, DMA, common QMI, common BMI initializations and
-+ run-time control routines. This module must be initialized always
-+ when working with any of the FM modules.
-+ NOTE - We assumes that the FML will be initialize only by core No. 0!
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Description FM Exceptions
-+*//***************************************************************************/
-+typedef enum ioc_fm_exceptions {
-+ e_IOC_FM_EX_DMA_BUS_ERROR, /**< DMA bus error. */
-+ e_IOC_EX_DMA_READ_ECC, /**< Read Buffer ECC error (Valid for FM rev < 6)*/
-+ e_IOC_EX_DMA_SYSTEM_WRITE_ECC, /**< Write Buffer ECC error on system side (Valid for FM rev < 6)*/
-+ e_IOC_EX_DMA_FM_WRITE_ECC, /**< Write Buffer ECC error on FM side (Valid for FM rev < 6)*/
-+ e_IOC_EX_DMA_SINGLE_PORT_ECC, /**< Single Port ECC error on FM side (Valid for FM rev > 6)*/
-+ e_IOC_EX_FPM_STALL_ON_TASKS, /**< Stall of tasks on FPM */
-+ e_IOC_EX_FPM_SINGLE_ECC, /**< Single ECC on FPM. */
-+ e_IOC_EX_FPM_DOUBLE_ECC, /**< Double ECC error on FPM ram access */
-+ e_IOC_EX_QMI_SINGLE_ECC, /**< Single ECC on QMI. */
-+ e_IOC_EX_QMI_DOUBLE_ECC, /**< Double bit ECC occurred on QMI */
-+ e_IOC_EX_QMI_DEQ_FROM_UNKNOWN_PORTID,/**< Dequeue from unknown port id */
-+ e_IOC_EX_BMI_LIST_RAM_ECC, /**< Linked List RAM ECC error */
-+ e_IOC_EX_BMI_STORAGE_PROFILE_ECC, /**< Storage Profile ECC Error */
-+ e_IOC_EX_BMI_STATISTICS_RAM_ECC, /**< Statistics Count RAM ECC Error Enable */
-+ e_IOC_EX_BMI_DISPATCH_RAM_ECC, /**< Dispatch RAM ECC Error Enable */
-+ e_IOC_EX_IRAM_ECC, /**< Double bit ECC occurred on IRAM*/
-+ e_IOC_EX_MURAM_ECC /**< Double bit ECC occurred on MURAM*/
-+} ioc_fm_exceptions;
-+
-+/**************************************************************************//**
-+ @Group lnx_ioctl_FM_runtime_control_grp FM Runtime Control Unit
-+
-+ @Description FM Runtime control unit API functions, definitions and enums.
-+ The FM driver provides a set of control routines for each module.
-+ These routines may only be called after the module was fully
-+ initialized (both configuration and initialization routines were
-+ called). They are typically used to get information from hardware
-+ (status, counters/statistics, revision etc.), to modify a current
-+ state or to force/enable a required action. Run-time control may
-+ be called whenever necessary and as many times as needed.
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Collection General FM defines.
-+ *//***************************************************************************/
-+#define IOC_FM_MAX_NUM_OF_VALID_PORTS (FM_MAX_NUM_OF_OH_PORTS + \
-+ FM_MAX_NUM_OF_1G_RX_PORTS + \
-+ FM_MAX_NUM_OF_10G_RX_PORTS + \
-+ FM_MAX_NUM_OF_1G_TX_PORTS + \
-+ FM_MAX_NUM_OF_10G_TX_PORTS)
-+/* @} */
-+
-+/**************************************************************************//**
-+ @Description Structure for Port bandwidth requirement. Port is identified
-+ by type and relative id.
-+ (must be identical to t_FmPortBandwidth defined in fm_ext.h)
-+*//***************************************************************************/
-+typedef struct ioc_fm_port_bandwidth_t {
-+ ioc_fm_port_type type; /**< FM port type */
-+ uint8_t relative_port_id; /**< Type relative port id */
-+ uint8_t bandwidth; /**< bandwidth - (in term of percents) */
-+} ioc_fm_port_bandwidth_t;
-+
-+/**************************************************************************//**
-+ @Description A Structure containing an array of Port bandwidth requirements.
-+ The user should state the ports requiring bandwidth in terms of
-+ percentage - i.e. all port's bandwidths in the array must add
-+ up to 100.
-+ (must be identical to t_FmPortsBandwidthParams defined in fm_ext.h)
-+*//***************************************************************************/
-+typedef struct ioc_fm_port_bandwidth_params {
-+ uint8_t num_of_ports;
-+ /**< num of ports listed in the array below */
-+ ioc_fm_port_bandwidth_t ports_bandwidths[IOC_FM_MAX_NUM_OF_VALID_PORTS];
-+ /**< for each port, it's bandwidth (all port's
-+ bandwidths must add up to 100.*/
-+} ioc_fm_port_bandwidth_params;
-+
-+/**************************************************************************//**
-+ @Description enum for defining FM counters
-+*//***************************************************************************/
-+typedef enum ioc_fm_counters {
-+ e_IOC_FM_COUNTERS_ENQ_TOTAL_FRAME, /**< QMI total enqueued frames counter */
-+ e_IOC_FM_COUNTERS_DEQ_TOTAL_FRAME, /**< QMI total dequeued frames counter */
-+ e_IOC_FM_COUNTERS_DEQ_0, /**< QMI 0 frames from QMan counter */
-+ e_IOC_FM_COUNTERS_DEQ_1, /**< QMI 1 frames from QMan counter */
-+ e_IOC_FM_COUNTERS_DEQ_2, /**< QMI 2 frames from QMan counter */
-+ e_IOC_FM_COUNTERS_DEQ_3, /**< QMI 3 frames from QMan counter */
-+ e_IOC_FM_COUNTERS_DEQ_FROM_DEFAULT, /**< QMI dequeue from default queue counter */
-+ e_IOC_FM_COUNTERS_DEQ_FROM_CONTEXT, /**< QMI dequeue from FQ context counter */
-+ e_IOC_FM_COUNTERS_DEQ_FROM_FD, /**< QMI dequeue from FD command field counter */
-+ e_IOC_FM_COUNTERS_DEQ_CONFIRM, /**< QMI dequeue confirm counter */
-+} ioc_fm_counters;
-+
-+typedef struct ioc_fm_obj_t {
-+ void *obj;
-+} ioc_fm_obj_t;
-+
-+/**************************************************************************//**
-+ @Description A structure for returning revision information
-+ (must match struct t_FmRevisionInfo declared in fm_ext.h)
-+*//***************************************************************************/
-+typedef struct ioc_fm_revision_info_t {
-+ uint8_t major; /**< Major revision */
-+ uint8_t minor; /**< Minor revision */
-+} ioc_fm_revision_info_t;
-+
-+/**************************************************************************//**
-+ @Description A structure for FM counters
-+*//***************************************************************************/
-+typedef struct ioc_fm_counters_params_t {
-+ ioc_fm_counters cnt; /**< The requested counter */
-+ uint32_t val; /**< The requested value to get/set from/into the counter */
-+} ioc_fm_counters_params_t;
-+
-+typedef union ioc_fm_api_version_t {
-+ struct {
-+ uint8_t major;
-+ uint8_t minor;
-+ uint8_t respin;
-+ uint8_t reserved;
-+ } version;
-+ uint32_t ver;
-+} ioc_fm_api_version_t;
-+
-+#if (DPAA_VERSION >= 11)
-+/**************************************************************************//**
-+ @Description A structure of information about each of the external
-+ buffer pools used by a port or storage-profile.
-+ (must be identical to t_FmExtPoolParams defined in fm_ext.h)
-+*//***************************************************************************/
-+typedef struct ioc_fm_ext_pool_params {
-+ uint8_t id; /**< External buffer pool id */
-+ uint16_t size; /**< External buffer pool buffer size */
-+} ioc_fm_ext_pool_params;
-+
-+/**************************************************************************//**
-+ @Description A structure for informing the driver about the external
-+ buffer pools allocated in the BM and used by a port or a
-+ storage-profile.
-+ (must be identical to t_FmExtPools defined in fm_ext.h)
-+*//***************************************************************************/
-+typedef struct ioc_fm_ext_pools {
-+ uint8_t num_of_pools_used; /**< Number of pools use by this port */
-+ ioc_fm_ext_pool_params ext_buf_pool[FM_PORT_MAX_NUM_OF_EXT_POOLS];
-+ /**< Parameters for each port */
-+} ioc_fm_ext_pools;
-+
-+typedef struct ioc_fm_vsp_params_t {
-+ void *p_fm; /**< A handle to the FM object this VSP related to */
-+ ioc_fm_ext_pools ext_buf_pools; /**< Which external buffer pools are used
-+ (up to FM_PORT_MAX_NUM_OF_EXT_POOLS), and their sizes.
-+ parameter associated with Rx / OP port */
-+ uint16_t liodn_offset; /**< VSP's LIODN offset */
-+ struct {
-+ ioc_fm_port_type port_type; /**< Port type */
-+ uint8_t port_id; /**< Port Id - relative to type */
-+ } port_params;
-+ uint8_t relative_profile_id; /**< VSP Id - relative to VSP's range
-+ defined in relevant FM object */
-+ void *id; /**< return value */
-+} ioc_fm_vsp_params_t;
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+/**************************************************************************//**
-+ @Description A structure for defining BM pool depletion criteria
-+*//***************************************************************************/
-+typedef struct ioc_fm_buf_pool_depletion_t {
-+ bool pools_grp_mode_enable; /**< select mode in which pause frames will be sent after
-+ a number of pools (all together!) are depleted */
-+ uint8_t num_of_pools; /**< the number of depleted pools that will invoke
-+ pause frames transmission. */
-+ bool pools_to_consider[BM_MAX_NUM_OF_POOLS];
-+ /**< For each pool, TRUE if it should be considered for
-+ depletion (Note - this pool must be used by this port!). */
-+ bool single_pool_mode_enable; /**< select mode in which pause frames will be sent after
-+ a single-pool is depleted; */
-+ bool pools_to_consider_for_single_mode[BM_MAX_NUM_OF_POOLS];
-+ /**< For each pool, TRUE if it should be considered for
-+ depletion (Note - this pool must be used by this port!) */
-+#if (DPAA_VERSION >= 11)
-+ bool pfc_priorities_en[FM_MAX_NUM_OF_PFC_PRIORITIES];
-+ /**< This field is used by the MAC as the Priority Enable Vector in the PFC frame
-+ which is transmitted */
-+#endif /* (DPAA_VERSION >= 11) */
-+} ioc_fm_buf_pool_depletion_t;
-+
-+#if (DPAA_VERSION >= 11)
-+typedef struct ioc_fm_buf_pool_depletion_params_t {
-+ void *p_fm_vsp;
-+ ioc_fm_buf_pool_depletion_t fm_buf_pool_depletion;
-+} ioc_fm_buf_pool_depletion_params_t;
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+typedef struct ioc_fm_buffer_prefix_content_t {
-+ uint16_t priv_data_size; /**< Number of bytes to be left at the beginning
-+ of the external buffer; Note that the private-area will
-+ start from the base of the buffer address. */
-+ bool pass_prs_result; /**< TRUE to pass the parse result to/from the FM;
-+ User may use FM_PORT_GetBufferPrsResult() in order to
-+ get the parser-result from a buffer. */
-+ bool pass_time_stamp; /**< TRUE to pass the timeStamp to/from the FM
-+ User may use FM_PORT_GetBufferTimeStamp() in order to
-+ get the parser-result from a buffer. */
-+ bool pass_hash_result; /**< TRUE to pass the KG hash result to/from the FM
-+ User may use FM_PORT_GetBufferHashResult() in order to
-+ get the parser-result from a buffer. */
-+ bool pass_all_other_pcd_info; /**< Add all other Internal-Context information:
-+ AD, hash-result, key, etc. */
-+ uint16_t data_align; /**< 0 to use driver's default alignment [64],
-+ other value for selecting a data alignment (must be a power of 2);
-+ if write optimization is used, must be >= 16. */
-+ uint8_t manip_extra_space; /**< Maximum extra size needed (insertion-size minus removal-size);
-+ Note that this field impacts the size of the buffer-prefix
-+ (i.e. it pushes the data offset);
-+ This field is irrelevant if DPAA_VERSION==10 */
-+} ioc_fm_buffer_prefix_content_t;
-+
-+typedef struct ioc_fm_buffer_prefix_content_params_t {
-+ void *p_fm_vsp;
-+ ioc_fm_buffer_prefix_content_t fm_buffer_prefix_content;
-+} ioc_fm_buffer_prefix_content_params_t;
-+
-+#if (DPAA_VERSION >= 11)
-+typedef struct ioc_fm_vsp_config_no_sg_params_t {
-+ void *p_fm_vsp;
-+ bool no_sg;
-+} ioc_fm_vsp_config_no_sg_params_t;
-+
-+typedef struct ioc_fm_vsp_prs_result_params_t {
-+ void *p_fm_vsp;
-+ void *p_data;
-+} ioc_fm_vsp_prs_result_params_t;
-+#endif
-+
-+typedef struct fm_ctrl_mon_t {
-+ uint8_t percent_cnt[2];
-+} fm_ctrl_mon_t;
-+
-+typedef struct ioc_fm_ctrl_mon_counters_params_t {
-+ uint8_t fm_ctrl_index;
-+ fm_ctrl_mon_t *p_mon;
-+} ioc_fm_ctrl_mon_counters_params_t;
-+
-+/**************************************************************************//**
-+ @Function FM_IOC_SET_PORTS_BANDWIDTH
-+
-+ @Description Sets relative weights between ports when accessing common resources.
-+
-+ @Param[in] ioc_fm_port_bandwidth_params Port bandwidth percentages,
-+ their sum must equal 100.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_Init().
-+*//***************************************************************************/
-+#define FM_IOC_SET_PORTS_BANDWIDTH _IOW(FM_IOC_TYPE_BASE, FM_IOC_NUM(2), ioc_fm_port_bandwidth_params)
-+
-+/**************************************************************************//**
-+ @Function FM_IOC_GET_REVISION
-+
-+ @Description Returns the FM revision
-+
-+ @Param[out] ioc_fm_revision_info_t A structure of revision information parameters.
-+
-+ @Return None.
-+
-+ @Cautions Allowed only following FM_Init().
-+*//***************************************************************************/
-+#define FM_IOC_GET_REVISION _IOR(FM_IOC_TYPE_BASE, FM_IOC_NUM(3), ioc_fm_revision_info_t)
-+
-+/**************************************************************************//**
-+ @Function FM_IOC_GET_COUNTER
-+
-+ @Description Reads one of the FM counters.
-+
-+ @Param[in,out] ioc_fm_counters_params_t The requested counter parameters.
-+
-+ @Return Counter's current value.
-+
-+ @Cautions Allowed only following FM_Init().
-+ Note that it is user's responsibilty to call this routine only
-+ for enabled counters, and there will be no indication if a
-+ disabled counter is accessed.
-+*//***************************************************************************/
-+#define FM_IOC_GET_COUNTER _IOWR(FM_IOC_TYPE_BASE, FM_IOC_NUM(4), ioc_fm_counters_params_t)
-+
-+/**************************************************************************//**
-+ @Function FM_IOC_SET_COUNTER
-+
-+ @Description Sets a value to an enabled counter. Use "0" to reset the counter.
-+
-+ @Param[in] ioc_fm_counters_params_t The requested counter parameters.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_Init().
-+*//***************************************************************************/
-+#define FM_IOC_SET_COUNTER _IOW(FM_IOC_TYPE_BASE, FM_IOC_NUM(5), ioc_fm_counters_params_t)
-+
-+/**************************************************************************//**
-+ @Function FM_IOC_FORCE_INTR
-+
-+ @Description Causes an interrupt event on the requested source.
-+
-+ @Param[in] ioc_fm_exceptions An exception to be forced.
-+
-+ @Return E_OK on success; Error code if the exception is not enabled,
-+ or is not able to create interrupt.
-+
-+ @Cautions Allowed only following FM_Init().
-+*//***************************************************************************/
-+#define FM_IOC_FORCE_INTR _IOW(FM_IOC_TYPE_BASE, FM_IOC_NUM(6), ioc_fm_exceptions)
-+
-+/**************************************************************************//**
-+ @Function FM_IOC_GET_API_VERSION
-+
-+ @Description Reads the FMD IOCTL API version.
-+
-+ @Param[in,out] ioc_fm_api_version_t The requested counter parameters.
-+
-+ @Return Version's value.
-+*//***************************************************************************/
-+#define FM_IOC_GET_API_VERSION _IOR(FM_IOC_TYPE_BASE, FM_IOC_NUM(7), ioc_fm_api_version_t)
-+
-+#if (DPAA_VERSION >= 11)
-+/**************************************************************************//**
-+ @Function FM_VSP_Config
-+
-+ @Description Creates descriptor for the FM VSP module.
-+
-+ The routine returns a handle (descriptor) to the FM VSP object.
-+ This descriptor must be passed as first parameter to all other
-+ FM VSP function calls.
-+
-+ No actual initialization or configuration of FM hardware is
-+ done by this routine.
-+
-+@Param[in] p_FmVspParams Pointer to data structure of parameters
-+
-+ @Retval Handle to FM VSP object, or NULL for Failure.
-+*//***************************************************************************/
-+#if defined(CONFIG_COMPAT)
-+#define FM_IOC_VSP_CONFIG_COMPAT _IOWR(FM_IOC_TYPE_BASE, FM_IOC_NUM(8), ioc_compat_fm_vsp_params_t)
-+#endif
-+#define FM_IOC_VSP_CONFIG _IOWR(FM_IOC_TYPE_BASE, FM_IOC_NUM(8), ioc_fm_vsp_params_t)
-+
-+/**************************************************************************//**
-+ @Function FM_VSP_Init
-+
-+ @Description Initializes the FM VSP module
-+
-+ @Param[in] h_FmVsp - FM VSP module descriptor
-+
-+ @Return E_OK on success; Error code otherwise.
-+*//***************************************************************************/
-+#if defined(CONFIG_COMPAT)
-+#define FM_IOC_VSP_INIT_COMPAT _IOW(FM_IOC_TYPE_BASE, FM_IOC_NUM(9), ioc_compat_fm_obj_t)
-+#endif
-+#define FM_IOC_VSP_INIT _IOW(FM_IOC_TYPE_BASE, FM_IOC_NUM(9), ioc_fm_obj_t)
-+
-+/**************************************************************************//**
-+ @Function FM_VSP_Free
-+
-+ @Description Frees all resources that were assigned to FM VSP module.
-+
-+ Calling this routine invalidates the descriptor.
-+
-+ @Param[in] h_FmVsp - FM VSP module descriptor
-+
-+ @Return E_OK on success; Error code otherwise.
-+*//***************************************************************************/
-+#if defined(CONFIG_COMPAT)
-+#define FM_IOC_VSP_FREE_COMPAT _IOW(FM_IOC_TYPE_BASE, FM_IOC_NUM(10), ioc_compat_fm_obj_t)
-+#endif
-+#define FM_IOC_VSP_FREE _IOW(FM_IOC_TYPE_BASE, FM_IOC_NUM(10), ioc_fm_obj_t)
-+
-+/**************************************************************************//**
-+ @Function FM_VSP_ConfigPoolDepletion
-+
-+ @Description Calling this routine enables pause frame generation depending on the
-+ depletion status of BM pools. It also defines the conditions to activate
-+ this functionality. By default, this functionality is disabled.
-+
-+ @Param[in] ioc_fm_buf_pool_depletion_params_t A structure holding the required parameters.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_VSP_Config() and before FM_VSP_Init().
-+*//***************************************************************************/
-+#if defined(CONFIG_COMPAT)
-+#define FM_IOC_VSP_CONFIG_POOL_DEPLETION_COMPAT _IOW(FM_IOC_TYPE_BASE, FM_IOC_NUM(11), ioc_compat_fm_buf_pool_depletion_params_t)
-+#endif
-+#define FM_IOC_VSP_CONFIG_POOL_DEPLETION _IOW(FM_IOC_TYPE_BASE, FM_IOC_NUM(11), ioc_fm_buf_pool_depletion_params_t)
-+
-+/**************************************************************************//**
-+ @Function FM_VSP_ConfigBufferPrefixContent
-+
-+ @Description Defines the structure, size and content of the application buffer.
-+
-+ The prefix will
-+ In VSPs defined for Tx ports, if 'passPrsResult', the application
-+ should set a value to their offsets in the prefix of
-+ the FM will save the first 'privDataSize', than,
-+ depending on 'passPrsResult' and 'passTimeStamp', copy parse result
-+ and timeStamp, and the packet itself (in this order), to the
-+ application buffer, and to offset.
-+
-+ Calling this routine changes the buffer margins definitions
-+ in the internal driver data base from its default
-+ configuration: Data size: [DEFAULT_FM_SP_bufferPrefixContent_privDataSize]
-+ Pass Parser result: [DEFAULT_FM_SP_bufferPrefixContent_passPrsResult].
-+ Pass timestamp: [DEFAULT_FM_SP_bufferPrefixContent_passTimeStamp].
-+
-+ @Param[in] ioc_fm_buffer_prefix_content_params_t A structure holding the required parameters.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_VSP_Config() and before FM_VSP_Init().
-+*//***************************************************************************/
-+#if defined(CONFIG_COMPAT)
-+#define FM_IOC_VSP_CONFIG_BUFFER_PREFIX_CONTENT_COMPAT _IOW(FM_IOC_TYPE_BASE, FM_IOC_NUM(12), ioc_compat_fm_buffer_prefix_content_params_t)
-+#endif
-+#define FM_IOC_VSP_CONFIG_BUFFER_PREFIX_CONTENT _IOW(FM_IOC_TYPE_BASE, FM_IOC_NUM(12), ioc_fm_buffer_prefix_content_params_t)
-+
-+/**************************************************************************//**
-+ @Function FM_VSP_ConfigNoScatherGather
-+
-+ @Description Calling this routine changes the possibility to receive S/G frame
-+ in the internal driver data base
-+ from its default configuration: optimize = [DEFAULT_FM_SP_noScatherGather]
-+
-+ @Param[in] ioc_fm_vsp_config_no_sg_params_t A structure holding the required parameters.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_VSP_Config() and before FM_VSP_Init().
-+*//***************************************************************************/
-+#if defined(CONFIG_COMPAT)
-+#define FM_IOC_VSP_CONFIG_NO_SG_COMPAT _IOW(FM_IOC_TYPE_BASE, FM_IOC_NUM(13), ioc_compat_fm_vsp_config_no_sg_params_t)
-+#endif
-+#define FM_IOC_VSP_CONFIG_NO_SG _IOW(FM_IOC_TYPE_BASE, FM_IOC_NUM(13), ioc_fm_vsp_config_no_sg_params_t)
-+
-+/**************************************************************************//**
-+ @Function FM_VSP_GetBufferPrsResult
-+
-+ @Description Returns the pointer to the parse result in the data buffer.
-+ In Rx ports this is relevant after reception, if parse
-+ result is configured to be part of the data passed to the
-+ application. For non Rx ports it may be used to get the pointer
-+ of the area in the buffer where parse result should be
-+ initialized - if so configured.
-+ See FM_VSP_ConfigBufferPrefixContent for data buffer prefix
-+ configuration.
-+
-+ @Param[in] ioc_fm_vsp_prs_result_params_t A structure holding the required parameters.
-+
-+ @Return Parse result pointer on success, NULL if parse result was not
-+ configured for this port.
-+
-+ @Cautions Allowed only following FM_VSP_Init().
-+*//***************************************************************************/
-+#if defined(CONFIG_COMPAT)
-+#define FM_IOC_VSP_GET_BUFFER_PRS_RESULT_COMPAT _IOWR(FM_IOC_TYPE_BASE, FM_IOC_NUM(14), ioc_compat_fm_vsp_prs_result_params_t)
-+#endif
-+#define FM_IOC_VSP_GET_BUFFER_PRS_RESULT _IOWR(FM_IOC_TYPE_BASE, FM_IOC_NUM(14), ioc_fm_vsp_prs_result_params_t)
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+/**************************************************************************//**
-+ @Function FM_CtrlMonStart
-+
-+ @Description Start monitoring utilization of all available FM controllers.
-+
-+ In order to obtain FM controllers utilization the following sequence
-+ should be used:
-+ -# FM_CtrlMonStart()
-+ -# FM_CtrlMonStop()
-+ -# FM_CtrlMonGetCounters() - issued for each FM controller
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_Init().
-+*//***************************************************************************/
-+#define FM_IOC_CTRL_MON_START _IO(FM_IOC_TYPE_BASE, FM_IOC_NUM(15))
-+
-+
-+/**************************************************************************//**
-+ @Function FM_CtrlMonStop
-+
-+ @Description Stop monitoring utilization of all available FM controllers.
-+
-+ In order to obtain FM controllers utilization the following sequence
-+ should be used:
-+ -# FM_CtrlMonStart()
-+ -# FM_CtrlMonStop()
-+ -# FM_CtrlMonGetCounters() - issued for each FM controller
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_Init().
-+*//***************************************************************************/
-+#define FM_IOC_CTRL_MON_STOP _IO(FM_IOC_TYPE_BASE, FM_IOC_NUM(16))
-+
-+/**************************************************************************//**
-+ @Function FM_CtrlMonGetCounters
-+
-+ @Description Obtain FM controller utilization parameters.
-+
-+ In order to obtain FM controllers utilization the following sequence
-+ should be used:
-+ -# FM_CtrlMonStart()
-+ -# FM_CtrlMonStop()
-+ -# FM_CtrlMonGetCounters() - issued for each FM controller
-+
-+ @Param[in] ioc_fm_ctrl_mon_counters_params_t A structure holding the required parameters.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_Init().
-+*//***************************************************************************/
-+#if defined(CONFIG_COMPAT)
-+#define FM_IOC_CTRL_MON_GET_COUNTERS_COMPAT _IOW(FM_IOC_TYPE_BASE, FM_IOC_NUM(17), ioc_compat_fm_ctrl_mon_counters_params_t)
-+#endif
-+#define FM_IOC_CTRL_MON_GET_COUNTERS _IOW(FM_IOC_TYPE_BASE, FM_IOC_NUM(17), ioc_fm_ctrl_mon_counters_params_t)
-+
-+/** @} */ /* end of lnx_ioctl_FM_runtime_control_grp group */
-+/** @} */ /* end of lnx_ioctl_FM_lib_grp group */
-+/** @} */ /* end of lnx_ioctl_FM_grp */
-+
-+#define FMD_API_VERSION_MAJOR 21
-+#define FMD_API_VERSION_MINOR 1
-+#define FMD_API_VERSION_RESPIN 0
-+
-+#endif /* __FM_IOCTLS_H */
---- /dev/null
-+++ b/include/uapi/linux/fmd/Peripherals/fm_pcd_ioctls.h
-@@ -0,0 +1,3084 @@
-+/* Copyright (c) 2008-2012 Freescale Semiconductor, Inc.
-+ * All rights reserved.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/******************************************************************************
-+ @File fm_pcd_ioctls.h
-+
-+ @Description FM PCD ...
-+*//***************************************************************************/
-+#ifndef __FM_PCD_IOCTLS_H
-+#define __FM_PCD_IOCTLS_H
-+
-+#include "net_ioctls.h"
-+#include "fm_ioctls.h"
-+
-+
-+/**************************************************************************//**
-+
-+ @Group lnx_ioctl_FM_grp Frame Manager Linux IOCTL API
-+
-+ @Description Frame Manager Linux ioctls definitions and enums
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Group lnx_ioctl_FM_PCD_grp FM PCD
-+
-+ @Description Frame Manager PCD API functions, definitions and enums
-+
-+ The FM PCD module is responsible for the initialization of all
-+ global classifying FM modules. This includes the parser general and
-+ common registers, the key generator global and common registers,
-+ and the policer global and common registers.
-+ In addition, the FM PCD SW module will initialize all required
-+ key generator schemes, coarse classification flows, and policer
-+ profiles. When an FM module is configured to work with one of these
-+ entities, it will register to it using the FM PORT API. The PCD
-+ module will manage the PCD resources - i.e. resource management of
-+ KeyGen schemes, etc.
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Collection General PCD defines
-+*//***************************************************************************/
-+#define IOC_FM_PCD_MAX_NUM_OF_PRIVATE_HDRS 2 /**< Number of units/headers saved for user */
-+
-+#define IOC_FM_PCD_PRS_NUM_OF_HDRS 16 /**< Number of headers supported by HW parser */
-+#define IOC_FM_PCD_MAX_NUM_OF_DISTINCTION_UNITS (32 - IOC_FM_PCD_MAX_NUM_OF_PRIVATE_HDRS)
-+ /**< Number of distinction units is limited by
-+ register size (32 bits) minus reserved bits
-+ for private headers. */
-+#define IOC_FM_PCD_MAX_NUM_OF_INTERCHANGEABLE_HDRS 4 /**< Maximum number of interchangeable headers
-+ in a distinction unit */
-+#define IOC_FM_PCD_KG_NUM_OF_GENERIC_REGS 8 /**< Total number of generic KeyGen registers */
-+#define IOC_FM_PCD_KG_MAX_NUM_OF_EXTRACTS_PER_KEY 35 /**< Max number allowed on any configuration;
-+ For HW implementation reasons, in most
-+ cases less than this will be allowed; The
-+ driver will return an initialization error
-+ if resource is unavailable. */
-+#define IOC_FM_PCD_KG_NUM_OF_EXTRACT_MASKS 4 /**< Total number of masks allowed on KeyGen extractions. */
-+#define IOC_FM_PCD_KG_NUM_OF_DEFAULT_GROUPS 16 /**< Number of default value logical groups */
-+
-+#define IOC_FM_PCD_PRS_NUM_OF_LABELS 32 /**< Maximum number of SW parser labels */
-+#define IOC_FM_PCD_SW_PRS_SIZE 0x00000800 /**< Total size of SW parser area */
-+
-+#define IOC_FM_PCD_MAX_MANIP_INSRT_TEMPLATE_SIZE 128 /**< Maximum size of insertion template for
-+ insert manipulation */
-+
-+#if DPAA_VERSION >= 11
-+#define IOC_FM_PCD_FRM_REPLIC_MAX_NUM_OF_ENTRIES 64 /**< Maximum possible entries for frame replicator group */
-+#endif /* DPAA_VERSION >= 11 */
-+/* @} */
-+
-+#ifdef FM_CAPWAP_SUPPORT
-+#error "FM_CAPWAP_SUPPORT not implemented!"
-+#endif
-+
-+
-+/**************************************************************************//**
-+ @Group lnx_ioctl_FM_PCD_init_grp FM PCD Initialization Unit
-+
-+ @Description Frame Manager PCD Initialization Unit API
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Description PCD counters
-+ (must match enum e_FmPcdCounters defined in fm_pcd_ext.h)
-+*//***************************************************************************/
-+typedef enum ioc_fm_pcd_counters {
-+ e_IOC_FM_PCD_KG_COUNTERS_TOTAL, /**< KeyGen counter */
-+ e_IOC_FM_PCD_PLCR_COUNTERS_RED, /**< Policer counter - counts the total number of RED packets that exit the Policer. */
-+ e_IOC_FM_PCD_PLCR_COUNTERS_YELLOW, /**< Policer counter - counts the total number of YELLOW packets that exit the Policer. */
-+ e_IOC_FM_PCD_PLCR_COUNTERS_RECOLORED_TO_RED, /**< Policer counter - counts the number of packets that changed color to RED by the Policer;
-+ This is a subset of e_IOC_FM_PCD_PLCR_COUNTERS_RED packet count, indicating active color changes. */
-+ e_IOC_FM_PCD_PLCR_COUNTERS_RECOLORED_TO_YELLOW, /**< Policer counter - counts the number of packets that changed color to YELLOW by the Policer;
-+ This is a subset of e_IOC_FM_PCD_PLCR_COUNTERS_YELLOW packet count, indicating active color changes. */
-+ e_IOC_FM_PCD_PLCR_COUNTERS_TOTAL, /**< Policer counter - counts the total number of packets passed in the Policer. */
-+ e_IOC_FM_PCD_PLCR_COUNTERS_LENGTH_MISMATCH, /**< Policer counter - counts the number of packets with length mismatch. */
-+ e_IOC_FM_PCD_PRS_COUNTERS_PARSE_DISPATCH, /**< Parser counter - counts the number of times the parser block is dispatched. */
-+ e_IOC_FM_PCD_PRS_COUNTERS_L2_PARSE_RESULT_RETURNED, /**< Parser counter - counts the number of times L2 parse result is returned (including errors). */
-+ e_IOC_FM_PCD_PRS_COUNTERS_L3_PARSE_RESULT_RETURNED, /**< Parser counter - counts the number of times L3 parse result is returned (including errors). */
-+ e_IOC_FM_PCD_PRS_COUNTERS_L4_PARSE_RESULT_RETURNED, /**< Parser counter - counts the number of times L4 parse result is returned (including errors). */
-+ e_IOC_FM_PCD_PRS_COUNTERS_SHIM_PARSE_RESULT_RETURNED, /**< Parser counter - counts the number of times SHIM parse result is returned (including errors). */
-+ e_IOC_FM_PCD_PRS_COUNTERS_L2_PARSE_RESULT_RETURNED_WITH_ERR, /**< Parser counter - counts the number of times L2 parse result is returned with errors. */
-+ e_IOC_FM_PCD_PRS_COUNTERS_L3_PARSE_RESULT_RETURNED_WITH_ERR, /**< Parser counter - counts the number of times L3 parse result is returned with errors. */
-+ e_IOC_FM_PCD_PRS_COUNTERS_L4_PARSE_RESULT_RETURNED_WITH_ERR, /**< Parser counter - counts the number of times L4 parse result is returned with errors. */
-+ e_IOC_FM_PCD_PRS_COUNTERS_SHIM_PARSE_RESULT_RETURNED_WITH_ERR, /**< Parser counter - counts the number of times SHIM parse result is returned with errors. */
-+ e_IOC_FM_PCD_PRS_COUNTERS_SOFT_PRS_CYCLES, /**< Parser counter - counts the number of cycles spent executing soft parser instruction (including stall cycles). */
-+ e_IOC_FM_PCD_PRS_COUNTERS_SOFT_PRS_STALL_CYCLES, /**< Parser counter - counts the number of cycles stalled waiting for parser internal memory reads while executing soft parser instruction. */
-+ e_IOC_FM_PCD_PRS_COUNTERS_HARD_PRS_CYCLE_INCL_STALL_CYCLES, /**< Parser counter - counts the number of cycles spent executing hard parser (including stall cycles). */
-+ e_IOC_FM_PCD_PRS_COUNTERS_MURAM_READ_CYCLES, /**< MURAM counter - counts the number of cycles while performing FMan Memory read. */
-+ e_IOC_FM_PCD_PRS_COUNTERS_MURAM_READ_STALL_CYCLES, /**< MURAM counter - counts the number of cycles stalled while performing FMan Memory read. */
-+ e_IOC_FM_PCD_PRS_COUNTERS_MURAM_WRITE_CYCLES, /**< MURAM counter - counts the number of cycles while performing FMan Memory write. */
-+ e_IOC_FM_PCD_PRS_COUNTERS_MURAM_WRITE_STALL_CYCLES, /**< MURAM counter - counts the number of cycles stalled while performing FMan Memory write. */
-+ e_IOC_FM_PCD_PRS_COUNTERS_FPM_COMMAND_STALL_CYCLES /**< FPM counter - counts the number of cycles stalled while performing a FPM Command. */
-+} ioc_fm_pcd_counters;
-+
-+/**************************************************************************//**
-+ @Description PCD interrupts
-+ (must match enum e_FmPcdExceptions defined in fm_pcd_ext.h)
-+*//***************************************************************************/
-+typedef enum ioc_fm_pcd_exceptions {
-+ e_IOC_FM_PCD_KG_EXCEPTION_DOUBLE_ECC, /**< KeyGen double-bit ECC error is detected on internal memory read access. */
-+ e_IOC_FM_PCD_KG_EXCEPTION_KEYSIZE_OVERFLOW, /**< KeyGen scheme configuration error indicating a key size larger than 56 bytes. */
-+ e_IOC_FM_PCD_PLCR_EXCEPTION_DOUBLE_ECC, /**< Policer double-bit ECC error has been detected on PRAM read access. */
-+ e_IOC_FM_PCD_PLCR_EXCEPTION_INIT_ENTRY_ERROR, /**< Policer access to a non-initialized profile has been detected. */
-+ e_IOC_FM_PCD_PLCR_EXCEPTION_PRAM_SELF_INIT_COMPLETE, /**< Policer RAM self-initialization complete */
-+ e_IOC_FM_PCD_PLCR_EXCEPTION_ATOMIC_ACTION_COMPLETE, /**< Policer atomic action complete */
-+ e_IOC_FM_PCD_PRS_EXCEPTION_DOUBLE_ECC, /**< Parser double-bit ECC error */
-+ e_IOC_FM_PCD_PRS_EXCEPTION_SINGLE_ECC /**< Parser single-bit ECC error */
-+} ioc_fm_pcd_exceptions;
-+
-+/** @} */ /* end of lnx_ioctl_FM_PCD_init_grp group */
-+
-+
-+/**************************************************************************//**
-+ @Group lnx_ioctl_FM_PCD_Runtime_grp FM PCD Runtime Unit
-+
-+ @Description Frame Manager PCD Runtime Unit
-+
-+ The runtime control allows creation of PCD infrastructure modules
-+ such as Network Environment Characteristics, Classification Plan
-+ Groups and Coarse Classification Trees.
-+ It also allows on-the-fly initialization, modification and removal
-+ of PCD modules such as KeyGen schemes, coarse classification nodes
-+ and Policer profiles.
-+
-+ In order to explain the programming model of the PCD driver interface
-+ a few terms should be explained, and will be used below.
-+ - Distinction Header - One of the 16 protocols supported by the FM parser,
-+ or one of the SHIM headers (1 or 2). May be a header with a special
-+ option (see below).
-+ - Interchangeable Headers Group - This is a group of Headers recognized
-+ by either one of them. For example, if in a specific context the user
-+ chooses to treat IPv4 and IPV6 in the same way, they may create an
-+ interchangeable Headers Unit consisting of these 2 headers.
-+ - A Distinction Unit - a Distinction Header or an Interchangeable Headers
-+ Group.
-+ - Header with special option - applies to Ethernet, MPLS, VLAN, IPv4 and
-+ IPv6, includes multicast, broadcast and other protocol specific options.
-+ In terms of hardware it relates to the options available in the classification
-+ plan.
-+ - Network Environment Characteristics - a set of Distinction Units that define
-+ the total recognizable header selection for a certain environment. This is
-+ NOT the list of all headers that will ever appear in a flow, but rather
-+ everything that needs distinction in a flow, where distinction is made by KeyGen
-+ schemes and coarse classification action descriptors.
-+
-+ The PCD runtime modules initialization is done in stages. The first stage after
-+ initializing the PCD module itself is to establish a Network Flows Environment
-+ Definition. The application may choose to establish one or more such environments.
-+ Later, when needed, the application will have to state, for some of its modules,
-+ to which single environment it belongs.
-+
-+ @{
-+*//***************************************************************************/
-+
-+
-+/**************************************************************************//**
-+ @Description structure for FM counters
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_counters_params_t {
-+ ioc_fm_pcd_counters cnt; /**< The requested counter */
-+ uint32_t val; /**< The requested value to get/set from/into the counter */
-+} ioc_fm_pcd_counters_params_t;
-+
-+/**************************************************************************//**
-+ @Description structure for FM exception definitios
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_exception_params_t {
-+ ioc_fm_pcd_exceptions exception; /**< The requested exception */
-+ bool enable; /**< TRUE to enable interrupt, FALSE to mask it. */
-+} ioc_fm_pcd_exception_params_t;
-+
-+/**************************************************************************//**
-+ @Description A structure for SW parser labels
-+ (must be identical to struct t_FmPcdPrsLabelParams defined in fm_pcd_ext.h)
-+ *//***************************************************************************/
-+typedef struct ioc_fm_pcd_prs_label_params_t {
-+ uint32_t instruction_offset; /**< SW parser label instruction offset (2 bytes
-+ resolution), relative to Parser RAM. */
-+ ioc_net_header_type hdr; /**< The existence of this header will invoke
-+ the SW parser code. */
-+ uint8_t index_per_hdr; /**< Normally 0, if more than one SW parser
-+ attachments for the same header, use this
-+ index to distinguish between them. */
-+} ioc_fm_pcd_prs_label_params_t;
-+
-+/**************************************************************************//**
-+ @Description A structure for SW parser
-+ (Must match struct t_FmPcdPrsSwParams defined in fm_pcd_ext.h)
-+ *//***************************************************************************/
-+typedef struct ioc_fm_pcd_prs_sw_params_t {
-+ bool override; /**< FALSE to invoke a check that nothing else
-+ was loaded to this address, including
-+ internal patches.
-+ TRUE to override any existing code.*/
-+ uint32_t size; /**< SW parser code size */
-+ uint16_t base; /**< SW parser base (in instruction counts!
-+ must be larger than 0x20)*/
-+ uint8_t *p_code; /**< SW parser code */
-+ uint32_t sw_prs_data_params[IOC_FM_PCD_PRS_NUM_OF_HDRS];
-+ /**< SW parser data (parameters) */
-+ uint8_t num_of_labels; /**< Number of labels for SW parser. */
-+ ioc_fm_pcd_prs_label_params_t labels_table[IOC_FM_PCD_PRS_NUM_OF_LABELS];
-+ /**< SW parser labels table,
-+ containing num_of_labels entries */
-+} ioc_fm_pcd_prs_sw_params_t;
-+
-+/**************************************************************************//**
-+ @Description A structure to set the a KeyGen default value
-+ *//***************************************************************************/
-+typedef struct ioc_fm_pcd_kg_dflt_value_params_t {
-+ uint8_t valueId; /**< 0,1 - one of 2 global default values */
-+ uint32_t value; /**< The requested default value */
-+} ioc_fm_pcd_kg_dflt_value_params_t;
-+
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_Enable
-+
-+ @Description This routine should be called after PCD is initialized for enabling all
-+ PCD engines according to their existing configuration.
-+
-+ @Return 0 on success; Error code otherwise.
-+
-+ @Cautions Allowed only when PCD is disabled.
-+*//***************************************************************************/
-+#define FM_PCD_IOC_ENABLE _IO(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(1))
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_Disable
-+
-+ @Description This routine may be called when PCD is enabled in order to
-+ disable all PCD engines. It may be called
-+ only when none of the ports in the system are using the PCD.
-+
-+ @Return 0 on success; Error code otherwise.
-+
-+ @Cautions Allowed only when PCD is enabled.
-+*//***************************************************************************/
-+#define FM_PCD_IOC_DISABLE _IO(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(2))
-+
-+ /**************************************************************************//**
-+ @Function FM_PCD_PrsLoadSw
-+
-+ @Description This routine may be called only when all ports in the
-+ system are actively using the classification plan scheme.
-+ In such cases it is recommended in order to save resources.
-+ The driver automatically saves 8 classification plans for
-+ ports that do NOT use the classification plan mechanism, to
-+ avoid this (in order to save those entries) this routine may
-+ be called.
-+
-+ @Param[in] ioc_fm_pcd_prs_sw_params_t A pointer to the image of the software parser code.
-+
-+ @Return 0 on success; Error code otherwise.
-+
-+ @Cautions Allowed only when PCD is disabled.
-+*//***************************************************************************/
-+#if defined(CONFIG_COMPAT)
-+#define FM_PCD_IOC_PRS_LOAD_SW_COMPAT _IOW(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(3), ioc_compat_fm_pcd_prs_sw_params_t)
-+#endif
-+#define FM_PCD_IOC_PRS_LOAD_SW _IOW(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(3), ioc_fm_pcd_prs_sw_params_t)
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_KgSetDfltValue
-+
-+ @Description Calling this routine sets a global default value to be used
-+ by the KeyGen when parser does not recognize a required
-+ field/header.
-+ By default default values are 0.
-+
-+ @Param[in] ioc_fm_pcd_kg_dflt_value_params_t A pointer to a structure with the relevant parameters
-+
-+ @Return 0 on success; Error code otherwise.
-+
-+ @Cautions Allowed only when PCD is disabled.
-+*//***************************************************************************/
-+#define FM_PCD_IOC_KG_SET_DFLT_VALUE _IOW(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(6), ioc_fm_pcd_kg_dflt_value_params_t)
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_KgSetAdditionalDataAfterParsing
-+
-+ @Description Calling this routine allows the keygen to access data past
-+ the parser finishing point.
-+
-+ @Param[in] uint8_t payload-offset; the number of bytes beyond the parser location.
-+
-+ @Return 0 on success; Error code otherwise.
-+
-+ @Cautions Allowed only when PCD is disabled.
-+*//***************************************************************************/
-+#define FM_PCD_IOC_KG_SET_ADDITIONAL_DATA_AFTER_PARSING _IOW(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(7), uint8_t)
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_SetException
-+
-+ @Description Calling this routine enables/disables PCD interrupts.
-+
-+ @Param[in] ioc_fm_pcd_exception_params_t Arguments struct with exception to be enabled/disabled.
-+
-+ @Return 0 on success; Error code otherwise.
-+*//***************************************************************************/
-+#define FM_PCD_IOC_SET_EXCEPTION _IOW(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(8), ioc_fm_pcd_exception_params_t)
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_GetCounter
-+
-+ @Description Reads one of the FM PCD counters.
-+
-+ @Param[in,out] ioc_fm_pcd_counters_params_t The requested counter parameters.
-+
-+ @Return 0 on success; Error code otherwise.
-+
-+ @Cautions Note that it is user's responsibilty to call this routine only
-+ for enabled counters, and there will be no indication if a
-+ disabled counter is accessed.
-+*//***************************************************************************/
-+#define FM_PCD_IOC_GET_COUNTER _IOWR(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(9), ioc_fm_pcd_counters_params_t)
-+
-+/**************************************************************************//**
-+
-+ @Function FM_PCD_KgSchemeGetCounter
-+
-+ @Description Reads scheme packet counter.
-+
-+ @Param[in] h_Scheme scheme handle as returned by FM_PCD_KgSchemeSet().
-+
-+ @Return Counter's current value.
-+
-+ @Cautions Allowed only following FM_PCD_Init() & FM_PCD_KgSchemeSet().
-+*//***************************************************************************/
-+#if defined(CONFIG_COMPAT)
-+#define FM_PCD_IOC_KG_SCHEME_GET_CNTR_COMPAT _IOR(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(4), ioc_compat_fm_pcd_kg_scheme_spc_t)
-+#endif
-+#define FM_PCD_IOC_KG_SCHEME_GET_CNTR _IOR(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(4), ioc_fm_pcd_kg_scheme_spc_t)
-+
-+#if 0
-+TODO: unused IOCTL
-+/**************************************************************************//**
-+ @Function FM_PCD_ModifyCounter
-+
-+ @Description Writes a value to an enabled counter. Use "0" to reset the counter.
-+
-+ @Param[in] ioc_fm_pcd_counters_params_t - The requested counter parameters.
-+
-+ @Return 0 on success; Error code otherwise.
-+*//***************************************************************************/
-+#define FM_PCD_IOC_MODIFY_COUNTER _IOW(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(10), ioc_fm_pcd_counters_params_t)
-+#define FM_PCD_IOC_SET_COUNTER FM_PCD_IOC_MODIFY_COUNTER
-+#endif
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_ForceIntr
-+
-+ @Description Causes an interrupt event on the requested source.
-+
-+ @Param[in] ioc_fm_pcd_exceptions - An exception to be forced.
-+
-+ @Return 0 on success; error code if the exception is not enabled,
-+ or is not able to create interrupt.
-+*//***************************************************************************/
-+#define FM_PCD_IOC_FORCE_INTR _IOW(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(11), ioc_fm_pcd_exceptions)
-+
-+/**************************************************************************//**
-+ @Collection Definitions of coarse classification parameters as required by KeyGen
-+ (when coarse classification is the next engine after this scheme).
-+*//***************************************************************************/
-+#define IOC_FM_PCD_MAX_NUM_OF_CC_TREES 8
-+#define IOC_FM_PCD_MAX_NUM_OF_CC_GROUPS 16
-+#define IOC_FM_PCD_MAX_NUM_OF_CC_UNITS 4
-+#define IOC_FM_PCD_MAX_NUM_OF_KEYS 256
-+#define IOC_FM_PCD_MAX_NUM_OF_FLOWS (4*KILOBYTE)
-+#define IOC_FM_PCD_MAX_SIZE_OF_KEY 56
-+#define IOC_FM_PCD_MAX_NUM_OF_CC_ENTRIES_IN_GRP 16
-+#define IOC_FM_PCD_LAST_KEY_INDEX 0xffff
-+#define IOC_FM_PCD_MANIP_DSCP_VALUES 64
-+/* @} */
-+
-+/**************************************************************************//**
-+ @Collection A set of definitions to allow protocol
-+ special option description.
-+*//***************************************************************************/
-+typedef uint32_t ioc_protocol_opt_t; /**< A general type to define a protocol option. */
-+
-+typedef ioc_protocol_opt_t ioc_eth_protocol_opt_t; /**< Ethernet protocol options. */
-+#define IOC_ETH_BROADCAST 0x80000000 /**< Ethernet Broadcast. */
-+#define IOC_ETH_MULTICAST 0x40000000 /**< Ethernet Multicast. */
-+
-+typedef ioc_protocol_opt_t ioc_vlan_protocol_opt_t; /**< Vlan protocol options. */
-+#define IOC_VLAN_STACKED 0x20000000 /**< Stacked VLAN. */
-+
-+typedef ioc_protocol_opt_t ioc_mpls_protocol_opt_t; /**< MPLS protocol options. */
-+#define IOC_MPLS_STACKED 0x10000000 /**< Stacked MPLS. */
-+
-+typedef ioc_protocol_opt_t ioc_ipv4_protocol_opt_t; /**< IPv4 protocol options. */
-+#define IOC_IPV4_BROADCAST_1 0x08000000 /**< IPv4 Broadcast. */
-+#define IOC_IPV4_MULTICAST_1 0x04000000 /**< IPv4 Multicast. */
-+#define IOC_IPV4_UNICAST_2 0x02000000 /**< Tunneled IPv4 - Unicast. */
-+#define IOC_IPV4_MULTICAST_BROADCAST_2 0x01000000 /**< Tunneled IPv4 - Broadcast/Multicast. */
-+
-+#define IOC_IPV4_FRAG_1 0x00000008 /**< IPV4 reassembly option.
-+ IPV4 Reassembly manipulation requires network
-+ environment with IPV4 header and IPV4_FRAG_1 option */
-+
-+typedef ioc_protocol_opt_t ioc_ipv6_protocol_opt_t; /**< IPv6 protocol options. */
-+#define IOC_IPV6_MULTICAST_1 0x00800000 /**< IPv6 Multicast. */
-+#define IOC_IPV6_UNICAST_2 0x00400000 /**< Tunneled IPv6 - Unicast. */
-+#define IOC_IPV6_MULTICAST_2 0x00200000 /**< Tunneled IPv6 - Multicast. */
-+
-+#define IOC_IPV6_FRAG_1 0x00000004 /**< IPV6 reassembly option.
-+ IPV6 Reassembly manipulation requires network
-+ environment with IPV6 header and IPV6_FRAG_1 option */
-+#if (DPAA_VERSION >= 11)
-+typedef ioc_protocol_opt_t ioc_capwap_protocol_opt_t; /**< CAPWAP protocol options. */
-+#define CAPWAP_FRAG_1 0x00000008 /**< CAPWAP reassembly option.
-+ CAPWAP Reassembly manipulation requires network
-+ environment with CAPWAP header and CAPWAP_FRAG_1 option;
-+ in case where fragment found, the fragment-extension offset
-+ may be found at 'shim2' (in parser-result). */
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+/* @} */
-+
-+#define IOC_FM_PCD_MANIP_MAX_HDR_SIZE 256
-+#define IOC_FM_PCD_MANIP_DSCP_TO_VLAN_TRANS 64
-+/**************************************************************************//**
-+ @Collection A set of definitions to support Header Manipulation selection.
-+*//***************************************************************************/
-+typedef uint32_t ioc_hdr_manip_flags_t; /**< A general type to define a HMan update command flags. */
-+
-+typedef ioc_hdr_manip_flags_t ioc_ipv4_hdr_manip_update_flags_t; /**< IPv4 protocol HMan update command flags. */
-+
-+#define IOC_HDR_MANIP_IPV4_TOS 0x80000000 /**< update TOS with the given value ('tos' field
-+ of ioc_fm_pcd_manip_hdr_field_update_ipv4_t) */
-+#define IOC_HDR_MANIP_IPV4_ID 0x40000000 /**< update IP ID with the given value ('id' field
-+ of ioc_fm_pcd_manip_hdr_field_update_ipv4_t) */
-+#define IOC_HDR_MANIP_IPV4_TTL 0x20000000 /**< Decrement TTL by 1 */
-+#define IOC_HDR_MANIP_IPV4_SRC 0x10000000 /**< update IP source address with the given value
-+ ('src' field of ioc_fm_pcd_manip_hdr_field_update_ipv4_t) */
-+#define IOC_HDR_MANIP_IPV4_DST 0x08000000 /**< update IP destination address with the given value
-+ ('dst' field of ioc_fm_pcd_manip_hdr_field_update_ipv4_t) */
-+
-+typedef ioc_hdr_manip_flags_t ioc_ipv6_hdr_manip_update_flags_t; /**< IPv6 protocol HMan update command flags. */
-+
-+#define IOC_HDR_MANIP_IPV6_TC 0x80000000 /**< update Traffic Class address with the given value
-+ ('traffic_class' field of ioc_fm_pcd_manip_hdr_field_update_ipv6_t) */
-+#define IOC_HDR_MANIP_IPV6_HL 0x40000000 /**< Decrement Hop Limit by 1 */
-+#define IOC_HDR_MANIP_IPV6_SRC 0x20000000 /**< update IP source address with the given value
-+ ('src' field of ioc_fm_pcd_manip_hdr_field_update_ipv6_t) */
-+#define IOC_HDR_MANIP_IPV6_DST 0x10000000 /**< update IP destination address with the given value
-+ ('dst' field of ioc_fm_pcd_manip_hdr_field_update_ipv6_t) */
-+
-+typedef ioc_hdr_manip_flags_t ioc_tcp_udp_hdr_manip_update_flags_t;/**< TCP/UDP protocol HMan update command flags. */
-+
-+#define IOC_HDR_MANIP_TCP_UDP_SRC 0x80000000 /**< update TCP/UDP source address with the given value
-+ ('src' field of ioc_fm_pcd_manip_hdr_field_update_tcp_udp_t) */
-+#define IOC_HDR_MANIP_TCP_UDP_DST 0x40000000 /**< update TCP/UDP destination address with the given value
-+ ('dst' field of ioc_fm_pcd_manip_hdr_field_update_tcp_udp_t) */
-+#define IOC_HDR_MANIP_TCP_UDP_CHECKSUM 0x20000000 /**< update TCP/UDP checksum */
-+
-+/* @} */
-+
-+/**************************************************************************//**
-+ @Description A type used for returning the order of the key extraction.
-+ each value in this array represents the index of the extraction
-+ command as defined by the user in the initialization extraction array.
-+ The valid size of this array is the user define number of extractions
-+ required (also marked by the second '0' in this array).
-+*//***************************************************************************/
-+typedef uint8_t ioc_fm_pcd_kg_key_order_t [IOC_FM_PCD_KG_MAX_NUM_OF_EXTRACTS_PER_KEY];
-+
-+/**************************************************************************//**
-+ @Description All PCD engines
-+ (must match enum e_FmPcdEngine defined in fm_pcd_ext.h)
-+*//***************************************************************************/
-+typedef enum ioc_fm_pcd_engine {
-+ e_IOC_FM_PCD_INVALID = 0, /**< Invalid PCD engine */
-+ e_IOC_FM_PCD_DONE, /**< No PCD Engine indicated */
-+ e_IOC_FM_PCD_KG, /**< KeyGen */
-+ e_IOC_FM_PCD_CC, /**< Coarse Classifier */
-+ e_IOC_FM_PCD_PLCR, /**< Policer */
-+ e_IOC_FM_PCD_PRS, /**< Parser */
-+#if DPAA_VERSION >= 11
-+ e_IOC_FM_PCD_FR, /**< Frame Replicator */
-+#endif /* DPAA_VERSION >= 11 */
-+ e_IOC_FM_PCD_HASH /**< Hash Table */
-+} ioc_fm_pcd_engine;
-+
-+/**************************************************************************//**
-+ @Description An enum for selecting extraction by header types
-+ (Must match enum e_FmPcdExtractByHdrType defined in fm_pcd_ext.h)
-+*//***************************************************************************/
-+typedef enum ioc_fm_pcd_extract_by_hdr_type {
-+ e_IOC_FM_PCD_EXTRACT_FROM_HDR, /**< Extract bytes from header */
-+ e_IOC_FM_PCD_EXTRACT_FROM_FIELD, /**< Extract bytes from header field */
-+ e_IOC_FM_PCD_EXTRACT_FULL_FIELD /**< Extract a full field */
-+} ioc_fm_pcd_extract_by_hdr_type;
-+
-+/**************************************************************************//**
-+ @Description An enum for selecting extraction source (when it is not the header)
-+ (Must match enum e_FmPcdExtractFrom defined in fm_pcd_ext.h)
-+*//***************************************************************************/
-+typedef enum ioc_fm_pcd_extract_from {
-+ e_IOC_FM_PCD_EXTRACT_FROM_FRAME_START, /**< KG & CC: Extract from beginning of frame */
-+ e_IOC_FM_PCD_EXTRACT_FROM_DFLT_VALUE, /**< KG only: Extract from a default value */
-+ e_IOC_FM_PCD_EXTRACT_FROM_CURR_END_OF_PARSE, /**< KG only: Extract from the point where parsing had finished */
-+ e_IOC_FM_PCD_EXTRACT_FROM_KEY, /**< CC only: Field where saved KEY */
-+ e_IOC_FM_PCD_EXTRACT_FROM_HASH, /**< CC only: Field where saved HASH */
-+ e_IOC_FM_PCD_EXTRACT_FROM_PARSE_RESULT, /**< KG & CC: Extract from the parser result */
-+ e_IOC_FM_PCD_EXTRACT_FROM_ENQ_FQID, /**< KG & CC: Extract from enqueue FQID */
-+ e_IOC_FM_PCD_EXTRACT_FROM_FLOW_ID /**< CC only: Field where saved Dequeue FQID */
-+} ioc_fm_pcd_extract_from;
-+
-+/**************************************************************************//**
-+ @Description An enum for selecting extraction type
-+*//***************************************************************************/
-+typedef enum ioc_fm_pcd_extract_type {
-+ e_IOC_FM_PCD_EXTRACT_BY_HDR, /**< Extract according to header */
-+ e_IOC_FM_PCD_EXTRACT_NON_HDR, /**< Extract from data that is not the header */
-+ e_IOC_FM_PCD_KG_EXTRACT_PORT_PRIVATE_INFO /**< Extract private info as specified by user */
-+} ioc_fm_pcd_extract_type;
-+
-+/**************************************************************************//**
-+ @Description An enum for selecting a default
-+*//***************************************************************************/
-+typedef enum ioc_fm_pcd_kg_extract_dflt_select {
-+ e_IOC_FM_PCD_KG_DFLT_GBL_0, /**< Default selection is KG register 0 */
-+ e_IOC_FM_PCD_KG_DFLT_GBL_1, /**< Default selection is KG register 1 */
-+ e_IOC_FM_PCD_KG_DFLT_PRIVATE_0, /**< Default selection is a per scheme register 0 */
-+ e_IOC_FM_PCD_KG_DFLT_PRIVATE_1, /**< Default selection is a per scheme register 1 */
-+ e_IOC_FM_PCD_KG_DFLT_ILLEGAL /**< Illegal selection */
-+} ioc_fm_pcd_kg_extract_dflt_select;
-+
-+/**************************************************************************//**
-+ @Description Enumeration type defining all default groups - each group shares
-+ a default value, one of four user-initialized values.
-+*//***************************************************************************/
-+typedef enum ioc_fm_pcd_kg_known_fields_dflt_types {
-+ e_IOC_FM_PCD_KG_MAC_ADDR, /**< MAC Address */
-+ e_IOC_FM_PCD_KG_TCI, /**< TCI field */
-+ e_IOC_FM_PCD_KG_ENET_TYPE, /**< ENET Type */
-+ e_IOC_FM_PCD_KG_PPP_SESSION_ID, /**< PPP Session id */
-+ e_IOC_FM_PCD_KG_PPP_PROTOCOL_ID, /**< PPP Protocol id */
-+ e_IOC_FM_PCD_KG_MPLS_LABEL, /**< MPLS label */
-+ e_IOC_FM_PCD_KG_IP_ADDR, /**< IP addr */
-+ e_IOC_FM_PCD_KG_PROTOCOL_TYPE, /**< Protocol type */
-+ e_IOC_FM_PCD_KG_IP_TOS_TC, /**< TOS or TC */
-+ e_IOC_FM_PCD_KG_IPV6_FLOW_LABEL, /**< IPV6 flow label */
-+ e_IOC_FM_PCD_KG_IPSEC_SPI, /**< IPSEC SPI */
-+ e_IOC_FM_PCD_KG_L4_PORT, /**< L4 Port */
-+ e_IOC_FM_PCD_KG_TCP_FLAG, /**< TCP Flag */
-+ e_IOC_FM_PCD_KG_GENERIC_FROM_DATA, /**< grouping implemented by SW,
-+ any data extraction that is not the full
-+ field described above */
-+ e_IOC_FM_PCD_KG_GENERIC_FROM_DATA_NO_V, /**< grouping implemented by SW,
-+ any data extraction without validation */
-+ e_IOC_FM_PCD_KG_GENERIC_NOT_FROM_DATA /**< grouping implemented by SW,
-+ extraction from parser result or
-+ direct use of default value */
-+} ioc_fm_pcd_kg_known_fields_dflt_types;
-+
-+/**************************************************************************//**
-+ @Description Enumeration type for defining header index for scenarios with
-+ multiple (tunneled) headers
-+*//***************************************************************************/
-+typedef enum ioc_fm_pcd_hdr_index {
-+ e_IOC_FM_PCD_HDR_INDEX_NONE = 0, /**< used when multiple headers not used, also
-+ to specify regular IP (not tunneled). */
-+ e_IOC_FM_PCD_HDR_INDEX_1, /**< may be used for VLAN, MPLS, tunneled IP */
-+ e_IOC_FM_PCD_HDR_INDEX_2, /**< may be used for MPLS, tunneled IP */
-+ e_IOC_FM_PCD_HDR_INDEX_3, /**< may be used for MPLS */
-+ e_IOC_FM_PCD_HDR_INDEX_LAST = 0xFF /**< may be used for VLAN, MPLS */
-+} ioc_fm_pcd_hdr_index;
-+
-+/**************************************************************************//**
-+ @Description Enumeration type for selecting the policer profile functional type
-+*//***************************************************************************/
-+typedef enum ioc_fm_pcd_profile_type_selection {
-+ e_IOC_FM_PCD_PLCR_PORT_PRIVATE, /**< Port dedicated profile */
-+ e_IOC_FM_PCD_PLCR_SHARED /**< Shared profile (shared within partition) */
-+} ioc_fm_pcd_profile_type_selection;
-+
-+/**************************************************************************//**
-+ @Description Enumeration type for selecting the policer profile algorithm
-+*//***************************************************************************/
-+typedef enum ioc_fm_pcd_plcr_algorithm_selection {
-+ e_IOC_FM_PCD_PLCR_PASS_THROUGH, /**< Policer pass through */
-+ e_IOC_FM_PCD_PLCR_RFC_2698, /**< Policer algorithm RFC 2698 */
-+ e_IOC_FM_PCD_PLCR_RFC_4115 /**< Policer algorithm RFC 4115 */
-+} ioc_fm_pcd_plcr_algorithm_selection;
-+
-+/**************************************************************************//**
-+ @Description Enumeration type for selecting a policer profile color mode
-+*//***************************************************************************/
-+typedef enum ioc_fm_pcd_plcr_color_mode {
-+ e_IOC_FM_PCD_PLCR_COLOR_BLIND, /**< Color blind */
-+ e_IOC_FM_PCD_PLCR_COLOR_AWARE /**< Color aware */
-+} ioc_fm_pcd_plcr_color_mode;
-+
-+/**************************************************************************//**
-+ @Description Enumeration type for selecting a policer profile color
-+*//***************************************************************************/
-+typedef enum ioc_fm_pcd_plcr_color {
-+ e_IOC_FM_PCD_PLCR_GREEN, /**< Green */
-+ e_IOC_FM_PCD_PLCR_YELLOW, /**< Yellow */
-+ e_IOC_FM_PCD_PLCR_RED, /**< Red */
-+ e_IOC_FM_PCD_PLCR_OVERRIDE /**< Color override */
-+} ioc_fm_pcd_plcr_color;
-+
-+/**************************************************************************//**
-+ @Description Enumeration type for selecting the policer profile packet frame length selector
-+*//***************************************************************************/
-+typedef enum ioc_fm_pcd_plcr_frame_length_select {
-+ e_IOC_FM_PCD_PLCR_L2_FRM_LEN, /**< L2 frame length */
-+ e_IOC_FM_PCD_PLCR_L3_FRM_LEN, /**< L3 frame length */
-+ e_IOC_FM_PCD_PLCR_L4_FRM_LEN, /**< L4 frame length */
-+ e_IOC_FM_PCD_PLCR_FULL_FRM_LEN /**< Full frame length */
-+} ioc_fm_pcd_plcr_frame_length_select;
-+
-+/**************************************************************************//**
-+ @Description Enumeration type for selecting roll-back frame
-+*//***************************************************************************/
-+typedef enum ioc_fm_pcd_plcr_roll_back_frame_select {
-+ e_IOC_FM_PCD_PLCR_ROLLBACK_L2_FRM_LEN, /**< Rollback L2 frame length */
-+ e_IOC_FM_PCD_PLCR_ROLLBACK_FULL_FRM_LEN /**< Rollback Full frame length */
-+} ioc_fm_pcd_plcr_roll_back_frame_select;
-+
-+/**************************************************************************//**
-+ @Description Enumeration type for selecting the policer profile packet or byte mode
-+*//***************************************************************************/
-+typedef enum ioc_fm_pcd_plcr_rate_mode {
-+ e_IOC_FM_PCD_PLCR_BYTE_MODE, /**< Byte mode */
-+ e_IOC_FM_PCD_PLCR_PACKET_MODE /**< Packet mode */
-+} ioc_fm_pcd_plcr_rate_mode;
-+
-+/**************************************************************************//**
-+ @Description Enumeration type for defining action of frame
-+*//***************************************************************************/
-+typedef enum ioc_fm_pcd_done_action {
-+ e_IOC_FM_PCD_ENQ_FRAME = 0, /**< Enqueue frame */
-+ e_IOC_FM_PCD_DROP_FRAME /**< Drop frame */
-+} ioc_fm_pcd_done_action;
-+
-+/**************************************************************************//**
-+ @Description Enumeration type for selecting the policer counter
-+*//***************************************************************************/
-+typedef enum ioc_fm_pcd_plcr_profile_counters {
-+ e_IOC_FM_PCD_PLCR_PROFILE_GREEN_PACKET_TOTAL_COUNTER, /**< Green packets counter */
-+ e_IOC_FM_PCD_PLCR_PROFILE_YELLOW_PACKET_TOTAL_COUNTER, /**< Yellow packets counter */
-+ e_IOC_FM_PCD_PLCR_PROFILE_RED_PACKET_TOTAL_COUNTER, /**< Red packets counter */
-+ e_IOC_FM_PCD_PLCR_PROFILE_RECOLOURED_YELLOW_PACKET_TOTAL_COUNTER, /**< Recolored yellow packets counter */
-+ e_IOC_FM_PCD_PLCR_PROFILE_RECOLOURED_RED_PACKET_TOTAL_COUNTER /**< Recolored red packets counter */
-+} ioc_fm_pcd_plcr_profile_counters;
-+
-+/**************************************************************************//**
-+ @Description Enumeration type for selecting the PCD action after extraction
-+*//***************************************************************************/
-+typedef enum ioc_fm_pcd_action {
-+ e_IOC_FM_PCD_ACTION_NONE, /**< NONE */
-+ e_IOC_FM_PCD_ACTION_EXACT_MATCH, /**< Exact match on the selected extraction*/
-+ e_IOC_FM_PCD_ACTION_INDEXED_LOOKUP /**< Indexed lookup on the selected extraction*/
-+} ioc_fm_pcd_action;
-+
-+/**************************************************************************//**
-+ @Description Enumeration type for selecting type of insert manipulation
-+*//***************************************************************************/
-+typedef enum ioc_fm_pcd_manip_hdr_insrt_type {
-+ e_IOC_FM_PCD_MANIP_INSRT_GENERIC, /**< Insert according to offset & size */
-+ e_IOC_FM_PCD_MANIP_INSRT_BY_HDR, /**< Insert according to protocol */
-+#if (defined(FM_CAPWAP_SUPPORT) && (DPAA_VERSION == 10))
-+ e_IOC_FM_PCD_MANIP_INSRT_BY_TEMPLATE /**< Insert template to start of frame */
-+#endif /* FM_CAPWAP_SUPPORT */
-+} ioc_fm_pcd_manip_hdr_insrt_type;
-+
-+/**************************************************************************//**
-+ @Description Enumeration type for selecting type of remove manipulation
-+*//***************************************************************************/
-+typedef enum ioc_fm_pcd_manip_hdr_rmv_type {
-+ e_IOC_FM_PCD_MANIP_RMV_GENERIC, /**< Remove according to offset & size */
-+ e_IOC_FM_PCD_MANIP_RMV_BY_HDR /**< Remove according to offset & size */
-+} ioc_fm_pcd_manip_hdr_rmv_type;
-+
-+/**************************************************************************//**
-+ @Description An enum for selecting specific L2 fields removal
-+*//***************************************************************************/
-+typedef enum ioc_fm_pcd_manip_hdr_rmv_specific_l2 {
-+ e_IOC_FM_PCD_MANIP_HDR_RMV_ETHERNET, /**< Ethernet/802.3 MAC */
-+ e_IOC_FM_PCD_MANIP_HDR_RMV_STACKED_QTAGS, /**< stacked QTags */
-+ e_IOC_FM_PCD_MANIP_HDR_RMV_ETHERNET_AND_MPLS, /**< MPLS and Ethernet/802.3 MAC header until
-+ the header which follows the MPLS header */
-+ e_IOC_FM_PCD_MANIP_HDR_RMV_MPLS /**< Remove MPLS header (Unlimited MPLS labels) */
-+} ioc_fm_pcd_manip_hdr_rmv_specific_l2;
-+
-+/**************************************************************************//**
-+ @Description Enumeration type for selecting specific fields updates
-+*//***************************************************************************/
-+typedef enum ioc_fm_pcd_manip_hdr_field_update_type {
-+ e_IOC_FM_PCD_MANIP_HDR_FIELD_UPDATE_VLAN, /**< VLAN updates */
-+ e_IOC_FM_PCD_MANIP_HDR_FIELD_UPDATE_IPV4, /**< IPV4 updates */
-+ e_IOC_FM_PCD_MANIP_HDR_FIELD_UPDATE_IPV6, /**< IPV6 updates */
-+ e_IOC_FM_PCD_MANIP_HDR_FIELD_UPDATE_TCP_UDP, /**< TCP_UDP updates */
-+} ioc_fm_pcd_manip_hdr_field_update_type;
-+
-+/**************************************************************************//**
-+ @Description Enumeration type for selecting VLAN updates
-+*//***************************************************************************/
-+typedef enum ioc_fm_pcd_manip_hdr_field_update_vlan {
-+ e_IOC_FM_PCD_MANIP_HDR_FIELD_UPDATE_VLAN_VPRI, /**< Replace VPri of outer most VLAN tag. */
-+ e_IOC_FM_PCD_MANIP_HDR_FIELD_UPDATE_DSCP_TO_VLAN /**< DSCP to VLAN priority bits translation */
-+} ioc_fm_pcd_manip_hdr_field_update_vlan;
-+
-+/**************************************************************************//**
-+ @Description Enumeration type for selecting specific L2 fields removal
-+*//***************************************************************************/
-+typedef enum ioc_fm_pcd_manip_hdr_insrt_specific_l2 {
-+ e_IOC_FM_PCD_MANIP_HDR_INSRT_MPLS /**< Insert MPLS header (Unlimited MPLS labels) */
-+} ioc_fm_pcd_manip_hdr_insrt_specific_l2;
-+
-+#if (DPAA_VERSION >= 11)
-+/**************************************************************************//**
-+ @Description Enumeration type for selecting QoS mapping mode
-+
-+ Note: In all cases except 'e_FM_PCD_MANIP_HDR_QOS_MAPPING_NONE'
-+ User should instruct the port to read the parser-result
-+*//***************************************************************************/
-+typedef enum ioc_fm_pcd_manip_hdr_qos_mapping_mode {
-+ e_IOC_FM_PCD_MANIP_HDR_QOS_MAPPING_NONE = 0, /**< No mapping, QoS field will not be changed */
-+ e_IOC_FM_PCD_MANIP_HDR_QOS_MAPPING_AS_IS, /**< QoS field will be overwritten by the last byte in the parser-result. */
-+} ioc_fm_pcd_manip_hdr_qos_mapping_mode;
-+
-+/**************************************************************************//**
-+ @Description Enumeration type for selecting QoS source
-+
-+ Note: In all cases except 'e_FM_PCD_MANIP_HDR_QOS_SRC_NONE'
-+ User should left room for the parser-result on input/output buffer
-+ and instruct the port to read/write the parser-result to the buffer (RPD should be set)
-+*//***************************************************************************/
-+typedef enum ioc_fm_pcd_manip_hdr_qos_src {
-+ e_IOC_FM_PCD_MANIP_HDR_QOS_SRC_NONE = 0, /**< TODO */
-+ e_IOC_FM_PCD_MANIP_HDR_QOS_SRC_USER_DEFINED, /**< QoS will be taken from the last byte in the parser-result. */
-+} ioc_fm_pcd_manip_hdr_qos_src;
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+/**************************************************************************//**
-+ @Description Enumeration type for selecting type of header insertion
-+*//***************************************************************************/
-+typedef enum ioc_fm_pcd_manip_hdr_insrt_by_hdr_type {
-+ e_IOC_FM_PCD_MANIP_INSRT_BY_HDR_SPECIFIC_L2, /**< Specific L2 fields insertion */
-+#if (DPAA_VERSION >= 11)
-+ e_IOC_FM_PCD_MANIP_INSRT_BY_HDR_IP, /**< IP insertion */
-+ e_IOC_FM_PCD_MANIP_INSRT_BY_HDR_UDP, /**< UDP insertion */
-+ e_IOC_FM_PCD_MANIP_INSRT_BY_HDR_UDP_LITE, /**< UDP lite insertion */
-+ e_IOC_FM_PCD_MANIP_INSRT_BY_HDR_CAPWAP /**< CAPWAP insertion */
-+#endif /* (DPAA_VERSION >= 11) */
-+} ioc_fm_pcd_manip_hdr_insrt_by_hdr_type;
-+
-+/**************************************************************************//**
-+ @Description Enumeration type for selecting specific custom command
-+*//***************************************************************************/
-+typedef enum ioc_fm_pcd_manip_hdr_custom_type {
-+ e_IOC_FM_PCD_MANIP_HDR_CUSTOM_IP_REPLACE, /**< Replace IPv4/IPv6 */
-+} ioc_fm_pcd_manip_hdr_custom_type;
-+
-+/**************************************************************************//**
-+ @Description Enumeration type for selecting specific custom command
-+*//***************************************************************************/
-+typedef enum ioc_fm_pcd_manip_hdr_custom_ip_replace {
-+ e_IOC_FM_PCD_MANIP_HDR_CUSTOM_REPLACE_IPV4_BY_IPV6, /**< Replace IPv4 by IPv6 */
-+ e_IOC_FM_PCD_MANIP_HDR_CUSTOM_REPLACE_IPV6_BY_IPV4 /**< Replace IPv6 by IPv4 */
-+} ioc_fm_pcd_manip_hdr_custom_ip_replace;
-+
-+/**************************************************************************//**
-+ @Description Enumeration type for selecting type of header removal
-+*//***************************************************************************/
-+typedef enum ioc_fm_pcd_manip_hdr_rmv_by_hdr_type {
-+ e_IOC_FM_PCD_MANIP_RMV_BY_HDR_SPECIFIC_L2 = 0, /**< Specific L2 fields removal */
-+#if (DPAA_VERSION >= 11)
-+ e_IOC_FM_PCD_MANIP_RMV_BY_HDR_CAPWAP, /**< CAPWAP removal */
-+#endif /* (DPAA_VERSION >= 11) */
-+#if (DPAA_VERSION >= 11) || ((DPAA_VERSION == 10) && defined(FM_CAPWAP_SUPPORT))
-+ e_IOC_FM_PCD_MANIP_RMV_BY_HDR_FROM_START, /**< Locate from data that is not the header */
-+#endif /* (DPAA_VERSION >= 11) || ((DPAA_VERSION == 10) && defined(FM_CAPWAP_SUPPORT)) */
-+} ioc_fm_pcd_manip_hdr_rmv_by_hdr_type;
-+
-+/**************************************************************************//**
-+ @Description Enumeration type for selecting type of timeout mode
-+*//***************************************************************************/
-+typedef enum ioc_fm_pcd_manip_reassem_time_out_mode {
-+ e_IOC_FM_PCD_MANIP_TIME_OUT_BETWEEN_FRAMES, /**< Limits the time of the reassembly process
-+ from the first fragment to the last */
-+ e_IOC_FM_PCD_MANIP_TIME_OUT_BETWEEN_FRAG /**< Limits the time of receiving the fragment */
-+} ioc_fm_pcd_manip_reassem_time_out_mode;
-+
-+/**************************************************************************//**
-+ @Description Enumeration type for selecting type of WaysNumber mode
-+*//***************************************************************************/
-+typedef enum ioc_fm_pcd_manip_reassem_ways_number {
-+ e_IOC_FM_PCD_MANIP_ONE_WAY_HASH = 1, /**< One way hash */
-+ e_IOC_FM_PCD_MANIP_TWO_WAYS_HASH, /**< Two ways hash */
-+ e_IOC_FM_PCD_MANIP_THREE_WAYS_HASH, /**< Three ways hash */
-+ e_IOC_FM_PCD_MANIP_FOUR_WAYS_HASH, /**< Four ways hash */
-+ e_IOC_FM_PCD_MANIP_FIVE_WAYS_HASH, /**< Five ways hash */
-+ e_IOC_FM_PCD_MANIP_SIX_WAYS_HASH, /**< Six ways hash */
-+ e_IOC_FM_PCD_MANIP_SEVEN_WAYS_HASH, /**< Seven ways hash */
-+ e_IOC_FM_PCD_MANIP_EIGHT_WAYS_HASH /**< Eight ways hash */
-+} ioc_fm_pcd_manip_reassem_ways_number;
-+
-+#if (defined(FM_CAPWAP_SUPPORT) && (DPAA_VERSION == 10))
-+/**************************************************************************//**
-+ @Description Enumeration type for selecting type of statistics mode
-+*//***************************************************************************/
-+typedef enum ioc_fm_pcd_stats {
-+ e_IOC_FM_PCD_STATS_PER_FLOWID = 0 /**< Flow ID is used as index for getting statistics */
-+} ioc_fm_pcd_stats;
-+#endif
-+
-+/**************************************************************************//**
-+ @Description Enumeration type for selecting manipulation type
-+*//***************************************************************************/
-+typedef enum ioc_fm_pcd_manip_type {
-+ e_IOC_FM_PCD_MANIP_HDR = 0, /**< Header manipulation */
-+ e_IOC_FM_PCD_MANIP_REASSEM, /**< Reassembly */
-+ e_IOC_FM_PCD_MANIP_FRAG, /**< Fragmentation */
-+ e_IOC_FM_PCD_MANIP_SPECIAL_OFFLOAD /**< Special Offloading */
-+} ioc_fm_pcd_manip_type;
-+
-+/**************************************************************************//**
-+ @Description Enumeration type for selecting type of statistics mode
-+*//***************************************************************************/
-+typedef enum ioc_fm_pcd_cc_stats_mode {
-+ e_IOC_FM_PCD_CC_STATS_MODE_NONE = 0, /**< No statistics support */
-+ e_IOC_FM_PCD_CC_STATS_MODE_FRAME, /**< Frame count statistics */
-+ e_IOC_FM_PCD_CC_STATS_MODE_BYTE_AND_FRAME, /**< Byte and frame count statistics */
-+#if (DPAA_VERSION >= 11)
-+ e_IOC_FM_PCD_CC_STATS_MODE_RMON, /**< Byte and frame length range count statistics */
-+#endif /* (DPAA_VERSION >= 11) */
-+} ioc_fm_pcd_cc_stats_mode;
-+
-+/**************************************************************************//**
-+ @Description Enumeration type for determining the action in case an IP packet
-+ is larger than MTU but its DF (Don't Fragment) bit is set.
-+*//***************************************************************************/
-+typedef enum ioc_fm_pcd_manip_dont_frag_action {
-+ e_IOC_FM_PCD_MANIP_DISCARD_PACKET = 0, /**< Discard packet */
-+ e_IOC_FM_PCD_MANIP_ENQ_TO_ERR_Q_OR_DISCARD_PACKET = e_IOC_FM_PCD_MANIP_DISCARD_PACKET,
-+ /**< Obsolete, cannot enqueue to error queue;
-+ In practice, selects to discard packets;
-+ Will be removed in the future */
-+ e_IOC_FM_PCD_MANIP_FRAGMENT_PACKECT, /**< Fragment packet and continue normal processing */
-+ e_IOC_FM_PCD_MANIP_CONTINUE_WITHOUT_FRAG /**< Continue normal processing without fragmenting the packet */
-+} ioc_fm_pcd_manip_dont_frag_action;
-+
-+/**************************************************************************//**
-+ @Description Enumeration type for selecting type of special offload manipulation
-+*//***************************************************************************/
-+typedef enum ioc_fm_pcd_manip_special_offload_type {
-+ e_IOC_FM_PCD_MANIP_SPECIAL_OFFLOAD_IPSEC, /**< IPSec offload manipulation */
-+#if (DPAA_VERSION >= 11)
-+ e_IOC_FM_PCD_MANIP_SPECIAL_OFFLOAD_CAPWAP /**< CAPWAP offload manipulation */
-+#endif /* (DPAA_VERSION >= 11) */
-+} ioc_fm_pcd_manip_special_offload_type;
-+
-+/**************************************************************************//**
-+ @Description A union of protocol dependent special options
-+ (Must match union u_FmPcdHdrProtocolOpt defined in fm_pcd_ext.h)
-+*//***************************************************************************/
-+typedef union ioc_fm_pcd_hdr_protocol_opt_u {
-+ ioc_eth_protocol_opt_t eth_opt; /**< Ethernet options */
-+ ioc_vlan_protocol_opt_t vlan_opt; /**< Vlan options */
-+ ioc_mpls_protocol_opt_t mpls_opt; /**< MPLS options */
-+ ioc_ipv4_protocol_opt_t ipv4_opt; /**< IPv4 options */
-+ ioc_ipv6_protocol_opt_t ipv6_opt; /**< IPv6 options */
-+#if (DPAA_VERSION >= 11)
-+ ioc_capwap_protocol_opt_t capwap_opt; /**< CAPWAP options */
-+#endif /* (DPAA_VERSION >= 11) */
-+} ioc_fm_pcd_hdr_protocol_opt_u;
-+
-+/**************************************************************************//**
-+ @Description A union holding all known protocol fields
-+*//***************************************************************************/
-+typedef union ioc_fm_pcd_fields_u {
-+ ioc_header_field_eth_t eth; /**< Ethernet */
-+ ioc_header_field_vlan_t vlan; /**< VLAN */
-+ ioc_header_field_llc_snap_t llc_snap; /**< LLC SNAP */
-+ ioc_header_field_pppoe_t pppoe; /**< PPPoE */
-+ ioc_header_field_mpls_t mpls; /**< MPLS */
-+ ioc_header_field_ip_t ip; /**< IP */
-+ ioc_header_field_ipv4_t ipv4; /**< IPv4 */
-+ ioc_header_field_ipv6_t ipv6; /**< IPv6 */
-+ ioc_header_field_udp_t udp; /**< UDP */
-+ ioc_header_field_udp_lite_t udp_lite; /**< UDP_Lite */
-+ ioc_header_field_tcp_t tcp; /**< TCP */
-+ ioc_header_field_sctp_t sctp; /**< SCTP */
-+ ioc_header_field_dccp_t dccp; /**< DCCP */
-+ ioc_header_field_gre_t gre; /**< GRE */
-+ ioc_header_field_minencap_t minencap; /**< Minimal Encapsulation */
-+ ioc_header_field_ipsec_ah_t ipsec_ah; /**< IPSec AH */
-+ ioc_header_field_ipsec_esp_t ipsec_esp; /**< IPSec ESP */
-+ ioc_header_field_udp_encap_esp_t udp_encap_esp; /**< UDP Encapsulation ESP */
-+} ioc_fm_pcd_fields_u;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining header extraction for key generation
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_from_hdr_t {
-+ uint8_t size; /**< Size in byte */
-+ uint8_t offset; /**< Byte offset */
-+} ioc_fm_pcd_from_hdr_t;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining field extraction for key generation
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_from_field_t {
-+ ioc_fm_pcd_fields_u field; /**< Field selection */
-+ uint8_t size; /**< Size in byte */
-+ uint8_t offset; /**< Byte offset */
-+} ioc_fm_pcd_from_field_t;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining a single network environment unit
-+ A distinction unit should be defined if it will later be used
-+ by one or more PCD engines to distinguish between flows.
-+ (Must match struct t_FmPcdDistinctionUnit defined in fm_pcd_ext.h)
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_distinction_unit_t {
-+ struct {
-+ ioc_net_header_type hdr; /**< One of the headers supported by the FM */
-+ ioc_fm_pcd_hdr_protocol_opt_u opt; /**< Select only one option! */
-+ } hdrs[IOC_FM_PCD_MAX_NUM_OF_INTERCHANGEABLE_HDRS];
-+} ioc_fm_pcd_distinction_unit_t;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining all different distinction units supported
-+ by a specific PCD Network Environment Characteristics module.
-+
-+ Each unit represent a protocol or a group of protocols that may
-+ be used later by the different PCD engines to distinguish between flows.
-+ (Must match struct t_FmPcdNetEnvParams defined in fm_pcd_ext.h)
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_net_env_params_t {
-+ uint8_t num_of_distinction_units;/**< Number of different units to be identified */
-+ ioc_fm_pcd_distinction_unit_t units[IOC_FM_PCD_MAX_NUM_OF_DISTINCTION_UNITS];
-+ /**< An array of num_of_distinction_units of the
-+ different units to be identified */
-+ void *id; /**< Output parameter; Returns the net-env Id to be used */
-+} ioc_fm_pcd_net_env_params_t;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining a single extraction action when
-+ creating a key
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_extract_entry_t {
-+ ioc_fm_pcd_extract_type type; /**< Extraction type select */
-+ union {
-+ struct {
-+ ioc_net_header_type hdr; /**< Header selection */
-+ bool ignore_protocol_validation;
-+ /**< Ignore protocol validation */
-+ ioc_fm_pcd_hdr_index hdr_index; /**< Relevant only for MPLS, VLAN and tunneled
-+ IP. Otherwise should be cleared.*/
-+ ioc_fm_pcd_extract_by_hdr_type type; /**< Header extraction type select */
-+ union {
-+ ioc_fm_pcd_from_hdr_t from_hdr; /**< Extract bytes from header parameters */
-+ ioc_fm_pcd_from_field_t from_field; /**< Extract bytes from field parameters */
-+ ioc_fm_pcd_fields_u full_field; /**< Extract full field parameters */
-+ } extract_by_hdr_type;
-+ } extract_by_hdr; /**< Used when type = e_IOC_FM_PCD_KG_EXTRACT_BY_HDR */
-+ struct {
-+ ioc_fm_pcd_extract_from src; /**< Non-header extraction source */
-+ ioc_fm_pcd_action action; /**< Relevant for CC Only */
-+ uint16_t ic_indx_mask; /**< Relevant only for CC when
-+ action = e_IOC_FM_PCD_ACTION_INDEXED_LOOKUP;
-+ Note that the number of bits that are set within
-+ this mask must be log2 of the CC-node 'num_of_keys'.
-+ Note that the mask cannot be set on the lower bits. */
-+ uint8_t offset; /**< Byte offset */
-+ uint8_t size; /**< Size in bytes */
-+ } extract_non_hdr; /**< Used when type = e_IOC_FM_PCD_KG_EXTRACT_NON_HDR */
-+ } extract_params;
-+} ioc_fm_pcd_extract_entry_t;
-+
-+/**************************************************************************//**
-+ @Description A structure for defining masks for each extracted
-+ field in the key.
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_kg_extract_mask_t {
-+ uint8_t extract_array_index; /**< Index in the extraction array, as initialized by user */
-+ uint8_t offset; /**< Byte offset */
-+ uint8_t mask; /**< A byte mask (selected bits will be ignored) */
-+} ioc_fm_pcd_kg_extract_mask_t;
-+
-+/**************************************************************************//**
-+ @Description A structure for defining default selection per groups
-+ of fields
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_kg_extract_dflt_t {
-+ ioc_fm_pcd_kg_known_fields_dflt_types type; /**< Default type select*/
-+ ioc_fm_pcd_kg_extract_dflt_select dflt_select; /**< Default register select */
-+} ioc_fm_pcd_kg_extract_dflt_t;
-+
-+
-+/**************************************************************************//**
-+ @Description A structure for defining all parameters needed for
-+ generation a key and using a hash function
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_kg_key_extract_and_hash_params_t {
-+ uint32_t private_dflt0; /**< Scheme default register 0 */
-+ uint32_t private_dflt1; /**< Scheme default register 1 */
-+ uint8_t num_of_used_extracts; /**< defines the valid size of the following array */
-+ ioc_fm_pcd_extract_entry_t extract_array [IOC_FM_PCD_KG_MAX_NUM_OF_EXTRACTS_PER_KEY];
-+ /**< An array of extraction definitions. */
-+ uint8_t num_of_used_dflts; /**< defines the valid size of the following array */
-+ ioc_fm_pcd_kg_extract_dflt_t dflts[IOC_FM_PCD_KG_NUM_OF_DEFAULT_GROUPS];
-+ /**< For each extraction used in this scheme, specify the required
-+ default register to be used when header is not found.
-+ types not specified in this array will get undefined value. */
-+ uint8_t num_of_used_masks; /**< Defines the valid size of the following array */
-+ ioc_fm_pcd_kg_extract_mask_t masks[IOC_FM_PCD_KG_NUM_OF_EXTRACT_MASKS];
-+ uint8_t hash_shift; /**< Hash result right shift.
-+ Selects the 24 bits out of the 64 hash result.
-+ 0 means using the 24 LSB's, otherwise use the
-+ 24 LSB's after shifting right.*/
-+ uint32_t hash_distribution_num_of_fqids; /**< must be > 1 and a power of 2. Represents the range
-+ of queues for the key and hash functionality */
-+ uint8_t hash_distribution_fqids_shift; /**< selects the FQID bits that will be effected by the hash */
-+ bool symmetric_hash; /**< TRUE to generate the same hash for frames with swapped source and
-+ destination fields on all layers; If TRUE, driver will check that for
-+ all layers, if SRC extraction is selected, DST extraction must also be
-+ selected, and vice versa. */
-+} ioc_fm_pcd_kg_key_extract_and_hash_params_t;
-+
-+/**************************************************************************//**
-+ @Description A structure of parameters for defining a single
-+ Qid mask (extracted OR).
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_kg_extracted_or_params_t {
-+ ioc_fm_pcd_extract_type type; /**< Extraction type select */
-+ union {
-+ struct { /**< used when type = e_IOC_FM_PCD_KG_EXTRACT_BY_HDR */
-+ ioc_net_header_type hdr;
-+ ioc_fm_pcd_hdr_index hdr_index; /**< Relevant only for MPLS, VLAN and tunneled
-+ IP. Otherwise should be cleared.*/
-+ bool ignore_protocol_validation;
-+
-+ } extract_by_hdr;
-+ ioc_fm_pcd_extract_from src; /**< used when type = e_IOC_FM_PCD_KG_EXTRACT_NON_HDR */
-+ } extract_params;
-+ uint8_t extraction_offset; /**< Offset for extraction */
-+ ioc_fm_pcd_kg_extract_dflt_select dflt_value; /**< Select register from which extraction is taken if
-+ field not found */
-+ uint8_t mask; /**< Mask LSB byte of extraction (specified bits are ignored) */
-+ uint8_t bit_offset_in_fqid; /**< 0-31, Selects which bits of the 24 FQID bits to effect using
-+ the extracted byte; Assume byte is placed as the 8 MSB's in
-+ a 32 bit word where the lower bits
-+ are the FQID; i.e if bitOffsetInFqid=1 than its LSB
-+ will effect the FQID MSB, if bitOffsetInFqid=24 than the
-+ extracted byte will effect the 8 LSB's of the FQID,
-+ if bitOffsetInFqid=31 than the byte's MSB will effect
-+ the FQID's LSB; 0 means - no effect on FQID;
-+ Note that one, and only one of
-+ bitOffsetInFqid or bitOffsetInPlcrProfile must be set (i.e,
-+ extracted byte must effect either FQID or Policer profile).*/
-+ uint8_t bit_offset_in_plcr_profile;
-+ /**< 0-15, Selects which bits of the 8 policer profile id bits to
-+ effect using the extracted byte; Assume byte is placed
-+ as the 8 MSB's in a 16 bit word where the lower bits
-+ are the policer profile id; i.e if bitOffsetInPlcrProfile=1
-+ than its LSB will effect the profile MSB, if bitOffsetInFqid=8
-+ than the extracted byte will effect the whole policer profile id,
-+ if bitOffsetInFqid=15 than the byte's MSB will effect
-+ the Policer Profile id's LSB;
-+ 0 means - no effect on policer profile; Note that one, and only one of
-+ bitOffsetInFqid or bitOffsetInPlcrProfile must be set (i.e,
-+ extracted byte must effect either FQID or Policer profile).*/
-+} ioc_fm_pcd_kg_extracted_or_params_t;
-+
-+/**************************************************************************//**
-+ @Description A structure for configuring scheme counter
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_kg_scheme_counter_t {
-+ bool update; /**< FALSE to keep the current counter state
-+ and continue from that point, TRUE to update/reset
-+ the counter when the scheme is written. */
-+ uint32_t value; /**< If update=TRUE, this value will be written into the
-+ counter; clear this field to reset the counter. */
-+} ioc_fm_pcd_kg_scheme_counter_t;
-+
-+
-+/**************************************************************************//**
-+ @Description A structure for retrieving FMKG_SE_SPC
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_kg_scheme_spc_t {
-+ uint32_t val; /**< return value */
-+ void *id; /**< scheme handle */
-+} ioc_fm_pcd_kg_scheme_spc_t;
-+
-+/**************************************************************************//**
-+ @Description A structure for defining policer profile parameters as required by keygen
-+ (when policer is the next engine after this scheme).
-+ (Must match struct t_FmPcdKgPlcrProfile defined in fm_pcd_ext.h)
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_kg_plcr_profile_t {
-+ bool shared_profile; /**< TRUE if this profile is shared between ports
-+ (i.e. managed by master partition) May not be TRUE
-+ if profile is after Coarse Classification*/
-+ bool direct; /**< If TRUE, direct_relative_profile_id only selects the profile
-+ id, if FALSE fqid_offset_relative_profile_id_base is used
-+ together with fqid_offset_shift and num_of_profiles
-+ parameters, to define a range of profiles from
-+ which the KeyGen result will determine the
-+ destination policer profile. */
-+ union {
-+ uint16_t direct_relative_profile_id; /**< Used if 'direct' is TRUE, to select policer profile.
-+ This parameter should indicate the policer profile offset within the port's
-+ policer profiles or SHARED window. */
-+ struct {
-+ uint8_t fqid_offset_shift; /**< Shift of KG results without the qid base */
-+ uint8_t fqid_offset_relative_profile_id_base;
-+ /**< OR of KG results without the qid base
-+ This parameter should indicate the policer profile
-+ offset within the port's policer profiles window
-+ or SHARED window depends on shared_profile */
-+ uint8_t num_of_profiles; /**< Range of profiles starting at base */
-+ } indirect_profile; /**< Indirect profile parameters */
-+ } profile_select; /**< Direct/indirect profile selection and parameters */
-+} ioc_fm_pcd_kg_plcr_profile_t;
-+
-+#if DPAA_VERSION >= 11
-+/**************************************************************************//**
-+ @Description Parameters for configuring a storage profile for a KeyGen scheme.
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_kg_storage_profile_t {
-+ bool direct; /**< If TRUE, directRelativeProfileId only selects the
-+ profile id;
-+ If FALSE, fqidOffsetRelativeProfileIdBase is used
-+ together with fqidOffsetShift and numOfProfiles
-+ parameters to define a range of profiles from which
-+ the KeyGen result will determine the destination
-+ storage profile. */
-+ union {
-+ uint16_t direct_relative_profileId; /**< Used when 'direct' is TRUE, to select a storage profile;
-+ should indicate the storage profile offset within the
-+ port's storage profiles window. */
-+ struct {
-+ uint8_t fqid_offset_shift; /**< Shift of KeyGen results without the FQID base */
-+ uint8_t fqid_offset_relative_profile_id_base;
-+ /**< OR of KeyGen results without the FQID base;
-+ should indicate the policer profile offset within the
-+ port's storage profiles window. */
-+ uint8_t num_of_profiles; /**< Range of profiles starting at base. */
-+ } indirect_profile; /**< Indirect profile parameters. */
-+ } profile_select; /**< Direct/indirect profile selection and parameters. */
-+} ioc_fm_pcd_kg_storage_profile_t;
-+#endif /* DPAA_VERSION >= 11 */
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining CC as the next engine after KeyGen
-+ (Must match struct t_FmPcdKgCc defined in fm_pcd_ext.h)
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_kg_cc_t {
-+ void *tree_id; /**< CC Tree id */
-+ uint8_t grp_id; /**< CC group id within the CC tree */
-+ bool plcr_next; /**< TRUE if after CC, in case of data frame,
-+ policing is required. */
-+ bool bypass_plcr_profile_generation;
-+ /**< TRUE to bypass KeyGen policer profile generation;
-+ selected profile is the one set at port initialization. */
-+ ioc_fm_pcd_kg_plcr_profile_t plcr_profile; /**< Valid only if plcr_next = TRUE and
-+ bypass_plcr_profile_generation = FALSE */
-+} ioc_fm_pcd_kg_cc_t;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining initializing a KeyGen scheme
-+ (Must match struct t_FmPcdKgSchemeParams defined in fm_pcd_ext.h)
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_kg_scheme_params_t {
-+ bool modify; /**< TRUE to change an existing scheme */
-+ union {
-+ uint8_t relative_scheme_id;
-+ /**< if modify=FALSE: partition-relative scheme id */
-+ void *scheme_id; /**< if modify=TRUE: the id of an existing scheme */
-+ } scm_id;
-+ bool always_direct; /**< This scheme is reached only directly, i.e. no need
-+ for match vector; KeyGen will ignore it when matching */
-+ struct { /**< HL relevant only if always_direct=FALSE */
-+ void *net_env_id; /**< The id of the Network Environment as returned
-+ by FM_PCD_NetEnvCharacteristicsSet() */
-+ uint8_t num_of_distinction_units;
-+ /**< Number of NetEnv units listed in unit_ids array */
-+ uint8_t unit_ids[IOC_FM_PCD_MAX_NUM_OF_DISTINCTION_UNITS];
-+ /**< Indexes as passed to SetNetEnvCharacteristics (?) array */
-+ } net_env_params;
-+ bool use_hash; /**< use the KG Hash functionality */
-+ ioc_fm_pcd_kg_key_extract_and_hash_params_t key_extract_and_hash_params;
-+ /**< used only if useHash = TRUE */
-+ bool bypass_fqid_generation;
-+ /**< Normally - FALSE, TRUE to avoid FQID update in the IC;
-+ In such a case FQID after KG will be the default FQID
-+ defined for the relevant port, or the FQID defined by CC
-+ in cases where CC was the previous engine. */
-+ uint32_t base_fqid; /**< Base FQID; Relevant only if bypass_fqid_generation = FALSE;
-+ If hash is used and an even distribution is expected
-+ according to hash_distribution_num_of_fqids, base_fqid must be aligned to
-+ hash_distribution_num_of_fqids. */
-+ uint8_t num_of_used_extracted_ors;
-+ /**< Number of FQID masks listed in extracted_ors array*/
-+ ioc_fm_pcd_kg_extracted_or_params_t extracted_ors[IOC_FM_PCD_KG_NUM_OF_GENERIC_REGS];
-+ /**< IOC_FM_PCD_KG_NUM_OF_GENERIC_REGS
-+ registers are shared between qid_masks
-+ functionality and some of the extraction
-+ actions; Normally only some will be used
-+ for qid_mask. Driver will return error if
-+ resource is full at initialization time. */
-+#if DPAA_VERSION >= 11
-+ bool override_storage_profile;
-+ /**< TRUE if KeyGen override previously decided storage profile */
-+ ioc_fm_pcd_kg_storage_profile_t storage_profile;/**< Used when override_storage_profile=TRUE */
-+#endif /* DPAA_VERSION >= 11 */
-+ ioc_fm_pcd_engine next_engine; /**< may be BMI, PLCR or CC */
-+ union { /**< depends on nextEngine */
-+ ioc_fm_pcd_done_action done_action; /**< Used when next engine is BMI (done) */
-+ ioc_fm_pcd_kg_plcr_profile_t plcr_profile; /**< Used when next engine is PLCR */
-+ ioc_fm_pcd_kg_cc_t cc; /**< Used when next engine is CC */
-+ } kg_next_engine_params;
-+ ioc_fm_pcd_kg_scheme_counter_t scheme_counter; /**< A structure of parameters for updating
-+ the scheme counter */
-+ void *id; /**< Returns the scheme Id to be used */
-+} ioc_fm_pcd_kg_scheme_params_t;
-+
-+/**************************************************************************//**
-+ @Collection
-+*//***************************************************************************/
-+#if DPAA_VERSION >= 11
-+#define IOC_FM_PCD_CC_STATS_MAX_NUM_OF_FLR 10 /* Maximal supported number of frame length ranges */
-+#define IOC_FM_PCD_CC_STATS_FLR_SIZE 2 /* Size in bytes of a frame length range limit */
-+#endif /* DPAA_VERSION >= 11 */
-+#define IOC_FM_PCD_CC_STATS_FLR_COUNT_SIZE 4 /* Size in bytes of a frame length range counter */
-+/* @} */
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining CC as the next engine after a CC node.
-+ (Must match struct t_FmPcdCcNextCcParams defined in fm_pcd_ext.h)
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_cc_next_cc_params_t {
-+ void *cc_node_id; /**< Id of the next CC node */
-+} ioc_fm_pcd_cc_next_cc_params_t;
-+
-+#if DPAA_VERSION >= 11
-+/**************************************************************************//**
-+ @Description A structure for defining Frame Replicator as the next engine after a CC node.
-+ (Must match struct t_FmPcdCcNextFrParams defined in fm_pcd_ext.h)
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_cc_next_fr_params_t {
-+ void* frm_replic_id; /**< The id of the next frame replicator group */
-+} ioc_fm_pcd_cc_next_fr_params_t;
-+#endif /* DPAA_VERSION >= 11 */
-+
-+/**************************************************************************//**
-+ @Description A structure for defining PLCR params when PLCR is the
-+ next engine after a CC node
-+ (Must match struct t_FmPcdCcNextPlcrParams defined in fm_pcd_ext.h)
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_cc_next_plcr_params_t {
-+ bool override_params; /**< TRUE if CC override previously decided parameters*/
-+ bool shared_profile; /**< Relevant only if overrideParams=TRUE:
-+ TRUE if this profile is shared between ports */
-+ uint16_t new_relative_profile_id; /**< Relevant only if overrideParams=TRUE:
-+ (otherwise profile id is taken from keygen);
-+ This parameter should indicate the policer
-+ profile offset within the port's
-+ policer profiles or from SHARED window.*/
-+ uint32_t new_fqid; /**< Relevant only if overrideParams=TRUE:
-+ FQID for enquing the frame;
-+ In earlier chips if policer next engine is KEYGEN,
-+ this parameter can be 0, because the KEYGEN always decides
-+ the enqueue FQID.*/
-+#if DPAA_VERSION >= 11
-+ uint8_t new_relative_storage_profile_id;
-+ /**< Indicates the relative storage profile offset within
-+ the port's storage profiles window;
-+ Relevant only if the port was configured with VSP. */
-+#endif /* DPAA_VERSION >= 11 */
-+} ioc_fm_pcd_cc_next_plcr_params_t;
-+
-+/**************************************************************************//**
-+ @Description A structure for defining enqueue params when BMI is the
-+ next engine after a CC node
-+ (Must match struct t_FmPcdCcNextEnqueueParams defined in fm_pcd_ext.h)
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_cc_next_enqueue_params_t {
-+ ioc_fm_pcd_done_action action; /**< Action - when next engine is BMI (done) */
-+ bool override_fqid; /**< TRUE if CC override previously decided fqid and vspid,
-+ relevant if action = e_IOC_FM_PCD_ENQ_FRAME */
-+ uint32_t new_fqid; /**< Valid if overrideFqid=TRUE, FQID for enqueuing the frame
-+ (otherwise FQID is taken from KeyGen),
-+ relevant if action = e_IOC_FM_PCD_ENQ_FRAME*/
-+#if DPAA_VERSION >= 11
-+ uint8_t new_relative_storage_profile_id;
-+ /**< Valid if override_fqid=TRUE, Indicates the relative virtual
-+ storage profile offset within the port's storage profiles
-+ window; Relevant only if the port was configured with VSP. */
-+#endif /* DPAA_VERSION >= 11 */
-+
-+} ioc_fm_pcd_cc_next_enqueue_params_t;
-+
-+/**************************************************************************//**
-+ @Description A structure for defining KG params when KG is the next engine after a CC node
-+ (Must match struct t_FmPcdCcNextKgParams defined in fm_pcd_ext.h)
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_cc_next_kg_params_t {
-+ bool override_fqid; /**< TRUE if CC override previously decided fqid and vspid,
-+ Note - this parameters are irrelevant for earlier chips */
-+ uint32_t new_fqid; /**< Valid if overrideFqid=TRUE, FQID for enqueuing the frame
-+ (otherwise FQID is taken from KeyGen),
-+ Note - this parameters are irrelevant for earlier chips */
-+#if DPAA_VERSION >= 11
-+ uint8_t new_relative_storage_profile_id;
-+ /**< Valid if override_fqid=TRUE, Indicates the relative virtual
-+ storage profile offset within the port's storage profiles
-+ window; Relevant only if the port was configured with VSP. */
-+#endif /* DPAA_VERSION >= 11 */
-+ void *p_direct_scheme; /**< Direct scheme id to go to. */
-+} ioc_fm_pcd_cc_next_kg_params_t;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining the next engine after a CC node.
-+ (Must match struct t_FmPcdCcNextEngineParams defined in fm_pcd_ext.h)
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_cc_next_engine_params_t {
-+ ioc_fm_pcd_engine next_engine; /**< User has to initialize parameters
-+ according to nextEngine definition */
-+ union {
-+ ioc_fm_pcd_cc_next_cc_params_t cc_params; /**< Parameters in case next engine is CC */
-+ ioc_fm_pcd_cc_next_plcr_params_t plcr_params; /**< Parameters in case next engine is PLCR */
-+ ioc_fm_pcd_cc_next_enqueue_params_t enqueue_params; /**< Parameters in case next engine is BMI */
-+ ioc_fm_pcd_cc_next_kg_params_t kg_params; /**< Parameters in case next engine is KG */
-+#if DPAA_VERSION >= 11
-+ ioc_fm_pcd_cc_next_fr_params_t fr_params; /**< Parameters in case next engine is FR */
-+#endif /* DPAA_VERSION >= 11 */
-+ } params; /**< Union used for all the next-engine parameters options */
-+ void *manip_id; /**< Handle to Manipulation object.
-+ Relevant if next engine is of type result
-+ (e_IOC_FM_PCD_PLCR, e_IOC_FM_PCD_KG, e_IOC_FM_PCD_DONE) */
-+ bool statistics_en; /**< If TRUE, statistics counters are incremented
-+ for each frame passing through this
-+ Coarse Classification entry. */
-+} ioc_fm_pcd_cc_next_engine_params_t;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining a single CC key
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_cc_key_params_t {
-+ uint8_t *p_key; /**< pointer to the key of the size defined in key_size */
-+ uint8_t *p_mask; /**< pointer to the Mask per key of the size defined
-+ in keySize. p_key and p_mask (if defined) has to be
-+ of the same size defined in the key_size */
-+ ioc_fm_pcd_cc_next_engine_params_t cc_next_engine_params;
-+ /**< parameters for the next for the defined Key in p_key */
-+
-+} ioc_fm_pcd_cc_key_params_t;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining CC keys parameters
-+ The driver supports two methods for CC node allocation: dynamic and static.
-+ Static mode was created in order to prevent runtime alloc/free
-+ of FMan memory (MURAM), which may cause fragmentation; in this mode,
-+ the driver automatically allocates the memory according to
-+ 'max_num_of_keys' parameter. The driver calculates the maximal memory
-+ size that may be used for this CC-Node taking into consideration
-+ 'mask_support' and 'statistics_mode' parameters.
-+ When 'action' = e_IOC_FM_PCD_ACTION_INDEXED_LOOKUP in the extraction
-+ parameters of this node, 'max_num_of_keys' must be equal to 'num_of_keys'.
-+ In dynamic mode, 'max_num_of_keys' must be zero. At initialization,
-+ all required structures are allocated according to 'num_of_keys'
-+ parameter. During runtime modification, these structures are
-+ re-allocated according to the updated number of keys.
-+
-+ Please note that 'action' and 'ic_indx_mask' mentioned in the
-+ specific parameter explanations are passed in the extraction
-+ parameters of the node (fields of extractccparams.extractnonhdr).
-+*//***************************************************************************/
-+typedef struct ioc_keys_params_t {
-+ uint16_t max_num_of_keys;/**< Maximum number of keys that will (ever) be used in this CC-Node;
-+ A value of zero may be used for dynamic memory allocation. */
-+ bool mask_support; /**< This parameter is relevant only if a node is initialized with
-+ action = e_IOC_FM_PCD_ACTION_EXACT_MATCH and max_num_of_keys > 0;
-+ Should be TRUE to reserve table memory for key masks, even if
-+ initial keys do not contain masks, or if the node was initialized
-+ as 'empty' (without keys); this will allow user to add keys with
-+ masks at runtime. */
-+ ioc_fm_pcd_cc_stats_mode statistics_mode;/**< Determines the supported statistics mode for all node's keys.
-+ To enable statistics gathering, statistics should be enabled per
-+ every key, using 'statistics_en' in next engine parameters structure
-+ of that key;
-+ If 'max_num_of_keys' is set, all required structures will be
-+ preallocated for all keys. */
-+#if (DPAA_VERSION >= 11)
-+ uint16_t frame_length_ranges[IOC_FM_PCD_CC_STATS_MAX_NUM_OF_FLR];
-+ /**< Relevant only for 'RMON' statistics mode
-+ (this feature is supported only on B4860 device);
-+ Holds a list of programmable thresholds. For each received frame,
-+ its length in bytes is examined against these range thresholds and
-+ the appropriate counter is incremented by 1. For example, to belong
-+ to range i, the following should hold:
-+ range i-1 threshold < frame length <= range i threshold
-+ Each range threshold must be larger then its preceding range
-+ threshold. Last range threshold must be 0xFFFF. */
-+#endif /* (DPAA_VERSION >= 11) */
-+ uint16_t num_of_keys; /**< Number of initial keys;
-+ Note that in case of 'action' = e_IOC_FM_PCD_ACTION_INDEXED_LOOKUP,
-+ this field should be power-of-2 of the number of bits that are
-+ set in 'ic_indx_mask'. */
-+ uint8_t key_size; /**< Size of key - for extraction of type FULL_FIELD, 'key_size' has
-+ to be the standard size of the selected key; For other extraction
-+ types, 'key_size' has to be as size of extraction; When 'action' =
-+ e_IOC_FM_PCD_ACTION_INDEXED_LOOKUP, 'keySize' must be 2. */
-+ ioc_fm_pcd_cc_key_params_t key_params[IOC_FM_PCD_MAX_NUM_OF_KEYS];
-+ /**< An array with 'num_of_keys' entries, each entry specifies the
-+ corresponding key parameters;
-+ When 'action' = e_IOC_FM_PCD_ACTION_EXACT_MATCH, this value must not
-+ exceed 255 (IOC_FM_PCD_MAX_NUM_OF_KEYS-1) as the last entry is saved
-+ for the 'miss' entry. */
-+ ioc_fm_pcd_cc_next_engine_params_t cc_next_engine_params_for_miss;
-+ /**< Parameters for defining the next engine when a key is not matched;
-+ Not relevant if action = e_IOC_FM_PCD_ACTION_INDEXED_LOOKUP. */
-+} ioc_keys_params_t;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining a CC node
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_cc_node_params_t {
-+ ioc_fm_pcd_extract_entry_t extract_cc_params; /**< Extraction parameters */
-+ ioc_keys_params_t keys_params; /**< Keys definition matching the selected extraction */
-+ void *id; /**< Output parameter; returns the CC node Id to be used */
-+} ioc_fm_pcd_cc_node_params_t;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining a hash table
-+ (Must match struct t_FmPcdHashTableParams defined in fm_pcd_ext.h)
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_hash_table_params_t {
-+ uint16_t max_num_of_keys; /**< Maximum Number Of Keys that will (ever) be used in this Hash-table */
-+ ioc_fm_pcd_cc_stats_mode statistics_mode; /**< If not e_IOC_FM_PCD_CC_STATS_MODE_NONE, the required structures for the
-+ requested statistics mode will be allocated according to max_num_of_keys. */
-+ uint8_t kg_hash_shift; /**< KG-Hash-shift as it was configured in the KG-scheme
-+ that leads to this hash-table. */
-+ uint16_t hash_res_mask; /**< Mask that will be used on the hash-result;
-+ The number-of-sets for this hash will be calculated
-+ as (2^(number of bits set in 'hash_res_mask'));
-+ The 4 lower bits must be cleared. */
-+ uint8_t hash_shift; /**< Byte offset from the beginning of the KeyGen hash result to the
-+ 2-bytes to be used as hash index. */
-+ uint8_t match_key_size; /**< Size of the exact match keys held by the hash buckets */
-+
-+ ioc_fm_pcd_cc_next_engine_params_t cc_next_engine_params_for_miss;
-+ /**< Parameters for defining the next engine when a key is not matched */
-+ void *id;
-+} ioc_fm_pcd_hash_table_params_t;
-+
-+/**************************************************************************//**
-+ @Description A structure with the arguments for the FM_PCD_HashTableAddKey ioctl() call
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_hash_table_add_key_params_t {
-+ void *p_hash_tbl;
-+ uint8_t key_size;
-+ ioc_fm_pcd_cc_key_params_t key_params;
-+} ioc_fm_pcd_hash_table_add_key_params_t;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining a CC tree group.
-+
-+ This structure defines a CC group in terms of NetEnv units
-+ and the action to be taken in each case. The unit_ids list must
-+ be given in order from low to high indices.
-+
-+ ioc_fm_pcd_cc_next_engine_params_t is a list of 2^num_of_distinction_units
-+ structures where each defines the next action to be taken for
-+ each units combination. for example:
-+ num_of_distinction_units = 2
-+ unit_ids = {1,3}
-+ next_engine_per_entries_in_grp[0] = ioc_fm_pcd_cc_next_engine_params_t for the case that
-+ unit 1 - not found; unit 3 - not found;
-+ next_engine_per_entries_in_grp[1] = ioc_fm_pcd_cc_next_engine_params_t for the case that
-+ unit 1 - not found; unit 3 - found;
-+ next_engine_per_entries_in_grp[2] = ioc_fm_pcd_cc_next_engine_params_t for the case that
-+ unit 1 - found; unit 3 - not found;
-+ next_engine_per_entries_in_grp[3] = ioc_fm_pcd_cc_next_engine_params_t for the case that
-+ unit 1 - found; unit 3 - found;
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_cc_grp_params_t {
-+ uint8_t num_of_distinction_units; /**< Up to 4 */
-+ uint8_t unit_ids [IOC_FM_PCD_MAX_NUM_OF_CC_UNITS];
-+ /**< Indexes of the units as defined in
-+ FM_PCD_NetEnvCharacteristicsSet() */
-+ ioc_fm_pcd_cc_next_engine_params_t next_engine_per_entries_in_grp[IOC_FM_PCD_MAX_NUM_OF_CC_ENTRIES_IN_GRP];
-+ /**< Maximum entries per group is 16 */
-+} ioc_fm_pcd_cc_grp_params_t;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining the CC tree groups
-+ (Must match struct t_FmPcdCcTreeParams defined in fm_pcd_ext.h)
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_cc_tree_params_t {
-+ void *net_env_id; /**< Id of the Network Environment as returned
-+ by FM_PCD_NetEnvCharacteristicsSet() */
-+ uint8_t num_of_groups; /**< Number of CC groups within the CC tree */
-+ ioc_fm_pcd_cc_grp_params_t fm_pcd_cc_group_params [IOC_FM_PCD_MAX_NUM_OF_CC_GROUPS];
-+ /**< Parameters for each group. */
-+ void *id; /**< Output parameter; Returns the tree Id to be used */
-+} ioc_fm_pcd_cc_tree_params_t;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining policer byte rate
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_plcr_byte_rate_mode_param_t {
-+ ioc_fm_pcd_plcr_frame_length_select frame_length_selection; /**< Frame length selection */
-+ ioc_fm_pcd_plcr_roll_back_frame_select roll_back_frame_selection; /**< relevant option only e_IOC_FM_PCD_PLCR_L2_FRM_LEN,
-+ e_IOC_FM_PCD_PLCR_FULL_FRM_LEN */
-+} ioc_fm_pcd_plcr_byte_rate_mode_param_t;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining the policer profile (based on
-+ RFC-2698 or RFC-4115 attributes).
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_plcr_non_passthrough_alg_param_t {
-+ ioc_fm_pcd_plcr_rate_mode rate_mode; /**< Byte / Packet */
-+ ioc_fm_pcd_plcr_byte_rate_mode_param_t byte_mode_param; /**< Valid for Byte NULL for Packet */
-+ uint32_t committed_info_rate; /**< KBits/Sec or Packets/Sec */
-+ uint32_t committed_burst_size; /**< KBits or Packets */
-+ uint32_t peak_or_excess_info_rate; /**< KBits/Sec or Packets/Sec */
-+ uint32_t peak_or_excess_burst_size; /**< KBits or Packets */
-+} ioc_fm_pcd_plcr_non_passthrough_alg_param_t;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining the next engine after policer
-+*//***************************************************************************/
-+typedef union ioc_fm_pcd_plcr_next_engine_params_u {
-+ ioc_fm_pcd_done_action action; /**< Action - when next engine is BMI (done) */
-+ void *p_profile; /**< Policer profile handle - used when next engine
-+ is PLCR, must be a SHARED profile */
-+ void *p_direct_scheme; /**< Direct scheme select - when next engine is Keygen */
-+} ioc_fm_pcd_plcr_next_engine_params_u;
-+
-+typedef struct ioc_fm_pcd_port_params_t {
-+ ioc_fm_port_type port_type; /**< Type of port for this profile */
-+ uint8_t port_id; /**< FM-Port id of port for this profile */
-+} ioc_fm_pcd_port_params_t;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining the policer profile entry
-+ (Must match struct t_FmPcdPlcrProfileParams defined in fm_pcd_ext.h)
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_plcr_profile_params_t {
-+ bool modify; /**< TRUE to change an existing profile */
-+ union {
-+ struct {
-+ ioc_fm_pcd_profile_type_selection profile_type; /**< Type of policer profile */
-+ ioc_fm_pcd_port_params_t *p_fm_port; /**< Relevant for per-port profiles only */
-+ uint16_t relative_profile_id; /**< Profile id - relative to shared group or to port */
-+ } new_params; /**< Use it when modify = FALSE */
-+ void *p_profile; /**< A handle to a profile - use it when modify=TRUE */
-+ } profile_select;
-+ ioc_fm_pcd_plcr_algorithm_selection alg_selection; /**< Profile Algorithm PASS_THROUGH, RFC_2698, RFC_4115 */
-+ ioc_fm_pcd_plcr_color_mode color_mode; /**< COLOR_BLIND, COLOR_AWARE */
-+
-+ union {
-+ ioc_fm_pcd_plcr_color dflt_color; /**< For Color-Blind Pass-Through mode; the policer will re-color
-+ any incoming packet with the default value. */
-+ ioc_fm_pcd_plcr_color override; /**< For Color-Aware modes; the profile response to a
-+ pre-color value of 2'b11. */
-+ } color;
-+
-+ ioc_fm_pcd_plcr_non_passthrough_alg_param_t non_passthrough_alg_param; /**< RFC2698 or RFC4115 parameters */
-+
-+ ioc_fm_pcd_engine next_engine_on_green; /**< Next engine for green-colored frames */
-+ ioc_fm_pcd_plcr_next_engine_params_u params_on_green; /**< Next engine parameters for green-colored frames */
-+
-+ ioc_fm_pcd_engine next_engine_on_yellow; /**< Next engine for yellow-colored frames */
-+ ioc_fm_pcd_plcr_next_engine_params_u params_on_yellow; /**< Next engine parameters for yellow-colored frames */
-+
-+ ioc_fm_pcd_engine next_engine_on_red; /**< Next engine for red-colored frames */
-+ ioc_fm_pcd_plcr_next_engine_params_u params_on_red; /**< Next engine parameters for red-colored frames */
-+
-+ bool trap_profile_on_flow_A; /**< Obsolete - do not use */
-+ bool trap_profile_on_flow_B; /**< Obsolete - do not use */
-+ bool trap_profile_on_flow_C; /**< Obsolete - do not use */
-+
-+ void *id; /**< output parameter; Returns the profile Id to be used */
-+} ioc_fm_pcd_plcr_profile_params_t;
-+
-+/**************************************************************************//**
-+ @Description A structure for modifying CC tree next engine
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_cc_tree_modify_next_engine_params_t {
-+ void *id; /**< CC tree Id to be used */
-+ uint8_t grp_indx; /**< A Group index in the tree */
-+ uint8_t indx; /**< Entry index in the group defined by grp_index */
-+ ioc_fm_pcd_cc_next_engine_params_t cc_next_engine_params;
-+ /**< Parameters for the next for the defined Key in the p_Key */
-+} ioc_fm_pcd_cc_tree_modify_next_engine_params_t;
-+
-+/**************************************************************************//**
-+ @Description A structure for modifying CC node next engine
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_cc_node_modify_next_engine_params_t {
-+ void *id; /**< CC node Id to be used */
-+ uint16_t key_indx; /**< Key index for Next Engine Params modifications;
-+ NOTE: This parameter is IGNORED for miss-key! */
-+ uint8_t key_size; /**< Key size of added key */
-+ ioc_fm_pcd_cc_next_engine_params_t cc_next_engine_params;
-+ /**< parameters for the next for the defined Key in the p_Key */
-+} ioc_fm_pcd_cc_node_modify_next_engine_params_t;
-+
-+/**************************************************************************//**
-+ @Description A structure for remove CC node key
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_cc_node_remove_key_params_t {
-+ void *id; /**< CC node Id to be used */
-+ uint16_t key_indx; /**< Key index for Next Engine Params modifications;
-+ NOTE: This parameter is IGNORED for miss-key! */
-+} ioc_fm_pcd_cc_node_remove_key_params_t;
-+
-+/**************************************************************************//**
-+ @Description A structure for modifying CC node key and next engine
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_cc_node_modify_key_and_next_engine_params_t {
-+ void *id; /**< CC node Id to be used */
-+ uint16_t key_indx; /**< Key index for Next Engine Params modifications;
-+ NOTE: This parameter is IGNORED for miss-key! */
-+ uint8_t key_size; /**< Key size of added key */
-+ ioc_fm_pcd_cc_key_params_t key_params; /**< it's array with numOfKeys entries each entry in
-+ the array of the type ioc_fm_pcd_cc_key_params_t */
-+} ioc_fm_pcd_cc_node_modify_key_and_next_engine_params_t;
-+
-+/**************************************************************************//**
-+ @Description A structure for modifying CC node key
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_cc_node_modify_key_params_t {
-+ void *id; /**< CC node Id to be used */
-+ uint16_t key_indx; /**< Key index for Next Engine Params modifications;
-+ NOTE: This parameter is IGNORED for miss-key! */
-+ uint8_t key_size; /**< Key size of added key */
-+ uint8_t *p_key; /**< Pointer to the key of the size defined in key_size */
-+ uint8_t *p_mask; /**< Pointer to the Mask per key of the size defined
-+ in keySize. p_Key and p_Mask (if defined) have to be
-+ of the same size as defined in the key_size */
-+} ioc_fm_pcd_cc_node_modify_key_params_t;
-+
-+/**************************************************************************//**
-+ @Description A structure with the arguments for the FM_PCD_HashTableRemoveKey ioctl() call
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_hash_table_remove_key_params_t {
-+ void *p_hash_tbl; /**< The id of the hash table */
-+ uint8_t key_size; /**< The size of the key to remove */
-+ uint8_t *p_key; /**< Pointer to the key to remove */
-+} ioc_fm_pcd_hash_table_remove_key_params_t;
-+
-+/**************************************************************************//**
-+ @Description Parameters for selecting a location for requested manipulation
-+*//***************************************************************************/
-+typedef struct ioc_fm_manip_hdr_info_t {
-+ ioc_net_header_type hdr; /**< Header selection */
-+ ioc_fm_pcd_hdr_index hdr_index; /**< Relevant only for MPLS, VLAN and tunneled IP. Otherwise should be cleared. */
-+ bool by_field; /**< TRUE if the location of manipulation is according to some field in the specific header*/
-+ ioc_fm_pcd_fields_u full_field; /**< Relevant only when by_field = TRUE: Extract field */
-+} ioc_fm_manip_hdr_info_t;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining header removal by header type
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_manip_hdr_rmv_by_hdr_params_t {
-+ ioc_fm_pcd_manip_hdr_rmv_by_hdr_type type; /**< Selection of header removal location */
-+ union {
-+#if ((DPAA_VERSION == 10) && defined(FM_CAPWAP_SUPPORT))
-+ struct {
-+ bool include;/**< If FALSE, remove until the specified header (not including the header);
-+ If TRUE, remove also the specified header. */
-+ ioc_fm_manip_hdr_info_t hdr_info;
-+ } from_start_by_hdr; /**< Relevant when type = e_IOC_FM_PCD_MANIP_RMV_BY_HDR_FROM_START */
-+#endif /* FM_CAPWAP_SUPPORT */
-+#if (DPAA_VERSION >= 11)
-+ ioc_fm_manip_hdr_info_t hdr_info; /**< Relevant when type = e_FM_PCD_MANIP_RMV_BY_HDR_FROM_START */
-+#endif /* (DPAA_VERSION >= 11) */
-+ ioc_fm_pcd_manip_hdr_rmv_specific_l2 specific_l2;/**< Relevant when type = e_IOC_FM_PCD_MANIP_BY_HDR_SPECIFIC_L2;
-+ Defines which L2 headers to remove. */
-+ } u;
-+} ioc_fm_pcd_manip_hdr_rmv_by_hdr_params_t;
-+
-+/**************************************************************************//**
-+ @Description Parameters for configuring IP fragmentation manipulation
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_manip_frag_ip_params_t {
-+ uint16_t size_for_fragmentation; /**< If length of the frame is greater than this value,
-+ IP fragmentation will be executed.*/
-+#if DPAA_VERSION == 10
-+ uint8_t scratch_bpid; /**< Absolute buffer pool id according to BM configuration.*/
-+#endif /* DPAA_VERSION == 10 */
-+ bool sg_bpid_en; /**< Enable a dedicated buffer pool id for the Scatter/Gather buffer allocation;
-+ If disabled, the Scatter/Gather buffer will be allocated from the same pool as the
-+ received frame's buffer. */
-+ uint8_t sg_bpid; /**< Scatter/Gather buffer pool id;
-+ This parameter is relevant when 'sg_bpid_en=TRUE';
-+ Same LIODN number is used for these buffers as for the received frames buffers, so buffers
-+ of this pool need to be allocated in the same memory area as the received buffers.
-+ If the received buffers arrive from different sources, the Scatter/Gather BP id should be
-+ mutual to all these sources. */
-+ ioc_fm_pcd_manip_dont_frag_action dont_frag_action; /**< Dont Fragment Action - If an IP packet is larger
-+ than MTU and its DF bit is set, then this field will
-+ determine the action to be taken.*/
-+} ioc_fm_pcd_manip_frag_ip_params_t;
-+
-+/**************************************************************************//**
-+ @Description Parameters for configuring IP reassembly manipulation.
-+
-+ This is a common structure for both IPv4 and IPv6 reassembly
-+ manipulation. For reassembly of both IPv4 and IPv6, make sure to
-+ set the 'hdr' field in ioc_fm_pcd_manip_reassem_params_t to IOC_HEADER_TYPE_IPv6.
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_manip_reassem_ip_params_t {
-+ uint8_t relative_scheme_id[2]; /**< Partition relative scheme id:
-+ relativeSchemeId[0] - Relative scheme ID for IPV4 Reassembly manipulation;
-+ relativeSchemeId[1] - Relative scheme ID for IPV6 Reassembly manipulation;
-+ NOTE: The following comment is relevant only for FMAN v2 devices:
-+ Relative scheme ID for IPv4/IPv6 Reassembly manipulation must be smaller than
-+ the user schemes id to ensure that the reassembly's schemes will be first match.
-+ The remaining schemes, if defined, should have higher relative scheme ID. */
-+#if DPAA_VERSION >= 11
-+ uint32_t non_consistent_sp_fqid; /**< In case that other fragments of the frame corresponds to different storage
-+ profile than the opening fragment (Non-Consistent-SP state)
-+ then one of two possible scenarios occurs:
-+ if 'nonConsistentSpFqid != 0', the reassembled frame will be enqueued to
-+ this fqid, otherwise a 'Non Consistent SP' bit will be set in the FD[status].*/
-+#else
-+ uint8_t sg_bpid; /**< Buffer pool id for the S/G frame created by the reassembly process */
-+#endif /* DPAA_VERSION >= 11 */
-+ uint8_t data_mem_id; /**< Memory partition ID for the IPR's external tables structure */
-+ uint16_t data_liodn_offset; /**< LIODN offset for access the IPR's external tables structure. */
-+ uint16_t min_frag_size[2]; /**< Minimum fragment size:
-+ minFragSize[0] - for ipv4, minFragSize[1] - for ipv6 */
-+ ioc_fm_pcd_manip_reassem_ways_number num_of_frames_per_hash_entry[2];
-+ /**< Number of frames per hash entry needed for reassembly process:
-+ numOfFramesPerHashEntry[0] - for ipv4 (max value is e_IOC_FM_PCD_MANIP_EIGHT_WAYS_HASH);
-+ numOfFramesPerHashEntry[1] - for ipv6 (max value is e_IOC_FM_PCD_MANIP_SIX_WAYS_HASH). */
-+ uint16_t max_num_frames_in_process;/**< Number of frames which can be processed by Reassembly in the same time;
-+ Must be power of 2;
-+ In the case numOfFramesPerHashEntry == e_IOC_FM_PCD_MANIP_FOUR_WAYS_HASH,
-+ maxNumFramesInProcess has to be in the range of 4 - 512;
-+ In the case numOfFramesPerHashEntry == e_IOC_FM_PCD_MANIP_EIGHT_WAYS_HASH,
-+ maxNumFramesInProcess has to be in the range of 8 - 2048. */
-+ ioc_fm_pcd_manip_reassem_time_out_mode time_out_mode; /**< Expiration delay initialized by Reassembly process */
-+ uint32_t fqid_for_time_out_frames;/**< FQID in which time out frames will enqueue during Time Out Process */
-+ uint32_t timeout_threshold_for_reassm_process;
-+ /**< Represents the time interval in microseconds which defines
-+ if opened frame (at least one fragment was processed but not all the fragments)is found as too old*/
-+} ioc_fm_pcd_manip_reassem_ip_params_t;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining IPSEC manipulation
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_manip_special_offload_ipsec_params_t {
-+ bool decryption; /**< TRUE if being used in decryption direction;
-+ FALSE if being used in encryption direction. */
-+ bool ecn_copy; /**< TRUE to copy the ECN bits from inner/outer to outer/inner
-+ (direction depends on the 'decryption' field). */
-+ bool dscp_copy; /**< TRUE to copy the DSCP bits from inner/outer to outer/inner
-+ (direction depends on the 'decryption' field). */
-+ bool variable_ip_hdr_len; /**< TRUE for supporting variable IP header length in decryption. */
-+ bool variable_ip_version; /**< TRUE for supporting both IP version on the same SA in encryption */
-+ uint8_t outer_ip_hdr_len; /**< If 'variable_ip_version == TRUE' than this field must be set to non-zero value;
-+ It is specifies the length of the outer IP header that was configured in the
-+ corresponding SA. */
-+ uint16_t arw_size; /**< if <> '0' then will perform ARW check for this SA;
-+ The value must be a multiplication of 16 */
-+ void *arw_addr; /**< if arwSize <> '0' then this field must be set to non-zero value;
-+ MUST be allocated from FMAN's MURAM that the post-sec op-port belong
-+ Must be 4B aligned. Required MURAM size is '(NEXT_POWER_OF_2(arwSize+32))/8+4' Bytes */
-+} ioc_fm_pcd_manip_special_offload_ipsec_params_t;
-+
-+#if (DPAA_VERSION >= 11)
-+/**************************************************************************//**
-+ @Description Parameters for configuring CAPWAP fragmentation manipulation
-+
-+ Restrictions:
-+ - Maximum number of fragments per frame is 16.
-+ - Transmit confirmation is not supported.
-+ - Fragmentation nodes must be set as the last PCD action (i.e. the
-+ corresponding CC node key must have next engine set to e_FM_PCD_DONE).
-+ - Only BMan buffers shall be used for frames to be fragmented.
-+ - NOTE: The following comment is relevant only for FMAN v3 devices: IPF
-+ does not support VSP. Therefore, on the same port where we have IPF we
-+ cannot support VSP.
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_manip_frag_capwap_params_t {
-+ uint16_t size_for_fragmentation; /**< If length of the frame is greater than this value,
-+ CAPWAP fragmentation will be executed.*/
-+ bool sg_bpid_en; /**< Enable a dedicated buffer pool id for the Scatter/Gather buffer allocation;
-+ If disabled, the Scatter/Gather buffer will be allocated from the same pool as the
-+ received frame's buffer. */
-+ uint8_t sg_bpid; /**< Scatter/Gather buffer pool id;
-+ This parameters is relevant when 'sgBpidEn=TRUE';
-+ Same LIODN number is used for these buffers as for the received frames buffers, so buffers
-+ of this pool need to be allocated in the same memory area as the received buffers.
-+ If the received buffers arrive from different sources, the Scatter/Gather BP id should be
-+ mutual to all these sources. */
-+ bool compress_mode_en; /**< CAPWAP Header Options Compress Enable mode;
-+ When this mode is enabled then only the first fragment include the CAPWAP header options
-+ field (if user provides it in the input frame) and all other fragments exclude the CAPWAP
-+ options field (CAPWAP header is updated accordingly).*/
-+} ioc_fm_pcd_manip_frag_capwap_params_t;
-+
-+/**************************************************************************//**
-+ @Description Parameters for configuring CAPWAP reassembly manipulation.
-+
-+ Restrictions:
-+ - Application must define one scheme to catch the reassembled frames.
-+ - Maximum number of fragments per frame is 16.
-+
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_manip_reassem_capwap_params_t {
-+ uint8_t relative_scheme_id; /**< Partition relative scheme id;
-+ NOTE: this id must be smaller than the user schemes id to ensure that the reassembly scheme will be first match;
-+ Rest schemes, if defined, should have higher relative scheme ID. */
-+ uint8_t data_mem_id; /**< Memory partition ID for the IPR's external tables structure */
-+ uint16_t data_liodn_offset; /**< LIODN offset for access the IPR's external tables structure. */
-+ uint16_t max_reassembled_frame_length;/**< The maximum CAPWAP reassembled frame length in bytes;
-+ If maxReassembledFrameLength == 0, any successful reassembled frame length is
-+ considered as a valid length;
-+ if maxReassembledFrameLength > 0, a successful reassembled frame which its length
-+ exceeds this value is considered as an error frame (FD status[CRE] bit is set). */
-+ ioc_fm_pcd_manip_reassem_ways_number num_of_frames_per_hash_entry;
-+ /**< Number of frames per hash entry needed for reassembly process */
-+ uint16_t max_num_frames_in_process; /**< Number of frames which can be processed by reassembly in the same time;
-+ Must be power of 2;
-+ In the case numOfFramesPerHashEntry == e_FM_PCD_MANIP_FOUR_WAYS_HASH,
-+ maxNumFramesInProcess has to be in the range of 4 - 512;
-+ In the case numOfFramesPerHashEntry == e_FM_PCD_MANIP_EIGHT_WAYS_HASH,
-+ maxNumFramesInProcess has to be in the range of 8 - 2048. */
-+ ioc_fm_pcd_manip_reassem_time_out_mode time_out_mode; /**< Expiration delay initialized by Reassembly process */
-+ uint32_t fqid_for_time_out_frames; /**< FQID in which time out frames will enqueue during Time Out Process;
-+ Recommended value for this field is 0; in this way timed-out frames will be discarded */
-+ uint32_t timeout_threshold_for_reassm_process;
-+ /**< Represents the time interval in microseconds which defines
-+ if opened frame (at least one fragment was processed but not all the fragments)is found as too old*/
-+} ioc_fm_pcd_manip_reassem_capwap_params_t;
-+
-+/**************************************************************************//**
-+ @Description structure for defining CAPWAP manipulation
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_manip_special_offload_capwap_params_t {
-+ bool dtls; /**< TRUE if continue to SEC DTLS encryption */
-+ ioc_fm_pcd_manip_hdr_qos_src qos_src; /**< TODO */
-+} ioc_fm_pcd_manip_special_offload_capwap_params_t;
-+
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining special offload manipulation
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_manip_special_offload_params_t {
-+ ioc_fm_pcd_manip_special_offload_type type; /**< Type of special offload manipulation */
-+ union
-+ {
-+ ioc_fm_pcd_manip_special_offload_ipsec_params_t ipsec; /**< Parameters for IPSec; Relevant when
-+ type = e_IOC_FM_PCD_MANIP_SPECIAL_OFFLOAD_IPSEC */
-+
-+#if (DPAA_VERSION >= 11)
-+ ioc_fm_pcd_manip_special_offload_capwap_params_t capwap; /**< Parameters for CAPWAP; Relevant when
-+ type = e_FM_PCD_MANIP_SPECIAL_OFFLOAD_CAPWAP */
-+#endif /* (DPAA_VERSION >= 11) */
-+ } u;
-+} ioc_fm_pcd_manip_special_offload_params_t;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining generic removal manipulation
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_manip_hdr_rmv_generic_params_t {
-+ uint8_t offset; /**< Offset from beginning of header to the start
-+ location of the removal */
-+ uint8_t size; /**< Size of removed section */
-+} ioc_fm_pcd_manip_hdr_rmv_generic_params_t;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining insertion manipulation
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_manip_hdr_insrt_t {
-+ uint8_t size; /**< size of inserted section */
-+ uint8_t *p_data; /**< data to be inserted */
-+} ioc_fm_pcd_manip_hdr_insrt_t;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining generic insertion manipulation
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_manip_hdr_insrt_generic_params_t {
-+ uint8_t offset; /**< Offset from beginning of header to the start
-+ location of the insertion */
-+ uint8_t size; /**< Size of inserted section */
-+ bool replace; /**< TRUE to override (replace) existing data at
-+ 'offset', FALSE to insert */
-+ uint8_t *p_data; /**< Pointer to data to be inserted */
-+} ioc_fm_pcd_manip_hdr_insrt_generic_params_t;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining header manipulation VLAN DSCP To Vpri translation
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_manip_hdr_field_update_vlan_dscp_to_vpri_t {
-+ uint8_t dscp_to_vpri_table[IOC_FM_PCD_MANIP_DSCP_TO_VLAN_TRANS];
-+ /**< A table of VPri values for each DSCP value;
-+ The index is the D_SCP value (0-0x3F) and the
-+ value is the corresponding VPRI (0-15). */
-+ uint8_t vpri_def_val; /**< 0-7, Relevant only if if update_type =
-+ e_IOC_FM_PCD_MANIP_HDR_FIELD_UPDATE_DSCP_TO_VLAN,
-+ this field is the Q Tag default value if the
-+ IP header is not found. */
-+} ioc_fm_pcd_manip_hdr_field_update_vlan_dscp_to_vpri_t;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining header manipulation VLAN fields updates
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_manip_hdr_field_update_vlan_t {
-+ ioc_fm_pcd_manip_hdr_field_update_vlan update_type; /**< Selects VLAN update type */
-+ union {
-+ uint8_t vpri; /**< 0-7, Relevant only if If update_type =
-+ e_IOC_FM_PCD_MANIP_HDR_FIELD_UPDATE_VLAN_PRI, this
-+ is the new VLAN pri. */
-+ ioc_fm_pcd_manip_hdr_field_update_vlan_dscp_to_vpri_t dscp_to_vpri;
-+ /**< Parameters structure, Relevant only if update_type =
-+ e_IOC_FM_PCD_MANIP_HDR_FIELD_UPDATE_DSCP_TO_VLAN. */
-+ } u;
-+} ioc_fm_pcd_manip_hdr_field_update_vlan_t;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining header manipulation IPV4 fields updates
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_manip_hdr_field_update_ipv4_t {
-+ ioc_ipv4_hdr_manip_update_flags_t valid_updates; /**< ORed flag, selecting the required updates */
-+ uint8_t tos; /**< 8 bit New TOS; Relevant if valid_updates contains
-+ IOC_HDR_MANIP_IPV4_TOS */
-+ uint16_t id; /**< 16 bit New IP ID; Relevant only if valid_updates
-+ contains IOC_HDR_MANIP_IPV4_ID */
-+ uint32_t src; /**< 32 bit New IP SRC; Relevant only if valid_updates
-+ contains IOC_HDR_MANIP_IPV4_SRC */
-+ uint32_t dst; /**< 32 bit New IP DST; Relevant only if valid_updates
-+ contains IOC_HDR_MANIP_IPV4_DST */
-+} ioc_fm_pcd_manip_hdr_field_update_ipv4_t;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining header manipulation IPV6 fields updates
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_manip_hdr_field_update_ipv6_t {
-+ ioc_ipv6_hdr_manip_update_flags_t valid_updates; /**< ORed flag, selecting the required updates */
-+ uint8_t traffic_class; /**< 8 bit New Traffic Class; Relevant if valid_updates contains
-+ IOC_HDR_MANIP_IPV6_TC */
-+ uint8_t src[IOC_NET_HEADER_FIELD_IPv6_ADDR_SIZE];
-+ /**< 16 byte new IP SRC; Relevant only if valid_updates
-+ contains IOC_HDR_MANIP_IPV6_SRC */
-+ uint8_t dst[IOC_NET_HEADER_FIELD_IPv6_ADDR_SIZE];
-+ /**< 16 byte new IP DST; Relevant only if valid_updates
-+ contains IOC_HDR_MANIP_IPV6_DST */
-+} ioc_fm_pcd_manip_hdr_field_update_ipv6_t;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining header manipulation TCP/UDP fields updates
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_manip_hdr_field_update_tcp_udp_t {
-+ ioc_tcp_udp_hdr_manip_update_flags_t valid_updates; /**< ORed flag, selecting the required updates */
-+ uint16_t src; /**< 16 bit New TCP/UDP SRC; Relevant only if valid_updates
-+ contains IOC_HDR_MANIP_TCP_UDP_SRC */
-+ uint16_t dst; /**< 16 bit New TCP/UDP DST; Relevant only if valid_updates
-+ contains IOC_HDR_MANIP_TCP_UDP_DST */
-+} ioc_fm_pcd_manip_hdr_field_update_tcp_udp_t;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining header manipulation fields updates
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_manip_hdr_field_update_params_t {
-+ ioc_fm_pcd_manip_hdr_field_update_type type; /**< Type of header field update manipulation */
-+ union {
-+ ioc_fm_pcd_manip_hdr_field_update_vlan_t vlan; /**< Parameters for VLAN update. Relevant when
-+ type = e_IOC_FM_PCD_MANIP_HDR_FIELD_UPDATE_VLAN */
-+ ioc_fm_pcd_manip_hdr_field_update_ipv4_t ipv4; /**< Parameters for IPv4 update. Relevant when
-+ type = e_IOC_FM_PCD_MANIP_HDR_FIELD_UPDATE_IPV4 */
-+ ioc_fm_pcd_manip_hdr_field_update_ipv6_t ipv6; /**< Parameters for IPv6 update. Relevant when
-+ type = e_IOC_FM_PCD_MANIP_HDR_FIELD_UPDATE_IPV6 */
-+ ioc_fm_pcd_manip_hdr_field_update_tcp_udp_t tcp_udp;/**< Parameters for TCP/UDP update. Relevant when
-+ type = e_IOC_FM_PCD_MANIP_HDR_FIELD_UPDATE_TCP_UDP */
-+ } u;
-+} ioc_fm_pcd_manip_hdr_field_update_params_t;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining custom header manipulation for IP replacement
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_manip_hdr_custom_ip_hdr_replace_t {
-+ ioc_fm_pcd_manip_hdr_custom_ip_replace replace_type; /**< Selects replace update type */
-+ bool dec_ttl_hl; /**< Decrement TTL (IPV4) or Hop limit (IPV6) by 1 */
-+ bool update_ipv4_id; /**< Relevant when replace_type =
-+ e_IOC_FM_PCD_MANIP_HDR_CUSTOM_REPLACE_IPV6_BY_IPV4 */
-+ uint16_t id; /**< 16 bit New IP ID; Relevant only if
-+ update_ipv4_id = TRUE */
-+ uint8_t hdr_size; /**< The size of the new IP header */
-+ uint8_t hdr[IOC_FM_PCD_MANIP_MAX_HDR_SIZE];
-+ /**< The new IP header */
-+} ioc_fm_pcd_manip_hdr_custom_ip_hdr_replace_t;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining custom header manipulation
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_manip_hdr_custom_params_t {
-+ ioc_fm_pcd_manip_hdr_custom_type type; /**< Type of header field update manipulation */
-+ union {
-+ ioc_fm_pcd_manip_hdr_custom_ip_hdr_replace_t ip_hdr_replace;
-+ /**< Parameters IP header replacement */
-+ } u;
-+} ioc_fm_pcd_manip_hdr_custom_params_t;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining specific L2 insertion manipulation
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_manip_hdr_insrt_specific_l2_params_t {
-+ ioc_fm_pcd_manip_hdr_insrt_specific_l2 specific_l2; /**< Selects which L2 headers to insert */
-+ bool update; /**< TRUE to update MPLS header */
-+ uint8_t size; /**< size of inserted section */
-+ uint8_t *p_data; /**< data to be inserted */
-+} ioc_fm_pcd_manip_hdr_insrt_specific_l2_params_t;
-+
-+#if (DPAA_VERSION >= 11)
-+/**************************************************************************//**
-+ @Description Parameters for defining IP insertion manipulation
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_manip_hdr_insrt_ip_params_t {
-+ bool calc_l4_checksum; /**< Calculate L4 checksum. */
-+ ioc_fm_pcd_manip_hdr_qos_mapping_mode mapping_mode; /**< TODO */
-+ uint8_t last_pid_offset; /**< the offset of the last Protocol within
-+ the inserted header */
-+ uint16_t id; /**< 16 bit New IP ID */
-+ bool dont_frag_overwrite;
-+ /**< IPv4 only. DF is overwritten with the hash-result next-to-last byte.
-+ * This byte is configured to be overwritten when RPD is set. */
-+ uint8_t last_dst_offset;
-+ /**< IPv6 only. if routing extension exist, user should set the offset of the destination address
-+ * in order to calculate UDP checksum pseudo header;
-+ * Otherwise set it to '0'. */
-+ ioc_fm_pcd_manip_hdr_insrt_t insrt; /**< size and data to be inserted. */
-+} ioc_fm_pcd_manip_hdr_insrt_ip_params_t;
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining header insertion manipulation by header type
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_manip_hdr_insrt_by_hdr_params_t {
-+ ioc_fm_pcd_manip_hdr_insrt_by_hdr_type type; /**< Selects manipulation type */
-+ union {
-+ ioc_fm_pcd_manip_hdr_insrt_specific_l2_params_t specific_l2_params;
-+ /**< Used when type = e_IOC_FM_PCD_MANIP_INSRT_BY_HDR_SPECIFIC_L2:
-+ Selects which L2 headers to remove */
-+#if (DPAA_VERSION >= 11)
-+ ioc_fm_pcd_manip_hdr_insrt_ip_params_t ip_params; /**< Used when type = e_FM_PCD_MANIP_INSRT_BY_HDR_IP */
-+ ioc_fm_pcd_manip_hdr_insrt_t insrt; /**< Used when type is one of e_FM_PCD_MANIP_INSRT_BY_HDR_UDP,
-+ e_FM_PCD_MANIP_INSRT_BY_HDR_UDP_LITE, or
-+ e_FM_PCD_MANIP_INSRT_BY_HDR_CAPWAP */
-+#endif /* (DPAA_VERSION >= 11) */
-+ } u;
-+} ioc_fm_pcd_manip_hdr_insrt_by_hdr_params_t;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining header insertion manipulation
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_manip_hdr_insrt_params_t {
-+ ioc_fm_pcd_manip_hdr_insrt_type type; /**< Type of insertion manipulation */
-+ union {
-+ ioc_fm_pcd_manip_hdr_insrt_by_hdr_params_t by_hdr; /**< Parameters for defining header insertion manipulation by header type,
-+ relevant if 'type' = e_IOC_FM_PCD_MANIP_INSRT_BY_HDR */
-+ ioc_fm_pcd_manip_hdr_insrt_generic_params_t generic;/**< Parameters for defining generic header insertion manipulation,
-+ relevant if type = e_IOC_FM_PCD_MANIP_INSRT_GENERIC */
-+#if (defined(FM_CAPWAP_SUPPORT) && (DPAA_VERSION == 10))
-+ ioc_fm_pcd_manip_hdr_insrt_by_template_params_t by_template;
-+ /**< Parameters for defining header insertion manipulation by template,
-+ relevant if 'type' = e_IOC_FM_PCD_MANIP_INSRT_BY_TEMPLATE */
-+#endif /* FM_CAPWAP_SUPPORT */
-+ } u;
-+} ioc_fm_pcd_manip_hdr_insrt_params_t;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining header removal manipulation
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_manip_hdr_rmv_params_t {
-+ ioc_fm_pcd_manip_hdr_rmv_type type; /**< Type of header removal manipulation */
-+ union {
-+ ioc_fm_pcd_manip_hdr_rmv_by_hdr_params_t by_hdr; /**< Parameters for defining header removal manipulation by header type,
-+ relevant if type = e_IOC_FM_PCD_MANIP_RMV_BY_HDR */
-+ ioc_fm_pcd_manip_hdr_rmv_generic_params_t generic; /**< Parameters for defining generic header removal manipulation,
-+ relevant if type = e_IOC_FM_PCD_MANIP_RMV_GENERIC */
-+ } u;
-+} ioc_fm_pcd_manip_hdr_rmv_params_t;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining header manipulation node
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_manip_hdr_params_t {
-+ bool rmv; /**< TRUE, to define removal manipulation */
-+ ioc_fm_pcd_manip_hdr_rmv_params_t rmv_params; /**< Parameters for removal manipulation, relevant if 'rmv' = TRUE */
-+
-+ bool insrt; /**< TRUE, to define insertion manipulation */
-+ ioc_fm_pcd_manip_hdr_insrt_params_t insrt_params; /**< Parameters for insertion manipulation, relevant if 'insrt' = TRUE */
-+
-+ bool field_update; /**< TRUE, to define field update manipulation */
-+ ioc_fm_pcd_manip_hdr_field_update_params_t field_update_params; /**< Parameters for field update manipulation, relevant if 'fieldUpdate' = TRUE */
-+
-+ bool custom; /**< TRUE, to define custom manipulation */
-+ ioc_fm_pcd_manip_hdr_custom_params_t custom_params; /**< Parameters for custom manipulation, relevant if 'custom' = TRUE */
-+
-+ bool dont_parse_after_manip;/**< FALSE to activate the parser a second time after
-+ completing the manipulation on the frame */
-+} ioc_fm_pcd_manip_hdr_params_t;
-+
-+
-+/**************************************************************************//**
-+ @Description structure for defining fragmentation manipulation
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_manip_frag_params_t {
-+ ioc_net_header_type hdr; /**< Header selection */
-+ union {
-+#if (DPAA_VERSION >= 11)
-+ ioc_fm_pcd_manip_frag_capwap_params_t capwap_frag; /**< Parameters for defining CAPWAP fragmentation,
-+ relevant if 'hdr' = HEADER_TYPE_CAPWAP */
-+#endif /* (DPAA_VERSION >= 11) */
-+ ioc_fm_pcd_manip_frag_ip_params_t ip_frag; /**< Parameters for defining IP fragmentation,
-+ relevant if 'hdr' = HEADER_TYPE_Ipv4 or HEADER_TYPE_Ipv6 */
-+ } u;
-+} ioc_fm_pcd_manip_frag_params_t;
-+
-+/**************************************************************************//**
-+ @Description structure for defining reassemble manipulation
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_manip_reassem_params_t {
-+ ioc_net_header_type hdr; /**< Header selection */
-+ union {
-+#if (DPAA_VERSION >= 11)
-+ ioc_fm_pcd_manip_reassem_capwap_params_t capwap_reassem; /**< Parameters for defining CAPWAP reassembly,
-+ relevant if 'hdr' = HEADER_TYPE_CAPWAP */
-+#endif /* (DPAA_VERSION >= 11) */
-+ ioc_fm_pcd_manip_reassem_ip_params_t ip_reassem; /**< Parameters for defining IP reassembly,
-+ relevant if 'hdr' = HEADER_TYPE_Ipv4 or HEADER_TYPE_Ipv6 */
-+ } u;
-+} ioc_fm_pcd_manip_reassem_params_t;
-+
-+/**************************************************************************//**
-+ @Description Parameters for defining a manipulation node
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_manip_params_t {
-+ ioc_fm_pcd_manip_type type; /**< Selects type of manipulation node */
-+ union {
-+ ioc_fm_pcd_manip_hdr_params_t hdr; /**< Parameters for defining header manipulation node */
-+ ioc_fm_pcd_manip_reassem_params_t reassem;/**< Parameters for defining reassembly manipulation node */
-+ ioc_fm_pcd_manip_frag_params_t frag; /**< Parameters for defining fragmentation manipulation node */
-+ ioc_fm_pcd_manip_special_offload_params_t special_offload;/**< Parameters for defining special offload manipulation node */
-+ } u;
-+ void *p_next_manip;/**< Handle to another (previously defined) manipulation node;
-+ Allows concatenation of manipulation actions
-+ This parameter is optional and may be NULL. */
-+#if (defined(FM_CAPWAP_SUPPORT) && (DPAA_VERSION == 10))
-+ bool frag_or_reasm;/**< TRUE, if defined fragmentation/reassembly manipulation */
-+ ioc_fm_pcd_manip_frag_or_reasm_params_t frag_or_reasm_params;/**< Parameters for fragmentation/reassembly manipulation,
-+ relevant if frag_or_reasm = TRUE */
-+#endif /* FM_CAPWAP_SUPPORT */
-+ void *id;
-+} ioc_fm_pcd_manip_params_t;
-+
-+/**************************************************************************//**
-+ @Description Structure for retrieving IP reassembly statistics
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_manip_reassem_ip_stats_t {
-+ /* common counters for both IPv4 and IPv6 */
-+ uint32_t timeout; /**< Counts the number of TimeOut occurrences */
-+ uint32_t rfd_pool_busy; /**< Counts the number of failed attempts to allocate
-+ a Reassembly Frame Descriptor */
-+ uint32_t internal_buffer_busy; /**< Counts the number of times an internal buffer busy occurred */
-+ uint32_t external_buffer_busy; /**< Counts the number of times external buffer busy occurred */
-+ uint32_t sg_fragments; /**< Counts the number of Scatter/Gather fragments */
-+ uint32_t dma_semaphore_depletion; /**< Counts the number of failed attempts to allocate a DMA semaphore */
-+#if (DPAA_VERSION >= 11)
-+ uint32_t non_consistent_sp; /**< Counts the number of Non Consistent Storage Profile events for
-+ successfully reassembled frames */
-+#endif /* (DPAA_VERSION >= 11) */
-+struct {
-+ uint32_t successfully_reassembled; /**< Counts the number of successfully reassembled frames */
-+ uint32_t valid_fragments; /**< Counts the total number of valid fragments that
-+ have been processed for all frames */
-+ uint32_t processed_fragments; /**< Counts the number of processed fragments
-+ (valid and error fragments) for all frames */
-+ uint32_t malformed_fragments; /**< Counts the number of malformed fragments processed for all frames */
-+ uint32_t discarded_fragments; /**< Counts the number of fragments discarded by the reassembly process */
-+ uint32_t auto_learn_busy; /**< Counts the number of times a busy condition occurs when attempting
-+ to access an IP-Reassembly Automatic Learning Hash set */
-+ uint32_t more_than16fragments; /**< Counts the fragment occurrences in which the number of fragments-per-frame
-+ exceeds 16 */
-+ } specific_hdr_statistics[2]; /**< slot '0' is for IPv4, slot '1' is for IPv6 */
-+} ioc_fm_pcd_manip_reassem_ip_stats_t;
-+
-+/**************************************************************************//**
-+ @Description Structure for retrieving IP fragmentation statistics
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_manip_frag_ip_stats_t {
-+ uint32_t total_frames; /**< Number of frames that passed through the manipulation node */
-+ uint32_t fragmented_frames; /**< Number of frames that were fragmented */
-+ uint32_t generated_fragments; /**< Number of fragments that were generated */
-+} ioc_fm_pcd_manip_frag_ip_stats_t;
-+
-+#if (DPAA_VERSION >= 11)
-+/**************************************************************************//**
-+ @Description Structure for retrieving CAPWAP reassembly statistics
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_manip_reassem_capwap_stats_t {
-+ uint32_t timeout; /**< Counts the number of timeout occurrences */
-+ uint32_t rfd_pool_busy; /**< Counts the number of failed attempts to allocate
-+ a Reassembly Frame Descriptor */
-+ uint32_t internal_buffer_busy; /**< Counts the number of times an internal buffer busy occurred */
-+ uint32_t external_buffer_busy; /**< Counts the number of times external buffer busy occurred */
-+ uint32_t sg_fragments; /**< Counts the number of Scatter/Gather fragments */
-+ uint32_t dma_semaphore_depletion; /**< Counts the number of failed attempts to allocate a DMA semaphore */
-+ uint32_t successfully_reassembled; /**< Counts the number of successfully reassembled frames */
-+ uint32_t valid_fragments; /**< Counts the total number of valid fragments that
-+ have been processed for all frames */
-+ uint32_t processed_fragments; /**< Counts the number of processed fragments
-+ (valid and error fragments) for all frames */
-+ uint32_t malformed_fragments; /**< Counts the number of malformed fragments processed for all frames */
-+ uint32_t autoLearn_busy; /**< Counts the number of times a busy condition occurs when attempting
-+ to access an Reassembly Automatic Learning Hash set */
-+ uint32_t discarded_fragments; /**< Counts the number of fragments discarded by the reassembly process */
-+ uint32_t more_than16fragments; /**< Counts the fragment occurrences in which the number of fragments-per-frame
-+ exceeds 16 */
-+ uint32_t exceed_max_reassembly_frame_len;/**< ounts the number of times that a successful reassembled frame
-+ length exceeds MaxReassembledFrameLength value */
-+} ioc_fm_pcd_manip_reassem_capwap_stats_t;
-+
-+/**************************************************************************//**
-+ @Description Structure for retrieving CAPWAP fragmentation statistics
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_manip_frag_capwap_stats_t {
-+ uint32_t total_frames; /**< Number of frames that passed through the manipulation node */
-+ uint32_t fragmented_frames; /**< Number of frames that were fragmented */
-+ uint32_t generated_fragments; /**< Number of fragments that were generated */
-+#if (defined(DEBUG_ERRORS) && (DEBUG_ERRORS > 0))
-+ uint8_t sg_allocation_failure; /**< Number of allocation failure of s/g buffers */
-+#endif /* (defined(DEBUG_ERRORS) && (DEBUG_ERRORS > 0)) */
-+} ioc_fm_pcd_manip_frag_capwap_stats_t;
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+/**************************************************************************//**
-+ @Description Structure for retrieving reassembly statistics
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_manip_reassem_stats_t {
-+ union {
-+ ioc_fm_pcd_manip_reassem_ip_stats_t ip_reassem; /**< Structure for IP reassembly statistics */
-+#if (DPAA_VERSION >= 11)
-+ ioc_fm_pcd_manip_reassem_capwap_stats_t capwap_reassem; /**< Structure for CAPWAP reassembly statistics */
-+#endif /* (DPAA_VERSION >= 11) */
-+ } u;
-+} ioc_fm_pcd_manip_reassem_stats_t;
-+
-+/**************************************************************************//**
-+ @Description structure for retrieving fragmentation statistics
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_manip_frag_stats_t {
-+ union {
-+ ioc_fm_pcd_manip_frag_ip_stats_t ip_frag; /**< Structure for IP fragmentation statistics */
-+#if (DPAA_VERSION >= 11)
-+ ioc_fm_pcd_manip_frag_capwap_stats_t capwap_frag; /**< Structure for CAPWAP fragmentation statistics */
-+#endif /* (DPAA_VERSION >= 11) */
-+ } u;
-+} ioc_fm_pcd_manip_frag_stats_t;
-+
-+/**************************************************************************//**
-+ @Description structure for defining manipulation statistics
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_manip_stats_t {
-+ union {
-+ ioc_fm_pcd_manip_reassem_stats_t reassem; /**< Structure for reassembly statistics */
-+ ioc_fm_pcd_manip_frag_stats_t frag; /**< Structure for fragmentation statistics */
-+ } u;
-+} ioc_fm_pcd_manip_stats_t;
-+
-+/**************************************************************************//**
-+ @Description Parameters for acquiring manipulation statistics
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_manip_get_stats_t {
-+ void *id;
-+ ioc_fm_pcd_manip_stats_t stats;
-+} ioc_fm_pcd_manip_get_stats_t;
-+
-+#if DPAA_VERSION >= 11
-+/**************************************************************************//**
-+ @Description Parameters for defining frame replicator group and its members
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_frm_replic_group_params_t {
-+ uint8_t max_num_of_entries; /**< Maximal number of members in the group - must be at least two */
-+ uint8_t num_of_entries; /**< Number of members in the group - must be at least 1 */
-+ ioc_fm_pcd_cc_next_engine_params_t next_engine_params[IOC_FM_PCD_FRM_REPLIC_MAX_NUM_OF_ENTRIES];
-+ /**< Array of members' parameters */
-+ void *id;
-+} ioc_fm_pcd_frm_replic_group_params_t;
-+
-+typedef struct ioc_fm_pcd_frm_replic_member_t {
-+ void *h_replic_group;
-+ uint16_t member_index;
-+} ioc_fm_pcd_frm_replic_member_t;
-+
-+typedef struct ioc_fm_pcd_frm_replic_member_params_t {
-+ ioc_fm_pcd_frm_replic_member_t member;
-+ ioc_fm_pcd_cc_next_engine_params_t next_engine_params;
-+} ioc_fm_pcd_frm_replic_member_params_t;
-+#endif /* DPAA_VERSION >= 11 */
-+
-+
-+typedef struct ioc_fm_pcd_cc_key_statistics_t {
-+ uint32_t byte_count; /**< This counter reflects byte count of frames that
-+ were matched by this key. */
-+ uint32_t frame_count; /**< This counter reflects count of frames that
-+ were matched by this key. */
-+#if (DPAA_VERSION >= 11)
-+ uint32_t frame_length_range_count[IOC_FM_PCD_CC_STATS_MAX_NUM_OF_FLR];
-+ /**< These counters reflect how many frames matched
-+ this key in 'RMON' statistics mode:
-+ Each counter holds the number of frames of a
-+ specific frames length range, according to the
-+ ranges provided at initialization. */
-+#endif /* (DPAA_VERSION >= 11) */
-+} ioc_fm_pcd_cc_key_statistics_t;
-+
-+
-+typedef struct ioc_fm_pcd_cc_tbl_get_stats_t {
-+ void *id;
-+ uint16_t key_index;
-+ ioc_fm_pcd_cc_key_statistics_t statistics;
-+} ioc_fm_pcd_cc_tbl_get_stats_t;
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_MatchTableGetKeyStatistics
-+
-+ @Description This routine may be used to get statistics counters of specific key
-+ in a CC Node.
-+
-+ If 'e_FM_PCD_CC_STATS_MODE_FRAME' and
-+ 'e_FM_PCD_CC_STATS_MODE_BYTE_AND_FRAME' were set for this node,
-+ these counters reflect how many frames passed that were matched
-+ this key; The total frames count will be returned in the counter
-+ of the first range (as only one frame length range was defined).
-+ If 'e_FM_PCD_CC_STATS_MODE_RMON' was set for this node, the total
-+ frame count will be separated to frame length counters, based on
-+ provided frame length ranges.
-+
-+ @Param[in] h_CcNode A handle to the node
-+ @Param[in] keyIndex Key index for adding
-+ @Param[out] p_KeyStatistics Key statistics counters
-+
-+ @Return The specific key statistics.
-+
-+ @Cautions Allowed only following FM_PCD_MatchTableSet().
-+*//***************************************************************************/
-+
-+#if defined(CONFIG_COMPAT)
-+#define FM_PCD_IOC_MATCH_TABLE_GET_KEY_STAT_COMPAT _IOWR(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(12), ioc_compat_fm_pcd_cc_tbl_get_stats_t)
-+#endif
-+#define FM_PCD_IOC_MATCH_TABLE_GET_KEY_STAT _IOWR(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(12), ioc_fm_pcd_cc_tbl_get_stats_t)
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_MatchTableGetMissStatistics
-+
-+ @Description This routine may be used to get statistics counters of miss entry
-+ in a CC Node.
-+
-+ If 'e_FM_PCD_CC_STATS_MODE_FRAME' and
-+ 'e_FM_PCD_CC_STATS_MODE_BYTE_AND_FRAME' were set for this node,
-+ these counters reflect how many frames were not matched to any
-+ existing key and therefore passed through the miss entry; The
-+ total frames count will be returned in the counter of the
-+ first range (as only one frame length range was defined).
-+
-+ @Param[in] h_CcNode A handle to the node
-+ @Param[out] p_MissStatistics Statistics counters for 'miss'
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_MatchTableSet().
-+*//***************************************************************************/
-+
-+#if defined(CONFIG_COMPAT)
-+#define FM_PCD_IOC_MATCH_TABLE_GET_MISS_STAT_COMPAT _IOWR(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(13), ioc_compat_fm_pcd_cc_tbl_get_stats_t)
-+#endif
-+#define FM_PCD_IOC_MATCH_TABLE_GET_MISS_STAT _IOWR(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(13), ioc_fm_pcd_cc_tbl_get_stats_t)
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_HashTableGetMissStatistics
-+
-+ @Description This routine may be used to get statistics counters of 'miss'
-+ entry of the a hash table.
-+
-+ If 'e_FM_PCD_CC_STATS_MODE_FRAME' and
-+ 'e_FM_PCD_CC_STATS_MODE_BYTE_AND_FRAME' were set for this node,
-+ these counters reflect how many frames were not matched to any
-+ existing key and therefore passed through the miss entry;
-+
-+ @Param[in] h_HashTbl A handle to a hash table
-+ @Param[out] p_MissStatistics Statistics counters for 'miss'
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_HashTableSet().
-+*//***************************************************************************/
-+
-+#if defined(CONFIG_COMPAT)
-+#define FM_PCD_IOC_HASH_TABLE_GET_MISS_STAT_COMPAT _IOWR(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(14), ioc_compat_fm_pcd_cc_tbl_get_stats_t)
-+#endif
-+#define FM_PCD_IOC_HASH_TABLE_GET_MISS_STAT _IOWR(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(14), ioc_fm_pcd_cc_tbl_get_stats_t)
-+
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_NetEnvCharacteristicsSet
-+
-+ @Description Define a set of Network Environment Characteristics.
-+
-+ When setting an environment it is important to understand its
-+ application. It is not meant to describe the flows that will run
-+ on the ports using this environment, but what the user means TO DO
-+ with the PCD mechanisms in order to parse-classify-distribute those
-+ frames.
-+ By specifying a distinction unit, the user means it would use that option
-+ for distinction between frames at either a KeyGen scheme or a coarse
-+ classification action descriptor. Using interchangeable headers to define a
-+ unit means that the user is indifferent to which of the interchangeable
-+ headers is present in the frame, and wants the distinction to be based
-+ on the presence of either one of them.
-+
-+ Depending on context, there are limitations to the use of environments. A
-+ port using the PCD functionality is bound to an environment. Some or even
-+ all ports may share an environment but also an environment per port is
-+ possible. When initializing a scheme, a classification plan group (see below),
-+ or a coarse classification tree, one of the initialized environments must be
-+ stated and related to. When a port is bound to a scheme, a classification
-+ plan group, or a coarse classification tree, it MUST be bound to the same
-+ environment.
-+
-+ The different PCD modules, may relate (for flows definition) ONLY on
-+ distinction units as defined by their environment. When initializing a
-+ scheme for example, it may not choose to select IPV4 as a match for
-+ recognizing flows unless it was defined in the relating environment. In
-+ fact, to guide the user through the configuration of the PCD, each module's
-+ characterization in terms of flows is not done using protocol names, but using
-+ environment indexes.
-+
-+ In terms of HW implementation, the list of distinction units sets the LCV vectors
-+ and later used for match vector, classification plan vectors and coarse classification
-+ indexing.
-+
-+ @Param[in,out] ioc_fm_pcd_net_env_params_t A structure defining the distiction units for this configuration.
-+
-+ @Return 0 on success; Error code otherwise.
-+*//***************************************************************************/
-+#if defined(CONFIG_COMPAT)
-+#define FM_PCD_IOC_NET_ENV_CHARACTERISTICS_SET_COMPAT _IOWR(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(20), ioc_compat_fm_pcd_net_env_params_t)
-+#endif
-+#define FM_PCD_IOC_NET_ENV_CHARACTERISTICS_SET _IOWR(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(20), ioc_fm_pcd_net_env_params_t)
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_NetEnvCharacteristicsDelete
-+
-+ @Description Deletes a set of Network Environment Charecteristics.
-+
-+ @Param[in] ioc_fm_obj_t - The id of a Network Environment object.
-+
-+ @Return 0 on success; Error code otherwise.
-+*//***************************************************************************/
-+#if defined(CONFIG_COMPAT)
-+#define FM_PCD_IOC_NET_ENV_CHARACTERISTICS_DELETE_COMPAT _IOW(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(21), ioc_compat_fm_obj_t)
-+#endif
-+#define FM_PCD_IOC_NET_ENV_CHARACTERISTICS_DELETE _IOW(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(21), ioc_fm_obj_t)
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_KgSchemeSet
-+
-+ @Description Initializing or modifying and enabling a scheme for the KeyGen.
-+ This routine should be called for adding or modifying a scheme.
-+ When a scheme needs modifying, the API requires that it will be
-+ rewritten. In such a case 'modify' should be TRUE. If the
-+ routine is called for a valid scheme and 'modify' is FALSE,
-+ it will return error.
-+
-+ @Param[in,out] ioc_fm_pcd_kg_scheme_params_t A structure of parameters for defining the scheme
-+
-+ @Return 0 on success; Error code otherwise.
-+*//***************************************************************************/
-+#if defined(CONFIG_COMPAT)
-+#define FM_PCD_IOC_KG_SCHEME_SET_COMPAT _IOWR(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(24), ioc_compat_fm_pcd_kg_scheme_params_t)
-+#endif
-+#define FM_PCD_IOC_KG_SCHEME_SET _IOWR(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(24), ioc_fm_pcd_kg_scheme_params_t)
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_KgSchemeDelete
-+
-+ @Description Deleting an initialized scheme.
-+
-+ @Param[in] ioc_fm_obj_t scheme id as initalized by application at FM_PCD_IOC_KG_SET_SCHEME
-+
-+ @Return 0 on success; Error code otherwise.
-+*//***************************************************************************/
-+#if defined(CONFIG_COMPAT)
-+#define FM_PCD_IOC_KG_SCHEME_DELETE_COMPAT _IOW(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(25), ioc_compat_fm_obj_t)
-+#endif
-+#define FM_PCD_IOC_KG_SCHEME_DELETE _IOW(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(25), ioc_fm_obj_t)
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_CcRootBuild
-+
-+ @Description This routine must be called to define a complete coarse
-+ classification tree. This is the way to define coarse
-+ classification to a certain flow - the KeyGen schemes
-+ may point only to trees defined in this way.
-+
-+ @Param[in,out] ioc_fm_pcd_cc_tree_params_t A structure of parameters to define the tree.
-+
-+ @Return 0 on success; Error code otherwise.
-+*//***************************************************************************/
-+#if defined(CONFIG_COMPAT)
-+#define FM_PCD_IOC_CC_ROOT_BUILD_COMPAT _IOWR(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(26), compat_uptr_t)
-+#endif
-+#define FM_PCD_IOC_CC_ROOT_BUILD _IOWR(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(26), void *) /* workaround ...*/
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_CcRootDelete
-+
-+ @Description Deleting a built tree.
-+
-+ @Param[in] ioc_fm_obj_t - The id of a CC tree.
-+*//***************************************************************************/
-+#if defined(CONFIG_COMPAT)
-+#define FM_PCD_IOC_CC_ROOT_DELETE_COMPAT _IOW(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(27), ioc_compat_fm_obj_t)
-+#endif
-+#define FM_PCD_IOC_CC_ROOT_DELETE _IOW(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(27), ioc_fm_obj_t)
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_MatchTableSet
-+
-+ @Description This routine should be called for each CC (coarse classification)
-+ node. The whole CC tree should be built bottom up so that each
-+ node points to already defined nodes. p_NodeId returns the node
-+ Id to be used by other nodes.
-+
-+ @Param[in,out] ioc_fm_pcd_cc_node_params_t A structure for defining the CC node params
-+
-+ @Return 0 on success; Error code otherwise.
-+*//***************************************************************************/
-+#if defined(CONFIG_COMPAT)
-+#define FM_PCD_IOC_MATCH_TABLE_SET_COMPAT _IOWR(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(28), compat_uptr_t)
-+#endif
-+#define FM_PCD_IOC_MATCH_TABLE_SET _IOWR(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(28), void *) /* workaround ...*/
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_MatchTableDelete
-+
-+ @Description Deleting a built node.
-+
-+ @Param[in] ioc_fm_obj_t - The id of a CC node.
-+
-+ @Return 0 on success; Error code otherwise.
-+*//***************************************************************************/
-+#if defined(CONFIG_COMPAT)
-+#define FM_PCD_IOC_MATCH_TABLE_DELETE_COMPAT _IOW(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(29), ioc_compat_fm_obj_t)
-+#endif
-+#define FM_PCD_IOC_MATCH_TABLE_DELETE _IOW(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(29), ioc_fm_obj_t)
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_CcRootModifyNextEngine
-+
-+ @Description Modify the Next Engine Parameters in the entry of the tree.
-+
-+ @Param[in] ioc_fm_pcd_cc_tree_modify_next_engine_params_t - Pointer to a structure with the relevant parameters
-+
-+ @Return 0 on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_CcRootBuild().
-+*//***************************************************************************/
-+#if defined(CONFIG_COMPAT)
-+#define FM_PCD_IOC_CC_ROOT_MODIFY_NEXT_ENGINE_COMPAT _IOW(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(30), ioc_compat_fm_pcd_cc_tree_modify_next_engine_params_t)
-+#endif
-+#define FM_PCD_IOC_CC_ROOT_MODIFY_NEXT_ENGINE _IOW(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(30), ioc_fm_pcd_cc_tree_modify_next_engine_params_t)
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_MatchTableModifyNextEngine
-+
-+ @Description Modify the Next Engine Parameters in the relevant key entry of the node.
-+
-+ @Param[in] ioc_fm_pcd_cc_node_modify_next_engine_params_t A pointer to a structure with the relevant parameters
-+
-+ @Return 0 on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_MatchTableSet().
-+*//***************************************************************************/
-+#if defined(CONFIG_COMPAT)
-+#define FM_PCD_IOC_MATCH_TABLE_MODIFY_NEXT_ENGINE_COMPAT _IOW(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(31), ioc_compat_fm_pcd_cc_node_modify_next_engine_params_t)
-+#endif
-+#define FM_PCD_IOC_MATCH_TABLE_MODIFY_NEXT_ENGINE _IOW(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(31), ioc_fm_pcd_cc_node_modify_next_engine_params_t)
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_MatchTableModifyMissNextEngine
-+
-+ @Description Modify the Next Engine Parameters of the Miss key case of the node.
-+
-+ @Param[in] ioc_fm_pcd_cc_node_modify_next_engine_params_t - Pointer to a structure with the relevant parameters
-+
-+ @Return 0 on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_MatchTableSet().
-+*//***************************************************************************/
-+#if defined(CONFIG_COMPAT)
-+#define FM_PCD_IOC_MATCH_TABLE_MODIFY_MISS_NEXT_ENGINE_COMPAT _IOW(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(32), ioc_compat_fm_pcd_cc_node_modify_next_engine_params_t)
-+#endif
-+#define FM_PCD_IOC_MATCH_TABLE_MODIFY_MISS_NEXT_ENGINE _IOW(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(32), ioc_fm_pcd_cc_node_modify_next_engine_params_t)
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_MatchTableRemoveKey
-+
-+ @Description Remove the key (including next engine parameters of this key)
-+ defined by the index of the relevant node.
-+
-+ @Param[in] ioc_fm_pcd_cc_node_remove_key_params_t A pointer to a structure with the relevant parameters
-+
-+ @Return 0 on success; Error code otherwise.
-+
-+ @Cautions Allowed only after FM_PCD_MatchTableSet() has been called for this
-+ node and for all of the nodes that lead to it.
-+*//***************************************************************************/
-+#if defined(CONFIG_COMPAT)
-+#define FM_PCD_IOC_MATCH_TABLE_REMOVE_KEY_COMPAT _IOW(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(33), ioc_compat_fm_pcd_cc_node_remove_key_params_t)
-+#endif
-+#define FM_PCD_IOC_MATCH_TABLE_REMOVE_KEY _IOW(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(33), ioc_fm_pcd_cc_node_remove_key_params_t)
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_MatchTableAddKey
-+
-+ @Description Add the key (including next engine parameters of this key in the
-+ index defined by the keyIndex. Note that 'FM_PCD_LAST_KEY_INDEX'
-+ may be used when the user doesn't care about the position of the
-+ key in the table - in that case, the key will be automatically
-+ added by the driver in the last available entry.
-+
-+ @Param[in] ioc_fm_pcd_cc_node_modify_key_and_next_engine_params_t A pointer to a structure with the relevant parameters
-+
-+ @Return 0 on success; Error code otherwise.
-+
-+ @Cautions Allowed only after FM_PCD_MatchTableSet() has been called for this
-+ node and for all of the nodes that lead to it.
-+*//***************************************************************************/
-+#if defined(CONFIG_COMPAT)
-+#define FM_PCD_IOC_MATCH_TABLE_ADD_KEY_COMPAT _IOW(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(34), ioc_compat_fm_pcd_cc_node_modify_key_and_next_engine_params_t)
-+#endif
-+#define FM_PCD_IOC_MATCH_TABLE_ADD_KEY _IOW(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(34), ioc_fm_pcd_cc_node_modify_key_and_next_engine_params_t)
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_MatchTableModifyKeyAndNextEngine
-+
-+ @Description Modify the key and Next Engine Parameters of this key in the index defined by key_index.
-+
-+ @Param[in] ioc_fm_pcd_cc_node_modify_key_and_next_engine_params_t A pointer to a structure with the relevant parameters
-+
-+ @Return 0 on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_MatchTableSet() not only of the relevnt node but also
-+ the node that points to this node
-+*//***************************************************************************/
-+#if defined(CONFIG_COMPAT)
-+#define FM_PCD_IOC_MATCH_TABLE_MODIFY_KEY_AND_NEXT_ENGINE_COMPAT _IOW(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(35), ioc_compat_fm_pcd_cc_node_modify_key_and_next_engine_params_t)
-+#endif
-+#define FM_PCD_IOC_MATCH_TABLE_MODIFY_KEY_AND_NEXT_ENGINE _IOW(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(35), ioc_fm_pcd_cc_node_modify_key_and_next_engine_params_t)
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_MatchTableModifyKey
-+
-+ @Description Modify the key at the index defined by key_index.
-+
-+ @Param[in] ioc_fm_pcd_cc_node_modify_key_params_t - Pointer to a structure with the relevant parameters
-+
-+ @Return 0 on success; Error code otherwise.
-+
-+ @Cautions Allowed only after FM_PCD_MatchTableSet() has been called for this
-+ node and for all of the nodes that lead to it.
-+*//***************************************************************************/
-+#if defined(CONFIG_COMPAT)
-+#define FM_PCD_IOC_MATCH_TABLE_MODIFY_KEY_COMPAT _IOW(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(36), ioc_compat_fm_pcd_cc_node_modify_key_params_t)
-+#endif
-+#define FM_PCD_IOC_MATCH_TABLE_MODIFY_KEY _IOW(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(36), ioc_fm_pcd_cc_node_modify_key_params_t)
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_HashTableSet
-+
-+ @Description This routine initializes a hash table structure.
-+ KeyGen hash result determines the hash bucket.
-+ Next, KeyGen key is compared against all keys of this
-+ bucket (exact match).
-+ Number of sets (number of buckets) of the hash equals to the
-+ number of 1-s in 'hash_res_mask' in the provided parameters.
-+ Number of hash table ways is then calculated by dividing
-+ 'max_num_of_keys' equally between the hash sets. This is the maximal
-+ number of keys that a hash bucket may hold.
-+ The hash table is initialized empty and keys may be
-+ added to it following the initialization. Keys masks are not
-+ supported in current hash table implementation.
-+ The initialized hash table can be integrated as a node in a
-+ CC tree.
-+
-+ @Param[in,out] ioc_fm_pcd_hash_table_params_t - Pointer to a structure with the relevant parameters
-+
-+ @Return 0 on success; Error code otherwise.
-+*//***************************************************************************/
-+#if defined(CONFIG_COMPAT)
-+#define FM_PCD_IOC_HASH_TABLE_SET_COMPAT _IOWR(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(37), ioc_compat_fm_pcd_hash_table_params_t)
-+#endif
-+#define FM_PCD_IOC_HASH_TABLE_SET _IOWR(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(37), ioc_fm_pcd_hash_table_params_t)
-+
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_HashTableDelete
-+
-+ @Description This routine deletes the provided hash table and released all
-+ its allocated resources.
-+
-+ @Param[in] ioc_fm_obj_t - The ID of a hash table.
-+
-+ @Return 0 on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_HashTableSet().
-+*//***************************************************************************/
-+#if defined(CONFIG_COMPAT)
-+#define FM_PCD_IOC_HASH_TABLE_DELETE_COMPAT _IOW(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(37), ioc_compat_fm_obj_t)
-+#endif
-+#define FM_PCD_IOC_HASH_TABLE_DELETE _IOW(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(37), ioc_fm_obj_t)
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_HashTableAddKey
-+
-+ @Description This routine adds the provided key (including next engine
-+ parameters of this key) to the hash table.
-+ The key is added as the last key of the bucket that it is
-+ mapped to.
-+
-+ @Param[in] ioc_fm_pcd_hash_table_add_key_params_t - Pointer to a structure with the relevant parameters
-+
-+ @Return 0 on success; error code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_HashTableSet().
-+*//***************************************************************************/
-+#if defined(CONFIG_COMPAT)
-+#define FM_PCD_IOC_HASH_TABLE_ADD_KEY_COMPAT _IOW(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(39), ioc_compat_fm_pcd_hash_table_add_key_params_t)
-+#endif
-+#define FM_PCD_IOC_HASH_TABLE_ADD_KEY _IOW(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(39), ioc_fm_pcd_hash_table_add_key_params_t)
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_HashTableRemoveKey
-+
-+ @Description This routine removes the requested key (including next engine
-+ parameters of this key) from the hash table.
-+
-+ @Param[in] ioc_fm_pcd_hash_table_remove_key_params_t - Pointer to a structure with the relevant parameters
-+
-+ @Return 0 on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_HashTableSet().
-+*//***************************************************************************/
-+#if defined(CONFIG_COMPAT)
-+#define FM_PCD_IOC_HASH_TABLE_REMOVE_KEY_COMPAT _IOW(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(40), ioc_compat_fm_pcd_hash_table_remove_key_params_t)
-+#endif
-+#define FM_PCD_IOC_HASH_TABLE_REMOVE_KEY _IOW(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(40), ioc_fm_pcd_hash_table_remove_key_params_t)
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_PlcrProfileSet
-+
-+ @Description Sets a profile entry in the policer profile table.
-+ The routine overrides any existing value.
-+
-+ @Param[in,out] ioc_fm_pcd_plcr_profile_params_t A structure of parameters for defining a
-+ policer profile entry.
-+
-+ @Return 0 on success; Error code otherwise.
-+*//***************************************************************************/
-+#if defined(CONFIG_COMPAT)
-+#define FM_PCD_IOC_PLCR_PROFILE_SET_COMPAT _IOWR(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(41), ioc_compat_fm_pcd_plcr_profile_params_t)
-+#endif
-+#define FM_PCD_IOC_PLCR_PROFILE_SET _IOWR(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(41), ioc_fm_pcd_plcr_profile_params_t)
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_PlcrProfileDelete
-+
-+ @Description Delete a profile entry in the policer profile table.
-+ The routine set entry to invalid.
-+
-+ @Param[in] ioc_fm_obj_t The id of a policer profile.
-+
-+ @Return 0 on success; Error code otherwise.
-+*//***************************************************************************/
-+#if defined(CONFIG_COMPAT)
-+#define FM_PCD_IOC_PLCR_PROFILE_DELETE_COMPAT _IOW(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(41), ioc_compat_fm_obj_t)
-+#endif
-+#define FM_PCD_IOC_PLCR_PROFILE_DELETE _IOW(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(41), ioc_fm_obj_t)
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_ManipNodeSet
-+
-+ @Description This routine should be called for defining a manipulation
-+ node. A manipulation node must be defined before the CC node
-+ that precedes it.
-+
-+ @Param[in] ioc_fm_pcd_manip_params_t - A structure of parameters defining the manipulation
-+
-+ @Return A handle to the initialized object on success; NULL code otherwise.
-+*//***************************************************************************/
-+#if defined(CONFIG_COMPAT)
-+#define FM_PCD_IOC_MANIP_NODE_SET_COMPAT _IOWR(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(43), ioc_compat_fm_pcd_manip_params_t)
-+#endif
-+#define FM_PCD_IOC_MANIP_NODE_SET _IOWR(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(43), ioc_fm_pcd_manip_params_t)
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_ManipNodeReplace
-+
-+ @Description Change existing manipulation node to be according to new requirement.
-+ (Here, it's implemented as a variant of the same IOCTL as for
-+ FM_PCD_ManipNodeSet(), and one that when called, the 'id' member
-+ in its 'ioc_fm_pcd_manip_params_t' argument is set to contain
-+ the manip node's handle)
-+
-+ @Param[in] ioc_fm_pcd_manip_params_t - A structure of parameters defining the manipulation
-+
-+ @Return 0 on success; error code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_ManipNodeSet().
-+*//***************************************************************************/
-+#if defined(CONFIG_COMPAT)
-+#define FM_PCD_IOC_MANIP_NODE_REPLACE_COMPAT FM_PCD_IOC_MANIP_NODE_SET_COMPAT
-+#endif
-+#define FM_PCD_IOC_MANIP_NODE_REPLACE FM_PCD_IOC_MANIP_NODE_SET
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_ManipNodeDelete
-+
-+ @Description Delete an existing manipulation node.
-+
-+ @Param[in] ioc_fm_obj_t The id of the manipulation node to delete.
-+
-+ @Return 0 on success; error code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_ManipNodeSet().
-+*//***************************************************************************/
-+#if defined(CONFIG_COMPAT)
-+#define FM_PCD_IOC_MANIP_NODE_DELETE_COMPAT _IOW(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(44), ioc_compat_fm_obj_t)
-+#endif
-+#define FM_PCD_IOC_MANIP_NODE_DELETE _IOW(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(44), ioc_fm_obj_t)
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_ManipGetStatistics
-+
-+ @Description Retrieve the manipulation statistics.
-+
-+ @Param[in] h_ManipNode A handle to a manipulation node.
-+ @Param[out] p_FmPcdManipStats A structure for retrieving the manipulation statistics
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_ManipNodeSet().
-+*//***************************************************************************/
-+#if defined(CONFIG_COMPAT)
-+#define FM_PCD_IOC_MANIP_GET_STATS_COMPAT _IOWR(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(50), ioc_compat_fm_pcd_manip_get_stats_t)
-+#endif
-+#define FM_PCD_IOC_MANIP_GET_STATS _IOWR(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(50), ioc_fm_pcd_manip_get_stats_t)
-+
-+/**************************************************************************//**
-+@Function FM_PCD_SetAdvancedOffloadSupport
-+
-+@Description This routine must be called in order to support the following features:
-+ IP-fragmentation, IP-reassembly, IPsec, Header-manipulation, frame-replicator.
-+
-+@Param[in] h_FmPcd FM PCD module descriptor.
-+
-+@Return 0 on success; error code otherwise.
-+
-+@Cautions Allowed only when PCD is disabled.
-+*//***************************************************************************/
-+#define FM_PCD_IOC_SET_ADVANCED_OFFLOAD_SUPPORT _IO(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(45))
-+
-+#if (DPAA_VERSION >= 11)
-+/**************************************************************************//**
-+ @Function FM_PCD_FrmReplicSetGroup
-+
-+ @Description Initialize a Frame Replicator group.
-+
-+ @Param[in] h_FmPcd FM PCD module descriptor.
-+ @Param[in] p_FrmReplicGroupParam A structure of parameters for the initialization of
-+ the frame replicator group.
-+
-+ @Return A handle to the initialized object on success; NULL code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_Init().
-+*//***************************************************************************/
-+#if defined(CONFIG_COMPAT)
-+#define FM_PCD_IOC_FRM_REPLIC_GROUP_SET_COMPAT _IOWR(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(46), ioc_compat_fm_pcd_frm_replic_group_params_t)
-+#endif
-+#define FM_PCD_IOC_FRM_REPLIC_GROUP_SET _IOWR(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(46), ioc_fm_pcd_frm_replic_group_params_t)
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_FrmReplicDeleteGroup
-+
-+ @Description Delete a Frame Replicator group.
-+
-+ @Param[in] h_FrmReplicGroup A handle to the frame replicator group.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_FrmReplicSetGroup().
-+*//***************************************************************************/
-+#if defined(CONFIG_COMPAT)
-+#define FM_PCD_IOC_FRM_REPLIC_GROUP_DELETE_COMPAT _IOWR(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(47), ioc_compat_fm_obj_t)
-+#endif
-+#define FM_PCD_IOC_FRM_REPLIC_GROUP_DELETE _IOWR(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(47), ioc_fm_obj_t)
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_FrmReplicAddMember
-+
-+ @Description Add the member in the index defined by the memberIndex.
-+
-+ @Param[in] h_FrmReplicGroup A handle to the frame replicator group.
-+ @Param[in] memberIndex member index for adding.
-+ @Param[in] p_MemberParams A pointer to the new member parameters.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_FrmReplicSetGroup() of this group.
-+*//***************************************************************************/
-+#if defined(CONFIG_COMPAT)
-+#define FM_PCD_IOC_FRM_REPLIC_MEMBER_ADD_COMPAT _IOWR(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(48), ioc_compat_fm_pcd_frm_replic_member_params_t)
-+#endif
-+#define FM_PCD_IOC_FRM_REPLIC_MEMBER_ADD _IOWR(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(48), ioc_fm_pcd_frm_replic_member_params_t)
-+
-+/**************************************************************************//**
-+ @Function FM_PCD_FrmReplicRemoveMember
-+
-+ @Description Remove the member defined by the index from the relevant group.
-+
-+ @Param[in] h_FrmReplicGroup A handle to the frame replicator group.
-+ @Param[in] memberIndex member index for removing.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PCD_FrmReplicSetGroup() of this group.
-+*//***************************************************************************/
-+#if defined(CONFIG_COMPAT)
-+#define FM_PCD_IOC_FRM_REPLIC_MEMBER_REMOVE_COMPAT _IOWR(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(49), ioc_compat_fm_pcd_frm_replic_member_t)
-+#endif
-+#define FM_PCD_IOC_FRM_REPLIC_MEMBER_REMOVE _IOWR(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(49), ioc_fm_pcd_frm_replic_member_t)
-+
-+#endif
-+
-+#if (defined(FM_CAPWAP_SUPPORT) && (DPAA_VERSION == 10))
-+/**************************************************************************//**
-+ @Function FM_PCD_StatisticsSetNode
-+
-+ @Description This routine should be called for defining a statistics node.
-+
-+ @Param[in,out] ioc_fm_pcd_stats_params_t A structure of parameters defining the statistics
-+
-+ @Return 0 on success; Error code otherwise.
-+*//***************************************************************************/
-+#if defined(CONFIG_COMPAT)
-+#define FM_PCD_IOC_STATISTICS_SET_NODE_COMPAT _IOWR(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(45), void *)
-+#endif
-+#define FM_PCD_IOC_STATISTICS_SET_NODE _IOWR(FM_IOC_TYPE_BASE, FM_PCD_IOC_NUM(45), void *)
-+
-+#endif /* FM_CAPWAP_SUPPORT */
-+
-+#ifdef NCSW_BACKWARD_COMPATIBLE_API
-+#if defined(CONFIG_COMPAT)
-+#define FM_PCD_IOC_SET_NET_ENV_CHARACTERISTICS_COMPAT \
-+ FM_PCD_IOC_NET_ENV_CHARACTERISTICS_SET_COMPAT
-+#define FM_PCD_IOC_DELETE_NET_ENV_CHARACTERISTICS_COMPAT \
-+ FM_PCD_IOC_NET_ENV_CHARACTERISTICS_DELETE_COMPAT
-+#define FM_PCD_IOC_KG_SET_SCHEME_COMPAT FM_PCD_IOC_KG_SCHEME_SET_COMPAT
-+#define FM_PCD_IOC_KG_DEL_SCHEME_COMPAT FM_PCD_IOC_KG_SCHEME_DELETE_COMPAT
-+#define FM_PCD_IOC_CC_BUILD_TREE_COMPAT FM_PCD_IOC_CC_ROOT_BUILD_COMPAT
-+#define FM_PCD_IOC_CC_DELETE_TREE_COMPAT FM_PCD_IOC_CC_ROOT_DELETE_COMPAT
-+#define FM_PCD_IOC_CC_DELETE_NODE_COMPAT FM_PCD_IOC_MATCH_TABLE_DELETE_COMPAT
-+#define FM_PCD_IOC_CC_TREE_MODIFY_NEXT_ENGINE_COMPAT \
-+ FM_PCD_IOC_CC_ROOT_MODIFY_NEXT_ENGINE_COMPAT
-+#define FM_PCD_IOC_CC_NODE_MODIFY_NEXT_ENGINE_COMPAT \
-+ FM_PCD_IOC_MATCH_TABLE_MODIFY_NEXT_ENGINE_COMPAT
-+#define FM_PCD_IOC_CC_NODE_MODIFY_MISS_NEXT_ENGINE_COMPAT \
-+ FM_PCD_IOC_MATCH_TABLE_MODIFY_MISS_NEXT_ENGINE_COMPAT
-+#define FM_PCD_IOC_CC_NODE_REMOVE_KEY_COMPAT FM_PCD_IOC_MATCH_TABLE_REMOVE_KEY_COMPAT
-+#define FM_PCD_IOC_CC_NODE_ADD_KEY_COMPAT FM_PCD_IOC_MATCH_TABLE_ADD_KEY_COMPAT
-+#define FM_PCD_IOC_CC_NODE_MODIFY_KEY_AND_NEXT_ENGINE_COMPAT \
-+ FM_PCD_IOC_MATCH_TABLE_MODIFY_KEY_AND_NEXT_ENGINE_COMPAT
-+#define FM_PCD_IOC_CC_NODE_MODIFY_KEY_COMPAT FM_PCD_IOC_MATCH_TABLE_MODIFY_KEY_COMPAT
-+#define FM_PCD_IOC_PLCR_SET_PROFILE_COMPAT FM_PCD_IOC_PLCR_PROFILE_SET_COMPAT
-+#define FM_PCD_IOC_PLCR_DEL_PROFILE_COMPAT FM_PCD_IOC_PLCR_PROFILE_DELETE_COMPAT
-+#define FM_PCD_IOC_MANIP_SET_NODE_COMPAT FM_PCD_IOC_MANIP_NODE_SET_COMPAT
-+#define FM_PCD_IOC_MANIP_DELETE_NODE_COMPAT FM_PCD_IOC_MANIP_NODE_DELETE_COMPAT
-+#endif
-+#define FM_PCD_IOC_SET_NET_ENV_CHARACTERISTICS FM_PCD_IOC_NET_ENV_CHARACTERISTICS_SET
-+#define FM_PCD_IOC_DELETE_NET_ENV_CHARACTERISTICS \
-+ FM_PCD_IOC_NET_ENV_CHARACTERISTICS_DELETE
-+#define FM_PCD_IOC_KG_SET_SCHEME FM_PCD_IOC_KG_SCHEME_SET
-+#define FM_PCD_IOC_KG_DEL_SCHEME FM_PCD_IOC_KG_SCHEME_DELETE
-+#define FM_PCD_IOC_CC_BUILD_TREE FM_PCD_IOC_CC_ROOT_BUILD
-+#define FM_PCD_IOC_CC_DELETE_TREE FM_PCD_IOC_CC_ROOT_DELETE
-+#define FM_PCD_IOC_CC_DELETE_NODE FM_PCD_IOC_MATCH_TABLE_DELETE
-+#define FM_PCD_IOC_CC_TREE_MODIFY_NEXT_ENGINE FM_PCD_IOC_CC_ROOT_MODIFY_NEXT_ENGINE
-+#define FM_PCD_IOC_CC_NODE_MODIFY_NEXT_ENGINE FM_PCD_IOC_MATCH_TABLE_MODIFY_NEXT_ENGINE
-+#define FM_PCD_IOC_CC_NODE_MODIFY_MISS_NEXT_ENGINE \
-+ FM_PCD_IOC_MATCH_TABLE_MODIFY_MISS_NEXT_ENGINE
-+#define FM_PCD_IOC_CC_NODE_REMOVE_KEY FM_PCD_IOC_MATCH_TABLE_REMOVE_KEY
-+#define FM_PCD_IOC_CC_NODE_ADD_KEY FM_PCD_IOC_MATCH_TABLE_ADD_KEY
-+#define FM_PCD_IOC_CC_NODE_MODIFY_KEY_AND_NEXT_ENGINE \
-+ FM_PCD_IOC_MATCH_TABLE_MODIFY_KEY_AND_NEXT_ENGINE
-+#define FM_PCD_IOC_CC_NODE_MODIFY_KEY FM_PCD_IOC_MATCH_TABLE_MODIFY_KEY
-+#define FM_PCD_IOC_PLCR_SET_PROFILE FM_PCD_IOC_PLCR_PROFILE_SET
-+#define FM_PCD_IOC_PLCR_DEL_PROFILE FM_PCD_IOC_PLCR_PROFILE_DELETE
-+#define FM_PCD_IOC_MANIP_SET_NODE FM_PCD_IOC_MANIP_NODE_SET
-+#define FM_PCD_IOC_MANIP_DELETE_NODE FM_PCD_IOC_MANIP_NODE_DELETE
-+#endif /* NCSW_BACKWARD_COMPATIBLE_API */
-+
-+#endif /* __FM_PCD_IOCTLS_H */
-+/** @} */ /* end of lnx_ioctl_FM_PCD_Runtime_grp group */
-+/** @} */ /* end of lnx_ioctl_FM_PCD_grp group */
-+/** @} */ /* end of lnx_ioctl_FM_grp group */
---- /dev/null
-+++ b/include/uapi/linux/fmd/Peripherals/fm_port_ioctls.h
-@@ -0,0 +1,973 @@
-+/* Copyright (c) 2008-2012 Freescale Semiconductor, Inc.
-+ * All rights reserved.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+/******************************************************************************
-+ @File fm_port_ioctls.h
-+
-+ @Description FM Port routines
-+*//***************************************************************************/
-+#ifndef __FM_PORT_IOCTLS_H
-+#define __FM_PORT_IOCTLS_H
-+
-+#include "enet_ext.h"
-+#include "net_ioctls.h"
-+#include "fm_ioctls.h"
-+#include "fm_pcd_ioctls.h"
-+
-+
-+/**************************************************************************//**
-+
-+ @Group lnx_ioctl_FM_grp Frame Manager Linux IOCTL API
-+
-+ @Description FM Linux ioctls definitions and enums
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Group lnx_ioctl_FM_PORT_grp FM Port
-+
-+ @Description FM Port API
-+
-+ The FM uses a general module called "port" to represent a Tx port
-+ (MAC), an Rx port (MAC), offline parsing flow or host command
-+ flow. There may be up to 17 (may change) ports in an FM - 5 Tx
-+ ports (4 for the 1G MACs, 1 for the 10G MAC), 5 Rx Ports, and 7
-+ Host command/Offline parsing ports. The SW driver manages these
-+ ports as sub-modules of the FM, i.e. after an FM is initialized,
-+ its ports may be initialized and operated upon.
-+
-+ The port is initialized aware of its type, but other functions on
-+ a port may be indifferent to its type. When necessary, the driver
-+ verifies coherency and returns error if applicable.
-+
-+ On initialization, user specifies the port type and it's index
-+ (relative to the port's type). Host command and Offline parsing
-+ ports share the same id range, I.e user may not initialized host
-+ command port 0 and offline parsing port 0.
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Description An enum for defining port PCD modes.
-+ (Must match enum e_FmPortPcdSupport defined in fm_port_ext.h)
-+
-+ This enum defines the superset of PCD engines support - i.e. not
-+ all engines have to be used, but all have to be enabled. The real
-+ flow of a specific frame depends on the PCD configuration and the
-+ frame headers and payload.
-+ Note: the first engine and the first engine after the parser (if
-+ exists) should be in order, the order is important as it will
-+ define the flow of the port. However, as for the rest engines
-+ (the ones that follows), the order is not important anymore as
-+ it is defined by the PCD graph itself.
-+*//***************************************************************************/
-+typedef enum ioc_fm_port_pcd_support {
-+ e_IOC_FM_PORT_PCD_SUPPORT_NONE = 0 /**< BMI to BMI, PCD is not used */
-+ , e_IOC_FM_PORT_PCD_SUPPORT_PRS_ONLY /**< Use only Parser */
-+ , e_IOC_FM_PORT_PCD_SUPPORT_PLCR_ONLY /**< Use only Policer */
-+ , e_IOC_FM_PORT_PCD_SUPPORT_PRS_AND_PLCR /**< Use Parser and Policer */
-+ , e_IOC_FM_PORT_PCD_SUPPORT_PRS_AND_KG /**< Use Parser and Keygen */
-+ , e_IOC_FM_PORT_PCD_SUPPORT_PRS_AND_KG_AND_CC /**< Use Parser, Keygen and Coarse Classification */
-+ , e_IOC_FM_PORT_PCD_SUPPORT_PRS_AND_KG_AND_CC_AND_PLCR
-+ /**< Use all PCD engines */
-+ , e_IOC_FM_PORT_PCD_SUPPORT_PRS_AND_KG_AND_PLCR /**< Use Parser, Keygen and Policer */
-+ , e_IOC_FM_PORT_PCD_SUPPORT_PRS_AND_CC /**< Use Parser and Coarse Classification */
-+ , e_IOC_FM_PORT_PCD_SUPPORT_PRS_AND_CC_AND_PLCR /**< Use Parser and Coarse Classification and Policer */
-+ , e_IOC_FM_PORT_PCD_SUPPORT_CC_ONLY /**< Use only Coarse Classification */
-+#if (defined(FM_CAPWAP_SUPPORT) && (DPAA_VERSION == 10))
-+ , e_IOC_FM_PORT_PCD_SUPPORT_CC_AND_KG /**< Use Coarse Classification,and Keygen */
-+ , e_IOC_FM_PORT_PCD_SUPPORT_CC_AND_KG_AND_PLCR /**< Use Coarse Classification, Keygen and Policer */
-+#endif /* FM_CAPWAP_SUPPORT */
-+} ioc_fm_port_pcd_support;
-+
-+
-+/**************************************************************************//**
-+ @Collection FM Frame error
-+*//***************************************************************************/
-+typedef uint32_t ioc_fm_port_frame_err_select_t; /**< typedef for defining Frame Descriptor errors */
-+
-+/* @} */
-+
-+
-+/**************************************************************************//**
-+ @Description An enum for defining Dual Tx rate limiting scale.
-+ (Must match e_FmPortDualRateLimiterScaleDown defined in fm_port_ext.h)
-+*//***************************************************************************/
-+typedef enum ioc_fm_port_dual_rate_limiter_scale_down {
-+ e_IOC_FM_PORT_DUAL_RATE_LIMITER_NONE = 0, /**< Use only single rate limiter */
-+ e_IOC_FM_PORT_DUAL_RATE_LIMITER_SCALE_DOWN_BY_2, /**< Divide high rate limiter by 2 */
-+ e_IOC_FM_PORT_DUAL_RATE_LIMITER_SCALE_DOWN_BY_4, /**< Divide high rate limiter by 4 */
-+ e_IOC_FM_PORT_DUAL_RATE_LIMITER_SCALE_DOWN_BY_8 /**< Divide high rate limiter by 8 */
-+} ioc_fm_port_dual_rate_limiter_scale_down;
-+
-+/**************************************************************************//**
-+ @Description A structure for defining Tx rate limiting
-+ (Must match struct t_FmPortRateLimit defined in fm_port_ext.h)
-+*//***************************************************************************/
-+typedef struct ioc_fm_port_rate_limit_t {
-+ uint16_t max_burst_size; /**< in KBytes for Tx ports, in frames
-+ for offline parsing ports. (note that
-+ for early chips burst size is
-+ rounded up to a multiply of 1000 frames).*/
-+ uint32_t rate_limit; /**< in Kb/sec for Tx ports, in frame/sec for
-+ offline parsing ports. Rate limit refers to
-+ data rate (rather than line rate). */
-+ ioc_fm_port_dual_rate_limiter_scale_down rate_limit_divider; /**< For offline parsing ports only. Not-valid
-+ for some earlier chip revisions */
-+} ioc_fm_port_rate_limit_t;
-+
-+
-+
-+/**************************************************************************//**
-+ @Group lnx_ioctl_FM_PORT_runtime_control_grp FM Port Runtime Control Unit
-+
-+ @Description FM Port Runtime control unit API functions, definitions and enums.
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Description An enum for defining FM Port counters.
-+ (Must match enum e_FmPortCounters defined in fm_port_ext.h)
-+*//***************************************************************************/
-+typedef enum ioc_fm_port_counters {
-+ e_IOC_FM_PORT_COUNTERS_CYCLE, /**< BMI performance counter */
-+ e_IOC_FM_PORT_COUNTERS_TASK_UTIL, /**< BMI performance counter */
-+ e_IOC_FM_PORT_COUNTERS_QUEUE_UTIL, /**< BMI performance counter */
-+ e_IOC_FM_PORT_COUNTERS_DMA_UTIL, /**< BMI performance counter */
-+ e_IOC_FM_PORT_COUNTERS_FIFO_UTIL, /**< BMI performance counter */
-+ e_IOC_FM_PORT_COUNTERS_RX_PAUSE_ACTIVATION, /**< BMI Rx only performance counter */
-+ e_IOC_FM_PORT_COUNTERS_FRAME, /**< BMI statistics counter */
-+ e_IOC_FM_PORT_COUNTERS_DISCARD_FRAME, /**< BMI statistics counter */
-+ e_IOC_FM_PORT_COUNTERS_DEALLOC_BUF, /**< BMI deallocate buffer statistics counter */
-+ e_IOC_FM_PORT_COUNTERS_RX_BAD_FRAME, /**< BMI Rx only statistics counter */
-+ e_IOC_FM_PORT_COUNTERS_RX_LARGE_FRAME, /**< BMI Rx only statistics counter */
-+ e_IOC_FM_PORT_COUNTERS_RX_FILTER_FRAME, /**< BMI Rx & OP only statistics counter */
-+ e_IOC_FM_PORT_COUNTERS_RX_LIST_DMA_ERR, /**< BMI Rx, OP & HC only statistics counter */
-+ e_IOC_FM_PORT_COUNTERS_RX_OUT_OF_BUFFERS_DISCARD, /**< BMI Rx, OP & HC statistics counter */
-+ e_IOC_FM_PORT_COUNTERS_PREPARE_TO_ENQUEUE_COUNTER, /**< BMI Rx, OP & HC only statistics counter */
-+ e_IOC_FM_PORT_COUNTERS_WRED_DISCARD, /**< BMI OP & HC only statistics counter */
-+ e_IOC_FM_PORT_COUNTERS_LENGTH_ERR, /**< BMI non-Rx statistics counter */
-+ e_IOC_FM_PORT_COUNTERS_UNSUPPRTED_FORMAT, /**< BMI non-Rx statistics counter */
-+ e_IOC_FM_PORT_COUNTERS_DEQ_TOTAL, /**< QMI total QM dequeues counter */
-+ e_IOC_FM_PORT_COUNTERS_ENQ_TOTAL, /**< QMI total QM enqueues counter */
-+ e_IOC_FM_PORT_COUNTERS_DEQ_FROM_DEFAULT, /**< QMI counter */
-+ e_IOC_FM_PORT_COUNTERS_DEQ_CONFIRM /**< QMI counter */
-+} ioc_fm_port_counters;
-+
-+typedef struct ioc_fm_port_bmi_stats_t {
-+ uint32_t cnt_cycle;
-+ uint32_t cnt_task_util;
-+ uint32_t cnt_queue_util;
-+ uint32_t cnt_dma_util;
-+ uint32_t cnt_fifo_util;
-+ uint32_t cnt_rx_pause_activation;
-+ uint32_t cnt_frame;
-+ uint32_t cnt_discard_frame;
-+ uint32_t cnt_dealloc_buf;
-+ uint32_t cnt_rx_bad_frame;
-+ uint32_t cnt_rx_large_frame;
-+ uint32_t cnt_rx_filter_frame;
-+ uint32_t cnt_rx_list_dma_err;
-+ uint32_t cnt_rx_out_of_buffers_discard;
-+ uint32_t cnt_wred_discard;
-+ uint32_t cnt_length_err;
-+ uint32_t cnt_unsupported_format;
-+} ioc_fm_port_bmi_stats_t;
-+
-+/**************************************************************************//**
-+ @Description Structure for Port id parameters.
-+ (Description may be inaccurate;
-+ must match struct t_FmPortCongestionGrps defined in fm_port_ext.h)
-+
-+ Fields commented 'IN' are passed by the port module to be used
-+ by the FM module.
-+ Fields commented 'OUT' will be filled by FM before returning to port.
-+*//***************************************************************************/
-+typedef struct ioc_fm_port_congestion_groups_t {
-+ uint16_t num_of_congestion_grps_to_consider; /**< The number of required congestion groups
-+ to define the size of the following array */
-+ uint8_t congestion_grps_to_consider [FM_PORT_NUM_OF_CONGESTION_GRPS];
-+ /**< An array of CG indexes;
-+ Note that the size of the array should be
-+ 'num_of_congestion_grps_to_consider'. */
-+#if DPAA_VERSION >= 11
-+ bool pfc_priorities_enable[FM_PORT_NUM_OF_CONGESTION_GRPS][FM_MAX_NUM_OF_PFC_PRIORITIES];
-+ /**< A matrix that represents the map between the CG ids
-+ defined in 'congestion_grps_to_consider' to the priorities
-+ mapping array. */
-+#endif /* DPAA_VERSION >= 11 */
-+} ioc_fm_port_congestion_groups_t;
-+
-+
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_Disable
-+
-+ @Description Gracefully disable an FM port. The port will not start new tasks after all
-+ tasks associated with the port are terminated.
-+
-+ @Return 0 on success; error code otherwise.
-+
-+ @Cautions This is a blocking routine, it returns after port is
-+ gracefully stopped, i.e. the port will not except new frames,
-+ but it will finish all frames or tasks which were already began
-+*//***************************************************************************/
-+#define FM_PORT_IOC_DISABLE _IO(FM_IOC_TYPE_BASE, FM_PORT_IOC_NUM(1))
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_Enable
-+
-+ @Description A runtime routine provided to allow disable/enable of port.
-+
-+ @Return 0 on success; error code otherwise.
-+*//***************************************************************************/
-+#define FM_PORT_IOC_ENABLE _IO(FM_IOC_TYPE_BASE, FM_PORT_IOC_NUM(2))
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_SetRateLimit
-+
-+ @Description Calling this routine enables rate limit algorithm.
-+ By default, this functionality is disabled.
-+ Note that rate-limit mechanism uses the FM time stamp.
-+ The selected rate limit specified here would be
-+ rounded DOWN to the nearest 16M.
-+
-+ May be used for Tx and offline parsing ports only
-+
-+ @Param[in] ioc_fm_port_rate_limit A structure of rate limit parameters
-+
-+ @Return 0 on success; error code otherwise.
-+*//***************************************************************************/
-+#define FM_PORT_IOC_SET_RATE_LIMIT _IOW(FM_IOC_TYPE_BASE, FM_PORT_IOC_NUM(3), ioc_fm_port_rate_limit_t)
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_DeleteRateLimit
-+
-+ @Description Calling this routine disables the previously enabled rate limit.
-+
-+ May be used for Tx and offline parsing ports only
-+
-+ @Return 0 on success; error code otherwise.
-+*//***************************************************************************/
-+#define FM_PORT_IOC_DELETE_RATE_LIMIT _IO(FM_IOC_TYPE_BASE, FM_PORT_IOC_NUM(5))
-+#define FM_PORT_IOC_REMOVE_RATE_LIMIT FM_PORT_IOC_DELETE_RATE_LIMIT
-+
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_AddCongestionGrps
-+
-+ @Description This routine effects the corresponding Tx port.
-+ It should be called in order to enable pause
-+ frame transmission in case of congestion in one or more
-+ of the congestion groups relevant to this port.
-+ Each call to this routine may add one or more congestion
-+ groups to be considered relevant to this port.
-+
-+ May be used for Rx, or RX+OP ports only (depending on chip)
-+
-+ @Param[in] ioc_fm_port_congestion_groups_t - A pointer to an array of
-+ congestion group ids to consider.
-+
-+ @Return 0 on success; error code otherwise.
-+*//***************************************************************************/
-+#define FM_PORT_IOC_ADD_CONGESTION_GRPS _IOW(FM_IOC_TYPE_BASE, FM_PORT_IOC_NUM(34), ioc_fm_port_congestion_groups_t)
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_RemoveCongestionGrps
-+
-+ @Description This routine effects the corresponding Tx port. It should be
-+ called when congestion groups were
-+ defined for this port and are no longer relevant, or pause
-+ frames transmitting is not required on their behalf.
-+ Each call to this routine may remove one or more congestion
-+ groups to be considered relevant to this port.
-+
-+ May be used for Rx, or RX+OP ports only (depending on chip)
-+
-+ @Param[in] ioc_fm_port_congestion_groups_t - A pointer to an array of
-+ congestion group ids to consider.
-+
-+ @Return 0 on success; error code otherwise.
-+*//***************************************************************************/
-+#define FM_PORT_IOC_REMOVE_CONGESTION_GRPS _IOW(FM_IOC_TYPE_BASE, FM_PORT_IOC_NUM(35), ioc_fm_port_congestion_groups_t)
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_SetErrorsRoute
-+
-+ @Description Errors selected for this routine will cause a frame with that error
-+ to be enqueued to error queue.
-+ Errors not selected for this routine will cause a frame with that error
-+ to be enqueued to the one of the other port queues.
-+ By default all errors are defined to be enqueued to error queue.
-+ Errors that were configured to be discarded (at initialization)
-+ may not be selected here.
-+
-+ May be used for Rx and offline parsing ports only
-+
-+ @Param[in] ioc_fm_port_frame_err_select_t A list of errors to enqueue to error queue
-+
-+ @Return 0 on success; error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Config() and before FM_PORT_Init().
-+ (szbs001: How is it possible to have one function that needs to be
-+ called BEFORE FM_PORT_Init() implemented as an ioctl,
-+ which will ALWAYS be called AFTER the FM_PORT_Init()
-+ for that port!?!?!?!???!?!??!?!?)
-+*//***************************************************************************/
-+#define FM_PORT_IOC_SET_ERRORS_ROUTE _IOW(FM_IOC_TYPE_BASE, FM_PORT_IOC_NUM(4), ioc_fm_port_frame_err_select_t)
-+
-+
-+/**************************************************************************//**
-+ @Group lnx_ioctl_FM_PORT_pcd_runtime_control_grp FM Port PCD Runtime Control Unit
-+
-+ @Description FM Port PCD Runtime control unit API functions, definitions and enums.
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Description A structure defining the KG scheme after the parser.
-+ (Must match struct t_FmPcdKgSchemeSelect defined in fm_port_ext.h)
-+
-+ This is relevant only to change scheme selection mode - from
-+ direct to indirect and vice versa, or when the scheme is selected directly,
-+ to select the scheme id.
-+
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_kg_scheme_select_t {
-+ bool direct; /**< TRUE to use 'scheme_id' directly, FALSE to use LCV.*/
-+ void *scheme_id; /**< Relevant for 'direct'=TRUE only.
-+ 'scheme_id' selects the scheme after parser. */
-+} ioc_fm_pcd_kg_scheme_select_t;
-+
-+/**************************************************************************//**
-+ @Description Scheme IDs structure
-+ (Must match struct t_FmPcdPortSchemesParams defined in fm_port_ext.h)
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_port_schemes_params_t {
-+ uint8_t num_of_schemes; /**< Number of schemes for port to be bound to. */
-+ void *scheme_ids[FM_PCD_KG_NUM_OF_SCHEMES]; /**< Array of 'num_of_schemes' schemes for the
-+ port to be bound to */
-+} ioc_fm_pcd_port_schemes_params_t;
-+
-+/**************************************************************************//**
-+ @Description A union for defining port protocol parameters for parser
-+ (Must match union u_FmPcdHdrPrsOpts defined in fm_port_ext.h)
-+*//***************************************************************************/
-+typedef union ioc_fm_pcd_hdr_prs_opts_u {
-+ /* MPLS */
-+ struct {
-+ bool label_interpretation_enable;/**< When this bit is set, the last MPLS label will be
-+ interpreted as described in HW spec table. When the bit
-+ is cleared, the parser will advance to MPLS next parse */
-+ ioc_net_header_type next_parse; /**< must be equal or higher than IPv4 */
-+ } mpls_prs_options;
-+
-+ /* VLAN */
-+ struct {
-+ uint16_t tag_protocol_id1; /**< User defined Tag Protocol Identifier, to be recognized
-+ on VLAN TAG on top of 0x8100 and 0x88A8 */
-+ uint16_t tag_protocol_id2; /**< User defined Tag Protocol Identifier, to be recognized
-+ on VLAN TAG on top of 0x8100 and 0x88A8 */
-+ } vlan_prs_options;
-+
-+ /* PPP */
-+ struct{
-+ bool enable_mtu_check; /**< Check validity of MTU according to RFC2516 */
-+ } pppoe_prs_options;
-+
-+ /* IPV6 */
-+ struct {
-+ bool routing_hdr_disable; /**< Disable routing header */
-+ } ipv6_prs_options;
-+
-+ /* UDP */
-+ struct {
-+ bool pad_ignore_checksum; /**< TRUE to ignore pad in checksum */
-+ } udp_prs_options;
-+
-+ /* TCP */
-+ struct {
-+ bool pad_ignore_checksum; /**< TRUE to ignore pad in checksum */
-+ } tcp_prs_options;
-+} ioc_fm_pcd_hdr_prs_opts_u;
-+
-+/**************************************************************************//**
-+ @Description A structure for defining each header for the parser
-+ (must match struct t_FmPcdPrsAdditionalHdrParams defined in fm_port_ext.h)
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_prs_additional_hdr_params_t {
-+ ioc_net_header_type hdr; /**< Selected header */
-+ bool err_disable; /**< TRUE to disable error indication */
-+ bool soft_prs_enable; /**< Enable jump to SW parser when this
-+ header is recognized by the HW parser. */
-+ uint8_t index_per_hdr; /**< Normally 0, if more than one sw parser
-+ attachments exists for the same header,
-+ (in the main sw parser code) use this
-+ index to distinguish between them. */
-+ bool use_prs_opts; /**< TRUE to use parser options. */
-+ ioc_fm_pcd_hdr_prs_opts_u prs_opts; /**< A unuion according to header type,
-+ defining the parser options selected.*/
-+} ioc_fm_pcd_prs_additional_hdr_params_t;
-+
-+/**************************************************************************//**
-+ @Description A structure for defining port PCD parameters
-+ (Must match t_FmPortPcdPrsParams defined in fm_port_ext.h)
-+*//***************************************************************************/
-+typedef struct ioc_fm_port_pcd_prs_params_t {
-+ uint8_t prs_res_priv_info; /**< The private info provides a method of inserting
-+ port information into the parser result. This information
-+ may be extracted by KeyGen and be used for frames
-+ distribution when a per-port distinction is required,
-+ it may also be used as a port logical id for analyzing
-+ incoming frames. */
-+ uint8_t parsing_offset; /**< Number of bytes from begining of packet to start parsing */
-+ ioc_net_header_type first_prs_hdr; /**< The type of the first header axpected at 'parsing_offset' */
-+ bool include_in_prs_statistics; /**< TRUE to include this port in the parser statistics */
-+ uint8_t num_of_hdrs_with_additional_params;
-+ /**< Normally 0, some headers may get special parameters */
-+ ioc_fm_pcd_prs_additional_hdr_params_t additional_params[IOC_FM_PCD_PRS_NUM_OF_HDRS];
-+ /**< 'num_of_hdrs_with_additional_params' structures
-+ additional parameters for each header that requires them */
-+ bool set_vlan_tpid1; /**< TRUE to configure user selection of Ethertype to
-+ indicate a VLAN tag (in addition to the TPID values
-+ 0x8100 and 0x88A8). */
-+ uint16_t vlan_tpid1; /**< extra tag to use if set_vlan_tpid1=TRUE. */
-+ bool set_vlan_tpid2; /**< TRUE to configure user selection of Ethertype to
-+ indicate a VLAN tag (in addition to the TPID values
-+ 0x8100 and 0x88A8). */
-+ uint16_t vlan_tpid2; /**< extra tag to use if set_vlan_tpid1=TRUE. */
-+} ioc_fm_port_pcd_prs_params_t;
-+
-+/**************************************************************************//**
-+ @Description A structure for defining coarse alassification parameters
-+ (Must match t_FmPortPcdCcParams defined in fm_port_ext.h)
-+*//***************************************************************************/
-+typedef struct ioc_fm_port_pcd_cc_params_t {
-+ void *cc_tree_id; /**< CC tree id */
-+} ioc_fm_port_pcd_cc_params_t;
-+
-+/**************************************************************************//**
-+ @Description A structure for defining keygen parameters
-+ (Must match t_FmPortPcdKgParams defined in fm_port_ext.h)
-+*//***************************************************************************/
-+typedef struct ioc_fm_port_pcd_kg_params_t {
-+ uint8_t num_of_schemes; /**< Number of schemes for port to be bound to. */
-+ void *scheme_ids[FM_PCD_KG_NUM_OF_SCHEMES];
-+ /**< Array of 'num_of_schemes' schemes for the
-+ port to be bound to */
-+ bool direct_scheme; /**< TRUE for going from parser to a specific scheme,
-+ regardless of parser result */
-+ void *direct_scheme_id; /**< Scheme id, as returned by FM_PCD_KgSetScheme;
-+ relevant only if direct=TRUE. */
-+} ioc_fm_port_pcd_kg_params_t;
-+
-+/**************************************************************************//**
-+ @Description A structure for defining policer parameters
-+ (Must match t_FmPortPcdPlcrParams defined in fm_port_ext.h)
-+*//***************************************************************************/
-+typedef struct ioc_fm_port_pcd_plcr_params_t {
-+ void *plcr_profile_id; /**< Selected profile handle;
-+ relevant in one of the following cases:
-+ e_IOC_FM_PORT_PCD_SUPPORT_PLCR_ONLY or
-+ e_IOC_FM_PORT_PCD_SUPPORT_PRS_AND_PLCR were selected,
-+ or if any flow uses a KG scheme where policer
-+ profile is not generated (bypass_plcr_profile_generation selected) */
-+} ioc_fm_port_pcd_plcr_params_t;
-+
-+/**************************************************************************//**
-+ @Description A structure for defining port PCD parameters
-+ (Must match struct t_FmPortPcdParams defined in fm_port_ext.h)
-+*//***************************************************************************/
-+typedef struct ioc_fm_port_pcd_params_t {
-+ ioc_fm_port_pcd_support pcd_support; /**< Relevant for Rx and offline ports only.
-+ Describes the active PCD engines for this port. */
-+ void *net_env_id; /**< HL Unused in PLCR only mode */
-+ ioc_fm_port_pcd_prs_params_t *p_prs_params; /**< Parser parameters for this port */
-+ ioc_fm_port_pcd_cc_params_t *p_cc_params; /**< Coarse classification parameters for this port */
-+ ioc_fm_port_pcd_kg_params_t *p_kg_params; /**< Keygen parameters for this port */
-+ ioc_fm_port_pcd_plcr_params_t *p_plcr_params; /**< Policer parameters for this port */
-+ void *p_ip_reassembly_manip;/**< IP Reassembly manipulation */
-+#if (DPAA_VERSION >= 11)
-+ void *p_capwap_reassembly_manip;/**< CAPWAP Reassembly manipulation */
-+#endif /* (DPAA_VERSION >= 11) */
-+} ioc_fm_port_pcd_params_t;
-+
-+/**************************************************************************//**
-+ @Description A structure for defining the Parser starting point
-+ (Must match struct t_FmPcdPrsStart defined in fm_port_ext.h)
-+*//***************************************************************************/
-+typedef struct ioc_fm_pcd_prs_start_t {
-+ uint8_t parsing_offset; /**< Number of bytes from begining of packet to
-+ start parsing */
-+ ioc_net_header_type first_prs_hdr; /**< The type of the first header axpected at
-+ 'parsing_offset' */
-+} ioc_fm_pcd_prs_start_t;
-+
-+
-+/**************************************************************************//**
-+ @Description FQID parameters structure
-+*//***************************************************************************/
-+typedef struct ioc_fm_port_pcd_fqids_params_t {
-+ uint32_t num_fqids; /**< Number of fqids to be allocated for the port */
-+ uint8_t alignment; /**< Alignment required for this port */
-+ uint32_t base_fqid; /**< output parameter - the base fqid */
-+} ioc_fm_port_pcd_fqids_params_t;
-+
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_IOC_ALLOC_PCD_FQIDS
-+
-+ @Description Allocates FQID's
-+
-+ May be used for Rx and offline parsing ports only
-+
-+ @Param[in,out] ioc_fm_port_pcd_fqids_params_t Parameters for allocating FQID's
-+
-+ @Return 0 on success; error code otherwise.
-+*//***************************************************************************/
-+#define FM_PORT_IOC_ALLOC_PCD_FQIDS _IOWR(FM_IOC_TYPE_BASE, FM_PORT_IOC_NUM(19), ioc_fm_port_pcd_fqids_params_t)
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_IOC_FREE_PCD_FQIDS
-+
-+ @Description Frees previously-allocated FQIDs
-+
-+ May be used for Rx and offline parsing ports only
-+
-+ @Param[in] uint32_t Base FQID of previously allocated range.
-+
-+ @Return 0 on success; error code otherwise.
-+*//***************************************************************************/
-+#define FM_PORT_IOC_FREE_PCD_FQIDS _IOW(FM_IOC_TYPE_BASE, FM_PORT_IOC_NUM(19), uint32_t)
-+
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_SetPCD
-+
-+ @Description Calling this routine defines the port's PCD configuration.
-+ It changes it from its default configuration which is PCD
-+ disabled (BMI to BMI) and configures it according to the passed
-+ parameters.
-+
-+ May be used for Rx and offline parsing ports only
-+
-+ @Param[in] ioc_fm_port_pcd_params_t A Structure of parameters defining the port's PCD
-+ configuration.
-+
-+ @Return 0 on success; error code otherwise.
-+*//***************************************************************************/
-+#if defined(CONFIG_COMPAT)
-+#define FM_PORT_IOC_SET_PCD_COMPAT _IOW(FM_IOC_TYPE_BASE, FM_PORT_IOC_NUM(20), ioc_compat_fm_port_pcd_params_t)
-+#endif
-+#define FM_PORT_IOC_SET_PCD _IOW(FM_IOC_TYPE_BASE, FM_PORT_IOC_NUM(20), ioc_fm_port_pcd_params_t)
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_DeletePCD
-+
-+ @Description Calling this routine releases the port's PCD configuration.
-+ The port returns to its default configuration which is PCD
-+ disabled (BMI to BMI) and all PCD configuration is removed.
-+
-+ May be used for Rx and offline parsing ports which are
-+ in PCD mode only
-+
-+ @Return 0 on success; error code otherwise.
-+*//***************************************************************************/
-+#define FM_PORT_IOC_DELETE_PCD _IO(FM_IOC_TYPE_BASE, FM_PORT_IOC_NUM(21))
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_AttachPCD
-+
-+ @Description This routine may be called after FM_PORT_DetachPCD was called,
-+ to return to the originally configured PCD support flow.
-+ The couple of routines are used to allow PCD configuration changes
-+ that demand that PCD will not be used while changes take place.
-+
-+ May be used for Rx and offline parsing ports which are
-+ in PCD mode only
-+
-+ @Return 0 on success; error code otherwise.
-+*//***************************************************************************/
-+#define FM_PORT_IOC_ATTACH_PCD _IO(FM_IOC_TYPE_BASE, FM_PORT_IOC_NUM(23))
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_DetachPCD
-+
-+ @Description Calling this routine detaches the port from its PCD functionality.
-+ The port returns to its default flow which is BMI to BMI.
-+
-+ May be used for Rx and offline parsing ports which are
-+ in PCD mode only
-+
-+ @Return 0 on success; error code otherwise.
-+*//***************************************************************************/
-+#define FM_PORT_IOC_DETACH_PCD _IO(FM_IOC_TYPE_BASE, FM_PORT_IOC_NUM(22))
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_PcdPlcrAllocProfiles
-+
-+ @Description This routine may be called only for ports that use the Policer in
-+ order to allocate private policer profiles.
-+
-+ @Param[in] uint16_t The number of required policer profiles
-+
-+ @Return 0 on success; error code otherwise.
-+
-+ @Cautions Allowed before FM_PORT_SetPCD() only.
-+*//***************************************************************************/
-+#define FM_PORT_IOC_PCD_PLCR_ALLOC_PROFILES _IOW(FM_IOC_TYPE_BASE, FM_PORT_IOC_NUM(24), uint16_t)
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_PcdPlcrFreeProfiles
-+
-+ @Description This routine should be called for freeing private policer profiles.
-+
-+ @Return 0 on success; error code otherwise.
-+
-+ @Cautions Allowed before FM_PORT_SetPCD() only.
-+*//***************************************************************************/
-+#define FM_PORT_IOC_PCD_PLCR_FREE_PROFILES _IO(FM_IOC_TYPE_BASE, FM_PORT_IOC_NUM(25))
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_PcdKgModifyInitialScheme
-+
-+ @Description This routine may be called only for ports that use the keygen in
-+ order to change the initial scheme frame should be routed to.
-+ The change may be of a scheme id (in case of direct mode),
-+ from direct to indirect, or from indirect to direct - specifying the scheme id.
-+
-+ @Param[in] ioc_fm_pcd_kg_scheme_select_t A structure of parameters for defining whether
-+ a scheme is direct/indirect, and if direct - scheme id.
-+
-+ @Return 0 on success; error code otherwise.
-+*//***************************************************************************/
-+#if defined(CONFIG_COMPAT)
-+#define FM_PORT_IOC_PCD_KG_MODIFY_INITIAL_SCHEME_COMPAT _IOW(FM_IOC_TYPE_BASE, FM_PORT_IOC_NUM(26), ioc_compat_fm_pcd_kg_scheme_select_t)
-+#endif
-+#define FM_PORT_IOC_PCD_KG_MODIFY_INITIAL_SCHEME _IOW(FM_IOC_TYPE_BASE, FM_PORT_IOC_NUM(26), ioc_fm_pcd_kg_scheme_select_t)
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_PcdPlcrModifyInitialProfile
-+
-+ @Description This routine may be called for ports with flows
-+ e_IOC_FM_PCD_SUPPORT_PLCR_ONLY or e_IOC_FM_PCD_SUPPORT_PRS_AND_PLCR only,
-+ to change the initial Policer profile frame should be routed to.
-+ The change may be of a profile and/or absolute/direct mode selection.
-+
-+ @Param[in] ioc_fm_obj_t Policer profile Id as returned from FM_PCD_PlcrSetProfile.
-+
-+ @Return 0 on success; error code otherwise.
-+*//***************************************************************************/
-+#if defined(CONFIG_COMPAT)
-+#define FM_PORT_IOC_PCD_PLCR_MODIFY_INITIAL_PROFILE_COMPAT _IOW(FM_IOC_TYPE_BASE, FM_PORT_IOC_NUM(27), ioc_compat_fm_obj_t)
-+#endif
-+#define FM_PORT_IOC_PCD_PLCR_MODIFY_INITIAL_PROFILE _IOW(FM_IOC_TYPE_BASE, FM_PORT_IOC_NUM(27), ioc_fm_obj_t)
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_PcdCcModifyTree
-+
-+ @Description This routine may be called to change this port connection to
-+ a pre-initializes coarse classification Tree.
-+
-+ @Param[in] ioc_fm_obj_t Id of new coarse classification tree selected for this port.
-+
-+ @Return 0 on success; error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_SetPCD() and FM_PORT_DetachPCD()
-+*//***************************************************************************/
-+#if defined(CONFIG_COMPAT)
-+#define FM_PORT_IOC_PCD_CC_MODIFY_TREE_COMPAT _IOW(FM_IOC_TYPE_BASE, FM_PORT_IOC_NUM(28), ioc_compat_fm_obj_t)
-+#endif
-+#define FM_PORT_IOC_PCD_CC_MODIFY_TREE _IOW(FM_IOC_TYPE_BASE, FM_PORT_IOC_NUM(28), ioc_fm_obj_t)
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_PcdKgBindSchemes
-+
-+ @Description These routines may be called for modifying the binding of ports
-+ to schemes. The scheme itself is not added,
-+ just this specific port starts using it.
-+
-+ @Param[in] ioc_fm_pcd_port_schemes_params_t Schemes parameters structre
-+
-+ @Return 0 on success; error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_SetPCD().
-+*//***************************************************************************/
-+#if defined(CONFIG_COMPAT)
-+#define FM_PORT_IOC_PCD_KG_BIND_SCHEMES_COMPAT _IOW(FM_IOC_TYPE_BASE, FM_PORT_IOC_NUM(30), ioc_compat_fm_pcd_port_schemes_params_t)
-+#endif
-+#define FM_PORT_IOC_PCD_KG_BIND_SCHEMES _IOW(FM_IOC_TYPE_BASE, FM_PORT_IOC_NUM(30), ioc_fm_pcd_port_schemes_params_t)
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_PcdKgUnbindSchemes
-+
-+ @Description These routines may be called for modifying the binding of ports
-+ to schemes. The scheme itself is not removed or invalidated,
-+ just this specific port stops using it.
-+
-+ @Param[in] ioc_fm_pcd_port_schemes_params_t Schemes parameters structre
-+
-+ @Return 0 on success; error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_SetPCD().
-+*//***************************************************************************/
-+#if defined(CONFIG_COMPAT)
-+#define FM_PORT_IOC_PCD_KG_UNBIND_SCHEMES_COMPAT _IOW(FM_IOC_TYPE_BASE, FM_PORT_IOC_NUM(31), ioc_compat_fm_pcd_port_schemes_params_t)
-+#endif
-+#define FM_PORT_IOC_PCD_KG_UNBIND_SCHEMES _IOW(FM_IOC_TYPE_BASE, FM_PORT_IOC_NUM(31), ioc_fm_pcd_port_schemes_params_t)
-+
-+typedef struct ioc_fm_port_mac_addr_params_t {
-+ uint8_t addr[ENET_NUM_OCTETS_PER_ADDRESS];
-+} ioc_fm_port_mac_addr_params_t;
-+
-+/**************************************************************************//**
-+ @Function FM_MAC_AddHashMacAddr
-+
-+ @Description Add an Address to the hash table. This is for filter purpose only.
-+
-+ @Param[in] ioc_fm_port_mac_addr_params_t - Ethernet Mac address
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MAC_Init(). It is a filter only address.
-+ @Cautions Some address need to be filtered out in upper FM blocks.
-+*//***************************************************************************/
-+#define FM_PORT_IOC_ADD_RX_HASH_MAC_ADDR _IOW(FM_IOC_TYPE_BASE, FM_PORT_IOC_NUM(36), ioc_fm_port_mac_addr_params_t)
-+
-+/**************************************************************************//**
-+ @Function FM_MAC_RemoveHashMacAddr
-+
-+ @Description Delete an Address to the hash table. This is for filter purpose only.
-+
-+ @Param[in] ioc_fm_port_mac_addr_params_t - Ethernet Mac address
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MAC_Init().
-+*//***************************************************************************/
-+#define FM_PORT_IOC_REMOVE_RX_HASH_MAC_ADDR _IOW(FM_IOC_TYPE_BASE, FM_PORT_IOC_NUM(37), ioc_fm_port_mac_addr_params_t)
-+
-+typedef struct ioc_fm_port_tx_pause_frames_params_t {
-+ uint8_t priority;
-+ uint16_t pause_time;
-+ uint16_t thresh_time;
-+} ioc_fm_port_tx_pause_frames_params_t;
-+
-+/**************************************************************************//**
-+ @Function FM_MAC_SetTxPauseFrames
-+
-+ @Description Enable/Disable transmission of Pause-Frames.
-+ The routine changes the default configuration:
-+ pause-time - [0xf000]
-+ threshold-time - [0]
-+
-+ @Param[in] ioc_fm_port_tx_pause_frames_params_t A structure holding the required parameters.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_MAC_Init().
-+ PFC is supported only on new mEMAC; i.e. in MACs that don't have
-+ PFC support (10G-MAC and dTSEC), user should use 'FM_MAC_NO_PFC'
-+ in the 'priority' field.
-+*//***************************************************************************/
-+#define FM_PORT_IOC_SET_TX_PAUSE_FRAMES _IOW(FM_IOC_TYPE_BASE, FM_PORT_IOC_NUM(40), ioc_fm_port_tx_pause_frames_params_t)
-+
-+typedef struct ioc_fm_port_mac_statistics_t {
-+ /* RMON */
-+ uint64_t e_stat_pkts_64; /**< r-10G tr-DT 64 byte frame counter */
-+ uint64_t e_stat_pkts_65_to_127; /**< r-10G 65 to 127 byte frame counter */
-+ uint64_t e_stat_pkts_128_to_255; /**< r-10G 128 to 255 byte frame counter */
-+ uint64_t e_stat_pkts_256_to_511; /**< r-10G 256 to 511 byte frame counter */
-+ uint64_t e_stat_pkts_512_to_1023; /**< r-10G 512 to 1023 byte frame counter */
-+ uint64_t e_stat_pkts_1024_to_1518; /**< r-10G 1024 to 1518 byte frame counter */
-+ uint64_t e_stat_pkts_1519_to_1522; /**< r-10G 1519 to 1522 byte good frame count */
-+ /* */
-+ uint64_t e_stat_fragments; /**< Total number of packets that were less than 64 octets long with a wrong CRC.*/
-+ uint64_t e_stat_jabbers; /**< Total number of packets longer than valid maximum length octets */
-+ uint64_t e_stat_drop_events; /**< number of dropped packets due to internal errors of the MAC Client (during recieve). */
-+ uint64_t e_stat_CRC_align_errors; /**< Incremented when frames of correct length but with CRC error are received.*/
-+ uint64_t e_stat_undersize_pkts; /**< Incremented for frames under 64 bytes with a valid FCS and otherwise well formed;
-+ This count does not include range length errors */
-+ uint64_t e_stat_oversize_pkts; /**< Incremented for frames which exceed 1518 (non VLAN) or 1522 (VLAN) and contains
-+ a valid FCS and otherwise well formed */
-+ /* Pause */
-+ uint64_t te_stat_pause; /**< Pause MAC Control received */
-+ uint64_t re_stat_pause; /**< Pause MAC Control sent */
-+ /* MIB II */
-+ uint64_t if_in_octets; /**< Total number of byte received. */
-+ uint64_t if_in_pkts; /**< Total number of packets received.*/
-+ uint64_t if_in_ucast_pkts; /**< Total number of unicast frame received;
-+ NOTE: this counter is not supported on dTSEC MAC */
-+ uint64_t if_in_mcast_pkts; /**< Total number of multicast frame received*/
-+ uint64_t if_in_bcast_pkts; /**< Total number of broadcast frame received */
-+ uint64_t if_in_discards; /**< Frames received, but discarded due to problems within the MAC RX. */
-+ uint64_t if_in_errors; /**< Number of frames received with error:
-+ - FIFO Overflow Error
-+ - CRC Error
-+ - Frame Too Long Error
-+ - Alignment Error
-+ - The dedicated Error Code (0xfe, not a code error) was received */
-+ uint64_t if_out_octets; /**< Total number of byte sent. */
-+ uint64_t if_out_pkts; /**< Total number of packets sent .*/
-+ uint64_t if_out_ucast_pkts; /**< Total number of unicast frame sent;
-+ NOTE: this counter is not supported on dTSEC MAC */
-+ uint64_t if_out_mcast_pkts; /**< Total number of multicast frame sent */
-+ uint64_t if_out_bcast_pkts; /**< Total number of multicast frame sent */
-+ uint64_t if_out_discards; /**< Frames received, but discarded due to problems within the MAC TX N/A!.*/
-+ uint64_t if_out_errors; /**< Number of frames transmitted with error:
-+ - FIFO Overflow Error
-+ - FIFO Underflow Error
-+ - Other */
-+} ioc_fm_port_mac_statistics_t;
-+
-+/**************************************************************************//**
-+ @Function FM_MAC_GetStatistics
-+
-+ @Description get all MAC statistics counters
-+
-+ @Param[out] ioc_fm_port_mac_statistics_t A structure holding the statistics
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_Init().
-+*//***************************************************************************/
-+#define FM_PORT_IOC_GET_MAC_STATISTICS _IOR(FM_IOC_TYPE_BASE, FM_PORT_IOC_NUM(41), ioc_fm_port_mac_statistics_t)
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_ConfigBufferPrefixContent
-+
-+ @Description Defines the structure, size and content of the application buffer.
-+ The prefix will
-+ In Tx ports, if 'passPrsResult', the application
-+ should set a value to their offsets in the prefix of
-+ the FM will save the first 'privDataSize', than,
-+ depending on 'passPrsResult' and 'passTimeStamp', copy parse result
-+ and timeStamp, and the packet itself (in this order), to the
-+ application buffer, and to offset.
-+ Calling this routine changes the buffer margins definitions
-+ in the internal driver data base from its default
-+ configuration: Data size: [DEFAULT_FM_SP_bufferPrefixContent_privDataSize]
-+ Pass Parser result: [DEFAULT_FM_SP_bufferPrefixContent_passPrsResult].
-+ Pass timestamp: [DEFAULT_FM_SP_bufferPrefixContent_passTimeStamp].
-+
-+ May be used for all ports
-+
-+ @Param[in] ioc_fm_buffer_prefix_content_t A structure holding the required parameters.
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Config() and before FM_PORT_Init().
-+*//***************************************************************************/
-+#define FM_PORT_IOC_CONFIG_BUFFER_PREFIX_CONTENT _IOW(FM_IOC_TYPE_BASE, FM_PORT_IOC_NUM(39), ioc_fm_buffer_prefix_content_t)
-+
-+#if (DPAA_VERSION >= 11)
-+typedef struct ioc_fm_port_vsp_alloc_params_t {
-+ uint8_t num_of_profiles; /**< Number of Virtual Storage Profiles */
-+ uint8_t dflt_relative_id; /**< The default Virtual-Storage-Profile-id dedicated to Rx/OP port
-+ The same default Virtual-Storage-Profile-id will be for coupled Tx port
-+ if relevant function called for Rx port */
-+ void *p_fm_tx_port; /**< Handle to coupled Tx Port; not relevant for OP port. */
-+}ioc_fm_port_vsp_alloc_params_t;
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_VSPAlloc
-+
-+ @Description This routine allocated VSPs per port and forces the port to work
-+ in VSP mode. Note that the port is initialized by default with the
-+ physical-storage-profile only.
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[in] p_Params A structure of parameters for allocation VSP's per port
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Init(), and before FM_PORT_SetPCD()
-+ and also before FM_PORT_Enable() (i.e. the port should be disabled).
-+*//***************************************************************************/
-+#if defined(CONFIG_COMPAT)
-+#define FM_PORT_IOC_VSP_ALLOC_COMPAT _IOW(FM_IOC_TYPE_BASE, FM_PORT_IOC_NUM(38), ioc_compat_fm_port_vsp_alloc_params_t)
-+#endif
-+#define FM_PORT_IOC_VSP_ALLOC _IOW(FM_IOC_TYPE_BASE, FM_PORT_IOC_NUM(38), ioc_fm_port_vsp_alloc_params_t)
-+#endif /* (DPAA_VERSION >= 11) */
-+
-+/**************************************************************************//**
-+ @Function FM_PORT_GetBmiCounters
-+
-+ @Description Read port's BMI stat counters and place them into
-+ a designated structure of counters.
-+
-+ @Param[in] h_FmPort A handle to a FM Port module.
-+ @Param[out] p_BmiStats counters structure
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_PORT_Init().
-+*//***************************************************************************/
-+
-+#define FM_PORT_IOC_GET_BMI_COUNTERS _IOR(FM_IOC_TYPE_BASE, FM_PORT_IOC_NUM(42), ioc_fm_port_bmi_stats_t)
-+
-+typedef struct ioc_fm_port_mac_frame_size_counters_t {
-+
-+ e_CommMode type;
-+ uint64_t count_pkts_64; /**< 64 byte frame counter */
-+ uint64_t count_pkts_65_to_127; /**< 65 to 127 byte frame counter */
-+ uint64_t count_pkts_128_to_255; /**< 128 to 255 byte frame counter */
-+ uint64_t count_pkts_256_to_511; /**< 256 to 511 byte frame counter */
-+ uint64_t count_pkts_512_to_1023; /**< 512 to 1023 byte frame counter */
-+ uint64_t count_pkts_1024_to_1518; /**< 1024 to 1518 byte frame counter */
-+ uint64_t count_pkts_1519_to_1522; /**< 1519 to 1522 byte good frame count */
-+} ioc_fm_port_mac_frame_size_counters_t;
-+
-+/**************************************************************************//**
-+ @Function FM_MAC_GetFrameSizeCounters
-+
-+ @Description get MAC statistics counters for different frame size
-+
-+ @Param[out] ioc_fm_port_mac_frame_size_counters_t A structure holding the counters
-+
-+ @Return E_OK on success; Error code otherwise.
-+
-+ @Cautions Allowed only following FM_Init().
-+*//***************************************************************************/
-+#define FM_PORT_IOC_GET_MAC_FRAME_SIZE_COUNTERS _IOR(FM_IOC_TYPE_BASE, FM_PORT_IOC_NUM(43), ioc_fm_port_mac_frame_size_counters_t)
-+
-+
-+/** @} */ /* end of lnx_ioctl_FM_PORT_pcd_runtime_control_grp group */
-+/** @} */ /* end of lnx_ioctl_FM_PORT_runtime_control_grp group */
-+
-+/** @} */ /* end of lnx_ioctl_FM_PORT_grp group */
-+/** @} */ /* end of lnx_ioctl_FM_grp group */
-+#endif /* __FM_PORT_IOCTLS_H */
---- /dev/null
-+++ b/include/uapi/linux/fmd/Peripherals/fm_test_ioctls.h
-@@ -0,0 +1,208 @@
-+/* Copyright (c) 2008-2012 Freescale Semiconductor, Inc.
-+ * All rights reserved.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+/**************************************************************************//**
-+ @File fm_test_ioctls.h
-+
-+ @Description FM Char device ioctls
-+*//***************************************************************************/
-+#ifndef __FM_TEST_IOCTLS_H
-+#define __FM_TEST_IOCTLS_H
-+
-+#include "ioctls.h"
-+
-+
-+/**************************************************************************//**
-+ @Group lnx_ioctl_FMT_grp Frame Manager Test Linux IOCTL API
-+
-+ @Description FM-Test Linux ioctls definitions and enums
-+
-+ @{
-+*//***************************************************************************/
-+
-+#define IOC_FMT_MAX_NUM_OF_PORTS 26
-+
-+/**************************************************************************//**
-+ @Collection TEST Parameters
-+*//***************************************************************************/
-+/**************************************************************************//**
-+ @Description: Name of the FM-Test chardev
-+*//***************************************************************************/
-+#define DEV_FM_TEST_NAME "fm-test-port"
-+
-+#define DEV_FM_TEST_PORTS_MINOR_BASE 0
-+#define DEV_FM_TEST_MAX_MINORS (DEV_FM_TEST_PORTS_MINOR_BASE + IOC_FMT_MAX_NUM_OF_PORTS)
-+
-+#define FMT_PORT_IOC_NUM(n) n
-+/* @} */
-+
-+/**************************************************************************//**
-+ @Group lnx_ioctl_FMT_lib_grp FM-Test library
-+
-+ @Description TODO
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Description TODO
-+*//***************************************************************************/
-+typedef uint8_t ioc_fmt_xxx_t;
-+
-+#define FM_PRS_MAX 32
-+#define FM_TIME_STAMP_MAX 8
-+
-+/**************************************************************************//**
-+ @Description FM Port buffer content description
-+*//***************************************************************************/
-+typedef struct ioc_fmt_buff_context_t {
-+ void *p_user_priv;
-+ uint8_t fm_prs_res[FM_PRS_MAX];
-+ uint8_t fm_time_stamp[FM_TIME_STAMP_MAX];
-+} ioc_fmt_buff_context_t;
-+
-+#if defined(__KERNEL__) && defined(CONFIG_COMPAT)
-+typedef struct ioc_fmt_compat_buff_context_t {
-+ compat_uptr_t p_user_priv;
-+ uint8_t fm_prs_res[FM_PRS_MAX];
-+ uint8_t fm_time_stamp[FM_TIME_STAMP_MAX];
-+} ioc_fmt_compat_buff_context_t;
-+#endif
-+
-+/**************************************************************************//**
-+ @Description Buffer descriptor
-+*//***************************************************************************/
-+typedef struct ioc_fmt_buff_desc_t {
-+ uint32_t qid;
-+ void *p_data;
-+ uint32_t size;
-+ uint32_t status;
-+ ioc_fmt_buff_context_t buff_context;
-+} ioc_fmt_buff_desc_t;
-+
-+#if defined(__KERNEL__) && defined(CONFIG_COMPAT)
-+typedef struct ioc_fmt_compat_buff_desc_t {
-+ uint32_t qid;
-+ compat_uptr_t p_data;
-+ uint32_t size;
-+ uint32_t status;
-+ ioc_fmt_compat_buff_context_t buff_context;
-+} ioc_fmt_compat_buff_desc_t;
-+#endif
-+
-+/**************************************************************************//**
-+ @Group lnx_ioctl_FMT_runtime_control_grp FM-Test Runtime Control Unit
-+
-+ @Description TODO
-+ @{
-+*//***************************************************************************/
-+
-+/** @} */ /* end of lnx_ioctl_FMT_runtime_control_grp group */
-+
-+
-+/**************************************************************************//**
-+ @Group lnx_ioctl_FMTP_lib_grp FM-Port-Test library
-+
-+ @Description TODO
-+
-+ @{
-+*//***************************************************************************/
-+
-+/**************************************************************************//**
-+ @Description FM-Test FM port type
-+*//***************************************************************************/
-+typedef enum ioc_fmt_port_type {
-+ e_IOC_FMT_PORT_T_RXTX, /**< Standard port */
-+ e_IOC_FMT_PORT_T_OP, /**< Offline-parsing port */
-+} ioc_fmt_port_type;
-+
-+/**************************************************************************//**
-+ @Description TODO
-+*//***************************************************************************/
-+typedef struct ioc_fmt_port_param_t {
-+ uint8_t fm_id;
-+ ioc_fmt_port_type fm_port_type;
-+ uint8_t fm_port_id;
-+ uint32_t num_tx_queues;
-+} ioc_fmt_port_param_t;
-+
-+
-+/**************************************************************************//**
-+ @Function FMT_PORT_IOC_INIT
-+
-+ @Description TODO
-+
-+ @Param[in] ioc_fmt_port_param_t TODO
-+
-+ @Cautions Allowed only after the FM equivalent port is already initialized.
-+*//***************************************************************************/
-+#define FMT_PORT_IOC_INIT _IOW(FMT_IOC_TYPE_BASE, FMT_PORT_IOC_NUM(0), ioc_fmt_port_param_t)
-+
-+/**************************************************************************//**
-+ @Function FMT_PORT_IOC_SET_DIAG_MODE
-+
-+ @Description TODO
-+
-+ @Param[in] ioc_diag_mode TODO
-+
-+ @Cautions Allowed only following FMT_PORT_IOC_INIT().
-+*//***************************************************************************/
-+#define FMT_PORT_IOC_SET_DIAG_MODE _IOW(FMT_IOC_TYPE_BASE, FMT_PORT_IOC_NUM(1), ioc_diag_mode)
-+
-+/**************************************************************************//**
-+ @Function FMT_PORT_IOC_SET_IP_HEADER_MANIP
-+
-+ @Description Set IP header manipulations for this port.
-+
-+ @Param[in] int 1 to enable; 0 to disable
-+
-+ @Cautions Allowed only following FMT_PORT_IOC_INIT().
-+*//***************************************************************************/
-+#define FMT_PORT_IOC_SET_IP_HEADER_MANIP _IOW(FMT_IOC_TYPE_BASE, FMT_PORT_IOC_NUM(2), int)
-+
-+/**************************************************************************//**
-+ @Function FMT_PORT_IOC_SET_DPAECHO_MODE
-+
-+ @Description Set DPA in echo mode - all frame are sent back.
-+
-+ @Param[in] int 1 to enable; 0 to disable
-+
-+ @Cautions Allowed only following FMT_PORT_IOC_INIT().
-+*//***************************************************************************/
-+#define FMT_PORT_IOC_SET_DPAECHO_MODE _IOW(FMT_IOC_TYPE_BASE, FMT_PORT_IOC_NUM(3), int)
-+
-+/** @} */ /* end of lnx_ioctl_FMTP_lib_grp group */
-+/** @} */ /* end of lnx_ioctl_FMT_lib_grp group */
-+/** @} */ /* end of lnx_ioctl_FMT_grp */
-+
-+
-+#endif /* __FM_TEST_IOCTLS_H */
---- /dev/null
-+++ b/include/uapi/linux/fmd/integrations/Kbuild
-@@ -0,0 +1 @@
-+header-y += integration_ioctls.h
---- /dev/null
-+++ b/include/uapi/linux/fmd/integrations/integration_ioctls.h
-@@ -0,0 +1,56 @@
-+/* Copyright (c) 2008-2012 Freescale Semiconductor, Inc.
-+ * All rights reserved.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+/**************************************************************************//**
-+ @File integration_ioctls.h
-+
-+ @Description External header file for Integration unit routines.
-+*//***************************************************************************/
-+
-+#ifndef __INTG_IOCTLS_H
-+#define __INTG_IOCTLS_H
-+
-+
-+#define FM_IOC_TYPE_BASE (NCSW_IOC_TYPE_BASE+1)
-+#define FMT_IOC_TYPE_BASE (NCSW_IOC_TYPE_BASE+3)
-+
-+/*#define FM_IOCTL_DBG*/
-+
-+#if defined(FM_IOCTL_DBG)
-+ #define _fm_ioctl_dbg(format, arg...) \
-+ printk("fm ioctl [%s:%u](cpu:%u) - " format, \
-+ __func__, __LINE__, smp_processor_id(), ##arg)
-+#else
-+# define _fm_ioctl_dbg(arg...)
-+#endif
-+
-+#endif /* __INTG_IOCTLS_H */
---- /dev/null
-+++ b/include/uapi/linux/fmd/ioctls.h
-@@ -0,0 +1,96 @@
-+/* Copyright (c) 2008-2012 Freescale Semiconductor, Inc.
-+ * All rights reserved.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+/**************************************************************************//**
-+ @File ioctls.h
-+
-+ @Description Structures and definitions for Command Relay Ioctls
-+*//***************************************************************************/
-+
-+#ifndef __IOCTLS_H__
-+#define __IOCTLS_H__
-+
-+#include <asm/ioctl.h>
-+
-+#include "integration_ioctls.h"
-+
-+
-+/**************************************************************************//**
-+ @Group lnx_ioctl_ncsw_grp NetCommSw Linux User-Space (IOCTL) API
-+ @{
-+*//***************************************************************************/
-+
-+#define NCSW_IOC_TYPE_BASE 0xe0 /**< defines the IOCTL type for all
-+ the NCSW Linux module commands */
-+
-+
-+/**************************************************************************//**
-+ @Description IOCTL Memory allocation types.
-+*//***************************************************************************/
-+typedef enum ioc_mem_type {
-+ e_IOC_MEM_INVALID = 0x00000000, /**< Invalid memory type (error) */
-+ e_IOC_MEM_CACHABLE_SYS = 0x00000001, /**< Primary DDR, cacheable segment */
-+ e_IOC_MEM_NOCACHE_SYS = 0x00000004, /**< Primary DDR, non-cacheable segment */
-+ e_IOC_MEM_SECONDARY = 0x00000002, /**< Either secondary DDR or SDRAM */
-+ e_IOC_MEM_PRAM = 0x00000008 /**< Multi-user RAM identifier */
-+} ioc_mem_type;
-+
-+/**************************************************************************//**
-+ @Description Enumeration (bit flags) of communication modes (Transmit,
-+ receive or both).
-+*//***************************************************************************/
-+typedef enum ioc_comm_mode {
-+ e_IOC_COMM_MODE_NONE = 0 /**< No transmit/receive communication */
-+ , e_IOC_COMM_MODE_RX = 1 /**< Only receive communication */
-+ , e_IOC_COMM_MODE_TX = 2 /**< Only transmit communication */
-+ , e_IOC_COMM_MODE_RX_AND_TX = 3 /**< Both transmit and receive communication */
-+} ioc_comm_mode;
-+
-+/**************************************************************************//**
-+ @Description General Diagnostic Mode
-+*//***************************************************************************/
-+typedef enum ioc_diag_mode
-+{
-+ e_IOC_DIAG_MODE_NONE = 0,
-+ e_IOC_DIAG_MODE_CTRL_LOOPBACK, /**< loopback in the controller; E.g. MAC, TDM, etc. */
-+ e_IOC_DIAG_MODE_CHIP_LOOPBACK, /**< loopback in the chip but not in controller;
-+ E.g. IO-pins, SerDes, etc. */
-+ e_IOC_DIAG_MODE_PHY_LOOPBACK, /**< loopback in the external PHY */
-+ e_IOC_DIAG_MODE_LINE_LOOPBACK, /**< loopback in the external line */
-+ e_IOC_DIAG_MODE_CTRL_ECHO, /**< */
-+ e_IOC_DIAG_MODE_PHY_ECHO /**< */
-+} ioc_diag_mode;
-+
-+/** @} */ /* end of lnx_ioctl_ncsw_grp */
-+
-+
-+#endif /* __IOCTLS_H__ */
---- /dev/null
-+++ b/include/uapi/linux/fmd/net_ioctls.h
-@@ -0,0 +1,430 @@
-+/* Copyright (c) 2008-2012 Freescale Semiconductor, Inc.
-+ * All rights reserved.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+
-+/**************************************************************************//**
-+ @File net_ioctls.h
-+
-+ @Description This file contains common and general netcomm headers definitions.
-+*//***************************************************************************/
-+#ifndef __NET_IOCTLS_H
-+#define __NET_IOCTLS_H
-+
-+#include "ioctls.h"
-+
-+
-+typedef uint8_t ioc_header_field_ppp_t;
-+
-+#define IOC_NET_HEADER_FIELD_PPP_PID (1)
-+#define IOC_NET_HEADER_FIELD_PPP_COMPRESSED (IOC_NET_HEADER_FIELD_PPP_PID << 1)
-+#define IOC_NET_HEADER_FIELD_PPP_ALL_FIELDS ((IOC_NET_HEADER_FIELD_PPP_PID << 2) - 1)
-+
-+
-+typedef uint8_t ioc_header_field_pppoe_t;
-+
-+#define IOC_NET_HEADER_FIELD_PPPoE_VER (1)
-+#define IOC_NET_HEADER_FIELD_PPPoE_TYPE (IOC_NET_HEADER_FIELD_PPPoE_VER << 1)
-+#define IOC_NET_HEADER_FIELD_PPPoE_CODE (IOC_NET_HEADER_FIELD_PPPoE_VER << 2)
-+#define IOC_NET_HEADER_FIELD_PPPoE_SID (IOC_NET_HEADER_FIELD_PPPoE_VER << 3)
-+#define IOC_NET_HEADER_FIELD_PPPoE_LEN (IOC_NET_HEADER_FIELD_PPPoE_VER << 4)
-+#define IOC_NET_HEADER_FIELD_PPPoE_SESSION (IOC_NET_HEADER_FIELD_PPPoE_VER << 5)
-+#define IOC_NET_HEADER_FIELD_PPPoE_PID (IOC_NET_HEADER_FIELD_PPPoE_VER << 6)
-+#define IOC_NET_HEADER_FIELD_PPPoE_ALL_FIELDS ((IOC_NET_HEADER_FIELD_PPPoE_VER << 7) - 1)
-+
-+#define IOC_NET_HEADER_FIELD_PPPMUX_PID (1)
-+#define IOC_NET_HEADER_FIELD_PPPMUX_CKSUM (IOC_NET_HEADER_FIELD_PPPMUX_PID << 1)
-+#define IOC_NET_HEADER_FIELD_PPPMUX_COMPRESSED (IOC_NET_HEADER_FIELD_PPPMUX_PID << 2)
-+#define IOC_NET_HEADER_FIELD_PPPMUX_ALL_FIELDS ((IOC_NET_HEADER_FIELD_PPPMUX_PID << 3) - 1)
-+
-+#define IOC_NET_HEADER_FIELD_PPPMUX_SUBFRAME_PFF (1)
-+#define IOC_NET_HEADER_FIELD_PPPMUX_SUBFRAME_LXT (IOC_NET_HEADER_FIELD_PPPMUX_SUBFRAME_PFF << 1)
-+#define IOC_NET_HEADER_FIELD_PPPMUX_SUBFRAME_LEN (IOC_NET_HEADER_FIELD_PPPMUX_SUBFRAME_PFF << 2)
-+#define IOC_NET_HEADER_FIELD_PPPMUX_SUBFRAME_PID (IOC_NET_HEADER_FIELD_PPPMUX_SUBFRAME_PFF << 3)
-+#define IOC_NET_HEADER_FIELD_PPPMUX_SUBFRAME_USE_PID (IOC_NET_HEADER_FIELD_PPPMUX_SUBFRAME_PFF << 4)
-+#define IOC_NET_HEADER_FIELD_PPPMUX_SUBFRAME_ALL_FIELDS ((IOC_NET_HEADER_FIELD_PPPMUX_SUBFRAME_PFF << 5) - 1)
-+
-+
-+typedef uint8_t ioc_header_field_eth_t;
-+
-+#define IOC_NET_HEADER_FIELD_ETH_DA (1)
-+#define IOC_NET_HEADER_FIELD_ETH_SA (IOC_NET_HEADER_FIELD_ETH_DA << 1)
-+#define IOC_NET_HEADER_FIELD_ETH_LENGTH (IOC_NET_HEADER_FIELD_ETH_DA << 2)
-+#define IOC_NET_HEADER_FIELD_ETH_TYPE (IOC_NET_HEADER_FIELD_ETH_DA << 3)
-+#define IOC_NET_HEADER_FIELD_ETH_FINAL_CKSUM (IOC_NET_HEADER_FIELD_ETH_DA << 4)
-+#define IOC_NET_HEADER_FIELD_ETH_PADDING (IOC_NET_HEADER_FIELD_ETH_DA << 5)
-+#define IOC_NET_HEADER_FIELD_ETH_ALL_FIELDS ((IOC_NET_HEADER_FIELD_ETH_DA << 6) - 1)
-+
-+#define IOC_NET_HEADER_FIELD_ETH_ADDR_SIZE 6
-+
-+typedef uint16_t ioc_header_field_ip_t;
-+
-+#define IOC_NET_HEADER_FIELD_IP_VER (1)
-+#define IOC_NET_HEADER_FIELD_IP_DSCP (IOC_NET_HEADER_FIELD_IP_VER << 2)
-+#define IOC_NET_HEADER_FIELD_IP_ECN (IOC_NET_HEADER_FIELD_IP_VER << 3)
-+#define IOC_NET_HEADER_FIELD_IP_PROTO (IOC_NET_HEADER_FIELD_IP_VER << 4)
-+
-+#define IOC_NET_HEADER_FIELD_IP_PROTO_SIZE 1
-+
-+typedef uint16_t ioc_header_field_ipv4_t;
-+
-+#define IOC_NET_HEADER_FIELD_IPv4_VER (1)
-+#define IOC_NET_HEADER_FIELD_IPv4_HDR_LEN (IOC_NET_HEADER_FIELD_IPv4_VER << 1)
-+#define IOC_NET_HEADER_FIELD_IPv4_TOS (IOC_NET_HEADER_FIELD_IPv4_VER << 2)
-+#define IOC_NET_HEADER_FIELD_IPv4_TOTAL_LEN (IOC_NET_HEADER_FIELD_IPv4_VER << 3)
-+#define IOC_NET_HEADER_FIELD_IPv4_ID (IOC_NET_HEADER_FIELD_IPv4_VER << 4)
-+#define IOC_NET_HEADER_FIELD_IPv4_FLAG_D (IOC_NET_HEADER_FIELD_IPv4_VER << 5)
-+#define IOC_NET_HEADER_FIELD_IPv4_FLAG_M (IOC_NET_HEADER_FIELD_IPv4_VER << 6)
-+#define IOC_NET_HEADER_FIELD_IPv4_OFFSET (IOC_NET_HEADER_FIELD_IPv4_VER << 7)
-+#define IOC_NET_HEADER_FIELD_IPv4_TTL (IOC_NET_HEADER_FIELD_IPv4_VER << 8)
-+#define IOC_NET_HEADER_FIELD_IPv4_PROTO (IOC_NET_HEADER_FIELD_IPv4_VER << 9)
-+#define IOC_NET_HEADER_FIELD_IPv4_CKSUM (IOC_NET_HEADER_FIELD_IPv4_VER << 10)
-+#define IOC_NET_HEADER_FIELD_IPv4_SRC_IP (IOC_NET_HEADER_FIELD_IPv4_VER << 11)
-+#define IOC_NET_HEADER_FIELD_IPv4_DST_IP (IOC_NET_HEADER_FIELD_IPv4_VER << 12)
-+#define IOC_NET_HEADER_FIELD_IPv4_OPTS (IOC_NET_HEADER_FIELD_IPv4_VER << 13)
-+#define IOC_NET_HEADER_FIELD_IPv4_OPTS_COUNT (IOC_NET_HEADER_FIELD_IPv4_VER << 14)
-+#define IOC_NET_HEADER_FIELD_IPv4_ALL_FIELDS ((IOC_NET_HEADER_FIELD_IPv4_VER << 15) - 1)
-+
-+#define IOC_NET_HEADER_FIELD_IPv4_ADDR_SIZE 4
-+#define IOC_NET_HEADER_FIELD_IPv4_PROTO_SIZE 1
-+
-+
-+typedef uint8_t ioc_header_field_ipv6_t;
-+
-+#define IOC_NET_HEADER_FIELD_IPv6_VER (1)
-+#define IOC_NET_HEADER_FIELD_IPv6_TC (IOC_NET_HEADER_FIELD_IPv6_VER << 1)
-+#define IOC_NET_HEADER_FIELD_IPv6_SRC_IP (IOC_NET_HEADER_FIELD_IPv6_VER << 2)
-+#define IOC_NET_HEADER_FIELD_IPv6_DST_IP (IOC_NET_HEADER_FIELD_IPv6_VER << 3)
-+#define IOC_NET_HEADER_FIELD_IPv6_NEXT_HDR (IOC_NET_HEADER_FIELD_IPv6_VER << 4)
-+#define IOC_NET_HEADER_FIELD_IPv6_FL (IOC_NET_HEADER_FIELD_IPv6_VER << 5)
-+#define IOC_NET_HEADER_FIELD_IPv6_HOP_LIMIT (IOC_NET_HEADER_FIELD_IPv6_VER << 6)
-+#define IOC_NET_HEADER_FIELD_IPv6_ALL_FIELDS ((IOC_NET_HEADER_FIELD_IPv6_VER << 7) - 1)
-+
-+#define IOC_NET_HEADER_FIELD_IPv6_ADDR_SIZE 16
-+#define IOC_NET_HEADER_FIELD_IPv6_NEXT_HDR_SIZE 1
-+
-+#define IOC_NET_HEADER_FIELD_ICMP_TYPE (1)
-+#define IOC_NET_HEADER_FIELD_ICMP_CODE (IOC_NET_HEADER_FIELD_ICMP_TYPE << 1)
-+#define IOC_NET_HEADER_FIELD_ICMP_CKSUM (IOC_NET_HEADER_FIELD_ICMP_TYPE << 2)
-+#define IOC_NET_HEADER_FIELD_ICMP_ID (IOC_NET_HEADER_FIELD_ICMP_TYPE << 3)
-+#define IOC_NET_HEADER_FIELD_ICMP_SQ_NUM (IOC_NET_HEADER_FIELD_ICMP_TYPE << 4)
-+#define IOC_NET_HEADER_FIELD_ICMP_ALL_FIELDS ((IOC_NET_HEADER_FIELD_ICMP_TYPE << 5) - 1)
-+
-+#define IOC_NET_HEADER_FIELD_ICMP_CODE_SIZE 1
-+#define IOC_NET_HEADER_FIELD_ICMP_TYPE_SIZE 1
-+
-+#define IOC_NET_HEADER_FIELD_IGMP_VERSION (1)
-+#define IOC_NET_HEADER_FIELD_IGMP_TYPE (IOC_NET_HEADER_FIELD_IGMP_VERSION << 1)
-+#define IOC_NET_HEADER_FIELD_IGMP_CKSUM (IOC_NET_HEADER_FIELD_IGMP_VERSION << 2)
-+#define IOC_NET_HEADER_FIELD_IGMP_DATA (IOC_NET_HEADER_FIELD_IGMP_VERSION << 3)
-+#define IOC_NET_HEADER_FIELD_IGMP_ALL_FIELDS ((IOC_NET_HEADER_FIELD_IGMP_VERSION << 4) - 1)
-+
-+
-+typedef uint16_t ioc_header_field_tcp_t;
-+
-+#define IOC_NET_HEADER_FIELD_TCP_PORT_SRC (1)
-+#define IOC_NET_HEADER_FIELD_TCP_PORT_DST (IOC_NET_HEADER_FIELD_TCP_PORT_SRC << 1)
-+#define IOC_NET_HEADER_FIELD_TCP_SEQ (IOC_NET_HEADER_FIELD_TCP_PORT_SRC << 2)
-+#define IOC_NET_HEADER_FIELD_TCP_ACK (IOC_NET_HEADER_FIELD_TCP_PORT_SRC << 3)
-+#define IOC_NET_HEADER_FIELD_TCP_OFFSET (IOC_NET_HEADER_FIELD_TCP_PORT_SRC << 4)
-+#define IOC_NET_HEADER_FIELD_TCP_FLAGS (IOC_NET_HEADER_FIELD_TCP_PORT_SRC << 5)
-+#define IOC_NET_HEADER_FIELD_TCP_WINDOW (IOC_NET_HEADER_FIELD_TCP_PORT_SRC << 6)
-+#define IOC_NET_HEADER_FIELD_TCP_CKSUM (IOC_NET_HEADER_FIELD_TCP_PORT_SRC << 7)
-+#define IOC_NET_HEADER_FIELD_TCP_URGPTR (IOC_NET_HEADER_FIELD_TCP_PORT_SRC << 8)
-+#define IOC_NET_HEADER_FIELD_TCP_OPTS (IOC_NET_HEADER_FIELD_TCP_PORT_SRC << 9)
-+#define IOC_NET_HEADER_FIELD_TCP_OPTS_COUNT (IOC_NET_HEADER_FIELD_TCP_PORT_SRC << 10)
-+#define IOC_NET_HEADER_FIELD_TCP_ALL_FIELDS ((IOC_NET_HEADER_FIELD_TCP_PORT_SRC << 11) - 1)
-+
-+#define IOC_NET_HEADER_FIELD_TCP_PORT_SIZE 2
-+
-+
-+typedef uint8_t ioc_header_field_sctp_t;
-+
-+#define IOC_NET_HEADER_FIELD_SCTP_PORT_SRC (1)
-+#define IOC_NET_HEADER_FIELD_SCTP_PORT_DST (IOC_NET_HEADER_FIELD_SCTP_PORT_SRC << 1)
-+#define IOC_NET_HEADER_FIELD_SCTP_VER_TAG (IOC_NET_HEADER_FIELD_SCTP_PORT_SRC << 2)
-+#define IOC_NET_HEADER_FIELD_SCTP_CKSUM (IOC_NET_HEADER_FIELD_SCTP_PORT_SRC << 3)
-+#define IOC_NET_HEADER_FIELD_SCTP_ALL_FIELDS ((IOC_NET_HEADER_FIELD_SCTP_PORT_SRC << 4) - 1)
-+
-+#define IOC_NET_HEADER_FIELD_SCTP_PORT_SIZE 2
-+
-+typedef uint8_t ioc_header_field_dccp_t;
-+
-+#define IOC_NET_HEADER_FIELD_DCCP_PORT_SRC (1)
-+#define IOC_NET_HEADER_FIELD_DCCP_PORT_DST (IOC_NET_HEADER_FIELD_DCCP_PORT_SRC << 1)
-+#define IOC_NET_HEADER_FIELD_DCCP_ALL_FIELDS ((IOC_NET_HEADER_FIELD_DCCP_PORT_SRC << 2) - 1)
-+
-+#define IOC_NET_HEADER_FIELD_DCCP_PORT_SIZE 2
-+
-+
-+typedef uint8_t ioc_header_field_udp_t;
-+
-+#define IOC_NET_HEADER_FIELD_UDP_PORT_SRC (1)
-+#define IOC_NET_HEADER_FIELD_UDP_PORT_DST (IOC_NET_HEADER_FIELD_UDP_PORT_SRC << 1)
-+#define IOC_NET_HEADER_FIELD_UDP_LEN (IOC_NET_HEADER_FIELD_UDP_PORT_SRC << 2)
-+#define IOC_NET_HEADER_FIELD_UDP_CKSUM (IOC_NET_HEADER_FIELD_UDP_PORT_SRC << 3)
-+#define IOC_NET_HEADER_FIELD_UDP_ALL_FIELDS ((IOC_NET_HEADER_FIELD_UDP_PORT_SRC << 4) - 1)
-+
-+#define IOC_NET_HEADER_FIELD_UDP_PORT_SIZE 2
-+
-+typedef uint8_t ioc_header_field_udp_lite_t;
-+
-+#define IOC_NET_HEADER_FIELD_UDP_LITE_PORT_SRC (1)
-+#define IOC_NET_HEADER_FIELD_UDP_LITE_PORT_DST (IOC_NET_HEADER_FIELD_UDP_LITE_PORT_SRC << 1)
-+#define IOC_NET_HEADER_FIELD_UDP_LITE_ALL_FIELDS ((IOC_NET_HEADER_FIELD_UDP_LITE_PORT_SRC << 2) - 1)
-+
-+#define IOC_NET_HEADER_FIELD_UDP_LITE_PORT_SIZE 2
-+
-+typedef uint8_t ioc_header_field_udp_encap_esp_t;
-+
-+#define IOC_NET_HEADER_FIELD_UDP_ENCAP_ESP_PORT_SRC (1)
-+#define IOC_NET_HEADER_FIELD_UDP_ENCAP_ESP_PORT_DST (IOC_NET_HEADER_FIELD_UDP_ENCAP_ESP_PORT_SRC << 1)
-+#define IOC_NET_HEADER_FIELD_UDP_ENCAP_ESP_LEN (IOC_NET_HEADER_FIELD_UDP_ENCAP_ESP_PORT_SRC << 2)
-+#define IOC_NET_HEADER_FIELD_UDP_ENCAP_ESP_CKSUM (IOC_NET_HEADER_FIELD_UDP_ENCAP_ESP_PORT_SRC << 3)
-+#define IOC_NET_HEADER_FIELD_UDP_ENCAP_ESP_SPI (IOC_NET_HEADER_FIELD_UDP_ENCAP_ESP_PORT_SRC << 4)
-+#define IOC_NET_HEADER_FIELD_UDP_ENCAP_ESP_SEQUENCE_NUM (IOC_NET_HEADER_FIELD_UDP_ENCAP_ESP_PORT_SRC << 5)
-+#define IOC_NET_HEADER_FIELD_UDP_ENCAP_ESP_ALL_FIELDS ((IOC_NET_HEADER_FIELD_UDP_ENCAP_ESP_PORT_SRC << 6) - 1)
-+
-+#define IOC_NET_HEADER_FIELD_UDP_ENCAP_ESP_PORT_SIZE 2
-+#define IOC_NET_HEADER_FIELD_UDP_ENCAP_ESP_SPI_SIZE 4
-+
-+#define IOC_NET_HEADER_FIELD_IPHC_CID (1)
-+#define IOC_NET_HEADER_FIELD_IPHC_CID_TYPE (IOC_NET_HEADER_FIELD_IPHC_CID << 1)
-+#define IOC_NET_HEADER_FIELD_IPHC_HCINDEX (IOC_NET_HEADER_FIELD_IPHC_CID << 2)
-+#define IOC_NET_HEADER_FIELD_IPHC_GEN (IOC_NET_HEADER_FIELD_IPHC_CID << 3)
-+#define IOC_NET_HEADER_FIELD_IPHC_D_BIT (IOC_NET_HEADER_FIELD_IPHC_CID << 4)
-+#define IOC_NET_HEADER_FIELD_IPHC_ALL_FIELDS ((IOC_NET_HEADER_FIELD_IPHC_CID << 5) - 1)
-+
-+#define IOC_NET_HEADER_FIELD_SCTP_CHUNK_DATA_TYPE (1)
-+#define IOC_NET_HEADER_FIELD_SCTP_CHUNK_DATA_FLAGS (IOC_NET_HEADER_FIELD_SCTP_CHUNK_DATA_TYPE << 1)
-+#define IOC_NET_HEADER_FIELD_SCTP_CHUNK_DATA_LENGTH (IOC_NET_HEADER_FIELD_SCTP_CHUNK_DATA_TYPE << 2)
-+#define IOC_NET_HEADER_FIELD_SCTP_CHUNK_DATA_TSN (IOC_NET_HEADER_FIELD_SCTP_CHUNK_DATA_TYPE << 3)
-+#define IOC_NET_HEADER_FIELD_SCTP_CHUNK_DATA_STREAM_ID (IOC_NET_HEADER_FIELD_SCTP_CHUNK_DATA_TYPE << 4)
-+#define IOC_NET_HEADER_FIELD_SCTP_CHUNK_DATA_STREAM_SQN (IOC_NET_HEADER_FIELD_SCTP_CHUNK_DATA_TYPE << 5)
-+#define IOC_NET_HEADER_FIELD_SCTP_CHUNK_DATA_PAYLOAD_PID (IOC_NET_HEADER_FIELD_SCTP_CHUNK_DATA_TYPE << 6)
-+#define IOC_NET_HEADER_FIELD_SCTP_CHUNK_DATA_UNORDERED (IOC_NET_HEADER_FIELD_SCTP_CHUNK_DATA_TYPE << 7)
-+#define IOC_NET_HEADER_FIELD_SCTP_CHUNK_DATA_BEGGINING (IOC_NET_HEADER_FIELD_SCTP_CHUNK_DATA_TYPE << 8)
-+#define IOC_NET_HEADER_FIELD_SCTP_CHUNK_DATA_END (IOC_NET_HEADER_FIELD_SCTP_CHUNK_DATA_TYPE << 9)
-+#define IOC_NET_HEADER_FIELD_SCTP_CHUNK_DATA_ALL_FIELDS ((IOC_NET_HEADER_FIELD_SCTP_CHUNK_DATA_TYPE << 10) - 1)
-+
-+#define IOC_NET_HEADER_FIELD_L2TPv2_TYPE_BIT (1)
-+#define IOC_NET_HEADER_FIELD_L2TPv2_LENGTH_BIT (IOC_NET_HEADER_FIELD_L2TPv2_TYPE_BIT << 1)
-+#define IOC_NET_HEADER_FIELD_L2TPv2_SEQUENCE_BIT (IOC_NET_HEADER_FIELD_L2TPv2_TYPE_BIT << 2)
-+#define IOC_NET_HEADER_FIELD_L2TPv2_OFFSET_BIT (IOC_NET_HEADER_FIELD_L2TPv2_TYPE_BIT << 3)
-+#define IOC_NET_HEADER_FIELD_L2TPv2_PRIORITY_BIT (IOC_NET_HEADER_FIELD_L2TPv2_TYPE_BIT << 4)
-+#define IOC_NET_HEADER_FIELD_L2TPv2_VERSION (IOC_NET_HEADER_FIELD_L2TPv2_TYPE_BIT << 5)
-+#define IOC_NET_HEADER_FIELD_L2TPv2_LEN (IOC_NET_HEADER_FIELD_L2TPv2_TYPE_BIT << 6)
-+#define IOC_NET_HEADER_FIELD_L2TPv2_TUNNEL_ID (IOC_NET_HEADER_FIELD_L2TPv2_TYPE_BIT << 7)
-+#define IOC_NET_HEADER_FIELD_L2TPv2_SESSION_ID (IOC_NET_HEADER_FIELD_L2TPv2_TYPE_BIT << 8)
-+#define IOC_NET_HEADER_FIELD_L2TPv2_NS (IOC_NET_HEADER_FIELD_L2TPv2_TYPE_BIT << 9)
-+#define IOC_NET_HEADER_FIELD_L2TPv2_NR (IOC_NET_HEADER_FIELD_L2TPv2_TYPE_BIT << 10)
-+#define IOC_NET_HEADER_FIELD_L2TPv2_OFFSET_SIZE (IOC_NET_HEADER_FIELD_L2TPv2_TYPE_BIT << 11)
-+#define IOC_NET_HEADER_FIELD_L2TPv2_FIRST_BYTE (IOC_NET_HEADER_FIELD_L2TPv2_TYPE_BIT << 12)
-+#define IOC_NET_HEADER_FIELD_L2TPv2_ALL_FIELDS ((IOC_NET_HEADER_FIELD_L2TPv2_TYPE_BIT << 13) - 1)
-+
-+#define IOC_NET_HEADER_FIELD_L2TPv3_CTRL_TYPE_BIT (1)
-+#define IOC_NET_HEADER_FIELD_L2TPv3_CTRL_LENGTH_BIT (IOC_NET_HEADER_FIELD_L2TPv3_CTRL_TYPE_BIT << 1)
-+#define IOC_NET_HEADER_FIELD_L2TPv3_CTRL_SEQUENCE_BIT (IOC_NET_HEADER_FIELD_L2TPv3_CTRL_TYPE_BIT << 2)
-+#define IOC_NET_HEADER_FIELD_L2TPv3_CTRL_VERSION (IOC_NET_HEADER_FIELD_L2TPv3_CTRL_TYPE_BIT << 3)
-+#define IOC_NET_HEADER_FIELD_L2TPv3_CTRL_LENGTH (IOC_NET_HEADER_FIELD_L2TPv3_CTRL_TYPE_BIT << 4)
-+#define IOC_NET_HEADER_FIELD_L2TPv3_CTRL_CONTROL (IOC_NET_HEADER_FIELD_L2TPv3_CTRL_TYPE_BIT << 5)
-+#define IOC_NET_HEADER_FIELD_L2TPv3_CTRL_SENT (IOC_NET_HEADER_FIELD_L2TPv3_CTRL_TYPE_BIT << 6)
-+#define IOC_NET_HEADER_FIELD_L2TPv3_CTRL_RECV (IOC_NET_HEADER_FIELD_L2TPv3_CTRL_TYPE_BIT << 7)
-+#define IOC_NET_HEADER_FIELD_L2TPv3_CTRL_FIRST_BYTE (IOC_NET_HEADER_FIELD_L2TPv3_CTRL_TYPE_BIT << 8)
-+#define IOC_NET_HEADER_FIELD_L2TPv3_CTRL_ALL_FIELDS ((IOC_NET_HEADER_FIELD_L2TPv3_CTRL_TYPE_BIT << 9) - 1)
-+
-+#define IOC_NET_HEADER_FIELD_L2TPv3_SESS_TYPE_BIT (1)
-+#define IOC_NET_HEADER_FIELD_L2TPv3_SESS_VERSION (IOC_NET_HEADER_FIELD_L2TPv3_SESS_TYPE_BIT << 1)
-+#define IOC_NET_HEADER_FIELD_L2TPv3_SESS_ID (IOC_NET_HEADER_FIELD_L2TPv3_SESS_TYPE_BIT << 2)
-+#define IOC_NET_HEADER_FIELD_L2TPv3_SESS_COOKIE (IOC_NET_HEADER_FIELD_L2TPv3_SESS_TYPE_BIT << 3)
-+#define IOC_NET_HEADER_FIELD_L2TPv3_SESS_ALL_FIELDS ((IOC_NET_HEADER_FIELD_L2TPv3_SESS_TYPE_BIT << 4) - 1)
-+
-+
-+typedef uint8_t ioc_header_field_vlan_t;
-+
-+#define IOC_NET_HEADER_FIELD_VLAN_VPRI (1)
-+#define IOC_NET_HEADER_FIELD_VLAN_CFI (IOC_NET_HEADER_FIELD_VLAN_VPRI << 1)
-+#define IOC_NET_HEADER_FIELD_VLAN_VID (IOC_NET_HEADER_FIELD_VLAN_VPRI << 2)
-+#define IOC_NET_HEADER_FIELD_VLAN_LENGTH (IOC_NET_HEADER_FIELD_VLAN_VPRI << 3)
-+#define IOC_NET_HEADER_FIELD_VLAN_TYPE (IOC_NET_HEADER_FIELD_VLAN_VPRI << 4)
-+#define IOC_NET_HEADER_FIELD_VLAN_ALL_FIELDS ((IOC_NET_HEADER_FIELD_VLAN_VPRI << 5) - 1)
-+
-+#define IOC_NET_HEADER_FIELD_VLAN_TCI (IOC_NET_HEADER_FIELD_VLAN_VPRI | \
-+ IOC_NET_HEADER_FIELD_VLAN_CFI | \
-+ IOC_NET_HEADER_FIELD_VLAN_VID)
-+
-+
-+typedef uint8_t ioc_header_field_llc_t;
-+
-+#define IOC_NET_HEADER_FIELD_LLC_DSAP (1)
-+#define IOC_NET_HEADER_FIELD_LLC_SSAP (IOC_NET_HEADER_FIELD_LLC_DSAP << 1)
-+#define IOC_NET_HEADER_FIELD_LLC_CTRL (IOC_NET_HEADER_FIELD_LLC_DSAP << 2)
-+#define IOC_NET_HEADER_FIELD_LLC_ALL_FIELDS ((IOC_NET_HEADER_FIELD_LLC_DSAP << 3) - 1)
-+
-+#define IOC_NET_HEADER_FIELD_NLPID_NLPID (1)
-+#define IOC_NET_HEADER_FIELD_NLPID_ALL_FIELDS ((IOC_NET_HEADER_FIELD_NLPID_NLPID << 1) - 1)
-+
-+
-+typedef uint8_t ioc_header_field_snap_t;
-+
-+#define IOC_NET_HEADER_FIELD_SNAP_OUI (1)
-+#define IOC_NET_HEADER_FIELD_SNAP_PID (IOC_NET_HEADER_FIELD_SNAP_OUI << 1)
-+#define IOC_NET_HEADER_FIELD_SNAP_ALL_FIELDS ((IOC_NET_HEADER_FIELD_SNAP_OUI << 2) - 1)
-+
-+
-+typedef uint8_t ioc_header_field_llc_snap_t;
-+
-+#define IOC_NET_HEADER_FIELD_LLC_SNAP_TYPE (1)
-+#define IOC_NET_HEADER_FIELD_LLC_SNAP_ALL_FIELDS ((IOC_NET_HEADER_FIELD_LLC_SNAP_TYPE << 1) - 1)
-+
-+#define IOC_NET_HEADER_FIELD_ARP_HTYPE (1)
-+#define IOC_NET_HEADER_FIELD_ARP_PTYPE (IOC_NET_HEADER_FIELD_ARP_HTYPE << 1)
-+#define IOC_NET_HEADER_FIELD_ARP_HLEN (IOC_NET_HEADER_FIELD_ARP_HTYPE << 2)
-+#define IOC_NET_HEADER_FIELD_ARP_PLEN (IOC_NET_HEADER_FIELD_ARP_HTYPE << 3)
-+#define IOC_NET_HEADER_FIELD_ARP_OPER (IOC_NET_HEADER_FIELD_ARP_HTYPE << 4)
-+#define IOC_NET_HEADER_FIELD_ARP_SHA (IOC_NET_HEADER_FIELD_ARP_HTYPE << 5)
-+#define IOC_NET_HEADER_FIELD_ARP_SPA (IOC_NET_HEADER_FIELD_ARP_HTYPE << 6)
-+#define IOC_NET_HEADER_FIELD_ARP_THA (IOC_NET_HEADER_FIELD_ARP_HTYPE << 7)
-+#define IOC_NET_HEADER_FIELD_ARP_TPA (IOC_NET_HEADER_FIELD_ARP_HTYPE << 8)
-+#define IOC_NET_HEADER_FIELD_ARP_ALL_FIELDS ((IOC_NET_HEADER_FIELD_ARP_HTYPE << 9) - 1)
-+
-+#define IOC_NET_HEADER_FIELD_RFC2684_LLC (1)
-+#define IOC_NET_HEADER_FIELD_RFC2684_NLPID (IOC_NET_HEADER_FIELD_RFC2684_LLC << 1)
-+#define IOC_NET_HEADER_FIELD_RFC2684_OUI (IOC_NET_HEADER_FIELD_RFC2684_LLC << 2)
-+#define IOC_NET_HEADER_FIELD_RFC2684_PID (IOC_NET_HEADER_FIELD_RFC2684_LLC << 3)
-+#define IOC_NET_HEADER_FIELD_RFC2684_VPN_OUI (IOC_NET_HEADER_FIELD_RFC2684_LLC << 4)
-+#define IOC_NET_HEADER_FIELD_RFC2684_VPN_IDX (IOC_NET_HEADER_FIELD_RFC2684_LLC << 5)
-+#define IOC_NET_HEADER_FIELD_RFC2684_ALL_FIELDS ((IOC_NET_HEADER_FIELD_RFC2684_LLC << 6) - 1)
-+
-+#define IOC_NET_HEADER_FIELD_USER_DEFINED_SRCPORT (1)
-+#define IOC_NET_HEADER_FIELD_USER_DEFINED_PCDID (IOC_NET_HEADER_FIELD_USER_DEFINED_SRCPORT << 1)
-+#define IOC_NET_HEADER_FIELD_USER_DEFINED_ALL_FIELDS ((IOC_NET_HEADER_FIELD_USER_DEFINED_SRCPORT << 2) - 1)
-+
-+#define IOC_NET_HEADER_FIELD_PAYLOAD_BUFFER (1)
-+#define IOC_NET_HEADER_FIELD_PAYLOAD_SIZE (IOC_NET_HEADER_FIELD_PAYLOAD_BUFFER << 1)
-+#define IOC_NET_HEADER_FIELD_MAX_FRM_SIZE (IOC_NET_HEADER_FIELD_PAYLOAD_BUFFER << 2)
-+#define IOC_NET_HEADER_FIELD_MIN_FRM_SIZE (IOC_NET_HEADER_FIELD_PAYLOAD_BUFFER << 3)
-+#define IOC_NET_HEADER_FIELD_PAYLOAD_TYPE (IOC_NET_HEADER_FIELD_PAYLOAD_BUFFER << 4)
-+#define IOC_NET_HEADER_FIELD_FRAME_SIZE (IOC_NET_HEADER_FIELD_PAYLOAD_BUFFER << 5)
-+#define IOC_NET_HEADER_FIELD_PAYLOAD_ALL_FIELDS ((IOC_NET_HEADER_FIELD_PAYLOAD_BUFFER << 6) - 1)
-+
-+
-+typedef uint8_t ioc_header_field_gre_t;
-+
-+#define IOC_NET_HEADER_FIELD_GRE_TYPE (1)
-+#define IOC_NET_HEADER_FIELD_GRE_ALL_FIELDS ((IOC_NET_HEADER_FIELD_GRE_TYPE << 1) - 1)
-+
-+
-+typedef uint8_t ioc_header_field_minencap_t;
-+
-+#define IOC_NET_HEADER_FIELD_MINENCAP_SRC_IP (1)
-+#define IOC_NET_HEADER_FIELD_MINENCAP_DST_IP (IOC_NET_HEADER_FIELD_MINENCAP_SRC_IP << 1)
-+#define IOC_NET_HEADER_FIELD_MINENCAP_TYPE (IOC_NET_HEADER_FIELD_MINENCAP_SRC_IP << 2)
-+#define IOC_NET_HEADER_FIELD_MINENCAP_ALL_FIELDS ((IOC_NET_HEADER_FIELD_MINENCAP_SRC_IP << 3) - 1)
-+
-+
-+typedef uint8_t ioc_header_field_ipsec_ah_t;
-+
-+#define IOC_NET_HEADER_FIELD_IPSEC_AH_SPI (1)
-+#define IOC_NET_HEADER_FIELD_IPSEC_AH_NH (IOC_NET_HEADER_FIELD_IPSEC_AH_SPI << 1)
-+#define IOC_NET_HEADER_FIELD_IPSEC_AH_ALL_FIELDS ((IOC_NET_HEADER_FIELD_IPSEC_AH_SPI << 2) - 1)
-+
-+
-+typedef uint8_t ioc_header_field_ipsec_esp_t;
-+
-+#define IOC_NET_HEADER_FIELD_IPSEC_ESP_SPI (1)
-+#define IOC_NET_HEADER_FIELD_IPSEC_ESP_SEQUENCE_NUM (IOC_NET_HEADER_FIELD_IPSEC_ESP_SPI << 1)
-+#define IOC_NET_HEADER_FIELD_IPSEC_ESP_ALL_FIELDS ((IOC_NET_HEADER_FIELD_IPSEC_ESP_SPI << 2) - 1)
-+
-+#define IOC_NET_HEADER_FIELD_IPSEC_ESP_SPI_SIZE 4
-+
-+
-+typedef uint8_t ioc_header_field_mpls_t;
-+
-+#define IOC_NET_HEADER_FIELD_MPLS_LABEL_STACK (1)
-+#define IOC_NET_HEADER_FIELD_MPLS_LABEL_STACK_ALL_FIELDS ((IOC_NET_HEADER_FIELD_MPLS_LABEL_STACK << 1) - 1)
-+
-+
-+typedef uint8_t ioc_header_field_macsec_t;
-+
-+#define IOC_NET_HEADER_FIELD_MACSEC_SECTAG (1)
-+#define IOC_NET_HEADER_FIELD_MACSEC_ALL_FIELDS ((IOC_NET_HEADER_FIELD_MACSEC_SECTAG << 1) - 1)
-+
-+
-+typedef enum {
-+ e_IOC_NET_HEADER_TYPE_NONE = 0,
-+ e_IOC_NET_HEADER_TYPE_PAYLOAD,
-+ e_IOC_NET_HEADER_TYPE_ETH,
-+ e_IOC_NET_HEADER_TYPE_VLAN,
-+ e_IOC_NET_HEADER_TYPE_IPv4,
-+ e_IOC_NET_HEADER_TYPE_IPv6,
-+ e_IOC_NET_HEADER_TYPE_IP,
-+ e_IOC_NET_HEADER_TYPE_TCP,
-+ e_IOC_NET_HEADER_TYPE_UDP,
-+ e_IOC_NET_HEADER_TYPE_UDP_LITE,
-+ e_IOC_NET_HEADER_TYPE_IPHC,
-+ e_IOC_NET_HEADER_TYPE_SCTP,
-+ e_IOC_NET_HEADER_TYPE_SCTP_CHUNK_DATA,
-+ e_IOC_NET_HEADER_TYPE_PPPoE,
-+ e_IOC_NET_HEADER_TYPE_PPP,
-+ e_IOC_NET_HEADER_TYPE_PPPMUX,
-+ e_IOC_NET_HEADER_TYPE_PPPMUX_SUBFRAME,
-+ e_IOC_NET_HEADER_TYPE_L2TPv2,
-+ e_IOC_NET_HEADER_TYPE_L2TPv3_CTRL,
-+ e_IOC_NET_HEADER_TYPE_L2TPv3_SESS,
-+ e_IOC_NET_HEADER_TYPE_LLC,
-+ e_IOC_NET_HEADER_TYPE_LLC_SNAP,
-+ e_IOC_NET_HEADER_TYPE_NLPID,
-+ e_IOC_NET_HEADER_TYPE_SNAP,
-+ e_IOC_NET_HEADER_TYPE_MPLS,
-+ e_IOC_NET_HEADER_TYPE_IPSEC_AH,
-+ e_IOC_NET_HEADER_TYPE_IPSEC_ESP,
-+ e_IOC_NET_HEADER_TYPE_UDP_ENCAP_ESP, /* RFC 3948 */
-+ e_IOC_NET_HEADER_TYPE_MACSEC,
-+ e_IOC_NET_HEADER_TYPE_GRE,
-+ e_IOC_NET_HEADER_TYPE_MINENCAP,
-+ e_IOC_NET_HEADER_TYPE_DCCP,
-+ e_IOC_NET_HEADER_TYPE_ICMP,
-+ e_IOC_NET_HEADER_TYPE_IGMP,
-+ e_IOC_NET_HEADER_TYPE_ARP,
-+ e_IOC_NET_HEADER_TYPE_CAPWAP,
-+ e_IOC_NET_HEADER_TYPE_CAPWAP_DTLS,
-+ e_IOC_NET_HEADER_TYPE_RFC2684,
-+ e_IOC_NET_HEADER_TYPE_USER_DEFINED_L2,
-+ e_IOC_NET_HEADER_TYPE_USER_DEFINED_L3,
-+ e_IOC_NET_HEADER_TYPE_USER_DEFINED_L4,
-+ e_IOC_NET_HEADER_TYPE_USER_DEFINED_SHIM1,
-+ e_IOC_NET_HEADER_TYPE_USER_DEFINED_SHIM2,
-+ e_IOC_NET_MAX_HEADER_TYPE_COUNT
-+} ioc_net_header_type;
-+
-+
-+#endif /* __NET_IOCTLS_H */
diff --git a/target/linux/layerscape/patches-4.9/702-pci-support-layerscape.patch b/target/linux/layerscape/patches-4.9/702-pci-support-layerscape.patch
deleted file mode 100644
index 6f34cfe58e..0000000000
--- a/target/linux/layerscape/patches-4.9/702-pci-support-layerscape.patch
+++ /dev/null
@@ -1,2098 +0,0 @@
-From b2ee6e29bad31facbbf5ac1ce98235ac163d9fa9 Mon Sep 17 00:00:00 2001
-From: Yangbo Lu <yangbo.lu@nxp.com>
-Date: Thu, 5 Jul 2018 16:26:47 +0800
-Subject: [PATCH 08/32] pci: support layerscape
-
-This is an integrated patch for layerscape pcie support.
-
-Signed-off-by: Po Liu <po.liu@nxp.com>
-Signed-off-by: Liu Gang <Gang.Liu@nxp.com>
-Signed-off-by: Minghuan Lian <Minghuan.Lian@freescale.com>
-Signed-off-by: hongbo.wang <hongbo.wang@nxp.com>
-Signed-off-by: Bjorn Helgaas <bhelgaas@google.com>
-Signed-off-by: Hou Zhiqiang <Zhiqiang.Hou@nxp.com>
-Signed-off-by: Mingkai Hu <mingkai.hu@nxp.com>
-Signed-off-by: Christoph Hellwig <hch@lst.de>
-Signed-off-by: Yangbo Lu <yangbo.lu@nxp.com>
----
- drivers/irqchip/irq-ls-scfg-msi.c | 257 ++++++-
- drivers/pci/host/Makefile | 2 +-
- drivers/pci/host/pci-layerscape-ep-debugfs.c | 758 +++++++++++++++++++
- drivers/pci/host/pci-layerscape-ep.c | 309 ++++++++
- drivers/pci/host/pci-layerscape-ep.h | 115 +++
- drivers/pci/host/pci-layerscape.c | 48 +-
- drivers/pci/host/pcie-designware.c | 6 +
- drivers/pci/host/pcie-designware.h | 1 +
- drivers/pci/pci.c | 2 +-
- drivers/pci/pcie/portdrv_core.c | 181 ++---
- drivers/pci/quirks.c | 15 +
- include/linux/pci.h | 1 +
- 12 files changed, 1546 insertions(+), 149 deletions(-)
- create mode 100644 drivers/pci/host/pci-layerscape-ep-debugfs.c
- create mode 100644 drivers/pci/host/pci-layerscape-ep.c
- create mode 100644 drivers/pci/host/pci-layerscape-ep.h
-
---- a/drivers/irqchip/irq-ls-scfg-msi.c
-+++ b/drivers/irqchip/irq-ls-scfg-msi.c
-@@ -17,13 +17,32 @@
- #include <linux/irq.h>
- #include <linux/irqchip/chained_irq.h>
- #include <linux/irqdomain.h>
-+#include <linux/of_irq.h>
- #include <linux/of_pci.h>
- #include <linux/of_platform.h>
- #include <linux/spinlock.h>
-
--#define MSI_MAX_IRQS 32
--#define MSI_IBS_SHIFT 3
--#define MSIR 4
-+#define MSI_IRQS_PER_MSIR 32
-+#define MSI_MSIR_OFFSET 4
-+
-+#define MSI_LS1043V1_1_IRQS_PER_MSIR 8
-+#define MSI_LS1043V1_1_MSIR_OFFSET 0x10
-+
-+struct ls_scfg_msi_cfg {
-+ u32 ibs_shift; /* Shift of interrupt bit select */
-+ u32 msir_irqs; /* The irq number per MSIR */
-+ u32 msir_base; /* The base address of MSIR */
-+};
-+
-+struct ls_scfg_msir {
-+ struct ls_scfg_msi *msi_data;
-+ unsigned int index;
-+ unsigned int gic_irq;
-+ unsigned int bit_start;
-+ unsigned int bit_end;
-+ unsigned int srs; /* Shared interrupt register select */
-+ void __iomem *reg;
-+};
-
- struct ls_scfg_msi {
- spinlock_t lock;
-@@ -32,8 +51,11 @@ struct ls_scfg_msi {
- struct irq_domain *msi_domain;
- void __iomem *regs;
- phys_addr_t msiir_addr;
-- int irq;
-- DECLARE_BITMAP(used, MSI_MAX_IRQS);
-+ struct ls_scfg_msi_cfg *cfg;
-+ u32 msir_num;
-+ struct ls_scfg_msir *msir;
-+ u32 irqs_num;
-+ unsigned long *used;
- };
-
- static struct irq_chip ls_scfg_msi_irq_chip = {
-@@ -49,19 +71,56 @@ static struct msi_domain_info ls_scfg_ms
- .chip = &ls_scfg_msi_irq_chip,
- };
-
-+static int msi_affinity_flag = 1;
-+
-+static int __init early_parse_ls_scfg_msi(char *p)
-+{
-+ if (p && strncmp(p, "no-affinity", 11) == 0)
-+ msi_affinity_flag = 0;
-+ else
-+ msi_affinity_flag = 1;
-+
-+ return 0;
-+}
-+early_param("lsmsi", early_parse_ls_scfg_msi);
-+
- static void ls_scfg_msi_compose_msg(struct irq_data *data, struct msi_msg *msg)
- {
- struct ls_scfg_msi *msi_data = irq_data_get_irq_chip_data(data);
-
- msg->address_hi = upper_32_bits(msi_data->msiir_addr);
- msg->address_lo = lower_32_bits(msi_data->msiir_addr);
-- msg->data = data->hwirq << MSI_IBS_SHIFT;
-+ msg->data = data->hwirq;
-+
-+ if (msi_affinity_flag)
-+ msg->data |= cpumask_first(data->common->affinity);
- }
-
- static int ls_scfg_msi_set_affinity(struct irq_data *irq_data,
- const struct cpumask *mask, bool force)
- {
-- return -EINVAL;
-+ struct ls_scfg_msi *msi_data = irq_data_get_irq_chip_data(irq_data);
-+ u32 cpu;
-+
-+ if (!msi_affinity_flag)
-+ return -EINVAL;
-+
-+ if (!force)
-+ cpu = cpumask_any_and(mask, cpu_online_mask);
-+ else
-+ cpu = cpumask_first(mask);
-+
-+ if (cpu >= msi_data->msir_num)
-+ return -EINVAL;
-+
-+ if (msi_data->msir[cpu].gic_irq <= 0) {
-+ pr_warn("cannot bind the irq to cpu%d\n", cpu);
-+ return -EINVAL;
-+ }
-+
-+ cpumask_copy(irq_data->common->affinity, mask);
-+
-+ return IRQ_SET_MASK_OK;
- }
-
- static struct irq_chip ls_scfg_msi_parent_chip = {
-@@ -81,8 +140,8 @@ static int ls_scfg_msi_domain_irq_alloc(
- WARN_ON(nr_irqs != 1);
-
- spin_lock(&msi_data->lock);
-- pos = find_first_zero_bit(msi_data->used, MSI_MAX_IRQS);
-- if (pos < MSI_MAX_IRQS)
-+ pos = find_first_zero_bit(msi_data->used, msi_data->irqs_num);
-+ if (pos < msi_data->irqs_num)
- __set_bit(pos, msi_data->used);
- else
- err = -ENOSPC;
-@@ -106,7 +165,7 @@ static void ls_scfg_msi_domain_irq_free(
- int pos;
-
- pos = d->hwirq;
-- if (pos < 0 || pos >= MSI_MAX_IRQS) {
-+ if (pos < 0 || pos >= msi_data->irqs_num) {
- pr_err("failed to teardown msi. Invalid hwirq %d\n", pos);
- return;
- }
-@@ -123,15 +182,22 @@ static const struct irq_domain_ops ls_sc
-
- static void ls_scfg_msi_irq_handler(struct irq_desc *desc)
- {
-- struct ls_scfg_msi *msi_data = irq_desc_get_handler_data(desc);
-+ struct ls_scfg_msir *msir = irq_desc_get_handler_data(desc);
-+ struct ls_scfg_msi *msi_data = msir->msi_data;
- unsigned long val;
-- int pos, virq;
-+ int pos, size, virq, hwirq;
-
- chained_irq_enter(irq_desc_get_chip(desc), desc);
-
-- val = ioread32be(msi_data->regs + MSIR);
-- for_each_set_bit(pos, &val, MSI_MAX_IRQS) {
-- virq = irq_find_mapping(msi_data->parent, (31 - pos));
-+ val = ioread32be(msir->reg);
-+
-+ pos = msir->bit_start;
-+ size = msir->bit_end + 1;
-+
-+ for_each_set_bit_from(pos, &val, size) {
-+ hwirq = ((msir->bit_end - pos) << msi_data->cfg->ibs_shift) |
-+ msir->srs;
-+ virq = irq_find_mapping(msi_data->parent, hwirq);
- if (virq)
- generic_handle_irq(virq);
- }
-@@ -143,7 +209,7 @@ static int ls_scfg_msi_domains_init(stru
- {
- /* Initialize MSI domain parent */
- msi_data->parent = irq_domain_add_linear(NULL,
-- MSI_MAX_IRQS,
-+ msi_data->irqs_num,
- &ls_scfg_msi_domain_ops,
- msi_data);
- if (!msi_data->parent) {
-@@ -164,16 +230,118 @@ static int ls_scfg_msi_domains_init(stru
- return 0;
- }
-
-+static int ls_scfg_msi_setup_hwirq(struct ls_scfg_msi *msi_data, int index)
-+{
-+ struct ls_scfg_msir *msir;
-+ int virq, i, hwirq;
-+
-+ virq = platform_get_irq(msi_data->pdev, index);
-+ if (virq <= 0)
-+ return -ENODEV;
-+
-+ msir = &msi_data->msir[index];
-+ msir->index = index;
-+ msir->msi_data = msi_data;
-+ msir->gic_irq = virq;
-+ msir->reg = msi_data->regs + msi_data->cfg->msir_base + 4 * index;
-+
-+ if (msi_data->cfg->msir_irqs == MSI_LS1043V1_1_IRQS_PER_MSIR) {
-+ msir->bit_start = 32 - ((msir->index + 1) *
-+ MSI_LS1043V1_1_IRQS_PER_MSIR);
-+ msir->bit_end = msir->bit_start +
-+ MSI_LS1043V1_1_IRQS_PER_MSIR - 1;
-+ } else {
-+ msir->bit_start = 0;
-+ msir->bit_end = msi_data->cfg->msir_irqs - 1;
-+ }
-+
-+ irq_set_chained_handler_and_data(msir->gic_irq,
-+ ls_scfg_msi_irq_handler,
-+ msir);
-+
-+ if (msi_affinity_flag) {
-+ /* Associate MSIR interrupt to the cpu */
-+ irq_set_affinity(msir->gic_irq, get_cpu_mask(index));
-+ msir->srs = 0; /* This value is determined by the CPU */
-+ } else
-+ msir->srs = index;
-+
-+ /* Release the hwirqs corresponding to this MSIR */
-+ if (!msi_affinity_flag || msir->index == 0) {
-+ for (i = 0; i < msi_data->cfg->msir_irqs; i++) {
-+ hwirq = i << msi_data->cfg->ibs_shift | msir->index;
-+ bitmap_clear(msi_data->used, hwirq, 1);
-+ }
-+ }
-+
-+ return 0;
-+}
-+
-+static int ls_scfg_msi_teardown_hwirq(struct ls_scfg_msir *msir)
-+{
-+ struct ls_scfg_msi *msi_data = msir->msi_data;
-+ int i, hwirq;
-+
-+ if (msir->gic_irq > 0)
-+ irq_set_chained_handler_and_data(msir->gic_irq, NULL, NULL);
-+
-+ for (i = 0; i < msi_data->cfg->msir_irqs; i++) {
-+ hwirq = i << msi_data->cfg->ibs_shift | msir->index;
-+ bitmap_set(msi_data->used, hwirq, 1);
-+ }
-+
-+ return 0;
-+}
-+
-+static struct ls_scfg_msi_cfg ls1021_msi_cfg = {
-+ .ibs_shift = 3,
-+ .msir_irqs = MSI_IRQS_PER_MSIR,
-+ .msir_base = MSI_MSIR_OFFSET,
-+};
-+
-+static struct ls_scfg_msi_cfg ls1046_msi_cfg = {
-+ .ibs_shift = 2,
-+ .msir_irqs = MSI_IRQS_PER_MSIR,
-+ .msir_base = MSI_MSIR_OFFSET,
-+};
-+
-+static struct ls_scfg_msi_cfg ls1043_v1_1_msi_cfg = {
-+ .ibs_shift = 2,
-+ .msir_irqs = MSI_LS1043V1_1_IRQS_PER_MSIR,
-+ .msir_base = MSI_LS1043V1_1_MSIR_OFFSET,
-+};
-+
-+static const struct of_device_id ls_scfg_msi_id[] = {
-+ /* The following two misspelled compatibles are obsolete */
-+ { .compatible = "fsl,1s1021a-msi", .data = &ls1021_msi_cfg},
-+ { .compatible = "fsl,1s1043a-msi", .data = &ls1021_msi_cfg},
-+
-+ { .compatible = "fsl,ls1012a-msi", .data = &ls1021_msi_cfg },
-+ { .compatible = "fsl,ls1021a-msi", .data = &ls1021_msi_cfg },
-+ { .compatible = "fsl,ls1043a-msi", .data = &ls1021_msi_cfg },
-+ { .compatible = "fsl,ls1043a-v1.1-msi", .data = &ls1043_v1_1_msi_cfg },
-+ { .compatible = "fsl,ls1046a-msi", .data = &ls1046_msi_cfg },
-+ {},
-+};
-+MODULE_DEVICE_TABLE(of, ls_scfg_msi_id);
-+
- static int ls_scfg_msi_probe(struct platform_device *pdev)
- {
-+ const struct of_device_id *match;
- struct ls_scfg_msi *msi_data;
- struct resource *res;
-- int ret;
-+ int i, ret;
-+
-+ match = of_match_device(ls_scfg_msi_id, &pdev->dev);
-+ if (!match)
-+ return -ENODEV;
-
- msi_data = devm_kzalloc(&pdev->dev, sizeof(*msi_data), GFP_KERNEL);
- if (!msi_data)
- return -ENOMEM;
-
-+ msi_data->cfg = (struct ls_scfg_msi_cfg *) match->data;
-+
- res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
- msi_data->regs = devm_ioremap_resource(&pdev->dev, res);
- if (IS_ERR(msi_data->regs)) {
-@@ -182,23 +350,48 @@ static int ls_scfg_msi_probe(struct plat
- }
- msi_data->msiir_addr = res->start;
-
-- msi_data->irq = platform_get_irq(pdev, 0);
-- if (msi_data->irq <= 0) {
-- dev_err(&pdev->dev, "failed to get MSI irq\n");
-- return -ENODEV;
-- }
--
- msi_data->pdev = pdev;
- spin_lock_init(&msi_data->lock);
-
-+ msi_data->irqs_num = MSI_IRQS_PER_MSIR *
-+ (1 << msi_data->cfg->ibs_shift);
-+ msi_data->used = devm_kcalloc(&pdev->dev,
-+ BITS_TO_LONGS(msi_data->irqs_num),
-+ sizeof(*msi_data->used),
-+ GFP_KERNEL);
-+ if (!msi_data->used)
-+ return -ENOMEM;
-+ /*
-+ * Reserve all the hwirqs
-+ * The available hwirqs will be released in ls1_msi_setup_hwirq()
-+ */
-+ bitmap_set(msi_data->used, 0, msi_data->irqs_num);
-+
-+ msi_data->msir_num = of_irq_count(pdev->dev.of_node);
-+
-+ if (msi_affinity_flag) {
-+ u32 cpu_num;
-+
-+ cpu_num = num_possible_cpus();
-+ if (msi_data->msir_num >= cpu_num)
-+ msi_data->msir_num = cpu_num;
-+ else
-+ msi_affinity_flag = 0;
-+ }
-+
-+ msi_data->msir = devm_kcalloc(&pdev->dev, msi_data->msir_num,
-+ sizeof(*msi_data->msir),
-+ GFP_KERNEL);
-+ if (!msi_data->msir)
-+ return -ENOMEM;
-+
-+ for (i = 0; i < msi_data->msir_num; i++)
-+ ls_scfg_msi_setup_hwirq(msi_data, i);
-+
- ret = ls_scfg_msi_domains_init(msi_data);
- if (ret)
- return ret;
-
-- irq_set_chained_handler_and_data(msi_data->irq,
-- ls_scfg_msi_irq_handler,
-- msi_data);
--
- platform_set_drvdata(pdev, msi_data);
-
- return 0;
-@@ -207,8 +400,10 @@ static int ls_scfg_msi_probe(struct plat
- static int ls_scfg_msi_remove(struct platform_device *pdev)
- {
- struct ls_scfg_msi *msi_data = platform_get_drvdata(pdev);
-+ int i;
-
-- irq_set_chained_handler_and_data(msi_data->irq, NULL, NULL);
-+ for (i = 0; i < msi_data->msir_num; i++)
-+ ls_scfg_msi_teardown_hwirq(&msi_data->msir[i]);
-
- irq_domain_remove(msi_data->msi_domain);
- irq_domain_remove(msi_data->parent);
-@@ -218,12 +413,6 @@ static int ls_scfg_msi_remove(struct pla
- return 0;
- }
-
--static const struct of_device_id ls_scfg_msi_id[] = {
-- { .compatible = "fsl,1s1021a-msi", },
-- { .compatible = "fsl,1s1043a-msi", },
-- {},
--};
--
- static struct platform_driver ls_scfg_msi_driver = {
- .driver = {
- .name = "ls-scfg-msi",
---- a/drivers/pci/host/Makefile
-+++ b/drivers/pci/host/Makefile
-@@ -17,7 +17,7 @@ obj-$(CONFIG_PCIE_XILINX) += pcie-xilinx
- obj-$(CONFIG_PCIE_XILINX_NWL) += pcie-xilinx-nwl.o
- obj-$(CONFIG_PCI_XGENE) += pci-xgene.o
- obj-$(CONFIG_PCI_XGENE_MSI) += pci-xgene-msi.o
--obj-$(CONFIG_PCI_LAYERSCAPE) += pci-layerscape.o
-+obj-$(CONFIG_PCI_LAYERSCAPE) += pci-layerscape.o pci-layerscape-ep.o pci-layerscape-ep-debugfs.o
- obj-$(CONFIG_PCI_VERSATILE) += pci-versatile.o
- obj-$(CONFIG_PCIE_IPROC) += pcie-iproc.o
- obj-$(CONFIG_PCIE_IPROC_MSI) += pcie-iproc-msi.o
---- /dev/null
-+++ b/drivers/pci/host/pci-layerscape-ep-debugfs.c
-@@ -0,0 +1,758 @@
-+/*
-+ * PCIe Endpoint driver for Freescale Layerscape SoCs
-+ *
-+ * Copyright (C) 2015 Freescale Semiconductor.
-+ *
-+ * Author: Minghuan Lian <Minghuan.Lian@freescale.com>
-+ *
-+ * This program is free software; you can redistribute it and/or modify
-+ * it under the terms of the GNU General Public License version 2 as
-+ * published by the Free Software Foundation.
-+ */
-+
-+#include <linux/kernel.h>
-+#include <linux/module.h>
-+#include <linux/debugfs.h>
-+#include <linux/time.h>
-+#include <linux/uaccess.h>
-+#include <linux/kthread.h>
-+#include <linux/slab.h>
-+#include <linux/dmaengine.h>
-+#include <linux/dma-mapping.h>
-+#include <linux/freezer.h>
-+
-+#include <linux/completion.h>
-+
-+#include "pci-layerscape-ep.h"
-+
-+#define PCIE_ATU_INDEX3 (0x3 << 0)
-+#define PCIE_ATU_INDEX2 (0x2 << 0)
-+#define PCIE_ATU_INDEX1 (0x1 << 0)
-+#define PCIE_ATU_INDEX0 (0x0 << 0)
-+
-+#define PCIE_BAR0_SIZE (4 * 1024) /* 4K */
-+#define PCIE_BAR1_SIZE (8 * 1024) /* 8K for MSIX */
-+#define PCIE_BAR2_SIZE (4 * 1024) /* 4K */
-+#define PCIE_BAR4_SIZE (1 * 1024 * 1024) /* 1M */
-+#define PCIE_MSI_OB_SIZE (4 * 1024) /* 4K */
-+
-+#define PCIE_MSI_MSG_ADDR_OFF 0x54
-+#define PCIE_MSI_MSG_DATA_OFF 0x5c
-+
-+enum test_type {
-+ TEST_TYPE_DMA,
-+ TEST_TYPE_MEMCPY
-+};
-+
-+enum test_dirt {
-+ TEST_DIRT_READ,
-+ TEST_DIRT_WRITE
-+};
-+
-+enum test_status {
-+ TEST_IDLE,
-+ TEST_BUSY
-+};
-+
-+struct ls_ep_test {
-+ struct ls_ep_dev *ep;
-+ void __iomem *cfg;
-+ void __iomem *buf;
-+ void __iomem *out;
-+ void __iomem *msi;
-+ dma_addr_t cfg_addr;
-+ dma_addr_t buf_addr;
-+ dma_addr_t out_addr;
-+ dma_addr_t bus_addr;
-+ dma_addr_t msi_addr;
-+ u64 msi_msg_addr;
-+ u16 msi_msg_data;
-+ struct task_struct *thread;
-+ spinlock_t lock;
-+ struct completion done;
-+ u32 len;
-+ int loop;
-+ char data;
-+ enum test_dirt dirt;
-+ enum test_type type;
-+ enum test_status status;
-+ u64 result; /* Mbps */
-+ char cmd[256];
-+};
-+
-+static int ls_pcie_ep_trigger_msi(struct ls_ep_test *test)
-+{
-+ if (!test->msi)
-+ return -EINVAL;
-+
-+ iowrite32(test->msi_msg_data, test->msi);
-+
-+ return 0;
-+}
-+
-+static int ls_pcie_ep_test_try_run(struct ls_ep_test *test)
-+{
-+ int ret;
-+
-+ spin_lock(&test->lock);
-+ if (test->status == TEST_IDLE) {
-+ test->status = TEST_BUSY;
-+ ret = 0;
-+ } else
-+ ret = -EBUSY;
-+ spin_unlock(&test->lock);
-+
-+ return ret;
-+}
-+
-+static void ls_pcie_ep_test_done(struct ls_ep_test *test)
-+{
-+ spin_lock(&test->lock);
-+ test->status = TEST_IDLE;
-+ spin_unlock(&test->lock);
-+}
-+
-+static void ls_pcie_ep_test_dma_cb(void *arg)
-+{
-+ struct ls_ep_test *test = arg;
-+
-+ complete(&test->done);
-+}
-+
-+static int ls_pcie_ep_test_dma(struct ls_ep_test *test)
-+{
-+ dma_cap_mask_t mask;
-+ struct dma_chan *chan;
-+ struct dma_device *dma_dev;
-+ dma_addr_t src, dst;
-+ enum dma_data_direction direction;
-+ enum dma_ctrl_flags dma_flags = DMA_CTRL_ACK | DMA_PREP_INTERRUPT;
-+ struct timespec start, end, period;
-+ int i = 0;
-+
-+ dma_cap_zero(mask);
-+ dma_cap_set(DMA_MEMCPY, mask);
-+
-+ chan = dma_request_channel(mask, NULL, test);
-+ if (!chan) {
-+ pr_err("failed to request dma channel\n");
-+ return -EINVAL;
-+ }
-+
-+ memset(test->buf, test->data, test->len);
-+
-+ if (test->dirt == TEST_DIRT_WRITE) {
-+ src = test->buf_addr;
-+ dst = test->out_addr;
-+ direction = DMA_TO_DEVICE;
-+ } else {
-+ src = test->out_addr;
-+ dst = test->buf_addr;
-+ direction = DMA_FROM_DEVICE;
-+ }
-+
-+ dma_dev = chan->device;
-+ dma_flags = DMA_CTRL_ACK | DMA_PREP_INTERRUPT;
-+
-+ dma_sync_single_for_device(&test->ep->dev, test->buf_addr,
-+ test->len, direction);
-+
-+ set_freezable();
-+
-+ getrawmonotonic(&start);
-+ while (!kthread_should_stop() && (i < test->loop)) {
-+ struct dma_async_tx_descriptor *dma_desc;
-+ dma_cookie_t dma_cookie = {0};
-+ unsigned long tmo;
-+ int status;
-+
-+ init_completion(&test->done);
-+
-+ dma_desc = dma_dev->device_prep_dma_memcpy(chan,
-+ dst, src,
-+ test->len,
-+ dma_flags);
-+ if (!dma_desc) {
-+ pr_err("DMA desc constr failed...\n");
-+ goto _err;
-+ }
-+
-+ dma_desc->callback = ls_pcie_ep_test_dma_cb;
-+ dma_desc->callback_param = test;
-+ dma_cookie = dmaengine_submit(dma_desc);
-+
-+ if (dma_submit_error(dma_cookie)) {
-+ pr_err("DMA submit error....\n");
-+ goto _err;
-+ }
-+
-+ /* Trigger the transaction */
-+ dma_async_issue_pending(chan);
-+
-+ tmo = wait_for_completion_timeout(&test->done,
-+ msecs_to_jiffies(5 * test->len));
-+ if (tmo == 0) {
-+ pr_err("Self-test copy timed out, disabling\n");
-+ goto _err;
-+ }
-+
-+ status = dma_async_is_tx_complete(chan, dma_cookie,
-+ NULL, NULL);
-+ if (status != DMA_COMPLETE) {
-+ pr_err("got completion callback, but status is %s\n",
-+ status == DMA_ERROR ? "error" : "in progress");
-+ goto _err;
-+ }
-+
-+ i++;
-+ }
-+
-+ getrawmonotonic(&end);
-+ period = timespec_sub(end, start);
-+ test->result = test->len * 8ULL * i * 1000;
-+ do_div(test->result, period.tv_sec * 1000 * 1000 * 1000 + period.tv_nsec);
-+ dma_release_channel(chan);
-+
-+ return 0;
-+
-+_err:
-+ dma_release_channel(chan);
-+ test->result = 0;
-+ return -EINVAL;
-+}
-+
-+static int ls_pcie_ep_test_cpy(struct ls_ep_test *test)
-+{
-+ void *dst, *src;
-+ struct timespec start, end, period;
-+ int i = 0;
-+
-+ memset(test->buf, test->data, test->len);
-+
-+ if (test->dirt == TEST_DIRT_WRITE) {
-+ dst = test->out;
-+ src = test->buf;
-+ } else {
-+ dst = test->buf;
-+ src = test->out;
-+ }
-+
-+ getrawmonotonic(&start);
-+ while (!kthread_should_stop() && i < test->loop) {
-+ memcpy(dst, src, test->len);
-+ i++;
-+ }
-+ getrawmonotonic(&end);
-+
-+ period = timespec_sub(end, start);
-+ test->result = test->len * 8ULL * i * 1000;
-+ do_div(test->result, period.tv_sec * 1000 * 1000 * 1000 + period.tv_nsec);
-+
-+ return 0;
-+}
-+
-+int ls_pcie_ep_test_thread(void *arg)
-+{
-+ int ret;
-+
-+ struct ls_ep_test *test = arg;
-+
-+ if (test->type == TEST_TYPE_DMA)
-+ ret = ls_pcie_ep_test_dma(test);
-+ else
-+ ret = ls_pcie_ep_test_cpy(test);
-+
-+ if (ret) {
-+ pr_err("\n%s \ttest failed\n",
-+ test->cmd);
-+ test->result = 0;
-+ } else
-+ pr_err("\n%s \tthroughput:%lluMbps\n",
-+ test->cmd, test->result);
-+
-+ ls_pcie_ep_test_done(test);
-+
-+ ls_pcie_ep_trigger_msi(test);
-+
-+ do_exit(0);
-+}
-+
-+static int ls_pcie_ep_free_test(struct ls_ep_dev *ep)
-+{
-+ struct ls_ep_test *test = ep->driver_data;
-+
-+ if (!test)
-+ return 0;
-+
-+ if (test->status == TEST_BUSY) {
-+ kthread_stop(test->thread);
-+ dev_info(&ep->dev,
-+ "test is running please wait and run again\n");
-+ return -EBUSY;
-+ }
-+
-+ if (test->buf)
-+ free_pages((unsigned long)test->buf,
-+ get_order(PCIE_BAR4_SIZE));
-+
-+ if (test->cfg)
-+ free_pages((unsigned long)test->cfg,
-+ get_order(PCIE_BAR2_SIZE));
-+
-+ if (test->out)
-+ iounmap(test->out);
-+
-+ kfree(test);
-+ ep->driver_data = NULL;
-+
-+ return 0;
-+}
-+
-+static int ls_pcie_ep_init_test(struct ls_ep_dev *ep, u64 bus_addr)
-+{
-+ struct ls_pcie *pcie = ep->pcie;
-+ struct ls_ep_test *test = ep->driver_data;
-+ int err;
-+
-+ if (test) {
-+ dev_info(&ep->dev,
-+ "Please use 'free' to remove the exiting test\n");
-+ return -EBUSY;
-+ }
-+
-+ test = kzalloc(sizeof(*test), GFP_KERNEL);
-+ if (!test)
-+ return -ENOMEM;
-+ ep->driver_data = test;
-+ test->ep = ep;
-+ spin_lock_init(&test->lock);
-+ test->status = TEST_IDLE;
-+
-+ test->buf = dma_alloc_coherent(pcie->dev, get_order(PCIE_BAR4_SIZE),
-+ &test->buf_addr,
-+ GFP_KERNEL);
-+ if (!test->buf) {
-+ dev_info(&ep->dev, "failed to get mem for bar4\n");
-+ err = -ENOMEM;
-+ goto _err;
-+ }
-+
-+ test->cfg = (void *)__get_free_pages(GFP_KERNEL | __GFP_ZERO,
-+ get_order(PCIE_BAR2_SIZE));
-+ if (!test->cfg) {
-+ dev_info(&ep->dev, "failed to get mem for bar4\n");
-+ err = -ENOMEM;
-+ goto _err;
-+ }
-+ test->cfg_addr = virt_to_phys(test->cfg);
-+
-+ test->out_addr = pcie->out_base;
-+ test->out = ioremap(test->out_addr, PCIE_BAR4_SIZE);
-+ if (!test->out) {
-+ dev_info(&ep->dev, "failed to map out\n");
-+ err = -ENOMEM;
-+ goto _err;
-+ }
-+
-+ test->bus_addr = bus_addr;
-+
-+ test->msi_addr = test->out_addr + PCIE_BAR4_SIZE;
-+ test->msi = ioremap(test->msi_addr, PCIE_MSI_OB_SIZE);
-+ if (!test->msi)
-+ dev_info(&ep->dev, "failed to map MSI outbound region\n");
-+
-+ test->msi_msg_addr = ioread32(pcie->dbi + PCIE_MSI_MSG_ADDR_OFF) |
-+ (((u64)ioread32(pcie->dbi + PCIE_MSI_MSG_ADDR_OFF + 4)) << 32);
-+ test->msi_msg_data = ioread16(pcie->dbi + PCIE_MSI_MSG_DATA_OFF);
-+
-+ ls_pcie_ep_dev_cfg_enable(ep);
-+
-+ /* outbound iATU for memory */
-+ ls_pcie_iatu_outbound_set(pcie, 0, PCIE_ATU_TYPE_MEM,
-+ test->out_addr, bus_addr, PCIE_BAR4_SIZE);
-+ /* outbound iATU for MSI */
-+ ls_pcie_iatu_outbound_set(pcie, 1, PCIE_ATU_TYPE_MEM,
-+ test->msi_addr, test->msi_msg_addr,
-+ PCIE_MSI_OB_SIZE);
-+
-+ /* ATU 0 : INBOUND : map BAR0 */
-+ ls_pcie_iatu_inbound_set(pcie, 0, 0, test->cfg_addr);
-+ /* ATU 2 : INBOUND : map BAR2 */
-+ ls_pcie_iatu_inbound_set(pcie, 2, 2, test->cfg_addr);
-+ /* ATU 3 : INBOUND : map BAR4 */
-+ ls_pcie_iatu_inbound_set(pcie, 3, 4, test->buf_addr);
-+
-+ return 0;
-+
-+_err:
-+ ls_pcie_ep_free_test(ep);
-+ return err;
-+}
-+
-+static int ls_pcie_ep_start_test(struct ls_ep_dev *ep, char *cmd)
-+{
-+ struct ls_ep_test *test = ep->driver_data;
-+ enum test_type type;
-+ enum test_dirt dirt;
-+ u32 cnt, len, loop;
-+ unsigned int data;
-+ char dirt_str[2];
-+ int ret;
-+
-+ if (strncmp(cmd, "dma", 3) == 0)
-+ type = TEST_TYPE_DMA;
-+ else
-+ type = TEST_TYPE_MEMCPY;
-+
-+ cnt = sscanf(&cmd[4], "%1s %u %u %x", dirt_str, &len, &loop, &data);
-+ if (cnt != 4) {
-+ dev_info(&ep->dev, "format error %s", cmd);
-+ dev_info(&ep->dev, "dma/cpy <r/w> <packet_size> <loop> <data>\n");
-+ return -EINVAL;
-+ }
-+
-+ if (strncmp(dirt_str, "r", 1) == 0)
-+ dirt = TEST_DIRT_READ;
-+ else
-+ dirt = TEST_DIRT_WRITE;
-+
-+ if (len > PCIE_BAR4_SIZE) {
-+ dev_err(&ep->dev, "max len is %d", PCIE_BAR4_SIZE);
-+ return -EINVAL;
-+ }
-+
-+ if (!test) {
-+ dev_err(&ep->dev, "Please first run init command\n");
-+ return -EINVAL;
-+ }
-+
-+ if (ls_pcie_ep_test_try_run(test)) {
-+ dev_err(&ep->dev, "There is already a test running\n");
-+ return -EINVAL;
-+ }
-+
-+ test->len = len;
-+ test->loop = loop;
-+ test->type = type;
-+ test->data = (char)data;
-+ test->dirt = dirt;
-+ strcpy(test->cmd, cmd);
-+ test->thread = kthread_run(ls_pcie_ep_test_thread, test,
-+ "pcie ep test");
-+ if (IS_ERR(test->thread)) {
-+ dev_err(&ep->dev, "fork failed for pcie ep test\n");
-+ ls_pcie_ep_test_done(test);
-+ ret = PTR_ERR(test->thread);
-+ }
-+
-+ return ret;
-+}
-+
-+
-+/**
-+ * ls_pcie_reg_ops_read - read for regs data
-+ * @filp: the opened file
-+ * @buffer: where to write the data for the user to read
-+ * @count: the size of the user's buffer
-+ * @ppos: file position offset
-+ **/
-+static ssize_t ls_pcie_ep_dbg_regs_read(struct file *filp, char __user *buffer,
-+ size_t count, loff_t *ppos)
-+{
-+ struct ls_ep_dev *ep = filp->private_data;
-+ struct ls_pcie *pcie = ep->pcie;
-+ char *buf;
-+ int desc = 0, i, len;
-+
-+ buf = kmalloc(4 * 1024, GFP_KERNEL);
-+ if (!buf)
-+ return -ENOMEM;
-+
-+ ls_pcie_ep_dev_cfg_enable(ep);
-+
-+ desc += sprintf(buf + desc, "%s", "reg info:");
-+ for (i = 0; i < 0x200; i += 4) {
-+ if (i % 16 == 0)
-+ desc += sprintf(buf + desc, "\n%08x:", i);
-+ desc += sprintf(buf + desc, " %08x", readl(pcie->dbi + i));
-+ }
-+
-+ desc += sprintf(buf + desc, "\n%s", "outbound iATU info:\n");
-+ for (i = 0; i < 6; i++) {
-+ writel(PCIE_ATU_REGION_OUTBOUND | i,
-+ pcie->dbi + PCIE_ATU_VIEWPORT);
-+ desc += sprintf(buf + desc, "iATU%d", i);
-+ desc += sprintf(buf + desc, "\tLOWER PHYS 0x%08x\n",
-+ readl(pcie->dbi + PCIE_ATU_LOWER_BASE));
-+ desc += sprintf(buf + desc, "\tUPPER PHYS 0x%08x\n",
-+ readl(pcie->dbi + PCIE_ATU_UPPER_BASE));
-+ desc += sprintf(buf + desc, "\tLOWER BUS 0x%08x\n",
-+ readl(pcie->dbi + PCIE_ATU_LOWER_TARGET));
-+ desc += sprintf(buf + desc, "\tUPPER BUS 0x%08x\n",
-+ readl(pcie->dbi + PCIE_ATU_UPPER_TARGET));
-+ desc += sprintf(buf + desc, "\tLIMIT 0x%08x\n",
-+ readl(pcie->dbi + PCIE_ATU_LIMIT));
-+ desc += sprintf(buf + desc, "\tCR1 0x%08x\n",
-+ readl(pcie->dbi + PCIE_ATU_CR1));
-+ desc += sprintf(buf + desc, "\tCR2 0x%08x\n",
-+ readl(pcie->dbi + PCIE_ATU_CR2));
-+ }
-+
-+ desc += sprintf(buf + desc, "\n%s", "inbound iATU info:\n");
-+ for (i = 0; i < 6; i++) {
-+ writel(PCIE_ATU_REGION_INBOUND | i,
-+ pcie->dbi + PCIE_ATU_VIEWPORT);
-+ desc += sprintf(buf + desc, "iATU%d", i);
-+ desc += sprintf(buf + desc, "\tLOWER BUS 0x%08x\n",
-+ readl(pcie->dbi + PCIE_ATU_LOWER_BASE));
-+ desc += sprintf(buf + desc, "\tUPPER BUSs 0x%08x\n",
-+ readl(pcie->dbi + PCIE_ATU_UPPER_BASE));
-+ desc += sprintf(buf + desc, "\tLOWER PHYS 0x%08x\n",
-+ readl(pcie->dbi + PCIE_ATU_LOWER_TARGET));
-+ desc += sprintf(buf + desc, "\tUPPER PHYS 0x%08x\n",
-+ readl(pcie->dbi + PCIE_ATU_UPPER_TARGET));
-+ desc += sprintf(buf + desc, "\tLIMIT 0x%08x\n",
-+ readl(pcie->dbi + PCIE_ATU_LIMIT));
-+ desc += sprintf(buf + desc, "\tCR1 0x%08x\n",
-+ readl(pcie->dbi + PCIE_ATU_CR1));
-+ desc += sprintf(buf + desc, "\tCR2 0x%08x\n",
-+ readl(pcie->dbi + PCIE_ATU_CR2));
-+ }
-+
-+ len = simple_read_from_buffer(buffer, count, ppos, buf, desc);
-+ kfree(buf);
-+
-+ return len;
-+}
-+
-+/**
-+ * ls_pcie_ep_dbg_regs_write - write into regs datum
-+ * @filp: the opened file
-+ * @buffer: where to find the user's data
-+ * @count: the length of the user's data
-+ * @ppos: file position offset
-+ **/
-+static ssize_t ls_pcie_ep_dbg_regs_write(struct file *filp,
-+ const char __user *buffer,
-+ size_t count, loff_t *ppos)
-+{
-+ struct ls_ep_dev *ep = filp->private_data;
-+ struct ls_pcie *pcie = ep->pcie;
-+ char buf[256];
-+
-+ if (count >= sizeof(buf))
-+ return -ENOSPC;
-+
-+ memset(buf, 0, sizeof(buf));
-+
-+ if (copy_from_user(buf, buffer, count))
-+ return -EFAULT;
-+
-+ ls_pcie_ep_dev_cfg_enable(ep);
-+
-+ if (strncmp(buf, "reg", 3) == 0) {
-+ u32 reg, value;
-+ int cnt;
-+
-+ cnt = sscanf(&buf[3], "%x %x", &reg, &value);
-+ if (cnt == 2) {
-+ writel(value, pcie->dbi + reg);
-+ value = readl(pcie->dbi + reg);
-+ dev_info(&ep->dev, "reg 0x%08x: 0x%08x\n",
-+ reg, value);
-+ } else {
-+ dev_info(&ep->dev, "reg <reg> <value>\n");
-+ }
-+ } else if (strncmp(buf, "atu", 3) == 0) {
-+ /* to do */
-+ dev_info(&ep->dev, " Not support atu command\n");
-+ } else {
-+ dev_info(&ep->dev, "Unknown command %s\n", buf);
-+ dev_info(&ep->dev, "Available commands:\n");
-+ dev_info(&ep->dev, " reg <reg> <value>\n");
-+ }
-+
-+ return count;
-+}
-+
-+static const struct file_operations ls_pcie_ep_dbg_regs_fops = {
-+ .owner = THIS_MODULE,
-+ .open = simple_open,
-+ .read = ls_pcie_ep_dbg_regs_read,
-+ .write = ls_pcie_ep_dbg_regs_write,
-+};
-+
-+static ssize_t ls_pcie_ep_dbg_test_read(struct file *filp,
-+ char __user *buffer,
-+ size_t count, loff_t *ppos)
-+{
-+ struct ls_ep_dev *ep = filp->private_data;
-+ struct ls_ep_test *test = ep->driver_data;
-+ char buf[512];
-+ int desc = 0, len;
-+
-+ if (!test) {
-+ dev_info(&ep->dev, " there is NO test\n");
-+ return 0;
-+ }
-+
-+ if (test->status != TEST_IDLE) {
-+ dev_info(&ep->dev, "test %s is running\n", test->cmd);
-+ return 0;
-+ }
-+
-+ desc = sprintf(buf, "MSI ADDR:0x%llx MSI DATA:0x%x\n",
-+ test->msi_msg_addr, test->msi_msg_data);
-+
-+ desc += sprintf(buf + desc, "%s throughput:%lluMbps\n",
-+ test->cmd, test->result);
-+
-+ len = simple_read_from_buffer(buffer, count, ppos,
-+ buf, desc);
-+
-+ return len;
-+}
-+
-+static ssize_t ls_pcie_ep_dbg_test_write(struct file *filp,
-+ const char __user *buffer,
-+ size_t count, loff_t *ppos)
-+{
-+ struct ls_ep_dev *ep = filp->private_data;
-+ char buf[256];
-+
-+ if (count >= sizeof(buf))
-+ return -ENOSPC;
-+
-+ memset(buf, 0, sizeof(buf));
-+
-+ if (copy_from_user(buf, buffer, count))
-+ return -EFAULT;
-+
-+ if (strncmp(buf, "init", 4) == 0) {
-+ int i = 4;
-+ u64 bus_addr;
-+
-+ while (buf[i] == ' ')
-+ i++;
-+
-+ if (kstrtou64(&buf[i], 0, &bus_addr))
-+ dev_info(&ep->dev, "command: init <bus_addr>\n");
-+ else {
-+ if (ls_pcie_ep_init_test(ep, bus_addr))
-+ dev_info(&ep->dev, "failed to init test\n");
-+ }
-+ } else if (strncmp(buf, "free", 4) == 0)
-+ ls_pcie_ep_free_test(ep);
-+ else if (strncmp(buf, "dma", 3) == 0 ||
-+ strncmp(buf, "cpy", 3) == 0)
-+ ls_pcie_ep_start_test(ep, buf);
-+ else {
-+ dev_info(&ep->dev, "Unknown command: %s\n", buf);
-+ dev_info(&ep->dev, "Available commands:\n");
-+ dev_info(&ep->dev, "\tinit <bus_addr>\n");
-+ dev_info(&ep->dev, "\t<dma/cpy> <r/w> <packet_size> <loop>\n");
-+ dev_info(&ep->dev, "\tfree\n");
-+ }
-+
-+ return count;
-+}
-+
-+static const struct file_operations ls_pcie_ep_dbg_test_fops = {
-+ .owner = THIS_MODULE,
-+ .open = simple_open,
-+ .read = ls_pcie_ep_dbg_test_read,
-+ .write = ls_pcie_ep_dbg_test_write,
-+};
-+
-+static ssize_t ls_pcie_ep_dbg_dump_read(struct file *filp,
-+ char __user *buffer,
-+ size_t count, loff_t *ppos)
-+{
-+ struct ls_ep_dev *ep = filp->private_data;
-+ struct ls_ep_test *test = ep->driver_data;
-+ char *buf;
-+ int desc = 0, i, len;
-+
-+ buf = kmalloc(4 * 1024, GFP_KERNEL);
-+ if (!buf)
-+ return -ENOMEM;
-+
-+ if (!test) {
-+ dev_info(&ep->dev, " there is NO test\n");
-+ kfree(buf);
-+ return 0;
-+ }
-+
-+ desc += sprintf(buf + desc, "%s", "dump info:");
-+ for (i = 0; i < 256; i += 4) {
-+ if (i % 16 == 0)
-+ desc += sprintf(buf + desc, "\n%08x:", i);
-+ desc += sprintf(buf + desc, " %08x", readl(test->buf + i));
-+ }
-+
-+ desc += sprintf(buf + desc, "\n");
-+ len = simple_read_from_buffer(buffer, count, ppos, buf, desc);
-+
-+ kfree(buf);
-+
-+ return len;
-+}
-+
-+static const struct file_operations ls_pcie_ep_dbg_dump_fops = {
-+ .owner = THIS_MODULE,
-+ .open = simple_open,
-+ .read = ls_pcie_ep_dbg_dump_read,
-+};
-+
-+static int ls_pcie_ep_dev_dbgfs_init(struct ls_ep_dev *ep)
-+{
-+ struct ls_pcie *pcie = ep->pcie;
-+ struct dentry *pfile;
-+
-+ ls_pcie_ep_dev_cfg_enable(ep);
-+
-+ ep->dir = debugfs_create_dir(dev_name(&ep->dev), pcie->dir);
-+ if (!ep->dir)
-+ return -ENOMEM;
-+
-+ pfile = debugfs_create_file("regs", 0600, ep->dir, ep,
-+ &ls_pcie_ep_dbg_regs_fops);
-+ if (!pfile)
-+ dev_info(&ep->dev, "debugfs regs for failed\n");
-+
-+ pfile = debugfs_create_file("test", 0600, ep->dir, ep,
-+ &ls_pcie_ep_dbg_test_fops);
-+ if (!pfile)
-+ dev_info(&ep->dev, "debugfs test for failed\n");
-+
-+ pfile = debugfs_create_file("dump", 0600, ep->dir, ep,
-+ &ls_pcie_ep_dbg_dump_fops);
-+ if (!pfile)
-+ dev_info(&ep->dev, "debugfs dump for failed\n");
-+
-+ return 0;
-+}
-+
-+int ls_pcie_ep_dbgfs_init(struct ls_pcie *pcie)
-+{
-+ struct ls_ep_dev *ep;
-+
-+ pcie->dir = debugfs_create_dir(dev_name(pcie->dev), NULL);
-+ if (!pcie->dir)
-+ return -ENOMEM;
-+
-+ list_for_each_entry(ep, &pcie->ep_list, node)
-+ ls_pcie_ep_dev_dbgfs_init(ep);
-+
-+ return 0;
-+}
-+
-+int ls_pcie_ep_dbgfs_remove(struct ls_pcie *pcie)
-+{
-+ debugfs_remove_recursive(pcie->dir);
-+ return 0;
-+}
-+
-+MODULE_AUTHOR("Minghuan Lian <Minghuan.Lian@freescale.com>");
-+MODULE_DESCRIPTION("Freescale Layerscape PCIe EP controller driver");
-+MODULE_LICENSE("GPL v2");
---- /dev/null
-+++ b/drivers/pci/host/pci-layerscape-ep.c
-@@ -0,0 +1,309 @@
-+/*
-+ * PCIe Endpoint driver for Freescale Layerscape SoCs
-+ *
-+ * Copyright (C) 2015 Freescale Semiconductor.
-+ *
-+ * Author: Minghuan Lian <Minghuan.Lian@freescale.com>
-+ *
-+ * This program is free software; you can redistribute it and/or modify
-+ * it under the terms of the GNU General Public License version 2 as
-+ * published by the Free Software Foundation.
-+ */
-+
-+#include <linux/kernel.h>
-+#include <linux/delay.h>
-+#include <linux/interrupt.h>
-+#include <linux/module.h>
-+#include <linux/of_pci.h>
-+#include <linux/of_platform.h>
-+#include <linux/of_irq.h>
-+#include <linux/of_address.h>
-+#include <linux/pci.h>
-+#include <linux/platform_device.h>
-+#include <linux/resource.h>
-+#include <linux/debugfs.h>
-+#include <linux/time.h>
-+#include <linux/uaccess.h>
-+
-+#include "pci-layerscape-ep.h"
-+
-+struct ls_ep_dev *
-+ls_pci_ep_find(struct ls_pcie *pcie, int dev_id)
-+{
-+ struct ls_ep_dev *ep;
-+
-+ list_for_each_entry(ep, &pcie->ep_list, node) {
-+ if (ep->dev_id == dev_id)
-+ return ep;
-+ }
-+
-+ return NULL;
-+}
-+
-+static void ls_pcie_try_cfg2(struct ls_pcie *pcie, int pf, int vf)
-+{
-+ if (pcie->sriov)
-+ writel(PCIE_LCTRL0_VAL(pf, vf),
-+ pcie->dbi + PCIE_LUT_BASE + PCIE_LUT_LCTRL0);
-+}
-+
-+static bool ls_pcie_is_bridge(struct ls_pcie *pcie)
-+{
-+ u32 header_type = 0;
-+
-+ header_type = readl(pcie->dbi + (PCI_HEADER_TYPE & ~0x3));
-+ header_type = (header_type >> 16) & 0x7f;
-+
-+ return header_type == PCI_HEADER_TYPE_BRIDGE;
-+}
-+
-+void ls_pcie_iatu_outbound_set(struct ls_pcie *pcie, int idx, int type,
-+ u64 cpu_addr, u64 pci_addr, u32 size)
-+{
-+ writel(PCIE_ATU_REGION_OUTBOUND | idx,
-+ pcie->dbi + PCIE_ATU_VIEWPORT);
-+ writel(lower_32_bits(cpu_addr),
-+ pcie->dbi + PCIE_ATU_LOWER_BASE);
-+ writel(upper_32_bits(cpu_addr),
-+ pcie->dbi + PCIE_ATU_UPPER_BASE);
-+ writel(lower_32_bits(cpu_addr + size - 1),
-+ pcie->dbi + PCIE_ATU_LIMIT);
-+ writel(lower_32_bits(pci_addr),
-+ pcie->dbi + PCIE_ATU_LOWER_TARGET);
-+ writel(upper_32_bits(pci_addr),
-+ pcie->dbi + PCIE_ATU_UPPER_TARGET);
-+ writel(type, pcie->dbi + PCIE_ATU_CR1);
-+ writel(PCIE_ATU_ENABLE, pcie->dbi + PCIE_ATU_CR2);
-+}
-+
-+/* Use bar match mode and MEM type as default */
-+void ls_pcie_iatu_inbound_set(struct ls_pcie *pcie, int idx,
-+ int bar, u64 phys)
-+{
-+ writel(PCIE_ATU_REGION_INBOUND | idx, pcie->dbi + PCIE_ATU_VIEWPORT);
-+ writel((u32)phys, pcie->dbi + PCIE_ATU_LOWER_TARGET);
-+ writel(phys >> 32, pcie->dbi + PCIE_ATU_UPPER_TARGET);
-+ writel(PCIE_ATU_TYPE_MEM, pcie->dbi + PCIE_ATU_CR1);
-+ writel(PCIE_ATU_ENABLE | PCIE_ATU_BAR_MODE_ENABLE |
-+ PCIE_ATU_BAR_NUM(bar), pcie->dbi + PCIE_ATU_CR2);
-+}
-+
-+void ls_pcie_ep_dev_cfg_enable(struct ls_ep_dev *ep)
-+{
-+ ls_pcie_try_cfg2(ep->pcie, ep->pf_idx, ep->vf_idx);
-+}
-+
-+void ls_pcie_ep_setup_bar(void *bar_base, int bar, u32 size)
-+{
-+ if (size < 4 * 1024)
-+ return;
-+
-+ switch (bar) {
-+ case 0:
-+ writel(size - 1, bar_base + PCI_BASE_ADDRESS_0);
-+ break;
-+ case 1:
-+ writel(size - 1, bar_base + PCI_BASE_ADDRESS_1);
-+ break;
-+ case 2:
-+ writel(size - 1, bar_base + PCI_BASE_ADDRESS_2);
-+ writel(0, bar_base + PCI_BASE_ADDRESS_3);
-+ break;
-+ case 4:
-+ writel(size - 1, bar_base + PCI_BASE_ADDRESS_4);
-+ writel(0, bar_base + PCI_BASE_ADDRESS_5);
-+ break;
-+ default:
-+ break;
-+ }
-+}
-+
-+void ls_pcie_ep_dev_setup_bar(struct ls_ep_dev *ep, int bar, u32 size)
-+{
-+ struct ls_pcie *pcie = ep->pcie;
-+ void *bar_base;
-+
-+ if (size < 4 * 1024)
-+ return;
-+
-+ if (pcie->sriov)
-+ bar_base = pcie->dbi;
-+ else
-+ bar_base = pcie->dbi + PCIE_NO_SRIOV_BAR_BASE;
-+
-+ ls_pcie_ep_dev_cfg_enable(ep);
-+ ls_pcie_ep_setup_bar(bar_base, bar, size);
-+}
-+
-+static int ls_pcie_ep_dev_init(struct ls_pcie *pcie, int pf_idx, int vf_idx)
-+{
-+ struct ls_ep_dev *ep;
-+
-+ ep = devm_kzalloc(pcie->dev, sizeof(*ep), GFP_KERNEL);
-+ if (!ep)
-+ return -ENOMEM;
-+
-+ ep->pcie = pcie;
-+ ep->pf_idx = pf_idx;
-+ ep->vf_idx = vf_idx;
-+ if (vf_idx)
-+ ep->dev_id = pf_idx + 4 + 4 * (vf_idx - 1);
-+ else
-+ ep->dev_id = pf_idx;
-+
-+ if (ep->vf_idx)
-+ dev_set_name(&ep->dev, "pf%d-vf%d",
-+ ep->pf_idx,
-+ ep->vf_idx);
-+ else
-+ dev_set_name(&ep->dev, "pf%d",
-+ ep->pf_idx);
-+
-+ list_add_tail(&ep->node, &pcie->ep_list);
-+
-+ return 0;
-+}
-+
-+static int ls_pcie_ep_init(struct ls_pcie *pcie)
-+{
-+ u32 sriov_header;
-+ int pf, vf, i, j;
-+
-+ sriov_header = readl(pcie->dbi + PCIE_SRIOV_POS);
-+
-+ if (PCI_EXT_CAP_ID(sriov_header) == PCI_EXT_CAP_ID_SRIOV) {
-+ pcie->sriov = PCIE_SRIOV_POS;
-+ pf = PCIE_PF_NUM;
-+ vf = PCIE_VF_NUM;
-+ } else {
-+ pcie->sriov = 0;
-+ pf = 1;
-+ vf = 0;
-+ }
-+
-+ for (i = 0; i < pf; i++) {
-+ for (j = 0; j <= vf; j++)
-+ ls_pcie_ep_dev_init(pcie, i, j);
-+ }
-+
-+ return 0;
-+}
-+
-+static struct ls_pcie_ep_drvdata ls1043_drvdata = {
-+ .lut_offset = 0x10000,
-+ .ltssm_shift = 24,
-+ .lut_dbg = 0x7fc,
-+};
-+
-+static struct ls_pcie_ep_drvdata ls1046_drvdata = {
-+ .lut_offset = 0x80000,
-+ .ltssm_shift = 24,
-+ .lut_dbg = 0x407fc,
-+};
-+
-+static struct ls_pcie_ep_drvdata ls2080_drvdata = {
-+ .lut_offset = 0x80000,
-+ .ltssm_shift = 0,
-+ .lut_dbg = 0x7fc,
-+};
-+
-+static const struct of_device_id ls_pcie_ep_of_match[] = {
-+ { .compatible = "fsl,ls1021a-pcie", },
-+ { .compatible = "fsl,ls1043a-pcie", .data = &ls1043_drvdata },
-+ { .compatible = "fsl,ls1046a-pcie", .data = &ls1046_drvdata },
-+ { .compatible = "fsl,ls2080a-pcie", .data = &ls2080_drvdata },
-+ { .compatible = "fsl,ls2085a-pcie", .data = &ls2080_drvdata },
-+ { },
-+};
-+MODULE_DEVICE_TABLE(of, ls_pcie_ep_of_match);
-+
-+static int ls_pcie_ep_probe(struct platform_device *pdev)
-+{
-+ struct ls_pcie *pcie;
-+ struct resource *dbi_base, *cfg_res;
-+ const struct of_device_id *match;
-+ int ret;
-+
-+ match = of_match_device(ls_pcie_ep_of_match, &pdev->dev);
-+ if (!match)
-+ return -ENODEV;
-+
-+ pcie = devm_kzalloc(&pdev->dev, sizeof(*pcie), GFP_KERNEL);
-+ if (!pcie)
-+ return -ENOMEM;
-+
-+ pcie->dev = &pdev->dev;
-+ INIT_LIST_HEAD(&pcie->ep_list);
-+
-+ dbi_base = platform_get_resource_byname(pdev, IORESOURCE_MEM, "regs");
-+ pcie->dbi = devm_ioremap_resource(&pdev->dev, dbi_base);
-+ if (IS_ERR(pcie->dbi)) {
-+ dev_err(&pdev->dev, "missing *regs* space\n");
-+ return PTR_ERR(pcie->dbi);
-+ }
-+
-+ pcie->drvdata = match->data;
-+ pcie->lut = pcie->dbi + pcie->drvdata->lut_offset;
-+
-+ if (ls_pcie_is_bridge(pcie))
-+ return -ENODEV;
-+
-+ dev_info(pcie->dev, "in EP mode\n");
-+
-+ cfg_res = platform_get_resource_byname(pdev, IORESOURCE_MEM, "config");
-+ if (cfg_res)
-+ pcie->out_base = cfg_res->start;
-+ else {
-+ dev_err(&pdev->dev, "missing *config* space\n");
-+ return -ENODEV;
-+ }
-+
-+ ret = ls_pcie_ep_init(pcie);
-+ if (ret)
-+ return ret;
-+
-+ ls_pcie_ep_dbgfs_init(pcie);
-+
-+ platform_set_drvdata(pdev, pcie);
-+
-+ return 0;
-+}
-+
-+static int ls_pcie_ep_dev_remove(struct ls_ep_dev *ep)
-+{
-+ list_del(&ep->node);
-+
-+ return 0;
-+}
-+
-+static int ls_pcie_ep_remove(struct platform_device *pdev)
-+{
-+ struct ls_pcie *pcie = platform_get_drvdata(pdev);
-+ struct ls_ep_dev *ep, *tmp;
-+
-+ if (!pcie)
-+ return 0;
-+
-+ ls_pcie_ep_dbgfs_remove(pcie);
-+
-+ list_for_each_entry_safe(ep, tmp, &pcie->ep_list, node)
-+ ls_pcie_ep_dev_remove(ep);
-+
-+ return 0;
-+}
-+
-+static struct platform_driver ls_pcie_ep_driver = {
-+ .driver = {
-+ .name = "ls-pcie-ep",
-+ .owner = THIS_MODULE,
-+ .of_match_table = ls_pcie_ep_of_match,
-+ },
-+ .probe = ls_pcie_ep_probe,
-+ .remove = ls_pcie_ep_remove,
-+};
-+
-+module_platform_driver(ls_pcie_ep_driver);
-+
-+MODULE_AUTHOR("Minghuan Lian <Minghuan.Lian@freescale.com>");
-+MODULE_DESCRIPTION("Freescale Layerscape PCIe EP driver");
-+MODULE_LICENSE("GPL v2");
---- /dev/null
-+++ b/drivers/pci/host/pci-layerscape-ep.h
-@@ -0,0 +1,115 @@
-+/*
-+ * PCIe Endpoint driver for Freescale Layerscape SoCs
-+ *
-+ * Copyright (C) 2015 Freescale Semiconductor.
-+ *
-+ * Author: Minghuan Lian <Minghuan.Lian@freescale.com>
-+ *
-+ * This program is free software; you can redistribute it and/or modify
-+ * it under the terms of the GNU General Public License version 2 as
-+ * published by the Free Software Foundation.
-+ */
-+
-+
-+#ifndef _PCIE_LAYERSCAPE_EP_H
-+#define _PCIE_LAYERSCAPE_EP_H
-+
-+#include <linux/device.h>
-+
-+/* Synopsis specific PCIE configuration registers */
-+#define PCIE_ATU_VIEWPORT 0x900
-+#define PCIE_ATU_REGION_INBOUND (0x1 << 31)
-+#define PCIE_ATU_REGION_OUTBOUND (0x0 << 31)
-+#define PCIE_ATU_REGION_INDEX3 (0x3 << 0)
-+#define PCIE_ATU_REGION_INDEX2 (0x2 << 0)
-+#define PCIE_ATU_REGION_INDEX1 (0x1 << 0)
-+#define PCIE_ATU_REGION_INDEX0 (0x0 << 0)
-+#define PCIE_ATU_CR1 0x904
-+#define PCIE_ATU_TYPE_MEM (0x0 << 0)
-+#define PCIE_ATU_TYPE_IO (0x2 << 0)
-+#define PCIE_ATU_TYPE_CFG0 (0x4 << 0)
-+#define PCIE_ATU_TYPE_CFG1 (0x5 << 0)
-+#define PCIE_ATU_CR2 0x908
-+#define PCIE_ATU_ENABLE (0x1 << 31)
-+#define PCIE_ATU_BAR_MODE_ENABLE (0x1 << 30)
-+#define PCIE_ATU_LOWER_BASE 0x90C
-+#define PCIE_ATU_UPPER_BASE 0x910
-+#define PCIE_ATU_LIMIT 0x914
-+#define PCIE_ATU_LOWER_TARGET 0x918
-+#define PCIE_ATU_BUS(x) (((x) & 0xff) << 24)
-+#define PCIE_ATU_DEV(x) (((x) & 0x1f) << 19)
-+#define PCIE_ATU_FUNC(x) (((x) & 0x7) << 16)
-+#define PCIE_ATU_UPPER_TARGET 0x91C
-+
-+/* PEX internal configuration registers */
-+#define PCIE_DBI_RO_WR_EN 0x8bc /* DBI Read-Only Write Enable Register */
-+
-+/* PEX LUT registers */
-+#define PCIE_LUT_BASE 0x80000
-+#define PCIE_LUT_DBG 0x7FC /* PEX LUT Debug register */
-+
-+#define PCIE_LUT_LCTRL0 0x7F8
-+
-+#define PCIE_ATU_BAR_NUM(bar) ((bar) << 8)
-+#define PCIE_LCTRL0_CFG2_ENABLE (1 << 31)
-+#define PCIE_LCTRL0_VF(vf) ((vf) << 22)
-+#define PCIE_LCTRL0_PF(pf) ((pf) << 16)
-+#define PCIE_LCTRL0_VF_ACTIVE (1 << 21)
-+#define PCIE_LCTRL0_VAL(pf, vf) (PCIE_LCTRL0_PF(pf) | \
-+ PCIE_LCTRL0_VF(vf) | \
-+ ((vf) == 0 ? 0 : PCIE_LCTRL0_VF_ACTIVE) | \
-+ PCIE_LCTRL0_CFG2_ENABLE)
-+
-+#define PCIE_NO_SRIOV_BAR_BASE 0x1000
-+
-+#define PCIE_SRIOV_POS 0x178
-+#define PCIE_PF_NUM 2
-+#define PCIE_VF_NUM 64
-+
-+struct ls_pcie_ep_drvdata {
-+ u32 lut_offset;
-+ u32 ltssm_shift;
-+ u32 lut_dbg;
-+};
-+
-+struct ls_pcie {
-+ struct list_head ep_list;
-+ struct device *dev;
-+ struct dentry *dir;
-+ const struct ls_pcie_ep_drvdata *drvdata;
-+ void __iomem *dbi;
-+ void __iomem *lut;
-+ phys_addr_t out_base;
-+ int sriov;
-+ int index;
-+};
-+
-+struct ls_ep_dev {
-+ struct list_head node;
-+ struct ls_pcie *pcie;
-+ struct device dev;
-+ struct dentry *dir;
-+ int pf_idx;
-+ int vf_idx;
-+ int dev_id;
-+ void *driver_data;
-+};
-+
-+struct ls_ep_dev *ls_pci_ep_find(struct ls_pcie *pcie, int dev_id);
-+
-+void ls_pcie_iatu_outbound_set(struct ls_pcie *pcie, int idx, int type,
-+ u64 cpu_addr, u64 pci_addr, u32 size);
-+
-+/* Use bar match mode and MEM type as default */
-+void ls_pcie_iatu_inbound_set(struct ls_pcie *pcie, int idx,
-+ int bar, u64 phys);
-+
-+void ls_pcie_ep_dev_setup_bar(struct ls_ep_dev *ep, int bar, u32 size);
-+
-+
-+void ls_pcie_ep_dev_cfg_enable(struct ls_ep_dev *ep);
-+
-+int ls_pcie_ep_dbgfs_init(struct ls_pcie *pcie);
-+int ls_pcie_ep_dbgfs_remove(struct ls_pcie *pcie);
-+
-+#endif /* _PCIE_LAYERSCAPE_EP_H */
---- a/drivers/pci/host/pci-layerscape.c
-+++ b/drivers/pci/host/pci-layerscape.c
-@@ -33,14 +33,18 @@
-
- /* PEX Internal Configuration Registers */
- #define PCIE_STRFMR1 0x71c /* Symbol Timer & Filter Mask Register1 */
-+#define PCIE_ABSERR 0x8d0 /* Bridge Slave Error Response Register */
-+#define PCIE_ABSERR_SETTING 0x9401 /* Forward error of non-posted request */
- #define PCIE_DBI_RO_WR_EN 0x8bc /* DBI Read-Only Write Enable Register */
-
--/* PEX LUT registers */
--#define PCIE_LUT_DBG 0x7FC /* PEX LUT Debug Register */
-+#define PCIE_IATU_NUM 6
-+
-+static void ls_pcie_host_init(struct pcie_port *pp);
-
- struct ls_pcie_drvdata {
- u32 lut_offset;
- u32 ltssm_shift;
-+ u32 lut_dbg;
- struct pcie_host_ops *ops;
- };
-
-@@ -86,6 +90,14 @@ static void ls_pcie_drop_msg_tlp(struct
- iowrite32(val, pcie->pp.dbi_base + PCIE_STRFMR1);
- }
-
-+static void ls_pcie_disable_outbound_atus(struct ls_pcie *pcie)
-+{
-+ int i;
-+
-+ for (i = 0; i < PCIE_IATU_NUM; i++)
-+ dw_pcie_disable_outbound_atu(&pcie->pp, i);
-+}
-+
- static int ls1021_pcie_link_up(struct pcie_port *pp)
- {
- u32 state;
-@@ -134,7 +146,7 @@ static int ls_pcie_link_up(struct pcie_p
- struct ls_pcie *pcie = to_ls_pcie(pp);
- u32 state;
-
-- state = (ioread32(pcie->lut + PCIE_LUT_DBG) >>
-+ state = (ioread32(pcie->lut + pcie->drvdata->lut_dbg) >>
- pcie->drvdata->ltssm_shift) &
- LTSSM_STATE_MASK;
-
-@@ -144,6 +156,12 @@ static int ls_pcie_link_up(struct pcie_p
- return 1;
- }
-
-+/* Forward error response of outbound non-posted requests */
-+static void ls_pcie_fix_error_response(struct ls_pcie *pcie)
-+{
-+ iowrite32(PCIE_ABSERR_SETTING, pcie->pp.dbi_base + PCIE_ABSERR);
-+}
-+
- static void ls_pcie_host_init(struct pcie_port *pp)
- {
- struct ls_pcie *pcie = to_ls_pcie(pp);
-@@ -153,6 +171,10 @@ static void ls_pcie_host_init(struct pci
- ls_pcie_clear_multifunction(pcie);
- ls_pcie_drop_msg_tlp(pcie);
- iowrite32(0, pcie->pp.dbi_base + PCIE_DBI_RO_WR_EN);
-+
-+ ls_pcie_disable_outbound_atus(pcie);
-+ ls_pcie_fix_error_response(pcie);
-+ dw_pcie_setup_rc(pp);
- }
-
- static int ls_pcie_msi_host_init(struct pcie_port *pp,
-@@ -196,20 +218,40 @@ static struct ls_pcie_drvdata ls1021_drv
- static struct ls_pcie_drvdata ls1043_drvdata = {
- .lut_offset = 0x10000,
- .ltssm_shift = 24,
-+ .lut_dbg = 0x7fc,
-+ .ops = &ls_pcie_host_ops,
-+};
-+
-+static struct ls_pcie_drvdata ls1046_drvdata = {
-+ .lut_offset = 0x80000,
-+ .ltssm_shift = 24,
-+ .lut_dbg = 0x407fc,
- .ops = &ls_pcie_host_ops,
- };
-
- static struct ls_pcie_drvdata ls2080_drvdata = {
- .lut_offset = 0x80000,
- .ltssm_shift = 0,
-+ .lut_dbg = 0x7fc,
-+ .ops = &ls_pcie_host_ops,
-+};
-+
-+static struct ls_pcie_drvdata ls2088_drvdata = {
-+ .lut_offset = 0x80000,
-+ .ltssm_shift = 0,
-+ .lut_dbg = 0x407fc,
- .ops = &ls_pcie_host_ops,
- };
-
- static const struct of_device_id ls_pcie_of_match[] = {
-+ { .compatible = "fsl,ls1012a-pcie", .data = &ls1046_drvdata },
- { .compatible = "fsl,ls1021a-pcie", .data = &ls1021_drvdata },
- { .compatible = "fsl,ls1043a-pcie", .data = &ls1043_drvdata },
-+ { .compatible = "fsl,ls1046a-pcie", .data = &ls1046_drvdata },
- { .compatible = "fsl,ls2080a-pcie", .data = &ls2080_drvdata },
- { .compatible = "fsl,ls2085a-pcie", .data = &ls2080_drvdata },
-+ { .compatible = "fsl,ls2088a-pcie", .data = &ls2088_drvdata },
-+ { .compatible = "fsl,ls1088a-pcie", .data = &ls2088_drvdata },
- { },
- };
-
---- a/drivers/pci/host/pcie-designware.c
-+++ b/drivers/pci/host/pcie-designware.c
-@@ -478,6 +478,12 @@ int dw_pcie_wait_for_link(struct pcie_po
- return -ETIMEDOUT;
- }
-
-+void dw_pcie_disable_outbound_atu(struct pcie_port *pp, int index)
-+{
-+ dw_pcie_writel_rc(pp, PCIE_ATU_VIEWPORT, PCIE_ATU_REGION_OUTBOUND | index);
-+ dw_pcie_writel_rc(pp, PCIE_ATU_CR2, 0);
-+}
-+
- int dw_pcie_link_up(struct pcie_port *pp)
- {
- u32 val;
---- a/drivers/pci/host/pcie-designware.h
-+++ b/drivers/pci/host/pcie-designware.h
-@@ -82,5 +82,6 @@ int dw_pcie_wait_for_link(struct pcie_po
- int dw_pcie_link_up(struct pcie_port *pp);
- void dw_pcie_setup_rc(struct pcie_port *pp);
- int dw_pcie_host_init(struct pcie_port *pp);
-+void dw_pcie_disable_outbound_atu(struct pcie_port *pp, int index);
-
- #endif /* _PCIE_DESIGNWARE_H */
---- a/drivers/pci/pci.c
-+++ b/drivers/pci/pci.c
-@@ -454,7 +454,7 @@ struct resource *pci_find_parent_resourc
- pci_bus_for_each_resource(bus, r, i) {
- if (!r)
- continue;
-- if (res->start && resource_contains(r, res)) {
-+ if (resource_contains(r, res)) {
-
- /*
- * If the window is prefetchable but the BAR is
---- a/drivers/pci/pcie/portdrv_core.c
-+++ b/drivers/pci/pcie/portdrv_core.c
-@@ -44,52 +44,30 @@ static void release_pcie_device(struct d
- }
-
- /**
-- * pcie_port_msix_add_entry - add entry to given array of MSI-X entries
-- * @entries: Array of MSI-X entries
-- * @new_entry: Index of the entry to add to the array
-- * @nr_entries: Number of entries already in the array
-+ * pcibios_check_service_irqs - check irqs in the device tree
-+ * @dev: PCI Express port to handle
-+ * @irqs: Array of irqs to populate
-+ * @mask: Bitmask of port capabilities returned by get_port_device_capability()
-+ *
-+ * Return value: 0 means no service irqs in the device tree
- *
-- * Return value: Position of the added entry in the array
- */
--static int pcie_port_msix_add_entry(
-- struct msix_entry *entries, int new_entry, int nr_entries)
-+int __weak pcibios_check_service_irqs(struct pci_dev *dev, int *irqs, int mask)
- {
-- int j;
--
-- for (j = 0; j < nr_entries; j++)
-- if (entries[j].entry == new_entry)
-- return j;
--
-- entries[j].entry = new_entry;
-- return j;
-+ return 0;
- }
-
- /**
- * pcie_port_enable_msix - try to set up MSI-X as interrupt mode for given port
- * @dev: PCI Express port to handle
-- * @vectors: Array of interrupt vectors to populate
-+ * @irqs: Array of interrupt vectors to populate
- * @mask: Bitmask of port capabilities returned by get_port_device_capability()
- *
- * Return value: 0 on success, error code on failure
- */
--static int pcie_port_enable_msix(struct pci_dev *dev, int *vectors, int mask)
-+static int pcie_port_enable_msix(struct pci_dev *dev, int *irqs, int mask)
- {
-- struct msix_entry *msix_entries;
-- int idx[PCIE_PORT_DEVICE_MAXSERVICES];
-- int nr_entries, status, pos, i, nvec;
-- u16 reg16;
-- u32 reg32;
--
-- nr_entries = pci_msix_vec_count(dev);
-- if (nr_entries < 0)
-- return nr_entries;
-- BUG_ON(!nr_entries);
-- if (nr_entries > PCIE_PORT_MAX_MSIX_ENTRIES)
-- nr_entries = PCIE_PORT_MAX_MSIX_ENTRIES;
--
-- msix_entries = kzalloc(sizeof(*msix_entries) * nr_entries, GFP_KERNEL);
-- if (!msix_entries)
-- return -ENOMEM;
-+ int nr_entries, entry, nvec = 0;
-
- /*
- * Allocate as many entries as the port wants, so that we can check
-@@ -97,20 +75,13 @@ static int pcie_port_enable_msix(struct
- * equal to the number of entries this port actually uses, we'll happily
- * go through without any tricks.
- */
-- for (i = 0; i < nr_entries; i++)
-- msix_entries[i].entry = i;
--
-- status = pci_enable_msix_exact(dev, msix_entries, nr_entries);
-- if (status)
-- goto Exit;
--
-- for (i = 0; i < PCIE_PORT_DEVICE_MAXSERVICES; i++)
-- idx[i] = -1;
-- status = -EIO;
-- nvec = 0;
-+ nr_entries = pci_alloc_irq_vectors(dev, 1, PCIE_PORT_MAX_MSIX_ENTRIES,
-+ PCI_IRQ_MSIX);
-+ if (nr_entries < 0)
-+ return nr_entries;
-
- if (mask & (PCIE_PORT_SERVICE_PME | PCIE_PORT_SERVICE_HP)) {
-- int entry;
-+ u16 reg16;
-
- /*
- * The code below follows the PCI Express Base Specification 2.0
-@@ -125,18 +96,16 @@ static int pcie_port_enable_msix(struct
- pcie_capability_read_word(dev, PCI_EXP_FLAGS, &reg16);
- entry = (reg16 & PCI_EXP_FLAGS_IRQ) >> 9;
- if (entry >= nr_entries)
-- goto Error;
-+ goto out_free_irqs;
-
-- i = pcie_port_msix_add_entry(msix_entries, entry, nvec);
-- if (i == nvec)
-- nvec++;
-+ irqs[PCIE_PORT_SERVICE_PME_SHIFT] = pci_irq_vector(dev, entry);
-+ irqs[PCIE_PORT_SERVICE_HP_SHIFT] = pci_irq_vector(dev, entry);
-
-- idx[PCIE_PORT_SERVICE_PME_SHIFT] = i;
-- idx[PCIE_PORT_SERVICE_HP_SHIFT] = i;
-+ nvec = max(nvec, entry + 1);
- }
-
- if (mask & PCIE_PORT_SERVICE_AER) {
-- int entry;
-+ u32 reg32, pos;
-
- /*
- * The code below follows Section 7.10.10 of the PCI Express
-@@ -151,13 +120,11 @@ static int pcie_port_enable_msix(struct
- pci_read_config_dword(dev, pos + PCI_ERR_ROOT_STATUS, &reg32);
- entry = reg32 >> 27;
- if (entry >= nr_entries)
-- goto Error;
-+ goto out_free_irqs;
-
-- i = pcie_port_msix_add_entry(msix_entries, entry, nvec);
-- if (i == nvec)
-- nvec++;
-+ irqs[PCIE_PORT_SERVICE_AER_SHIFT] = pci_irq_vector(dev, entry);
-
-- idx[PCIE_PORT_SERVICE_AER_SHIFT] = i;
-+ nvec = max(nvec, entry + 1);
- }
-
- /*
-@@ -165,41 +132,54 @@ static int pcie_port_enable_msix(struct
- * what we have. Otherwise, the port has some extra entries not for the
- * services we know and we need to work around that.
- */
-- if (nvec == nr_entries) {
-- status = 0;
-- } else {
-+ if (nvec != nr_entries) {
- /* Drop the temporary MSI-X setup */
-- pci_disable_msix(dev);
-+ pci_free_irq_vectors(dev);
-
- /* Now allocate the MSI-X vectors for real */
-- status = pci_enable_msix_exact(dev, msix_entries, nvec);
-- if (status)
-- goto Exit;
-+ nr_entries = pci_alloc_irq_vectors(dev, nvec, nvec,
-+ PCI_IRQ_MSIX);
-+ if (nr_entries < 0)
-+ return nr_entries;
- }
-
-- for (i = 0; i < PCIE_PORT_DEVICE_MAXSERVICES; i++)
-- vectors[i] = idx[i] >= 0 ? msix_entries[idx[i]].vector : -1;
--
-- Exit:
-- kfree(msix_entries);
-- return status;
-+ return 0;
-
-- Error:
-- pci_disable_msix(dev);
-- goto Exit;
-+out_free_irqs:
-+ pci_free_irq_vectors(dev);
-+ return -EIO;
- }
-
- /**
-- * init_service_irqs - initialize irqs for PCI Express port services
-+ * pcie_init_service_irqs - initialize irqs for PCI Express port services
- * @dev: PCI Express port to handle
- * @irqs: Array of irqs to populate
- * @mask: Bitmask of port capabilities returned by get_port_device_capability()
- *
- * Return value: Interrupt mode associated with the port
- */
--static int init_service_irqs(struct pci_dev *dev, int *irqs, int mask)
-+static int pcie_init_service_irqs(struct pci_dev *dev, int *irqs, int mask)
- {
-- int i, irq = -1;
-+ unsigned flags = PCI_IRQ_LEGACY | PCI_IRQ_MSI;
-+ int ret, i;
-+ int irq = -1;
-+
-+ for (i = 0; i < PCIE_PORT_DEVICE_MAXSERVICES; i++)
-+ irqs[i] = -1;
-+
-+ /* Check if some platforms owns independent irq pins for AER/PME etc.
-+ * Some platforms may own independent AER/PME interrupts and set
-+ * them in the device tree file.
-+ */
-+ ret = pcibios_check_service_irqs(dev, irqs, mask);
-+ if (ret) {
-+ if (dev->irq)
-+ irq = dev->irq;
-+ for (i = 0; i < PCIE_PORT_DEVICE_MAXSERVICES; i++)
-+ if (irqs[i] == -1 && i != PCIE_PORT_SERVICE_VC_SHIFT)
-+ irqs[i] = irq;
-+ return 0;
-+ }
-
- /*
- * If MSI cannot be used for PCIe PME or hotplug, we have to use
-@@ -207,41 +187,25 @@ static int init_service_irqs(struct pci_
- */
- if (((mask & PCIE_PORT_SERVICE_PME) && pcie_pme_no_msi()) ||
- ((mask & PCIE_PORT_SERVICE_HP) && pciehp_no_msi())) {
-- if (dev->irq)
-- irq = dev->irq;
-- goto no_msi;
-+ flags &= ~PCI_IRQ_MSI;
-+ } else {
-+ /* Try to use MSI-X if supported */
-+ if (!pcie_port_enable_msix(dev, irqs, mask))
-+ return 0;
- }
-
-- /* Try to use MSI-X if supported */
-- if (!pcie_port_enable_msix(dev, irqs, mask))
-- return 0;
--
-- /*
-- * We're not going to use MSI-X, so try MSI and fall back to INTx.
-- * If neither MSI/MSI-X nor INTx available, try other interrupt. On
-- * some platforms, root port doesn't support MSI/MSI-X/INTx in RC mode.
-- */
-- if (!pci_enable_msi(dev) || dev->irq)
-- irq = dev->irq;
-+ ret = pci_alloc_irq_vectors(dev, 1, 1, flags);
-+ if (ret < 0)
-+ return -ENODEV;
-
-- no_msi:
-- for (i = 0; i < PCIE_PORT_DEVICE_MAXSERVICES; i++)
-- irqs[i] = irq;
-- irqs[PCIE_PORT_SERVICE_VC_SHIFT] = -1;
-+ for (i = 0; i < PCIE_PORT_DEVICE_MAXSERVICES; i++) {
-+ if (i != PCIE_PORT_SERVICE_VC_SHIFT)
-+ irqs[i] = pci_irq_vector(dev, 0);
-+ }
-
-- if (irq < 0)
-- return -ENODEV;
- return 0;
- }
-
--static void cleanup_service_irqs(struct pci_dev *dev)
--{
-- if (dev->msix_enabled)
-- pci_disable_msix(dev);
-- else if (dev->msi_enabled)
-- pci_disable_msi(dev);
--}
--
- /**
- * get_port_device_capability - discover capabilities of a PCI Express port
- * @dev: PCI Express port to examine
-@@ -378,7 +342,7 @@ int pcie_port_device_register(struct pci
- * that can be used in the absence of irqs. Allow them to determine
- * if that is to be used.
- */
-- status = init_service_irqs(dev, irqs, capabilities);
-+ status = pcie_init_service_irqs(dev, irqs, capabilities);
- if (status) {
- capabilities &= PCIE_PORT_SERVICE_VC | PCIE_PORT_SERVICE_HP;
- if (!capabilities)
-@@ -401,7 +365,7 @@ int pcie_port_device_register(struct pci
- return 0;
-
- error_cleanup_irqs:
-- cleanup_service_irqs(dev);
-+ pci_free_irq_vectors(dev);
- error_disable:
- pci_disable_device(dev);
- return status;
-@@ -469,7 +433,7 @@ static int remove_iter(struct device *de
- void pcie_port_device_remove(struct pci_dev *dev)
- {
- device_for_each_child(&dev->dev, NULL, remove_iter);
-- cleanup_service_irqs(dev);
-+ pci_free_irq_vectors(dev);
- pci_disable_device(dev);
- }
-
-@@ -499,7 +463,6 @@ static int pcie_port_probe_service(struc
- if (status)
- return status;
-
-- dev_printk(KERN_DEBUG, dev, "service driver %s loaded\n", driver->name);
- get_device(dev);
- return 0;
- }
-@@ -524,8 +487,6 @@ static int pcie_port_remove_service(stru
- pciedev = to_pcie_device(dev);
- driver = to_service_driver(dev->driver);
- if (driver && driver->remove) {
-- dev_printk(KERN_DEBUG, dev, "unloading service driver %s\n",
-- driver->name);
- driver->remove(pciedev);
- put_device(dev);
- }
---- a/drivers/pci/quirks.c
-+++ b/drivers/pci/quirks.c
-@@ -3333,6 +3333,13 @@ DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_A
- DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_ATHEROS, 0x003c, quirk_no_bus_reset);
- DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_ATHEROS, 0x0033, quirk_no_bus_reset);
-
-+/*
-+ * NXP (Freescale Vendor ID) LS1088 chips do not behave correctly after
-+ * bus reset. Link state of device does not comes UP and so config space
-+ * never accessible again.
-+ */
-+DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_FREESCALE, 0x80c0, quirk_no_bus_reset);
-+
- static void quirk_no_pm_reset(struct pci_dev *dev)
- {
- /*
-@@ -4677,3 +4684,11 @@ DECLARE_PCI_FIXUP_EARLY(PCI_VENDOR_ID_IN
- DECLARE_PCI_FIXUP_EARLY(PCI_VENDOR_ID_INTEL, 0x2031, quirk_no_aersid);
- DECLARE_PCI_FIXUP_EARLY(PCI_VENDOR_ID_INTEL, 0x2032, quirk_no_aersid);
- DECLARE_PCI_FIXUP_EARLY(PCI_VENDOR_ID_INTEL, 0x2033, quirk_no_aersid);
-+
-+/* Freescale PCIe doesn't support MSI in RC mode */
-+static void quirk_fsl_no_msi(struct pci_dev *pdev)
-+{
-+ if (pci_pcie_type(pdev) == PCI_EXP_TYPE_ROOT_PORT)
-+ pdev->no_msi = 1;
-+}
-+DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_FREESCALE, PCI_ANY_ID, quirk_fsl_no_msi);
---- a/include/linux/pci.h
-+++ b/include/linux/pci.h
-@@ -1825,6 +1825,7 @@ void pcibios_release_device(struct pci_d
- void pcibios_penalize_isa_irq(int irq, int active);
- int pcibios_alloc_irq(struct pci_dev *dev);
- void pcibios_free_irq(struct pci_dev *dev);
-+int pcibios_check_service_irqs(struct pci_dev *dev, int *irqs, int mask);
-
- #ifdef CONFIG_HIBERNATE_CALLBACKS
- extern struct dev_pm_ops pcibios_pm_ops;
diff --git a/target/linux/layerscape/patches-4.9/703-phy-support-layerscape.patch b/target/linux/layerscape/patches-4.9/703-phy-support-layerscape.patch
deleted file mode 100644
index 9c23279188..0000000000
--- a/target/linux/layerscape/patches-4.9/703-phy-support-layerscape.patch
+++ /dev/null
@@ -1,1763 +0,0 @@
-From 8949ebc0c5b982eab7ca493dad7b86c30befa6ec Mon Sep 17 00:00:00 2001
-From: Yangbo Lu <yangbo.lu@nxp.com>
-Date: Wed, 17 Jan 2018 15:01:30 +0800
-Subject: [PATCH 09/30] phy: support layerscape
-
-This is an integrated patch for layerscape mdio-phy support.
-
-Signed-off-by: Bogdan Purcareata <bogdan.purcareata@nxp.com>
-Signed-off-by: Zhang Ying-22455 <ying.zhang22455@nxp.com>
-Signed-off-by: costi <constantin.tudor@freescale.com>
-Signed-off-by: Madalin Bucur <madalin.bucur@freescale.com>
-Signed-off-by: Shaohui Xie <Shaohui.Xie@nxp.com>
-Signed-off-by: Florian Fainelli <f.fainelli@gmail.com>
-Signed-off-by: Yangbo Lu <yangbo.lu@nxp.com>
----
- drivers/net/phy/Kconfig | 11 +
- drivers/net/phy/Makefile | 2 +
- drivers/net/phy/aquantia.c | 28 +
- drivers/net/phy/cortina.c | 118 ++++
- drivers/net/phy/fsl_backplane.c | 1358 +++++++++++++++++++++++++++++++++++++++
- drivers/net/phy/marvell.c | 2 +-
- drivers/net/phy/phy.c | 23 +-
- drivers/net/phy/phy_device.c | 6 +-
- drivers/net/phy/swphy.c | 1 +
- include/linux/phy.h | 6 +
- 10 files changed, 1547 insertions(+), 8 deletions(-)
- create mode 100644 drivers/net/phy/cortina.c
- create mode 100644 drivers/net/phy/fsl_backplane.c
-
---- a/drivers/net/phy/Kconfig
-+++ b/drivers/net/phy/Kconfig
-@@ -89,6 +89,12 @@ config MDIO_BUS_MUX_MMIOREG
- config MDIO_CAVIUM
- tristate
-
-+config MDIO_FSL_BACKPLANE
-+ tristate "Support for backplane on Freescale XFI interface"
-+ depends on OF_MDIO
-+ help
-+ This module provides a driver for Freescale XFI's backplane.
-+
- config MDIO_GPIO
- tristate "GPIO lib-based bitbanged MDIO buses"
- depends on MDIO_BITBANG && GPIOLIB
-@@ -298,6 +304,11 @@ config CICADA_PHY
- ---help---
- Currently supports the cis8204
-
-+config CORTINA_PHY
-+ tristate "Cortina EDC CDR 10G Ethernet PHY"
-+ ---help---
-+ Currently supports the CS4340 phy.
-+
- config DAVICOM_PHY
- tristate "Davicom PHYs"
- ---help---
---- a/drivers/net/phy/Makefile
-+++ b/drivers/net/phy/Makefile
-@@ -30,6 +30,7 @@ obj-$(CONFIG_MDIO_BUS_MUX_BCM_IPROC) +=
- obj-$(CONFIG_MDIO_BUS_MUX_GPIO) += mdio-mux-gpio.o
- obj-$(CONFIG_MDIO_BUS_MUX_MMIOREG) += mdio-mux-mmioreg.o
- obj-$(CONFIG_MDIO_CAVIUM) += mdio-cavium.o
-+obj-$(CONFIG_MDIO_FSL_BACKPLANE) += fsl_backplane.o
- obj-$(CONFIG_MDIO_GPIO) += mdio-gpio.o
- obj-$(CONFIG_MDIO_HISI_FEMAC) += mdio-hisi-femac.o
- obj-$(CONFIG_MDIO_MOXART) += mdio-moxart.o
-@@ -48,6 +49,7 @@ obj-$(CONFIG_BCM_CYGNUS_PHY) += bcm-cygn
- obj-$(CONFIG_BCM_NET_PHYLIB) += bcm-phy-lib.o
- obj-$(CONFIG_BROADCOM_PHY) += broadcom.o
- obj-$(CONFIG_CICADA_PHY) += cicada.o
-+obj-$(CONFIG_CORTINA_PHY) += cortina.o
- obj-$(CONFIG_DAVICOM_PHY) += davicom.o
- obj-$(CONFIG_DP83640_PHY) += dp83640.o
- obj-$(CONFIG_DP83848_PHY) += dp83848.o
---- a/drivers/net/phy/aquantia.c
-+++ b/drivers/net/phy/aquantia.c
-@@ -21,6 +21,8 @@
- #define PHY_ID_AQ1202 0x03a1b445
- #define PHY_ID_AQ2104 0x03a1b460
- #define PHY_ID_AQR105 0x03a1b4a2
-+#define PHY_ID_AQR106 0x03a1b4d0
-+#define PHY_ID_AQR107 0x03a1b4e0
- #define PHY_ID_AQR405 0x03a1b4b0
-
- #define PHY_AQUANTIA_FEATURES (SUPPORTED_10000baseT_Full | \
-@@ -154,6 +156,30 @@ static struct phy_driver aquantia_driver
- .read_status = aquantia_read_status,
- },
- {
-+ .phy_id = PHY_ID_AQR106,
-+ .phy_id_mask = 0xfffffff0,
-+ .name = "Aquantia AQR106",
-+ .features = PHY_AQUANTIA_FEATURES,
-+ .flags = PHY_HAS_INTERRUPT,
-+ .aneg_done = aquantia_aneg_done,
-+ .config_aneg = aquantia_config_aneg,
-+ .config_intr = aquantia_config_intr,
-+ .ack_interrupt = aquantia_ack_interrupt,
-+ .read_status = aquantia_read_status,
-+},
-+{
-+ .phy_id = PHY_ID_AQR107,
-+ .phy_id_mask = 0xfffffff0,
-+ .name = "Aquantia AQR107",
-+ .features = PHY_AQUANTIA_FEATURES,
-+ .flags = PHY_HAS_INTERRUPT,
-+ .aneg_done = aquantia_aneg_done,
-+ .config_aneg = aquantia_config_aneg,
-+ .config_intr = aquantia_config_intr,
-+ .ack_interrupt = aquantia_ack_interrupt,
-+ .read_status = aquantia_read_status,
-+},
-+{
- .phy_id = PHY_ID_AQR405,
- .phy_id_mask = 0xfffffff0,
- .name = "Aquantia AQR405",
-@@ -173,6 +199,8 @@ static struct mdio_device_id __maybe_unu
- { PHY_ID_AQ1202, 0xfffffff0 },
- { PHY_ID_AQ2104, 0xfffffff0 },
- { PHY_ID_AQR105, 0xfffffff0 },
-+ { PHY_ID_AQR106, 0xfffffff0 },
-+ { PHY_ID_AQR107, 0xfffffff0 },
- { PHY_ID_AQR405, 0xfffffff0 },
- { }
- };
---- /dev/null
-+++ b/drivers/net/phy/cortina.c
-@@ -0,0 +1,118 @@
-+/*
-+ * Copyright 2017 NXP
-+ *
-+ * This program is free software; you can redistribute it and/or modify
-+ * it under the terms of the GNU General Public License as published by
-+ * the Free Software Foundation; either version 2 of the License, or
-+ * (at your option) any later version.
-+ *
-+ * This program is distributed in the hope that it will be useful,
-+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
-+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
-+ * GNU General Public License for more details.
-+ *
-+ * CORTINA is a registered trademark of Cortina Systems, Inc.
-+ *
-+ */
-+#include <linux/module.h>
-+#include <linux/phy.h>
-+
-+#define PHY_ID_CS4340 0x13e51002
-+
-+#define VILLA_GLOBAL_CHIP_ID_LSB 0x0
-+#define VILLA_GLOBAL_CHIP_ID_MSB 0x1
-+
-+#define VILLA_GLOBAL_GPIO_1_INTS 0x017
-+
-+static int cortina_read_reg(struct phy_device *phydev, u16 regnum)
-+{
-+ return mdiobus_read(phydev->mdio.bus, phydev->mdio.addr,
-+ MII_ADDR_C45 | regnum);
-+}
-+
-+static int cortina_config_aneg(struct phy_device *phydev)
-+{
-+ phydev->supported = SUPPORTED_10000baseT_Full;
-+ phydev->advertising = SUPPORTED_10000baseT_Full;
-+
-+ return 0;
-+}
-+
-+static int cortina_read_status(struct phy_device *phydev)
-+{
-+ int gpio_int_status, ret = 0;
-+
-+ gpio_int_status = cortina_read_reg(phydev, VILLA_GLOBAL_GPIO_1_INTS);
-+ if (gpio_int_status < 0) {
-+ ret = gpio_int_status;
-+ goto err;
-+ }
-+
-+ if (gpio_int_status & 0x8) {
-+ /* up when edc_convergedS set */
-+ phydev->speed = SPEED_10000;
-+ phydev->duplex = DUPLEX_FULL;
-+ phydev->link = 1;
-+ } else {
-+ phydev->link = 0;
-+ }
-+
-+err:
-+ return ret;
-+}
-+
-+static int cortina_soft_reset(struct phy_device *phydev)
-+{
-+ return 0;
-+}
-+
-+static int cortina_probe(struct phy_device *phydev)
-+{
-+ u32 phy_id = 0;
-+ int id_lsb = 0, id_msb = 0;
-+
-+ /* Read device id from phy registers. */
-+ id_lsb = cortina_read_reg(phydev, VILLA_GLOBAL_CHIP_ID_LSB);
-+ if (id_lsb < 0)
-+ return -ENXIO;
-+
-+ phy_id = id_lsb << 16;
-+
-+ id_msb = cortina_read_reg(phydev, VILLA_GLOBAL_CHIP_ID_MSB);
-+ if (id_msb < 0)
-+ return -ENXIO;
-+
-+ phy_id |= id_msb;
-+
-+ /* Make sure the device tree binding matched the driver with the
-+ * right device.
-+ */
-+ if (phy_id != phydev->drv->phy_id) {
-+ phydev_err(phydev, "Error matching phy with %s driver\n",
-+ phydev->drv->name);
-+ return -ENODEV;
-+ }
-+
-+ return 0;
-+}
-+
-+static struct phy_driver cortina_driver[] = {
-+{
-+ .phy_id = PHY_ID_CS4340,
-+ .phy_id_mask = 0xffffffff,
-+ .name = "Cortina CS4340",
-+ .config_aneg = cortina_config_aneg,
-+ .read_status = cortina_read_status,
-+ .soft_reset = cortina_soft_reset,
-+ .probe = cortina_probe,
-+},
-+};
-+
-+module_phy_driver(cortina_driver);
-+
-+static struct mdio_device_id __maybe_unused cortina_tbl[] = {
-+ { PHY_ID_CS4340, 0xffffffff},
-+ {},
-+};
-+
-+MODULE_DEVICE_TABLE(mdio, cortina_tbl);
---- /dev/null
-+++ b/drivers/net/phy/fsl_backplane.c
-@@ -0,0 +1,1358 @@
-+/* Freescale backplane driver.
-+ * Author: Shaohui Xie <Shaohui.Xie@freescale.com>
-+ *
-+ * Copyright 2015 Freescale Semiconductor, Inc.
-+ *
-+ * Licensed under the GPL-2 or later.
-+ */
-+
-+#include <linux/kernel.h>
-+#include <linux/module.h>
-+#include <linux/mii.h>
-+#include <linux/mdio.h>
-+#include <linux/ethtool.h>
-+#include <linux/phy.h>
-+#include <linux/io.h>
-+#include <linux/of.h>
-+#include <linux/of_net.h>
-+#include <linux/of_address.h>
-+#include <linux/of_platform.h>
-+#include <linux/timer.h>
-+#include <linux/delay.h>
-+#include <linux/workqueue.h>
-+
-+/* XFI PCS Device Identifier */
-+#define FSL_PCS_PHY_ID 0x0083e400
-+
-+/* Freescale KR PMD registers */
-+#define FSL_KR_PMD_CTRL 0x96
-+#define FSL_KR_PMD_STATUS 0x97
-+#define FSL_KR_LP_CU 0x98
-+#define FSL_KR_LP_STATUS 0x99
-+#define FSL_KR_LD_CU 0x9a
-+#define FSL_KR_LD_STATUS 0x9b
-+
-+/* Freescale KR PMD defines */
-+#define PMD_RESET 0x1
-+#define PMD_STATUS_SUP_STAT 0x4
-+#define PMD_STATUS_FRAME_LOCK 0x2
-+#define TRAIN_EN 0x3
-+#define TRAIN_DISABLE 0x1
-+#define RX_STAT 0x1
-+
-+#define FSL_KR_RX_LINK_STAT_MASK 0x1000
-+#define FSL_XFI_PCS_10GR_SR1 0x20
-+
-+/* Freescale KX PCS mode register */
-+#define FSL_PCS_IF_MODE 0x8014
-+
-+/* Freescale KX PCS mode register init value */
-+#define IF_MODE_INIT 0x8
-+
-+/* Freescale KX/KR AN registers */
-+#define FSL_AN_AD1 0x11
-+#define FSL_AN_BP_STAT 0x30
-+
-+/* Freescale KX/KR AN registers defines */
-+#define AN_CTRL_INIT 0x1200
-+#define KX_AN_AD1_INIT 0x25
-+#define KR_AN_AD1_INIT 0x85
-+#define AN_LNK_UP_MASK 0x4
-+#define KR_AN_MASK 0x8
-+#define TRAIN_FAIL 0x8
-+
-+/* C(-1) */
-+#define BIN_M1 0
-+/* C(1) */
-+#define BIN_LONG 1
-+#define BIN_M1_SEL 6
-+#define BIN_Long_SEL 7
-+#define CDR_SEL_MASK 0x00070000
-+#define BIN_SNAPSHOT_NUM 5
-+#define BIN_M1_THRESHOLD 3
-+#define BIN_LONG_THRESHOLD 2
-+
-+#define PRE_COE_SHIFT 22
-+#define POST_COE_SHIFT 16
-+#define ZERO_COE_SHIFT 8
-+
-+#define PRE_COE_MAX 0x0
-+#define PRE_COE_MIN 0x8
-+#define POST_COE_MAX 0x0
-+#define POST_COE_MIN 0x10
-+#define ZERO_COE_MAX 0x30
-+#define ZERO_COE_MIN 0x0
-+
-+#define TECR0_INIT 0x24200000
-+#define RATIO_PREQ 0x3
-+#define RATIO_PST1Q 0xd
-+#define RATIO_EQ 0x20
-+
-+#define GCR0_RESET_MASK 0x600000
-+#define GCR1_SNP_START_MASK 0x00000040
-+#define GCR1_CTL_SNP_START_MASK 0x00002000
-+#define GCR1_REIDL_TH_MASK 0x00700000
-+#define GCR1_REIDL_EX_SEL_MASK 0x000c0000
-+#define GCR1_REIDL_ET_MAS_MASK 0x00004000
-+#define TECR0_AMP_RED_MASK 0x0000003f
-+
-+#define RECR1_CTL_SNP_DONE_MASK 0x00000002
-+#define RECR1_SNP_DONE_MASK 0x00000004
-+#define TCSR1_SNP_DATA_MASK 0x0000ffc0
-+#define TCSR1_SNP_DATA_SHIFT 6
-+#define TCSR1_EQ_SNPBIN_SIGN_MASK 0x100
-+
-+#define RECR1_GAINK2_MASK 0x0f000000
-+#define RECR1_GAINK2_SHIFT 24
-+#define RECR1_GAINK3_MASK 0x000f0000
-+#define RECR1_GAINK3_SHIFT 16
-+#define RECR1_OFFSET_MASK 0x00003f80
-+#define RECR1_OFFSET_SHIFT 7
-+#define RECR1_BLW_MASK 0x00000f80
-+#define RECR1_BLW_SHIFT 7
-+#define EYE_CTRL_SHIFT 12
-+#define BASE_WAND_SHIFT 10
-+
-+#define XGKR_TIMEOUT 1050
-+
-+#define INCREMENT 1
-+#define DECREMENT 2
-+#define TIMEOUT_LONG 3
-+#define TIMEOUT_M1 3
-+
-+#define RX_READY_MASK 0x8000
-+#define PRESET_MASK 0x2000
-+#define INIT_MASK 0x1000
-+#define COP1_MASK 0x30
-+#define COP1_SHIFT 4
-+#define COZ_MASK 0xc
-+#define COZ_SHIFT 2
-+#define COM1_MASK 0x3
-+#define COM1_SHIFT 0
-+#define REQUEST_MASK 0x3f
-+#define LD_ALL_MASK (PRESET_MASK | INIT_MASK | \
-+ COP1_MASK | COZ_MASK | COM1_MASK)
-+
-+#define NEW_ALGORITHM_TRAIN_TX
-+#ifdef NEW_ALGORITHM_TRAIN_TX
-+#define FORCE_INC_COP1_NUMBER 0
-+#define FORCE_INC_COM1_NUMBER 1
-+#endif
-+
-+#define VAL_INVALID 0xff
-+
-+static const u32 preq_table[] = {0x0, 0x1, 0x3, 0x5,
-+ 0x7, 0x9, 0xb, 0xc, VAL_INVALID};
-+static const u32 pst1q_table[] = {0x0, 0x1, 0x3, 0x5, 0x7,
-+ 0x9, 0xb, 0xd, 0xf, 0x10, VAL_INVALID};
-+
-+enum backplane_mode {
-+ PHY_BACKPLANE_1000BASE_KX,
-+ PHY_BACKPLANE_10GBASE_KR,
-+ PHY_BACKPLANE_INVAL
-+};
-+
-+enum coe_filed {
-+ COE_COP1,
-+ COE_COZ,
-+ COE_COM
-+};
-+
-+enum coe_update {
-+ COE_NOTUPDATED,
-+ COE_UPDATED,
-+ COE_MIN,
-+ COE_MAX,
-+ COE_INV
-+};
-+
-+enum train_state {
-+ DETECTING_LP,
-+ TRAINED,
-+};
-+
-+struct per_lane_ctrl_status {
-+ __be32 gcr0; /* 0x.000 - General Control Register 0 */
-+ __be32 gcr1; /* 0x.004 - General Control Register 1 */
-+ __be32 gcr2; /* 0x.008 - General Control Register 2 */
-+ __be32 resv1; /* 0x.00C - Reserved */
-+ __be32 recr0; /* 0x.010 - Receive Equalization Control Register 0 */
-+ __be32 recr1; /* 0x.014 - Receive Equalization Control Register 1 */
-+ __be32 tecr0; /* 0x.018 - Transmit Equalization Control Register 0 */
-+ __be32 resv2; /* 0x.01C - Reserved */
-+ __be32 tlcr0; /* 0x.020 - TTL Control Register 0 */
-+ __be32 tlcr1; /* 0x.024 - TTL Control Register 1 */
-+ __be32 tlcr2; /* 0x.028 - TTL Control Register 2 */
-+ __be32 tlcr3; /* 0x.02C - TTL Control Register 3 */
-+ __be32 tcsr0; /* 0x.030 - Test Control/Status Register 0 */
-+ __be32 tcsr1; /* 0x.034 - Test Control/Status Register 1 */
-+ __be32 tcsr2; /* 0x.038 - Test Control/Status Register 2 */
-+ __be32 tcsr3; /* 0x.03C - Test Control/Status Register 3 */
-+};
-+
-+struct tx_condition {
-+ bool bin_m1_late_early;
-+ bool bin_long_late_early;
-+ bool bin_m1_stop;
-+ bool bin_long_stop;
-+ bool tx_complete;
-+ bool sent_init;
-+ int m1_min_max_cnt;
-+ int long_min_max_cnt;
-+#ifdef NEW_ALGORITHM_TRAIN_TX
-+ int pre_inc;
-+ int post_inc;
-+#endif
-+};
-+
-+struct fsl_xgkr_inst {
-+ void *reg_base;
-+ struct phy_device *phydev;
-+ struct tx_condition tx_c;
-+ struct delayed_work xgkr_wk;
-+ enum train_state state;
-+ u32 ld_update;
-+ u32 ld_status;
-+ u32 ratio_preq;
-+ u32 ratio_pst1q;
-+ u32 adpt_eq;
-+};
-+
-+static void tx_condition_init(struct tx_condition *tx_c)
-+{
-+ tx_c->bin_m1_late_early = true;
-+ tx_c->bin_long_late_early = false;
-+ tx_c->bin_m1_stop = false;
-+ tx_c->bin_long_stop = false;
-+ tx_c->tx_complete = false;
-+ tx_c->sent_init = false;
-+ tx_c->m1_min_max_cnt = 0;
-+ tx_c->long_min_max_cnt = 0;
-+#ifdef NEW_ALGORITHM_TRAIN_TX
-+ tx_c->pre_inc = FORCE_INC_COM1_NUMBER;
-+ tx_c->post_inc = FORCE_INC_COP1_NUMBER;
-+#endif
-+}
-+
-+void tune_tecr0(struct fsl_xgkr_inst *inst)
-+{
-+ struct per_lane_ctrl_status *reg_base = inst->reg_base;
-+ u32 val;
-+
-+ val = TECR0_INIT |
-+ inst->adpt_eq << ZERO_COE_SHIFT |
-+ inst->ratio_preq << PRE_COE_SHIFT |
-+ inst->ratio_pst1q << POST_COE_SHIFT;
-+
-+ /* reset the lane */
-+ iowrite32(ioread32(&reg_base->gcr0) & ~GCR0_RESET_MASK,
-+ &reg_base->gcr0);
-+ udelay(1);
-+ iowrite32(val, &reg_base->tecr0);
-+ udelay(1);
-+ /* unreset the lane */
-+ iowrite32(ioread32(&reg_base->gcr0) | GCR0_RESET_MASK,
-+ &reg_base->gcr0);
-+ udelay(1);
-+}
-+
-+static void start_lt(struct phy_device *phydev)
-+{
-+ phy_write_mmd(phydev, MDIO_MMD_PMAPMD, FSL_KR_PMD_CTRL, TRAIN_EN);
-+}
-+
-+static void stop_lt(struct phy_device *phydev)
-+{
-+ phy_write_mmd(phydev, MDIO_MMD_PMAPMD, FSL_KR_PMD_CTRL, TRAIN_DISABLE);
-+}
-+
-+static void reset_gcr0(struct fsl_xgkr_inst *inst)
-+{
-+ struct per_lane_ctrl_status *reg_base = inst->reg_base;
-+
-+ iowrite32(ioread32(&reg_base->gcr0) & ~GCR0_RESET_MASK,
-+ &reg_base->gcr0);
-+ udelay(1);
-+ iowrite32(ioread32(&reg_base->gcr0) | GCR0_RESET_MASK,
-+ &reg_base->gcr0);
-+ udelay(1);
-+}
-+
-+void lane_set_1gkx(void *reg)
-+{
-+ struct per_lane_ctrl_status *reg_base = reg;
-+ u32 val;
-+
-+ /* reset the lane */
-+ iowrite32(ioread32(&reg_base->gcr0) & ~GCR0_RESET_MASK,
-+ &reg_base->gcr0);
-+ udelay(1);
-+
-+ /* set gcr1 for 1GKX */
-+ val = ioread32(&reg_base->gcr1);
-+ val &= ~(GCR1_REIDL_TH_MASK | GCR1_REIDL_EX_SEL_MASK |
-+ GCR1_REIDL_ET_MAS_MASK);
-+ iowrite32(val, &reg_base->gcr1);
-+ udelay(1);
-+
-+ /* set tecr0 for 1GKX */
-+ val = ioread32(&reg_base->tecr0);
-+ val &= ~TECR0_AMP_RED_MASK;
-+ iowrite32(val, &reg_base->tecr0);
-+ udelay(1);
-+
-+ /* unreset the lane */
-+ iowrite32(ioread32(&reg_base->gcr0) | GCR0_RESET_MASK,
-+ &reg_base->gcr0);
-+ udelay(1);
-+}
-+
-+static void reset_lt(struct phy_device *phydev)
-+{
-+ phy_write_mmd(phydev, MDIO_MMD_PMAPMD, MDIO_CTRL1, PMD_RESET);
-+ phy_write_mmd(phydev, MDIO_MMD_PMAPMD, FSL_KR_PMD_CTRL, TRAIN_DISABLE);
-+ phy_write_mmd(phydev, MDIO_MMD_PMAPMD, FSL_KR_LD_CU, 0);
-+ phy_write_mmd(phydev, MDIO_MMD_PMAPMD, FSL_KR_LD_STATUS, 0);
-+ phy_write_mmd(phydev, MDIO_MMD_PMAPMD, FSL_KR_PMD_STATUS, 0);
-+ phy_write_mmd(phydev, MDIO_MMD_PMAPMD, FSL_KR_LP_CU, 0);
-+ phy_write_mmd(phydev, MDIO_MMD_PMAPMD, FSL_KR_LP_STATUS, 0);
-+}
-+
-+static void start_xgkr_state_machine(struct delayed_work *work)
-+{
-+ queue_delayed_work(system_power_efficient_wq, work,
-+ msecs_to_jiffies(XGKR_TIMEOUT));
-+}
-+
-+static void start_xgkr_an(struct phy_device *phydev)
-+{
-+ struct fsl_xgkr_inst *inst;
-+
-+ reset_lt(phydev);
-+ phy_write_mmd(phydev, MDIO_MMD_AN, FSL_AN_AD1, KR_AN_AD1_INIT);
-+ phy_write_mmd(phydev, MDIO_MMD_AN, MDIO_CTRL1, AN_CTRL_INIT);
-+
-+ inst = phydev->priv;
-+
-+ /* start state machine*/
-+ start_xgkr_state_machine(&inst->xgkr_wk);
-+}
-+
-+static void start_1gkx_an(struct phy_device *phydev)
-+{
-+ phy_write_mmd(phydev, MDIO_MMD_PCS, FSL_PCS_IF_MODE, IF_MODE_INIT);
-+ phy_write_mmd(phydev, MDIO_MMD_AN, FSL_AN_AD1, KX_AN_AD1_INIT);
-+ phy_read_mmd(phydev, MDIO_MMD_AN, MDIO_STAT1);
-+ phy_write_mmd(phydev, MDIO_MMD_AN, MDIO_CTRL1, AN_CTRL_INIT);
-+}
-+
-+static void ld_coe_status(struct fsl_xgkr_inst *inst)
-+{
-+ phy_write_mmd(inst->phydev, MDIO_MMD_PMAPMD,
-+ FSL_KR_LD_STATUS, inst->ld_status);
-+}
-+
-+static void ld_coe_update(struct fsl_xgkr_inst *inst)
-+{
-+ dev_dbg(&inst->phydev->mdio.dev, "sending request: %x\n", inst->ld_update);
-+ phy_write_mmd(inst->phydev, MDIO_MMD_PMAPMD,
-+ FSL_KR_LD_CU, inst->ld_update);
-+}
-+
-+static void init_inst(struct fsl_xgkr_inst *inst, int reset)
-+{
-+ if (reset) {
-+ inst->ratio_preq = RATIO_PREQ;
-+ inst->ratio_pst1q = RATIO_PST1Q;
-+ inst->adpt_eq = RATIO_EQ;
-+ tune_tecr0(inst);
-+ }
-+
-+ tx_condition_init(&inst->tx_c);
-+ inst->state = DETECTING_LP;
-+ inst->ld_status &= RX_READY_MASK;
-+ ld_coe_status(inst);
-+ inst->ld_update = 0;
-+ inst->ld_status &= ~RX_READY_MASK;
-+ ld_coe_status(inst);
-+}
-+
-+#ifdef NEW_ALGORITHM_TRAIN_TX
-+static int get_median_gaink2(u32 *reg)
-+{
-+ int gaink2_snap_shot[BIN_SNAPSHOT_NUM];
-+ u32 rx_eq_snp;
-+ struct per_lane_ctrl_status *reg_base;
-+ int timeout;
-+ int i, j, tmp, pos;
-+
-+ reg_base = (struct per_lane_ctrl_status *)reg;
-+
-+ for (i = 0; i < BIN_SNAPSHOT_NUM; i++) {
-+ /* wait RECR1_CTL_SNP_DONE_MASK has cleared */
-+ timeout = 100;
-+ while (ioread32(&reg_base->recr1) &
-+ RECR1_CTL_SNP_DONE_MASK) {
-+ udelay(1);
-+ timeout--;
-+ if (timeout == 0)
-+ break;
-+ }
-+
-+ /* start snap shot */
-+ iowrite32((ioread32(&reg_base->gcr1) |
-+ GCR1_CTL_SNP_START_MASK),
-+ &reg_base->gcr1);
-+
-+ /* wait for SNP done */
-+ timeout = 100;
-+ while (!(ioread32(&reg_base->recr1) &
-+ RECR1_CTL_SNP_DONE_MASK)) {
-+ udelay(1);
-+ timeout--;
-+ if (timeout == 0)
-+ break;
-+ }
-+
-+ /* read and save the snap shot */
-+ rx_eq_snp = ioread32(&reg_base->recr1);
-+ gaink2_snap_shot[i] = (rx_eq_snp & RECR1_GAINK2_MASK) >>
-+ RECR1_GAINK2_SHIFT;
-+
-+ /* terminate the snap shot by setting GCR1[REQ_CTL_SNP] */
-+ iowrite32((ioread32(&reg_base->gcr1) &
-+ ~GCR1_CTL_SNP_START_MASK),
-+ &reg_base->gcr1);
-+ }
-+
-+ /* get median of the 5 snap shot */
-+ for (i = 0; i < BIN_SNAPSHOT_NUM - 1; i++) {
-+ tmp = gaink2_snap_shot[i];
-+ pos = i;
-+ for (j = i + 1; j < BIN_SNAPSHOT_NUM; j++) {
-+ if (gaink2_snap_shot[j] < tmp) {
-+ tmp = gaink2_snap_shot[j];
-+ pos = j;
-+ }
-+ }
-+
-+ gaink2_snap_shot[pos] = gaink2_snap_shot[i];
-+ gaink2_snap_shot[i] = tmp;
-+ }
-+
-+ return gaink2_snap_shot[2];
-+}
-+#endif
-+
-+static bool is_bin_early(int bin_sel, void *reg)
-+{
-+ bool early = false;
-+ int bin_snap_shot[BIN_SNAPSHOT_NUM];
-+ int i, negative_count = 0;
-+ struct per_lane_ctrl_status *reg_base = reg;
-+ int timeout;
-+
-+ for (i = 0; i < BIN_SNAPSHOT_NUM; i++) {
-+ /* wait RECR1_SNP_DONE_MASK has cleared */
-+ timeout = 100;
-+ while ((ioread32(&reg_base->recr1) & RECR1_SNP_DONE_MASK)) {
-+ udelay(1);
-+ timeout--;
-+ if (timeout == 0)
-+ break;
-+ }
-+
-+ /* set TCSR1[CDR_SEL] to BinM1/BinLong */
-+ if (bin_sel == BIN_M1) {
-+ iowrite32((ioread32(&reg_base->tcsr1) &
-+ ~CDR_SEL_MASK) | BIN_M1_SEL,
-+ &reg_base->tcsr1);
-+ } else {
-+ iowrite32((ioread32(&reg_base->tcsr1) &
-+ ~CDR_SEL_MASK) | BIN_Long_SEL,
-+ &reg_base->tcsr1);
-+ }
-+
-+ /* start snap shot */
-+ iowrite32(ioread32(&reg_base->gcr1) | GCR1_SNP_START_MASK,
-+ &reg_base->gcr1);
-+
-+ /* wait for SNP done */
-+ timeout = 100;
-+ while (!(ioread32(&reg_base->recr1) & RECR1_SNP_DONE_MASK)) {
-+ udelay(1);
-+ timeout--;
-+ if (timeout == 0)
-+ break;
-+ }
-+
-+ /* read and save the snap shot */
-+ bin_snap_shot[i] = (ioread32(&reg_base->tcsr1) &
-+ TCSR1_SNP_DATA_MASK) >> TCSR1_SNP_DATA_SHIFT;
-+ if (bin_snap_shot[i] & TCSR1_EQ_SNPBIN_SIGN_MASK)
-+ negative_count++;
-+
-+ /* terminate the snap shot by setting GCR1[REQ_CTL_SNP] */
-+ iowrite32(ioread32(&reg_base->gcr1) & ~GCR1_SNP_START_MASK,
-+ &reg_base->gcr1);
-+ }
-+
-+ if (((bin_sel == BIN_M1) && (negative_count > BIN_M1_THRESHOLD)) ||
-+ ((bin_sel == BIN_LONG && (negative_count > BIN_LONG_THRESHOLD)))) {
-+ early = true;
-+ }
-+
-+ return early;
-+}
-+
-+static void train_tx(struct fsl_xgkr_inst *inst)
-+{
-+ struct phy_device *phydev = inst->phydev;
-+ struct tx_condition *tx_c = &inst->tx_c;
-+ bool bin_m1_early, bin_long_early;
-+ u32 lp_status, old_ld_update;
-+ u32 status_cop1, status_coz, status_com1;
-+ u32 req_cop1, req_coz, req_com1, req_preset, req_init;
-+ u32 temp;
-+#ifdef NEW_ALGORITHM_TRAIN_TX
-+ u32 median_gaink2;
-+#endif
-+
-+recheck:
-+ if (tx_c->bin_long_stop && tx_c->bin_m1_stop) {
-+ tx_c->tx_complete = true;
-+ inst->ld_status |= RX_READY_MASK;
-+ ld_coe_status(inst);
-+ /* tell LP we are ready */
-+ phy_write_mmd(phydev, MDIO_MMD_PMAPMD,
-+ FSL_KR_PMD_STATUS, RX_STAT);
-+ return;
-+ }
-+
-+ /* We start by checking the current LP status. If we got any responses,
-+ * we can clear up the appropriate update request so that the
-+ * subsequent code may easily issue new update requests if needed.
-+ */
-+ lp_status = phy_read_mmd(phydev, MDIO_MMD_PMAPMD, FSL_KR_LP_STATUS) &
-+ REQUEST_MASK;
-+ status_cop1 = (lp_status & COP1_MASK) >> COP1_SHIFT;
-+ status_coz = (lp_status & COZ_MASK) >> COZ_SHIFT;
-+ status_com1 = (lp_status & COM1_MASK) >> COM1_SHIFT;
-+
-+ old_ld_update = inst->ld_update;
-+ req_cop1 = (old_ld_update & COP1_MASK) >> COP1_SHIFT;
-+ req_coz = (old_ld_update & COZ_MASK) >> COZ_SHIFT;
-+ req_com1 = (old_ld_update & COM1_MASK) >> COM1_SHIFT;
-+ req_preset = old_ld_update & PRESET_MASK;
-+ req_init = old_ld_update & INIT_MASK;
-+
-+ /* IEEE802.3-2008, 72.6.10.2.3.1
-+ * We may clear PRESET when all coefficients show UPDATED or MAX.
-+ */
-+ if (req_preset) {
-+ if ((status_cop1 == COE_UPDATED || status_cop1 == COE_MAX) &&
-+ (status_coz == COE_UPDATED || status_coz == COE_MAX) &&
-+ (status_com1 == COE_UPDATED || status_com1 == COE_MAX)) {
-+ inst->ld_update &= ~PRESET_MASK;
-+ }
-+ }
-+
-+ /* IEEE802.3-2008, 72.6.10.2.3.2
-+ * We may clear INITIALIZE when no coefficients show NOT UPDATED.
-+ */
-+ if (req_init) {
-+ if (status_cop1 != COE_NOTUPDATED &&
-+ status_coz != COE_NOTUPDATED &&
-+ status_com1 != COE_NOTUPDATED) {
-+ inst->ld_update &= ~INIT_MASK;
-+ }
-+ }
-+
-+ /* IEEE802.3-2008, 72.6.10.2.3.2
-+ * we send initialize to the other side to ensure default settings
-+ * for the LP. Naturally, we should do this only once.
-+ */
-+ if (!tx_c->sent_init) {
-+ if (!lp_status && !(old_ld_update & (LD_ALL_MASK))) {
-+ inst->ld_update = INIT_MASK;
-+ tx_c->sent_init = true;
-+ }
-+ }
-+
-+ /* IEEE802.3-2008, 72.6.10.2.3.3
-+ * We set coefficient requests to HOLD when we get the information
-+ * about any updates On clearing our prior response, we also update
-+ * our internal status.
-+ */
-+ if (status_cop1 != COE_NOTUPDATED) {
-+ if (req_cop1) {
-+ inst->ld_update &= ~COP1_MASK;
-+#ifdef NEW_ALGORITHM_TRAIN_TX
-+ if (tx_c->post_inc) {
-+ if (req_cop1 == INCREMENT &&
-+ status_cop1 == COE_MAX) {
-+ tx_c->post_inc = 0;
-+ tx_c->bin_long_stop = true;
-+ tx_c->bin_m1_stop = true;
-+ } else {
-+ tx_c->post_inc -= 1;
-+ }
-+
-+ ld_coe_update(inst);
-+ goto recheck;
-+ }
-+#endif
-+ if ((req_cop1 == DECREMENT && status_cop1 == COE_MIN) ||
-+ (req_cop1 == INCREMENT && status_cop1 == COE_MAX)) {
-+ dev_dbg(&inst->phydev->mdio.dev, "COP1 hit limit %s",
-+ (status_cop1 == COE_MIN) ?
-+ "DEC MIN" : "INC MAX");
-+ tx_c->long_min_max_cnt++;
-+ if (tx_c->long_min_max_cnt >= TIMEOUT_LONG) {
-+ tx_c->bin_long_stop = true;
-+ ld_coe_update(inst);
-+ goto recheck;
-+ }
-+ }
-+ }
-+ }
-+
-+ if (status_coz != COE_NOTUPDATED) {
-+ if (req_coz)
-+ inst->ld_update &= ~COZ_MASK;
-+ }
-+
-+ if (status_com1 != COE_NOTUPDATED) {
-+ if (req_com1) {
-+ inst->ld_update &= ~COM1_MASK;
-+#ifdef NEW_ALGORITHM_TRAIN_TX
-+ if (tx_c->pre_inc) {
-+ if (req_com1 == INCREMENT &&
-+ status_com1 == COE_MAX)
-+ tx_c->pre_inc = 0;
-+ else
-+ tx_c->pre_inc -= 1;
-+
-+ ld_coe_update(inst);
-+ goto recheck;
-+ }
-+#endif
-+ /* Stop If we have reached the limit for a parameter. */
-+ if ((req_com1 == DECREMENT && status_com1 == COE_MIN) ||
-+ (req_com1 == INCREMENT && status_com1 == COE_MAX)) {
-+ dev_dbg(&inst->phydev->mdio.dev, "COM1 hit limit %s",
-+ (status_com1 == COE_MIN) ?
-+ "DEC MIN" : "INC MAX");
-+ tx_c->m1_min_max_cnt++;
-+ if (tx_c->m1_min_max_cnt >= TIMEOUT_M1) {
-+ tx_c->bin_m1_stop = true;
-+ ld_coe_update(inst);
-+ goto recheck;
-+ }
-+ }
-+ }
-+ }
-+
-+ if (old_ld_update != inst->ld_update) {
-+ ld_coe_update(inst);
-+ /* Redo these status checks and updates until we have no more
-+ * changes, to speed up the overall process.
-+ */
-+ goto recheck;
-+ }
-+
-+ /* Do nothing if we have pending request. */
-+ if ((req_coz || req_com1 || req_cop1))
-+ return;
-+ else if (lp_status)
-+ /* No pending request but LP status was not reverted to
-+ * not updated.
-+ */
-+ return;
-+
-+#ifdef NEW_ALGORITHM_TRAIN_TX
-+ if (!(inst->ld_update & (PRESET_MASK | INIT_MASK))) {
-+ if (tx_c->pre_inc) {
-+ inst->ld_update = INCREMENT << COM1_SHIFT;
-+ ld_coe_update(inst);
-+ return;
-+ }
-+
-+ if (status_cop1 != COE_MAX) {
-+ median_gaink2 = get_median_gaink2(inst->reg_base);
-+ if (median_gaink2 == 0xf) {
-+ tx_c->post_inc = 1;
-+ } else {
-+ /* Gaink2 median lower than "F" */
-+ tx_c->bin_m1_stop = true;
-+ tx_c->bin_long_stop = true;
-+ goto recheck;
-+ }
-+ } else {
-+ /* C1 MAX */
-+ tx_c->bin_m1_stop = true;
-+ tx_c->bin_long_stop = true;
-+ goto recheck;
-+ }
-+
-+ if (tx_c->post_inc) {
-+ inst->ld_update = INCREMENT << COP1_SHIFT;
-+ ld_coe_update(inst);
-+ return;
-+ }
-+ }
-+#endif
-+
-+ /* snapshot and select bin */
-+ bin_m1_early = is_bin_early(BIN_M1, inst->reg_base);
-+ bin_long_early = is_bin_early(BIN_LONG, inst->reg_base);
-+
-+ if (!tx_c->bin_m1_stop && !tx_c->bin_m1_late_early && bin_m1_early) {
-+ tx_c->bin_m1_stop = true;
-+ goto recheck;
-+ }
-+
-+ if (!tx_c->bin_long_stop &&
-+ tx_c->bin_long_late_early && !bin_long_early) {
-+ tx_c->bin_long_stop = true;
-+ goto recheck;
-+ }
-+
-+ /* IEEE802.3-2008, 72.6.10.2.3.3
-+ * We only request coefficient updates when no PRESET/INITIALIZE is
-+ * pending. We also only request coefficient updates when the
-+ * corresponding status is NOT UPDATED and nothing is pending.
-+ */
-+ if (!(inst->ld_update & (PRESET_MASK | INIT_MASK))) {
-+ if (!tx_c->bin_long_stop) {
-+ /* BinM1 correction means changing COM1 */
-+ if (!status_com1 && !(inst->ld_update & COM1_MASK)) {
-+ /* Avoid BinM1Late by requesting an
-+ * immediate decrement.
-+ */
-+ if (!bin_m1_early) {
-+ /* request decrement c(-1) */
-+ temp = DECREMENT << COM1_SHIFT;
-+ inst->ld_update = temp;
-+ ld_coe_update(inst);
-+ tx_c->bin_m1_late_early = bin_m1_early;
-+ return;
-+ }
-+ }
-+
-+ /* BinLong correction means changing COP1 */
-+ if (!status_cop1 && !(inst->ld_update & COP1_MASK)) {
-+ /* Locate BinLong transition point (if any)
-+ * while avoiding BinM1Late.
-+ */
-+ if (bin_long_early) {
-+ /* request increment c(1) */
-+ temp = INCREMENT << COP1_SHIFT;
-+ inst->ld_update = temp;
-+ } else {
-+ /* request decrement c(1) */
-+ temp = DECREMENT << COP1_SHIFT;
-+ inst->ld_update = temp;
-+ }
-+
-+ ld_coe_update(inst);
-+ tx_c->bin_long_late_early = bin_long_early;
-+ }
-+ /* We try to finish BinLong before we do BinM1 */
-+ return;
-+ }
-+
-+ if (!tx_c->bin_m1_stop) {
-+ /* BinM1 correction means changing COM1 */
-+ if (!status_com1 && !(inst->ld_update & COM1_MASK)) {
-+ /* Locate BinM1 transition point (if any) */
-+ if (bin_m1_early) {
-+ /* request increment c(-1) */
-+ temp = INCREMENT << COM1_SHIFT;
-+ inst->ld_update = temp;
-+ } else {
-+ /* request decrement c(-1) */
-+ temp = DECREMENT << COM1_SHIFT;
-+ inst->ld_update = temp;
-+ }
-+
-+ ld_coe_update(inst);
-+ tx_c->bin_m1_late_early = bin_m1_early;
-+ }
-+ }
-+ }
-+}
-+
-+static int is_link_up(struct phy_device *phydev)
-+{
-+ int val;
-+
-+ phy_read_mmd(phydev, MDIO_MMD_PCS, FSL_XFI_PCS_10GR_SR1);
-+ val = phy_read_mmd(phydev, MDIO_MMD_PCS, FSL_XFI_PCS_10GR_SR1);
-+
-+ return (val & FSL_KR_RX_LINK_STAT_MASK) ? 1 : 0;
-+}
-+
-+static int is_link_training_fail(struct phy_device *phydev)
-+{
-+ int val;
-+ int timeout = 100;
-+
-+ val = phy_read_mmd(phydev, MDIO_MMD_PMAPMD, FSL_KR_PMD_STATUS);
-+ if (!(val & TRAIN_FAIL) && (val & RX_STAT)) {
-+ /* check LNK_STAT for sure */
-+ while (timeout--) {
-+ if (is_link_up(phydev))
-+ return 0;
-+
-+ usleep_range(100, 500);
-+ }
-+ }
-+
-+ return 1;
-+}
-+
-+static int check_rx(struct phy_device *phydev)
-+{
-+ return phy_read_mmd(phydev, MDIO_MMD_PMAPMD, FSL_KR_LP_STATUS) &
-+ RX_READY_MASK;
-+}
-+
-+/* Coefficient values have hardware restrictions */
-+static int is_ld_valid(struct fsl_xgkr_inst *inst)
-+{
-+ u32 ratio_pst1q = inst->ratio_pst1q;
-+ u32 adpt_eq = inst->adpt_eq;
-+ u32 ratio_preq = inst->ratio_preq;
-+
-+ if ((ratio_pst1q + adpt_eq + ratio_preq) > 48)
-+ return 0;
-+
-+ if (((ratio_pst1q + adpt_eq + ratio_preq) * 4) >=
-+ ((adpt_eq - ratio_pst1q - ratio_preq) * 17))
-+ return 0;
-+
-+ if (ratio_preq > ratio_pst1q)
-+ return 0;
-+
-+ if (ratio_preq > 8)
-+ return 0;
-+
-+ if (adpt_eq < 26)
-+ return 0;
-+
-+ if (ratio_pst1q > 16)
-+ return 0;
-+
-+ return 1;
-+}
-+
-+static int is_value_allowed(const u32 *val_table, u32 val)
-+{
-+ int i;
-+
-+ for (i = 0;; i++) {
-+ if (*(val_table + i) == VAL_INVALID)
-+ return 0;
-+ if (*(val_table + i) == val)
-+ return 1;
-+ }
-+}
-+
-+static int inc_dec(struct fsl_xgkr_inst *inst, int field, int request)
-+{
-+ u32 ld_limit[3], ld_coe[3], step[3];
-+
-+ ld_coe[0] = inst->ratio_pst1q;
-+ ld_coe[1] = inst->adpt_eq;
-+ ld_coe[2] = inst->ratio_preq;
-+
-+ /* Information specific to the Freescale SerDes for 10GBase-KR:
-+ * Incrementing C(+1) means *decrementing* RATIO_PST1Q
-+ * Incrementing C(0) means incrementing ADPT_EQ
-+ * Incrementing C(-1) means *decrementing* RATIO_PREQ
-+ */
-+ step[0] = -1;
-+ step[1] = 1;
-+ step[2] = -1;
-+
-+ switch (request) {
-+ case INCREMENT:
-+ ld_limit[0] = POST_COE_MAX;
-+ ld_limit[1] = ZERO_COE_MAX;
-+ ld_limit[2] = PRE_COE_MAX;
-+ if (ld_coe[field] != ld_limit[field])
-+ ld_coe[field] += step[field];
-+ else
-+ /* MAX */
-+ return 2;
-+ break;
-+ case DECREMENT:
-+ ld_limit[0] = POST_COE_MIN;
-+ ld_limit[1] = ZERO_COE_MIN;
-+ ld_limit[2] = PRE_COE_MIN;
-+ if (ld_coe[field] != ld_limit[field])
-+ ld_coe[field] -= step[field];
-+ else
-+ /* MIN */
-+ return 1;
-+ break;
-+ default:
-+ break;
-+ }
-+
-+ if (is_ld_valid(inst)) {
-+ /* accept new ld */
-+ inst->ratio_pst1q = ld_coe[0];
-+ inst->adpt_eq = ld_coe[1];
-+ inst->ratio_preq = ld_coe[2];
-+ /* only some values for preq and pst1q can be used.
-+ * for preq: 0x0, 0x1, 0x3, 0x5, 0x7, 0x9, 0xb, 0xc.
-+ * for pst1q: 0x0, 0x1, 0x3, 0x5, 0x7, 0x9, 0xb, 0xd, 0xf, 0x10.
-+ */
-+ if (!is_value_allowed((const u32 *)&preq_table, ld_coe[2])) {
-+ dev_dbg(&inst->phydev->mdio.dev,
-+ "preq skipped value: %d\n", ld_coe[2]);
-+ return 0;
-+ }
-+
-+ if (!is_value_allowed((const u32 *)&pst1q_table, ld_coe[0])) {
-+ dev_dbg(&inst->phydev->mdio.dev,
-+ "pst1q skipped value: %d\n", ld_coe[0]);
-+ return 0;
-+ }
-+
-+ tune_tecr0(inst);
-+ } else {
-+ if (request == DECREMENT)
-+ /* MIN */
-+ return 1;
-+ if (request == INCREMENT)
-+ /* MAX */
-+ return 2;
-+ }
-+
-+ return 0;
-+}
-+
-+static void min_max_updated(struct fsl_xgkr_inst *inst, int field, int new_ld)
-+{
-+ u32 ld_coe[] = {COE_UPDATED, COE_MIN, COE_MAX};
-+ u32 mask, val;
-+
-+ switch (field) {
-+ case COE_COP1:
-+ mask = COP1_MASK;
-+ val = ld_coe[new_ld] << COP1_SHIFT;
-+ break;
-+ case COE_COZ:
-+ mask = COZ_MASK;
-+ val = ld_coe[new_ld] << COZ_SHIFT;
-+ break;
-+ case COE_COM:
-+ mask = COM1_MASK;
-+ val = ld_coe[new_ld] << COM1_SHIFT;
-+ break;
-+ default:
-+ return;
-+ }
-+
-+ inst->ld_status &= ~mask;
-+ inst->ld_status |= val;
-+}
-+
-+static void check_request(struct fsl_xgkr_inst *inst, int request)
-+{
-+ int cop1_req, coz_req, com_req;
-+ int old_status, new_ld_sta;
-+
-+ cop1_req = (request & COP1_MASK) >> COP1_SHIFT;
-+ coz_req = (request & COZ_MASK) >> COZ_SHIFT;
-+ com_req = (request & COM1_MASK) >> COM1_SHIFT;
-+
-+ /* IEEE802.3-2008, 72.6.10.2.5
-+ * Ensure we only act on INCREMENT/DECREMENT when we are in NOT UPDATED
-+ */
-+ old_status = inst->ld_status;
-+
-+ if (cop1_req && !(inst->ld_status & COP1_MASK)) {
-+ new_ld_sta = inc_dec(inst, COE_COP1, cop1_req);
-+ min_max_updated(inst, COE_COP1, new_ld_sta);
-+ }
-+
-+ if (coz_req && !(inst->ld_status & COZ_MASK)) {
-+ new_ld_sta = inc_dec(inst, COE_COZ, coz_req);
-+ min_max_updated(inst, COE_COZ, new_ld_sta);
-+ }
-+
-+ if (com_req && !(inst->ld_status & COM1_MASK)) {
-+ new_ld_sta = inc_dec(inst, COE_COM, com_req);
-+ min_max_updated(inst, COE_COM, new_ld_sta);
-+ }
-+
-+ if (old_status != inst->ld_status)
-+ ld_coe_status(inst);
-+}
-+
-+static void preset(struct fsl_xgkr_inst *inst)
-+{
-+ /* These are all MAX values from the IEEE802.3 perspective. */
-+ inst->ratio_pst1q = POST_COE_MAX;
-+ inst->adpt_eq = ZERO_COE_MAX;
-+ inst->ratio_preq = PRE_COE_MAX;
-+
-+ tune_tecr0(inst);
-+ inst->ld_status &= ~(COP1_MASK | COZ_MASK | COM1_MASK);
-+ inst->ld_status |= COE_MAX << COP1_SHIFT |
-+ COE_MAX << COZ_SHIFT |
-+ COE_MAX << COM1_SHIFT;
-+ ld_coe_status(inst);
-+}
-+
-+static void initialize(struct fsl_xgkr_inst *inst)
-+{
-+ inst->ratio_preq = RATIO_PREQ;
-+ inst->ratio_pst1q = RATIO_PST1Q;
-+ inst->adpt_eq = RATIO_EQ;
-+
-+ tune_tecr0(inst);
-+ inst->ld_status &= ~(COP1_MASK | COZ_MASK | COM1_MASK);
-+ inst->ld_status |= COE_UPDATED << COP1_SHIFT |
-+ COE_UPDATED << COZ_SHIFT |
-+ COE_UPDATED << COM1_SHIFT;
-+ ld_coe_status(inst);
-+}
-+
-+static void train_rx(struct fsl_xgkr_inst *inst)
-+{
-+ struct phy_device *phydev = inst->phydev;
-+ int request, old_ld_status;
-+
-+ /* get request from LP */
-+ request = phy_read_mmd(phydev, MDIO_MMD_PMAPMD, FSL_KR_LP_CU) &
-+ (LD_ALL_MASK);
-+ old_ld_status = inst->ld_status;
-+
-+ /* IEEE802.3-2008, 72.6.10.2.5
-+ * Ensure we always go to NOT UDPATED for status reporting in
-+ * response to HOLD requests.
-+ * IEEE802.3-2008, 72.6.10.2.3.1/2
-+ * ... but only if PRESET/INITIALIZE are not active to ensure
-+ * we keep status until they are released.
-+ */
-+ if (!(request & (PRESET_MASK | INIT_MASK))) {
-+ if (!(request & COP1_MASK))
-+ inst->ld_status &= ~COP1_MASK;
-+
-+ if (!(request & COZ_MASK))
-+ inst->ld_status &= ~COZ_MASK;
-+
-+ if (!(request & COM1_MASK))
-+ inst->ld_status &= ~COM1_MASK;
-+
-+ if (old_ld_status != inst->ld_status)
-+ ld_coe_status(inst);
-+ }
-+
-+ /* As soon as the LP shows ready, no need to do any more updates. */
-+ if (check_rx(phydev)) {
-+ /* LP receiver is ready */
-+ if (inst->ld_status & (COP1_MASK | COZ_MASK | COM1_MASK)) {
-+ inst->ld_status &= ~(COP1_MASK | COZ_MASK | COM1_MASK);
-+ ld_coe_status(inst);
-+ }
-+ } else {
-+ /* IEEE802.3-2008, 72.6.10.2.3.1/2
-+ * only act on PRESET/INITIALIZE if all status is NOT UPDATED.
-+ */
-+ if (request & (PRESET_MASK | INIT_MASK)) {
-+ if (!(inst->ld_status &
-+ (COP1_MASK | COZ_MASK | COM1_MASK))) {
-+ if (request & PRESET_MASK)
-+ preset(inst);
-+
-+ if (request & INIT_MASK)
-+ initialize(inst);
-+ }
-+ }
-+
-+ /* LP Coefficient are not in HOLD */
-+ if (request & REQUEST_MASK)
-+ check_request(inst, request & REQUEST_MASK);
-+ }
-+}
-+
-+static void xgkr_start_train(struct phy_device *phydev)
-+{
-+ struct fsl_xgkr_inst *inst = phydev->priv;
-+ struct tx_condition *tx_c = &inst->tx_c;
-+ int val = 0, i;
-+ int lt_state;
-+ unsigned long dead_line;
-+ int rx_ok, tx_ok;
-+
-+ init_inst(inst, 0);
-+ start_lt(phydev);
-+
-+ for (i = 0; i < 2;) {
-+ dead_line = jiffies + msecs_to_jiffies(500);
-+ while (time_before(jiffies, dead_line)) {
-+ val = phy_read_mmd(phydev, MDIO_MMD_PMAPMD,
-+ FSL_KR_PMD_STATUS);
-+ if (val & TRAIN_FAIL) {
-+ /* LT failed already, reset lane to avoid
-+ * it run into hanging, then start LT again.
-+ */
-+ reset_gcr0(inst);
-+ start_lt(phydev);
-+ } else if ((val & PMD_STATUS_SUP_STAT) &&
-+ (val & PMD_STATUS_FRAME_LOCK))
-+ break;
-+ usleep_range(100, 500);
-+ }
-+
-+ if (!((val & PMD_STATUS_FRAME_LOCK) &&
-+ (val & PMD_STATUS_SUP_STAT))) {
-+ i++;
-+ continue;
-+ }
-+
-+ /* init process */
-+ rx_ok = false;
-+ tx_ok = false;
-+ /* the LT should be finished in 500ms, failed or OK. */
-+ dead_line = jiffies + msecs_to_jiffies(500);
-+
-+ while (time_before(jiffies, dead_line)) {
-+ /* check if the LT is already failed */
-+ lt_state = phy_read_mmd(phydev, MDIO_MMD_PMAPMD,
-+ FSL_KR_PMD_STATUS);
-+ if (lt_state & TRAIN_FAIL) {
-+ reset_gcr0(inst);
-+ break;
-+ }
-+
-+ rx_ok = check_rx(phydev);
-+ tx_ok = tx_c->tx_complete;
-+
-+ if (rx_ok && tx_ok)
-+ break;
-+
-+ if (!rx_ok)
-+ train_rx(inst);
-+
-+ if (!tx_ok)
-+ train_tx(inst);
-+
-+ usleep_range(100, 500);
-+ }
-+
-+ i++;
-+ /* check LT result */
-+ if (is_link_training_fail(phydev)) {
-+ init_inst(inst, 0);
-+ continue;
-+ } else {
-+ stop_lt(phydev);
-+ inst->state = TRAINED;
-+ break;
-+ }
-+ }
-+}
-+
-+static void xgkr_state_machine(struct work_struct *work)
-+{
-+ struct delayed_work *dwork = to_delayed_work(work);
-+ struct fsl_xgkr_inst *inst = container_of(dwork,
-+ struct fsl_xgkr_inst,
-+ xgkr_wk);
-+ struct phy_device *phydev = inst->phydev;
-+ int an_state;
-+ bool needs_train = false;
-+
-+ mutex_lock(&phydev->lock);
-+
-+ switch (inst->state) {
-+ case DETECTING_LP:
-+ phy_read_mmd(phydev, MDIO_MMD_AN, FSL_AN_BP_STAT);
-+ an_state = phy_read_mmd(phydev, MDIO_MMD_AN, FSL_AN_BP_STAT);
-+ if ((an_state & KR_AN_MASK))
-+ needs_train = true;
-+ break;
-+ case TRAINED:
-+ if (!is_link_up(phydev)) {
-+ dev_info(&phydev->mdio.dev,
-+ "Detect hotplug, restart training\n");
-+ init_inst(inst, 1);
-+ start_xgkr_an(phydev);
-+ inst->state = DETECTING_LP;
-+ }
-+ break;
-+ }
-+
-+ if (needs_train)
-+ xgkr_start_train(phydev);
-+
-+ mutex_unlock(&phydev->lock);
-+ queue_delayed_work(system_power_efficient_wq, &inst->xgkr_wk,
-+ msecs_to_jiffies(XGKR_TIMEOUT));
-+}
-+
-+static int fsl_backplane_probe(struct phy_device *phydev)
-+{
-+ struct fsl_xgkr_inst *xgkr_inst;
-+ struct device_node *phy_node, *lane_node;
-+ struct resource res_lane;
-+ const char *bm;
-+ int ret;
-+ int bp_mode;
-+ u32 lane[2];
-+
-+ phy_node = phydev->mdio.dev.of_node;
-+ bp_mode = of_property_read_string(phy_node, "backplane-mode", &bm);
-+ if (bp_mode < 0)
-+ return 0;
-+
-+ if (!strcasecmp(bm, "1000base-kx")) {
-+ bp_mode = PHY_BACKPLANE_1000BASE_KX;
-+ } else if (!strcasecmp(bm, "10gbase-kr")) {
-+ bp_mode = PHY_BACKPLANE_10GBASE_KR;
-+ } else {
-+ dev_err(&phydev->mdio.dev, "Unknown backplane-mode\n");
-+ return -EINVAL;
-+ }
-+
-+ lane_node = of_parse_phandle(phy_node, "fsl,lane-handle", 0);
-+ if (!lane_node) {
-+ dev_err(&phydev->mdio.dev, "parse fsl,lane-handle failed\n");
-+ return -EINVAL;
-+ }
-+
-+ ret = of_address_to_resource(lane_node, 0, &res_lane);
-+ if (ret) {
-+ dev_err(&phydev->mdio.dev, "could not obtain memory map\n");
-+ return ret;
-+ }
-+
-+ of_node_put(lane_node);
-+ ret = of_property_read_u32_array(phy_node, "fsl,lane-reg",
-+ (u32 *)&lane, 2);
-+ if (ret) {
-+ dev_err(&phydev->mdio.dev, "could not get fsl,lane-reg\n");
-+ return -EINVAL;
-+ }
-+
-+ phydev->priv = devm_ioremap_nocache(&phydev->mdio.dev,
-+ res_lane.start + lane[0],
-+ lane[1]);
-+ if (!phydev->priv) {
-+ dev_err(&phydev->mdio.dev, "ioremap_nocache failed\n");
-+ return -ENOMEM;
-+ }
-+
-+ if (bp_mode == PHY_BACKPLANE_1000BASE_KX) {
-+ phydev->speed = SPEED_1000;
-+ /* configure the lane for 1000BASE-KX */
-+ lane_set_1gkx(phydev->priv);
-+ return 0;
-+ }
-+
-+ xgkr_inst = devm_kzalloc(&phydev->mdio.dev,
-+ sizeof(*xgkr_inst), GFP_KERNEL);
-+ if (!xgkr_inst)
-+ return -ENOMEM;
-+
-+ xgkr_inst->reg_base = phydev->priv;
-+ xgkr_inst->phydev = phydev;
-+ phydev->priv = xgkr_inst;
-+
-+ if (bp_mode == PHY_BACKPLANE_10GBASE_KR) {
-+ phydev->speed = SPEED_10000;
-+ INIT_DELAYED_WORK(&xgkr_inst->xgkr_wk, xgkr_state_machine);
-+ }
-+
-+ return 0;
-+}
-+
-+static int fsl_backplane_aneg_done(struct phy_device *phydev)
-+{
-+ return 1;
-+}
-+
-+static int fsl_backplane_config_aneg(struct phy_device *phydev)
-+{
-+ if (phydev->speed == SPEED_10000) {
-+ phydev->supported |= SUPPORTED_10000baseKR_Full;
-+ start_xgkr_an(phydev);
-+ } else if (phydev->speed == SPEED_1000) {
-+ phydev->supported |= SUPPORTED_1000baseKX_Full;
-+ start_1gkx_an(phydev);
-+ }
-+
-+ phydev->advertising = phydev->supported;
-+ phydev->duplex = 1;
-+
-+ return 0;
-+}
-+
-+static int fsl_backplane_suspend(struct phy_device *phydev)
-+{
-+ if (phydev->speed == SPEED_10000) {
-+ struct fsl_xgkr_inst *xgkr_inst = phydev->priv;
-+
-+ cancel_delayed_work_sync(&xgkr_inst->xgkr_wk);
-+ }
-+ return 0;
-+}
-+
-+static int fsl_backplane_resume(struct phy_device *phydev)
-+{
-+ if (phydev->speed == SPEED_10000) {
-+ struct fsl_xgkr_inst *xgkr_inst = phydev->priv;
-+
-+ init_inst(xgkr_inst, 1);
-+ queue_delayed_work(system_power_efficient_wq,
-+ &xgkr_inst->xgkr_wk,
-+ msecs_to_jiffies(XGKR_TIMEOUT));
-+ }
-+ return 0;
-+}
-+
-+static int fsl_backplane_read_status(struct phy_device *phydev)
-+{
-+ if (is_link_up(phydev))
-+ phydev->link = 1;
-+ else
-+ phydev->link = 0;
-+
-+ return 0;
-+}
-+
-+static struct phy_driver fsl_backplane_driver[] = {
-+ {
-+ .phy_id = FSL_PCS_PHY_ID,
-+ .name = "Freescale Backplane",
-+ .phy_id_mask = 0xffffffff,
-+ .features = SUPPORTED_Backplane | SUPPORTED_Autoneg |
-+ SUPPORTED_MII,
-+ .probe = fsl_backplane_probe,
-+ .aneg_done = fsl_backplane_aneg_done,
-+ .config_aneg = fsl_backplane_config_aneg,
-+ .read_status = fsl_backplane_read_status,
-+ .suspend = fsl_backplane_suspend,
-+ .resume = fsl_backplane_resume,
-+ },
-+};
-+
-+module_phy_driver(fsl_backplane_driver);
-+
-+static struct mdio_device_id __maybe_unused freescale_tbl[] = {
-+ { FSL_PCS_PHY_ID, 0xffffffff },
-+ { }
-+};
-+
-+MODULE_DEVICE_TABLE(mdio, freescale_tbl);
-+
-+MODULE_DESCRIPTION("Freescale Backplane driver");
-+MODULE_AUTHOR("Shaohui Xie <Shaohui.Xie@freescale.com>");
-+MODULE_LICENSE("GPL v2");
---- a/drivers/net/phy/phy.c
-+++ b/drivers/net/phy/phy.c
-@@ -591,7 +591,7 @@ int phy_mii_ioctl(struct phy_device *phy
- return 0;
-
- case SIOCSHWTSTAMP:
-- if (phydev->drv->hwtstamp)
-+ if (phydev->drv && phydev->drv->hwtstamp)
- return phydev->drv->hwtstamp(phydev, ifr);
- /* fall through */
-
-@@ -616,6 +616,9 @@ static int phy_start_aneg_priv(struct ph
- bool trigger = 0;
- int err;
-
-+ if (!phydev->drv)
-+ return -EIO;
-+
- mutex_lock(&phydev->lock);
-
- if (AUTONEG_DISABLE == phydev->autoneg)
-@@ -1015,7 +1018,7 @@ void phy_state_machine(struct work_struc
-
- old_state = phydev->state;
-
-- if (phydev->drv->link_change_notify)
-+ if (phydev->drv && phydev->drv->link_change_notify)
- phydev->drv->link_change_notify(phydev);
-
- switch (phydev->state) {
-@@ -1317,6 +1320,9 @@ EXPORT_SYMBOL(phy_write_mmd_indirect);
- */
- int phy_init_eee(struct phy_device *phydev, bool clk_stop_enable)
- {
-+ if (!phydev->drv)
-+ return -EIO;
-+
- /* According to 802.3az,the EEE is supported only in full duplex-mode.
- * Also EEE feature is active when core is operating with MII, GMII
- * or RGMII (all kinds). Internal PHYs are also allowed to proceed and
-@@ -1394,6 +1400,9 @@ EXPORT_SYMBOL(phy_init_eee);
- */
- int phy_get_eee_err(struct phy_device *phydev)
- {
-+ if (!phydev->drv)
-+ return -EIO;
-+
- return phy_read_mmd_indirect(phydev, MDIO_PCS_EEE_WK_ERR, MDIO_MMD_PCS);
- }
- EXPORT_SYMBOL(phy_get_eee_err);
-@@ -1410,6 +1419,9 @@ int phy_ethtool_get_eee(struct phy_devic
- {
- int val;
-
-+ if (!phydev->drv)
-+ return -EIO;
-+
- /* Get Supported EEE */
- val = phy_read_mmd_indirect(phydev, MDIO_PCS_EEE_ABLE, MDIO_MMD_PCS);
- if (val < 0)
-@@ -1443,6 +1455,9 @@ int phy_ethtool_set_eee(struct phy_devic
- {
- int val = ethtool_adv_to_mmd_eee_adv_t(data->advertised);
-
-+ if (!phydev->drv)
-+ return -EIO;
-+
- /* Mask prohibited EEE modes */
- val &= ~phydev->eee_broken_modes;
-
-@@ -1454,7 +1469,7 @@ EXPORT_SYMBOL(phy_ethtool_set_eee);
-
- int phy_ethtool_set_wol(struct phy_device *phydev, struct ethtool_wolinfo *wol)
- {
-- if (phydev->drv->set_wol)
-+ if (phydev->drv && phydev->drv->set_wol)
- return phydev->drv->set_wol(phydev, wol);
-
- return -EOPNOTSUPP;
-@@ -1463,7 +1478,7 @@ EXPORT_SYMBOL(phy_ethtool_set_wol);
-
- void phy_ethtool_get_wol(struct phy_device *phydev, struct ethtool_wolinfo *wol)
- {
-- if (phydev->drv->get_wol)
-+ if (phydev->drv && phydev->drv->get_wol)
- phydev->drv->get_wol(phydev, wol);
- }
- EXPORT_SYMBOL(phy_ethtool_get_wol);
---- a/drivers/net/phy/phy_device.c
-+++ b/drivers/net/phy/phy_device.c
-@@ -1046,7 +1046,7 @@ int phy_suspend(struct phy_device *phyde
- if (wol.wolopts)
- return -EBUSY;
-
-- if (phydrv->suspend)
-+ if (phydev->drv && phydrv->suspend)
- ret = phydrv->suspend(phydev);
-
- if (ret)
-@@ -1063,7 +1063,7 @@ int phy_resume(struct phy_device *phydev
- struct phy_driver *phydrv = to_phy_driver(phydev->mdio.dev.driver);
- int ret = 0;
-
-- if (phydrv->resume)
-+ if (phydev->drv && phydrv->resume)
- ret = phydrv->resume(phydev);
-
- if (ret)
-@@ -1720,7 +1720,7 @@ static int phy_remove(struct device *dev
- phydev->state = PHY_DOWN;
- mutex_unlock(&phydev->lock);
-
-- if (phydev->drv->remove)
-+ if (phydev->drv && phydev->drv->remove)
- phydev->drv->remove(phydev);
- phydev->drv = NULL;
-
---- a/drivers/net/phy/swphy.c
-+++ b/drivers/net/phy/swphy.c
-@@ -77,6 +77,7 @@ static const struct swmii_regs duplex[]
- static int swphy_decode_speed(int speed)
- {
- switch (speed) {
-+ case 10000:
- case 1000:
- return SWMII_SPEED_1000;
- case 100:
---- a/include/linux/phy.h
-+++ b/include/linux/phy.h
-@@ -81,6 +81,7 @@ typedef enum {
- PHY_INTERFACE_MODE_MOCA,
- PHY_INTERFACE_MODE_QSGMII,
- PHY_INTERFACE_MODE_TRGMII,
-+ PHY_INTERFACE_MODE_2500SGMII,
- PHY_INTERFACE_MODE_MAX,
- } phy_interface_t;
-
-@@ -126,6 +127,8 @@ static inline const char *phy_modes(phy_
- return "qsgmii";
- case PHY_INTERFACE_MODE_TRGMII:
- return "trgmii";
-+ case PHY_INTERFACE_MODE_2500SGMII:
-+ return "sgmii-2500";
- default:
- return "unknown";
- }
-@@ -791,6 +794,9 @@ int phy_stop_interrupts(struct phy_devic
-
- static inline int phy_read_status(struct phy_device *phydev)
- {
-+ if (!phydev->drv)
-+ return -EIO;
-+
- return phydev->drv->read_status(phydev);
- }
-
diff --git a/target/linux/layerscape/patches-4.9/704-fsl-mc-layerscape-support.patch b/target/linux/layerscape/patches-4.9/704-fsl-mc-layerscape-support.patch
deleted file mode 100644
index 52c099203c..0000000000
--- a/target/linux/layerscape/patches-4.9/704-fsl-mc-layerscape-support.patch
+++ /dev/null
@@ -1,20043 +0,0 @@
-From ab7b47676f9334bb55f80e0ac096c7aa289810e2 Mon Sep 17 00:00:00 2001
-From: Yangbo Lu <yangbo.lu@nxp.com>
-Date: Thu, 5 Jul 2018 16:44:34 +0800
-Subject: [PATCH 10/32] fsl-mc: layerscape support
-MIME-Version: 1.0
-Content-Type: text/plain; charset=UTF-8
-Content-Transfer-Encoding: 8bit
-
-This is an integrated patch for layerscape mc-bus support.
-
-Signed-off-by: Stuart Yoder <stuart.yoder@nxp.com>
-Signed-off-by: Bharat Bhushan <Bharat.Bhushan@nxp.com>
-Signed-off-by: Arnd Bergmann <arnd@arndb.de>
-Signed-off-by: Laurentiu Tudor <laurentiu.tudor@nxp.com>
-Signed-off-by: Roy Pledge <roy.pledge@nxp.com>
-Signed-off-by: Shiva Kerdel <shiva@exdev.nl>
-Signed-off-by: Nipun Gupta <nipun.gupta@nxp.com>
-Signed-off-by: Ioana Ciornei <ioana.ciornei@nxp.com>
-Signed-off-by: Horia Geantă <horia.geanta@nxp.com>
-Signed-off-by: Yangbo Lu <yangbo.lu@nxp.com>
----
- Documentation/ABI/stable/sysfs-bus-fsl-mc | 13 +
- Documentation/ioctl/ioctl-number.txt | 1 +
- Documentation/networking/dpaa2/index.rst | 8 +
- Documentation/networking/dpaa2/overview.rst | 408 +++++
- MAINTAINERS | 11 +-
- drivers/bus/Kconfig | 3 +
- drivers/bus/Makefile | 4 +
- drivers/bus/fsl-mc/Kconfig | 23 +
- drivers/bus/fsl-mc/Makefile | 22 +
- drivers/bus/fsl-mc/dpbp.c | 186 +++
- drivers/bus/fsl-mc/dpcon.c | 222 +++
- drivers/bus/fsl-mc/dpmcp.c | 99 ++
- .../fsl-mc/bus => bus/fsl-mc}/dprc-driver.c | 180 ++-
- drivers/bus/fsl-mc/dprc.c | 575 +++++++
- .../bus => bus/fsl-mc}/fsl-mc-allocator.c | 195 ++-
- .../fsl-mc/bus => bus/fsl-mc}/fsl-mc-bus.c | 523 +++++--
- drivers/bus/fsl-mc/fsl-mc-iommu.c | 78 +
- .../fsl-mc/bus => bus/fsl-mc}/fsl-mc-msi.c | 34 +-
- drivers/bus/fsl-mc/fsl-mc-private.h | 223 +++
- drivers/bus/fsl-mc/fsl-mc-restool.c | 219 +++
- .../fsl-mc/bus => bus/fsl-mc}/mc-io.c | 80 +-
- .../fsl-mc/bus => bus/fsl-mc}/mc-sys.c | 105 +-
- drivers/irqchip/Kconfig | 6 +
- drivers/irqchip/Makefile | 1 +
- .../irq-gic-v3-its-fsl-mc-msi.c | 52 +-
- drivers/staging/fsl-mc/Kconfig | 1 +
- drivers/staging/fsl-mc/Makefile | 1 +
- drivers/staging/fsl-mc/TODO | 18 -
- drivers/staging/fsl-mc/bus/Kconfig | 37 +-
- drivers/staging/fsl-mc/bus/Makefile | 17 +-
- drivers/staging/fsl-mc/bus/dpbp.c | 691 --------
- drivers/staging/fsl-mc/bus/dpio/Makefile | 8 +
- drivers/staging/fsl-mc/bus/dpio/dpio-cmd.h | 50 +
- drivers/staging/fsl-mc/bus/dpio/dpio-driver.c | 278 ++++
- .../staging/fsl-mc/bus/dpio/dpio-service.c | 780 +++++++++
- drivers/staging/fsl-mc/bus/dpio/dpio.c | 221 +++
- drivers/staging/fsl-mc/bus/dpio/dpio.h | 87 ++
- .../staging/fsl-mc/bus/dpio/qbman-portal.c | 1164 ++++++++++++++
- .../staging/fsl-mc/bus/dpio/qbman-portal.h | 505 ++++++
- drivers/staging/fsl-mc/bus/dpmcp-cmd.h | 140 --
- drivers/staging/fsl-mc/bus/dpmcp.c | 504 ------
- drivers/staging/fsl-mc/bus/dpmcp.h | 159 --
- drivers/staging/fsl-mc/bus/dpmng-cmd.h | 58 -
- drivers/staging/fsl-mc/bus/dpmng.c | 107 --
- drivers/staging/fsl-mc/bus/dprc-cmd.h | 465 ------
- drivers/staging/fsl-mc/bus/dprc.c | 1388 -----------------
- drivers/staging/fsl-mc/bus/fsl-mc-private.h | 52 -
- drivers/staging/fsl-mc/include/dpaa2-fd.h | 681 ++++++++
- drivers/staging/fsl-mc/include/dpaa2-global.h | 177 +++
- drivers/staging/fsl-mc/include/dpaa2-io.h | 178 +++
- drivers/staging/fsl-mc/include/dpbp-cmd.h | 185 ---
- drivers/staging/fsl-mc/include/dpbp.h | 220 ---
- drivers/staging/fsl-mc/include/dpcon-cmd.h | 62 -
- drivers/staging/fsl-mc/include/dpmng.h | 69 -
- drivers/staging/fsl-mc/include/dpopr.h | 112 ++
- drivers/staging/fsl-mc/include/dprc.h | 544 -------
- drivers/staging/fsl-mc/include/mc-bus.h | 111 --
- drivers/staging/fsl-mc/include/mc-cmd.h | 108 --
- drivers/staging/fsl-mc/include/mc-sys.h | 98 --
- drivers/staging/fsl-mc/include/mc.h | 201 ---
- include/linux/fsl/mc.h | 1025 ++++++++++++
- include/uapi/linux/fsl_mc.h | 31 +
- 62 files changed, 8068 insertions(+), 5736 deletions(-)
- create mode 100644 Documentation/ABI/stable/sysfs-bus-fsl-mc
- create mode 100644 Documentation/networking/dpaa2/index.rst
- create mode 100644 Documentation/networking/dpaa2/overview.rst
- create mode 100644 drivers/bus/fsl-mc/Kconfig
- create mode 100644 drivers/bus/fsl-mc/Makefile
- create mode 100644 drivers/bus/fsl-mc/dpbp.c
- create mode 100644 drivers/bus/fsl-mc/dpcon.c
- create mode 100644 drivers/bus/fsl-mc/dpmcp.c
- rename drivers/{staging/fsl-mc/bus => bus/fsl-mc}/dprc-driver.c (84%)
- create mode 100644 drivers/bus/fsl-mc/dprc.c
- rename drivers/{staging/fsl-mc/bus => bus/fsl-mc}/fsl-mc-allocator.c (71%)
- rename drivers/{staging/fsl-mc/bus => bus/fsl-mc}/fsl-mc-bus.c (64%)
- create mode 100644 drivers/bus/fsl-mc/fsl-mc-iommu.c
- rename drivers/{staging/fsl-mc/bus => bus/fsl-mc}/fsl-mc-msi.c (89%)
- create mode 100644 drivers/bus/fsl-mc/fsl-mc-private.h
- create mode 100644 drivers/bus/fsl-mc/fsl-mc-restool.c
- rename drivers/{staging/fsl-mc/bus => bus/fsl-mc}/mc-io.c (68%)
- rename drivers/{staging/fsl-mc/bus => bus/fsl-mc}/mc-sys.c (66%)
- rename drivers/{staging/fsl-mc/bus => irqchip}/irq-gic-v3-its-fsl-mc-msi.c (60%)
- delete mode 100644 drivers/staging/fsl-mc/TODO
- delete mode 100644 drivers/staging/fsl-mc/bus/dpbp.c
- create mode 100644 drivers/staging/fsl-mc/bus/dpio/Makefile
- create mode 100644 drivers/staging/fsl-mc/bus/dpio/dpio-cmd.h
- create mode 100644 drivers/staging/fsl-mc/bus/dpio/dpio-driver.c
- create mode 100644 drivers/staging/fsl-mc/bus/dpio/dpio-service.c
- create mode 100644 drivers/staging/fsl-mc/bus/dpio/dpio.c
- create mode 100644 drivers/staging/fsl-mc/bus/dpio/dpio.h
- create mode 100644 drivers/staging/fsl-mc/bus/dpio/qbman-portal.c
- create mode 100644 drivers/staging/fsl-mc/bus/dpio/qbman-portal.h
- delete mode 100644 drivers/staging/fsl-mc/bus/dpmcp-cmd.h
- delete mode 100644 drivers/staging/fsl-mc/bus/dpmcp.c
- delete mode 100644 drivers/staging/fsl-mc/bus/dpmcp.h
- delete mode 100644 drivers/staging/fsl-mc/bus/dpmng-cmd.h
- delete mode 100644 drivers/staging/fsl-mc/bus/dpmng.c
- delete mode 100644 drivers/staging/fsl-mc/bus/dprc-cmd.h
- delete mode 100644 drivers/staging/fsl-mc/bus/dprc.c
- delete mode 100644 drivers/staging/fsl-mc/bus/fsl-mc-private.h
- create mode 100644 drivers/staging/fsl-mc/include/dpaa2-fd.h
- create mode 100644 drivers/staging/fsl-mc/include/dpaa2-global.h
- create mode 100644 drivers/staging/fsl-mc/include/dpaa2-io.h
- delete mode 100644 drivers/staging/fsl-mc/include/dpbp-cmd.h
- delete mode 100644 drivers/staging/fsl-mc/include/dpbp.h
- delete mode 100644 drivers/staging/fsl-mc/include/dpcon-cmd.h
- delete mode 100644 drivers/staging/fsl-mc/include/dpmng.h
- create mode 100644 drivers/staging/fsl-mc/include/dpopr.h
- delete mode 100644 drivers/staging/fsl-mc/include/dprc.h
- delete mode 100644 drivers/staging/fsl-mc/include/mc-bus.h
- delete mode 100644 drivers/staging/fsl-mc/include/mc-cmd.h
- delete mode 100644 drivers/staging/fsl-mc/include/mc-sys.h
- delete mode 100644 drivers/staging/fsl-mc/include/mc.h
- create mode 100644 include/linux/fsl/mc.h
- create mode 100644 include/uapi/linux/fsl_mc.h
-
---- /dev/null
-+++ b/Documentation/ABI/stable/sysfs-bus-fsl-mc
-@@ -0,0 +1,13 @@
-+What: /sys/bus/fsl-mc/devices/dprc.*/rescan
-+Date: March. 2018
-+KernelVersion: 4.16
-+Contact: Ioana Ciornei <ioana.ciornei@nxp.com>
-+Description: Root dprc rescan attribute
-+Users: Userspace drivers and management tools
-+
-+What: /sys/bus/fsl-mc/rescan
-+Date: March. 2018
-+KernelVersion: 4.16
-+Contact: Ioana Ciornei <ioana.ciornei@nxp.com>
-+Description: Bus rescan attribute
-+Users: Userspace drivers and management tools
---- a/Documentation/ioctl/ioctl-number.txt
-+++ b/Documentation/ioctl/ioctl-number.txt
-@@ -170,6 +170,7 @@ Code Seq#(hex) Include File Comments
- 'R' 00-1F linux/random.h conflict!
- 'R' 01 linux/rfkill.h conflict!
- 'R' C0-DF net/bluetooth/rfcomm.h
-+'R' E0 uapi/linux/fsl_mc.h
- 'S' all linux/cdrom.h conflict!
- 'S' 80-81 scsi/scsi_ioctl.h conflict!
- 'S' 82-FF scsi/scsi.h conflict!
---- /dev/null
-+++ b/Documentation/networking/dpaa2/index.rst
-@@ -0,0 +1,8 @@
-+===================
-+DPAA2 Documentation
-+===================
-+
-+.. toctree::
-+ :maxdepth: 1
-+
-+ overview
---- /dev/null
-+++ b/Documentation/networking/dpaa2/overview.rst
-@@ -0,0 +1,408 @@
-+.. include:: <isonum.txt>
-+
-+DPAA2 (Data Path Acceleration Architecture Gen2) Overview
-+=========================================================
-+
-+:Copyright: |copy| 2015 Freescale Semiconductor Inc.
-+:Copyright: |copy| 2018 NXP
-+
-+This document provides an overview of the Freescale DPAA2 architecture
-+and how it is integrated into the Linux kernel.
-+
-+Introduction
-+============
-+
-+DPAA2 is a hardware architecture designed for high-speeed network
-+packet processing. DPAA2 consists of sophisticated mechanisms for
-+processing Ethernet packets, queue management, buffer management,
-+autonomous L2 switching, virtual Ethernet bridging, and accelerator
-+(e.g. crypto) sharing.
-+
-+A DPAA2 hardware component called the Management Complex (or MC) manages the
-+DPAA2 hardware resources. The MC provides an object-based abstraction for
-+software drivers to use the DPAA2 hardware.
-+The MC uses DPAA2 hardware resources such as queues, buffer pools, and
-+network ports to create functional objects/devices such as network
-+interfaces, an L2 switch, or accelerator instances.
-+The MC provides memory-mapped I/O command interfaces (MC portals)
-+which DPAA2 software drivers use to operate on DPAA2 objects.
-+
-+The diagram below shows an overview of the DPAA2 resource management
-+architecture::
-+
-+ +--------------------------------------+
-+ | OS |
-+ | DPAA2 drivers |
-+ | | |
-+ +-----------------------------|--------+
-+ |
-+ | (create,discover,connect
-+ | config,use,destroy)
-+ |
-+ DPAA2 |
-+ +------------------------| mc portal |-+
-+ | | |
-+ | +- - - - - - - - - - - - -V- - -+ |
-+ | | | |
-+ | | Management Complex (MC) | |
-+ | | | |
-+ | +- - - - - - - - - - - - - - - -+ |
-+ | |
-+ | Hardware Hardware |
-+ | Resources Objects |
-+ | --------- ------- |
-+ | -queues -DPRC |
-+ | -buffer pools -DPMCP |
-+ | -Eth MACs/ports -DPIO |
-+ | -network interface -DPNI |
-+ | profiles -DPMAC |
-+ | -queue portals -DPBP |
-+ | -MC portals ... |
-+ | ... |
-+ | |
-+ +--------------------------------------+
-+
-+
-+The MC mediates operations such as create, discover,
-+connect, configuration, and destroy. Fast-path operations
-+on data, such as packet transmit/receive, are not mediated by
-+the MC and are done directly using memory mapped regions in
-+DPIO objects.
-+
-+Overview of DPAA2 Objects
-+=========================
-+
-+The section provides a brief overview of some key DPAA2 objects.
-+A simple scenario is described illustrating the objects involved
-+in creating a network interfaces.
-+
-+DPRC (Datapath Resource Container)
-+----------------------------------
-+
-+A DPRC is a container object that holds all the other
-+types of DPAA2 objects. In the example diagram below there
-+are 8 objects of 5 types (DPMCP, DPIO, DPBP, DPNI, and DPMAC)
-+in the container.
-+
-+::
-+
-+ +---------------------------------------------------------+
-+ | DPRC |
-+ | |
-+ | +-------+ +-------+ +-------+ +-------+ +-------+ |
-+ | | DPMCP | | DPIO | | DPBP | | DPNI | | DPMAC | |
-+ | +-------+ +-------+ +-------+ +---+---+ +---+---+ |
-+ | | DPMCP | | DPIO | |
-+ | +-------+ +-------+ |
-+ | | DPMCP | |
-+ | +-------+ |
-+ | |
-+ +---------------------------------------------------------+
-+
-+From the point of view of an OS, a DPRC behaves similar to a plug and
-+play bus, like PCI. DPRC commands can be used to enumerate the contents
-+of the DPRC, discover the hardware objects present (including mappable
-+regions and interrupts).
-+
-+::
-+
-+ DPRC.1 (bus)
-+ |
-+ +--+--------+-------+-------+-------+
-+ | | | | |
-+ DPMCP.1 DPIO.1 DPBP.1 DPNI.1 DPMAC.1
-+ DPMCP.2 DPIO.2
-+ DPMCP.3
-+
-+Hardware objects can be created and destroyed dynamically, providing
-+the ability to hot plug/unplug objects in and out of the DPRC.
-+
-+A DPRC has a mappable MMIO region (an MC portal) that can be used
-+to send MC commands. It has an interrupt for status events (like
-+hotplug).
-+All objects in a container share the same hardware "isolation context".
-+This means that with respect to an IOMMU the isolation granularity
-+is at the DPRC (container) level, not at the individual object
-+level.
-+
-+DPRCs can be defined statically and populated with objects
-+via a config file passed to the MC when firmware starts it.
-+There is also a Linux user space tool called "restool" that can be
-+used to create/destroy containers and objects dynamically. The latest
-+version of restool can be found at:
-+ https://github.com/qoriq-open-source/restool
-+
-+DPAA2 Objects for an Ethernet Network Interface
-+-----------------------------------------------
-+
-+A typical Ethernet NIC is monolithic-- the NIC device contains TX/RX
-+queuing mechanisms, configuration mechanisms, buffer management,
-+physical ports, and interrupts. DPAA2 uses a more granular approach
-+utilizing multiple hardware objects. Each object provides specialized
-+functions. Groups of these objects are used by software to provide
-+Ethernet network interface functionality. This approach provides
-+efficient use of finite hardware resources, flexibility, and
-+performance advantages.
-+
-+The diagram below shows the objects needed for a simple
-+network interface configuration on a system with 2 CPUs.
-+
-+::
-+
-+ +---+---+ +---+---+
-+ CPU0 CPU1
-+ +---+---+ +---+---+
-+ | |
-+ +---+---+ +---+---+
-+ DPIO DPIO
-+ +---+---+ +---+---+
-+ \ /
-+ \ /
-+ \ /
-+ +---+---+
-+ DPNI --- DPBP,DPMCP
-+ +---+---+
-+ |
-+ |
-+ +---+---+
-+ DPMAC
-+ +---+---+
-+ |
-+ port/PHY
-+
-+Below the objects are described. For each object a brief description
-+is provided along with a summary of the kinds of operations the object
-+supports and a summary of key resources of the object (MMIO regions
-+and IRQs).
-+
-+DPMAC (Datapath Ethernet MAC)
-+~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
-+Represents an Ethernet MAC, a hardware device that connects to an Ethernet
-+PHY and allows physical transmission and reception of Ethernet frames.
-+
-+- MMIO regions: none
-+- IRQs: DPNI link change
-+- commands: set link up/down, link config, get stats,
-+ IRQ config, enable, reset
-+
-+DPNI (Datapath Network Interface)
-+Contains TX/RX queues, network interface configuration, and RX buffer pool
-+configuration mechanisms. The TX/RX queues are in memory and are identified
-+by queue number.
-+
-+- MMIO regions: none
-+- IRQs: link state
-+- commands: port config, offload config, queue config,
-+ parse/classify config, IRQ config, enable, reset
-+
-+DPIO (Datapath I/O)
-+~~~~~~~~~~~~~~~~~~~
-+Provides interfaces to enqueue and dequeue
-+packets and do hardware buffer pool management operations. The DPAA2
-+architecture separates the mechanism to access queues (the DPIO object)
-+from the queues themselves. The DPIO provides an MMIO interface to
-+enqueue/dequeue packets. To enqueue something a descriptor is written
-+to the DPIO MMIO region, which includes the target queue number.
-+There will typically be one DPIO assigned to each CPU. This allows all
-+CPUs to simultaneously perform enqueue/dequeued operations. DPIOs are
-+expected to be shared by different DPAA2 drivers.
-+
-+- MMIO regions: queue operations, buffer management
-+- IRQs: data availability, congestion notification, buffer
-+ pool depletion
-+- commands: IRQ config, enable, reset
-+
-+DPBP (Datapath Buffer Pool)
-+~~~~~~~~~~~~~~~~~~~~~~~~~~~
-+Represents a hardware buffer pool.
-+
-+- MMIO regions: none
-+- IRQs: none
-+- commands: enable, reset
-+
-+DPMCP (Datapath MC Portal)
-+~~~~~~~~~~~~~~~~~~~~~~~~~~
-+Provides an MC command portal.
-+Used by drivers to send commands to the MC to manage
-+objects.
-+
-+- MMIO regions: MC command portal
-+- IRQs: command completion
-+- commands: IRQ config, enable, reset
-+
-+Object Connections
-+==================
-+Some objects have explicit relationships that must
-+be configured:
-+
-+- DPNI <--> DPMAC
-+- DPNI <--> DPNI
-+- DPNI <--> L2-switch-port
-+
-+ A DPNI must be connected to something such as a DPMAC,
-+ another DPNI, or L2 switch port. The DPNI connection
-+ is made via a DPRC command.
-+
-+::
-+
-+ +-------+ +-------+
-+ | DPNI | | DPMAC |
-+ +---+---+ +---+---+
-+ | |
-+ +==========+
-+
-+- DPNI <--> DPBP
-+
-+ A network interface requires a 'buffer pool' (DPBP
-+ object) which provides a list of pointers to memory
-+ where received Ethernet data is to be copied. The
-+ Ethernet driver configures the DPBPs associated with
-+ the network interface.
-+
-+Interrupts
-+==========
-+All interrupts generated by DPAA2 objects are message
-+interrupts. At the hardware level message interrupts
-+generated by devices will normally have 3 components--
-+1) a non-spoofable 'device-id' expressed on the hardware
-+bus, 2) an address, 3) a data value.
-+
-+In the case of DPAA2 devices/objects, all objects in the
-+same container/DPRC share the same 'device-id'.
-+For ARM-based SoC this is the same as the stream ID.
-+
-+
-+DPAA2 Linux Drivers Overview
-+============================
-+
-+This section provides an overview of the Linux kernel drivers for
-+DPAA2-- 1) the bus driver and associated "DPAA2 infrastructure"
-+drivers and 2) functional object drivers (such as Ethernet).
-+
-+As described previously, a DPRC is a container that holds the other
-+types of DPAA2 objects. It is functionally similar to a plug-and-play
-+bus controller.
-+Each object in the DPRC is a Linux "device" and is bound to a driver.
-+The diagram below shows the Linux drivers involved in a networking
-+scenario and the objects bound to each driver. A brief description
-+of each driver follows.
-+
-+::
-+
-+ +------------+
-+ | OS Network |
-+ | Stack |
-+ +------------+ +------------+
-+ | Allocator |. . . . . . . | Ethernet |
-+ |(DPMCP,DPBP)| | (DPNI) |
-+ +-.----------+ +---+---+----+
-+ . . ^ |
-+ . . <data avail, | | <enqueue,
-+ . . tx confirm> | | dequeue>
-+ +-------------+ . | |
-+ | DPRC driver | . +---+---V----+ +---------+
-+ | (DPRC) | . . . . . .| DPIO driver| | MAC |
-+ +----------+--+ | (DPIO) | | (DPMAC) |
-+ | +------+-----+ +-----+---+
-+ |<dev add/remove> | |
-+ | | |
-+ +--------+----------+ | +--+---+
-+ | MC-bus driver | | | PHY |
-+ | | | |driver|
-+ | /bus/fsl-mc | | +--+---+
-+ +-------------------+ | |
-+ | |
-+ ========================= HARDWARE =========|=================|======
-+ DPIO |
-+ | |
-+ DPNI---DPBP |
-+ | |
-+ DPMAC |
-+ | |
-+ PHY ---------------+
-+ ============================================|========================
-+
-+A brief description of each driver is provided below.
-+
-+MC-bus driver
-+-------------
-+The MC-bus driver is a platform driver and is probed from a
-+node in the device tree (compatible "fsl,qoriq-mc") passed in by boot
-+firmware. It is responsible for bootstrapping the DPAA2 kernel
-+infrastructure.
-+Key functions include:
-+
-+- registering a new bus type named "fsl-mc" with the kernel,
-+ and implementing bus call-backs (e.g. match/uevent/dev_groups)
-+- implementing APIs for DPAA2 driver registration and for device
-+ add/remove
-+- creates an MSI IRQ domain
-+- doing a 'device add' to expose the 'root' DPRC, in turn triggering
-+ a bind of the root DPRC to the DPRC driver
-+
-+The binding for the MC-bus device-tree node can be consulted at
-+*Documentation/devicetree/bindings/misc/fsl,qoriq-mc.txt*.
-+The sysfs bind/unbind interfaces for the MC-bus can be consulted at
-+*Documentation/ABI/testing/sysfs-bus-fsl-mc*.
-+
-+DPRC driver
-+-----------
-+The DPRC driver is bound to DPRC objects and does runtime management
-+of a bus instance. It performs the initial bus scan of the DPRC
-+and handles interrupts for container events such as hot plug by
-+re-scanning the DPRC.
-+
-+Allocator
-+---------
-+Certain objects such as DPMCP and DPBP are generic and fungible,
-+and are intended to be used by other drivers. For example,
-+the DPAA2 Ethernet driver needs:
-+
-+- DPMCPs to send MC commands, to configure network interfaces
-+- DPBPs for network buffer pools
-+
-+The allocator driver registers for these allocatable object types
-+and those objects are bound to the allocator when the bus is probed.
-+The allocator maintains a pool of objects that are available for
-+allocation by other DPAA2 drivers.
-+
-+DPIO driver
-+-----------
-+The DPIO driver is bound to DPIO objects and provides services that allow
-+other drivers such as the Ethernet driver to enqueue and dequeue data for
-+their respective objects.
-+Key services include:
-+
-+- data availability notifications
-+- hardware queuing operations (enqueue and dequeue of data)
-+- hardware buffer pool management
-+
-+To transmit a packet the Ethernet driver puts data on a queue and
-+invokes a DPIO API. For receive, the Ethernet driver registers
-+a data availability notification callback. To dequeue a packet
-+a DPIO API is used.
-+There is typically one DPIO object per physical CPU for optimum
-+performance, allowing different CPUs to simultaneously enqueue
-+and dequeue data.
-+
-+The DPIO driver operates on behalf of all DPAA2 drivers
-+active in the kernel-- Ethernet, crypto, compression,
-+etc.
-+
-+Ethernet driver
-+---------------
-+The Ethernet driver is bound to a DPNI and implements the kernel
-+interfaces needed to connect the DPAA2 network interface to
-+the network stack.
-+Each DPNI corresponds to a Linux network interface.
-+
-+MAC driver
-+----------
-+An Ethernet PHY is an off-chip, board specific component and is managed
-+by the appropriate PHY driver via an mdio bus. The MAC driver
-+plays a role of being a proxy between the PHY driver and the
-+MC. It does this proxy via the MC commands to a DPMAC object.
-+If the PHY driver signals a link change, the MAC driver notifies
-+the MC via a DPMAC command. If a network interface is brought
-+up or down, the MC notifies the DPMAC driver via an interrupt and
-+the driver can take appropriate action.
---- a/MAINTAINERS
-+++ b/MAINTAINERS
-@@ -3980,6 +3980,12 @@ S: Maintained
- F: drivers/char/dtlk.c
- F: include/linux/dtlk.h
-
-+DPAA2 DATAPATH I/O (DPIO) DRIVER
-+M: Roy Pledge <Roy.Pledge@nxp.com>
-+L: linux-kernel@vger.kernel.org
-+S: Maintained
-+F: drivers/staging/fsl-mc/bus/dpio
-+
- DPT_I2O SCSI RAID DRIVER
- M: Adaptec OEM Raid Solutions <aacraid@adaptec.com>
- L: linux-scsi@vger.kernel.org
-@@ -5110,7 +5116,10 @@ M: "J. German Rivera" <German.Rivera@fre
- M: Stuart Yoder <stuart.yoder@nxp.com>
- L: linux-kernel@vger.kernel.org
- S: Maintained
--F: drivers/staging/fsl-mc/
-+F: drivers/bus/fsl-mc/
-+F: Documentation/networking/dpaa2/overview.rst
-+F: include/uapi/linux/fsl_mc.h
-+F: Documentation/ABI/stable/sysfs-bus-fsl-mc
-
- FREEVXFS FILESYSTEM
- M: Christoph Hellwig <hch@infradead.org>
---- a/drivers/bus/Kconfig
-+++ b/drivers/bus/Kconfig
-@@ -167,4 +167,7 @@ config VEXPRESS_CONFIG
- help
- Platform configuration infrastructure for the ARM Ltd.
- Versatile Express.
-+
-+source "drivers/bus/fsl-mc/Kconfig"
-+
- endmenu
---- a/drivers/bus/Makefile
-+++ b/drivers/bus/Makefile
-@@ -7,6 +7,10 @@ obj-$(CONFIG_ARM_CCI) += arm-cci.o
- obj-$(CONFIG_ARM_CCN) += arm-ccn.o
-
- obj-$(CONFIG_BRCMSTB_GISB_ARB) += brcmstb_gisb.o
-+
-+# DPAA2 fsl-mc bus
-+obj-$(CONFIG_FSL_MC_BUS) += fsl-mc/
-+
- obj-$(CONFIG_IMX_WEIM) += imx-weim.o
- obj-$(CONFIG_MIPS_CDMM) += mips_cdmm.o
- obj-$(CONFIG_MVEBU_MBUS) += mvebu-mbus.o
---- /dev/null
-+++ b/drivers/bus/fsl-mc/Kconfig
-@@ -0,0 +1,23 @@
-+# SPDX-License-Identifier: GPL-2.0
-+#
-+# DPAA2 fsl-mc bus
-+#
-+# Copyright (C) 2014-2016 Freescale Semiconductor, Inc.
-+#
-+
-+config FSL_MC_BUS
-+ bool "QorIQ DPAA2 fsl-mc bus driver"
-+ depends on OF && (ARCH_LAYERSCAPE || (COMPILE_TEST && (ARM || ARM64 || X86 || PPC)))
-+ select GENERIC_MSI_IRQ_DOMAIN
-+ help
-+ Driver to enable the bus infrastructure for the QorIQ DPAA2
-+ architecture. The fsl-mc bus driver handles discovery of
-+ DPAA2 objects (which are represented as Linux devices) and
-+ binding objects to drivers.
-+
-+config FSL_MC_RESTOOL
-+ bool "Management Complex (MC) restool support"
-+ depends on FSL_MC_BUS
-+ help
-+ Provides kernel support for the Management Complex resource
-+ manager user-space tool - restool.
---- /dev/null
-+++ b/drivers/bus/fsl-mc/Makefile
-@@ -0,0 +1,22 @@
-+# SPDX-License-Identifier: GPL-2.0
-+#
-+# Freescale Management Complex (MC) bus drivers
-+#
-+# Copyright (C) 2014 Freescale Semiconductor, Inc.
-+#
-+obj-$(CONFIG_FSL_MC_BUS) += mc-bus-driver.o
-+
-+mc-bus-driver-objs := fsl-mc-bus.o \
-+ mc-sys.o \
-+ mc-io.o \
-+ dpbp.o \
-+ dpcon.o \
-+ dprc.o \
-+ dprc-driver.o \
-+ fsl-mc-allocator.o \
-+ fsl-mc-msi.o \
-+ dpmcp.o \
-+ fsl-mc-iommu.o
-+
-+# MC restool kernel support
-+obj-$(CONFIG_FSL_MC_RESTOOL) += fsl-mc-restool.o
---- /dev/null
-+++ b/drivers/bus/fsl-mc/dpbp.c
-@@ -0,0 +1,186 @@
-+// SPDX-License-Identifier: (GPL-2.0+ OR BSD-3-Clause)
-+/*
-+ * Copyright 2013-2016 Freescale Semiconductor Inc.
-+ *
-+ */
-+#include <linux/kernel.h>
-+#include <linux/fsl/mc.h>
-+#include <linux/fsl/mc.h>
-+
-+#include "fsl-mc-private.h"
-+
-+/**
-+ * dpbp_open() - Open a control session for the specified object.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @dpbp_id: DPBP unique ID
-+ * @token: Returned token; use in subsequent API calls
-+ *
-+ * This function can be used to open a control session for an
-+ * already created object; an object may have been declared in
-+ * the DPL or by calling the dpbp_create function.
-+ * This function returns a unique authentication token,
-+ * associated with the specific object ID and the specific MC
-+ * portal; this token must be used in all subsequent commands for
-+ * this specific object
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpbp_open(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ int dpbp_id,
-+ u16 *token)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpbp_cmd_open *cmd_params;
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPBP_CMDID_OPEN,
-+ cmd_flags, 0);
-+ cmd_params = (struct dpbp_cmd_open *)cmd.params;
-+ cmd_params->dpbp_id = cpu_to_le32(dpbp_id);
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ *token = mc_cmd_hdr_read_token(&cmd);
-+
-+ return err;
-+}
-+EXPORT_SYMBOL_GPL(dpbp_open);
-+
-+/**
-+ * dpbp_close() - Close the control session of the object
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPBP object
-+ *
-+ * After this function is called, no further operations are
-+ * allowed on the object without opening a new control session.
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpbp_close(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPBP_CMDID_CLOSE, cmd_flags,
-+ token);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+EXPORT_SYMBOL_GPL(dpbp_close);
-+
-+/**
-+ * dpbp_enable() - Enable the DPBP.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPBP object
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpbp_enable(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPBP_CMDID_ENABLE, cmd_flags,
-+ token);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+EXPORT_SYMBOL_GPL(dpbp_enable);
-+
-+/**
-+ * dpbp_disable() - Disable the DPBP.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPBP object
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpbp_disable(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPBP_CMDID_DISABLE,
-+ cmd_flags, token);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+EXPORT_SYMBOL_GPL(dpbp_disable);
-+
-+/**
-+ * dpbp_reset() - Reset the DPBP, returns the object to initial state.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPBP object
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpbp_reset(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPBP_CMDID_RESET,
-+ cmd_flags, token);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+EXPORT_SYMBOL_GPL(dpbp_reset);
-+
-+/**
-+ * dpbp_get_attributes - Retrieve DPBP attributes.
-+ *
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPBP object
-+ * @attr: Returned object's attributes
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpbp_get_attributes(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ struct dpbp_attr *attr)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpbp_rsp_get_attributes *rsp_params;
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPBP_CMDID_GET_ATTR,
-+ cmd_flags, token);
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ rsp_params = (struct dpbp_rsp_get_attributes *)cmd.params;
-+ attr->bpid = le16_to_cpu(rsp_params->bpid);
-+ attr->id = le32_to_cpu(rsp_params->id);
-+
-+ return 0;
-+}
-+EXPORT_SYMBOL_GPL(dpbp_get_attributes);
---- /dev/null
-+++ b/drivers/bus/fsl-mc/dpcon.c
-@@ -0,0 +1,222 @@
-+// SPDX-License-Identifier: (GPL-2.0+ OR BSD-3-Clause)
-+/*
-+ * Copyright 2013-2016 Freescale Semiconductor Inc.
-+ *
-+ */
-+#include <linux/kernel.h>
-+#include <linux/fsl/mc.h>
-+#include <linux/fsl/mc.h>
-+
-+#include "fsl-mc-private.h"
-+
-+/**
-+ * dpcon_open() - Open a control session for the specified object
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @dpcon_id: DPCON unique ID
-+ * @token: Returned token; use in subsequent API calls
-+ *
-+ * This function can be used to open a control session for an
-+ * already created object; an object may have been declared in
-+ * the DPL or by calling the dpcon_create() function.
-+ * This function returns a unique authentication token,
-+ * associated with the specific object ID and the specific MC
-+ * portal; this token must be used in all subsequent commands for
-+ * this specific object.
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpcon_open(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ int dpcon_id,
-+ u16 *token)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpcon_cmd_open *dpcon_cmd;
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPCON_CMDID_OPEN,
-+ cmd_flags,
-+ 0);
-+ dpcon_cmd = (struct dpcon_cmd_open *)cmd.params;
-+ dpcon_cmd->dpcon_id = cpu_to_le32(dpcon_id);
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ *token = mc_cmd_hdr_read_token(&cmd);
-+
-+ return 0;
-+}
-+EXPORT_SYMBOL_GPL(dpcon_open);
-+
-+/**
-+ * dpcon_close() - Close the control session of the object
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPCON object
-+ *
-+ * After this function is called, no further operations are
-+ * allowed on the object without opening a new control session.
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpcon_close(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPCON_CMDID_CLOSE,
-+ cmd_flags,
-+ token);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+EXPORT_SYMBOL_GPL(dpcon_close);
-+
-+/**
-+ * dpcon_enable() - Enable the DPCON
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPCON object
-+ *
-+ * Return: '0' on Success; Error code otherwise
-+ */
-+int dpcon_enable(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPCON_CMDID_ENABLE,
-+ cmd_flags,
-+ token);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+EXPORT_SYMBOL_GPL(dpcon_enable);
-+
-+/**
-+ * dpcon_disable() - Disable the DPCON
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPCON object
-+ *
-+ * Return: '0' on Success; Error code otherwise
-+ */
-+int dpcon_disable(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPCON_CMDID_DISABLE,
-+ cmd_flags,
-+ token);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+EXPORT_SYMBOL_GPL(dpcon_disable);
-+
-+/**
-+ * dpcon_reset() - Reset the DPCON, returns the object to initial state.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPCON object
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpcon_reset(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPCON_CMDID_RESET,
-+ cmd_flags, token);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+EXPORT_SYMBOL_GPL(dpcon_reset);
-+
-+/**
-+ * dpcon_get_attributes() - Retrieve DPCON attributes.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPCON object
-+ * @attr: Object's attributes
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpcon_get_attributes(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ struct dpcon_attr *attr)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpcon_rsp_get_attr *dpcon_rsp;
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPCON_CMDID_GET_ATTR,
-+ cmd_flags,
-+ token);
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ dpcon_rsp = (struct dpcon_rsp_get_attr *)cmd.params;
-+ attr->id = le32_to_cpu(dpcon_rsp->id);
-+ attr->qbman_ch_id = le16_to_cpu(dpcon_rsp->qbman_ch_id);
-+ attr->num_priorities = dpcon_rsp->num_priorities;
-+
-+ return 0;
-+}
-+EXPORT_SYMBOL_GPL(dpcon_get_attributes);
-+
-+/**
-+ * dpcon_set_notification() - Set DPCON notification destination
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPCON object
-+ * @cfg: Notification parameters
-+ *
-+ * Return: '0' on Success; Error code otherwise
-+ */
-+int dpcon_set_notification(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ struct dpcon_notification_cfg *cfg)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpcon_cmd_set_notification *dpcon_cmd;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPCON_CMDID_SET_NOTIFICATION,
-+ cmd_flags,
-+ token);
-+ dpcon_cmd = (struct dpcon_cmd_set_notification *)cmd.params;
-+ dpcon_cmd->dpio_id = cpu_to_le32(cfg->dpio_id);
-+ dpcon_cmd->priority = cfg->priority;
-+ dpcon_cmd->user_ctx = cpu_to_le64(cfg->user_ctx);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+EXPORT_SYMBOL_GPL(dpcon_set_notification);
---- /dev/null
-+++ b/drivers/bus/fsl-mc/dpmcp.c
-@@ -0,0 +1,99 @@
-+// SPDX-License-Identifier: (GPL-2.0+ OR BSD-3-Clause)
-+/*
-+ * Copyright 2013-2016 Freescale Semiconductor Inc.
-+ *
-+ */
-+#include <linux/kernel.h>
-+#include <linux/fsl/mc.h>
-+
-+#include "fsl-mc-private.h"
-+
-+/**
-+ * dpmcp_open() - Open a control session for the specified object.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @dpmcp_id: DPMCP unique ID
-+ * @token: Returned token; use in subsequent API calls
-+ *
-+ * This function can be used to open a control session for an
-+ * already created object; an object may have been declared in
-+ * the DPL or by calling the dpmcp_create function.
-+ * This function returns a unique authentication token,
-+ * associated with the specific object ID and the specific MC
-+ * portal; this token must be used in all subsequent commands for
-+ * this specific object
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpmcp_open(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ int dpmcp_id,
-+ u16 *token)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpmcp_cmd_open *cmd_params;
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPMCP_CMDID_OPEN,
-+ cmd_flags, 0);
-+ cmd_params = (struct dpmcp_cmd_open *)cmd.params;
-+ cmd_params->dpmcp_id = cpu_to_le32(dpmcp_id);
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ *token = mc_cmd_hdr_read_token(&cmd);
-+
-+ return err;
-+}
-+
-+/**
-+ * dpmcp_close() - Close the control session of the object
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPMCP object
-+ *
-+ * After this function is called, no further operations are
-+ * allowed on the object without opening a new control session.
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpmcp_close(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPMCP_CMDID_CLOSE,
-+ cmd_flags, token);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpmcp_reset() - Reset the DPMCP, returns the object to initial state.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPMCP object
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpmcp_reset(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPMCP_CMDID_RESET,
-+ cmd_flags, token);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
---- a/drivers/staging/fsl-mc/bus/dprc-driver.c
-+++ /dev/null
-@@ -1,805 +0,0 @@
--/*
-- * Freescale data path resource container (DPRC) driver
-- *
-- * Copyright (C) 2014 Freescale Semiconductor, Inc.
-- * Author: German Rivera <German.Rivera@freescale.com>
-- *
-- * This file is licensed under the terms of the GNU General Public
-- * License version 2. This program is licensed "as is" without any
-- * warranty of any kind, whether express or implied.
-- */
--
--#include <linux/module.h>
--#include <linux/slab.h>
--#include <linux/interrupt.h>
--#include <linux/msi.h>
--#include "../include/mc-bus.h"
--#include "../include/mc-sys.h"
--
--#include "dprc-cmd.h"
--#include "fsl-mc-private.h"
--
--#define FSL_MC_DPRC_DRIVER_NAME "fsl_mc_dprc"
--
--#define FSL_MC_DEVICE_MATCH(_mc_dev, _obj_desc) \
-- (strcmp((_mc_dev)->obj_desc.type, (_obj_desc)->type) == 0 && \
-- (_mc_dev)->obj_desc.id == (_obj_desc)->id)
--
--struct dprc_child_objs {
-- int child_count;
-- struct dprc_obj_desc *child_array;
--};
--
--static int __fsl_mc_device_remove_if_not_in_mc(struct device *dev, void *data)
--{
-- int i;
-- struct dprc_child_objs *objs;
-- struct fsl_mc_device *mc_dev;
--
-- WARN_ON(!dev);
-- WARN_ON(!data);
-- mc_dev = to_fsl_mc_device(dev);
-- objs = data;
--
-- for (i = 0; i < objs->child_count; i++) {
-- struct dprc_obj_desc *obj_desc = &objs->child_array[i];
--
-- if (strlen(obj_desc->type) != 0 &&
-- FSL_MC_DEVICE_MATCH(mc_dev, obj_desc))
-- break;
-- }
--
-- if (i == objs->child_count)
-- fsl_mc_device_remove(mc_dev);
--
-- return 0;
--}
--
--static int __fsl_mc_device_remove(struct device *dev, void *data)
--{
-- WARN_ON(!dev);
-- WARN_ON(data);
-- fsl_mc_device_remove(to_fsl_mc_device(dev));
-- return 0;
--}
--
--/**
-- * dprc_remove_devices - Removes devices for objects removed from a DPRC
-- *
-- * @mc_bus_dev: pointer to the fsl-mc device that represents a DPRC object
-- * @obj_desc_array: array of object descriptors for child objects currently
-- * present in the DPRC in the MC.
-- * @num_child_objects_in_mc: number of entries in obj_desc_array
-- *
-- * Synchronizes the state of the Linux bus driver with the actual state of
-- * the MC by removing devices that represent MC objects that have
-- * been dynamically removed in the physical DPRC.
-- */
--static void dprc_remove_devices(struct fsl_mc_device *mc_bus_dev,
-- struct dprc_obj_desc *obj_desc_array,
-- int num_child_objects_in_mc)
--{
-- if (num_child_objects_in_mc != 0) {
-- /*
-- * Remove child objects that are in the DPRC in Linux,
-- * but not in the MC:
-- */
-- struct dprc_child_objs objs;
--
-- objs.child_count = num_child_objects_in_mc;
-- objs.child_array = obj_desc_array;
-- device_for_each_child(&mc_bus_dev->dev, &objs,
-- __fsl_mc_device_remove_if_not_in_mc);
-- } else {
-- /*
-- * There are no child objects for this DPRC in the MC.
-- * So, remove all the child devices from Linux:
-- */
-- device_for_each_child(&mc_bus_dev->dev, NULL,
-- __fsl_mc_device_remove);
-- }
--}
--
--static int __fsl_mc_device_match(struct device *dev, void *data)
--{
-- struct dprc_obj_desc *obj_desc = data;
-- struct fsl_mc_device *mc_dev = to_fsl_mc_device(dev);
--
-- return FSL_MC_DEVICE_MATCH(mc_dev, obj_desc);
--}
--
--static struct fsl_mc_device *fsl_mc_device_lookup(struct dprc_obj_desc
-- *obj_desc,
-- struct fsl_mc_device
-- *mc_bus_dev)
--{
-- struct device *dev;
--
-- dev = device_find_child(&mc_bus_dev->dev, obj_desc,
-- __fsl_mc_device_match);
--
-- return dev ? to_fsl_mc_device(dev) : NULL;
--}
--
--/**
-- * check_plugged_state_change - Check change in an MC object's plugged state
-- *
-- * @mc_dev: pointer to the fsl-mc device for a given MC object
-- * @obj_desc: pointer to the MC object's descriptor in the MC
-- *
-- * If the plugged state has changed from unplugged to plugged, the fsl-mc
-- * device is bound to the corresponding device driver.
-- * If the plugged state has changed from plugged to unplugged, the fsl-mc
-- * device is unbound from the corresponding device driver.
-- */
--static void check_plugged_state_change(struct fsl_mc_device *mc_dev,
-- struct dprc_obj_desc *obj_desc)
--{
-- int error;
-- u32 plugged_flag_at_mc =
-- obj_desc->state & DPRC_OBJ_STATE_PLUGGED;
--
-- if (plugged_flag_at_mc !=
-- (mc_dev->obj_desc.state & DPRC_OBJ_STATE_PLUGGED)) {
-- if (plugged_flag_at_mc) {
-- mc_dev->obj_desc.state |= DPRC_OBJ_STATE_PLUGGED;
-- error = device_attach(&mc_dev->dev);
-- if (error < 0) {
-- dev_err(&mc_dev->dev,
-- "device_attach() failed: %d\n",
-- error);
-- }
-- } else {
-- mc_dev->obj_desc.state &= ~DPRC_OBJ_STATE_PLUGGED;
-- device_release_driver(&mc_dev->dev);
-- }
-- }
--}
--
--/**
-- * dprc_add_new_devices - Adds devices to the logical bus for a DPRC
-- *
-- * @mc_bus_dev: pointer to the fsl-mc device that represents a DPRC object
-- * @obj_desc_array: array of device descriptors for child devices currently
-- * present in the physical DPRC.
-- * @num_child_objects_in_mc: number of entries in obj_desc_array
-- *
-- * Synchronizes the state of the Linux bus driver with the actual
-- * state of the MC by adding objects that have been newly discovered
-- * in the physical DPRC.
-- */
--static void dprc_add_new_devices(struct fsl_mc_device *mc_bus_dev,
-- struct dprc_obj_desc *obj_desc_array,
-- int num_child_objects_in_mc)
--{
-- int error;
-- int i;
--
-- for (i = 0; i < num_child_objects_in_mc; i++) {
-- struct fsl_mc_device *child_dev;
-- struct dprc_obj_desc *obj_desc = &obj_desc_array[i];
--
-- if (strlen(obj_desc->type) == 0)
-- continue;
--
-- /*
-- * Check if device is already known to Linux:
-- */
-- child_dev = fsl_mc_device_lookup(obj_desc, mc_bus_dev);
-- if (child_dev) {
-- check_plugged_state_change(child_dev, obj_desc);
-- continue;
-- }
--
-- error = fsl_mc_device_add(obj_desc, NULL, &mc_bus_dev->dev,
-- &child_dev);
-- if (error < 0)
-- continue;
-- }
--}
--
--/**
-- * dprc_scan_objects - Discover objects in a DPRC
-- *
-- * @mc_bus_dev: pointer to the fsl-mc device that represents a DPRC object
-- * @total_irq_count: total number of IRQs needed by objects in the DPRC.
-- *
-- * Detects objects added and removed from a DPRC and synchronizes the
-- * state of the Linux bus driver, MC by adding and removing
-- * devices accordingly.
-- * Two types of devices can be found in a DPRC: allocatable objects (e.g.,
-- * dpbp, dpmcp) and non-allocatable devices (e.g., dprc, dpni).
-- * All allocatable devices needed to be probed before all non-allocatable
-- * devices, to ensure that device drivers for non-allocatable
-- * devices can allocate any type of allocatable devices.
-- * That is, we need to ensure that the corresponding resource pools are
-- * populated before they can get allocation requests from probe callbacks
-- * of the device drivers for the non-allocatable devices.
-- */
--int dprc_scan_objects(struct fsl_mc_device *mc_bus_dev,
-- unsigned int *total_irq_count)
--{
-- int num_child_objects;
-- int dprc_get_obj_failures;
-- int error;
-- unsigned int irq_count = mc_bus_dev->obj_desc.irq_count;
-- struct dprc_obj_desc *child_obj_desc_array = NULL;
--
-- error = dprc_get_obj_count(mc_bus_dev->mc_io,
-- 0,
-- mc_bus_dev->mc_handle,
-- &num_child_objects);
-- if (error < 0) {
-- dev_err(&mc_bus_dev->dev, "dprc_get_obj_count() failed: %d\n",
-- error);
-- return error;
-- }
--
-- if (num_child_objects != 0) {
-- int i;
--
-- child_obj_desc_array =
-- devm_kmalloc_array(&mc_bus_dev->dev, num_child_objects,
-- sizeof(*child_obj_desc_array),
-- GFP_KERNEL);
-- if (!child_obj_desc_array)
-- return -ENOMEM;
--
-- /*
-- * Discover objects currently present in the physical DPRC:
-- */
-- dprc_get_obj_failures = 0;
-- for (i = 0; i < num_child_objects; i++) {
-- struct dprc_obj_desc *obj_desc =
-- &child_obj_desc_array[i];
--
-- error = dprc_get_obj(mc_bus_dev->mc_io,
-- 0,
-- mc_bus_dev->mc_handle,
-- i, obj_desc);
-- if (error < 0) {
-- dev_err(&mc_bus_dev->dev,
-- "dprc_get_obj(i=%d) failed: %d\n",
-- i, error);
-- /*
-- * Mark the obj entry as "invalid", by using the
-- * empty string as obj type:
-- */
-- obj_desc->type[0] = '\0';
-- obj_desc->id = error;
-- dprc_get_obj_failures++;
-- continue;
-- }
--
-- /*
-- * add a quirk for all versions of dpsec < 4.0...none
-- * are coherent regardless of what the MC reports.
-- */
-- if ((strcmp(obj_desc->type, "dpseci") == 0) &&
-- (obj_desc->ver_major < 4))
-- obj_desc->flags |=
-- DPRC_OBJ_FLAG_NO_MEM_SHAREABILITY;
--
-- irq_count += obj_desc->irq_count;
-- dev_dbg(&mc_bus_dev->dev,
-- "Discovered object: type %s, id %d\n",
-- obj_desc->type, obj_desc->id);
-- }
--
-- if (dprc_get_obj_failures != 0) {
-- dev_err(&mc_bus_dev->dev,
-- "%d out of %d devices could not be retrieved\n",
-- dprc_get_obj_failures, num_child_objects);
-- }
-- }
--
-- *total_irq_count = irq_count;
-- dprc_remove_devices(mc_bus_dev, child_obj_desc_array,
-- num_child_objects);
--
-- dprc_add_new_devices(mc_bus_dev, child_obj_desc_array,
-- num_child_objects);
--
-- if (child_obj_desc_array)
-- devm_kfree(&mc_bus_dev->dev, child_obj_desc_array);
--
-- return 0;
--}
--EXPORT_SYMBOL_GPL(dprc_scan_objects);
--
--/**
-- * dprc_scan_container - Scans a physical DPRC and synchronizes Linux bus state
-- *
-- * @mc_bus_dev: pointer to the fsl-mc device that represents a DPRC object
-- *
-- * Scans the physical DPRC and synchronizes the state of the Linux
-- * bus driver with the actual state of the MC by adding and removing
-- * devices as appropriate.
-- */
--int dprc_scan_container(struct fsl_mc_device *mc_bus_dev)
--{
-- int error;
-- unsigned int irq_count;
-- struct fsl_mc_bus *mc_bus = to_fsl_mc_bus(mc_bus_dev);
--
-- fsl_mc_init_all_resource_pools(mc_bus_dev);
--
-- /*
-- * Discover objects in the DPRC:
-- */
-- mutex_lock(&mc_bus->scan_mutex);
-- error = dprc_scan_objects(mc_bus_dev, &irq_count);
-- mutex_unlock(&mc_bus->scan_mutex);
-- if (error < 0)
-- goto error;
--
-- if (dev_get_msi_domain(&mc_bus_dev->dev) && !mc_bus->irq_resources) {
-- if (irq_count > FSL_MC_IRQ_POOL_MAX_TOTAL_IRQS) {
-- dev_warn(&mc_bus_dev->dev,
-- "IRQs needed (%u) exceed IRQs preallocated (%u)\n",
-- irq_count, FSL_MC_IRQ_POOL_MAX_TOTAL_IRQS);
-- }
--
-- error = fsl_mc_populate_irq_pool(
-- mc_bus,
-- FSL_MC_IRQ_POOL_MAX_TOTAL_IRQS);
-- if (error < 0)
-- goto error;
-- }
--
-- return 0;
--error:
-- fsl_mc_cleanup_all_resource_pools(mc_bus_dev);
-- return error;
--}
--EXPORT_SYMBOL_GPL(dprc_scan_container);
--
--/**
-- * dprc_irq0_handler - Regular ISR for DPRC interrupt 0
-- *
-- * @irq: IRQ number of the interrupt being handled
-- * @arg: Pointer to device structure
-- */
--static irqreturn_t dprc_irq0_handler(int irq_num, void *arg)
--{
-- return IRQ_WAKE_THREAD;
--}
--
--/**
-- * dprc_irq0_handler_thread - Handler thread function for DPRC interrupt 0
-- *
-- * @irq: IRQ number of the interrupt being handled
-- * @arg: Pointer to device structure
-- */
--static irqreturn_t dprc_irq0_handler_thread(int irq_num, void *arg)
--{
-- int error;
-- u32 status;
-- struct device *dev = arg;
-- struct fsl_mc_device *mc_dev = to_fsl_mc_device(dev);
-- struct fsl_mc_bus *mc_bus = to_fsl_mc_bus(mc_dev);
-- struct fsl_mc_io *mc_io = mc_dev->mc_io;
-- struct msi_desc *msi_desc = mc_dev->irqs[0]->msi_desc;
--
-- dev_dbg(dev, "DPRC IRQ %d triggered on CPU %u\n",
-- irq_num, smp_processor_id());
--
-- if (WARN_ON(!(mc_dev->flags & FSL_MC_IS_DPRC)))
-- return IRQ_HANDLED;
--
-- mutex_lock(&mc_bus->scan_mutex);
-- if (WARN_ON(!msi_desc || msi_desc->irq != (u32)irq_num))
-- goto out;
--
-- status = 0;
-- error = dprc_get_irq_status(mc_io, 0, mc_dev->mc_handle, 0,
-- &status);
-- if (error < 0) {
-- dev_err(dev,
-- "dprc_get_irq_status() failed: %d\n", error);
-- goto out;
-- }
--
-- error = dprc_clear_irq_status(mc_io, 0, mc_dev->mc_handle, 0,
-- status);
-- if (error < 0) {
-- dev_err(dev,
-- "dprc_clear_irq_status() failed: %d\n", error);
-- goto out;
-- }
--
-- if (status & (DPRC_IRQ_EVENT_OBJ_ADDED |
-- DPRC_IRQ_EVENT_OBJ_REMOVED |
-- DPRC_IRQ_EVENT_CONTAINER_DESTROYED |
-- DPRC_IRQ_EVENT_OBJ_DESTROYED |
-- DPRC_IRQ_EVENT_OBJ_CREATED)) {
-- unsigned int irq_count;
--
-- error = dprc_scan_objects(mc_dev, &irq_count);
-- if (error < 0) {
-- /*
-- * If the error is -ENXIO, we ignore it, as it indicates
-- * that the object scan was aborted, as we detected that
-- * an object was removed from the DPRC in the MC, while
-- * we were scanning the DPRC.
-- */
-- if (error != -ENXIO) {
-- dev_err(dev, "dprc_scan_objects() failed: %d\n",
-- error);
-- }
--
-- goto out;
-- }
--
-- if (irq_count > FSL_MC_IRQ_POOL_MAX_TOTAL_IRQS) {
-- dev_warn(dev,
-- "IRQs needed (%u) exceed IRQs preallocated (%u)\n",
-- irq_count, FSL_MC_IRQ_POOL_MAX_TOTAL_IRQS);
-- }
-- }
--
--out:
-- mutex_unlock(&mc_bus->scan_mutex);
-- return IRQ_HANDLED;
--}
--
--/*
-- * Disable and clear interrupt for a given DPRC object
-- */
--static int disable_dprc_irq(struct fsl_mc_device *mc_dev)
--{
-- int error;
-- struct fsl_mc_io *mc_io = mc_dev->mc_io;
--
-- WARN_ON(mc_dev->obj_desc.irq_count != 1);
--
-- /*
-- * Disable generation of interrupt, while we configure it:
-- */
-- error = dprc_set_irq_enable(mc_io, 0, mc_dev->mc_handle, 0, 0);
-- if (error < 0) {
-- dev_err(&mc_dev->dev,
-- "Disabling DPRC IRQ failed: dprc_set_irq_enable() failed: %d\n",
-- error);
-- return error;
-- }
--
-- /*
-- * Disable all interrupt causes for the interrupt:
-- */
-- error = dprc_set_irq_mask(mc_io, 0, mc_dev->mc_handle, 0, 0x0);
-- if (error < 0) {
-- dev_err(&mc_dev->dev,
-- "Disabling DPRC IRQ failed: dprc_set_irq_mask() failed: %d\n",
-- error);
-- return error;
-- }
--
-- /*
-- * Clear any leftover interrupts:
-- */
-- error = dprc_clear_irq_status(mc_io, 0, mc_dev->mc_handle, 0, ~0x0U);
-- if (error < 0) {
-- dev_err(&mc_dev->dev,
-- "Disabling DPRC IRQ failed: dprc_clear_irq_status() failed: %d\n",
-- error);
-- return error;
-- }
--
-- return 0;
--}
--
--static int register_dprc_irq_handler(struct fsl_mc_device *mc_dev)
--{
-- int error;
-- struct fsl_mc_device_irq *irq = mc_dev->irqs[0];
--
-- WARN_ON(mc_dev->obj_desc.irq_count != 1);
--
-- /*
-- * NOTE: devm_request_threaded_irq() invokes the device-specific
-- * function that programs the MSI physically in the device
-- */
-- error = devm_request_threaded_irq(&mc_dev->dev,
-- irq->msi_desc->irq,
-- dprc_irq0_handler,
-- dprc_irq0_handler_thread,
-- IRQF_NO_SUSPEND | IRQF_ONESHOT,
-- "FSL MC DPRC irq0",
-- &mc_dev->dev);
-- if (error < 0) {
-- dev_err(&mc_dev->dev,
-- "devm_request_threaded_irq() failed: %d\n",
-- error);
-- return error;
-- }
--
-- return 0;
--}
--
--static int enable_dprc_irq(struct fsl_mc_device *mc_dev)
--{
-- int error;
--
-- /*
-- * Enable all interrupt causes for the interrupt:
-- */
-- error = dprc_set_irq_mask(mc_dev->mc_io, 0, mc_dev->mc_handle, 0,
-- ~0x0u);
-- if (error < 0) {
-- dev_err(&mc_dev->dev,
-- "Enabling DPRC IRQ failed: dprc_set_irq_mask() failed: %d\n",
-- error);
--
-- return error;
-- }
--
-- /*
-- * Enable generation of the interrupt:
-- */
-- error = dprc_set_irq_enable(mc_dev->mc_io, 0, mc_dev->mc_handle, 0, 1);
-- if (error < 0) {
-- dev_err(&mc_dev->dev,
-- "Enabling DPRC IRQ failed: dprc_set_irq_enable() failed: %d\n",
-- error);
--
-- return error;
-- }
--
-- return 0;
--}
--
--/*
-- * Setup interrupt for a given DPRC device
-- */
--static int dprc_setup_irq(struct fsl_mc_device *mc_dev)
--{
-- int error;
--
-- error = fsl_mc_allocate_irqs(mc_dev);
-- if (error < 0)
-- return error;
--
-- error = disable_dprc_irq(mc_dev);
-- if (error < 0)
-- goto error_free_irqs;
--
-- error = register_dprc_irq_handler(mc_dev);
-- if (error < 0)
-- goto error_free_irqs;
--
-- error = enable_dprc_irq(mc_dev);
-- if (error < 0)
-- goto error_free_irqs;
--
-- return 0;
--
--error_free_irqs:
-- fsl_mc_free_irqs(mc_dev);
-- return error;
--}
--
--/**
-- * dprc_probe - callback invoked when a DPRC is being bound to this driver
-- *
-- * @mc_dev: Pointer to fsl-mc device representing a DPRC
-- *
-- * It opens the physical DPRC in the MC.
-- * It scans the DPRC to discover the MC objects contained in it.
-- * It creates the interrupt pool for the MC bus associated with the DPRC.
-- * It configures the interrupts for the DPRC device itself.
-- */
--static int dprc_probe(struct fsl_mc_device *mc_dev)
--{
-- int error;
-- size_t region_size;
-- struct device *parent_dev = mc_dev->dev.parent;
-- struct fsl_mc_bus *mc_bus = to_fsl_mc_bus(mc_dev);
-- bool mc_io_created = false;
-- bool msi_domain_set = false;
--
-- if (WARN_ON(strcmp(mc_dev->obj_desc.type, "dprc") != 0))
-- return -EINVAL;
--
-- if (WARN_ON(dev_get_msi_domain(&mc_dev->dev)))
-- return -EINVAL;
--
-- if (!mc_dev->mc_io) {
-- /*
-- * This is a child DPRC:
-- */
-- if (WARN_ON(!dev_is_fsl_mc(parent_dev)))
-- return -EINVAL;
--
-- if (WARN_ON(mc_dev->obj_desc.region_count == 0))
-- return -EINVAL;
--
-- region_size = mc_dev->regions[0].end -
-- mc_dev->regions[0].start + 1;
--
-- error = fsl_create_mc_io(&mc_dev->dev,
-- mc_dev->regions[0].start,
-- region_size,
-- NULL,
-- FSL_MC_IO_ATOMIC_CONTEXT_PORTAL,
-- &mc_dev->mc_io);
-- if (error < 0)
-- return error;
--
-- mc_io_created = true;
--
-- /*
-- * Inherit parent MSI domain:
-- */
-- dev_set_msi_domain(&mc_dev->dev,
-- dev_get_msi_domain(parent_dev));
-- msi_domain_set = true;
-- } else {
-- /*
-- * This is a root DPRC
-- */
-- struct irq_domain *mc_msi_domain;
--
-- if (WARN_ON(dev_is_fsl_mc(parent_dev)))
-- return -EINVAL;
--
-- error = fsl_mc_find_msi_domain(parent_dev,
-- &mc_msi_domain);
-- if (error < 0) {
-- dev_warn(&mc_dev->dev,
-- "WARNING: MC bus without interrupt support\n");
-- } else {
-- dev_set_msi_domain(&mc_dev->dev, mc_msi_domain);
-- msi_domain_set = true;
-- }
-- }
--
-- error = dprc_open(mc_dev->mc_io, 0, mc_dev->obj_desc.id,
-- &mc_dev->mc_handle);
-- if (error < 0) {
-- dev_err(&mc_dev->dev, "dprc_open() failed: %d\n", error);
-- goto error_cleanup_msi_domain;
-- }
--
-- error = dprc_get_attributes(mc_dev->mc_io, 0, mc_dev->mc_handle,
-- &mc_bus->dprc_attr);
-- if (error < 0) {
-- dev_err(&mc_dev->dev, "dprc_get_attributes() failed: %d\n",
-- error);
-- goto error_cleanup_open;
-- }
--
-- if (mc_bus->dprc_attr.version.major < DPRC_MIN_VER_MAJOR ||
-- (mc_bus->dprc_attr.version.major == DPRC_MIN_VER_MAJOR &&
-- mc_bus->dprc_attr.version.minor < DPRC_MIN_VER_MINOR)) {
-- dev_err(&mc_dev->dev,
-- "ERROR: DPRC version %d.%d not supported\n",
-- mc_bus->dprc_attr.version.major,
-- mc_bus->dprc_attr.version.minor);
-- error = -ENOTSUPP;
-- goto error_cleanup_open;
-- }
--
-- mutex_init(&mc_bus->scan_mutex);
--
-- /*
-- * Discover MC objects in DPRC object:
-- */
-- error = dprc_scan_container(mc_dev);
-- if (error < 0)
-- goto error_cleanup_open;
--
-- /*
-- * Configure interrupt for the DPRC object associated with this MC bus:
-- */
-- error = dprc_setup_irq(mc_dev);
-- if (error < 0)
-- goto error_cleanup_open;
--
-- dev_info(&mc_dev->dev, "DPRC device bound to driver");
-- return 0;
--
--error_cleanup_open:
-- (void)dprc_close(mc_dev->mc_io, 0, mc_dev->mc_handle);
--
--error_cleanup_msi_domain:
-- if (msi_domain_set)
-- dev_set_msi_domain(&mc_dev->dev, NULL);
--
-- if (mc_io_created) {
-- fsl_destroy_mc_io(mc_dev->mc_io);
-- mc_dev->mc_io = NULL;
-- }
--
-- return error;
--}
--
--/*
-- * Tear down interrupt for a given DPRC object
-- */
--static void dprc_teardown_irq(struct fsl_mc_device *mc_dev)
--{
-- struct fsl_mc_device_irq *irq = mc_dev->irqs[0];
--
-- (void)disable_dprc_irq(mc_dev);
--
-- devm_free_irq(&mc_dev->dev, irq->msi_desc->irq, &mc_dev->dev);
--
-- fsl_mc_free_irqs(mc_dev);
--}
--
--/**
-- * dprc_remove - callback invoked when a DPRC is being unbound from this driver
-- *
-- * @mc_dev: Pointer to fsl-mc device representing the DPRC
-- *
-- * It removes the DPRC's child objects from Linux (not from the MC) and
-- * closes the DPRC device in the MC.
-- * It tears down the interrupts that were configured for the DPRC device.
-- * It destroys the interrupt pool associated with this MC bus.
-- */
--static int dprc_remove(struct fsl_mc_device *mc_dev)
--{
-- int error;
-- struct fsl_mc_bus *mc_bus = to_fsl_mc_bus(mc_dev);
--
-- if (WARN_ON(strcmp(mc_dev->obj_desc.type, "dprc") != 0))
-- return -EINVAL;
-- if (WARN_ON(!mc_dev->mc_io))
-- return -EINVAL;
--
-- if (WARN_ON(!mc_bus->irq_resources))
-- return -EINVAL;
--
-- if (dev_get_msi_domain(&mc_dev->dev))
-- dprc_teardown_irq(mc_dev);
--
-- device_for_each_child(&mc_dev->dev, NULL, __fsl_mc_device_remove);
--
-- if (dev_get_msi_domain(&mc_dev->dev)) {
-- fsl_mc_cleanup_irq_pool(mc_bus);
-- dev_set_msi_domain(&mc_dev->dev, NULL);
-- }
--
-- fsl_mc_cleanup_all_resource_pools(mc_dev);
--
-- error = dprc_close(mc_dev->mc_io, 0, mc_dev->mc_handle);
-- if (error < 0)
-- dev_err(&mc_dev->dev, "dprc_close() failed: %d\n", error);
--
-- if (!fsl_mc_is_root_dprc(&mc_dev->dev)) {
-- fsl_destroy_mc_io(mc_dev->mc_io);
-- mc_dev->mc_io = NULL;
-- }
--
-- dev_info(&mc_dev->dev, "DPRC device unbound from driver");
-- return 0;
--}
--
--static const struct fsl_mc_device_id match_id_table[] = {
-- {
-- .vendor = FSL_MC_VENDOR_FREESCALE,
-- .obj_type = "dprc"},
-- {.vendor = 0x0},
--};
--
--static struct fsl_mc_driver dprc_driver = {
-- .driver = {
-- .name = FSL_MC_DPRC_DRIVER_NAME,
-- .owner = THIS_MODULE,
-- .pm = NULL,
-- },
-- .match_id_table = match_id_table,
-- .probe = dprc_probe,
-- .remove = dprc_remove,
--};
--
--int __init dprc_driver_init(void)
--{
-- return fsl_mc_driver_register(&dprc_driver);
--}
--
--void dprc_driver_exit(void)
--{
-- fsl_mc_driver_unregister(&dprc_driver);
--}
---- /dev/null
-+++ b/drivers/bus/fsl-mc/dprc-driver.c
-@@ -0,0 +1,815 @@
-+// SPDX-License-Identifier: GPL-2.0
-+/*
-+ * Freescale data path resource container (DPRC) driver
-+ *
-+ * Copyright (C) 2014-2016 Freescale Semiconductor, Inc.
-+ * Author: German Rivera <German.Rivera@freescale.com>
-+ *
-+ */
-+
-+#include <linux/module.h>
-+#include <linux/slab.h>
-+#include <linux/interrupt.h>
-+#include <linux/msi.h>
-+#include <linux/fsl/mc.h>
-+
-+#include "fsl-mc-private.h"
-+
-+#define FSL_MC_DPRC_DRIVER_NAME "fsl_mc_dprc"
-+
-+struct fsl_mc_child_objs {
-+ int child_count;
-+ struct fsl_mc_obj_desc *child_array;
-+};
-+
-+static bool fsl_mc_device_match(struct fsl_mc_device *mc_dev,
-+ struct fsl_mc_obj_desc *obj_desc)
-+{
-+ return mc_dev->obj_desc.id == obj_desc->id &&
-+ strcmp(mc_dev->obj_desc.type, obj_desc->type) == 0;
-+
-+}
-+
-+static int __fsl_mc_device_remove_if_not_in_mc(struct device *dev, void *data)
-+{
-+ int i;
-+ struct fsl_mc_child_objs *objs;
-+ struct fsl_mc_device *mc_dev;
-+
-+ mc_dev = to_fsl_mc_device(dev);
-+ objs = data;
-+
-+ for (i = 0; i < objs->child_count; i++) {
-+ struct fsl_mc_obj_desc *obj_desc = &objs->child_array[i];
-+
-+ if (strlen(obj_desc->type) != 0 &&
-+ fsl_mc_device_match(mc_dev, obj_desc))
-+ break;
-+ }
-+
-+ if (i == objs->child_count)
-+ fsl_mc_device_remove(mc_dev);
-+
-+ return 0;
-+}
-+
-+static int __fsl_mc_device_remove(struct device *dev, void *data)
-+{
-+ fsl_mc_device_remove(to_fsl_mc_device(dev));
-+ return 0;
-+}
-+
-+/**
-+ * dprc_remove_devices - Removes devices for objects removed from a DPRC
-+ *
-+ * @mc_bus_dev: pointer to the fsl-mc device that represents a DPRC object
-+ * @obj_desc_array: array of object descriptors for child objects currently
-+ * present in the DPRC in the MC.
-+ * @num_child_objects_in_mc: number of entries in obj_desc_array
-+ *
-+ * Synchronizes the state of the Linux bus driver with the actual state of
-+ * the MC by removing devices that represent MC objects that have
-+ * been dynamically removed in the physical DPRC.
-+ */
-+static void dprc_remove_devices(struct fsl_mc_device *mc_bus_dev,
-+ struct fsl_mc_obj_desc *obj_desc_array,
-+ int num_child_objects_in_mc)
-+{
-+ if (num_child_objects_in_mc != 0) {
-+ /*
-+ * Remove child objects that are in the DPRC in Linux,
-+ * but not in the MC:
-+ */
-+ struct fsl_mc_child_objs objs;
-+
-+ objs.child_count = num_child_objects_in_mc;
-+ objs.child_array = obj_desc_array;
-+ device_for_each_child(&mc_bus_dev->dev, &objs,
-+ __fsl_mc_device_remove_if_not_in_mc);
-+ } else {
-+ /*
-+ * There are no child objects for this DPRC in the MC.
-+ * So, remove all the child devices from Linux:
-+ */
-+ device_for_each_child(&mc_bus_dev->dev, NULL,
-+ __fsl_mc_device_remove);
-+ }
-+}
-+
-+static int __fsl_mc_device_match(struct device *dev, void *data)
-+{
-+ struct fsl_mc_obj_desc *obj_desc = data;
-+ struct fsl_mc_device *mc_dev = to_fsl_mc_device(dev);
-+
-+ return fsl_mc_device_match(mc_dev, obj_desc);
-+}
-+
-+static struct fsl_mc_device *fsl_mc_device_lookup(struct fsl_mc_obj_desc
-+ *obj_desc,
-+ struct fsl_mc_device
-+ *mc_bus_dev)
-+{
-+ struct device *dev;
-+
-+ dev = device_find_child(&mc_bus_dev->dev, obj_desc,
-+ __fsl_mc_device_match);
-+
-+ return dev ? to_fsl_mc_device(dev) : NULL;
-+}
-+
-+/**
-+ * check_plugged_state_change - Check change in an MC object's plugged state
-+ *
-+ * @mc_dev: pointer to the fsl-mc device for a given MC object
-+ * @obj_desc: pointer to the MC object's descriptor in the MC
-+ *
-+ * If the plugged state has changed from unplugged to plugged, the fsl-mc
-+ * device is bound to the corresponding device driver.
-+ * If the plugged state has changed from plugged to unplugged, the fsl-mc
-+ * device is unbound from the corresponding device driver.
-+ */
-+static void check_plugged_state_change(struct fsl_mc_device *mc_dev,
-+ struct fsl_mc_obj_desc *obj_desc)
-+{
-+ int error;
-+ u32 plugged_flag_at_mc =
-+ obj_desc->state & FSL_MC_OBJ_STATE_PLUGGED;
-+
-+ if (plugged_flag_at_mc !=
-+ (mc_dev->obj_desc.state & FSL_MC_OBJ_STATE_PLUGGED)) {
-+ if (plugged_flag_at_mc) {
-+ mc_dev->obj_desc.state |= FSL_MC_OBJ_STATE_PLUGGED;
-+ error = device_attach(&mc_dev->dev);
-+ if (error < 0) {
-+ dev_err(&mc_dev->dev,
-+ "device_attach() failed: %d\n",
-+ error);
-+ }
-+ } else {
-+ mc_dev->obj_desc.state &= ~FSL_MC_OBJ_STATE_PLUGGED;
-+ device_release_driver(&mc_dev->dev);
-+ }
-+ }
-+}
-+
-+/**
-+ * dprc_add_new_devices - Adds devices to the logical bus for a DPRC
-+ *
-+ * @mc_bus_dev: pointer to the fsl-mc device that represents a DPRC object
-+ * @driver_override: driver override to apply to new objects found in the
-+ * DPRC, or NULL, if none.
-+ * @obj_desc_array: array of device descriptors for child devices currently
-+ * present in the physical DPRC.
-+ * @num_child_objects_in_mc: number of entries in obj_desc_array
-+ *
-+ * Synchronizes the state of the Linux bus driver with the actual
-+ * state of the MC by adding objects that have been newly discovered
-+ * in the physical DPRC.
-+ */
-+static void dprc_add_new_devices(struct fsl_mc_device *mc_bus_dev,
-+ const char *driver_override,
-+ struct fsl_mc_obj_desc *obj_desc_array,
-+ int num_child_objects_in_mc)
-+{
-+ int error;
-+ int i;
-+
-+ for (i = 0; i < num_child_objects_in_mc; i++) {
-+ struct fsl_mc_device *child_dev;
-+ struct fsl_mc_obj_desc *obj_desc = &obj_desc_array[i];
-+
-+ if (strlen(obj_desc->type) == 0)
-+ continue;
-+
-+ /*
-+ * Check if device is already known to Linux:
-+ */
-+ child_dev = fsl_mc_device_lookup(obj_desc, mc_bus_dev);
-+ if (child_dev) {
-+ check_plugged_state_change(child_dev, obj_desc);
-+ put_device(&child_dev->dev);
-+ continue;
-+ }
-+
-+ error = fsl_mc_device_add(obj_desc, NULL, &mc_bus_dev->dev,
-+ driver_override, &child_dev);
-+ if (error < 0)
-+ continue;
-+ }
-+}
-+
-+/**
-+ * dprc_scan_objects - Discover objects in a DPRC
-+ *
-+ * @mc_bus_dev: pointer to the fsl-mc device that represents a DPRC object
-+ * @driver_override: driver override to apply to new objects found in the
-+ * DPRC, or NULL, if none.
-+ * @total_irq_count: If argument is provided the function populates the
-+ * total number of IRQs created by objects in the DPRC.
-+ *
-+ * Detects objects added and removed from a DPRC and synchronizes the
-+ * state of the Linux bus driver, MC by adding and removing
-+ * devices accordingly.
-+ * Two types of devices can be found in a DPRC: allocatable objects (e.g.,
-+ * dpbp, dpmcp) and non-allocatable devices (e.g., dprc, dpni).
-+ * All allocatable devices needed to be probed before all non-allocatable
-+ * devices, to ensure that device drivers for non-allocatable
-+ * devices can allocate any type of allocatable devices.
-+ * That is, we need to ensure that the corresponding resource pools are
-+ * populated before they can get allocation requests from probe callbacks
-+ * of the device drivers for the non-allocatable devices.
-+ */
-+int dprc_scan_objects(struct fsl_mc_device *mc_bus_dev,
-+ const char *driver_override,
-+ unsigned int *total_irq_count)
-+{
-+ int num_child_objects;
-+ int dprc_get_obj_failures;
-+ int error;
-+ unsigned int irq_count = mc_bus_dev->obj_desc.irq_count;
-+ struct fsl_mc_obj_desc *child_obj_desc_array = NULL;
-+ struct fsl_mc_bus *mc_bus = to_fsl_mc_bus(mc_bus_dev);
-+
-+ error = dprc_get_obj_count(mc_bus_dev->mc_io,
-+ 0,
-+ mc_bus_dev->mc_handle,
-+ &num_child_objects);
-+ if (error < 0) {
-+ dev_err(&mc_bus_dev->dev, "dprc_get_obj_count() failed: %d\n",
-+ error);
-+ return error;
-+ }
-+
-+ if (num_child_objects != 0) {
-+ int i;
-+
-+ child_obj_desc_array =
-+ devm_kmalloc_array(&mc_bus_dev->dev, num_child_objects,
-+ sizeof(*child_obj_desc_array),
-+ GFP_KERNEL);
-+ if (!child_obj_desc_array)
-+ return -ENOMEM;
-+
-+ /*
-+ * Discover objects currently present in the physical DPRC:
-+ */
-+ dprc_get_obj_failures = 0;
-+ for (i = 0; i < num_child_objects; i++) {
-+ struct fsl_mc_obj_desc *obj_desc =
-+ &child_obj_desc_array[i];
-+
-+ error = dprc_get_obj(mc_bus_dev->mc_io,
-+ 0,
-+ mc_bus_dev->mc_handle,
-+ i, obj_desc);
-+ if (error < 0) {
-+ dev_err(&mc_bus_dev->dev,
-+ "dprc_get_obj(i=%d) failed: %d\n",
-+ i, error);
-+ /*
-+ * Mark the obj entry as "invalid", by using the
-+ * empty string as obj type:
-+ */
-+ obj_desc->type[0] = '\0';
-+ obj_desc->id = error;
-+ dprc_get_obj_failures++;
-+ continue;
-+ }
-+
-+ /*
-+ * add a quirk for all versions of dpsec < 4.0...none
-+ * are coherent regardless of what the MC reports.
-+ */
-+ if ((strcmp(obj_desc->type, "dpseci") == 0) &&
-+ (obj_desc->ver_major < 4))
-+ obj_desc->flags |=
-+ FSL_MC_OBJ_FLAG_NO_MEM_SHAREABILITY;
-+
-+ irq_count += obj_desc->irq_count;
-+ dev_dbg(&mc_bus_dev->dev,
-+ "Discovered object: type %s, id %d\n",
-+ obj_desc->type, obj_desc->id);
-+ }
-+
-+ if (dprc_get_obj_failures != 0) {
-+ dev_err(&mc_bus_dev->dev,
-+ "%d out of %d devices could not be retrieved\n",
-+ dprc_get_obj_failures, num_child_objects);
-+ }
-+ }
-+
-+ /*
-+ * Allocate IRQ's before binding the scanned devices with their
-+ * respective drivers.
-+ */
-+ if (dev_get_msi_domain(&mc_bus_dev->dev) && !mc_bus->irq_resources) {
-+ if (irq_count > FSL_MC_IRQ_POOL_MAX_TOTAL_IRQS) {
-+ dev_warn(&mc_bus_dev->dev,
-+ "IRQs needed (%u) exceed IRQs preallocated (%u)\n",
-+ irq_count, FSL_MC_IRQ_POOL_MAX_TOTAL_IRQS);
-+ }
-+
-+ error = fsl_mc_populate_irq_pool(mc_bus,
-+ FSL_MC_IRQ_POOL_MAX_TOTAL_IRQS);
-+ if (error < 0)
-+ return error;
-+ }
-+
-+ if (total_irq_count)
-+ *total_irq_count = irq_count;
-+
-+ dprc_remove_devices(mc_bus_dev, child_obj_desc_array,
-+ num_child_objects);
-+
-+ dprc_add_new_devices(mc_bus_dev, driver_override, child_obj_desc_array,
-+ num_child_objects);
-+
-+ if (child_obj_desc_array)
-+ devm_kfree(&mc_bus_dev->dev, child_obj_desc_array);
-+
-+ return 0;
-+}
-+
-+/**
-+ * dprc_scan_container - Scans a physical DPRC and synchronizes Linux bus state
-+ *
-+ * @mc_bus_dev: pointer to the fsl-mc device that represents a DPRC object
-+ *
-+ * Scans the physical DPRC and synchronizes the state of the Linux
-+ * bus driver with the actual state of the MC by adding and removing
-+ * devices as appropriate.
-+ */
-+static int dprc_scan_container(struct fsl_mc_device *mc_bus_dev)
-+{
-+ int error;
-+ struct fsl_mc_bus *mc_bus = to_fsl_mc_bus(mc_bus_dev);
-+
-+ fsl_mc_init_all_resource_pools(mc_bus_dev);
-+
-+ /*
-+ * Discover objects in the DPRC:
-+ */
-+ mutex_lock(&mc_bus->scan_mutex);
-+ error = dprc_scan_objects(mc_bus_dev, NULL, NULL);
-+ mutex_unlock(&mc_bus->scan_mutex);
-+ if (error < 0) {
-+ fsl_mc_cleanup_all_resource_pools(mc_bus_dev);
-+ return error;
-+ }
-+
-+ return 0;
-+}
-+
-+/**
-+ * dprc_irq0_handler - Regular ISR for DPRC interrupt 0
-+ *
-+ * @irq: IRQ number of the interrupt being handled
-+ * @arg: Pointer to device structure
-+ */
-+static irqreturn_t dprc_irq0_handler(int irq_num, void *arg)
-+{
-+ return IRQ_WAKE_THREAD;
-+}
-+
-+/**
-+ * dprc_irq0_handler_thread - Handler thread function for DPRC interrupt 0
-+ *
-+ * @irq: IRQ number of the interrupt being handled
-+ * @arg: Pointer to device structure
-+ */
-+static irqreturn_t dprc_irq0_handler_thread(int irq_num, void *arg)
-+{
-+ int error;
-+ u32 status;
-+ struct device *dev = arg;
-+ struct fsl_mc_device *mc_dev = to_fsl_mc_device(dev);
-+ struct fsl_mc_bus *mc_bus = to_fsl_mc_bus(mc_dev);
-+ struct fsl_mc_io *mc_io = mc_dev->mc_io;
-+ struct msi_desc *msi_desc = mc_dev->irqs[0]->msi_desc;
-+
-+ dev_dbg(dev, "DPRC IRQ %d triggered on CPU %u\n",
-+ irq_num, smp_processor_id());
-+
-+ if (!(mc_dev->flags & FSL_MC_IS_DPRC))
-+ return IRQ_HANDLED;
-+
-+ mutex_lock(&mc_bus->scan_mutex);
-+ if (!msi_desc || msi_desc->irq != (u32)irq_num)
-+ goto out;
-+
-+ status = 0;
-+ error = dprc_get_irq_status(mc_io, 0, mc_dev->mc_handle, 0,
-+ &status);
-+ if (error < 0) {
-+ dev_err(dev,
-+ "dprc_get_irq_status() failed: %d\n", error);
-+ goto out;
-+ }
-+
-+ error = dprc_clear_irq_status(mc_io, 0, mc_dev->mc_handle, 0,
-+ status);
-+ if (error < 0) {
-+ dev_err(dev,
-+ "dprc_clear_irq_status() failed: %d\n", error);
-+ goto out;
-+ }
-+
-+ if (status & (DPRC_IRQ_EVENT_OBJ_ADDED |
-+ DPRC_IRQ_EVENT_OBJ_REMOVED |
-+ DPRC_IRQ_EVENT_CONTAINER_DESTROYED |
-+ DPRC_IRQ_EVENT_OBJ_DESTROYED |
-+ DPRC_IRQ_EVENT_OBJ_CREATED)) {
-+ unsigned int irq_count;
-+
-+ error = dprc_scan_objects(mc_dev, NULL, &irq_count);
-+ if (error < 0) {
-+ /*
-+ * If the error is -ENXIO, we ignore it, as it indicates
-+ * that the object scan was aborted, as we detected that
-+ * an object was removed from the DPRC in the MC, while
-+ * we were scanning the DPRC.
-+ */
-+ if (error != -ENXIO) {
-+ dev_err(dev, "dprc_scan_objects() failed: %d\n",
-+ error);
-+ }
-+
-+ goto out;
-+ }
-+
-+ if (irq_count > FSL_MC_IRQ_POOL_MAX_TOTAL_IRQS) {
-+ dev_warn(dev,
-+ "IRQs needed (%u) exceed IRQs preallocated (%u)\n",
-+ irq_count, FSL_MC_IRQ_POOL_MAX_TOTAL_IRQS);
-+ }
-+ }
-+
-+out:
-+ mutex_unlock(&mc_bus->scan_mutex);
-+ return IRQ_HANDLED;
-+}
-+
-+/*
-+ * Disable and clear interrupt for a given DPRC object
-+ */
-+static int disable_dprc_irq(struct fsl_mc_device *mc_dev)
-+{
-+ int error;
-+ struct fsl_mc_io *mc_io = mc_dev->mc_io;
-+
-+ /*
-+ * Disable generation of interrupt, while we configure it:
-+ */
-+ error = dprc_set_irq_enable(mc_io, 0, mc_dev->mc_handle, 0, 0);
-+ if (error < 0) {
-+ dev_err(&mc_dev->dev,
-+ "Disabling DPRC IRQ failed: dprc_set_irq_enable() failed: %d\n",
-+ error);
-+ return error;
-+ }
-+
-+ /*
-+ * Disable all interrupt causes for the interrupt:
-+ */
-+ error = dprc_set_irq_mask(mc_io, 0, mc_dev->mc_handle, 0, 0x0);
-+ if (error < 0) {
-+ dev_err(&mc_dev->dev,
-+ "Disabling DPRC IRQ failed: dprc_set_irq_mask() failed: %d\n",
-+ error);
-+ return error;
-+ }
-+
-+ /*
-+ * Clear any leftover interrupts:
-+ */
-+ error = dprc_clear_irq_status(mc_io, 0, mc_dev->mc_handle, 0, ~0x0U);
-+ if (error < 0) {
-+ dev_err(&mc_dev->dev,
-+ "Disabling DPRC IRQ failed: dprc_clear_irq_status() failed: %d\n",
-+ error);
-+ return error;
-+ }
-+
-+ return 0;
-+}
-+
-+static int register_dprc_irq_handler(struct fsl_mc_device *mc_dev)
-+{
-+ int error;
-+ struct fsl_mc_device_irq *irq = mc_dev->irqs[0];
-+
-+ /*
-+ * NOTE: devm_request_threaded_irq() invokes the device-specific
-+ * function that programs the MSI physically in the device
-+ */
-+ error = devm_request_threaded_irq(&mc_dev->dev,
-+ irq->msi_desc->irq,
-+ dprc_irq0_handler,
-+ dprc_irq0_handler_thread,
-+ IRQF_NO_SUSPEND | IRQF_ONESHOT,
-+ dev_name(&mc_dev->dev),
-+ &mc_dev->dev);
-+ if (error < 0) {
-+ dev_err(&mc_dev->dev,
-+ "devm_request_threaded_irq() failed: %d\n",
-+ error);
-+ return error;
-+ }
-+
-+ return 0;
-+}
-+
-+static int enable_dprc_irq(struct fsl_mc_device *mc_dev)
-+{
-+ int error;
-+
-+ /*
-+ * Enable all interrupt causes for the interrupt:
-+ */
-+ error = dprc_set_irq_mask(mc_dev->mc_io, 0, mc_dev->mc_handle, 0,
-+ ~0x0u);
-+ if (error < 0) {
-+ dev_err(&mc_dev->dev,
-+ "Enabling DPRC IRQ failed: dprc_set_irq_mask() failed: %d\n",
-+ error);
-+
-+ return error;
-+ }
-+
-+ /*
-+ * Enable generation of the interrupt:
-+ */
-+ error = dprc_set_irq_enable(mc_dev->mc_io, 0, mc_dev->mc_handle, 0, 1);
-+ if (error < 0) {
-+ dev_err(&mc_dev->dev,
-+ "Enabling DPRC IRQ failed: dprc_set_irq_enable() failed: %d\n",
-+ error);
-+
-+ return error;
-+ }
-+
-+ return 0;
-+}
-+
-+/*
-+ * Setup interrupt for a given DPRC device
-+ */
-+static int dprc_setup_irq(struct fsl_mc_device *mc_dev)
-+{
-+ int error;
-+
-+ error = fsl_mc_allocate_irqs(mc_dev);
-+ if (error < 0)
-+ return error;
-+
-+ error = disable_dprc_irq(mc_dev);
-+ if (error < 0)
-+ goto error_free_irqs;
-+
-+ error = register_dprc_irq_handler(mc_dev);
-+ if (error < 0)
-+ goto error_free_irqs;
-+
-+ error = enable_dprc_irq(mc_dev);
-+ if (error < 0)
-+ goto error_free_irqs;
-+
-+ return 0;
-+
-+error_free_irqs:
-+ fsl_mc_free_irqs(mc_dev);
-+ return error;
-+}
-+
-+/**
-+ * dprc_probe - callback invoked when a DPRC is being bound to this driver
-+ *
-+ * @mc_dev: Pointer to fsl-mc device representing a DPRC
-+ *
-+ * It opens the physical DPRC in the MC.
-+ * It scans the DPRC to discover the MC objects contained in it.
-+ * It creates the interrupt pool for the MC bus associated with the DPRC.
-+ * It configures the interrupts for the DPRC device itself.
-+ */
-+static int dprc_probe(struct fsl_mc_device *mc_dev)
-+{
-+ int error;
-+ size_t region_size;
-+ struct device *parent_dev = mc_dev->dev.parent;
-+ struct fsl_mc_bus *mc_bus = to_fsl_mc_bus(mc_dev);
-+ bool mc_io_created = false;
-+ bool msi_domain_set = false;
-+ u16 major_ver, minor_ver;
-+
-+ if (!is_fsl_mc_bus_dprc(mc_dev))
-+ return -EINVAL;
-+
-+ if (dev_get_msi_domain(&mc_dev->dev))
-+ return -EINVAL;
-+
-+ if (!mc_dev->mc_io) {
-+ /*
-+ * This is a child DPRC:
-+ */
-+ if (!dev_is_fsl_mc(parent_dev))
-+ return -EINVAL;
-+
-+ if (mc_dev->obj_desc.region_count == 0)
-+ return -EINVAL;
-+
-+ region_size = resource_size(mc_dev->regions);
-+
-+ error = fsl_create_mc_io(&mc_dev->dev,
-+ mc_dev->regions[0].start,
-+ region_size,
-+ NULL,
-+ FSL_MC_IO_ATOMIC_CONTEXT_PORTAL,
-+ &mc_dev->mc_io);
-+ if (error < 0)
-+ return error;
-+
-+ mc_io_created = true;
-+
-+ /*
-+ * Inherit parent MSI domain:
-+ */
-+ dev_set_msi_domain(&mc_dev->dev,
-+ dev_get_msi_domain(parent_dev));
-+ msi_domain_set = true;
-+ } else {
-+ /*
-+ * This is a root DPRC
-+ */
-+ struct irq_domain *mc_msi_domain;
-+
-+ if (dev_is_fsl_mc(parent_dev))
-+ return -EINVAL;
-+
-+ error = fsl_mc_find_msi_domain(parent_dev,
-+ &mc_msi_domain);
-+ if (error < 0) {
-+ dev_warn(&mc_dev->dev,
-+ "WARNING: MC bus without interrupt support\n");
-+ } else {
-+ dev_set_msi_domain(&mc_dev->dev, mc_msi_domain);
-+ msi_domain_set = true;
-+ }
-+ }
-+
-+ error = dprc_open(mc_dev->mc_io, 0, mc_dev->obj_desc.id,
-+ &mc_dev->mc_handle);
-+ if (error < 0) {
-+ dev_err(&mc_dev->dev, "dprc_open() failed: %d\n", error);
-+ goto error_cleanup_msi_domain;
-+ }
-+
-+ error = dprc_get_attributes(mc_dev->mc_io, 0, mc_dev->mc_handle,
-+ &mc_bus->dprc_attr);
-+ if (error < 0) {
-+ dev_err(&mc_dev->dev, "dprc_get_attributes() failed: %d\n",
-+ error);
-+ goto error_cleanup_open;
-+ }
-+
-+ error = dprc_get_api_version(mc_dev->mc_io, 0,
-+ &major_ver,
-+ &minor_ver);
-+ if (error < 0) {
-+ dev_err(&mc_dev->dev, "dprc_get_api_version() failed: %d\n",
-+ error);
-+ goto error_cleanup_open;
-+ }
-+
-+ if (major_ver < DPRC_MIN_VER_MAJOR ||
-+ (major_ver == DPRC_MIN_VER_MAJOR &&
-+ minor_ver < DPRC_MIN_VER_MINOR)) {
-+ dev_err(&mc_dev->dev,
-+ "ERROR: DPRC version %d.%d not supported\n",
-+ major_ver, minor_ver);
-+ error = -ENOTSUPP;
-+ goto error_cleanup_open;
-+ }
-+
-+ mutex_init(&mc_bus->scan_mutex);
-+
-+ /*
-+ * Discover MC objects in DPRC object:
-+ */
-+ error = dprc_scan_container(mc_dev);
-+ if (error < 0)
-+ goto error_cleanup_open;
-+
-+ /*
-+ * Configure interrupt for the DPRC object associated with this MC bus:
-+ */
-+ error = dprc_setup_irq(mc_dev);
-+ if (error < 0)
-+ goto error_cleanup_open;
-+
-+ dev_info(&mc_dev->dev, "DPRC device bound to driver");
-+ return 0;
-+
-+error_cleanup_open:
-+ (void)dprc_close(mc_dev->mc_io, 0, mc_dev->mc_handle);
-+
-+error_cleanup_msi_domain:
-+ if (msi_domain_set)
-+ dev_set_msi_domain(&mc_dev->dev, NULL);
-+
-+ if (mc_io_created) {
-+ fsl_destroy_mc_io(mc_dev->mc_io);
-+ mc_dev->mc_io = NULL;
-+ }
-+
-+ return error;
-+}
-+
-+/*
-+ * Tear down interrupt for a given DPRC object
-+ */
-+static void dprc_teardown_irq(struct fsl_mc_device *mc_dev)
-+{
-+ struct fsl_mc_device_irq *irq = mc_dev->irqs[0];
-+
-+ (void)disable_dprc_irq(mc_dev);
-+
-+ devm_free_irq(&mc_dev->dev, irq->msi_desc->irq, &mc_dev->dev);
-+
-+ fsl_mc_free_irqs(mc_dev);
-+}
-+
-+/**
-+ * dprc_remove - callback invoked when a DPRC is being unbound from this driver
-+ *
-+ * @mc_dev: Pointer to fsl-mc device representing the DPRC
-+ *
-+ * It removes the DPRC's child objects from Linux (not from the MC) and
-+ * closes the DPRC device in the MC.
-+ * It tears down the interrupts that were configured for the DPRC device.
-+ * It destroys the interrupt pool associated with this MC bus.
-+ */
-+static int dprc_remove(struct fsl_mc_device *mc_dev)
-+{
-+ int error;
-+ struct fsl_mc_bus *mc_bus = to_fsl_mc_bus(mc_dev);
-+
-+ if (!is_fsl_mc_bus_dprc(mc_dev))
-+ return -EINVAL;
-+ if (!mc_dev->mc_io)
-+ return -EINVAL;
-+
-+ if (!mc_bus->irq_resources)
-+ return -EINVAL;
-+
-+ if (dev_get_msi_domain(&mc_dev->dev))
-+ dprc_teardown_irq(mc_dev);
-+
-+ device_for_each_child(&mc_dev->dev, NULL, __fsl_mc_device_remove);
-+
-+ if (dev_get_msi_domain(&mc_dev->dev)) {
-+ fsl_mc_cleanup_irq_pool(mc_bus);
-+ dev_set_msi_domain(&mc_dev->dev, NULL);
-+ }
-+
-+ fsl_mc_cleanup_all_resource_pools(mc_dev);
-+
-+ error = dprc_close(mc_dev->mc_io, 0, mc_dev->mc_handle);
-+ if (error < 0)
-+ dev_err(&mc_dev->dev, "dprc_close() failed: %d\n", error);
-+
-+ if (!fsl_mc_is_root_dprc(&mc_dev->dev)) {
-+ fsl_destroy_mc_io(mc_dev->mc_io);
-+ mc_dev->mc_io = NULL;
-+ }
-+
-+ dev_info(&mc_dev->dev, "DPRC device unbound from driver");
-+ return 0;
-+}
-+
-+static const struct fsl_mc_device_id match_id_table[] = {
-+ {
-+ .vendor = FSL_MC_VENDOR_FREESCALE,
-+ .obj_type = "dprc"},
-+ {.vendor = 0x0},
-+};
-+
-+static struct fsl_mc_driver dprc_driver = {
-+ .driver = {
-+ .name = FSL_MC_DPRC_DRIVER_NAME,
-+ .owner = THIS_MODULE,
-+ .pm = NULL,
-+ },
-+ .match_id_table = match_id_table,
-+ .probe = dprc_probe,
-+ .remove = dprc_remove,
-+};
-+
-+int __init dprc_driver_init(void)
-+{
-+ return fsl_mc_driver_register(&dprc_driver);
-+}
-+
-+void dprc_driver_exit(void)
-+{
-+ fsl_mc_driver_unregister(&dprc_driver);
-+}
---- /dev/null
-+++ b/drivers/bus/fsl-mc/dprc.c
-@@ -0,0 +1,575 @@
-+// SPDX-License-Identifier: (GPL-2.0+ OR BSD-3-Clause)
-+/*
-+ * Copyright 2013-2016 Freescale Semiconductor Inc.
-+ *
-+ */
-+#include <linux/kernel.h>
-+#include <linux/fsl/mc.h>
-+
-+#include "fsl-mc-private.h"
-+
-+/**
-+ * dprc_open() - Open DPRC object for use
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @container_id: Container ID to open
-+ * @token: Returned token of DPRC object
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ *
-+ * @warning Required before any operation on the object.
-+ */
-+int dprc_open(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ int container_id,
-+ u16 *token)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dprc_cmd_open *cmd_params;
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPRC_CMDID_OPEN, cmd_flags,
-+ 0);
-+ cmd_params = (struct dprc_cmd_open *)cmd.params;
-+ cmd_params->container_id = cpu_to_le32(container_id);
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ *token = mc_cmd_hdr_read_token(&cmd);
-+
-+ return 0;
-+}
-+EXPORT_SYMBOL_GPL(dprc_open);
-+
-+/**
-+ * dprc_close() - Close the control session of the object
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPRC object
-+ *
-+ * After this function is called, no further operations are
-+ * allowed on the object without opening a new control session.
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dprc_close(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPRC_CMDID_CLOSE, cmd_flags,
-+ token);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+EXPORT_SYMBOL_GPL(dprc_close);
-+
-+/**
-+ * dprc_reset_container - Reset child container.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPRC object
-+ * @child_container_id: ID of the container to reset
-+ *
-+ * In case a software context crashes or becomes non-responsive, the parent
-+ * may wish to reset its resources container before the software context is
-+ * restarted.
-+ *
-+ * This routine informs all objects assigned to the child container that the
-+ * container is being reset, so they may perform any cleanup operations that are
-+ * needed. All objects handles that were owned by the child container shall be
-+ * closed.
-+ *
-+ * Note that such request may be submitted even if the child software context
-+ * has not crashed, but the resulting object cleanup operations will not be
-+ * aware of that.
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dprc_reset_container(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ int child_container_id)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dprc_cmd_reset_container *cmd_params;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPRC_CMDID_RESET_CONT,
-+ cmd_flags, token);
-+ cmd_params = (struct dprc_cmd_reset_container *)cmd.params;
-+ cmd_params->child_container_id = cpu_to_le32(child_container_id);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+EXPORT_SYMBOL_GPL(dprc_reset_container);
-+
-+/**
-+ * dprc_set_irq() - Set IRQ information for the DPRC to trigger an interrupt.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPRC object
-+ * @irq_index: Identifies the interrupt index to configure
-+ * @irq_cfg: IRQ configuration
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dprc_set_irq(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 irq_index,
-+ struct dprc_irq_cfg *irq_cfg)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dprc_cmd_set_irq *cmd_params;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPRC_CMDID_SET_IRQ,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dprc_cmd_set_irq *)cmd.params;
-+ cmd_params->irq_val = cpu_to_le32(irq_cfg->val);
-+ cmd_params->irq_index = irq_index;
-+ cmd_params->irq_addr = cpu_to_le64(irq_cfg->paddr);
-+ cmd_params->irq_num = cpu_to_le32(irq_cfg->irq_num);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dprc_set_irq_enable() - Set overall interrupt state.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPRC object
-+ * @irq_index: The interrupt index to configure
-+ * @en: Interrupt state - enable = 1, disable = 0
-+ *
-+ * Allows GPP software to control when interrupts are generated.
-+ * Each interrupt can have up to 32 causes. The enable/disable control's the
-+ * overall interrupt state. if the interrupt is disabled no causes will cause
-+ * an interrupt.
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dprc_set_irq_enable(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 irq_index,
-+ u8 en)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dprc_cmd_set_irq_enable *cmd_params;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPRC_CMDID_SET_IRQ_ENABLE,
-+ cmd_flags, token);
-+ cmd_params = (struct dprc_cmd_set_irq_enable *)cmd.params;
-+ cmd_params->enable = en & DPRC_ENABLE;
-+ cmd_params->irq_index = irq_index;
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dprc_set_irq_mask() - Set interrupt mask.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPRC object
-+ * @irq_index: The interrupt index to configure
-+ * @mask: event mask to trigger interrupt;
-+ * each bit:
-+ * 0 = ignore event
-+ * 1 = consider event for asserting irq
-+ *
-+ * Every interrupt can have up to 32 causes and the interrupt model supports
-+ * masking/unmasking each cause independently
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dprc_set_irq_mask(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 irq_index,
-+ u32 mask)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dprc_cmd_set_irq_mask *cmd_params;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPRC_CMDID_SET_IRQ_MASK,
-+ cmd_flags, token);
-+ cmd_params = (struct dprc_cmd_set_irq_mask *)cmd.params;
-+ cmd_params->mask = cpu_to_le32(mask);
-+ cmd_params->irq_index = irq_index;
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dprc_get_irq_status() - Get the current status of any pending interrupts.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPRC object
-+ * @irq_index: The interrupt index to configure
-+ * @status: Returned interrupts status - one bit per cause:
-+ * 0 = no interrupt pending
-+ * 1 = interrupt pending
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dprc_get_irq_status(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 irq_index,
-+ u32 *status)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dprc_cmd_get_irq_status *cmd_params;
-+ struct dprc_rsp_get_irq_status *rsp_params;
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPRC_CMDID_GET_IRQ_STATUS,
-+ cmd_flags, token);
-+ cmd_params = (struct dprc_cmd_get_irq_status *)cmd.params;
-+ cmd_params->status = cpu_to_le32(*status);
-+ cmd_params->irq_index = irq_index;
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ rsp_params = (struct dprc_rsp_get_irq_status *)cmd.params;
-+ *status = le32_to_cpu(rsp_params->status);
-+
-+ return 0;
-+}
-+
-+/**
-+ * dprc_clear_irq_status() - Clear a pending interrupt's status
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPRC object
-+ * @irq_index: The interrupt index to configure
-+ * @status: bits to clear (W1C) - one bit per cause:
-+ * 0 = don't change
-+ * 1 = clear status bit
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dprc_clear_irq_status(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 irq_index,
-+ u32 status)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dprc_cmd_clear_irq_status *cmd_params;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPRC_CMDID_CLEAR_IRQ_STATUS,
-+ cmd_flags, token);
-+ cmd_params = (struct dprc_cmd_clear_irq_status *)cmd.params;
-+ cmd_params->status = cpu_to_le32(status);
-+ cmd_params->irq_index = irq_index;
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dprc_get_attributes() - Obtains container attributes
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPRC object
-+ * @attributes Returned container attributes
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dprc_get_attributes(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ struct dprc_attributes *attr)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dprc_rsp_get_attributes *rsp_params;
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPRC_CMDID_GET_ATTR,
-+ cmd_flags,
-+ token);
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ rsp_params = (struct dprc_rsp_get_attributes *)cmd.params;
-+ attr->container_id = le32_to_cpu(rsp_params->container_id);
-+ attr->icid = le32_to_cpu(rsp_params->icid);
-+ attr->options = le32_to_cpu(rsp_params->options);
-+ attr->portal_id = le32_to_cpu(rsp_params->portal_id);
-+
-+ return 0;
-+}
-+
-+/**
-+ * dprc_get_obj_count() - Obtains the number of objects in the DPRC
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPRC object
-+ * @obj_count: Number of objects assigned to the DPRC
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dprc_get_obj_count(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ int *obj_count)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dprc_rsp_get_obj_count *rsp_params;
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPRC_CMDID_GET_OBJ_COUNT,
-+ cmd_flags, token);
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ rsp_params = (struct dprc_rsp_get_obj_count *)cmd.params;
-+ *obj_count = le32_to_cpu(rsp_params->obj_count);
-+
-+ return 0;
-+}
-+EXPORT_SYMBOL_GPL(dprc_get_obj_count);
-+
-+/**
-+ * dprc_get_obj() - Get general information on an object
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPRC object
-+ * @obj_index: Index of the object to be queried (< obj_count)
-+ * @obj_desc: Returns the requested object descriptor
-+ *
-+ * The object descriptors are retrieved one by one by incrementing
-+ * obj_index up to (not including) the value of obj_count returned
-+ * from dprc_get_obj_count(). dprc_get_obj_count() must
-+ * be called prior to dprc_get_obj().
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dprc_get_obj(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ int obj_index,
-+ struct fsl_mc_obj_desc *obj_desc)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dprc_cmd_get_obj *cmd_params;
-+ struct dprc_rsp_get_obj *rsp_params;
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPRC_CMDID_GET_OBJ,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dprc_cmd_get_obj *)cmd.params;
-+ cmd_params->obj_index = cpu_to_le32(obj_index);
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ rsp_params = (struct dprc_rsp_get_obj *)cmd.params;
-+ obj_desc->id = le32_to_cpu(rsp_params->id);
-+ obj_desc->vendor = le16_to_cpu(rsp_params->vendor);
-+ obj_desc->irq_count = rsp_params->irq_count;
-+ obj_desc->region_count = rsp_params->region_count;
-+ obj_desc->state = le32_to_cpu(rsp_params->state);
-+ obj_desc->ver_major = le16_to_cpu(rsp_params->version_major);
-+ obj_desc->ver_minor = le16_to_cpu(rsp_params->version_minor);
-+ obj_desc->flags = le16_to_cpu(rsp_params->flags);
-+ strncpy(obj_desc->type, rsp_params->type, 16);
-+ obj_desc->type[15] = '\0';
-+ strncpy(obj_desc->label, rsp_params->label, 16);
-+ obj_desc->label[15] = '\0';
-+ return 0;
-+}
-+EXPORT_SYMBOL_GPL(dprc_get_obj);
-+
-+/**
-+ * dprc_set_obj_irq() - Set IRQ information for object to trigger an interrupt.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPRC object
-+ * @obj_type: Type of the object to set its IRQ
-+ * @obj_id: ID of the object to set its IRQ
-+ * @irq_index: The interrupt index to configure
-+ * @irq_cfg: IRQ configuration
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dprc_set_obj_irq(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ char *obj_type,
-+ int obj_id,
-+ u8 irq_index,
-+ struct dprc_irq_cfg *irq_cfg)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dprc_cmd_set_obj_irq *cmd_params;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPRC_CMDID_SET_OBJ_IRQ,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dprc_cmd_set_obj_irq *)cmd.params;
-+ cmd_params->irq_val = cpu_to_le32(irq_cfg->val);
-+ cmd_params->irq_index = irq_index;
-+ cmd_params->irq_addr = cpu_to_le64(irq_cfg->paddr);
-+ cmd_params->irq_num = cpu_to_le32(irq_cfg->irq_num);
-+ cmd_params->obj_id = cpu_to_le32(obj_id);
-+ strncpy(cmd_params->obj_type, obj_type, 16);
-+ cmd_params->obj_type[15] = '\0';
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+EXPORT_SYMBOL_GPL(dprc_set_obj_irq);
-+
-+/**
-+ * dprc_get_obj_region() - Get region information for a specified object.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPRC object
-+ * @obj_type; Object type as returned in dprc_get_obj()
-+ * @obj_id: Unique object instance as returned in dprc_get_obj()
-+ * @region_index: The specific region to query
-+ * @region_desc: Returns the requested region descriptor
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dprc_get_obj_region(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ char *obj_type,
-+ int obj_id,
-+ u8 region_index,
-+ struct dprc_region_desc *region_desc)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dprc_cmd_get_obj_region *cmd_params;
-+ struct dprc_rsp_get_obj_region *rsp_params;
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPRC_CMDID_GET_OBJ_REG,
-+ cmd_flags, token);
-+ cmd_params = (struct dprc_cmd_get_obj_region *)cmd.params;
-+ cmd_params->obj_id = cpu_to_le32(obj_id);
-+ cmd_params->region_index = region_index;
-+ strncpy(cmd_params->obj_type, obj_type, 16);
-+ cmd_params->obj_type[15] = '\0';
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ rsp_params = (struct dprc_rsp_get_obj_region *)cmd.params;
-+ region_desc->base_offset = le32_to_cpu(rsp_params->base_addr);
-+ region_desc->size = le32_to_cpu(rsp_params->size);
-+ region_desc->type = rsp_params->type;
-+ region_desc->flags = le32_to_cpu(rsp_params->flags);
-+
-+ return 0;
-+}
-+EXPORT_SYMBOL_GPL(dprc_get_obj_region);
-+
-+/**
-+ * dprc_get_api_version - Get Data Path Resource Container API version
-+ * @mc_io: Pointer to Mc portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @major_ver: Major version of Data Path Resource Container API
-+ * @minor_ver: Minor version of Data Path Resource Container API
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dprc_get_api_version(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 *major_ver,
-+ u16 *minor_ver)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPRC_CMDID_GET_API_VERSION,
-+ cmd_flags, 0);
-+
-+ /* send command to mc */
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ mc_cmd_read_api_version(&cmd, major_ver, minor_ver);
-+
-+ return 0;
-+}
-+
-+/**
-+ * dprc_get_container_id - Get container ID associated with a given portal.
-+ * @mc_io: Pointer to Mc portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @container_id: Requested container id
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dprc_get_container_id(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ int *container_id)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPRC_CMDID_GET_CONT_ID,
-+ cmd_flags,
-+ 0);
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ *container_id = (int)mc_cmd_read_object_id(&cmd);
-+
-+ return 0;
-+}
---- a/drivers/staging/fsl-mc/bus/fsl-mc-allocator.c
-+++ /dev/null
-@@ -1,668 +0,0 @@
--/*
-- * Freescale MC object device allocator driver
-- *
-- * Copyright (C) 2013 Freescale Semiconductor, Inc.
-- *
-- * This file is licensed under the terms of the GNU General Public
-- * License version 2. This program is licensed "as is" without any
-- * warranty of any kind, whether express or implied.
-- */
--
--#include <linux/module.h>
--#include <linux/msi.h>
--#include "../include/mc-bus.h"
--#include "../include/mc-sys.h"
--#include "../include/dpbp-cmd.h"
--#include "../include/dpcon-cmd.h"
--
--#include "fsl-mc-private.h"
--
--#define FSL_MC_IS_ALLOCATABLE(_obj_type) \
-- (strcmp(_obj_type, "dpbp") == 0 || \
-- strcmp(_obj_type, "dpmcp") == 0 || \
-- strcmp(_obj_type, "dpcon") == 0)
--
--/**
-- * fsl_mc_resource_pool_add_device - add allocatable device to a resource
-- * pool of a given MC bus
-- *
-- * @mc_bus: pointer to the MC bus
-- * @pool_type: MC bus pool type
-- * @mc_dev: Pointer to allocatable MC object device
-- *
-- * It adds an allocatable MC object device to a container's resource pool of
-- * the given resource type
-- */
--static int __must_check fsl_mc_resource_pool_add_device(struct fsl_mc_bus
-- *mc_bus,
-- enum fsl_mc_pool_type
-- pool_type,
-- struct fsl_mc_device
-- *mc_dev)
--{
-- struct fsl_mc_resource_pool *res_pool;
-- struct fsl_mc_resource *resource;
-- struct fsl_mc_device *mc_bus_dev = &mc_bus->mc_dev;
-- int error = -EINVAL;
--
-- if (WARN_ON(pool_type < 0 || pool_type >= FSL_MC_NUM_POOL_TYPES))
-- goto out;
-- if (WARN_ON(!FSL_MC_IS_ALLOCATABLE(mc_dev->obj_desc.type)))
-- goto out;
-- if (WARN_ON(mc_dev->resource))
-- goto out;
--
-- res_pool = &mc_bus->resource_pools[pool_type];
-- if (WARN_ON(res_pool->type != pool_type))
-- goto out;
-- if (WARN_ON(res_pool->mc_bus != mc_bus))
-- goto out;
--
-- mutex_lock(&res_pool->mutex);
--
-- if (WARN_ON(res_pool->max_count < 0))
-- goto out_unlock;
-- if (WARN_ON(res_pool->free_count < 0 ||
-- res_pool->free_count > res_pool->max_count))
-- goto out_unlock;
--
-- resource = devm_kzalloc(&mc_bus_dev->dev, sizeof(*resource),
-- GFP_KERNEL);
-- if (!resource) {
-- error = -ENOMEM;
-- dev_err(&mc_bus_dev->dev,
-- "Failed to allocate memory for fsl_mc_resource\n");
-- goto out_unlock;
-- }
--
-- resource->type = pool_type;
-- resource->id = mc_dev->obj_desc.id;
-- resource->data = mc_dev;
-- resource->parent_pool = res_pool;
-- INIT_LIST_HEAD(&resource->node);
-- list_add_tail(&resource->node, &res_pool->free_list);
-- mc_dev->resource = resource;
-- res_pool->free_count++;
-- res_pool->max_count++;
-- error = 0;
--out_unlock:
-- mutex_unlock(&res_pool->mutex);
--out:
-- return error;
--}
--
--/**
-- * fsl_mc_resource_pool_remove_device - remove an allocatable device from a
-- * resource pool
-- *
-- * @mc_dev: Pointer to allocatable MC object device
-- *
-- * It permanently removes an allocatable MC object device from the resource
-- * pool, the device is currently in, as long as it is in the pool's free list.
-- */
--static int __must_check fsl_mc_resource_pool_remove_device(struct fsl_mc_device
-- *mc_dev)
--{
-- struct fsl_mc_device *mc_bus_dev;
-- struct fsl_mc_bus *mc_bus;
-- struct fsl_mc_resource_pool *res_pool;
-- struct fsl_mc_resource *resource;
-- int error = -EINVAL;
--
-- if (WARN_ON(!FSL_MC_IS_ALLOCATABLE(mc_dev->obj_desc.type)))
-- goto out;
--
-- resource = mc_dev->resource;
-- if (WARN_ON(!resource || resource->data != mc_dev))
-- goto out;
--
-- mc_bus_dev = to_fsl_mc_device(mc_dev->dev.parent);
-- mc_bus = to_fsl_mc_bus(mc_bus_dev);
-- res_pool = resource->parent_pool;
-- if (WARN_ON(res_pool != &mc_bus->resource_pools[resource->type]))
-- goto out;
--
-- mutex_lock(&res_pool->mutex);
--
-- if (WARN_ON(res_pool->max_count <= 0))
-- goto out_unlock;
-- if (WARN_ON(res_pool->free_count <= 0 ||
-- res_pool->free_count > res_pool->max_count))
-- goto out_unlock;
--
-- /*
-- * If the device is currently allocated, its resource is not
-- * in the free list and thus, the device cannot be removed.
-- */
-- if (list_empty(&resource->node)) {
-- error = -EBUSY;
-- dev_err(&mc_bus_dev->dev,
-- "Device %s cannot be removed from resource pool\n",
-- dev_name(&mc_dev->dev));
-- goto out_unlock;
-- }
--
-- list_del_init(&resource->node);
-- res_pool->free_count--;
-- res_pool->max_count--;
--
-- devm_kfree(&mc_bus_dev->dev, resource);
-- mc_dev->resource = NULL;
-- error = 0;
--out_unlock:
-- mutex_unlock(&res_pool->mutex);
--out:
-- return error;
--}
--
--static const char *const fsl_mc_pool_type_strings[] = {
-- [FSL_MC_POOL_DPMCP] = "dpmcp",
-- [FSL_MC_POOL_DPBP] = "dpbp",
-- [FSL_MC_POOL_DPCON] = "dpcon",
-- [FSL_MC_POOL_IRQ] = "irq",
--};
--
--static int __must_check object_type_to_pool_type(const char *object_type,
-- enum fsl_mc_pool_type
-- *pool_type)
--{
-- unsigned int i;
--
-- for (i = 0; i < ARRAY_SIZE(fsl_mc_pool_type_strings); i++) {
-- if (strcmp(object_type, fsl_mc_pool_type_strings[i]) == 0) {
-- *pool_type = i;
-- return 0;
-- }
-- }
--
-- return -EINVAL;
--}
--
--int __must_check fsl_mc_resource_allocate(struct fsl_mc_bus *mc_bus,
-- enum fsl_mc_pool_type pool_type,
-- struct fsl_mc_resource **new_resource)
--{
-- struct fsl_mc_resource_pool *res_pool;
-- struct fsl_mc_resource *resource;
-- struct fsl_mc_device *mc_bus_dev = &mc_bus->mc_dev;
-- int error = -EINVAL;
--
-- BUILD_BUG_ON(ARRAY_SIZE(fsl_mc_pool_type_strings) !=
-- FSL_MC_NUM_POOL_TYPES);
--
-- *new_resource = NULL;
-- if (WARN_ON(pool_type < 0 || pool_type >= FSL_MC_NUM_POOL_TYPES))
-- goto out;
--
-- res_pool = &mc_bus->resource_pools[pool_type];
-- if (WARN_ON(res_pool->mc_bus != mc_bus))
-- goto out;
--
-- mutex_lock(&res_pool->mutex);
-- resource = list_first_entry_or_null(&res_pool->free_list,
-- struct fsl_mc_resource, node);
--
-- if (!resource) {
-- WARN_ON(res_pool->free_count != 0);
-- error = -ENXIO;
-- dev_err(&mc_bus_dev->dev,
-- "No more resources of type %s left\n",
-- fsl_mc_pool_type_strings[pool_type]);
-- goto out_unlock;
-- }
--
-- if (WARN_ON(resource->type != pool_type))
-- goto out_unlock;
-- if (WARN_ON(resource->parent_pool != res_pool))
-- goto out_unlock;
-- if (WARN_ON(res_pool->free_count <= 0 ||
-- res_pool->free_count > res_pool->max_count))
-- goto out_unlock;
--
-- list_del_init(&resource->node);
--
-- res_pool->free_count--;
-- error = 0;
--out_unlock:
-- mutex_unlock(&res_pool->mutex);
-- *new_resource = resource;
--out:
-- return error;
--}
--EXPORT_SYMBOL_GPL(fsl_mc_resource_allocate);
--
--void fsl_mc_resource_free(struct fsl_mc_resource *resource)
--{
-- struct fsl_mc_resource_pool *res_pool;
--
-- res_pool = resource->parent_pool;
-- if (WARN_ON(resource->type != res_pool->type))
-- return;
--
-- mutex_lock(&res_pool->mutex);
-- if (WARN_ON(res_pool->free_count < 0 ||
-- res_pool->free_count >= res_pool->max_count))
-- goto out_unlock;
--
-- if (WARN_ON(!list_empty(&resource->node)))
-- goto out_unlock;
--
-- list_add_tail(&resource->node, &res_pool->free_list);
-- res_pool->free_count++;
--out_unlock:
-- mutex_unlock(&res_pool->mutex);
--}
--EXPORT_SYMBOL_GPL(fsl_mc_resource_free);
--
--/**
-- * fsl_mc_object_allocate - Allocates a MC object device of the given
-- * pool type from a given MC bus
-- *
-- * @mc_dev: MC device for which the MC object device is to be allocated
-- * @pool_type: MC bus resource pool type
-- * @new_mc_dev: Pointer to area where the pointer to the allocated
-- * MC object device is to be returned
-- *
-- * This function allocates a MC object device from the device's parent DPRC,
-- * from the corresponding MC bus' pool of allocatable MC object devices of
-- * the given resource type. mc_dev cannot be a DPRC itself.
-- *
-- * NOTE: pool_type must be different from FSL_MC_POOL_MCP, since MC
-- * portals are allocated using fsl_mc_portal_allocate(), instead of
-- * this function.
-- */
--int __must_check fsl_mc_object_allocate(struct fsl_mc_device *mc_dev,
-- enum fsl_mc_pool_type pool_type,
-- struct fsl_mc_device **new_mc_adev)
--{
-- struct fsl_mc_device *mc_bus_dev;
-- struct fsl_mc_bus *mc_bus;
-- struct fsl_mc_device *mc_adev;
-- int error = -EINVAL;
-- struct fsl_mc_resource *resource = NULL;
--
-- *new_mc_adev = NULL;
-- if (WARN_ON(mc_dev->flags & FSL_MC_IS_DPRC))
-- goto error;
--
-- if (WARN_ON(!dev_is_fsl_mc(mc_dev->dev.parent)))
-- goto error;
--
-- if (WARN_ON(pool_type == FSL_MC_POOL_DPMCP))
-- goto error;
--
-- mc_bus_dev = to_fsl_mc_device(mc_dev->dev.parent);
-- mc_bus = to_fsl_mc_bus(mc_bus_dev);
-- error = fsl_mc_resource_allocate(mc_bus, pool_type, &resource);
-- if (error < 0)
-- goto error;
--
-- mc_adev = resource->data;
-- if (WARN_ON(!mc_adev))
-- goto error;
--
-- *new_mc_adev = mc_adev;
-- return 0;
--error:
-- if (resource)
-- fsl_mc_resource_free(resource);
--
-- return error;
--}
--EXPORT_SYMBOL_GPL(fsl_mc_object_allocate);
--
--/**
-- * fsl_mc_object_free - Returns an allocatable MC object device to the
-- * corresponding resource pool of a given MC bus.
-- *
-- * @mc_adev: Pointer to the MC object device
-- */
--void fsl_mc_object_free(struct fsl_mc_device *mc_adev)
--{
-- struct fsl_mc_resource *resource;
--
-- resource = mc_adev->resource;
-- if (WARN_ON(resource->type == FSL_MC_POOL_DPMCP))
-- return;
-- if (WARN_ON(resource->data != mc_adev))
-- return;
--
-- fsl_mc_resource_free(resource);
--}
--EXPORT_SYMBOL_GPL(fsl_mc_object_free);
--
--/*
-- * Initialize the interrupt pool associated with a MC bus.
-- * It allocates a block of IRQs from the GIC-ITS
-- */
--int fsl_mc_populate_irq_pool(struct fsl_mc_bus *mc_bus,
-- unsigned int irq_count)
--{
-- unsigned int i;
-- struct msi_desc *msi_desc;
-- struct fsl_mc_device_irq *irq_resources;
-- struct fsl_mc_device_irq *mc_dev_irq;
-- int error;
-- struct fsl_mc_device *mc_bus_dev = &mc_bus->mc_dev;
-- struct fsl_mc_resource_pool *res_pool =
-- &mc_bus->resource_pools[FSL_MC_POOL_IRQ];
--
-- if (WARN_ON(irq_count == 0 ||
-- irq_count > FSL_MC_IRQ_POOL_MAX_TOTAL_IRQS))
-- return -EINVAL;
--
-- error = fsl_mc_msi_domain_alloc_irqs(&mc_bus_dev->dev, irq_count);
-- if (error < 0)
-- return error;
--
-- irq_resources = devm_kzalloc(&mc_bus_dev->dev,
-- sizeof(*irq_resources) * irq_count,
-- GFP_KERNEL);
-- if (!irq_resources) {
-- error = -ENOMEM;
-- goto cleanup_msi_irqs;
-- }
--
-- for (i = 0; i < irq_count; i++) {
-- mc_dev_irq = &irq_resources[i];
--
-- /*
-- * NOTE: This mc_dev_irq's MSI addr/value pair will be set
-- * by the fsl_mc_msi_write_msg() callback
-- */
-- mc_dev_irq->resource.type = res_pool->type;
-- mc_dev_irq->resource.data = mc_dev_irq;
-- mc_dev_irq->resource.parent_pool = res_pool;
-- INIT_LIST_HEAD(&mc_dev_irq->resource.node);
-- list_add_tail(&mc_dev_irq->resource.node, &res_pool->free_list);
-- }
--
-- for_each_msi_entry(msi_desc, &mc_bus_dev->dev) {
-- mc_dev_irq = &irq_resources[msi_desc->fsl_mc.msi_index];
-- mc_dev_irq->msi_desc = msi_desc;
-- mc_dev_irq->resource.id = msi_desc->irq;
-- }
--
-- res_pool->max_count = irq_count;
-- res_pool->free_count = irq_count;
-- mc_bus->irq_resources = irq_resources;
-- return 0;
--
--cleanup_msi_irqs:
-- fsl_mc_msi_domain_free_irqs(&mc_bus_dev->dev);
-- return error;
--}
--EXPORT_SYMBOL_GPL(fsl_mc_populate_irq_pool);
--
--/**
-- * Teardown the interrupt pool associated with an MC bus.
-- * It frees the IRQs that were allocated to the pool, back to the GIC-ITS.
-- */
--void fsl_mc_cleanup_irq_pool(struct fsl_mc_bus *mc_bus)
--{
-- struct fsl_mc_device *mc_bus_dev = &mc_bus->mc_dev;
-- struct fsl_mc_resource_pool *res_pool =
-- &mc_bus->resource_pools[FSL_MC_POOL_IRQ];
--
-- if (WARN_ON(!mc_bus->irq_resources))
-- return;
--
-- if (WARN_ON(res_pool->max_count == 0))
-- return;
--
-- if (WARN_ON(res_pool->free_count != res_pool->max_count))
-- return;
--
-- INIT_LIST_HEAD(&res_pool->free_list);
-- res_pool->max_count = 0;
-- res_pool->free_count = 0;
-- mc_bus->irq_resources = NULL;
-- fsl_mc_msi_domain_free_irqs(&mc_bus_dev->dev);
--}
--EXPORT_SYMBOL_GPL(fsl_mc_cleanup_irq_pool);
--
--/**
-- * It allocates the IRQs required by a given MC object device. The
-- * IRQs are allocated from the interrupt pool associated with the
-- * MC bus that contains the device, if the device is not a DPRC device.
-- * Otherwise, the IRQs are allocated from the interrupt pool associated
-- * with the MC bus that represents the DPRC device itself.
-- */
--int __must_check fsl_mc_allocate_irqs(struct fsl_mc_device *mc_dev)
--{
-- int i;
-- int irq_count;
-- int res_allocated_count = 0;
-- int error = -EINVAL;
-- struct fsl_mc_device_irq **irqs = NULL;
-- struct fsl_mc_bus *mc_bus;
-- struct fsl_mc_resource_pool *res_pool;
--
-- if (WARN_ON(mc_dev->irqs))
-- return -EINVAL;
--
-- irq_count = mc_dev->obj_desc.irq_count;
-- if (WARN_ON(irq_count == 0))
-- return -EINVAL;
--
-- if (strcmp(mc_dev->obj_desc.type, "dprc") == 0)
-- mc_bus = to_fsl_mc_bus(mc_dev);
-- else
-- mc_bus = to_fsl_mc_bus(to_fsl_mc_device(mc_dev->dev.parent));
--
-- if (WARN_ON(!mc_bus->irq_resources))
-- return -EINVAL;
--
-- res_pool = &mc_bus->resource_pools[FSL_MC_POOL_IRQ];
-- if (res_pool->free_count < irq_count) {
-- dev_err(&mc_dev->dev,
-- "Not able to allocate %u irqs for device\n", irq_count);
-- return -ENOSPC;
-- }
--
-- irqs = devm_kzalloc(&mc_dev->dev, irq_count * sizeof(irqs[0]),
-- GFP_KERNEL);
-- if (!irqs)
-- return -ENOMEM;
--
-- for (i = 0; i < irq_count; i++) {
-- struct fsl_mc_resource *resource;
--
-- error = fsl_mc_resource_allocate(mc_bus, FSL_MC_POOL_IRQ,
-- &resource);
-- if (error < 0)
-- goto error_resource_alloc;
--
-- irqs[i] = to_fsl_mc_irq(resource);
-- res_allocated_count++;
--
-- WARN_ON(irqs[i]->mc_dev);
-- irqs[i]->mc_dev = mc_dev;
-- irqs[i]->dev_irq_index = i;
-- }
--
-- mc_dev->irqs = irqs;
-- return 0;
--
--error_resource_alloc:
-- for (i = 0; i < res_allocated_count; i++) {
-- irqs[i]->mc_dev = NULL;
-- fsl_mc_resource_free(&irqs[i]->resource);
-- }
--
-- return error;
--}
--EXPORT_SYMBOL_GPL(fsl_mc_allocate_irqs);
--
--/*
-- * It frees the IRQs that were allocated for a MC object device, by
-- * returning them to the corresponding interrupt pool.
-- */
--void fsl_mc_free_irqs(struct fsl_mc_device *mc_dev)
--{
-- int i;
-- int irq_count;
-- struct fsl_mc_bus *mc_bus;
-- struct fsl_mc_device_irq **irqs = mc_dev->irqs;
--
-- if (WARN_ON(!irqs))
-- return;
--
-- irq_count = mc_dev->obj_desc.irq_count;
--
-- if (strcmp(mc_dev->obj_desc.type, "dprc") == 0)
-- mc_bus = to_fsl_mc_bus(mc_dev);
-- else
-- mc_bus = to_fsl_mc_bus(to_fsl_mc_device(mc_dev->dev.parent));
--
-- if (WARN_ON(!mc_bus->irq_resources))
-- return;
--
-- for (i = 0; i < irq_count; i++) {
-- WARN_ON(!irqs[i]->mc_dev);
-- irqs[i]->mc_dev = NULL;
-- fsl_mc_resource_free(&irqs[i]->resource);
-- }
--
-- mc_dev->irqs = NULL;
--}
--EXPORT_SYMBOL_GPL(fsl_mc_free_irqs);
--
--void fsl_mc_init_all_resource_pools(struct fsl_mc_device *mc_bus_dev)
--{
-- int pool_type;
-- struct fsl_mc_bus *mc_bus = to_fsl_mc_bus(mc_bus_dev);
--
-- for (pool_type = 0; pool_type < FSL_MC_NUM_POOL_TYPES; pool_type++) {
-- struct fsl_mc_resource_pool *res_pool =
-- &mc_bus->resource_pools[pool_type];
--
-- res_pool->type = pool_type;
-- res_pool->max_count = 0;
-- res_pool->free_count = 0;
-- res_pool->mc_bus = mc_bus;
-- INIT_LIST_HEAD(&res_pool->free_list);
-- mutex_init(&res_pool->mutex);
-- }
--}
--
--static void fsl_mc_cleanup_resource_pool(struct fsl_mc_device *mc_bus_dev,
-- enum fsl_mc_pool_type pool_type)
--{
-- struct fsl_mc_resource *resource;
-- struct fsl_mc_resource *next;
-- struct fsl_mc_bus *mc_bus = to_fsl_mc_bus(mc_bus_dev);
-- struct fsl_mc_resource_pool *res_pool =
-- &mc_bus->resource_pools[pool_type];
-- int free_count = 0;
--
-- WARN_ON(res_pool->type != pool_type);
-- WARN_ON(res_pool->free_count != res_pool->max_count);
--
-- list_for_each_entry_safe(resource, next, &res_pool->free_list, node) {
-- free_count++;
-- WARN_ON(resource->type != res_pool->type);
-- WARN_ON(resource->parent_pool != res_pool);
-- devm_kfree(&mc_bus_dev->dev, resource);
-- }
--
-- WARN_ON(free_count != res_pool->free_count);
--}
--
--void fsl_mc_cleanup_all_resource_pools(struct fsl_mc_device *mc_bus_dev)
--{
-- int pool_type;
--
-- for (pool_type = 0; pool_type < FSL_MC_NUM_POOL_TYPES; pool_type++)
-- fsl_mc_cleanup_resource_pool(mc_bus_dev, pool_type);
--}
--
--/**
-- * fsl_mc_allocator_probe - callback invoked when an allocatable device is
-- * being added to the system
-- */
--static int fsl_mc_allocator_probe(struct fsl_mc_device *mc_dev)
--{
-- enum fsl_mc_pool_type pool_type;
-- struct fsl_mc_device *mc_bus_dev;
-- struct fsl_mc_bus *mc_bus;
-- int error;
--
-- if (WARN_ON(!FSL_MC_IS_ALLOCATABLE(mc_dev->obj_desc.type)))
-- return -EINVAL;
--
-- mc_bus_dev = to_fsl_mc_device(mc_dev->dev.parent);
-- if (WARN_ON(!dev_is_fsl_mc(&mc_bus_dev->dev)))
-- return -EINVAL;
--
-- mc_bus = to_fsl_mc_bus(mc_bus_dev);
-- error = object_type_to_pool_type(mc_dev->obj_desc.type, &pool_type);
-- if (error < 0)
-- return error;
--
-- error = fsl_mc_resource_pool_add_device(mc_bus, pool_type, mc_dev);
-- if (error < 0)
-- return error;
--
-- dev_dbg(&mc_dev->dev,
-- "Allocatable MC object device bound to fsl_mc_allocator driver");
-- return 0;
--}
--
--/**
-- * fsl_mc_allocator_remove - callback invoked when an allocatable device is
-- * being removed from the system
-- */
--static int fsl_mc_allocator_remove(struct fsl_mc_device *mc_dev)
--{
-- int error;
--
-- if (WARN_ON(!FSL_MC_IS_ALLOCATABLE(mc_dev->obj_desc.type)))
-- return -EINVAL;
--
-- if (mc_dev->resource) {
-- error = fsl_mc_resource_pool_remove_device(mc_dev);
-- if (error < 0)
-- return error;
-- }
--
-- dev_dbg(&mc_dev->dev,
-- "Allocatable MC object device unbound from fsl_mc_allocator driver");
-- return 0;
--}
--
--static const struct fsl_mc_device_id match_id_table[] = {
-- {
-- .vendor = FSL_MC_VENDOR_FREESCALE,
-- .obj_type = "dpbp",
-- },
-- {
-- .vendor = FSL_MC_VENDOR_FREESCALE,
-- .obj_type = "dpmcp",
-- },
-- {
-- .vendor = FSL_MC_VENDOR_FREESCALE,
-- .obj_type = "dpcon",
-- },
-- {.vendor = 0x0},
--};
--
--static struct fsl_mc_driver fsl_mc_allocator_driver = {
-- .driver = {
-- .name = "fsl_mc_allocator",
-- .pm = NULL,
-- },
-- .match_id_table = match_id_table,
-- .probe = fsl_mc_allocator_probe,
-- .remove = fsl_mc_allocator_remove,
--};
--
--int __init fsl_mc_allocator_driver_init(void)
--{
-- return fsl_mc_driver_register(&fsl_mc_allocator_driver);
--}
--
--void fsl_mc_allocator_driver_exit(void)
--{
-- fsl_mc_driver_unregister(&fsl_mc_allocator_driver);
--}
---- /dev/null
-+++ b/drivers/bus/fsl-mc/fsl-mc-allocator.c
-@@ -0,0 +1,655 @@
-+// SPDX-License-Identifier: GPL-2.0
-+/*
-+ * fsl-mc object allocator driver
-+ *
-+ * Copyright (C) 2013-2016 Freescale Semiconductor, Inc.
-+ *
-+ */
-+
-+#include <linux/module.h>
-+#include <linux/msi.h>
-+#include <linux/fsl/mc.h>
-+
-+#include "fsl-mc-private.h"
-+
-+static bool __must_check fsl_mc_is_allocatable(struct fsl_mc_device *mc_dev)
-+{
-+ return is_fsl_mc_bus_dpbp(mc_dev) ||
-+ is_fsl_mc_bus_dpmcp(mc_dev) ||
-+ is_fsl_mc_bus_dpcon(mc_dev);
-+}
-+
-+/**
-+ * fsl_mc_resource_pool_add_device - add allocatable object to a resource
-+ * pool of a given fsl-mc bus
-+ *
-+ * @mc_bus: pointer to the fsl-mc bus
-+ * @pool_type: pool type
-+ * @mc_dev: pointer to allocatable fsl-mc device
-+ */
-+static int __must_check fsl_mc_resource_pool_add_device(struct fsl_mc_bus
-+ *mc_bus,
-+ enum fsl_mc_pool_type
-+ pool_type,
-+ struct fsl_mc_device
-+ *mc_dev)
-+{
-+ struct fsl_mc_resource_pool *res_pool;
-+ struct fsl_mc_resource *resource;
-+ struct fsl_mc_device *mc_bus_dev = &mc_bus->mc_dev;
-+ int error = -EINVAL;
-+
-+ if (pool_type < 0 || pool_type >= FSL_MC_NUM_POOL_TYPES)
-+ goto out;
-+ if (!fsl_mc_is_allocatable(mc_dev))
-+ goto out;
-+ if (mc_dev->resource)
-+ goto out;
-+
-+ res_pool = &mc_bus->resource_pools[pool_type];
-+ if (res_pool->type != pool_type)
-+ goto out;
-+ if (res_pool->mc_bus != mc_bus)
-+ goto out;
-+
-+ mutex_lock(&res_pool->mutex);
-+
-+ if (res_pool->max_count < 0)
-+ goto out_unlock;
-+ if (res_pool->free_count < 0 ||
-+ res_pool->free_count > res_pool->max_count)
-+ goto out_unlock;
-+
-+ resource = devm_kzalloc(&mc_bus_dev->dev, sizeof(*resource),
-+ GFP_KERNEL);
-+ if (!resource) {
-+ error = -ENOMEM;
-+ dev_err(&mc_bus_dev->dev,
-+ "Failed to allocate memory for fsl_mc_resource\n");
-+ goto out_unlock;
-+ }
-+
-+ resource->type = pool_type;
-+ resource->id = mc_dev->obj_desc.id;
-+ resource->data = mc_dev;
-+ resource->parent_pool = res_pool;
-+ INIT_LIST_HEAD(&resource->node);
-+ list_add_tail(&resource->node, &res_pool->free_list);
-+ mc_dev->resource = resource;
-+ res_pool->free_count++;
-+ res_pool->max_count++;
-+ error = 0;
-+out_unlock:
-+ mutex_unlock(&res_pool->mutex);
-+out:
-+ return error;
-+}
-+
-+/**
-+ * fsl_mc_resource_pool_remove_device - remove an allocatable device from a
-+ * resource pool
-+ *
-+ * @mc_dev: pointer to allocatable fsl-mc device
-+ *
-+ * It permanently removes an allocatable fsl-mc device from the resource
-+ * pool. It's an error if the device is in use.
-+ */
-+static int __must_check fsl_mc_resource_pool_remove_device(struct fsl_mc_device
-+ *mc_dev)
-+{
-+ struct fsl_mc_device *mc_bus_dev;
-+ struct fsl_mc_bus *mc_bus;
-+ struct fsl_mc_resource_pool *res_pool;
-+ struct fsl_mc_resource *resource;
-+ int error = -EINVAL;
-+
-+ if (!fsl_mc_is_allocatable(mc_dev))
-+ goto out;
-+
-+ resource = mc_dev->resource;
-+ if (!resource || resource->data != mc_dev)
-+ goto out;
-+
-+ mc_bus_dev = to_fsl_mc_device(mc_dev->dev.parent);
-+ mc_bus = to_fsl_mc_bus(mc_bus_dev);
-+ res_pool = resource->parent_pool;
-+ if (res_pool != &mc_bus->resource_pools[resource->type])
-+ goto out;
-+
-+ mutex_lock(&res_pool->mutex);
-+
-+ if (res_pool->max_count <= 0)
-+ goto out_unlock;
-+ if (res_pool->free_count <= 0 ||
-+ res_pool->free_count > res_pool->max_count)
-+ goto out_unlock;
-+
-+ /*
-+ * If the device is currently allocated, its resource is not
-+ * in the free list and thus, the device cannot be removed.
-+ */
-+ if (list_empty(&resource->node)) {
-+ error = -EBUSY;
-+ dev_err(&mc_bus_dev->dev,
-+ "Device %s cannot be removed from resource pool\n",
-+ dev_name(&mc_dev->dev));
-+ goto out_unlock;
-+ }
-+
-+ list_del_init(&resource->node);
-+ res_pool->free_count--;
-+ res_pool->max_count--;
-+
-+ devm_kfree(&mc_bus_dev->dev, resource);
-+ mc_dev->resource = NULL;
-+ error = 0;
-+out_unlock:
-+ mutex_unlock(&res_pool->mutex);
-+out:
-+ return error;
-+}
-+
-+static const char *const fsl_mc_pool_type_strings[] = {
-+ [FSL_MC_POOL_DPMCP] = "dpmcp",
-+ [FSL_MC_POOL_DPBP] = "dpbp",
-+ [FSL_MC_POOL_DPCON] = "dpcon",
-+ [FSL_MC_POOL_IRQ] = "irq",
-+};
-+
-+static int __must_check object_type_to_pool_type(const char *object_type,
-+ enum fsl_mc_pool_type
-+ *pool_type)
-+{
-+ unsigned int i;
-+
-+ for (i = 0; i < ARRAY_SIZE(fsl_mc_pool_type_strings); i++) {
-+ if (strcmp(object_type, fsl_mc_pool_type_strings[i]) == 0) {
-+ *pool_type = i;
-+ return 0;
-+ }
-+ }
-+
-+ return -EINVAL;
-+}
-+
-+int __must_check fsl_mc_resource_allocate(struct fsl_mc_bus *mc_bus,
-+ enum fsl_mc_pool_type pool_type,
-+ struct fsl_mc_resource **new_resource)
-+{
-+ struct fsl_mc_resource_pool *res_pool;
-+ struct fsl_mc_resource *resource;
-+ struct fsl_mc_device *mc_bus_dev = &mc_bus->mc_dev;
-+ int error = -EINVAL;
-+
-+ BUILD_BUG_ON(ARRAY_SIZE(fsl_mc_pool_type_strings) !=
-+ FSL_MC_NUM_POOL_TYPES);
-+
-+ *new_resource = NULL;
-+ if (pool_type < 0 || pool_type >= FSL_MC_NUM_POOL_TYPES)
-+ goto out;
-+
-+ res_pool = &mc_bus->resource_pools[pool_type];
-+ if (res_pool->mc_bus != mc_bus)
-+ goto out;
-+
-+ mutex_lock(&res_pool->mutex);
-+ resource = list_first_entry_or_null(&res_pool->free_list,
-+ struct fsl_mc_resource, node);
-+
-+ if (!resource) {
-+ error = -ENXIO;
-+ dev_err(&mc_bus_dev->dev,
-+ "No more resources of type %s left\n",
-+ fsl_mc_pool_type_strings[pool_type]);
-+ goto out_unlock;
-+ }
-+
-+ if (resource->type != pool_type)
-+ goto out_unlock;
-+ if (resource->parent_pool != res_pool)
-+ goto out_unlock;
-+ if (res_pool->free_count <= 0 ||
-+ res_pool->free_count > res_pool->max_count)
-+ goto out_unlock;
-+
-+ list_del_init(&resource->node);
-+
-+ res_pool->free_count--;
-+ error = 0;
-+out_unlock:
-+ mutex_unlock(&res_pool->mutex);
-+ *new_resource = resource;
-+out:
-+ return error;
-+}
-+EXPORT_SYMBOL_GPL(fsl_mc_resource_allocate);
-+
-+void fsl_mc_resource_free(struct fsl_mc_resource *resource)
-+{
-+ struct fsl_mc_resource_pool *res_pool;
-+
-+ res_pool = resource->parent_pool;
-+ if (resource->type != res_pool->type)
-+ return;
-+
-+ mutex_lock(&res_pool->mutex);
-+ if (res_pool->free_count < 0 ||
-+ res_pool->free_count >= res_pool->max_count)
-+ goto out_unlock;
-+
-+ if (!list_empty(&resource->node))
-+ goto out_unlock;
-+
-+ list_add_tail(&resource->node, &res_pool->free_list);
-+ res_pool->free_count++;
-+out_unlock:
-+ mutex_unlock(&res_pool->mutex);
-+}
-+EXPORT_SYMBOL_GPL(fsl_mc_resource_free);
-+
-+/**
-+ * fsl_mc_object_allocate - Allocates an fsl-mc object of the given
-+ * pool type from a given fsl-mc bus instance
-+ *
-+ * @mc_dev: fsl-mc device which is used in conjunction with the
-+ * allocated object
-+ * @pool_type: pool type
-+ * @new_mc_dev: pointer to area where the pointer to the allocated device
-+ * is to be returned
-+ *
-+ * Allocatable objects are always used in conjunction with some functional
-+ * device. This function allocates an object of the specified type from
-+ * the DPRC containing the functional device.
-+ *
-+ * NOTE: pool_type must be different from FSL_MC_POOL_MCP, since MC
-+ * portals are allocated using fsl_mc_portal_allocate(), instead of
-+ * this function.
-+ */
-+int __must_check fsl_mc_object_allocate(struct fsl_mc_device *mc_dev,
-+ enum fsl_mc_pool_type pool_type,
-+ struct fsl_mc_device **new_mc_adev)
-+{
-+ struct fsl_mc_device *mc_bus_dev;
-+ struct fsl_mc_bus *mc_bus;
-+ struct fsl_mc_device *mc_adev;
-+ int error = -EINVAL;
-+ struct fsl_mc_resource *resource = NULL;
-+
-+ *new_mc_adev = NULL;
-+ if (mc_dev->flags & FSL_MC_IS_DPRC)
-+ goto error;
-+
-+ if (!dev_is_fsl_mc(mc_dev->dev.parent))
-+ goto error;
-+
-+ if (pool_type == FSL_MC_POOL_DPMCP)
-+ goto error;
-+
-+ mc_bus_dev = to_fsl_mc_device(mc_dev->dev.parent);
-+ mc_bus = to_fsl_mc_bus(mc_bus_dev);
-+ error = fsl_mc_resource_allocate(mc_bus, pool_type, &resource);
-+ if (error < 0)
-+ goto error;
-+
-+ mc_adev = resource->data;
-+ if (!mc_adev)
-+ goto error;
-+
-+ *new_mc_adev = mc_adev;
-+ return 0;
-+error:
-+ if (resource)
-+ fsl_mc_resource_free(resource);
-+
-+ return error;
-+}
-+EXPORT_SYMBOL_GPL(fsl_mc_object_allocate);
-+
-+/**
-+ * fsl_mc_object_free - Returns an fsl-mc object to the resource
-+ * pool where it came from.
-+ * @mc_adev: Pointer to the fsl-mc device
-+ */
-+void fsl_mc_object_free(struct fsl_mc_device *mc_adev)
-+{
-+ struct fsl_mc_resource *resource;
-+
-+ resource = mc_adev->resource;
-+ if (resource->type == FSL_MC_POOL_DPMCP)
-+ return;
-+ if (resource->data != mc_adev)
-+ return;
-+
-+ fsl_mc_resource_free(resource);
-+}
-+EXPORT_SYMBOL_GPL(fsl_mc_object_free);
-+
-+/*
-+ * A DPRC and the devices in the DPRC all share the same GIC-ITS device
-+ * ID. A block of IRQs is pre-allocated and maintained in a pool
-+ * from which devices can allocate them when needed.
-+ */
-+
-+/*
-+ * Initialize the interrupt pool associated with an fsl-mc bus.
-+ * It allocates a block of IRQs from the GIC-ITS.
-+ */
-+int fsl_mc_populate_irq_pool(struct fsl_mc_bus *mc_bus,
-+ unsigned int irq_count)
-+{
-+ unsigned int i;
-+ struct msi_desc *msi_desc;
-+ struct fsl_mc_device_irq *irq_resources;
-+ struct fsl_mc_device_irq *mc_dev_irq;
-+ int error;
-+ struct fsl_mc_device *mc_bus_dev = &mc_bus->mc_dev;
-+ struct fsl_mc_resource_pool *res_pool =
-+ &mc_bus->resource_pools[FSL_MC_POOL_IRQ];
-+
-+ if (irq_count == 0 ||
-+ irq_count > FSL_MC_IRQ_POOL_MAX_TOTAL_IRQS)
-+ return -EINVAL;
-+
-+ error = fsl_mc_msi_domain_alloc_irqs(&mc_bus_dev->dev, irq_count);
-+ if (error < 0)
-+ return error;
-+
-+ irq_resources = devm_kzalloc(&mc_bus_dev->dev,
-+ sizeof(*irq_resources) * irq_count,
-+ GFP_KERNEL);
-+ if (!irq_resources) {
-+ error = -ENOMEM;
-+ goto cleanup_msi_irqs;
-+ }
-+
-+ for (i = 0; i < irq_count; i++) {
-+ mc_dev_irq = &irq_resources[i];
-+
-+ /*
-+ * NOTE: This mc_dev_irq's MSI addr/value pair will be set
-+ * by the fsl_mc_msi_write_msg() callback
-+ */
-+ mc_dev_irq->resource.type = res_pool->type;
-+ mc_dev_irq->resource.data = mc_dev_irq;
-+ mc_dev_irq->resource.parent_pool = res_pool;
-+ INIT_LIST_HEAD(&mc_dev_irq->resource.node);
-+ list_add_tail(&mc_dev_irq->resource.node, &res_pool->free_list);
-+ }
-+
-+ for_each_msi_entry(msi_desc, &mc_bus_dev->dev) {
-+ mc_dev_irq = &irq_resources[msi_desc->fsl_mc.msi_index];
-+ mc_dev_irq->msi_desc = msi_desc;
-+ mc_dev_irq->resource.id = msi_desc->irq;
-+ }
-+
-+ res_pool->max_count = irq_count;
-+ res_pool->free_count = irq_count;
-+ mc_bus->irq_resources = irq_resources;
-+ return 0;
-+
-+cleanup_msi_irqs:
-+ fsl_mc_msi_domain_free_irqs(&mc_bus_dev->dev);
-+ return error;
-+}
-+EXPORT_SYMBOL_GPL(fsl_mc_populate_irq_pool);
-+
-+/**
-+ * Teardown the interrupt pool associated with an fsl-mc bus.
-+ * It frees the IRQs that were allocated to the pool, back to the GIC-ITS.
-+ */
-+void fsl_mc_cleanup_irq_pool(struct fsl_mc_bus *mc_bus)
-+{
-+ struct fsl_mc_device *mc_bus_dev = &mc_bus->mc_dev;
-+ struct fsl_mc_resource_pool *res_pool =
-+ &mc_bus->resource_pools[FSL_MC_POOL_IRQ];
-+
-+ if (!mc_bus->irq_resources)
-+ return;
-+
-+ if (res_pool->max_count == 0)
-+ return;
-+
-+ if (res_pool->free_count != res_pool->max_count)
-+ return;
-+
-+ INIT_LIST_HEAD(&res_pool->free_list);
-+ res_pool->max_count = 0;
-+ res_pool->free_count = 0;
-+ mc_bus->irq_resources = NULL;
-+ fsl_mc_msi_domain_free_irqs(&mc_bus_dev->dev);
-+}
-+EXPORT_SYMBOL_GPL(fsl_mc_cleanup_irq_pool);
-+
-+/**
-+ * Allocate the IRQs required by a given fsl-mc device.
-+ */
-+int __must_check fsl_mc_allocate_irqs(struct fsl_mc_device *mc_dev)
-+{
-+ int i;
-+ int irq_count;
-+ int res_allocated_count = 0;
-+ int error = -EINVAL;
-+ struct fsl_mc_device_irq **irqs = NULL;
-+ struct fsl_mc_bus *mc_bus;
-+ struct fsl_mc_resource_pool *res_pool;
-+
-+ if (mc_dev->irqs)
-+ return -EINVAL;
-+
-+ irq_count = mc_dev->obj_desc.irq_count;
-+ if (irq_count == 0)
-+ return -EINVAL;
-+
-+ if (is_fsl_mc_bus_dprc(mc_dev))
-+ mc_bus = to_fsl_mc_bus(mc_dev);
-+ else
-+ mc_bus = to_fsl_mc_bus(to_fsl_mc_device(mc_dev->dev.parent));
-+
-+ if (!mc_bus->irq_resources)
-+ return -EINVAL;
-+
-+ res_pool = &mc_bus->resource_pools[FSL_MC_POOL_IRQ];
-+ if (res_pool->free_count < irq_count) {
-+ dev_err(&mc_dev->dev,
-+ "Not able to allocate %u irqs for device\n", irq_count);
-+ return -ENOSPC;
-+ }
-+
-+ irqs = devm_kzalloc(&mc_dev->dev, irq_count * sizeof(irqs[0]),
-+ GFP_KERNEL);
-+ if (!irqs)
-+ return -ENOMEM;
-+
-+ for (i = 0; i < irq_count; i++) {
-+ struct fsl_mc_resource *resource;
-+
-+ error = fsl_mc_resource_allocate(mc_bus, FSL_MC_POOL_IRQ,
-+ &resource);
-+ if (error < 0)
-+ goto error_resource_alloc;
-+
-+ irqs[i] = to_fsl_mc_irq(resource);
-+ res_allocated_count++;
-+
-+ irqs[i]->mc_dev = mc_dev;
-+ irqs[i]->dev_irq_index = i;
-+ }
-+
-+ mc_dev->irqs = irqs;
-+ return 0;
-+
-+error_resource_alloc:
-+ for (i = 0; i < res_allocated_count; i++) {
-+ irqs[i]->mc_dev = NULL;
-+ fsl_mc_resource_free(&irqs[i]->resource);
-+ }
-+
-+ return error;
-+}
-+EXPORT_SYMBOL_GPL(fsl_mc_allocate_irqs);
-+
-+/*
-+ * Frees the IRQs that were allocated for an fsl-mc device.
-+ */
-+void fsl_mc_free_irqs(struct fsl_mc_device *mc_dev)
-+{
-+ int i;
-+ int irq_count;
-+ struct fsl_mc_bus *mc_bus;
-+ struct fsl_mc_device_irq **irqs = mc_dev->irqs;
-+
-+ if (!irqs)
-+ return;
-+
-+ irq_count = mc_dev->obj_desc.irq_count;
-+
-+ if (is_fsl_mc_bus_dprc(mc_dev))
-+ mc_bus = to_fsl_mc_bus(mc_dev);
-+ else
-+ mc_bus = to_fsl_mc_bus(to_fsl_mc_device(mc_dev->dev.parent));
-+
-+ if (!mc_bus->irq_resources)
-+ return;
-+
-+ for (i = 0; i < irq_count; i++) {
-+ irqs[i]->mc_dev = NULL;
-+ fsl_mc_resource_free(&irqs[i]->resource);
-+ }
-+
-+ mc_dev->irqs = NULL;
-+}
-+EXPORT_SYMBOL_GPL(fsl_mc_free_irqs);
-+
-+void fsl_mc_init_all_resource_pools(struct fsl_mc_device *mc_bus_dev)
-+{
-+ int pool_type;
-+ struct fsl_mc_bus *mc_bus = to_fsl_mc_bus(mc_bus_dev);
-+
-+ for (pool_type = 0; pool_type < FSL_MC_NUM_POOL_TYPES; pool_type++) {
-+ struct fsl_mc_resource_pool *res_pool =
-+ &mc_bus->resource_pools[pool_type];
-+
-+ res_pool->type = pool_type;
-+ res_pool->max_count = 0;
-+ res_pool->free_count = 0;
-+ res_pool->mc_bus = mc_bus;
-+ INIT_LIST_HEAD(&res_pool->free_list);
-+ mutex_init(&res_pool->mutex);
-+ }
-+}
-+EXPORT_SYMBOL_GPL(fsl_mc_init_all_resource_pools);
-+
-+static void fsl_mc_cleanup_resource_pool(struct fsl_mc_device *mc_bus_dev,
-+ enum fsl_mc_pool_type pool_type)
-+{
-+ struct fsl_mc_resource *resource;
-+ struct fsl_mc_resource *next;
-+ struct fsl_mc_bus *mc_bus = to_fsl_mc_bus(mc_bus_dev);
-+ struct fsl_mc_resource_pool *res_pool =
-+ &mc_bus->resource_pools[pool_type];
-+ int free_count = 0;
-+
-+ list_for_each_entry_safe(resource, next, &res_pool->free_list, node) {
-+ free_count++;
-+ devm_kfree(&mc_bus_dev->dev, resource);
-+ }
-+}
-+
-+void fsl_mc_cleanup_all_resource_pools(struct fsl_mc_device *mc_bus_dev)
-+{
-+ int pool_type;
-+
-+ for (pool_type = 0; pool_type < FSL_MC_NUM_POOL_TYPES; pool_type++)
-+ fsl_mc_cleanup_resource_pool(mc_bus_dev, pool_type);
-+}
-+EXPORT_SYMBOL_GPL(fsl_mc_cleanup_all_resource_pools);
-+
-+/**
-+ * fsl_mc_allocator_probe - callback invoked when an allocatable device is
-+ * being added to the system
-+ */
-+static int fsl_mc_allocator_probe(struct fsl_mc_device *mc_dev)
-+{
-+ enum fsl_mc_pool_type pool_type;
-+ struct fsl_mc_device *mc_bus_dev;
-+ struct fsl_mc_bus *mc_bus;
-+ int error;
-+
-+ if (!fsl_mc_is_allocatable(mc_dev))
-+ return -EINVAL;
-+
-+ mc_bus_dev = to_fsl_mc_device(mc_dev->dev.parent);
-+ if (!dev_is_fsl_mc(&mc_bus_dev->dev))
-+ return -EINVAL;
-+
-+ mc_bus = to_fsl_mc_bus(mc_bus_dev);
-+ error = object_type_to_pool_type(mc_dev->obj_desc.type, &pool_type);
-+ if (error < 0)
-+ return error;
-+
-+ error = fsl_mc_resource_pool_add_device(mc_bus, pool_type, mc_dev);
-+ if (error < 0)
-+ return error;
-+
-+ dev_dbg(&mc_dev->dev,
-+ "Allocatable fsl-mc device bound to fsl_mc_allocator driver");
-+ return 0;
-+}
-+
-+/**
-+ * fsl_mc_allocator_remove - callback invoked when an allocatable device is
-+ * being removed from the system
-+ */
-+static int fsl_mc_allocator_remove(struct fsl_mc_device *mc_dev)
-+{
-+ int error;
-+
-+ if (!fsl_mc_is_allocatable(mc_dev))
-+ return -EINVAL;
-+
-+ if (mc_dev->resource) {
-+ error = fsl_mc_resource_pool_remove_device(mc_dev);
-+ if (error < 0)
-+ return error;
-+ }
-+
-+ dev_dbg(&mc_dev->dev,
-+ "Allocatable fsl-mc device unbound from fsl_mc_allocator driver");
-+ return 0;
-+}
-+
-+static const struct fsl_mc_device_id match_id_table[] = {
-+ {
-+ .vendor = FSL_MC_VENDOR_FREESCALE,
-+ .obj_type = "dpbp",
-+ },
-+ {
-+ .vendor = FSL_MC_VENDOR_FREESCALE,
-+ .obj_type = "dpmcp",
-+ },
-+ {
-+ .vendor = FSL_MC_VENDOR_FREESCALE,
-+ .obj_type = "dpcon",
-+ },
-+ {.vendor = 0x0},
-+};
-+
-+static struct fsl_mc_driver fsl_mc_allocator_driver = {
-+ .driver = {
-+ .name = "fsl_mc_allocator",
-+ .pm = NULL,
-+ },
-+ .match_id_table = match_id_table,
-+ .probe = fsl_mc_allocator_probe,
-+ .remove = fsl_mc_allocator_remove,
-+};
-+
-+int __init fsl_mc_allocator_driver_init(void)
-+{
-+ return fsl_mc_driver_register(&fsl_mc_allocator_driver);
-+}
-+
-+void fsl_mc_allocator_driver_exit(void)
-+{
-+ fsl_mc_driver_unregister(&fsl_mc_allocator_driver);
-+}
---- a/drivers/staging/fsl-mc/bus/fsl-mc-bus.c
-+++ /dev/null
-@@ -1,920 +0,0 @@
--/*
-- * Freescale Management Complex (MC) bus driver
-- *
-- * Copyright (C) 2014 Freescale Semiconductor, Inc.
-- * Author: German Rivera <German.Rivera@freescale.com>
-- *
-- * This file is licensed under the terms of the GNU General Public
-- * License version 2. This program is licensed "as is" without any
-- * warranty of any kind, whether express or implied.
-- */
--
--#include <linux/module.h>
--#include <linux/of_device.h>
--#include <linux/of_address.h>
--#include <linux/ioport.h>
--#include <linux/slab.h>
--#include <linux/limits.h>
--#include <linux/bitops.h>
--#include <linux/msi.h>
--#include <linux/dma-mapping.h>
--#include "../include/mc-bus.h"
--#include "../include/dpmng.h"
--#include "../include/mc-sys.h"
--
--#include "fsl-mc-private.h"
--#include "dprc-cmd.h"
--
--static struct kmem_cache *mc_dev_cache;
--
--/**
-- * Default DMA mask for devices on a fsl-mc bus
-- */
--#define FSL_MC_DEFAULT_DMA_MASK (~0ULL)
--
--/**
-- * struct fsl_mc - Private data of a "fsl,qoriq-mc" platform device
-- * @root_mc_bus_dev: MC object device representing the root DPRC
-- * @num_translation_ranges: number of entries in addr_translation_ranges
-- * @translation_ranges: array of bus to system address translation ranges
-- */
--struct fsl_mc {
-- struct fsl_mc_device *root_mc_bus_dev;
-- u8 num_translation_ranges;
-- struct fsl_mc_addr_translation_range *translation_ranges;
--};
--
--/**
-- * struct fsl_mc_addr_translation_range - bus to system address translation
-- * range
-- * @mc_region_type: Type of MC region for the range being translated
-- * @start_mc_offset: Start MC offset of the range being translated
-- * @end_mc_offset: MC offset of the first byte after the range (last MC
-- * offset of the range is end_mc_offset - 1)
-- * @start_phys_addr: system physical address corresponding to start_mc_addr
-- */
--struct fsl_mc_addr_translation_range {
-- enum dprc_region_type mc_region_type;
-- u64 start_mc_offset;
-- u64 end_mc_offset;
-- phys_addr_t start_phys_addr;
--};
--
--/**
-- * fsl_mc_bus_match - device to driver matching callback
-- * @dev: the MC object device structure to match against
-- * @drv: the device driver to search for matching MC object device id
-- * structures
-- *
-- * Returns 1 on success, 0 otherwise.
-- */
--static int fsl_mc_bus_match(struct device *dev, struct device_driver *drv)
--{
-- const struct fsl_mc_device_id *id;
-- struct fsl_mc_device *mc_dev = to_fsl_mc_device(dev);
-- struct fsl_mc_driver *mc_drv = to_fsl_mc_driver(drv);
-- bool found = false;
--
-- if (WARN_ON(!fsl_mc_bus_exists()))
-- goto out;
--
-- if (!mc_drv->match_id_table)
-- goto out;
--
-- /*
-- * If the object is not 'plugged' don't match.
-- * Only exception is the root DPRC, which is a special case.
-- */
-- if ((mc_dev->obj_desc.state & DPRC_OBJ_STATE_PLUGGED) == 0 &&
-- !fsl_mc_is_root_dprc(&mc_dev->dev))
-- goto out;
--
-- /*
-- * Traverse the match_id table of the given driver, trying to find
-- * a matching for the given MC object device.
-- */
-- for (id = mc_drv->match_id_table; id->vendor != 0x0; id++) {
-- if (id->vendor == mc_dev->obj_desc.vendor &&
-- strcmp(id->obj_type, mc_dev->obj_desc.type) == 0) {
-- found = true;
--
-- break;
-- }
-- }
--
--out:
-- dev_dbg(dev, "%smatched\n", found ? "" : "not ");
-- return found;
--}
--
--/**
-- * fsl_mc_bus_uevent - callback invoked when a device is added
-- */
--static int fsl_mc_bus_uevent(struct device *dev, struct kobj_uevent_env *env)
--{
-- struct fsl_mc_device *mc_dev = to_fsl_mc_device(dev);
--
-- if (add_uevent_var(env, "MODALIAS=fsl-mc:v%08Xd%s",
-- mc_dev->obj_desc.vendor,
-- mc_dev->obj_desc.type))
-- return -ENOMEM;
--
-- return 0;
--}
--
--static ssize_t modalias_show(struct device *dev, struct device_attribute *attr,
-- char *buf)
--{
-- struct fsl_mc_device *mc_dev = to_fsl_mc_device(dev);
--
-- return sprintf(buf, "fsl-mc:v%08Xd%s\n", mc_dev->obj_desc.vendor,
-- mc_dev->obj_desc.type);
--}
--static DEVICE_ATTR_RO(modalias);
--
--static struct attribute *fsl_mc_dev_attrs[] = {
-- &dev_attr_modalias.attr,
-- NULL,
--};
--
--ATTRIBUTE_GROUPS(fsl_mc_dev);
--
--struct bus_type fsl_mc_bus_type = {
-- .name = "fsl-mc",
-- .match = fsl_mc_bus_match,
-- .uevent = fsl_mc_bus_uevent,
-- .dev_groups = fsl_mc_dev_groups,
--};
--EXPORT_SYMBOL_GPL(fsl_mc_bus_type);
--
--static atomic_t root_dprc_count = ATOMIC_INIT(0);
--
--static int fsl_mc_driver_probe(struct device *dev)
--{
-- struct fsl_mc_driver *mc_drv;
-- struct fsl_mc_device *mc_dev = to_fsl_mc_device(dev);
-- int error;
--
-- if (WARN_ON(!dev->driver))
-- return -EINVAL;
--
-- mc_drv = to_fsl_mc_driver(dev->driver);
-- if (WARN_ON(!mc_drv->probe))
-- return -EINVAL;
--
-- error = mc_drv->probe(mc_dev);
-- if (error < 0) {
-- dev_err(dev, "MC object device probe callback failed: %d\n",
-- error);
-- return error;
-- }
--
-- return 0;
--}
--
--static int fsl_mc_driver_remove(struct device *dev)
--{
-- struct fsl_mc_driver *mc_drv = to_fsl_mc_driver(dev->driver);
-- struct fsl_mc_device *mc_dev = to_fsl_mc_device(dev);
-- int error;
--
-- if (WARN_ON(!dev->driver))
-- return -EINVAL;
--
-- error = mc_drv->remove(mc_dev);
-- if (error < 0) {
-- dev_err(dev,
-- "MC object device remove callback failed: %d\n",
-- error);
-- return error;
-- }
--
-- return 0;
--}
--
--static void fsl_mc_driver_shutdown(struct device *dev)
--{
-- struct fsl_mc_driver *mc_drv = to_fsl_mc_driver(dev->driver);
-- struct fsl_mc_device *mc_dev = to_fsl_mc_device(dev);
--
-- mc_drv->shutdown(mc_dev);
--}
--
--/**
-- * __fsl_mc_driver_register - registers a child device driver with the
-- * MC bus
-- *
-- * This function is implicitly invoked from the registration function of
-- * fsl_mc device drivers, which is generated by the
-- * module_fsl_mc_driver() macro.
-- */
--int __fsl_mc_driver_register(struct fsl_mc_driver *mc_driver,
-- struct module *owner)
--{
-- int error;
--
-- mc_driver->driver.owner = owner;
-- mc_driver->driver.bus = &fsl_mc_bus_type;
--
-- if (mc_driver->probe)
-- mc_driver->driver.probe = fsl_mc_driver_probe;
--
-- if (mc_driver->remove)
-- mc_driver->driver.remove = fsl_mc_driver_remove;
--
-- if (mc_driver->shutdown)
-- mc_driver->driver.shutdown = fsl_mc_driver_shutdown;
--
-- error = driver_register(&mc_driver->driver);
-- if (error < 0) {
-- pr_err("driver_register() failed for %s: %d\n",
-- mc_driver->driver.name, error);
-- return error;
-- }
--
-- pr_info("MC object device driver %s registered\n",
-- mc_driver->driver.name);
-- return 0;
--}
--EXPORT_SYMBOL_GPL(__fsl_mc_driver_register);
--
--/**
-- * fsl_mc_driver_unregister - unregisters a device driver from the
-- * MC bus
-- */
--void fsl_mc_driver_unregister(struct fsl_mc_driver *mc_driver)
--{
-- driver_unregister(&mc_driver->driver);
--}
--EXPORT_SYMBOL_GPL(fsl_mc_driver_unregister);
--
--/**
-- * fsl_mc_bus_exists - check if a root dprc exists
-- */
--bool fsl_mc_bus_exists(void)
--{
-- return atomic_read(&root_dprc_count) > 0;
--}
--EXPORT_SYMBOL_GPL(fsl_mc_bus_exists);
--
--/**
-- * fsl_mc_get_root_dprc - function to traverse to the root dprc
-- */
--void fsl_mc_get_root_dprc(struct device *dev,
-- struct device **root_dprc_dev)
--{
-- if (WARN_ON(!dev)) {
-- *root_dprc_dev = NULL;
-- } else if (WARN_ON(!dev_is_fsl_mc(dev))) {
-- *root_dprc_dev = NULL;
-- } else {
-- *root_dprc_dev = dev;
-- while (dev_is_fsl_mc((*root_dprc_dev)->parent))
-- *root_dprc_dev = (*root_dprc_dev)->parent;
-- }
--}
--EXPORT_SYMBOL_GPL(fsl_mc_get_root_dprc);
--
--static int get_dprc_attr(struct fsl_mc_io *mc_io,
-- int container_id, struct dprc_attributes *attr)
--{
-- u16 dprc_handle;
-- int error;
--
-- error = dprc_open(mc_io, 0, container_id, &dprc_handle);
-- if (error < 0) {
-- dev_err(mc_io->dev, "dprc_open() failed: %d\n", error);
-- return error;
-- }
--
-- memset(attr, 0, sizeof(struct dprc_attributes));
-- error = dprc_get_attributes(mc_io, 0, dprc_handle, attr);
-- if (error < 0) {
-- dev_err(mc_io->dev, "dprc_get_attributes() failed: %d\n",
-- error);
-- goto common_cleanup;
-- }
--
-- error = 0;
--
--common_cleanup:
-- (void)dprc_close(mc_io, 0, dprc_handle);
-- return error;
--}
--
--static int get_dprc_icid(struct fsl_mc_io *mc_io,
-- int container_id, u16 *icid)
--{
-- struct dprc_attributes attr;
-- int error;
--
-- error = get_dprc_attr(mc_io, container_id, &attr);
-- if (error == 0)
-- *icid = attr.icid;
--
-- return error;
--}
--
--static int get_dprc_version(struct fsl_mc_io *mc_io,
-- int container_id, u16 *major, u16 *minor)
--{
-- struct dprc_attributes attr;
-- int error;
--
-- error = get_dprc_attr(mc_io, container_id, &attr);
-- if (error == 0) {
-- *major = attr.version.major;
-- *minor = attr.version.minor;
-- }
--
-- return error;
--}
--
--static int translate_mc_addr(struct fsl_mc_device *mc_dev,
-- enum dprc_region_type mc_region_type,
-- u64 mc_offset, phys_addr_t *phys_addr)
--{
-- int i;
-- struct device *root_dprc_dev;
-- struct fsl_mc *mc;
--
-- fsl_mc_get_root_dprc(&mc_dev->dev, &root_dprc_dev);
-- if (WARN_ON(!root_dprc_dev))
-- return -EINVAL;
-- mc = dev_get_drvdata(root_dprc_dev->parent);
--
-- if (mc->num_translation_ranges == 0) {
-- /*
-- * Do identity mapping:
-- */
-- *phys_addr = mc_offset;
-- return 0;
-- }
--
-- for (i = 0; i < mc->num_translation_ranges; i++) {
-- struct fsl_mc_addr_translation_range *range =
-- &mc->translation_ranges[i];
--
-- if (mc_region_type == range->mc_region_type &&
-- mc_offset >= range->start_mc_offset &&
-- mc_offset < range->end_mc_offset) {
-- *phys_addr = range->start_phys_addr +
-- (mc_offset - range->start_mc_offset);
-- return 0;
-- }
-- }
--
-- return -EFAULT;
--}
--
--static int fsl_mc_device_get_mmio_regions(struct fsl_mc_device *mc_dev,
-- struct fsl_mc_device *mc_bus_dev)
--{
-- int i;
-- int error;
-- struct resource *regions;
-- struct dprc_obj_desc *obj_desc = &mc_dev->obj_desc;
-- struct device *parent_dev = mc_dev->dev.parent;
-- enum dprc_region_type mc_region_type;
--
-- if (strcmp(obj_desc->type, "dprc") == 0 ||
-- strcmp(obj_desc->type, "dpmcp") == 0) {
-- mc_region_type = DPRC_REGION_TYPE_MC_PORTAL;
-- } else if (strcmp(obj_desc->type, "dpio") == 0) {
-- mc_region_type = DPRC_REGION_TYPE_QBMAN_PORTAL;
-- } else {
-- /*
-- * This function should not have been called for this MC object
-- * type, as this object type is not supposed to have MMIO
-- * regions
-- */
-- WARN_ON(true);
-- return -EINVAL;
-- }
--
-- regions = kmalloc_array(obj_desc->region_count,
-- sizeof(regions[0]), GFP_KERNEL);
-- if (!regions)
-- return -ENOMEM;
--
-- for (i = 0; i < obj_desc->region_count; i++) {
-- struct dprc_region_desc region_desc;
--
-- error = dprc_get_obj_region(mc_bus_dev->mc_io,
-- 0,
-- mc_bus_dev->mc_handle,
-- obj_desc->type,
-- obj_desc->id, i, &region_desc);
-- if (error < 0) {
-- dev_err(parent_dev,
-- "dprc_get_obj_region() failed: %d\n", error);
-- goto error_cleanup_regions;
-- }
--
-- WARN_ON(region_desc.size == 0);
-- error = translate_mc_addr(mc_dev, mc_region_type,
-- region_desc.base_offset,
-- &regions[i].start);
-- if (error < 0) {
-- dev_err(parent_dev,
-- "Invalid MC offset: %#x (for %s.%d\'s region %d)\n",
-- region_desc.base_offset,
-- obj_desc->type, obj_desc->id, i);
-- goto error_cleanup_regions;
-- }
--
-- regions[i].end = regions[i].start + region_desc.size - 1;
-- regions[i].name = "fsl-mc object MMIO region";
-- regions[i].flags = IORESOURCE_IO;
-- if (region_desc.flags & DPRC_REGION_CACHEABLE)
-- regions[i].flags |= IORESOURCE_CACHEABLE;
-- }
--
-- mc_dev->regions = regions;
-- return 0;
--
--error_cleanup_regions:
-- kfree(regions);
-- return error;
--}
--
--/**
-- * fsl_mc_is_root_dprc - function to check if a given device is a root dprc
-- */
--bool fsl_mc_is_root_dprc(struct device *dev)
--{
-- struct device *root_dprc_dev;
--
-- fsl_mc_get_root_dprc(dev, &root_dprc_dev);
-- if (!root_dprc_dev)
-- return false;
-- return dev == root_dprc_dev;
--}
--
--/**
-- * Add a newly discovered MC object device to be visible in Linux
-- */
--int fsl_mc_device_add(struct dprc_obj_desc *obj_desc,
-- struct fsl_mc_io *mc_io,
-- struct device *parent_dev,
-- struct fsl_mc_device **new_mc_dev)
--{
-- int error;
-- struct fsl_mc_device *mc_dev = NULL;
-- struct fsl_mc_bus *mc_bus = NULL;
-- struct fsl_mc_device *parent_mc_dev;
--
-- if (dev_is_fsl_mc(parent_dev))
-- parent_mc_dev = to_fsl_mc_device(parent_dev);
-- else
-- parent_mc_dev = NULL;
--
-- if (strcmp(obj_desc->type, "dprc") == 0) {
-- /*
-- * Allocate an MC bus device object:
-- */
-- mc_bus = devm_kzalloc(parent_dev, sizeof(*mc_bus), GFP_KERNEL);
-- if (!mc_bus)
-- return -ENOMEM;
--
-- mc_dev = &mc_bus->mc_dev;
-- } else {
-- /*
-- * Allocate a regular fsl_mc_device object:
-- */
-- mc_dev = kmem_cache_zalloc(mc_dev_cache, GFP_KERNEL);
-- if (!mc_dev)
-- return -ENOMEM;
-- }
--
-- mc_dev->obj_desc = *obj_desc;
-- mc_dev->mc_io = mc_io;
-- device_initialize(&mc_dev->dev);
-- mc_dev->dev.parent = parent_dev;
-- mc_dev->dev.bus = &fsl_mc_bus_type;
-- dev_set_name(&mc_dev->dev, "%s.%d", obj_desc->type, obj_desc->id);
--
-- if (strcmp(obj_desc->type, "dprc") == 0) {
-- struct fsl_mc_io *mc_io2;
--
-- mc_dev->flags |= FSL_MC_IS_DPRC;
--
-- /*
-- * To get the DPRC's ICID, we need to open the DPRC
-- * in get_dprc_icid(). For child DPRCs, we do so using the
-- * parent DPRC's MC portal instead of the child DPRC's MC
-- * portal, in case the child DPRC is already opened with
-- * its own portal (e.g., the DPRC used by AIOP).
-- *
-- * NOTE: There cannot be more than one active open for a
-- * given MC object, using the same MC portal.
-- */
-- if (parent_mc_dev) {
-- /*
-- * device being added is a child DPRC device
-- */
-- mc_io2 = parent_mc_dev->mc_io;
-- } else {
-- /*
-- * device being added is the root DPRC device
-- */
-- if (WARN_ON(!mc_io)) {
-- error = -EINVAL;
-- goto error_cleanup_dev;
-- }
--
-- mc_io2 = mc_io;
--
-- atomic_inc(&root_dprc_count);
-- }
--
-- error = get_dprc_icid(mc_io2, obj_desc->id, &mc_dev->icid);
-- if (error < 0)
-- goto error_cleanup_dev;
-- } else {
-- /*
-- * A non-DPRC MC object device has to be a child of another
-- * MC object (specifically a DPRC object)
-- */
-- mc_dev->icid = parent_mc_dev->icid;
-- mc_dev->dma_mask = FSL_MC_DEFAULT_DMA_MASK;
-- mc_dev->dev.dma_mask = &mc_dev->dma_mask;
-- dev_set_msi_domain(&mc_dev->dev,
-- dev_get_msi_domain(&parent_mc_dev->dev));
-- }
--
-- /*
-- * Get MMIO regions for the device from the MC:
-- *
-- * NOTE: the root DPRC is a special case as its MMIO region is
-- * obtained from the device tree
-- */
-- if (parent_mc_dev && obj_desc->region_count != 0) {
-- error = fsl_mc_device_get_mmio_regions(mc_dev,
-- parent_mc_dev);
-- if (error < 0)
-- goto error_cleanup_dev;
-- }
--
-- /* Objects are coherent, unless 'no shareability' flag set. */
-- if (!(obj_desc->flags & DPRC_OBJ_FLAG_NO_MEM_SHAREABILITY))
-- arch_setup_dma_ops(&mc_dev->dev, 0, 0, NULL, true);
--
-- /*
-- * The device-specific probe callback will get invoked by device_add()
-- */
-- error = device_add(&mc_dev->dev);
-- if (error < 0) {
-- dev_err(parent_dev,
-- "device_add() failed for device %s: %d\n",
-- dev_name(&mc_dev->dev), error);
-- goto error_cleanup_dev;
-- }
--
-- (void)get_device(&mc_dev->dev);
-- dev_dbg(parent_dev, "Added MC object device %s\n",
-- dev_name(&mc_dev->dev));
--
-- *new_mc_dev = mc_dev;
-- return 0;
--
--error_cleanup_dev:
-- kfree(mc_dev->regions);
-- if (mc_bus)
-- devm_kfree(parent_dev, mc_bus);
-- else
-- kmem_cache_free(mc_dev_cache, mc_dev);
--
-- return error;
--}
--EXPORT_SYMBOL_GPL(fsl_mc_device_add);
--
--/**
-- * fsl_mc_device_remove - Remove a MC object device from being visible to
-- * Linux
-- *
-- * @mc_dev: Pointer to a MC object device object
-- */
--void fsl_mc_device_remove(struct fsl_mc_device *mc_dev)
--{
-- struct fsl_mc_bus *mc_bus = NULL;
--
-- kfree(mc_dev->regions);
--
-- /*
-- * The device-specific remove callback will get invoked by device_del()
-- */
-- device_del(&mc_dev->dev);
-- put_device(&mc_dev->dev);
--
-- if (strcmp(mc_dev->obj_desc.type, "dprc") == 0) {
-- mc_bus = to_fsl_mc_bus(mc_dev);
--
-- if (fsl_mc_is_root_dprc(&mc_dev->dev)) {
-- if (atomic_read(&root_dprc_count) > 0)
-- atomic_dec(&root_dprc_count);
-- else
-- WARN_ON(1);
-- }
-- }
--
-- if (mc_bus)
-- devm_kfree(mc_dev->dev.parent, mc_bus);
-- else
-- kmem_cache_free(mc_dev_cache, mc_dev);
--}
--EXPORT_SYMBOL_GPL(fsl_mc_device_remove);
--
--static int parse_mc_ranges(struct device *dev,
-- int *paddr_cells,
-- int *mc_addr_cells,
-- int *mc_size_cells,
-- const __be32 **ranges_start,
-- u8 *num_ranges)
--{
-- const __be32 *prop;
-- int range_tuple_cell_count;
-- int ranges_len;
-- int tuple_len;
-- struct device_node *mc_node = dev->of_node;
--
-- *ranges_start = of_get_property(mc_node, "ranges", &ranges_len);
-- if (!(*ranges_start) || !ranges_len) {
-- dev_warn(dev,
-- "missing or empty ranges property for device tree node '%s'\n",
-- mc_node->name);
--
-- *num_ranges = 0;
-- return 0;
-- }
--
-- *paddr_cells = of_n_addr_cells(mc_node);
--
-- prop = of_get_property(mc_node, "#address-cells", NULL);
-- if (prop)
-- *mc_addr_cells = be32_to_cpup(prop);
-- else
-- *mc_addr_cells = *paddr_cells;
--
-- prop = of_get_property(mc_node, "#size-cells", NULL);
-- if (prop)
-- *mc_size_cells = be32_to_cpup(prop);
-- else
-- *mc_size_cells = of_n_size_cells(mc_node);
--
-- range_tuple_cell_count = *paddr_cells + *mc_addr_cells +
-- *mc_size_cells;
--
-- tuple_len = range_tuple_cell_count * sizeof(__be32);
-- if (ranges_len % tuple_len != 0) {
-- dev_err(dev, "malformed ranges property '%s'\n", mc_node->name);
-- return -EINVAL;
-- }
--
-- *num_ranges = ranges_len / tuple_len;
-- return 0;
--}
--
--static int get_mc_addr_translation_ranges(struct device *dev,
-- struct fsl_mc_addr_translation_range
-- **ranges,
-- u8 *num_ranges)
--{
-- int error;
-- int paddr_cells;
-- int mc_addr_cells;
-- int mc_size_cells;
-- int i;
-- const __be32 *ranges_start;
-- const __be32 *cell;
--
-- error = parse_mc_ranges(dev,
-- &paddr_cells,
-- &mc_addr_cells,
-- &mc_size_cells,
-- &ranges_start,
-- num_ranges);
-- if (error < 0)
-- return error;
--
-- if (!(*num_ranges)) {
-- /*
-- * Missing or empty ranges property ("ranges;") for the
-- * 'fsl,qoriq-mc' node. In this case, identity mapping
-- * will be used.
-- */
-- *ranges = NULL;
-- return 0;
-- }
--
-- *ranges = devm_kcalloc(dev, *num_ranges,
-- sizeof(struct fsl_mc_addr_translation_range),
-- GFP_KERNEL);
-- if (!(*ranges))
-- return -ENOMEM;
--
-- cell = ranges_start;
-- for (i = 0; i < *num_ranges; ++i) {
-- struct fsl_mc_addr_translation_range *range = &(*ranges)[i];
--
-- range->mc_region_type = of_read_number(cell, 1);
-- range->start_mc_offset = of_read_number(cell + 1,
-- mc_addr_cells - 1);
-- cell += mc_addr_cells;
-- range->start_phys_addr = of_read_number(cell, paddr_cells);
-- cell += paddr_cells;
-- range->end_mc_offset = range->start_mc_offset +
-- of_read_number(cell, mc_size_cells);
--
-- cell += mc_size_cells;
-- }
--
-- return 0;
--}
--
--/**
-- * fsl_mc_bus_probe - callback invoked when the root MC bus is being
-- * added
-- */
--static int fsl_mc_bus_probe(struct platform_device *pdev)
--{
-- struct dprc_obj_desc obj_desc;
-- int error;
-- struct fsl_mc *mc;
-- struct fsl_mc_device *mc_bus_dev = NULL;
-- struct fsl_mc_io *mc_io = NULL;
-- int container_id;
-- phys_addr_t mc_portal_phys_addr;
-- u32 mc_portal_size;
-- struct mc_version mc_version;
-- struct resource res;
--
-- dev_info(&pdev->dev, "Root MC bus device probed");
--
-- mc = devm_kzalloc(&pdev->dev, sizeof(*mc), GFP_KERNEL);
-- if (!mc)
-- return -ENOMEM;
--
-- platform_set_drvdata(pdev, mc);
--
-- /*
-- * Get physical address of MC portal for the root DPRC:
-- */
-- error = of_address_to_resource(pdev->dev.of_node, 0, &res);
-- if (error < 0) {
-- dev_err(&pdev->dev,
-- "of_address_to_resource() failed for %s\n",
-- pdev->dev.of_node->full_name);
-- return error;
-- }
--
-- mc_portal_phys_addr = res.start;
-- mc_portal_size = resource_size(&res);
-- error = fsl_create_mc_io(&pdev->dev, mc_portal_phys_addr,
-- mc_portal_size, NULL,
-- FSL_MC_IO_ATOMIC_CONTEXT_PORTAL, &mc_io);
-- if (error < 0)
-- return error;
--
-- error = mc_get_version(mc_io, 0, &mc_version);
-- if (error != 0) {
-- dev_err(&pdev->dev,
-- "mc_get_version() failed with error %d\n", error);
-- goto error_cleanup_mc_io;
-- }
--
-- dev_info(&pdev->dev,
-- "Freescale Management Complex Firmware version: %u.%u.%u\n",
-- mc_version.major, mc_version.minor, mc_version.revision);
--
-- error = get_mc_addr_translation_ranges(&pdev->dev,
-- &mc->translation_ranges,
-- &mc->num_translation_ranges);
-- if (error < 0)
-- goto error_cleanup_mc_io;
--
-- error = dpmng_get_container_id(mc_io, 0, &container_id);
-- if (error < 0) {
-- dev_err(&pdev->dev,
-- "dpmng_get_container_id() failed: %d\n", error);
-- goto error_cleanup_mc_io;
-- }
--
-- memset(&obj_desc, 0, sizeof(struct dprc_obj_desc));
-- error = get_dprc_version(mc_io, container_id,
-- &obj_desc.ver_major, &obj_desc.ver_minor);
-- if (error < 0)
-- goto error_cleanup_mc_io;
--
-- obj_desc.vendor = FSL_MC_VENDOR_FREESCALE;
-- strcpy(obj_desc.type, "dprc");
-- obj_desc.id = container_id;
-- obj_desc.irq_count = 1;
-- obj_desc.region_count = 0;
--
-- error = fsl_mc_device_add(&obj_desc, mc_io, &pdev->dev, &mc_bus_dev);
-- if (error < 0)
-- goto error_cleanup_mc_io;
--
-- mc->root_mc_bus_dev = mc_bus_dev;
-- return 0;
--
--error_cleanup_mc_io:
-- fsl_destroy_mc_io(mc_io);
-- return error;
--}
--
--/**
-- * fsl_mc_bus_remove - callback invoked when the root MC bus is being
-- * removed
-- */
--static int fsl_mc_bus_remove(struct platform_device *pdev)
--{
-- struct fsl_mc *mc = platform_get_drvdata(pdev);
--
-- if (WARN_ON(!fsl_mc_is_root_dprc(&mc->root_mc_bus_dev->dev)))
-- return -EINVAL;
--
-- fsl_mc_device_remove(mc->root_mc_bus_dev);
--
-- fsl_destroy_mc_io(mc->root_mc_bus_dev->mc_io);
-- mc->root_mc_bus_dev->mc_io = NULL;
--
-- dev_info(&pdev->dev, "Root MC bus device removed");
-- return 0;
--}
--
--static const struct of_device_id fsl_mc_bus_match_table[] = {
-- {.compatible = "fsl,qoriq-mc",},
-- {},
--};
--
--MODULE_DEVICE_TABLE(of, fsl_mc_bus_match_table);
--
--static struct platform_driver fsl_mc_bus_driver = {
-- .driver = {
-- .name = "fsl_mc_bus",
-- .pm = NULL,
-- .of_match_table = fsl_mc_bus_match_table,
-- },
-- .probe = fsl_mc_bus_probe,
-- .remove = fsl_mc_bus_remove,
--};
--
--static int __init fsl_mc_bus_driver_init(void)
--{
-- int error;
--
-- mc_dev_cache = kmem_cache_create("fsl_mc_device",
-- sizeof(struct fsl_mc_device), 0, 0,
-- NULL);
-- if (!mc_dev_cache) {
-- pr_err("Could not create fsl_mc_device cache\n");
-- return -ENOMEM;
-- }
--
-- error = bus_register(&fsl_mc_bus_type);
-- if (error < 0) {
-- pr_err("fsl-mc bus type registration failed: %d\n", error);
-- goto error_cleanup_cache;
-- }
--
-- pr_info("fsl-mc bus type registered\n");
--
-- error = platform_driver_register(&fsl_mc_bus_driver);
-- if (error < 0) {
-- pr_err("platform_driver_register() failed: %d\n", error);
-- goto error_cleanup_bus;
-- }
--
-- error = dprc_driver_init();
-- if (error < 0)
-- goto error_cleanup_driver;
--
-- error = fsl_mc_allocator_driver_init();
-- if (error < 0)
-- goto error_cleanup_dprc_driver;
--
-- error = its_fsl_mc_msi_init();
-- if (error < 0)
-- goto error_cleanup_mc_allocator;
--
-- return 0;
--
--error_cleanup_mc_allocator:
-- fsl_mc_allocator_driver_exit();
--
--error_cleanup_dprc_driver:
-- dprc_driver_exit();
--
--error_cleanup_driver:
-- platform_driver_unregister(&fsl_mc_bus_driver);
--
--error_cleanup_bus:
-- bus_unregister(&fsl_mc_bus_type);
--
--error_cleanup_cache:
-- kmem_cache_destroy(mc_dev_cache);
-- return error;
--}
--postcore_initcall(fsl_mc_bus_driver_init);
---- /dev/null
-+++ b/drivers/bus/fsl-mc/fsl-mc-bus.c
-@@ -0,0 +1,1151 @@
-+// SPDX-License-Identifier: GPL-2.0
-+/*
-+ * Freescale Management Complex (MC) bus driver
-+ *
-+ * Copyright (C) 2014-2016 Freescale Semiconductor, Inc.
-+ * Author: German Rivera <German.Rivera@freescale.com>
-+ *
-+ */
-+
-+#define pr_fmt(fmt) "fsl-mc: " fmt
-+
-+#include <linux/module.h>
-+#include <linux/of_device.h>
-+#include <linux/of_address.h>
-+#include <linux/ioport.h>
-+#include <linux/slab.h>
-+#include <linux/limits.h>
-+#include <linux/bitops.h>
-+#include <linux/msi.h>
-+#include <linux/dma-mapping.h>
-+#include <linux/fsl/mc.h>
-+
-+#include "fsl-mc-private.h"
-+
-+/**
-+ * Default DMA mask for devices on a fsl-mc bus
-+ */
-+#define FSL_MC_DEFAULT_DMA_MASK (~0ULL)
-+
-+/**
-+ * struct fsl_mc - Private data of a "fsl,qoriq-mc" platform device
-+ * @root_mc_bus_dev: fsl-mc device representing the root DPRC
-+ * @num_translation_ranges: number of entries in addr_translation_ranges
-+ * @translation_ranges: array of bus to system address translation ranges
-+ */
-+struct fsl_mc {
-+ struct fsl_mc_device *root_mc_bus_dev;
-+ u8 num_translation_ranges;
-+ struct fsl_mc_addr_translation_range *translation_ranges;
-+};
-+
-+/**
-+ * struct fsl_mc_addr_translation_range - bus to system address translation
-+ * range
-+ * @mc_region_type: Type of MC region for the range being translated
-+ * @start_mc_offset: Start MC offset of the range being translated
-+ * @end_mc_offset: MC offset of the first byte after the range (last MC
-+ * offset of the range is end_mc_offset - 1)
-+ * @start_phys_addr: system physical address corresponding to start_mc_addr
-+ */
-+struct fsl_mc_addr_translation_range {
-+ enum dprc_region_type mc_region_type;
-+ u64 start_mc_offset;
-+ u64 end_mc_offset;
-+ phys_addr_t start_phys_addr;
-+};
-+
-+/**
-+ * struct mc_version
-+ * @major: Major version number: incremented on API compatibility changes
-+ * @minor: Minor version number: incremented on API additions (that are
-+ * backward compatible); reset when major version is incremented
-+ * @revision: Internal revision number: incremented on implementation changes
-+ * and/or bug fixes that have no impact on API
-+ */
-+struct mc_version {
-+ u32 major;
-+ u32 minor;
-+ u32 revision;
-+};
-+
-+/**
-+ * fsl_mc_bus_match - device to driver matching callback
-+ * @dev: the fsl-mc device to match against
-+ * @drv: the device driver to search for matching fsl-mc object type
-+ * structures
-+ *
-+ * Returns 1 on success, 0 otherwise.
-+ */
-+static int fsl_mc_bus_match(struct device *dev, struct device_driver *drv)
-+{
-+ const struct fsl_mc_device_id *id;
-+ struct fsl_mc_device *mc_dev = to_fsl_mc_device(dev);
-+ struct fsl_mc_driver *mc_drv = to_fsl_mc_driver(drv);
-+ bool found = false;
-+
-+ /* When driver_override is set, only bind to the matching driver */
-+ if (mc_dev->driver_override) {
-+ found = !strcmp(mc_dev->driver_override, mc_drv->driver.name);
-+ goto out;
-+ }
-+
-+ if (!mc_drv->match_id_table)
-+ goto out;
-+
-+ /*
-+ * If the object is not 'plugged' don't match.
-+ * Only exception is the root DPRC, which is a special case.
-+ */
-+ if ((mc_dev->obj_desc.state & FSL_MC_OBJ_STATE_PLUGGED) == 0 &&
-+ !fsl_mc_is_root_dprc(&mc_dev->dev))
-+ goto out;
-+
-+ /*
-+ * Traverse the match_id table of the given driver, trying to find
-+ * a matching for the given device.
-+ */
-+ for (id = mc_drv->match_id_table; id->vendor != 0x0; id++) {
-+ if (id->vendor == mc_dev->obj_desc.vendor &&
-+ strcmp(id->obj_type, mc_dev->obj_desc.type) == 0) {
-+ found = true;
-+
-+ break;
-+ }
-+ }
-+
-+out:
-+ dev_dbg(dev, "%smatched\n", found ? "" : "not ");
-+ return found;
-+}
-+
-+/**
-+ * fsl_mc_bus_uevent - callback invoked when a device is added
-+ */
-+static int fsl_mc_bus_uevent(struct device *dev, struct kobj_uevent_env *env)
-+{
-+ struct fsl_mc_device *mc_dev = to_fsl_mc_device(dev);
-+
-+ if (add_uevent_var(env, "MODALIAS=fsl-mc:v%08Xd%s",
-+ mc_dev->obj_desc.vendor,
-+ mc_dev->obj_desc.type))
-+ return -ENOMEM;
-+
-+ return 0;
-+}
-+
-+static ssize_t modalias_show(struct device *dev, struct device_attribute *attr,
-+ char *buf)
-+{
-+ struct fsl_mc_device *mc_dev = to_fsl_mc_device(dev);
-+
-+ return sprintf(buf, "fsl-mc:v%08Xd%s\n", mc_dev->obj_desc.vendor,
-+ mc_dev->obj_desc.type);
-+}
-+static DEVICE_ATTR_RO(modalias);
-+
-+static ssize_t rescan_store(struct device *dev,
-+ struct device_attribute *attr,
-+ const char *buf, size_t count)
-+{
-+ struct fsl_mc_device *root_mc_dev;
-+ struct fsl_mc_bus *root_mc_bus;
-+ unsigned long val;
-+
-+ if (!fsl_mc_is_root_dprc(dev))
-+ return -EINVAL;
-+
-+ root_mc_dev = to_fsl_mc_device(dev);
-+ root_mc_bus = to_fsl_mc_bus(root_mc_dev);
-+
-+ if (kstrtoul(buf, 0, &val) < 0)
-+ return -EINVAL;
-+
-+ if (val) {
-+ mutex_lock(&root_mc_bus->scan_mutex);
-+ dprc_scan_objects(root_mc_dev, NULL, NULL);
-+ mutex_unlock(&root_mc_bus->scan_mutex);
-+ }
-+
-+ return count;
-+}
-+static DEVICE_ATTR_WO(rescan);
-+
-+static ssize_t driver_override_store(struct device *dev,
-+ struct device_attribute *attr,
-+ const char *buf, size_t count)
-+{
-+ struct fsl_mc_device *mc_dev = to_fsl_mc_device(dev);
-+ const char *driver_override, *old = mc_dev->driver_override;
-+ char *cp;
-+
-+ if (WARN_ON(dev->bus != &fsl_mc_bus_type))
-+ return -EINVAL;
-+
-+ if (count >= (PAGE_SIZE - 1))
-+ return -EINVAL;
-+
-+ driver_override = kstrndup(buf, count, GFP_KERNEL);
-+ if (!driver_override)
-+ return -ENOMEM;
-+
-+ cp = strchr(driver_override, '\n');
-+ if (cp)
-+ *cp = '\0';
-+
-+ if (strlen(driver_override)) {
-+ mc_dev->driver_override = driver_override;
-+ } else {
-+ kfree(driver_override);
-+ mc_dev->driver_override = NULL;
-+ }
-+
-+ kfree(old);
-+
-+ return count;
-+}
-+
-+static ssize_t driver_override_show(struct device *dev,
-+ struct device_attribute *attr, char *buf)
-+{
-+ struct fsl_mc_device *mc_dev = to_fsl_mc_device(dev);
-+
-+ return snprintf(buf, PAGE_SIZE, "%s\n", mc_dev->driver_override);
-+}
-+static DEVICE_ATTR_RW(driver_override);
-+
-+static struct attribute *fsl_mc_dev_attrs[] = {
-+ &dev_attr_modalias.attr,
-+ &dev_attr_rescan.attr,
-+ &dev_attr_driver_override.attr,
-+ NULL,
-+};
-+
-+ATTRIBUTE_GROUPS(fsl_mc_dev);
-+
-+static int scan_fsl_mc_bus(struct device *dev, void *data)
-+{
-+ struct fsl_mc_device *root_mc_dev;
-+ struct fsl_mc_bus *root_mc_bus;
-+
-+ if (!fsl_mc_is_root_dprc(dev))
-+ goto exit;
-+
-+ root_mc_dev = to_fsl_mc_device(dev);
-+ root_mc_bus = to_fsl_mc_bus(root_mc_dev);
-+ mutex_lock(&root_mc_bus->scan_mutex);
-+ dprc_scan_objects(root_mc_dev, NULL, NULL);
-+ mutex_unlock(&root_mc_bus->scan_mutex);
-+
-+exit:
-+ return 0;
-+}
-+
-+static ssize_t bus_rescan_store(struct bus_type *bus,
-+ const char *buf, size_t count)
-+{
-+ unsigned long val;
-+
-+ if (kstrtoul(buf, 0, &val) < 0)
-+ return -EINVAL;
-+
-+ if (val)
-+ bus_for_each_dev(bus, NULL, NULL, scan_fsl_mc_bus);
-+
-+ return count;
-+}
-+static BUS_ATTR(rescan, 0220, NULL, bus_rescan_store);
-+
-+static struct attribute *fsl_mc_bus_attrs[] = {
-+ &bus_attr_rescan.attr,
-+ NULL,
-+};
-+
-+static const struct attribute_group fsl_mc_bus_group = {
-+ .attrs = fsl_mc_bus_attrs,
-+};
-+
-+static const struct attribute_group *fsl_mc_bus_groups[] = {
-+ &fsl_mc_bus_group,
-+ NULL,
-+};
-+
-+struct bus_type fsl_mc_bus_type = {
-+ .name = "fsl-mc",
-+ .match = fsl_mc_bus_match,
-+ .uevent = fsl_mc_bus_uevent,
-+ .dev_groups = fsl_mc_dev_groups,
-+ .bus_groups = fsl_mc_bus_groups,
-+};
-+EXPORT_SYMBOL_GPL(fsl_mc_bus_type);
-+
-+struct device_type fsl_mc_bus_dprc_type = {
-+ .name = "fsl_mc_bus_dprc"
-+};
-+
-+struct device_type fsl_mc_bus_dpni_type = {
-+ .name = "fsl_mc_bus_dpni"
-+};
-+
-+struct device_type fsl_mc_bus_dpio_type = {
-+ .name = "fsl_mc_bus_dpio"
-+};
-+
-+struct device_type fsl_mc_bus_dpsw_type = {
-+ .name = "fsl_mc_bus_dpsw"
-+};
-+
-+struct device_type fsl_mc_bus_dpdmux_type = {
-+ .name = "fsl_mc_bus_dpdmux"
-+};
-+
-+struct device_type fsl_mc_bus_dpbp_type = {
-+ .name = "fsl_mc_bus_dpbp"
-+};
-+
-+struct device_type fsl_mc_bus_dpcon_type = {
-+ .name = "fsl_mc_bus_dpcon"
-+};
-+
-+struct device_type fsl_mc_bus_dpmcp_type = {
-+ .name = "fsl_mc_bus_dpmcp"
-+};
-+
-+struct device_type fsl_mc_bus_dpmac_type = {
-+ .name = "fsl_mc_bus_dpmac"
-+};
-+
-+struct device_type fsl_mc_bus_dprtc_type = {
-+ .name = "fsl_mc_bus_dprtc"
-+};
-+
-+struct device_type fsl_mc_bus_dpseci_type = {
-+ .name = "fsl_mc_bus_dpseci"
-+};
-+
-+struct device_type fsl_mc_bus_dpdcei_type = {
-+ .name = "fsl_mc_bus_dpdcei"
-+};
-+
-+struct device_type fsl_mc_bus_dpaiop_type = {
-+ .name = "fsl_mc_bus_dpaiop"
-+};
-+
-+struct device_type fsl_mc_bus_dpci_type = {
-+ .name = "fsl_mc_bus_dpci"
-+};
-+
-+struct device_type fsl_mc_bus_dpdmai_type = {
-+ .name = "fsl_mc_bus_dpdmai"
-+};
-+
-+static struct device_type *fsl_mc_get_device_type(const char *type)
-+{
-+ static const struct {
-+ struct device_type *dev_type;
-+ const char *type;
-+ } dev_types[] = {
-+ { &fsl_mc_bus_dprc_type, "dprc" },
-+ { &fsl_mc_bus_dpni_type, "dpni" },
-+ { &fsl_mc_bus_dpio_type, "dpio" },
-+ { &fsl_mc_bus_dpsw_type, "dpsw" },
-+ { &fsl_mc_bus_dpdmux_type, "dpdmux" },
-+ { &fsl_mc_bus_dpbp_type, "dpbp" },
-+ { &fsl_mc_bus_dpcon_type, "dpcon" },
-+ { &fsl_mc_bus_dpmcp_type, "dpmcp" },
-+ { &fsl_mc_bus_dpmac_type, "dpmac" },
-+ { &fsl_mc_bus_dprtc_type, "dprtc" },
-+ { &fsl_mc_bus_dpseci_type, "dpseci" },
-+ { &fsl_mc_bus_dpdcei_type, "dpdcei" },
-+ { &fsl_mc_bus_dpaiop_type, "dpaiop" },
-+ { &fsl_mc_bus_dpci_type, "dpci" },
-+ { &fsl_mc_bus_dpdmai_type, "dpdmai" },
-+ { NULL, NULL }
-+ };
-+ int i;
-+
-+ for (i = 0; dev_types[i].dev_type; i++)
-+ if (!strcmp(dev_types[i].type, type))
-+ return dev_types[i].dev_type;
-+
-+ return NULL;
-+}
-+
-+static int fsl_mc_driver_probe(struct device *dev)
-+{
-+ struct fsl_mc_driver *mc_drv;
-+ struct fsl_mc_device *mc_dev = to_fsl_mc_device(dev);
-+ int error;
-+
-+ mc_drv = to_fsl_mc_driver(dev->driver);
-+
-+ error = mc_drv->probe(mc_dev);
-+ if (error < 0) {
-+ if (error != -EPROBE_DEFER)
-+ dev_err(dev, "%s failed: %d\n", __func__, error);
-+ return error;
-+ }
-+
-+ return 0;
-+}
-+
-+static int fsl_mc_driver_remove(struct device *dev)
-+{
-+ struct fsl_mc_driver *mc_drv = to_fsl_mc_driver(dev->driver);
-+ struct fsl_mc_device *mc_dev = to_fsl_mc_device(dev);
-+ int error;
-+
-+ error = mc_drv->remove(mc_dev);
-+ if (error < 0) {
-+ dev_err(dev, "%s failed: %d\n", __func__, error);
-+ return error;
-+ }
-+
-+ return 0;
-+}
-+
-+static void fsl_mc_driver_shutdown(struct device *dev)
-+{
-+ struct fsl_mc_driver *mc_drv = to_fsl_mc_driver(dev->driver);
-+ struct fsl_mc_device *mc_dev = to_fsl_mc_device(dev);
-+
-+ mc_drv->shutdown(mc_dev);
-+}
-+
-+/**
-+ * __fsl_mc_driver_register - registers a child device driver with the
-+ * MC bus
-+ *
-+ * This function is implicitly invoked from the registration function of
-+ * fsl_mc device drivers, which is generated by the
-+ * module_fsl_mc_driver() macro.
-+ */
-+int __fsl_mc_driver_register(struct fsl_mc_driver *mc_driver,
-+ struct module *owner)
-+{
-+ int error;
-+
-+ mc_driver->driver.owner = owner;
-+ mc_driver->driver.bus = &fsl_mc_bus_type;
-+
-+ if (mc_driver->probe)
-+ mc_driver->driver.probe = fsl_mc_driver_probe;
-+
-+ if (mc_driver->remove)
-+ mc_driver->driver.remove = fsl_mc_driver_remove;
-+
-+ if (mc_driver->shutdown)
-+ mc_driver->driver.shutdown = fsl_mc_driver_shutdown;
-+
-+ error = driver_register(&mc_driver->driver);
-+ if (error < 0) {
-+ pr_err("driver_register() failed for %s: %d\n",
-+ mc_driver->driver.name, error);
-+ return error;
-+ }
-+
-+ return 0;
-+}
-+EXPORT_SYMBOL_GPL(__fsl_mc_driver_register);
-+
-+/**
-+ * fsl_mc_driver_unregister - unregisters a device driver from the
-+ * MC bus
-+ */
-+void fsl_mc_driver_unregister(struct fsl_mc_driver *mc_driver)
-+{
-+ driver_unregister(&mc_driver->driver);
-+}
-+EXPORT_SYMBOL_GPL(fsl_mc_driver_unregister);
-+
-+/**
-+ * mc_get_version() - Retrieves the Management Complex firmware
-+ * version information
-+ * @mc_io: Pointer to opaque I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @mc_ver_info: Returned version information structure
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+static int mc_get_version(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ struct mc_version *mc_ver_info)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpmng_rsp_get_version *rsp_params;
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPMNG_CMDID_GET_VERSION,
-+ cmd_flags,
-+ 0);
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ rsp_params = (struct dpmng_rsp_get_version *)cmd.params;
-+ mc_ver_info->revision = le32_to_cpu(rsp_params->revision);
-+ mc_ver_info->major = le32_to_cpu(rsp_params->version_major);
-+ mc_ver_info->minor = le32_to_cpu(rsp_params->version_minor);
-+
-+ return 0;
-+}
-+
-+/**
-+ * fsl_mc_get_root_dprc - function to traverse to the root dprc
-+ */
-+void fsl_mc_get_root_dprc(struct device *dev,
-+ struct device **root_dprc_dev)
-+{
-+ if (!dev) {
-+ *root_dprc_dev = NULL;
-+ } else if (!dev_is_fsl_mc(dev)) {
-+ *root_dprc_dev = NULL;
-+ } else {
-+ *root_dprc_dev = dev;
-+ while (dev_is_fsl_mc((*root_dprc_dev)->parent))
-+ *root_dprc_dev = (*root_dprc_dev)->parent;
-+ }
-+}
-+EXPORT_SYMBOL_GPL(fsl_mc_get_root_dprc);
-+
-+static int get_dprc_attr(struct fsl_mc_io *mc_io,
-+ int container_id, struct dprc_attributes *attr)
-+{
-+ u16 dprc_handle;
-+ int error;
-+
-+ error = dprc_open(mc_io, 0, container_id, &dprc_handle);
-+ if (error < 0) {
-+ dev_err(mc_io->dev, "dprc_open() failed: %d\n", error);
-+ return error;
-+ }
-+
-+ memset(attr, 0, sizeof(struct dprc_attributes));
-+ error = dprc_get_attributes(mc_io, 0, dprc_handle, attr);
-+ if (error < 0) {
-+ dev_err(mc_io->dev, "dprc_get_attributes() failed: %d\n",
-+ error);
-+ goto common_cleanup;
-+ }
-+
-+ error = 0;
-+
-+common_cleanup:
-+ (void)dprc_close(mc_io, 0, dprc_handle);
-+ return error;
-+}
-+
-+static int get_dprc_icid(struct fsl_mc_io *mc_io,
-+ int container_id, u32 *icid)
-+{
-+ struct dprc_attributes attr;
-+ int error;
-+
-+ error = get_dprc_attr(mc_io, container_id, &attr);
-+ if (error == 0)
-+ *icid = attr.icid;
-+
-+ return error;
-+}
-+
-+static int translate_mc_addr(struct fsl_mc_device *mc_dev,
-+ enum dprc_region_type mc_region_type,
-+ u64 mc_offset, phys_addr_t *phys_addr)
-+{
-+ int i;
-+ struct device *root_dprc_dev;
-+ struct fsl_mc *mc;
-+
-+ fsl_mc_get_root_dprc(&mc_dev->dev, &root_dprc_dev);
-+ mc = dev_get_drvdata(root_dprc_dev->parent);
-+
-+ if (mc->num_translation_ranges == 0) {
-+ /*
-+ * Do identity mapping:
-+ */
-+ *phys_addr = mc_offset;
-+ return 0;
-+ }
-+
-+ for (i = 0; i < mc->num_translation_ranges; i++) {
-+ struct fsl_mc_addr_translation_range *range =
-+ &mc->translation_ranges[i];
-+
-+ if (mc_region_type == range->mc_region_type &&
-+ mc_offset >= range->start_mc_offset &&
-+ mc_offset < range->end_mc_offset) {
-+ *phys_addr = range->start_phys_addr +
-+ (mc_offset - range->start_mc_offset);
-+ return 0;
-+ }
-+ }
-+
-+ return -EFAULT;
-+}
-+
-+static int fsl_mc_device_get_mmio_regions(struct fsl_mc_device *mc_dev,
-+ struct fsl_mc_device *mc_bus_dev)
-+{
-+ int i;
-+ int error;
-+ struct resource *regions;
-+ struct fsl_mc_obj_desc *obj_desc = &mc_dev->obj_desc;
-+ struct device *parent_dev = mc_dev->dev.parent;
-+ enum dprc_region_type mc_region_type;
-+
-+ if (is_fsl_mc_bus_dprc(mc_dev) ||
-+ is_fsl_mc_bus_dpmcp(mc_dev)) {
-+ mc_region_type = DPRC_REGION_TYPE_MC_PORTAL;
-+ } else if (is_fsl_mc_bus_dpio(mc_dev)) {
-+ mc_region_type = DPRC_REGION_TYPE_QBMAN_PORTAL;
-+ } else {
-+ /*
-+ * This function should not have been called for this MC object
-+ * type, as this object type is not supposed to have MMIO
-+ * regions
-+ */
-+ return -EINVAL;
-+ }
-+
-+ regions = kmalloc_array(obj_desc->region_count,
-+ sizeof(regions[0]), GFP_KERNEL);
-+ if (!regions)
-+ return -ENOMEM;
-+
-+ for (i = 0; i < obj_desc->region_count; i++) {
-+ struct dprc_region_desc region_desc;
-+
-+ error = dprc_get_obj_region(mc_bus_dev->mc_io,
-+ 0,
-+ mc_bus_dev->mc_handle,
-+ obj_desc->type,
-+ obj_desc->id, i, &region_desc);
-+ if (error < 0) {
-+ dev_err(parent_dev,
-+ "dprc_get_obj_region() failed: %d\n", error);
-+ goto error_cleanup_regions;
-+ }
-+
-+ error = translate_mc_addr(mc_dev, mc_region_type,
-+ region_desc.base_offset,
-+ &regions[i].start);
-+ if (error < 0) {
-+ dev_err(parent_dev,
-+ "Invalid MC offset: %#x (for %s.%d\'s region %d)\n",
-+ region_desc.base_offset,
-+ obj_desc->type, obj_desc->id, i);
-+ goto error_cleanup_regions;
-+ }
-+
-+ regions[i].end = regions[i].start + region_desc.size - 1;
-+ regions[i].name = "fsl-mc object MMIO region";
-+ regions[i].flags = IORESOURCE_IO;
-+ if (region_desc.flags & DPRC_REGION_CACHEABLE)
-+ regions[i].flags |= IORESOURCE_CACHEABLE;
-+ }
-+
-+ mc_dev->regions = regions;
-+ return 0;
-+
-+error_cleanup_regions:
-+ kfree(regions);
-+ return error;
-+}
-+
-+/**
-+ * fsl_mc_is_root_dprc - function to check if a given device is a root dprc
-+ */
-+bool fsl_mc_is_root_dprc(struct device *dev)
-+{
-+ struct device *root_dprc_dev;
-+
-+ fsl_mc_get_root_dprc(dev, &root_dprc_dev);
-+ if (!root_dprc_dev)
-+ return false;
-+ return dev == root_dprc_dev;
-+}
-+
-+static void fsl_mc_device_release(struct device *dev)
-+{
-+ struct fsl_mc_device *mc_dev = to_fsl_mc_device(dev);
-+
-+ kfree(mc_dev->regions);
-+
-+ if (is_fsl_mc_bus_dprc(mc_dev))
-+ kfree(to_fsl_mc_bus(mc_dev));
-+ else
-+ kfree(mc_dev);
-+}
-+
-+/**
-+ * Add a newly discovered fsl-mc device to be visible in Linux
-+ */
-+int fsl_mc_device_add(struct fsl_mc_obj_desc *obj_desc,
-+ struct fsl_mc_io *mc_io,
-+ struct device *parent_dev,
-+ const char *driver_override,
-+ struct fsl_mc_device **new_mc_dev)
-+{
-+ int error;
-+ struct fsl_mc_device *mc_dev = NULL;
-+ struct fsl_mc_bus *mc_bus = NULL;
-+ struct fsl_mc_device *parent_mc_dev;
-+ struct device *fsl_mc_platform_dev;
-+ struct device_node *fsl_mc_platform_node;
-+
-+ if (dev_is_fsl_mc(parent_dev))
-+ parent_mc_dev = to_fsl_mc_device(parent_dev);
-+ else
-+ parent_mc_dev = NULL;
-+
-+ if (strcmp(obj_desc->type, "dprc") == 0) {
-+ /*
-+ * Allocate an MC bus device object:
-+ */
-+ mc_bus = kzalloc(sizeof(*mc_bus), GFP_KERNEL);
-+ if (!mc_bus)
-+ return -ENOMEM;
-+
-+ mc_dev = &mc_bus->mc_dev;
-+ } else {
-+ /*
-+ * Allocate a regular fsl_mc_device object:
-+ */
-+ mc_dev = kzalloc(sizeof(*mc_dev), GFP_KERNEL);
-+ if (!mc_dev)
-+ return -ENOMEM;
-+ }
-+
-+ mc_dev->obj_desc = *obj_desc;
-+ mc_dev->mc_io = mc_io;
-+
-+ if (driver_override) {
-+ /*
-+ * We trust driver_override, so we don't need to use
-+ * kstrndup() here
-+ */
-+ mc_dev->driver_override = kstrdup(driver_override, GFP_KERNEL);
-+ if (!mc_dev->driver_override) {
-+ error = -ENOMEM;
-+ goto error_cleanup_dev;
-+ }
-+ }
-+
-+ device_initialize(&mc_dev->dev);
-+ mc_dev->dev.parent = parent_dev;
-+ mc_dev->dev.bus = &fsl_mc_bus_type;
-+ mc_dev->dev.release = fsl_mc_device_release;
-+ mc_dev->dev.type = fsl_mc_get_device_type(obj_desc->type);
-+ if (!mc_dev->dev.type) {
-+ error = -ENODEV;
-+ dev_err(parent_dev, "unknown device type %s\n", obj_desc->type);
-+ goto error_cleanup_dev;
-+ }
-+ dev_set_name(&mc_dev->dev, "%s.%d", obj_desc->type, obj_desc->id);
-+
-+ if (strcmp(obj_desc->type, "dprc") == 0) {
-+ struct fsl_mc_io *mc_io2;
-+
-+ mc_dev->flags |= FSL_MC_IS_DPRC;
-+
-+ /*
-+ * To get the DPRC's ICID, we need to open the DPRC
-+ * in get_dprc_icid(). For child DPRCs, we do so using the
-+ * parent DPRC's MC portal instead of the child DPRC's MC
-+ * portal, in case the child DPRC is already opened with
-+ * its own portal (e.g., the DPRC used by AIOP).
-+ *
-+ * NOTE: There cannot be more than one active open for a
-+ * given MC object, using the same MC portal.
-+ */
-+ if (parent_mc_dev) {
-+ /*
-+ * device being added is a child DPRC device
-+ */
-+ mc_io2 = parent_mc_dev->mc_io;
-+ } else {
-+ /*
-+ * device being added is the root DPRC device
-+ */
-+ if (!mc_io) {
-+ error = -EINVAL;
-+ goto error_cleanup_dev;
-+ }
-+
-+ mc_io2 = mc_io;
-+ }
-+
-+ error = get_dprc_icid(mc_io2, obj_desc->id, &mc_dev->icid);
-+ if (error < 0)
-+ goto error_cleanup_dev;
-+ } else {
-+ /*
-+ * A non-DPRC object has to be a child of a DPRC, use the
-+ * parent's ICID and interrupt domain.
-+ */
-+ mc_dev->icid = parent_mc_dev->icid;
-+ mc_dev->dma_mask = FSL_MC_DEFAULT_DMA_MASK;
-+ mc_dev->dev.dma_mask = &mc_dev->dma_mask;
-+ mc_dev->dev.coherent_dma_mask = mc_dev->dma_mask;
-+ dev_set_msi_domain(&mc_dev->dev,
-+ dev_get_msi_domain(&parent_mc_dev->dev));
-+ }
-+
-+ /*
-+ * Get MMIO regions for the device from the MC:
-+ *
-+ * NOTE: the root DPRC is a special case as its MMIO region is
-+ * obtained from the device tree
-+ */
-+ if (parent_mc_dev && obj_desc->region_count != 0) {
-+ error = fsl_mc_device_get_mmio_regions(mc_dev,
-+ parent_mc_dev);
-+ if (error < 0)
-+ goto error_cleanup_dev;
-+ }
-+
-+ fsl_mc_platform_dev = &mc_dev->dev;
-+ while (dev_is_fsl_mc(fsl_mc_platform_dev))
-+ fsl_mc_platform_dev = fsl_mc_platform_dev->parent;
-+ fsl_mc_platform_node = fsl_mc_platform_dev->of_node;
-+
-+ /* Set up the iommu configuration for the devices. */
-+ fsl_mc_dma_configure(mc_dev, fsl_mc_platform_node,
-+ !(obj_desc->flags & DPRC_OBJ_FLAG_NO_MEM_SHAREABILITY));
-+
-+ /*
-+ * The device-specific probe callback will get invoked by device_add()
-+ */
-+ error = device_add(&mc_dev->dev);
-+ if (error < 0) {
-+ dev_err(parent_dev,
-+ "device_add() failed for device %s: %d\n",
-+ dev_name(&mc_dev->dev), error);
-+ goto error_cleanup_dev;
-+ }
-+
-+ dev_dbg(parent_dev, "added %s\n", dev_name(&mc_dev->dev));
-+
-+ *new_mc_dev = mc_dev;
-+ return 0;
-+
-+error_cleanup_dev:
-+ kfree(mc_dev->regions);
-+ kfree(mc_bus);
-+ kfree(mc_dev);
-+
-+ return error;
-+}
-+EXPORT_SYMBOL_GPL(fsl_mc_device_add);
-+
-+/**
-+ * fsl_mc_device_remove - Remove an fsl-mc device from being visible to
-+ * Linux
-+ *
-+ * @mc_dev: Pointer to an fsl-mc device
-+ */
-+void fsl_mc_device_remove(struct fsl_mc_device *mc_dev)
-+{
-+ kfree(mc_dev->driver_override);
-+ mc_dev->driver_override = NULL;
-+
-+ /*
-+ * The device-specific remove callback will get invoked by device_del()
-+ */
-+ device_del(&mc_dev->dev);
-+ put_device(&mc_dev->dev);
-+}
-+EXPORT_SYMBOL_GPL(fsl_mc_device_remove);
-+
-+static int parse_mc_ranges(struct device *dev,
-+ int *paddr_cells,
-+ int *mc_addr_cells,
-+ int *mc_size_cells,
-+ const __be32 **ranges_start)
-+{
-+ const __be32 *prop;
-+ int range_tuple_cell_count;
-+ int ranges_len;
-+ int tuple_len;
-+ struct device_node *mc_node = dev->of_node;
-+
-+ *ranges_start = of_get_property(mc_node, "ranges", &ranges_len);
-+ if (!(*ranges_start) || !ranges_len) {
-+ dev_warn(dev,
-+ "missing or empty ranges property for device tree node '%s'\n",
-+ mc_node->name);
-+ return 0;
-+ }
-+
-+ *paddr_cells = of_n_addr_cells(mc_node);
-+
-+ prop = of_get_property(mc_node, "#address-cells", NULL);
-+ if (prop)
-+ *mc_addr_cells = be32_to_cpup(prop);
-+ else
-+ *mc_addr_cells = *paddr_cells;
-+
-+ prop = of_get_property(mc_node, "#size-cells", NULL);
-+ if (prop)
-+ *mc_size_cells = be32_to_cpup(prop);
-+ else
-+ *mc_size_cells = of_n_size_cells(mc_node);
-+
-+ range_tuple_cell_count = *paddr_cells + *mc_addr_cells +
-+ *mc_size_cells;
-+
-+ tuple_len = range_tuple_cell_count * sizeof(__be32);
-+ if (ranges_len % tuple_len != 0) {
-+ dev_err(dev, "malformed ranges property '%s'\n", mc_node->name);
-+ return -EINVAL;
-+ }
-+
-+ return ranges_len / tuple_len;
-+}
-+
-+static int get_mc_addr_translation_ranges(struct device *dev,
-+ struct fsl_mc_addr_translation_range
-+ **ranges,
-+ u8 *num_ranges)
-+{
-+ int ret;
-+ int paddr_cells;
-+ int mc_addr_cells;
-+ int mc_size_cells;
-+ int i;
-+ const __be32 *ranges_start;
-+ const __be32 *cell;
-+
-+ ret = parse_mc_ranges(dev,
-+ &paddr_cells,
-+ &mc_addr_cells,
-+ &mc_size_cells,
-+ &ranges_start);
-+ if (ret < 0)
-+ return ret;
-+
-+ *num_ranges = ret;
-+ if (!ret) {
-+ /*
-+ * Missing or empty ranges property ("ranges;") for the
-+ * 'fsl,qoriq-mc' node. In this case, identity mapping
-+ * will be used.
-+ */
-+ *ranges = NULL;
-+ return 0;
-+ }
-+
-+ *ranges = devm_kcalloc(dev, *num_ranges,
-+ sizeof(struct fsl_mc_addr_translation_range),
-+ GFP_KERNEL);
-+ if (!(*ranges))
-+ return -ENOMEM;
-+
-+ cell = ranges_start;
-+ for (i = 0; i < *num_ranges; ++i) {
-+ struct fsl_mc_addr_translation_range *range = &(*ranges)[i];
-+
-+ range->mc_region_type = of_read_number(cell, 1);
-+ range->start_mc_offset = of_read_number(cell + 1,
-+ mc_addr_cells - 1);
-+ cell += mc_addr_cells;
-+ range->start_phys_addr = of_read_number(cell, paddr_cells);
-+ cell += paddr_cells;
-+ range->end_mc_offset = range->start_mc_offset +
-+ of_read_number(cell, mc_size_cells);
-+
-+ cell += mc_size_cells;
-+ }
-+
-+ return 0;
-+}
-+
-+/**
-+ * fsl_mc_bus_probe - callback invoked when the root MC bus is being
-+ * added
-+ */
-+static int fsl_mc_bus_probe(struct platform_device *pdev)
-+{
-+ struct fsl_mc_obj_desc obj_desc;
-+ int error;
-+ struct fsl_mc *mc;
-+ struct fsl_mc_device *mc_bus_dev = NULL;
-+ struct fsl_mc_io *mc_io = NULL;
-+ struct fsl_mc_bus *mc_bus = NULL;
-+ int container_id;
-+ phys_addr_t mc_portal_phys_addr;
-+ u32 mc_portal_size;
-+ struct mc_version mc_version;
-+ struct resource res;
-+
-+ mc = devm_kzalloc(&pdev->dev, sizeof(*mc), GFP_KERNEL);
-+ if (!mc)
-+ return -ENOMEM;
-+
-+ platform_set_drvdata(pdev, mc);
-+
-+ /*
-+ * Get physical address of MC portal for the root DPRC:
-+ */
-+ error = of_address_to_resource(pdev->dev.of_node, 0, &res);
-+ if (error < 0) {
-+ dev_err(&pdev->dev,
-+ "of_address_to_resource() failed for %pOF\n",
-+ pdev->dev.of_node);
-+ return error;
-+ }
-+
-+ mc_portal_phys_addr = res.start;
-+ mc_portal_size = resource_size(&res);
-+ error = fsl_create_mc_io(&pdev->dev, mc_portal_phys_addr,
-+ mc_portal_size, NULL,
-+ FSL_MC_IO_ATOMIC_CONTEXT_PORTAL, &mc_io);
-+ if (error < 0)
-+ return error;
-+
-+ error = mc_get_version(mc_io, 0, &mc_version);
-+ if (error != 0) {
-+ dev_err(&pdev->dev,
-+ "mc_get_version() failed with error %d\n", error);
-+ goto error_cleanup_mc_io;
-+ }
-+
-+ dev_info(&pdev->dev, "MC firmware version: %u.%u.%u\n",
-+ mc_version.major, mc_version.minor, mc_version.revision);
-+
-+ error = get_mc_addr_translation_ranges(&pdev->dev,
-+ &mc->translation_ranges,
-+ &mc->num_translation_ranges);
-+ if (error < 0)
-+ goto error_cleanup_mc_io;
-+
-+ error = dprc_get_container_id(mc_io, 0, &container_id);
-+ if (error < 0) {
-+ dev_err(&pdev->dev,
-+ "dprc_get_container_id() failed: %d\n", error);
-+ goto error_cleanup_mc_io;
-+ }
-+
-+ memset(&obj_desc, 0, sizeof(struct fsl_mc_obj_desc));
-+ error = dprc_get_api_version(mc_io, 0,
-+ &obj_desc.ver_major,
-+ &obj_desc.ver_minor);
-+ if (error < 0)
-+ goto error_cleanup_mc_io;
-+
-+ obj_desc.vendor = FSL_MC_VENDOR_FREESCALE;
-+ strcpy(obj_desc.type, "dprc");
-+ obj_desc.id = container_id;
-+ obj_desc.irq_count = 1;
-+ obj_desc.region_count = 0;
-+
-+ error = fsl_mc_device_add(&obj_desc, mc_io, &pdev->dev, NULL,
-+ &mc_bus_dev);
-+ if (error < 0)
-+ goto error_cleanup_mc_io;
-+
-+ mc_bus = to_fsl_mc_bus(mc_bus_dev);
-+ error = fsl_mc_restool_create_device_file(mc_bus);
-+ if (error < 0)
-+ goto error_cleanup_device;
-+
-+ mc->root_mc_bus_dev = mc_bus_dev;
-+
-+ return 0;
-+
-+error_cleanup_device:
-+ fsl_mc_device_remove(mc_bus_dev);
-+
-+error_cleanup_mc_io:
-+ fsl_destroy_mc_io(mc_io);
-+ return error;
-+}
-+
-+/**
-+ * fsl_mc_bus_remove - callback invoked when the root MC bus is being
-+ * removed
-+ */
-+static int fsl_mc_bus_remove(struct platform_device *pdev)
-+{
-+ struct fsl_mc *mc = platform_get_drvdata(pdev);
-+ struct fsl_mc_bus *mc_bus = to_fsl_mc_bus(mc->root_mc_bus_dev);
-+
-+ if (!fsl_mc_is_root_dprc(&mc->root_mc_bus_dev->dev))
-+ return -EINVAL;
-+
-+ fsl_mc_restool_remove_device_file(mc_bus);
-+ fsl_mc_device_remove(mc->root_mc_bus_dev);
-+
-+ fsl_destroy_mc_io(mc->root_mc_bus_dev->mc_io);
-+ mc->root_mc_bus_dev->mc_io = NULL;
-+
-+ return 0;
-+}
-+
-+static const struct of_device_id fsl_mc_bus_match_table[] = {
-+ {.compatible = "fsl,qoriq-mc",},
-+ {},
-+};
-+
-+MODULE_DEVICE_TABLE(of, fsl_mc_bus_match_table);
-+
-+static struct platform_driver fsl_mc_bus_driver = {
-+ .driver = {
-+ .name = "fsl_mc_bus",
-+ .pm = NULL,
-+ .of_match_table = fsl_mc_bus_match_table,
-+ },
-+ .probe = fsl_mc_bus_probe,
-+ .remove = fsl_mc_bus_remove,
-+};
-+
-+static int __init fsl_mc_bus_driver_init(void)
-+{
-+ int error;
-+
-+ error = bus_register(&fsl_mc_bus_type);
-+ if (error < 0) {
-+ pr_err("bus type registration failed: %d\n", error);
-+ goto error_cleanup_cache;
-+ }
-+
-+ error = platform_driver_register(&fsl_mc_bus_driver);
-+ if (error < 0) {
-+ pr_err("platform_driver_register() failed: %d\n", error);
-+ goto error_cleanup_bus;
-+ }
-+
-+ error = dprc_driver_init();
-+ if (error < 0)
-+ goto error_cleanup_driver;
-+
-+ error = fsl_mc_allocator_driver_init();
-+ if (error < 0)
-+ goto error_cleanup_dprc_driver;
-+
-+ error = fsl_mc_restool_init();
-+ if (error < 0)
-+ goto error_cleanup_mc_allocator;
-+
-+ return 0;
-+
-+error_cleanup_mc_allocator:
-+ fsl_mc_allocator_driver_exit();
-+
-+error_cleanup_dprc_driver:
-+ dprc_driver_exit();
-+
-+error_cleanup_driver:
-+ platform_driver_unregister(&fsl_mc_bus_driver);
-+
-+error_cleanup_bus:
-+ bus_unregister(&fsl_mc_bus_type);
-+
-+error_cleanup_cache:
-+ return error;
-+}
-+postcore_initcall(fsl_mc_bus_driver_init);
---- /dev/null
-+++ b/drivers/bus/fsl-mc/fsl-mc-iommu.c
-@@ -0,0 +1,78 @@
-+/*
-+ * Copyright 2016 Freescale Semiconductor, Inc.
-+ * Copyright 2017 NXP
-+ * Author: Nipun Gupta <nipun.gupta@nxp.com>
-+ *
-+ * This file is licensed under the terms of the GNU General Public
-+ * License version 2. This program is licensed "as is" without any
-+ * warranty of any kind, whether express or implied.
-+ */
-+
-+#include <linux/iommu.h>
-+#include <linux/of.h>
-+#include <linux/of_iommu.h>
-+#include <linux/fsl/mc.h>
-+
-+/* Setup the IOMMU for the DPRC container */
-+static const struct iommu_ops
-+*fsl_mc_iommu_configure(struct fsl_mc_device *mc_dev,
-+ struct device_node *fsl_mc_platform_node)
-+{
-+ struct of_phandle_args iommu_spec;
-+ const struct iommu_ops *ops;
-+ u32 iommu_phandle;
-+ struct device_node *iommu_node;
-+ const __be32 *map = NULL;
-+ int iommu_cells, map_len, ret;
-+
-+ map = of_get_property(fsl_mc_platform_node, "iommu-map", &map_len);
-+ if (!map)
-+ return NULL;
-+
-+ ops = mc_dev->dev.bus->iommu_ops;
-+ if (!ops || !ops->of_xlate)
-+ return NULL;
-+
-+ iommu_phandle = be32_to_cpup(map + 1);
-+ iommu_node = of_find_node_by_phandle(iommu_phandle);
-+
-+ if (of_property_read_u32(iommu_node, "#iommu-cells", &iommu_cells)) {
-+ pr_err("%s: missing #iommu-cells property\n", iommu_node->name);
-+ return NULL;
-+ }
-+
-+ /* Initialize the fwspec */
-+ ret = iommu_fwspec_init(&mc_dev->dev, &iommu_node->fwnode, ops);
-+ if (ret)
-+ return NULL;
-+
-+ /*
-+ * Fill in the required stream-id before calling the iommu's
-+ * ops->xlate callback.
-+ */
-+ iommu_spec.np = iommu_node;
-+ iommu_spec.args[0] = mc_dev->icid;
-+ iommu_spec.args_count = 1;
-+
-+ ret = ops->of_xlate(&mc_dev->dev, &iommu_spec);
-+ if (ret)
-+ return NULL;
-+
-+ of_node_put(iommu_spec.np);
-+
-+ return ops;
-+}
-+
-+/* Set up DMA configuration for fsl-mc devices */
-+void fsl_mc_dma_configure(struct fsl_mc_device *mc_dev,
-+ struct device_node *fsl_mc_platform_node, int coherent)
-+{
-+ const struct iommu_ops *ops;
-+
-+ ops = fsl_mc_iommu_configure(mc_dev, fsl_mc_platform_node);
-+
-+ mc_dev->dev.coherent_dma_mask = DMA_BIT_MASK(48);
-+ mc_dev->dev.dma_mask = &mc_dev->dev.coherent_dma_mask;
-+ arch_setup_dma_ops(&mc_dev->dev, 0,
-+ mc_dev->dev.coherent_dma_mask + 1, ops, coherent);
-+}
---- a/drivers/staging/fsl-mc/bus/fsl-mc-msi.c
-+++ /dev/null
-@@ -1,285 +0,0 @@
--/*
-- * Freescale Management Complex (MC) bus driver MSI support
-- *
-- * Copyright (C) 2015 Freescale Semiconductor, Inc.
-- * Author: German Rivera <German.Rivera@freescale.com>
-- *
-- * This file is licensed under the terms of the GNU General Public
-- * License version 2. This program is licensed "as is" without any
-- * warranty of any kind, whether express or implied.
-- */
--
--#include <linux/of_device.h>
--#include <linux/of_address.h>
--#include <linux/irqchip/arm-gic-v3.h>
--#include <linux/of_irq.h>
--#include <linux/irq.h>
--#include <linux/irqdomain.h>
--#include <linux/msi.h>
--#include "../include/mc-bus.h"
--#include "fsl-mc-private.h"
--
--/*
-- * Generate a unique ID identifying the interrupt (only used within the MSI
-- * irqdomain. Combine the icid with the interrupt index.
-- */
--static irq_hw_number_t fsl_mc_domain_calc_hwirq(struct fsl_mc_device *dev,
-- struct msi_desc *desc)
--{
-- /*
-- * Make the base hwirq value for ICID*10000 so it is readable
-- * as a decimal value in /proc/interrupts.
-- */
-- return (irq_hw_number_t)(desc->fsl_mc.msi_index + (dev->icid * 10000));
--}
--
--static void fsl_mc_msi_set_desc(msi_alloc_info_t *arg,
-- struct msi_desc *desc)
--{
-- arg->desc = desc;
-- arg->hwirq = fsl_mc_domain_calc_hwirq(to_fsl_mc_device(desc->dev),
-- desc);
--}
--
--static void fsl_mc_msi_update_dom_ops(struct msi_domain_info *info)
--{
-- struct msi_domain_ops *ops = info->ops;
--
-- if (WARN_ON(!ops))
-- return;
--
-- /*
-- * set_desc should not be set by the caller
-- */
-- if (!ops->set_desc)
-- ops->set_desc = fsl_mc_msi_set_desc;
--}
--
--static void __fsl_mc_msi_write_msg(struct fsl_mc_device *mc_bus_dev,
-- struct fsl_mc_device_irq *mc_dev_irq)
--{
-- int error;
-- struct fsl_mc_device *owner_mc_dev = mc_dev_irq->mc_dev;
-- struct msi_desc *msi_desc = mc_dev_irq->msi_desc;
-- struct dprc_irq_cfg irq_cfg;
--
-- /*
-- * msi_desc->msg.address is 0x0 when this function is invoked in
-- * the free_irq() code path. In this case, for the MC, we don't
-- * really need to "unprogram" the MSI, so we just return.
-- */
-- if (msi_desc->msg.address_lo == 0x0 && msi_desc->msg.address_hi == 0x0)
-- return;
--
-- if (WARN_ON(!owner_mc_dev))
-- return;
--
-- irq_cfg.paddr = ((u64)msi_desc->msg.address_hi << 32) |
-- msi_desc->msg.address_lo;
-- irq_cfg.val = msi_desc->msg.data;
-- irq_cfg.irq_num = msi_desc->irq;
--
-- if (owner_mc_dev == mc_bus_dev) {
-- /*
-- * IRQ is for the mc_bus_dev's DPRC itself
-- */
-- error = dprc_set_irq(mc_bus_dev->mc_io,
-- MC_CMD_FLAG_INTR_DIS | MC_CMD_FLAG_PRI,
-- mc_bus_dev->mc_handle,
-- mc_dev_irq->dev_irq_index,
-- &irq_cfg);
-- if (error < 0) {
-- dev_err(&owner_mc_dev->dev,
-- "dprc_set_irq() failed: %d\n", error);
-- }
-- } else {
-- /*
-- * IRQ is for for a child device of mc_bus_dev
-- */
-- error = dprc_set_obj_irq(mc_bus_dev->mc_io,
-- MC_CMD_FLAG_INTR_DIS | MC_CMD_FLAG_PRI,
-- mc_bus_dev->mc_handle,
-- owner_mc_dev->obj_desc.type,
-- owner_mc_dev->obj_desc.id,
-- mc_dev_irq->dev_irq_index,
-- &irq_cfg);
-- if (error < 0) {
-- dev_err(&owner_mc_dev->dev,
-- "dprc_obj_set_irq() failed: %d\n", error);
-- }
-- }
--}
--
--/*
-- * NOTE: This function is invoked with interrupts disabled
-- */
--static void fsl_mc_msi_write_msg(struct irq_data *irq_data,
-- struct msi_msg *msg)
--{
-- struct msi_desc *msi_desc = irq_data_get_msi_desc(irq_data);
-- struct fsl_mc_device *mc_bus_dev = to_fsl_mc_device(msi_desc->dev);
-- struct fsl_mc_bus *mc_bus = to_fsl_mc_bus(mc_bus_dev);
-- struct fsl_mc_device_irq *mc_dev_irq =
-- &mc_bus->irq_resources[msi_desc->fsl_mc.msi_index];
--
-- WARN_ON(mc_dev_irq->msi_desc != msi_desc);
-- msi_desc->msg = *msg;
--
-- /*
-- * Program the MSI (paddr, value) pair in the device:
-- */
-- __fsl_mc_msi_write_msg(mc_bus_dev, mc_dev_irq);
--}
--
--static void fsl_mc_msi_update_chip_ops(struct msi_domain_info *info)
--{
-- struct irq_chip *chip = info->chip;
--
-- if (WARN_ON((!chip)))
-- return;
--
-- /*
-- * irq_write_msi_msg should not be set by the caller
-- */
-- if (!chip->irq_write_msi_msg)
-- chip->irq_write_msi_msg = fsl_mc_msi_write_msg;
--}
--
--/**
-- * fsl_mc_msi_create_irq_domain - Create a fsl-mc MSI interrupt domain
-- * @np: Optional device-tree node of the interrupt controller
-- * @info: MSI domain info
-- * @parent: Parent irq domain
-- *
-- * Updates the domain and chip ops and creates a fsl-mc MSI
-- * interrupt domain.
-- *
-- * Returns:
-- * A domain pointer or NULL in case of failure.
-- */
--struct irq_domain *fsl_mc_msi_create_irq_domain(struct fwnode_handle *fwnode,
-- struct msi_domain_info *info,
-- struct irq_domain *parent)
--{
-- struct irq_domain *domain;
--
-- if (info->flags & MSI_FLAG_USE_DEF_DOM_OPS)
-- fsl_mc_msi_update_dom_ops(info);
-- if (info->flags & MSI_FLAG_USE_DEF_CHIP_OPS)
-- fsl_mc_msi_update_chip_ops(info);
--
-- domain = msi_create_irq_domain(fwnode, info, parent);
-- if (domain)
-- domain->bus_token = DOMAIN_BUS_FSL_MC_MSI;
--
-- return domain;
--}
--
--int fsl_mc_find_msi_domain(struct device *mc_platform_dev,
-- struct irq_domain **mc_msi_domain)
--{
-- struct irq_domain *msi_domain;
-- struct device_node *mc_of_node = mc_platform_dev->of_node;
--
-- msi_domain = of_msi_get_domain(mc_platform_dev, mc_of_node,
-- DOMAIN_BUS_FSL_MC_MSI);
-- if (!msi_domain) {
-- pr_err("Unable to find fsl-mc MSI domain for %s\n",
-- mc_of_node->full_name);
--
-- return -ENOENT;
-- }
--
-- *mc_msi_domain = msi_domain;
-- return 0;
--}
--
--static void fsl_mc_msi_free_descs(struct device *dev)
--{
-- struct msi_desc *desc, *tmp;
--
-- list_for_each_entry_safe(desc, tmp, dev_to_msi_list(dev), list) {
-- list_del(&desc->list);
-- free_msi_entry(desc);
-- }
--}
--
--static int fsl_mc_msi_alloc_descs(struct device *dev, unsigned int irq_count)
--
--{
-- unsigned int i;
-- int error;
-- struct msi_desc *msi_desc;
--
-- for (i = 0; i < irq_count; i++) {
-- msi_desc = alloc_msi_entry(dev, 1, NULL);
-- if (!msi_desc) {
-- dev_err(dev, "Failed to allocate msi entry\n");
-- error = -ENOMEM;
-- goto cleanup_msi_descs;
-- }
--
-- msi_desc->fsl_mc.msi_index = i;
-- INIT_LIST_HEAD(&msi_desc->list);
-- list_add_tail(&msi_desc->list, dev_to_msi_list(dev));
-- }
--
-- return 0;
--
--cleanup_msi_descs:
-- fsl_mc_msi_free_descs(dev);
-- return error;
--}
--
--int fsl_mc_msi_domain_alloc_irqs(struct device *dev,
-- unsigned int irq_count)
--{
-- struct irq_domain *msi_domain;
-- int error;
--
-- if (WARN_ON(!list_empty(dev_to_msi_list(dev))))
-- return -EINVAL;
--
-- error = fsl_mc_msi_alloc_descs(dev, irq_count);
-- if (error < 0)
-- return error;
--
-- msi_domain = dev_get_msi_domain(dev);
-- if (WARN_ON(!msi_domain)) {
-- error = -EINVAL;
-- goto cleanup_msi_descs;
-- }
--
-- /*
-- * NOTE: Calling this function will trigger the invocation of the
-- * its_fsl_mc_msi_prepare() callback
-- */
-- error = msi_domain_alloc_irqs(msi_domain, dev, irq_count);
--
-- if (error) {
-- dev_err(dev, "Failed to allocate IRQs\n");
-- goto cleanup_msi_descs;
-- }
--
-- return 0;
--
--cleanup_msi_descs:
-- fsl_mc_msi_free_descs(dev);
-- return error;
--}
--
--void fsl_mc_msi_domain_free_irqs(struct device *dev)
--{
-- struct irq_domain *msi_domain;
--
-- msi_domain = dev_get_msi_domain(dev);
-- if (WARN_ON(!msi_domain))
-- return;
--
-- msi_domain_free_irqs(msi_domain, dev);
--
-- if (WARN_ON(list_empty(dev_to_msi_list(dev))))
-- return;
--
-- fsl_mc_msi_free_descs(dev);
--}
---- /dev/null
-+++ b/drivers/bus/fsl-mc/fsl-mc-msi.c
-@@ -0,0 +1,285 @@
-+// SPDX-License-Identifier: GPL-2.0
-+/*
-+ * Freescale Management Complex (MC) bus driver MSI support
-+ *
-+ * Copyright (C) 2015-2016 Freescale Semiconductor, Inc.
-+ * Author: German Rivera <German.Rivera@freescale.com>
-+ *
-+ */
-+
-+#include <linux/of_device.h>
-+#include <linux/of_address.h>
-+#include <linux/of_irq.h>
-+#include <linux/irq.h>
-+#include <linux/irqdomain.h>
-+#include <linux/msi.h>
-+
-+#include "fsl-mc-private.h"
-+
-+#ifdef GENERIC_MSI_DOMAIN_OPS
-+/*
-+ * Generate a unique ID identifying the interrupt (only used within the MSI
-+ * irqdomain. Combine the icid with the interrupt index.
-+ */
-+static irq_hw_number_t fsl_mc_domain_calc_hwirq(struct fsl_mc_device *dev,
-+ struct msi_desc *desc)
-+{
-+ /*
-+ * Make the base hwirq value for ICID*10000 so it is readable
-+ * as a decimal value in /proc/interrupts.
-+ */
-+ return (irq_hw_number_t)(desc->fsl_mc.msi_index + (dev->icid * 10000));
-+}
-+
-+static void fsl_mc_msi_set_desc(msi_alloc_info_t *arg,
-+ struct msi_desc *desc)
-+{
-+ arg->desc = desc;
-+ arg->hwirq = fsl_mc_domain_calc_hwirq(to_fsl_mc_device(desc->dev),
-+ desc);
-+}
-+#else
-+#define fsl_mc_msi_set_desc NULL
-+#endif
-+
-+static void fsl_mc_msi_update_dom_ops(struct msi_domain_info *info)
-+{
-+ struct msi_domain_ops *ops = info->ops;
-+
-+ if (!ops)
-+ return;
-+
-+ /*
-+ * set_desc should not be set by the caller
-+ */
-+ if (!ops->set_desc)
-+ ops->set_desc = fsl_mc_msi_set_desc;
-+}
-+
-+static void __fsl_mc_msi_write_msg(struct fsl_mc_device *mc_bus_dev,
-+ struct fsl_mc_device_irq *mc_dev_irq)
-+{
-+ int error;
-+ struct fsl_mc_device *owner_mc_dev = mc_dev_irq->mc_dev;
-+ struct msi_desc *msi_desc = mc_dev_irq->msi_desc;
-+ struct dprc_irq_cfg irq_cfg;
-+
-+ /*
-+ * msi_desc->msg.address is 0x0 when this function is invoked in
-+ * the free_irq() code path. In this case, for the MC, we don't
-+ * really need to "unprogram" the MSI, so we just return.
-+ */
-+ if (msi_desc->msg.address_lo == 0x0 && msi_desc->msg.address_hi == 0x0)
-+ return;
-+
-+ if (!owner_mc_dev)
-+ return;
-+
-+ irq_cfg.paddr = ((u64)msi_desc->msg.address_hi << 32) |
-+ msi_desc->msg.address_lo;
-+ irq_cfg.val = msi_desc->msg.data;
-+ irq_cfg.irq_num = msi_desc->irq;
-+
-+ if (owner_mc_dev == mc_bus_dev) {
-+ /*
-+ * IRQ is for the mc_bus_dev's DPRC itself
-+ */
-+ error = dprc_set_irq(mc_bus_dev->mc_io,
-+ MC_CMD_FLAG_INTR_DIS | MC_CMD_FLAG_PRI,
-+ mc_bus_dev->mc_handle,
-+ mc_dev_irq->dev_irq_index,
-+ &irq_cfg);
-+ if (error < 0) {
-+ dev_err(&owner_mc_dev->dev,
-+ "dprc_set_irq() failed: %d\n", error);
-+ }
-+ } else {
-+ /*
-+ * IRQ is for for a child device of mc_bus_dev
-+ */
-+ error = dprc_set_obj_irq(mc_bus_dev->mc_io,
-+ MC_CMD_FLAG_INTR_DIS | MC_CMD_FLAG_PRI,
-+ mc_bus_dev->mc_handle,
-+ owner_mc_dev->obj_desc.type,
-+ owner_mc_dev->obj_desc.id,
-+ mc_dev_irq->dev_irq_index,
-+ &irq_cfg);
-+ if (error < 0) {
-+ dev_err(&owner_mc_dev->dev,
-+ "dprc_obj_set_irq() failed: %d\n", error);
-+ }
-+ }
-+}
-+
-+/*
-+ * NOTE: This function is invoked with interrupts disabled
-+ */
-+static void fsl_mc_msi_write_msg(struct irq_data *irq_data,
-+ struct msi_msg *msg)
-+{
-+ struct msi_desc *msi_desc = irq_data_get_msi_desc(irq_data);
-+ struct fsl_mc_device *mc_bus_dev = to_fsl_mc_device(msi_desc->dev);
-+ struct fsl_mc_bus *mc_bus = to_fsl_mc_bus(mc_bus_dev);
-+ struct fsl_mc_device_irq *mc_dev_irq =
-+ &mc_bus->irq_resources[msi_desc->fsl_mc.msi_index];
-+
-+ msi_desc->msg = *msg;
-+
-+ /*
-+ * Program the MSI (paddr, value) pair in the device:
-+ */
-+ __fsl_mc_msi_write_msg(mc_bus_dev, mc_dev_irq);
-+}
-+
-+static void fsl_mc_msi_update_chip_ops(struct msi_domain_info *info)
-+{
-+ struct irq_chip *chip = info->chip;
-+
-+ if (!chip)
-+ return;
-+
-+ /*
-+ * irq_write_msi_msg should not be set by the caller
-+ */
-+ if (!chip->irq_write_msi_msg)
-+ chip->irq_write_msi_msg = fsl_mc_msi_write_msg;
-+}
-+
-+/**
-+ * fsl_mc_msi_create_irq_domain - Create a fsl-mc MSI interrupt domain
-+ * @np: Optional device-tree node of the interrupt controller
-+ * @info: MSI domain info
-+ * @parent: Parent irq domain
-+ *
-+ * Updates the domain and chip ops and creates a fsl-mc MSI
-+ * interrupt domain.
-+ *
-+ * Returns:
-+ * A domain pointer or NULL in case of failure.
-+ */
-+struct irq_domain *fsl_mc_msi_create_irq_domain(struct fwnode_handle *fwnode,
-+ struct msi_domain_info *info,
-+ struct irq_domain *parent)
-+{
-+ struct irq_domain *domain;
-+
-+ if (info->flags & MSI_FLAG_USE_DEF_DOM_OPS)
-+ fsl_mc_msi_update_dom_ops(info);
-+ if (info->flags & MSI_FLAG_USE_DEF_CHIP_OPS)
-+ fsl_mc_msi_update_chip_ops(info);
-+
-+ domain = msi_create_irq_domain(fwnode, info, parent);
-+ if (domain)
-+ irq_domain_update_bus_token(domain, DOMAIN_BUS_FSL_MC_MSI);
-+
-+ return domain;
-+}
-+
-+int fsl_mc_find_msi_domain(struct device *mc_platform_dev,
-+ struct irq_domain **mc_msi_domain)
-+{
-+ struct irq_domain *msi_domain;
-+ struct device_node *mc_of_node = mc_platform_dev->of_node;
-+
-+ msi_domain = of_msi_get_domain(mc_platform_dev, mc_of_node,
-+ DOMAIN_BUS_FSL_MC_MSI);
-+ if (!msi_domain) {
-+ pr_err("Unable to find fsl-mc MSI domain for %pOF\n",
-+ mc_of_node);
-+
-+ return -ENOENT;
-+ }
-+
-+ *mc_msi_domain = msi_domain;
-+ return 0;
-+}
-+
-+static void fsl_mc_msi_free_descs(struct device *dev)
-+{
-+ struct msi_desc *desc, *tmp;
-+
-+ list_for_each_entry_safe(desc, tmp, dev_to_msi_list(dev), list) {
-+ list_del(&desc->list);
-+ free_msi_entry(desc);
-+ }
-+}
-+
-+static int fsl_mc_msi_alloc_descs(struct device *dev, unsigned int irq_count)
-+
-+{
-+ unsigned int i;
-+ int error;
-+ struct msi_desc *msi_desc;
-+
-+ for (i = 0; i < irq_count; i++) {
-+ msi_desc = alloc_msi_entry(dev, 1, NULL);
-+ if (!msi_desc) {
-+ dev_err(dev, "Failed to allocate msi entry\n");
-+ error = -ENOMEM;
-+ goto cleanup_msi_descs;
-+ }
-+
-+ msi_desc->fsl_mc.msi_index = i;
-+ INIT_LIST_HEAD(&msi_desc->list);
-+ list_add_tail(&msi_desc->list, dev_to_msi_list(dev));
-+ }
-+
-+ return 0;
-+
-+cleanup_msi_descs:
-+ fsl_mc_msi_free_descs(dev);
-+ return error;
-+}
-+
-+int fsl_mc_msi_domain_alloc_irqs(struct device *dev,
-+ unsigned int irq_count)
-+{
-+ struct irq_domain *msi_domain;
-+ int error;
-+
-+ if (!list_empty(dev_to_msi_list(dev)))
-+ return -EINVAL;
-+
-+ error = fsl_mc_msi_alloc_descs(dev, irq_count);
-+ if (error < 0)
-+ return error;
-+
-+ msi_domain = dev_get_msi_domain(dev);
-+ if (!msi_domain) {
-+ error = -EINVAL;
-+ goto cleanup_msi_descs;
-+ }
-+
-+ /*
-+ * NOTE: Calling this function will trigger the invocation of the
-+ * its_fsl_mc_msi_prepare() callback
-+ */
-+ error = msi_domain_alloc_irqs(msi_domain, dev, irq_count);
-+
-+ if (error) {
-+ dev_err(dev, "Failed to allocate IRQs\n");
-+ goto cleanup_msi_descs;
-+ }
-+
-+ return 0;
-+
-+cleanup_msi_descs:
-+ fsl_mc_msi_free_descs(dev);
-+ return error;
-+}
-+
-+void fsl_mc_msi_domain_free_irqs(struct device *dev)
-+{
-+ struct irq_domain *msi_domain;
-+
-+ msi_domain = dev_get_msi_domain(dev);
-+ if (!msi_domain)
-+ return;
-+
-+ msi_domain_free_irqs(msi_domain, dev);
-+
-+ if (list_empty(dev_to_msi_list(dev)))
-+ return;
-+
-+ fsl_mc_msi_free_descs(dev);
-+}
---- /dev/null
-+++ b/drivers/bus/fsl-mc/fsl-mc-private.h
-@@ -0,0 +1,223 @@
-+/* SPDX-License-Identifier: GPL-2.0 */
-+/*
-+ * Freescale Management Complex (MC) bus private declarations
-+ *
-+ * Copyright (C) 2016 Freescale Semiconductor, Inc.
-+ *
-+ */
-+#ifndef _FSL_MC_PRIVATE_H_
-+#define _FSL_MC_PRIVATE_H_
-+
-+#include <linux/fsl/mc.h>
-+#include <linux/mutex.h>
-+#include <linux/cdev.h>
-+#include <linux/ioctl.h>
-+
-+/*
-+ * Data Path Management Complex (DPMNG) General API
-+ */
-+
-+/* DPMNG command versioning */
-+#define DPMNG_CMD_BASE_VERSION 1
-+#define DPMNG_CMD_ID_OFFSET 4
-+
-+#define DPMNG_CMD(id) (((id) << DPMNG_CMD_ID_OFFSET) | DPMNG_CMD_BASE_VERSION)
-+
-+/* DPMNG command IDs */
-+#define DPMNG_CMDID_GET_VERSION DPMNG_CMD(0x831)
-+
-+struct dpmng_rsp_get_version {
-+ __le32 revision;
-+ __le32 version_major;
-+ __le32 version_minor;
-+};
-+
-+/*
-+ * Data Path Management Command Portal (DPMCP) API
-+ */
-+
-+/* Minimal supported DPMCP Version */
-+#define DPMCP_MIN_VER_MAJOR 3
-+#define DPMCP_MIN_VER_MINOR 0
-+
-+/* DPMCP command versioning */
-+#define DPMCP_CMD_BASE_VERSION 1
-+#define DPMCP_CMD_ID_OFFSET 4
-+
-+#define DPMCP_CMD(id) (((id) << DPMCP_CMD_ID_OFFSET) | DPMCP_CMD_BASE_VERSION)
-+
-+/* DPMCP command IDs */
-+#define DPMCP_CMDID_CLOSE DPMCP_CMD(0x800)
-+#define DPMCP_CMDID_OPEN DPMCP_CMD(0x80b)
-+#define DPMCP_CMDID_RESET DPMCP_CMD(0x005)
-+
-+struct dpmcp_cmd_open {
-+ __le32 dpmcp_id;
-+};
-+
-+/*
-+ * Initialization and runtime control APIs for DPMCP
-+ */
-+int dpmcp_open(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ int dpmcp_id,
-+ u16 *token);
-+
-+int dpmcp_close(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token);
-+
-+int dpmcp_reset(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token);
-+
-+/*
-+ * Data Path Buffer Pool (DPBP) API
-+ */
-+
-+/* DPBP Version */
-+#define DPBP_VER_MAJOR 3
-+#define DPBP_VER_MINOR 2
-+
-+/* Command versioning */
-+#define DPBP_CMD_BASE_VERSION 1
-+#define DPBP_CMD_ID_OFFSET 4
-+
-+#define DPBP_CMD(id) (((id) << DPBP_CMD_ID_OFFSET) | DPBP_CMD_BASE_VERSION)
-+
-+/* Command IDs */
-+#define DPBP_CMDID_CLOSE DPBP_CMD(0x800)
-+#define DPBP_CMDID_OPEN DPBP_CMD(0x804)
-+
-+#define DPBP_CMDID_ENABLE DPBP_CMD(0x002)
-+#define DPBP_CMDID_DISABLE DPBP_CMD(0x003)
-+#define DPBP_CMDID_GET_ATTR DPBP_CMD(0x004)
-+#define DPBP_CMDID_RESET DPBP_CMD(0x005)
-+
-+struct dpbp_cmd_open {
-+ __le32 dpbp_id;
-+};
-+
-+#define DPBP_ENABLE 0x1
-+
-+struct dpbp_rsp_get_attributes {
-+ /* response word 0 */
-+ __le16 pad;
-+ __le16 bpid;
-+ __le32 id;
-+ /* response word 1 */
-+ __le16 version_major;
-+ __le16 version_minor;
-+};
-+
-+/*
-+ * Data Path Concentrator (DPCON) API
-+ */
-+
-+/* DPCON Version */
-+#define DPCON_VER_MAJOR 3
-+#define DPCON_VER_MINOR 2
-+
-+/* Command versioning */
-+#define DPCON_CMD_BASE_VERSION 1
-+#define DPCON_CMD_ID_OFFSET 4
-+
-+#define DPCON_CMD(id) (((id) << DPCON_CMD_ID_OFFSET) | DPCON_CMD_BASE_VERSION)
-+
-+/* Command IDs */
-+#define DPCON_CMDID_CLOSE DPCON_CMD(0x800)
-+#define DPCON_CMDID_OPEN DPCON_CMD(0x808)
-+
-+#define DPCON_CMDID_ENABLE DPCON_CMD(0x002)
-+#define DPCON_CMDID_DISABLE DPCON_CMD(0x003)
-+#define DPCON_CMDID_GET_ATTR DPCON_CMD(0x004)
-+#define DPCON_CMDID_RESET DPCON_CMD(0x005)
-+
-+#define DPCON_CMDID_SET_NOTIFICATION DPCON_CMD(0x100)
-+
-+struct dpcon_cmd_open {
-+ __le32 dpcon_id;
-+};
-+
-+#define DPCON_ENABLE 1
-+
-+struct dpcon_rsp_get_attr {
-+ /* response word 0 */
-+ __le32 id;
-+ __le16 qbman_ch_id;
-+ u8 num_priorities;
-+ u8 pad;
-+};
-+
-+struct dpcon_cmd_set_notification {
-+ /* cmd word 0 */
-+ __le32 dpio_id;
-+ u8 priority;
-+ u8 pad[3];
-+ /* cmd word 1 */
-+ __le64 user_ctx;
-+};
-+
-+int __must_check fsl_mc_device_add(struct fsl_mc_obj_desc *obj_desc,
-+ struct fsl_mc_io *mc_io,
-+ struct device *parent_dev,
-+ const char *driver_override,
-+ struct fsl_mc_device **new_mc_dev);
-+
-+int __init dprc_driver_init(void);
-+
-+void dprc_driver_exit(void);
-+
-+int __init fsl_mc_allocator_driver_init(void);
-+
-+void fsl_mc_allocator_driver_exit(void);
-+
-+int __must_check fsl_mc_resource_allocate(struct fsl_mc_bus *mc_bus,
-+ enum fsl_mc_pool_type pool_type,
-+ struct fsl_mc_resource
-+ **new_resource);
-+
-+void fsl_mc_resource_free(struct fsl_mc_resource *resource);
-+
-+int fsl_mc_msi_domain_alloc_irqs(struct device *dev,
-+ unsigned int irq_count);
-+
-+void fsl_mc_msi_domain_free_irqs(struct device *dev);
-+
-+int __must_check fsl_create_mc_io(struct device *dev,
-+ phys_addr_t mc_portal_phys_addr,
-+ u32 mc_portal_size,
-+ struct fsl_mc_device *dpmcp_dev,
-+ u32 flags, struct fsl_mc_io **new_mc_io);
-+
-+void fsl_destroy_mc_io(struct fsl_mc_io *mc_io);
-+
-+bool fsl_mc_is_root_dprc(struct device *dev);
-+
-+#ifdef CONFIG_FSL_MC_RESTOOL
-+
-+int fsl_mc_restool_create_device_file(struct fsl_mc_bus *mc_bus);
-+
-+void fsl_mc_restool_remove_device_file(struct fsl_mc_bus *mc_bus);
-+
-+int fsl_mc_restool_init(void);
-+
-+#else
-+
-+static inline int fsl_mc_restool_create_device_file(struct fsl_mc_bus *mc_bus)
-+{
-+ return 0;
-+}
-+
-+static inline void fsl_mc_restool_remove_device_file(struct fsl_mc_bus *mc_bus)
-+{
-+}
-+
-+static inline int fsl_mc_restool_init(void)
-+{
-+ return 0;
-+}
-+
-+#endif
-+
-+#endif /* _FSL_MC_PRIVATE_H_ */
---- /dev/null
-+++ b/drivers/bus/fsl-mc/fsl-mc-restool.c
-@@ -0,0 +1,219 @@
-+// SPDX-License-Identifier: GPL-2.0
-+/*
-+ * Management Complex (MC) restool support
-+ *
-+ * Copyright 2018 NXP
-+ *
-+ */
-+
-+#include <linux/slab.h>
-+#include <linux/cdev.h>
-+#include <linux/fs.h>
-+#include <linux/uaccess.h>
-+
-+#include "fsl-mc-private.h"
-+
-+#define FSL_MC_BUS_MAX_MINORS 1
-+
-+static struct class *fsl_mc_bus_class;
-+static int fsl_mc_bus_major;
-+
-+static int fsl_mc_restool_send_command(unsigned long arg,
-+ struct fsl_mc_io *mc_io)
-+{
-+ struct fsl_mc_command mc_cmd;
-+ int error;
-+
-+ error = copy_from_user(&mc_cmd, (void __user *)arg, sizeof(mc_cmd));
-+ if (error)
-+ return -EFAULT;
-+
-+ error = mc_send_command(mc_io, &mc_cmd);
-+ if (error)
-+ return error;
-+
-+ error = copy_to_user((void __user *)arg, &mc_cmd, sizeof(mc_cmd));
-+ if (error)
-+ return -EFAULT;
-+
-+ return 0;
-+}
-+
-+int fsl_mc_restool_init(void)
-+{
-+ dev_t dev;
-+ int error;
-+
-+ fsl_mc_bus_class = class_create(THIS_MODULE, "fsl_mc_bus");
-+ if (IS_ERR(fsl_mc_bus_class)) {
-+ error = PTR_ERR(fsl_mc_bus_class);
-+ return error;
-+ }
-+
-+ error = alloc_chrdev_region(&dev, 0,
-+ FSL_MC_BUS_MAX_MINORS,
-+ "fsl_mc_bus");
-+ if (error < 0)
-+ return error;
-+
-+ fsl_mc_bus_major = MAJOR(dev);
-+
-+ return 0;
-+}
-+
-+static int fsl_mc_restool_dev_open(struct inode *inode, struct file *filep)
-+{
-+ struct fsl_mc_device *root_mc_device;
-+ struct fsl_mc_restool *mc_restool;
-+ struct fsl_mc_bus *mc_bus;
-+ struct fsl_mc_io *dynamic_mc_io;
-+ int error;
-+
-+ mc_restool = container_of(inode->i_cdev, struct fsl_mc_restool, cdev);
-+ mc_bus = container_of(mc_restool, struct fsl_mc_bus, restool_misc);
-+ root_mc_device = &mc_bus->mc_dev;
-+
-+ mutex_lock(&mc_restool->mutex);
-+
-+ if (!mc_restool->local_instance_in_use) {
-+ filep->private_data = root_mc_device->mc_io;
-+ mc_restool->local_instance_in_use = true;
-+ } else {
-+ dynamic_mc_io = kzalloc(sizeof(*dynamic_mc_io), GFP_KERNEL);
-+ if (!dynamic_mc_io) {
-+ error = -ENOMEM;
-+ goto error_alloc_mc_io;
-+ }
-+
-+ error = fsl_mc_portal_allocate(root_mc_device, 0,
-+ &dynamic_mc_io);
-+ if (error) {
-+ pr_err("Could not allocate MC portal\n");
-+ goto error_portal_allocate;
-+ }
-+
-+ mc_restool->dynamic_instance_count++;
-+ filep->private_data = dynamic_mc_io;
-+ }
-+
-+ mutex_unlock(&mc_restool->mutex);
-+
-+ return 0;
-+
-+error_portal_allocate:
-+ kfree(dynamic_mc_io);
-+
-+error_alloc_mc_io:
-+ mutex_unlock(&mc_restool->mutex);
-+
-+ return error;
-+}
-+
-+static int fsl_mc_restool_dev_release(struct inode *inode, struct file *filep)
-+{
-+ struct fsl_mc_device *root_mc_device;
-+ struct fsl_mc_restool *mc_restool;
-+ struct fsl_mc_bus *mc_bus;
-+ struct fsl_mc_io *mc_io;
-+
-+ mc_restool = container_of(inode->i_cdev, struct fsl_mc_restool, cdev);
-+ mc_bus = container_of(mc_restool, struct fsl_mc_bus, restool_misc);
-+ root_mc_device = &mc_bus->mc_dev;
-+ mc_io = filep->private_data;
-+
-+ mutex_lock(&mc_restool->mutex);
-+
-+ if (WARN_ON(!mc_restool->local_instance_in_use &&
-+ mc_restool->dynamic_instance_count == 0)) {
-+ mutex_unlock(&mc_restool->mutex);
-+ return -EINVAL;
-+ }
-+
-+ if (filep->private_data == root_mc_device->mc_io) {
-+ mc_restool->local_instance_in_use = false;
-+ } else {
-+ fsl_mc_portal_free(mc_io);
-+ kfree(mc_io);
-+ mc_restool->dynamic_instance_count--;
-+ }
-+
-+ filep->private_data = NULL;
-+ mutex_unlock(&mc_restool->mutex);
-+
-+ return 0;
-+}
-+
-+static long fsl_mc_restool_dev_ioctl(struct file *file,
-+ unsigned int cmd,
-+ unsigned long arg)
-+{
-+ int error;
-+
-+ switch (cmd) {
-+ case RESTOOL_SEND_MC_COMMAND:
-+ error = fsl_mc_restool_send_command(arg, file->private_data);
-+ break;
-+ default:
-+ pr_err("%s: unexpected ioctl call number\n", __func__);
-+ error = -EINVAL;
-+ }
-+
-+ return error;
-+}
-+
-+static const struct file_operations fsl_mc_restool_dev_fops = {
-+ .owner = THIS_MODULE,
-+ .open = fsl_mc_restool_dev_open,
-+ .release = fsl_mc_restool_dev_release,
-+ .unlocked_ioctl = fsl_mc_restool_dev_ioctl,
-+};
-+
-+int fsl_mc_restool_create_device_file(struct fsl_mc_bus *mc_bus)
-+{
-+ struct fsl_mc_device *mc_dev = &mc_bus->mc_dev;
-+ struct fsl_mc_restool *mc_restool = &mc_bus->restool_misc;
-+ int error;
-+
-+ mc_restool = &mc_bus->restool_misc;
-+ mc_restool->dev = MKDEV(fsl_mc_bus_major, 0);
-+ cdev_init(&mc_restool->cdev, &fsl_mc_restool_dev_fops);
-+
-+ error = cdev_add(&mc_restool->cdev,
-+ mc_restool->dev,
-+ FSL_MC_BUS_MAX_MINORS);
-+ if (error)
-+ return error;
-+
-+ mc_restool->device = device_create(fsl_mc_bus_class,
-+ NULL,
-+ mc_restool->dev,
-+ NULL,
-+ "%s",
-+ dev_name(&mc_dev->dev));
-+ if (IS_ERR(mc_restool->device)) {
-+ error = PTR_ERR(mc_restool->device);
-+ goto error_device_create;
-+ }
-+
-+ mutex_init(&mc_restool->mutex);
-+
-+ return 0;
-+
-+error_device_create:
-+ cdev_del(&mc_restool->cdev);
-+
-+ return error;
-+}
-+
-+void fsl_mc_restool_remove_device_file(struct fsl_mc_bus *mc_bus)
-+{
-+ struct fsl_mc_restool *mc_restool = &mc_bus->restool_misc;
-+
-+ if (WARN_ON(mc_restool->local_instance_in_use))
-+ return;
-+
-+ if (WARN_ON(mc_restool->dynamic_instance_count != 0))
-+ return;
-+
-+ cdev_del(&mc_restool->cdev);
-+}
---- a/drivers/staging/fsl-mc/bus/mc-io.c
-+++ /dev/null
-@@ -1,320 +0,0 @@
--/* Copyright 2013-2016 Freescale Semiconductor Inc.
-- *
-- * Redistribution and use in source and binary forms, with or without
-- * modification, are permitted provided that the following conditions are met:
-- * * Redistributions of source code must retain the above copyright
-- * notice, this list of conditions and the following disclaimer.
-- * * Redistributions in binary form must reproduce the above copyright
-- * notice, this list of conditions and the following disclaimer in the
-- * documentation and/or other materials provided with the distribution.
-- * * Neither the name of the above-listed copyright holders nor the
-- * names of any contributors may be used to endorse or promote products
-- * derived from this software without specific prior written permission.
-- *
-- *
-- * ALTERNATIVELY, this software may be distributed under the terms of the
-- * GNU General Public License ("GPL") as published by the Free Software
-- * Foundation, either version 2 of that License or (at your option) any
-- * later version.
-- *
-- * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
-- * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
-- * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
-- * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE
-- * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
-- * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
-- * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
-- * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
-- * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
-- * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
-- * POSSIBILITY OF SUCH DAMAGE.
-- */
--
--#include <linux/io.h>
--#include "../include/mc-bus.h"
--#include "../include/mc-sys.h"
--
--#include "fsl-mc-private.h"
--#include "dpmcp.h"
--#include "dpmcp-cmd.h"
--
--static int fsl_mc_io_set_dpmcp(struct fsl_mc_io *mc_io,
-- struct fsl_mc_device *dpmcp_dev)
--{
-- int error;
--
-- if (WARN_ON(!dpmcp_dev))
-- return -EINVAL;
--
-- if (WARN_ON(mc_io->dpmcp_dev))
-- return -EINVAL;
--
-- if (WARN_ON(dpmcp_dev->mc_io))
-- return -EINVAL;
--
-- error = dpmcp_open(mc_io,
-- 0,
-- dpmcp_dev->obj_desc.id,
-- &dpmcp_dev->mc_handle);
-- if (error < 0)
-- return error;
--
-- mc_io->dpmcp_dev = dpmcp_dev;
-- dpmcp_dev->mc_io = mc_io;
-- return 0;
--}
--
--static void fsl_mc_io_unset_dpmcp(struct fsl_mc_io *mc_io)
--{
-- int error;
-- struct fsl_mc_device *dpmcp_dev = mc_io->dpmcp_dev;
--
-- if (WARN_ON(!dpmcp_dev))
-- return;
--
-- if (WARN_ON(dpmcp_dev->mc_io != mc_io))
-- return;
--
-- error = dpmcp_close(mc_io,
-- 0,
-- dpmcp_dev->mc_handle);
-- if (error < 0) {
-- dev_err(&dpmcp_dev->dev, "dpmcp_close() failed: %d\n",
-- error);
-- }
--
-- mc_io->dpmcp_dev = NULL;
-- dpmcp_dev->mc_io = NULL;
--}
--
--/**
-- * Creates an MC I/O object
-- *
-- * @dev: device to be associated with the MC I/O object
-- * @mc_portal_phys_addr: physical address of the MC portal to use
-- * @mc_portal_size: size in bytes of the MC portal
-- * @dpmcp-dev: Pointer to the DPMCP object associated with this MC I/O
-- * object or NULL if none.
-- * @flags: flags for the new MC I/O object
-- * @new_mc_io: Area to return pointer to newly created MC I/O object
-- *
-- * Returns '0' on Success; Error code otherwise.
-- */
--int __must_check fsl_create_mc_io(struct device *dev,
-- phys_addr_t mc_portal_phys_addr,
-- u32 mc_portal_size,
-- struct fsl_mc_device *dpmcp_dev,
-- u32 flags, struct fsl_mc_io **new_mc_io)
--{
-- int error;
-- struct fsl_mc_io *mc_io;
-- void __iomem *mc_portal_virt_addr;
-- struct resource *res;
--
-- mc_io = devm_kzalloc(dev, sizeof(*mc_io), GFP_KERNEL);
-- if (!mc_io)
-- return -ENOMEM;
--
-- mc_io->dev = dev;
-- mc_io->flags = flags;
-- mc_io->portal_phys_addr = mc_portal_phys_addr;
-- mc_io->portal_size = mc_portal_size;
-- if (flags & FSL_MC_IO_ATOMIC_CONTEXT_PORTAL)
-- spin_lock_init(&mc_io->spinlock);
-- else
-- mutex_init(&mc_io->mutex);
--
-- res = devm_request_mem_region(dev,
-- mc_portal_phys_addr,
-- mc_portal_size,
-- "mc_portal");
-- if (!res) {
-- dev_err(dev,
-- "devm_request_mem_region failed for MC portal %#llx\n",
-- mc_portal_phys_addr);
-- return -EBUSY;
-- }
--
-- mc_portal_virt_addr = devm_ioremap_nocache(dev,
-- mc_portal_phys_addr,
-- mc_portal_size);
-- if (!mc_portal_virt_addr) {
-- dev_err(dev,
-- "devm_ioremap_nocache failed for MC portal %#llx\n",
-- mc_portal_phys_addr);
-- return -ENXIO;
-- }
--
-- mc_io->portal_virt_addr = mc_portal_virt_addr;
-- if (dpmcp_dev) {
-- error = fsl_mc_io_set_dpmcp(mc_io, dpmcp_dev);
-- if (error < 0)
-- goto error_destroy_mc_io;
-- }
--
-- *new_mc_io = mc_io;
-- return 0;
--
--error_destroy_mc_io:
-- fsl_destroy_mc_io(mc_io);
-- return error;
--}
--
--/**
-- * Destroys an MC I/O object
-- *
-- * @mc_io: MC I/O object to destroy
-- */
--void fsl_destroy_mc_io(struct fsl_mc_io *mc_io)
--{
-- struct fsl_mc_device *dpmcp_dev = mc_io->dpmcp_dev;
--
-- if (dpmcp_dev)
-- fsl_mc_io_unset_dpmcp(mc_io);
--
-- devm_iounmap(mc_io->dev, mc_io->portal_virt_addr);
-- devm_release_mem_region(mc_io->dev,
-- mc_io->portal_phys_addr,
-- mc_io->portal_size);
--
-- mc_io->portal_virt_addr = NULL;
-- devm_kfree(mc_io->dev, mc_io);
--}
--
--/**
-- * fsl_mc_portal_allocate - Allocates an MC portal
-- *
-- * @mc_dev: MC device for which the MC portal is to be allocated
-- * @mc_io_flags: Flags for the fsl_mc_io object that wraps the allocated
-- * MC portal.
-- * @new_mc_io: Pointer to area where the pointer to the fsl_mc_io object
-- * that wraps the allocated MC portal is to be returned
-- *
-- * This function allocates an MC portal from the device's parent DPRC,
-- * from the corresponding MC bus' pool of MC portals and wraps
-- * it in a new fsl_mc_io object. If 'mc_dev' is a DPRC itself, the
-- * portal is allocated from its own MC bus.
-- */
--int __must_check fsl_mc_portal_allocate(struct fsl_mc_device *mc_dev,
-- u16 mc_io_flags,
-- struct fsl_mc_io **new_mc_io)
--{
-- struct fsl_mc_device *mc_bus_dev;
-- struct fsl_mc_bus *mc_bus;
-- phys_addr_t mc_portal_phys_addr;
-- size_t mc_portal_size;
-- struct fsl_mc_device *dpmcp_dev;
-- int error = -EINVAL;
-- struct fsl_mc_resource *resource = NULL;
-- struct fsl_mc_io *mc_io = NULL;
--
-- if (mc_dev->flags & FSL_MC_IS_DPRC) {
-- mc_bus_dev = mc_dev;
-- } else {
-- if (WARN_ON(!dev_is_fsl_mc(mc_dev->dev.parent)))
-- return error;
--
-- mc_bus_dev = to_fsl_mc_device(mc_dev->dev.parent);
-- }
--
-- mc_bus = to_fsl_mc_bus(mc_bus_dev);
-- *new_mc_io = NULL;
-- error = fsl_mc_resource_allocate(mc_bus, FSL_MC_POOL_DPMCP, &resource);
-- if (error < 0)
-- return error;
--
-- error = -EINVAL;
-- dpmcp_dev = resource->data;
-- if (WARN_ON(!dpmcp_dev))
-- goto error_cleanup_resource;
--
-- if (dpmcp_dev->obj_desc.ver_major < DPMCP_MIN_VER_MAJOR ||
-- (dpmcp_dev->obj_desc.ver_major == DPMCP_MIN_VER_MAJOR &&
-- dpmcp_dev->obj_desc.ver_minor < DPMCP_MIN_VER_MINOR)) {
-- dev_err(&dpmcp_dev->dev,
-- "ERROR: Version %d.%d of DPMCP not supported.\n",
-- dpmcp_dev->obj_desc.ver_major,
-- dpmcp_dev->obj_desc.ver_minor);
-- error = -ENOTSUPP;
-- goto error_cleanup_resource;
-- }
--
-- if (WARN_ON(dpmcp_dev->obj_desc.region_count == 0))
-- goto error_cleanup_resource;
--
-- mc_portal_phys_addr = dpmcp_dev->regions[0].start;
-- mc_portal_size = dpmcp_dev->regions[0].end -
-- dpmcp_dev->regions[0].start + 1;
--
-- if (WARN_ON(mc_portal_size != mc_bus_dev->mc_io->portal_size))
-- goto error_cleanup_resource;
--
-- error = fsl_create_mc_io(&mc_bus_dev->dev,
-- mc_portal_phys_addr,
-- mc_portal_size, dpmcp_dev,
-- mc_io_flags, &mc_io);
-- if (error < 0)
-- goto error_cleanup_resource;
--
-- *new_mc_io = mc_io;
-- return 0;
--
--error_cleanup_resource:
-- fsl_mc_resource_free(resource);
-- return error;
--}
--EXPORT_SYMBOL_GPL(fsl_mc_portal_allocate);
--
--/**
-- * fsl_mc_portal_free - Returns an MC portal to the pool of free MC portals
-- * of a given MC bus
-- *
-- * @mc_io: Pointer to the fsl_mc_io object that wraps the MC portal to free
-- */
--void fsl_mc_portal_free(struct fsl_mc_io *mc_io)
--{
-- struct fsl_mc_device *dpmcp_dev;
-- struct fsl_mc_resource *resource;
--
-- /*
-- * Every mc_io obtained by calling fsl_mc_portal_allocate() is supposed
-- * to have a DPMCP object associated with.
-- */
-- dpmcp_dev = mc_io->dpmcp_dev;
-- if (WARN_ON(!dpmcp_dev))
-- return;
--
-- resource = dpmcp_dev->resource;
-- if (WARN_ON(!resource || resource->type != FSL_MC_POOL_DPMCP))
-- return;
--
-- if (WARN_ON(resource->data != dpmcp_dev))
-- return;
--
-- fsl_destroy_mc_io(mc_io);
-- fsl_mc_resource_free(resource);
--}
--EXPORT_SYMBOL_GPL(fsl_mc_portal_free);
--
--/**
-- * fsl_mc_portal_reset - Resets the dpmcp object for a given fsl_mc_io object
-- *
-- * @mc_io: Pointer to the fsl_mc_io object that wraps the MC portal to free
-- */
--int fsl_mc_portal_reset(struct fsl_mc_io *mc_io)
--{
-- int error;
-- struct fsl_mc_device *dpmcp_dev = mc_io->dpmcp_dev;
--
-- if (WARN_ON(!dpmcp_dev))
-- return -EINVAL;
--
-- error = dpmcp_reset(mc_io, 0, dpmcp_dev->mc_handle);
-- if (error < 0) {
-- dev_err(&dpmcp_dev->dev, "dpmcp_reset() failed: %d\n", error);
-- return error;
-- }
--
-- return 0;
--}
--EXPORT_SYMBOL_GPL(fsl_mc_portal_reset);
---- /dev/null
-+++ b/drivers/bus/fsl-mc/mc-io.c
-@@ -0,0 +1,268 @@
-+// SPDX-License-Identifier: (GPL-2.0+ OR BSD-3-Clause)
-+/*
-+ * Copyright 2013-2016 Freescale Semiconductor Inc.
-+ *
-+ */
-+
-+#include <linux/io.h>
-+#include <linux/fsl/mc.h>
-+
-+#include "fsl-mc-private.h"
-+
-+static int fsl_mc_io_set_dpmcp(struct fsl_mc_io *mc_io,
-+ struct fsl_mc_device *dpmcp_dev)
-+{
-+ int error;
-+
-+ if (mc_io->dpmcp_dev)
-+ return -EINVAL;
-+
-+ if (dpmcp_dev->mc_io)
-+ return -EINVAL;
-+
-+ error = dpmcp_open(mc_io,
-+ 0,
-+ dpmcp_dev->obj_desc.id,
-+ &dpmcp_dev->mc_handle);
-+ if (error < 0)
-+ return error;
-+
-+ mc_io->dpmcp_dev = dpmcp_dev;
-+ dpmcp_dev->mc_io = mc_io;
-+ return 0;
-+}
-+
-+static void fsl_mc_io_unset_dpmcp(struct fsl_mc_io *mc_io)
-+{
-+ int error;
-+ struct fsl_mc_device *dpmcp_dev = mc_io->dpmcp_dev;
-+
-+ error = dpmcp_close(mc_io,
-+ 0,
-+ dpmcp_dev->mc_handle);
-+ if (error < 0) {
-+ dev_err(&dpmcp_dev->dev, "dpmcp_close() failed: %d\n",
-+ error);
-+ }
-+
-+ mc_io->dpmcp_dev = NULL;
-+ dpmcp_dev->mc_io = NULL;
-+}
-+
-+/**
-+ * Creates an MC I/O object
-+ *
-+ * @dev: device to be associated with the MC I/O object
-+ * @mc_portal_phys_addr: physical address of the MC portal to use
-+ * @mc_portal_size: size in bytes of the MC portal
-+ * @dpmcp-dev: Pointer to the DPMCP object associated with this MC I/O
-+ * object or NULL if none.
-+ * @flags: flags for the new MC I/O object
-+ * @new_mc_io: Area to return pointer to newly created MC I/O object
-+ *
-+ * Returns '0' on Success; Error code otherwise.
-+ */
-+int __must_check fsl_create_mc_io(struct device *dev,
-+ phys_addr_t mc_portal_phys_addr,
-+ u32 mc_portal_size,
-+ struct fsl_mc_device *dpmcp_dev,
-+ u32 flags, struct fsl_mc_io **new_mc_io)
-+{
-+ int error;
-+ struct fsl_mc_io *mc_io;
-+ void __iomem *mc_portal_virt_addr;
-+ struct resource *res;
-+
-+ mc_io = devm_kzalloc(dev, sizeof(*mc_io), GFP_KERNEL);
-+ if (!mc_io)
-+ return -ENOMEM;
-+
-+ mc_io->dev = dev;
-+ mc_io->flags = flags;
-+ mc_io->portal_phys_addr = mc_portal_phys_addr;
-+ mc_io->portal_size = mc_portal_size;
-+ if (flags & FSL_MC_IO_ATOMIC_CONTEXT_PORTAL)
-+ spin_lock_init(&mc_io->spinlock);
-+ else
-+ mutex_init(&mc_io->mutex);
-+
-+ res = devm_request_mem_region(dev,
-+ mc_portal_phys_addr,
-+ mc_portal_size,
-+ "mc_portal");
-+ if (!res) {
-+ dev_err(dev,
-+ "devm_request_mem_region failed for MC portal %pa\n",
-+ &mc_portal_phys_addr);
-+ return -EBUSY;
-+ }
-+
-+ mc_portal_virt_addr = devm_ioremap_nocache(dev,
-+ mc_portal_phys_addr,
-+ mc_portal_size);
-+ if (!mc_portal_virt_addr) {
-+ dev_err(dev,
-+ "devm_ioremap_nocache failed for MC portal %pa\n",
-+ &mc_portal_phys_addr);
-+ return -ENXIO;
-+ }
-+
-+ mc_io->portal_virt_addr = mc_portal_virt_addr;
-+ if (dpmcp_dev) {
-+ error = fsl_mc_io_set_dpmcp(mc_io, dpmcp_dev);
-+ if (error < 0)
-+ goto error_destroy_mc_io;
-+ }
-+
-+ *new_mc_io = mc_io;
-+ return 0;
-+
-+error_destroy_mc_io:
-+ fsl_destroy_mc_io(mc_io);
-+ return error;
-+}
-+
-+/**
-+ * Destroys an MC I/O object
-+ *
-+ * @mc_io: MC I/O object to destroy
-+ */
-+void fsl_destroy_mc_io(struct fsl_mc_io *mc_io)
-+{
-+ struct fsl_mc_device *dpmcp_dev = mc_io->dpmcp_dev;
-+
-+ if (dpmcp_dev)
-+ fsl_mc_io_unset_dpmcp(mc_io);
-+
-+ devm_iounmap(mc_io->dev, mc_io->portal_virt_addr);
-+ devm_release_mem_region(mc_io->dev,
-+ mc_io->portal_phys_addr,
-+ mc_io->portal_size);
-+
-+ mc_io->portal_virt_addr = NULL;
-+ devm_kfree(mc_io->dev, mc_io);
-+}
-+
-+/**
-+ * fsl_mc_portal_allocate - Allocates an MC portal
-+ *
-+ * @mc_dev: MC device for which the MC portal is to be allocated
-+ * @mc_io_flags: Flags for the fsl_mc_io object that wraps the allocated
-+ * MC portal.
-+ * @new_mc_io: Pointer to area where the pointer to the fsl_mc_io object
-+ * that wraps the allocated MC portal is to be returned
-+ *
-+ * This function allocates an MC portal from the device's parent DPRC,
-+ * from the corresponding MC bus' pool of MC portals and wraps
-+ * it in a new fsl_mc_io object. If 'mc_dev' is a DPRC itself, the
-+ * portal is allocated from its own MC bus.
-+ */
-+int __must_check fsl_mc_portal_allocate(struct fsl_mc_device *mc_dev,
-+ u16 mc_io_flags,
-+ struct fsl_mc_io **new_mc_io)
-+{
-+ struct fsl_mc_device *mc_bus_dev;
-+ struct fsl_mc_bus *mc_bus;
-+ phys_addr_t mc_portal_phys_addr;
-+ size_t mc_portal_size;
-+ struct fsl_mc_device *dpmcp_dev;
-+ int error = -EINVAL;
-+ struct fsl_mc_resource *resource = NULL;
-+ struct fsl_mc_io *mc_io = NULL;
-+
-+ if (mc_dev->flags & FSL_MC_IS_DPRC) {
-+ mc_bus_dev = mc_dev;
-+ } else {
-+ if (!dev_is_fsl_mc(mc_dev->dev.parent))
-+ return error;
-+
-+ mc_bus_dev = to_fsl_mc_device(mc_dev->dev.parent);
-+ }
-+
-+ mc_bus = to_fsl_mc_bus(mc_bus_dev);
-+ *new_mc_io = NULL;
-+ error = fsl_mc_resource_allocate(mc_bus, FSL_MC_POOL_DPMCP, &resource);
-+ if (error < 0)
-+ return error;
-+
-+ error = -EINVAL;
-+ dpmcp_dev = resource->data;
-+
-+ if (dpmcp_dev->obj_desc.ver_major < DPMCP_MIN_VER_MAJOR ||
-+ (dpmcp_dev->obj_desc.ver_major == DPMCP_MIN_VER_MAJOR &&
-+ dpmcp_dev->obj_desc.ver_minor < DPMCP_MIN_VER_MINOR)) {
-+ dev_err(&dpmcp_dev->dev,
-+ "ERROR: Version %d.%d of DPMCP not supported.\n",
-+ dpmcp_dev->obj_desc.ver_major,
-+ dpmcp_dev->obj_desc.ver_minor);
-+ error = -ENOTSUPP;
-+ goto error_cleanup_resource;
-+ }
-+
-+ mc_portal_phys_addr = dpmcp_dev->regions[0].start;
-+ mc_portal_size = resource_size(dpmcp_dev->regions);
-+
-+ error = fsl_create_mc_io(&mc_bus_dev->dev,
-+ mc_portal_phys_addr,
-+ mc_portal_size, dpmcp_dev,
-+ mc_io_flags, &mc_io);
-+ if (error < 0)
-+ goto error_cleanup_resource;
-+
-+ *new_mc_io = mc_io;
-+ return 0;
-+
-+error_cleanup_resource:
-+ fsl_mc_resource_free(resource);
-+ return error;
-+}
-+EXPORT_SYMBOL_GPL(fsl_mc_portal_allocate);
-+
-+/**
-+ * fsl_mc_portal_free - Returns an MC portal to the pool of free MC portals
-+ * of a given MC bus
-+ *
-+ * @mc_io: Pointer to the fsl_mc_io object that wraps the MC portal to free
-+ */
-+void fsl_mc_portal_free(struct fsl_mc_io *mc_io)
-+{
-+ struct fsl_mc_device *dpmcp_dev;
-+ struct fsl_mc_resource *resource;
-+
-+ /*
-+ * Every mc_io obtained by calling fsl_mc_portal_allocate() is supposed
-+ * to have a DPMCP object associated with.
-+ */
-+ dpmcp_dev = mc_io->dpmcp_dev;
-+
-+ resource = dpmcp_dev->resource;
-+ if (!resource || resource->type != FSL_MC_POOL_DPMCP)
-+ return;
-+
-+ if (resource->data != dpmcp_dev)
-+ return;
-+
-+ fsl_destroy_mc_io(mc_io);
-+ fsl_mc_resource_free(resource);
-+}
-+EXPORT_SYMBOL_GPL(fsl_mc_portal_free);
-+
-+/**
-+ * fsl_mc_portal_reset - Resets the dpmcp object for a given fsl_mc_io object
-+ *
-+ * @mc_io: Pointer to the fsl_mc_io object that wraps the MC portal to free
-+ */
-+int fsl_mc_portal_reset(struct fsl_mc_io *mc_io)
-+{
-+ int error;
-+ struct fsl_mc_device *dpmcp_dev = mc_io->dpmcp_dev;
-+
-+ error = dpmcp_reset(mc_io, 0, dpmcp_dev->mc_handle);
-+ if (error < 0) {
-+ dev_err(&dpmcp_dev->dev, "dpmcp_reset() failed: %d\n", error);
-+ return error;
-+ }
-+
-+ return 0;
-+}
-+EXPORT_SYMBOL_GPL(fsl_mc_portal_reset);
---- a/drivers/staging/fsl-mc/bus/mc-sys.c
-+++ /dev/null
-@@ -1,317 +0,0 @@
--/* Copyright 2013-2014 Freescale Semiconductor Inc.
-- *
-- * I/O services to send MC commands to the MC hardware
-- *
-- * Redistribution and use in source and binary forms, with or without
-- * modification, are permitted provided that the following conditions are met:
-- * * Redistributions of source code must retain the above copyright
-- * notice, this list of conditions and the following disclaimer.
-- * * Redistributions in binary form must reproduce the above copyright
-- * notice, this list of conditions and the following disclaimer in the
-- * documentation and/or other materials provided with the distribution.
-- * * Neither the name of the above-listed copyright holders nor the
-- * names of any contributors may be used to endorse or promote products
-- * derived from this software without specific prior written permission.
-- *
-- *
-- * ALTERNATIVELY, this software may be distributed under the terms of the
-- * GNU General Public License ("GPL") as published by the Free Software
-- * Foundation, either version 2 of that License or (at your option) any
-- * later version.
-- *
-- * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
-- * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
-- * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
-- * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE
-- * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
-- * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
-- * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
-- * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
-- * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
-- * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
-- * POSSIBILITY OF SUCH DAMAGE.
-- */
--
--#include <linux/delay.h>
--#include <linux/slab.h>
--#include <linux/ioport.h>
--#include <linux/device.h>
--#include <linux/io.h>
--#include "../include/mc-sys.h"
--#include "../include/mc-cmd.h"
--#include "../include/mc.h"
--
--#include "dpmcp.h"
--
--/**
-- * Timeout in milliseconds to wait for the completion of an MC command
-- */
--#define MC_CMD_COMPLETION_TIMEOUT_MS 500
--
--/*
-- * usleep_range() min and max values used to throttle down polling
-- * iterations while waiting for MC command completion
-- */
--#define MC_CMD_COMPLETION_POLLING_MIN_SLEEP_USECS 10
--#define MC_CMD_COMPLETION_POLLING_MAX_SLEEP_USECS 500
--
--static enum mc_cmd_status mc_cmd_hdr_read_status(struct mc_command *cmd)
--{
-- struct mc_cmd_header *hdr = (struct mc_cmd_header *)&cmd->header;
--
-- return (enum mc_cmd_status)hdr->status;
--}
--
--static u16 mc_cmd_hdr_read_cmdid(struct mc_command *cmd)
--{
-- struct mc_cmd_header *hdr = (struct mc_cmd_header *)&cmd->header;
-- u16 cmd_id = le16_to_cpu(hdr->cmd_id);
--
-- return (cmd_id & MC_CMD_HDR_CMDID_MASK) >> MC_CMD_HDR_CMDID_SHIFT;
--}
--
--static int mc_status_to_error(enum mc_cmd_status status)
--{
-- static const int mc_status_to_error_map[] = {
-- [MC_CMD_STATUS_OK] = 0,
-- [MC_CMD_STATUS_AUTH_ERR] = -EACCES,
-- [MC_CMD_STATUS_NO_PRIVILEGE] = -EPERM,
-- [MC_CMD_STATUS_DMA_ERR] = -EIO,
-- [MC_CMD_STATUS_CONFIG_ERR] = -ENXIO,
-- [MC_CMD_STATUS_TIMEOUT] = -ETIMEDOUT,
-- [MC_CMD_STATUS_NO_RESOURCE] = -ENAVAIL,
-- [MC_CMD_STATUS_NO_MEMORY] = -ENOMEM,
-- [MC_CMD_STATUS_BUSY] = -EBUSY,
-- [MC_CMD_STATUS_UNSUPPORTED_OP] = -ENOTSUPP,
-- [MC_CMD_STATUS_INVALID_STATE] = -ENODEV,
-- };
--
-- if (WARN_ON((u32)status >= ARRAY_SIZE(mc_status_to_error_map)))
-- return -EINVAL;
--
-- return mc_status_to_error_map[status];
--}
--
--static const char *mc_status_to_string(enum mc_cmd_status status)
--{
-- static const char *const status_strings[] = {
-- [MC_CMD_STATUS_OK] = "Command completed successfully",
-- [MC_CMD_STATUS_READY] = "Command ready to be processed",
-- [MC_CMD_STATUS_AUTH_ERR] = "Authentication error",
-- [MC_CMD_STATUS_NO_PRIVILEGE] = "No privilege",
-- [MC_CMD_STATUS_DMA_ERR] = "DMA or I/O error",
-- [MC_CMD_STATUS_CONFIG_ERR] = "Configuration error",
-- [MC_CMD_STATUS_TIMEOUT] = "Operation timed out",
-- [MC_CMD_STATUS_NO_RESOURCE] = "No resources",
-- [MC_CMD_STATUS_NO_MEMORY] = "No memory available",
-- [MC_CMD_STATUS_BUSY] = "Device is busy",
-- [MC_CMD_STATUS_UNSUPPORTED_OP] = "Unsupported operation",
-- [MC_CMD_STATUS_INVALID_STATE] = "Invalid state"
-- };
--
-- if ((unsigned int)status >= ARRAY_SIZE(status_strings))
-- return "Unknown MC error";
--
-- return status_strings[status];
--}
--
--/**
-- * mc_write_command - writes a command to a Management Complex (MC) portal
-- *
-- * @portal: pointer to an MC portal
-- * @cmd: pointer to a filled command
-- */
--static inline void mc_write_command(struct mc_command __iomem *portal,
-- struct mc_command *cmd)
--{
-- int i;
--
-- /* copy command parameters into the portal */
-- for (i = 0; i < MC_CMD_NUM_OF_PARAMS; i++)
-- __raw_writeq(cmd->params[i], &portal->params[i]);
-- __iowmb();
--
-- /* submit the command by writing the header */
-- __raw_writeq(cmd->header, &portal->header);
--}
--
--/**
-- * mc_read_response - reads the response for the last MC command from a
-- * Management Complex (MC) portal
-- *
-- * @portal: pointer to an MC portal
-- * @resp: pointer to command response buffer
-- *
-- * Returns MC_CMD_STATUS_OK on Success; Error code otherwise.
-- */
--static inline enum mc_cmd_status mc_read_response(struct mc_command __iomem *
-- portal,
-- struct mc_command *resp)
--{
-- int i;
-- enum mc_cmd_status status;
--
-- /* Copy command response header from MC portal: */
-- __iormb();
-- resp->header = __raw_readq(&portal->header);
-- __iormb();
-- status = mc_cmd_hdr_read_status(resp);
-- if (status != MC_CMD_STATUS_OK)
-- return status;
--
-- /* Copy command response data from MC portal: */
-- for (i = 0; i < MC_CMD_NUM_OF_PARAMS; i++)
-- resp->params[i] = __raw_readq(&portal->params[i]);
-- __iormb();
--
-- return status;
--}
--
--/**
-- * Waits for the completion of an MC command doing preemptible polling.
-- * uslepp_range() is called between polling iterations.
-- *
-- * @mc_io: MC I/O object to be used
-- * @cmd: command buffer to receive MC response
-- * @mc_status: MC command completion status
-- */
--static int mc_polling_wait_preemptible(struct fsl_mc_io *mc_io,
-- struct mc_command *cmd,
-- enum mc_cmd_status *mc_status)
--{
-- enum mc_cmd_status status;
-- unsigned long jiffies_until_timeout =
-- jiffies + msecs_to_jiffies(MC_CMD_COMPLETION_TIMEOUT_MS);
--
-- /*
-- * Wait for response from the MC hardware:
-- */
-- for (;;) {
-- status = mc_read_response(mc_io->portal_virt_addr, cmd);
-- if (status != MC_CMD_STATUS_READY)
-- break;
--
-- /*
-- * TODO: When MC command completion interrupts are supported
-- * call wait function here instead of usleep_range()
-- */
-- usleep_range(MC_CMD_COMPLETION_POLLING_MIN_SLEEP_USECS,
-- MC_CMD_COMPLETION_POLLING_MAX_SLEEP_USECS);
--
-- if (time_after_eq(jiffies, jiffies_until_timeout)) {
-- dev_dbg(mc_io->dev,
-- "MC command timed out (portal: %#llx, obj handle: %#x, command: %#x)\n",
-- mc_io->portal_phys_addr,
-- (unsigned int)mc_cmd_hdr_read_token(cmd),
-- (unsigned int)mc_cmd_hdr_read_cmdid(cmd));
--
-- return -ETIMEDOUT;
-- }
-- }
--
-- *mc_status = status;
-- return 0;
--}
--
--/**
-- * Waits for the completion of an MC command doing atomic polling.
-- * udelay() is called between polling iterations.
-- *
-- * @mc_io: MC I/O object to be used
-- * @cmd: command buffer to receive MC response
-- * @mc_status: MC command completion status
-- */
--static int mc_polling_wait_atomic(struct fsl_mc_io *mc_io,
-- struct mc_command *cmd,
-- enum mc_cmd_status *mc_status)
--{
-- enum mc_cmd_status status;
-- unsigned long timeout_usecs = MC_CMD_COMPLETION_TIMEOUT_MS * 1000;
--
-- BUILD_BUG_ON((MC_CMD_COMPLETION_TIMEOUT_MS * 1000) %
-- MC_CMD_COMPLETION_POLLING_MAX_SLEEP_USECS != 0);
--
-- for (;;) {
-- status = mc_read_response(mc_io->portal_virt_addr, cmd);
-- if (status != MC_CMD_STATUS_READY)
-- break;
--
-- udelay(MC_CMD_COMPLETION_POLLING_MAX_SLEEP_USECS);
-- timeout_usecs -= MC_CMD_COMPLETION_POLLING_MAX_SLEEP_USECS;
-- if (timeout_usecs == 0) {
-- dev_dbg(mc_io->dev,
-- "MC command timed out (portal: %#llx, obj handle: %#x, command: %#x)\n",
-- mc_io->portal_phys_addr,
-- (unsigned int)mc_cmd_hdr_read_token(cmd),
-- (unsigned int)mc_cmd_hdr_read_cmdid(cmd));
--
-- return -ETIMEDOUT;
-- }
-- }
--
-- *mc_status = status;
-- return 0;
--}
--
--/**
-- * Sends a command to the MC device using the given MC I/O object
-- *
-- * @mc_io: MC I/O object to be used
-- * @cmd: command to be sent
-- *
-- * Returns '0' on Success; Error code otherwise.
-- */
--int mc_send_command(struct fsl_mc_io *mc_io, struct mc_command *cmd)
--{
-- int error;
-- enum mc_cmd_status status;
-- unsigned long irq_flags = 0;
--
-- if (WARN_ON(in_irq() &&
-- !(mc_io->flags & FSL_MC_IO_ATOMIC_CONTEXT_PORTAL)))
-- return -EINVAL;
--
-- if (mc_io->flags & FSL_MC_IO_ATOMIC_CONTEXT_PORTAL)
-- spin_lock_irqsave(&mc_io->spinlock, irq_flags);
-- else
-- mutex_lock(&mc_io->mutex);
--
-- /*
-- * Send command to the MC hardware:
-- */
-- mc_write_command(mc_io->portal_virt_addr, cmd);
--
-- /*
-- * Wait for response from the MC hardware:
-- */
-- if (!(mc_io->flags & FSL_MC_IO_ATOMIC_CONTEXT_PORTAL))
-- error = mc_polling_wait_preemptible(mc_io, cmd, &status);
-- else
-- error = mc_polling_wait_atomic(mc_io, cmd, &status);
--
-- if (error < 0)
-- goto common_exit;
--
-- if (status != MC_CMD_STATUS_OK) {
-- dev_dbg(mc_io->dev,
-- "MC command failed: portal: %#llx, obj handle: %#x, command: %#x, status: %s (%#x)\n",
-- mc_io->portal_phys_addr,
-- (unsigned int)mc_cmd_hdr_read_token(cmd),
-- (unsigned int)mc_cmd_hdr_read_cmdid(cmd),
-- mc_status_to_string(status),
-- (unsigned int)status);
--
-- error = mc_status_to_error(status);
-- goto common_exit;
-- }
--
-- error = 0;
--common_exit:
-- if (mc_io->flags & FSL_MC_IO_ATOMIC_CONTEXT_PORTAL)
-- spin_unlock_irqrestore(&mc_io->spinlock, irq_flags);
-- else
-- mutex_unlock(&mc_io->mutex);
--
-- return error;
--}
--EXPORT_SYMBOL(mc_send_command);
---- /dev/null
-+++ b/drivers/bus/fsl-mc/mc-sys.c
-@@ -0,0 +1,296 @@
-+// SPDX-License-Identifier: (GPL-2.0+ OR BSD-3-Clause)
-+/*
-+ * Copyright 2013-2016 Freescale Semiconductor Inc.
-+ *
-+ * I/O services to send MC commands to the MC hardware
-+ *
-+ */
-+
-+#include <linux/delay.h>
-+#include <linux/slab.h>
-+#include <linux/ioport.h>
-+#include <linux/device.h>
-+#include <linux/io.h>
-+#include <linux/io-64-nonatomic-hi-lo.h>
-+#include <linux/fsl/mc.h>
-+
-+#include "fsl-mc-private.h"
-+
-+/**
-+ * Timeout in milliseconds to wait for the completion of an MC command
-+ */
-+#define MC_CMD_COMPLETION_TIMEOUT_MS 15000
-+
-+/*
-+ * usleep_range() min and max values used to throttle down polling
-+ * iterations while waiting for MC command completion
-+ */
-+#define MC_CMD_COMPLETION_POLLING_MIN_SLEEP_USECS 10
-+#define MC_CMD_COMPLETION_POLLING_MAX_SLEEP_USECS 500
-+
-+static enum mc_cmd_status mc_cmd_hdr_read_status(struct fsl_mc_command *cmd)
-+{
-+ struct mc_cmd_header *hdr = (struct mc_cmd_header *)&cmd->header;
-+
-+ return (enum mc_cmd_status)hdr->status;
-+}
-+
-+static u16 mc_cmd_hdr_read_cmdid(struct fsl_mc_command *cmd)
-+{
-+ struct mc_cmd_header *hdr = (struct mc_cmd_header *)&cmd->header;
-+ u16 cmd_id = le16_to_cpu(hdr->cmd_id);
-+
-+ return cmd_id;
-+}
-+
-+static int mc_status_to_error(enum mc_cmd_status status)
-+{
-+ static const int mc_status_to_error_map[] = {
-+ [MC_CMD_STATUS_OK] = 0,
-+ [MC_CMD_STATUS_AUTH_ERR] = -EACCES,
-+ [MC_CMD_STATUS_NO_PRIVILEGE] = -EPERM,
-+ [MC_CMD_STATUS_DMA_ERR] = -EIO,
-+ [MC_CMD_STATUS_CONFIG_ERR] = -ENXIO,
-+ [MC_CMD_STATUS_TIMEOUT] = -ETIMEDOUT,
-+ [MC_CMD_STATUS_NO_RESOURCE] = -ENAVAIL,
-+ [MC_CMD_STATUS_NO_MEMORY] = -ENOMEM,
-+ [MC_CMD_STATUS_BUSY] = -EBUSY,
-+ [MC_CMD_STATUS_UNSUPPORTED_OP] = -ENOTSUPP,
-+ [MC_CMD_STATUS_INVALID_STATE] = -ENODEV,
-+ };
-+
-+ if ((u32)status >= ARRAY_SIZE(mc_status_to_error_map))
-+ return -EINVAL;
-+
-+ return mc_status_to_error_map[status];
-+}
-+
-+static const char *mc_status_to_string(enum mc_cmd_status status)
-+{
-+ static const char *const status_strings[] = {
-+ [MC_CMD_STATUS_OK] = "Command completed successfully",
-+ [MC_CMD_STATUS_READY] = "Command ready to be processed",
-+ [MC_CMD_STATUS_AUTH_ERR] = "Authentication error",
-+ [MC_CMD_STATUS_NO_PRIVILEGE] = "No privilege",
-+ [MC_CMD_STATUS_DMA_ERR] = "DMA or I/O error",
-+ [MC_CMD_STATUS_CONFIG_ERR] = "Configuration error",
-+ [MC_CMD_STATUS_TIMEOUT] = "Operation timed out",
-+ [MC_CMD_STATUS_NO_RESOURCE] = "No resources",
-+ [MC_CMD_STATUS_NO_MEMORY] = "No memory available",
-+ [MC_CMD_STATUS_BUSY] = "Device is busy",
-+ [MC_CMD_STATUS_UNSUPPORTED_OP] = "Unsupported operation",
-+ [MC_CMD_STATUS_INVALID_STATE] = "Invalid state"
-+ };
-+
-+ if ((unsigned int)status >= ARRAY_SIZE(status_strings))
-+ return "Unknown MC error";
-+
-+ return status_strings[status];
-+}
-+
-+/**
-+ * mc_write_command - writes a command to a Management Complex (MC) portal
-+ *
-+ * @portal: pointer to an MC portal
-+ * @cmd: pointer to a filled command
-+ */
-+static inline void mc_write_command(struct fsl_mc_command __iomem *portal,
-+ struct fsl_mc_command *cmd)
-+{
-+ int i;
-+
-+ /* copy command parameters into the portal */
-+ for (i = 0; i < MC_CMD_NUM_OF_PARAMS; i++)
-+ /*
-+ * Data is already in the expected LE byte-order. Do an
-+ * extra LE -> CPU conversion so that the CPU -> LE done in
-+ * the device io write api puts it back in the right order.
-+ */
-+ writeq_relaxed(le64_to_cpu(cmd->params[i]), &portal->params[i]);
-+
-+ /* submit the command by writing the header */
-+ writeq(le64_to_cpu(cmd->header), &portal->header);
-+}
-+
-+/**
-+ * mc_read_response - reads the response for the last MC command from a
-+ * Management Complex (MC) portal
-+ *
-+ * @portal: pointer to an MC portal
-+ * @resp: pointer to command response buffer
-+ *
-+ * Returns MC_CMD_STATUS_OK on Success; Error code otherwise.
-+ */
-+static inline enum mc_cmd_status mc_read_response(struct fsl_mc_command __iomem
-+ *portal,
-+ struct fsl_mc_command *resp)
-+{
-+ int i;
-+ enum mc_cmd_status status;
-+
-+ /* Copy command response header from MC portal: */
-+ resp->header = cpu_to_le64(readq_relaxed(&portal->header));
-+ status = mc_cmd_hdr_read_status(resp);
-+ if (status != MC_CMD_STATUS_OK)
-+ return status;
-+
-+ /* Copy command response data from MC portal: */
-+ for (i = 0; i < MC_CMD_NUM_OF_PARAMS; i++)
-+ /*
-+ * Data is expected to be in LE byte-order. Do an
-+ * extra CPU -> LE to revert the LE -> CPU done in
-+ * the device io read api.
-+ */
-+ resp->params[i] =
-+ cpu_to_le64(readq_relaxed(&portal->params[i]));
-+
-+ return status;
-+}
-+
-+/**
-+ * Waits for the completion of an MC command doing preemptible polling.
-+ * uslepp_range() is called between polling iterations.
-+ *
-+ * @mc_io: MC I/O object to be used
-+ * @cmd: command buffer to receive MC response
-+ * @mc_status: MC command completion status
-+ */
-+static int mc_polling_wait_preemptible(struct fsl_mc_io *mc_io,
-+ struct fsl_mc_command *cmd,
-+ enum mc_cmd_status *mc_status)
-+{
-+ enum mc_cmd_status status;
-+ unsigned long jiffies_until_timeout =
-+ jiffies + msecs_to_jiffies(MC_CMD_COMPLETION_TIMEOUT_MS);
-+
-+ /*
-+ * Wait for response from the MC hardware:
-+ */
-+ for (;;) {
-+ status = mc_read_response(mc_io->portal_virt_addr, cmd);
-+ if (status != MC_CMD_STATUS_READY)
-+ break;
-+
-+ /*
-+ * TODO: When MC command completion interrupts are supported
-+ * call wait function here instead of usleep_range()
-+ */
-+ usleep_range(MC_CMD_COMPLETION_POLLING_MIN_SLEEP_USECS,
-+ MC_CMD_COMPLETION_POLLING_MAX_SLEEP_USECS);
-+
-+ if (time_after_eq(jiffies, jiffies_until_timeout)) {
-+ dev_dbg(mc_io->dev,
-+ "MC command timed out (portal: %pa, dprc handle: %#x, command: %#x)\n",
-+ &mc_io->portal_phys_addr,
-+ (unsigned int)mc_cmd_hdr_read_token(cmd),
-+ (unsigned int)mc_cmd_hdr_read_cmdid(cmd));
-+
-+ return -ETIMEDOUT;
-+ }
-+ }
-+
-+ *mc_status = status;
-+ return 0;
-+}
-+
-+/**
-+ * Waits for the completion of an MC command doing atomic polling.
-+ * udelay() is called between polling iterations.
-+ *
-+ * @mc_io: MC I/O object to be used
-+ * @cmd: command buffer to receive MC response
-+ * @mc_status: MC command completion status
-+ */
-+static int mc_polling_wait_atomic(struct fsl_mc_io *mc_io,
-+ struct fsl_mc_command *cmd,
-+ enum mc_cmd_status *mc_status)
-+{
-+ enum mc_cmd_status status;
-+ unsigned long timeout_usecs = MC_CMD_COMPLETION_TIMEOUT_MS * 1000;
-+
-+ BUILD_BUG_ON((MC_CMD_COMPLETION_TIMEOUT_MS * 1000) %
-+ MC_CMD_COMPLETION_POLLING_MAX_SLEEP_USECS != 0);
-+
-+ for (;;) {
-+ status = mc_read_response(mc_io->portal_virt_addr, cmd);
-+ if (status != MC_CMD_STATUS_READY)
-+ break;
-+
-+ udelay(MC_CMD_COMPLETION_POLLING_MAX_SLEEP_USECS);
-+ timeout_usecs -= MC_CMD_COMPLETION_POLLING_MAX_SLEEP_USECS;
-+ if (timeout_usecs == 0) {
-+ dev_dbg(mc_io->dev,
-+ "MC command timed out (portal: %pa, dprc handle: %#x, command: %#x)\n",
-+ &mc_io->portal_phys_addr,
-+ (unsigned int)mc_cmd_hdr_read_token(cmd),
-+ (unsigned int)mc_cmd_hdr_read_cmdid(cmd));
-+
-+ return -ETIMEDOUT;
-+ }
-+ }
-+
-+ *mc_status = status;
-+ return 0;
-+}
-+
-+/**
-+ * Sends a command to the MC device using the given MC I/O object
-+ *
-+ * @mc_io: MC I/O object to be used
-+ * @cmd: command to be sent
-+ *
-+ * Returns '0' on Success; Error code otherwise.
-+ */
-+int mc_send_command(struct fsl_mc_io *mc_io, struct fsl_mc_command *cmd)
-+{
-+ int error;
-+ enum mc_cmd_status status;
-+ unsigned long irq_flags = 0;
-+
-+ if (in_irq() && !(mc_io->flags & FSL_MC_IO_ATOMIC_CONTEXT_PORTAL))
-+ return -EINVAL;
-+
-+ if (mc_io->flags & FSL_MC_IO_ATOMIC_CONTEXT_PORTAL)
-+ spin_lock_irqsave(&mc_io->spinlock, irq_flags);
-+ else
-+ mutex_lock(&mc_io->mutex);
-+
-+ /*
-+ * Send command to the MC hardware:
-+ */
-+ mc_write_command(mc_io->portal_virt_addr, cmd);
-+
-+ /*
-+ * Wait for response from the MC hardware:
-+ */
-+ if (!(mc_io->flags & FSL_MC_IO_ATOMIC_CONTEXT_PORTAL))
-+ error = mc_polling_wait_preemptible(mc_io, cmd, &status);
-+ else
-+ error = mc_polling_wait_atomic(mc_io, cmd, &status);
-+
-+ if (error < 0)
-+ goto common_exit;
-+
-+ if (status != MC_CMD_STATUS_OK) {
-+ dev_dbg(mc_io->dev,
-+ "MC command failed: portal: %pa, dprc handle: %#x, command: %#x, status: %s (%#x)\n",
-+ &mc_io->portal_phys_addr,
-+ (unsigned int)mc_cmd_hdr_read_token(cmd),
-+ (unsigned int)mc_cmd_hdr_read_cmdid(cmd),
-+ mc_status_to_string(status),
-+ (unsigned int)status);
-+
-+ error = mc_status_to_error(status);
-+ goto common_exit;
-+ }
-+
-+ error = 0;
-+common_exit:
-+ if (mc_io->flags & FSL_MC_IO_ATOMIC_CONTEXT_PORTAL)
-+ spin_unlock_irqrestore(&mc_io->spinlock, irq_flags);
-+ else
-+ mutex_unlock(&mc_io->mutex);
-+
-+ return error;
-+}
-+EXPORT_SYMBOL_GPL(mc_send_command);
---- a/drivers/irqchip/Kconfig
-+++ b/drivers/irqchip/Kconfig
-@@ -41,6 +41,12 @@ config ARM_GIC_V3_ITS
- depends on PCI_MSI
- select ACPI_IORT if ACPI
-
-+config ARM_GIC_V3_ITS_FSL_MC
-+ bool
-+ depends on ARM_GIC_V3_ITS
-+ depends on FSL_MC_BUS
-+ default ARM_GIC_V3_ITS
-+
- config ARM_NVIC
- bool
- select IRQ_DOMAIN
---- a/drivers/irqchip/Makefile
-+++ b/drivers/irqchip/Makefile
-@@ -29,6 +29,7 @@ obj-$(CONFIG_ARCH_REALVIEW) += irq-gic-
- obj-$(CONFIG_ARM_GIC_V2M) += irq-gic-v2m.o
- obj-$(CONFIG_ARM_GIC_V3) += irq-gic-v3.o irq-gic-common.o
- obj-$(CONFIG_ARM_GIC_V3_ITS) += irq-gic-v3-its.o irq-gic-v3-its-pci-msi.o irq-gic-v3-its-platform-msi.o
-+obj-$(CONFIG_ARM_GIC_V3_ITS_FSL_MC) += irq-gic-v3-its-fsl-mc-msi.o
- obj-$(CONFIG_PARTITION_PERCPU) += irq-partition-percpu.o
- obj-$(CONFIG_HISILICON_IRQ_MBIGEN) += irq-mbigen.o
- obj-$(CONFIG_ARM_NVIC) += irq-nvic.o
---- a/drivers/staging/fsl-mc/bus/irq-gic-v3-its-fsl-mc-msi.c
-+++ /dev/null
-@@ -1,126 +0,0 @@
--/*
-- * Freescale Management Complex (MC) bus driver MSI support
-- *
-- * Copyright (C) 2015 Freescale Semiconductor, Inc.
-- * Author: German Rivera <German.Rivera@freescale.com>
-- *
-- * This file is licensed under the terms of the GNU General Public
-- * License version 2. This program is licensed "as is" without any
-- * warranty of any kind, whether express or implied.
-- */
--
--#include <linux/of_device.h>
--#include <linux/of_address.h>
--#include <linux/irqchip/arm-gic-v3.h>
--#include <linux/irq.h>
--#include <linux/msi.h>
--#include <linux/of.h>
--#include <linux/of_irq.h>
--#include "../include/mc-bus.h"
--#include "fsl-mc-private.h"
--
--static struct irq_chip its_msi_irq_chip = {
-- .name = "fsl-mc-bus-msi",
-- .irq_mask = irq_chip_mask_parent,
-- .irq_unmask = irq_chip_unmask_parent,
-- .irq_eoi = irq_chip_eoi_parent,
-- .irq_set_affinity = msi_domain_set_affinity
--};
--
--static int its_fsl_mc_msi_prepare(struct irq_domain *msi_domain,
-- struct device *dev,
-- int nvec, msi_alloc_info_t *info)
--{
-- struct fsl_mc_device *mc_bus_dev;
-- struct msi_domain_info *msi_info;
--
-- if (WARN_ON(!dev_is_fsl_mc(dev)))
-- return -EINVAL;
--
-- mc_bus_dev = to_fsl_mc_device(dev);
-- if (WARN_ON(!(mc_bus_dev->flags & FSL_MC_IS_DPRC)))
-- return -EINVAL;
--
-- /*
-- * Set the device Id to be passed to the GIC-ITS:
-- *
-- * NOTE: This device id corresponds to the IOMMU stream ID
-- * associated with the DPRC object (ICID).
-- */
-- info->scratchpad[0].ul = mc_bus_dev->icid;
-- msi_info = msi_get_domain_info(msi_domain->parent);
-- return msi_info->ops->msi_prepare(msi_domain->parent, dev, nvec, info);
--}
--
--static struct msi_domain_ops its_fsl_mc_msi_ops = {
-- .msi_prepare = its_fsl_mc_msi_prepare,
--};
--
--static struct msi_domain_info its_fsl_mc_msi_domain_info = {
-- .flags = (MSI_FLAG_USE_DEF_DOM_OPS | MSI_FLAG_USE_DEF_CHIP_OPS),
-- .ops = &its_fsl_mc_msi_ops,
-- .chip = &its_msi_irq_chip,
--};
--
--static const struct of_device_id its_device_id[] = {
-- { .compatible = "arm,gic-v3-its", },
-- {},
--};
--
--int __init its_fsl_mc_msi_init(void)
--{
-- struct device_node *np;
-- struct irq_domain *parent;
-- struct irq_domain *mc_msi_domain;
--
-- for (np = of_find_matching_node(NULL, its_device_id); np;
-- np = of_find_matching_node(np, its_device_id)) {
-- if (!of_device_is_available(np))
-- continue;
-- if (!of_property_read_bool(np, "msi-controller"))
-- continue;
--
-- parent = irq_find_matching_host(np, DOMAIN_BUS_NEXUS);
-- if (!parent || !msi_get_domain_info(parent)) {
-- pr_err("%s: unable to locate ITS domain\n",
-- np->full_name);
-- continue;
-- }
--
-- mc_msi_domain = fsl_mc_msi_create_irq_domain(
-- of_node_to_fwnode(np),
-- &its_fsl_mc_msi_domain_info,
-- parent);
-- if (!mc_msi_domain) {
-- pr_err("%s: unable to create fsl-mc domain\n",
-- np->full_name);
-- continue;
-- }
--
-- WARN_ON(mc_msi_domain->
-- host_data != &its_fsl_mc_msi_domain_info);
--
-- pr_info("fsl-mc MSI: %s domain created\n", np->full_name);
-- }
--
-- return 0;
--}
--
--void its_fsl_mc_msi_cleanup(void)
--{
-- struct device_node *np;
--
-- for (np = of_find_matching_node(NULL, its_device_id); np;
-- np = of_find_matching_node(np, its_device_id)) {
-- struct irq_domain *mc_msi_domain = irq_find_matching_host(
-- np,
-- DOMAIN_BUS_FSL_MC_MSI);
--
-- if (!of_property_read_bool(np, "msi-controller"))
-- continue;
--
-- if (mc_msi_domain &&
-- mc_msi_domain->host_data == &its_fsl_mc_msi_domain_info)
-- irq_domain_remove(mc_msi_domain);
-- }
--}
---- /dev/null
-+++ b/drivers/irqchip/irq-gic-v3-its-fsl-mc-msi.c
-@@ -0,0 +1,98 @@
-+// SPDX-License-Identifier: GPL-2.0
-+/*
-+ * Freescale Management Complex (MC) bus driver MSI support
-+ *
-+ * Copyright (C) 2015-2016 Freescale Semiconductor, Inc.
-+ * Author: German Rivera <German.Rivera@freescale.com>
-+ *
-+ */
-+
-+#include <linux/of_device.h>
-+#include <linux/of_address.h>
-+#include <linux/irq.h>
-+#include <linux/msi.h>
-+#include <linux/of.h>
-+#include <linux/of_irq.h>
-+#include <linux/fsl/mc.h>
-+
-+static struct irq_chip its_msi_irq_chip = {
-+ .name = "ITS-fMSI",
-+ .irq_mask = irq_chip_mask_parent,
-+ .irq_unmask = irq_chip_unmask_parent,
-+ .irq_eoi = irq_chip_eoi_parent,
-+ .irq_set_affinity = msi_domain_set_affinity
-+};
-+
-+static int its_fsl_mc_msi_prepare(struct irq_domain *msi_domain,
-+ struct device *dev,
-+ int nvec, msi_alloc_info_t *info)
-+{
-+ struct fsl_mc_device *mc_bus_dev;
-+ struct msi_domain_info *msi_info;
-+
-+ if (!dev_is_fsl_mc(dev))
-+ return -EINVAL;
-+
-+ mc_bus_dev = to_fsl_mc_device(dev);
-+ if (!(mc_bus_dev->flags & FSL_MC_IS_DPRC))
-+ return -EINVAL;
-+
-+ /*
-+ * Set the device Id to be passed to the GIC-ITS:
-+ *
-+ * NOTE: This device id corresponds to the IOMMU stream ID
-+ * associated with the DPRC object (ICID).
-+ */
-+ info->scratchpad[0].ul = mc_bus_dev->icid;
-+ msi_info = msi_get_domain_info(msi_domain->parent);
-+ return msi_info->ops->msi_prepare(msi_domain->parent, dev, nvec, info);
-+}
-+
-+static struct msi_domain_ops its_fsl_mc_msi_ops __ro_after_init = {
-+ .msi_prepare = its_fsl_mc_msi_prepare,
-+};
-+
-+static struct msi_domain_info its_fsl_mc_msi_domain_info = {
-+ .flags = (MSI_FLAG_USE_DEF_DOM_OPS | MSI_FLAG_USE_DEF_CHIP_OPS),
-+ .ops = &its_fsl_mc_msi_ops,
-+ .chip = &its_msi_irq_chip,
-+};
-+
-+static const struct of_device_id its_device_id[] = {
-+ { .compatible = "arm,gic-v3-its", },
-+ {},
-+};
-+
-+static int __init its_fsl_mc_msi_init(void)
-+{
-+ struct device_node *np;
-+ struct irq_domain *parent;
-+ struct irq_domain *mc_msi_domain;
-+
-+ for (np = of_find_matching_node(NULL, its_device_id); np;
-+ np = of_find_matching_node(np, its_device_id)) {
-+ if (!of_property_read_bool(np, "msi-controller"))
-+ continue;
-+
-+ parent = irq_find_matching_host(np, DOMAIN_BUS_NEXUS);
-+ if (!parent || !msi_get_domain_info(parent)) {
-+ pr_err("%pOF: unable to locate ITS domain\n", np);
-+ continue;
-+ }
-+
-+ mc_msi_domain = fsl_mc_msi_create_irq_domain(
-+ of_node_to_fwnode(np),
-+ &its_fsl_mc_msi_domain_info,
-+ parent);
-+ if (!mc_msi_domain) {
-+ pr_err("%pOF: unable to create fsl-mc domain\n", np);
-+ continue;
-+ }
-+
-+ pr_info("fsl-mc MSI: %pOF domain created\n", np);
-+ }
-+
-+ return 0;
-+}
-+
-+early_initcall(its_fsl_mc_msi_init);
---- a/drivers/staging/fsl-mc/Kconfig
-+++ b/drivers/staging/fsl-mc/Kconfig
-@@ -1 +1,2 @@
-+# SPDX-License-Identifier: GPL-2.0
- source "drivers/staging/fsl-mc/bus/Kconfig"
---- a/drivers/staging/fsl-mc/Makefile
-+++ b/drivers/staging/fsl-mc/Makefile
-@@ -1,2 +1,3 @@
-+# SPDX-License-Identifier: GPL-2.0
- # Freescale Management Complex (MC) bus drivers
- obj-$(CONFIG_FSL_MC_BUS) += bus/
---- a/drivers/staging/fsl-mc/TODO
-+++ /dev/null
-@@ -1,18 +0,0 @@
--* Add at least one device driver for a DPAA2 object (child device of the
-- fsl-mc bus). Most likely candidate for this is adding DPAA2 Ethernet
-- driver support, which depends on drivers for several objects: DPNI,
-- DPIO, DPMAC. Other pre-requisites include:
--
-- * MC firmware uprev. The MC firmware upon which the fsl-mc
-- bus driver and DPAA2 object drivers are based is continuing
-- to evolve, so minor updates are needed to keep in sync with binary
-- interface changes to the MC.
--
--* Cleanup
--
--Please send any patches to Greg Kroah-Hartman <gregkh@linuxfoundation.org>,
--german.rivera@freescale.com, devel@driverdev.osuosl.org,
--linux-kernel@vger.kernel.org
--
--[1] https://lkml.org/lkml/2015/7/9/93
--[2] https://lkml.org/lkml/2015/7/7/712
---- a/drivers/staging/fsl-mc/bus/Kconfig
-+++ b/drivers/staging/fsl-mc/bus/Kconfig
-@@ -1,25 +1,22 @@
-+# SPDX-License-Identifier: GPL-2.0
- #
--# Freescale Management Complex (MC) bus drivers
-+# DPAA2 fsl-mc bus
- #
--# Copyright (C) 2014 Freescale Semiconductor, Inc.
-+# Copyright (C) 2014-2016 Freescale Semiconductor, Inc.
- #
--# This file is released under the GPLv2
--#
--
--config FSL_MC_BUS
-- bool "Freescale Management Complex (MC) bus driver"
-- depends on OF && ARM64
-- select GENERIC_MSI_IRQ_DOMAIN
-- help
-- Driver to enable the bus infrastructure for the Freescale
-- QorIQ Management Complex (fsl-mc). The fsl-mc is a hardware
-- module of the QorIQ LS2 SoCs, that does resource management
-- for hardware building-blocks in the SoC that can be used
-- to dynamically create networking hardware objects such as
-- network interfaces (NICs), crypto accelerator instances,
-- or L2 switches.
--
-- Only enable this option when building the kernel for
-- Freescale QorQIQ LS2xxxx SoCs.
-
-+config FSL_MC_DPIO
-+ tristate "QorIQ DPAA2 DPIO driver"
-+ depends on FSL_MC_BUS
-+ help
-+ Driver for the DPAA2 DPIO object. A DPIO provides queue and
-+ buffer management facilities for software to interact with
-+ other DPAA2 objects. This driver does not expose the DPIO
-+ objects individually, but groups them under a service layer
-+ API.
-
-+config FSL_QBMAN_DEBUG
-+ tristate "Freescale QBMAN Debug APIs"
-+ depends on FSL_MC_DPIO
-+ help
-+ QBMan debug assistant APIs.
---- a/drivers/staging/fsl-mc/bus/Makefile
-+++ b/drivers/staging/fsl-mc/bus/Makefile
-@@ -1,20 +1,9 @@
-+# SPDX-License-Identifier: GPL-2.0
- #
- # Freescale Management Complex (MC) bus drivers
- #
- # Copyright (C) 2014 Freescale Semiconductor, Inc.
- #
--# This file is released under the GPLv2
--#
--obj-$(CONFIG_FSL_MC_BUS) += mc-bus-driver.o
-
--mc-bus-driver-objs := fsl-mc-bus.o \
-- mc-sys.o \
-- mc-io.o \
-- dprc.o \
-- dpmng.o \
-- dprc-driver.o \
-- fsl-mc-allocator.o \
-- fsl-mc-msi.o \
-- irq-gic-v3-its-fsl-mc-msi.o \
-- dpmcp.o \
-- dpbp.o
-+# MC DPIO driver
-+obj-$(CONFIG_FSL_MC_DPIO) += dpio/
---- a/drivers/staging/fsl-mc/bus/dpbp.c
-+++ /dev/null
-@@ -1,691 +0,0 @@
--/* Copyright 2013-2016 Freescale Semiconductor Inc.
-- *
-- * Redistribution and use in source and binary forms, with or without
-- * modification, are permitted provided that the following conditions are met:
-- * * Redistributions of source code must retain the above copyright
-- * notice, this list of conditions and the following disclaimer.
-- * * Redistributions in binary form must reproduce the above copyright
-- * notice, this list of conditions and the following disclaimer in the
-- * documentation and/or other materials provided with the distribution.
-- * * Neither the name of the above-listed copyright holders nor the
-- * names of any contributors may be used to endorse or promote products
-- * derived from this software without specific prior written permission.
-- *
-- *
-- * ALTERNATIVELY, this software may be distributed under the terms of the
-- * GNU General Public License ("GPL") as published by the Free Software
-- * Foundation, either version 2 of that License or (at your option) any
-- * later version.
-- *
-- * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
-- * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
-- * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
-- * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE
-- * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
-- * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
-- * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
-- * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
-- * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
-- * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
-- * POSSIBILITY OF SUCH DAMAGE.
-- */
--#include "../include/mc-sys.h"
--#include "../include/mc-cmd.h"
--#include "../include/dpbp.h"
--#include "../include/dpbp-cmd.h"
--
--/**
-- * dpbp_open() - Open a control session for the specified object.
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @dpbp_id: DPBP unique ID
-- * @token: Returned token; use in subsequent API calls
-- *
-- * This function can be used to open a control session for an
-- * already created object; an object may have been declared in
-- * the DPL or by calling the dpbp_create function.
-- * This function returns a unique authentication token,
-- * associated with the specific object ID and the specific MC
-- * portal; this token must be used in all subsequent commands for
-- * this specific object
-- *
-- * Return: '0' on Success; Error code otherwise.
-- */
--int dpbp_open(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- int dpbp_id,
-- u16 *token)
--{
-- struct mc_command cmd = { 0 };
-- struct dpbp_cmd_open *cmd_params;
-- int err;
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPBP_CMDID_OPEN,
-- cmd_flags, 0);
-- cmd_params = (struct dpbp_cmd_open *)cmd.params;
-- cmd_params->dpbp_id = cpu_to_le32(dpbp_id);
--
-- /* send command to mc*/
-- err = mc_send_command(mc_io, &cmd);
-- if (err)
-- return err;
--
-- /* retrieve response parameters */
-- *token = mc_cmd_hdr_read_token(&cmd);
--
-- return err;
--}
--EXPORT_SYMBOL(dpbp_open);
--
--/**
-- * dpbp_close() - Close the control session of the object
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @token: Token of DPBP object
-- *
-- * After this function is called, no further operations are
-- * allowed on the object without opening a new control session.
-- *
-- * Return: '0' on Success; Error code otherwise.
-- */
--int dpbp_close(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token)
--{
-- struct mc_command cmd = { 0 };
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPBP_CMDID_CLOSE, cmd_flags,
-- token);
--
-- /* send command to mc*/
-- return mc_send_command(mc_io, &cmd);
--}
--EXPORT_SYMBOL(dpbp_close);
--
--/**
-- * dpbp_create() - Create the DPBP object.
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @cfg: Configuration structure
-- * @token: Returned token; use in subsequent API calls
-- *
-- * Create the DPBP object, allocate required resources and
-- * perform required initialization.
-- *
-- * The object can be created either by declaring it in the
-- * DPL file, or by calling this function.
-- * This function returns a unique authentication token,
-- * associated with the specific object ID and the specific MC
-- * portal; this token must be used in all subsequent calls to
-- * this specific object. For objects that are created using the
-- * DPL file, call dpbp_open function to get an authentication
-- * token first.
-- *
-- * Return: '0' on Success; Error code otherwise.
-- */
--int dpbp_create(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- const struct dpbp_cfg *cfg,
-- u16 *token)
--{
-- struct mc_command cmd = { 0 };
-- int err;
--
-- (void)(cfg); /* unused */
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPBP_CMDID_CREATE,
-- cmd_flags, 0);
--
-- /* send command to mc*/
-- err = mc_send_command(mc_io, &cmd);
-- if (err)
-- return err;
--
-- /* retrieve response parameters */
-- *token = mc_cmd_hdr_read_token(&cmd);
--
-- return 0;
--}
--
--/**
-- * dpbp_destroy() - Destroy the DPBP object and release all its resources.
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @token: Token of DPBP object
-- *
-- * Return: '0' on Success; error code otherwise.
-- */
--int dpbp_destroy(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token)
--{
-- struct mc_command cmd = { 0 };
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPBP_CMDID_DESTROY,
-- cmd_flags, token);
--
-- /* send command to mc*/
-- return mc_send_command(mc_io, &cmd);
--}
--
--/**
-- * dpbp_enable() - Enable the DPBP.
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @token: Token of DPBP object
-- *
-- * Return: '0' on Success; Error code otherwise.
-- */
--int dpbp_enable(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token)
--{
-- struct mc_command cmd = { 0 };
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPBP_CMDID_ENABLE, cmd_flags,
-- token);
--
-- /* send command to mc*/
-- return mc_send_command(mc_io, &cmd);
--}
--EXPORT_SYMBOL(dpbp_enable);
--
--/**
-- * dpbp_disable() - Disable the DPBP.
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @token: Token of DPBP object
-- *
-- * Return: '0' on Success; Error code otherwise.
-- */
--int dpbp_disable(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token)
--{
-- struct mc_command cmd = { 0 };
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPBP_CMDID_DISABLE,
-- cmd_flags, token);
--
-- /* send command to mc*/
-- return mc_send_command(mc_io, &cmd);
--}
--EXPORT_SYMBOL(dpbp_disable);
--
--/**
-- * dpbp_is_enabled() - Check if the DPBP is enabled.
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @token: Token of DPBP object
-- * @en: Returns '1' if object is enabled; '0' otherwise
-- *
-- * Return: '0' on Success; Error code otherwise.
-- */
--int dpbp_is_enabled(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- int *en)
--{
-- struct mc_command cmd = { 0 };
-- struct dpbp_rsp_is_enabled *rsp_params;
-- int err;
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPBP_CMDID_IS_ENABLED, cmd_flags,
-- token);
--
-- /* send command to mc*/
-- err = mc_send_command(mc_io, &cmd);
-- if (err)
-- return err;
--
-- /* retrieve response parameters */
-- rsp_params = (struct dpbp_rsp_is_enabled *)cmd.params;
-- *en = rsp_params->enabled & DPBP_ENABLE;
--
-- return 0;
--}
--
--/**
-- * dpbp_reset() - Reset the DPBP, returns the object to initial state.
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @token: Token of DPBP object
-- *
-- * Return: '0' on Success; Error code otherwise.
-- */
--int dpbp_reset(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token)
--{
-- struct mc_command cmd = { 0 };
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPBP_CMDID_RESET,
-- cmd_flags, token);
--
-- /* send command to mc*/
-- return mc_send_command(mc_io, &cmd);
--}
--
--/**
-- * dpbp_set_irq() - Set IRQ information for the DPBP to trigger an interrupt.
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @token: Token of DPBP object
-- * @irq_index: Identifies the interrupt index to configure
-- * @irq_cfg: IRQ configuration
-- *
-- * Return: '0' on Success; Error code otherwise.
-- */
--int dpbp_set_irq(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- u8 irq_index,
-- struct dpbp_irq_cfg *irq_cfg)
--{
-- struct mc_command cmd = { 0 };
-- struct dpbp_cmd_set_irq *cmd_params;
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPBP_CMDID_SET_IRQ,
-- cmd_flags, token);
-- cmd_params = (struct dpbp_cmd_set_irq *)cmd.params;
-- cmd_params->irq_index = irq_index;
-- cmd_params->irq_val = cpu_to_le32(irq_cfg->val);
-- cmd_params->irq_addr = cpu_to_le64(irq_cfg->addr);
-- cmd_params->irq_num = cpu_to_le32(irq_cfg->irq_num);
--
-- /* send command to mc*/
-- return mc_send_command(mc_io, &cmd);
--}
--
--/**
-- * dpbp_get_irq() - Get IRQ information from the DPBP.
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @token: Token of DPBP object
-- * @irq_index: The interrupt index to configure
-- * @type: Interrupt type: 0 represents message interrupt
-- * type (both irq_addr and irq_val are valid)
-- * @irq_cfg: IRQ attributes
-- *
-- * Return: '0' on Success; Error code otherwise.
-- */
--int dpbp_get_irq(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- u8 irq_index,
-- int *type,
-- struct dpbp_irq_cfg *irq_cfg)
--{
-- struct mc_command cmd = { 0 };
-- struct dpbp_cmd_get_irq *cmd_params;
-- struct dpbp_rsp_get_irq *rsp_params;
-- int err;
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPBP_CMDID_GET_IRQ,
-- cmd_flags, token);
-- cmd_params = (struct dpbp_cmd_get_irq *)cmd.params;
-- cmd_params->irq_index = irq_index;
--
-- /* send command to mc*/
-- err = mc_send_command(mc_io, &cmd);
-- if (err)
-- return err;
--
-- /* retrieve response parameters */
-- rsp_params = (struct dpbp_rsp_get_irq *)cmd.params;
-- irq_cfg->val = le32_to_cpu(rsp_params->irq_val);
-- irq_cfg->addr = le64_to_cpu(rsp_params->irq_addr);
-- irq_cfg->irq_num = le32_to_cpu(rsp_params->irq_num);
-- *type = le32_to_cpu(rsp_params->type);
--
-- return 0;
--}
--
--/**
-- * dpbp_set_irq_enable() - Set overall interrupt state.
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @token: Token of DPBP object
-- * @irq_index: The interrupt index to configure
-- * @en: Interrupt state - enable = 1, disable = 0
-- *
-- * Allows GPP software to control when interrupts are generated.
-- * Each interrupt can have up to 32 causes. The enable/disable control's the
-- * overall interrupt state. if the interrupt is disabled no causes will cause
-- * an interrupt.
-- *
-- * Return: '0' on Success; Error code otherwise.
-- */
--int dpbp_set_irq_enable(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- u8 irq_index,
-- u8 en)
--{
-- struct mc_command cmd = { 0 };
-- struct dpbp_cmd_set_irq_enable *cmd_params;
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPBP_CMDID_SET_IRQ_ENABLE,
-- cmd_flags, token);
-- cmd_params = (struct dpbp_cmd_set_irq_enable *)cmd.params;
-- cmd_params->enable = en & DPBP_ENABLE;
-- cmd_params->irq_index = irq_index;
--
-- /* send command to mc*/
-- return mc_send_command(mc_io, &cmd);
--}
--
--/**
-- * dpbp_get_irq_enable() - Get overall interrupt state
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @token: Token of DPBP object
-- * @irq_index: The interrupt index to configure
-- * @en: Returned interrupt state - enable = 1, disable = 0
-- *
-- * Return: '0' on Success; Error code otherwise.
-- */
--int dpbp_get_irq_enable(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- u8 irq_index,
-- u8 *en)
--{
-- struct mc_command cmd = { 0 };
-- struct dpbp_cmd_get_irq_enable *cmd_params;
-- struct dpbp_rsp_get_irq_enable *rsp_params;
-- int err;
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPBP_CMDID_GET_IRQ_ENABLE,
-- cmd_flags, token);
-- cmd_params = (struct dpbp_cmd_get_irq_enable *)cmd.params;
-- cmd_params->irq_index = irq_index;
--
-- /* send command to mc*/
-- err = mc_send_command(mc_io, &cmd);
-- if (err)
-- return err;
--
-- /* retrieve response parameters */
-- rsp_params = (struct dpbp_rsp_get_irq_enable *)cmd.params;
-- *en = rsp_params->enabled & DPBP_ENABLE;
-- return 0;
--}
--
--/**
-- * dpbp_set_irq_mask() - Set interrupt mask.
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @token: Token of DPBP object
-- * @irq_index: The interrupt index to configure
-- * @mask: Event mask to trigger interrupt;
-- * each bit:
-- * 0 = ignore event
-- * 1 = consider event for asserting IRQ
-- *
-- * Every interrupt can have up to 32 causes and the interrupt model supports
-- * masking/unmasking each cause independently
-- *
-- * Return: '0' on Success; Error code otherwise.
-- */
--int dpbp_set_irq_mask(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- u8 irq_index,
-- u32 mask)
--{
-- struct mc_command cmd = { 0 };
-- struct dpbp_cmd_set_irq_mask *cmd_params;
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPBP_CMDID_SET_IRQ_MASK,
-- cmd_flags, token);
-- cmd_params = (struct dpbp_cmd_set_irq_mask *)cmd.params;
-- cmd_params->mask = cpu_to_le32(mask);
-- cmd_params->irq_index = irq_index;
--
-- /* send command to mc*/
-- return mc_send_command(mc_io, &cmd);
--}
--
--/**
-- * dpbp_get_irq_mask() - Get interrupt mask.
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @token: Token of DPBP object
-- * @irq_index: The interrupt index to configure
-- * @mask: Returned event mask to trigger interrupt
-- *
-- * Every interrupt can have up to 32 causes and the interrupt model supports
-- * masking/unmasking each cause independently
-- *
-- * Return: '0' on Success; Error code otherwise.
-- */
--int dpbp_get_irq_mask(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- u8 irq_index,
-- u32 *mask)
--{
-- struct mc_command cmd = { 0 };
-- struct dpbp_cmd_get_irq_mask *cmd_params;
-- struct dpbp_rsp_get_irq_mask *rsp_params;
-- int err;
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPBP_CMDID_GET_IRQ_MASK,
-- cmd_flags, token);
-- cmd_params = (struct dpbp_cmd_get_irq_mask *)cmd.params;
-- cmd_params->irq_index = irq_index;
--
-- /* send command to mc*/
-- err = mc_send_command(mc_io, &cmd);
-- if (err)
-- return err;
--
-- /* retrieve response parameters */
-- rsp_params = (struct dpbp_rsp_get_irq_mask *)cmd.params;
-- *mask = le32_to_cpu(rsp_params->mask);
--
-- return 0;
--}
--
--/**
-- * dpbp_get_irq_status() - Get the current status of any pending interrupts.
-- *
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @token: Token of DPBP object
-- * @irq_index: The interrupt index to configure
-- * @status: Returned interrupts status - one bit per cause:
-- * 0 = no interrupt pending
-- * 1 = interrupt pending
-- *
-- * Return: '0' on Success; Error code otherwise.
-- */
--int dpbp_get_irq_status(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- u8 irq_index,
-- u32 *status)
--{
-- struct mc_command cmd = { 0 };
-- struct dpbp_cmd_get_irq_status *cmd_params;
-- struct dpbp_rsp_get_irq_status *rsp_params;
-- int err;
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPBP_CMDID_GET_IRQ_STATUS,
-- cmd_flags, token);
-- cmd_params = (struct dpbp_cmd_get_irq_status *)cmd.params;
-- cmd_params->status = cpu_to_le32(*status);
-- cmd_params->irq_index = irq_index;
--
-- /* send command to mc*/
-- err = mc_send_command(mc_io, &cmd);
-- if (err)
-- return err;
--
-- /* retrieve response parameters */
-- rsp_params = (struct dpbp_rsp_get_irq_status *)cmd.params;
-- *status = le32_to_cpu(rsp_params->status);
--
-- return 0;
--}
--
--/**
-- * dpbp_clear_irq_status() - Clear a pending interrupt's status
-- *
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @token: Token of DPBP object
-- * @irq_index: The interrupt index to configure
-- * @status: Bits to clear (W1C) - one bit per cause:
-- * 0 = don't change
-- * 1 = clear status bit
-- *
-- * Return: '0' on Success; Error code otherwise.
-- */
--int dpbp_clear_irq_status(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- u8 irq_index,
-- u32 status)
--{
-- struct mc_command cmd = { 0 };
-- struct dpbp_cmd_clear_irq_status *cmd_params;
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPBP_CMDID_CLEAR_IRQ_STATUS,
-- cmd_flags, token);
-- cmd_params = (struct dpbp_cmd_clear_irq_status *)cmd.params;
-- cmd_params->status = cpu_to_le32(status);
-- cmd_params->irq_index = irq_index;
--
-- /* send command to mc*/
-- return mc_send_command(mc_io, &cmd);
--}
--
--/**
-- * dpbp_get_attributes - Retrieve DPBP attributes.
-- *
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @token: Token of DPBP object
-- * @attr: Returned object's attributes
-- *
-- * Return: '0' on Success; Error code otherwise.
-- */
--int dpbp_get_attributes(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- struct dpbp_attr *attr)
--{
-- struct mc_command cmd = { 0 };
-- struct dpbp_rsp_get_attributes *rsp_params;
-- int err;
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPBP_CMDID_GET_ATTR,
-- cmd_flags, token);
--
-- /* send command to mc*/
-- err = mc_send_command(mc_io, &cmd);
-- if (err)
-- return err;
--
-- /* retrieve response parameters */
-- rsp_params = (struct dpbp_rsp_get_attributes *)cmd.params;
-- attr->bpid = le16_to_cpu(rsp_params->bpid);
-- attr->id = le32_to_cpu(rsp_params->id);
-- attr->version.major = le16_to_cpu(rsp_params->version_major);
-- attr->version.minor = le16_to_cpu(rsp_params->version_minor);
--
-- return 0;
--}
--EXPORT_SYMBOL(dpbp_get_attributes);
--
--/**
-- * dpbp_set_notifications() - Set notifications towards software
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @token: Token of DPBP object
-- * @cfg: notifications configuration
-- *
-- * Return: '0' on Success; Error code otherwise.
-- */
--int dpbp_set_notifications(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- struct dpbp_notification_cfg *cfg)
--{
-- struct mc_command cmd = { 0 };
-- struct dpbp_cmd_set_notifications *cmd_params;
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPBP_CMDID_SET_NOTIFICATIONS,
-- cmd_flags, token);
-- cmd_params = (struct dpbp_cmd_set_notifications *)cmd.params;
-- cmd_params->depletion_entry = cpu_to_le32(cfg->depletion_entry);
-- cmd_params->depletion_exit = cpu_to_le32(cfg->depletion_exit);
-- cmd_params->surplus_entry = cpu_to_le32(cfg->surplus_entry);
-- cmd_params->surplus_exit = cpu_to_le32(cfg->surplus_exit);
-- cmd_params->options = cpu_to_le16(cfg->options);
-- cmd_params->message_ctx = cpu_to_le64(cfg->message_ctx);
-- cmd_params->message_iova = cpu_to_le64(cfg->message_iova);
--
-- /* send command to mc*/
-- return mc_send_command(mc_io, &cmd);
--}
--
--/**
-- * dpbp_get_notifications() - Get the notifications configuration
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @token: Token of DPBP object
-- * @cfg: notifications configuration
-- *
-- * Return: '0' on Success; Error code otherwise.
-- */
--int dpbp_get_notifications(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- struct dpbp_notification_cfg *cfg)
--{
-- struct mc_command cmd = { 0 };
-- struct dpbp_rsp_get_notifications *rsp_params;
-- int err;
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPBP_CMDID_GET_NOTIFICATIONS,
-- cmd_flags,
-- token);
--
-- /* send command to mc*/
-- err = mc_send_command(mc_io, &cmd);
-- if (err)
-- return err;
--
-- /* retrieve response parameters */
-- rsp_params = (struct dpbp_rsp_get_notifications *)cmd.params;
-- cfg->depletion_entry = le32_to_cpu(rsp_params->depletion_entry);
-- cfg->depletion_exit = le32_to_cpu(rsp_params->depletion_exit);
-- cfg->surplus_entry = le32_to_cpu(rsp_params->surplus_entry);
-- cfg->surplus_exit = le32_to_cpu(rsp_params->surplus_exit);
-- cfg->options = le16_to_cpu(rsp_params->options);
-- cfg->message_ctx = le64_to_cpu(rsp_params->message_ctx);
-- cfg->message_iova = le64_to_cpu(rsp_params->message_iova);
--
-- return 0;
--}
---- /dev/null
-+++ b/drivers/staging/fsl-mc/bus/dpio/Makefile
-@@ -0,0 +1,8 @@
-+# SPDX-License-Identifier: GPL-2.0
-+#
-+# QorIQ DPAA2 DPIO driver
-+#
-+
-+obj-$(CONFIG_FSL_MC_DPIO) += fsl-mc-dpio.o
-+
-+fsl-mc-dpio-objs := dpio.o qbman-portal.o dpio-service.o dpio-driver.o
---- /dev/null
-+++ b/drivers/staging/fsl-mc/bus/dpio/dpio-cmd.h
-@@ -0,0 +1,50 @@
-+/* SPDX-License-Identifier: (GPL-2.0+ OR BSD-3-Clause) */
-+/*
-+ * Copyright 2013-2016 Freescale Semiconductor Inc.
-+ * Copyright 2016 NXP
-+ *
-+ */
-+#ifndef _FSL_DPIO_CMD_H
-+#define _FSL_DPIO_CMD_H
-+
-+/* DPIO Version */
-+#define DPIO_VER_MAJOR 4
-+#define DPIO_VER_MINOR 2
-+
-+/* Command Versioning */
-+
-+#define DPIO_CMD_ID_OFFSET 4
-+#define DPIO_CMD_BASE_VERSION 1
-+
-+#define DPIO_CMD(id) (((id) << DPIO_CMD_ID_OFFSET) | DPIO_CMD_BASE_VERSION)
-+
-+/* Command IDs */
-+#define DPIO_CMDID_CLOSE DPIO_CMD(0x800)
-+#define DPIO_CMDID_OPEN DPIO_CMD(0x803)
-+#define DPIO_CMDID_GET_API_VERSION DPIO_CMD(0xa03)
-+#define DPIO_CMDID_ENABLE DPIO_CMD(0x002)
-+#define DPIO_CMDID_DISABLE DPIO_CMD(0x003)
-+#define DPIO_CMDID_GET_ATTR DPIO_CMD(0x004)
-+#define DPIO_CMDID_RESET DPIO_CMD(0x005)
-+
-+struct dpio_cmd_open {
-+ __le32 dpio_id;
-+};
-+
-+#define DPIO_CHANNEL_MODE_MASK 0x3
-+
-+struct dpio_rsp_get_attr {
-+ /* cmd word 0 */
-+ __le32 id;
-+ __le16 qbman_portal_id;
-+ u8 num_priorities;
-+ u8 channel_mode;
-+ /* cmd word 1 */
-+ __le64 qbman_portal_ce_addr;
-+ /* cmd word 2 */
-+ __le64 qbman_portal_ci_addr;
-+ /* cmd word 3 */
-+ __le32 qbman_version;
-+};
-+
-+#endif /* _FSL_DPIO_CMD_H */
---- /dev/null
-+++ b/drivers/staging/fsl-mc/bus/dpio/dpio-driver.c
-@@ -0,0 +1,278 @@
-+// SPDX-License-Identifier: (GPL-2.0+ OR BSD-3-Clause)
-+/*
-+ * Copyright 2014-2016 Freescale Semiconductor Inc.
-+ * Copyright 2016 NXP
-+ *
-+ */
-+
-+#include <linux/types.h>
-+#include <linux/init.h>
-+#include <linux/module.h>
-+#include <linux/platform_device.h>
-+#include <linux/interrupt.h>
-+#include <linux/msi.h>
-+#include <linux/dma-mapping.h>
-+#include <linux/delay.h>
-+
-+#include <linux/fsl/mc.h>
-+#include "../../include/dpaa2-io.h"
-+
-+#include "qbman-portal.h"
-+#include "dpio.h"
-+#include "dpio-cmd.h"
-+
-+MODULE_LICENSE("Dual BSD/GPL");
-+MODULE_AUTHOR("Freescale Semiconductor, Inc");
-+MODULE_DESCRIPTION("DPIO Driver");
-+
-+struct dpio_priv {
-+ struct dpaa2_io *io;
-+};
-+
-+static irqreturn_t dpio_irq_handler(int irq_num, void *arg)
-+{
-+ struct device *dev = (struct device *)arg;
-+ struct dpio_priv *priv = dev_get_drvdata(dev);
-+
-+ return dpaa2_io_irq(priv->io);
-+}
-+
-+static void unregister_dpio_irq_handlers(struct fsl_mc_device *dpio_dev)
-+{
-+ struct fsl_mc_device_irq *irq;
-+
-+ irq = dpio_dev->irqs[0];
-+
-+ /* clear the affinity hint */
-+ irq_set_affinity_hint(irq->msi_desc->irq, NULL);
-+}
-+
-+static int register_dpio_irq_handlers(struct fsl_mc_device *dpio_dev, int cpu)
-+{
-+ struct dpio_priv *priv;
-+ int error;
-+ struct fsl_mc_device_irq *irq;
-+ cpumask_t mask;
-+
-+ priv = dev_get_drvdata(&dpio_dev->dev);
-+
-+ irq = dpio_dev->irqs[0];
-+ error = devm_request_irq(&dpio_dev->dev,
-+ irq->msi_desc->irq,
-+ dpio_irq_handler,
-+ 0,
-+ dev_name(&dpio_dev->dev),
-+ &dpio_dev->dev);
-+ if (error < 0) {
-+ dev_err(&dpio_dev->dev,
-+ "devm_request_irq() failed: %d\n",
-+ error);
-+ return error;
-+ }
-+
-+ /* set the affinity hint */
-+ cpumask_clear(&mask);
-+ cpumask_set_cpu(cpu, &mask);
-+ if (irq_set_affinity_hint(irq->msi_desc->irq, &mask))
-+ dev_err(&dpio_dev->dev,
-+ "irq_set_affinity failed irq %d cpu %d\n",
-+ irq->msi_desc->irq, cpu);
-+
-+ return 0;
-+}
-+
-+static int dpaa2_dpio_probe(struct fsl_mc_device *dpio_dev)
-+{
-+ struct dpio_attr dpio_attrs;
-+ struct dpaa2_io_desc desc;
-+ struct dpio_priv *priv;
-+ int err = -ENOMEM;
-+ struct device *dev = &dpio_dev->dev;
-+ static int next_cpu = -1;
-+
-+ priv = devm_kzalloc(dev, sizeof(*priv), GFP_KERNEL);
-+ if (!priv)
-+ goto err_priv_alloc;
-+
-+ dev_set_drvdata(dev, priv);
-+
-+ err = fsl_mc_portal_allocate(dpio_dev, 0, &dpio_dev->mc_io);
-+ if (err) {
-+ dev_dbg(dev, "MC portal allocation failed\n");
-+ err = -EPROBE_DEFER;
-+ goto err_mcportal;
-+ }
-+
-+ err = dpio_open(dpio_dev->mc_io, 0, dpio_dev->obj_desc.id,
-+ &dpio_dev->mc_handle);
-+ if (err) {
-+ dev_err(dev, "dpio_open() failed\n");
-+ goto err_open;
-+ }
-+
-+ err = dpio_reset(dpio_dev->mc_io, 0, dpio_dev->mc_handle);
-+ if (err) {
-+ dev_err(dev, "dpio_reset() failed\n");
-+ goto err_reset;
-+ }
-+
-+ err = dpio_get_attributes(dpio_dev->mc_io, 0, dpio_dev->mc_handle,
-+ &dpio_attrs);
-+ if (err) {
-+ dev_err(dev, "dpio_get_attributes() failed %d\n", err);
-+ goto err_get_attr;
-+ }
-+ desc.qman_version = dpio_attrs.qbman_version;
-+
-+ err = dpio_enable(dpio_dev->mc_io, 0, dpio_dev->mc_handle);
-+ if (err) {
-+ dev_err(dev, "dpio_enable() failed %d\n", err);
-+ goto err_get_attr;
-+ }
-+
-+ /* initialize DPIO descriptor */
-+ desc.receives_notifications = dpio_attrs.num_priorities ? 1 : 0;
-+ desc.has_8prio = dpio_attrs.num_priorities == 8 ? 1 : 0;
-+ desc.dpio_id = dpio_dev->obj_desc.id;
-+
-+ /* get the cpu to use for the affinity hint */
-+ if (next_cpu == -1)
-+ next_cpu = cpumask_first(cpu_online_mask);
-+ else
-+ next_cpu = cpumask_next(next_cpu, cpu_online_mask);
-+
-+ if (!cpu_possible(next_cpu)) {
-+ dev_err(dev, "probe failed. Number of DPIOs exceeds NR_CPUS.\n");
-+ err = -ERANGE;
-+ goto err_allocate_irqs;
-+ }
-+ desc.cpu = next_cpu;
-+
-+ /*
-+ * Set the CENA regs to be the cache enabled area of the portal to
-+ * achieve the best performance.
-+ */
-+ desc.regs_cena = ioremap_cache_ns(dpio_dev->regions[0].start,
-+ resource_size(&dpio_dev->regions[0]));
-+ desc.regs_cinh = ioremap(dpio_dev->regions[1].start,
-+ resource_size(&dpio_dev->regions[1]));
-+
-+ err = fsl_mc_allocate_irqs(dpio_dev);
-+ if (err) {
-+ dev_err(dev, "fsl_mc_allocate_irqs failed. err=%d\n", err);
-+ goto err_allocate_irqs;
-+ }
-+
-+ err = register_dpio_irq_handlers(dpio_dev, desc.cpu);
-+ if (err)
-+ goto err_register_dpio_irq;
-+
-+ priv->io = dpaa2_io_create(&desc);
-+ if (!priv->io) {
-+ dev_err(dev, "dpaa2_io_create failed\n");
-+ goto err_dpaa2_io_create;
-+ }
-+
-+ dev_info(dev, "probed\n");
-+ dev_dbg(dev, " receives_notifications = %d\n",
-+ desc.receives_notifications);
-+ dpio_close(dpio_dev->mc_io, 0, dpio_dev->mc_handle);
-+ fsl_mc_portal_free(dpio_dev->mc_io);
-+
-+ return 0;
-+
-+err_dpaa2_io_create:
-+ unregister_dpio_irq_handlers(dpio_dev);
-+err_register_dpio_irq:
-+ fsl_mc_free_irqs(dpio_dev);
-+err_allocate_irqs:
-+ dpio_disable(dpio_dev->mc_io, 0, dpio_dev->mc_handle);
-+err_get_attr:
-+err_reset:
-+ dpio_close(dpio_dev->mc_io, 0, dpio_dev->mc_handle);
-+err_open:
-+ fsl_mc_portal_free(dpio_dev->mc_io);
-+err_mcportal:
-+ dev_set_drvdata(dev, NULL);
-+err_priv_alloc:
-+ return err;
-+}
-+
-+/* Tear down interrupts for a given DPIO object */
-+static void dpio_teardown_irqs(struct fsl_mc_device *dpio_dev)
-+{
-+ unregister_dpio_irq_handlers(dpio_dev);
-+ fsl_mc_free_irqs(dpio_dev);
-+}
-+
-+static int dpaa2_dpio_remove(struct fsl_mc_device *dpio_dev)
-+{
-+ struct device *dev;
-+ struct dpio_priv *priv;
-+ int err;
-+
-+ dev = &dpio_dev->dev;
-+ priv = dev_get_drvdata(dev);
-+
-+ dpaa2_io_down(priv->io);
-+
-+ dpio_teardown_irqs(dpio_dev);
-+
-+ err = fsl_mc_portal_allocate(dpio_dev, 0, &dpio_dev->mc_io);
-+ if (err) {
-+ dev_err(dev, "MC portal allocation failed\n");
-+ goto err_mcportal;
-+ }
-+
-+ err = dpio_open(dpio_dev->mc_io, 0, dpio_dev->obj_desc.id,
-+ &dpio_dev->mc_handle);
-+ if (err) {
-+ dev_err(dev, "dpio_open() failed\n");
-+ goto err_open;
-+ }
-+
-+ dpio_disable(dpio_dev->mc_io, 0, dpio_dev->mc_handle);
-+
-+ dpio_close(dpio_dev->mc_io, 0, dpio_dev->mc_handle);
-+
-+ fsl_mc_portal_free(dpio_dev->mc_io);
-+
-+ dev_set_drvdata(dev, NULL);
-+
-+ return 0;
-+
-+err_open:
-+ fsl_mc_portal_free(dpio_dev->mc_io);
-+err_mcportal:
-+ return err;
-+}
-+
-+static const struct fsl_mc_device_id dpaa2_dpio_match_id_table[] = {
-+ {
-+ .vendor = FSL_MC_VENDOR_FREESCALE,
-+ .obj_type = "dpio",
-+ },
-+ { .vendor = 0x0 }
-+};
-+
-+static struct fsl_mc_driver dpaa2_dpio_driver = {
-+ .driver = {
-+ .name = KBUILD_MODNAME,
-+ .owner = THIS_MODULE,
-+ },
-+ .probe = dpaa2_dpio_probe,
-+ .remove = dpaa2_dpio_remove,
-+ .match_id_table = dpaa2_dpio_match_id_table
-+};
-+
-+static int dpio_driver_init(void)
-+{
-+ return fsl_mc_driver_register(&dpaa2_dpio_driver);
-+}
-+
-+static void dpio_driver_exit(void)
-+{
-+ fsl_mc_driver_unregister(&dpaa2_dpio_driver);
-+}
-+module_init(dpio_driver_init);
-+module_exit(dpio_driver_exit);
---- /dev/null
-+++ b/drivers/staging/fsl-mc/bus/dpio/dpio-service.c
-@@ -0,0 +1,780 @@
-+// SPDX-License-Identifier: (GPL-2.0+ OR BSD-3-Clause)
-+/*
-+ * Copyright 2014-2016 Freescale Semiconductor Inc.
-+ * Copyright 2016 NXP
-+ *
-+ */
-+#include <linux/types.h>
-+#include <linux/fsl/mc.h>
-+#include "../../include/dpaa2-io.h"
-+#include <linux/init.h>
-+#include <linux/module.h>
-+#include <linux/platform_device.h>
-+#include <linux/interrupt.h>
-+#include <linux/dma-mapping.h>
-+#include <linux/slab.h>
-+
-+#include "dpio.h"
-+#include "qbman-portal.h"
-+
-+struct dpaa2_io {
-+ struct dpaa2_io_desc dpio_desc;
-+ struct qbman_swp_desc swp_desc;
-+ struct qbman_swp *swp;
-+ struct list_head node;
-+ /* protect against multiple management commands */
-+ spinlock_t lock_mgmt_cmd;
-+ /* protect notifications list */
-+ spinlock_t lock_notifications;
-+ struct list_head notifications;
-+};
-+
-+struct dpaa2_io_store {
-+ unsigned int max;
-+ dma_addr_t paddr;
-+ struct dpaa2_dq *vaddr;
-+ void *alloced_addr; /* unaligned value from kmalloc() */
-+ unsigned int idx; /* position of the next-to-be-returned entry */
-+ struct qbman_swp *swp; /* portal used to issue VDQCR */
-+ struct device *dev; /* device used for DMA mapping */
-+};
-+
-+/* keep a per cpu array of DPIOs for fast access */
-+static struct dpaa2_io *dpio_by_cpu[NR_CPUS];
-+static struct list_head dpio_list = LIST_HEAD_INIT(dpio_list);
-+static DEFINE_SPINLOCK(dpio_list_lock);
-+
-+static inline struct dpaa2_io *service_select_by_cpu(struct dpaa2_io *d,
-+ int cpu)
-+{
-+ if (d)
-+ return d;
-+
-+ if (cpu != DPAA2_IO_ANY_CPU && cpu >= num_possible_cpus())
-+ return NULL;
-+
-+ /*
-+ * If cpu == -1, choose the current cpu, with no guarantees about
-+ * potentially being migrated away.
-+ */
-+ if (cpu < 0)
-+ cpu = smp_processor_id();
-+
-+ /* If a specific cpu was requested, pick it up immediately */
-+ return dpio_by_cpu[cpu];
-+}
-+
-+static inline struct dpaa2_io *service_select(struct dpaa2_io *d)
-+{
-+ if (d)
-+ return d;
-+
-+ d = service_select_by_cpu(d, -1);
-+ if (d)
-+ return d;
-+
-+ spin_lock(&dpio_list_lock);
-+ d = list_entry(dpio_list.next, struct dpaa2_io, node);
-+ list_del(&d->node);
-+ list_add_tail(&d->node, &dpio_list);
-+ spin_unlock(&dpio_list_lock);
-+
-+ return d;
-+}
-+
-+/**
-+ * dpaa2_io_service_select() - return a dpaa2_io service affined to this cpu
-+ * @cpu: the cpu id
-+ *
-+ * Return the affine dpaa2_io service, or NULL if there is no service affined
-+ * to the specified cpu. If DPAA2_IO_ANY_CPU is used, return the next available
-+ * service.
-+ */
-+struct dpaa2_io *dpaa2_io_service_select(int cpu)
-+{
-+ if (cpu == DPAA2_IO_ANY_CPU)
-+ return service_select(NULL);
-+
-+ return service_select_by_cpu(NULL, cpu);
-+}
-+EXPORT_SYMBOL_GPL(dpaa2_io_service_select);
-+
-+/**
-+ * dpaa2_io_create() - create a dpaa2_io object.
-+ * @desc: the dpaa2_io descriptor
-+ *
-+ * Activates a "struct dpaa2_io" corresponding to the given config of an actual
-+ * DPIO object.
-+ *
-+ * Return a valid dpaa2_io object for success, or NULL for failure.
-+ */
-+struct dpaa2_io *dpaa2_io_create(const struct dpaa2_io_desc *desc)
-+{
-+ struct dpaa2_io *obj = kmalloc(sizeof(*obj), GFP_KERNEL);
-+
-+ if (!obj)
-+ return NULL;
-+
-+ /* check if CPU is out of range (-1 means any cpu) */
-+ if (desc->cpu != DPAA2_IO_ANY_CPU && desc->cpu >= num_possible_cpus()) {
-+ kfree(obj);
-+ return NULL;
-+ }
-+
-+ obj->dpio_desc = *desc;
-+ obj->swp_desc.cena_bar = obj->dpio_desc.regs_cena;
-+ obj->swp_desc.cinh_bar = obj->dpio_desc.regs_cinh;
-+ obj->swp_desc.qman_version = obj->dpio_desc.qman_version;
-+ obj->swp = qbman_swp_init(&obj->swp_desc);
-+
-+ if (!obj->swp) {
-+ kfree(obj);
-+ return NULL;
-+ }
-+
-+ INIT_LIST_HEAD(&obj->node);
-+ spin_lock_init(&obj->lock_mgmt_cmd);
-+ spin_lock_init(&obj->lock_notifications);
-+ INIT_LIST_HEAD(&obj->notifications);
-+
-+ /* For now only enable DQRR interrupts */
-+ qbman_swp_interrupt_set_trigger(obj->swp,
-+ QBMAN_SWP_INTERRUPT_DQRI);
-+ qbman_swp_interrupt_clear_status(obj->swp, 0xffffffff);
-+ if (obj->dpio_desc.receives_notifications)
-+ qbman_swp_push_set(obj->swp, 0, 1);
-+
-+ spin_lock(&dpio_list_lock);
-+ list_add_tail(&obj->node, &dpio_list);
-+ if (desc->cpu >= 0 && !dpio_by_cpu[desc->cpu])
-+ dpio_by_cpu[desc->cpu] = obj;
-+ spin_unlock(&dpio_list_lock);
-+
-+ return obj;
-+}
-+
-+/**
-+ * dpaa2_io_down() - release the dpaa2_io object.
-+ * @d: the dpaa2_io object to be released.
-+ *
-+ * The "struct dpaa2_io" type can represent an individual DPIO object (as
-+ * described by "struct dpaa2_io_desc") or an instance of a "DPIO service",
-+ * which can be used to group/encapsulate multiple DPIO objects. In all cases,
-+ * each handle obtained should be released using this function.
-+ */
-+void dpaa2_io_down(struct dpaa2_io *d)
-+{
-+ kfree(d);
-+}
-+
-+#define DPAA_POLL_MAX 32
-+
-+/**
-+ * dpaa2_io_irq() - ISR for DPIO interrupts
-+ *
-+ * @obj: the given DPIO object.
-+ *
-+ * Return IRQ_HANDLED for success or IRQ_NONE if there
-+ * were no pending interrupts.
-+ */
-+irqreturn_t dpaa2_io_irq(struct dpaa2_io *obj)
-+{
-+ const struct dpaa2_dq *dq;
-+ int max = 0;
-+ struct qbman_swp *swp;
-+ u32 status;
-+
-+ swp = obj->swp;
-+ status = qbman_swp_interrupt_read_status(swp);
-+ if (!status)
-+ return IRQ_NONE;
-+
-+ dq = qbman_swp_dqrr_next(swp);
-+ while (dq) {
-+ if (qbman_result_is_SCN(dq)) {
-+ struct dpaa2_io_notification_ctx *ctx;
-+ u64 q64;
-+
-+ q64 = qbman_result_SCN_ctx(dq);
-+ ctx = (void *)(uintptr_t)q64;
-+ ctx->cb(ctx);
-+ } else {
-+ pr_crit("fsl-mc-dpio: Unrecognised/ignored DQRR entry\n");
-+ }
-+ qbman_swp_dqrr_consume(swp, dq);
-+ ++max;
-+ if (max > DPAA_POLL_MAX)
-+ goto done;
-+ dq = qbman_swp_dqrr_next(swp);
-+ }
-+done:
-+ qbman_swp_interrupt_clear_status(swp, status);
-+ qbman_swp_interrupt_set_inhibit(swp, 0);
-+ return IRQ_HANDLED;
-+}
-+
-+/**
-+ * dpaa2_io_service_register() - Prepare for servicing of FQDAN or CDAN
-+ * notifications on the given DPIO service.
-+ * @d: the given DPIO service.
-+ * @ctx: the notification context.
-+ *
-+ * The caller should make the MC command to attach a DPAA2 object to
-+ * a DPIO after this function completes successfully. In that way:
-+ * (a) The DPIO service is "ready" to handle a notification arrival
-+ * (which might happen before the "attach" command to MC has
-+ * returned control of execution back to the caller)
-+ * (b) The DPIO service can provide back to the caller the 'dpio_id' and
-+ * 'qman64' parameters that it should pass along in the MC command
-+ * in order for the object to be configured to produce the right
-+ * notification fields to the DPIO service.
-+ *
-+ * Return 0 for success, or -ENODEV for failure.
-+ */
-+int dpaa2_io_service_register(struct dpaa2_io *d,
-+ struct dpaa2_io_notification_ctx *ctx)
-+{
-+ unsigned long irqflags;
-+
-+ d = service_select_by_cpu(d, ctx->desired_cpu);
-+ if (!d)
-+ return -ENODEV;
-+
-+ ctx->dpio_id = d->dpio_desc.dpio_id;
-+ ctx->qman64 = (u64)(uintptr_t)ctx;
-+ ctx->dpio_private = d;
-+ spin_lock_irqsave(&d->lock_notifications, irqflags);
-+ list_add(&ctx->node, &d->notifications);
-+ spin_unlock_irqrestore(&d->lock_notifications, irqflags);
-+
-+ /* Enable the generation of CDAN notifications */
-+ if (ctx->is_cdan)
-+ return qbman_swp_CDAN_set_context_enable(d->swp,
-+ (u16)ctx->id,
-+ ctx->qman64);
-+ return 0;
-+}
-+EXPORT_SYMBOL_GPL(dpaa2_io_service_register);
-+
-+/**
-+ * dpaa2_io_service_deregister - The opposite of 'register'.
-+ * @service: the given DPIO service.
-+ * @ctx: the notification context.
-+ *
-+ * This function should be called only after sending the MC command to
-+ * to detach the notification-producing device from the DPIO.
-+ */
-+void dpaa2_io_service_deregister(struct dpaa2_io *service,
-+ struct dpaa2_io_notification_ctx *ctx)
-+{
-+ struct dpaa2_io *d = ctx->dpio_private;
-+ unsigned long irqflags;
-+
-+ if (ctx->is_cdan)
-+ qbman_swp_CDAN_disable(d->swp, (u16)ctx->id);
-+
-+ spin_lock_irqsave(&d->lock_notifications, irqflags);
-+ list_del(&ctx->node);
-+ spin_unlock_irqrestore(&d->lock_notifications, irqflags);
-+}
-+EXPORT_SYMBOL_GPL(dpaa2_io_service_deregister);
-+
-+/**
-+ * dpaa2_io_service_rearm() - Rearm the notification for the given DPIO service.
-+ * @d: the given DPIO service.
-+ * @ctx: the notification context.
-+ *
-+ * Once a FQDAN/CDAN has been produced, the corresponding FQ/channel is
-+ * considered "disarmed". Ie. the user can issue pull dequeue operations on that
-+ * traffic source for as long as it likes. Eventually it may wish to "rearm"
-+ * that source to allow it to produce another FQDAN/CDAN, that's what this
-+ * function achieves.
-+ *
-+ * Return 0 for success.
-+ */
-+int dpaa2_io_service_rearm(struct dpaa2_io *d,
-+ struct dpaa2_io_notification_ctx *ctx)
-+{
-+ unsigned long irqflags;
-+ int err;
-+
-+ d = service_select_by_cpu(d, ctx->desired_cpu);
-+ if (!unlikely(d))
-+ return -ENODEV;
-+
-+ spin_lock_irqsave(&d->lock_mgmt_cmd, irqflags);
-+ if (ctx->is_cdan)
-+ err = qbman_swp_CDAN_enable(d->swp, (u16)ctx->id);
-+ else
-+ err = qbman_swp_fq_schedule(d->swp, ctx->id);
-+ spin_unlock_irqrestore(&d->lock_mgmt_cmd, irqflags);
-+
-+ return err;
-+}
-+EXPORT_SYMBOL_GPL(dpaa2_io_service_rearm);
-+
-+/**
-+ * dpaa2_io_service_pull_fq() - pull dequeue functions from a fq.
-+ * @d: the given DPIO service.
-+ * @fqid: the given frame queue id.
-+ * @s: the dpaa2_io_store object for the result.
-+ *
-+ * Return 0 for success, or error code for failure.
-+ */
-+int dpaa2_io_service_pull_fq(struct dpaa2_io *d, u32 fqid,
-+ struct dpaa2_io_store *s)
-+{
-+ struct qbman_pull_desc pd;
-+ int err;
-+
-+ qbman_pull_desc_clear(&pd);
-+ qbman_pull_desc_set_storage(&pd, s->vaddr, s->paddr, 1);
-+ qbman_pull_desc_set_numframes(&pd, (u8)s->max);
-+ qbman_pull_desc_set_fq(&pd, fqid);
-+
-+ d = service_select(d);
-+ if (!d)
-+ return -ENODEV;
-+ s->swp = d->swp;
-+ err = qbman_swp_pull(d->swp, &pd);
-+ if (err)
-+ s->swp = NULL;
-+
-+ return err;
-+}
-+EXPORT_SYMBOL(dpaa2_io_service_pull_fq);
-+
-+/**
-+ * dpaa2_io_service_pull_channel() - pull dequeue functions from a channel.
-+ * @d: the given DPIO service.
-+ * @channelid: the given channel id.
-+ * @s: the dpaa2_io_store object for the result.
-+ *
-+ * Return 0 for success, or error code for failure.
-+ */
-+int dpaa2_io_service_pull_channel(struct dpaa2_io *d, u32 channelid,
-+ struct dpaa2_io_store *s)
-+{
-+ struct qbman_pull_desc pd;
-+ int err;
-+
-+ qbman_pull_desc_clear(&pd);
-+ qbman_pull_desc_set_storage(&pd, s->vaddr, s->paddr, 1);
-+ qbman_pull_desc_set_numframes(&pd, (u8)s->max);
-+ qbman_pull_desc_set_channel(&pd, channelid, qbman_pull_type_prio);
-+
-+ d = service_select(d);
-+ if (!d)
-+ return -ENODEV;
-+
-+ s->swp = d->swp;
-+ err = qbman_swp_pull(d->swp, &pd);
-+ if (err)
-+ s->swp = NULL;
-+
-+ return err;
-+}
-+EXPORT_SYMBOL_GPL(dpaa2_io_service_pull_channel);
-+
-+/**
-+ * dpaa2_io_service_enqueue_fq() - Enqueue a frame to a frame queue.
-+ * @d: the given DPIO service.
-+ * @fqid: the given frame queue id.
-+ * @fd: the frame descriptor which is enqueued.
-+ *
-+ * Return 0 for successful enqueue, -EBUSY if the enqueue ring is not ready,
-+ * or -ENODEV if there is no dpio service.
-+ */
-+int dpaa2_io_service_enqueue_fq(struct dpaa2_io *d,
-+ u32 fqid,
-+ const struct dpaa2_fd *fd)
-+{
-+ struct qbman_eq_desc ed;
-+
-+ d = service_select(d);
-+ if (!d)
-+ return -ENODEV;
-+
-+ qbman_eq_desc_clear(&ed);
-+ qbman_eq_desc_set_no_orp(&ed, 0);
-+ qbman_eq_desc_set_fq(&ed, fqid);
-+
-+ return qbman_swp_enqueue(d->swp, &ed, fd);
-+}
-+EXPORT_SYMBOL(dpaa2_io_service_enqueue_fq);
-+
-+/**
-+ * dpaa2_io_service_enqueue_qd() - Enqueue a frame to a QD.
-+ * @d: the given DPIO service.
-+ * @qdid: the given queuing destination id.
-+ * @prio: the given queuing priority.
-+ * @qdbin: the given queuing destination bin.
-+ * @fd: the frame descriptor which is enqueued.
-+ *
-+ * Return 0 for successful enqueue, or -EBUSY if the enqueue ring is not ready,
-+ * or -ENODEV if there is no dpio service.
-+ */
-+int dpaa2_io_service_enqueue_qd(struct dpaa2_io *d,
-+ u32 qdid, u8 prio, u16 qdbin,
-+ const struct dpaa2_fd *fd)
-+{
-+ struct qbman_eq_desc ed;
-+
-+ d = service_select(d);
-+ if (!d)
-+ return -ENODEV;
-+
-+ qbman_eq_desc_clear(&ed);
-+ qbman_eq_desc_set_no_orp(&ed, 0);
-+ qbman_eq_desc_set_qd(&ed, qdid, qdbin, prio);
-+
-+ return qbman_swp_enqueue(d->swp, &ed, fd);
-+}
-+EXPORT_SYMBOL_GPL(dpaa2_io_service_enqueue_qd);
-+
-+/**
-+ * dpaa2_io_service_release() - Release buffers to a buffer pool.
-+ * @d: the given DPIO object.
-+ * @bpid: the buffer pool id.
-+ * @buffers: the buffers to be released.
-+ * @num_buffers: the number of the buffers to be released.
-+ *
-+ * Return 0 for success, and negative error code for failure.
-+ */
-+int dpaa2_io_service_release(struct dpaa2_io *d,
-+ u32 bpid,
-+ const u64 *buffers,
-+ unsigned int num_buffers)
-+{
-+ struct qbman_release_desc rd;
-+
-+ d = service_select(d);
-+ if (!d)
-+ return -ENODEV;
-+
-+ qbman_release_desc_clear(&rd);
-+ qbman_release_desc_set_bpid(&rd, bpid);
-+
-+ return qbman_swp_release(d->swp, &rd, buffers, num_buffers);
-+}
-+EXPORT_SYMBOL_GPL(dpaa2_io_service_release);
-+
-+/**
-+ * dpaa2_io_service_acquire() - Acquire buffers from a buffer pool.
-+ * @d: the given DPIO object.
-+ * @bpid: the buffer pool id.
-+ * @buffers: the buffer addresses for acquired buffers.
-+ * @num_buffers: the expected number of the buffers to acquire.
-+ *
-+ * Return a negative error code if the command failed, otherwise it returns
-+ * the number of buffers acquired, which may be less than the number requested.
-+ * Eg. if the buffer pool is empty, this will return zero.
-+ */
-+int dpaa2_io_service_acquire(struct dpaa2_io *d,
-+ u32 bpid,
-+ u64 *buffers,
-+ unsigned int num_buffers)
-+{
-+ unsigned long irqflags;
-+ int err;
-+
-+ d = service_select(d);
-+ if (!d)
-+ return -ENODEV;
-+
-+ spin_lock_irqsave(&d->lock_mgmt_cmd, irqflags);
-+ err = qbman_swp_acquire(d->swp, bpid, buffers, num_buffers);
-+ spin_unlock_irqrestore(&d->lock_mgmt_cmd, irqflags);
-+
-+ return err;
-+}
-+EXPORT_SYMBOL_GPL(dpaa2_io_service_acquire);
-+
-+/*
-+ * 'Stores' are reusable memory blocks for holding dequeue results, and to
-+ * assist with parsing those results.
-+ */
-+
-+/**
-+ * dpaa2_io_store_create() - Create the dma memory storage for dequeue result.
-+ * @max_frames: the maximum number of dequeued result for frames, must be <= 16.
-+ * @dev: the device to allow mapping/unmapping the DMAable region.
-+ *
-+ * The size of the storage is "max_frames*sizeof(struct dpaa2_dq)".
-+ * The 'dpaa2_io_store' returned is a DPIO service managed object.
-+ *
-+ * Return pointer to dpaa2_io_store struct for successfuly created storage
-+ * memory, or NULL on error.
-+ */
-+struct dpaa2_io_store *dpaa2_io_store_create(unsigned int max_frames,
-+ struct device *dev)
-+{
-+ struct dpaa2_io_store *ret;
-+ size_t size;
-+
-+ if (!max_frames || (max_frames > 16))
-+ return NULL;
-+
-+ ret = kmalloc(sizeof(*ret), GFP_KERNEL);
-+ if (!ret)
-+ return NULL;
-+
-+ ret->max = max_frames;
-+ size = max_frames * sizeof(struct dpaa2_dq) + 64;
-+ ret->alloced_addr = kzalloc(size, GFP_KERNEL);
-+ if (!ret->alloced_addr) {
-+ kfree(ret);
-+ return NULL;
-+ }
-+
-+ ret->vaddr = PTR_ALIGN(ret->alloced_addr, 64);
-+ ret->paddr = dma_map_single(dev, ret->vaddr,
-+ sizeof(struct dpaa2_dq) * max_frames,
-+ DMA_FROM_DEVICE);
-+ if (dma_mapping_error(dev, ret->paddr)) {
-+ kfree(ret->alloced_addr);
-+ kfree(ret);
-+ return NULL;
-+ }
-+
-+ ret->idx = 0;
-+ ret->dev = dev;
-+
-+ return ret;
-+}
-+EXPORT_SYMBOL_GPL(dpaa2_io_store_create);
-+
-+/**
-+ * dpaa2_io_store_destroy() - Frees the dma memory storage for dequeue
-+ * result.
-+ * @s: the storage memory to be destroyed.
-+ */
-+void dpaa2_io_store_destroy(struct dpaa2_io_store *s)
-+{
-+ dma_unmap_single(s->dev, s->paddr, sizeof(struct dpaa2_dq) * s->max,
-+ DMA_FROM_DEVICE);
-+ kfree(s->alloced_addr);
-+ kfree(s);
-+}
-+EXPORT_SYMBOL_GPL(dpaa2_io_store_destroy);
-+
-+/**
-+ * dpaa2_io_store_next() - Determine when the next dequeue result is available.
-+ * @s: the dpaa2_io_store object.
-+ * @is_last: indicate whether this is the last frame in the pull command.
-+ *
-+ * When an object driver performs dequeues to a dpaa2_io_store, this function
-+ * can be used to determine when the next frame result is available. Once
-+ * this function returns non-NULL, a subsequent call to it will try to find
-+ * the next dequeue result.
-+ *
-+ * Note that if a pull-dequeue has a NULL result because the target FQ/channel
-+ * was empty, then this function will also return NULL (rather than expecting
-+ * the caller to always check for this. As such, "is_last" can be used to
-+ * differentiate between "end-of-empty-dequeue" and "still-waiting".
-+ *
-+ * Return dequeue result for a valid dequeue result, or NULL for empty dequeue.
-+ */
-+struct dpaa2_dq *dpaa2_io_store_next(struct dpaa2_io_store *s, int *is_last)
-+{
-+ int match;
-+ struct dpaa2_dq *ret = &s->vaddr[s->idx];
-+
-+ match = qbman_result_has_new_result(s->swp, ret);
-+ if (!match) {
-+ *is_last = 0;
-+ return NULL;
-+ }
-+
-+ s->idx++;
-+
-+ if (dpaa2_dq_is_pull_complete(ret)) {
-+ *is_last = 1;
-+ s->idx = 0;
-+ /*
-+ * If we get an empty dequeue result to terminate a zero-results
-+ * vdqcr, return NULL to the caller rather than expecting him to
-+ * check non-NULL results every time.
-+ */
-+ if (!(dpaa2_dq_flags(ret) & DPAA2_DQ_STAT_VALIDFRAME))
-+ ret = NULL;
-+ } else {
-+ *is_last = 0;
-+ }
-+
-+ return ret;
-+}
-+EXPORT_SYMBOL_GPL(dpaa2_io_store_next);
-+
-+/**
-+ * dpaa2_io_query_fq_count() - Get the frame and byte count for a given fq.
-+ * @d: the given DPIO object.
-+ * @fqid: the id of frame queue to be queried.
-+ * @fcnt: the queried frame count.
-+ * @bcnt: the queried byte count.
-+ *
-+ * Knowing the FQ count at run-time can be useful in debugging situations.
-+ * The instantaneous frame- and byte-count are hereby returned.
-+ *
-+ * Return 0 for a successful query, and negative error code if query fails.
-+ */
-+int dpaa2_io_query_fq_count(struct dpaa2_io *d, u32 fqid,
-+ u32 *fcnt, u32 *bcnt)
-+{
-+ struct qbman_fq_query_np_rslt state;
-+ struct qbman_swp *swp;
-+ unsigned long irqflags;
-+ int ret;
-+
-+ d = service_select(d);
-+ if (!d)
-+ return -ENODEV;
-+
-+ swp = d->swp;
-+ spin_lock_irqsave(&d->lock_mgmt_cmd, irqflags);
-+ ret = qbman_fq_query_state(swp, fqid, &state);
-+ spin_unlock_irqrestore(&d->lock_mgmt_cmd, irqflags);
-+ if (ret)
-+ return ret;
-+ *fcnt = qbman_fq_state_frame_count(&state);
-+ *bcnt = qbman_fq_state_byte_count(&state);
-+
-+ return 0;
-+}
-+EXPORT_SYMBOL(dpaa2_io_query_fq_count);
-+
-+/**
-+ * dpaa2_io_query_bp_count() - Query the number of buffers currently in a
-+ * buffer pool.
-+ * @d: the given DPIO object.
-+ * @bpid: the index of buffer pool to be queried.
-+ * @num: the queried number of buffers in the buffer pool.
-+ *
-+ * Return 0 for a successful query, and negative error code if query fails.
-+ */
-+int dpaa2_io_query_bp_count(struct dpaa2_io *d, u32 bpid, u32 *num)
-+{
-+ struct qbman_bp_query_rslt state;
-+ struct qbman_swp *swp;
-+ unsigned long irqflags;
-+ int ret;
-+
-+ d = service_select(d);
-+ if (!d)
-+ return -ENODEV;
-+
-+ swp = d->swp;
-+ spin_lock_irqsave(&d->lock_mgmt_cmd, irqflags);
-+ ret = qbman_bp_query(swp, bpid, &state);
-+ spin_unlock_irqrestore(&d->lock_mgmt_cmd, irqflags);
-+ if (ret)
-+ return ret;
-+ *num = qbman_bp_info_num_free_bufs(&state);
-+ return 0;
-+}
-+EXPORT_SYMBOL(dpaa2_io_query_bp_count);
-+
-+/**
-+ * dpaa2_io_service_enqueue_orp_fq() - Enqueue a frame to a frame queue with
-+ * order restoration
-+ * @d: the given DPIO service.
-+ * @fqid: the given frame queue id.
-+ * @fd: the frame descriptor which is enqueued.
-+ * @orpid: the order restoration point ID
-+ * @seqnum: the order sequence number
-+ * @last: must be set for the final frame if seqnum is shared (spilt frame)
-+ *
-+ * Performs an enqueue to a frame queue using the specified order restoration
-+ * point. The QMan device will ensure the order of frames placed on the
-+ * queue will be ordered as per the sequence number.
-+ *
-+ * In the case a frame is split it is possible to enqueue using the same
-+ * sequence number more than once. The final frame in a shared sequence number
-+ * most be indicated by setting last = 1. For non shared sequence numbers
-+ * last = 1 must always be set.
-+ *
-+ * Return 0 for successful enqueue, or -EBUSY if the enqueue ring is not ready,
-+ * or -ENODEV if there is no dpio service.
-+ */
-+int dpaa2_io_service_enqueue_orp_fq(struct dpaa2_io *d, u32 fqid,
-+ const struct dpaa2_fd *fd, u16 orpid,
-+ u16 seqnum, int last)
-+{
-+ struct qbman_eq_desc ed;
-+
-+ d = service_select(d);
-+ if (!d)
-+ return -ENODEV;
-+ qbman_eq_desc_clear(&ed);
-+ qbman_eq_desc_set_orp(&ed, 0, orpid, seqnum, !last);
-+ qbman_eq_desc_set_fq(&ed, fqid);
-+ return qbman_swp_enqueue(d->swp, &ed, fd);
-+}
-+EXPORT_SYMBOL(dpaa2_io_service_enqueue_orp_fq);
-+
-+/**
-+ * dpaa2_io_service_enqueue_orp_qd() - Enqueue a frame to a queueing destination
-+ * with order restoration
-+ * @d: the given DPIO service.
-+ * @qdid: the given queuing destination id.
-+ * @fd: the frame descriptor which is enqueued.
-+ * @orpid: the order restoration point ID
-+ * @seqnum: the order sequence number
-+ * @last: must be set for the final frame if seqnum is shared (spilt frame)
-+ *
-+ * Performs an enqueue to a frame queue using the specified order restoration
-+ * point. The QMan device will ensure the order of frames placed on the
-+ * queue will be ordered as per the sequence number.
-+ *
-+ * In the case a frame is split it is possible to enqueue using the same
-+ * sequence number more than once. The final frame in a shared sequence number
-+ * most be indicated by setting last = 1. For non shared sequence numbers
-+ * last = 1 must always be set.
-+ *
-+ * Return 0 for successful enqueue, or -EBUSY if the enqueue ring is not ready,
-+ * or -ENODEV if there is no dpio service.
-+ */
-+int dpaa2_io_service_enqueue_orp_qd(struct dpaa2_io *d, u32 qdid, u8 prio,
-+ u16 qdbin, const struct dpaa2_fd *fd,
-+ u16 orpid, u16 seqnum, int last)
-+{
-+ struct qbman_eq_desc ed;
-+
-+ d = service_select(d);
-+ if (!d)
-+ return -ENODEV;
-+ qbman_eq_desc_clear(&ed);
-+ qbman_eq_desc_set_orp(&ed, 0, orpid, seqnum, !last);
-+ qbman_eq_desc_set_qd(&ed, qdid, qdbin, prio);
-+ return qbman_swp_enqueue(d->swp, &ed, fd);
-+}
-+EXPORT_SYMBOL_GPL(dpaa2_io_service_enqueue_orp_qd);
-+
-+/**
-+ * dpaa2_io_service_orp_seqnum_drop() - Remove a sequence number from
-+ * an order restoration list
-+ * @d: the given DPIO service.
-+ * @orpid: Order restoration point to remove a sequence number from
-+ * @seqnum: Sequence number to remove
-+ *
-+ * Removes a frames sequence number from an order restoration point without
-+ * enqueing the frame. Used to indicate that the order restoration hardware
-+ * should not expect to see this sequence number. Typically used to indicate
-+ * a frame was terminated or dropped from a flow.
-+ *
-+ * Return 0 for successful enqueue, or -EBUSY if the enqueue ring is not ready,
-+ * or -ENODEV if there is no dpio service.
-+ */
-+int dpaa2_io_service_orp_seqnum_drop(struct dpaa2_io *d, u16 orpid, u16 seqnum)
-+{
-+ struct qbman_eq_desc ed;
-+ struct dpaa2_fd fd;
-+
-+ d = service_select(d);
-+ if (!d)
-+ return -ENODEV;
-+ qbman_eq_desc_clear(&ed);
-+ qbman_eq_desc_set_orp_hole(&ed, orpid, seqnum);
-+ return qbman_swp_enqueue(d->swp, &ed, &fd);
-+}
-+EXPORT_SYMBOL_GPL(dpaa2_io_service_orp_seqnum_drop);
---- /dev/null
-+++ b/drivers/staging/fsl-mc/bus/dpio/dpio.c
-@@ -0,0 +1,221 @@
-+// SPDX-License-Identifier: (GPL-2.0+ OR BSD-3-Clause)
-+/*
-+ * Copyright 2013-2016 Freescale Semiconductor Inc.
-+ * Copyright 2016 NXP
-+ *
-+ */
-+#include <linux/kernel.h>
-+#include <linux/fsl/mc.h>
-+
-+#include "dpio.h"
-+#include "dpio-cmd.h"
-+
-+/*
-+ * Data Path I/O Portal API
-+ * Contains initialization APIs and runtime control APIs for DPIO
-+ */
-+
-+/**
-+ * dpio_open() - Open a control session for the specified object
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @dpio_id: DPIO unique ID
-+ * @token: Returned token; use in subsequent API calls
-+ *
-+ * This function can be used to open a control session for an
-+ * already created object; an object may have been declared in
-+ * the DPL or by calling the dpio_create() function.
-+ * This function returns a unique authentication token,
-+ * associated with the specific object ID and the specific MC
-+ * portal; this token must be used in all subsequent commands for
-+ * this specific object.
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpio_open(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ int dpio_id,
-+ u16 *token)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpio_cmd_open *dpio_cmd;
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPIO_CMDID_OPEN,
-+ cmd_flags,
-+ 0);
-+ dpio_cmd = (struct dpio_cmd_open *)cmd.params;
-+ dpio_cmd->dpio_id = cpu_to_le32(dpio_id);
-+
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ *token = mc_cmd_hdr_read_token(&cmd);
-+
-+ return 0;
-+}
-+
-+/**
-+ * dpio_close() - Close the control session of the object
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPIO object
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpio_close(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPIO_CMDID_CLOSE,
-+ cmd_flags,
-+ token);
-+
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpio_enable() - Enable the DPIO, allow I/O portal operations.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPIO object
-+ *
-+ * Return: '0' on Success; Error code otherwise
-+ */
-+int dpio_enable(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPIO_CMDID_ENABLE,
-+ cmd_flags,
-+ token);
-+
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpio_disable() - Disable the DPIO, stop any I/O portal operation.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPIO object
-+ *
-+ * Return: '0' on Success; Error code otherwise
-+ */
-+int dpio_disable(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPIO_CMDID_DISABLE,
-+ cmd_flags,
-+ token);
-+
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpio_get_attributes() - Retrieve DPIO attributes
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPIO object
-+ * @attr: Returned object's attributes
-+ *
-+ * Return: '0' on Success; Error code otherwise
-+ */
-+int dpio_get_attributes(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ struct dpio_attr *attr)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpio_rsp_get_attr *dpio_rsp;
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPIO_CMDID_GET_ATTR,
-+ cmd_flags,
-+ token);
-+
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ dpio_rsp = (struct dpio_rsp_get_attr *)cmd.params;
-+ attr->id = le32_to_cpu(dpio_rsp->id);
-+ attr->qbman_portal_id = le16_to_cpu(dpio_rsp->qbman_portal_id);
-+ attr->num_priorities = dpio_rsp->num_priorities;
-+ attr->channel_mode = dpio_rsp->channel_mode & DPIO_CHANNEL_MODE_MASK;
-+ attr->qbman_portal_ce_offset =
-+ le64_to_cpu(dpio_rsp->qbman_portal_ce_addr);
-+ attr->qbman_portal_ci_offset =
-+ le64_to_cpu(dpio_rsp->qbman_portal_ci_addr);
-+ attr->qbman_version = le32_to_cpu(dpio_rsp->qbman_version);
-+
-+ return 0;
-+}
-+
-+/**
-+ * dpio_get_api_version - Get Data Path I/O API version
-+ * @mc_io: Pointer to MC portal's DPIO object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @major_ver: Major version of DPIO API
-+ * @minor_ver: Minor version of DPIO API
-+ *
-+ * Return: '0' on Success; Error code otherwise
-+ */
-+int dpio_get_api_version(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 *major_ver,
-+ u16 *minor_ver)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPIO_CMDID_GET_API_VERSION,
-+ cmd_flags, 0);
-+
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ mc_cmd_read_api_version(&cmd, major_ver, minor_ver);
-+
-+ return 0;
-+}
-+
-+/**
-+ * dpio_reset() - Reset the DPIO, returns the object to initial state.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPIO object
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpio_reset(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPIO_CMDID_RESET,
-+ cmd_flags,
-+ token);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
---- /dev/null
-+++ b/drivers/staging/fsl-mc/bus/dpio/dpio.h
-@@ -0,0 +1,87 @@
-+/* SPDX-License-Identifier: (GPL-2.0+ OR BSD-3-Clause) */
-+/*
-+ * Copyright 2013-2016 Freescale Semiconductor Inc.
-+ * Copyright 2016 NXP
-+ *
-+ */
-+#ifndef __FSL_DPIO_H
-+#define __FSL_DPIO_H
-+
-+struct fsl_mc_io;
-+
-+int dpio_open(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ int dpio_id,
-+ u16 *token);
-+
-+int dpio_close(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token);
-+
-+/**
-+ * enum dpio_channel_mode - DPIO notification channel mode
-+ * @DPIO_NO_CHANNEL: No support for notification channel
-+ * @DPIO_LOCAL_CHANNEL: Notifications on data availability can be received by a
-+ * dedicated channel in the DPIO; user should point the queue's
-+ * destination in the relevant interface to this DPIO
-+ */
-+enum dpio_channel_mode {
-+ DPIO_NO_CHANNEL = 0,
-+ DPIO_LOCAL_CHANNEL = 1,
-+};
-+
-+/**
-+ * struct dpio_cfg - Structure representing DPIO configuration
-+ * @channel_mode: Notification channel mode
-+ * @num_priorities: Number of priorities for the notification channel (1-8);
-+ * relevant only if 'channel_mode = DPIO_LOCAL_CHANNEL'
-+ */
-+struct dpio_cfg {
-+ enum dpio_channel_mode channel_mode;
-+ u8 num_priorities;
-+};
-+
-+int dpio_enable(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token);
-+
-+int dpio_disable(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token);
-+
-+/**
-+ * struct dpio_attr - Structure representing DPIO attributes
-+ * @id: DPIO object ID
-+ * @qbman_portal_ce_offset: offset of the software portal cache-enabled area
-+ * @qbman_portal_ci_offset: offset of the software portal cache-inhibited area
-+ * @qbman_portal_id: Software portal ID
-+ * @channel_mode: Notification channel mode
-+ * @num_priorities: Number of priorities for the notification channel (1-8);
-+ * relevant only if 'channel_mode = DPIO_LOCAL_CHANNEL'
-+ * @qbman_version: QBMAN version
-+ */
-+struct dpio_attr {
-+ int id;
-+ u64 qbman_portal_ce_offset;
-+ u64 qbman_portal_ci_offset;
-+ u16 qbman_portal_id;
-+ enum dpio_channel_mode channel_mode;
-+ u8 num_priorities;
-+ u32 qbman_version;
-+};
-+
-+int dpio_get_attributes(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ struct dpio_attr *attr);
-+
-+int dpio_get_api_version(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 *major_ver,
-+ u16 *minor_ver);
-+
-+int dpio_reset(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token);
-+
-+#endif /* __FSL_DPIO_H */
---- /dev/null
-+++ b/drivers/staging/fsl-mc/bus/dpio/qbman-portal.c
-@@ -0,0 +1,1164 @@
-+// SPDX-License-Identifier: (GPL-2.0+ OR BSD-3-Clause)
-+/*
-+ * Copyright (C) 2014-2016 Freescale Semiconductor, Inc.
-+ * Copyright 2016 NXP
-+ *
-+ */
-+
-+#include <asm/cacheflush.h>
-+#include <linux/io.h>
-+#include <linux/slab.h>
-+#include "../../include/dpaa2-global.h"
-+
-+#include "qbman-portal.h"
-+
-+#define QMAN_REV_4000 0x04000000
-+#define QMAN_REV_4100 0x04010000
-+#define QMAN_REV_4101 0x04010001
-+#define QMAN_REV_MASK 0xffff0000
-+
-+/* All QBMan command and result structures use this "valid bit" encoding */
-+#define QB_VALID_BIT ((u32)0x80)
-+
-+/* QBMan portal management command codes */
-+#define QBMAN_MC_ACQUIRE 0x30
-+#define QBMAN_WQCHAN_CONFIGURE 0x46
-+
-+/* CINH register offsets */
-+#define QBMAN_CINH_SWP_EQAR 0x8c0
-+#define QBMAN_CINH_SWP_DQPI 0xa00
-+#define QBMAN_CINH_SWP_DCAP 0xac0
-+#define QBMAN_CINH_SWP_SDQCR 0xb00
-+#define QBMAN_CINH_SWP_RAR 0xcc0
-+#define QBMAN_CINH_SWP_ISR 0xe00
-+#define QBMAN_CINH_SWP_IER 0xe40
-+#define QBMAN_CINH_SWP_ISDR 0xe80
-+#define QBMAN_CINH_SWP_IIR 0xec0
-+
-+/* CENA register offsets */
-+#define QBMAN_CENA_SWP_EQCR(n) (0x000 + ((u32)(n) << 6))
-+#define QBMAN_CENA_SWP_DQRR(n) (0x200 + ((u32)(n) << 6))
-+#define QBMAN_CENA_SWP_RCR(n) (0x400 + ((u32)(n) << 6))
-+#define QBMAN_CENA_SWP_CR 0x600
-+#define QBMAN_CENA_SWP_RR(vb) (0x700 + ((u32)(vb) >> 1))
-+#define QBMAN_CENA_SWP_VDQCR 0x780
-+
-+/* Reverse mapping of QBMAN_CENA_SWP_DQRR() */
-+#define QBMAN_IDX_FROM_DQRR(p) (((unsigned long)(p) & 0x1ff) >> 6)
-+
-+/* Define token used to determine if response written to memory is valid */
-+#define QMAN_DQ_TOKEN_VALID 1
-+
-+/* SDQCR attribute codes */
-+#define QB_SDQCR_FC_SHIFT 29
-+#define QB_SDQCR_FC_MASK 0x1
-+#define QB_SDQCR_DCT_SHIFT 24
-+#define QB_SDQCR_DCT_MASK 0x3
-+#define QB_SDQCR_TOK_SHIFT 16
-+#define QB_SDQCR_TOK_MASK 0xff
-+#define QB_SDQCR_SRC_SHIFT 0
-+#define QB_SDQCR_SRC_MASK 0xffff
-+
-+/* opaque token for static dequeues */
-+#define QMAN_SDQCR_TOKEN 0xbb
-+
-+enum qbman_sdqcr_dct {
-+ qbman_sdqcr_dct_null = 0,
-+ qbman_sdqcr_dct_prio_ics,
-+ qbman_sdqcr_dct_active_ics,
-+ qbman_sdqcr_dct_active
-+};
-+
-+enum qbman_sdqcr_fc {
-+ qbman_sdqcr_fc_one = 0,
-+ qbman_sdqcr_fc_up_to_3 = 1
-+};
-+
-+#define dccvac(p) { asm volatile("dc cvac, %0;" : : "r" (p) : "memory"); }
-+#define dcivac(p) { asm volatile("dc ivac, %0" : : "r"(p) : "memory"); }
-+static inline void qbman_inval_prefetch(struct qbman_swp *p, uint32_t offset)
-+{
-+ dcivac(p->addr_cena + offset);
-+ prefetch(p->addr_cena + offset);
-+}
-+
-+/* Portal Access */
-+
-+static inline u32 qbman_read_register(struct qbman_swp *p, u32 offset)
-+{
-+ return readl_relaxed(p->addr_cinh + offset);
-+}
-+
-+static inline void qbman_write_register(struct qbman_swp *p, u32 offset,
-+ u32 value)
-+{
-+ writel_relaxed(value, p->addr_cinh + offset);
-+}
-+
-+static inline void *qbman_get_cmd(struct qbman_swp *p, u32 offset)
-+{
-+ return p->addr_cena + offset;
-+}
-+
-+#define QBMAN_CINH_SWP_CFG 0xd00
-+
-+#define SWP_CFG_DQRR_MF_SHIFT 20
-+#define SWP_CFG_EST_SHIFT 16
-+#define SWP_CFG_WN_SHIFT 14
-+#define SWP_CFG_RPM_SHIFT 12
-+#define SWP_CFG_DCM_SHIFT 10
-+#define SWP_CFG_EPM_SHIFT 8
-+#define SWP_CFG_SD_SHIFT 5
-+#define SWP_CFG_SP_SHIFT 4
-+#define SWP_CFG_SE_SHIFT 3
-+#define SWP_CFG_DP_SHIFT 2
-+#define SWP_CFG_DE_SHIFT 1
-+#define SWP_CFG_EP_SHIFT 0
-+
-+static inline u32 qbman_set_swp_cfg(u8 max_fill, u8 wn, u8 est, u8 rpm, u8 dcm,
-+ u8 epm, int sd, int sp, int se,
-+ int dp, int de, int ep)
-+{
-+ return (max_fill << SWP_CFG_DQRR_MF_SHIFT |
-+ est << SWP_CFG_EST_SHIFT |
-+ wn << SWP_CFG_WN_SHIFT |
-+ rpm << SWP_CFG_RPM_SHIFT |
-+ dcm << SWP_CFG_DCM_SHIFT |
-+ epm << SWP_CFG_EPM_SHIFT |
-+ sd << SWP_CFG_SD_SHIFT |
-+ sp << SWP_CFG_SP_SHIFT |
-+ se << SWP_CFG_SE_SHIFT |
-+ dp << SWP_CFG_DP_SHIFT |
-+ de << SWP_CFG_DE_SHIFT |
-+ ep << SWP_CFG_EP_SHIFT);
-+}
-+
-+/**
-+ * qbman_swp_init() - Create a functional object representing the given
-+ * QBMan portal descriptor.
-+ * @d: the given qbman swp descriptor
-+ *
-+ * Return qbman_swp portal for success, NULL if the object cannot
-+ * be created.
-+ */
-+struct qbman_swp *qbman_swp_init(const struct qbman_swp_desc *d)
-+{
-+ struct qbman_swp *p = kmalloc(sizeof(*p), GFP_KERNEL);
-+ u32 reg;
-+
-+ if (!p)
-+ return NULL;
-+ p->desc = d;
-+ p->mc.valid_bit = QB_VALID_BIT;
-+ p->sdq = 0;
-+ p->sdq |= qbman_sdqcr_dct_prio_ics << QB_SDQCR_DCT_SHIFT;
-+ p->sdq |= qbman_sdqcr_fc_up_to_3 << QB_SDQCR_FC_SHIFT;
-+ p->sdq |= QMAN_SDQCR_TOKEN << QB_SDQCR_TOK_SHIFT;
-+
-+ atomic_set(&p->vdq.available, 1);
-+ p->vdq.valid_bit = QB_VALID_BIT;
-+ p->dqrr.next_idx = 0;
-+ p->dqrr.valid_bit = QB_VALID_BIT;
-+
-+ if ((p->desc->qman_version & QMAN_REV_MASK) < QMAN_REV_4100) {
-+ p->dqrr.dqrr_size = 4;
-+ p->dqrr.reset_bug = 1;
-+ } else {
-+ p->dqrr.dqrr_size = 8;
-+ p->dqrr.reset_bug = 0;
-+ }
-+
-+ p->addr_cena = d->cena_bar;
-+ p->addr_cinh = d->cinh_bar;
-+
-+ reg = qbman_set_swp_cfg(p->dqrr.dqrr_size,
-+ 0, /* Writes cacheable */
-+ 0, /* EQCR_CI stashing threshold */
-+ 3, /* RPM: Valid bit mode, RCR in array mode */
-+ 2, /* DCM: Discrete consumption ack mode */
-+ 3, /* EPM: Valid bit mode, EQCR in array mode */
-+ 0, /* mem stashing drop enable == FALSE */
-+ 1, /* mem stashing priority == TRUE */
-+ 0, /* mem stashing enable == FALSE */
-+ 1, /* dequeue stashing priority == TRUE */
-+ 0, /* dequeue stashing enable == FALSE */
-+ 0); /* EQCR_CI stashing priority == FALSE */
-+
-+ qbman_write_register(p, QBMAN_CINH_SWP_CFG, reg);
-+ reg = qbman_read_register(p, QBMAN_CINH_SWP_CFG);
-+ if (!reg) {
-+ pr_err("qbman: the portal is not enabled!\n");
-+ return NULL;
-+ }
-+
-+ /*
-+ * SDQCR needs to be initialized to 0 when no channels are
-+ * being dequeued from or else the QMan HW will indicate an
-+ * error. The values that were calculated above will be
-+ * applied when dequeues from a specific channel are enabled.
-+ */
-+ qbman_write_register(p, QBMAN_CINH_SWP_SDQCR, 0);
-+ return p;
-+}
-+
-+/**
-+ * qbman_swp_finish() - Create and destroy a functional object representing
-+ * the given QBMan portal descriptor.
-+ * @p: the qbman_swp object to be destroyed
-+ */
-+void qbman_swp_finish(struct qbman_swp *p)
-+{
-+ kfree(p);
-+}
-+
-+/**
-+ * qbman_swp_interrupt_read_status()
-+ * @p: the given software portal
-+ *
-+ * Return the value in the SWP_ISR register.
-+ */
-+u32 qbman_swp_interrupt_read_status(struct qbman_swp *p)
-+{
-+ return qbman_read_register(p, QBMAN_CINH_SWP_ISR);
-+}
-+
-+/**
-+ * qbman_swp_interrupt_clear_status()
-+ * @p: the given software portal
-+ * @mask: The mask to clear in SWP_ISR register
-+ */
-+void qbman_swp_interrupt_clear_status(struct qbman_swp *p, u32 mask)
-+{
-+ qbman_write_register(p, QBMAN_CINH_SWP_ISR, mask);
-+}
-+
-+/**
-+ * qbman_swp_interrupt_get_trigger() - read interrupt enable register
-+ * @p: the given software portal
-+ *
-+ * Return the value in the SWP_IER register.
-+ */
-+u32 qbman_swp_interrupt_get_trigger(struct qbman_swp *p)
-+{
-+ return qbman_read_register(p, QBMAN_CINH_SWP_IER);
-+}
-+
-+/**
-+ * qbman_swp_interrupt_set_trigger() - enable interrupts for a swp
-+ * @p: the given software portal
-+ * @mask: The mask of bits to enable in SWP_IER
-+ */
-+void qbman_swp_interrupt_set_trigger(struct qbman_swp *p, u32 mask)
-+{
-+ qbman_write_register(p, QBMAN_CINH_SWP_IER, mask);
-+}
-+
-+/**
-+ * qbman_swp_interrupt_get_inhibit() - read interrupt mask register
-+ * @p: the given software portal object
-+ *
-+ * Return the value in the SWP_IIR register.
-+ */
-+int qbman_swp_interrupt_get_inhibit(struct qbman_swp *p)
-+{
-+ return qbman_read_register(p, QBMAN_CINH_SWP_IIR);
-+}
-+
-+/**
-+ * qbman_swp_interrupt_set_inhibit() - write interrupt mask register
-+ * @p: the given software portal object
-+ * @mask: The mask to set in SWP_IIR register
-+ */
-+void qbman_swp_interrupt_set_inhibit(struct qbman_swp *p, int inhibit)
-+{
-+ qbman_write_register(p, QBMAN_CINH_SWP_IIR, inhibit ? 0xffffffff : 0);
-+}
-+
-+/*
-+ * Different management commands all use this common base layer of code to issue
-+ * commands and poll for results.
-+ */
-+
-+/*
-+ * Returns a pointer to where the caller should fill in their management command
-+ * (caller should ignore the verb byte)
-+ */
-+void *qbman_swp_mc_start(struct qbman_swp *p)
-+{
-+ return qbman_get_cmd(p, QBMAN_CENA_SWP_CR);
-+}
-+
-+/*
-+ * Commits merges in the caller-supplied command verb (which should not include
-+ * the valid-bit) and submits the command to hardware
-+ */
-+void qbman_swp_mc_submit(struct qbman_swp *p, void *cmd, u8 cmd_verb)
-+{
-+ u8 *v = cmd;
-+
-+ dma_wmb();
-+ *v = cmd_verb | p->mc.valid_bit;
-+ dccvac(cmd);
-+}
-+
-+/*
-+ * Checks for a completed response (returns non-NULL if only if the response
-+ * is complete).
-+ */
-+void *qbman_swp_mc_result(struct qbman_swp *p)
-+{
-+ u32 *ret, verb;
-+
-+ qbman_inval_prefetch(p, QBMAN_CENA_SWP_RR(p->mc.valid_bit));
-+ ret = qbman_get_cmd(p, QBMAN_CENA_SWP_RR(p->mc.valid_bit));
-+
-+ /* Remove the valid-bit - command completed if the rest is non-zero */
-+ verb = ret[0] & ~QB_VALID_BIT;
-+ if (!verb)
-+ return NULL;
-+ p->mc.valid_bit ^= QB_VALID_BIT;
-+ return ret;
-+}
-+
-+#define QB_ENQUEUE_CMD_OPTIONS_SHIFT 0
-+enum qb_enqueue_commands {
-+ enqueue_empty = 0,
-+ enqueue_response_always = 1,
-+ enqueue_rejects_to_fq = 2
-+};
-+
-+#define QB_ENQUEUE_CMD_ORP_ENABLE_SHIFT 2
-+#define QB_ENQUEUE_CMD_IRQ_ON_DISPATCH_SHIFT 3
-+#define QB_ENQUEUE_CMD_TARGET_TYPE_SHIFT 4
-+
-+/**
-+ * qbman_eq_desc_clear() - Clear the contents of a descriptor to
-+ * default/starting state.
-+ */
-+void qbman_eq_desc_clear(struct qbman_eq_desc *d)
-+{
-+ memset(d, 0, sizeof(*d));
-+}
-+
-+/**
-+ * qbman_eq_desc_set_no_orp() - Set enqueue descriptor without orp
-+ * @d: the enqueue descriptor.
-+ * @response_success: 1 = enqueue with response always; 0 = enqueue with
-+ * rejections returned on a FQ.
-+ */
-+void qbman_eq_desc_set_no_orp(struct qbman_eq_desc *d, int respond_success)
-+{
-+ d->verb &= ~(1 << QB_ENQUEUE_CMD_ORP_ENABLE_SHIFT);
-+ if (respond_success)
-+ d->verb |= enqueue_response_always;
-+ else
-+ d->verb |= enqueue_rejects_to_fq;
-+}
-+
-+/**
-+ * qbman_eq_desc_set_orp() - Set order-restoration in the enqueue descriptor
-+ * @d: the enqueue descriptor.
-+ * @response_success: 1 = enqueue with response always; 0 = enqueue with
-+ * rejections returned on a FQ.
-+ * @oprid: the order point record id.
-+ * @seqnum: the order restoration sequence number.
-+ * @incomplete: indicates whether this is the last fragments using the same
-+ * sequence number.
-+ */
-+void qbman_eq_desc_set_orp(struct qbman_eq_desc *d, int respond_success,
-+ u16 oprid, u16 seqnum, int incomplete)
-+{
-+ d->verb |= (1 << QB_ENQUEUE_CMD_ORP_ENABLE_SHIFT);
-+ if (respond_success)
-+ d->verb |= enqueue_response_always;
-+ else
-+ d->verb |= enqueue_rejects_to_fq;
-+ d->orpid = cpu_to_le16(oprid);
-+ d->seqnum = cpu_to_le16((!!incomplete << 14) | seqnum);
-+}
-+
-+/**
-+ * qbman_eq_desc_set_orp_hole() - fill a hole in the order-restoration sequence
-+ * without any enqueue
-+ * @d: the enqueue descriptor.
-+ * @oprid: the order point record id.
-+ * @seqnum: the order restoration sequence number.
-+ */
-+void qbman_eq_desc_set_orp_hole(struct qbman_eq_desc *d, u16 oprid,
-+ u16 seqnum)
-+{
-+ d->verb |= (1 << QB_ENQUEUE_CMD_ORP_ENABLE_SHIFT) | enqueue_empty;
-+ d->orpid = cpu_to_le16(oprid);
-+ d->seqnum = cpu_to_le16(seqnum);
-+}
-+
-+/*
-+ * Exactly one of the following descriptor "targets" should be set. (Calling any
-+ * one of these will replace the effect of any prior call to one of these.)
-+ * -enqueue to a frame queue
-+ * -enqueue to a queuing destination
-+ */
-+
-+/**
-+ * qbman_eq_desc_set_fq() - set the FQ for the enqueue command
-+ * @d: the enqueue descriptor
-+ * @fqid: the id of the frame queue to be enqueued
-+ */
-+void qbman_eq_desc_set_fq(struct qbman_eq_desc *d, u32 fqid)
-+{
-+ d->verb &= ~(1 << QB_ENQUEUE_CMD_TARGET_TYPE_SHIFT);
-+ d->tgtid = cpu_to_le32(fqid);
-+}
-+
-+/**
-+ * qbman_eq_desc_set_qd() - Set Queuing Destination for the enqueue command
-+ * @d: the enqueue descriptor
-+ * @qdid: the id of the queuing destination to be enqueued
-+ * @qd_bin: the queuing destination bin
-+ * @qd_prio: the queuing destination priority
-+ */
-+void qbman_eq_desc_set_qd(struct qbman_eq_desc *d, u32 qdid,
-+ u32 qd_bin, u32 qd_prio)
-+{
-+ d->verb |= 1 << QB_ENQUEUE_CMD_TARGET_TYPE_SHIFT;
-+ d->tgtid = cpu_to_le32(qdid);
-+ d->qdbin = cpu_to_le16(qd_bin);
-+ d->qpri = qd_prio;
-+}
-+
-+#define EQAR_IDX(eqar) ((eqar) & 0x7)
-+#define EQAR_VB(eqar) ((eqar) & 0x80)
-+#define EQAR_SUCCESS(eqar) ((eqar) & 0x100)
-+
-+/**
-+ * qbman_swp_enqueue() - Issue an enqueue command
-+ * @s: the software portal used for enqueue
-+ * @d: the enqueue descriptor
-+ * @fd: the frame descriptor to be enqueued
-+ *
-+ * Please note that 'fd' should only be NULL if the "action" of the
-+ * descriptor is "orp_hole" or "orp_nesn".
-+ *
-+ * Return 0 for successful enqueue, -EBUSY if the EQCR is not ready.
-+ */
-+int qbman_swp_enqueue(struct qbman_swp *s, const struct qbman_eq_desc *d,
-+ const struct dpaa2_fd *fd)
-+{
-+ struct qbman_eq_desc *p;
-+ u32 eqar = qbman_read_register(s, QBMAN_CINH_SWP_EQAR);
-+
-+ if (!EQAR_SUCCESS(eqar))
-+ return -EBUSY;
-+
-+ p = qbman_get_cmd(s, QBMAN_CENA_SWP_EQCR(EQAR_IDX(eqar)));
-+ /* This is mapped as DEVICE type memory, writes are
-+ * with address alignment:
-+ * desc.dca address alignment = 1
-+ * desc.seqnum address alignment = 2
-+ * desc.orpid address alignment = 4
-+ * desc.tgtid address alignment = 8
-+ */
-+ p->dca = d->dca;
-+ p->seqnum = d->seqnum;
-+ p->orpid = d->orpid;
-+ memcpy(&p->tgtid, &d->tgtid, 24);
-+ memcpy(&p->fd, fd, sizeof(*fd));
-+
-+ /* Set the verb byte, have to substitute in the valid-bit */
-+ dma_wmb();
-+ p->verb = d->verb | EQAR_VB(eqar);
-+ dccvac(p);
-+
-+ return 0;
-+}
-+
-+/* Static (push) dequeue */
-+
-+/**
-+ * qbman_swp_push_get() - Get the push dequeue setup
-+ * @p: the software portal object
-+ * @channel_idx: the channel index to query
-+ * @enabled: returned boolean to show whether the push dequeue is enabled
-+ * for the given channel
-+ */
-+void qbman_swp_push_get(struct qbman_swp *s, u8 channel_idx, int *enabled)
-+{
-+ u16 src = (s->sdq >> QB_SDQCR_SRC_SHIFT) & QB_SDQCR_SRC_MASK;
-+
-+ WARN_ON(channel_idx > 15);
-+ *enabled = src | (1 << channel_idx);
-+}
-+
-+/**
-+ * qbman_swp_push_set() - Enable or disable push dequeue
-+ * @p: the software portal object
-+ * @channel_idx: the channel index (0 to 15)
-+ * @enable: enable or disable push dequeue
-+ */
-+void qbman_swp_push_set(struct qbman_swp *s, u8 channel_idx, int enable)
-+{
-+ u16 dqsrc;
-+
-+ WARN_ON(channel_idx > 15);
-+ if (enable)
-+ s->sdq |= 1 << channel_idx;
-+ else
-+ s->sdq &= ~(1 << channel_idx);
-+
-+ /* Read make the complete src map. If no channels are enabled
-+ * the SDQCR must be 0 or else QMan will assert errors
-+ */
-+ dqsrc = (s->sdq >> QB_SDQCR_SRC_SHIFT) & QB_SDQCR_SRC_MASK;
-+ if (dqsrc != 0)
-+ qbman_write_register(s, QBMAN_CINH_SWP_SDQCR, s->sdq);
-+ else
-+ qbman_write_register(s, QBMAN_CINH_SWP_SDQCR, 0);
-+}
-+
-+#define QB_VDQCR_VERB_DCT_SHIFT 0
-+#define QB_VDQCR_VERB_DT_SHIFT 2
-+#define QB_VDQCR_VERB_RLS_SHIFT 4
-+#define QB_VDQCR_VERB_WAE_SHIFT 5
-+
-+enum qb_pull_dt_e {
-+ qb_pull_dt_channel,
-+ qb_pull_dt_workqueue,
-+ qb_pull_dt_framequeue
-+};
-+
-+/**
-+ * qbman_pull_desc_clear() - Clear the contents of a descriptor to
-+ * default/starting state
-+ * @d: the pull dequeue descriptor to be cleared
-+ */
-+void qbman_pull_desc_clear(struct qbman_pull_desc *d)
-+{
-+ memset(d, 0, sizeof(*d));
-+}
-+
-+/**
-+ * qbman_pull_desc_set_storage()- Set the pull dequeue storage
-+ * @d: the pull dequeue descriptor to be set
-+ * @storage: the pointer of the memory to store the dequeue result
-+ * @storage_phys: the physical address of the storage memory
-+ * @stash: to indicate whether write allocate is enabled
-+ *
-+ * If not called, or if called with 'storage' as NULL, the result pull dequeues
-+ * will produce results to DQRR. If 'storage' is non-NULL, then results are
-+ * produced to the given memory location (using the DMA address which
-+ * the caller provides in 'storage_phys'), and 'stash' controls whether or not
-+ * those writes to main-memory express a cache-warming attribute.
-+ */
-+void qbman_pull_desc_set_storage(struct qbman_pull_desc *d,
-+ struct dpaa2_dq *storage,
-+ dma_addr_t storage_phys,
-+ int stash)
-+{
-+ /* save the virtual address */
-+ d->rsp_addr_virt = (u64)(uintptr_t)storage;
-+
-+ if (!storage) {
-+ d->verb &= ~(1 << QB_VDQCR_VERB_RLS_SHIFT);
-+ return;
-+ }
-+ d->verb |= 1 << QB_VDQCR_VERB_RLS_SHIFT;
-+ if (stash)
-+ d->verb |= 1 << QB_VDQCR_VERB_WAE_SHIFT;
-+ else
-+ d->verb &= ~(1 << QB_VDQCR_VERB_WAE_SHIFT);
-+
-+ d->rsp_addr = cpu_to_le64(storage_phys);
-+}
-+
-+/**
-+ * qbman_pull_desc_set_numframes() - Set the number of frames to be dequeued
-+ * @d: the pull dequeue descriptor to be set
-+ * @numframes: number of frames to be set, must be between 1 and 16, inclusive
-+ */
-+void qbman_pull_desc_set_numframes(struct qbman_pull_desc *d, u8 numframes)
-+{
-+ d->numf = numframes - 1;
-+}
-+
-+void qbman_pull_desc_set_token(struct qbman_pull_desc *d, u8 token)
-+{
-+ d->tok = token;
-+}
-+
-+/*
-+ * Exactly one of the following descriptor "actions" should be set. (Calling any
-+ * one of these will replace the effect of any prior call to one of these.)
-+ * - pull dequeue from the given frame queue (FQ)
-+ * - pull dequeue from any FQ in the given work queue (WQ)
-+ * - pull dequeue from any FQ in any WQ in the given channel
-+ */
-+
-+/**
-+ * qbman_pull_desc_set_fq() - Set fqid from which the dequeue command dequeues
-+ * @fqid: the frame queue index of the given FQ
-+ */
-+void qbman_pull_desc_set_fq(struct qbman_pull_desc *d, u32 fqid)
-+{
-+ d->verb |= 1 << QB_VDQCR_VERB_DCT_SHIFT;
-+ d->verb |= qb_pull_dt_framequeue << QB_VDQCR_VERB_DT_SHIFT;
-+ d->dq_src = cpu_to_le32(fqid);
-+}
-+
-+/**
-+ * qbman_pull_desc_set_wq() - Set wqid from which the dequeue command dequeues
-+ * @wqid: composed of channel id and wqid within the channel
-+ * @dct: the dequeue command type
-+ */
-+void qbman_pull_desc_set_wq(struct qbman_pull_desc *d, u32 wqid,
-+ enum qbman_pull_type_e dct)
-+{
-+ d->verb |= dct << QB_VDQCR_VERB_DCT_SHIFT;
-+ d->verb |= qb_pull_dt_workqueue << QB_VDQCR_VERB_DT_SHIFT;
-+ d->dq_src = cpu_to_le32(wqid);
-+}
-+
-+/**
-+ * qbman_pull_desc_set_channel() - Set channelid from which the dequeue command
-+ * dequeues
-+ * @chid: the channel id to be dequeued
-+ * @dct: the dequeue command type
-+ */
-+void qbman_pull_desc_set_channel(struct qbman_pull_desc *d, u32 chid,
-+ enum qbman_pull_type_e dct)
-+{
-+ d->verb |= dct << QB_VDQCR_VERB_DCT_SHIFT;
-+ d->verb |= qb_pull_dt_channel << QB_VDQCR_VERB_DT_SHIFT;
-+ d->dq_src = cpu_to_le32(chid);
-+}
-+
-+/**
-+ * qbman_swp_pull() - Issue the pull dequeue command
-+ * @s: the software portal object
-+ * @d: the software portal descriptor which has been configured with
-+ * the set of qbman_pull_desc_set_*() calls
-+ *
-+ * Return 0 for success, and -EBUSY if the software portal is not ready
-+ * to do pull dequeue.
-+ */
-+int qbman_swp_pull(struct qbman_swp *s, struct qbman_pull_desc *d)
-+{
-+ struct qbman_pull_desc *p;
-+
-+ if (!atomic_dec_and_test(&s->vdq.available)) {
-+ atomic_inc(&s->vdq.available);
-+ return -EBUSY;
-+ }
-+ s->vdq.storage = (void *)(uintptr_t)d->rsp_addr_virt;
-+ p = qbman_get_cmd(s, QBMAN_CENA_SWP_VDQCR);
-+ p->numf = d->numf;
-+ p->tok = QMAN_DQ_TOKEN_VALID;
-+ p->dq_src = d->dq_src;
-+ p->rsp_addr = d->rsp_addr;
-+ p->rsp_addr_virt = d->rsp_addr_virt;
-+ dma_wmb();
-+
-+ /* Set the verb byte, have to substitute in the valid-bit */
-+ p->verb = d->verb | s->vdq.valid_bit;
-+ s->vdq.valid_bit ^= QB_VALID_BIT;
-+ dccvac(p);
-+
-+ return 0;
-+}
-+
-+#define QMAN_DQRR_PI_MASK 0xf
-+
-+/**
-+ * qbman_swp_dqrr_next() - Get an valid DQRR entry
-+ * @s: the software portal object
-+ *
-+ * Return NULL if there are no unconsumed DQRR entries. Return a DQRR entry
-+ * only once, so repeated calls can return a sequence of DQRR entries, without
-+ * requiring they be consumed immediately or in any particular order.
-+ */
-+const struct dpaa2_dq *qbman_swp_dqrr_next(struct qbman_swp *s)
-+{
-+ u32 verb;
-+ u32 response_verb;
-+ u32 flags;
-+ struct dpaa2_dq *p;
-+
-+ /* Before using valid-bit to detect if something is there, we have to
-+ * handle the case of the DQRR reset bug...
-+ */
-+ if (unlikely(s->dqrr.reset_bug)) {
-+ /*
-+ * We pick up new entries by cache-inhibited producer index,
-+ * which means that a non-coherent mapping would require us to
-+ * invalidate and read *only* once that PI has indicated that
-+ * there's an entry here. The first trip around the DQRR ring
-+ * will be much less efficient than all subsequent trips around
-+ * it...
-+ */
-+ u8 pi = qbman_read_register(s, QBMAN_CINH_SWP_DQPI) &
-+ QMAN_DQRR_PI_MASK;
-+
-+ /* there are new entries if pi != next_idx */
-+ if (pi == s->dqrr.next_idx)
-+ return NULL;
-+
-+ /*
-+ * if next_idx is/was the last ring index, and 'pi' is
-+ * different, we can disable the workaround as all the ring
-+ * entries have now been DMA'd to so valid-bit checking is
-+ * repaired. Note: this logic needs to be based on next_idx
-+ * (which increments one at a time), rather than on pi (which
-+ * can burst and wrap-around between our snapshots of it).
-+ */
-+ if (s->dqrr.next_idx == (s->dqrr.dqrr_size - 1)) {
-+ pr_debug("next_idx=%d, pi=%d, clear reset bug\n",
-+ s->dqrr.next_idx, pi);
-+ s->dqrr.reset_bug = 0;
-+ }
-+ qbman_inval_prefetch(s, QBMAN_CENA_SWP_DQRR(s->dqrr.next_idx));
-+ }
-+
-+ p = qbman_get_cmd(s, QBMAN_CENA_SWP_DQRR(s->dqrr.next_idx));
-+ verb = p->dq.verb;
-+
-+ /*
-+ * If the valid-bit isn't of the expected polarity, nothing there. Note,
-+ * in the DQRR reset bug workaround, we shouldn't need to skip these
-+ * check, because we've already determined that a new entry is available
-+ * and we've invalidated the cacheline before reading it, so the
-+ * valid-bit behaviour is repaired and should tell us what we already
-+ * knew from reading PI.
-+ */
-+ if ((verb & QB_VALID_BIT) != s->dqrr.valid_bit) {
-+ qbman_inval_prefetch(s, QBMAN_CENA_SWP_DQRR(s->dqrr.next_idx));
-+ return NULL;
-+ }
-+ /*
-+ * There's something there. Move "next_idx" attention to the next ring
-+ * entry (and prefetch it) before returning what we found.
-+ */
-+ s->dqrr.next_idx++;
-+ s->dqrr.next_idx &= s->dqrr.dqrr_size - 1; /* Wrap around */
-+ if (!s->dqrr.next_idx)
-+ s->dqrr.valid_bit ^= QB_VALID_BIT;
-+
-+ /*
-+ * If this is the final response to a volatile dequeue command
-+ * indicate that the vdq is available
-+ */
-+ flags = p->dq.stat;
-+ response_verb = verb & QBMAN_RESULT_MASK;
-+ if ((response_verb == QBMAN_RESULT_DQ) &&
-+ (flags & DPAA2_DQ_STAT_VOLATILE) &&
-+ (flags & DPAA2_DQ_STAT_EXPIRED))
-+ atomic_inc(&s->vdq.available);
-+
-+ qbman_inval_prefetch(s, QBMAN_CENA_SWP_DQRR(s->dqrr.next_idx));
-+
-+ return p;
-+}
-+
-+/**
-+ * qbman_swp_dqrr_consume() - Consume DQRR entries previously returned from
-+ * qbman_swp_dqrr_next().
-+ * @s: the software portal object
-+ * @dq: the DQRR entry to be consumed
-+ */
-+void qbman_swp_dqrr_consume(struct qbman_swp *s, const struct dpaa2_dq *dq)
-+{
-+ qbman_write_register(s, QBMAN_CINH_SWP_DCAP, QBMAN_IDX_FROM_DQRR(dq));
-+}
-+
-+/**
-+ * qbman_result_has_new_result() - Check and get the dequeue response from the
-+ * dq storage memory set in pull dequeue command
-+ * @s: the software portal object
-+ * @dq: the dequeue result read from the memory
-+ *
-+ * Return 1 for getting a valid dequeue result, or 0 for not getting a valid
-+ * dequeue result.
-+ *
-+ * Only used for user-provided storage of dequeue results, not DQRR. For
-+ * efficiency purposes, the driver will perform any required endianness
-+ * conversion to ensure that the user's dequeue result storage is in host-endian
-+ * format. As such, once the user has called qbman_result_has_new_result() and
-+ * been returned a valid dequeue result, they should not call it again on
-+ * the same memory location (except of course if another dequeue command has
-+ * been executed to produce a new result to that location).
-+ */
-+int qbman_result_has_new_result(struct qbman_swp *s, const struct dpaa2_dq *dq)
-+{
-+ if (dq->dq.tok != QMAN_DQ_TOKEN_VALID)
-+ return 0;
-+
-+ /*
-+ * Set token to be 0 so we will detect change back to 1
-+ * next time the looping is traversed. Const is cast away here
-+ * as we want users to treat the dequeue responses as read only.
-+ */
-+ ((struct dpaa2_dq *)dq)->dq.tok = 0;
-+
-+ /*
-+ * Determine whether VDQCR is available based on whether the
-+ * current result is sitting in the first storage location of
-+ * the busy command.
-+ */
-+ if (s->vdq.storage == dq) {
-+ s->vdq.storage = NULL;
-+ atomic_inc(&s->vdq.available);
-+ }
-+
-+ return 1;
-+}
-+
-+/**
-+ * qbman_release_desc_clear() - Clear the contents of a descriptor to
-+ * default/starting state.
-+ */
-+void qbman_release_desc_clear(struct qbman_release_desc *d)
-+{
-+ memset(d, 0, sizeof(*d));
-+ d->verb = 1 << 5; /* Release Command Valid */
-+}
-+
-+/**
-+ * qbman_release_desc_set_bpid() - Set the ID of the buffer pool to release to
-+ */
-+void qbman_release_desc_set_bpid(struct qbman_release_desc *d, u16 bpid)
-+{
-+ d->bpid = cpu_to_le16(bpid);
-+}
-+
-+/**
-+ * qbman_release_desc_set_rcdi() - Determines whether or not the portal's RCDI
-+ * interrupt source should be asserted after the release command is completed.
-+ */
-+void qbman_release_desc_set_rcdi(struct qbman_release_desc *d, int enable)
-+{
-+ if (enable)
-+ d->verb |= 1 << 6;
-+ else
-+ d->verb &= ~(1 << 6);
-+}
-+
-+#define RAR_IDX(rar) ((rar) & 0x7)
-+#define RAR_VB(rar) ((rar) & 0x80)
-+#define RAR_SUCCESS(rar) ((rar) & 0x100)
-+
-+/**
-+ * qbman_swp_release() - Issue a buffer release command
-+ * @s: the software portal object
-+ * @d: the release descriptor
-+ * @buffers: a pointer pointing to the buffer address to be released
-+ * @num_buffers: number of buffers to be released, must be less than 8
-+ *
-+ * Return 0 for success, -EBUSY if the release command ring is not ready.
-+ */
-+int qbman_swp_release(struct qbman_swp *s, const struct qbman_release_desc *d,
-+ const u64 *buffers, unsigned int num_buffers)
-+{
-+ int i;
-+ struct qbman_release_desc *p;
-+ u32 rar;
-+
-+ if (!num_buffers || (num_buffers > 7))
-+ return -EINVAL;
-+
-+ rar = qbman_read_register(s, QBMAN_CINH_SWP_RAR);
-+ if (!RAR_SUCCESS(rar))
-+ return -EBUSY;
-+
-+ /* Start the release command */
-+ p = qbman_get_cmd(s, QBMAN_CENA_SWP_RCR(RAR_IDX(rar)));
-+ /* Copy the caller's buffer pointers to the command */
-+ for (i = 0; i < num_buffers; i++)
-+ p->buf[i] = cpu_to_le64(buffers[i]);
-+ p->bpid = d->bpid;
-+
-+ /*
-+ * Set the verb byte, have to substitute in the valid-bit and the number
-+ * of buffers.
-+ */
-+ dma_wmb();
-+ p->verb = d->verb | RAR_VB(rar) | num_buffers;
-+ dccvac(p);
-+
-+ return 0;
-+}
-+
-+struct qbman_acquire_desc {
-+ u8 verb;
-+ u8 reserved;
-+ __le16 bpid;
-+ u8 num;
-+ u8 reserved2[59];
-+};
-+
-+struct qbman_acquire_rslt {
-+ u8 verb;
-+ u8 rslt;
-+ __le16 reserved;
-+ u8 num;
-+ u8 reserved2[3];
-+ __le64 buf[7];
-+};
-+
-+/**
-+ * qbman_swp_acquire() - Issue a buffer acquire command
-+ * @s: the software portal object
-+ * @bpid: the buffer pool index
-+ * @buffers: a pointer pointing to the acquired buffer addresses
-+ * @num_buffers: number of buffers to be acquired, must be less than 8
-+ *
-+ * Return 0 for success, or negative error code if the acquire command
-+ * fails.
-+ */
-+int qbman_swp_acquire(struct qbman_swp *s, u16 bpid, u64 *buffers,
-+ unsigned int num_buffers)
-+{
-+ struct qbman_acquire_desc *p;
-+ struct qbman_acquire_rslt *r;
-+ int i;
-+
-+ if (!num_buffers || (num_buffers > 7))
-+ return -EINVAL;
-+
-+ /* Start the management command */
-+ p = qbman_swp_mc_start(s);
-+
-+ if (!p)
-+ return -EBUSY;
-+
-+ /* Encode the caller-provided attributes */
-+ p->bpid = cpu_to_le16(bpid);
-+ p->num = num_buffers;
-+
-+ /* Complete the management command */
-+ r = qbman_swp_mc_complete(s, p, QBMAN_MC_ACQUIRE);
-+ if (unlikely(!r)) {
-+ pr_err("qbman: acquire from BPID %d failed, no response\n",
-+ bpid);
-+ return -EIO;
-+ }
-+
-+ /* Decode the outcome */
-+ WARN_ON((r->verb & 0x7f) != QBMAN_MC_ACQUIRE);
-+
-+ /* Determine success or failure */
-+ if (unlikely(r->rslt != QBMAN_MC_RSLT_OK)) {
-+ pr_err("qbman: acquire from BPID 0x%x failed, code=0x%02x\n",
-+ bpid, r->rslt);
-+ return -EIO;
-+ }
-+
-+ WARN_ON(r->num > num_buffers);
-+
-+ /* Copy the acquired buffers to the caller's array */
-+ for (i = 0; i < r->num; i++)
-+ buffers[i] = le64_to_cpu(r->buf[i]);
-+
-+ return (int)r->num;
-+}
-+
-+struct qbman_alt_fq_state_desc {
-+ u8 verb;
-+ u8 reserved[3];
-+ __le32 fqid;
-+ u8 reserved2[56];
-+};
-+
-+struct qbman_alt_fq_state_rslt {
-+ u8 verb;
-+ u8 rslt;
-+ u8 reserved[62];
-+};
-+
-+#define ALT_FQ_FQID_MASK 0x00FFFFFF
-+
-+int qbman_swp_alt_fq_state(struct qbman_swp *s, u32 fqid,
-+ u8 alt_fq_verb)
-+{
-+ struct qbman_alt_fq_state_desc *p;
-+ struct qbman_alt_fq_state_rslt *r;
-+
-+ /* Start the management command */
-+ p = qbman_swp_mc_start(s);
-+ if (!p)
-+ return -EBUSY;
-+
-+ p->fqid = cpu_to_le32(fqid & ALT_FQ_FQID_MASK);
-+
-+ /* Complete the management command */
-+ r = qbman_swp_mc_complete(s, p, alt_fq_verb);
-+ if (unlikely(!r)) {
-+ pr_err("qbman: mgmt cmd failed, no response (verb=0x%x)\n",
-+ alt_fq_verb);
-+ return -EIO;
-+ }
-+
-+ /* Decode the outcome */
-+ WARN_ON((r->verb & QBMAN_RESULT_MASK) != alt_fq_verb);
-+
-+ /* Determine success or failure */
-+ if (unlikely(r->rslt != QBMAN_MC_RSLT_OK)) {
-+ pr_err("qbman: ALT FQID %d failed: verb = 0x%08x code = 0x%02x\n",
-+ fqid, r->verb, r->rslt);
-+ return -EIO;
-+ }
-+
-+ return 0;
-+}
-+
-+struct qbman_cdan_ctrl_desc {
-+ u8 verb;
-+ u8 reserved;
-+ __le16 ch;
-+ u8 we;
-+ u8 ctrl;
-+ __le16 reserved2;
-+ __le64 cdan_ctx;
-+ u8 reserved3[48];
-+
-+};
-+
-+struct qbman_cdan_ctrl_rslt {
-+ u8 verb;
-+ u8 rslt;
-+ __le16 ch;
-+ u8 reserved[60];
-+};
-+
-+int qbman_swp_CDAN_set(struct qbman_swp *s, u16 channelid,
-+ u8 we_mask, u8 cdan_en,
-+ u64 ctx)
-+{
-+ struct qbman_cdan_ctrl_desc *p = NULL;
-+ struct qbman_cdan_ctrl_rslt *r = NULL;
-+
-+ /* Start the management command */
-+ p = qbman_swp_mc_start(s);
-+ if (!p)
-+ return -EBUSY;
-+
-+ /* Encode the caller-provided attributes */
-+ p->ch = cpu_to_le16(channelid);
-+ p->we = we_mask;
-+ if (cdan_en)
-+ p->ctrl = 1;
-+ else
-+ p->ctrl = 0;
-+ p->cdan_ctx = cpu_to_le64(ctx);
-+
-+ /* Complete the management command */
-+ r = qbman_swp_mc_complete(s, p, QBMAN_WQCHAN_CONFIGURE);
-+ if (unlikely(!r)) {
-+ pr_err("qbman: wqchan config failed, no response\n");
-+ return -EIO;
-+ }
-+
-+ WARN_ON((r->verb & 0x7f) != QBMAN_WQCHAN_CONFIGURE);
-+
-+ /* Determine success or failure */
-+ if (unlikely(r->rslt != QBMAN_MC_RSLT_OK)) {
-+ pr_err("qbman: CDAN cQID %d failed: code = 0x%02x\n",
-+ channelid, r->rslt);
-+ return -EIO;
-+ }
-+
-+ return 0;
-+}
-+
-+#define QBMAN_RESPONSE_VERB_MASK 0x7f
-+#define QBMAN_FQ_QUERY_NP 0x45
-+#define QBMAN_BP_QUERY 0x32
-+
-+struct qbman_fq_query_desc {
-+ u8 verb;
-+ u8 reserved[3];
-+ u32 fqid;
-+ u8 reserved2[56];
-+};
-+
-+int qbman_fq_query_state(struct qbman_swp *s, u32 fqid,
-+ struct qbman_fq_query_np_rslt *r)
-+{
-+ struct qbman_fq_query_desc *p;
-+ void *resp;
-+
-+ p = (struct qbman_fq_query_desc *)qbman_swp_mc_start(s);
-+ if (!p)
-+ return -EBUSY;
-+
-+ /* FQID is a 24 bit value */
-+ p->fqid = cpu_to_le32(fqid) & 0x00FFFFFF;
-+ resp = qbman_swp_mc_complete(s, p, QBMAN_FQ_QUERY_NP);
-+ if (!resp) {
-+ pr_err("qbman: Query FQID %d NP fields failed, no response\n",
-+ fqid);
-+ return -EIO;
-+ }
-+ *r = *(struct qbman_fq_query_np_rslt *)resp;
-+ /* Decode the outcome */
-+ WARN_ON((r->verb & QBMAN_RESPONSE_VERB_MASK) != QBMAN_FQ_QUERY_NP);
-+
-+ /* Determine success or failure */
-+ if (r->rslt != QBMAN_MC_RSLT_OK) {
-+ pr_err("Query NP fields of FQID 0x%x failed, code=0x%02x\n",
-+ p->fqid, r->rslt);
-+ return -EIO;
-+ }
-+
-+ return 0;
-+}
-+
-+u32 qbman_fq_state_frame_count(const struct qbman_fq_query_np_rslt *r)
-+{
-+ return (r->frm_cnt & 0x00FFFFFF);
-+}
-+
-+u32 qbman_fq_state_byte_count(const struct qbman_fq_query_np_rslt *r)
-+{
-+ return r->byte_cnt;
-+}
-+
-+struct qbman_bp_query_desc {
-+ u8 verb;
-+ u8 reserved;
-+ u16 bpid;
-+ u8 reserved2[60];
-+};
-+
-+int qbman_bp_query(struct qbman_swp *s, u32 bpid,
-+ struct qbman_bp_query_rslt *r)
-+{
-+ struct qbman_bp_query_desc *p;
-+ void *resp;
-+
-+ p = (struct qbman_bp_query_desc *)qbman_swp_mc_start(s);
-+ if (!p)
-+ return -EBUSY;
-+
-+ p->bpid = bpid;
-+ resp = qbman_swp_mc_complete(s, p, QBMAN_BP_QUERY);
-+ if (!resp) {
-+ pr_err("qbman: Query BPID %d fields failed, no response\n",
-+ bpid);
-+ return -EIO;
-+ }
-+ *r = *(struct qbman_bp_query_rslt *)resp;
-+ /* Decode the outcome */
-+ WARN_ON((r->verb & QBMAN_RESPONSE_VERB_MASK) != QBMAN_BP_QUERY);
-+
-+ /* Determine success or failure */
-+ if (r->rslt != QBMAN_MC_RSLT_OK) {
-+ pr_err("Query fields of BPID 0x%x failed, code=0x%02x\n",
-+ bpid, r->rslt);
-+ return -EIO;
-+ }
-+
-+ return 0;
-+}
-+
-+u32 qbman_bp_info_num_free_bufs(struct qbman_bp_query_rslt *a)
-+{
-+ return a->fill;
-+}
---- /dev/null
-+++ b/drivers/staging/fsl-mc/bus/dpio/qbman-portal.h
-@@ -0,0 +1,505 @@
-+/* SPDX-License-Identifier: (GPL-2.0+ OR BSD-3-Clause) */
-+/*
-+ * Copyright (C) 2014-2016 Freescale Semiconductor, Inc.
-+ * Copyright 2016 NXP
-+ *
-+ */
-+#ifndef __FSL_QBMAN_PORTAL_H
-+#define __FSL_QBMAN_PORTAL_H
-+
-+#include "../../include/dpaa2-fd.h"
-+
-+struct dpaa2_dq;
-+struct qbman_swp;
-+
-+/* qbman software portal descriptor structure */
-+struct qbman_swp_desc {
-+ void *cena_bar; /* Cache-enabled portal base address */
-+ void *cinh_bar; /* Cache-inhibited portal base address */
-+ u32 qman_version;
-+};
-+
-+#define QBMAN_SWP_INTERRUPT_EQRI 0x01
-+#define QBMAN_SWP_INTERRUPT_EQDI 0x02
-+#define QBMAN_SWP_INTERRUPT_DQRI 0x04
-+#define QBMAN_SWP_INTERRUPT_RCRI 0x08
-+#define QBMAN_SWP_INTERRUPT_RCDI 0x10
-+#define QBMAN_SWP_INTERRUPT_VDCI 0x20
-+
-+/* the structure for pull dequeue descriptor */
-+struct qbman_pull_desc {
-+ u8 verb;
-+ u8 numf;
-+ u8 tok;
-+ u8 reserved;
-+ __le32 dq_src;
-+ __le64 rsp_addr;
-+ u64 rsp_addr_virt;
-+ u8 padding[40];
-+};
-+
-+enum qbman_pull_type_e {
-+ /* dequeue with priority precedence, respect intra-class scheduling */
-+ qbman_pull_type_prio = 1,
-+ /* dequeue with active FQ precedence, respect ICS */
-+ qbman_pull_type_active,
-+ /* dequeue with active FQ precedence, no ICS */
-+ qbman_pull_type_active_noics
-+};
-+
-+/* Definitions for parsing dequeue entries */
-+#define QBMAN_RESULT_MASK 0x7f
-+#define QBMAN_RESULT_DQ 0x60
-+#define QBMAN_RESULT_FQRN 0x21
-+#define QBMAN_RESULT_FQRNI 0x22
-+#define QBMAN_RESULT_FQPN 0x24
-+#define QBMAN_RESULT_FQDAN 0x25
-+#define QBMAN_RESULT_CDAN 0x26
-+#define QBMAN_RESULT_CSCN_MEM 0x27
-+#define QBMAN_RESULT_CGCU 0x28
-+#define QBMAN_RESULT_BPSCN 0x29
-+#define QBMAN_RESULT_CSCN_WQ 0x2a
-+
-+/* QBMan FQ management command codes */
-+#define QBMAN_FQ_SCHEDULE 0x48
-+#define QBMAN_FQ_FORCE 0x49
-+#define QBMAN_FQ_XON 0x4d
-+#define QBMAN_FQ_XOFF 0x4e
-+
-+/* structure of enqueue descriptor */
-+struct qbman_eq_desc {
-+ u8 verb;
-+ u8 dca;
-+ __le16 seqnum;
-+ __le16 orpid;
-+ __le16 reserved1;
-+ __le32 tgtid;
-+ __le32 tag;
-+ __le16 qdbin;
-+ u8 qpri;
-+ u8 reserved[3];
-+ u8 wae;
-+ u8 rspid;
-+ __le64 rsp_addr;
-+ u8 fd[32];
-+};
-+
-+/* buffer release descriptor */
-+struct qbman_release_desc {
-+ u8 verb;
-+ u8 reserved;
-+ __le16 bpid;
-+ __le32 reserved2;
-+ __le64 buf[7];
-+};
-+
-+/* Management command result codes */
-+#define QBMAN_MC_RSLT_OK 0xf0
-+
-+#define CODE_CDAN_WE_EN 0x1
-+#define CODE_CDAN_WE_CTX 0x4
-+
-+/* portal data structure */
-+struct qbman_swp {
-+ const struct qbman_swp_desc *desc;
-+ void __iomem *addr_cena;
-+ void __iomem *addr_cinh;
-+
-+ /* Management commands */
-+ struct {
-+ u32 valid_bit; /* 0x00 or 0x80 */
-+ } mc;
-+
-+ /* Push dequeues */
-+ u32 sdq;
-+
-+ /* Volatile dequeues */
-+ struct {
-+ atomic_t available; /* indicates if a command can be sent */
-+ u32 valid_bit; /* 0x00 or 0x80 */
-+ struct dpaa2_dq *storage; /* NULL if DQRR */
-+ } vdq;
-+
-+ /* DQRR */
-+ struct {
-+ u32 next_idx;
-+ u32 valid_bit;
-+ u8 dqrr_size;
-+ int reset_bug; /* indicates dqrr reset workaround is needed */
-+ } dqrr;
-+};
-+
-+struct qbman_swp *qbman_swp_init(const struct qbman_swp_desc *d);
-+void qbman_swp_finish(struct qbman_swp *p);
-+u32 qbman_swp_interrupt_read_status(struct qbman_swp *p);
-+void qbman_swp_interrupt_clear_status(struct qbman_swp *p, u32 mask);
-+u32 qbman_swp_interrupt_get_trigger(struct qbman_swp *p);
-+void qbman_swp_interrupt_set_trigger(struct qbman_swp *p, u32 mask);
-+int qbman_swp_interrupt_get_inhibit(struct qbman_swp *p);
-+void qbman_swp_interrupt_set_inhibit(struct qbman_swp *p, int inhibit);
-+
-+void qbman_swp_push_get(struct qbman_swp *p, u8 channel_idx, int *enabled);
-+void qbman_swp_push_set(struct qbman_swp *p, u8 channel_idx, int enable);
-+
-+void qbman_pull_desc_clear(struct qbman_pull_desc *d);
-+void qbman_pull_desc_set_storage(struct qbman_pull_desc *d,
-+ struct dpaa2_dq *storage,
-+ dma_addr_t storage_phys,
-+ int stash);
-+void qbman_pull_desc_set_numframes(struct qbman_pull_desc *d, u8 numframes);
-+void qbman_pull_desc_set_fq(struct qbman_pull_desc *d, u32 fqid);
-+void qbman_pull_desc_set_wq(struct qbman_pull_desc *d, u32 wqid,
-+ enum qbman_pull_type_e dct);
-+void qbman_pull_desc_set_channel(struct qbman_pull_desc *d, u32 chid,
-+ enum qbman_pull_type_e dct);
-+
-+int qbman_swp_pull(struct qbman_swp *p, struct qbman_pull_desc *d);
-+
-+const struct dpaa2_dq *qbman_swp_dqrr_next(struct qbman_swp *s);
-+void qbman_swp_dqrr_consume(struct qbman_swp *s, const struct dpaa2_dq *dq);
-+
-+int qbman_result_has_new_result(struct qbman_swp *p, const struct dpaa2_dq *dq);
-+
-+void qbman_eq_desc_clear(struct qbman_eq_desc *d);
-+void qbman_eq_desc_set_no_orp(struct qbman_eq_desc *d, int respond_success);
-+void qbman_eq_desc_set_orp(struct qbman_eq_desc *d, int respond_success,
-+ u16 oprid, u16 seqnum, int incomplete);
-+void qbman_eq_desc_set_orp_hole(struct qbman_eq_desc *d, u16 oprid, u16 seqnum);
-+void qbman_eq_desc_set_token(struct qbman_eq_desc *d, u8 token);
-+void qbman_eq_desc_set_fq(struct qbman_eq_desc *d, u32 fqid);
-+void qbman_eq_desc_set_qd(struct qbman_eq_desc *d, u32 qdid,
-+ u32 qd_bin, u32 qd_prio);
-+
-+int qbman_swp_enqueue(struct qbman_swp *p, const struct qbman_eq_desc *d,
-+ const struct dpaa2_fd *fd);
-+
-+void qbman_release_desc_clear(struct qbman_release_desc *d);
-+void qbman_release_desc_set_bpid(struct qbman_release_desc *d, u16 bpid);
-+void qbman_release_desc_set_rcdi(struct qbman_release_desc *d, int enable);
-+
-+int qbman_swp_release(struct qbman_swp *s, const struct qbman_release_desc *d,
-+ const u64 *buffers, unsigned int num_buffers);
-+int qbman_swp_acquire(struct qbman_swp *s, u16 bpid, u64 *buffers,
-+ unsigned int num_buffers);
-+int qbman_swp_alt_fq_state(struct qbman_swp *s, u32 fqid,
-+ u8 alt_fq_verb);
-+int qbman_swp_CDAN_set(struct qbman_swp *s, u16 channelid,
-+ u8 we_mask, u8 cdan_en,
-+ u64 ctx);
-+
-+void *qbman_swp_mc_start(struct qbman_swp *p);
-+void qbman_swp_mc_submit(struct qbman_swp *p, void *cmd, u8 cmd_verb);
-+void *qbman_swp_mc_result(struct qbman_swp *p);
-+
-+/**
-+ * qbman_result_is_DQ() - check if the dequeue result is a dequeue response
-+ * @dq: the dequeue result to be checked
-+ *
-+ * DQRR entries may contain non-dequeue results, ie. notifications
-+ */
-+static inline int qbman_result_is_DQ(const struct dpaa2_dq *dq)
-+{
-+ return ((dq->dq.verb & QBMAN_RESULT_MASK) == QBMAN_RESULT_DQ);
-+}
-+
-+/**
-+ * qbman_result_is_SCN() - Check the dequeue result is notification or not
-+ * @dq: the dequeue result to be checked
-+ *
-+ */
-+static inline int qbman_result_is_SCN(const struct dpaa2_dq *dq)
-+{
-+ return !qbman_result_is_DQ(dq);
-+}
-+
-+/* FQ Data Availability */
-+static inline int qbman_result_is_FQDAN(const struct dpaa2_dq *dq)
-+{
-+ return ((dq->dq.verb & QBMAN_RESULT_MASK) == QBMAN_RESULT_FQDAN);
-+}
-+
-+/* Channel Data Availability */
-+static inline int qbman_result_is_CDAN(const struct dpaa2_dq *dq)
-+{
-+ return ((dq->dq.verb & QBMAN_RESULT_MASK) == QBMAN_RESULT_CDAN);
-+}
-+
-+/* Congestion State Change */
-+static inline int qbman_result_is_CSCN(const struct dpaa2_dq *dq)
-+{
-+ return ((dq->dq.verb & QBMAN_RESULT_MASK) == QBMAN_RESULT_CSCN_WQ);
-+}
-+
-+/* Buffer Pool State Change */
-+static inline int qbman_result_is_BPSCN(const struct dpaa2_dq *dq)
-+{
-+ return ((dq->dq.verb & QBMAN_RESULT_MASK) == QBMAN_RESULT_BPSCN);
-+}
-+
-+/* Congestion Group Count Update */
-+static inline int qbman_result_is_CGCU(const struct dpaa2_dq *dq)
-+{
-+ return ((dq->dq.verb & QBMAN_RESULT_MASK) == QBMAN_RESULT_CGCU);
-+}
-+
-+/* Retirement */
-+static inline int qbman_result_is_FQRN(const struct dpaa2_dq *dq)
-+{
-+ return ((dq->dq.verb & QBMAN_RESULT_MASK) == QBMAN_RESULT_FQRN);
-+}
-+
-+/* Retirement Immediate */
-+static inline int qbman_result_is_FQRNI(const struct dpaa2_dq *dq)
-+{
-+ return ((dq->dq.verb & QBMAN_RESULT_MASK) == QBMAN_RESULT_FQRNI);
-+}
-+
-+ /* Park */
-+static inline int qbman_result_is_FQPN(const struct dpaa2_dq *dq)
-+{
-+ return ((dq->dq.verb & QBMAN_RESULT_MASK) == QBMAN_RESULT_FQPN);
-+}
-+
-+/**
-+ * qbman_result_SCN_state() - Get the state field in State-change notification
-+ */
-+static inline u8 qbman_result_SCN_state(const struct dpaa2_dq *scn)
-+{
-+ return scn->scn.state;
-+}
-+
-+#define SCN_RID_MASK 0x00FFFFFF
-+
-+/**
-+ * qbman_result_SCN_rid() - Get the resource id in State-change notification
-+ */
-+static inline u32 qbman_result_SCN_rid(const struct dpaa2_dq *scn)
-+{
-+ return le32_to_cpu(scn->scn.rid_tok) & SCN_RID_MASK;
-+}
-+
-+/**
-+ * qbman_result_SCN_ctx() - Get the context data in State-change notification
-+ */
-+static inline u64 qbman_result_SCN_ctx(const struct dpaa2_dq *scn)
-+{
-+ return le64_to_cpu(scn->scn.ctx);
-+}
-+
-+/**
-+ * qbman_swp_fq_schedule() - Move the fq to the scheduled state
-+ * @s: the software portal object
-+ * @fqid: the index of frame queue to be scheduled
-+ *
-+ * There are a couple of different ways that a FQ can end up parked state,
-+ * This schedules it.
-+ *
-+ * Return 0 for success, or negative error code for failure.
-+ */
-+static inline int qbman_swp_fq_schedule(struct qbman_swp *s, u32 fqid)
-+{
-+ return qbman_swp_alt_fq_state(s, fqid, QBMAN_FQ_SCHEDULE);
-+}
-+
-+/**
-+ * qbman_swp_fq_force() - Force the FQ to fully scheduled state
-+ * @s: the software portal object
-+ * @fqid: the index of frame queue to be forced
-+ *
-+ * Force eligible will force a tentatively-scheduled FQ to be fully-scheduled
-+ * and thus be available for selection by any channel-dequeuing behaviour (push
-+ * or pull). If the FQ is subsequently "dequeued" from the channel and is still
-+ * empty at the time this happens, the resulting dq_entry will have no FD.
-+ * (qbman_result_DQ_fd() will return NULL.)
-+ *
-+ * Return 0 for success, or negative error code for failure.
-+ */
-+static inline int qbman_swp_fq_force(struct qbman_swp *s, u32 fqid)
-+{
-+ return qbman_swp_alt_fq_state(s, fqid, QBMAN_FQ_FORCE);
-+}
-+
-+/**
-+ * qbman_swp_fq_xon() - sets FQ flow-control to XON
-+ * @s: the software portal object
-+ * @fqid: the index of frame queue
-+ *
-+ * This setting doesn't affect enqueues to the FQ, just dequeues.
-+ *
-+ * Return 0 for success, or negative error code for failure.
-+ */
-+static inline int qbman_swp_fq_xon(struct qbman_swp *s, u32 fqid)
-+{
-+ return qbman_swp_alt_fq_state(s, fqid, QBMAN_FQ_XON);
-+}
-+
-+/**
-+ * qbman_swp_fq_xoff() - sets FQ flow-control to XOFF
-+ * @s: the software portal object
-+ * @fqid: the index of frame queue
-+ *
-+ * This setting doesn't affect enqueues to the FQ, just dequeues.
-+ * XOFF FQs will remain in the tenatively-scheduled state, even when
-+ * non-empty, meaning they won't be selected for scheduled dequeuing.
-+ * If a FQ is changed to XOFF after it had already become truly-scheduled
-+ * to a channel, and a pull dequeue of that channel occurs that selects
-+ * that FQ for dequeuing, then the resulting dq_entry will have no FD.
-+ * (qbman_result_DQ_fd() will return NULL.)
-+ *
-+ * Return 0 for success, or negative error code for failure.
-+ */
-+static inline int qbman_swp_fq_xoff(struct qbman_swp *s, u32 fqid)
-+{
-+ return qbman_swp_alt_fq_state(s, fqid, QBMAN_FQ_XOFF);
-+}
-+
-+/* If the user has been allocated a channel object that is going to generate
-+ * CDANs to another channel, then the qbman_swp_CDAN* functions will be
-+ * necessary.
-+ *
-+ * CDAN-enabled channels only generate a single CDAN notification, after which
-+ * they need to be reenabled before they'll generate another. The idea is
-+ * that pull dequeuing will occur in reaction to the CDAN, followed by a
-+ * reenable step. Each function generates a distinct command to hardware, so a
-+ * combination function is provided if the user wishes to modify the "context"
-+ * (which shows up in each CDAN message) each time they reenable, as a single
-+ * command to hardware.
-+ */
-+
-+/**
-+ * qbman_swp_CDAN_set_context() - Set CDAN context
-+ * @s: the software portal object
-+ * @channelid: the channel index
-+ * @ctx: the context to be set in CDAN
-+ *
-+ * Return 0 for success, or negative error code for failure.
-+ */
-+static inline int qbman_swp_CDAN_set_context(struct qbman_swp *s, u16 channelid,
-+ u64 ctx)
-+{
-+ return qbman_swp_CDAN_set(s, channelid,
-+ CODE_CDAN_WE_CTX,
-+ 0, ctx);
-+}
-+
-+/**
-+ * qbman_swp_CDAN_enable() - Enable CDAN for the channel
-+ * @s: the software portal object
-+ * @channelid: the index of the channel to generate CDAN
-+ *
-+ * Return 0 for success, or negative error code for failure.
-+ */
-+static inline int qbman_swp_CDAN_enable(struct qbman_swp *s, u16 channelid)
-+{
-+ return qbman_swp_CDAN_set(s, channelid,
-+ CODE_CDAN_WE_EN,
-+ 1, 0);
-+}
-+
-+/**
-+ * qbman_swp_CDAN_disable() - disable CDAN for the channel
-+ * @s: the software portal object
-+ * @channelid: the index of the channel to generate CDAN
-+ *
-+ * Return 0 for success, or negative error code for failure.
-+ */
-+static inline int qbman_swp_CDAN_disable(struct qbman_swp *s, u16 channelid)
-+{
-+ return qbman_swp_CDAN_set(s, channelid,
-+ CODE_CDAN_WE_EN,
-+ 0, 0);
-+}
-+
-+/**
-+ * qbman_swp_CDAN_set_context_enable() - Set CDAN contest and enable CDAN
-+ * @s: the software portal object
-+ * @channelid: the index of the channel to generate CDAN
-+ * @ctx:i the context set in CDAN
-+ *
-+ * Return 0 for success, or negative error code for failure.
-+ */
-+static inline int qbman_swp_CDAN_set_context_enable(struct qbman_swp *s,
-+ u16 channelid,
-+ u64 ctx)
-+{
-+ return qbman_swp_CDAN_set(s, channelid,
-+ CODE_CDAN_WE_EN | CODE_CDAN_WE_CTX,
-+ 1, ctx);
-+}
-+
-+/* Wraps up submit + poll-for-result */
-+static inline void *qbman_swp_mc_complete(struct qbman_swp *swp, void *cmd,
-+ u8 cmd_verb)
-+{
-+ int loopvar = 1000;
-+
-+ qbman_swp_mc_submit(swp, cmd, cmd_verb);
-+
-+ do {
-+ cmd = qbman_swp_mc_result(swp);
-+ } while (!cmd && loopvar--);
-+
-+ WARN_ON(!loopvar);
-+
-+ return cmd;
-+}
-+
-+/* Query APIs */
-+struct qbman_fq_query_np_rslt {
-+ u8 verb;
-+ u8 rslt;
-+ u8 st1;
-+ u8 st2;
-+ u8 reserved[2];
-+ u16 od1_sfdr;
-+ u16 od2_sfdr;
-+ u16 od3_sfdr;
-+ u16 ra1_sfdr;
-+ u16 ra2_sfdr;
-+ u32 pfdr_hptr;
-+ u32 pfdr_tptr;
-+ u32 frm_cnt;
-+ u32 byte_cnt;
-+ u16 ics_surp;
-+ u8 is;
-+ u8 reserved2[29];
-+};
-+
-+int qbman_fq_query_state(struct qbman_swp *s, u32 fqid,
-+ struct qbman_fq_query_np_rslt *r);
-+u32 qbman_fq_state_frame_count(const struct qbman_fq_query_np_rslt *r);
-+u32 qbman_fq_state_byte_count(const struct qbman_fq_query_np_rslt *r);
-+
-+struct qbman_bp_query_rslt {
-+ u8 verb;
-+ u8 rslt;
-+ u8 reserved[4];
-+ u8 bdi;
-+ u8 state;
-+ u32 fill;
-+ u32 hdotr;
-+ u16 swdet;
-+ u16 swdxt;
-+ u16 hwdet;
-+ u16 hwdxt;
-+ u16 swset;
-+ u16 swsxt;
-+ u16 vbpid;
-+ u16 icid;
-+ u64 bpscn_addr;
-+ u64 bpscn_ctx;
-+ u16 hw_targ;
-+ u8 dbe;
-+ u8 reserved2;
-+ u8 sdcnt;
-+ u8 hdcnt;
-+ u8 sscnt;
-+ u8 reserved3[9];
-+};
-+
-+int qbman_bp_query(struct qbman_swp *s, u32 bpid,
-+ struct qbman_bp_query_rslt *r);
-+
-+u32 qbman_bp_info_num_free_bufs(struct qbman_bp_query_rslt *a);
-+
-+#endif /* __FSL_QBMAN_PORTAL_H */
---- a/drivers/staging/fsl-mc/bus/dpmcp-cmd.h
-+++ /dev/null
-@@ -1,140 +0,0 @@
--/* Copyright 2013-2016 Freescale Semiconductor Inc.
-- *
-- * Redistribution and use in source and binary forms, with or without
-- * modification, are permitted provided that the following conditions are met:
-- * * Redistributions of source code must retain the above copyright
-- * notice, this list of conditions and the following disclaimer.
-- * * Redistributions in binary form must reproduce the above copyright
-- * notice, this list of conditions and the following disclaimer in the
-- * documentation and/or other materials provided with the distribution.
-- * * Neither the name of the above-listed copyright holders nor the
-- * names of any contributors may be used to endorse or promote products
-- * derived from this software without specific prior written permission.
-- *
-- *
-- * ALTERNATIVELY, this software may be distributed under the terms of the
-- * GNU General Public License ("GPL") as published by the Free Software
-- * Foundation, either version 2 of that License or (at your option) any
-- * later version.
-- *
-- * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
-- * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
-- * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
-- * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE
-- * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
-- * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
-- * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
-- * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
-- * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
-- * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
-- * POSSIBILITY OF SUCH DAMAGE.
-- */
--#ifndef _FSL_DPMCP_CMD_H
--#define _FSL_DPMCP_CMD_H
--
--/* Minimal supported DPMCP Version */
--#define DPMCP_MIN_VER_MAJOR 3
--#define DPMCP_MIN_VER_MINOR 0
--
--/* Command IDs */
--#define DPMCP_CMDID_CLOSE 0x800
--#define DPMCP_CMDID_OPEN 0x80b
--#define DPMCP_CMDID_CREATE 0x90b
--#define DPMCP_CMDID_DESTROY 0x900
--
--#define DPMCP_CMDID_GET_ATTR 0x004
--#define DPMCP_CMDID_RESET 0x005
--
--#define DPMCP_CMDID_SET_IRQ 0x010
--#define DPMCP_CMDID_GET_IRQ 0x011
--#define DPMCP_CMDID_SET_IRQ_ENABLE 0x012
--#define DPMCP_CMDID_GET_IRQ_ENABLE 0x013
--#define DPMCP_CMDID_SET_IRQ_MASK 0x014
--#define DPMCP_CMDID_GET_IRQ_MASK 0x015
--#define DPMCP_CMDID_GET_IRQ_STATUS 0x016
--
--struct dpmcp_cmd_open {
-- __le32 dpmcp_id;
--};
--
--struct dpmcp_cmd_create {
-- __le32 portal_id;
--};
--
--struct dpmcp_cmd_set_irq {
-- /* cmd word 0 */
-- u8 irq_index;
-- u8 pad[3];
-- __le32 irq_val;
-- /* cmd word 1 */
-- __le64 irq_addr;
-- /* cmd word 2 */
-- __le32 irq_num;
--};
--
--struct dpmcp_cmd_get_irq {
-- __le32 pad;
-- u8 irq_index;
--};
--
--struct dpmcp_rsp_get_irq {
-- /* cmd word 0 */
-- __le32 irq_val;
-- __le32 pad;
-- /* cmd word 1 */
-- __le64 irq_paddr;
-- /* cmd word 2 */
-- __le32 irq_num;
-- __le32 type;
--};
--
--#define DPMCP_ENABLE 0x1
--
--struct dpmcp_cmd_set_irq_enable {
-- u8 enable;
-- u8 pad[3];
-- u8 irq_index;
--};
--
--struct dpmcp_cmd_get_irq_enable {
-- __le32 pad;
-- u8 irq_index;
--};
--
--struct dpmcp_rsp_get_irq_enable {
-- u8 enabled;
--};
--
--struct dpmcp_cmd_set_irq_mask {
-- __le32 mask;
-- u8 irq_index;
--};
--
--struct dpmcp_cmd_get_irq_mask {
-- __le32 pad;
-- u8 irq_index;
--};
--
--struct dpmcp_rsp_get_irq_mask {
-- __le32 mask;
--};
--
--struct dpmcp_cmd_get_irq_status {
-- __le32 status;
-- u8 irq_index;
--};
--
--struct dpmcp_rsp_get_irq_status {
-- __le32 status;
--};
--
--struct dpmcp_rsp_get_attributes {
-- /* response word 0 */
-- __le32 pad;
-- __le32 id;
-- /* response word 1 */
-- __le16 version_major;
-- __le16 version_minor;
--};
--
--#endif /* _FSL_DPMCP_CMD_H */
---- a/drivers/staging/fsl-mc/bus/dpmcp.c
-+++ /dev/null
-@@ -1,504 +0,0 @@
--/* Copyright 2013-2016 Freescale Semiconductor Inc.
-- *
-- * Redistribution and use in source and binary forms, with or without
-- * modification, are permitted provided that the following conditions are met:
-- * * Redistributions of source code must retain the above copyright
-- * notice, this list of conditions and the following disclaimer.
-- * * Redistributions in binary form must reproduce the above copyright
-- * notice, this list of conditions and the following disclaimer in the
-- * documentation and/or other materials provided with the distribution.
-- * * Neither the name of the above-listed copyright holders nor the
-- * names of any contributors may be used to endorse or promote products
-- * derived from this software without specific prior written permission.
-- *
-- *
-- * ALTERNATIVELY, this software may be distributed under the terms of the
-- * GNU General Public License ("GPL") as published by the Free Software
-- * Foundation, either version 2 of that License or (at your option) any
-- * later version.
-- *
-- * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
-- * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
-- * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
-- * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE
-- * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
-- * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
-- * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
-- * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
-- * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
-- * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
-- * POSSIBILITY OF SUCH DAMAGE.
-- */
--#include "../include/mc-sys.h"
--#include "../include/mc-cmd.h"
--
--#include "dpmcp.h"
--#include "dpmcp-cmd.h"
--
--/**
-- * dpmcp_open() - Open a control session for the specified object.
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @dpmcp_id: DPMCP unique ID
-- * @token: Returned token; use in subsequent API calls
-- *
-- * This function can be used to open a control session for an
-- * already created object; an object may have been declared in
-- * the DPL or by calling the dpmcp_create function.
-- * This function returns a unique authentication token,
-- * associated with the specific object ID and the specific MC
-- * portal; this token must be used in all subsequent commands for
-- * this specific object
-- *
-- * Return: '0' on Success; Error code otherwise.
-- */
--int dpmcp_open(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- int dpmcp_id,
-- u16 *token)
--{
-- struct mc_command cmd = { 0 };
-- struct dpmcp_cmd_open *cmd_params;
-- int err;
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPMCP_CMDID_OPEN,
-- cmd_flags, 0);
-- cmd_params = (struct dpmcp_cmd_open *)cmd.params;
-- cmd_params->dpmcp_id = cpu_to_le32(dpmcp_id);
--
-- /* send command to mc*/
-- err = mc_send_command(mc_io, &cmd);
-- if (err)
-- return err;
--
-- /* retrieve response parameters */
-- *token = mc_cmd_hdr_read_token(&cmd);
--
-- return err;
--}
--
--/**
-- * dpmcp_close() - Close the control session of the object
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @token: Token of DPMCP object
-- *
-- * After this function is called, no further operations are
-- * allowed on the object without opening a new control session.
-- *
-- * Return: '0' on Success; Error code otherwise.
-- */
--int dpmcp_close(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token)
--{
-- struct mc_command cmd = { 0 };
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPMCP_CMDID_CLOSE,
-- cmd_flags, token);
--
-- /* send command to mc*/
-- return mc_send_command(mc_io, &cmd);
--}
--
--/**
-- * dpmcp_create() - Create the DPMCP object.
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @cfg: Configuration structure
-- * @token: Returned token; use in subsequent API calls
-- *
-- * Create the DPMCP object, allocate required resources and
-- * perform required initialization.
-- *
-- * The object can be created either by declaring it in the
-- * DPL file, or by calling this function.
-- * This function returns a unique authentication token,
-- * associated with the specific object ID and the specific MC
-- * portal; this token must be used in all subsequent calls to
-- * this specific object. For objects that are created using the
-- * DPL file, call dpmcp_open function to get an authentication
-- * token first.
-- *
-- * Return: '0' on Success; Error code otherwise.
-- */
--int dpmcp_create(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- const struct dpmcp_cfg *cfg,
-- u16 *token)
--{
-- struct mc_command cmd = { 0 };
-- struct dpmcp_cmd_create *cmd_params;
--
-- int err;
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPMCP_CMDID_CREATE,
-- cmd_flags, 0);
-- cmd_params = (struct dpmcp_cmd_create *)cmd.params;
-- cmd_params->portal_id = cpu_to_le32(cfg->portal_id);
--
-- /* send command to mc*/
-- err = mc_send_command(mc_io, &cmd);
-- if (err)
-- return err;
--
-- /* retrieve response parameters */
-- *token = mc_cmd_hdr_read_token(&cmd);
--
-- return 0;
--}
--
--/**
-- * dpmcp_destroy() - Destroy the DPMCP object and release all its resources.
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @token: Token of DPMCP object
-- *
-- * Return: '0' on Success; error code otherwise.
-- */
--int dpmcp_destroy(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token)
--{
-- struct mc_command cmd = { 0 };
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPMCP_CMDID_DESTROY,
-- cmd_flags, token);
--
-- /* send command to mc*/
-- return mc_send_command(mc_io, &cmd);
--}
--
--/**
-- * dpmcp_reset() - Reset the DPMCP, returns the object to initial state.
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @token: Token of DPMCP object
-- *
-- * Return: '0' on Success; Error code otherwise.
-- */
--int dpmcp_reset(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token)
--{
-- struct mc_command cmd = { 0 };
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPMCP_CMDID_RESET,
-- cmd_flags, token);
--
-- /* send command to mc*/
-- return mc_send_command(mc_io, &cmd);
--}
--
--/**
-- * dpmcp_set_irq() - Set IRQ information for the DPMCP to trigger an interrupt.
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @token: Token of DPMCP object
-- * @irq_index: Identifies the interrupt index to configure
-- * @irq_cfg: IRQ configuration
-- *
-- * Return: '0' on Success; Error code otherwise.
-- */
--int dpmcp_set_irq(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- u8 irq_index,
-- struct dpmcp_irq_cfg *irq_cfg)
--{
-- struct mc_command cmd = { 0 };
-- struct dpmcp_cmd_set_irq *cmd_params;
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPMCP_CMDID_SET_IRQ,
-- cmd_flags, token);
-- cmd_params = (struct dpmcp_cmd_set_irq *)cmd.params;
-- cmd_params->irq_index = irq_index;
-- cmd_params->irq_val = cpu_to_le32(irq_cfg->val);
-- cmd_params->irq_addr = cpu_to_le64(irq_cfg->paddr);
-- cmd_params->irq_num = cpu_to_le32(irq_cfg->irq_num);
--
-- /* send command to mc*/
-- return mc_send_command(mc_io, &cmd);
--}
--
--/**
-- * dpmcp_get_irq() - Get IRQ information from the DPMCP.
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @token: Token of DPMCP object
-- * @irq_index: The interrupt index to configure
-- * @type: Interrupt type: 0 represents message interrupt
-- * type (both irq_addr and irq_val are valid)
-- * @irq_cfg: IRQ attributes
-- *
-- * Return: '0' on Success; Error code otherwise.
-- */
--int dpmcp_get_irq(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- u8 irq_index,
-- int *type,
-- struct dpmcp_irq_cfg *irq_cfg)
--{
-- struct mc_command cmd = { 0 };
-- struct dpmcp_cmd_get_irq *cmd_params;
-- struct dpmcp_rsp_get_irq *rsp_params;
-- int err;
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPMCP_CMDID_GET_IRQ,
-- cmd_flags, token);
-- cmd_params = (struct dpmcp_cmd_get_irq *)cmd.params;
-- cmd_params->irq_index = irq_index;
--
-- /* send command to mc*/
-- err = mc_send_command(mc_io, &cmd);
-- if (err)
-- return err;
--
-- /* retrieve response parameters */
-- rsp_params = (struct dpmcp_rsp_get_irq *)cmd.params;
-- irq_cfg->val = le32_to_cpu(rsp_params->irq_val);
-- irq_cfg->paddr = le64_to_cpu(rsp_params->irq_paddr);
-- irq_cfg->irq_num = le32_to_cpu(rsp_params->irq_num);
-- *type = le32_to_cpu(rsp_params->type);
-- return 0;
--}
--
--/**
-- * dpmcp_set_irq_enable() - Set overall interrupt state.
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @token: Token of DPMCP object
-- * @irq_index: The interrupt index to configure
-- * @en: Interrupt state - enable = 1, disable = 0
-- *
-- * Allows GPP software to control when interrupts are generated.
-- * Each interrupt can have up to 32 causes. The enable/disable control's the
-- * overall interrupt state. if the interrupt is disabled no causes will cause
-- * an interrupt.
-- *
-- * Return: '0' on Success; Error code otherwise.
-- */
--int dpmcp_set_irq_enable(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- u8 irq_index,
-- u8 en)
--{
-- struct mc_command cmd = { 0 };
-- struct dpmcp_cmd_set_irq_enable *cmd_params;
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPMCP_CMDID_SET_IRQ_ENABLE,
-- cmd_flags, token);
-- cmd_params = (struct dpmcp_cmd_set_irq_enable *)cmd.params;
-- cmd_params->enable = en & DPMCP_ENABLE;
-- cmd_params->irq_index = irq_index;
--
-- /* send command to mc*/
-- return mc_send_command(mc_io, &cmd);
--}
--
--/**
-- * dpmcp_get_irq_enable() - Get overall interrupt state
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @token: Token of DPMCP object
-- * @irq_index: The interrupt index to configure
-- * @en: Returned interrupt state - enable = 1, disable = 0
-- *
-- * Return: '0' on Success; Error code otherwise.
-- */
--int dpmcp_get_irq_enable(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- u8 irq_index,
-- u8 *en)
--{
-- struct mc_command cmd = { 0 };
-- struct dpmcp_cmd_get_irq_enable *cmd_params;
-- struct dpmcp_rsp_get_irq_enable *rsp_params;
-- int err;
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPMCP_CMDID_GET_IRQ_ENABLE,
-- cmd_flags, token);
-- cmd_params = (struct dpmcp_cmd_get_irq_enable *)cmd.params;
-- cmd_params->irq_index = irq_index;
--
-- /* send command to mc*/
-- err = mc_send_command(mc_io, &cmd);
-- if (err)
-- return err;
--
-- /* retrieve response parameters */
-- rsp_params = (struct dpmcp_rsp_get_irq_enable *)cmd.params;
-- *en = rsp_params->enabled & DPMCP_ENABLE;
-- return 0;
--}
--
--/**
-- * dpmcp_set_irq_mask() - Set interrupt mask.
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @token: Token of DPMCP object
-- * @irq_index: The interrupt index to configure
-- * @mask: Event mask to trigger interrupt;
-- * each bit:
-- * 0 = ignore event
-- * 1 = consider event for asserting IRQ
-- *
-- * Every interrupt can have up to 32 causes and the interrupt model supports
-- * masking/unmasking each cause independently
-- *
-- * Return: '0' on Success; Error code otherwise.
-- */
--int dpmcp_set_irq_mask(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- u8 irq_index,
-- u32 mask)
--{
-- struct mc_command cmd = { 0 };
-- struct dpmcp_cmd_set_irq_mask *cmd_params;
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPMCP_CMDID_SET_IRQ_MASK,
-- cmd_flags, token);
-- cmd_params = (struct dpmcp_cmd_set_irq_mask *)cmd.params;
-- cmd_params->mask = cpu_to_le32(mask);
-- cmd_params->irq_index = irq_index;
--
-- /* send command to mc*/
-- return mc_send_command(mc_io, &cmd);
--}
--
--/**
-- * dpmcp_get_irq_mask() - Get interrupt mask.
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @token: Token of DPMCP object
-- * @irq_index: The interrupt index to configure
-- * @mask: Returned event mask to trigger interrupt
-- *
-- * Every interrupt can have up to 32 causes and the interrupt model supports
-- * masking/unmasking each cause independently
-- *
-- * Return: '0' on Success; Error code otherwise.
-- */
--int dpmcp_get_irq_mask(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- u8 irq_index,
-- u32 *mask)
--{
-- struct mc_command cmd = { 0 };
-- struct dpmcp_cmd_get_irq_mask *cmd_params;
-- struct dpmcp_rsp_get_irq_mask *rsp_params;
--
-- int err;
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPMCP_CMDID_GET_IRQ_MASK,
-- cmd_flags, token);
-- cmd_params = (struct dpmcp_cmd_get_irq_mask *)cmd.params;
-- cmd_params->irq_index = irq_index;
--
-- /* send command to mc*/
-- err = mc_send_command(mc_io, &cmd);
-- if (err)
-- return err;
--
-- /* retrieve response parameters */
-- rsp_params = (struct dpmcp_rsp_get_irq_mask *)cmd.params;
-- *mask = le32_to_cpu(rsp_params->mask);
--
-- return 0;
--}
--
--/**
-- * dpmcp_get_irq_status() - Get the current status of any pending interrupts.
-- *
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @token: Token of DPMCP object
-- * @irq_index: The interrupt index to configure
-- * @status: Returned interrupts status - one bit per cause:
-- * 0 = no interrupt pending
-- * 1 = interrupt pending
-- *
-- * Return: '0' on Success; Error code otherwise.
-- */
--int dpmcp_get_irq_status(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- u8 irq_index,
-- u32 *status)
--{
-- struct mc_command cmd = { 0 };
-- struct dpmcp_cmd_get_irq_status *cmd_params;
-- struct dpmcp_rsp_get_irq_status *rsp_params;
-- int err;
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPMCP_CMDID_GET_IRQ_STATUS,
-- cmd_flags, token);
-- cmd_params = (struct dpmcp_cmd_get_irq_status *)cmd.params;
-- cmd_params->status = cpu_to_le32(*status);
-- cmd_params->irq_index = irq_index;
--
-- /* send command to mc*/
-- err = mc_send_command(mc_io, &cmd);
-- if (err)
-- return err;
--
-- /* retrieve response parameters */
-- rsp_params = (struct dpmcp_rsp_get_irq_status *)cmd.params;
-- *status = le32_to_cpu(rsp_params->status);
--
-- return 0;
--}
--
--/**
-- * dpmcp_get_attributes - Retrieve DPMCP attributes.
-- *
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @token: Token of DPMCP object
-- * @attr: Returned object's attributes
-- *
-- * Return: '0' on Success; Error code otherwise.
-- */
--int dpmcp_get_attributes(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- struct dpmcp_attr *attr)
--{
-- struct mc_command cmd = { 0 };
-- struct dpmcp_rsp_get_attributes *rsp_params;
-- int err;
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPMCP_CMDID_GET_ATTR,
-- cmd_flags, token);
--
-- /* send command to mc*/
-- err = mc_send_command(mc_io, &cmd);
-- if (err)
-- return err;
--
-- /* retrieve response parameters */
-- rsp_params = (struct dpmcp_rsp_get_attributes *)cmd.params;
-- attr->id = le32_to_cpu(rsp_params->id);
-- attr->version.major = le16_to_cpu(rsp_params->version_major);
-- attr->version.minor = le16_to_cpu(rsp_params->version_minor);
--
-- return 0;
--}
---- a/drivers/staging/fsl-mc/bus/dpmcp.h
-+++ /dev/null
-@@ -1,159 +0,0 @@
--/* Copyright 2013-2015 Freescale Semiconductor Inc.
-- *
-- * Redistribution and use in source and binary forms, with or without
-- * modification, are permitted provided that the following conditions are met:
-- * * Redistributions of source code must retain the above copyright
-- * notice, this list of conditions and the following disclaimer.
-- * * Redistributions in binary form must reproduce the above copyright
-- * notice, this list of conditions and the following disclaimer in the
-- * documentation and/or other materials provided with the distribution.
-- * * Neither the name of the above-listed copyright holders nor the
-- * names of any contributors may be used to endorse or promote products
-- * derived from this software without specific prior written permission.
-- *
-- *
-- * ALTERNATIVELY, this software may be distributed under the terms of the
-- * GNU General Public License ("GPL") as published by the Free Software
-- * Foundation, either version 2 of that License or (at your option) any
-- * later version.
-- *
-- * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
-- * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
-- * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
-- * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE
-- * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
-- * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
-- * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
-- * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
-- * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
-- * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
-- * POSSIBILITY OF SUCH DAMAGE.
-- */
--#ifndef __FSL_DPMCP_H
--#define __FSL_DPMCP_H
--
--/* Data Path Management Command Portal API
-- * Contains initialization APIs and runtime control APIs for DPMCP
-- */
--
--struct fsl_mc_io;
--
--int dpmcp_open(struct fsl_mc_io *mc_io,
-- uint32_t cmd_flags,
-- int dpmcp_id,
-- uint16_t *token);
--
--/* Get portal ID from pool */
--#define DPMCP_GET_PORTAL_ID_FROM_POOL (-1)
--
--int dpmcp_close(struct fsl_mc_io *mc_io,
-- uint32_t cmd_flags,
-- uint16_t token);
--
--/**
-- * struct dpmcp_cfg - Structure representing DPMCP configuration
-- * @portal_id: Portal ID; 'DPMCP_GET_PORTAL_ID_FROM_POOL' to get the portal ID
-- * from pool
-- */
--struct dpmcp_cfg {
-- int portal_id;
--};
--
--int dpmcp_create(struct fsl_mc_io *mc_io,
-- uint32_t cmd_flags,
-- const struct dpmcp_cfg *cfg,
-- uint16_t *token);
--
--int dpmcp_destroy(struct fsl_mc_io *mc_io,
-- uint32_t cmd_flags,
-- uint16_t token);
--
--int dpmcp_reset(struct fsl_mc_io *mc_io,
-- uint32_t cmd_flags,
-- uint16_t token);
--
--/* IRQ */
--/* IRQ Index */
--#define DPMCP_IRQ_INDEX 0
--/* irq event - Indicates that the link state changed */
--#define DPMCP_IRQ_EVENT_CMD_DONE 0x00000001
--
--/**
-- * struct dpmcp_irq_cfg - IRQ configuration
-- * @paddr: Address that must be written to signal a message-based interrupt
-- * @val: Value to write into irq_addr address
-- * @irq_num: A user defined number associated with this IRQ
-- */
--struct dpmcp_irq_cfg {
-- uint64_t paddr;
-- uint32_t val;
-- int irq_num;
--};
--
--int dpmcp_set_irq(struct fsl_mc_io *mc_io,
-- uint32_t cmd_flags,
-- uint16_t token,
-- uint8_t irq_index,
-- struct dpmcp_irq_cfg *irq_cfg);
--
--int dpmcp_get_irq(struct fsl_mc_io *mc_io,
-- uint32_t cmd_flags,
-- uint16_t token,
-- uint8_t irq_index,
-- int *type,
-- struct dpmcp_irq_cfg *irq_cfg);
--
--int dpmcp_set_irq_enable(struct fsl_mc_io *mc_io,
-- uint32_t cmd_flags,
-- uint16_t token,
-- uint8_t irq_index,
-- uint8_t en);
--
--int dpmcp_get_irq_enable(struct fsl_mc_io *mc_io,
-- uint32_t cmd_flags,
-- uint16_t token,
-- uint8_t irq_index,
-- uint8_t *en);
--
--int dpmcp_set_irq_mask(struct fsl_mc_io *mc_io,
-- uint32_t cmd_flags,
-- uint16_t token,
-- uint8_t irq_index,
-- uint32_t mask);
--
--int dpmcp_get_irq_mask(struct fsl_mc_io *mc_io,
-- uint32_t cmd_flags,
-- uint16_t token,
-- uint8_t irq_index,
-- uint32_t *mask);
--
--int dpmcp_get_irq_status(struct fsl_mc_io *mc_io,
-- uint32_t cmd_flags,
-- uint16_t token,
-- uint8_t irq_index,
-- uint32_t *status);
--
--/**
-- * struct dpmcp_attr - Structure representing DPMCP attributes
-- * @id: DPMCP object ID
-- * @version: DPMCP version
-- */
--struct dpmcp_attr {
-- int id;
-- /**
-- * struct version - Structure representing DPMCP version
-- * @major: DPMCP major version
-- * @minor: DPMCP minor version
-- */
-- struct {
-- uint16_t major;
-- uint16_t minor;
-- } version;
--};
--
--int dpmcp_get_attributes(struct fsl_mc_io *mc_io,
-- uint32_t cmd_flags,
-- uint16_t token,
-- struct dpmcp_attr *attr);
--
--#endif /* __FSL_DPMCP_H */
---- a/drivers/staging/fsl-mc/bus/dpmng-cmd.h
-+++ /dev/null
-@@ -1,58 +0,0 @@
--/*
-- * Copyright 2013-2016 Freescale Semiconductor Inc.
-- *
-- * Redistribution and use in source and binary forms, with or without
-- * modification, are permitted provided that the following conditions are met:
-- * * Redistributions of source code must retain the above copyright
-- * notice, this list of conditions and the following disclaimer.
-- * * Redistributions in binary form must reproduce the above copyright
-- * notice, this list of conditions and the following disclaimer in the
-- * documentation and/or other materials provided with the distribution.
-- * * Neither the name of the above-listed copyright holders nor the
-- * names of any contributors may be used to endorse or promote products
-- * derived from this software without specific prior written permission.
-- *
-- *
-- * ALTERNATIVELY, this software may be distributed under the terms of the
-- * GNU General Public License ("GPL") as published by the Free Software
-- * Foundation, either version 2 of that License or (at your option) any
-- * later version.
-- *
-- * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
-- * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
-- * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
-- * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE
-- * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
-- * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
-- * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
-- * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
-- * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
-- * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
-- * POSSIBILITY OF SUCH DAMAGE.
-- */
--
--/*
-- * dpmng-cmd.h
-- *
-- * defines portal commands
-- *
-- */
--
--#ifndef __FSL_DPMNG_CMD_H
--#define __FSL_DPMNG_CMD_H
--
--/* Command IDs */
--#define DPMNG_CMDID_GET_CONT_ID 0x830
--#define DPMNG_CMDID_GET_VERSION 0x831
--
--struct dpmng_rsp_get_container_id {
-- __le32 container_id;
--};
--
--struct dpmng_rsp_get_version {
-- __le32 revision;
-- __le32 version_major;
-- __le32 version_minor;
--};
--
--#endif /* __FSL_DPMNG_CMD_H */
---- a/drivers/staging/fsl-mc/bus/dpmng.c
-+++ /dev/null
-@@ -1,107 +0,0 @@
--/* Copyright 2013-2016 Freescale Semiconductor Inc.
-- *
-- * Redistribution and use in source and binary forms, with or without
-- * modification, are permitted provided that the following conditions are met:
-- * * Redistributions of source code must retain the above copyright
-- * notice, this list of conditions and the following disclaimer.
-- * * Redistributions in binary form must reproduce the above copyright
-- * notice, this list of conditions and the following disclaimer in the
-- * documentation and/or other materials provided with the distribution.
-- * * Neither the name of the above-listed copyright holders nor the
-- * names of any contributors may be used to endorse or promote products
-- * derived from this software without specific prior written permission.
-- *
-- *
-- * ALTERNATIVELY, this software may be distributed under the terms of the
-- * GNU General Public License ("GPL") as published by the Free Software
-- * Foundation, either version 2 of that License or (at your option) any
-- * later version.
-- *
-- * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
-- * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
-- * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
-- * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE
-- * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
-- * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
-- * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
-- * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
-- * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
-- * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
-- * POSSIBILITY OF SUCH DAMAGE.
-- */
--#include "../include/mc-sys.h"
--#include "../include/mc-cmd.h"
--#include "../include/dpmng.h"
--
--#include "dpmng-cmd.h"
--
--/**
-- * mc_get_version() - Retrieves the Management Complex firmware
-- * version information
-- * @mc_io: Pointer to opaque I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @mc_ver_info: Returned version information structure
-- *
-- * Return: '0' on Success; Error code otherwise.
-- */
--int mc_get_version(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- struct mc_version *mc_ver_info)
--{
-- struct mc_command cmd = { 0 };
-- struct dpmng_rsp_get_version *rsp_params;
-- int err;
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPMNG_CMDID_GET_VERSION,
-- cmd_flags,
-- 0);
--
-- /* send command to mc*/
-- err = mc_send_command(mc_io, &cmd);
-- if (err)
-- return err;
--
-- /* retrieve response parameters */
-- rsp_params = (struct dpmng_rsp_get_version *)cmd.params;
-- mc_ver_info->revision = le32_to_cpu(rsp_params->revision);
-- mc_ver_info->major = le32_to_cpu(rsp_params->version_major);
-- mc_ver_info->minor = le32_to_cpu(rsp_params->version_minor);
--
-- return 0;
--}
--EXPORT_SYMBOL(mc_get_version);
--
--/**
-- * dpmng_get_container_id() - Get container ID associated with a given portal.
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @container_id: Requested container ID
-- *
-- * Return: '0' on Success; Error code otherwise.
-- */
--int dpmng_get_container_id(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- int *container_id)
--{
-- struct mc_command cmd = { 0 };
-- struct dpmng_rsp_get_container_id *rsp_params;
-- int err;
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPMNG_CMDID_GET_CONT_ID,
-- cmd_flags,
-- 0);
--
-- /* send command to mc*/
-- err = mc_send_command(mc_io, &cmd);
-- if (err)
-- return err;
--
-- /* retrieve response parameters */
-- rsp_params = (struct dpmng_rsp_get_container_id *)cmd.params;
-- *container_id = le32_to_cpu(rsp_params->container_id);
--
-- return 0;
--}
--
---- a/drivers/staging/fsl-mc/bus/dprc-cmd.h
-+++ /dev/null
-@@ -1,465 +0,0 @@
--/*
-- * Copyright 2013-2016 Freescale Semiconductor Inc.
-- *
-- * Redistribution and use in source and binary forms, with or without
-- * modification, are permitted provided that the following conditions are met:
-- * * Redistributions of source code must retain the above copyright
-- * notice, this list of conditions and the following disclaimer.
-- * * Redistributions in binary form must reproduce the above copyright
-- * notice, this list of conditions and the following disclaimer in the
-- * documentation and/or other materials provided with the distribution.
-- * * Neither the name of the above-listed copyright holders nor the
-- * names of any contributors may be used to endorse or promote products
-- * derived from this software without specific prior written permission.
-- *
-- *
-- * ALTERNATIVELY, this software may be distributed under the terms of the
-- * GNU General Public License ("GPL") as published by the Free Software
-- * Foundation, either version 2 of that License or (at your option) any
-- * later version.
-- *
-- * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
-- * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
-- * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
-- * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE
-- * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
-- * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
-- * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
-- * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
-- * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
-- * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
-- * POSSIBILITY OF SUCH DAMAGE.
-- */
--
--/*
-- * dprc-cmd.h
-- *
-- * defines dprc portal commands
-- *
-- */
--
--#ifndef _FSL_DPRC_CMD_H
--#define _FSL_DPRC_CMD_H
--
--/* Minimal supported DPRC Version */
--#define DPRC_MIN_VER_MAJOR 5
--#define DPRC_MIN_VER_MINOR 0
--
--/* Command IDs */
--#define DPRC_CMDID_CLOSE 0x800
--#define DPRC_CMDID_OPEN 0x805
--#define DPRC_CMDID_CREATE 0x905
--
--#define DPRC_CMDID_GET_ATTR 0x004
--#define DPRC_CMDID_RESET_CONT 0x005
--
--#define DPRC_CMDID_SET_IRQ 0x010
--#define DPRC_CMDID_GET_IRQ 0x011
--#define DPRC_CMDID_SET_IRQ_ENABLE 0x012
--#define DPRC_CMDID_GET_IRQ_ENABLE 0x013
--#define DPRC_CMDID_SET_IRQ_MASK 0x014
--#define DPRC_CMDID_GET_IRQ_MASK 0x015
--#define DPRC_CMDID_GET_IRQ_STATUS 0x016
--#define DPRC_CMDID_CLEAR_IRQ_STATUS 0x017
--
--#define DPRC_CMDID_CREATE_CONT 0x151
--#define DPRC_CMDID_DESTROY_CONT 0x152
--#define DPRC_CMDID_SET_RES_QUOTA 0x155
--#define DPRC_CMDID_GET_RES_QUOTA 0x156
--#define DPRC_CMDID_ASSIGN 0x157
--#define DPRC_CMDID_UNASSIGN 0x158
--#define DPRC_CMDID_GET_OBJ_COUNT 0x159
--#define DPRC_CMDID_GET_OBJ 0x15A
--#define DPRC_CMDID_GET_RES_COUNT 0x15B
--#define DPRC_CMDID_GET_RES_IDS 0x15C
--#define DPRC_CMDID_GET_OBJ_REG 0x15E
--#define DPRC_CMDID_SET_OBJ_IRQ 0x15F
--#define DPRC_CMDID_GET_OBJ_IRQ 0x160
--#define DPRC_CMDID_SET_OBJ_LABEL 0x161
--#define DPRC_CMDID_GET_OBJ_DESC 0x162
--
--#define DPRC_CMDID_CONNECT 0x167
--#define DPRC_CMDID_DISCONNECT 0x168
--#define DPRC_CMDID_GET_POOL 0x169
--#define DPRC_CMDID_GET_POOL_COUNT 0x16A
--
--#define DPRC_CMDID_GET_CONNECTION 0x16C
--
--struct dprc_cmd_open {
-- __le32 container_id;
--};
--
--struct dprc_cmd_create_container {
-- /* cmd word 0 */
-- __le32 options;
-- __le16 icid;
-- __le16 pad0;
-- /* cmd word 1 */
-- __le32 pad1;
-- __le32 portal_id;
-- /* cmd words 2-3 */
-- u8 label[16];
--};
--
--struct dprc_rsp_create_container {
-- /* response word 0 */
-- __le64 pad0;
-- /* response word 1 */
-- __le32 child_container_id;
-- __le32 pad1;
-- /* response word 2 */
-- __le64 child_portal_addr;
--};
--
--struct dprc_cmd_destroy_container {
-- __le32 child_container_id;
--};
--
--struct dprc_cmd_reset_container {
-- __le32 child_container_id;
--};
--
--struct dprc_cmd_set_irq {
-- /* cmd word 0 */
-- __le32 irq_val;
-- u8 irq_index;
-- u8 pad[3];
-- /* cmd word 1 */
-- __le64 irq_addr;
-- /* cmd word 2 */
-- __le32 irq_num;
--};
--
--struct dprc_cmd_get_irq {
-- __le32 pad;
-- u8 irq_index;
--};
--
--struct dprc_rsp_get_irq {
-- /* response word 0 */
-- __le32 irq_val;
-- __le32 pad;
-- /* response word 1 */
-- __le64 irq_addr;
-- /* response word 2 */
-- __le32 irq_num;
-- __le32 type;
--};
--
--#define DPRC_ENABLE 0x1
--
--struct dprc_cmd_set_irq_enable {
-- u8 enable;
-- u8 pad[3];
-- u8 irq_index;
--};
--
--struct dprc_cmd_get_irq_enable {
-- __le32 pad;
-- u8 irq_index;
--};
--
--struct dprc_rsp_get_irq_enable {
-- u8 enabled;
--};
--
--struct dprc_cmd_set_irq_mask {
-- __le32 mask;
-- u8 irq_index;
--};
--
--struct dprc_cmd_get_irq_mask {
-- __le32 pad;
-- u8 irq_index;
--};
--
--struct dprc_rsp_get_irq_mask {
-- __le32 mask;
--};
--
--struct dprc_cmd_get_irq_status {
-- __le32 status;
-- u8 irq_index;
--};
--
--struct dprc_rsp_get_irq_status {
-- __le32 status;
--};
--
--struct dprc_cmd_clear_irq_status {
-- __le32 status;
-- u8 irq_index;
--};
--
--struct dprc_rsp_get_attributes {
-- /* response word 0 */
-- __le32 container_id;
-- __le16 icid;
-- __le16 pad;
-- /* response word 1 */
-- __le32 options;
-- __le32 portal_id;
-- /* response word 2 */
-- __le16 version_major;
-- __le16 version_minor;
--};
--
--struct dprc_cmd_set_res_quota {
-- /* cmd word 0 */
-- __le32 child_container_id;
-- __le16 quota;
-- __le16 pad;
-- /* cmd words 1-2 */
-- u8 type[16];
--};
--
--struct dprc_cmd_get_res_quota {
-- /* cmd word 0 */
-- __le32 child_container_id;
-- __le32 pad;
-- /* cmd word 1-2 */
-- u8 type[16];
--};
--
--struct dprc_rsp_get_res_quota {
-- __le32 pad;
-- __le16 quota;
--};
--
--struct dprc_cmd_assign {
-- /* cmd word 0 */
-- __le32 container_id;
-- __le32 options;
-- /* cmd word 1 */
-- __le32 num;
-- __le32 id_base_align;
-- /* cmd word 2-3 */
-- u8 type[16];
--};
--
--struct dprc_cmd_unassign {
-- /* cmd word 0 */
-- __le32 child_container_id;
-- __le32 options;
-- /* cmd word 1 */
-- __le32 num;
-- __le32 id_base_align;
-- /* cmd word 2-3 */
-- u8 type[16];
--};
--
--struct dprc_rsp_get_pool_count {
-- __le32 pool_count;
--};
--
--struct dprc_cmd_get_pool {
-- __le32 pool_index;
--};
--
--struct dprc_rsp_get_pool {
-- /* response word 0 */
-- __le64 pad;
-- /* response word 1-2 */
-- u8 type[16];
--};
--
--struct dprc_rsp_get_obj_count {
-- __le32 pad;
-- __le32 obj_count;
--};
--
--struct dprc_cmd_get_obj {
-- __le32 obj_index;
--};
--
--struct dprc_rsp_get_obj {
-- /* response word 0 */
-- __le32 pad0;
-- __le32 id;
-- /* response word 1 */
-- __le16 vendor;
-- u8 irq_count;
-- u8 region_count;
-- __le32 state;
-- /* response word 2 */
-- __le16 version_major;
-- __le16 version_minor;
-- __le16 flags;
-- __le16 pad1;
-- /* response word 3-4 */
-- u8 type[16];
-- /* response word 5-6 */
-- u8 label[16];
--};
--
--struct dprc_cmd_get_obj_desc {
-- /* cmd word 0 */
-- __le32 obj_id;
-- __le32 pad;
-- /* cmd word 1-2 */
-- u8 type[16];
--};
--
--struct dprc_rsp_get_obj_desc {
-- /* response word 0 */
-- __le32 pad0;
-- __le32 id;
-- /* response word 1 */
-- __le16 vendor;
-- u8 irq_count;
-- u8 region_count;
-- __le32 state;
-- /* response word 2 */
-- __le16 version_major;
-- __le16 version_minor;
-- __le16 flags;
-- __le16 pad1;
-- /* response word 3-4 */
-- u8 type[16];
-- /* response word 5-6 */
-- u8 label[16];
--};
--
--struct dprc_cmd_get_res_count {
-- /* cmd word 0 */
-- __le64 pad;
-- /* cmd word 1-2 */
-- u8 type[16];
--};
--
--struct dprc_rsp_get_res_count {
-- __le32 res_count;
--};
--
--struct dprc_cmd_get_res_ids {
-- /* cmd word 0 */
-- u8 pad0[5];
-- u8 iter_status;
-- __le16 pad1;
-- /* cmd word 1 */
-- __le32 base_id;
-- __le32 last_id;
-- /* cmd word 2-3 */
-- u8 type[16];
--};
--
--struct dprc_rsp_get_res_ids {
-- /* response word 0 */
-- u8 pad0[5];
-- u8 iter_status;
-- __le16 pad1;
-- /* response word 1 */
-- __le32 base_id;
-- __le32 last_id;
--};
--
--struct dprc_cmd_get_obj_region {
-- /* cmd word 0 */
-- __le32 obj_id;
-- __le16 pad0;
-- u8 region_index;
-- u8 pad1;
-- /* cmd word 1-2 */
-- __le64 pad2[2];
-- /* cmd word 3-4 */
-- u8 obj_type[16];
--};
--
--struct dprc_rsp_get_obj_region {
-- /* response word 0 */
-- __le64 pad;
-- /* response word 1 */
-- __le64 base_addr;
-- /* response word 2 */
-- __le32 size;
--};
--
--struct dprc_cmd_set_obj_label {
-- /* cmd word 0 */
-- __le32 obj_id;
-- __le32 pad;
-- /* cmd word 1-2 */
-- u8 label[16];
-- /* cmd word 3-4 */
-- u8 obj_type[16];
--};
--
--struct dprc_cmd_set_obj_irq {
-- /* cmd word 0 */
-- __le32 irq_val;
-- u8 irq_index;
-- u8 pad[3];
-- /* cmd word 1 */
-- __le64 irq_addr;
-- /* cmd word 2 */
-- __le32 irq_num;
-- __le32 obj_id;
-- /* cmd word 3-4 */
-- u8 obj_type[16];
--};
--
--struct dprc_cmd_get_obj_irq {
-- /* cmd word 0 */
-- __le32 obj_id;
-- u8 irq_index;
-- u8 pad[3];
-- /* cmd word 1-2 */
-- u8 obj_type[16];
--};
--
--struct dprc_rsp_get_obj_irq {
-- /* response word 0 */
-- __le32 irq_val;
-- __le32 pad;
-- /* response word 1 */
-- __le64 irq_addr;
-- /* response word 2 */
-- __le32 irq_num;
-- __le32 type;
--};
--
--struct dprc_cmd_connect {
-- /* cmd word 0 */
-- __le32 ep1_id;
-- __le32 ep1_interface_id;
-- /* cmd word 1 */
-- __le32 ep2_id;
-- __le32 ep2_interface_id;
-- /* cmd word 2-3 */
-- u8 ep1_type[16];
-- /* cmd word 4 */
-- __le32 max_rate;
-- __le32 committed_rate;
-- /* cmd word 5-6 */
-- u8 ep2_type[16];
--};
--
--struct dprc_cmd_disconnect {
-- /* cmd word 0 */
-- __le32 id;
-- __le32 interface_id;
-- /* cmd word 1-2 */
-- u8 type[16];
--};
--
--struct dprc_cmd_get_connection {
-- /* cmd word 0 */
-- __le32 ep1_id;
-- __le32 ep1_interface_id;
-- /* cmd word 1-2 */
-- u8 ep1_type[16];
--};
--
--struct dprc_rsp_get_connection {
-- /* response word 0-2 */
-- __le64 pad[3];
-- /* response word 3 */
-- __le32 ep2_id;
-- __le32 ep2_interface_id;
-- /* response word 4-5 */
-- u8 ep2_type[16];
-- /* response word 6 */
-- __le32 state;
--};
--
--#endif /* _FSL_DPRC_CMD_H */
---- a/drivers/staging/fsl-mc/bus/dprc.c
-+++ /dev/null
-@@ -1,1388 +0,0 @@
--/* Copyright 2013-2016 Freescale Semiconductor Inc.
-- *
-- * Redistribution and use in source and binary forms, with or without
-- * modification, are permitted provided that the following conditions are met:
-- * * Redistributions of source code must retain the above copyright
-- * notice, this list of conditions and the following disclaimer.
-- * * Redistributions in binary form must reproduce the above copyright
-- * notice, this list of conditions and the following disclaimer in the
-- * documentation and/or other materials provided with the distribution.
-- * * Neither the name of the above-listed copyright holders nor the
-- * names of any contributors may be used to endorse or promote products
-- * derived from this software without specific prior written permission.
-- *
-- *
-- * ALTERNATIVELY, this software may be distributed under the terms of the
-- * GNU General Public License ("GPL") as published by the Free Software
-- * Foundation, either version 2 of that License or (at your option) any
-- * later version.
-- *
-- * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
-- * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
-- * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
-- * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE
-- * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
-- * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
-- * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
-- * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
-- * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
-- * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
-- * POSSIBILITY OF SUCH DAMAGE.
-- */
--#include "../include/mc-sys.h"
--#include "../include/mc-cmd.h"
--#include "../include/dprc.h"
--
--#include "dprc-cmd.h"
--
--/**
-- * dprc_open() - Open DPRC object for use
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @container_id: Container ID to open
-- * @token: Returned token of DPRC object
-- *
-- * Return: '0' on Success; Error code otherwise.
-- *
-- * @warning Required before any operation on the object.
-- */
--int dprc_open(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- int container_id,
-- u16 *token)
--{
-- struct mc_command cmd = { 0 };
-- struct dprc_cmd_open *cmd_params;
-- int err;
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPRC_CMDID_OPEN, cmd_flags,
-- 0);
-- cmd_params = (struct dprc_cmd_open *)cmd.params;
-- cmd_params->container_id = cpu_to_le32(container_id);
--
-- /* send command to mc*/
-- err = mc_send_command(mc_io, &cmd);
-- if (err)
-- return err;
--
-- /* retrieve response parameters */
-- *token = mc_cmd_hdr_read_token(&cmd);
--
-- return 0;
--}
--EXPORT_SYMBOL(dprc_open);
--
--/**
-- * dprc_close() - Close the control session of the object
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @token: Token of DPRC object
-- *
-- * After this function is called, no further operations are
-- * allowed on the object without opening a new control session.
-- *
-- * Return: '0' on Success; Error code otherwise.
-- */
--int dprc_close(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token)
--{
-- struct mc_command cmd = { 0 };
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPRC_CMDID_CLOSE, cmd_flags,
-- token);
--
-- /* send command to mc*/
-- return mc_send_command(mc_io, &cmd);
--}
--EXPORT_SYMBOL(dprc_close);
--
--/**
-- * dprc_create_container() - Create child container
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @token: Token of DPRC object
-- * @cfg: Child container configuration
-- * @child_container_id: Returned child container ID
-- * @child_portal_offset: Returned child portal offset from MC portal base
-- *
-- * Return: '0' on Success; Error code otherwise.
-- */
--int dprc_create_container(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- struct dprc_cfg *cfg,
-- int *child_container_id,
-- u64 *child_portal_offset)
--{
-- struct mc_command cmd = { 0 };
-- struct dprc_cmd_create_container *cmd_params;
-- struct dprc_rsp_create_container *rsp_params;
-- int err;
--
-- /* prepare command */
-- cmd_params = (struct dprc_cmd_create_container *)cmd.params;
-- cmd_params->options = cpu_to_le32(cfg->options);
-- cmd_params->icid = cpu_to_le16(cfg->icid);
-- cmd_params->portal_id = cpu_to_le32(cfg->portal_id);
-- strncpy(cmd_params->label, cfg->label, 16);
-- cmd_params->label[15] = '\0';
--
-- cmd.header = mc_encode_cmd_header(DPRC_CMDID_CREATE_CONT,
-- cmd_flags, token);
--
-- /* send command to mc*/
-- err = mc_send_command(mc_io, &cmd);
-- if (err)
-- return err;
--
-- /* retrieve response parameters */
-- rsp_params = (struct dprc_rsp_create_container *)cmd.params;
-- *child_container_id = le32_to_cpu(rsp_params->child_container_id);
-- *child_portal_offset = le64_to_cpu(rsp_params->child_portal_addr);
--
-- return 0;
--}
--
--/**
-- * dprc_destroy_container() - Destroy child container.
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @token: Token of DPRC object
-- * @child_container_id: ID of the container to destroy
-- *
-- * This function terminates the child container, so following this call the
-- * child container ID becomes invalid.
-- *
-- * Notes:
-- * - All resources and objects of the destroyed container are returned to the
-- * parent container or destroyed if were created be the destroyed container.
-- * - This function destroy all the child containers of the specified
-- * container prior to destroying the container itself.
-- *
-- * warning: Only the parent container is allowed to destroy a child policy
-- * Container 0 can't be destroyed
-- *
-- * Return: '0' on Success; Error code otherwise.
-- *
-- */
--int dprc_destroy_container(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- int child_container_id)
--{
-- struct mc_command cmd = { 0 };
-- struct dprc_cmd_destroy_container *cmd_params;
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPRC_CMDID_DESTROY_CONT,
-- cmd_flags, token);
-- cmd_params = (struct dprc_cmd_destroy_container *)cmd.params;
-- cmd_params->child_container_id = cpu_to_le32(child_container_id);
--
-- /* send command to mc*/
-- return mc_send_command(mc_io, &cmd);
--}
--
--/**
-- * dprc_reset_container - Reset child container.
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @token: Token of DPRC object
-- * @child_container_id: ID of the container to reset
-- *
-- * In case a software context crashes or becomes non-responsive, the parent
-- * may wish to reset its resources container before the software context is
-- * restarted.
-- *
-- * This routine informs all objects assigned to the child container that the
-- * container is being reset, so they may perform any cleanup operations that are
-- * needed. All objects handles that were owned by the child container shall be
-- * closed.
-- *
-- * Note that such request may be submitted even if the child software context
-- * has not crashed, but the resulting object cleanup operations will not be
-- * aware of that.
-- *
-- * Return: '0' on Success; Error code otherwise.
-- */
--int dprc_reset_container(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- int child_container_id)
--{
-- struct mc_command cmd = { 0 };
-- struct dprc_cmd_reset_container *cmd_params;
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPRC_CMDID_RESET_CONT,
-- cmd_flags, token);
-- cmd_params = (struct dprc_cmd_reset_container *)cmd.params;
-- cmd_params->child_container_id = cpu_to_le32(child_container_id);
--
-- /* send command to mc*/
-- return mc_send_command(mc_io, &cmd);
--}
--
--/**
-- * dprc_get_irq() - Get IRQ information from the DPRC.
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @token: Token of DPRC object
-- * @irq_index: The interrupt index to configure
-- * @type: Interrupt type: 0 represents message interrupt
-- * type (both irq_addr and irq_val are valid)
-- * @irq_cfg: IRQ attributes
-- *
-- * Return: '0' on Success; Error code otherwise.
-- */
--int dprc_get_irq(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- u8 irq_index,
-- int *type,
-- struct dprc_irq_cfg *irq_cfg)
--{
-- struct mc_command cmd = { 0 };
-- struct dprc_cmd_get_irq *cmd_params;
-- struct dprc_rsp_get_irq *rsp_params;
-- int err;
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPRC_CMDID_GET_IRQ,
-- cmd_flags,
-- token);
-- cmd_params = (struct dprc_cmd_get_irq *)cmd.params;
-- cmd_params->irq_index = irq_index;
--
-- /* send command to mc*/
-- err = mc_send_command(mc_io, &cmd);
-- if (err)
-- return err;
--
-- /* retrieve response parameters */
-- rsp_params = (struct dprc_rsp_get_irq *)cmd.params;
-- irq_cfg->val = le32_to_cpu(rsp_params->irq_val);
-- irq_cfg->paddr = le64_to_cpu(rsp_params->irq_addr);
-- irq_cfg->irq_num = le32_to_cpu(rsp_params->irq_num);
-- *type = le32_to_cpu(rsp_params->type);
--
-- return 0;
--}
--
--/**
-- * dprc_set_irq() - Set IRQ information for the DPRC to trigger an interrupt.
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @token: Token of DPRC object
-- * @irq_index: Identifies the interrupt index to configure
-- * @irq_cfg: IRQ configuration
-- *
-- * Return: '0' on Success; Error code otherwise.
-- */
--int dprc_set_irq(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- u8 irq_index,
-- struct dprc_irq_cfg *irq_cfg)
--{
-- struct mc_command cmd = { 0 };
-- struct dprc_cmd_set_irq *cmd_params;
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPRC_CMDID_SET_IRQ,
-- cmd_flags,
-- token);
-- cmd_params = (struct dprc_cmd_set_irq *)cmd.params;
-- cmd_params->irq_val = cpu_to_le32(irq_cfg->val);
-- cmd_params->irq_index = irq_index;
-- cmd_params->irq_addr = cpu_to_le64(irq_cfg->paddr);
-- cmd_params->irq_num = cpu_to_le32(irq_cfg->irq_num);
--
-- /* send command to mc*/
-- return mc_send_command(mc_io, &cmd);
--}
--
--/**
-- * dprc_get_irq_enable() - Get overall interrupt state.
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @token: Token of DPRC object
-- * @irq_index: The interrupt index to configure
-- * @en: Returned interrupt state - enable = 1, disable = 0
-- *
-- * Return: '0' on Success; Error code otherwise.
-- */
--int dprc_get_irq_enable(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- u8 irq_index,
-- u8 *en)
--{
-- struct mc_command cmd = { 0 };
-- struct dprc_cmd_get_irq_enable *cmd_params;
-- struct dprc_rsp_get_irq_enable *rsp_params;
-- int err;
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPRC_CMDID_GET_IRQ_ENABLE,
-- cmd_flags, token);
-- cmd_params = (struct dprc_cmd_get_irq_enable *)cmd.params;
-- cmd_params->irq_index = irq_index;
--
-- /* send command to mc*/
-- err = mc_send_command(mc_io, &cmd);
-- if (err)
-- return err;
--
-- /* retrieve response parameters */
-- rsp_params = (struct dprc_rsp_get_irq_enable *)cmd.params;
-- *en = rsp_params->enabled & DPRC_ENABLE;
--
-- return 0;
--}
--
--/**
-- * dprc_set_irq_enable() - Set overall interrupt state.
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @token: Token of DPRC object
-- * @irq_index: The interrupt index to configure
-- * @en: Interrupt state - enable = 1, disable = 0
-- *
-- * Allows GPP software to control when interrupts are generated.
-- * Each interrupt can have up to 32 causes. The enable/disable control's the
-- * overall interrupt state. if the interrupt is disabled no causes will cause
-- * an interrupt.
-- *
-- * Return: '0' on Success; Error code otherwise.
-- */
--int dprc_set_irq_enable(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- u8 irq_index,
-- u8 en)
--{
-- struct mc_command cmd = { 0 };
-- struct dprc_cmd_set_irq_enable *cmd_params;
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPRC_CMDID_SET_IRQ_ENABLE,
-- cmd_flags, token);
-- cmd_params = (struct dprc_cmd_set_irq_enable *)cmd.params;
-- cmd_params->enable = en & DPRC_ENABLE;
-- cmd_params->irq_index = irq_index;
--
-- /* send command to mc*/
-- return mc_send_command(mc_io, &cmd);
--}
--
--/**
-- * dprc_get_irq_mask() - Get interrupt mask.
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @token: Token of DPRC object
-- * @irq_index: The interrupt index to configure
-- * @mask: Returned event mask to trigger interrupt
-- *
-- * Every interrupt can have up to 32 causes and the interrupt model supports
-- * masking/unmasking each cause independently
-- *
-- * Return: '0' on Success; Error code otherwise.
-- */
--int dprc_get_irq_mask(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- u8 irq_index,
-- u32 *mask)
--{
-- struct mc_command cmd = { 0 };
-- struct dprc_cmd_get_irq_mask *cmd_params;
-- struct dprc_rsp_get_irq_mask *rsp_params;
-- int err;
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPRC_CMDID_GET_IRQ_MASK,
-- cmd_flags, token);
-- cmd_params = (struct dprc_cmd_get_irq_mask *)cmd.params;
-- cmd_params->irq_index = irq_index;
--
-- /* send command to mc*/
-- err = mc_send_command(mc_io, &cmd);
-- if (err)
-- return err;
--
-- /* retrieve response parameters */
-- rsp_params = (struct dprc_rsp_get_irq_mask *)cmd.params;
-- *mask = le32_to_cpu(rsp_params->mask);
--
-- return 0;
--}
--
--/**
-- * dprc_set_irq_mask() - Set interrupt mask.
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @token: Token of DPRC object
-- * @irq_index: The interrupt index to configure
-- * @mask: event mask to trigger interrupt;
-- * each bit:
-- * 0 = ignore event
-- * 1 = consider event for asserting irq
-- *
-- * Every interrupt can have up to 32 causes and the interrupt model supports
-- * masking/unmasking each cause independently
-- *
-- * Return: '0' on Success; Error code otherwise.
-- */
--int dprc_set_irq_mask(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- u8 irq_index,
-- u32 mask)
--{
-- struct mc_command cmd = { 0 };
-- struct dprc_cmd_set_irq_mask *cmd_params;
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPRC_CMDID_SET_IRQ_MASK,
-- cmd_flags, token);
-- cmd_params = (struct dprc_cmd_set_irq_mask *)cmd.params;
-- cmd_params->mask = cpu_to_le32(mask);
-- cmd_params->irq_index = irq_index;
--
-- /* send command to mc*/
-- return mc_send_command(mc_io, &cmd);
--}
--
--/**
-- * dprc_get_irq_status() - Get the current status of any pending interrupts.
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @token: Token of DPRC object
-- * @irq_index: The interrupt index to configure
-- * @status: Returned interrupts status - one bit per cause:
-- * 0 = no interrupt pending
-- * 1 = interrupt pending
-- *
-- * Return: '0' on Success; Error code otherwise.
-- */
--int dprc_get_irq_status(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- u8 irq_index,
-- u32 *status)
--{
-- struct mc_command cmd = { 0 };
-- struct dprc_cmd_get_irq_status *cmd_params;
-- struct dprc_rsp_get_irq_status *rsp_params;
-- int err;
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPRC_CMDID_GET_IRQ_STATUS,
-- cmd_flags, token);
-- cmd_params = (struct dprc_cmd_get_irq_status *)cmd.params;
-- cmd_params->status = cpu_to_le32(*status);
-- cmd_params->irq_index = irq_index;
--
-- /* send command to mc*/
-- err = mc_send_command(mc_io, &cmd);
-- if (err)
-- return err;
--
-- /* retrieve response parameters */
-- rsp_params = (struct dprc_rsp_get_irq_status *)cmd.params;
-- *status = le32_to_cpu(rsp_params->status);
--
-- return 0;
--}
--
--/**
-- * dprc_clear_irq_status() - Clear a pending interrupt's status
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @token: Token of DPRC object
-- * @irq_index: The interrupt index to configure
-- * @status: bits to clear (W1C) - one bit per cause:
-- * 0 = don't change
-- * 1 = clear status bit
-- *
-- * Return: '0' on Success; Error code otherwise.
-- */
--int dprc_clear_irq_status(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- u8 irq_index,
-- u32 status)
--{
-- struct mc_command cmd = { 0 };
-- struct dprc_cmd_clear_irq_status *cmd_params;
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPRC_CMDID_CLEAR_IRQ_STATUS,
-- cmd_flags, token);
-- cmd_params = (struct dprc_cmd_clear_irq_status *)cmd.params;
-- cmd_params->status = cpu_to_le32(status);
-- cmd_params->irq_index = irq_index;
--
-- /* send command to mc*/
-- return mc_send_command(mc_io, &cmd);
--}
--
--/**
-- * dprc_get_attributes() - Obtains container attributes
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @token: Token of DPRC object
-- * @attributes Returned container attributes
-- *
-- * Return: '0' on Success; Error code otherwise.
-- */
--int dprc_get_attributes(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- struct dprc_attributes *attr)
--{
-- struct mc_command cmd = { 0 };
-- struct dprc_rsp_get_attributes *rsp_params;
-- int err;
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPRC_CMDID_GET_ATTR,
-- cmd_flags,
-- token);
--
-- /* send command to mc*/
-- err = mc_send_command(mc_io, &cmd);
-- if (err)
-- return err;
--
-- /* retrieve response parameters */
-- rsp_params = (struct dprc_rsp_get_attributes *)cmd.params;
-- attr->container_id = le32_to_cpu(rsp_params->container_id);
-- attr->icid = le16_to_cpu(rsp_params->icid);
-- attr->options = le32_to_cpu(rsp_params->options);
-- attr->portal_id = le32_to_cpu(rsp_params->portal_id);
-- attr->version.major = le16_to_cpu(rsp_params->version_major);
-- attr->version.minor = le16_to_cpu(rsp_params->version_minor);
--
-- return 0;
--}
--
--/**
-- * dprc_set_res_quota() - Set allocation policy for a specific resource/object
-- * type in a child container
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @token: Token of DPRC object
-- * @child_container_id: ID of the child container
-- * @type: Resource/object type
-- * @quota: Sets the maximum number of resources of the selected type
-- * that the child container is allowed to allocate from its parent;
-- * when quota is set to -1, the policy is the same as container's
-- * general policy.
-- *
-- * Allocation policy determines whether or not a container may allocate
-- * resources from its parent. Each container has a 'global' allocation policy
-- * that is set when the container is created.
-- *
-- * This function sets allocation policy for a specific resource type.
-- * The default policy for all resource types matches the container's 'global'
-- * allocation policy.
-- *
-- * Return: '0' on Success; Error code otherwise.
-- *
-- * @warning Only the parent container is allowed to change a child policy.
-- */
--int dprc_set_res_quota(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- int child_container_id,
-- char *type,
-- u16 quota)
--{
-- struct mc_command cmd = { 0 };
-- struct dprc_cmd_set_res_quota *cmd_params;
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPRC_CMDID_SET_RES_QUOTA,
-- cmd_flags, token);
-- cmd_params = (struct dprc_cmd_set_res_quota *)cmd.params;
-- cmd_params->child_container_id = cpu_to_le32(child_container_id);
-- cmd_params->quota = cpu_to_le16(quota);
-- strncpy(cmd_params->type, type, 16);
-- cmd_params->type[15] = '\0';
--
-- /* send command to mc*/
-- return mc_send_command(mc_io, &cmd);
--}
--
--/**
-- * dprc_get_res_quota() - Gets the allocation policy of a specific
-- * resource/object type in a child container
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @token: Token of DPRC object
-- * @child_container_id; ID of the child container
-- * @type: resource/object type
-- * @quota: Returnes the maximum number of resources of the selected type
-- * that the child container is allowed to allocate from the parent;
-- * when quota is set to -1, the policy is the same as container's
-- * general policy.
-- *
-- * Return: '0' on Success; Error code otherwise.
-- */
--int dprc_get_res_quota(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- int child_container_id,
-- char *type,
-- u16 *quota)
--{
-- struct mc_command cmd = { 0 };
-- struct dprc_cmd_get_res_quota *cmd_params;
-- struct dprc_rsp_get_res_quota *rsp_params;
-- int err;
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPRC_CMDID_GET_RES_QUOTA,
-- cmd_flags, token);
-- cmd_params = (struct dprc_cmd_get_res_quota *)cmd.params;
-- cmd_params->child_container_id = cpu_to_le32(child_container_id);
-- strncpy(cmd_params->type, type, 16);
-- cmd_params->type[15] = '\0';
--
-- /* send command to mc*/
-- err = mc_send_command(mc_io, &cmd);
-- if (err)
-- return err;
--
-- /* retrieve response parameters */
-- rsp_params = (struct dprc_rsp_get_res_quota *)cmd.params;
-- *quota = le16_to_cpu(rsp_params->quota);
--
-- return 0;
--}
--
--/**
-- * dprc_assign() - Assigns objects or resource to a child container.
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @token: Token of DPRC object
-- * @container_id: ID of the child container
-- * @res_req: Describes the type and amount of resources to
-- * assign to the given container
-- *
-- * Assignment is usually done by a parent (this DPRC) to one of its child
-- * containers.
-- *
-- * According to the DPRC allocation policy, the assigned resources may be taken
-- * (allocated) from the container's ancestors, if not enough resources are
-- * available in the container itself.
-- *
-- * The type of assignment depends on the dprc_res_req options, as follows:
-- * - DPRC_RES_REQ_OPT_EXPLICIT: indicates that assigned resources should have
-- * the explicit base ID specified at the id_base_align field of res_req.
-- * - DPRC_RES_REQ_OPT_ALIGNED: indicates that the assigned resources should be
-- * aligned to the value given at id_base_align field of res_req.
-- * - DPRC_RES_REQ_OPT_PLUGGED: Relevant only for object assignment,
-- * and indicates that the object must be set to the plugged state.
-- *
-- * A container may use this function with its own ID in order to change a
-- * object state to plugged or unplugged.
-- *
-- * If IRQ information has been set in the child DPRC, it will signal an
-- * interrupt following every change in its object assignment.
-- *
-- * Return: '0' on Success; Error code otherwise.
-- */
--int dprc_assign(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- int container_id,
-- struct dprc_res_req *res_req)
--{
-- struct mc_command cmd = { 0 };
-- struct dprc_cmd_assign *cmd_params;
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPRC_CMDID_ASSIGN,
-- cmd_flags, token);
-- cmd_params = (struct dprc_cmd_assign *)cmd.params;
-- cmd_params->container_id = cpu_to_le32(container_id);
-- cmd_params->options = cpu_to_le32(res_req->options);
-- cmd_params->num = cpu_to_le32(res_req->num);
-- cmd_params->id_base_align = cpu_to_le32(res_req->id_base_align);
-- strncpy(cmd_params->type, res_req->type, 16);
-- cmd_params->type[15] = '\0';
--
-- /* send command to mc*/
-- return mc_send_command(mc_io, &cmd);
--}
--
--/**
-- * dprc_unassign() - Un-assigns objects or resources from a child container
-- * and moves them into this (parent) DPRC.
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @token: Token of DPRC object
-- * @child_container_id: ID of the child container
-- * @res_req: Describes the type and amount of resources to un-assign from
-- * the child container
-- *
-- * Un-assignment of objects can succeed only if the object is not in the
-- * plugged or opened state.
-- *
-- * Return: '0' on Success; Error code otherwise.
-- */
--int dprc_unassign(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- int child_container_id,
-- struct dprc_res_req *res_req)
--{
-- struct mc_command cmd = { 0 };
-- struct dprc_cmd_unassign *cmd_params;
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPRC_CMDID_UNASSIGN,
-- cmd_flags,
-- token);
-- cmd_params = (struct dprc_cmd_unassign *)cmd.params;
-- cmd_params->child_container_id = cpu_to_le32(child_container_id);
-- cmd_params->options = cpu_to_le32(res_req->options);
-- cmd_params->num = cpu_to_le32(res_req->num);
-- cmd_params->id_base_align = cpu_to_le32(res_req->id_base_align);
-- strncpy(cmd_params->type, res_req->type, 16);
-- cmd_params->type[15] = '\0';
--
-- /* send command to mc*/
-- return mc_send_command(mc_io, &cmd);
--}
--
--/**
-- * dprc_get_pool_count() - Get the number of dprc's pools
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @mc_io: Pointer to MC portal's I/O object
-- * @token: Token of DPRC object
-- * @pool_count: Returned number of resource pools in the dprc
-- *
-- * Return: '0' on Success; Error code otherwise.
-- */
--int dprc_get_pool_count(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- int *pool_count)
--{
-- struct mc_command cmd = { 0 };
-- struct dprc_rsp_get_pool_count *rsp_params;
-- int err;
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPRC_CMDID_GET_POOL_COUNT,
-- cmd_flags, token);
--
-- /* send command to mc*/
-- err = mc_send_command(mc_io, &cmd);
-- if (err)
-- return err;
--
-- /* retrieve response parameters */
-- rsp_params = (struct dprc_rsp_get_pool_count *)cmd.params;
-- *pool_count = le32_to_cpu(rsp_params->pool_count);
--
-- return 0;
--}
--
--/**
-- * dprc_get_pool() - Get the type (string) of a certain dprc's pool
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @token: Token of DPRC object
-- * @pool_index; Index of the pool to be queried (< pool_count)
-- * @type: The type of the pool
-- *
-- * The pool types retrieved one by one by incrementing
-- * pool_index up to (not including) the value of pool_count returned
-- * from dprc_get_pool_count(). dprc_get_pool_count() must
-- * be called prior to dprc_get_pool().
-- *
-- * Return: '0' on Success; Error code otherwise.
-- */
--int dprc_get_pool(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- int pool_index,
-- char *type)
--{
-- struct mc_command cmd = { 0 };
-- struct dprc_cmd_get_pool *cmd_params;
-- struct dprc_rsp_get_pool *rsp_params;
-- int err;
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPRC_CMDID_GET_POOL,
-- cmd_flags,
-- token);
-- cmd_params = (struct dprc_cmd_get_pool *)cmd.params;
-- cmd_params->pool_index = cpu_to_le32(pool_index);
--
-- /* send command to mc*/
-- err = mc_send_command(mc_io, &cmd);
-- if (err)
-- return err;
--
-- /* retrieve response parameters */
-- rsp_params = (struct dprc_rsp_get_pool *)cmd.params;
-- strncpy(type, rsp_params->type, 16);
-- type[15] = '\0';
--
-- return 0;
--}
--
--/**
-- * dprc_get_obj_count() - Obtains the number of objects in the DPRC
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @token: Token of DPRC object
-- * @obj_count: Number of objects assigned to the DPRC
-- *
-- * Return: '0' on Success; Error code otherwise.
-- */
--int dprc_get_obj_count(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- int *obj_count)
--{
-- struct mc_command cmd = { 0 };
-- struct dprc_rsp_get_obj_count *rsp_params;
-- int err;
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPRC_CMDID_GET_OBJ_COUNT,
-- cmd_flags, token);
--
-- /* send command to mc*/
-- err = mc_send_command(mc_io, &cmd);
-- if (err)
-- return err;
--
-- /* retrieve response parameters */
-- rsp_params = (struct dprc_rsp_get_obj_count *)cmd.params;
-- *obj_count = le32_to_cpu(rsp_params->obj_count);
--
-- return 0;
--}
--EXPORT_SYMBOL(dprc_get_obj_count);
--
--/**
-- * dprc_get_obj() - Get general information on an object
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @token: Token of DPRC object
-- * @obj_index: Index of the object to be queried (< obj_count)
-- * @obj_desc: Returns the requested object descriptor
-- *
-- * The object descriptors are retrieved one by one by incrementing
-- * obj_index up to (not including) the value of obj_count returned
-- * from dprc_get_obj_count(). dprc_get_obj_count() must
-- * be called prior to dprc_get_obj().
-- *
-- * Return: '0' on Success; Error code otherwise.
-- */
--int dprc_get_obj(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- int obj_index,
-- struct dprc_obj_desc *obj_desc)
--{
-- struct mc_command cmd = { 0 };
-- struct dprc_cmd_get_obj *cmd_params;
-- struct dprc_rsp_get_obj *rsp_params;
-- int err;
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPRC_CMDID_GET_OBJ,
-- cmd_flags,
-- token);
-- cmd_params = (struct dprc_cmd_get_obj *)cmd.params;
-- cmd_params->obj_index = cpu_to_le32(obj_index);
--
-- /* send command to mc*/
-- err = mc_send_command(mc_io, &cmd);
-- if (err)
-- return err;
--
-- /* retrieve response parameters */
-- rsp_params = (struct dprc_rsp_get_obj *)cmd.params;
-- obj_desc->id = le32_to_cpu(rsp_params->id);
-- obj_desc->vendor = le16_to_cpu(rsp_params->vendor);
-- obj_desc->irq_count = rsp_params->irq_count;
-- obj_desc->region_count = rsp_params->region_count;
-- obj_desc->state = le32_to_cpu(rsp_params->state);
-- obj_desc->ver_major = le16_to_cpu(rsp_params->version_major);
-- obj_desc->ver_minor = le16_to_cpu(rsp_params->version_minor);
-- obj_desc->flags = le16_to_cpu(rsp_params->flags);
-- strncpy(obj_desc->type, rsp_params->type, 16);
-- obj_desc->type[15] = '\0';
-- strncpy(obj_desc->label, rsp_params->label, 16);
-- obj_desc->label[15] = '\0';
-- return 0;
--}
--EXPORT_SYMBOL(dprc_get_obj);
--
--/**
-- * dprc_get_obj_desc() - Get object descriptor.
-- *
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @token: Token of DPRC object
-- * @obj_type: The type of the object to get its descriptor.
-- * @obj_id: The id of the object to get its descriptor
-- * @obj_desc: The returned descriptor to fill and return to the user
-- *
-- * Return: '0' on Success; Error code otherwise.
-- *
-- */
--int dprc_get_obj_desc(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- char *obj_type,
-- int obj_id,
-- struct dprc_obj_desc *obj_desc)
--{
-- struct mc_command cmd = { 0 };
-- struct dprc_cmd_get_obj_desc *cmd_params;
-- struct dprc_rsp_get_obj_desc *rsp_params;
-- int err;
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPRC_CMDID_GET_OBJ_DESC,
-- cmd_flags,
-- token);
-- cmd_params = (struct dprc_cmd_get_obj_desc *)cmd.params;
-- cmd_params->obj_id = cpu_to_le32(obj_id);
-- strncpy(cmd_params->type, obj_type, 16);
-- cmd_params->type[15] = '\0';
--
-- /* send command to mc*/
-- err = mc_send_command(mc_io, &cmd);
-- if (err)
-- return err;
--
-- /* retrieve response parameters */
-- rsp_params = (struct dprc_rsp_get_obj_desc *)cmd.params;
-- obj_desc->id = le32_to_cpu(rsp_params->id);
-- obj_desc->vendor = le16_to_cpu(rsp_params->vendor);
-- obj_desc->irq_count = rsp_params->irq_count;
-- obj_desc->region_count = rsp_params->region_count;
-- obj_desc->state = le32_to_cpu(rsp_params->state);
-- obj_desc->ver_major = le16_to_cpu(rsp_params->version_major);
-- obj_desc->ver_minor = le16_to_cpu(rsp_params->version_minor);
-- obj_desc->flags = le16_to_cpu(rsp_params->flags);
-- strncpy(obj_desc->type, rsp_params->type, 16);
-- obj_desc->type[15] = '\0';
-- strncpy(obj_desc->label, rsp_params->label, 16);
-- obj_desc->label[15] = '\0';
--
-- return 0;
--}
--EXPORT_SYMBOL(dprc_get_obj_desc);
--
--/**
-- * dprc_set_obj_irq() - Set IRQ information for object to trigger an interrupt.
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @token: Token of DPRC object
-- * @obj_type: Type of the object to set its IRQ
-- * @obj_id: ID of the object to set its IRQ
-- * @irq_index: The interrupt index to configure
-- * @irq_cfg: IRQ configuration
-- *
-- * Return: '0' on Success; Error code otherwise.
-- */
--int dprc_set_obj_irq(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- char *obj_type,
-- int obj_id,
-- u8 irq_index,
-- struct dprc_irq_cfg *irq_cfg)
--{
-- struct mc_command cmd = { 0 };
-- struct dprc_cmd_set_obj_irq *cmd_params;
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPRC_CMDID_SET_OBJ_IRQ,
-- cmd_flags,
-- token);
-- cmd_params = (struct dprc_cmd_set_obj_irq *)cmd.params;
-- cmd_params->irq_val = cpu_to_le32(irq_cfg->val);
-- cmd_params->irq_index = irq_index;
-- cmd_params->irq_addr = cpu_to_le64(irq_cfg->paddr);
-- cmd_params->irq_num = cpu_to_le32(irq_cfg->irq_num);
-- cmd_params->obj_id = cpu_to_le32(obj_id);
-- strncpy(cmd_params->obj_type, obj_type, 16);
-- cmd_params->obj_type[15] = '\0';
--
-- /* send command to mc*/
-- return mc_send_command(mc_io, &cmd);
--}
--EXPORT_SYMBOL(dprc_set_obj_irq);
--
--/**
-- * dprc_get_obj_irq() - Get IRQ information from object.
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @token: Token of DPRC object
-- * @obj_type: Type od the object to get its IRQ
-- * @obj_id: ID of the object to get its IRQ
-- * @irq_index: The interrupt index to configure
-- * @type: Interrupt type: 0 represents message interrupt
-- * type (both irq_addr and irq_val are valid)
-- * @irq_cfg: The returned IRQ attributes
-- *
-- * Return: '0' on Success; Error code otherwise.
-- */
--int dprc_get_obj_irq(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- char *obj_type,
-- int obj_id,
-- u8 irq_index,
-- int *type,
-- struct dprc_irq_cfg *irq_cfg)
--{
-- struct mc_command cmd = { 0 };
-- struct dprc_cmd_get_obj_irq *cmd_params;
-- struct dprc_rsp_get_obj_irq *rsp_params;
-- int err;
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPRC_CMDID_GET_OBJ_IRQ,
-- cmd_flags,
-- token);
-- cmd_params = (struct dprc_cmd_get_obj_irq *)cmd.params;
-- cmd_params->obj_id = cpu_to_le32(obj_id);
-- cmd_params->irq_index = irq_index;
-- strncpy(cmd_params->obj_type, obj_type, 16);
-- cmd_params->obj_type[15] = '\0';
--
-- /* send command to mc*/
-- err = mc_send_command(mc_io, &cmd);
-- if (err)
-- return err;
--
-- /* retrieve response parameters */
-- rsp_params = (struct dprc_rsp_get_obj_irq *)cmd.params;
-- irq_cfg->val = le32_to_cpu(rsp_params->irq_val);
-- irq_cfg->paddr = le64_to_cpu(rsp_params->irq_addr);
-- irq_cfg->irq_num = le32_to_cpu(rsp_params->irq_num);
-- *type = le32_to_cpu(rsp_params->type);
--
-- return 0;
--}
--EXPORT_SYMBOL(dprc_get_obj_irq);
--
--/**
-- * dprc_get_res_count() - Obtains the number of free resources that are assigned
-- * to this container, by pool type
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @token: Token of DPRC object
-- * @type: pool type
-- * @res_count: Returned number of free resources of the given
-- * resource type that are assigned to this DPRC
-- *
-- * Return: '0' on Success; Error code otherwise.
-- */
--int dprc_get_res_count(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- char *type,
-- int *res_count)
--{
-- struct mc_command cmd = { 0 };
-- struct dprc_cmd_get_res_count *cmd_params;
-- struct dprc_rsp_get_res_count *rsp_params;
-- int err;
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPRC_CMDID_GET_RES_COUNT,
-- cmd_flags, token);
-- cmd_params = (struct dprc_cmd_get_res_count *)cmd.params;
-- strncpy(cmd_params->type, type, 16);
-- cmd_params->type[15] = '\0';
--
-- /* send command to mc*/
-- err = mc_send_command(mc_io, &cmd);
-- if (err)
-- return err;
--
-- /* retrieve response parameters */
-- rsp_params = (struct dprc_rsp_get_res_count *)cmd.params;
-- *res_count = le32_to_cpu(rsp_params->res_count);
--
-- return 0;
--}
--EXPORT_SYMBOL(dprc_get_res_count);
--
--/**
-- * dprc_get_res_ids() - Obtains IDs of free resources in the container
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @token: Token of DPRC object
-- * @type: pool type
-- * @range_desc: range descriptor
-- *
-- * Return: '0' on Success; Error code otherwise.
-- */
--int dprc_get_res_ids(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- char *type,
-- struct dprc_res_ids_range_desc *range_desc)
--{
-- struct mc_command cmd = { 0 };
-- struct dprc_cmd_get_res_ids *cmd_params;
-- struct dprc_rsp_get_res_ids *rsp_params;
-- int err;
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPRC_CMDID_GET_RES_IDS,
-- cmd_flags, token);
-- cmd_params = (struct dprc_cmd_get_res_ids *)cmd.params;
-- cmd_params->iter_status = range_desc->iter_status;
-- cmd_params->base_id = cpu_to_le32(range_desc->base_id);
-- cmd_params->last_id = cpu_to_le32(range_desc->last_id);
-- strncpy(cmd_params->type, type, 16);
-- cmd_params->type[15] = '\0';
--
-- /* send command to mc*/
-- err = mc_send_command(mc_io, &cmd);
-- if (err)
-- return err;
--
-- /* retrieve response parameters */
-- rsp_params = (struct dprc_rsp_get_res_ids *)cmd.params;
-- range_desc->iter_status = rsp_params->iter_status;
-- range_desc->base_id = le32_to_cpu(rsp_params->base_id);
-- range_desc->last_id = le32_to_cpu(rsp_params->last_id);
--
-- return 0;
--}
--EXPORT_SYMBOL(dprc_get_res_ids);
--
--/**
-- * dprc_get_obj_region() - Get region information for a specified object.
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @token: Token of DPRC object
-- * @obj_type; Object type as returned in dprc_get_obj()
-- * @obj_id: Unique object instance as returned in dprc_get_obj()
-- * @region_index: The specific region to query
-- * @region_desc: Returns the requested region descriptor
-- *
-- * Return: '0' on Success; Error code otherwise.
-- */
--int dprc_get_obj_region(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- char *obj_type,
-- int obj_id,
-- u8 region_index,
-- struct dprc_region_desc *region_desc)
--{
-- struct mc_command cmd = { 0 };
-- struct dprc_cmd_get_obj_region *cmd_params;
-- struct dprc_rsp_get_obj_region *rsp_params;
-- int err;
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPRC_CMDID_GET_OBJ_REG,
-- cmd_flags, token);
-- cmd_params = (struct dprc_cmd_get_obj_region *)cmd.params;
-- cmd_params->obj_id = cpu_to_le32(obj_id);
-- cmd_params->region_index = region_index;
-- strncpy(cmd_params->obj_type, obj_type, 16);
-- cmd_params->obj_type[15] = '\0';
--
-- /* send command to mc*/
-- err = mc_send_command(mc_io, &cmd);
-- if (err)
-- return err;
--
-- /* retrieve response parameters */
-- rsp_params = (struct dprc_rsp_get_obj_region *)cmd.params;
-- region_desc->base_offset = le64_to_cpu(rsp_params->base_addr);
-- region_desc->size = le32_to_cpu(rsp_params->size);
--
-- return 0;
--}
--EXPORT_SYMBOL(dprc_get_obj_region);
--
--/**
-- * dprc_set_obj_label() - Set object label.
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @token: Token of DPRC object
-- * @obj_type: Object's type
-- * @obj_id: Object's ID
-- * @label: The required label. The maximum length is 16 chars.
-- *
-- * Return: '0' on Success; Error code otherwise.
-- */
--int dprc_set_obj_label(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- char *obj_type,
-- int obj_id,
-- char *label)
--{
-- struct mc_command cmd = { 0 };
-- struct dprc_cmd_set_obj_label *cmd_params;
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPRC_CMDID_SET_OBJ_LABEL,
-- cmd_flags,
-- token);
-- cmd_params = (struct dprc_cmd_set_obj_label *)cmd.params;
-- cmd_params->obj_id = cpu_to_le32(obj_id);
-- strncpy(cmd_params->label, label, 16);
-- cmd_params->label[15] = '\0';
-- strncpy(cmd_params->obj_type, obj_type, 16);
-- cmd_params->obj_type[15] = '\0';
--
-- /* send command to mc*/
-- return mc_send_command(mc_io, &cmd);
--}
--EXPORT_SYMBOL(dprc_set_obj_label);
--
--/**
-- * dprc_connect() - Connect two endpoints to create a network link between them
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @token: Token of DPRC object
-- * @endpoint1: Endpoint 1 configuration parameters
-- * @endpoint2: Endpoint 2 configuration parameters
-- * @cfg: Connection configuration. The connection configuration is ignored for
-- * connections made to DPMAC objects, where rate is retrieved from the
-- * MAC configuration.
-- *
-- * Return: '0' on Success; Error code otherwise.
-- */
--int dprc_connect(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- const struct dprc_endpoint *endpoint1,
-- const struct dprc_endpoint *endpoint2,
-- const struct dprc_connection_cfg *cfg)
--{
-- struct mc_command cmd = { 0 };
-- struct dprc_cmd_connect *cmd_params;
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPRC_CMDID_CONNECT,
-- cmd_flags,
-- token);
-- cmd_params = (struct dprc_cmd_connect *)cmd.params;
-- cmd_params->ep1_id = cpu_to_le32(endpoint1->id);
-- cmd_params->ep1_interface_id = cpu_to_le32(endpoint1->if_id);
-- cmd_params->ep2_id = cpu_to_le32(endpoint2->id);
-- cmd_params->ep2_interface_id = cpu_to_le32(endpoint2->if_id);
-- strncpy(cmd_params->ep1_type, endpoint1->type, 16);
-- cmd_params->ep1_type[15] = '\0';
-- cmd_params->max_rate = cpu_to_le32(cfg->max_rate);
-- cmd_params->committed_rate = cpu_to_le32(cfg->committed_rate);
-- strncpy(cmd_params->ep2_type, endpoint2->type, 16);
-- cmd_params->ep2_type[15] = '\0';
--
-- /* send command to mc*/
-- return mc_send_command(mc_io, &cmd);
--}
--
--/**
-- * dprc_disconnect() - Disconnect one endpoint to remove its network connection
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @token: Token of DPRC object
-- * @endpoint: Endpoint configuration parameters
-- *
-- * Return: '0' on Success; Error code otherwise.
-- */
--int dprc_disconnect(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- const struct dprc_endpoint *endpoint)
--{
-- struct mc_command cmd = { 0 };
-- struct dprc_cmd_disconnect *cmd_params;
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPRC_CMDID_DISCONNECT,
-- cmd_flags,
-- token);
-- cmd_params = (struct dprc_cmd_disconnect *)cmd.params;
-- cmd_params->id = cpu_to_le32(endpoint->id);
-- cmd_params->interface_id = cpu_to_le32(endpoint->if_id);
-- strncpy(cmd_params->type, endpoint->type, 16);
-- cmd_params->type[15] = '\0';
--
-- /* send command to mc*/
-- return mc_send_command(mc_io, &cmd);
--}
--
--/**
-- * dprc_get_connection() - Get connected endpoint and link status if connection
-- * exists.
-- * @mc_io: Pointer to MC portal's I/O object
-- * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-- * @token: Token of DPRC object
-- * @endpoint1: Endpoint 1 configuration parameters
-- * @endpoint2: Returned endpoint 2 configuration parameters
-- * @state: Returned link state:
-- * 1 - link is up;
-- * 0 - link is down;
-- * -1 - no connection (endpoint2 information is irrelevant)
-- *
-- * Return: '0' on Success; -ENAVAIL if connection does not exist.
-- */
--int dprc_get_connection(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- const struct dprc_endpoint *endpoint1,
-- struct dprc_endpoint *endpoint2,
-- int *state)
--{
-- struct mc_command cmd = { 0 };
-- struct dprc_cmd_get_connection *cmd_params;
-- struct dprc_rsp_get_connection *rsp_params;
-- int err;
--
-- /* prepare command */
-- cmd.header = mc_encode_cmd_header(DPRC_CMDID_GET_CONNECTION,
-- cmd_flags,
-- token);
-- cmd_params = (struct dprc_cmd_get_connection *)cmd.params;
-- cmd_params->ep1_id = cpu_to_le32(endpoint1->id);
-- cmd_params->ep1_interface_id = cpu_to_le32(endpoint1->if_id);
-- strncpy(cmd_params->ep1_type, endpoint1->type, 16);
-- cmd_params->ep1_type[15] = '\0';
--
-- /* send command to mc*/
-- err = mc_send_command(mc_io, &cmd);
-- if (err)
-- return err;
--
-- /* retrieve response parameters */
-- rsp_params = (struct dprc_rsp_get_connection *)cmd.params;
-- endpoint2->id = le32_to_cpu(rsp_params->ep2_id);
-- endpoint2->if_id = le32_to_cpu(rsp_params->ep2_interface_id);
-- strncpy(endpoint2->type, rsp_params->ep2_type, 16);
-- endpoint2->type[15] = '\0';
-- *state = le32_to_cpu(rsp_params->state);
--
-- return 0;
--}
---- a/drivers/staging/fsl-mc/bus/fsl-mc-private.h
-+++ /dev/null
-@@ -1,52 +0,0 @@
--/*
-- * Freescale Management Complex (MC) bus private declarations
-- *
-- * Copyright (C) 2016 Freescale Semiconductor, Inc.
-- *
-- * This file is licensed under the terms of the GNU General Public
-- * License version 2. This program is licensed "as is" without any
-- * warranty of any kind, whether express or implied.
-- */
--#ifndef _FSL_MC_PRIVATE_H_
--#define _FSL_MC_PRIVATE_H_
--
--int __must_check fsl_mc_device_add(struct dprc_obj_desc *obj_desc,
-- struct fsl_mc_io *mc_io,
-- struct device *parent_dev,
-- struct fsl_mc_device **new_mc_dev);
--
--void fsl_mc_device_remove(struct fsl_mc_device *mc_dev);
--
--int __init dprc_driver_init(void);
--
--void dprc_driver_exit(void);
--
--int __init fsl_mc_allocator_driver_init(void);
--
--void fsl_mc_allocator_driver_exit(void);
--
--int __must_check fsl_mc_resource_allocate(struct fsl_mc_bus *mc_bus,
-- enum fsl_mc_pool_type pool_type,
-- struct fsl_mc_resource
-- **new_resource);
--
--void fsl_mc_resource_free(struct fsl_mc_resource *resource);
--
--int fsl_mc_msi_domain_alloc_irqs(struct device *dev,
-- unsigned int irq_count);
--
--void fsl_mc_msi_domain_free_irqs(struct device *dev);
--
--int __init its_fsl_mc_msi_init(void);
--
--void its_fsl_mc_msi_cleanup(void);
--
--int __must_check fsl_create_mc_io(struct device *dev,
-- phys_addr_t mc_portal_phys_addr,
-- u32 mc_portal_size,
-- struct fsl_mc_device *dpmcp_dev,
-- u32 flags, struct fsl_mc_io **new_mc_io);
--
--void fsl_destroy_mc_io(struct fsl_mc_io *mc_io);
--
--#endif /* _FSL_MC_PRIVATE_H_ */
---- /dev/null
-+++ b/drivers/staging/fsl-mc/include/dpaa2-fd.h
-@@ -0,0 +1,681 @@
-+/* SPDX-License-Identifier: (GPL-2.0+ OR BSD-3-Clause) */
-+/*
-+ * Copyright 2014-2016 Freescale Semiconductor Inc.
-+ * Copyright 2016 NXP
-+ *
-+ */
-+#ifndef __FSL_DPAA2_FD_H
-+#define __FSL_DPAA2_FD_H
-+
-+#include <linux/kernel.h>
-+
-+/**
-+ * DOC: DPAA2 FD - Frame Descriptor APIs for DPAA2
-+ *
-+ * Frame Descriptors (FDs) are used to describe frame data in the DPAA2.
-+ * Frames can be enqueued and dequeued to Frame Queues (FQs) which are consumed
-+ * by the various DPAA accelerators (WRIOP, SEC, PME, DCE)
-+ *
-+ * There are three types of frames: single, scatter gather, and frame lists.
-+ *
-+ * The set of APIs in this file must be used to create, manipulate and
-+ * query Frame Descriptors.
-+ */
-+
-+/**
-+ * struct dpaa2_fd - Struct describing FDs
-+ * @words: for easier/faster copying the whole FD structure
-+ * @addr: address in the FD
-+ * @len: length in the FD
-+ * @bpid: buffer pool ID
-+ * @format_offset: format, offset, and short-length fields
-+ * @frc: frame context
-+ * @ctrl: control bits...including dd, sc, va, err, etc
-+ * @flc: flow context address
-+ *
-+ * This structure represents the basic Frame Descriptor used in the system.
-+ */
-+struct dpaa2_fd {
-+ union {
-+ u32 words[8];
-+ struct dpaa2_fd_simple {
-+ __le64 addr;
-+ __le32 len;
-+ __le16 bpid;
-+ __le16 format_offset;
-+ __le32 frc;
-+ __le32 ctrl;
-+ __le64 flc;
-+ } simple;
-+ };
-+};
-+
-+#define FD_SHORT_LEN_FLAG_MASK 0x1
-+#define FD_SHORT_LEN_FLAG_SHIFT 14
-+#define FD_SHORT_LEN_MASK 0x3FFFF
-+#define FD_OFFSET_MASK 0x0FFF
-+#define FD_FORMAT_MASK 0x3
-+#define FD_FORMAT_SHIFT 12
-+#define FD_BPID_MASK 0x3FFF
-+#define SG_SHORT_LEN_FLAG_MASK 0x1
-+#define SG_SHORT_LEN_FLAG_SHIFT 14
-+#define SG_SHORT_LEN_MASK 0x1FFFF
-+#define SG_OFFSET_MASK 0x0FFF
-+#define SG_FORMAT_MASK 0x3
-+#define SG_FORMAT_SHIFT 12
-+#define SG_BPID_MASK 0x3FFF
-+#define SG_FINAL_FLAG_MASK 0x1
-+#define SG_FINAL_FLAG_SHIFT 15
-+#define FL_SHORT_LEN_FLAG_MASK 0x1
-+#define FL_SHORT_LEN_FLAG_SHIFT 14
-+#define FL_SHORT_LEN_MASK 0x3FFFF
-+#define FL_OFFSET_MASK 0x0FFF
-+#define FL_FORMAT_MASK 0x3
-+#define FL_FORMAT_SHIFT 12
-+#define FL_BPID_MASK 0x3FFF
-+#define FL_FINAL_FLAG_MASK 0x1
-+#define FL_FINAL_FLAG_SHIFT 15
-+
-+/* Error bits in FD CTRL */
-+#define FD_CTRL_ERR_MASK 0x000000FF
-+#define FD_CTRL_UFD 0x00000004
-+#define FD_CTRL_SBE 0x00000008
-+#define FD_CTRL_FLC 0x00000010
-+#define FD_CTRL_FSE 0x00000020
-+#define FD_CTRL_FAERR 0x00000040
-+
-+/* Annotation bits in FD CTRL */
-+#define FD_CTRL_PTA 0x00800000
-+#define FD_CTRL_PTV1 0x00400000
-+
-+enum dpaa2_fd_format {
-+ dpaa2_fd_single = 0,
-+ dpaa2_fd_list,
-+ dpaa2_fd_sg
-+};
-+
-+/**
-+ * dpaa2_fd_get_addr() - get the addr field of frame descriptor
-+ * @fd: the given frame descriptor
-+ *
-+ * Return the address in the frame descriptor.
-+ */
-+static inline dma_addr_t dpaa2_fd_get_addr(const struct dpaa2_fd *fd)
-+{
-+ return (dma_addr_t)le64_to_cpu(fd->simple.addr);
-+}
-+
-+/**
-+ * dpaa2_fd_set_addr() - Set the addr field of frame descriptor
-+ * @fd: the given frame descriptor
-+ * @addr: the address needs to be set in frame descriptor
-+ */
-+static inline void dpaa2_fd_set_addr(struct dpaa2_fd *fd, dma_addr_t addr)
-+{
-+ fd->simple.addr = cpu_to_le64(addr);
-+}
-+
-+/**
-+ * dpaa2_fd_get_frc() - Get the frame context in the frame descriptor
-+ * @fd: the given frame descriptor
-+ *
-+ * Return the frame context field in the frame descriptor.
-+ */
-+static inline u32 dpaa2_fd_get_frc(const struct dpaa2_fd *fd)
-+{
-+ return le32_to_cpu(fd->simple.frc);
-+}
-+
-+/**
-+ * dpaa2_fd_set_frc() - Set the frame context in the frame descriptor
-+ * @fd: the given frame descriptor
-+ * @frc: the frame context needs to be set in frame descriptor
-+ */
-+static inline void dpaa2_fd_set_frc(struct dpaa2_fd *fd, u32 frc)
-+{
-+ fd->simple.frc = cpu_to_le32(frc);
-+}
-+
-+/**
-+ * dpaa2_fd_get_ctrl() - Get the control bits in the frame descriptor
-+ * @fd: the given frame descriptor
-+ *
-+ * Return the control bits field in the frame descriptor.
-+ */
-+static inline u32 dpaa2_fd_get_ctrl(const struct dpaa2_fd *fd)
-+{
-+ return le32_to_cpu(fd->simple.ctrl);
-+}
-+
-+/**
-+ * dpaa2_fd_set_ctrl() - Set the control bits in the frame descriptor
-+ * @fd: the given frame descriptor
-+ * @ctrl: the control bits to be set in the frame descriptor
-+ */
-+static inline void dpaa2_fd_set_ctrl(struct dpaa2_fd *fd, u32 ctrl)
-+{
-+ fd->simple.ctrl = cpu_to_le32(ctrl);
-+}
-+
-+/**
-+ * dpaa2_fd_get_flc() - Get the flow context in the frame descriptor
-+ * @fd: the given frame descriptor
-+ *
-+ * Return the flow context in the frame descriptor.
-+ */
-+static inline dma_addr_t dpaa2_fd_get_flc(const struct dpaa2_fd *fd)
-+{
-+ return (dma_addr_t)le64_to_cpu(fd->simple.flc);
-+}
-+
-+/**
-+ * dpaa2_fd_set_flc() - Set the flow context field of frame descriptor
-+ * @fd: the given frame descriptor
-+ * @flc_addr: the flow context needs to be set in frame descriptor
-+ */
-+static inline void dpaa2_fd_set_flc(struct dpaa2_fd *fd, dma_addr_t flc_addr)
-+{
-+ fd->simple.flc = cpu_to_le64(flc_addr);
-+}
-+
-+static inline bool dpaa2_fd_short_len(const struct dpaa2_fd *fd)
-+{
-+ return !!((le16_to_cpu(fd->simple.format_offset) >>
-+ FD_SHORT_LEN_FLAG_SHIFT) & FD_SHORT_LEN_FLAG_MASK);
-+}
-+
-+/**
-+ * dpaa2_fd_get_len() - Get the length in the frame descriptor
-+ * @fd: the given frame descriptor
-+ *
-+ * Return the length field in the frame descriptor.
-+ */
-+static inline u32 dpaa2_fd_get_len(const struct dpaa2_fd *fd)
-+{
-+ if (dpaa2_fd_short_len(fd))
-+ return le32_to_cpu(fd->simple.len) & FD_SHORT_LEN_MASK;
-+
-+ return le32_to_cpu(fd->simple.len);
-+}
-+
-+/**
-+ * dpaa2_fd_set_len() - Set the length field of frame descriptor
-+ * @fd: the given frame descriptor
-+ * @len: the length needs to be set in frame descriptor
-+ */
-+static inline void dpaa2_fd_set_len(struct dpaa2_fd *fd, u32 len)
-+{
-+ fd->simple.len = cpu_to_le32(len);
-+}
-+
-+/**
-+ * dpaa2_fd_get_offset() - Get the offset field in the frame descriptor
-+ * @fd: the given frame descriptor
-+ *
-+ * Return the offset.
-+ */
-+static inline uint16_t dpaa2_fd_get_offset(const struct dpaa2_fd *fd)
-+{
-+ return le16_to_cpu(fd->simple.format_offset) & FD_OFFSET_MASK;
-+}
-+
-+/**
-+ * dpaa2_fd_set_offset() - Set the offset field of frame descriptor
-+ * @fd: the given frame descriptor
-+ * @offset: the offset needs to be set in frame descriptor
-+ */
-+static inline void dpaa2_fd_set_offset(struct dpaa2_fd *fd, uint16_t offset)
-+{
-+ fd->simple.format_offset &= cpu_to_le16(~FD_OFFSET_MASK);
-+ fd->simple.format_offset |= cpu_to_le16(offset);
-+}
-+
-+/**
-+ * dpaa2_fd_get_format() - Get the format field in the frame descriptor
-+ * @fd: the given frame descriptor
-+ *
-+ * Return the format.
-+ */
-+static inline enum dpaa2_fd_format dpaa2_fd_get_format(
-+ const struct dpaa2_fd *fd)
-+{
-+ return (enum dpaa2_fd_format)((le16_to_cpu(fd->simple.format_offset)
-+ >> FD_FORMAT_SHIFT) & FD_FORMAT_MASK);
-+}
-+
-+/**
-+ * dpaa2_fd_set_format() - Set the format field of frame descriptor
-+ * @fd: the given frame descriptor
-+ * @format: the format needs to be set in frame descriptor
-+ */
-+static inline void dpaa2_fd_set_format(struct dpaa2_fd *fd,
-+ enum dpaa2_fd_format format)
-+{
-+ fd->simple.format_offset &=
-+ cpu_to_le16(~(FD_FORMAT_MASK << FD_FORMAT_SHIFT));
-+ fd->simple.format_offset |= cpu_to_le16(format << FD_FORMAT_SHIFT);
-+}
-+
-+/**
-+ * dpaa2_fd_get_bpid() - Get the bpid field in the frame descriptor
-+ * @fd: the given frame descriptor
-+ *
-+ * Return the buffer pool id.
-+ */
-+static inline uint16_t dpaa2_fd_get_bpid(const struct dpaa2_fd *fd)
-+{
-+ return le16_to_cpu(fd->simple.bpid) & FD_BPID_MASK;
-+}
-+
-+/**
-+ * dpaa2_fd_set_bpid() - Set the bpid field of frame descriptor
-+ * @fd: the given frame descriptor
-+ * @bpid: buffer pool id to be set
-+ */
-+static inline void dpaa2_fd_set_bpid(struct dpaa2_fd *fd, uint16_t bpid)
-+{
-+ fd->simple.bpid &= cpu_to_le16(~(FD_BPID_MASK));
-+ fd->simple.bpid |= cpu_to_le16(bpid);
-+}
-+
-+/**
-+ * struct dpaa2_sg_entry - the scatter-gathering structure
-+ * @addr: address of the sg entry
-+ * @len: length in this sg entry
-+ * @bpid: buffer pool id
-+ * @format_offset: format and offset fields
-+ */
-+struct dpaa2_sg_entry {
-+ __le64 addr;
-+ __le32 len;
-+ __le16 bpid;
-+ __le16 format_offset;
-+};
-+
-+enum dpaa2_sg_format {
-+ dpaa2_sg_single = 0,
-+ dpaa2_sg_frame_data,
-+ dpaa2_sg_sgt_ext
-+};
-+
-+/* Accessors for SG entry fields */
-+
-+/**
-+ * dpaa2_sg_get_addr() - Get the address from SG entry
-+ * @sg: the given scatter-gathering object
-+ *
-+ * Return the address.
-+ */
-+static inline dma_addr_t dpaa2_sg_get_addr(const struct dpaa2_sg_entry *sg)
-+{
-+ return (dma_addr_t)le64_to_cpu(sg->addr);
-+}
-+
-+/**
-+ * dpaa2_sg_set_addr() - Set the address in SG entry
-+ * @sg: the given scatter-gathering object
-+ * @addr: the address to be set
-+ */
-+static inline void dpaa2_sg_set_addr(struct dpaa2_sg_entry *sg, dma_addr_t addr)
-+{
-+ sg->addr = cpu_to_le64(addr);
-+}
-+
-+static inline bool dpaa2_sg_short_len(const struct dpaa2_sg_entry *sg)
-+{
-+ return !!((le16_to_cpu(sg->format_offset) >> SG_SHORT_LEN_FLAG_SHIFT)
-+ & SG_SHORT_LEN_FLAG_MASK);
-+}
-+
-+/**
-+ * dpaa2_sg_get_len() - Get the length in SG entry
-+ * @sg: the given scatter-gathering object
-+ *
-+ * Return the length.
-+ */
-+static inline u32 dpaa2_sg_get_len(const struct dpaa2_sg_entry *sg)
-+{
-+ if (dpaa2_sg_short_len(sg))
-+ return le32_to_cpu(sg->len) & SG_SHORT_LEN_MASK;
-+
-+ return le32_to_cpu(sg->len);
-+}
-+
-+/**
-+ * dpaa2_sg_set_len() - Set the length in SG entry
-+ * @sg: the given scatter-gathering object
-+ * @len: the length to be set
-+ */
-+static inline void dpaa2_sg_set_len(struct dpaa2_sg_entry *sg, u32 len)
-+{
-+ sg->len = cpu_to_le32(len);
-+}
-+
-+/**
-+ * dpaa2_sg_get_offset() - Get the offset in SG entry
-+ * @sg: the given scatter-gathering object
-+ *
-+ * Return the offset.
-+ */
-+static inline u16 dpaa2_sg_get_offset(const struct dpaa2_sg_entry *sg)
-+{
-+ return le16_to_cpu(sg->format_offset) & SG_OFFSET_MASK;
-+}
-+
-+/**
-+ * dpaa2_sg_set_offset() - Set the offset in SG entry
-+ * @sg: the given scatter-gathering object
-+ * @offset: the offset to be set
-+ */
-+static inline void dpaa2_sg_set_offset(struct dpaa2_sg_entry *sg,
-+ u16 offset)
-+{
-+ sg->format_offset &= cpu_to_le16(~SG_OFFSET_MASK);
-+ sg->format_offset |= cpu_to_le16(offset);
-+}
-+
-+/**
-+ * dpaa2_sg_get_format() - Get the SG format in SG entry
-+ * @sg: the given scatter-gathering object
-+ *
-+ * Return the format.
-+ */
-+static inline enum dpaa2_sg_format
-+ dpaa2_sg_get_format(const struct dpaa2_sg_entry *sg)
-+{
-+ return (enum dpaa2_sg_format)((le16_to_cpu(sg->format_offset)
-+ >> SG_FORMAT_SHIFT) & SG_FORMAT_MASK);
-+}
-+
-+/**
-+ * dpaa2_sg_set_format() - Set the SG format in SG entry
-+ * @sg: the given scatter-gathering object
-+ * @format: the format to be set
-+ */
-+static inline void dpaa2_sg_set_format(struct dpaa2_sg_entry *sg,
-+ enum dpaa2_sg_format format)
-+{
-+ sg->format_offset &= cpu_to_le16(~(SG_FORMAT_MASK << SG_FORMAT_SHIFT));
-+ sg->format_offset |= cpu_to_le16(format << SG_FORMAT_SHIFT);
-+}
-+
-+/**
-+ * dpaa2_sg_get_bpid() - Get the buffer pool id in SG entry
-+ * @sg: the given scatter-gathering object
-+ *
-+ * Return the bpid.
-+ */
-+static inline u16 dpaa2_sg_get_bpid(const struct dpaa2_sg_entry *sg)
-+{
-+ return le16_to_cpu(sg->bpid) & SG_BPID_MASK;
-+}
-+
-+/**
-+ * dpaa2_sg_set_bpid() - Set the buffer pool id in SG entry
-+ * @sg: the given scatter-gathering object
-+ * @bpid: the bpid to be set
-+ */
-+static inline void dpaa2_sg_set_bpid(struct dpaa2_sg_entry *sg, u16 bpid)
-+{
-+ sg->bpid &= cpu_to_le16(~(SG_BPID_MASK));
-+ sg->bpid |= cpu_to_le16(bpid);
-+}
-+
-+/**
-+ * dpaa2_sg_is_final() - Check final bit in SG entry
-+ * @sg: the given scatter-gathering object
-+ *
-+ * Return bool.
-+ */
-+static inline bool dpaa2_sg_is_final(const struct dpaa2_sg_entry *sg)
-+{
-+ return !!(le16_to_cpu(sg->format_offset) >> SG_FINAL_FLAG_SHIFT);
-+}
-+
-+/**
-+ * dpaa2_sg_set_final() - Set the final bit in SG entry
-+ * @sg: the given scatter-gathering object
-+ * @final: the final boolean to be set
-+ */
-+static inline void dpaa2_sg_set_final(struct dpaa2_sg_entry *sg, bool final)
-+{
-+ sg->format_offset &= cpu_to_le16((~(SG_FINAL_FLAG_MASK
-+ << SG_FINAL_FLAG_SHIFT)) & 0xFFFF);
-+ sg->format_offset |= cpu_to_le16(final << SG_FINAL_FLAG_SHIFT);
-+}
-+
-+/**
-+ * struct dpaa2_fl_entry - structure for frame list entry.
-+ * @addr: address in the FLE
-+ * @len: length in the FLE
-+ * @bpid: buffer pool ID
-+ * @format_offset: format, offset, and short-length fields
-+ * @frc: frame context
-+ * @ctrl: control bits...including pta, pvt1, pvt2, err, etc
-+ * @flc: flow context address
-+ */
-+struct dpaa2_fl_entry {
-+ __le64 addr;
-+ __le32 len;
-+ __le16 bpid;
-+ __le16 format_offset;
-+ __le32 frc;
-+ __le32 ctrl;
-+ __le64 flc;
-+};
-+
-+enum dpaa2_fl_format {
-+ dpaa2_fl_single = 0,
-+ dpaa2_fl_res,
-+ dpaa2_fl_sg
-+};
-+
-+/**
-+ * dpaa2_fl_get_addr() - get the addr field of FLE
-+ * @fle: the given frame list entry
-+ *
-+ * Return the address in the frame list entry.
-+ */
-+static inline dma_addr_t dpaa2_fl_get_addr(const struct dpaa2_fl_entry *fle)
-+{
-+ return (dma_addr_t)le64_to_cpu(fle->addr);
-+}
-+
-+/**
-+ * dpaa2_fl_set_addr() - Set the addr field of FLE
-+ * @fle: the given frame list entry
-+ * @addr: the address needs to be set in frame list entry
-+ */
-+static inline void dpaa2_fl_set_addr(struct dpaa2_fl_entry *fle,
-+ dma_addr_t addr)
-+{
-+ fle->addr = cpu_to_le64(addr);
-+}
-+
-+/**
-+ * dpaa2_fl_get_frc() - Get the frame context in the FLE
-+ * @fle: the given frame list entry
-+ *
-+ * Return the frame context field in the frame lsit entry.
-+ */
-+static inline u32 dpaa2_fl_get_frc(const struct dpaa2_fl_entry *fle)
-+{
-+ return le32_to_cpu(fle->frc);
-+}
-+
-+/**
-+ * dpaa2_fl_set_frc() - Set the frame context in the FLE
-+ * @fle: the given frame list entry
-+ * @frc: the frame context needs to be set in frame list entry
-+ */
-+static inline void dpaa2_fl_set_frc(struct dpaa2_fl_entry *fle, u32 frc)
-+{
-+ fle->frc = cpu_to_le32(frc);
-+}
-+
-+/**
-+ * dpaa2_fl_get_ctrl() - Get the control bits in the FLE
-+ * @fle: the given frame list entry
-+ *
-+ * Return the control bits field in the frame list entry.
-+ */
-+static inline u32 dpaa2_fl_get_ctrl(const struct dpaa2_fl_entry *fle)
-+{
-+ return le32_to_cpu(fle->ctrl);
-+}
-+
-+/**
-+ * dpaa2_fl_set_ctrl() - Set the control bits in the FLE
-+ * @fle: the given frame list entry
-+ * @ctrl: the control bits to be set in the frame list entry
-+ */
-+static inline void dpaa2_fl_set_ctrl(struct dpaa2_fl_entry *fle, u32 ctrl)
-+{
-+ fle->ctrl = cpu_to_le32(ctrl);
-+}
-+
-+/**
-+ * dpaa2_fl_get_flc() - Get the flow context in the FLE
-+ * @fle: the given frame list entry
-+ *
-+ * Return the flow context in the frame list entry.
-+ */
-+static inline dma_addr_t dpaa2_fl_get_flc(const struct dpaa2_fl_entry *fle)
-+{
-+ return (dma_addr_t)le64_to_cpu(fle->flc);
-+}
-+
-+/**
-+ * dpaa2_fl_set_flc() - Set the flow context field of FLE
-+ * @fle: the given frame list entry
-+ * @flc_addr: the flow context needs to be set in frame list entry
-+ */
-+static inline void dpaa2_fl_set_flc(struct dpaa2_fl_entry *fle,
-+ dma_addr_t flc_addr)
-+{
-+ fle->flc = cpu_to_le64(flc_addr);
-+}
-+
-+static inline bool dpaa2_fl_short_len(const struct dpaa2_fl_entry *fle)
-+{
-+ return !!((le16_to_cpu(fle->format_offset) >>
-+ FL_SHORT_LEN_FLAG_SHIFT) & FL_SHORT_LEN_FLAG_MASK);
-+}
-+
-+/**
-+ * dpaa2_fl_get_len() - Get the length in the FLE
-+ * @fle: the given frame list entry
-+ *
-+ * Return the length field in the frame list entry.
-+ */
-+static inline u32 dpaa2_fl_get_len(const struct dpaa2_fl_entry *fle)
-+{
-+ if (dpaa2_fl_short_len(fle))
-+ return le32_to_cpu(fle->len) & FL_SHORT_LEN_MASK;
-+
-+ return le32_to_cpu(fle->len);
-+}
-+
-+/**
-+ * dpaa2_fl_set_len() - Set the length field of FLE
-+ * @fle: the given frame list entry
-+ * @len: the length needs to be set in frame list entry
-+ */
-+static inline void dpaa2_fl_set_len(struct dpaa2_fl_entry *fle, u32 len)
-+{
-+ fle->len = cpu_to_le32(len);
-+}
-+
-+/**
-+ * dpaa2_fl_get_offset() - Get the offset field in the frame list entry
-+ * @fle: the given frame list entry
-+ *
-+ * Return the offset.
-+ */
-+static inline u16 dpaa2_fl_get_offset(const struct dpaa2_fl_entry *fle)
-+{
-+ return le16_to_cpu(fle->format_offset) & FL_OFFSET_MASK;
-+}
-+
-+/**
-+ * dpaa2_fl_set_offset() - Set the offset field of FLE
-+ * @fle: the given frame list entry
-+ * @offset: the offset needs to be set in frame list entry
-+ */
-+static inline void dpaa2_fl_set_offset(struct dpaa2_fl_entry *fle, u16 offset)
-+{
-+ fle->format_offset &= cpu_to_le16(~FL_OFFSET_MASK);
-+ fle->format_offset |= cpu_to_le16(offset);
-+}
-+
-+/**
-+ * dpaa2_fl_get_format() - Get the format field in the FLE
-+ * @fle: the given frame list entry
-+ *
-+ * Return the format.
-+ */
-+static inline enum dpaa2_fl_format dpaa2_fl_get_format(
-+ const struct dpaa2_fl_entry *fle)
-+{
-+ return (enum dpaa2_fl_format)((le16_to_cpu(fle->format_offset) >>
-+ FL_FORMAT_SHIFT) & FL_FORMAT_MASK);
-+}
-+
-+/**
-+ * dpaa2_fl_set_format() - Set the format field of FLE
-+ * @fle: the given frame list entry
-+ * @format: the format needs to be set in frame list entry
-+ */
-+static inline void dpaa2_fl_set_format(struct dpaa2_fl_entry *fle,
-+ enum dpaa2_fl_format format)
-+{
-+ fle->format_offset &= cpu_to_le16(~(FL_FORMAT_MASK << FL_FORMAT_SHIFT));
-+ fle->format_offset |= cpu_to_le16(format << FL_FORMAT_SHIFT);
-+}
-+
-+/**
-+ * dpaa2_fl_get_bpid() - Get the bpid field in the FLE
-+ * @fle: the given frame list entry
-+ *
-+ * Return the buffer pool id.
-+ */
-+static inline u16 dpaa2_fl_get_bpid(const struct dpaa2_fl_entry *fle)
-+{
-+ return le16_to_cpu(fle->bpid) & FL_BPID_MASK;
-+}
-+
-+/**
-+ * dpaa2_fl_set_bpid() - Set the bpid field of FLE
-+ * @fle: the given frame list entry
-+ * @bpid: buffer pool id to be set
-+ */
-+static inline void dpaa2_fl_set_bpid(struct dpaa2_fl_entry *fle, u16 bpid)
-+{
-+ fle->bpid &= cpu_to_le16(~(FL_BPID_MASK));
-+ fle->bpid |= cpu_to_le16(bpid);
-+}
-+
-+/**
-+ * dpaa2_fl_is_final() - Check final bit in FLE
-+ * @fle: the given frame list entry
-+ *
-+ * Return bool.
-+ */
-+static inline bool dpaa2_fl_is_final(const struct dpaa2_fl_entry *fle)
-+{
-+ return !!(le16_to_cpu(fle->format_offset) >> FL_FINAL_FLAG_SHIFT);
-+}
-+
-+/**
-+ * dpaa2_fl_set_final() - Set the final bit in FLE
-+ * @fle: the given frame list entry
-+ * @final: the final boolean to be set
-+ */
-+static inline void dpaa2_fl_set_final(struct dpaa2_fl_entry *fle, bool final)
-+{
-+ fle->format_offset &= cpu_to_le16(~(FL_FINAL_FLAG_MASK <<
-+ FL_FINAL_FLAG_SHIFT));
-+ fle->format_offset |= cpu_to_le16(final << FL_FINAL_FLAG_SHIFT);
-+}
-+
-+#endif /* __FSL_DPAA2_FD_H */
---- /dev/null
-+++ b/drivers/staging/fsl-mc/include/dpaa2-global.h
-@@ -0,0 +1,177 @@
-+/* SPDX-License-Identifier: (GPL-2.0+ OR BSD-3-Clause) */
-+/*
-+ * Copyright 2014-2016 Freescale Semiconductor Inc.
-+ * Copyright 2016 NXP
-+ *
-+ */
-+#ifndef __FSL_DPAA2_GLOBAL_H
-+#define __FSL_DPAA2_GLOBAL_H
-+
-+#include <linux/types.h>
-+#include <linux/cpumask.h>
-+#include "dpaa2-fd.h"
-+
-+struct dpaa2_dq {
-+ union {
-+ struct common {
-+ u8 verb;
-+ u8 reserved[63];
-+ } common;
-+ struct dq {
-+ u8 verb;
-+ u8 stat;
-+ __le16 seqnum;
-+ __le16 oprid;
-+ u8 reserved;
-+ u8 tok;
-+ __le32 fqid;
-+ u32 reserved2;
-+ __le32 fq_byte_cnt;
-+ __le32 fq_frm_cnt;
-+ __le64 fqd_ctx;
-+ u8 fd[32];
-+ } dq;
-+ struct scn {
-+ u8 verb;
-+ u8 stat;
-+ u8 state;
-+ u8 reserved;
-+ __le32 rid_tok;
-+ __le64 ctx;
-+ } scn;
-+ };
-+};
-+
-+/* Parsing frame dequeue results */
-+/* FQ empty */
-+#define DPAA2_DQ_STAT_FQEMPTY 0x80
-+/* FQ held active */
-+#define DPAA2_DQ_STAT_HELDACTIVE 0x40
-+/* FQ force eligible */
-+#define DPAA2_DQ_STAT_FORCEELIGIBLE 0x20
-+/* valid frame */
-+#define DPAA2_DQ_STAT_VALIDFRAME 0x10
-+/* FQ ODP enable */
-+#define DPAA2_DQ_STAT_ODPVALID 0x04
-+/* volatile dequeue */
-+#define DPAA2_DQ_STAT_VOLATILE 0x02
-+/* volatile dequeue command is expired */
-+#define DPAA2_DQ_STAT_EXPIRED 0x01
-+
-+#define DQ_FQID_MASK 0x00FFFFFF
-+#define DQ_FRAME_COUNT_MASK 0x00FFFFFF
-+
-+/**
-+ * dpaa2_dq_flags() - Get the stat field of dequeue response
-+ * @dq: the dequeue result.
-+ */
-+static inline u32 dpaa2_dq_flags(const struct dpaa2_dq *dq)
-+{
-+ return dq->dq.stat;
-+}
-+
-+/**
-+ * dpaa2_dq_is_pull() - Check whether the dq response is from a pull
-+ * command.
-+ * @dq: the dequeue result
-+ *
-+ * Return 1 for volatile(pull) dequeue, 0 for static dequeue.
-+ */
-+static inline int dpaa2_dq_is_pull(const struct dpaa2_dq *dq)
-+{
-+ return (int)(dpaa2_dq_flags(dq) & DPAA2_DQ_STAT_VOLATILE);
-+}
-+
-+/**
-+ * dpaa2_dq_is_pull_complete() - Check whether the pull command is completed.
-+ * @dq: the dequeue result
-+ *
-+ * Return boolean.
-+ */
-+static inline bool dpaa2_dq_is_pull_complete(const struct dpaa2_dq *dq)
-+{
-+ return !!(dpaa2_dq_flags(dq) & DPAA2_DQ_STAT_EXPIRED);
-+}
-+
-+/**
-+ * dpaa2_dq_seqnum() - Get the seqnum field in dequeue response
-+ * @dq: the dequeue result
-+ *
-+ * seqnum is valid only if VALIDFRAME flag is TRUE
-+ *
-+ * Return seqnum.
-+ */
-+static inline u16 dpaa2_dq_seqnum(const struct dpaa2_dq *dq)
-+{
-+ return le16_to_cpu(dq->dq.seqnum);
-+}
-+
-+/**
-+ * dpaa2_dq_odpid() - Get the odpid field in dequeue response
-+ * @dq: the dequeue result
-+ *
-+ * odpid is valid only if ODPVALID flag is TRUE.
-+ *
-+ * Return odpid.
-+ */
-+static inline u16 dpaa2_dq_odpid(const struct dpaa2_dq *dq)
-+{
-+ return le16_to_cpu(dq->dq.oprid);
-+}
-+
-+/**
-+ * dpaa2_dq_fqid() - Get the fqid in dequeue response
-+ * @dq: the dequeue result
-+ *
-+ * Return fqid.
-+ */
-+static inline u32 dpaa2_dq_fqid(const struct dpaa2_dq *dq)
-+{
-+ return le32_to_cpu(dq->dq.fqid) & DQ_FQID_MASK;
-+}
-+
-+/**
-+ * dpaa2_dq_byte_count() - Get the byte count in dequeue response
-+ * @dq: the dequeue result
-+ *
-+ * Return the byte count remaining in the FQ.
-+ */
-+static inline u32 dpaa2_dq_byte_count(const struct dpaa2_dq *dq)
-+{
-+ return le32_to_cpu(dq->dq.fq_byte_cnt);
-+}
-+
-+/**
-+ * dpaa2_dq_frame_count() - Get the frame count in dequeue response
-+ * @dq: the dequeue result
-+ *
-+ * Return the frame count remaining in the FQ.
-+ */
-+static inline u32 dpaa2_dq_frame_count(const struct dpaa2_dq *dq)
-+{
-+ return le32_to_cpu(dq->dq.fq_frm_cnt) & DQ_FRAME_COUNT_MASK;
-+}
-+
-+/**
-+ * dpaa2_dq_fd_ctx() - Get the frame queue context in dequeue response
-+ * @dq: the dequeue result
-+ *
-+ * Return the frame queue context.
-+ */
-+static inline u64 dpaa2_dq_fqd_ctx(const struct dpaa2_dq *dq)
-+{
-+ return le64_to_cpu(dq->dq.fqd_ctx);
-+}
-+
-+/**
-+ * dpaa2_dq_fd() - Get the frame descriptor in dequeue response
-+ * @dq: the dequeue result
-+ *
-+ * Return the frame descriptor.
-+ */
-+static inline const struct dpaa2_fd *dpaa2_dq_fd(const struct dpaa2_dq *dq)
-+{
-+ return (const struct dpaa2_fd *)&dq->dq.fd[0];
-+}
-+
-+#endif /* __FSL_DPAA2_GLOBAL_H */
---- /dev/null
-+++ b/drivers/staging/fsl-mc/include/dpaa2-io.h
-@@ -0,0 +1,178 @@
-+/* SPDX-License-Identifier: (GPL-2.0+ OR BSD-3-Clause) */
-+/*
-+ * Copyright 2014-2016 Freescale Semiconductor Inc.
-+ * Copyright 2017 NXP
-+ *
-+ */
-+#ifndef __FSL_DPAA2_IO_H
-+#define __FSL_DPAA2_IO_H
-+
-+#include <linux/types.h>
-+#include <linux/cpumask.h>
-+
-+#include "dpaa2-fd.h"
-+#include "dpaa2-global.h"
-+
-+struct dpaa2_io;
-+struct dpaa2_io_store;
-+struct device;
-+
-+/**
-+ * DOC: DPIO Service
-+ *
-+ * The DPIO service provides APIs for users to interact with the datapath
-+ * by enqueueing and dequeing frame descriptors.
-+ *
-+ * The following set of APIs can be used to enqueue and dequeue frames
-+ * as well as producing notification callbacks when data is available
-+ * for dequeue.
-+ */
-+
-+#define DPAA2_IO_ANY_CPU -1
-+
-+/**
-+ * struct dpaa2_io_desc - The DPIO descriptor
-+ * @receives_notifications: Use notificaton mode. Non-zero if the DPIO
-+ * has a channel.
-+ * @has_8prio: Set to non-zero for channel with 8 priority WQs. Ignored
-+ * unless receives_notification is TRUE.
-+ * @cpu: The cpu index that at least interrupt handlers will
-+ * execute on.
-+ * @stash_affinity: The stash affinity for this portal favour 'cpu'
-+ * @regs_cena: The cache enabled regs.
-+ * @regs_cinh: The cache inhibited regs
-+ * @dpio_id: The dpio index
-+ * @qman_version: The qman version
-+ *
-+ * Describes the attributes and features of the DPIO object.
-+ */
-+struct dpaa2_io_desc {
-+ int receives_notifications;
-+ int has_8prio;
-+ int cpu;
-+ void *regs_cena;
-+ void *regs_cinh;
-+ int dpio_id;
-+ u32 qman_version;
-+};
-+
-+struct dpaa2_io *dpaa2_io_create(const struct dpaa2_io_desc *desc);
-+
-+void dpaa2_io_down(struct dpaa2_io *d);
-+
-+irqreturn_t dpaa2_io_irq(struct dpaa2_io *obj);
-+
-+struct dpaa2_io *dpaa2_io_service_select(int cpu);
-+
-+/**
-+ * struct dpaa2_io_notification_ctx - The DPIO notification context structure
-+ * @cb: The callback to be invoked when the notification arrives
-+ * @is_cdan: Zero for FQDAN, non-zero for CDAN
-+ * @id: FQID or channel ID, needed for rearm
-+ * @desired_cpu: The cpu on which the notifications will show up. Use
-+ * DPAA2_IO_ANY_CPU if don't care
-+ * @dpio_id: The dpio index
-+ * @qman64: The 64-bit context value shows up in the FQDAN/CDAN.
-+ * @node: The list node
-+ * @dpio_private: The dpio object internal to dpio_service
-+ *
-+ * Used when a FQDAN/CDAN registration is made by drivers.
-+ */
-+struct dpaa2_io_notification_ctx {
-+ void (*cb)(struct dpaa2_io_notification_ctx *ctx);
-+ int is_cdan;
-+ u32 id;
-+ int desired_cpu;
-+ int dpio_id;
-+ u64 qman64;
-+ struct list_head node;
-+ void *dpio_private;
-+};
-+
-+int dpaa2_io_service_register(struct dpaa2_io *service,
-+ struct dpaa2_io_notification_ctx *ctx);
-+void dpaa2_io_service_deregister(struct dpaa2_io *service,
-+ struct dpaa2_io_notification_ctx *ctx);
-+int dpaa2_io_service_rearm(struct dpaa2_io *service,
-+ struct dpaa2_io_notification_ctx *ctx);
-+
-+int dpaa2_io_service_pull_fq(struct dpaa2_io *d, u32 fqid,
-+ struct dpaa2_io_store *s);
-+int dpaa2_io_service_pull_channel(struct dpaa2_io *d, u32 channelid,
-+ struct dpaa2_io_store *s);
-+
-+int dpaa2_io_service_enqueue_fq(struct dpaa2_io *d, u32 fqid,
-+ const struct dpaa2_fd *fd);
-+int dpaa2_io_service_enqueue_qd(struct dpaa2_io *d, u32 qdid, u8 prio,
-+ u16 qdbin, const struct dpaa2_fd *fd);
-+int dpaa2_io_service_release(struct dpaa2_io *d, u32 bpid,
-+ const u64 *buffers, unsigned int num_buffers);
-+int dpaa2_io_service_acquire(struct dpaa2_io *d, u32 bpid,
-+ u64 *buffers, unsigned int num_buffers);
-+
-+struct dpaa2_io_store *dpaa2_io_store_create(unsigned int max_frames,
-+ struct device *dev);
-+void dpaa2_io_store_destroy(struct dpaa2_io_store *s);
-+struct dpaa2_dq *dpaa2_io_store_next(struct dpaa2_io_store *s, int *is_last);
-+
-+/* Order Restoration Support */
-+int dpaa2_io_service_enqueue_orp_fq(struct dpaa2_io *d, u32 fqid,
-+ const struct dpaa2_fd *fd, u16 orpid,
-+ u16 seqnum, int last);
-+
-+int dpaa2_io_service_enqueue_orp_qd(struct dpaa2_io *d, u32 qdid, u8 prio,
-+ u16 qdbin, const struct dpaa2_fd *fd,
-+ u16 orpid, u16 seqnum, int last);
-+
-+int dpaa2_io_service_orp_seqnum_drop(struct dpaa2_io *d, u16 orpid,
-+ u16 seqnum);
-+
-+/***************/
-+/* CSCN */
-+/***************/
-+
-+/**
-+ * struct dpaa2_cscn - The CSCN message format
-+ * @verb: identifies the type of message (should be 0x27).
-+ * @stat: status bits related to dequeuing response (not used)
-+ * @state: bit 0 = 0/1 if CG is no/is congested
-+ * @reserved: reserved byte
-+ * @cgid: congest grp ID - the first 16 bits
-+ * @ctx: context data
-+ *
-+ * Congestion management can be implemented in software through
-+ * the use of Congestion State Change Notifications (CSCN). These
-+ * are messages written by DPAA2 hardware to memory whenever the
-+ * instantaneous count (I_CNT field in the CG) exceeds the
-+ * Congestion State (CS) entrance threshold, signifying congestion
-+ * entrance, or when the instantaneous count returns below exit
-+ * threshold, signifying congestion exit. The format of the message
-+ * is given by the dpaa2_cscn structure. Bit 0 of the state field
-+ * represents congestion state written by the hardware.
-+ */
-+struct dpaa2_cscn {
-+ u8 verb;
-+ u8 stat;
-+ u8 state;
-+ u8 reserved;
-+ __le32 cgid;
-+ __le64 ctx;
-+};
-+
-+#define DPAA2_CSCN_SIZE 64
-+#define DPAA2_CSCN_ALIGN 16
-+
-+#define DPAA2_CSCN_STATE_MASK 0x1
-+#define DPAA2_CSCN_CONGESTED 1
-+
-+static inline bool dpaa2_cscn_state_congested(struct dpaa2_cscn *cscn)
-+{
-+ return ((cscn->state & DPAA2_CSCN_STATE_MASK) == DPAA2_CSCN_CONGESTED);
-+}
-+
-+int dpaa2_io_query_fq_count(struct dpaa2_io *d, u32 fqid,
-+ u32 *fcnt, u32 *bcnt);
-+int dpaa2_io_query_bp_count(struct dpaa2_io *d, u32 bpid,
-+ u32 *num);
-+
-+#endif /* __FSL_DPAA2_IO_H */
---- a/drivers/staging/fsl-mc/include/dpbp-cmd.h
-+++ /dev/null
-@@ -1,185 +0,0 @@
--/* Copyright 2013-2016 Freescale Semiconductor Inc.
-- *
-- * Redistribution and use in source and binary forms, with or without
-- * modification, are permitted provided that the following conditions are met:
-- * * Redistributions of source code must retain the above copyright
-- * notice, this list of conditions and the following disclaimer.
-- * * Redistributions in binary form must reproduce the above copyright
-- * notice, this list of conditions and the following disclaimer in the
-- * documentation and/or other materials provided with the distribution.
-- * * Neither the name of the above-listed copyright holders nor the
-- * names of any contributors may be used to endorse or promote products
-- * derived from this software without specific prior written permission.
-- *
-- *
-- * ALTERNATIVELY, this software may be distributed under the terms of the
-- * GNU General Public License ("GPL") as published by the Free Software
-- * Foundation, either version 2 of that License or (at your option) any
-- * later version.
-- *
-- * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
-- * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
-- * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
-- * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE
-- * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
-- * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
-- * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
-- * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
-- * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
-- * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
-- * POSSIBILITY OF SUCH DAMAGE.
-- */
--#ifndef _FSL_DPBP_CMD_H
--#define _FSL_DPBP_CMD_H
--
--/* DPBP Version */
--#define DPBP_VER_MAJOR 2
--#define DPBP_VER_MINOR 2
--
--/* Command IDs */
--#define DPBP_CMDID_CLOSE 0x800
--#define DPBP_CMDID_OPEN 0x804
--#define DPBP_CMDID_CREATE 0x904
--#define DPBP_CMDID_DESTROY 0x900
--
--#define DPBP_CMDID_ENABLE 0x002
--#define DPBP_CMDID_DISABLE 0x003
--#define DPBP_CMDID_GET_ATTR 0x004
--#define DPBP_CMDID_RESET 0x005
--#define DPBP_CMDID_IS_ENABLED 0x006
--
--#define DPBP_CMDID_SET_IRQ 0x010
--#define DPBP_CMDID_GET_IRQ 0x011
--#define DPBP_CMDID_SET_IRQ_ENABLE 0x012
--#define DPBP_CMDID_GET_IRQ_ENABLE 0x013
--#define DPBP_CMDID_SET_IRQ_MASK 0x014
--#define DPBP_CMDID_GET_IRQ_MASK 0x015
--#define DPBP_CMDID_GET_IRQ_STATUS 0x016
--#define DPBP_CMDID_CLEAR_IRQ_STATUS 0x017
--
--#define DPBP_CMDID_SET_NOTIFICATIONS 0x01b0
--#define DPBP_CMDID_GET_NOTIFICATIONS 0x01b1
--
--struct dpbp_cmd_open {
-- __le32 dpbp_id;
--};
--
--#define DPBP_ENABLE 0x1
--
--struct dpbp_rsp_is_enabled {
-- u8 enabled;
--};
--
--struct dpbp_cmd_set_irq {
-- /* cmd word 0 */
-- u8 irq_index;
-- u8 pad[3];
-- __le32 irq_val;
-- /* cmd word 1 */
-- __le64 irq_addr;
-- /* cmd word 2 */
-- __le32 irq_num;
--};
--
--struct dpbp_cmd_get_irq {
-- __le32 pad;
-- u8 irq_index;
--};
--
--struct dpbp_rsp_get_irq {
-- /* response word 0 */
-- __le32 irq_val;
-- __le32 pad;
-- /* response word 1 */
-- __le64 irq_addr;
-- /* response word 2 */
-- __le32 irq_num;
-- __le32 type;
--};
--
--struct dpbp_cmd_set_irq_enable {
-- u8 enable;
-- u8 pad[3];
-- u8 irq_index;
--};
--
--struct dpbp_cmd_get_irq_enable {
-- __le32 pad;
-- u8 irq_index;
--};
--
--struct dpbp_rsp_get_irq_enable {
-- u8 enabled;
--};
--
--struct dpbp_cmd_set_irq_mask {
-- __le32 mask;
-- u8 irq_index;
--};
--
--struct dpbp_cmd_get_irq_mask {
-- __le32 pad;
-- u8 irq_index;
--};
--
--struct dpbp_rsp_get_irq_mask {
-- __le32 mask;
--};
--
--struct dpbp_cmd_get_irq_status {
-- __le32 status;
-- u8 irq_index;
--};
--
--struct dpbp_rsp_get_irq_status {
-- __le32 status;
--};
--
--struct dpbp_cmd_clear_irq_status {
-- __le32 status;
-- u8 irq_index;
--};
--
--struct dpbp_rsp_get_attributes {
-- /* response word 0 */
-- __le16 pad;
-- __le16 bpid;
-- __le32 id;
-- /* response word 1 */
-- __le16 version_major;
-- __le16 version_minor;
--};
--
--struct dpbp_cmd_set_notifications {
-- /* cmd word 0 */
-- __le32 depletion_entry;
-- __le32 depletion_exit;
-- /* cmd word 1 */
-- __le32 surplus_entry;
-- __le32 surplus_exit;
-- /* cmd word 2 */
-- __le16 options;
-- __le16 pad[3];
-- /* cmd word 3 */
-- __le64 message_ctx;
-- /* cmd word 4 */
-- __le64 message_iova;
--};
--
--struct dpbp_rsp_get_notifications {
-- /* response word 0 */
-- __le32 depletion_entry;
-- __le32 depletion_exit;
-- /* response word 1 */
-- __le32 surplus_entry;
-- __le32 surplus_exit;
-- /* response word 2 */
-- __le16 options;
-- __le16 pad[3];
-- /* response word 3 */
-- __le64 message_ctx;
-- /* response word 4 */
-- __le64 message_iova;
--};
--
--#endif /* _FSL_DPBP_CMD_H */
---- a/drivers/staging/fsl-mc/include/dpbp.h
-+++ /dev/null
-@@ -1,220 +0,0 @@
--/* Copyright 2013-2015 Freescale Semiconductor Inc.
-- *
-- * Redistribution and use in source and binary forms, with or without
-- * modification, are permitted provided that the following conditions are met:
-- * * Redistributions of source code must retain the above copyright
-- * notice, this list of conditions and the following disclaimer.
-- * * Redistributions in binary form must reproduce the above copyright
-- * notice, this list of conditions and the following disclaimer in the
-- * documentation and/or other materials provided with the distribution.
-- * * Neither the name of the above-listed copyright holders nor the
-- * names of any contributors may be used to endorse or promote products
-- * derived from this software without specific prior written permission.
-- *
-- *
-- * ALTERNATIVELY, this software may be distributed under the terms of the
-- * GNU General Public License ("GPL") as published by the Free Software
-- * Foundation, either version 2 of that License or (at your option) any
-- * later version.
-- *
-- * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
-- * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
-- * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
-- * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE
-- * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
-- * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
-- * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
-- * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
-- * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
-- * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
-- * POSSIBILITY OF SUCH DAMAGE.
-- */
--#ifndef __FSL_DPBP_H
--#define __FSL_DPBP_H
--
--/* Data Path Buffer Pool API
-- * Contains initialization APIs and runtime control APIs for DPBP
-- */
--
--struct fsl_mc_io;
--
--int dpbp_open(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- int dpbp_id,
-- u16 *token);
--
--int dpbp_close(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token);
--
--/**
-- * struct dpbp_cfg - Structure representing DPBP configuration
-- * @options: place holder
-- */
--struct dpbp_cfg {
-- u32 options;
--};
--
--int dpbp_create(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- const struct dpbp_cfg *cfg,
-- u16 *token);
--
--int dpbp_destroy(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token);
--
--int dpbp_enable(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token);
--
--int dpbp_disable(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token);
--
--int dpbp_is_enabled(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- int *en);
--
--int dpbp_reset(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token);
--
--/**
-- * struct dpbp_irq_cfg - IRQ configuration
-- * @addr: Address that must be written to signal a message-based interrupt
-- * @val: Value to write into irq_addr address
-- * @irq_num: A user defined number associated with this IRQ
-- */
--struct dpbp_irq_cfg {
-- u64 addr;
-- u32 val;
-- int irq_num;
--};
--
--int dpbp_set_irq(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- u8 irq_index,
-- struct dpbp_irq_cfg *irq_cfg);
--
--int dpbp_get_irq(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- u8 irq_index,
-- int *type,
-- struct dpbp_irq_cfg *irq_cfg);
--
--int dpbp_set_irq_enable(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- u8 irq_index,
-- u8 en);
--
--int dpbp_get_irq_enable(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- u8 irq_index,
-- u8 *en);
--
--int dpbp_set_irq_mask(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- u8 irq_index,
-- u32 mask);
--
--int dpbp_get_irq_mask(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- u8 irq_index,
-- u32 *mask);
--
--int dpbp_get_irq_status(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- u8 irq_index,
-- u32 *status);
--
--int dpbp_clear_irq_status(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- u8 irq_index,
-- u32 status);
--
--/**
-- * struct dpbp_attr - Structure representing DPBP attributes
-- * @id: DPBP object ID
-- * @version: DPBP version
-- * @bpid: Hardware buffer pool ID; should be used as an argument in
-- * acquire/release operations on buffers
-- */
--struct dpbp_attr {
-- int id;
-- /**
-- * struct version - Structure representing DPBP version
-- * @major: DPBP major version
-- * @minor: DPBP minor version
-- */
-- struct {
-- u16 major;
-- u16 minor;
-- } version;
-- u16 bpid;
--};
--
--int dpbp_get_attributes(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- struct dpbp_attr *attr);
--
--/**
-- * DPBP notifications options
-- */
--
--/**
-- * BPSCN write will attempt to allocate into a cache (coherent write)
-- */
--#define DPBP_NOTIF_OPT_COHERENT_WRITE 0x00000001
--
--/**
-- * struct dpbp_notification_cfg - Structure representing DPBP notifications
-- * towards software
-- * @depletion_entry: below this threshold the pool is "depleted";
-- * set it to '0' to disable it
-- * @depletion_exit: greater than or equal to this threshold the pool exit its
-- * "depleted" state
-- * @surplus_entry: above this threshold the pool is in "surplus" state;
-- * set it to '0' to disable it
-- * @surplus_exit: less than or equal to this threshold the pool exit its
-- * "surplus" state
-- * @message_iova: MUST be given if either 'depletion_entry' or 'surplus_entry'
-- * is not '0' (enable); I/O virtual address (must be in DMA-able memory),
-- * must be 16B aligned.
-- * @message_ctx: The context that will be part of the BPSCN message and will
-- * be written to 'message_iova'
-- * @options: Mask of available options; use 'DPBP_NOTIF_OPT_<X>' values
-- */
--struct dpbp_notification_cfg {
-- u32 depletion_entry;
-- u32 depletion_exit;
-- u32 surplus_entry;
-- u32 surplus_exit;
-- u64 message_iova;
-- u64 message_ctx;
-- u16 options;
--};
--
--int dpbp_set_notifications(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- struct dpbp_notification_cfg *cfg);
--
--int dpbp_get_notifications(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- struct dpbp_notification_cfg *cfg);
--
--/** @} */
--
--#endif /* __FSL_DPBP_H */
---- a/drivers/staging/fsl-mc/include/dpcon-cmd.h
-+++ /dev/null
-@@ -1,62 +0,0 @@
--/* Copyright 2013-2015 Freescale Semiconductor Inc.
-- *
-- * Redistribution and use in source and binary forms, with or without
-- * modification, are permitted provided that the following conditions are met:
-- * * Redistributions of source code must retain the above copyright
-- * notice, this list of conditions and the following disclaimer.
-- * * Redistributions in binary form must reproduce the above copyright
-- * notice, this list of conditions and the following disclaimer in the
-- * documentation and/or other materials provided with the distribution.
-- * * Neither the name of the above-listed copyright holders nor the
-- * names of any contributors may be used to endorse or promote products
-- * derived from this software without specific prior written permission.
-- *
-- *
-- * ALTERNATIVELY, this software may be distributed under the terms of the
-- * GNU General Public License ("GPL") as published by the Free Software
-- * Foundation, either version 2 of that License or (at your option) any
-- * later version.
-- *
-- * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
-- * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
-- * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
-- * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE
-- * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
-- * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
-- * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
-- * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
-- * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
-- * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
-- * POSSIBILITY OF SUCH DAMAGE.
-- */
--#ifndef _FSL_DPCON_CMD_H
--#define _FSL_DPCON_CMD_H
--
--/* DPCON Version */
--#define DPCON_VER_MAJOR 2
--#define DPCON_VER_MINOR 1
--
--/* Command IDs */
--#define DPCON_CMDID_CLOSE 0x800
--#define DPCON_CMDID_OPEN 0x808
--#define DPCON_CMDID_CREATE 0x908
--#define DPCON_CMDID_DESTROY 0x900
--
--#define DPCON_CMDID_ENABLE 0x002
--#define DPCON_CMDID_DISABLE 0x003
--#define DPCON_CMDID_GET_ATTR 0x004
--#define DPCON_CMDID_RESET 0x005
--#define DPCON_CMDID_IS_ENABLED 0x006
--
--#define DPCON_CMDID_SET_IRQ 0x010
--#define DPCON_CMDID_GET_IRQ 0x011
--#define DPCON_CMDID_SET_IRQ_ENABLE 0x012
--#define DPCON_CMDID_GET_IRQ_ENABLE 0x013
--#define DPCON_CMDID_SET_IRQ_MASK 0x014
--#define DPCON_CMDID_GET_IRQ_MASK 0x015
--#define DPCON_CMDID_GET_IRQ_STATUS 0x016
--#define DPCON_CMDID_CLEAR_IRQ_STATUS 0x017
--
--#define DPCON_CMDID_SET_NOTIFICATION 0x100
--
--#endif /* _FSL_DPCON_CMD_H */
---- a/drivers/staging/fsl-mc/include/dpmng.h
-+++ /dev/null
-@@ -1,69 +0,0 @@
--/* Copyright 2013-2015 Freescale Semiconductor Inc.
-- *
-- * Redistribution and use in source and binary forms, with or without
-- * modification, are permitted provided that the following conditions are met:
-- * * Redistributions of source code must retain the above copyright
-- * notice, this list of conditions and the following disclaimer.
-- * * Redistributions in binary form must reproduce the above copyright
-- * notice, this list of conditions and the following disclaimer in the
-- * documentation and/or other materials provided with the distribution.
-- * * Neither the name of the above-listed copyright holders nor the
-- * names of any contributors may be used to endorse or promote products
-- * derived from this software without specific prior written permission.
-- *
-- *
-- * ALTERNATIVELY, this software may be distributed under the terms of the
-- * GNU General Public License ("GPL") as published by the Free Software
-- * Foundation, either version 2 of that License or (at your option) any
-- * later version.
-- *
-- * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
-- * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
-- * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
-- * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE
-- * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
-- * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
-- * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
-- * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
-- * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
-- * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
-- * POSSIBILITY OF SUCH DAMAGE.
-- */
--#ifndef __FSL_DPMNG_H
--#define __FSL_DPMNG_H
--
--/* Management Complex General API
-- * Contains general API for the Management Complex firmware
-- */
--
--struct fsl_mc_io;
--
--/**
-- * Management Complex firmware version information
-- */
--#define MC_VER_MAJOR 8
--#define MC_VER_MINOR 0
--
--/**
-- * struct mc_version
-- * @major: Major version number: incremented on API compatibility changes
-- * @minor: Minor version number: incremented on API additions (that are
-- * backward compatible); reset when major version is incremented
-- * @revision: Internal revision number: incremented on implementation changes
-- * and/or bug fixes that have no impact on API
-- */
--struct mc_version {
-- u32 major;
-- u32 minor;
-- u32 revision;
--};
--
--int mc_get_version(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- struct mc_version *mc_ver_info);
--
--int dpmng_get_container_id(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- int *container_id);
--
--#endif /* __FSL_DPMNG_H */
---- /dev/null
-+++ b/drivers/staging/fsl-mc/include/dpopr.h
-@@ -0,0 +1,112 @@
-+/*
-+ * Copyright 2017 NXP
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of the above-listed copyright holders nor the
-+ * names of any contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
-+ * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
-+ * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
-+ * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE
-+ * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
-+ * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
-+ * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
-+ * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
-+ * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
-+ * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
-+ * POSSIBILITY OF SUCH DAMAGE.
-+ */
-+#ifndef __FSL_DPOPR_H_
-+#define __FSL_DPOPR_H_
-+
-+#include <linux/types.h>
-+
-+/* Data Path Order Restoration API
-+ * Contains initialization APIs and runtime APIs for the Order Restoration
-+ */
-+
-+/** Order Restoration properties */
-+
-+/**
-+ * Create a new Order Point Record option
-+ */
-+#define OPR_OPT_CREATE 0x1
-+/**
-+ * Retire an existing Order Point Record option
-+ */
-+#define OPR_OPT_RETIRE 0x2
-+
-+/**
-+ * struct opr_cfg - Structure representing OPR configuration
-+ * @oprrws: Order point record (OPR) restoration window size (0 to 5)
-+ * 0 - Window size is 32 frames.
-+ * 1 - Window size is 64 frames.
-+ * 2 - Window size is 128 frames.
-+ * 3 - Window size is 256 frames.
-+ * 4 - Window size is 512 frames.
-+ * 5 - Window size is 1024 frames.
-+ * @oa: OPR auto advance NESN window size (0 disabled, 1 enabled)
-+ * @olws: OPR acceptable late arrival window size (0 to 3)
-+ * 0 - Disabled. Late arrivals are always rejected.
-+ * 1 - Window size is 32 frames.
-+ * 2 - Window size is the same as the OPR restoration
-+ * window size configured in the OPRRWS field.
-+ * 3 - Window size is 8192 frames. Late arrivals are
-+ * always accepted.
-+ * @oeane: Order restoration list (ORL) resource exhaustion
-+ * advance NESN enable (0 disabled, 1 enabled)
-+ * @oloe: OPR loose ordering enable (0 disabled, 1 enabled)
-+ */
-+struct opr_cfg {
-+ u8 oprrws;
-+ u8 oa;
-+ u8 olws;
-+ u8 oeane;
-+ u8 oloe;
-+};
-+
-+/**
-+ * struct opr_qry - Structure representing OPR configuration
-+ * @enable: Enabled state
-+ * @rip: Retirement In Progress
-+ * @ndsn: Next dispensed sequence number
-+ * @nesn: Next expected sequence number
-+ * @ea_hseq: Early arrival head sequence number
-+ * @hseq_nlis: HSEQ not last in sequence
-+ * @ea_tseq: Early arrival tail sequence number
-+ * @tseq_nlis: TSEQ not last in sequence
-+ * @ea_tptr: Early arrival tail pointer
-+ * @ea_hptr: Early arrival head pointer
-+ * @opr_id: Order Point Record ID
-+ * @opr_vid: Order Point Record Virtual ID
-+ */
-+struct opr_qry {
-+ char enable;
-+ char rip;
-+ u16 ndsn;
-+ u16 nesn;
-+ u16 ea_hseq;
-+ char hseq_nlis;
-+ u16 ea_tseq;
-+ char tseq_nlis;
-+ u16 ea_tptr;
-+ u16 ea_hptr;
-+ u16 opr_id;
-+ u16 opr_vid;
-+};
-+
-+#endif /* __FSL_DPOPR_H_ */
---- a/drivers/staging/fsl-mc/include/dprc.h
-+++ /dev/null
-@@ -1,544 +0,0 @@
--/* Copyright 2013-2015 Freescale Semiconductor Inc.
-- *
-- * Redistribution and use in source and binary forms, with or without
-- * modification, are permitted provided that the following conditions are met:
-- * * Redistributions of source code must retain the above copyright
-- * notice, this list of conditions and the following disclaimer.
-- * * Redistributions in binary form must reproduce the above copyright
-- * notice, this list of conditions and the following disclaimer in the
-- * documentation and/or other materials provided with the distribution.
-- * * Neither the name of the above-listed copyright holders nor the
-- * names of any contributors may be used to endorse or promote products
-- * derived from this software without specific prior written permission.
-- *
-- *
-- * ALTERNATIVELY, this software may be distributed under the terms of the
-- * GNU General Public License ("GPL") as published by the Free Software
-- * Foundation, either version 2 of that License or (at your option) any
-- * later version.
-- *
-- * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
-- * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
-- * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
-- * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE
-- * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
-- * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
-- * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
-- * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
-- * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
-- * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
-- * POSSIBILITY OF SUCH DAMAGE.
-- */
--#ifndef _FSL_DPRC_H
--#define _FSL_DPRC_H
--
--#include "mc-cmd.h"
--
--/* Data Path Resource Container API
-- * Contains DPRC API for managing and querying DPAA resources
-- */
--
--struct fsl_mc_io;
--
--/**
-- * Set this value as the icid value in dprc_cfg structure when creating a
-- * container, in case the ICID is not selected by the user and should be
-- * allocated by the DPRC from the pool of ICIDs.
-- */
--#define DPRC_GET_ICID_FROM_POOL (u16)(~(0))
--
--/**
-- * Set this value as the portal_id value in dprc_cfg structure when creating a
-- * container, in case the portal ID is not specifically selected by the
-- * user and should be allocated by the DPRC from the pool of portal ids.
-- */
--#define DPRC_GET_PORTAL_ID_FROM_POOL (int)(~(0))
--
--int dprc_open(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- int container_id,
-- u16 *token);
--
--int dprc_close(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token);
--
--/**
-- * Container general options
-- *
-- * These options may be selected at container creation by the container creator
-- * and can be retrieved using dprc_get_attributes()
-- */
--
--/* Spawn Policy Option allowed - Indicates that the new container is allowed
-- * to spawn and have its own child containers.
-- */
--#define DPRC_CFG_OPT_SPAWN_ALLOWED 0x00000001
--
--/* General Container allocation policy - Indicates that the new container is
-- * allowed to allocate requested resources from its parent container; if not
-- * set, the container is only allowed to use resources in its own pools; Note
-- * that this is a container's global policy, but the parent container may
-- * override it and set specific quota per resource type.
-- */
--#define DPRC_CFG_OPT_ALLOC_ALLOWED 0x00000002
--
--/* Object initialization allowed - software context associated with this
-- * container is allowed to invoke object initialization operations.
-- */
--#define DPRC_CFG_OPT_OBJ_CREATE_ALLOWED 0x00000004
--
--/* Topology change allowed - software context associated with this
-- * container is allowed to invoke topology operations, such as attach/detach
-- * of network objects.
-- */
--#define DPRC_CFG_OPT_TOPOLOGY_CHANGES_ALLOWED 0x00000008
--
--/* AIOP - Indicates that container belongs to AIOP. */
--#define DPRC_CFG_OPT_AIOP 0x00000020
--
--/* IRQ Config - Indicates that the container allowed to configure its IRQs. */
--#define DPRC_CFG_OPT_IRQ_CFG_ALLOWED 0x00000040
--
--/**
-- * struct dprc_cfg - Container configuration options
-- * @icid: Container's ICID; if set to 'DPRC_GET_ICID_FROM_POOL', a free
-- * ICID value is allocated by the DPRC
-- * @portal_id: Portal ID; if set to 'DPRC_GET_PORTAL_ID_FROM_POOL', a free
-- * portal ID is allocated by the DPRC
-- * @options: Combination of 'DPRC_CFG_OPT_<X>' options
-- * @label: Object's label
-- */
--struct dprc_cfg {
-- u16 icid;
-- int portal_id;
-- u64 options;
-- char label[16];
--};
--
--int dprc_create_container(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- struct dprc_cfg *cfg,
-- int *child_container_id,
-- u64 *child_portal_offset);
--
--int dprc_destroy_container(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- int child_container_id);
--
--int dprc_reset_container(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- int child_container_id);
--
--/* IRQ */
--
--/* IRQ index */
--#define DPRC_IRQ_INDEX 0
--
--/* Number of dprc's IRQs */
--#define DPRC_NUM_OF_IRQS 1
--
--/* DPRC IRQ events */
--
--/* IRQ event - Indicates that a new object added to the container */
--#define DPRC_IRQ_EVENT_OBJ_ADDED 0x00000001
--/* IRQ event - Indicates that an object was removed from the container */
--#define DPRC_IRQ_EVENT_OBJ_REMOVED 0x00000002
--/* IRQ event - Indicates that resources added to the container */
--#define DPRC_IRQ_EVENT_RES_ADDED 0x00000004
--/* IRQ event - Indicates that resources removed from the container */
--#define DPRC_IRQ_EVENT_RES_REMOVED 0x00000008
--/* IRQ event - Indicates that one of the descendant containers that opened by
-- * this container is destroyed
-- */
--#define DPRC_IRQ_EVENT_CONTAINER_DESTROYED 0x00000010
--
--/* IRQ event - Indicates that on one of the container's opened object is
-- * destroyed
-- */
--#define DPRC_IRQ_EVENT_OBJ_DESTROYED 0x00000020
--
--/* Irq event - Indicates that object is created at the container */
--#define DPRC_IRQ_EVENT_OBJ_CREATED 0x00000040
--
--/**
-- * struct dprc_irq_cfg - IRQ configuration
-- * @paddr: Address that must be written to signal a message-based interrupt
-- * @val: Value to write into irq_addr address
-- * @irq_num: A user defined number associated with this IRQ
-- */
--struct dprc_irq_cfg {
-- phys_addr_t paddr;
-- u32 val;
-- int irq_num;
--};
--
--int dprc_set_irq(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- u8 irq_index,
-- struct dprc_irq_cfg *irq_cfg);
--
--int dprc_get_irq(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- u8 irq_index,
-- int *type,
-- struct dprc_irq_cfg *irq_cfg);
--
--int dprc_set_irq_enable(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- u8 irq_index,
-- u8 en);
--
--int dprc_get_irq_enable(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- u8 irq_index,
-- u8 *en);
--
--int dprc_set_irq_mask(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- u8 irq_index,
-- u32 mask);
--
--int dprc_get_irq_mask(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- u8 irq_index,
-- u32 *mask);
--
--int dprc_get_irq_status(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- u8 irq_index,
-- u32 *status);
--
--int dprc_clear_irq_status(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- u8 irq_index,
-- u32 status);
--
--/**
-- * struct dprc_attributes - Container attributes
-- * @container_id: Container's ID
-- * @icid: Container's ICID
-- * @portal_id: Container's portal ID
-- * @options: Container's options as set at container's creation
-- * @version: DPRC version
-- */
--struct dprc_attributes {
-- int container_id;
-- u16 icid;
-- int portal_id;
-- u64 options;
-- /**
-- * struct version - DPRC version
-- * @major: DPRC major version
-- * @minor: DPRC minor version
-- */
-- struct {
-- u16 major;
-- u16 minor;
-- } version;
--};
--
--int dprc_get_attributes(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- struct dprc_attributes *attributes);
--
--int dprc_set_res_quota(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- int child_container_id,
-- char *type,
-- u16 quota);
--
--int dprc_get_res_quota(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- int child_container_id,
-- char *type,
-- u16 *quota);
--
--/* Resource request options */
--
--/* Explicit resource ID request - The requested objects/resources
-- * are explicit and sequential (in case of resources).
-- * The base ID is given at res_req at base_align field
-- */
--#define DPRC_RES_REQ_OPT_EXPLICIT 0x00000001
--
--/* Aligned resources request - Relevant only for resources
-- * request (and not objects). Indicates that resources base ID should be
-- * sequential and aligned to the value given at dprc_res_req base_align field
-- */
--#define DPRC_RES_REQ_OPT_ALIGNED 0x00000002
--
--/* Plugged Flag - Relevant only for object assignment request.
-- * Indicates that after all objects assigned. An interrupt will be invoked at
-- * the relevant GPP. The assigned object will be marked as plugged.
-- * plugged objects can't be assigned from their container
-- */
--#define DPRC_RES_REQ_OPT_PLUGGED 0x00000004
--
--/**
-- * struct dprc_res_req - Resource request descriptor, to be used in assignment
-- * or un-assignment of resources and objects.
-- * @type: Resource/object type: Represent as a NULL terminated string.
-- * This string may received by using dprc_get_pool() to get resource
-- * type and dprc_get_obj() to get object type;
-- * Note: it is not possible to assign/un-assign DPRC objects
-- * @num: Number of resources
-- * @options: Request options: combination of DPRC_RES_REQ_OPT_ options
-- * @id_base_align: In case of explicit assignment (DPRC_RES_REQ_OPT_EXPLICIT
-- * is set at option), this field represents the required base ID
-- * for resource allocation; In case of aligned assignment
-- * (DPRC_RES_REQ_OPT_ALIGNED is set at option), this field
-- * indicates the required alignment for the resource ID(s) -
-- * use 0 if there is no alignment or explicit ID requirements
-- */
--struct dprc_res_req {
-- char type[16];
-- u32 num;
-- u32 options;
-- int id_base_align;
--};
--
--int dprc_assign(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- int container_id,
-- struct dprc_res_req *res_req);
--
--int dprc_unassign(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- int child_container_id,
-- struct dprc_res_req *res_req);
--
--int dprc_get_pool_count(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- int *pool_count);
--
--int dprc_get_pool(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- int pool_index,
-- char *type);
--
--int dprc_get_obj_count(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- int *obj_count);
--
--/* Objects Attributes Flags */
--
--/* Opened state - Indicates that an object is open by at least one owner */
--#define DPRC_OBJ_STATE_OPEN 0x00000001
--/* Plugged state - Indicates that the object is plugged */
--#define DPRC_OBJ_STATE_PLUGGED 0x00000002
--
--/**
-- * Shareability flag - Object flag indicating no memory shareability.
-- * the object generates memory accesses that are non coherent with other
-- * masters;
-- * user is responsible for proper memory handling through IOMMU configuration.
-- */
--#define DPRC_OBJ_FLAG_NO_MEM_SHAREABILITY 0x0001
--
--/**
-- * struct dprc_obj_desc - Object descriptor, returned from dprc_get_obj()
-- * @type: Type of object: NULL terminated string
-- * @id: ID of logical object resource
-- * @vendor: Object vendor identifier
-- * @ver_major: Major version number
-- * @ver_minor: Minor version number
-- * @irq_count: Number of interrupts supported by the object
-- * @region_count: Number of mappable regions supported by the object
-- * @state: Object state: combination of DPRC_OBJ_STATE_ states
-- * @label: Object label
-- * @flags: Object's flags
-- */
--struct dprc_obj_desc {
-- char type[16];
-- int id;
-- u16 vendor;
-- u16 ver_major;
-- u16 ver_minor;
-- u8 irq_count;
-- u8 region_count;
-- u32 state;
-- char label[16];
-- u16 flags;
--};
--
--int dprc_get_obj(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- int obj_index,
-- struct dprc_obj_desc *obj_desc);
--
--int dprc_get_obj_desc(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- char *obj_type,
-- int obj_id,
-- struct dprc_obj_desc *obj_desc);
--
--int dprc_set_obj_irq(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- char *obj_type,
-- int obj_id,
-- u8 irq_index,
-- struct dprc_irq_cfg *irq_cfg);
--
--int dprc_get_obj_irq(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- char *obj_type,
-- int obj_id,
-- u8 irq_index,
-- int *type,
-- struct dprc_irq_cfg *irq_cfg);
--
--int dprc_get_res_count(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- char *type,
-- int *res_count);
--
--/**
-- * enum dprc_iter_status - Iteration status
-- * @DPRC_ITER_STATUS_FIRST: Perform first iteration
-- * @DPRC_ITER_STATUS_MORE: Indicates more/next iteration is needed
-- * @DPRC_ITER_STATUS_LAST: Indicates last iteration
-- */
--enum dprc_iter_status {
-- DPRC_ITER_STATUS_FIRST = 0,
-- DPRC_ITER_STATUS_MORE = 1,
-- DPRC_ITER_STATUS_LAST = 2
--};
--
--/**
-- * struct dprc_res_ids_range_desc - Resource ID range descriptor
-- * @base_id: Base resource ID of this range
-- * @last_id: Last resource ID of this range
-- * @iter_status: Iteration status - should be set to DPRC_ITER_STATUS_FIRST at
-- * first iteration; while the returned marker is DPRC_ITER_STATUS_MORE,
-- * additional iterations are needed, until the returned marker is
-- * DPRC_ITER_STATUS_LAST
-- */
--struct dprc_res_ids_range_desc {
-- int base_id;
-- int last_id;
-- enum dprc_iter_status iter_status;
--};
--
--int dprc_get_res_ids(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- char *type,
-- struct dprc_res_ids_range_desc *range_desc);
--
--/* Region flags */
--/* Cacheable - Indicates that region should be mapped as cacheable */
--#define DPRC_REGION_CACHEABLE 0x00000001
--
--/**
-- * enum dprc_region_type - Region type
-- * @DPRC_REGION_TYPE_MC_PORTAL: MC portal region
-- * @DPRC_REGION_TYPE_QBMAN_PORTAL: Qbman portal region
-- */
--enum dprc_region_type {
-- DPRC_REGION_TYPE_MC_PORTAL,
-- DPRC_REGION_TYPE_QBMAN_PORTAL
--};
--
--/**
-- * struct dprc_region_desc - Mappable region descriptor
-- * @base_offset: Region offset from region's base address.
-- * For DPMCP and DPRC objects, region base is offset from SoC MC portals
-- * base address; For DPIO, region base is offset from SoC QMan portals
-- * base address
-- * @size: Region size (in bytes)
-- * @flags: Region attributes
-- * @type: Portal region type
-- */
--struct dprc_region_desc {
-- u32 base_offset;
-- u32 size;
-- u32 flags;
-- enum dprc_region_type type;
--};
--
--int dprc_get_obj_region(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- char *obj_type,
-- int obj_id,
-- u8 region_index,
-- struct dprc_region_desc *region_desc);
--
--int dprc_set_obj_label(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- char *obj_type,
-- int obj_id,
-- char *label);
--
--/**
-- * struct dprc_endpoint - Endpoint description for link connect/disconnect
-- * operations
-- * @type: Endpoint object type: NULL terminated string
-- * @id: Endpoint object ID
-- * @if_id: Interface ID; should be set for endpoints with multiple
-- * interfaces ("dpsw", "dpdmux"); for others, always set to 0
-- */
--struct dprc_endpoint {
-- char type[16];
-- int id;
-- int if_id;
--};
--
--/**
-- * struct dprc_connection_cfg - Connection configuration.
-- * Used for virtual connections only
-- * @committed_rate: Committed rate (Mbits/s)
-- * @max_rate: Maximum rate (Mbits/s)
-- */
--struct dprc_connection_cfg {
-- u32 committed_rate;
-- u32 max_rate;
--};
--
--int dprc_connect(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- const struct dprc_endpoint *endpoint1,
-- const struct dprc_endpoint *endpoint2,
-- const struct dprc_connection_cfg *cfg);
--
--int dprc_disconnect(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- const struct dprc_endpoint *endpoint);
--
--int dprc_get_connection(struct fsl_mc_io *mc_io,
-- u32 cmd_flags,
-- u16 token,
-- const struct dprc_endpoint *endpoint1,
-- struct dprc_endpoint *endpoint2,
-- int *state);
--
--#endif /* _FSL_DPRC_H */
--
---- a/drivers/staging/fsl-mc/include/mc-bus.h
-+++ /dev/null
-@@ -1,111 +0,0 @@
--/*
-- * Freescale Management Complex (MC) bus declarations
-- *
-- * Copyright (C) 2014 Freescale Semiconductor, Inc.
-- * Author: German Rivera <German.Rivera@freescale.com>
-- *
-- * This file is licensed under the terms of the GNU General Public
-- * License version 2. This program is licensed "as is" without any
-- * warranty of any kind, whether express or implied.
-- */
--#ifndef _FSL_MC_MCBUS_H_
--#define _FSL_MC_MCBUS_H_
--
--#include "../include/mc.h"
--#include <linux/mutex.h>
--
--struct irq_domain;
--struct msi_domain_info;
--
--/**
-- * Maximum number of total IRQs that can be pre-allocated for an MC bus'
-- * IRQ pool
-- */
--#define FSL_MC_IRQ_POOL_MAX_TOTAL_IRQS 256
--
--#ifdef CONFIG_FSL_MC_BUS
--#define dev_is_fsl_mc(_dev) ((_dev)->bus == &fsl_mc_bus_type)
--#else
--/* If fsl-mc bus is not present device cannot belong to fsl-mc bus */
--#define dev_is_fsl_mc(_dev) (0)
--#endif
--
--/**
-- * struct fsl_mc_resource_pool - Pool of MC resources of a given
-- * type
-- * @type: type of resources in the pool
-- * @max_count: maximum number of resources in the pool
-- * @free_count: number of free resources in the pool
-- * @mutex: mutex to serialize access to the pool's free list
-- * @free_list: anchor node of list of free resources in the pool
-- * @mc_bus: pointer to the MC bus that owns this resource pool
-- */
--struct fsl_mc_resource_pool {
-- enum fsl_mc_pool_type type;
-- int16_t max_count;
-- int16_t free_count;
-- struct mutex mutex; /* serializes access to free_list */
-- struct list_head free_list;
-- struct fsl_mc_bus *mc_bus;
--};
--
--/**
-- * struct fsl_mc_bus - logical bus that corresponds to a physical DPRC
-- * @mc_dev: fsl-mc device for the bus device itself.
-- * @resource_pools: array of resource pools (one pool per resource type)
-- * for this MC bus. These resources represent allocatable entities
-- * from the physical DPRC.
-- * @irq_resources: Pointer to array of IRQ objects for the IRQ pool
-- * @scan_mutex: Serializes bus scanning
-- * @dprc_attr: DPRC attributes
-- */
--struct fsl_mc_bus {
-- struct fsl_mc_device mc_dev;
-- struct fsl_mc_resource_pool resource_pools[FSL_MC_NUM_POOL_TYPES];
-- struct fsl_mc_device_irq *irq_resources;
-- struct mutex scan_mutex; /* serializes bus scanning */
-- struct dprc_attributes dprc_attr;
--};
--
--#define to_fsl_mc_bus(_mc_dev) \
-- container_of(_mc_dev, struct fsl_mc_bus, mc_dev)
--
--int dprc_scan_container(struct fsl_mc_device *mc_bus_dev);
--
--int dprc_scan_objects(struct fsl_mc_device *mc_bus_dev,
-- unsigned int *total_irq_count);
--
--int __init dprc_driver_init(void);
--
--void dprc_driver_exit(void);
--
--int __init fsl_mc_allocator_driver_init(void);
--
--void fsl_mc_allocator_driver_exit(void);
--
--struct irq_domain *fsl_mc_msi_create_irq_domain(struct fwnode_handle *fwnode,
-- struct msi_domain_info *info,
-- struct irq_domain *parent);
--
--int fsl_mc_find_msi_domain(struct device *mc_platform_dev,
-- struct irq_domain **mc_msi_domain);
--
--int fsl_mc_populate_irq_pool(struct fsl_mc_bus *mc_bus,
-- unsigned int irq_count);
--
--void fsl_mc_cleanup_irq_pool(struct fsl_mc_bus *mc_bus);
--
--void fsl_mc_init_all_resource_pools(struct fsl_mc_device *mc_bus_dev);
--
--void fsl_mc_cleanup_all_resource_pools(struct fsl_mc_device *mc_bus_dev);
--
--bool fsl_mc_bus_exists(void);
--
--void fsl_mc_get_root_dprc(struct device *dev,
-- struct device **root_dprc_dev);
--
--bool fsl_mc_is_root_dprc(struct device *dev);
--
--extern struct bus_type fsl_mc_bus_type;
--
--#endif /* _FSL_MC_MCBUS_H_ */
---- a/drivers/staging/fsl-mc/include/mc-cmd.h
-+++ /dev/null
-@@ -1,108 +0,0 @@
--/* Copyright 2013-2015 Freescale Semiconductor Inc.
-- *
-- * Redistribution and use in source and binary forms, with or without
-- * modification, are permitted provided that the following conditions are met:
-- * * Redistributions of source code must retain the above copyright
-- * notice, this list of conditions and the following disclaimer.
-- * * Redistributions in binary form must reproduce the above copyright
-- * notice, this list of conditions and the following disclaimer in the
-- * documentation and/or other materials provided with the distribution.
-- * * Neither the name of the above-listed copyright holders nor the
-- * names of any contributors may be used to endorse or promote products
-- * derived from this software without specific prior written permission.
-- *
-- *
-- * ALTERNATIVELY, this software may be distributed under the terms of the
-- * GNU General Public License ("GPL") as published by the Free Software
-- * Foundation, either version 2 of that License or (at your option) any
-- * later version.
-- *
-- * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
-- * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
-- * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
-- * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE
-- * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
-- * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
-- * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
-- * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
-- * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
-- * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
-- * POSSIBILITY OF SUCH DAMAGE.
-- */
--#ifndef __FSL_MC_CMD_H
--#define __FSL_MC_CMD_H
--
--#define MC_CMD_NUM_OF_PARAMS 7
--
--struct mc_cmd_header {
-- u8 src_id;
-- u8 flags_hw;
-- u8 status;
-- u8 flags_sw;
-- __le16 token;
-- __le16 cmd_id;
--};
--
--struct mc_command {
-- u64 header;
-- u64 params[MC_CMD_NUM_OF_PARAMS];
--};
--
--enum mc_cmd_status {
-- MC_CMD_STATUS_OK = 0x0, /* Completed successfully */
-- MC_CMD_STATUS_READY = 0x1, /* Ready to be processed */
-- MC_CMD_STATUS_AUTH_ERR = 0x3, /* Authentication error */
-- MC_CMD_STATUS_NO_PRIVILEGE = 0x4, /* No privilege */
-- MC_CMD_STATUS_DMA_ERR = 0x5, /* DMA or I/O error */
-- MC_CMD_STATUS_CONFIG_ERR = 0x6, /* Configuration error */
-- MC_CMD_STATUS_TIMEOUT = 0x7, /* Operation timed out */
-- MC_CMD_STATUS_NO_RESOURCE = 0x8, /* No resources */
-- MC_CMD_STATUS_NO_MEMORY = 0x9, /* No memory available */
-- MC_CMD_STATUS_BUSY = 0xA, /* Device is busy */
-- MC_CMD_STATUS_UNSUPPORTED_OP = 0xB, /* Unsupported operation */
-- MC_CMD_STATUS_INVALID_STATE = 0xC /* Invalid state */
--};
--
--/*
-- * MC command flags
-- */
--
--/* High priority flag */
--#define MC_CMD_FLAG_PRI 0x80
--/* Command completion flag */
--#define MC_CMD_FLAG_INTR_DIS 0x01
--
--#define MC_CMD_HDR_CMDID_MASK 0xFFF0
--#define MC_CMD_HDR_CMDID_SHIFT 4
--#define MC_CMD_HDR_TOKEN_MASK 0xFFC0
--#define MC_CMD_HDR_TOKEN_SHIFT 6
--
--static inline u64 mc_encode_cmd_header(u16 cmd_id,
-- u32 cmd_flags,
-- u16 token)
--{
-- u64 header = 0;
-- struct mc_cmd_header *hdr = (struct mc_cmd_header *)&header;
--
-- hdr->cmd_id = cpu_to_le16((cmd_id << MC_CMD_HDR_CMDID_SHIFT) &
-- MC_CMD_HDR_CMDID_MASK);
-- hdr->token = cpu_to_le16((token << MC_CMD_HDR_TOKEN_SHIFT) &
-- MC_CMD_HDR_TOKEN_MASK);
-- hdr->status = MC_CMD_STATUS_READY;
-- if (cmd_flags & MC_CMD_FLAG_PRI)
-- hdr->flags_hw = MC_CMD_FLAG_PRI;
-- if (cmd_flags & MC_CMD_FLAG_INTR_DIS)
-- hdr->flags_sw = MC_CMD_FLAG_INTR_DIS;
--
-- return header;
--}
--
--static inline u16 mc_cmd_hdr_read_token(struct mc_command *cmd)
--{
-- struct mc_cmd_header *hdr = (struct mc_cmd_header *)&cmd->header;
-- u16 token = le16_to_cpu(hdr->token);
--
-- return (token & MC_CMD_HDR_TOKEN_MASK) >> MC_CMD_HDR_TOKEN_SHIFT;
--}
--
--#endif /* __FSL_MC_CMD_H */
---- a/drivers/staging/fsl-mc/include/mc-sys.h
-+++ /dev/null
-@@ -1,98 +0,0 @@
--/* Copyright 2013-2014 Freescale Semiconductor Inc.
-- *
-- * Interface of the I/O services to send MC commands to the MC hardware
-- *
-- * Redistribution and use in source and binary forms, with or without
-- * modification, are permitted provided that the following conditions are met:
-- * * Redistributions of source code must retain the above copyright
-- * notice, this list of conditions and the following disclaimer.
-- * * Redistributions in binary form must reproduce the above copyright
-- * notice, this list of conditions and the following disclaimer in the
-- * documentation and/or other materials provided with the distribution.
-- * * Neither the name of the above-listed copyright holders nor the
-- * names of any contributors may be used to endorse or promote products
-- * derived from this software without specific prior written permission.
-- *
-- *
-- * ALTERNATIVELY, this software may be distributed under the terms of the
-- * GNU General Public License ("GPL") as published by the Free Software
-- * Foundation, either version 2 of that License or (at your option) any
-- * later version.
-- *
-- * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
-- * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
-- * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
-- * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE
-- * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
-- * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
-- * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
-- * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
-- * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
-- * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
-- * POSSIBILITY OF SUCH DAMAGE.
-- */
--
--#ifndef _FSL_MC_SYS_H
--#define _FSL_MC_SYS_H
--
--#include <linux/types.h>
--#include <linux/errno.h>
--#include <linux/mutex.h>
--#include <linux/spinlock.h>
--
--/**
-- * Bit masks for a MC I/O object (struct fsl_mc_io) flags
-- */
--#define FSL_MC_IO_ATOMIC_CONTEXT_PORTAL 0x0001
--
--struct fsl_mc_resource;
--struct mc_command;
--
--/**
-- * struct fsl_mc_io - MC I/O object to be passed-in to mc_send_command()
-- * @dev: device associated with this Mc I/O object
-- * @flags: flags for mc_send_command()
-- * @portal_size: MC command portal size in bytes
-- * @portal_phys_addr: MC command portal physical address
-- * @portal_virt_addr: MC command portal virtual address
-- * @dpmcp_dev: pointer to the DPMCP device associated with the MC portal.
-- *
-- * Fields are only meaningful if the FSL_MC_IO_ATOMIC_CONTEXT_PORTAL flag is not
-- * set:
-- * @mutex: Mutex to serialize mc_send_command() calls that use the same MC
-- * portal, if the fsl_mc_io object was created with the
-- * FSL_MC_IO_ATOMIC_CONTEXT_PORTAL flag off. mc_send_command() calls for this
-- * fsl_mc_io object must be made only from non-atomic context.
-- *
-- * Fields are only meaningful if the FSL_MC_IO_ATOMIC_CONTEXT_PORTAL flag is
-- * set:
-- * @spinlock: Spinlock to serialize mc_send_command() calls that use the same MC
-- * portal, if the fsl_mc_io object was created with the
-- * FSL_MC_IO_ATOMIC_CONTEXT_PORTAL flag on. mc_send_command() calls for this
-- * fsl_mc_io object can be made from atomic or non-atomic context.
-- */
--struct fsl_mc_io {
-- struct device *dev;
-- u16 flags;
-- u16 portal_size;
-- phys_addr_t portal_phys_addr;
-- void __iomem *portal_virt_addr;
-- struct fsl_mc_device *dpmcp_dev;
-- union {
-- /*
-- * This field is only meaningful if the
-- * FSL_MC_IO_ATOMIC_CONTEXT_PORTAL flag is not set
-- */
-- struct mutex mutex; /* serializes mc_send_command() */
--
-- /*
-- * This field is only meaningful if the
-- * FSL_MC_IO_ATOMIC_CONTEXT_PORTAL flag is set
-- */
-- spinlock_t spinlock; /* serializes mc_send_command() */
-- };
--};
--
--int mc_send_command(struct fsl_mc_io *mc_io, struct mc_command *cmd);
--
--#endif /* _FSL_MC_SYS_H */
---- a/drivers/staging/fsl-mc/include/mc.h
-+++ /dev/null
-@@ -1,201 +0,0 @@
--/*
-- * Freescale Management Complex (MC) bus public interface
-- *
-- * Copyright (C) 2014 Freescale Semiconductor, Inc.
-- * Author: German Rivera <German.Rivera@freescale.com>
-- *
-- * This file is licensed under the terms of the GNU General Public
-- * License version 2. This program is licensed "as is" without any
-- * warranty of any kind, whether express or implied.
-- */
--#ifndef _FSL_MC_H_
--#define _FSL_MC_H_
--
--#include <linux/device.h>
--#include <linux/mod_devicetable.h>
--#include <linux/interrupt.h>
--#include "../include/dprc.h"
--
--#define FSL_MC_VENDOR_FREESCALE 0x1957
--
--struct fsl_mc_device;
--struct fsl_mc_io;
--
--/**
-- * struct fsl_mc_driver - MC object device driver object
-- * @driver: Generic device driver
-- * @match_id_table: table of supported device matching Ids
-- * @probe: Function called when a device is added
-- * @remove: Function called when a device is removed
-- * @shutdown: Function called at shutdown time to quiesce the device
-- * @suspend: Function called when a device is stopped
-- * @resume: Function called when a device is resumed
-- *
-- * Generic DPAA device driver object for device drivers that are registered
-- * with a DPRC bus. This structure is to be embedded in each device-specific
-- * driver structure.
-- */
--struct fsl_mc_driver {
-- struct device_driver driver;
-- const struct fsl_mc_device_id *match_id_table;
-- int (*probe)(struct fsl_mc_device *dev);
-- int (*remove)(struct fsl_mc_device *dev);
-- void (*shutdown)(struct fsl_mc_device *dev);
-- int (*suspend)(struct fsl_mc_device *dev, pm_message_t state);
-- int (*resume)(struct fsl_mc_device *dev);
--};
--
--#define to_fsl_mc_driver(_drv) \
-- container_of(_drv, struct fsl_mc_driver, driver)
--
--/**
-- * enum fsl_mc_pool_type - Types of allocatable MC bus resources
-- *
-- * Entries in these enum are used as indices in the array of resource
-- * pools of an fsl_mc_bus object.
-- */
--enum fsl_mc_pool_type {
-- FSL_MC_POOL_DPMCP = 0x0, /* corresponds to "dpmcp" in the MC */
-- FSL_MC_POOL_DPBP, /* corresponds to "dpbp" in the MC */
-- FSL_MC_POOL_DPCON, /* corresponds to "dpcon" in the MC */
-- FSL_MC_POOL_IRQ,
--
-- /*
-- * NOTE: New resource pool types must be added before this entry
-- */
-- FSL_MC_NUM_POOL_TYPES
--};
--
--/**
-- * struct fsl_mc_resource - MC generic resource
-- * @type: type of resource
-- * @id: unique MC resource Id within the resources of the same type
-- * @data: pointer to resource-specific data if the resource is currently
-- * allocated, or NULL if the resource is not currently allocated.
-- * @parent_pool: pointer to the parent resource pool from which this
-- * resource is allocated from.
-- * @node: Node in the free list of the corresponding resource pool
-- *
-- * NOTE: This structure is to be embedded as a field of specific
-- * MC resource structures.
-- */
--struct fsl_mc_resource {
-- enum fsl_mc_pool_type type;
-- int32_t id;
-- void *data;
-- struct fsl_mc_resource_pool *parent_pool;
-- struct list_head node;
--};
--
--/**
-- * struct fsl_mc_device_irq - MC object device message-based interrupt
-- * @msi_desc: pointer to MSI descriptor allocated by fsl_mc_msi_alloc_descs()
-- * @mc_dev: MC object device that owns this interrupt
-- * @dev_irq_index: device-relative IRQ index
-- * @resource: MC generic resource associated with the interrupt
-- */
--struct fsl_mc_device_irq {
-- struct msi_desc *msi_desc;
-- struct fsl_mc_device *mc_dev;
-- u8 dev_irq_index;
-- struct fsl_mc_resource resource;
--};
--
--#define to_fsl_mc_irq(_mc_resource) \
-- container_of(_mc_resource, struct fsl_mc_device_irq, resource)
--
--/**
-- * Bit masks for a MC object device (struct fsl_mc_device) flags
-- */
--#define FSL_MC_IS_DPRC 0x0001
--
--/**
-- * struct fsl_mc_device - MC object device object
-- * @dev: Linux driver model device object
-- * @dma_mask: Default DMA mask
-- * @flags: MC object device flags
-- * @icid: Isolation context ID for the device
-- * @mc_handle: MC handle for the corresponding MC object opened
-- * @mc_io: Pointer to MC IO object assigned to this device or
-- * NULL if none.
-- * @obj_desc: MC description of the DPAA device
-- * @regions: pointer to array of MMIO region entries
-- * @irqs: pointer to array of pointers to interrupts allocated to this device
-- * @resource: generic resource associated with this MC object device, if any.
-- *
-- * Generic device object for MC object devices that are "attached" to a
-- * MC bus.
-- *
-- * NOTES:
-- * - For a non-DPRC object its icid is the same as its parent DPRC's icid.
-- * - The SMMU notifier callback gets invoked after device_add() has been
-- * called for an MC object device, but before the device-specific probe
-- * callback gets called.
-- * - DP_OBJ_DPRC objects are the only MC objects that have built-in MC
-- * portals. For all other MC objects, their device drivers are responsible for
-- * allocating MC portals for them by calling fsl_mc_portal_allocate().
-- * - Some types of MC objects (e.g., DP_OBJ_DPBP, DP_OBJ_DPCON) are
-- * treated as resources that can be allocated/deallocated from the
-- * corresponding resource pool in the object's parent DPRC, using the
-- * fsl_mc_object_allocate()/fsl_mc_object_free() functions. These MC objects
-- * are known as "allocatable" objects. For them, the corresponding
-- * fsl_mc_device's 'resource' points to the associated resource object.
-- * For MC objects that are not allocatable (e.g., DP_OBJ_DPRC, DP_OBJ_DPNI),
-- * 'resource' is NULL.
-- */
--struct fsl_mc_device {
-- struct device dev;
-- u64 dma_mask;
-- u16 flags;
-- u16 icid;
-- u16 mc_handle;
-- struct fsl_mc_io *mc_io;
-- struct dprc_obj_desc obj_desc;
-- struct resource *regions;
-- struct fsl_mc_device_irq **irqs;
-- struct fsl_mc_resource *resource;
--};
--
--#define to_fsl_mc_device(_dev) \
-- container_of(_dev, struct fsl_mc_device, dev)
--
--/*
-- * module_fsl_mc_driver() - Helper macro for drivers that don't do
-- * anything special in module init/exit. This eliminates a lot of
-- * boilerplate. Each module may only use this macro once, and
-- * calling it replaces module_init() and module_exit()
-- */
--#define module_fsl_mc_driver(__fsl_mc_driver) \
-- module_driver(__fsl_mc_driver, fsl_mc_driver_register, \
-- fsl_mc_driver_unregister)
--
--/*
-- * Macro to avoid include chaining to get THIS_MODULE
-- */
--#define fsl_mc_driver_register(drv) \
-- __fsl_mc_driver_register(drv, THIS_MODULE)
--
--int __must_check __fsl_mc_driver_register(struct fsl_mc_driver *fsl_mc_driver,
-- struct module *owner);
--
--void fsl_mc_driver_unregister(struct fsl_mc_driver *driver);
--
--int __must_check fsl_mc_portal_allocate(struct fsl_mc_device *mc_dev,
-- u16 mc_io_flags,
-- struct fsl_mc_io **new_mc_io);
--
--void fsl_mc_portal_free(struct fsl_mc_io *mc_io);
--
--int fsl_mc_portal_reset(struct fsl_mc_io *mc_io);
--
--int __must_check fsl_mc_object_allocate(struct fsl_mc_device *mc_dev,
-- enum fsl_mc_pool_type pool_type,
-- struct fsl_mc_device **new_mc_adev);
--
--void fsl_mc_object_free(struct fsl_mc_device *mc_adev);
--
--int __must_check fsl_mc_allocate_irqs(struct fsl_mc_device *mc_dev);
--
--void fsl_mc_free_irqs(struct fsl_mc_device *mc_dev);
--
--#endif /* _FSL_MC_H_ */
---- /dev/null
-+++ b/include/linux/fsl/mc.h
-@@ -0,0 +1,1025 @@
-+/* SPDX-License-Identifier: GPL-2.0 */
-+/*
-+ * Freescale Management Complex (MC) bus public interface
-+ *
-+ * Copyright (C) 2014-2016 Freescale Semiconductor, Inc.
-+ * Author: German Rivera <German.Rivera@freescale.com>
-+ *
-+ */
-+#ifndef _FSL_MC_H_
-+#define _FSL_MC_H_
-+
-+#include <linux/device.h>
-+#include <linux/mod_devicetable.h>
-+#include <linux/interrupt.h>
-+#include <linux/cdev.h>
-+#include <uapi/linux/fsl_mc.h>
-+
-+#define FSL_MC_VENDOR_FREESCALE 0x1957
-+
-+struct irq_domain;
-+struct msi_domain_info;
-+
-+struct fsl_mc_device;
-+struct fsl_mc_io;
-+
-+/**
-+ * struct fsl_mc_driver - MC object device driver object
-+ * @driver: Generic device driver
-+ * @match_id_table: table of supported device matching Ids
-+ * @probe: Function called when a device is added
-+ * @remove: Function called when a device is removed
-+ * @shutdown: Function called at shutdown time to quiesce the device
-+ * @suspend: Function called when a device is stopped
-+ * @resume: Function called when a device is resumed
-+ *
-+ * Generic DPAA device driver object for device drivers that are registered
-+ * with a DPRC bus. This structure is to be embedded in each device-specific
-+ * driver structure.
-+ */
-+struct fsl_mc_driver {
-+ struct device_driver driver;
-+ const struct fsl_mc_device_id *match_id_table;
-+ int (*probe)(struct fsl_mc_device *dev);
-+ int (*remove)(struct fsl_mc_device *dev);
-+ void (*shutdown)(struct fsl_mc_device *dev);
-+ int (*suspend)(struct fsl_mc_device *dev, pm_message_t state);
-+ int (*resume)(struct fsl_mc_device *dev);
-+};
-+
-+#define to_fsl_mc_driver(_drv) \
-+ container_of(_drv, struct fsl_mc_driver, driver)
-+
-+#define to_fsl_mc_bus(_mc_dev) \
-+ container_of(_mc_dev, struct fsl_mc_bus, mc_dev)
-+
-+/**
-+ * enum fsl_mc_pool_type - Types of allocatable MC bus resources
-+ *
-+ * Entries in these enum are used as indices in the array of resource
-+ * pools of an fsl_mc_bus object.
-+ */
-+enum fsl_mc_pool_type {
-+ FSL_MC_POOL_DPMCP = 0x0, /* corresponds to "dpmcp" in the MC */
-+ FSL_MC_POOL_DPBP, /* corresponds to "dpbp" in the MC */
-+ FSL_MC_POOL_DPCON, /* corresponds to "dpcon" in the MC */
-+ FSL_MC_POOL_IRQ,
-+
-+ /*
-+ * NOTE: New resource pool types must be added before this entry
-+ */
-+ FSL_MC_NUM_POOL_TYPES
-+};
-+
-+/**
-+ * struct fsl_mc_resource - MC generic resource
-+ * @type: type of resource
-+ * @id: unique MC resource Id within the resources of the same type
-+ * @data: pointer to resource-specific data if the resource is currently
-+ * allocated, or NULL if the resource is not currently allocated.
-+ * @parent_pool: pointer to the parent resource pool from which this
-+ * resource is allocated from.
-+ * @node: Node in the free list of the corresponding resource pool
-+ *
-+ * NOTE: This structure is to be embedded as a field of specific
-+ * MC resource structures.
-+ */
-+struct fsl_mc_resource {
-+ enum fsl_mc_pool_type type;
-+ s32 id;
-+ void *data;
-+ struct fsl_mc_resource_pool *parent_pool;
-+ struct list_head node;
-+};
-+
-+/**
-+ * struct fsl_mc_device_irq - MC object device message-based interrupt
-+ * @msi_desc: pointer to MSI descriptor allocated by fsl_mc_msi_alloc_descs()
-+ * @mc_dev: MC object device that owns this interrupt
-+ * @dev_irq_index: device-relative IRQ index
-+ * @resource: MC generic resource associated with the interrupt
-+ */
-+struct fsl_mc_device_irq {
-+ struct msi_desc *msi_desc;
-+ struct fsl_mc_device *mc_dev;
-+ u8 dev_irq_index;
-+ struct fsl_mc_resource resource;
-+};
-+
-+#define to_fsl_mc_irq(_mc_resource) \
-+ container_of(_mc_resource, struct fsl_mc_device_irq, resource)
-+
-+/* Opened state - Indicates that an object is open by at least one owner */
-+#define FSL_MC_OBJ_STATE_OPEN 0x00000001
-+/* Plugged state - Indicates that the object is plugged */
-+#define FSL_MC_OBJ_STATE_PLUGGED 0x00000002
-+
-+/**
-+ * Shareability flag - Object flag indicating no memory shareability.
-+ * the object generates memory accesses that are non coherent with other
-+ * masters;
-+ * user is responsible for proper memory handling through IOMMU configuration.
-+ */
-+#define FSL_MC_OBJ_FLAG_NO_MEM_SHAREABILITY 0x0001
-+
-+/**
-+ * struct fsl_mc_obj_desc - Object descriptor
-+ * @type: Type of object: NULL terminated string
-+ * @id: ID of logical object resource
-+ * @vendor: Object vendor identifier
-+ * @ver_major: Major version number
-+ * @ver_minor: Minor version number
-+ * @irq_count: Number of interrupts supported by the object
-+ * @region_count: Number of mappable regions supported by the object
-+ * @state: Object state: combination of FSL_MC_OBJ_STATE_ states
-+ * @label: Object label: NULL terminated string
-+ * @flags: Object's flags
-+ */
-+struct fsl_mc_obj_desc {
-+ char type[16];
-+ int id;
-+ u16 vendor;
-+ u16 ver_major;
-+ u16 ver_minor;
-+ u8 irq_count;
-+ u8 region_count;
-+ u32 state;
-+ char label[16];
-+ u16 flags;
-+};
-+
-+/**
-+ * Bit masks for a MC object device (struct fsl_mc_device) flags
-+ */
-+#define FSL_MC_IS_DPRC 0x0001
-+
-+/**
-+ * struct fsl_mc_device - MC object device object
-+ * @dev: Linux driver model device object
-+ * @dma_mask: Default DMA mask
-+ * @flags: MC object device flags
-+ * @icid: Isolation context ID for the device
-+ * @mc_handle: MC handle for the corresponding MC object opened
-+ * @mc_io: Pointer to MC IO object assigned to this device or
-+ * NULL if none.
-+ * @obj_desc: MC description of the DPAA device
-+ * @regions: pointer to array of MMIO region entries
-+ * @irqs: pointer to array of pointers to interrupts allocated to this device
-+ * @resource: generic resource associated with this MC object device, if any.
-+ * @driver_override: Driver name to force a match
-+ *
-+ * Generic device object for MC object devices that are "attached" to a
-+ * MC bus.
-+ *
-+ * NOTES:
-+ * - For a non-DPRC object its icid is the same as its parent DPRC's icid.
-+ * - The SMMU notifier callback gets invoked after device_add() has been
-+ * called for an MC object device, but before the device-specific probe
-+ * callback gets called.
-+ * - DP_OBJ_DPRC objects are the only MC objects that have built-in MC
-+ * portals. For all other MC objects, their device drivers are responsible for
-+ * allocating MC portals for them by calling fsl_mc_portal_allocate().
-+ * - Some types of MC objects (e.g., DP_OBJ_DPBP, DP_OBJ_DPCON) are
-+ * treated as resources that can be allocated/deallocated from the
-+ * corresponding resource pool in the object's parent DPRC, using the
-+ * fsl_mc_object_allocate()/fsl_mc_object_free() functions. These MC objects
-+ * are known as "allocatable" objects. For them, the corresponding
-+ * fsl_mc_device's 'resource' points to the associated resource object.
-+ * For MC objects that are not allocatable (e.g., DP_OBJ_DPRC, DP_OBJ_DPNI),
-+ * 'resource' is NULL.
-+ */
-+struct fsl_mc_device {
-+ struct device dev;
-+ u64 dma_mask;
-+ u16 flags;
-+ u32 icid;
-+ u16 mc_handle;
-+ struct fsl_mc_io *mc_io;
-+ struct fsl_mc_obj_desc obj_desc;
-+ struct resource *regions;
-+ struct fsl_mc_device_irq **irqs;
-+ struct fsl_mc_resource *resource;
-+ const char *driver_override;
-+};
-+
-+#define to_fsl_mc_device(_dev) \
-+ container_of(_dev, struct fsl_mc_device, dev)
-+
-+struct mc_cmd_header {
-+ u8 src_id;
-+ u8 flags_hw;
-+ u8 status;
-+ u8 flags_sw;
-+ __le16 token;
-+ __le16 cmd_id;
-+};
-+
-+enum mc_cmd_status {
-+ MC_CMD_STATUS_OK = 0x0, /* Completed successfully */
-+ MC_CMD_STATUS_READY = 0x1, /* Ready to be processed */
-+ MC_CMD_STATUS_AUTH_ERR = 0x3, /* Authentication error */
-+ MC_CMD_STATUS_NO_PRIVILEGE = 0x4, /* No privilege */
-+ MC_CMD_STATUS_DMA_ERR = 0x5, /* DMA or I/O error */
-+ MC_CMD_STATUS_CONFIG_ERR = 0x6, /* Configuration error */
-+ MC_CMD_STATUS_TIMEOUT = 0x7, /* Operation timed out */
-+ MC_CMD_STATUS_NO_RESOURCE = 0x8, /* No resources */
-+ MC_CMD_STATUS_NO_MEMORY = 0x9, /* No memory available */
-+ MC_CMD_STATUS_BUSY = 0xA, /* Device is busy */
-+ MC_CMD_STATUS_UNSUPPORTED_OP = 0xB, /* Unsupported operation */
-+ MC_CMD_STATUS_INVALID_STATE = 0xC /* Invalid state */
-+};
-+
-+/*
-+ * MC command flags
-+ */
-+
-+/* High priority flag */
-+#define MC_CMD_FLAG_PRI 0x80
-+/* Command completion flag */
-+#define MC_CMD_FLAG_INTR_DIS 0x01
-+
-+static inline u64 mc_encode_cmd_header(u16 cmd_id,
-+ u32 cmd_flags,
-+ u16 token)
-+{
-+ u64 header = 0;
-+ struct mc_cmd_header *hdr = (struct mc_cmd_header *)&header;
-+
-+ hdr->cmd_id = cpu_to_le16(cmd_id);
-+ hdr->token = cpu_to_le16(token);
-+ hdr->status = MC_CMD_STATUS_READY;
-+ if (cmd_flags & MC_CMD_FLAG_PRI)
-+ hdr->flags_hw = MC_CMD_FLAG_PRI;
-+ if (cmd_flags & MC_CMD_FLAG_INTR_DIS)
-+ hdr->flags_sw = MC_CMD_FLAG_INTR_DIS;
-+
-+ return header;
-+}
-+
-+static inline u16 mc_cmd_hdr_read_token(struct fsl_mc_command *cmd)
-+{
-+ struct mc_cmd_header *hdr = (struct mc_cmd_header *)&cmd->header;
-+ u16 token = le16_to_cpu(hdr->token);
-+
-+ return token;
-+}
-+
-+struct mc_rsp_create {
-+ __le32 object_id;
-+};
-+
-+struct mc_rsp_api_ver {
-+ __le16 major_ver;
-+ __le16 minor_ver;
-+};
-+
-+static inline u32 mc_cmd_read_object_id(struct fsl_mc_command *cmd)
-+{
-+ struct mc_rsp_create *rsp_params;
-+
-+ rsp_params = (struct mc_rsp_create *)cmd->params;
-+ return le32_to_cpu(rsp_params->object_id);
-+}
-+
-+static inline void mc_cmd_read_api_version(struct fsl_mc_command *cmd,
-+ u16 *major_ver,
-+ u16 *minor_ver)
-+{
-+ struct mc_rsp_api_ver *rsp_params;
-+
-+ rsp_params = (struct mc_rsp_api_ver *)cmd->params;
-+ *major_ver = le16_to_cpu(rsp_params->major_ver);
-+ *minor_ver = le16_to_cpu(rsp_params->minor_ver);
-+}
-+
-+/**
-+ * Bit masks for a MC I/O object (struct fsl_mc_io) flags
-+ */
-+#define FSL_MC_IO_ATOMIC_CONTEXT_PORTAL 0x0001
-+
-+/**
-+ * struct fsl_mc_io - MC I/O object to be passed-in to mc_send_command()
-+ * @dev: device associated with this Mc I/O object
-+ * @flags: flags for mc_send_command()
-+ * @portal_size: MC command portal size in bytes
-+ * @portal_phys_addr: MC command portal physical address
-+ * @portal_virt_addr: MC command portal virtual address
-+ * @dpmcp_dev: pointer to the DPMCP device associated with the MC portal.
-+ *
-+ * Fields are only meaningful if the FSL_MC_IO_ATOMIC_CONTEXT_PORTAL flag is not
-+ * set:
-+ * @mutex: Mutex to serialize mc_send_command() calls that use the same MC
-+ * portal, if the fsl_mc_io object was created with the
-+ * FSL_MC_IO_ATOMIC_CONTEXT_PORTAL flag off. mc_send_command() calls for this
-+ * fsl_mc_io object must be made only from non-atomic context.
-+ *
-+ * Fields are only meaningful if the FSL_MC_IO_ATOMIC_CONTEXT_PORTAL flag is
-+ * set:
-+ * @spinlock: Spinlock to serialize mc_send_command() calls that use the same MC
-+ * portal, if the fsl_mc_io object was created with the
-+ * FSL_MC_IO_ATOMIC_CONTEXT_PORTAL flag on. mc_send_command() calls for this
-+ * fsl_mc_io object can be made from atomic or non-atomic context.
-+ */
-+struct fsl_mc_io {
-+ struct device *dev;
-+ u16 flags;
-+ u32 portal_size;
-+ phys_addr_t portal_phys_addr;
-+ void __iomem *portal_virt_addr;
-+ struct fsl_mc_device *dpmcp_dev;
-+ union {
-+ /*
-+ * This field is only meaningful if the
-+ * FSL_MC_IO_ATOMIC_CONTEXT_PORTAL flag is not set
-+ */
-+ struct mutex mutex; /* serializes mc_send_command() */
-+
-+ /*
-+ * This field is only meaningful if the
-+ * FSL_MC_IO_ATOMIC_CONTEXT_PORTAL flag is set
-+ */
-+ spinlock_t spinlock; /* serializes mc_send_command() */
-+ };
-+};
-+
-+int mc_send_command(struct fsl_mc_io *mc_io, struct fsl_mc_command *cmd);
-+
-+#ifdef CONFIG_FSL_MC_BUS
-+#define dev_is_fsl_mc(_dev) ((_dev)->bus == &fsl_mc_bus_type)
-+#else
-+/* If fsl-mc bus is not present device cannot belong to fsl-mc bus */
-+#define dev_is_fsl_mc(_dev) (0)
-+#endif
-+
-+/* Macro to check if a device is a container device */
-+#define fsl_mc_is_cont_dev(_dev) (to_fsl_mc_device(_dev)->flags & \
-+ FSL_MC_IS_DPRC)
-+
-+/* Macro to get the container device of a MC device */
-+#define fsl_mc_cont_dev(_dev) (fsl_mc_is_cont_dev(_dev) ? \
-+ (_dev) : (_dev)->parent)
-+
-+#define fsl_mc_is_dev_coherent(_dev) \
-+ (!((to_fsl_mc_device(_dev))->obj_desc.flags & \
-+ FSL_MC_OBJ_FLAG_NO_MEM_SHAREABILITY))
-+
-+/*
-+ * module_fsl_mc_driver() - Helper macro for drivers that don't do
-+ * anything special in module init/exit. This eliminates a lot of
-+ * boilerplate. Each module may only use this macro once, and
-+ * calling it replaces module_init() and module_exit()
-+ */
-+#define module_fsl_mc_driver(__fsl_mc_driver) \
-+ module_driver(__fsl_mc_driver, fsl_mc_driver_register, \
-+ fsl_mc_driver_unregister)
-+
-+void fsl_mc_device_remove(struct fsl_mc_device *mc_dev);
-+
-+/*
-+ * Macro to avoid include chaining to get THIS_MODULE
-+ */
-+#define fsl_mc_driver_register(drv) \
-+ __fsl_mc_driver_register(drv, THIS_MODULE)
-+
-+int __must_check __fsl_mc_driver_register(struct fsl_mc_driver *fsl_mc_driver,
-+ struct module *owner);
-+
-+void fsl_mc_driver_unregister(struct fsl_mc_driver *driver);
-+
-+int __must_check fsl_mc_portal_allocate(struct fsl_mc_device *mc_dev,
-+ u16 mc_io_flags,
-+ struct fsl_mc_io **new_mc_io);
-+
-+void fsl_mc_portal_free(struct fsl_mc_io *mc_io);
-+
-+int fsl_mc_portal_reset(struct fsl_mc_io *mc_io);
-+
-+int __must_check fsl_mc_object_allocate(struct fsl_mc_device *mc_dev,
-+ enum fsl_mc_pool_type pool_type,
-+ struct fsl_mc_device **new_mc_adev);
-+
-+void fsl_mc_object_free(struct fsl_mc_device *mc_adev);
-+
-+struct irq_domain *fsl_mc_msi_create_irq_domain(struct fwnode_handle *fwnode,
-+ struct msi_domain_info *info,
-+ struct irq_domain *parent);
-+
-+int __must_check fsl_mc_allocate_irqs(struct fsl_mc_device *mc_dev);
-+
-+void fsl_mc_free_irqs(struct fsl_mc_device *mc_dev);
-+
-+void fsl_mc_dma_configure(struct fsl_mc_device *mc_dev,
-+ struct device_node *fsl_mc_platform_node, int coherent);
-+
-+extern struct bus_type fsl_mc_bus_type;
-+
-+extern struct device_type fsl_mc_bus_dprc_type;
-+extern struct device_type fsl_mc_bus_dpni_type;
-+extern struct device_type fsl_mc_bus_dpio_type;
-+extern struct device_type fsl_mc_bus_dpsw_type;
-+extern struct device_type fsl_mc_bus_dpdmux_type;
-+extern struct device_type fsl_mc_bus_dpbp_type;
-+extern struct device_type fsl_mc_bus_dpcon_type;
-+extern struct device_type fsl_mc_bus_dpmcp_type;
-+extern struct device_type fsl_mc_bus_dpmac_type;
-+extern struct device_type fsl_mc_bus_dprtc_type;
-+extern struct device_type fsl_mc_bus_dpseci_type;
-+extern struct device_type fsl_mc_bus_dpdcei_type;
-+extern struct device_type fsl_mc_bus_dpaiop_type;
-+extern struct device_type fsl_mc_bus_dpci_type;
-+extern struct device_type fsl_mc_bus_dpdmai_type;
-+
-+static inline bool is_fsl_mc_bus_dprc(const struct fsl_mc_device *mc_dev)
-+{
-+ return mc_dev->dev.type == &fsl_mc_bus_dprc_type;
-+}
-+
-+static inline bool is_fsl_mc_bus_dpni(const struct fsl_mc_device *mc_dev)
-+{
-+ return mc_dev->dev.type == &fsl_mc_bus_dpni_type;
-+}
-+
-+static inline bool is_fsl_mc_bus_dpio(const struct fsl_mc_device *mc_dev)
-+{
-+ return mc_dev->dev.type == &fsl_mc_bus_dpio_type;
-+}
-+
-+static inline bool is_fsl_mc_bus_dpsw(const struct fsl_mc_device *mc_dev)
-+{
-+ return mc_dev->dev.type == &fsl_mc_bus_dpsw_type;
-+}
-+
-+static inline bool is_fsl_mc_bus_dpdmux(const struct fsl_mc_device *mc_dev)
-+{
-+ return mc_dev->dev.type == &fsl_mc_bus_dpdmux_type;
-+}
-+
-+static inline bool is_fsl_mc_bus_dpbp(const struct fsl_mc_device *mc_dev)
-+{
-+ return mc_dev->dev.type == &fsl_mc_bus_dpbp_type;
-+}
-+
-+static inline bool is_fsl_mc_bus_dpcon(const struct fsl_mc_device *mc_dev)
-+{
-+ return mc_dev->dev.type == &fsl_mc_bus_dpcon_type;
-+}
-+
-+static inline bool is_fsl_mc_bus_dpmcp(const struct fsl_mc_device *mc_dev)
-+{
-+ return mc_dev->dev.type == &fsl_mc_bus_dpmcp_type;
-+}
-+
-+static inline bool is_fsl_mc_bus_dpmac(const struct fsl_mc_device *mc_dev)
-+{
-+ return mc_dev->dev.type == &fsl_mc_bus_dpmac_type;
-+}
-+
-+static inline bool is_fsl_mc_bus_dprtc(const struct fsl_mc_device *mc_dev)
-+{
-+ return mc_dev->dev.type == &fsl_mc_bus_dprtc_type;
-+}
-+
-+static inline bool is_fsl_mc_bus_dpseci(const struct fsl_mc_device *mc_dev)
-+{
-+ return mc_dev->dev.type == &fsl_mc_bus_dpseci_type;
-+}
-+
-+static inline bool is_fsl_mc_bus_dpdcei(const struct fsl_mc_device *mc_dev)
-+{
-+ return mc_dev->dev.type == &fsl_mc_bus_dpdcei_type;
-+}
-+
-+static inline bool is_fsl_mc_bus_dpaiop(const struct fsl_mc_device *mc_dev)
-+{
-+ return mc_dev->dev.type == &fsl_mc_bus_dpaiop_type;
-+}
-+
-+static inline bool is_fsl_mc_bus_dpci(const struct fsl_mc_device *mc_dev)
-+{
-+ return mc_dev->dev.type == &fsl_mc_bus_dpci_type;
-+}
-+
-+static inline bool is_fsl_mc_bus_dpdmai(const struct fsl_mc_device *mc_dev)
-+{
-+ return mc_dev->dev.type == &fsl_mc_bus_dpdmai_type;
-+}
-+
-+/*
-+ * Data Path Resource Container (DPRC) API
-+ */
-+
-+/* Minimal supported DPRC Version */
-+#define DPRC_MIN_VER_MAJOR 6
-+#define DPRC_MIN_VER_MINOR 0
-+
-+/* DPRC command versioning */
-+#define DPRC_CMD_BASE_VERSION 1
-+#define DPRC_CMD_ID_OFFSET 4
-+
-+#define DPRC_CMD(id) (((id) << DPRC_CMD_ID_OFFSET) | DPRC_CMD_BASE_VERSION)
-+
-+/* DPRC command IDs */
-+#define DPRC_CMDID_CLOSE DPRC_CMD(0x800)
-+#define DPRC_CMDID_OPEN DPRC_CMD(0x805)
-+#define DPRC_CMDID_GET_API_VERSION DPRC_CMD(0xa05)
-+
-+#define DPRC_CMDID_GET_ATTR DPRC_CMD(0x004)
-+#define DPRC_CMDID_RESET_CONT DPRC_CMD(0x005)
-+
-+#define DPRC_CMDID_SET_IRQ DPRC_CMD(0x010)
-+#define DPRC_CMDID_SET_IRQ_ENABLE DPRC_CMD(0x012)
-+#define DPRC_CMDID_SET_IRQ_MASK DPRC_CMD(0x014)
-+#define DPRC_CMDID_GET_IRQ_STATUS DPRC_CMD(0x016)
-+#define DPRC_CMDID_CLEAR_IRQ_STATUS DPRC_CMD(0x017)
-+
-+#define DPRC_CMDID_GET_CONT_ID DPRC_CMD(0x830)
-+#define DPRC_CMDID_GET_OBJ_COUNT DPRC_CMD(0x159)
-+#define DPRC_CMDID_GET_OBJ DPRC_CMD(0x15A)
-+#define DPRC_CMDID_GET_OBJ_REG DPRC_CMD(0x15E)
-+#define DPRC_CMDID_SET_OBJ_IRQ DPRC_CMD(0x15F)
-+
-+struct dprc_cmd_open {
-+ __le32 container_id;
-+};
-+
-+struct dprc_cmd_reset_container {
-+ __le32 child_container_id;
-+};
-+
-+struct dprc_cmd_set_irq {
-+ /* cmd word 0 */
-+ __le32 irq_val;
-+ u8 irq_index;
-+ u8 pad[3];
-+ /* cmd word 1 */
-+ __le64 irq_addr;
-+ /* cmd word 2 */
-+ __le32 irq_num;
-+};
-+
-+#define DPRC_ENABLE 0x1
-+
-+struct dprc_cmd_set_irq_enable {
-+ u8 enable;
-+ u8 pad[3];
-+ u8 irq_index;
-+};
-+
-+struct dprc_cmd_set_irq_mask {
-+ __le32 mask;
-+ u8 irq_index;
-+};
-+
-+struct dprc_cmd_get_irq_status {
-+ __le32 status;
-+ u8 irq_index;
-+};
-+
-+struct dprc_rsp_get_irq_status {
-+ __le32 status;
-+};
-+
-+struct dprc_cmd_clear_irq_status {
-+ __le32 status;
-+ u8 irq_index;
-+};
-+
-+struct dprc_rsp_get_attributes {
-+ /* response word 0 */
-+ __le32 container_id;
-+ __le32 icid;
-+ /* response word 1 */
-+ __le32 options;
-+ __le32 portal_id;
-+};
-+
-+struct dprc_rsp_get_obj_count {
-+ __le32 pad;
-+ __le32 obj_count;
-+};
-+
-+struct dprc_cmd_get_obj {
-+ __le32 obj_index;
-+};
-+
-+struct dprc_rsp_get_obj {
-+ /* response word 0 */
-+ __le32 pad0;
-+ __le32 id;
-+ /* response word 1 */
-+ __le16 vendor;
-+ u8 irq_count;
-+ u8 region_count;
-+ __le32 state;
-+ /* response word 2 */
-+ __le16 version_major;
-+ __le16 version_minor;
-+ __le16 flags;
-+ __le16 pad1;
-+ /* response word 3-4 */
-+ u8 type[16];
-+ /* response word 5-6 */
-+ u8 label[16];
-+};
-+
-+struct dprc_cmd_get_obj_region {
-+ /* cmd word 0 */
-+ __le32 obj_id;
-+ __le16 pad0;
-+ u8 region_index;
-+ u8 pad1;
-+ /* cmd word 1-2 */
-+ __le64 pad2[2];
-+ /* cmd word 3-4 */
-+ u8 obj_type[16];
-+};
-+
-+struct dprc_rsp_get_obj_region {
-+ /* response word 0 */
-+ __le64 pad0;
-+ /* response word 1 */
-+ __le32 base_addr;
-+ __le32 pad1;
-+ /* response word 2 */
-+ __le32 size;
-+ u8 type;
-+ u8 pad2[3];
-+ /* response word 3 */
-+ __le32 flags;
-+};
-+
-+struct dprc_cmd_set_obj_irq {
-+ /* cmd word 0 */
-+ __le32 irq_val;
-+ u8 irq_index;
-+ u8 pad[3];
-+ /* cmd word 1 */
-+ __le64 irq_addr;
-+ /* cmd word 2 */
-+ __le32 irq_num;
-+ __le32 obj_id;
-+ /* cmd word 3-4 */
-+ u8 obj_type[16];
-+};
-+
-+/*
-+ * DPRC API for managing and querying DPAA resources
-+ */
-+int dprc_open(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ int container_id,
-+ u16 *token);
-+
-+int dprc_close(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token);
-+
-+/* DPRC IRQ events */
-+
-+/* IRQ event - Indicates that a new object added to the container */
-+#define DPRC_IRQ_EVENT_OBJ_ADDED 0x00000001
-+/* IRQ event - Indicates that an object was removed from the container */
-+#define DPRC_IRQ_EVENT_OBJ_REMOVED 0x00000002
-+/*
-+ * IRQ event - Indicates that one of the descendant containers that opened by
-+ * this container is destroyed
-+ */
-+#define DPRC_IRQ_EVENT_CONTAINER_DESTROYED 0x00000010
-+
-+/*
-+ * IRQ event - Indicates that on one of the container's opened object is
-+ * destroyed
-+ */
-+#define DPRC_IRQ_EVENT_OBJ_DESTROYED 0x00000020
-+
-+/* Irq event - Indicates that object is created at the container */
-+#define DPRC_IRQ_EVENT_OBJ_CREATED 0x00000040
-+
-+/**
-+ * struct dprc_irq_cfg - IRQ configuration
-+ * @paddr: Address that must be written to signal a message-based interrupt
-+ * @val: Value to write into irq_addr address
-+ * @irq_num: A user defined number associated with this IRQ
-+ */
-+struct dprc_irq_cfg {
-+ phys_addr_t paddr;
-+ u32 val;
-+ int irq_num;
-+};
-+
-+int dprc_set_irq(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 irq_index,
-+ struct dprc_irq_cfg *irq_cfg);
-+
-+int dprc_set_irq_enable(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 irq_index,
-+ u8 en);
-+
-+int dprc_set_irq_mask(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 irq_index,
-+ u32 mask);
-+
-+int dprc_get_irq_status(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 irq_index,
-+ u32 *status);
-+
-+int dprc_clear_irq_status(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 irq_index,
-+ u32 status);
-+
-+/**
-+ * struct dprc_attributes - Container attributes
-+ * @container_id: Container's ID
-+ * @icid: Container's ICID
-+ * @portal_id: Container's portal ID
-+ * @options: Container's options as set at container's creation
-+ */
-+struct dprc_attributes {
-+ int container_id;
-+ u32 icid;
-+ int portal_id;
-+ u64 options;
-+};
-+
-+int dprc_get_attributes(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ struct dprc_attributes *attributes);
-+
-+int dprc_get_obj_count(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ int *obj_count);
-+
-+int dprc_get_obj(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ int obj_index,
-+ struct fsl_mc_obj_desc *obj_desc);
-+
-+int dprc_set_obj_irq(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ char *obj_type,
-+ int obj_id,
-+ u8 irq_index,
-+ struct dprc_irq_cfg *irq_cfg);
-+
-+/* Region flags */
-+/* Cacheable - Indicates that region should be mapped as cacheable */
-+#define DPRC_REGION_CACHEABLE 0x00000001
-+
-+/**
-+ * enum dprc_region_type - Region type
-+ * @DPRC_REGION_TYPE_MC_PORTAL: MC portal region
-+ * @DPRC_REGION_TYPE_QBMAN_PORTAL: Qbman portal region
-+ */
-+enum dprc_region_type {
-+ DPRC_REGION_TYPE_MC_PORTAL,
-+ DPRC_REGION_TYPE_QBMAN_PORTAL
-+};
-+
-+#define DPRC_OBJ_FLAG_NO_MEM_SHAREABILITY 0x0001
-+
-+/**
-+ * struct dprc_region_desc - Mappable region descriptor
-+ * @base_offset: Region offset from region's base address.
-+ * For DPMCP and DPRC objects, region base is offset from SoC MC portals
-+ * base address; For DPIO, region base is offset from SoC QMan portals
-+ * base address
-+ * @size: Region size (in bytes)
-+ * @flags: Region attributes
-+ * @type: Portal region type
-+ */
-+struct dprc_region_desc {
-+ u32 base_offset;
-+ u32 size;
-+ u32 flags;
-+ enum dprc_region_type type;
-+};
-+
-+int dprc_get_obj_region(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ char *obj_type,
-+ int obj_id,
-+ u8 region_index,
-+ struct dprc_region_desc *region_desc);
-+
-+int dprc_get_api_version(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 *major_ver,
-+ u16 *minor_ver);
-+
-+int dprc_get_container_id(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ int *container_id);
-+
-+int dprc_reset_container(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ int child_container_id);
-+
-+/*
-+ * Data Path Buffer Pool (DPBP) API
-+ * Contains initialization APIs and runtime control APIs for DPBP
-+ */
-+
-+int dpbp_open(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ int dpbp_id,
-+ u16 *token);
-+
-+int dpbp_close(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token);
-+
-+int dpbp_enable(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token);
-+
-+int dpbp_disable(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token);
-+
-+int dpbp_reset(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token);
-+
-+/**
-+ * struct dpbp_attr - Structure representing DPBP attributes
-+ * @id: DPBP object ID
-+ * @bpid: Hardware buffer pool ID; should be used as an argument in
-+ * acquire/release operations on buffers
-+ */
-+struct dpbp_attr {
-+ int id;
-+ u16 bpid;
-+};
-+
-+int dpbp_get_attributes(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ struct dpbp_attr *attr);
-+
-+/* Data Path Concentrator (DPCON) API
-+ * Contains initialization APIs and runtime control APIs for DPCON
-+ */
-+
-+/**
-+ * Use it to disable notifications; see dpcon_set_notification()
-+ */
-+#define DPCON_INVALID_DPIO_ID (int)(-1)
-+
-+int dpcon_open(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ int dpcon_id,
-+ u16 *token);
-+
-+int dpcon_close(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token);
-+
-+int dpcon_enable(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token);
-+
-+int dpcon_disable(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token);
-+
-+int dpcon_reset(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token);
-+
-+/**
-+ * struct dpcon_attr - Structure representing DPCON attributes
-+ * @id: DPCON object ID
-+ * @qbman_ch_id: Channel ID to be used by dequeue operation
-+ * @num_priorities: Number of priorities for the DPCON channel (1-8)
-+ */
-+struct dpcon_attr {
-+ int id;
-+ u16 qbman_ch_id;
-+ u8 num_priorities;
-+};
-+
-+int dpcon_get_attributes(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ struct dpcon_attr *attr);
-+
-+/**
-+ * struct dpcon_notification_cfg - Structure representing notification params
-+ * @dpio_id: DPIO object ID; must be configured with a notification channel;
-+ * to disable notifications set it to 'DPCON_INVALID_DPIO_ID';
-+ * @priority: Priority selection within the DPIO channel; valid values
-+ * are 0-7, depending on the number of priorities in that channel
-+ * @user_ctx: User context value provided with each CDAN message
-+ */
-+struct dpcon_notification_cfg {
-+ int dpio_id;
-+ u8 priority;
-+ u64 user_ctx;
-+};
-+
-+int dpcon_set_notification(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ struct dpcon_notification_cfg *cfg);
-+
-+struct irq_domain;
-+struct msi_domain_info;
-+
-+/**
-+ * Maximum number of total IRQs that can be pre-allocated for an MC bus'
-+ * IRQ pool
-+ */
-+#define FSL_MC_IRQ_POOL_MAX_TOTAL_IRQS 256
-+
-+/**
-+ * struct fsl_mc_resource_pool - Pool of MC resources of a given
-+ * type
-+ * @type: type of resources in the pool
-+ * @max_count: maximum number of resources in the pool
-+ * @free_count: number of free resources in the pool
-+ * @mutex: mutex to serialize access to the pool's free list
-+ * @free_list: anchor node of list of free resources in the pool
-+ * @mc_bus: pointer to the MC bus that owns this resource pool
-+ */
-+struct fsl_mc_resource_pool {
-+ enum fsl_mc_pool_type type;
-+ int max_count;
-+ int free_count;
-+ struct mutex mutex; /* serializes access to free_list */
-+ struct list_head free_list;
-+ struct fsl_mc_bus *mc_bus;
-+};
-+
-+/**
-+ * struct fsl_mc_restool - information associated with a restool device file
-+ * @cdev: struct char device linked to the root dprc
-+ * @dev: dev_t for the char device to be added
-+ * @device: newly created device in /dev
-+ * @mutex: mutex lock to serialize the open/release operations
-+ * @local_instance_in_use: local MC I/O instance in use or not
-+ * @dynamic_instance_count: number of dynamically created MC I/O instances
-+ */
-+struct fsl_mc_restool {
-+ struct cdev cdev;
-+ dev_t dev;
-+ struct device *device;
-+ struct mutex mutex; /* serialize open/release operations */
-+ bool local_instance_in_use;
-+ u32 dynamic_instance_count;
-+};
-+
-+/**
-+ * struct fsl_mc_bus - logical bus that corresponds to a physical DPRC
-+ * @mc_dev: fsl-mc device for the bus device itself.
-+ * @resource_pools: array of resource pools (one pool per resource type)
-+ * for this MC bus. These resources represent allocatable entities
-+ * from the physical DPRC.
-+ * @irq_resources: Pointer to array of IRQ objects for the IRQ pool
-+ * @scan_mutex: Serializes bus scanning
-+ * @dprc_attr: DPRC attributes
-+ * @restool_misc: struct that abstracts the interaction with userspace restool
-+ */
-+struct fsl_mc_bus {
-+ struct fsl_mc_device mc_dev;
-+ struct fsl_mc_resource_pool resource_pools[FSL_MC_NUM_POOL_TYPES];
-+ struct fsl_mc_device_irq *irq_resources;
-+ struct mutex scan_mutex; /* serializes bus scanning */
-+ struct dprc_attributes dprc_attr;
-+ struct fsl_mc_restool restool_misc;
-+};
-+
-+int dprc_scan_objects(struct fsl_mc_device *mc_bus_dev,
-+ const char *driver_override,
-+ unsigned int *total_irq_count);
-+
-+int fsl_mc_find_msi_domain(struct device *mc_platform_dev,
-+ struct irq_domain **mc_msi_domain);
-+
-+int fsl_mc_populate_irq_pool(struct fsl_mc_bus *mc_bus,
-+ unsigned int irq_count);
-+
-+void fsl_mc_cleanup_irq_pool(struct fsl_mc_bus *mc_bus);
-+
-+void fsl_mc_init_all_resource_pools(struct fsl_mc_device *mc_bus_dev);
-+
-+void fsl_mc_cleanup_all_resource_pools(struct fsl_mc_device *mc_bus_dev);
-+
-+void fsl_mc_get_root_dprc(struct device *dev, struct device **root_dprc_dev);
-+
-+#endif /* _FSL_MC_H_ */
---- /dev/null
-+++ b/include/uapi/linux/fsl_mc.h
-@@ -0,0 +1,31 @@
-+/* SPDX-License-Identifier: GPL-2.0 WITH Linux-syscall-note */
-+/*
-+ * Management Complex (MC) userspace public interface
-+ *
-+ * Copyright 2018 NXP
-+ *
-+ */
-+#ifndef _UAPI_FSL_MC_H_
-+#define _UAPI_FSL_MC_H_
-+
-+#define MC_CMD_NUM_OF_PARAMS 7
-+
-+/**
-+ * struct fsl_mc_command - Management Complex (MC) command structure
-+ * @header: MC command header
-+ * @params: MC command parameters
-+ *
-+ * Used by RESTOOL_SEND_MC_COMMAND
-+ */
-+struct fsl_mc_command {
-+ __u64 header;
-+ __u64 params[MC_CMD_NUM_OF_PARAMS];
-+};
-+
-+#define RESTOOL_IOCTL_TYPE 'R'
-+#define RESTOOL_IOCTL_SEQ 0xE0
-+
-+#define RESTOOL_SEND_MC_COMMAND \
-+ _IOWR(RESTOOL_IOCTL_TYPE, RESTOOL_IOCTL_SEQ, struct fsl_mc_command)
-+
-+#endif /* _UAPI_FSL_MC_H_ */
diff --git a/target/linux/layerscape/patches-4.9/705-dpaa2-support-layerscape.patch b/target/linux/layerscape/patches-4.9/705-dpaa2-support-layerscape.patch
deleted file mode 100644
index 23e4ff9202..0000000000
--- a/target/linux/layerscape/patches-4.9/705-dpaa2-support-layerscape.patch
+++ /dev/null
@@ -1,23038 +0,0 @@
-From 77cc39e936f87463f92f7fddaaf0de51eec3972f Mon Sep 17 00:00:00 2001
-From: Yangbo Lu <yangbo.lu@nxp.com>
-Date: Fri, 6 Jul 2018 15:30:21 +0800
-Subject: [PATCH] dpaa2: support layerscape
-
-This is an integrated patch for layerscape dpaa2 support.
-
-Signed-off-by: Bogdan Purcareata <bogdan.purcareata@nxp.com>
-Signed-off-by: Ioana Radulescu <ruxandra.radulescu@nxp.com>
-Signed-off-by: Razvan Stefanescu <razvan.stefanescu@nxp.com>
-Signed-off-by: costi <constantin.tudor@freescale.com>
-Signed-off-by: Catalin Horghidan <catalin.horghidan@nxp.com>
-Signed-off-by: Mathew McBride <matt@traverse.com.au>
-Signed-off-by: Yangbo Lu <yangbo.lu@nxp.com>
----
- drivers/soc/fsl/ls2-console/Kconfig | 4 +
- drivers/soc/fsl/ls2-console/Makefile | 1 +
- drivers/soc/fsl/ls2-console/ls2-console.c | 284 ++
- drivers/staging/fsl-dpaa2/ethernet/Makefile | 12 +
- drivers/staging/fsl-dpaa2/ethernet/README | 186 +
- drivers/staging/fsl-dpaa2/ethernet/TODO | 18 +
- .../fsl-dpaa2/ethernet/dpaa2-eth-ceetm.c | 1253 ++++++
- .../fsl-dpaa2/ethernet/dpaa2-eth-ceetm.h | 182 +
- .../fsl-dpaa2/ethernet/dpaa2-eth-debugfs.c | 357 ++
- .../fsl-dpaa2/ethernet/dpaa2-eth-debugfs.h | 60 +
- .../fsl-dpaa2/ethernet/dpaa2-eth-trace.h | 185 +
- .../staging/fsl-dpaa2/ethernet/dpaa2-eth.c | 3734 +++++++++++++++++
- .../staging/fsl-dpaa2/ethernet/dpaa2-eth.h | 601 +++
- .../fsl-dpaa2/ethernet/dpaa2-ethtool.c | 878 ++++
- drivers/staging/fsl-dpaa2/ethernet/dpkg.h | 176 +
- drivers/staging/fsl-dpaa2/ethernet/dpni-cmd.h | 719 ++++
- drivers/staging/fsl-dpaa2/ethernet/dpni.c | 2112 ++++++++++
- drivers/staging/fsl-dpaa2/ethernet/dpni.h | 1172 ++++++
- drivers/staging/fsl-dpaa2/ethernet/net.h | 480 +++
- drivers/staging/fsl-dpaa2/ethsw/Makefile | 10 +
- drivers/staging/fsl-dpaa2/ethsw/README | 106 +
- drivers/staging/fsl-dpaa2/ethsw/TODO | 14 +
- drivers/staging/fsl-dpaa2/ethsw/dpsw-cmd.h | 359 ++
- drivers/staging/fsl-dpaa2/ethsw/dpsw.c | 1165 +++++
- drivers/staging/fsl-dpaa2/ethsw/dpsw.h | 592 +++
- .../staging/fsl-dpaa2/ethsw/ethsw-ethtool.c | 206 +
- drivers/staging/fsl-dpaa2/ethsw/ethsw.c | 1438 +++++++
- drivers/staging/fsl-dpaa2/ethsw/ethsw.h | 90 +
- drivers/staging/fsl-dpaa2/evb/Kconfig | 7 +
- drivers/staging/fsl-dpaa2/evb/Makefile | 10 +
- drivers/staging/fsl-dpaa2/evb/dpdmux-cmd.h | 279 ++
- drivers/staging/fsl-dpaa2/evb/dpdmux.c | 1111 +++++
- drivers/staging/fsl-dpaa2/evb/dpdmux.h | 453 ++
- drivers/staging/fsl-dpaa2/evb/evb.c | 1354 ++++++
- drivers/staging/fsl-dpaa2/mac/Kconfig | 23 +
- drivers/staging/fsl-dpaa2/mac/Makefile | 10 +
- drivers/staging/fsl-dpaa2/mac/dpmac-cmd.h | 172 +
- drivers/staging/fsl-dpaa2/mac/dpmac.c | 619 +++
- drivers/staging/fsl-dpaa2/mac/dpmac.h | 342 ++
- drivers/staging/fsl-dpaa2/mac/mac.c | 673 +++
- drivers/staging/fsl-dpaa2/rtc/Makefile | 10 +
- drivers/staging/fsl-dpaa2/rtc/dprtc-cmd.h | 160 +
- drivers/staging/fsl-dpaa2/rtc/dprtc.c | 746 ++++
- drivers/staging/fsl-dpaa2/rtc/dprtc.h | 172 +
- drivers/staging/fsl-dpaa2/rtc/rtc.c | 242 ++
- include/linux/filter.h | 3 +
- 46 files changed, 22780 insertions(+)
- create mode 100644 drivers/soc/fsl/ls2-console/Kconfig
- create mode 100644 drivers/soc/fsl/ls2-console/Makefile
- create mode 100644 drivers/soc/fsl/ls2-console/ls2-console.c
- create mode 100644 drivers/staging/fsl-dpaa2/ethernet/Makefile
- create mode 100644 drivers/staging/fsl-dpaa2/ethernet/README
- create mode 100644 drivers/staging/fsl-dpaa2/ethernet/TODO
- create mode 100644 drivers/staging/fsl-dpaa2/ethernet/dpaa2-eth-ceetm.c
- create mode 100644 drivers/staging/fsl-dpaa2/ethernet/dpaa2-eth-ceetm.h
- create mode 100644 drivers/staging/fsl-dpaa2/ethernet/dpaa2-eth-debugfs.c
- create mode 100644 drivers/staging/fsl-dpaa2/ethernet/dpaa2-eth-debugfs.h
- create mode 100644 drivers/staging/fsl-dpaa2/ethernet/dpaa2-eth-trace.h
- create mode 100644 drivers/staging/fsl-dpaa2/ethernet/dpaa2-eth.c
- create mode 100644 drivers/staging/fsl-dpaa2/ethernet/dpaa2-eth.h
- create mode 100644 drivers/staging/fsl-dpaa2/ethernet/dpaa2-ethtool.c
- create mode 100644 drivers/staging/fsl-dpaa2/ethernet/dpkg.h
- create mode 100644 drivers/staging/fsl-dpaa2/ethernet/dpni-cmd.h
- create mode 100644 drivers/staging/fsl-dpaa2/ethernet/dpni.c
- create mode 100644 drivers/staging/fsl-dpaa2/ethernet/dpni.h
- create mode 100644 drivers/staging/fsl-dpaa2/ethernet/net.h
- create mode 100644 drivers/staging/fsl-dpaa2/ethsw/Makefile
- create mode 100644 drivers/staging/fsl-dpaa2/ethsw/README
- create mode 100644 drivers/staging/fsl-dpaa2/ethsw/TODO
- create mode 100644 drivers/staging/fsl-dpaa2/ethsw/dpsw-cmd.h
- create mode 100644 drivers/staging/fsl-dpaa2/ethsw/dpsw.c
- create mode 100644 drivers/staging/fsl-dpaa2/ethsw/dpsw.h
- create mode 100644 drivers/staging/fsl-dpaa2/ethsw/ethsw-ethtool.c
- create mode 100644 drivers/staging/fsl-dpaa2/ethsw/ethsw.c
- create mode 100644 drivers/staging/fsl-dpaa2/ethsw/ethsw.h
- create mode 100644 drivers/staging/fsl-dpaa2/evb/Kconfig
- create mode 100644 drivers/staging/fsl-dpaa2/evb/Makefile
- create mode 100644 drivers/staging/fsl-dpaa2/evb/dpdmux-cmd.h
- create mode 100644 drivers/staging/fsl-dpaa2/evb/dpdmux.c
- create mode 100644 drivers/staging/fsl-dpaa2/evb/dpdmux.h
- create mode 100644 drivers/staging/fsl-dpaa2/evb/evb.c
- create mode 100644 drivers/staging/fsl-dpaa2/mac/Kconfig
- create mode 100644 drivers/staging/fsl-dpaa2/mac/Makefile
- create mode 100644 drivers/staging/fsl-dpaa2/mac/dpmac-cmd.h
- create mode 100644 drivers/staging/fsl-dpaa2/mac/dpmac.c
- create mode 100644 drivers/staging/fsl-dpaa2/mac/dpmac.h
- create mode 100644 drivers/staging/fsl-dpaa2/mac/mac.c
- create mode 100644 drivers/staging/fsl-dpaa2/rtc/Makefile
- create mode 100644 drivers/staging/fsl-dpaa2/rtc/dprtc-cmd.h
- create mode 100644 drivers/staging/fsl-dpaa2/rtc/dprtc.c
- create mode 100644 drivers/staging/fsl-dpaa2/rtc/dprtc.h
- create mode 100644 drivers/staging/fsl-dpaa2/rtc/rtc.c
-
---- /dev/null
-+++ b/drivers/soc/fsl/ls2-console/Kconfig
-@@ -0,0 +1,4 @@
-+config FSL_LS2_CONSOLE
-+ tristate "Layerscape MC and AIOP console support"
-+ depends on ARCH_LAYERSCAPE
-+ default y
---- /dev/null
-+++ b/drivers/soc/fsl/ls2-console/Makefile
-@@ -0,0 +1 @@
-+obj-$(CONFIG_FSL_LS2_CONSOLE) += ls2-console.o
---- /dev/null
-+++ b/drivers/soc/fsl/ls2-console/ls2-console.c
-@@ -0,0 +1,284 @@
-+/* Copyright 2015-2016 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of the above-listed copyright holders nor the
-+ * names of any contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
-+ * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
-+ * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
-+ * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE
-+ * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
-+ * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
-+ * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
-+ * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
-+ * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
-+ * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
-+ * POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#include <linux/miscdevice.h>
-+#include <linux/uaccess.h>
-+#include <linux/poll.h>
-+#include <linux/compat.h>
-+#include <linux/module.h>
-+#include <linux/slab.h>
-+#include <linux/io.h>
-+
-+/* SoC address for the MC firmware base low/high registers */
-+#define SOC_CCSR_MC_FW_BASE_ADDR_REGS 0x8340020
-+#define SOC_CCSR_MC_FW_BASE_ADDR_REGS_SIZE 2
-+/* MC firmware base low/high registers indexes */
-+#define MCFBALR_OFFSET 0
-+#define MCFBAHR_OFFSET 1
-+
-+/* Bit mask used to obtain the most significant part of the MC base address */
-+#define MC_FW_HIGH_ADDR_MASK 0x1FFFF
-+/* Bit mask used to obtain the least significant part of the MC base address */
-+#define MC_FW_LOW_ADDR_MASK 0xE0000000
-+
-+#define MC_BUFFER_OFFSET 0x01000000
-+#define MC_BUFFER_SIZE (1024*1024*16)
-+#define MC_OFFSET_DELTA (MC_BUFFER_OFFSET)
-+
-+#define AIOP_BUFFER_OFFSET 0x06000000
-+#define AIOP_BUFFER_SIZE (1024*1024*16)
-+#define AIOP_OFFSET_DELTA (0)
-+
-+struct log_header {
-+ char magic_word[8]; /* magic word */
-+ uint32_t buf_start; /* holds the 32-bit little-endian
-+ * offset of the start of the buffer
-+ */
-+ uint32_t buf_length; /* holds the 32-bit little-endian
-+ * length of the buffer
-+ */
-+ uint32_t last_byte; /* holds the 32-bit little-endian offset
-+ * of the byte after the last byte that
-+ * was written
-+ */
-+ char reserved[44];
-+};
-+
-+#define LOG_HEADER_FLAG_BUFFER_WRAPAROUND 0x80000000
-+#define LOG_VERSION_MAJOR 1
-+#define LOG_VERSION_MINOR 0
-+
-+
-+#define invalidate(p) { asm volatile("dc ivac, %0" : : "r" (p) : "memory"); }
-+
-+struct console_data {
-+ char *map_addr;
-+ struct log_header *hdr;
-+ char *start_addr; /* Start of buffer */
-+ char *end_addr; /* End of buffer */
-+ char *end_of_data; /* Current end of data */
-+ char *cur_ptr; /* Last data sent to console */
-+};
-+
-+#define LAST_BYTE(a) ((a) & ~(LOG_HEADER_FLAG_BUFFER_WRAPAROUND))
-+
-+static inline void __adjust_end(struct console_data *cd)
-+{
-+ cd->end_of_data = cd->start_addr
-+ + LAST_BYTE(le32_to_cpu(cd->hdr->last_byte));
-+}
-+
-+static inline void adjust_end(struct console_data *cd)
-+{
-+ invalidate(cd->hdr);
-+ __adjust_end(cd);
-+}
-+
-+static inline uint64_t get_mc_fw_base_address(void)
-+{
-+ u32 *mcfbaregs = (u32 *) ioremap(SOC_CCSR_MC_FW_BASE_ADDR_REGS,
-+ SOC_CCSR_MC_FW_BASE_ADDR_REGS_SIZE);
-+ u64 mcfwbase = 0ULL;
-+
-+ mcfwbase = readl(mcfbaregs + MCFBAHR_OFFSET) & MC_FW_HIGH_ADDR_MASK;
-+ mcfwbase <<= 32;
-+ mcfwbase |= readl(mcfbaregs + MCFBALR_OFFSET) & MC_FW_LOW_ADDR_MASK;
-+ iounmap(mcfbaregs);
-+ pr_info("fsl-ls2-console: MC base address at 0x%016llx\n", mcfwbase);
-+ return mcfwbase;
-+}
-+
-+static int fsl_ls2_generic_console_open(struct inode *node, struct file *fp,
-+ u64 offset, u64 size,
-+ uint8_t *emagic, uint8_t magic_len,
-+ u32 offset_delta)
-+{
-+ struct console_data *cd;
-+ uint8_t *magic;
-+ uint32_t wrapped;
-+
-+ cd = kmalloc(sizeof(*cd), GFP_KERNEL);
-+ if (cd == NULL)
-+ return -ENOMEM;
-+ fp->private_data = cd;
-+ cd->map_addr = ioremap(get_mc_fw_base_address() + offset, size);
-+
-+ cd->hdr = (struct log_header *) cd->map_addr;
-+ invalidate(cd->hdr);
-+
-+ magic = cd->hdr->magic_word;
-+ if (memcmp(magic, emagic, magic_len)) {
-+ pr_info("magic didn't match!\n");
-+ pr_info("expected: %02x %02x %02x %02x %02x %02x %02x %02x\n",
-+ emagic[0], emagic[1], emagic[2], emagic[3],
-+ emagic[4], emagic[5], emagic[6], emagic[7]);
-+ pr_info(" seen: %02x %02x %02x %02x %02x %02x %02x %02x\n",
-+ magic[0], magic[1], magic[2], magic[3],
-+ magic[4], magic[5], magic[6], magic[7]);
-+ kfree(cd);
-+ iounmap(cd->map_addr);
-+ return -EIO;
-+ }
-+
-+ cd->start_addr = cd->map_addr
-+ + le32_to_cpu(cd->hdr->buf_start) - offset_delta;
-+ cd->end_addr = cd->start_addr + le32_to_cpu(cd->hdr->buf_length);
-+
-+ wrapped = le32_to_cpu(cd->hdr->last_byte)
-+ & LOG_HEADER_FLAG_BUFFER_WRAPAROUND;
-+
-+ __adjust_end(cd);
-+ if (wrapped && (cd->end_of_data != cd->end_addr))
-+ cd->cur_ptr = cd->end_of_data+1;
-+ else
-+ cd->cur_ptr = cd->start_addr;
-+
-+ return 0;
-+}
-+
-+static int fsl_ls2_mc_console_open(struct inode *node, struct file *fp)
-+{
-+ uint8_t magic_word[] = { 0, 1, 'C', 'M' };
-+
-+ return fsl_ls2_generic_console_open(node, fp,
-+ MC_BUFFER_OFFSET, MC_BUFFER_SIZE,
-+ magic_word, sizeof(magic_word),
-+ MC_OFFSET_DELTA);
-+}
-+
-+static int fsl_ls2_aiop_console_open(struct inode *node, struct file *fp)
-+{
-+ uint8_t magic_word[] = { 'P', 'O', 'I', 'A' };
-+
-+ return fsl_ls2_generic_console_open(node, fp,
-+ AIOP_BUFFER_OFFSET, AIOP_BUFFER_SIZE,
-+ magic_word, sizeof(magic_word),
-+ AIOP_OFFSET_DELTA);
-+}
-+
-+static int fsl_ls2_console_close(struct inode *node, struct file *fp)
-+{
-+ struct console_data *cd = fp->private_data;
-+
-+ iounmap(cd->map_addr);
-+ kfree(cd);
-+ return 0;
-+}
-+
-+ssize_t fsl_ls2_console_read(struct file *fp, char __user *buf, size_t count,
-+ loff_t *f_pos)
-+{
-+ struct console_data *cd = fp->private_data;
-+ size_t bytes = 0;
-+ char data;
-+
-+ /* Check if we need to adjust the end of data addr */
-+ adjust_end(cd);
-+
-+ while ((count != bytes) && (cd->end_of_data != cd->cur_ptr)) {
-+ if (((u64)cd->cur_ptr) % 64 == 0)
-+ invalidate(cd->cur_ptr);
-+
-+ data = *(cd->cur_ptr);
-+ if (copy_to_user(&buf[bytes], &data, 1))
-+ return -EFAULT;
-+ cd->cur_ptr++;
-+ if (cd->cur_ptr >= cd->end_addr)
-+ cd->cur_ptr = cd->start_addr;
-+ ++bytes;
-+ }
-+ return bytes;
-+}
-+
-+static const struct file_operations fsl_ls2_mc_console_fops = {
-+ .owner = THIS_MODULE,
-+ .open = fsl_ls2_mc_console_open,
-+ .release = fsl_ls2_console_close,
-+ .read = fsl_ls2_console_read,
-+};
-+
-+static struct miscdevice fsl_ls2_mc_console_dev = {
-+ .minor = MISC_DYNAMIC_MINOR,
-+ .name = "fsl_mc_console",
-+ .fops = &fsl_ls2_mc_console_fops
-+};
-+
-+static const struct file_operations fsl_ls2_aiop_console_fops = {
-+ .owner = THIS_MODULE,
-+ .open = fsl_ls2_aiop_console_open,
-+ .release = fsl_ls2_console_close,
-+ .read = fsl_ls2_console_read,
-+};
-+
-+static struct miscdevice fsl_ls2_aiop_console_dev = {
-+ .minor = MISC_DYNAMIC_MINOR,
-+ .name = "fsl_aiop_console",
-+ .fops = &fsl_ls2_aiop_console_fops
-+};
-+
-+static int __init fsl_ls2_console_init(void)
-+{
-+ int err = 0;
-+
-+ pr_info("Freescale LS2 console driver\n");
-+ err = misc_register(&fsl_ls2_mc_console_dev);
-+ if (err) {
-+ pr_err("fsl_mc_console: cannot register device\n");
-+ return err;
-+ }
-+ pr_info("fsl-ls2-console: device %s registered\n",
-+ fsl_ls2_mc_console_dev.name);
-+
-+ err = misc_register(&fsl_ls2_aiop_console_dev);
-+ if (err) {
-+ pr_err("fsl_aiop_console: cannot register device\n");
-+ return err;
-+ }
-+ pr_info("fsl-ls2-console: device %s registered\n",
-+ fsl_ls2_aiop_console_dev.name);
-+
-+ return 0;
-+}
-+
-+static void __exit fsl_ls2_console_exit(void)
-+{
-+ misc_deregister(&fsl_ls2_mc_console_dev);
-+
-+ misc_deregister(&fsl_ls2_aiop_console_dev);
-+}
-+
-+module_init(fsl_ls2_console_init);
-+module_exit(fsl_ls2_console_exit);
-+
-+MODULE_AUTHOR("Roy Pledge <roy.pledge@freescale.com>");
-+MODULE_LICENSE("Dual BSD/GPL");
-+MODULE_DESCRIPTION("Freescale LS2 console driver");
---- /dev/null
-+++ b/drivers/staging/fsl-dpaa2/ethernet/Makefile
-@@ -0,0 +1,12 @@
-+#
-+# Makefile for the Freescale DPAA2 Ethernet controller
-+#
-+
-+obj-$(CONFIG_FSL_DPAA2_ETH) += fsl-dpaa2-eth.o
-+
-+fsl-dpaa2-eth-objs := dpaa2-eth.o dpaa2-ethtool.o dpni.o
-+fsl-dpaa2-eth-${CONFIG_FSL_DPAA2_ETH_DEBUGFS} += dpaa2-eth-debugfs.o
-+fsl-dpaa2-eth-${CONFIG_FSL_DPAA2_ETH_CEETM} += dpaa2-eth-ceetm.o
-+
-+# Needed by the tracing framework
-+CFLAGS_dpaa2-eth.o := -I$(src)
---- /dev/null
-+++ b/drivers/staging/fsl-dpaa2/ethernet/README
-@@ -0,0 +1,186 @@
-+Freescale DPAA2 Ethernet driver
-+===============================
-+
-+This file provides documentation for the Freescale DPAA2 Ethernet driver.
-+
-+
-+Contents
-+========
-+ Supported Platforms
-+ Architecture Overview
-+ Creating a Network Interface
-+ Features & Offloads
-+
-+
-+Supported Platforms
-+===================
-+This driver provides networking support for Freescale DPAA2 SoCs, e.g.
-+LS2080A, LS2088A, LS1088A.
-+
-+
-+Architecture Overview
-+=====================
-+Unlike regular NICs, in the DPAA2 architecture there is no single hardware block
-+representing network interfaces; instead, several separate hardware resources
-+concur to provide the networking functionality:
-+ - network interfaces
-+ - queues, channels
-+ - buffer pools
-+ - MAC/PHY
-+
-+All hardware resources are allocated and configured through the Management
-+Complex (MC) portals. MC abstracts most of these resources as DPAA2 objects
-+and exposes ABIs through which they can be configured and controlled. A few
-+hardware resources, like queues, do not have a corresponding MC object and
-+are treated as internal resources of other objects.
-+
-+For a more detailed description of the DPAA2 architecture and its object
-+abstractions see:
-+ drivers/staging/fsl-mc/README.txt
-+
-+Each Linux net device is built on top of a Datapath Network Interface (DPNI)
-+object and uses Buffer Pools (DPBPs), I/O Portals (DPIOs) and Concentrators
-+(DPCONs).
-+
-+Configuration interface:
-+
-+ -----------------------
-+ | DPAA2 Ethernet Driver |
-+ -----------------------
-+ . . .
-+ . . .
-+ . . . . . . . . . . . .
-+ . . .
-+ . . .
-+ ---------- ---------- -----------
-+ | DPBP API | | DPNI API | | DPCON API |
-+ ---------- ---------- -----------
-+ . . . software
-+=========== . ========== . ============ . ===================
-+ . . . hardware
-+ ------------------------------------------
-+ | MC hardware portals |
-+ ------------------------------------------
-+ . . .
-+ . . .
-+ ------ ------ -------
-+ | DPBP | | DPNI | | DPCON |
-+ ------ ------ -------
-+
-+The DPNIs are network interfaces without a direct one-on-one mapping to PHYs.
-+DPBPs represent hardware buffer pools. Packet I/O is performed in the context
-+of DPCON objects, using DPIO portals for managing and communicating with the
-+hardware resources.
-+
-+Datapath (I/O) interface:
-+
-+ -----------------------------------------------
-+ | DPAA2 Ethernet Driver |
-+ -----------------------------------------------
-+ | ^ ^ | |
-+ | | | | |
-+ enqueue| dequeue| data | dequeue| seed |
-+ (Tx) | (Rx, TxC)| avail.| request| buffers|
-+ | | notify| | |
-+ | | | | |
-+ V | | V V
-+ -----------------------------------------------
-+ | DPIO Driver |
-+ -----------------------------------------------
-+ | | | | | software
-+ | | | | | ================
-+ | | | | | hardware
-+ -----------------------------------------------
-+ | I/O hardware portals |
-+ -----------------------------------------------
-+ | ^ ^ | |
-+ | | | | |
-+ | | | V |
-+ V | ================ V
-+ ---------------------- | -------------
-+ queues ---------------------- | | Buffer pool |
-+ ---------------------- | -------------
-+ =======================
-+ Channel
-+
-+Datapath I/O (DPIO) portals provide enqueue and dequeue services, data
-+availability notifications and buffer pool management. DPIOs are shared between
-+all DPAA2 objects (and implicitly all DPAA2 kernel drivers) that work with data
-+frames, but must be affine to the CPUs for the purpose of traffic distribution.
-+
-+Frames are transmitted and received through hardware frame queues, which can be
-+grouped in channels for the purpose of hardware scheduling. The Ethernet driver
-+enqueues TX frames on egress queues and after transmission is complete a TX
-+confirmation frame is sent back to the CPU.
-+
-+When frames are available on ingress queues, a data availability notification
-+is sent to the CPU; notifications are raised per channel, so even if multiple
-+queues in the same channel have available frames, only one notification is sent.
-+After a channel fires a notification, is must be explicitly rearmed.
-+
-+Each network interface can have multiple Rx, Tx and confirmation queues affined
-+to CPUs, and one channel (DPCON) for each CPU that services at least one queue.
-+DPCONs are used to distribute ingress traffic to different CPUs via the cores'
-+affine DPIOs.
-+
-+The role of hardware buffer pools is storage of ingress frame data. Each network
-+interface has a privately owned buffer pool which it seeds with kernel allocated
-+buffers.
-+
-+
-+DPNIs are decoupled from PHYs; a DPNI can be connected to a PHY through a DPMAC
-+object or to another DPNI through an internal link, but the connection is
-+managed by MC and completely transparent to the Ethernet driver.
-+
-+ --------- --------- ---------
-+ | eth if1 | | eth if2 | | eth ifn |
-+ --------- --------- ---------
-+ . . .
-+ . . .
-+ . . .
-+ ---------------------------
-+ | DPAA2 Ethernet Driver |
-+ ---------------------------
-+ . . .
-+ . . .
-+ . . .
-+ ------ ------ ------ -------
-+ | DPNI | | DPNI | | DPNI | | DPMAC |----+
-+ ------ ------ ------ ------- |
-+ | | | | |
-+ | | | | -----
-+ =========== ================== | PHY |
-+ -----
-+
-+Creating a Network Interface
-+============================
-+A net device is created for each DPNI object probed on the MC bus. Each DPNI has
-+a number of properties which determine the network interface configuration
-+options and associated hardware resources.
-+
-+DPNI objects (and the other DPAA2 objects needed for a network interface) can be
-+added to a container on the MC bus in one of two ways: statically, through a
-+Datapath Layout Binary file (DPL) that is parsed by MC at boot time; or created
-+dynamically at runtime, via the DPAA2 objects APIs.
-+
-+
-+Features & Offloads
-+===================
-+Hardware checksum offloading is supported for TCP and UDP over IPv4/6 frames.
-+The checksum offloads can be independently configured on RX and TX through
-+ethtool.
-+
-+Hardware offload of unicast and multicast MAC filtering is supported on the
-+ingress path and permanently enabled.
-+
-+Scatter-gather frames are supported on both RX and TX paths. On TX, SG support
-+is configurable via ethtool; on RX it is always enabled.
-+
-+The DPAA2 hardware can process jumbo Ethernet frames of up to 10K bytes.
-+
-+The Ethernet driver defines a static flow hashing scheme that distributes
-+traffic based on a 5-tuple key: src IP, dst IP, IP proto, L4 src port,
-+L4 dst port. No user configuration is supported for now.
-+
-+Hardware specific statistics for the network interface as well as some
-+non-standard driver stats can be consulted through ethtool -S option.
---- /dev/null
-+++ b/drivers/staging/fsl-dpaa2/ethernet/TODO
-@@ -0,0 +1,18 @@
-+* Add a DPAA2 MAC kernel driver in order to allow PHY management; currently
-+ the DPMAC objects and their link to DPNIs are handled by MC internally
-+ and all PHYs are seen as fixed-link
-+* add more debug support: decide how to expose detailed debug statistics,
-+ add ingress error queue support
-+* MC firmware uprev; the DPAA2 objects used by the Ethernet driver need to
-+ be kept in sync with binary interface changes in MC
-+* refine README file
-+* cleanup
-+
-+NOTE: None of the above is must-have before getting the DPAA2 Ethernet driver
-+out of staging. The main requirement for that is to have the drivers it
-+depends on, fsl-mc bus and DPIO driver, moved to drivers/bus and drivers/soc
-+respectively.
-+
-+ Please send any patches to Greg Kroah-Hartman <gregkh@linuxfoundation.org>,
-+ ruxandra.radulescu@nxp.com, devel@driverdev.osuosl.org,
-+ linux-kernel@vger.kernel.org
---- /dev/null
-+++ b/drivers/staging/fsl-dpaa2/ethernet/dpaa2-eth-ceetm.c
-@@ -0,0 +1,1253 @@
-+// SPDX-License-Identifier: (GPL-2.0+ OR BSD-3-Clause)
-+/*
-+ * Copyright 2017 NXP
-+ *
-+ */
-+
-+#include <linux/init.h>
-+#include <linux/module.h>
-+
-+#include "dpaa2-eth-ceetm.h"
-+#include "dpaa2-eth.h"
-+
-+#define DPAA2_CEETM_DESCRIPTION "FSL DPAA2 CEETM qdisc"
-+/* Conversion formula from userspace passed Bps to expected Mbit */
-+#define dpaa2_eth_bps_to_mbit(rate) (rate >> 17)
-+
-+static const struct nla_policy dpaa2_ceetm_policy[DPAA2_CEETM_TCA_MAX] = {
-+ [DPAA2_CEETM_TCA_COPT] = { .len = sizeof(struct dpaa2_ceetm_tc_copt) },
-+ [DPAA2_CEETM_TCA_QOPS] = { .len = sizeof(struct dpaa2_ceetm_tc_qopt) },
-+};
-+
-+struct Qdisc_ops dpaa2_ceetm_qdisc_ops;
-+
-+static inline int dpaa2_eth_set_ch_shaping(struct dpaa2_eth_priv *priv,
-+ struct dpni_tx_shaping_cfg *scfg,
-+ struct dpni_tx_shaping_cfg *ecfg,
-+ int coupled, int ch_id)
-+{
-+ int err = 0;
-+
-+ netdev_dbg(priv->net_dev, "%s: ch_id %d rate %d mbps\n", __func__,
-+ ch_id, scfg->rate_limit);
-+ err = dpni_set_tx_shaping(priv->mc_io, 0, priv->mc_token, scfg,
-+ ecfg, coupled);
-+ if (err)
-+ netdev_err(priv->net_dev, "dpni_set_tx_shaping err\n");
-+
-+ return err;
-+}
-+
-+static inline int dpaa2_eth_reset_ch_shaping(struct dpaa2_eth_priv *priv,
-+ int ch_id)
-+{
-+ struct dpni_tx_shaping_cfg cfg = { 0 };
-+
-+ return dpaa2_eth_set_ch_shaping(priv, &cfg, &cfg, 0, ch_id);
-+}
-+
-+static inline int
-+dpaa2_eth_update_shaping_cfg(struct net_device *dev,
-+ struct dpaa2_ceetm_shaping_cfg cfg,
-+ struct dpni_tx_shaping_cfg *scfg,
-+ struct dpni_tx_shaping_cfg *ecfg)
-+{
-+ scfg->rate_limit = dpaa2_eth_bps_to_mbit(cfg.cir);
-+ ecfg->rate_limit = dpaa2_eth_bps_to_mbit(cfg.eir);
-+
-+ if (cfg.cbs > DPAA2_ETH_MAX_BURST_SIZE) {
-+ netdev_err(dev, "Committed burst size must be under %d\n",
-+ DPAA2_ETH_MAX_BURST_SIZE);
-+ return -EINVAL;
-+ }
-+
-+ scfg->max_burst_size = cfg.cbs;
-+
-+ if (cfg.ebs > DPAA2_ETH_MAX_BURST_SIZE) {
-+ netdev_err(dev, "Excess burst size must be under %d\n",
-+ DPAA2_ETH_MAX_BURST_SIZE);
-+ return -EINVAL;
-+ }
-+
-+ ecfg->max_burst_size = cfg.ebs;
-+
-+ if ((!cfg.cir || !cfg.eir) && cfg.coupled) {
-+ netdev_err(dev, "Coupling can be set when both CIR and EIR are finite\n");
-+ return -EINVAL;
-+ }
-+
-+ return 0;
-+}
-+
-+enum update_tx_prio {
-+ DPAA2_ETH_ADD_CQ,
-+ DPAA2_ETH_DEL_CQ,
-+};
-+
-+/* Normalize weights based on max passed value */
-+static inline int dpaa2_eth_normalize_tx_prio(struct dpaa2_ceetm_qdisc *priv)
-+{
-+ struct dpni_tx_schedule_cfg *sched_cfg;
-+ struct dpaa2_ceetm_class *cl;
-+ u32 qpri;
-+ u16 weight_max = 0, increment;
-+ int i;
-+
-+ /* Check the boundaries of the provided values */
-+ for (i = 0; i < priv->clhash.hashsize; i++)
-+ hlist_for_each_entry(cl, &priv->clhash.hash[i], common.hnode)
-+ weight_max = (weight_max == 0 ? cl->prio.weight :
-+ (weight_max < cl->prio.weight ?
-+ cl->prio.weight : weight_max));
-+
-+ /* If there are no elements, there's nothing to do */
-+ if (weight_max == 0)
-+ return 0;
-+
-+ increment = (DPAA2_CEETM_MAX_WEIGHT - DPAA2_CEETM_MIN_WEIGHT) /
-+ weight_max;
-+
-+ for (i = 0; i < priv->clhash.hashsize; i++) {
-+ hlist_for_each_entry(cl, &priv->clhash.hash[i], common.hnode) {
-+ if (cl->prio.mode == STRICT_PRIORITY)
-+ continue;
-+
-+ qpri = cl->prio.qpri;
-+ sched_cfg = &priv->prio.tx_prio_cfg.tc_sched[qpri];
-+
-+ sched_cfg->delta_bandwidth =
-+ DPAA2_CEETM_MIN_WEIGHT +
-+ (cl->prio.weight * increment);
-+
-+ pr_debug("%s: Normalized CQ qpri %d weight to %d\n",
-+ __func__, qpri, sched_cfg->delta_bandwidth);
-+ }
-+ }
-+
-+ return 0;
-+}
-+
-+static inline int dpaa2_eth_update_tx_prio(struct dpaa2_eth_priv *priv,
-+ struct dpaa2_ceetm_class *cl,
-+ enum update_tx_prio type)
-+{
-+ struct dpaa2_ceetm_qdisc *sch = qdisc_priv(cl->parent);
-+ struct dpni_congestion_notification_cfg notif_cfg = {0};
-+ struct dpni_tx_schedule_cfg *sched_cfg;
-+ struct dpni_taildrop td = {0};
-+ u8 ch_id = 0, tc_id = 0;
-+ u32 qpri = 0;
-+ int err = 0;
-+
-+ qpri = cl->prio.qpri;
-+ tc_id = DPNI_BUILD_CH_TC(ch_id, qpri);
-+
-+ switch (type) {
-+ case DPAA2_ETH_ADD_CQ:
-+ /* Disable congestion notifications */
-+ notif_cfg.threshold_entry = 0;
-+ notif_cfg.threshold_exit = 0;
-+ err = dpni_set_congestion_notification(priv->mc_io, 0,
-+ priv->mc_token,
-+ DPNI_QUEUE_TX, tc_id,
-+ &notif_cfg);
-+ if (err) {
-+ netdev_err(priv->net_dev, "Error disabling congestion notifications %d\n",
-+ err);
-+ return err;
-+ }
-+ /* Enable taildrop */
-+ td.enable = 1;
-+ td.units = DPNI_CONGESTION_UNIT_FRAMES;
-+ td.threshold = DPAA2_CEETM_TD_THRESHOLD;
-+ err = dpni_set_taildrop(priv->mc_io, 0, priv->mc_token,
-+ DPNI_CP_GROUP, DPNI_QUEUE_TX, tc_id,
-+ 0, &td);
-+ if (err) {
-+ netdev_err(priv->net_dev, "Error enabling Tx taildrop %d\n",
-+ err);
-+ return err;
-+ }
-+ break;
-+ case DPAA2_ETH_DEL_CQ:
-+ /* Disable taildrop */
-+ td.enable = 0;
-+ err = dpni_set_taildrop(priv->mc_io, 0, priv->mc_token,
-+ DPNI_CP_GROUP, DPNI_QUEUE_TX, tc_id,
-+ 0, &td);
-+ if (err) {
-+ netdev_err(priv->net_dev, "Error disabling Tx taildrop %d\n",
-+ err);
-+ return err;
-+ }
-+ /* Enable congestion notifications */
-+ notif_cfg.units = DPNI_CONGESTION_UNIT_BYTES;
-+ notif_cfg.threshold_entry = DPAA2_ETH_TX_CONG_ENTRY_THRESH;
-+ notif_cfg.threshold_exit = DPAA2_ETH_TX_CONG_EXIT_THRESH;
-+ notif_cfg.message_ctx = (u64)priv;
-+ notif_cfg.message_iova = priv->cscn_dma;
-+ notif_cfg.notification_mode = DPNI_CONG_OPT_WRITE_MEM_ON_ENTER |
-+ DPNI_CONG_OPT_WRITE_MEM_ON_EXIT |
-+ DPNI_CONG_OPT_COHERENT_WRITE;
-+ err = dpni_set_congestion_notification(priv->mc_io, 0,
-+ priv->mc_token,
-+ DPNI_QUEUE_TX, tc_id,
-+ &notif_cfg);
-+ if (err) {
-+ netdev_err(priv->net_dev, "Error enabling congestion notifications %d\n",
-+ err);
-+ return err;
-+ }
-+ break;
-+ }
-+
-+ /* We can zero out the structure in the tx_prio_conf array */
-+ if (type == DPAA2_ETH_DEL_CQ) {
-+ sched_cfg = &sch->prio.tx_prio_cfg.tc_sched[qpri];
-+ memset(sched_cfg, 0, sizeof(*sched_cfg));
-+ }
-+
-+ /* Normalize priorities */
-+ err = dpaa2_eth_normalize_tx_prio(sch);
-+
-+ /* Debug print goes here */
-+ print_hex_dump_debug("tx_prio: ", DUMP_PREFIX_OFFSET, 16, 1,
-+ &sch->prio.tx_prio_cfg,
-+ sizeof(sch->prio.tx_prio_cfg), 0);
-+
-+ /* Call dpni_set_tx_priorities for the entire prio qdisc */
-+ err = dpni_set_tx_priorities(priv->mc_io, 0, priv->mc_token,
-+ &sch->prio.tx_prio_cfg);
-+ if (err)
-+ netdev_err(priv->net_dev, "dpni_set_tx_priorities err %d\n",
-+ err);
-+
-+ return err;
-+}
-+
-+static void dpaa2_eth_ceetm_enable(struct dpaa2_eth_priv *priv)
-+{
-+ priv->ceetm_en = true;
-+}
-+
-+static void dpaa2_eth_ceetm_disable(struct dpaa2_eth_priv *priv)
-+{
-+ priv->ceetm_en = false;
-+}
-+
-+/* Find class in qdisc hash table using given handle */
-+static inline struct dpaa2_ceetm_class *dpaa2_ceetm_find(u32 handle,
-+ struct Qdisc *sch)
-+{
-+ struct dpaa2_ceetm_qdisc *priv = qdisc_priv(sch);
-+ struct Qdisc_class_common *clc;
-+
-+ pr_debug(KBUILD_BASENAME " : %s : find class %X in qdisc %X\n",
-+ __func__, handle, sch->handle);
-+
-+ clc = qdisc_class_find(&priv->clhash, handle);
-+ return clc ? container_of(clc, struct dpaa2_ceetm_class, common) : NULL;
-+}
-+
-+/* Insert a class in the qdisc's class hash */
-+static void dpaa2_ceetm_link_class(struct Qdisc *sch,
-+ struct Qdisc_class_hash *clhash,
-+ struct Qdisc_class_common *common)
-+{
-+ sch_tree_lock(sch);
-+ qdisc_class_hash_insert(clhash, common);
-+ sch_tree_unlock(sch);
-+ qdisc_class_hash_grow(sch, clhash);
-+}
-+
-+/* Destroy a ceetm class */
-+static void dpaa2_ceetm_cls_destroy(struct Qdisc *sch,
-+ struct dpaa2_ceetm_class *cl)
-+{
-+ struct net_device *dev = qdisc_dev(sch);
-+ struct dpaa2_eth_priv *priv = netdev_priv(dev);
-+
-+ if (!cl)
-+ return;
-+
-+ pr_debug(KBUILD_BASENAME " : %s : destroy class %X from under %X\n",
-+ __func__, cl->common.classid, sch->handle);
-+
-+ /* Recurse into child first */
-+ if (cl->child) {
-+ qdisc_destroy(cl->child);
-+ cl->child = NULL;
-+ }
-+
-+ switch (cl->type) {
-+ case CEETM_ROOT:
-+ if (dpaa2_eth_reset_ch_shaping(priv, cl->root.ch_id))
-+ netdev_err(dev, "Error resetting channel shaping\n");
-+
-+ break;
-+
-+ case CEETM_PRIO:
-+ if (dpaa2_eth_update_tx_prio(priv, cl, DPAA2_ETH_DEL_CQ))
-+ netdev_err(dev, "Error resetting tx_priorities\n");
-+
-+ if (cl->prio.cstats)
-+ free_percpu(cl->prio.cstats);
-+
-+ break;
-+ }
-+
-+ tcf_destroy_chain(&cl->filter_list);
-+ kfree(cl);
-+}
-+
-+/* Destroy a ceetm qdisc */
-+static void dpaa2_ceetm_destroy(struct Qdisc *sch)
-+{
-+ unsigned int i;
-+ struct hlist_node *next;
-+ struct dpaa2_ceetm_class *cl;
-+ struct dpaa2_ceetm_qdisc *priv = qdisc_priv(sch);
-+ struct net_device *dev = qdisc_dev(sch);
-+ struct dpaa2_eth_priv *priv_eth = netdev_priv(dev);
-+
-+ pr_debug(KBUILD_BASENAME " : %s : destroy qdisc %X\n",
-+ __func__, sch->handle);
-+
-+ /* All filters need to be removed before destroying the classes */
-+ tcf_destroy_chain(&priv->filter_list);
-+
-+ for (i = 0; i < priv->clhash.hashsize; i++) {
-+ hlist_for_each_entry(cl, &priv->clhash.hash[i], common.hnode)
-+ tcf_destroy_chain(&cl->filter_list);
-+ }
-+
-+ for (i = 0; i < priv->clhash.hashsize; i++) {
-+ hlist_for_each_entry_safe(cl, next, &priv->clhash.hash[i],
-+ common.hnode)
-+ dpaa2_ceetm_cls_destroy(sch, cl);
-+ }
-+
-+ qdisc_class_hash_destroy(&priv->clhash);
-+
-+ switch (priv->type) {
-+ case CEETM_ROOT:
-+ dpaa2_eth_ceetm_disable(priv_eth);
-+
-+ if (priv->root.qstats)
-+ free_percpu(priv->root.qstats);
-+
-+ if (!priv->root.qdiscs)
-+ break;
-+
-+ /* Destroy the pfifo qdiscs in case they haven't been attached
-+ * to the netdev queues yet.
-+ */
-+ for (i = 0; i < dev->num_tx_queues; i++)
-+ if (priv->root.qdiscs[i])
-+ qdisc_destroy(priv->root.qdiscs[i]);
-+
-+ kfree(priv->root.qdiscs);
-+ break;
-+
-+ case CEETM_PRIO:
-+ if (priv->prio.parent)
-+ priv->prio.parent->child = NULL;
-+ break;
-+ }
-+}
-+
-+static int dpaa2_ceetm_dump(struct Qdisc *sch, struct sk_buff *skb)
-+{
-+ struct Qdisc *qdisc;
-+ unsigned int ntx, i;
-+ struct nlattr *nest;
-+ struct dpaa2_ceetm_tc_qopt qopt;
-+ struct dpaa2_ceetm_qdisc_stats *qstats;
-+ struct net_device *dev = qdisc_dev(sch);
-+ struct dpaa2_ceetm_qdisc *priv = qdisc_priv(sch);
-+
-+ pr_debug(KBUILD_BASENAME " : %s : qdisc %X\n", __func__, sch->handle);
-+
-+ sch_tree_lock(sch);
-+ memset(&qopt, 0, sizeof(qopt));
-+ qopt.type = priv->type;
-+ qopt.shaped = priv->shaped;
-+
-+ switch (priv->type) {
-+ case CEETM_ROOT:
-+ /* Gather statistics from the underlying pfifo qdiscs */
-+ sch->q.qlen = 0;
-+ memset(&sch->bstats, 0, sizeof(sch->bstats));
-+ memset(&sch->qstats, 0, sizeof(sch->qstats));
-+
-+ for (ntx = 0; ntx < dev->num_tx_queues; ntx++) {
-+ qdisc = netdev_get_tx_queue(dev, ntx)->qdisc_sleeping;
-+ sch->q.qlen += qdisc->q.qlen;
-+ sch->bstats.bytes += qdisc->bstats.bytes;
-+ sch->bstats.packets += qdisc->bstats.packets;
-+ sch->qstats.qlen += qdisc->qstats.qlen;
-+ sch->qstats.backlog += qdisc->qstats.backlog;
-+ sch->qstats.drops += qdisc->qstats.drops;
-+ sch->qstats.requeues += qdisc->qstats.requeues;
-+ sch->qstats.overlimits += qdisc->qstats.overlimits;
-+ }
-+
-+ for_each_online_cpu(i) {
-+ qstats = per_cpu_ptr(priv->root.qstats, i);
-+ sch->qstats.drops += qstats->drops;
-+ }
-+
-+ break;
-+
-+ case CEETM_PRIO:
-+ qopt.prio_group_A = priv->prio.tx_prio_cfg.prio_group_A;
-+ qopt.prio_group_B = priv->prio.tx_prio_cfg.prio_group_B;
-+ qopt.separate_groups = priv->prio.tx_prio_cfg.separate_groups;
-+ break;
-+
-+ default:
-+ pr_err(KBUILD_BASENAME " : %s : invalid qdisc\n", __func__);
-+ sch_tree_unlock(sch);
-+ return -EINVAL;
-+ }
-+
-+ nest = nla_nest_start(skb, TCA_OPTIONS);
-+ if (!nest)
-+ goto nla_put_failure;
-+ if (nla_put(skb, DPAA2_CEETM_TCA_QOPS, sizeof(qopt), &qopt))
-+ goto nla_put_failure;
-+ nla_nest_end(skb, nest);
-+
-+ sch_tree_unlock(sch);
-+ return skb->len;
-+
-+nla_put_failure:
-+ sch_tree_unlock(sch);
-+ nla_nest_cancel(skb, nest);
-+ return -EMSGSIZE;
-+}
-+
-+static int dpaa2_ceetm_change_prio(struct Qdisc *sch,
-+ struct dpaa2_ceetm_qdisc *priv,
-+ struct dpaa2_ceetm_tc_qopt *qopt)
-+{
-+ /* TODO: Once LX2 support is added */
-+ /* priv->shaped = parent_cl->shaped; */
-+ priv->prio.tx_prio_cfg.prio_group_A = qopt->prio_group_A;
-+ priv->prio.tx_prio_cfg.prio_group_B = qopt->prio_group_B;
-+ priv->prio.tx_prio_cfg.separate_groups = qopt->separate_groups;
-+
-+ return 0;
-+}
-+
-+/* Edit a ceetm qdisc */
-+static int dpaa2_ceetm_change(struct Qdisc *sch, struct nlattr *opt)
-+{
-+ struct dpaa2_ceetm_qdisc *priv = qdisc_priv(sch);
-+ struct nlattr *tb[DPAA2_CEETM_TCA_QOPS + 1];
-+ struct dpaa2_ceetm_tc_qopt *qopt;
-+ int err;
-+
-+ pr_debug(KBUILD_BASENAME " : %s : qdisc %X\n", __func__, sch->handle);
-+
-+ err = nla_parse_nested(tb, DPAA2_CEETM_TCA_QOPS, opt,
-+ dpaa2_ceetm_policy);
-+ if (err < 0) {
-+ pr_err(KBUILD_BASENAME " : %s : tc error in %s\n", __func__,
-+ "nla_parse_nested");
-+ return err;
-+ }
-+
-+ if (!tb[DPAA2_CEETM_TCA_QOPS]) {
-+ pr_err(KBUILD_BASENAME " : %s : tc error in %s\n", __func__,
-+ "tb");
-+ return -EINVAL;
-+ }
-+
-+ if (TC_H_MIN(sch->handle)) {
-+ pr_err("CEETM: a qdisc should not have a minor\n");
-+ return -EINVAL;
-+ }
-+
-+ qopt = nla_data(tb[DPAA2_CEETM_TCA_QOPS]);
-+
-+ if (priv->type != qopt->type) {
-+ pr_err("CEETM: qdisc %X is not of the provided type\n",
-+ sch->handle);
-+ return -EINVAL;
-+ }
-+
-+ switch (priv->type) {
-+ case CEETM_PRIO:
-+ err = dpaa2_ceetm_change_prio(sch, priv, qopt);
-+ break;
-+ default:
-+ pr_err(KBUILD_BASENAME " : %s : invalid qdisc\n", __func__);
-+ err = -EINVAL;
-+ }
-+
-+ return err;
-+}
-+
-+/* Configure a root ceetm qdisc */
-+static int dpaa2_ceetm_init_root(struct Qdisc *sch,
-+ struct dpaa2_ceetm_qdisc *priv,
-+ struct dpaa2_ceetm_tc_qopt *qopt)
-+{
-+ struct net_device *dev = qdisc_dev(sch);
-+ struct dpaa2_eth_priv *priv_eth = netdev_priv(dev);
-+ struct netdev_queue *dev_queue;
-+ unsigned int i, parent_id;
-+ struct Qdisc *qdisc;
-+ int err;
-+
-+ pr_debug(KBUILD_BASENAME " : %s : qdisc %X\n", __func__, sch->handle);
-+
-+ /* Validate inputs */
-+ if (sch->parent != TC_H_ROOT) {
-+ pr_err("CEETM: a root ceetm qdisc can not be attached to a class\n");
-+ tcf_destroy_chain(&priv->filter_list);
-+ qdisc_class_hash_destroy(&priv->clhash);
-+ return -EINVAL;
-+ }
-+
-+ /* Pre-allocate underlying pfifo qdiscs.
-+ *
-+ * We want to offload shaping and scheduling decisions to the hardware.
-+ * The pfifo qdiscs will be attached to the netdev queues and will
-+ * guide the traffic from the IP stack down to the driver with minimum
-+ * interference.
-+ *
-+ * The CEETM qdiscs and classes will be crossed when the traffic
-+ * reaches the driver.
-+ */
-+ priv->root.qdiscs = kcalloc(dev->num_tx_queues,
-+ sizeof(priv->root.qdiscs[0]),
-+ GFP_KERNEL);
-+ if (!priv->root.qdiscs) {
-+ err = -ENOMEM;
-+ goto err_init_root;
-+ }
-+
-+ for (i = 0; i < dev->num_tx_queues; i++) {
-+ dev_queue = netdev_get_tx_queue(dev, i);
-+ parent_id = TC_H_MAKE(TC_H_MAJ(sch->handle),
-+ TC_H_MIN(i + PFIFO_MIN_OFFSET));
-+
-+ qdisc = qdisc_create_dflt(dev_queue, &pfifo_qdisc_ops,
-+ parent_id);
-+ if (!qdisc) {
-+ err = -ENOMEM;
-+ goto err_init_root;
-+ }
-+
-+ priv->root.qdiscs[i] = qdisc;
-+ qdisc->flags |= TCQ_F_ONETXQUEUE;
-+ }
-+
-+ sch->flags |= TCQ_F_MQROOT;
-+
-+ priv->root.qstats = alloc_percpu(struct dpaa2_ceetm_qdisc_stats);
-+ if (!priv->root.qstats) {
-+ pr_err(KBUILD_BASENAME " : %s : alloc_percpu() failed\n",
-+ __func__);
-+ err = -ENOMEM;
-+ goto err_init_root;
-+ }
-+
-+ dpaa2_eth_ceetm_enable(priv_eth);
-+ return 0;
-+
-+err_init_root:
-+ dpaa2_ceetm_destroy(sch);
-+ return err;
-+}
-+
-+/* Configure a prio ceetm qdisc */
-+static int dpaa2_ceetm_init_prio(struct Qdisc *sch,
-+ struct dpaa2_ceetm_qdisc *priv,
-+ struct dpaa2_ceetm_tc_qopt *qopt)
-+{
-+ struct net_device *dev = qdisc_dev(sch);
-+ struct dpaa2_ceetm_class *parent_cl;
-+ struct Qdisc *parent_qdisc;
-+ int err;
-+
-+ pr_debug(KBUILD_BASENAME " : %s : qdisc %X\n", __func__, sch->handle);
-+
-+ if (sch->parent == TC_H_ROOT) {
-+ pr_err("CEETM: a prio ceetm qdisc can not be root\n");
-+ err = -EINVAL;
-+ goto err_init_prio;
-+ }
-+
-+ parent_qdisc = qdisc_lookup(dev, TC_H_MAJ(sch->parent));
-+ if (strcmp(parent_qdisc->ops->id, dpaa2_ceetm_qdisc_ops.id)) {
-+ pr_err("CEETM: a ceetm qdisc can not be attached to other qdisc/class types\n");
-+ err = -EINVAL;
-+ goto err_init_prio;
-+ }
-+
-+ /* Obtain the parent root ceetm_class */
-+ parent_cl = dpaa2_ceetm_find(sch->parent, parent_qdisc);
-+
-+ if (!parent_cl || parent_cl->type != CEETM_ROOT) {
-+ pr_err("CEETM: a prio ceetm qdiscs can be added only under a root ceetm class\n");
-+ err = -EINVAL;
-+ goto err_init_prio;
-+ }
-+
-+ priv->prio.parent = parent_cl;
-+ parent_cl->child = sch;
-+
-+ err = dpaa2_ceetm_change_prio(sch, priv, qopt);
-+
-+ return 0;
-+
-+err_init_prio:
-+ dpaa2_ceetm_destroy(sch);
-+ return err;
-+}
-+
-+/* Configure a generic ceetm qdisc */
-+static int dpaa2_ceetm_init(struct Qdisc *sch, struct nlattr *opt)
-+{
-+ struct dpaa2_ceetm_qdisc *priv = qdisc_priv(sch);
-+ struct net_device *dev = qdisc_dev(sch);
-+ struct nlattr *tb[DPAA2_CEETM_TCA_QOPS + 1];
-+ struct dpaa2_ceetm_tc_qopt *qopt;
-+ int err;
-+
-+ pr_debug(KBUILD_BASENAME " : %s : qdisc %X\n", __func__, sch->handle);
-+
-+ if (!netif_is_multiqueue(dev))
-+ return -EOPNOTSUPP;
-+
-+ RCU_INIT_POINTER(priv->filter_list, NULL);
-+
-+ if (!opt) {
-+ pr_err(KBUILD_BASENAME " : %s : tc error - opt = NULL\n",
-+ __func__);
-+ return -EINVAL;
-+ }
-+
-+ err = nla_parse_nested(tb, DPAA2_CEETM_TCA_QOPS, opt,
-+ dpaa2_ceetm_policy);
-+ if (err < 0) {
-+ pr_err(KBUILD_BASENAME " : %s : tc error in %s\n", __func__,
-+ "nla_parse_nested");
-+ return err;
-+ }
-+
-+ if (!tb[DPAA2_CEETM_TCA_QOPS]) {
-+ pr_err(KBUILD_BASENAME " : %s : tc error in %s\n", __func__,
-+ "tb");
-+ return -EINVAL;
-+ }
-+
-+ if (TC_H_MIN(sch->handle)) {
-+ pr_err("CEETM: a qdisc should not have a minor\n");
-+ return -EINVAL;
-+ }
-+
-+ qopt = nla_data(tb[DPAA2_CEETM_TCA_QOPS]);
-+
-+ /* Initialize the class hash list. Each qdisc has its own class hash */
-+ err = qdisc_class_hash_init(&priv->clhash);
-+ if (err < 0) {
-+ pr_err(KBUILD_BASENAME " : %s : qdisc_class_hash_init failed\n",
-+ __func__);
-+ return err;
-+ }
-+
-+ priv->type = qopt->type;
-+ priv->shaped = qopt->shaped;
-+
-+ switch (priv->type) {
-+ case CEETM_ROOT:
-+ err = dpaa2_ceetm_init_root(sch, priv, qopt);
-+ break;
-+ case CEETM_PRIO:
-+ err = dpaa2_ceetm_init_prio(sch, priv, qopt);
-+ break;
-+ default:
-+ pr_err(KBUILD_BASENAME " : %s : invalid qdisc\n", __func__);
-+ dpaa2_ceetm_destroy(sch);
-+ err = -EINVAL;
-+ }
-+
-+ return err;
-+}
-+
-+/* Attach the underlying pfifo qdiscs */
-+static void dpaa2_ceetm_attach(struct Qdisc *sch)
-+{
-+ struct net_device *dev = qdisc_dev(sch);
-+ struct dpaa2_ceetm_qdisc *priv = qdisc_priv(sch);
-+ struct Qdisc *qdisc, *old_qdisc;
-+ unsigned int i;
-+
-+ pr_debug(KBUILD_BASENAME " : %s : qdisc %X\n", __func__, sch->handle);
-+
-+ for (i = 0; i < dev->num_tx_queues; i++) {
-+ qdisc = priv->root.qdiscs[i];
-+ old_qdisc = dev_graft_qdisc(qdisc->dev_queue, qdisc);
-+ if (old_qdisc)
-+ qdisc_destroy(old_qdisc);
-+ }
-+
-+ /* Remove the references to the pfifo qdiscs since the kernel will
-+ * destroy them when needed. No cleanup from our part is required from
-+ * this point on.
-+ */
-+ kfree(priv->root.qdiscs);
-+ priv->root.qdiscs = NULL;
-+}
-+
-+static unsigned long dpaa2_ceetm_cls_get(struct Qdisc *sch, u32 classid)
-+{
-+ struct dpaa2_ceetm_class *cl;
-+
-+ pr_debug(KBUILD_BASENAME " : %s : classid %X from qdisc %X\n",
-+ __func__, classid, sch->handle);
-+ cl = dpaa2_ceetm_find(classid, sch);
-+
-+ if (cl)
-+ cl->refcnt++;
-+
-+ return (unsigned long)cl;
-+}
-+
-+static void dpaa2_ceetm_cls_put(struct Qdisc *sch, unsigned long arg)
-+{
-+ struct dpaa2_ceetm_class *cl = (struct dpaa2_ceetm_class *)arg;
-+ pr_debug(KBUILD_BASENAME " : %s : classid %X from qdisc %X\n",
-+ __func__, cl->common.classid, sch->handle);
-+ cl->refcnt--;
-+
-+ if (cl->refcnt == 0)
-+ dpaa2_ceetm_cls_destroy(sch, cl);
-+}
-+
-+static int dpaa2_ceetm_cls_change_root(struct dpaa2_ceetm_class *cl,
-+ struct dpaa2_ceetm_tc_copt *copt,
-+ struct net_device *dev)
-+{
-+ struct dpaa2_eth_priv *priv = netdev_priv(dev);
-+ struct dpni_tx_shaping_cfg scfg = { 0 }, ecfg = { 0 };
-+ int err = 0;
-+
-+ pr_debug(KBUILD_BASENAME " : %s : class %X\n", __func__,
-+ cl->common.classid);
-+
-+ if (!cl->shaped)
-+ return 0;
-+
-+ if (dpaa2_eth_update_shaping_cfg(dev, copt->shaping_cfg,
-+ &scfg, &ecfg))
-+ return -EINVAL;
-+
-+ err = dpaa2_eth_set_ch_shaping(priv, &scfg, &ecfg,
-+ copt->shaping_cfg.coupled,
-+ cl->root.ch_id);
-+ if (err)
-+ return err;
-+
-+ memcpy(&cl->root.shaping_cfg, &copt->shaping_cfg,
-+ sizeof(struct dpaa2_ceetm_shaping_cfg));
-+
-+ return err;
-+}
-+
-+static int dpaa2_ceetm_cls_change_prio(struct dpaa2_ceetm_class *cl,
-+ struct dpaa2_ceetm_tc_copt *copt,
-+ struct net_device *dev)
-+{
-+ struct dpaa2_ceetm_qdisc *sch = qdisc_priv(cl->parent);
-+ struct dpni_tx_schedule_cfg *sched_cfg;
-+ struct dpaa2_eth_priv *priv = netdev_priv(dev);
-+ int err;
-+
-+ pr_debug(KBUILD_BASENAME " : %s : class %X mode %d weight %d\n",
-+ __func__, cl->common.classid, copt->mode, copt->weight);
-+
-+ if (!cl->prio.cstats) {
-+ cl->prio.cstats = alloc_percpu(struct dpaa2_ceetm_class_stats);
-+ if (!cl->prio.cstats) {
-+ pr_err(KBUILD_BASENAME " : %s : alloc_percpu() failed\n",
-+ __func__);
-+ return -ENOMEM;
-+ }
-+ }
-+
-+ cl->prio.mode = copt->mode;
-+ cl->prio.weight = copt->weight;
-+
-+ sched_cfg = &sch->prio.tx_prio_cfg.tc_sched[cl->prio.qpri];
-+
-+ switch (copt->mode) {
-+ case STRICT_PRIORITY:
-+ sched_cfg->mode = DPNI_TX_SCHED_STRICT_PRIORITY;
-+ break;
-+ case WEIGHTED_A:
-+ sched_cfg->mode = DPNI_TX_SCHED_WEIGHTED_A;
-+ break;
-+ case WEIGHTED_B:
-+ sched_cfg->mode = DPNI_TX_SCHED_WEIGHTED_B;
-+ break;
-+ }
-+
-+ err = dpaa2_eth_update_tx_prio(priv, cl, DPAA2_ETH_ADD_CQ);
-+
-+ return err;
-+}
-+
-+/* Add a new ceetm class */
-+static int dpaa2_ceetm_cls_add(struct Qdisc *sch, u32 classid,
-+ struct dpaa2_ceetm_tc_copt *copt,
-+ unsigned long *arg)
-+{
-+ struct dpaa2_ceetm_qdisc *priv = qdisc_priv(sch);
-+ struct net_device *dev = qdisc_dev(sch);
-+ struct dpaa2_eth_priv *priv_eth = netdev_priv(dev);
-+ struct dpaa2_ceetm_class *cl;
-+ int err;
-+
-+ if (copt->type == CEETM_ROOT &&
-+ priv->clhash.hashelems == dpaa2_eth_ch_count(priv_eth)) {
-+ pr_err("CEETM: only %d channel%s per DPNI allowed, sorry\n",
-+ dpaa2_eth_ch_count(priv_eth),
-+ dpaa2_eth_ch_count(priv_eth) == 1 ? "" : "s");
-+ return -EINVAL;
-+ }
-+
-+ if (copt->type == CEETM_PRIO &&
-+ priv->clhash.hashelems == dpaa2_eth_tc_count(priv_eth)) {
-+ pr_err("CEETM: only %d queue%s per channel allowed, sorry\n",
-+ dpaa2_eth_tc_count(priv_eth),
-+ dpaa2_eth_tc_count(priv_eth) == 1 ? "" : "s");
-+ return -EINVAL;
-+ }
-+
-+ cl = kzalloc(sizeof(*cl), GFP_KERNEL);
-+ if (!cl)
-+ return -ENOMEM;
-+
-+ RCU_INIT_POINTER(cl->filter_list, NULL);
-+
-+ cl->common.classid = classid;
-+ cl->refcnt = 1;
-+ cl->parent = sch;
-+ cl->child = NULL;
-+
-+ /* Add class handle in Qdisc */
-+ dpaa2_ceetm_link_class(sch, &priv->clhash, &cl->common);
-+
-+ cl->shaped = copt->shaped;
-+ cl->type = copt->type;
-+
-+ /* Claim a CEETM channel / tc - DPAA2. will assume transition from
-+ * classid to qdid/qpri, starting from qdid / qpri 0
-+ */
-+ switch (copt->type) {
-+ case CEETM_ROOT:
-+ cl->root.ch_id = classid - sch->handle - 1;
-+ err = dpaa2_ceetm_cls_change_root(cl, copt, dev);
-+ break;
-+ case CEETM_PRIO:
-+ cl->prio.qpri = classid - sch->handle - 1;
-+ err = dpaa2_ceetm_cls_change_prio(cl, copt, dev);
-+ break;
-+ default:
-+ err = -EINVAL;
-+ break;
-+ }
-+
-+ if (err) {
-+ pr_err("%s: Unable to set new %s class\n", __func__,
-+ (copt->type == CEETM_ROOT ? "root" : "prio"));
-+ goto out_free;
-+ }
-+
-+ switch (copt->type) {
-+ case CEETM_ROOT:
-+ pr_debug(KBUILD_BASENAME " : %s : configured root class %X associated with channel qdid %d\n",
-+ __func__, classid, cl->root.ch_id);
-+ break;
-+ case CEETM_PRIO:
-+ pr_debug(KBUILD_BASENAME " : %s : configured prio class %X associated with queue qpri %d\n",
-+ __func__, classid, cl->prio.qpri);
-+ break;
-+ }
-+
-+ *arg = (unsigned long)cl;
-+ return 0;
-+
-+out_free:
-+ kfree(cl);
-+ return err;
-+}
-+
-+/* Add or configure a ceetm class */
-+static int dpaa2_ceetm_cls_change(struct Qdisc *sch, u32 classid, u32 parentid,
-+ struct nlattr **tca, unsigned long *arg)
-+{
-+ struct dpaa2_ceetm_qdisc *priv;
-+ struct dpaa2_ceetm_class *cl = (struct dpaa2_ceetm_class *)*arg;
-+ struct nlattr *opt = tca[TCA_OPTIONS];
-+ struct nlattr *tb[DPAA2_CEETM_TCA_MAX];
-+ struct dpaa2_ceetm_tc_copt *copt;
-+ struct net_device *dev = qdisc_dev(sch);
-+ int err;
-+
-+ pr_debug(KBUILD_BASENAME " : %s : classid %X under qdisc %X\n",
-+ __func__, classid, sch->handle);
-+
-+ if (strcmp(sch->ops->id, dpaa2_ceetm_qdisc_ops.id)) {
-+ pr_err("CEETM: a ceetm class can not be attached to other qdisc/class types\n");
-+ return -EINVAL;
-+ }
-+
-+ priv = qdisc_priv(sch);
-+
-+ if (!opt) {
-+ pr_err(KBUILD_BASENAME " : %s : tc error NULL opt\n", __func__);
-+ return -EINVAL;
-+ }
-+
-+ err = nla_parse_nested(tb, DPAA2_CEETM_TCA_COPT, opt,
-+ dpaa2_ceetm_policy);
-+ if (err < 0) {
-+ pr_err(KBUILD_BASENAME " : %s : tc error in %s\n", __func__,
-+ "nla_parse_nested");
-+ return -EINVAL;
-+ }
-+
-+ if (!tb[DPAA2_CEETM_TCA_COPT]) {
-+ pr_err(KBUILD_BASENAME " : %s : tc error in %s\n", __func__,
-+ "tb");
-+ return -EINVAL;
-+ }
-+
-+ copt = nla_data(tb[DPAA2_CEETM_TCA_COPT]);
-+
-+ /* Configure an existing ceetm class */
-+ if (cl) {
-+ if (copt->type != cl->type) {
-+ pr_err("CEETM: class %X is not of the provided type\n",
-+ cl->common.classid);
-+ return -EINVAL;
-+ }
-+
-+ switch (copt->type) {
-+ case CEETM_ROOT:
-+ return dpaa2_ceetm_cls_change_root(cl, copt, dev);
-+ case CEETM_PRIO:
-+ return dpaa2_ceetm_cls_change_prio(cl, copt, dev);
-+
-+ default:
-+ pr_err(KBUILD_BASENAME " : %s : invalid class\n",
-+ __func__);
-+ return -EINVAL;
-+ }
-+ }
-+
-+ return dpaa2_ceetm_cls_add(sch, classid, copt, arg);
-+}
-+
-+static void dpaa2_ceetm_cls_walk(struct Qdisc *sch, struct qdisc_walker *arg)
-+{
-+ struct dpaa2_ceetm_qdisc *priv = qdisc_priv(sch);
-+ struct dpaa2_ceetm_class *cl;
-+ unsigned int i;
-+
-+ pr_debug(KBUILD_BASENAME " : %s : qdisc %X\n", __func__, sch->handle);
-+
-+ if (arg->stop)
-+ return;
-+
-+ for (i = 0; i < priv->clhash.hashsize; i++) {
-+ hlist_for_each_entry(cl, &priv->clhash.hash[i], common.hnode) {
-+ if (arg->count < arg->skip) {
-+ arg->count++;
-+ continue;
-+ }
-+ if (arg->fn(sch, (unsigned long)cl, arg) < 0) {
-+ arg->stop = 1;
-+ return;
-+ }
-+ arg->count++;
-+ }
-+ }
-+}
-+
-+static int dpaa2_ceetm_cls_dump(struct Qdisc *sch, unsigned long arg,
-+ struct sk_buff *skb, struct tcmsg *tcm)
-+{
-+ struct dpaa2_ceetm_class *cl = (struct dpaa2_ceetm_class *)arg;
-+ struct nlattr *nest;
-+ struct dpaa2_ceetm_tc_copt copt;
-+
-+ pr_debug(KBUILD_BASENAME " : %s : class %X under qdisc %X\n",
-+ __func__, cl->common.classid, sch->handle);
-+
-+ sch_tree_lock(sch);
-+
-+ tcm->tcm_parent = ((struct Qdisc *)cl->parent)->handle;
-+ tcm->tcm_handle = cl->common.classid;
-+
-+ memset(&copt, 0, sizeof(copt));
-+
-+ copt.shaped = cl->shaped;
-+ copt.type = cl->type;
-+
-+ switch (cl->type) {
-+ case CEETM_ROOT:
-+ if (cl->child)
-+ tcm->tcm_info = cl->child->handle;
-+
-+ memcpy(&copt.shaping_cfg, &cl->root.shaping_cfg,
-+ sizeof(struct dpaa2_ceetm_shaping_cfg));
-+
-+ break;
-+
-+ case CEETM_PRIO:
-+ if (cl->child)
-+ tcm->tcm_info = cl->child->handle;
-+
-+ copt.mode = cl->prio.mode;
-+ copt.weight = cl->prio.weight;
-+
-+ break;
-+ }
-+
-+ nest = nla_nest_start(skb, TCA_OPTIONS);
-+ if (!nest)
-+ goto nla_put_failure;
-+ if (nla_put(skb, DPAA2_CEETM_TCA_COPT, sizeof(copt), &copt))
-+ goto nla_put_failure;
-+ nla_nest_end(skb, nest);
-+ sch_tree_unlock(sch);
-+ return skb->len;
-+
-+nla_put_failure:
-+ sch_tree_unlock(sch);
-+ nla_nest_cancel(skb, nest);
-+ return -EMSGSIZE;
-+}
-+
-+static int dpaa2_ceetm_cls_delete(struct Qdisc *sch, unsigned long arg)
-+{
-+ struct dpaa2_ceetm_qdisc *priv = qdisc_priv(sch);
-+ struct dpaa2_ceetm_class *cl = (struct dpaa2_ceetm_class *)arg;
-+
-+ pr_debug(KBUILD_BASENAME " : %s : class %X under qdisc %X\n",
-+ __func__, cl->common.classid, sch->handle);
-+
-+ sch_tree_lock(sch);
-+ qdisc_class_hash_remove(&priv->clhash, &cl->common);
-+ cl->refcnt--;
-+ WARN_ON(cl->refcnt == 0);
-+ sch_tree_unlock(sch);
-+ return 0;
-+}
-+
-+/* Get the class' child qdisc, if any */
-+static struct Qdisc *dpaa2_ceetm_cls_leaf(struct Qdisc *sch, unsigned long arg)
-+{
-+ struct dpaa2_ceetm_class *cl = (struct dpaa2_ceetm_class *)arg;
-+
-+ pr_debug(KBUILD_BASENAME " : %s : class %X under qdisc %X\n",
-+ __func__, cl->common.classid, sch->handle);
-+
-+ switch (cl->type) {
-+ case CEETM_ROOT:
-+ case CEETM_PRIO:
-+ return cl->child;
-+ }
-+
-+ return NULL;
-+}
-+
-+static int dpaa2_ceetm_cls_graft(struct Qdisc *sch, unsigned long arg,
-+ struct Qdisc *new, struct Qdisc **old)
-+{
-+ if (new && strcmp(new->ops->id, dpaa2_ceetm_qdisc_ops.id)) {
-+ pr_err("CEETM: only ceetm qdiscs can be attached to ceetm classes\n");
-+ return -EOPNOTSUPP;
-+ }
-+
-+ return 0;
-+}
-+
-+static int dpaa2_ceetm_cls_dump_stats(struct Qdisc *sch, unsigned long arg,
-+ struct gnet_dump *d)
-+{
-+ struct dpaa2_ceetm_class *cl = (struct dpaa2_ceetm_class *)arg;
-+ struct gnet_stats_basic_packed tmp_bstats;
-+ struct dpaa2_ceetm_tc_xstats xstats;
-+ union dpni_statistics dpni_stats;
-+ struct net_device *dev = qdisc_dev(sch);
-+ struct dpaa2_eth_priv *priv_eth = netdev_priv(dev);
-+ u8 ch_id = 0;
-+ int err;
-+
-+ memset(&xstats, 0, sizeof(xstats));
-+ memset(&tmp_bstats, 0, sizeof(tmp_bstats));
-+
-+ if (cl->type == CEETM_ROOT)
-+ return 0;
-+
-+ err = dpni_get_statistics(priv_eth->mc_io, 0, priv_eth->mc_token, 3,
-+ DPNI_BUILD_CH_TC(ch_id, cl->prio.qpri),
-+ &dpni_stats);
-+ if (err)
-+ netdev_warn(dev, "dpni_get_stats(%d) failed - %d\n", 3, err);
-+
-+ xstats.ceetm_dequeue_bytes = dpni_stats.page_3.ceetm_dequeue_bytes;
-+ xstats.ceetm_dequeue_frames = dpni_stats.page_3.ceetm_dequeue_frames;
-+ xstats.ceetm_reject_bytes = dpni_stats.page_3.ceetm_reject_bytes;
-+ xstats.ceetm_reject_frames = dpni_stats.page_3.ceetm_reject_frames;
-+
-+ return gnet_stats_copy_app(d, &xstats, sizeof(xstats));
-+}
-+
-+static struct tcf_proto __rcu **dpaa2_ceetm_tcf_chain(struct Qdisc *sch,
-+ unsigned long arg)
-+{
-+ struct dpaa2_ceetm_qdisc *priv = qdisc_priv(sch);
-+ struct dpaa2_ceetm_class *cl = (struct dpaa2_ceetm_class *)arg;
-+
-+ pr_debug(KBUILD_BASENAME " : %s : class %X under qdisc %X\n", __func__,
-+ cl ? cl->common.classid : 0, sch->handle);
-+ return cl ? &cl->filter_list : &priv->filter_list;
-+}
-+
-+static unsigned long dpaa2_ceetm_tcf_bind(struct Qdisc *sch,
-+ unsigned long parent,
-+ u32 classid)
-+{
-+ struct dpaa2_ceetm_class *cl = dpaa2_ceetm_find(classid, sch);
-+
-+ pr_debug(KBUILD_BASENAME " : %s : class %X under qdisc %X\n", __func__,
-+ cl ? cl->common.classid : 0, sch->handle);
-+ return (unsigned long)cl;
-+}
-+
-+static void dpaa2_ceetm_tcf_unbind(struct Qdisc *sch, unsigned long arg)
-+{
-+ struct dpaa2_ceetm_class *cl = (struct dpaa2_ceetm_class *)arg;
-+
-+ pr_debug(KBUILD_BASENAME " : %s : class %X under qdisc %X\n", __func__,
-+ cl ? cl->common.classid : 0, sch->handle);
-+}
-+
-+const struct Qdisc_class_ops dpaa2_ceetm_cls_ops = {
-+ .graft = dpaa2_ceetm_cls_graft,
-+ .leaf = dpaa2_ceetm_cls_leaf,
-+ .get = dpaa2_ceetm_cls_get,
-+ .put = dpaa2_ceetm_cls_put,
-+ .change = dpaa2_ceetm_cls_change,
-+ .delete = dpaa2_ceetm_cls_delete,
-+ .walk = dpaa2_ceetm_cls_walk,
-+ .tcf_chain = dpaa2_ceetm_tcf_chain,
-+ .bind_tcf = dpaa2_ceetm_tcf_bind,
-+ .unbind_tcf = dpaa2_ceetm_tcf_unbind,
-+ .dump = dpaa2_ceetm_cls_dump,
-+ .dump_stats = dpaa2_ceetm_cls_dump_stats,
-+};
-+
-+struct Qdisc_ops dpaa2_ceetm_qdisc_ops __read_mostly = {
-+ .id = "ceetm",
-+ .priv_size = sizeof(struct dpaa2_ceetm_qdisc),
-+ .cl_ops = &dpaa2_ceetm_cls_ops,
-+ .init = dpaa2_ceetm_init,
-+ .destroy = dpaa2_ceetm_destroy,
-+ .change = dpaa2_ceetm_change,
-+ .dump = dpaa2_ceetm_dump,
-+ .attach = dpaa2_ceetm_attach,
-+ .owner = THIS_MODULE,
-+};
-+
-+/* Run the filters and classifiers attached to the qdisc on the provided skb */
-+int dpaa2_ceetm_classify(struct sk_buff *skb, struct Qdisc *sch,
-+ int *qdid, int *qpri)
-+{
-+ struct dpaa2_ceetm_qdisc *priv = qdisc_priv(sch);
-+ struct dpaa2_ceetm_class *cl = NULL;
-+ struct tcf_result res;
-+ struct tcf_proto *tcf;
-+ int result;
-+
-+ tcf = rcu_dereference_bh(priv->filter_list);
-+ while (tcf && (result = tc_classify(skb, tcf, &res, false)) >= 0) {
-+#ifdef CONFIG_NET_CLS_ACT
-+ switch (result) {
-+ case TC_ACT_QUEUED:
-+ case TC_ACT_STOLEN:
-+ case TC_ACT_SHOT:
-+ /* No valid class found due to action */
-+ return -1;
-+ }
-+#endif
-+ cl = (void *)res.class;
-+ if (!cl) {
-+ /* The filter leads to the qdisc */
-+ if (res.classid == sch->handle)
-+ return 0;
-+
-+ cl = dpaa2_ceetm_find(res.classid, sch);
-+ /* The filter leads to an invalid class */
-+ if (!cl)
-+ break;
-+ }
-+
-+ /* The class might have its own filters attached */
-+ tcf = rcu_dereference_bh(cl->filter_list);
-+ }
-+
-+ /* No valid class found */
-+ if (!cl)
-+ return 0;
-+
-+ switch (cl->type) {
-+ case CEETM_ROOT:
-+ *qdid = cl->root.ch_id;
-+
-+ /* The root class does not have a child prio qdisc */
-+ if (!cl->child)
-+ return 0;
-+
-+ /* Run the prio qdisc classifiers */
-+ return dpaa2_ceetm_classify(skb, cl->child, qdid, qpri);
-+
-+ case CEETM_PRIO:
-+ *qpri = cl->prio.qpri;
-+ break;
-+ }
-+
-+ return 0;
-+}
-+
-+int __init dpaa2_ceetm_register(void)
-+{
-+ int err = 0;
-+
-+ pr_debug(KBUILD_MODNAME ": " DPAA2_CEETM_DESCRIPTION "\n");
-+
-+ err = register_qdisc(&dpaa2_ceetm_qdisc_ops);
-+ if (unlikely(err))
-+ pr_err(KBUILD_MODNAME
-+ ": %s:%hu:%s(): register_qdisc() = %d\n",
-+ KBUILD_BASENAME ".c", __LINE__, __func__, err);
-+
-+ return err;
-+}
-+
-+void __exit dpaa2_ceetm_unregister(void)
-+{
-+ pr_debug(KBUILD_MODNAME ": %s:%s() ->\n",
-+ KBUILD_BASENAME ".c", __func__);
-+
-+ unregister_qdisc(&dpaa2_ceetm_qdisc_ops);
-+}
---- /dev/null
-+++ b/drivers/staging/fsl-dpaa2/ethernet/dpaa2-eth-ceetm.h
-@@ -0,0 +1,182 @@
-+// SPDX-License-Identifier: (GPL-2.0+ OR BSD-3-Clause)
-+/*
-+ * Copyright 2017 NXP
-+ *
-+ */
-+
-+#ifndef __DPAA2_ETH_CEETM_H
-+#define __DPAA2_ETH_CEETM_H
-+
-+#include <net/pkt_sched.h>
-+#include <net/pkt_cls.h>
-+#include <net/netlink.h>
-+
-+#include "dpaa2-eth.h"
-+
-+/* For functional purposes, there are num_tx_queues pfifo qdiscs through which
-+ * frames reach the driver. Their handles start from 1:21. Handles 1:1 to 1:20
-+ * are reserved for the maximum 32 CEETM channels (majors and minors are in
-+ * hex).
-+ */
-+#define PFIFO_MIN_OFFSET 0x21
-+
-+#define DPAA2_CEETM_MIN_WEIGHT 100
-+#define DPAA2_CEETM_MAX_WEIGHT 24800
-+
-+#define DPAA2_CEETM_TD_THRESHOLD 1000
-+
-+enum wbfs_group_type {
-+ WBFS_GRP_A,
-+ WBFS_GRP_B,
-+ WBFS_GRP_LARGE
-+};
-+
-+enum {
-+ DPAA2_CEETM_TCA_UNSPEC,
-+ DPAA2_CEETM_TCA_COPT,
-+ DPAA2_CEETM_TCA_QOPS,
-+ DPAA2_CEETM_TCA_MAX,
-+};
-+
-+/* CEETM configuration types */
-+enum dpaa2_ceetm_type {
-+ CEETM_ROOT = 1,
-+ CEETM_PRIO,
-+};
-+
-+enum {
-+ STRICT_PRIORITY = 0,
-+ WEIGHTED_A,
-+ WEIGHTED_B,
-+};
-+
-+struct dpaa2_ceetm_shaping_cfg {
-+ __u64 cir; /* committed information rate */
-+ __u64 eir; /* excess information rate */
-+ __u16 cbs; /* committed burst size */
-+ __u16 ebs; /* excess burst size */
-+ __u8 coupled; /* shaper coupling */
-+};
-+
-+extern const struct nla_policy ceetm_policy[DPAA2_CEETM_TCA_MAX];
-+
-+struct dpaa2_ceetm_class;
-+struct dpaa2_ceetm_qdisc_stats;
-+struct dpaa2_ceetm_class_stats;
-+
-+/* corresponds to CEETM shaping at LNI level */
-+struct dpaa2_root_q {
-+ struct Qdisc **qdiscs;
-+ struct dpaa2_ceetm_qdisc_stats __percpu *qstats;
-+};
-+
-+/* corresponds to the number of priorities a channel serves */
-+struct dpaa2_prio_q {
-+ struct dpaa2_ceetm_class *parent;
-+ struct dpni_tx_priorities_cfg tx_prio_cfg;
-+};
-+
-+struct dpaa2_ceetm_qdisc {
-+ struct Qdisc_class_hash clhash;
-+ struct tcf_proto *filter_list; /* qdisc attached filters */
-+
-+ enum dpaa2_ceetm_type type; /* ROOT/PRIO */
-+ bool shaped;
-+ union {
-+ struct dpaa2_root_q root;
-+ struct dpaa2_prio_q prio;
-+ };
-+};
-+
-+/* CEETM Qdisc configuration parameters */
-+struct dpaa2_ceetm_tc_qopt {
-+ enum dpaa2_ceetm_type type;
-+ __u16 shaped;
-+ __u8 prio_group_A;
-+ __u8 prio_group_B;
-+ __u8 separate_groups;
-+};
-+
-+/* root class - corresponds to a channel */
-+struct dpaa2_root_c {
-+ struct dpaa2_ceetm_shaping_cfg shaping_cfg;
-+ u32 ch_id;
-+};
-+
-+/* prio class - corresponds to a strict priority queue (group) */
-+struct dpaa2_prio_c {
-+ struct dpaa2_ceetm_class_stats __percpu *cstats;
-+ u32 qpri;
-+ u8 mode;
-+ u16 weight;
-+};
-+
-+struct dpaa2_ceetm_class {
-+ struct Qdisc_class_common common;
-+ int refcnt;
-+ struct tcf_proto *filter_list; /* class attached filters */
-+ struct Qdisc *parent;
-+ struct Qdisc *child;
-+
-+ enum dpaa2_ceetm_type type; /* ROOT/PRIO */
-+ bool shaped;
-+ union {
-+ struct dpaa2_root_c root;
-+ struct dpaa2_prio_c prio;
-+ };
-+};
-+
-+/* CEETM Class configuration parameters */
-+struct dpaa2_ceetm_tc_copt {
-+ enum dpaa2_ceetm_type type;
-+ struct dpaa2_ceetm_shaping_cfg shaping_cfg;
-+ __u16 shaped;
-+ __u8 mode;
-+ __u16 weight;
-+};
-+
-+/* CEETM stats */
-+struct dpaa2_ceetm_qdisc_stats {
-+ __u32 drops;
-+};
-+
-+struct dpaa2_ceetm_class_stats {
-+ /* Software counters */
-+ struct gnet_stats_basic_packed bstats;
-+ __u32 ern_drop_count;
-+ __u32 congested_count;
-+};
-+
-+struct dpaa2_ceetm_tc_xstats {
-+ __u64 ceetm_dequeue_bytes;
-+ __u64 ceetm_dequeue_frames;
-+ __u64 ceetm_reject_bytes;
-+ __u64 ceetm_reject_frames;
-+};
-+
-+#ifdef CONFIG_FSL_DPAA2_ETH_CEETM
-+int __init dpaa2_ceetm_register(void);
-+void __exit dpaa2_ceetm_unregister(void);
-+int dpaa2_ceetm_classify(struct sk_buff *skb, struct Qdisc *sch,
-+ int *qdid, int *qpri);
-+#else
-+static inline int dpaa2_ceetm_register(void)
-+{
-+ return 0;
-+}
-+
-+static inline void dpaa2_ceetm_unregister(void) {}
-+
-+static inline int dpaa2_ceetm_classify(struct sk_buff *skb, struct Qdisc *sch,
-+ int *qdid, int *qpri)
-+{
-+ return 0;
-+}
-+#endif
-+
-+static inline bool dpaa2_eth_ceetm_is_enabled(struct dpaa2_eth_priv *priv)
-+{
-+ return priv->ceetm_en;
-+}
-+
-+#endif
---- /dev/null
-+++ b/drivers/staging/fsl-dpaa2/ethernet/dpaa2-eth-debugfs.c
-@@ -0,0 +1,357 @@
-+
-+/* Copyright 2015 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#include <linux/module.h>
-+#include <linux/debugfs.h>
-+#include "dpaa2-eth.h"
-+#include "dpaa2-eth-debugfs.h"
-+
-+#define DPAA2_ETH_DBG_ROOT "dpaa2-eth"
-+
-+static struct dentry *dpaa2_dbg_root;
-+
-+static int dpaa2_dbg_cpu_show(struct seq_file *file, void *offset)
-+{
-+ struct dpaa2_eth_priv *priv = (struct dpaa2_eth_priv *)file->private;
-+ struct rtnl_link_stats64 *stats;
-+ struct dpaa2_eth_drv_stats *extras;
-+ int i;
-+
-+ seq_printf(file, "Per-CPU stats for %s\n", priv->net_dev->name);
-+ seq_printf(file, "%s%16s%16s%16s%16s%16s%16s%16s%16s%16s\n",
-+ "CPU", "Rx", "Rx Err", "Rx SG", "Tx", "Tx Err", "Tx conf",
-+ "Tx SG", "Tx realloc", "Enq busy");
-+
-+ for_each_online_cpu(i) {
-+ stats = per_cpu_ptr(priv->percpu_stats, i);
-+ extras = per_cpu_ptr(priv->percpu_extras, i);
-+ seq_printf(file, "%3d%16llu%16llu%16llu%16llu%16llu%16llu%16llu%16llu%16llu\n",
-+ i,
-+ stats->rx_packets,
-+ stats->rx_errors,
-+ extras->rx_sg_frames,
-+ stats->tx_packets,
-+ stats->tx_errors,
-+ extras->tx_conf_frames,
-+ extras->tx_sg_frames,
-+ extras->tx_reallocs,
-+ extras->tx_portal_busy);
-+ }
-+
-+ return 0;
-+}
-+
-+static int dpaa2_dbg_cpu_open(struct inode *inode, struct file *file)
-+{
-+ int err;
-+ struct dpaa2_eth_priv *priv = (struct dpaa2_eth_priv *)inode->i_private;
-+
-+ err = single_open(file, dpaa2_dbg_cpu_show, priv);
-+ if (err < 0)
-+ netdev_err(priv->net_dev, "single_open() failed\n");
-+
-+ return err;
-+}
-+
-+static const struct file_operations dpaa2_dbg_cpu_ops = {
-+ .open = dpaa2_dbg_cpu_open,
-+ .read = seq_read,
-+ .llseek = seq_lseek,
-+ .release = single_release,
-+};
-+
-+static char *fq_type_to_str(struct dpaa2_eth_fq *fq)
-+{
-+ switch (fq->type) {
-+ case DPAA2_RX_FQ:
-+ return "Rx";
-+ case DPAA2_TX_CONF_FQ:
-+ return "Tx conf";
-+ case DPAA2_RX_ERR_FQ:
-+ return "Rx err";
-+ default:
-+ return "N/A";
-+ }
-+}
-+
-+static int dpaa2_dbg_fqs_show(struct seq_file *file, void *offset)
-+{
-+ struct dpaa2_eth_priv *priv = (struct dpaa2_eth_priv *)file->private;
-+ struct dpaa2_eth_fq *fq;
-+ u32 fcnt, bcnt;
-+ int i, err;
-+
-+ seq_printf(file, "non-zero FQ stats for %s:\n", priv->net_dev->name);
-+ seq_printf(file, "%s%16s%16s%16s%16s%16s%16s\n",
-+ "VFQID", "CPU", "Traffic Class", "Type", "Frames",
-+ "Pending frames", "Congestion");
-+
-+ for (i = 0; i < priv->num_fqs; i++) {
-+ fq = &priv->fq[i];
-+ err = dpaa2_io_query_fq_count(NULL, fq->fqid, &fcnt, &bcnt);
-+ if (err)
-+ fcnt = 0;
-+
-+ /* A lot of queues, no use displaying zero traffic ones */
-+ if (!fq->stats.frames && !fcnt)
-+ continue;
-+
-+ seq_printf(file, "%5d%16d%16d%16s%16llu%16u%16llu\n",
-+ fq->fqid,
-+ fq->target_cpu,
-+ fq->tc,
-+ fq_type_to_str(fq),
-+ fq->stats.frames,
-+ fcnt,
-+ fq->stats.congestion_entry);
-+ }
-+
-+ return 0;
-+}
-+
-+static int dpaa2_dbg_fqs_open(struct inode *inode, struct file *file)
-+{
-+ int err;
-+ struct dpaa2_eth_priv *priv = (struct dpaa2_eth_priv *)inode->i_private;
-+
-+ err = single_open(file, dpaa2_dbg_fqs_show, priv);
-+ if (err < 0)
-+ netdev_err(priv->net_dev, "single_open() failed\n");
-+
-+ return err;
-+}
-+
-+static const struct file_operations dpaa2_dbg_fq_ops = {
-+ .open = dpaa2_dbg_fqs_open,
-+ .read = seq_read,
-+ .llseek = seq_lseek,
-+ .release = single_release,
-+};
-+
-+static int dpaa2_dbg_ch_show(struct seq_file *file, void *offset)
-+{
-+ struct dpaa2_eth_priv *priv = (struct dpaa2_eth_priv *)file->private;
-+ struct dpaa2_eth_channel *ch;
-+ int i;
-+
-+ seq_printf(file, "Channel stats for %s:\n", priv->net_dev->name);
-+ seq_printf(file, "%s%16s%16s%16s%16s%16s%16s\n",
-+ "CHID", "CPU", "Deq busy", "Frames", "CDANs",
-+ "Avg frm/CDAN", "Buf count");
-+
-+ for (i = 0; i < priv->num_channels; i++) {
-+ ch = priv->channel[i];
-+ seq_printf(file, "%4d%16d%16llu%16llu%16llu%16llu%16d\n",
-+ ch->ch_id,
-+ ch->nctx.desired_cpu,
-+ ch->stats.dequeue_portal_busy,
-+ ch->stats.frames,
-+ ch->stats.cdan,
-+ ch->stats.frames / ch->stats.cdan,
-+ ch->buf_count);
-+ }
-+
-+ return 0;
-+}
-+
-+static int dpaa2_dbg_ch_open(struct inode *inode, struct file *file)
-+{
-+ int err;
-+ struct dpaa2_eth_priv *priv = (struct dpaa2_eth_priv *)inode->i_private;
-+
-+ err = single_open(file, dpaa2_dbg_ch_show, priv);
-+ if (err < 0)
-+ netdev_err(priv->net_dev, "single_open() failed\n");
-+
-+ return err;
-+}
-+
-+static const struct file_operations dpaa2_dbg_ch_ops = {
-+ .open = dpaa2_dbg_ch_open,
-+ .read = seq_read,
-+ .llseek = seq_lseek,
-+ .release = single_release,
-+};
-+
-+static ssize_t dpaa2_dbg_reset_write(struct file *file, const char __user *buf,
-+ size_t count, loff_t *offset)
-+{
-+ struct dpaa2_eth_priv *priv = file->private_data;
-+ struct rtnl_link_stats64 *percpu_stats;
-+ struct dpaa2_eth_drv_stats *percpu_extras;
-+ struct dpaa2_eth_fq *fq;
-+ struct dpaa2_eth_channel *ch;
-+ int i;
-+
-+ for_each_online_cpu(i) {
-+ percpu_stats = per_cpu_ptr(priv->percpu_stats, i);
-+ memset(percpu_stats, 0, sizeof(*percpu_stats));
-+
-+ percpu_extras = per_cpu_ptr(priv->percpu_extras, i);
-+ memset(percpu_extras, 0, sizeof(*percpu_extras));
-+ }
-+
-+ for (i = 0; i < priv->num_fqs; i++) {
-+ fq = &priv->fq[i];
-+ memset(&fq->stats, 0, sizeof(fq->stats));
-+ }
-+
-+ for (i = 0; i < priv->num_channels; i++) {
-+ ch = priv->channel[i];
-+ memset(&ch->stats, 0, sizeof(ch->stats));
-+ }
-+
-+ return count;
-+}
-+
-+static const struct file_operations dpaa2_dbg_reset_ops = {
-+ .open = simple_open,
-+ .write = dpaa2_dbg_reset_write,
-+};
-+
-+static ssize_t dpaa2_dbg_reset_mc_write(struct file *file,
-+ const char __user *buf,
-+ size_t count, loff_t *offset)
-+{
-+ struct dpaa2_eth_priv *priv = file->private_data;
-+ int err;
-+
-+ err = dpni_reset_statistics(priv->mc_io, 0, priv->mc_token);
-+ if (err)
-+ netdev_err(priv->net_dev,
-+ "dpni_reset_statistics() failed %d\n", err);
-+
-+ return count;
-+}
-+
-+static const struct file_operations dpaa2_dbg_reset_mc_ops = {
-+ .open = simple_open,
-+ .write = dpaa2_dbg_reset_mc_write,
-+};
-+
-+void dpaa2_dbg_add(struct dpaa2_eth_priv *priv)
-+{
-+ if (!dpaa2_dbg_root)
-+ return;
-+
-+ /* Create a directory for the interface */
-+ priv->dbg.dir = debugfs_create_dir(priv->net_dev->name,
-+ dpaa2_dbg_root);
-+ if (!priv->dbg.dir) {
-+ netdev_err(priv->net_dev, "debugfs_create_dir() failed\n");
-+ return;
-+ }
-+
-+ /* per-cpu stats file */
-+ priv->dbg.cpu_stats = debugfs_create_file("cpu_stats", 0444,
-+ priv->dbg.dir, priv,
-+ &dpaa2_dbg_cpu_ops);
-+ if (!priv->dbg.cpu_stats) {
-+ netdev_err(priv->net_dev, "debugfs_create_file() failed\n");
-+ goto err_cpu_stats;
-+ }
-+
-+ /* per-fq stats file */
-+ priv->dbg.fq_stats = debugfs_create_file("fq_stats", 0444,
-+ priv->dbg.dir, priv,
-+ &dpaa2_dbg_fq_ops);
-+ if (!priv->dbg.fq_stats) {
-+ netdev_err(priv->net_dev, "debugfs_create_file() failed\n");
-+ goto err_fq_stats;
-+ }
-+
-+ /* per-fq stats file */
-+ priv->dbg.ch_stats = debugfs_create_file("ch_stats", 0444,
-+ priv->dbg.dir, priv,
-+ &dpaa2_dbg_ch_ops);
-+ if (!priv->dbg.fq_stats) {
-+ netdev_err(priv->net_dev, "debugfs_create_file() failed\n");
-+ goto err_ch_stats;
-+ }
-+
-+ /* reset stats */
-+ priv->dbg.reset_stats = debugfs_create_file("reset_stats", 0200,
-+ priv->dbg.dir, priv,
-+ &dpaa2_dbg_reset_ops);
-+ if (!priv->dbg.reset_stats) {
-+ netdev_err(priv->net_dev, "debugfs_create_file() failed\n");
-+ goto err_reset_stats;
-+ }
-+
-+ /* reset MC stats */
-+ priv->dbg.reset_mc_stats = debugfs_create_file("reset_mc_stats",
-+ 0222, priv->dbg.dir, priv,
-+ &dpaa2_dbg_reset_mc_ops);
-+ if (!priv->dbg.reset_mc_stats) {
-+ netdev_err(priv->net_dev, "debugfs_create_file() failed\n");
-+ goto err_reset_mc_stats;
-+ }
-+
-+ return;
-+
-+err_reset_mc_stats:
-+ debugfs_remove(priv->dbg.reset_stats);
-+err_reset_stats:
-+ debugfs_remove(priv->dbg.ch_stats);
-+err_ch_stats:
-+ debugfs_remove(priv->dbg.fq_stats);
-+err_fq_stats:
-+ debugfs_remove(priv->dbg.cpu_stats);
-+err_cpu_stats:
-+ debugfs_remove(priv->dbg.dir);
-+}
-+
-+void dpaa2_dbg_remove(struct dpaa2_eth_priv *priv)
-+{
-+ debugfs_remove(priv->dbg.reset_mc_stats);
-+ debugfs_remove(priv->dbg.reset_stats);
-+ debugfs_remove(priv->dbg.fq_stats);
-+ debugfs_remove(priv->dbg.ch_stats);
-+ debugfs_remove(priv->dbg.cpu_stats);
-+ debugfs_remove(priv->dbg.dir);
-+}
-+
-+void dpaa2_eth_dbg_init(void)
-+{
-+ dpaa2_dbg_root = debugfs_create_dir(DPAA2_ETH_DBG_ROOT, NULL);
-+ if (!dpaa2_dbg_root) {
-+ pr_err("DPAA2-ETH: debugfs create failed\n");
-+ return;
-+ }
-+
-+ pr_info("DPAA2-ETH: debugfs created\n");
-+}
-+
-+void __exit dpaa2_eth_dbg_exit(void)
-+{
-+ debugfs_remove(dpaa2_dbg_root);
-+}
---- /dev/null
-+++ b/drivers/staging/fsl-dpaa2/ethernet/dpaa2-eth-debugfs.h
-@@ -0,0 +1,60 @@
-+/* Copyright 2015 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#ifndef DPAA2_ETH_DEBUGFS_H
-+#define DPAA2_ETH_DEBUGFS_H
-+
-+#include <linux/dcache.h>
-+
-+struct dpaa2_eth_priv;
-+
-+struct dpaa2_debugfs {
-+ struct dentry *dir;
-+ struct dentry *fq_stats;
-+ struct dentry *ch_stats;
-+ struct dentry *cpu_stats;
-+ struct dentry *reset_stats;
-+ struct dentry *reset_mc_stats;
-+};
-+
-+#ifdef CONFIG_FSL_DPAA2_ETH_DEBUGFS
-+void dpaa2_eth_dbg_init(void);
-+void dpaa2_eth_dbg_exit(void);
-+void dpaa2_dbg_add(struct dpaa2_eth_priv *priv);
-+void dpaa2_dbg_remove(struct dpaa2_eth_priv *priv);
-+#else
-+static inline void dpaa2_eth_dbg_init(void) {}
-+static inline void dpaa2_eth_dbg_exit(void) {}
-+static inline void dpaa2_dbg_add(struct dpaa2_eth_priv *priv) {}
-+static inline void dpaa2_dbg_remove(struct dpaa2_eth_priv *priv) {}
-+#endif /* CONFIG_FSL_DPAA2_ETH_DEBUGFS */
-+
-+#endif /* DPAA2_ETH_DEBUGFS_H */
---- /dev/null
-+++ b/drivers/staging/fsl-dpaa2/ethernet/dpaa2-eth-trace.h
-@@ -0,0 +1,185 @@
-+/* Copyright 2014-2015 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#undef TRACE_SYSTEM
-+#define TRACE_SYSTEM dpaa2_eth
-+
-+#if !defined(_DPAA2_ETH_TRACE_H) || defined(TRACE_HEADER_MULTI_READ)
-+#define _DPAA2_ETH_TRACE_H
-+
-+#include <linux/skbuff.h>
-+#include <linux/netdevice.h>
-+#include "dpaa2-eth.h"
-+#include <linux/tracepoint.h>
-+
-+#define TR_FMT "[%s] fd: addr=0x%llx, len=%u, off=%u"
-+/* trace_printk format for raw buffer event class */
-+#define TR_BUF_FMT "[%s] vaddr=%p size=%zu dma_addr=%pad map_size=%zu bpid=%d"
-+
-+/* This is used to declare a class of events.
-+ * individual events of this type will be defined below.
-+ */
-+
-+/* Store details about a frame descriptor */
-+DECLARE_EVENT_CLASS(dpaa2_eth_fd,
-+ /* Trace function prototype */
-+ TP_PROTO(struct net_device *netdev,
-+ const struct dpaa2_fd *fd),
-+
-+ /* Repeat argument list here */
-+ TP_ARGS(netdev, fd),
-+
-+ /* A structure containing the relevant information we want
-+ * to record. Declare name and type for each normal element,
-+ * name, type and size for arrays. Use __string for variable
-+ * length strings.
-+ */
-+ TP_STRUCT__entry(
-+ __field(u64, fd_addr)
-+ __field(u32, fd_len)
-+ __field(u16, fd_offset)
-+ __string(name, netdev->name)
-+ ),
-+
-+ /* The function that assigns values to the above declared
-+ * fields
-+ */
-+ TP_fast_assign(
-+ __entry->fd_addr = dpaa2_fd_get_addr(fd);
-+ __entry->fd_len = dpaa2_fd_get_len(fd);
-+ __entry->fd_offset = dpaa2_fd_get_offset(fd);
-+ __assign_str(name, netdev->name);
-+ ),
-+
-+ /* This is what gets printed when the trace event is
-+ * triggered.
-+ */
-+ TP_printk(TR_FMT,
-+ __get_str(name),
-+ __entry->fd_addr,
-+ __entry->fd_len,
-+ __entry->fd_offset)
-+);
-+
-+/* Now declare events of the above type. Format is:
-+ * DEFINE_EVENT(class, name, proto, args), with proto and args same as for class
-+ */
-+
-+/* Tx (egress) fd */
-+DEFINE_EVENT(dpaa2_eth_fd, dpaa2_tx_fd,
-+ TP_PROTO(struct net_device *netdev,
-+ const struct dpaa2_fd *fd),
-+
-+ TP_ARGS(netdev, fd)
-+);
-+
-+/* Rx fd */
-+DEFINE_EVENT(dpaa2_eth_fd, dpaa2_rx_fd,
-+ TP_PROTO(struct net_device *netdev,
-+ const struct dpaa2_fd *fd),
-+
-+ TP_ARGS(netdev, fd)
-+);
-+
-+/* Tx confirmation fd */
-+DEFINE_EVENT(dpaa2_eth_fd, dpaa2_tx_conf_fd,
-+ TP_PROTO(struct net_device *netdev,
-+ const struct dpaa2_fd *fd),
-+
-+ TP_ARGS(netdev, fd)
-+);
-+
-+/* Log data about raw buffers. Useful for tracing DPBP content. */
-+TRACE_EVENT(dpaa2_eth_buf_seed,
-+ /* Trace function prototype */
-+ TP_PROTO(struct net_device *netdev,
-+ /* virtual address and size */
-+ void *vaddr,
-+ size_t size,
-+ /* dma map address and size */
-+ dma_addr_t dma_addr,
-+ size_t map_size,
-+ /* buffer pool id, if relevant */
-+ u16 bpid),
-+
-+ /* Repeat argument list here */
-+ TP_ARGS(netdev, vaddr, size, dma_addr, map_size, bpid),
-+
-+ /* A structure containing the relevant information we want
-+ * to record. Declare name and type for each normal element,
-+ * name, type and size for arrays. Use __string for variable
-+ * length strings.
-+ */
-+ TP_STRUCT__entry(
-+ __field(void *, vaddr)
-+ __field(size_t, size)
-+ __field(dma_addr_t, dma_addr)
-+ __field(size_t, map_size)
-+ __field(u16, bpid)
-+ __string(name, netdev->name)
-+ ),
-+
-+ /* The function that assigns values to the above declared
-+ * fields
-+ */
-+ TP_fast_assign(
-+ __entry->vaddr = vaddr;
-+ __entry->size = size;
-+ __entry->dma_addr = dma_addr;
-+ __entry->map_size = map_size;
-+ __entry->bpid = bpid;
-+ __assign_str(name, netdev->name);
-+ ),
-+
-+ /* This is what gets printed when the trace event is
-+ * triggered.
-+ */
-+ TP_printk(TR_BUF_FMT,
-+ __get_str(name),
-+ __entry->vaddr,
-+ __entry->size,
-+ &__entry->dma_addr,
-+ __entry->map_size,
-+ __entry->bpid)
-+);
-+
-+/* If only one event of a certain type needs to be declared, use TRACE_EVENT().
-+ * The syntax is the same as for DECLARE_EVENT_CLASS().
-+ */
-+
-+#endif /* _DPAA2_ETH_TRACE_H */
-+
-+/* This must be outside ifdef _DPAA2_ETH_TRACE_H */
-+#undef TRACE_INCLUDE_PATH
-+#define TRACE_INCLUDE_PATH .
-+#undef TRACE_INCLUDE_FILE
-+#define TRACE_INCLUDE_FILE dpaa2-eth-trace
-+#include <trace/define_trace.h>
---- /dev/null
-+++ b/drivers/staging/fsl-dpaa2/ethernet/dpaa2-eth.c
-@@ -0,0 +1,3734 @@
-+/* Copyright 2014-2016 Freescale Semiconductor Inc.
-+ * Copyright 2016-2017 NXP
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+#include <linux/init.h>
-+#include <linux/module.h>
-+#include <linux/platform_device.h>
-+#include <linux/etherdevice.h>
-+#include <linux/of_net.h>
-+#include <linux/interrupt.h>
-+#include <linux/msi.h>
-+#include <linux/kthread.h>
-+#include <linux/iommu.h>
-+#include <linux/net_tstamp.h>
-+#include <linux/bpf.h>
-+#include <linux/filter.h>
-+#include <linux/atomic.h>
-+#include <net/sock.h>
-+#include <linux/fsl/mc.h>
-+#include "dpaa2-eth.h"
-+#include "dpaa2-eth-ceetm.h"
-+
-+/* CREATE_TRACE_POINTS only needs to be defined once. Other dpa files
-+ * using trace events only need to #include <trace/events/sched.h>
-+ */
-+#define CREATE_TRACE_POINTS
-+#include "dpaa2-eth-trace.h"
-+
-+MODULE_LICENSE("Dual BSD/GPL");
-+MODULE_AUTHOR("Freescale Semiconductor, Inc");
-+MODULE_DESCRIPTION("Freescale DPAA2 Ethernet Driver");
-+
-+const char dpaa2_eth_drv_version[] = "0.1";
-+
-+static void *dpaa2_iova_to_virt(struct iommu_domain *domain,
-+ dma_addr_t iova_addr)
-+{
-+ phys_addr_t phys_addr;
-+
-+ phys_addr = domain ? iommu_iova_to_phys(domain, iova_addr) : iova_addr;
-+
-+ return phys_to_virt(phys_addr);
-+}
-+
-+static void validate_rx_csum(struct dpaa2_eth_priv *priv,
-+ u32 fd_status,
-+ struct sk_buff *skb)
-+{
-+ skb_checksum_none_assert(skb);
-+
-+ /* HW checksum validation is disabled, nothing to do here */
-+ if (!(priv->net_dev->features & NETIF_F_RXCSUM))
-+ return;
-+
-+ /* Read checksum validation bits */
-+ if (!((fd_status & DPAA2_FAS_L3CV) &&
-+ (fd_status & DPAA2_FAS_L4CV)))
-+ return;
-+
-+ /* Inform the stack there's no need to compute L3/L4 csum anymore */
-+ skb->ip_summed = CHECKSUM_UNNECESSARY;
-+}
-+
-+/* Free a received FD.
-+ * Not to be used for Tx conf FDs or on any other paths.
-+ */
-+static void free_rx_fd(struct dpaa2_eth_priv *priv,
-+ const struct dpaa2_fd *fd,
-+ void *vaddr)
-+{
-+ struct device *dev = priv->net_dev->dev.parent;
-+ dma_addr_t addr = dpaa2_fd_get_addr(fd);
-+ u8 fd_format = dpaa2_fd_get_format(fd);
-+ struct dpaa2_sg_entry *sgt;
-+ void *sg_vaddr;
-+ int i;
-+
-+ /* If single buffer frame, just free the data buffer */
-+ if (fd_format == dpaa2_fd_single)
-+ goto free_buf;
-+ else if (fd_format != dpaa2_fd_sg)
-+ /* We don't support any other format */
-+ return;
-+
-+ /* For S/G frames, we first need to free all SG entries
-+ * except the first one, which was taken care of already
-+ */
-+ sgt = vaddr + dpaa2_fd_get_offset(fd);
-+ for (i = 1; i < DPAA2_ETH_MAX_SG_ENTRIES; i++) {
-+ addr = dpaa2_sg_get_addr(&sgt[i]);
-+ sg_vaddr = dpaa2_iova_to_virt(priv->iommu_domain, addr);
-+ dma_unmap_single(dev, addr, DPAA2_ETH_RX_BUF_SIZE,
-+ DMA_BIDIRECTIONAL);
-+
-+ skb_free_frag(sg_vaddr);
-+ if (dpaa2_sg_is_final(&sgt[i]))
-+ break;
-+ }
-+
-+free_buf:
-+ skb_free_frag(vaddr);
-+}
-+
-+/* Build a linear skb based on a single-buffer frame descriptor */
-+static struct sk_buff *build_linear_skb(struct dpaa2_eth_priv *priv,
-+ struct dpaa2_eth_channel *ch,
-+ const struct dpaa2_fd *fd,
-+ void *fd_vaddr)
-+{
-+ struct sk_buff *skb = NULL;
-+ u16 fd_offset = dpaa2_fd_get_offset(fd);
-+ u32 fd_length = dpaa2_fd_get_len(fd);
-+
-+ ch->buf_count--;
-+
-+ skb = build_skb(fd_vaddr, DPAA2_ETH_SKB_SIZE);
-+ if (unlikely(!skb))
-+ return NULL;
-+
-+ skb_reserve(skb, fd_offset);
-+ skb_put(skb, fd_length);
-+
-+ return skb;
-+}
-+
-+/* Build a non linear (fragmented) skb based on a S/G table */
-+static struct sk_buff *build_frag_skb(struct dpaa2_eth_priv *priv,
-+ struct dpaa2_eth_channel *ch,
-+ struct dpaa2_sg_entry *sgt)
-+{
-+ struct sk_buff *skb = NULL;
-+ struct device *dev = priv->net_dev->dev.parent;
-+ void *sg_vaddr;
-+ dma_addr_t sg_addr;
-+ u16 sg_offset;
-+ u32 sg_length;
-+ struct page *page, *head_page;
-+ int page_offset;
-+ int i;
-+
-+ for (i = 0; i < DPAA2_ETH_MAX_SG_ENTRIES; i++) {
-+ struct dpaa2_sg_entry *sge = &sgt[i];
-+
-+ /* NOTE: We only support SG entries in dpaa2_sg_single format,
-+ * but this is the only format we may receive from HW anyway
-+ */
-+
-+ /* Get the address and length from the S/G entry */
-+ sg_addr = dpaa2_sg_get_addr(sge);
-+ sg_vaddr = dpaa2_iova_to_virt(priv->iommu_domain, sg_addr);
-+ dma_unmap_single(dev, sg_addr, DPAA2_ETH_RX_BUF_SIZE,
-+ DMA_BIDIRECTIONAL);
-+
-+ sg_length = dpaa2_sg_get_len(sge);
-+
-+ if (i == 0) {
-+ /* We build the skb around the first data buffer */
-+ skb = build_skb(sg_vaddr, DPAA2_ETH_SKB_SIZE);
-+ if (unlikely(!skb)) {
-+ /* Free the first SG entry now, since we already
-+ * unmapped it and obtained the virtual address
-+ */
-+ skb_free_frag(sg_vaddr);
-+
-+ /* We still need to subtract the buffers used
-+ * by this FD from our software counter
-+ */
-+ while (!dpaa2_sg_is_final(&sgt[i]) &&
-+ i < DPAA2_ETH_MAX_SG_ENTRIES)
-+ i++;
-+ break;
-+ }
-+
-+ sg_offset = dpaa2_sg_get_offset(sge);
-+ skb_reserve(skb, sg_offset);
-+ skb_put(skb, sg_length);
-+ } else {
-+ /* Rest of the data buffers are stored as skb frags */
-+ page = virt_to_page(sg_vaddr);
-+ head_page = virt_to_head_page(sg_vaddr);
-+
-+ /* Offset in page (which may be compound).
-+ * Data in subsequent SG entries is stored from the
-+ * beginning of the buffer, so we don't need to add the
-+ * sg_offset.
-+ */
-+ page_offset = ((unsigned long)sg_vaddr &
-+ (PAGE_SIZE - 1)) +
-+ (page_address(page) - page_address(head_page));
-+
-+ skb_add_rx_frag(skb, i - 1, head_page, page_offset,
-+ sg_length, DPAA2_ETH_RX_BUF_SIZE);
-+ }
-+
-+ if (dpaa2_sg_is_final(sge))
-+ break;
-+ }
-+
-+ WARN_ONCE(i == DPAA2_ETH_MAX_SG_ENTRIES, "Final bit not set in SGT");
-+
-+ /* Count all data buffers + SG table buffer */
-+ ch->buf_count -= i + 2;
-+
-+ return skb;
-+}
-+
-+static int dpaa2_eth_xdp_tx(struct dpaa2_eth_priv *priv,
-+ struct dpaa2_fd *fd,
-+ void *buf_start,
-+ u16 queue_id)
-+{
-+ struct dpaa2_eth_fq *fq;
-+ struct rtnl_link_stats64 *percpu_stats;
-+ struct dpaa2_eth_drv_stats *percpu_extras;
-+ struct dpaa2_faead *faead;
-+ u32 ctrl, frc;
-+ int i, err;
-+
-+ /* Mark the egress frame annotation area as valid */
-+ frc = dpaa2_fd_get_frc(fd);
-+ dpaa2_fd_set_frc(fd, frc | DPAA2_FD_FRC_FAEADV);
-+ dpaa2_fd_set_ctrl(fd, DPAA2_FD_CTRL_ASAL);
-+
-+ ctrl = DPAA2_FAEAD_A4V | DPAA2_FAEAD_A2V | DPAA2_FAEAD_EBDDV;
-+ faead = dpaa2_get_faead(buf_start, false);
-+ faead->ctrl = cpu_to_le32(ctrl);
-+ faead->conf_fqid = 0;
-+
-+ percpu_stats = this_cpu_ptr(priv->percpu_stats);
-+ percpu_extras = this_cpu_ptr(priv->percpu_extras);
-+
-+ fq = &priv->fq[queue_id];
-+ for (i = 0; i < DPAA2_ETH_ENQUEUE_RETRIES; i++) {
-+ err = dpaa2_io_service_enqueue_qd(fq->channel->dpio,
-+ priv->tx_qdid, 0,
-+ fq->tx_qdbin, fd);
-+ if (err != -EBUSY)
-+ break;
-+ }
-+
-+ percpu_extras->tx_portal_busy += i;
-+ if (unlikely(err)) {
-+ percpu_stats->tx_errors++;
-+ } else {
-+ percpu_stats->tx_packets++;
-+ percpu_stats->tx_bytes += dpaa2_fd_get_len(fd);
-+ }
-+
-+ return err;
-+}
-+
-+static void free_bufs(struct dpaa2_eth_priv *priv, u64 *buf_array, int count)
-+{
-+ struct device *dev = priv->net_dev->dev.parent;
-+ void *vaddr;
-+ int i;
-+
-+ for (i = 0; i < count; i++) {
-+ /* Same logic as on regular Rx path */
-+ vaddr = dpaa2_iova_to_virt(priv->iommu_domain, buf_array[i]);
-+ dma_unmap_single(dev, buf_array[i], DPAA2_ETH_RX_BUF_SIZE,
-+ DMA_BIDIRECTIONAL);
-+ skb_free_frag(vaddr);
-+ }
-+}
-+
-+static void release_fd_buf(struct dpaa2_eth_priv *priv,
-+ struct dpaa2_eth_channel *ch,
-+ dma_addr_t addr)
-+{
-+ int err;
-+
-+ ch->rel_buf_array[ch->rel_buf_cnt++] = addr;
-+ if (likely(ch->rel_buf_cnt < DPAA2_ETH_BUFS_PER_CMD))
-+ return;
-+
-+ while ((err = dpaa2_io_service_release(ch->dpio, priv->bpid,
-+ ch->rel_buf_array,
-+ ch->rel_buf_cnt)) == -EBUSY)
-+ cpu_relax();
-+
-+ if (err)
-+ free_bufs(priv, ch->rel_buf_array, ch->rel_buf_cnt);
-+
-+ ch->rel_buf_cnt = 0;
-+}
-+
-+/* Main Rx frame processing routine */
-+static void dpaa2_eth_rx(struct dpaa2_eth_priv *priv,
-+ struct dpaa2_eth_channel *ch,
-+ const struct dpaa2_fd *fd,
-+ struct napi_struct *napi,
-+ u16 queue_id)
-+{
-+ dma_addr_t addr = dpaa2_fd_get_addr(fd);
-+ u8 fd_format = dpaa2_fd_get_format(fd);
-+ void *vaddr;
-+ struct sk_buff *skb;
-+ struct rtnl_link_stats64 *percpu_stats;
-+ struct dpaa2_eth_drv_stats *percpu_extras;
-+ struct device *dev = priv->net_dev->dev.parent;
-+ struct dpaa2_fas *fas;
-+ void *buf_data;
-+ u32 status = 0;
-+ struct bpf_prog *xdp_prog;
-+ struct xdp_buff xdp;
-+ u32 xdp_act;
-+
-+ /* Tracing point */
-+ trace_dpaa2_rx_fd(priv->net_dev, fd);
-+
-+ vaddr = dpaa2_iova_to_virt(priv->iommu_domain, addr);
-+ dma_sync_single_for_cpu(dev, addr, DPAA2_ETH_RX_BUF_SIZE,
-+ DMA_BIDIRECTIONAL);
-+
-+ fas = dpaa2_get_fas(vaddr, false);
-+ prefetch(fas);
-+ buf_data = vaddr + dpaa2_fd_get_offset(fd);
-+ prefetch(buf_data);
-+
-+ percpu_stats = this_cpu_ptr(priv->percpu_stats);
-+ percpu_extras = this_cpu_ptr(priv->percpu_extras);
-+
-+ xdp_prog = READ_ONCE(ch->xdp_prog);
-+
-+ if (fd_format == dpaa2_fd_single) {
-+ if (xdp_prog) {
-+ xdp.data = buf_data;
-+ xdp.data_end = buf_data + dpaa2_fd_get_len(fd);
-+ /* for now, we don't support changes in header size */
-+ xdp.data_hard_start = buf_data;
-+
-+ /* update stats here, as we won't reach the code
-+ * that does that for standard frames
-+ */
-+ percpu_stats->rx_packets++;
-+ percpu_stats->rx_bytes += dpaa2_fd_get_len(fd);
-+
-+ xdp_act = bpf_prog_run_xdp(xdp_prog, &xdp);
-+ switch (xdp_act) {
-+ case XDP_PASS:
-+ break;
-+ default:
-+ bpf_warn_invalid_xdp_action(xdp_act);
-+ case XDP_ABORTED:
-+ case XDP_DROP:
-+ release_fd_buf(priv, ch, addr);
-+ goto drop_cnt;
-+ case XDP_TX:
-+ if (dpaa2_eth_xdp_tx(priv, (struct dpaa2_fd *)fd, vaddr,
-+ queue_id)) {
-+ dma_unmap_single(dev, addr,
-+ DPAA2_ETH_RX_BUF_SIZE,
-+ DMA_BIDIRECTIONAL);
-+ free_rx_fd(priv, fd, vaddr);
-+ ch->buf_count--;
-+ }
-+ return;
-+ }
-+ }
-+ dma_unmap_single(dev, addr, DPAA2_ETH_RX_BUF_SIZE,
-+ DMA_BIDIRECTIONAL);
-+ skb = build_linear_skb(priv, ch, fd, vaddr);
-+ } else if (fd_format == dpaa2_fd_sg) {
-+ dma_unmap_single(dev, addr, DPAA2_ETH_RX_BUF_SIZE,
-+ DMA_BIDIRECTIONAL);
-+ skb = build_frag_skb(priv, ch, buf_data);
-+ skb_free_frag(vaddr);
-+ percpu_extras->rx_sg_frames++;
-+ percpu_extras->rx_sg_bytes += dpaa2_fd_get_len(fd);
-+ } else {
-+ /* We don't support any other format */
-+ goto drop_cnt;
-+ }
-+
-+ if (unlikely(!skb))
-+ goto drop_fd;
-+
-+ prefetch(skb->data);
-+
-+ /* Get the timestamp value */
-+ if (priv->ts_rx_en) {
-+ struct skb_shared_hwtstamps *shhwtstamps = skb_hwtstamps(skb);
-+ u64 *ns = dpaa2_get_ts(vaddr, false);
-+
-+ *ns = DPAA2_PTP_NOMINAL_FREQ_PERIOD_NS * le64_to_cpup(ns);
-+ memset(shhwtstamps, 0, sizeof(*shhwtstamps));
-+ shhwtstamps->hwtstamp = ns_to_ktime(*ns);
-+ }
-+
-+ /* Check if we need to validate the L4 csum */
-+ if (likely(dpaa2_fd_get_frc(fd) & DPAA2_FD_FRC_FASV)) {
-+ status = le32_to_cpu(fas->status);
-+ validate_rx_csum(priv, status, skb);
-+ }
-+
-+ skb->protocol = eth_type_trans(skb, priv->net_dev);
-+
-+ /* Record Rx queue - this will be used when picking a Tx queue to
-+ * forward the frames. We're keeping flow affinity through the
-+ * network stack.
-+ */
-+ skb_record_rx_queue(skb, queue_id);
-+
-+ percpu_stats->rx_packets++;
-+ percpu_stats->rx_bytes += dpaa2_fd_get_len(fd);
-+
-+ napi_gro_receive(napi, skb);
-+
-+ return;
-+
-+drop_fd:
-+ free_rx_fd(priv, fd, vaddr);
-+drop_cnt:
-+ percpu_stats->rx_dropped++;
-+}
-+
-+#ifdef CONFIG_FSL_DPAA2_ETH_USE_ERR_QUEUE
-+/* Processing of Rx frames received on the error FQ
-+ * We check and print the error bits and then free the frame
-+ */
-+static void dpaa2_eth_rx_err(struct dpaa2_eth_priv *priv,
-+ struct dpaa2_eth_channel *ch,
-+ const struct dpaa2_fd *fd,
-+ struct napi_struct *napi __always_unused,
-+ u16 queue_id __always_unused)
-+{
-+ struct device *dev = priv->net_dev->dev.parent;
-+ dma_addr_t addr = dpaa2_fd_get_addr(fd);
-+ void *vaddr;
-+ struct rtnl_link_stats64 *percpu_stats;
-+ struct dpaa2_fas *fas;
-+ u32 status = 0;
-+ u32 fd_errors;
-+ bool has_fas_errors = false;
-+
-+ vaddr = dpaa2_iova_to_virt(priv->iommu_domain, addr);
-+ dma_unmap_single(dev, addr, DPAA2_ETH_RX_BUF_SIZE, DMA_BIDIRECTIONAL);
-+
-+ /* check frame errors in the FD field */
-+ fd_errors = dpaa2_fd_get_ctrl(fd) & DPAA2_FD_RX_ERR_MASK;
-+ if (likely(fd_errors)) {
-+ has_fas_errors = (fd_errors & FD_CTRL_FAERR) &&
-+ !!(dpaa2_fd_get_frc(fd) & DPAA2_FD_FRC_FASV);
-+ if (net_ratelimit())
-+ netdev_dbg(priv->net_dev, "RX frame FD err: %08x\n",
-+ fd_errors);
-+ }
-+
-+ /* check frame errors in the FAS field */
-+ if (has_fas_errors) {
-+ fas = dpaa2_get_fas(vaddr, false);
-+ status = le32_to_cpu(fas->status);
-+ if (net_ratelimit())
-+ netdev_dbg(priv->net_dev, "Rx frame FAS err: 0x%08x\n",
-+ status & DPAA2_FAS_RX_ERR_MASK);
-+ }
-+ free_rx_fd(priv, fd, vaddr);
-+
-+ percpu_stats = this_cpu_ptr(priv->percpu_stats);
-+ percpu_stats->rx_errors++;
-+ ch->buf_count--;
-+}
-+#endif
-+
-+/* Consume all frames pull-dequeued into the store. This is the simplest way to
-+ * make sure we don't accidentally issue another volatile dequeue which would
-+ * overwrite (leak) frames already in the store.
-+ *
-+ * The number of frames is returned using the last 2 output arguments,
-+ * separately for Rx and Tx confirmations.
-+ *
-+ * Observance of NAPI budget is not our concern, leaving that to the caller.
-+ */
-+static bool consume_frames(struct dpaa2_eth_channel *ch, int *rx_cleaned,
-+ int *tx_conf_cleaned)
-+{
-+ struct dpaa2_eth_priv *priv = ch->priv;
-+ struct dpaa2_eth_fq *fq = NULL;
-+ struct dpaa2_dq *dq;
-+ const struct dpaa2_fd *fd;
-+ int cleaned = 0;
-+ int is_last;
-+
-+ do {
-+ dq = dpaa2_io_store_next(ch->store, &is_last);
-+ if (unlikely(!dq)) {
-+ /* If we're here, we *must* have placed a
-+ * volatile dequeue comnmand, so keep reading through
-+ * the store until we get some sort of valid response
-+ * token (either a valid frame or an "empty dequeue")
-+ */
-+ continue;
-+ }
-+
-+ fd = dpaa2_dq_fd(dq);
-+ prefetch(fd);
-+
-+ fq = (struct dpaa2_eth_fq *)(uintptr_t)dpaa2_dq_fqd_ctx(dq);
-+ fq->consume(priv, ch, fd, &ch->napi, fq->flowid);
-+ cleaned++;
-+ } while (!is_last);
-+
-+ if (!cleaned)
-+ return false;
-+
-+ /* All frames brought in store by a volatile dequeue
-+ * come from the same queue
-+ */
-+ if (fq->type == DPAA2_TX_CONF_FQ)
-+ *tx_conf_cleaned += cleaned;
-+ else
-+ *rx_cleaned += cleaned;
-+
-+ fq->stats.frames += cleaned;
-+ ch->stats.frames += cleaned;
-+
-+ return true;
-+}
-+
-+/* Configure the egress frame annotation for timestamp update */
-+static void enable_tx_tstamp(struct dpaa2_fd *fd, void *buf_start)
-+{
-+ struct dpaa2_faead *faead;
-+ u32 ctrl, frc;
-+
-+ /* Mark the egress frame annotation area as valid */
-+ frc = dpaa2_fd_get_frc(fd);
-+ dpaa2_fd_set_frc(fd, frc | DPAA2_FD_FRC_FAEADV);
-+
-+ /* Set hardware annotation size */
-+ ctrl = dpaa2_fd_get_ctrl(fd);
-+ dpaa2_fd_set_ctrl(fd, ctrl | DPAA2_FD_CTRL_ASAL);
-+
-+ /* enable UPD (update prepanded data) bit in FAEAD field of
-+ * hardware frame annotation area
-+ */
-+ ctrl = DPAA2_FAEAD_A2V | DPAA2_FAEAD_UPDV | DPAA2_FAEAD_UPD;
-+ faead = dpaa2_get_faead(buf_start, true);
-+ faead->ctrl = cpu_to_le32(ctrl);
-+}
-+
-+/* Create a frame descriptor based on a fragmented skb */
-+static int build_sg_fd(struct dpaa2_eth_priv *priv,
-+ struct sk_buff *skb,
-+ struct dpaa2_fd *fd)
-+{
-+ struct device *dev = priv->net_dev->dev.parent;
-+ void *sgt_buf = NULL;
-+ dma_addr_t addr;
-+ int nr_frags = skb_shinfo(skb)->nr_frags;
-+ struct dpaa2_sg_entry *sgt;
-+ int i, err;
-+ int sgt_buf_size;
-+ struct scatterlist *scl, *crt_scl;
-+ int num_sg;
-+ int num_dma_bufs;
-+ struct dpaa2_eth_swa *swa;
-+
-+ /* Create and map scatterlist.
-+ * We don't advertise NETIF_F_FRAGLIST, so skb_to_sgvec() will not have
-+ * to go beyond nr_frags+1.
-+ * Note: We don't support chained scatterlists
-+ */
-+ if (unlikely(PAGE_SIZE / sizeof(struct scatterlist) < nr_frags + 1))
-+ return -EINVAL;
-+
-+ scl = kcalloc(nr_frags + 1, sizeof(struct scatterlist), GFP_ATOMIC);
-+ if (unlikely(!scl))
-+ return -ENOMEM;
-+
-+ sg_init_table(scl, nr_frags + 1);
-+ num_sg = skb_to_sgvec(skb, scl, 0, skb->len);
-+ num_dma_bufs = dma_map_sg(dev, scl, num_sg, DMA_BIDIRECTIONAL);
-+ if (unlikely(!num_dma_bufs)) {
-+ err = -ENOMEM;
-+ goto dma_map_sg_failed;
-+ }
-+
-+ /* Prepare the HW SGT structure */
-+ sgt_buf_size = priv->tx_data_offset +
-+ sizeof(struct dpaa2_sg_entry) * num_dma_bufs;
-+ sgt_buf = netdev_alloc_frag(sgt_buf_size + DPAA2_ETH_TX_BUF_ALIGN);
-+ if (unlikely(!sgt_buf)) {
-+ err = -ENOMEM;
-+ goto sgt_buf_alloc_failed;
-+ }
-+ sgt_buf = PTR_ALIGN(sgt_buf, DPAA2_ETH_TX_BUF_ALIGN);
-+ memset(sgt_buf, 0, sgt_buf_size);
-+
-+ sgt = (struct dpaa2_sg_entry *)(sgt_buf + priv->tx_data_offset);
-+
-+ /* Fill in the HW SGT structure.
-+ *
-+ * sgt_buf is zeroed out, so the following fields are implicit
-+ * in all sgt entries:
-+ * - offset is 0
-+ * - format is 'dpaa2_sg_single'
-+ */
-+ for_each_sg(scl, crt_scl, num_dma_bufs, i) {
-+ dpaa2_sg_set_addr(&sgt[i], sg_dma_address(crt_scl));
-+ dpaa2_sg_set_len(&sgt[i], sg_dma_len(crt_scl));
-+ }
-+ dpaa2_sg_set_final(&sgt[i - 1], true);
-+
-+ /* Store the skb backpointer in the SGT buffer.
-+ * Fit the scatterlist and the number of buffers alongside the
-+ * skb backpointer in the software annotation area. We'll need
-+ * all of them on Tx Conf.
-+ */
-+ swa = (struct dpaa2_eth_swa *)sgt_buf;
-+ swa->type = DPAA2_ETH_SWA_SG;
-+ swa->sg.skb = skb;
-+ swa->sg.scl = scl;
-+ swa->sg.num_sg = num_sg;
-+ swa->sg.sgt_size = sgt_buf_size;
-+
-+ /* Separately map the SGT buffer */
-+ addr = dma_map_single(dev, sgt_buf, sgt_buf_size, DMA_BIDIRECTIONAL);
-+ if (unlikely(dma_mapping_error(dev, addr))) {
-+ err = -ENOMEM;
-+ goto dma_map_single_failed;
-+ }
-+ dpaa2_fd_set_offset(fd, priv->tx_data_offset);
-+ dpaa2_fd_set_format(fd, dpaa2_fd_sg);
-+ dpaa2_fd_set_addr(fd, addr);
-+ dpaa2_fd_set_len(fd, skb->len);
-+ dpaa2_fd_set_ctrl(fd, FD_CTRL_PTA);
-+
-+ if (priv->ts_tx_en && skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP)
-+ enable_tx_tstamp(fd, sgt_buf);
-+
-+ return 0;
-+
-+dma_map_single_failed:
-+ skb_free_frag(sgt_buf);
-+sgt_buf_alloc_failed:
-+ dma_unmap_sg(dev, scl, num_sg, DMA_BIDIRECTIONAL);
-+dma_map_sg_failed:
-+ kfree(scl);
-+ return err;
-+}
-+
-+/* Create a frame descriptor based on a linear skb */
-+static int build_single_fd(struct dpaa2_eth_priv *priv,
-+ struct sk_buff *skb,
-+ struct dpaa2_fd *fd)
-+{
-+ struct device *dev = priv->net_dev->dev.parent;
-+ u8 *buffer_start, *aligned_start;
-+ struct dpaa2_eth_swa *swa;
-+ dma_addr_t addr;
-+
-+ buffer_start = skb->data - dpaa2_eth_needed_headroom(priv, skb);
-+
-+ /* If there's enough room to align the FD address, do it.
-+ * It will help hardware optimize accesses.
-+ */
-+ aligned_start = PTR_ALIGN(buffer_start - DPAA2_ETH_TX_BUF_ALIGN,
-+ DPAA2_ETH_TX_BUF_ALIGN);
-+ if (aligned_start >= skb->head)
-+ buffer_start = aligned_start;
-+
-+ /* Store a backpointer to the skb at the beginning of the buffer
-+ * (in the private data area) such that we can release it
-+ * on Tx confirm
-+ */
-+ swa = (struct dpaa2_eth_swa *)buffer_start;
-+ swa->type = DPAA2_ETH_SWA_SINGLE;
-+ swa->single.skb = skb;
-+
-+ addr = dma_map_single(dev, buffer_start,
-+ skb_tail_pointer(skb) - buffer_start,
-+ DMA_BIDIRECTIONAL);
-+ if (unlikely(dma_mapping_error(dev, addr)))
-+ return -ENOMEM;
-+
-+ dpaa2_fd_set_addr(fd, addr);
-+ dpaa2_fd_set_offset(fd, (u16)(skb->data - buffer_start));
-+ dpaa2_fd_set_len(fd, skb->len);
-+ dpaa2_fd_set_format(fd, dpaa2_fd_single);
-+ dpaa2_fd_set_ctrl(fd, FD_CTRL_PTA);
-+
-+ if (priv->ts_tx_en && skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP)
-+ enable_tx_tstamp(fd, buffer_start);
-+
-+ return 0;
-+}
-+
-+/* FD freeing routine on the Tx path
-+ *
-+ * DMA-unmap and free FD and possibly SGT buffer allocated on Tx. The skb
-+ * back-pointed to is also freed.
-+ * This can be called either from dpaa2_eth_tx_conf() or on the error path of
-+ * dpaa2_eth_tx().
-+ * Optionally, return the frame annotation status word (FAS), which needs
-+ * to be checked if we're on the confirmation path.
-+ */
-+static void free_tx_fd(struct dpaa2_eth_priv *priv,
-+ const struct dpaa2_fd *fd,
-+ bool in_napi)
-+{
-+ struct device *dev = priv->net_dev->dev.parent;
-+ dma_addr_t fd_addr;
-+ struct sk_buff *skb = NULL;
-+ unsigned char *buffer_start;
-+ struct dpaa2_eth_swa *swa;
-+ u8 fd_format = dpaa2_fd_get_format(fd);
-+
-+ fd_addr = dpaa2_fd_get_addr(fd);
-+ buffer_start = dpaa2_iova_to_virt(priv->iommu_domain, fd_addr);
-+ swa = (struct dpaa2_eth_swa *)buffer_start;
-+
-+ if (fd_format == dpaa2_fd_single) {
-+ skb = swa->single.skb;
-+ /* Accessing the skb buffer is safe before dma unmap, because
-+ * we didn't map the actual skb shell.
-+ */
-+ dma_unmap_single(dev, fd_addr,
-+ skb_tail_pointer(skb) - buffer_start,
-+ DMA_BIDIRECTIONAL);
-+ } else if (fd_format == dpaa2_fd_sg) {
-+ skb = swa->sg.skb;
-+
-+ /* Unmap the scatterlist */
-+ dma_unmap_sg(dev, swa->sg.scl, swa->sg.num_sg, DMA_BIDIRECTIONAL);
-+ kfree(swa->sg.scl);
-+
-+ /* Unmap the SGT buffer */
-+ dma_unmap_single(dev, fd_addr, swa->sg.sgt_size,
-+ DMA_BIDIRECTIONAL);
-+ } else {
-+ netdev_dbg(priv->net_dev, "Invalid FD format\n");
-+ return;
-+ }
-+
-+ /* Get the timestamp value */
-+ if (priv->ts_tx_en && skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP) {
-+ struct skb_shared_hwtstamps shhwtstamps;
-+ u64 *ns;
-+
-+ memset(&shhwtstamps, 0, sizeof(shhwtstamps));
-+
-+ ns = dpaa2_get_ts(buffer_start, true);
-+ *ns = DPAA2_PTP_NOMINAL_FREQ_PERIOD_NS * le64_to_cpup(ns);
-+ shhwtstamps.hwtstamp = ns_to_ktime(*ns);
-+ skb_tstamp_tx(skb, &shhwtstamps);
-+ }
-+
-+ /* Free SGT buffer allocated on tx */
-+ if (fd_format != dpaa2_fd_single)
-+ skb_free_frag(buffer_start);
-+
-+ /* Move on with skb release */
-+ napi_consume_skb(skb, in_napi);
-+}
-+
-+static netdev_tx_t dpaa2_eth_tx(struct sk_buff *skb, struct net_device *net_dev)
-+{
-+ struct dpaa2_eth_priv *priv = netdev_priv(net_dev);
-+ struct device *dev = net_dev->dev.parent;
-+ struct dpaa2_fd fd;
-+ struct rtnl_link_stats64 *percpu_stats;
-+ struct dpaa2_eth_drv_stats *percpu_extras;
-+ unsigned int needed_headroom;
-+ struct dpaa2_eth_fq *fq;
-+ u16 queue_mapping;
-+ int err, i, ch_id = 0, qpri = 0;
-+
-+ queue_mapping = skb_get_queue_mapping(skb);
-+ fq = &priv->fq[queue_mapping];
-+
-+ /* If we're congested, stop this tx queue; transmission of
-+ * the current skb happens regardless of congestion state
-+ */
-+ dma_sync_single_for_cpu(dev, priv->cscn_dma,
-+ DPAA2_CSCN_SIZE, DMA_FROM_DEVICE);
-+ if (unlikely(dpaa2_cscn_state_congested(priv->cscn_mem))) {
-+ netif_stop_subqueue(net_dev, queue_mapping);
-+ fq->stats.congestion_entry++;
-+ }
-+
-+ percpu_stats = this_cpu_ptr(priv->percpu_stats);
-+ percpu_extras = this_cpu_ptr(priv->percpu_extras);
-+
-+ /* For non-linear skb we don't need a minimum headroom */
-+ needed_headroom = dpaa2_eth_needed_headroom(priv, skb);
-+ if (skb_headroom(skb) < needed_headroom) {
-+ struct sk_buff *ns;
-+
-+ ns = skb_realloc_headroom(skb, needed_headroom);
-+ if (unlikely(!ns)) {
-+ percpu_stats->tx_dropped++;
-+ goto err_alloc_headroom;
-+ }
-+ percpu_extras->tx_reallocs++;
-+ if (skb->sk)
-+ skb_set_owner_w(ns, skb->sk);
-+ dev_kfree_skb(skb);
-+ skb = ns;
-+ }
-+
-+ /* We'll be holding a back-reference to the skb until Tx Confirmation;
-+ * we don't want that overwritten by a concurrent Tx with a cloned skb.
-+ */
-+ skb = skb_unshare(skb, GFP_ATOMIC);
-+ if (unlikely(!skb)) {
-+ /* skb_unshare() has already freed the skb */
-+ percpu_stats->tx_dropped++;
-+ return NETDEV_TX_OK;
-+ }
-+
-+ /* Setup the FD fields */
-+ memset(&fd, 0, sizeof(fd));
-+
-+ if (skb_is_nonlinear(skb)) {
-+ err = build_sg_fd(priv, skb, &fd);
-+ percpu_extras->tx_sg_frames++;
-+ percpu_extras->tx_sg_bytes += skb->len;
-+ } else {
-+ err = build_single_fd(priv, skb, &fd);
-+ }
-+
-+ if (unlikely(err)) {
-+ percpu_stats->tx_dropped++;
-+ goto err_build_fd;
-+ }
-+
-+ /* Tracing point */
-+ trace_dpaa2_tx_fd(net_dev, &fd);
-+
-+ if (dpaa2_eth_ceetm_is_enabled(priv)) {
-+ err = dpaa2_ceetm_classify(skb, net_dev->qdisc, &ch_id, &qpri);
-+ if (err)
-+ goto err_ceetm_classify;
-+ }
-+
-+ for (i = 0; i < DPAA2_ETH_ENQUEUE_RETRIES; i++) {
-+ err = dpaa2_io_service_enqueue_qd(fq->channel->dpio,
-+ priv->tx_qdid, qpri,
-+ fq->tx_qdbin, &fd);
-+ if (err != -EBUSY)
-+ break;
-+ }
-+ percpu_extras->tx_portal_busy += i;
-+ if (unlikely(err < 0)) {
-+ percpu_stats->tx_errors++;
-+ /* Clean up everything, including freeing the skb */
-+ free_tx_fd(priv, &fd, false);
-+ } else {
-+ percpu_stats->tx_packets++;
-+ percpu_stats->tx_bytes += dpaa2_fd_get_len(&fd);
-+ }
-+
-+ return NETDEV_TX_OK;
-+
-+err_ceetm_classify:
-+ free_tx_fd(priv, &fd, false);
-+err_build_fd:
-+err_alloc_headroom:
-+ dev_kfree_skb(skb);
-+
-+ return NETDEV_TX_OK;
-+}
-+
-+/* Tx confirmation frame processing routine */
-+static void dpaa2_eth_tx_conf(struct dpaa2_eth_priv *priv,
-+ struct dpaa2_eth_channel *ch,
-+ const struct dpaa2_fd *fd,
-+ struct napi_struct *napi __always_unused,
-+ u16 queue_id)
-+{
-+ struct device *dev = priv->net_dev->dev.parent;
-+ struct rtnl_link_stats64 *percpu_stats;
-+ struct dpaa2_eth_drv_stats *percpu_extras;
-+ u32 fd_errors;
-+
-+ /* Tracing point */
-+ trace_dpaa2_tx_conf_fd(priv->net_dev, fd);
-+
-+ percpu_extras = this_cpu_ptr(priv->percpu_extras);
-+ percpu_extras->tx_conf_frames++;
-+ percpu_extras->tx_conf_bytes += dpaa2_fd_get_len(fd);
-+
-+ /* Check congestion state and wake all queues if necessary */
-+ if (unlikely(__netif_subqueue_stopped(priv->net_dev, queue_id))) {
-+ dma_sync_single_for_cpu(dev, priv->cscn_dma,
-+ DPAA2_CSCN_SIZE, DMA_FROM_DEVICE);
-+ if (!dpaa2_cscn_state_congested(priv->cscn_mem))
-+ netif_tx_wake_all_queues(priv->net_dev);
-+ }
-+
-+ /* Check frame errors in the FD field */
-+ fd_errors = dpaa2_fd_get_ctrl(fd) & DPAA2_FD_TX_ERR_MASK;
-+ free_tx_fd(priv, fd, true);
-+
-+ if (likely(!fd_errors))
-+ return;
-+
-+ if (net_ratelimit())
-+ netdev_dbg(priv->net_dev, "TX frame FD error: 0x%08x\n",
-+ fd_errors);
-+
-+ percpu_stats = this_cpu_ptr(priv->percpu_stats);
-+ /* Tx-conf logically pertains to the egress path. */
-+ percpu_stats->tx_errors++;
-+}
-+
-+static int set_rx_csum(struct dpaa2_eth_priv *priv, bool enable)
-+{
-+ int err;
-+
-+ err = dpni_set_offload(priv->mc_io, 0, priv->mc_token,
-+ DPNI_OFF_RX_L3_CSUM, enable);
-+ if (err) {
-+ netdev_err(priv->net_dev,
-+ "dpni_set_offload(RX_L3_CSUM) failed\n");
-+ return err;
-+ }
-+
-+ err = dpni_set_offload(priv->mc_io, 0, priv->mc_token,
-+ DPNI_OFF_RX_L4_CSUM, enable);
-+ if (err) {
-+ netdev_err(priv->net_dev,
-+ "dpni_set_offload(RX_L4_CSUM) failed\n");
-+ return err;
-+ }
-+
-+ return 0;
-+}
-+
-+static int set_tx_csum(struct dpaa2_eth_priv *priv, bool enable)
-+{
-+ int err;
-+
-+ err = dpni_set_offload(priv->mc_io, 0, priv->mc_token,
-+ DPNI_OFF_TX_L3_CSUM, enable);
-+ if (err) {
-+ netdev_err(priv->net_dev, "dpni_set_offload(TX_L3_CSUM) failed\n");
-+ return err;
-+ }
-+
-+ err = dpni_set_offload(priv->mc_io, 0, priv->mc_token,
-+ DPNI_OFF_TX_L4_CSUM, enable);
-+ if (err) {
-+ netdev_err(priv->net_dev, "dpni_set_offload(TX_L4_CSUM) failed\n");
-+ return err;
-+ }
-+
-+ return 0;
-+}
-+
-+/* Perform a single release command to add buffers
-+ * to the specified buffer pool
-+ */
-+static int add_bufs(struct dpaa2_eth_priv *priv,
-+ struct dpaa2_eth_channel *ch, u16 bpid)
-+{
-+ struct device *dev = priv->net_dev->dev.parent;
-+ u64 buf_array[DPAA2_ETH_BUFS_PER_CMD];
-+ void *buf;
-+ dma_addr_t addr;
-+ int i, err;
-+
-+ for (i = 0; i < DPAA2_ETH_BUFS_PER_CMD; i++) {
-+ /* Allocate buffer visible to WRIOP + skb shared info +
-+ * alignment padding
-+ */
-+ buf = napi_alloc_frag(dpaa2_eth_buf_raw_size(priv));
-+ if (unlikely(!buf))
-+ goto err_alloc;
-+
-+ buf = PTR_ALIGN(buf, priv->rx_buf_align);
-+
-+ addr = dma_map_single(dev, buf, DPAA2_ETH_RX_BUF_SIZE,
-+ DMA_BIDIRECTIONAL);
-+ if (unlikely(dma_mapping_error(dev, addr)))
-+ goto err_map;
-+
-+ buf_array[i] = addr;
-+
-+ /* tracing point */
-+ trace_dpaa2_eth_buf_seed(priv->net_dev,
-+ buf, dpaa2_eth_buf_raw_size(priv),
-+ addr, DPAA2_ETH_RX_BUF_SIZE,
-+ bpid);
-+ }
-+
-+release_bufs:
-+ /* In case the portal is busy, retry until successful */
-+ while ((err = dpaa2_io_service_release(ch->dpio, bpid,
-+ buf_array, i)) == -EBUSY)
-+ cpu_relax();
-+
-+ /* If release command failed, clean up and bail out; not much
-+ * else we can do about it
-+ */
-+ if (err) {
-+ free_bufs(priv, buf_array, i);
-+ return 0;
-+ }
-+
-+ return i;
-+
-+err_map:
-+ skb_free_frag(buf);
-+err_alloc:
-+ /* If we managed to allocate at least some buffers, release them */
-+ if (i)
-+ goto release_bufs;
-+
-+ return 0;
-+}
-+
-+static int seed_pool(struct dpaa2_eth_priv *priv, u16 bpid)
-+{
-+ int i, j;
-+ int new_count;
-+
-+ /* This is the lazy seeding of Rx buffer pools.
-+ * dpaa2_add_bufs() is also used on the Rx hotpath and calls
-+ * napi_alloc_frag(). The trouble with that is that it in turn ends up
-+ * calling this_cpu_ptr(), which mandates execution in atomic context.
-+ * Rather than splitting up the code, do a one-off preempt disable.
-+ */
-+ preempt_disable();
-+ for (j = 0; j < priv->num_channels; j++) {
-+ priv->channel[j]->buf_count = 0;
-+ for (i = 0; i < priv->max_bufs_per_ch;
-+ i += DPAA2_ETH_BUFS_PER_CMD) {
-+ new_count = add_bufs(priv, priv->channel[j], bpid);
-+ priv->channel[j]->buf_count += new_count;
-+
-+ if (new_count < DPAA2_ETH_BUFS_PER_CMD) {
-+ preempt_enable();
-+ return -ENOMEM;
-+ }
-+ }
-+ }
-+ preempt_enable();
-+
-+ return 0;
-+}
-+
-+/**
-+ * Drain the specified number of buffers from the DPNI's private buffer pool.
-+ * @count must not exceeed DPAA2_ETH_BUFS_PER_CMD
-+ */
-+static void drain_bufs(struct dpaa2_eth_priv *priv, int count)
-+{
-+ u64 buf_array[DPAA2_ETH_BUFS_PER_CMD];
-+ int ret;
-+
-+ do {
-+ ret = dpaa2_io_service_acquire(NULL, priv->bpid,
-+ buf_array, count);
-+ if (ret < 0) {
-+ netdev_err(priv->net_dev, "dpaa2_io_service_acquire() failed\n");
-+ return;
-+ }
-+ free_bufs(priv, buf_array, ret);
-+ } while (ret);
-+}
-+
-+static void drain_pool(struct dpaa2_eth_priv *priv)
-+{
-+ preempt_disable();
-+ drain_bufs(priv, DPAA2_ETH_BUFS_PER_CMD);
-+ drain_bufs(priv, 1);
-+ preempt_enable();
-+}
-+
-+/* Function is called from softirq context only, so we don't need to guard
-+ * the access to percpu count
-+ */
-+static int refill_pool(struct dpaa2_eth_priv *priv,
-+ struct dpaa2_eth_channel *ch,
-+ u16 bpid)
-+{
-+ int new_count;
-+
-+ if (likely(ch->buf_count >= priv->refill_thresh))
-+ return 0;
-+
-+ do {
-+ new_count = add_bufs(priv, ch, bpid);
-+ if (unlikely(!new_count)) {
-+ /* Out of memory; abort for now, we'll try later on */
-+ break;
-+ }
-+ ch->buf_count += new_count;
-+ } while (ch->buf_count < priv->max_bufs_per_ch);
-+
-+ if (unlikely(ch->buf_count < priv->max_bufs_per_ch))
-+ return -ENOMEM;
-+
-+ return 0;
-+}
-+
-+static int pull_channel(struct dpaa2_eth_channel *ch)
-+{
-+ int err;
-+ int dequeues = -1;
-+
-+ /* Retry while portal is busy */
-+ do {
-+ err = dpaa2_io_service_pull_channel(ch->dpio, ch->ch_id,
-+ ch->store);
-+ dequeues++;
-+ cpu_relax();
-+ } while (err == -EBUSY);
-+
-+ ch->stats.dequeue_portal_busy += dequeues;
-+ if (unlikely(err))
-+ ch->stats.pull_err++;
-+
-+ return err;
-+}
-+
-+/* NAPI poll routine
-+ *
-+ * Frames are dequeued from the QMan channel associated with this NAPI context.
-+ * Rx and (if configured) Rx error frames count towards the NAPI budget. Tx
-+ * confirmation frames are limited by a threshold per NAPI poll cycle.
-+ */
-+static int dpaa2_eth_poll(struct napi_struct *napi, int budget)
-+{
-+ struct dpaa2_eth_channel *ch;
-+ int rx_cleaned = 0, tx_conf_cleaned = 0;
-+ bool store_cleaned;
-+ struct dpaa2_eth_priv *priv;
-+ int err;
-+
-+ ch = container_of(napi, struct dpaa2_eth_channel, napi);
-+ priv = ch->priv;
-+
-+ do {
-+ err = pull_channel(ch);
-+ if (unlikely(err))
-+ break;
-+
-+ /* Refill pool if appropriate */
-+ refill_pool(priv, ch, priv->bpid);
-+
-+ store_cleaned = consume_frames(ch, &rx_cleaned,
-+ &tx_conf_cleaned);
-+
-+ /* If we've either consumed the budget with Rx frames,
-+ * or reached the Tx conf threshold, we're done.
-+ */
-+ if (rx_cleaned >= budget ||
-+ tx_conf_cleaned >= TX_CONF_PER_NAPI_POLL)
-+ return budget;
-+ } while (store_cleaned);
-+
-+ /* We didn't consume the entire budget, finish napi and
-+ * re-enable data availability notifications
-+ */
-+ napi_complete(napi);
-+ do {
-+ err = dpaa2_io_service_rearm(ch->dpio, &ch->nctx);
-+ cpu_relax();
-+ } while (err == -EBUSY);
-+ WARN_ONCE(err, "CDAN notifications rearm failed on core %d",
-+ ch->nctx.desired_cpu);
-+
-+ return max(rx_cleaned, 1);
-+}
-+
-+static void enable_ch_napi(struct dpaa2_eth_priv *priv)
-+{
-+ struct dpaa2_eth_channel *ch;
-+ int i;
-+
-+ for (i = 0; i < priv->num_channels; i++) {
-+ ch = priv->channel[i];
-+ napi_enable(&ch->napi);
-+ }
-+}
-+
-+static void disable_ch_napi(struct dpaa2_eth_priv *priv)
-+{
-+ struct dpaa2_eth_channel *ch;
-+ int i;
-+
-+ for (i = 0; i < priv->num_channels; i++) {
-+ ch = priv->channel[i];
-+ napi_disable(&ch->napi);
-+ }
-+}
-+
-+static int link_state_update(struct dpaa2_eth_priv *priv)
-+{
-+ struct dpni_link_state state;
-+ int err;
-+
-+ err = dpni_get_link_state(priv->mc_io, 0, priv->mc_token, &state);
-+ if (unlikely(err)) {
-+ netdev_err(priv->net_dev,
-+ "dpni_get_link_state() failed\n");
-+ return err;
-+ }
-+
-+ /* Chech link state; speed / duplex changes are not treated yet */
-+ if (priv->link_state.up == state.up)
-+ return 0;
-+
-+ priv->link_state = state;
-+ if (state.up) {
-+ netif_carrier_on(priv->net_dev);
-+ netif_tx_start_all_queues(priv->net_dev);
-+ } else {
-+ netif_tx_stop_all_queues(priv->net_dev);
-+ netif_carrier_off(priv->net_dev);
-+ }
-+
-+ netdev_info(priv->net_dev, "Link Event: state %s\n",
-+ state.up ? "up" : "down");
-+
-+ return 0;
-+}
-+
-+static int dpaa2_eth_open(struct net_device *net_dev)
-+{
-+ struct dpaa2_eth_priv *priv = netdev_priv(net_dev);
-+ int err;
-+
-+ /* We'll only start the txqs when the link is actually ready; make sure
-+ * we don't race against the link up notification, which may come
-+ * immediately after dpni_enable();
-+ */
-+ netif_tx_stop_all_queues(net_dev);
-+
-+ /* Also, explicitly set carrier off, otherwise netif_carrier_ok() will
-+ * return true and cause 'ip link show' to report the LOWER_UP flag,
-+ * even though the link notification wasn't even received.
-+ */
-+ netif_carrier_off(net_dev);
-+
-+ err = seed_pool(priv, priv->bpid);
-+ if (err) {
-+ /* Not much to do; the buffer pool, though not filled up,
-+ * may still contain some buffers which would enable us
-+ * to limp on.
-+ */
-+ netdev_err(net_dev, "Buffer seeding failed for DPBP %d (bpid=%d)\n",
-+ priv->dpbp_dev->obj_desc.id, priv->bpid);
-+ }
-+
-+ priv->refill_thresh = DPAA2_ETH_REFILL_THRESH(priv);
-+
-+ err = dpni_enable(priv->mc_io, 0, priv->mc_token);
-+ if (err < 0) {
-+ netdev_err(net_dev, "dpni_enable() failed\n");
-+ goto enable_err;
-+ }
-+
-+ /* If the DPMAC object has already processed the link up interrupt,
-+ * we have to learn the link state ourselves.
-+ */
-+ err = link_state_update(priv);
-+ if (err < 0) {
-+ netdev_err(net_dev, "Can't update link state\n");
-+ goto link_state_err;
-+ }
-+
-+ return 0;
-+
-+link_state_err:
-+enable_err:
-+ priv->refill_thresh = 0;
-+ drain_pool(priv);
-+ return err;
-+}
-+
-+static int dpaa2_eth_stop(struct net_device *net_dev)
-+{
-+ struct dpaa2_eth_priv *priv = netdev_priv(net_dev);
-+ int dpni_enabled;
-+ int retries = 10, i;
-+ int err = 0;
-+
-+ netif_tx_stop_all_queues(net_dev);
-+ netif_carrier_off(net_dev);
-+
-+ /* Loop while dpni_disable() attempts to drain the egress FQs
-+ * and confirm them back to us.
-+ */
-+ do {
-+ dpni_disable(priv->mc_io, 0, priv->mc_token);
-+ dpni_is_enabled(priv->mc_io, 0, priv->mc_token, &dpni_enabled);
-+ if (dpni_enabled)
-+ /* Allow the hardware some slack */
-+ msleep(100);
-+ } while (dpni_enabled && --retries);
-+ if (!retries) {
-+ netdev_warn(net_dev, "Retry count exceeded disabling DPNI\n");
-+ /* Must go on and finish processing pending frames, so we don't
-+ * crash at the next "ifconfig up"
-+ */
-+ err = -ETIMEDOUT;
-+ }
-+
-+ priv->refill_thresh = 0;
-+
-+ /* Wait for all running napi poll routines to finish, so that no
-+ * new refill operations are started
-+ */
-+ for (i = 0; i < priv->num_channels; i++)
-+ napi_synchronize(&priv->channel[i]->napi);
-+
-+ /* Empty the buffer pool */
-+ drain_pool(priv);
-+
-+ return err;
-+}
-+
-+static int dpaa2_eth_init(struct net_device *net_dev)
-+{
-+ u64 supported = 0;
-+ u64 not_supported = 0;
-+ struct dpaa2_eth_priv *priv = netdev_priv(net_dev);
-+ u32 options = priv->dpni_attrs.options;
-+
-+ /* Capabilities listing */
-+ supported |= IFF_LIVE_ADDR_CHANGE;
-+
-+ if (options & DPNI_OPT_NO_MAC_FILTER)
-+ not_supported |= IFF_UNICAST_FLT;
-+ else
-+ supported |= IFF_UNICAST_FLT;
-+
-+ net_dev->priv_flags |= supported;
-+ net_dev->priv_flags &= ~not_supported;
-+
-+ /* Features */
-+ net_dev->features = NETIF_F_RXCSUM |
-+ NETIF_F_IP_CSUM | NETIF_F_IPV6_CSUM |
-+ NETIF_F_SG | NETIF_F_HIGHDMA |
-+ NETIF_F_LLTX;
-+ net_dev->hw_features = net_dev->features;
-+
-+ return 0;
-+}
-+
-+static int dpaa2_eth_set_addr(struct net_device *net_dev, void *addr)
-+{
-+ struct dpaa2_eth_priv *priv = netdev_priv(net_dev);
-+ struct device *dev = net_dev->dev.parent;
-+ int err;
-+
-+ err = eth_mac_addr(net_dev, addr);
-+ if (err < 0) {
-+ dev_err(dev, "eth_mac_addr() failed (%d)\n", err);
-+ return err;
-+ }
-+
-+ err = dpni_set_primary_mac_addr(priv->mc_io, 0, priv->mc_token,
-+ net_dev->dev_addr);
-+ if (err) {
-+ dev_err(dev, "dpni_set_primary_mac_addr() failed (%d)\n", err);
-+ return err;
-+ }
-+
-+ return 0;
-+}
-+
-+/** Fill in counters maintained by the GPP driver. These may be different from
-+ * the hardware counters obtained by ethtool.
-+ */
-+static struct rtnl_link_stats64 *dpaa2_eth_get_stats(struct net_device *net_dev,
-+ struct rtnl_link_stats64 *stats)
-+{
-+ struct dpaa2_eth_priv *priv = netdev_priv(net_dev);
-+ struct rtnl_link_stats64 *percpu_stats;
-+ u64 *cpustats;
-+ u64 *netstats = (u64 *)stats;
-+ int i, j;
-+ int num = sizeof(struct rtnl_link_stats64) / sizeof(u64);
-+
-+ for_each_possible_cpu(i) {
-+ percpu_stats = per_cpu_ptr(priv->percpu_stats, i);
-+ cpustats = (u64 *)percpu_stats;
-+ for (j = 0; j < num; j++)
-+ netstats[j] += cpustats[j];
-+ }
-+ return stats;
-+}
-+
-+/* Copy mac unicast addresses from @net_dev to @priv.
-+ * Its sole purpose is to make dpaa2_eth_set_rx_mode() more readable.
-+ */
-+static void add_uc_hw_addr(const struct net_device *net_dev,
-+ struct dpaa2_eth_priv *priv)
-+{
-+ struct netdev_hw_addr *ha;
-+ int err;
-+
-+ netdev_for_each_uc_addr(ha, net_dev) {
-+ err = dpni_add_mac_addr(priv->mc_io, 0, priv->mc_token,
-+ ha->addr);
-+ if (err)
-+ netdev_warn(priv->net_dev,
-+ "Could not add ucast MAC %pM to the filtering table (err %d)\n",
-+ ha->addr, err);
-+ }
-+}
-+
-+/* Copy mac multicast addresses from @net_dev to @priv
-+ * Its sole purpose is to make dpaa2_eth_set_rx_mode() more readable.
-+ */
-+static void add_mc_hw_addr(const struct net_device *net_dev,
-+ struct dpaa2_eth_priv *priv)
-+{
-+ struct netdev_hw_addr *ha;
-+ int err;
-+
-+ netdev_for_each_mc_addr(ha, net_dev) {
-+ err = dpni_add_mac_addr(priv->mc_io, 0, priv->mc_token,
-+ ha->addr);
-+ if (err)
-+ netdev_warn(priv->net_dev,
-+ "Could not add mcast MAC %pM to the filtering table (err %d)\n",
-+ ha->addr, err);
-+ }
-+}
-+
-+static void dpaa2_eth_set_rx_mode(struct net_device *net_dev)
-+{
-+ struct dpaa2_eth_priv *priv = netdev_priv(net_dev);
-+ int uc_count = netdev_uc_count(net_dev);
-+ int mc_count = netdev_mc_count(net_dev);
-+ u8 max_mac = priv->dpni_attrs.mac_filter_entries;
-+ u32 options = priv->dpni_attrs.options;
-+ u16 mc_token = priv->mc_token;
-+ struct fsl_mc_io *mc_io = priv->mc_io;
-+ int err;
-+
-+ /* Basic sanity checks; these probably indicate a misconfiguration */
-+ if (options & DPNI_OPT_NO_MAC_FILTER && max_mac != 0)
-+ netdev_info(net_dev,
-+ "mac_filter_entries=%d, DPNI_OPT_NO_MAC_FILTER option must be disabled\n",
-+ max_mac);
-+
-+ /* Force promiscuous if the uc or mc counts exceed our capabilities. */
-+ if (uc_count > max_mac) {
-+ netdev_info(net_dev,
-+ "Unicast addr count reached %d, max allowed is %d; forcing promisc\n",
-+ uc_count, max_mac);
-+ goto force_promisc;
-+ }
-+ if (mc_count + uc_count > max_mac) {
-+ netdev_info(net_dev,
-+ "Unicast + multicast addr count reached %d, max allowed is %d; forcing promisc\n",
-+ uc_count + mc_count, max_mac);
-+ goto force_mc_promisc;
-+ }
-+
-+ /* Adjust promisc settings due to flag combinations */
-+ if (net_dev->flags & IFF_PROMISC)
-+ goto force_promisc;
-+ if (net_dev->flags & IFF_ALLMULTI) {
-+ /* First, rebuild unicast filtering table. This should be done
-+ * in promisc mode, in order to avoid frame loss while we
-+ * progressively add entries to the table.
-+ * We don't know whether we had been in promisc already, and
-+ * making an MC call to find out is expensive; so set uc promisc
-+ * nonetheless.
-+ */
-+ err = dpni_set_unicast_promisc(mc_io, 0, mc_token, 1);
-+ if (err)
-+ netdev_warn(net_dev, "Can't set uc promisc\n");
-+
-+ /* Actual uc table reconstruction. */
-+ err = dpni_clear_mac_filters(mc_io, 0, mc_token, 1, 0);
-+ if (err)
-+ netdev_warn(net_dev, "Can't clear uc filters\n");
-+ add_uc_hw_addr(net_dev, priv);
-+
-+ /* Finally, clear uc promisc and set mc promisc as requested. */
-+ err = dpni_set_unicast_promisc(mc_io, 0, mc_token, 0);
-+ if (err)
-+ netdev_warn(net_dev, "Can't clear uc promisc\n");
-+ goto force_mc_promisc;
-+ }
-+
-+ /* Neither unicast, nor multicast promisc will be on... eventually.
-+ * For now, rebuild mac filtering tables while forcing both of them on.
-+ */
-+ err = dpni_set_unicast_promisc(mc_io, 0, mc_token, 1);
-+ if (err)
-+ netdev_warn(net_dev, "Can't set uc promisc (%d)\n", err);
-+ err = dpni_set_multicast_promisc(mc_io, 0, mc_token, 1);
-+ if (err)
-+ netdev_warn(net_dev, "Can't set mc promisc (%d)\n", err);
-+
-+ /* Actual mac filtering tables reconstruction */
-+ err = dpni_clear_mac_filters(mc_io, 0, mc_token, 1, 1);
-+ if (err)
-+ netdev_warn(net_dev, "Can't clear mac filters\n");
-+ add_mc_hw_addr(net_dev, priv);
-+ add_uc_hw_addr(net_dev, priv);
-+
-+ /* Now we can clear both ucast and mcast promisc, without risking
-+ * to drop legitimate frames anymore.
-+ */
-+ err = dpni_set_unicast_promisc(mc_io, 0, mc_token, 0);
-+ if (err)
-+ netdev_warn(net_dev, "Can't clear ucast promisc\n");
-+ err = dpni_set_multicast_promisc(mc_io, 0, mc_token, 0);
-+ if (err)
-+ netdev_warn(net_dev, "Can't clear mcast promisc\n");
-+
-+ return;
-+
-+force_promisc:
-+ err = dpni_set_unicast_promisc(mc_io, 0, mc_token, 1);
-+ if (err)
-+ netdev_warn(net_dev, "Can't set ucast promisc\n");
-+force_mc_promisc:
-+ err = dpni_set_multicast_promisc(mc_io, 0, mc_token, 1);
-+ if (err)
-+ netdev_warn(net_dev, "Can't set mcast promisc\n");
-+}
-+
-+static int dpaa2_eth_set_features(struct net_device *net_dev,
-+ netdev_features_t features)
-+{
-+ struct dpaa2_eth_priv *priv = netdev_priv(net_dev);
-+ netdev_features_t changed = features ^ net_dev->features;
-+ bool enable;
-+ int err;
-+
-+ if (changed & NETIF_F_RXCSUM) {
-+ enable = !!(features & NETIF_F_RXCSUM);
-+ err = set_rx_csum(priv, enable);
-+ if (err)
-+ return err;
-+ }
-+
-+ if (changed & (NETIF_F_IP_CSUM | NETIF_F_IPV6_CSUM)) {
-+ enable = !!(features & (NETIF_F_IP_CSUM | NETIF_F_IPV6_CSUM));
-+ err = set_tx_csum(priv, enable);
-+ if (err)
-+ return err;
-+ }
-+
-+ return 0;
-+}
-+
-+static int dpaa2_eth_ts_ioctl(struct net_device *dev, struct ifreq *rq, int cmd)
-+{
-+ struct dpaa2_eth_priv *priv = netdev_priv(dev);
-+ struct hwtstamp_config config;
-+
-+ if (copy_from_user(&config, rq->ifr_data, sizeof(config)))
-+ return -EFAULT;
-+
-+ switch (config.tx_type) {
-+ case HWTSTAMP_TX_OFF:
-+ priv->ts_tx_en = false;
-+ break;
-+ case HWTSTAMP_TX_ON:
-+ priv->ts_tx_en = true;
-+ break;
-+ default:
-+ return -ERANGE;
-+ }
-+
-+ if (config.rx_filter == HWTSTAMP_FILTER_NONE) {
-+ priv->ts_rx_en = false;
-+ } else {
-+ priv->ts_rx_en = true;
-+ /* TS is set for all frame types, not only those requested */
-+ config.rx_filter = HWTSTAMP_FILTER_ALL;
-+ }
-+
-+ return copy_to_user(rq->ifr_data, &config, sizeof(config)) ?
-+ -EFAULT : 0;
-+}
-+
-+static int dpaa2_eth_ioctl(struct net_device *dev, struct ifreq *rq, int cmd)
-+{
-+ if (cmd == SIOCSHWTSTAMP)
-+ return dpaa2_eth_ts_ioctl(dev, rq, cmd);
-+
-+ return -EINVAL;
-+}
-+
-+static int set_buffer_layout(struct dpaa2_eth_priv *priv)
-+{
-+ struct device *dev = priv->net_dev->dev.parent;
-+ struct dpni_buffer_layout buf_layout = {0};
-+ int err;
-+
-+ /* We need to check for WRIOP version 1.0.0, but depending on the MC
-+ * version, this number is not always provided correctly on rev1.
-+ * We need to check for both alternatives in this situation.
-+ */
-+ if (priv->dpni_attrs.wriop_version == DPAA2_WRIOP_VERSION(0, 0, 0) ||
-+ priv->dpni_attrs.wriop_version == DPAA2_WRIOP_VERSION(1, 0, 0))
-+ priv->rx_buf_align = DPAA2_ETH_RX_BUF_ALIGN_REV1;
-+ else
-+ priv->rx_buf_align = DPAA2_ETH_RX_BUF_ALIGN;
-+
-+ /* tx buffer */
-+ buf_layout.pass_timestamp = true;
-+ buf_layout.private_data_size = DPAA2_ETH_SWA_SIZE;
-+ buf_layout.options = DPNI_BUF_LAYOUT_OPT_TIMESTAMP |
-+ DPNI_BUF_LAYOUT_OPT_PRIVATE_DATA_SIZE;
-+ err = dpni_set_buffer_layout(priv->mc_io, 0, priv->mc_token,
-+ DPNI_QUEUE_TX, &buf_layout);
-+ if (err) {
-+ dev_err(dev, "dpni_set_buffer_layout(TX) failed\n");
-+ return err;
-+ }
-+
-+ /* tx-confirm buffer */
-+ buf_layout.options = DPNI_BUF_LAYOUT_OPT_TIMESTAMP;
-+ err = dpni_set_buffer_layout(priv->mc_io, 0, priv->mc_token,
-+ DPNI_QUEUE_TX_CONFIRM, &buf_layout);
-+ if (err) {
-+ dev_err(dev, "dpni_set_buffer_layout(TX_CONF) failed\n");
-+ return err;
-+ }
-+
-+ /* Now that we've set our tx buffer layout, retrieve the minimum
-+ * required tx data offset.
-+ */
-+ err = dpni_get_tx_data_offset(priv->mc_io, 0, priv->mc_token,
-+ &priv->tx_data_offset);
-+ if (err) {
-+ dev_err(dev, "dpni_get_tx_data_offset() failed\n");
-+ return err;
-+ }
-+
-+ if ((priv->tx_data_offset % 64) != 0)
-+ dev_warn(dev, "Tx data offset (%d) not a multiple of 64B\n",
-+ priv->tx_data_offset);
-+
-+ /* rx buffer */
-+ buf_layout.pass_frame_status = true;
-+ buf_layout.pass_parser_result = true;
-+ buf_layout.data_align = priv->rx_buf_align;
-+ buf_layout.private_data_size = 0;
-+ buf_layout.data_head_room = dpaa2_eth_rx_headroom(priv);
-+ buf_layout.options = DPNI_BUF_LAYOUT_OPT_PARSER_RESULT |
-+ DPNI_BUF_LAYOUT_OPT_FRAME_STATUS |
-+ DPNI_BUF_LAYOUT_OPT_DATA_ALIGN |
-+ DPNI_BUF_LAYOUT_OPT_DATA_HEAD_ROOM |
-+ DPNI_BUF_LAYOUT_OPT_TIMESTAMP;
-+ err = dpni_set_buffer_layout(priv->mc_io, 0, priv->mc_token,
-+ DPNI_QUEUE_RX, &buf_layout);
-+ if (err) {
-+ dev_err(dev, "dpni_set_buffer_layout(RX) failed\n");
-+ return err;
-+ }
-+
-+ return 0;
-+}
-+
-+static int dpaa2_eth_set_xdp(struct net_device *net_dev, struct bpf_prog *prog)
-+{
-+ struct dpaa2_eth_priv *priv = netdev_priv(net_dev);
-+ struct dpaa2_eth_channel *ch;
-+ struct bpf_prog *old_prog;
-+ int i, err;
-+
-+ /* No support for SG frames */
-+ if (DPAA2_ETH_L2_MAX_FRM(net_dev->mtu) > DPAA2_ETH_RX_BUF_SIZE)
-+ return -EINVAL;
-+
-+ if (netif_running(net_dev)) {
-+ err = dpaa2_eth_stop(net_dev);
-+ if (err)
-+ return err;
-+ }
-+
-+ if (prog) {
-+ prog = bpf_prog_add(prog, priv->num_channels - 1);
-+ if (IS_ERR(prog))
-+ return PTR_ERR(prog);
-+ }
-+
-+ priv->has_xdp_prog = !!prog;
-+
-+ for (i = 0; i < priv->num_channels; i++) {
-+ ch = priv->channel[i];
-+ old_prog = xchg(&ch->xdp_prog, prog);
-+ if (old_prog)
-+ bpf_prog_put(old_prog);
-+ }
-+
-+ if (netif_running(net_dev)) {
-+ err = dpaa2_eth_open(net_dev);
-+ if (err)
-+ return err;
-+ }
-+
-+ return 0;
-+}
-+
-+static int dpaa2_eth_xdp(struct net_device *dev, struct netdev_xdp *xdp)
-+{
-+ struct dpaa2_eth_priv *priv = netdev_priv(dev);
-+
-+ switch (xdp->command) {
-+ case XDP_SETUP_PROG:
-+ return dpaa2_eth_set_xdp(dev, xdp->prog);
-+ case XDP_QUERY_PROG:
-+ xdp->prog_attached = priv->has_xdp_prog;
-+ return 0;
-+ default:
-+ return -EINVAL;
-+ }
-+}
-+
-+static const struct net_device_ops dpaa2_eth_ops = {
-+ .ndo_open = dpaa2_eth_open,
-+ .ndo_start_xmit = dpaa2_eth_tx,
-+ .ndo_stop = dpaa2_eth_stop,
-+ .ndo_init = dpaa2_eth_init,
-+ .ndo_set_mac_address = dpaa2_eth_set_addr,
-+ .ndo_get_stats64 = dpaa2_eth_get_stats,
-+ .ndo_set_rx_mode = dpaa2_eth_set_rx_mode,
-+ .ndo_set_features = dpaa2_eth_set_features,
-+ .ndo_do_ioctl = dpaa2_eth_ioctl,
-+ .ndo_xdp = dpaa2_eth_xdp,
-+};
-+
-+static void cdan_cb(struct dpaa2_io_notification_ctx *ctx)
-+{
-+ struct dpaa2_eth_channel *ch;
-+
-+ ch = container_of(ctx, struct dpaa2_eth_channel, nctx);
-+
-+ /* Update NAPI statistics */
-+ ch->stats.cdan++;
-+
-+ napi_schedule_irqoff(&ch->napi);
-+}
-+
-+/* Allocate and configure a DPCON object */
-+static struct fsl_mc_device *setup_dpcon(struct dpaa2_eth_priv *priv)
-+{
-+ struct fsl_mc_device *dpcon;
-+ struct device *dev = priv->net_dev->dev.parent;
-+ struct dpcon_attr attrs;
-+ int err;
-+
-+ err = fsl_mc_object_allocate(to_fsl_mc_device(dev),
-+ FSL_MC_POOL_DPCON, &dpcon);
-+ if (err) {
-+ dev_info(dev, "Not enough DPCONs, will go on as-is\n");
-+ return NULL;
-+ }
-+
-+ err = dpcon_open(priv->mc_io, 0, dpcon->obj_desc.id, &dpcon->mc_handle);
-+ if (err) {
-+ dev_err(dev, "dpcon_open() failed\n");
-+ goto free;
-+ }
-+
-+ err = dpcon_reset(priv->mc_io, 0, dpcon->mc_handle);
-+ if (err) {
-+ dev_err(dev, "dpcon_reset() failed\n");
-+ goto close;
-+ }
-+
-+ err = dpcon_get_attributes(priv->mc_io, 0, dpcon->mc_handle, &attrs);
-+ if (err) {
-+ dev_err(dev, "dpcon_get_attributes() failed\n");
-+ goto close;
-+ }
-+
-+ err = dpcon_enable(priv->mc_io, 0, dpcon->mc_handle);
-+ if (err) {
-+ dev_err(dev, "dpcon_enable() failed\n");
-+ goto close;
-+ }
-+
-+ return dpcon;
-+
-+close:
-+ dpcon_close(priv->mc_io, 0, dpcon->mc_handle);
-+free:
-+ fsl_mc_object_free(dpcon);
-+
-+ return NULL;
-+}
-+
-+static void free_dpcon(struct dpaa2_eth_priv *priv,
-+ struct fsl_mc_device *dpcon)
-+{
-+ dpcon_disable(priv->mc_io, 0, dpcon->mc_handle);
-+ dpcon_close(priv->mc_io, 0, dpcon->mc_handle);
-+ fsl_mc_object_free(dpcon);
-+}
-+
-+static struct dpaa2_eth_channel *
-+alloc_channel(struct dpaa2_eth_priv *priv)
-+{
-+ struct dpaa2_eth_channel *channel;
-+ struct dpcon_attr attr;
-+ struct device *dev = priv->net_dev->dev.parent;
-+ int err;
-+
-+ channel = kzalloc(sizeof(*channel), GFP_KERNEL);
-+ if (!channel)
-+ return NULL;
-+
-+ channel->dpcon = setup_dpcon(priv);
-+ if (!channel->dpcon)
-+ goto err_setup;
-+
-+ err = dpcon_get_attributes(priv->mc_io, 0, channel->dpcon->mc_handle,
-+ &attr);
-+ if (err) {
-+ dev_err(dev, "dpcon_get_attributes() failed\n");
-+ goto err_get_attr;
-+ }
-+
-+ channel->dpcon_id = attr.id;
-+ channel->ch_id = attr.qbman_ch_id;
-+ channel->priv = priv;
-+
-+ return channel;
-+
-+err_get_attr:
-+ free_dpcon(priv, channel->dpcon);
-+err_setup:
-+ kfree(channel);
-+ return NULL;
-+}
-+
-+static void free_channel(struct dpaa2_eth_priv *priv,
-+ struct dpaa2_eth_channel *channel)
-+{
-+ struct bpf_prog *prog;
-+
-+ free_dpcon(priv, channel->dpcon);
-+
-+ prog = READ_ONCE(channel->xdp_prog);
-+ if (prog)
-+ bpf_prog_put(prog);
-+
-+ kfree(channel);
-+}
-+
-+/* DPIO setup: allocate and configure QBMan channels, setup core affinity
-+ * and register data availability notifications
-+ */
-+static int setup_dpio(struct dpaa2_eth_priv *priv)
-+{
-+ struct dpaa2_io_notification_ctx *nctx;
-+ struct dpaa2_eth_channel *channel;
-+ struct dpcon_notification_cfg dpcon_notif_cfg;
-+ struct device *dev = priv->net_dev->dev.parent;
-+ int i, err;
-+
-+ /* We want the ability to spread ingress traffic (RX, TX conf) to as
-+ * many cores as possible, so we need one channel for each core
-+ * (unless there's fewer queues than cores, in which case the extra
-+ * channels would be wasted).
-+ * Allocate one channel per core and register it to the core's
-+ * affine DPIO. If not enough channels are available for all cores
-+ * or if some cores don't have an affine DPIO, there will be no
-+ * ingress frame processing on those cores.
-+ */
-+ cpumask_clear(&priv->dpio_cpumask);
-+ for_each_online_cpu(i) {
-+ /* Try to allocate a channel */
-+ channel = alloc_channel(priv);
-+ if (!channel) {
-+ dev_info(dev,
-+ "No affine channel for cpu %d and above\n", i);
-+ err = -ENODEV;
-+ goto err_alloc_ch;
-+ }
-+
-+ priv->channel[priv->num_channels] = channel;
-+
-+ nctx = &channel->nctx;
-+ nctx->is_cdan = 1;
-+ nctx->cb = cdan_cb;
-+ nctx->id = channel->ch_id;
-+ nctx->desired_cpu = i;
-+
-+ /* Register the new context */
-+ channel->dpio = dpaa2_io_service_select(i);
-+ err = dpaa2_io_service_register(channel->dpio, nctx);
-+ if (err) {
-+ dev_dbg(dev, "No affine DPIO for cpu %d\n", i);
-+ /* If no affine DPIO for this core, there's probably
-+ * none available for next cores either. Signal we want
-+ * to retry later, in case the DPIO devices weren't
-+ * probed yet.
-+ */
-+ err = -EPROBE_DEFER;
-+ goto err_service_reg;
-+ }
-+
-+ /* Register DPCON notification with MC */
-+ dpcon_notif_cfg.dpio_id = nctx->dpio_id;
-+ dpcon_notif_cfg.priority = 0;
-+ dpcon_notif_cfg.user_ctx = nctx->qman64;
-+ err = dpcon_set_notification(priv->mc_io, 0,
-+ channel->dpcon->mc_handle,
-+ &dpcon_notif_cfg);
-+ if (err) {
-+ dev_err(dev, "dpcon_set_notification failed()\n");
-+ goto err_set_cdan;
-+ }
-+
-+ /* If we managed to allocate a channel and also found an affine
-+ * DPIO for this core, add it to the final mask
-+ */
-+ cpumask_set_cpu(i, &priv->dpio_cpumask);
-+ priv->num_channels++;
-+
-+ /* Stop if we already have enough channels to accommodate all
-+ * RX and TX conf queues
-+ */
-+ if (priv->num_channels == dpaa2_eth_queue_count(priv))
-+ break;
-+ }
-+
-+ return 0;
-+
-+err_set_cdan:
-+ dpaa2_io_service_deregister(channel->dpio, nctx);
-+err_service_reg:
-+ free_channel(priv, channel);
-+err_alloc_ch:
-+ if (cpumask_empty(&priv->dpio_cpumask)) {
-+ dev_err(dev, "No cpu with an affine DPIO/DPCON\n");
-+ return err;
-+ }
-+
-+ dev_info(dev, "Cores %*pbl available for processing ingress traffic\n",
-+ cpumask_pr_args(&priv->dpio_cpumask));
-+
-+ return 0;
-+}
-+
-+static void free_dpio(struct dpaa2_eth_priv *priv)
-+{
-+ int i;
-+ struct dpaa2_eth_channel *ch;
-+
-+ /* deregister CDAN notifications and free channels */
-+ for (i = 0; i < priv->num_channels; i++) {
-+ ch = priv->channel[i];
-+ dpaa2_io_service_deregister(ch->dpio, &ch->nctx);
-+ free_channel(priv, ch);
-+ }
-+}
-+
-+static struct dpaa2_eth_channel *get_affine_channel(struct dpaa2_eth_priv *priv,
-+ int cpu)
-+{
-+ struct device *dev = priv->net_dev->dev.parent;
-+ int i;
-+
-+ for (i = 0; i < priv->num_channels; i++)
-+ if (priv->channel[i]->nctx.desired_cpu == cpu)
-+ return priv->channel[i];
-+
-+ /* We should never get here. Issue a warning and return
-+ * the first channel, because it's still better than nothing
-+ */
-+ dev_warn(dev, "No affine channel found for cpu %d\n", cpu);
-+
-+ return priv->channel[0];
-+}
-+
-+static void set_fq_affinity(struct dpaa2_eth_priv *priv)
-+{
-+ struct device *dev = priv->net_dev->dev.parent;
-+ struct cpumask xps_mask;
-+ struct dpaa2_eth_fq *fq;
-+ int rx_cpu, txc_cpu;
-+ int i, err;
-+
-+ /* For each FQ, pick one channel/CPU to deliver frames to.
-+ * This may well change at runtime, either through irqbalance or
-+ * through direct user intervention.
-+ */
-+ rx_cpu = txc_cpu = cpumask_first(&priv->dpio_cpumask);
-+
-+ for (i = 0; i < priv->num_fqs; i++) {
-+ fq = &priv->fq[i];
-+ switch (fq->type) {
-+ case DPAA2_RX_FQ:
-+ case DPAA2_RX_ERR_FQ:
-+ fq->target_cpu = rx_cpu;
-+ rx_cpu = cpumask_next(rx_cpu, &priv->dpio_cpumask);
-+ if (rx_cpu >= nr_cpu_ids)
-+ rx_cpu = cpumask_first(&priv->dpio_cpumask);
-+ break;
-+ case DPAA2_TX_CONF_FQ:
-+ fq->target_cpu = txc_cpu;
-+
-+ /* Tell the stack to affine to txc_cpu the Tx queue
-+ * associated with the confirmation one
-+ */
-+ cpumask_clear(&xps_mask);
-+ cpumask_set_cpu(txc_cpu, &xps_mask);
-+ err = netif_set_xps_queue(priv->net_dev, &xps_mask,
-+ fq->flowid);
-+ if (err)
-+ dev_info_once(dev, "Error setting XPS queue\n");
-+
-+ txc_cpu = cpumask_next(txc_cpu, &priv->dpio_cpumask);
-+ if (txc_cpu >= nr_cpu_ids)
-+ txc_cpu = cpumask_first(&priv->dpio_cpumask);
-+ break;
-+ default:
-+ dev_err(dev, "Unknown FQ type: %d\n", fq->type);
-+ }
-+ fq->channel = get_affine_channel(priv, fq->target_cpu);
-+ }
-+}
-+
-+static void setup_fqs(struct dpaa2_eth_priv *priv)
-+{
-+ int i, j;
-+
-+ /* We have one TxConf FQ per Tx flow.
-+ * The number of Tx and Rx queues is the same.
-+ * Tx queues come first in the fq array.
-+ */
-+ for (i = 0; i < dpaa2_eth_queue_count(priv); i++) {
-+ priv->fq[priv->num_fqs].type = DPAA2_TX_CONF_FQ;
-+ priv->fq[priv->num_fqs].consume = dpaa2_eth_tx_conf;
-+ priv->fq[priv->num_fqs++].flowid = (u16)i;
-+ }
-+
-+ for (i = 0; i < dpaa2_eth_tc_count(priv); i++)
-+ for (j = 0; j < dpaa2_eth_queue_count(priv); j++) {
-+ priv->fq[priv->num_fqs].type = DPAA2_RX_FQ;
-+ priv->fq[priv->num_fqs].consume = dpaa2_eth_rx;
-+ priv->fq[priv->num_fqs].tc = (u8)i;
-+ priv->fq[priv->num_fqs++].flowid = (u16)j;
-+ }
-+
-+#ifdef CONFIG_FSL_DPAA2_ETH_USE_ERR_QUEUE
-+ /* We have exactly one Rx error queue per DPNI */
-+ priv->fq[priv->num_fqs].type = DPAA2_RX_ERR_FQ;
-+ priv->fq[priv->num_fqs++].consume = dpaa2_eth_rx_err;
-+#endif
-+
-+ /* For each FQ, decide on which core to process incoming frames */
-+ set_fq_affinity(priv);
-+}
-+
-+/* Allocate and configure one buffer pool for each interface */
-+static int setup_dpbp(struct dpaa2_eth_priv *priv)
-+{
-+ int err;
-+ struct fsl_mc_device *dpbp_dev;
-+ struct device *dev = priv->net_dev->dev.parent;
-+ struct dpbp_attr dpbp_attrs;
-+
-+ err = fsl_mc_object_allocate(to_fsl_mc_device(dev), FSL_MC_POOL_DPBP,
-+ &dpbp_dev);
-+ if (err) {
-+ dev_err(dev, "DPBP device allocation failed\n");
-+ return err;
-+ }
-+
-+ priv->dpbp_dev = dpbp_dev;
-+
-+ err = dpbp_open(priv->mc_io, 0, priv->dpbp_dev->obj_desc.id,
-+ &dpbp_dev->mc_handle);
-+ if (err) {
-+ dev_err(dev, "dpbp_open() failed\n");
-+ goto err_open;
-+ }
-+
-+ err = dpbp_reset(priv->mc_io, 0, dpbp_dev->mc_handle);
-+ if (err) {
-+ dev_err(dev, "dpbp_reset() failed\n");
-+ goto err_reset;
-+ }
-+
-+ err = dpbp_enable(priv->mc_io, 0, dpbp_dev->mc_handle);
-+ if (err) {
-+ dev_err(dev, "dpbp_enable() failed\n");
-+ goto err_enable;
-+ }
-+
-+ err = dpbp_get_attributes(priv->mc_io, 0, dpbp_dev->mc_handle,
-+ &dpbp_attrs);
-+ if (err) {
-+ dev_err(dev, "dpbp_get_attributes() failed\n");
-+ goto err_get_attr;
-+ }
-+ priv->bpid = dpbp_attrs.bpid;
-+
-+ /* By default we start with flow control enabled */
-+ priv->max_bufs_per_ch = DPAA2_ETH_NUM_BUFS_FC / priv->num_channels;
-+
-+ return 0;
-+
-+err_get_attr:
-+ dpbp_disable(priv->mc_io, 0, dpbp_dev->mc_handle);
-+err_enable:
-+err_reset:
-+ dpbp_close(priv->mc_io, 0, dpbp_dev->mc_handle);
-+err_open:
-+ fsl_mc_object_free(dpbp_dev);
-+
-+ return err;
-+}
-+
-+static void free_dpbp(struct dpaa2_eth_priv *priv)
-+{
-+ drain_pool(priv);
-+ dpbp_disable(priv->mc_io, 0, priv->dpbp_dev->mc_handle);
-+ dpbp_close(priv->mc_io, 0, priv->dpbp_dev->mc_handle);
-+ fsl_mc_object_free(priv->dpbp_dev);
-+}
-+
-+static int setup_tx_congestion(struct dpaa2_eth_priv *priv)
-+{
-+ struct dpni_congestion_notification_cfg notif_cfg = {0};
-+ struct device *dev = priv->net_dev->dev.parent;
-+ int err;
-+
-+ priv->cscn_unaligned = kzalloc(DPAA2_CSCN_SIZE + DPAA2_CSCN_ALIGN,
-+ GFP_KERNEL);
-+
-+ if (!priv->cscn_unaligned)
-+ return -ENOMEM;
-+
-+ priv->cscn_mem = PTR_ALIGN(priv->cscn_unaligned, DPAA2_CSCN_ALIGN);
-+ priv->cscn_dma = dma_map_single(dev, priv->cscn_mem, DPAA2_CSCN_SIZE,
-+ DMA_FROM_DEVICE);
-+ if (dma_mapping_error(dev, priv->cscn_dma)) {
-+ dev_err(dev, "Error mapping CSCN memory area\n");
-+ err = -ENOMEM;
-+ goto err_dma_map;
-+ }
-+
-+ notif_cfg.units = DPNI_CONGESTION_UNIT_BYTES;
-+ notif_cfg.threshold_entry = DPAA2_ETH_TX_CONG_ENTRY_THRESH;
-+ notif_cfg.threshold_exit = DPAA2_ETH_TX_CONG_EXIT_THRESH;
-+ notif_cfg.message_ctx = (u64)priv;
-+ notif_cfg.message_iova = priv->cscn_dma;
-+ notif_cfg.notification_mode = DPNI_CONG_OPT_WRITE_MEM_ON_ENTER |
-+ DPNI_CONG_OPT_WRITE_MEM_ON_EXIT |
-+ DPNI_CONG_OPT_COHERENT_WRITE;
-+ err = dpni_set_congestion_notification(priv->mc_io, 0, priv->mc_token,
-+ DPNI_QUEUE_TX, 0, &notif_cfg);
-+ if (err) {
-+ dev_err(dev, "dpni_set_congestion_notification failed\n");
-+ goto err_set_cong;
-+ }
-+
-+ return 0;
-+
-+err_set_cong:
-+ dma_unmap_single(dev, priv->cscn_dma, DPAA2_CSCN_SIZE, DMA_FROM_DEVICE);
-+err_dma_map:
-+ kfree(priv->cscn_unaligned);
-+
-+ return err;
-+}
-+
-+/* Configure the DPNI object this interface is associated with */
-+static int setup_dpni(struct fsl_mc_device *ls_dev)
-+{
-+ struct device *dev = &ls_dev->dev;
-+ struct dpaa2_eth_priv *priv;
-+ struct net_device *net_dev;
-+ struct dpni_link_cfg cfg = {0};
-+ int err;
-+
-+ net_dev = dev_get_drvdata(dev);
-+ priv = netdev_priv(net_dev);
-+
-+ /* get a handle for the DPNI object */
-+ err = dpni_open(priv->mc_io, 0, ls_dev->obj_desc.id, &priv->mc_token);
-+ if (err) {
-+ dev_err(dev, "dpni_open() failed\n");
-+ return err;
-+ }
-+
-+ /* Check if we can work with this DPNI object */
-+ err = dpni_get_api_version(priv->mc_io, 0, &priv->dpni_ver_major,
-+ &priv->dpni_ver_minor);
-+ if (err) {
-+ dev_err(dev, "dpni_get_api_version() failed\n");
-+ goto close;
-+ }
-+ if (dpaa2_eth_cmp_dpni_ver(priv, DPNI_VER_MAJOR, DPNI_VER_MINOR) < 0) {
-+ dev_err(dev, "DPNI version %u.%u not supported, need >= %u.%u\n",
-+ priv->dpni_ver_major, priv->dpni_ver_minor,
-+ DPNI_VER_MAJOR, DPNI_VER_MINOR);
-+ err = -ENOTSUPP;
-+ goto close;
-+ }
-+
-+ ls_dev->mc_io = priv->mc_io;
-+ ls_dev->mc_handle = priv->mc_token;
-+
-+ err = dpni_reset(priv->mc_io, 0, priv->mc_token);
-+ if (err) {
-+ dev_err(dev, "dpni_reset() failed\n");
-+ goto close;
-+ }
-+
-+ err = dpni_get_attributes(priv->mc_io, 0, priv->mc_token,
-+ &priv->dpni_attrs);
-+ if (err) {
-+ dev_err(dev, "dpni_get_attributes() failed (err=%d)\n", err);
-+ goto close;
-+ }
-+
-+ err = set_buffer_layout(priv);
-+ if (err)
-+ goto close;
-+
-+ /* Enable congestion notifications for Tx queues */
-+ err = setup_tx_congestion(priv);
-+ if (err)
-+ goto close;
-+
-+ /* allocate classification rule space */
-+ priv->cls_rule = kzalloc(sizeof(*priv->cls_rule) *
-+ dpaa2_eth_fs_count(priv), GFP_KERNEL);
-+ if (!priv->cls_rule)
-+ goto close;
-+
-+ /* Enable flow control */
-+ cfg.options = DPNI_LINK_OPT_AUTONEG | DPNI_LINK_OPT_PAUSE;
-+ priv->tx_pause_frames = true;
-+ err = dpni_set_link_cfg(priv->mc_io, 0, priv->mc_token, &cfg);
-+ if (err) {
-+ dev_err(dev, "dpni_set_link_cfg() failed\n");
-+ goto cls_free;
-+ }
-+
-+ return 0;
-+
-+cls_free:
-+ kfree(priv->cls_rule);
-+close:
-+ dpni_close(priv->mc_io, 0, priv->mc_token);
-+
-+ return err;
-+}
-+
-+static void free_dpni(struct dpaa2_eth_priv *priv)
-+{
-+ struct device *dev = priv->net_dev->dev.parent;
-+ int err;
-+
-+ err = dpni_reset(priv->mc_io, 0, priv->mc_token);
-+ if (err)
-+ netdev_warn(priv->net_dev, "dpni_reset() failed (err %d)\n",
-+ err);
-+
-+ dpni_close(priv->mc_io, 0, priv->mc_token);
-+
-+ kfree(priv->cls_rule);
-+
-+ dma_unmap_single(dev, priv->cscn_dma, DPAA2_CSCN_SIZE, DMA_FROM_DEVICE);
-+ kfree(priv->cscn_unaligned);
-+}
-+
-+static int setup_rx_flow(struct dpaa2_eth_priv *priv,
-+ struct dpaa2_eth_fq *fq)
-+{
-+ struct device *dev = priv->net_dev->dev.parent;
-+ struct dpni_queue queue;
-+ struct dpni_queue_id qid;
-+ int err;
-+
-+ err = dpni_get_queue(priv->mc_io, 0, priv->mc_token,
-+ DPNI_QUEUE_RX, fq->tc, fq->flowid, &queue, &qid);
-+ if (err) {
-+ dev_err(dev, "dpni_get_queue(RX) failed\n");
-+ return err;
-+ }
-+
-+ fq->fqid = qid.fqid;
-+
-+ queue.destination.id = fq->channel->dpcon_id;
-+ queue.destination.type = DPNI_DEST_DPCON;
-+ queue.destination.priority = 1;
-+ queue.user_context = (u64)(uintptr_t)fq;
-+ err = dpni_set_queue(priv->mc_io, 0, priv->mc_token,
-+ DPNI_QUEUE_RX, fq->tc, fq->flowid,
-+ DPNI_QUEUE_OPT_USER_CTX | DPNI_QUEUE_OPT_DEST,
-+ &queue);
-+ if (err) {
-+ dev_err(dev, "dpni_set_queue(RX) failed\n");
-+ return err;
-+ }
-+
-+ return 0;
-+}
-+
-+static int set_queue_taildrop(struct dpaa2_eth_priv *priv,
-+ struct dpni_taildrop *td)
-+{
-+ struct device *dev = priv->net_dev->dev.parent;
-+ int i, err;
-+
-+ for (i = 0; i < priv->num_fqs; i++) {
-+ if (priv->fq[i].type != DPAA2_RX_FQ)
-+ continue;
-+
-+ err = dpni_set_taildrop(priv->mc_io, 0, priv->mc_token,
-+ DPNI_CP_QUEUE, DPNI_QUEUE_RX,
-+ priv->fq[i].tc, priv->fq[i].flowid,
-+ td);
-+ if (err) {
-+ dev_err(dev, "dpni_set_taildrop() failed (%d)\n", err);
-+ return err;
-+ }
-+
-+ dev_dbg(dev, "%s taildrop for Rx group tc %d\n",
-+ (td->enable ? "Enabled" : "Disabled"),
-+ i);
-+ }
-+
-+ return 0;
-+}
-+
-+static int set_group_taildrop(struct dpaa2_eth_priv *priv,
-+ struct dpni_taildrop *td)
-+{
-+ struct device *dev = priv->net_dev->dev.parent;
-+ struct dpni_taildrop disable_td, *tc_td;
-+ int i, err;
-+
-+ memset(&disable_td, 0, sizeof(struct dpni_taildrop));
-+ for (i = 0; i < dpaa2_eth_tc_count(priv); i++) {
-+ if (td->enable && dpaa2_eth_is_pfc_enabled(priv, i))
-+ /* Do not set taildrop thresholds for PFC-enabled
-+ * traffic classes. We will enable congestion
-+ * notifications for them.
-+ */
-+ tc_td = &disable_td;
-+ else
-+ tc_td = td;
-+
-+ err = dpni_set_taildrop(priv->mc_io, 0, priv->mc_token,
-+ DPNI_CP_GROUP, DPNI_QUEUE_RX,
-+ i, 0, tc_td);
-+ if (err) {
-+ dev_err(dev, "dpni_set_taildrop() failed (%d)\n", err);
-+ return err;
-+ }
-+
-+ dev_dbg(dev, "%s taildrop for Rx queue id %d tc %d\n",
-+ (tc_td->enable ? "Enabled" : "Disabled"),
-+ priv->fq[i].flowid, priv->fq[i].tc);
-+ }
-+
-+ return 0;
-+}
-+
-+/* Enable/disable Rx FQ taildrop
-+ *
-+ * Rx FQ taildrop is mutually exclusive with flow control and it only gets
-+ * disabled when FC is active. Depending on FC status, we need to compute
-+ * the maximum number of buffers in the pool differently, so use the
-+ * opportunity to update max number of buffers as well.
-+ */
-+int set_rx_taildrop(struct dpaa2_eth_priv *priv)
-+{
-+ enum dpaa2_eth_td_cfg cfg = dpaa2_eth_get_td_type(priv);
-+ struct dpni_taildrop td_queue, td_group;
-+ int err = 0;
-+
-+ switch (cfg) {
-+ case DPAA2_ETH_TD_NONE:
-+ memset(&td_queue, 0, sizeof(struct dpni_taildrop));
-+ memset(&td_group, 0, sizeof(struct dpni_taildrop));
-+ priv->max_bufs_per_ch = DPAA2_ETH_NUM_BUFS_FC /
-+ priv->num_channels;
-+ break;
-+ case DPAA2_ETH_TD_QUEUE:
-+ memset(&td_group, 0, sizeof(struct dpni_taildrop));
-+ td_queue.enable = 1;
-+ td_queue.units = DPNI_CONGESTION_UNIT_BYTES;
-+ td_queue.threshold = DPAA2_ETH_TAILDROP_THRESH /
-+ dpaa2_eth_tc_count(priv);
-+ priv->max_bufs_per_ch = DPAA2_ETH_NUM_BUFS_PER_CH;
-+ break;
-+ case DPAA2_ETH_TD_GROUP:
-+ memset(&td_queue, 0, sizeof(struct dpni_taildrop));
-+ td_group.enable = 1;
-+ td_group.units = DPNI_CONGESTION_UNIT_FRAMES;
-+ td_group.threshold = NAPI_POLL_WEIGHT *
-+ dpaa2_eth_queue_count(priv);
-+ priv->max_bufs_per_ch = NAPI_POLL_WEIGHT *
-+ dpaa2_eth_tc_count(priv);
-+ break;
-+ default:
-+ break;
-+ }
-+
-+ err = set_queue_taildrop(priv, &td_queue);
-+ if (err)
-+ return err;
-+
-+ err = set_group_taildrop(priv, &td_group);
-+ if (err)
-+ return err;
-+
-+ priv->refill_thresh = DPAA2_ETH_REFILL_THRESH(priv);
-+
-+ return 0;
-+}
-+
-+static int setup_tx_flow(struct dpaa2_eth_priv *priv,
-+ struct dpaa2_eth_fq *fq)
-+{
-+ struct device *dev = priv->net_dev->dev.parent;
-+ struct dpni_queue queue;
-+ struct dpni_queue_id qid;
-+ int err;
-+
-+ err = dpni_get_queue(priv->mc_io, 0, priv->mc_token,
-+ DPNI_QUEUE_TX, 0, fq->flowid, &queue, &qid);
-+ if (err) {
-+ dev_err(dev, "dpni_get_queue(TX) failed\n");
-+ return err;
-+ }
-+
-+ fq->tx_qdbin = qid.qdbin;
-+
-+ err = dpni_get_queue(priv->mc_io, 0, priv->mc_token,
-+ DPNI_QUEUE_TX_CONFIRM, 0, fq->flowid,
-+ &queue, &qid);
-+ if (err) {
-+ dev_err(dev, "dpni_get_queue(TX_CONF) failed\n");
-+ return err;
-+ }
-+
-+ fq->fqid = qid.fqid;
-+
-+ queue.destination.id = fq->channel->dpcon_id;
-+ queue.destination.type = DPNI_DEST_DPCON;
-+ queue.destination.priority = 0;
-+ queue.user_context = (u64)(uintptr_t)fq;
-+ err = dpni_set_queue(priv->mc_io, 0, priv->mc_token,
-+ DPNI_QUEUE_TX_CONFIRM, 0, fq->flowid,
-+ DPNI_QUEUE_OPT_USER_CTX | DPNI_QUEUE_OPT_DEST,
-+ &queue);
-+ if (err) {
-+ dev_err(dev, "dpni_set_queue(TX_CONF) failed\n");
-+ return err;
-+ }
-+
-+ return 0;
-+}
-+
-+#ifdef CONFIG_FSL_DPAA2_ETH_USE_ERR_QUEUE
-+static int setup_rx_err_flow(struct dpaa2_eth_priv *priv,
-+ struct dpaa2_eth_fq *fq)
-+{
-+ struct device *dev = priv->net_dev->dev.parent;
-+ struct dpni_queue q = { { 0 } };
-+ struct dpni_queue_id qid;
-+ u8 q_opt = DPNI_QUEUE_OPT_USER_CTX | DPNI_QUEUE_OPT_DEST;
-+ int err;
-+
-+ err = dpni_get_queue(priv->mc_io, 0, priv->mc_token,
-+ DPNI_QUEUE_RX_ERR, 0, 0, &q, &qid);
-+ if (err) {
-+ dev_err(dev, "dpni_get_queue() failed (%d)\n", err);
-+ return err;
-+ }
-+
-+ fq->fqid = qid.fqid;
-+
-+ q.destination.id = fq->channel->dpcon_id;
-+ q.destination.type = DPNI_DEST_DPCON;
-+ q.destination.priority = 1;
-+ q.user_context = (u64)fq;
-+ err = dpni_set_queue(priv->mc_io, 0, priv->mc_token,
-+ DPNI_QUEUE_RX_ERR, 0, 0, q_opt, &q);
-+ if (err) {
-+ dev_err(dev, "dpni_set_queue() failed (%d)\n", err);
-+ return err;
-+ }
-+
-+ return 0;
-+}
-+#endif
-+
-+/* default hash key fields */
-+static struct dpaa2_eth_dist_fields default_dist_fields[] = {
-+ {
-+ /* L2 header */
-+ .rxnfc_field = RXH_L2DA,
-+ .cls_prot = NET_PROT_ETH,
-+ .cls_field = NH_FLD_ETH_DA,
-+ .id = DPAA2_ETH_DIST_ETHDST,
-+ .size = 6,
-+ }, {
-+ .cls_prot = NET_PROT_ETH,
-+ .cls_field = NH_FLD_ETH_SA,
-+ .id = DPAA2_ETH_DIST_ETHSRC,
-+ .size = 6,
-+ }, {
-+ /* This is the last ethertype field parsed:
-+ * depending on frame format, it can be the MAC ethertype
-+ * or the VLAN etype.
-+ */
-+ .cls_prot = NET_PROT_ETH,
-+ .cls_field = NH_FLD_ETH_TYPE,
-+ .id = DPAA2_ETH_DIST_ETHTYPE,
-+ .size = 2,
-+ }, {
-+ /* VLAN header */
-+ .rxnfc_field = RXH_VLAN,
-+ .cls_prot = NET_PROT_VLAN,
-+ .cls_field = NH_FLD_VLAN_TCI,
-+ .id = DPAA2_ETH_DIST_VLAN,
-+ .size = 2,
-+ }, {
-+ /* IP header */
-+ .rxnfc_field = RXH_IP_SRC,
-+ .cls_prot = NET_PROT_IP,
-+ .cls_field = NH_FLD_IP_SRC,
-+ .id = DPAA2_ETH_DIST_IPSRC,
-+ .size = 4,
-+ }, {
-+ .rxnfc_field = RXH_IP_DST,
-+ .cls_prot = NET_PROT_IP,
-+ .cls_field = NH_FLD_IP_DST,
-+ .id = DPAA2_ETH_DIST_IPDST,
-+ .size = 4,
-+ }, {
-+ .rxnfc_field = RXH_L3_PROTO,
-+ .cls_prot = NET_PROT_IP,
-+ .cls_field = NH_FLD_IP_PROTO,
-+ .id = DPAA2_ETH_DIST_IPPROTO,
-+ .size = 1,
-+ }, {
-+ /* Using UDP ports, this is functionally equivalent to raw
-+ * byte pairs from L4 header.
-+ */
-+ .rxnfc_field = RXH_L4_B_0_1,
-+ .cls_prot = NET_PROT_UDP,
-+ .cls_field = NH_FLD_UDP_PORT_SRC,
-+ .id = DPAA2_ETH_DIST_L4SRC,
-+ .size = 2,
-+ }, {
-+ .rxnfc_field = RXH_L4_B_2_3,
-+ .cls_prot = NET_PROT_UDP,
-+ .cls_field = NH_FLD_UDP_PORT_DST,
-+ .id = DPAA2_ETH_DIST_L4DST,
-+ .size = 2,
-+ },
-+};
-+
-+static int legacy_config_dist_key(struct dpaa2_eth_priv *priv,
-+ dma_addr_t key_iova)
-+{
-+ struct device *dev = priv->net_dev->dev.parent;
-+ struct dpni_rx_tc_dist_cfg dist_cfg;
-+ int i, err;
-+
-+ /* In legacy mode, we can't configure flow steering independently */
-+ if (!dpaa2_eth_hash_enabled(priv))
-+ return -EOPNOTSUPP;
-+
-+ memset(&dist_cfg, 0, sizeof(dist_cfg));
-+
-+ dist_cfg.key_cfg_iova = key_iova;
-+ dist_cfg.dist_size = dpaa2_eth_queue_count(priv);
-+ if (dpaa2_eth_fs_enabled(priv)) {
-+ dist_cfg.dist_mode = DPNI_DIST_MODE_FS;
-+ dist_cfg.fs_cfg.miss_action = DPNI_FS_MISS_HASH;
-+ } else {
-+ dist_cfg.dist_mode = DPNI_DIST_MODE_HASH;
-+ }
-+
-+ for (i = 0; i < dpaa2_eth_tc_count(priv); i++) {
-+ err = dpni_set_rx_tc_dist(priv->mc_io, 0, priv->mc_token, i,
-+ &dist_cfg);
-+ if (err) {
-+ dev_err(dev, "dpni_set_rx_tc_dist failed\n");
-+ return err;
-+ }
-+ }
-+
-+ return 0;
-+}
-+
-+static int config_hash_key(struct dpaa2_eth_priv *priv, dma_addr_t key_iova)
-+{
-+ struct device *dev = priv->net_dev->dev.parent;
-+ struct dpni_rx_dist_cfg dist_cfg;
-+ int i, err;
-+
-+ if (!dpaa2_eth_hash_enabled(priv))
-+ return -EOPNOTSUPP;
-+
-+ memset(&dist_cfg, 0, sizeof(dist_cfg));
-+
-+ dist_cfg.key_cfg_iova = key_iova;
-+ dist_cfg.dist_size = dpaa2_eth_queue_count(priv);
-+ dist_cfg.enable = true;
-+
-+ for (i = 0; i < dpaa2_eth_tc_count(priv); i++) {
-+ dist_cfg.tc = i;
-+
-+ err = dpni_set_rx_hash_dist(priv->mc_io, 0,
-+ priv->mc_token, &dist_cfg);
-+ if (err) {
-+ dev_err(dev, "dpni_set_rx_hash_dist failed\n");
-+ return err;
-+ }
-+ }
-+
-+ return 0;
-+}
-+
-+static int config_fs_key(struct dpaa2_eth_priv *priv, dma_addr_t key_iova)
-+{
-+ struct device *dev = priv->net_dev->dev.parent;
-+ struct dpni_rx_dist_cfg dist_cfg;
-+ int i, err;
-+
-+ if (!dpaa2_eth_fs_enabled(priv))
-+ return -EOPNOTSUPP;
-+
-+ memset(&dist_cfg, 0, sizeof(dist_cfg));
-+
-+ dist_cfg.key_cfg_iova = key_iova;
-+ dist_cfg.dist_size = dpaa2_eth_queue_count(priv);
-+ dist_cfg.enable = true;
-+
-+ for (i = 0; i < dpaa2_eth_tc_count(priv); i++) {
-+ dist_cfg.tc = i;
-+
-+ err = dpni_set_rx_fs_dist(priv->mc_io, 0,
-+ priv->mc_token, &dist_cfg);
-+ if (err) {
-+ dev_err(dev, "dpni_set_rx_fs_dist failed\n");
-+ return err;
-+ }
-+ }
-+
-+ return 0;
-+}
-+
-+int dpaa2_eth_set_dist_key(struct dpaa2_eth_priv *priv,
-+ enum dpaa2_eth_rx_dist type, u32 key_fields)
-+{
-+ struct device *dev = priv->net_dev->dev.parent;
-+ struct dpkg_profile_cfg cls_cfg;
-+ struct dpkg_extract *key;
-+ u32 hash_fields = 0;
-+ dma_addr_t key_iova;
-+ u8 *key_mem;
-+ int i, err;
-+
-+ memset(&cls_cfg, 0, sizeof(cls_cfg));
-+
-+ for (i = 0; i < priv->num_dist_fields; i++) {
-+ if (!(key_fields & priv->dist_fields[i].id))
-+ continue;
-+
-+ key = &cls_cfg.extracts[cls_cfg.num_extracts];
-+ key->type = DPKG_EXTRACT_FROM_HDR;
-+ key->extract.from_hdr.prot = priv->dist_fields[i].cls_prot;
-+ key->extract.from_hdr.type = DPKG_FULL_FIELD;
-+ key->extract.from_hdr.field = priv->dist_fields[i].cls_field;
-+ cls_cfg.num_extracts++;
-+
-+ hash_fields |= priv->dist_fields[i].rxnfc_field;
-+ }
-+
-+ key_mem = kzalloc(DPAA2_CLASSIFIER_DMA_SIZE, GFP_KERNEL);
-+ if (!key_mem)
-+ return -ENOMEM;
-+
-+ err = dpni_prepare_key_cfg(&cls_cfg, key_mem);
-+ if (err) {
-+ dev_err(dev, "dpni_prepare_key_cfg error %d\n", err);
-+ goto free_key;
-+ }
-+
-+ key_iova = dma_map_single(dev, key_mem, DPAA2_CLASSIFIER_DMA_SIZE,
-+ DMA_TO_DEVICE);
-+ if (dma_mapping_error(dev, key_iova)) {
-+ dev_err(dev, "DMA mapping failed\n");
-+ err = -ENOMEM;
-+ goto free_key;
-+ }
-+
-+ switch (type) {
-+ case DPAA2_ETH_RX_DIST_LEGACY:
-+ err = legacy_config_dist_key(priv, key_iova);
-+ break;
-+ case DPAA2_ETH_RX_DIST_HASH:
-+ err = config_hash_key(priv, key_iova);
-+ break;
-+ case DPAA2_ETH_RX_DIST_FS:
-+ err = config_fs_key(priv, key_iova);
-+ break;
-+ default:
-+ err = -EINVAL;
-+ break;
-+ }
-+
-+ dma_unmap_single(dev, key_iova, DPAA2_CLASSIFIER_DMA_SIZE,
-+ DMA_TO_DEVICE);
-+ if (err) {
-+ if (err != -EOPNOTSUPP)
-+ dev_err(dev, "Distribution key config failed\n");
-+ goto free_key;
-+ }
-+
-+ if (type != DPAA2_ETH_RX_DIST_FS)
-+ priv->rx_hash_fields = hash_fields;
-+
-+free_key:
-+ kfree(key_mem);
-+ return err;
-+}
-+
-+/* Bind the DPNI to its needed objects and resources: buffer pool, DPIOs,
-+ * frame queues and channels
-+ */
-+static int bind_dpni(struct dpaa2_eth_priv *priv)
-+{
-+ struct net_device *net_dev = priv->net_dev;
-+ struct device *dev = net_dev->dev.parent;
-+ struct dpni_pools_cfg pools_params;
-+ struct dpni_error_cfg err_cfg;
-+ int err = 0;
-+ int i;
-+
-+ pools_params.num_dpbp = 1;
-+ pools_params.pools[0].dpbp_id = priv->dpbp_dev->obj_desc.id;
-+ pools_params.pools[0].backup_pool = 0;
-+ pools_params.pools[0].priority_mask = 0xff;
-+ pools_params.pools[0].buffer_size = DPAA2_ETH_RX_BUF_SIZE;
-+ err = dpni_set_pools(priv->mc_io, 0, priv->mc_token, &pools_params);
-+ if (err) {
-+ dev_err(dev, "dpni_set_pools() failed\n");
-+ return err;
-+ }
-+
-+ /* Verify classification options and disable hashing and/or
-+ * flow steering support in case of invalid configuration values
-+ */
-+ priv->dist_fields = default_dist_fields;
-+ priv->num_dist_fields = ARRAY_SIZE(default_dist_fields);
-+ check_cls_support(priv);
-+
-+ /* have the interface implicitly distribute traffic based on
-+ * a static hash key. Also configure flow steering key, if supported.
-+ * Errors here are not blocking, so just let the called function
-+ * print its error message and move along.
-+ */
-+ if (dpaa2_eth_has_legacy_dist(priv)) {
-+ dpaa2_eth_set_dist_key(priv, DPAA2_ETH_RX_DIST_LEGACY,
-+ DPAA2_ETH_DIST_ALL);
-+ } else {
-+ dpaa2_eth_set_dist_key(priv, DPAA2_ETH_RX_DIST_HASH,
-+ DPAA2_ETH_DIST_DEFAULT_HASH);
-+ dpaa2_eth_set_dist_key(priv, DPAA2_ETH_RX_DIST_FS,
-+ DPAA2_ETH_DIST_ALL);
-+ }
-+
-+ /* Configure handling of error frames */
-+ err_cfg.errors = DPAA2_FAS_RX_ERR_MASK;
-+ err_cfg.set_frame_annotation = 1;
-+#ifdef CONFIG_FSL_DPAA2_ETH_USE_ERR_QUEUE
-+ err_cfg.error_action = DPNI_ERROR_ACTION_SEND_TO_ERROR_QUEUE;
-+#else
-+ err_cfg.error_action = DPNI_ERROR_ACTION_DISCARD;
-+#endif
-+ err = dpni_set_errors_behavior(priv->mc_io, 0, priv->mc_token,
-+ &err_cfg);
-+ if (err) {
-+ dev_err(dev, "dpni_set_errors_behavior failed\n");
-+ return err;
-+ }
-+
-+ /* Configure Rx and Tx conf queues to generate CDANs */
-+ for (i = 0; i < priv->num_fqs; i++) {
-+ switch (priv->fq[i].type) {
-+ case DPAA2_RX_FQ:
-+ err = setup_rx_flow(priv, &priv->fq[i]);
-+ break;
-+ case DPAA2_TX_CONF_FQ:
-+ err = setup_tx_flow(priv, &priv->fq[i]);
-+ break;
-+#ifdef CONFIG_FSL_DPAA2_ETH_USE_ERR_QUEUE
-+ case DPAA2_RX_ERR_FQ:
-+ err = setup_rx_err_flow(priv, &priv->fq[i]);
-+ break;
-+#endif
-+ default:
-+ dev_err(dev, "Invalid FQ type %d\n", priv->fq[i].type);
-+ return -EINVAL;
-+ }
-+ if (err)
-+ return err;
-+ }
-+
-+ err = dpni_get_qdid(priv->mc_io, 0, priv->mc_token,
-+ DPNI_QUEUE_TX, &priv->tx_qdid);
-+ if (err) {
-+ dev_err(dev, "dpni_get_qdid() failed\n");
-+ return err;
-+ }
-+
-+ return 0;
-+}
-+
-+/* Allocate rings for storing incoming frame descriptors */
-+static int alloc_rings(struct dpaa2_eth_priv *priv)
-+{
-+ struct net_device *net_dev = priv->net_dev;
-+ struct device *dev = net_dev->dev.parent;
-+ int i;
-+
-+ for (i = 0; i < priv->num_channels; i++) {
-+ priv->channel[i]->store =
-+ dpaa2_io_store_create(DPAA2_ETH_STORE_SIZE, dev);
-+ if (!priv->channel[i]->store) {
-+ netdev_err(net_dev, "dpaa2_io_store_create() failed\n");
-+ goto err_ring;
-+ }
-+ }
-+
-+ return 0;
-+
-+err_ring:
-+ for (i = 0; i < priv->num_channels; i++) {
-+ if (!priv->channel[i]->store)
-+ break;
-+ dpaa2_io_store_destroy(priv->channel[i]->store);
-+ }
-+
-+ return -ENOMEM;
-+}
-+
-+static void free_rings(struct dpaa2_eth_priv *priv)
-+{
-+ int i;
-+
-+ for (i = 0; i < priv->num_channels; i++)
-+ dpaa2_io_store_destroy(priv->channel[i]->store);
-+}
-+
-+static int set_mac_addr(struct dpaa2_eth_priv *priv)
-+{
-+ struct net_device *net_dev = priv->net_dev;
-+ struct device *dev = net_dev->dev.parent;
-+ u8 mac_addr[ETH_ALEN], dpni_mac_addr[ETH_ALEN];
-+ int err;
-+
-+ /* Get firmware address, if any */
-+ err = dpni_get_port_mac_addr(priv->mc_io, 0, priv->mc_token, mac_addr);
-+ if (err) {
-+ dev_err(dev, "dpni_get_port_mac_addr() failed\n");
-+ return err;
-+ }
-+
-+ /* Get DPNI attributes address, if any */
-+ err = dpni_get_primary_mac_addr(priv->mc_io, 0, priv->mc_token,
-+ dpni_mac_addr);
-+ if (err) {
-+ dev_err(dev, "dpni_get_primary_mac_addr() failed\n");
-+ return err;
-+ }
-+
-+ /* First check if firmware has any address configured by bootloader */
-+ if (!is_zero_ether_addr(mac_addr)) {
-+ /* If the DPMAC addr != DPNI addr, update it */
-+ if (!ether_addr_equal(mac_addr, dpni_mac_addr)) {
-+ err = dpni_set_primary_mac_addr(priv->mc_io, 0,
-+ priv->mc_token,
-+ mac_addr);
-+ if (err) {
-+ dev_err(dev, "dpni_set_primary_mac_addr() failed\n");
-+ return err;
-+ }
-+ }
-+ memcpy(net_dev->dev_addr, mac_addr, net_dev->addr_len);
-+ } else if (is_zero_ether_addr(dpni_mac_addr)) {
-+ /* No MAC address configured, fill in net_dev->dev_addr
-+ * with a random one
-+ */
-+ eth_hw_addr_random(net_dev);
-+ dev_dbg_once(dev, "device(s) have all-zero hwaddr, replaced with random\n");
-+
-+ err = dpni_set_primary_mac_addr(priv->mc_io, 0, priv->mc_token,
-+ net_dev->dev_addr);
-+ if (err) {
-+ dev_err(dev, "dpni_set_primary_mac_addr() failed\n");
-+ return err;
-+ }
-+
-+ /* Override NET_ADDR_RANDOM set by eth_hw_addr_random(); for all
-+ * practical purposes, this will be our "permanent" mac address,
-+ * at least until the next reboot. This move will also permit
-+ * register_netdevice() to properly fill up net_dev->perm_addr.
-+ */
-+ net_dev->addr_assign_type = NET_ADDR_PERM;
-+ } else {
-+ /* NET_ADDR_PERM is default, all we have to do is
-+ * fill in the device addr.
-+ */
-+ memcpy(net_dev->dev_addr, dpni_mac_addr, net_dev->addr_len);
-+ }
-+
-+ return 0;
-+}
-+
-+static int netdev_init(struct net_device *net_dev)
-+{
-+ struct device *dev = net_dev->dev.parent;
-+ struct dpaa2_eth_priv *priv = netdev_priv(net_dev);
-+ u8 bcast_addr[ETH_ALEN];
-+ u8 num_queues;
-+ int err;
-+
-+ net_dev->netdev_ops = &dpaa2_eth_ops;
-+
-+ err = set_mac_addr(priv);
-+ if (err)
-+ return err;
-+
-+ /* Explicitly add the broadcast address to the MAC filtering table */
-+ eth_broadcast_addr(bcast_addr);
-+ err = dpni_add_mac_addr(priv->mc_io, 0, priv->mc_token, bcast_addr);
-+ if (err) {
-+ dev_err(dev, "dpni_add_mac_addr() failed\n");
-+ return err;
-+ }
-+
-+ /* Set MTU upper limit; lower limit is default (68B) */
-+ net_dev->max_mtu = DPAA2_ETH_MAX_MTU;
-+ err = dpni_set_max_frame_length(priv->mc_io, 0, priv->mc_token,
-+ (u16)DPAA2_ETH_MFL);
-+ if (err) {
-+ dev_err(dev, "dpni_set_max_frame_length() failed\n");
-+ return err;
-+ }
-+
-+ /* Set actual number of queues in the net device */
-+ num_queues = dpaa2_eth_queue_count(priv);
-+ err = netif_set_real_num_tx_queues(net_dev, num_queues);
-+ if (err) {
-+ dev_err(dev, "netif_set_real_num_tx_queues() failed\n");
-+ return err;
-+ }
-+ err = netif_set_real_num_rx_queues(net_dev, num_queues);
-+ if (err) {
-+ dev_err(dev, "netif_set_real_num_rx_queues() failed\n");
-+ return err;
-+ }
-+
-+ /* Our .ndo_init will be called herein */
-+ err = register_netdev(net_dev);
-+ if (err < 0) {
-+ dev_err(dev, "register_netdev() failed\n");
-+ return err;
-+ }
-+
-+ return 0;
-+}
-+
-+static int poll_link_state(void *arg)
-+{
-+ struct dpaa2_eth_priv *priv = (struct dpaa2_eth_priv *)arg;
-+ int err;
-+
-+ while (!kthread_should_stop()) {
-+ err = link_state_update(priv);
-+ if (unlikely(err))
-+ return err;
-+
-+ msleep(DPAA2_ETH_LINK_STATE_REFRESH);
-+ }
-+
-+ return 0;
-+}
-+
-+static irqreturn_t dpni_irq0_handler_thread(int irq_num, void *arg)
-+{
-+ u32 status = ~0;
-+ struct device *dev = (struct device *)arg;
-+ struct fsl_mc_device *dpni_dev = to_fsl_mc_device(dev);
-+ struct net_device *net_dev = dev_get_drvdata(dev);
-+ int err;
-+
-+ err = dpni_get_irq_status(dpni_dev->mc_io, 0, dpni_dev->mc_handle,
-+ DPNI_IRQ_INDEX, &status);
-+ if (unlikely(err)) {
-+ netdev_err(net_dev, "Can't get irq status (err %d)\n", err);
-+ return IRQ_HANDLED;
-+ }
-+
-+ if (status & DPNI_IRQ_EVENT_LINK_CHANGED)
-+ link_state_update(netdev_priv(net_dev));
-+
-+ return IRQ_HANDLED;
-+}
-+
-+static int setup_irqs(struct fsl_mc_device *ls_dev)
-+{
-+ int err = 0;
-+ struct fsl_mc_device_irq *irq;
-+
-+ err = fsl_mc_allocate_irqs(ls_dev);
-+ if (err) {
-+ dev_err(&ls_dev->dev, "MC irqs allocation failed\n");
-+ return err;
-+ }
-+
-+ irq = ls_dev->irqs[0];
-+ err = devm_request_threaded_irq(&ls_dev->dev, irq->msi_desc->irq,
-+ NULL, dpni_irq0_handler_thread,
-+ IRQF_NO_SUSPEND | IRQF_ONESHOT,
-+ dev_name(&ls_dev->dev), &ls_dev->dev);
-+ if (err < 0) {
-+ dev_err(&ls_dev->dev, "devm_request_threaded_irq(): %d\n", err);
-+ goto free_mc_irq;
-+ }
-+
-+ err = dpni_set_irq_mask(ls_dev->mc_io, 0, ls_dev->mc_handle,
-+ DPNI_IRQ_INDEX, DPNI_IRQ_EVENT_LINK_CHANGED);
-+ if (err < 0) {
-+ dev_err(&ls_dev->dev, "dpni_set_irq_mask(): %d\n", err);
-+ goto free_irq;
-+ }
-+
-+ err = dpni_set_irq_enable(ls_dev->mc_io, 0, ls_dev->mc_handle,
-+ DPNI_IRQ_INDEX, 1);
-+ if (err < 0) {
-+ dev_err(&ls_dev->dev, "dpni_set_irq_enable(): %d\n", err);
-+ goto free_irq;
-+ }
-+
-+ return 0;
-+
-+free_irq:
-+ devm_free_irq(&ls_dev->dev, irq->msi_desc->irq, &ls_dev->dev);
-+free_mc_irq:
-+ fsl_mc_free_irqs(ls_dev);
-+
-+ return err;
-+}
-+
-+static void add_ch_napi(struct dpaa2_eth_priv *priv)
-+{
-+ int i;
-+ struct dpaa2_eth_channel *ch;
-+
-+ for (i = 0; i < priv->num_channels; i++) {
-+ ch = priv->channel[i];
-+ /* NAPI weight *MUST* be a multiple of DPAA2_ETH_STORE_SIZE */
-+ netif_napi_add(priv->net_dev, &ch->napi, dpaa2_eth_poll,
-+ NAPI_POLL_WEIGHT);
-+ }
-+}
-+
-+static void del_ch_napi(struct dpaa2_eth_priv *priv)
-+{
-+ int i;
-+ struct dpaa2_eth_channel *ch;
-+
-+ for (i = 0; i < priv->num_channels; i++) {
-+ ch = priv->channel[i];
-+ netif_napi_del(&ch->napi);
-+ }
-+}
-+
-+/* SysFS support */
-+static ssize_t dpaa2_eth_show_tx_shaping(struct device *dev,
-+ struct device_attribute *attr,
-+ char *buf)
-+{
-+ struct dpaa2_eth_priv *priv = netdev_priv(to_net_dev(dev));
-+ /* No MC API for getting the shaping config. We're stateful. */
-+ struct dpni_tx_shaping_cfg *scfg = &priv->shaping_cfg;
-+
-+ return sprintf(buf, "%u %hu\n", scfg->rate_limit, scfg->max_burst_size);
-+}
-+
-+static ssize_t dpaa2_eth_write_tx_shaping(struct device *dev,
-+ struct device_attribute *attr,
-+ const char *buf,
-+ size_t count)
-+{
-+ int err, items;
-+ struct dpaa2_eth_priv *priv = netdev_priv(to_net_dev(dev));
-+ struct dpni_tx_shaping_cfg scfg, ercfg = { 0 };
-+
-+ items = sscanf(buf, "%u %hu", &scfg.rate_limit, &scfg.max_burst_size);
-+ if (items != 2) {
-+ pr_err("Expected format: \"rate_limit(Mbps) max_burst_size(bytes)\"\n");
-+ return -EINVAL;
-+ }
-+ /* Size restriction as per MC API documentation */
-+ if (scfg.max_burst_size > DPAA2_ETH_MAX_BURST_SIZE) {
-+ pr_err("max_burst_size must be <= %d\n",
-+ DPAA2_ETH_MAX_BURST_SIZE);
-+ return -EINVAL;
-+ }
-+
-+ err = dpni_set_tx_shaping(priv->mc_io, 0, priv->mc_token, &scfg,
-+ &ercfg, 0);
-+ if (err) {
-+ dev_err(dev, "dpni_set_tx_shaping() failed\n");
-+ return -EPERM;
-+ }
-+ /* If successful, save the current configuration for future inquiries */
-+ priv->shaping_cfg = scfg;
-+
-+ return count;
-+}
-+
-+static struct device_attribute dpaa2_eth_attrs[] = {
-+ __ATTR(tx_shaping,
-+ 0600,
-+ dpaa2_eth_show_tx_shaping,
-+ dpaa2_eth_write_tx_shaping),
-+};
-+
-+static void dpaa2_eth_sysfs_init(struct device *dev)
-+{
-+ int i, err;
-+
-+ for (i = 0; i < ARRAY_SIZE(dpaa2_eth_attrs); i++) {
-+ err = device_create_file(dev, &dpaa2_eth_attrs[i]);
-+ if (err) {
-+ dev_err(dev, "ERROR creating sysfs file\n");
-+ goto undo;
-+ }
-+ }
-+ return;
-+
-+undo:
-+ while (i > 0)
-+ device_remove_file(dev, &dpaa2_eth_attrs[--i]);
-+}
-+
-+static void dpaa2_eth_sysfs_remove(struct device *dev)
-+{
-+ int i;
-+
-+ for (i = 0; i < ARRAY_SIZE(dpaa2_eth_attrs); i++)
-+ device_remove_file(dev, &dpaa2_eth_attrs[i]);
-+}
-+
-+#ifdef CONFIG_FSL_DPAA2_ETH_DCB
-+static int dpaa2_eth_dcbnl_ieee_getpfc(struct net_device *net_dev,
-+ struct ieee_pfc *pfc)
-+{
-+ struct dpaa2_eth_priv *priv = netdev_priv(net_dev);
-+ struct dpni_congestion_notification_cfg notification_cfg;
-+ struct dpni_link_state state;
-+ int err, i;
-+
-+ priv->pfc.pfc_cap = dpaa2_eth_tc_count(priv);
-+
-+ err = dpni_get_link_state(priv->mc_io, 0, priv->mc_token, &state);
-+ if (err) {
-+ netdev_err(net_dev, "ERROR %d getting link state", err);
-+ return err;
-+ }
-+
-+ if (!(state.options & DPNI_LINK_OPT_PFC_PAUSE))
-+ return 0;
-+
-+ priv->pfc.pfc_en = 0;
-+ for (i = 0; i < dpaa2_eth_tc_count(priv); i++) {
-+ err = dpni_get_congestion_notification(priv->mc_io, 0,
-+ priv->mc_token,
-+ DPNI_QUEUE_RX,
-+ i, &notification_cfg);
-+ if (err) {
-+ netdev_err(net_dev, "Error %d getting congestion notif",
-+ err);
-+ return err;
-+ }
-+
-+ if (notification_cfg.threshold_entry)
-+ priv->pfc.pfc_en |= 1 << i;
-+ }
-+
-+ memcpy(pfc, &priv->pfc, sizeof(priv->pfc));
-+
-+ return 0;
-+}
-+
-+/* Configure ingress classification based on VLAN PCP */
-+static int set_vlan_qos(struct dpaa2_eth_priv *priv)
-+{
-+ struct device *dev = priv->net_dev->dev.parent;
-+ struct dpkg_profile_cfg kg_cfg = {0};
-+ struct dpni_qos_tbl_cfg qos_cfg = {0};
-+ struct dpni_rule_cfg key_params;
-+ u8 *params_iova, *key, *mask = NULL;
-+ /* We only need the trailing 16 bits, without the TPID */
-+ u8 key_size = VLAN_HLEN / 2;
-+ int err = 0, i, j = 0;
-+
-+ if (priv->vlan_clsf_set)
-+ return 0;
-+
-+ params_iova = kzalloc(DPAA2_CLASSIFIER_DMA_SIZE, GFP_KERNEL);
-+ if (!params_iova)
-+ return -ENOMEM;
-+
-+ kg_cfg.num_extracts = 1;
-+ kg_cfg.extracts[0].type = DPKG_EXTRACT_FROM_HDR;
-+ kg_cfg.extracts[0].extract.from_hdr.prot = NET_PROT_VLAN;
-+ kg_cfg.extracts[0].extract.from_hdr.type = DPKG_FULL_FIELD;
-+ kg_cfg.extracts[0].extract.from_hdr.field = NH_FLD_VLAN_TCI;
-+
-+ err = dpni_prepare_key_cfg(&kg_cfg, params_iova);
-+ if (err) {
-+ dev_err(dev, "dpkg_prepare_key_cfg failed: %d\n", err);
-+ goto out_free;
-+ }
-+
-+ /* Set QoS table */
-+ qos_cfg.default_tc = 0;
-+ qos_cfg.discard_on_miss = 0;
-+ qos_cfg.key_cfg_iova = dma_map_single(dev, params_iova,
-+ DPAA2_CLASSIFIER_DMA_SIZE,
-+ DMA_TO_DEVICE);
-+ if (dma_mapping_error(dev, qos_cfg.key_cfg_iova)) {
-+ dev_err(dev, "%s: DMA mapping failed\n", __func__);
-+ err = -ENOMEM;
-+ goto out_free;
-+ }
-+ err = dpni_set_qos_table(priv->mc_io, 0, priv->mc_token, &qos_cfg);
-+ dma_unmap_single(dev, qos_cfg.key_cfg_iova,
-+ DPAA2_CLASSIFIER_DMA_SIZE, DMA_TO_DEVICE);
-+
-+ if (err) {
-+ dev_err(dev, "dpni_set_qos_table failed: %d\n", err);
-+ goto out_free;
-+ }
-+
-+ key_params.key_size = key_size;
-+
-+ if (dpaa2_eth_fs_mask_enabled(priv)) {
-+ mask = kzalloc(key_size, GFP_KERNEL);
-+ if (!mask)
-+ goto out_free;
-+
-+ *mask = cpu_to_be16(VLAN_PRIO_MASK);
-+
-+ key_params.mask_iova = dma_map_single(dev, mask, key_size,
-+ DMA_TO_DEVICE);
-+ if (dma_mapping_error(dev, key_params.mask_iova)) {
-+ dev_err(dev, "DMA mapping failed %s\n", __func__);
-+ err = -ENOMEM;
-+ goto out_free_mask;
-+ }
-+ } else {
-+ key_params.mask_iova = 0;
-+ }
-+
-+ key = kzalloc(key_size, GFP_KERNEL);
-+ if (!key)
-+ goto out_cleanup_mask;
-+
-+ key_params.key_iova = dma_map_single(dev, key, key_size,
-+ DMA_TO_DEVICE);
-+ if (dma_mapping_error(dev, key_params.key_iova)) {
-+ dev_err(dev, "%s: DMA mapping failed\n", __func__);
-+ err = -ENOMEM;
-+ goto out_free_key;
-+ }
-+
-+ for (i = 0; i < dpaa2_eth_tc_count(priv); i++) {
-+ *key = cpu_to_be16(i << VLAN_PRIO_SHIFT);
-+
-+ dma_sync_single_for_device(dev, key_params.key_iova,
-+ key_size, DMA_TO_DEVICE);
-+
-+ err = dpni_add_qos_entry(priv->mc_io, 0, priv->mc_token,
-+ &key_params, i, j++);
-+ if (err) {
-+ dev_err(dev, "dpni_add_qos_entry failed: %d\n", err);
-+ goto out_remove;
-+ }
-+ }
-+
-+ priv->vlan_clsf_set = true;
-+ dev_dbg(dev, "Vlan PCP QoS classification set\n");
-+ goto out_cleanup;
-+
-+out_remove:
-+ for (j = 0; j < i; j++) {
-+ *key = cpu_to_be16(j << VLAN_PRIO_SHIFT);
-+
-+ dma_sync_single_for_device(dev, key_params.key_iova, key_size,
-+ DMA_TO_DEVICE);
-+
-+ err = dpni_remove_qos_entry(priv->mc_io, 0, priv->mc_token,
-+ &key_params);
-+ if (err)
-+ dev_err(dev, "dpni_remove_qos_entry failed: %d\n", err);
-+ }
-+
-+out_cleanup:
-+ dma_unmap_single(dev, key_params.key_iova, key_size, DMA_TO_DEVICE);
-+out_free_key:
-+ kfree(key);
-+out_cleanup_mask:
-+ if (key_params.mask_iova)
-+ dma_unmap_single(dev, key_params.mask_iova, key_size,
-+ DMA_TO_DEVICE);
-+out_free_mask:
-+ kfree(mask);
-+out_free:
-+ kfree(params_iova);
-+ return err;
-+}
-+
-+static int dpaa2_eth_dcbnl_ieee_setpfc(struct net_device *net_dev,
-+ struct ieee_pfc *pfc)
-+{
-+ struct dpaa2_eth_priv *priv = netdev_priv(net_dev);
-+ struct dpni_congestion_notification_cfg notification_cfg = {0};
-+ struct dpni_link_state state = {0};
-+ struct dpni_link_cfg cfg = {0};
-+ struct ieee_pfc old_pfc;
-+ int err = 0, i;
-+
-+ if (dpaa2_eth_tc_count(priv) == 1) {
-+ netdev_dbg(net_dev, "DPNI has 1 TC, PFC configuration N/A\n");
-+ return 0;
-+ }
-+
-+ /* Zero out pfc_enabled prios greater than tc_count */
-+ pfc->pfc_en &= (1 << dpaa2_eth_tc_count(priv)) - 1;
-+
-+ if (priv->pfc.pfc_en == pfc->pfc_en)
-+ /* Same enabled mask, nothing to be done */
-+ return 0;
-+
-+ err = set_vlan_qos(priv);
-+ if (err)
-+ return err;
-+
-+ err = dpni_get_link_state(priv->mc_io, 0, priv->mc_token, &state);
-+ if (err) {
-+ netdev_err(net_dev, "ERROR %d getting link state", err);
-+ return err;
-+ }
-+
-+ cfg.rate = state.rate;
-+ cfg.options = state.options;
-+ if (pfc->pfc_en)
-+ cfg.options |= DPNI_LINK_OPT_PFC_PAUSE;
-+ else
-+ cfg.options &= ~DPNI_LINK_OPT_PFC_PAUSE;
-+
-+ err = dpni_set_link_cfg(priv->mc_io, 0, priv->mc_token, &cfg);
-+ if (err) {
-+ netdev_err(net_dev, "ERROR %d setting link cfg", err);
-+ return err;
-+ }
-+
-+ memcpy(&old_pfc, &priv->pfc, sizeof(priv->pfc));
-+ memcpy(&priv->pfc, pfc, sizeof(priv->pfc));
-+
-+ err = set_rx_taildrop(priv);
-+ if (err)
-+ goto out_restore_config;
-+
-+ /* configure congestion notifications */
-+ notification_cfg.notification_mode = DPNI_CONG_OPT_FLOW_CONTROL;
-+ notification_cfg.units = DPNI_CONGESTION_UNIT_FRAMES;
-+ notification_cfg.message_iova = 0ULL;
-+ notification_cfg.message_ctx = 0ULL;
-+
-+ for (i = 0; i < dpaa2_eth_tc_count(priv); i++) {
-+ if (dpaa2_eth_is_pfc_enabled(priv, i)) {
-+ notification_cfg.threshold_entry = NAPI_POLL_WEIGHT;
-+ notification_cfg.threshold_exit = NAPI_POLL_WEIGHT / 2;
-+ } else {
-+ notification_cfg.threshold_entry = 0;
-+ notification_cfg.threshold_exit = 0;
-+ }
-+
-+ err = dpni_set_congestion_notification(priv->mc_io, 0,
-+ priv->mc_token,
-+ DPNI_QUEUE_RX,
-+ i, &notification_cfg);
-+ if (err) {
-+ netdev_err(net_dev, "Error %d setting congestion notif",
-+ err);
-+ goto out_restore_config;
-+ }
-+
-+ netdev_dbg(net_dev, "%s congestion notifications for tc %d\n",
-+ (notification_cfg.threshold_entry ?
-+ "Enabled" : "Disabled"), i);
-+ }
-+
-+ return 0;
-+
-+out_restore_config:
-+ memcpy(&priv->pfc, &old_pfc, sizeof(priv->pfc));
-+ return err;
-+}
-+
-+static u8 dpaa2_eth_dcbnl_getdcbx(struct net_device *net_dev)
-+{
-+ struct dpaa2_eth_priv *priv = netdev_priv(net_dev);
-+
-+ return priv->dcbx_mode;
-+}
-+
-+static u8 dpaa2_eth_dcbnl_setdcbx(struct net_device *net_dev, u8 mode)
-+{
-+ struct dpaa2_eth_priv *priv = netdev_priv(net_dev);
-+
-+ priv->dcbx_mode = mode;
-+ return 0;
-+}
-+
-+static u8 dpaa2_eth_dcbnl_getcap(struct net_device *net_dev, int capid, u8 *cap)
-+{
-+ struct dpaa2_eth_priv *priv = netdev_priv(net_dev);
-+
-+ switch (capid) {
-+ case DCB_CAP_ATTR_PFC:
-+ *cap = true;
-+ break;
-+ case DCB_CAP_ATTR_PFC_TCS:
-+ /* bitmap where each bit represents a number of traffic
-+ * classes the device can be configured to use for Priority
-+ * Flow Control
-+ */
-+ *cap = 1 << (dpaa2_eth_tc_count(priv) - 1);
-+ break;
-+ case DCB_CAP_ATTR_DCBX:
-+ *cap = priv->dcbx_mode;
-+ break;
-+ default:
-+ *cap = false;
-+ break;
-+ }
-+
-+ return 0;
-+}
-+
-+const struct dcbnl_rtnl_ops dpaa2_eth_dcbnl_ops = {
-+ .ieee_getpfc = dpaa2_eth_dcbnl_ieee_getpfc,
-+ .ieee_setpfc = dpaa2_eth_dcbnl_ieee_setpfc,
-+ .getdcbx = dpaa2_eth_dcbnl_getdcbx,
-+ .setdcbx = dpaa2_eth_dcbnl_setdcbx,
-+ .getcap = dpaa2_eth_dcbnl_getcap,
-+};
-+#endif
-+
-+static int dpaa2_eth_probe(struct fsl_mc_device *dpni_dev)
-+{
-+ struct device *dev;
-+ struct net_device *net_dev = NULL;
-+ struct dpaa2_eth_priv *priv = NULL;
-+ int err = 0;
-+
-+ dev = &dpni_dev->dev;
-+
-+ /* Net device */
-+ net_dev = alloc_etherdev_mq(sizeof(*priv), DPAA2_ETH_MAX_TX_QUEUES);
-+ if (!net_dev) {
-+ dev_err(dev, "alloc_etherdev_mq() failed\n");
-+ return -ENOMEM;
-+ }
-+
-+ SET_NETDEV_DEV(net_dev, dev);
-+ dev_set_drvdata(dev, net_dev);
-+
-+ priv = netdev_priv(net_dev);
-+ priv->net_dev = net_dev;
-+
-+ priv->iommu_domain = iommu_get_domain_for_dev(dev);
-+
-+ /* Obtain a MC portal */
-+ err = fsl_mc_portal_allocate(dpni_dev, FSL_MC_IO_ATOMIC_CONTEXT_PORTAL,
-+ &priv->mc_io);
-+ if (err) {
-+ if (err == -ENXIO)
-+ err = -EPROBE_DEFER;
-+ else
-+ dev_err(dev, "MC portal allocation failed\n");
-+ goto err_portal_alloc;
-+ }
-+
-+ /* MC objects initialization and configuration */
-+ err = setup_dpni(dpni_dev);
-+ if (err)
-+ goto err_dpni_setup;
-+
-+ err = setup_dpio(priv);
-+ if (err)
-+ goto err_dpio_setup;
-+
-+ setup_fqs(priv);
-+
-+ err = setup_dpbp(priv);
-+ if (err)
-+ goto err_dpbp_setup;
-+
-+ err = bind_dpni(priv);
-+ if (err)
-+ goto err_bind;
-+
-+ /* Percpu statistics */
-+ priv->percpu_stats = alloc_percpu(*priv->percpu_stats);
-+ if (!priv->percpu_stats) {
-+ dev_err(dev, "alloc_percpu(percpu_stats) failed\n");
-+ err = -ENOMEM;
-+ goto err_alloc_percpu_stats;
-+ }
-+ priv->percpu_extras = alloc_percpu(*priv->percpu_extras);
-+ if (!priv->percpu_extras) {
-+ dev_err(dev, "alloc_percpu(percpu_extras) failed\n");
-+ err = -ENOMEM;
-+ goto err_alloc_percpu_extras;
-+ }
-+
-+ err = netdev_init(net_dev);
-+ if (err)
-+ goto err_netdev_init;
-+
-+ /* Configure checksum offload based on current interface flags */
-+ err = set_rx_csum(priv, !!(net_dev->features & NETIF_F_RXCSUM));
-+ if (err)
-+ goto err_csum;
-+
-+ err = set_tx_csum(priv, !!(net_dev->features &
-+ (NETIF_F_IP_CSUM | NETIF_F_IPV6_CSUM)));
-+ if (err)
-+ goto err_csum;
-+
-+ err = alloc_rings(priv);
-+ if (err)
-+ goto err_alloc_rings;
-+
-+ net_dev->ethtool_ops = &dpaa2_ethtool_ops;
-+#ifdef CONFIG_FSL_DPAA2_ETH_DCB
-+ net_dev->dcbnl_ops = &dpaa2_eth_dcbnl_ops;
-+ priv->dcbx_mode = DCB_CAP_DCBX_HOST | DCB_CAP_DCBX_VER_IEEE;
-+#endif
-+
-+ /* Add a NAPI context for each channel */
-+ add_ch_napi(priv);
-+ enable_ch_napi(priv);
-+
-+ err = setup_irqs(dpni_dev);
-+ if (err) {
-+ netdev_warn(net_dev, "Failed to set link interrupt, fall back to polling\n");
-+ priv->poll_thread = kthread_run(poll_link_state, priv,
-+ "%s_poll_link", net_dev->name);
-+ if (IS_ERR(priv->poll_thread)) {
-+ netdev_err(net_dev, "Error starting polling thread\n");
-+ goto err_poll_thread;
-+ }
-+ priv->do_link_poll = true;
-+ }
-+
-+ dpaa2_eth_sysfs_init(&net_dev->dev);
-+#ifdef CONFIG_FSL_DPAA2_ETH_DEBUGFS
-+ dpaa2_dbg_add(priv);
-+#endif
-+
-+ dev_info(dev, "Probed interface %s\n", net_dev->name);
-+ return 0;
-+
-+err_poll_thread:
-+ free_rings(priv);
-+err_alloc_rings:
-+err_csum:
-+ unregister_netdev(net_dev);
-+err_netdev_init:
-+ free_percpu(priv->percpu_extras);
-+err_alloc_percpu_extras:
-+ free_percpu(priv->percpu_stats);
-+err_alloc_percpu_stats:
-+ disable_ch_napi(priv);
-+ del_ch_napi(priv);
-+err_bind:
-+ free_dpbp(priv);
-+err_dpbp_setup:
-+ free_dpio(priv);
-+err_dpio_setup:
-+ free_dpni(priv);
-+err_dpni_setup:
-+ fsl_mc_portal_free(priv->mc_io);
-+err_portal_alloc:
-+ dev_set_drvdata(dev, NULL);
-+ free_netdev(net_dev);
-+
-+ return err;
-+}
-+
-+static int dpaa2_eth_remove(struct fsl_mc_device *ls_dev)
-+{
-+ struct device *dev;
-+ struct net_device *net_dev;
-+ struct dpaa2_eth_priv *priv;
-+
-+ dev = &ls_dev->dev;
-+ net_dev = dev_get_drvdata(dev);
-+ priv = netdev_priv(net_dev);
-+
-+#ifdef CONFIG_FSL_DPAA2_ETH_DEBUGFS
-+ dpaa2_dbg_remove(priv);
-+#endif
-+ dpaa2_eth_sysfs_remove(&net_dev->dev);
-+
-+ unregister_netdev(net_dev);
-+
-+ disable_ch_napi(priv);
-+ del_ch_napi(priv);
-+
-+ if (priv->do_link_poll)
-+ kthread_stop(priv->poll_thread);
-+ else
-+ fsl_mc_free_irqs(ls_dev);
-+
-+ free_rings(priv);
-+ free_percpu(priv->percpu_stats);
-+ free_percpu(priv->percpu_extras);
-+ free_dpbp(priv);
-+ free_dpio(priv);
-+ free_dpni(priv);
-+
-+ fsl_mc_portal_free(priv->mc_io);
-+
-+ dev_set_drvdata(dev, NULL);
-+ free_netdev(net_dev);
-+
-+ dev_dbg(net_dev->dev.parent, "Removed interface %s\n", net_dev->name);
-+
-+ return 0;
-+}
-+
-+static const struct fsl_mc_device_id dpaa2_eth_match_id_table[] = {
-+ {
-+ .vendor = FSL_MC_VENDOR_FREESCALE,
-+ .obj_type = "dpni",
-+ },
-+ { .vendor = 0x0 }
-+};
-+MODULE_DEVICE_TABLE(fslmc, dpaa2_eth_match_id_table);
-+
-+static struct fsl_mc_driver dpaa2_eth_driver = {
-+ .driver = {
-+ .name = KBUILD_MODNAME,
-+ .owner = THIS_MODULE,
-+ },
-+ .probe = dpaa2_eth_probe,
-+ .remove = dpaa2_eth_remove,
-+ .match_id_table = dpaa2_eth_match_id_table
-+};
-+
-+static int __init dpaa2_eth_driver_init(void)
-+{
-+ int err;
-+
-+ dpaa2_eth_dbg_init();
-+ err = fsl_mc_driver_register(&dpaa2_eth_driver);
-+ if (err)
-+ goto out_debugfs_err;
-+
-+ err = dpaa2_ceetm_register();
-+ if (err)
-+ goto out_ceetm_err;
-+
-+ return 0;
-+
-+out_ceetm_err:
-+ fsl_mc_driver_unregister(&dpaa2_eth_driver);
-+out_debugfs_err:
-+ dpaa2_eth_dbg_exit();
-+ return err;
-+}
-+
-+static void __exit dpaa2_eth_driver_exit(void)
-+{
-+ dpaa2_ceetm_unregister();
-+ fsl_mc_driver_unregister(&dpaa2_eth_driver);
-+ dpaa2_eth_dbg_exit();
-+}
-+
-+module_init(dpaa2_eth_driver_init);
-+module_exit(dpaa2_eth_driver_exit);
---- /dev/null
-+++ b/drivers/staging/fsl-dpaa2/ethernet/dpaa2-eth.h
-@@ -0,0 +1,601 @@
-+/* Copyright 2014-2016 Freescale Semiconductor Inc.
-+ * Copyright 2016 NXP
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#ifndef __DPAA2_ETH_H
-+#define __DPAA2_ETH_H
-+
-+#include <linux/dcbnl.h>
-+#include <linux/netdevice.h>
-+#include <linux/if_vlan.h>
-+
-+#include "../../fsl-mc/include/dpaa2-io.h"
-+#include "../../fsl-mc/include/dpaa2-fd.h"
-+#include "dpni.h"
-+#include "dpni-cmd.h"
-+
-+#include "dpaa2-eth-trace.h"
-+#include "dpaa2-eth-debugfs.h"
-+
-+#define DPAA2_WRIOP_VERSION(x, y, z) ((x) << 10 | (y) << 5 | (z) << 0)
-+
-+#define DPAA2_ETH_STORE_SIZE 16
-+
-+/* We set a max threshold for how many Tx confirmations we should process
-+ * on a NAPI poll call, they take less processing time.
-+ */
-+#define TX_CONF_PER_NAPI_POLL 256
-+
-+/* Maximum number of scatter-gather entries in an ingress frame,
-+ * considering the maximum receive frame size is 64K
-+ */
-+#define DPAA2_ETH_MAX_SG_ENTRIES ((64 * 1024) / DPAA2_ETH_RX_BUF_SIZE)
-+
-+/* Maximum acceptable MTU value. It is in direct relation with the hardware
-+ * enforced Max Frame Length (currently 10k).
-+ */
-+#define DPAA2_ETH_MFL (10 * 1024)
-+#define DPAA2_ETH_MAX_MTU (DPAA2_ETH_MFL - VLAN_ETH_HLEN)
-+/* Convert L3 MTU to L2 MFL */
-+#define DPAA2_ETH_L2_MAX_FRM(mtu) ((mtu) + VLAN_ETH_HLEN)
-+
-+/* Maximum burst size value for Tx shaping */
-+#define DPAA2_ETH_MAX_BURST_SIZE 0xF7FF
-+
-+/* Maximum number of buffers that can be acquired/released through a single
-+ * QBMan command
-+ */
-+#define DPAA2_ETH_BUFS_PER_CMD 7
-+
-+/* Set the taildrop threshold (in bytes) to allow the enqueue of several jumbo
-+ * frames in the Rx queues (length of the current frame is not
-+ * taken into account when making the taildrop decision)
-+ */
-+#define DPAA2_ETH_TAILDROP_THRESH (64 * 1024)
-+
-+/* Buffer quota per queue. Must be large enough such that for minimum sized
-+ * frames taildrop kicks in before the bpool gets depleted, so we compute
-+ * how many 64B frames fit inside the taildrop threshold and add a margin
-+ * to accommodate the buffer refill delay.
-+ */
-+#define DPAA2_ETH_MAX_FRAMES_PER_QUEUE (DPAA2_ETH_TAILDROP_THRESH / 64)
-+#define DPAA2_ETH_NUM_BUFS_PER_CH (DPAA2_ETH_MAX_FRAMES_PER_QUEUE + 256)
-+#define DPAA2_ETH_REFILL_THRESH(priv) \
-+ ((priv)->max_bufs_per_ch - DPAA2_ETH_BUFS_PER_CMD)
-+
-+/* Global buffer quota in case flow control is enabled */
-+#define DPAA2_ETH_NUM_BUFS_FC 256
-+
-+/* Hardware requires alignment for ingress/egress buffer addresses */
-+#define DPAA2_ETH_TX_BUF_ALIGN 64
-+
-+/* Due to a limitation in WRIOP 1.0.0, the RX buffer data must be aligned
-+ * to 256B. For newer revisions, the requirement is only for 64B alignment
-+ */
-+#define DPAA2_ETH_RX_BUF_ALIGN_REV1 256
-+#define DPAA2_ETH_RX_BUF_ALIGN 64
-+
-+#define DPAA2_ETH_RX_BUF_SIZE 2048
-+#define DPAA2_ETH_SKB_SIZE \
-+ (DPAA2_ETH_RX_BUF_SIZE + SKB_DATA_ALIGN(sizeof(struct skb_shared_info)))
-+
-+/* PTP nominal frequency 1GHz */
-+#define DPAA2_PTP_NOMINAL_FREQ_PERIOD_NS 1
-+
-+/* Hardware annotation area in RX/TX buffers */
-+#define DPAA2_ETH_RX_HWA_SIZE 64
-+#define DPAA2_ETH_TX_HWA_SIZE 128
-+
-+/* We are accommodating a skb backpointer and some S/G info
-+ * in the frame's software annotation. The hardware
-+ * options are either 0 or 64, so we choose the latter.
-+ */
-+#define DPAA2_ETH_SWA_SIZE 64
-+
-+/* We store different information in the software annotation area of a Tx frame
-+ * based on what type of frame it is
-+ */
-+enum dpaa2_eth_swa_type {
-+ DPAA2_ETH_SWA_SINGLE,
-+ DPAA2_ETH_SWA_SG,
-+};
-+
-+/* Must keep this struct smaller than DPAA2_ETH_SWA_SIZE */
-+struct dpaa2_eth_swa {
-+ enum dpaa2_eth_swa_type type;
-+ union {
-+ struct {
-+ struct sk_buff *skb;
-+ } single;
-+ struct {
-+ struct sk_buff *skb;
-+ struct scatterlist *scl;
-+ int num_sg;
-+ int sgt_size;
-+ } sg;
-+ };
-+};
-+
-+/* Annotation valid bits in FD FRC */
-+#define DPAA2_FD_FRC_FASV 0x8000
-+#define DPAA2_FD_FRC_FAEADV 0x4000
-+#define DPAA2_FD_FRC_FAPRV 0x2000
-+#define DPAA2_FD_FRC_FAIADV 0x1000
-+#define DPAA2_FD_FRC_FASWOV 0x0800
-+#define DPAA2_FD_FRC_FAICFDV 0x0400
-+
-+#define DPAA2_FD_RX_ERR_MASK (FD_CTRL_SBE | FD_CTRL_FAERR)
-+#define DPAA2_FD_TX_ERR_MASK (FD_CTRL_UFD | \
-+ FD_CTRL_SBE | \
-+ FD_CTRL_FSE | \
-+ FD_CTRL_FAERR)
-+
-+/* Annotation bits in FD CTRL */
-+#define DPAA2_FD_CTRL_ASAL 0x00020000 /* ASAL = 128 */
-+
-+/* Frame annotation status */
-+struct dpaa2_fas {
-+ u8 reserved;
-+ u8 ppid;
-+ __le16 ifpid;
-+ __le32 status;
-+};
-+
-+/* Frame annotation status word is located in the first 8 bytes
-+ * of the buffer's hardware annoatation area
-+ */
-+#define DPAA2_FAS_OFFSET 0
-+#define DPAA2_FAS_SIZE (sizeof(struct dpaa2_fas))
-+
-+/* Timestamp is located in the next 8 bytes of the buffer's
-+ * hardware annotation area
-+ */
-+#define DPAA2_TS_OFFSET 0x8
-+
-+/* Frame annotation egress action descriptor */
-+#define DPAA2_FAEAD_OFFSET 0x58
-+
-+struct dpaa2_faead {
-+ __le32 conf_fqid;
-+ __le32 ctrl;
-+};
-+
-+#define DPAA2_FAEAD_A2V 0x20000000
-+#define DPAA2_FAEAD_A4V 0x08000000
-+#define DPAA2_FAEAD_UPDV 0x00001000
-+#define DPAA2_FAEAD_EBDDV 0x00002000
-+#define DPAA2_FAEAD_UPD 0x00000010
-+
-+/* Accessors for the hardware annotation fields that we use */
-+static inline void *dpaa2_get_hwa(void *buf_addr, bool swa)
-+{
-+ return buf_addr + (swa ? DPAA2_ETH_SWA_SIZE : 0);
-+}
-+
-+static inline struct dpaa2_fas *dpaa2_get_fas(void *buf_addr, bool swa)
-+{
-+ return dpaa2_get_hwa(buf_addr, swa) + DPAA2_FAS_OFFSET;
-+}
-+
-+static inline u64 *dpaa2_get_ts(void *buf_addr, bool swa)
-+{
-+ return dpaa2_get_hwa(buf_addr, swa) + DPAA2_TS_OFFSET;
-+}
-+
-+static inline struct dpaa2_faead *dpaa2_get_faead(void *buf_addr, bool swa)
-+{
-+ return dpaa2_get_hwa(buf_addr, swa) + DPAA2_FAEAD_OFFSET;
-+}
-+
-+/* Error and status bits in the frame annotation status word */
-+/* Debug frame, otherwise supposed to be discarded */
-+#define DPAA2_FAS_DISC 0x80000000
-+/* MACSEC frame */
-+#define DPAA2_FAS_MS 0x40000000
-+#define DPAA2_FAS_PTP 0x08000000
-+/* Ethernet multicast frame */
-+#define DPAA2_FAS_MC 0x04000000
-+/* Ethernet broadcast frame */
-+#define DPAA2_FAS_BC 0x02000000
-+#define DPAA2_FAS_KSE 0x00040000
-+#define DPAA2_FAS_EOFHE 0x00020000
-+#define DPAA2_FAS_MNLE 0x00010000
-+#define DPAA2_FAS_TIDE 0x00008000
-+#define DPAA2_FAS_PIEE 0x00004000
-+/* Frame length error */
-+#define DPAA2_FAS_FLE 0x00002000
-+/* Frame physical error */
-+#define DPAA2_FAS_FPE 0x00001000
-+#define DPAA2_FAS_PTE 0x00000080
-+#define DPAA2_FAS_ISP 0x00000040
-+#define DPAA2_FAS_PHE 0x00000020
-+#define DPAA2_FAS_BLE 0x00000010
-+/* L3 csum validation performed */
-+#define DPAA2_FAS_L3CV 0x00000008
-+/* L3 csum error */
-+#define DPAA2_FAS_L3CE 0x00000004
-+/* L4 csum validation performed */
-+#define DPAA2_FAS_L4CV 0x00000002
-+/* L4 csum error */
-+#define DPAA2_FAS_L4CE 0x00000001
-+/* Possible errors on the ingress path */
-+#define DPAA2_FAS_RX_ERR_MASK (DPAA2_FAS_KSE | \
-+ DPAA2_FAS_EOFHE | \
-+ DPAA2_FAS_MNLE | \
-+ DPAA2_FAS_TIDE | \
-+ DPAA2_FAS_PIEE | \
-+ DPAA2_FAS_FLE | \
-+ DPAA2_FAS_FPE | \
-+ DPAA2_FAS_PTE | \
-+ DPAA2_FAS_ISP | \
-+ DPAA2_FAS_PHE | \
-+ DPAA2_FAS_BLE | \
-+ DPAA2_FAS_L3CE | \
-+ DPAA2_FAS_L4CE)
-+
-+/* Time in milliseconds between link state updates */
-+#define DPAA2_ETH_LINK_STATE_REFRESH 1000
-+
-+/* Number of times to retry a frame enqueue before giving up.
-+ * Value determined empirically, in order to minimize the number
-+ * of frames dropped on Tx
-+ */
-+#define DPAA2_ETH_ENQUEUE_RETRIES 10
-+
-+/* Tx congestion entry & exit thresholds, in number of bytes.
-+ * We allow a maximum of 512KB worth of frames pending processing on the Tx
-+ * queues of an interface
-+ */
-+#define DPAA2_ETH_TX_CONG_ENTRY_THRESH (512 * 1024)
-+#define DPAA2_ETH_TX_CONG_EXIT_THRESH \
-+ (DPAA2_ETH_TX_CONG_ENTRY_THRESH * 9 / 10)
-+
-+/* Driver statistics, other than those in struct rtnl_link_stats64.
-+ * These are usually collected per-CPU and aggregated by ethtool.
-+ */
-+struct dpaa2_eth_drv_stats {
-+ __u64 tx_conf_frames;
-+ __u64 tx_conf_bytes;
-+ __u64 tx_sg_frames;
-+ __u64 tx_sg_bytes;
-+ __u64 tx_reallocs;
-+ __u64 rx_sg_frames;
-+ __u64 rx_sg_bytes;
-+ /* Enqueues retried due to portal busy */
-+ __u64 tx_portal_busy;
-+};
-+
-+/* Per-FQ statistics */
-+struct dpaa2_eth_fq_stats {
-+ /* Number of frames received on this queue */
-+ __u64 frames;
-+ /* Number of times this queue entered congestion */
-+ __u64 congestion_entry;
-+};
-+
-+/* Per-channel statistics */
-+struct dpaa2_eth_ch_stats {
-+ /* Volatile dequeues retried due to portal busy */
-+ __u64 dequeue_portal_busy;
-+ /* Number of CDANs; useful to estimate avg NAPI len */
-+ __u64 cdan;
-+ /* Number of frames received on queues from this channel */
-+ __u64 frames;
-+ /* Pull errors */
-+ __u64 pull_err;
-+};
-+
-+#define DPAA2_ETH_MAX_TCS 8
-+
-+/* Maximum number of queues associated with a DPNI */
-+#define DPAA2_ETH_MAX_RX_QUEUES (DPNI_MAX_DIST_SIZE * DPAA2_ETH_MAX_TCS)
-+#define DPAA2_ETH_MAX_TX_QUEUES DPNI_MAX_SENDERS
-+#define DPAA2_ETH_MAX_RX_ERR_QUEUES 1
-+#define DPAA2_ETH_MAX_QUEUES (DPAA2_ETH_MAX_RX_QUEUES + \
-+ DPAA2_ETH_MAX_TX_QUEUES + \
-+ DPAA2_ETH_MAX_RX_ERR_QUEUES)
-+
-+#define DPAA2_ETH_MAX_DPCONS 16
-+
-+enum dpaa2_eth_fq_type {
-+ DPAA2_RX_FQ = 0,
-+ DPAA2_TX_CONF_FQ,
-+ DPAA2_RX_ERR_FQ
-+};
-+
-+struct dpaa2_eth_priv;
-+
-+struct dpaa2_eth_fq {
-+ u32 fqid;
-+ u32 tx_qdbin;
-+ u16 flowid;
-+ u8 tc;
-+ int target_cpu;
-+ struct dpaa2_eth_channel *channel;
-+ enum dpaa2_eth_fq_type type;
-+
-+ void (*consume)(struct dpaa2_eth_priv *,
-+ struct dpaa2_eth_channel *,
-+ const struct dpaa2_fd *,
-+ struct napi_struct *,
-+ u16 queue_id);
-+ struct dpaa2_eth_fq_stats stats;
-+};
-+
-+struct dpaa2_eth_channel {
-+ struct dpaa2_io_notification_ctx nctx;
-+ struct fsl_mc_device *dpcon;
-+ int dpcon_id;
-+ int ch_id;
-+ struct napi_struct napi;
-+ struct dpaa2_io *dpio;
-+ struct dpaa2_io_store *store;
-+ struct dpaa2_eth_priv *priv;
-+ int buf_count;
-+ struct dpaa2_eth_ch_stats stats;
-+ struct bpf_prog *xdp_prog;
-+ u64 rel_buf_array[DPAA2_ETH_BUFS_PER_CMD];
-+ u8 rel_buf_cnt;
-+};
-+
-+struct dpaa2_eth_cls_rule {
-+ struct ethtool_rx_flow_spec fs;
-+ bool in_use;
-+};
-+
-+struct dpaa2_eth_dist_fields {
-+ u64 rxnfc_field;
-+ enum net_prot cls_prot;
-+ int cls_field;
-+ int offset;
-+ int size;
-+ u32 id;
-+};
-+
-+/* Driver private data */
-+struct dpaa2_eth_priv {
-+ struct net_device *net_dev;
-+ /* Standard statistics */
-+ struct rtnl_link_stats64 __percpu *percpu_stats;
-+ /* Extra stats, in addition to the ones known by the kernel */
-+ struct dpaa2_eth_drv_stats __percpu *percpu_extras;
-+ bool ts_tx_en; /* Tx timestamping enabled */
-+ bool ts_rx_en; /* Rx timestamping enabled */
-+ u16 tx_data_offset;
-+ u16 bpid;
-+ u16 tx_qdid;
-+ u16 rx_buf_align;
-+ struct iommu_domain *iommu_domain;
-+ int max_bufs_per_ch;
-+ int refill_thresh;
-+ bool has_xdp_prog;
-+
-+ void *cscn_mem; /* Tx congestion notifications are written here */
-+ void *cscn_unaligned;
-+ dma_addr_t cscn_dma;
-+
-+ u8 num_fqs;
-+ struct dpaa2_eth_fq fq[DPAA2_ETH_MAX_QUEUES];
-+
-+ u8 num_channels;
-+ struct dpaa2_eth_channel *channel[DPAA2_ETH_MAX_DPCONS];
-+
-+ struct dpni_attr dpni_attrs;
-+ u16 dpni_ver_major;
-+ u16 dpni_ver_minor;
-+ struct fsl_mc_device *dpbp_dev;
-+
-+ struct fsl_mc_io *mc_io;
-+ /* Cores which have an affine DPIO/DPCON.
-+ * This is the cpu set on which Rx and Tx conf frames are processed
-+ */
-+ struct cpumask dpio_cpumask;
-+
-+ u16 mc_token;
-+
-+ struct dpni_link_state link_state;
-+ bool do_link_poll;
-+ struct task_struct *poll_thread;
-+
-+ /* Rx distribution (hash and flow steering) header fields
-+ * supported by the driver
-+ */
-+ struct dpaa2_eth_dist_fields *dist_fields;
-+ u8 num_dist_fields;
-+ /* enabled ethtool hashing bits */
-+ u64 rx_hash_fields;
-+#ifdef CONFIG_FSL_DPAA2_ETH_DEBUGFS
-+ struct dpaa2_debugfs dbg;
-+#endif
-+ /* array of classification rules */
-+ struct dpaa2_eth_cls_rule *cls_rule;
-+ struct dpni_tx_shaping_cfg shaping_cfg;
-+
-+ u8 dcbx_mode;
-+ struct ieee_pfc pfc;
-+ bool vlan_clsf_set;
-+ bool tx_pause_frames;
-+
-+ bool ceetm_en;
-+};
-+
-+enum dpaa2_eth_rx_dist {
-+ DPAA2_ETH_RX_DIST_HASH,
-+ DPAA2_ETH_RX_DIST_FS,
-+ DPAA2_ETH_RX_DIST_LEGACY
-+};
-+
-+/* Supported Rx distribution field ids */
-+#define DPAA2_ETH_DIST_ETHSRC BIT(0)
-+#define DPAA2_ETH_DIST_ETHDST BIT(1)
-+#define DPAA2_ETH_DIST_ETHTYPE BIT(2)
-+#define DPAA2_ETH_DIST_VLAN BIT(3)
-+#define DPAA2_ETH_DIST_IPSRC BIT(4)
-+#define DPAA2_ETH_DIST_IPDST BIT(5)
-+#define DPAA2_ETH_DIST_IPPROTO BIT(6)
-+#define DPAA2_ETH_DIST_L4SRC BIT(7)
-+#define DPAA2_ETH_DIST_L4DST BIT(8)
-+#define DPAA2_ETH_DIST_ALL (~0U)
-+
-+/* Default Rx hash key */
-+#define DPAA2_ETH_DIST_DEFAULT_HASH \
-+ (DPAA2_ETH_DIST_IPPROTO | \
-+ DPAA2_ETH_DIST_IPSRC | DPAA2_ETH_DIST_IPDST | \
-+ DPAA2_ETH_DIST_L4SRC | DPAA2_ETH_DIST_L4DST)
-+
-+#define dpaa2_eth_hash_enabled(priv) \
-+ ((priv)->dpni_attrs.num_queues > 1)
-+
-+#define dpaa2_eth_fs_enabled(priv) \
-+ (!((priv)->dpni_attrs.options & DPNI_OPT_NO_FS))
-+
-+#define dpaa2_eth_fs_mask_enabled(priv) \
-+ ((priv)->dpni_attrs.options & DPNI_OPT_HAS_KEY_MASKING)
-+
-+#define dpaa2_eth_fs_count(priv) \
-+ ((priv)->dpni_attrs.fs_entries)
-+
-+/* Required by struct dpni_rx_tc_dist_cfg::key_cfg_iova */
-+#define DPAA2_CLASSIFIER_DMA_SIZE 256
-+
-+extern const struct ethtool_ops dpaa2_ethtool_ops;
-+extern const char dpaa2_eth_drv_version[];
-+
-+static inline int dpaa2_eth_cmp_dpni_ver(struct dpaa2_eth_priv *priv,
-+ u16 ver_major, u16 ver_minor)
-+{
-+ if (priv->dpni_ver_major == ver_major)
-+ return priv->dpni_ver_minor - ver_minor;
-+ return priv->dpni_ver_major - ver_major;
-+}
-+
-+#define DPNI_DIST_KEY_VER_MAJOR 7
-+#define DPNI_DIST_KEY_VER_MINOR 5
-+
-+static inline bool dpaa2_eth_has_legacy_dist(struct dpaa2_eth_priv *priv)
-+{
-+ return (dpaa2_eth_cmp_dpni_ver(priv, DPNI_DIST_KEY_VER_MAJOR,
-+ DPNI_DIST_KEY_VER_MINOR) < 0);
-+}
-+
-+/* Hardware only sees DPAA2_ETH_RX_BUF_SIZE, but the skb built around
-+ * the buffer also needs space for its shared info struct, and we need
-+ * to allocate enough to accommodate hardware alignment restrictions
-+ */
-+static inline unsigned int dpaa2_eth_buf_raw_size(struct dpaa2_eth_priv *priv)
-+{
-+ return DPAA2_ETH_SKB_SIZE + priv->rx_buf_align;
-+}
-+
-+/* Total headroom needed by the hardware in Tx frame buffers */
-+static inline unsigned int
-+dpaa2_eth_needed_headroom(struct dpaa2_eth_priv *priv, struct sk_buff *skb)
-+{
-+ unsigned int headroom = DPAA2_ETH_SWA_SIZE;
-+
-+ /* If we don't have an skb (e.g. XDP buffer), we only need space for
-+ * the software annotation area
-+ */
-+ if (!skb)
-+ return headroom;
-+
-+ /* For non-linear skbs we have no headroom requirement, as we build a
-+ * SG frame with a newly allocated SGT buffer
-+ */
-+ if (skb_is_nonlinear(skb))
-+ return 0;
-+
-+ /* If we have Tx timestamping, need 128B hardware annotation */
-+ if (priv->ts_tx_en && skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP)
-+ headroom += DPAA2_ETH_TX_HWA_SIZE;
-+
-+ return headroom;
-+}
-+
-+/* Extra headroom space requested to hardware, in order to make sure there's
-+ * no realloc'ing in forwarding scenarios. We need to reserve enough space
-+ * such that we can accommodate the maximum required Tx offset and alignment
-+ * in the ingress frame buffer
-+ */
-+static inline unsigned int dpaa2_eth_rx_headroom(struct dpaa2_eth_priv *priv)
-+{
-+ return priv->tx_data_offset + DPAA2_ETH_TX_BUF_ALIGN -
-+ DPAA2_ETH_RX_HWA_SIZE;
-+}
-+
-+static inline int dpaa2_eth_queue_count(struct dpaa2_eth_priv *priv)
-+{
-+ return priv->dpni_attrs.num_queues;
-+}
-+
-+static inline int dpaa2_eth_tc_count(struct dpaa2_eth_priv *priv)
-+{
-+ return priv->dpni_attrs.num_tcs;
-+}
-+
-+static inline bool dpaa2_eth_is_pfc_enabled(struct dpaa2_eth_priv *priv,
-+ int traffic_class)
-+{
-+ return priv->pfc.pfc_en & (1 << traffic_class);
-+}
-+
-+enum dpaa2_eth_td_cfg {
-+ DPAA2_ETH_TD_NONE,
-+ DPAA2_ETH_TD_QUEUE,
-+ DPAA2_ETH_TD_GROUP
-+};
-+
-+static inline enum dpaa2_eth_td_cfg
-+dpaa2_eth_get_td_type(struct dpaa2_eth_priv *priv)
-+{
-+ bool pfc_enabled = !!(priv->pfc.pfc_en);
-+
-+ if (pfc_enabled)
-+ return DPAA2_ETH_TD_GROUP;
-+ else if (priv->tx_pause_frames)
-+ return DPAA2_ETH_TD_NONE;
-+ else
-+ return DPAA2_ETH_TD_QUEUE;
-+}
-+
-+static inline int dpaa2_eth_ch_count(struct dpaa2_eth_priv *priv)
-+{
-+ return 1;
-+}
-+
-+void check_cls_support(struct dpaa2_eth_priv *priv);
-+
-+int set_rx_taildrop(struct dpaa2_eth_priv *priv);
-+
-+int dpaa2_eth_set_dist_key(struct dpaa2_eth_priv *priv,
-+ enum dpaa2_eth_rx_dist type, u32 key_fields);
-+
-+#endif /* __DPAA2_H */
---- /dev/null
-+++ b/drivers/staging/fsl-dpaa2/ethernet/dpaa2-ethtool.c
-@@ -0,0 +1,878 @@
-+/* Copyright 2014-2016 Freescale Semiconductor Inc.
-+ * Copyright 2016-2017 NXP
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#include "dpni.h" /* DPNI_LINK_OPT_* */
-+#include "dpaa2-eth.h"
-+
-+/* To be kept in sync with DPNI statistics */
-+static char dpaa2_ethtool_stats[][ETH_GSTRING_LEN] = {
-+ "[hw] rx frames",
-+ "[hw] rx bytes",
-+ "[hw] rx mcast frames",
-+ "[hw] rx mcast bytes",
-+ "[hw] rx bcast frames",
-+ "[hw] rx bcast bytes",
-+ "[hw] tx frames",
-+ "[hw] tx bytes",
-+ "[hw] tx mcast frames",
-+ "[hw] tx mcast bytes",
-+ "[hw] tx bcast frames",
-+ "[hw] tx bcast bytes",
-+ "[hw] rx filtered frames",
-+ "[hw] rx discarded frames",
-+ "[hw] rx nobuffer discards",
-+ "[hw] tx discarded frames",
-+ "[hw] tx confirmed frames",
-+};
-+
-+#define DPAA2_ETH_NUM_STATS ARRAY_SIZE(dpaa2_ethtool_stats)
-+
-+static char dpaa2_ethtool_extras[][ETH_GSTRING_LEN] = {
-+ /* per-cpu stats */
-+ "[drv] tx conf frames",
-+ "[drv] tx conf bytes",
-+ "[drv] tx sg frames",
-+ "[drv] tx sg bytes",
-+ "[drv] tx realloc frames",
-+ "[drv] rx sg frames",
-+ "[drv] rx sg bytes",
-+ "[drv] enqueue portal busy",
-+ /* Channel stats */
-+ "[drv] dequeue portal busy",
-+ "[drv] channel pull errors",
-+ "[drv] cdan",
-+ "[drv] tx congestion state",
-+#ifdef CONFIG_FSL_QBMAN_DEBUG
-+ /* FQ stats */
-+ "rx pending frames",
-+ "rx pending bytes",
-+ "tx conf pending frames",
-+ "tx conf pending bytes",
-+ "buffer count"
-+#endif
-+};
-+
-+#define DPAA2_ETH_NUM_EXTRA_STATS ARRAY_SIZE(dpaa2_ethtool_extras)
-+
-+static void dpaa2_eth_get_drvinfo(struct net_device *net_dev,
-+ struct ethtool_drvinfo *drvinfo)
-+{
-+ struct dpaa2_eth_priv *priv = netdev_priv(net_dev);
-+
-+ strlcpy(drvinfo->driver, KBUILD_MODNAME, sizeof(drvinfo->driver));
-+ strlcpy(drvinfo->version, dpaa2_eth_drv_version,
-+ sizeof(drvinfo->version));
-+
-+ snprintf(drvinfo->fw_version, sizeof(drvinfo->fw_version),
-+ "%u.%u", priv->dpni_ver_major, priv->dpni_ver_minor);
-+
-+ strlcpy(drvinfo->bus_info, dev_name(net_dev->dev.parent->parent),
-+ sizeof(drvinfo->bus_info));
-+}
-+
-+static int
-+dpaa2_eth_get_link_ksettings(struct net_device *net_dev,
-+ struct ethtool_link_ksettings *link_settings)
-+{
-+ struct dpni_link_state state = {0};
-+ int err = 0;
-+ struct dpaa2_eth_priv *priv = netdev_priv(net_dev);
-+
-+ err = dpni_get_link_state(priv->mc_io, 0, priv->mc_token, &state);
-+ if (err) {
-+ netdev_err(net_dev, "ERROR %d getting link state\n", err);
-+ goto out;
-+ }
-+
-+ /* At the moment, we have no way of interrogating the DPMAC
-+ * from the DPNI side - and for that matter there may exist
-+ * no DPMAC at all. So for now we just don't report anything
-+ * beyond the DPNI attributes.
-+ */
-+ if (state.options & DPNI_LINK_OPT_AUTONEG)
-+ link_settings->base.autoneg = AUTONEG_ENABLE;
-+ if (!(state.options & DPNI_LINK_OPT_HALF_DUPLEX))
-+ link_settings->base.duplex = DUPLEX_FULL;
-+ link_settings->base.speed = state.rate;
-+
-+out:
-+ return err;
-+}
-+
-+#define DPNI_DYNAMIC_LINK_SET_VER_MAJOR 7
-+#define DPNI_DYNAMIC_LINK_SET_VER_MINOR 1
-+static int
-+dpaa2_eth_set_link_ksettings(struct net_device *net_dev,
-+ const struct ethtool_link_ksettings *link_settings)
-+{
-+ struct dpaa2_eth_priv *priv = netdev_priv(net_dev);
-+ struct dpni_link_state state = {0};
-+ struct dpni_link_cfg cfg = {0};
-+ int err = 0;
-+
-+ /* If using an older MC version, the DPNI must be down
-+ * in order to be able to change link settings. Taking steps to let
-+ * the user know that.
-+ */
-+ if (dpaa2_eth_cmp_dpni_ver(priv, DPNI_DYNAMIC_LINK_SET_VER_MAJOR,
-+ DPNI_DYNAMIC_LINK_SET_VER_MINOR) < 0) {
-+ if (netif_running(net_dev)) {
-+ netdev_info(net_dev, "Interface must be brought down first.\n");
-+ return -EACCES;
-+ }
-+ }
-+
-+ /* Need to interrogate link state to get flow control params */
-+ err = dpni_get_link_state(priv->mc_io, 0, priv->mc_token, &state);
-+ if (err) {
-+ netdev_err(net_dev, "Error getting link state\n");
-+ goto out;
-+ }
-+
-+ cfg.options = state.options;
-+ cfg.rate = link_settings->base.speed;
-+ if (link_settings->base.autoneg == AUTONEG_ENABLE)
-+ cfg.options |= DPNI_LINK_OPT_AUTONEG;
-+ else
-+ cfg.options &= ~DPNI_LINK_OPT_AUTONEG;
-+ if (link_settings->base.duplex == DUPLEX_HALF)
-+ cfg.options |= DPNI_LINK_OPT_HALF_DUPLEX;
-+ else
-+ cfg.options &= ~DPNI_LINK_OPT_HALF_DUPLEX;
-+
-+ err = dpni_set_link_cfg(priv->mc_io, 0, priv->mc_token, &cfg);
-+ if (err)
-+ /* ethtool will be loud enough if we return an error; no point
-+ * in putting our own error message on the console by default
-+ */
-+ netdev_dbg(net_dev, "ERROR %d setting link cfg\n", err);
-+
-+out:
-+ return err;
-+}
-+
-+static void dpaa2_eth_get_pauseparam(struct net_device *net_dev,
-+ struct ethtool_pauseparam *pause)
-+{
-+ struct dpaa2_eth_priv *priv = netdev_priv(net_dev);
-+ struct dpni_link_state state = {0};
-+ int err;
-+
-+ err = dpni_get_link_state(priv->mc_io, 0, priv->mc_token, &state);
-+ if (err)
-+ netdev_dbg(net_dev, "Error getting link state\n");
-+
-+ /* Report general port autonegotiation status */
-+ pause->autoneg = !!(state.options & DPNI_LINK_OPT_AUTONEG);
-+ pause->rx_pause = !!(state.options & DPNI_LINK_OPT_PAUSE);
-+ pause->tx_pause = pause->rx_pause ^
-+ !!(state.options & DPNI_LINK_OPT_ASYM_PAUSE);
-+}
-+
-+static int dpaa2_eth_set_pauseparam(struct net_device *net_dev,
-+ struct ethtool_pauseparam *pause)
-+{
-+ struct dpaa2_eth_priv *priv = netdev_priv(net_dev);
-+ struct dpni_link_state state = {0};
-+ struct dpni_link_cfg cfg = {0};
-+ u32 current_tx_pause;
-+ int err = 0;
-+
-+ err = dpni_get_link_state(priv->mc_io, 0, priv->mc_token, &state);
-+ if (err) {
-+ netdev_dbg(net_dev, "Error getting link state\n");
-+ goto out;
-+ }
-+
-+ cfg.rate = state.rate;
-+ cfg.options = state.options;
-+ current_tx_pause = !!(cfg.options & DPNI_LINK_OPT_PAUSE) ^
-+ !!(cfg.options & DPNI_LINK_OPT_ASYM_PAUSE);
-+
-+ /* We don't support changing pause frame autonegotiation separately
-+ * from general port autoneg
-+ */
-+ if (pause->autoneg != !!(state.options & DPNI_LINK_OPT_AUTONEG))
-+ netdev_warn(net_dev,
-+ "Cannot change pause frame autoneg separately\n");
-+
-+ if (pause->rx_pause)
-+ cfg.options |= DPNI_LINK_OPT_PAUSE;
-+ else
-+ cfg.options &= ~DPNI_LINK_OPT_PAUSE;
-+
-+ if (pause->rx_pause ^ pause->tx_pause)
-+ cfg.options |= DPNI_LINK_OPT_ASYM_PAUSE;
-+ else
-+ cfg.options &= ~DPNI_LINK_OPT_ASYM_PAUSE;
-+
-+ err = dpni_set_link_cfg(priv->mc_io, 0, priv->mc_token, &cfg);
-+ if (err) {
-+ netdev_dbg(net_dev, "Error setting link\n");
-+ goto out;
-+ }
-+
-+ /* Enable/disable Rx FQ taildrop if Tx pause frames have changed */
-+ if (current_tx_pause == pause->tx_pause)
-+ goto out;
-+
-+ priv->tx_pause_frames = pause->tx_pause;
-+ err = set_rx_taildrop(priv);
-+ if (err)
-+ netdev_dbg(net_dev, "Error configuring taildrop\n");
-+
-+out:
-+ return err;
-+}
-+
-+static void dpaa2_eth_get_strings(struct net_device *netdev, u32 stringset,
-+ u8 *data)
-+{
-+ u8 *p = data;
-+ int i;
-+
-+ switch (stringset) {
-+ case ETH_SS_STATS:
-+ for (i = 0; i < DPAA2_ETH_NUM_STATS; i++) {
-+ strlcpy(p, dpaa2_ethtool_stats[i], ETH_GSTRING_LEN);
-+ p += ETH_GSTRING_LEN;
-+ }
-+ for (i = 0; i < DPAA2_ETH_NUM_EXTRA_STATS; i++) {
-+ strlcpy(p, dpaa2_ethtool_extras[i], ETH_GSTRING_LEN);
-+ p += ETH_GSTRING_LEN;
-+ }
-+ break;
-+ }
-+}
-+
-+static int dpaa2_eth_get_sset_count(struct net_device *net_dev, int sset)
-+{
-+ switch (sset) {
-+ case ETH_SS_STATS: /* ethtool_get_stats(), ethtool_get_drvinfo() */
-+ return DPAA2_ETH_NUM_STATS + DPAA2_ETH_NUM_EXTRA_STATS;
-+ default:
-+ return -EOPNOTSUPP;
-+ }
-+}
-+
-+/** Fill in hardware counters, as returned by MC.
-+ */
-+static void dpaa2_eth_get_ethtool_stats(struct net_device *net_dev,
-+ struct ethtool_stats *stats,
-+ u64 *data)
-+{
-+ int i = 0;
-+ int j, k, err;
-+ int num_cnt;
-+ union dpni_statistics dpni_stats;
-+
-+#ifdef CONFIG_FSL_QBMAN_DEBUG
-+ u32 fcnt, bcnt;
-+ u32 fcnt_rx_total = 0, fcnt_tx_total = 0;
-+ u32 bcnt_rx_total = 0, bcnt_tx_total = 0;
-+ u32 buf_cnt;
-+#endif
-+ u64 cdan = 0;
-+ u64 portal_busy = 0, pull_err = 0;
-+ struct dpaa2_eth_priv *priv = netdev_priv(net_dev);
-+ struct dpaa2_eth_drv_stats *extras;
-+ struct dpaa2_eth_ch_stats *ch_stats;
-+
-+ memset(data, 0,
-+ sizeof(u64) * (DPAA2_ETH_NUM_STATS + DPAA2_ETH_NUM_EXTRA_STATS));
-+
-+ /* Print standard counters, from DPNI statistics */
-+ for (j = 0; j <= 2; j++) {
-+ err = dpni_get_statistics(priv->mc_io, 0, priv->mc_token,
-+ j, 0, &dpni_stats);
-+ if (err != 0)
-+ netdev_warn(net_dev, "dpni_get_stats(%d) failed\n", j);
-+ switch (j) {
-+ case 0:
-+ num_cnt = sizeof(dpni_stats.page_0) / sizeof(u64);
-+ break;
-+ case 1:
-+ num_cnt = sizeof(dpni_stats.page_1) / sizeof(u64);
-+ break;
-+ case 2:
-+ num_cnt = sizeof(dpni_stats.page_2) / sizeof(u64);
-+ break;
-+ }
-+ for (k = 0; k < num_cnt; k++)
-+ *(data + i++) = dpni_stats.raw.counter[k];
-+ }
-+
-+ /* Print per-cpu extra stats */
-+ for_each_online_cpu(k) {
-+ extras = per_cpu_ptr(priv->percpu_extras, k);
-+ for (j = 0; j < sizeof(*extras) / sizeof(__u64); j++)
-+ *((__u64 *)data + i + j) += *((__u64 *)extras + j);
-+ }
-+ i += j;
-+
-+ for (j = 0; j < priv->num_channels; j++) {
-+ ch_stats = &priv->channel[j]->stats;
-+ cdan += ch_stats->cdan;
-+ portal_busy += ch_stats->dequeue_portal_busy;
-+ pull_err += ch_stats->pull_err;
-+ }
-+
-+ *(data + i++) = portal_busy;
-+ *(data + i++) = pull_err;
-+ *(data + i++) = cdan;
-+
-+ *(data + i++) = dpaa2_cscn_state_congested(priv->cscn_mem);
-+
-+#ifdef CONFIG_FSL_QBMAN_DEBUG
-+ for (j = 0; j < priv->num_fqs; j++) {
-+ /* Print FQ instantaneous counts */
-+ err = dpaa2_io_query_fq_count(NULL, priv->fq[j].fqid,
-+ &fcnt, &bcnt);
-+ if (err) {
-+ netdev_warn(net_dev, "FQ query error %d", err);
-+ return;
-+ }
-+
-+ if (priv->fq[j].type == DPAA2_TX_CONF_FQ) {
-+ fcnt_tx_total += fcnt;
-+ bcnt_tx_total += bcnt;
-+ } else {
-+ fcnt_rx_total += fcnt;
-+ bcnt_rx_total += bcnt;
-+ }
-+ }
-+
-+ *(data + i++) = fcnt_rx_total;
-+ *(data + i++) = bcnt_rx_total;
-+ *(data + i++) = fcnt_tx_total;
-+ *(data + i++) = bcnt_tx_total;
-+
-+ err = dpaa2_io_query_bp_count(NULL, priv->bpid, &buf_cnt);
-+ if (err) {
-+ netdev_warn(net_dev, "Buffer count query error %d\n", err);
-+ return;
-+ }
-+ *(data + i++) = buf_cnt;
-+#endif
-+}
-+
-+static int cls_key_off(struct dpaa2_eth_priv *priv, int prot, int field)
-+{
-+ int i, off = 0;
-+
-+ for (i = 0; i < priv->num_dist_fields; i++) {
-+ if (priv->dist_fields[i].cls_prot == prot &&
-+ priv->dist_fields[i].cls_field == field)
-+ return off;
-+ off += priv->dist_fields[i].size;
-+ }
-+
-+ return -1;
-+}
-+
-+static u8 cls_key_size(struct dpaa2_eth_priv *priv)
-+{
-+ u8 i, size = 0;
-+
-+ for (i = 0; i < priv->num_dist_fields; i++)
-+ size += priv->dist_fields[i].size;
-+
-+ return size;
-+}
-+
-+void check_cls_support(struct dpaa2_eth_priv *priv)
-+{
-+ u8 key_size = cls_key_size(priv);
-+ struct device *dev = priv->net_dev->dev.parent;
-+
-+ if (dpaa2_eth_hash_enabled(priv)) {
-+ if (priv->dpni_attrs.fs_key_size < key_size) {
-+ dev_info(dev, "max_dist_key_size = %d, expected %d. Hashing and steering are disabled\n",
-+ priv->dpni_attrs.fs_key_size,
-+ key_size);
-+ goto disable_fs;
-+ }
-+ if (priv->num_dist_fields > DPKG_MAX_NUM_OF_EXTRACTS) {
-+ dev_info(dev, "Too many key fields (max = %d). Hashing and steering are disabled\n",
-+ DPKG_MAX_NUM_OF_EXTRACTS);
-+ goto disable_fs;
-+ }
-+ }
-+
-+ if (dpaa2_eth_fs_enabled(priv)) {
-+ if (!dpaa2_eth_hash_enabled(priv)) {
-+ dev_info(dev, "Insufficient queues. Steering is disabled\n");
-+ goto disable_fs;
-+ }
-+
-+ if (!dpaa2_eth_fs_mask_enabled(priv)) {
-+ dev_info(dev, "Key masks not supported. Steering is disabled\n");
-+ goto disable_fs;
-+ }
-+ }
-+
-+ return;
-+
-+disable_fs:
-+ priv->dpni_attrs.options |= DPNI_OPT_NO_FS;
-+ priv->dpni_attrs.options &= ~DPNI_OPT_HAS_KEY_MASKING;
-+}
-+
-+static int prep_l4_rule(struct dpaa2_eth_priv *priv,
-+ struct ethtool_tcpip4_spec *l4_value,
-+ struct ethtool_tcpip4_spec *l4_mask,
-+ void *key, void *mask, u8 l4_proto)
-+{
-+ int offset;
-+
-+ if (l4_mask->tos) {
-+ netdev_err(priv->net_dev, "ToS is not supported for IPv4 L4\n");
-+ return -EOPNOTSUPP;
-+ }
-+
-+ if (l4_mask->ip4src) {
-+ offset = cls_key_off(priv, NET_PROT_IP, NH_FLD_IP_SRC);
-+ *(u32 *)(key + offset) = l4_value->ip4src;
-+ *(u32 *)(mask + offset) = l4_mask->ip4src;
-+ }
-+
-+ if (l4_mask->ip4dst) {
-+ offset = cls_key_off(priv, NET_PROT_IP, NH_FLD_IP_DST);
-+ *(u32 *)(key + offset) = l4_value->ip4dst;
-+ *(u32 *)(mask + offset) = l4_mask->ip4dst;
-+ }
-+
-+ if (l4_mask->psrc) {
-+ offset = cls_key_off(priv, NET_PROT_UDP, NH_FLD_UDP_PORT_SRC);
-+ *(u32 *)(key + offset) = l4_value->psrc;
-+ *(u32 *)(mask + offset) = l4_mask->psrc;
-+ }
-+
-+ if (l4_mask->pdst) {
-+ offset = cls_key_off(priv, NET_PROT_UDP, NH_FLD_UDP_PORT_DST);
-+ *(u32 *)(key + offset) = l4_value->pdst;
-+ *(u32 *)(mask + offset) = l4_mask->pdst;
-+ }
-+
-+ /* Only apply the rule for the user-specified L4 protocol
-+ * and if ethertype matches IPv4
-+ */
-+ offset = cls_key_off(priv, NET_PROT_ETH, NH_FLD_ETH_TYPE);
-+ *(u16 *)(key + offset) = htons(ETH_P_IP);
-+ *(u16 *)(mask + offset) = 0xFFFF;
-+
-+ offset = cls_key_off(priv, NET_PROT_IP, NH_FLD_IP_PROTO);
-+ *(u8 *)(key + offset) = l4_proto;
-+ *(u8 *)(mask + offset) = 0xFF;
-+
-+ /* TODO: check IP version */
-+
-+ return 0;
-+}
-+
-+static int prep_eth_rule(struct dpaa2_eth_priv *priv,
-+ struct ethhdr *eth_value, struct ethhdr *eth_mask,
-+ void *key, void *mask)
-+{
-+ int offset;
-+
-+ if (eth_mask->h_proto) {
-+ netdev_err(priv->net_dev, "Ethertype is not supported!\n");
-+ return -EOPNOTSUPP;
-+ }
-+
-+ if (!is_zero_ether_addr(eth_mask->h_source)) {
-+ offset = cls_key_off(priv, NET_PROT_ETH, NH_FLD_ETH_SA);
-+ ether_addr_copy(key + offset, eth_value->h_source);
-+ ether_addr_copy(mask + offset, eth_mask->h_source);
-+ }
-+
-+ if (!is_zero_ether_addr(eth_mask->h_dest)) {
-+ offset = cls_key_off(priv, NET_PROT_ETH, NH_FLD_ETH_DA);
-+ ether_addr_copy(key + offset, eth_value->h_dest);
-+ ether_addr_copy(mask + offset, eth_mask->h_dest);
-+ }
-+
-+ return 0;
-+}
-+
-+static int prep_user_ip_rule(struct dpaa2_eth_priv *priv,
-+ struct ethtool_usrip4_spec *uip_value,
-+ struct ethtool_usrip4_spec *uip_mask,
-+ void *key, void *mask)
-+{
-+ int offset;
-+
-+ if (uip_mask->tos)
-+ return -EOPNOTSUPP;
-+
-+ if (uip_mask->ip4src) {
-+ offset = cls_key_off(priv, NET_PROT_IP, NH_FLD_IP_SRC);
-+ *(u32 *)(key + offset) = uip_value->ip4src;
-+ *(u32 *)(mask + offset) = uip_mask->ip4src;
-+ }
-+
-+ if (uip_mask->ip4dst) {
-+ offset = cls_key_off(priv, NET_PROT_IP, NH_FLD_IP_DST);
-+ *(u32 *)(key + offset) = uip_value->ip4dst;
-+ *(u32 *)(mask + offset) = uip_mask->ip4dst;
-+ }
-+
-+ if (uip_mask->proto) {
-+ offset = cls_key_off(priv, NET_PROT_IP, NH_FLD_IP_PROTO);
-+ *(u32 *)(key + offset) = uip_value->proto;
-+ *(u32 *)(mask + offset) = uip_mask->proto;
-+ }
-+ if (uip_mask->l4_4_bytes) {
-+ offset = cls_key_off(priv, NET_PROT_UDP, NH_FLD_UDP_PORT_SRC);
-+ *(u16 *)(key + offset) = uip_value->l4_4_bytes << 16;
-+ *(u16 *)(mask + offset) = uip_mask->l4_4_bytes << 16;
-+
-+ offset = cls_key_off(priv, NET_PROT_UDP, NH_FLD_UDP_PORT_DST);
-+ *(u16 *)(key + offset) = uip_value->l4_4_bytes & 0xFFFF;
-+ *(u16 *)(mask + offset) = uip_mask->l4_4_bytes & 0xFFFF;
-+ }
-+
-+ /* Ethertype must be IP */
-+ offset = cls_key_off(priv, NET_PROT_ETH, NH_FLD_ETH_TYPE);
-+ *(u16 *)(key + offset) = htons(ETH_P_IP);
-+ *(u16 *)(mask + offset) = 0xFFFF;
-+
-+ return 0;
-+}
-+
-+static int prep_ext_rule(struct dpaa2_eth_priv *priv,
-+ struct ethtool_flow_ext *ext_value,
-+ struct ethtool_flow_ext *ext_mask,
-+ void *key, void *mask)
-+{
-+ int offset;
-+
-+ if (ext_mask->vlan_etype)
-+ return -EOPNOTSUPP;
-+
-+ if (ext_mask->vlan_tci) {
-+ offset = cls_key_off(priv, NET_PROT_VLAN, NH_FLD_VLAN_TCI);
-+ *(u16 *)(key + offset) = ext_value->vlan_tci;
-+ *(u16 *)(mask + offset) = ext_mask->vlan_tci;
-+ }
-+
-+ return 0;
-+}
-+
-+static int prep_mac_ext_rule(struct dpaa2_eth_priv *priv,
-+ struct ethtool_flow_ext *ext_value,
-+ struct ethtool_flow_ext *ext_mask,
-+ void *key, void *mask)
-+{
-+ int offset;
-+
-+ if (!is_zero_ether_addr(ext_mask->h_dest)) {
-+ offset = cls_key_off(priv, NET_PROT_ETH, NH_FLD_ETH_DA);
-+ ether_addr_copy(key + offset, ext_value->h_dest);
-+ ether_addr_copy(mask + offset, ext_mask->h_dest);
-+ }
-+
-+ return 0;
-+}
-+
-+static int prep_cls_rule(struct net_device *net_dev,
-+ struct ethtool_rx_flow_spec *fs,
-+ void *key)
-+{
-+ struct dpaa2_eth_priv *priv = netdev_priv(net_dev);
-+ const u8 key_size = cls_key_size(priv);
-+ void *msk = key + key_size;
-+ int err;
-+
-+ memset(key, 0, key_size * 2);
-+
-+ switch (fs->flow_type & 0xff) {
-+ case TCP_V4_FLOW:
-+ err = prep_l4_rule(priv, &fs->h_u.tcp_ip4_spec,
-+ &fs->m_u.tcp_ip4_spec, key, msk,
-+ IPPROTO_TCP);
-+ break;
-+ case UDP_V4_FLOW:
-+ err = prep_l4_rule(priv, &fs->h_u.udp_ip4_spec,
-+ &fs->m_u.udp_ip4_spec, key, msk,
-+ IPPROTO_UDP);
-+ break;
-+ case SCTP_V4_FLOW:
-+ err = prep_l4_rule(priv, &fs->h_u.sctp_ip4_spec,
-+ &fs->m_u.sctp_ip4_spec, key, msk,
-+ IPPROTO_SCTP);
-+ break;
-+ case ETHER_FLOW:
-+ err = prep_eth_rule(priv, &fs->h_u.ether_spec,
-+ &fs->m_u.ether_spec, key, msk);
-+ break;
-+ case IP_USER_FLOW:
-+ err = prep_user_ip_rule(priv, &fs->h_u.usr_ip4_spec,
-+ &fs->m_u.usr_ip4_spec, key, msk);
-+ break;
-+ default:
-+ /* TODO: AH, ESP */
-+ return -EOPNOTSUPP;
-+ }
-+ if (err)
-+ return err;
-+
-+ if (fs->flow_type & FLOW_EXT) {
-+ err = prep_ext_rule(priv, &fs->h_ext, &fs->m_ext, key, msk);
-+ if (err)
-+ return err;
-+ }
-+
-+ if (fs->flow_type & FLOW_MAC_EXT) {
-+ err = prep_mac_ext_rule(priv, &fs->h_ext, &fs->m_ext, key, msk);
-+ if (err)
-+ return err;
-+ }
-+
-+ return 0;
-+}
-+
-+static int del_cls(struct net_device *net_dev, int location);
-+
-+static int do_cls(struct net_device *net_dev,
-+ struct ethtool_rx_flow_spec *fs,
-+ bool add)
-+{
-+ struct dpaa2_eth_priv *priv = netdev_priv(net_dev);
-+ struct device *dev = net_dev->dev.parent;
-+ const int rule_cnt = dpaa2_eth_fs_count(priv);
-+ struct dpni_rule_cfg rule_cfg;
-+ struct dpni_fs_action_cfg fs_act = { 0 };
-+ void *dma_mem;
-+ int err = 0, tc;
-+
-+ if (!dpaa2_eth_fs_enabled(priv)) {
-+ netdev_err(net_dev, "dev does not support steering!\n");
-+ /* dev doesn't support steering */
-+ return -EOPNOTSUPP;
-+ }
-+
-+ if ((fs->ring_cookie != RX_CLS_FLOW_DISC &&
-+ fs->ring_cookie >= dpaa2_eth_queue_count(priv)) ||
-+ fs->location >= rule_cnt)
-+ return -EINVAL;
-+
-+ /* When adding a new rule, check if location if available
-+ * and if not, free the existing table entry before inserting
-+ * the new one
-+ */
-+ if (add && (priv->cls_rule[fs->location].in_use == true))
-+ del_cls(net_dev, fs->location);
-+
-+ memset(&rule_cfg, 0, sizeof(rule_cfg));
-+ rule_cfg.key_size = cls_key_size(priv);
-+
-+ /* allocate twice the key size, for the actual key and for mask */
-+ dma_mem = kzalloc(rule_cfg.key_size * 2, GFP_DMA | GFP_KERNEL);
-+ if (!dma_mem)
-+ return -ENOMEM;
-+
-+ err = prep_cls_rule(net_dev, fs, dma_mem);
-+ if (err)
-+ goto err_free_mem;
-+
-+ rule_cfg.key_iova = dma_map_single(dev, dma_mem,
-+ rule_cfg.key_size * 2,
-+ DMA_TO_DEVICE);
-+
-+ rule_cfg.mask_iova = rule_cfg.key_iova + rule_cfg.key_size;
-+
-+ if (fs->ring_cookie == RX_CLS_FLOW_DISC)
-+ fs_act.options |= DPNI_FS_OPT_DISCARD;
-+ else
-+ fs_act.flow_id = fs->ring_cookie;
-+
-+ for (tc = 0; tc < dpaa2_eth_tc_count(priv); tc++) {
-+ if (add)
-+ err = dpni_add_fs_entry(priv->mc_io, 0, priv->mc_token,
-+ tc, fs->location, &rule_cfg,
-+ &fs_act);
-+ else
-+ err = dpni_remove_fs_entry(priv->mc_io, 0,
-+ priv->mc_token, tc,
-+ &rule_cfg);
-+
-+ if (err)
-+ break;
-+ }
-+
-+ dma_unmap_single(dev, rule_cfg.key_iova,
-+ rule_cfg.key_size * 2, DMA_TO_DEVICE);
-+
-+ if (err)
-+ netdev_err(net_dev, "dpaa2_add/remove_cls() error %d\n", err);
-+
-+err_free_mem:
-+ kfree(dma_mem);
-+
-+ return err;
-+}
-+
-+static int add_cls(struct net_device *net_dev,
-+ struct ethtool_rx_flow_spec *fs)
-+{
-+ struct dpaa2_eth_priv *priv = netdev_priv(net_dev);
-+ int err;
-+
-+ err = do_cls(net_dev, fs, true);
-+ if (err)
-+ return err;
-+
-+ priv->cls_rule[fs->location].in_use = true;
-+ priv->cls_rule[fs->location].fs = *fs;
-+
-+ return 0;
-+}
-+
-+static int del_cls(struct net_device *net_dev, int location)
-+{
-+ struct dpaa2_eth_priv *priv = netdev_priv(net_dev);
-+ int err;
-+
-+ err = do_cls(net_dev, &priv->cls_rule[location].fs, false);
-+ if (err)
-+ return err;
-+
-+ priv->cls_rule[location].in_use = false;
-+
-+ return 0;
-+}
-+
-+static int set_hash(struct net_device *net_dev, u64 data)
-+{
-+ struct dpaa2_eth_priv *priv = netdev_priv(net_dev);
-+ u32 key = 0;
-+ int i;
-+
-+ if (data & RXH_DISCARD)
-+ return -EOPNOTSUPP;
-+
-+ for (i = 0; i < priv->num_dist_fields; i++)
-+ if (priv->dist_fields[i].rxnfc_field & data)
-+ key |= priv->dist_fields[i].id;
-+
-+ return dpaa2_eth_set_dist_key(priv, DPAA2_ETH_RX_DIST_HASH, key);
-+}
-+
-+static int dpaa2_eth_set_rxnfc(struct net_device *net_dev,
-+ struct ethtool_rxnfc *rxnfc)
-+{
-+ int err = 0;
-+
-+ switch (rxnfc->cmd) {
-+ case ETHTOOL_SRXCLSRLINS:
-+ err = add_cls(net_dev, &rxnfc->fs);
-+ break;
-+ case ETHTOOL_SRXCLSRLDEL:
-+ err = del_cls(net_dev, rxnfc->fs.location);
-+ break;
-+ case ETHTOOL_SRXFH:
-+ err = set_hash(net_dev, rxnfc->data);
-+ break;
-+ default:
-+ err = -EOPNOTSUPP;
-+ }
-+
-+ return err;
-+}
-+
-+static int dpaa2_eth_get_rxnfc(struct net_device *net_dev,
-+ struct ethtool_rxnfc *rxnfc, u32 *rule_locs)
-+{
-+ struct dpaa2_eth_priv *priv = netdev_priv(net_dev);
-+ const int rule_cnt = dpaa2_eth_fs_count(priv);
-+ int i, j;
-+
-+ switch (rxnfc->cmd) {
-+ case ETHTOOL_GRXFH:
-+ /* we purposely ignore cmd->flow_type for now, because the
-+ * classifier only supports a single set of fields for all
-+ * protocols
-+ */
-+ rxnfc->data = priv->rx_hash_fields;
-+ break;
-+ case ETHTOOL_GRXRINGS:
-+ rxnfc->data = dpaa2_eth_queue_count(priv);
-+ break;
-+
-+ case ETHTOOL_GRXCLSRLCNT:
-+ for (i = 0, rxnfc->rule_cnt = 0; i < rule_cnt; i++)
-+ if (priv->cls_rule[i].in_use)
-+ rxnfc->rule_cnt++;
-+ rxnfc->data = rule_cnt;
-+ break;
-+
-+ case ETHTOOL_GRXCLSRULE:
-+ if (!priv->cls_rule[rxnfc->fs.location].in_use)
-+ return -EINVAL;
-+
-+ rxnfc->fs = priv->cls_rule[rxnfc->fs.location].fs;
-+ break;
-+
-+ case ETHTOOL_GRXCLSRLALL:
-+ for (i = 0, j = 0; i < rule_cnt; i++) {
-+ if (!priv->cls_rule[i].in_use)
-+ continue;
-+ if (j == rxnfc->rule_cnt)
-+ return -EMSGSIZE;
-+ rule_locs[j++] = i;
-+ }
-+ rxnfc->rule_cnt = j;
-+ rxnfc->data = rule_cnt;
-+ break;
-+
-+ default:
-+ return -EOPNOTSUPP;
-+ }
-+
-+ return 0;
-+}
-+
-+const struct ethtool_ops dpaa2_ethtool_ops = {
-+ .get_drvinfo = dpaa2_eth_get_drvinfo,
-+ .get_link = ethtool_op_get_link,
-+ .get_link_ksettings = dpaa2_eth_get_link_ksettings,
-+ .set_link_ksettings = dpaa2_eth_set_link_ksettings,
-+ .get_pauseparam = dpaa2_eth_get_pauseparam,
-+ .set_pauseparam = dpaa2_eth_set_pauseparam,
-+ .get_sset_count = dpaa2_eth_get_sset_count,
-+ .get_ethtool_stats = dpaa2_eth_get_ethtool_stats,
-+ .get_strings = dpaa2_eth_get_strings,
-+ .get_rxnfc = dpaa2_eth_get_rxnfc,
-+ .set_rxnfc = dpaa2_eth_set_rxnfc,
-+};
---- /dev/null
-+++ b/drivers/staging/fsl-dpaa2/ethernet/dpkg.h
-@@ -0,0 +1,176 @@
-+/* Copyright 2013-2015 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of the above-listed copyright holders nor the
-+ * names of any contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
-+ * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
-+ * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
-+ * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE
-+ * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
-+ * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
-+ * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
-+ * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
-+ * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
-+ * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
-+ * POSSIBILITY OF SUCH DAMAGE.
-+ */
-+#ifndef __FSL_DPKG_H_
-+#define __FSL_DPKG_H_
-+
-+#include <linux/types.h>
-+#include "net.h"
-+
-+/* Data Path Key Generator API
-+ * Contains initialization APIs and runtime APIs for the Key Generator
-+ */
-+
-+/** Key Generator properties */
-+
-+/**
-+ * Number of masks per key extraction
-+ */
-+#define DPKG_NUM_OF_MASKS 4
-+/**
-+ * Number of extractions per key profile
-+ */
-+#define DPKG_MAX_NUM_OF_EXTRACTS 10
-+
-+/**
-+ * enum dpkg_extract_from_hdr_type - Selecting extraction by header types
-+ * @DPKG_FROM_HDR: Extract selected bytes from header, by offset
-+ * @DPKG_FROM_FIELD: Extract selected bytes from header, by offset from field
-+ * @DPKG_FULL_FIELD: Extract a full field
-+ */
-+enum dpkg_extract_from_hdr_type {
-+ DPKG_FROM_HDR = 0,
-+ DPKG_FROM_FIELD = 1,
-+ DPKG_FULL_FIELD = 2
-+};
-+
-+/**
-+ * enum dpkg_extract_type - Enumeration for selecting extraction type
-+ * @DPKG_EXTRACT_FROM_HDR: Extract from the header
-+ * @DPKG_EXTRACT_FROM_DATA: Extract from data not in specific header
-+ * @DPKG_EXTRACT_FROM_PARSE: Extract from parser-result;
-+ * e.g. can be used to extract header existence;
-+ * please refer to 'Parse Result definition' section in the parser BG
-+ */
-+enum dpkg_extract_type {
-+ DPKG_EXTRACT_FROM_HDR = 0,
-+ DPKG_EXTRACT_FROM_DATA = 1,
-+ DPKG_EXTRACT_FROM_PARSE = 3
-+};
-+
-+/**
-+ * struct dpkg_mask - A structure for defining a single extraction mask
-+ * @mask: Byte mask for the extracted content
-+ * @offset: Offset within the extracted content
-+ */
-+struct dpkg_mask {
-+ u8 mask;
-+ u8 offset;
-+};
-+
-+/**
-+ * struct dpkg_extract - A structure for defining a single extraction
-+ * @type: Determines how the union below is interpreted:
-+ * DPKG_EXTRACT_FROM_HDR: selects 'from_hdr';
-+ * DPKG_EXTRACT_FROM_DATA: selects 'from_data';
-+ * DPKG_EXTRACT_FROM_PARSE: selects 'from_parse'
-+ * @extract: Selects extraction method
-+ * @num_of_byte_masks: Defines the number of valid entries in the array below;
-+ * This is also the number of bytes to be used as masks
-+ * @masks: Masks parameters
-+ */
-+struct dpkg_extract {
-+ enum dpkg_extract_type type;
-+ /**
-+ * union extract - Selects extraction method
-+ * @from_hdr - Used when 'type = DPKG_EXTRACT_FROM_HDR'
-+ * @from_data - Used when 'type = DPKG_EXTRACT_FROM_DATA'
-+ * @from_parse - Used when 'type = DPKG_EXTRACT_FROM_PARSE'
-+ */
-+ union {
-+ /**
-+ * struct from_hdr - Used when 'type = DPKG_EXTRACT_FROM_HDR'
-+ * @prot: Any of the supported headers
-+ * @type: Defines the type of header extraction:
-+ * DPKG_FROM_HDR: use size & offset below;
-+ * DPKG_FROM_FIELD: use field, size and offset below;
-+ * DPKG_FULL_FIELD: use field below
-+ * @field: One of the supported fields (NH_FLD_)
-+ *
-+ * @size: Size in bytes
-+ * @offset: Byte offset
-+ * @hdr_index: Clear for cases not listed below;
-+ * Used for protocols that may have more than a single
-+ * header, 0 indicates an outer header;
-+ * Supported protocols (possible values):
-+ * NET_PROT_VLAN (0, HDR_INDEX_LAST);
-+ * NET_PROT_MPLS (0, 1, HDR_INDEX_LAST);
-+ * NET_PROT_IP(0, HDR_INDEX_LAST);
-+ * NET_PROT_IPv4(0, HDR_INDEX_LAST);
-+ * NET_PROT_IPv6(0, HDR_INDEX_LAST);
-+ */
-+
-+ struct {
-+ enum net_prot prot;
-+ enum dpkg_extract_from_hdr_type type;
-+ u32 field;
-+ u8 size;
-+ u8 offset;
-+ u8 hdr_index;
-+ } from_hdr;
-+ /**
-+ * struct from_data - Used when 'type = DPKG_EXTRACT_FROM_DATA'
-+ * @size: Size in bytes
-+ * @offset: Byte offset
-+ */
-+ struct {
-+ u8 size;
-+ u8 offset;
-+ } from_data;
-+
-+ /**
-+ * struct from_parse - Used when
-+ * 'type = DPKG_EXTRACT_FROM_PARSE'
-+ * @size: Size in bytes
-+ * @offset: Byte offset
-+ */
-+ struct {
-+ u8 size;
-+ u8 offset;
-+ } from_parse;
-+ } extract;
-+
-+ u8 num_of_byte_masks;
-+ struct dpkg_mask masks[DPKG_NUM_OF_MASKS];
-+};
-+
-+/**
-+ * struct dpkg_profile_cfg - A structure for defining a full Key Generation
-+ * profile (rule)
-+ * @num_extracts: Defines the number of valid entries in the array below
-+ * @extracts: Array of required extractions
-+ */
-+struct dpkg_profile_cfg {
-+ u8 num_extracts;
-+ struct dpkg_extract extracts[DPKG_MAX_NUM_OF_EXTRACTS];
-+};
-+
-+#endif /* __FSL_DPKG_H_ */
---- /dev/null
-+++ b/drivers/staging/fsl-dpaa2/ethernet/dpni-cmd.h
-@@ -0,0 +1,719 @@
-+/* Copyright 2013-2016 Freescale Semiconductor Inc.
-+ * Copyright 2016 NXP
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of the above-listed copyright holders nor the
-+ * names of any contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
-+ * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
-+ * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
-+ * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE
-+ * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
-+ * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
-+ * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
-+ * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
-+ * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
-+ * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
-+ * POSSIBILITY OF SUCH DAMAGE.
-+ */
-+#ifndef _FSL_DPNI_CMD_H
-+#define _FSL_DPNI_CMD_H
-+
-+#include "dpni.h"
-+
-+/* DPNI Version */
-+#define DPNI_VER_MAJOR 7
-+#define DPNI_VER_MINOR 0
-+#define DPNI_CMD_BASE_VERSION 1
-+#define DPNI_CMD_2ND_VERSION 2
-+#define DPNI_CMD_ID_OFFSET 4
-+
-+#define DPNI_CMD(id) (((id) << DPNI_CMD_ID_OFFSET) | DPNI_CMD_BASE_VERSION)
-+#define DPNI_CMD_V2(id) (((id) << DPNI_CMD_ID_OFFSET) | DPNI_CMD_2ND_VERSION)
-+
-+#define DPNI_CMDID_OPEN DPNI_CMD(0x801)
-+#define DPNI_CMDID_CLOSE DPNI_CMD(0x800)
-+#define DPNI_CMDID_CREATE DPNI_CMD(0x901)
-+#define DPNI_CMDID_DESTROY DPNI_CMD(0x900)
-+#define DPNI_CMDID_GET_API_VERSION DPNI_CMD(0xa01)
-+
-+#define DPNI_CMDID_ENABLE DPNI_CMD(0x002)
-+#define DPNI_CMDID_DISABLE DPNI_CMD(0x003)
-+#define DPNI_CMDID_GET_ATTR DPNI_CMD(0x004)
-+#define DPNI_CMDID_RESET DPNI_CMD(0x005)
-+#define DPNI_CMDID_IS_ENABLED DPNI_CMD(0x006)
-+
-+#define DPNI_CMDID_SET_IRQ DPNI_CMD(0x010)
-+#define DPNI_CMDID_GET_IRQ DPNI_CMD(0x011)
-+#define DPNI_CMDID_SET_IRQ_ENABLE DPNI_CMD(0x012)
-+#define DPNI_CMDID_GET_IRQ_ENABLE DPNI_CMD(0x013)
-+#define DPNI_CMDID_SET_IRQ_MASK DPNI_CMD(0x014)
-+#define DPNI_CMDID_GET_IRQ_MASK DPNI_CMD(0x015)
-+#define DPNI_CMDID_GET_IRQ_STATUS DPNI_CMD(0x016)
-+#define DPNI_CMDID_CLEAR_IRQ_STATUS DPNI_CMD(0x017)
-+
-+#define DPNI_CMDID_SET_POOLS DPNI_CMD_V2(0x200)
-+#define DPNI_CMDID_SET_ERRORS_BEHAVIOR DPNI_CMD(0x20B)
-+
-+#define DPNI_CMDID_GET_QDID DPNI_CMD(0x210)
-+#define DPNI_CMDID_GET_TX_DATA_OFFSET DPNI_CMD(0x212)
-+#define DPNI_CMDID_GET_LINK_STATE DPNI_CMD(0x215)
-+#define DPNI_CMDID_SET_MAX_FRAME_LENGTH DPNI_CMD(0x216)
-+#define DPNI_CMDID_GET_MAX_FRAME_LENGTH DPNI_CMD(0x217)
-+#define DPNI_CMDID_SET_LINK_CFG DPNI_CMD(0x21A)
-+#define DPNI_CMDID_SET_TX_SHAPING DPNI_CMD_V2(0x21B)
-+
-+#define DPNI_CMDID_SET_MCAST_PROMISC DPNI_CMD(0x220)
-+#define DPNI_CMDID_GET_MCAST_PROMISC DPNI_CMD(0x221)
-+#define DPNI_CMDID_SET_UNICAST_PROMISC DPNI_CMD(0x222)
-+#define DPNI_CMDID_GET_UNICAST_PROMISC DPNI_CMD(0x223)
-+#define DPNI_CMDID_SET_PRIM_MAC DPNI_CMD(0x224)
-+#define DPNI_CMDID_GET_PRIM_MAC DPNI_CMD(0x225)
-+#define DPNI_CMDID_ADD_MAC_ADDR DPNI_CMD(0x226)
-+#define DPNI_CMDID_REMOVE_MAC_ADDR DPNI_CMD(0x227)
-+#define DPNI_CMDID_CLR_MAC_FILTERS DPNI_CMD(0x228)
-+
-+#define DPNI_CMDID_SET_RX_TC_DIST DPNI_CMD(0x235)
-+
-+#define DPNI_CMDID_SET_QOS_TBL DPNI_CMD(0x240)
-+#define DPNI_CMDID_ADD_QOS_ENT DPNI_CMD(0x241)
-+#define DPNI_CMDID_REMOVE_QOS_ENT DPNI_CMD(0x242)
-+#define DPNI_CMDID_ADD_FS_ENT DPNI_CMD(0x244)
-+#define DPNI_CMDID_REMOVE_FS_ENT DPNI_CMD(0x245)
-+#define DPNI_CMDID_CLR_FS_ENT DPNI_CMD(0x246)
-+
-+#define DPNI_CMDID_SET_TX_PRIORITIES DPNI_CMD_V2(0x250)
-+#define DPNI_CMDID_GET_STATISTICS DPNI_CMD_V2(0x25D)
-+#define DPNI_CMDID_RESET_STATISTICS DPNI_CMD(0x25E)
-+#define DPNI_CMDID_GET_QUEUE DPNI_CMD(0x25F)
-+#define DPNI_CMDID_SET_QUEUE DPNI_CMD(0x260)
-+#define DPNI_CMDID_GET_TAILDROP DPNI_CMD(0x261)
-+#define DPNI_CMDID_SET_TAILDROP DPNI_CMD(0x262)
-+
-+#define DPNI_CMDID_GET_PORT_MAC_ADDR DPNI_CMD(0x263)
-+
-+#define DPNI_CMDID_GET_BUFFER_LAYOUT DPNI_CMD(0x264)
-+#define DPNI_CMDID_SET_BUFFER_LAYOUT DPNI_CMD(0x265)
-+
-+#define DPNI_CMDID_SET_TX_CONFIRMATION_MODE DPNI_CMD(0x266)
-+#define DPNI_CMDID_SET_CONGESTION_NOTIFICATION DPNI_CMD(0x267)
-+#define DPNI_CMDID_GET_CONGESTION_NOTIFICATION DPNI_CMD(0x268)
-+#define DPNI_CMDID_SET_EARLY_DROP DPNI_CMD(0x269)
-+#define DPNI_CMDID_GET_EARLY_DROP DPNI_CMD(0x26A)
-+#define DPNI_CMDID_GET_OFFLOAD DPNI_CMD(0x26B)
-+#define DPNI_CMDID_SET_OFFLOAD DPNI_CMD(0x26C)
-+
-+#define DPNI_CMDID_SET_RX_FS_DIST DPNI_CMD(0x273)
-+#define DPNI_CMDID_SET_RX_HASH_DIST DPNI_CMD(0x274)
-+
-+/* Macros for accessing command fields smaller than 1byte */
-+#define DPNI_MASK(field) \
-+ GENMASK(DPNI_##field##_SHIFT + DPNI_##field##_SIZE - 1, \
-+ DPNI_##field##_SHIFT)
-+
-+#define dpni_set_field(var, field, val) \
-+ ((var) |= (((val) << DPNI_##field##_SHIFT) & DPNI_MASK(field)))
-+#define dpni_get_field(var, field) \
-+ (((var) & DPNI_MASK(field)) >> DPNI_##field##_SHIFT)
-+
-+struct dpni_cmd_open {
-+ __le32 dpni_id;
-+};
-+
-+#define DPNI_BACKUP_POOL(val, order) (((val) & 0x1) << (order))
-+struct dpni_cmd_set_pools {
-+ u8 num_dpbp;
-+ u8 backup_pool_mask;
-+ __le16 pad;
-+ struct {
-+ __le16 dpbp_id;
-+ u8 priority_mask;
-+ u8 pad;
-+ } pool[DPNI_MAX_DPBP];
-+ __le16 buffer_size[DPNI_MAX_DPBP];
-+};
-+
-+/* The enable indication is always the least significant bit */
-+#define DPNI_ENABLE_SHIFT 0
-+#define DPNI_ENABLE_SIZE 1
-+
-+struct dpni_rsp_is_enabled {
-+ u8 enabled;
-+};
-+
-+struct dpni_rsp_get_irq {
-+ /* response word 0 */
-+ __le32 irq_val;
-+ __le32 pad;
-+ /* response word 1 */
-+ __le64 irq_addr;
-+ /* response word 2 */
-+ __le32 irq_num;
-+ __le32 type;
-+};
-+
-+struct dpni_cmd_set_irq_enable {
-+ u8 enable;
-+ u8 pad[3];
-+ u8 irq_index;
-+};
-+
-+struct dpni_cmd_get_irq_enable {
-+ __le32 pad;
-+ u8 irq_index;
-+};
-+
-+struct dpni_rsp_get_irq_enable {
-+ u8 enabled;
-+};
-+
-+struct dpni_cmd_set_irq_mask {
-+ __le32 mask;
-+ u8 irq_index;
-+};
-+
-+struct dpni_cmd_get_irq_mask {
-+ __le32 pad;
-+ u8 irq_index;
-+};
-+
-+struct dpni_rsp_get_irq_mask {
-+ __le32 mask;
-+};
-+
-+struct dpni_cmd_get_irq_status {
-+ __le32 status;
-+ u8 irq_index;
-+};
-+
-+struct dpni_rsp_get_irq_status {
-+ __le32 status;
-+};
-+
-+struct dpni_cmd_clear_irq_status {
-+ __le32 status;
-+ u8 irq_index;
-+};
-+
-+struct dpni_rsp_get_attr {
-+ /* response word 0 */
-+ __le32 options;
-+ u8 num_queues;
-+ u8 num_tcs;
-+ u8 mac_filter_entries;
-+ u8 pad0;
-+ /* response word 1 */
-+ u8 vlan_filter_entries;
-+ u8 pad1;
-+ u8 qos_entries;
-+ u8 pad2;
-+ __le16 fs_entries;
-+ __le16 pad3;
-+ /* response word 2 */
-+ u8 qos_key_size;
-+ u8 fs_key_size;
-+ __le16 wriop_version;
-+};
-+
-+#define DPNI_ERROR_ACTION_SHIFT 0
-+#define DPNI_ERROR_ACTION_SIZE 4
-+#define DPNI_FRAME_ANN_SHIFT 4
-+#define DPNI_FRAME_ANN_SIZE 1
-+
-+struct dpni_cmd_set_errors_behavior {
-+ __le32 errors;
-+ /* from least significant bit: error_action:4, set_frame_annotation:1 */
-+ u8 flags;
-+};
-+
-+/* There are 3 separate commands for configuring Rx, Tx and Tx confirmation
-+ * buffer layouts, but they all share the same parameters.
-+ * If one of the functions changes, below structure needs to be split.
-+ */
-+
-+#define DPNI_PASS_TS_SHIFT 0
-+#define DPNI_PASS_TS_SIZE 1
-+#define DPNI_PASS_PR_SHIFT 1
-+#define DPNI_PASS_PR_SIZE 1
-+#define DPNI_PASS_FS_SHIFT 2
-+#define DPNI_PASS_FS_SIZE 1
-+
-+struct dpni_cmd_get_buffer_layout {
-+ u8 qtype;
-+};
-+
-+struct dpni_rsp_get_buffer_layout {
-+ /* response word 0 */
-+ u8 pad0[6];
-+ /* from LSB: pass_timestamp:1, parser_result:1, frame_status:1 */
-+ u8 flags;
-+ u8 pad1;
-+ /* response word 1 */
-+ __le16 private_data_size;
-+ __le16 data_align;
-+ __le16 head_room;
-+ __le16 tail_room;
-+};
-+
-+struct dpni_cmd_set_buffer_layout {
-+ /* cmd word 0 */
-+ u8 qtype;
-+ u8 pad0[3];
-+ __le16 options;
-+ /* from LSB: pass_timestamp:1, parser_result:1, frame_status:1 */
-+ u8 flags;
-+ u8 pad1;
-+ /* cmd word 1 */
-+ __le16 private_data_size;
-+ __le16 data_align;
-+ __le16 head_room;
-+ __le16 tail_room;
-+};
-+
-+struct dpni_cmd_set_offload {
-+ u8 pad[3];
-+ u8 dpni_offload;
-+ __le32 config;
-+};
-+
-+struct dpni_cmd_get_offload {
-+ u8 pad[3];
-+ u8 dpni_offload;
-+};
-+
-+struct dpni_rsp_get_offload {
-+ __le32 pad;
-+ __le32 config;
-+};
-+
-+struct dpni_cmd_get_qdid {
-+ u8 qtype;
-+};
-+
-+struct dpni_rsp_get_qdid {
-+ __le16 qdid;
-+};
-+
-+struct dpni_rsp_get_tx_data_offset {
-+ __le16 data_offset;
-+};
-+
-+struct dpni_cmd_get_statistics {
-+ u8 page_number;
-+ u8 param;
-+};
-+
-+struct dpni_rsp_get_statistics {
-+ __le64 counter[DPNI_STATISTICS_CNT];
-+};
-+
-+struct dpni_cmd_set_link_cfg {
-+ /* cmd word 0 */
-+ __le64 pad0;
-+ /* cmd word 1 */
-+ __le32 rate;
-+ __le32 pad1;
-+ /* cmd word 2 */
-+ __le64 options;
-+};
-+
-+#define DPNI_LINK_STATE_SHIFT 0
-+#define DPNI_LINK_STATE_SIZE 1
-+
-+struct dpni_rsp_get_link_state {
-+ /* response word 0 */
-+ __le32 pad0;
-+ /* from LSB: up:1 */
-+ u8 flags;
-+ u8 pad1[3];
-+ /* response word 1 */
-+ __le32 rate;
-+ __le32 pad2;
-+ /* response word 2 */
-+ __le64 options;
-+};
-+
-+#define DPNI_COUPLED_SHIFT 0
-+#define DPNI_COUPLED_SIZE 1
-+
-+struct dpni_cmd_set_tx_shaping {
-+ /* cmd word 0 */
-+ __le16 tx_cr_max_burst_size;
-+ __le16 tx_er_max_burst_size;
-+ __le32 pad;
-+ /* cmd word 1 */
-+ __le32 tx_cr_rate_limit;
-+ __le32 tx_er_rate_limit;
-+ /* cmd word 2 */
-+ /* from LSB: coupled:1 */
-+ u8 coupled;
-+};
-+
-+struct dpni_cmd_set_max_frame_length {
-+ __le16 max_frame_length;
-+};
-+
-+struct dpni_rsp_get_max_frame_length {
-+ __le16 max_frame_length;
-+};
-+
-+struct dpni_cmd_set_multicast_promisc {
-+ u8 enable;
-+};
-+
-+struct dpni_rsp_get_multicast_promisc {
-+ u8 enabled;
-+};
-+
-+struct dpni_cmd_set_unicast_promisc {
-+ u8 enable;
-+};
-+
-+struct dpni_rsp_get_unicast_promisc {
-+ u8 enabled;
-+};
-+
-+struct dpni_cmd_set_primary_mac_addr {
-+ __le16 pad;
-+ u8 mac_addr[6];
-+};
-+
-+struct dpni_rsp_get_primary_mac_addr {
-+ __le16 pad;
-+ u8 mac_addr[6];
-+};
-+
-+struct dpni_rsp_get_port_mac_addr {
-+ __le16 pad;
-+ u8 mac_addr[6];
-+};
-+
-+struct dpni_cmd_add_mac_addr {
-+ __le16 pad;
-+ u8 mac_addr[6];
-+};
-+
-+struct dpni_cmd_remove_mac_addr {
-+ __le16 pad;
-+ u8 mac_addr[6];
-+};
-+
-+#define DPNI_UNICAST_FILTERS_SHIFT 0
-+#define DPNI_UNICAST_FILTERS_SIZE 1
-+#define DPNI_MULTICAST_FILTERS_SHIFT 1
-+#define DPNI_MULTICAST_FILTERS_SIZE 1
-+
-+struct dpni_cmd_clear_mac_filters {
-+ /* from LSB: unicast:1, multicast:1 */
-+ u8 flags;
-+};
-+
-+#define DPNI_SEPARATE_GRP_SHIFT 0
-+#define DPNI_SEPARATE_GRP_SIZE 1
-+#define DPNI_MODE_1_SHIFT 0
-+#define DPNI_MODE_1_SIZE 4
-+#define DPNI_MODE_2_SHIFT 4
-+#define DPNI_MODE_2_SIZE 4
-+
-+struct dpni_cmd_set_tx_priorities {
-+ __le16 flags;
-+ u8 prio_group_A;
-+ u8 prio_group_B;
-+ __le32 pad0;
-+ u8 modes[4];
-+ __le32 pad1;
-+ __le64 pad2;
-+ __le16 delta_bandwidth[8];
-+};
-+
-+#define DPNI_DIST_MODE_SHIFT 0
-+#define DPNI_DIST_MODE_SIZE 4
-+#define DPNI_MISS_ACTION_SHIFT 4
-+#define DPNI_MISS_ACTION_SIZE 4
-+
-+struct dpni_cmd_set_rx_tc_dist {
-+ /* cmd word 0 */
-+ __le16 dist_size;
-+ u8 tc_id;
-+ /* from LSB: dist_mode:4, miss_action:4 */
-+ u8 flags;
-+ __le16 pad0;
-+ __le16 default_flow_id;
-+ /* cmd word 1..5 */
-+ __le64 pad1[5];
-+ /* cmd word 6 */
-+ __le64 key_cfg_iova;
-+};
-+
-+/* dpni_set_rx_tc_dist extension (structure of the DMA-able memory at
-+ * key_cfg_iova)
-+ */
-+struct dpni_mask_cfg {
-+ u8 mask;
-+ u8 offset;
-+};
-+
-+#define DPNI_EFH_TYPE_SHIFT 0
-+#define DPNI_EFH_TYPE_SIZE 4
-+#define DPNI_EXTRACT_TYPE_SHIFT 0
-+#define DPNI_EXTRACT_TYPE_SIZE 4
-+
-+struct dpni_dist_extract {
-+ /* word 0 */
-+ u8 prot;
-+ /* EFH type stored in the 4 least significant bits */
-+ u8 efh_type;
-+ u8 size;
-+ u8 offset;
-+ __le32 field;
-+ /* word 1 */
-+ u8 hdr_index;
-+ u8 constant;
-+ u8 num_of_repeats;
-+ u8 num_of_byte_masks;
-+ /* Extraction type is stored in the 4 LSBs */
-+ u8 extract_type;
-+ u8 pad[3];
-+ /* word 2 */
-+ struct dpni_mask_cfg masks[4];
-+};
-+
-+struct dpni_ext_set_rx_tc_dist {
-+ /* extension word 0 */
-+ u8 num_extracts;
-+ u8 pad[7];
-+ /* words 1..25 */
-+ struct dpni_dist_extract extracts[DPKG_MAX_NUM_OF_EXTRACTS];
-+};
-+
-+struct dpni_cmd_get_queue {
-+ u8 qtype;
-+ u8 tc;
-+ u8 index;
-+};
-+
-+#define DPNI_DEST_TYPE_SHIFT 0
-+#define DPNI_DEST_TYPE_SIZE 4
-+#define DPNI_STASH_CTRL_SHIFT 6
-+#define DPNI_STASH_CTRL_SIZE 1
-+#define DPNI_HOLD_ACTIVE_SHIFT 7
-+#define DPNI_HOLD_ACTIVE_SIZE 1
-+
-+struct dpni_rsp_get_queue {
-+ /* response word 0 */
-+ __le64 pad0;
-+ /* response word 1 */
-+ __le32 dest_id;
-+ __le16 pad1;
-+ u8 dest_prio;
-+ /* From LSB: dest_type:4, pad:2, flc_stash_ctrl:1, hold_active:1 */
-+ u8 flags;
-+ /* response word 2 */
-+ __le64 flc;
-+ /* response word 3 */
-+ __le64 user_context;
-+ /* response word 4 */
-+ __le32 fqid;
-+ __le16 qdbin;
-+};
-+
-+struct dpni_cmd_set_queue {
-+ /* cmd word 0 */
-+ u8 qtype;
-+ u8 tc;
-+ u8 index;
-+ u8 options;
-+ __le32 pad0;
-+ /* cmd word 1 */
-+ __le32 dest_id;
-+ __le16 pad1;
-+ u8 dest_prio;
-+ u8 flags;
-+ /* cmd word 2 */
-+ __le64 flc;
-+ /* cmd word 3 */
-+ __le64 user_context;
-+};
-+
-+#define DPNI_DISCARD_ON_MISS_SHIFT 0
-+#define DPNI_DISCARD_ON_MISS_SIZE 1
-+
-+struct dpni_cmd_set_qos_table {
-+ __le32 pad;
-+ u8 default_tc;
-+ /* only the LSB */
-+ u8 discard_on_miss;
-+ __le16 pad1[21];
-+ __le64 key_cfg_iova;
-+};
-+
-+struct dpni_cmd_add_qos_entry {
-+ __le16 pad;
-+ u8 tc_id;
-+ u8 key_size;
-+ __le16 index;
-+ __le16 pad2;
-+ __le64 key_iova;
-+ __le64 mask_iova;
-+};
-+
-+struct dpni_cmd_remove_qos_entry {
-+ u8 pad1[3];
-+ u8 key_size;
-+ __le32 pad2;
-+ __le64 key_iova;
-+ __le64 mask_iova;
-+};
-+
-+struct dpni_cmd_add_fs_entry {
-+ /* cmd word 0 */
-+ __le16 options;
-+ u8 tc_id;
-+ u8 key_size;
-+ __le16 index;
-+ __le16 flow_id;
-+ /* cmd word 1 */
-+ __le64 key_iova;
-+ /* cmd word 2 */
-+ __le64 mask_iova;
-+ /* cmd word 3 */
-+ __le64 flc;
-+};
-+
-+struct dpni_cmd_remove_fs_entry {
-+ /* cmd word 0 */
-+ __le16 pad0;
-+ u8 tc_id;
-+ u8 key_size;
-+ __le32 pad1;
-+ /* cmd word 1 */
-+ __le64 key_iova;
-+ /* cmd word 2 */
-+ __le64 mask_iova;
-+};
-+
-+struct dpni_cmd_set_taildrop {
-+ /* cmd word 0 */
-+ u8 congestion_point;
-+ u8 qtype;
-+ u8 tc;
-+ u8 index;
-+ __le32 pad0;
-+ /* cmd word 1 */
-+ /* Only least significant bit is relevant */
-+ u8 enable;
-+ u8 pad1;
-+ u8 units;
-+ u8 pad2;
-+ __le32 threshold;
-+};
-+
-+struct dpni_cmd_get_taildrop {
-+ u8 congestion_point;
-+ u8 qtype;
-+ u8 tc;
-+ u8 index;
-+};
-+
-+struct dpni_rsp_get_taildrop {
-+ /* cmd word 0 */
-+ __le64 pad0;
-+ /* cmd word 1 */
-+ /* only least significant bit is relevant */
-+ u8 enable;
-+ u8 pad1;
-+ u8 units;
-+ u8 pad2;
-+ __le32 threshold;
-+};
-+
-+struct dpni_rsp_get_api_version {
-+ u16 major;
-+ u16 minor;
-+};
-+
-+#define DPNI_DEST_TYPE_SHIFT 0
-+#define DPNI_DEST_TYPE_SIZE 4
-+#define DPNI_CONG_UNITS_SHIFT 4
-+#define DPNI_CONG_UNITS_SIZE 2
-+
-+struct dpni_cmd_set_congestion_notification {
-+ /* cmd word 0 */
-+ u8 qtype;
-+ u8 tc;
-+ u8 pad[6];
-+ /* cmd word 1 */
-+ __le32 dest_id;
-+ __le16 notification_mode;
-+ u8 dest_priority;
-+ /* from LSB: dest_type: 4 units:2 */
-+ u8 type_units;
-+ /* cmd word 2 */
-+ __le64 message_iova;
-+ /* cmd word 3 */
-+ __le64 message_ctx;
-+ /* cmd word 4 */
-+ __le32 threshold_entry;
-+ __le32 threshold_exit;
-+};
-+
-+struct dpni_cmd_get_congestion_notification {
-+ /* cmd word 0 */
-+ u8 qtype;
-+ u8 tc;
-+};
-+
-+struct dpni_rsp_get_congestion_notification {
-+ /* cmd word 0 */
-+ __le64 pad;
-+ /* cmd word 1 */
-+ __le32 dest_id;
-+ __le16 notification_mode;
-+ u8 dest_priority;
-+ /* from LSB: dest_type: 4 units:2 */
-+ u8 type_units;
-+ /* cmd word 2 */
-+ __le64 message_iova;
-+ /* cmd word 3 */
-+ __le64 message_ctx;
-+ /* cmd word 4 */
-+ __le32 threshold_entry;
-+ __le32 threshold_exit;
-+};
-+
-+#define DPNI_RX_FS_DIST_ENABLE_SHIFT 0
-+#define DPNI_RX_FS_DIST_ENABLE_SIZE 1
-+struct dpni_cmd_set_rx_fs_dist {
-+ __le16 dist_size;
-+ u8 enable;
-+ u8 tc;
-+ __le16 miss_flow_id;
-+ __le16 pad;
-+ __le64 key_cfg_iova;
-+};
-+
-+#define DPNI_RX_HASH_DIST_ENABLE_SHIFT 0
-+#define DPNI_RX_HASH_DIST_ENABLE_SIZE 1
-+struct dpni_cmd_set_rx_hash_dist {
-+ __le16 dist_size;
-+ u8 enable;
-+ u8 tc;
-+ __le32 pad;
-+ __le64 key_cfg_iova;
-+};
-+
-+#endif /* _FSL_DPNI_CMD_H */
---- /dev/null
-+++ b/drivers/staging/fsl-dpaa2/ethernet/dpni.c
-@@ -0,0 +1,2112 @@
-+/* Copyright 2013-2016 Freescale Semiconductor Inc.
-+ * Copyright 2016 NXP
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of the above-listed copyright holders nor the
-+ * names of any contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
-+ * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
-+ * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
-+ * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE
-+ * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
-+ * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
-+ * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
-+ * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
-+ * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
-+ * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
-+ * POSSIBILITY OF SUCH DAMAGE.
-+ */
-+#include <linux/kernel.h>
-+#include <linux/errno.h>
-+#include <linux/fsl/mc.h>
-+#include "dpni.h"
-+#include "dpni-cmd.h"
-+
-+/**
-+ * dpni_prepare_key_cfg() - function prepare extract parameters
-+ * @cfg: defining a full Key Generation profile (rule)
-+ * @key_cfg_buf: Zeroed 256 bytes of memory before mapping it to DMA
-+ *
-+ * This function has to be called before the following functions:
-+ * - dpni_set_rx_tc_dist()
-+ * - dpni_set_qos_table()
-+ */
-+int dpni_prepare_key_cfg(const struct dpkg_profile_cfg *cfg, u8 *key_cfg_buf)
-+{
-+ int i, j;
-+ struct dpni_ext_set_rx_tc_dist *dpni_ext;
-+ struct dpni_dist_extract *extr;
-+
-+ if (cfg->num_extracts > DPKG_MAX_NUM_OF_EXTRACTS)
-+ return -EINVAL;
-+
-+ dpni_ext = (struct dpni_ext_set_rx_tc_dist *)key_cfg_buf;
-+ dpni_ext->num_extracts = cfg->num_extracts;
-+
-+ for (i = 0; i < cfg->num_extracts; i++) {
-+ extr = &dpni_ext->extracts[i];
-+
-+ switch (cfg->extracts[i].type) {
-+ case DPKG_EXTRACT_FROM_HDR:
-+ extr->prot = cfg->extracts[i].extract.from_hdr.prot;
-+ dpni_set_field(extr->efh_type, EFH_TYPE,
-+ cfg->extracts[i].extract.from_hdr.type);
-+ extr->size = cfg->extracts[i].extract.from_hdr.size;
-+ extr->offset = cfg->extracts[i].extract.from_hdr.offset;
-+ extr->field = cpu_to_le32(
-+ cfg->extracts[i].extract.from_hdr.field);
-+ extr->hdr_index =
-+ cfg->extracts[i].extract.from_hdr.hdr_index;
-+ break;
-+ case DPKG_EXTRACT_FROM_DATA:
-+ extr->size = cfg->extracts[i].extract.from_data.size;
-+ extr->offset =
-+ cfg->extracts[i].extract.from_data.offset;
-+ break;
-+ case DPKG_EXTRACT_FROM_PARSE:
-+ extr->size = cfg->extracts[i].extract.from_parse.size;
-+ extr->offset =
-+ cfg->extracts[i].extract.from_parse.offset;
-+ break;
-+ default:
-+ return -EINVAL;
-+ }
-+
-+ extr->num_of_byte_masks = cfg->extracts[i].num_of_byte_masks;
-+ dpni_set_field(extr->extract_type, EXTRACT_TYPE,
-+ cfg->extracts[i].type);
-+
-+ for (j = 0; j < DPKG_NUM_OF_MASKS; j++) {
-+ extr->masks[j].mask = cfg->extracts[i].masks[j].mask;
-+ extr->masks[j].offset =
-+ cfg->extracts[i].masks[j].offset;
-+ }
-+ }
-+
-+ return 0;
-+}
-+
-+/**
-+ * dpni_open() - Open a control session for the specified object
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @dpni_id: DPNI unique ID
-+ * @token: Returned token; use in subsequent API calls
-+ *
-+ * This function can be used to open a control session for an
-+ * already created object; an object may have been declared in
-+ * the DPL or by calling the dpni_create() function.
-+ * This function returns a unique authentication token,
-+ * associated with the specific object ID and the specific MC
-+ * portal; this token must be used in all subsequent commands for
-+ * this specific object.
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpni_open(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ int dpni_id,
-+ u16 *token)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpni_cmd_open *cmd_params;
-+
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPNI_CMDID_OPEN,
-+ cmd_flags,
-+ 0);
-+ cmd_params = (struct dpni_cmd_open *)cmd.params;
-+ cmd_params->dpni_id = cpu_to_le32(dpni_id);
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ *token = mc_cmd_hdr_read_token(&cmd);
-+
-+ return 0;
-+}
-+
-+/**
-+ * dpni_close() - Close the control session of the object
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPNI object
-+ *
-+ * After this function is called, no further operations are
-+ * allowed on the object without opening a new control session.
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpni_close(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPNI_CMDID_CLOSE,
-+ cmd_flags,
-+ token);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpni_set_pools() - Set buffer pools configuration
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPNI object
-+ * @cfg: Buffer pools configuration
-+ *
-+ * mandatory for DPNI operation
-+ * warning:Allowed only when DPNI is disabled
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpni_set_pools(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ const struct dpni_pools_cfg *cfg)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpni_cmd_set_pools *cmd_params;
-+ int i;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPNI_CMDID_SET_POOLS,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpni_cmd_set_pools *)cmd.params;
-+ cmd_params->num_dpbp = cfg->num_dpbp;
-+ for (i = 0; i < DPNI_MAX_DPBP; i++) {
-+ cmd_params->pool[i].dpbp_id =
-+ cpu_to_le16(cfg->pools[i].dpbp_id);
-+ cmd_params->pool[i].priority_mask =
-+ cfg->pools[i].priority_mask;
-+ cmd_params->buffer_size[i] =
-+ cpu_to_le16(cfg->pools[i].buffer_size);
-+ cmd_params->backup_pool_mask |=
-+ DPNI_BACKUP_POOL(cfg->pools[i].backup_pool, i);
-+ }
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpni_enable() - Enable the DPNI, allow sending and receiving frames.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPNI object
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpni_enable(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPNI_CMDID_ENABLE,
-+ cmd_flags,
-+ token);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpni_disable() - Disable the DPNI, stop sending and receiving frames.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPNI object
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpni_disable(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPNI_CMDID_DISABLE,
-+ cmd_flags,
-+ token);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpni_is_enabled() - Check if the DPNI is enabled.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPNI object
-+ * @en: Returns '1' if object is enabled; '0' otherwise
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpni_is_enabled(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ int *en)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpni_rsp_is_enabled *rsp_params;
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPNI_CMDID_IS_ENABLED,
-+ cmd_flags,
-+ token);
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ rsp_params = (struct dpni_rsp_is_enabled *)cmd.params;
-+ *en = dpni_get_field(rsp_params->enabled, ENABLE);
-+
-+ return 0;
-+}
-+
-+/**
-+ * dpni_reset() - Reset the DPNI, returns the object to initial state.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPNI object
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpni_reset(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPNI_CMDID_RESET,
-+ cmd_flags,
-+ token);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpni_set_irq_enable() - Set overall interrupt state.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPNI object
-+ * @irq_index: The interrupt index to configure
-+ * @en: Interrupt state: - enable = 1, disable = 0
-+ *
-+ * Allows GPP software to control when interrupts are generated.
-+ * Each interrupt can have up to 32 causes. The enable/disable control's the
-+ * overall interrupt state. if the interrupt is disabled no causes will cause
-+ * an interrupt.
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpni_set_irq_enable(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 irq_index,
-+ u8 en)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpni_cmd_set_irq_enable *cmd_params;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPNI_CMDID_SET_IRQ_ENABLE,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpni_cmd_set_irq_enable *)cmd.params;
-+ dpni_set_field(cmd_params->enable, ENABLE, en);
-+ cmd_params->irq_index = irq_index;
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpni_get_irq_enable() - Get overall interrupt state
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPNI object
-+ * @irq_index: The interrupt index to configure
-+ * @en: Returned interrupt state - enable = 1, disable = 0
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpni_get_irq_enable(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 irq_index,
-+ u8 *en)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpni_cmd_get_irq_enable *cmd_params;
-+ struct dpni_rsp_get_irq_enable *rsp_params;
-+
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPNI_CMDID_GET_IRQ_ENABLE,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpni_cmd_get_irq_enable *)cmd.params;
-+ cmd_params->irq_index = irq_index;
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ rsp_params = (struct dpni_rsp_get_irq_enable *)cmd.params;
-+ *en = dpni_get_field(rsp_params->enabled, ENABLE);
-+
-+ return 0;
-+}
-+
-+/**
-+ * dpni_set_irq_mask() - Set interrupt mask.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPNI object
-+ * @irq_index: The interrupt index to configure
-+ * @mask: event mask to trigger interrupt;
-+ * each bit:
-+ * 0 = ignore event
-+ * 1 = consider event for asserting IRQ
-+ *
-+ * Every interrupt can have up to 32 causes and the interrupt model supports
-+ * masking/unmasking each cause independently
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpni_set_irq_mask(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 irq_index,
-+ u32 mask)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpni_cmd_set_irq_mask *cmd_params;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPNI_CMDID_SET_IRQ_MASK,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpni_cmd_set_irq_mask *)cmd.params;
-+ cmd_params->mask = cpu_to_le32(mask);
-+ cmd_params->irq_index = irq_index;
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpni_get_irq_mask() - Get interrupt mask.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPNI object
-+ * @irq_index: The interrupt index to configure
-+ * @mask: Returned event mask to trigger interrupt
-+ *
-+ * Every interrupt can have up to 32 causes and the interrupt model supports
-+ * masking/unmasking each cause independently
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpni_get_irq_mask(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 irq_index,
-+ u32 *mask)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpni_cmd_get_irq_mask *cmd_params;
-+ struct dpni_rsp_get_irq_mask *rsp_params;
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPNI_CMDID_GET_IRQ_MASK,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpni_cmd_get_irq_mask *)cmd.params;
-+ cmd_params->irq_index = irq_index;
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ rsp_params = (struct dpni_rsp_get_irq_mask *)cmd.params;
-+ *mask = le32_to_cpu(rsp_params->mask);
-+
-+ return 0;
-+}
-+
-+/**
-+ * dpni_get_irq_status() - Get the current status of any pending interrupts.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPNI object
-+ * @irq_index: The interrupt index to configure
-+ * @status: Returned interrupts status - one bit per cause:
-+ * 0 = no interrupt pending
-+ * 1 = interrupt pending
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpni_get_irq_status(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 irq_index,
-+ u32 *status)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpni_cmd_get_irq_status *cmd_params;
-+ struct dpni_rsp_get_irq_status *rsp_params;
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPNI_CMDID_GET_IRQ_STATUS,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpni_cmd_get_irq_status *)cmd.params;
-+ cmd_params->status = cpu_to_le32(*status);
-+ cmd_params->irq_index = irq_index;
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ rsp_params = (struct dpni_rsp_get_irq_status *)cmd.params;
-+ *status = le32_to_cpu(rsp_params->status);
-+
-+ return 0;
-+}
-+
-+/**
-+ * dpni_clear_irq_status() - Clear a pending interrupt's status
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPNI object
-+ * @irq_index: The interrupt index to configure
-+ * @status: bits to clear (W1C) - one bit per cause:
-+ * 0 = don't change
-+ * 1 = clear status bit
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpni_clear_irq_status(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 irq_index,
-+ u32 status)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpni_cmd_clear_irq_status *cmd_params;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPNI_CMDID_CLEAR_IRQ_STATUS,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpni_cmd_clear_irq_status *)cmd.params;
-+ cmd_params->irq_index = irq_index;
-+ cmd_params->status = cpu_to_le32(status);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpni_get_attributes() - Retrieve DPNI attributes.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPNI object
-+ * @attr: Object's attributes
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpni_get_attributes(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ struct dpni_attr *attr)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpni_rsp_get_attr *rsp_params;
-+
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPNI_CMDID_GET_ATTR,
-+ cmd_flags,
-+ token);
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ rsp_params = (struct dpni_rsp_get_attr *)cmd.params;
-+ attr->options = le32_to_cpu(rsp_params->options);
-+ attr->num_queues = rsp_params->num_queues;
-+ attr->num_tcs = rsp_params->num_tcs;
-+ attr->mac_filter_entries = rsp_params->mac_filter_entries;
-+ attr->vlan_filter_entries = rsp_params->vlan_filter_entries;
-+ attr->qos_entries = rsp_params->qos_entries;
-+ attr->fs_entries = le16_to_cpu(rsp_params->fs_entries);
-+ attr->qos_key_size = rsp_params->qos_key_size;
-+ attr->fs_key_size = rsp_params->fs_key_size;
-+ attr->wriop_version = le16_to_cpu(rsp_params->wriop_version);
-+
-+ return 0;
-+}
-+
-+/**
-+ * dpni_set_errors_behavior() - Set errors behavior
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPNI object
-+ * @cfg: Errors configuration
-+ *
-+ * this function may be called numerous times with different
-+ * error masks
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpni_set_errors_behavior(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ struct dpni_error_cfg *cfg)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpni_cmd_set_errors_behavior *cmd_params;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPNI_CMDID_SET_ERRORS_BEHAVIOR,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpni_cmd_set_errors_behavior *)cmd.params;
-+ cmd_params->errors = cpu_to_le32(cfg->errors);
-+ dpni_set_field(cmd_params->flags, ERROR_ACTION, cfg->error_action);
-+ dpni_set_field(cmd_params->flags, FRAME_ANN, cfg->set_frame_annotation);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpni_get_buffer_layout() - Retrieve buffer layout attributes.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPNI object
-+ * @qtype: Type of queue to retrieve configuration for
-+ * @layout: Returns buffer layout attributes
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpni_get_buffer_layout(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ enum dpni_queue_type qtype,
-+ struct dpni_buffer_layout *layout)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpni_cmd_get_buffer_layout *cmd_params;
-+ struct dpni_rsp_get_buffer_layout *rsp_params;
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPNI_CMDID_GET_BUFFER_LAYOUT,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpni_cmd_get_buffer_layout *)cmd.params;
-+ cmd_params->qtype = qtype;
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ rsp_params = (struct dpni_rsp_get_buffer_layout *)cmd.params;
-+ layout->pass_timestamp = dpni_get_field(rsp_params->flags, PASS_TS);
-+ layout->pass_parser_result = dpni_get_field(rsp_params->flags, PASS_PR);
-+ layout->pass_frame_status = dpni_get_field(rsp_params->flags, PASS_FS);
-+ layout->private_data_size = le16_to_cpu(rsp_params->private_data_size);
-+ layout->data_align = le16_to_cpu(rsp_params->data_align);
-+ layout->data_head_room = le16_to_cpu(rsp_params->head_room);
-+ layout->data_tail_room = le16_to_cpu(rsp_params->tail_room);
-+
-+ return 0;
-+}
-+
-+/**
-+ * dpni_set_buffer_layout() - Set buffer layout configuration.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPNI object
-+ * @qtype: Type of queue this configuration applies to
-+ * @layout: Buffer layout configuration
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ *
-+ * @warning Allowed only when DPNI is disabled
-+ */
-+int dpni_set_buffer_layout(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ enum dpni_queue_type qtype,
-+ const struct dpni_buffer_layout *layout)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpni_cmd_set_buffer_layout *cmd_params;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPNI_CMDID_SET_BUFFER_LAYOUT,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpni_cmd_set_buffer_layout *)cmd.params;
-+ cmd_params->qtype = qtype;
-+ cmd_params->options = cpu_to_le16(layout->options);
-+ dpni_set_field(cmd_params->flags, PASS_TS, layout->pass_timestamp);
-+ dpni_set_field(cmd_params->flags, PASS_PR, layout->pass_parser_result);
-+ dpni_set_field(cmd_params->flags, PASS_FS, layout->pass_frame_status);
-+ cmd_params->private_data_size = cpu_to_le16(layout->private_data_size);
-+ cmd_params->data_align = cpu_to_le16(layout->data_align);
-+ cmd_params->head_room = cpu_to_le16(layout->data_head_room);
-+ cmd_params->tail_room = cpu_to_le16(layout->data_tail_room);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpni_set_offload() - Set DPNI offload configuration.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPNI object
-+ * @type: Type of DPNI offload
-+ * @config: Offload configuration.
-+ * For checksum offloads, non-zero value enables the offload
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ *
-+ * @warning Allowed only when DPNI is disabled
-+ */
-+
-+int dpni_set_offload(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ enum dpni_offload type,
-+ u32 config)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpni_cmd_set_offload *cmd_params;
-+
-+ cmd.header = mc_encode_cmd_header(DPNI_CMDID_SET_OFFLOAD,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpni_cmd_set_offload *)cmd.params;
-+ cmd_params->dpni_offload = type;
-+ cmd_params->config = cpu_to_le32(config);
-+
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+int dpni_get_offload(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ enum dpni_offload type,
-+ u32 *config)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpni_cmd_get_offload *cmd_params;
-+ struct dpni_rsp_get_offload *rsp_params;
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPNI_CMDID_GET_OFFLOAD,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpni_cmd_get_offload *)cmd.params;
-+ cmd_params->dpni_offload = type;
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ rsp_params = (struct dpni_rsp_get_offload *)cmd.params;
-+ *config = le32_to_cpu(rsp_params->config);
-+
-+ return 0;
-+}
-+
-+/**
-+ * dpni_get_qdid() - Get the Queuing Destination ID (QDID) that should be used
-+ * for enqueue operations
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPNI object
-+ * @qtype: Type of queue to receive QDID for
-+ * @qdid: Returned virtual QDID value that should be used as an argument
-+ * in all enqueue operations
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpni_get_qdid(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ enum dpni_queue_type qtype,
-+ u16 *qdid)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpni_cmd_get_qdid *cmd_params;
-+ struct dpni_rsp_get_qdid *rsp_params;
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPNI_CMDID_GET_QDID,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpni_cmd_get_qdid *)cmd.params;
-+ cmd_params->qtype = qtype;
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ rsp_params = (struct dpni_rsp_get_qdid *)cmd.params;
-+ *qdid = le16_to_cpu(rsp_params->qdid);
-+
-+ return 0;
-+}
-+
-+/**
-+ * dpni_get_tx_data_offset() - Get the Tx data offset (from start of buffer)
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPNI object
-+ * @data_offset: Tx data offset (from start of buffer)
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpni_get_tx_data_offset(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 *data_offset)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpni_rsp_get_tx_data_offset *rsp_params;
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPNI_CMDID_GET_TX_DATA_OFFSET,
-+ cmd_flags,
-+ token);
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ rsp_params = (struct dpni_rsp_get_tx_data_offset *)cmd.params;
-+ *data_offset = le16_to_cpu(rsp_params->data_offset);
-+
-+ return 0;
-+}
-+
-+/**
-+ * dpni_set_link_cfg() - set the link configuration.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPNI object
-+ * @cfg: Link configuration
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpni_set_link_cfg(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ const struct dpni_link_cfg *cfg)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpni_cmd_set_link_cfg *cmd_params;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPNI_CMDID_SET_LINK_CFG,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpni_cmd_set_link_cfg *)cmd.params;
-+ cmd_params->rate = cpu_to_le32(cfg->rate);
-+ cmd_params->options = cpu_to_le64(cfg->options);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpni_get_link_state() - Return the link state (either up or down)
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPNI object
-+ * @state: Returned link state;
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpni_get_link_state(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ struct dpni_link_state *state)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpni_rsp_get_link_state *rsp_params;
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPNI_CMDID_GET_LINK_STATE,
-+ cmd_flags,
-+ token);
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ rsp_params = (struct dpni_rsp_get_link_state *)cmd.params;
-+ state->up = dpni_get_field(rsp_params->flags, LINK_STATE);
-+ state->rate = le32_to_cpu(rsp_params->rate);
-+ state->options = le64_to_cpu(rsp_params->options);
-+
-+ return 0;
-+}
-+
-+/**
-+ * dpni_set_tx_shaping() - Set the transmit shaping
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPNI object
-+ * @tx_cr_shaper: TX committed rate shaping configuration
-+ * @tx_er_shaper: TX excess rate shaping configuration
-+ * @coupled: Committed and excess rate shapers are coupled
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpni_set_tx_shaping(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ const struct dpni_tx_shaping_cfg *tx_cr_shaper,
-+ const struct dpni_tx_shaping_cfg *tx_er_shaper,
-+ int coupled)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpni_cmd_set_tx_shaping *cmd_params;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPNI_CMDID_SET_TX_SHAPING,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpni_cmd_set_tx_shaping *)cmd.params;
-+ cmd_params->tx_cr_max_burst_size =
-+ cpu_to_le16(tx_cr_shaper->max_burst_size);
-+ cmd_params->tx_er_max_burst_size =
-+ cpu_to_le16(tx_er_shaper->max_burst_size);
-+ cmd_params->tx_cr_rate_limit = cpu_to_le32(tx_cr_shaper->rate_limit);
-+ cmd_params->tx_er_rate_limit = cpu_to_le32(tx_er_shaper->rate_limit);
-+ dpni_set_field(cmd_params->coupled, COUPLED, coupled);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpni_set_max_frame_length() - Set the maximum received frame length.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPNI object
-+ * @max_frame_length: Maximum received frame length (in
-+ * bytes); frame is discarded if its
-+ * length exceeds this value
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpni_set_max_frame_length(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 max_frame_length)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpni_cmd_set_max_frame_length *cmd_params;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPNI_CMDID_SET_MAX_FRAME_LENGTH,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpni_cmd_set_max_frame_length *)cmd.params;
-+ cmd_params->max_frame_length = cpu_to_le16(max_frame_length);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpni_get_max_frame_length() - Get the maximum received frame length.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPNI object
-+ * @max_frame_length: Maximum received frame length (in
-+ * bytes); frame is discarded if its
-+ * length exceeds this value
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpni_get_max_frame_length(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 *max_frame_length)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpni_rsp_get_max_frame_length *rsp_params;
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPNI_CMDID_GET_MAX_FRAME_LENGTH,
-+ cmd_flags,
-+ token);
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ rsp_params = (struct dpni_rsp_get_max_frame_length *)cmd.params;
-+ *max_frame_length = le16_to_cpu(rsp_params->max_frame_length);
-+
-+ return 0;
-+}
-+
-+/**
-+ * dpni_set_multicast_promisc() - Enable/disable multicast promiscuous mode
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPNI object
-+ * @en: Set to '1' to enable; '0' to disable
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpni_set_multicast_promisc(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ int en)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpni_cmd_set_multicast_promisc *cmd_params;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPNI_CMDID_SET_MCAST_PROMISC,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpni_cmd_set_multicast_promisc *)cmd.params;
-+ dpni_set_field(cmd_params->enable, ENABLE, en);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpni_get_multicast_promisc() - Get multicast promiscuous mode
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPNI object
-+ * @en: Returns '1' if enabled; '0' otherwise
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpni_get_multicast_promisc(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ int *en)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpni_rsp_get_multicast_promisc *rsp_params;
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPNI_CMDID_GET_MCAST_PROMISC,
-+ cmd_flags,
-+ token);
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ rsp_params = (struct dpni_rsp_get_multicast_promisc *)cmd.params;
-+ *en = dpni_get_field(rsp_params->enabled, ENABLE);
-+
-+ return 0;
-+}
-+
-+/**
-+ * dpni_set_unicast_promisc() - Enable/disable unicast promiscuous mode
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPNI object
-+ * @en: Set to '1' to enable; '0' to disable
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpni_set_unicast_promisc(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ int en)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpni_cmd_set_unicast_promisc *cmd_params;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPNI_CMDID_SET_UNICAST_PROMISC,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpni_cmd_set_unicast_promisc *)cmd.params;
-+ dpni_set_field(cmd_params->enable, ENABLE, en);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpni_get_unicast_promisc() - Get unicast promiscuous mode
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPNI object
-+ * @en: Returns '1' if enabled; '0' otherwise
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpni_get_unicast_promisc(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ int *en)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpni_rsp_get_unicast_promisc *rsp_params;
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPNI_CMDID_GET_UNICAST_PROMISC,
-+ cmd_flags,
-+ token);
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ rsp_params = (struct dpni_rsp_get_unicast_promisc *)cmd.params;
-+ *en = dpni_get_field(rsp_params->enabled, ENABLE);
-+
-+ return 0;
-+}
-+
-+/**
-+ * dpni_set_primary_mac_addr() - Set the primary MAC address
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPNI object
-+ * @mac_addr: MAC address to set as primary address
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpni_set_primary_mac_addr(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ const u8 mac_addr[6])
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpni_cmd_set_primary_mac_addr *cmd_params;
-+ int i;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPNI_CMDID_SET_PRIM_MAC,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpni_cmd_set_primary_mac_addr *)cmd.params;
-+ for (i = 0; i < 6; i++)
-+ cmd_params->mac_addr[i] = mac_addr[5 - i];
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpni_get_primary_mac_addr() - Get the primary MAC address
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPNI object
-+ * @mac_addr: Returned MAC address
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpni_get_primary_mac_addr(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 mac_addr[6])
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpni_rsp_get_primary_mac_addr *rsp_params;
-+ int i, err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPNI_CMDID_GET_PRIM_MAC,
-+ cmd_flags,
-+ token);
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ rsp_params = (struct dpni_rsp_get_primary_mac_addr *)cmd.params;
-+ for (i = 0; i < 6; i++)
-+ mac_addr[5 - i] = rsp_params->mac_addr[i];
-+
-+ return 0;
-+}
-+
-+/**
-+ * dpni_get_port_mac_addr() - Retrieve MAC address associated to the physical
-+ * port the DPNI is attached to
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPNI object
-+ * @mac_addr: MAC address of the physical port, if any, otherwise 0
-+ *
-+ * The primary MAC address is not cleared by this operation.
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpni_get_port_mac_addr(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 mac_addr[6])
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpni_rsp_get_port_mac_addr *rsp_params;
-+ int i, err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPNI_CMDID_GET_PORT_MAC_ADDR,
-+ cmd_flags,
-+ token);
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ rsp_params = (struct dpni_rsp_get_port_mac_addr *)cmd.params;
-+ for (i = 0; i < 6; i++)
-+ mac_addr[5 - i] = rsp_params->mac_addr[i];
-+
-+ return 0;
-+}
-+
-+/**
-+ * dpni_add_mac_addr() - Add MAC address filter
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPNI object
-+ * @mac_addr: MAC address to add
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpni_add_mac_addr(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ const u8 mac_addr[6])
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpni_cmd_add_mac_addr *cmd_params;
-+ int i;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPNI_CMDID_ADD_MAC_ADDR,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpni_cmd_add_mac_addr *)cmd.params;
-+ for (i = 0; i < 6; i++)
-+ cmd_params->mac_addr[i] = mac_addr[5 - i];
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpni_remove_mac_addr() - Remove MAC address filter
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPNI object
-+ * @mac_addr: MAC address to remove
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpni_remove_mac_addr(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ const u8 mac_addr[6])
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpni_cmd_remove_mac_addr *cmd_params;
-+ int i;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPNI_CMDID_REMOVE_MAC_ADDR,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpni_cmd_remove_mac_addr *)cmd.params;
-+ for (i = 0; i < 6; i++)
-+ cmd_params->mac_addr[i] = mac_addr[5 - i];
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpni_clear_mac_filters() - Clear all unicast and/or multicast MAC filters
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPNI object
-+ * @unicast: Set to '1' to clear unicast addresses
-+ * @multicast: Set to '1' to clear multicast addresses
-+ *
-+ * The primary MAC address is not cleared by this operation.
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpni_clear_mac_filters(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ int unicast,
-+ int multicast)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpni_cmd_clear_mac_filters *cmd_params;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPNI_CMDID_CLR_MAC_FILTERS,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpni_cmd_clear_mac_filters *)cmd.params;
-+ dpni_set_field(cmd_params->flags, UNICAST_FILTERS, unicast);
-+ dpni_set_field(cmd_params->flags, MULTICAST_FILTERS, multicast);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpni_set_tx_priorities() - Set transmission TC priority configuration
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPNI object
-+ * @cfg: Transmission selection configuration
-+ *
-+ * warning: Allowed only when DPNI is disabled
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpni_set_tx_priorities(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ const struct dpni_tx_priorities_cfg *cfg)
-+{
-+ struct dpni_cmd_set_tx_priorities *cmd_params;
-+ struct fsl_mc_command cmd = { 0 };
-+ int i;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPNI_CMDID_SET_TX_PRIORITIES,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpni_cmd_set_tx_priorities *)cmd.params;
-+ dpni_set_field(cmd_params->flags,
-+ SEPARATE_GRP,
-+ cfg->separate_groups);
-+ cmd_params->prio_group_A = cfg->prio_group_A;
-+ cmd_params->prio_group_B = cfg->prio_group_B;
-+
-+ for (i = 0; i + 1 < DPNI_MAX_TC; i += 2) {
-+ dpni_set_field(cmd_params->modes[i / 2],
-+ MODE_1,
-+ cfg->tc_sched[i].mode);
-+ dpni_set_field(cmd_params->modes[i / 2],
-+ MODE_2,
-+ cfg->tc_sched[i + 1].mode);
-+ }
-+
-+ for (i = 0; i < DPNI_MAX_TC; i++) {
-+ cmd_params->delta_bandwidth[i] =
-+ cpu_to_le16(cfg->tc_sched[i].delta_bandwidth);
-+ }
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpni_set_rx_tc_dist() - Set Rx traffic class distribution configuration
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPNI object
-+ * @tc_id: Traffic class selection (0-7)
-+ * @cfg: Traffic class distribution configuration
-+ *
-+ * warning: if 'dist_mode != DPNI_DIST_MODE_NONE', call dpni_prepare_key_cfg()
-+ * first to prepare the key_cfg_iova parameter
-+ *
-+ * Return: '0' on Success; error code otherwise.
-+ */
-+int dpni_set_rx_tc_dist(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 tc_id,
-+ const struct dpni_rx_tc_dist_cfg *cfg)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpni_cmd_set_rx_tc_dist *cmd_params;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPNI_CMDID_SET_RX_TC_DIST,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpni_cmd_set_rx_tc_dist *)cmd.params;
-+ cmd_params->dist_size = cpu_to_le16(cfg->dist_size);
-+ cmd_params->tc_id = tc_id;
-+ dpni_set_field(cmd_params->flags, DIST_MODE, cfg->dist_mode);
-+ dpni_set_field(cmd_params->flags, MISS_ACTION, cfg->fs_cfg.miss_action);
-+ cmd_params->default_flow_id = cpu_to_le16(cfg->fs_cfg.default_flow_id);
-+ cmd_params->key_cfg_iova = cpu_to_le64(cfg->key_cfg_iova);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/*
-+ * dpni_set_qos_table() - Set QoS mapping table
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPNI object
-+ * @cfg: QoS table configuration
-+ *
-+ * This function and all QoS-related functions require that
-+ *'max_tcs > 1' was set at DPNI creation.
-+ *
-+ * warning: Before calling this function, call dpkg_prepare_key_cfg() to
-+ * prepare the key_cfg_iova parameter
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpni_set_qos_table(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ const struct dpni_qos_tbl_cfg *cfg)
-+{
-+ struct dpni_cmd_set_qos_table *cmd_params;
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPNI_CMDID_SET_QOS_TBL,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpni_cmd_set_qos_table *)cmd.params;
-+ cmd_params->default_tc = cfg->default_tc;
-+ cmd_params->key_cfg_iova = cpu_to_le64(cfg->key_cfg_iova);
-+ dpni_set_field(cmd_params->discard_on_miss,
-+ ENABLE,
-+ cfg->discard_on_miss);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpni_add_qos_entry() - Add QoS mapping entry (to select a traffic class)
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPNI object
-+ * @cfg: QoS rule to add
-+ * @tc_id: Traffic class selection (0-7)
-+ * @index: Location in the QoS table where to insert the entry.
-+ * Only relevant if MASKING is enabled for QoS classification on
-+ * this DPNI, it is ignored for exact match.
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpni_add_qos_entry(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ const struct dpni_rule_cfg *cfg,
-+ u8 tc_id,
-+ u16 index)
-+{
-+ struct dpni_cmd_add_qos_entry *cmd_params;
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPNI_CMDID_ADD_QOS_ENT,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpni_cmd_add_qos_entry *)cmd.params;
-+ cmd_params->tc_id = tc_id;
-+ cmd_params->key_size = cfg->key_size;
-+ cmd_params->index = cpu_to_le16(index);
-+ cmd_params->key_iova = cpu_to_le64(cfg->key_iova);
-+ cmd_params->mask_iova = cpu_to_le64(cfg->mask_iova);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpni_remove_qos_entry() - Remove QoS mapping entry
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPNI object
-+ * @cfg: QoS rule to remove
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpni_remove_qos_entry(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ const struct dpni_rule_cfg *cfg)
-+{
-+ struct dpni_cmd_remove_qos_entry *cmd_params;
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPNI_CMDID_REMOVE_QOS_ENT,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpni_cmd_remove_qos_entry *)cmd.params;
-+ cmd_params->key_size = cfg->key_size;
-+ cmd_params->key_iova = cpu_to_le64(cfg->key_iova);
-+ cmd_params->mask_iova = cpu_to_le64(cfg->mask_iova);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpni_add_fs_entry() - Add Flow Steering entry for a specific traffic class
-+ * (to select a flow ID)
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPNI object
-+ * @tc_id: Traffic class selection (0-7)
-+ * @index: Location in the QoS table where to insert the entry.
-+ * Only relevant if MASKING is enabled for QoS
-+ * classification on this DPNI, it is ignored for exact match.
-+ * @cfg: Flow steering rule to add
-+ * @action: Action to be taken as result of a classification hit
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpni_add_fs_entry(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 tc_id,
-+ u16 index,
-+ const struct dpni_rule_cfg *cfg,
-+ const struct dpni_fs_action_cfg *action)
-+{
-+ struct dpni_cmd_add_fs_entry *cmd_params;
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPNI_CMDID_ADD_FS_ENT,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpni_cmd_add_fs_entry *)cmd.params;
-+ cmd_params->tc_id = tc_id;
-+ cmd_params->key_size = cfg->key_size;
-+ cmd_params->index = cpu_to_le16(index);
-+ cmd_params->key_iova = cpu_to_le64(cfg->key_iova);
-+ cmd_params->mask_iova = cpu_to_le64(cfg->mask_iova);
-+ cmd_params->options = cpu_to_le16(action->options);
-+ cmd_params->flow_id = cpu_to_le16(action->flow_id);
-+ cmd_params->flc = cpu_to_le64(action->flc);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpni_remove_fs_entry() - Remove Flow Steering entry from a specific
-+ * traffic class
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPNI object
-+ * @tc_id: Traffic class selection (0-7)
-+ * @cfg: Flow steering rule to remove
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpni_remove_fs_entry(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 tc_id,
-+ const struct dpni_rule_cfg *cfg)
-+{
-+ struct dpni_cmd_remove_fs_entry *cmd_params;
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPNI_CMDID_REMOVE_FS_ENT,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpni_cmd_remove_fs_entry *)cmd.params;
-+ cmd_params->tc_id = tc_id;
-+ cmd_params->key_size = cfg->key_size;
-+ cmd_params->key_iova = cpu_to_le64(cfg->key_iova);
-+ cmd_params->mask_iova = cpu_to_le64(cfg->mask_iova);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpni_set_congestion_notification() - Set traffic class congestion
-+ * notification configuration
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPNI object
-+ * @qtype: Type of queue - Rx, Tx and Tx confirm types are supported
-+ * @tc_id: Traffic class selection (0-7)
-+ * @cfg: Congestion notification configuration
-+ *
-+ * Return: '0' on Success; error code otherwise.
-+ */
-+int dpni_set_congestion_notification(
-+ struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ enum dpni_queue_type qtype,
-+ u8 tc_id,
-+ const struct dpni_congestion_notification_cfg *cfg)
-+{
-+ struct dpni_cmd_set_congestion_notification *cmd_params;
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(
-+ DPNI_CMDID_SET_CONGESTION_NOTIFICATION,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpni_cmd_set_congestion_notification *)cmd.params;
-+ cmd_params->qtype = qtype;
-+ cmd_params->tc = tc_id;
-+ cmd_params->dest_id = cpu_to_le32(cfg->dest_cfg.dest_id);
-+ cmd_params->notification_mode = cpu_to_le16(cfg->notification_mode);
-+ cmd_params->dest_priority = cfg->dest_cfg.priority;
-+ dpni_set_field(cmd_params->type_units, DEST_TYPE,
-+ cfg->dest_cfg.dest_type);
-+ dpni_set_field(cmd_params->type_units, CONG_UNITS, cfg->units);
-+ cmd_params->message_iova = cpu_to_le64(cfg->message_iova);
-+ cmd_params->message_ctx = cpu_to_le64(cfg->message_ctx);
-+ cmd_params->threshold_entry = cpu_to_le32(cfg->threshold_entry);
-+ cmd_params->threshold_exit = cpu_to_le32(cfg->threshold_exit);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpni_get_congestion_notification() - Get traffic class congestion
-+ * notification configuration
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPNI object
-+ * @qtype: Type of queue - Rx, Tx and Tx confirm types are supported
-+ * @tc_id: bits 7-4 contain ceetm channel index (valid only for TX);
-+ * bits 3-0 contain traffic class.
-+ * Use macro DPNI_BUILD_CH_TC() to build correct value for
-+ * tc_id parameter.
-+ * @cfg: congestion notification configuration
-+ *
-+ * Return: '0' on Success; error code otherwise.
-+ */
-+int dpni_get_congestion_notification(
-+ struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ enum dpni_queue_type qtype,
-+ u8 tc_id,
-+ struct dpni_congestion_notification_cfg *cfg)
-+{
-+ struct dpni_rsp_get_congestion_notification *rsp_params;
-+ struct dpni_cmd_get_congestion_notification *cmd_params;
-+ struct fsl_mc_command cmd = { 0 };
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(
-+ DPNI_CMDID_GET_CONGESTION_NOTIFICATION,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpni_cmd_get_congestion_notification *)cmd.params;
-+ cmd_params->qtype = qtype;
-+ cmd_params->tc = tc_id;
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ rsp_params = (struct dpni_rsp_get_congestion_notification *)cmd.params;
-+ cfg->units = dpni_get_field(rsp_params->type_units, CONG_UNITS);
-+ cfg->threshold_entry = le32_to_cpu(rsp_params->threshold_entry);
-+ cfg->threshold_exit = le32_to_cpu(rsp_params->threshold_exit);
-+ cfg->message_ctx = le64_to_cpu(rsp_params->message_ctx);
-+ cfg->message_iova = le64_to_cpu(rsp_params->message_iova);
-+ cfg->notification_mode = le16_to_cpu(rsp_params->notification_mode);
-+ cfg->dest_cfg.dest_id = le32_to_cpu(rsp_params->dest_id);
-+ cfg->dest_cfg.priority = rsp_params->dest_priority;
-+ cfg->dest_cfg.dest_type = dpni_get_field(rsp_params->type_units,
-+ DEST_TYPE);
-+
-+ return 0;
-+}
-+
-+/**
-+ * dpni_set_queue() - Set queue parameters
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPNI object
-+ * @qtype: Type of queue - all queue types are supported, although
-+ * the command is ignored for Tx
-+ * @tc: Traffic class, in range 0 to NUM_TCS - 1
-+ * @index: Selects the specific queue out of the set allocated for the
-+ * same TC. Value must be in range 0 to NUM_QUEUES - 1
-+ * @options: A combination of DPNI_QUEUE_OPT_ values that control what
-+ * configuration options are set on the queue
-+ * @queue: Queue structure
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpni_set_queue(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ enum dpni_queue_type qtype,
-+ u8 tc,
-+ u8 index,
-+ u8 options,
-+ const struct dpni_queue *queue)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpni_cmd_set_queue *cmd_params;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPNI_CMDID_SET_QUEUE,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpni_cmd_set_queue *)cmd.params;
-+ cmd_params->qtype = qtype;
-+ cmd_params->tc = tc;
-+ cmd_params->index = index;
-+ cmd_params->options = options;
-+ cmd_params->dest_id = cpu_to_le32(queue->destination.id);
-+ cmd_params->dest_prio = queue->destination.priority;
-+ dpni_set_field(cmd_params->flags, DEST_TYPE, queue->destination.type);
-+ dpni_set_field(cmd_params->flags, STASH_CTRL, queue->flc.stash_control);
-+ dpni_set_field(cmd_params->flags, HOLD_ACTIVE,
-+ queue->destination.hold_active);
-+ cmd_params->flc = cpu_to_le64(queue->flc.value);
-+ cmd_params->user_context = cpu_to_le64(queue->user_context);
-+
-+ /* send command to mc */
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpni_get_queue() - Get queue parameters
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPNI object
-+ * @qtype: Type of queue - all queue types are supported
-+ * @tc: Traffic class, in range 0 to NUM_TCS - 1
-+ * @index: Selects the specific queue out of the set allocated for the
-+ * same TC. Value must be in range 0 to NUM_QUEUES - 1
-+ * @queue: Queue configuration structure
-+ * @qid: Queue identification
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpni_get_queue(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ enum dpni_queue_type qtype,
-+ u8 tc,
-+ u8 index,
-+ struct dpni_queue *queue,
-+ struct dpni_queue_id *qid)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpni_cmd_get_queue *cmd_params;
-+ struct dpni_rsp_get_queue *rsp_params;
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPNI_CMDID_GET_QUEUE,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpni_cmd_get_queue *)cmd.params;
-+ cmd_params->qtype = qtype;
-+ cmd_params->tc = tc;
-+ cmd_params->index = index;
-+
-+ /* send command to mc */
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ rsp_params = (struct dpni_rsp_get_queue *)cmd.params;
-+ queue->destination.id = le32_to_cpu(rsp_params->dest_id);
-+ queue->destination.priority = rsp_params->dest_prio;
-+ queue->destination.type = dpni_get_field(rsp_params->flags,
-+ DEST_TYPE);
-+ queue->flc.stash_control = dpni_get_field(rsp_params->flags,
-+ STASH_CTRL);
-+ queue->destination.hold_active = dpni_get_field(rsp_params->flags,
-+ HOLD_ACTIVE);
-+ queue->flc.value = le64_to_cpu(rsp_params->flc);
-+ queue->user_context = le64_to_cpu(rsp_params->user_context);
-+ qid->fqid = le32_to_cpu(rsp_params->fqid);
-+ qid->qdbin = le16_to_cpu(rsp_params->qdbin);
-+
-+ return 0;
-+}
-+
-+/**
-+ * dpni_get_statistics() - Get DPNI statistics
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPNI object
-+ * @page: Selects the statistics page to retrieve, see
-+ * DPNI_GET_STATISTICS output. Pages are numbered 0 to 2.
-+ * @param: Custom parameter for some pages used to select a certain
-+ * statistic source, for example the TC.
-+ * @stat: Structure containing the statistics
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpni_get_statistics(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 page,
-+ u8 param,
-+ union dpni_statistics *stat)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpni_cmd_get_statistics *cmd_params;
-+ struct dpni_rsp_get_statistics *rsp_params;
-+ int i, err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPNI_CMDID_GET_STATISTICS,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpni_cmd_get_statistics *)cmd.params;
-+ cmd_params->page_number = page;
-+ cmd_params->param = param;
-+
-+ /* send command to mc */
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ rsp_params = (struct dpni_rsp_get_statistics *)cmd.params;
-+ for (i = 0; i < DPNI_STATISTICS_CNT; i++)
-+ stat->raw.counter[i] = le64_to_cpu(rsp_params->counter[i]);
-+
-+ return 0;
-+}
-+
-+/**
-+ * dpni_reset_statistics() - Clears DPNI statistics
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPNI object
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpni_reset_statistics(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPNI_CMDID_RESET_STATISTICS,
-+ cmd_flags,
-+ token);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpni_set_taildrop() - Set taildrop per queue or TC
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPNI object
-+ * @cg_point: Congestion point
-+ * @q_type: Queue type on which the taildrop is configured.
-+ * Only Rx queues are supported for now
-+ * @tc: bits 7-4 contain ceetm channel index (valid only for TX);
-+ * bits 3-0 contain traffic class.
-+ * Use macro DPNI_BUILD_CH_TC() to build correct value for
-+ * tc parameter.
-+ * @q_index: Index of the queue if the DPNI supports multiple queues for
-+ * traffic distribution. Ignored if CONGESTION_POINT is not 0.
-+ * @taildrop: Taildrop structure
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpni_set_taildrop(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ enum dpni_congestion_point cg_point,
-+ enum dpni_queue_type qtype,
-+ u8 tc,
-+ u8 index,
-+ struct dpni_taildrop *taildrop)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpni_cmd_set_taildrop *cmd_params;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPNI_CMDID_SET_TAILDROP,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpni_cmd_set_taildrop *)cmd.params;
-+ cmd_params->congestion_point = cg_point;
-+ cmd_params->qtype = qtype;
-+ cmd_params->tc = tc;
-+ cmd_params->index = index;
-+ dpni_set_field(cmd_params->enable, ENABLE, taildrop->enable);
-+ cmd_params->units = taildrop->units;
-+ cmd_params->threshold = cpu_to_le32(taildrop->threshold);
-+
-+ /* send command to mc */
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpni_get_taildrop() - Get taildrop information
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPNI object
-+ * @cg_point: Congestion point
-+ * @q_type: Queue type on which the taildrop is configured.
-+ * Only Rx queues are supported for now
-+ * @tc: bits 7-4 contain ceetm channel index (valid only for TX);
-+ * bits 3-0 contain traffic class.
-+ * Use macro DPNI_BUILD_CH_TC() to build correct value for
-+ * tc parameter.
-+ * @q_index: Index of the queue if the DPNI supports multiple queues for
-+ * traffic distribution. Ignored if CONGESTION_POINT is not 0.
-+ * @taildrop: Taildrop structure
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpni_get_taildrop(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ enum dpni_congestion_point cg_point,
-+ enum dpni_queue_type qtype,
-+ u8 tc,
-+ u8 index,
-+ struct dpni_taildrop *taildrop)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpni_cmd_get_taildrop *cmd_params;
-+ struct dpni_rsp_get_taildrop *rsp_params;
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPNI_CMDID_GET_TAILDROP,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpni_cmd_get_taildrop *)cmd.params;
-+ cmd_params->congestion_point = cg_point;
-+ cmd_params->qtype = qtype;
-+ cmd_params->tc = tc;
-+ cmd_params->index = index;
-+
-+ /* send command to mc */
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ rsp_params = (struct dpni_rsp_get_taildrop *)cmd.params;
-+ taildrop->enable = dpni_get_field(rsp_params->enable, ENABLE);
-+ taildrop->units = rsp_params->units;
-+ taildrop->threshold = le32_to_cpu(rsp_params->threshold);
-+
-+ return 0;
-+}
-+
-+/**
-+ * dpni_get_api_version() - Get Data Path Network Interface API version
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @major_ver: Major version of data path network interface API
-+ * @minor_ver: Minor version of data path network interface API
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpni_get_api_version(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 *major_ver,
-+ u16 *minor_ver)
-+{
-+ struct dpni_rsp_get_api_version *rsp_params;
-+ struct fsl_mc_command cmd = { 0 };
-+ int err;
-+
-+ cmd.header = mc_encode_cmd_header(DPNI_CMDID_GET_API_VERSION,
-+ cmd_flags, 0);
-+
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ rsp_params = (struct dpni_rsp_get_api_version *)cmd.params;
-+ *major_ver = le16_to_cpu(rsp_params->major);
-+ *minor_ver = le16_to_cpu(rsp_params->minor);
-+
-+ return 0;
-+}
-+
-+/**
-+ * dpni_set_rx_fs_dist() - Set Rx traffic class FS distribution
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPNI object
-+ * @cfg: Distribution configuration
-+ * If the FS is already enabled with a previous call the classification
-+ * key will be changed but all the table rules are kept. If the
-+ * existing rules do not match the key the results will not be
-+ * predictable. It is the user responsibility to keep key integrity.
-+ * If cfg.enable is set to 1 the command will create a flow steering table
-+ * and will classify packets according to this table. The packets that
-+ * miss all the table rules will be classified according to settings
-+ * made in dpni_set_rx_hash_dist()
-+ * If cfg.enable is set to 0 the command will clear flow steering table.
-+ * The packets will be classified according to settings made in
-+ * dpni_set_rx_hash_dist()
-+ */
-+int dpni_set_rx_fs_dist(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ const struct dpni_rx_dist_cfg *cfg)
-+{
-+ struct dpni_cmd_set_rx_fs_dist *cmd_params;
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPNI_CMDID_SET_RX_FS_DIST,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpni_cmd_set_rx_fs_dist *)cmd.params;
-+ cmd_params->dist_size = le16_to_cpu(cfg->dist_size);
-+ dpni_set_field(cmd_params->enable, RX_FS_DIST_ENABLE, cfg->enable);
-+ cmd_params->tc = cfg->tc;
-+ cmd_params->miss_flow_id = le16_to_cpu(cfg->fs_miss_flow_id);
-+ cmd_params->key_cfg_iova = le64_to_cpu(cfg->key_cfg_iova);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpni_set_rx_hash_dist() - Set Rx traffic class HASH distribution
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPNI object
-+ * @cfg: Distribution configuration
-+ * If cfg.enable is set to 1 the packets will be classified using a hash
-+ * function based on the key received in cfg.key_cfg_iova parameter.
-+ * If cfg.enable is set to 0 the packets will be sent to the queue configured
-+ * in dpni_set_rx_dist_default_queue() call
-+ */
-+int dpni_set_rx_hash_dist(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ const struct dpni_rx_dist_cfg *cfg)
-+{
-+ struct dpni_cmd_set_rx_hash_dist *cmd_params;
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPNI_CMDID_SET_RX_HASH_DIST,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpni_cmd_set_rx_hash_dist *)cmd.params;
-+ cmd_params->dist_size = le16_to_cpu(cfg->dist_size);
-+ dpni_set_field(cmd_params->enable, RX_FS_DIST_ENABLE, cfg->enable);
-+ cmd_params->tc = cfg->tc;
-+ cmd_params->key_cfg_iova = le64_to_cpu(cfg->key_cfg_iova);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
---- /dev/null
-+++ b/drivers/staging/fsl-dpaa2/ethernet/dpni.h
-@@ -0,0 +1,1172 @@
-+/* Copyright 2013-2016 Freescale Semiconductor Inc.
-+ * Copyright 2016 NXP
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of the above-listed copyright holders nor the
-+ * names of any contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
-+ * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
-+ * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
-+ * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE
-+ * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
-+ * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
-+ * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
-+ * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
-+ * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
-+ * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
-+ * POSSIBILITY OF SUCH DAMAGE.
-+ */
-+#ifndef __FSL_DPNI_H
-+#define __FSL_DPNI_H
-+
-+#include "dpkg.h"
-+
-+struct fsl_mc_io;
-+
-+/**
-+ * Data Path Network Interface API
-+ * Contains initialization APIs and runtime control APIs for DPNI
-+ */
-+
-+/** General DPNI macros */
-+
-+/**
-+ * Maximum number of traffic classes
-+ */
-+#define DPNI_MAX_TC 8
-+/**
-+ * Maximum number of buffer pools per DPNI
-+ */
-+#define DPNI_MAX_DPBP 8
-+/**
-+ * Maximum number of senders
-+ */
-+#define DPNI_MAX_SENDERS 16
-+/**
-+ * Maximum distribution size
-+ */
-+#define DPNI_MAX_DIST_SIZE 16
-+
-+/**
-+ * All traffic classes considered; see dpni_set_queue()
-+ */
-+#define DPNI_ALL_TCS (u8)(-1)
-+/**
-+ * All flows within traffic class considered; see dpni_set_queue()
-+ */
-+#define DPNI_ALL_TC_FLOWS (u16)(-1)
-+/**
-+ * Generate new flow ID; see dpni_set_queue()
-+ */
-+#define DPNI_NEW_FLOW_ID (u16)(-1)
-+
-+/**
-+ * Tx traffic is always released to a buffer pool on transmit, there are no
-+ * resources allocated to have the frames confirmed back to the source after
-+ * transmission.
-+ */
-+#define DPNI_OPT_TX_FRM_RELEASE 0x000001
-+/**
-+ * Disables support for MAC address filtering for addresses other than primary
-+ * MAC address. This affects both unicast and multicast. Promiscuous mode can
-+ * still be enabled/disabled for both unicast and multicast. If promiscuous mode
-+ * is disabled, only traffic matching the primary MAC address will be accepted.
-+ */
-+#define DPNI_OPT_NO_MAC_FILTER 0x000002
-+/**
-+ * Allocate policers for this DPNI. They can be used to rate-limit traffic per
-+ * traffic class (TC) basis.
-+ */
-+#define DPNI_OPT_HAS_POLICING 0x000004
-+/**
-+ * Congestion can be managed in several ways, allowing the buffer pool to
-+ * deplete on ingress, taildrop on each queue or use congestion groups for sets
-+ * of queues. If set, it configures a single congestion groups across all TCs.
-+ * If reset, a congestion group is allocated for each TC. Only relevant if the
-+ * DPNI has multiple traffic classes.
-+ */
-+#define DPNI_OPT_SHARED_CONGESTION 0x000008
-+/**
-+ * Enables TCAM for Flow Steering and QoS look-ups. If not specified, all
-+ * look-ups are exact match. Note that TCAM is not available on LS1088 and its
-+ * variants. Setting this bit on these SoCs will trigger an error.
-+ */
-+#define DPNI_OPT_HAS_KEY_MASKING 0x000010
-+/**
-+ * Disables the flow steering table.
-+ */
-+#define DPNI_OPT_NO_FS 0x000020
-+
-+int dpni_open(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ int dpni_id,
-+ u16 *token);
-+
-+int dpni_close(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token);
-+
-+/**
-+ * struct dpni_pools_cfg - Structure representing buffer pools configuration
-+ * @num_dpbp: Number of DPBPs
-+ * @pools: Array of buffer pools parameters; The number of valid entries
-+ * must match 'num_dpbp' value
-+ */
-+struct dpni_pools_cfg {
-+ u8 num_dpbp;
-+ /**
-+ * struct pools - Buffer pools parameters
-+ * @dpbp_id: DPBP object ID
-+ * @priority_mask: priorities served by DPBP
-+ * @buffer_size: Buffer size
-+ * @backup_pool: Backup pool
-+ */
-+ struct {
-+ u16 dpbp_id;
-+ u8 priority_mask;
-+ u16 buffer_size;
-+ u8 backup_pool;
-+ } pools[DPNI_MAX_DPBP];
-+};
-+
-+int dpni_set_pools(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ const struct dpni_pools_cfg *cfg);
-+
-+int dpni_enable(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token);
-+
-+int dpni_disable(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token);
-+
-+int dpni_is_enabled(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ int *en);
-+
-+int dpni_reset(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token);
-+
-+/**
-+ * DPNI IRQ Index and Events
-+ */
-+
-+/**
-+ * IRQ index
-+ */
-+#define DPNI_IRQ_INDEX 0
-+/**
-+ * IRQ event - indicates a change in link state
-+ */
-+#define DPNI_IRQ_EVENT_LINK_CHANGED 0x00000001
-+
-+int dpni_set_irq_enable(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 irq_index,
-+ u8 en);
-+
-+int dpni_get_irq_enable(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 irq_index,
-+ u8 *en);
-+
-+int dpni_set_irq_mask(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 irq_index,
-+ u32 mask);
-+
-+int dpni_get_irq_mask(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 irq_index,
-+ u32 *mask);
-+
-+int dpni_get_irq_status(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 irq_index,
-+ u32 *status);
-+
-+int dpni_clear_irq_status(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 irq_index,
-+ u32 status);
-+
-+/**
-+ * struct dpni_attr - Structure representing DPNI attributes
-+ * @options: Any combination of the following options:
-+ * DPNI_OPT_TX_FRM_RELEASE
-+ * DPNI_OPT_NO_MAC_FILTER
-+ * DPNI_OPT_HAS_POLICING
-+ * DPNI_OPT_SHARED_CONGESTION
-+ * DPNI_OPT_HAS_KEY_MASKING
-+ * DPNI_OPT_NO_FS
-+ * @num_queues: Number of Tx and Rx queues used for traffic distribution.
-+ * @num_tcs: Number of traffic classes (TCs), reserved for the DPNI.
-+ * @mac_filter_entries: Number of entries in the MAC address filtering table.
-+ * @vlan_filter_entries: Number of entries in the VLAN address filtering table.
-+ * @qos_entries: Number of entries in the QoS classification table.
-+ * @fs_entries: Number of entries in the flow steering table.
-+ * @qos_key_size: Size, in bytes, of the QoS look-up key. Defining a key larger
-+ * than this when adding QoS entries will result in an error.
-+ * @fs_key_size: Size, in bytes, of the flow steering look-up key. Defining a
-+ * key larger than this when composing the hash + FS key will
-+ * result in an error.
-+ * @wriop_version: Version of WRIOP HW block. The 3 version values are stored
-+ * on 6, 5, 5 bits respectively.
-+ */
-+struct dpni_attr {
-+ u32 options;
-+ u8 num_queues;
-+ u8 num_tcs;
-+ u8 mac_filter_entries;
-+ u8 vlan_filter_entries;
-+ u8 qos_entries;
-+ u16 fs_entries;
-+ u8 qos_key_size;
-+ u8 fs_key_size;
-+ u16 wriop_version;
-+};
-+
-+int dpni_get_attributes(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ struct dpni_attr *attr);
-+
-+/**
-+ * DPNI errors
-+ */
-+
-+/**
-+ * Extract out of frame header error
-+ */
-+#define DPNI_ERROR_EOFHE 0x00020000
-+/**
-+ * Frame length error
-+ */
-+#define DPNI_ERROR_FLE 0x00002000
-+/**
-+ * Frame physical error
-+ */
-+#define DPNI_ERROR_FPE 0x00001000
-+/**
-+ * Parsing header error
-+ */
-+#define DPNI_ERROR_PHE 0x00000020
-+/**
-+ * Parser L3 checksum error
-+ */
-+#define DPNI_ERROR_L3CE 0x00000004
-+/**
-+ * Parser L3 checksum error
-+ */
-+#define DPNI_ERROR_L4CE 0x00000001
-+
-+/**
-+ * enum dpni_error_action - Defines DPNI behavior for errors
-+ * @DPNI_ERROR_ACTION_DISCARD: Discard the frame
-+ * @DPNI_ERROR_ACTION_CONTINUE: Continue with the normal flow
-+ * @DPNI_ERROR_ACTION_SEND_TO_ERROR_QUEUE: Send the frame to the error queue
-+ */
-+enum dpni_error_action {
-+ DPNI_ERROR_ACTION_DISCARD = 0,
-+ DPNI_ERROR_ACTION_CONTINUE = 1,
-+ DPNI_ERROR_ACTION_SEND_TO_ERROR_QUEUE = 2
-+};
-+
-+/**
-+ * struct dpni_error_cfg - Structure representing DPNI errors treatment
-+ * @errors: Errors mask; use 'DPNI_ERROR__<X>
-+ * @error_action: The desired action for the errors mask
-+ * @set_frame_annotation: Set to '1' to mark the errors in frame annotation
-+ * status (FAS); relevant only for the non-discard action
-+ */
-+struct dpni_error_cfg {
-+ u32 errors;
-+ enum dpni_error_action error_action;
-+ int set_frame_annotation;
-+};
-+
-+int dpni_set_errors_behavior(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ struct dpni_error_cfg *cfg);
-+
-+/**
-+ * DPNI buffer layout modification options
-+ */
-+
-+/**
-+ * Select to modify the time-stamp setting
-+ */
-+#define DPNI_BUF_LAYOUT_OPT_TIMESTAMP 0x00000001
-+/**
-+ * Select to modify the parser-result setting; not applicable for Tx
-+ */
-+#define DPNI_BUF_LAYOUT_OPT_PARSER_RESULT 0x00000002
-+/**
-+ * Select to modify the frame-status setting
-+ */
-+#define DPNI_BUF_LAYOUT_OPT_FRAME_STATUS 0x00000004
-+/**
-+ * Select to modify the private-data-size setting
-+ */
-+#define DPNI_BUF_LAYOUT_OPT_PRIVATE_DATA_SIZE 0x00000008
-+/**
-+ * Select to modify the data-alignment setting
-+ */
-+#define DPNI_BUF_LAYOUT_OPT_DATA_ALIGN 0x00000010
-+/**
-+ * Select to modify the data-head-room setting
-+ */
-+#define DPNI_BUF_LAYOUT_OPT_DATA_HEAD_ROOM 0x00000020
-+/**
-+ * Select to modify the data-tail-room setting
-+ */
-+#define DPNI_BUF_LAYOUT_OPT_DATA_TAIL_ROOM 0x00000040
-+
-+/**
-+ * struct dpni_buffer_layout - Structure representing DPNI buffer layout
-+ * @options: Flags representing the suggested modifications to the buffer
-+ * layout; Use any combination of 'DPNI_BUF_LAYOUT_OPT_<X>' flags
-+ * @pass_timestamp: Pass timestamp value
-+ * @pass_parser_result: Pass parser results
-+ * @pass_frame_status: Pass frame status
-+ * @private_data_size: Size kept for private data (in bytes)
-+ * @data_align: Data alignment
-+ * @data_head_room: Data head room
-+ * @data_tail_room: Data tail room
-+ */
-+struct dpni_buffer_layout {
-+ u32 options;
-+ int pass_timestamp;
-+ int pass_parser_result;
-+ int pass_frame_status;
-+ u16 private_data_size;
-+ u16 data_align;
-+ u16 data_head_room;
-+ u16 data_tail_room;
-+};
-+
-+/**
-+ * enum dpni_queue_type - Identifies a type of queue targeted by the command
-+ * @DPNI_QUEUE_RX: Rx queue
-+ * @DPNI_QUEUE_TX: Tx queue
-+ * @DPNI_QUEUE_TX_CONFIRM: Tx confirmation queue
-+ * @DPNI_QUEUE_RX_ERR: Rx error queue
-+ */enum dpni_queue_type {
-+ DPNI_QUEUE_RX,
-+ DPNI_QUEUE_TX,
-+ DPNI_QUEUE_TX_CONFIRM,
-+ DPNI_QUEUE_RX_ERR,
-+};
-+
-+int dpni_get_buffer_layout(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ enum dpni_queue_type qtype,
-+ struct dpni_buffer_layout *layout);
-+
-+int dpni_set_buffer_layout(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ enum dpni_queue_type qtype,
-+ const struct dpni_buffer_layout *layout);
-+
-+/**
-+ * enum dpni_offload - Identifies a type of offload targeted by the command
-+ * @DPNI_OFF_RX_L3_CSUM: Rx L3 checksum validation
-+ * @DPNI_OFF_RX_L4_CSUM: Rx L4 checksum validation
-+ * @DPNI_OFF_TX_L3_CSUM: Tx L3 checksum generation
-+ * @DPNI_OFF_TX_L4_CSUM: Tx L4 checksum generation
-+ */
-+enum dpni_offload {
-+ DPNI_OFF_RX_L3_CSUM,
-+ DPNI_OFF_RX_L4_CSUM,
-+ DPNI_OFF_TX_L3_CSUM,
-+ DPNI_OFF_TX_L4_CSUM,
-+};
-+
-+int dpni_set_offload(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ enum dpni_offload type,
-+ u32 config);
-+
-+int dpni_get_offload(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ enum dpni_offload type,
-+ u32 *config);
-+
-+int dpni_get_qdid(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ enum dpni_queue_type qtype,
-+ u16 *qdid);
-+
-+int dpni_get_tx_data_offset(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 *data_offset);
-+
-+#define DPNI_STATISTICS_CNT 7
-+
-+union dpni_statistics {
-+ /**
-+ * struct page_0 - Page_0 statistics structure
-+ * @ingress_all_frames: Ingress frame count
-+ * @ingress_all_bytes: Ingress byte count
-+ * @ingress_multicast_frames: Ingress multicast frame count
-+ * @ingress_multicast_bytes: Ingress multicast byte count
-+ * @ingress_broadcast_frames: Ingress broadcast frame count
-+ * @ingress_broadcast_bytes: Ingress broadcast byte count
-+ */
-+ struct {
-+ u64 ingress_all_frames;
-+ u64 ingress_all_bytes;
-+ u64 ingress_multicast_frames;
-+ u64 ingress_multicast_bytes;
-+ u64 ingress_broadcast_frames;
-+ u64 ingress_broadcast_bytes;
-+ } page_0;
-+ /**
-+ * struct page_1 - Page_1 statistics structure
-+ * @egress_all_frames: Egress frame count
-+ * @egress_all_bytes: Egress byte count
-+ * @egress_multicast_frames: Egress multicast frame count
-+ * @egress_multicast_bytes: Egress multicast byte count
-+ * @egress_broadcast_frames: Egress broadcast frame count
-+ * @egress_broadcast_bytes: Egress broadcast byte count
-+ */
-+ struct {
-+ u64 egress_all_frames;
-+ u64 egress_all_bytes;
-+ u64 egress_multicast_frames;
-+ u64 egress_multicast_bytes;
-+ u64 egress_broadcast_frames;
-+ u64 egress_broadcast_bytes;
-+ } page_1;
-+ /**
-+ * struct page_2 - Page_2 statistics structure
-+ * @ingress_filtered_frames: Ingress filtered frame count
-+ * @ingress_discarded_frames: Ingress discarded frame count
-+ * @ingress_nobuffer_discards: Ingress discarded frame count
-+ * due to lack of buffers
-+ * @egress_discarded_frames: Egress discarded frame count
-+ * @egress_confirmed_frames: Egress confirmed frame count
-+ */
-+ struct {
-+ u64 ingress_filtered_frames;
-+ u64 ingress_discarded_frames;
-+ u64 ingress_nobuffer_discards;
-+ u64 egress_discarded_frames;
-+ u64 egress_confirmed_frames;
-+ } page_2;
-+ /**
-+ * struct page_3 - Page_3 statistics structure with values for the
-+ * selected TC
-+ * @ceetm_dequeue_bytes: Cumulative count of the number of bytes
-+ * dequeued
-+ * @ceetm_dequeue_frames: Cumulative count of the number of frames
-+ * dequeued
-+ * @ceetm_reject_bytes: Cumulative count of the number of bytes in all
-+ * frames whose enqueue was rejected
-+ * @ceetm_reject_frames: Cumulative count of all frame enqueues
-+ * rejected
-+ */
-+ struct {
-+ u64 ceetm_dequeue_bytes;
-+ u64 ceetm_dequeue_frames;
-+ u64 ceetm_reject_bytes;
-+ u64 ceetm_reject_frames;
-+ } page_3;
-+ /**
-+ * struct raw - raw statistics structure
-+ */
-+ struct {
-+ u64 counter[DPNI_STATISTICS_CNT];
-+ } raw;
-+};
-+
-+int dpni_get_statistics(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 page,
-+ u8 param,
-+ union dpni_statistics *stat);
-+
-+int dpni_reset_statistics(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token);
-+
-+/**
-+ * Enable auto-negotiation
-+ */
-+#define DPNI_LINK_OPT_AUTONEG 0x0000000000000001ULL
-+/**
-+ * Enable half-duplex mode
-+ */
-+#define DPNI_LINK_OPT_HALF_DUPLEX 0x0000000000000002ULL
-+/**
-+ * Enable pause frames
-+ */
-+#define DPNI_LINK_OPT_PAUSE 0x0000000000000004ULL
-+/**
-+ * Enable a-symmetric pause frames
-+ */
-+#define DPNI_LINK_OPT_ASYM_PAUSE 0x0000000000000008ULL
-+/**
-+ * Enable priority flow control pause frames
-+ */
-+#define DPNI_LINK_OPT_PFC_PAUSE 0x0000000000000010ULL
-+
-+/**
-+ * struct - Structure representing DPNI link configuration
-+ * @rate: Rate
-+ * @options: Mask of available options; use 'DPNI_LINK_OPT_<X>' values
-+ */
-+struct dpni_link_cfg {
-+ u32 rate;
-+ u64 options;
-+};
-+
-+int dpni_set_link_cfg(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ const struct dpni_link_cfg *cfg);
-+
-+/**
-+ * struct dpni_link_state - Structure representing DPNI link state
-+ * @rate: Rate
-+ * @options: Mask of available options; use 'DPNI_LINK_OPT_<X>' values
-+ * @up: Link state; '0' for down, '1' for up
-+ */
-+struct dpni_link_state {
-+ u32 rate;
-+ u64 options;
-+ int up;
-+};
-+
-+int dpni_get_link_state(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ struct dpni_link_state *state);
-+
-+/**
-+ * struct dpni_tx_shaping - Structure representing DPNI tx shaping configuration
-+ * @rate_limit: rate in Mbps
-+ * @max_burst_size: burst size in bytes (up to 64KB)
-+ */
-+struct dpni_tx_shaping_cfg {
-+ u32 rate_limit;
-+ u16 max_burst_size;
-+};
-+
-+int dpni_set_tx_shaping(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ const struct dpni_tx_shaping_cfg *tx_cr_shaper,
-+ const struct dpni_tx_shaping_cfg *tx_er_shaper,
-+ int coupled);
-+
-+int dpni_set_max_frame_length(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 max_frame_length);
-+
-+int dpni_get_max_frame_length(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 *max_frame_length);
-+
-+int dpni_set_multicast_promisc(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ int en);
-+
-+int dpni_get_multicast_promisc(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ int *en);
-+
-+int dpni_set_unicast_promisc(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ int en);
-+
-+int dpni_get_unicast_promisc(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ int *en);
-+
-+int dpni_set_primary_mac_addr(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ const u8 mac_addr[6]);
-+
-+int dpni_get_primary_mac_addr(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 mac_addr[6]);
-+
-+int dpni_get_port_mac_addr(struct fsl_mc_io *mc_io,
-+ u32 cm_flags,
-+ u16 token,
-+ u8 mac_addr[6]);
-+
-+int dpni_add_mac_addr(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ const u8 mac_addr[6]);
-+
-+int dpni_remove_mac_addr(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ const u8 mac_addr[6]);
-+
-+int dpni_clear_mac_filters(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ int unicast,
-+ int multicast);
-+
-+/**
-+ * enum dpni_dist_mode - DPNI distribution mode
-+ * @DPNI_DIST_MODE_NONE: No distribution
-+ * @DPNI_DIST_MODE_HASH: Use hash distribution; only relevant if
-+ * the 'DPNI_OPT_DIST_HASH' option was set at DPNI creation
-+ * @DPNI_DIST_MODE_FS: Use explicit flow steering; only relevant if
-+ * the 'DPNI_OPT_DIST_FS' option was set at DPNI creation
-+ */
-+enum dpni_dist_mode {
-+ DPNI_DIST_MODE_NONE = 0,
-+ DPNI_DIST_MODE_HASH = 1,
-+ DPNI_DIST_MODE_FS = 2
-+};
-+
-+/**
-+ * enum dpni_fs_miss_action - DPNI Flow Steering miss action
-+ * @DPNI_FS_MISS_DROP: In case of no-match, drop the frame
-+ * @DPNI_FS_MISS_EXPLICIT_FLOWID: In case of no-match, use explicit flow-id
-+ * @DPNI_FS_MISS_HASH: In case of no-match, distribute using hash
-+ */
-+enum dpni_fs_miss_action {
-+ DPNI_FS_MISS_DROP = 0,
-+ DPNI_FS_MISS_EXPLICIT_FLOWID = 1,
-+ DPNI_FS_MISS_HASH = 2
-+};
-+
-+/**
-+ * struct dpni_fs_tbl_cfg - Flow Steering table configuration
-+ * @miss_action: Miss action selection
-+ * @default_flow_id: Used when 'miss_action = DPNI_FS_MISS_EXPLICIT_FLOWID'
-+ */
-+struct dpni_fs_tbl_cfg {
-+ enum dpni_fs_miss_action miss_action;
-+ u16 default_flow_id;
-+};
-+
-+int dpni_prepare_key_cfg(const struct dpkg_profile_cfg *cfg,
-+ u8 *key_cfg_buf);
-+
-+/**
-+ * struct dpni_qos_tbl_cfg - Structure representing QOS table configuration
-+ * @key_cfg_iova: I/O virtual address of 256 bytes DMA-able memory filled with
-+ * key extractions to be used as the QoS criteria by calling
-+ * dpkg_prepare_key_cfg()
-+ * @discard_on_miss: Set to '1' to discard frames in case of no match (miss);
-+ * '0' to use the 'default_tc' in such cases
-+ * @default_tc: Used in case of no-match and 'discard_on_miss'= 0
-+ */
-+struct dpni_qos_tbl_cfg {
-+ u64 key_cfg_iova;
-+ int discard_on_miss;
-+ u8 default_tc;
-+};
-+
-+int dpni_set_qos_table(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ const struct dpni_qos_tbl_cfg *cfg);
-+
-+/**
-+ * enum dpni_tx_schedule_mode - DPNI Tx scheduling mode
-+ * @DPNI_TX_SCHED_STRICT_PRIORITY: strict priority
-+ * @DPNI_TX_SCHED_WEIGHTED_A: weighted based scheduling in group A
-+ * @DPNI_TX_SCHED_WEIGHTED_B: weighted based scheduling in group B
-+ */
-+enum dpni_tx_schedule_mode {
-+ DPNI_TX_SCHED_STRICT_PRIORITY = 0,
-+ DPNI_TX_SCHED_WEIGHTED_A,
-+ DPNI_TX_SCHED_WEIGHTED_B,
-+};
-+
-+/**
-+ * struct dpni_tx_schedule_cfg - Structure representing Tx scheduling conf
-+ * @mode: Scheduling mode
-+ * @delta_bandwidth: Bandwidth represented in weights from 100 to 10000;
-+ * not applicable for 'strict-priority' mode;
-+ */
-+struct dpni_tx_schedule_cfg {
-+ enum dpni_tx_schedule_mode mode;
-+ u16 delta_bandwidth;
-+};
-+
-+/**
-+ * struct dpni_tx_priorities_cfg - Structure representing transmission
-+ * priorities for DPNI TCs
-+ * @tc_sched: An array of traffic-classes
-+ * @prio_group_A: Priority of group A
-+ * @prio_group_B: Priority of group B
-+ * @separate_groups: Treat A and B groups as separate
-+ * @ceetm_ch_idx: ceetm channel index to apply the changes
-+ */
-+struct dpni_tx_priorities_cfg {
-+ struct dpni_tx_schedule_cfg tc_sched[DPNI_MAX_TC];
-+ u8 prio_group_A;
-+ u8 prio_group_B;
-+ u8 separate_groups;
-+};
-+
-+int dpni_set_tx_priorities(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ const struct dpni_tx_priorities_cfg *cfg);
-+
-+/**
-+ * struct dpni_rx_tc_dist_cfg - Rx traffic class distribution configuration
-+ * @dist_size: Set the distribution size;
-+ * supported values: 1,2,3,4,6,7,8,12,14,16,24,28,32,48,56,64,96,
-+ * 112,128,192,224,256,384,448,512,768,896,1024
-+ * @dist_mode: Distribution mode
-+ * @key_cfg_iova: I/O virtual address of 256 bytes DMA-able memory filled with
-+ * the extractions to be used for the distribution key by calling
-+ * dpni_prepare_key_cfg() relevant only when
-+ * 'dist_mode != DPNI_DIST_MODE_NONE', otherwise it can be '0'
-+ * @fs_cfg: Flow Steering table configuration; only relevant if
-+ * 'dist_mode = DPNI_DIST_MODE_FS'
-+ */
-+struct dpni_rx_tc_dist_cfg {
-+ u16 dist_size;
-+ enum dpni_dist_mode dist_mode;
-+ u64 key_cfg_iova;
-+ struct dpni_fs_tbl_cfg fs_cfg;
-+};
-+
-+int dpni_set_rx_tc_dist(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 tc_id,
-+ const struct dpni_rx_tc_dist_cfg *cfg);
-+
-+/**
-+ * enum dpni_dest - DPNI destination types
-+ * @DPNI_DEST_NONE: Unassigned destination; The queue is set in parked mode and
-+ * does not generate FQDAN notifications; user is expected to
-+ * dequeue from the queue based on polling or other user-defined
-+ * method
-+ * @DPNI_DEST_DPIO: The queue is set in schedule mode and generates FQDAN
-+ * notifications to the specified DPIO; user is expected to dequeue
-+ * from the queue only after notification is received
-+ * @DPNI_DEST_DPCON: The queue is set in schedule mode and does not generate
-+ * FQDAN notifications, but is connected to the specified DPCON
-+ * object; user is expected to dequeue from the DPCON channel
-+ */
-+enum dpni_dest {
-+ DPNI_DEST_NONE = 0,
-+ DPNI_DEST_DPIO = 1,
-+ DPNI_DEST_DPCON = 2
-+};
-+
-+/**
-+ * struct dpni_queue - Queue structure
-+ * @user_context: User data, presented to the user along with any frames from
-+ * this queue. Not relevant for Tx queues.
-+ */
-+struct dpni_queue {
-+/**
-+ * struct destination - Destination structure
-+ * @id: ID of the destination, only relevant if DEST_TYPE is > 0.
-+ * Identifies either a DPIO or a DPCON object. Not relevant for
-+ * Tx queues.
-+ * @type: May be one of the following:
-+ * 0 - No destination, queue can be manually queried, but will not
-+ * push traffic or notifications to a DPIO;
-+ * 1 - The destination is a DPIO. When traffic becomes available in
-+ * the queue a FQDAN (FQ data available notification) will be
-+ * generated to selected DPIO;
-+ * 2 - The destination is a DPCON. The queue is associated with a
-+ * DPCON object for the purpose of scheduling between multiple
-+ * queues. The DPCON may be independently configured to
-+ * generate notifications. Not relevant for Tx queues.
-+ * @hold_active: Hold active, maintains a queue scheduled for longer
-+ * in a DPIO during dequeue to reduce spread of traffic.
-+ * Only relevant if queues are not affined to a single DPIO.
-+ */
-+ struct {
-+ u16 id;
-+ enum dpni_dest type;
-+ char hold_active;
-+ u8 priority;
-+ } destination;
-+ u64 user_context;
-+ struct {
-+ u64 value;
-+ char stash_control;
-+ } flc;
-+};
-+
-+/**
-+ * struct dpni_queue_id - Queue identification, used for enqueue commands
-+ * or queue control
-+ * @fqid: FQID used for enqueueing to and/or configuration of this specific FQ
-+ * @qdbin: Queueing bin, used to enqueue using QDID, DQBIN, QPRI. Only relevant
-+ * for Tx queues.
-+ */
-+struct dpni_queue_id {
-+ u32 fqid;
-+ u16 qdbin;
-+};
-+
-+/**
-+ * Set User Context
-+ */
-+#define DPNI_QUEUE_OPT_USER_CTX 0x00000001
-+#define DPNI_QUEUE_OPT_DEST 0x00000002
-+#define DPNI_QUEUE_OPT_FLC 0x00000004
-+#define DPNI_QUEUE_OPT_HOLD_ACTIVE 0x00000008
-+
-+int dpni_set_queue(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ enum dpni_queue_type qtype,
-+ u8 tc,
-+ u8 index,
-+ u8 options,
-+ const struct dpni_queue *queue);
-+
-+int dpni_get_queue(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ enum dpni_queue_type qtype,
-+ u8 tc,
-+ u8 index,
-+ struct dpni_queue *queue,
-+ struct dpni_queue_id *qid);
-+
-+/**
-+ * enum dpni_congestion_unit - DPNI congestion units
-+ * @DPNI_CONGESTION_UNIT_BYTES: bytes units
-+ * @DPNI_CONGESTION_UNIT_FRAMES: frames units
-+ */
-+enum dpni_congestion_unit {
-+ DPNI_CONGESTION_UNIT_BYTES = 0,
-+ DPNI_CONGESTION_UNIT_FRAMES
-+};
-+
-+/**
-+ * enum dpni_congestion_point - Structure representing congestion point
-+ * @DPNI_CP_QUEUE: Set taildrop per queue, identified by QUEUE_TYPE, TC and
-+ * QUEUE_INDEX
-+ * @DPNI_CP_GROUP: Set taildrop per queue group. Depending on options used to
-+ * define the DPNI this can be either per TC (default) or per
-+ * interface (DPNI_OPT_SHARED_CONGESTION set at DPNI create).
-+ * QUEUE_INDEX is ignored if this type is used.
-+ */
-+enum dpni_congestion_point {
-+ DPNI_CP_QUEUE,
-+ DPNI_CP_GROUP,
-+};
-+
-+/**
-+ * struct dpni_dest_cfg - Structure representing DPNI destination parameters
-+ * @dest_type: Destination type
-+ * @dest_id: Either DPIO ID or DPCON ID, depending on the destination type
-+ * @priority: Priority selection within the DPIO or DPCON channel; valid
-+ * values are 0-1 or 0-7, depending on the number of priorities
-+ * in that channel; not relevant for 'DPNI_DEST_NONE' option
-+ */
-+struct dpni_dest_cfg {
-+ enum dpni_dest dest_type;
-+ int dest_id;
-+ u8 priority;
-+};
-+
-+/* DPNI congestion options */
-+
-+/**
-+ * CSCN message is written to message_iova once entering a
-+ * congestion state (see 'threshold_entry')
-+ */
-+#define DPNI_CONG_OPT_WRITE_MEM_ON_ENTER 0x00000001
-+/**
-+ * CSCN message is written to message_iova once exiting a
-+ * congestion state (see 'threshold_exit')
-+ */
-+#define DPNI_CONG_OPT_WRITE_MEM_ON_EXIT 0x00000002
-+/**
-+ * CSCN write will attempt to allocate into a cache (coherent write);
-+ * valid only if 'DPNI_CONG_OPT_WRITE_MEM_<X>' is selected
-+ */
-+#define DPNI_CONG_OPT_COHERENT_WRITE 0x00000004
-+/**
-+ * if 'dest_cfg.dest_type != DPNI_DEST_NONE' CSCN message is sent to
-+ * DPIO/DPCON's WQ channel once entering a congestion state
-+ * (see 'threshold_entry')
-+ */
-+#define DPNI_CONG_OPT_NOTIFY_DEST_ON_ENTER 0x00000008
-+/**
-+ * if 'dest_cfg.dest_type != DPNI_DEST_NONE' CSCN message is sent to
-+ * DPIO/DPCON's WQ channel once exiting a congestion state
-+ * (see 'threshold_exit')
-+ */
-+#define DPNI_CONG_OPT_NOTIFY_DEST_ON_EXIT 0x00000010
-+/**
-+ * if 'dest_cfg.dest_type != DPNI_DEST_NONE' when the CSCN is written to the
-+ * sw-portal's DQRR, the DQRI interrupt is asserted immediately (if enabled)
-+ */
-+#define DPNI_CONG_OPT_INTR_COALESCING_DISABLED 0x00000020
-+/**
-+ * This congestion will trigger flow control or priority flow control.
-+ * This will have effect only if flow control is enabled with
-+ * dpni_set_link_cfg().
-+ */
-+#define DPNI_CONG_OPT_FLOW_CONTROL 0x00000040
-+
-+/**
-+ * struct dpni_congestion_notification_cfg - congestion notification
-+ * configuration
-+ * @units: Units type
-+ * @threshold_entry: Above this threshold we enter a congestion state.
-+ * set it to '0' to disable it
-+ * @threshold_exit: Below this threshold we exit the congestion state.
-+ * @message_ctx: The context that will be part of the CSCN message
-+ * @message_iova: I/O virtual address (must be in DMA-able memory),
-+ * must be 16B aligned; valid only if 'DPNI_CONG_OPT_WRITE_MEM_<X>'
-+ * is contained in 'options'
-+ * @dest_cfg: CSCN can be send to either DPIO or DPCON WQ channel
-+ * @notification_mode: Mask of available options; use 'DPNI_CONG_OPT_<X>' values
-+ */
-+
-+struct dpni_congestion_notification_cfg {
-+ enum dpni_congestion_unit units;
-+ u32 threshold_entry;
-+ u32 threshold_exit;
-+ u64 message_ctx;
-+ u64 message_iova;
-+ struct dpni_dest_cfg dest_cfg;
-+ u16 notification_mode;
-+};
-+
-+/** Compose TC parameter for function dpni_set_congestion_notification()
-+ * and dpni_get_congestion_notification().
-+ */
-+#define DPNI_BUILD_CH_TC(ceetm_ch_idx, tc) \
-+ ((((ceetm_ch_idx) & 0x0F) << 4) | ((tc) & 0x0F))
-+
-+int dpni_set_congestion_notification(
-+ struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ enum dpni_queue_type qtype,
-+ u8 tc_id,
-+ const struct dpni_congestion_notification_cfg *cfg);
-+
-+int dpni_get_congestion_notification(
-+ struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ enum dpni_queue_type qtype,
-+ u8 tc_id,
-+ struct dpni_congestion_notification_cfg *cfg);
-+
-+/**
-+ * struct dpni_taildrop - Structure representing the taildrop
-+ * @enable: Indicates whether the taildrop is active or not.
-+ * @units: Indicates the unit of THRESHOLD. Queue taildrop only supports
-+ * byte units, this field is ignored and assumed = 0 if
-+ * CONGESTION_POINT is 0.
-+ * @threshold: Threshold value, in units identified by UNITS field. Value 0
-+ * cannot be used as a valid taildrop threshold, THRESHOLD must
-+ * be > 0 if the taildrop is enabled.
-+ */
-+struct dpni_taildrop {
-+ char enable;
-+ enum dpni_congestion_unit units;
-+ u32 threshold;
-+};
-+
-+int dpni_set_taildrop(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ enum dpni_congestion_point cg_point,
-+ enum dpni_queue_type q_type,
-+ u8 tc,
-+ u8 q_index,
-+ struct dpni_taildrop *taildrop);
-+
-+int dpni_get_taildrop(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ enum dpni_congestion_point cg_point,
-+ enum dpni_queue_type q_type,
-+ u8 tc,
-+ u8 q_index,
-+ struct dpni_taildrop *taildrop);
-+
-+/**
-+ * struct dpni_rule_cfg - Rule configuration for table lookup
-+ * @key_iova: I/O virtual address of the key (must be in DMA-able memory)
-+ * @mask_iova: I/O virtual address of the mask (must be in DMA-able memory)
-+ * @key_size: key and mask size (in bytes)
-+ */
-+struct dpni_rule_cfg {
-+ u64 key_iova;
-+ u64 mask_iova;
-+ u8 key_size;
-+};
-+
-+int dpni_get_api_version(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 *major_ver,
-+ u16 *minor_ver);
-+
-+int dpni_add_qos_entry(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ const struct dpni_rule_cfg *cfg,
-+ u8 tc_id,
-+ u16 index);
-+
-+int dpni_remove_qos_entry(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ const struct dpni_rule_cfg *cfg);
-+
-+int dpni_clear_qos_table(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token);
-+
-+/**
-+ * Discard matching traffic. If set, this takes precedence over any other
-+ * configuration and matching traffic is always discarded.
-+ */
-+ #define DPNI_FS_OPT_DISCARD 0x1
-+
-+/**
-+ * Set FLC value. If set, flc member of struct dpni_fs_action_cfg is used to
-+ * override the FLC value set per queue.
-+ * For more details check the Frame Descriptor section in the hardware
-+ * documentation.
-+ */
-+#define DPNI_FS_OPT_SET_FLC 0x2
-+
-+/*
-+ * Indicates whether the 6 lowest significant bits of FLC are used for stash
-+ * control. If set, the 6 least significant bits in value are interpreted as
-+ * follows:
-+ * - bits 0-1: indicates the number of 64 byte units of context that are
-+ * stashed. FLC value is interpreted as a memory address in this case,
-+ * excluding the 6 LS bits.
-+ * - bits 2-3: indicates the number of 64 byte units of frame annotation
-+ * to be stashed. Annotation is placed at FD[ADDR].
-+ * - bits 4-5: indicates the number of 64 byte units of frame data to be
-+ * stashed. Frame data is placed at FD[ADDR] + FD[OFFSET].
-+ * This flag is ignored if DPNI_FS_OPT_SET_FLC is not specified.
-+ */
-+#define DPNI_FS_OPT_SET_STASH_CONTROL 0x4
-+
-+/**
-+ * struct dpni_fs_action_cfg - Action configuration for table look-up
-+ * @flc: FLC value for traffic matching this rule. Please check the
-+ * Frame Descriptor section in the hardware documentation for
-+ * more information.
-+ * @flow_id: Identifies the Rx queue used for matching traffic. Supported
-+ * values are in range 0 to num_queue-1.
-+ * @options: Any combination of DPNI_FS_OPT_ values.
-+ */
-+struct dpni_fs_action_cfg {
-+ u64 flc;
-+ u16 flow_id;
-+ u16 options;
-+};
-+
-+int dpni_add_fs_entry(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 tc_id,
-+ u16 index,
-+ const struct dpni_rule_cfg *cfg,
-+ const struct dpni_fs_action_cfg *action);
-+
-+int dpni_remove_fs_entry(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 tc_id,
-+ const struct dpni_rule_cfg *cfg);
-+
-+/**
-+ * When used for queue_idx in function dpni_set_rx_dist_default_queue
-+ * will signal to dpni to drop all unclassified frames
-+ */
-+#define DPNI_FS_MISS_DROP ((uint16_t)-1)
-+
-+/**
-+ * struct dpni_rx_dist_cfg - distribution configuration
-+ * @dist_size: distribution size; supported values: 1,2,3,4,6,7,8,
-+ * 12,14,16,24,28,32,48,56,64,96,112,128,192,224,256,384,448,
-+ * 512,768,896,1024
-+ * @key_cfg_iova: I/O virtual address of 256 bytes DMA-able memory filled with
-+ * the extractions to be used for the distribution key by calling
-+ * dpkg_prepare_key_cfg() relevant only when enable!=0 otherwise
-+ * it can be '0'
-+ * @enable: enable/disable the distribution.
-+ * @tc: TC id for which distribution is set
-+ * @fs_miss_flow_id: when packet misses all rules from flow steering table and
-+ * hash is disabled it will be put into this queue id; use
-+ * DPNI_FS_MISS_DROP to drop frames. The value of this field is
-+ * used only when flow steering distribution is enabled and hash
-+ * distribution is disabled
-+ */
-+struct dpni_rx_dist_cfg {
-+ u16 dist_size;
-+ u64 key_cfg_iova;
-+ u8 enable;
-+ u8 tc;
-+ u16 fs_miss_flow_id;
-+};
-+
-+int dpni_set_rx_fs_dist(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ const struct dpni_rx_dist_cfg *cfg);
-+
-+int dpni_set_rx_hash_dist(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ const struct dpni_rx_dist_cfg *cfg);
-+
-+#endif /* __FSL_DPNI_H */
---- /dev/null
-+++ b/drivers/staging/fsl-dpaa2/ethernet/net.h
-@@ -0,0 +1,480 @@
-+/* Copyright 2013-2015 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of the above-listed copyright holders nor the
-+ * names of any contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
-+ * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
-+ * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
-+ * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE
-+ * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
-+ * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
-+ * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
-+ * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
-+ * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
-+ * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
-+ * POSSIBILITY OF SUCH DAMAGE.
-+ */
-+#ifndef __FSL_NET_H
-+#define __FSL_NET_H
-+
-+#define LAST_HDR_INDEX 0xFFFFFFFF
-+
-+/*****************************************************************************/
-+/* Protocol fields */
-+/*****************************************************************************/
-+
-+/************************* Ethernet fields *********************************/
-+#define NH_FLD_ETH_DA (1)
-+#define NH_FLD_ETH_SA (NH_FLD_ETH_DA << 1)
-+#define NH_FLD_ETH_LENGTH (NH_FLD_ETH_DA << 2)
-+#define NH_FLD_ETH_TYPE (NH_FLD_ETH_DA << 3)
-+#define NH_FLD_ETH_FINAL_CKSUM (NH_FLD_ETH_DA << 4)
-+#define NH_FLD_ETH_PADDING (NH_FLD_ETH_DA << 5)
-+#define NH_FLD_ETH_ALL_FIELDS ((NH_FLD_ETH_DA << 6) - 1)
-+
-+#define NH_FLD_ETH_ADDR_SIZE 6
-+
-+/*************************** VLAN fields ***********************************/
-+#define NH_FLD_VLAN_VPRI (1)
-+#define NH_FLD_VLAN_CFI (NH_FLD_VLAN_VPRI << 1)
-+#define NH_FLD_VLAN_VID (NH_FLD_VLAN_VPRI << 2)
-+#define NH_FLD_VLAN_LENGTH (NH_FLD_VLAN_VPRI << 3)
-+#define NH_FLD_VLAN_TYPE (NH_FLD_VLAN_VPRI << 4)
-+#define NH_FLD_VLAN_ALL_FIELDS ((NH_FLD_VLAN_VPRI << 5) - 1)
-+
-+#define NH_FLD_VLAN_TCI (NH_FLD_VLAN_VPRI | \
-+ NH_FLD_VLAN_CFI | \
-+ NH_FLD_VLAN_VID)
-+
-+/************************ IP (generic) fields ******************************/
-+#define NH_FLD_IP_VER (1)
-+#define NH_FLD_IP_DSCP (NH_FLD_IP_VER << 2)
-+#define NH_FLD_IP_ECN (NH_FLD_IP_VER << 3)
-+#define NH_FLD_IP_PROTO (NH_FLD_IP_VER << 4)
-+#define NH_FLD_IP_SRC (NH_FLD_IP_VER << 5)
-+#define NH_FLD_IP_DST (NH_FLD_IP_VER << 6)
-+#define NH_FLD_IP_TOS_TC (NH_FLD_IP_VER << 7)
-+#define NH_FLD_IP_ID (NH_FLD_IP_VER << 8)
-+#define NH_FLD_IP_ALL_FIELDS ((NH_FLD_IP_VER << 9) - 1)
-+
-+#define NH_FLD_IP_PROTO_SIZE 1
-+
-+/***************************** IPV4 fields *********************************/
-+#define NH_FLD_IPV4_VER (1)
-+#define NH_FLD_IPV4_HDR_LEN (NH_FLD_IPV4_VER << 1)
-+#define NH_FLD_IPV4_TOS (NH_FLD_IPV4_VER << 2)
-+#define NH_FLD_IPV4_TOTAL_LEN (NH_FLD_IPV4_VER << 3)
-+#define NH_FLD_IPV4_ID (NH_FLD_IPV4_VER << 4)
-+#define NH_FLD_IPV4_FLAG_D (NH_FLD_IPV4_VER << 5)
-+#define NH_FLD_IPV4_FLAG_M (NH_FLD_IPV4_VER << 6)
-+#define NH_FLD_IPV4_OFFSET (NH_FLD_IPV4_VER << 7)
-+#define NH_FLD_IPV4_TTL (NH_FLD_IPV4_VER << 8)
-+#define NH_FLD_IPV4_PROTO (NH_FLD_IPV4_VER << 9)
-+#define NH_FLD_IPV4_CKSUM (NH_FLD_IPV4_VER << 10)
-+#define NH_FLD_IPV4_SRC_IP (NH_FLD_IPV4_VER << 11)
-+#define NH_FLD_IPV4_DST_IP (NH_FLD_IPV4_VER << 12)
-+#define NH_FLD_IPV4_OPTS (NH_FLD_IPV4_VER << 13)
-+#define NH_FLD_IPV4_OPTS_COUNT (NH_FLD_IPV4_VER << 14)
-+#define NH_FLD_IPV4_ALL_FIELDS ((NH_FLD_IPV4_VER << 15) - 1)
-+
-+#define NH_FLD_IPV4_ADDR_SIZE 4
-+#define NH_FLD_IPV4_PROTO_SIZE 1
-+
-+/***************************** IPV6 fields *********************************/
-+#define NH_FLD_IPV6_VER (1)
-+#define NH_FLD_IPV6_TC (NH_FLD_IPV6_VER << 1)
-+#define NH_FLD_IPV6_SRC_IP (NH_FLD_IPV6_VER << 2)
-+#define NH_FLD_IPV6_DST_IP (NH_FLD_IPV6_VER << 3)
-+#define NH_FLD_IPV6_NEXT_HDR (NH_FLD_IPV6_VER << 4)
-+#define NH_FLD_IPV6_FL (NH_FLD_IPV6_VER << 5)
-+#define NH_FLD_IPV6_HOP_LIMIT (NH_FLD_IPV6_VER << 6)
-+#define NH_FLD_IPV6_ID (NH_FLD_IPV6_VER << 7)
-+#define NH_FLD_IPV6_ALL_FIELDS ((NH_FLD_IPV6_VER << 8) - 1)
-+
-+#define NH_FLD_IPV6_ADDR_SIZE 16
-+#define NH_FLD_IPV6_NEXT_HDR_SIZE 1
-+
-+/***************************** ICMP fields *********************************/
-+#define NH_FLD_ICMP_TYPE (1)
-+#define NH_FLD_ICMP_CODE (NH_FLD_ICMP_TYPE << 1)
-+#define NH_FLD_ICMP_CKSUM (NH_FLD_ICMP_TYPE << 2)
-+#define NH_FLD_ICMP_ID (NH_FLD_ICMP_TYPE << 3)
-+#define NH_FLD_ICMP_SQ_NUM (NH_FLD_ICMP_TYPE << 4)
-+#define NH_FLD_ICMP_ALL_FIELDS ((NH_FLD_ICMP_TYPE << 5) - 1)
-+
-+#define NH_FLD_ICMP_CODE_SIZE 1
-+#define NH_FLD_ICMP_TYPE_SIZE 1
-+
-+/***************************** IGMP fields *********************************/
-+#define NH_FLD_IGMP_VERSION (1)
-+#define NH_FLD_IGMP_TYPE (NH_FLD_IGMP_VERSION << 1)
-+#define NH_FLD_IGMP_CKSUM (NH_FLD_IGMP_VERSION << 2)
-+#define NH_FLD_IGMP_DATA (NH_FLD_IGMP_VERSION << 3)
-+#define NH_FLD_IGMP_ALL_FIELDS ((NH_FLD_IGMP_VERSION << 4) - 1)
-+
-+/***************************** TCP fields **********************************/
-+#define NH_FLD_TCP_PORT_SRC (1)
-+#define NH_FLD_TCP_PORT_DST (NH_FLD_TCP_PORT_SRC << 1)
-+#define NH_FLD_TCP_SEQ (NH_FLD_TCP_PORT_SRC << 2)
-+#define NH_FLD_TCP_ACK (NH_FLD_TCP_PORT_SRC << 3)
-+#define NH_FLD_TCP_OFFSET (NH_FLD_TCP_PORT_SRC << 4)
-+#define NH_FLD_TCP_FLAGS (NH_FLD_TCP_PORT_SRC << 5)
-+#define NH_FLD_TCP_WINDOW (NH_FLD_TCP_PORT_SRC << 6)
-+#define NH_FLD_TCP_CKSUM (NH_FLD_TCP_PORT_SRC << 7)
-+#define NH_FLD_TCP_URGPTR (NH_FLD_TCP_PORT_SRC << 8)
-+#define NH_FLD_TCP_OPTS (NH_FLD_TCP_PORT_SRC << 9)
-+#define NH_FLD_TCP_OPTS_COUNT (NH_FLD_TCP_PORT_SRC << 10)
-+#define NH_FLD_TCP_ALL_FIELDS ((NH_FLD_TCP_PORT_SRC << 11) - 1)
-+
-+#define NH_FLD_TCP_PORT_SIZE 2
-+
-+/***************************** UDP fields **********************************/
-+#define NH_FLD_UDP_PORT_SRC (1)
-+#define NH_FLD_UDP_PORT_DST (NH_FLD_UDP_PORT_SRC << 1)
-+#define NH_FLD_UDP_LEN (NH_FLD_UDP_PORT_SRC << 2)
-+#define NH_FLD_UDP_CKSUM (NH_FLD_UDP_PORT_SRC << 3)
-+#define NH_FLD_UDP_ALL_FIELDS ((NH_FLD_UDP_PORT_SRC << 4) - 1)
-+
-+#define NH_FLD_UDP_PORT_SIZE 2
-+
-+/*************************** UDP-lite fields *******************************/
-+#define NH_FLD_UDP_LITE_PORT_SRC (1)
-+#define NH_FLD_UDP_LITE_PORT_DST (NH_FLD_UDP_LITE_PORT_SRC << 1)
-+#define NH_FLD_UDP_LITE_ALL_FIELDS \
-+ ((NH_FLD_UDP_LITE_PORT_SRC << 2) - 1)
-+
-+#define NH_FLD_UDP_LITE_PORT_SIZE 2
-+
-+/*************************** UDP-encap-ESP fields **************************/
-+#define NH_FLD_UDP_ENC_ESP_PORT_SRC (1)
-+#define NH_FLD_UDP_ENC_ESP_PORT_DST (NH_FLD_UDP_ENC_ESP_PORT_SRC << 1)
-+#define NH_FLD_UDP_ENC_ESP_LEN (NH_FLD_UDP_ENC_ESP_PORT_SRC << 2)
-+#define NH_FLD_UDP_ENC_ESP_CKSUM (NH_FLD_UDP_ENC_ESP_PORT_SRC << 3)
-+#define NH_FLD_UDP_ENC_ESP_SPI (NH_FLD_UDP_ENC_ESP_PORT_SRC << 4)
-+#define NH_FLD_UDP_ENC_ESP_SEQUENCE_NUM (NH_FLD_UDP_ENC_ESP_PORT_SRC << 5)
-+#define NH_FLD_UDP_ENC_ESP_ALL_FIELDS \
-+ ((NH_FLD_UDP_ENC_ESP_PORT_SRC << 6) - 1)
-+
-+#define NH_FLD_UDP_ENC_ESP_PORT_SIZE 2
-+#define NH_FLD_UDP_ENC_ESP_SPI_SIZE 4
-+
-+/***************************** SCTP fields *********************************/
-+#define NH_FLD_SCTP_PORT_SRC (1)
-+#define NH_FLD_SCTP_PORT_DST (NH_FLD_SCTP_PORT_SRC << 1)
-+#define NH_FLD_SCTP_VER_TAG (NH_FLD_SCTP_PORT_SRC << 2)
-+#define NH_FLD_SCTP_CKSUM (NH_FLD_SCTP_PORT_SRC << 3)
-+#define NH_FLD_SCTP_ALL_FIELDS ((NH_FLD_SCTP_PORT_SRC << 4) - 1)
-+
-+#define NH_FLD_SCTP_PORT_SIZE 2
-+
-+/***************************** DCCP fields *********************************/
-+#define NH_FLD_DCCP_PORT_SRC (1)
-+#define NH_FLD_DCCP_PORT_DST (NH_FLD_DCCP_PORT_SRC << 1)
-+#define NH_FLD_DCCP_ALL_FIELDS ((NH_FLD_DCCP_PORT_SRC << 2) - 1)
-+
-+#define NH_FLD_DCCP_PORT_SIZE 2
-+
-+/***************************** IPHC fields *********************************/
-+#define NH_FLD_IPHC_CID (1)
-+#define NH_FLD_IPHC_CID_TYPE (NH_FLD_IPHC_CID << 1)
-+#define NH_FLD_IPHC_HCINDEX (NH_FLD_IPHC_CID << 2)
-+#define NH_FLD_IPHC_GEN (NH_FLD_IPHC_CID << 3)
-+#define NH_FLD_IPHC_D_BIT (NH_FLD_IPHC_CID << 4)
-+#define NH_FLD_IPHC_ALL_FIELDS ((NH_FLD_IPHC_CID << 5) - 1)
-+
-+/***************************** SCTP fields *********************************/
-+#define NH_FLD_SCTP_CHUNK_DATA_TYPE (1)
-+#define NH_FLD_SCTP_CHUNK_DATA_FLAGS (NH_FLD_SCTP_CHUNK_DATA_TYPE << 1)
-+#define NH_FLD_SCTP_CHUNK_DATA_LENGTH (NH_FLD_SCTP_CHUNK_DATA_TYPE << 2)
-+#define NH_FLD_SCTP_CHUNK_DATA_TSN (NH_FLD_SCTP_CHUNK_DATA_TYPE << 3)
-+#define NH_FLD_SCTP_CHUNK_DATA_STREAM_ID (NH_FLD_SCTP_CHUNK_DATA_TYPE << 4)
-+#define NH_FLD_SCTP_CHUNK_DATA_STREAM_SQN (NH_FLD_SCTP_CHUNK_DATA_TYPE << 5)
-+#define NH_FLD_SCTP_CHUNK_DATA_PAYLOAD_PID (NH_FLD_SCTP_CHUNK_DATA_TYPE << 6)
-+#define NH_FLD_SCTP_CHUNK_DATA_UNORDERED (NH_FLD_SCTP_CHUNK_DATA_TYPE << 7)
-+#define NH_FLD_SCTP_CHUNK_DATA_BEGGINING (NH_FLD_SCTP_CHUNK_DATA_TYPE << 8)
-+#define NH_FLD_SCTP_CHUNK_DATA_END (NH_FLD_SCTP_CHUNK_DATA_TYPE << 9)
-+#define NH_FLD_SCTP_CHUNK_DATA_ALL_FIELDS \
-+ ((NH_FLD_SCTP_CHUNK_DATA_TYPE << 10) - 1)
-+
-+/*************************** L2TPV2 fields *********************************/
-+#define NH_FLD_L2TPV2_TYPE_BIT (1)
-+#define NH_FLD_L2TPV2_LENGTH_BIT (NH_FLD_L2TPV2_TYPE_BIT << 1)
-+#define NH_FLD_L2TPV2_SEQUENCE_BIT (NH_FLD_L2TPV2_TYPE_BIT << 2)
-+#define NH_FLD_L2TPV2_OFFSET_BIT (NH_FLD_L2TPV2_TYPE_BIT << 3)
-+#define NH_FLD_L2TPV2_PRIORITY_BIT (NH_FLD_L2TPV2_TYPE_BIT << 4)
-+#define NH_FLD_L2TPV2_VERSION (NH_FLD_L2TPV2_TYPE_BIT << 5)
-+#define NH_FLD_L2TPV2_LEN (NH_FLD_L2TPV2_TYPE_BIT << 6)
-+#define NH_FLD_L2TPV2_TUNNEL_ID (NH_FLD_L2TPV2_TYPE_BIT << 7)
-+#define NH_FLD_L2TPV2_SESSION_ID (NH_FLD_L2TPV2_TYPE_BIT << 8)
-+#define NH_FLD_L2TPV2_NS (NH_FLD_L2TPV2_TYPE_BIT << 9)
-+#define NH_FLD_L2TPV2_NR (NH_FLD_L2TPV2_TYPE_BIT << 10)
-+#define NH_FLD_L2TPV2_OFFSET_SIZE (NH_FLD_L2TPV2_TYPE_BIT << 11)
-+#define NH_FLD_L2TPV2_FIRST_BYTE (NH_FLD_L2TPV2_TYPE_BIT << 12)
-+#define NH_FLD_L2TPV2_ALL_FIELDS \
-+ ((NH_FLD_L2TPV2_TYPE_BIT << 13) - 1)
-+
-+/*************************** L2TPV3 fields *********************************/
-+#define NH_FLD_L2TPV3_CTRL_TYPE_BIT (1)
-+#define NH_FLD_L2TPV3_CTRL_LENGTH_BIT (NH_FLD_L2TPV3_CTRL_TYPE_BIT << 1)
-+#define NH_FLD_L2TPV3_CTRL_SEQUENCE_BIT (NH_FLD_L2TPV3_CTRL_TYPE_BIT << 2)
-+#define NH_FLD_L2TPV3_CTRL_VERSION (NH_FLD_L2TPV3_CTRL_TYPE_BIT << 3)
-+#define NH_FLD_L2TPV3_CTRL_LENGTH (NH_FLD_L2TPV3_CTRL_TYPE_BIT << 4)
-+#define NH_FLD_L2TPV3_CTRL_CONTROL (NH_FLD_L2TPV3_CTRL_TYPE_BIT << 5)
-+#define NH_FLD_L2TPV3_CTRL_SENT (NH_FLD_L2TPV3_CTRL_TYPE_BIT << 6)
-+#define NH_FLD_L2TPV3_CTRL_RECV (NH_FLD_L2TPV3_CTRL_TYPE_BIT << 7)
-+#define NH_FLD_L2TPV3_CTRL_FIRST_BYTE (NH_FLD_L2TPV3_CTRL_TYPE_BIT << 8)
-+#define NH_FLD_L2TPV3_CTRL_ALL_FIELDS \
-+ ((NH_FLD_L2TPV3_CTRL_TYPE_BIT << 9) - 1)
-+
-+#define NH_FLD_L2TPV3_SESS_TYPE_BIT (1)
-+#define NH_FLD_L2TPV3_SESS_VERSION (NH_FLD_L2TPV3_SESS_TYPE_BIT << 1)
-+#define NH_FLD_L2TPV3_SESS_ID (NH_FLD_L2TPV3_SESS_TYPE_BIT << 2)
-+#define NH_FLD_L2TPV3_SESS_COOKIE (NH_FLD_L2TPV3_SESS_TYPE_BIT << 3)
-+#define NH_FLD_L2TPV3_SESS_ALL_FIELDS \
-+ ((NH_FLD_L2TPV3_SESS_TYPE_BIT << 4) - 1)
-+
-+/**************************** PPP fields ***********************************/
-+#define NH_FLD_PPP_PID (1)
-+#define NH_FLD_PPP_COMPRESSED (NH_FLD_PPP_PID << 1)
-+#define NH_FLD_PPP_ALL_FIELDS ((NH_FLD_PPP_PID << 2) - 1)
-+
-+/************************** PPPoE fields ***********************************/
-+#define NH_FLD_PPPOE_VER (1)
-+#define NH_FLD_PPPOE_TYPE (NH_FLD_PPPOE_VER << 1)
-+#define NH_FLD_PPPOE_CODE (NH_FLD_PPPOE_VER << 2)
-+#define NH_FLD_PPPOE_SID (NH_FLD_PPPOE_VER << 3)
-+#define NH_FLD_PPPOE_LEN (NH_FLD_PPPOE_VER << 4)
-+#define NH_FLD_PPPOE_SESSION (NH_FLD_PPPOE_VER << 5)
-+#define NH_FLD_PPPOE_PID (NH_FLD_PPPOE_VER << 6)
-+#define NH_FLD_PPPOE_ALL_FIELDS ((NH_FLD_PPPOE_VER << 7) - 1)
-+
-+/************************* PPP-Mux fields **********************************/
-+#define NH_FLD_PPPMUX_PID (1)
-+#define NH_FLD_PPPMUX_CKSUM (NH_FLD_PPPMUX_PID << 1)
-+#define NH_FLD_PPPMUX_COMPRESSED (NH_FLD_PPPMUX_PID << 2)
-+#define NH_FLD_PPPMUX_ALL_FIELDS ((NH_FLD_PPPMUX_PID << 3) - 1)
-+
-+/*********************** PPP-Mux sub-frame fields **************************/
-+#define NH_FLD_PPPMUX_SUBFRM_PFF (1)
-+#define NH_FLD_PPPMUX_SUBFRM_LXT (NH_FLD_PPPMUX_SUBFRM_PFF << 1)
-+#define NH_FLD_PPPMUX_SUBFRM_LEN (NH_FLD_PPPMUX_SUBFRM_PFF << 2)
-+#define NH_FLD_PPPMUX_SUBFRM_PID (NH_FLD_PPPMUX_SUBFRM_PFF << 3)
-+#define NH_FLD_PPPMUX_SUBFRM_USE_PID (NH_FLD_PPPMUX_SUBFRM_PFF << 4)
-+#define NH_FLD_PPPMUX_SUBFRM_ALL_FIELDS \
-+ ((NH_FLD_PPPMUX_SUBFRM_PFF << 5) - 1)
-+
-+/*************************** LLC fields ************************************/
-+#define NH_FLD_LLC_DSAP (1)
-+#define NH_FLD_LLC_SSAP (NH_FLD_LLC_DSAP << 1)
-+#define NH_FLD_LLC_CTRL (NH_FLD_LLC_DSAP << 2)
-+#define NH_FLD_LLC_ALL_FIELDS ((NH_FLD_LLC_DSAP << 3) - 1)
-+
-+/*************************** NLPID fields **********************************/
-+#define NH_FLD_NLPID_NLPID (1)
-+#define NH_FLD_NLPID_ALL_FIELDS ((NH_FLD_NLPID_NLPID << 1) - 1)
-+
-+/*************************** SNAP fields ***********************************/
-+#define NH_FLD_SNAP_OUI (1)
-+#define NH_FLD_SNAP_PID (NH_FLD_SNAP_OUI << 1)
-+#define NH_FLD_SNAP_ALL_FIELDS ((NH_FLD_SNAP_OUI << 2) - 1)
-+
-+/*************************** LLC SNAP fields *******************************/
-+#define NH_FLD_LLC_SNAP_TYPE (1)
-+#define NH_FLD_LLC_SNAP_ALL_FIELDS ((NH_FLD_LLC_SNAP_TYPE << 1) - 1)
-+
-+#define NH_FLD_ARP_HTYPE (1)
-+#define NH_FLD_ARP_PTYPE (NH_FLD_ARP_HTYPE << 1)
-+#define NH_FLD_ARP_HLEN (NH_FLD_ARP_HTYPE << 2)
-+#define NH_FLD_ARP_PLEN (NH_FLD_ARP_HTYPE << 3)
-+#define NH_FLD_ARP_OPER (NH_FLD_ARP_HTYPE << 4)
-+#define NH_FLD_ARP_SHA (NH_FLD_ARP_HTYPE << 5)
-+#define NH_FLD_ARP_SPA (NH_FLD_ARP_HTYPE << 6)
-+#define NH_FLD_ARP_THA (NH_FLD_ARP_HTYPE << 7)
-+#define NH_FLD_ARP_TPA (NH_FLD_ARP_HTYPE << 8)
-+#define NH_FLD_ARP_ALL_FIELDS ((NH_FLD_ARP_HTYPE << 9) - 1)
-+
-+/*************************** RFC2684 fields ********************************/
-+#define NH_FLD_RFC2684_LLC (1)
-+#define NH_FLD_RFC2684_NLPID (NH_FLD_RFC2684_LLC << 1)
-+#define NH_FLD_RFC2684_OUI (NH_FLD_RFC2684_LLC << 2)
-+#define NH_FLD_RFC2684_PID (NH_FLD_RFC2684_LLC << 3)
-+#define NH_FLD_RFC2684_VPN_OUI (NH_FLD_RFC2684_LLC << 4)
-+#define NH_FLD_RFC2684_VPN_IDX (NH_FLD_RFC2684_LLC << 5)
-+#define NH_FLD_RFC2684_ALL_FIELDS ((NH_FLD_RFC2684_LLC << 6) - 1)
-+
-+/*************************** User defined fields ***************************/
-+#define NH_FLD_USER_DEFINED_SRCPORT (1)
-+#define NH_FLD_USER_DEFINED_PCDID (NH_FLD_USER_DEFINED_SRCPORT << 1)
-+#define NH_FLD_USER_DEFINED_ALL_FIELDS \
-+ ((NH_FLD_USER_DEFINED_SRCPORT << 2) - 1)
-+
-+/*************************** Payload fields ********************************/
-+#define NH_FLD_PAYLOAD_BUFFER (1)
-+#define NH_FLD_PAYLOAD_SIZE (NH_FLD_PAYLOAD_BUFFER << 1)
-+#define NH_FLD_MAX_FRM_SIZE (NH_FLD_PAYLOAD_BUFFER << 2)
-+#define NH_FLD_MIN_FRM_SIZE (NH_FLD_PAYLOAD_BUFFER << 3)
-+#define NH_FLD_PAYLOAD_TYPE (NH_FLD_PAYLOAD_BUFFER << 4)
-+#define NH_FLD_FRAME_SIZE (NH_FLD_PAYLOAD_BUFFER << 5)
-+#define NH_FLD_PAYLOAD_ALL_FIELDS ((NH_FLD_PAYLOAD_BUFFER << 6) - 1)
-+
-+/*************************** GRE fields ************************************/
-+#define NH_FLD_GRE_TYPE (1)
-+#define NH_FLD_GRE_ALL_FIELDS ((NH_FLD_GRE_TYPE << 1) - 1)
-+
-+/*************************** MINENCAP fields *******************************/
-+#define NH_FLD_MINENCAP_SRC_IP (1)
-+#define NH_FLD_MINENCAP_DST_IP (NH_FLD_MINENCAP_SRC_IP << 1)
-+#define NH_FLD_MINENCAP_TYPE (NH_FLD_MINENCAP_SRC_IP << 2)
-+#define NH_FLD_MINENCAP_ALL_FIELDS \
-+ ((NH_FLD_MINENCAP_SRC_IP << 3) - 1)
-+
-+/*************************** IPSEC AH fields *******************************/
-+#define NH_FLD_IPSEC_AH_SPI (1)
-+#define NH_FLD_IPSEC_AH_NH (NH_FLD_IPSEC_AH_SPI << 1)
-+#define NH_FLD_IPSEC_AH_ALL_FIELDS ((NH_FLD_IPSEC_AH_SPI << 2) - 1)
-+
-+/*************************** IPSEC ESP fields ******************************/
-+#define NH_FLD_IPSEC_ESP_SPI (1)
-+#define NH_FLD_IPSEC_ESP_SEQUENCE_NUM (NH_FLD_IPSEC_ESP_SPI << 1)
-+#define NH_FLD_IPSEC_ESP_ALL_FIELDS ((NH_FLD_IPSEC_ESP_SPI << 2) - 1)
-+
-+#define NH_FLD_IPSEC_ESP_SPI_SIZE 4
-+
-+/*************************** MPLS fields ***********************************/
-+#define NH_FLD_MPLS_LABEL_STACK (1)
-+#define NH_FLD_MPLS_LABEL_STACK_ALL_FIELDS \
-+ ((NH_FLD_MPLS_LABEL_STACK << 1) - 1)
-+
-+/*************************** MACSEC fields *********************************/
-+#define NH_FLD_MACSEC_SECTAG (1)
-+#define NH_FLD_MACSEC_ALL_FIELDS ((NH_FLD_MACSEC_SECTAG << 1) - 1)
-+
-+/*************************** GTP fields ************************************/
-+#define NH_FLD_GTP_TEID (1)
-+
-+/* Protocol options */
-+
-+/* Ethernet options */
-+#define NH_OPT_ETH_BROADCAST 1
-+#define NH_OPT_ETH_MULTICAST 2
-+#define NH_OPT_ETH_UNICAST 3
-+#define NH_OPT_ETH_BPDU 4
-+
-+#define NH_ETH_IS_MULTICAST_ADDR(addr) (addr[0] & 0x01)
-+/* also applicable for broadcast */
-+
-+/* VLAN options */
-+#define NH_OPT_VLAN_CFI 1
-+
-+/* IPV4 options */
-+#define NH_OPT_IPV4_UNICAST 1
-+#define NH_OPT_IPV4_MULTICAST 2
-+#define NH_OPT_IPV4_BROADCAST 3
-+#define NH_OPT_IPV4_OPTION 4
-+#define NH_OPT_IPV4_FRAG 5
-+#define NH_OPT_IPV4_INITIAL_FRAG 6
-+
-+/* IPV6 options */
-+#define NH_OPT_IPV6_UNICAST 1
-+#define NH_OPT_IPV6_MULTICAST 2
-+#define NH_OPT_IPV6_OPTION 3
-+#define NH_OPT_IPV6_FRAG 4
-+#define NH_OPT_IPV6_INITIAL_FRAG 5
-+
-+/* General IP options (may be used for any version) */
-+#define NH_OPT_IP_FRAG 1
-+#define NH_OPT_IP_INITIAL_FRAG 2
-+#define NH_OPT_IP_OPTION 3
-+
-+/* Minenc. options */
-+#define NH_OPT_MINENCAP_SRC_ADDR_PRESENT 1
-+
-+/* GRE. options */
-+#define NH_OPT_GRE_ROUTING_PRESENT 1
-+
-+/* TCP options */
-+#define NH_OPT_TCP_OPTIONS 1
-+#define NH_OPT_TCP_CONTROL_HIGH_BITS 2
-+#define NH_OPT_TCP_CONTROL_LOW_BITS 3
-+
-+/* CAPWAP options */
-+#define NH_OPT_CAPWAP_DTLS 1
-+
-+enum net_prot {
-+ NET_PROT_NONE = 0,
-+ NET_PROT_PAYLOAD,
-+ NET_PROT_ETH,
-+ NET_PROT_VLAN,
-+ NET_PROT_IPV4,
-+ NET_PROT_IPV6,
-+ NET_PROT_IP,
-+ NET_PROT_TCP,
-+ NET_PROT_UDP,
-+ NET_PROT_UDP_LITE,
-+ NET_PROT_IPHC,
-+ NET_PROT_SCTP,
-+ NET_PROT_SCTP_CHUNK_DATA,
-+ NET_PROT_PPPOE,
-+ NET_PROT_PPP,
-+ NET_PROT_PPPMUX,
-+ NET_PROT_PPPMUX_SUBFRM,
-+ NET_PROT_L2TPV2,
-+ NET_PROT_L2TPV3_CTRL,
-+ NET_PROT_L2TPV3_SESS,
-+ NET_PROT_LLC,
-+ NET_PROT_LLC_SNAP,
-+ NET_PROT_NLPID,
-+ NET_PROT_SNAP,
-+ NET_PROT_MPLS,
-+ NET_PROT_IPSEC_AH,
-+ NET_PROT_IPSEC_ESP,
-+ NET_PROT_UDP_ENC_ESP, /* RFC 3948 */
-+ NET_PROT_MACSEC,
-+ NET_PROT_GRE,
-+ NET_PROT_MINENCAP,
-+ NET_PROT_DCCP,
-+ NET_PROT_ICMP,
-+ NET_PROT_IGMP,
-+ NET_PROT_ARP,
-+ NET_PROT_CAPWAP_DATA,
-+ NET_PROT_CAPWAP_CTRL,
-+ NET_PROT_RFC2684,
-+ NET_PROT_ICMPV6,
-+ NET_PROT_FCOE,
-+ NET_PROT_FIP,
-+ NET_PROT_ISCSI,
-+ NET_PROT_GTP,
-+ NET_PROT_USER_DEFINED_L2,
-+ NET_PROT_USER_DEFINED_L3,
-+ NET_PROT_USER_DEFINED_L4,
-+ NET_PROT_USER_DEFINED_L5,
-+ NET_PROT_USER_DEFINED_SHIM1,
-+ NET_PROT_USER_DEFINED_SHIM2,
-+
-+ NET_PROT_DUMMY_LAST
-+};
-+
-+/*! IEEE8021.Q */
-+#define NH_IEEE8021Q_ETYPE 0x8100
-+#define NH_IEEE8021Q_HDR(etype, pcp, dei, vlan_id) \
-+ ((((u32)((etype) & 0xFFFF)) << 16) | \
-+ (((u32)((pcp) & 0x07)) << 13) | \
-+ (((u32)((dei) & 0x01)) << 12) | \
-+ (((u32)((vlan_id) & 0xFFF))))
-+
-+#endif /* __FSL_NET_H */
---- /dev/null
-+++ b/drivers/staging/fsl-dpaa2/ethsw/Makefile
-@@ -0,0 +1,10 @@
-+# SPDX-License-Identifier: GPL-2.0
-+#
-+# Makefile for the Freescale DPAA2 Ethernet Switch
-+#
-+# Copyright 2014-2017 Freescale Semiconductor, Inc.
-+# Copyright 2017-2018 NXP
-+
-+obj-$(CONFIG_FSL_DPAA2_ETHSW) += dpaa2-ethsw.o
-+
-+dpaa2-ethsw-objs := ethsw.o ethsw-ethtool.o dpsw.o
---- /dev/null
-+++ b/drivers/staging/fsl-dpaa2/ethsw/README
-@@ -0,0 +1,106 @@
-+DPAA2 Ethernet Switch driver
-+============================
-+
-+This file provides documentation for the DPAA2 Ethernet Switch driver
-+
-+
-+Contents
-+========
-+ Supported Platforms
-+ Architecture Overview
-+ Creating an Ethernet Switch
-+ Features
-+
-+
-+ Supported Platforms
-+===================
-+This driver provides networking support for Freescale LS2085A, LS2088A
-+DPAA2 SoCs.
-+
-+
-+Architecture Overview
-+=====================
-+The Ethernet Switch in the DPAA2 architecture consists of several hardware
-+resources that provide the functionality. These are allocated and
-+configured via the Management Complex (MC) portals. MC abstracts most of
-+these resources as DPAA2 objects and exposes ABIs through which they can
-+be configured and controlled.
-+
-+For a more detailed description of the DPAA2 architecture and its object
-+abstractions see:
-+ drivers/staging/fsl-mc/README.txt
-+
-+The Ethernet Switch is built on top of a Datapath Switch (DPSW) object.
-+
-+Configuration interface:
-+
-+ ---------------------
-+ | DPAA2 Switch driver |
-+ ---------------------
-+ .
-+ .
-+ ----------
-+ | DPSW API |
-+ ----------
-+ . software
-+ ================= . ==============
-+ . hardware
-+ ---------------------
-+ | MC hardware portals |
-+ ---------------------
-+ .
-+ .
-+ ------
-+ | DPSW |
-+ ------
-+
-+Driver uses the switch device driver model and exposes each switch port as
-+a network interface, which can be included in a bridge. Traffic switched
-+between ports is offloaded into the hardware. Exposed network interfaces
-+are not used for I/O, they are used just for configuration. This
-+limitation is going to be addressed in the future.
-+
-+The DPSW can have ports connected to DPNIs or to PHYs via DPMACs.
-+
-+
-+ [ethA] [ethB] [ethC] [ethD] [ethE] [ethF]
-+ : : : : : :
-+ : : : : : :
-+[eth drv] [eth drv] [ ethsw drv ]
-+ : : : : : : kernel
-+========================================================================
-+ : : : : : : hardware
-+ [DPNI] [DPNI] [============= DPSW =================]
-+ | | | | | |
-+ | ---------- | [DPMAC] [DPMAC]
-+ ------------------------------- | |
-+ | |
-+ [PHY] [PHY]
-+
-+For a more detailed description of the Ethernet switch device driver model
-+see:
-+ Documentation/networking/switchdev.txt
-+
-+Creating an Ethernet Switch
-+===========================
-+A device is created for the switch objects probed on the MC bus. Each DPSW
-+has a number of properties which determine the configuration options and
-+associated hardware resources.
-+
-+A DPSW object (and the other DPAA2 objects needed for a DPAA2 switch) can
-+be added to a container on the MC bus in one of two ways: statically,
-+through a Datapath Layout Binary file (DPL) that is parsed by MC at boot
-+time; or created dynamically at runtime, via the DPAA2 objects APIs.
-+
-+Features
-+========
-+Driver configures DPSW to perform hardware switching offload of
-+unicast/multicast/broadcast (VLAN tagged or untagged) traffic between its
-+ports.
-+
-+It allows configuration of hardware learning, flooding, multicast groups,
-+port VLAN configuration and STP state.
-+
-+Static entries can be added/removed from the FDB.
-+
-+Hardware statistics for each port are provided through ethtool -S option.
---- /dev/null
-+++ b/drivers/staging/fsl-dpaa2/ethsw/TODO
-@@ -0,0 +1,14 @@
-+* Add I/O capabilities on switch port netdevices. This will allow control
-+traffic to reach the CPU.
-+* Add ACL to redirect control traffic to CPU.
-+* Add support for displaying learned FDB entries
-+* MC firmware uprev; the DPAA2 objects used by the Ethernet Switch driver
-+need to be kept in sync with binary interface changes in MC
-+* refine README file
-+* cleanup
-+
-+NOTE: At least first three of the above are required before getting the
-+DPAA2 Ethernet Switch driver out of staging. Another requirement is that
-+the fsl-mc bus driver is moved to drivers/bus and dpio driver is moved to
-+drivers/soc (this is required for I/O).
-+
---- /dev/null
-+++ b/drivers/staging/fsl-dpaa2/ethsw/dpsw-cmd.h
-@@ -0,0 +1,359 @@
-+// SPDX-License-Identifier: GPL-2.0
-+/*
-+ * Copyright 2013-2016 Freescale Semiconductor, Inc.
-+ * Copyright 2017-2018 NXP
-+ *
-+ */
-+
-+#ifndef __FSL_DPSW_CMD_H
-+#define __FSL_DPSW_CMD_H
-+
-+/* DPSW Version */
-+#define DPSW_VER_MAJOR 8
-+#define DPSW_VER_MINOR 0
-+
-+#define DPSW_CMD_BASE_VERSION 1
-+#define DPSW_CMD_ID_OFFSET 4
-+
-+#define DPSW_CMD_ID(id) (((id) << DPSW_CMD_ID_OFFSET) | DPSW_CMD_BASE_VERSION)
-+
-+/* Command IDs */
-+#define DPSW_CMDID_CLOSE DPSW_CMD_ID(0x800)
-+#define DPSW_CMDID_OPEN DPSW_CMD_ID(0x802)
-+
-+#define DPSW_CMDID_GET_API_VERSION DPSW_CMD_ID(0xa02)
-+
-+#define DPSW_CMDID_ENABLE DPSW_CMD_ID(0x002)
-+#define DPSW_CMDID_DISABLE DPSW_CMD_ID(0x003)
-+#define DPSW_CMDID_GET_ATTR DPSW_CMD_ID(0x004)
-+#define DPSW_CMDID_RESET DPSW_CMD_ID(0x005)
-+
-+#define DPSW_CMDID_SET_IRQ_ENABLE DPSW_CMD_ID(0x012)
-+
-+#define DPSW_CMDID_SET_IRQ_MASK DPSW_CMD_ID(0x014)
-+
-+#define DPSW_CMDID_GET_IRQ_STATUS DPSW_CMD_ID(0x016)
-+#define DPSW_CMDID_CLEAR_IRQ_STATUS DPSW_CMD_ID(0x017)
-+
-+#define DPSW_CMDID_IF_SET_TCI DPSW_CMD_ID(0x030)
-+#define DPSW_CMDID_IF_SET_STP DPSW_CMD_ID(0x031)
-+
-+#define DPSW_CMDID_IF_GET_COUNTER DPSW_CMD_ID(0x034)
-+
-+#define DPSW_CMDID_IF_ENABLE DPSW_CMD_ID(0x03D)
-+#define DPSW_CMDID_IF_DISABLE DPSW_CMD_ID(0x03E)
-+
-+#define DPSW_CMDID_IF_SET_MAX_FRAME_LENGTH DPSW_CMD_ID(0x044)
-+
-+#define DPSW_CMDID_IF_GET_LINK_STATE DPSW_CMD_ID(0x046)
-+#define DPSW_CMDID_IF_SET_FLOODING DPSW_CMD_ID(0x047)
-+#define DPSW_CMDID_IF_SET_BROADCAST DPSW_CMD_ID(0x048)
-+
-+#define DPSW_CMDID_IF_GET_TCI DPSW_CMD_ID(0x04A)
-+
-+#define DPSW_CMDID_IF_SET_LINK_CFG DPSW_CMD_ID(0x04C)
-+
-+#define DPSW_CMDID_VLAN_ADD DPSW_CMD_ID(0x060)
-+#define DPSW_CMDID_VLAN_ADD_IF DPSW_CMD_ID(0x061)
-+#define DPSW_CMDID_VLAN_ADD_IF_UNTAGGED DPSW_CMD_ID(0x062)
-+
-+#define DPSW_CMDID_VLAN_REMOVE_IF DPSW_CMD_ID(0x064)
-+#define DPSW_CMDID_VLAN_REMOVE_IF_UNTAGGED DPSW_CMD_ID(0x065)
-+#define DPSW_CMDID_VLAN_REMOVE_IF_FLOODING DPSW_CMD_ID(0x066)
-+#define DPSW_CMDID_VLAN_REMOVE DPSW_CMD_ID(0x067)
-+
-+#define DPSW_CMDID_FDB_ADD_UNICAST DPSW_CMD_ID(0x084)
-+#define DPSW_CMDID_FDB_REMOVE_UNICAST DPSW_CMD_ID(0x085)
-+#define DPSW_CMDID_FDB_ADD_MULTICAST DPSW_CMD_ID(0x086)
-+#define DPSW_CMDID_FDB_REMOVE_MULTICAST DPSW_CMD_ID(0x087)
-+#define DPSW_CMDID_FDB_SET_LEARNING_MODE DPSW_CMD_ID(0x088)
-+
-+/* Macros for accessing command fields smaller than 1byte */
-+#define DPSW_MASK(field) \
-+ GENMASK(DPSW_##field##_SHIFT + DPSW_##field##_SIZE - 1, \
-+ DPSW_##field##_SHIFT)
-+#define dpsw_set_field(var, field, val) \
-+ ((var) |= (((val) << DPSW_##field##_SHIFT) & DPSW_MASK(field)))
-+#define dpsw_get_field(var, field) \
-+ (((var) & DPSW_MASK(field)) >> DPSW_##field##_SHIFT)
-+#define dpsw_get_bit(var, bit) \
-+ (((var) >> (bit)) & GENMASK(0, 0))
-+
-+struct dpsw_cmd_open {
-+ __le32 dpsw_id;
-+};
-+
-+#define DPSW_COMPONENT_TYPE_SHIFT 0
-+#define DPSW_COMPONENT_TYPE_SIZE 4
-+
-+struct dpsw_cmd_create {
-+ /* cmd word 0 */
-+ __le16 num_ifs;
-+ u8 max_fdbs;
-+ u8 max_meters_per_if;
-+ /* from LSB: only the first 4 bits */
-+ u8 component_type;
-+ u8 pad[3];
-+ /* cmd word 1 */
-+ __le16 max_vlans;
-+ __le16 max_fdb_entries;
-+ __le16 fdb_aging_time;
-+ __le16 max_fdb_mc_groups;
-+ /* cmd word 2 */
-+ __le64 options;
-+};
-+
-+struct dpsw_cmd_destroy {
-+ __le32 dpsw_id;
-+};
-+
-+#define DPSW_ENABLE_SHIFT 0
-+#define DPSW_ENABLE_SIZE 1
-+
-+struct dpsw_rsp_is_enabled {
-+ /* from LSB: enable:1 */
-+ u8 enabled;
-+};
-+
-+struct dpsw_cmd_set_irq_enable {
-+ u8 enable_state;
-+ u8 pad[3];
-+ u8 irq_index;
-+};
-+
-+struct dpsw_cmd_get_irq_enable {
-+ __le32 pad;
-+ u8 irq_index;
-+};
-+
-+struct dpsw_rsp_get_irq_enable {
-+ u8 enable_state;
-+};
-+
-+struct dpsw_cmd_set_irq_mask {
-+ __le32 mask;
-+ u8 irq_index;
-+};
-+
-+struct dpsw_cmd_get_irq_mask {
-+ __le32 pad;
-+ u8 irq_index;
-+};
-+
-+struct dpsw_rsp_get_irq_mask {
-+ __le32 mask;
-+};
-+
-+struct dpsw_cmd_get_irq_status {
-+ __le32 status;
-+ u8 irq_index;
-+};
-+
-+struct dpsw_rsp_get_irq_status {
-+ __le32 status;
-+};
-+
-+struct dpsw_cmd_clear_irq_status {
-+ __le32 status;
-+ u8 irq_index;
-+};
-+
-+#define DPSW_COMPONENT_TYPE_SHIFT 0
-+#define DPSW_COMPONENT_TYPE_SIZE 4
-+
-+struct dpsw_rsp_get_attr {
-+ /* cmd word 0 */
-+ __le16 num_ifs;
-+ u8 max_fdbs;
-+ u8 num_fdbs;
-+ __le16 max_vlans;
-+ __le16 num_vlans;
-+ /* cmd word 1 */
-+ __le16 max_fdb_entries;
-+ __le16 fdb_aging_time;
-+ __le32 dpsw_id;
-+ /* cmd word 2 */
-+ __le16 mem_size;
-+ __le16 max_fdb_mc_groups;
-+ u8 max_meters_per_if;
-+ /* from LSB only the first 4 bits */
-+ u8 component_type;
-+ __le16 pad;
-+ /* cmd word 3 */
-+ __le64 options;
-+};
-+
-+struct dpsw_cmd_if_set_flooding {
-+ __le16 if_id;
-+ /* from LSB: enable:1 */
-+ u8 enable;
-+};
-+
-+struct dpsw_cmd_if_set_broadcast {
-+ __le16 if_id;
-+ /* from LSB: enable:1 */
-+ u8 enable;
-+};
-+
-+#define DPSW_VLAN_ID_SHIFT 0
-+#define DPSW_VLAN_ID_SIZE 12
-+#define DPSW_DEI_SHIFT 12
-+#define DPSW_DEI_SIZE 1
-+#define DPSW_PCP_SHIFT 13
-+#define DPSW_PCP_SIZE 3
-+
-+struct dpsw_cmd_if_set_tci {
-+ __le16 if_id;
-+ /* from LSB: VLAN_ID:12 DEI:1 PCP:3 */
-+ __le16 conf;
-+};
-+
-+struct dpsw_cmd_if_get_tci {
-+ __le16 if_id;
-+};
-+
-+struct dpsw_rsp_if_get_tci {
-+ __le16 pad;
-+ __le16 vlan_id;
-+ u8 dei;
-+ u8 pcp;
-+};
-+
-+#define DPSW_STATE_SHIFT 0
-+#define DPSW_STATE_SIZE 4
-+
-+struct dpsw_cmd_if_set_stp {
-+ __le16 if_id;
-+ __le16 vlan_id;
-+ /* only the first LSB 4 bits */
-+ u8 state;
-+};
-+
-+#define DPSW_COUNTER_TYPE_SHIFT 0
-+#define DPSW_COUNTER_TYPE_SIZE 5
-+
-+struct dpsw_cmd_if_get_counter {
-+ __le16 if_id;
-+ /* from LSB: type:5 */
-+ u8 type;
-+};
-+
-+struct dpsw_rsp_if_get_counter {
-+ __le64 pad;
-+ __le64 counter;
-+};
-+
-+struct dpsw_cmd_if {
-+ __le16 if_id;
-+};
-+
-+struct dpsw_cmd_if_set_max_frame_length {
-+ __le16 if_id;
-+ __le16 frame_length;
-+};
-+
-+struct dpsw_cmd_if_set_link_cfg {
-+ /* cmd word 0 */
-+ __le16 if_id;
-+ u8 pad[6];
-+ /* cmd word 1 */
-+ __le32 rate;
-+ __le32 pad1;
-+ /* cmd word 2 */
-+ __le64 options;
-+};
-+
-+struct dpsw_cmd_if_get_link_state {
-+ __le16 if_id;
-+};
-+
-+#define DPSW_UP_SHIFT 0
-+#define DPSW_UP_SIZE 1
-+
-+struct dpsw_rsp_if_get_link_state {
-+ /* cmd word 0 */
-+ __le32 pad0;
-+ u8 up;
-+ u8 pad1[3];
-+ /* cmd word 1 */
-+ __le32 rate;
-+ __le32 pad2;
-+ /* cmd word 2 */
-+ __le64 options;
-+};
-+
-+struct dpsw_vlan_add {
-+ __le16 fdb_id;
-+ __le16 vlan_id;
-+};
-+
-+struct dpsw_cmd_vlan_manage_if {
-+ /* cmd word 0 */
-+ __le16 pad0;
-+ __le16 vlan_id;
-+ __le32 pad1;
-+ /* cmd word 1-4 */
-+ __le64 if_id[4];
-+};
-+
-+struct dpsw_cmd_vlan_remove {
-+ __le16 pad;
-+ __le16 vlan_id;
-+};
-+
-+struct dpsw_cmd_fdb_add {
-+ __le32 pad;
-+ __le16 fdb_aging_time;
-+ __le16 num_fdb_entries;
-+};
-+
-+struct dpsw_rsp_fdb_add {
-+ __le16 fdb_id;
-+};
-+
-+struct dpsw_cmd_fdb_remove {
-+ __le16 fdb_id;
-+};
-+
-+#define DPSW_ENTRY_TYPE_SHIFT 0
-+#define DPSW_ENTRY_TYPE_SIZE 4
-+
-+struct dpsw_cmd_fdb_unicast_op {
-+ /* cmd word 0 */
-+ __le16 fdb_id;
-+ u8 mac_addr[6];
-+ /* cmd word 1 */
-+ __le16 if_egress;
-+ /* only the first 4 bits from LSB */
-+ u8 type;
-+};
-+
-+struct dpsw_cmd_fdb_multicast_op {
-+ /* cmd word 0 */
-+ __le16 fdb_id;
-+ __le16 num_ifs;
-+ /* only the first 4 bits from LSB */
-+ u8 type;
-+ u8 pad[3];
-+ /* cmd word 1 */
-+ u8 mac_addr[6];
-+ __le16 pad2;
-+ /* cmd word 2-5 */
-+ __le64 if_id[4];
-+};
-+
-+#define DPSW_LEARNING_MODE_SHIFT 0
-+#define DPSW_LEARNING_MODE_SIZE 4
-+
-+struct dpsw_cmd_fdb_set_learning_mode {
-+ __le16 fdb_id;
-+ /* only the first 4 bits from LSB */
-+ u8 mode;
-+};
-+
-+struct dpsw_rsp_get_api_version {
-+ __le16 version_major;
-+ __le16 version_minor;
-+};
-+
-+#endif /* __FSL_DPSW_CMD_H */
---- /dev/null
-+++ b/drivers/staging/fsl-dpaa2/ethsw/dpsw.c
-@@ -0,0 +1,1165 @@
-+// SPDX-License-Identifier: GPL-2.0
-+/*
-+ * Copyright 2013-2016 Freescale Semiconductor, Inc.
-+ * Copyright 2017-2018 NXP
-+ *
-+ */
-+
-+#include <linux/fsl/mc.h>
-+#include "dpsw.h"
-+#include "dpsw-cmd.h"
-+
-+static void build_if_id_bitmap(__le64 *bmap,
-+ const u16 *id,
-+ const u16 num_ifs)
-+{
-+ int i;
-+
-+ for (i = 0; (i < num_ifs) && (i < DPSW_MAX_IF); i++) {
-+ if (id[i] < DPSW_MAX_IF)
-+ bmap[id[i] / 64] |= cpu_to_le64(BIT_MASK(id[i] % 64));
-+ }
-+}
-+
-+/**
-+ * dpsw_open() - Open a control session for the specified object
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @dpsw_id: DPSW unique ID
-+ * @token: Returned token; use in subsequent API calls
-+ *
-+ * This function can be used to open a control session for an
-+ * already created object; an object may have been declared in
-+ * the DPL or by calling the dpsw_create() function.
-+ * This function returns a unique authentication token,
-+ * associated with the specific object ID and the specific MC
-+ * portal; this token must be used in all subsequent commands for
-+ * this specific object
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpsw_open(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ int dpsw_id,
-+ u16 *token)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpsw_cmd_open *cmd_params;
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPSW_CMDID_OPEN,
-+ cmd_flags,
-+ 0);
-+ cmd_params = (struct dpsw_cmd_open *)cmd.params;
-+ cmd_params->dpsw_id = cpu_to_le32(dpsw_id);
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ *token = mc_cmd_hdr_read_token(&cmd);
-+
-+ return 0;
-+}
-+
-+/**
-+ * dpsw_close() - Close the control session of the object
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPSW object
-+ *
-+ * After this function is called, no further operations are
-+ * allowed on the object without opening a new control session.
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpsw_close(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPSW_CMDID_CLOSE,
-+ cmd_flags,
-+ token);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpsw_enable() - Enable DPSW functionality
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPSW object
-+ *
-+ * Return: Completion status. '0' on Success; Error code otherwise.
-+ */
-+int dpsw_enable(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPSW_CMDID_ENABLE,
-+ cmd_flags,
-+ token);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpsw_disable() - Disable DPSW functionality
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPSW object
-+ *
-+ * Return: Completion status. '0' on Success; Error code otherwise.
-+ */
-+int dpsw_disable(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPSW_CMDID_DISABLE,
-+ cmd_flags,
-+ token);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpsw_reset() - Reset the DPSW, returns the object to initial state.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPSW object
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpsw_reset(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPSW_CMDID_RESET,
-+ cmd_flags,
-+ token);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpsw_set_irq_enable() - Set overall interrupt state.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPCI object
-+ * @irq_index: The interrupt index to configure
-+ * @en: Interrupt state - enable = 1, disable = 0
-+ *
-+ * Allows GPP software to control when interrupts are generated.
-+ * Each interrupt can have up to 32 causes. The enable/disable control's the
-+ * overall interrupt state. if the interrupt is disabled no causes will cause
-+ * an interrupt
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpsw_set_irq_enable(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 irq_index,
-+ u8 en)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpsw_cmd_set_irq_enable *cmd_params;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPSW_CMDID_SET_IRQ_ENABLE,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpsw_cmd_set_irq_enable *)cmd.params;
-+ dpsw_set_field(cmd_params->enable_state, ENABLE, en);
-+ cmd_params->irq_index = irq_index;
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpsw_set_irq_mask() - Set interrupt mask.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPCI object
-+ * @irq_index: The interrupt index to configure
-+ * @mask: Event mask to trigger interrupt;
-+ * each bit:
-+ * 0 = ignore event
-+ * 1 = consider event for asserting IRQ
-+ *
-+ * Every interrupt can have up to 32 causes and the interrupt model supports
-+ * masking/unmasking each cause independently
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpsw_set_irq_mask(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 irq_index,
-+ u32 mask)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpsw_cmd_set_irq_mask *cmd_params;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPSW_CMDID_SET_IRQ_MASK,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpsw_cmd_set_irq_mask *)cmd.params;
-+ cmd_params->mask = cpu_to_le32(mask);
-+ cmd_params->irq_index = irq_index;
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpsw_get_irq_status() - Get the current status of any pending interrupts
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPSW object
-+ * @irq_index: The interrupt index to configure
-+ * @status: Returned interrupts status - one bit per cause:
-+ * 0 = no interrupt pending
-+ * 1 = interrupt pending
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpsw_get_irq_status(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 irq_index,
-+ u32 *status)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpsw_cmd_get_irq_status *cmd_params;
-+ struct dpsw_rsp_get_irq_status *rsp_params;
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPSW_CMDID_GET_IRQ_STATUS,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpsw_cmd_get_irq_status *)cmd.params;
-+ cmd_params->status = cpu_to_le32(*status);
-+ cmd_params->irq_index = irq_index;
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ rsp_params = (struct dpsw_rsp_get_irq_status *)cmd.params;
-+ *status = le32_to_cpu(rsp_params->status);
-+
-+ return 0;
-+}
-+
-+/**
-+ * dpsw_clear_irq_status() - Clear a pending interrupt's status
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPCI object
-+ * @irq_index: The interrupt index to configure
-+ * @status: bits to clear (W1C) - one bit per cause:
-+ * 0 = don't change
-+ * 1 = clear status bit
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpsw_clear_irq_status(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 irq_index,
-+ u32 status)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpsw_cmd_clear_irq_status *cmd_params;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPSW_CMDID_CLEAR_IRQ_STATUS,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpsw_cmd_clear_irq_status *)cmd.params;
-+ cmd_params->status = cpu_to_le32(status);
-+ cmd_params->irq_index = irq_index;
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpsw_get_attributes() - Retrieve DPSW attributes
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPSW object
-+ * @attr: Returned DPSW attributes
-+ *
-+ * Return: Completion status. '0' on Success; Error code otherwise.
-+ */
-+int dpsw_get_attributes(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ struct dpsw_attr *attr)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpsw_rsp_get_attr *rsp_params;
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPSW_CMDID_GET_ATTR,
-+ cmd_flags,
-+ token);
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ rsp_params = (struct dpsw_rsp_get_attr *)cmd.params;
-+ attr->num_ifs = le16_to_cpu(rsp_params->num_ifs);
-+ attr->max_fdbs = rsp_params->max_fdbs;
-+ attr->num_fdbs = rsp_params->num_fdbs;
-+ attr->max_vlans = le16_to_cpu(rsp_params->max_vlans);
-+ attr->num_vlans = le16_to_cpu(rsp_params->num_vlans);
-+ attr->max_fdb_entries = le16_to_cpu(rsp_params->max_fdb_entries);
-+ attr->fdb_aging_time = le16_to_cpu(rsp_params->fdb_aging_time);
-+ attr->id = le32_to_cpu(rsp_params->dpsw_id);
-+ attr->mem_size = le16_to_cpu(rsp_params->mem_size);
-+ attr->max_fdb_mc_groups = le16_to_cpu(rsp_params->max_fdb_mc_groups);
-+ attr->max_meters_per_if = rsp_params->max_meters_per_if;
-+ attr->options = le64_to_cpu(rsp_params->options);
-+ attr->component_type = dpsw_get_field(rsp_params->component_type,
-+ COMPONENT_TYPE);
-+
-+ return 0;
-+}
-+
-+/**
-+ * dpsw_if_set_link_cfg() - Set the link configuration.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPSW object
-+ * @if_id: Interface id
-+ * @cfg: Link configuration
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpsw_if_set_link_cfg(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 if_id,
-+ struct dpsw_link_cfg *cfg)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpsw_cmd_if_set_link_cfg *cmd_params;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPSW_CMDID_IF_SET_LINK_CFG,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpsw_cmd_if_set_link_cfg *)cmd.params;
-+ cmd_params->if_id = cpu_to_le16(if_id);
-+ cmd_params->rate = cpu_to_le32(cfg->rate);
-+ cmd_params->options = cpu_to_le64(cfg->options);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpsw_if_get_link_state - Return the link state
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPSW object
-+ * @if_id: Interface id
-+ * @state: Link state 1 - linkup, 0 - link down or disconnected
-+ *
-+ * @Return '0' on Success; Error code otherwise.
-+ */
-+int dpsw_if_get_link_state(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 if_id,
-+ struct dpsw_link_state *state)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpsw_cmd_if_get_link_state *cmd_params;
-+ struct dpsw_rsp_if_get_link_state *rsp_params;
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPSW_CMDID_IF_GET_LINK_STATE,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpsw_cmd_if_get_link_state *)cmd.params;
-+ cmd_params->if_id = cpu_to_le16(if_id);
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ rsp_params = (struct dpsw_rsp_if_get_link_state *)cmd.params;
-+ state->rate = le32_to_cpu(rsp_params->rate);
-+ state->options = le64_to_cpu(rsp_params->options);
-+ state->up = dpsw_get_field(rsp_params->up, UP);
-+
-+ return 0;
-+}
-+
-+/**
-+ * dpsw_if_set_flooding() - Enable Disable flooding for particular interface
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPSW object
-+ * @if_id: Interface Identifier
-+ * @en: 1 - enable, 0 - disable
-+ *
-+ * Return: Completion status. '0' on Success; Error code otherwise.
-+ */
-+int dpsw_if_set_flooding(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 if_id,
-+ u8 en)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpsw_cmd_if_set_flooding *cmd_params;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPSW_CMDID_IF_SET_FLOODING,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpsw_cmd_if_set_flooding *)cmd.params;
-+ cmd_params->if_id = cpu_to_le16(if_id);
-+ dpsw_set_field(cmd_params->enable, ENABLE, en);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpsw_if_set_broadcast() - Enable/disable broadcast for particular interface
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPSW object
-+ * @if_id: Interface Identifier
-+ * @en: 1 - enable, 0 - disable
-+ *
-+ * Return: Completion status. '0' on Success; Error code otherwise.
-+ */
-+int dpsw_if_set_broadcast(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 if_id,
-+ u8 en)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpsw_cmd_if_set_broadcast *cmd_params;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPSW_CMDID_IF_SET_BROADCAST,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpsw_cmd_if_set_broadcast *)cmd.params;
-+ cmd_params->if_id = cpu_to_le16(if_id);
-+ dpsw_set_field(cmd_params->enable, ENABLE, en);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpsw_if_set_tci() - Set default VLAN Tag Control Information (TCI)
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPSW object
-+ * @if_id: Interface Identifier
-+ * @cfg: Tag Control Information Configuration
-+ *
-+ * Return: Completion status. '0' on Success; Error code otherwise.
-+ */
-+int dpsw_if_set_tci(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 if_id,
-+ const struct dpsw_tci_cfg *cfg)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpsw_cmd_if_set_tci *cmd_params;
-+ u16 tmp_conf = 0;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPSW_CMDID_IF_SET_TCI,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpsw_cmd_if_set_tci *)cmd.params;
-+ cmd_params->if_id = cpu_to_le16(if_id);
-+ dpsw_set_field(tmp_conf, VLAN_ID, cfg->vlan_id);
-+ dpsw_set_field(tmp_conf, DEI, cfg->dei);
-+ dpsw_set_field(tmp_conf, PCP, cfg->pcp);
-+ cmd_params->conf = cpu_to_le16(tmp_conf);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpsw_if_get_tci() - Get default VLAN Tag Control Information (TCI)
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPSW object
-+ * @if_id: Interface Identifier
-+ * @cfg: Tag Control Information Configuration
-+ *
-+ * Return: Completion status. '0' on Success; Error code otherwise.
-+ */
-+int dpsw_if_get_tci(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 if_id,
-+ struct dpsw_tci_cfg *cfg)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpsw_cmd_if_get_tci *cmd_params;
-+ struct dpsw_rsp_if_get_tci *rsp_params;
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPSW_CMDID_IF_GET_TCI,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpsw_cmd_if_get_tci *)cmd.params;
-+ cmd_params->if_id = cpu_to_le16(if_id);
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ rsp_params = (struct dpsw_rsp_if_get_tci *)cmd.params;
-+ cfg->pcp = rsp_params->pcp;
-+ cfg->dei = rsp_params->dei;
-+ cfg->vlan_id = le16_to_cpu(rsp_params->vlan_id);
-+
-+ return 0;
-+}
-+
-+/**
-+ * dpsw_if_set_stp() - Function sets Spanning Tree Protocol (STP) state.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPSW object
-+ * @if_id: Interface Identifier
-+ * @cfg: STP State configuration parameters
-+ *
-+ * The following STP states are supported -
-+ * blocking, listening, learning, forwarding and disabled.
-+ *
-+ * Return: Completion status. '0' on Success; Error code otherwise.
-+ */
-+int dpsw_if_set_stp(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 if_id,
-+ const struct dpsw_stp_cfg *cfg)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpsw_cmd_if_set_stp *cmd_params;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPSW_CMDID_IF_SET_STP,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpsw_cmd_if_set_stp *)cmd.params;
-+ cmd_params->if_id = cpu_to_le16(if_id);
-+ cmd_params->vlan_id = cpu_to_le16(cfg->vlan_id);
-+ dpsw_set_field(cmd_params->state, STATE, cfg->state);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpsw_if_get_counter() - Get specific counter of particular interface
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPSW object
-+ * @if_id: Interface Identifier
-+ * @type: Counter type
-+ * @counter: return value
-+ *
-+ * Return: Completion status. '0' on Success; Error code otherwise.
-+ */
-+int dpsw_if_get_counter(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 if_id,
-+ enum dpsw_counter type,
-+ u64 *counter)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpsw_cmd_if_get_counter *cmd_params;
-+ struct dpsw_rsp_if_get_counter *rsp_params;
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPSW_CMDID_IF_GET_COUNTER,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpsw_cmd_if_get_counter *)cmd.params;
-+ cmd_params->if_id = cpu_to_le16(if_id);
-+ dpsw_set_field(cmd_params->type, COUNTER_TYPE, type);
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ rsp_params = (struct dpsw_rsp_if_get_counter *)cmd.params;
-+ *counter = le64_to_cpu(rsp_params->counter);
-+
-+ return 0;
-+}
-+
-+/**
-+ * dpsw_if_enable() - Enable Interface
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPSW object
-+ * @if_id: Interface Identifier
-+ *
-+ * Return: Completion status. '0' on Success; Error code otherwise.
-+ */
-+int dpsw_if_enable(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 if_id)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpsw_cmd_if *cmd_params;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPSW_CMDID_IF_ENABLE,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpsw_cmd_if *)cmd.params;
-+ cmd_params->if_id = cpu_to_le16(if_id);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpsw_if_disable() - Disable Interface
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPSW object
-+ * @if_id: Interface Identifier
-+ *
-+ * Return: Completion status. '0' on Success; Error code otherwise.
-+ */
-+int dpsw_if_disable(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 if_id)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpsw_cmd_if *cmd_params;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPSW_CMDID_IF_DISABLE,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpsw_cmd_if *)cmd.params;
-+ cmd_params->if_id = cpu_to_le16(if_id);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpsw_if_set_max_frame_length() - Set Maximum Receive frame length.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPSW object
-+ * @if_id: Interface Identifier
-+ * @frame_length: Maximum Frame Length
-+ *
-+ * Return: Completion status. '0' on Success; Error code otherwise.
-+ */
-+int dpsw_if_set_max_frame_length(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 if_id,
-+ u16 frame_length)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpsw_cmd_if_set_max_frame_length *cmd_params;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPSW_CMDID_IF_SET_MAX_FRAME_LENGTH,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpsw_cmd_if_set_max_frame_length *)cmd.params;
-+ cmd_params->if_id = cpu_to_le16(if_id);
-+ cmd_params->frame_length = cpu_to_le16(frame_length);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpsw_vlan_add() - Adding new VLAN to DPSW.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPSW object
-+ * @vlan_id: VLAN Identifier
-+ * @cfg: VLAN configuration
-+ *
-+ * Only VLAN ID and FDB ID are required parameters here.
-+ * 12 bit VLAN ID is defined in IEEE802.1Q.
-+ * Adding a duplicate VLAN ID is not allowed.
-+ * FDB ID can be shared across multiple VLANs. Shared learning
-+ * is obtained by calling dpsw_vlan_add for multiple VLAN IDs
-+ * with same fdb_id
-+ *
-+ * Return: Completion status. '0' on Success; Error code otherwise.
-+ */
-+int dpsw_vlan_add(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 vlan_id,
-+ const struct dpsw_vlan_cfg *cfg)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpsw_vlan_add *cmd_params;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPSW_CMDID_VLAN_ADD,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpsw_vlan_add *)cmd.params;
-+ cmd_params->fdb_id = cpu_to_le16(cfg->fdb_id);
-+ cmd_params->vlan_id = cpu_to_le16(vlan_id);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpsw_vlan_add_if() - Adding a set of interfaces to an existing VLAN.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPSW object
-+ * @vlan_id: VLAN Identifier
-+ * @cfg: Set of interfaces to add
-+ *
-+ * It adds only interfaces not belonging to this VLAN yet,
-+ * otherwise an error is generated and an entire command is
-+ * ignored. This function can be called numerous times always
-+ * providing required interfaces delta.
-+ *
-+ * Return: Completion status. '0' on Success; Error code otherwise.
-+ */
-+int dpsw_vlan_add_if(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 vlan_id,
-+ const struct dpsw_vlan_if_cfg *cfg)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpsw_cmd_vlan_manage_if *cmd_params;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPSW_CMDID_VLAN_ADD_IF,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpsw_cmd_vlan_manage_if *)cmd.params;
-+ cmd_params->vlan_id = cpu_to_le16(vlan_id);
-+ build_if_id_bitmap(cmd_params->if_id, cfg->if_id, cfg->num_ifs);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpsw_vlan_add_if_untagged() - Defining a set of interfaces that should be
-+ * transmitted as untagged.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPSW object
-+ * @vlan_id: VLAN Identifier
-+ * @cfg: Set of interfaces that should be transmitted as untagged
-+ *
-+ * These interfaces should already belong to this VLAN.
-+ * By default all interfaces are transmitted as tagged.
-+ * Providing un-existing interface or untagged interface that is
-+ * configured untagged already generates an error and the entire
-+ * command is ignored.
-+ *
-+ * Return: Completion status. '0' on Success; Error code otherwise.
-+ */
-+int dpsw_vlan_add_if_untagged(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 vlan_id,
-+ const struct dpsw_vlan_if_cfg *cfg)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpsw_cmd_vlan_manage_if *cmd_params;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPSW_CMDID_VLAN_ADD_IF_UNTAGGED,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpsw_cmd_vlan_manage_if *)cmd.params;
-+ cmd_params->vlan_id = cpu_to_le16(vlan_id);
-+ build_if_id_bitmap(cmd_params->if_id, cfg->if_id, cfg->num_ifs);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpsw_vlan_remove_if() - Remove interfaces from an existing VLAN.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPSW object
-+ * @vlan_id: VLAN Identifier
-+ * @cfg: Set of interfaces that should be removed
-+ *
-+ * Interfaces must belong to this VLAN, otherwise an error
-+ * is returned and an the command is ignored
-+ *
-+ * Return: Completion status. '0' on Success; Error code otherwise.
-+ */
-+int dpsw_vlan_remove_if(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 vlan_id,
-+ const struct dpsw_vlan_if_cfg *cfg)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpsw_cmd_vlan_manage_if *cmd_params;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPSW_CMDID_VLAN_REMOVE_IF,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpsw_cmd_vlan_manage_if *)cmd.params;
-+ cmd_params->vlan_id = cpu_to_le16(vlan_id);
-+ build_if_id_bitmap(cmd_params->if_id, cfg->if_id, cfg->num_ifs);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpsw_vlan_remove_if_untagged() - Define a set of interfaces that should be
-+ * converted from transmitted as untagged to transmit as tagged.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPSW object
-+ * @vlan_id: VLAN Identifier
-+ * @cfg: Set of interfaces that should be removed
-+ *
-+ * Interfaces provided by API have to belong to this VLAN and
-+ * configured untagged, otherwise an error is returned and the
-+ * command is ignored
-+ *
-+ * Return: Completion status. '0' on Success; Error code otherwise.
-+ */
-+int dpsw_vlan_remove_if_untagged(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 vlan_id,
-+ const struct dpsw_vlan_if_cfg *cfg)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpsw_cmd_vlan_manage_if *cmd_params;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPSW_CMDID_VLAN_REMOVE_IF_UNTAGGED,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpsw_cmd_vlan_manage_if *)cmd.params;
-+ cmd_params->vlan_id = cpu_to_le16(vlan_id);
-+ build_if_id_bitmap(cmd_params->if_id, cfg->if_id, cfg->num_ifs);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpsw_vlan_remove() - Remove an entire VLAN
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPSW object
-+ * @vlan_id: VLAN Identifier
-+ *
-+ * Return: Completion status. '0' on Success; Error code otherwise.
-+ */
-+int dpsw_vlan_remove(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 vlan_id)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpsw_cmd_vlan_remove *cmd_params;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPSW_CMDID_VLAN_REMOVE,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpsw_cmd_vlan_remove *)cmd.params;
-+ cmd_params->vlan_id = cpu_to_le16(vlan_id);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpsw_fdb_add_unicast() - Function adds an unicast entry into MAC lookup table
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPSW object
-+ * @fdb_id: Forwarding Database Identifier
-+ * @cfg: Unicast entry configuration
-+ *
-+ * Return: Completion status. '0' on Success; Error code otherwise.
-+ */
-+int dpsw_fdb_add_unicast(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 fdb_id,
-+ const struct dpsw_fdb_unicast_cfg *cfg)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpsw_cmd_fdb_unicast_op *cmd_params;
-+ int i;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPSW_CMDID_FDB_ADD_UNICAST,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpsw_cmd_fdb_unicast_op *)cmd.params;
-+ cmd_params->fdb_id = cpu_to_le16(fdb_id);
-+ cmd_params->if_egress = cpu_to_le16(cfg->if_egress);
-+ for (i = 0; i < 6; i++)
-+ cmd_params->mac_addr[i] = cfg->mac_addr[5 - i];
-+ dpsw_set_field(cmd_params->type, ENTRY_TYPE, cfg->type);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpsw_fdb_remove_unicast() - removes an entry from MAC lookup table
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPSW object
-+ * @fdb_id: Forwarding Database Identifier
-+ * @cfg: Unicast entry configuration
-+ *
-+ * Return: Completion status. '0' on Success; Error code otherwise.
-+ */
-+int dpsw_fdb_remove_unicast(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 fdb_id,
-+ const struct dpsw_fdb_unicast_cfg *cfg)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpsw_cmd_fdb_unicast_op *cmd_params;
-+ int i;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPSW_CMDID_FDB_REMOVE_UNICAST,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpsw_cmd_fdb_unicast_op *)cmd.params;
-+ cmd_params->fdb_id = cpu_to_le16(fdb_id);
-+ for (i = 0; i < 6; i++)
-+ cmd_params->mac_addr[i] = cfg->mac_addr[5 - i];
-+ cmd_params->if_egress = cpu_to_le16(cfg->if_egress);
-+ dpsw_set_field(cmd_params->type, ENTRY_TYPE, cfg->type);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpsw_fdb_add_multicast() - Add a set of egress interfaces to multi-cast group
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPSW object
-+ * @fdb_id: Forwarding Database Identifier
-+ * @cfg: Multicast entry configuration
-+ *
-+ * If group doesn't exist, it will be created.
-+ * It adds only interfaces not belonging to this multicast group
-+ * yet, otherwise error will be generated and the command is
-+ * ignored.
-+ * This function may be called numerous times always providing
-+ * required interfaces delta.
-+ *
-+ * Return: Completion status. '0' on Success; Error code otherwise.
-+ */
-+int dpsw_fdb_add_multicast(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 fdb_id,
-+ const struct dpsw_fdb_multicast_cfg *cfg)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpsw_cmd_fdb_multicast_op *cmd_params;
-+ int i;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPSW_CMDID_FDB_ADD_MULTICAST,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpsw_cmd_fdb_multicast_op *)cmd.params;
-+ cmd_params->fdb_id = cpu_to_le16(fdb_id);
-+ cmd_params->num_ifs = cpu_to_le16(cfg->num_ifs);
-+ dpsw_set_field(cmd_params->type, ENTRY_TYPE, cfg->type);
-+ build_if_id_bitmap(cmd_params->if_id, cfg->if_id, cfg->num_ifs);
-+ for (i = 0; i < 6; i++)
-+ cmd_params->mac_addr[i] = cfg->mac_addr[5 - i];
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpsw_fdb_remove_multicast() - Removing interfaces from an existing multicast
-+ * group.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPSW object
-+ * @fdb_id: Forwarding Database Identifier
-+ * @cfg: Multicast entry configuration
-+ *
-+ * Interfaces provided by this API have to exist in the group,
-+ * otherwise an error will be returned and an entire command
-+ * ignored. If there is no interface left in the group,
-+ * an entire group is deleted
-+ *
-+ * Return: Completion status. '0' on Success; Error code otherwise.
-+ */
-+int dpsw_fdb_remove_multicast(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 fdb_id,
-+ const struct dpsw_fdb_multicast_cfg *cfg)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpsw_cmd_fdb_multicast_op *cmd_params;
-+ int i;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPSW_CMDID_FDB_REMOVE_MULTICAST,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpsw_cmd_fdb_multicast_op *)cmd.params;
-+ cmd_params->fdb_id = cpu_to_le16(fdb_id);
-+ cmd_params->num_ifs = cpu_to_le16(cfg->num_ifs);
-+ dpsw_set_field(cmd_params->type, ENTRY_TYPE, cfg->type);
-+ build_if_id_bitmap(cmd_params->if_id, cfg->if_id, cfg->num_ifs);
-+ for (i = 0; i < 6; i++)
-+ cmd_params->mac_addr[i] = cfg->mac_addr[5 - i];
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpsw_fdb_set_learning_mode() - Define FDB learning mode
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPSW object
-+ * @fdb_id: Forwarding Database Identifier
-+ * @mode: Learning mode
-+ *
-+ * Return: Completion status. '0' on Success; Error code otherwise.
-+ */
-+int dpsw_fdb_set_learning_mode(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 fdb_id,
-+ enum dpsw_fdb_learning_mode mode)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpsw_cmd_fdb_set_learning_mode *cmd_params;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPSW_CMDID_FDB_SET_LEARNING_MODE,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpsw_cmd_fdb_set_learning_mode *)cmd.params;
-+ cmd_params->fdb_id = cpu_to_le16(fdb_id);
-+ dpsw_set_field(cmd_params->mode, LEARNING_MODE, mode);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpsw_get_api_version() - Get Data Path Switch API version
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @major_ver: Major version of data path switch API
-+ * @minor_ver: Minor version of data path switch API
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpsw_get_api_version(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 *major_ver,
-+ u16 *minor_ver)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpsw_rsp_get_api_version *rsp_params;
-+ int err;
-+
-+ cmd.header = mc_encode_cmd_header(DPSW_CMDID_GET_API_VERSION,
-+ cmd_flags,
-+ 0);
-+
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ rsp_params = (struct dpsw_rsp_get_api_version *)cmd.params;
-+ *major_ver = le16_to_cpu(rsp_params->version_major);
-+ *minor_ver = le16_to_cpu(rsp_params->version_minor);
-+
-+ return 0;
-+}
---- /dev/null
-+++ b/drivers/staging/fsl-dpaa2/ethsw/dpsw.h
-@@ -0,0 +1,592 @@
-+// SPDX-License-Identifier: GPL-2.0
-+/*
-+ * Copyright 2013-2016 Freescale Semiconductor, Inc.
-+ * Copyright 2017-2018 NXP
-+ *
-+ */
-+
-+#ifndef __FSL_DPSW_H
-+#define __FSL_DPSW_H
-+
-+/* Data Path L2-Switch API
-+ * Contains API for handling DPSW topology and functionality
-+ */
-+
-+struct fsl_mc_io;
-+
-+/**
-+ * DPSW general definitions
-+ */
-+
-+/**
-+ * Maximum number of traffic class priorities
-+ */
-+#define DPSW_MAX_PRIORITIES 8
-+/**
-+ * Maximum number of interfaces
-+ */
-+#define DPSW_MAX_IF 64
-+
-+int dpsw_open(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ int dpsw_id,
-+ u16 *token);
-+
-+int dpsw_close(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token);
-+
-+/**
-+ * DPSW options
-+ */
-+
-+/**
-+ * Disable flooding
-+ */
-+#define DPSW_OPT_FLOODING_DIS 0x0000000000000001ULL
-+/**
-+ * Disable Multicast
-+ */
-+#define DPSW_OPT_MULTICAST_DIS 0x0000000000000004ULL
-+/**
-+ * Support control interface
-+ */
-+#define DPSW_OPT_CTRL_IF_DIS 0x0000000000000010ULL
-+/**
-+ * Disable flooding metering
-+ */
-+#define DPSW_OPT_FLOODING_METERING_DIS 0x0000000000000020ULL
-+/**
-+ * Enable metering
-+ */
-+#define DPSW_OPT_METERING_EN 0x0000000000000040ULL
-+
-+/**
-+ * enum dpsw_component_type - component type of a bridge
-+ * @DPSW_COMPONENT_TYPE_C_VLAN: A C-VLAN component of an
-+ * enterprise VLAN bridge or of a Provider Bridge used
-+ * to process C-tagged frames
-+ * @DPSW_COMPONENT_TYPE_S_VLAN: An S-VLAN component of a
-+ * Provider Bridge
-+ *
-+ */
-+enum dpsw_component_type {
-+ DPSW_COMPONENT_TYPE_C_VLAN = 0,
-+ DPSW_COMPONENT_TYPE_S_VLAN
-+};
-+
-+/**
-+ * struct dpsw_cfg - DPSW configuration
-+ * @num_ifs: Number of external and internal interfaces
-+ * @adv: Advanced parameters; default is all zeros;
-+ * use this structure to change default settings
-+ */
-+struct dpsw_cfg {
-+ u16 num_ifs;
-+ /**
-+ * struct adv - Advanced parameters
-+ * @options: Enable/Disable DPSW features (bitmap)
-+ * @max_vlans: Maximum Number of VLAN's; 0 - indicates default 16
-+ * @max_meters_per_if: Number of meters per interface
-+ * @max_fdbs: Maximum Number of FDB's; 0 - indicates default 16
-+ * @max_fdb_entries: Number of FDB entries for default FDB table;
-+ * 0 - indicates default 1024 entries.
-+ * @fdb_aging_time: Default FDB aging time for default FDB table;
-+ * 0 - indicates default 300 seconds
-+ * @max_fdb_mc_groups: Number of multicast groups in each FDB table;
-+ * 0 - indicates default 32
-+ * @component_type: Indicates the component type of this bridge
-+ */
-+ struct {
-+ u64 options;
-+ u16 max_vlans;
-+ u8 max_meters_per_if;
-+ u8 max_fdbs;
-+ u16 max_fdb_entries;
-+ u16 fdb_aging_time;
-+ u16 max_fdb_mc_groups;
-+ enum dpsw_component_type component_type;
-+ } adv;
-+};
-+
-+int dpsw_enable(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token);
-+
-+int dpsw_disable(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token);
-+
-+int dpsw_reset(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token);
-+
-+/**
-+ * DPSW IRQ Index and Events
-+ */
-+
-+#define DPSW_IRQ_INDEX_IF 0x0000
-+#define DPSW_IRQ_INDEX_L2SW 0x0001
-+
-+/**
-+ * IRQ event - Indicates that the link state changed
-+ */
-+#define DPSW_IRQ_EVENT_LINK_CHANGED 0x0001
-+
-+/**
-+ * struct dpsw_irq_cfg - IRQ configuration
-+ * @addr: Address that must be written to signal a message-based interrupt
-+ * @val: Value to write into irq_addr address
-+ * @irq_num: A user defined number associated with this IRQ
-+ */
-+struct dpsw_irq_cfg {
-+ u64 addr;
-+ u32 val;
-+ int irq_num;
-+};
-+
-+int dpsw_set_irq_enable(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 irq_index,
-+ u8 en);
-+
-+int dpsw_set_irq_mask(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 irq_index,
-+ u32 mask);
-+
-+int dpsw_get_irq_status(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 irq_index,
-+ u32 *status);
-+
-+int dpsw_clear_irq_status(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 irq_index,
-+ u32 status);
-+
-+/**
-+ * struct dpsw_attr - Structure representing DPSW attributes
-+ * @id: DPSW object ID
-+ * @options: Enable/Disable DPSW features
-+ * @max_vlans: Maximum Number of VLANs
-+ * @max_meters_per_if: Number of meters per interface
-+ * @max_fdbs: Maximum Number of FDBs
-+ * @max_fdb_entries: Number of FDB entries for default FDB table;
-+ * 0 - indicates default 1024 entries.
-+ * @fdb_aging_time: Default FDB aging time for default FDB table;
-+ * 0 - indicates default 300 seconds
-+ * @max_fdb_mc_groups: Number of multicast groups in each FDB table;
-+ * 0 - indicates default 32
-+ * @mem_size: DPSW frame storage memory size
-+ * @num_ifs: Number of interfaces
-+ * @num_vlans: Current number of VLANs
-+ * @num_fdbs: Current number of FDBs
-+ * @component_type: Component type of this bridge
-+ */
-+struct dpsw_attr {
-+ int id;
-+ u64 options;
-+ u16 max_vlans;
-+ u8 max_meters_per_if;
-+ u8 max_fdbs;
-+ u16 max_fdb_entries;
-+ u16 fdb_aging_time;
-+ u16 max_fdb_mc_groups;
-+ u16 num_ifs;
-+ u16 mem_size;
-+ u16 num_vlans;
-+ u8 num_fdbs;
-+ enum dpsw_component_type component_type;
-+};
-+
-+int dpsw_get_attributes(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ struct dpsw_attr *attr);
-+
-+/**
-+ * enum dpsw_action - Action selection for special/control frames
-+ * @DPSW_ACTION_DROP: Drop frame
-+ * @DPSW_ACTION_REDIRECT: Redirect frame to control port
-+ */
-+enum dpsw_action {
-+ DPSW_ACTION_DROP = 0,
-+ DPSW_ACTION_REDIRECT = 1
-+};
-+
-+/**
-+ * Enable auto-negotiation
-+ */
-+#define DPSW_LINK_OPT_AUTONEG 0x0000000000000001ULL
-+/**
-+ * Enable half-duplex mode
-+ */
-+#define DPSW_LINK_OPT_HALF_DUPLEX 0x0000000000000002ULL
-+/**
-+ * Enable pause frames
-+ */
-+#define DPSW_LINK_OPT_PAUSE 0x0000000000000004ULL
-+/**
-+ * Enable a-symmetric pause frames
-+ */
-+#define DPSW_LINK_OPT_ASYM_PAUSE 0x0000000000000008ULL
-+
-+/**
-+ * struct dpsw_link_cfg - Structure representing DPSW link configuration
-+ * @rate: Rate
-+ * @options: Mask of available options; use 'DPSW_LINK_OPT_<X>' values
-+ */
-+struct dpsw_link_cfg {
-+ u32 rate;
-+ u64 options;
-+};
-+
-+int dpsw_if_set_link_cfg(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 if_id,
-+ struct dpsw_link_cfg *cfg);
-+/**
-+ * struct dpsw_link_state - Structure representing DPSW link state
-+ * @rate: Rate
-+ * @options: Mask of available options; use 'DPSW_LINK_OPT_<X>' values
-+ * @up: 0 - covers two cases: down and disconnected, 1 - up
-+ */
-+struct dpsw_link_state {
-+ u32 rate;
-+ u64 options;
-+ u8 up;
-+};
-+
-+int dpsw_if_get_link_state(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 if_id,
-+ struct dpsw_link_state *state);
-+
-+int dpsw_if_set_flooding(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 if_id,
-+ u8 en);
-+
-+int dpsw_if_set_broadcast(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 if_id,
-+ u8 en);
-+
-+/**
-+ * struct dpsw_tci_cfg - Tag Control Information (TCI) configuration
-+ * @pcp: Priority Code Point (PCP): a 3-bit field which refers
-+ * to the IEEE 802.1p priority
-+ * @dei: Drop Eligible Indicator (DEI): a 1-bit field. May be used
-+ * separately or in conjunction with PCP to indicate frames
-+ * eligible to be dropped in the presence of congestion
-+ * @vlan_id: VLAN Identifier (VID): a 12-bit field specifying the VLAN
-+ * to which the frame belongs. The hexadecimal values
-+ * of 0x000 and 0xFFF are reserved;
-+ * all other values may be used as VLAN identifiers,
-+ * allowing up to 4,094 VLANs
-+ */
-+struct dpsw_tci_cfg {
-+ u8 pcp;
-+ u8 dei;
-+ u16 vlan_id;
-+};
-+
-+int dpsw_if_set_tci(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 if_id,
-+ const struct dpsw_tci_cfg *cfg);
-+
-+int dpsw_if_get_tci(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 if_id,
-+ struct dpsw_tci_cfg *cfg);
-+
-+/**
-+ * enum dpsw_stp_state - Spanning Tree Protocol (STP) states
-+ * @DPSW_STP_STATE_BLOCKING: Blocking state
-+ * @DPSW_STP_STATE_LISTENING: Listening state
-+ * @DPSW_STP_STATE_LEARNING: Learning state
-+ * @DPSW_STP_STATE_FORWARDING: Forwarding state
-+ *
-+ */
-+enum dpsw_stp_state {
-+ DPSW_STP_STATE_DISABLED = 0,
-+ DPSW_STP_STATE_LISTENING = 1,
-+ DPSW_STP_STATE_LEARNING = 2,
-+ DPSW_STP_STATE_FORWARDING = 3,
-+ DPSW_STP_STATE_BLOCKING = 0
-+};
-+
-+/**
-+ * struct dpsw_stp_cfg - Spanning Tree Protocol (STP) Configuration
-+ * @vlan_id: VLAN ID STP state
-+ * @state: STP state
-+ */
-+struct dpsw_stp_cfg {
-+ u16 vlan_id;
-+ enum dpsw_stp_state state;
-+};
-+
-+int dpsw_if_set_stp(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 if_id,
-+ const struct dpsw_stp_cfg *cfg);
-+
-+/**
-+ * enum dpsw_accepted_frames - Types of frames to accept
-+ * @DPSW_ADMIT_ALL: The device accepts VLAN tagged, untagged and
-+ * priority tagged frames
-+ * @DPSW_ADMIT_ONLY_VLAN_TAGGED: The device discards untagged frames or
-+ * Priority-Tagged frames received on this interface.
-+ *
-+ */
-+enum dpsw_accepted_frames {
-+ DPSW_ADMIT_ALL = 1,
-+ DPSW_ADMIT_ONLY_VLAN_TAGGED = 3
-+};
-+
-+/**
-+ * enum dpsw_counter - Counters types
-+ * @DPSW_CNT_ING_FRAME: Counts ingress frames
-+ * @DPSW_CNT_ING_BYTE: Counts ingress bytes
-+ * @DPSW_CNT_ING_FLTR_FRAME: Counts filtered ingress frames
-+ * @DPSW_CNT_ING_FRAME_DISCARD: Counts discarded ingress frame
-+ * @DPSW_CNT_ING_MCAST_FRAME: Counts ingress multicast frames
-+ * @DPSW_CNT_ING_MCAST_BYTE: Counts ingress multicast bytes
-+ * @DPSW_CNT_ING_BCAST_FRAME: Counts ingress broadcast frames
-+ * @DPSW_CNT_ING_BCAST_BYTES: Counts ingress broadcast bytes
-+ * @DPSW_CNT_EGR_FRAME: Counts egress frames
-+ * @DPSW_CNT_EGR_BYTE: Counts eEgress bytes
-+ * @DPSW_CNT_EGR_FRAME_DISCARD: Counts discarded egress frames
-+ * @DPSW_CNT_EGR_STP_FRAME_DISCARD: Counts egress STP discarded frames
-+ */
-+enum dpsw_counter {
-+ DPSW_CNT_ING_FRAME = 0x0,
-+ DPSW_CNT_ING_BYTE = 0x1,
-+ DPSW_CNT_ING_FLTR_FRAME = 0x2,
-+ DPSW_CNT_ING_FRAME_DISCARD = 0x3,
-+ DPSW_CNT_ING_MCAST_FRAME = 0x4,
-+ DPSW_CNT_ING_MCAST_BYTE = 0x5,
-+ DPSW_CNT_ING_BCAST_FRAME = 0x6,
-+ DPSW_CNT_ING_BCAST_BYTES = 0x7,
-+ DPSW_CNT_EGR_FRAME = 0x8,
-+ DPSW_CNT_EGR_BYTE = 0x9,
-+ DPSW_CNT_EGR_FRAME_DISCARD = 0xa,
-+ DPSW_CNT_EGR_STP_FRAME_DISCARD = 0xb
-+};
-+
-+int dpsw_if_get_counter(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 if_id,
-+ enum dpsw_counter type,
-+ u64 *counter);
-+
-+int dpsw_if_enable(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 if_id);
-+
-+int dpsw_if_disable(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 if_id);
-+
-+int dpsw_if_set_max_frame_length(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 if_id,
-+ u16 frame_length);
-+
-+/**
-+ * struct dpsw_vlan_cfg - VLAN Configuration
-+ * @fdb_id: Forwarding Data Base
-+ */
-+struct dpsw_vlan_cfg {
-+ u16 fdb_id;
-+};
-+
-+int dpsw_vlan_add(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 vlan_id,
-+ const struct dpsw_vlan_cfg *cfg);
-+
-+/**
-+ * struct dpsw_vlan_if_cfg - Set of VLAN Interfaces
-+ * @num_ifs: The number of interfaces that are assigned to the egress
-+ * list for this VLAN
-+ * @if_id: The set of interfaces that are
-+ * assigned to the egress list for this VLAN
-+ */
-+struct dpsw_vlan_if_cfg {
-+ u16 num_ifs;
-+ u16 if_id[DPSW_MAX_IF];
-+};
-+
-+int dpsw_vlan_add_if(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 vlan_id,
-+ const struct dpsw_vlan_if_cfg *cfg);
-+
-+int dpsw_vlan_add_if_untagged(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 vlan_id,
-+ const struct dpsw_vlan_if_cfg *cfg);
-+
-+int dpsw_vlan_remove_if(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 vlan_id,
-+ const struct dpsw_vlan_if_cfg *cfg);
-+
-+int dpsw_vlan_remove_if_untagged(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 vlan_id,
-+ const struct dpsw_vlan_if_cfg *cfg);
-+
-+int dpsw_vlan_remove(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 vlan_id);
-+
-+/**
-+ * enum dpsw_fdb_entry_type - FDB Entry type - Static/Dynamic
-+ * @DPSW_FDB_ENTRY_STATIC: Static entry
-+ * @DPSW_FDB_ENTRY_DINAMIC: Dynamic entry
-+ */
-+enum dpsw_fdb_entry_type {
-+ DPSW_FDB_ENTRY_STATIC = 0,
-+ DPSW_FDB_ENTRY_DINAMIC = 1
-+};
-+
-+/**
-+ * struct dpsw_fdb_unicast_cfg - Unicast entry configuration
-+ * @type: Select static or dynamic entry
-+ * @mac_addr: MAC address
-+ * @if_egress: Egress interface ID
-+ */
-+struct dpsw_fdb_unicast_cfg {
-+ enum dpsw_fdb_entry_type type;
-+ u8 mac_addr[6];
-+ u16 if_egress;
-+};
-+
-+int dpsw_fdb_add_unicast(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 fdb_id,
-+ const struct dpsw_fdb_unicast_cfg *cfg);
-+
-+int dpsw_fdb_remove_unicast(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 fdb_id,
-+ const struct dpsw_fdb_unicast_cfg *cfg);
-+
-+/**
-+ * struct dpsw_fdb_multicast_cfg - Multi-cast entry configuration
-+ * @type: Select static or dynamic entry
-+ * @mac_addr: MAC address
-+ * @num_ifs: Number of external and internal interfaces
-+ * @if_id: Egress interface IDs
-+ */
-+struct dpsw_fdb_multicast_cfg {
-+ enum dpsw_fdb_entry_type type;
-+ u8 mac_addr[6];
-+ u16 num_ifs;
-+ u16 if_id[DPSW_MAX_IF];
-+};
-+
-+int dpsw_fdb_add_multicast(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 fdb_id,
-+ const struct dpsw_fdb_multicast_cfg *cfg);
-+
-+int dpsw_fdb_remove_multicast(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 fdb_id,
-+ const struct dpsw_fdb_multicast_cfg *cfg);
-+
-+/**
-+ * enum dpsw_fdb_learning_mode - Auto-learning modes
-+ * @DPSW_FDB_LEARNING_MODE_DIS: Disable Auto-learning
-+ * @DPSW_FDB_LEARNING_MODE_HW: Enable HW auto-Learning
-+ * @DPSW_FDB_LEARNING_MODE_NON_SECURE: Enable None secure learning by CPU
-+ * @DPSW_FDB_LEARNING_MODE_SECURE: Enable secure learning by CPU
-+ *
-+ * NONE - SECURE LEARNING
-+ * SMAC found DMAC found CTLU Action
-+ * v v Forward frame to
-+ * 1. DMAC destination
-+ * - v Forward frame to
-+ * 1. DMAC destination
-+ * 2. Control interface
-+ * v - Forward frame to
-+ * 1. Flooding list of interfaces
-+ * - - Forward frame to
-+ * 1. Flooding list of interfaces
-+ * 2. Control interface
-+ * SECURE LEARING
-+ * SMAC found DMAC found CTLU Action
-+ * v v Forward frame to
-+ * 1. DMAC destination
-+ * - v Forward frame to
-+ * 1. Control interface
-+ * v - Forward frame to
-+ * 1. Flooding list of interfaces
-+ * - - Forward frame to
-+ * 1. Control interface
-+ */
-+enum dpsw_fdb_learning_mode {
-+ DPSW_FDB_LEARNING_MODE_DIS = 0,
-+ DPSW_FDB_LEARNING_MODE_HW = 1,
-+ DPSW_FDB_LEARNING_MODE_NON_SECURE = 2,
-+ DPSW_FDB_LEARNING_MODE_SECURE = 3
-+};
-+
-+int dpsw_fdb_set_learning_mode(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 fdb_id,
-+ enum dpsw_fdb_learning_mode mode);
-+
-+/**
-+ * struct dpsw_fdb_attr - FDB Attributes
-+ * @max_fdb_entries: Number of FDB entries
-+ * @fdb_aging_time: Aging time in seconds
-+ * @learning_mode: Learning mode
-+ * @num_fdb_mc_groups: Current number of multicast groups
-+ * @max_fdb_mc_groups: Maximum number of multicast groups
-+ */
-+struct dpsw_fdb_attr {
-+ u16 max_fdb_entries;
-+ u16 fdb_aging_time;
-+ enum dpsw_fdb_learning_mode learning_mode;
-+ u16 num_fdb_mc_groups;
-+ u16 max_fdb_mc_groups;
-+};
-+
-+int dpsw_get_api_version(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 *major_ver,
-+ u16 *minor_ver);
-+
-+#endif /* __FSL_DPSW_H */
---- /dev/null
-+++ b/drivers/staging/fsl-dpaa2/ethsw/ethsw-ethtool.c
-@@ -0,0 +1,206 @@
-+/* Copyright 2014-2016 Freescale Semiconductor Inc.
-+ * Copyright 2017 NXP
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of the above-listed copyright holders nor the
-+ * names of any contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
-+ * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
-+ * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
-+ * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE
-+ * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
-+ * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
-+ * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
-+ * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
-+ * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
-+ * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
-+ * POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#include "ethsw.h"
-+
-+static struct {
-+ enum dpsw_counter id;
-+ char name[ETH_GSTRING_LEN];
-+} ethsw_ethtool_counters[] = {
-+ {DPSW_CNT_ING_FRAME, "rx frames"},
-+ {DPSW_CNT_ING_BYTE, "rx bytes"},
-+ {DPSW_CNT_ING_FLTR_FRAME, "rx filtered frames"},
-+ {DPSW_CNT_ING_FRAME_DISCARD, "rx discarded frames"},
-+ {DPSW_CNT_ING_BCAST_FRAME, "rx b-cast frames"},
-+ {DPSW_CNT_ING_BCAST_BYTES, "rx b-cast bytes"},
-+ {DPSW_CNT_ING_MCAST_FRAME, "rx m-cast frames"},
-+ {DPSW_CNT_ING_MCAST_BYTE, "rx m-cast bytes"},
-+ {DPSW_CNT_EGR_FRAME, "tx frames"},
-+ {DPSW_CNT_EGR_BYTE, "tx bytes"},
-+ {DPSW_CNT_EGR_FRAME_DISCARD, "tx discarded frames"},
-+
-+};
-+
-+#define ETHSW_NUM_COUNTERS ARRAY_SIZE(ethsw_ethtool_counters)
-+
-+static void ethsw_get_drvinfo(struct net_device *netdev,
-+ struct ethtool_drvinfo *drvinfo)
-+{
-+ struct ethsw_port_priv *port_priv = netdev_priv(netdev);
-+ u16 version_major, version_minor;
-+ int err;
-+
-+ strlcpy(drvinfo->driver, KBUILD_MODNAME, sizeof(drvinfo->driver));
-+
-+ err = dpsw_get_api_version(port_priv->ethsw_data->mc_io, 0,
-+ &version_major,
-+ &version_minor);
-+ if (err)
-+ strlcpy(drvinfo->fw_version, "N/A",
-+ sizeof(drvinfo->fw_version));
-+ else
-+ snprintf(drvinfo->fw_version, sizeof(drvinfo->fw_version),
-+ "%u.%u", version_major, version_minor);
-+
-+ strlcpy(drvinfo->bus_info, dev_name(netdev->dev.parent->parent),
-+ sizeof(drvinfo->bus_info));
-+}
-+
-+static int
-+ethsw_get_link_ksettings(struct net_device *netdev,
-+ struct ethtool_link_ksettings *link_ksettings)
-+{
-+ struct ethsw_port_priv *port_priv = netdev_priv(netdev);
-+ struct dpsw_link_state state = {0};
-+ int err = 0;
-+
-+ err = dpsw_if_get_link_state(port_priv->ethsw_data->mc_io, 0,
-+ port_priv->ethsw_data->dpsw_handle,
-+ port_priv->idx,
-+ &state);
-+ if (err) {
-+ netdev_err(netdev, "ERROR %d getting link state", err);
-+ goto out;
-+ }
-+
-+ /* At the moment, we have no way of interrogating the DPMAC
-+ * from the DPSW side or there may not exist a DPMAC at all.
-+ * Report only autoneg state, duplexity and speed.
-+ */
-+ if (state.options & DPSW_LINK_OPT_AUTONEG)
-+ link_ksettings->base.autoneg = AUTONEG_ENABLE;
-+ if (!(state.options & DPSW_LINK_OPT_HALF_DUPLEX))
-+ link_ksettings->base.duplex = DUPLEX_FULL;
-+ link_ksettings->base.speed = state.rate;
-+
-+out:
-+ return err;
-+}
-+
-+static int
-+ethsw_set_link_ksettings(struct net_device *netdev,
-+ const struct ethtool_link_ksettings *link_ksettings)
-+{
-+ struct ethsw_port_priv *port_priv = netdev_priv(netdev);
-+ struct dpsw_link_cfg cfg = {0};
-+ int err = 0;
-+
-+ netdev_dbg(netdev, "Setting link parameters...");
-+
-+ /* Due to a temporary MC limitation, the DPSW port must be down
-+ * in order to be able to change link settings. Taking steps to let
-+ * the user know that.
-+ */
-+ if (netif_running(netdev)) {
-+ netdev_info(netdev, "Sorry, interface must be brought down first.\n");
-+ return -EACCES;
-+ }
-+
-+ cfg.rate = link_ksettings->base.speed;
-+ if (link_ksettings->base.autoneg == AUTONEG_ENABLE)
-+ cfg.options |= DPSW_LINK_OPT_AUTONEG;
-+ else
-+ cfg.options &= ~DPSW_LINK_OPT_AUTONEG;
-+ if (link_ksettings->base.duplex == DUPLEX_HALF)
-+ cfg.options |= DPSW_LINK_OPT_HALF_DUPLEX;
-+ else
-+ cfg.options &= ~DPSW_LINK_OPT_HALF_DUPLEX;
-+
-+ err = dpsw_if_set_link_cfg(port_priv->ethsw_data->mc_io, 0,
-+ port_priv->ethsw_data->dpsw_handle,
-+ port_priv->idx,
-+ &cfg);
-+ if (err)
-+ /* ethtool will be loud enough if we return an error; no point
-+ * in putting our own error message on the console by default
-+ */
-+ netdev_dbg(netdev, "ERROR %d setting link cfg", err);
-+
-+ return err;
-+}
-+
-+static int ethsw_ethtool_get_sset_count(struct net_device *dev, int sset)
-+{
-+ switch (sset) {
-+ case ETH_SS_STATS:
-+ return ETHSW_NUM_COUNTERS;
-+ default:
-+ return -EOPNOTSUPP;
-+ }
-+}
-+
-+static void ethsw_ethtool_get_strings(struct net_device *netdev,
-+ u32 stringset, u8 *data)
-+{
-+ int i;
-+
-+ switch (stringset) {
-+ case ETH_SS_STATS:
-+ for (i = 0; i < ETHSW_NUM_COUNTERS; i++)
-+ memcpy(data + i * ETH_GSTRING_LEN,
-+ ethsw_ethtool_counters[i].name, ETH_GSTRING_LEN);
-+ break;
-+ }
-+}
-+
-+static void ethsw_ethtool_get_stats(struct net_device *netdev,
-+ struct ethtool_stats *stats,
-+ u64 *data)
-+{
-+ struct ethsw_port_priv *port_priv = netdev_priv(netdev);
-+ int i, err;
-+
-+ memset(data, 0,
-+ sizeof(u64) * ETHSW_NUM_COUNTERS);
-+
-+ for (i = 0; i < ETHSW_NUM_COUNTERS; i++) {
-+ err = dpsw_if_get_counter(port_priv->ethsw_data->mc_io, 0,
-+ port_priv->ethsw_data->dpsw_handle,
-+ port_priv->idx,
-+ ethsw_ethtool_counters[i].id,
-+ &data[i]);
-+ if (err)
-+ netdev_err(netdev, "dpsw_if_get_counter[%s] err %d\n",
-+ ethsw_ethtool_counters[i].name, err);
-+ }
-+}
-+
-+const struct ethtool_ops ethsw_port_ethtool_ops = {
-+ .get_drvinfo = ethsw_get_drvinfo,
-+ .get_link = ethtool_op_get_link,
-+ .get_link_ksettings = ethsw_get_link_ksettings,
-+ .set_link_ksettings = ethsw_set_link_ksettings,
-+ .get_strings = ethsw_ethtool_get_strings,
-+ .get_ethtool_stats = ethsw_ethtool_get_stats,
-+ .get_sset_count = ethsw_ethtool_get_sset_count,
-+};
---- /dev/null
-+++ b/drivers/staging/fsl-dpaa2/ethsw/ethsw.c
-@@ -0,0 +1,1438 @@
-+/* Copyright 2014-2016 Freescale Semiconductor Inc.
-+ * Copyright 2017 NXP
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of the above-listed copyright holders nor the
-+ * names of any contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
-+ * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
-+ * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
-+ * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE
-+ * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
-+ * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
-+ * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
-+ * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
-+ * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
-+ * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
-+ * POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#include <linux/module.h>
-+
-+#include <linux/interrupt.h>
-+#include <linux/msi.h>
-+#include <linux/kthread.h>
-+#include <linux/workqueue.h>
-+
-+#include <linux/fsl/mc.h>
-+
-+#include "ethsw.h"
-+
-+static struct workqueue_struct *ethsw_owq;
-+
-+/* Minimal supported DPSW version */
-+#define DPSW_MIN_VER_MAJOR 8
-+#define DPSW_MIN_VER_MINOR 0
-+
-+#define DEFAULT_VLAN_ID 1
-+
-+static int ethsw_add_vlan(struct ethsw_core *ethsw, u16 vid)
-+{
-+ int err;
-+
-+ struct dpsw_vlan_cfg vcfg = {
-+ .fdb_id = 0,
-+ };
-+
-+ if (ethsw->vlans[vid]) {
-+ dev_err(ethsw->dev, "VLAN already configured\n");
-+ return -EEXIST;
-+ }
-+
-+ err = dpsw_vlan_add(ethsw->mc_io, 0,
-+ ethsw->dpsw_handle, vid, &vcfg);
-+ if (err) {
-+ dev_err(ethsw->dev, "dpsw_vlan_add err %d\n", err);
-+ return err;
-+ }
-+ ethsw->vlans[vid] = ETHSW_VLAN_MEMBER;
-+
-+ return 0;
-+}
-+
-+static int ethsw_port_set_pvid(struct ethsw_port_priv *port_priv, u16 pvid)
-+{
-+ struct ethsw_core *ethsw = port_priv->ethsw_data;
-+ struct net_device *netdev = port_priv->netdev;
-+ struct dpsw_tci_cfg tci_cfg = { 0 };
-+ bool is_oper;
-+ int err, ret;
-+
-+ err = dpsw_if_get_tci(ethsw->mc_io, 0, ethsw->dpsw_handle,
-+ port_priv->idx, &tci_cfg);
-+ if (err) {
-+ netdev_err(netdev, "dpsw_if_get_tci err %d\n", err);
-+ return err;
-+ }
-+
-+ tci_cfg.vlan_id = pvid;
-+
-+ /* Interface needs to be down to change PVID */
-+ is_oper = netif_oper_up(netdev);
-+ if (is_oper) {
-+ err = dpsw_if_disable(ethsw->mc_io, 0,
-+ ethsw->dpsw_handle,
-+ port_priv->idx);
-+ if (err) {
-+ netdev_err(netdev, "dpsw_if_disable err %d\n", err);
-+ return err;
-+ }
-+ }
-+
-+ err = dpsw_if_set_tci(ethsw->mc_io, 0, ethsw->dpsw_handle,
-+ port_priv->idx, &tci_cfg);
-+ if (err) {
-+ netdev_err(netdev, "dpsw_if_set_tci err %d\n", err);
-+ goto set_tci_error;
-+ }
-+
-+ /* Delete previous PVID info and mark the new one */
-+ port_priv->vlans[port_priv->pvid] &= ~ETHSW_VLAN_PVID;
-+ port_priv->vlans[pvid] |= ETHSW_VLAN_PVID;
-+ port_priv->pvid = pvid;
-+
-+set_tci_error:
-+ if (is_oper) {
-+ ret = dpsw_if_enable(ethsw->mc_io, 0,
-+ ethsw->dpsw_handle,
-+ port_priv->idx);
-+ if (ret) {
-+ netdev_err(netdev, "dpsw_if_enable err %d\n", ret);
-+ return ret;
-+ }
-+ }
-+
-+ return err;
-+}
-+
-+static int ethsw_port_add_vlan(struct ethsw_port_priv *port_priv,
-+ u16 vid, u16 flags)
-+{
-+ struct ethsw_core *ethsw = port_priv->ethsw_data;
-+ struct net_device *netdev = port_priv->netdev;
-+ struct dpsw_vlan_if_cfg vcfg;
-+ int err;
-+
-+ if (port_priv->vlans[vid]) {
-+ netdev_warn(netdev, "VLAN %d already configured\n", vid);
-+ return -EEXIST;
-+ }
-+
-+ vcfg.num_ifs = 1;
-+ vcfg.if_id[0] = port_priv->idx;
-+ err = dpsw_vlan_add_if(ethsw->mc_io, 0, ethsw->dpsw_handle, vid, &vcfg);
-+ if (err) {
-+ netdev_err(netdev, "dpsw_vlan_add_if err %d\n", err);
-+ return err;
-+ }
-+
-+ port_priv->vlans[vid] = ETHSW_VLAN_MEMBER;
-+
-+ if (flags & BRIDGE_VLAN_INFO_UNTAGGED) {
-+ err = dpsw_vlan_add_if_untagged(ethsw->mc_io, 0,
-+ ethsw->dpsw_handle,
-+ vid, &vcfg);
-+ if (err) {
-+ netdev_err(netdev,
-+ "dpsw_vlan_add_if_untagged err %d\n", err);
-+ return err;
-+ }
-+ port_priv->vlans[vid] |= ETHSW_VLAN_UNTAGGED;
-+ }
-+
-+ if (flags & BRIDGE_VLAN_INFO_PVID) {
-+ err = ethsw_port_set_pvid(port_priv, vid);
-+ if (err)
-+ return err;
-+ }
-+
-+ return 0;
-+}
-+
-+static int ethsw_set_learning(struct ethsw_core *ethsw, u8 flag)
-+{
-+ enum dpsw_fdb_learning_mode learn_mode;
-+ int err;
-+
-+ if (flag)
-+ learn_mode = DPSW_FDB_LEARNING_MODE_HW;
-+ else
-+ learn_mode = DPSW_FDB_LEARNING_MODE_DIS;
-+
-+ err = dpsw_fdb_set_learning_mode(ethsw->mc_io, 0, ethsw->dpsw_handle, 0,
-+ learn_mode);
-+ if (err) {
-+ dev_err(ethsw->dev, "dpsw_fdb_set_learning_mode err %d\n", err);
-+ return err;
-+ }
-+ ethsw->learning = !!flag;
-+
-+ return 0;
-+}
-+
-+static int ethsw_port_set_flood(struct ethsw_port_priv *port_priv, u8 flag)
-+{
-+ int err;
-+
-+ err = dpsw_if_set_flooding(port_priv->ethsw_data->mc_io, 0,
-+ port_priv->ethsw_data->dpsw_handle,
-+ port_priv->idx, flag);
-+ if (err) {
-+ netdev_err(port_priv->netdev,
-+ "dpsw_fdb_set_learning_mode err %d\n", err);
-+ return err;
-+ }
-+ port_priv->flood = !!flag;
-+
-+ return 0;
-+}
-+
-+static int ethsw_port_set_stp_state(struct ethsw_port_priv *port_priv, u8 state)
-+{
-+ struct dpsw_stp_cfg stp_cfg = {
-+ .vlan_id = DEFAULT_VLAN_ID,
-+ .state = state,
-+ };
-+ int err;
-+
-+ if (!netif_oper_up(port_priv->netdev) || state == port_priv->stp_state)
-+ return 0; /* Nothing to do */
-+
-+ err = dpsw_if_set_stp(port_priv->ethsw_data->mc_io, 0,
-+ port_priv->ethsw_data->dpsw_handle,
-+ port_priv->idx, &stp_cfg);
-+ if (err) {
-+ netdev_err(port_priv->netdev,
-+ "dpsw_if_set_stp err %d\n", err);
-+ return err;
-+ }
-+
-+ port_priv->stp_state = state;
-+
-+ return 0;
-+}
-+
-+static int ethsw_dellink_switch(struct ethsw_core *ethsw, u16 vid)
-+{
-+ struct ethsw_port_priv *ppriv_local = NULL;
-+ int i, err;
-+
-+ if (!ethsw->vlans[vid])
-+ return -ENOENT;
-+
-+ err = dpsw_vlan_remove(ethsw->mc_io, 0, ethsw->dpsw_handle, vid);
-+ if (err) {
-+ dev_err(ethsw->dev, "dpsw_vlan_remove err %d\n", err);
-+ return err;
-+ }
-+ ethsw->vlans[vid] = 0;
-+
-+ for (i = 0; i < ethsw->sw_attr.num_ifs; i++) {
-+ ppriv_local = ethsw->ports[i];
-+ ppriv_local->vlans[vid] = 0;
-+ }
-+
-+ return 0;
-+}
-+
-+static int ethsw_port_fdb_add_uc(struct ethsw_port_priv *port_priv,
-+ const unsigned char *addr)
-+{
-+ struct dpsw_fdb_unicast_cfg entry = {0};
-+ int err;
-+
-+ entry.if_egress = port_priv->idx;
-+ entry.type = DPSW_FDB_ENTRY_STATIC;
-+ ether_addr_copy(entry.mac_addr, addr);
-+
-+ err = dpsw_fdb_add_unicast(port_priv->ethsw_data->mc_io, 0,
-+ port_priv->ethsw_data->dpsw_handle,
-+ 0, &entry);
-+ if (err)
-+ netdev_err(port_priv->netdev,
-+ "dpsw_fdb_add_unicast err %d\n", err);
-+ return err;
-+}
-+
-+static int ethsw_port_fdb_del_uc(struct ethsw_port_priv *port_priv,
-+ const unsigned char *addr)
-+{
-+ struct dpsw_fdb_unicast_cfg entry = {0};
-+ int err;
-+
-+ entry.if_egress = port_priv->idx;
-+ entry.type = DPSW_FDB_ENTRY_STATIC;
-+ ether_addr_copy(entry.mac_addr, addr);
-+
-+ err = dpsw_fdb_remove_unicast(port_priv->ethsw_data->mc_io, 0,
-+ port_priv->ethsw_data->dpsw_handle,
-+ 0, &entry);
-+ /* Silently discard calling multiple times the del command */
-+ if (err && err != -ENXIO)
-+ netdev_err(port_priv->netdev,
-+ "dpsw_fdb_remove_unicast err %d\n", err);
-+ return err;
-+}
-+
-+static int ethsw_port_fdb_add_mc(struct ethsw_port_priv *port_priv,
-+ const unsigned char *addr)
-+{
-+ struct dpsw_fdb_multicast_cfg entry = {0};
-+ int err;
-+
-+ ether_addr_copy(entry.mac_addr, addr);
-+ entry.type = DPSW_FDB_ENTRY_STATIC;
-+ entry.num_ifs = 1;
-+ entry.if_id[0] = port_priv->idx;
-+
-+ err = dpsw_fdb_add_multicast(port_priv->ethsw_data->mc_io, 0,
-+ port_priv->ethsw_data->dpsw_handle,
-+ 0, &entry);
-+ /* Silently discard calling multiple times the add command */
-+ if (err && err != -ENXIO)
-+ netdev_err(port_priv->netdev, "dpsw_fdb_add_multicast err %d\n",
-+ err);
-+ return err;
-+}
-+
-+static int ethsw_port_fdb_del_mc(struct ethsw_port_priv *port_priv,
-+ const unsigned char *addr)
-+{
-+ struct dpsw_fdb_multicast_cfg entry = {0};
-+ int err;
-+
-+ ether_addr_copy(entry.mac_addr, addr);
-+ entry.type = DPSW_FDB_ENTRY_STATIC;
-+ entry.num_ifs = 1;
-+ entry.if_id[0] = port_priv->idx;
-+
-+ err = dpsw_fdb_remove_multicast(port_priv->ethsw_data->mc_io, 0,
-+ port_priv->ethsw_data->dpsw_handle,
-+ 0, &entry);
-+ /* Silently discard calling multiple times the del command */
-+ if (err && err != -ENAVAIL)
-+ netdev_err(port_priv->netdev,
-+ "dpsw_fdb_remove_multicast err %d\n", err);
-+ return err;
-+}
-+
-+static void port_get_stats(struct net_device *netdev,
-+ struct rtnl_link_stats64 *stats)
-+{
-+ struct ethsw_port_priv *port_priv = netdev_priv(netdev);
-+ u64 tmp;
-+ int err;
-+
-+ err = dpsw_if_get_counter(port_priv->ethsw_data->mc_io, 0,
-+ port_priv->ethsw_data->dpsw_handle,
-+ port_priv->idx,
-+ DPSW_CNT_ING_FRAME, &stats->rx_packets);
-+ if (err)
-+ goto error;
-+
-+ err = dpsw_if_get_counter(port_priv->ethsw_data->mc_io, 0,
-+ port_priv->ethsw_data->dpsw_handle,
-+ port_priv->idx,
-+ DPSW_CNT_EGR_FRAME, &stats->tx_packets);
-+ if (err)
-+ goto error;
-+
-+ err = dpsw_if_get_counter(port_priv->ethsw_data->mc_io, 0,
-+ port_priv->ethsw_data->dpsw_handle,
-+ port_priv->idx,
-+ DPSW_CNT_ING_BYTE, &stats->rx_bytes);
-+ if (err)
-+ goto error;
-+
-+ err = dpsw_if_get_counter(port_priv->ethsw_data->mc_io, 0,
-+ port_priv->ethsw_data->dpsw_handle,
-+ port_priv->idx,
-+ DPSW_CNT_EGR_BYTE, &stats->tx_bytes);
-+ if (err)
-+ goto error;
-+
-+ err = dpsw_if_get_counter(port_priv->ethsw_data->mc_io, 0,
-+ port_priv->ethsw_data->dpsw_handle,
-+ port_priv->idx,
-+ DPSW_CNT_ING_FRAME_DISCARD,
-+ &stats->rx_dropped);
-+ if (err)
-+ goto error;
-+
-+ err = dpsw_if_get_counter(port_priv->ethsw_data->mc_io, 0,
-+ port_priv->ethsw_data->dpsw_handle,
-+ port_priv->idx,
-+ DPSW_CNT_ING_FLTR_FRAME,
-+ &tmp);
-+ if (err)
-+ goto error;
-+ stats->rx_dropped += tmp;
-+
-+ err = dpsw_if_get_counter(port_priv->ethsw_data->mc_io, 0,
-+ port_priv->ethsw_data->dpsw_handle,
-+ port_priv->idx,
-+ DPSW_CNT_EGR_FRAME_DISCARD,
-+ &stats->tx_dropped);
-+ if (err)
-+ goto error;
-+
-+ return;
-+
-+error:
-+ netdev_err(netdev, "dpsw_if_get_counter err %d\n", err);
-+}
-+
-+static bool port_has_offload_stats(const struct net_device *netdev,
-+ int attr_id)
-+{
-+ return (attr_id == IFLA_OFFLOAD_XSTATS_CPU_HIT);
-+}
-+
-+static int port_get_offload_stats(int attr_id,
-+ const struct net_device *netdev,
-+ void *sp)
-+{
-+ switch (attr_id) {
-+ case IFLA_OFFLOAD_XSTATS_CPU_HIT:
-+ port_get_stats((struct net_device *)netdev, sp);
-+ return 0;
-+ }
-+
-+ return -EINVAL;
-+}
-+
-+static int port_change_mtu(struct net_device *netdev, int mtu)
-+{
-+ struct ethsw_port_priv *port_priv = netdev_priv(netdev);
-+ int err;
-+
-+ err = dpsw_if_set_max_frame_length(port_priv->ethsw_data->mc_io,
-+ 0,
-+ port_priv->ethsw_data->dpsw_handle,
-+ port_priv->idx,
-+ (u16)ETHSW_L2_MAX_FRM(mtu));
-+ if (err) {
-+ netdev_err(netdev,
-+ "dpsw_if_set_max_frame_length() err %d\n", err);
-+ return err;
-+ }
-+
-+ netdev->mtu = mtu;
-+ return 0;
-+}
-+
-+static int port_carrier_state_sync(struct net_device *netdev)
-+{
-+ struct ethsw_port_priv *port_priv = netdev_priv(netdev);
-+ struct dpsw_link_state state;
-+ int err;
-+
-+ err = dpsw_if_get_link_state(port_priv->ethsw_data->mc_io, 0,
-+ port_priv->ethsw_data->dpsw_handle,
-+ port_priv->idx, &state);
-+ if (err) {
-+ netdev_err(netdev, "dpsw_if_get_link_state() err %d\n", err);
-+ return err;
-+ }
-+
-+ WARN_ONCE(state.up > 1, "Garbage read into link_state");
-+
-+ if (state.up != port_priv->link_state) {
-+ if (state.up)
-+ netif_carrier_on(netdev);
-+ else
-+ netif_carrier_off(netdev);
-+ port_priv->link_state = state.up;
-+ }
-+ return 0;
-+}
-+
-+static int port_open(struct net_device *netdev)
-+{
-+ struct ethsw_port_priv *port_priv = netdev_priv(netdev);
-+ int err;
-+
-+ /* No need to allow Tx as control interface is disabled */
-+ netif_tx_stop_all_queues(netdev);
-+
-+ err = dpsw_if_enable(port_priv->ethsw_data->mc_io, 0,
-+ port_priv->ethsw_data->dpsw_handle,
-+ port_priv->idx);
-+ if (err) {
-+ netdev_err(netdev, "dpsw_if_enable err %d\n", err);
-+ return err;
-+ }
-+
-+ /* sync carrier state */
-+ err = port_carrier_state_sync(netdev);
-+ if (err) {
-+ netdev_err(netdev,
-+ "port_carrier_state_sync err %d\n", err);
-+ goto err_carrier_sync;
-+ }
-+
-+ return 0;
-+
-+err_carrier_sync:
-+ dpsw_if_disable(port_priv->ethsw_data->mc_io, 0,
-+ port_priv->ethsw_data->dpsw_handle,
-+ port_priv->idx);
-+ return err;
-+}
-+
-+static int port_stop(struct net_device *netdev)
-+{
-+ struct ethsw_port_priv *port_priv = netdev_priv(netdev);
-+ int err;
-+
-+ err = dpsw_if_disable(port_priv->ethsw_data->mc_io, 0,
-+ port_priv->ethsw_data->dpsw_handle,
-+ port_priv->idx);
-+ if (err) {
-+ netdev_err(netdev, "dpsw_if_disable err %d\n", err);
-+ return err;
-+ }
-+
-+ return 0;
-+}
-+
-+static netdev_tx_t port_dropframe(struct sk_buff *skb,
-+ struct net_device *netdev)
-+{
-+ /* we don't support I/O for now, drop the frame */
-+ dev_kfree_skb_any(skb);
-+
-+ return NETDEV_TX_OK;
-+}
-+
-+static const struct net_device_ops ethsw_port_ops = {
-+ .ndo_open = port_open,
-+ .ndo_stop = port_stop,
-+
-+ .ndo_set_mac_address = eth_mac_addr,
-+ .ndo_change_mtu = port_change_mtu,
-+ .ndo_has_offload_stats = port_has_offload_stats,
-+ .ndo_get_offload_stats = port_get_offload_stats,
-+
-+ .ndo_start_xmit = port_dropframe,
-+};
-+
-+static void ethsw_links_state_update(struct ethsw_core *ethsw)
-+{
-+ int i;
-+
-+ for (i = 0; i < ethsw->sw_attr.num_ifs; i++)
-+ port_carrier_state_sync(ethsw->ports[i]->netdev);
-+}
-+
-+static irqreturn_t ethsw_irq0_handler(int irq_num, void *arg)
-+{
-+ return IRQ_WAKE_THREAD;
-+}
-+
-+static irqreturn_t ethsw_irq0_handler_thread(int irq_num, void *arg)
-+{
-+ struct device *dev = (struct device *)arg;
-+ struct ethsw_core *ethsw = dev_get_drvdata(dev);
-+
-+ /* Mask the events and the if_id reserved bits to be cleared on read */
-+ u32 status = DPSW_IRQ_EVENT_LINK_CHANGED | 0xFFFF0000;
-+ int err;
-+
-+ err = dpsw_get_irq_status(ethsw->mc_io, 0, ethsw->dpsw_handle,
-+ DPSW_IRQ_INDEX_IF, &status);
-+ if (err) {
-+ dev_err(dev, "Can't get irq status (err %d)", err);
-+
-+ err = dpsw_clear_irq_status(ethsw->mc_io, 0, ethsw->dpsw_handle,
-+ DPSW_IRQ_INDEX_IF, 0xFFFFFFFF);
-+ if (err)
-+ dev_err(dev, "Can't clear irq status (err %d)", err);
-+ goto out;
-+ }
-+
-+ if (status & DPSW_IRQ_EVENT_LINK_CHANGED)
-+ ethsw_links_state_update(ethsw);
-+
-+out:
-+ return IRQ_HANDLED;
-+}
-+
-+static int ethsw_setup_irqs(struct fsl_mc_device *sw_dev)
-+{
-+ struct device *dev = &sw_dev->dev;
-+ struct ethsw_core *ethsw = dev_get_drvdata(dev);
-+ u32 mask = DPSW_IRQ_EVENT_LINK_CHANGED;
-+ struct fsl_mc_device_irq *irq;
-+ int err;
-+
-+ err = fsl_mc_allocate_irqs(sw_dev);
-+ if (err) {
-+ dev_err(dev, "MC irqs allocation failed\n");
-+ return err;
-+ }
-+
-+ if (WARN_ON(sw_dev->obj_desc.irq_count != DPSW_IRQ_NUM)) {
-+ err = -EINVAL;
-+ goto free_irq;
-+ }
-+
-+ err = dpsw_set_irq_enable(ethsw->mc_io, 0, ethsw->dpsw_handle,
-+ DPSW_IRQ_INDEX_IF, 0);
-+ if (err) {
-+ dev_err(dev, "dpsw_set_irq_enable err %d\n", err);
-+ goto free_irq;
-+ }
-+
-+ irq = sw_dev->irqs[DPSW_IRQ_INDEX_IF];
-+
-+ err = devm_request_threaded_irq(dev, irq->msi_desc->irq,
-+ ethsw_irq0_handler,
-+ ethsw_irq0_handler_thread,
-+ IRQF_NO_SUSPEND | IRQF_ONESHOT,
-+ dev_name(dev), dev);
-+ if (err) {
-+ dev_err(dev, "devm_request_threaded_irq(): %d", err);
-+ goto free_irq;
-+ }
-+
-+ err = dpsw_set_irq_mask(ethsw->mc_io, 0, ethsw->dpsw_handle,
-+ DPSW_IRQ_INDEX_IF, mask);
-+ if (err) {
-+ dev_err(dev, "dpsw_set_irq_mask(): %d", err);
-+ goto free_devm_irq;
-+ }
-+
-+ err = dpsw_set_irq_enable(ethsw->mc_io, 0, ethsw->dpsw_handle,
-+ DPSW_IRQ_INDEX_IF, 1);
-+ if (err) {
-+ dev_err(dev, "dpsw_set_irq_enable(): %d", err);
-+ goto free_devm_irq;
-+ }
-+
-+ return 0;
-+
-+free_devm_irq:
-+ devm_free_irq(dev, irq->msi_desc->irq, dev);
-+free_irq:
-+ fsl_mc_free_irqs(sw_dev);
-+ return err;
-+}
-+
-+static void ethsw_teardown_irqs(struct fsl_mc_device *sw_dev)
-+{
-+ struct device *dev = &sw_dev->dev;
-+ struct ethsw_core *ethsw = dev_get_drvdata(dev);
-+ struct fsl_mc_device_irq *irq;
-+ int err;
-+
-+ irq = sw_dev->irqs[DPSW_IRQ_INDEX_IF];
-+ err = dpsw_set_irq_enable(ethsw->mc_io, 0, ethsw->dpsw_handle,
-+ DPSW_IRQ_INDEX_IF, 0);
-+ if (err)
-+ dev_err(dev, "dpsw_set_irq_enable err %d\n", err);
-+
-+ fsl_mc_free_irqs(sw_dev);
-+}
-+
-+static int swdev_port_attr_get(struct net_device *netdev,
-+ struct switchdev_attr *attr)
-+{
-+ struct ethsw_port_priv *port_priv = netdev_priv(netdev);
-+
-+ switch (attr->id) {
-+ case SWITCHDEV_ATTR_ID_PORT_PARENT_ID:
-+ attr->u.ppid.id_len = 1;
-+ attr->u.ppid.id[0] = port_priv->ethsw_data->dev_id;
-+ break;
-+ case SWITCHDEV_ATTR_ID_PORT_BRIDGE_FLAGS:
-+ attr->u.brport_flags =
-+ (port_priv->ethsw_data->learning ? BR_LEARNING : 0) |
-+ (port_priv->flood ? BR_FLOOD : 0);
-+ break;
-+ case SWITCHDEV_ATTR_ID_PORT_BRIDGE_FLAGS_SUPPORT:
-+ attr->u.brport_flags_support = BR_LEARNING | BR_FLOOD;
-+ break;
-+ default:
-+ return -EOPNOTSUPP;
-+ }
-+
-+ return 0;
-+}
-+
-+static int port_attr_stp_state_set(struct net_device *netdev,
-+ struct switchdev_trans *trans,
-+ u8 state)
-+{
-+ struct ethsw_port_priv *port_priv = netdev_priv(netdev);
-+
-+ if (switchdev_trans_ph_prepare(trans))
-+ return 0;
-+
-+ return ethsw_port_set_stp_state(port_priv, state);
-+}
-+
-+static int port_attr_br_flags_set(struct net_device *netdev,
-+ struct switchdev_trans *trans,
-+ unsigned long flags)
-+{
-+ struct ethsw_port_priv *port_priv = netdev_priv(netdev);
-+ int err = 0;
-+
-+ if (switchdev_trans_ph_prepare(trans))
-+ return 0;
-+
-+ /* Learning is enabled per switch */
-+ err = ethsw_set_learning(port_priv->ethsw_data, !!(flags & BR_LEARNING));
-+ if (err)
-+ goto exit;
-+
-+ err = ethsw_port_set_flood(port_priv, !!(flags & BR_FLOOD));
-+
-+exit:
-+ return err;
-+}
-+
-+static int swdev_port_attr_set(struct net_device *netdev,
-+ const struct switchdev_attr *attr,
-+ struct switchdev_trans *trans)
-+{
-+ int err = 0;
-+
-+ switch (attr->id) {
-+ case SWITCHDEV_ATTR_ID_PORT_STP_STATE:
-+ err = port_attr_stp_state_set(netdev, trans,
-+ attr->u.stp_state);
-+ break;
-+ case SWITCHDEV_ATTR_ID_PORT_BRIDGE_FLAGS:
-+ err = port_attr_br_flags_set(netdev, trans,
-+ attr->u.brport_flags);
-+ break;
-+ case SWITCHDEV_ATTR_ID_BRIDGE_VLAN_FILTERING:
-+ /* VLANs are supported by default */
-+ break;
-+ default:
-+ err = -EOPNOTSUPP;
-+ break;
-+ }
-+
-+ return err;
-+}
-+
-+static int port_vlans_add(struct net_device *netdev,
-+ const struct switchdev_obj_port_vlan *vlan,
-+ struct switchdev_trans *trans)
-+{
-+ struct ethsw_port_priv *port_priv = netdev_priv(netdev);
-+ int vid, err;
-+
-+ if (switchdev_trans_ph_prepare(trans))
-+ return 0;
-+
-+ for (vid = vlan->vid_begin; vid <= vlan->vid_end; vid++) {
-+ if (!port_priv->ethsw_data->vlans[vid]) {
-+ /* this is a new VLAN */
-+ err = ethsw_add_vlan(port_priv->ethsw_data, vid);
-+ if (err)
-+ return err;
-+
-+ port_priv->ethsw_data->vlans[vid] |= ETHSW_VLAN_GLOBAL;
-+ }
-+ err = ethsw_port_add_vlan(port_priv, vid, vlan->flags);
-+ if (err)
-+ break;
-+ }
-+
-+ return err;
-+}
-+
-+static int swdev_port_obj_add(struct net_device *netdev,
-+ const struct switchdev_obj *obj,
-+ struct switchdev_trans *trans)
-+{
-+ int err;
-+
-+ switch (obj->id) {
-+ case SWITCHDEV_OBJ_ID_PORT_VLAN:
-+ err = port_vlans_add(netdev,
-+ SWITCHDEV_OBJ_PORT_VLAN(obj),
-+ trans);
-+ break;
-+ default:
-+ err = -EOPNOTSUPP;
-+ break;
-+ }
-+
-+ return err;
-+}
-+
-+static int ethsw_port_del_vlan(struct ethsw_port_priv *port_priv, u16 vid)
-+{
-+ struct ethsw_core *ethsw = port_priv->ethsw_data;
-+ struct net_device *netdev = port_priv->netdev;
-+ struct dpsw_vlan_if_cfg vcfg;
-+ int i, err;
-+
-+ if (!port_priv->vlans[vid])
-+ return -ENOENT;
-+
-+ if (port_priv->vlans[vid] & ETHSW_VLAN_PVID) {
-+ err = ethsw_port_set_pvid(port_priv, 0);
-+ if (err)
-+ return err;
-+ }
-+
-+ vcfg.num_ifs = 1;
-+ vcfg.if_id[0] = port_priv->idx;
-+ if (port_priv->vlans[vid] & ETHSW_VLAN_UNTAGGED) {
-+ err = dpsw_vlan_remove_if_untagged(ethsw->mc_io, 0,
-+ ethsw->dpsw_handle,
-+ vid, &vcfg);
-+ if (err) {
-+ netdev_err(netdev,
-+ "dpsw_vlan_remove_if_untagged err %d\n",
-+ err);
-+ }
-+ port_priv->vlans[vid] &= ~ETHSW_VLAN_UNTAGGED;
-+ }
-+
-+ if (port_priv->vlans[vid] & ETHSW_VLAN_MEMBER) {
-+ err = dpsw_vlan_remove_if(ethsw->mc_io, 0, ethsw->dpsw_handle,
-+ vid, &vcfg);
-+ if (err) {
-+ netdev_err(netdev,
-+ "dpsw_vlan_remove_if err %d\n", err);
-+ return err;
-+ }
-+ port_priv->vlans[vid] &= ~ETHSW_VLAN_MEMBER;
-+
-+ /* Delete VLAN from switch if it is no longer configured on
-+ * any port
-+ */
-+ for (i = 0; i < ethsw->sw_attr.num_ifs; i++)
-+ if (ethsw->ports[i]->vlans[vid] & ETHSW_VLAN_MEMBER)
-+ return 0; /* Found a port member in VID */
-+
-+ ethsw->vlans[vid] &= ~ETHSW_VLAN_GLOBAL;
-+
-+ err = ethsw_dellink_switch(ethsw, vid);
-+ if (err)
-+ return err;
-+ }
-+
-+ return 0;
-+}
-+
-+static int port_vlans_del(struct net_device *netdev,
-+ const struct switchdev_obj_port_vlan *vlan)
-+{
-+ struct ethsw_port_priv *port_priv = netdev_priv(netdev);
-+ int vid, err;
-+
-+ for (vid = vlan->vid_begin; vid <= vlan->vid_end; vid++) {
-+ err = ethsw_port_del_vlan(port_priv, vid);
-+ if (err)
-+ break;
-+ }
-+
-+ return err;
-+}
-+
-+static int swdev_port_obj_del(struct net_device *netdev,
-+ const struct switchdev_obj *obj)
-+{
-+ int err;
-+
-+ switch (obj->id) {
-+ case SWITCHDEV_OBJ_ID_PORT_VLAN:
-+ err = port_vlans_del(netdev, SWITCHDEV_OBJ_PORT_VLAN(obj));
-+ break;
-+ default:
-+ err = -EOPNOTSUPP;
-+ break;
-+ }
-+ return err;
-+}
-+
-+static const struct switchdev_ops ethsw_port_switchdev_ops = {
-+ .switchdev_port_attr_get = swdev_port_attr_get,
-+ .switchdev_port_attr_set = swdev_port_attr_set,
-+ .switchdev_port_obj_add = swdev_port_obj_add,
-+ .switchdev_port_obj_del = swdev_port_obj_del,
-+};
-+
-+/* For the moment, only flood setting needs to be updated */
-+static int port_bridge_join(struct net_device *netdev)
-+{
-+ struct ethsw_port_priv *port_priv = netdev_priv(netdev);
-+
-+ /* Enable flooding */
-+ return ethsw_port_set_flood(port_priv, 1);
-+}
-+
-+static int port_bridge_leave(struct net_device *netdev)
-+{
-+ struct ethsw_port_priv *port_priv = netdev_priv(netdev);
-+
-+ /* Disable flooding */
-+ return ethsw_port_set_flood(port_priv, 0);
-+}
-+
-+static int port_netdevice_event(struct notifier_block *unused,
-+ unsigned long event, void *ptr)
-+{
-+ struct net_device *netdev = netdev_notifier_info_to_dev(ptr);
-+ struct netdev_notifier_changeupper_info *info = ptr;
-+ struct net_device *upper_dev;
-+ int err = 0;
-+
-+ if (netdev->netdev_ops != &ethsw_port_ops)
-+ return NOTIFY_DONE;
-+
-+ /* Handle just upper dev link/unlink for the moment */
-+ if (event == NETDEV_CHANGEUPPER) {
-+ upper_dev = info->upper_dev;
-+ if (netif_is_bridge_master(upper_dev)) {
-+ if (info->linking)
-+ err = port_bridge_join(netdev);
-+ else
-+ err = port_bridge_leave(netdev);
-+ }
-+ }
-+
-+ return notifier_from_errno(err);
-+}
-+
-+static struct notifier_block port_nb __read_mostly = {
-+ .notifier_call = port_netdevice_event,
-+};
-+
-+struct ethsw_switchdev_event_work {
-+ struct work_struct work;
-+ struct switchdev_notifier_fdb_info fdb_info;
-+ struct net_device *dev;
-+ unsigned long event;
-+};
-+
-+static void ethsw_switchdev_event_work(struct work_struct *work)
-+{
-+ struct ethsw_switchdev_event_work *switchdev_work =
-+ container_of(work, struct ethsw_switchdev_event_work, work);
-+ struct net_device *dev = switchdev_work->dev;
-+ struct switchdev_notifier_fdb_info *fdb_info;
-+ struct ethsw_port_priv *port_priv;
-+
-+ rtnl_lock();
-+ port_priv = netdev_priv(dev);
-+ fdb_info = &switchdev_work->fdb_info;
-+
-+ switch (switchdev_work->event) {
-+ case SWITCHDEV_FDB_ADD_TO_DEVICE:
-+ if (is_unicast_ether_addr(fdb_info->addr))
-+ ethsw_port_fdb_add_uc(netdev_priv(dev), fdb_info->addr);
-+ else
-+ ethsw_port_fdb_add_mc(netdev_priv(dev), fdb_info->addr);
-+ break;
-+ case SWITCHDEV_FDB_DEL_TO_DEVICE:
-+ if (is_unicast_ether_addr(fdb_info->addr))
-+ ethsw_port_fdb_del_uc(netdev_priv(dev), fdb_info->addr);
-+ else
-+ ethsw_port_fdb_del_mc(netdev_priv(dev), fdb_info->addr);
-+ break;
-+ }
-+
-+ rtnl_unlock();
-+ kfree(switchdev_work->fdb_info.addr);
-+ kfree(switchdev_work);
-+ dev_put(dev);
-+}
-+
-+/* Called under rcu_read_lock() */
-+static int port_switchdev_event(struct notifier_block *unused,
-+ unsigned long event, void *ptr)
-+{
-+ struct net_device *dev = switchdev_notifier_info_to_dev(ptr);
-+ struct ethsw_switchdev_event_work *switchdev_work;
-+ struct switchdev_notifier_fdb_info *fdb_info = ptr;
-+
-+ switchdev_work = kzalloc(sizeof(*switchdev_work), GFP_ATOMIC);
-+ if (!switchdev_work)
-+ return NOTIFY_BAD;
-+
-+ INIT_WORK(&switchdev_work->work, ethsw_switchdev_event_work);
-+ switchdev_work->dev = dev;
-+ switchdev_work->event = event;
-+
-+ switch (event) {
-+ case SWITCHDEV_FDB_ADD_TO_DEVICE:
-+ case SWITCHDEV_FDB_DEL_TO_DEVICE:
-+ memcpy(&switchdev_work->fdb_info, ptr,
-+ sizeof(switchdev_work->fdb_info));
-+ switchdev_work->fdb_info.addr = kzalloc(ETH_ALEN, GFP_ATOMIC);
-+ if (!switchdev_work->fdb_info.addr)
-+ goto err_addr_alloc;
-+
-+ ether_addr_copy((u8 *)switchdev_work->fdb_info.addr,
-+ fdb_info->addr);
-+
-+ /* Take a reference on the device to avoid being freed. */
-+ dev_hold(dev);
-+ break;
-+ default:
-+ return NOTIFY_DONE;
-+ }
-+
-+ queue_work(ethsw_owq, &switchdev_work->work);
-+
-+ return NOTIFY_DONE;
-+
-+err_addr_alloc:
-+ kfree(switchdev_work);
-+ return NOTIFY_BAD;
-+}
-+
-+static struct notifier_block port_switchdev_nb = {
-+ .notifier_call = port_switchdev_event,
-+};
-+
-+static int ethsw_register_notifier(struct device *dev)
-+{
-+ int err;
-+
-+ err = register_netdevice_notifier(&port_nb);
-+ if (err) {
-+ dev_err(dev, "Failed to register netdev notifier\n");
-+ return err;
-+ }
-+
-+ err = register_switchdev_notifier(&port_switchdev_nb);
-+ if (err) {
-+ dev_err(dev, "Failed to register switchdev notifier\n");
-+ goto err_switchdev_nb;
-+ }
-+
-+ return 0;
-+
-+err_switchdev_nb:
-+ unregister_netdevice_notifier(&port_nb);
-+ return err;
-+}
-+
-+static int ethsw_open(struct ethsw_core *ethsw)
-+{
-+ struct ethsw_port_priv *port_priv = NULL;
-+ int i, err;
-+
-+ err = dpsw_enable(ethsw->mc_io, 0, ethsw->dpsw_handle);
-+ if (err) {
-+ dev_err(ethsw->dev, "dpsw_enable err %d\n", err);
-+ return err;
-+ }
-+
-+ for (i = 0; i < ethsw->sw_attr.num_ifs; i++) {
-+ port_priv = ethsw->ports[i];
-+ err = dev_open(port_priv->netdev);
-+ if (err) {
-+ netdev_err(port_priv->netdev, "dev_open err %d\n", err);
-+ return err;
-+ }
-+ }
-+
-+ return 0;
-+}
-+
-+static int ethsw_stop(struct ethsw_core *ethsw)
-+{
-+ struct ethsw_port_priv *port_priv = NULL;
-+ int i, err;
-+
-+ for (i = 0; i < ethsw->sw_attr.num_ifs; i++) {
-+ port_priv = ethsw->ports[i];
-+ dev_close(port_priv->netdev);
-+ }
-+
-+ err = dpsw_disable(ethsw->mc_io, 0, ethsw->dpsw_handle);
-+ if (err) {
-+ dev_err(ethsw->dev, "dpsw_disable err %d\n", err);
-+ return err;
-+ }
-+
-+ return 0;
-+}
-+
-+static int ethsw_init(struct fsl_mc_device *sw_dev)
-+{
-+ struct device *dev = &sw_dev->dev;
-+ struct ethsw_core *ethsw = dev_get_drvdata(dev);
-+ u16 version_major, version_minor, i;
-+ struct dpsw_stp_cfg stp_cfg;
-+ int err;
-+
-+ ethsw->dev_id = sw_dev->obj_desc.id;
-+
-+ err = dpsw_open(ethsw->mc_io, 0, ethsw->dev_id, &ethsw->dpsw_handle);
-+ if (err) {
-+ dev_err(dev, "dpsw_open err %d\n", err);
-+ return err;
-+ }
-+
-+ err = dpsw_get_attributes(ethsw->mc_io, 0, ethsw->dpsw_handle,
-+ &ethsw->sw_attr);
-+ if (err) {
-+ dev_err(dev, "dpsw_get_attributes err %d\n", err);
-+ goto err_close;
-+ }
-+
-+ err = dpsw_get_api_version(ethsw->mc_io, 0,
-+ &version_major,
-+ &version_minor);
-+ if (err) {
-+ dev_err(dev, "dpsw_get_api_version err %d\n", err);
-+ goto err_close;
-+ }
-+
-+ /* Minimum supported DPSW version check */
-+ if (version_major < DPSW_MIN_VER_MAJOR ||
-+ (version_major == DPSW_MIN_VER_MAJOR &&
-+ version_minor < DPSW_MIN_VER_MINOR)) {
-+ dev_err(dev, "DPSW version %d:%d not supported. Use %d.%d or greater.\n",
-+ version_major,
-+ version_minor,
-+ DPSW_MIN_VER_MAJOR, DPSW_MIN_VER_MINOR);
-+ err = -ENOTSUPP;
-+ goto err_close;
-+ }
-+
-+ err = dpsw_reset(ethsw->mc_io, 0, ethsw->dpsw_handle);
-+ if (err) {
-+ dev_err(dev, "dpsw_reset err %d\n", err);
-+ goto err_close;
-+ }
-+
-+ err = dpsw_fdb_set_learning_mode(ethsw->mc_io, 0, ethsw->dpsw_handle, 0,
-+ DPSW_FDB_LEARNING_MODE_HW);
-+ if (err) {
-+ dev_err(dev, "dpsw_fdb_set_learning_mode err %d\n", err);
-+ goto err_close;
-+ }
-+
-+ stp_cfg.vlan_id = DEFAULT_VLAN_ID;
-+ stp_cfg.state = DPSW_STP_STATE_FORWARDING;
-+
-+ for (i = 0; i < ethsw->sw_attr.num_ifs; i++) {
-+ err = dpsw_if_set_stp(ethsw->mc_io, 0, ethsw->dpsw_handle, i,
-+ &stp_cfg);
-+ if (err) {
-+ dev_err(dev, "dpsw_if_set_stp err %d for port %d\n",
-+ err, i);
-+ goto err_close;
-+ }
-+
-+ err = dpsw_if_set_broadcast(ethsw->mc_io, 0,
-+ ethsw->dpsw_handle, i, 1);
-+ if (err) {
-+ dev_err(dev,
-+ "dpsw_if_set_broadcast err %d for port %d\n",
-+ err, i);
-+ goto err_close;
-+ }
-+ }
-+
-+ ethsw_owq = alloc_ordered_workqueue("%s_ordered", WQ_MEM_RECLAIM,
-+ "ethsw");
-+ if (!ethsw_owq) {
-+ err = -ENOMEM;
-+ goto err_close;
-+ }
-+
-+ err = ethsw_register_notifier(dev);
-+ if (err)
-+ goto err_destroy_ordered_workqueue;
-+
-+ return 0;
-+
-+err_destroy_ordered_workqueue:
-+ destroy_workqueue(ethsw_owq);
-+
-+err_close:
-+ dpsw_close(ethsw->mc_io, 0, ethsw->dpsw_handle);
-+ return err;
-+}
-+
-+static int ethsw_port_init(struct ethsw_port_priv *port_priv, u16 port)
-+{
-+ const char def_mcast[ETH_ALEN] = {0x01, 0x00, 0x5e, 0x00, 0x00, 0x01};
-+ struct net_device *netdev = port_priv->netdev;
-+ struct ethsw_core *ethsw = port_priv->ethsw_data;
-+ struct dpsw_vlan_if_cfg vcfg;
-+ int err;
-+
-+ /* Switch starts with all ports configured to VLAN 1. Need to
-+ * remove this setting to allow configuration at bridge join
-+ */
-+ vcfg.num_ifs = 1;
-+ vcfg.if_id[0] = port_priv->idx;
-+
-+ err = dpsw_vlan_remove_if_untagged(ethsw->mc_io, 0, ethsw->dpsw_handle,
-+ DEFAULT_VLAN_ID, &vcfg);
-+ if (err) {
-+ netdev_err(netdev, "dpsw_vlan_remove_if_untagged err %d\n",
-+ err);
-+ return err;
-+ }
-+
-+ err = ethsw_port_set_pvid(port_priv, 0);
-+ if (err)
-+ return err;
-+
-+ err = dpsw_vlan_remove_if(ethsw->mc_io, 0, ethsw->dpsw_handle,
-+ DEFAULT_VLAN_ID, &vcfg);
-+ if (err) {
-+ netdev_err(netdev, "dpsw_vlan_remove_if err %d\n", err);
-+ return err;
-+ }
-+
-+ err = ethsw_port_fdb_add_mc(port_priv, def_mcast);
-+
-+ return err;
-+}
-+
-+static void ethsw_unregister_notifier(struct device *dev)
-+{
-+ int err;
-+
-+ err = unregister_switchdev_notifier(&port_switchdev_nb);
-+ if (err)
-+ dev_err(dev,
-+ "Failed to unregister switchdev notifier (%d)\n", err);
-+
-+ err = unregister_netdevice_notifier(&port_nb);
-+ if (err)
-+ dev_err(dev,
-+ "Failed to unregister netdev notifier (%d)\n", err);
-+}
-+
-+static void ethsw_takedown(struct fsl_mc_device *sw_dev)
-+{
-+ struct device *dev = &sw_dev->dev;
-+ struct ethsw_core *ethsw = dev_get_drvdata(dev);
-+ int err;
-+
-+ ethsw_unregister_notifier(dev);
-+
-+ err = dpsw_close(ethsw->mc_io, 0, ethsw->dpsw_handle);
-+ if (err)
-+ dev_warn(dev, "dpsw_close err %d\n", err);
-+}
-+
-+static int ethsw_remove(struct fsl_mc_device *sw_dev)
-+{
-+ struct ethsw_port_priv *port_priv;
-+ struct ethsw_core *ethsw;
-+ struct device *dev;
-+ int i;
-+
-+ dev = &sw_dev->dev;
-+ ethsw = dev_get_drvdata(dev);
-+
-+ ethsw_teardown_irqs(sw_dev);
-+
-+ destroy_workqueue(ethsw_owq);
-+
-+ rtnl_lock();
-+ ethsw_stop(ethsw);
-+ rtnl_unlock();
-+
-+ for (i = 0; i < ethsw->sw_attr.num_ifs; i++) {
-+ port_priv = ethsw->ports[i];
-+ unregister_netdev(port_priv->netdev);
-+ free_netdev(port_priv->netdev);
-+ }
-+ kfree(ethsw->ports);
-+
-+ ethsw_takedown(sw_dev);
-+ fsl_mc_portal_free(ethsw->mc_io);
-+
-+ kfree(ethsw);
-+
-+ dev_set_drvdata(dev, NULL);
-+
-+ return 0;
-+}
-+
-+static int ethsw_probe_port(struct ethsw_core *ethsw, u16 port_idx)
-+{
-+ struct ethsw_port_priv *port_priv;
-+ struct device *dev = ethsw->dev;
-+ struct net_device *port_netdev;
-+ int err;
-+
-+ port_netdev = alloc_etherdev(sizeof(struct ethsw_port_priv));
-+ if (!port_netdev) {
-+ dev_err(dev, "alloc_etherdev error\n");
-+ return -ENOMEM;
-+ }
-+
-+ port_priv = netdev_priv(port_netdev);
-+ port_priv->netdev = port_netdev;
-+ port_priv->ethsw_data = ethsw;
-+
-+ port_priv->idx = port_idx;
-+ port_priv->stp_state = BR_STATE_FORWARDING;
-+
-+ /* Flooding is implicitly enabled */
-+ port_priv->flood = true;
-+
-+ SET_NETDEV_DEV(port_netdev, dev);
-+ port_netdev->netdev_ops = &ethsw_port_ops;
-+ port_netdev->ethtool_ops = &ethsw_port_ethtool_ops;
-+ port_netdev->switchdev_ops = &ethsw_port_switchdev_ops;
-+
-+ /* Set MTU limits */
-+ port_netdev->min_mtu = ETH_MIN_MTU;
-+ port_netdev->max_mtu = ETHSW_MAX_FRAME_LENGTH;
-+
-+ err = register_netdev(port_netdev);
-+ if (err < 0) {
-+ dev_err(dev, "register_netdev error %d\n", err);
-+ free_netdev(port_netdev);
-+ return err;
-+ }
-+
-+ ethsw->ports[port_idx] = port_priv;
-+
-+ return ethsw_port_init(port_priv, port_idx);
-+}
-+
-+static int ethsw_probe(struct fsl_mc_device *sw_dev)
-+{
-+ struct device *dev = &sw_dev->dev;
-+ struct ethsw_core *ethsw;
-+ int i, err;
-+
-+ /* Allocate switch core*/
-+ ethsw = kzalloc(sizeof(*ethsw), GFP_KERNEL);
-+
-+ if (!ethsw)
-+ return -ENOMEM;
-+
-+ ethsw->dev = dev;
-+ dev_set_drvdata(dev, ethsw);
-+
-+ err = fsl_mc_portal_allocate(sw_dev, 0, &ethsw->mc_io);
-+ if (err) {
-+ dev_err(dev, "fsl_mc_portal_allocate err %d\n", err);
-+ goto err_free_drvdata;
-+ }
-+
-+ err = ethsw_init(sw_dev);
-+ if (err)
-+ goto err_free_cmdport;
-+
-+ /* DEFAULT_VLAN_ID is implicitly configured on the switch */
-+ ethsw->vlans[DEFAULT_VLAN_ID] = ETHSW_VLAN_MEMBER;
-+
-+ /* Learning is implicitly enabled */
-+ ethsw->learning = true;
-+
-+ ethsw->ports = kcalloc(ethsw->sw_attr.num_ifs, sizeof(*ethsw->ports),
-+ GFP_KERNEL);
-+ if (!(ethsw->ports)) {
-+ err = -ENOMEM;
-+ goto err_takedown;
-+ }
-+
-+ for (i = 0; i < ethsw->sw_attr.num_ifs; i++) {
-+ err = ethsw_probe_port(ethsw, i);
-+ if (err)
-+ goto err_free_ports;
-+ }
-+
-+ /* Switch starts up enabled */
-+ rtnl_lock();
-+ err = ethsw_open(ethsw);
-+ rtnl_unlock();
-+ if (err)
-+ goto err_free_ports;
-+
-+ /* Setup IRQs */
-+ err = ethsw_setup_irqs(sw_dev);
-+ if (err)
-+ goto err_stop;
-+
-+ dev_info(dev, "probed %d port switch\n", ethsw->sw_attr.num_ifs);
-+ return 0;
-+
-+err_stop:
-+ rtnl_lock();
-+ ethsw_stop(ethsw);
-+ rtnl_unlock();
-+
-+err_free_ports:
-+ /* Cleanup registered ports only */
-+ for (i--; i >= 0; i--) {
-+ unregister_netdev(ethsw->ports[i]->netdev);
-+ free_netdev(ethsw->ports[i]->netdev);
-+ }
-+ kfree(ethsw->ports);
-+
-+err_takedown:
-+ ethsw_takedown(sw_dev);
-+
-+err_free_cmdport:
-+ fsl_mc_portal_free(ethsw->mc_io);
-+
-+err_free_drvdata:
-+ kfree(ethsw);
-+ dev_set_drvdata(dev, NULL);
-+
-+ return err;
-+}
-+
-+static const struct fsl_mc_device_id ethsw_match_id_table[] = {
-+ {
-+ .vendor = FSL_MC_VENDOR_FREESCALE,
-+ .obj_type = "dpsw",
-+ },
-+ { .vendor = 0x0 }
-+};
-+MODULE_DEVICE_TABLE(fslmc, ethsw_match_id_table);
-+
-+static struct fsl_mc_driver eth_sw_drv = {
-+ .driver = {
-+ .name = KBUILD_MODNAME,
-+ .owner = THIS_MODULE,
-+ },
-+ .probe = ethsw_probe,
-+ .remove = ethsw_remove,
-+ .match_id_table = ethsw_match_id_table
-+};
-+
-+module_fsl_mc_driver(eth_sw_drv);
-+
-+MODULE_LICENSE("Dual BSD/GPL");
-+MODULE_DESCRIPTION("DPAA2 Ethernet Switch Driver");
---- /dev/null
-+++ b/drivers/staging/fsl-dpaa2/ethsw/ethsw.h
-@@ -0,0 +1,90 @@
-+/* Copyright 2014-2017 Freescale Semiconductor Inc.
-+ * Copyright 2017 NXP
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of the above-listed copyright holders nor the
-+ * names of any contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
-+ * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
-+ * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
-+ * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE
-+ * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
-+ * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
-+ * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
-+ * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
-+ * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
-+ * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
-+ * POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#ifndef __ETHSW_H
-+#define __ETHSW_H
-+
-+#include <linux/netdevice.h>
-+#include <linux/etherdevice.h>
-+#include <linux/rtnetlink.h>
-+#include <linux/if_vlan.h>
-+#include <uapi/linux/if_bridge.h>
-+#include <net/switchdev.h>
-+#include <linux/if_bridge.h>
-+
-+#include "dpsw.h"
-+
-+/* Number of IRQs supported */
-+#define DPSW_IRQ_NUM 2
-+
-+#define ETHSW_VLAN_MEMBER 1
-+#define ETHSW_VLAN_UNTAGGED 2
-+#define ETHSW_VLAN_PVID 4
-+#define ETHSW_VLAN_GLOBAL 8
-+
-+/* Maximum Frame Length supported by HW (currently 10k) */
-+#define DPAA2_MFL (10 * 1024)
-+#define ETHSW_MAX_FRAME_LENGTH (DPAA2_MFL - VLAN_ETH_HLEN - ETH_FCS_LEN)
-+#define ETHSW_L2_MAX_FRM(mtu) ((mtu) + VLAN_ETH_HLEN + ETH_FCS_LEN)
-+
-+extern const struct ethtool_ops ethsw_port_ethtool_ops;
-+
-+struct ethsw_core;
-+
-+/* Per port private data */
-+struct ethsw_port_priv {
-+ struct net_device *netdev;
-+ u16 idx;
-+ struct ethsw_core *ethsw_data;
-+ u8 link_state;
-+ u8 stp_state;
-+ bool flood;
-+
-+ u8 vlans[VLAN_VID_MASK + 1];
-+ u16 pvid;
-+};
-+
-+/* Switch data */
-+struct ethsw_core {
-+ struct device *dev;
-+ struct fsl_mc_io *mc_io;
-+ u16 dpsw_handle;
-+ struct dpsw_attr sw_attr;
-+ int dev_id;
-+ struct ethsw_port_priv **ports;
-+
-+ u8 vlans[VLAN_VID_MASK + 1];
-+ bool learning;
-+};
-+
-+#endif /* __ETHSW_H */
---- /dev/null
-+++ b/drivers/staging/fsl-dpaa2/evb/Kconfig
-@@ -0,0 +1,7 @@
-+config FSL_DPAA2_EVB
-+ tristate "DPAA2 Edge Virtual Bridge"
-+ depends on FSL_MC_BUS && FSL_DPAA2
-+ select VLAN_8021Q
-+ default y
-+ ---help---
-+ Prototype driver for DPAA2 Edge Virtual Bridge.
---- /dev/null
-+++ b/drivers/staging/fsl-dpaa2/evb/Makefile
-@@ -0,0 +1,10 @@
-+
-+obj-$(CONFIG_FSL_DPAA2_EVB) += dpaa2-evb.o
-+
-+dpaa2-evb-objs := evb.o dpdmux.o
-+
-+all:
-+ make -C /lib/modules/$(shell uname -r)/build M=$(PWD) modules
-+
-+clean:
-+ make -C /lib/modules/$(shell uname -r)/build M=$(PWD) clean
---- /dev/null
-+++ b/drivers/staging/fsl-dpaa2/evb/dpdmux-cmd.h
-@@ -0,0 +1,279 @@
-+/* Copyright 2013-2016 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of the above-listed copyright holders nor the
-+ * names of any contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
-+ * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
-+ * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
-+ * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE
-+ * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
-+ * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
-+ * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
-+ * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
-+ * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
-+ * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
-+ * POSSIBILITY OF SUCH DAMAGE.
-+ */
-+#ifndef _FSL_DPDMUX_CMD_H
-+#define _FSL_DPDMUX_CMD_H
-+
-+/* DPDMUX Version */
-+#define DPDMUX_VER_MAJOR 6
-+#define DPDMUX_VER_MINOR 1
-+
-+#define DPDMUX_CMD_BASE_VER 1
-+#define DPDMUX_CMD_ID_OFFSET 4
-+
-+#define DPDMUX_CMD(id) (((id) << DPDMUX_CMD_ID_OFFSET) | DPDMUX_CMD_BASE_VER)
-+
-+/* Command IDs */
-+#define DPDMUX_CMDID_CLOSE DPDMUX_CMD(0x800)
-+#define DPDMUX_CMDID_OPEN DPDMUX_CMD(0x806)
-+#define DPDMUX_CMDID_CREATE DPDMUX_CMD(0x906)
-+#define DPDMUX_CMDID_DESTROY DPDMUX_CMD(0x986)
-+#define DPDMUX_CMDID_GET_API_VERSION DPDMUX_CMD(0xa06)
-+
-+#define DPDMUX_CMDID_ENABLE DPDMUX_CMD(0x002)
-+#define DPDMUX_CMDID_DISABLE DPDMUX_CMD(0x003)
-+#define DPDMUX_CMDID_GET_ATTR DPDMUX_CMD(0x004)
-+#define DPDMUX_CMDID_RESET DPDMUX_CMD(0x005)
-+#define DPDMUX_CMDID_IS_ENABLED DPDMUX_CMD(0x006)
-+
-+#define DPDMUX_CMDID_SET_IRQ_ENABLE DPDMUX_CMD(0x012)
-+#define DPDMUX_CMDID_GET_IRQ_ENABLE DPDMUX_CMD(0x013)
-+#define DPDMUX_CMDID_SET_IRQ_MASK DPDMUX_CMD(0x014)
-+#define DPDMUX_CMDID_GET_IRQ_MASK DPDMUX_CMD(0x015)
-+#define DPDMUX_CMDID_GET_IRQ_STATUS DPDMUX_CMD(0x016)
-+#define DPDMUX_CMDID_CLEAR_IRQ_STATUS DPDMUX_CMD(0x017)
-+
-+#define DPDMUX_CMDID_SET_MAX_FRAME_LENGTH DPDMUX_CMD(0x0a1)
-+
-+#define DPDMUX_CMDID_UL_RESET_COUNTERS DPDMUX_CMD(0x0a3)
-+
-+#define DPDMUX_CMDID_IF_SET_ACCEPTED_FRAMES DPDMUX_CMD(0x0a7)
-+#define DPDMUX_CMDID_IF_GET_ATTR DPDMUX_CMD(0x0a8)
-+#define DPDMUX_CMDID_IF_ENABLE DPDMUX_CMD(0x0a9)
-+#define DPDMUX_CMDID_IF_DISABLE DPDMUX_CMD(0x0aa)
-+
-+#define DPDMUX_CMDID_IF_ADD_L2_RULE DPDMUX_CMD(0x0b0)
-+#define DPDMUX_CMDID_IF_REMOVE_L2_RULE DPDMUX_CMD(0x0b1)
-+#define DPDMUX_CMDID_IF_GET_COUNTER DPDMUX_CMD(0x0b2)
-+#define DPDMUX_CMDID_IF_SET_LINK_CFG DPDMUX_CMD(0x0b3)
-+#define DPDMUX_CMDID_IF_GET_LINK_STATE DPDMUX_CMD(0x0b4)
-+
-+#define DPDMUX_CMDID_SET_CUSTOM_KEY DPDMUX_CMD(0x0b5)
-+#define DPDMUX_CMDID_ADD_CUSTOM_CLS_ENTRY DPDMUX_CMD(0x0b6)
-+#define DPDMUX_CMDID_REMOVE_CUSTOM_CLS_ENTRY DPDMUX_CMD(0x0b7)
-+
-+#define DPDMUX_MASK(field) \
-+ GENMASK(DPDMUX_##field##_SHIFT + DPDMUX_##field##_SIZE - 1, \
-+ DPDMUX_##field##_SHIFT)
-+#define dpdmux_set_field(var, field, val) \
-+ ((var) |= (((val) << DPDMUX_##field##_SHIFT) & DPDMUX_MASK(field)))
-+#define dpdmux_get_field(var, field) \
-+ (((var) & DPDMUX_MASK(field)) >> DPDMUX_##field##_SHIFT)
-+
-+struct dpdmux_cmd_open {
-+ u32 dpdmux_id;
-+};
-+
-+struct dpdmux_cmd_create {
-+ u8 method;
-+ u8 manip;
-+ u16 num_ifs;
-+ u32 pad;
-+
-+ u16 adv_max_dmat_entries;
-+ u16 adv_max_mc_groups;
-+ u16 adv_max_vlan_ids;
-+ u16 pad1;
-+
-+ u64 options;
-+};
-+
-+struct dpdmux_cmd_destroy {
-+ u32 dpdmux_id;
-+};
-+
-+#define DPDMUX_ENABLE_SHIFT 0
-+#define DPDMUX_ENABLE_SIZE 1
-+
-+struct dpdmux_rsp_is_enabled {
-+ u8 en;
-+};
-+
-+struct dpdmux_cmd_set_irq_enable {
-+ u8 enable;
-+ u8 pad[3];
-+ u8 irq_index;
-+};
-+
-+struct dpdmux_cmd_get_irq_enable {
-+ u32 pad;
-+ u8 irq_index;
-+};
-+
-+struct dpdmux_rsp_get_irq_enable {
-+ u8 enable;
-+};
-+
-+struct dpdmux_cmd_set_irq_mask {
-+ u32 mask;
-+ u8 irq_index;
-+};
-+
-+struct dpdmux_cmd_get_irq_mask {
-+ u32 pad;
-+ u8 irq_index;
-+};
-+
-+struct dpdmux_rsp_get_irq_mask {
-+ u32 mask;
-+};
-+
-+struct dpdmux_cmd_get_irq_status {
-+ u32 status;
-+ u8 irq_index;
-+};
-+
-+struct dpdmux_rsp_get_irq_status {
-+ u32 status;
-+};
-+
-+struct dpdmux_cmd_clear_irq_status {
-+ u32 status;
-+ u8 irq_index;
-+};
-+
-+struct dpdmux_rsp_get_attr {
-+ u8 method;
-+ u8 manip;
-+ u16 num_ifs;
-+ u16 mem_size;
-+ u16 pad;
-+
-+ u64 pad1;
-+
-+ u32 id;
-+ u32 pad2;
-+
-+ u64 options;
-+};
-+
-+struct dpdmux_cmd_set_max_frame_length {
-+ u16 max_frame_length;
-+};
-+
-+#define DPDMUX_ACCEPTED_FRAMES_TYPE_SHIFT 0
-+#define DPDMUX_ACCEPTED_FRAMES_TYPE_SIZE 4
-+#define DPDMUX_UNACCEPTED_FRAMES_ACTION_SHIFT 4
-+#define DPDMUX_UNACCEPTED_FRAMES_ACTION_SIZE 4
-+
-+struct dpdmux_cmd_if_set_accepted_frames {
-+ u16 if_id;
-+ u8 frames_options;
-+};
-+
-+struct dpdmux_cmd_if {
-+ u16 if_id;
-+};
-+
-+struct dpdmux_rsp_if_get_attr {
-+ u8 pad[3];
-+ u8 enabled;
-+ u8 pad1[3];
-+ u8 accepted_frames_type;
-+ u32 rate;
-+};
-+
-+struct dpdmux_cmd_if_l2_rule {
-+ u16 if_id;
-+ u8 mac_addr5;
-+ u8 mac_addr4;
-+ u8 mac_addr3;
-+ u8 mac_addr2;
-+ u8 mac_addr1;
-+ u8 mac_addr0;
-+
-+ u32 pad;
-+ u16 vlan_id;
-+};
-+
-+struct dpdmux_cmd_if_get_counter {
-+ u16 if_id;
-+ u8 counter_type;
-+};
-+
-+struct dpdmux_rsp_if_get_counter {
-+ u64 pad;
-+ u64 counter;
-+};
-+
-+struct dpdmux_cmd_if_set_link_cfg {
-+ u16 if_id;
-+ u16 pad[3];
-+
-+ u32 rate;
-+ u32 pad1;
-+
-+ u64 options;
-+};
-+
-+struct dpdmux_cmd_if_get_link_state {
-+ u16 if_id;
-+};
-+
-+struct dpdmux_rsp_if_get_link_state {
-+ u32 pad;
-+ u8 up;
-+ u8 pad1[3];
-+
-+ u32 rate;
-+ u32 pad2;
-+
-+ u64 options;
-+};
-+
-+struct dpdmux_rsp_get_api_version {
-+ u16 major;
-+ u16 minor;
-+};
-+
-+struct dpdmux_set_custom_key {
-+ u64 pad[6];
-+ u64 key_cfg_iova;
-+};
-+
-+struct dpdmux_cmd_add_custom_cls_entry {
-+ u8 pad[3];
-+ u8 key_size;
-+ u16 pad1;
-+ u16 dest_if;
-+ u64 key_iova;
-+ u64 mask_iova;
-+};
-+
-+struct dpdmux_cmd_remove_custom_cls_entry {
-+ u8 pad[3];
-+ u8 key_size;
-+ u32 pad1;
-+ u64 key_iova;
-+ u64 mask_iova;
-+};
-+
-+#endif /* _FSL_DPDMUX_CMD_H */
---- /dev/null
-+++ b/drivers/staging/fsl-dpaa2/evb/dpdmux.c
-@@ -0,0 +1,1111 @@
-+/* Copyright 2013-2016 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of the above-listed copyright holders nor the
-+ * names of any contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
-+ * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
-+ * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
-+ * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE
-+ * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
-+ * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
-+ * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
-+ * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
-+ * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
-+ * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
-+ * POSSIBILITY OF SUCH DAMAGE.
-+ */
-+#include <linux/fsl/mc.h>
-+#include "dpdmux.h"
-+#include "dpdmux-cmd.h"
-+
-+/**
-+ * dpdmux_open() - Open a control session for the specified object
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @dpdmux_id: DPDMUX unique ID
-+ * @token: Returned token; use in subsequent API calls
-+ *
-+ * This function can be used to open a control session for an
-+ * already created object; an object may have been declared in
-+ * the DPL or by calling the dpdmux_create() function.
-+ * This function returns a unique authentication token,
-+ * associated with the specific object ID and the specific MC
-+ * portal; this token must be used in all subsequent commands for
-+ * this specific object.
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpdmux_open(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ int dpdmux_id,
-+ u16 *token)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpdmux_cmd_open *cmd_params;
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPDMUX_CMDID_OPEN,
-+ cmd_flags,
-+ 0);
-+ cmd_params = (struct dpdmux_cmd_open *)cmd.params;
-+ cmd_params->dpdmux_id = cpu_to_le32(dpdmux_id);
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ *token = mc_cmd_hdr_read_token(&cmd);
-+
-+ return 0;
-+}
-+
-+/**
-+ * dpdmux_close() - Close the control session of the object
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPDMUX object
-+ *
-+ * After this function is called, no further operations are
-+ * allowed on the object without opening a new control session.
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpdmux_close(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPDMUX_CMDID_CLOSE,
-+ cmd_flags,
-+ token);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpdmux_create() - Create the DPDMUX object
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @dprc_token: Parent container token; '0' for default container
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @cfg: Configuration structure
-+ * @obj_id: returned object id
-+ *
-+ * Create the DPDMUX object, allocate required resources and
-+ * perform required initialization.
-+ *
-+ * The object can be created either by declaring it in the
-+ * DPL file, or by calling this function.
-+ *
-+ * The function accepts an authentication token of a parent
-+ * container that this object should be assigned to. The token
-+ * can be '0' so the object will be assigned to the default container.
-+ * The newly created object can be opened with the returned
-+ * object id and using the container's associated tokens and MC portals.
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpdmux_create(struct fsl_mc_io *mc_io,
-+ u16 dprc_token,
-+ u32 cmd_flags,
-+ const struct dpdmux_cfg *cfg,
-+ u32 *obj_id)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpdmux_cmd_create *cmd_params;
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPDMUX_CMDID_CREATE,
-+ cmd_flags,
-+ dprc_token);
-+ cmd_params = (struct dpdmux_cmd_create *)cmd.params;
-+ cmd_params->method = cfg->method;
-+ cmd_params->manip = cfg->manip;
-+ cmd_params->num_ifs = cpu_to_le16(cfg->num_ifs);
-+ cmd_params->adv_max_dmat_entries =
-+ cpu_to_le16(cfg->adv.max_dmat_entries);
-+ cmd_params->adv_max_mc_groups = cpu_to_le16(cfg->adv.max_mc_groups);
-+ cmd_params->adv_max_vlan_ids = cpu_to_le16(cfg->adv.max_vlan_ids);
-+ cmd_params->options = cpu_to_le64(cfg->adv.options);
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ *obj_id = mc_cmd_hdr_read_token(&cmd);
-+
-+ return 0;
-+}
-+
-+/**
-+ * dpdmux_destroy() - Destroy the DPDMUX object and release all its resources.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @dprc_token: Parent container token; '0' for default container
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @object_id: The object id; it must be a valid id within the container that
-+ * created this object;
-+ *
-+ * The function accepts the authentication token of the parent container that
-+ * created the object (not the one that currently owns the object). The object
-+ * is searched within parent using the provided 'object_id'.
-+ * All tokens to the object must be closed before calling destroy.
-+ *
-+ * Return: '0' on Success; error code otherwise.
-+ */
-+int dpdmux_destroy(struct fsl_mc_io *mc_io,
-+ u16 dprc_token,
-+ u32 cmd_flags,
-+ u32 object_id)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpdmux_cmd_destroy *cmd_params;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPDMUX_CMDID_DESTROY,
-+ cmd_flags,
-+ dprc_token);
-+ cmd_params = (struct dpdmux_cmd_destroy *)cmd.params;
-+ cmd_params->dpdmux_id = cpu_to_le32(object_id);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpdmux_enable() - Enable DPDMUX functionality
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPDMUX object
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpdmux_enable(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPDMUX_CMDID_ENABLE,
-+ cmd_flags,
-+ token);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpdmux_disable() - Disable DPDMUX functionality
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPDMUX object
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpdmux_disable(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPDMUX_CMDID_DISABLE,
-+ cmd_flags,
-+ token);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpdmux_is_enabled() - Check if the DPDMUX is enabled.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPDMUX object
-+ * @en: Returns '1' if object is enabled; '0' otherwise
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpdmux_is_enabled(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ int *en)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpdmux_rsp_is_enabled *rsp_params;
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPDMUX_CMDID_IS_ENABLED,
-+ cmd_flags,
-+ token);
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ rsp_params = (struct dpdmux_rsp_is_enabled *)cmd.params;
-+ *en = dpdmux_get_field(rsp_params->en, ENABLE);
-+
-+ return 0;
-+}
-+
-+/**
-+ * dpdmux_reset() - Reset the DPDMUX, returns the object to initial state.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPDMUX object
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpdmux_reset(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPDMUX_CMDID_RESET,
-+ cmd_flags,
-+ token);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpdmux_set_irq_enable() - Set overall interrupt state.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPDMUX object
-+ * @irq_index: The interrupt index to configure
-+ * @en: Interrupt state - enable = 1, disable = 0
-+ *
-+ * Allows GPP software to control when interrupts are generated.
-+ * Each interrupt can have up to 32 causes. The enable/disable control's the
-+ * overall interrupt state. if the interrupt is disabled no causes will cause
-+ * an interrupt.
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpdmux_set_irq_enable(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 irq_index,
-+ u8 en)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpdmux_cmd_set_irq_enable *cmd_params;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPDMUX_CMDID_SET_IRQ_ENABLE,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpdmux_cmd_set_irq_enable *)cmd.params;
-+ cmd_params->enable = en;
-+ cmd_params->irq_index = irq_index;
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpdmux_get_irq_enable() - Get overall interrupt state.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPDMUX object
-+ * @irq_index: The interrupt index to configure
-+ * @en: Returned interrupt state - enable = 1, disable = 0
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpdmux_get_irq_enable(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 irq_index,
-+ u8 *en)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpdmux_cmd_get_irq_enable *cmd_params;
-+ struct dpdmux_rsp_get_irq_enable *rsp_params;
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPDMUX_CMDID_GET_IRQ_ENABLE,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpdmux_cmd_get_irq_enable *)cmd.params;
-+ cmd_params->irq_index = irq_index;
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ rsp_params = (struct dpdmux_rsp_get_irq_enable *)cmd.params;
-+ *en = rsp_params->enable;
-+
-+ return 0;
-+}
-+
-+/**
-+ * dpdmux_set_irq_mask() - Set interrupt mask.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPDMUX object
-+ * @irq_index: The interrupt index to configure
-+ * @mask: event mask to trigger interrupt;
-+ * each bit:
-+ * 0 = ignore event
-+ * 1 = consider event for asserting IRQ
-+ *
-+ * Every interrupt can have up to 32 causes and the interrupt model supports
-+ * masking/unmasking each cause independently
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpdmux_set_irq_mask(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 irq_index,
-+ u32 mask)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpdmux_cmd_set_irq_mask *cmd_params;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPDMUX_CMDID_SET_IRQ_MASK,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpdmux_cmd_set_irq_mask *)cmd.params;
-+ cmd_params->mask = cpu_to_le32(mask);
-+ cmd_params->irq_index = irq_index;
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpdmux_get_irq_mask() - Get interrupt mask.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPDMUX object
-+ * @irq_index: The interrupt index to configure
-+ * @mask: Returned event mask to trigger interrupt
-+ *
-+ * Every interrupt can have up to 32 causes and the interrupt model supports
-+ * masking/unmasking each cause independently
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpdmux_get_irq_mask(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 irq_index,
-+ u32 *mask)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpdmux_cmd_get_irq_mask *cmd_params;
-+ struct dpdmux_rsp_get_irq_mask *rsp_params;
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPDMUX_CMDID_GET_IRQ_MASK,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpdmux_cmd_get_irq_mask *)cmd.params;
-+ cmd_params->irq_index = irq_index;
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ rsp_params = (struct dpdmux_rsp_get_irq_mask *)cmd.params;
-+ *mask = le32_to_cpu(rsp_params->mask);
-+
-+ return 0;
-+}
-+
-+/**
-+ * dpdmux_get_irq_status() - Get the current status of any pending interrupts.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPDMUX object
-+ * @irq_index: The interrupt index to configure
-+ * @status: Returned interrupts status - one bit per cause:
-+ * 0 = no interrupt pending
-+ * 1 = interrupt pending
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpdmux_get_irq_status(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 irq_index,
-+ u32 *status)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpdmux_cmd_get_irq_status *cmd_params;
-+ struct dpdmux_rsp_get_irq_status *rsp_params;
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPDMUX_CMDID_GET_IRQ_STATUS,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpdmux_cmd_get_irq_status *)cmd.params;
-+ cmd_params->status = cpu_to_le32(*status);
-+ cmd_params->irq_index = irq_index;
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ rsp_params = (struct dpdmux_rsp_get_irq_status *)cmd.params;
-+ *status = le32_to_cpu(rsp_params->status);
-+
-+ return 0;
-+}
-+
-+/**
-+ * dpdmux_clear_irq_status() - Clear a pending interrupt's status
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPDMUX object
-+ * @irq_index: The interrupt index to configure
-+ * @status: bits to clear (W1C) - one bit per cause:
-+ * 0 = don't change
-+ * 1 = clear status bit
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpdmux_clear_irq_status(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 irq_index,
-+ u32 status)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpdmux_cmd_clear_irq_status *cmd_params;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPDMUX_CMDID_CLEAR_IRQ_STATUS,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpdmux_cmd_clear_irq_status *)cmd.params;
-+ cmd_params->status = cpu_to_le32(status);
-+ cmd_params->irq_index = irq_index;
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpdmux_get_attributes() - Retrieve DPDMUX attributes
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPDMUX object
-+ * @attr: Returned object's attributes
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpdmux_get_attributes(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ struct dpdmux_attr *attr)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpdmux_rsp_get_attr *rsp_params;
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPDMUX_CMDID_GET_ATTR,
-+ cmd_flags,
-+ token);
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ rsp_params = (struct dpdmux_rsp_get_attr *)cmd.params;
-+ attr->id = le32_to_cpu(rsp_params->id);
-+ attr->options = le64_to_cpu(rsp_params->options);
-+ attr->method = rsp_params->method;
-+ attr->manip = rsp_params->manip;
-+ attr->num_ifs = le16_to_cpu(rsp_params->num_ifs);
-+ attr->mem_size = le16_to_cpu(rsp_params->mem_size);
-+
-+ return 0;
-+}
-+
-+/**
-+ * dpdmux_if_enable() - Enable Interface
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPDMUX object
-+ * @if_id: Interface Identifier
-+ *
-+ * Return: Completion status. '0' on Success; Error code otherwise.
-+ */
-+int dpdmux_if_enable(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 if_id)
-+{
-+ struct dpdmux_cmd_if *cmd_params;
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPDMUX_CMDID_IF_ENABLE,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpdmux_cmd_if *)cmd.params;
-+ cmd_params->if_id = cpu_to_le16(if_id);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpdmux_if_disable() - Disable Interface
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPDMUX object
-+ * @if_id: Interface Identifier
-+ *
-+ * Return: Completion status. '0' on Success; Error code otherwise.
-+ */
-+int dpdmux_if_disable(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 if_id)
-+{
-+ struct dpdmux_cmd_if *cmd_params;
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPDMUX_CMDID_IF_DISABLE,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpdmux_cmd_if *)cmd.params;
-+ cmd_params->if_id = cpu_to_le16(if_id);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpdmux_set_max_frame_length() - Set the maximum frame length in DPDMUX
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPDMUX object
-+ * @max_frame_length: The required maximum frame length
-+ *
-+ * Update the maximum frame length on all DMUX interfaces.
-+ * In case of VEPA, the maximum frame length on all dmux interfaces
-+ * will be updated with the minimum value of the mfls of the connected
-+ * dpnis and the actual value of dmux mfl.
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpdmux_set_max_frame_length(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 max_frame_length)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpdmux_cmd_set_max_frame_length *cmd_params;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPDMUX_CMDID_SET_MAX_FRAME_LENGTH,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpdmux_cmd_set_max_frame_length *)cmd.params;
-+ cmd_params->max_frame_length = cpu_to_le16(max_frame_length);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpdmux_ul_reset_counters() - Function resets the uplink counter
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPDMUX object
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpdmux_ul_reset_counters(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPDMUX_CMDID_UL_RESET_COUNTERS,
-+ cmd_flags,
-+ token);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpdmux_if_set_accepted_frames() - Set the accepted frame types
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPDMUX object
-+ * @if_id: Interface ID (0 for uplink, or 1-num_ifs);
-+ * @cfg: Frame types configuration
-+ *
-+ * if 'DPDMUX_ADMIT_ONLY_VLAN_TAGGED' is set - untagged frames or
-+ * priority-tagged frames are discarded.
-+ * if 'DPDMUX_ADMIT_ONLY_UNTAGGED' is set - untagged frames or
-+ * priority-tagged frames are accepted.
-+ * if 'DPDMUX_ADMIT_ALL' is set (default mode) - all VLAN tagged,
-+ * untagged and priority-tagged frame are accepted;
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpdmux_if_set_accepted_frames(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 if_id,
-+ const struct dpdmux_accepted_frames *cfg)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpdmux_cmd_if_set_accepted_frames *cmd_params;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPDMUX_CMDID_IF_SET_ACCEPTED_FRAMES,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpdmux_cmd_if_set_accepted_frames *)cmd.params;
-+ cmd_params->if_id = cpu_to_le16(if_id);
-+ dpdmux_set_field(cmd_params->frames_options, ACCEPTED_FRAMES_TYPE,
-+ cfg->type);
-+ dpdmux_set_field(cmd_params->frames_options, UNACCEPTED_FRAMES_ACTION,
-+ cfg->unaccept_act);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpdmux_if_get_attributes() - Obtain DPDMUX interface attributes
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPDMUX object
-+ * @if_id: Interface ID (0 for uplink, or 1-num_ifs);
-+ * @attr: Interface attributes
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpdmux_if_get_attributes(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 if_id,
-+ struct dpdmux_if_attr *attr)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpdmux_cmd_if *cmd_params;
-+ struct dpdmux_rsp_if_get_attr *rsp_params;
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPDMUX_CMDID_IF_GET_ATTR,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpdmux_cmd_if *)cmd.params;
-+ cmd_params->if_id = cpu_to_le16(if_id);
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ rsp_params = (struct dpdmux_rsp_if_get_attr *)cmd.params;
-+ attr->rate = le32_to_cpu(rsp_params->rate);
-+ attr->enabled = dpdmux_get_field(rsp_params->enabled, ENABLE);
-+ attr->accept_frame_type =
-+ dpdmux_get_field(rsp_params->accepted_frames_type,
-+ ACCEPTED_FRAMES_TYPE);
-+
-+ return 0;
-+}
-+
-+/**
-+ * dpdmux_if_remove_l2_rule() - Remove L2 rule from DPDMUX table
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPDMUX object
-+ * @if_id: Destination interface ID
-+ * @rule: L2 rule
-+ *
-+ * Function removes a L2 rule from DPDMUX table
-+ * or adds an interface to an existing multicast address
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpdmux_if_remove_l2_rule(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 if_id,
-+ const struct dpdmux_l2_rule *rule)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpdmux_cmd_if_l2_rule *cmd_params;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPDMUX_CMDID_IF_REMOVE_L2_RULE,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpdmux_cmd_if_l2_rule *)cmd.params;
-+ cmd_params->if_id = cpu_to_le16(if_id);
-+ cmd_params->vlan_id = cpu_to_le16(rule->vlan_id);
-+ cmd_params->mac_addr5 = rule->mac_addr[5];
-+ cmd_params->mac_addr4 = rule->mac_addr[4];
-+ cmd_params->mac_addr3 = rule->mac_addr[3];
-+ cmd_params->mac_addr2 = rule->mac_addr[2];
-+ cmd_params->mac_addr1 = rule->mac_addr[1];
-+ cmd_params->mac_addr0 = rule->mac_addr[0];
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpdmux_if_add_l2_rule() - Add L2 rule into DPDMUX table
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPDMUX object
-+ * @if_id: Destination interface ID
-+ * @rule: L2 rule
-+ *
-+ * Function adds a L2 rule into DPDMUX table
-+ * or adds an interface to an existing multicast address
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpdmux_if_add_l2_rule(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 if_id,
-+ const struct dpdmux_l2_rule *rule)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpdmux_cmd_if_l2_rule *cmd_params;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPDMUX_CMDID_IF_ADD_L2_RULE,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpdmux_cmd_if_l2_rule *)cmd.params;
-+ cmd_params->if_id = cpu_to_le16(if_id);
-+ cmd_params->vlan_id = cpu_to_le16(rule->vlan_id);
-+ cmd_params->mac_addr5 = rule->mac_addr[5];
-+ cmd_params->mac_addr4 = rule->mac_addr[4];
-+ cmd_params->mac_addr3 = rule->mac_addr[3];
-+ cmd_params->mac_addr2 = rule->mac_addr[2];
-+ cmd_params->mac_addr1 = rule->mac_addr[1];
-+ cmd_params->mac_addr0 = rule->mac_addr[0];
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpdmux_if_get_counter() - Functions obtains specific counter of an interface
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPDMUX object
-+ * @if_id: Interface Id
-+ * @counter_type: counter type
-+ * @counter: Returned specific counter information
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpdmux_if_get_counter(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 if_id,
-+ enum dpdmux_counter_type counter_type,
-+ u64 *counter)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpdmux_cmd_if_get_counter *cmd_params;
-+ struct dpdmux_rsp_if_get_counter *rsp_params;
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPDMUX_CMDID_IF_GET_COUNTER,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpdmux_cmd_if_get_counter *)cmd.params;
-+ cmd_params->if_id = cpu_to_le16(if_id);
-+ cmd_params->counter_type = counter_type;
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ rsp_params = (struct dpdmux_rsp_if_get_counter *)cmd.params;
-+ *counter = le64_to_cpu(rsp_params->counter);
-+
-+ return 0;
-+}
-+
-+/**
-+ * dpdmux_if_set_link_cfg() - set the link configuration.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPSW object
-+ * @if_id: interface id
-+ * @cfg: Link configuration
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpdmux_if_set_link_cfg(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 if_id,
-+ struct dpdmux_link_cfg *cfg)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpdmux_cmd_if_set_link_cfg *cmd_params;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPDMUX_CMDID_IF_SET_LINK_CFG,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpdmux_cmd_if_set_link_cfg *)cmd.params;
-+ cmd_params->if_id = cpu_to_le16(if_id);
-+ cmd_params->rate = cpu_to_le32(cfg->rate);
-+ cmd_params->options = cpu_to_le64(cfg->options);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpdmux_if_get_link_state - Return the link state
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPSW object
-+ * @if_id: interface id
-+ * @state: link state
-+ *
-+ * @returns '0' on Success; Error code otherwise.
-+ */
-+int dpdmux_if_get_link_state(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 if_id,
-+ struct dpdmux_link_state *state)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpdmux_cmd_if_get_link_state *cmd_params;
-+ struct dpdmux_rsp_if_get_link_state *rsp_params;
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPDMUX_CMDID_IF_GET_LINK_STATE,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpdmux_cmd_if_get_link_state *)cmd.params;
-+ cmd_params->if_id = cpu_to_le16(if_id);
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ rsp_params = (struct dpdmux_rsp_if_get_link_state *)cmd.params;
-+ state->rate = le32_to_cpu(rsp_params->rate);
-+ state->options = le64_to_cpu(rsp_params->options);
-+ state->up = dpdmux_get_field(rsp_params->up, ENABLE);
-+
-+ return 0;
-+}
-+
-+/**
-+ * dpdmux_set_custom_key - Set a custom classification key.
-+ *
-+ * This API is only available for DPDMUX instance created with
-+ * DPDMUX_METHOD_CUSTOM. This API must be called before populating the
-+ * classification table using dpdmux_add_custom_cls_entry.
-+ *
-+ * Calls to dpdmux_set_custom_key remove all existing classification entries
-+ * that may have been added previously using dpdmux_add_custom_cls_entry.
-+ *
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPSW object
-+ * @if_id: interface id
-+ * @key_cfg_iova: DMA address of a configuration structure set up using
-+ * dpkg_prepare_key_cfg. Maximum key size is 24 bytes.
-+ *
-+ * @returns '0' on Success; Error code otherwise.
-+ */
-+int dpdmux_set_custom_key(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u64 key_cfg_iova)
-+{
-+ struct dpdmux_set_custom_key *cmd_params;
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPDMUX_CMDID_SET_CUSTOM_KEY,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpdmux_set_custom_key *)cmd.params;
-+ cmd_params->key_cfg_iova = cpu_to_le64(key_cfg_iova);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpdmux_add_custom_cls_entry - Adds a custom classification entry.
-+ *
-+ * This API is only available for DPDMUX instances created with
-+ * DPDMUX_METHOD_CUSTOM. Before calling this function a classification key
-+ * composition rule must be set up using dpdmux_set_custom_key.
-+ *
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPSW object
-+ * @rule: Classification rule to insert. Rules cannot be duplicated, if a
-+ * matching rule already exists, the action will be replaced.
-+ * @action: Action to perform for matching traffic.
-+ *
-+ * @returns '0' on Success; Error code otherwise.
-+ */
-+int dpdmux_add_custom_cls_entry(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ struct dpdmux_rule_cfg *rule,
-+ struct dpdmux_cls_action *action)
-+{
-+ struct dpdmux_cmd_add_custom_cls_entry *cmd_params;
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPDMUX_CMDID_ADD_CUSTOM_CLS_ENTRY,
-+ cmd_flags,
-+ token);
-+
-+ cmd_params = (struct dpdmux_cmd_add_custom_cls_entry *)cmd.params;
-+ cmd_params->key_size = rule->key_size;
-+ cmd_params->dest_if = cpu_to_le16(action->dest_if);
-+ cmd_params->key_iova = cpu_to_le64(rule->key_iova);
-+ cmd_params->mask_iova = cpu_to_le64(rule->mask_iova);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpdmux_remove_custom_cls_entry - Removes a custom classification entry.
-+ *
-+ * This API is only available for DPDMUX instances created with
-+ * DPDMUX_METHOD_CUSTOM. The API can be used to remove classification
-+ * entries previously inserted using dpdmux_add_custom_cls_entry.
-+ *
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPSW object
-+ * @rule: Classification rule to remove
-+ *
-+ * @returns '0' on Success; Error code otherwise.
-+ */
-+int dpdmux_remove_custom_cls_entry(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ struct dpdmux_rule_cfg *rule)
-+{
-+ struct dpdmux_cmd_remove_custom_cls_entry *cmd_params;
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPDMUX_CMDID_REMOVE_CUSTOM_CLS_ENTRY,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpdmux_cmd_remove_custom_cls_entry *)cmd.params;
-+ cmd_params->key_size = rule->key_size;
-+ cmd_params->key_iova = cpu_to_le64(rule->key_iova);
-+ cmd_params->mask_iova = cpu_to_le64(rule->mask_iova);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpdmux_get_api_version() - Get Data Path Demux API version
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @major_ver: Major version of data path demux API
-+ * @minor_ver: Minor version of data path demux API
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpdmux_get_api_version(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 *major_ver,
-+ u16 *minor_ver)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpdmux_rsp_get_api_version *rsp_params;
-+ int err;
-+
-+ cmd.header = mc_encode_cmd_header(DPDMUX_CMDID_GET_API_VERSION,
-+ cmd_flags,
-+ 0);
-+
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ rsp_params = (struct dpdmux_rsp_get_api_version *)cmd.params;
-+ *major_ver = le16_to_cpu(rsp_params->major);
-+ *minor_ver = le16_to_cpu(rsp_params->minor);
-+
-+ return 0;
-+}
---- /dev/null
-+++ b/drivers/staging/fsl-dpaa2/evb/dpdmux.h
-@@ -0,0 +1,453 @@
-+/* Copyright 2013-2015 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of the above-listed copyright holders nor the
-+ * names of any contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
-+ * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
-+ * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
-+ * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE
-+ * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
-+ * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
-+ * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
-+ * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
-+ * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
-+ * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
-+ * POSSIBILITY OF SUCH DAMAGE.
-+ */
-+#ifndef __FSL_DPDMUX_H
-+#define __FSL_DPDMUX_H
-+
-+struct fsl_mc_io;
-+
-+/* Data Path Demux API
-+ * Contains API for handling DPDMUX topology and functionality
-+ */
-+
-+int dpdmux_open(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ int dpdmux_id,
-+ u16 *token);
-+
-+int dpdmux_close(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token);
-+
-+/**
-+ * DPDMUX general options
-+ */
-+
-+/**
-+ * Enable bridging between internal interfaces
-+ */
-+#define DPDMUX_OPT_BRIDGE_EN 0x0000000000000002ULL
-+
-+/**
-+ * Mask support for classification
-+ */
-+#define DPDMUX_OPT_CLS_MASK_SUPPORT 0x0000000000000020ULL
-+
-+#define DPDMUX_IRQ_INDEX_IF 0x0000
-+#define DPDMUX_IRQ_INDEX 0x0001
-+
-+/**
-+ * IRQ event - Indicates that the link state changed
-+ */
-+#define DPDMUX_IRQ_EVENT_LINK_CHANGED 0x0001
-+
-+/**
-+ * enum dpdmux_manip - DPDMUX manipulation operations
-+ * @DPDMUX_MANIP_NONE: No manipulation on frames
-+ * @DPDMUX_MANIP_ADD_REMOVE_S_VLAN: Add S-VLAN on egress, remove it on ingress
-+ */
-+enum dpdmux_manip {
-+ DPDMUX_MANIP_NONE = 0x0,
-+ DPDMUX_MANIP_ADD_REMOVE_S_VLAN = 0x1
-+};
-+
-+/**
-+ * enum dpdmux_method - DPDMUX method options
-+ * @DPDMUX_METHOD_NONE: no DPDMUX method
-+ * @DPDMUX_METHOD_C_VLAN_MAC: DPDMUX based on C-VLAN and MAC address
-+ * @DPDMUX_METHOD_MAC: DPDMUX based on MAC address
-+ * @DPDMUX_METHOD_C_VLAN: DPDMUX based on C-VLAN
-+ * @DPDMUX_METHOD_S_VLAN: DPDMUX based on S-VLAN
-+ */
-+enum dpdmux_method {
-+ DPDMUX_METHOD_NONE = 0x0,
-+ DPDMUX_METHOD_C_VLAN_MAC = 0x1,
-+ DPDMUX_METHOD_MAC = 0x2,
-+ DPDMUX_METHOD_C_VLAN = 0x3,
-+ DPDMUX_METHOD_S_VLAN = 0x4,
-+ DPDMUX_METHOD_CUSTOM = 0x5
-+};
-+
-+/**
-+ * struct dpdmux_cfg - DPDMUX configuration parameters
-+ * @method: Defines the operation method for the DPDMUX address table
-+ * @manip: Required manipulation operation
-+ * @num_ifs: Number of interfaces (excluding the uplink interface)
-+ * @adv: Advanced parameters; default is all zeros;
-+ * use this structure to change default settings
-+ */
-+struct dpdmux_cfg {
-+ enum dpdmux_method method;
-+ enum dpdmux_manip manip;
-+ u16 num_ifs;
-+ /**
-+ * struct adv - Advanced parameters
-+ * @options: DPDMUX options - combination of 'DPDMUX_OPT_<X>' flags
-+ * @max_dmat_entries: Maximum entries in DPDMUX address table
-+ * 0 - indicates default: 64 entries per interface.
-+ * @max_mc_groups: Number of multicast groups in DPDMUX table
-+ * 0 - indicates default: 32 multicast groups
-+ * @max_vlan_ids: max vlan ids allowed in the system -
-+ * relevant only case of working in mac+vlan method.
-+ * 0 - indicates default 16 vlan ids.
-+ */
-+ struct {
-+ u64 options;
-+ u16 max_dmat_entries;
-+ u16 max_mc_groups;
-+ u16 max_vlan_ids;
-+ } adv;
-+};
-+
-+int dpdmux_create(struct fsl_mc_io *mc_io,
-+ u16 dprc_token,
-+ u32 cmd_flags,
-+ const struct dpdmux_cfg *cfg,
-+ u32 *obj_id);
-+
-+int dpdmux_destroy(struct fsl_mc_io *mc_io,
-+ u16 dprc_token,
-+ u32 cmd_flags,
-+ u32 object_id);
-+
-+int dpdmux_enable(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token);
-+
-+int dpdmux_disable(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token);
-+
-+int dpdmux_is_enabled(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ int *en);
-+
-+int dpdmux_reset(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token);
-+
-+int dpdmux_set_irq_enable(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 irq_index,
-+ u8 en);
-+
-+int dpdmux_get_irq_enable(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 irq_index,
-+ u8 *en);
-+
-+int dpdmux_set_irq_mask(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 irq_index,
-+ u32 mask);
-+
-+int dpdmux_get_irq_mask(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 irq_index,
-+ u32 *mask);
-+
-+int dpdmux_get_irq_status(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 irq_index,
-+ u32 *status);
-+
-+int dpdmux_clear_irq_status(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 irq_index,
-+ u32 status);
-+
-+/**
-+ * struct dpdmux_attr - Structure representing DPDMUX attributes
-+ * @id: DPDMUX object ID
-+ * @options: Configuration options (bitmap)
-+ * @method: DPDMUX address table method
-+ * @manip: DPDMUX manipulation type
-+ * @num_ifs: Number of interfaces (excluding the uplink interface)
-+ * @mem_size: DPDMUX frame storage memory size
-+ */
-+struct dpdmux_attr {
-+ int id;
-+ u64 options;
-+ enum dpdmux_method method;
-+ enum dpdmux_manip manip;
-+ u16 num_ifs;
-+ u16 mem_size;
-+};
-+
-+int dpdmux_get_attributes(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ struct dpdmux_attr *attr);
-+
-+int dpdmux_set_max_frame_length(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 max_frame_length);
-+
-+/**
-+ * enum dpdmux_counter_type - Counter types
-+ * @DPDMUX_CNT_ING_FRAME: Counts ingress frames
-+ * @DPDMUX_CNT_ING_BYTE: Counts ingress bytes
-+ * @DPDMUX_CNT_ING_FLTR_FRAME: Counts filtered ingress frames
-+ * @DPDMUX_CNT_ING_FRAME_DISCARD: Counts discarded ingress frames
-+ * @DPDMUX_CNT_ING_MCAST_FRAME: Counts ingress multicast frames
-+ * @DPDMUX_CNT_ING_MCAST_BYTE: Counts ingress multicast bytes
-+ * @DPDMUX_CNT_ING_BCAST_FRAME: Counts ingress broadcast frames
-+ * @DPDMUX_CNT_ING_BCAST_BYTES: Counts ingress broadcast bytes
-+ * @DPDMUX_CNT_EGR_FRAME: Counts egress frames
-+ * @DPDMUX_CNT_EGR_BYTE: Counts egress bytes
-+ * @DPDMUX_CNT_EGR_FRAME_DISCARD: Counts discarded egress frames
-+ */
-+enum dpdmux_counter_type {
-+ DPDMUX_CNT_ING_FRAME = 0x0,
-+ DPDMUX_CNT_ING_BYTE = 0x1,
-+ DPDMUX_CNT_ING_FLTR_FRAME = 0x2,
-+ DPDMUX_CNT_ING_FRAME_DISCARD = 0x3,
-+ DPDMUX_CNT_ING_MCAST_FRAME = 0x4,
-+ DPDMUX_CNT_ING_MCAST_BYTE = 0x5,
-+ DPDMUX_CNT_ING_BCAST_FRAME = 0x6,
-+ DPDMUX_CNT_ING_BCAST_BYTES = 0x7,
-+ DPDMUX_CNT_EGR_FRAME = 0x8,
-+ DPDMUX_CNT_EGR_BYTE = 0x9,
-+ DPDMUX_CNT_EGR_FRAME_DISCARD = 0xa
-+};
-+
-+/**
-+ * enum dpdmux_accepted_frames_type - DPDMUX frame types
-+ * @DPDMUX_ADMIT_ALL: The device accepts VLAN tagged, untagged and
-+ * priority-tagged frames
-+ * @DPDMUX_ADMIT_ONLY_VLAN_TAGGED: The device discards untagged frames or
-+ * priority-tagged frames that are received on this
-+ * interface
-+ * @DPDMUX_ADMIT_ONLY_UNTAGGED: Untagged frames or priority-tagged frames
-+ * received on this interface are accepted
-+ */
-+enum dpdmux_accepted_frames_type {
-+ DPDMUX_ADMIT_ALL = 0,
-+ DPDMUX_ADMIT_ONLY_VLAN_TAGGED = 1,
-+ DPDMUX_ADMIT_ONLY_UNTAGGED = 2
-+};
-+
-+/**
-+ * enum dpdmux_action - DPDMUX action for un-accepted frames
-+ * @DPDMUX_ACTION_DROP: Drop un-accepted frames
-+ * @DPDMUX_ACTION_REDIRECT_TO_CTRL: Redirect un-accepted frames to the
-+ * control interface
-+ */
-+enum dpdmux_action {
-+ DPDMUX_ACTION_DROP = 0,
-+ DPDMUX_ACTION_REDIRECT_TO_CTRL = 1
-+};
-+
-+/**
-+ * struct dpdmux_accepted_frames - Frame types configuration
-+ * @type: Defines ingress accepted frames
-+ * @unaccept_act: Defines action on frames not accepted
-+ */
-+struct dpdmux_accepted_frames {
-+ enum dpdmux_accepted_frames_type type;
-+ enum dpdmux_action unaccept_act;
-+};
-+
-+int dpdmux_if_set_accepted_frames(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 if_id,
-+ const struct dpdmux_accepted_frames *cfg);
-+
-+/**
-+ * struct dpdmux_if_attr - Structure representing frame types configuration
-+ * @rate: Configured interface rate (in bits per second)
-+ * @enabled: Indicates if interface is enabled
-+ * @accept_frame_type: Indicates type of accepted frames for the interface
-+ */
-+struct dpdmux_if_attr {
-+ u32 rate;
-+ int enabled;
-+ enum dpdmux_accepted_frames_type accept_frame_type;
-+};
-+
-+int dpdmux_if_get_attributes(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 if_id,
-+ struct dpdmux_if_attr *attr);
-+
-+int dpdmux_if_enable(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 if_id);
-+
-+int dpdmux_if_disable(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 if_id);
-+
-+/**
-+ * struct dpdmux_l2_rule - Structure representing L2 rule
-+ * @mac_addr: MAC address
-+ * @vlan_id: VLAN ID
-+ */
-+struct dpdmux_l2_rule {
-+ u8 mac_addr[6];
-+ u16 vlan_id;
-+};
-+
-+int dpdmux_if_remove_l2_rule(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 if_id,
-+ const struct dpdmux_l2_rule *rule);
-+
-+int dpdmux_if_add_l2_rule(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 if_id,
-+ const struct dpdmux_l2_rule *rule);
-+
-+int dpdmux_if_get_counter(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 if_id,
-+ enum dpdmux_counter_type counter_type,
-+ u64 *counter);
-+
-+int dpdmux_ul_reset_counters(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token);
-+
-+/**
-+ * Enable auto-negotiation
-+ */
-+#define DPDMUX_LINK_OPT_AUTONEG 0x0000000000000001ULL
-+/**
-+ * Enable half-duplex mode
-+ */
-+#define DPDMUX_LINK_OPT_HALF_DUPLEX 0x0000000000000002ULL
-+/**
-+ * Enable pause frames
-+ */
-+#define DPDMUX_LINK_OPT_PAUSE 0x0000000000000004ULL
-+/**
-+ * Enable a-symmetric pause frames
-+ */
-+#define DPDMUX_LINK_OPT_ASYM_PAUSE 0x0000000000000008ULL
-+
-+/**
-+ * struct dpdmux_link_cfg - Structure representing DPDMUX link configuration
-+ * @rate: Rate
-+ * @options: Mask of available options; use 'DPDMUX_LINK_OPT_<X>' values
-+ */
-+struct dpdmux_link_cfg {
-+ u32 rate;
-+ u64 options;
-+};
-+
-+int dpdmux_if_set_link_cfg(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 if_id,
-+ struct dpdmux_link_cfg *cfg);
-+/**
-+ * struct dpdmux_link_state - Structure representing DPDMUX link state
-+ * @rate: Rate
-+ * @options: Mask of available options; use 'DPDMUX_LINK_OPT_<X>' values
-+ * @up: 0 - down, 1 - up
-+ */
-+struct dpdmux_link_state {
-+ u32 rate;
-+ u64 options;
-+ int up;
-+};
-+
-+int dpdmux_if_get_link_state(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u16 if_id,
-+ struct dpdmux_link_state *state);
-+
-+int dpdmux_set_custom_key(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u64 key_cfg_iova);
-+
-+/**
-+ * struct dpdmux_rule_cfg - Custom classification rule.
-+ *
-+ * @key_iova: DMA address of buffer storing the look-up value
-+ * @mask_iova: DMA address of the mask used for TCAM classification
-+ * @key_size: size, in bytes, of the look-up value. This must match the size
-+ * of the look-up key defined using dpdmux_set_custom_key, otherwise the
-+ * entry will never be hit
-+ */
-+struct dpdmux_rule_cfg {
-+ u64 key_iova;
-+ u64 mask_iova;
-+ u8 key_size;
-+};
-+
-+/**
-+ * struct dpdmux_cls_action - Action to execute for frames matching the
-+ * classification entry
-+ *
-+ * @dest_if: Interface to forward the frames to. Port numbering is similar to
-+ * the one used to connect interfaces:
-+ * - 0 is the uplink port,
-+ * - all others are downlink ports.
-+ */
-+struct dpdmux_cls_action {
-+ u16 dest_if;
-+};
-+
-+int dpdmux_add_custom_cls_entry(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ struct dpdmux_rule_cfg *rule,
-+ struct dpdmux_cls_action *action);
-+
-+int dpdmux_remove_custom_cls_entry(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ struct dpdmux_rule_cfg *rule);
-+
-+int dpdmux_get_api_version(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 *major_ver,
-+ u16 *minor_ver);
-+
-+#endif /* __FSL_DPDMUX_H */
---- /dev/null
-+++ b/drivers/staging/fsl-dpaa2/evb/evb.c
-@@ -0,0 +1,1354 @@
-+/* Copyright 2015 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+#include <linux/module.h>
-+#include <linux/msi.h>
-+#include <linux/netdevice.h>
-+#include <linux/etherdevice.h>
-+#include <linux/rtnetlink.h>
-+#include <linux/if_vlan.h>
-+
-+#include <uapi/linux/if_bridge.h>
-+#include <net/netlink.h>
-+
-+#include <linux/fsl/mc.h>
-+
-+#include "dpdmux.h"
-+#include "dpdmux-cmd.h"
-+
-+static const char evb_drv_version[] = "0.1";
-+
-+/* Minimal supported DPDMUX version */
-+#define DPDMUX_MIN_VER_MAJOR 6
-+#define DPDMUX_MIN_VER_MINOR 0
-+
-+/* IRQ index */
-+#define DPDMUX_MAX_IRQ_NUM 2
-+
-+/* MAX FRAME LENGTH (currently 10k) */
-+#define EVB_MAX_FRAME_LENGTH (10 * 1024)
-+#define EVB_MAX_MTU (EVB_MAX_FRAME_LENGTH - VLAN_ETH_HLEN)
-+#define EVB_MIN_MTU 68
-+
-+struct evb_port_priv {
-+ struct net_device *netdev;
-+ struct list_head list;
-+ u16 port_index;
-+ struct evb_priv *evb_priv;
-+ u8 vlans[VLAN_VID_MASK + 1];
-+};
-+
-+struct evb_priv {
-+ /* keep first */
-+ struct evb_port_priv uplink;
-+
-+ struct fsl_mc_io *mc_io;
-+ struct list_head port_list;
-+ struct dpdmux_attr attr;
-+ u16 mux_handle;
-+ int dev_id;
-+};
-+
-+static int _evb_port_carrier_state_sync(struct net_device *netdev)
-+{
-+ struct evb_port_priv *port_priv = netdev_priv(netdev);
-+ struct dpdmux_link_state state;
-+ int err;
-+
-+ err = dpdmux_if_get_link_state(port_priv->evb_priv->mc_io, 0,
-+ port_priv->evb_priv->mux_handle,
-+ port_priv->port_index, &state);
-+ if (unlikely(err)) {
-+ netdev_err(netdev, "dpdmux_if_get_link_state() err %d\n", err);
-+ return err;
-+ }
-+
-+ WARN_ONCE(state.up > 1, "Garbage read into link_state");
-+
-+ if (state.up)
-+ netif_carrier_on(port_priv->netdev);
-+ else
-+ netif_carrier_off(port_priv->netdev);
-+
-+ return 0;
-+}
-+
-+static int evb_port_open(struct net_device *netdev)
-+{
-+ int err;
-+
-+ /* FIXME: enable port when support added */
-+
-+ err = _evb_port_carrier_state_sync(netdev);
-+ if (err) {
-+ netdev_err(netdev, "ethsw_port_carrier_state_sync err %d\n",
-+ err);
-+ return err;
-+ }
-+
-+ return 0;
-+}
-+
-+static netdev_tx_t evb_dropframe(struct sk_buff *skb, struct net_device *dev)
-+{
-+ /* we don't support I/O for now, drop the frame */
-+ dev_kfree_skb_any(skb);
-+ return NETDEV_TX_OK;
-+}
-+
-+static int evb_links_state_update(struct evb_priv *priv)
-+{
-+ struct evb_port_priv *port_priv;
-+ struct list_head *pos;
-+ int err;
-+
-+ list_for_each(pos, &priv->port_list) {
-+ port_priv = list_entry(pos, struct evb_port_priv, list);
-+
-+ err = _evb_port_carrier_state_sync(port_priv->netdev);
-+ if (err)
-+ netdev_err(port_priv->netdev,
-+ "_evb_port_carrier_state_sync err %d\n",
-+ err);
-+ }
-+
-+ return 0;
-+}
-+
-+static irqreturn_t evb_irq0_handler(int irq_num, void *arg)
-+{
-+ return IRQ_WAKE_THREAD;
-+}
-+
-+static irqreturn_t _evb_irq0_handler_thread(int irq_num, void *arg)
-+{
-+ struct device *dev = (struct device *)arg;
-+ struct fsl_mc_device *evb_dev = to_fsl_mc_device(dev);
-+ struct net_device *netdev = dev_get_drvdata(dev);
-+ struct evb_priv *priv = netdev_priv(netdev);
-+ struct fsl_mc_io *io = priv->mc_io;
-+ u16 token = priv->mux_handle;
-+ int irq_index = DPDMUX_IRQ_INDEX_IF;
-+
-+ /* Mask the events and the if_id reserved bits to be cleared on read */
-+ u32 status = DPDMUX_IRQ_EVENT_LINK_CHANGED | 0xFFFF0000;
-+ int err;
-+
-+ /* Sanity check */
-+ if (WARN_ON(!evb_dev || !evb_dev->irqs || !evb_dev->irqs[irq_index]))
-+ goto out;
-+ if (WARN_ON(evb_dev->irqs[irq_index]->msi_desc->irq != (u32)irq_num))
-+ goto out;
-+
-+ err = dpdmux_get_irq_status(io, 0, token, irq_index, &status);
-+ if (unlikely(err)) {
-+ netdev_err(netdev, "Can't get irq status (err %d)", err);
-+ err = dpdmux_clear_irq_status(io, 0, token, irq_index,
-+ 0xFFFFFFFF);
-+ if (unlikely(err))
-+ netdev_err(netdev, "Can't clear irq status (err %d)",
-+ err);
-+ goto out;
-+ }
-+
-+ if (status & DPDMUX_IRQ_EVENT_LINK_CHANGED) {
-+ err = evb_links_state_update(priv);
-+ if (unlikely(err))
-+ goto out;
-+ }
-+
-+out:
-+ return IRQ_HANDLED;
-+}
-+
-+static int evb_setup_irqs(struct fsl_mc_device *evb_dev)
-+{
-+ struct device *dev = &evb_dev->dev;
-+ struct net_device *netdev = dev_get_drvdata(dev);
-+ struct evb_priv *priv = netdev_priv(netdev);
-+ int err = 0;
-+ struct fsl_mc_device_irq *irq;
-+ const int irq_index = DPDMUX_IRQ_INDEX_IF;
-+ u32 mask = DPDMUX_IRQ_EVENT_LINK_CHANGED;
-+
-+ err = fsl_mc_allocate_irqs(evb_dev);
-+ if (unlikely(err)) {
-+ dev_err(dev, "MC irqs allocation failed\n");
-+ return err;
-+ }
-+
-+ if (WARN_ON(evb_dev->obj_desc.irq_count != DPDMUX_MAX_IRQ_NUM)) {
-+ err = -EINVAL;
-+ goto free_irq;
-+ }
-+
-+ err = dpdmux_set_irq_enable(priv->mc_io, 0, priv->mux_handle,
-+ irq_index, 0);
-+ if (unlikely(err)) {
-+ dev_err(dev, "dpdmux_set_irq_enable err %d\n", err);
-+ goto free_irq;
-+ }
-+
-+ irq = evb_dev->irqs[irq_index];
-+
-+ err = devm_request_threaded_irq(dev, irq->msi_desc->irq,
-+ evb_irq0_handler,
-+ _evb_irq0_handler_thread,
-+ IRQF_NO_SUSPEND | IRQF_ONESHOT,
-+ dev_name(dev), dev);
-+ if (unlikely(err)) {
-+ dev_err(dev, "devm_request_threaded_irq(): %d", err);
-+ goto free_irq;
-+ }
-+
-+ err = dpdmux_set_irq_mask(priv->mc_io, 0, priv->mux_handle,
-+ irq_index, mask);
-+ if (unlikely(err)) {
-+ dev_err(dev, "dpdmux_set_irq_mask(): %d", err);
-+ goto free_devm_irq;
-+ }
-+
-+ err = dpdmux_set_irq_enable(priv->mc_io, 0, priv->mux_handle,
-+ irq_index, 1);
-+ if (unlikely(err)) {
-+ dev_err(dev, "dpdmux_set_irq_enable(): %d", err);
-+ goto free_devm_irq;
-+ }
-+
-+ return 0;
-+
-+free_devm_irq:
-+ devm_free_irq(dev, irq->msi_desc->irq, dev);
-+free_irq:
-+ fsl_mc_free_irqs(evb_dev);
-+ return err;
-+}
-+
-+static void evb_teardown_irqs(struct fsl_mc_device *evb_dev)
-+{
-+ struct device *dev = &evb_dev->dev;
-+ struct net_device *netdev = dev_get_drvdata(dev);
-+ struct evb_priv *priv = netdev_priv(netdev);
-+
-+ dpdmux_set_irq_enable(priv->mc_io, 0, priv->mux_handle,
-+ DPDMUX_IRQ_INDEX_IF, 0);
-+
-+ devm_free_irq(dev,
-+ evb_dev->irqs[DPDMUX_IRQ_INDEX_IF]->msi_desc->irq,
-+ dev);
-+ fsl_mc_free_irqs(evb_dev);
-+}
-+
-+static int evb_port_add_rule(struct net_device *netdev,
-+ const unsigned char *addr, u16 vid)
-+{
-+ struct evb_port_priv *port_priv = netdev_priv(netdev);
-+ struct dpdmux_l2_rule rule = { .vlan_id = vid };
-+ int err;
-+
-+ if (addr)
-+ ether_addr_copy(rule.mac_addr, addr);
-+
-+ err = dpdmux_if_add_l2_rule(port_priv->evb_priv->mc_io,
-+ 0,
-+ port_priv->evb_priv->mux_handle,
-+ port_priv->port_index, &rule);
-+ if (unlikely(err))
-+ netdev_err(netdev, "dpdmux_if_add_l2_rule err %d\n", err);
-+ return err;
-+}
-+
-+static int evb_port_del_rule(struct net_device *netdev,
-+ const unsigned char *addr, u16 vid)
-+{
-+ struct evb_port_priv *port_priv = netdev_priv(netdev);
-+ struct dpdmux_l2_rule rule = { .vlan_id = vid };
-+ int err;
-+
-+ if (addr)
-+ ether_addr_copy(rule.mac_addr, addr);
-+
-+ err = dpdmux_if_remove_l2_rule(port_priv->evb_priv->mc_io,
-+ 0,
-+ port_priv->evb_priv->mux_handle,
-+ port_priv->port_index, &rule);
-+ if (unlikely(err))
-+ netdev_err(netdev, "dpdmux_if_remove_l2_rule err %d\n", err);
-+ return err;
-+}
-+
-+static bool _lookup_address(struct net_device *netdev,
-+ const unsigned char *addr)
-+{
-+ struct netdev_hw_addr *ha;
-+ struct netdev_hw_addr_list *list = (is_unicast_ether_addr(addr)) ?
-+ &netdev->uc : &netdev->mc;
-+
-+ netif_addr_lock_bh(netdev);
-+ list_for_each_entry(ha, &list->list, list) {
-+ if (ether_addr_equal(ha->addr, addr)) {
-+ netif_addr_unlock_bh(netdev);
-+ return true;
-+ }
-+ }
-+ netif_addr_unlock_bh(netdev);
-+ return false;
-+}
-+
-+static inline int evb_port_fdb_prep(struct nlattr *tb[],
-+ struct net_device *netdev,
-+ const unsigned char *addr, u16 *vid,
-+ bool del)
-+{
-+ struct evb_port_priv *port_priv = netdev_priv(netdev);
-+ struct evb_priv *evb_priv = port_priv->evb_priv;
-+
-+ *vid = 0;
-+
-+ if (evb_priv->attr.method != DPDMUX_METHOD_MAC &&
-+ evb_priv->attr.method != DPDMUX_METHOD_C_VLAN_MAC) {
-+ netdev_err(netdev,
-+ "EVB mode does not support MAC classification\n");
-+ return -EOPNOTSUPP;
-+ }
-+
-+ /* check if the address is configured on this port */
-+ if (_lookup_address(netdev, addr)) {
-+ if (!del)
-+ return -EEXIST;
-+ } else {
-+ if (del)
-+ return -ENOENT;
-+ }
-+
-+ if (tb[NDA_VLAN] && evb_priv->attr.method == DPDMUX_METHOD_C_VLAN_MAC) {
-+ if (nla_len(tb[NDA_VLAN]) != sizeof(unsigned short)) {
-+ netdev_err(netdev, "invalid vlan size %d\n",
-+ nla_len(tb[NDA_VLAN]));
-+ return -EINVAL;
-+ }
-+
-+ *vid = nla_get_u16(tb[NDA_VLAN]);
-+
-+ if (!*vid || *vid >= VLAN_VID_MASK) {
-+ netdev_err(netdev, "invalid vid value 0x%04x\n", *vid);
-+ return -EINVAL;
-+ }
-+ } else if (evb_priv->attr.method == DPDMUX_METHOD_C_VLAN_MAC) {
-+ netdev_err(netdev,
-+ "EVB mode requires explicit VLAN configuration\n");
-+ return -EINVAL;
-+ } else if (tb[NDA_VLAN]) {
-+ netdev_warn(netdev, "VLAN not supported, argument ignored\n");
-+ }
-+
-+ return 0;
-+}
-+
-+static int evb_port_fdb_add(struct ndmsg *ndm, struct nlattr *tb[],
-+ struct net_device *netdev,
-+ const unsigned char *addr, u16 vid, u16 flags)
-+{
-+ u16 _vid;
-+ int err;
-+
-+ /* TODO: add replace support when added to iproute bridge */
-+ if (!(flags & NLM_F_REQUEST)) {
-+ netdev_err(netdev,
-+ "evb_port_fdb_add unexpected flags value %08x\n",
-+ flags);
-+ return -EINVAL;
-+ }
-+
-+ err = evb_port_fdb_prep(tb, netdev, addr, &_vid, 0);
-+ if (unlikely(err))
-+ return err;
-+
-+ err = evb_port_add_rule(netdev, addr, _vid);
-+ if (unlikely(err))
-+ return err;
-+
-+ if (is_unicast_ether_addr(addr)) {
-+ err = dev_uc_add(netdev, addr);
-+ if (unlikely(err)) {
-+ netdev_err(netdev, "dev_uc_add err %d\n", err);
-+ return err;
-+ }
-+ } else {
-+ err = dev_mc_add(netdev, addr);
-+ if (unlikely(err)) {
-+ netdev_err(netdev, "dev_mc_add err %d\n", err);
-+ return err;
-+ }
-+ }
-+
-+ return 0;
-+}
-+
-+static int evb_port_fdb_del(struct ndmsg *ndm, struct nlattr *tb[],
-+ struct net_device *netdev,
-+ const unsigned char *addr, u16 vid)
-+{
-+ u16 _vid;
-+ int err;
-+
-+ err = evb_port_fdb_prep(tb, netdev, addr, &_vid, 1);
-+ if (unlikely(err))
-+ return err;
-+
-+ err = evb_port_del_rule(netdev, addr, _vid);
-+ if (unlikely(err))
-+ return err;
-+
-+ if (is_unicast_ether_addr(addr)) {
-+ err = dev_uc_del(netdev, addr);
-+ if (unlikely(err)) {
-+ netdev_err(netdev, "dev_uc_del err %d\n", err);
-+ return err;
-+ }
-+ } else {
-+ err = dev_mc_del(netdev, addr);
-+ if (unlikely(err)) {
-+ netdev_err(netdev, "dev_mc_del err %d\n", err);
-+ return err;
-+ }
-+ }
-+
-+ return 0;
-+}
-+
-+static int evb_change_mtu(struct net_device *netdev,
-+ int mtu)
-+{
-+ struct evb_port_priv *port_priv = netdev_priv(netdev);
-+ struct evb_priv *evb_priv = port_priv->evb_priv;
-+ struct list_head *pos;
-+ int err = 0;
-+
-+ /* This operation is not permitted on downlinks */
-+ if (port_priv->port_index > 0)
-+ return -EPERM;
-+
-+ err = dpdmux_set_max_frame_length(evb_priv->mc_io,
-+ 0,
-+ evb_priv->mux_handle,
-+ (uint16_t)(mtu + VLAN_ETH_HLEN));
-+
-+ if (unlikely(err)) {
-+ netdev_err(netdev, "dpdmux_ul_set_max_frame_length err %d\n",
-+ err);
-+ return err;
-+ }
-+
-+ /* Update the max frame length for downlinks */
-+ list_for_each(pos, &evb_priv->port_list) {
-+ port_priv = list_entry(pos, struct evb_port_priv, list);
-+ port_priv->netdev->mtu = mtu;
-+ }
-+
-+ netdev->mtu = mtu;
-+ return 0;
-+}
-+
-+static const struct nla_policy ifla_br_policy[IFLA_MAX + 1] = {
-+ [IFLA_BRIDGE_FLAGS] = { .type = NLA_U16 },
-+ [IFLA_BRIDGE_MODE] = { .type = NLA_U16 },
-+ [IFLA_BRIDGE_VLAN_INFO] = { .type = NLA_BINARY,
-+ .len = sizeof(struct bridge_vlan_info), },
-+};
-+
-+static int evb_setlink_af_spec(struct net_device *netdev,
-+ struct nlattr **tb)
-+{
-+ struct bridge_vlan_info *vinfo;
-+ struct evb_port_priv *port_priv = netdev_priv(netdev);
-+ int err = 0;
-+
-+ if (!tb[IFLA_BRIDGE_VLAN_INFO]) {
-+ netdev_err(netdev, "no VLAN INFO in nlmsg\n");
-+ return -EOPNOTSUPP;
-+ }
-+
-+ vinfo = nla_data(tb[IFLA_BRIDGE_VLAN_INFO]);
-+
-+ if (!vinfo->vid || vinfo->vid > VLAN_VID_MASK)
-+ return -EINVAL;
-+
-+ err = evb_port_add_rule(netdev, NULL, vinfo->vid);
-+ if (unlikely(err))
-+ return err;
-+
-+ port_priv->vlans[vinfo->vid] = 1;
-+
-+ return 0;
-+}
-+
-+static int evb_setlink(struct net_device *netdev,
-+ struct nlmsghdr *nlh,
-+ u16 flags)
-+{
-+ struct evb_port_priv *port_priv = netdev_priv(netdev);
-+ struct evb_priv *evb_priv = port_priv->evb_priv;
-+ struct nlattr *attr;
-+ struct nlattr *tb[(IFLA_BRIDGE_MAX > IFLA_BRPORT_MAX) ?
-+ IFLA_BRIDGE_MAX : IFLA_BRPORT_MAX + 1];
-+ int err = 0;
-+
-+ if (evb_priv->attr.method != DPDMUX_METHOD_C_VLAN &&
-+ evb_priv->attr.method != DPDMUX_METHOD_S_VLAN) {
-+ netdev_err(netdev,
-+ "EVB mode does not support VLAN only classification\n");
-+ return -EOPNOTSUPP;
-+ }
-+
-+ attr = nlmsg_find_attr(nlh, sizeof(struct ifinfomsg), IFLA_AF_SPEC);
-+ if (attr) {
-+ err = nla_parse_nested(tb, IFLA_BRIDGE_MAX, attr,
-+ ifla_br_policy);
-+ if (unlikely(err)) {
-+ netdev_err(netdev,
-+ "nla_parse_nested for br_policy err %d\n",
-+ err);
-+ return err;
-+ }
-+
-+ err = evb_setlink_af_spec(netdev, tb);
-+ return err;
-+ }
-+
-+ netdev_err(netdev, "nlmsg_find_attr found no AF_SPEC\n");
-+ return -EOPNOTSUPP;
-+}
-+
-+static int __nla_put_netdev(struct sk_buff *skb, struct net_device *netdev)
-+{
-+ struct evb_port_priv *port_priv = netdev_priv(netdev);
-+ struct evb_priv *evb_priv = port_priv->evb_priv;
-+ u8 operstate = netif_running(netdev) ?
-+ netdev->operstate : IF_OPER_DOWN;
-+ int iflink;
-+ int err;
-+
-+ err = nla_put_string(skb, IFLA_IFNAME, netdev->name);
-+ if (unlikely(err))
-+ goto nla_put_err;
-+ err = nla_put_u32(skb, IFLA_MASTER, evb_priv->uplink.netdev->ifindex);
-+ if (unlikely(err))
-+ goto nla_put_err;
-+ err = nla_put_u32(skb, IFLA_MTU, netdev->mtu);
-+ if (unlikely(err))
-+ goto nla_put_err;
-+ err = nla_put_u8(skb, IFLA_OPERSTATE, operstate);
-+ if (unlikely(err))
-+ goto nla_put_err;
-+ if (netdev->addr_len) {
-+ err = nla_put(skb, IFLA_ADDRESS, netdev->addr_len,
-+ netdev->dev_addr);
-+ if (unlikely(err))
-+ goto nla_put_err;
-+ }
-+
-+ iflink = dev_get_iflink(netdev);
-+ if (netdev->ifindex != iflink) {
-+ err = nla_put_u32(skb, IFLA_LINK, iflink);
-+ if (unlikely(err))
-+ goto nla_put_err;
-+ }
-+
-+ return 0;
-+
-+nla_put_err:
-+ netdev_err(netdev, "nla_put_ err %d\n", err);
-+ return err;
-+}
-+
-+static int __nla_put_port(struct sk_buff *skb, struct net_device *netdev)
-+{
-+ struct nlattr *nest;
-+ int err;
-+
-+ nest = nla_nest_start(skb, IFLA_PROTINFO | NLA_F_NESTED);
-+ if (!nest) {
-+ netdev_err(netdev, "nla_nest_start failed\n");
-+ return -ENOMEM;
-+ }
-+
-+ err = nla_put_u8(skb, IFLA_BRPORT_STATE, BR_STATE_FORWARDING);
-+ if (unlikely(err))
-+ goto nla_put_err;
-+ err = nla_put_u16(skb, IFLA_BRPORT_PRIORITY, 0);
-+ if (unlikely(err))
-+ goto nla_put_err;
-+ err = nla_put_u32(skb, IFLA_BRPORT_COST, 0);
-+ if (unlikely(err))
-+ goto nla_put_err;
-+ err = nla_put_u8(skb, IFLA_BRPORT_MODE, 0);
-+ if (unlikely(err))
-+ goto nla_put_err;
-+ err = nla_put_u8(skb, IFLA_BRPORT_GUARD, 0);
-+ if (unlikely(err))
-+ goto nla_put_err;
-+ err = nla_put_u8(skb, IFLA_BRPORT_PROTECT, 0);
-+ if (unlikely(err))
-+ goto nla_put_err;
-+ err = nla_put_u8(skb, IFLA_BRPORT_FAST_LEAVE, 0);
-+ if (unlikely(err))
-+ goto nla_put_err;
-+ err = nla_put_u8(skb, IFLA_BRPORT_LEARNING, 0);
-+ if (unlikely(err))
-+ goto nla_put_err;
-+ err = nla_put_u8(skb, IFLA_BRPORT_UNICAST_FLOOD, 1);
-+ if (unlikely(err))
-+ goto nla_put_err;
-+ nla_nest_end(skb, nest);
-+
-+ return 0;
-+
-+nla_put_err:
-+ netdev_err(netdev, "nla_put_ err %d\n", err);
-+ nla_nest_cancel(skb, nest);
-+ return err;
-+}
-+
-+static int __nla_put_vlan(struct sk_buff *skb, struct net_device *netdev)
-+{
-+ struct evb_port_priv *port_priv = netdev_priv(netdev);
-+ struct nlattr *nest;
-+ struct bridge_vlan_info vinfo;
-+ const u8 *vlans = port_priv->vlans;
-+ u16 i;
-+ int err;
-+
-+ nest = nla_nest_start(skb, IFLA_AF_SPEC);
-+ if (!nest) {
-+ netdev_err(netdev, "nla_nest_start failed");
-+ return -ENOMEM;
-+ }
-+
-+ for (i = 0; i < VLAN_VID_MASK + 1; i++) {
-+ if (!vlans[i])
-+ continue;
-+
-+ vinfo.flags = 0;
-+ vinfo.vid = i;
-+
-+ err = nla_put(skb, IFLA_BRIDGE_VLAN_INFO,
-+ sizeof(vinfo), &vinfo);
-+ if (unlikely(err))
-+ goto nla_put_err;
-+ }
-+
-+ nla_nest_end(skb, nest);
-+
-+ return 0;
-+
-+nla_put_err:
-+ netdev_err(netdev, "nla_put_ err %d\n", err);
-+ nla_nest_cancel(skb, nest);
-+ return err;
-+}
-+
-+static int evb_getlink(struct sk_buff *skb, u32 pid, u32 seq,
-+ struct net_device *netdev, u32 filter_mask, int nlflags)
-+{
-+ struct evb_port_priv *port_priv = netdev_priv(netdev);
-+ struct evb_priv *evb_priv = port_priv->evb_priv;
-+ struct ifinfomsg *hdr;
-+ struct nlmsghdr *nlh;
-+ int err;
-+
-+ if (evb_priv->attr.method != DPDMUX_METHOD_C_VLAN &&
-+ evb_priv->attr.method != DPDMUX_METHOD_S_VLAN) {
-+ return 0;
-+ }
-+
-+ nlh = nlmsg_put(skb, pid, seq, RTM_NEWLINK, sizeof(*hdr), NLM_F_MULTI);
-+ if (!nlh)
-+ return -EMSGSIZE;
-+
-+ hdr = nlmsg_data(nlh);
-+ memset(hdr, 0, sizeof(*hdr));
-+ hdr->ifi_family = AF_BRIDGE;
-+ hdr->ifi_type = netdev->type;
-+ hdr->ifi_index = netdev->ifindex;
-+ hdr->ifi_flags = dev_get_flags(netdev);
-+
-+ err = __nla_put_netdev(skb, netdev);
-+ if (unlikely(err))
-+ goto nla_put_err;
-+
-+ err = __nla_put_port(skb, netdev);
-+ if (unlikely(err))
-+ goto nla_put_err;
-+
-+ /* Check if the VID information is requested */
-+ if (filter_mask & RTEXT_FILTER_BRVLAN) {
-+ err = __nla_put_vlan(skb, netdev);
-+ if (unlikely(err))
-+ goto nla_put_err;
-+ }
-+
-+ nlmsg_end(skb, nlh);
-+ return skb->len;
-+
-+nla_put_err:
-+ nlmsg_cancel(skb, nlh);
-+ return -EMSGSIZE;
-+}
-+
-+static int evb_dellink(struct net_device *netdev,
-+ struct nlmsghdr *nlh,
-+ u16 flags)
-+{
-+ struct nlattr *tb[IFLA_BRIDGE_MAX + 1];
-+ struct nlattr *spec;
-+ struct bridge_vlan_info *vinfo;
-+ struct evb_port_priv *port_priv = netdev_priv(netdev);
-+ int err = 0;
-+
-+ spec = nlmsg_find_attr(nlh, sizeof(struct ifinfomsg), IFLA_AF_SPEC);
-+ if (!spec)
-+ return 0;
-+
-+ err = nla_parse_nested(tb, IFLA_BRIDGE_MAX, spec, ifla_br_policy);
-+ if (unlikely(err))
-+ return err;
-+
-+ if (!tb[IFLA_BRIDGE_VLAN_INFO])
-+ return -EOPNOTSUPP;
-+
-+ vinfo = nla_data(tb[IFLA_BRIDGE_VLAN_INFO]);
-+
-+ if (!vinfo->vid || vinfo->vid > VLAN_VID_MASK)
-+ return -EINVAL;
-+
-+ err = evb_port_del_rule(netdev, NULL, vinfo->vid);
-+ if (unlikely(err)) {
-+ netdev_err(netdev, "evb_port_del_rule err %d\n", err);
-+ return err;
-+ }
-+ port_priv->vlans[vinfo->vid] = 0;
-+
-+ return 0;
-+}
-+
-+struct rtnl_link_stats64 *evb_port_get_stats(struct net_device *netdev,
-+ struct rtnl_link_stats64 *storage)
-+{
-+ struct evb_port_priv *port_priv = netdev_priv(netdev);
-+ u64 tmp;
-+ int err;
-+
-+ err = dpdmux_if_get_counter(port_priv->evb_priv->mc_io,
-+ 0,
-+ port_priv->evb_priv->mux_handle,
-+ port_priv->port_index,
-+ DPDMUX_CNT_ING_FRAME, &storage->rx_packets);
-+ if (unlikely(err))
-+ goto error;
-+
-+ err = dpdmux_if_get_counter(port_priv->evb_priv->mc_io,
-+ 0,
-+ port_priv->evb_priv->mux_handle,
-+ port_priv->port_index,
-+ DPDMUX_CNT_ING_BYTE, &storage->rx_bytes);
-+ if (unlikely(err))
-+ goto error;
-+
-+ err = dpdmux_if_get_counter(port_priv->evb_priv->mc_io,
-+ 0,
-+ port_priv->evb_priv->mux_handle,
-+ port_priv->port_index,
-+ DPDMUX_CNT_ING_FLTR_FRAME, &tmp);
-+ if (unlikely(err))
-+ goto error;
-+
-+ err = dpdmux_if_get_counter(port_priv->evb_priv->mc_io,
-+ 0,
-+ port_priv->evb_priv->mux_handle,
-+ port_priv->port_index,
-+ DPDMUX_CNT_ING_FRAME_DISCARD,
-+ &storage->rx_dropped);
-+ if (unlikely(err)) {
-+ storage->rx_dropped = tmp;
-+ goto error;
-+ }
-+ storage->rx_dropped += tmp;
-+
-+ err = dpdmux_if_get_counter(port_priv->evb_priv->mc_io,
-+ 0,
-+ port_priv->evb_priv->mux_handle,
-+ port_priv->port_index,
-+ DPDMUX_CNT_ING_MCAST_FRAME,
-+ &storage->multicast);
-+ if (unlikely(err))
-+ goto error;
-+
-+ err = dpdmux_if_get_counter(port_priv->evb_priv->mc_io,
-+ 0,
-+ port_priv->evb_priv->mux_handle,
-+ port_priv->port_index,
-+ DPDMUX_CNT_EGR_FRAME, &storage->tx_packets);
-+ if (unlikely(err))
-+ goto error;
-+
-+ err = dpdmux_if_get_counter(port_priv->evb_priv->mc_io,
-+ 0,
-+ port_priv->evb_priv->mux_handle,
-+ port_priv->port_index,
-+ DPDMUX_CNT_EGR_BYTE, &storage->tx_bytes);
-+ if (unlikely(err))
-+ goto error;
-+
-+ err = dpdmux_if_get_counter(port_priv->evb_priv->mc_io,
-+ 0,
-+ port_priv->evb_priv->mux_handle,
-+ port_priv->port_index,
-+ DPDMUX_CNT_EGR_FRAME_DISCARD,
-+ &storage->tx_dropped);
-+ if (unlikely(err))
-+ goto error;
-+
-+ return storage;
-+
-+error:
-+ netdev_err(netdev, "dpdmux_if_get_counter err %d\n", err);
-+ return storage;
-+}
-+
-+static const struct net_device_ops evb_port_ops = {
-+ .ndo_open = &evb_port_open,
-+
-+ .ndo_start_xmit = &evb_dropframe,
-+
-+ .ndo_fdb_add = &evb_port_fdb_add,
-+ .ndo_fdb_del = &evb_port_fdb_del,
-+
-+ .ndo_get_stats64 = &evb_port_get_stats,
-+ .ndo_change_mtu = &evb_change_mtu,
-+};
-+
-+static void evb_get_drvinfo(struct net_device *netdev,
-+ struct ethtool_drvinfo *drvinfo)
-+{
-+ struct evb_port_priv *port_priv = netdev_priv(netdev);
-+ u16 version_major, version_minor;
-+ int err;
-+
-+ strlcpy(drvinfo->driver, KBUILD_MODNAME, sizeof(drvinfo->driver));
-+ strlcpy(drvinfo->version, evb_drv_version, sizeof(drvinfo->version));
-+
-+ err = dpdmux_get_api_version(port_priv->evb_priv->mc_io, 0,
-+ &version_major,
-+ &version_minor);
-+ if (err)
-+ strlcpy(drvinfo->fw_version, "N/A",
-+ sizeof(drvinfo->fw_version));
-+ else
-+ snprintf(drvinfo->fw_version, sizeof(drvinfo->fw_version),
-+ "%u.%u", version_major, version_minor);
-+
-+ strlcpy(drvinfo->bus_info, dev_name(netdev->dev.parent->parent),
-+ sizeof(drvinfo->bus_info));
-+}
-+
-+static int evb_get_settings(struct net_device *netdev,
-+ struct ethtool_cmd *cmd)
-+{
-+ struct evb_port_priv *port_priv = netdev_priv(netdev);
-+ struct dpdmux_link_state state = {0};
-+ int err = 0;
-+
-+ err = dpdmux_if_get_link_state(port_priv->evb_priv->mc_io, 0,
-+ port_priv->evb_priv->mux_handle,
-+ port_priv->port_index,
-+ &state);
-+ if (err) {
-+ netdev_err(netdev, "ERROR %d getting link state", err);
-+ goto out;
-+ }
-+
-+ /* At the moment, we have no way of interrogating the DPMAC
-+ * from the DPDMUX side or there may not exist a DPMAC at all.
-+ * Report only autoneg state, duplexity and speed.
-+ */
-+ if (state.options & DPDMUX_LINK_OPT_AUTONEG)
-+ cmd->autoneg = AUTONEG_ENABLE;
-+ if (!(state.options & DPDMUX_LINK_OPT_HALF_DUPLEX))
-+ cmd->duplex = DUPLEX_FULL;
-+ ethtool_cmd_speed_set(cmd, state.rate);
-+
-+out:
-+ return err;
-+}
-+
-+static int evb_set_settings(struct net_device *netdev,
-+ struct ethtool_cmd *cmd)
-+{
-+ struct evb_port_priv *port_priv = netdev_priv(netdev);
-+ struct dpdmux_link_state state = {0};
-+ struct dpdmux_link_cfg cfg = {0};
-+ int err = 0;
-+
-+ netdev_dbg(netdev, "Setting link parameters...");
-+
-+ err = dpdmux_if_get_link_state(port_priv->evb_priv->mc_io, 0,
-+ port_priv->evb_priv->mux_handle,
-+ port_priv->port_index,
-+ &state);
-+ if (err) {
-+ netdev_err(netdev, "ERROR %d getting link state", err);
-+ goto out;
-+ }
-+
-+ /* Due to a temporary MC limitation, the DPDMUX port must be down
-+ * in order to be able to change link settings. Taking steps to let
-+ * the user know that.
-+ */
-+ if (netif_running(netdev)) {
-+ netdev_info(netdev,
-+ "Sorry, interface must be brought down first.\n");
-+ return -EACCES;
-+ }
-+
-+ cfg.options = state.options;
-+ cfg.rate = ethtool_cmd_speed(cmd);
-+ if (cmd->autoneg == AUTONEG_ENABLE)
-+ cfg.options |= DPDMUX_LINK_OPT_AUTONEG;
-+ else
-+ cfg.options &= ~DPDMUX_LINK_OPT_AUTONEG;
-+ if (cmd->duplex == DUPLEX_HALF)
-+ cfg.options |= DPDMUX_LINK_OPT_HALF_DUPLEX;
-+ else
-+ cfg.options &= ~DPDMUX_LINK_OPT_HALF_DUPLEX;
-+
-+ err = dpdmux_if_set_link_cfg(port_priv->evb_priv->mc_io, 0,
-+ port_priv->evb_priv->mux_handle,
-+ port_priv->port_index,
-+ &cfg);
-+ if (err)
-+ /* ethtool will be loud enough if we return an error; no point
-+ * in putting our own error message on the console by default
-+ */
-+ netdev_dbg(netdev, "ERROR %d setting link cfg", err);
-+
-+out:
-+ return err;
-+}
-+
-+static struct {
-+ enum dpdmux_counter_type id;
-+ char name[ETH_GSTRING_LEN];
-+} evb_ethtool_counters[] = {
-+ {DPDMUX_CNT_ING_FRAME, "rx frames"},
-+ {DPDMUX_CNT_ING_BYTE, "rx bytes"},
-+ {DPDMUX_CNT_ING_FLTR_FRAME, "rx filtered frames"},
-+ {DPDMUX_CNT_ING_FRAME_DISCARD, "rx discarded frames"},
-+ {DPDMUX_CNT_ING_BCAST_FRAME, "rx b-cast frames"},
-+ {DPDMUX_CNT_ING_BCAST_BYTES, "rx b-cast bytes"},
-+ {DPDMUX_CNT_ING_MCAST_FRAME, "rx m-cast frames"},
-+ {DPDMUX_CNT_ING_MCAST_BYTE, "rx m-cast bytes"},
-+ {DPDMUX_CNT_EGR_FRAME, "tx frames"},
-+ {DPDMUX_CNT_EGR_BYTE, "tx bytes"},
-+ {DPDMUX_CNT_EGR_FRAME_DISCARD, "tx discarded frames"},
-+};
-+
-+static int evb_ethtool_get_sset_count(struct net_device *dev, int sset)
-+{
-+ switch (sset) {
-+ case ETH_SS_STATS:
-+ return ARRAY_SIZE(evb_ethtool_counters);
-+ default:
-+ return -EOPNOTSUPP;
-+ }
-+}
-+
-+static void evb_ethtool_get_strings(struct net_device *netdev,
-+ u32 stringset, u8 *data)
-+{
-+ u32 i;
-+
-+ switch (stringset) {
-+ case ETH_SS_STATS:
-+ for (i = 0; i < ARRAY_SIZE(evb_ethtool_counters); i++)
-+ memcpy(data + i * ETH_GSTRING_LEN,
-+ evb_ethtool_counters[i].name, ETH_GSTRING_LEN);
-+ break;
-+ }
-+}
-+
-+static void evb_ethtool_get_stats(struct net_device *netdev,
-+ struct ethtool_stats *stats,
-+ u64 *data)
-+{
-+ struct evb_port_priv *port_priv = netdev_priv(netdev);
-+ u32 i;
-+ int err;
-+
-+ for (i = 0; i < ARRAY_SIZE(evb_ethtool_counters); i++) {
-+ err = dpdmux_if_get_counter(port_priv->evb_priv->mc_io,
-+ 0,
-+ port_priv->evb_priv->mux_handle,
-+ port_priv->port_index,
-+ evb_ethtool_counters[i].id,
-+ &data[i]);
-+ if (err)
-+ netdev_err(netdev, "dpdmux_if_get_counter[%s] err %d\n",
-+ evb_ethtool_counters[i].name, err);
-+ }
-+}
-+
-+static const struct ethtool_ops evb_port_ethtool_ops = {
-+ .get_drvinfo = &evb_get_drvinfo,
-+ .get_link = &ethtool_op_get_link,
-+ .get_settings = &evb_get_settings,
-+ .set_settings = &evb_set_settings,
-+ .get_strings = &evb_ethtool_get_strings,
-+ .get_ethtool_stats = &evb_ethtool_get_stats,
-+ .get_sset_count = &evb_ethtool_get_sset_count,
-+};
-+
-+static int evb_open(struct net_device *netdev)
-+{
-+ struct evb_priv *priv = netdev_priv(netdev);
-+ int err = 0;
-+
-+ err = dpdmux_enable(priv->mc_io, 0, priv->mux_handle);
-+ if (unlikely(err))
-+ netdev_err(netdev, "dpdmux_enable err %d\n", err);
-+
-+ return err;
-+}
-+
-+static int evb_close(struct net_device *netdev)
-+{
-+ struct evb_priv *priv = netdev_priv(netdev);
-+ int err = 0;
-+
-+ err = dpdmux_disable(priv->mc_io, 0, priv->mux_handle);
-+ if (unlikely(err))
-+ netdev_err(netdev, "dpdmux_disable err %d\n", err);
-+
-+ return err;
-+}
-+
-+static const struct net_device_ops evb_ops = {
-+ .ndo_start_xmit = &evb_dropframe,
-+ .ndo_open = &evb_open,
-+ .ndo_stop = &evb_close,
-+
-+ .ndo_bridge_setlink = &evb_setlink,
-+ .ndo_bridge_getlink = &evb_getlink,
-+ .ndo_bridge_dellink = &evb_dellink,
-+
-+ .ndo_get_stats64 = &evb_port_get_stats,
-+ .ndo_change_mtu = &evb_change_mtu,
-+};
-+
-+static int evb_takedown(struct fsl_mc_device *evb_dev)
-+{
-+ struct device *dev = &evb_dev->dev;
-+ struct net_device *netdev = dev_get_drvdata(dev);
-+ struct evb_priv *priv = netdev_priv(netdev);
-+ int err;
-+
-+ err = dpdmux_close(priv->mc_io, 0, priv->mux_handle);
-+ if (unlikely(err))
-+ dev_warn(dev, "dpdmux_close err %d\n", err);
-+
-+ return 0;
-+}
-+
-+static int evb_init(struct fsl_mc_device *evb_dev)
-+{
-+ struct device *dev = &evb_dev->dev;
-+ struct net_device *netdev = dev_get_drvdata(dev);
-+ struct evb_priv *priv = netdev_priv(netdev);
-+ u16 version_major;
-+ u16 version_minor;
-+ int err = 0;
-+
-+ priv->dev_id = evb_dev->obj_desc.id;
-+
-+ err = dpdmux_open(priv->mc_io, 0, priv->dev_id, &priv->mux_handle);
-+ if (unlikely(err)) {
-+ dev_err(dev, "dpdmux_open err %d\n", err);
-+ goto err_exit;
-+ }
-+ if (!priv->mux_handle) {
-+ dev_err(dev, "dpdmux_open returned null handle but no error\n");
-+ err = -EFAULT;
-+ goto err_exit;
-+ }
-+
-+ err = dpdmux_get_attributes(priv->mc_io, 0, priv->mux_handle,
-+ &priv->attr);
-+ if (unlikely(err)) {
-+ dev_err(dev, "dpdmux_get_attributes err %d\n", err);
-+ goto err_close;
-+ }
-+
-+ err = dpdmux_get_api_version(priv->mc_io, 0,
-+ &version_major,
-+ &version_minor);
-+ if (unlikely(err)) {
-+ dev_err(dev, "dpdmux_get_api_version err %d\n", err);
-+ goto err_close;
-+ }
-+
-+ /* Minimum supported DPDMUX version check */
-+ if (version_major < DPDMUX_MIN_VER_MAJOR ||
-+ (version_major == DPDMUX_MIN_VER_MAJOR &&
-+ version_minor < DPDMUX_MIN_VER_MINOR)) {
-+ dev_err(dev, "DPDMUX version %d.%d not supported. Use %d.%d or greater.\n",
-+ version_major, version_minor,
-+ DPDMUX_MIN_VER_MAJOR, DPDMUX_MIN_VER_MAJOR);
-+ err = -ENOTSUPP;
-+ goto err_close;
-+ }
-+
-+ err = dpdmux_reset(priv->mc_io, 0, priv->mux_handle);
-+ if (unlikely(err)) {
-+ dev_err(dev, "dpdmux_reset err %d\n", err);
-+ goto err_close;
-+ }
-+
-+ return 0;
-+
-+err_close:
-+ dpdmux_close(priv->mc_io, 0, priv->mux_handle);
-+err_exit:
-+ return err;
-+}
-+
-+static int evb_remove(struct fsl_mc_device *evb_dev)
-+{
-+ struct device *dev = &evb_dev->dev;
-+ struct net_device *netdev = dev_get_drvdata(dev);
-+ struct evb_priv *priv = netdev_priv(netdev);
-+ struct evb_port_priv *port_priv;
-+ struct list_head *pos;
-+
-+ list_for_each(pos, &priv->port_list) {
-+ port_priv = list_entry(pos, struct evb_port_priv, list);
-+
-+ rtnl_lock();
-+ netdev_upper_dev_unlink(port_priv->netdev, netdev);
-+ rtnl_unlock();
-+
-+ unregister_netdev(port_priv->netdev);
-+ free_netdev(port_priv->netdev);
-+ }
-+
-+ evb_teardown_irqs(evb_dev);
-+
-+ unregister_netdev(netdev);
-+
-+ evb_takedown(evb_dev);
-+ fsl_mc_portal_free(priv->mc_io);
-+
-+ dev_set_drvdata(dev, NULL);
-+ free_netdev(netdev);
-+
-+ return 0;
-+}
-+
-+static int evb_probe(struct fsl_mc_device *evb_dev)
-+{
-+ struct device *dev;
-+ struct evb_priv *priv = NULL;
-+ struct net_device *netdev = NULL;
-+ char port_name[IFNAMSIZ];
-+ int i;
-+ int err = 0;
-+
-+ dev = &evb_dev->dev;
-+
-+ /* register switch device, it's for management only - no I/O */
-+ netdev = alloc_etherdev(sizeof(*priv));
-+ if (!netdev) {
-+ dev_err(dev, "alloc_etherdev error\n");
-+ return -ENOMEM;
-+ }
-+ netdev->netdev_ops = &evb_ops;
-+
-+ dev_set_drvdata(dev, netdev);
-+
-+ priv = netdev_priv(netdev);
-+
-+ err = fsl_mc_portal_allocate(evb_dev, FSL_MC_IO_ATOMIC_CONTEXT_PORTAL,
-+ &priv->mc_io);
-+ if (err) {
-+ if (err == -ENXIO)
-+ err = -EPROBE_DEFER;
-+ else
-+ dev_err(dev, "fsl_mc_portal_allocate err %d\n", err);
-+ goto err_free_netdev;
-+ }
-+
-+ if (!priv->mc_io) {
-+ dev_err(dev, "fsl_mc_portal_allocate returned null handle but no error\n");
-+ err = -EFAULT;
-+ goto err_free_netdev;
-+ }
-+
-+ err = evb_init(evb_dev);
-+ if (unlikely(err)) {
-+ dev_err(dev, "evb init err %d\n", err);
-+ goto err_free_cmdport;
-+ }
-+
-+ INIT_LIST_HEAD(&priv->port_list);
-+ netdev->flags |= IFF_PROMISC | IFF_MASTER;
-+
-+ dev_alloc_name(netdev, "evb%d");
-+
-+ /* register switch ports */
-+ snprintf(port_name, IFNAMSIZ, "%sp%%d", netdev->name);
-+
-+ /* only register downlinks? */
-+ for (i = 0; i < priv->attr.num_ifs + 1; i++) {
-+ struct net_device *port_netdev;
-+ struct evb_port_priv *port_priv;
-+
-+ if (i) {
-+ port_netdev =
-+ alloc_etherdev(sizeof(struct evb_port_priv));
-+ if (!port_netdev) {
-+ dev_err(dev, "alloc_etherdev error\n");
-+ goto err_takedown;
-+ }
-+
-+ port_priv = netdev_priv(port_netdev);
-+
-+ port_netdev->flags |= IFF_PROMISC | IFF_SLAVE;
-+
-+ dev_alloc_name(port_netdev, port_name);
-+ } else {
-+ port_netdev = netdev;
-+ port_priv = &priv->uplink;
-+ }
-+
-+ port_priv->netdev = port_netdev;
-+ port_priv->evb_priv = priv;
-+ port_priv->port_index = i;
-+
-+ SET_NETDEV_DEV(port_netdev, dev);
-+
-+ if (i) {
-+ port_netdev->netdev_ops = &evb_port_ops;
-+
-+ err = register_netdev(port_netdev);
-+ if (err < 0) {
-+ dev_err(dev, "register_netdev err %d\n", err);
-+ free_netdev(port_netdev);
-+ goto err_takedown;
-+ }
-+
-+ rtnl_lock();
-+ err = netdev_master_upper_dev_link(port_netdev, netdev,
-+ NULL, NULL);
-+ if (unlikely(err)) {
-+ dev_err(dev, "netdev_master_upper_dev_link err %d\n",
-+ err);
-+ unregister_netdev(port_netdev);
-+ free_netdev(port_netdev);
-+ rtnl_unlock();
-+ goto err_takedown;
-+ }
-+ rtmsg_ifinfo(RTM_NEWLINK, port_netdev,
-+ IFF_SLAVE, GFP_KERNEL);
-+ rtnl_unlock();
-+
-+ list_add(&port_priv->list, &priv->port_list);
-+ } else {
-+ /* Set MTU limits only on uplink */
-+ port_netdev->min_mtu = EVB_MIN_MTU;
-+ port_netdev->max_mtu = EVB_MAX_MTU;
-+
-+ err = register_netdev(netdev);
-+
-+ if (err < 0) {
-+ dev_err(dev, "register_netdev error %d\n", err);
-+ goto err_takedown;
-+ }
-+ }
-+
-+ port_netdev->ethtool_ops = &evb_port_ethtool_ops;
-+
-+ /* ports are up from init */
-+ rtnl_lock();
-+ err = dev_open(port_netdev);
-+ rtnl_unlock();
-+ if (unlikely(err))
-+ dev_warn(dev, "dev_open err %d\n", err);
-+ }
-+
-+ /* setup irqs */
-+ err = evb_setup_irqs(evb_dev);
-+ if (unlikely(err)) {
-+ dev_warn(dev, "evb_setup_irqs err %d\n", err);
-+ goto err_takedown;
-+ }
-+
-+ dev_info(dev, "probed evb device with %d ports\n",
-+ priv->attr.num_ifs);
-+ return 0;
-+
-+err_takedown:
-+ evb_remove(evb_dev);
-+err_free_cmdport:
-+ fsl_mc_portal_free(priv->mc_io);
-+err_free_netdev:
-+ return err;
-+}
-+
-+static const struct fsl_mc_device_id evb_match_id_table[] = {
-+ {
-+ .vendor = FSL_MC_VENDOR_FREESCALE,
-+ .obj_type = "dpdmux",
-+ },
-+ {}
-+};
-+
-+static struct fsl_mc_driver evb_drv = {
-+ .driver = {
-+ .name = KBUILD_MODNAME,
-+ .owner = THIS_MODULE,
-+ },
-+ .probe = evb_probe,
-+ .remove = evb_remove,
-+ .match_id_table = evb_match_id_table,
-+};
-+
-+module_fsl_mc_driver(evb_drv);
-+
-+MODULE_LICENSE("GPL");
-+MODULE_DESCRIPTION("Layerscape DPAA Edge Virtual Bridge driver (prototype)");
---- /dev/null
-+++ b/drivers/staging/fsl-dpaa2/mac/Kconfig
-@@ -0,0 +1,23 @@
-+config FSL_DPAA2_MAC
-+ tristate "DPAA2 MAC / PHY interface"
-+ depends on FSL_MC_BUS && FSL_DPAA2
-+ select MDIO_BUS_MUX_MMIOREG
-+ select FSL_XGMAC_MDIO
-+ select FIXED_PHY
-+ ---help---
-+ Prototype driver for DPAA2 MAC / PHY interface object.
-+ This driver works as a proxy between phylib including phy drivers and
-+ the MC firmware. It receives updates on link state changes from PHY
-+ lib and forwards them to MC and receives interrupt from MC whenever
-+ a request is made to change the link state.
-+
-+
-+config FSL_DPAA2_MAC_NETDEVS
-+ bool "Expose net interfaces for PHYs"
-+ default n
-+ depends on FSL_DPAA2_MAC
-+ ---help---
-+ Exposes macX net interfaces which allow direct control over MACs and
-+ PHYs.
-+ .
-+ Leave disabled if unsure.
---- /dev/null
-+++ b/drivers/staging/fsl-dpaa2/mac/Makefile
-@@ -0,0 +1,10 @@
-+
-+obj-$(CONFIG_FSL_DPAA2_MAC) += dpaa2-mac.o
-+
-+dpaa2-mac-objs := mac.o dpmac.o
-+
-+all:
-+ make -C /lib/modules/$(shell uname -r)/build M=$(PWD) modules
-+
-+clean:
-+ make -C /lib/modules/$(shell uname -r)/build M=$(PWD) clean
---- /dev/null
-+++ b/drivers/staging/fsl-dpaa2/mac/dpmac-cmd.h
-@@ -0,0 +1,172 @@
-+/* Copyright 2013-2016 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of the above-listed copyright holders nor the
-+ * names of any contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
-+ * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
-+ * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
-+ * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE
-+ * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
-+ * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
-+ * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
-+ * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
-+ * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
-+ * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
-+ * POSSIBILITY OF SUCH DAMAGE.
-+ */
-+#ifndef _FSL_DPMAC_CMD_H
-+#define _FSL_DPMAC_CMD_H
-+
-+/* DPMAC Version */
-+#define DPMAC_VER_MAJOR 4
-+#define DPMAC_VER_MINOR 2
-+#define DPMAC_CMD_BASE_VERSION 1
-+#define DPMAC_CMD_ID_OFFSET 4
-+
-+#define DPMAC_CMD(id) (((id) << DPMAC_CMD_ID_OFFSET) | DPMAC_CMD_BASE_VERSION)
-+
-+/* Command IDs */
-+#define DPMAC_CMDID_CLOSE DPMAC_CMD(0x800)
-+#define DPMAC_CMDID_OPEN DPMAC_CMD(0x80c)
-+#define DPMAC_CMDID_CREATE DPMAC_CMD(0x90c)
-+#define DPMAC_CMDID_DESTROY DPMAC_CMD(0x98c)
-+#define DPMAC_CMDID_GET_API_VERSION DPMAC_CMD(0xa0c)
-+
-+#define DPMAC_CMDID_GET_ATTR DPMAC_CMD(0x004)
-+#define DPMAC_CMDID_RESET DPMAC_CMD(0x005)
-+
-+#define DPMAC_CMDID_SET_IRQ_ENABLE DPMAC_CMD(0x012)
-+#define DPMAC_CMDID_GET_IRQ_ENABLE DPMAC_CMD(0x013)
-+#define DPMAC_CMDID_SET_IRQ_MASK DPMAC_CMD(0x014)
-+#define DPMAC_CMDID_GET_IRQ_MASK DPMAC_CMD(0x015)
-+#define DPMAC_CMDID_GET_IRQ_STATUS DPMAC_CMD(0x016)
-+#define DPMAC_CMDID_CLEAR_IRQ_STATUS DPMAC_CMD(0x017)
-+
-+#define DPMAC_CMDID_GET_LINK_CFG DPMAC_CMD(0x0c2)
-+#define DPMAC_CMDID_SET_LINK_STATE DPMAC_CMD(0x0c3)
-+#define DPMAC_CMDID_GET_COUNTER DPMAC_CMD(0x0c4)
-+
-+#define DPMAC_CMDID_SET_PORT_MAC_ADDR DPMAC_CMD(0x0c5)
-+
-+/* Macros for accessing command fields smaller than 1byte */
-+#define DPMAC_MASK(field) \
-+ GENMASK(DPMAC_##field##_SHIFT + DPMAC_##field##_SIZE - 1, \
-+ DPMAC_##field##_SHIFT)
-+#define dpmac_set_field(var, field, val) \
-+ ((var) |= (((val) << DPMAC_##field##_SHIFT) & DPMAC_MASK(field)))
-+#define dpmac_get_field(var, field) \
-+ (((var) & DPMAC_MASK(field)) >> DPMAC_##field##_SHIFT)
-+
-+struct dpmac_cmd_open {
-+ u32 dpmac_id;
-+};
-+
-+struct dpmac_cmd_create {
-+ u32 mac_id;
-+};
-+
-+struct dpmac_cmd_destroy {
-+ u32 dpmac_id;
-+};
-+
-+struct dpmac_cmd_set_irq_enable {
-+ u8 enable;
-+ u8 pad[3];
-+ u8 irq_index;
-+};
-+
-+struct dpmac_cmd_get_irq_enable {
-+ u32 pad;
-+ u8 irq_index;
-+};
-+
-+struct dpmac_rsp_get_irq_enable {
-+ u8 enabled;
-+};
-+
-+struct dpmac_cmd_set_irq_mask {
-+ u32 mask;
-+ u8 irq_index;
-+};
-+
-+struct dpmac_cmd_get_irq_mask {
-+ u32 pad;
-+ u8 irq_index;
-+};
-+
-+struct dpmac_rsp_get_irq_mask {
-+ u32 mask;
-+};
-+
-+struct dpmac_cmd_get_irq_status {
-+ u32 status;
-+ u8 irq_index;
-+};
-+
-+struct dpmac_rsp_get_irq_status {
-+ u32 status;
-+};
-+
-+struct dpmac_cmd_clear_irq_status {
-+ u32 status;
-+ u8 irq_index;
-+};
-+
-+struct dpmac_rsp_get_attributes {
-+ u8 eth_if;
-+ u8 link_type;
-+ u16 id;
-+ u32 max_rate;
-+};
-+
-+struct dpmac_rsp_get_link_cfg {
-+ u64 options;
-+ u32 rate;
-+};
-+
-+#define DPMAC_STATE_SIZE 1
-+#define DPMAC_STATE_SHIFT 0
-+
-+struct dpmac_cmd_set_link_state {
-+ u64 options;
-+ u32 rate;
-+ u32 pad;
-+ /* only least significant bit is valid */
-+ u8 up;
-+};
-+
-+struct dpmac_cmd_get_counter {
-+ u8 type;
-+};
-+
-+struct dpmac_rsp_get_counter {
-+ u64 pad;
-+ u64 counter;
-+};
-+
-+struct dpmac_rsp_get_api_version {
-+ u16 major;
-+ u16 minor;
-+};
-+
-+struct dpmac_cmd_set_port_mac_addr {
-+ u8 pad[2];
-+ u8 addr[6];
-+};
-+
-+#endif /* _FSL_DPMAC_CMD_H */
---- /dev/null
-+++ b/drivers/staging/fsl-dpaa2/mac/dpmac.c
-@@ -0,0 +1,619 @@
-+/* Copyright 2013-2016 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of the above-listed copyright holders nor the
-+ * names of any contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
-+ * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
-+ * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
-+ * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE
-+ * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
-+ * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
-+ * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
-+ * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
-+ * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
-+ * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
-+ * POSSIBILITY OF SUCH DAMAGE.
-+ */
-+#include <linux/fsl/mc.h>
-+#include "dpmac.h"
-+#include "dpmac-cmd.h"
-+
-+/**
-+ * dpmac_open() - Open a control session for the specified object.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @dpmac_id: DPMAC unique ID
-+ * @token: Returned token; use in subsequent API calls
-+ *
-+ * This function can be used to open a control session for an
-+ * already created object; an object may have been declared in
-+ * the DPL or by calling the dpmac_create function.
-+ * This function returns a unique authentication token,
-+ * associated with the specific object ID and the specific MC
-+ * portal; this token must be used in all subsequent commands for
-+ * this specific object
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpmac_open(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ int dpmac_id,
-+ u16 *token)
-+{
-+ struct dpmac_cmd_open *cmd_params;
-+ struct fsl_mc_command cmd = { 0 };
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPMAC_CMDID_OPEN,
-+ cmd_flags,
-+ 0);
-+ cmd_params = (struct dpmac_cmd_open *)cmd.params;
-+ cmd_params->dpmac_id = cpu_to_le32(dpmac_id);
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ *token = mc_cmd_hdr_read_token(&cmd);
-+
-+ return err;
-+}
-+
-+/**
-+ * dpmac_close() - Close the control session of the object
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPMAC object
-+ *
-+ * After this function is called, no further operations are
-+ * allowed on the object without opening a new control session.
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpmac_close(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPMAC_CMDID_CLOSE, cmd_flags,
-+ token);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpmac_create() - Create the DPMAC object.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @dprc_token: Parent container token; '0' for default container
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @cfg: Configuration structure
-+ * @obj_id: Returned object id
-+ *
-+ * Create the DPMAC object, allocate required resources and
-+ * perform required initialization.
-+ *
-+ * The function accepts an authentication token of a parent
-+ * container that this object should be assigned to. The token
-+ * can be '0' so the object will be assigned to the default container.
-+ * The newly created object can be opened with the returned
-+ * object id and using the container's associated tokens and MC portals.
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpmac_create(struct fsl_mc_io *mc_io,
-+ u16 dprc_token,
-+ u32 cmd_flags,
-+ const struct dpmac_cfg *cfg,
-+ u32 *obj_id)
-+{
-+ struct dpmac_cmd_create *cmd_params;
-+ struct fsl_mc_command cmd = { 0 };
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPMAC_CMDID_CREATE,
-+ cmd_flags,
-+ dprc_token);
-+ cmd_params = (struct dpmac_cmd_create *)cmd.params;
-+ cmd_params->mac_id = cpu_to_le32(cfg->mac_id);
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ *obj_id = mc_cmd_read_object_id(&cmd);
-+
-+ return 0;
-+}
-+
-+/**
-+ * dpmac_destroy() - Destroy the DPMAC object and release all its resources.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @dprc_token: Parent container token; '0' for default container
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @object_id: The object id; it must be a valid id within the container that
-+ * created this object;
-+ *
-+ * The function accepts the authentication token of the parent container that
-+ * created the object (not the one that currently owns the object). The object
-+ * is searched within parent using the provided 'object_id'.
-+ * All tokens to the object must be closed before calling destroy.
-+ *
-+ * Return: '0' on Success; error code otherwise.
-+ */
-+int dpmac_destroy(struct fsl_mc_io *mc_io,
-+ u16 dprc_token,
-+ u32 cmd_flags,
-+ u32 object_id)
-+{
-+ struct dpmac_cmd_destroy *cmd_params;
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPMAC_CMDID_DESTROY,
-+ cmd_flags,
-+ dprc_token);
-+ cmd_params = (struct dpmac_cmd_destroy *)cmd.params;
-+ cmd_params->dpmac_id = cpu_to_le32(object_id);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpmac_set_irq_enable() - Set overall interrupt state.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPMAC object
-+ * @irq_index: The interrupt index to configure
-+ * @en: Interrupt state - enable = 1, disable = 0
-+ *
-+ * Allows GPP software to control when interrupts are generated.
-+ * Each interrupt can have up to 32 causes. The enable/disable control's the
-+ * overall interrupt state. if the interrupt is disabled no causes will cause
-+ * an interrupt.
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpmac_set_irq_enable(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 irq_index,
-+ u8 en)
-+{
-+ struct dpmac_cmd_set_irq_enable *cmd_params;
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPMAC_CMDID_SET_IRQ_ENABLE,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpmac_cmd_set_irq_enable *)cmd.params;
-+ cmd_params->irq_index = irq_index;
-+ cmd_params->enable = en;
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpmac_get_irq_enable() - Get overall interrupt state
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPMAC object
-+ * @irq_index: The interrupt index to configure
-+ * @en: Returned interrupt state - enable = 1, disable = 0
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpmac_get_irq_enable(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 irq_index,
-+ u8 *en)
-+{
-+ struct dpmac_cmd_get_irq_enable *cmd_params;
-+ struct dpmac_rsp_get_irq_enable *rsp_params;
-+ struct fsl_mc_command cmd = { 0 };
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPMAC_CMDID_GET_IRQ_ENABLE,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpmac_cmd_get_irq_enable *)cmd.params;
-+ cmd_params->irq_index = irq_index;
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ rsp_params = (struct dpmac_rsp_get_irq_enable *)cmd.params;
-+ *en = rsp_params->enabled;
-+
-+ return 0;
-+}
-+
-+/**
-+ * dpmac_set_irq_mask() - Set interrupt mask.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPMAC object
-+ * @irq_index: The interrupt index to configure
-+ * @mask: Event mask to trigger interrupt;
-+ * each bit:
-+ * 0 = ignore event
-+ * 1 = consider event for asserting IRQ
-+ *
-+ * Every interrupt can have up to 32 causes and the interrupt model supports
-+ * masking/unmasking each cause independently
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpmac_set_irq_mask(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 irq_index,
-+ u32 mask)
-+{
-+ struct dpmac_cmd_set_irq_mask *cmd_params;
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPMAC_CMDID_SET_IRQ_MASK,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpmac_cmd_set_irq_mask *)cmd.params;
-+ cmd_params->mask = cpu_to_le32(mask);
-+ cmd_params->irq_index = irq_index;
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpmac_get_irq_mask() - Get interrupt mask.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPMAC object
-+ * @irq_index: The interrupt index to configure
-+ * @mask: Returned event mask to trigger interrupt
-+ *
-+ * Every interrupt can have up to 32 causes and the interrupt model supports
-+ * masking/unmasking each cause independently
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpmac_get_irq_mask(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 irq_index,
-+ u32 *mask)
-+{
-+ struct dpmac_cmd_get_irq_mask *cmd_params;
-+ struct dpmac_rsp_get_irq_mask *rsp_params;
-+ struct fsl_mc_command cmd = { 0 };
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPMAC_CMDID_GET_IRQ_MASK,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpmac_cmd_get_irq_mask *)cmd.params;
-+ cmd_params->irq_index = irq_index;
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ rsp_params = (struct dpmac_rsp_get_irq_mask *)cmd.params;
-+ *mask = le32_to_cpu(rsp_params->mask);
-+
-+ return 0;
-+}
-+
-+/**
-+ * dpmac_get_irq_status() - Get the current status of any pending interrupts.
-+ *
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPMAC object
-+ * @irq_index: The interrupt index to configure
-+ * @status: Returned interrupts status - one bit per cause:
-+ * 0 = no interrupt pending
-+ * 1 = interrupt pending
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpmac_get_irq_status(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 irq_index,
-+ u32 *status)
-+{
-+ struct dpmac_cmd_get_irq_status *cmd_params;
-+ struct dpmac_rsp_get_irq_status *rsp_params;
-+ struct fsl_mc_command cmd = { 0 };
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPMAC_CMDID_GET_IRQ_STATUS,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpmac_cmd_get_irq_status *)cmd.params;
-+ cmd_params->status = cpu_to_le32(*status);
-+ cmd_params->irq_index = irq_index;
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ rsp_params = (struct dpmac_rsp_get_irq_status *)cmd.params;
-+ *status = le32_to_cpu(rsp_params->status);
-+
-+ return 0;
-+}
-+
-+/**
-+ * dpmac_clear_irq_status() - Clear a pending interrupt's status
-+ *
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPMAC object
-+ * @irq_index: The interrupt index to configure
-+ * @status: Bits to clear (W1C) - one bit per cause:
-+ * 0 = don't change
-+ * 1 = clear status bit
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpmac_clear_irq_status(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 irq_index,
-+ u32 status)
-+{
-+ struct dpmac_cmd_clear_irq_status *cmd_params;
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPMAC_CMDID_CLEAR_IRQ_STATUS,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpmac_cmd_clear_irq_status *)cmd.params;
-+ cmd_params->status = cpu_to_le32(status);
-+ cmd_params->irq_index = irq_index;
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpmac_get_attributes - Retrieve DPMAC attributes.
-+ *
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPMAC object
-+ * @attr: Returned object's attributes
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpmac_get_attributes(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ struct dpmac_attr *attr)
-+{
-+ struct dpmac_rsp_get_attributes *rsp_params;
-+ struct fsl_mc_command cmd = { 0 };
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPMAC_CMDID_GET_ATTR,
-+ cmd_flags,
-+ token);
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ rsp_params = (struct dpmac_rsp_get_attributes *)cmd.params;
-+ attr->eth_if = rsp_params->eth_if;
-+ attr->link_type = rsp_params->link_type;
-+ attr->id = le16_to_cpu(rsp_params->id);
-+ attr->max_rate = le32_to_cpu(rsp_params->max_rate);
-+
-+ return 0;
-+}
-+
-+/**
-+ * dpmac_get_link_cfg() - Get Ethernet link configuration
-+ * @mc_io: Pointer to opaque I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPMAC object
-+ * @cfg: Returned structure with the link configuration
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpmac_get_link_cfg(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ struct dpmac_link_cfg *cfg)
-+{
-+ struct dpmac_rsp_get_link_cfg *rsp_params;
-+ struct fsl_mc_command cmd = { 0 };
-+ int err = 0;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPMAC_CMDID_GET_LINK_CFG,
-+ cmd_flags,
-+ token);
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ rsp_params = (struct dpmac_rsp_get_link_cfg *)cmd.params;
-+ cfg->options = le64_to_cpu(rsp_params->options);
-+ cfg->rate = le32_to_cpu(rsp_params->rate);
-+
-+ return 0;
-+}
-+
-+/**
-+ * dpmac_set_link_state() - Set the Ethernet link status
-+ * @mc_io: Pointer to opaque I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPMAC object
-+ * @link_state: Link state configuration
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpmac_set_link_state(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ struct dpmac_link_state *link_state)
-+{
-+ struct dpmac_cmd_set_link_state *cmd_params;
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPMAC_CMDID_SET_LINK_STATE,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpmac_cmd_set_link_state *)cmd.params;
-+ cmd_params->options = cpu_to_le64(link_state->options);
-+ cmd_params->rate = cpu_to_le32(link_state->rate);
-+ cmd_params->up = dpmac_get_field(link_state->up, STATE);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpmac_get_counter() - Read a specific DPMAC counter
-+ * @mc_io: Pointer to opaque I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPMAC object
-+ * @type: The requested counter
-+ * @counter: Returned counter value
-+ *
-+ * Return: The requested counter; '0' otherwise.
-+ */
-+int dpmac_get_counter(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ enum dpmac_counter type,
-+ u64 *counter)
-+{
-+ struct dpmac_cmd_get_counter *dpmac_cmd;
-+ struct dpmac_rsp_get_counter *dpmac_rsp;
-+ struct fsl_mc_command cmd = { 0 };
-+ int err = 0;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPMAC_CMDID_GET_COUNTER,
-+ cmd_flags,
-+ token);
-+ dpmac_cmd = (struct dpmac_cmd_get_counter *)cmd.params;
-+ dpmac_cmd->type = type;
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ dpmac_rsp = (struct dpmac_rsp_get_counter *)cmd.params;
-+ *counter = le64_to_cpu(dpmac_rsp->counter);
-+
-+ return 0;
-+}
-+
-+/* untested */
-+int dpmac_set_port_mac_addr(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ const u8 addr[6])
-+{
-+ struct dpmac_cmd_set_port_mac_addr *dpmac_cmd;
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPMAC_CMDID_SET_PORT_MAC_ADDR,
-+ cmd_flags,
-+ token);
-+ dpmac_cmd = (struct dpmac_cmd_set_port_mac_addr *)cmd.params;
-+ dpmac_cmd->addr[0] = addr[5];
-+ dpmac_cmd->addr[1] = addr[4];
-+ dpmac_cmd->addr[2] = addr[3];
-+ dpmac_cmd->addr[3] = addr[2];
-+ dpmac_cmd->addr[4] = addr[1];
-+ dpmac_cmd->addr[5] = addr[0];
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpmac_get_api_version() - Get Data Path MAC version
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @major_ver: Major version of data path mac API
-+ * @minor_ver: Minor version of data path mac API
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpmac_get_api_version(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 *major_ver,
-+ u16 *minor_ver)
-+{
-+ struct dpmac_rsp_get_api_version *rsp_params;
-+ struct fsl_mc_command cmd = { 0 };
-+ int err;
-+
-+ cmd.header = mc_encode_cmd_header(DPMAC_CMDID_GET_API_VERSION,
-+ cmd_flags,
-+ 0);
-+
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ rsp_params = (struct dpmac_rsp_get_api_version *)cmd.params;
-+ *major_ver = le16_to_cpu(rsp_params->major);
-+ *minor_ver = le16_to_cpu(rsp_params->minor);
-+
-+ return 0;
-+}
---- /dev/null
-+++ b/drivers/staging/fsl-dpaa2/mac/dpmac.h
-@@ -0,0 +1,342 @@
-+/* Copyright 2013-2016 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of the above-listed copyright holders nor the
-+ * names of any contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
-+ * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
-+ * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
-+ * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE
-+ * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
-+ * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
-+ * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
-+ * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
-+ * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
-+ * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
-+ * POSSIBILITY OF SUCH DAMAGE.
-+ */
-+#ifndef __FSL_DPMAC_H
-+#define __FSL_DPMAC_H
-+
-+/* Data Path MAC API
-+ * Contains initialization APIs and runtime control APIs for DPMAC
-+ */
-+
-+struct fsl_mc_io;
-+
-+int dpmac_open(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ int dpmac_id,
-+ u16 *token);
-+
-+int dpmac_close(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token);
-+
-+/**
-+ * enum dpmac_link_type - DPMAC link type
-+ * @DPMAC_LINK_TYPE_NONE: No link
-+ * @DPMAC_LINK_TYPE_FIXED: Link is fixed type
-+ * @DPMAC_LINK_TYPE_PHY: Link by PHY ID
-+ * @DPMAC_LINK_TYPE_BACKPLANE: Backplane link type
-+ */
-+enum dpmac_link_type {
-+ DPMAC_LINK_TYPE_NONE,
-+ DPMAC_LINK_TYPE_FIXED,
-+ DPMAC_LINK_TYPE_PHY,
-+ DPMAC_LINK_TYPE_BACKPLANE
-+};
-+
-+/**
-+ * enum dpmac_eth_if - DPMAC Ethrnet interface
-+ * @DPMAC_ETH_IF_MII: MII interface
-+ * @DPMAC_ETH_IF_RMII: RMII interface
-+ * @DPMAC_ETH_IF_SMII: SMII interface
-+ * @DPMAC_ETH_IF_GMII: GMII interface
-+ * @DPMAC_ETH_IF_RGMII: RGMII interface
-+ * @DPMAC_ETH_IF_SGMII: SGMII interface
-+ * @DPMAC_ETH_IF_QSGMII: QSGMII interface
-+ * @DPMAC_ETH_IF_XAUI: XAUI interface
-+ * @DPMAC_ETH_IF_XFI: XFI interface
-+ */
-+enum dpmac_eth_if {
-+ DPMAC_ETH_IF_MII,
-+ DPMAC_ETH_IF_RMII,
-+ DPMAC_ETH_IF_SMII,
-+ DPMAC_ETH_IF_GMII,
-+ DPMAC_ETH_IF_RGMII,
-+ DPMAC_ETH_IF_SGMII,
-+ DPMAC_ETH_IF_QSGMII,
-+ DPMAC_ETH_IF_XAUI,
-+ DPMAC_ETH_IF_XFI
-+};
-+
-+/**
-+ * struct dpmac_cfg - Structure representing DPMAC configuration
-+ * @mac_id: Represents the Hardware MAC ID; in case of multiple WRIOP,
-+ * the MAC IDs are continuous.
-+ * For example: 2 WRIOPs, 16 MACs in each:
-+ * MAC IDs for the 1st WRIOP: 1-16,
-+ * MAC IDs for the 2nd WRIOP: 17-32.
-+ */
-+struct dpmac_cfg {
-+ u16 mac_id;
-+};
-+
-+int dpmac_create(struct fsl_mc_io *mc_io,
-+ u16 dprc_token,
-+ u32 cmd_flags,
-+ const struct dpmac_cfg *cfg,
-+ u32 *obj_id);
-+
-+int dpmac_destroy(struct fsl_mc_io *mc_io,
-+ u16 dprc_token,
-+ u32 cmd_flags,
-+ u32 object_id);
-+
-+/**
-+ * DPMAC IRQ Index and Events
-+ */
-+
-+/**
-+ * IRQ index
-+ */
-+#define DPMAC_IRQ_INDEX 0
-+/**
-+ * IRQ event - indicates a change in link state
-+ */
-+#define DPMAC_IRQ_EVENT_LINK_CFG_REQ 0x00000001
-+/**
-+ * IRQ event - Indicates that the link state changed
-+ */
-+#define DPMAC_IRQ_EVENT_LINK_CHANGED 0x00000002
-+
-+int dpmac_set_irq_enable(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 irq_index,
-+ u8 en);
-+
-+int dpmac_get_irq_enable(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 irq_index,
-+ u8 *en);
-+
-+int dpmac_set_irq_mask(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 irq_index,
-+ u32 mask);
-+
-+int dpmac_get_irq_mask(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 irq_index,
-+ u32 *mask);
-+
-+int dpmac_get_irq_status(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 irq_index,
-+ u32 *status);
-+
-+int dpmac_clear_irq_status(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ u8 irq_index,
-+ u32 status);
-+
-+/**
-+ * struct dpmac_attr - Structure representing DPMAC attributes
-+ * @id: DPMAC object ID
-+ * @max_rate: Maximum supported rate - in Mbps
-+ * @eth_if: Ethernet interface
-+ * @link_type: link type
-+ */
-+struct dpmac_attr {
-+ u16 id;
-+ u32 max_rate;
-+ enum dpmac_eth_if eth_if;
-+ enum dpmac_link_type link_type;
-+};
-+
-+int dpmac_get_attributes(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ struct dpmac_attr *attr);
-+
-+/**
-+ * DPMAC link configuration/state options
-+ */
-+
-+/**
-+ * Enable auto-negotiation
-+ */
-+#define DPMAC_LINK_OPT_AUTONEG 0x0000000000000001ULL
-+/**
-+ * Enable half-duplex mode
-+ */
-+#define DPMAC_LINK_OPT_HALF_DUPLEX 0x0000000000000002ULL
-+/**
-+ * Enable pause frames
-+ */
-+#define DPMAC_LINK_OPT_PAUSE 0x0000000000000004ULL
-+/**
-+ * Enable a-symmetric pause frames
-+ */
-+#define DPMAC_LINK_OPT_ASYM_PAUSE 0x0000000000000008ULL
-+
-+/**
-+ * struct dpmac_link_cfg - Structure representing DPMAC link configuration
-+ * @rate: Link's rate - in Mbps
-+ * @options: Enable/Disable DPMAC link cfg features (bitmap)
-+ */
-+struct dpmac_link_cfg {
-+ u32 rate;
-+ u64 options;
-+};
-+
-+int dpmac_get_link_cfg(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ struct dpmac_link_cfg *cfg);
-+
-+/**
-+ * struct dpmac_link_state - DPMAC link configuration request
-+ * @rate: Rate in Mbps
-+ * @options: Enable/Disable DPMAC link cfg features (bitmap)
-+ * @up: Link state
-+ */
-+struct dpmac_link_state {
-+ u32 rate;
-+ u64 options;
-+ int up;
-+};
-+
-+int dpmac_set_link_state(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ struct dpmac_link_state *link_state);
-+
-+/**
-+ * enum dpmac_counter - DPMAC counter types
-+ * @DPMAC_CNT_ING_FRAME_64: counts 64-bytes frames, good or bad.
-+ * @DPMAC_CNT_ING_FRAME_127: counts 65- to 127-bytes frames, good or bad.
-+ * @DPMAC_CNT_ING_FRAME_255: counts 128- to 255-bytes frames, good or bad.
-+ * @DPMAC_CNT_ING_FRAME_511: counts 256- to 511-bytes frames, good or bad.
-+ * @DPMAC_CNT_ING_FRAME_1023: counts 512- to 1023-bytes frames, good or bad.
-+ * @DPMAC_CNT_ING_FRAME_1518: counts 1024- to 1518-bytes frames, good or bad.
-+ * @DPMAC_CNT_ING_FRAME_1519_MAX: counts 1519-bytes frames and larger
-+ * (up to max frame length specified),
-+ * good or bad.
-+ * @DPMAC_CNT_ING_FRAG: counts frames which are shorter than 64 bytes received
-+ * with a wrong CRC
-+ * @DPMAC_CNT_ING_JABBER: counts frames longer than the maximum frame length
-+ * specified, with a bad frame check sequence.
-+ * @DPMAC_CNT_ING_FRAME_DISCARD: counts dropped frames due to internal errors.
-+ * Occurs when a receive FIFO overflows.
-+ * Includes also frames truncated as a result of
-+ * the receive FIFO overflow.
-+ * @DPMAC_CNT_ING_ALIGN_ERR: counts frames with an alignment error
-+ * (optional used for wrong SFD).
-+ * @DPMAC_CNT_EGR_UNDERSIZED: counts frames transmitted that was less than 64
-+ * bytes long with a good CRC.
-+ * @DPMAC_CNT_ING_OVERSIZED: counts frames longer than the maximum frame length
-+ * specified, with a good frame check sequence.
-+ * @DPMAC_CNT_ING_VALID_PAUSE_FRAME: counts valid pause frames (regular and PFC)
-+ * @DPMAC_CNT_EGR_VALID_PAUSE_FRAME: counts valid pause frames transmitted
-+ * (regular and PFC).
-+ * @DPMAC_CNT_ING_BYTE: counts bytes received except preamble for all valid
-+ * frames and valid pause frames.
-+ * @DPMAC_CNT_ING_MCAST_FRAME: counts received multicast frames.
-+ * @DPMAC_CNT_ING_BCAST_FRAME: counts received broadcast frames.
-+ * @DPMAC_CNT_ING_ALL_FRAME: counts each good or bad frames received.
-+ * @DPMAC_CNT_ING_UCAST_FRAME: counts received unicast frames.
-+ * @DPMAC_CNT_ING_ERR_FRAME: counts frames received with an error
-+ * (except for undersized/fragment frame).
-+ * @DPMAC_CNT_EGR_BYTE: counts bytes transmitted except preamble for all valid
-+ * frames and valid pause frames transmitted.
-+ * @DPMAC_CNT_EGR_MCAST_FRAME: counts transmitted multicast frames.
-+ * @DPMAC_CNT_EGR_BCAST_FRAME: counts transmitted broadcast frames.
-+ * @DPMAC_CNT_EGR_UCAST_FRAME: counts transmitted unicast frames.
-+ * @DPMAC_CNT_EGR_ERR_FRAME: counts frames transmitted with an error.
-+ * @DPMAC_CNT_ING_GOOD_FRAME: counts frames received without error, including
-+ * pause frames.
-+ * @DPMAC_CNT_ENG_GOOD_FRAME: counts frames transmitted without error, including
-+ * pause frames.
-+ */
-+enum dpmac_counter {
-+ DPMAC_CNT_ING_FRAME_64,
-+ DPMAC_CNT_ING_FRAME_127,
-+ DPMAC_CNT_ING_FRAME_255,
-+ DPMAC_CNT_ING_FRAME_511,
-+ DPMAC_CNT_ING_FRAME_1023,
-+ DPMAC_CNT_ING_FRAME_1518,
-+ DPMAC_CNT_ING_FRAME_1519_MAX,
-+ DPMAC_CNT_ING_FRAG,
-+ DPMAC_CNT_ING_JABBER,
-+ DPMAC_CNT_ING_FRAME_DISCARD,
-+ DPMAC_CNT_ING_ALIGN_ERR,
-+ DPMAC_CNT_EGR_UNDERSIZED,
-+ DPMAC_CNT_ING_OVERSIZED,
-+ DPMAC_CNT_ING_VALID_PAUSE_FRAME,
-+ DPMAC_CNT_EGR_VALID_PAUSE_FRAME,
-+ DPMAC_CNT_ING_BYTE,
-+ DPMAC_CNT_ING_MCAST_FRAME,
-+ DPMAC_CNT_ING_BCAST_FRAME,
-+ DPMAC_CNT_ING_ALL_FRAME,
-+ DPMAC_CNT_ING_UCAST_FRAME,
-+ DPMAC_CNT_ING_ERR_FRAME,
-+ DPMAC_CNT_EGR_BYTE,
-+ DPMAC_CNT_EGR_MCAST_FRAME,
-+ DPMAC_CNT_EGR_BCAST_FRAME,
-+ DPMAC_CNT_EGR_UCAST_FRAME,
-+ DPMAC_CNT_EGR_ERR_FRAME,
-+ DPMAC_CNT_ING_GOOD_FRAME,
-+ DPMAC_CNT_ENG_GOOD_FRAME
-+};
-+
-+int dpmac_get_counter(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ enum dpmac_counter type,
-+ u64 *counter);
-+
-+/**
-+ * dpmac_set_port_mac_addr() - Set a MAC address associated with the physical
-+ * port. This is not used for filtering, MAC is always in
-+ * promiscuous mode, it is passed to DPNIs through DPNI API for
-+ * application used.
-+ * @mc_io: Pointer to opaque I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPMAC object
-+ * @addr: MAC address to set
-+ *
-+ * Return: The requested counter; '0' otherwise.
-+ */
-+int dpmac_set_port_mac_addr(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 token,
-+ const u8 addr[6]);
-+
-+int dpmac_get_api_version(struct fsl_mc_io *mc_io,
-+ u32 cmd_flags,
-+ u16 *major_ver,
-+ u16 *minor_ver);
-+
-+#endif /* __FSL_DPMAC_H */
---- /dev/null
-+++ b/drivers/staging/fsl-dpaa2/mac/mac.c
-@@ -0,0 +1,673 @@
-+/* Copyright 2015 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#include <linux/module.h>
-+
-+#include <linux/netdevice.h>
-+#include <linux/etherdevice.h>
-+#include <linux/msi.h>
-+#include <linux/rtnetlink.h>
-+#include <linux/if_vlan.h>
-+
-+#include <uapi/linux/if_bridge.h>
-+#include <net/netlink.h>
-+
-+#include <linux/of.h>
-+#include <linux/of_mdio.h>
-+#include <linux/of_net.h>
-+#include <linux/phy.h>
-+#include <linux/phy_fixed.h>
-+
-+#include <linux/fsl/mc.h>
-+
-+#include "dpmac.h"
-+#include "dpmac-cmd.h"
-+
-+struct dpaa2_mac_priv {
-+ struct net_device *netdev;
-+ struct fsl_mc_device *mc_dev;
-+ struct dpmac_attr attr;
-+ struct dpmac_link_state old_state;
-+};
-+
-+/* TODO: fix the 10G modes, mapping can't be right:
-+ * XGMII is paralel
-+ * XAUI is serial, using 8b/10b encoding
-+ * XFI is also serial but using 64b/66b encoding
-+ * they can't all map to XGMII...
-+ *
-+ * This must be kept in sync with enum dpmac_eth_if.
-+ */
-+static phy_interface_t dpaa2_mac_iface_mode[] = {
-+ PHY_INTERFACE_MODE_MII, /* DPMAC_ETH_IF_MII */
-+ PHY_INTERFACE_MODE_RMII, /* DPMAC_ETH_IF_RMII */
-+ PHY_INTERFACE_MODE_SMII, /* DPMAC_ETH_IF_SMII */
-+ PHY_INTERFACE_MODE_GMII, /* DPMAC_ETH_IF_GMII */
-+ PHY_INTERFACE_MODE_RGMII, /* DPMAC_ETH_IF_RGMII */
-+ PHY_INTERFACE_MODE_SGMII, /* DPMAC_ETH_IF_SGMII */
-+ PHY_INTERFACE_MODE_QSGMII, /* DPMAC_ETH_IF_QSGMII */
-+ PHY_INTERFACE_MODE_XGMII, /* DPMAC_ETH_IF_XAUI */
-+ PHY_INTERFACE_MODE_XGMII, /* DPMAC_ETH_IF_XFI */
-+};
-+
-+static void dpaa2_mac_link_changed(struct net_device *netdev)
-+{
-+ struct phy_device *phydev;
-+ struct dpmac_link_state state = { 0 };
-+ struct dpaa2_mac_priv *priv = netdev_priv(netdev);
-+ int err;
-+
-+ /* the PHY just notified us of link state change */
-+ phydev = netdev->phydev;
-+
-+ state.up = !!phydev->link;
-+ if (phydev->link) {
-+ state.rate = phydev->speed;
-+
-+ if (!phydev->duplex)
-+ state.options |= DPMAC_LINK_OPT_HALF_DUPLEX;
-+ if (phydev->autoneg)
-+ state.options |= DPMAC_LINK_OPT_AUTONEG;
-+
-+ netif_carrier_on(netdev);
-+ } else {
-+ netif_carrier_off(netdev);
-+ }
-+
-+ if (priv->old_state.up != state.up ||
-+ priv->old_state.rate != state.rate ||
-+ priv->old_state.options != state.options) {
-+ priv->old_state = state;
-+ phy_print_status(phydev);
-+ }
-+
-+ /* We must interrogate MC at all times, because we don't know
-+ * when and whether a potential DPNI may have read the link state.
-+ */
-+ err = dpmac_set_link_state(priv->mc_dev->mc_io, 0,
-+ priv->mc_dev->mc_handle, &state);
-+ if (unlikely(err))
-+ dev_err(&priv->mc_dev->dev, "dpmac_set_link_state: %d\n", err);
-+}
-+
-+static int dpaa2_mac_open(struct net_device *netdev)
-+{
-+ /* start PHY state machine */
-+ phy_start(netdev->phydev);
-+
-+ return 0;
-+}
-+
-+static int dpaa2_mac_stop(struct net_device *netdev)
-+{
-+ if (!netdev->phydev)
-+ goto done;
-+
-+ /* stop PHY state machine */
-+ phy_stop(netdev->phydev);
-+
-+ /* signal link down to firmware */
-+ netdev->phydev->link = 0;
-+ dpaa2_mac_link_changed(netdev);
-+
-+done:
-+ return 0;
-+}
-+
-+#ifdef CONFIG_FSL_DPAA2_MAC_NETDEVS
-+static netdev_tx_t dpaa2_mac_drop_frame(struct sk_buff *skb,
-+ struct net_device *dev)
-+{
-+ /* we don't support I/O for now, drop the frame */
-+ dev_kfree_skb_any(skb);
-+ return NETDEV_TX_OK;
-+}
-+
-+static int dpaa2_mac_get_link_ksettings(struct net_device *netdev,
-+ struct ethtool_link_ksettings *ks)
-+{
-+ phy_ethtool_ksettings_get(netdev->phydev, ks);
-+
-+ return 0;
-+}
-+
-+static int dpaa2_mac_set_link_ksettings(struct net_device *netdev,
-+ const struct ethtool_link_ksettings *ks)
-+{
-+ return phy_ethtool_ksettings_set(netdev->phydev, ks);
-+}
-+
-+static struct rtnl_link_stats64 *dpaa2_mac_get_stats(struct net_device *netdev,
-+ struct rtnl_link_stats64 *storage)
-+{
-+ struct dpaa2_mac_priv *priv = netdev_priv(netdev);
-+ u64 tmp;
-+ int err;
-+
-+ err = dpmac_get_counter(priv->mc_dev->mc_io, 0, priv->mc_dev->mc_handle,
-+ DPMAC_CNT_EGR_MCAST_FRAME,
-+ &storage->tx_packets);
-+ if (err)
-+ goto error;
-+ err = dpmac_get_counter(priv->mc_dev->mc_io, 0, priv->mc_dev->mc_handle,
-+ DPMAC_CNT_EGR_BCAST_FRAME, &tmp);
-+ if (err)
-+ goto error;
-+ storage->tx_packets += tmp;
-+ err = dpmac_get_counter(priv->mc_dev->mc_io, 0, priv->mc_dev->mc_handle,
-+ DPMAC_CNT_EGR_UCAST_FRAME, &tmp);
-+ if (err)
-+ goto error;
-+ storage->tx_packets += tmp;
-+
-+ err = dpmac_get_counter(priv->mc_dev->mc_io, 0, priv->mc_dev->mc_handle,
-+ DPMAC_CNT_EGR_UNDERSIZED, &storage->tx_dropped);
-+ if (err)
-+ goto error;
-+ err = dpmac_get_counter(priv->mc_dev->mc_io, 0, priv->mc_dev->mc_handle,
-+ DPMAC_CNT_EGR_BYTE, &storage->tx_bytes);
-+ if (err)
-+ goto error;
-+ err = dpmac_get_counter(priv->mc_dev->mc_io, 0, priv->mc_dev->mc_handle,
-+ DPMAC_CNT_EGR_ERR_FRAME, &storage->tx_errors);
-+ if (err)
-+ goto error;
-+
-+ err = dpmac_get_counter(priv->mc_dev->mc_io, 0, priv->mc_dev->mc_handle,
-+ DPMAC_CNT_ING_ALL_FRAME, &storage->rx_packets);
-+ if (err)
-+ goto error;
-+ err = dpmac_get_counter(priv->mc_dev->mc_io, 0, priv->mc_dev->mc_handle,
-+ DPMAC_CNT_ING_MCAST_FRAME, &storage->multicast);
-+ if (err)
-+ goto error;
-+ err = dpmac_get_counter(priv->mc_dev->mc_io, 0, priv->mc_dev->mc_handle,
-+ DPMAC_CNT_ING_FRAME_DISCARD,
-+ &storage->rx_dropped);
-+ if (err)
-+ goto error;
-+ err = dpmac_get_counter(priv->mc_dev->mc_io, 0, priv->mc_dev->mc_handle,
-+ DPMAC_CNT_ING_ALIGN_ERR, &storage->rx_errors);
-+ if (err)
-+ goto error;
-+ err = dpmac_get_counter(priv->mc_dev->mc_io, 0, priv->mc_dev->mc_handle,
-+ DPMAC_CNT_ING_OVERSIZED, &tmp);
-+ if (err)
-+ goto error;
-+ storage->rx_errors += tmp;
-+ err = dpmac_get_counter(priv->mc_dev->mc_io, 0, priv->mc_dev->mc_handle,
-+ DPMAC_CNT_ING_BYTE, &storage->rx_bytes);
-+ if (err)
-+ goto error;
-+
-+ return storage;
-+error:
-+ netdev_err(netdev, "dpmac_get_counter err %d\n", err);
-+ return storage;
-+}
-+
-+static struct {
-+ enum dpmac_counter id;
-+ char name[ETH_GSTRING_LEN];
-+} dpaa2_mac_counters[] = {
-+ {DPMAC_CNT_ING_ALL_FRAME, "rx all frames"},
-+ {DPMAC_CNT_ING_GOOD_FRAME, "rx frames ok"},
-+ {DPMAC_CNT_ING_ERR_FRAME, "rx frame errors"},
-+ {DPMAC_CNT_ING_FRAME_DISCARD, "rx frame discards"},
-+ {DPMAC_CNT_ING_UCAST_FRAME, "rx u-cast"},
-+ {DPMAC_CNT_ING_BCAST_FRAME, "rx b-cast"},
-+ {DPMAC_CNT_ING_MCAST_FRAME, "rx m-cast"},
-+ {DPMAC_CNT_ING_FRAME_64, "rx 64 bytes"},
-+ {DPMAC_CNT_ING_FRAME_127, "rx 65-127 bytes"},
-+ {DPMAC_CNT_ING_FRAME_255, "rx 128-255 bytes"},
-+ {DPMAC_CNT_ING_FRAME_511, "rx 256-511 bytes"},
-+ {DPMAC_CNT_ING_FRAME_1023, "rx 512-1023 bytes"},
-+ {DPMAC_CNT_ING_FRAME_1518, "rx 1024-1518 bytes"},
-+ {DPMAC_CNT_ING_FRAME_1519_MAX, "rx 1519-max bytes"},
-+ {DPMAC_CNT_ING_FRAG, "rx frags"},
-+ {DPMAC_CNT_ING_JABBER, "rx jabber"},
-+ {DPMAC_CNT_ING_ALIGN_ERR, "rx align errors"},
-+ {DPMAC_CNT_ING_OVERSIZED, "rx oversized"},
-+ {DPMAC_CNT_ING_VALID_PAUSE_FRAME, "rx pause"},
-+ {DPMAC_CNT_ING_BYTE, "rx bytes"},
-+ {DPMAC_CNT_ENG_GOOD_FRAME, "tx frames ok"},
-+ {DPMAC_CNT_EGR_UCAST_FRAME, "tx u-cast"},
-+ {DPMAC_CNT_EGR_MCAST_FRAME, "tx m-cast"},
-+ {DPMAC_CNT_EGR_BCAST_FRAME, "tx b-cast"},
-+ {DPMAC_CNT_EGR_ERR_FRAME, "tx frame errors"},
-+ {DPMAC_CNT_EGR_UNDERSIZED, "tx undersized"},
-+ {DPMAC_CNT_EGR_VALID_PAUSE_FRAME, "tx b-pause"},
-+ {DPMAC_CNT_EGR_BYTE, "tx bytes"},
-+
-+};
-+
-+static void dpaa2_mac_get_strings(struct net_device *netdev,
-+ u32 stringset, u8 *data)
-+{
-+ int i;
-+
-+ switch (stringset) {
-+ case ETH_SS_STATS:
-+ for (i = 0; i < ARRAY_SIZE(dpaa2_mac_counters); i++)
-+ memcpy(data + i * ETH_GSTRING_LEN,
-+ dpaa2_mac_counters[i].name,
-+ ETH_GSTRING_LEN);
-+ break;
-+ }
-+}
-+
-+static void dpaa2_mac_get_ethtool_stats(struct net_device *netdev,
-+ struct ethtool_stats *stats,
-+ u64 *data)
-+{
-+ struct dpaa2_mac_priv *priv = netdev_priv(netdev);
-+ int i;
-+ int err;
-+
-+ for (i = 0; i < ARRAY_SIZE(dpaa2_mac_counters); i++) {
-+ err = dpmac_get_counter(priv->mc_dev->mc_io,
-+ 0,
-+ priv->mc_dev->mc_handle,
-+ dpaa2_mac_counters[i].id, &data[i]);
-+ if (err)
-+ netdev_err(netdev, "dpmac_get_counter[%s] err %d\n",
-+ dpaa2_mac_counters[i].name, err);
-+ }
-+}
-+
-+static int dpaa2_mac_get_sset_count(struct net_device *dev, int sset)
-+{
-+ switch (sset) {
-+ case ETH_SS_STATS:
-+ return ARRAY_SIZE(dpaa2_mac_counters);
-+ default:
-+ return -EOPNOTSUPP;
-+ }
-+}
-+
-+static const struct net_device_ops dpaa2_mac_ndo_ops = {
-+ .ndo_open = &dpaa2_mac_open,
-+ .ndo_stop = &dpaa2_mac_stop,
-+ .ndo_start_xmit = &dpaa2_mac_drop_frame,
-+ .ndo_get_stats64 = &dpaa2_mac_get_stats,
-+};
-+
-+static const struct ethtool_ops dpaa2_mac_ethtool_ops = {
-+ .get_link_ksettings = &dpaa2_mac_get_link_ksettings,
-+ .set_link_ksettings = &dpaa2_mac_set_link_ksettings,
-+ .get_strings = &dpaa2_mac_get_strings,
-+ .get_ethtool_stats = &dpaa2_mac_get_ethtool_stats,
-+ .get_sset_count = &dpaa2_mac_get_sset_count,
-+};
-+#endif /* CONFIG_FSL_DPAA2_MAC_NETDEVS */
-+
-+static void configure_link(struct dpaa2_mac_priv *priv,
-+ struct dpmac_link_cfg *cfg)
-+{
-+ struct phy_device *phydev = priv->netdev->phydev;
-+
-+ if (unlikely(!phydev))
-+ return;
-+
-+ phydev->speed = cfg->rate;
-+ phydev->duplex = !!(cfg->options & DPMAC_LINK_OPT_HALF_DUPLEX);
-+
-+ if (cfg->options & DPMAC_LINK_OPT_AUTONEG) {
-+ phydev->autoneg = 1;
-+ phydev->advertising |= ADVERTISED_Autoneg;
-+ } else {
-+ phydev->autoneg = 0;
-+ phydev->advertising &= ~ADVERTISED_Autoneg;
-+ }
-+
-+ phy_start_aneg(phydev);
-+}
-+
-+static irqreturn_t dpaa2_mac_irq_handler(int irq_num, void *arg)
-+{
-+ struct device *dev = (struct device *)arg;
-+ struct fsl_mc_device *mc_dev = to_fsl_mc_device(dev);
-+ struct dpaa2_mac_priv *priv = dev_get_drvdata(dev);
-+ struct dpmac_link_cfg link_cfg;
-+ u32 status;
-+ int err;
-+
-+ err = dpmac_get_irq_status(mc_dev->mc_io, 0, mc_dev->mc_handle,
-+ DPMAC_IRQ_INDEX, &status);
-+ if (unlikely(err || !status))
-+ return IRQ_NONE;
-+
-+ /* DPNI-initiated link configuration; 'ifconfig up' also calls this */
-+ if (status & DPMAC_IRQ_EVENT_LINK_CFG_REQ) {
-+ err = dpmac_get_link_cfg(mc_dev->mc_io, 0, mc_dev->mc_handle,
-+ &link_cfg);
-+ if (unlikely(err))
-+ goto out;
-+
-+ configure_link(priv, &link_cfg);
-+ }
-+
-+out:
-+ dpmac_clear_irq_status(mc_dev->mc_io, 0, mc_dev->mc_handle,
-+ DPMAC_IRQ_INDEX, status);
-+
-+ return IRQ_HANDLED;
-+}
-+
-+static int setup_irqs(struct fsl_mc_device *mc_dev)
-+{
-+ int err = 0;
-+ struct fsl_mc_device_irq *irq;
-+
-+ err = fsl_mc_allocate_irqs(mc_dev);
-+ if (err) {
-+ dev_err(&mc_dev->dev, "fsl_mc_allocate_irqs err %d\n", err);
-+ return err;
-+ }
-+
-+ irq = mc_dev->irqs[0];
-+ err = devm_request_threaded_irq(&mc_dev->dev, irq->msi_desc->irq,
-+ NULL, &dpaa2_mac_irq_handler,
-+ IRQF_NO_SUSPEND | IRQF_ONESHOT,
-+ dev_name(&mc_dev->dev), &mc_dev->dev);
-+ if (err) {
-+ dev_err(&mc_dev->dev, "devm_request_threaded_irq err %d\n",
-+ err);
-+ goto free_irq;
-+ }
-+
-+ err = dpmac_set_irq_mask(mc_dev->mc_io, 0, mc_dev->mc_handle,
-+ DPMAC_IRQ_INDEX, DPMAC_IRQ_EVENT_LINK_CFG_REQ);
-+ if (err) {
-+ dev_err(&mc_dev->dev, "dpmac_set_irq_mask err %d\n", err);
-+ goto free_irq;
-+ }
-+ err = dpmac_set_irq_enable(mc_dev->mc_io, 0, mc_dev->mc_handle,
-+ DPMAC_IRQ_INDEX, 1);
-+ if (err) {
-+ dev_err(&mc_dev->dev, "dpmac_set_irq_enable err %d\n", err);
-+ goto free_irq;
-+ }
-+
-+ return 0;
-+
-+free_irq:
-+ fsl_mc_free_irqs(mc_dev);
-+
-+ return err;
-+}
-+
-+static void teardown_irqs(struct fsl_mc_device *mc_dev)
-+{
-+ int err;
-+
-+ err = dpmac_set_irq_enable(mc_dev->mc_io, 0, mc_dev->mc_handle,
-+ DPMAC_IRQ_INDEX, 0);
-+ if (err)
-+ dev_err(&mc_dev->dev, "dpmac_set_irq_enable err %d\n", err);
-+
-+ fsl_mc_free_irqs(mc_dev);
-+}
-+
-+static struct device_node *find_dpmac_node(struct device *dev, u16 dpmac_id)
-+{
-+ struct device_node *dpmacs, *dpmac = NULL;
-+ struct device_node *mc_node = dev->of_node;
-+ u32 id;
-+ int err;
-+
-+ dpmacs = of_find_node_by_name(mc_node, "dpmacs");
-+ if (!dpmacs) {
-+ dev_err(dev, "No dpmacs subnode in device-tree\n");
-+ return NULL;
-+ }
-+
-+ while ((dpmac = of_get_next_child(dpmacs, dpmac))) {
-+ err = of_property_read_u32(dpmac, "reg", &id);
-+ if (err)
-+ continue;
-+ if (id == dpmac_id)
-+ return dpmac;
-+ }
-+
-+ return NULL;
-+}
-+
-+static int dpaa2_mac_probe(struct fsl_mc_device *mc_dev)
-+{
-+ struct device *dev;
-+ struct dpaa2_mac_priv *priv = NULL;
-+ struct device_node *phy_node, *dpmac_node;
-+ struct net_device *netdev;
-+ phy_interface_t if_mode;
-+ int err = 0;
-+
-+ dev = &mc_dev->dev;
-+
-+ /* prepare a net_dev structure to make the phy lib API happy */
-+ netdev = alloc_etherdev(sizeof(*priv));
-+ if (!netdev) {
-+ dev_err(dev, "alloc_etherdev error\n");
-+ err = -ENOMEM;
-+ goto err_exit;
-+ }
-+ priv = netdev_priv(netdev);
-+ priv->mc_dev = mc_dev;
-+ priv->netdev = netdev;
-+
-+ SET_NETDEV_DEV(netdev, dev);
-+
-+#ifdef CONFIG_FSL_DPAA2_MAC_NETDEVS
-+ snprintf(netdev->name, IFNAMSIZ, "mac%d", mc_dev->obj_desc.id);
-+#endif
-+
-+ dev_set_drvdata(dev, priv);
-+
-+ /* We may need to issue MC commands while in atomic context */
-+ err = fsl_mc_portal_allocate(mc_dev, FSL_MC_IO_ATOMIC_CONTEXT_PORTAL,
-+ &mc_dev->mc_io);
-+ if (err || !mc_dev->mc_io) {
-+ dev_dbg(dev, "fsl_mc_portal_allocate error: %d\n", err);
-+ err = -EPROBE_DEFER;
-+ goto err_free_netdev;
-+ }
-+
-+ err = dpmac_open(mc_dev->mc_io, 0, mc_dev->obj_desc.id,
-+ &mc_dev->mc_handle);
-+ if (err || !mc_dev->mc_handle) {
-+ dev_err(dev, "dpmac_open error: %d\n", err);
-+ err = -ENODEV;
-+ goto err_free_mcp;
-+ }
-+
-+ err = dpmac_get_attributes(mc_dev->mc_io, 0,
-+ mc_dev->mc_handle, &priv->attr);
-+ if (err) {
-+ dev_err(dev, "dpmac_get_attributes err %d\n", err);
-+ err = -EINVAL;
-+ goto err_close;
-+ }
-+
-+ /* Look up the DPMAC node in the device-tree. */
-+ dpmac_node = find_dpmac_node(dev, priv->attr.id);
-+ if (!dpmac_node) {
-+ dev_err(dev, "No dpmac@%d subnode found.\n", priv->attr.id);
-+ err = -ENODEV;
-+ goto err_close;
-+ }
-+
-+ err = setup_irqs(mc_dev);
-+ if (err) {
-+ err = -EFAULT;
-+ goto err_close;
-+ }
-+
-+#ifdef CONFIG_FSL_DPAA2_MAC_NETDEVS
-+ /* OPTIONAL, register netdev just to make it visible to the user */
-+ netdev->netdev_ops = &dpaa2_mac_ndo_ops;
-+ netdev->ethtool_ops = &dpaa2_mac_ethtool_ops;
-+
-+ /* phy starts up enabled so netdev should be up too */
-+ netdev->flags |= IFF_UP;
-+
-+ err = register_netdev(priv->netdev);
-+ if (err < 0) {
-+ dev_err(dev, "register_netdev error %d\n", err);
-+ err = -ENODEV;
-+ goto err_free_irq;
-+ }
-+#endif /* CONFIG_FSL_DPAA2_MAC_NETDEVS */
-+
-+ /* probe the PHY as a fixed-link if there's a phy-handle defined
-+ * in the device tree
-+ */
-+ phy_node = of_parse_phandle(dpmac_node, "phy-handle", 0);
-+ if (!phy_node) {
-+ goto probe_fixed_link;
-+ }
-+
-+ if (priv->attr.eth_if < ARRAY_SIZE(dpaa2_mac_iface_mode)) {
-+ if_mode = dpaa2_mac_iface_mode[priv->attr.eth_if];
-+ dev_dbg(dev, "\tusing if mode %s for eth_if %d\n",
-+ phy_modes(if_mode), priv->attr.eth_if);
-+ } else {
-+ dev_warn(dev, "Unexpected interface mode %d, will probe as fixed link\n",
-+ priv->attr.eth_if);
-+ goto probe_fixed_link;
-+ }
-+
-+ /* try to connect to the PHY */
-+ netdev->phydev = of_phy_connect(netdev, phy_node,
-+ &dpaa2_mac_link_changed, 0, if_mode);
-+ if (!netdev->phydev) {
-+ /* No need for dev_err(); the kernel's loud enough as it is. */
-+ dev_dbg(dev, "Can't of_phy_connect() now.\n");
-+ /* We might be waiting for the MDIO MUX to probe, so defer
-+ * our own probing.
-+ */
-+ err = -EPROBE_DEFER;
-+ goto err_defer;
-+ }
-+ dev_info(dev, "Connected to %s PHY.\n", phy_modes(if_mode));
-+
-+probe_fixed_link:
-+ if (!netdev->phydev) {
-+ struct fixed_phy_status status = {
-+ .link = 1,
-+ /* fixed-phys don't support 10Gbps speed for now */
-+ .speed = 1000,
-+ .duplex = 1,
-+ };
-+
-+ /* try to register a fixed link phy */
-+ netdev->phydev = fixed_phy_register(PHY_POLL, &status, -1,
-+ NULL);
-+ if (!netdev->phydev || IS_ERR(netdev->phydev)) {
-+ dev_err(dev, "error trying to register fixed PHY\n");
-+ /* So we don't crash unregister_netdev() later on */
-+ netdev->phydev = NULL;
-+ err = -EFAULT;
-+ goto err_no_phy;
-+ }
-+ dev_info(dev, "Registered fixed PHY.\n");
-+ }
-+
-+ dpaa2_mac_open(netdev);
-+
-+ return 0;
-+
-+err_defer:
-+err_no_phy:
-+#ifdef CONFIG_FSL_DPAA2_MAC_NETDEVS
-+ unregister_netdev(netdev);
-+err_free_irq:
-+#endif
-+ teardown_irqs(mc_dev);
-+err_close:
-+ dpmac_close(mc_dev->mc_io, 0, mc_dev->mc_handle);
-+err_free_mcp:
-+ fsl_mc_portal_free(mc_dev->mc_io);
-+err_free_netdev:
-+ free_netdev(netdev);
-+err_exit:
-+ return err;
-+}
-+
-+static int dpaa2_mac_remove(struct fsl_mc_device *mc_dev)
-+{
-+ struct device *dev = &mc_dev->dev;
-+ struct dpaa2_mac_priv *priv = dev_get_drvdata(dev);
-+ struct net_device *netdev = priv->netdev;
-+
-+ dpaa2_mac_stop(netdev);
-+
-+ if (phy_is_pseudo_fixed_link(netdev->phydev))
-+ fixed_phy_unregister(netdev->phydev);
-+ else
-+ phy_disconnect(netdev->phydev);
-+ netdev->phydev = NULL;
-+
-+#ifdef CONFIG_FSL_DPAA2_MAC_NETDEVS
-+ unregister_netdev(priv->netdev);
-+#endif
-+ teardown_irqs(priv->mc_dev);
-+ dpmac_close(priv->mc_dev->mc_io, 0, priv->mc_dev->mc_handle);
-+ fsl_mc_portal_free(priv->mc_dev->mc_io);
-+ free_netdev(priv->netdev);
-+
-+ dev_set_drvdata(dev, NULL);
-+
-+ return 0;
-+}
-+
-+static const struct fsl_mc_device_id dpaa2_mac_match_id_table[] = {
-+ {
-+ .vendor = FSL_MC_VENDOR_FREESCALE,
-+ .obj_type = "dpmac",
-+ },
-+ { .vendor = 0x0 }
-+};
-+MODULE_DEVICE_TABLE(fslmc, dpaa2_mac_match_id_table);
-+
-+static struct fsl_mc_driver dpaa2_mac_drv = {
-+ .driver = {
-+ .name = KBUILD_MODNAME,
-+ .owner = THIS_MODULE,
-+ },
-+ .probe = dpaa2_mac_probe,
-+ .remove = dpaa2_mac_remove,
-+ .match_id_table = dpaa2_mac_match_id_table,
-+};
-+
-+module_fsl_mc_driver(dpaa2_mac_drv);
-+
-+MODULE_LICENSE("GPL");
-+MODULE_DESCRIPTION("DPAA2 PHY proxy interface driver");
---- /dev/null
-+++ b/drivers/staging/fsl-dpaa2/rtc/Makefile
-@@ -0,0 +1,10 @@
-+
-+obj-$(CONFIG_PTP_1588_CLOCK_DPAA2) += dpaa2-rtc.o
-+
-+dpaa2-rtc-objs := rtc.o dprtc.o
-+
-+all:
-+ make -C /lib/modules/$(shell uname -r)/build M=$(PWD) modules
-+
-+clean:
-+ make -C /lib/modules/$(shell uname -r)/build M=$(PWD) clean
---- /dev/null
-+++ b/drivers/staging/fsl-dpaa2/rtc/dprtc-cmd.h
-@@ -0,0 +1,160 @@
-+/* Copyright 2013-2016 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of the above-listed copyright holders nor the
-+ * names of any contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
-+ * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
-+ * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
-+ * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE
-+ * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
-+ * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
-+ * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
-+ * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
-+ * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
-+ * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
-+ * POSSIBILITY OF SUCH DAMAGE.
-+ */
-+#ifndef _FSL_DPRTC_CMD_H
-+#define _FSL_DPRTC_CMD_H
-+
-+/* DPRTC Version */
-+#define DPRTC_VER_MAJOR 2
-+#define DPRTC_VER_MINOR 0
-+
-+/* Command versioning */
-+#define DPRTC_CMD_BASE_VERSION 1
-+#define DPRTC_CMD_ID_OFFSET 4
-+
-+#define DPRTC_CMD(id) (((id) << DPRTC_CMD_ID_OFFSET) | DPRTC_CMD_BASE_VERSION)
-+
-+/* Command IDs */
-+#define DPRTC_CMDID_CLOSE DPRTC_CMD(0x800)
-+#define DPRTC_CMDID_OPEN DPRTC_CMD(0x810)
-+#define DPRTC_CMDID_CREATE DPRTC_CMD(0x910)
-+#define DPRTC_CMDID_DESTROY DPRTC_CMD(0x990)
-+#define DPRTC_CMDID_GET_API_VERSION DPRTC_CMD(0xa10)
-+
-+#define DPRTC_CMDID_ENABLE DPRTC_CMD(0x002)
-+#define DPRTC_CMDID_DISABLE DPRTC_CMD(0x003)
-+#define DPRTC_CMDID_GET_ATTR DPRTC_CMD(0x004)
-+#define DPRTC_CMDID_RESET DPRTC_CMD(0x005)
-+#define DPRTC_CMDID_IS_ENABLED DPRTC_CMD(0x006)
-+
-+#define DPRTC_CMDID_SET_IRQ_ENABLE DPRTC_CMD(0x012)
-+#define DPRTC_CMDID_GET_IRQ_ENABLE DPRTC_CMD(0x013)
-+#define DPRTC_CMDID_SET_IRQ_MASK DPRTC_CMD(0x014)
-+#define DPRTC_CMDID_GET_IRQ_MASK DPRTC_CMD(0x015)
-+#define DPRTC_CMDID_GET_IRQ_STATUS DPRTC_CMD(0x016)
-+#define DPRTC_CMDID_CLEAR_IRQ_STATUS DPRTC_CMD(0x017)
-+
-+#define DPRTC_CMDID_SET_CLOCK_OFFSET DPRTC_CMD(0x1d0)
-+#define DPRTC_CMDID_SET_FREQ_COMPENSATION DPRTC_CMD(0x1d1)
-+#define DPRTC_CMDID_GET_FREQ_COMPENSATION DPRTC_CMD(0x1d2)
-+#define DPRTC_CMDID_GET_TIME DPRTC_CMD(0x1d3)
-+#define DPRTC_CMDID_SET_TIME DPRTC_CMD(0x1d4)
-+#define DPRTC_CMDID_SET_ALARM DPRTC_CMD(0x1d5)
-+#define DPRTC_CMDID_SET_PERIODIC_PULSE DPRTC_CMD(0x1d6)
-+#define DPRTC_CMDID_CLEAR_PERIODIC_PULSE DPRTC_CMD(0x1d7)
-+#define DPRTC_CMDID_SET_EXT_TRIGGER DPRTC_CMD(0x1d8)
-+#define DPRTC_CMDID_CLEAR_EXT_TRIGGER DPRTC_CMD(0x1d9)
-+#define DPRTC_CMDID_GET_EXT_TRIGGER_TIMESTAMP DPRTC_CMD(0x1dA)
-+
-+/* Macros for accessing command fields smaller than 1byte */
-+#define DPRTC_MASK(field) \
-+ GENMASK(DPRTC_##field##_SHIFT + DPRTC_##field##_SIZE - 1, \
-+ DPRTC_##field##_SHIFT)
-+#define dprtc_get_field(var, field) \
-+ (((var) & DPRTC_MASK(field)) >> DPRTC_##field##_SHIFT)
-+
-+#pragma pack(push, 1)
-+struct dprtc_cmd_open {
-+ uint32_t dprtc_id;
-+};
-+
-+struct dprtc_cmd_destroy {
-+ uint32_t object_id;
-+};
-+
-+#define DPRTC_ENABLE_SHIFT 0
-+#define DPRTC_ENABLE_SIZE 1
-+
-+struct dprtc_rsp_is_enabled {
-+ uint8_t en;
-+};
-+
-+struct dprtc_cmd_get_irq {
-+ uint32_t pad;
-+ uint8_t irq_index;
-+};
-+
-+struct dprtc_cmd_set_irq_enable {
-+ uint8_t en;
-+ uint8_t pad[3];
-+ uint8_t irq_index;
-+};
-+
-+struct dprtc_rsp_get_irq_enable {
-+ uint8_t en;
-+};
-+
-+struct dprtc_cmd_set_irq_mask {
-+ uint32_t mask;
-+ uint8_t irq_index;
-+};
-+
-+struct dprtc_rsp_get_irq_mask {
-+ uint32_t mask;
-+};
-+
-+struct dprtc_cmd_get_irq_status {
-+ uint32_t status;
-+ uint8_t irq_index;
-+};
-+
-+struct dprtc_rsp_get_irq_status {
-+ uint32_t status;
-+};
-+
-+struct dprtc_cmd_clear_irq_status {
-+ uint32_t status;
-+ uint8_t irq_index;
-+};
-+
-+struct dprtc_rsp_get_attributes {
-+ uint32_t pad;
-+ uint32_t id;
-+};
-+
-+struct dprtc_cmd_set_clock_offset {
-+ uint64_t offset;
-+};
-+
-+struct dprtc_get_freq_compensation {
-+ uint32_t freq_compensation;
-+};
-+
-+struct dprtc_time {
-+ uint64_t time;
-+};
-+
-+struct dprtc_rsp_get_api_version {
-+ uint16_t major;
-+ uint16_t minor;
-+};
-+#pragma pack(pop)
-+#endif /* _FSL_DPRTC_CMD_H */
---- /dev/null
-+++ b/drivers/staging/fsl-dpaa2/rtc/dprtc.c
-@@ -0,0 +1,746 @@
-+/* Copyright 2013-2016 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of the above-listed copyright holders nor the
-+ * names of any contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
-+ * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
-+ * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
-+ * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE
-+ * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
-+ * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
-+ * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
-+ * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
-+ * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
-+ * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
-+ * POSSIBILITY OF SUCH DAMAGE.
-+ */
-+#include <linux/fsl/mc.h>
-+
-+#include "dprtc.h"
-+#include "dprtc-cmd.h"
-+
-+/**
-+ * dprtc_open() - Open a control session for the specified object.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @dprtc_id: DPRTC unique ID
-+ * @token: Returned token; use in subsequent API calls
-+ *
-+ * This function can be used to open a control session for an
-+ * already created object; an object may have been declared in
-+ * the DPL or by calling the dprtc_create function.
-+ * This function returns a unique authentication token,
-+ * associated with the specific object ID and the specific MC
-+ * portal; this token must be used in all subsequent commands for
-+ * this specific object
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dprtc_open(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ int dprtc_id,
-+ uint16_t *token)
-+{
-+ struct dprtc_cmd_open *cmd_params;
-+ struct fsl_mc_command cmd = { 0 };
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPRTC_CMDID_OPEN,
-+ cmd_flags,
-+ 0);
-+ cmd_params = (struct dprtc_cmd_open *)cmd.params;
-+ cmd_params->dprtc_id = cpu_to_le32(dprtc_id);
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ *token = mc_cmd_hdr_read_token(&cmd);
-+
-+ return err;
-+}
-+
-+/**
-+ * dprtc_close() - Close the control session of the object
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPRTC object
-+ *
-+ * After this function is called, no further operations are
-+ * allowed on the object without opening a new control session.
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dprtc_close(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPRTC_CMDID_CLOSE, cmd_flags,
-+ token);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dprtc_create() - Create the DPRTC object.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @dprc_token: Parent container token; '0' for default container
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @cfg: Configuration structure
-+ * @obj_id: Returned object id
-+ *
-+ * Create the DPRTC object, allocate required resources and
-+ * perform required initialization.
-+ *
-+ * The function accepts an authentication token of a parent
-+ * container that this object should be assigned to. The token
-+ * can be '0' so the object will be assigned to the default container.
-+ * The newly created object can be opened with the returned
-+ * object id and using the container's associated tokens and MC portals.
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dprtc_create(struct fsl_mc_io *mc_io,
-+ uint16_t dprc_token,
-+ uint32_t cmd_flags,
-+ const struct dprtc_cfg *cfg,
-+ uint32_t *obj_id)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ int err;
-+
-+ (void)(cfg); /* unused */
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPRTC_CMDID_CREATE,
-+ cmd_flags,
-+ dprc_token);
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ *obj_id = mc_cmd_read_object_id(&cmd);
-+
-+ return 0;
-+}
-+
-+/**
-+ * dprtc_destroy() - Destroy the DPRTC object and release all its resources.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @dprc_token: Parent container token; '0' for default container
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @object_id: The object id; it must be a valid id within the container that
-+ * created this object;
-+ *
-+ * The function accepts the authentication token of the parent container that
-+ * created the object (not the one that currently owns the object). The object
-+ * is searched within parent using the provided 'object_id'.
-+ * All tokens to the object must be closed before calling destroy.
-+ *
-+ * Return: '0' on Success; error code otherwise.
-+ */
-+int dprtc_destroy(struct fsl_mc_io *mc_io,
-+ uint16_t dprc_token,
-+ uint32_t cmd_flags,
-+ uint32_t object_id)
-+{
-+ struct dprtc_cmd_destroy *cmd_params;
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPRTC_CMDID_DESTROY,
-+ cmd_flags,
-+ dprc_token);
-+ cmd_params = (struct dprtc_cmd_destroy *)cmd.params;
-+ cmd_params->object_id = cpu_to_le32(object_id);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+int dprtc_enable(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPRTC_CMDID_ENABLE, cmd_flags,
-+ token);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+int dprtc_disable(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPRTC_CMDID_DISABLE,
-+ cmd_flags,
-+ token);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+int dprtc_is_enabled(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token,
-+ int *en)
-+{
-+ struct dprtc_rsp_is_enabled *rsp_params;
-+ struct fsl_mc_command cmd = { 0 };
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPRTC_CMDID_IS_ENABLED, cmd_flags,
-+ token);
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ rsp_params = (struct dprtc_rsp_is_enabled *)cmd.params;
-+ *en = dprtc_get_field(rsp_params->en, ENABLE);
-+
-+ return 0;
-+}
-+
-+int dprtc_reset(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPRTC_CMDID_RESET,
-+ cmd_flags,
-+ token);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dprtc_set_irq_enable() - Set overall interrupt state.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPRTC object
-+ * @irq_index: The interrupt index to configure
-+ * @en: Interrupt state - enable = 1, disable = 0
-+ *
-+ * Allows GPP software to control when interrupts are generated.
-+ * Each interrupt can have up to 32 causes. The enable/disable control's the
-+ * overall interrupt state. if the interrupt is disabled no causes will cause
-+ * an interrupt.
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dprtc_set_irq_enable(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token,
-+ uint8_t irq_index,
-+ uint8_t en)
-+{
-+ struct dprtc_cmd_set_irq_enable *cmd_params;
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPRTC_CMDID_SET_IRQ_ENABLE,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dprtc_cmd_set_irq_enable *)cmd.params;
-+ cmd_params->irq_index = irq_index;
-+ cmd_params->en = en;
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dprtc_get_irq_enable() - Get overall interrupt state
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPRTC object
-+ * @irq_index: The interrupt index to configure
-+ * @en: Returned interrupt state - enable = 1, disable = 0
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dprtc_get_irq_enable(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token,
-+ uint8_t irq_index,
-+ uint8_t *en)
-+{
-+ struct dprtc_rsp_get_irq_enable *rsp_params;
-+ struct dprtc_cmd_get_irq *cmd_params;
-+ struct fsl_mc_command cmd = { 0 };
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPRTC_CMDID_GET_IRQ_ENABLE,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dprtc_cmd_get_irq *)cmd.params;
-+ cmd_params->irq_index = irq_index;
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ rsp_params = (struct dprtc_rsp_get_irq_enable *)cmd.params;
-+ *en = rsp_params->en;
-+
-+ return 0;
-+}
-+
-+/**
-+ * dprtc_set_irq_mask() - Set interrupt mask.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPRTC object
-+ * @irq_index: The interrupt index to configure
-+ * @mask: Event mask to trigger interrupt;
-+ * each bit:
-+ * 0 = ignore event
-+ * 1 = consider event for asserting IRQ
-+ *
-+ * Every interrupt can have up to 32 causes and the interrupt model supports
-+ * masking/unmasking each cause independently
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dprtc_set_irq_mask(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token,
-+ uint8_t irq_index,
-+ uint32_t mask)
-+{
-+ struct dprtc_cmd_set_irq_mask *cmd_params;
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPRTC_CMDID_SET_IRQ_MASK,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dprtc_cmd_set_irq_mask *)cmd.params;
-+ cmd_params->mask = cpu_to_le32(mask);
-+ cmd_params->irq_index = irq_index;
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dprtc_get_irq_mask() - Get interrupt mask.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPRTC object
-+ * @irq_index: The interrupt index to configure
-+ * @mask: Returned event mask to trigger interrupt
-+ *
-+ * Every interrupt can have up to 32 causes and the interrupt model supports
-+ * masking/unmasking each cause independently
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dprtc_get_irq_mask(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token,
-+ uint8_t irq_index,
-+ uint32_t *mask)
-+{
-+ struct dprtc_rsp_get_irq_mask *rsp_params;
-+ struct dprtc_cmd_get_irq *cmd_params;
-+ struct fsl_mc_command cmd = { 0 };
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPRTC_CMDID_GET_IRQ_MASK,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dprtc_cmd_get_irq *)cmd.params;
-+ cmd_params->irq_index = irq_index;
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ rsp_params = (struct dprtc_rsp_get_irq_mask *)cmd.params;
-+ *mask = le32_to_cpu(rsp_params->mask);
-+
-+ return 0;
-+}
-+
-+/**
-+ * dprtc_get_irq_status() - Get the current status of any pending interrupts.
-+ *
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPRTC object
-+ * @irq_index: The interrupt index to configure
-+ * @status: Returned interrupts status - one bit per cause:
-+ * 0 = no interrupt pending
-+ * 1 = interrupt pending
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dprtc_get_irq_status(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token,
-+ uint8_t irq_index,
-+ uint32_t *status)
-+{
-+ struct dprtc_cmd_get_irq_status *cmd_params;
-+ struct dprtc_rsp_get_irq_status *rsp_params;
-+ struct fsl_mc_command cmd = { 0 };
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPRTC_CMDID_GET_IRQ_STATUS,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dprtc_cmd_get_irq_status *)cmd.params;
-+ cmd_params->status = cpu_to_le32(*status);
-+ cmd_params->irq_index = irq_index;
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ rsp_params = (struct dprtc_rsp_get_irq_status *)cmd.params;
-+ *status = rsp_params->status;
-+
-+ return 0;
-+}
-+
-+/**
-+ * dprtc_clear_irq_status() - Clear a pending interrupt's status
-+ *
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPRTC object
-+ * @irq_index: The interrupt index to configure
-+ * @status: Bits to clear (W1C) - one bit per cause:
-+ * 0 = don't change
-+ * 1 = clear status bit
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dprtc_clear_irq_status(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token,
-+ uint8_t irq_index,
-+ uint32_t status)
-+{
-+ struct dprtc_cmd_clear_irq_status *cmd_params;
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPRTC_CMDID_CLEAR_IRQ_STATUS,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dprtc_cmd_clear_irq_status *)cmd.params;
-+ cmd_params->irq_index = irq_index;
-+ cmd_params->status = cpu_to_le32(status);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dprtc_get_attributes - Retrieve DPRTC attributes.
-+ *
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPRTC object
-+ * @attr: Returned object's attributes
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dprtc_get_attributes(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token,
-+ struct dprtc_attr *attr)
-+{
-+ struct dprtc_rsp_get_attributes *rsp_params;
-+ struct fsl_mc_command cmd = { 0 };
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPRTC_CMDID_GET_ATTR,
-+ cmd_flags,
-+ token);
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ rsp_params = (struct dprtc_rsp_get_attributes *)cmd.params;
-+ attr->id = le32_to_cpu(rsp_params->id);
-+
-+ return 0;
-+}
-+
-+/**
-+ * dprtc_set_clock_offset() - Sets the clock's offset
-+ * (usually relative to another clock).
-+ *
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPRTC object
-+ * @offset: New clock offset (in nanoseconds).
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dprtc_set_clock_offset(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token,
-+ int64_t offset)
-+{
-+ struct dprtc_cmd_set_clock_offset *cmd_params;
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPRTC_CMDID_SET_CLOCK_OFFSET,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dprtc_cmd_set_clock_offset *)cmd.params;
-+ cmd_params->offset = cpu_to_le64(offset);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dprtc_set_freq_compensation() - Sets a new frequency compensation value.
-+ *
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPRTC object
-+ * @freq_compensation: The new frequency compensation value to set.
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dprtc_set_freq_compensation(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token,
-+ uint32_t freq_compensation)
-+{
-+ struct dprtc_get_freq_compensation *cmd_params;
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPRTC_CMDID_SET_FREQ_COMPENSATION,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dprtc_get_freq_compensation *)cmd.params;
-+ cmd_params->freq_compensation = cpu_to_le32(freq_compensation);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dprtc_get_freq_compensation() - Retrieves the frequency compensation value
-+ *
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPRTC object
-+ * @freq_compensation: Frequency compensation value
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dprtc_get_freq_compensation(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token,
-+ uint32_t *freq_compensation)
-+{
-+ struct dprtc_get_freq_compensation *rsp_params;
-+ struct fsl_mc_command cmd = { 0 };
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPRTC_CMDID_GET_FREQ_COMPENSATION,
-+ cmd_flags,
-+ token);
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ rsp_params = (struct dprtc_get_freq_compensation *)cmd.params;
-+ *freq_compensation = le32_to_cpu(rsp_params->freq_compensation);
-+
-+ return 0;
-+}
-+
-+/**
-+ * dprtc_get_time() - Returns the current RTC time.
-+ *
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPRTC object
-+ * @time: Current RTC time.
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dprtc_get_time(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token,
-+ uint64_t *time)
-+{
-+ struct dprtc_time *rsp_params;
-+ struct fsl_mc_command cmd = { 0 };
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPRTC_CMDID_GET_TIME,
-+ cmd_flags,
-+ token);
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ rsp_params = (struct dprtc_time *)cmd.params;
-+ *time = le64_to_cpu(rsp_params->time);
-+
-+ return 0;
-+}
-+
-+/**
-+ * dprtc_set_time() - Updates current RTC time.
-+ *
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPRTC object
-+ * @time: New RTC time.
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dprtc_set_time(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token,
-+ uint64_t time)
-+{
-+ struct dprtc_time *cmd_params;
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPRTC_CMDID_SET_TIME,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dprtc_time *)cmd.params;
-+ cmd_params->time = cpu_to_le64(time);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dprtc_set_alarm() - Defines and sets alarm.
-+ *
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPRTC object
-+ * @time: In nanoseconds, the time when the alarm
-+ * should go off - must be a multiple of
-+ * 1 microsecond
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dprtc_set_alarm(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token, uint64_t time)
-+{
-+ struct dprtc_time *cmd_params;
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPRTC_CMDID_SET_ALARM,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dprtc_time *)cmd.params;
-+ cmd_params->time = cpu_to_le64(time);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dprtc_get_api_version() - Get Data Path Real Time Counter API version
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @major_ver: Major version of data path real time counter API
-+ * @minor_ver: Minor version of data path real time counter API
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dprtc_get_api_version(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t *major_ver,
-+ uint16_t *minor_ver)
-+{
-+ struct dprtc_rsp_get_api_version *rsp_params;
-+ struct fsl_mc_command cmd = { 0 };
-+ int err;
-+
-+ cmd.header = mc_encode_cmd_header(DPRTC_CMDID_GET_API_VERSION,
-+ cmd_flags,
-+ 0);
-+
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ rsp_params = (struct dprtc_rsp_get_api_version *)cmd.params;
-+ *major_ver = le16_to_cpu(rsp_params->major);
-+ *minor_ver = le16_to_cpu(rsp_params->minor);
-+
-+ return 0;
-+}
---- /dev/null
-+++ b/drivers/staging/fsl-dpaa2/rtc/dprtc.h
-@@ -0,0 +1,172 @@
-+/* Copyright 2013-2016 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of the above-listed copyright holders nor the
-+ * names of any contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
-+ * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
-+ * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
-+ * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE
-+ * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
-+ * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
-+ * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
-+ * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
-+ * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
-+ * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
-+ * POSSIBILITY OF SUCH DAMAGE.
-+ */
-+#ifndef __FSL_DPRTC_H
-+#define __FSL_DPRTC_H
-+
-+/* Data Path Real Time Counter API
-+ * Contains initialization APIs and runtime control APIs for RTC
-+ */
-+
-+struct fsl_mc_io;
-+
-+/**
-+ * Number of irq's
-+ */
-+#define DPRTC_MAX_IRQ_NUM 1
-+#define DPRTC_IRQ_INDEX 0
-+
-+/**
-+ * Interrupt event masks:
-+ */
-+
-+/**
-+ * Interrupt event mask indicating alarm event had occurred
-+ */
-+#define DPRTC_EVENT_ALARM 0x40000000
-+/**
-+ * Interrupt event mask indicating periodic pulse event had occurred
-+ */
-+#define DPRTC_EVENT_PPS 0x08000000
-+
-+int dprtc_open(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ int dprtc_id,
-+ uint16_t *token);
-+
-+int dprtc_close(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token);
-+
-+/**
-+ * struct dprtc_cfg - Structure representing DPRTC configuration
-+ * @options: place holder
-+ */
-+struct dprtc_cfg {
-+ uint32_t options;
-+};
-+
-+int dprtc_create(struct fsl_mc_io *mc_io,
-+ uint16_t dprc_token,
-+ uint32_t cmd_flags,
-+ const struct dprtc_cfg *cfg,
-+ uint32_t *obj_id);
-+
-+int dprtc_destroy(struct fsl_mc_io *mc_io,
-+ uint16_t dprc_token,
-+ uint32_t cmd_flags,
-+ uint32_t object_id);
-+
-+int dprtc_set_clock_offset(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token,
-+ int64_t offset);
-+
-+int dprtc_set_freq_compensation(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token,
-+ uint32_t freq_compensation);
-+
-+int dprtc_get_freq_compensation(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token,
-+ uint32_t *freq_compensation);
-+
-+int dprtc_get_time(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token,
-+ uint64_t *time);
-+
-+int dprtc_set_time(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token,
-+ uint64_t time);
-+
-+int dprtc_set_alarm(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token,
-+ uint64_t time);
-+
-+int dprtc_set_irq_enable(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token,
-+ uint8_t irq_index,
-+ uint8_t en);
-+
-+int dprtc_get_irq_enable(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token,
-+ uint8_t irq_index,
-+ uint8_t *en);
-+
-+int dprtc_set_irq_mask(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token,
-+ uint8_t irq_index,
-+ uint32_t mask);
-+
-+int dprtc_get_irq_mask(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token,
-+ uint8_t irq_index,
-+ uint32_t *mask);
-+
-+int dprtc_get_irq_status(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token,
-+ uint8_t irq_index,
-+ uint32_t *status);
-+
-+int dprtc_clear_irq_status(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token,
-+ uint8_t irq_index,
-+ uint32_t status);
-+
-+/**
-+ * struct dprtc_attr - Structure representing DPRTC attributes
-+ * @id: DPRTC object ID
-+ */
-+struct dprtc_attr {
-+ int id;
-+};
-+
-+int dprtc_get_attributes(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token,
-+ struct dprtc_attr *attr);
-+
-+int dprtc_get_api_version(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t *major_ver,
-+ uint16_t *minor_ver);
-+
-+#endif /* __FSL_DPRTC_H */
---- /dev/null
-+++ b/drivers/staging/fsl-dpaa2/rtc/rtc.c
-@@ -0,0 +1,242 @@
-+/* Copyright 2013-2015 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of the above-listed copyright holders nor the
-+ * names of any contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
-+ * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
-+ * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
-+ * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE
-+ * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
-+ * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
-+ * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
-+ * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
-+ * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
-+ * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
-+ * POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#include <linux/module.h>
-+#include <linux/ptp_clock_kernel.h>
-+
-+#include <linux/fsl/mc.h>
-+
-+#include "dprtc.h"
-+#include "dprtc-cmd.h"
-+
-+#define N_EXT_TS 2
-+
-+struct ptp_clock *clock;
-+struct fsl_mc_device *rtc_mc_dev;
-+u32 freqCompensation;
-+
-+/* PTP clock operations */
-+static int ptp_dpaa2_adjfreq(struct ptp_clock_info *ptp, s32 ppb)
-+{
-+ u64 adj;
-+ u32 diff, tmr_add;
-+ int neg_adj = 0;
-+ int err = 0;
-+ struct fsl_mc_device *mc_dev = rtc_mc_dev;
-+ struct device *dev = &mc_dev->dev;
-+
-+ if (ppb < 0) {
-+ neg_adj = 1;
-+ ppb = -ppb;
-+ }
-+
-+ tmr_add = freqCompensation;
-+ adj = tmr_add;
-+ adj *= ppb;
-+ diff = div_u64(adj, 1000000000ULL);
-+
-+ tmr_add = neg_adj ? tmr_add - diff : tmr_add + diff;
-+
-+ err = dprtc_set_freq_compensation(mc_dev->mc_io, 0,
-+ mc_dev->mc_handle, tmr_add);
-+ if (err)
-+ dev_err(dev, "dprtc_set_freq_compensation err %d\n", err);
-+ return 0;
-+}
-+
-+static int ptp_dpaa2_adjtime(struct ptp_clock_info *ptp, s64 delta)
-+{
-+ s64 now;
-+ int err = 0;
-+ struct fsl_mc_device *mc_dev = rtc_mc_dev;
-+ struct device *dev = &mc_dev->dev;
-+
-+ err = dprtc_get_time(mc_dev->mc_io, 0, mc_dev->mc_handle, &now);
-+ if (err) {
-+ dev_err(dev, "dprtc_get_time err %d\n", err);
-+ return 0;
-+ }
-+
-+ now += delta;
-+
-+ err = dprtc_set_time(mc_dev->mc_io, 0, mc_dev->mc_handle, now);
-+ if (err) {
-+ dev_err(dev, "dprtc_set_time err %d\n", err);
-+ return 0;
-+ }
-+ return 0;
-+}
-+
-+static int ptp_dpaa2_gettime(struct ptp_clock_info *ptp, struct timespec *ts)
-+{
-+ u64 ns;
-+ u32 remainder;
-+ int err = 0;
-+ struct fsl_mc_device *mc_dev = rtc_mc_dev;
-+ struct device *dev = &mc_dev->dev;
-+
-+ err = dprtc_get_time(mc_dev->mc_io, 0, mc_dev->mc_handle, &ns);
-+ if (err) {
-+ dev_err(dev, "dprtc_get_time err %d\n", err);
-+ return 0;
-+ }
-+
-+ ts->tv_sec = div_u64_rem(ns, 1000000000, &remainder);
-+ ts->tv_nsec = remainder;
-+ return 0;
-+}
-+
-+static int ptp_dpaa2_settime(struct ptp_clock_info *ptp,
-+ const struct timespec *ts)
-+{
-+ u64 ns;
-+ int err = 0;
-+ struct fsl_mc_device *mc_dev = rtc_mc_dev;
-+ struct device *dev = &mc_dev->dev;
-+
-+ ns = ts->tv_sec * 1000000000ULL;
-+ ns += ts->tv_nsec;
-+
-+ err = dprtc_set_time(mc_dev->mc_io, 0, mc_dev->mc_handle, ns);
-+ if (err)
-+ dev_err(dev, "dprtc_set_time err %d\n", err);
-+ return 0;
-+}
-+
-+static struct ptp_clock_info ptp_dpaa2_caps = {
-+ .owner = THIS_MODULE,
-+ .name = "dpaa2 clock",
-+ .max_adj = 512000,
-+ .n_alarm = 0,
-+ .n_ext_ts = N_EXT_TS,
-+ .n_per_out = 0,
-+ .n_pins = 0,
-+ .pps = 1,
-+ .adjfreq = ptp_dpaa2_adjfreq,
-+ .adjtime = ptp_dpaa2_adjtime,
-+ .gettime64 = ptp_dpaa2_gettime,
-+ .settime64 = ptp_dpaa2_settime,
-+};
-+
-+static int rtc_probe(struct fsl_mc_device *mc_dev)
-+{
-+ struct device *dev;
-+ int err = 0;
-+ int dpaa2_phc_index;
-+ u32 tmr_add = 0;
-+
-+ if (!mc_dev)
-+ return -EFAULT;
-+
-+ dev = &mc_dev->dev;
-+
-+ err = fsl_mc_portal_allocate(mc_dev, 0, &mc_dev->mc_io);
-+ if (unlikely(err)) {
-+ dev_err(dev, "fsl_mc_portal_allocate err %d\n", err);
-+ goto err_exit;
-+ }
-+ if (!mc_dev->mc_io) {
-+ dev_err(dev,
-+ "fsl_mc_portal_allocate returned null handle but no error\n");
-+ err = -EFAULT;
-+ goto err_exit;
-+ }
-+
-+ err = dprtc_open(mc_dev->mc_io, 0, mc_dev->obj_desc.id,
-+ &mc_dev->mc_handle);
-+ if (err) {
-+ dev_err(dev, "dprtc_open err %d\n", err);
-+ goto err_free_mcp;
-+ }
-+ if (!mc_dev->mc_handle) {
-+ dev_err(dev, "dprtc_open returned null handle but no error\n");
-+ err = -EFAULT;
-+ goto err_free_mcp;
-+ }
-+
-+ rtc_mc_dev = mc_dev;
-+
-+ err = dprtc_get_freq_compensation(mc_dev->mc_io, 0,
-+ mc_dev->mc_handle, &tmr_add);
-+ if (err) {
-+ dev_err(dev, "dprtc_get_freq_compensation err %d\n", err);
-+ goto err_close;
-+ }
-+ freqCompensation = tmr_add;
-+
-+ clock = ptp_clock_register(&ptp_dpaa2_caps, dev);
-+ if (IS_ERR(clock)) {
-+ err = PTR_ERR(clock);
-+ goto err_close;
-+ }
-+ dpaa2_phc_index = ptp_clock_index(clock);
-+
-+ return 0;
-+err_close:
-+ dprtc_close(mc_dev->mc_io, 0, mc_dev->mc_handle);
-+err_free_mcp:
-+ fsl_mc_portal_free(mc_dev->mc_io);
-+err_exit:
-+ return err;
-+}
-+
-+static int rtc_remove(struct fsl_mc_device *mc_dev)
-+{
-+ ptp_clock_unregister(clock);
-+ dprtc_close(mc_dev->mc_io, 0, mc_dev->mc_handle);
-+ fsl_mc_portal_free(mc_dev->mc_io);
-+
-+ return 0;
-+}
-+
-+static const struct fsl_mc_device_id rtc_match_id_table[] = {
-+ {
-+ .vendor = FSL_MC_VENDOR_FREESCALE,
-+ .obj_type = "dprtc",
-+ },
-+ {}
-+};
-+
-+static struct fsl_mc_driver rtc_drv = {
-+ .driver = {
-+ .name = KBUILD_MODNAME,
-+ .owner = THIS_MODULE,
-+ },
-+ .probe = rtc_probe,
-+ .remove = rtc_remove,
-+ .match_id_table = rtc_match_id_table,
-+};
-+
-+module_fsl_mc_driver(rtc_drv);
-+
-+MODULE_LICENSE("GPL");
-+MODULE_DESCRIPTION("DPAA2 RTC (PTP 1588 clock) driver (prototype)");
---- a/include/linux/filter.h
-+++ b/include/linux/filter.h
-@@ -429,12 +429,15 @@ struct sk_filter {
-
- struct bpf_skb_data_end {
- struct qdisc_skb_cb qdisc_cb;
-+ void *data_meta;
- void *data_end;
- };
-
- struct xdp_buff {
- void *data;
- void *data_end;
-+ void *data_meta;
-+ void *data_hard_start;
- };
-
- /* compute the linear packet data range [data, data_end) which
diff --git a/target/linux/layerscape/patches-4.9/706-fsl_ppfe-support-layercape.patch b/target/linux/layerscape/patches-4.9/706-fsl_ppfe-support-layercape.patch
deleted file mode 100644
index e71c85c7f5..0000000000
--- a/target/linux/layerscape/patches-4.9/706-fsl_ppfe-support-layercape.patch
+++ /dev/null
@@ -1,10654 +0,0 @@
-From 50fb2f2e93aeae0baed156eb4794a2f358376b77 Mon Sep 17 00:00:00 2001
-From: Yangbo Lu <yangbo.lu@nxp.com>
-Date: Thu, 5 Jul 2018 17:19:20 +0800
-Subject: [PATCH 12/32] fsl_ppfe: support layercape
-
-This is an integrated patch for layerscape pfe support.
-
-Calvin Johnson <calvin.johnson@nxp.com>
-Signed-off-by: Yangbo Lu <yangbo.lu@nxp.com>
----
- drivers/staging/fsl_ppfe/Kconfig | 20 +
- drivers/staging/fsl_ppfe/Makefile | 19 +
- drivers/staging/fsl_ppfe/TODO | 2 +
- drivers/staging/fsl_ppfe/include/pfe/cbus.h | 78 +
- .../staging/fsl_ppfe/include/pfe/cbus/bmu.h | 55 +
- .../fsl_ppfe/include/pfe/cbus/class_csr.h | 289 ++
- .../fsl_ppfe/include/pfe/cbus/emac_mtip.h | 242 ++
- .../staging/fsl_ppfe/include/pfe/cbus/gpi.h | 86 +
- .../staging/fsl_ppfe/include/pfe/cbus/hif.h | 100 +
- .../fsl_ppfe/include/pfe/cbus/hif_nocpy.h | 50 +
- .../fsl_ppfe/include/pfe/cbus/tmu_csr.h | 168 ++
- .../fsl_ppfe/include/pfe/cbus/util_csr.h | 61 +
- drivers/staging/fsl_ppfe/include/pfe/pfe.h | 372 +++
- drivers/staging/fsl_ppfe/pfe_ctrl.c | 238 ++
- drivers/staging/fsl_ppfe/pfe_ctrl.h | 112 +
- drivers/staging/fsl_ppfe/pfe_debugfs.c | 111 +
- drivers/staging/fsl_ppfe/pfe_debugfs.h | 25 +
- drivers/staging/fsl_ppfe/pfe_eth.c | 2491 +++++++++++++++++
- drivers/staging/fsl_ppfe/pfe_eth.h | 184 ++
- drivers/staging/fsl_ppfe/pfe_firmware.c | 314 +++
- drivers/staging/fsl_ppfe/pfe_firmware.h | 32 +
- drivers/staging/fsl_ppfe/pfe_hal.c | 1516 ++++++++++
- drivers/staging/fsl_ppfe/pfe_hif.c | 1072 +++++++
- drivers/staging/fsl_ppfe/pfe_hif.h | 211 ++
- drivers/staging/fsl_ppfe/pfe_hif_lib.c | 640 +++++
- drivers/staging/fsl_ppfe/pfe_hif_lib.h | 241 ++
- drivers/staging/fsl_ppfe/pfe_hw.c | 176 ++
- drivers/staging/fsl_ppfe/pfe_hw.h | 27 +
- .../staging/fsl_ppfe/pfe_ls1012a_platform.c | 385 +++
- drivers/staging/fsl_ppfe/pfe_mod.c | 156 ++
- drivers/staging/fsl_ppfe/pfe_mod.h | 114 +
- drivers/staging/fsl_ppfe/pfe_perfmon.h | 38 +
- drivers/staging/fsl_ppfe/pfe_sysfs.c | 818 ++++++
- drivers/staging/fsl_ppfe/pfe_sysfs.h | 29 +
- 34 files changed, 10472 insertions(+)
- create mode 100644 drivers/staging/fsl_ppfe/Kconfig
- create mode 100644 drivers/staging/fsl_ppfe/Makefile
- create mode 100644 drivers/staging/fsl_ppfe/TODO
- create mode 100644 drivers/staging/fsl_ppfe/include/pfe/cbus.h
- create mode 100644 drivers/staging/fsl_ppfe/include/pfe/cbus/bmu.h
- create mode 100644 drivers/staging/fsl_ppfe/include/pfe/cbus/class_csr.h
- create mode 100644 drivers/staging/fsl_ppfe/include/pfe/cbus/emac_mtip.h
- create mode 100644 drivers/staging/fsl_ppfe/include/pfe/cbus/gpi.h
- create mode 100644 drivers/staging/fsl_ppfe/include/pfe/cbus/hif.h
- create mode 100644 drivers/staging/fsl_ppfe/include/pfe/cbus/hif_nocpy.h
- create mode 100644 drivers/staging/fsl_ppfe/include/pfe/cbus/tmu_csr.h
- create mode 100644 drivers/staging/fsl_ppfe/include/pfe/cbus/util_csr.h
- create mode 100644 drivers/staging/fsl_ppfe/include/pfe/pfe.h
- create mode 100644 drivers/staging/fsl_ppfe/pfe_ctrl.c
- create mode 100644 drivers/staging/fsl_ppfe/pfe_ctrl.h
- create mode 100644 drivers/staging/fsl_ppfe/pfe_debugfs.c
- create mode 100644 drivers/staging/fsl_ppfe/pfe_debugfs.h
- create mode 100644 drivers/staging/fsl_ppfe/pfe_eth.c
- create mode 100644 drivers/staging/fsl_ppfe/pfe_eth.h
- create mode 100644 drivers/staging/fsl_ppfe/pfe_firmware.c
- create mode 100644 drivers/staging/fsl_ppfe/pfe_firmware.h
- create mode 100644 drivers/staging/fsl_ppfe/pfe_hal.c
- create mode 100644 drivers/staging/fsl_ppfe/pfe_hif.c
- create mode 100644 drivers/staging/fsl_ppfe/pfe_hif.h
- create mode 100644 drivers/staging/fsl_ppfe/pfe_hif_lib.c
- create mode 100644 drivers/staging/fsl_ppfe/pfe_hif_lib.h
- create mode 100644 drivers/staging/fsl_ppfe/pfe_hw.c
- create mode 100644 drivers/staging/fsl_ppfe/pfe_hw.h
- create mode 100644 drivers/staging/fsl_ppfe/pfe_ls1012a_platform.c
- create mode 100644 drivers/staging/fsl_ppfe/pfe_mod.c
- create mode 100644 drivers/staging/fsl_ppfe/pfe_mod.h
- create mode 100644 drivers/staging/fsl_ppfe/pfe_perfmon.h
- create mode 100644 drivers/staging/fsl_ppfe/pfe_sysfs.c
- create mode 100644 drivers/staging/fsl_ppfe/pfe_sysfs.h
-
---- /dev/null
-+++ b/drivers/staging/fsl_ppfe/Kconfig
-@@ -0,0 +1,20 @@
-+#
-+# Freescale Programmable Packet Forwarding Engine driver
-+#
-+config FSL_PPFE
-+ bool "Freescale PPFE Driver"
-+ default n
-+ ---help---
-+ Freescale LS1012A SoC has a Programmable Packet Forwarding Engine.
-+ It provides two high performance ethernet interfaces.
-+ This driver initializes, programs and controls the PPFE.
-+ Use this driver to enable network connectivity on LS1012A platforms.
-+
-+if FSL_PPFE
-+
-+config FSL_PPFE_UTIL_DISABLED
-+ bool "Disable PPFE UTIL Processor Engine"
-+ ---help---
-+ UTIL PE has to be enabled only if required.
-+
-+endif # FSL_PPFE
---- /dev/null
-+++ b/drivers/staging/fsl_ppfe/Makefile
-@@ -0,0 +1,19 @@
-+#
-+# Makefile for Freesecale PPFE driver
-+#
-+
-+ccflags-y += -I$(src)/include -I$(src)
-+
-+obj-m += pfe.o
-+
-+pfe-y += pfe_mod.o \
-+ pfe_hw.o \
-+ pfe_firmware.o \
-+ pfe_ctrl.o \
-+ pfe_hif.o \
-+ pfe_hif_lib.o\
-+ pfe_eth.o \
-+ pfe_sysfs.o \
-+ pfe_debugfs.o \
-+ pfe_ls1012a_platform.o \
-+ pfe_hal.o
---- /dev/null
-+++ b/drivers/staging/fsl_ppfe/TODO
-@@ -0,0 +1,2 @@
-+TODO:
-+ - provide pfe pe monitoring support
---- /dev/null
-+++ b/drivers/staging/fsl_ppfe/include/pfe/cbus.h
-@@ -0,0 +1,78 @@
-+/*
-+ * Copyright 2015-2016 Freescale Semiconductor, Inc.
-+ * Copyright 2017 NXP
-+ *
-+ * This program is free software; you can redistribute it and/or modify
-+ * it under the terms of the GNU General Public License as published by
-+ * the Free Software Foundation; either version 2 of the License, or
-+ * (at your option) any later version.
-+ *
-+ * This program is distributed in the hope that it will be useful,
-+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
-+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
-+ * GNU General Public License for more details.
-+ *
-+ * You should have received a copy of the GNU General Public License
-+ * along with this program. If not, see <http://www.gnu.org/licenses/>.
-+ */
-+
-+#ifndef _CBUS_H_
-+#define _CBUS_H_
-+
-+#define EMAC1_BASE_ADDR (CBUS_BASE_ADDR + 0x200000)
-+#define EGPI1_BASE_ADDR (CBUS_BASE_ADDR + 0x210000)
-+#define EMAC2_BASE_ADDR (CBUS_BASE_ADDR + 0x220000)
-+#define EGPI2_BASE_ADDR (CBUS_BASE_ADDR + 0x230000)
-+#define BMU1_BASE_ADDR (CBUS_BASE_ADDR + 0x240000)
-+#define BMU2_BASE_ADDR (CBUS_BASE_ADDR + 0x250000)
-+#define ARB_BASE_ADDR (CBUS_BASE_ADDR + 0x260000)
-+#define DDR_CONFIG_BASE_ADDR (CBUS_BASE_ADDR + 0x270000)
-+#define HIF_BASE_ADDR (CBUS_BASE_ADDR + 0x280000)
-+#define HGPI_BASE_ADDR (CBUS_BASE_ADDR + 0x290000)
-+#define LMEM_BASE_ADDR (CBUS_BASE_ADDR + 0x300000)
-+#define LMEM_SIZE 0x10000
-+#define LMEM_END (LMEM_BASE_ADDR + LMEM_SIZE)
-+#define TMU_CSR_BASE_ADDR (CBUS_BASE_ADDR + 0x310000)
-+#define CLASS_CSR_BASE_ADDR (CBUS_BASE_ADDR + 0x320000)
-+#define HIF_NOCPY_BASE_ADDR (CBUS_BASE_ADDR + 0x350000)
-+#define UTIL_CSR_BASE_ADDR (CBUS_BASE_ADDR + 0x360000)
-+#define CBUS_GPT_BASE_ADDR (CBUS_BASE_ADDR + 0x370000)
-+
-+/*
-+ * defgroup XXX_MEM_ACCESS_ADDR PE memory access through CSR
-+ * XXX_MEM_ACCESS_ADDR register bit definitions.
-+ */
-+#define PE_MEM_ACCESS_WRITE BIT(31) /* Internal Memory Write. */
-+#define PE_MEM_ACCESS_IMEM BIT(15)
-+#define PE_MEM_ACCESS_DMEM BIT(16)
-+
-+/* Byte Enables of the Internal memory access. These are interpred in BE */
-+#define PE_MEM_ACCESS_BYTE_ENABLE(offset, size) \
-+ ({ typeof(size) size_ = (size); \
-+ (((BIT(size_) - 1) << (4 - (offset) - (size_))) & 0xf) << 24; })
-+
-+#include "cbus/emac_mtip.h"
-+#include "cbus/gpi.h"
-+#include "cbus/bmu.h"
-+#include "cbus/hif.h"
-+#include "cbus/tmu_csr.h"
-+#include "cbus/class_csr.h"
-+#include "cbus/hif_nocpy.h"
-+#include "cbus/util_csr.h"
-+
-+/* PFE cores states */
-+#define CORE_DISABLE 0x00000000
-+#define CORE_ENABLE 0x00000001
-+#define CORE_SW_RESET 0x00000002
-+
-+/* LMEM defines */
-+#define LMEM_HDR_SIZE 0x0010
-+#define LMEM_BUF_SIZE_LN2 0x7
-+#define LMEM_BUF_SIZE BIT(LMEM_BUF_SIZE_LN2)
-+
-+/* DDR defines */
-+#define DDR_HDR_SIZE 0x0100
-+#define DDR_BUF_SIZE_LN2 0xb
-+#define DDR_BUF_SIZE BIT(DDR_BUF_SIZE_LN2)
-+
-+#endif /* _CBUS_H_ */
---- /dev/null
-+++ b/drivers/staging/fsl_ppfe/include/pfe/cbus/bmu.h
-@@ -0,0 +1,55 @@
-+/*
-+ * Copyright 2015-2016 Freescale Semiconductor, Inc.
-+ * Copyright 2017 NXP
-+ *
-+ * This program is free software; you can redistribute it and/or modify
-+ * it under the terms of the GNU General Public License as published by
-+ * the Free Software Foundation; either version 2 of the License, or
-+ * (at your option) any later version.
-+ *
-+ * This program is distributed in the hope that it will be useful,
-+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
-+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
-+ * GNU General Public License for more details.
-+ *
-+ * You should have received a copy of the GNU General Public License
-+ * along with this program. If not, see <http://www.gnu.org/licenses/>.
-+ */
-+
-+#ifndef _BMU_H_
-+#define _BMU_H_
-+
-+#define BMU_VERSION 0x000
-+#define BMU_CTRL 0x004
-+#define BMU_UCAST_CONFIG 0x008
-+#define BMU_UCAST_BASE_ADDR 0x00c
-+#define BMU_BUF_SIZE 0x010
-+#define BMU_BUF_CNT 0x014
-+#define BMU_THRES 0x018
-+#define BMU_INT_SRC 0x020
-+#define BMU_INT_ENABLE 0x024
-+#define BMU_ALLOC_CTRL 0x030
-+#define BMU_FREE_CTRL 0x034
-+#define BMU_FREE_ERR_ADDR 0x038
-+#define BMU_CURR_BUF_CNT 0x03c
-+#define BMU_MCAST_CNT 0x040
-+#define BMU_MCAST_ALLOC_CTRL 0x044
-+#define BMU_REM_BUF_CNT 0x048
-+#define BMU_LOW_WATERMARK 0x050
-+#define BMU_HIGH_WATERMARK 0x054
-+#define BMU_INT_MEM_ACCESS 0x100
-+
-+struct BMU_CFG {
-+ unsigned long baseaddr;
-+ u32 count;
-+ u32 size;
-+ u32 low_watermark;
-+ u32 high_watermark;
-+};
-+
-+#define BMU1_BUF_SIZE LMEM_BUF_SIZE_LN2
-+#define BMU2_BUF_SIZE DDR_BUF_SIZE_LN2
-+
-+#define BMU2_MCAST_ALLOC_CTRL (BMU2_BASE_ADDR + BMU_MCAST_ALLOC_CTRL)
-+
-+#endif /* _BMU_H_ */
---- /dev/null
-+++ b/drivers/staging/fsl_ppfe/include/pfe/cbus/class_csr.h
-@@ -0,0 +1,289 @@
-+/*
-+ * Copyright 2015-2016 Freescale Semiconductor, Inc.
-+ * Copyright 2017 NXP
-+ *
-+ * This program is free software; you can redistribute it and/or modify
-+ * it under the terms of the GNU General Public License as published by
-+ * the Free Software Foundation; either version 2 of the License, or
-+ * (at your option) any later version.
-+ *
-+ * This program is distributed in the hope that it will be useful,
-+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
-+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
-+ * GNU General Public License for more details.
-+ *
-+ * You should have received a copy of the GNU General Public License
-+ * along with this program. If not, see <http://www.gnu.org/licenses/>.
-+ */
-+
-+#ifndef _CLASS_CSR_H_
-+#define _CLASS_CSR_H_
-+
-+/* @file class_csr.h.
-+ * class_csr - block containing all the classifier control and status register.
-+ * Mapped on CBUS and accessible from all PE's and ARM.
-+ */
-+#define CLASS_VERSION (CLASS_CSR_BASE_ADDR + 0x000)
-+#define CLASS_TX_CTRL (CLASS_CSR_BASE_ADDR + 0x004)
-+#define CLASS_INQ_PKTPTR (CLASS_CSR_BASE_ADDR + 0x010)
-+
-+/* (ddr_hdr_size[24:16], lmem_hdr_size[5:0]) */
-+#define CLASS_HDR_SIZE (CLASS_CSR_BASE_ADDR + 0x014)
-+
-+/* LMEM header size for the Classifier block.\ Data in the LMEM
-+ * is written from this offset.
-+ */
-+#define CLASS_HDR_SIZE_LMEM(off) ((off) & 0x3f)
-+
-+/* DDR header size for the Classifier block.\ Data in the DDR
-+ * is written from this offset.
-+ */
-+#define CLASS_HDR_SIZE_DDR(off) (((off) & 0x1ff) << 16)
-+
-+#define CLASS_PE0_QB_DM_ADDR0 (CLASS_CSR_BASE_ADDR + 0x020)
-+
-+/* DMEM address of first [15:0] and second [31:16] buffers on QB side. */
-+#define CLASS_PE0_QB_DM_ADDR1 (CLASS_CSR_BASE_ADDR + 0x024)
-+
-+/* DMEM address of third [15:0] and fourth [31:16] buffers on QB side. */
-+#define CLASS_PE0_RO_DM_ADDR0 (CLASS_CSR_BASE_ADDR + 0x060)
-+
-+/* DMEM address of first [15:0] and second [31:16] buffers on RO side. */
-+#define CLASS_PE0_RO_DM_ADDR1 (CLASS_CSR_BASE_ADDR + 0x064)
-+
-+/* DMEM address of third [15:0] and fourth [31:16] buffers on RO side. */
-+
-+/* @name Class PE memory access. Allows external PE's and HOST to
-+ * read/write PMEM/DMEM memory ranges for each classifier PE.
-+ */
-+/* {sr_pe_mem_cmd[31], csr_pe_mem_wren[27:24], csr_pe_mem_addr[23:0]},
-+ * See \ref XXX_MEM_ACCESS_ADDR for details.
-+ */
-+#define CLASS_MEM_ACCESS_ADDR (CLASS_CSR_BASE_ADDR + 0x100)
-+
-+/* Internal Memory Access Write Data [31:0] */
-+#define CLASS_MEM_ACCESS_WDATA (CLASS_CSR_BASE_ADDR + 0x104)
-+
-+/* Internal Memory Access Read Data [31:0] */
-+#define CLASS_MEM_ACCESS_RDATA (CLASS_CSR_BASE_ADDR + 0x108)
-+#define CLASS_TM_INQ_ADDR (CLASS_CSR_BASE_ADDR + 0x114)
-+#define CLASS_PE_STATUS (CLASS_CSR_BASE_ADDR + 0x118)
-+
-+#define CLASS_PHY1_RX_PKTS (CLASS_CSR_BASE_ADDR + 0x11c)
-+#define CLASS_PHY1_TX_PKTS (CLASS_CSR_BASE_ADDR + 0x120)
-+#define CLASS_PHY1_LP_FAIL_PKTS (CLASS_CSR_BASE_ADDR + 0x124)
-+#define CLASS_PHY1_INTF_FAIL_PKTS (CLASS_CSR_BASE_ADDR + 0x128)
-+#define CLASS_PHY1_INTF_MATCH_PKTS (CLASS_CSR_BASE_ADDR + 0x12c)
-+#define CLASS_PHY1_L3_FAIL_PKTS (CLASS_CSR_BASE_ADDR + 0x130)
-+#define CLASS_PHY1_V4_PKTS (CLASS_CSR_BASE_ADDR + 0x134)
-+#define CLASS_PHY1_V6_PKTS (CLASS_CSR_BASE_ADDR + 0x138)
-+#define CLASS_PHY1_CHKSUM_ERR_PKTS (CLASS_CSR_BASE_ADDR + 0x13c)
-+#define CLASS_PHY1_TTL_ERR_PKTS (CLASS_CSR_BASE_ADDR + 0x140)
-+#define CLASS_PHY2_RX_PKTS (CLASS_CSR_BASE_ADDR + 0x144)
-+#define CLASS_PHY2_TX_PKTS (CLASS_CSR_BASE_ADDR + 0x148)
-+#define CLASS_PHY2_LP_FAIL_PKTS (CLASS_CSR_BASE_ADDR + 0x14c)
-+#define CLASS_PHY2_INTF_FAIL_PKTS (CLASS_CSR_BASE_ADDR + 0x150)
-+#define CLASS_PHY2_INTF_MATCH_PKTS (CLASS_CSR_BASE_ADDR + 0x154)
-+#define CLASS_PHY2_L3_FAIL_PKTS (CLASS_CSR_BASE_ADDR + 0x158)
-+#define CLASS_PHY2_V4_PKTS (CLASS_CSR_BASE_ADDR + 0x15c)
-+#define CLASS_PHY2_V6_PKTS (CLASS_CSR_BASE_ADDR + 0x160)
-+#define CLASS_PHY2_CHKSUM_ERR_PKTS (CLASS_CSR_BASE_ADDR + 0x164)
-+#define CLASS_PHY2_TTL_ERR_PKTS (CLASS_CSR_BASE_ADDR + 0x168)
-+#define CLASS_PHY3_RX_PKTS (CLASS_CSR_BASE_ADDR + 0x16c)
-+#define CLASS_PHY3_TX_PKTS (CLASS_CSR_BASE_ADDR + 0x170)
-+#define CLASS_PHY3_LP_FAIL_PKTS (CLASS_CSR_BASE_ADDR + 0x174)
-+#define CLASS_PHY3_INTF_FAIL_PKTS (CLASS_CSR_BASE_ADDR + 0x178)
-+#define CLASS_PHY3_INTF_MATCH_PKTS (CLASS_CSR_BASE_ADDR + 0x17c)
-+#define CLASS_PHY3_L3_FAIL_PKTS (CLASS_CSR_BASE_ADDR + 0x180)
-+#define CLASS_PHY3_V4_PKTS (CLASS_CSR_BASE_ADDR + 0x184)
-+#define CLASS_PHY3_V6_PKTS (CLASS_CSR_BASE_ADDR + 0x188)
-+#define CLASS_PHY3_CHKSUM_ERR_PKTS (CLASS_CSR_BASE_ADDR + 0x18c)
-+#define CLASS_PHY3_TTL_ERR_PKTS (CLASS_CSR_BASE_ADDR + 0x190)
-+#define CLASS_PHY1_ICMP_PKTS (CLASS_CSR_BASE_ADDR + 0x194)
-+#define CLASS_PHY1_IGMP_PKTS (CLASS_CSR_BASE_ADDR + 0x198)
-+#define CLASS_PHY1_TCP_PKTS (CLASS_CSR_BASE_ADDR + 0x19c)
-+#define CLASS_PHY1_UDP_PKTS (CLASS_CSR_BASE_ADDR + 0x1a0)
-+#define CLASS_PHY2_ICMP_PKTS (CLASS_CSR_BASE_ADDR + 0x1a4)
-+#define CLASS_PHY2_IGMP_PKTS (CLASS_CSR_BASE_ADDR + 0x1a8)
-+#define CLASS_PHY2_TCP_PKTS (CLASS_CSR_BASE_ADDR + 0x1ac)
-+#define CLASS_PHY2_UDP_PKTS (CLASS_CSR_BASE_ADDR + 0x1b0)
-+#define CLASS_PHY3_ICMP_PKTS (CLASS_CSR_BASE_ADDR + 0x1b4)
-+#define CLASS_PHY3_IGMP_PKTS (CLASS_CSR_BASE_ADDR + 0x1b8)
-+#define CLASS_PHY3_TCP_PKTS (CLASS_CSR_BASE_ADDR + 0x1bc)
-+#define CLASS_PHY3_UDP_PKTS (CLASS_CSR_BASE_ADDR + 0x1c0)
-+#define CLASS_PHY4_ICMP_PKTS (CLASS_CSR_BASE_ADDR + 0x1c4)
-+#define CLASS_PHY4_IGMP_PKTS (CLASS_CSR_BASE_ADDR + 0x1c8)
-+#define CLASS_PHY4_TCP_PKTS (CLASS_CSR_BASE_ADDR + 0x1cc)
-+#define CLASS_PHY4_UDP_PKTS (CLASS_CSR_BASE_ADDR + 0x1d0)
-+#define CLASS_PHY4_RX_PKTS (CLASS_CSR_BASE_ADDR + 0x1d4)
-+#define CLASS_PHY4_TX_PKTS (CLASS_CSR_BASE_ADDR + 0x1d8)
-+#define CLASS_PHY4_LP_FAIL_PKTS (CLASS_CSR_BASE_ADDR + 0x1dc)
-+#define CLASS_PHY4_INTF_FAIL_PKTS (CLASS_CSR_BASE_ADDR + 0x1e0)
-+#define CLASS_PHY4_INTF_MATCH_PKTS (CLASS_CSR_BASE_ADDR + 0x1e4)
-+#define CLASS_PHY4_L3_FAIL_PKTS (CLASS_CSR_BASE_ADDR + 0x1e8)
-+#define CLASS_PHY4_V4_PKTS (CLASS_CSR_BASE_ADDR + 0x1ec)
-+#define CLASS_PHY4_V6_PKTS (CLASS_CSR_BASE_ADDR + 0x1f0)
-+#define CLASS_PHY4_CHKSUM_ERR_PKTS (CLASS_CSR_BASE_ADDR + 0x1f4)
-+#define CLASS_PHY4_TTL_ERR_PKTS (CLASS_CSR_BASE_ADDR + 0x1f8)
-+
-+#define CLASS_PE_SYS_CLK_RATIO (CLASS_CSR_BASE_ADDR + 0x200)
-+#define CLASS_AFULL_THRES (CLASS_CSR_BASE_ADDR + 0x204)
-+#define CLASS_GAP_BETWEEN_READS (CLASS_CSR_BASE_ADDR + 0x208)
-+#define CLASS_MAX_BUF_CNT (CLASS_CSR_BASE_ADDR + 0x20c)
-+#define CLASS_TSQ_FIFO_THRES (CLASS_CSR_BASE_ADDR + 0x210)
-+#define CLASS_TSQ_MAX_CNT (CLASS_CSR_BASE_ADDR + 0x214)
-+#define CLASS_IRAM_DATA_0 (CLASS_CSR_BASE_ADDR + 0x218)
-+#define CLASS_IRAM_DATA_1 (CLASS_CSR_BASE_ADDR + 0x21c)
-+#define CLASS_IRAM_DATA_2 (CLASS_CSR_BASE_ADDR + 0x220)
-+#define CLASS_IRAM_DATA_3 (CLASS_CSR_BASE_ADDR + 0x224)
-+
-+#define CLASS_BUS_ACCESS_ADDR (CLASS_CSR_BASE_ADDR + 0x228)
-+
-+#define CLASS_BUS_ACCESS_WDATA (CLASS_CSR_BASE_ADDR + 0x22c)
-+#define CLASS_BUS_ACCESS_RDATA (CLASS_CSR_BASE_ADDR + 0x230)
-+
-+/* (route_entry_size[9:0], route_hash_size[23:16]
-+ * (this is actually ln2(size)))
-+ */
-+#define CLASS_ROUTE_HASH_ENTRY_SIZE (CLASS_CSR_BASE_ADDR + 0x234)
-+
-+#define CLASS_ROUTE_ENTRY_SIZE(size) ((size) & 0x1ff)
-+#define CLASS_ROUTE_HASH_SIZE(hash_bits) (((hash_bits) & 0xff) << 16)
-+
-+#define CLASS_ROUTE_TABLE_BASE (CLASS_CSR_BASE_ADDR + 0x238)
-+
-+#define CLASS_ROUTE_MULTI (CLASS_CSR_BASE_ADDR + 0x23c)
-+#define CLASS_SMEM_OFFSET (CLASS_CSR_BASE_ADDR + 0x240)
-+#define CLASS_LMEM_BUF_SIZE (CLASS_CSR_BASE_ADDR + 0x244)
-+#define CLASS_VLAN_ID (CLASS_CSR_BASE_ADDR + 0x248)
-+#define CLASS_BMU1_BUF_FREE (CLASS_CSR_BASE_ADDR + 0x24c)
-+#define CLASS_USE_TMU_INQ (CLASS_CSR_BASE_ADDR + 0x250)
-+#define CLASS_VLAN_ID1 (CLASS_CSR_BASE_ADDR + 0x254)
-+
-+#define CLASS_BUS_ACCESS_BASE (CLASS_CSR_BASE_ADDR + 0x258)
-+#define CLASS_BUS_ACCESS_BASE_MASK (0xFF000000)
-+/* bit 31:24 of PE peripheral address are stored in CLASS_BUS_ACCESS_BASE */
-+
-+#define CLASS_HIF_PARSE (CLASS_CSR_BASE_ADDR + 0x25c)
-+
-+#define CLASS_HOST_PE0_GP (CLASS_CSR_BASE_ADDR + 0x260)
-+#define CLASS_PE0_GP (CLASS_CSR_BASE_ADDR + 0x264)
-+#define CLASS_HOST_PE1_GP (CLASS_CSR_BASE_ADDR + 0x268)
-+#define CLASS_PE1_GP (CLASS_CSR_BASE_ADDR + 0x26c)
-+#define CLASS_HOST_PE2_GP (CLASS_CSR_BASE_ADDR + 0x270)
-+#define CLASS_PE2_GP (CLASS_CSR_BASE_ADDR + 0x274)
-+#define CLASS_HOST_PE3_GP (CLASS_CSR_BASE_ADDR + 0x278)
-+#define CLASS_PE3_GP (CLASS_CSR_BASE_ADDR + 0x27c)
-+#define CLASS_HOST_PE4_GP (CLASS_CSR_BASE_ADDR + 0x280)
-+#define CLASS_PE4_GP (CLASS_CSR_BASE_ADDR + 0x284)
-+#define CLASS_HOST_PE5_GP (CLASS_CSR_BASE_ADDR + 0x288)
-+#define CLASS_PE5_GP (CLASS_CSR_BASE_ADDR + 0x28c)
-+
-+#define CLASS_PE_INT_SRC (CLASS_CSR_BASE_ADDR + 0x290)
-+#define CLASS_PE_INT_ENABLE (CLASS_CSR_BASE_ADDR + 0x294)
-+
-+#define CLASS_TPID0_TPID1 (CLASS_CSR_BASE_ADDR + 0x298)
-+#define CLASS_TPID2 (CLASS_CSR_BASE_ADDR + 0x29c)
-+
-+#define CLASS_L4_CHKSUM_ADDR (CLASS_CSR_BASE_ADDR + 0x2a0)
-+
-+#define CLASS_PE0_DEBUG (CLASS_CSR_BASE_ADDR + 0x2a4)
-+#define CLASS_PE1_DEBUG (CLASS_CSR_BASE_ADDR + 0x2a8)
-+#define CLASS_PE2_DEBUG (CLASS_CSR_BASE_ADDR + 0x2ac)
-+#define CLASS_PE3_DEBUG (CLASS_CSR_BASE_ADDR + 0x2b0)
-+#define CLASS_PE4_DEBUG (CLASS_CSR_BASE_ADDR + 0x2b4)
-+#define CLASS_PE5_DEBUG (CLASS_CSR_BASE_ADDR + 0x2b8)
-+
-+#define CLASS_STATE (CLASS_CSR_BASE_ADDR + 0x2bc)
-+
-+/* CLASS defines */
-+#define CLASS_PBUF_SIZE 0x100 /* Fixed by hardware */
-+#define CLASS_PBUF_HEADER_OFFSET 0x80 /* Can be configured */
-+
-+/* Can be configured */
-+#define CLASS_PBUF0_BASE_ADDR 0x000
-+/* Can be configured */
-+#define CLASS_PBUF1_BASE_ADDR (CLASS_PBUF0_BASE_ADDR + CLASS_PBUF_SIZE)
-+/* Can be configured */
-+#define CLASS_PBUF2_BASE_ADDR (CLASS_PBUF1_BASE_ADDR + CLASS_PBUF_SIZE)
-+/* Can be configured */
-+#define CLASS_PBUF3_BASE_ADDR (CLASS_PBUF2_BASE_ADDR + CLASS_PBUF_SIZE)
-+
-+#define CLASS_PBUF0_HEADER_BASE_ADDR (CLASS_PBUF0_BASE_ADDR + \
-+ CLASS_PBUF_HEADER_OFFSET)
-+#define CLASS_PBUF1_HEADER_BASE_ADDR (CLASS_PBUF1_BASE_ADDR + \
-+ CLASS_PBUF_HEADER_OFFSET)
-+#define CLASS_PBUF2_HEADER_BASE_ADDR (CLASS_PBUF2_BASE_ADDR + \
-+ CLASS_PBUF_HEADER_OFFSET)
-+#define CLASS_PBUF3_HEADER_BASE_ADDR (CLASS_PBUF3_BASE_ADDR + \
-+ CLASS_PBUF_HEADER_OFFSET)
-+
-+#define CLASS_PE0_RO_DM_ADDR0_VAL ((CLASS_PBUF1_BASE_ADDR << 16) | \
-+ CLASS_PBUF0_BASE_ADDR)
-+#define CLASS_PE0_RO_DM_ADDR1_VAL ((CLASS_PBUF3_BASE_ADDR << 16) | \
-+ CLASS_PBUF2_BASE_ADDR)
-+
-+#define CLASS_PE0_QB_DM_ADDR0_VAL ((CLASS_PBUF1_HEADER_BASE_ADDR << 16) |\
-+ CLASS_PBUF0_HEADER_BASE_ADDR)
-+#define CLASS_PE0_QB_DM_ADDR1_VAL ((CLASS_PBUF3_HEADER_BASE_ADDR << 16) |\
-+ CLASS_PBUF2_HEADER_BASE_ADDR)
-+
-+#define CLASS_ROUTE_SIZE 128
-+#define CLASS_MAX_ROUTE_SIZE 256
-+#define CLASS_ROUTE_HASH_BITS 20
-+#define CLASS_ROUTE_HASH_MASK (BIT(CLASS_ROUTE_HASH_BITS) - 1)
-+
-+/* Can be configured */
-+#define CLASS_ROUTE0_BASE_ADDR 0x400
-+/* Can be configured */
-+#define CLASS_ROUTE1_BASE_ADDR (CLASS_ROUTE0_BASE_ADDR + CLASS_ROUTE_SIZE)
-+/* Can be configured */
-+#define CLASS_ROUTE2_BASE_ADDR (CLASS_ROUTE1_BASE_ADDR + CLASS_ROUTE_SIZE)
-+/* Can be configured */
-+#define CLASS_ROUTE3_BASE_ADDR (CLASS_ROUTE2_BASE_ADDR + CLASS_ROUTE_SIZE)
-+
-+#define CLASS_SA_SIZE 128
-+#define CLASS_IPSEC_SA0_BASE_ADDR 0x600
-+/* not used */
-+#define CLASS_IPSEC_SA1_BASE_ADDR (CLASS_IPSEC_SA0_BASE_ADDR + CLASS_SA_SIZE)
-+/* not used */
-+#define CLASS_IPSEC_SA2_BASE_ADDR (CLASS_IPSEC_SA1_BASE_ADDR + CLASS_SA_SIZE)
-+/* not used */
-+#define CLASS_IPSEC_SA3_BASE_ADDR (CLASS_IPSEC_SA2_BASE_ADDR + CLASS_SA_SIZE)
-+
-+/* generic purpose free dmem buffer, last portion of 2K dmem pbuf */
-+#define CLASS_GP_DMEM_BUF_SIZE (2048 - (CLASS_PBUF_SIZE * 4) - \
-+ (CLASS_ROUTE_SIZE * 4) - (CLASS_SA_SIZE))
-+#define CLASS_GP_DMEM_BUF ((void *)(CLASS_IPSEC_SA0_BASE_ADDR + \
-+ CLASS_SA_SIZE))
-+
-+#define TWO_LEVEL_ROUTE BIT(0)
-+#define PHYNO_IN_HASH BIT(1)
-+#define HW_ROUTE_FETCH BIT(3)
-+#define HW_BRIDGE_FETCH BIT(5)
-+#define IP_ALIGNED BIT(6)
-+#define ARC_HIT_CHECK_EN BIT(7)
-+#define CLASS_TOE BIT(11)
-+#define HASH_NORMAL (0 << 12)
-+#define HASH_CRC_PORT BIT(12)
-+#define HASH_CRC_IP (2 << 12)
-+#define HASH_CRC_PORT_IP (3 << 12)
-+#define QB2BUS_LE BIT(15)
-+
-+#define TCP_CHKSUM_DROP BIT(0)
-+#define UDP_CHKSUM_DROP BIT(1)
-+#define IPV4_CHKSUM_DROP BIT(9)
-+
-+/*CLASS_HIF_PARSE bits*/
-+#define HIF_PKT_CLASS_EN BIT(0)
-+#define HIF_PKT_OFFSET(ofst) (((ofst) & 0xF) << 1)
-+
-+struct class_cfg {
-+ u32 toe_mode;
-+ unsigned long route_table_baseaddr;
-+ u32 route_table_hash_bits;
-+ u32 pe_sys_clk_ratio;
-+ u32 resume;
-+};
-+
-+#endif /* _CLASS_CSR_H_ */
---- /dev/null
-+++ b/drivers/staging/fsl_ppfe/include/pfe/cbus/emac_mtip.h
-@@ -0,0 +1,242 @@
-+/*
-+ * Copyright 2015-2016 Freescale Semiconductor, Inc.
-+ * Copyright 2017 NXP
-+ *
-+ * This program is free software; you can redistribute it and/or modify
-+ * it under the terms of the GNU General Public License as published by
-+ * the Free Software Foundation; either version 2 of the License, or
-+ * (at your option) any later version.
-+ *
-+ * This program is distributed in the hope that it will be useful,
-+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
-+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
-+ * GNU General Public License for more details.
-+ *
-+ * You should have received a copy of the GNU General Public License
-+ * along with this program. If not, see <http://www.gnu.org/licenses/>.
-+ */
-+
-+#ifndef _EMAC_H_
-+#define _EMAC_H_
-+
-+#include <linux/ethtool.h>
-+
-+#define EMAC_IEVENT_REG 0x004
-+#define EMAC_IMASK_REG 0x008
-+#define EMAC_R_DES_ACTIVE_REG 0x010
-+#define EMAC_X_DES_ACTIVE_REG 0x014
-+#define EMAC_ECNTRL_REG 0x024
-+#define EMAC_MII_DATA_REG 0x040
-+#define EMAC_MII_CTRL_REG 0x044
-+#define EMAC_MIB_CTRL_STS_REG 0x064
-+#define EMAC_RCNTRL_REG 0x084
-+#define EMAC_TCNTRL_REG 0x0C4
-+#define EMAC_PHY_ADDR_LOW 0x0E4
-+#define EMAC_PHY_ADDR_HIGH 0x0E8
-+#define EMAC_GAUR 0x120
-+#define EMAC_GALR 0x124
-+#define EMAC_TFWR_STR_FWD 0x144
-+#define EMAC_RX_SECTION_FULL 0x190
-+#define EMAC_RX_SECTION_EMPTY 0x194
-+#define EMAC_TX_SECTION_EMPTY 0x1A0
-+#define EMAC_TRUNC_FL 0x1B0
-+
-+#define RMON_T_DROP 0x200 /* Count of frames not cntd correctly */
-+#define RMON_T_PACKETS 0x204 /* RMON TX packet count */
-+#define RMON_T_BC_PKT 0x208 /* RMON TX broadcast pkts */
-+#define RMON_T_MC_PKT 0x20c /* RMON TX multicast pkts */
-+#define RMON_T_CRC_ALIGN 0x210 /* RMON TX pkts with CRC align err */
-+#define RMON_T_UNDERSIZE 0x214 /* RMON TX pkts < 64 bytes, good CRC */
-+#define RMON_T_OVERSIZE 0x218 /* RMON TX pkts > MAX_FL bytes good CRC */
-+#define RMON_T_FRAG 0x21c /* RMON TX pkts < 64 bytes, bad CRC */
-+#define RMON_T_JAB 0x220 /* RMON TX pkts > MAX_FL bytes, bad CRC */
-+#define RMON_T_COL 0x224 /* RMON TX collision count */
-+#define RMON_T_P64 0x228 /* RMON TX 64 byte pkts */
-+#define RMON_T_P65TO127 0x22c /* RMON TX 65 to 127 byte pkts */
-+#define RMON_T_P128TO255 0x230 /* RMON TX 128 to 255 byte pkts */
-+#define RMON_T_P256TO511 0x234 /* RMON TX 256 to 511 byte pkts */
-+#define RMON_T_P512TO1023 0x238 /* RMON TX 512 to 1023 byte pkts */
-+#define RMON_T_P1024TO2047 0x23c /* RMON TX 1024 to 2047 byte pkts */
-+#define RMON_T_P_GTE2048 0x240 /* RMON TX pkts > 2048 bytes */
-+#define RMON_T_OCTETS 0x244 /* RMON TX octets */
-+#define IEEE_T_DROP 0x248 /* Count of frames not counted crtly */
-+#define IEEE_T_FRAME_OK 0x24c /* Frames tx'd OK */
-+#define IEEE_T_1COL 0x250 /* Frames tx'd with single collision */
-+#define IEEE_T_MCOL 0x254 /* Frames tx'd with multiple collision */
-+#define IEEE_T_DEF 0x258 /* Frames tx'd after deferral delay */
-+#define IEEE_T_LCOL 0x25c /* Frames tx'd with late collision */
-+#define IEEE_T_EXCOL 0x260 /* Frames tx'd with excesv collisions */
-+#define IEEE_T_MACERR 0x264 /* Frames tx'd with TX FIFO underrun */
-+#define IEEE_T_CSERR 0x268 /* Frames tx'd with carrier sense err */
-+#define IEEE_T_SQE 0x26c /* Frames tx'd with SQE err */
-+#define IEEE_T_FDXFC 0x270 /* Flow control pause frames tx'd */
-+#define IEEE_T_OCTETS_OK 0x274 /* Octet count for frames tx'd w/o err */
-+#define RMON_R_PACKETS 0x284 /* RMON RX packet count */
-+#define RMON_R_BC_PKT 0x288 /* RMON RX broadcast pkts */
-+#define RMON_R_MC_PKT 0x28c /* RMON RX multicast pkts */
-+#define RMON_R_CRC_ALIGN 0x290 /* RMON RX pkts with CRC alignment err */
-+#define RMON_R_UNDERSIZE 0x294 /* RMON RX pkts < 64 bytes, good CRC */
-+#define RMON_R_OVERSIZE 0x298 /* RMON RX pkts > MAX_FL bytes good CRC */
-+#define RMON_R_FRAG 0x29c /* RMON RX pkts < 64 bytes, bad CRC */
-+#define RMON_R_JAB 0x2a0 /* RMON RX pkts > MAX_FL bytes, bad CRC */
-+#define RMON_R_RESVD_O 0x2a4 /* Reserved */
-+#define RMON_R_P64 0x2a8 /* RMON RX 64 byte pkts */
-+#define RMON_R_P65TO127 0x2ac /* RMON RX 65 to 127 byte pkts */
-+#define RMON_R_P128TO255 0x2b0 /* RMON RX 128 to 255 byte pkts */
-+#define RMON_R_P256TO511 0x2b4 /* RMON RX 256 to 511 byte pkts */
-+#define RMON_R_P512TO1023 0x2b8 /* RMON RX 512 to 1023 byte pkts */
-+#define RMON_R_P1024TO2047 0x2bc /* RMON RX 1024 to 2047 byte pkts */
-+#define RMON_R_P_GTE2048 0x2c0 /* RMON RX pkts > 2048 bytes */
-+#define RMON_R_OCTETS 0x2c4 /* RMON RX octets */
-+#define IEEE_R_DROP 0x2c8 /* Count frames not counted correctly */
-+#define IEEE_R_FRAME_OK 0x2cc /* Frames rx'd OK */
-+#define IEEE_R_CRC 0x2d0 /* Frames rx'd with CRC err */
-+#define IEEE_R_ALIGN 0x2d4 /* Frames rx'd with alignment err */
-+#define IEEE_R_MACERR 0x2d8 /* Receive FIFO overflow count */
-+#define IEEE_R_FDXFC 0x2dc /* Flow control pause frames rx'd */
-+#define IEEE_R_OCTETS_OK 0x2e0 /* Octet cnt for frames rx'd w/o err */
-+
-+#define EMAC_SMAC_0_0 0x500 /*Supplemental MAC Address 0 (RW).*/
-+#define EMAC_SMAC_0_1 0x504 /*Supplemental MAC Address 0 (RW).*/
-+
-+/* GEMAC definitions and settings */
-+
-+#define EMAC_PORT_0 0
-+#define EMAC_PORT_1 1
-+
-+/* GEMAC Bit definitions */
-+#define EMAC_IEVENT_HBERR 0x80000000
-+#define EMAC_IEVENT_BABR 0x40000000
-+#define EMAC_IEVENT_BABT 0x20000000
-+#define EMAC_IEVENT_GRA 0x10000000
-+#define EMAC_IEVENT_TXF 0x08000000
-+#define EMAC_IEVENT_TXB 0x04000000
-+#define EMAC_IEVENT_RXF 0x02000000
-+#define EMAC_IEVENT_RXB 0x01000000
-+#define EMAC_IEVENT_MII 0x00800000
-+#define EMAC_IEVENT_EBERR 0x00400000
-+#define EMAC_IEVENT_LC 0x00200000
-+#define EMAC_IEVENT_RL 0x00100000
-+#define EMAC_IEVENT_UN 0x00080000
-+
-+#define EMAC_IMASK_HBERR 0x80000000
-+#define EMAC_IMASK_BABR 0x40000000
-+#define EMAC_IMASKT_BABT 0x20000000
-+#define EMAC_IMASK_GRA 0x10000000
-+#define EMAC_IMASKT_TXF 0x08000000
-+#define EMAC_IMASK_TXB 0x04000000
-+#define EMAC_IMASKT_RXF 0x02000000
-+#define EMAC_IMASK_RXB 0x01000000
-+#define EMAC_IMASK_MII 0x00800000
-+#define EMAC_IMASK_EBERR 0x00400000
-+#define EMAC_IMASK_LC 0x00200000
-+#define EMAC_IMASKT_RL 0x00100000
-+#define EMAC_IMASK_UN 0x00080000
-+
-+#define EMAC_RCNTRL_MAX_FL_SHIFT 16
-+#define EMAC_RCNTRL_LOOP 0x00000001
-+#define EMAC_RCNTRL_DRT 0x00000002
-+#define EMAC_RCNTRL_MII_MODE 0x00000004
-+#define EMAC_RCNTRL_PROM 0x00000008
-+#define EMAC_RCNTRL_BC_REJ 0x00000010
-+#define EMAC_RCNTRL_FCE 0x00000020
-+#define EMAC_RCNTRL_RGMII 0x00000040
-+#define EMAC_RCNTRL_SGMII 0x00000080
-+#define EMAC_RCNTRL_RMII 0x00000100
-+#define EMAC_RCNTRL_RMII_10T 0x00000200
-+#define EMAC_RCNTRL_CRC_FWD 0x00004000
-+
-+#define EMAC_TCNTRL_GTS 0x00000001
-+#define EMAC_TCNTRL_HBC 0x00000002
-+#define EMAC_TCNTRL_FDEN 0x00000004
-+#define EMAC_TCNTRL_TFC_PAUSE 0x00000008
-+#define EMAC_TCNTRL_RFC_PAUSE 0x00000010
-+
-+#define EMAC_ECNTRL_RESET 0x00000001 /* reset the EMAC */
-+#define EMAC_ECNTRL_ETHER_EN 0x00000002 /* enable the EMAC */
-+#define EMAC_ECNTRL_MAGIC_ENA 0x00000004
-+#define EMAC_ECNTRL_SLEEP 0x00000008
-+#define EMAC_ECNTRL_SPEED 0x00000020
-+#define EMAC_ECNTRL_DBSWAP 0x00000100
-+
-+#define EMAC_X_WMRK_STRFWD 0x00000100
-+
-+#define EMAC_X_DES_ACTIVE_TDAR 0x01000000
-+#define EMAC_R_DES_ACTIVE_RDAR 0x01000000
-+
-+#define EMAC_RX_SECTION_EMPTY_V 0x00010006
-+/*
-+ * The possible operating speeds of the MAC, currently supporting 10, 100 and
-+ * 1000Mb modes.
-+ */
-+enum mac_speed {SPEED_10M, SPEED_100M, SPEED_1000M, SPEED_1000M_PCS};
-+
-+/* MII-related definitios */
-+#define EMAC_MII_DATA_ST 0x40000000 /* Start of frame delimiter */
-+#define EMAC_MII_DATA_OP_RD 0x20000000 /* Perform a read operation */
-+#define EMAC_MII_DATA_OP_CL45_RD 0x30000000 /* Perform a read operation */
-+#define EMAC_MII_DATA_OP_WR 0x10000000 /* Perform a write operation */
-+#define EMAC_MII_DATA_OP_CL45_WR 0x10000000 /* Perform a write operation */
-+#define EMAC_MII_DATA_PA_MSK 0x0f800000 /* PHY Address field mask */
-+#define EMAC_MII_DATA_RA_MSK 0x007c0000 /* PHY Register field mask */
-+#define EMAC_MII_DATA_TA 0x00020000 /* Turnaround */
-+#define EMAC_MII_DATA_DATAMSK 0x0000ffff /* PHY data field */
-+
-+#define EMAC_MII_DATA_RA_SHIFT 18 /* MII Register address bits */
-+#define EMAC_MII_DATA_RA_MASK 0x1F /* MII Register address mask */
-+#define EMAC_MII_DATA_PA_SHIFT 23 /* MII PHY address bits */
-+#define EMAC_MII_DATA_PA_MASK 0x1F /* MII PHY address mask */
-+
-+#define EMAC_MII_DATA_RA(v) (((v) & EMAC_MII_DATA_RA_MASK) << \
-+ EMAC_MII_DATA_RA_SHIFT)
-+#define EMAC_MII_DATA_PA(v) (((v) & EMAC_MII_DATA_RA_MASK) << \
-+ EMAC_MII_DATA_PA_SHIFT)
-+#define EMAC_MII_DATA(v) ((v) & 0xffff)
-+
-+#define EMAC_MII_SPEED_SHIFT 1
-+#define EMAC_HOLDTIME_SHIFT 8
-+#define EMAC_HOLDTIME_MASK 0x7
-+#define EMAC_HOLDTIME(v) (((v) & EMAC_HOLDTIME_MASK) << \
-+ EMAC_HOLDTIME_SHIFT)
-+
-+/*
-+ * The Address organisation for the MAC device. All addresses are split into
-+ * two 32-bit register fields. The first one (bottom) is the lower 32-bits of
-+ * the address and the other field are the high order bits - this may be 16-bits
-+ * in the case of MAC addresses, or 32-bits for the hash address.
-+ * In terms of memory storage, the first item (bottom) is assumed to be at a
-+ * lower address location than 'top'. i.e. top should be at address location of
-+ * 'bottom' + 4 bytes.
-+ */
-+struct pfe_mac_addr {
-+ u32 bottom; /* Lower 32-bits of address. */
-+ u32 top; /* Upper 32-bits of address. */
-+};
-+
-+/*
-+ * The following is the organisation of the address filters section of the MAC
-+ * registers. The Cadence MAC contains four possible specific address match
-+ * addresses, if an incoming frame corresponds to any one of these four
-+ * addresses then the frame will be copied to memory.
-+ * It is not necessary for all four of the address match registers to be
-+ * programmed, this is application dependent.
-+ */
-+struct spec_addr {
-+ struct pfe_mac_addr one; /* Specific address register 1. */
-+ struct pfe_mac_addr two; /* Specific address register 2. */
-+ struct pfe_mac_addr three; /* Specific address register 3. */
-+ struct pfe_mac_addr four; /* Specific address register 4. */
-+};
-+
-+struct gemac_cfg {
-+ u32 mode;
-+ u32 speed;
-+ u32 duplex;
-+};
-+
-+/* EMAC Hash size */
-+#define EMAC_HASH_REG_BITS 64
-+
-+#define EMAC_SPEC_ADDR_MAX 4
-+
-+#endif /* _EMAC_H_ */
---- /dev/null
-+++ b/drivers/staging/fsl_ppfe/include/pfe/cbus/gpi.h
-@@ -0,0 +1,86 @@
-+/*
-+ * Copyright 2015-2016 Freescale Semiconductor, Inc.
-+ * Copyright 2017 NXP
-+ *
-+ * This program is free software; you can redistribute it and/or modify
-+ * it under the terms of the GNU General Public License as published by
-+ * the Free Software Foundation; either version 2 of the License, or
-+ * (at your option) any later version.
-+ *
-+ * This program is distributed in the hope that it will be useful,
-+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
-+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
-+ * GNU General Public License for more details.
-+ *
-+ * You should have received a copy of the GNU General Public License
-+ * along with this program. If not, see <http://www.gnu.org/licenses/>.
-+ */
-+
-+#ifndef _GPI_H_
-+#define _GPI_H_
-+
-+#define GPI_VERSION 0x00
-+#define GPI_CTRL 0x04
-+#define GPI_RX_CONFIG 0x08
-+#define GPI_HDR_SIZE 0x0c
-+#define GPI_BUF_SIZE 0x10
-+#define GPI_LMEM_ALLOC_ADDR 0x14
-+#define GPI_LMEM_FREE_ADDR 0x18
-+#define GPI_DDR_ALLOC_ADDR 0x1c
-+#define GPI_DDR_FREE_ADDR 0x20
-+#define GPI_CLASS_ADDR 0x24
-+#define GPI_DRX_FIFO 0x28
-+#define GPI_TRX_FIFO 0x2c
-+#define GPI_INQ_PKTPTR 0x30
-+#define GPI_DDR_DATA_OFFSET 0x34
-+#define GPI_LMEM_DATA_OFFSET 0x38
-+#define GPI_TMLF_TX 0x4c
-+#define GPI_DTX_ASEQ 0x50
-+#define GPI_FIFO_STATUS 0x54
-+#define GPI_FIFO_DEBUG 0x58
-+#define GPI_TX_PAUSE_TIME 0x5c
-+#define GPI_LMEM_SEC_BUF_DATA_OFFSET 0x60
-+#define GPI_DDR_SEC_BUF_DATA_OFFSET 0x64
-+#define GPI_TOE_CHKSUM_EN 0x68
-+#define GPI_OVERRUN_DROPCNT 0x6c
-+#define GPI_CSR_MTIP_PAUSE_REG 0x74
-+#define GPI_CSR_MTIP_PAUSE_QUANTUM 0x78
-+#define GPI_CSR_RX_CNT 0x7c
-+#define GPI_CSR_TX_CNT 0x80
-+#define GPI_CSR_DEBUG1 0x84
-+#define GPI_CSR_DEBUG2 0x88
-+
-+struct gpi_cfg {
-+ u32 lmem_rtry_cnt;
-+ u32 tmlf_txthres;
-+ u32 aseq_len;
-+ u32 mtip_pause_reg;
-+};
-+
-+/* GPI commons defines */
-+#define GPI_LMEM_BUF_EN 0x1
-+#define GPI_DDR_BUF_EN 0x1
-+
-+/* EGPI 1 defines */
-+#define EGPI1_LMEM_RTRY_CNT 0x40
-+#define EGPI1_TMLF_TXTHRES 0xBC
-+#define EGPI1_ASEQ_LEN 0x50
-+
-+/* EGPI 2 defines */
-+#define EGPI2_LMEM_RTRY_CNT 0x40
-+#define EGPI2_TMLF_TXTHRES 0xBC
-+#define EGPI2_ASEQ_LEN 0x40
-+
-+/* EGPI 3 defines */
-+#define EGPI3_LMEM_RTRY_CNT 0x40
-+#define EGPI3_TMLF_TXTHRES 0xBC
-+#define EGPI3_ASEQ_LEN 0x40
-+
-+/* HGPI defines */
-+#define HGPI_LMEM_RTRY_CNT 0x40
-+#define HGPI_TMLF_TXTHRES 0xBC
-+#define HGPI_ASEQ_LEN 0x40
-+
-+#define EGPI_PAUSE_TIME 0x000007D0
-+#define EGPI_PAUSE_ENABLE 0x40000000
-+#endif /* _GPI_H_ */
---- /dev/null
-+++ b/drivers/staging/fsl_ppfe/include/pfe/cbus/hif.h
-@@ -0,0 +1,100 @@
-+/*
-+ * Copyright 2015-2016 Freescale Semiconductor, Inc.
-+ * Copyright 2017 NXP
-+ *
-+ * This program is free software; you can redistribute it and/or modify
-+ * it under the terms of the GNU General Public License as published by
-+ * the Free Software Foundation; either version 2 of the License, or
-+ * (at your option) any later version.
-+ *
-+ * This program is distributed in the hope that it will be useful,
-+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
-+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
-+ * GNU General Public License for more details.
-+ *
-+ * You should have received a copy of the GNU General Public License
-+ * along with this program. If not, see <http://www.gnu.org/licenses/>.
-+ */
-+
-+#ifndef _HIF_H_
-+#define _HIF_H_
-+
-+/* @file hif.h.
-+ * hif - PFE hif block control and status register.
-+ * Mapped on CBUS and accessible from all PE's and ARM.
-+ */
-+#define HIF_VERSION (HIF_BASE_ADDR + 0x00)
-+#define HIF_TX_CTRL (HIF_BASE_ADDR + 0x04)
-+#define HIF_TX_CURR_BD_ADDR (HIF_BASE_ADDR + 0x08)
-+#define HIF_TX_ALLOC (HIF_BASE_ADDR + 0x0c)
-+#define HIF_TX_BDP_ADDR (HIF_BASE_ADDR + 0x10)
-+#define HIF_TX_STATUS (HIF_BASE_ADDR + 0x14)
-+#define HIF_RX_CTRL (HIF_BASE_ADDR + 0x20)
-+#define HIF_RX_BDP_ADDR (HIF_BASE_ADDR + 0x24)
-+#define HIF_RX_STATUS (HIF_BASE_ADDR + 0x30)
-+#define HIF_INT_SRC (HIF_BASE_ADDR + 0x34)
-+#define HIF_INT_ENABLE (HIF_BASE_ADDR + 0x38)
-+#define HIF_POLL_CTRL (HIF_BASE_ADDR + 0x3c)
-+#define HIF_RX_CURR_BD_ADDR (HIF_BASE_ADDR + 0x40)
-+#define HIF_RX_ALLOC (HIF_BASE_ADDR + 0x44)
-+#define HIF_TX_DMA_STATUS (HIF_BASE_ADDR + 0x48)
-+#define HIF_RX_DMA_STATUS (HIF_BASE_ADDR + 0x4c)
-+#define HIF_INT_COAL (HIF_BASE_ADDR + 0x50)
-+
-+/* HIF_INT_SRC/ HIF_INT_ENABLE control bits */
-+#define HIF_INT BIT(0)
-+#define HIF_RXBD_INT BIT(1)
-+#define HIF_RXPKT_INT BIT(2)
-+#define HIF_TXBD_INT BIT(3)
-+#define HIF_TXPKT_INT BIT(4)
-+
-+/* HIF_TX_CTRL bits */
-+#define HIF_CTRL_DMA_EN BIT(0)
-+#define HIF_CTRL_BDP_POLL_CTRL_EN BIT(1)
-+#define HIF_CTRL_BDP_CH_START_WSTB BIT(2)
-+
-+/* HIF_RX_STATUS bits */
-+#define BDP_CSR_RX_DMA_ACTV BIT(16)
-+
-+/* HIF_INT_ENABLE bits */
-+#define HIF_INT_EN BIT(0)
-+#define HIF_RXBD_INT_EN BIT(1)
-+#define HIF_RXPKT_INT_EN BIT(2)
-+#define HIF_TXBD_INT_EN BIT(3)
-+#define HIF_TXPKT_INT_EN BIT(4)
-+
-+/* HIF_POLL_CTRL bits*/
-+#define HIF_RX_POLL_CTRL_CYCLE 0x0400
-+#define HIF_TX_POLL_CTRL_CYCLE 0x0400
-+
-+/* HIF_INT_COAL bits*/
-+#define HIF_INT_COAL_ENABLE BIT(31)
-+
-+/* Buffer descriptor control bits */
-+#define BD_CTRL_BUFLEN_MASK 0x3fff
-+#define BD_BUF_LEN(x) ((x) & BD_CTRL_BUFLEN_MASK)
-+#define BD_CTRL_CBD_INT_EN BIT(16)
-+#define BD_CTRL_PKT_INT_EN BIT(17)
-+#define BD_CTRL_LIFM BIT(18)
-+#define BD_CTRL_LAST_BD BIT(19)
-+#define BD_CTRL_DIR BIT(20)
-+#define BD_CTRL_LMEM_CPY BIT(21) /* Valid only for HIF_NOCPY */
-+#define BD_CTRL_PKT_XFER BIT(24)
-+#define BD_CTRL_DESC_EN BIT(31)
-+#define BD_CTRL_PARSE_DISABLE BIT(25)
-+#define BD_CTRL_BRFETCH_DISABLE BIT(26)
-+#define BD_CTRL_RTFETCH_DISABLE BIT(27)
-+
-+/* Buffer descriptor status bits*/
-+#define BD_STATUS_CONN_ID(x) ((x) & 0xffff)
-+#define BD_STATUS_DIR_PROC_ID BIT(16)
-+#define BD_STATUS_CONN_ID_EN BIT(17)
-+#define BD_STATUS_PE2PROC_ID(x) (((x) & 7) << 18)
-+#define BD_STATUS_LE_DATA BIT(21)
-+#define BD_STATUS_CHKSUM_EN BIT(22)
-+
-+/* HIF Buffer descriptor status bits */
-+#define DIR_PROC_ID BIT(16)
-+#define PROC_ID(id) ((id) << 18)
-+
-+#endif /* _HIF_H_ */
---- /dev/null
-+++ b/drivers/staging/fsl_ppfe/include/pfe/cbus/hif_nocpy.h
-@@ -0,0 +1,50 @@
-+/*
-+ * Copyright 2015-2016 Freescale Semiconductor, Inc.
-+ * Copyright 2017 NXP
-+ *
-+ * This program is free software; you can redistribute it and/or modify
-+ * it under the terms of the GNU General Public License as published by
-+ * the Free Software Foundation; either version 2 of the License, or
-+ * (at your option) any later version.
-+ *
-+ * This program is distributed in the hope that it will be useful,
-+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
-+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
-+ * GNU General Public License for more details.
-+ *
-+ * You should have received a copy of the GNU General Public License
-+ * along with this program. If not, see <http://www.gnu.org/licenses/>.
-+ */
-+
-+#ifndef _HIF_NOCPY_H_
-+#define _HIF_NOCPY_H_
-+
-+#define HIF_NOCPY_VERSION (HIF_NOCPY_BASE_ADDR + 0x00)
-+#define HIF_NOCPY_TX_CTRL (HIF_NOCPY_BASE_ADDR + 0x04)
-+#define HIF_NOCPY_TX_CURR_BD_ADDR (HIF_NOCPY_BASE_ADDR + 0x08)
-+#define HIF_NOCPY_TX_ALLOC (HIF_NOCPY_BASE_ADDR + 0x0c)
-+#define HIF_NOCPY_TX_BDP_ADDR (HIF_NOCPY_BASE_ADDR + 0x10)
-+#define HIF_NOCPY_TX_STATUS (HIF_NOCPY_BASE_ADDR + 0x14)
-+#define HIF_NOCPY_RX_CTRL (HIF_NOCPY_BASE_ADDR + 0x20)
-+#define HIF_NOCPY_RX_BDP_ADDR (HIF_NOCPY_BASE_ADDR + 0x24)
-+#define HIF_NOCPY_RX_STATUS (HIF_NOCPY_BASE_ADDR + 0x30)
-+#define HIF_NOCPY_INT_SRC (HIF_NOCPY_BASE_ADDR + 0x34)
-+#define HIF_NOCPY_INT_ENABLE (HIF_NOCPY_BASE_ADDR + 0x38)
-+#define HIF_NOCPY_POLL_CTRL (HIF_NOCPY_BASE_ADDR + 0x3c)
-+#define HIF_NOCPY_RX_CURR_BD_ADDR (HIF_NOCPY_BASE_ADDR + 0x40)
-+#define HIF_NOCPY_RX_ALLOC (HIF_NOCPY_BASE_ADDR + 0x44)
-+#define HIF_NOCPY_TX_DMA_STATUS (HIF_NOCPY_BASE_ADDR + 0x48)
-+#define HIF_NOCPY_RX_DMA_STATUS (HIF_NOCPY_BASE_ADDR + 0x4c)
-+#define HIF_NOCPY_RX_INQ0_PKTPTR (HIF_NOCPY_BASE_ADDR + 0x50)
-+#define HIF_NOCPY_RX_INQ1_PKTPTR (HIF_NOCPY_BASE_ADDR + 0x54)
-+#define HIF_NOCPY_TX_PORT_NO (HIF_NOCPY_BASE_ADDR + 0x60)
-+#define HIF_NOCPY_LMEM_ALLOC_ADDR (HIF_NOCPY_BASE_ADDR + 0x64)
-+#define HIF_NOCPY_CLASS_ADDR (HIF_NOCPY_BASE_ADDR + 0x68)
-+#define HIF_NOCPY_TMU_PORT0_ADDR (HIF_NOCPY_BASE_ADDR + 0x70)
-+#define HIF_NOCPY_TMU_PORT1_ADDR (HIF_NOCPY_BASE_ADDR + 0x74)
-+#define HIF_NOCPY_TMU_PORT2_ADDR (HIF_NOCPY_BASE_ADDR + 0x7c)
-+#define HIF_NOCPY_TMU_PORT3_ADDR (HIF_NOCPY_BASE_ADDR + 0x80)
-+#define HIF_NOCPY_TMU_PORT4_ADDR (HIF_NOCPY_BASE_ADDR + 0x84)
-+#define HIF_NOCPY_INT_COAL (HIF_NOCPY_BASE_ADDR + 0x90)
-+
-+#endif /* _HIF_NOCPY_H_ */
---- /dev/null
-+++ b/drivers/staging/fsl_ppfe/include/pfe/cbus/tmu_csr.h
-@@ -0,0 +1,168 @@
-+/*
-+ * Copyright 2015-2016 Freescale Semiconductor, Inc.
-+ * Copyright 2017 NXP
-+ *
-+ * This program is free software; you can redistribute it and/or modify
-+ * it under the terms of the GNU General Public License as published by
-+ * the Free Software Foundation; either version 2 of the License, or
-+ * (at your option) any later version.
-+ *
-+ * This program is distributed in the hope that it will be useful,
-+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
-+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
-+ * GNU General Public License for more details.
-+ *
-+ * You should have received a copy of the GNU General Public License
-+ * along with this program. If not, see <http://www.gnu.org/licenses/>.
-+ */
-+
-+#ifndef _TMU_CSR_H_
-+#define _TMU_CSR_H_
-+
-+#define TMU_VERSION (TMU_CSR_BASE_ADDR + 0x000)
-+#define TMU_INQ_WATERMARK (TMU_CSR_BASE_ADDR + 0x004)
-+#define TMU_PHY_INQ_PKTPTR (TMU_CSR_BASE_ADDR + 0x008)
-+#define TMU_PHY_INQ_PKTINFO (TMU_CSR_BASE_ADDR + 0x00c)
-+#define TMU_PHY_INQ_FIFO_CNT (TMU_CSR_BASE_ADDR + 0x010)
-+#define TMU_SYS_GENERIC_CONTROL (TMU_CSR_BASE_ADDR + 0x014)
-+#define TMU_SYS_GENERIC_STATUS (TMU_CSR_BASE_ADDR + 0x018)
-+#define TMU_SYS_GEN_CON0 (TMU_CSR_BASE_ADDR + 0x01c)
-+#define TMU_SYS_GEN_CON1 (TMU_CSR_BASE_ADDR + 0x020)
-+#define TMU_SYS_GEN_CON2 (TMU_CSR_BASE_ADDR + 0x024)
-+#define TMU_SYS_GEN_CON3 (TMU_CSR_BASE_ADDR + 0x028)
-+#define TMU_SYS_GEN_CON4 (TMU_CSR_BASE_ADDR + 0x02c)
-+#define TMU_TEQ_DISABLE_DROPCHK (TMU_CSR_BASE_ADDR + 0x030)
-+#define TMU_TEQ_CTRL (TMU_CSR_BASE_ADDR + 0x034)
-+#define TMU_TEQ_QCFG (TMU_CSR_BASE_ADDR + 0x038)
-+#define TMU_TEQ_DROP_STAT (TMU_CSR_BASE_ADDR + 0x03c)
-+#define TMU_TEQ_QAVG (TMU_CSR_BASE_ADDR + 0x040)
-+#define TMU_TEQ_WREG_PROB (TMU_CSR_BASE_ADDR + 0x044)
-+#define TMU_TEQ_TRANS_STAT (TMU_CSR_BASE_ADDR + 0x048)
-+#define TMU_TEQ_HW_PROB_CFG0 (TMU_CSR_BASE_ADDR + 0x04c)
-+#define TMU_TEQ_HW_PROB_CFG1 (TMU_CSR_BASE_ADDR + 0x050)
-+#define TMU_TEQ_HW_PROB_CFG2 (TMU_CSR_BASE_ADDR + 0x054)
-+#define TMU_TEQ_HW_PROB_CFG3 (TMU_CSR_BASE_ADDR + 0x058)
-+#define TMU_TEQ_HW_PROB_CFG4 (TMU_CSR_BASE_ADDR + 0x05c)
-+#define TMU_TEQ_HW_PROB_CFG5 (TMU_CSR_BASE_ADDR + 0x060)
-+#define TMU_TEQ_HW_PROB_CFG6 (TMU_CSR_BASE_ADDR + 0x064)
-+#define TMU_TEQ_HW_PROB_CFG7 (TMU_CSR_BASE_ADDR + 0x068)
-+#define TMU_TEQ_HW_PROB_CFG8 (TMU_CSR_BASE_ADDR + 0x06c)
-+#define TMU_TEQ_HW_PROB_CFG9 (TMU_CSR_BASE_ADDR + 0x070)
-+#define TMU_TEQ_HW_PROB_CFG10 (TMU_CSR_BASE_ADDR + 0x074)
-+#define TMU_TEQ_HW_PROB_CFG11 (TMU_CSR_BASE_ADDR + 0x078)
-+#define TMU_TEQ_HW_PROB_CFG12 (TMU_CSR_BASE_ADDR + 0x07c)
-+#define TMU_TEQ_HW_PROB_CFG13 (TMU_CSR_BASE_ADDR + 0x080)
-+#define TMU_TEQ_HW_PROB_CFG14 (TMU_CSR_BASE_ADDR + 0x084)
-+#define TMU_TEQ_HW_PROB_CFG15 (TMU_CSR_BASE_ADDR + 0x088)
-+#define TMU_TEQ_HW_PROB_CFG16 (TMU_CSR_BASE_ADDR + 0x08c)
-+#define TMU_TEQ_HW_PROB_CFG17 (TMU_CSR_BASE_ADDR + 0x090)
-+#define TMU_TEQ_HW_PROB_CFG18 (TMU_CSR_BASE_ADDR + 0x094)
-+#define TMU_TEQ_HW_PROB_CFG19 (TMU_CSR_BASE_ADDR + 0x098)
-+#define TMU_TEQ_HW_PROB_CFG20 (TMU_CSR_BASE_ADDR + 0x09c)
-+#define TMU_TEQ_HW_PROB_CFG21 (TMU_CSR_BASE_ADDR + 0x0a0)
-+#define TMU_TEQ_HW_PROB_CFG22 (TMU_CSR_BASE_ADDR + 0x0a4)
-+#define TMU_TEQ_HW_PROB_CFG23 (TMU_CSR_BASE_ADDR + 0x0a8)
-+#define TMU_TEQ_HW_PROB_CFG24 (TMU_CSR_BASE_ADDR + 0x0ac)
-+#define TMU_TEQ_HW_PROB_CFG25 (TMU_CSR_BASE_ADDR + 0x0b0)
-+#define TMU_TDQ_IIFG_CFG (TMU_CSR_BASE_ADDR + 0x0b4)
-+/* [9:0] Scheduler Enable for each of the scheduler in the TDQ.
-+ * This is a global Enable for all schedulers in PHY0
-+ */
-+#define TMU_TDQ0_SCH_CTRL (TMU_CSR_BASE_ADDR + 0x0b8)
-+
-+#define TMU_LLM_CTRL (TMU_CSR_BASE_ADDR + 0x0bc)
-+#define TMU_LLM_BASE_ADDR (TMU_CSR_BASE_ADDR + 0x0c0)
-+#define TMU_LLM_QUE_LEN (TMU_CSR_BASE_ADDR + 0x0c4)
-+#define TMU_LLM_QUE_HEADPTR (TMU_CSR_BASE_ADDR + 0x0c8)
-+#define TMU_LLM_QUE_TAILPTR (TMU_CSR_BASE_ADDR + 0x0cc)
-+#define TMU_LLM_QUE_DROPCNT (TMU_CSR_BASE_ADDR + 0x0d0)
-+#define TMU_INT_EN (TMU_CSR_BASE_ADDR + 0x0d4)
-+#define TMU_INT_SRC (TMU_CSR_BASE_ADDR + 0x0d8)
-+#define TMU_INQ_STAT (TMU_CSR_BASE_ADDR + 0x0dc)
-+#define TMU_CTRL (TMU_CSR_BASE_ADDR + 0x0e0)
-+
-+/* [31] Mem Access Command. 0 = Internal Memory Read, 1 = Internal memory
-+ * Write [27:24] Byte Enables of the Internal memory access [23:0] Address of
-+ * the internal memory. This address is used to access both the PM and DM of
-+ * all the PE's
-+ */
-+#define TMU_MEM_ACCESS_ADDR (TMU_CSR_BASE_ADDR + 0x0e4)
-+
-+/* Internal Memory Access Write Data */
-+#define TMU_MEM_ACCESS_WDATA (TMU_CSR_BASE_ADDR + 0x0e8)
-+/* Internal Memory Access Read Data. The commands are blocked
-+ * at the mem_access only
-+ */
-+#define TMU_MEM_ACCESS_RDATA (TMU_CSR_BASE_ADDR + 0x0ec)
-+
-+/* [31:0] PHY0 in queue address (must be initialized with one of the
-+ * xxx_INQ_PKTPTR cbus addresses)
-+ */
-+#define TMU_PHY0_INQ_ADDR (TMU_CSR_BASE_ADDR + 0x0f0)
-+/* [31:0] PHY1 in queue address (must be initialized with one of the
-+ * xxx_INQ_PKTPTR cbus addresses)
-+ */
-+#define TMU_PHY1_INQ_ADDR (TMU_CSR_BASE_ADDR + 0x0f4)
-+/* [31:0] PHY2 in queue address (must be initialized with one of the
-+ * xxx_INQ_PKTPTR cbus addresses)
-+ */
-+#define TMU_PHY2_INQ_ADDR (TMU_CSR_BASE_ADDR + 0x0f8)
-+/* [31:0] PHY3 in queue address (must be initialized with one of the
-+ * xxx_INQ_PKTPTR cbus addresses)
-+ */
-+#define TMU_PHY3_INQ_ADDR (TMU_CSR_BASE_ADDR + 0x0fc)
-+#define TMU_BMU_INQ_ADDR (TMU_CSR_BASE_ADDR + 0x100)
-+#define TMU_TX_CTRL (TMU_CSR_BASE_ADDR + 0x104)
-+
-+#define TMU_BUS_ACCESS_WDATA (TMU_CSR_BASE_ADDR + 0x108)
-+#define TMU_BUS_ACCESS (TMU_CSR_BASE_ADDR + 0x10c)
-+#define TMU_BUS_ACCESS_RDATA (TMU_CSR_BASE_ADDR + 0x110)
-+
-+#define TMU_PE_SYS_CLK_RATIO (TMU_CSR_BASE_ADDR + 0x114)
-+#define TMU_PE_STATUS (TMU_CSR_BASE_ADDR + 0x118)
-+#define TMU_TEQ_MAX_THRESHOLD (TMU_CSR_BASE_ADDR + 0x11c)
-+/* [31:0] PHY4 in queue address (must be initialized with one of the
-+ * xxx_INQ_PKTPTR cbus addresses)
-+ */
-+#define TMU_PHY4_INQ_ADDR (TMU_CSR_BASE_ADDR + 0x134)
-+/* [9:0] Scheduler Enable for each of the scheduler in the TDQ.
-+ * This is a global Enable for all schedulers in PHY1
-+ */
-+#define TMU_TDQ1_SCH_CTRL (TMU_CSR_BASE_ADDR + 0x138)
-+/* [9:0] Scheduler Enable for each of the scheduler in the TDQ.
-+ * This is a global Enable for all schedulers in PHY2
-+ */
-+#define TMU_TDQ2_SCH_CTRL (TMU_CSR_BASE_ADDR + 0x13c)
-+/* [9:0] Scheduler Enable for each of the scheduler in the TDQ.
-+ * This is a global Enable for all schedulers in PHY3
-+ */
-+#define TMU_TDQ3_SCH_CTRL (TMU_CSR_BASE_ADDR + 0x140)
-+#define TMU_BMU_BUF_SIZE (TMU_CSR_BASE_ADDR + 0x144)
-+/* [31:0] PHY5 in queue address (must be initialized with one of the
-+ * xxx_INQ_PKTPTR cbus addresses)
-+ */
-+#define TMU_PHY5_INQ_ADDR (TMU_CSR_BASE_ADDR + 0x148)
-+
-+#define SW_RESET BIT(0) /* Global software reset */
-+#define INQ_RESET BIT(2)
-+#define TEQ_RESET BIT(3)
-+#define TDQ_RESET BIT(4)
-+#define PE_RESET BIT(5)
-+#define MEM_INIT BIT(6)
-+#define MEM_INIT_DONE BIT(7)
-+#define LLM_INIT BIT(8)
-+#define LLM_INIT_DONE BIT(9)
-+#define ECC_MEM_INIT_DONE BIT(10)
-+
-+struct tmu_cfg {
-+ u32 pe_sys_clk_ratio;
-+ unsigned long llm_base_addr;
-+ u32 llm_queue_len;
-+};
-+
-+/* Not HW related for pfe_ctrl / pfe common defines */
-+#define DEFAULT_MAX_QDEPTH 80
-+#define DEFAULT_Q0_QDEPTH 511 /*We keep one large queue for host tx qos */
-+#define DEFAULT_TMU3_QDEPTH 127
-+
-+#endif /* _TMU_CSR_H_ */
---- /dev/null
-+++ b/drivers/staging/fsl_ppfe/include/pfe/cbus/util_csr.h
-@@ -0,0 +1,61 @@
-+/*
-+ * Copyright 2015-2016 Freescale Semiconductor, Inc.
-+ * Copyright 2017 NXP
-+ *
-+ * This program is free software; you can redistribute it and/or modify
-+ * it under the terms of the GNU General Public License as published by
-+ * the Free Software Foundation; either version 2 of the License, or
-+ * (at your option) any later version.
-+ *
-+ * This program is distributed in the hope that it will be useful,
-+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
-+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
-+ * GNU General Public License for more details.
-+ *
-+ * You should have received a copy of the GNU General Public License
-+ * along with this program. If not, see <http://www.gnu.org/licenses/>.
-+ */
-+
-+#ifndef _UTIL_CSR_H_
-+#define _UTIL_CSR_H_
-+
-+#define UTIL_VERSION (UTIL_CSR_BASE_ADDR + 0x000)
-+#define UTIL_TX_CTRL (UTIL_CSR_BASE_ADDR + 0x004)
-+#define UTIL_INQ_PKTPTR (UTIL_CSR_BASE_ADDR + 0x010)
-+
-+#define UTIL_HDR_SIZE (UTIL_CSR_BASE_ADDR + 0x014)
-+
-+#define UTIL_PE0_QB_DM_ADDR0 (UTIL_CSR_BASE_ADDR + 0x020)
-+#define UTIL_PE0_QB_DM_ADDR1 (UTIL_CSR_BASE_ADDR + 0x024)
-+#define UTIL_PE0_RO_DM_ADDR0 (UTIL_CSR_BASE_ADDR + 0x060)
-+#define UTIL_PE0_RO_DM_ADDR1 (UTIL_CSR_BASE_ADDR + 0x064)
-+
-+#define UTIL_MEM_ACCESS_ADDR (UTIL_CSR_BASE_ADDR + 0x100)
-+#define UTIL_MEM_ACCESS_WDATA (UTIL_CSR_BASE_ADDR + 0x104)
-+#define UTIL_MEM_ACCESS_RDATA (UTIL_CSR_BASE_ADDR + 0x108)
-+
-+#define UTIL_TM_INQ_ADDR (UTIL_CSR_BASE_ADDR + 0x114)
-+#define UTIL_PE_STATUS (UTIL_CSR_BASE_ADDR + 0x118)
-+
-+#define UTIL_PE_SYS_CLK_RATIO (UTIL_CSR_BASE_ADDR + 0x200)
-+#define UTIL_AFULL_THRES (UTIL_CSR_BASE_ADDR + 0x204)
-+#define UTIL_GAP_BETWEEN_READS (UTIL_CSR_BASE_ADDR + 0x208)
-+#define UTIL_MAX_BUF_CNT (UTIL_CSR_BASE_ADDR + 0x20c)
-+#define UTIL_TSQ_FIFO_THRES (UTIL_CSR_BASE_ADDR + 0x210)
-+#define UTIL_TSQ_MAX_CNT (UTIL_CSR_BASE_ADDR + 0x214)
-+#define UTIL_IRAM_DATA_0 (UTIL_CSR_BASE_ADDR + 0x218)
-+#define UTIL_IRAM_DATA_1 (UTIL_CSR_BASE_ADDR + 0x21c)
-+#define UTIL_IRAM_DATA_2 (UTIL_CSR_BASE_ADDR + 0x220)
-+#define UTIL_IRAM_DATA_3 (UTIL_CSR_BASE_ADDR + 0x224)
-+
-+#define UTIL_BUS_ACCESS_ADDR (UTIL_CSR_BASE_ADDR + 0x228)
-+#define UTIL_BUS_ACCESS_WDATA (UTIL_CSR_BASE_ADDR + 0x22c)
-+#define UTIL_BUS_ACCESS_RDATA (UTIL_CSR_BASE_ADDR + 0x230)
-+
-+#define UTIL_INQ_AFULL_THRES (UTIL_CSR_BASE_ADDR + 0x234)
-+
-+struct util_cfg {
-+ u32 pe_sys_clk_ratio;
-+};
-+
-+#endif /* _UTIL_CSR_H_ */
---- /dev/null
-+++ b/drivers/staging/fsl_ppfe/include/pfe/pfe.h
-@@ -0,0 +1,372 @@
-+/*
-+ * Copyright 2015-2016 Freescale Semiconductor, Inc.
-+ * Copyright 2017 NXP
-+ *
-+ * This program is free software; you can redistribute it and/or modify
-+ * it under the terms of the GNU General Public License as published by
-+ * the Free Software Foundation; either version 2 of the License, or
-+ * (at your option) any later version.
-+ *
-+ * This program is distributed in the hope that it will be useful,
-+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
-+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
-+ * GNU General Public License for more details.
-+ *
-+ * You should have received a copy of the GNU General Public License
-+ * along with this program. If not, see <http://www.gnu.org/licenses/>.
-+ */
-+
-+#ifndef _PFE_H_
-+#define _PFE_H_
-+
-+#include "cbus.h"
-+
-+#define CLASS_DMEM_BASE_ADDR(i) (0x00000000 | ((i) << 20))
-+/*
-+ * Only valid for mem access register interface
-+ */
-+#define CLASS_IMEM_BASE_ADDR(i) (0x00000000 | ((i) << 20))
-+#define CLASS_DMEM_SIZE 0x00002000
-+#define CLASS_IMEM_SIZE 0x00008000
-+
-+#define TMU_DMEM_BASE_ADDR(i) (0x00000000 + ((i) << 20))
-+/*
-+ * Only valid for mem access register interface
-+ */
-+#define TMU_IMEM_BASE_ADDR(i) (0x00000000 + ((i) << 20))
-+#define TMU_DMEM_SIZE 0x00000800
-+#define TMU_IMEM_SIZE 0x00002000
-+
-+#define UTIL_DMEM_BASE_ADDR 0x00000000
-+#define UTIL_DMEM_SIZE 0x00002000
-+
-+#define PE_LMEM_BASE_ADDR 0xc3010000
-+#define PE_LMEM_SIZE 0x8000
-+#define PE_LMEM_END (PE_LMEM_BASE_ADDR + PE_LMEM_SIZE)
-+
-+#define DMEM_BASE_ADDR 0x00000000
-+#define DMEM_SIZE 0x2000 /* TMU has less... */
-+#define DMEM_END (DMEM_BASE_ADDR + DMEM_SIZE)
-+
-+#define PMEM_BASE_ADDR 0x00010000
-+#define PMEM_SIZE 0x8000 /* TMU has less... */
-+#define PMEM_END (PMEM_BASE_ADDR + PMEM_SIZE)
-+
-+/* These check memory ranges from PE point of view/memory map */
-+#define IS_DMEM(addr, len) \
-+ ({ typeof(addr) addr_ = (addr); \
-+ ((unsigned long)(addr_) >= DMEM_BASE_ADDR) && \
-+ (((unsigned long)(addr_) + (len)) <= DMEM_END); })
-+
-+#define IS_PMEM(addr, len) \
-+ ({ typeof(addr) addr_ = (addr); \
-+ ((unsigned long)(addr_) >= PMEM_BASE_ADDR) && \
-+ (((unsigned long)(addr_) + (len)) <= PMEM_END); })
-+
-+#define IS_PE_LMEM(addr, len) \
-+ ({ typeof(addr) addr_ = (addr); \
-+ ((unsigned long)(addr_) >= \
-+ PE_LMEM_BASE_ADDR) && \
-+ (((unsigned long)(addr_) + \
-+ (len)) <= PE_LMEM_END); })
-+
-+#define IS_PFE_LMEM(addr, len) \
-+ ({ typeof(addr) addr_ = (addr); \
-+ ((unsigned long)(addr_) >= \
-+ CBUS_VIRT_TO_PFE(LMEM_BASE_ADDR)) && \
-+ (((unsigned long)(addr_) + (len)) <= \
-+ CBUS_VIRT_TO_PFE(LMEM_END)); })
-+
-+#define __IS_PHYS_DDR(addr, len) \
-+ ({ typeof(addr) addr_ = (addr); \
-+ ((unsigned long)(addr_) >= \
-+ DDR_PHYS_BASE_ADDR) && \
-+ (((unsigned long)(addr_) + (len)) <= \
-+ DDR_PHYS_END); })
-+
-+#define IS_PHYS_DDR(addr, len) __IS_PHYS_DDR(DDR_PFE_TO_PHYS(addr), len)
-+
-+/*
-+ * If using a run-time virtual address for the cbus base address use this code
-+ */
-+extern void *cbus_base_addr;
-+extern void *ddr_base_addr;
-+extern unsigned long ddr_phys_base_addr;
-+extern unsigned int ddr_size;
-+
-+#define CBUS_BASE_ADDR cbus_base_addr
-+#define DDR_PHYS_BASE_ADDR ddr_phys_base_addr
-+#define DDR_BASE_ADDR ddr_base_addr
-+#define DDR_SIZE ddr_size
-+
-+#define DDR_PHYS_END (DDR_PHYS_BASE_ADDR + DDR_SIZE)
-+
-+#define LS1012A_PFE_RESET_WA /*
-+ * PFE doesn't have global reset and re-init
-+ * should takecare few things to make PFE
-+ * functional after reset
-+ */
-+#define PFE_CBUS_PHYS_BASE_ADDR 0xc0000000 /* CBUS physical base address
-+ * as seen by PE's.
-+ */
-+/* CBUS physical base address as seen by PE's. */
-+#define PFE_CBUS_PHYS_BASE_ADDR_FROM_PFE 0xc0000000
-+
-+#define DDR_PHYS_TO_PFE(p) (((unsigned long int)(p)) & 0x7FFFFFFF)
-+#define DDR_PFE_TO_PHYS(p) (((unsigned long int)(p)) | 0x80000000)
-+#define CBUS_PHYS_TO_PFE(p) (((p) - PFE_CBUS_PHYS_BASE_ADDR) + \
-+ PFE_CBUS_PHYS_BASE_ADDR_FROM_PFE)
-+/* Translates to PFE address map */
-+
-+#define DDR_PHYS_TO_VIRT(p) (((p) - DDR_PHYS_BASE_ADDR) + DDR_BASE_ADDR)
-+#define DDR_VIRT_TO_PHYS(v) (((v) - DDR_BASE_ADDR) + DDR_PHYS_BASE_ADDR)
-+#define DDR_VIRT_TO_PFE(p) (DDR_PHYS_TO_PFE(DDR_VIRT_TO_PHYS(p)))
-+
-+#define CBUS_VIRT_TO_PFE(v) (((v) - CBUS_BASE_ADDR) + \
-+ PFE_CBUS_PHYS_BASE_ADDR)
-+#define CBUS_PFE_TO_VIRT(p) (((unsigned long int)(p) - \
-+ PFE_CBUS_PHYS_BASE_ADDR) + CBUS_BASE_ADDR)
-+
-+/* The below part of the code is used in QOS control driver from host */
-+#define TMU_APB_BASE_ADDR 0xc1000000 /* TMU base address seen by
-+ * pe's
-+ */
-+
-+enum {
-+ CLASS0_ID = 0,
-+ CLASS1_ID,
-+ CLASS2_ID,
-+ CLASS3_ID,
-+ CLASS4_ID,
-+ CLASS5_ID,
-+ TMU0_ID,
-+ TMU1_ID,
-+ TMU2_ID,
-+ TMU3_ID,
-+#if !defined(CONFIG_FSL_PPFE_UTIL_DISABLED)
-+ UTIL_ID,
-+#endif
-+ MAX_PE
-+};
-+
-+#define CLASS_MASK (BIT(CLASS0_ID) | BIT(CLASS1_ID) |\
-+ BIT(CLASS2_ID) | BIT(CLASS3_ID) |\
-+ BIT(CLASS4_ID) | BIT(CLASS5_ID))
-+#define CLASS_MAX_ID CLASS5_ID
-+
-+#define TMU_MASK (BIT(TMU0_ID) | BIT(TMU1_ID) |\
-+ BIT(TMU3_ID))
-+
-+#define TMU_MAX_ID TMU3_ID
-+
-+#if !defined(CONFIG_FSL_PPFE_UTIL_DISABLED)
-+#define UTIL_MASK BIT(UTIL_ID)
-+#endif
-+
-+struct pe_status {
-+ u32 cpu_state;
-+ u32 activity_counter;
-+ u32 rx;
-+ union {
-+ u32 tx;
-+ u32 tmu_qstatus;
-+ };
-+ u32 drop;
-+#if defined(CFG_PE_DEBUG)
-+ u32 debug_indicator;
-+ u32 debug[16];
-+#endif
-+} __aligned(16);
-+
-+struct pe_sync_mailbox {
-+ u32 stop;
-+ u32 stopped;
-+};
-+
-+/* Drop counter definitions */
-+
-+#define CLASS_NUM_DROP_COUNTERS 13
-+#define UTIL_NUM_DROP_COUNTERS 8
-+
-+/* PE information.
-+ * Structure containing PE's specific information. It is used to create
-+ * generic C functions common to all PE's.
-+ * Before using the library functions this structure needs to be initialized
-+ * with the different registers virtual addresses
-+ * (according to the ARM MMU mmaping). The default initialization supports a
-+ * virtual == physical mapping.
-+ */
-+struct pe_info {
-+ u32 dmem_base_addr; /* PE's dmem base address */
-+ u32 pmem_base_addr; /* PE's pmem base address */
-+ u32 pmem_size; /* PE's pmem size */
-+
-+ void *mem_access_wdata; /* PE's _MEM_ACCESS_WDATA register
-+ * address
-+ */
-+ void *mem_access_addr; /* PE's _MEM_ACCESS_ADDR register
-+ * address
-+ */
-+ void *mem_access_rdata; /* PE's _MEM_ACCESS_RDATA register
-+ * address
-+ */
-+};
-+
-+void pe_lmem_read(u32 *dst, u32 len, u32 offset);
-+void pe_lmem_write(u32 *src, u32 len, u32 offset);
-+
-+void pe_dmem_memcpy_to32(int id, u32 dst, const void *src, unsigned int len);
-+void pe_pmem_memcpy_to32(int id, u32 dst, const void *src, unsigned int len);
-+
-+u32 pe_pmem_read(int id, u32 addr, u8 size);
-+
-+void pe_dmem_write(int id, u32 val, u32 addr, u8 size);
-+u32 pe_dmem_read(int id, u32 addr, u8 size);
-+void class_pe_lmem_memcpy_to32(u32 dst, const void *src, unsigned int len);
-+void class_pe_lmem_memset(u32 dst, int val, unsigned int len);
-+void class_bus_write(u32 val, u32 addr, u8 size);
-+u32 class_bus_read(u32 addr, u8 size);
-+
-+#define class_bus_readl(addr) class_bus_read(addr, 4)
-+#define class_bus_readw(addr) class_bus_read(addr, 2)
-+#define class_bus_readb(addr) class_bus_read(addr, 1)
-+
-+#define class_bus_writel(val, addr) class_bus_write(val, addr, 4)
-+#define class_bus_writew(val, addr) class_bus_write(val, addr, 2)
-+#define class_bus_writeb(val, addr) class_bus_write(val, addr, 1)
-+
-+#define pe_dmem_readl(id, addr) pe_dmem_read(id, addr, 4)
-+#define pe_dmem_readw(id, addr) pe_dmem_read(id, addr, 2)
-+#define pe_dmem_readb(id, addr) pe_dmem_read(id, addr, 1)
-+
-+#define pe_dmem_writel(id, val, addr) pe_dmem_write(id, val, addr, 4)
-+#define pe_dmem_writew(id, val, addr) pe_dmem_write(id, val, addr, 2)
-+#define pe_dmem_writeb(id, val, addr) pe_dmem_write(id, val, addr, 1)
-+
-+/*int pe_load_elf_section(int id, const void *data, elf32_shdr *shdr); */
-+int pe_load_elf_section(int id, const void *data, struct elf32_shdr *shdr,
-+ struct device *dev);
-+
-+void pfe_lib_init(void *cbus_base, void *ddr_base, unsigned long ddr_phys_base,
-+ unsigned int ddr_size);
-+void bmu_init(void *base, struct BMU_CFG *cfg);
-+void bmu_reset(void *base);
-+void bmu_enable(void *base);
-+void bmu_disable(void *base);
-+void bmu_set_config(void *base, struct BMU_CFG *cfg);
-+
-+/*
-+ * An enumerated type for loopback values. This can be one of three values, no
-+ * loopback -normal operation, local loopback with internal loopback module of
-+ * MAC or PHY loopback which is through the external PHY.
-+ */
-+#ifndef __MAC_LOOP_ENUM__
-+#define __MAC_LOOP_ENUM__
-+enum mac_loop {LB_NONE, LB_EXT, LB_LOCAL};
-+#endif
-+
-+void gemac_init(void *base, void *config);
-+void gemac_disable_rx_checksum_offload(void *base);
-+void gemac_enable_rx_checksum_offload(void *base);
-+void gemac_set_mdc_div(void *base, int mdc_div);
-+void gemac_set_speed(void *base, enum mac_speed gem_speed);
-+void gemac_set_duplex(void *base, int duplex);
-+void gemac_set_mode(void *base, int mode);
-+void gemac_enable(void *base);
-+void gemac_tx_disable(void *base);
-+void gemac_tx_enable(void *base);
-+void gemac_disable(void *base);
-+void gemac_reset(void *base);
-+void gemac_set_address(void *base, struct spec_addr *addr);
-+struct spec_addr gemac_get_address(void *base);
-+void gemac_set_loop(void *base, enum mac_loop gem_loop);
-+void gemac_set_laddr1(void *base, struct pfe_mac_addr *address);
-+void gemac_set_laddr2(void *base, struct pfe_mac_addr *address);
-+void gemac_set_laddr3(void *base, struct pfe_mac_addr *address);
-+void gemac_set_laddr4(void *base, struct pfe_mac_addr *address);
-+void gemac_set_laddrN(void *base, struct pfe_mac_addr *address,
-+ unsigned int entry_index);
-+void gemac_clear_laddr1(void *base);
-+void gemac_clear_laddr2(void *base);
-+void gemac_clear_laddr3(void *base);
-+void gemac_clear_laddr4(void *base);
-+void gemac_clear_laddrN(void *base, unsigned int entry_index);
-+struct pfe_mac_addr gemac_get_hash(void *base);
-+void gemac_set_hash(void *base, struct pfe_mac_addr *hash);
-+struct pfe_mac_addr gem_get_laddr1(void *base);
-+struct pfe_mac_addr gem_get_laddr2(void *base);
-+struct pfe_mac_addr gem_get_laddr3(void *base);
-+struct pfe_mac_addr gem_get_laddr4(void *base);
-+struct pfe_mac_addr gem_get_laddrN(void *base, unsigned int entry_index);
-+void gemac_set_config(void *base, struct gemac_cfg *cfg);
-+void gemac_allow_broadcast(void *base);
-+void gemac_no_broadcast(void *base);
-+void gemac_enable_1536_rx(void *base);
-+void gemac_disable_1536_rx(void *base);
-+void gemac_enable_rx_jmb(void *base);
-+void gemac_disable_rx_jmb(void *base);
-+void gemac_enable_stacked_vlan(void *base);
-+void gemac_disable_stacked_vlan(void *base);
-+void gemac_enable_pause_rx(void *base);
-+void gemac_disable_pause_rx(void *base);
-+void gemac_enable_copy_all(void *base);
-+void gemac_disable_copy_all(void *base);
-+void gemac_set_bus_width(void *base, int width);
-+void gemac_set_wol(void *base, u32 wol_conf);
-+
-+void gpi_init(void *base, struct gpi_cfg *cfg);
-+void gpi_reset(void *base);
-+void gpi_enable(void *base);
-+void gpi_disable(void *base);
-+void gpi_set_config(void *base, struct gpi_cfg *cfg);
-+
-+void class_init(struct class_cfg *cfg);
-+void class_reset(void);
-+void class_enable(void);
-+void class_disable(void);
-+void class_set_config(struct class_cfg *cfg);
-+
-+void tmu_reset(void);
-+void tmu_init(struct tmu_cfg *cfg);
-+void tmu_enable(u32 pe_mask);
-+void tmu_disable(u32 pe_mask);
-+u32 tmu_qstatus(u32 if_id);
-+u32 tmu_pkts_processed(u32 if_id);
-+
-+void util_init(struct util_cfg *cfg);
-+void util_reset(void);
-+void util_enable(void);
-+void util_disable(void);
-+
-+void hif_init(void);
-+void hif_tx_enable(void);
-+void hif_tx_disable(void);
-+void hif_rx_enable(void);
-+void hif_rx_disable(void);
-+
-+/* Get Chip Revision level
-+ *
-+ */
-+static inline unsigned int CHIP_REVISION(void)
-+{
-+ /*For LS1012A return always 1 */
-+ return 1;
-+}
-+
-+/* Start HIF rx DMA
-+ *
-+ */
-+static inline void hif_rx_dma_start(void)
-+{
-+ writel(HIF_CTRL_DMA_EN | HIF_CTRL_BDP_CH_START_WSTB, HIF_RX_CTRL);
-+}
-+
-+/* Start HIF tx DMA
-+ *
-+ */
-+static inline void hif_tx_dma_start(void)
-+{
-+ writel(HIF_CTRL_DMA_EN | HIF_CTRL_BDP_CH_START_WSTB, HIF_TX_CTRL);
-+}
-+
-+#endif /* _PFE_H_ */
---- /dev/null
-+++ b/drivers/staging/fsl_ppfe/pfe_ctrl.c
-@@ -0,0 +1,238 @@
-+/*
-+ * Copyright 2015-2016 Freescale Semiconductor, Inc.
-+ * Copyright 2017 NXP
-+ *
-+ * This program is free software; you can redistribute it and/or modify
-+ * it under the terms of the GNU General Public License as published by
-+ * the Free Software Foundation; either version 2 of the License, or
-+ * (at your option) any later version.
-+ *
-+ * This program is distributed in the hope that it will be useful,
-+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
-+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
-+ * GNU General Public License for more details.
-+ *
-+ * You should have received a copy of the GNU General Public License
-+ * along with this program. If not, see <http://www.gnu.org/licenses/>.
-+ */
-+
-+#include <linux/kernel.h>
-+#include <linux/sched.h>
-+#include <linux/module.h>
-+#include <linux/list.h>
-+#include <linux/kthread.h>
-+
-+#include "pfe_mod.h"
-+#include "pfe_ctrl.h"
-+
-+#define TIMEOUT_MS 1000
-+
-+int relax(unsigned long end)
-+{
-+ if (time_after(jiffies, end)) {
-+ if (time_after(jiffies, end + (TIMEOUT_MS * HZ) / 1000))
-+ return -1;
-+
-+ if (need_resched())
-+ schedule();
-+ }
-+
-+ return 0;
-+}
-+
-+void pfe_ctrl_suspend(struct pfe_ctrl *ctrl)
-+{
-+ int id;
-+
-+ mutex_lock(&ctrl->mutex);
-+
-+ for (id = CLASS0_ID; id <= CLASS_MAX_ID; id++)
-+ pe_dmem_write(id, cpu_to_be32(0x1), CLASS_DM_RESUME, 4);
-+
-+ for (id = TMU0_ID; id <= TMU_MAX_ID; id++) {
-+ if (id == TMU2_ID)
-+ continue;
-+ pe_dmem_write(id, cpu_to_be32(0x1), TMU_DM_RESUME, 4);
-+ }
-+
-+#if !defined(CONFIG_FSL_PPFE_UTIL_DISABLED)
-+ pe_dmem_write(UTIL_ID, cpu_to_be32(0x1), UTIL_DM_RESUME, 4);
-+#endif
-+ mutex_unlock(&ctrl->mutex);
-+}
-+
-+void pfe_ctrl_resume(struct pfe_ctrl *ctrl)
-+{
-+ int pe_mask = CLASS_MASK | TMU_MASK;
-+
-+#if !defined(CONFIG_FSL_PPFE_UTIL_DISABLED)
-+ pe_mask |= UTIL_MASK;
-+#endif
-+ mutex_lock(&ctrl->mutex);
-+ pe_start(&pfe->ctrl, pe_mask);
-+ mutex_unlock(&ctrl->mutex);
-+}
-+
-+/* PE sync stop.
-+ * Stops packet processing for a list of PE's (specified using a bitmask).
-+ * The caller must hold ctrl->mutex.
-+ *
-+ * @param ctrl Control context
-+ * @param pe_mask Mask of PE id's to stop
-+ *
-+ */
-+int pe_sync_stop(struct pfe_ctrl *ctrl, int pe_mask)
-+{
-+ struct pe_sync_mailbox *mbox;
-+ int pe_stopped = 0;
-+ unsigned long end = jiffies + 2;
-+ int i;
-+
-+ pe_mask &= 0x2FF; /*Exclude Util + TMU2 */
-+
-+ for (i = 0; i < MAX_PE; i++)
-+ if (pe_mask & (1 << i)) {
-+ mbox = (void *)ctrl->sync_mailbox_baseaddr[i];
-+
-+ pe_dmem_write(i, cpu_to_be32(0x1), (unsigned
-+ long)&mbox->stop, 4);
-+ }
-+
-+ while (pe_stopped != pe_mask) {
-+ for (i = 0; i < MAX_PE; i++)
-+ if ((pe_mask & (1 << i)) && !(pe_stopped & (1 << i))) {
-+ mbox = (void *)ctrl->sync_mailbox_baseaddr[i];
-+
-+ if (pe_dmem_read(i, (unsigned
-+ long)&mbox->stopped, 4) &
-+ cpu_to_be32(0x1))
-+ pe_stopped |= (1 << i);
-+ }
-+
-+ if (relax(end) < 0)
-+ goto err;
-+ }
-+
-+ return 0;
-+
-+err:
-+ pr_err("%s: timeout, %x %x\n", __func__, pe_mask, pe_stopped);
-+
-+ for (i = 0; i < MAX_PE; i++)
-+ if (pe_mask & (1 << i)) {
-+ mbox = (void *)ctrl->sync_mailbox_baseaddr[i];
-+
-+ pe_dmem_write(i, cpu_to_be32(0x0), (unsigned
-+ long)&mbox->stop, 4);
-+ }
-+
-+ return -EIO;
-+}
-+
-+/* PE start.
-+ * Starts packet processing for a list of PE's (specified using a bitmask).
-+ * The caller must hold ctrl->mutex.
-+ *
-+ * @param ctrl Control context
-+ * @param pe_mask Mask of PE id's to start
-+ *
-+ */
-+void pe_start(struct pfe_ctrl *ctrl, int pe_mask)
-+{
-+ struct pe_sync_mailbox *mbox;
-+ int i;
-+
-+ for (i = 0; i < MAX_PE; i++)
-+ if (pe_mask & (1 << i)) {
-+ mbox = (void *)ctrl->sync_mailbox_baseaddr[i];
-+
-+ pe_dmem_write(i, cpu_to_be32(0x0), (unsigned
-+ long)&mbox->stop, 4);
-+ }
-+}
-+
-+/* This function will ensure all PEs are put in to idle state */
-+int pe_reset_all(struct pfe_ctrl *ctrl)
-+{
-+ struct pe_sync_mailbox *mbox;
-+ int pe_stopped = 0;
-+ unsigned long end = jiffies + 2;
-+ int i;
-+ int pe_mask = CLASS_MASK | TMU_MASK;
-+
-+#if !defined(CONFIG_FSL_PPFE_UTIL_DISABLED)
-+ pe_mask |= UTIL_MASK;
-+#endif
-+
-+ for (i = 0; i < MAX_PE; i++)
-+ if (pe_mask & (1 << i)) {
-+ mbox = (void *)ctrl->sync_mailbox_baseaddr[i];
-+
-+ pe_dmem_write(i, cpu_to_be32(0x2), (unsigned
-+ long)&mbox->stop, 4);
-+ }
-+
-+ while (pe_stopped != pe_mask) {
-+ for (i = 0; i < MAX_PE; i++)
-+ if ((pe_mask & (1 << i)) && !(pe_stopped & (1 << i))) {
-+ mbox = (void *)ctrl->sync_mailbox_baseaddr[i];
-+
-+ if (pe_dmem_read(i, (unsigned long)
-+ &mbox->stopped, 4) &
-+ cpu_to_be32(0x1))
-+ pe_stopped |= (1 << i);
-+ }
-+
-+ if (relax(end) < 0)
-+ goto err;
-+ }
-+
-+ return 0;
-+
-+err:
-+ pr_err("%s: timeout, %x %x\n", __func__, pe_mask, pe_stopped);
-+ return -EIO;
-+}
-+
-+int pfe_ctrl_init(struct pfe *pfe)
-+{
-+ struct pfe_ctrl *ctrl = &pfe->ctrl;
-+ int id;
-+
-+ pr_info("%s\n", __func__);
-+
-+ mutex_init(&ctrl->mutex);
-+ spin_lock_init(&ctrl->lock);
-+
-+ for (id = CLASS0_ID; id <= CLASS_MAX_ID; id++) {
-+ ctrl->sync_mailbox_baseaddr[id] = CLASS_DM_SYNC_MBOX;
-+ ctrl->msg_mailbox_baseaddr[id] = CLASS_DM_MSG_MBOX;
-+ }
-+
-+ for (id = TMU0_ID; id <= TMU_MAX_ID; id++) {
-+ if (id == TMU2_ID)
-+ continue;
-+ ctrl->sync_mailbox_baseaddr[id] = TMU_DM_SYNC_MBOX;
-+ ctrl->msg_mailbox_baseaddr[id] = TMU_DM_MSG_MBOX;
-+ }
-+
-+#if !defined(CONFIG_FSL_PPFE_UTIL_DISABLED)
-+ ctrl->sync_mailbox_baseaddr[UTIL_ID] = UTIL_DM_SYNC_MBOX;
-+ ctrl->msg_mailbox_baseaddr[UTIL_ID] = UTIL_DM_MSG_MBOX;
-+#endif
-+
-+ ctrl->hash_array_baseaddr = pfe->ddr_baseaddr + ROUTE_TABLE_BASEADDR;
-+ ctrl->hash_array_phys_baseaddr = pfe->ddr_phys_baseaddr +
-+ ROUTE_TABLE_BASEADDR;
-+
-+ ctrl->dev = pfe->dev;
-+
-+ pr_info("%s finished\n", __func__);
-+
-+ return 0;
-+}
-+
-+void pfe_ctrl_exit(struct pfe *pfe)
-+{
-+ pr_info("%s\n", __func__);
-+}
---- /dev/null
-+++ b/drivers/staging/fsl_ppfe/pfe_ctrl.h
-@@ -0,0 +1,112 @@
-+/*
-+ * Copyright 2015-2016 Freescale Semiconductor, Inc.
-+ * Copyright 2017 NXP
-+ *
-+ * This program is free software; you can redistribute it and/or modify
-+ * it under the terms of the GNU General Public License as published by
-+ * the Free Software Foundation; either version 2 of the License, or
-+ * (at your option) any later version.
-+ *
-+ * This program is distributed in the hope that it will be useful,
-+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
-+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
-+ * GNU General Public License for more details.
-+ *
-+ * You should have received a copy of the GNU General Public License
-+ * along with this program. If not, see <http://www.gnu.org/licenses/>.
-+ */
-+
-+#ifndef _PFE_CTRL_H_
-+#define _PFE_CTRL_H_
-+
-+#include <linux/dmapool.h>
-+
-+#include "pfe_mod.h"
-+#include "pfe/pfe.h"
-+
-+#define DMA_BUF_SIZE_128 0x80 /* enough for 1 conntracks */
-+#define DMA_BUF_SIZE_256 0x100
-+/* enough for 2 conntracks, 1 bridge entry or 1 multicast entry */
-+#define DMA_BUF_SIZE_512 0x200
-+/* 512bytes dma allocated buffers used by rtp relay feature */
-+#define DMA_BUF_MIN_ALIGNMENT 8
-+#define DMA_BUF_BOUNDARY (4 * 1024)
-+/* bursts can not cross 4k boundary */
-+
-+#define CMD_TX_ENABLE 0x0501
-+#define CMD_TX_DISABLE 0x0502
-+
-+#define CMD_RX_LRO 0x0011
-+#define CMD_PKTCAP_ENABLE 0x0d01
-+#define CMD_QM_EXPT_RATE 0x020c
-+
-+#define CLASS_DM_SH_STATIC (0x800)
-+#define CLASS_DM_CPU_TICKS (CLASS_DM_SH_STATIC)
-+#define CLASS_DM_SYNC_MBOX (0x808)
-+#define CLASS_DM_MSG_MBOX (0x810)
-+#define CLASS_DM_DROP_CNTR (0x820)
-+#define CLASS_DM_RESUME (0x854)
-+#define CLASS_DM_PESTATUS (0x860)
-+
-+#define TMU_DM_SH_STATIC (0x80)
-+#define TMU_DM_CPU_TICKS (TMU_DM_SH_STATIC)
-+#define TMU_DM_SYNC_MBOX (0x88)
-+#define TMU_DM_MSG_MBOX (0x90)
-+#define TMU_DM_RESUME (0xA0)
-+#define TMU_DM_PESTATUS (0xB0)
-+#define TMU_DM_CONTEXT (0x300)
-+#define TMU_DM_TX_TRANS (0x480)
-+
-+#define UTIL_DM_SH_STATIC (0x0)
-+#define UTIL_DM_CPU_TICKS (UTIL_DM_SH_STATIC)
-+#define UTIL_DM_SYNC_MBOX (0x8)
-+#define UTIL_DM_MSG_MBOX (0x10)
-+#define UTIL_DM_DROP_CNTR (0x20)
-+#define UTIL_DM_RESUME (0x40)
-+#define UTIL_DM_PESTATUS (0x50)
-+
-+struct pfe_ctrl {
-+ struct mutex mutex; /* to serialize pfe control access */
-+ spinlock_t lock;
-+
-+ void *dma_pool;
-+ void *dma_pool_512;
-+ void *dma_pool_128;
-+
-+ struct device *dev;
-+
-+ void *hash_array_baseaddr; /*
-+ * Virtual base address of
-+ * the conntrack hash array
-+ */
-+ unsigned long hash_array_phys_baseaddr; /*
-+ * Physical base address of
-+ * the conntrack hash array
-+ */
-+
-+ int (*event_cb)(u16, u16, u16*);
-+
-+ unsigned long sync_mailbox_baseaddr[MAX_PE]; /*
-+ * Sync mailbox PFE
-+ * internal address,
-+ * initialized
-+ * when parsing elf images
-+ */
-+ unsigned long msg_mailbox_baseaddr[MAX_PE]; /*
-+ * Msg mailbox PFE internal
-+ * address, initialized
-+ * when parsing elf images
-+ */
-+ unsigned int sys_clk; /* AXI clock value, in KHz */
-+};
-+
-+int pfe_ctrl_init(struct pfe *pfe);
-+void pfe_ctrl_exit(struct pfe *pfe);
-+int pe_sync_stop(struct pfe_ctrl *ctrl, int pe_mask);
-+void pe_start(struct pfe_ctrl *ctrl, int pe_mask);
-+int pe_reset_all(struct pfe_ctrl *ctrl);
-+void pfe_ctrl_suspend(struct pfe_ctrl *ctrl);
-+void pfe_ctrl_resume(struct pfe_ctrl *ctrl);
-+int relax(unsigned long end);
-+
-+#endif /* _PFE_CTRL_H_ */
---- /dev/null
-+++ b/drivers/staging/fsl_ppfe/pfe_debugfs.c
-@@ -0,0 +1,111 @@
-+/*
-+ * Copyright 2015-2016 Freescale Semiconductor, Inc.
-+ * Copyright 2017 NXP
-+ *
-+ * This program is free software; you can redistribute it and/or modify
-+ * it under the terms of the GNU General Public License as published by
-+ * the Free Software Foundation; either version 2 of the License, or
-+ * (at your option) any later version.
-+ *
-+ * This program is distributed in the hope that it will be useful,
-+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
-+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
-+ * GNU General Public License for more details.
-+ *
-+ * You should have received a copy of the GNU General Public License
-+ * along with this program. If not, see <http://www.gnu.org/licenses/>.
-+ */
-+
-+#include <linux/module.h>
-+#include <linux/debugfs.h>
-+#include <linux/platform_device.h>
-+
-+#include "pfe_mod.h"
-+
-+static int dmem_show(struct seq_file *s, void *unused)
-+{
-+ u32 dmem_addr, val;
-+ int id = (long int)s->private;
-+ int i;
-+
-+ for (dmem_addr = 0; dmem_addr < CLASS_DMEM_SIZE; dmem_addr += 8 * 4) {
-+ seq_printf(s, "%04x:", dmem_addr);
-+
-+ for (i = 0; i < 8; i++) {
-+ val = pe_dmem_read(id, dmem_addr + i * 4, 4);
-+ seq_printf(s, " %02x %02x %02x %02x", val & 0xff,
-+ (val >> 8) & 0xff, (val >> 16) & 0xff,
-+ (val >> 24) & 0xff);
-+ }
-+
-+ seq_puts(s, "\n");
-+ }
-+
-+ return 0;
-+}
-+
-+static int dmem_open(struct inode *inode, struct file *file)
-+{
-+ return single_open(file, dmem_show, inode->i_private);
-+}
-+
-+static const struct file_operations dmem_fops = {
-+ .open = dmem_open,
-+ .read = seq_read,
-+ .llseek = seq_lseek,
-+ .release = single_release,
-+};
-+
-+int pfe_debugfs_init(struct pfe *pfe)
-+{
-+ struct dentry *d;
-+
-+ pr_info("%s\n", __func__);
-+
-+ pfe->dentry = debugfs_create_dir("pfe", NULL);
-+ if (IS_ERR_OR_NULL(pfe->dentry))
-+ goto err_dir;
-+
-+ d = debugfs_create_file("pe0_dmem", 0444, pfe->dentry, (void *)0,
-+ &dmem_fops);
-+ if (IS_ERR_OR_NULL(d))
-+ goto err_pe;
-+
-+ d = debugfs_create_file("pe1_dmem", 0444, pfe->dentry, (void *)1,
-+ &dmem_fops);
-+ if (IS_ERR_OR_NULL(d))
-+ goto err_pe;
-+
-+ d = debugfs_create_file("pe2_dmem", 0444, pfe->dentry, (void *)2,
-+ &dmem_fops);
-+ if (IS_ERR_OR_NULL(d))
-+ goto err_pe;
-+
-+ d = debugfs_create_file("pe3_dmem", 0444, pfe->dentry, (void *)3,
-+ &dmem_fops);
-+ if (IS_ERR_OR_NULL(d))
-+ goto err_pe;
-+
-+ d = debugfs_create_file("pe4_dmem", 0444, pfe->dentry, (void *)4,
-+ &dmem_fops);
-+ if (IS_ERR_OR_NULL(d))
-+ goto err_pe;
-+
-+ d = debugfs_create_file("pe5_dmem", 0444, pfe->dentry, (void *)5,
-+ &dmem_fops);
-+ if (IS_ERR_OR_NULL(d))
-+ goto err_pe;
-+
-+ return 0;
-+
-+err_pe:
-+ debugfs_remove_recursive(pfe->dentry);
-+
-+err_dir:
-+ return -1;
-+}
-+
-+void pfe_debugfs_exit(struct pfe *pfe)
-+{
-+ debugfs_remove_recursive(pfe->dentry);
-+}
---- /dev/null
-+++ b/drivers/staging/fsl_ppfe/pfe_debugfs.h
-@@ -0,0 +1,25 @@
-+/*
-+ * Copyright 2015-2016 Freescale Semiconductor, Inc.
-+ * Copyright 2017 NXP
-+ *
-+ * This program is free software; you can redistribute it and/or modify
-+ * it under the terms of the GNU General Public License as published by
-+ * the Free Software Foundation; either version 2 of the License, or
-+ * (at your option) any later version.
-+ *
-+ * This program is distributed in the hope that it will be useful,
-+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
-+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
-+ * GNU General Public License for more details.
-+ *
-+ * You should have received a copy of the GNU General Public License
-+ * along with this program. If not, see <http://www.gnu.org/licenses/>.
-+ */
-+
-+#ifndef _PFE_DEBUGFS_H_
-+#define _PFE_DEBUGFS_H_
-+
-+int pfe_debugfs_init(struct pfe *pfe);
-+void pfe_debugfs_exit(struct pfe *pfe);
-+
-+#endif /* _PFE_DEBUGFS_H_ */
---- /dev/null
-+++ b/drivers/staging/fsl_ppfe/pfe_eth.c
-@@ -0,0 +1,2491 @@
-+/*
-+ * Copyright 2015-2016 Freescale Semiconductor, Inc.
-+ * Copyright 2017 NXP
-+ *
-+ * This program is free software; you can redistribute it and/or modify
-+ * it under the terms of the GNU General Public License as published by
-+ * the Free Software Foundation; either version 2 of the License, or
-+ * (at your option) any later version.
-+ *
-+ * This program is distributed in the hope that it will be useful,
-+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
-+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
-+ * GNU General Public License for more details.
-+ *
-+ * You should have received a copy of the GNU General Public License
-+ * along with this program. If not, see <http://www.gnu.org/licenses/>.
-+ */
-+
-+/* @pfe_eth.c.
-+ * Ethernet driver for to handle exception path for PFE.
-+ * - uses HIF functions to send/receive packets.
-+ * - uses ctrl function to start/stop interfaces.
-+ * - uses direct register accesses to control phy operation.
-+ */
-+#include <linux/version.h>
-+#include <linux/kernel.h>
-+#include <linux/interrupt.h>
-+#include <linux/dma-mapping.h>
-+#include <linux/dmapool.h>
-+#include <linux/netdevice.h>
-+#include <linux/etherdevice.h>
-+#include <linux/ethtool.h>
-+#include <linux/mii.h>
-+#include <linux/phy.h>
-+#include <linux/timer.h>
-+#include <linux/hrtimer.h>
-+#include <linux/platform_device.h>
-+
-+#include <net/ip.h>
-+#include <net/sock.h>
-+
-+#include <linux/io.h>
-+#include <asm/irq.h>
-+#include <linux/delay.h>
-+#include <linux/regmap.h>
-+#include <linux/i2c.h>
-+
-+#if defined(CONFIG_NF_CONNTRACK_MARK)
-+#include <net/netfilter/nf_conntrack.h>
-+#endif
-+
-+#include "pfe_mod.h"
-+#include "pfe_eth.h"
-+
-+static void *cbus_emac_base[3];
-+static void *cbus_gpi_base[3];
-+
-+/* Forward Declaration */
-+static void pfe_eth_exit_one(struct pfe_eth_priv_s *priv);
-+static void pfe_eth_flush_tx(struct pfe_eth_priv_s *priv);
-+static void pfe_eth_flush_txQ(struct pfe_eth_priv_s *priv, int tx_q_num, int
-+ from_tx, int n_desc);
-+
-+unsigned int gemac_regs[] = {
-+ 0x0004, /* Interrupt event */
-+ 0x0008, /* Interrupt mask */
-+ 0x0024, /* Ethernet control */
-+ 0x0064, /* MIB Control/Status */
-+ 0x0084, /* Receive control/status */
-+ 0x00C4, /* Transmit control */
-+ 0x00E4, /* Physical address low */
-+ 0x00E8, /* Physical address high */
-+ 0x0144, /* Transmit FIFO Watermark and Store and Forward Control*/
-+ 0x0190, /* Receive FIFO Section Full Threshold */
-+ 0x01A0, /* Transmit FIFO Section Empty Threshold */
-+ 0x01B0, /* Frame Truncation Length */
-+};
-+
-+/********************************************************************/
-+/* SYSFS INTERFACE */
-+/********************************************************************/
-+
-+#ifdef PFE_ETH_NAPI_STATS
-+/*
-+ * pfe_eth_show_napi_stats
-+ */
-+static ssize_t pfe_eth_show_napi_stats(struct device *dev,
-+ struct device_attribute *attr,
-+ char *buf)
-+{
-+ struct pfe_eth_priv_s *priv = netdev_priv(to_net_dev(dev));
-+ ssize_t len = 0;
-+
-+ len += sprintf(buf + len, "sched: %u\n",
-+ priv->napi_counters[NAPI_SCHED_COUNT]);
-+ len += sprintf(buf + len, "poll: %u\n",
-+ priv->napi_counters[NAPI_POLL_COUNT]);
-+ len += sprintf(buf + len, "packet: %u\n",
-+ priv->napi_counters[NAPI_PACKET_COUNT]);
-+ len += sprintf(buf + len, "budget: %u\n",
-+ priv->napi_counters[NAPI_FULL_BUDGET_COUNT]);
-+ len += sprintf(buf + len, "desc: %u\n",
-+ priv->napi_counters[NAPI_DESC_COUNT]);
-+
-+ return len;
-+}
-+
-+/*
-+ * pfe_eth_set_napi_stats
-+ */
-+static ssize_t pfe_eth_set_napi_stats(struct device *dev,
-+ struct device_attribute *attr,
-+ const char *buf, size_t count)
-+{
-+ struct pfe_eth_priv_s *priv = netdev_priv(to_net_dev(dev));
-+
-+ memset(priv->napi_counters, 0, sizeof(priv->napi_counters));
-+
-+ return count;
-+}
-+#endif
-+#ifdef PFE_ETH_TX_STATS
-+/* pfe_eth_show_tx_stats
-+ *
-+ */
-+static ssize_t pfe_eth_show_tx_stats(struct device *dev,
-+ struct device_attribute *attr,
-+ char *buf)
-+{
-+ struct pfe_eth_priv_s *priv = netdev_priv(to_net_dev(dev));
-+ ssize_t len = 0;
-+ int i;
-+
-+ len += sprintf(buf + len, "TX queues stats:\n");
-+
-+ for (i = 0; i < emac_txq_cnt; i++) {
-+ struct netdev_queue *tx_queue = netdev_get_tx_queue(priv->ndev,
-+ i);
-+
-+ len += sprintf(buf + len, "\n");
-+ __netif_tx_lock_bh(tx_queue);
-+
-+ hif_tx_lock(&pfe->hif);
-+ len += sprintf(buf + len,
-+ "Queue %2d : credits = %10d\n"
-+ , i, hif_lib_tx_credit_avail(pfe, priv->id, i));
-+ len += sprintf(buf + len,
-+ " tx packets = %10d\n"
-+ , pfe->tmu_credit.tx_packets[priv->id][i]);
-+ hif_tx_unlock(&pfe->hif);
-+
-+ /* Don't output additionnal stats if queue never used */
-+ if (!pfe->tmu_credit.tx_packets[priv->id][i])
-+ goto skip;
-+
-+ len += sprintf(buf + len,
-+ " clean_fail = %10d\n"
-+ , priv->clean_fail[i]);
-+ len += sprintf(buf + len,
-+ " stop_queue = %10d\n"
-+ , priv->stop_queue_total[i]);
-+ len += sprintf(buf + len,
-+ " stop_queue_hif = %10d\n"
-+ , priv->stop_queue_hif[i]);
-+ len += sprintf(buf + len,
-+ " stop_queue_hif_client = %10d\n"
-+ , priv->stop_queue_hif_client[i]);
-+ len += sprintf(buf + len,
-+ " stop_queue_credit = %10d\n"
-+ , priv->stop_queue_credit[i]);
-+skip:
-+ __netif_tx_unlock_bh(tx_queue);
-+ }
-+ return len;
-+}
-+
-+/* pfe_eth_set_tx_stats
-+ *
-+ */
-+static ssize_t pfe_eth_set_tx_stats(struct device *dev,
-+ struct device_attribute *attr,
-+ const char *buf, size_t count)
-+{
-+ struct pfe_eth_priv_s *priv = netdev_priv(to_net_dev(dev));
-+ int i;
-+
-+ for (i = 0; i < emac_txq_cnt; i++) {
-+ struct netdev_queue *tx_queue = netdev_get_tx_queue(priv->ndev,
-+ i);
-+
-+ __netif_tx_lock_bh(tx_queue);
-+ priv->clean_fail[i] = 0;
-+ priv->stop_queue_total[i] = 0;
-+ priv->stop_queue_hif[i] = 0;
-+ priv->stop_queue_hif_client[i] = 0;
-+ priv->stop_queue_credit[i] = 0;
-+ __netif_tx_unlock_bh(tx_queue);
-+ }
-+
-+ return count;
-+}
-+#endif
-+/* pfe_eth_show_txavail
-+ *
-+ */
-+static ssize_t pfe_eth_show_txavail(struct device *dev,
-+ struct device_attribute *attr,
-+ char *buf)
-+{
-+ struct pfe_eth_priv_s *priv = netdev_priv(to_net_dev(dev));
-+ ssize_t len = 0;
-+ int i;
-+
-+ for (i = 0; i < emac_txq_cnt; i++) {
-+ struct netdev_queue *tx_queue = netdev_get_tx_queue(priv->ndev,
-+ i);
-+
-+ __netif_tx_lock_bh(tx_queue);
-+
-+ len += sprintf(buf + len, "%d",
-+ hif_lib_tx_avail(&priv->client, i));
-+
-+ __netif_tx_unlock_bh(tx_queue);
-+
-+ if (i == (emac_txq_cnt - 1))
-+ len += sprintf(buf + len, "\n");
-+ else
-+ len += sprintf(buf + len, " ");
-+ }
-+
-+ return len;
-+}
-+
-+/* pfe_eth_show_default_priority
-+ *
-+ */
-+static ssize_t pfe_eth_show_default_priority(struct device *dev,
-+ struct device_attribute *attr,
-+ char *buf)
-+{
-+ struct pfe_eth_priv_s *priv = netdev_priv(to_net_dev(dev));
-+ unsigned long flags;
-+ int rc;
-+
-+ spin_lock_irqsave(&priv->lock, flags);
-+ rc = sprintf(buf, "%d\n", priv->default_priority);
-+ spin_unlock_irqrestore(&priv->lock, flags);
-+
-+ return rc;
-+}
-+
-+/* pfe_eth_set_default_priority
-+ *
-+ */
-+
-+static ssize_t pfe_eth_set_default_priority(struct device *dev,
-+ struct device_attribute *attr,
-+ const char *buf, size_t count)
-+{
-+ struct pfe_eth_priv_s *priv = netdev_priv(to_net_dev(dev));
-+ unsigned long flags;
-+
-+ spin_lock_irqsave(&priv->lock, flags);
-+ priv->default_priority = kstrtoul(buf, 0, 0);
-+ spin_unlock_irqrestore(&priv->lock, flags);
-+
-+ return count;
-+}
-+
-+static DEVICE_ATTR(txavail, 0444, pfe_eth_show_txavail, NULL);
-+static DEVICE_ATTR(default_priority, 0644, pfe_eth_show_default_priority,
-+ pfe_eth_set_default_priority);
-+
-+#ifdef PFE_ETH_NAPI_STATS
-+static DEVICE_ATTR(napi_stats, 0644, pfe_eth_show_napi_stats,
-+ pfe_eth_set_napi_stats);
-+#endif
-+
-+#ifdef PFE_ETH_TX_STATS
-+static DEVICE_ATTR(tx_stats, 0644, pfe_eth_show_tx_stats,
-+ pfe_eth_set_tx_stats);
-+#endif
-+
-+/*
-+ * pfe_eth_sysfs_init
-+ *
-+ */
-+static int pfe_eth_sysfs_init(struct net_device *ndev)
-+{
-+ struct pfe_eth_priv_s *priv = netdev_priv(ndev);
-+ int err;
-+
-+ /* Initialize the default values */
-+
-+ /*
-+ * By default, packets without conntrack will use this default low
-+ * priority queue
-+ */
-+ priv->default_priority = 0;
-+
-+ /* Create our sysfs files */
-+ err = device_create_file(&ndev->dev, &dev_attr_default_priority);
-+ if (err) {
-+ netdev_err(ndev,
-+ "failed to create default_priority sysfs files\n");
-+ goto err_priority;
-+ }
-+
-+ err = device_create_file(&ndev->dev, &dev_attr_txavail);
-+ if (err) {
-+ netdev_err(ndev,
-+ "failed to create default_priority sysfs files\n");
-+ goto err_txavail;
-+ }
-+
-+#ifdef PFE_ETH_NAPI_STATS
-+ err = device_create_file(&ndev->dev, &dev_attr_napi_stats);
-+ if (err) {
-+ netdev_err(ndev, "failed to create napi stats sysfs files\n");
-+ goto err_napi;
-+ }
-+#endif
-+
-+#ifdef PFE_ETH_TX_STATS
-+ err = device_create_file(&ndev->dev, &dev_attr_tx_stats);
-+ if (err) {
-+ netdev_err(ndev, "failed to create tx stats sysfs files\n");
-+ goto err_tx;
-+ }
-+#endif
-+
-+ return 0;
-+
-+#ifdef PFE_ETH_TX_STATS
-+err_tx:
-+#endif
-+#ifdef PFE_ETH_NAPI_STATS
-+ device_remove_file(&ndev->dev, &dev_attr_napi_stats);
-+
-+err_napi:
-+#endif
-+ device_remove_file(&ndev->dev, &dev_attr_txavail);
-+
-+err_txavail:
-+ device_remove_file(&ndev->dev, &dev_attr_default_priority);
-+
-+err_priority:
-+ return -1;
-+}
-+
-+/* pfe_eth_sysfs_exit
-+ *
-+ */
-+void pfe_eth_sysfs_exit(struct net_device *ndev)
-+{
-+#ifdef PFE_ETH_TX_STATS
-+ device_remove_file(&ndev->dev, &dev_attr_tx_stats);
-+#endif
-+
-+#ifdef PFE_ETH_NAPI_STATS
-+ device_remove_file(&ndev->dev, &dev_attr_napi_stats);
-+#endif
-+ device_remove_file(&ndev->dev, &dev_attr_txavail);
-+ device_remove_file(&ndev->dev, &dev_attr_default_priority);
-+}
-+
-+/*************************************************************************/
-+/* ETHTOOL INTERCAE */
-+/*************************************************************************/
-+
-+/*MTIP GEMAC */
-+static const struct fec_stat {
-+ char name[ETH_GSTRING_LEN];
-+ u16 offset;
-+} fec_stats[] = {
-+ /* RMON TX */
-+ { "tx_dropped", RMON_T_DROP },
-+ { "tx_packets", RMON_T_PACKETS },
-+ { "tx_broadcast", RMON_T_BC_PKT },
-+ { "tx_multicast", RMON_T_MC_PKT },
-+ { "tx_crc_errors", RMON_T_CRC_ALIGN },
-+ { "tx_undersize", RMON_T_UNDERSIZE },
-+ { "tx_oversize", RMON_T_OVERSIZE },
-+ { "tx_fragment", RMON_T_FRAG },
-+ { "tx_jabber", RMON_T_JAB },
-+ { "tx_collision", RMON_T_COL },
-+ { "tx_64byte", RMON_T_P64 },
-+ { "tx_65to127byte", RMON_T_P65TO127 },
-+ { "tx_128to255byte", RMON_T_P128TO255 },
-+ { "tx_256to511byte", RMON_T_P256TO511 },
-+ { "tx_512to1023byte", RMON_T_P512TO1023 },
-+ { "tx_1024to2047byte", RMON_T_P1024TO2047 },
-+ { "tx_GTE2048byte", RMON_T_P_GTE2048 },
-+ { "tx_octets", RMON_T_OCTETS },
-+
-+ /* IEEE TX */
-+ { "IEEE_tx_drop", IEEE_T_DROP },
-+ { "IEEE_tx_frame_ok", IEEE_T_FRAME_OK },
-+ { "IEEE_tx_1col", IEEE_T_1COL },
-+ { "IEEE_tx_mcol", IEEE_T_MCOL },
-+ { "IEEE_tx_def", IEEE_T_DEF },
-+ { "IEEE_tx_lcol", IEEE_T_LCOL },
-+ { "IEEE_tx_excol", IEEE_T_EXCOL },
-+ { "IEEE_tx_macerr", IEEE_T_MACERR },
-+ { "IEEE_tx_cserr", IEEE_T_CSERR },
-+ { "IEEE_tx_sqe", IEEE_T_SQE },
-+ { "IEEE_tx_fdxfc", IEEE_T_FDXFC },
-+ { "IEEE_tx_octets_ok", IEEE_T_OCTETS_OK },
-+
-+ /* RMON RX */
-+ { "rx_packets", RMON_R_PACKETS },
-+ { "rx_broadcast", RMON_R_BC_PKT },
-+ { "rx_multicast", RMON_R_MC_PKT },
-+ { "rx_crc_errors", RMON_R_CRC_ALIGN },
-+ { "rx_undersize", RMON_R_UNDERSIZE },
-+ { "rx_oversize", RMON_R_OVERSIZE },
-+ { "rx_fragment", RMON_R_FRAG },
-+ { "rx_jabber", RMON_R_JAB },
-+ { "rx_64byte", RMON_R_P64 },
-+ { "rx_65to127byte", RMON_R_P65TO127 },
-+ { "rx_128to255byte", RMON_R_P128TO255 },
-+ { "rx_256to511byte", RMON_R_P256TO511 },
-+ { "rx_512to1023byte", RMON_R_P512TO1023 },
-+ { "rx_1024to2047byte", RMON_R_P1024TO2047 },
-+ { "rx_GTE2048byte", RMON_R_P_GTE2048 },
-+ { "rx_octets", RMON_R_OCTETS },
-+
-+ /* IEEE RX */
-+ { "IEEE_rx_drop", IEEE_R_DROP },
-+ { "IEEE_rx_frame_ok", IEEE_R_FRAME_OK },
-+ { "IEEE_rx_crc", IEEE_R_CRC },
-+ { "IEEE_rx_align", IEEE_R_ALIGN },
-+ { "IEEE_rx_macerr", IEEE_R_MACERR },
-+ { "IEEE_rx_fdxfc", IEEE_R_FDXFC },
-+ { "IEEE_rx_octets_ok", IEEE_R_OCTETS_OK },
-+};
-+
-+static void pfe_eth_fill_stats(struct net_device *ndev, struct ethtool_stats
-+ *stats, u64 *data)
-+{
-+ struct pfe_eth_priv_s *priv = netdev_priv(ndev);
-+ int i;
-+
-+ for (i = 0; i < ARRAY_SIZE(fec_stats); i++)
-+ data[i] = readl(priv->EMAC_baseaddr + fec_stats[i].offset);
-+}
-+
-+static void pfe_eth_gstrings(struct net_device *netdev,
-+ u32 stringset, u8 *data)
-+{
-+ int i;
-+
-+ switch (stringset) {
-+ case ETH_SS_STATS:
-+ for (i = 0; i < ARRAY_SIZE(fec_stats); i++)
-+ memcpy(data + i * ETH_GSTRING_LEN,
-+ fec_stats[i].name, ETH_GSTRING_LEN);
-+ break;
-+ }
-+}
-+
-+static int pfe_eth_stats_count(struct net_device *ndev, int sset)
-+{
-+ switch (sset) {
-+ case ETH_SS_STATS:
-+ return ARRAY_SIZE(fec_stats);
-+ default:
-+ return -EOPNOTSUPP;
-+ }
-+}
-+
-+/*
-+ * pfe_eth_gemac_reglen - Return the length of the register structure.
-+ *
-+ */
-+static int pfe_eth_gemac_reglen(struct net_device *ndev)
-+{
-+ pr_info("%s()\n", __func__);
-+ return (sizeof(gemac_regs) / sizeof(u32));
-+}
-+
-+/*
-+ * pfe_eth_gemac_get_regs - Return the gemac register structure.
-+ *
-+ */
-+static void pfe_eth_gemac_get_regs(struct net_device *ndev, struct ethtool_regs
-+ *regs, void *regbuf)
-+{
-+ int i;
-+
-+ struct pfe_eth_priv_s *priv = netdev_priv(ndev);
-+ u32 *buf = (u32 *)regbuf;
-+
-+ pr_info("%s()\n", __func__);
-+ for (i = 0; i < sizeof(gemac_regs) / sizeof(u32); i++)
-+ buf[i] = readl(priv->EMAC_baseaddr + gemac_regs[i]);
-+}
-+
-+/*
-+ * pfe_eth_set_wol - Set the magic packet option, in WoL register.
-+ *
-+ */
-+static int pfe_eth_set_wol(struct net_device *ndev, struct ethtool_wolinfo *wol)
-+{
-+ struct pfe_eth_priv_s *priv = netdev_priv(ndev);
-+
-+ if (wol->wolopts & ~WAKE_MAGIC)
-+ return -EOPNOTSUPP;
-+
-+ /* for MTIP we store wol->wolopts */
-+ priv->wol = wol->wolopts;
-+
-+ device_set_wakeup_enable(&ndev->dev, wol->wolopts & WAKE_MAGIC);
-+
-+ return 0;
-+}
-+
-+/*
-+ *
-+ * pfe_eth_get_wol - Get the WoL options.
-+ *
-+ */
-+static void pfe_eth_get_wol(struct net_device *ndev, struct ethtool_wolinfo
-+ *wol)
-+{
-+ struct pfe_eth_priv_s *priv = netdev_priv(ndev);
-+
-+ wol->supported = WAKE_MAGIC;
-+ wol->wolopts = 0;
-+
-+ if (priv->wol & WAKE_MAGIC)
-+ wol->wolopts = WAKE_MAGIC;
-+
-+ memset(&wol->sopass, 0, sizeof(wol->sopass));
-+}
-+
-+/*
-+ * pfe_eth_get_drvinfo - Fills in the drvinfo structure with some basic info
-+ *
-+ */
-+static void pfe_eth_get_drvinfo(struct net_device *ndev, struct ethtool_drvinfo
-+ *drvinfo)
-+{
-+ strlcpy(drvinfo->driver, DRV_NAME, sizeof(drvinfo->driver));
-+ strlcpy(drvinfo->version, DRV_VERSION, sizeof(drvinfo->version));
-+ strlcpy(drvinfo->fw_version, "N/A", sizeof(drvinfo->fw_version));
-+ strlcpy(drvinfo->bus_info, "N/A", sizeof(drvinfo->bus_info));
-+}
-+
-+/*
-+ * pfe_eth_set_settings - Used to send commands to PHY.
-+ *
-+ */
-+static int pfe_eth_set_settings(struct net_device *ndev,
-+ const struct ethtool_link_ksettings *cmd)
-+{
-+ struct pfe_eth_priv_s *priv = netdev_priv(ndev);
-+ struct phy_device *phydev = priv->phydev;
-+
-+ if (!phydev)
-+ return -ENODEV;
-+
-+ return phy_ethtool_ksettings_set(phydev, cmd);
-+}
-+
-+/*
-+ * pfe_eth_getsettings - Return the current settings in the ethtool_cmd
-+ * structure.
-+ *
-+ */
-+static int pfe_eth_get_settings(struct net_device *ndev,
-+ struct ethtool_link_ksettings *cmd)
-+{
-+ struct pfe_eth_priv_s *priv = netdev_priv(ndev);
-+ struct phy_device *phydev = priv->phydev;
-+
-+ if (!phydev)
-+ return -ENODEV;
-+
-+ phy_ethtool_ksettings_get(phydev, cmd);
-+
-+ return 0;
-+}
-+
-+/*
-+ * pfe_eth_get_msglevel - Gets the debug message mask.
-+ *
-+ */
-+static uint32_t pfe_eth_get_msglevel(struct net_device *ndev)
-+{
-+ struct pfe_eth_priv_s *priv = netdev_priv(ndev);
-+
-+ return priv->msg_enable;
-+}
-+
-+/*
-+ * pfe_eth_set_msglevel - Sets the debug message mask.
-+ *
-+ */
-+static void pfe_eth_set_msglevel(struct net_device *ndev, uint32_t data)
-+{
-+ struct pfe_eth_priv_s *priv = netdev_priv(ndev);
-+
-+ priv->msg_enable = data;
-+}
-+
-+#define HIF_RX_COAL_MAX_CLKS (~(1 << 31))
-+#define HIF_RX_COAL_CLKS_PER_USEC (pfe->ctrl.sys_clk / 1000)
-+#define HIF_RX_COAL_MAX_USECS (HIF_RX_COAL_MAX_CLKS / \
-+ HIF_RX_COAL_CLKS_PER_USEC)
-+
-+/*
-+ * pfe_eth_set_coalesce - Sets rx interrupt coalescing timer.
-+ *
-+ */
-+static int pfe_eth_set_coalesce(struct net_device *ndev,
-+ struct ethtool_coalesce *ec)
-+{
-+ if (ec->rx_coalesce_usecs > HIF_RX_COAL_MAX_USECS)
-+ return -EINVAL;
-+
-+ if (!ec->rx_coalesce_usecs) {
-+ writel(0, HIF_INT_COAL);
-+ return 0;
-+ }
-+
-+ writel((ec->rx_coalesce_usecs * HIF_RX_COAL_CLKS_PER_USEC) |
-+ HIF_INT_COAL_ENABLE, HIF_INT_COAL);
-+
-+ return 0;
-+}
-+
-+/*
-+ * pfe_eth_get_coalesce - Gets rx interrupt coalescing timer value.
-+ *
-+ */
-+static int pfe_eth_get_coalesce(struct net_device *ndev,
-+ struct ethtool_coalesce *ec)
-+{
-+ int reg_val = readl(HIF_INT_COAL);
-+
-+ if (reg_val & HIF_INT_COAL_ENABLE)
-+ ec->rx_coalesce_usecs = (reg_val & HIF_RX_COAL_MAX_CLKS) /
-+ HIF_RX_COAL_CLKS_PER_USEC;
-+ else
-+ ec->rx_coalesce_usecs = 0;
-+
-+ return 0;
-+}
-+
-+/*
-+ * pfe_eth_set_pauseparam - Sets pause parameters
-+ *
-+ */
-+static int pfe_eth_set_pauseparam(struct net_device *ndev,
-+ struct ethtool_pauseparam *epause)
-+{
-+ struct pfe_eth_priv_s *priv = netdev_priv(ndev);
-+
-+ if (epause->tx_pause != epause->rx_pause) {
-+ netdev_info(ndev,
-+ "hardware only support enable/disable both tx and rx\n");
-+ return -EINVAL;
-+ }
-+
-+ priv->pause_flag = 0;
-+ priv->pause_flag |= epause->rx_pause ? PFE_PAUSE_FLAG_ENABLE : 0;
-+ priv->pause_flag |= epause->autoneg ? PFE_PAUSE_FLAG_AUTONEG : 0;
-+
-+ if (epause->rx_pause || epause->autoneg) {
-+ gemac_enable_pause_rx(priv->EMAC_baseaddr);
-+ writel((readl(priv->GPI_baseaddr + GPI_TX_PAUSE_TIME) |
-+ EGPI_PAUSE_ENABLE),
-+ priv->GPI_baseaddr + GPI_TX_PAUSE_TIME);
-+ if (priv->phydev) {
-+ priv->phydev->supported |= ADVERTISED_Pause |
-+ ADVERTISED_Asym_Pause;
-+ priv->phydev->advertising |= ADVERTISED_Pause |
-+ ADVERTISED_Asym_Pause;
-+ }
-+ } else {
-+ gemac_disable_pause_rx(priv->EMAC_baseaddr);
-+ writel((readl(priv->GPI_baseaddr + GPI_TX_PAUSE_TIME) &
-+ ~EGPI_PAUSE_ENABLE),
-+ priv->GPI_baseaddr + GPI_TX_PAUSE_TIME);
-+ if (priv->phydev) {
-+ priv->phydev->supported &= ~(ADVERTISED_Pause |
-+ ADVERTISED_Asym_Pause);
-+ priv->phydev->advertising &= ~(ADVERTISED_Pause |
-+ ADVERTISED_Asym_Pause);
-+ }
-+ }
-+
-+ return 0;
-+}
-+
-+/*
-+ * pfe_eth_get_pauseparam - Gets pause parameters
-+ *
-+ */
-+static void pfe_eth_get_pauseparam(struct net_device *ndev,
-+ struct ethtool_pauseparam *epause)
-+{
-+ struct pfe_eth_priv_s *priv = netdev_priv(ndev);
-+
-+ epause->autoneg = (priv->pause_flag & PFE_PAUSE_FLAG_AUTONEG) != 0;
-+ epause->tx_pause = (priv->pause_flag & PFE_PAUSE_FLAG_ENABLE) != 0;
-+ epause->rx_pause = epause->tx_pause;
-+}
-+
-+/*
-+ * pfe_eth_get_hash
-+ */
-+#define PFE_HASH_BITS 6 /* #bits in hash */
-+#define CRC32_POLY 0xEDB88320
-+
-+static int pfe_eth_get_hash(u8 *addr)
-+{
-+ unsigned int i, bit, data, crc, hash;
-+
-+ /* calculate crc32 value of mac address */
-+ crc = 0xffffffff;
-+
-+ for (i = 0; i < 6; i++) {
-+ data = addr[i];
-+ for (bit = 0; bit < 8; bit++, data >>= 1) {
-+ crc = (crc >> 1) ^
-+ (((crc ^ data) & 1) ? CRC32_POLY : 0);
-+ }
-+ }
-+
-+ /*
-+ * only upper 6 bits (PFE_HASH_BITS) are used
-+ * which point to specific bit in the hash registers
-+ */
-+ hash = (crc >> (32 - PFE_HASH_BITS)) & 0x3f;
-+
-+ return hash;
-+}
-+
-+const struct ethtool_ops pfe_ethtool_ops = {
-+ .get_drvinfo = pfe_eth_get_drvinfo,
-+ .get_regs_len = pfe_eth_gemac_reglen,
-+ .get_regs = pfe_eth_gemac_get_regs,
-+ .get_link = ethtool_op_get_link,
-+ .get_wol = pfe_eth_get_wol,
-+ .set_wol = pfe_eth_set_wol,
-+ .set_pauseparam = pfe_eth_set_pauseparam,
-+ .get_pauseparam = pfe_eth_get_pauseparam,
-+ .get_strings = pfe_eth_gstrings,
-+ .get_sset_count = pfe_eth_stats_count,
-+ .get_ethtool_stats = pfe_eth_fill_stats,
-+ .get_msglevel = pfe_eth_get_msglevel,
-+ .set_msglevel = pfe_eth_set_msglevel,
-+ .set_coalesce = pfe_eth_set_coalesce,
-+ .get_coalesce = pfe_eth_get_coalesce,
-+ .get_link_ksettings = pfe_eth_get_settings,
-+ .set_link_ksettings = pfe_eth_set_settings,
-+};
-+
-+/* pfe_eth_mdio_reset
-+ */
-+int pfe_eth_mdio_reset(struct mii_bus *bus)
-+{
-+ struct pfe_eth_priv_s *priv = (struct pfe_eth_priv_s *)bus->priv;
-+ u32 phy_speed;
-+
-+ netif_info(priv, hw, priv->ndev, "%s\n", __func__);
-+
-+ mutex_lock(&bus->mdio_lock);
-+
-+ /*
-+ * Set MII speed to 2.5 MHz (= clk_get_rate() / 2 * phy_speed)
-+ *
-+ * The formula for FEC MDC is 'ref_freq / (MII_SPEED x 2)' while
-+ * for ENET-MAC is 'ref_freq / ((MII_SPEED + 1) x 2)'.
-+ */
-+ phy_speed = (DIV_ROUND_UP((pfe->ctrl.sys_clk * 1000), 4000000)
-+ << EMAC_MII_SPEED_SHIFT);
-+ phy_speed |= EMAC_HOLDTIME(0x5);
-+ __raw_writel(phy_speed, priv->PHY_baseaddr + EMAC_MII_CTRL_REG);
-+
-+ mutex_unlock(&bus->mdio_lock);
-+
-+ return 0;
-+}
-+
-+/* pfe_eth_gemac_phy_timeout
-+ *
-+ */
-+static int pfe_eth_gemac_phy_timeout(struct pfe_eth_priv_s *priv, int timeout)
-+{
-+ while (!(__raw_readl(priv->PHY_baseaddr + EMAC_IEVENT_REG) &
-+ EMAC_IEVENT_MII)) {
-+ if (timeout-- <= 0)
-+ return -1;
-+ usleep_range(10, 20);
-+ }
-+ __raw_writel(EMAC_IEVENT_MII, priv->PHY_baseaddr + EMAC_IEVENT_REG);
-+ return 0;
-+}
-+
-+static int pfe_eth_mdio_mux(u8 muxval)
-+{
-+ struct i2c_adapter *a;
-+ struct i2c_msg msg;
-+ unsigned char buf[2];
-+ int ret;
-+
-+ a = i2c_get_adapter(0);
-+ if (!a)
-+ return -ENODEV;
-+
-+ /* set bit 1 (the second bit) of chip at 0x09, register 0x13 */
-+ buf[0] = 0x54; /* reg number */
-+ buf[1] = (muxval << 6) | 0x3; /* data */
-+ msg.addr = 0x66;
-+ msg.buf = buf;
-+ msg.len = 2;
-+ msg.flags = 0;
-+ ret = i2c_transfer(a, &msg, 1);
-+ i2c_put_adapter(a);
-+ if (ret != 1)
-+ return -ENODEV;
-+ return 0;
-+}
-+
-+static int pfe_eth_mdio_write_addr(struct mii_bus *bus, int mii_id,
-+ int dev_addr, int regnum)
-+{
-+ struct pfe_eth_priv_s *priv = (struct pfe_eth_priv_s *)bus->priv;
-+
-+ __raw_writel(EMAC_MII_DATA_PA(mii_id) |
-+ EMAC_MII_DATA_RA(dev_addr) |
-+ EMAC_MII_DATA_TA | EMAC_MII_DATA(regnum),
-+ priv->PHY_baseaddr + EMAC_MII_DATA_REG);
-+
-+ if (pfe_eth_gemac_phy_timeout(priv, EMAC_MDIO_TIMEOUT)) {
-+ netdev_err(priv->ndev, "%s: phy MDIO address write timeout\n",
-+ __func__);
-+ return -1;
-+ }
-+
-+ return 0;
-+}
-+
-+static int pfe_eth_mdio_write(struct mii_bus *bus, int mii_id, int regnum,
-+ u16 value)
-+{
-+ struct pfe_eth_priv_s *priv = (struct pfe_eth_priv_s *)bus->priv;
-+
-+ /*To access external PHYs on QDS board mux needs to be configured*/
-+ if ((mii_id) && (pfe->mdio_muxval[mii_id]))
-+ pfe_eth_mdio_mux(pfe->mdio_muxval[mii_id]);
-+
-+ if (regnum & MII_ADDR_C45) {
-+ pfe_eth_mdio_write_addr(bus, mii_id, (regnum >> 16) & 0x1f,
-+ regnum & 0xffff);
-+ __raw_writel(EMAC_MII_DATA_OP_CL45_WR |
-+ EMAC_MII_DATA_PA(mii_id) |
-+ EMAC_MII_DATA_RA((regnum >> 16) & 0x1f) |
-+ EMAC_MII_DATA_TA | EMAC_MII_DATA(value),
-+ priv->PHY_baseaddr + EMAC_MII_DATA_REG);
-+ } else {
-+ /* start a write op */
-+ __raw_writel(EMAC_MII_DATA_ST | EMAC_MII_DATA_OP_WR |
-+ EMAC_MII_DATA_PA(mii_id) |
-+ EMAC_MII_DATA_RA(regnum) |
-+ EMAC_MII_DATA_TA | EMAC_MII_DATA(value),
-+ priv->PHY_baseaddr + EMAC_MII_DATA_REG);
-+ }
-+
-+ if (pfe_eth_gemac_phy_timeout(priv, EMAC_MDIO_TIMEOUT)) {
-+ netdev_err(priv->ndev, "%s: phy MDIO write timeout\n",
-+ __func__);
-+ return -1;
-+ }
-+ netif_info(priv, hw, priv->ndev, "%s: phy %x reg %x val %x\n", __func__,
-+ mii_id, regnum, value);
-+
-+ return 0;
-+}
-+
-+static int pfe_eth_mdio_read(struct mii_bus *bus, int mii_id, int regnum)
-+{
-+ struct pfe_eth_priv_s *priv = (struct pfe_eth_priv_s *)bus->priv;
-+ u16 value = 0;
-+
-+ /*To access external PHYs on QDS board mux needs to be configured*/
-+ if ((mii_id) && (pfe->mdio_muxval[mii_id]))
-+ pfe_eth_mdio_mux(pfe->mdio_muxval[mii_id]);
-+
-+ if (regnum & MII_ADDR_C45) {
-+ pfe_eth_mdio_write_addr(bus, mii_id, (regnum >> 16) & 0x1f,
-+ regnum & 0xffff);
-+ __raw_writel(EMAC_MII_DATA_OP_CL45_RD |
-+ EMAC_MII_DATA_PA(mii_id) |
-+ EMAC_MII_DATA_RA((regnum >> 16) & 0x1f) |
-+ EMAC_MII_DATA_TA,
-+ priv->PHY_baseaddr + EMAC_MII_DATA_REG);
-+ } else {
-+ /* start a read op */
-+ __raw_writel(EMAC_MII_DATA_ST | EMAC_MII_DATA_OP_RD |
-+ EMAC_MII_DATA_PA(mii_id) |
-+ EMAC_MII_DATA_RA(regnum) |
-+ EMAC_MII_DATA_TA, priv->PHY_baseaddr +
-+ EMAC_MII_DATA_REG);
-+ }
-+
-+ if (pfe_eth_gemac_phy_timeout(priv, EMAC_MDIO_TIMEOUT)) {
-+ netdev_err(priv->ndev, "%s: phy MDIO read timeout\n", __func__);
-+ return -1;
-+ }
-+
-+ value = EMAC_MII_DATA(__raw_readl(priv->PHY_baseaddr +
-+ EMAC_MII_DATA_REG));
-+ netif_info(priv, hw, priv->ndev, "%s: phy %x reg %x val %x\n", __func__,
-+ mii_id, regnum, value);
-+ return value;
-+}
-+
-+static int pfe_eth_mdio_init(struct pfe_eth_priv_s *priv,
-+ struct ls1012a_mdio_platform_data *minfo)
-+{
-+ struct mii_bus *bus;
-+ int rc, ii;
-+ struct phy_device *phydev;
-+
-+ netif_info(priv, drv, priv->ndev, "%s\n", __func__);
-+ pr_info("%s\n", __func__);
-+
-+ bus = mdiobus_alloc();
-+ if (!bus) {
-+ netdev_err(priv->ndev, "mdiobus_alloc() failed\n");
-+ rc = -ENOMEM;
-+ goto err0;
-+ }
-+
-+ bus->name = "ls1012a MDIO Bus";
-+ bus->read = &pfe_eth_mdio_read;
-+ bus->write = &pfe_eth_mdio_write;
-+ bus->reset = &pfe_eth_mdio_reset;
-+ snprintf(bus->id, MII_BUS_ID_SIZE, "ls1012a-%x", priv->id);
-+ bus->priv = priv;
-+
-+ bus->phy_mask = minfo->phy_mask;
-+ priv->mdc_div = minfo->mdc_div;
-+
-+ if (!priv->mdc_div)
-+ priv->mdc_div = 64;
-+
-+ bus->irq[0] = minfo->irq[0];
-+
-+ bus->parent = priv->pfe->dev;
-+
-+ netif_info(priv, drv, priv->ndev, "%s: mdc_div: %d, phy_mask: %x\n",
-+ __func__, priv->mdc_div, bus->phy_mask);
-+ rc = mdiobus_register(bus);
-+ if (rc) {
-+ netdev_err(priv->ndev, "mdiobus_register(%s) failed\n",
-+ bus->name);
-+ goto err1;
-+ }
-+
-+ priv->mii_bus = bus;
-+
-+ /* For clause 45 we need to call get_phy_device() with it's
-+ * 3rd argument as true and then register the phy device
-+ * via phy_device_register()
-+ */
-+
-+ if (priv->einfo->mii_config == PHY_INTERFACE_MODE_2500SGMII) {
-+ for (ii = 0; ii < NUM_GEMAC_SUPPORT; ii++) {
-+ phydev = get_phy_device(priv->mii_bus,
-+ priv->einfo->phy_id + ii, true);
-+ if (!phydev || IS_ERR(phydev)) {
-+ rc = -EIO;
-+ netdev_err(priv->ndev, "fail to get device\n");
-+ goto err1;
-+ }
-+ rc = phy_device_register(phydev);
-+ if (rc) {
-+ phy_device_free(phydev);
-+ netdev_err(priv->ndev,
-+ "phy_device_register() failed\n");
-+ goto err1;
-+ }
-+ }
-+ }
-+
-+ pfe_eth_mdio_reset(bus);
-+
-+ return 0;
-+
-+err1:
-+ mdiobus_free(bus);
-+err0:
-+ return rc;
-+}
-+
-+/* pfe_eth_mdio_exit
-+ */
-+static void pfe_eth_mdio_exit(struct mii_bus *bus)
-+{
-+ if (!bus)
-+ return;
-+
-+ netif_info((struct pfe_eth_priv_s *)bus->priv, drv, ((struct
-+ pfe_eth_priv_s *)(bus->priv))->ndev, "%s\n", __func__);
-+
-+ mdiobus_unregister(bus);
-+ mdiobus_free(bus);
-+}
-+
-+/* pfe_get_phydev_speed
-+ */
-+static int pfe_get_phydev_speed(struct phy_device *phydev)
-+{
-+ switch (phydev->speed) {
-+ case 10:
-+ return SPEED_10M;
-+ case 100:
-+ return SPEED_100M;
-+ case 1000:
-+ default:
-+ return SPEED_1000M;
-+ }
-+}
-+
-+/* pfe_set_rgmii_speed
-+ */
-+#define RGMIIPCR 0x434
-+/* RGMIIPCR bit definitions*/
-+#define SCFG_RGMIIPCR_EN_AUTO (0x00000008)
-+#define SCFG_RGMIIPCR_SETSP_1000M (0x00000004)
-+#define SCFG_RGMIIPCR_SETSP_100M (0x00000000)
-+#define SCFG_RGMIIPCR_SETSP_10M (0x00000002)
-+#define SCFG_RGMIIPCR_SETFD (0x00000001)
-+
-+static void pfe_set_rgmii_speed(struct phy_device *phydev)
-+{
-+ u32 rgmii_pcr;
-+
-+ regmap_read(pfe->scfg, RGMIIPCR, &rgmii_pcr);
-+ rgmii_pcr &= ~(SCFG_RGMIIPCR_SETSP_1000M | SCFG_RGMIIPCR_SETSP_10M);
-+
-+ switch (phydev->speed) {
-+ case 10:
-+ rgmii_pcr |= SCFG_RGMIIPCR_SETSP_10M;
-+ break;
-+ case 1000:
-+ rgmii_pcr |= SCFG_RGMIIPCR_SETSP_1000M;
-+ break;
-+ case 100:
-+ default:
-+ /* Default is 100M */
-+ break;
-+ }
-+ regmap_write(pfe->scfg, RGMIIPCR, rgmii_pcr);
-+}
-+
-+/* pfe_get_phydev_duplex
-+ */
-+static int pfe_get_phydev_duplex(struct phy_device *phydev)
-+{
-+ /*return (phydev->duplex == DUPLEX_HALF) ? DUP_HALF:DUP_FULL ; */
-+ return DUPLEX_FULL;
-+}
-+
-+/* pfe_eth_adjust_link
-+ */
-+static void pfe_eth_adjust_link(struct net_device *ndev)
-+{
-+ struct pfe_eth_priv_s *priv = netdev_priv(ndev);
-+ unsigned long flags;
-+ struct phy_device *phydev = priv->phydev;
-+ int new_state = 0;
-+
-+ netif_info(priv, drv, ndev, "%s\n", __func__);
-+
-+ spin_lock_irqsave(&priv->lock, flags);
-+
-+ if (phydev->link) {
-+ /*
-+ * Now we make sure that we can be in full duplex mode.
-+ * If not, we operate in half-duplex mode.
-+ */
-+ if (phydev->duplex != priv->oldduplex) {
-+ new_state = 1;
-+ gemac_set_duplex(priv->EMAC_baseaddr,
-+ pfe_get_phydev_duplex(phydev));
-+ priv->oldduplex = phydev->duplex;
-+ }
-+
-+ if (phydev->speed != priv->oldspeed) {
-+ new_state = 1;
-+ gemac_set_speed(priv->EMAC_baseaddr,
-+ pfe_get_phydev_speed(phydev));
-+ if (priv->einfo->mii_config == PHY_INTERFACE_MODE_RGMII_TXID)
-+ pfe_set_rgmii_speed(phydev);
-+ priv->oldspeed = phydev->speed;
-+ }
-+
-+ if (!priv->oldlink) {
-+ new_state = 1;
-+ priv->oldlink = 1;
-+ }
-+
-+ } else if (priv->oldlink) {
-+ new_state = 1;
-+ priv->oldlink = 0;
-+ priv->oldspeed = 0;
-+ priv->oldduplex = -1;
-+ }
-+
-+ if (new_state && netif_msg_link(priv))
-+ phy_print_status(phydev);
-+
-+ spin_unlock_irqrestore(&priv->lock, flags);
-+}
-+
-+/* pfe_phy_exit
-+ */
-+static void pfe_phy_exit(struct net_device *ndev)
-+{
-+ struct pfe_eth_priv_s *priv = netdev_priv(ndev);
-+
-+ netif_info(priv, drv, ndev, "%s\n", __func__);
-+
-+ phy_disconnect(priv->phydev);
-+ priv->phydev = NULL;
-+}
-+
-+/* pfe_eth_stop
-+ */
-+static void pfe_eth_stop(struct net_device *ndev, int wake)
-+{
-+ struct pfe_eth_priv_s *priv = netdev_priv(ndev);
-+
-+ netif_info(priv, drv, ndev, "%s\n", __func__);
-+
-+ if (wake) {
-+ gemac_tx_disable(priv->EMAC_baseaddr);
-+ } else {
-+ gemac_disable(priv->EMAC_baseaddr);
-+ gpi_disable(priv->GPI_baseaddr);
-+
-+ if (priv->phydev)
-+ phy_stop(priv->phydev);
-+ }
-+}
-+
-+/* pfe_eth_start
-+ */
-+static int pfe_eth_start(struct pfe_eth_priv_s *priv)
-+{
-+ netif_info(priv, drv, priv->ndev, "%s\n", __func__);
-+
-+ if (priv->phydev)
-+ phy_start(priv->phydev);
-+
-+ gpi_enable(priv->GPI_baseaddr);
-+ gemac_enable(priv->EMAC_baseaddr);
-+
-+ return 0;
-+}
-+
-+/*
-+ * Configure on chip serdes through mdio
-+ */
-+static void ls1012a_configure_serdes(struct net_device *ndev)
-+{
-+ struct pfe_eth_priv_s *priv = pfe->eth.eth_priv[0];
-+ int sgmii_2500 = 0;
-+ struct mii_bus *bus = priv->mii_bus;
-+ u16 value = 0;
-+
-+ if (priv->einfo->mii_config == PHY_INTERFACE_MODE_2500SGMII)
-+ sgmii_2500 = 1;
-+
-+ netif_info(priv, drv, ndev, "%s\n", __func__);
-+ /* PCS configuration done with corresponding GEMAC */
-+
-+ pfe_eth_mdio_read(bus, 0, 0);
-+ pfe_eth_mdio_read(bus, 0, 1);
-+
-+ /*These settings taken from validtion team */
-+ pfe_eth_mdio_write(bus, 0, 0x0, 0x8000);
-+ if (sgmii_2500) {
-+ pfe_eth_mdio_write(bus, 0, 0x14, 0x9);
-+ pfe_eth_mdio_write(bus, 0, 0x4, 0x4001);
-+ pfe_eth_mdio_write(bus, 0, 0x12, 0xa120);
-+ pfe_eth_mdio_write(bus, 0, 0x13, 0x7);
-+ /* Autonegotiation need to be disabled for 2.5G SGMII mode*/
-+ value = 0x0140;
-+ pfe_eth_mdio_write(bus, 0, 0x0, value);
-+ } else {
-+ pfe_eth_mdio_write(bus, 0, 0x14, 0xb);
-+ pfe_eth_mdio_write(bus, 0, 0x4, 0x1a1);
-+ pfe_eth_mdio_write(bus, 0, 0x12, 0x400);
-+ pfe_eth_mdio_write(bus, 0, 0x13, 0x0);
-+ pfe_eth_mdio_write(bus, 0, 0x0, 0x1140);
-+ }
-+}
-+
-+/*
-+ * pfe_phy_init
-+ *
-+ */
-+static int pfe_phy_init(struct net_device *ndev)
-+{
-+ struct pfe_eth_priv_s *priv = netdev_priv(ndev);
-+ struct phy_device *phydev;
-+ char phy_id[MII_BUS_ID_SIZE + 3];
-+ char bus_id[MII_BUS_ID_SIZE];
-+ phy_interface_t interface;
-+
-+ priv->oldlink = 0;
-+ priv->oldspeed = 0;
-+ priv->oldduplex = -1;
-+
-+ snprintf(bus_id, MII_BUS_ID_SIZE, "ls1012a-%d", 0);
-+ snprintf(phy_id, MII_BUS_ID_SIZE + 3, PHY_ID_FMT, bus_id,
-+ priv->einfo->phy_id);
-+
-+ netif_info(priv, drv, ndev, "%s: %s\n", __func__, phy_id);
-+ interface = priv->einfo->mii_config;
-+ if ((interface == PHY_INTERFACE_MODE_SGMII) ||
-+ (interface == PHY_INTERFACE_MODE_2500SGMII)) {
-+ /*Configure SGMII PCS */
-+ if (pfe->scfg) {
-+ /*Config MDIO from serdes */
-+ regmap_write(pfe->scfg, 0x484, 0x00000000);
-+ }
-+ ls1012a_configure_serdes(ndev);
-+ }
-+
-+ if (pfe->scfg) {
-+ /*Config MDIO from PAD */
-+ regmap_write(pfe->scfg, 0x484, 0x80000000);
-+ }
-+
-+ priv->oldlink = 0;
-+ priv->oldspeed = 0;
-+ priv->oldduplex = -1;
-+ pr_info("%s interface %x\n", __func__, interface);
-+ phydev = phy_connect(ndev, phy_id, &pfe_eth_adjust_link, interface);
-+
-+ if (IS_ERR(phydev)) {
-+ netdev_err(ndev, "phy_connect() failed\n");
-+ return PTR_ERR(phydev);
-+ }
-+
-+ priv->phydev = phydev;
-+ phydev->irq = PHY_POLL;
-+
-+ return 0;
-+}
-+
-+/* pfe_gemac_init
-+ */
-+static int pfe_gemac_init(struct pfe_eth_priv_s *priv)
-+{
-+ struct gemac_cfg cfg;
-+
-+ netif_info(priv, ifup, priv->ndev, "%s\n", __func__);
-+
-+ cfg.speed = SPEED_1000M;
-+ cfg.duplex = DUPLEX_FULL;
-+
-+ gemac_set_config(priv->EMAC_baseaddr, &cfg);
-+ gemac_allow_broadcast(priv->EMAC_baseaddr);
-+ gemac_enable_1536_rx(priv->EMAC_baseaddr);
-+ gemac_enable_rx_jmb(priv->EMAC_baseaddr);
-+ gemac_enable_stacked_vlan(priv->EMAC_baseaddr);
-+ gemac_enable_pause_rx(priv->EMAC_baseaddr);
-+ gemac_set_bus_width(priv->EMAC_baseaddr, 64);
-+
-+ /*GEM will perform checksum verifications*/
-+ if (priv->ndev->features & NETIF_F_RXCSUM)
-+ gemac_enable_rx_checksum_offload(priv->EMAC_baseaddr);
-+ else
-+ gemac_disable_rx_checksum_offload(priv->EMAC_baseaddr);
-+
-+ return 0;
-+}
-+
-+/* pfe_eth_event_handler
-+ */
-+static int pfe_eth_event_handler(void *data, int event, int qno)
-+{
-+ struct pfe_eth_priv_s *priv = data;
-+
-+ switch (event) {
-+ case EVENT_RX_PKT_IND:
-+
-+ if (qno == 0) {
-+ if (napi_schedule_prep(&priv->high_napi)) {
-+ netif_info(priv, intr, priv->ndev,
-+ "%s: schedule high prio poll\n"
-+ , __func__);
-+
-+#ifdef PFE_ETH_NAPI_STATS
-+ priv->napi_counters[NAPI_SCHED_COUNT]++;
-+#endif
-+
-+ __napi_schedule(&priv->high_napi);
-+ }
-+ } else if (qno == 1) {
-+ if (napi_schedule_prep(&priv->low_napi)) {
-+ netif_info(priv, intr, priv->ndev,
-+ "%s: schedule low prio poll\n"
-+ , __func__);
-+
-+#ifdef PFE_ETH_NAPI_STATS
-+ priv->napi_counters[NAPI_SCHED_COUNT]++;
-+#endif
-+ __napi_schedule(&priv->low_napi);
-+ }
-+ } else if (qno == 2) {
-+ if (napi_schedule_prep(&priv->lro_napi)) {
-+ netif_info(priv, intr, priv->ndev,
-+ "%s: schedule lro prio poll\n"
-+ , __func__);
-+
-+#ifdef PFE_ETH_NAPI_STATS
-+ priv->napi_counters[NAPI_SCHED_COUNT]++;
-+#endif
-+ __napi_schedule(&priv->lro_napi);
-+ }
-+ }
-+
-+ break;
-+
-+ case EVENT_TXDONE_IND:
-+ pfe_eth_flush_tx(priv);
-+ hif_lib_event_handler_start(&priv->client, EVENT_TXDONE_IND, 0);
-+ break;
-+ case EVENT_HIGH_RX_WM:
-+ default:
-+ break;
-+ }
-+
-+ return 0;
-+}
-+
-+/* pfe_eth_open
-+ */
-+static int pfe_eth_open(struct net_device *ndev)
-+{
-+ struct pfe_eth_priv_s *priv = netdev_priv(ndev);
-+ struct hif_client_s *client;
-+ int rc;
-+
-+ netif_info(priv, ifup, ndev, "%s\n", __func__);
-+
-+ /* Register client driver with HIF */
-+ client = &priv->client;
-+ memset(client, 0, sizeof(*client));
-+ client->id = PFE_CL_GEM0 + priv->id;
-+ client->tx_qn = emac_txq_cnt;
-+ client->rx_qn = EMAC_RXQ_CNT;
-+ client->priv = priv;
-+ client->pfe = priv->pfe;
-+ client->event_handler = pfe_eth_event_handler;
-+
-+ client->tx_qsize = EMAC_TXQ_DEPTH;
-+ client->rx_qsize = EMAC_RXQ_DEPTH;
-+
-+ rc = hif_lib_client_register(client);
-+ if (rc) {
-+ netdev_err(ndev, "%s: hif_lib_client_register(%d) failed\n",
-+ __func__, client->id);
-+ goto err0;
-+ }
-+
-+ netif_info(priv, drv, ndev, "%s: registered client: %p\n", __func__,
-+ client);
-+
-+ pfe_gemac_init(priv);
-+
-+ if (!is_valid_ether_addr(ndev->dev_addr)) {
-+ netdev_err(ndev, "%s: invalid MAC address\n", __func__);
-+ rc = -EADDRNOTAVAIL;
-+ goto err1;
-+ }
-+
-+ gemac_set_laddrN(priv->EMAC_baseaddr,
-+ (struct pfe_mac_addr *)ndev->dev_addr, 1);
-+
-+ napi_enable(&priv->high_napi);
-+ napi_enable(&priv->low_napi);
-+ napi_enable(&priv->lro_napi);
-+
-+ rc = pfe_eth_start(priv);
-+
-+ netif_tx_wake_all_queues(ndev);
-+
-+ return rc;
-+
-+err1:
-+ hif_lib_client_unregister(&priv->client);
-+
-+err0:
-+ return rc;
-+}
-+
-+/*
-+ * pfe_eth_shutdown
-+ */
-+int pfe_eth_shutdown(struct net_device *ndev, int wake)
-+{
-+ struct pfe_eth_priv_s *priv = netdev_priv(ndev);
-+ int i, qstatus;
-+ unsigned long next_poll = jiffies + 1, end = jiffies +
-+ (TX_POLL_TIMEOUT_MS * HZ) / 1000;
-+ int tx_pkts, prv_tx_pkts;
-+
-+ netif_info(priv, ifdown, ndev, "%s\n", __func__);
-+
-+ for (i = 0; i < emac_txq_cnt; i++)
-+ hrtimer_cancel(&priv->fast_tx_timeout[i].timer);
-+
-+ netif_tx_stop_all_queues(ndev);
-+
-+ do {
-+ tx_pkts = 0;
-+ pfe_eth_flush_tx(priv);
-+
-+ for (i = 0; i < emac_txq_cnt; i++)
-+ tx_pkts += hif_lib_tx_pending(&priv->client, i);
-+
-+ if (tx_pkts) {
-+ /*Don't wait forever, break if we cross max timeout */
-+ if (time_after(jiffies, end)) {
-+ pr_err(
-+ "(%s)Tx is not complete after %dmsec\n",
-+ ndev->name, TX_POLL_TIMEOUT_MS);
-+ break;
-+ }
-+
-+ pr_info("%s : (%s) Waiting for tx packets to free. Pending tx pkts = %d.\n"
-+ , __func__, ndev->name, tx_pkts);
-+ if (need_resched())
-+ schedule();
-+ }
-+
-+ } while (tx_pkts);
-+
-+ end = jiffies + (TX_POLL_TIMEOUT_MS * HZ) / 1000;
-+
-+ prv_tx_pkts = tmu_pkts_processed(priv->id);
-+ /*
-+ * Wait till TMU transmits all pending packets
-+ * poll tmu_qstatus and pkts processed by TMU for every 10ms
-+ * Consider TMU is busy, If we see TMU qeueu pending or any packets
-+ * processed by TMU
-+ */
-+ while (1) {
-+ if (time_after(jiffies, next_poll)) {
-+ tx_pkts = tmu_pkts_processed(priv->id);
-+ qstatus = tmu_qstatus(priv->id) & 0x7ffff;
-+
-+ if (!qstatus && (tx_pkts == prv_tx_pkts))
-+ break;
-+ /* Don't wait forever, break if we cross max
-+ * timeout(TX_POLL_TIMEOUT_MS)
-+ */
-+ if (time_after(jiffies, end)) {
-+ pr_err("TMU%d is busy after %dmsec\n",
-+ priv->id, TX_POLL_TIMEOUT_MS);
-+ break;
-+ }
-+ prv_tx_pkts = tx_pkts;
-+ next_poll++;
-+ }
-+ if (need_resched())
-+ schedule();
-+ }
-+ /* Wait for some more time to complete transmitting packet if any */
-+ next_poll = jiffies + 1;
-+ while (1) {
-+ if (time_after(jiffies, next_poll))
-+ break;
-+ if (need_resched())
-+ schedule();
-+ }
-+
-+ pfe_eth_stop(ndev, wake);
-+
-+ napi_disable(&priv->lro_napi);
-+ napi_disable(&priv->low_napi);
-+ napi_disable(&priv->high_napi);
-+
-+ hif_lib_client_unregister(&priv->client);
-+
-+ return 0;
-+}
-+
-+/* pfe_eth_close
-+ *
-+ */
-+static int pfe_eth_close(struct net_device *ndev)
-+{
-+ pfe_eth_shutdown(ndev, 0);
-+
-+ return 0;
-+}
-+
-+/* pfe_eth_suspend
-+ *
-+ * return value : 1 if netdevice is configured to wakeup system
-+ * 0 otherwise
-+ */
-+int pfe_eth_suspend(struct net_device *ndev)
-+{
-+ struct pfe_eth_priv_s *priv = netdev_priv(ndev);
-+ int retval = 0;
-+
-+ if (priv->wol) {
-+ gemac_set_wol(priv->EMAC_baseaddr, priv->wol);
-+ retval = 1;
-+ }
-+ pfe_eth_shutdown(ndev, priv->wol);
-+
-+ return retval;
-+}
-+
-+/* pfe_eth_resume
-+ *
-+ */
-+int pfe_eth_resume(struct net_device *ndev)
-+{
-+ struct pfe_eth_priv_s *priv = netdev_priv(ndev);
-+
-+ if (priv->wol)
-+ gemac_set_wol(priv->EMAC_baseaddr, 0);
-+ gemac_tx_enable(priv->EMAC_baseaddr);
-+
-+ return pfe_eth_open(ndev);
-+}
-+
-+/* pfe_eth_get_queuenum
-+ */
-+static int pfe_eth_get_queuenum(struct pfe_eth_priv_s *priv, struct sk_buff
-+ *skb)
-+{
-+ int queuenum = 0;
-+ unsigned long flags;
-+
-+ /* Get the Fast Path queue number */
-+ /*
-+ * Use conntrack mark (if conntrack exists), then packet mark (if any),
-+ * then fallback to default
-+ */
-+#if defined(CONFIG_IP_NF_CONNTRACK_MARK) || defined(CONFIG_NF_CONNTRACK_MARK)
-+ if (skb->nfct) {
-+ enum ip_conntrack_info cinfo;
-+ struct nf_conn *ct;
-+
-+ ct = nf_ct_get(skb, &cinfo);
-+
-+ if (ct) {
-+ u32 connmark;
-+
-+ connmark = ct->mark;
-+
-+ if ((connmark & 0x80000000) && priv->id != 0)
-+ connmark >>= 16;
-+
-+ queuenum = connmark & EMAC_QUEUENUM_MASK;
-+ }
-+ } else {/* continued after #endif ... */
-+#endif
-+ if (skb->mark) {
-+ queuenum = skb->mark & EMAC_QUEUENUM_MASK;
-+ } else {
-+ spin_lock_irqsave(&priv->lock, flags);
-+ queuenum = priv->default_priority & EMAC_QUEUENUM_MASK;
-+ spin_unlock_irqrestore(&priv->lock, flags);
-+ }
-+#if defined(CONFIG_IP_NF_CONNTRACK_MARK) || defined(CONFIG_NF_CONNTRACK_MARK)
-+ }
-+#endif
-+ return queuenum;
-+}
-+
-+/* pfe_eth_might_stop_tx
-+ *
-+ */
-+static int pfe_eth_might_stop_tx(struct pfe_eth_priv_s *priv, int queuenum,
-+ struct netdev_queue *tx_queue,
-+ unsigned int n_desc,
-+ unsigned int n_segs)
-+{
-+ ktime_t kt;
-+ int tried = 0;
-+
-+try_again:
-+ if (unlikely((__hif_tx_avail(&pfe->hif) < n_desc) ||
-+ (hif_lib_tx_avail(&priv->client, queuenum) < n_desc) ||
-+ (hif_lib_tx_credit_avail(pfe, priv->id, queuenum) < n_segs))) {
-+ if (!tried) {
-+ __hif_lib_update_credit(&priv->client, queuenum);
-+ tried = 1;
-+ goto try_again;
-+ }
-+#ifdef PFE_ETH_TX_STATS
-+ if (__hif_tx_avail(&pfe->hif) < n_desc) {
-+ priv->stop_queue_hif[queuenum]++;
-+ } else if (hif_lib_tx_avail(&priv->client, queuenum) < n_desc) {
-+ priv->stop_queue_hif_client[queuenum]++;
-+ } else if (hif_lib_tx_credit_avail(pfe, priv->id, queuenum) <
-+ n_segs) {
-+ priv->stop_queue_credit[queuenum]++;
-+ }
-+ priv->stop_queue_total[queuenum]++;
-+#endif
-+ netif_tx_stop_queue(tx_queue);
-+
-+ kt = ktime_set(0, LS1012A_TX_FAST_RECOVERY_TIMEOUT_MS *
-+ NSEC_PER_MSEC);
-+ hrtimer_start(&priv->fast_tx_timeout[queuenum].timer, kt,
-+ HRTIMER_MODE_REL);
-+ return -1;
-+ } else {
-+ return 0;
-+ }
-+}
-+
-+#define SA_MAX_OP 2
-+/* pfe_hif_send_packet
-+ *
-+ * At this level if TX fails we drop the packet
-+ */
-+static void pfe_hif_send_packet(struct sk_buff *skb, struct pfe_eth_priv_s
-+ *priv, int queuenum)
-+{
-+ struct skb_shared_info *sh = skb_shinfo(skb);
-+ unsigned int nr_frags;
-+ u32 ctrl = 0;
-+
-+ netif_info(priv, tx_queued, priv->ndev, "%s\n", __func__);
-+
-+ if (skb_is_gso(skb)) {
-+ priv->stats.tx_dropped++;
-+ return;
-+ }
-+
-+ if (skb->ip_summed == CHECKSUM_PARTIAL)
-+ ctrl = HIF_CTRL_TX_CHECKSUM;
-+
-+ nr_frags = sh->nr_frags;
-+
-+ if (nr_frags) {
-+ skb_frag_t *f;
-+ int i;
-+
-+ __hif_lib_xmit_pkt(&priv->client, queuenum, skb->data,
-+ skb_headlen(skb), ctrl, HIF_FIRST_BUFFER,
-+ skb);
-+
-+ for (i = 0; i < nr_frags - 1; i++) {
-+ f = &sh->frags[i];
-+ __hif_lib_xmit_pkt(&priv->client, queuenum,
-+ skb_frag_address(f),
-+ skb_frag_size(f),
-+ 0x0, 0x0, skb);
-+ }
-+
-+ f = &sh->frags[i];
-+
-+ __hif_lib_xmit_pkt(&priv->client, queuenum,
-+ skb_frag_address(f), skb_frag_size(f),
-+ 0x0, HIF_LAST_BUFFER | HIF_DATA_VALID,
-+ skb);
-+
-+ netif_info(priv, tx_queued, priv->ndev,
-+ "%s: pkt sent successfully skb:%p nr_frags:%d len:%d\n",
-+ __func__, skb, nr_frags, skb->len);
-+ } else {
-+ __hif_lib_xmit_pkt(&priv->client, queuenum, skb->data,
-+ skb->len, ctrl, HIF_FIRST_BUFFER |
-+ HIF_LAST_BUFFER | HIF_DATA_VALID,
-+ skb);
-+ netif_info(priv, tx_queued, priv->ndev,
-+ "%s: pkt sent successfully skb:%p len:%d\n",
-+ __func__, skb, skb->len);
-+ }
-+ hif_tx_dma_start();
-+ priv->stats.tx_packets++;
-+ priv->stats.tx_bytes += skb->len;
-+ hif_lib_tx_credit_use(pfe, priv->id, queuenum, 1);
-+}
-+
-+/* pfe_eth_flush_txQ
-+ */
-+static void pfe_eth_flush_txQ(struct pfe_eth_priv_s *priv, int tx_q_num, int
-+ from_tx, int n_desc)
-+{
-+ struct sk_buff *skb;
-+ struct netdev_queue *tx_queue = netdev_get_tx_queue(priv->ndev,
-+ tx_q_num);
-+ unsigned int flags;
-+
-+ netif_info(priv, tx_done, priv->ndev, "%s\n", __func__);
-+
-+ if (!from_tx)
-+ __netif_tx_lock_bh(tx_queue);
-+
-+ /* Clean HIF and client queue */
-+ while ((skb = hif_lib_tx_get_next_complete(&priv->client,
-+ tx_q_num, &flags,
-+ HIF_TX_DESC_NT))) {
-+ if (flags & HIF_DATA_VALID)
-+ dev_kfree_skb_any(skb);
-+ }
-+ if (!from_tx)
-+ __netif_tx_unlock_bh(tx_queue);
-+}
-+
-+/* pfe_eth_flush_tx
-+ */
-+static void pfe_eth_flush_tx(struct pfe_eth_priv_s *priv)
-+{
-+ int ii;
-+
-+ netif_info(priv, tx_done, priv->ndev, "%s\n", __func__);
-+
-+ for (ii = 0; ii < emac_txq_cnt; ii++) {
-+ pfe_eth_flush_txQ(priv, ii, 0, 0);
-+ __hif_lib_update_credit(&priv->client, ii);
-+ }
-+}
-+
-+void pfe_tx_get_req_desc(struct sk_buff *skb, unsigned int *n_desc, unsigned int
-+ *n_segs)
-+{
-+ struct skb_shared_info *sh = skb_shinfo(skb);
-+
-+ /* Scattered data */
-+ if (sh->nr_frags) {
-+ *n_desc = sh->nr_frags + 1;
-+ *n_segs = 1;
-+ /* Regular case */
-+ } else {
-+ *n_desc = 1;
-+ *n_segs = 1;
-+ }
-+}
-+
-+/* pfe_eth_send_packet
-+ */
-+static int pfe_eth_send_packet(struct sk_buff *skb, struct net_device *ndev)
-+{
-+ struct pfe_eth_priv_s *priv = netdev_priv(ndev);
-+ int tx_q_num = skb_get_queue_mapping(skb);
-+ int n_desc, n_segs;
-+ struct netdev_queue *tx_queue = netdev_get_tx_queue(priv->ndev,
-+ tx_q_num);
-+
-+ netif_info(priv, tx_queued, ndev, "%s\n", __func__);
-+
-+ if ((!skb_is_gso(skb)) && (skb_headroom(skb) < (PFE_PKT_HEADER_SZ +
-+ sizeof(unsigned long)))) {
-+ netif_warn(priv, tx_err, priv->ndev, "%s: copying skb\n",
-+ __func__);
-+
-+ if (pskb_expand_head(skb, (PFE_PKT_HEADER_SZ + sizeof(unsigned
-+ long)), 0, GFP_ATOMIC)) {
-+ /* No need to re-transmit, no way to recover*/
-+ kfree_skb(skb);
-+ priv->stats.tx_dropped++;
-+ return NETDEV_TX_OK;
-+ }
-+ }
-+
-+ pfe_tx_get_req_desc(skb, &n_desc, &n_segs);
-+
-+ hif_tx_lock(&pfe->hif);
-+ if (unlikely(pfe_eth_might_stop_tx(priv, tx_q_num, tx_queue, n_desc,
-+ n_segs))) {
-+#ifdef PFE_ETH_TX_STATS
-+ if (priv->was_stopped[tx_q_num]) {
-+ priv->clean_fail[tx_q_num]++;
-+ priv->was_stopped[tx_q_num] = 0;
-+ }
-+#endif
-+ hif_tx_unlock(&pfe->hif);
-+ return NETDEV_TX_BUSY;
-+ }
-+
-+ pfe_hif_send_packet(skb, priv, tx_q_num);
-+
-+ hif_tx_unlock(&pfe->hif);
-+
-+ tx_queue->trans_start = jiffies;
-+
-+#ifdef PFE_ETH_TX_STATS
-+ priv->was_stopped[tx_q_num] = 0;
-+#endif
-+
-+ return NETDEV_TX_OK;
-+}
-+
-+/* pfe_eth_select_queue
-+ *
-+ */
-+static u16 pfe_eth_select_queue(struct net_device *ndev, struct sk_buff *skb,
-+ void *accel_priv,
-+ select_queue_fallback_t fallback)
-+{
-+ struct pfe_eth_priv_s *priv = netdev_priv(ndev);
-+
-+ return pfe_eth_get_queuenum(priv, skb);
-+}
-+
-+/* pfe_eth_get_stats
-+ */
-+static struct net_device_stats *pfe_eth_get_stats(struct net_device *ndev)
-+{
-+ struct pfe_eth_priv_s *priv = netdev_priv(ndev);
-+
-+ netif_info(priv, drv, ndev, "%s\n", __func__);
-+
-+ return &priv->stats;
-+}
-+
-+/* pfe_eth_set_mac_address
-+ */
-+static int pfe_eth_set_mac_address(struct net_device *ndev, void *addr)
-+{
-+ struct pfe_eth_priv_s *priv = netdev_priv(ndev);
-+ struct sockaddr *sa = addr;
-+
-+ netif_info(priv, drv, ndev, "%s\n", __func__);
-+
-+ if (!is_valid_ether_addr(sa->sa_data))
-+ return -EADDRNOTAVAIL;
-+
-+ memcpy(ndev->dev_addr, sa->sa_data, ETH_ALEN);
-+
-+ gemac_set_laddrN(priv->EMAC_baseaddr,
-+ (struct pfe_mac_addr *)ndev->dev_addr, 1);
-+
-+ return 0;
-+}
-+
-+/* pfe_eth_enet_addr_byte_mac
-+ */
-+int pfe_eth_enet_addr_byte_mac(u8 *enet_byte_addr,
-+ struct pfe_mac_addr *enet_addr)
-+{
-+ if (!enet_byte_addr || !enet_addr) {
-+ return -1;
-+
-+ } else {
-+ enet_addr->bottom = enet_byte_addr[0] |
-+ (enet_byte_addr[1] << 8) |
-+ (enet_byte_addr[2] << 16) |
-+ (enet_byte_addr[3] << 24);
-+ enet_addr->top = enet_byte_addr[4] |
-+ (enet_byte_addr[5] << 8);
-+ return 0;
-+ }
-+}
-+
-+/* pfe_eth_set_multi
-+ */
-+static void pfe_eth_set_multi(struct net_device *ndev)
-+{
-+ struct pfe_eth_priv_s *priv = netdev_priv(ndev);
-+ struct pfe_mac_addr hash_addr; /* hash register structure */
-+ /* specific mac address register structure */
-+ struct pfe_mac_addr spec_addr;
-+ int result; /* index into hash register to set.. */
-+ int uc_count = 0;
-+ struct netdev_hw_addr *ha;
-+
-+ if (ndev->flags & IFF_PROMISC) {
-+ netif_info(priv, drv, ndev, "entering promiscuous mode\n");
-+
-+ priv->promisc = 1;
-+ gemac_enable_copy_all(priv->EMAC_baseaddr);
-+ } else {
-+ priv->promisc = 0;
-+ gemac_disable_copy_all(priv->EMAC_baseaddr);
-+ }
-+
-+ /* Enable broadcast frame reception if required. */
-+ if (ndev->flags & IFF_BROADCAST) {
-+ gemac_allow_broadcast(priv->EMAC_baseaddr);
-+ } else {
-+ netif_info(priv, drv, ndev,
-+ "disabling broadcast frame reception\n");
-+
-+ gemac_no_broadcast(priv->EMAC_baseaddr);
-+ }
-+
-+ if (ndev->flags & IFF_ALLMULTI) {
-+ /* Set the hash to rx all multicast frames */
-+ hash_addr.bottom = 0xFFFFFFFF;
-+ hash_addr.top = 0xFFFFFFFF;
-+ gemac_set_hash(priv->EMAC_baseaddr, &hash_addr);
-+ netdev_for_each_uc_addr(ha, ndev) {
-+ if (uc_count >= MAX_UC_SPEC_ADDR_REG)
-+ break;
-+ pfe_eth_enet_addr_byte_mac(ha->addr, &spec_addr);
-+ gemac_set_laddrN(priv->EMAC_baseaddr, &spec_addr,
-+ uc_count + 2);
-+ uc_count++;
-+ }
-+ } else if ((netdev_mc_count(ndev) > 0) || (netdev_uc_count(ndev))) {
-+ u8 *addr;
-+
-+ hash_addr.bottom = 0;
-+ hash_addr.top = 0;
-+
-+ netdev_for_each_mc_addr(ha, ndev) {
-+ addr = ha->addr;
-+
-+ netif_info(priv, drv, ndev,
-+ "adding multicast address %X:%X:%X:%X:%X:%X to gem filter\n",
-+ addr[0], addr[1], addr[2],
-+ addr[3], addr[4], addr[5]);
-+
-+ result = pfe_eth_get_hash(addr);
-+
-+ if (result < EMAC_HASH_REG_BITS) {
-+ if (result < 32)
-+ hash_addr.bottom |= (1 << result);
-+ else
-+ hash_addr.top |= (1 << (result - 32));
-+ } else {
-+ break;
-+ }
-+ }
-+
-+ uc_count = -1;
-+ netdev_for_each_uc_addr(ha, ndev) {
-+ addr = ha->addr;
-+
-+ if (++uc_count < MAX_UC_SPEC_ADDR_REG) {
-+ netdev_info(ndev,
-+ "adding unicast address %02x:%02x:%02x:%02x:%02x:%02x to gem filter\n",
-+ addr[0], addr[1], addr[2],
-+ addr[3], addr[4], addr[5]);
-+ pfe_eth_enet_addr_byte_mac(addr, &spec_addr);
-+ gemac_set_laddrN(priv->EMAC_baseaddr,
-+ &spec_addr, uc_count + 2);
-+ } else {
-+ netif_info(priv, drv, ndev,
-+ "adding unicast address %02x:%02x:%02x:%02x:%02x:%02x to gem hash\n",
-+ addr[0], addr[1], addr[2],
-+ addr[3], addr[4], addr[5]);
-+
-+ result = pfe_eth_get_hash(addr);
-+ if (result >= EMAC_HASH_REG_BITS) {
-+ break;
-+
-+ } else {
-+ if (result < 32)
-+ hash_addr.bottom |= (1 <<
-+ result);
-+ else
-+ hash_addr.top |= (1 <<
-+ (result - 32));
-+ }
-+ }
-+ }
-+
-+ gemac_set_hash(priv->EMAC_baseaddr, &hash_addr);
-+ }
-+
-+ if (!(netdev_uc_count(ndev) >= MAX_UC_SPEC_ADDR_REG)) {
-+ /*
-+ * Check if there are any specific address HW registers that
-+ * need to be flushed
-+ */
-+ for (uc_count = netdev_uc_count(ndev); uc_count <
-+ MAX_UC_SPEC_ADDR_REG; uc_count++)
-+ gemac_clear_laddrN(priv->EMAC_baseaddr, uc_count + 2);
-+ }
-+
-+ if (ndev->flags & IFF_LOOPBACK)
-+ gemac_set_loop(priv->EMAC_baseaddr, LB_LOCAL);
-+}
-+
-+/* pfe_eth_set_features
-+ */
-+static int pfe_eth_set_features(struct net_device *ndev, netdev_features_t
-+ features)
-+{
-+ struct pfe_eth_priv_s *priv = netdev_priv(ndev);
-+ int rc = 0;
-+
-+ if (features & NETIF_F_RXCSUM)
-+ gemac_enable_rx_checksum_offload(priv->EMAC_baseaddr);
-+ else
-+ gemac_disable_rx_checksum_offload(priv->EMAC_baseaddr);
-+ return rc;
-+}
-+
-+/* pfe_eth_fast_tx_timeout
-+ */
-+static enum hrtimer_restart pfe_eth_fast_tx_timeout(struct hrtimer *timer)
-+{
-+ struct pfe_eth_fast_timer *fast_tx_timeout = container_of(timer, struct
-+ pfe_eth_fast_timer,
-+ timer);
-+ struct pfe_eth_priv_s *priv = container_of(fast_tx_timeout->base,
-+ struct pfe_eth_priv_s,
-+ fast_tx_timeout);
-+ struct netdev_queue *tx_queue = netdev_get_tx_queue(priv->ndev,
-+ fast_tx_timeout->queuenum);
-+
-+ if (netif_tx_queue_stopped(tx_queue)) {
-+#ifdef PFE_ETH_TX_STATS
-+ priv->was_stopped[fast_tx_timeout->queuenum] = 1;
-+#endif
-+ netif_tx_wake_queue(tx_queue);
-+ }
-+
-+ return HRTIMER_NORESTART;
-+}
-+
-+/* pfe_eth_fast_tx_timeout_init
-+ */
-+static void pfe_eth_fast_tx_timeout_init(struct pfe_eth_priv_s *priv)
-+{
-+ int i;
-+
-+ for (i = 0; i < emac_txq_cnt; i++) {
-+ priv->fast_tx_timeout[i].queuenum = i;
-+ hrtimer_init(&priv->fast_tx_timeout[i].timer, CLOCK_MONOTONIC,
-+ HRTIMER_MODE_REL);
-+ priv->fast_tx_timeout[i].timer.function =
-+ pfe_eth_fast_tx_timeout;
-+ priv->fast_tx_timeout[i].base = priv->fast_tx_timeout;
-+ }
-+}
-+
-+static struct sk_buff *pfe_eth_rx_skb(struct net_device *ndev,
-+ struct pfe_eth_priv_s *priv,
-+ unsigned int qno)
-+{
-+ void *buf_addr;
-+ unsigned int rx_ctrl;
-+ unsigned int desc_ctrl = 0;
-+ struct hif_ipsec_hdr *ipsec_hdr = NULL;
-+ struct sk_buff *skb;
-+ struct sk_buff *skb_frag, *skb_frag_last = NULL;
-+ int length = 0, offset;
-+
-+ skb = priv->skb_inflight[qno];
-+
-+ if (skb) {
-+ skb_frag_last = skb_shinfo(skb)->frag_list;
-+ if (skb_frag_last) {
-+ while (skb_frag_last->next)
-+ skb_frag_last = skb_frag_last->next;
-+ }
-+ }
-+
-+ while (!(desc_ctrl & CL_DESC_LAST)) {
-+ buf_addr = hif_lib_receive_pkt(&priv->client, qno, &length,
-+ &offset, &rx_ctrl, &desc_ctrl,
-+ (void **)&ipsec_hdr);
-+ if (!buf_addr)
-+ goto incomplete;
-+
-+#ifdef PFE_ETH_NAPI_STATS
-+ priv->napi_counters[NAPI_DESC_COUNT]++;
-+#endif
-+
-+ /* First frag */
-+ if (desc_ctrl & CL_DESC_FIRST) {
-+ skb = build_skb(buf_addr, 0);
-+ if (unlikely(!skb))
-+ goto pkt_drop;
-+
-+ skb_reserve(skb, offset);
-+ skb_put(skb, length);
-+ skb->dev = ndev;
-+
-+ if ((ndev->features & NETIF_F_RXCSUM) && (rx_ctrl &
-+ HIF_CTRL_RX_CHECKSUMMED))
-+ skb->ip_summed = CHECKSUM_UNNECESSARY;
-+ else
-+ skb_checksum_none_assert(skb);
-+
-+ } else {
-+ /* Next frags */
-+ if (unlikely(!skb)) {
-+ pr_err("%s: NULL skb_inflight\n",
-+ __func__);
-+ goto pkt_drop;
-+ }
-+
-+ skb_frag = build_skb(buf_addr, 0);
-+
-+ if (unlikely(!skb_frag)) {
-+ kfree(buf_addr);
-+ goto pkt_drop;
-+ }
-+
-+ skb_reserve(skb_frag, offset);
-+ skb_put(skb_frag, length);
-+
-+ skb_frag->dev = ndev;
-+
-+ if (skb_shinfo(skb)->frag_list)
-+ skb_frag_last->next = skb_frag;
-+ else
-+ skb_shinfo(skb)->frag_list = skb_frag;
-+
-+ skb->truesize += skb_frag->truesize;
-+ skb->data_len += length;
-+ skb->len += length;
-+ skb_frag_last = skb_frag;
-+ }
-+ }
-+
-+ priv->skb_inflight[qno] = NULL;
-+ return skb;
-+
-+incomplete:
-+ priv->skb_inflight[qno] = skb;
-+ return NULL;
-+
-+pkt_drop:
-+ priv->skb_inflight[qno] = NULL;
-+
-+ if (skb)
-+ kfree_skb(skb);
-+ else
-+ kfree(buf_addr);
-+
-+ priv->stats.rx_errors++;
-+
-+ return NULL;
-+}
-+
-+/* pfe_eth_poll
-+ */
-+static int pfe_eth_poll(struct pfe_eth_priv_s *priv, struct napi_struct *napi,
-+ unsigned int qno, int budget)
-+{
-+ struct net_device *ndev = priv->ndev;
-+ struct sk_buff *skb;
-+ int work_done = 0;
-+ unsigned int len;
-+
-+ netif_info(priv, intr, priv->ndev, "%s\n", __func__);
-+
-+#ifdef PFE_ETH_NAPI_STATS
-+ priv->napi_counters[NAPI_POLL_COUNT]++;
-+#endif
-+
-+ do {
-+ skb = pfe_eth_rx_skb(ndev, priv, qno);
-+
-+ if (!skb)
-+ break;
-+
-+ len = skb->len;
-+
-+ /* Packet will be processed */
-+ skb->protocol = eth_type_trans(skb, ndev);
-+
-+ netif_receive_skb(skb);
-+
-+ priv->stats.rx_packets++;
-+ priv->stats.rx_bytes += len;
-+
-+ work_done++;
-+
-+#ifdef PFE_ETH_NAPI_STATS
-+ priv->napi_counters[NAPI_PACKET_COUNT]++;
-+#endif
-+
-+ } while (work_done < budget);
-+
-+ /*
-+ * If no Rx receive nor cleanup work was done, exit polling mode.
-+ * No more netif_running(dev) check is required here , as this is
-+ * checked in net/core/dev.c (2.6.33.5 kernel specific).
-+ */
-+ if (work_done < budget) {
-+ napi_complete(napi);
-+
-+ hif_lib_event_handler_start(&priv->client, EVENT_RX_PKT_IND,
-+ qno);
-+ }
-+#ifdef PFE_ETH_NAPI_STATS
-+ else
-+ priv->napi_counters[NAPI_FULL_BUDGET_COUNT]++;
-+#endif
-+
-+ return work_done;
-+}
-+
-+/*
-+ * pfe_eth_lro_poll
-+ */
-+static int pfe_eth_lro_poll(struct napi_struct *napi, int budget)
-+{
-+ struct pfe_eth_priv_s *priv = container_of(napi, struct pfe_eth_priv_s,
-+ lro_napi);
-+
-+ netif_info(priv, intr, priv->ndev, "%s\n", __func__);
-+
-+ return pfe_eth_poll(priv, napi, 2, budget);
-+}
-+
-+/* pfe_eth_low_poll
-+ */
-+static int pfe_eth_low_poll(struct napi_struct *napi, int budget)
-+{
-+ struct pfe_eth_priv_s *priv = container_of(napi, struct pfe_eth_priv_s,
-+ low_napi);
-+
-+ netif_info(priv, intr, priv->ndev, "%s\n", __func__);
-+
-+ return pfe_eth_poll(priv, napi, 1, budget);
-+}
-+
-+/* pfe_eth_high_poll
-+ */
-+static int pfe_eth_high_poll(struct napi_struct *napi, int budget)
-+{
-+ struct pfe_eth_priv_s *priv = container_of(napi, struct pfe_eth_priv_s,
-+ high_napi);
-+
-+ netif_info(priv, intr, priv->ndev, "%s\n", __func__);
-+
-+ return pfe_eth_poll(priv, napi, 0, budget);
-+}
-+
-+static const struct net_device_ops pfe_netdev_ops = {
-+ .ndo_open = pfe_eth_open,
-+ .ndo_stop = pfe_eth_close,
-+ .ndo_start_xmit = pfe_eth_send_packet,
-+ .ndo_select_queue = pfe_eth_select_queue,
-+ .ndo_get_stats = pfe_eth_get_stats,
-+ .ndo_set_mac_address = pfe_eth_set_mac_address,
-+ .ndo_set_rx_mode = pfe_eth_set_multi,
-+ .ndo_set_features = pfe_eth_set_features,
-+ .ndo_validate_addr = eth_validate_addr,
-+};
-+
-+/* pfe_eth_init_one
-+ */
-+static int pfe_eth_init_one(struct pfe *pfe, int id)
-+{
-+ struct net_device *ndev = NULL;
-+ struct pfe_eth_priv_s *priv = NULL;
-+ struct ls1012a_eth_platform_data *einfo;
-+ struct ls1012a_mdio_platform_data *minfo;
-+ struct ls1012a_pfe_platform_data *pfe_info;
-+ int err;
-+
-+ /* Extract pltform data */
-+ pfe_info = (struct ls1012a_pfe_platform_data *)
-+ pfe->dev->platform_data;
-+ if (!pfe_info) {
-+ pr_err(
-+ "%s: pfe missing additional platform data\n"
-+ , __func__);
-+ err = -ENODEV;
-+ goto err0;
-+ }
-+
-+ einfo = (struct ls1012a_eth_platform_data *)
-+ pfe_info->ls1012a_eth_pdata;
-+
-+ /* einfo never be NULL, but no harm in having this check */
-+ if (!einfo) {
-+ pr_err(
-+ "%s: pfe missing additional gemacs platform data\n"
-+ , __func__);
-+ err = -ENODEV;
-+ goto err0;
-+ }
-+
-+ minfo = (struct ls1012a_mdio_platform_data *)
-+ pfe_info->ls1012a_mdio_pdata;
-+
-+ /* einfo never be NULL, but no harm in having this check */
-+ if (!minfo) {
-+ pr_err(
-+ "%s: pfe missing additional mdios platform data\n",
-+ __func__);
-+ err = -ENODEV;
-+ goto err0;
-+ }
-+
-+ if (us)
-+ emac_txq_cnt = EMAC_TXQ_CNT;
-+ /* Create an ethernet device instance */
-+ ndev = alloc_etherdev_mq(sizeof(*priv), emac_txq_cnt);
-+
-+ if (!ndev) {
-+ pr_err("%s: gemac %d device allocation failed\n",
-+ __func__, einfo[id].gem_id);
-+ err = -ENOMEM;
-+ goto err0;
-+ }
-+
-+ priv = netdev_priv(ndev);
-+ priv->ndev = ndev;
-+ priv->id = einfo[id].gem_id;
-+ priv->pfe = pfe;
-+
-+ SET_NETDEV_DEV(priv->ndev, priv->pfe->dev);
-+
-+ pfe->eth.eth_priv[id] = priv;
-+
-+ /* Set the info in the priv to the current info */
-+ priv->einfo = &einfo[id];
-+ priv->EMAC_baseaddr = cbus_emac_base[id];
-+ priv->PHY_baseaddr = cbus_emac_base[0];
-+ priv->GPI_baseaddr = cbus_gpi_base[id];
-+
-+#define HIF_GEMAC_TMUQ_BASE 6
-+ priv->low_tmu_q = HIF_GEMAC_TMUQ_BASE + (id * 2);
-+ priv->high_tmu_q = priv->low_tmu_q + 1;
-+
-+ spin_lock_init(&priv->lock);
-+
-+ pfe_eth_fast_tx_timeout_init(priv);
-+
-+ /* Copy the station address into the dev structure, */
-+ memcpy(ndev->dev_addr, einfo[id].mac_addr, ETH_ALEN);
-+
-+ /* Initialize mdio */
-+ if (minfo[id].enabled) {
-+ err = pfe_eth_mdio_init(priv, &minfo[id]);
-+ if (err) {
-+ netdev_err(ndev, "%s: pfe_eth_mdio_init() failed\n",
-+ __func__);
-+ goto err2;
-+ }
-+ }
-+
-+ if (us)
-+ goto phy_init;
-+
-+ ndev->mtu = 1500;
-+
-+ /* Set MTU limits */
-+ ndev->min_mtu = ETH_MIN_MTU;
-+ ndev->max_mtu = JUMBO_FRAME_SIZE;
-+
-+ /* supported features */
-+ ndev->hw_features = NETIF_F_SG;
-+
-+ /*Enable after checksum offload is validated */
-+ ndev->hw_features = NETIF_F_RXCSUM | NETIF_F_IP_CSUM |
-+ NETIF_F_IPV6_CSUM | NETIF_F_SG;
-+
-+ /* enabled by default */
-+ ndev->features = ndev->hw_features;
-+
-+ priv->usr_features = ndev->features;
-+
-+ ndev->netdev_ops = &pfe_netdev_ops;
-+
-+ ndev->ethtool_ops = &pfe_ethtool_ops;
-+
-+ /* Enable basic messages by default */
-+ priv->msg_enable = NETIF_MSG_IFUP | NETIF_MSG_IFDOWN | NETIF_MSG_LINK |
-+ NETIF_MSG_PROBE;
-+
-+ netif_napi_add(ndev, &priv->low_napi, pfe_eth_low_poll,
-+ HIF_RX_POLL_WEIGHT - 16);
-+ netif_napi_add(ndev, &priv->high_napi, pfe_eth_high_poll,
-+ HIF_RX_POLL_WEIGHT - 16);
-+ netif_napi_add(ndev, &priv->lro_napi, pfe_eth_lro_poll,
-+ HIF_RX_POLL_WEIGHT - 16);
-+
-+ err = register_netdev(ndev);
-+
-+ if (err) {
-+ netdev_err(ndev, "register_netdev() failed\n");
-+ goto err3;
-+ }
-+
-+phy_init:
-+ device_init_wakeup(&ndev->dev, WAKE_MAGIC);
-+
-+ if (!(priv->einfo->phy_flags & GEMAC_NO_PHY)) {
-+ err = pfe_phy_init(ndev);
-+ if (err) {
-+ netdev_err(ndev, "%s: pfe_phy_init() failed\n",
-+ __func__);
-+ goto err4;
-+ }
-+ }
-+
-+ if (us) {
-+ if (priv->phydev)
-+ phy_start(priv->phydev);
-+ return 0;
-+ }
-+
-+ netif_carrier_on(ndev);
-+
-+ /* Create all the sysfs files */
-+ if (pfe_eth_sysfs_init(ndev))
-+ goto err4;
-+
-+ netif_info(priv, probe, ndev, "%s: created interface, baseaddr: %p\n",
-+ __func__, priv->EMAC_baseaddr);
-+
-+ return 0;
-+err4:
-+ if (us)
-+ goto err3;
-+ unregister_netdev(ndev);
-+err3:
-+ pfe_eth_mdio_exit(priv->mii_bus);
-+err2:
-+ free_netdev(priv->ndev);
-+err0:
-+ return err;
-+}
-+
-+/* pfe_eth_init
-+ */
-+int pfe_eth_init(struct pfe *pfe)
-+{
-+ int ii = 0;
-+ int err;
-+
-+ pr_info("%s\n", __func__);
-+
-+ cbus_emac_base[0] = EMAC1_BASE_ADDR;
-+ cbus_emac_base[1] = EMAC2_BASE_ADDR;
-+
-+ cbus_gpi_base[0] = EGPI1_BASE_ADDR;
-+ cbus_gpi_base[1] = EGPI2_BASE_ADDR;
-+
-+ for (ii = 0; ii < NUM_GEMAC_SUPPORT; ii++) {
-+ err = pfe_eth_init_one(pfe, ii);
-+ if (err)
-+ goto err0;
-+ }
-+
-+ return 0;
-+
-+err0:
-+ while (ii--)
-+ pfe_eth_exit_one(pfe->eth.eth_priv[ii]);
-+
-+ /* Register three network devices in the kernel */
-+ return err;
-+}
-+
-+/* pfe_eth_exit_one
-+ */
-+static void pfe_eth_exit_one(struct pfe_eth_priv_s *priv)
-+{
-+ netif_info(priv, probe, priv->ndev, "%s\n", __func__);
-+
-+ if (!us)
-+ pfe_eth_sysfs_exit(priv->ndev);
-+
-+ if (!(priv->einfo->phy_flags & GEMAC_NO_PHY))
-+ pfe_phy_exit(priv->ndev);
-+
-+ if (!us)
-+ unregister_netdev(priv->ndev);
-+
-+ if (priv->mii_bus)
-+ pfe_eth_mdio_exit(priv->mii_bus);
-+
-+ free_netdev(priv->ndev);
-+}
-+
-+/* pfe_eth_exit
-+ */
-+void pfe_eth_exit(struct pfe *pfe)
-+{
-+ int ii;
-+
-+ pr_info("%s\n", __func__);
-+
-+ for (ii = NUM_GEMAC_SUPPORT - 1; ii >= 0; ii--)
-+ pfe_eth_exit_one(pfe->eth.eth_priv[ii]);
-+}
---- /dev/null
-+++ b/drivers/staging/fsl_ppfe/pfe_eth.h
-@@ -0,0 +1,184 @@
-+/*
-+ * Copyright 2015-2016 Freescale Semiconductor, Inc.
-+ * Copyright 2017 NXP
-+ *
-+ * This program is free software; you can redistribute it and/or modify
-+ * it under the terms of the GNU General Public License as published by
-+ * the Free Software Foundation; either version 2 of the License, or
-+ * (at your option) any later version.
-+ *
-+ * This program is distributed in the hope that it will be useful,
-+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
-+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
-+ * GNU General Public License for more details.
-+ *
-+ * You should have received a copy of the GNU General Public License
-+ * along with this program. If not, see <http://www.gnu.org/licenses/>.
-+ */
-+
-+#ifndef _PFE_ETH_H_
-+#define _PFE_ETH_H_
-+#include <linux/kernel.h>
-+#include <linux/netdevice.h>
-+#include <linux/etherdevice.h>
-+#include <linux/ethtool.h>
-+#include <linux/mii.h>
-+#include <linux/phy.h>
-+#include <linux/clk.h>
-+#include <linux/interrupt.h>
-+#include <linux/time.h>
-+
-+#define PFE_ETH_NAPI_STATS
-+#define PFE_ETH_TX_STATS
-+
-+#define PFE_ETH_FRAGS_MAX (65536 / HIF_RX_PKT_MIN_SIZE)
-+#define LRO_LEN_COUNT_MAX 32
-+#define LRO_NB_COUNT_MAX 32
-+
-+#define PFE_PAUSE_FLAG_ENABLE 1
-+#define PFE_PAUSE_FLAG_AUTONEG 2
-+
-+/* GEMAC configured by SW */
-+/* GEMAC configured by phy lines (not for MII/GMII) */
-+
-+#define GEMAC_SW_FULL_DUPLEX BIT(9)
-+#define GEMAC_SW_SPEED_10M (0 << 12)
-+#define GEMAC_SW_SPEED_100M BIT(12)
-+#define GEMAC_SW_SPEED_1G (2 << 12)
-+
-+#define GEMAC_NO_PHY BIT(0)
-+
-+struct ls1012a_eth_platform_data {
-+ /* device specific information */
-+ u32 device_flags;
-+ char name[16];
-+
-+ /* board specific information */
-+ u32 mii_config;
-+ u32 phy_flags;
-+ u32 gem_id;
-+ u32 bus_id;
-+ u32 phy_id;
-+ u32 mdio_muxval;
-+ u8 mac_addr[ETH_ALEN];
-+};
-+
-+struct ls1012a_mdio_platform_data {
-+ int enabled;
-+ int irq[32];
-+ u32 phy_mask;
-+ int mdc_div;
-+};
-+
-+struct ls1012a_pfe_platform_data {
-+ struct ls1012a_eth_platform_data ls1012a_eth_pdata[3];
-+ struct ls1012a_mdio_platform_data ls1012a_mdio_pdata[3];
-+};
-+
-+#define NUM_GEMAC_SUPPORT 2
-+#define DRV_NAME "pfe-eth"
-+#define DRV_VERSION "1.0"
-+
-+#define LS1012A_TX_FAST_RECOVERY_TIMEOUT_MS 3
-+#define TX_POLL_TIMEOUT_MS 1000
-+
-+#define EMAC_TXQ_CNT 16
-+#define EMAC_TXQ_DEPTH (HIF_TX_DESC_NT)
-+
-+#define JUMBO_FRAME_SIZE 10258
-+/*
-+ * Client Tx queue threshold, for txQ flush condition.
-+ * It must be smaller than the queue size (in case we ever change it in the
-+ * future).
-+ */
-+#define HIF_CL_TX_FLUSH_MARK 32
-+
-+/*
-+ * Max number of TX resources (HIF descriptors or skbs) that will be released
-+ * in a single go during batch recycling.
-+ * Should be lower than the flush mark so the SW can provide the HW with a
-+ * continuous stream of packets instead of bursts.
-+ */
-+#define TX_FREE_MAX_COUNT 16
-+#define EMAC_RXQ_CNT 3
-+#define EMAC_RXQ_DEPTH HIF_RX_DESC_NT
-+/* make sure clients can receive a full burst of packets */
-+#define EMAC_RMON_TXBYTES_POS 0x00
-+#define EMAC_RMON_RXBYTES_POS 0x14
-+
-+#define EMAC_QUEUENUM_MASK (emac_txq_cnt - 1)
-+#define EMAC_MDIO_TIMEOUT 1000
-+#define MAX_UC_SPEC_ADDR_REG 31
-+
-+struct pfe_eth_fast_timer {
-+ int queuenum;
-+ struct hrtimer timer;
-+ void *base;
-+};
-+
-+struct pfe_eth_priv_s {
-+ struct pfe *pfe;
-+ struct hif_client_s client;
-+ struct napi_struct lro_napi;
-+ struct napi_struct low_napi;
-+ struct napi_struct high_napi;
-+ int low_tmu_q;
-+ int high_tmu_q;
-+ struct net_device_stats stats;
-+ struct net_device *ndev;
-+ int id;
-+ int promisc;
-+ unsigned int msg_enable;
-+ unsigned int usr_features;
-+
-+ spinlock_t lock; /* protect member variables */
-+ unsigned int event_status;
-+ int irq;
-+ void *EMAC_baseaddr;
-+ /* This points to the EMAC base from where we access PHY */
-+ void *PHY_baseaddr;
-+ void *GPI_baseaddr;
-+ /* PHY stuff */
-+ struct phy_device *phydev;
-+ int oldspeed;
-+ int oldduplex;
-+ int oldlink;
-+ /* mdio info */
-+ int mdc_div;
-+ struct mii_bus *mii_bus;
-+ struct clk *gemtx_clk;
-+ int wol;
-+ int pause_flag;
-+
-+ int default_priority;
-+ struct pfe_eth_fast_timer fast_tx_timeout[EMAC_TXQ_CNT];
-+
-+ struct ls1012a_eth_platform_data *einfo;
-+ struct sk_buff *skb_inflight[EMAC_RXQ_CNT + 6];
-+
-+#ifdef PFE_ETH_TX_STATS
-+ unsigned int stop_queue_total[EMAC_TXQ_CNT];
-+ unsigned int stop_queue_hif[EMAC_TXQ_CNT];
-+ unsigned int stop_queue_hif_client[EMAC_TXQ_CNT];
-+ unsigned int stop_queue_credit[EMAC_TXQ_CNT];
-+ unsigned int clean_fail[EMAC_TXQ_CNT];
-+ unsigned int was_stopped[EMAC_TXQ_CNT];
-+#endif
-+
-+#ifdef PFE_ETH_NAPI_STATS
-+ unsigned int napi_counters[NAPI_MAX_COUNT];
-+#endif
-+ unsigned int frags_inflight[EMAC_RXQ_CNT + 6];
-+};
-+
-+struct pfe_eth {
-+ struct pfe_eth_priv_s *eth_priv[3];
-+};
-+
-+int pfe_eth_init(struct pfe *pfe);
-+void pfe_eth_exit(struct pfe *pfe);
-+int pfe_eth_suspend(struct net_device *dev);
-+int pfe_eth_resume(struct net_device *dev);
-+int pfe_eth_mdio_reset(struct mii_bus *bus);
-+
-+#endif /* _PFE_ETH_H_ */
---- /dev/null
-+++ b/drivers/staging/fsl_ppfe/pfe_firmware.c
-@@ -0,0 +1,314 @@
-+/*
-+ * Copyright 2015-2016 Freescale Semiconductor, Inc.
-+ * Copyright 2017 NXP
-+ *
-+ * This program is free software; you can redistribute it and/or modify
-+ * it under the terms of the GNU General Public License as published by
-+ * the Free Software Foundation; either version 2 of the License, or
-+ * (at your option) any later version.
-+ *
-+ * This program is distributed in the hope that it will be useful,
-+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
-+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
-+ * GNU General Public License for more details.
-+ *
-+ * You should have received a copy of the GNU General Public License
-+ * along with this program. If not, see <http://www.gnu.org/licenses/>.
-+ */
-+
-+/*
-+ * @file
-+ * Contains all the functions to handle parsing and loading of PE firmware
-+ * files.
-+ */
-+#include <linux/firmware.h>
-+
-+#include "pfe_mod.h"
-+#include "pfe_firmware.h"
-+#include "pfe/pfe.h"
-+
-+static struct elf32_shdr *get_elf_section_header(const struct firmware *fw,
-+ const char *section)
-+{
-+ struct elf32_hdr *elf_hdr = (struct elf32_hdr *)fw->data;
-+ struct elf32_shdr *shdr;
-+ struct elf32_shdr *shdr_shstr;
-+ Elf32_Off e_shoff = be32_to_cpu(elf_hdr->e_shoff);
-+ Elf32_Half e_shentsize = be16_to_cpu(elf_hdr->e_shentsize);
-+ Elf32_Half e_shnum = be16_to_cpu(elf_hdr->e_shnum);
-+ Elf32_Half e_shstrndx = be16_to_cpu(elf_hdr->e_shstrndx);
-+ Elf32_Off shstr_offset;
-+ Elf32_Word sh_name;
-+ const char *name;
-+ int i;
-+
-+ /* Section header strings */
-+ shdr_shstr = (struct elf32_shdr *)(fw->data + e_shoff + e_shstrndx *
-+ e_shentsize);
-+ shstr_offset = be32_to_cpu(shdr_shstr->sh_offset);
-+
-+ for (i = 0; i < e_shnum; i++) {
-+ shdr = (struct elf32_shdr *)(fw->data + e_shoff
-+ + i * e_shentsize);
-+
-+ sh_name = be32_to_cpu(shdr->sh_name);
-+
-+ name = (const char *)(fw->data + shstr_offset + sh_name);
-+
-+ if (!strcmp(name, section))
-+ return shdr;
-+ }
-+
-+ pr_err("%s: didn't find section %s\n", __func__, section);
-+
-+ return NULL;
-+}
-+
-+#if defined(CFG_DIAGS)
-+static int pfe_get_diags_info(const struct firmware *fw, struct pfe_diags_info
-+ *diags_info)
-+{
-+ struct elf32_shdr *shdr;
-+ unsigned long offset, size;
-+
-+ shdr = get_elf_section_header(fw, ".pfe_diags_str");
-+ if (shdr) {
-+ offset = be32_to_cpu(shdr->sh_offset);
-+ size = be32_to_cpu(shdr->sh_size);
-+ diags_info->diags_str_base = be32_to_cpu(shdr->sh_addr);
-+ diags_info->diags_str_size = size;
-+ diags_info->diags_str_array = kmalloc(size, GFP_KERNEL);
-+ memcpy(diags_info->diags_str_array, fw->data + offset, size);
-+
-+ return 0;
-+ } else {
-+ return -1;
-+ }
-+}
-+#endif
-+
-+static void pfe_check_version_info(const struct firmware *fw)
-+{
-+ /*static char *version = NULL;*/
-+ static char *version;
-+
-+ struct elf32_shdr *shdr = get_elf_section_header(fw, ".version");
-+
-+ if (shdr) {
-+ if (!version) {
-+ /*
-+ * this is the first fw we load, use its version
-+ * string as reference (whatever it is)
-+ */
-+ version = (char *)(fw->data +
-+ be32_to_cpu(shdr->sh_offset));
-+
-+ pr_info("PFE binary version: %s\n", version);
-+ } else {
-+ /*
-+ * already have loaded at least one firmware, check
-+ * sequence can start now
-+ */
-+ if (strcmp(version, (char *)(fw->data +
-+ be32_to_cpu(shdr->sh_offset)))) {
-+ pr_info(
-+ "WARNING: PFE firmware binaries from incompatible version\n");
-+ }
-+ }
-+ } else {
-+ /*
-+ * version cannot be verified, a potential issue that should
-+ * be reported
-+ */
-+ pr_info(
-+ "WARNING: PFE firmware binaries from incompatible version\n");
-+ }
-+}
-+
-+/* PFE elf firmware loader.
-+ * Loads an elf firmware image into a list of PE's (specified using a bitmask)
-+ *
-+ * @param pe_mask Mask of PE id's to load firmware to
-+ * @param fw Pointer to the firmware image
-+ *
-+ * @return 0 on success, a negative value on error
-+ *
-+ */
-+int pfe_load_elf(int pe_mask, const struct firmware *fw, struct pfe *pfe)
-+{
-+ struct elf32_hdr *elf_hdr = (struct elf32_hdr *)fw->data;
-+ Elf32_Half sections = be16_to_cpu(elf_hdr->e_shnum);
-+ struct elf32_shdr *shdr = (struct elf32_shdr *)(fw->data +
-+ be32_to_cpu(elf_hdr->e_shoff));
-+ int id, section;
-+ int rc;
-+
-+ pr_info("%s\n", __func__);
-+
-+ /* Some sanity checks */
-+ if (strncmp(&elf_hdr->e_ident[EI_MAG0], ELFMAG, SELFMAG)) {
-+ pr_err("%s: incorrect elf magic number\n", __func__);
-+ return -EINVAL;
-+ }
-+
-+ if (elf_hdr->e_ident[EI_CLASS] != ELFCLASS32) {
-+ pr_err("%s: incorrect elf class(%x)\n", __func__,
-+ elf_hdr->e_ident[EI_CLASS]);
-+ return -EINVAL;
-+ }
-+
-+ if (elf_hdr->e_ident[EI_DATA] != ELFDATA2MSB) {
-+ pr_err("%s: incorrect elf data(%x)\n", __func__,
-+ elf_hdr->e_ident[EI_DATA]);
-+ return -EINVAL;
-+ }
-+
-+ if (be16_to_cpu(elf_hdr->e_type) != ET_EXEC) {
-+ pr_err("%s: incorrect elf file type(%x)\n", __func__,
-+ be16_to_cpu(elf_hdr->e_type));
-+ return -EINVAL;
-+ }
-+
-+ for (section = 0; section < sections; section++, shdr++) {
-+ if (!(be32_to_cpu(shdr->sh_flags) & (SHF_WRITE | SHF_ALLOC |
-+ SHF_EXECINSTR)))
-+ continue;
-+
-+ for (id = 0; id < MAX_PE; id++)
-+ if (pe_mask & (1 << id)) {
-+ rc = pe_load_elf_section(id, fw->data, shdr,
-+ pfe->dev);
-+ if (rc < 0)
-+ goto err;
-+ }
-+ }
-+
-+ pfe_check_version_info(fw);
-+
-+ return 0;
-+
-+err:
-+ return rc;
-+}
-+
-+/* PFE firmware initialization.
-+ * Loads different firmware files from filesystem.
-+ * Initializes PE IMEM/DMEM and UTIL-PE DDR
-+ * Initializes control path symbol addresses (by looking them up in the elf
-+ * firmware files
-+ * Takes PE's out of reset
-+ *
-+ * @return 0 on success, a negative value on error
-+ *
-+ */
-+int pfe_firmware_init(struct pfe *pfe)
-+{
-+ const struct firmware *class_fw, *tmu_fw;
-+ int rc = 0;
-+#if !defined(CONFIG_FSL_PPFE_UTIL_DISABLED)
-+ const char *util_fw_name;
-+ const struct firmware *util_fw;
-+#endif
-+
-+ pr_info("%s\n", __func__);
-+
-+ if (request_firmware(&class_fw, CLASS_FIRMWARE_FILENAME, pfe->dev)) {
-+ pr_err("%s: request firmware %s failed\n", __func__,
-+ CLASS_FIRMWARE_FILENAME);
-+ rc = -ETIMEDOUT;
-+ goto err0;
-+ }
-+
-+ if (request_firmware(&tmu_fw, TMU_FIRMWARE_FILENAME, pfe->dev)) {
-+ pr_err("%s: request firmware %s failed\n", __func__,
-+ TMU_FIRMWARE_FILENAME);
-+ rc = -ETIMEDOUT;
-+ goto err1;
-+}
-+
-+#if !defined(CONFIG_FSL_PPFE_UTIL_DISABLED)
-+ util_fw_name = UTIL_FIRMWARE_FILENAME;
-+
-+ if (request_firmware(&util_fw, util_fw_name, pfe->dev)) {
-+ pr_err("%s: request firmware %s failed\n", __func__,
-+ util_fw_name);
-+ rc = -ETIMEDOUT;
-+ goto err2;
-+ }
-+#endif
-+ rc = pfe_load_elf(CLASS_MASK, class_fw, pfe);
-+ if (rc < 0) {
-+ pr_err("%s: class firmware load failed\n", __func__);
-+ goto err3;
-+ }
-+
-+#if defined(CFG_DIAGS)
-+ rc = pfe_get_diags_info(class_fw, &pfe->diags.class_diags_info);
-+ if (rc < 0) {
-+ pr_warn(
-+ "PFE diags won't be available for class PEs\n");
-+ rc = 0;
-+ }
-+#endif
-+
-+ rc = pfe_load_elf(TMU_MASK, tmu_fw, pfe);
-+ if (rc < 0) {
-+ pr_err("%s: tmu firmware load failed\n", __func__);
-+ goto err3;
-+ }
-+
-+#if !defined(CONFIG_FSL_PPFE_UTIL_DISABLED)
-+ rc = pfe_load_elf(UTIL_MASK, util_fw, pfe);
-+ if (rc < 0) {
-+ pr_err("%s: util firmware load failed\n", __func__);
-+ goto err3;
-+ }
-+
-+#if defined(CFG_DIAGS)
-+ rc = pfe_get_diags_info(util_fw, &pfe->diags.util_diags_info);
-+ if (rc < 0) {
-+ pr_warn(
-+ "PFE diags won't be available for util PE\n");
-+ rc = 0;
-+ }
-+#endif
-+
-+ util_enable();
-+#endif
-+
-+ tmu_enable(0xf);
-+ class_enable();
-+
-+err3:
-+#if !defined(CONFIG_FSL_PPFE_UTIL_DISABLED)
-+ release_firmware(util_fw);
-+
-+err2:
-+#endif
-+ release_firmware(tmu_fw);
-+
-+err1:
-+ release_firmware(class_fw);
-+
-+err0:
-+ return rc;
-+}
-+
-+/* PFE firmware cleanup
-+ * Puts PE's in reset
-+ *
-+ *
-+ */
-+void pfe_firmware_exit(struct pfe *pfe)
-+{
-+ pr_info("%s\n", __func__);
-+
-+ if (pe_reset_all(&pfe->ctrl) != 0)
-+ pr_err("Error: Failed to stop PEs, PFE reload may not work correctly\n");
-+
-+ class_disable();
-+ tmu_disable(0xf);
-+#if !defined(CONFIG_FSL_PPFE_UTIL_DISABLED)
-+ util_disable();
-+#endif
-+}
---- /dev/null
-+++ b/drivers/staging/fsl_ppfe/pfe_firmware.h
-@@ -0,0 +1,32 @@
-+/*
-+ * Copyright 2015-2016 Freescale Semiconductor, Inc.
-+ * Copyright 2017 NXP
-+ *
-+ * This program is free software; you can redistribute it and/or modify
-+ * it under the terms of the GNU General Public License as published by
-+ * the Free Software Foundation; either version 2 of the License, or
-+ * (at your option) any later version.
-+ *
-+ * This program is distributed in the hope that it will be useful,
-+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
-+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
-+ * GNU General Public License for more details.
-+ *
-+ * You should have received a copy of the GNU General Public License
-+ * along with this program. If not, see <http://www.gnu.org/licenses/>.
-+ */
-+
-+#ifndef _PFE_FIRMWARE_H_
-+#define _PFE_FIRMWARE_H_
-+
-+#define CLASS_FIRMWARE_FILENAME "ppfe_class_ls1012a.elf"
-+#define TMU_FIRMWARE_FILENAME "ppfe_tmu_ls1012a.elf"
-+
-+#define PFE_FW_CHECK_PASS 0
-+#define PFE_FW_CHECK_FAIL 1
-+#define NUM_PFE_FW 3
-+
-+int pfe_firmware_init(struct pfe *pfe);
-+void pfe_firmware_exit(struct pfe *pfe);
-+
-+#endif /* _PFE_FIRMWARE_H_ */
---- /dev/null
-+++ b/drivers/staging/fsl_ppfe/pfe_hal.c
-@@ -0,0 +1,1516 @@
-+/*
-+ * Copyright 2015-2016 Freescale Semiconductor, Inc.
-+ * Copyright 2017 NXP
-+ *
-+ * This program is free software; you can redistribute it and/or modify
-+ * it under the terms of the GNU General Public License as published by
-+ * the Free Software Foundation; either version 2 of the License, or
-+ * (at your option) any later version.
-+ *
-+ * This program is distributed in the hope that it will be useful,
-+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
-+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
-+ * GNU General Public License for more details.
-+ *
-+ * You should have received a copy of the GNU General Public License
-+ * along with this program. If not, see <http://www.gnu.org/licenses/>.
-+ */
-+
-+#include "pfe_mod.h"
-+#include "pfe/pfe.h"
-+
-+void *cbus_base_addr;
-+void *ddr_base_addr;
-+unsigned long ddr_phys_base_addr;
-+unsigned int ddr_size;
-+
-+static struct pe_info pe[MAX_PE];
-+
-+/* Initializes the PFE library.
-+ * Must be called before using any of the library functions.
-+ *
-+ * @param[in] cbus_base CBUS virtual base address (as mapped in
-+ * the host CPU address space)
-+ * @param[in] ddr_base PFE DDR range virtual base address (as
-+ * mapped in the host CPU address space)
-+ * @param[in] ddr_phys_base PFE DDR range physical base address (as
-+ * mapped in platform)
-+ * @param[in] size PFE DDR range size (as defined by the host
-+ * software)
-+ */
-+void pfe_lib_init(void *cbus_base, void *ddr_base, unsigned long ddr_phys_base,
-+ unsigned int size)
-+{
-+ cbus_base_addr = cbus_base;
-+ ddr_base_addr = ddr_base;
-+ ddr_phys_base_addr = ddr_phys_base;
-+ ddr_size = size;
-+
-+ pe[CLASS0_ID].dmem_base_addr = CLASS_DMEM_BASE_ADDR(0);
-+ pe[CLASS0_ID].pmem_base_addr = CLASS_IMEM_BASE_ADDR(0);
-+ pe[CLASS0_ID].pmem_size = CLASS_IMEM_SIZE;
-+ pe[CLASS0_ID].mem_access_wdata = CLASS_MEM_ACCESS_WDATA;
-+ pe[CLASS0_ID].mem_access_addr = CLASS_MEM_ACCESS_ADDR;
-+ pe[CLASS0_ID].mem_access_rdata = CLASS_MEM_ACCESS_RDATA;
-+
-+ pe[CLASS1_ID].dmem_base_addr = CLASS_DMEM_BASE_ADDR(1);
-+ pe[CLASS1_ID].pmem_base_addr = CLASS_IMEM_BASE_ADDR(1);
-+ pe[CLASS1_ID].pmem_size = CLASS_IMEM_SIZE;
-+ pe[CLASS1_ID].mem_access_wdata = CLASS_MEM_ACCESS_WDATA;
-+ pe[CLASS1_ID].mem_access_addr = CLASS_MEM_ACCESS_ADDR;
-+ pe[CLASS1_ID].mem_access_rdata = CLASS_MEM_ACCESS_RDATA;
-+
-+ pe[CLASS2_ID].dmem_base_addr = CLASS_DMEM_BASE_ADDR(2);
-+ pe[CLASS2_ID].pmem_base_addr = CLASS_IMEM_BASE_ADDR(2);
-+ pe[CLASS2_ID].pmem_size = CLASS_IMEM_SIZE;
-+ pe[CLASS2_ID].mem_access_wdata = CLASS_MEM_ACCESS_WDATA;
-+ pe[CLASS2_ID].mem_access_addr = CLASS_MEM_ACCESS_ADDR;
-+ pe[CLASS2_ID].mem_access_rdata = CLASS_MEM_ACCESS_RDATA;
-+
-+ pe[CLASS3_ID].dmem_base_addr = CLASS_DMEM_BASE_ADDR(3);
-+ pe[CLASS3_ID].pmem_base_addr = CLASS_IMEM_BASE_ADDR(3);
-+ pe[CLASS3_ID].pmem_size = CLASS_IMEM_SIZE;
-+ pe[CLASS3_ID].mem_access_wdata = CLASS_MEM_ACCESS_WDATA;
-+ pe[CLASS3_ID].mem_access_addr = CLASS_MEM_ACCESS_ADDR;
-+ pe[CLASS3_ID].mem_access_rdata = CLASS_MEM_ACCESS_RDATA;
-+
-+ pe[CLASS4_ID].dmem_base_addr = CLASS_DMEM_BASE_ADDR(4);
-+ pe[CLASS4_ID].pmem_base_addr = CLASS_IMEM_BASE_ADDR(4);
-+ pe[CLASS4_ID].pmem_size = CLASS_IMEM_SIZE;
-+ pe[CLASS4_ID].mem_access_wdata = CLASS_MEM_ACCESS_WDATA;
-+ pe[CLASS4_ID].mem_access_addr = CLASS_MEM_ACCESS_ADDR;
-+ pe[CLASS4_ID].mem_access_rdata = CLASS_MEM_ACCESS_RDATA;
-+
-+ pe[CLASS5_ID].dmem_base_addr = CLASS_DMEM_BASE_ADDR(5);
-+ pe[CLASS5_ID].pmem_base_addr = CLASS_IMEM_BASE_ADDR(5);
-+ pe[CLASS5_ID].pmem_size = CLASS_IMEM_SIZE;
-+ pe[CLASS5_ID].mem_access_wdata = CLASS_MEM_ACCESS_WDATA;
-+ pe[CLASS5_ID].mem_access_addr = CLASS_MEM_ACCESS_ADDR;
-+ pe[CLASS5_ID].mem_access_rdata = CLASS_MEM_ACCESS_RDATA;
-+
-+ pe[TMU0_ID].dmem_base_addr = TMU_DMEM_BASE_ADDR(0);
-+ pe[TMU0_ID].pmem_base_addr = TMU_IMEM_BASE_ADDR(0);
-+ pe[TMU0_ID].pmem_size = TMU_IMEM_SIZE;
-+ pe[TMU0_ID].mem_access_wdata = TMU_MEM_ACCESS_WDATA;
-+ pe[TMU0_ID].mem_access_addr = TMU_MEM_ACCESS_ADDR;
-+ pe[TMU0_ID].mem_access_rdata = TMU_MEM_ACCESS_RDATA;
-+
-+ pe[TMU1_ID].dmem_base_addr = TMU_DMEM_BASE_ADDR(1);
-+ pe[TMU1_ID].pmem_base_addr = TMU_IMEM_BASE_ADDR(1);
-+ pe[TMU1_ID].pmem_size = TMU_IMEM_SIZE;
-+ pe[TMU1_ID].mem_access_wdata = TMU_MEM_ACCESS_WDATA;
-+ pe[TMU1_ID].mem_access_addr = TMU_MEM_ACCESS_ADDR;
-+ pe[TMU1_ID].mem_access_rdata = TMU_MEM_ACCESS_RDATA;
-+
-+ pe[TMU3_ID].dmem_base_addr = TMU_DMEM_BASE_ADDR(3);
-+ pe[TMU3_ID].pmem_base_addr = TMU_IMEM_BASE_ADDR(3);
-+ pe[TMU3_ID].pmem_size = TMU_IMEM_SIZE;
-+ pe[TMU3_ID].mem_access_wdata = TMU_MEM_ACCESS_WDATA;
-+ pe[TMU3_ID].mem_access_addr = TMU_MEM_ACCESS_ADDR;
-+ pe[TMU3_ID].mem_access_rdata = TMU_MEM_ACCESS_RDATA;
-+
-+#if !defined(CONFIG_FSL_PPFE_UTIL_DISABLED)
-+ pe[UTIL_ID].dmem_base_addr = UTIL_DMEM_BASE_ADDR;
-+ pe[UTIL_ID].mem_access_wdata = UTIL_MEM_ACCESS_WDATA;
-+ pe[UTIL_ID].mem_access_addr = UTIL_MEM_ACCESS_ADDR;
-+ pe[UTIL_ID].mem_access_rdata = UTIL_MEM_ACCESS_RDATA;
-+#endif
-+}
-+
-+/* Writes a buffer to PE internal memory from the host
-+ * through indirect access registers.
-+ *
-+ * @param[in] id PE identification (CLASS0_ID, ..., TMU0_ID,
-+ * ..., UTIL_ID)
-+ * @param[in] src Buffer source address
-+ * @param[in] mem_access_addr DMEM destination address (must be 32bit
-+ * aligned)
-+ * @param[in] len Number of bytes to copy
-+ */
-+void pe_mem_memcpy_to32(int id, u32 mem_access_addr, const void *src, unsigned
-+int len)
-+{
-+ u32 offset = 0, val, addr;
-+ unsigned int len32 = len >> 2;
-+ int i;
-+
-+ addr = mem_access_addr | PE_MEM_ACCESS_WRITE |
-+ PE_MEM_ACCESS_BYTE_ENABLE(0, 4);
-+
-+ for (i = 0; i < len32; i++, offset += 4, src += 4) {
-+ val = *(u32 *)src;
-+ writel(cpu_to_be32(val), pe[id].mem_access_wdata);
-+ writel(addr + offset, pe[id].mem_access_addr);
-+ }
-+
-+ len = (len & 0x3);
-+ if (len) {
-+ val = 0;
-+
-+ addr = (mem_access_addr | PE_MEM_ACCESS_WRITE |
-+ PE_MEM_ACCESS_BYTE_ENABLE(0, len)) + offset;
-+
-+ for (i = 0; i < len; i++, src++)
-+ val |= (*(u8 *)src) << (8 * i);
-+
-+ writel(cpu_to_be32(val), pe[id].mem_access_wdata);
-+ writel(addr, pe[id].mem_access_addr);
-+ }
-+}
-+
-+/* Writes a buffer to PE internal data memory (DMEM) from the host
-+ * through indirect access registers.
-+ * @param[in] id PE identification (CLASS0_ID, ..., TMU0_ID,
-+ * ..., UTIL_ID)
-+ * @param[in] src Buffer source address
-+ * @param[in] dst DMEM destination address (must be 32bit
-+ * aligned)
-+ * @param[in] len Number of bytes to copy
-+ */
-+void pe_dmem_memcpy_to32(int id, u32 dst, const void *src, unsigned int len)
-+{
-+ pe_mem_memcpy_to32(id, pe[id].dmem_base_addr | dst |
-+ PE_MEM_ACCESS_DMEM, src, len);
-+}
-+
-+/* Writes a buffer to PE internal program memory (PMEM) from the host
-+ * through indirect access registers.
-+ * @param[in] id PE identification (CLASS0_ID, ..., TMU0_ID,
-+ * ..., TMU3_ID)
-+ * @param[in] src Buffer source address
-+ * @param[in] dst PMEM destination address (must be 32bit
-+ * aligned)
-+ * @param[in] len Number of bytes to copy
-+ */
-+void pe_pmem_memcpy_to32(int id, u32 dst, const void *src, unsigned int len)
-+{
-+ pe_mem_memcpy_to32(id, pe[id].pmem_base_addr | (dst & (pe[id].pmem_size
-+ - 1)) | PE_MEM_ACCESS_IMEM, src, len);
-+}
-+
-+/* Reads PE internal program memory (IMEM) from the host
-+ * through indirect access registers.
-+ * @param[in] id PE identification (CLASS0_ID, ..., TMU0_ID,
-+ * ..., TMU3_ID)
-+ * @param[in] addr PMEM read address (must be aligned on size)
-+ * @param[in] size Number of bytes to read (maximum 4, must not
-+ * cross 32bit boundaries)
-+ * @return the data read (in PE endianness, i.e BE).
-+ */
-+u32 pe_pmem_read(int id, u32 addr, u8 size)
-+{
-+ u32 offset = addr & 0x3;
-+ u32 mask = 0xffffffff >> ((4 - size) << 3);
-+ u32 val;
-+
-+ addr = pe[id].pmem_base_addr | ((addr & ~0x3) & (pe[id].pmem_size - 1))
-+ | PE_MEM_ACCESS_IMEM | PE_MEM_ACCESS_BYTE_ENABLE(offset, size);
-+
-+ writel(addr, pe[id].mem_access_addr);
-+ val = be32_to_cpu(readl(pe[id].mem_access_rdata));
-+
-+ return (val >> (offset << 3)) & mask;
-+}
-+
-+/* Writes PE internal data memory (DMEM) from the host
-+ * through indirect access registers.
-+ * @param[in] id PE identification (CLASS0_ID, ..., TMU0_ID,
-+ * ..., UTIL_ID)
-+ * @param[in] addr DMEM write address (must be aligned on size)
-+ * @param[in] val Value to write (in PE endianness, i.e BE)
-+ * @param[in] size Number of bytes to write (maximum 4, must not
-+ * cross 32bit boundaries)
-+ */
-+void pe_dmem_write(int id, u32 val, u32 addr, u8 size)
-+{
-+ u32 offset = addr & 0x3;
-+
-+ addr = pe[id].dmem_base_addr | (addr & ~0x3) | PE_MEM_ACCESS_WRITE |
-+ PE_MEM_ACCESS_DMEM | PE_MEM_ACCESS_BYTE_ENABLE(offset, size);
-+
-+ /* Indirect access interface is byte swapping data being written */
-+ writel(cpu_to_be32(val << (offset << 3)), pe[id].mem_access_wdata);
-+ writel(addr, pe[id].mem_access_addr);
-+}
-+
-+/* Reads PE internal data memory (DMEM) from the host
-+ * through indirect access registers.
-+ * @param[in] id PE identification (CLASS0_ID, ..., TMU0_ID,
-+ * ..., UTIL_ID)
-+ * @param[in] addr DMEM read address (must be aligned on size)
-+ * @param[in] size Number of bytes to read (maximum 4, must not
-+ * cross 32bit boundaries)
-+ * @return the data read (in PE endianness, i.e BE).
-+ */
-+u32 pe_dmem_read(int id, u32 addr, u8 size)
-+{
-+ u32 offset = addr & 0x3;
-+ u32 mask = 0xffffffff >> ((4 - size) << 3);
-+ u32 val;
-+
-+ addr = pe[id].dmem_base_addr | (addr & ~0x3) | PE_MEM_ACCESS_DMEM |
-+ PE_MEM_ACCESS_BYTE_ENABLE(offset, size);
-+
-+ writel(addr, pe[id].mem_access_addr);
-+
-+ /* Indirect access interface is byte swapping data being read */
-+ val = be32_to_cpu(readl(pe[id].mem_access_rdata));
-+
-+ return (val >> (offset << 3)) & mask;
-+}
-+
-+/* This function is used to write to CLASS internal bus peripherals (ccu,
-+ * pe-lem) from the host
-+ * through indirect access registers.
-+ * @param[in] val value to write
-+ * @param[in] addr Address to write to (must be aligned on size)
-+ * @param[in] size Number of bytes to write (1, 2 or 4)
-+ *
-+ */
-+void class_bus_write(u32 val, u32 addr, u8 size)
-+{
-+ u32 offset = addr & 0x3;
-+
-+ writel((addr & CLASS_BUS_ACCESS_BASE_MASK), CLASS_BUS_ACCESS_BASE);
-+
-+ addr = (addr & ~CLASS_BUS_ACCESS_BASE_MASK) | PE_MEM_ACCESS_WRITE |
-+ (size << 24);
-+
-+ writel(cpu_to_be32(val << (offset << 3)), CLASS_BUS_ACCESS_WDATA);
-+ writel(addr, CLASS_BUS_ACCESS_ADDR);
-+}
-+
-+/* Reads from CLASS internal bus peripherals (ccu, pe-lem) from the host
-+ * through indirect access registers.
-+ * @param[in] addr Address to read from (must be aligned on size)
-+ * @param[in] size Number of bytes to read (1, 2 or 4)
-+ * @return the read data
-+ *
-+ */
-+u32 class_bus_read(u32 addr, u8 size)
-+{
-+ u32 offset = addr & 0x3;
-+ u32 mask = 0xffffffff >> ((4 - size) << 3);
-+ u32 val;
-+
-+ writel((addr & CLASS_BUS_ACCESS_BASE_MASK), CLASS_BUS_ACCESS_BASE);
-+
-+ addr = (addr & ~CLASS_BUS_ACCESS_BASE_MASK) | (size << 24);
-+
-+ writel(addr, CLASS_BUS_ACCESS_ADDR);
-+ val = be32_to_cpu(readl(CLASS_BUS_ACCESS_RDATA));
-+
-+ return (val >> (offset << 3)) & mask;
-+}
-+
-+/* Writes data to the cluster memory (PE_LMEM)
-+ * @param[in] dst PE LMEM destination address (must be 32bit aligned)
-+ * @param[in] src Buffer source address
-+ * @param[in] len Number of bytes to copy
-+ */
-+void class_pe_lmem_memcpy_to32(u32 dst, const void *src, unsigned int len)
-+{
-+ u32 len32 = len >> 2;
-+ int i;
-+
-+ for (i = 0; i < len32; i++, src += 4, dst += 4)
-+ class_bus_write(*(u32 *)src, dst, 4);
-+
-+ if (len & 0x2) {
-+ class_bus_write(*(u16 *)src, dst, 2);
-+ src += 2;
-+ dst += 2;
-+ }
-+
-+ if (len & 0x1) {
-+ class_bus_write(*(u8 *)src, dst, 1);
-+ src++;
-+ dst++;
-+ }
-+}
-+
-+/* Writes value to the cluster memory (PE_LMEM)
-+ * @param[in] dst PE LMEM destination address (must be 32bit aligned)
-+ * @param[in] val Value to write
-+ * @param[in] len Number of bytes to write
-+ */
-+void class_pe_lmem_memset(u32 dst, int val, unsigned int len)
-+{
-+ u32 len32 = len >> 2;
-+ int i;
-+
-+ val = val | (val << 8) | (val << 16) | (val << 24);
-+
-+ for (i = 0; i < len32; i++, dst += 4)
-+ class_bus_write(val, dst, 4);
-+
-+ if (len & 0x2) {
-+ class_bus_write(val, dst, 2);
-+ dst += 2;
-+ }
-+
-+ if (len & 0x1) {
-+ class_bus_write(val, dst, 1);
-+ dst++;
-+ }
-+}
-+
-+#if !defined(CONFIG_FSL_PPFE_UTIL_DISABLED)
-+
-+/* Writes UTIL program memory (DDR) from the host.
-+ *
-+ * @param[in] addr Address to write (virtual, must be aligned on size)
-+ * @param[in] val Value to write (in PE endianness, i.e BE)
-+ * @param[in] size Number of bytes to write (2 or 4)
-+ */
-+static void util_pmem_write(u32 val, void *addr, u8 size)
-+{
-+ void *addr64 = (void *)((unsigned long)addr & ~0x7);
-+ unsigned long off = 8 - ((unsigned long)addr & 0x7) - size;
-+
-+ /*
-+ * IMEM should be loaded as a 64bit swapped value in a 64bit aligned
-+ * location
-+ */
-+ if (size == 4)
-+ writel(be32_to_cpu(val), addr64 + off);
-+ else
-+ writew(be16_to_cpu((u16)val), addr64 + off);
-+}
-+
-+/* Writes a buffer to UTIL program memory (DDR) from the host.
-+ *
-+ * @param[in] dst Address to write (virtual, must be at least 16bit
-+ * aligned)
-+ * @param[in] src Buffer to write (in PE endianness, i.e BE, must have
-+ * same alignment as dst)
-+ * @param[in] len Number of bytes to write (must be at least 16bit
-+ * aligned)
-+ */
-+static void util_pmem_memcpy(void *dst, const void *src, unsigned int len)
-+{
-+ unsigned int len32;
-+ int i;
-+
-+ if ((unsigned long)src & 0x2) {
-+ util_pmem_write(*(u16 *)src, dst, 2);
-+ src += 2;
-+ dst += 2;
-+ len -= 2;
-+ }
-+
-+ len32 = len >> 2;
-+
-+ for (i = 0; i < len32; i++, dst += 4, src += 4)
-+ util_pmem_write(*(u32 *)src, dst, 4);
-+
-+ if (len & 0x2)
-+ util_pmem_write(*(u16 *)src, dst, len & 0x2);
-+}
-+#endif
-+
-+/* Loads an elf section into pmem
-+ * Code needs to be at least 16bit aligned and only PROGBITS sections are
-+ * supported
-+ *
-+ * @param[in] id PE identification (CLASS0_ID, ..., TMU0_ID, ...,
-+ * TMU3_ID)
-+ * @param[in] data pointer to the elf firmware
-+ * @param[in] shdr pointer to the elf section header
-+ *
-+ */
-+static int pe_load_pmem_section(int id, const void *data,
-+ struct elf32_shdr *shdr)
-+{
-+ u32 offset = be32_to_cpu(shdr->sh_offset);
-+ u32 addr = be32_to_cpu(shdr->sh_addr);
-+ u32 size = be32_to_cpu(shdr->sh_size);
-+ u32 type = be32_to_cpu(shdr->sh_type);
-+
-+#if !defined(CONFIG_FSL_PPFE_UTIL_DISABLED)
-+ if (id == UTIL_ID) {
-+ pr_err("%s: unsupported pmem section for UTIL\n",
-+ __func__);
-+ return -EINVAL;
-+ }
-+#endif
-+
-+ if (((unsigned long)(data + offset) & 0x3) != (addr & 0x3)) {
-+ pr_err(
-+ "%s: load address(%x) and elf file address(%lx) don't have the same alignment\n"
-+ , __func__, addr, (unsigned long)data + offset);
-+
-+ return -EINVAL;
-+ }
-+
-+ if (addr & 0x1) {
-+ pr_err("%s: load address(%x) is not 16bit aligned\n",
-+ __func__, addr);
-+ return -EINVAL;
-+ }
-+
-+ if (size & 0x1) {
-+ pr_err("%s: load size(%x) is not 16bit aligned\n",
-+ __func__, size);
-+ return -EINVAL;
-+ }
-+
-+ switch (type) {
-+ case SHT_PROGBITS:
-+ pe_pmem_memcpy_to32(id, addr, data + offset, size);
-+
-+ break;
-+
-+ default:
-+ pr_err("%s: unsupported section type(%x)\n", __func__,
-+ type);
-+ return -EINVAL;
-+ }
-+
-+ return 0;
-+}
-+
-+/* Loads an elf section into dmem
-+ * Data needs to be at least 32bit aligned, NOBITS sections are correctly
-+ * initialized to 0
-+ *
-+ * @param[in] id PE identification (CLASS0_ID, ..., TMU0_ID,
-+ * ..., UTIL_ID)
-+ * @param[in] data pointer to the elf firmware
-+ * @param[in] shdr pointer to the elf section header
-+ *
-+ */
-+static int pe_load_dmem_section(int id, const void *data,
-+ struct elf32_shdr *shdr)
-+{
-+ u32 offset = be32_to_cpu(shdr->sh_offset);
-+ u32 addr = be32_to_cpu(shdr->sh_addr);
-+ u32 size = be32_to_cpu(shdr->sh_size);
-+ u32 type = be32_to_cpu(shdr->sh_type);
-+ u32 size32 = size >> 2;
-+ int i;
-+
-+ if (((unsigned long)(data + offset) & 0x3) != (addr & 0x3)) {
-+ pr_err(
-+ "%s: load address(%x) and elf file address(%lx) don't have the same alignment\n",
-+ __func__, addr, (unsigned long)data + offset);
-+
-+ return -EINVAL;
-+ }
-+
-+ if (addr & 0x3) {
-+ pr_err("%s: load address(%x) is not 32bit aligned\n",
-+ __func__, addr);
-+ return -EINVAL;
-+ }
-+
-+ switch (type) {
-+ case SHT_PROGBITS:
-+ pe_dmem_memcpy_to32(id, addr, data + offset, size);
-+ break;
-+
-+ case SHT_NOBITS:
-+ for (i = 0; i < size32; i++, addr += 4)
-+ pe_dmem_write(id, 0, addr, 4);
-+
-+ if (size & 0x3)
-+ pe_dmem_write(id, 0, addr, size & 0x3);
-+
-+ break;
-+
-+ default:
-+ pr_err("%s: unsupported section type(%x)\n", __func__,
-+ type);
-+ return -EINVAL;
-+ }
-+
-+ return 0;
-+}
-+
-+/* Loads an elf section into DDR
-+ * Data needs to be at least 32bit aligned, NOBITS sections are correctly
-+ * initialized to 0
-+ *
-+ * @param[in] id PE identification (CLASS0_ID, ..., TMU0_ID,
-+ * ..., UTIL_ID)
-+ * @param[in] data pointer to the elf firmware
-+ * @param[in] shdr pointer to the elf section header
-+ *
-+ */
-+static int pe_load_ddr_section(int id, const void *data,
-+ struct elf32_shdr *shdr,
-+ struct device *dev) {
-+ u32 offset = be32_to_cpu(shdr->sh_offset);
-+ u32 addr = be32_to_cpu(shdr->sh_addr);
-+ u32 size = be32_to_cpu(shdr->sh_size);
-+ u32 type = be32_to_cpu(shdr->sh_type);
-+ u32 flags = be32_to_cpu(shdr->sh_flags);
-+
-+ switch (type) {
-+ case SHT_PROGBITS:
-+ if (flags & SHF_EXECINSTR) {
-+ if (id <= CLASS_MAX_ID) {
-+ /* DO the loading only once in DDR */
-+ if (id == CLASS0_ID) {
-+ pr_err(
-+ "%s: load address(%x) and elf file address(%lx) rcvd\n",
-+ __func__, addr,
-+ (unsigned long)data + offset);
-+ if (((unsigned long)(data + offset)
-+ & 0x3) != (addr & 0x3)) {
-+ pr_err(
-+ "%s: load address(%x) and elf file address(%lx) don't have the same alignment\n"
-+ , __func__, addr,
-+ (unsigned long)data + offset);
-+
-+ return -EINVAL;
-+ }
-+
-+ if (addr & 0x1) {
-+ pr_err(
-+ "%s: load address(%x) is not 16bit aligned\n"
-+ , __func__, addr);
-+ return -EINVAL;
-+ }
-+
-+ if (size & 0x1) {
-+ pr_err(
-+ "%s: load length(%x) is not 16bit aligned\n"
-+ , __func__, size);
-+ return -EINVAL;
-+ }
-+ memcpy(DDR_PHYS_TO_VIRT(
-+ DDR_PFE_TO_PHYS(addr)),
-+ data + offset, size);
-+ }
-+#if !defined(CONFIG_FSL_PPFE_UTIL_DISABLED)
-+ } else if (id == UTIL_ID) {
-+ if (((unsigned long)(data + offset) & 0x3)
-+ != (addr & 0x3)) {
-+ pr_err(
-+ "%s: load address(%x) and elf file address(%lx) don't have the same alignment\n"
-+ , __func__, addr,
-+ (unsigned long)data + offset);
-+
-+ return -EINVAL;
-+ }
-+
-+ if (addr & 0x1) {
-+ pr_err(
-+ "%s: load address(%x) is not 16bit aligned\n"
-+ , __func__, addr);
-+ return -EINVAL;
-+ }
-+
-+ if (size & 0x1) {
-+ pr_err(
-+ "%s: load length(%x) is not 16bit aligned\n"
-+ , __func__, size);
-+ return -EINVAL;
-+ }
-+
-+ util_pmem_memcpy(DDR_PHYS_TO_VIRT(
-+ DDR_PFE_TO_PHYS(addr)),
-+ data + offset, size);
-+ }
-+#endif
-+ } else {
-+ pr_err(
-+ "%s: unsupported ddr section type(%x) for PE(%d)\n"
-+ , __func__, type, id);
-+ return -EINVAL;
-+ }
-+
-+ } else {
-+ memcpy(DDR_PHYS_TO_VIRT(DDR_PFE_TO_PHYS(addr)), data
-+ + offset, size);
-+ }
-+
-+ break;
-+
-+ case SHT_NOBITS:
-+ memset(DDR_PHYS_TO_VIRT(DDR_PFE_TO_PHYS(addr)), 0, size);
-+
-+ break;
-+
-+ default:
-+ pr_err("%s: unsupported section type(%x)\n", __func__,
-+ type);
-+ return -EINVAL;
-+ }
-+
-+ return 0;
-+}
-+
-+/* Loads an elf section into pe lmem
-+ * Data needs to be at least 32bit aligned, NOBITS sections are correctly
-+ * initialized to 0
-+ *
-+ * @param[in] id PE identification (CLASS0_ID,..., CLASS5_ID)
-+ * @param[in] data pointer to the elf firmware
-+ * @param[in] shdr pointer to the elf section header
-+ *
-+ */
-+static int pe_load_pe_lmem_section(int id, const void *data,
-+ struct elf32_shdr *shdr)
-+{
-+ u32 offset = be32_to_cpu(shdr->sh_offset);
-+ u32 addr = be32_to_cpu(shdr->sh_addr);
-+ u32 size = be32_to_cpu(shdr->sh_size);
-+ u32 type = be32_to_cpu(shdr->sh_type);
-+
-+ if (id > CLASS_MAX_ID) {
-+ pr_err(
-+ "%s: unsupported pe-lmem section type(%x) for PE(%d)\n",
-+ __func__, type, id);
-+ return -EINVAL;
-+ }
-+
-+ if (((unsigned long)(data + offset) & 0x3) != (addr & 0x3)) {
-+ pr_err(
-+ "%s: load address(%x) and elf file address(%lx) don't have the same alignment\n",
-+ __func__, addr, (unsigned long)data + offset);
-+
-+ return -EINVAL;
-+ }
-+
-+ if (addr & 0x3) {
-+ pr_err("%s: load address(%x) is not 32bit aligned\n",
-+ __func__, addr);
-+ return -EINVAL;
-+ }
-+
-+ switch (type) {
-+ case SHT_PROGBITS:
-+ class_pe_lmem_memcpy_to32(addr, data + offset, size);
-+ break;
-+
-+ case SHT_NOBITS:
-+ class_pe_lmem_memset(addr, 0, size);
-+ break;
-+
-+ default:
-+ pr_err("%s: unsupported section type(%x)\n", __func__,
-+ type);
-+ return -EINVAL;
-+ }
-+
-+ return 0;
-+}
-+
-+/* Loads an elf section into a PE
-+ * For now only supports loading a section to dmem (all PE's), pmem (class and
-+ * tmu PE's),
-+ * DDDR (util PE code)
-+ *
-+ * @param[in] id PE identification (CLASS0_ID, ..., TMU0_ID,
-+ * ..., UTIL_ID)
-+ * @param[in] data pointer to the elf firmware
-+ * @param[in] shdr pointer to the elf section header
-+ *
-+ */
-+int pe_load_elf_section(int id, const void *data, struct elf32_shdr *shdr,
-+ struct device *dev) {
-+ u32 addr = be32_to_cpu(shdr->sh_addr);
-+ u32 size = be32_to_cpu(shdr->sh_size);
-+
-+ if (IS_DMEM(addr, size))
-+ return pe_load_dmem_section(id, data, shdr);
-+ else if (IS_PMEM(addr, size))
-+ return pe_load_pmem_section(id, data, shdr);
-+ else if (IS_PFE_LMEM(addr, size))
-+ return 0;
-+ else if (IS_PHYS_DDR(addr, size))
-+ return pe_load_ddr_section(id, data, shdr, dev);
-+ else if (IS_PE_LMEM(addr, size))
-+ return pe_load_pe_lmem_section(id, data, shdr);
-+
-+ pr_err("%s: unsupported memory range(%x)\n", __func__,
-+ addr);
-+ return 0;
-+}
-+
-+/**************************** BMU ***************************/
-+
-+/* Initializes a BMU block.
-+ * @param[in] base BMU block base address
-+ * @param[in] cfg BMU configuration
-+ */
-+void bmu_init(void *base, struct BMU_CFG *cfg)
-+{
-+ bmu_disable(base);
-+
-+ bmu_set_config(base, cfg);
-+
-+ bmu_reset(base);
-+}
-+
-+/* Resets a BMU block.
-+ * @param[in] base BMU block base address
-+ */
-+void bmu_reset(void *base)
-+{
-+ writel(CORE_SW_RESET, base + BMU_CTRL);
-+
-+ /* Wait for self clear */
-+ while (readl(base + BMU_CTRL) & CORE_SW_RESET)
-+ ;
-+}
-+
-+/* Enabled a BMU block.
-+ * @param[in] base BMU block base address
-+ */
-+void bmu_enable(void *base)
-+{
-+ writel(CORE_ENABLE, base + BMU_CTRL);
-+}
-+
-+/* Disables a BMU block.
-+ * @param[in] base BMU block base address
-+ */
-+void bmu_disable(void *base)
-+{
-+ writel(CORE_DISABLE, base + BMU_CTRL);
-+}
-+
-+/* Sets the configuration of a BMU block.
-+ * @param[in] base BMU block base address
-+ * @param[in] cfg BMU configuration
-+ */
-+void bmu_set_config(void *base, struct BMU_CFG *cfg)
-+{
-+ writel(cfg->baseaddr, base + BMU_UCAST_BASE_ADDR);
-+ writel(cfg->count & 0xffff, base + BMU_UCAST_CONFIG);
-+ writel(cfg->size & 0xffff, base + BMU_BUF_SIZE);
-+
-+ /* Interrupts are never used */
-+ writel(cfg->low_watermark, base + BMU_LOW_WATERMARK);
-+ writel(cfg->high_watermark, base + BMU_HIGH_WATERMARK);
-+ writel(0x0, base + BMU_INT_ENABLE);
-+}
-+
-+/**************************** MTIP GEMAC ***************************/
-+
-+/* Enable Rx Checksum Engine. With this enabled, Frame with bad IP,
-+ * TCP or UDP checksums are discarded
-+ *
-+ * @param[in] base GEMAC base address.
-+ */
-+void gemac_enable_rx_checksum_offload(void *base)
-+{
-+ /*Do not find configuration to do this */
-+}
-+
-+/* Disable Rx Checksum Engine.
-+ *
-+ * @param[in] base GEMAC base address.
-+ */
-+void gemac_disable_rx_checksum_offload(void *base)
-+{
-+ /*Do not find configuration to do this */
-+}
-+
-+/* GEMAC set speed.
-+ * @param[in] base GEMAC base address
-+ * @param[in] speed GEMAC speed (10, 100 or 1000 Mbps)
-+ */
-+void gemac_set_speed(void *base, enum mac_speed gem_speed)
-+{
-+ u32 ecr = readl(base + EMAC_ECNTRL_REG) & ~EMAC_ECNTRL_SPEED;
-+ u32 rcr = readl(base + EMAC_RCNTRL_REG) & ~EMAC_RCNTRL_RMII_10T;
-+
-+ switch (gem_speed) {
-+ case SPEED_10M:
-+ rcr |= EMAC_RCNTRL_RMII_10T;
-+ break;
-+
-+ case SPEED_1000M:
-+ ecr |= EMAC_ECNTRL_SPEED;
-+ break;
-+
-+ case SPEED_100M:
-+ default:
-+ /*It is in 100M mode */
-+ break;
-+ }
-+ writel(ecr, (base + EMAC_ECNTRL_REG));
-+ writel(rcr, (base + EMAC_RCNTRL_REG));
-+}
-+
-+/* GEMAC set duplex.
-+ * @param[in] base GEMAC base address
-+ * @param[in] duplex GEMAC duplex mode (Full, Half)
-+ */
-+void gemac_set_duplex(void *base, int duplex)
-+{
-+ if (duplex == DUPLEX_HALF) {
-+ writel(readl(base + EMAC_TCNTRL_REG) & ~EMAC_TCNTRL_FDEN, base
-+ + EMAC_TCNTRL_REG);
-+ writel(readl(base + EMAC_RCNTRL_REG) | EMAC_RCNTRL_DRT, (base
-+ + EMAC_RCNTRL_REG));
-+ } else{
-+ writel(readl(base + EMAC_TCNTRL_REG) | EMAC_TCNTRL_FDEN, base
-+ + EMAC_TCNTRL_REG);
-+ writel(readl(base + EMAC_RCNTRL_REG) & ~EMAC_RCNTRL_DRT, (base
-+ + EMAC_RCNTRL_REG));
-+ }
-+}
-+
-+/* GEMAC set mode.
-+ * @param[in] base GEMAC base address
-+ * @param[in] mode GEMAC operation mode (MII, RMII, RGMII, SGMII)
-+ */
-+void gemac_set_mode(void *base, int mode)
-+{
-+ u32 val = readl(base + EMAC_RCNTRL_REG);
-+
-+ /*Remove loopbank*/
-+ val &= ~EMAC_RCNTRL_LOOP;
-+
-+ /*Enable flow control and MII mode*/
-+ val |= (EMAC_RCNTRL_FCE | EMAC_RCNTRL_MII_MODE);
-+
-+ writel(val, base + EMAC_RCNTRL_REG);
-+}
-+
-+/* GEMAC enable function.
-+ * @param[in] base GEMAC base address
-+ */
-+void gemac_enable(void *base)
-+{
-+ writel(readl(base + EMAC_ECNTRL_REG) | EMAC_ECNTRL_ETHER_EN, base +
-+ EMAC_ECNTRL_REG);
-+}
-+
-+/* GEMAC disable function.
-+ * @param[in] base GEMAC base address
-+ */
-+void gemac_disable(void *base)
-+{
-+ writel(readl(base + EMAC_ECNTRL_REG) & ~EMAC_ECNTRL_ETHER_EN, base +
-+ EMAC_ECNTRL_REG);
-+}
-+
-+/* GEMAC TX disable function.
-+ * @param[in] base GEMAC base address
-+ */
-+void gemac_tx_disable(void *base)
-+{
-+ writel(readl(base + EMAC_TCNTRL_REG) | EMAC_TCNTRL_GTS, base +
-+ EMAC_TCNTRL_REG);
-+}
-+
-+void gemac_tx_enable(void *base)
-+{
-+ writel(readl(base + EMAC_TCNTRL_REG) & ~EMAC_TCNTRL_GTS, base +
-+ EMAC_TCNTRL_REG);
-+}
-+
-+/* Sets the hash register of the MAC.
-+ * This register is used for matching unicast and multicast frames.
-+ *
-+ * @param[in] base GEMAC base address.
-+ * @param[in] hash 64-bit hash to be configured.
-+ */
-+void gemac_set_hash(void *base, struct pfe_mac_addr *hash)
-+{
-+ writel(hash->bottom, base + EMAC_GALR);
-+ writel(hash->top, base + EMAC_GAUR);
-+}
-+
-+void gemac_set_laddrN(void *base, struct pfe_mac_addr *address,
-+ unsigned int entry_index)
-+{
-+ if ((entry_index < 1) || (entry_index > EMAC_SPEC_ADDR_MAX))
-+ return;
-+
-+ entry_index = entry_index - 1;
-+ if (entry_index < 1) {
-+ writel(htonl(address->bottom), base + EMAC_PHY_ADDR_LOW);
-+ writel((htonl(address->top) | 0x8808), base +
-+ EMAC_PHY_ADDR_HIGH);
-+ } else {
-+ writel(htonl(address->bottom), base + ((entry_index - 1) * 8)
-+ + EMAC_SMAC_0_0);
-+ writel((htonl(address->top) | 0x8808), base + ((entry_index -
-+ 1) * 8) + EMAC_SMAC_0_1);
-+ }
-+}
-+
-+void gemac_clear_laddrN(void *base, unsigned int entry_index)
-+{
-+ if ((entry_index < 1) || (entry_index > EMAC_SPEC_ADDR_MAX))
-+ return;
-+
-+ entry_index = entry_index - 1;
-+ if (entry_index < 1) {
-+ writel(0, base + EMAC_PHY_ADDR_LOW);
-+ writel(0, base + EMAC_PHY_ADDR_HIGH);
-+ } else {
-+ writel(0, base + ((entry_index - 1) * 8) + EMAC_SMAC_0_0);
-+ writel(0, base + ((entry_index - 1) * 8) + EMAC_SMAC_0_1);
-+ }
-+}
-+
-+/* Set the loopback mode of the MAC. This can be either no loopback for
-+ * normal operation, local loopback through MAC internal loopback module or PHY
-+ * loopback for external loopback through a PHY. This asserts the external
-+ * loop pin.
-+ *
-+ * @param[in] base GEMAC base address.
-+ * @param[in] gem_loop Loopback mode to be enabled. LB_LOCAL - MAC
-+ * Loopback,
-+ * LB_EXT - PHY Loopback.
-+ */
-+void gemac_set_loop(void *base, enum mac_loop gem_loop)
-+{
-+ pr_info("%s()\n", __func__);
-+ writel(readl(base + EMAC_RCNTRL_REG) | EMAC_RCNTRL_LOOP, (base +
-+ EMAC_RCNTRL_REG));
-+}
-+
-+/* GEMAC allow frames
-+ * @param[in] base GEMAC base address
-+ */
-+void gemac_enable_copy_all(void *base)
-+{
-+ writel(readl(base + EMAC_RCNTRL_REG) | EMAC_RCNTRL_PROM, (base +
-+ EMAC_RCNTRL_REG));
-+}
-+
-+/* GEMAC do not allow frames
-+ * @param[in] base GEMAC base address
-+ */
-+void gemac_disable_copy_all(void *base)
-+{
-+ writel(readl(base + EMAC_RCNTRL_REG) & ~EMAC_RCNTRL_PROM, (base +
-+ EMAC_RCNTRL_REG));
-+}
-+
-+/* GEMAC allow broadcast function.
-+ * @param[in] base GEMAC base address
-+ */
-+void gemac_allow_broadcast(void *base)
-+{
-+ writel(readl(base + EMAC_RCNTRL_REG) & ~EMAC_RCNTRL_BC_REJ, base +
-+ EMAC_RCNTRL_REG);
-+}
-+
-+/* GEMAC no broadcast function.
-+ * @param[in] base GEMAC base address
-+ */
-+void gemac_no_broadcast(void *base)
-+{
-+ writel(readl(base + EMAC_RCNTRL_REG) | EMAC_RCNTRL_BC_REJ, base +
-+ EMAC_RCNTRL_REG);
-+}
-+
-+/* GEMAC enable 1536 rx function.
-+ * @param[in] base GEMAC base address
-+ */
-+void gemac_enable_1536_rx(void *base)
-+{
-+ /* Set 1536 as Maximum frame length */
-+ writel(readl(base + EMAC_RCNTRL_REG) | (1536 << 16), base +
-+ EMAC_RCNTRL_REG);
-+}
-+
-+/* GEMAC enable jumbo function.
-+ * @param[in] base GEMAC base address
-+ */
-+void gemac_enable_rx_jmb(void *base)
-+{
-+ writel(readl(base + EMAC_RCNTRL_REG) | (JUMBO_FRAME_SIZE << 16), base
-+ + EMAC_RCNTRL_REG);
-+}
-+
-+/* GEMAC enable stacked vlan function.
-+ * @param[in] base GEMAC base address
-+ */
-+void gemac_enable_stacked_vlan(void *base)
-+{
-+ /* MTIP doesn't support stacked vlan */
-+}
-+
-+/* GEMAC enable pause rx function.
-+ * @param[in] base GEMAC base address
-+ */
-+void gemac_enable_pause_rx(void *base)
-+{
-+ writel(readl(base + EMAC_RCNTRL_REG) | EMAC_RCNTRL_FCE,
-+ base + EMAC_RCNTRL_REG);
-+}
-+
-+/* GEMAC disable pause rx function.
-+ * @param[in] base GEMAC base address
-+ */
-+void gemac_disable_pause_rx(void *base)
-+{
-+ writel(readl(base + EMAC_RCNTRL_REG) & ~EMAC_RCNTRL_FCE,
-+ base + EMAC_RCNTRL_REG);
-+}
-+
-+/* GEMAC enable pause tx function.
-+ * @param[in] base GEMAC base address
-+ */
-+void gemac_enable_pause_tx(void *base)
-+{
-+ writel(EMAC_RX_SECTION_EMPTY_V, base + EMAC_RX_SECTION_EMPTY);
-+}
-+
-+/* GEMAC disable pause tx function.
-+ * @param[in] base GEMAC base address
-+ */
-+void gemac_disable_pause_tx(void *base)
-+{
-+ writel(0x0, base + EMAC_RX_SECTION_EMPTY);
-+}
-+
-+/* GEMAC wol configuration
-+ * @param[in] base GEMAC base address
-+ * @param[in] wol_conf WoL register configuration
-+ */
-+void gemac_set_wol(void *base, u32 wol_conf)
-+{
-+ u32 val = readl(base + EMAC_ECNTRL_REG);
-+
-+ if (wol_conf)
-+ val |= (EMAC_ECNTRL_MAGIC_ENA | EMAC_ECNTRL_SLEEP);
-+ else
-+ val &= ~(EMAC_ECNTRL_MAGIC_ENA | EMAC_ECNTRL_SLEEP);
-+ writel(val, base + EMAC_ECNTRL_REG);
-+}
-+
-+/* Sets Gemac bus width to 64bit
-+ * @param[in] base GEMAC base address
-+ * @param[in] width gemac bus width to be set possible values are 32/64/128
-+ */
-+void gemac_set_bus_width(void *base, int width)
-+{
-+}
-+
-+/* Sets Gemac configuration.
-+ * @param[in] base GEMAC base address
-+ * @param[in] cfg GEMAC configuration
-+ */
-+void gemac_set_config(void *base, struct gemac_cfg *cfg)
-+{
-+ /*GEMAC config taken from VLSI */
-+ writel(0x00000004, base + EMAC_TFWR_STR_FWD);
-+ writel(0x00000005, base + EMAC_RX_SECTION_FULL);
-+ writel(0x00003fff, base + EMAC_TRUNC_FL);
-+ writel(0x00000030, base + EMAC_TX_SECTION_EMPTY);
-+ writel(0x00000000, base + EMAC_MIB_CTRL_STS_REG);
-+
-+ gemac_set_mode(base, cfg->mode);
-+
-+ gemac_set_speed(base, cfg->speed);
-+
-+ gemac_set_duplex(base, cfg->duplex);
-+}
-+
-+/**************************** GPI ***************************/
-+
-+/* Initializes a GPI block.
-+ * @param[in] base GPI base address
-+ * @param[in] cfg GPI configuration
-+ */
-+void gpi_init(void *base, struct gpi_cfg *cfg)
-+{
-+ gpi_reset(base);
-+
-+ gpi_disable(base);
-+
-+ gpi_set_config(base, cfg);
-+}
-+
-+/* Resets a GPI block.
-+ * @param[in] base GPI base address
-+ */
-+void gpi_reset(void *base)
-+{
-+ writel(CORE_SW_RESET, base + GPI_CTRL);
-+}
-+
-+/* Enables a GPI block.
-+ * @param[in] base GPI base address
-+ */
-+void gpi_enable(void *base)
-+{
-+ writel(CORE_ENABLE, base + GPI_CTRL);
-+}
-+
-+/* Disables a GPI block.
-+ * @param[in] base GPI base address
-+ */
-+void gpi_disable(void *base)
-+{
-+ writel(CORE_DISABLE, base + GPI_CTRL);
-+}
-+
-+/* Sets the configuration of a GPI block.
-+ * @param[in] base GPI base address
-+ * @param[in] cfg GPI configuration
-+ */
-+void gpi_set_config(void *base, struct gpi_cfg *cfg)
-+{
-+ writel(CBUS_VIRT_TO_PFE(BMU1_BASE_ADDR + BMU_ALLOC_CTRL), base
-+ + GPI_LMEM_ALLOC_ADDR);
-+ writel(CBUS_VIRT_TO_PFE(BMU1_BASE_ADDR + BMU_FREE_CTRL), base
-+ + GPI_LMEM_FREE_ADDR);
-+ writel(CBUS_VIRT_TO_PFE(BMU2_BASE_ADDR + BMU_ALLOC_CTRL), base
-+ + GPI_DDR_ALLOC_ADDR);
-+ writel(CBUS_VIRT_TO_PFE(BMU2_BASE_ADDR + BMU_FREE_CTRL), base
-+ + GPI_DDR_FREE_ADDR);
-+ writel(CBUS_VIRT_TO_PFE(CLASS_INQ_PKTPTR), base + GPI_CLASS_ADDR);
-+ writel(DDR_HDR_SIZE, base + GPI_DDR_DATA_OFFSET);
-+ writel(LMEM_HDR_SIZE, base + GPI_LMEM_DATA_OFFSET);
-+ writel(0, base + GPI_LMEM_SEC_BUF_DATA_OFFSET);
-+ writel(0, base + GPI_DDR_SEC_BUF_DATA_OFFSET);
-+ writel((DDR_HDR_SIZE << 16) | LMEM_HDR_SIZE, base + GPI_HDR_SIZE);
-+ writel((DDR_BUF_SIZE << 16) | LMEM_BUF_SIZE, base + GPI_BUF_SIZE);
-+
-+ writel(((cfg->lmem_rtry_cnt << 16) | (GPI_DDR_BUF_EN << 1) |
-+ GPI_LMEM_BUF_EN), base + GPI_RX_CONFIG);
-+ writel(cfg->tmlf_txthres, base + GPI_TMLF_TX);
-+ writel(cfg->aseq_len, base + GPI_DTX_ASEQ);
-+ writel(1, base + GPI_TOE_CHKSUM_EN);
-+
-+ if (cfg->mtip_pause_reg) {
-+ writel(cfg->mtip_pause_reg, base + GPI_CSR_MTIP_PAUSE_REG);
-+ writel(EGPI_PAUSE_TIME, base + GPI_TX_PAUSE_TIME);
-+ }
-+}
-+
-+/**************************** CLASSIFIER ***************************/
-+
-+/* Initializes CLASSIFIER block.
-+ * @param[in] cfg CLASSIFIER configuration
-+ */
-+void class_init(struct class_cfg *cfg)
-+{
-+ class_reset();
-+
-+ class_disable();
-+
-+ class_set_config(cfg);
-+}
-+
-+/* Resets CLASSIFIER block.
-+ *
-+ */
-+void class_reset(void)
-+{
-+ writel(CORE_SW_RESET, CLASS_TX_CTRL);
-+}
-+
-+/* Enables all CLASS-PE's cores.
-+ *
-+ */
-+void class_enable(void)
-+{
-+ writel(CORE_ENABLE, CLASS_TX_CTRL);
-+}
-+
-+/* Disables all CLASS-PE's cores.
-+ *
-+ */
-+void class_disable(void)
-+{
-+ writel(CORE_DISABLE, CLASS_TX_CTRL);
-+}
-+
-+/*
-+ * Sets the configuration of the CLASSIFIER block.
-+ * @param[in] cfg CLASSIFIER configuration
-+ */
-+void class_set_config(struct class_cfg *cfg)
-+{
-+ u32 val;
-+
-+ /* Initialize route table */
-+ if (!cfg->resume)
-+ memset(DDR_PHYS_TO_VIRT(cfg->route_table_baseaddr), 0, (1 <<
-+ cfg->route_table_hash_bits) * CLASS_ROUTE_SIZE);
-+
-+#if !defined(LS1012A_PFE_RESET_WA)
-+ writel(cfg->pe_sys_clk_ratio, CLASS_PE_SYS_CLK_RATIO);
-+#endif
-+
-+ writel((DDR_HDR_SIZE << 16) | LMEM_HDR_SIZE, CLASS_HDR_SIZE);
-+ writel(LMEM_BUF_SIZE, CLASS_LMEM_BUF_SIZE);
-+ writel(CLASS_ROUTE_ENTRY_SIZE(CLASS_ROUTE_SIZE) |
-+ CLASS_ROUTE_HASH_SIZE(cfg->route_table_hash_bits),
-+ CLASS_ROUTE_HASH_ENTRY_SIZE);
-+ writel(HIF_PKT_CLASS_EN | HIF_PKT_OFFSET(sizeof(struct hif_hdr)),
-+ CLASS_HIF_PARSE);
-+
-+ val = HASH_CRC_PORT_IP | QB2BUS_LE;
-+
-+#if defined(CONFIG_IP_ALIGNED)
-+ val |= IP_ALIGNED;
-+#endif
-+
-+ /*
-+ * Class PE packet steering will only work if TOE mode, bridge fetch or
-+ * route fetch are enabled (see class/qb_fet.v). Route fetch would
-+ * trigger additional memory copies (likely from DDR because of hash
-+ * table size, which cannot be reduced because PE software still
-+ * relies on hash value computed in HW), so when not in TOE mode we
-+ * simply enable HW bridge fetch even though we don't use it.
-+ */
-+ if (cfg->toe_mode)
-+ val |= CLASS_TOE;
-+ else
-+ val |= HW_BRIDGE_FETCH;
-+
-+ writel(val, CLASS_ROUTE_MULTI);
-+
-+ writel(DDR_PHYS_TO_PFE(cfg->route_table_baseaddr),
-+ CLASS_ROUTE_TABLE_BASE);
-+ writel(CLASS_PE0_RO_DM_ADDR0_VAL, CLASS_PE0_RO_DM_ADDR0);
-+ writel(CLASS_PE0_RO_DM_ADDR1_VAL, CLASS_PE0_RO_DM_ADDR1);
-+ writel(CLASS_PE0_QB_DM_ADDR0_VAL, CLASS_PE0_QB_DM_ADDR0);
-+ writel(CLASS_PE0_QB_DM_ADDR1_VAL, CLASS_PE0_QB_DM_ADDR1);
-+ writel(CBUS_VIRT_TO_PFE(TMU_PHY_INQ_PKTPTR), CLASS_TM_INQ_ADDR);
-+
-+ writel(23, CLASS_AFULL_THRES);
-+ writel(23, CLASS_TSQ_FIFO_THRES);
-+
-+ writel(24, CLASS_MAX_BUF_CNT);
-+ writel(24, CLASS_TSQ_MAX_CNT);
-+}
-+
-+/**************************** TMU ***************************/
-+
-+void tmu_reset(void)
-+{
-+ writel(SW_RESET, TMU_CTRL);
-+}
-+
-+/* Initializes TMU block.
-+ * @param[in] cfg TMU configuration
-+ */
-+void tmu_init(struct tmu_cfg *cfg)
-+{
-+ int q, phyno;
-+
-+ tmu_disable(0xF);
-+ mdelay(10);
-+
-+#if !defined(LS1012A_PFE_RESET_WA)
-+ /* keep in soft reset */
-+ writel(SW_RESET, TMU_CTRL);
-+#endif
-+ writel(0x3, TMU_SYS_GENERIC_CONTROL);
-+ writel(750, TMU_INQ_WATERMARK);
-+ writel(CBUS_VIRT_TO_PFE(EGPI1_BASE_ADDR +
-+ GPI_INQ_PKTPTR), TMU_PHY0_INQ_ADDR);
-+ writel(CBUS_VIRT_TO_PFE(EGPI2_BASE_ADDR +
-+ GPI_INQ_PKTPTR), TMU_PHY1_INQ_ADDR);
-+ writel(CBUS_VIRT_TO_PFE(HGPI_BASE_ADDR +
-+ GPI_INQ_PKTPTR), TMU_PHY3_INQ_ADDR);
-+ writel(CBUS_VIRT_TO_PFE(HIF_NOCPY_RX_INQ0_PKTPTR), TMU_PHY4_INQ_ADDR);
-+ writel(CBUS_VIRT_TO_PFE(UTIL_INQ_PKTPTR), TMU_PHY5_INQ_ADDR);
-+ writel(CBUS_VIRT_TO_PFE(BMU2_BASE_ADDR + BMU_FREE_CTRL),
-+ TMU_BMU_INQ_ADDR);
-+
-+ writel(0x3FF, TMU_TDQ0_SCH_CTRL); /*
-+ * enabling all 10
-+ * schedulers [9:0] of each TDQ
-+ */
-+ writel(0x3FF, TMU_TDQ1_SCH_CTRL);
-+ writel(0x3FF, TMU_TDQ3_SCH_CTRL);
-+
-+#if !defined(LS1012A_PFE_RESET_WA)
-+ writel(cfg->pe_sys_clk_ratio, TMU_PE_SYS_CLK_RATIO);
-+#endif
-+
-+#if !defined(LS1012A_PFE_RESET_WA)
-+ writel(DDR_PHYS_TO_PFE(cfg->llm_base_addr), TMU_LLM_BASE_ADDR);
-+ /* Extra packet pointers will be stored from this address onwards */
-+
-+ writel(cfg->llm_queue_len, TMU_LLM_QUE_LEN);
-+ writel(5, TMU_TDQ_IIFG_CFG);
-+ writel(DDR_BUF_SIZE, TMU_BMU_BUF_SIZE);
-+
-+ writel(0x0, TMU_CTRL);
-+
-+ /* MEM init */
-+ pr_info("%s: mem init\n", __func__);
-+ writel(MEM_INIT, TMU_CTRL);
-+
-+ while (!(readl(TMU_CTRL) & MEM_INIT_DONE))
-+ ;
-+
-+ /* LLM init */
-+ pr_info("%s: lmem init\n", __func__);
-+ writel(LLM_INIT, TMU_CTRL);
-+
-+ while (!(readl(TMU_CTRL) & LLM_INIT_DONE))
-+ ;
-+#endif
-+ /* set up each queue for tail drop */
-+ for (phyno = 0; phyno < 4; phyno++) {
-+ if (phyno == 2)
-+ continue;
-+ for (q = 0; q < 16; q++) {
-+ u32 qdepth;
-+
-+ writel((phyno << 8) | q, TMU_TEQ_CTRL);
-+ writel(1 << 22, TMU_TEQ_QCFG); /*Enable tail drop */
-+
-+ if (phyno == 3)
-+ qdepth = DEFAULT_TMU3_QDEPTH;
-+ else
-+ qdepth = (q == 0) ? DEFAULT_Q0_QDEPTH :
-+ DEFAULT_MAX_QDEPTH;
-+
-+ /* LOG: 68855 */
-+ /*
-+ * The following is a workaround for the reordered
-+ * packet and BMU2 buffer leakage issue.
-+ */
-+ if (CHIP_REVISION() == 0)
-+ qdepth = 31;
-+
-+ writel(qdepth << 18, TMU_TEQ_HW_PROB_CFG2);
-+ writel(qdepth >> 14, TMU_TEQ_HW_PROB_CFG3);
-+ }
-+ }
-+
-+#ifdef CFG_LRO
-+ /* Set TMU-3 queue 5 (LRO) in no-drop mode */
-+ writel((3 << 8) | TMU_QUEUE_LRO, TMU_TEQ_CTRL);
-+ writel(0, TMU_TEQ_QCFG);
-+#endif
-+
-+ writel(0x05, TMU_TEQ_DISABLE_DROPCHK);
-+
-+ writel(0x0, TMU_CTRL);
-+}
-+
-+/* Enables TMU-PE cores.
-+ * @param[in] pe_mask TMU PE mask
-+ */
-+void tmu_enable(u32 pe_mask)
-+{
-+ writel(readl(TMU_TX_CTRL) | (pe_mask & 0xF), TMU_TX_CTRL);
-+}
-+
-+/* Disables TMU cores.
-+ * @param[in] pe_mask TMU PE mask
-+ */
-+void tmu_disable(u32 pe_mask)
-+{
-+ writel(readl(TMU_TX_CTRL) & ~(pe_mask & 0xF), TMU_TX_CTRL);
-+}
-+
-+/* This will return the tmu queue status
-+ * @param[in] if_id gem interface id or TMU index
-+ * @return returns the bit mask of busy queues, zero means all
-+ * queues are empty
-+ */
-+u32 tmu_qstatus(u32 if_id)
-+{
-+ return cpu_to_be32(pe_dmem_read(TMU0_ID + if_id, TMU_DM_PESTATUS +
-+ offsetof(struct pe_status, tmu_qstatus), 4));
-+}
-+
-+u32 tmu_pkts_processed(u32 if_id)
-+{
-+ return cpu_to_be32(pe_dmem_read(TMU0_ID + if_id, TMU_DM_PESTATUS +
-+ offsetof(struct pe_status, rx), 4));
-+}
-+
-+/**************************** UTIL ***************************/
-+
-+/* Resets UTIL block.
-+ */
-+void util_reset(void)
-+{
-+ writel(CORE_SW_RESET, UTIL_TX_CTRL);
-+}
-+
-+/* Initializes UTIL block.
-+ * @param[in] cfg UTIL configuration
-+ */
-+void util_init(struct util_cfg *cfg)
-+{
-+ writel(cfg->pe_sys_clk_ratio, UTIL_PE_SYS_CLK_RATIO);
-+}
-+
-+/* Enables UTIL-PE core.
-+ *
-+ */
-+void util_enable(void)
-+{
-+ writel(CORE_ENABLE, UTIL_TX_CTRL);
-+}
-+
-+/* Disables UTIL-PE core.
-+ *
-+ */
-+void util_disable(void)
-+{
-+ writel(CORE_DISABLE, UTIL_TX_CTRL);
-+}
-+
-+/**************************** HIF ***************************/
-+/* Initializes HIF copy block.
-+ *
-+ */
-+void hif_init(void)
-+{
-+ /*Initialize HIF registers*/
-+ writel((HIF_RX_POLL_CTRL_CYCLE << 16) | HIF_TX_POLL_CTRL_CYCLE,
-+ HIF_POLL_CTRL);
-+}
-+
-+/* Enable hif tx DMA and interrupt
-+ *
-+ */
-+void hif_tx_enable(void)
-+{
-+ writel(HIF_CTRL_DMA_EN, HIF_TX_CTRL);
-+ writel((readl(HIF_INT_ENABLE) | HIF_INT_EN | HIF_TXPKT_INT_EN),
-+ HIF_INT_ENABLE);
-+}
-+
-+/* Disable hif tx DMA and interrupt
-+ *
-+ */
-+void hif_tx_disable(void)
-+{
-+ u32 hif_int;
-+
-+ writel(0, HIF_TX_CTRL);
-+
-+ hif_int = readl(HIF_INT_ENABLE);
-+ hif_int &= HIF_TXPKT_INT_EN;
-+ writel(hif_int, HIF_INT_ENABLE);
-+}
-+
-+/* Enable hif rx DMA and interrupt
-+ *
-+ */
-+void hif_rx_enable(void)
-+{
-+ hif_rx_dma_start();
-+ writel((readl(HIF_INT_ENABLE) | HIF_INT_EN | HIF_RXPKT_INT_EN),
-+ HIF_INT_ENABLE);
-+}
-+
-+/* Disable hif rx DMA and interrupt
-+ *
-+ */
-+void hif_rx_disable(void)
-+{
-+ u32 hif_int;
-+
-+ writel(0, HIF_RX_CTRL);
-+
-+ hif_int = readl(HIF_INT_ENABLE);
-+ hif_int &= HIF_RXPKT_INT_EN;
-+ writel(hif_int, HIF_INT_ENABLE);
-+}
---- /dev/null
-+++ b/drivers/staging/fsl_ppfe/pfe_hif.c
-@@ -0,0 +1,1072 @@
-+/*
-+ * Copyright 2015-2016 Freescale Semiconductor, Inc.
-+ * Copyright 2017 NXP
-+ *
-+ * This program is free software; you can redistribute it and/or modify
-+ * it under the terms of the GNU General Public License as published by
-+ * the Free Software Foundation; either version 2 of the License, or
-+ * (at your option) any later version.
-+ *
-+ * This program is distributed in the hope that it will be useful,
-+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
-+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
-+ * GNU General Public License for more details.
-+ *
-+ * You should have received a copy of the GNU General Public License
-+ * along with this program. If not, see <http://www.gnu.org/licenses/>.
-+ */
-+
-+#include <linux/kernel.h>
-+#include <linux/interrupt.h>
-+#include <linux/dma-mapping.h>
-+#include <linux/dmapool.h>
-+#include <linux/sched.h>
-+#include <linux/module.h>
-+#include <linux/list.h>
-+#include <linux/kthread.h>
-+#include <linux/slab.h>
-+
-+#include <linux/io.h>
-+#include <asm/irq.h>
-+
-+#include "pfe_mod.h"
-+
-+#define HIF_INT_MASK (HIF_INT | HIF_RXPKT_INT | HIF_TXPKT_INT)
-+
-+unsigned char napi_first_batch;
-+
-+static void pfe_tx_do_cleanup(unsigned long data);
-+
-+static int pfe_hif_alloc_descr(struct pfe_hif *hif)
-+{
-+ void *addr;
-+ dma_addr_t dma_addr;
-+ int err = 0;
-+
-+ pr_info("%s\n", __func__);
-+ addr = dma_alloc_coherent(pfe->dev,
-+ HIF_RX_DESC_NT * sizeof(struct hif_desc) +
-+ HIF_TX_DESC_NT * sizeof(struct hif_desc),
-+ &dma_addr, GFP_KERNEL);
-+
-+ if (!addr) {
-+ pr_err("%s: Could not allocate buffer descriptors!\n"
-+ , __func__);
-+ err = -ENOMEM;
-+ goto err0;
-+ }
-+
-+ hif->descr_baseaddr_p = dma_addr;
-+ hif->descr_baseaddr_v = addr;
-+ hif->rx_ring_size = HIF_RX_DESC_NT;
-+ hif->tx_ring_size = HIF_TX_DESC_NT;
-+
-+ return 0;
-+
-+err0:
-+ return err;
-+}
-+
-+#if defined(LS1012A_PFE_RESET_WA)
-+static void pfe_hif_disable_rx_desc(struct pfe_hif *hif)
-+{
-+ int ii;
-+ struct hif_desc *desc = hif->rx_base;
-+
-+ /*Mark all descriptors as LAST_BD */
-+ for (ii = 0; ii < hif->rx_ring_size; ii++) {
-+ desc->ctrl |= BD_CTRL_LAST_BD;
-+ desc++;
-+ }
-+}
-+
-+struct class_rx_hdr_t {
-+ u32 next_ptr; /* ptr to the start of the first DDR buffer */
-+ u16 length; /* total packet length */
-+ u16 phyno; /* input physical port number */
-+ u32 status; /* gemac status bits */
-+ u32 status2; /* reserved for software usage */
-+};
-+
-+/* STATUS_BAD_FRAME_ERR is set for all errors (including checksums if enabled)
-+ * except overflow
-+ */
-+#define STATUS_BAD_FRAME_ERR BIT(16)
-+#define STATUS_LENGTH_ERR BIT(17)
-+#define STATUS_CRC_ERR BIT(18)
-+#define STATUS_TOO_SHORT_ERR BIT(19)
-+#define STATUS_TOO_LONG_ERR BIT(20)
-+#define STATUS_CODE_ERR BIT(21)
-+#define STATUS_MC_HASH_MATCH BIT(22)
-+#define STATUS_CUMULATIVE_ARC_HIT BIT(23)
-+#define STATUS_UNICAST_HASH_MATCH BIT(24)
-+#define STATUS_IP_CHECKSUM_CORRECT BIT(25)
-+#define STATUS_TCP_CHECKSUM_CORRECT BIT(26)
-+#define STATUS_UDP_CHECKSUM_CORRECT BIT(27)
-+#define STATUS_OVERFLOW_ERR BIT(28) /* GPI error */
-+#define MIN_PKT_SIZE 64
-+
-+static inline void copy_to_lmem(u32 *dst, u32 *src, int len)
-+{
-+ int i;
-+
-+ for (i = 0; i < len; i += sizeof(u32)) {
-+ *dst = htonl(*src);
-+ dst++; src++;
-+ }
-+}
-+
-+static void send_dummy_pkt_to_hif(void)
-+{
-+ void *lmem_ptr, *ddr_ptr, *lmem_virt_addr;
-+ u32 physaddr;
-+ struct class_rx_hdr_t local_hdr;
-+ static u32 dummy_pkt[] = {
-+ 0x33221100, 0x2b785544, 0xd73093cb, 0x01000608,
-+ 0x04060008, 0x2b780200, 0xd73093cb, 0x0a01a8c0,
-+ 0x33221100, 0xa8c05544, 0x00000301, 0x00000000,
-+ 0x00000000, 0x00000000, 0x00000000, 0xbe86c51f };
-+
-+ ddr_ptr = (void *)((u64)readl(BMU2_BASE_ADDR + BMU_ALLOC_CTRL));
-+ if (!ddr_ptr)
-+ return;
-+
-+ lmem_ptr = (void *)((u64)readl(BMU1_BASE_ADDR + BMU_ALLOC_CTRL));
-+ if (!lmem_ptr)
-+ return;
-+
-+ pr_info("Sending a dummy pkt to HIF %p %p\n", ddr_ptr, lmem_ptr);
-+ physaddr = (u32)DDR_VIRT_TO_PFE(ddr_ptr);
-+
-+ lmem_virt_addr = (void *)CBUS_PFE_TO_VIRT((unsigned long int)lmem_ptr);
-+
-+ local_hdr.phyno = htons(0); /* RX_PHY_0 */
-+ local_hdr.length = htons(MIN_PKT_SIZE);
-+
-+ local_hdr.next_ptr = htonl((u32)physaddr);
-+ /*Mark checksum is correct */
-+ local_hdr.status = htonl((STATUS_IP_CHECKSUM_CORRECT |
-+ STATUS_UDP_CHECKSUM_CORRECT |
-+ STATUS_TCP_CHECKSUM_CORRECT |
-+ STATUS_UNICAST_HASH_MATCH |
-+ STATUS_CUMULATIVE_ARC_HIT));
-+ copy_to_lmem((u32 *)lmem_virt_addr, (u32 *)&local_hdr,
-+ sizeof(local_hdr));
-+
-+ copy_to_lmem((u32 *)(lmem_virt_addr + LMEM_HDR_SIZE), (u32 *)dummy_pkt,
-+ 0x40);
-+
-+ writel((unsigned long int)lmem_ptr, CLASS_INQ_PKTPTR);
-+}
-+
-+void pfe_hif_rx_idle(struct pfe_hif *hif)
-+{
-+ int hif_stop_loop = 10;
-+ u32 rx_status;
-+
-+ pfe_hif_disable_rx_desc(hif);
-+ pr_info("Bringing hif to idle state...");
-+ writel(0, HIF_INT_ENABLE);
-+ /*If HIF Rx BDP is busy send a dummy packet */
-+ do {
-+ rx_status = readl(HIF_RX_STATUS);
-+ if (rx_status & BDP_CSR_RX_DMA_ACTV)
-+ send_dummy_pkt_to_hif();
-+
-+ usleep_range(100, 150);
-+ } while (--hif_stop_loop);
-+
-+ if (readl(HIF_RX_STATUS) & BDP_CSR_RX_DMA_ACTV)
-+ pr_info("Failed\n");
-+ else
-+ pr_info("Done\n");
-+}
-+#endif
-+
-+static void pfe_hif_free_descr(struct pfe_hif *hif)
-+{
-+ pr_info("%s\n", __func__);
-+
-+ dma_free_coherent(pfe->dev,
-+ hif->rx_ring_size * sizeof(struct hif_desc) +
-+ hif->tx_ring_size * sizeof(struct hif_desc),
-+ hif->descr_baseaddr_v, hif->descr_baseaddr_p);
-+}
-+
-+void pfe_hif_desc_dump(struct pfe_hif *hif)
-+{
-+ struct hif_desc *desc;
-+ unsigned long desc_p;
-+ int ii = 0;
-+
-+ pr_info("%s\n", __func__);
-+
-+ desc = hif->rx_base;
-+ desc_p = (u32)((u64)desc - (u64)hif->descr_baseaddr_v +
-+ hif->descr_baseaddr_p);
-+
-+ pr_info("HIF Rx desc base %p physical %x\n", desc, (u32)desc_p);
-+ for (ii = 0; ii < hif->rx_ring_size; ii++) {
-+ pr_info("status: %08x, ctrl: %08x, data: %08x, next: %x\n",
-+ readl(&desc->status), readl(&desc->ctrl),
-+ readl(&desc->data), readl(&desc->next));
-+ desc++;
-+ }
-+
-+ desc = hif->tx_base;
-+ desc_p = ((u64)desc - (u64)hif->descr_baseaddr_v +
-+ hif->descr_baseaddr_p);
-+
-+ pr_info("HIF Tx desc base %p physical %x\n", desc, (u32)desc_p);
-+ for (ii = 0; ii < hif->tx_ring_size; ii++) {
-+ pr_info("status: %08x, ctrl: %08x, data: %08x, next: %x\n",
-+ readl(&desc->status), readl(&desc->ctrl),
-+ readl(&desc->data), readl(&desc->next));
-+ desc++;
-+ }
-+}
-+
-+/* pfe_hif_release_buffers */
-+static void pfe_hif_release_buffers(struct pfe_hif *hif)
-+{
-+ struct hif_desc *desc;
-+ int i = 0;
-+
-+ hif->rx_base = hif->descr_baseaddr_v;
-+
-+ pr_info("%s\n", __func__);
-+
-+ /*Free Rx buffers */
-+ desc = hif->rx_base;
-+ for (i = 0; i < hif->rx_ring_size; i++) {
-+ if (readl(&desc->data)) {
-+ if ((i < hif->shm->rx_buf_pool_cnt) &&
-+ (!hif->shm->rx_buf_pool[i])) {
-+ /*
-+ * dma_unmap_single(hif->dev, desc->data,
-+ * hif->rx_buf_len[i], DMA_FROM_DEVICE);
-+ */
-+ dma_unmap_single(hif->dev,
-+ DDR_PFE_TO_PHYS(
-+ readl(&desc->data)),
-+ hif->rx_buf_len[i],
-+ DMA_FROM_DEVICE);
-+ hif->shm->rx_buf_pool[i] = hif->rx_buf_addr[i];
-+ } else {
-+ pr_err("%s: buffer pool already full\n"
-+ , __func__);
-+ }
-+ }
-+
-+ writel(0, &desc->data);
-+ writel(0, &desc->status);
-+ writel(0, &desc->ctrl);
-+ desc++;
-+ }
-+}
-+
-+/*
-+ * pfe_hif_init_buffers
-+ * This function initializes the HIF Rx/Tx ring descriptors and
-+ * initialize Rx queue with buffers.
-+ */
-+static int pfe_hif_init_buffers(struct pfe_hif *hif)
-+{
-+ struct hif_desc *desc, *first_desc_p;
-+ u32 data;
-+ int i = 0;
-+
-+ pr_info("%s\n", __func__);
-+
-+ /* Check enough Rx buffers available in the shared memory */
-+ if (hif->shm->rx_buf_pool_cnt < hif->rx_ring_size)
-+ return -ENOMEM;
-+
-+ hif->rx_base = hif->descr_baseaddr_v;
-+ memset(hif->rx_base, 0, hif->rx_ring_size * sizeof(struct hif_desc));
-+
-+ /*Initialize Rx descriptors */
-+ desc = hif->rx_base;
-+ first_desc_p = (struct hif_desc *)hif->descr_baseaddr_p;
-+
-+ for (i = 0; i < hif->rx_ring_size; i++) {
-+ /* Initialize Rx buffers from the shared memory */
-+
-+ data = (u32)dma_map_single(hif->dev, hif->shm->rx_buf_pool[i],
-+ pfe_pkt_size, DMA_FROM_DEVICE);
-+ hif->rx_buf_addr[i] = hif->shm->rx_buf_pool[i];
-+ hif->rx_buf_len[i] = pfe_pkt_size;
-+ hif->shm->rx_buf_pool[i] = NULL;
-+
-+ if (likely(dma_mapping_error(hif->dev, data) == 0)) {
-+ writel(DDR_PHYS_TO_PFE(data), &desc->data);
-+ } else {
-+ pr_err("%s : low on mem\n", __func__);
-+
-+ goto err;
-+ }
-+
-+ writel(0, &desc->status);
-+
-+ /*
-+ * Ensure everything else is written to DDR before
-+ * writing bd->ctrl
-+ */
-+ wmb();
-+
-+ writel((BD_CTRL_PKT_INT_EN | BD_CTRL_LIFM
-+ | BD_CTRL_DIR | BD_CTRL_DESC_EN
-+ | BD_BUF_LEN(pfe_pkt_size)), &desc->ctrl);
-+
-+ /* Chain descriptors */
-+ writel((u32)DDR_PHYS_TO_PFE(first_desc_p + i + 1), &desc->next);
-+ desc++;
-+ }
-+
-+ /* Overwrite last descriptor to chain it to first one*/
-+ desc--;
-+ writel((u32)DDR_PHYS_TO_PFE(first_desc_p), &desc->next);
-+
-+ hif->rxtoclean_index = 0;
-+
-+ /*Initialize Rx buffer descriptor ring base address */
-+ writel(DDR_PHYS_TO_PFE(hif->descr_baseaddr_p), HIF_RX_BDP_ADDR);
-+
-+ hif->tx_base = hif->rx_base + hif->rx_ring_size;
-+ first_desc_p = (struct hif_desc *)hif->descr_baseaddr_p +
-+ hif->rx_ring_size;
-+ memset(hif->tx_base, 0, hif->tx_ring_size * sizeof(struct hif_desc));
-+
-+ /*Initialize tx descriptors */
-+ desc = hif->tx_base;
-+
-+ for (i = 0; i < hif->tx_ring_size; i++) {
-+ /* Chain descriptors */
-+ writel((u32)DDR_PHYS_TO_PFE(first_desc_p + i + 1), &desc->next);
-+ writel(0, &desc->ctrl);
-+ desc++;
-+ }
-+
-+ /* Overwrite last descriptor to chain it to first one */
-+ desc--;
-+ writel((u32)DDR_PHYS_TO_PFE(first_desc_p), &desc->next);
-+ hif->txavail = hif->tx_ring_size;
-+ hif->txtosend = 0;
-+ hif->txtoclean = 0;
-+ hif->txtoflush = 0;
-+
-+ /*Initialize Tx buffer descriptor ring base address */
-+ writel((u32)DDR_PHYS_TO_PFE(first_desc_p), HIF_TX_BDP_ADDR);
-+
-+ return 0;
-+
-+err:
-+ pfe_hif_release_buffers(hif);
-+ return -ENOMEM;
-+}
-+
-+/*
-+ * pfe_hif_client_register
-+ *
-+ * This function used to register a client driver with the HIF driver.
-+ *
-+ * Return value:
-+ * 0 - on Successful registration
-+ */
-+static int pfe_hif_client_register(struct pfe_hif *hif, u32 client_id,
-+ struct hif_client_shm *client_shm)
-+{
-+ struct hif_client *client = &hif->client[client_id];
-+ u32 i, cnt;
-+ struct rx_queue_desc *rx_qbase;
-+ struct tx_queue_desc *tx_qbase;
-+ struct hif_rx_queue *rx_queue;
-+ struct hif_tx_queue *tx_queue;
-+ int err = 0;
-+
-+ pr_info("%s\n", __func__);
-+
-+ spin_lock_bh(&hif->tx_lock);
-+
-+ if (test_bit(client_id, &hif->shm->g_client_status[0])) {
-+ pr_err("%s: client %d already registered\n",
-+ __func__, client_id);
-+ err = -1;
-+ goto unlock;
-+ }
-+
-+ memset(client, 0, sizeof(struct hif_client));
-+
-+ /* Initialize client Rx queues baseaddr, size */
-+
-+ cnt = CLIENT_CTRL_RX_Q_CNT(client_shm->ctrl);
-+ /* Check if client is requesting for more queues than supported */
-+ if (cnt > HIF_CLIENT_QUEUES_MAX)
-+ cnt = HIF_CLIENT_QUEUES_MAX;
-+
-+ client->rx_qn = cnt;
-+ rx_qbase = (struct rx_queue_desc *)client_shm->rx_qbase;
-+ for (i = 0; i < cnt; i++) {
-+ rx_queue = &client->rx_q[i];
-+ rx_queue->base = rx_qbase + i * client_shm->rx_qsize;
-+ rx_queue->size = client_shm->rx_qsize;
-+ rx_queue->write_idx = 0;
-+ }
-+
-+ /* Initialize client Tx queues baseaddr, size */
-+ cnt = CLIENT_CTRL_TX_Q_CNT(client_shm->ctrl);
-+
-+ /* Check if client is requesting for more queues than supported */
-+ if (cnt > HIF_CLIENT_QUEUES_MAX)
-+ cnt = HIF_CLIENT_QUEUES_MAX;
-+
-+ client->tx_qn = cnt;
-+ tx_qbase = (struct tx_queue_desc *)client_shm->tx_qbase;
-+ for (i = 0; i < cnt; i++) {
-+ tx_queue = &client->tx_q[i];
-+ tx_queue->base = tx_qbase + i * client_shm->tx_qsize;
-+ tx_queue->size = client_shm->tx_qsize;
-+ tx_queue->ack_idx = 0;
-+ }
-+
-+ set_bit(client_id, &hif->shm->g_client_status[0]);
-+
-+unlock:
-+ spin_unlock_bh(&hif->tx_lock);
-+
-+ return err;
-+}
-+
-+/*
-+ * pfe_hif_client_unregister
-+ *
-+ * This function used to unregister a client from the HIF driver.
-+ *
-+ */
-+static void pfe_hif_client_unregister(struct pfe_hif *hif, u32 client_id)
-+{
-+ pr_info("%s\n", __func__);
-+
-+ /*
-+ * Mark client as no longer available (which prevents further packet
-+ * receive for this client)
-+ */
-+ spin_lock_bh(&hif->tx_lock);
-+
-+ if (!test_bit(client_id, &hif->shm->g_client_status[0])) {
-+ pr_err("%s: client %d not registered\n", __func__,
-+ client_id);
-+
-+ spin_unlock_bh(&hif->tx_lock);
-+ return;
-+ }
-+
-+ clear_bit(client_id, &hif->shm->g_client_status[0]);
-+
-+ spin_unlock_bh(&hif->tx_lock);
-+}
-+
-+/*
-+ * client_put_rxpacket-
-+ * This functions puts the Rx pkt in the given client Rx queue.
-+ * It actually swap the Rx pkt in the client Rx descriptor buffer
-+ * and returns the free buffer from it.
-+ *
-+ * If the function returns NULL means client Rx queue is full and
-+ * packet couldn't send to client queue.
-+ */
-+static void *client_put_rxpacket(struct hif_rx_queue *queue, void *pkt, u32 len,
-+ u32 flags, u32 client_ctrl, u32 *rem_len)
-+{
-+ void *free_pkt = NULL;
-+ struct rx_queue_desc *desc = queue->base + queue->write_idx;
-+
-+ if (readl(&desc->ctrl) & CL_DESC_OWN) {
-+ if (page_mode) {
-+ int rem_page_size = PAGE_SIZE -
-+ PRESENT_OFST_IN_PAGE(pkt);
-+ int cur_pkt_size = ROUND_MIN_RX_SIZE(len +
-+ pfe_pkt_headroom);
-+ *rem_len = (rem_page_size - cur_pkt_size);
-+ if (*rem_len) {
-+ free_pkt = pkt + cur_pkt_size;
-+ get_page(virt_to_page(free_pkt));
-+ } else {
-+ free_pkt = (void
-+ *)__get_free_page(GFP_ATOMIC | GFP_DMA_PFE);
-+ *rem_len = pfe_pkt_size;
-+ }
-+ } else {
-+ free_pkt = kmalloc(PFE_BUF_SIZE, GFP_ATOMIC |
-+ GFP_DMA_PFE);
-+ *rem_len = PFE_BUF_SIZE - pfe_pkt_headroom;
-+ }
-+
-+ if (free_pkt) {
-+ desc->data = pkt;
-+ desc->client_ctrl = client_ctrl;
-+ /*
-+ * Ensure everything else is written to DDR before
-+ * writing bd->ctrl
-+ */
-+ smp_wmb();
-+ writel(CL_DESC_BUF_LEN(len) | flags, &desc->ctrl);
-+ queue->write_idx = (queue->write_idx + 1)
-+ & (queue->size - 1);
-+
-+ free_pkt += pfe_pkt_headroom;
-+ }
-+ }
-+
-+ return free_pkt;
-+}
-+
-+/*
-+ * pfe_hif_rx_process-
-+ * This function does pfe hif rx queue processing.
-+ * Dequeue packet from Rx queue and send it to corresponding client queue
-+ */
-+static int pfe_hif_rx_process(struct pfe_hif *hif, int budget)
-+{
-+ struct hif_desc *desc;
-+ struct hif_hdr *pkt_hdr;
-+ struct __hif_hdr hif_hdr;
-+ void *free_buf;
-+ int rtc, len, rx_processed = 0;
-+ struct __hif_desc local_desc;
-+ int flags;
-+ unsigned int desc_p;
-+ unsigned int buf_size = 0;
-+
-+ spin_lock_bh(&hif->lock);
-+
-+ rtc = hif->rxtoclean_index;
-+
-+ while (rx_processed < budget) {
-+ desc = hif->rx_base + rtc;
-+
-+ __memcpy12(&local_desc, desc);
-+
-+ /* ACK pending Rx interrupt */
-+ if (local_desc.ctrl & BD_CTRL_DESC_EN) {
-+ writel(HIF_INT | HIF_RXPKT_INT, HIF_INT_SRC);
-+
-+ if (rx_processed == 0) {
-+ if (napi_first_batch == 1) {
-+ desc_p = hif->descr_baseaddr_p +
-+ ((unsigned long int)(desc) -
-+ (unsigned long
-+ int)hif->descr_baseaddr_v);
-+ napi_first_batch = 0;
-+ }
-+ }
-+
-+ __memcpy12(&local_desc, desc);
-+
-+ if (local_desc.ctrl & BD_CTRL_DESC_EN)
-+ break;
-+ }
-+
-+ napi_first_batch = 0;
-+
-+#ifdef HIF_NAPI_STATS
-+ hif->napi_counters[NAPI_DESC_COUNT]++;
-+#endif
-+ len = BD_BUF_LEN(local_desc.ctrl);
-+ /*
-+ * dma_unmap_single(hif->dev, DDR_PFE_TO_PHYS(local_desc.data),
-+ * hif->rx_buf_len[rtc], DMA_FROM_DEVICE);
-+ */
-+ dma_unmap_single(hif->dev, DDR_PFE_TO_PHYS(local_desc.data),
-+ hif->rx_buf_len[rtc], DMA_FROM_DEVICE);
-+
-+ pkt_hdr = (struct hif_hdr *)hif->rx_buf_addr[rtc];
-+
-+ /* Track last HIF header received */
-+ if (!hif->started) {
-+ hif->started = 1;
-+
-+ __memcpy8(&hif_hdr, pkt_hdr);
-+
-+ hif->qno = hif_hdr.hdr.q_num;
-+ hif->client_id = hif_hdr.hdr.client_id;
-+ hif->client_ctrl = (hif_hdr.hdr.client_ctrl1 << 16) |
-+ hif_hdr.hdr.client_ctrl;
-+ flags = CL_DESC_FIRST;
-+
-+ } else {
-+ flags = 0;
-+ }
-+
-+ if (local_desc.ctrl & BD_CTRL_LIFM)
-+ flags |= CL_DESC_LAST;
-+
-+ /* Check for valid client id and still registered */
-+ if ((hif->client_id >= HIF_CLIENTS_MAX) ||
-+ !(test_bit(hif->client_id,
-+ &hif->shm->g_client_status[0]))) {
-+ printk_ratelimited("%s: packet with invalid client id %d q_num %d\n",
-+ __func__,
-+ hif->client_id,
-+ hif->qno);
-+
-+ free_buf = pkt_hdr;
-+
-+ goto pkt_drop;
-+ }
-+
-+ /* Check to valid queue number */
-+ if (hif->client[hif->client_id].rx_qn <= hif->qno) {
-+ pr_info("%s: packet with invalid queue: %d\n"
-+ , __func__, hif->qno);
-+ hif->qno = 0;
-+ }
-+
-+ free_buf =
-+ client_put_rxpacket(&hif->client[hif->client_id].rx_q[hif->qno],
-+ (void *)pkt_hdr, len, flags,
-+ hif->client_ctrl, &buf_size);
-+
-+ hif_lib_indicate_client(hif->client_id, EVENT_RX_PKT_IND,
-+ hif->qno);
-+
-+ if (unlikely(!free_buf)) {
-+#ifdef HIF_NAPI_STATS
-+ hif->napi_counters[NAPI_CLIENT_FULL_COUNT]++;
-+#endif
-+ /*
-+ * If we want to keep in polling mode to retry later,
-+ * we need to tell napi that we consumed
-+ * the full budget or we will hit a livelock scenario.
-+ * The core code keeps this napi instance
-+ * at the head of the list and none of the other
-+ * instances get to run
-+ */
-+ rx_processed = budget;
-+
-+ if (flags & CL_DESC_FIRST)
-+ hif->started = 0;
-+
-+ break;
-+ }
-+
-+pkt_drop:
-+ /*Fill free buffer in the descriptor */
-+ hif->rx_buf_addr[rtc] = free_buf;
-+ hif->rx_buf_len[rtc] = min(pfe_pkt_size, buf_size);
-+ writel((DDR_PHYS_TO_PFE
-+ ((u32)dma_map_single(hif->dev,
-+ free_buf, hif->rx_buf_len[rtc], DMA_FROM_DEVICE))),
-+ &desc->data);
-+ /*
-+ * Ensure everything else is written to DDR before
-+ * writing bd->ctrl
-+ */
-+ wmb();
-+ writel((BD_CTRL_PKT_INT_EN | BD_CTRL_LIFM | BD_CTRL_DIR |
-+ BD_CTRL_DESC_EN | BD_BUF_LEN(hif->rx_buf_len[rtc])),
-+ &desc->ctrl);
-+
-+ rtc = (rtc + 1) & (hif->rx_ring_size - 1);
-+
-+ if (local_desc.ctrl & BD_CTRL_LIFM) {
-+ if (!(hif->client_ctrl & HIF_CTRL_RX_CONTINUED)) {
-+ rx_processed++;
-+
-+#ifdef HIF_NAPI_STATS
-+ hif->napi_counters[NAPI_PACKET_COUNT]++;
-+#endif
-+ }
-+ hif->started = 0;
-+ }
-+ }
-+
-+ hif->rxtoclean_index = rtc;
-+ spin_unlock_bh(&hif->lock);
-+
-+ /* we made some progress, re-start rx dma in case it stopped */
-+ hif_rx_dma_start();
-+
-+ return rx_processed;
-+}
-+
-+/*
-+ * client_ack_txpacket-
-+ * This function ack the Tx packet in the give client Tx queue by resetting
-+ * ownership bit in the descriptor.
-+ */
-+static int client_ack_txpacket(struct pfe_hif *hif, unsigned int client_id,
-+ unsigned int q_no)
-+{
-+ struct hif_tx_queue *queue = &hif->client[client_id].tx_q[q_no];
-+ struct tx_queue_desc *desc = queue->base + queue->ack_idx;
-+
-+ if (readl(&desc->ctrl) & CL_DESC_OWN) {
-+ writel((readl(&desc->ctrl) & ~CL_DESC_OWN), &desc->ctrl);
-+ queue->ack_idx = (queue->ack_idx + 1) & (queue->size - 1);
-+
-+ return 0;
-+
-+ } else {
-+ /*This should not happen */
-+ pr_err("%s: %d %d %d %d %d %p %d\n", __func__,
-+ hif->txtosend, hif->txtoclean, hif->txavail,
-+ client_id, q_no, queue, queue->ack_idx);
-+ WARN(1, "%s: doesn't own this descriptor", __func__);
-+ return 1;
-+ }
-+}
-+
-+void __hif_tx_done_process(struct pfe_hif *hif, int count)
-+{
-+ struct hif_desc *desc;
-+ struct hif_desc_sw *desc_sw;
-+ int ttc, tx_avl;
-+ int pkts_done[HIF_CLIENTS_MAX] = {0, 0};
-+
-+ ttc = hif->txtoclean;
-+ tx_avl = hif->txavail;
-+
-+ while ((tx_avl < hif->tx_ring_size) && count--) {
-+ desc = hif->tx_base + ttc;
-+
-+ if (readl(&desc->ctrl) & BD_CTRL_DESC_EN)
-+ break;
-+
-+ desc_sw = &hif->tx_sw_queue[ttc];
-+
-+ if (desc_sw->data) {
-+ /*
-+ * dmap_unmap_single(hif->dev, desc_sw->data,
-+ * desc_sw->len, DMA_TO_DEVICE);
-+ */
-+ dma_unmap_single(hif->dev, desc_sw->data,
-+ desc_sw->len, DMA_TO_DEVICE);
-+ }
-+
-+ if (desc_sw->client_id > HIF_CLIENTS_MAX)
-+ pr_err("Invalid cl id %d\n", desc_sw->client_id);
-+
-+ pkts_done[desc_sw->client_id]++;
-+
-+ client_ack_txpacket(hif, desc_sw->client_id, desc_sw->q_no);
-+
-+ ttc = (ttc + 1) & (hif->tx_ring_size - 1);
-+ tx_avl++;
-+ }
-+
-+ if (pkts_done[0])
-+ hif_lib_indicate_client(0, EVENT_TXDONE_IND, 0);
-+ if (pkts_done[1])
-+ hif_lib_indicate_client(1, EVENT_TXDONE_IND, 0);
-+
-+ hif->txtoclean = ttc;
-+ hif->txavail = tx_avl;
-+
-+ if (!count) {
-+ tasklet_schedule(&hif->tx_cleanup_tasklet);
-+ } else {
-+ /*Enable Tx done interrupt */
-+ writel(readl_relaxed(HIF_INT_ENABLE) | HIF_TXPKT_INT,
-+ HIF_INT_ENABLE);
-+ }
-+}
-+
-+static void pfe_tx_do_cleanup(unsigned long data)
-+{
-+ struct pfe_hif *hif = (struct pfe_hif *)data;
-+
-+ writel(HIF_INT | HIF_TXPKT_INT, HIF_INT_SRC);
-+
-+ hif_tx_done_process(hif, 64);
-+}
-+
-+/*
-+ * __hif_xmit_pkt -
-+ * This function puts one packet in the HIF Tx queue
-+ */
-+void __hif_xmit_pkt(struct pfe_hif *hif, unsigned int client_id, unsigned int
-+ q_no, void *data, u32 len, unsigned int flags)
-+{
-+ struct hif_desc *desc;
-+ struct hif_desc_sw *desc_sw;
-+
-+ desc = hif->tx_base + hif->txtosend;
-+ desc_sw = &hif->tx_sw_queue[hif->txtosend];
-+
-+ desc_sw->len = len;
-+ desc_sw->client_id = client_id;
-+ desc_sw->q_no = q_no;
-+ desc_sw->flags = flags;
-+
-+ if (flags & HIF_DONT_DMA_MAP) {
-+ desc_sw->data = 0;
-+ writel((u32)DDR_PHYS_TO_PFE(data), &desc->data);
-+ } else {
-+ desc_sw->data = dma_map_single(hif->dev, data, len,
-+ DMA_TO_DEVICE);
-+ writel((u32)DDR_PHYS_TO_PFE(desc_sw->data), &desc->data);
-+ }
-+
-+ hif->txtosend = (hif->txtosend + 1) & (hif->tx_ring_size - 1);
-+ hif->txavail--;
-+
-+ if ((!((flags & HIF_DATA_VALID) && (flags &
-+ HIF_LAST_BUFFER))))
-+ goto skip_tx;
-+
-+ /*
-+ * Ensure everything else is written to DDR before
-+ * writing bd->ctrl
-+ */
-+ wmb();
-+
-+ do {
-+ desc_sw = &hif->tx_sw_queue[hif->txtoflush];
-+ desc = hif->tx_base + hif->txtoflush;
-+
-+ if (desc_sw->flags & HIF_LAST_BUFFER) {
-+ writel((BD_CTRL_LIFM |
-+ BD_CTRL_BRFETCH_DISABLE | BD_CTRL_RTFETCH_DISABLE
-+ | BD_CTRL_PARSE_DISABLE | BD_CTRL_DESC_EN |
-+ BD_CTRL_PKT_INT_EN | BD_BUF_LEN(desc_sw->len)),
-+ &desc->ctrl);
-+ } else {
-+ writel((BD_CTRL_DESC_EN |
-+ BD_BUF_LEN(desc_sw->len)), &desc->ctrl);
-+ }
-+ hif->txtoflush = (hif->txtoflush + 1) & (hif->tx_ring_size - 1);
-+ }
-+ while (hif->txtoflush != hif->txtosend)
-+ ;
-+
-+skip_tx:
-+ return;
-+}
-+
-+static irqreturn_t wol_isr(int irq, void *dev_id)
-+{
-+ pr_info("WoL\n");
-+ gemac_set_wol(EMAC1_BASE_ADDR, 0);
-+ gemac_set_wol(EMAC2_BASE_ADDR, 0);
-+ return IRQ_HANDLED;
-+}
-+
-+/*
-+ * hif_isr-
-+ * This ISR routine processes Rx/Tx done interrupts from the HIF hardware block
-+ */
-+static irqreturn_t hif_isr(int irq, void *dev_id)
-+{
-+ struct pfe_hif *hif = (struct pfe_hif *)dev_id;
-+ int int_status;
-+ int int_enable_mask;
-+
-+ /*Read hif interrupt source register */
-+ int_status = readl_relaxed(HIF_INT_SRC);
-+ int_enable_mask = readl_relaxed(HIF_INT_ENABLE);
-+
-+ if ((int_status & HIF_INT) == 0)
-+ return IRQ_NONE;
-+
-+ int_status &= ~(HIF_INT);
-+
-+ if (int_status & HIF_RXPKT_INT) {
-+ int_status &= ~(HIF_RXPKT_INT);
-+ int_enable_mask &= ~(HIF_RXPKT_INT);
-+
-+ napi_first_batch = 1;
-+
-+ if (napi_schedule_prep(&hif->napi)) {
-+#ifdef HIF_NAPI_STATS
-+ hif->napi_counters[NAPI_SCHED_COUNT]++;
-+#endif
-+ __napi_schedule(&hif->napi);
-+ }
-+ }
-+
-+ if (int_status & HIF_TXPKT_INT) {
-+ int_status &= ~(HIF_TXPKT_INT);
-+ int_enable_mask &= ~(HIF_TXPKT_INT);
-+ /*Schedule tx cleanup tassklet */
-+ tasklet_schedule(&hif->tx_cleanup_tasklet);
-+ }
-+
-+ /*Disable interrupts, they will be enabled after they are serviced */
-+ writel_relaxed(int_enable_mask, HIF_INT_ENABLE);
-+
-+ if (int_status) {
-+ pr_info("%s : Invalid interrupt : %d\n", __func__,
-+ int_status);
-+ writel(int_status, HIF_INT_SRC);
-+ }
-+
-+ return IRQ_HANDLED;
-+}
-+
-+void hif_process_client_req(struct pfe_hif *hif, int req, int data1, int data2)
-+{
-+ unsigned int client_id = data1;
-+
-+ if (client_id >= HIF_CLIENTS_MAX) {
-+ pr_err("%s: client id %d out of bounds\n", __func__,
-+ client_id);
-+ return;
-+ }
-+
-+ switch (req) {
-+ case REQUEST_CL_REGISTER:
-+ /* Request for register a client */
-+ pr_info("%s: register client_id %d\n",
-+ __func__, client_id);
-+ pfe_hif_client_register(hif, client_id, (struct
-+ hif_client_shm *)&hif->shm->client[client_id]);
-+ break;
-+
-+ case REQUEST_CL_UNREGISTER:
-+ pr_info("%s: unregister client_id %d\n",
-+ __func__, client_id);
-+
-+ /* Request for unregister a client */
-+ pfe_hif_client_unregister(hif, client_id);
-+
-+ break;
-+
-+ default:
-+ pr_err("%s: unsupported request %d\n",
-+ __func__, req);
-+ break;
-+ }
-+
-+ /*
-+ * Process client Tx queues
-+ * Currently we don't have checking for tx pending
-+ */
-+}
-+
-+/*
-+ * pfe_hif_rx_poll
-+ * This function is NAPI poll function to process HIF Rx queue.
-+ */
-+static int pfe_hif_rx_poll(struct napi_struct *napi, int budget)
-+{
-+ struct pfe_hif *hif = container_of(napi, struct pfe_hif, napi);
-+ int work_done;
-+
-+#ifdef HIF_NAPI_STATS
-+ hif->napi_counters[NAPI_POLL_COUNT]++;
-+#endif
-+
-+ work_done = pfe_hif_rx_process(hif, budget);
-+
-+ if (work_done < budget) {
-+ napi_complete(napi);
-+ writel(readl_relaxed(HIF_INT_ENABLE) | HIF_RXPKT_INT,
-+ HIF_INT_ENABLE);
-+ }
-+#ifdef HIF_NAPI_STATS
-+ else
-+ hif->napi_counters[NAPI_FULL_BUDGET_COUNT]++;
-+#endif
-+
-+ return work_done;
-+}
-+
-+/*
-+ * pfe_hif_init
-+ * This function initializes the baseaddresses and irq, etc.
-+ */
-+int pfe_hif_init(struct pfe *pfe)
-+{
-+ struct pfe_hif *hif = &pfe->hif;
-+ int err;
-+
-+ pr_info("%s\n", __func__);
-+
-+ hif->dev = pfe->dev;
-+ hif->irq = pfe->hif_irq;
-+
-+ err = pfe_hif_alloc_descr(hif);
-+ if (err)
-+ goto err0;
-+
-+ if (pfe_hif_init_buffers(hif)) {
-+ pr_err("%s: Could not initialize buffer descriptors\n"
-+ , __func__);
-+ err = -ENOMEM;
-+ goto err1;
-+ }
-+
-+ /* Initialize NAPI for Rx processing */
-+ init_dummy_netdev(&hif->dummy_dev);
-+ netif_napi_add(&hif->dummy_dev, &hif->napi, pfe_hif_rx_poll,
-+ HIF_RX_POLL_WEIGHT);
-+ napi_enable(&hif->napi);
-+
-+ spin_lock_init(&hif->tx_lock);
-+ spin_lock_init(&hif->lock);
-+
-+ hif_init();
-+ hif_rx_enable();
-+ hif_tx_enable();
-+
-+ /* Disable tx done interrupt */
-+ writel(HIF_INT_MASK, HIF_INT_ENABLE);
-+
-+ gpi_enable(HGPI_BASE_ADDR);
-+
-+ err = request_irq(hif->irq, hif_isr, 0, "pfe_hif", hif);
-+ if (err) {
-+ pr_err("%s: failed to get the hif IRQ = %d\n",
-+ __func__, hif->irq);
-+ goto err1;
-+ }
-+
-+ err = request_irq(pfe->wol_irq, wol_isr, 0, "pfe_wol", pfe);
-+ if (err) {
-+ pr_err("%s: failed to get the wol IRQ = %d\n",
-+ __func__, pfe->wol_irq);
-+ goto err1;
-+ }
-+
-+ tasklet_init(&hif->tx_cleanup_tasklet,
-+ (void(*)(unsigned long))pfe_tx_do_cleanup,
-+ (unsigned long)hif);
-+
-+ return 0;
-+err1:
-+ pfe_hif_free_descr(hif);
-+err0:
-+ return err;
-+}
-+
-+/* pfe_hif_exit- */
-+void pfe_hif_exit(struct pfe *pfe)
-+{
-+ struct pfe_hif *hif = &pfe->hif;
-+
-+ pr_info("%s\n", __func__);
-+
-+ tasklet_kill(&hif->tx_cleanup_tasklet);
-+
-+ spin_lock_bh(&hif->lock);
-+ hif->shm->g_client_status[0] = 0;
-+ /* Make sure all clients are disabled*/
-+ hif->shm->g_client_status[1] = 0;
-+
-+ spin_unlock_bh(&hif->lock);
-+
-+ /*Disable Rx/Tx */
-+ gpi_disable(HGPI_BASE_ADDR);
-+ hif_rx_disable();
-+ hif_tx_disable();
-+
-+ napi_disable(&hif->napi);
-+ netif_napi_del(&hif->napi);
-+
-+ free_irq(pfe->wol_irq, pfe);
-+ free_irq(hif->irq, hif);
-+
-+ pfe_hif_release_buffers(hif);
-+ pfe_hif_free_descr(hif);
-+}
---- /dev/null
-+++ b/drivers/staging/fsl_ppfe/pfe_hif.h
-@@ -0,0 +1,211 @@
-+/*
-+ * Copyright 2015-2016 Freescale Semiconductor, Inc.
-+ * Copyright 2017 NXP
-+ *
-+ * This program is free software; you can redistribute it and/or modify
-+ * it under the terms of the GNU General Public License as published by
-+ * the Free Software Foundation; either version 2 of the License, or
-+ * (at your option) any later version.
-+ *
-+ * This program is distributed in the hope that it will be useful,
-+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
-+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
-+ * GNU General Public License for more details.
-+ *
-+ * You should have received a copy of the GNU General Public License
-+ * along with this program. If not, see <http://www.gnu.org/licenses/>.
-+ */
-+
-+#ifndef _PFE_HIF_H_
-+#define _PFE_HIF_H_
-+
-+#include <linux/netdevice.h>
-+
-+#define HIF_NAPI_STATS
-+
-+#define HIF_CLIENT_QUEUES_MAX 16
-+#define HIF_RX_POLL_WEIGHT 64
-+
-+#define HIF_RX_PKT_MIN_SIZE 0x800 /* 2KB */
-+#define HIF_RX_PKT_MIN_SIZE_MASK ~(HIF_RX_PKT_MIN_SIZE - 1)
-+#define ROUND_MIN_RX_SIZE(_sz) (((_sz) + (HIF_RX_PKT_MIN_SIZE - 1)) \
-+ & HIF_RX_PKT_MIN_SIZE_MASK)
-+#define PRESENT_OFST_IN_PAGE(_buf) (((unsigned long int)(_buf) & (PAGE_SIZE \
-+ - 1)) & HIF_RX_PKT_MIN_SIZE_MASK)
-+
-+enum {
-+ NAPI_SCHED_COUNT = 0,
-+ NAPI_POLL_COUNT,
-+ NAPI_PACKET_COUNT,
-+ NAPI_DESC_COUNT,
-+ NAPI_FULL_BUDGET_COUNT,
-+ NAPI_CLIENT_FULL_COUNT,
-+ NAPI_MAX_COUNT
-+};
-+
-+/*
-+ * HIF_TX_DESC_NT value should be always greter than 4,
-+ * Otherwise HIF_TX_POLL_MARK will become zero.
-+ */
-+#define HIF_RX_DESC_NT 256
-+#define HIF_TX_DESC_NT 2048
-+
-+#define HIF_FIRST_BUFFER BIT(0)
-+#define HIF_LAST_BUFFER BIT(1)
-+#define HIF_DONT_DMA_MAP BIT(2)
-+#define HIF_DATA_VALID BIT(3)
-+#define HIF_TSO BIT(4)
-+
-+enum {
-+ PFE_CL_GEM0 = 0,
-+ PFE_CL_GEM1,
-+ HIF_CLIENTS_MAX
-+};
-+
-+/*structure to store client queue info */
-+struct hif_rx_queue {
-+ struct rx_queue_desc *base;
-+ u32 size;
-+ u32 write_idx;
-+};
-+
-+struct hif_tx_queue {
-+ struct tx_queue_desc *base;
-+ u32 size;
-+ u32 ack_idx;
-+};
-+
-+/*Structure to store the client info */
-+struct hif_client {
-+ int rx_qn;
-+ struct hif_rx_queue rx_q[HIF_CLIENT_QUEUES_MAX];
-+ int tx_qn;
-+ struct hif_tx_queue tx_q[HIF_CLIENT_QUEUES_MAX];
-+};
-+
-+/*HIF hardware buffer descriptor */
-+struct hif_desc {
-+ u32 ctrl;
-+ u32 status;
-+ u32 data;
-+ u32 next;
-+};
-+
-+struct __hif_desc {
-+ u32 ctrl;
-+ u32 status;
-+ u32 data;
-+};
-+
-+struct hif_desc_sw {
-+ dma_addr_t data;
-+ u16 len;
-+ u8 client_id;
-+ u8 q_no;
-+ u16 flags;
-+};
-+
-+struct hif_hdr {
-+ u8 client_id;
-+ u8 q_num;
-+ u16 client_ctrl;
-+ u16 client_ctrl1;
-+};
-+
-+struct __hif_hdr {
-+ union {
-+ struct hif_hdr hdr;
-+ u32 word[2];
-+ };
-+};
-+
-+struct hif_ipsec_hdr {
-+ u16 sa_handle[2];
-+} __packed;
-+
-+/* HIF_CTRL_TX... defines */
-+#define HIF_CTRL_TX_CHECKSUM BIT(2)
-+
-+/* HIF_CTRL_RX... defines */
-+#define HIF_CTRL_RX_OFFSET_OFST (24)
-+#define HIF_CTRL_RX_CHECKSUMMED BIT(2)
-+#define HIF_CTRL_RX_CONTINUED BIT(1)
-+
-+struct pfe_hif {
-+ /* To store registered clients in hif layer */
-+ struct hif_client client[HIF_CLIENTS_MAX];
-+ struct hif_shm *shm;
-+ int irq;
-+
-+ void *descr_baseaddr_v;
-+ unsigned long descr_baseaddr_p;
-+
-+ struct hif_desc *rx_base;
-+ u32 rx_ring_size;
-+ u32 rxtoclean_index;
-+ void *rx_buf_addr[HIF_RX_DESC_NT];
-+ int rx_buf_len[HIF_RX_DESC_NT];
-+ unsigned int qno;
-+ unsigned int client_id;
-+ unsigned int client_ctrl;
-+ unsigned int started;
-+
-+ struct hif_desc *tx_base;
-+ u32 tx_ring_size;
-+ u32 txtosend;
-+ u32 txtoclean;
-+ u32 txavail;
-+ u32 txtoflush;
-+ struct hif_desc_sw tx_sw_queue[HIF_TX_DESC_NT];
-+
-+/* tx_lock synchronizes hif packet tx as well as pfe_hif structure access */
-+ spinlock_t tx_lock;
-+/* lock synchronizes hif rx queue processing */
-+ spinlock_t lock;
-+ struct net_device dummy_dev;
-+ struct napi_struct napi;
-+ struct device *dev;
-+
-+#ifdef HIF_NAPI_STATS
-+ unsigned int napi_counters[NAPI_MAX_COUNT];
-+#endif
-+ struct tasklet_struct tx_cleanup_tasklet;
-+};
-+
-+void __hif_xmit_pkt(struct pfe_hif *hif, unsigned int client_id, unsigned int
-+ q_no, void *data, u32 len, unsigned int flags);
-+int hif_xmit_pkt(struct pfe_hif *hif, unsigned int client_id, unsigned int q_no,
-+ void *data, unsigned int len);
-+void __hif_tx_done_process(struct pfe_hif *hif, int count);
-+void hif_process_client_req(struct pfe_hif *hif, int req, int data1, int
-+ data2);
-+int pfe_hif_init(struct pfe *pfe);
-+void pfe_hif_exit(struct pfe *pfe);
-+void pfe_hif_rx_idle(struct pfe_hif *hif);
-+static inline void hif_tx_done_process(struct pfe_hif *hif, int count)
-+{
-+ spin_lock_bh(&hif->tx_lock);
-+ __hif_tx_done_process(hif, count);
-+ spin_unlock_bh(&hif->tx_lock);
-+}
-+
-+static inline void hif_tx_lock(struct pfe_hif *hif)
-+{
-+ spin_lock_bh(&hif->tx_lock);
-+}
-+
-+static inline void hif_tx_unlock(struct pfe_hif *hif)
-+{
-+ spin_unlock_bh(&hif->tx_lock);
-+}
-+
-+static inline int __hif_tx_avail(struct pfe_hif *hif)
-+{
-+ return hif->txavail;
-+}
-+
-+#define __memcpy8(dst, src) memcpy(dst, src, 8)
-+#define __memcpy12(dst, src) memcpy(dst, src, 12)
-+#define __memcpy(dst, src, len) memcpy(dst, src, len)
-+
-+#endif /* _PFE_HIF_H_ */
---- /dev/null
-+++ b/drivers/staging/fsl_ppfe/pfe_hif_lib.c
-@@ -0,0 +1,640 @@
-+/*
-+ * Copyright 2015-2016 Freescale Semiconductor, Inc.
-+ * Copyright 2017 NXP
-+ *
-+ * This program is free software; you can redistribute it and/or modify
-+ * it under the terms of the GNU General Public License as published by
-+ * the Free Software Foundation; either version 2 of the License, or
-+ * (at your option) any later version.
-+ *
-+ * This program is distributed in the hope that it will be useful,
-+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
-+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
-+ * GNU General Public License for more details.
-+ *
-+ * You should have received a copy of the GNU General Public License
-+ * along with this program. If not, see <http://www.gnu.org/licenses/>.
-+ */
-+
-+#include <linux/version.h>
-+#include <linux/kernel.h>
-+#include <linux/slab.h>
-+#include <linux/interrupt.h>
-+#include <linux/workqueue.h>
-+#include <linux/dma-mapping.h>
-+#include <linux/dmapool.h>
-+#include <linux/sched.h>
-+#include <linux/skbuff.h>
-+#include <linux/moduleparam.h>
-+#include <linux/cpu.h>
-+
-+#include "pfe_mod.h"
-+#include "pfe_hif.h"
-+#include "pfe_hif_lib.h"
-+
-+unsigned int lro_mode;
-+unsigned int page_mode;
-+unsigned int tx_qos = 1;
-+module_param(tx_qos, uint, 0444);
-+MODULE_PARM_DESC(tx_qos, "0: disable ,\n"
-+ "1: enable (default), guarantee no packet drop at TMU level\n");
-+unsigned int pfe_pkt_size;
-+unsigned int pfe_pkt_headroom;
-+unsigned int emac_txq_cnt;
-+
-+/*
-+ * @pfe_hal_lib.c.
-+ * Common functions used by HIF client drivers
-+ */
-+
-+/*HIF shared memory Global variable */
-+struct hif_shm ghif_shm;
-+
-+/* Cleanup the HIF shared memory, release HIF rx_buffer_pool.
-+ * This function should be called after pfe_hif_exit
-+ *
-+ * @param[in] hif_shm Shared memory address location in DDR
-+ */
-+static void pfe_hif_shm_clean(struct hif_shm *hif_shm)
-+{
-+ int i;
-+ void *pkt;
-+
-+ for (i = 0; i < hif_shm->rx_buf_pool_cnt; i++) {
-+ pkt = hif_shm->rx_buf_pool[i];
-+ if (pkt) {
-+ hif_shm->rx_buf_pool[i] = NULL;
-+ pkt -= pfe_pkt_headroom;
-+
-+ if (page_mode)
-+ put_page(virt_to_page(pkt));
-+ else
-+ kfree(pkt);
-+ }
-+ }
-+}
-+
-+/* Initialize shared memory used between HIF driver and clients,
-+ * allocate rx_buffer_pool required for HIF Rx descriptors.
-+ * This function should be called before initializing HIF driver.
-+ *
-+ * @param[in] hif_shm Shared memory address location in DDR
-+ * @rerurn 0 - on succes, <0 on fail to initialize
-+ */
-+static int pfe_hif_shm_init(struct hif_shm *hif_shm)
-+{
-+ int i;
-+ void *pkt;
-+
-+ memset(hif_shm, 0, sizeof(struct hif_shm));
-+ hif_shm->rx_buf_pool_cnt = HIF_RX_DESC_NT;
-+
-+ for (i = 0; i < hif_shm->rx_buf_pool_cnt; i++) {
-+ if (page_mode) {
-+ pkt = (void *)__get_free_page(GFP_KERNEL |
-+ GFP_DMA_PFE);
-+ } else {
-+ pkt = kmalloc(PFE_BUF_SIZE, GFP_KERNEL | GFP_DMA_PFE);
-+ }
-+
-+ if (pkt)
-+ hif_shm->rx_buf_pool[i] = pkt + pfe_pkt_headroom;
-+ else
-+ goto err0;
-+ }
-+
-+ return 0;
-+
-+err0:
-+ pr_err("%s Low memory\n", __func__);
-+ pfe_hif_shm_clean(hif_shm);
-+ return -ENOMEM;
-+}
-+
-+/*This function sends indication to HIF driver
-+ *
-+ * @param[in] hif hif context
-+ */
-+static void hif_lib_indicate_hif(struct pfe_hif *hif, int req, int data1, int
-+ data2)
-+{
-+ hif_process_client_req(hif, req, data1, data2);
-+}
-+
-+void hif_lib_indicate_client(int client_id, int event_type, int qno)
-+{
-+ struct hif_client_s *client = pfe->hif_client[client_id];
-+
-+ if (!client || (event_type >= HIF_EVENT_MAX) || (qno >=
-+ HIF_CLIENT_QUEUES_MAX))
-+ return;
-+
-+ if (!test_and_set_bit(qno, &client->queue_mask[event_type]))
-+ client->event_handler(client->priv, event_type, qno);
-+}
-+
-+/*This function releases Rx queue descriptors memory and pre-filled buffers
-+ *
-+ * @param[in] client hif_client context
-+ */
-+static void hif_lib_client_release_rx_buffers(struct hif_client_s *client)
-+{
-+ struct rx_queue_desc *desc;
-+ int qno, ii;
-+ void *buf;
-+
-+ for (qno = 0; qno < client->rx_qn; qno++) {
-+ desc = client->rx_q[qno].base;
-+
-+ for (ii = 0; ii < client->rx_q[qno].size; ii++) {
-+ buf = (void *)desc->data;
-+ if (buf) {
-+ buf -= pfe_pkt_headroom;
-+
-+ if (page_mode)
-+ free_page((unsigned long)buf);
-+ else
-+ kfree(buf);
-+
-+ desc->ctrl = 0;
-+ }
-+
-+ desc++;
-+ }
-+ }
-+
-+ kfree(client->rx_qbase);
-+}
-+
-+/*This function allocates memory for the rxq descriptors and pre-fill rx queues
-+ * with buffers.
-+ * @param[in] client client context
-+ * @param[in] q_size size of the rxQ, all queues are of same size
-+ */
-+static int hif_lib_client_init_rx_buffers(struct hif_client_s *client, int
-+ q_size)
-+{
-+ struct rx_queue_desc *desc;
-+ struct hif_client_rx_queue *queue;
-+ int ii, qno;
-+
-+ /*Allocate memory for the client queues */
-+ client->rx_qbase = kzalloc(client->rx_qn * q_size * sizeof(struct
-+ rx_queue_desc), GFP_KERNEL);
-+ if (!client->rx_qbase)
-+ goto err;
-+
-+ for (qno = 0; qno < client->rx_qn; qno++) {
-+ queue = &client->rx_q[qno];
-+
-+ queue->base = client->rx_qbase + qno * q_size * sizeof(struct
-+ rx_queue_desc);
-+ queue->size = q_size;
-+ queue->read_idx = 0;
-+ queue->write_idx = 0;
-+
-+ pr_debug("rx queue: %d, base: %p, size: %d\n", qno,
-+ queue->base, queue->size);
-+ }
-+
-+ for (qno = 0; qno < client->rx_qn; qno++) {
-+ queue = &client->rx_q[qno];
-+ desc = queue->base;
-+
-+ for (ii = 0; ii < queue->size; ii++) {
-+ desc->ctrl = CL_DESC_BUF_LEN(pfe_pkt_size) |
-+ CL_DESC_OWN;
-+ desc++;
-+ }
-+ }
-+
-+ return 0;
-+
-+err:
-+ return 1;
-+}
-+
-+
-+static void hif_lib_client_cleanup_tx_queue(struct hif_client_tx_queue *queue)
-+{
-+ pr_debug("%s\n", __func__);
-+
-+ /*
-+ * Check if there are any pending packets. Client must flush the tx
-+ * queues before unregistering, by calling by calling
-+ * hif_lib_tx_get_next_complete()
-+ *
-+ * Hif no longer calls since we are no longer registered
-+ */
-+ if (queue->tx_pending)
-+ pr_err("%s: pending transmit packets\n", __func__);
-+}
-+
-+static void hif_lib_client_release_tx_buffers(struct hif_client_s *client)
-+{
-+ int qno;
-+
-+ pr_debug("%s\n", __func__);
-+
-+ for (qno = 0; qno < client->tx_qn; qno++)
-+ hif_lib_client_cleanup_tx_queue(&client->tx_q[qno]);
-+
-+ kfree(client->tx_qbase);
-+}
-+
-+static int hif_lib_client_init_tx_buffers(struct hif_client_s *client, int
-+ q_size)
-+{
-+ struct hif_client_tx_queue *queue;
-+ int qno;
-+
-+ client->tx_qbase = kzalloc(client->tx_qn * q_size * sizeof(struct
-+ tx_queue_desc), GFP_KERNEL);
-+ if (!client->tx_qbase)
-+ return 1;
-+
-+ for (qno = 0; qno < client->tx_qn; qno++) {
-+ queue = &client->tx_q[qno];
-+
-+ queue->base = client->tx_qbase + qno * q_size * sizeof(struct
-+ tx_queue_desc);
-+ queue->size = q_size;
-+ queue->read_idx = 0;
-+ queue->write_idx = 0;
-+ queue->tx_pending = 0;
-+ queue->nocpy_flag = 0;
-+ queue->prev_tmu_tx_pkts = 0;
-+ queue->done_tmu_tx_pkts = 0;
-+
-+ pr_debug("tx queue: %d, base: %p, size: %d\n", qno,
-+ queue->base, queue->size);
-+ }
-+
-+ return 0;
-+}
-+
-+static int hif_lib_event_dummy(void *priv, int event_type, int qno)
-+{
-+ return 0;
-+}
-+
-+int hif_lib_client_register(struct hif_client_s *client)
-+{
-+ struct hif_shm *hif_shm;
-+ struct hif_client_shm *client_shm;
-+ int err, i;
-+ /* int loop_cnt = 0; */
-+
-+ pr_debug("%s\n", __func__);
-+
-+ /*Allocate memory before spin_lock*/
-+ if (hif_lib_client_init_rx_buffers(client, client->rx_qsize)) {
-+ err = -ENOMEM;
-+ goto err_rx;
-+ }
-+
-+ if (hif_lib_client_init_tx_buffers(client, client->tx_qsize)) {
-+ err = -ENOMEM;
-+ goto err_tx;
-+ }
-+
-+ spin_lock_bh(&pfe->hif.lock);
-+ if (!(client->pfe) || (client->id >= HIF_CLIENTS_MAX) ||
-+ (pfe->hif_client[client->id])) {
-+ err = -EINVAL;
-+ goto err;
-+ }
-+
-+ hif_shm = client->pfe->hif.shm;
-+
-+ if (!client->event_handler)
-+ client->event_handler = hif_lib_event_dummy;
-+
-+ /*Initialize client specific shared memory */
-+ client_shm = (struct hif_client_shm *)&hif_shm->client[client->id];
-+ client_shm->rx_qbase = (unsigned long int)client->rx_qbase;
-+ client_shm->rx_qsize = client->rx_qsize;
-+ client_shm->tx_qbase = (unsigned long int)client->tx_qbase;
-+ client_shm->tx_qsize = client->tx_qsize;
-+ client_shm->ctrl = (client->tx_qn << CLIENT_CTRL_TX_Q_CNT_OFST) |
-+ (client->rx_qn << CLIENT_CTRL_RX_Q_CNT_OFST);
-+ /* spin_lock_init(&client->rx_lock); */
-+
-+ for (i = 0; i < HIF_EVENT_MAX; i++) {
-+ client->queue_mask[i] = 0; /*
-+ * By default all events are
-+ * unmasked
-+ */
-+ }
-+
-+ /*Indicate to HIF driver*/
-+ hif_lib_indicate_hif(&pfe->hif, REQUEST_CL_REGISTER, client->id, 0);
-+
-+ pr_debug("%s: client: %p, client_id: %d, tx_qsize: %d, rx_qsize: %d\n",
-+ __func__, client, client->id, client->tx_qsize,
-+ client->rx_qsize);
-+
-+ client->cpu_id = -1;
-+
-+ pfe->hif_client[client->id] = client;
-+ spin_unlock_bh(&pfe->hif.lock);
-+
-+ return 0;
-+
-+err:
-+ spin_unlock_bh(&pfe->hif.lock);
-+ hif_lib_client_release_tx_buffers(client);
-+
-+err_tx:
-+ hif_lib_client_release_rx_buffers(client);
-+
-+err_rx:
-+ return err;
-+}
-+
-+int hif_lib_client_unregister(struct hif_client_s *client)
-+{
-+ struct pfe *pfe = client->pfe;
-+ u32 client_id = client->id;
-+
-+ pr_info(
-+ "%s : client: %p, client_id: %d, txQ_depth: %d, rxQ_depth: %d\n"
-+ , __func__, client, client->id, client->tx_qsize,
-+ client->rx_qsize);
-+
-+ spin_lock_bh(&pfe->hif.lock);
-+ hif_lib_indicate_hif(&pfe->hif, REQUEST_CL_UNREGISTER, client->id, 0);
-+
-+ hif_lib_client_release_tx_buffers(client);
-+ hif_lib_client_release_rx_buffers(client);
-+ pfe->hif_client[client_id] = NULL;
-+ spin_unlock_bh(&pfe->hif.lock);
-+
-+ return 0;
-+}
-+
-+int hif_lib_event_handler_start(struct hif_client_s *client, int event,
-+ int qno)
-+{
-+ struct hif_client_rx_queue *queue = &client->rx_q[qno];
-+ struct rx_queue_desc *desc = queue->base + queue->read_idx;
-+
-+ if ((event >= HIF_EVENT_MAX) || (qno >= HIF_CLIENT_QUEUES_MAX)) {
-+ pr_debug("%s: Unsupported event : %d queue number : %d\n",
-+ __func__, event, qno);
-+ return -1;
-+ }
-+
-+ test_and_clear_bit(qno, &client->queue_mask[event]);
-+
-+ switch (event) {
-+ case EVENT_RX_PKT_IND:
-+ if (!(desc->ctrl & CL_DESC_OWN))
-+ hif_lib_indicate_client(client->id,
-+ EVENT_RX_PKT_IND, qno);
-+ break;
-+
-+ case EVENT_HIGH_RX_WM:
-+ case EVENT_TXDONE_IND:
-+ default:
-+ break;
-+ }
-+
-+ return 0;
-+}
-+
-+/*
-+ * This function gets one packet from the specified client queue
-+ * It also refill the rx buffer
-+ */
-+void *hif_lib_receive_pkt(struct hif_client_s *client, int qno, int *len, int
-+ *ofst, unsigned int *rx_ctrl,
-+ unsigned int *desc_ctrl, void **priv_data)
-+{
-+ struct hif_client_rx_queue *queue = &client->rx_q[qno];
-+ struct rx_queue_desc *desc;
-+ void *pkt = NULL;
-+
-+ /*
-+ * Following lock is to protect rx queue access from,
-+ * hif_lib_event_handler_start.
-+ * In general below lock is not required, because hif_lib_xmit_pkt and
-+ * hif_lib_event_handler_start are called from napi poll and which is
-+ * not re-entrant. But if some client use in different way this lock is
-+ * required.
-+ */
-+ /*spin_lock_irqsave(&client->rx_lock, flags); */
-+ desc = queue->base + queue->read_idx;
-+ if (!(desc->ctrl & CL_DESC_OWN)) {
-+ pkt = desc->data - pfe_pkt_headroom;
-+
-+ *rx_ctrl = desc->client_ctrl;
-+ *desc_ctrl = desc->ctrl;
-+
-+ if (desc->ctrl & CL_DESC_FIRST) {
-+ u16 size = *rx_ctrl >> HIF_CTRL_RX_OFFSET_OFST;
-+
-+ if (size) {
-+ size += PFE_PARSE_INFO_SIZE;
-+ *len = CL_DESC_BUF_LEN(desc->ctrl) -
-+ PFE_PKT_HEADER_SZ - size;
-+ *ofst = pfe_pkt_headroom + PFE_PKT_HEADER_SZ
-+ + size;
-+ *priv_data = desc->data + PFE_PKT_HEADER_SZ;
-+ } else {
-+ *len = CL_DESC_BUF_LEN(desc->ctrl) -
-+ PFE_PKT_HEADER_SZ - PFE_PARSE_INFO_SIZE;
-+ *ofst = pfe_pkt_headroom
-+ + PFE_PKT_HEADER_SZ
-+ + PFE_PARSE_INFO_SIZE;
-+ *priv_data = NULL;
-+ }
-+
-+ } else {
-+ *len = CL_DESC_BUF_LEN(desc->ctrl);
-+ *ofst = pfe_pkt_headroom;
-+ }
-+
-+ /*
-+ * Needed so we don't free a buffer/page
-+ * twice on module_exit
-+ */
-+ desc->data = NULL;
-+
-+ /*
-+ * Ensure everything else is written to DDR before
-+ * writing bd->ctrl
-+ */
-+ smp_wmb();
-+
-+ desc->ctrl = CL_DESC_BUF_LEN(pfe_pkt_size) | CL_DESC_OWN;
-+ queue->read_idx = (queue->read_idx + 1) & (queue->size - 1);
-+ }
-+
-+ /*spin_unlock_irqrestore(&client->rx_lock, flags); */
-+ return pkt;
-+}
-+
-+static inline void hif_hdr_write(struct hif_hdr *pkt_hdr, unsigned int
-+ client_id, unsigned int qno,
-+ u32 client_ctrl)
-+{
-+ /* Optimize the write since the destinaton may be non-cacheable */
-+ if (!((unsigned long)pkt_hdr & 0x3)) {
-+ ((u32 *)pkt_hdr)[0] = (client_ctrl << 16) | (qno << 8) |
-+ client_id;
-+ } else {
-+ ((u16 *)pkt_hdr)[0] = (qno << 8) | (client_id & 0xFF);
-+ ((u16 *)pkt_hdr)[1] = (client_ctrl & 0xFFFF);
-+ }
-+}
-+
-+/*This function puts the given packet in the specific client queue */
-+void __hif_lib_xmit_pkt(struct hif_client_s *client, unsigned int qno, void
-+ *data, unsigned int len, u32 client_ctrl,
-+ unsigned int flags, void *client_data)
-+{
-+ struct hif_client_tx_queue *queue = &client->tx_q[qno];
-+ struct tx_queue_desc *desc = queue->base + queue->write_idx;
-+
-+ /* First buffer */
-+ if (flags & HIF_FIRST_BUFFER) {
-+ data -= sizeof(struct hif_hdr);
-+ len += sizeof(struct hif_hdr);
-+
-+ hif_hdr_write(data, client->id, qno, client_ctrl);
-+ }
-+
-+ desc->data = client_data;
-+ desc->ctrl = CL_DESC_OWN | CL_DESC_FLAGS(flags);
-+
-+ __hif_xmit_pkt(&pfe->hif, client->id, qno, data, len, flags);
-+
-+ queue->write_idx = (queue->write_idx + 1) & (queue->size - 1);
-+ queue->tx_pending++;
-+ queue->jiffies_last_packet = jiffies;
-+}
-+
-+void *hif_lib_tx_get_next_complete(struct hif_client_s *client, int qno,
-+ unsigned int *flags, int count)
-+{
-+ struct hif_client_tx_queue *queue = &client->tx_q[qno];
-+ struct tx_queue_desc *desc = queue->base + queue->read_idx;
-+
-+ pr_debug("%s: qno : %d rd_indx: %d pending:%d\n", __func__, qno,
-+ queue->read_idx, queue->tx_pending);
-+
-+ if (!queue->tx_pending)
-+ return NULL;
-+
-+ if (queue->nocpy_flag && !queue->done_tmu_tx_pkts) {
-+ u32 tmu_tx_pkts = be32_to_cpu(pe_dmem_read(TMU0_ID +
-+ client->id, TMU_DM_TX_TRANS, 4));
-+
-+ if (queue->prev_tmu_tx_pkts > tmu_tx_pkts)
-+ queue->done_tmu_tx_pkts = UINT_MAX -
-+ queue->prev_tmu_tx_pkts + tmu_tx_pkts;
-+ else
-+ queue->done_tmu_tx_pkts = tmu_tx_pkts -
-+ queue->prev_tmu_tx_pkts;
-+
-+ queue->prev_tmu_tx_pkts = tmu_tx_pkts;
-+
-+ if (!queue->done_tmu_tx_pkts)
-+ return NULL;
-+ }
-+
-+ if (desc->ctrl & CL_DESC_OWN)
-+ return NULL;
-+
-+ queue->read_idx = (queue->read_idx + 1) & (queue->size - 1);
-+ queue->tx_pending--;
-+
-+ *flags = CL_DESC_GET_FLAGS(desc->ctrl);
-+
-+ if (queue->done_tmu_tx_pkts && (*flags & HIF_LAST_BUFFER))
-+ queue->done_tmu_tx_pkts--;
-+
-+ return desc->data;
-+}
-+
-+static void hif_lib_tmu_credit_init(struct pfe *pfe)
-+{
-+ int i, q;
-+
-+ for (i = 0; i < NUM_GEMAC_SUPPORT; i++)
-+ for (q = 0; q < emac_txq_cnt; q++) {
-+ pfe->tmu_credit.tx_credit_max[i][q] = (q == 0) ?
-+ DEFAULT_Q0_QDEPTH : DEFAULT_MAX_QDEPTH;
-+ pfe->tmu_credit.tx_credit[i][q] =
-+ pfe->tmu_credit.tx_credit_max[i][q];
-+ }
-+}
-+
-+/* __hif_lib_update_credit
-+ *
-+ * @param[in] client hif client context
-+ * @param[in] queue queue number in match with TMU
-+ */
-+void __hif_lib_update_credit(struct hif_client_s *client, unsigned int queue)
-+{
-+ unsigned int tmu_tx_packets, tmp;
-+
-+ if (tx_qos) {
-+ tmu_tx_packets = be32_to_cpu(pe_dmem_read(TMU0_ID +
-+ client->id, (TMU_DM_TX_TRANS + (queue * 4)), 4));
-+
-+ /* tx_packets counter overflowed */
-+ if (tmu_tx_packets >
-+ pfe->tmu_credit.tx_packets[client->id][queue]) {
-+ tmp = UINT_MAX - tmu_tx_packets +
-+ pfe->tmu_credit.tx_packets[client->id][queue];
-+
-+ pfe->tmu_credit.tx_credit[client->id][queue] =
-+ pfe->tmu_credit.tx_credit_max[client->id][queue] - tmp;
-+ } else {
-+ /* TMU tx <= pfe_eth tx, normal case or both OF since
-+ * last time
-+ */
-+ pfe->tmu_credit.tx_credit[client->id][queue] =
-+ pfe->tmu_credit.tx_credit_max[client->id][queue] -
-+ (pfe->tmu_credit.tx_packets[client->id][queue] -
-+ tmu_tx_packets);
-+ }
-+ }
-+}
-+
-+int pfe_hif_lib_init(struct pfe *pfe)
-+{
-+ int rc;
-+
-+ pr_info("%s\n", __func__);
-+
-+ if (lro_mode) {
-+ page_mode = 1;
-+ pfe_pkt_size = min(PAGE_SIZE, MAX_PFE_PKT_SIZE);
-+ pfe_pkt_headroom = 0;
-+ } else {
-+ page_mode = 0;
-+ pfe_pkt_size = PFE_PKT_SIZE;
-+ pfe_pkt_headroom = PFE_PKT_HEADROOM;
-+ }
-+
-+ if (tx_qos)
-+ emac_txq_cnt = EMAC_TXQ_CNT / 2;
-+ else
-+ emac_txq_cnt = EMAC_TXQ_CNT;
-+
-+ hif_lib_tmu_credit_init(pfe);
-+ pfe->hif.shm = &ghif_shm;
-+ rc = pfe_hif_shm_init(pfe->hif.shm);
-+
-+ return rc;
-+}
-+
-+void pfe_hif_lib_exit(struct pfe *pfe)
-+{
-+ pr_info("%s\n", __func__);
-+
-+ pfe_hif_shm_clean(pfe->hif.shm);
-+}
---- /dev/null
-+++ b/drivers/staging/fsl_ppfe/pfe_hif_lib.h
-@@ -0,0 +1,241 @@
-+/*
-+ * Copyright 2015-2016 Freescale Semiconductor, Inc.
-+ * Copyright 2017 NXP
-+ *
-+ * This program is free software; you can redistribute it and/or modify
-+ * it under the terms of the GNU General Public License as published by
-+ * the Free Software Foundation; either version 2 of the License, or
-+ * (at your option) any later version.
-+ *
-+ * This program is distributed in the hope that it will be useful,
-+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
-+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
-+ * GNU General Public License for more details.
-+ *
-+ * You should have received a copy of the GNU General Public License
-+ * along with this program. If not, see <http://www.gnu.org/licenses/>.
-+ */
-+
-+#ifndef _PFE_HIF_LIB_H_
-+#define _PFE_HIF_LIB_H_
-+
-+#include "pfe_hif.h"
-+
-+#define HIF_CL_REQ_TIMEOUT 10
-+#define GFP_DMA_PFE 0
-+#define PFE_PARSE_INFO_SIZE 16
-+
-+enum {
-+ REQUEST_CL_REGISTER = 0,
-+ REQUEST_CL_UNREGISTER,
-+ HIF_REQUEST_MAX
-+};
-+
-+enum {
-+ /* Event to indicate that client rx queue is reached water mark level */
-+ EVENT_HIGH_RX_WM = 0,
-+ /* Event to indicate that, packet received for client */
-+ EVENT_RX_PKT_IND,
-+ /* Event to indicate that, packet tx done for client */
-+ EVENT_TXDONE_IND,
-+ HIF_EVENT_MAX
-+};
-+
-+/*structure to store client queue info */
-+
-+/*structure to store client queue info */
-+struct hif_client_rx_queue {
-+ struct rx_queue_desc *base;
-+ u32 size;
-+ u32 read_idx;
-+ u32 write_idx;
-+};
-+
-+struct hif_client_tx_queue {
-+ struct tx_queue_desc *base;
-+ u32 size;
-+ u32 read_idx;
-+ u32 write_idx;
-+ u32 tx_pending;
-+ unsigned long jiffies_last_packet;
-+ u32 nocpy_flag;
-+ u32 prev_tmu_tx_pkts;
-+ u32 done_tmu_tx_pkts;
-+};
-+
-+struct hif_client_s {
-+ int id;
-+ int tx_qn;
-+ int rx_qn;
-+ void *rx_qbase;
-+ void *tx_qbase;
-+ int tx_qsize;
-+ int rx_qsize;
-+ int cpu_id;
-+ struct hif_client_tx_queue tx_q[HIF_CLIENT_QUEUES_MAX];
-+ struct hif_client_rx_queue rx_q[HIF_CLIENT_QUEUES_MAX];
-+ int (*event_handler)(void *priv, int event, int data);
-+ unsigned long queue_mask[HIF_EVENT_MAX];
-+ struct pfe *pfe;
-+ void *priv;
-+};
-+
-+/*
-+ * Client specific shared memory
-+ * It contains number of Rx/Tx queues, base addresses and queue sizes
-+ */
-+struct hif_client_shm {
-+ u32 ctrl; /*0-7: number of Rx queues, 8-15: number of tx queues */
-+ unsigned long rx_qbase; /*Rx queue base address */
-+ u32 rx_qsize; /*each Rx queue size, all Rx queues are of same size */
-+ unsigned long tx_qbase; /* Tx queue base address */
-+ u32 tx_qsize; /*each Tx queue size, all Tx queues are of same size */
-+};
-+
-+/*Client shared memory ctrl bit description */
-+#define CLIENT_CTRL_RX_Q_CNT_OFST 0
-+#define CLIENT_CTRL_TX_Q_CNT_OFST 8
-+#define CLIENT_CTRL_RX_Q_CNT(ctrl) (((ctrl) >> CLIENT_CTRL_RX_Q_CNT_OFST) \
-+ & 0xFF)
-+#define CLIENT_CTRL_TX_Q_CNT(ctrl) (((ctrl) >> CLIENT_CTRL_TX_Q_CNT_OFST) \
-+ & 0xFF)
-+
-+/*
-+ * Shared memory used to communicate between HIF driver and host/client drivers
-+ * Before starting the hif driver rx_buf_pool ans rx_buf_pool_cnt should be
-+ * initialized with host buffers and buffers count in the pool.
-+ * rx_buf_pool_cnt should be >= HIF_RX_DESC_NT.
-+ *
-+ */
-+struct hif_shm {
-+ u32 rx_buf_pool_cnt; /*Number of rx buffers available*/
-+ /*Rx buffers required to initialize HIF rx descriptors */
-+ void *rx_buf_pool[HIF_RX_DESC_NT];
-+ unsigned long g_client_status[2]; /*Global client status bit mask */
-+ /* Client specific shared memory */
-+ struct hif_client_shm client[HIF_CLIENTS_MAX];
-+};
-+
-+#define CL_DESC_OWN BIT(31)
-+/* This sets owner ship to HIF driver */
-+#define CL_DESC_LAST BIT(30)
-+/* This indicates last packet for multi buffers handling */
-+#define CL_DESC_FIRST BIT(29)
-+/* This indicates first packet for multi buffers handling */
-+
-+#define CL_DESC_BUF_LEN(x) ((x) & 0xFFFF)
-+#define CL_DESC_FLAGS(x) (((x) & 0xF) << 16)
-+#define CL_DESC_GET_FLAGS(x) (((x) >> 16) & 0xF)
-+
-+struct rx_queue_desc {
-+ void *data;
-+ u32 ctrl; /*0-15bit len, 16-20bit flags, 31bit owner*/
-+ u32 client_ctrl;
-+};
-+
-+struct tx_queue_desc {
-+ void *data;
-+ u32 ctrl; /*0-15bit len, 16-20bit flags, 31bit owner*/
-+};
-+
-+/* HIF Rx is not working properly for 2-byte aligned buffers and
-+ * ip_header should be 4byte aligned for better iperformance.
-+ * "ip_header = 64 + 6(hif_header) + 14 (MAC Header)" will be 4byte aligned.
-+ */
-+#define PFE_PKT_HEADER_SZ sizeof(struct hif_hdr)
-+/* must be big enough for headroom, pkt size and skb shared info */
-+#define PFE_BUF_SIZE 2048
-+#define PFE_PKT_HEADROOM 128
-+
-+#define SKB_SHARED_INFO_SIZE SKB_DATA_ALIGN(sizeof(struct skb_shared_info))
-+#define PFE_PKT_SIZE (PFE_BUF_SIZE - PFE_PKT_HEADROOM \
-+ - SKB_SHARED_INFO_SIZE)
-+#define MAX_L2_HDR_SIZE 14 /* Not correct for VLAN/PPPoE */
-+#define MAX_L3_HDR_SIZE 20 /* Not correct for IPv6 */
-+#define MAX_L4_HDR_SIZE 60 /* TCP with maximum options */
-+#define MAX_HDR_SIZE (MAX_L2_HDR_SIZE + MAX_L3_HDR_SIZE \
-+ + MAX_L4_HDR_SIZE)
-+/* Used in page mode to clamp packet size to the maximum supported by the hif
-+ *hw interface (<16KiB)
-+ */
-+#define MAX_PFE_PKT_SIZE 16380UL
-+
-+extern unsigned int pfe_pkt_size;
-+extern unsigned int pfe_pkt_headroom;
-+extern unsigned int page_mode;
-+extern unsigned int lro_mode;
-+extern unsigned int tx_qos;
-+extern unsigned int emac_txq_cnt;
-+
-+int pfe_hif_lib_init(struct pfe *pfe);
-+void pfe_hif_lib_exit(struct pfe *pfe);
-+int hif_lib_client_register(struct hif_client_s *client);
-+int hif_lib_client_unregister(struct hif_client_s *client);
-+void __hif_lib_xmit_pkt(struct hif_client_s *client, unsigned int qno, void
-+ *data, unsigned int len, u32 client_ctrl,
-+ unsigned int flags, void *client_data);
-+int hif_lib_xmit_pkt(struct hif_client_s *client, unsigned int qno, void *data,
-+ unsigned int len, u32 client_ctrl, void *client_data);
-+void hif_lib_indicate_client(int cl_id, int event, int data);
-+int hif_lib_event_handler_start(struct hif_client_s *client, int event, int
-+ data);
-+int hif_lib_tmu_queue_start(struct hif_client_s *client, int qno);
-+int hif_lib_tmu_queue_stop(struct hif_client_s *client, int qno);
-+void *hif_lib_tx_get_next_complete(struct hif_client_s *client, int qno,
-+ unsigned int *flags, int count);
-+void *hif_lib_receive_pkt(struct hif_client_s *client, int qno, int *len, int
-+ *ofst, unsigned int *rx_ctrl,
-+ unsigned int *desc_ctrl, void **priv_data);
-+void __hif_lib_update_credit(struct hif_client_s *client, unsigned int queue);
-+void hif_lib_set_rx_cpu_affinity(struct hif_client_s *client, int cpu_id);
-+void hif_lib_set_tx_queue_nocpy(struct hif_client_s *client, int qno, int
-+ enable);
-+static inline int hif_lib_tx_avail(struct hif_client_s *client, unsigned int
-+ qno)
-+{
-+ struct hif_client_tx_queue *queue = &client->tx_q[qno];
-+
-+ return (queue->size - queue->tx_pending);
-+}
-+
-+static inline int hif_lib_get_tx_wr_index(struct hif_client_s *client, unsigned
-+ int qno)
-+{
-+ struct hif_client_tx_queue *queue = &client->tx_q[qno];
-+
-+ return queue->write_idx;
-+}
-+
-+static inline int hif_lib_tx_pending(struct hif_client_s *client, unsigned int
-+ qno)
-+{
-+ struct hif_client_tx_queue *queue = &client->tx_q[qno];
-+
-+ return queue->tx_pending;
-+}
-+
-+#define hif_lib_tx_credit_avail(pfe, id, qno) \
-+ ((pfe)->tmu_credit.tx_credit[id][qno])
-+
-+#define hif_lib_tx_credit_max(pfe, id, qno) \
-+ ((pfe)->tmu_credit.tx_credit_max[id][qno])
-+
-+/*
-+ * Test comment
-+ */
-+#define hif_lib_tx_credit_use(pfe, id, qno, credit) \
-+ ({ typeof(pfe) pfe_ = pfe; \
-+ typeof(id) id_ = id; \
-+ typeof(qno) qno_ = qno_; \
-+ typeof(credit) credit_ = credit; \
-+ do { \
-+ if (tx_qos) { \
-+ (pfe_)->tmu_credit.tx_credit[id_][qno_]\
-+ -= credit_; \
-+ (pfe_)->tmu_credit.tx_packets[id_][qno_]\
-+ += credit_; \
-+ } \
-+ } while (0); \
-+ })
-+
-+#endif /* _PFE_HIF_LIB_H_ */
---- /dev/null
-+++ b/drivers/staging/fsl_ppfe/pfe_hw.c
-@@ -0,0 +1,176 @@
-+/*
-+ * Copyright 2015-2016 Freescale Semiconductor, Inc.
-+ * Copyright 2017 NXP
-+ *
-+ * This program is free software; you can redistribute it and/or modify
-+ * it under the terms of the GNU General Public License as published by
-+ * the Free Software Foundation; either version 2 of the License, or
-+ * (at your option) any later version.
-+ *
-+ * This program is distributed in the hope that it will be useful,
-+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
-+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
-+ * GNU General Public License for more details.
-+ *
-+ * You should have received a copy of the GNU General Public License
-+ * along with this program. If not, see <http://www.gnu.org/licenses/>.
-+ */
-+
-+#include "pfe_mod.h"
-+#include "pfe_hw.h"
-+
-+/* Functions to handle most of pfe hw register initialization */
-+int pfe_hw_init(struct pfe *pfe, int resume)
-+{
-+ struct class_cfg class_cfg = {
-+ .pe_sys_clk_ratio = PE_SYS_CLK_RATIO,
-+ .route_table_baseaddr = pfe->ddr_phys_baseaddr +
-+ ROUTE_TABLE_BASEADDR,
-+ .route_table_hash_bits = ROUTE_TABLE_HASH_BITS,
-+ };
-+
-+ struct tmu_cfg tmu_cfg = {
-+ .pe_sys_clk_ratio = PE_SYS_CLK_RATIO,
-+ .llm_base_addr = pfe->ddr_phys_baseaddr + TMU_LLM_BASEADDR,
-+ .llm_queue_len = TMU_LLM_QUEUE_LEN,
-+ };
-+
-+#if !defined(CONFIG_FSL_PPFE_UTIL_DISABLED)
-+ struct util_cfg util_cfg = {
-+ .pe_sys_clk_ratio = PE_SYS_CLK_RATIO,
-+ };
-+#endif
-+
-+ struct BMU_CFG bmu1_cfg = {
-+ .baseaddr = CBUS_VIRT_TO_PFE(LMEM_BASE_ADDR +
-+ BMU1_LMEM_BASEADDR),
-+ .count = BMU1_BUF_COUNT,
-+ .size = BMU1_BUF_SIZE,
-+ .low_watermark = 10,
-+ .high_watermark = 15,
-+ };
-+
-+ struct BMU_CFG bmu2_cfg = {
-+ .baseaddr = DDR_PHYS_TO_PFE(pfe->ddr_phys_baseaddr +
-+ BMU2_DDR_BASEADDR),
-+ .count = BMU2_BUF_COUNT,
-+ .size = BMU2_BUF_SIZE,
-+ .low_watermark = 250,
-+ .high_watermark = 253,
-+ };
-+
-+ struct gpi_cfg egpi1_cfg = {
-+ .lmem_rtry_cnt = EGPI1_LMEM_RTRY_CNT,
-+ .tmlf_txthres = EGPI1_TMLF_TXTHRES,
-+ .aseq_len = EGPI1_ASEQ_LEN,
-+ .mtip_pause_reg = CBUS_VIRT_TO_PFE(EMAC1_BASE_ADDR +
-+ EMAC_TCNTRL_REG),
-+ };
-+
-+ struct gpi_cfg egpi2_cfg = {
-+ .lmem_rtry_cnt = EGPI2_LMEM_RTRY_CNT,
-+ .tmlf_txthres = EGPI2_TMLF_TXTHRES,
-+ .aseq_len = EGPI2_ASEQ_LEN,
-+ .mtip_pause_reg = CBUS_VIRT_TO_PFE(EMAC2_BASE_ADDR +
-+ EMAC_TCNTRL_REG),
-+ };
-+
-+ struct gpi_cfg hgpi_cfg = {
-+ .lmem_rtry_cnt = HGPI_LMEM_RTRY_CNT,
-+ .tmlf_txthres = HGPI_TMLF_TXTHRES,
-+ .aseq_len = HGPI_ASEQ_LEN,
-+ .mtip_pause_reg = 0,
-+ };
-+
-+ pr_info("%s\n", __func__);
-+
-+#if !defined(LS1012A_PFE_RESET_WA)
-+ /* LS1012A needs this to make PE work correctly */
-+ writel(0x3, CLASS_PE_SYS_CLK_RATIO);
-+ writel(0x3, TMU_PE_SYS_CLK_RATIO);
-+ writel(0x3, UTIL_PE_SYS_CLK_RATIO);
-+ usleep_range(10, 20);
-+#endif
-+
-+ pr_info("CLASS version: %x\n", readl(CLASS_VERSION));
-+ pr_info("TMU version: %x\n", readl(TMU_VERSION));
-+
-+ pr_info("BMU1 version: %x\n", readl(BMU1_BASE_ADDR +
-+ BMU_VERSION));
-+ pr_info("BMU2 version: %x\n", readl(BMU2_BASE_ADDR +
-+ BMU_VERSION));
-+
-+ pr_info("EGPI1 version: %x\n", readl(EGPI1_BASE_ADDR +
-+ GPI_VERSION));
-+ pr_info("EGPI2 version: %x\n", readl(EGPI2_BASE_ADDR +
-+ GPI_VERSION));
-+ pr_info("HGPI version: %x\n", readl(HGPI_BASE_ADDR +
-+ GPI_VERSION));
-+
-+ pr_info("HIF version: %x\n", readl(HIF_VERSION));
-+ pr_info("HIF NOPCY version: %x\n", readl(HIF_NOCPY_VERSION));
-+
-+#if !defined(CONFIG_FSL_PPFE_UTIL_DISABLED)
-+ pr_info("UTIL version: %x\n", readl(UTIL_VERSION));
-+#endif
-+ while (!(readl(TMU_CTRL) & ECC_MEM_INIT_DONE))
-+ ;
-+
-+ hif_rx_disable();
-+ hif_tx_disable();
-+
-+ bmu_init(BMU1_BASE_ADDR, &bmu1_cfg);
-+
-+ pr_info("bmu_init(1) done\n");
-+
-+ bmu_init(BMU2_BASE_ADDR, &bmu2_cfg);
-+
-+ pr_info("bmu_init(2) done\n");
-+
-+ class_cfg.resume = resume ? 1 : 0;
-+
-+ class_init(&class_cfg);
-+
-+ pr_info("class_init() done\n");
-+
-+ tmu_init(&tmu_cfg);
-+
-+ pr_info("tmu_init() done\n");
-+#if !defined(CONFIG_FSL_PPFE_UTIL_DISABLED)
-+ util_init(&util_cfg);
-+
-+ pr_info("util_init() done\n");
-+#endif
-+ gpi_init(EGPI1_BASE_ADDR, &egpi1_cfg);
-+
-+ pr_info("gpi_init(1) done\n");
-+
-+ gpi_init(EGPI2_BASE_ADDR, &egpi2_cfg);
-+
-+ pr_info("gpi_init(2) done\n");
-+
-+ gpi_init(HGPI_BASE_ADDR, &hgpi_cfg);
-+
-+ pr_info("gpi_init(hif) done\n");
-+
-+ bmu_enable(BMU1_BASE_ADDR);
-+
-+ pr_info("bmu_enable(1) done\n");
-+
-+ bmu_enable(BMU2_BASE_ADDR);
-+
-+ pr_info("bmu_enable(2) done\n");
-+
-+ return 0;
-+}
-+
-+void pfe_hw_exit(struct pfe *pfe)
-+{
-+ pr_info("%s\n", __func__);
-+
-+ bmu_disable(BMU1_BASE_ADDR);
-+ bmu_reset(BMU1_BASE_ADDR);
-+
-+ bmu_disable(BMU2_BASE_ADDR);
-+ bmu_reset(BMU2_BASE_ADDR);
-+}
---- /dev/null
-+++ b/drivers/staging/fsl_ppfe/pfe_hw.h
-@@ -0,0 +1,27 @@
-+/*
-+ * Copyright 2015-2016 Freescale Semiconductor, Inc.
-+ * Copyright 2017 NXP
-+ *
-+ * This program is free software; you can redistribute it and/or modify
-+ * it under the terms of the GNU General Public License as published by
-+ * the Free Software Foundation; either version 2 of the License, or
-+ * (at your option) any later version.
-+ *
-+ * This program is distributed in the hope that it will be useful,
-+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
-+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
-+ * GNU General Public License for more details.
-+ *
-+ * You should have received a copy of the GNU General Public License
-+ * along with this program. If not, see <http://www.gnu.org/licenses/>.
-+ */
-+
-+#ifndef _PFE_HW_H_
-+#define _PFE_HW_H_
-+
-+#define PE_SYS_CLK_RATIO 1 /* SYS/AXI = 250MHz, HFE = 500MHz */
-+
-+int pfe_hw_init(struct pfe *pfe, int resume);
-+void pfe_hw_exit(struct pfe *pfe);
-+
-+#endif /* _PFE_HW_H_ */
---- /dev/null
-+++ b/drivers/staging/fsl_ppfe/pfe_ls1012a_platform.c
-@@ -0,0 +1,385 @@
-+/*
-+ * Copyright 2015-2016 Freescale Semiconductor, Inc.
-+ * Copyright 2017 NXP
-+ *
-+ * This program is free software; you can redistribute it and/or modify
-+ * it under the terms of the GNU General Public License as published by
-+ * the Free Software Foundation; either version 2 of the License, or
-+ * (at your option) any later version.
-+ *
-+ * This program is distributed in the hope that it will be useful,
-+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
-+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
-+ * GNU General Public License for more details.
-+ *
-+ * You should have received a copy of the GNU General Public License
-+ * along with this program. If not, see <http://www.gnu.org/licenses/>.
-+ */
-+
-+#include <linux/module.h>
-+#include <linux/device.h>
-+#include <linux/of_net.h>
-+#include <linux/of_address.h>
-+#include <linux/platform_device.h>
-+#include <linux/slab.h>
-+#include <linux/clk.h>
-+#include <linux/mfd/syscon.h>
-+#include <linux/regmap.h>
-+
-+#include "pfe_mod.h"
-+
-+struct ls1012a_pfe_platform_data pfe_platform_data;
-+
-+static int pfe_get_gemac_if_proprties(struct device_node *parent, int port, int
-+ if_cnt,
-+ struct ls1012a_pfe_platform_data
-+ *pdata)
-+{
-+ struct device_node *gem = NULL, *phy = NULL;
-+ int size;
-+ int ii = 0, phy_id = 0;
-+ const u32 *addr;
-+ const void *mac_addr;
-+
-+ for (ii = 0; ii < if_cnt; ii++) {
-+ gem = of_get_next_child(parent, gem);
-+ if (!gem)
-+ goto err;
-+ addr = of_get_property(gem, "reg", &size);
-+ if (addr && (be32_to_cpup(addr) == port))
-+ break;
-+ }
-+
-+ if (ii >= if_cnt) {
-+ pr_err("%s:%d Failed to find interface = %d\n",
-+ __func__, __LINE__, if_cnt);
-+ goto err;
-+ }
-+
-+ pdata->ls1012a_eth_pdata[port].gem_id = port;
-+
-+ mac_addr = of_get_mac_address(gem);
-+
-+ if (mac_addr) {
-+ memcpy(pdata->ls1012a_eth_pdata[port].mac_addr, mac_addr,
-+ ETH_ALEN);
-+ }
-+
-+ pdata->ls1012a_eth_pdata[port].mii_config = of_get_phy_mode(gem);
-+
-+ if ((pdata->ls1012a_eth_pdata[port].mii_config) < 0)
-+ pr_err("%s:%d Incorrect Phy mode....\n", __func__,
-+ __LINE__);
-+
-+ addr = of_get_property(gem, "fsl,gemac-bus-id", &size);
-+ if (!addr)
-+ pr_err("%s:%d Invalid gemac-bus-id....\n", __func__,
-+ __LINE__);
-+ else
-+ pdata->ls1012a_eth_pdata[port].bus_id = be32_to_cpup(addr);
-+
-+ addr = of_get_property(gem, "fsl,gemac-phy-id", &size);
-+ if (!addr) {
-+ pr_err("%s:%d Invalid gemac-phy-id....\n", __func__,
-+ __LINE__);
-+ } else {
-+ phy_id = be32_to_cpup(addr);
-+ pdata->ls1012a_eth_pdata[port].phy_id = phy_id;
-+ pdata->ls1012a_mdio_pdata[0].phy_mask &= ~(1 << phy_id);
-+ }
-+
-+ addr = of_get_property(gem, "fsl,mdio-mux-val", &size);
-+ if (!addr)
-+ pr_err("%s: Invalid mdio-mux-val....\n", __func__);
-+ else
-+ phy_id = be32_to_cpup(addr);
-+ pdata->ls1012a_eth_pdata[port].mdio_muxval = phy_id;
-+
-+ if (pdata->ls1012a_eth_pdata[port].phy_id < 32)
-+ pfe->mdio_muxval[pdata->ls1012a_eth_pdata[port].phy_id] =
-+ pdata->ls1012a_eth_pdata[port].mdio_muxval;
-+
-+ addr = of_get_property(gem, "fsl,pfe-phy-if-flags", &size);
-+ if (!addr)
-+ pr_err("%s:%d Invalid pfe-phy-if-flags....\n",
-+ __func__, __LINE__);
-+ else
-+ pdata->ls1012a_eth_pdata[port].phy_flags = be32_to_cpup(addr);
-+
-+ /* If PHY is enabled, read mdio properties */
-+ if (pdata->ls1012a_eth_pdata[port].phy_flags & GEMAC_NO_PHY)
-+ goto done;
-+
-+ phy = of_get_next_child(gem, NULL);
-+
-+ addr = of_get_property(phy, "reg", &size);
-+
-+ if (!addr)
-+ pr_err("%s:%d Invalid phy enable flag....\n",
-+ __func__, __LINE__);
-+ else
-+ pdata->ls1012a_mdio_pdata[port].enabled = be32_to_cpup(addr);
-+
-+ pdata->ls1012a_mdio_pdata[port].irq[0] = PHY_POLL;
-+
-+done:
-+
-+ return 0;
-+
-+err:
-+ return -1;
-+}
-+
-+/*
-+ *
-+ * pfe_platform_probe -
-+ *
-+ *
-+ */
-+static int pfe_platform_probe(struct platform_device *pdev)
-+{
-+ struct resource res;
-+ int ii, rc, interface_count = 0, size = 0;
-+ const u32 *prop;
-+ struct device_node *np;
-+ struct clk *pfe_clk;
-+
-+ np = pdev->dev.of_node;
-+
-+ if (!np) {
-+ pr_err("Invalid device node\n");
-+ return -EINVAL;
-+ }
-+
-+ pfe = kzalloc(sizeof(*pfe), GFP_KERNEL);
-+ if (!pfe) {
-+ rc = -ENOMEM;
-+ goto err_alloc;
-+ }
-+
-+ platform_set_drvdata(pdev, pfe);
-+
-+ dma_set_mask_and_coherent(&pdev->dev, DMA_BIT_MASK(32));
-+
-+ if (of_address_to_resource(np, 1, &res)) {
-+ rc = -ENOMEM;
-+ pr_err("failed to get ddr resource\n");
-+ goto err_ddr;
-+ }
-+
-+ pfe->ddr_phys_baseaddr = res.start;
-+ pfe->ddr_size = resource_size(&res);
-+ pfe->ddr_baseaddr = phys_to_virt(res.start);
-+
-+ pfe->scfg =
-+ syscon_regmap_lookup_by_phandle(pdev->dev.of_node,
-+ "fsl,pfe-scfg");
-+ if (IS_ERR(pfe->scfg)) {
-+ dev_err(&pdev->dev, "No syscfg phandle specified\n");
-+ return PTR_ERR(pfe->scfg);
-+ }
-+
-+ pfe->cbus_baseaddr = of_iomap(np, 0);
-+ if (!pfe->cbus_baseaddr) {
-+ rc = -ENOMEM;
-+ pr_err("failed to get axi resource\n");
-+ goto err_axi;
-+ }
-+
-+ pfe->hif_irq = platform_get_irq(pdev, 0);
-+ if (pfe->hif_irq < 0) {
-+ pr_err("platform_get_irq for hif failed\n");
-+ rc = pfe->hif_irq;
-+ goto err_hif_irq;
-+ }
-+
-+ pfe->wol_irq = platform_get_irq(pdev, 2);
-+ if (pfe->wol_irq < 0) {
-+ pr_err("platform_get_irq for WoL failed\n");
-+ rc = pfe->wol_irq;
-+ goto err_hif_irq;
-+ }
-+
-+ /* Read interface count */
-+ prop = of_get_property(np, "fsl,pfe-num-interfaces", &size);
-+ if (!prop) {
-+ pr_err("Failed to read number of interfaces\n");
-+ rc = -ENXIO;
-+ goto err_prop;
-+ }
-+
-+ interface_count = be32_to_cpup(prop);
-+ if (interface_count <= 0) {
-+ pr_err("No ethernet interface count : %d\n",
-+ interface_count);
-+ rc = -ENXIO;
-+ goto err_prop;
-+ }
-+
-+ pfe_platform_data.ls1012a_mdio_pdata[0].phy_mask = 0xffffffff;
-+
-+ for (ii = 0; ii < interface_count; ii++) {
-+ pfe_get_gemac_if_proprties(np, ii, interface_count,
-+ &pfe_platform_data);
-+ }
-+
-+ pfe->dev = &pdev->dev;
-+
-+ pfe->dev->platform_data = &pfe_platform_data;
-+
-+ /* declare WoL capabilities */
-+ device_init_wakeup(&pdev->dev, true);
-+
-+ /* find the clocks */
-+ pfe_clk = devm_clk_get(pfe->dev, "pfe");
-+ if (IS_ERR(pfe_clk))
-+ return PTR_ERR(pfe_clk);
-+
-+ /* PFE clock is (platform clock / 2) */
-+ /* save sys_clk value as KHz */
-+ pfe->ctrl.sys_clk = clk_get_rate(pfe_clk) / (2 * 1000);
-+
-+ rc = pfe_probe(pfe);
-+ if (rc < 0)
-+ goto err_probe;
-+
-+ return 0;
-+
-+err_probe:
-+err_prop:
-+err_hif_irq:
-+ iounmap(pfe->cbus_baseaddr);
-+
-+err_axi:
-+err_ddr:
-+ platform_set_drvdata(pdev, NULL);
-+
-+ kfree(pfe);
-+
-+err_alloc:
-+ return rc;
-+}
-+
-+/*
-+ * pfe_platform_remove -
-+ */
-+static int pfe_platform_remove(struct platform_device *pdev)
-+{
-+ struct pfe *pfe = platform_get_drvdata(pdev);
-+ int rc;
-+
-+ pr_info("%s\n", __func__);
-+
-+ rc = pfe_remove(pfe);
-+
-+ iounmap(pfe->cbus_baseaddr);
-+
-+ platform_set_drvdata(pdev, NULL);
-+
-+ kfree(pfe);
-+
-+ return rc;
-+}
-+
-+#ifdef CONFIG_PM
-+#ifdef CONFIG_PM_SLEEP
-+int pfe_platform_suspend(struct device *dev)
-+{
-+ struct pfe *pfe = platform_get_drvdata(to_platform_device(dev));
-+ struct net_device *netdev;
-+ int i;
-+
-+ pfe->wake = 0;
-+
-+ for (i = 0; i < (NUM_GEMAC_SUPPORT); i++) {
-+ netdev = pfe->eth.eth_priv[i]->ndev;
-+
-+ netif_device_detach(netdev);
-+
-+ if (netif_running(netdev))
-+ if (pfe_eth_suspend(netdev))
-+ pfe->wake = 1;
-+ }
-+
-+ /* Shutdown PFE only if we're not waking up the system */
-+ if (!pfe->wake) {
-+#if defined(LS1012A_PFE_RESET_WA)
-+ pfe_hif_rx_idle(&pfe->hif);
-+#endif
-+ pfe_ctrl_suspend(&pfe->ctrl);
-+ pfe_firmware_exit(pfe);
-+
-+ pfe_hif_exit(pfe);
-+ pfe_hif_lib_exit(pfe);
-+
-+ pfe_hw_exit(pfe);
-+ }
-+
-+ return 0;
-+}
-+
-+static int pfe_platform_resume(struct device *dev)
-+{
-+ struct pfe *pfe = platform_get_drvdata(to_platform_device(dev));
-+ struct net_device *netdev;
-+ int i;
-+
-+ if (!pfe->wake) {
-+ pfe_hw_init(pfe, 1);
-+ pfe_hif_lib_init(pfe);
-+ pfe_hif_init(pfe);
-+#if !defined(CONFIG_FSL_PPFE_UTIL_DISABLED)
-+ util_enable();
-+#endif
-+ tmu_enable(0xf);
-+ class_enable();
-+ pfe_ctrl_resume(&pfe->ctrl);
-+ }
-+
-+ for (i = 0; i < (NUM_GEMAC_SUPPORT); i++) {
-+ netdev = pfe->eth.eth_priv[i]->ndev;
-+
-+ if (pfe->eth.eth_priv[i]->mii_bus)
-+ pfe_eth_mdio_reset(pfe->eth.eth_priv[i]->mii_bus);
-+
-+ if (netif_running(netdev))
-+ pfe_eth_resume(netdev);
-+
-+ netif_device_attach(netdev);
-+ }
-+ return 0;
-+}
-+#else
-+#define pfe_platform_suspend NULL
-+#define pfe_platform_resume NULL
-+#endif
-+
-+static const struct dev_pm_ops pfe_platform_pm_ops = {
-+ SET_SYSTEM_SLEEP_PM_OPS(pfe_platform_suspend, pfe_platform_resume)
-+};
-+#endif
-+
-+static const struct of_device_id pfe_match[] = {
-+ {
-+ .compatible = "fsl,pfe",
-+ },
-+ {},
-+};
-+MODULE_DEVICE_TABLE(of, pfe_match);
-+
-+static struct platform_driver pfe_platform_driver = {
-+ .probe = pfe_platform_probe,
-+ .remove = pfe_platform_remove,
-+ .driver = {
-+ .name = "pfe",
-+ .of_match_table = pfe_match,
-+#ifdef CONFIG_PM
-+ .pm = &pfe_platform_pm_ops,
-+#endif
-+ },
-+};
-+
-+module_platform_driver(pfe_platform_driver);
-+MODULE_LICENSE("GPL");
-+MODULE_DESCRIPTION("PFE Ethernet driver");
-+MODULE_AUTHOR("NXP DNCPE");
---- /dev/null
-+++ b/drivers/staging/fsl_ppfe/pfe_mod.c
-@@ -0,0 +1,156 @@
-+/*
-+ * Copyright 2015-2016 Freescale Semiconductor, Inc.
-+ * Copyright 2017 NXP
-+ *
-+ * This program is free software; you can redistribute it and/or modify
-+ * it under the terms of the GNU General Public License as published by
-+ * the Free Software Foundation; either version 2 of the License, or
-+ * (at your option) any later version.
-+ *
-+ * This program is distributed in the hope that it will be useful,
-+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
-+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
-+ * GNU General Public License for more details.
-+ *
-+ * You should have received a copy of the GNU General Public License
-+ * along with this program. If not, see <http://www.gnu.org/licenses/>.
-+ */
-+
-+#include <linux/dma-mapping.h>
-+#include "pfe_mod.h"
-+
-+unsigned int us;
-+module_param(us, uint, 0444);
-+MODULE_PARM_DESC(us, "0: module enabled for kernel networking (DEFAULT)\n"
-+ "1: module enabled for userspace networking\n");
-+struct pfe *pfe;
-+
-+/*
-+ * pfe_probe -
-+ */
-+int pfe_probe(struct pfe *pfe)
-+{
-+ int rc;
-+
-+ if (pfe->ddr_size < DDR_MAX_SIZE) {
-+ pr_err("%s: required DDR memory (%x) above platform ddr memory (%x)\n",
-+ __func__, (unsigned int)DDR_MAX_SIZE, pfe->ddr_size);
-+ rc = -ENOMEM;
-+ goto err_hw;
-+ }
-+
-+ if (((int)(pfe->ddr_phys_baseaddr + BMU2_DDR_BASEADDR) &
-+ (8 * SZ_1M - 1)) != 0) {
-+ pr_err("%s: BMU2 base address (0x%x) must be aligned on 8MB boundary\n",
-+ __func__, (int)pfe->ddr_phys_baseaddr +
-+ BMU2_DDR_BASEADDR);
-+ rc = -ENOMEM;
-+ goto err_hw;
-+ }
-+
-+ pr_info("cbus_baseaddr: %lx, ddr_baseaddr: %lx, ddr_phys_baseaddr: %lx, ddr_size: %x\n",
-+ (unsigned long)pfe->cbus_baseaddr,
-+ (unsigned long)pfe->ddr_baseaddr,
-+ pfe->ddr_phys_baseaddr, pfe->ddr_size);
-+
-+ pfe_lib_init(pfe->cbus_baseaddr, pfe->ddr_baseaddr,
-+ pfe->ddr_phys_baseaddr, pfe->ddr_size);
-+
-+ rc = pfe_hw_init(pfe, 0);
-+ if (rc < 0)
-+ goto err_hw;
-+
-+ if (us)
-+ goto firmware_init;
-+
-+ rc = pfe_hif_lib_init(pfe);
-+ if (rc < 0)
-+ goto err_hif_lib;
-+
-+ rc = pfe_hif_init(pfe);
-+ if (rc < 0)
-+ goto err_hif;
-+
-+firmware_init:
-+ rc = pfe_firmware_init(pfe);
-+ if (rc < 0)
-+ goto err_firmware;
-+
-+ rc = pfe_ctrl_init(pfe);
-+ if (rc < 0)
-+ goto err_ctrl;
-+
-+ rc = pfe_eth_init(pfe);
-+ if (rc < 0)
-+ goto err_eth;
-+
-+ rc = pfe_sysfs_init(pfe);
-+ if (rc < 0)
-+ goto err_sysfs;
-+
-+ rc = pfe_debugfs_init(pfe);
-+ if (rc < 0)
-+ goto err_debugfs;
-+
-+ return 0;
-+
-+err_debugfs:
-+ pfe_sysfs_exit(pfe);
-+
-+err_sysfs:
-+ pfe_eth_exit(pfe);
-+
-+err_eth:
-+ pfe_ctrl_exit(pfe);
-+
-+err_ctrl:
-+ pfe_firmware_exit(pfe);
-+
-+err_firmware:
-+ if (us)
-+ goto err_hif_lib;
-+
-+ pfe_hif_exit(pfe);
-+
-+err_hif:
-+ pfe_hif_lib_exit(pfe);
-+
-+err_hif_lib:
-+ pfe_hw_exit(pfe);
-+
-+err_hw:
-+ return rc;
-+}
-+
-+/*
-+ * pfe_remove -
-+ */
-+int pfe_remove(struct pfe *pfe)
-+{
-+ pr_info("%s\n", __func__);
-+
-+ pfe_debugfs_exit(pfe);
-+
-+ pfe_sysfs_exit(pfe);
-+
-+ pfe_eth_exit(pfe);
-+
-+ pfe_ctrl_exit(pfe);
-+
-+#if defined(LS1012A_PFE_RESET_WA)
-+ pfe_hif_rx_idle(&pfe->hif);
-+#endif
-+ pfe_firmware_exit(pfe);
-+
-+ if (us)
-+ goto hw_exit;
-+
-+ pfe_hif_exit(pfe);
-+
-+ pfe_hif_lib_exit(pfe);
-+
-+hw_exit:
-+ pfe_hw_exit(pfe);
-+
-+ return 0;
-+}
---- /dev/null
-+++ b/drivers/staging/fsl_ppfe/pfe_mod.h
-@@ -0,0 +1,114 @@
-+/*
-+ * Copyright 2015-2016 Freescale Semiconductor, Inc.
-+ * Copyright 2017 NXP
-+ *
-+ * This program is free software; you can redistribute it and/or modify
-+ * it under the terms of the GNU General Public License as published by
-+ * the Free Software Foundation; either version 2 of the License, or
-+ * (at your option) any later version.
-+ *
-+ * This program is distributed in the hope that it will be useful,
-+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
-+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
-+ * GNU General Public License for more details.
-+ *
-+ * You should have received a copy of the GNU General Public License
-+ * along with this program. If not, see <http://www.gnu.org/licenses/>.
-+ */
-+
-+#ifndef _PFE_MOD_H_
-+#define _PFE_MOD_H_
-+
-+#include <linux/device.h>
-+#include <linux/elf.h>
-+
-+extern unsigned int us;
-+
-+struct pfe;
-+
-+#include "pfe_hw.h"
-+#include "pfe_firmware.h"
-+#include "pfe_ctrl.h"
-+#include "pfe_hif.h"
-+#include "pfe_hif_lib.h"
-+#include "pfe_eth.h"
-+#include "pfe_sysfs.h"
-+#include "pfe_perfmon.h"
-+#include "pfe_debugfs.h"
-+
-+#define PHYID_MAX_VAL 32
-+
-+struct pfe_tmu_credit {
-+ /* Number of allowed TX packet in-flight, matches TMU queue size */
-+ unsigned int tx_credit[NUM_GEMAC_SUPPORT][EMAC_TXQ_CNT];
-+ unsigned int tx_credit_max[NUM_GEMAC_SUPPORT][EMAC_TXQ_CNT];
-+ unsigned int tx_packets[NUM_GEMAC_SUPPORT][EMAC_TXQ_CNT];
-+};
-+
-+struct pfe {
-+ struct regmap *scfg;
-+ unsigned long ddr_phys_baseaddr;
-+ void *ddr_baseaddr;
-+ unsigned int ddr_size;
-+ void *cbus_baseaddr;
-+ void *apb_baseaddr;
-+ unsigned long iram_phys_baseaddr;
-+ void *iram_baseaddr;
-+ unsigned long ipsec_phys_baseaddr;
-+ void *ipsec_baseaddr;
-+ int hif_irq;
-+ int wol_irq;
-+ int hif_client_irq;
-+ struct device *dev;
-+ struct dentry *dentry;
-+ struct pfe_ctrl ctrl;
-+ struct pfe_hif hif;
-+ struct pfe_eth eth;
-+ struct hif_client_s *hif_client[HIF_CLIENTS_MAX];
-+#if defined(CFG_DIAGS)
-+ struct pfe_diags diags;
-+#endif
-+ struct pfe_tmu_credit tmu_credit;
-+ struct pfe_cpumon cpumon;
-+ struct pfe_memmon memmon;
-+ int wake;
-+ int mdio_muxval[PHYID_MAX_VAL];
-+ struct clk *hfe_clock;
-+};
-+
-+extern struct pfe *pfe;
-+
-+int pfe_probe(struct pfe *pfe);
-+int pfe_remove(struct pfe *pfe);
-+
-+/* DDR Mapping in reserved memory*/
-+#define ROUTE_TABLE_BASEADDR 0
-+#define ROUTE_TABLE_HASH_BITS 15 /* 32K entries */
-+#define ROUTE_TABLE_SIZE ((1 << ROUTE_TABLE_HASH_BITS) \
-+ * CLASS_ROUTE_SIZE)
-+#define BMU2_DDR_BASEADDR (ROUTE_TABLE_BASEADDR + ROUTE_TABLE_SIZE)
-+#define BMU2_BUF_COUNT (4096 - 256)
-+/* This is to get a total DDR size of 12MiB */
-+#define BMU2_DDR_SIZE (DDR_BUF_SIZE * BMU2_BUF_COUNT)
-+#define UTIL_CODE_BASEADDR (BMU2_DDR_BASEADDR + BMU2_DDR_SIZE)
-+#define UTIL_CODE_SIZE (128 * SZ_1K)
-+#define UTIL_DDR_DATA_BASEADDR (UTIL_CODE_BASEADDR + UTIL_CODE_SIZE)
-+#define UTIL_DDR_DATA_SIZE (64 * SZ_1K)
-+#define CLASS_DDR_DATA_BASEADDR (UTIL_DDR_DATA_BASEADDR + UTIL_DDR_DATA_SIZE)
-+#define CLASS_DDR_DATA_SIZE (32 * SZ_1K)
-+#define TMU_DDR_DATA_BASEADDR (CLASS_DDR_DATA_BASEADDR + CLASS_DDR_DATA_SIZE)
-+#define TMU_DDR_DATA_SIZE (32 * SZ_1K)
-+#define TMU_LLM_BASEADDR (TMU_DDR_DATA_BASEADDR + TMU_DDR_DATA_SIZE)
-+#define TMU_LLM_QUEUE_LEN (8 * 512)
-+/* Must be power of two and at least 16 * 8 = 128 bytes */
-+#define TMU_LLM_SIZE (4 * 16 * TMU_LLM_QUEUE_LEN)
-+/* (4 TMU's x 16 queues x queue_len) */
-+
-+#define DDR_MAX_SIZE (TMU_LLM_BASEADDR + TMU_LLM_SIZE)
-+
-+/* LMEM Mapping */
-+#define BMU1_LMEM_BASEADDR 0
-+#define BMU1_BUF_COUNT 256
-+#define BMU1_LMEM_SIZE (LMEM_BUF_SIZE * BMU1_BUF_COUNT)
-+
-+#endif /* _PFE_MOD_H */
---- /dev/null
-+++ b/drivers/staging/fsl_ppfe/pfe_perfmon.h
-@@ -0,0 +1,38 @@
-+/*
-+ * Copyright 2015-2016 Freescale Semiconductor, Inc.
-+ * Copyright 2017 NXP
-+ *
-+ * This program is free software; you can redistribute it and/or modify
-+ * it under the terms of the GNU General Public License as published by
-+ * the Free Software Foundation; either version 2 of the License, or
-+ * (at your option) any later version.
-+ *
-+ * This program is distributed in the hope that it will be useful,
-+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
-+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
-+ * GNU General Public License for more details.
-+ *
-+ * You should have received a copy of the GNU General Public License
-+ * along with this program. If not, see <http://www.gnu.org/licenses/>.
-+ */
-+
-+#ifndef _PFE_PERFMON_H_
-+#define _PFE_PERFMON_H_
-+
-+#include "pfe/pfe.h"
-+
-+#define CT_CPUMON_INTERVAL (1 * TIMER_TICKS_PER_SEC)
-+
-+struct pfe_cpumon {
-+ u32 cpu_usage_pct[MAX_PE];
-+ u32 class_usage_pct;
-+};
-+
-+struct pfe_memmon {
-+ u32 kernel_memory_allocated;
-+};
-+
-+int pfe_perfmon_init(struct pfe *pfe);
-+void pfe_perfmon_exit(struct pfe *pfe);
-+
-+#endif /* _PFE_PERFMON_H_ */
---- /dev/null
-+++ b/drivers/staging/fsl_ppfe/pfe_sysfs.c
-@@ -0,0 +1,818 @@
-+/*
-+ * Copyright 2015-2016 Freescale Semiconductor, Inc.
-+ * Copyright 2017 NXP
-+ *
-+ * This program is free software; you can redistribute it and/or modify
-+ * it under the terms of the GNU General Public License as published by
-+ * the Free Software Foundation; either version 2 of the License, or
-+ * (at your option) any later version.
-+ *
-+ * This program is distributed in the hope that it will be useful,
-+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
-+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
-+ * GNU General Public License for more details.
-+ *
-+ * You should have received a copy of the GNU General Public License
-+ * along with this program. If not, see <http://www.gnu.org/licenses/>.
-+ */
-+
-+#include <linux/module.h>
-+#include <linux/platform_device.h>
-+
-+#include "pfe_mod.h"
-+
-+#define PE_EXCEPTION_DUMP_ADDRESS 0x1fa8
-+#define NUM_QUEUES 16
-+
-+static char register_name[20][5] = {
-+ "EPC", "ECAS", "EID", "ED",
-+ "r0", "r1", "r2", "r3",
-+ "r4", "r5", "r6", "r7",
-+ "r8", "r9", "r10", "r11",
-+ "r12", "r13", "r14", "r15",
-+};
-+
-+static char exception_name[14][20] = {
-+ "Reset",
-+ "HardwareFailure",
-+ "NMI",
-+ "InstBreakpoint",
-+ "DataBreakpoint",
-+ "Unsupported",
-+ "PrivilegeViolation",
-+ "InstBusError",
-+ "DataBusError",
-+ "AlignmentError",
-+ "ArithmeticError",
-+ "SystemCall",
-+ "MemoryManagement",
-+ "Interrupt",
-+};
-+
-+static unsigned long class_do_clear;
-+static unsigned long tmu_do_clear;
-+#if !defined(CONFIG_FSL_PPFE_UTIL_DISABLED)
-+static unsigned long util_do_clear;
-+#endif
-+
-+static ssize_t display_pe_status(char *buf, int id, u32 dmem_addr, unsigned long
-+ do_clear)
-+{
-+ ssize_t len = 0;
-+ u32 val;
-+ char statebuf[5];
-+ struct pfe_cpumon *cpumon = &pfe->cpumon;
-+ u32 debug_indicator;
-+ u32 debug[20];
-+
-+ *(u32 *)statebuf = pe_dmem_read(id, dmem_addr, 4);
-+ dmem_addr += 4;
-+
-+ statebuf[4] = '\0';
-+ len += sprintf(buf + len, "state=%4s ", statebuf);
-+
-+ val = pe_dmem_read(id, dmem_addr, 4);
-+ dmem_addr += 4;
-+ len += sprintf(buf + len, "ctr=%08x ", cpu_to_be32(val));
-+
-+ val = pe_dmem_read(id, dmem_addr, 4);
-+ if (do_clear && val)
-+ pe_dmem_write(id, 0, dmem_addr, 4);
-+ dmem_addr += 4;
-+ len += sprintf(buf + len, "rx=%u ", cpu_to_be32(val));
-+
-+ val = pe_dmem_read(id, dmem_addr, 4);
-+ if (do_clear && val)
-+ pe_dmem_write(id, 0, dmem_addr, 4);
-+ dmem_addr += 4;
-+ if (id >= TMU0_ID && id <= TMU_MAX_ID)
-+ len += sprintf(buf + len, "qstatus=%x", cpu_to_be32(val));
-+ else
-+ len += sprintf(buf + len, "tx=%u", cpu_to_be32(val));
-+
-+ val = pe_dmem_read(id, dmem_addr, 4);
-+ if (do_clear && val)
-+ pe_dmem_write(id, 0, dmem_addr, 4);
-+ dmem_addr += 4;
-+ if (val)
-+ len += sprintf(buf + len, " drop=%u", cpu_to_be32(val));
-+
-+ len += sprintf(buf + len, " load=%d%%", cpumon->cpu_usage_pct[id]);
-+
-+ len += sprintf(buf + len, "\n");
-+
-+ debug_indicator = pe_dmem_read(id, dmem_addr, 4);
-+ dmem_addr += 4;
-+ if (!strncmp((char *)&debug_indicator, "DBUG", 4)) {
-+ int j, last = 0;
-+
-+ for (j = 0; j < 16; j++) {
-+ debug[j] = pe_dmem_read(id, dmem_addr, 4);
-+ if (debug[j]) {
-+ if (do_clear)
-+ pe_dmem_write(id, 0, dmem_addr, 4);
-+ last = j + 1;
-+ }
-+ dmem_addr += 4;
-+ }
-+ for (j = 0; j < last; j++) {
-+ len += sprintf(buf + len, "%08x%s",
-+ cpu_to_be32(debug[j]),
-+ (j & 0x7) == 0x7 || j == last - 1 ? "\n" : " ");
-+ }
-+ }
-+
-+ if (!strncmp(statebuf, "DEAD", 4)) {
-+ u32 i, dump = PE_EXCEPTION_DUMP_ADDRESS;
-+
-+ len += sprintf(buf + len, "Exception details:\n");
-+ for (i = 0; i < 20; i++) {
-+ debug[i] = pe_dmem_read(id, dump, 4);
-+ dump += 4;
-+ if (i == 2)
-+ len += sprintf(buf + len, "%4s = %08x (=%s) ",
-+ register_name[i], cpu_to_be32(debug[i]),
-+ exception_name[min((u32)
-+ cpu_to_be32(debug[i]), (u32)13)]);
-+ else
-+ len += sprintf(buf + len, "%4s = %08x%s",
-+ register_name[i], cpu_to_be32(debug[i]),
-+ (i & 0x3) == 0x3 || i == 19 ? "\n" : " ");
-+ }
-+ }
-+
-+ return len;
-+}
-+
-+static ssize_t class_phy_stats(char *buf, int phy)
-+{
-+ ssize_t len = 0;
-+ int off1 = phy * 0x28;
-+ int off2 = phy * 0x10;
-+
-+ if (phy == 3)
-+ off1 = CLASS_PHY4_RX_PKTS - CLASS_PHY1_RX_PKTS;
-+
-+ len += sprintf(buf + len, "phy: %d\n", phy);
-+ len += sprintf(buf + len,
-+ " rx: %10u, tx: %10u, intf: %10u, ipv4: %10u, ipv6: %10u\n",
-+ readl(CLASS_PHY1_RX_PKTS + off1),
-+ readl(CLASS_PHY1_TX_PKTS + off1),
-+ readl(CLASS_PHY1_INTF_MATCH_PKTS + off1),
-+ readl(CLASS_PHY1_V4_PKTS + off1),
-+ readl(CLASS_PHY1_V6_PKTS + off1));
-+
-+ len += sprintf(buf + len,
-+ " icmp: %10u, igmp: %10u, tcp: %10u, udp: %10u\n",
-+ readl(CLASS_PHY1_ICMP_PKTS + off2),
-+ readl(CLASS_PHY1_IGMP_PKTS + off2),
-+ readl(CLASS_PHY1_TCP_PKTS + off2),
-+ readl(CLASS_PHY1_UDP_PKTS + off2));
-+
-+ len += sprintf(buf + len, " err\n");
-+ len += sprintf(buf + len,
-+ " lp: %10u, intf: %10u, l3: %10u, chcksum: %10u, ttl: %10u\n",
-+ readl(CLASS_PHY1_LP_FAIL_PKTS + off1),
-+ readl(CLASS_PHY1_INTF_FAIL_PKTS + off1),
-+ readl(CLASS_PHY1_L3_FAIL_PKTS + off1),
-+ readl(CLASS_PHY1_CHKSUM_ERR_PKTS + off1),
-+ readl(CLASS_PHY1_TTL_ERR_PKTS + off1));
-+
-+ return len;
-+}
-+
-+/* qm_read_drop_stat
-+ * This function is used to read the drop statistics from the TMU
-+ * hw drop counter. Since the hw counter is always cleared afer
-+ * reading, this function maintains the previous drop count, and
-+ * adds the new value to it. That value can be retrieved by
-+ * passing a pointer to it with the total_drops arg.
-+ *
-+ * @param tmu TMU number (0 - 3)
-+ * @param queue queue number (0 - 15)
-+ * @param total_drops pointer to location to store total drops (or NULL)
-+ * @param do_reset if TRUE, clear total drops after updating
-+ */
-+u32 qm_read_drop_stat(u32 tmu, u32 queue, u32 *total_drops, int do_reset)
-+{
-+ static u32 qtotal[TMU_MAX_ID + 1][NUM_QUEUES];
-+ u32 val;
-+
-+ writel((tmu << 8) | queue, TMU_TEQ_CTRL);
-+ writel((tmu << 8) | queue, TMU_LLM_CTRL);
-+ val = readl(TMU_TEQ_DROP_STAT);
-+ qtotal[tmu][queue] += val;
-+ if (total_drops)
-+ *total_drops = qtotal[tmu][queue];
-+ if (do_reset)
-+ qtotal[tmu][queue] = 0;
-+ return val;
-+}
-+
-+static ssize_t tmu_queue_stats(char *buf, int tmu, int queue)
-+{
-+ ssize_t len = 0;
-+ u32 drops;
-+
-+ len += sprintf(buf + len, "%d-%02d, ", tmu, queue);
-+
-+ drops = qm_read_drop_stat(tmu, queue, NULL, 0);
-+
-+ /* Select queue */
-+ writel((tmu << 8) | queue, TMU_TEQ_CTRL);
-+ writel((tmu << 8) | queue, TMU_LLM_CTRL);
-+
-+ len += sprintf(buf + len,
-+ "(teq) drop: %10u, tx: %10u (llm) head: %08x, tail: %08x, drop: %10u\n",
-+ drops, readl(TMU_TEQ_TRANS_STAT),
-+ readl(TMU_LLM_QUE_HEADPTR), readl(TMU_LLM_QUE_TAILPTR),
-+ readl(TMU_LLM_QUE_DROPCNT));
-+
-+ return len;
-+}
-+
-+static ssize_t tmu_queues(char *buf, int tmu)
-+{
-+ ssize_t len = 0;
-+ int queue;
-+
-+ for (queue = 0; queue < 16; queue++)
-+ len += tmu_queue_stats(buf + len, tmu, queue);
-+
-+ return len;
-+}
-+
-+static ssize_t block_version(char *buf, void *addr)
-+{
-+ ssize_t len = 0;
-+ u32 val;
-+
-+ val = readl(addr);
-+ len += sprintf(buf + len, "revision: %x, version: %x, id: %x\n",
-+ (val >> 24) & 0xff, (val >> 16) & 0xff, val & 0xffff);
-+
-+ return len;
-+}
-+
-+static ssize_t bmu(char *buf, int id, void *base)
-+{
-+ ssize_t len = 0;
-+
-+ len += sprintf(buf + len, "%s: %d\n ", __func__, id);
-+
-+ len += block_version(buf + len, base + BMU_VERSION);
-+
-+ len += sprintf(buf + len, " buf size: %x\n", (1 << readl(base +
-+ BMU_BUF_SIZE)));
-+ len += sprintf(buf + len, " buf count: %x\n", readl(base +
-+ BMU_BUF_CNT));
-+ len += sprintf(buf + len, " buf rem: %x\n", readl(base +
-+ BMU_REM_BUF_CNT));
-+ len += sprintf(buf + len, " buf curr: %x\n", readl(base +
-+ BMU_CURR_BUF_CNT));
-+ len += sprintf(buf + len, " free err: %x\n", readl(base +
-+ BMU_FREE_ERR_ADDR));
-+
-+ return len;
-+}
-+
-+static ssize_t gpi(char *buf, int id, void *base)
-+{
-+ ssize_t len = 0;
-+ u32 val;
-+
-+ len += sprintf(buf + len, "%s%d:\n ", __func__, id);
-+ len += block_version(buf + len, base + GPI_VERSION);
-+
-+ len += sprintf(buf + len, " tx under stick: %x\n", readl(base +
-+ GPI_FIFO_STATUS));
-+ val = readl(base + GPI_FIFO_DEBUG);
-+ len += sprintf(buf + len, " tx pkts: %x\n", (val >> 23) &
-+ 0x3f);
-+ len += sprintf(buf + len, " rx pkts: %x\n", (val >> 18) &
-+ 0x3f);
-+ len += sprintf(buf + len, " tx bytes: %x\n", (val >> 9) &
-+ 0x1ff);
-+ len += sprintf(buf + len, " rx bytes: %x\n", (val >> 0) &
-+ 0x1ff);
-+ len += sprintf(buf + len, " overrun: %x\n", readl(base +
-+ GPI_OVERRUN_DROPCNT));
-+
-+ return len;
-+}
-+
-+static ssize_t pfe_set_class(struct device *dev, struct device_attribute *attr,
-+ const char *buf, size_t count)
-+{
-+ class_do_clear = kstrtoul(buf, 0, 0);
-+ return count;
-+}
-+
-+static ssize_t pfe_show_class(struct device *dev, struct device_attribute *attr,
-+ char *buf)
-+{
-+ ssize_t len = 0;
-+ int id;
-+ u32 val;
-+ struct pfe_cpumon *cpumon = &pfe->cpumon;
-+
-+ len += block_version(buf + len, CLASS_VERSION);
-+
-+ for (id = CLASS0_ID; id <= CLASS_MAX_ID; id++) {
-+ len += sprintf(buf + len, "%d: ", id - CLASS0_ID);
-+
-+ val = readl(CLASS_PE0_DEBUG + id * 4);
-+ len += sprintf(buf + len, "pc=1%04x ", val & 0xffff);
-+
-+ len += display_pe_status(buf + len, id, CLASS_DM_PESTATUS,
-+ class_do_clear);
-+ }
-+ len += sprintf(buf + len, "aggregate load=%d%%\n\n",
-+ cpumon->class_usage_pct);
-+
-+ len += sprintf(buf + len, "pe status: 0x%x\n",
-+ readl(CLASS_PE_STATUS));
-+ len += sprintf(buf + len, "max buf cnt: 0x%x afull thres: 0x%x\n",
-+ readl(CLASS_MAX_BUF_CNT), readl(CLASS_AFULL_THRES));
-+ len += sprintf(buf + len, "tsq max cnt: 0x%x tsq fifo thres: 0x%x\n",
-+ readl(CLASS_TSQ_MAX_CNT), readl(CLASS_TSQ_FIFO_THRES));
-+ len += sprintf(buf + len, "state: 0x%x\n", readl(CLASS_STATE));
-+
-+ len += class_phy_stats(buf + len, 0);
-+ len += class_phy_stats(buf + len, 1);
-+ len += class_phy_stats(buf + len, 2);
-+ len += class_phy_stats(buf + len, 3);
-+
-+ return len;
-+}
-+
-+static ssize_t pfe_set_tmu(struct device *dev, struct device_attribute *attr,
-+ const char *buf, size_t count)
-+{
-+ tmu_do_clear = kstrtoul(buf, 0, 0);
-+ return count;
-+}
-+
-+static ssize_t pfe_show_tmu(struct device *dev, struct device_attribute *attr,
-+ char *buf)
-+{
-+ ssize_t len = 0;
-+ int id;
-+ u32 val;
-+
-+ len += block_version(buf + len, TMU_VERSION);
-+
-+ for (id = TMU0_ID; id <= TMU_MAX_ID; id++) {
-+ if (id == TMU2_ID)
-+ continue;
-+ len += sprintf(buf + len, "%d: ", id - TMU0_ID);
-+
-+ len += display_pe_status(buf + len, id, TMU_DM_PESTATUS,
-+ tmu_do_clear);
-+ }
-+
-+ len += sprintf(buf + len, "pe status: %x\n", readl(TMU_PE_STATUS));
-+ len += sprintf(buf + len, "inq fifo cnt: %x\n",
-+ readl(TMU_PHY_INQ_FIFO_CNT));
-+ val = readl(TMU_INQ_STAT);
-+ len += sprintf(buf + len, "inq wr ptr: %x\n", val & 0x3ff);
-+ len += sprintf(buf + len, "inq rd ptr: %x\n", val >> 10);
-+
-+ return len;
-+}
-+
-+static unsigned long drops_do_clear;
-+static u32 class_drop_counter[CLASS_NUM_DROP_COUNTERS];
-+#if !defined(CONFIG_FSL_PPFE_UTIL_DISABLED)
-+static u32 util_drop_counter[UTIL_NUM_DROP_COUNTERS];
-+#endif
-+
-+char *class_drop_description[CLASS_NUM_DROP_COUNTERS] = {
-+ "ICC",
-+ "Host Pkt Error",
-+ "Rx Error",
-+ "IPsec Outbound",
-+ "IPsec Inbound",
-+ "EXPT IPsec Error",
-+ "Reassembly",
-+ "Fragmenter",
-+ "NAT-T",
-+ "Socket",
-+ "Multicast",
-+ "NAT-PT",
-+ "Tx Disabled",
-+};
-+
-+#if !defined(CONFIG_FSL_PPFE_UTIL_DISABLED)
-+char *util_drop_description[UTIL_NUM_DROP_COUNTERS] = {
-+ "IPsec Outbound",
-+ "IPsec Inbound",
-+ "IPsec Rate Limiter",
-+ "Fragmenter",
-+ "Socket",
-+ "Tx Disabled",
-+ "Rx Error",
-+};
-+#endif
-+
-+static ssize_t pfe_set_drops(struct device *dev, struct device_attribute *attr,
-+ const char *buf, size_t count)
-+{
-+ drops_do_clear = kstrtoul(buf, 0, 0);
-+ return count;
-+}
-+
-+static u32 tmu_drops[4][16];
-+static ssize_t pfe_show_drops(struct device *dev, struct device_attribute *attr,
-+ char *buf)
-+{
-+ ssize_t len = 0;
-+ int id, dropnum;
-+ int tmu, queue;
-+ u32 val;
-+ u32 dmem_addr;
-+ int num_class_drops = 0, num_tmu_drops = 0, num_util_drops = 0;
-+ struct pfe_ctrl *ctrl = &pfe->ctrl;
-+
-+ memset(class_drop_counter, 0, sizeof(class_drop_counter));
-+ for (id = CLASS0_ID; id <= CLASS_MAX_ID; id++) {
-+ if (drops_do_clear)
-+ pe_sync_stop(ctrl, (1 << id));
-+ for (dropnum = 0; dropnum < CLASS_NUM_DROP_COUNTERS;
-+ dropnum++) {
-+ dmem_addr = CLASS_DM_DROP_CNTR;
-+ val = be32_to_cpu(pe_dmem_read(id, dmem_addr, 4));
-+ class_drop_counter[dropnum] += val;
-+ num_class_drops += val;
-+ if (drops_do_clear)
-+ pe_dmem_write(id, 0, dmem_addr, 4);
-+ }
-+ if (drops_do_clear)
-+ pe_start(ctrl, (1 << id));
-+ }
-+
-+#if !defined(CONFIG_FSL_PPFE_UTIL_DISABLED)
-+ if (drops_do_clear)
-+ pe_sync_stop(ctrl, (1 << UTIL_ID));
-+ for (dropnum = 0; dropnum < UTIL_NUM_DROP_COUNTERS; dropnum++) {
-+ dmem_addr = UTIL_DM_DROP_CNTR;
-+ val = be32_to_cpu(pe_dmem_read(UTIL_ID, dmem_addr, 4));
-+ util_drop_counter[dropnum] = val;
-+ num_util_drops += val;
-+ if (drops_do_clear)
-+ pe_dmem_write(UTIL_ID, 0, dmem_addr, 4);
-+ }
-+ if (drops_do_clear)
-+ pe_start(ctrl, (1 << UTIL_ID));
-+#endif
-+ for (tmu = 0; tmu < 4; tmu++) {
-+ for (queue = 0; queue < 16; queue++) {
-+ qm_read_drop_stat(tmu, queue, &tmu_drops[tmu][queue],
-+ drops_do_clear);
-+ num_tmu_drops += tmu_drops[tmu][queue];
-+ }
-+ }
-+
-+ if (num_class_drops == 0 && num_util_drops == 0 && num_tmu_drops == 0)
-+ len += sprintf(buf + len, "No PE drops\n\n");
-+
-+ if (num_class_drops > 0) {
-+ len += sprintf(buf + len, "Class PE drops --\n");
-+ for (dropnum = 0; dropnum < CLASS_NUM_DROP_COUNTERS;
-+ dropnum++) {
-+ if (class_drop_counter[dropnum] > 0)
-+ len += sprintf(buf + len, " %s: %d\n",
-+ class_drop_description[dropnum],
-+ class_drop_counter[dropnum]);
-+ }
-+ len += sprintf(buf + len, "\n");
-+ }
-+
-+#if !defined(CONFIG_FSL_PPFE_UTIL_DISABLED)
-+ if (num_util_drops > 0) {
-+ len += sprintf(buf + len, "Util PE drops --\n");
-+ for (dropnum = 0; dropnum < UTIL_NUM_DROP_COUNTERS; dropnum++) {
-+ if (util_drop_counter[dropnum] > 0)
-+ len += sprintf(buf + len, " %s: %d\n",
-+ util_drop_description[dropnum],
-+ util_drop_counter[dropnum]);
-+ }
-+ len += sprintf(buf + len, "\n");
-+ }
-+#endif
-+ if (num_tmu_drops > 0) {
-+ len += sprintf(buf + len, "TMU drops --\n");
-+ for (tmu = 0; tmu < 4; tmu++) {
-+ for (queue = 0; queue < 16; queue++) {
-+ if (tmu_drops[tmu][queue] > 0)
-+ len += sprintf(buf + len,
-+ " TMU%d-Q%d: %d\n"
-+ , tmu, queue, tmu_drops[tmu][queue]);
-+ }
-+ }
-+ len += sprintf(buf + len, "\n");
-+ }
-+
-+ return len;
-+}
-+
-+static ssize_t pfe_show_tmu0_queues(struct device *dev, struct device_attribute
-+ *attr, char *buf)
-+{
-+ return tmu_queues(buf, 0);
-+}
-+
-+static ssize_t pfe_show_tmu1_queues(struct device *dev, struct device_attribute
-+ *attr, char *buf)
-+{
-+ return tmu_queues(buf, 1);
-+}
-+
-+static ssize_t pfe_show_tmu2_queues(struct device *dev, struct device_attribute
-+ *attr, char *buf)
-+{
-+ return tmu_queues(buf, 2);
-+}
-+
-+static ssize_t pfe_show_tmu3_queues(struct device *dev, struct device_attribute
-+ *attr, char *buf)
-+{
-+ return tmu_queues(buf, 3);
-+}
-+
-+#if !defined(CONFIG_FSL_PPFE_UTIL_DISABLED)
-+static ssize_t pfe_set_util(struct device *dev, struct device_attribute *attr,
-+ const char *buf, size_t count)
-+{
-+ util_do_clear = kstrtoul(buf, NULL, 0);
-+ return count;
-+}
-+
-+static ssize_t pfe_show_util(struct device *dev, struct device_attribute *attr,
-+ char *buf)
-+{
-+ ssize_t len = 0;
-+ struct pfe_ctrl *ctrl = &pfe->ctrl;
-+
-+ len += block_version(buf + len, UTIL_VERSION);
-+
-+ pe_sync_stop(ctrl, (1 << UTIL_ID));
-+ len += display_pe_status(buf + len, UTIL_ID, UTIL_DM_PESTATUS,
-+ util_do_clear);
-+ pe_start(ctrl, (1 << UTIL_ID));
-+
-+ len += sprintf(buf + len, "pe status: %x\n", readl(UTIL_PE_STATUS));
-+ len += sprintf(buf + len, "max buf cnt: %x\n",
-+ readl(UTIL_MAX_BUF_CNT));
-+ len += sprintf(buf + len, "tsq max cnt: %x\n",
-+ readl(UTIL_TSQ_MAX_CNT));
-+
-+ return len;
-+}
-+#endif
-+
-+static ssize_t pfe_show_bmu(struct device *dev, struct device_attribute *attr,
-+ char *buf)
-+{
-+ ssize_t len = 0;
-+
-+ len += bmu(buf + len, 1, BMU1_BASE_ADDR);
-+ len += bmu(buf + len, 2, BMU2_BASE_ADDR);
-+
-+ return len;
-+}
-+
-+static ssize_t pfe_show_hif(struct device *dev, struct device_attribute *attr,
-+ char *buf)
-+{
-+ ssize_t len = 0;
-+
-+ len += sprintf(buf + len, "hif:\n ");
-+ len += block_version(buf + len, HIF_VERSION);
-+
-+ len += sprintf(buf + len, " tx curr bd: %x\n",
-+ readl(HIF_TX_CURR_BD_ADDR));
-+ len += sprintf(buf + len, " tx status: %x\n",
-+ readl(HIF_TX_STATUS));
-+ len += sprintf(buf + len, " tx dma status: %x\n",
-+ readl(HIF_TX_DMA_STATUS));
-+
-+ len += sprintf(buf + len, " rx curr bd: %x\n",
-+ readl(HIF_RX_CURR_BD_ADDR));
-+ len += sprintf(buf + len, " rx status: %x\n",
-+ readl(HIF_RX_STATUS));
-+ len += sprintf(buf + len, " rx dma status: %x\n",
-+ readl(HIF_RX_DMA_STATUS));
-+
-+ len += sprintf(buf + len, "hif nocopy:\n ");
-+ len += block_version(buf + len, HIF_NOCPY_VERSION);
-+
-+ len += sprintf(buf + len, " tx curr bd: %x\n",
-+ readl(HIF_NOCPY_TX_CURR_BD_ADDR));
-+ len += sprintf(buf + len, " tx status: %x\n",
-+ readl(HIF_NOCPY_TX_STATUS));
-+ len += sprintf(buf + len, " tx dma status: %x\n",
-+ readl(HIF_NOCPY_TX_DMA_STATUS));
-+
-+ len += sprintf(buf + len, " rx curr bd: %x\n",
-+ readl(HIF_NOCPY_RX_CURR_BD_ADDR));
-+ len += sprintf(buf + len, " rx status: %x\n",
-+ readl(HIF_NOCPY_RX_STATUS));
-+ len += sprintf(buf + len, " rx dma status: %x\n",
-+ readl(HIF_NOCPY_RX_DMA_STATUS));
-+
-+ return len;
-+}
-+
-+static ssize_t pfe_show_gpi(struct device *dev, struct device_attribute *attr,
-+ char *buf)
-+{
-+ ssize_t len = 0;
-+
-+ len += gpi(buf + len, 0, EGPI1_BASE_ADDR);
-+ len += gpi(buf + len, 1, EGPI2_BASE_ADDR);
-+ len += gpi(buf + len, 3, HGPI_BASE_ADDR);
-+
-+ return len;
-+}
-+
-+static ssize_t pfe_show_pfemem(struct device *dev, struct device_attribute
-+ *attr, char *buf)
-+{
-+ ssize_t len = 0;
-+ struct pfe_memmon *memmon = &pfe->memmon;
-+
-+ len += sprintf(buf + len, "Kernel Memory: %d Bytes (%d KB)\n",
-+ memmon->kernel_memory_allocated,
-+ (memmon->kernel_memory_allocated + 1023) / 1024);
-+
-+ return len;
-+}
-+
-+#ifdef HIF_NAPI_STATS
-+static ssize_t pfe_show_hif_napi_stats(struct device *dev,
-+ struct device_attribute *attr,
-+ char *buf)
-+{
-+ struct platform_device *pdev = to_platform_device(dev);
-+ struct pfe *pfe = platform_get_drvdata(pdev);
-+ ssize_t len = 0;
-+
-+ len += sprintf(buf + len, "sched: %u\n",
-+ pfe->hif.napi_counters[NAPI_SCHED_COUNT]);
-+ len += sprintf(buf + len, "poll: %u\n",
-+ pfe->hif.napi_counters[NAPI_POLL_COUNT]);
-+ len += sprintf(buf + len, "packet: %u\n",
-+ pfe->hif.napi_counters[NAPI_PACKET_COUNT]);
-+ len += sprintf(buf + len, "budget: %u\n",
-+ pfe->hif.napi_counters[NAPI_FULL_BUDGET_COUNT]);
-+ len += sprintf(buf + len, "desc: %u\n",
-+ pfe->hif.napi_counters[NAPI_DESC_COUNT]);
-+ len += sprintf(buf + len, "full: %u\n",
-+ pfe->hif.napi_counters[NAPI_CLIENT_FULL_COUNT]);
-+
-+ return len;
-+}
-+
-+static ssize_t pfe_set_hif_napi_stats(struct device *dev,
-+ struct device_attribute *attr,
-+ const char *buf, size_t count)
-+{
-+ struct platform_device *pdev = to_platform_device(dev);
-+ struct pfe *pfe = platform_get_drvdata(pdev);
-+
-+ memset(pfe->hif.napi_counters, 0, sizeof(pfe->hif.napi_counters));
-+
-+ return count;
-+}
-+
-+static DEVICE_ATTR(hif_napi_stats, 0644, pfe_show_hif_napi_stats,
-+ pfe_set_hif_napi_stats);
-+#endif
-+
-+static DEVICE_ATTR(class, 0644, pfe_show_class, pfe_set_class);
-+static DEVICE_ATTR(tmu, 0644, pfe_show_tmu, pfe_set_tmu);
-+#if !defined(CONFIG_FSL_PPFE_UTIL_DISABLED)
-+static DEVICE_ATTR(util, 0644, pfe_show_util, pfe_set_util);
-+#endif
-+static DEVICE_ATTR(bmu, 0444, pfe_show_bmu, NULL);
-+static DEVICE_ATTR(hif, 0444, pfe_show_hif, NULL);
-+static DEVICE_ATTR(gpi, 0444, pfe_show_gpi, NULL);
-+static DEVICE_ATTR(drops, 0644, pfe_show_drops, pfe_set_drops);
-+static DEVICE_ATTR(tmu0_queues, 0444, pfe_show_tmu0_queues, NULL);
-+static DEVICE_ATTR(tmu1_queues, 0444, pfe_show_tmu1_queues, NULL);
-+static DEVICE_ATTR(tmu2_queues, 0444, pfe_show_tmu2_queues, NULL);
-+static DEVICE_ATTR(tmu3_queues, 0444, pfe_show_tmu3_queues, NULL);
-+static DEVICE_ATTR(pfemem, 0444, pfe_show_pfemem, NULL);
-+
-+int pfe_sysfs_init(struct pfe *pfe)
-+{
-+ if (device_create_file(pfe->dev, &dev_attr_class))
-+ goto err_class;
-+
-+ if (device_create_file(pfe->dev, &dev_attr_tmu))
-+ goto err_tmu;
-+
-+#if !defined(CONFIG_FSL_PPFE_UTIL_DISABLED)
-+ if (device_create_file(pfe->dev, &dev_attr_util))
-+ goto err_util;
-+#endif
-+
-+ if (device_create_file(pfe->dev, &dev_attr_bmu))
-+ goto err_bmu;
-+
-+ if (device_create_file(pfe->dev, &dev_attr_hif))
-+ goto err_hif;
-+
-+ if (device_create_file(pfe->dev, &dev_attr_gpi))
-+ goto err_gpi;
-+
-+ if (device_create_file(pfe->dev, &dev_attr_drops))
-+ goto err_drops;
-+
-+ if (device_create_file(pfe->dev, &dev_attr_tmu0_queues))
-+ goto err_tmu0_queues;
-+
-+ if (device_create_file(pfe->dev, &dev_attr_tmu1_queues))
-+ goto err_tmu1_queues;
-+
-+ if (device_create_file(pfe->dev, &dev_attr_tmu2_queues))
-+ goto err_tmu2_queues;
-+
-+ if (device_create_file(pfe->dev, &dev_attr_tmu3_queues))
-+ goto err_tmu3_queues;
-+
-+ if (device_create_file(pfe->dev, &dev_attr_pfemem))
-+ goto err_pfemem;
-+
-+#ifdef HIF_NAPI_STATS
-+ if (device_create_file(pfe->dev, &dev_attr_hif_napi_stats))
-+ goto err_hif_napi_stats;
-+#endif
-+
-+ return 0;
-+
-+#ifdef HIF_NAPI_STATS
-+err_hif_napi_stats:
-+ device_remove_file(pfe->dev, &dev_attr_pfemem);
-+#endif
-+
-+err_pfemem:
-+ device_remove_file(pfe->dev, &dev_attr_tmu3_queues);
-+
-+err_tmu3_queues:
-+ device_remove_file(pfe->dev, &dev_attr_tmu2_queues);
-+
-+err_tmu2_queues:
-+ device_remove_file(pfe->dev, &dev_attr_tmu1_queues);
-+
-+err_tmu1_queues:
-+ device_remove_file(pfe->dev, &dev_attr_tmu0_queues);
-+
-+err_tmu0_queues:
-+ device_remove_file(pfe->dev, &dev_attr_drops);
-+
-+err_drops:
-+ device_remove_file(pfe->dev, &dev_attr_gpi);
-+
-+err_gpi:
-+ device_remove_file(pfe->dev, &dev_attr_hif);
-+
-+err_hif:
-+ device_remove_file(pfe->dev, &dev_attr_bmu);
-+
-+err_bmu:
-+#if !defined(CONFIG_FSL_PPFE_UTIL_DISABLED)
-+ device_remove_file(pfe->dev, &dev_attr_util);
-+
-+err_util:
-+#endif
-+ device_remove_file(pfe->dev, &dev_attr_tmu);
-+
-+err_tmu:
-+ device_remove_file(pfe->dev, &dev_attr_class);
-+
-+err_class:
-+ return -1;
-+}
-+
-+void pfe_sysfs_exit(struct pfe *pfe)
-+{
-+#ifdef HIF_NAPI_STATS
-+ device_remove_file(pfe->dev, &dev_attr_hif_napi_stats);
-+#endif
-+ device_remove_file(pfe->dev, &dev_attr_pfemem);
-+ device_remove_file(pfe->dev, &dev_attr_tmu3_queues);
-+ device_remove_file(pfe->dev, &dev_attr_tmu2_queues);
-+ device_remove_file(pfe->dev, &dev_attr_tmu1_queues);
-+ device_remove_file(pfe->dev, &dev_attr_tmu0_queues);
-+ device_remove_file(pfe->dev, &dev_attr_drops);
-+ device_remove_file(pfe->dev, &dev_attr_gpi);
-+ device_remove_file(pfe->dev, &dev_attr_hif);
-+ device_remove_file(pfe->dev, &dev_attr_bmu);
-+#if !defined(CONFIG_FSL_PPFE_UTIL_DISABLED)
-+ device_remove_file(pfe->dev, &dev_attr_util);
-+#endif
-+ device_remove_file(pfe->dev, &dev_attr_tmu);
-+ device_remove_file(pfe->dev, &dev_attr_class);
-+}
---- /dev/null
-+++ b/drivers/staging/fsl_ppfe/pfe_sysfs.h
-@@ -0,0 +1,29 @@
-+/*
-+ * Copyright 2015-2016 Freescale Semiconductor, Inc.
-+ * Copyright 2017 NXP
-+ *
-+ * This program is free software; you can redistribute it and/or modify
-+ * it under the terms of the GNU General Public License as published by
-+ * the Free Software Foundation; either version 2 of the License, or
-+ * (at your option) any later version.
-+ *
-+ * This program is distributed in the hope that it will be useful,
-+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
-+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
-+ * GNU General Public License for more details.
-+ *
-+ * You should have received a copy of the GNU General Public License
-+ * along with this program. If not, see <http://www.gnu.org/licenses/>.
-+ */
-+
-+#ifndef _PFE_SYSFS_H_
-+#define _PFE_SYSFS_H_
-+
-+#include <linux/proc_fs.h>
-+
-+u32 qm_read_drop_stat(u32 tmu, u32 queue, u32 *total_drops, int do_reset);
-+
-+int pfe_sysfs_init(struct pfe *pfe);
-+void pfe_sysfs_exit(struct pfe *pfe);
-+
-+#endif /* _PFE_SYSFS_H_ */
diff --git a/target/linux/layerscape/patches-4.9/801-ata-support-layerscape.patch b/target/linux/layerscape/patches-4.9/801-ata-support-layerscape.patch
deleted file mode 100644
index e9d34d36cc..0000000000
--- a/target/linux/layerscape/patches-4.9/801-ata-support-layerscape.patch
+++ /dev/null
@@ -1,144 +0,0 @@
-From 4c3979602db05bca439bfc98db88dc14a8663db0 Mon Sep 17 00:00:00 2001
-From: Yangbo Lu <yangbo.lu@nxp.com>
-Date: Wed, 17 Jan 2018 15:14:57 +0800
-Subject: [PATCH 13/30] ata: support layerscape
-
-This is an integrated patch for layerscape sata support.
-
-Signed-off-by: Tang Yuantian <Yuantian.Tang@nxp.com>
-Signed-off-by: Yangbo Lu <yangbo.lu@nxp.com>
----
- drivers/ata/ahci_qoriq.c | 63 ++++++++++++++++++++++++++++++++++++++++++------
- 1 file changed, 56 insertions(+), 7 deletions(-)
-
---- a/drivers/ata/ahci_qoriq.c
-+++ b/drivers/ata/ahci_qoriq.c
-@@ -1,7 +1,7 @@
- /*
- * Freescale QorIQ AHCI SATA platform driver
- *
-- * Copyright 2015 Freescale, Inc.
-+ * Copyright (C) 2015 Freescale Semiconductor, Inc.
- * Tang Yuantian <Yuantian.Tang@freescale.com>
- *
- * This program is free software; you can redistribute it and/or modify
-@@ -46,23 +46,32 @@
- #define LS1021A_AXICC_ADDR 0xC0
-
- #define SATA_ECC_DISABLE 0x00020000
-+#define ECC_DIS_ARMV8_CH2 0x80000000
-+#define ECC_DIS_LS1088A 0x40000000
-
- enum ahci_qoriq_type {
- AHCI_LS1021A,
- AHCI_LS1043A,
- AHCI_LS2080A,
-+ AHCI_LS1046A,
-+ AHCI_LS1088A,
-+ AHCI_LS2088A,
- };
-
- struct ahci_qoriq_priv {
- struct ccsr_ahci *reg_base;
- enum ahci_qoriq_type type;
- void __iomem *ecc_addr;
-+ bool is_dmacoherent;
- };
-
- static const struct of_device_id ahci_qoriq_of_match[] = {
- { .compatible = "fsl,ls1021a-ahci", .data = (void *)AHCI_LS1021A},
- { .compatible = "fsl,ls1043a-ahci", .data = (void *)AHCI_LS1043A},
- { .compatible = "fsl,ls2080a-ahci", .data = (void *)AHCI_LS2080A},
-+ { .compatible = "fsl,ls1046a-ahci", .data = (void *)AHCI_LS1046A},
-+ { .compatible = "fsl,ls1088a-ahci", .data = (void *)AHCI_LS1088A},
-+ { .compatible = "fsl,ls2088a-ahci", .data = (void *)AHCI_LS2088A},
- {},
- };
- MODULE_DEVICE_TABLE(of, ahci_qoriq_of_match);
-@@ -154,6 +163,8 @@ static int ahci_qoriq_phy_init(struct ah
-
- switch (qpriv->type) {
- case AHCI_LS1021A:
-+ if (!qpriv->ecc_addr)
-+ return -EINVAL;
- writel(SATA_ECC_DISABLE, qpriv->ecc_addr);
- writel(AHCI_PORT_PHY_1_CFG, reg_base + PORT_PHY1);
- writel(LS1021A_PORT_PHY2, reg_base + PORT_PHY2);
-@@ -161,19 +172,56 @@ static int ahci_qoriq_phy_init(struct ah
- writel(LS1021A_PORT_PHY4, reg_base + PORT_PHY4);
- writel(LS1021A_PORT_PHY5, reg_base + PORT_PHY5);
- writel(AHCI_PORT_TRANS_CFG, reg_base + PORT_TRANS);
-- writel(AHCI_PORT_AXICC_CFG, reg_base + LS1021A_AXICC_ADDR);
-+ if (qpriv->is_dmacoherent)
-+ writel(AHCI_PORT_AXICC_CFG,
-+ reg_base + LS1021A_AXICC_ADDR);
- break;
-
- case AHCI_LS1043A:
-+ if (!qpriv->ecc_addr)
-+ return -EINVAL;
-+ writel(readl(qpriv->ecc_addr) | ECC_DIS_ARMV8_CH2,
-+ qpriv->ecc_addr);
- writel(AHCI_PORT_PHY_1_CFG, reg_base + PORT_PHY1);
- writel(AHCI_PORT_TRANS_CFG, reg_base + PORT_TRANS);
-- writel(AHCI_PORT_AXICC_CFG, reg_base + PORT_AXICC);
-+ if (qpriv->is_dmacoherent)
-+ writel(AHCI_PORT_AXICC_CFG, reg_base + PORT_AXICC);
- break;
-
- case AHCI_LS2080A:
- writel(AHCI_PORT_PHY_1_CFG, reg_base + PORT_PHY1);
- writel(AHCI_PORT_TRANS_CFG, reg_base + PORT_TRANS);
-- writel(AHCI_PORT_AXICC_CFG, reg_base + PORT_AXICC);
-+ if (qpriv->is_dmacoherent)
-+ writel(AHCI_PORT_AXICC_CFG, reg_base + PORT_AXICC);
-+ break;
-+
-+ case AHCI_LS1046A:
-+ if (!qpriv->ecc_addr)
-+ return -EINVAL;
-+ writel(readl(qpriv->ecc_addr) | ECC_DIS_ARMV8_CH2,
-+ qpriv->ecc_addr);
-+ writel(AHCI_PORT_PHY_1_CFG, reg_base + PORT_PHY1);
-+ writel(AHCI_PORT_TRANS_CFG, reg_base + PORT_TRANS);
-+ if (qpriv->is_dmacoherent)
-+ writel(AHCI_PORT_AXICC_CFG, reg_base + PORT_AXICC);
-+ break;
-+
-+ case AHCI_LS1088A:
-+ if (!qpriv->ecc_addr)
-+ return -EINVAL;
-+ writel(readl(qpriv->ecc_addr) | ECC_DIS_LS1088A,
-+ qpriv->ecc_addr);
-+ writel(AHCI_PORT_PHY_1_CFG, reg_base + PORT_PHY1);
-+ writel(AHCI_PORT_TRANS_CFG, reg_base + PORT_TRANS);
-+ if (qpriv->is_dmacoherent)
-+ writel(AHCI_PORT_AXICC_CFG, reg_base + PORT_AXICC);
-+ break;
-+
-+ case AHCI_LS2088A:
-+ writel(AHCI_PORT_PHY_1_CFG, reg_base + PORT_PHY1);
-+ writel(AHCI_PORT_TRANS_CFG, reg_base + PORT_TRANS);
-+ if (qpriv->is_dmacoherent)
-+ writel(AHCI_PORT_AXICC_CFG, reg_base + PORT_AXICC);
- break;
- }
-
-@@ -204,13 +252,14 @@ static int ahci_qoriq_probe(struct platf
-
- qoriq_priv->type = (enum ahci_qoriq_type)of_id->data;
-
-- if (qoriq_priv->type == AHCI_LS1021A) {
-- res = platform_get_resource_byname(pdev, IORESOURCE_MEM,
-- "sata-ecc");
-+ res = platform_get_resource_byname(pdev, IORESOURCE_MEM,
-+ "sata-ecc");
-+ if (res) {
- qoriq_priv->ecc_addr = devm_ioremap_resource(dev, res);
- if (IS_ERR(qoriq_priv->ecc_addr))
- return PTR_ERR(qoriq_priv->ecc_addr);
- }
-+ qoriq_priv->is_dmacoherent = of_dma_is_coherent(np);
-
- rc = ahci_platform_enable_resources(hpriv);
- if (rc)
diff --git a/target/linux/layerscape/patches-4.9/802-clk-support-layerscape.patch b/target/linux/layerscape/patches-4.9/802-clk-support-layerscape.patch
deleted file mode 100644
index 0d05dc78ab..0000000000
--- a/target/linux/layerscape/patches-4.9/802-clk-support-layerscape.patch
+++ /dev/null
@@ -1,323 +0,0 @@
-From 82a391a067491f4c46b75d0dfe2bf9e5a11aca8e Mon Sep 17 00:00:00 2001
-From: Yangbo Lu <yangbo.lu@nxp.com>
-Date: Wed, 17 Jan 2018 15:15:44 +0800
-Subject: [PATCH 14/30] clk: support layerscape
-
-This is an integrated patch for layerscape clock support.
-
-Signed-off-by: Yuantian Tang <andy.tang@nxp.com>
-Signed-off-by: Mingkai Hu <mingkai.hu@nxp.com>
-Signed-off-by: Scott Wood <oss@buserror.net>
-Signed-off-by: Yangbo Lu <yangbo.lu@nxp.com>
----
- drivers/clk/clk-qoriq.c | 179 ++++++++++++++++++++++++++++++++++++++++++++----
- 1 file changed, 164 insertions(+), 15 deletions(-)
-
---- a/drivers/clk/clk-qoriq.c
-+++ b/drivers/clk/clk-qoriq.c
-@@ -12,6 +12,7 @@
-
- #include <linux/clk.h>
- #include <linux/clk-provider.h>
-+#include <linux/clkdev.h>
- #include <linux/fsl/guts.h>
- #include <linux/io.h>
- #include <linux/kernel.h>
-@@ -40,7 +41,7 @@ struct clockgen_pll_div {
- };
-
- struct clockgen_pll {
-- struct clockgen_pll_div div[4];
-+ struct clockgen_pll_div div[8];
- };
-
- #define CLKSEL_VALID 1
-@@ -87,7 +88,7 @@ struct clockgen {
- struct device_node *node;
- void __iomem *regs;
- struct clockgen_chipinfo info; /* mutable copy */
-- struct clk *sysclk;
-+ struct clk *sysclk, *coreclk;
- struct clockgen_pll pll[6];
- struct clk *cmux[NUM_CMUX];
- struct clk *hwaccel[NUM_HWACCEL];
-@@ -266,6 +267,39 @@ static const struct clockgen_muxinfo ls1
- },
- };
-
-+static const struct clockgen_muxinfo ls1046a_hwa1 = {
-+ {
-+ {},
-+ {},
-+ { CLKSEL_VALID, CGA_PLL1, PLL_DIV2 },
-+ { CLKSEL_VALID, CGA_PLL1, PLL_DIV3 },
-+ { CLKSEL_VALID, CGA_PLL1, PLL_DIV4 },
-+ { CLKSEL_VALID, PLATFORM_PLL, PLL_DIV1 },
-+ { CLKSEL_VALID, CGA_PLL2, PLL_DIV2 },
-+ { CLKSEL_VALID, CGA_PLL2, PLL_DIV3 },
-+ },
-+};
-+
-+static const struct clockgen_muxinfo ls1046a_hwa2 = {
-+ {
-+ {},
-+ { CLKSEL_VALID, CGA_PLL2, PLL_DIV1 },
-+ { CLKSEL_VALID, CGA_PLL2, PLL_DIV2 },
-+ { CLKSEL_VALID, CGA_PLL2, PLL_DIV3 },
-+ {},
-+ {},
-+ { CLKSEL_VALID, CGA_PLL1, PLL_DIV2 },
-+ },
-+};
-+
-+static const struct clockgen_muxinfo ls1012a_cmux = {
-+ {
-+ [0] = { CLKSEL_VALID, CGA_PLL1, PLL_DIV1 },
-+ {},
-+ [2] = { CLKSEL_VALID, CGA_PLL1, PLL_DIV2 },
-+ }
-+};
-+
- static const struct clockgen_muxinfo t1023_hwa1 = {
- {
- {},
-@@ -489,6 +523,42 @@ static const struct clockgen_chipinfo ch
- .flags = CG_PLL_8BIT,
- },
- {
-+ .compat = "fsl,ls1046a-clockgen",
-+ .init_periph = t2080_init_periph,
-+ .cmux_groups = {
-+ &t1040_cmux
-+ },
-+ .hwaccel = {
-+ &ls1046a_hwa1, &ls1046a_hwa2
-+ },
-+ .cmux_to_group = {
-+ 0, -1
-+ },
-+ .pll_mask = 0x07,
-+ .flags = CG_PLL_8BIT,
-+ },
-+ {
-+ .compat = "fsl,ls1088a-clockgen",
-+ .cmux_groups = {
-+ &clockgen2_cmux_cga12
-+ },
-+ .cmux_to_group = {
-+ 0, 0, -1
-+ },
-+ .pll_mask = 0x07,
-+ .flags = CG_VER3 | CG_LITTLE_ENDIAN,
-+ },
-+ {
-+ .compat = "fsl,ls1012a-clockgen",
-+ .cmux_groups = {
-+ &ls1012a_cmux
-+ },
-+ .cmux_to_group = {
-+ 0, -1
-+ },
-+ .pll_mask = 0x03,
-+ },
-+ {
- .compat = "fsl,ls2080a-clockgen",
- .cmux_groups = {
- &clockgen2_cmux_cga12, &clockgen2_cmux_cgb
-@@ -846,7 +916,12 @@ static void __init create_muxes(struct c
-
- static void __init clockgen_init(struct device_node *np);
-
--/* Legacy nodes may get probed before the parent clockgen node */
-+/*
-+ * Legacy nodes may get probed before the parent clockgen node.
-+ * It is assumed that device trees with legacy nodes will not
-+ * contain a "clocks" property -- otherwise the input clocks may
-+ * not be initialized at this point.
-+ */
- static void __init legacy_init_clockgen(struct device_node *np)
- {
- if (!clockgen.node)
-@@ -887,18 +962,13 @@ static struct clk __init
- return clk_register_fixed_rate(NULL, name, NULL, 0, rate);
- }
-
--static struct clk *sysclk_from_parent(const char *name)
-+static struct clk __init *input_clock(const char *name, struct clk *clk)
- {
-- struct clk *clk;
-- const char *parent_name;
--
-- clk = of_clk_get(clockgen.node, 0);
-- if (IS_ERR(clk))
-- return clk;
-+ const char *input_name;
-
- /* Register the input clock under the desired name. */
-- parent_name = __clk_get_name(clk);
-- clk = clk_register_fixed_factor(NULL, name, parent_name,
-+ input_name = __clk_get_name(clk);
-+ clk = clk_register_fixed_factor(NULL, name, input_name,
- 0, 1, 1);
- if (IS_ERR(clk))
- pr_err("%s: Couldn't register %s: %ld\n", __func__, name,
-@@ -907,6 +977,29 @@ static struct clk *sysclk_from_parent(co
- return clk;
- }
-
-+static struct clk __init *input_clock_by_name(const char *name,
-+ const char *dtname)
-+{
-+ struct clk *clk;
-+
-+ clk = of_clk_get_by_name(clockgen.node, dtname);
-+ if (IS_ERR(clk))
-+ return clk;
-+
-+ return input_clock(name, clk);
-+}
-+
-+static struct clk __init *input_clock_by_index(const char *name, int idx)
-+{
-+ struct clk *clk;
-+
-+ clk = of_clk_get(clockgen.node, 0);
-+ if (IS_ERR(clk))
-+ return clk;
-+
-+ return input_clock(name, clk);
-+}
-+
- static struct clk * __init create_sysclk(const char *name)
- {
- struct device_node *sysclk;
-@@ -916,7 +1009,11 @@ static struct clk * __init create_sysclk
- if (!IS_ERR(clk))
- return clk;
-
-- clk = sysclk_from_parent(name);
-+ clk = input_clock_by_name(name, "sysclk");
-+ if (!IS_ERR(clk))
-+ return clk;
-+
-+ clk = input_clock_by_index(name, 0);
- if (!IS_ERR(clk))
- return clk;
-
-@@ -927,7 +1024,27 @@ static struct clk * __init create_sysclk
- return clk;
- }
-
-- pr_err("%s: No input clock\n", __func__);
-+ pr_err("%s: No input sysclk\n", __func__);
-+ return NULL;
-+}
-+
-+static struct clk * __init create_coreclk(const char *name)
-+{
-+ struct clk *clk;
-+
-+ clk = input_clock_by_name(name, "coreclk");
-+ if (!IS_ERR(clk))
-+ return clk;
-+
-+ /*
-+ * This indicates a mix of legacy nodes with the new coreclk
-+ * mechanism, which should never happen. If this error occurs,
-+ * don't use the wrong input clock just because coreclk isn't
-+ * ready yet.
-+ */
-+ if (WARN_ON(PTR_ERR(clk) == -EPROBE_DEFER))
-+ return clk;
-+
- return NULL;
- }
-
-@@ -950,11 +1067,19 @@ static void __init create_one_pll(struct
- u32 __iomem *reg;
- u32 mult;
- struct clockgen_pll *pll = &cg->pll[idx];
-+ const char *input = "cg-sysclk";
- int i;
-
- if (!(cg->info.pll_mask & (1 << idx)))
- return;
-
-+ if (cg->coreclk && idx != PLATFORM_PLL) {
-+ if (IS_ERR(cg->coreclk))
-+ return;
-+
-+ input = "cg-coreclk";
-+ }
-+
- if (cg->info.flags & CG_VER3) {
- switch (idx) {
- case PLATFORM_PLL:
-@@ -1000,12 +1125,20 @@ static void __init create_one_pll(struct
-
- for (i = 0; i < ARRAY_SIZE(pll->div); i++) {
- struct clk *clk;
-+ int ret;
-+
-+ /*
-+ * For platform PLL, there are 8 divider clocks.
-+ * For core PLL, there are 4 divider clocks at most.
-+ */
-+ if (idx != 0 && i >= 4)
-+ break;
-
- snprintf(pll->div[i].name, sizeof(pll->div[i].name),
- "cg-pll%d-div%d", idx, i + 1);
-
- clk = clk_register_fixed_factor(NULL,
-- pll->div[i].name, "cg-sysclk", 0, mult, i + 1);
-+ pll->div[i].name, input, 0, mult, i + 1);
- if (IS_ERR(clk)) {
- pr_err("%s: %s: register failed %ld\n",
- __func__, pll->div[i].name, PTR_ERR(clk));
-@@ -1013,6 +1146,11 @@ static void __init create_one_pll(struct
- }
-
- pll->div[i].clk = clk;
-+ ret = clk_register_clkdev(clk, pll->div[i].name, NULL);
-+ if (ret != 0)
-+ pr_err("%s: %s: register to lookup table failed %ld\n",
-+ __func__, pll->div[i].name, PTR_ERR(clk));
-+
- }
- }
-
-@@ -1142,6 +1280,13 @@ static struct clk *clockgen_clk_get(stru
- goto bad_args;
- clk = pll->div[idx].clk;
- break;
-+ case 5:
-+ if (idx != 0)
-+ goto bad_args;
-+ clk = cg->coreclk;
-+ if (IS_ERR(clk))
-+ clk = NULL;
-+ break;
- default:
- goto bad_args;
- }
-@@ -1253,6 +1398,7 @@ static void __init clockgen_init(struct
- clockgen.info.flags |= CG_CMUX_GE_PLAT;
-
- clockgen.sysclk = create_sysclk("cg-sysclk");
-+ clockgen.coreclk = create_coreclk("cg-coreclk");
- create_plls(&clockgen);
- create_muxes(&clockgen);
-
-@@ -1273,8 +1419,11 @@ err:
-
- CLK_OF_DECLARE(qoriq_clockgen_1, "fsl,qoriq-clockgen-1.0", clockgen_init);
- CLK_OF_DECLARE(qoriq_clockgen_2, "fsl,qoriq-clockgen-2.0", clockgen_init);
-+CLK_OF_DECLARE(qoriq_clockgen_ls1012a, "fsl,ls1012a-clockgen", clockgen_init);
- CLK_OF_DECLARE(qoriq_clockgen_ls1021a, "fsl,ls1021a-clockgen", clockgen_init);
- CLK_OF_DECLARE(qoriq_clockgen_ls1043a, "fsl,ls1043a-clockgen", clockgen_init);
-+CLK_OF_DECLARE(qoriq_clockgen_ls1046a, "fsl,ls1046a-clockgen", clockgen_init);
-+CLK_OF_DECLARE(qoriq_clockgen_ls1088a, "fsl,ls1088a-clockgen", clockgen_init);
- CLK_OF_DECLARE(qoriq_clockgen_ls2080a, "fsl,ls2080a-clockgen", clockgen_init);
-
- /* Legacy nodes */
diff --git a/target/linux/layerscape/patches-4.9/803-cpufreq-support-layerscape.patch b/target/linux/layerscape/patches-4.9/803-cpufreq-support-layerscape.patch
deleted file mode 100644
index cf2b42a547..0000000000
--- a/target/linux/layerscape/patches-4.9/803-cpufreq-support-layerscape.patch
+++ /dev/null
@@ -1,1030 +0,0 @@
-From d78d78ccbaded757e8bea0d13c4120518bdd4660 Mon Sep 17 00:00:00 2001
-From: Yangbo Lu <yangbo.lu@nxp.com>
-Date: Thu, 5 Jul 2018 17:21:38 +0800
-Subject: [PATCH 15/32] cpufreq: support layerscape
-
-This is an integrated patch for layerscape pm support.
-
-Signed-off-by: Tang Yuantian <Yuantian.Tang@nxp.com>
-Signed-off-by: Yangbo Lu <yangbo.lu@nxp.com>
----
- .../devicetree/bindings/powerpc/fsl/pmc.txt | 59 ++--
- drivers/cpufreq/Kconfig | 2 +-
- drivers/cpufreq/qoriq-cpufreq.c | 176 +++++------
- drivers/firmware/psci.c | 12 +-
- drivers/soc/fsl/rcpm.c | 158 ++++++++++
- drivers/soc/fsl/sleep_fsm.c | 279 ++++++++++++++++++
- drivers/soc/fsl/sleep_fsm.h | 130 ++++++++
- 7 files changed, 678 insertions(+), 138 deletions(-)
- create mode 100644 drivers/soc/fsl/rcpm.c
- create mode 100644 drivers/soc/fsl/sleep_fsm.c
- create mode 100644 drivers/soc/fsl/sleep_fsm.h
-
---- a/Documentation/devicetree/bindings/powerpc/fsl/pmc.txt
-+++ b/Documentation/devicetree/bindings/powerpc/fsl/pmc.txt
-@@ -9,15 +9,20 @@ Properties:
-
- "fsl,mpc8548-pmc" should be listed for any chip whose PMC is
- compatible. "fsl,mpc8536-pmc" should also be listed for any chip
-- whose PMC is compatible, and implies deep-sleep capability.
-+ whose PMC is compatible, and implies deep-sleep capability and
-+ wake on user defined packet(wakeup on ARP).
-+
-+ "fsl,p1022-pmc" should be listed for any chip whose PMC is
-+ compatible, and implies lossless Ethernet capability during sleep.
-
- "fsl,mpc8641d-pmc" should be listed for any chip whose PMC is
- compatible; all statements below that apply to "fsl,mpc8548-pmc" also
- apply to "fsl,mpc8641d-pmc".
-
- Compatibility does not include bit assignments in SCCR/PMCDR/DEVDISR; these
-- bit assignments are indicated via the sleep specifier in each device's
-- sleep property.
-+ bit assignments are indicated via the clock nodes. Device which has a
-+ controllable clock source should have a "fsl,pmc-handle" property pointing
-+ to the clock node.
-
- - reg: For devices compatible with "fsl,mpc8349-pmc", the first resource
- is the PMC block, and the second resource is the Clock Configuration
-@@ -33,31 +38,35 @@ Properties:
- this is a phandle to an "fsl,gtm" node on which timer 4 can be used as
- a wakeup source from deep sleep.
-
--Sleep specifiers:
--
-- fsl,mpc8349-pmc: Sleep specifiers consist of one cell. For each bit
-- that is set in the cell, the corresponding bit in SCCR will be saved
-- and cleared on suspend, and restored on resume. This sleep controller
-- supports disabling and resuming devices at any time.
--
-- fsl,mpc8536-pmc: Sleep specifiers consist of three cells, the third of
-- which will be ORed into PMCDR upon suspend, and cleared from PMCDR
-- upon resume. The first two cells are as described for fsl,mpc8578-pmc.
-- This sleep controller only supports disabling devices during system
-- sleep, or permanently.
--
-- fsl,mpc8548-pmc: Sleep specifiers consist of one or two cells, the
-- first of which will be ORed into DEVDISR (and the second into
-- DEVDISR2, if present -- this cell should be zero or absent if the
-- hardware does not have DEVDISR2) upon a request for permanent device
-- disabling. This sleep controller does not support configuring devices
-- to disable during system sleep (unless supported by another compatible
-- match), or dynamically.
-+Clock nodes:
-+The clock nodes are to describe the masks in PM controller registers for each
-+soc clock.
-+- fsl,pmcdr-mask: For "fsl,mpc8548-pmc"-compatible devices, the mask will be
-+ ORed into PMCDR before suspend if the device using this clock is the wake-up
-+ source and need to be running during low power mode; clear the mask if
-+ otherwise.
-+
-+- fsl,sccr-mask: For "fsl,mpc8349-pmc"-compatible devices, the corresponding
-+ bit specified by the mask in SCCR will be saved and cleared on suspend, and
-+ restored on resume.
-+
-+- fsl,devdisr-mask: Contain one or two cells, depending on the availability of
-+ DEVDISR2 register. For compatible devices, the mask will be ORed into DEVDISR
-+ or DEVDISR2 when the clock should be permenently disabled.
-
- Example:
-
-- power@b00 {
-- compatible = "fsl,mpc8313-pmc", "fsl,mpc8349-pmc";
-- reg = <0xb00 0x100 0xa00 0x100>;
-- interrupts = <80 8>;
-+ power@e0070 {
-+ compatible = "fsl,mpc8536-pmc", "fsl,mpc8548-pmc";
-+ reg = <0xe0070 0x20>;
-+
-+ etsec1_clk: soc-clk@24 {
-+ fsl,pmcdr-mask = <0x00000080>;
-+ };
-+ etsec2_clk: soc-clk@25 {
-+ fsl,pmcdr-mask = <0x00000040>;
-+ };
-+ etsec3_clk: soc-clk@26 {
-+ fsl,pmcdr-mask = <0x00000020>;
-+ };
- };
---- a/drivers/cpufreq/Kconfig
-+++ b/drivers/cpufreq/Kconfig
-@@ -334,7 +334,7 @@ endif
-
- config QORIQ_CPUFREQ
- tristate "CPU frequency scaling driver for Freescale QorIQ SoCs"
-- depends on OF && COMMON_CLK && (PPC_E500MC || ARM)
-+ depends on OF && COMMON_CLK && (PPC_E500MC || ARM || ARM64)
- depends on !CPU_THERMAL || THERMAL
- select CLK_QORIQ
- help
---- a/drivers/cpufreq/qoriq-cpufreq.c
-+++ b/drivers/cpufreq/qoriq-cpufreq.c
-@@ -11,6 +11,7 @@
- #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
-
- #include <linux/clk.h>
-+#include <linux/clk-provider.h>
- #include <linux/cpufreq.h>
- #include <linux/cpu_cooling.h>
- #include <linux/errno.h>
-@@ -22,10 +23,6 @@
- #include <linux/slab.h>
- #include <linux/smp.h>
-
--#if !defined(CONFIG_ARM)
--#include <asm/smp.h> /* for get_hard_smp_processor_id() in UP configs */
--#endif
--
- /**
- * struct cpu_data
- * @pclk: the parent clock of cpu
-@@ -37,73 +34,51 @@ struct cpu_data {
- struct thermal_cooling_device *cdev;
- };
-
-+/*
-+ * Don't use cpufreq on this SoC -- used when the SoC would have otherwise
-+ * matched a more generic compatible.
-+ */
-+#define SOC_BLACKLIST 1
-+
- /**
- * struct soc_data - SoC specific data
-- * @freq_mask: mask the disallowed frequencies
-- * @flag: unique flags
-+ * @flags: SOC_xxx
- */
- struct soc_data {
-- u32 freq_mask[4];
-- u32 flag;
--};
--
--#define FREQ_MASK 1
--/* see hardware specification for the allowed frqeuencies */
--static const struct soc_data sdata[] = {
-- { /* used by p2041 and p3041 */
-- .freq_mask = {0x8, 0x8, 0x2, 0x2},
-- .flag = FREQ_MASK,
-- },
-- { /* used by p5020 */
-- .freq_mask = {0x8, 0x2},
-- .flag = FREQ_MASK,
-- },
-- { /* used by p4080, p5040 */
-- .freq_mask = {0},
-- .flag = 0,
-- },
-+ u32 flags;
- };
-
--/*
-- * the minimum allowed core frequency, in Hz
-- * for chassis v1.0, >= platform frequency
-- * for chassis v2.0, >= platform frequency / 2
-- */
--static u32 min_cpufreq;
--static const u32 *fmask;
--
--#if defined(CONFIG_ARM)
--static int get_cpu_physical_id(int cpu)
--{
-- return topology_core_id(cpu);
--}
--#else
--static int get_cpu_physical_id(int cpu)
--{
-- return get_hard_smp_processor_id(cpu);
--}
--#endif
--
- static u32 get_bus_freq(void)
- {
- struct device_node *soc;
- u32 sysfreq;
-+ struct clk *pltclk;
-+ int ret;
-
-+ /* get platform freq by searching bus-frequency property */
- soc = of_find_node_by_type(NULL, "soc");
-- if (!soc)
-- return 0;
--
-- if (of_property_read_u32(soc, "bus-frequency", &sysfreq))
-- sysfreq = 0;
-+ if (soc) {
-+ ret = of_property_read_u32(soc, "bus-frequency", &sysfreq);
-+ of_node_put(soc);
-+ if (!ret)
-+ return sysfreq;
-+ }
-
-- of_node_put(soc);
-+ /* get platform freq by its clock name */
-+ pltclk = clk_get(NULL, "cg-pll0-div1");
-+ if (IS_ERR(pltclk)) {
-+ pr_err("%s: can't get bus frequency %ld\n",
-+ __func__, PTR_ERR(pltclk));
-+ return PTR_ERR(pltclk);
-+ }
-
-- return sysfreq;
-+ return clk_get_rate(pltclk);
- }
-
--static struct device_node *cpu_to_clk_node(int cpu)
-+static struct clk *cpu_to_clk(int cpu)
- {
-- struct device_node *np, *clk_np;
-+ struct device_node *np;
-+ struct clk *clk;
-
- if (!cpu_present(cpu))
- return NULL;
-@@ -112,37 +87,28 @@ static struct device_node *cpu_to_clk_no
- if (!np)
- return NULL;
-
-- clk_np = of_parse_phandle(np, "clocks", 0);
-- if (!clk_np)
-- return NULL;
--
-+ clk = of_clk_get(np, 0);
- of_node_put(np);
--
-- return clk_np;
-+ return clk;
- }
-
- /* traverse cpu nodes to get cpu mask of sharing clock wire */
- static void set_affected_cpus(struct cpufreq_policy *policy)
- {
-- struct device_node *np, *clk_np;
- struct cpumask *dstp = policy->cpus;
-+ struct clk *clk;
- int i;
-
-- np = cpu_to_clk_node(policy->cpu);
-- if (!np)
-- return;
--
- for_each_present_cpu(i) {
-- clk_np = cpu_to_clk_node(i);
-- if (!clk_np)
-+ clk = cpu_to_clk(i);
-+ if (IS_ERR(clk)) {
-+ pr_err("%s: no clock for cpu %d\n", __func__, i);
- continue;
-+ }
-
-- if (clk_np == np)
-+ if (clk_is_match(policy->clk, clk))
- cpumask_set_cpu(i, dstp);
--
-- of_node_put(clk_np);
- }
-- of_node_put(np);
- }
-
- /* reduce the duplicated frequencies in frequency table */
-@@ -198,10 +164,11 @@ static void freq_table_sort(struct cpufr
-
- static int qoriq_cpufreq_cpu_init(struct cpufreq_policy *policy)
- {
-- struct device_node *np, *pnode;
-+ struct device_node *np;
- int i, count, ret;
-- u32 freq, mask;
-+ u32 freq;
- struct clk *clk;
-+ const struct clk_hw *hwclk;
- struct cpufreq_frequency_table *table;
- struct cpu_data *data;
- unsigned int cpu = policy->cpu;
-@@ -221,17 +188,13 @@ static int qoriq_cpufreq_cpu_init(struct
- goto err_nomem2;
- }
-
-- pnode = of_parse_phandle(np, "clocks", 0);
-- if (!pnode) {
-- pr_err("%s: could not get clock information\n", __func__);
-- goto err_nomem2;
-- }
-+ hwclk = __clk_get_hw(policy->clk);
-+ count = clk_hw_get_num_parents(hwclk);
-
-- count = of_property_count_strings(pnode, "clock-names");
- data->pclk = kcalloc(count, sizeof(struct clk *), GFP_KERNEL);
- if (!data->pclk) {
- pr_err("%s: no memory\n", __func__);
-- goto err_node;
-+ goto err_nomem2;
- }
-
- table = kcalloc(count + 1, sizeof(*table), GFP_KERNEL);
-@@ -240,23 +203,11 @@ static int qoriq_cpufreq_cpu_init(struct
- goto err_pclk;
- }
-
-- if (fmask)
-- mask = fmask[get_cpu_physical_id(cpu)];
-- else
-- mask = 0x0;
--
- for (i = 0; i < count; i++) {
-- clk = of_clk_get(pnode, i);
-+ clk = clk_hw_get_parent_by_index(hwclk, i)->clk;
- data->pclk[i] = clk;
- freq = clk_get_rate(clk);
-- /*
-- * the clock is valid if its frequency is not masked
-- * and large than minimum allowed frequency.
-- */
-- if (freq < min_cpufreq || (mask & (1 << i)))
-- table[i].frequency = CPUFREQ_ENTRY_INVALID;
-- else
-- table[i].frequency = freq / 1000;
-+ table[i].frequency = freq / 1000;
- table[i].driver_data = i;
- }
- freq_table_redup(table, count);
-@@ -282,7 +233,6 @@ static int qoriq_cpufreq_cpu_init(struct
- policy->cpuinfo.transition_latency = u64temp + 1;
-
- of_node_put(np);
-- of_node_put(pnode);
-
- return 0;
-
-@@ -290,10 +240,7 @@ err_nomem1:
- kfree(table);
- err_pclk:
- kfree(data->pclk);
--err_node:
-- of_node_put(pnode);
- err_nomem2:
-- policy->driver_data = NULL;
- kfree(data);
- err_np:
- of_node_put(np);
-@@ -357,12 +304,25 @@ static struct cpufreq_driver qoriq_cpufr
- .attr = cpufreq_generic_attr,
- };
-
-+static const struct soc_data blacklist = {
-+ .flags = SOC_BLACKLIST,
-+};
-+
- static const struct of_device_id node_matches[] __initconst = {
-- { .compatible = "fsl,p2041-clockgen", .data = &sdata[0], },
-- { .compatible = "fsl,p3041-clockgen", .data = &sdata[0], },
-- { .compatible = "fsl,p5020-clockgen", .data = &sdata[1], },
-- { .compatible = "fsl,p4080-clockgen", .data = &sdata[2], },
-- { .compatible = "fsl,p5040-clockgen", .data = &sdata[2], },
-+ /* e6500 cannot use cpufreq due to erratum A-008083 */
-+ { .compatible = "fsl,b4420-clockgen", &blacklist },
-+ { .compatible = "fsl,b4860-clockgen", &blacklist },
-+ { .compatible = "fsl,t2080-clockgen", &blacklist },
-+ { .compatible = "fsl,t4240-clockgen", &blacklist },
-+
-+ { .compatible = "fsl,ls1012a-clockgen", },
-+ { .compatible = "fsl,ls1021a-clockgen", },
-+ { .compatible = "fsl,ls1043a-clockgen", },
-+ { .compatible = "fsl,ls1046a-clockgen", },
-+ { .compatible = "fsl,ls1088a-clockgen", },
-+ { .compatible = "fsl,ls2080a-clockgen", },
-+ { .compatible = "fsl,p4080-clockgen", },
-+ { .compatible = "fsl,qoriq-clockgen-1.0", },
- { .compatible = "fsl,qoriq-clockgen-2.0", },
- {}
- };
-@@ -380,16 +340,12 @@ static int __init qoriq_cpufreq_init(voi
-
- match = of_match_node(node_matches, np);
- data = match->data;
-- if (data) {
-- if (data->flag)
-- fmask = data->freq_mask;
-- min_cpufreq = get_bus_freq();
-- } else {
-- min_cpufreq = get_bus_freq() / 2;
-- }
-
- of_node_put(np);
-
-+ if (data && data->flags & SOC_BLACKLIST)
-+ return -ENODEV;
-+
- ret = cpufreq_register_driver(&qoriq_cpufreq_driver);
- if (!ret)
- pr_info("Freescale QorIQ CPU frequency scaling driver\n");
---- a/drivers/firmware/psci.c
-+++ b/drivers/firmware/psci.c
-@@ -437,8 +437,12 @@ CPUIDLE_METHOD_OF_DECLARE(psci, "psci",
-
- static int psci_system_suspend(unsigned long unused)
- {
-- return invoke_psci_fn(PSCI_FN_NATIVE(1_0, SYSTEM_SUSPEND),
-- virt_to_phys(cpu_resume), 0, 0);
-+ u32 state;
-+
-+ state = ( 2 << PSCI_0_2_POWER_STATE_AFFL_SHIFT) |
-+ (1 << PSCI_0_2_POWER_STATE_TYPE_SHIFT);
-+
-+ return psci_cpu_suspend(state, virt_to_phys(cpu_resume));
- }
-
- static int psci_system_suspend_enter(suspend_state_t state)
-@@ -458,6 +462,8 @@ static void __init psci_init_system_susp
- if (!IS_ENABLED(CONFIG_SUSPEND))
- return;
-
-+ suspend_set_ops(&psci_suspend_ops);
-+
- ret = psci_features(PSCI_FN_NATIVE(1_0, SYSTEM_SUSPEND));
-
- if (ret != PSCI_RET_NOT_SUPPORTED)
-@@ -562,6 +568,8 @@ static void __init psci_0_2_set_function
- arm_pm_restart = psci_sys_reset;
-
- pm_power_off = psci_sys_poweroff;
-+ psci_init_system_suspend();
-+ suspend_set_ops(&psci_suspend_ops);
- }
-
- /*
---- /dev/null
-+++ b/drivers/soc/fsl/rcpm.c
-@@ -0,0 +1,158 @@
-+/*
-+ * Run Control and Power Management (RCPM) driver
-+ *
-+ * Copyright 2016 NXP
-+ *
-+ * This program is free software; you can redistribute it and/or modify
-+ * it under the terms of the GNU General Public License as published by
-+ * the Free Software Foundation; either version 2 of the License, or
-+ * (at your option) any later version.
-+ *
-+ * This program is distributed in the hope that it will be useful,
-+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
-+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
-+ * GNU General Public License for more details.
-+ *
-+ */
-+#define pr_fmt(fmt) "RCPM: %s: " fmt, __func__
-+
-+#include <linux/kernel.h>
-+#include <linux/io.h>
-+#include <linux/of_platform.h>
-+#include <linux/of_address.h>
-+#include <linux/suspend.h>
-+
-+/* RCPM register offset */
-+#define RCPM_IPPDEXPCR0 0x140
-+
-+#define RCPM_WAKEUP_CELL_SIZE 2
-+
-+struct rcpm_config {
-+ int ipp_num;
-+ int ippdexpcr_offset;
-+ u32 ippdexpcr[2];
-+ void *rcpm_reg_base;
-+};
-+
-+static struct rcpm_config *rcpm;
-+
-+static inline void rcpm_reg_write(u32 offset, u32 value)
-+{
-+ iowrite32be(value, rcpm->rcpm_reg_base + offset);
-+}
-+
-+static inline u32 rcpm_reg_read(u32 offset)
-+{
-+ return ioread32be(rcpm->rcpm_reg_base + offset);
-+}
-+
-+static void rcpm_wakeup_fixup(struct device *dev, void *data)
-+{
-+ struct device_node *node = dev ? dev->of_node : NULL;
-+ u32 value[RCPM_WAKEUP_CELL_SIZE];
-+ int ret, i;
-+
-+ if (!dev || !node || !device_may_wakeup(dev))
-+ return;
-+
-+ /*
-+ * Get the values in the "rcpm-wakeup" property.
-+ * Three values are:
-+ * The first is a pointer to the RCPM node.
-+ * The second is the value of the ippdexpcr0 register.
-+ * The third is the value of the ippdexpcr1 register.
-+ */
-+ ret = of_property_read_u32_array(node, "fsl,rcpm-wakeup",
-+ value, RCPM_WAKEUP_CELL_SIZE);
-+ if (ret)
-+ return;
-+
-+ pr_debug("wakeup source: the device %s\n", node->full_name);
-+
-+ for (i = 0; i < rcpm->ipp_num; i++)
-+ rcpm->ippdexpcr[i] |= value[i + 1];
-+}
-+
-+static int rcpm_suspend_prepare(void)
-+{
-+ int i;
-+ u32 val;
-+
-+ BUG_ON(!rcpm);
-+
-+ for (i = 0; i < rcpm->ipp_num; i++)
-+ rcpm->ippdexpcr[i] = 0;
-+
-+ dpm_for_each_dev(NULL, rcpm_wakeup_fixup);
-+
-+ for (i = 0; i < rcpm->ipp_num; i++) {
-+ if (rcpm->ippdexpcr[i]) {
-+ val = rcpm_reg_read(rcpm->ippdexpcr_offset + 4 * i);
-+ rcpm_reg_write(rcpm->ippdexpcr_offset + 4 * i,
-+ val | rcpm->ippdexpcr[i]);
-+ pr_debug("ippdexpcr%d = 0x%x\n", i, rcpm->ippdexpcr[i]);
-+ }
-+ }
-+
-+ return 0;
-+}
-+
-+static int rcpm_suspend_notifier_call(struct notifier_block *bl,
-+ unsigned long state,
-+ void *unused)
-+{
-+ switch (state) {
-+ case PM_SUSPEND_PREPARE:
-+ rcpm_suspend_prepare();
-+ break;
-+ }
-+
-+ return NOTIFY_DONE;
-+}
-+
-+static struct rcpm_config rcpm_default_config = {
-+ .ipp_num = 1,
-+ .ippdexpcr_offset = RCPM_IPPDEXPCR0,
-+};
-+
-+static const struct of_device_id rcpm_matches[] = {
-+ {
-+ .compatible = "fsl,qoriq-rcpm-2.1",
-+ .data = &rcpm_default_config,
-+ },
-+ {}
-+};
-+
-+static struct notifier_block rcpm_suspend_notifier = {
-+ .notifier_call = rcpm_suspend_notifier_call,
-+};
-+
-+static int __init layerscape_rcpm_init(void)
-+{
-+ const struct of_device_id *match;
-+ struct device_node *np;
-+
-+ np = of_find_matching_node_and_match(NULL, rcpm_matches, &match);
-+ if (!np) {
-+ pr_err("Can't find the RCPM node.\n");
-+ return -EINVAL;
-+ }
-+
-+ if (match->data)
-+ rcpm = (struct rcpm_config *)match->data;
-+ else
-+ return -EINVAL;
-+
-+ rcpm->rcpm_reg_base = of_iomap(np, 0);
-+ of_node_put(np);
-+ if (!rcpm->rcpm_reg_base)
-+ return -ENOMEM;
-+
-+ register_pm_notifier(&rcpm_suspend_notifier);
-+
-+ pr_info("The RCPM driver initialized.\n");
-+
-+ return 0;
-+}
-+
-+subsys_initcall(layerscape_rcpm_init);
---- /dev/null
-+++ b/drivers/soc/fsl/sleep_fsm.c
-@@ -0,0 +1,279 @@
-+/*
-+ * deep sleep FSM (finite-state machine) configuration
-+ *
-+ * Copyright 2018 NXP
-+ *
-+ * Author: Hongbo Zhang <hongbo.zhang@freescale.com>
-+ * Chenhui Zhao <chenhui.zhao@freescale.com>
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of the above-listed copyright holders nor the
-+ * names of any contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
-+ * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
-+ * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
-+ * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE
-+ * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
-+ * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
-+ * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
-+ * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
-+ * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
-+ * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
-+ * POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#include <linux/kernel.h>
-+#include <linux/io.h>
-+#include <linux/types.h>
-+
-+#include "sleep_fsm.h"
-+/*
-+ * These values are from chip's reference manual. For example,
-+ * the values for T1040 can be found in "8.4.3.8 Programming
-+ * supporting deep sleep mode" of Chapter 8 "Run Control and
-+ * Power Management (RCPM)".
-+ * The default value can be applied to T104x, LS1021.
-+ */
-+struct fsm_reg_vals epu_default_val[] = {
-+ /* EPGCR (Event Processor Global Control Register) */
-+ {EPGCR, 0},
-+ /* EPECR (Event Processor Event Control Registers) */
-+ {EPECR0 + EPECR_STRIDE * 0, 0},
-+ {EPECR0 + EPECR_STRIDE * 1, 0},
-+ {EPECR0 + EPECR_STRIDE * 2, 0xF0004004},
-+ {EPECR0 + EPECR_STRIDE * 3, 0x80000084},
-+ {EPECR0 + EPECR_STRIDE * 4, 0x20000084},
-+ {EPECR0 + EPECR_STRIDE * 5, 0x08000004},
-+ {EPECR0 + EPECR_STRIDE * 6, 0x80000084},
-+ {EPECR0 + EPECR_STRIDE * 7, 0x80000084},
-+ {EPECR0 + EPECR_STRIDE * 8, 0x60000084},
-+ {EPECR0 + EPECR_STRIDE * 9, 0x08000084},
-+ {EPECR0 + EPECR_STRIDE * 10, 0x42000084},
-+ {EPECR0 + EPECR_STRIDE * 11, 0x90000084},
-+ {EPECR0 + EPECR_STRIDE * 12, 0x80000084},
-+ {EPECR0 + EPECR_STRIDE * 13, 0x08000084},
-+ {EPECR0 + EPECR_STRIDE * 14, 0x02000084},
-+ {EPECR0 + EPECR_STRIDE * 15, 0x00000004},
-+ /*
-+ * EPEVTCR (Event Processor EVT Pin Control Registers)
-+ * SCU8 triger EVT2, and SCU11 triger EVT9
-+ */
-+ {EPEVTCR0 + EPEVTCR_STRIDE * 0, 0},
-+ {EPEVTCR0 + EPEVTCR_STRIDE * 1, 0},
-+ {EPEVTCR0 + EPEVTCR_STRIDE * 2, 0x80000001},
-+ {EPEVTCR0 + EPEVTCR_STRIDE * 3, 0},
-+ {EPEVTCR0 + EPEVTCR_STRIDE * 4, 0},
-+ {EPEVTCR0 + EPEVTCR_STRIDE * 5, 0},
-+ {EPEVTCR0 + EPEVTCR_STRIDE * 6, 0},
-+ {EPEVTCR0 + EPEVTCR_STRIDE * 7, 0},
-+ {EPEVTCR0 + EPEVTCR_STRIDE * 8, 0},
-+ {EPEVTCR0 + EPEVTCR_STRIDE * 9, 0xB0000001},
-+ /* EPCMPR (Event Processor Counter Compare Registers) */
-+ {EPCMPR0 + EPCMPR_STRIDE * 0, 0},
-+ {EPCMPR0 + EPCMPR_STRIDE * 1, 0},
-+ {EPCMPR0 + EPCMPR_STRIDE * 2, 0x000000FF},
-+ {EPCMPR0 + EPCMPR_STRIDE * 3, 0},
-+ {EPCMPR0 + EPCMPR_STRIDE * 4, 0x000000FF},
-+ {EPCMPR0 + EPCMPR_STRIDE * 5, 0x00000020},
-+ {EPCMPR0 + EPCMPR_STRIDE * 6, 0},
-+ {EPCMPR0 + EPCMPR_STRIDE * 7, 0},
-+ {EPCMPR0 + EPCMPR_STRIDE * 8, 0x000000FF},
-+ {EPCMPR0 + EPCMPR_STRIDE * 9, 0x000000FF},
-+ {EPCMPR0 + EPCMPR_STRIDE * 10, 0x000000FF},
-+ {EPCMPR0 + EPCMPR_STRIDE * 11, 0x000000FF},
-+ {EPCMPR0 + EPCMPR_STRIDE * 12, 0x000000FF},
-+ {EPCMPR0 + EPCMPR_STRIDE * 13, 0},
-+ {EPCMPR0 + EPCMPR_STRIDE * 14, 0x000000FF},
-+ {EPCMPR0 + EPCMPR_STRIDE * 15, 0x000000FF},
-+ /* EPCCR (Event Processor Counter Control Registers) */
-+ {EPCCR0 + EPCCR_STRIDE * 0, 0},
-+ {EPCCR0 + EPCCR_STRIDE * 1, 0},
-+ {EPCCR0 + EPCCR_STRIDE * 2, 0x92840000},
-+ {EPCCR0 + EPCCR_STRIDE * 3, 0},
-+ {EPCCR0 + EPCCR_STRIDE * 4, 0x92840000},
-+ {EPCCR0 + EPCCR_STRIDE * 5, 0x92840000},
-+ {EPCCR0 + EPCCR_STRIDE * 6, 0},
-+ {EPCCR0 + EPCCR_STRIDE * 7, 0},
-+ {EPCCR0 + EPCCR_STRIDE * 8, 0x92840000},
-+ {EPCCR0 + EPCCR_STRIDE * 9, 0x92840000},
-+ {EPCCR0 + EPCCR_STRIDE * 10, 0x92840000},
-+ {EPCCR0 + EPCCR_STRIDE * 11, 0x92840000},
-+ {EPCCR0 + EPCCR_STRIDE * 12, 0x92840000},
-+ {EPCCR0 + EPCCR_STRIDE * 13, 0},
-+ {EPCCR0 + EPCCR_STRIDE * 14, 0x92840000},
-+ {EPCCR0 + EPCCR_STRIDE * 15, 0x92840000},
-+ /* EPSMCR (Event Processor SCU Mux Control Registers) */
-+ {EPSMCR0 + EPSMCR_STRIDE * 0, 0},
-+ {EPSMCR0 + EPSMCR_STRIDE * 1, 0},
-+ {EPSMCR0 + EPSMCR_STRIDE * 2, 0x6C700000},
-+ {EPSMCR0 + EPSMCR_STRIDE * 3, 0x2F000000},
-+ {EPSMCR0 + EPSMCR_STRIDE * 4, 0x002F0000},
-+ {EPSMCR0 + EPSMCR_STRIDE * 5, 0x00002E00},
-+ {EPSMCR0 + EPSMCR_STRIDE * 6, 0x7C000000},
-+ {EPSMCR0 + EPSMCR_STRIDE * 7, 0x30000000},
-+ {EPSMCR0 + EPSMCR_STRIDE * 8, 0x64300000},
-+ {EPSMCR0 + EPSMCR_STRIDE * 9, 0x00003000},
-+ {EPSMCR0 + EPSMCR_STRIDE * 10, 0x65000030},
-+ {EPSMCR0 + EPSMCR_STRIDE * 11, 0x31740000},
-+ {EPSMCR0 + EPSMCR_STRIDE * 12, 0x7F000000},
-+ {EPSMCR0 + EPSMCR_STRIDE * 13, 0x00003100},
-+ {EPSMCR0 + EPSMCR_STRIDE * 14, 0x00000031},
-+ {EPSMCR0 + EPSMCR_STRIDE * 15, 0x76000000},
-+ /* EPACR (Event Processor Action Control Registers) */
-+ {EPACR0 + EPACR_STRIDE * 0, 0},
-+ {EPACR0 + EPACR_STRIDE * 1, 0},
-+ {EPACR0 + EPACR_STRIDE * 2, 0},
-+ {EPACR0 + EPACR_STRIDE * 3, 0x00000080},
-+ {EPACR0 + EPACR_STRIDE * 4, 0},
-+ {EPACR0 + EPACR_STRIDE * 5, 0x00000040},
-+ {EPACR0 + EPACR_STRIDE * 6, 0},
-+ {EPACR0 + EPACR_STRIDE * 7, 0},
-+ {EPACR0 + EPACR_STRIDE * 8, 0},
-+ {EPACR0 + EPACR_STRIDE * 9, 0x0000001C},
-+ {EPACR0 + EPACR_STRIDE * 10, 0x00000020},
-+ {EPACR0 + EPACR_STRIDE * 11, 0},
-+ {EPACR0 + EPACR_STRIDE * 12, 0x00000003},
-+ {EPACR0 + EPACR_STRIDE * 13, 0x06000000},
-+ {EPACR0 + EPACR_STRIDE * 14, 0x04000000},
-+ {EPACR0 + EPACR_STRIDE * 15, 0x02000000},
-+ /* EPIMCR (Event Processor Input Mux Control Registers) */
-+ {EPIMCR0 + EPIMCR_STRIDE * 0, 0},
-+ {EPIMCR0 + EPIMCR_STRIDE * 1, 0},
-+ {EPIMCR0 + EPIMCR_STRIDE * 2, 0},
-+ {EPIMCR0 + EPIMCR_STRIDE * 3, 0},
-+ {EPIMCR0 + EPIMCR_STRIDE * 4, 0x44000000},
-+ {EPIMCR0 + EPIMCR_STRIDE * 5, 0x40000000},
-+ {EPIMCR0 + EPIMCR_STRIDE * 6, 0},
-+ {EPIMCR0 + EPIMCR_STRIDE * 7, 0},
-+ {EPIMCR0 + EPIMCR_STRIDE * 8, 0},
-+ {EPIMCR0 + EPIMCR_STRIDE * 9, 0},
-+ {EPIMCR0 + EPIMCR_STRIDE * 10, 0},
-+ {EPIMCR0 + EPIMCR_STRIDE * 11, 0},
-+ {EPIMCR0 + EPIMCR_STRIDE * 12, 0x44000000},
-+ {EPIMCR0 + EPIMCR_STRIDE * 13, 0},
-+ {EPIMCR0 + EPIMCR_STRIDE * 14, 0},
-+ {EPIMCR0 + EPIMCR_STRIDE * 15, 0},
-+ {EPIMCR0 + EPIMCR_STRIDE * 16, 0x6A000000},
-+ {EPIMCR0 + EPIMCR_STRIDE * 17, 0},
-+ {EPIMCR0 + EPIMCR_STRIDE * 18, 0},
-+ {EPIMCR0 + EPIMCR_STRIDE * 19, 0},
-+ {EPIMCR0 + EPIMCR_STRIDE * 20, 0x48000000},
-+ {EPIMCR0 + EPIMCR_STRIDE * 21, 0},
-+ {EPIMCR0 + EPIMCR_STRIDE * 22, 0x6C000000},
-+ {EPIMCR0 + EPIMCR_STRIDE * 23, 0},
-+ {EPIMCR0 + EPIMCR_STRIDE * 24, 0},
-+ {EPIMCR0 + EPIMCR_STRIDE * 25, 0},
-+ {EPIMCR0 + EPIMCR_STRIDE * 26, 0},
-+ {EPIMCR0 + EPIMCR_STRIDE * 27, 0},
-+ {EPIMCR0 + EPIMCR_STRIDE * 28, 0x76000000},
-+ {EPIMCR0 + EPIMCR_STRIDE * 29, 0},
-+ {EPIMCR0 + EPIMCR_STRIDE * 30, 0},
-+ {EPIMCR0 + EPIMCR_STRIDE * 31, 0x76000000},
-+ /* EPXTRIGCR (Event Processor Crosstrigger Control Register) */
-+ {EPXTRIGCR, 0x0000FFDF},
-+ /* end */
-+ {FSM_END_FLAG, 0},
-+};
-+
-+struct fsm_reg_vals npc_default_val[] = {
-+ /* NPC triggered Memory-Mapped Access Registers */
-+ {NCR, 0x80000000},
-+ {MCCR1, 0},
-+ {MCSR1, 0},
-+ {MMAR1LO, 0},
-+ {MMAR1HI, 0},
-+ {MMDR1, 0},
-+ {MCSR2, 0},
-+ {MMAR2LO, 0},
-+ {MMAR2HI, 0},
-+ {MMDR2, 0},
-+ {MCSR3, 0x80000000},
-+ {MMAR3LO, 0x000E2130},
-+ {MMAR3HI, 0x00030000},
-+ {MMDR3, 0x00020000},
-+ /* end */
-+ {FSM_END_FLAG, 0},
-+};
-+
-+/**
-+ * fsl_fsm_setup - Configure EPU's FSM registers
-+ * @base: the base address of registers
-+ * @val: Pointer to address-value pairs for FSM registers
-+ */
-+void fsl_fsm_setup(void __iomem *base, struct fsm_reg_vals *val)
-+{
-+ struct fsm_reg_vals *data = val;
-+
-+ WARN_ON(!base || !data);
-+ while (data->offset != FSM_END_FLAG) {
-+ iowrite32be(data->value, base + data->offset);
-+ data++;
-+ }
-+}
-+
-+void fsl_epu_setup_default(void __iomem *epu_base)
-+{
-+ fsl_fsm_setup(epu_base, epu_default_val);
-+}
-+
-+void fsl_npc_setup_default(void __iomem *npc_base)
-+{
-+ fsl_fsm_setup(npc_base, npc_default_val);
-+}
-+
-+void fsl_epu_clean_default(void __iomem *epu_base)
-+{
-+ u32 offset;
-+
-+ /* follow the exact sequence to clear the registers */
-+ /* Clear EPACRn */
-+ for (offset = EPACR0; offset <= EPACR15; offset += EPACR_STRIDE)
-+ iowrite32be(0, epu_base + offset);
-+
-+ /* Clear EPEVTCRn */
-+ for (offset = EPEVTCR0; offset <= EPEVTCR9; offset += EPEVTCR_STRIDE)
-+ iowrite32be(0, epu_base + offset);
-+
-+ /* Clear EPGCR */
-+ iowrite32be(0, epu_base + EPGCR);
-+
-+ /* Clear EPSMCRn */
-+ for (offset = EPSMCR0; offset <= EPSMCR15; offset += EPSMCR_STRIDE)
-+ iowrite32be(0, epu_base + offset);
-+
-+ /* Clear EPCCRn */
-+ for (offset = EPCCR0; offset <= EPCCR31; offset += EPCCR_STRIDE)
-+ iowrite32be(0, epu_base + offset);
-+
-+ /* Clear EPCMPRn */
-+ for (offset = EPCMPR0; offset <= EPCMPR31; offset += EPCMPR_STRIDE)
-+ iowrite32be(0, epu_base + offset);
-+
-+ /* Clear EPCTRn */
-+ for (offset = EPCTR0; offset <= EPCTR31; offset += EPCTR_STRIDE)
-+ iowrite32be(0, epu_base + offset);
-+
-+ /* Clear EPIMCRn */
-+ for (offset = EPIMCR0; offset <= EPIMCR31; offset += EPIMCR_STRIDE)
-+ iowrite32be(0, epu_base + offset);
-+
-+ /* Clear EPXTRIGCRn */
-+ iowrite32be(0, epu_base + EPXTRIGCR);
-+
-+ /* Clear EPECRn */
-+ for (offset = EPECR0; offset <= EPECR15; offset += EPECR_STRIDE)
-+ iowrite32be(0, epu_base + offset);
-+}
---- /dev/null
-+++ b/drivers/soc/fsl/sleep_fsm.h
-@@ -0,0 +1,130 @@
-+/*
-+ * deep sleep FSM (finite-state machine) configuration
-+ *
-+ * Copyright 2018 NXP
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of the above-listed copyright holders nor the
-+ * names of any contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
-+ * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
-+ * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
-+ * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE
-+ * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
-+ * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
-+ * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
-+ * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
-+ * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
-+ * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
-+ * POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#ifndef _FSL_SLEEP_FSM_H
-+#define _FSL_SLEEP_FSM_H
-+
-+#define FSL_STRIDE_4B 4
-+#define FSL_STRIDE_8B 8
-+
-+/* End flag */
-+#define FSM_END_FLAG 0xFFFFFFFFUL
-+
-+/* Block offsets */
-+#define RCPM_BLOCK_OFFSET 0x00022000
-+#define EPU_BLOCK_OFFSET 0x00000000
-+#define NPC_BLOCK_OFFSET 0x00001000
-+
-+/* EPGCR (Event Processor Global Control Register) */
-+#define EPGCR 0x000
-+
-+/* EPEVTCR0-9 (Event Processor EVT Pin Control Registers) */
-+#define EPEVTCR0 0x050
-+#define EPEVTCR9 0x074
-+#define EPEVTCR_STRIDE FSL_STRIDE_4B
-+
-+/* EPXTRIGCR (Event Processor Crosstrigger Control Register) */
-+#define EPXTRIGCR 0x090
-+
-+/* EPIMCR0-31 (Event Processor Input Mux Control Registers) */
-+#define EPIMCR0 0x100
-+#define EPIMCR31 0x17C
-+#define EPIMCR_STRIDE FSL_STRIDE_4B
-+
-+/* EPSMCR0-15 (Event Processor SCU Mux Control Registers) */
-+#define EPSMCR0 0x200
-+#define EPSMCR15 0x278
-+#define EPSMCR_STRIDE FSL_STRIDE_8B
-+
-+/* EPECR0-15 (Event Processor Event Control Registers) */
-+#define EPECR0 0x300
-+#define EPECR15 0x33C
-+#define EPECR_STRIDE FSL_STRIDE_4B
-+
-+/* EPACR0-15 (Event Processor Action Control Registers) */
-+#define EPACR0 0x400
-+#define EPACR15 0x43C
-+#define EPACR_STRIDE FSL_STRIDE_4B
-+
-+/* EPCCRi0-15 (Event Processor Counter Control Registers) */
-+#define EPCCR0 0x800
-+#define EPCCR15 0x83C
-+#define EPCCR31 0x87C
-+#define EPCCR_STRIDE FSL_STRIDE_4B
-+
-+/* EPCMPR0-15 (Event Processor Counter Compare Registers) */
-+#define EPCMPR0 0x900
-+#define EPCMPR15 0x93C
-+#define EPCMPR31 0x97C
-+#define EPCMPR_STRIDE FSL_STRIDE_4B
-+
-+/* EPCTR0-31 (Event Processor Counter Register) */
-+#define EPCTR0 0xA00
-+#define EPCTR31 0xA7C
-+#define EPCTR_STRIDE FSL_STRIDE_4B
-+
-+/* NPC triggered Memory-Mapped Access Registers */
-+#define NCR 0x000
-+#define MCCR1 0x0CC
-+#define MCSR1 0x0D0
-+#define MMAR1LO 0x0D4
-+#define MMAR1HI 0x0D8
-+#define MMDR1 0x0DC
-+#define MCSR2 0x0E0
-+#define MMAR2LO 0x0E4
-+#define MMAR2HI 0x0E8
-+#define MMDR2 0x0EC
-+#define MCSR3 0x0F0
-+#define MMAR3LO 0x0F4
-+#define MMAR3HI 0x0F8
-+#define MMDR3 0x0FC
-+
-+/* RCPM Core State Action Control Register 0 */
-+#define CSTTACR0 0xB00
-+
-+/* RCPM Core Group 1 Configuration Register 0 */
-+#define CG1CR0 0x31C
-+
-+struct fsm_reg_vals {
-+ u32 offset;
-+ u32 value;
-+};
-+
-+void fsl_fsm_setup(void __iomem *base, struct fsm_reg_vals *val);
-+void fsl_epu_setup_default(void __iomem *epu_base);
-+void fsl_npc_setup_default(void __iomem *npc_base);
-+void fsl_epu_clean_default(void __iomem *epu_base);
-+
-+#endif /* _FSL_SLEEP_FSM_H */
diff --git a/target/linux/layerscape/patches-4.9/804-crypto-support-layerscape.patch b/target/linux/layerscape/patches-4.9/804-crypto-support-layerscape.patch
deleted file mode 100644
index 714fba3a7e..0000000000
--- a/target/linux/layerscape/patches-4.9/804-crypto-support-layerscape.patch
+++ /dev/null
@@ -1,29223 +0,0 @@
-From 2a0aa9bd187f6f5693982a8f79665585af772237 Mon Sep 17 00:00:00 2001
-From: Yangbo Lu <yangbo.lu@nxp.com>
-Date: Thu, 5 Jul 2018 17:29:41 +0800
-Subject: [PATCH 16/32] crypto: support layerscape
-MIME-Version: 1.0
-Content-Type: text/plain; charset=UTF-8
-Content-Transfer-Encoding: 8bit
-
-This is an integrated patch for layerscape sec support.
-
-Signed-off-by: Radu Alexe <radu.alexe@nxp.com>
-Signed-off-by: Fabio Estevam <festevam@gmail.com>
-Signed-off-by: Arnd Bergmann <arnd@arndb.de>
-Signed-off-by: Radu Alexe <radu.alexe@nxp.com>
-Signed-off-by: Tudor Ambarus <tudor-dan.ambarus@nxp.com>
-Signed-off-by: Eric Biggers <ebiggers@google.com>
-Signed-off-by: Giovanni Cabiddu <giovanni.cabiddu@intel.com>
-Signed-off-by: Xulin Sun <xulin.sun@windriver.com>
-Signed-off-by: Ard Biesheuvel <ard.biesheuvel@linaro.org>
-Signed-off-by: Marcus Folkesson <marcus.folkesson@gmail.com>
-Signed-off-by: Tudor Ambarus <tudor-dan.ambarus@nxp.com>
-Signed-off-by: Andrew Lutomirski <luto@kernel.org>
-Signed-off-by: Wei Yongjun <weiyongjun1@huawei.com>
-Signed-off-by: Masahiro Yamada <yamada.masahiro@socionext.com>
-Signed-off-by: Marcelo Cerri <marcelo.cerri@canonical.com>
-Signed-off-by: Arvind Yadav <arvind.yadav.cs@gmail.com>
-Signed-off-by: Herbert Xu <herbert@gondor.apana.org.au>
-Signed-off-by: Laura Abbott <labbott@redhat.com>
-Signed-off-by: Horia Geantă <horia.geanta@nxp.com>
-Singed-off-by: Yangbo Lu <yangbo.lu@nxp.com>
----
- crypto/Kconfig | 30 +
- crypto/Makefile | 4 +
- crypto/acompress.c | 169 +
- crypto/algboss.c | 12 +-
- crypto/crypto_user.c | 19 +
- crypto/scompress.c | 356 +
- crypto/tcrypt.c | 17 +-
- crypto/testmgr.c | 1708 ++---
- crypto/testmgr.h | 1125 ++--
- crypto/tls.c | 607 ++
- drivers/crypto/caam/Kconfig | 77 +-
- drivers/crypto/caam/Makefile | 16 +-
- drivers/crypto/caam/caamalg.c | 2185 ++----
- drivers/crypto/caam/caamalg_desc.c | 1961 ++++++
- drivers/crypto/caam/caamalg_desc.h | 127 +
- drivers/crypto/caam/caamalg_qi.c | 3321 +++++++++
- drivers/crypto/caam/caamalg_qi2.c | 5938 +++++++++++++++++
- drivers/crypto/caam/caamalg_qi2.h | 283 +
- drivers/crypto/caam/caamhash.c | 555 +-
- drivers/crypto/caam/caamhash_desc.c | 108 +
- drivers/crypto/caam/caamhash_desc.h | 49 +
- drivers/crypto/caam/caampkc.c | 471 +-
- drivers/crypto/caam/caampkc.h | 58 +
- drivers/crypto/caam/caamrng.c | 16 +-
- drivers/crypto/caam/compat.h | 1 +
- drivers/crypto/caam/ctrl.c | 358 +-
- drivers/crypto/caam/ctrl.h | 2 +
- drivers/crypto/caam/desc.h | 84 +-
- drivers/crypto/caam/desc_constr.h | 180 +-
- drivers/crypto/caam/dpseci.c | 858 +++
- drivers/crypto/caam/dpseci.h | 395 ++
- drivers/crypto/caam/dpseci_cmd.h | 261 +
- drivers/crypto/caam/error.c | 127 +-
- drivers/crypto/caam/error.h | 10 +-
- drivers/crypto/caam/intern.h | 31 +-
- drivers/crypto/caam/jr.c | 72 +-
- drivers/crypto/caam/jr.h | 2 +
- drivers/crypto/caam/key_gen.c | 32 +-
- drivers/crypto/caam/key_gen.h | 36 +-
- drivers/crypto/caam/pdb.h | 62 +
- drivers/crypto/caam/pkc_desc.c | 36 +
- drivers/crypto/caam/qi.c | 804 +++
- drivers/crypto/caam/qi.h | 204 +
- drivers/crypto/caam/regs.h | 63 +-
- drivers/crypto/caam/sg_sw_qm.h | 126 +
- drivers/crypto/caam/sg_sw_qm2.h | 81 +
- drivers/crypto/caam/sg_sw_sec4.h | 60 +-
- drivers/crypto/talitos.c | 8 +
- drivers/net/wireless/rsi/rsi_91x_usb.c | 2 +-
- drivers/staging/wilc1000/linux_wlan.c | 2 +-
- .../staging/wilc1000/wilc_wfi_cfgoperations.c | 2 +-
- include/crypto/acompress.h | 269 +
- include/crypto/internal/acompress.h | 81 +
- include/crypto/internal/scompress.h | 136 +
- include/linux/crypto.h | 3 +
- include/uapi/linux/cryptouser.h | 5 +
- scripts/spelling.txt | 3 +
- sound/soc/amd/acp-pcm-dma.c | 2 +-
- 58 files changed, 19620 insertions(+), 3990 deletions(-)
- create mode 100644 crypto/acompress.c
- create mode 100644 crypto/scompress.c
- create mode 100644 crypto/tls.c
- create mode 100644 drivers/crypto/caam/caamalg_desc.c
- create mode 100644 drivers/crypto/caam/caamalg_desc.h
- create mode 100644 drivers/crypto/caam/caamalg_qi.c
- create mode 100644 drivers/crypto/caam/caamalg_qi2.c
- create mode 100644 drivers/crypto/caam/caamalg_qi2.h
- create mode 100644 drivers/crypto/caam/caamhash_desc.c
- create mode 100644 drivers/crypto/caam/caamhash_desc.h
- create mode 100644 drivers/crypto/caam/dpseci.c
- create mode 100644 drivers/crypto/caam/dpseci.h
- create mode 100644 drivers/crypto/caam/dpseci_cmd.h
- create mode 100644 drivers/crypto/caam/qi.c
- create mode 100644 drivers/crypto/caam/qi.h
- create mode 100644 drivers/crypto/caam/sg_sw_qm.h
- create mode 100644 drivers/crypto/caam/sg_sw_qm2.h
- create mode 100644 include/crypto/acompress.h
- create mode 100644 include/crypto/internal/acompress.h
- create mode 100644 include/crypto/internal/scompress.h
-
---- a/crypto/Kconfig
-+++ b/crypto/Kconfig
-@@ -102,6 +102,15 @@ config CRYPTO_KPP
- select CRYPTO_ALGAPI
- select CRYPTO_KPP2
-
-+config CRYPTO_ACOMP2
-+ tristate
-+ select CRYPTO_ALGAPI2
-+
-+config CRYPTO_ACOMP
-+ tristate
-+ select CRYPTO_ALGAPI
-+ select CRYPTO_ACOMP2
-+
- config CRYPTO_RSA
- tristate "RSA algorithm"
- select CRYPTO_AKCIPHER
-@@ -138,6 +147,7 @@ config CRYPTO_MANAGER2
- select CRYPTO_BLKCIPHER2 if !CRYPTO_MANAGER_DISABLE_TESTS
- select CRYPTO_AKCIPHER2 if !CRYPTO_MANAGER_DISABLE_TESTS
- select CRYPTO_KPP2 if !CRYPTO_MANAGER_DISABLE_TESTS
-+ select CRYPTO_ACOMP2 if !CRYPTO_MANAGER_DISABLE_TESTS
-
- config CRYPTO_USER
- tristate "Userspace cryptographic algorithm configuration"
-@@ -295,6 +305,26 @@ config CRYPTO_ECHAINIV
- a sequence number xored with a salt. This is the default
- algorithm for CBC.
-
-+config CRYPTO_TLS
-+ tristate "TLS support"
-+ select CRYPTO_AEAD
-+ select CRYPTO_BLKCIPHER
-+ select CRYPTO_MANAGER
-+ select CRYPTO_HASH
-+ select CRYPTO_NULL
-+ select CRYPTO_AUTHENC
-+ help
-+ Support for TLS 1.0 record encryption and decryption
-+
-+ This module adds support for encryption/decryption of TLS 1.0 frames
-+ using blockcipher algorithms. The name of the resulting algorithm is
-+ "tls10(hmac(<digest>),cbc(<cipher>))". By default, the generic base
-+ algorithms are used (e.g. aes-generic, sha1-generic), but hardware
-+ accelerated versions will be used automatically if available.
-+
-+ User-space applications (OpenSSL, GnuTLS) can offload TLS 1.0
-+ operations through AF_ALG or cryptodev interfaces
-+
- comment "Block modes"
-
- config CRYPTO_CBC
---- a/crypto/Makefile
-+++ b/crypto/Makefile
-@@ -51,6 +51,9 @@ rsa_generic-y += rsa_helper.o
- rsa_generic-y += rsa-pkcs1pad.o
- obj-$(CONFIG_CRYPTO_RSA) += rsa_generic.o
-
-+obj-$(CONFIG_CRYPTO_ACOMP2) += acompress.o
-+obj-$(CONFIG_CRYPTO_ACOMP2) += scompress.o
-+
- cryptomgr-y := algboss.o testmgr.o
-
- obj-$(CONFIG_CRYPTO_MANAGER2) += cryptomgr.o
-@@ -115,6 +118,7 @@ obj-$(CONFIG_CRYPTO_CRC32C) += crc32c_ge
- obj-$(CONFIG_CRYPTO_CRC32) += crc32_generic.o
- obj-$(CONFIG_CRYPTO_CRCT10DIF) += crct10dif_common.o crct10dif_generic.o
- obj-$(CONFIG_CRYPTO_AUTHENC) += authenc.o authencesn.o
-+obj-$(CONFIG_CRYPTO_TLS) += tls.o
- obj-$(CONFIG_CRYPTO_LZO) += lzo.o
- obj-$(CONFIG_CRYPTO_LZ4) += lz4.o
- obj-$(CONFIG_CRYPTO_LZ4HC) += lz4hc.o
---- /dev/null
-+++ b/crypto/acompress.c
-@@ -0,0 +1,169 @@
-+/*
-+ * Asynchronous Compression operations
-+ *
-+ * Copyright (c) 2016, Intel Corporation
-+ * Authors: Weigang Li <weigang.li@intel.com>
-+ * Giovanni Cabiddu <giovanni.cabiddu@intel.com>
-+ *
-+ * This program is free software; you can redistribute it and/or modify it
-+ * under the terms of the GNU General Public License as published by the Free
-+ * Software Foundation; either version 2 of the License, or (at your option)
-+ * any later version.
-+ *
-+ */
-+#include <linux/errno.h>
-+#include <linux/kernel.h>
-+#include <linux/module.h>
-+#include <linux/seq_file.h>
-+#include <linux/slab.h>
-+#include <linux/string.h>
-+#include <linux/crypto.h>
-+#include <crypto/algapi.h>
-+#include <linux/cryptouser.h>
-+#include <net/netlink.h>
-+#include <crypto/internal/acompress.h>
-+#include <crypto/internal/scompress.h>
-+#include "internal.h"
-+
-+static const struct crypto_type crypto_acomp_type;
-+
-+#ifdef CONFIG_NET
-+static int crypto_acomp_report(struct sk_buff *skb, struct crypto_alg *alg)
-+{
-+ struct crypto_report_acomp racomp;
-+
-+ strncpy(racomp.type, "acomp", sizeof(racomp.type));
-+
-+ if (nla_put(skb, CRYPTOCFGA_REPORT_ACOMP,
-+ sizeof(struct crypto_report_acomp), &racomp))
-+ goto nla_put_failure;
-+ return 0;
-+
-+nla_put_failure:
-+ return -EMSGSIZE;
-+}
-+#else
-+static int crypto_acomp_report(struct sk_buff *skb, struct crypto_alg *alg)
-+{
-+ return -ENOSYS;
-+}
-+#endif
-+
-+static void crypto_acomp_show(struct seq_file *m, struct crypto_alg *alg)
-+ __attribute__ ((unused));
-+
-+static void crypto_acomp_show(struct seq_file *m, struct crypto_alg *alg)
-+{
-+ seq_puts(m, "type : acomp\n");
-+}
-+
-+static void crypto_acomp_exit_tfm(struct crypto_tfm *tfm)
-+{
-+ struct crypto_acomp *acomp = __crypto_acomp_tfm(tfm);
-+ struct acomp_alg *alg = crypto_acomp_alg(acomp);
-+
-+ alg->exit(acomp);
-+}
-+
-+static int crypto_acomp_init_tfm(struct crypto_tfm *tfm)
-+{
-+ struct crypto_acomp *acomp = __crypto_acomp_tfm(tfm);
-+ struct acomp_alg *alg = crypto_acomp_alg(acomp);
-+
-+ if (tfm->__crt_alg->cra_type != &crypto_acomp_type)
-+ return crypto_init_scomp_ops_async(tfm);
-+
-+ acomp->compress = alg->compress;
-+ acomp->decompress = alg->decompress;
-+ acomp->dst_free = alg->dst_free;
-+ acomp->reqsize = alg->reqsize;
-+
-+ if (alg->exit)
-+ acomp->base.exit = crypto_acomp_exit_tfm;
-+
-+ if (alg->init)
-+ return alg->init(acomp);
-+
-+ return 0;
-+}
-+
-+static unsigned int crypto_acomp_extsize(struct crypto_alg *alg)
-+{
-+ int extsize = crypto_alg_extsize(alg);
-+
-+ if (alg->cra_type != &crypto_acomp_type)
-+ extsize += sizeof(struct crypto_scomp *);
-+
-+ return extsize;
-+}
-+
-+static const struct crypto_type crypto_acomp_type = {
-+ .extsize = crypto_acomp_extsize,
-+ .init_tfm = crypto_acomp_init_tfm,
-+#ifdef CONFIG_PROC_FS
-+ .show = crypto_acomp_show,
-+#endif
-+ .report = crypto_acomp_report,
-+ .maskclear = ~CRYPTO_ALG_TYPE_MASK,
-+ .maskset = CRYPTO_ALG_TYPE_ACOMPRESS_MASK,
-+ .type = CRYPTO_ALG_TYPE_ACOMPRESS,
-+ .tfmsize = offsetof(struct crypto_acomp, base),
-+};
-+
-+struct crypto_acomp *crypto_alloc_acomp(const char *alg_name, u32 type,
-+ u32 mask)
-+{
-+ return crypto_alloc_tfm(alg_name, &crypto_acomp_type, type, mask);
-+}
-+EXPORT_SYMBOL_GPL(crypto_alloc_acomp);
-+
-+struct acomp_req *acomp_request_alloc(struct crypto_acomp *acomp)
-+{
-+ struct crypto_tfm *tfm = crypto_acomp_tfm(acomp);
-+ struct acomp_req *req;
-+
-+ req = __acomp_request_alloc(acomp);
-+ if (req && (tfm->__crt_alg->cra_type != &crypto_acomp_type))
-+ return crypto_acomp_scomp_alloc_ctx(req);
-+
-+ return req;
-+}
-+EXPORT_SYMBOL_GPL(acomp_request_alloc);
-+
-+void acomp_request_free(struct acomp_req *req)
-+{
-+ struct crypto_acomp *acomp = crypto_acomp_reqtfm(req);
-+ struct crypto_tfm *tfm = crypto_acomp_tfm(acomp);
-+
-+ if (tfm->__crt_alg->cra_type != &crypto_acomp_type)
-+ crypto_acomp_scomp_free_ctx(req);
-+
-+ if (req->flags & CRYPTO_ACOMP_ALLOC_OUTPUT) {
-+ acomp->dst_free(req->dst);
-+ req->dst = NULL;
-+ }
-+
-+ __acomp_request_free(req);
-+}
-+EXPORT_SYMBOL_GPL(acomp_request_free);
-+
-+int crypto_register_acomp(struct acomp_alg *alg)
-+{
-+ struct crypto_alg *base = &alg->base;
-+
-+ base->cra_type = &crypto_acomp_type;
-+ base->cra_flags &= ~CRYPTO_ALG_TYPE_MASK;
-+ base->cra_flags |= CRYPTO_ALG_TYPE_ACOMPRESS;
-+
-+ return crypto_register_alg(base);
-+}
-+EXPORT_SYMBOL_GPL(crypto_register_acomp);
-+
-+int crypto_unregister_acomp(struct acomp_alg *alg)
-+{
-+ return crypto_unregister_alg(&alg->base);
-+}
-+EXPORT_SYMBOL_GPL(crypto_unregister_acomp);
-+
-+MODULE_LICENSE("GPL");
-+MODULE_DESCRIPTION("Asynchronous compression type");
---- a/crypto/algboss.c
-+++ b/crypto/algboss.c
-@@ -247,17 +247,9 @@ static int cryptomgr_schedule_test(struc
- memcpy(param->alg, alg->cra_name, sizeof(param->alg));
- type = alg->cra_flags;
-
-- /* This piece of crap needs to disappear into per-type test hooks. */
--#ifdef CONFIG_CRYPTO_MANAGER_DISABLE_TESTS
-- type |= CRYPTO_ALG_TESTED;
--#else
-- if (!((type ^ CRYPTO_ALG_TYPE_BLKCIPHER) &
-- CRYPTO_ALG_TYPE_BLKCIPHER_MASK) && !(type & CRYPTO_ALG_GENIV) &&
-- ((alg->cra_flags & CRYPTO_ALG_TYPE_MASK) ==
-- CRYPTO_ALG_TYPE_BLKCIPHER ? alg->cra_blkcipher.ivsize :
-- alg->cra_ablkcipher.ivsize))
-+ /* Do not test internal algorithms. */
-+ if (type & CRYPTO_ALG_INTERNAL)
- type |= CRYPTO_ALG_TESTED;
--#endif
-
- param->type = type;
-
---- a/crypto/crypto_user.c
-+++ b/crypto/crypto_user.c
-@@ -112,6 +112,21 @@ nla_put_failure:
- return -EMSGSIZE;
- }
-
-+static int crypto_report_acomp(struct sk_buff *skb, struct crypto_alg *alg)
-+{
-+ struct crypto_report_acomp racomp;
-+
-+ strncpy(racomp.type, "acomp", sizeof(racomp.type));
-+
-+ if (nla_put(skb, CRYPTOCFGA_REPORT_ACOMP,
-+ sizeof(struct crypto_report_acomp), &racomp))
-+ goto nla_put_failure;
-+ return 0;
-+
-+nla_put_failure:
-+ return -EMSGSIZE;
-+}
-+
- static int crypto_report_akcipher(struct sk_buff *skb, struct crypto_alg *alg)
- {
- struct crypto_report_akcipher rakcipher;
-@@ -186,7 +201,11 @@ static int crypto_report_one(struct cryp
- goto nla_put_failure;
-
- break;
-+ case CRYPTO_ALG_TYPE_ACOMPRESS:
-+ if (crypto_report_acomp(skb, alg))
-+ goto nla_put_failure;
-
-+ break;
- case CRYPTO_ALG_TYPE_AKCIPHER:
- if (crypto_report_akcipher(skb, alg))
- goto nla_put_failure;
---- /dev/null
-+++ b/crypto/scompress.c
-@@ -0,0 +1,356 @@
-+/*
-+ * Synchronous Compression operations
-+ *
-+ * Copyright 2015 LG Electronics Inc.
-+ * Copyright (c) 2016, Intel Corporation
-+ * Author: Giovanni Cabiddu <giovanni.cabiddu@intel.com>
-+ *
-+ * This program is free software; you can redistribute it and/or modify it
-+ * under the terms of the GNU General Public License as published by the Free
-+ * Software Foundation; either version 2 of the License, or (at your option)
-+ * any later version.
-+ *
-+ */
-+#include <linux/errno.h>
-+#include <linux/kernel.h>
-+#include <linux/module.h>
-+#include <linux/seq_file.h>
-+#include <linux/slab.h>
-+#include <linux/string.h>
-+#include <linux/crypto.h>
-+#include <linux/vmalloc.h>
-+#include <crypto/algapi.h>
-+#include <linux/cryptouser.h>
-+#include <net/netlink.h>
-+#include <linux/scatterlist.h>
-+#include <crypto/scatterwalk.h>
-+#include <crypto/internal/acompress.h>
-+#include <crypto/internal/scompress.h>
-+#include "internal.h"
-+
-+static const struct crypto_type crypto_scomp_type;
-+static void * __percpu *scomp_src_scratches;
-+static void * __percpu *scomp_dst_scratches;
-+static int scomp_scratch_users;
-+static DEFINE_MUTEX(scomp_lock);
-+
-+#ifdef CONFIG_NET
-+static int crypto_scomp_report(struct sk_buff *skb, struct crypto_alg *alg)
-+{
-+ struct crypto_report_comp rscomp;
-+
-+ strncpy(rscomp.type, "scomp", sizeof(rscomp.type));
-+
-+ if (nla_put(skb, CRYPTOCFGA_REPORT_COMPRESS,
-+ sizeof(struct crypto_report_comp), &rscomp))
-+ goto nla_put_failure;
-+ return 0;
-+
-+nla_put_failure:
-+ return -EMSGSIZE;
-+}
-+#else
-+static int crypto_scomp_report(struct sk_buff *skb, struct crypto_alg *alg)
-+{
-+ return -ENOSYS;
-+}
-+#endif
-+
-+static void crypto_scomp_show(struct seq_file *m, struct crypto_alg *alg)
-+ __attribute__ ((unused));
-+
-+static void crypto_scomp_show(struct seq_file *m, struct crypto_alg *alg)
-+{
-+ seq_puts(m, "type : scomp\n");
-+}
-+
-+static int crypto_scomp_init_tfm(struct crypto_tfm *tfm)
-+{
-+ return 0;
-+}
-+
-+static void crypto_scomp_free_scratches(void * __percpu *scratches)
-+{
-+ int i;
-+
-+ if (!scratches)
-+ return;
-+
-+ for_each_possible_cpu(i)
-+ vfree(*per_cpu_ptr(scratches, i));
-+
-+ free_percpu(scratches);
-+}
-+
-+static void * __percpu *crypto_scomp_alloc_scratches(void)
-+{
-+ void * __percpu *scratches;
-+ int i;
-+
-+ scratches = alloc_percpu(void *);
-+ if (!scratches)
-+ return NULL;
-+
-+ for_each_possible_cpu(i) {
-+ void *scratch;
-+
-+ scratch = vmalloc_node(SCOMP_SCRATCH_SIZE, cpu_to_node(i));
-+ if (!scratch)
-+ goto error;
-+ *per_cpu_ptr(scratches, i) = scratch;
-+ }
-+
-+ return scratches;
-+
-+error:
-+ crypto_scomp_free_scratches(scratches);
-+ return NULL;
-+}
-+
-+static void crypto_scomp_free_all_scratches(void)
-+{
-+ if (!--scomp_scratch_users) {
-+ crypto_scomp_free_scratches(scomp_src_scratches);
-+ crypto_scomp_free_scratches(scomp_dst_scratches);
-+ scomp_src_scratches = NULL;
-+ scomp_dst_scratches = NULL;
-+ }
-+}
-+
-+static int crypto_scomp_alloc_all_scratches(void)
-+{
-+ if (!scomp_scratch_users++) {
-+ scomp_src_scratches = crypto_scomp_alloc_scratches();
-+ if (!scomp_src_scratches)
-+ return -ENOMEM;
-+ scomp_dst_scratches = crypto_scomp_alloc_scratches();
-+ if (!scomp_dst_scratches)
-+ return -ENOMEM;
-+ }
-+ return 0;
-+}
-+
-+static void crypto_scomp_sg_free(struct scatterlist *sgl)
-+{
-+ int i, n;
-+ struct page *page;
-+
-+ if (!sgl)
-+ return;
-+
-+ n = sg_nents(sgl);
-+ for_each_sg(sgl, sgl, n, i) {
-+ page = sg_page(sgl);
-+ if (page)
-+ __free_page(page);
-+ }
-+
-+ kfree(sgl);
-+}
-+
-+static struct scatterlist *crypto_scomp_sg_alloc(size_t size, gfp_t gfp)
-+{
-+ struct scatterlist *sgl;
-+ struct page *page;
-+ int i, n;
-+
-+ n = ((size - 1) >> PAGE_SHIFT) + 1;
-+
-+ sgl = kmalloc_array(n, sizeof(struct scatterlist), gfp);
-+ if (!sgl)
-+ return NULL;
-+
-+ sg_init_table(sgl, n);
-+
-+ for (i = 0; i < n; i++) {
-+ page = alloc_page(gfp);
-+ if (!page)
-+ goto err;
-+ sg_set_page(sgl + i, page, PAGE_SIZE, 0);
-+ }
-+
-+ return sgl;
-+
-+err:
-+ sg_mark_end(sgl + i);
-+ crypto_scomp_sg_free(sgl);
-+ return NULL;
-+}
-+
-+static int scomp_acomp_comp_decomp(struct acomp_req *req, int dir)
-+{
-+ struct crypto_acomp *tfm = crypto_acomp_reqtfm(req);
-+ void **tfm_ctx = acomp_tfm_ctx(tfm);
-+ struct crypto_scomp *scomp = *tfm_ctx;
-+ void **ctx = acomp_request_ctx(req);
-+ const int cpu = get_cpu();
-+ u8 *scratch_src = *per_cpu_ptr(scomp_src_scratches, cpu);
-+ u8 *scratch_dst = *per_cpu_ptr(scomp_dst_scratches, cpu);
-+ int ret;
-+
-+ if (!req->src || !req->slen || req->slen > SCOMP_SCRATCH_SIZE) {
-+ ret = -EINVAL;
-+ goto out;
-+ }
-+
-+ if (req->dst && !req->dlen) {
-+ ret = -EINVAL;
-+ goto out;
-+ }
-+
-+ if (!req->dlen || req->dlen > SCOMP_SCRATCH_SIZE)
-+ req->dlen = SCOMP_SCRATCH_SIZE;
-+
-+ scatterwalk_map_and_copy(scratch_src, req->src, 0, req->slen, 0);
-+ if (dir)
-+ ret = crypto_scomp_compress(scomp, scratch_src, req->slen,
-+ scratch_dst, &req->dlen, *ctx);
-+ else
-+ ret = crypto_scomp_decompress(scomp, scratch_src, req->slen,
-+ scratch_dst, &req->dlen, *ctx);
-+ if (!ret) {
-+ if (!req->dst) {
-+ req->dst = crypto_scomp_sg_alloc(req->dlen,
-+ req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP ?
-+ GFP_KERNEL : GFP_ATOMIC);
-+ if (!req->dst)
-+ goto out;
-+ }
-+ scatterwalk_map_and_copy(scratch_dst, req->dst, 0, req->dlen,
-+ 1);
-+ }
-+out:
-+ put_cpu();
-+ return ret;
-+}
-+
-+static int scomp_acomp_compress(struct acomp_req *req)
-+{
-+ return scomp_acomp_comp_decomp(req, 1);
-+}
-+
-+static int scomp_acomp_decompress(struct acomp_req *req)
-+{
-+ return scomp_acomp_comp_decomp(req, 0);
-+}
-+
-+static void crypto_exit_scomp_ops_async(struct crypto_tfm *tfm)
-+{
-+ struct crypto_scomp **ctx = crypto_tfm_ctx(tfm);
-+
-+ crypto_free_scomp(*ctx);
-+}
-+
-+int crypto_init_scomp_ops_async(struct crypto_tfm *tfm)
-+{
-+ struct crypto_alg *calg = tfm->__crt_alg;
-+ struct crypto_acomp *crt = __crypto_acomp_tfm(tfm);
-+ struct crypto_scomp **ctx = crypto_tfm_ctx(tfm);
-+ struct crypto_scomp *scomp;
-+
-+ if (!crypto_mod_get(calg))
-+ return -EAGAIN;
-+
-+ scomp = crypto_create_tfm(calg, &crypto_scomp_type);
-+ if (IS_ERR(scomp)) {
-+ crypto_mod_put(calg);
-+ return PTR_ERR(scomp);
-+ }
-+
-+ *ctx = scomp;
-+ tfm->exit = crypto_exit_scomp_ops_async;
-+
-+ crt->compress = scomp_acomp_compress;
-+ crt->decompress = scomp_acomp_decompress;
-+ crt->dst_free = crypto_scomp_sg_free;
-+ crt->reqsize = sizeof(void *);
-+
-+ return 0;
-+}
-+
-+struct acomp_req *crypto_acomp_scomp_alloc_ctx(struct acomp_req *req)
-+{
-+ struct crypto_acomp *acomp = crypto_acomp_reqtfm(req);
-+ struct crypto_tfm *tfm = crypto_acomp_tfm(acomp);
-+ struct crypto_scomp **tfm_ctx = crypto_tfm_ctx(tfm);
-+ struct crypto_scomp *scomp = *tfm_ctx;
-+ void *ctx;
-+
-+ ctx = crypto_scomp_alloc_ctx(scomp);
-+ if (IS_ERR(ctx)) {
-+ kfree(req);
-+ return NULL;
-+ }
-+
-+ *req->__ctx = ctx;
-+
-+ return req;
-+}
-+
-+void crypto_acomp_scomp_free_ctx(struct acomp_req *req)
-+{
-+ struct crypto_acomp *acomp = crypto_acomp_reqtfm(req);
-+ struct crypto_tfm *tfm = crypto_acomp_tfm(acomp);
-+ struct crypto_scomp **tfm_ctx = crypto_tfm_ctx(tfm);
-+ struct crypto_scomp *scomp = *tfm_ctx;
-+ void *ctx = *req->__ctx;
-+
-+ if (ctx)
-+ crypto_scomp_free_ctx(scomp, ctx);
-+}
-+
-+static const struct crypto_type crypto_scomp_type = {
-+ .extsize = crypto_alg_extsize,
-+ .init_tfm = crypto_scomp_init_tfm,
-+#ifdef CONFIG_PROC_FS
-+ .show = crypto_scomp_show,
-+#endif
-+ .report = crypto_scomp_report,
-+ .maskclear = ~CRYPTO_ALG_TYPE_MASK,
-+ .maskset = CRYPTO_ALG_TYPE_MASK,
-+ .type = CRYPTO_ALG_TYPE_SCOMPRESS,
-+ .tfmsize = offsetof(struct crypto_scomp, base),
-+};
-+
-+int crypto_register_scomp(struct scomp_alg *alg)
-+{
-+ struct crypto_alg *base = &alg->base;
-+ int ret = -ENOMEM;
-+
-+ mutex_lock(&scomp_lock);
-+ if (crypto_scomp_alloc_all_scratches())
-+ goto error;
-+
-+ base->cra_type = &crypto_scomp_type;
-+ base->cra_flags &= ~CRYPTO_ALG_TYPE_MASK;
-+ base->cra_flags |= CRYPTO_ALG_TYPE_SCOMPRESS;
-+
-+ ret = crypto_register_alg(base);
-+ if (ret)
-+ goto error;
-+
-+ mutex_unlock(&scomp_lock);
-+ return ret;
-+
-+error:
-+ crypto_scomp_free_all_scratches();
-+ mutex_unlock(&scomp_lock);
-+ return ret;
-+}
-+EXPORT_SYMBOL_GPL(crypto_register_scomp);
-+
-+int crypto_unregister_scomp(struct scomp_alg *alg)
-+{
-+ int ret;
-+
-+ mutex_lock(&scomp_lock);
-+ ret = crypto_unregister_alg(&alg->base);
-+ crypto_scomp_free_all_scratches();
-+ mutex_unlock(&scomp_lock);
-+
-+ return ret;
-+}
-+EXPORT_SYMBOL_GPL(crypto_unregister_scomp);
-+
-+MODULE_LICENSE("GPL");
-+MODULE_DESCRIPTION("Synchronous compression type");
---- a/crypto/tcrypt.c
-+++ b/crypto/tcrypt.c
-@@ -74,7 +74,7 @@ static char *check[] = {
- "khazad", "wp512", "wp384", "wp256", "tnepres", "xeta", "fcrypt",
- "camellia", "seed", "salsa20", "rmd128", "rmd160", "rmd256", "rmd320",
- "lzo", "cts", "zlib", "sha3-224", "sha3-256", "sha3-384", "sha3-512",
-- NULL
-+ "rsa", NULL
- };
-
- struct tcrypt_result {
-@@ -1336,6 +1336,10 @@ static int do_test(const char *alg, u32
- ret += tcrypt_test("hmac(sha3-512)");
- break;
-
-+ case 115:
-+ ret += tcrypt_test("rsa");
-+ break;
-+
- case 150:
- ret += tcrypt_test("ansi_cprng");
- break;
-@@ -1397,6 +1401,9 @@ static int do_test(const char *alg, u32
- case 190:
- ret += tcrypt_test("authenc(hmac(sha512),cbc(des3_ede))");
- break;
-+ case 191:
-+ ret += tcrypt_test("tls10(hmac(sha1),cbc(aes))");
-+ break;
- case 200:
- test_cipher_speed("ecb(aes)", ENCRYPT, sec, NULL, 0,
- speed_template_16_24_32);
-@@ -1411,9 +1418,9 @@ static int do_test(const char *alg, u32
- test_cipher_speed("lrw(aes)", DECRYPT, sec, NULL, 0,
- speed_template_32_40_48);
- test_cipher_speed("xts(aes)", ENCRYPT, sec, NULL, 0,
-- speed_template_32_48_64);
-+ speed_template_32_64);
- test_cipher_speed("xts(aes)", DECRYPT, sec, NULL, 0,
-- speed_template_32_48_64);
-+ speed_template_32_64);
- test_cipher_speed("cts(cbc(aes))", ENCRYPT, sec, NULL, 0,
- speed_template_16_24_32);
- test_cipher_speed("cts(cbc(aes))", DECRYPT, sec, NULL, 0,
-@@ -1844,9 +1851,9 @@ static int do_test(const char *alg, u32
- test_acipher_speed("lrw(aes)", DECRYPT, sec, NULL, 0,
- speed_template_32_40_48);
- test_acipher_speed("xts(aes)", ENCRYPT, sec, NULL, 0,
-- speed_template_32_48_64);
-+ speed_template_32_64);
- test_acipher_speed("xts(aes)", DECRYPT, sec, NULL, 0,
-- speed_template_32_48_64);
-+ speed_template_32_64);
- test_acipher_speed("cts(cbc(aes))", ENCRYPT, sec, NULL, 0,
- speed_template_16_24_32);
- test_acipher_speed("cts(cbc(aes))", DECRYPT, sec, NULL, 0,
---- a/crypto/testmgr.c
-+++ b/crypto/testmgr.c
-@@ -33,6 +33,7 @@
- #include <crypto/drbg.h>
- #include <crypto/akcipher.h>
- #include <crypto/kpp.h>
-+#include <crypto/acompress.h>
-
- #include "internal.h"
-
-@@ -62,7 +63,7 @@ int alg_test(const char *driver, const c
- */
- #define IDX1 32
- #define IDX2 32400
--#define IDX3 1
-+#define IDX3 1511
- #define IDX4 8193
- #define IDX5 22222
- #define IDX6 17101
-@@ -82,47 +83,54 @@ struct tcrypt_result {
-
- struct aead_test_suite {
- struct {
-- struct aead_testvec *vecs;
-+ const struct aead_testvec *vecs;
- unsigned int count;
- } enc, dec;
- };
-
- struct cipher_test_suite {
- struct {
-- struct cipher_testvec *vecs;
-+ const struct cipher_testvec *vecs;
- unsigned int count;
- } enc, dec;
- };
-
- struct comp_test_suite {
- struct {
-- struct comp_testvec *vecs;
-+ const struct comp_testvec *vecs;
- unsigned int count;
- } comp, decomp;
- };
-
- struct hash_test_suite {
-- struct hash_testvec *vecs;
-+ const struct hash_testvec *vecs;
- unsigned int count;
- };
-
- struct cprng_test_suite {
-- struct cprng_testvec *vecs;
-+ const struct cprng_testvec *vecs;
- unsigned int count;
- };
-
- struct drbg_test_suite {
-- struct drbg_testvec *vecs;
-+ const struct drbg_testvec *vecs;
- unsigned int count;
- };
-
-+struct tls_test_suite {
-+ struct {
-+ struct tls_testvec *vecs;
-+ unsigned int count;
-+ } enc, dec;
-+};
-+
- struct akcipher_test_suite {
-- struct akcipher_testvec *vecs;
-+ const struct akcipher_testvec *vecs;
- unsigned int count;
- };
-
- struct kpp_test_suite {
-- struct kpp_testvec *vecs;
-+ const struct kpp_testvec *vecs;
- unsigned int count;
- };
-
-@@ -139,12 +147,14 @@ struct alg_test_desc {
- struct hash_test_suite hash;
- struct cprng_test_suite cprng;
- struct drbg_test_suite drbg;
-+ struct tls_test_suite tls;
- struct akcipher_test_suite akcipher;
- struct kpp_test_suite kpp;
- } suite;
- };
-
--static unsigned int IDX[8] = { IDX1, IDX2, IDX3, IDX4, IDX5, IDX6, IDX7, IDX8 };
-+static const unsigned int IDX[8] = {
-+ IDX1, IDX2, IDX3, IDX4, IDX5, IDX6, IDX7, IDX8 };
-
- static void hexdump(unsigned char *buf, unsigned int len)
- {
-@@ -202,7 +212,7 @@ static int wait_async_op(struct tcrypt_r
- }
-
- static int ahash_partial_update(struct ahash_request **preq,
-- struct crypto_ahash *tfm, struct hash_testvec *template,
-+ struct crypto_ahash *tfm, const struct hash_testvec *template,
- void *hash_buff, int k, int temp, struct scatterlist *sg,
- const char *algo, char *result, struct tcrypt_result *tresult)
- {
-@@ -259,11 +269,12 @@ out_nostate:
- return ret;
- }
-
--static int __test_hash(struct crypto_ahash *tfm, struct hash_testvec *template,
-- unsigned int tcount, bool use_digest,
-- const int align_offset)
-+static int __test_hash(struct crypto_ahash *tfm,
-+ const struct hash_testvec *template, unsigned int tcount,
-+ bool use_digest, const int align_offset)
- {
- const char *algo = crypto_tfm_alg_driver_name(crypto_ahash_tfm(tfm));
-+ size_t digest_size = crypto_ahash_digestsize(tfm);
- unsigned int i, j, k, temp;
- struct scatterlist sg[8];
- char *result;
-@@ -274,7 +285,7 @@ static int __test_hash(struct crypto_aha
- char *xbuf[XBUFSIZE];
- int ret = -ENOMEM;
-
-- result = kmalloc(MAX_DIGEST_SIZE, GFP_KERNEL);
-+ result = kmalloc(digest_size, GFP_KERNEL);
- if (!result)
- return ret;
- key = kmalloc(MAX_KEYLEN, GFP_KERNEL);
-@@ -304,7 +315,7 @@ static int __test_hash(struct crypto_aha
- goto out;
-
- j++;
-- memset(result, 0, MAX_DIGEST_SIZE);
-+ memset(result, 0, digest_size);
-
- hash_buff = xbuf[0];
- hash_buff += align_offset;
-@@ -379,7 +390,7 @@ static int __test_hash(struct crypto_aha
- continue;
-
- j++;
-- memset(result, 0, MAX_DIGEST_SIZE);
-+ memset(result, 0, digest_size);
-
- temp = 0;
- sg_init_table(sg, template[i].np);
-@@ -457,7 +468,7 @@ static int __test_hash(struct crypto_aha
- continue;
-
- j++;
-- memset(result, 0, MAX_DIGEST_SIZE);
-+ memset(result, 0, digest_size);
-
- ret = -EINVAL;
- hash_buff = xbuf[0];
-@@ -536,7 +547,8 @@ out_nobuf:
- return ret;
- }
-
--static int test_hash(struct crypto_ahash *tfm, struct hash_testvec *template,
-+static int test_hash(struct crypto_ahash *tfm,
-+ const struct hash_testvec *template,
- unsigned int tcount, bool use_digest)
- {
- unsigned int alignmask;
-@@ -564,7 +576,7 @@ static int test_hash(struct crypto_ahash
- }
-
- static int __test_aead(struct crypto_aead *tfm, int enc,
-- struct aead_testvec *template, unsigned int tcount,
-+ const struct aead_testvec *template, unsigned int tcount,
- const bool diff_dst, const int align_offset)
- {
- const char *algo = crypto_tfm_alg_driver_name(crypto_aead_tfm(tfm));
-@@ -955,7 +967,7 @@ out_noxbuf:
- }
-
- static int test_aead(struct crypto_aead *tfm, int enc,
-- struct aead_testvec *template, unsigned int tcount)
-+ const struct aead_testvec *template, unsigned int tcount)
- {
- unsigned int alignmask;
- int ret;
-@@ -987,8 +999,236 @@ static int test_aead(struct crypto_aead
- return 0;
- }
-
-+static int __test_tls(struct crypto_aead *tfm, int enc,
-+ struct tls_testvec *template, unsigned int tcount,
-+ const bool diff_dst)
-+{
-+ const char *algo = crypto_tfm_alg_driver_name(crypto_aead_tfm(tfm));
-+ unsigned int i, k, authsize;
-+ char *q;
-+ struct aead_request *req;
-+ struct scatterlist *sg;
-+ struct scatterlist *sgout;
-+ const char *e, *d;
-+ struct tcrypt_result result;
-+ void *input;
-+ void *output;
-+ void *assoc;
-+ char *iv;
-+ char *key;
-+ char *xbuf[XBUFSIZE];
-+ char *xoutbuf[XBUFSIZE];
-+ char *axbuf[XBUFSIZE];
-+ int ret = -ENOMEM;
-+
-+ if (testmgr_alloc_buf(xbuf))
-+ goto out_noxbuf;
-+
-+ if (diff_dst && testmgr_alloc_buf(xoutbuf))
-+ goto out_nooutbuf;
-+
-+ if (testmgr_alloc_buf(axbuf))
-+ goto out_noaxbuf;
-+
-+ iv = kzalloc(MAX_IVLEN, GFP_KERNEL);
-+ if (!iv)
-+ goto out_noiv;
-+
-+ key = kzalloc(MAX_KEYLEN, GFP_KERNEL);
-+ if (!key)
-+ goto out_nokey;
-+
-+ sg = kmalloc(sizeof(*sg) * 8 * (diff_dst ? 2 : 1), GFP_KERNEL);
-+ if (!sg)
-+ goto out_nosg;
-+
-+ sgout = sg + 8;
-+
-+ d = diff_dst ? "-ddst" : "";
-+ e = enc ? "encryption" : "decryption";
-+
-+ init_completion(&result.completion);
-+
-+ req = aead_request_alloc(tfm, GFP_KERNEL);
-+ if (!req) {
-+ pr_err("alg: tls%s: Failed to allocate request for %s\n",
-+ d, algo);
-+ goto out;
-+ }
-+
-+ aead_request_set_callback(req, CRYPTO_TFM_REQ_MAY_BACKLOG,
-+ tcrypt_complete, &result);
-+
-+ for (i = 0; i < tcount; i++) {
-+ input = xbuf[0];
-+ assoc = axbuf[0];
-+
-+ ret = -EINVAL;
-+ if (WARN_ON(template[i].ilen > PAGE_SIZE ||
-+ template[i].alen > PAGE_SIZE))
-+ goto out;
-+
-+ memcpy(assoc, template[i].assoc, template[i].alen);
-+ memcpy(input, template[i].input, template[i].ilen);
-+
-+ if (template[i].iv)
-+ memcpy(iv, template[i].iv, MAX_IVLEN);
-+ else
-+ memset(iv, 0, MAX_IVLEN);
-+
-+ crypto_aead_clear_flags(tfm, ~0);
-+
-+ if (template[i].klen > MAX_KEYLEN) {
-+ pr_err("alg: aead%s: setkey failed on test %d for %s: key size %d > %d\n",
-+ d, i, algo, template[i].klen, MAX_KEYLEN);
-+ ret = -EINVAL;
-+ goto out;
-+ }
-+ memcpy(key, template[i].key, template[i].klen);
-+
-+ ret = crypto_aead_setkey(tfm, key, template[i].klen);
-+ if (!ret == template[i].fail) {
-+ pr_err("alg: tls%s: setkey failed on test %d for %s: flags=%x\n",
-+ d, i, algo, crypto_aead_get_flags(tfm));
-+ goto out;
-+ } else if (ret)
-+ continue;
-+
-+ authsize = 20;
-+ ret = crypto_aead_setauthsize(tfm, authsize);
-+ if (ret) {
-+ pr_err("alg: aead%s: Failed to set authsize to %u on test %d for %s\n",
-+ d, authsize, i, algo);
-+ goto out;
-+ }
-+
-+ k = !!template[i].alen;
-+ sg_init_table(sg, k + 1);
-+ sg_set_buf(&sg[0], assoc, template[i].alen);
-+ sg_set_buf(&sg[k], input, (enc ? template[i].rlen :
-+ template[i].ilen));
-+ output = input;
-+
-+ if (diff_dst) {
-+ sg_init_table(sgout, k + 1);
-+ sg_set_buf(&sgout[0], assoc, template[i].alen);
-+
-+ output = xoutbuf[0];
-+ sg_set_buf(&sgout[k], output,
-+ (enc ? template[i].rlen : template[i].ilen));
-+ }
-+
-+ aead_request_set_crypt(req, sg, (diff_dst) ? sgout : sg,
-+ template[i].ilen, iv);
-+
-+ aead_request_set_ad(req, template[i].alen);
-+
-+ ret = enc ? crypto_aead_encrypt(req) : crypto_aead_decrypt(req);
-+
-+ switch (ret) {
-+ case 0:
-+ if (template[i].novrfy) {
-+ /* verification was supposed to fail */
-+ pr_err("alg: tls%s: %s failed on test %d for %s: ret was 0, expected -EBADMSG\n",
-+ d, e, i, algo);
-+ /* so really, we got a bad message */
-+ ret = -EBADMSG;
-+ goto out;
-+ }
-+ break;
-+ case -EINPROGRESS:
-+ case -EBUSY:
-+ wait_for_completion(&result.completion);
-+ reinit_completion(&result.completion);
-+ ret = result.err;
-+ if (!ret)
-+ break;
-+ case -EBADMSG:
-+ /* verification failure was expected */
-+ if (template[i].novrfy)
-+ continue;
-+ /* fall through */
-+ default:
-+ pr_err("alg: tls%s: %s failed on test %d for %s: ret=%d\n",
-+ d, e, i, algo, -ret);
-+ goto out;
-+ }
-+
-+ q = output;
-+ if (memcmp(q, template[i].result, template[i].rlen)) {
-+ pr_err("alg: tls%s: Test %d failed on %s for %s\n",
-+ d, i, e, algo);
-+ hexdump(q, template[i].rlen);
-+ pr_err("should be:\n");
-+ hexdump(template[i].result, template[i].rlen);
-+ ret = -EINVAL;
-+ goto out;
-+ }
-+ }
-+
-+out:
-+ aead_request_free(req);
-+
-+ kfree(sg);
-+out_nosg:
-+ kfree(key);
-+out_nokey:
-+ kfree(iv);
-+out_noiv:
-+ testmgr_free_buf(axbuf);
-+out_noaxbuf:
-+ if (diff_dst)
-+ testmgr_free_buf(xoutbuf);
-+out_nooutbuf:
-+ testmgr_free_buf(xbuf);
-+out_noxbuf:
-+ return ret;
-+}
-+
-+static int test_tls(struct crypto_aead *tfm, int enc,
-+ struct tls_testvec *template, unsigned int tcount)
-+{
-+ int ret;
-+ /* test 'dst == src' case */
-+ ret = __test_tls(tfm, enc, template, tcount, false);
-+ if (ret)
-+ return ret;
-+ /* test 'dst != src' case */
-+ return __test_tls(tfm, enc, template, tcount, true);
-+}
-+
-+static int alg_test_tls(const struct alg_test_desc *desc, const char *driver,
-+ u32 type, u32 mask)
-+{
-+ struct crypto_aead *tfm;
-+ int err = 0;
-+
-+ tfm = crypto_alloc_aead(driver, type, mask);
-+ if (IS_ERR(tfm)) {
-+ pr_err("alg: aead: Failed to load transform for %s: %ld\n",
-+ driver, PTR_ERR(tfm));
-+ return PTR_ERR(tfm);
-+ }
-+
-+ if (desc->suite.tls.enc.vecs) {
-+ err = test_tls(tfm, ENCRYPT, desc->suite.tls.enc.vecs,
-+ desc->suite.tls.enc.count);
-+ if (err)
-+ goto out;
-+ }
-+
-+ if (!err && desc->suite.tls.dec.vecs)
-+ err = test_tls(tfm, DECRYPT, desc->suite.tls.dec.vecs,
-+ desc->suite.tls.dec.count);
-+
-+out:
-+ crypto_free_aead(tfm);
-+ return err;
-+}
-+
- static int test_cipher(struct crypto_cipher *tfm, int enc,
-- struct cipher_testvec *template, unsigned int tcount)
-+ const struct cipher_testvec *template,
-+ unsigned int tcount)
- {
- const char *algo = crypto_tfm_alg_driver_name(crypto_cipher_tfm(tfm));
- unsigned int i, j, k;
-@@ -1066,7 +1306,8 @@ out_nobuf:
- }
-
- static int __test_skcipher(struct crypto_skcipher *tfm, int enc,
-- struct cipher_testvec *template, unsigned int tcount,
-+ const struct cipher_testvec *template,
-+ unsigned int tcount,
- const bool diff_dst, const int align_offset)
- {
- const char *algo =
-@@ -1079,12 +1320,16 @@ static int __test_skcipher(struct crypto
- const char *e, *d;
- struct tcrypt_result result;
- void *data;
-- char iv[MAX_IVLEN];
-+ char *iv;
- char *xbuf[XBUFSIZE];
- char *xoutbuf[XBUFSIZE];
- int ret = -ENOMEM;
- unsigned int ivsize = crypto_skcipher_ivsize(tfm);
-
-+ iv = kmalloc(MAX_IVLEN, GFP_KERNEL);
-+ if (!iv)
-+ return ret;
-+
- if (testmgr_alloc_buf(xbuf))
- goto out_nobuf;
-
-@@ -1325,12 +1570,14 @@ out:
- testmgr_free_buf(xoutbuf);
- out_nooutbuf:
- testmgr_free_buf(xbuf);
-+ kfree(iv);
- out_nobuf:
- return ret;
- }
-
- static int test_skcipher(struct crypto_skcipher *tfm, int enc,
-- struct cipher_testvec *template, unsigned int tcount)
-+ const struct cipher_testvec *template,
-+ unsigned int tcount)
- {
- unsigned int alignmask;
- int ret;
-@@ -1362,8 +1609,10 @@ static int test_skcipher(struct crypto_s
- return 0;
- }
-
--static int test_comp(struct crypto_comp *tfm, struct comp_testvec *ctemplate,
-- struct comp_testvec *dtemplate, int ctcount, int dtcount)
-+static int test_comp(struct crypto_comp *tfm,
-+ const struct comp_testvec *ctemplate,
-+ const struct comp_testvec *dtemplate,
-+ int ctcount, int dtcount)
- {
- const char *algo = crypto_tfm_alg_driver_name(crypto_comp_tfm(tfm));
- unsigned int i;
-@@ -1442,7 +1691,154 @@ out:
- return ret;
- }
-
--static int test_cprng(struct crypto_rng *tfm, struct cprng_testvec *template,
-+static int test_acomp(struct crypto_acomp *tfm,
-+ const struct comp_testvec *ctemplate,
-+ const struct comp_testvec *dtemplate,
-+ int ctcount, int dtcount)
-+{
-+ const char *algo = crypto_tfm_alg_driver_name(crypto_acomp_tfm(tfm));
-+ unsigned int i;
-+ char *output;
-+ int ret;
-+ struct scatterlist src, dst;
-+ struct acomp_req *req;
-+ struct tcrypt_result result;
-+
-+ output = kmalloc(COMP_BUF_SIZE, GFP_KERNEL);
-+ if (!output)
-+ return -ENOMEM;
-+
-+ for (i = 0; i < ctcount; i++) {
-+ unsigned int dlen = COMP_BUF_SIZE;
-+ int ilen = ctemplate[i].inlen;
-+ void *input_vec;
-+
-+ input_vec = kmemdup(ctemplate[i].input, ilen, GFP_KERNEL);
-+ if (!input_vec) {
-+ ret = -ENOMEM;
-+ goto out;
-+ }
-+
-+ memset(output, 0, dlen);
-+ init_completion(&result.completion);
-+ sg_init_one(&src, input_vec, ilen);
-+ sg_init_one(&dst, output, dlen);
-+
-+ req = acomp_request_alloc(tfm);
-+ if (!req) {
-+ pr_err("alg: acomp: request alloc failed for %s\n",
-+ algo);
-+ kfree(input_vec);
-+ ret = -ENOMEM;
-+ goto out;
-+ }
-+
-+ acomp_request_set_params(req, &src, &dst, ilen, dlen);
-+ acomp_request_set_callback(req, CRYPTO_TFM_REQ_MAY_BACKLOG,
-+ tcrypt_complete, &result);
-+
-+ ret = wait_async_op(&result, crypto_acomp_compress(req));
-+ if (ret) {
-+ pr_err("alg: acomp: compression failed on test %d for %s: ret=%d\n",
-+ i + 1, algo, -ret);
-+ kfree(input_vec);
-+ acomp_request_free(req);
-+ goto out;
-+ }
-+
-+ if (req->dlen != ctemplate[i].outlen) {
-+ pr_err("alg: acomp: Compression test %d failed for %s: output len = %d\n",
-+ i + 1, algo, req->dlen);
-+ ret = -EINVAL;
-+ kfree(input_vec);
-+ acomp_request_free(req);
-+ goto out;
-+ }
-+
-+ if (memcmp(output, ctemplate[i].output, req->dlen)) {
-+ pr_err("alg: acomp: Compression test %d failed for %s\n",
-+ i + 1, algo);
-+ hexdump(output, req->dlen);
-+ ret = -EINVAL;
-+ kfree(input_vec);
-+ acomp_request_free(req);
-+ goto out;
-+ }
-+
-+ kfree(input_vec);
-+ acomp_request_free(req);
-+ }
-+
-+ for (i = 0; i < dtcount; i++) {
-+ unsigned int dlen = COMP_BUF_SIZE;
-+ int ilen = dtemplate[i].inlen;
-+ void *input_vec;
-+
-+ input_vec = kmemdup(dtemplate[i].input, ilen, GFP_KERNEL);
-+ if (!input_vec) {
-+ ret = -ENOMEM;
-+ goto out;
-+ }
-+
-+ memset(output, 0, dlen);
-+ init_completion(&result.completion);
-+ sg_init_one(&src, input_vec, ilen);
-+ sg_init_one(&dst, output, dlen);
-+
-+ req = acomp_request_alloc(tfm);
-+ if (!req) {
-+ pr_err("alg: acomp: request alloc failed for %s\n",
-+ algo);
-+ kfree(input_vec);
-+ ret = -ENOMEM;
-+ goto out;
-+ }
-+
-+ acomp_request_set_params(req, &src, &dst, ilen, dlen);
-+ acomp_request_set_callback(req, CRYPTO_TFM_REQ_MAY_BACKLOG,
-+ tcrypt_complete, &result);
-+
-+ ret = wait_async_op(&result, crypto_acomp_decompress(req));
-+ if (ret) {
-+ pr_err("alg: acomp: decompression failed on test %d for %s: ret=%d\n",
-+ i + 1, algo, -ret);
-+ kfree(input_vec);
-+ acomp_request_free(req);
-+ goto out;
-+ }
-+
-+ if (req->dlen != dtemplate[i].outlen) {
-+ pr_err("alg: acomp: Decompression test %d failed for %s: output len = %d\n",
-+ i + 1, algo, req->dlen);
-+ ret = -EINVAL;
-+ kfree(input_vec);
-+ acomp_request_free(req);
-+ goto out;
-+ }
-+
-+ if (memcmp(output, dtemplate[i].output, req->dlen)) {
-+ pr_err("alg: acomp: Decompression test %d failed for %s\n",
-+ i + 1, algo);
-+ hexdump(output, req->dlen);
-+ ret = -EINVAL;
-+ kfree(input_vec);
-+ acomp_request_free(req);
-+ goto out;
-+ }
-+
-+ kfree(input_vec);
-+ acomp_request_free(req);
-+ }
-+
-+ ret = 0;
-+
-+out:
-+ kfree(output);
-+ return ret;
-+}
-+
-+static int test_cprng(struct crypto_rng *tfm,
-+ const struct cprng_testvec *template,
- unsigned int tcount)
- {
- const char *algo = crypto_tfm_alg_driver_name(crypto_rng_tfm(tfm));
-@@ -1509,7 +1905,7 @@ static int alg_test_aead(const struct al
- struct crypto_aead *tfm;
- int err = 0;
-
-- tfm = crypto_alloc_aead(driver, type | CRYPTO_ALG_INTERNAL, mask);
-+ tfm = crypto_alloc_aead(driver, type, mask);
- if (IS_ERR(tfm)) {
- printk(KERN_ERR "alg: aead: Failed to load transform for %s: "
- "%ld\n", driver, PTR_ERR(tfm));
-@@ -1538,7 +1934,7 @@ static int alg_test_cipher(const struct
- struct crypto_cipher *tfm;
- int err = 0;
-
-- tfm = crypto_alloc_cipher(driver, type | CRYPTO_ALG_INTERNAL, mask);
-+ tfm = crypto_alloc_cipher(driver, type, mask);
- if (IS_ERR(tfm)) {
- printk(KERN_ERR "alg: cipher: Failed to load transform for "
- "%s: %ld\n", driver, PTR_ERR(tfm));
-@@ -1567,7 +1963,7 @@ static int alg_test_skcipher(const struc
- struct crypto_skcipher *tfm;
- int err = 0;
-
-- tfm = crypto_alloc_skcipher(driver, type | CRYPTO_ALG_INTERNAL, mask);
-+ tfm = crypto_alloc_skcipher(driver, type, mask);
- if (IS_ERR(tfm)) {
- printk(KERN_ERR "alg: skcipher: Failed to load transform for "
- "%s: %ld\n", driver, PTR_ERR(tfm));
-@@ -1593,22 +1989,38 @@ out:
- static int alg_test_comp(const struct alg_test_desc *desc, const char *driver,
- u32 type, u32 mask)
- {
-- struct crypto_comp *tfm;
-+ struct crypto_comp *comp;
-+ struct crypto_acomp *acomp;
- int err;
-+ u32 algo_type = type & CRYPTO_ALG_TYPE_ACOMPRESS_MASK;
-
-- tfm = crypto_alloc_comp(driver, type, mask);
-- if (IS_ERR(tfm)) {
-- printk(KERN_ERR "alg: comp: Failed to load transform for %s: "
-- "%ld\n", driver, PTR_ERR(tfm));
-- return PTR_ERR(tfm);
-- }
-+ if (algo_type == CRYPTO_ALG_TYPE_ACOMPRESS) {
-+ acomp = crypto_alloc_acomp(driver, type, mask);
-+ if (IS_ERR(acomp)) {
-+ pr_err("alg: acomp: Failed to load transform for %s: %ld\n",
-+ driver, PTR_ERR(acomp));
-+ return PTR_ERR(acomp);
-+ }
-+ err = test_acomp(acomp, desc->suite.comp.comp.vecs,
-+ desc->suite.comp.decomp.vecs,
-+ desc->suite.comp.comp.count,
-+ desc->suite.comp.decomp.count);
-+ crypto_free_acomp(acomp);
-+ } else {
-+ comp = crypto_alloc_comp(driver, type, mask);
-+ if (IS_ERR(comp)) {
-+ pr_err("alg: comp: Failed to load transform for %s: %ld\n",
-+ driver, PTR_ERR(comp));
-+ return PTR_ERR(comp);
-+ }
-
-- err = test_comp(tfm, desc->suite.comp.comp.vecs,
-- desc->suite.comp.decomp.vecs,
-- desc->suite.comp.comp.count,
-- desc->suite.comp.decomp.count);
-+ err = test_comp(comp, desc->suite.comp.comp.vecs,
-+ desc->suite.comp.decomp.vecs,
-+ desc->suite.comp.comp.count,
-+ desc->suite.comp.decomp.count);
-
-- crypto_free_comp(tfm);
-+ crypto_free_comp(comp);
-+ }
- return err;
- }
-
-@@ -1618,7 +2030,7 @@ static int alg_test_hash(const struct al
- struct crypto_ahash *tfm;
- int err;
-
-- tfm = crypto_alloc_ahash(driver, type | CRYPTO_ALG_INTERNAL, mask);
-+ tfm = crypto_alloc_ahash(driver, type, mask);
- if (IS_ERR(tfm)) {
- printk(KERN_ERR "alg: hash: Failed to load transform for %s: "
- "%ld\n", driver, PTR_ERR(tfm));
-@@ -1646,7 +2058,7 @@ static int alg_test_crc32c(const struct
- if (err)
- goto out;
-
-- tfm = crypto_alloc_shash(driver, type | CRYPTO_ALG_INTERNAL, mask);
-+ tfm = crypto_alloc_shash(driver, type, mask);
- if (IS_ERR(tfm)) {
- printk(KERN_ERR "alg: crc32c: Failed to load transform for %s: "
- "%ld\n", driver, PTR_ERR(tfm));
-@@ -1688,7 +2100,7 @@ static int alg_test_cprng(const struct a
- struct crypto_rng *rng;
- int err;
-
-- rng = crypto_alloc_rng(driver, type | CRYPTO_ALG_INTERNAL, mask);
-+ rng = crypto_alloc_rng(driver, type, mask);
- if (IS_ERR(rng)) {
- printk(KERN_ERR "alg: cprng: Failed to load transform for %s: "
- "%ld\n", driver, PTR_ERR(rng));
-@@ -1703,7 +2115,7 @@ static int alg_test_cprng(const struct a
- }
-
-
--static int drbg_cavs_test(struct drbg_testvec *test, int pr,
-+static int drbg_cavs_test(const struct drbg_testvec *test, int pr,
- const char *driver, u32 type, u32 mask)
- {
- int ret = -EAGAIN;
-@@ -1715,7 +2127,7 @@ static int drbg_cavs_test(struct drbg_te
- if (!buf)
- return -ENOMEM;
-
-- drng = crypto_alloc_rng(driver, type | CRYPTO_ALG_INTERNAL, mask);
-+ drng = crypto_alloc_rng(driver, type, mask);
- if (IS_ERR(drng)) {
- printk(KERN_ERR "alg: drbg: could not allocate DRNG handle for "
- "%s\n", driver);
-@@ -1777,7 +2189,7 @@ static int alg_test_drbg(const struct al
- int err = 0;
- int pr = 0;
- int i = 0;
-- struct drbg_testvec *template = desc->suite.drbg.vecs;
-+ const struct drbg_testvec *template = desc->suite.drbg.vecs;
- unsigned int tcount = desc->suite.drbg.count;
-
- if (0 == memcmp(driver, "drbg_pr_", 8))
-@@ -1796,7 +2208,7 @@ static int alg_test_drbg(const struct al
-
- }
-
--static int do_test_kpp(struct crypto_kpp *tfm, struct kpp_testvec *vec,
-+static int do_test_kpp(struct crypto_kpp *tfm, const struct kpp_testvec *vec,
- const char *alg)
- {
- struct kpp_request *req;
-@@ -1888,7 +2300,7 @@ free_req:
- }
-
- static int test_kpp(struct crypto_kpp *tfm, const char *alg,
-- struct kpp_testvec *vecs, unsigned int tcount)
-+ const struct kpp_testvec *vecs, unsigned int tcount)
- {
- int ret, i;
-
-@@ -1909,7 +2321,7 @@ static int alg_test_kpp(const struct alg
- struct crypto_kpp *tfm;
- int err = 0;
-
-- tfm = crypto_alloc_kpp(driver, type | CRYPTO_ALG_INTERNAL, mask);
-+ tfm = crypto_alloc_kpp(driver, type, mask);
- if (IS_ERR(tfm)) {
- pr_err("alg: kpp: Failed to load tfm for %s: %ld\n",
- driver, PTR_ERR(tfm));
-@@ -1924,7 +2336,7 @@ static int alg_test_kpp(const struct alg
- }
-
- static int test_akcipher_one(struct crypto_akcipher *tfm,
-- struct akcipher_testvec *vecs)
-+ const struct akcipher_testvec *vecs)
- {
- char *xbuf[XBUFSIZE];
- struct akcipher_request *req;
-@@ -2044,7 +2456,8 @@ free_xbuf:
- }
-
- static int test_akcipher(struct crypto_akcipher *tfm, const char *alg,
-- struct akcipher_testvec *vecs, unsigned int tcount)
-+ const struct akcipher_testvec *vecs,
-+ unsigned int tcount)
- {
- const char *algo =
- crypto_tfm_alg_driver_name(crypto_akcipher_tfm(tfm));
-@@ -2068,7 +2481,7 @@ static int alg_test_akcipher(const struc
- struct crypto_akcipher *tfm;
- int err = 0;
-
-- tfm = crypto_alloc_akcipher(driver, type | CRYPTO_ALG_INTERNAL, mask);
-+ tfm = crypto_alloc_akcipher(driver, type, mask);
- if (IS_ERR(tfm)) {
- pr_err("alg: akcipher: Failed to load tfm for %s: %ld\n",
- driver, PTR_ERR(tfm));
-@@ -2088,112 +2501,23 @@ static int alg_test_null(const struct al
- return 0;
- }
-
-+#define __VECS(tv) { .vecs = tv, .count = ARRAY_SIZE(tv) }
-+
- /* Please keep this list sorted by algorithm name. */
- static const struct alg_test_desc alg_test_descs[] = {
- {
-- .alg = "__cbc-cast5-avx",
-- .test = alg_test_null,
-- }, {
-- .alg = "__cbc-cast6-avx",
-- .test = alg_test_null,
-- }, {
-- .alg = "__cbc-serpent-avx",
-- .test = alg_test_null,
-- }, {
-- .alg = "__cbc-serpent-avx2",
-- .test = alg_test_null,
-- }, {
-- .alg = "__cbc-serpent-sse2",
-- .test = alg_test_null,
-- }, {
-- .alg = "__cbc-twofish-avx",
-- .test = alg_test_null,
-- }, {
-- .alg = "__driver-cbc-aes-aesni",
-- .test = alg_test_null,
-- .fips_allowed = 1,
-- }, {
-- .alg = "__driver-cbc-camellia-aesni",
-- .test = alg_test_null,
-- }, {
-- .alg = "__driver-cbc-camellia-aesni-avx2",
-- .test = alg_test_null,
-- }, {
-- .alg = "__driver-cbc-cast5-avx",
-- .test = alg_test_null,
-- }, {
-- .alg = "__driver-cbc-cast6-avx",
-- .test = alg_test_null,
-- }, {
-- .alg = "__driver-cbc-serpent-avx",
-- .test = alg_test_null,
-- }, {
-- .alg = "__driver-cbc-serpent-avx2",
-- .test = alg_test_null,
-- }, {
-- .alg = "__driver-cbc-serpent-sse2",
-- .test = alg_test_null,
-- }, {
-- .alg = "__driver-cbc-twofish-avx",
-- .test = alg_test_null,
-- }, {
-- .alg = "__driver-ecb-aes-aesni",
-- .test = alg_test_null,
-- .fips_allowed = 1,
-- }, {
-- .alg = "__driver-ecb-camellia-aesni",
-- .test = alg_test_null,
-- }, {
-- .alg = "__driver-ecb-camellia-aesni-avx2",
-- .test = alg_test_null,
-- }, {
-- .alg = "__driver-ecb-cast5-avx",
-- .test = alg_test_null,
-- }, {
-- .alg = "__driver-ecb-cast6-avx",
-- .test = alg_test_null,
-- }, {
-- .alg = "__driver-ecb-serpent-avx",
-- .test = alg_test_null,
-- }, {
-- .alg = "__driver-ecb-serpent-avx2",
-- .test = alg_test_null,
-- }, {
-- .alg = "__driver-ecb-serpent-sse2",
-- .test = alg_test_null,
-- }, {
-- .alg = "__driver-ecb-twofish-avx",
-- .test = alg_test_null,
-- }, {
-- .alg = "__driver-gcm-aes-aesni",
-- .test = alg_test_null,
-- .fips_allowed = 1,
-- }, {
-- .alg = "__ghash-pclmulqdqni",
-- .test = alg_test_null,
-- .fips_allowed = 1,
-- }, {
- .alg = "ansi_cprng",
- .test = alg_test_cprng,
- .suite = {
-- .cprng = {
-- .vecs = ansi_cprng_aes_tv_template,
-- .count = ANSI_CPRNG_AES_TEST_VECTORS
-- }
-+ .cprng = __VECS(ansi_cprng_aes_tv_template)
- }
- }, {
- .alg = "authenc(hmac(md5),ecb(cipher_null))",
- .test = alg_test_aead,
- .suite = {
- .aead = {
-- .enc = {
-- .vecs = hmac_md5_ecb_cipher_null_enc_tv_template,
-- .count = HMAC_MD5_ECB_CIPHER_NULL_ENC_TEST_VECTORS
-- },
-- .dec = {
-- .vecs = hmac_md5_ecb_cipher_null_dec_tv_template,
-- .count = HMAC_MD5_ECB_CIPHER_NULL_DEC_TEST_VECTORS
-- }
-+ .enc = __VECS(hmac_md5_ecb_cipher_null_enc_tv_template),
-+ .dec = __VECS(hmac_md5_ecb_cipher_null_dec_tv_template)
- }
- }
- }, {
-@@ -2201,12 +2525,7 @@ static const struct alg_test_desc alg_te
- .test = alg_test_aead,
- .suite = {
- .aead = {
-- .enc = {
-- .vecs =
-- hmac_sha1_aes_cbc_enc_tv_temp,
-- .count =
-- HMAC_SHA1_AES_CBC_ENC_TEST_VEC
-- }
-+ .enc = __VECS(hmac_sha1_aes_cbc_enc_tv_temp)
- }
- }
- }, {
-@@ -2214,12 +2533,7 @@ static const struct alg_test_desc alg_te
- .test = alg_test_aead,
- .suite = {
- .aead = {
-- .enc = {
-- .vecs =
-- hmac_sha1_des_cbc_enc_tv_temp,
-- .count =
-- HMAC_SHA1_DES_CBC_ENC_TEST_VEC
-- }
-+ .enc = __VECS(hmac_sha1_des_cbc_enc_tv_temp)
- }
- }
- }, {
-@@ -2228,12 +2542,7 @@ static const struct alg_test_desc alg_te
- .fips_allowed = 1,
- .suite = {
- .aead = {
-- .enc = {
-- .vecs =
-- hmac_sha1_des3_ede_cbc_enc_tv_temp,
-- .count =
-- HMAC_SHA1_DES3_EDE_CBC_ENC_TEST_VEC
-- }
-+ .enc = __VECS(hmac_sha1_des3_ede_cbc_enc_tv_temp)
- }
- }
- }, {
-@@ -2245,18 +2554,8 @@ static const struct alg_test_desc alg_te
- .test = alg_test_aead,
- .suite = {
- .aead = {
-- .enc = {
-- .vecs =
-- hmac_sha1_ecb_cipher_null_enc_tv_temp,
-- .count =
-- HMAC_SHA1_ECB_CIPHER_NULL_ENC_TEST_VEC
-- },
-- .dec = {
-- .vecs =
-- hmac_sha1_ecb_cipher_null_dec_tv_temp,
-- .count =
-- HMAC_SHA1_ECB_CIPHER_NULL_DEC_TEST_VEC
-- }
-+ .enc = __VECS(hmac_sha1_ecb_cipher_null_enc_tv_temp),
-+ .dec = __VECS(hmac_sha1_ecb_cipher_null_dec_tv_temp)
- }
- }
- }, {
-@@ -2268,12 +2567,7 @@ static const struct alg_test_desc alg_te
- .test = alg_test_aead,
- .suite = {
- .aead = {
-- .enc = {
-- .vecs =
-- hmac_sha224_des_cbc_enc_tv_temp,
-- .count =
-- HMAC_SHA224_DES_CBC_ENC_TEST_VEC
-- }
-+ .enc = __VECS(hmac_sha224_des_cbc_enc_tv_temp)
- }
- }
- }, {
-@@ -2282,12 +2576,7 @@ static const struct alg_test_desc alg_te
- .fips_allowed = 1,
- .suite = {
- .aead = {
-- .enc = {
-- .vecs =
-- hmac_sha224_des3_ede_cbc_enc_tv_temp,
-- .count =
-- HMAC_SHA224_DES3_EDE_CBC_ENC_TEST_VEC
-- }
-+ .enc = __VECS(hmac_sha224_des3_ede_cbc_enc_tv_temp)
- }
- }
- }, {
-@@ -2296,12 +2585,7 @@ static const struct alg_test_desc alg_te
- .fips_allowed = 1,
- .suite = {
- .aead = {
-- .enc = {
-- .vecs =
-- hmac_sha256_aes_cbc_enc_tv_temp,
-- .count =
-- HMAC_SHA256_AES_CBC_ENC_TEST_VEC
-- }
-+ .enc = __VECS(hmac_sha256_aes_cbc_enc_tv_temp)
- }
- }
- }, {
-@@ -2309,12 +2593,7 @@ static const struct alg_test_desc alg_te
- .test = alg_test_aead,
- .suite = {
- .aead = {
-- .enc = {
-- .vecs =
-- hmac_sha256_des_cbc_enc_tv_temp,
-- .count =
-- HMAC_SHA256_DES_CBC_ENC_TEST_VEC
-- }
-+ .enc = __VECS(hmac_sha256_des_cbc_enc_tv_temp)
- }
- }
- }, {
-@@ -2323,12 +2602,7 @@ static const struct alg_test_desc alg_te
- .fips_allowed = 1,
- .suite = {
- .aead = {
-- .enc = {
-- .vecs =
-- hmac_sha256_des3_ede_cbc_enc_tv_temp,
-- .count =
-- HMAC_SHA256_DES3_EDE_CBC_ENC_TEST_VEC
-- }
-+ .enc = __VECS(hmac_sha256_des3_ede_cbc_enc_tv_temp)
- }
- }
- }, {
-@@ -2344,12 +2618,7 @@ static const struct alg_test_desc alg_te
- .test = alg_test_aead,
- .suite = {
- .aead = {
-- .enc = {
-- .vecs =
-- hmac_sha384_des_cbc_enc_tv_temp,
-- .count =
-- HMAC_SHA384_DES_CBC_ENC_TEST_VEC
-- }
-+ .enc = __VECS(hmac_sha384_des_cbc_enc_tv_temp)
- }
- }
- }, {
-@@ -2358,12 +2627,7 @@ static const struct alg_test_desc alg_te
- .fips_allowed = 1,
- .suite = {
- .aead = {
-- .enc = {
-- .vecs =
-- hmac_sha384_des3_ede_cbc_enc_tv_temp,
-- .count =
-- HMAC_SHA384_DES3_EDE_CBC_ENC_TEST_VEC
-- }
-+ .enc = __VECS(hmac_sha384_des3_ede_cbc_enc_tv_temp)
- }
- }
- }, {
-@@ -2380,12 +2644,7 @@ static const struct alg_test_desc alg_te
- .test = alg_test_aead,
- .suite = {
- .aead = {
-- .enc = {
-- .vecs =
-- hmac_sha512_aes_cbc_enc_tv_temp,
-- .count =
-- HMAC_SHA512_AES_CBC_ENC_TEST_VEC
-- }
-+ .enc = __VECS(hmac_sha512_aes_cbc_enc_tv_temp)
- }
- }
- }, {
-@@ -2393,12 +2652,7 @@ static const struct alg_test_desc alg_te
- .test = alg_test_aead,
- .suite = {
- .aead = {
-- .enc = {
-- .vecs =
-- hmac_sha512_des_cbc_enc_tv_temp,
-- .count =
-- HMAC_SHA512_DES_CBC_ENC_TEST_VEC
-- }
-+ .enc = __VECS(hmac_sha512_des_cbc_enc_tv_temp)
- }
- }
- }, {
-@@ -2407,12 +2661,7 @@ static const struct alg_test_desc alg_te
- .fips_allowed = 1,
- .suite = {
- .aead = {
-- .enc = {
-- .vecs =
-- hmac_sha512_des3_ede_cbc_enc_tv_temp,
-- .count =
-- HMAC_SHA512_DES3_EDE_CBC_ENC_TEST_VEC
-- }
-+ .enc = __VECS(hmac_sha512_des3_ede_cbc_enc_tv_temp)
- }
- }
- }, {
-@@ -2429,14 +2678,8 @@ static const struct alg_test_desc alg_te
- .fips_allowed = 1,
- .suite = {
- .cipher = {
-- .enc = {
-- .vecs = aes_cbc_enc_tv_template,
-- .count = AES_CBC_ENC_TEST_VECTORS
-- },
-- .dec = {
-- .vecs = aes_cbc_dec_tv_template,
-- .count = AES_CBC_DEC_TEST_VECTORS
-- }
-+ .enc = __VECS(aes_cbc_enc_tv_template),
-+ .dec = __VECS(aes_cbc_dec_tv_template)
- }
- }
- }, {
-@@ -2444,14 +2687,8 @@ static const struct alg_test_desc alg_te
- .test = alg_test_skcipher,
- .suite = {
- .cipher = {
-- .enc = {
-- .vecs = anubis_cbc_enc_tv_template,
-- .count = ANUBIS_CBC_ENC_TEST_VECTORS
-- },
-- .dec = {
-- .vecs = anubis_cbc_dec_tv_template,
-- .count = ANUBIS_CBC_DEC_TEST_VECTORS
-- }
-+ .enc = __VECS(anubis_cbc_enc_tv_template),
-+ .dec = __VECS(anubis_cbc_dec_tv_template)
- }
- }
- }, {
-@@ -2459,14 +2696,8 @@ static const struct alg_test_desc alg_te
- .test = alg_test_skcipher,
- .suite = {
- .cipher = {
-- .enc = {
-- .vecs = bf_cbc_enc_tv_template,
-- .count = BF_CBC_ENC_TEST_VECTORS
-- },
-- .dec = {
-- .vecs = bf_cbc_dec_tv_template,
-- .count = BF_CBC_DEC_TEST_VECTORS
-- }
-+ .enc = __VECS(bf_cbc_enc_tv_template),
-+ .dec = __VECS(bf_cbc_dec_tv_template)
- }
- }
- }, {
-@@ -2474,14 +2705,8 @@ static const struct alg_test_desc alg_te
- .test = alg_test_skcipher,
- .suite = {
- .cipher = {
-- .enc = {
-- .vecs = camellia_cbc_enc_tv_template,
-- .count = CAMELLIA_CBC_ENC_TEST_VECTORS
-- },
-- .dec = {
-- .vecs = camellia_cbc_dec_tv_template,
-- .count = CAMELLIA_CBC_DEC_TEST_VECTORS
-- }
-+ .enc = __VECS(camellia_cbc_enc_tv_template),
-+ .dec = __VECS(camellia_cbc_dec_tv_template)
- }
- }
- }, {
-@@ -2489,14 +2714,8 @@ static const struct alg_test_desc alg_te
- .test = alg_test_skcipher,
- .suite = {
- .cipher = {
-- .enc = {
-- .vecs = cast5_cbc_enc_tv_template,
-- .count = CAST5_CBC_ENC_TEST_VECTORS
-- },
-- .dec = {
-- .vecs = cast5_cbc_dec_tv_template,
-- .count = CAST5_CBC_DEC_TEST_VECTORS
-- }
-+ .enc = __VECS(cast5_cbc_enc_tv_template),
-+ .dec = __VECS(cast5_cbc_dec_tv_template)
- }
- }
- }, {
-@@ -2504,14 +2723,8 @@ static const struct alg_test_desc alg_te
- .test = alg_test_skcipher,
- .suite = {
- .cipher = {
-- .enc = {
-- .vecs = cast6_cbc_enc_tv_template,
-- .count = CAST6_CBC_ENC_TEST_VECTORS
-- },
-- .dec = {
-- .vecs = cast6_cbc_dec_tv_template,
-- .count = CAST6_CBC_DEC_TEST_VECTORS
-- }
-+ .enc = __VECS(cast6_cbc_enc_tv_template),
-+ .dec = __VECS(cast6_cbc_dec_tv_template)
- }
- }
- }, {
-@@ -2519,14 +2732,8 @@ static const struct alg_test_desc alg_te
- .test = alg_test_skcipher,
- .suite = {
- .cipher = {
-- .enc = {
-- .vecs = des_cbc_enc_tv_template,
-- .count = DES_CBC_ENC_TEST_VECTORS
-- },
-- .dec = {
-- .vecs = des_cbc_dec_tv_template,
-- .count = DES_CBC_DEC_TEST_VECTORS
-- }
-+ .enc = __VECS(des_cbc_enc_tv_template),
-+ .dec = __VECS(des_cbc_dec_tv_template)
- }
- }
- }, {
-@@ -2535,14 +2742,8 @@ static const struct alg_test_desc alg_te
- .fips_allowed = 1,
- .suite = {
- .cipher = {
-- .enc = {
-- .vecs = des3_ede_cbc_enc_tv_template,
-- .count = DES3_EDE_CBC_ENC_TEST_VECTORS
-- },
-- .dec = {
-- .vecs = des3_ede_cbc_dec_tv_template,
-- .count = DES3_EDE_CBC_DEC_TEST_VECTORS
-- }
-+ .enc = __VECS(des3_ede_cbc_enc_tv_template),
-+ .dec = __VECS(des3_ede_cbc_dec_tv_template)
- }
- }
- }, {
-@@ -2550,14 +2751,8 @@ static const struct alg_test_desc alg_te
- .test = alg_test_skcipher,
- .suite = {
- .cipher = {
-- .enc = {
-- .vecs = serpent_cbc_enc_tv_template,
-- .count = SERPENT_CBC_ENC_TEST_VECTORS
-- },
-- .dec = {
-- .vecs = serpent_cbc_dec_tv_template,
-- .count = SERPENT_CBC_DEC_TEST_VECTORS
-- }
-+ .enc = __VECS(serpent_cbc_enc_tv_template),
-+ .dec = __VECS(serpent_cbc_dec_tv_template)
- }
- }
- }, {
-@@ -2565,30 +2760,25 @@ static const struct alg_test_desc alg_te
- .test = alg_test_skcipher,
- .suite = {
- .cipher = {
-- .enc = {
-- .vecs = tf_cbc_enc_tv_template,
-- .count = TF_CBC_ENC_TEST_VECTORS
-- },
-- .dec = {
-- .vecs = tf_cbc_dec_tv_template,
-- .count = TF_CBC_DEC_TEST_VECTORS
-- }
-+ .enc = __VECS(tf_cbc_enc_tv_template),
-+ .dec = __VECS(tf_cbc_dec_tv_template)
- }
- }
- }, {
-+ .alg = "cbcmac(aes)",
-+ .fips_allowed = 1,
-+ .test = alg_test_hash,
-+ .suite = {
-+ .hash = __VECS(aes_cbcmac_tv_template)
-+ }
-+ }, {
- .alg = "ccm(aes)",
- .test = alg_test_aead,
- .fips_allowed = 1,
- .suite = {
- .aead = {
-- .enc = {
-- .vecs = aes_ccm_enc_tv_template,
-- .count = AES_CCM_ENC_TEST_VECTORS
-- },
-- .dec = {
-- .vecs = aes_ccm_dec_tv_template,
-- .count = AES_CCM_DEC_TEST_VECTORS
-- }
-+ .enc = __VECS(aes_ccm_enc_tv_template),
-+ .dec = __VECS(aes_ccm_dec_tv_template)
- }
- }
- }, {
-@@ -2596,14 +2786,8 @@ static const struct alg_test_desc alg_te
- .test = alg_test_skcipher,
- .suite = {
- .cipher = {
-- .enc = {
-- .vecs = chacha20_enc_tv_template,
-- .count = CHACHA20_ENC_TEST_VECTORS
-- },
-- .dec = {
-- .vecs = chacha20_enc_tv_template,
-- .count = CHACHA20_ENC_TEST_VECTORS
-- },
-+ .enc = __VECS(chacha20_enc_tv_template),
-+ .dec = __VECS(chacha20_enc_tv_template),
- }
- }
- }, {
-@@ -2611,20 +2795,14 @@ static const struct alg_test_desc alg_te
- .fips_allowed = 1,
- .test = alg_test_hash,
- .suite = {
-- .hash = {
-- .vecs = aes_cmac128_tv_template,
-- .count = CMAC_AES_TEST_VECTORS
-- }
-+ .hash = __VECS(aes_cmac128_tv_template)
- }
- }, {
- .alg = "cmac(des3_ede)",
- .fips_allowed = 1,
- .test = alg_test_hash,
- .suite = {
-- .hash = {
-- .vecs = des3_ede_cmac64_tv_template,
-- .count = CMAC_DES3_EDE_TEST_VECTORS
-- }
-+ .hash = __VECS(des3_ede_cmac64_tv_template)
- }
- }, {
- .alg = "compress_null",
-@@ -2633,94 +2811,30 @@ static const struct alg_test_desc alg_te
- .alg = "crc32",
- .test = alg_test_hash,
- .suite = {
-- .hash = {
-- .vecs = crc32_tv_template,
-- .count = CRC32_TEST_VECTORS
-- }
-+ .hash = __VECS(crc32_tv_template)
- }
- }, {
- .alg = "crc32c",
- .test = alg_test_crc32c,
- .fips_allowed = 1,
- .suite = {
-- .hash = {
-- .vecs = crc32c_tv_template,
-- .count = CRC32C_TEST_VECTORS
-- }
-+ .hash = __VECS(crc32c_tv_template)
- }
- }, {
- .alg = "crct10dif",
- .test = alg_test_hash,
- .fips_allowed = 1,
- .suite = {
-- .hash = {
-- .vecs = crct10dif_tv_template,
-- .count = CRCT10DIF_TEST_VECTORS
-- }
-+ .hash = __VECS(crct10dif_tv_template)
- }
- }, {
-- .alg = "cryptd(__driver-cbc-aes-aesni)",
-- .test = alg_test_null,
-- .fips_allowed = 1,
-- }, {
-- .alg = "cryptd(__driver-cbc-camellia-aesni)",
-- .test = alg_test_null,
-- }, {
-- .alg = "cryptd(__driver-cbc-camellia-aesni-avx2)",
-- .test = alg_test_null,
-- }, {
-- .alg = "cryptd(__driver-cbc-serpent-avx2)",
-- .test = alg_test_null,
-- }, {
-- .alg = "cryptd(__driver-ecb-aes-aesni)",
-- .test = alg_test_null,
-- .fips_allowed = 1,
-- }, {
-- .alg = "cryptd(__driver-ecb-camellia-aesni)",
-- .test = alg_test_null,
-- }, {
-- .alg = "cryptd(__driver-ecb-camellia-aesni-avx2)",
-- .test = alg_test_null,
-- }, {
-- .alg = "cryptd(__driver-ecb-cast5-avx)",
-- .test = alg_test_null,
-- }, {
-- .alg = "cryptd(__driver-ecb-cast6-avx)",
-- .test = alg_test_null,
-- }, {
-- .alg = "cryptd(__driver-ecb-serpent-avx)",
-- .test = alg_test_null,
-- }, {
-- .alg = "cryptd(__driver-ecb-serpent-avx2)",
-- .test = alg_test_null,
-- }, {
-- .alg = "cryptd(__driver-ecb-serpent-sse2)",
-- .test = alg_test_null,
-- }, {
-- .alg = "cryptd(__driver-ecb-twofish-avx)",
-- .test = alg_test_null,
-- }, {
-- .alg = "cryptd(__driver-gcm-aes-aesni)",
-- .test = alg_test_null,
-- .fips_allowed = 1,
-- }, {
-- .alg = "cryptd(__ghash-pclmulqdqni)",
-- .test = alg_test_null,
-- .fips_allowed = 1,
-- }, {
- .alg = "ctr(aes)",
- .test = alg_test_skcipher,
- .fips_allowed = 1,
- .suite = {
- .cipher = {
-- .enc = {
-- .vecs = aes_ctr_enc_tv_template,
-- .count = AES_CTR_ENC_TEST_VECTORS
-- },
-- .dec = {
-- .vecs = aes_ctr_dec_tv_template,
-- .count = AES_CTR_DEC_TEST_VECTORS
-- }
-+ .enc = __VECS(aes_ctr_enc_tv_template),
-+ .dec = __VECS(aes_ctr_dec_tv_template)
- }
- }
- }, {
-@@ -2728,14 +2842,8 @@ static const struct alg_test_desc alg_te
- .test = alg_test_skcipher,
- .suite = {
- .cipher = {
-- .enc = {
-- .vecs = bf_ctr_enc_tv_template,
-- .count = BF_CTR_ENC_TEST_VECTORS
-- },
-- .dec = {
-- .vecs = bf_ctr_dec_tv_template,
-- .count = BF_CTR_DEC_TEST_VECTORS
-- }
-+ .enc = __VECS(bf_ctr_enc_tv_template),
-+ .dec = __VECS(bf_ctr_dec_tv_template)
- }
- }
- }, {
-@@ -2743,14 +2851,8 @@ static const struct alg_test_desc alg_te
- .test = alg_test_skcipher,
- .suite = {
- .cipher = {
-- .enc = {
-- .vecs = camellia_ctr_enc_tv_template,
-- .count = CAMELLIA_CTR_ENC_TEST_VECTORS
-- },
-- .dec = {
-- .vecs = camellia_ctr_dec_tv_template,
-- .count = CAMELLIA_CTR_DEC_TEST_VECTORS
-- }
-+ .enc = __VECS(camellia_ctr_enc_tv_template),
-+ .dec = __VECS(camellia_ctr_dec_tv_template)
- }
- }
- }, {
-@@ -2758,14 +2860,8 @@ static const struct alg_test_desc alg_te
- .test = alg_test_skcipher,
- .suite = {
- .cipher = {
-- .enc = {
-- .vecs = cast5_ctr_enc_tv_template,
-- .count = CAST5_CTR_ENC_TEST_VECTORS
-- },
-- .dec = {
-- .vecs = cast5_ctr_dec_tv_template,
-- .count = CAST5_CTR_DEC_TEST_VECTORS
-- }
-+ .enc = __VECS(cast5_ctr_enc_tv_template),
-+ .dec = __VECS(cast5_ctr_dec_tv_template)
- }
- }
- }, {
-@@ -2773,14 +2869,8 @@ static const struct alg_test_desc alg_te
- .test = alg_test_skcipher,
- .suite = {
- .cipher = {
-- .enc = {
-- .vecs = cast6_ctr_enc_tv_template,
-- .count = CAST6_CTR_ENC_TEST_VECTORS
-- },
-- .dec = {
-- .vecs = cast6_ctr_dec_tv_template,
-- .count = CAST6_CTR_DEC_TEST_VECTORS
-- }
-+ .enc = __VECS(cast6_ctr_enc_tv_template),
-+ .dec = __VECS(cast6_ctr_dec_tv_template)
- }
- }
- }, {
-@@ -2788,29 +2878,18 @@ static const struct alg_test_desc alg_te
- .test = alg_test_skcipher,
- .suite = {
- .cipher = {
-- .enc = {
-- .vecs = des_ctr_enc_tv_template,
-- .count = DES_CTR_ENC_TEST_VECTORS
-- },
-- .dec = {
-- .vecs = des_ctr_dec_tv_template,
-- .count = DES_CTR_DEC_TEST_VECTORS
-- }
-+ .enc = __VECS(des_ctr_enc_tv_template),
-+ .dec = __VECS(des_ctr_dec_tv_template)
- }
- }
- }, {
- .alg = "ctr(des3_ede)",
- .test = alg_test_skcipher,
-+ .fips_allowed = 1,
- .suite = {
- .cipher = {
-- .enc = {
-- .vecs = des3_ede_ctr_enc_tv_template,
-- .count = DES3_EDE_CTR_ENC_TEST_VECTORS
-- },
-- .dec = {
-- .vecs = des3_ede_ctr_dec_tv_template,
-- .count = DES3_EDE_CTR_DEC_TEST_VECTORS
-- }
-+ .enc = __VECS(des3_ede_ctr_enc_tv_template),
-+ .dec = __VECS(des3_ede_ctr_dec_tv_template)
- }
- }
- }, {
-@@ -2818,14 +2897,8 @@ static const struct alg_test_desc alg_te
- .test = alg_test_skcipher,
- .suite = {
- .cipher = {
-- .enc = {
-- .vecs = serpent_ctr_enc_tv_template,
-- .count = SERPENT_CTR_ENC_TEST_VECTORS
-- },
-- .dec = {
-- .vecs = serpent_ctr_dec_tv_template,
-- .count = SERPENT_CTR_DEC_TEST_VECTORS
-- }
-+ .enc = __VECS(serpent_ctr_enc_tv_template),
-+ .dec = __VECS(serpent_ctr_dec_tv_template)
- }
- }
- }, {
-@@ -2833,14 +2906,8 @@ static const struct alg_test_desc alg_te
- .test = alg_test_skcipher,
- .suite = {
- .cipher = {
-- .enc = {
-- .vecs = tf_ctr_enc_tv_template,
-- .count = TF_CTR_ENC_TEST_VECTORS
-- },
-- .dec = {
-- .vecs = tf_ctr_dec_tv_template,
-- .count = TF_CTR_DEC_TEST_VECTORS
-- }
-+ .enc = __VECS(tf_ctr_enc_tv_template),
-+ .dec = __VECS(tf_ctr_dec_tv_template)
- }
- }
- }, {
-@@ -2848,14 +2915,8 @@ static const struct alg_test_desc alg_te
- .test = alg_test_skcipher,
- .suite = {
- .cipher = {
-- .enc = {
-- .vecs = cts_mode_enc_tv_template,
-- .count = CTS_MODE_ENC_TEST_VECTORS
-- },
-- .dec = {
-- .vecs = cts_mode_dec_tv_template,
-- .count = CTS_MODE_DEC_TEST_VECTORS
-- }
-+ .enc = __VECS(cts_mode_enc_tv_template),
-+ .dec = __VECS(cts_mode_dec_tv_template)
- }
- }
- }, {
-@@ -2864,14 +2925,8 @@ static const struct alg_test_desc alg_te
- .fips_allowed = 1,
- .suite = {
- .comp = {
-- .comp = {
-- .vecs = deflate_comp_tv_template,
-- .count = DEFLATE_COMP_TEST_VECTORS
-- },
-- .decomp = {
-- .vecs = deflate_decomp_tv_template,
-- .count = DEFLATE_DECOMP_TEST_VECTORS
-- }
-+ .comp = __VECS(deflate_comp_tv_template),
-+ .decomp = __VECS(deflate_decomp_tv_template)
- }
- }
- }, {
-@@ -2879,10 +2934,7 @@ static const struct alg_test_desc alg_te
- .test = alg_test_kpp,
- .fips_allowed = 1,
- .suite = {
-- .kpp = {
-- .vecs = dh_tv_template,
-- .count = DH_TEST_VECTORS
-- }
-+ .kpp = __VECS(dh_tv_template)
- }
- }, {
- .alg = "digest_null",
-@@ -2892,30 +2944,21 @@ static const struct alg_test_desc alg_te
- .test = alg_test_drbg,
- .fips_allowed = 1,
- .suite = {
-- .drbg = {
-- .vecs = drbg_nopr_ctr_aes128_tv_template,
-- .count = ARRAY_SIZE(drbg_nopr_ctr_aes128_tv_template)
-- }
-+ .drbg = __VECS(drbg_nopr_ctr_aes128_tv_template)
- }
- }, {
- .alg = "drbg_nopr_ctr_aes192",
- .test = alg_test_drbg,
- .fips_allowed = 1,
- .suite = {
-- .drbg = {
-- .vecs = drbg_nopr_ctr_aes192_tv_template,
-- .count = ARRAY_SIZE(drbg_nopr_ctr_aes192_tv_template)
-- }
-+ .drbg = __VECS(drbg_nopr_ctr_aes192_tv_template)
- }
- }, {
- .alg = "drbg_nopr_ctr_aes256",
- .test = alg_test_drbg,
- .fips_allowed = 1,
- .suite = {
-- .drbg = {
-- .vecs = drbg_nopr_ctr_aes256_tv_template,
-- .count = ARRAY_SIZE(drbg_nopr_ctr_aes256_tv_template)
-- }
-+ .drbg = __VECS(drbg_nopr_ctr_aes256_tv_template)
- }
- }, {
- /*
-@@ -2930,11 +2973,7 @@ static const struct alg_test_desc alg_te
- .test = alg_test_drbg,
- .fips_allowed = 1,
- .suite = {
-- .drbg = {
-- .vecs = drbg_nopr_hmac_sha256_tv_template,
-- .count =
-- ARRAY_SIZE(drbg_nopr_hmac_sha256_tv_template)
-- }
-+ .drbg = __VECS(drbg_nopr_hmac_sha256_tv_template)
- }
- }, {
- /* covered by drbg_nopr_hmac_sha256 test */
-@@ -2954,10 +2993,7 @@ static const struct alg_test_desc alg_te
- .test = alg_test_drbg,
- .fips_allowed = 1,
- .suite = {
-- .drbg = {
-- .vecs = drbg_nopr_sha256_tv_template,
-- .count = ARRAY_SIZE(drbg_nopr_sha256_tv_template)
-- }
-+ .drbg = __VECS(drbg_nopr_sha256_tv_template)
- }
- }, {
- /* covered by drbg_nopr_sha256 test */
-@@ -2973,10 +3009,7 @@ static const struct alg_test_desc alg_te
- .test = alg_test_drbg,
- .fips_allowed = 1,
- .suite = {
-- .drbg = {
-- .vecs = drbg_pr_ctr_aes128_tv_template,
-- .count = ARRAY_SIZE(drbg_pr_ctr_aes128_tv_template)
-- }
-+ .drbg = __VECS(drbg_pr_ctr_aes128_tv_template)
- }
- }, {
- /* covered by drbg_pr_ctr_aes128 test */
-@@ -2996,10 +3029,7 @@ static const struct alg_test_desc alg_te
- .test = alg_test_drbg,
- .fips_allowed = 1,
- .suite = {
-- .drbg = {
-- .vecs = drbg_pr_hmac_sha256_tv_template,
-- .count = ARRAY_SIZE(drbg_pr_hmac_sha256_tv_template)
-- }
-+ .drbg = __VECS(drbg_pr_hmac_sha256_tv_template)
- }
- }, {
- /* covered by drbg_pr_hmac_sha256 test */
-@@ -3019,10 +3049,7 @@ static const struct alg_test_desc alg_te
- .test = alg_test_drbg,
- .fips_allowed = 1,
- .suite = {
-- .drbg = {
-- .vecs = drbg_pr_sha256_tv_template,
-- .count = ARRAY_SIZE(drbg_pr_sha256_tv_template)
-- }
-+ .drbg = __VECS(drbg_pr_sha256_tv_template)
- }
- }, {
- /* covered by drbg_pr_sha256 test */
-@@ -3034,23 +3061,13 @@ static const struct alg_test_desc alg_te
- .fips_allowed = 1,
- .test = alg_test_null,
- }, {
-- .alg = "ecb(__aes-aesni)",
-- .test = alg_test_null,
-- .fips_allowed = 1,
-- }, {
- .alg = "ecb(aes)",
- .test = alg_test_skcipher,
- .fips_allowed = 1,
- .suite = {
- .cipher = {
-- .enc = {
-- .vecs = aes_enc_tv_template,
-- .count = AES_ENC_TEST_VECTORS
-- },
-- .dec = {
-- .vecs = aes_dec_tv_template,
-- .count = AES_DEC_TEST_VECTORS
-- }
-+ .enc = __VECS(aes_enc_tv_template),
-+ .dec = __VECS(aes_dec_tv_template)
- }
- }
- }, {
-@@ -3058,14 +3075,8 @@ static const struct alg_test_desc alg_te
- .test = alg_test_skcipher,
- .suite = {
- .cipher = {
-- .enc = {
-- .vecs = anubis_enc_tv_template,
-- .count = ANUBIS_ENC_TEST_VECTORS
-- },
-- .dec = {
-- .vecs = anubis_dec_tv_template,
-- .count = ANUBIS_DEC_TEST_VECTORS
-- }
-+ .enc = __VECS(anubis_enc_tv_template),
-+ .dec = __VECS(anubis_dec_tv_template)
- }
- }
- }, {
-@@ -3073,14 +3084,8 @@ static const struct alg_test_desc alg_te
- .test = alg_test_skcipher,
- .suite = {
- .cipher = {
-- .enc = {
-- .vecs = arc4_enc_tv_template,
-- .count = ARC4_ENC_TEST_VECTORS
-- },
-- .dec = {
-- .vecs = arc4_dec_tv_template,
-- .count = ARC4_DEC_TEST_VECTORS
-- }
-+ .enc = __VECS(arc4_enc_tv_template),
-+ .dec = __VECS(arc4_dec_tv_template)
- }
- }
- }, {
-@@ -3088,14 +3093,8 @@ static const struct alg_test_desc alg_te
- .test = alg_test_skcipher,
- .suite = {
- .cipher = {
-- .enc = {
-- .vecs = bf_enc_tv_template,
-- .count = BF_ENC_TEST_VECTORS
-- },
-- .dec = {
-- .vecs = bf_dec_tv_template,
-- .count = BF_DEC_TEST_VECTORS
-- }
-+ .enc = __VECS(bf_enc_tv_template),
-+ .dec = __VECS(bf_dec_tv_template)
- }
- }
- }, {
-@@ -3103,14 +3102,8 @@ static const struct alg_test_desc alg_te
- .test = alg_test_skcipher,
- .suite = {
- .cipher = {
-- .enc = {
-- .vecs = camellia_enc_tv_template,
-- .count = CAMELLIA_ENC_TEST_VECTORS
-- },
-- .dec = {
-- .vecs = camellia_dec_tv_template,
-- .count = CAMELLIA_DEC_TEST_VECTORS
-- }
-+ .enc = __VECS(camellia_enc_tv_template),
-+ .dec = __VECS(camellia_dec_tv_template)
- }
- }
- }, {
-@@ -3118,14 +3111,8 @@ static const struct alg_test_desc alg_te
- .test = alg_test_skcipher,
- .suite = {
- .cipher = {
-- .enc = {
-- .vecs = cast5_enc_tv_template,
-- .count = CAST5_ENC_TEST_VECTORS
-- },
-- .dec = {
-- .vecs = cast5_dec_tv_template,
-- .count = CAST5_DEC_TEST_VECTORS
-- }
-+ .enc = __VECS(cast5_enc_tv_template),
-+ .dec = __VECS(cast5_dec_tv_template)
- }
- }
- }, {
-@@ -3133,14 +3120,8 @@ static const struct alg_test_desc alg_te
- .test = alg_test_skcipher,
- .suite = {
- .cipher = {
-- .enc = {
-- .vecs = cast6_enc_tv_template,
-- .count = CAST6_ENC_TEST_VECTORS
-- },
-- .dec = {
-- .vecs = cast6_dec_tv_template,
-- .count = CAST6_DEC_TEST_VECTORS
-- }
-+ .enc = __VECS(cast6_enc_tv_template),
-+ .dec = __VECS(cast6_dec_tv_template)
- }
- }
- }, {
-@@ -3151,14 +3132,8 @@ static const struct alg_test_desc alg_te
- .test = alg_test_skcipher,
- .suite = {
- .cipher = {
-- .enc = {
-- .vecs = des_enc_tv_template,
-- .count = DES_ENC_TEST_VECTORS
-- },
-- .dec = {
-- .vecs = des_dec_tv_template,
-- .count = DES_DEC_TEST_VECTORS
-- }
-+ .enc = __VECS(des_enc_tv_template),
-+ .dec = __VECS(des_dec_tv_template)
- }
- }
- }, {
-@@ -3167,14 +3142,8 @@ static const struct alg_test_desc alg_te
- .fips_allowed = 1,
- .suite = {
- .cipher = {
-- .enc = {
-- .vecs = des3_ede_enc_tv_template,
-- .count = DES3_EDE_ENC_TEST_VECTORS
-- },
-- .dec = {
-- .vecs = des3_ede_dec_tv_template,
-- .count = DES3_EDE_DEC_TEST_VECTORS
-- }
-+ .enc = __VECS(des3_ede_enc_tv_template),
-+ .dec = __VECS(des3_ede_dec_tv_template)
- }
- }
- }, {
-@@ -3197,14 +3166,8 @@ static const struct alg_test_desc alg_te
- .test = alg_test_skcipher,
- .suite = {
- .cipher = {
-- .enc = {
-- .vecs = khazad_enc_tv_template,
-- .count = KHAZAD_ENC_TEST_VECTORS
-- },
-- .dec = {
-- .vecs = khazad_dec_tv_template,
-- .count = KHAZAD_DEC_TEST_VECTORS
-- }
-+ .enc = __VECS(khazad_enc_tv_template),
-+ .dec = __VECS(khazad_dec_tv_template)
- }
- }
- }, {
-@@ -3212,14 +3175,8 @@ static const struct alg_test_desc alg_te
- .test = alg_test_skcipher,
- .suite = {
- .cipher = {
-- .enc = {
-- .vecs = seed_enc_tv_template,
-- .count = SEED_ENC_TEST_VECTORS
-- },
-- .dec = {
-- .vecs = seed_dec_tv_template,
-- .count = SEED_DEC_TEST_VECTORS
-- }
-+ .enc = __VECS(seed_enc_tv_template),
-+ .dec = __VECS(seed_dec_tv_template)
- }
- }
- }, {
-@@ -3227,14 +3184,8 @@ static const struct alg_test_desc alg_te
- .test = alg_test_skcipher,
- .suite = {
- .cipher = {
-- .enc = {
-- .vecs = serpent_enc_tv_template,
-- .count = SERPENT_ENC_TEST_VECTORS
-- },
-- .dec = {
-- .vecs = serpent_dec_tv_template,
-- .count = SERPENT_DEC_TEST_VECTORS
-- }
-+ .enc = __VECS(serpent_enc_tv_template),
-+ .dec = __VECS(serpent_dec_tv_template)
- }
- }
- }, {
-@@ -3242,14 +3193,8 @@ static const struct alg_test_desc alg_te
- .test = alg_test_skcipher,
- .suite = {
- .cipher = {
-- .enc = {
-- .vecs = tea_enc_tv_template,
-- .count = TEA_ENC_TEST_VECTORS
-- },
-- .dec = {
-- .vecs = tea_dec_tv_template,
-- .count = TEA_DEC_TEST_VECTORS
-- }
-+ .enc = __VECS(tea_enc_tv_template),
-+ .dec = __VECS(tea_dec_tv_template)
- }
- }
- }, {
-@@ -3257,14 +3202,8 @@ static const struct alg_test_desc alg_te
- .test = alg_test_skcipher,
- .suite = {
- .cipher = {
-- .enc = {
-- .vecs = tnepres_enc_tv_template,
-- .count = TNEPRES_ENC_TEST_VECTORS
-- },
-- .dec = {
-- .vecs = tnepres_dec_tv_template,
-- .count = TNEPRES_DEC_TEST_VECTORS
-- }
-+ .enc = __VECS(tnepres_enc_tv_template),
-+ .dec = __VECS(tnepres_dec_tv_template)
- }
- }
- }, {
-@@ -3272,14 +3211,8 @@ static const struct alg_test_desc alg_te
- .test = alg_test_skcipher,
- .suite = {
- .cipher = {
-- .enc = {
-- .vecs = tf_enc_tv_template,
-- .count = TF_ENC_TEST_VECTORS
-- },
-- .dec = {
-- .vecs = tf_dec_tv_template,
-- .count = TF_DEC_TEST_VECTORS
-- }
-+ .enc = __VECS(tf_enc_tv_template),
-+ .dec = __VECS(tf_dec_tv_template)
- }
- }
- }, {
-@@ -3287,14 +3220,8 @@ static const struct alg_test_desc alg_te
- .test = alg_test_skcipher,
- .suite = {
- .cipher = {
-- .enc = {
-- .vecs = xeta_enc_tv_template,
-- .count = XETA_ENC_TEST_VECTORS
-- },
-- .dec = {
-- .vecs = xeta_dec_tv_template,
-- .count = XETA_DEC_TEST_VECTORS
-- }
-+ .enc = __VECS(xeta_enc_tv_template),
-+ .dec = __VECS(xeta_dec_tv_template)
- }
- }
- }, {
-@@ -3302,14 +3229,8 @@ static const struct alg_test_desc alg_te
- .test = alg_test_skcipher,
- .suite = {
- .cipher = {
-- .enc = {
-- .vecs = xtea_enc_tv_template,
-- .count = XTEA_ENC_TEST_VECTORS
-- },
-- .dec = {
-- .vecs = xtea_dec_tv_template,
-- .count = XTEA_DEC_TEST_VECTORS
-- }
-+ .enc = __VECS(xtea_enc_tv_template),
-+ .dec = __VECS(xtea_dec_tv_template)
- }
- }
- }, {
-@@ -3317,10 +3238,7 @@ static const struct alg_test_desc alg_te
- .test = alg_test_kpp,
- .fips_allowed = 1,
- .suite = {
-- .kpp = {
-- .vecs = ecdh_tv_template,
-- .count = ECDH_TEST_VECTORS
-- }
-+ .kpp = __VECS(ecdh_tv_template)
- }
- }, {
- .alg = "gcm(aes)",
-@@ -3328,14 +3246,8 @@ static const struct alg_test_desc alg_te
- .fips_allowed = 1,
- .suite = {
- .aead = {
-- .enc = {
-- .vecs = aes_gcm_enc_tv_template,
-- .count = AES_GCM_ENC_TEST_VECTORS
-- },
-- .dec = {
-- .vecs = aes_gcm_dec_tv_template,
-- .count = AES_GCM_DEC_TEST_VECTORS
-- }
-+ .enc = __VECS(aes_gcm_enc_tv_template),
-+ .dec = __VECS(aes_gcm_dec_tv_template)
- }
- }
- }, {
-@@ -3343,136 +3255,94 @@ static const struct alg_test_desc alg_te
- .test = alg_test_hash,
- .fips_allowed = 1,
- .suite = {
-- .hash = {
-- .vecs = ghash_tv_template,
-- .count = GHASH_TEST_VECTORS
-- }
-+ .hash = __VECS(ghash_tv_template)
- }
- }, {
- .alg = "hmac(crc32)",
- .test = alg_test_hash,
- .suite = {
-- .hash = {
-- .vecs = bfin_crc_tv_template,
-- .count = BFIN_CRC_TEST_VECTORS
-- }
-+ .hash = __VECS(bfin_crc_tv_template)
- }
- }, {
- .alg = "hmac(md5)",
- .test = alg_test_hash,
- .suite = {
-- .hash = {
-- .vecs = hmac_md5_tv_template,
-- .count = HMAC_MD5_TEST_VECTORS
-- }
-+ .hash = __VECS(hmac_md5_tv_template)
- }
- }, {
- .alg = "hmac(rmd128)",
- .test = alg_test_hash,
- .suite = {
-- .hash = {
-- .vecs = hmac_rmd128_tv_template,
-- .count = HMAC_RMD128_TEST_VECTORS
-- }
-+ .hash = __VECS(hmac_rmd128_tv_template)
- }
- }, {
- .alg = "hmac(rmd160)",
- .test = alg_test_hash,
- .suite = {
-- .hash = {
-- .vecs = hmac_rmd160_tv_template,
-- .count = HMAC_RMD160_TEST_VECTORS
-- }
-+ .hash = __VECS(hmac_rmd160_tv_template)
- }
- }, {
- .alg = "hmac(sha1)",
- .test = alg_test_hash,
- .fips_allowed = 1,
- .suite = {
-- .hash = {
-- .vecs = hmac_sha1_tv_template,
-- .count = HMAC_SHA1_TEST_VECTORS
-- }
-+ .hash = __VECS(hmac_sha1_tv_template)
- }
- }, {
- .alg = "hmac(sha224)",
- .test = alg_test_hash,
- .fips_allowed = 1,
- .suite = {
-- .hash = {
-- .vecs = hmac_sha224_tv_template,
-- .count = HMAC_SHA224_TEST_VECTORS
-- }
-+ .hash = __VECS(hmac_sha224_tv_template)
- }
- }, {
- .alg = "hmac(sha256)",
- .test = alg_test_hash,
- .fips_allowed = 1,
- .suite = {
-- .hash = {
-- .vecs = hmac_sha256_tv_template,
-- .count = HMAC_SHA256_TEST_VECTORS
-- }
-+ .hash = __VECS(hmac_sha256_tv_template)
- }
- }, {
- .alg = "hmac(sha3-224)",
- .test = alg_test_hash,
- .fips_allowed = 1,
- .suite = {
-- .hash = {
-- .vecs = hmac_sha3_224_tv_template,
-- .count = HMAC_SHA3_224_TEST_VECTORS
-- }
-+ .hash = __VECS(hmac_sha3_224_tv_template)
- }
- }, {
- .alg = "hmac(sha3-256)",
- .test = alg_test_hash,
- .fips_allowed = 1,
- .suite = {
-- .hash = {
-- .vecs = hmac_sha3_256_tv_template,
-- .count = HMAC_SHA3_256_TEST_VECTORS
-- }
-+ .hash = __VECS(hmac_sha3_256_tv_template)
- }
- }, {
- .alg = "hmac(sha3-384)",
- .test = alg_test_hash,
- .fips_allowed = 1,
- .suite = {
-- .hash = {
-- .vecs = hmac_sha3_384_tv_template,
-- .count = HMAC_SHA3_384_TEST_VECTORS
-- }
-+ .hash = __VECS(hmac_sha3_384_tv_template)
- }
- }, {
- .alg = "hmac(sha3-512)",
- .test = alg_test_hash,
- .fips_allowed = 1,
- .suite = {
-- .hash = {
-- .vecs = hmac_sha3_512_tv_template,
-- .count = HMAC_SHA3_512_TEST_VECTORS
-- }
-+ .hash = __VECS(hmac_sha3_512_tv_template)
- }
- }, {
- .alg = "hmac(sha384)",
- .test = alg_test_hash,
- .fips_allowed = 1,
- .suite = {
-- .hash = {
-- .vecs = hmac_sha384_tv_template,
-- .count = HMAC_SHA384_TEST_VECTORS
-- }
-+ .hash = __VECS(hmac_sha384_tv_template)
- }
- }, {
- .alg = "hmac(sha512)",
- .test = alg_test_hash,
- .fips_allowed = 1,
- .suite = {
-- .hash = {
-- .vecs = hmac_sha512_tv_template,
-- .count = HMAC_SHA512_TEST_VECTORS
-- }
-+ .hash = __VECS(hmac_sha512_tv_template)
- }
- }, {
- .alg = "jitterentropy_rng",
-@@ -3484,14 +3354,8 @@ static const struct alg_test_desc alg_te
- .fips_allowed = 1,
- .suite = {
- .cipher = {
-- .enc = {
-- .vecs = aes_kw_enc_tv_template,
-- .count = ARRAY_SIZE(aes_kw_enc_tv_template)
-- },
-- .dec = {
-- .vecs = aes_kw_dec_tv_template,
-- .count = ARRAY_SIZE(aes_kw_dec_tv_template)
-- }
-+ .enc = __VECS(aes_kw_enc_tv_template),
-+ .dec = __VECS(aes_kw_dec_tv_template)
- }
- }
- }, {
-@@ -3499,14 +3363,8 @@ static const struct alg_test_desc alg_te
- .test = alg_test_skcipher,
- .suite = {
- .cipher = {
-- .enc = {
-- .vecs = aes_lrw_enc_tv_template,
-- .count = AES_LRW_ENC_TEST_VECTORS
-- },
-- .dec = {
-- .vecs = aes_lrw_dec_tv_template,
-- .count = AES_LRW_DEC_TEST_VECTORS
-- }
-+ .enc = __VECS(aes_lrw_enc_tv_template),
-+ .dec = __VECS(aes_lrw_dec_tv_template)
- }
- }
- }, {
-@@ -3514,14 +3372,8 @@ static const struct alg_test_desc alg_te
- .test = alg_test_skcipher,
- .suite = {
- .cipher = {
-- .enc = {
-- .vecs = camellia_lrw_enc_tv_template,
-- .count = CAMELLIA_LRW_ENC_TEST_VECTORS
-- },
-- .dec = {
-- .vecs = camellia_lrw_dec_tv_template,
-- .count = CAMELLIA_LRW_DEC_TEST_VECTORS
-- }
-+ .enc = __VECS(camellia_lrw_enc_tv_template),
-+ .dec = __VECS(camellia_lrw_dec_tv_template)
- }
- }
- }, {
-@@ -3529,14 +3381,8 @@ static const struct alg_test_desc alg_te
- .test = alg_test_skcipher,
- .suite = {
- .cipher = {
-- .enc = {
-- .vecs = cast6_lrw_enc_tv_template,
-- .count = CAST6_LRW_ENC_TEST_VECTORS
-- },
-- .dec = {
-- .vecs = cast6_lrw_dec_tv_template,
-- .count = CAST6_LRW_DEC_TEST_VECTORS
-- }
-+ .enc = __VECS(cast6_lrw_enc_tv_template),
-+ .dec = __VECS(cast6_lrw_dec_tv_template)
- }
- }
- }, {
-@@ -3544,14 +3390,8 @@ static const struct alg_test_desc alg_te
- .test = alg_test_skcipher,
- .suite = {
- .cipher = {
-- .enc = {
-- .vecs = serpent_lrw_enc_tv_template,
-- .count = SERPENT_LRW_ENC_TEST_VECTORS
-- },
-- .dec = {
-- .vecs = serpent_lrw_dec_tv_template,
-- .count = SERPENT_LRW_DEC_TEST_VECTORS
-- }
-+ .enc = __VECS(serpent_lrw_enc_tv_template),
-+ .dec = __VECS(serpent_lrw_dec_tv_template)
- }
- }
- }, {
-@@ -3559,14 +3399,8 @@ static const struct alg_test_desc alg_te
- .test = alg_test_skcipher,
- .suite = {
- .cipher = {
-- .enc = {
-- .vecs = tf_lrw_enc_tv_template,
-- .count = TF_LRW_ENC_TEST_VECTORS
-- },
-- .dec = {
-- .vecs = tf_lrw_dec_tv_template,
-- .count = TF_LRW_DEC_TEST_VECTORS
-- }
-+ .enc = __VECS(tf_lrw_enc_tv_template),
-+ .dec = __VECS(tf_lrw_dec_tv_template)
- }
- }
- }, {
-@@ -3575,14 +3409,8 @@ static const struct alg_test_desc alg_te
- .fips_allowed = 1,
- .suite = {
- .comp = {
-- .comp = {
-- .vecs = lz4_comp_tv_template,
-- .count = LZ4_COMP_TEST_VECTORS
-- },
-- .decomp = {
-- .vecs = lz4_decomp_tv_template,
-- .count = LZ4_DECOMP_TEST_VECTORS
-- }
-+ .comp = __VECS(lz4_comp_tv_template),
-+ .decomp = __VECS(lz4_decomp_tv_template)
- }
- }
- }, {
-@@ -3591,14 +3419,8 @@ static const struct alg_test_desc alg_te
- .fips_allowed = 1,
- .suite = {
- .comp = {
-- .comp = {
-- .vecs = lz4hc_comp_tv_template,
-- .count = LZ4HC_COMP_TEST_VECTORS
-- },
-- .decomp = {
-- .vecs = lz4hc_decomp_tv_template,
-- .count = LZ4HC_DECOMP_TEST_VECTORS
-- }
-+ .comp = __VECS(lz4hc_comp_tv_template),
-+ .decomp = __VECS(lz4hc_decomp_tv_template)
- }
- }
- }, {
-@@ -3607,42 +3429,27 @@ static const struct alg_test_desc alg_te
- .fips_allowed = 1,
- .suite = {
- .comp = {
-- .comp = {
-- .vecs = lzo_comp_tv_template,
-- .count = LZO_COMP_TEST_VECTORS
-- },
-- .decomp = {
-- .vecs = lzo_decomp_tv_template,
-- .count = LZO_DECOMP_TEST_VECTORS
-- }
-+ .comp = __VECS(lzo_comp_tv_template),
-+ .decomp = __VECS(lzo_decomp_tv_template)
- }
- }
- }, {
- .alg = "md4",
- .test = alg_test_hash,
- .suite = {
-- .hash = {
-- .vecs = md4_tv_template,
-- .count = MD4_TEST_VECTORS
-- }
-+ .hash = __VECS(md4_tv_template)
- }
- }, {
- .alg = "md5",
- .test = alg_test_hash,
- .suite = {
-- .hash = {
-- .vecs = md5_tv_template,
-- .count = MD5_TEST_VECTORS
-- }
-+ .hash = __VECS(md5_tv_template)
- }
- }, {
- .alg = "michael_mic",
- .test = alg_test_hash,
- .suite = {
-- .hash = {
-- .vecs = michael_mic_tv_template,
-- .count = MICHAEL_MIC_TEST_VECTORS
-- }
-+ .hash = __VECS(michael_mic_tv_template)
- }
- }, {
- .alg = "ofb(aes)",
-@@ -3650,14 +3457,8 @@ static const struct alg_test_desc alg_te
- .fips_allowed = 1,
- .suite = {
- .cipher = {
-- .enc = {
-- .vecs = aes_ofb_enc_tv_template,
-- .count = AES_OFB_ENC_TEST_VECTORS
-- },
-- .dec = {
-- .vecs = aes_ofb_dec_tv_template,
-- .count = AES_OFB_DEC_TEST_VECTORS
-- }
-+ .enc = __VECS(aes_ofb_enc_tv_template),
-+ .dec = __VECS(aes_ofb_dec_tv_template)
- }
- }
- }, {
-@@ -3665,24 +3466,15 @@ static const struct alg_test_desc alg_te
- .test = alg_test_skcipher,
- .suite = {
- .cipher = {
-- .enc = {
-- .vecs = fcrypt_pcbc_enc_tv_template,
-- .count = FCRYPT_ENC_TEST_VECTORS
-- },
-- .dec = {
-- .vecs = fcrypt_pcbc_dec_tv_template,
-- .count = FCRYPT_DEC_TEST_VECTORS
-- }
-+ .enc = __VECS(fcrypt_pcbc_enc_tv_template),
-+ .dec = __VECS(fcrypt_pcbc_dec_tv_template)
- }
- }
- }, {
- .alg = "poly1305",
- .test = alg_test_hash,
- .suite = {
-- .hash = {
-- .vecs = poly1305_tv_template,
-- .count = POLY1305_TEST_VECTORS
-- }
-+ .hash = __VECS(poly1305_tv_template)
- }
- }, {
- .alg = "rfc3686(ctr(aes))",
-@@ -3690,14 +3482,8 @@ static const struct alg_test_desc alg_te
- .fips_allowed = 1,
- .suite = {
- .cipher = {
-- .enc = {
-- .vecs = aes_ctr_rfc3686_enc_tv_template,
-- .count = AES_CTR_3686_ENC_TEST_VECTORS
-- },
-- .dec = {
-- .vecs = aes_ctr_rfc3686_dec_tv_template,
-- .count = AES_CTR_3686_DEC_TEST_VECTORS
-- }
-+ .enc = __VECS(aes_ctr_rfc3686_enc_tv_template),
-+ .dec = __VECS(aes_ctr_rfc3686_dec_tv_template)
- }
- }
- }, {
-@@ -3706,14 +3492,8 @@ static const struct alg_test_desc alg_te
- .fips_allowed = 1,
- .suite = {
- .aead = {
-- .enc = {
-- .vecs = aes_gcm_rfc4106_enc_tv_template,
-- .count = AES_GCM_4106_ENC_TEST_VECTORS
-- },
-- .dec = {
-- .vecs = aes_gcm_rfc4106_dec_tv_template,
-- .count = AES_GCM_4106_DEC_TEST_VECTORS
-- }
-+ .enc = __VECS(aes_gcm_rfc4106_enc_tv_template),
-+ .dec = __VECS(aes_gcm_rfc4106_dec_tv_template)
- }
- }
- }, {
-@@ -3722,14 +3502,8 @@ static const struct alg_test_desc alg_te
- .fips_allowed = 1,
- .suite = {
- .aead = {
-- .enc = {
-- .vecs = aes_ccm_rfc4309_enc_tv_template,
-- .count = AES_CCM_4309_ENC_TEST_VECTORS
-- },
-- .dec = {
-- .vecs = aes_ccm_rfc4309_dec_tv_template,
-- .count = AES_CCM_4309_DEC_TEST_VECTORS
-- }
-+ .enc = __VECS(aes_ccm_rfc4309_enc_tv_template),
-+ .dec = __VECS(aes_ccm_rfc4309_dec_tv_template)
- }
- }
- }, {
-@@ -3737,14 +3511,8 @@ static const struct alg_test_desc alg_te
- .test = alg_test_aead,
- .suite = {
- .aead = {
-- .enc = {
-- .vecs = aes_gcm_rfc4543_enc_tv_template,
-- .count = AES_GCM_4543_ENC_TEST_VECTORS
-- },
-- .dec = {
-- .vecs = aes_gcm_rfc4543_dec_tv_template,
-- .count = AES_GCM_4543_DEC_TEST_VECTORS
-- },
-+ .enc = __VECS(aes_gcm_rfc4543_enc_tv_template),
-+ .dec = __VECS(aes_gcm_rfc4543_dec_tv_template),
- }
- }
- }, {
-@@ -3752,14 +3520,8 @@ static const struct alg_test_desc alg_te
- .test = alg_test_aead,
- .suite = {
- .aead = {
-- .enc = {
-- .vecs = rfc7539_enc_tv_template,
-- .count = RFC7539_ENC_TEST_VECTORS
-- },
-- .dec = {
-- .vecs = rfc7539_dec_tv_template,
-- .count = RFC7539_DEC_TEST_VECTORS
-- },
-+ .enc = __VECS(rfc7539_enc_tv_template),
-+ .dec = __VECS(rfc7539_dec_tv_template),
- }
- }
- }, {
-@@ -3767,71 +3529,47 @@ static const struct alg_test_desc alg_te
- .test = alg_test_aead,
- .suite = {
- .aead = {
-- .enc = {
-- .vecs = rfc7539esp_enc_tv_template,
-- .count = RFC7539ESP_ENC_TEST_VECTORS
-- },
-- .dec = {
-- .vecs = rfc7539esp_dec_tv_template,
-- .count = RFC7539ESP_DEC_TEST_VECTORS
-- },
-+ .enc = __VECS(rfc7539esp_enc_tv_template),
-+ .dec = __VECS(rfc7539esp_dec_tv_template),
- }
- }
- }, {
- .alg = "rmd128",
- .test = alg_test_hash,
- .suite = {
-- .hash = {
-- .vecs = rmd128_tv_template,
-- .count = RMD128_TEST_VECTORS
-- }
-+ .hash = __VECS(rmd128_tv_template)
- }
- }, {
- .alg = "rmd160",
- .test = alg_test_hash,
- .suite = {
-- .hash = {
-- .vecs = rmd160_tv_template,
-- .count = RMD160_TEST_VECTORS
-- }
-+ .hash = __VECS(rmd160_tv_template)
- }
- }, {
- .alg = "rmd256",
- .test = alg_test_hash,
- .suite = {
-- .hash = {
-- .vecs = rmd256_tv_template,
-- .count = RMD256_TEST_VECTORS
-- }
-+ .hash = __VECS(rmd256_tv_template)
- }
- }, {
- .alg = "rmd320",
- .test = alg_test_hash,
- .suite = {
-- .hash = {
-- .vecs = rmd320_tv_template,
-- .count = RMD320_TEST_VECTORS
-- }
-+ .hash = __VECS(rmd320_tv_template)
- }
- }, {
- .alg = "rsa",
- .test = alg_test_akcipher,
- .fips_allowed = 1,
- .suite = {
-- .akcipher = {
-- .vecs = rsa_tv_template,
-- .count = RSA_TEST_VECTORS
-- }
-+ .akcipher = __VECS(rsa_tv_template)
- }
- }, {
- .alg = "salsa20",
- .test = alg_test_skcipher,
- .suite = {
- .cipher = {
-- .enc = {
-- .vecs = salsa20_stream_enc_tv_template,
-- .count = SALSA20_STREAM_ENC_TEST_VECTORS
-- }
-+ .enc = __VECS(salsa20_stream_enc_tv_template)
- }
- }
- }, {
-@@ -3839,162 +3577,120 @@ static const struct alg_test_desc alg_te
- .test = alg_test_hash,
- .fips_allowed = 1,
- .suite = {
-- .hash = {
-- .vecs = sha1_tv_template,
-- .count = SHA1_TEST_VECTORS
-- }
-+ .hash = __VECS(sha1_tv_template)
- }
- }, {
- .alg = "sha224",
- .test = alg_test_hash,
- .fips_allowed = 1,
- .suite = {
-- .hash = {
-- .vecs = sha224_tv_template,
-- .count = SHA224_TEST_VECTORS
-- }
-+ .hash = __VECS(sha224_tv_template)
- }
- }, {
- .alg = "sha256",
- .test = alg_test_hash,
- .fips_allowed = 1,
- .suite = {
-- .hash = {
-- .vecs = sha256_tv_template,
-- .count = SHA256_TEST_VECTORS
-- }
-+ .hash = __VECS(sha256_tv_template)
- }
- }, {
- .alg = "sha3-224",
- .test = alg_test_hash,
- .fips_allowed = 1,
- .suite = {
-- .hash = {
-- .vecs = sha3_224_tv_template,
-- .count = SHA3_224_TEST_VECTORS
-- }
-+ .hash = __VECS(sha3_224_tv_template)
- }
- }, {
- .alg = "sha3-256",
- .test = alg_test_hash,
- .fips_allowed = 1,
- .suite = {
-- .hash = {
-- .vecs = sha3_256_tv_template,
-- .count = SHA3_256_TEST_VECTORS
-- }
-+ .hash = __VECS(sha3_256_tv_template)
- }
- }, {
- .alg = "sha3-384",
- .test = alg_test_hash,
- .fips_allowed = 1,
- .suite = {
-- .hash = {
-- .vecs = sha3_384_tv_template,
-- .count = SHA3_384_TEST_VECTORS
-- }
-+ .hash = __VECS(sha3_384_tv_template)
- }
- }, {
- .alg = "sha3-512",
- .test = alg_test_hash,
- .fips_allowed = 1,
- .suite = {
-- .hash = {
-- .vecs = sha3_512_tv_template,
-- .count = SHA3_512_TEST_VECTORS
-- }
-+ .hash = __VECS(sha3_512_tv_template)
- }
- }, {
- .alg = "sha384",
- .test = alg_test_hash,
- .fips_allowed = 1,
- .suite = {
-- .hash = {
-- .vecs = sha384_tv_template,
-- .count = SHA384_TEST_VECTORS
-- }
-+ .hash = __VECS(sha384_tv_template)
- }
- }, {
- .alg = "sha512",
- .test = alg_test_hash,
- .fips_allowed = 1,
- .suite = {
-- .hash = {
-- .vecs = sha512_tv_template,
-- .count = SHA512_TEST_VECTORS
-- }
-+ .hash = __VECS(sha512_tv_template)
- }
- }, {
- .alg = "tgr128",
- .test = alg_test_hash,
- .suite = {
-- .hash = {
-- .vecs = tgr128_tv_template,
-- .count = TGR128_TEST_VECTORS
-- }
-+ .hash = __VECS(tgr128_tv_template)
- }
- }, {
- .alg = "tgr160",
- .test = alg_test_hash,
- .suite = {
-- .hash = {
-- .vecs = tgr160_tv_template,
-- .count = TGR160_TEST_VECTORS
-- }
-+ .hash = __VECS(tgr160_tv_template)
- }
- }, {
- .alg = "tgr192",
- .test = alg_test_hash,
- .suite = {
-- .hash = {
-- .vecs = tgr192_tv_template,
-- .count = TGR192_TEST_VECTORS
-+ .hash = __VECS(tgr192_tv_template)
-+ }
-+ }, {
-+ .alg = "tls10(hmac(sha1),cbc(aes))",
-+ .test = alg_test_tls,
-+ .suite = {
-+ .tls = {
-+ .enc = __VECS(tls_enc_tv_template),
-+ .dec = __VECS(tls_dec_tv_template)
- }
- }
- }, {
- .alg = "vmac(aes)",
- .test = alg_test_hash,
- .suite = {
-- .hash = {
-- .vecs = aes_vmac128_tv_template,
-- .count = VMAC_AES_TEST_VECTORS
-- }
-+ .hash = __VECS(aes_vmac128_tv_template)
- }
- }, {
- .alg = "wp256",
- .test = alg_test_hash,
- .suite = {
-- .hash = {
-- .vecs = wp256_tv_template,
-- .count = WP256_TEST_VECTORS
-- }
-+ .hash = __VECS(wp256_tv_template)
- }
- }, {
- .alg = "wp384",
- .test = alg_test_hash,
- .suite = {
-- .hash = {
-- .vecs = wp384_tv_template,
-- .count = WP384_TEST_VECTORS
-- }
-+ .hash = __VECS(wp384_tv_template)
- }
- }, {
- .alg = "wp512",
- .test = alg_test_hash,
- .suite = {
-- .hash = {
-- .vecs = wp512_tv_template,
-- .count = WP512_TEST_VECTORS
-- }
-+ .hash = __VECS(wp512_tv_template)
- }
- }, {
- .alg = "xcbc(aes)",
- .test = alg_test_hash,
- .suite = {
-- .hash = {
-- .vecs = aes_xcbc128_tv_template,
-- .count = XCBC_AES_TEST_VECTORS
-- }
-+ .hash = __VECS(aes_xcbc128_tv_template)
- }
- }, {
- .alg = "xts(aes)",
-@@ -4002,14 +3698,8 @@ static const struct alg_test_desc alg_te
- .fips_allowed = 1,
- .suite = {
- .cipher = {
-- .enc = {
-- .vecs = aes_xts_enc_tv_template,
-- .count = AES_XTS_ENC_TEST_VECTORS
-- },
-- .dec = {
-- .vecs = aes_xts_dec_tv_template,
-- .count = AES_XTS_DEC_TEST_VECTORS
-- }
-+ .enc = __VECS(aes_xts_enc_tv_template),
-+ .dec = __VECS(aes_xts_dec_tv_template)
- }
- }
- }, {
-@@ -4017,14 +3707,8 @@ static const struct alg_test_desc alg_te
- .test = alg_test_skcipher,
- .suite = {
- .cipher = {
-- .enc = {
-- .vecs = camellia_xts_enc_tv_template,
-- .count = CAMELLIA_XTS_ENC_TEST_VECTORS
-- },
-- .dec = {
-- .vecs = camellia_xts_dec_tv_template,
-- .count = CAMELLIA_XTS_DEC_TEST_VECTORS
-- }
-+ .enc = __VECS(camellia_xts_enc_tv_template),
-+ .dec = __VECS(camellia_xts_dec_tv_template)
- }
- }
- }, {
-@@ -4032,14 +3716,8 @@ static const struct alg_test_desc alg_te
- .test = alg_test_skcipher,
- .suite = {
- .cipher = {
-- .enc = {
-- .vecs = cast6_xts_enc_tv_template,
-- .count = CAST6_XTS_ENC_TEST_VECTORS
-- },
-- .dec = {
-- .vecs = cast6_xts_dec_tv_template,
-- .count = CAST6_XTS_DEC_TEST_VECTORS
-- }
-+ .enc = __VECS(cast6_xts_enc_tv_template),
-+ .dec = __VECS(cast6_xts_dec_tv_template)
- }
- }
- }, {
-@@ -4047,14 +3725,8 @@ static const struct alg_test_desc alg_te
- .test = alg_test_skcipher,
- .suite = {
- .cipher = {
-- .enc = {
-- .vecs = serpent_xts_enc_tv_template,
-- .count = SERPENT_XTS_ENC_TEST_VECTORS
-- },
-- .dec = {
-- .vecs = serpent_xts_dec_tv_template,
-- .count = SERPENT_XTS_DEC_TEST_VECTORS
-- }
-+ .enc = __VECS(serpent_xts_enc_tv_template),
-+ .dec = __VECS(serpent_xts_dec_tv_template)
- }
- }
- }, {
-@@ -4062,14 +3734,8 @@ static const struct alg_test_desc alg_te
- .test = alg_test_skcipher,
- .suite = {
- .cipher = {
-- .enc = {
-- .vecs = tf_xts_enc_tv_template,
-- .count = TF_XTS_ENC_TEST_VECTORS
-- },
-- .dec = {
-- .vecs = tf_xts_dec_tv_template,
-- .count = TF_XTS_DEC_TEST_VECTORS
-- }
-+ .enc = __VECS(tf_xts_enc_tv_template),
-+ .dec = __VECS(tf_xts_dec_tv_template)
- }
- }
- }
---- a/crypto/testmgr.h
-+++ b/crypto/testmgr.h
-@@ -34,9 +34,9 @@
-
- struct hash_testvec {
- /* only used with keyed hash algorithms */
-- char *key;
-- char *plaintext;
-- char *digest;
-+ const char *key;
-+ const char *plaintext;
-+ const char *digest;
- unsigned char tap[MAX_TAP];
- unsigned short psize;
- unsigned char np;
-@@ -63,11 +63,11 @@ struct hash_testvec {
- */
-
- struct cipher_testvec {
-- char *key;
-- char *iv;
-- char *iv_out;
-- char *input;
-- char *result;
-+ const char *key;
-+ const char *iv;
-+ const char *iv_out;
-+ const char *input;
-+ const char *result;
- unsigned short tap[MAX_TAP];
- int np;
- unsigned char also_non_np;
-@@ -80,11 +80,11 @@ struct cipher_testvec {
- };
-
- struct aead_testvec {
-- char *key;
-- char *iv;
-- char *input;
-- char *assoc;
-- char *result;
-+ const char *key;
-+ const char *iv;
-+ const char *input;
-+ const char *assoc;
-+ const char *result;
- unsigned char tap[MAX_TAP];
- unsigned char atap[MAX_TAP];
- int np;
-@@ -99,10 +99,10 @@ struct aead_testvec {
- };
-
- struct cprng_testvec {
-- char *key;
-- char *dt;
-- char *v;
-- char *result;
-+ const char *key;
-+ const char *dt;
-+ const char *v;
-+ const char *result;
- unsigned char klen;
- unsigned short dtlen;
- unsigned short vlen;
-@@ -111,24 +111,38 @@ struct cprng_testvec {
- };
-
- struct drbg_testvec {
-- unsigned char *entropy;
-+ const unsigned char *entropy;
- size_t entropylen;
-- unsigned char *entpra;
-- unsigned char *entprb;
-+ const unsigned char *entpra;
-+ const unsigned char *entprb;
- size_t entprlen;
-- unsigned char *addtla;
-- unsigned char *addtlb;
-+ const unsigned char *addtla;
-+ const unsigned char *addtlb;
- size_t addtllen;
-- unsigned char *pers;
-+ const unsigned char *pers;
- size_t perslen;
-- unsigned char *expected;
-+ const unsigned char *expected;
- size_t expectedlen;
- };
-
-+struct tls_testvec {
-+ char *key; /* wrapped keys for encryption and authentication */
-+ char *iv; /* initialization vector */
-+ char *input; /* input data */
-+ char *assoc; /* associated data: seq num, type, version, input len */
-+ char *result; /* result data */
-+ unsigned char fail; /* the test failure is expected */
-+ unsigned char novrfy; /* dec verification failure expected */
-+ unsigned char klen; /* key length */
-+ unsigned short ilen; /* input data length */
-+ unsigned short alen; /* associated data length */
-+ unsigned short rlen; /* result length */
-+};
-+
- struct akcipher_testvec {
-- unsigned char *key;
-- unsigned char *m;
-- unsigned char *c;
-+ const unsigned char *key;
-+ const unsigned char *m;
-+ const unsigned char *c;
- unsigned int key_len;
- unsigned int m_size;
- unsigned int c_size;
-@@ -136,27 +150,227 @@ struct akcipher_testvec {
- };
-
- struct kpp_testvec {
-- unsigned char *secret;
-- unsigned char *b_public;
-- unsigned char *expected_a_public;
-- unsigned char *expected_ss;
-+ const unsigned char *secret;
-+ const unsigned char *b_public;
-+ const unsigned char *expected_a_public;
-+ const unsigned char *expected_ss;
- unsigned short secret_size;
- unsigned short b_public_size;
- unsigned short expected_a_public_size;
- unsigned short expected_ss_size;
- };
-
--static char zeroed_string[48];
-+static const char zeroed_string[48];
-
- /*
-- * RSA test vectors. Borrowed from openSSL.
-+ * TLS1.0 synthetic test vectors
- */
--#ifdef CONFIG_CRYPTO_FIPS
--#define RSA_TEST_VECTORS 2
-+static struct tls_testvec tls_enc_tv_template[] = {
-+ {
-+#ifdef __LITTLE_ENDIAN
-+ .key = "\x08\x00" /* rta length */
-+ "\x01\x00" /* rta type */
-+#else
-+ .key = "\x00\x08" /* rta length */
-+ "\x00\x01" /* rta type */
-+#endif
-+ "\x00\x00\x00\x10" /* enc key length */
-+ "authenticationkey20benckeyis16_bytes",
-+ .klen = 8 + 20 + 16,
-+ .iv = "iv0123456789abcd",
-+ .input = "Single block msg",
-+ .ilen = 16,
-+ .assoc = "\x00\x01\x02\x03\x04\x05\x06\x07"
-+ "\x00\x03\x01\x00\x10",
-+ .alen = 13,
-+ .result = "\xd5\xac\xb\xd2\xac\xad\x3f\xb1"
-+ "\x59\x79\x1e\x91\x5f\x52\x14\x9c"
-+ "\xc0\x75\xd8\x4c\x97\x0f\x07\x73"
-+ "\xdc\x89\x47\x49\x49\xcb\x30\x6b"
-+ "\x1b\x45\x23\xa1\xd0\x51\xcf\x02"
-+ "\x2e\xa8\x5d\xa0\xfe\xca\x82\x61",
-+ .rlen = 16 + 20 + 12,
-+ }, {
-+#ifdef __LITTLE_ENDIAN
-+ .key = "\x08\x00" /* rta length */
-+ "\x01\x00" /* rta type */
-+#else
-+ .key = "\x00\x08" /* rta length */
-+ "\x00\x01" /* rta type */
-+#endif
-+ "\x00\x00\x00\x10" /* enc key length */
-+ "authenticationkey20benckeyis16_bytes",
-+ .klen = 8 + 20 + 16,
-+ .iv = "iv0123456789abcd",
-+ .input = "",
-+ .ilen = 0,
-+ .assoc = "\x00\x01\x02\x03\x04\x05\x06\x07"
-+ "\x00\x03\x01\x00\x00",
-+ .alen = 13,
-+ .result = "\x58\x2a\x11\xc\x86\x8e\x4b\x67"
-+ "\x2d\x16\x26\x1a\xac\x4b\xe2\x1a"
-+ "\xe9\x6a\xcc\x4d\x6f\x79\x8a\x45"
-+ "\x1f\x4e\x27\xf2\xa7\x59\xb4\x5a",
-+ .rlen = 20 + 12,
-+ }, {
-+#ifdef __LITTLE_ENDIAN
-+ .key = "\x08\x00" /* rta length */
-+ "\x01\x00" /* rta type */
-+#else
-+ .key = "\x00\x08" /* rta length */
-+ "\x00\x01" /* rta type */
-+#endif
-+ "\x00\x00\x00\x10" /* enc key length */
-+ "authenticationkey20benckeyis16_bytes",
-+ .klen = 8 + 20 + 16,
-+ .iv = "iv0123456789abcd",
-+ .input = "285 bytes plaintext285 bytes plaintext285 bytes"
-+ " plaintext285 bytes plaintext285 bytes plaintext285"
-+ " bytes plaintext285 bytes plaintext285 bytes"
-+ " plaintext285 bytes plaintext285 bytes plaintext285"
-+ " bytes plaintext285 bytes plaintext285 bytes"
-+ " plaintext285 bytes plaintext285 bytes plaintext285"
-+ " bytes plaintext285 bytes plaintext",
-+ .ilen = 285,
-+ .assoc = "\x00\x01\x02\x03\x04\x05\x06\x07"
-+ "\x00\x03\x01\x01\x1d",
-+ .alen = 13,
-+ .result = "\x80\x23\x82\x44\x14\x2a\x1d\x94\xc\xc2\x1d\xd"
-+ "\x3a\x32\x89\x4c\x57\x30\xa8\x89\x76\x46\xcc\x90"
-+ "\x1d\x88\xb8\xa6\x1a\x58\xe\x2d\xeb\x2c\xc7\x3a"
-+ "\x52\x4e\xdb\xb3\x1e\x83\x11\xf5\x3c\xce\x6e\x94"
-+ "\xd3\x26\x6a\x9a\xd\xbd\xc7\x98\xb9\xb3\x3a\x51"
-+ "\x1e\x4\x84\x8a\x8f\x54\x9a\x51\x69\x9c\xce\x31"
-+ "\x8d\x5d\x8b\xee\x5f\x70\xc\xc9\xb8\x50\x54\xf8"
-+ "\xb2\x4a\x7a\xcd\xeb\x7a\x82\x81\xc6\x41\xc8\x50"
-+ "\x91\x8d\xc8\xed\xcd\x40\x8f\x55\xd1\xec\xc9\xac"
-+ "\x15\x18\xf9\x20\xa0\xed\x18\xa1\xe3\x56\xe3\x14"
-+ "\xe5\xe8\x66\x63\x20\xed\xe4\x62\x9d\xa3\xa4\x1d"
-+ "\x81\x89\x18\xf2\x36\xae\xc8\x8a\x2b\xbc\xc3\xb8"
-+ "\x80\xf\x97\x21\x36\x39\x8\x84\x23\x18\x9e\x9c"
-+ "\x72\x32\x75\x2d\x2e\xf9\x60\xb\xe8\xcc\xd9\x74"
-+ "\x4\x1b\x8e\x99\xc1\x94\xee\xd0\xac\x4e\xfc\x7e"
-+ "\xf1\x96\xb3\xe7\x14\xb8\xf2\xc\x25\x97\x82\x6b"
-+ "\xbd\x0\x65\xab\x5c\xe3\x16\xfb\x68\xef\xea\x9d"
-+ "\xff\x44\x1d\x2a\x44\xf5\xc8\x56\x77\xb7\xbf\x13"
-+ "\xc8\x54\xdb\x92\xfe\x16\x4c\xbe\x18\xe9\xb\x8d"
-+ "\xb\xd4\x43\x58\x43\xaa\xf4\x3\x80\x97\x62\xd5"
-+ "\xdf\x3c\x28\xaa\xee\x48\x4b\x55\x41\x1b\x31\x2"
-+ "\xbe\xa0\x1c\xbd\xb7\x22\x2a\xe5\x53\x72\x73\x20"
-+ "\x44\x4f\xe6\x1\x2b\x34\x33\x11\x7d\xfb\x10\xc1"
-+ "\x66\x7c\xa6\xf4\x48\x36\x5e\x2\xda\x41\x4b\x3e"
-+ "\xe7\x80\x17\x17\xce\xf1\x3e\x6a\x8e\x26\xf3\xb7"
-+ "\x2b\x85\xd\x31\x8d\xba\x6c\x22\xb4\x28\x55\x7e"
-+ "\x2a\x9e\x26\xf1\x3d\x21\xac\x65",
-+ .rlen = 285 + 20 + 15,
-+ }
-+};
-+
-+static struct tls_testvec tls_dec_tv_template[] = {
-+ {
-+#ifdef __LITTLE_ENDIAN
-+ .key = "\x08\x00" /* rta length */
-+ "\x01\x00" /* rta type */
-+#else
-+ .key = "\x00\x08" /* rta length */
-+ "\x00\x01" /* rta type */
-+#endif
-+ "\x00\x00\x00\x10" /* enc key length */
-+ "authenticationkey20benckeyis16_bytes",
-+ .klen = 8 + 20 + 16,
-+ .iv = "iv0123456789abcd",
-+ .input = "\xd5\xac\xb\xd2\xac\xad\x3f\xb1"
-+ "\x59\x79\x1e\x91\x5f\x52\x14\x9c"
-+ "\xc0\x75\xd8\x4c\x97\x0f\x07\x73"
-+ "\xdc\x89\x47\x49\x49\xcb\x30\x6b"
-+ "\x1b\x45\x23\xa1\xd0\x51\xcf\x02"
-+ "\x2e\xa8\x5d\xa0\xfe\xca\x82\x61",
-+ .ilen = 16 + 20 + 12,
-+ .assoc = "\x00\x01\x02\x03\x04\x05\x06\x07"
-+ "\x00\x03\x01\x00\x30",
-+ .alen = 13,
-+ .result = "Single block msg",
-+ .rlen = 16,
-+ }, {
-+#ifdef __LITTLE_ENDIAN
-+ .key = "\x08\x00" /* rta length */
-+ "\x01\x00" /* rta type */
- #else
--#define RSA_TEST_VECTORS 5
-+ .key = "\x00\x08" /* rta length */
-+ "\x00\x01" /* rta type */
- #endif
--static struct akcipher_testvec rsa_tv_template[] = {
-+ "\x00\x00\x00\x10" /* enc key length */
-+ "authenticationkey20benckeyis16_bytes",
-+ .klen = 8 + 20 + 16,
-+ .iv = "iv0123456789abcd",
-+ .input = "\x58\x2a\x11\xc\x86\x8e\x4b\x67"
-+ "\x2d\x16\x26\x1a\xac\x4b\xe2\x1a"
-+ "\xe9\x6a\xcc\x4d\x6f\x79\x8a\x45"
-+ "\x1f\x4e\x27\xf2\xa7\x59\xb4\x5a",
-+ .ilen = 20 + 12,
-+ .assoc = "\x00\x01\x02\x03\x04\x05\x06\x07"
-+ "\x00\x03\x01\x00\x20",
-+ .alen = 13,
-+ .result = "",
-+ .rlen = 0,
-+ }, {
-+#ifdef __LITTLE_ENDIAN
-+ .key = "\x08\x00" /* rta length */
-+ "\x01\x00" /* rta type */
-+#else
-+ .key = "\x00\x08" /* rta length */
-+ "\x00\x01" /* rta type */
-+#endif
-+ "\x00\x00\x00\x10" /* enc key length */
-+ "authenticationkey20benckeyis16_bytes",
-+ .klen = 8 + 20 + 16,
-+ .iv = "iv0123456789abcd",
-+ .input = "\x80\x23\x82\x44\x14\x2a\x1d\x94\xc\xc2\x1d\xd"
-+ "\x3a\x32\x89\x4c\x57\x30\xa8\x89\x76\x46\xcc\x90"
-+ "\x1d\x88\xb8\xa6\x1a\x58\xe\x2d\xeb\x2c\xc7\x3a"
-+ "\x52\x4e\xdb\xb3\x1e\x83\x11\xf5\x3c\xce\x6e\x94"
-+ "\xd3\x26\x6a\x9a\xd\xbd\xc7\x98\xb9\xb3\x3a\x51"
-+ "\x1e\x4\x84\x8a\x8f\x54\x9a\x51\x69\x9c\xce\x31"
-+ "\x8d\x5d\x8b\xee\x5f\x70\xc\xc9\xb8\x50\x54\xf8"
-+ "\xb2\x4a\x7a\xcd\xeb\x7a\x82\x81\xc6\x41\xc8\x50"
-+ "\x91\x8d\xc8\xed\xcd\x40\x8f\x55\xd1\xec\xc9\xac"
-+ "\x15\x18\xf9\x20\xa0\xed\x18\xa1\xe3\x56\xe3\x14"
-+ "\xe5\xe8\x66\x63\x20\xed\xe4\x62\x9d\xa3\xa4\x1d"
-+ "\x81\x89\x18\xf2\x36\xae\xc8\x8a\x2b\xbc\xc3\xb8"
-+ "\x80\xf\x97\x21\x36\x39\x8\x84\x23\x18\x9e\x9c"
-+ "\x72\x32\x75\x2d\x2e\xf9\x60\xb\xe8\xcc\xd9\x74"
-+ "\x4\x1b\x8e\x99\xc1\x94\xee\xd0\xac\x4e\xfc\x7e"
-+ "\xf1\x96\xb3\xe7\x14\xb8\xf2\xc\x25\x97\x82\x6b"
-+ "\xbd\x0\x65\xab\x5c\xe3\x16\xfb\x68\xef\xea\x9d"
-+ "\xff\x44\x1d\x2a\x44\xf5\xc8\x56\x77\xb7\xbf\x13"
-+ "\xc8\x54\xdb\x92\xfe\x16\x4c\xbe\x18\xe9\xb\x8d"
-+ "\xb\xd4\x43\x58\x43\xaa\xf4\x3\x80\x97\x62\xd5"
-+ "\xdf\x3c\x28\xaa\xee\x48\x4b\x55\x41\x1b\x31\x2"
-+ "\xbe\xa0\x1c\xbd\xb7\x22\x2a\xe5\x53\x72\x73\x20"
-+ "\x44\x4f\xe6\x1\x2b\x34\x33\x11\x7d\xfb\x10\xc1"
-+ "\x66\x7c\xa6\xf4\x48\x36\x5e\x2\xda\x41\x4b\x3e"
-+ "\xe7\x80\x17\x17\xce\xf1\x3e\x6a\x8e\x26\xf3\xb7"
-+ "\x2b\x85\xd\x31\x8d\xba\x6c\x22\xb4\x28\x55\x7e"
-+ "\x2a\x9e\x26\xf1\x3d\x21\xac\x65",
-+
-+ .ilen = 285 + 20 + 15,
-+ .assoc = "\x00\x01\x02\x03\x04\x05\x06\x07"
-+ "\x00\x03\x01\x01\x40",
-+ .alen = 13,
-+ .result = "285 bytes plaintext285 bytes plaintext285 bytes"
-+ " plaintext285 bytes plaintext285 bytes plaintext285"
-+ " bytes plaintext285 bytes plaintext285 bytes"
-+ " plaintext285 bytes plaintext285 bytes plaintext285"
-+ " bytes plaintext285 bytes plaintext285 bytes"
-+ " plaintext285 bytes plaintext285 bytes plaintext",
-+ .rlen = 285,
-+ }
-+};
-+
-+/*
-+ * RSA test vectors. Borrowed from openSSL.
-+ */
-+static const struct akcipher_testvec rsa_tv_template[] = {
- {
- #ifndef CONFIG_CRYPTO_FIPS
- .key =
-@@ -340,6 +554,7 @@ static struct akcipher_testvec rsa_tv_te
- .m_size = 8,
- .c_size = 256,
- .public_key_vec = true,
-+#ifndef CONFIG_CRYPTO_FIPS
- }, {
- .key =
- "\x30\x82\x09\x29" /* sequence of 2345 bytes */
-@@ -538,12 +753,11 @@ static struct akcipher_testvec rsa_tv_te
- .key_len = 2349,
- .m_size = 8,
- .c_size = 512,
-+#endif
- }
- };
-
--#define DH_TEST_VECTORS 2
--
--struct kpp_testvec dh_tv_template[] = {
-+static const struct kpp_testvec dh_tv_template[] = {
- {
- .secret =
- #ifdef __LITTLE_ENDIAN
-@@ -760,12 +974,7 @@ struct kpp_testvec dh_tv_template[] = {
- }
- };
-
--#ifdef CONFIG_CRYPTO_FIPS
--#define ECDH_TEST_VECTORS 1
--#else
--#define ECDH_TEST_VECTORS 2
--#endif
--struct kpp_testvec ecdh_tv_template[] = {
-+static const struct kpp_testvec ecdh_tv_template[] = {
- {
- #ifndef CONFIG_CRYPTO_FIPS
- .secret =
-@@ -856,9 +1065,7 @@ struct kpp_testvec ecdh_tv_template[] =
- /*
- * MD4 test vectors from RFC1320
- */
--#define MD4_TEST_VECTORS 7
--
--static struct hash_testvec md4_tv_template [] = {
-+static const struct hash_testvec md4_tv_template[] = {
- {
- .plaintext = "",
- .digest = "\x31\xd6\xcf\xe0\xd1\x6a\xe9\x31"
-@@ -899,8 +1106,7 @@ static struct hash_testvec md4_tv_templa
- },
- };
-
--#define SHA3_224_TEST_VECTORS 3
--static struct hash_testvec sha3_224_tv_template[] = {
-+static const struct hash_testvec sha3_224_tv_template[] = {
- {
- .plaintext = "",
- .digest = "\x6b\x4e\x03\x42\x36\x67\xdb\xb7"
-@@ -925,8 +1131,7 @@ static struct hash_testvec sha3_224_tv_t
- },
- };
-
--#define SHA3_256_TEST_VECTORS 3
--static struct hash_testvec sha3_256_tv_template[] = {
-+static const struct hash_testvec sha3_256_tv_template[] = {
- {
- .plaintext = "",
- .digest = "\xa7\xff\xc6\xf8\xbf\x1e\xd7\x66"
-@@ -952,8 +1157,7 @@ static struct hash_testvec sha3_256_tv_t
- };
-
-
--#define SHA3_384_TEST_VECTORS 3
--static struct hash_testvec sha3_384_tv_template[] = {
-+static const struct hash_testvec sha3_384_tv_template[] = {
- {
- .plaintext = "",
- .digest = "\x0c\x63\xa7\x5b\x84\x5e\x4f\x7d"
-@@ -985,8 +1189,7 @@ static struct hash_testvec sha3_384_tv_t
- };
-
-
--#define SHA3_512_TEST_VECTORS 3
--static struct hash_testvec sha3_512_tv_template[] = {
-+static const struct hash_testvec sha3_512_tv_template[] = {
- {
- .plaintext = "",
- .digest = "\xa6\x9f\x73\xcc\xa2\x3a\x9a\xc5"
-@@ -1027,9 +1230,7 @@ static struct hash_testvec sha3_512_tv_t
- /*
- * MD5 test vectors from RFC1321
- */
--#define MD5_TEST_VECTORS 7
--
--static struct hash_testvec md5_tv_template[] = {
-+static const struct hash_testvec md5_tv_template[] = {
- {
- .digest = "\xd4\x1d\x8c\xd9\x8f\x00\xb2\x04"
- "\xe9\x80\x09\x98\xec\xf8\x42\x7e",
-@@ -1073,9 +1274,7 @@ static struct hash_testvec md5_tv_templa
- /*
- * RIPEMD-128 test vectors from ISO/IEC 10118-3:2004(E)
- */
--#define RMD128_TEST_VECTORS 10
--
--static struct hash_testvec rmd128_tv_template[] = {
-+static const struct hash_testvec rmd128_tv_template[] = {
- {
- .digest = "\xcd\xf2\x62\x13\xa1\x50\xdc\x3e"
- "\xcb\x61\x0f\x18\xf6\xb3\x8b\x46",
-@@ -1137,9 +1336,7 @@ static struct hash_testvec rmd128_tv_tem
- /*
- * RIPEMD-160 test vectors from ISO/IEC 10118-3:2004(E)
- */
--#define RMD160_TEST_VECTORS 10
--
--static struct hash_testvec rmd160_tv_template[] = {
-+static const struct hash_testvec rmd160_tv_template[] = {
- {
- .digest = "\x9c\x11\x85\xa5\xc5\xe9\xfc\x54\x61\x28"
- "\x08\x97\x7e\xe8\xf5\x48\xb2\x25\x8d\x31",
-@@ -1201,9 +1398,7 @@ static struct hash_testvec rmd160_tv_tem
- /*
- * RIPEMD-256 test vectors
- */
--#define RMD256_TEST_VECTORS 8
--
--static struct hash_testvec rmd256_tv_template[] = {
-+static const struct hash_testvec rmd256_tv_template[] = {
- {
- .digest = "\x02\xba\x4c\x4e\x5f\x8e\xcd\x18"
- "\x77\xfc\x52\xd6\x4d\x30\xe3\x7a"
-@@ -1269,9 +1464,7 @@ static struct hash_testvec rmd256_tv_tem
- /*
- * RIPEMD-320 test vectors
- */
--#define RMD320_TEST_VECTORS 8
--
--static struct hash_testvec rmd320_tv_template[] = {
-+static const struct hash_testvec rmd320_tv_template[] = {
- {
- .digest = "\x22\xd6\x5d\x56\x61\x53\x6c\xdc\x75\xc1"
- "\xfd\xf5\xc6\xde\x7b\x41\xb9\xf2\x73\x25"
-@@ -1334,36 +1527,49 @@ static struct hash_testvec rmd320_tv_tem
- }
- };
-
--#define CRCT10DIF_TEST_VECTORS 3
--static struct hash_testvec crct10dif_tv_template[] = {
-+static const struct hash_testvec crct10dif_tv_template[] = {
- {
-- .plaintext = "abc",
-- .psize = 3,
--#ifdef __LITTLE_ENDIAN
-- .digest = "\x3b\x44",
--#else
-- .digest = "\x44\x3b",
--#endif
-- }, {
-- .plaintext = "1234567890123456789012345678901234567890"
-- "123456789012345678901234567890123456789",
-- .psize = 79,
--#ifdef __LITTLE_ENDIAN
-- .digest = "\x70\x4b",
--#else
-- .digest = "\x4b\x70",
--#endif
-- }, {
-- .plaintext =
-- "abcddddddddddddddddddddddddddddddddddddddddddddddddddddd",
-- .psize = 56,
--#ifdef __LITTLE_ENDIAN
-- .digest = "\xe3\x9c",
--#else
-- .digest = "\x9c\xe3",
--#endif
-- .np = 2,
-- .tap = { 28, 28 }
-+ .plaintext = "abc",
-+ .psize = 3,
-+ .digest = (u8 *)(u16 []){ 0x443b },
-+ }, {
-+ .plaintext = "1234567890123456789012345678901234567890"
-+ "123456789012345678901234567890123456789",
-+ .psize = 79,
-+ .digest = (u8 *)(u16 []){ 0x4b70 },
-+ .np = 2,
-+ .tap = { 63, 16 },
-+ }, {
-+ .plaintext = "abcdddddddddddddddddddddddddddddddddddddddd"
-+ "ddddddddddddd",
-+ .psize = 56,
-+ .digest = (u8 *)(u16 []){ 0x9ce3 },
-+ .np = 8,
-+ .tap = { 1, 2, 28, 7, 6, 5, 4, 3 },
-+ }, {
-+ .plaintext = "1234567890123456789012345678901234567890"
-+ "1234567890123456789012345678901234567890"
-+ "1234567890123456789012345678901234567890"
-+ "1234567890123456789012345678901234567890"
-+ "1234567890123456789012345678901234567890"
-+ "1234567890123456789012345678901234567890"
-+ "1234567890123456789012345678901234567890"
-+ "123456789012345678901234567890123456789",
-+ .psize = 319,
-+ .digest = (u8 *)(u16 []){ 0x44c6 },
-+ }, {
-+ .plaintext = "1234567890123456789012345678901234567890"
-+ "1234567890123456789012345678901234567890"
-+ "1234567890123456789012345678901234567890"
-+ "1234567890123456789012345678901234567890"
-+ "1234567890123456789012345678901234567890"
-+ "1234567890123456789012345678901234567890"
-+ "1234567890123456789012345678901234567890"
-+ "123456789012345678901234567890123456789",
-+ .psize = 319,
-+ .digest = (u8 *)(u16 []){ 0x44c6 },
-+ .np = 4,
-+ .tap = { 1, 255, 57, 6 },
- }
- };
-
-@@ -1371,9 +1577,7 @@ static struct hash_testvec crct10dif_tv_
- * SHA1 test vectors from from FIPS PUB 180-1
- * Long vector from CAVS 5.0
- */
--#define SHA1_TEST_VECTORS 6
--
--static struct hash_testvec sha1_tv_template[] = {
-+static const struct hash_testvec sha1_tv_template[] = {
- {
- .plaintext = "",
- .psize = 0,
-@@ -1563,9 +1767,7 @@ static struct hash_testvec sha1_tv_templ
- /*
- * SHA224 test vectors from from FIPS PUB 180-2
- */
--#define SHA224_TEST_VECTORS 5
--
--static struct hash_testvec sha224_tv_template[] = {
-+static const struct hash_testvec sha224_tv_template[] = {
- {
- .plaintext = "",
- .psize = 0,
-@@ -1737,9 +1939,7 @@ static struct hash_testvec sha224_tv_tem
- /*
- * SHA256 test vectors from from NIST
- */
--#define SHA256_TEST_VECTORS 5
--
--static struct hash_testvec sha256_tv_template[] = {
-+static const struct hash_testvec sha256_tv_template[] = {
- {
- .plaintext = "",
- .psize = 0,
-@@ -1910,9 +2110,7 @@ static struct hash_testvec sha256_tv_tem
- /*
- * SHA384 test vectors from from NIST and kerneli
- */
--#define SHA384_TEST_VECTORS 6
--
--static struct hash_testvec sha384_tv_template[] = {
-+static const struct hash_testvec sha384_tv_template[] = {
- {
- .plaintext = "",
- .psize = 0,
-@@ -2104,9 +2302,7 @@ static struct hash_testvec sha384_tv_tem
- /*
- * SHA512 test vectors from from NIST and kerneli
- */
--#define SHA512_TEST_VECTORS 6
--
--static struct hash_testvec sha512_tv_template[] = {
-+static const struct hash_testvec sha512_tv_template[] = {
- {
- .plaintext = "",
- .psize = 0,
-@@ -2313,9 +2509,7 @@ static struct hash_testvec sha512_tv_tem
- * by Vincent Rijmen and Paulo S. L. M. Barreto as part of the NESSIE
- * submission
- */
--#define WP512_TEST_VECTORS 8
--
--static struct hash_testvec wp512_tv_template[] = {
-+static const struct hash_testvec wp512_tv_template[] = {
- {
- .plaintext = "",
- .psize = 0,
-@@ -2411,9 +2605,7 @@ static struct hash_testvec wp512_tv_temp
- },
- };
-
--#define WP384_TEST_VECTORS 8
--
--static struct hash_testvec wp384_tv_template[] = {
-+static const struct hash_testvec wp384_tv_template[] = {
- {
- .plaintext = "",
- .psize = 0,
-@@ -2493,9 +2685,7 @@ static struct hash_testvec wp384_tv_temp
- },
- };
-
--#define WP256_TEST_VECTORS 8
--
--static struct hash_testvec wp256_tv_template[] = {
-+static const struct hash_testvec wp256_tv_template[] = {
- {
- .plaintext = "",
- .psize = 0,
-@@ -2562,9 +2752,7 @@ static struct hash_testvec wp256_tv_temp
- /*
- * TIGER test vectors from Tiger website
- */
--#define TGR192_TEST_VECTORS 6
--
--static struct hash_testvec tgr192_tv_template[] = {
-+static const struct hash_testvec tgr192_tv_template[] = {
- {
- .plaintext = "",
- .psize = 0,
-@@ -2607,9 +2795,7 @@ static struct hash_testvec tgr192_tv_tem
- },
- };
-
--#define TGR160_TEST_VECTORS 6
--
--static struct hash_testvec tgr160_tv_template[] = {
-+static const struct hash_testvec tgr160_tv_template[] = {
- {
- .plaintext = "",
- .psize = 0,
-@@ -2652,9 +2838,7 @@ static struct hash_testvec tgr160_tv_tem
- },
- };
-
--#define TGR128_TEST_VECTORS 6
--
--static struct hash_testvec tgr128_tv_template[] = {
-+static const struct hash_testvec tgr128_tv_template[] = {
- {
- .plaintext = "",
- .psize = 0,
-@@ -2691,9 +2875,7 @@ static struct hash_testvec tgr128_tv_tem
- },
- };
-
--#define GHASH_TEST_VECTORS 6
--
--static struct hash_testvec ghash_tv_template[] =
-+static const struct hash_testvec ghash_tv_template[] =
- {
- {
- .key = "\xdf\xa6\xbf\x4d\xed\x81\xdb\x03"
-@@ -2808,9 +2990,7 @@ static struct hash_testvec ghash_tv_temp
- * HMAC-MD5 test vectors from RFC2202
- * (These need to be fixed to not use strlen).
- */
--#define HMAC_MD5_TEST_VECTORS 7
--
--static struct hash_testvec hmac_md5_tv_template[] =
-+static const struct hash_testvec hmac_md5_tv_template[] =
- {
- {
- .key = "\x0b\x0b\x0b\x0b\x0b\x0b\x0b\x0b\x0b\x0b\x0b\x0b\x0b\x0b\x0b\x0b",
-@@ -2890,9 +3070,7 @@ static struct hash_testvec hmac_md5_tv_t
- /*
- * HMAC-RIPEMD128 test vectors from RFC2286
- */
--#define HMAC_RMD128_TEST_VECTORS 7
--
--static struct hash_testvec hmac_rmd128_tv_template[] = {
-+static const struct hash_testvec hmac_rmd128_tv_template[] = {
- {
- .key = "\x0b\x0b\x0b\x0b\x0b\x0b\x0b\x0b\x0b\x0b\x0b\x0b\x0b\x0b\x0b\x0b",
- .ksize = 16,
-@@ -2971,9 +3149,7 @@ static struct hash_testvec hmac_rmd128_t
- /*
- * HMAC-RIPEMD160 test vectors from RFC2286
- */
--#define HMAC_RMD160_TEST_VECTORS 7
--
--static struct hash_testvec hmac_rmd160_tv_template[] = {
-+static const struct hash_testvec hmac_rmd160_tv_template[] = {
- {
- .key = "\x0b\x0b\x0b\x0b\x0b\x0b\x0b\x0b\x0b\x0b\x0b\x0b\x0b\x0b\x0b\x0b\x0b\x0b\x0b\x0b",
- .ksize = 20,
-@@ -3052,9 +3228,7 @@ static struct hash_testvec hmac_rmd160_t
- /*
- * HMAC-SHA1 test vectors from RFC2202
- */
--#define HMAC_SHA1_TEST_VECTORS 7
--
--static struct hash_testvec hmac_sha1_tv_template[] = {
-+static const struct hash_testvec hmac_sha1_tv_template[] = {
- {
- .key = "\x0b\x0b\x0b\x0b\x0b\x0b\x0b\x0b\x0b\x0b\x0b\x0b\x0b\x0b\x0b\x0b\x0b\x0b\x0b\x0b",
- .ksize = 20,
-@@ -3135,9 +3309,7 @@ static struct hash_testvec hmac_sha1_tv_
- /*
- * SHA224 HMAC test vectors from RFC4231
- */
--#define HMAC_SHA224_TEST_VECTORS 4
--
--static struct hash_testvec hmac_sha224_tv_template[] = {
-+static const struct hash_testvec hmac_sha224_tv_template[] = {
- {
- .key = "\x0b\x0b\x0b\x0b\x0b\x0b\x0b\x0b"
- "\x0b\x0b\x0b\x0b\x0b\x0b\x0b\x0b"
-@@ -3250,9 +3422,7 @@ static struct hash_testvec hmac_sha224_t
- * HMAC-SHA256 test vectors from
- * draft-ietf-ipsec-ciph-sha-256-01.txt
- */
--#define HMAC_SHA256_TEST_VECTORS 10
--
--static struct hash_testvec hmac_sha256_tv_template[] = {
-+static const struct hash_testvec hmac_sha256_tv_template[] = {
- {
- .key = "\x01\x02\x03\x04\x05\x06\x07\x08"
- "\x09\x0a\x0b\x0c\x0d\x0e\x0f\x10"
-@@ -3387,9 +3557,7 @@ static struct hash_testvec hmac_sha256_t
- },
- };
-
--#define CMAC_AES_TEST_VECTORS 6
--
--static struct hash_testvec aes_cmac128_tv_template[] = {
-+static const struct hash_testvec aes_cmac128_tv_template[] = {
- { /* From NIST Special Publication 800-38B, AES-128 */
- .key = "\x2b\x7e\x15\x16\x28\xae\xd2\xa6"
- "\xab\xf7\x15\x88\x09\xcf\x4f\x3c",
-@@ -3464,9 +3632,67 @@ static struct hash_testvec aes_cmac128_t
- }
- };
-
--#define CMAC_DES3_EDE_TEST_VECTORS 4
-+static const struct hash_testvec aes_cbcmac_tv_template[] = {
-+ {
-+ .key = "\x2b\x7e\x15\x16\x28\xae\xd2\xa6"
-+ "\xab\xf7\x15\x88\x09\xcf\x4f\x3c",
-+ .plaintext = "\x6b\xc1\xbe\xe2\x2e\x40\x9f\x96"
-+ "\xe9\x3d\x7e\x11\x73\x93\x17\x2a",
-+ .digest = "\x3a\xd7\x7b\xb4\x0d\x7a\x36\x60"
-+ "\xa8\x9e\xca\xf3\x24\x66\xef\x97",
-+ .psize = 16,
-+ .ksize = 16,
-+ }, {
-+ .key = "\x2b\x7e\x15\x16\x28\xae\xd2\xa6"
-+ "\xab\xf7\x15\x88\x09\xcf\x4f\x3c",
-+ .plaintext = "\x6b\xc1\xbe\xe2\x2e\x40\x9f\x96"
-+ "\xe9\x3d\x7e\x11\x73\x93\x17\x2a"
-+ "\xae\x2d\x8a\x57\x1e\x03\xac\x9c"
-+ "\x9e\xb7\x6f\xac\x45\xaf\x8e\x51"
-+ "\x30",
-+ .digest = "\x9d\x0d\xd0\x63\xfb\xcb\x24\x43"
-+ "\xf8\xf2\x76\x03\xac\x39\xb0\x9d",
-+ .psize = 33,
-+ .ksize = 16,
-+ .np = 2,
-+ .tap = { 7, 26 },
-+ }, {
-+ .key = "\x2b\x7e\x15\x16\x28\xae\xd2\xa6"
-+ "\xab\xf7\x15\x88\x09\xcf\x4f\x3c",
-+ .plaintext = "\x6b\xc1\xbe\xe2\x2e\x40\x9f\x96"
-+ "\xe9\x3d\x7e\x11\x73\x93\x17\x2a"
-+ "\xae\x2d\x8a\x57\x1e\x03\xac\x9c"
-+ "\x9e\xb7\x6f\xac\x45\xaf\x8e\x51"
-+ "\x30\xc8\x1c\x46\xa3\x5c\xe4\x11"
-+ "\xe5\xfb\xc1\x19\x1a\x0a\x52\xef"
-+ "\xf6\x9f\x24\x45\xdf\x4f\x9b\x17"
-+ "\xad\x2b\x41\x7b\xe6\x6c\x37",
-+ .digest = "\xc0\x71\x73\xb8\xa0\x2c\x11\x7c"
-+ "\xaf\xdc\xb2\xf8\x89\x32\xa3\x3a",
-+ .psize = 63,
-+ .ksize = 16,
-+ }, {
-+ .key = "\x60\x3d\xeb\x10\x15\xca\x71\xbe"
-+ "\x2b\x73\xae\xf0\x85\x7d\x77\x81"
-+ "\x1f\x35\x2c\x07\x3b\x61\x08\xd7"
-+ "\x2d\x98\x10\xa3\x09\x14\xdf\xf4",
-+ .plaintext = "\x6b\xc1\xbe\xe2\x2e\x40\x9f\x96"
-+ "\xe9\x3d\x7e\x11\x73\x93\x17\x2a"
-+ "\xae\x2d\x8a\x57\x1e\x03\xac\x9c"
-+ "\x9e\xb7\x6f\xac\x45\xaf\x8e\x51"
-+ "\x30\xc8\x1c\x46\xa3\x5c\xe4\x11"
-+ "\xe5\xfb\xc1\x19\x1a\x0a\x52\xef"
-+ "\xf6\x9f\x24\x45\xdf\x4f\x9b\x17"
-+ "\xad\x2b\x41\x7b\xe6\x6c\x37\x10"
-+ "\x1c",
-+ .digest = "\x6a\x4e\xdb\x21\x47\x51\xdf\x4f"
-+ "\xa8\x4d\x4c\x10\x3b\x72\x7d\xd6",
-+ .psize = 65,
-+ .ksize = 32,
-+ }
-+};
-
--static struct hash_testvec des3_ede_cmac64_tv_template[] = {
-+static const struct hash_testvec des3_ede_cmac64_tv_template[] = {
- /*
- * From NIST Special Publication 800-38B, Three Key TDEA
- * Corrected test vectors from:
-@@ -3512,9 +3738,7 @@ static struct hash_testvec des3_ede_cmac
- }
- };
-
--#define XCBC_AES_TEST_VECTORS 6
--
--static struct hash_testvec aes_xcbc128_tv_template[] = {
-+static const struct hash_testvec aes_xcbc128_tv_template[] = {
- {
- .key = "\x00\x01\x02\x03\x04\x05\x06\x07"
- "\x08\x09\x0a\x0b\x0c\x0d\x0e\x0f",
-@@ -3580,36 +3804,35 @@ static struct hash_testvec aes_xcbc128_t
- }
- };
-
--#define VMAC_AES_TEST_VECTORS 11
--static char vmac_string1[128] = {'\x01', '\x01', '\x01', '\x01',
-- '\x02', '\x03', '\x02', '\x02',
-- '\x02', '\x04', '\x01', '\x07',
-- '\x04', '\x01', '\x04', '\x03',};
--static char vmac_string2[128] = {'a', 'b', 'c',};
--static char vmac_string3[128] = {'a', 'b', 'c', 'a', 'b', 'c',
-- 'a', 'b', 'c', 'a', 'b', 'c',
-- 'a', 'b', 'c', 'a', 'b', 'c',
-- 'a', 'b', 'c', 'a', 'b', 'c',
-- 'a', 'b', 'c', 'a', 'b', 'c',
-- 'a', 'b', 'c', 'a', 'b', 'c',
-- 'a', 'b', 'c', 'a', 'b', 'c',
-- 'a', 'b', 'c', 'a', 'b', 'c',
-- };
--
--static char vmac_string4[17] = {'b', 'c', 'e', 'f',
-- 'i', 'j', 'l', 'm',
-- 'o', 'p', 'r', 's',
-- 't', 'u', 'w', 'x', 'z'};
--
--static char vmac_string5[127] = {'r', 'm', 'b', 't', 'c',
-- 'o', 'l', 'k', ']', '%',
-- '9', '2', '7', '!', 'A'};
--
--static char vmac_string6[129] = {'p', 't', '*', '7', 'l',
-- 'i', '!', '#', 'w', '0',
-- 'z', '/', '4', 'A', 'n'};
-+static const char vmac_string1[128] = {'\x01', '\x01', '\x01', '\x01',
-+ '\x02', '\x03', '\x02', '\x02',
-+ '\x02', '\x04', '\x01', '\x07',
-+ '\x04', '\x01', '\x04', '\x03',};
-+static const char vmac_string2[128] = {'a', 'b', 'c',};
-+static const char vmac_string3[128] = {'a', 'b', 'c', 'a', 'b', 'c',
-+ 'a', 'b', 'c', 'a', 'b', 'c',
-+ 'a', 'b', 'c', 'a', 'b', 'c',
-+ 'a', 'b', 'c', 'a', 'b', 'c',
-+ 'a', 'b', 'c', 'a', 'b', 'c',
-+ 'a', 'b', 'c', 'a', 'b', 'c',
-+ 'a', 'b', 'c', 'a', 'b', 'c',
-+ 'a', 'b', 'c', 'a', 'b', 'c',
-+ };
-+
-+static const char vmac_string4[17] = {'b', 'c', 'e', 'f',
-+ 'i', 'j', 'l', 'm',
-+ 'o', 'p', 'r', 's',
-+ 't', 'u', 'w', 'x', 'z'};
-+
-+static const char vmac_string5[127] = {'r', 'm', 'b', 't', 'c',
-+ 'o', 'l', 'k', ']', '%',
-+ '9', '2', '7', '!', 'A'};
-+
-+static const char vmac_string6[129] = {'p', 't', '*', '7', 'l',
-+ 'i', '!', '#', 'w', '0',
-+ 'z', '/', '4', 'A', 'n'};
-
--static struct hash_testvec aes_vmac128_tv_template[] = {
-+static const struct hash_testvec aes_vmac128_tv_template[] = {
- {
- .key = "\x00\x01\x02\x03\x04\x05\x06\x07"
- "\x08\x09\x0a\x0b\x0c\x0d\x0e\x0f",
-@@ -3687,9 +3910,7 @@ static struct hash_testvec aes_vmac128_t
- * SHA384 HMAC test vectors from RFC4231
- */
-
--#define HMAC_SHA384_TEST_VECTORS 4
--
--static struct hash_testvec hmac_sha384_tv_template[] = {
-+static const struct hash_testvec hmac_sha384_tv_template[] = {
- {
- .key = "\x0b\x0b\x0b\x0b\x0b\x0b\x0b\x0b"
- "\x0b\x0b\x0b\x0b\x0b\x0b\x0b\x0b"
-@@ -3787,9 +4008,7 @@ static struct hash_testvec hmac_sha384_t
- * SHA512 HMAC test vectors from RFC4231
- */
-
--#define HMAC_SHA512_TEST_VECTORS 4
--
--static struct hash_testvec hmac_sha512_tv_template[] = {
-+static const struct hash_testvec hmac_sha512_tv_template[] = {
- {
- .key = "\x0b\x0b\x0b\x0b\x0b\x0b\x0b\x0b"
- "\x0b\x0b\x0b\x0b\x0b\x0b\x0b\x0b"
-@@ -3894,9 +4113,7 @@ static struct hash_testvec hmac_sha512_t
- },
- };
-
--#define HMAC_SHA3_224_TEST_VECTORS 4
--
--static struct hash_testvec hmac_sha3_224_tv_template[] = {
-+static const struct hash_testvec hmac_sha3_224_tv_template[] = {
- {
- .key = "\x0b\x0b\x0b\x0b\x0b\x0b\x0b\x0b"
- "\x0b\x0b\x0b\x0b\x0b\x0b\x0b\x0b"
-@@ -3985,9 +4202,7 @@ static struct hash_testvec hmac_sha3_224
- },
- };
-
--#define HMAC_SHA3_256_TEST_VECTORS 4
--
--static struct hash_testvec hmac_sha3_256_tv_template[] = {
-+static const struct hash_testvec hmac_sha3_256_tv_template[] = {
- {
- .key = "\x0b\x0b\x0b\x0b\x0b\x0b\x0b\x0b"
- "\x0b\x0b\x0b\x0b\x0b\x0b\x0b\x0b"
-@@ -4076,9 +4291,7 @@ static struct hash_testvec hmac_sha3_256
- },
- };
-
--#define HMAC_SHA3_384_TEST_VECTORS 4
--
--static struct hash_testvec hmac_sha3_384_tv_template[] = {
-+static const struct hash_testvec hmac_sha3_384_tv_template[] = {
- {
- .key = "\x0b\x0b\x0b\x0b\x0b\x0b\x0b\x0b"
- "\x0b\x0b\x0b\x0b\x0b\x0b\x0b\x0b"
-@@ -4175,9 +4388,7 @@ static struct hash_testvec hmac_sha3_384
- },
- };
-
--#define HMAC_SHA3_512_TEST_VECTORS 4
--
--static struct hash_testvec hmac_sha3_512_tv_template[] = {
-+static const struct hash_testvec hmac_sha3_512_tv_template[] = {
- {
- .key = "\x0b\x0b\x0b\x0b\x0b\x0b\x0b\x0b"
- "\x0b\x0b\x0b\x0b\x0b\x0b\x0b\x0b"
-@@ -4286,9 +4497,7 @@ static struct hash_testvec hmac_sha3_512
- * Poly1305 test vectors from RFC7539 A.3.
- */
-
--#define POLY1305_TEST_VECTORS 11
--
--static struct hash_testvec poly1305_tv_template[] = {
-+static const struct hash_testvec poly1305_tv_template[] = {
- { /* Test Vector #1 */
- .plaintext = "\x00\x00\x00\x00\x00\x00\x00\x00"
- "\x00\x00\x00\x00\x00\x00\x00\x00"
-@@ -4533,20 +4742,7 @@ static struct hash_testvec poly1305_tv_t
- /*
- * DES test vectors.
- */
--#define DES_ENC_TEST_VECTORS 11
--#define DES_DEC_TEST_VECTORS 5
--#define DES_CBC_ENC_TEST_VECTORS 6
--#define DES_CBC_DEC_TEST_VECTORS 5
--#define DES_CTR_ENC_TEST_VECTORS 2
--#define DES_CTR_DEC_TEST_VECTORS 2
--#define DES3_EDE_ENC_TEST_VECTORS 4
--#define DES3_EDE_DEC_TEST_VECTORS 4
--#define DES3_EDE_CBC_ENC_TEST_VECTORS 2
--#define DES3_EDE_CBC_DEC_TEST_VECTORS 2
--#define DES3_EDE_CTR_ENC_TEST_VECTORS 2
--#define DES3_EDE_CTR_DEC_TEST_VECTORS 2
--
--static struct cipher_testvec des_enc_tv_template[] = {
-+static const struct cipher_testvec des_enc_tv_template[] = {
- { /* From Applied Cryptography */
- .key = "\x01\x23\x45\x67\x89\xab\xcd\xef",
- .klen = 8,
-@@ -4720,7 +4916,7 @@ static struct cipher_testvec des_enc_tv_
- },
- };
-
--static struct cipher_testvec des_dec_tv_template[] = {
-+static const struct cipher_testvec des_dec_tv_template[] = {
- { /* From Applied Cryptography */
- .key = "\x01\x23\x45\x67\x89\xab\xcd\xef",
- .klen = 8,
-@@ -4830,7 +5026,7 @@ static struct cipher_testvec des_dec_tv_
- },
- };
-
--static struct cipher_testvec des_cbc_enc_tv_template[] = {
-+static const struct cipher_testvec des_cbc_enc_tv_template[] = {
- { /* From OpenSSL */
- .key = "\x01\x23\x45\x67\x89\xab\xcd\xef",
- .klen = 8,
-@@ -4956,7 +5152,7 @@ static struct cipher_testvec des_cbc_enc
- },
- };
-
--static struct cipher_testvec des_cbc_dec_tv_template[] = {
-+static const struct cipher_testvec des_cbc_dec_tv_template[] = {
- { /* FIPS Pub 81 */
- .key = "\x01\x23\x45\x67\x89\xab\xcd\xef",
- .klen = 8,
-@@ -5065,7 +5261,7 @@ static struct cipher_testvec des_cbc_dec
- },
- };
-
--static struct cipher_testvec des_ctr_enc_tv_template[] = {
-+static const struct cipher_testvec des_ctr_enc_tv_template[] = {
- { /* Generated with Crypto++ */
- .key = "\xC9\x83\xA6\xC9\xEC\x0F\x32\x55",
- .klen = 8,
-@@ -5211,7 +5407,7 @@ static struct cipher_testvec des_ctr_enc
- },
- };
-
--static struct cipher_testvec des_ctr_dec_tv_template[] = {
-+static const struct cipher_testvec des_ctr_dec_tv_template[] = {
- { /* Generated with Crypto++ */
- .key = "\xC9\x83\xA6\xC9\xEC\x0F\x32\x55",
- .klen = 8,
-@@ -5357,7 +5553,7 @@ static struct cipher_testvec des_ctr_dec
- },
- };
-
--static struct cipher_testvec des3_ede_enc_tv_template[] = {
-+static const struct cipher_testvec des3_ede_enc_tv_template[] = {
- { /* These are from openssl */
- .key = "\x01\x23\x45\x67\x89\xab\xcd\xef"
- "\x55\x55\x55\x55\x55\x55\x55\x55"
-@@ -5522,7 +5718,7 @@ static struct cipher_testvec des3_ede_en
- },
- };
-
--static struct cipher_testvec des3_ede_dec_tv_template[] = {
-+static const struct cipher_testvec des3_ede_dec_tv_template[] = {
- { /* These are from openssl */
- .key = "\x01\x23\x45\x67\x89\xab\xcd\xef"
- "\x55\x55\x55\x55\x55\x55\x55\x55"
-@@ -5687,7 +5883,7 @@ static struct cipher_testvec des3_ede_de
- },
- };
-
--static struct cipher_testvec des3_ede_cbc_enc_tv_template[] = {
-+static const struct cipher_testvec des3_ede_cbc_enc_tv_template[] = {
- { /* Generated from openssl */
- .key = "\xE9\xC0\xFF\x2E\x76\x0B\x64\x24"
- "\x44\x4D\x99\x5A\x12\xD6\x40\xC0"
-@@ -5867,7 +6063,7 @@ static struct cipher_testvec des3_ede_cb
- },
- };
-
--static struct cipher_testvec des3_ede_cbc_dec_tv_template[] = {
-+static const struct cipher_testvec des3_ede_cbc_dec_tv_template[] = {
- { /* Generated from openssl */
- .key = "\xE9\xC0\xFF\x2E\x76\x0B\x64\x24"
- "\x44\x4D\x99\x5A\x12\xD6\x40\xC0"
-@@ -6047,7 +6243,7 @@ static struct cipher_testvec des3_ede_cb
- },
- };
-
--static struct cipher_testvec des3_ede_ctr_enc_tv_template[] = {
-+static const struct cipher_testvec des3_ede_ctr_enc_tv_template[] = {
- { /* Generated with Crypto++ */
- .key = "\x9C\xD6\xF3\x9C\xB9\x5A\x67\x00"
- "\x5A\x67\x00\x2D\xCE\xEB\x2D\xCE"
-@@ -6325,7 +6521,7 @@ static struct cipher_testvec des3_ede_ct
- },
- };
-
--static struct cipher_testvec des3_ede_ctr_dec_tv_template[] = {
-+static const struct cipher_testvec des3_ede_ctr_dec_tv_template[] = {
- { /* Generated with Crypto++ */
- .key = "\x9C\xD6\xF3\x9C\xB9\x5A\x67\x00"
- "\x5A\x67\x00\x2D\xCE\xEB\x2D\xCE"
-@@ -6606,14 +6802,7 @@ static struct cipher_testvec des3_ede_ct
- /*
- * Blowfish test vectors.
- */
--#define BF_ENC_TEST_VECTORS 7
--#define BF_DEC_TEST_VECTORS 7
--#define BF_CBC_ENC_TEST_VECTORS 2
--#define BF_CBC_DEC_TEST_VECTORS 2
--#define BF_CTR_ENC_TEST_VECTORS 2
--#define BF_CTR_DEC_TEST_VECTORS 2
--
--static struct cipher_testvec bf_enc_tv_template[] = {
-+static const struct cipher_testvec bf_enc_tv_template[] = {
- { /* DES test vectors from OpenSSL */
- .key = "\x00\x00\x00\x00\x00\x00\x00\x00",
- .klen = 8,
-@@ -6805,7 +6994,7 @@ static struct cipher_testvec bf_enc_tv_t
- },
- };
-
--static struct cipher_testvec bf_dec_tv_template[] = {
-+static const struct cipher_testvec bf_dec_tv_template[] = {
- { /* DES test vectors from OpenSSL */
- .key = "\x00\x00\x00\x00\x00\x00\x00\x00",
- .klen = 8,
-@@ -6997,7 +7186,7 @@ static struct cipher_testvec bf_dec_tv_t
- },
- };
-
--static struct cipher_testvec bf_cbc_enc_tv_template[] = {
-+static const struct cipher_testvec bf_cbc_enc_tv_template[] = {
- { /* From OpenSSL */
- .key = "\x01\x23\x45\x67\x89\xab\xcd\xef"
- "\xf0\xe1\xd2\xc3\xb4\xa5\x96\x87",
-@@ -7154,7 +7343,7 @@ static struct cipher_testvec bf_cbc_enc_
- },
- };
-
--static struct cipher_testvec bf_cbc_dec_tv_template[] = {
-+static const struct cipher_testvec bf_cbc_dec_tv_template[] = {
- { /* From OpenSSL */
- .key = "\x01\x23\x45\x67\x89\xab\xcd\xef"
- "\xf0\xe1\xd2\xc3\xb4\xa5\x96\x87",
-@@ -7311,7 +7500,7 @@ static struct cipher_testvec bf_cbc_dec_
- },
- };
-
--static struct cipher_testvec bf_ctr_enc_tv_template[] = {
-+static const struct cipher_testvec bf_ctr_enc_tv_template[] = {
- { /* Generated with Crypto++ */
- .key = "\x85\x62\x3F\x1C\xF9\xD6\x1C\xF9"
- "\xD6\xB3\x90\x6D\x4A\x90\x6D\x4A"
-@@ -7723,7 +7912,7 @@ static struct cipher_testvec bf_ctr_enc_
- },
- };
-
--static struct cipher_testvec bf_ctr_dec_tv_template[] = {
-+static const struct cipher_testvec bf_ctr_dec_tv_template[] = {
- { /* Generated with Crypto++ */
- .key = "\x85\x62\x3F\x1C\xF9\xD6\x1C\xF9"
- "\xD6\xB3\x90\x6D\x4A\x90\x6D\x4A"
-@@ -8138,18 +8327,7 @@ static struct cipher_testvec bf_ctr_dec_
- /*
- * Twofish test vectors.
- */
--#define TF_ENC_TEST_VECTORS 4
--#define TF_DEC_TEST_VECTORS 4
--#define TF_CBC_ENC_TEST_VECTORS 5
--#define TF_CBC_DEC_TEST_VECTORS 5
--#define TF_CTR_ENC_TEST_VECTORS 2
--#define TF_CTR_DEC_TEST_VECTORS 2
--#define TF_LRW_ENC_TEST_VECTORS 8
--#define TF_LRW_DEC_TEST_VECTORS 8
--#define TF_XTS_ENC_TEST_VECTORS 5
--#define TF_XTS_DEC_TEST_VECTORS 5
--
--static struct cipher_testvec tf_enc_tv_template[] = {
-+static const struct cipher_testvec tf_enc_tv_template[] = {
- {
- .key = zeroed_string,
- .klen = 16,
-@@ -8317,7 +8495,7 @@ static struct cipher_testvec tf_enc_tv_t
- },
- };
-
--static struct cipher_testvec tf_dec_tv_template[] = {
-+static const struct cipher_testvec tf_dec_tv_template[] = {
- {
- .key = zeroed_string,
- .klen = 16,
-@@ -8485,7 +8663,7 @@ static struct cipher_testvec tf_dec_tv_t
- },
- };
-
--static struct cipher_testvec tf_cbc_enc_tv_template[] = {
-+static const struct cipher_testvec tf_cbc_enc_tv_template[] = {
- { /* Generated with Nettle */
- .key = zeroed_string,
- .klen = 16,
-@@ -8668,7 +8846,7 @@ static struct cipher_testvec tf_cbc_enc_
- },
- };
-
--static struct cipher_testvec tf_cbc_dec_tv_template[] = {
-+static const struct cipher_testvec tf_cbc_dec_tv_template[] = {
- { /* Reverse of the first four above */
- .key = zeroed_string,
- .klen = 16,
-@@ -8851,7 +9029,7 @@ static struct cipher_testvec tf_cbc_dec_
- },
- };
-
--static struct cipher_testvec tf_ctr_enc_tv_template[] = {
-+static const struct cipher_testvec tf_ctr_enc_tv_template[] = {
- { /* Generated with Crypto++ */
- .key = "\x85\x62\x3F\x1C\xF9\xD6\x1C\xF9"
- "\xD6\xB3\x90\x6D\x4A\x90\x6D\x4A"
-@@ -9262,7 +9440,7 @@ static struct cipher_testvec tf_ctr_enc_
- },
- };
-
--static struct cipher_testvec tf_ctr_dec_tv_template[] = {
-+static const struct cipher_testvec tf_ctr_dec_tv_template[] = {
- { /* Generated with Crypto++ */
- .key = "\x85\x62\x3F\x1C\xF9\xD6\x1C\xF9"
- "\xD6\xB3\x90\x6D\x4A\x90\x6D\x4A"
-@@ -9673,7 +9851,7 @@ static struct cipher_testvec tf_ctr_dec_
- },
- };
-
--static struct cipher_testvec tf_lrw_enc_tv_template[] = {
-+static const struct cipher_testvec tf_lrw_enc_tv_template[] = {
- /* Generated from AES-LRW test vectors */
- {
- .key = "\x45\x62\xac\x25\xf8\x28\x17\x6d"
-@@ -9925,7 +10103,7 @@ static struct cipher_testvec tf_lrw_enc_
- },
- };
-
--static struct cipher_testvec tf_lrw_dec_tv_template[] = {
-+static const struct cipher_testvec tf_lrw_dec_tv_template[] = {
- /* Generated from AES-LRW test vectors */
- /* same as enc vectors with input and result reversed */
- {
-@@ -10178,7 +10356,7 @@ static struct cipher_testvec tf_lrw_dec_
- },
- };
-
--static struct cipher_testvec tf_xts_enc_tv_template[] = {
-+static const struct cipher_testvec tf_xts_enc_tv_template[] = {
- /* Generated from AES-XTS test vectors */
- {
- .key = "\x00\x00\x00\x00\x00\x00\x00\x00"
-@@ -10520,7 +10698,7 @@ static struct cipher_testvec tf_xts_enc_
- },
- };
-
--static struct cipher_testvec tf_xts_dec_tv_template[] = {
-+static const struct cipher_testvec tf_xts_dec_tv_template[] = {
- /* Generated from AES-XTS test vectors */
- /* same as enc vectors with input and result reversed */
- {
-@@ -10867,25 +11045,7 @@ static struct cipher_testvec tf_xts_dec_
- * Serpent test vectors. These are backwards because Serpent writes
- * octet sequences in right-to-left mode.
- */
--#define SERPENT_ENC_TEST_VECTORS 5
--#define SERPENT_DEC_TEST_VECTORS 5
--
--#define TNEPRES_ENC_TEST_VECTORS 4
--#define TNEPRES_DEC_TEST_VECTORS 4
--
--#define SERPENT_CBC_ENC_TEST_VECTORS 1
--#define SERPENT_CBC_DEC_TEST_VECTORS 1
--
--#define SERPENT_CTR_ENC_TEST_VECTORS 2
--#define SERPENT_CTR_DEC_TEST_VECTORS 2
--
--#define SERPENT_LRW_ENC_TEST_VECTORS 8
--#define SERPENT_LRW_DEC_TEST_VECTORS 8
--
--#define SERPENT_XTS_ENC_TEST_VECTORS 5
--#define SERPENT_XTS_DEC_TEST_VECTORS 5
--
--static struct cipher_testvec serpent_enc_tv_template[] = {
-+static const struct cipher_testvec serpent_enc_tv_template[] = {
- {
- .input = "\x00\x01\x02\x03\x04\x05\x06\x07"
- "\x08\x09\x0a\x0b\x0c\x0d\x0e\x0f",
-@@ -11061,7 +11221,7 @@ static struct cipher_testvec serpent_enc
- },
- };
-
--static struct cipher_testvec tnepres_enc_tv_template[] = {
-+static const struct cipher_testvec tnepres_enc_tv_template[] = {
- { /* KeySize=128, PT=0, I=1 */
- .input = "\x00\x00\x00\x00\x00\x00\x00\x00"
- "\x00\x00\x00\x00\x00\x00\x00\x00",
-@@ -11111,7 +11271,7 @@ static struct cipher_testvec tnepres_enc
- };
-
-
--static struct cipher_testvec serpent_dec_tv_template[] = {
-+static const struct cipher_testvec serpent_dec_tv_template[] = {
- {
- .input = "\x12\x07\xfc\xce\x9b\xd0\xd6\x47"
- "\x6a\xe9\x8f\xbe\xd1\x43\xa0\xe2",
-@@ -11287,7 +11447,7 @@ static struct cipher_testvec serpent_dec
- },
- };
-
--static struct cipher_testvec tnepres_dec_tv_template[] = {
-+static const struct cipher_testvec tnepres_dec_tv_template[] = {
- {
- .input = "\x41\xcc\x6b\x31\x59\x31\x45\x97"
- "\x6d\x6f\xbb\x38\x4b\x37\x21\x28",
-@@ -11328,7 +11488,7 @@ static struct cipher_testvec tnepres_dec
- },
- };
-
--static struct cipher_testvec serpent_cbc_enc_tv_template[] = {
-+static const struct cipher_testvec serpent_cbc_enc_tv_template[] = {
- { /* Generated with Crypto++ */
- .key = "\x85\x62\x3F\x1C\xF9\xD6\x1C\xF9"
- "\xD6\xB3\x90\x6D\x4A\x90\x6D\x4A"
-@@ -11469,7 +11629,7 @@ static struct cipher_testvec serpent_cbc
- },
- };
-
--static struct cipher_testvec serpent_cbc_dec_tv_template[] = {
-+static const struct cipher_testvec serpent_cbc_dec_tv_template[] = {
- { /* Generated with Crypto++ */
- .key = "\x85\x62\x3F\x1C\xF9\xD6\x1C\xF9"
- "\xD6\xB3\x90\x6D\x4A\x90\x6D\x4A"
-@@ -11610,7 +11770,7 @@ static struct cipher_testvec serpent_cbc
- },
- };
-
--static struct cipher_testvec serpent_ctr_enc_tv_template[] = {
-+static const struct cipher_testvec serpent_ctr_enc_tv_template[] = {
- { /* Generated with Crypto++ */
- .key = "\x85\x62\x3F\x1C\xF9\xD6\x1C\xF9"
- "\xD6\xB3\x90\x6D\x4A\x90\x6D\x4A"
-@@ -12021,7 +12181,7 @@ static struct cipher_testvec serpent_ctr
- },
- };
-
--static struct cipher_testvec serpent_ctr_dec_tv_template[] = {
-+static const struct cipher_testvec serpent_ctr_dec_tv_template[] = {
- { /* Generated with Crypto++ */
- .key = "\x85\x62\x3F\x1C\xF9\xD6\x1C\xF9"
- "\xD6\xB3\x90\x6D\x4A\x90\x6D\x4A"
-@@ -12432,7 +12592,7 @@ static struct cipher_testvec serpent_ctr
- },
- };
-
--static struct cipher_testvec serpent_lrw_enc_tv_template[] = {
-+static const struct cipher_testvec serpent_lrw_enc_tv_template[] = {
- /* Generated from AES-LRW test vectors */
- {
- .key = "\x45\x62\xac\x25\xf8\x28\x17\x6d"
-@@ -12684,7 +12844,7 @@ static struct cipher_testvec serpent_lrw
- },
- };
-
--static struct cipher_testvec serpent_lrw_dec_tv_template[] = {
-+static const struct cipher_testvec serpent_lrw_dec_tv_template[] = {
- /* Generated from AES-LRW test vectors */
- /* same as enc vectors with input and result reversed */
- {
-@@ -12937,7 +13097,7 @@ static struct cipher_testvec serpent_lrw
- },
- };
-
--static struct cipher_testvec serpent_xts_enc_tv_template[] = {
-+static const struct cipher_testvec serpent_xts_enc_tv_template[] = {
- /* Generated from AES-XTS test vectors */
- {
- .key = "\x00\x00\x00\x00\x00\x00\x00\x00"
-@@ -13279,7 +13439,7 @@ static struct cipher_testvec serpent_xts
- },
- };
-
--static struct cipher_testvec serpent_xts_dec_tv_template[] = {
-+static const struct cipher_testvec serpent_xts_dec_tv_template[] = {
- /* Generated from AES-XTS test vectors */
- /* same as enc vectors with input and result reversed */
- {
-@@ -13623,18 +13783,7 @@ static struct cipher_testvec serpent_xts
- };
-
- /* Cast6 test vectors from RFC 2612 */
--#define CAST6_ENC_TEST_VECTORS 4
--#define CAST6_DEC_TEST_VECTORS 4
--#define CAST6_CBC_ENC_TEST_VECTORS 1
--#define CAST6_CBC_DEC_TEST_VECTORS 1
--#define CAST6_CTR_ENC_TEST_VECTORS 2
--#define CAST6_CTR_DEC_TEST_VECTORS 2
--#define CAST6_LRW_ENC_TEST_VECTORS 1
--#define CAST6_LRW_DEC_TEST_VECTORS 1
--#define CAST6_XTS_ENC_TEST_VECTORS 1
--#define CAST6_XTS_DEC_TEST_VECTORS 1
--
--static struct cipher_testvec cast6_enc_tv_template[] = {
-+static const struct cipher_testvec cast6_enc_tv_template[] = {
- {
- .key = "\x23\x42\xbb\x9e\xfa\x38\x54\x2c"
- "\x0a\xf7\x56\x47\xf2\x9f\x61\x5d",
-@@ -13805,7 +13954,7 @@ static struct cipher_testvec cast6_enc_t
- },
- };
-
--static struct cipher_testvec cast6_dec_tv_template[] = {
-+static const struct cipher_testvec cast6_dec_tv_template[] = {
- {
- .key = "\x23\x42\xbb\x9e\xfa\x38\x54\x2c"
- "\x0a\xf7\x56\x47\xf2\x9f\x61\x5d",
-@@ -13976,7 +14125,7 @@ static struct cipher_testvec cast6_dec_t
- },
- };
-
--static struct cipher_testvec cast6_cbc_enc_tv_template[] = {
-+static const struct cipher_testvec cast6_cbc_enc_tv_template[] = {
- { /* Generated from TF test vectors */
- .key = "\x85\x62\x3F\x1C\xF9\xD6\x1C\xF9"
- "\xD6\xB3\x90\x6D\x4A\x90\x6D\x4A"
-@@ -14117,7 +14266,7 @@ static struct cipher_testvec cast6_cbc_e
- },
- };
-
--static struct cipher_testvec cast6_cbc_dec_tv_template[] = {
-+static const struct cipher_testvec cast6_cbc_dec_tv_template[] = {
- { /* Generated from TF test vectors */
- .key = "\x85\x62\x3F\x1C\xF9\xD6\x1C\xF9"
- "\xD6\xB3\x90\x6D\x4A\x90\x6D\x4A"
-@@ -14258,7 +14407,7 @@ static struct cipher_testvec cast6_cbc_d
- },
- };
-
--static struct cipher_testvec cast6_ctr_enc_tv_template[] = {
-+static const struct cipher_testvec cast6_ctr_enc_tv_template[] = {
- { /* Generated from TF test vectors */
- .key = "\x85\x62\x3F\x1C\xF9\xD6\x1C\xF9"
- "\xD6\xB3\x90\x6D\x4A\x90\x6D\x4A"
-@@ -14415,7 +14564,7 @@ static struct cipher_testvec cast6_ctr_e
- },
- };
-
--static struct cipher_testvec cast6_ctr_dec_tv_template[] = {
-+static const struct cipher_testvec cast6_ctr_dec_tv_template[] = {
- { /* Generated from TF test vectors */
- .key = "\x85\x62\x3F\x1C\xF9\xD6\x1C\xF9"
- "\xD6\xB3\x90\x6D\x4A\x90\x6D\x4A"
-@@ -14572,7 +14721,7 @@ static struct cipher_testvec cast6_ctr_d
- },
- };
-
--static struct cipher_testvec cast6_lrw_enc_tv_template[] = {
-+static const struct cipher_testvec cast6_lrw_enc_tv_template[] = {
- { /* Generated from TF test vectors */
- .key = "\xf8\xd4\x76\xff\xd6\x46\xee\x6c"
- "\x23\x84\xcb\x1c\x77\xd6\x19\x5d"
-@@ -14719,7 +14868,7 @@ static struct cipher_testvec cast6_lrw_e
- },
- };
-
--static struct cipher_testvec cast6_lrw_dec_tv_template[] = {
-+static const struct cipher_testvec cast6_lrw_dec_tv_template[] = {
- { /* Generated from TF test vectors */
- .key = "\xf8\xd4\x76\xff\xd6\x46\xee\x6c"
- "\x23\x84\xcb\x1c\x77\xd6\x19\x5d"
-@@ -14866,7 +15015,7 @@ static struct cipher_testvec cast6_lrw_d
- },
- };
-
--static struct cipher_testvec cast6_xts_enc_tv_template[] = {
-+static const struct cipher_testvec cast6_xts_enc_tv_template[] = {
- { /* Generated from TF test vectors */
- .key = "\x27\x18\x28\x18\x28\x45\x90\x45"
- "\x23\x53\x60\x28\x74\x71\x35\x26"
-@@ -15015,7 +15164,7 @@ static struct cipher_testvec cast6_xts_e
- },
- };
-
--static struct cipher_testvec cast6_xts_dec_tv_template[] = {
-+static const struct cipher_testvec cast6_xts_dec_tv_template[] = {
- { /* Generated from TF test vectors */
- .key = "\x27\x18\x28\x18\x28\x45\x90\x45"
- "\x23\x53\x60\x28\x74\x71\x35\x26"
-@@ -15168,39 +15317,7 @@ static struct cipher_testvec cast6_xts_d
- /*
- * AES test vectors.
- */
--#define AES_ENC_TEST_VECTORS 4
--#define AES_DEC_TEST_VECTORS 4
--#define AES_CBC_ENC_TEST_VECTORS 5
--#define AES_CBC_DEC_TEST_VECTORS 5
--#define HMAC_MD5_ECB_CIPHER_NULL_ENC_TEST_VECTORS 2
--#define HMAC_MD5_ECB_CIPHER_NULL_DEC_TEST_VECTORS 2
--#define HMAC_SHA1_ECB_CIPHER_NULL_ENC_TEST_VEC 2
--#define HMAC_SHA1_ECB_CIPHER_NULL_DEC_TEST_VEC 2
--#define HMAC_SHA1_AES_CBC_ENC_TEST_VEC 7
--#define HMAC_SHA256_AES_CBC_ENC_TEST_VEC 7
--#define HMAC_SHA512_AES_CBC_ENC_TEST_VEC 7
--#define AES_LRW_ENC_TEST_VECTORS 8
--#define AES_LRW_DEC_TEST_VECTORS 8
--#define AES_XTS_ENC_TEST_VECTORS 5
--#define AES_XTS_DEC_TEST_VECTORS 5
--#define AES_CTR_ENC_TEST_VECTORS 5
--#define AES_CTR_DEC_TEST_VECTORS 5
--#define AES_OFB_ENC_TEST_VECTORS 1
--#define AES_OFB_DEC_TEST_VECTORS 1
--#define AES_CTR_3686_ENC_TEST_VECTORS 7
--#define AES_CTR_3686_DEC_TEST_VECTORS 6
--#define AES_GCM_ENC_TEST_VECTORS 9
--#define AES_GCM_DEC_TEST_VECTORS 8
--#define AES_GCM_4106_ENC_TEST_VECTORS 23
--#define AES_GCM_4106_DEC_TEST_VECTORS 23
--#define AES_GCM_4543_ENC_TEST_VECTORS 1
--#define AES_GCM_4543_DEC_TEST_VECTORS 2
--#define AES_CCM_ENC_TEST_VECTORS 8
--#define AES_CCM_DEC_TEST_VECTORS 7
--#define AES_CCM_4309_ENC_TEST_VECTORS 7
--#define AES_CCM_4309_DEC_TEST_VECTORS 10
--
--static struct cipher_testvec aes_enc_tv_template[] = {
-+static const struct cipher_testvec aes_enc_tv_template[] = {
- { /* From FIPS-197 */
- .key = "\x00\x01\x02\x03\x04\x05\x06\x07"
- "\x08\x09\x0a\x0b\x0c\x0d\x0e\x0f",
-@@ -15372,7 +15489,7 @@ static struct cipher_testvec aes_enc_tv_
- },
- };
-
--static struct cipher_testvec aes_dec_tv_template[] = {
-+static const struct cipher_testvec aes_dec_tv_template[] = {
- { /* From FIPS-197 */
- .key = "\x00\x01\x02\x03\x04\x05\x06\x07"
- "\x08\x09\x0a\x0b\x0c\x0d\x0e\x0f",
-@@ -15544,7 +15661,7 @@ static struct cipher_testvec aes_dec_tv_
- },
- };
-
--static struct cipher_testvec aes_cbc_enc_tv_template[] = {
-+static const struct cipher_testvec aes_cbc_enc_tv_template[] = {
- { /* From RFC 3602 */
- .key = "\x06\xa9\x21\x40\x36\xb8\xa1\x5b"
- "\x51\x2e\x03\xd5\x34\x12\x00\x06",
-@@ -15766,7 +15883,7 @@ static struct cipher_testvec aes_cbc_enc
- },
- };
-
--static struct cipher_testvec aes_cbc_dec_tv_template[] = {
-+static const struct cipher_testvec aes_cbc_dec_tv_template[] = {
- { /* From RFC 3602 */
- .key = "\x06\xa9\x21\x40\x36\xb8\xa1\x5b"
- "\x51\x2e\x03\xd5\x34\x12\x00\x06",
-@@ -15988,7 +16105,7 @@ static struct cipher_testvec aes_cbc_dec
- },
- };
-
--static struct aead_testvec hmac_md5_ecb_cipher_null_enc_tv_template[] = {
-+static const struct aead_testvec hmac_md5_ecb_cipher_null_enc_tv_template[] = {
- { /* Input data from RFC 2410 Case 1 */
- #ifdef __LITTLE_ENDIAN
- .key = "\x08\x00" /* rta length */
-@@ -16030,7 +16147,7 @@ static struct aead_testvec hmac_md5_ecb_
- },
- };
-
--static struct aead_testvec hmac_md5_ecb_cipher_null_dec_tv_template[] = {
-+static const struct aead_testvec hmac_md5_ecb_cipher_null_dec_tv_template[] = {
- {
- #ifdef __LITTLE_ENDIAN
- .key = "\x08\x00" /* rta length */
-@@ -16072,7 +16189,7 @@ static struct aead_testvec hmac_md5_ecb_
- },
- };
-
--static struct aead_testvec hmac_sha1_aes_cbc_enc_tv_temp[] = {
-+static const struct aead_testvec hmac_sha1_aes_cbc_enc_tv_temp[] = {
- { /* RFC 3602 Case 1 */
- #ifdef __LITTLE_ENDIAN
- .key = "\x08\x00" /* rta length */
-@@ -16341,7 +16458,7 @@ static struct aead_testvec hmac_sha1_aes
- },
- };
-
--static struct aead_testvec hmac_sha1_ecb_cipher_null_enc_tv_temp[] = {
-+static const struct aead_testvec hmac_sha1_ecb_cipher_null_enc_tv_temp[] = {
- { /* Input data from RFC 2410 Case 1 */
- #ifdef __LITTLE_ENDIAN
- .key = "\x08\x00" /* rta length */
-@@ -16387,7 +16504,7 @@ static struct aead_testvec hmac_sha1_ecb
- },
- };
-
--static struct aead_testvec hmac_sha1_ecb_cipher_null_dec_tv_temp[] = {
-+static const struct aead_testvec hmac_sha1_ecb_cipher_null_dec_tv_temp[] = {
- {
- #ifdef __LITTLE_ENDIAN
- .key = "\x08\x00" /* rta length */
-@@ -16433,7 +16550,7 @@ static struct aead_testvec hmac_sha1_ecb
- },
- };
-
--static struct aead_testvec hmac_sha256_aes_cbc_enc_tv_temp[] = {
-+static const struct aead_testvec hmac_sha256_aes_cbc_enc_tv_temp[] = {
- { /* RFC 3602 Case 1 */
- #ifdef __LITTLE_ENDIAN
- .key = "\x08\x00" /* rta length */
-@@ -16716,7 +16833,7 @@ static struct aead_testvec hmac_sha256_a
- },
- };
-
--static struct aead_testvec hmac_sha512_aes_cbc_enc_tv_temp[] = {
-+static const struct aead_testvec hmac_sha512_aes_cbc_enc_tv_temp[] = {
- { /* RFC 3602 Case 1 */
- #ifdef __LITTLE_ENDIAN
- .key = "\x08\x00" /* rta length */
-@@ -17055,9 +17172,7 @@ static struct aead_testvec hmac_sha512_a
- },
- };
-
--#define HMAC_SHA1_DES_CBC_ENC_TEST_VEC 1
--
--static struct aead_testvec hmac_sha1_des_cbc_enc_tv_temp[] = {
-+static const struct aead_testvec hmac_sha1_des_cbc_enc_tv_temp[] = {
- { /*Generated with cryptopp*/
- #ifdef __LITTLE_ENDIAN
- .key = "\x08\x00" /* rta length */
-@@ -17116,9 +17231,7 @@ static struct aead_testvec hmac_sha1_des
- },
- };
-
--#define HMAC_SHA224_DES_CBC_ENC_TEST_VEC 1
--
--static struct aead_testvec hmac_sha224_des_cbc_enc_tv_temp[] = {
-+static const struct aead_testvec hmac_sha224_des_cbc_enc_tv_temp[] = {
- { /*Generated with cryptopp*/
- #ifdef __LITTLE_ENDIAN
- .key = "\x08\x00" /* rta length */
-@@ -17177,9 +17290,7 @@ static struct aead_testvec hmac_sha224_d
- },
- };
-
--#define HMAC_SHA256_DES_CBC_ENC_TEST_VEC 1
--
--static struct aead_testvec hmac_sha256_des_cbc_enc_tv_temp[] = {
-+static const struct aead_testvec hmac_sha256_des_cbc_enc_tv_temp[] = {
- { /*Generated with cryptopp*/
- #ifdef __LITTLE_ENDIAN
- .key = "\x08\x00" /* rta length */
-@@ -17240,9 +17351,7 @@ static struct aead_testvec hmac_sha256_d
- },
- };
-
--#define HMAC_SHA384_DES_CBC_ENC_TEST_VEC 1
--
--static struct aead_testvec hmac_sha384_des_cbc_enc_tv_temp[] = {
-+static const struct aead_testvec hmac_sha384_des_cbc_enc_tv_temp[] = {
- { /*Generated with cryptopp*/
- #ifdef __LITTLE_ENDIAN
- .key = "\x08\x00" /* rta length */
-@@ -17307,9 +17416,7 @@ static struct aead_testvec hmac_sha384_d
- },
- };
-
--#define HMAC_SHA512_DES_CBC_ENC_TEST_VEC 1
--
--static struct aead_testvec hmac_sha512_des_cbc_enc_tv_temp[] = {
-+static const struct aead_testvec hmac_sha512_des_cbc_enc_tv_temp[] = {
- { /*Generated with cryptopp*/
- #ifdef __LITTLE_ENDIAN
- .key = "\x08\x00" /* rta length */
-@@ -17378,9 +17485,7 @@ static struct aead_testvec hmac_sha512_d
- },
- };
-
--#define HMAC_SHA1_DES3_EDE_CBC_ENC_TEST_VEC 1
--
--static struct aead_testvec hmac_sha1_des3_ede_cbc_enc_tv_temp[] = {
-+static const struct aead_testvec hmac_sha1_des3_ede_cbc_enc_tv_temp[] = {
- { /*Generated with cryptopp*/
- #ifdef __LITTLE_ENDIAN
- .key = "\x08\x00" /* rta length */
-@@ -17441,9 +17546,7 @@ static struct aead_testvec hmac_sha1_des
- },
- };
-
--#define HMAC_SHA224_DES3_EDE_CBC_ENC_TEST_VEC 1
--
--static struct aead_testvec hmac_sha224_des3_ede_cbc_enc_tv_temp[] = {
-+static const struct aead_testvec hmac_sha224_des3_ede_cbc_enc_tv_temp[] = {
- { /*Generated with cryptopp*/
- #ifdef __LITTLE_ENDIAN
- .key = "\x08\x00" /* rta length */
-@@ -17504,9 +17607,7 @@ static struct aead_testvec hmac_sha224_d
- },
- };
-
--#define HMAC_SHA256_DES3_EDE_CBC_ENC_TEST_VEC 1
--
--static struct aead_testvec hmac_sha256_des3_ede_cbc_enc_tv_temp[] = {
-+static const struct aead_testvec hmac_sha256_des3_ede_cbc_enc_tv_temp[] = {
- { /*Generated with cryptopp*/
- #ifdef __LITTLE_ENDIAN
- .key = "\x08\x00" /* rta length */
-@@ -17569,9 +17670,7 @@ static struct aead_testvec hmac_sha256_d
- },
- };
-
--#define HMAC_SHA384_DES3_EDE_CBC_ENC_TEST_VEC 1
--
--static struct aead_testvec hmac_sha384_des3_ede_cbc_enc_tv_temp[] = {
-+static const struct aead_testvec hmac_sha384_des3_ede_cbc_enc_tv_temp[] = {
- { /*Generated with cryptopp*/
- #ifdef __LITTLE_ENDIAN
- .key = "\x08\x00" /* rta length */
-@@ -17638,9 +17737,7 @@ static struct aead_testvec hmac_sha384_d
- },
- };
-
--#define HMAC_SHA512_DES3_EDE_CBC_ENC_TEST_VEC 1
--
--static struct aead_testvec hmac_sha512_des3_ede_cbc_enc_tv_temp[] = {
-+static const struct aead_testvec hmac_sha512_des3_ede_cbc_enc_tv_temp[] = {
- { /*Generated with cryptopp*/
- #ifdef __LITTLE_ENDIAN
- .key = "\x08\x00" /* rta length */
-@@ -17711,7 +17808,7 @@ static struct aead_testvec hmac_sha512_d
- },
- };
-
--static struct cipher_testvec aes_lrw_enc_tv_template[] = {
-+static const struct cipher_testvec aes_lrw_enc_tv_template[] = {
- /* from http://grouper.ieee.org/groups/1619/email/pdf00017.pdf */
- { /* LRW-32-AES 1 */
- .key = "\x45\x62\xac\x25\xf8\x28\x17\x6d"
-@@ -17964,7 +18061,7 @@ static struct cipher_testvec aes_lrw_enc
- }
- };
-
--static struct cipher_testvec aes_lrw_dec_tv_template[] = {
-+static const struct cipher_testvec aes_lrw_dec_tv_template[] = {
- /* from http://grouper.ieee.org/groups/1619/email/pdf00017.pdf */
- /* same as enc vectors with input and result reversed */
- { /* LRW-32-AES 1 */
-@@ -18218,7 +18315,7 @@ static struct cipher_testvec aes_lrw_dec
- }
- };
-
--static struct cipher_testvec aes_xts_enc_tv_template[] = {
-+static const struct cipher_testvec aes_xts_enc_tv_template[] = {
- /* http://grouper.ieee.org/groups/1619/email/pdf00086.pdf */
- { /* XTS-AES 1 */
- .key = "\x00\x00\x00\x00\x00\x00\x00\x00"
-@@ -18561,7 +18658,7 @@ static struct cipher_testvec aes_xts_enc
- }
- };
-
--static struct cipher_testvec aes_xts_dec_tv_template[] = {
-+static const struct cipher_testvec aes_xts_dec_tv_template[] = {
- /* http://grouper.ieee.org/groups/1619/email/pdf00086.pdf */
- { /* XTS-AES 1 */
- .key = "\x00\x00\x00\x00\x00\x00\x00\x00"
-@@ -18905,7 +19002,7 @@ static struct cipher_testvec aes_xts_dec
- };
-
-
--static struct cipher_testvec aes_ctr_enc_tv_template[] = {
-+static const struct cipher_testvec aes_ctr_enc_tv_template[] = {
- { /* From NIST Special Publication 800-38A, Appendix F.5 */
- .key = "\x2b\x7e\x15\x16\x28\xae\xd2\xa6"
- "\xab\xf7\x15\x88\x09\xcf\x4f\x3c",
-@@ -19260,7 +19357,7 @@ static struct cipher_testvec aes_ctr_enc
- },
- };
-
--static struct cipher_testvec aes_ctr_dec_tv_template[] = {
-+static const struct cipher_testvec aes_ctr_dec_tv_template[] = {
- { /* From NIST Special Publication 800-38A, Appendix F.5 */
- .key = "\x2b\x7e\x15\x16\x28\xae\xd2\xa6"
- "\xab\xf7\x15\x88\x09\xcf\x4f\x3c",
-@@ -19615,7 +19712,7 @@ static struct cipher_testvec aes_ctr_dec
- },
- };
-
--static struct cipher_testvec aes_ctr_rfc3686_enc_tv_template[] = {
-+static const struct cipher_testvec aes_ctr_rfc3686_enc_tv_template[] = {
- { /* From RFC 3686 */
- .key = "\xae\x68\x52\xf8\x12\x10\x67\xcc"
- "\x4b\xf7\xa5\x76\x55\x77\xf3\x9e"
-@@ -20747,7 +20844,7 @@ static struct cipher_testvec aes_ctr_rfc
- },
- };
-
--static struct cipher_testvec aes_ctr_rfc3686_dec_tv_template[] = {
-+static const struct cipher_testvec aes_ctr_rfc3686_dec_tv_template[] = {
- { /* From RFC 3686 */
- .key = "\xae\x68\x52\xf8\x12\x10\x67\xcc"
- "\x4b\xf7\xa5\x76\x55\x77\xf3\x9e"
-@@ -20838,7 +20935,7 @@ static struct cipher_testvec aes_ctr_rfc
- },
- };
-
--static struct cipher_testvec aes_ofb_enc_tv_template[] = {
-+static const struct cipher_testvec aes_ofb_enc_tv_template[] = {
- /* From NIST Special Publication 800-38A, Appendix F.5 */
- {
- .key = "\x2b\x7e\x15\x16\x28\xae\xd2\xa6"
-@@ -20867,7 +20964,7 @@ static struct cipher_testvec aes_ofb_enc
- }
- };
-
--static struct cipher_testvec aes_ofb_dec_tv_template[] = {
-+static const struct cipher_testvec aes_ofb_dec_tv_template[] = {
- /* From NIST Special Publication 800-38A, Appendix F.5 */
- {
- .key = "\x2b\x7e\x15\x16\x28\xae\xd2\xa6"
-@@ -20896,7 +20993,7 @@ static struct cipher_testvec aes_ofb_dec
- }
- };
-
--static struct aead_testvec aes_gcm_enc_tv_template[] = {
-+static const struct aead_testvec aes_gcm_enc_tv_template[] = {
- { /* From McGrew & Viega - http://citeseer.ist.psu.edu/656989.html */
- .key = zeroed_string,
- .klen = 16,
-@@ -21056,7 +21153,7 @@ static struct aead_testvec aes_gcm_enc_t
- }
- };
-
--static struct aead_testvec aes_gcm_dec_tv_template[] = {
-+static const struct aead_testvec aes_gcm_dec_tv_template[] = {
- { /* From McGrew & Viega - http://citeseer.ist.psu.edu/656989.html */
- .key = zeroed_string,
- .klen = 32,
-@@ -21258,7 +21355,7 @@ static struct aead_testvec aes_gcm_dec_t
- }
- };
-
--static struct aead_testvec aes_gcm_rfc4106_enc_tv_template[] = {
-+static const struct aead_testvec aes_gcm_rfc4106_enc_tv_template[] = {
- { /* Generated using Crypto++ */
- .key = zeroed_string,
- .klen = 20,
-@@ -21871,7 +21968,7 @@ static struct aead_testvec aes_gcm_rfc41
- }
- };
-
--static struct aead_testvec aes_gcm_rfc4106_dec_tv_template[] = {
-+static const struct aead_testvec aes_gcm_rfc4106_dec_tv_template[] = {
- { /* Generated using Crypto++ */
- .key = zeroed_string,
- .klen = 20,
-@@ -22485,7 +22582,7 @@ static struct aead_testvec aes_gcm_rfc41
- }
- };
-
--static struct aead_testvec aes_gcm_rfc4543_enc_tv_template[] = {
-+static const struct aead_testvec aes_gcm_rfc4543_enc_tv_template[] = {
- { /* From draft-mcgrew-gcm-test-01 */
- .key = "\x4c\x80\xcd\xef\xbb\x5d\x10\xda"
- "\x90\x6a\xc7\x3c\x36\x13\xa6\x34"
-@@ -22516,7 +22613,7 @@ static struct aead_testvec aes_gcm_rfc45
- }
- };
-
--static struct aead_testvec aes_gcm_rfc4543_dec_tv_template[] = {
-+static const struct aead_testvec aes_gcm_rfc4543_dec_tv_template[] = {
- { /* From draft-mcgrew-gcm-test-01 */
- .key = "\x4c\x80\xcd\xef\xbb\x5d\x10\xda"
- "\x90\x6a\xc7\x3c\x36\x13\xa6\x34"
-@@ -22575,7 +22672,7 @@ static struct aead_testvec aes_gcm_rfc45
- },
- };
-
--static struct aead_testvec aes_ccm_enc_tv_template[] = {
-+static const struct aead_testvec aes_ccm_enc_tv_template[] = {
- { /* From RFC 3610 */
- .key = "\xc0\xc1\xc2\xc3\xc4\xc5\xc6\xc7"
- "\xc8\xc9\xca\xcb\xcc\xcd\xce\xcf",
-@@ -22859,7 +22956,7 @@ static struct aead_testvec aes_ccm_enc_t
- }
- };
-
--static struct aead_testvec aes_ccm_dec_tv_template[] = {
-+static const struct aead_testvec aes_ccm_dec_tv_template[] = {
- { /* From RFC 3610 */
- .key = "\xc0\xc1\xc2\xc3\xc4\xc5\xc6\xc7"
- "\xc8\xc9\xca\xcb\xcc\xcd\xce\xcf",
-@@ -23191,7 +23288,7 @@ static struct aead_testvec aes_ccm_dec_t
- * These vectors are copied/generated from the ones for rfc4106 with
- * the key truncated by one byte..
- */
--static struct aead_testvec aes_ccm_rfc4309_enc_tv_template[] = {
-+static const struct aead_testvec aes_ccm_rfc4309_enc_tv_template[] = {
- { /* Generated using Crypto++ */
- .key = zeroed_string,
- .klen = 19,
-@@ -23804,7 +23901,7 @@ static struct aead_testvec aes_ccm_rfc43
- }
- };
-
--static struct aead_testvec aes_ccm_rfc4309_dec_tv_template[] = {
-+static const struct aead_testvec aes_ccm_rfc4309_dec_tv_template[] = {
- { /* Generated using Crypto++ */
- .key = zeroed_string,
- .klen = 19,
-@@ -24420,9 +24517,7 @@ static struct aead_testvec aes_ccm_rfc43
- /*
- * ChaCha20-Poly1305 AEAD test vectors from RFC7539 2.8.2./A.5.
- */
--#define RFC7539_ENC_TEST_VECTORS 2
--#define RFC7539_DEC_TEST_VECTORS 2
--static struct aead_testvec rfc7539_enc_tv_template[] = {
-+static const struct aead_testvec rfc7539_enc_tv_template[] = {
- {
- .key = "\x80\x81\x82\x83\x84\x85\x86\x87"
- "\x88\x89\x8a\x8b\x8c\x8d\x8e\x8f"
-@@ -24554,7 +24649,7 @@ static struct aead_testvec rfc7539_enc_t
- },
- };
-
--static struct aead_testvec rfc7539_dec_tv_template[] = {
-+static const struct aead_testvec rfc7539_dec_tv_template[] = {
- {
- .key = "\x80\x81\x82\x83\x84\x85\x86\x87"
- "\x88\x89\x8a\x8b\x8c\x8d\x8e\x8f"
-@@ -24689,9 +24784,7 @@ static struct aead_testvec rfc7539_dec_t
- /*
- * draft-irtf-cfrg-chacha20-poly1305
- */
--#define RFC7539ESP_DEC_TEST_VECTORS 1
--#define RFC7539ESP_ENC_TEST_VECTORS 1
--static struct aead_testvec rfc7539esp_enc_tv_template[] = {
-+static const struct aead_testvec rfc7539esp_enc_tv_template[] = {
- {
- .key = "\x1c\x92\x40\xa5\xeb\x55\xd3\x8a"
- "\xf3\x33\x88\x86\x04\xf6\xb5\xf0"
-@@ -24779,7 +24872,7 @@ static struct aead_testvec rfc7539esp_en
- },
- };
-
--static struct aead_testvec rfc7539esp_dec_tv_template[] = {
-+static const struct aead_testvec rfc7539esp_dec_tv_template[] = {
- {
- .key = "\x1c\x92\x40\xa5\xeb\x55\xd3\x8a"
- "\xf3\x33\x88\x86\x04\xf6\xb5\xf0"
-@@ -24875,7 +24968,7 @@ static struct aead_testvec rfc7539esp_de
- * semiblock of the ciphertext from the test vector. For decryption, iv is
- * the first semiblock of the ciphertext.
- */
--static struct cipher_testvec aes_kw_enc_tv_template[] = {
-+static const struct cipher_testvec aes_kw_enc_tv_template[] = {
- {
- .key = "\x75\x75\xda\x3a\x93\x60\x7c\xc2"
- "\xbf\xd8\xce\xc7\xaa\xdf\xd9\xa6",
-@@ -24890,7 +24983,7 @@ static struct cipher_testvec aes_kw_enc_
- },
- };
-
--static struct cipher_testvec aes_kw_dec_tv_template[] = {
-+static const struct cipher_testvec aes_kw_dec_tv_template[] = {
- {
- .key = "\x80\xaa\x99\x73\x27\xa4\x80\x6b"
- "\x6a\x7a\x41\xa5\x2b\x86\xc3\x71"
-@@ -24913,9 +25006,7 @@ static struct cipher_testvec aes_kw_dec_
- * http://csrc.nist.gov/groups/STM/cavp/documents/rng/RNGVS.pdf
- * Only AES-128 is supported at this time.
- */
--#define ANSI_CPRNG_AES_TEST_VECTORS 6
--
--static struct cprng_testvec ansi_cprng_aes_tv_template[] = {
-+static const struct cprng_testvec ansi_cprng_aes_tv_template[] = {
- {
- .key = "\xf3\xb1\x66\x6d\x13\x60\x72\x42"
- "\xed\x06\x1c\xab\xb8\xd4\x62\x02",
-@@ -25011,7 +25102,7 @@ static struct cprng_testvec ansi_cprng_a
- * (Hash, HMAC, CTR) are tested with all permutations of use cases (w/ and
- * w/o personalization string, w/ and w/o additional input string).
- */
--static struct drbg_testvec drbg_pr_sha256_tv_template[] = {
-+static const struct drbg_testvec drbg_pr_sha256_tv_template[] = {
- {
- .entropy = (unsigned char *)
- "\x72\x88\x4c\xcd\x6c\x85\x57\x70\xf7\x0b\x8b\x86"
-@@ -25169,7 +25260,7 @@ static struct drbg_testvec drbg_pr_sha25
- },
- };
-
--static struct drbg_testvec drbg_pr_hmac_sha256_tv_template[] = {
-+static const struct drbg_testvec drbg_pr_hmac_sha256_tv_template[] = {
- {
- .entropy = (unsigned char *)
- "\x99\x69\xe5\x4b\x47\x03\xff\x31\x78\x5b\x87\x9a"
-@@ -25327,7 +25418,7 @@ static struct drbg_testvec drbg_pr_hmac_
- },
- };
-
--static struct drbg_testvec drbg_pr_ctr_aes128_tv_template[] = {
-+static const struct drbg_testvec drbg_pr_ctr_aes128_tv_template[] = {
- {
- .entropy = (unsigned char *)
- "\xd1\x44\xc6\x61\x81\x6d\xca\x9d\x15\x28\x8a\x42"
-@@ -25451,7 +25542,7 @@ static struct drbg_testvec drbg_pr_ctr_a
- * (Hash, HMAC, CTR) are tested with all permutations of use cases (w/ and
- * w/o personalization string, w/ and w/o additional input string).
- */
--static struct drbg_testvec drbg_nopr_sha256_tv_template[] = {
-+static const struct drbg_testvec drbg_nopr_sha256_tv_template[] = {
- {
- .entropy = (unsigned char *)
- "\xa6\x5a\xd0\xf3\x45\xdb\x4e\x0e\xff\xe8\x75\xc3"
-@@ -25573,7 +25664,7 @@ static struct drbg_testvec drbg_nopr_sha
- },
- };
-
--static struct drbg_testvec drbg_nopr_hmac_sha256_tv_template[] = {
-+static const struct drbg_testvec drbg_nopr_hmac_sha256_tv_template[] = {
- {
- .entropy = (unsigned char *)
- "\xca\x85\x19\x11\x34\x93\x84\xbf\xfe\x89\xde\x1c"
-@@ -25695,7 +25786,7 @@ static struct drbg_testvec drbg_nopr_hma
- },
- };
-
--static struct drbg_testvec drbg_nopr_ctr_aes192_tv_template[] = {
-+static const struct drbg_testvec drbg_nopr_ctr_aes192_tv_template[] = {
- {
- .entropy = (unsigned char *)
- "\xc3\x5c\x2f\xa2\xa8\x9d\x52\xa1\x1f\xa3\x2a\xa9"
-@@ -25719,7 +25810,7 @@ static struct drbg_testvec drbg_nopr_ctr
- },
- };
-
--static struct drbg_testvec drbg_nopr_ctr_aes256_tv_template[] = {
-+static const struct drbg_testvec drbg_nopr_ctr_aes256_tv_template[] = {
- {
- .entropy = (unsigned char *)
- "\x36\x40\x19\x40\xfa\x8b\x1f\xba\x91\xa1\x66\x1f"
-@@ -25743,7 +25834,7 @@ static struct drbg_testvec drbg_nopr_ctr
- },
- };
-
--static struct drbg_testvec drbg_nopr_ctr_aes128_tv_template[] = {
-+static const struct drbg_testvec drbg_nopr_ctr_aes128_tv_template[] = {
- {
- .entropy = (unsigned char *)
- "\x87\xe1\xc5\x32\x99\x7f\x57\xa3\x5c\x28\x6d\xe8"
-@@ -25832,14 +25923,7 @@ static struct drbg_testvec drbg_nopr_ctr
- };
-
- /* Cast5 test vectors from RFC 2144 */
--#define CAST5_ENC_TEST_VECTORS 4
--#define CAST5_DEC_TEST_VECTORS 4
--#define CAST5_CBC_ENC_TEST_VECTORS 1
--#define CAST5_CBC_DEC_TEST_VECTORS 1
--#define CAST5_CTR_ENC_TEST_VECTORS 2
--#define CAST5_CTR_DEC_TEST_VECTORS 2
--
--static struct cipher_testvec cast5_enc_tv_template[] = {
-+static const struct cipher_testvec cast5_enc_tv_template[] = {
- {
- .key = "\x01\x23\x45\x67\x12\x34\x56\x78"
- "\x23\x45\x67\x89\x34\x56\x78\x9a",
-@@ -26000,7 +26084,7 @@ static struct cipher_testvec cast5_enc_t
- },
- };
-
--static struct cipher_testvec cast5_dec_tv_template[] = {
-+static const struct cipher_testvec cast5_dec_tv_template[] = {
- {
- .key = "\x01\x23\x45\x67\x12\x34\x56\x78"
- "\x23\x45\x67\x89\x34\x56\x78\x9a",
-@@ -26161,7 +26245,7 @@ static struct cipher_testvec cast5_dec_t
- },
- };
-
--static struct cipher_testvec cast5_cbc_enc_tv_template[] = {
-+static const struct cipher_testvec cast5_cbc_enc_tv_template[] = {
- { /* Generated from TF test vectors */
- .key = "\x85\x62\x3F\x1C\xF9\xD6\x1C\xF9"
- "\xD6\xB3\x90\x6D\x4A\x90\x6D\x4A",
-@@ -26299,7 +26383,7 @@ static struct cipher_testvec cast5_cbc_e
- },
- };
-
--static struct cipher_testvec cast5_cbc_dec_tv_template[] = {
-+static const struct cipher_testvec cast5_cbc_dec_tv_template[] = {
- { /* Generated from TF test vectors */
- .key = "\x85\x62\x3F\x1C\xF9\xD6\x1C\xF9"
- "\xD6\xB3\x90\x6D\x4A\x90\x6D\x4A",
-@@ -26437,7 +26521,7 @@ static struct cipher_testvec cast5_cbc_d
- },
- };
-
--static struct cipher_testvec cast5_ctr_enc_tv_template[] = {
-+static const struct cipher_testvec cast5_ctr_enc_tv_template[] = {
- { /* Generated from TF test vectors */
- .key = "\x85\x62\x3F\x1C\xF9\xD6\x1C\xF9"
- "\xD6\xB3\x90\x6D\x4A\x90\x6D\x4A",
-@@ -26588,7 +26672,7 @@ static struct cipher_testvec cast5_ctr_e
- },
- };
-
--static struct cipher_testvec cast5_ctr_dec_tv_template[] = {
-+static const struct cipher_testvec cast5_ctr_dec_tv_template[] = {
- { /* Generated from TF test vectors */
- .key = "\x85\x62\x3F\x1C\xF9\xD6\x1C\xF9"
- "\xD6\xB3\x90\x6D\x4A\x90\x6D\x4A",
-@@ -26742,10 +26826,7 @@ static struct cipher_testvec cast5_ctr_d
- /*
- * ARC4 test vectors from OpenSSL
- */
--#define ARC4_ENC_TEST_VECTORS 7
--#define ARC4_DEC_TEST_VECTORS 7
--
--static struct cipher_testvec arc4_enc_tv_template[] = {
-+static const struct cipher_testvec arc4_enc_tv_template[] = {
- {
- .key = "\x01\x23\x45\x67\x89\xab\xcd\xef",
- .klen = 8,
-@@ -26811,7 +26892,7 @@ static struct cipher_testvec arc4_enc_tv
- },
- };
-
--static struct cipher_testvec arc4_dec_tv_template[] = {
-+static const struct cipher_testvec arc4_dec_tv_template[] = {
- {
- .key = "\x01\x23\x45\x67\x89\xab\xcd\xef",
- .klen = 8,
-@@ -26880,10 +26961,7 @@ static struct cipher_testvec arc4_dec_tv
- /*
- * TEA test vectors
- */
--#define TEA_ENC_TEST_VECTORS 4
--#define TEA_DEC_TEST_VECTORS 4
--
--static struct cipher_testvec tea_enc_tv_template[] = {
-+static const struct cipher_testvec tea_enc_tv_template[] = {
- {
- .key = zeroed_string,
- .klen = 16,
-@@ -26926,7 +27004,7 @@ static struct cipher_testvec tea_enc_tv_
- }
- };
-
--static struct cipher_testvec tea_dec_tv_template[] = {
-+static const struct cipher_testvec tea_dec_tv_template[] = {
- {
- .key = zeroed_string,
- .klen = 16,
-@@ -26972,10 +27050,7 @@ static struct cipher_testvec tea_dec_tv_
- /*
- * XTEA test vectors
- */
--#define XTEA_ENC_TEST_VECTORS 4
--#define XTEA_DEC_TEST_VECTORS 4
--
--static struct cipher_testvec xtea_enc_tv_template[] = {
-+static const struct cipher_testvec xtea_enc_tv_template[] = {
- {
- .key = zeroed_string,
- .klen = 16,
-@@ -27018,7 +27093,7 @@ static struct cipher_testvec xtea_enc_tv
- }
- };
-
--static struct cipher_testvec xtea_dec_tv_template[] = {
-+static const struct cipher_testvec xtea_dec_tv_template[] = {
- {
- .key = zeroed_string,
- .klen = 16,
-@@ -27064,10 +27139,7 @@ static struct cipher_testvec xtea_dec_tv
- /*
- * KHAZAD test vectors.
- */
--#define KHAZAD_ENC_TEST_VECTORS 5
--#define KHAZAD_DEC_TEST_VECTORS 5
--
--static struct cipher_testvec khazad_enc_tv_template[] = {
-+static const struct cipher_testvec khazad_enc_tv_template[] = {
- {
- .key = "\x80\x00\x00\x00\x00\x00\x00\x00"
- "\x00\x00\x00\x00\x00\x00\x00\x00",
-@@ -27113,7 +27185,7 @@ static struct cipher_testvec khazad_enc_
- },
- };
-
--static struct cipher_testvec khazad_dec_tv_template[] = {
-+static const struct cipher_testvec khazad_dec_tv_template[] = {
- {
- .key = "\x80\x00\x00\x00\x00\x00\x00\x00"
- "\x00\x00\x00\x00\x00\x00\x00\x00",
-@@ -27163,12 +27235,7 @@ static struct cipher_testvec khazad_dec_
- * Anubis test vectors.
- */
-
--#define ANUBIS_ENC_TEST_VECTORS 5
--#define ANUBIS_DEC_TEST_VECTORS 5
--#define ANUBIS_CBC_ENC_TEST_VECTORS 2
--#define ANUBIS_CBC_DEC_TEST_VECTORS 2
--
--static struct cipher_testvec anubis_enc_tv_template[] = {
-+static const struct cipher_testvec anubis_enc_tv_template[] = {
- {
- .key = "\xfe\xfe\xfe\xfe\xfe\xfe\xfe\xfe"
- "\xfe\xfe\xfe\xfe\xfe\xfe\xfe\xfe",
-@@ -27231,7 +27298,7 @@ static struct cipher_testvec anubis_enc_
- },
- };
-
--static struct cipher_testvec anubis_dec_tv_template[] = {
-+static const struct cipher_testvec anubis_dec_tv_template[] = {
- {
- .key = "\xfe\xfe\xfe\xfe\xfe\xfe\xfe\xfe"
- "\xfe\xfe\xfe\xfe\xfe\xfe\xfe\xfe",
-@@ -27294,7 +27361,7 @@ static struct cipher_testvec anubis_dec_
- },
- };
-
--static struct cipher_testvec anubis_cbc_enc_tv_template[] = {
-+static const struct cipher_testvec anubis_cbc_enc_tv_template[] = {
- {
- .key = "\xfe\xfe\xfe\xfe\xfe\xfe\xfe\xfe"
- "\xfe\xfe\xfe\xfe\xfe\xfe\xfe\xfe",
-@@ -27329,7 +27396,7 @@ static struct cipher_testvec anubis_cbc_
- },
- };
-
--static struct cipher_testvec anubis_cbc_dec_tv_template[] = {
-+static const struct cipher_testvec anubis_cbc_dec_tv_template[] = {
- {
- .key = "\xfe\xfe\xfe\xfe\xfe\xfe\xfe\xfe"
- "\xfe\xfe\xfe\xfe\xfe\xfe\xfe\xfe",
-@@ -27367,10 +27434,7 @@ static struct cipher_testvec anubis_cbc_
- /*
- * XETA test vectors
- */
--#define XETA_ENC_TEST_VECTORS 4
--#define XETA_DEC_TEST_VECTORS 4
--
--static struct cipher_testvec xeta_enc_tv_template[] = {
-+static const struct cipher_testvec xeta_enc_tv_template[] = {
- {
- .key = zeroed_string,
- .klen = 16,
-@@ -27413,7 +27477,7 @@ static struct cipher_testvec xeta_enc_tv
- }
- };
-
--static struct cipher_testvec xeta_dec_tv_template[] = {
-+static const struct cipher_testvec xeta_dec_tv_template[] = {
- {
- .key = zeroed_string,
- .klen = 16,
-@@ -27459,10 +27523,7 @@ static struct cipher_testvec xeta_dec_tv
- /*
- * FCrypt test vectors
- */
--#define FCRYPT_ENC_TEST_VECTORS ARRAY_SIZE(fcrypt_pcbc_enc_tv_template)
--#define FCRYPT_DEC_TEST_VECTORS ARRAY_SIZE(fcrypt_pcbc_dec_tv_template)
--
--static struct cipher_testvec fcrypt_pcbc_enc_tv_template[] = {
-+static const struct cipher_testvec fcrypt_pcbc_enc_tv_template[] = {
- { /* http://www.openafs.org/pipermail/openafs-devel/2000-December/005320.html */
- .key = "\x00\x00\x00\x00\x00\x00\x00\x00",
- .klen = 8,
-@@ -27523,7 +27584,7 @@ static struct cipher_testvec fcrypt_pcbc
- }
- };
-
--static struct cipher_testvec fcrypt_pcbc_dec_tv_template[] = {
-+static const struct cipher_testvec fcrypt_pcbc_dec_tv_template[] = {
- { /* http://www.openafs.org/pipermail/openafs-devel/2000-December/005320.html */
- .key = "\x00\x00\x00\x00\x00\x00\x00\x00",
- .klen = 8,
-@@ -27587,18 +27648,7 @@ static struct cipher_testvec fcrypt_pcbc
- /*
- * CAMELLIA test vectors.
- */
--#define CAMELLIA_ENC_TEST_VECTORS 4
--#define CAMELLIA_DEC_TEST_VECTORS 4
--#define CAMELLIA_CBC_ENC_TEST_VECTORS 3
--#define CAMELLIA_CBC_DEC_TEST_VECTORS 3
--#define CAMELLIA_CTR_ENC_TEST_VECTORS 2
--#define CAMELLIA_CTR_DEC_TEST_VECTORS 2
--#define CAMELLIA_LRW_ENC_TEST_VECTORS 8
--#define CAMELLIA_LRW_DEC_TEST_VECTORS 8
--#define CAMELLIA_XTS_ENC_TEST_VECTORS 5
--#define CAMELLIA_XTS_DEC_TEST_VECTORS 5
--
--static struct cipher_testvec camellia_enc_tv_template[] = {
-+static const struct cipher_testvec camellia_enc_tv_template[] = {
- {
- .key = "\x01\x23\x45\x67\x89\xab\xcd\xef"
- "\xfe\xdc\xba\x98\x76\x54\x32\x10",
-@@ -27898,7 +27948,7 @@ static struct cipher_testvec camellia_en
- },
- };
-
--static struct cipher_testvec camellia_dec_tv_template[] = {
-+static const struct cipher_testvec camellia_dec_tv_template[] = {
- {
- .key = "\x01\x23\x45\x67\x89\xab\xcd\xef"
- "\xfe\xdc\xba\x98\x76\x54\x32\x10",
-@@ -28198,7 +28248,7 @@ static struct cipher_testvec camellia_de
- },
- };
-
--static struct cipher_testvec camellia_cbc_enc_tv_template[] = {
-+static const struct cipher_testvec camellia_cbc_enc_tv_template[] = {
- {
- .key = "\x06\xa9\x21\x40\x36\xb8\xa1\x5b"
- "\x51\x2e\x03\xd5\x34\x12\x00\x06",
-@@ -28494,7 +28544,7 @@ static struct cipher_testvec camellia_cb
- },
- };
-
--static struct cipher_testvec camellia_cbc_dec_tv_template[] = {
-+static const struct cipher_testvec camellia_cbc_dec_tv_template[] = {
- {
- .key = "\x06\xa9\x21\x40\x36\xb8\xa1\x5b"
- "\x51\x2e\x03\xd5\x34\x12\x00\x06",
-@@ -28790,7 +28840,7 @@ static struct cipher_testvec camellia_cb
- },
- };
-
--static struct cipher_testvec camellia_ctr_enc_tv_template[] = {
-+static const struct cipher_testvec camellia_ctr_enc_tv_template[] = {
- { /* Generated with Crypto++ */
- .key = "\x85\x62\x3F\x1C\xF9\xD6\x1C\xF9"
- "\xD6\xB3\x90\x6D\x4A\x90\x6D\x4A"
-@@ -29457,7 +29507,7 @@ static struct cipher_testvec camellia_ct
- },
- };
-
--static struct cipher_testvec camellia_ctr_dec_tv_template[] = {
-+static const struct cipher_testvec camellia_ctr_dec_tv_template[] = {
- { /* Generated with Crypto++ */
- .key = "\x85\x62\x3F\x1C\xF9\xD6\x1C\xF9"
- "\xD6\xB3\x90\x6D\x4A\x90\x6D\x4A"
-@@ -30124,7 +30174,7 @@ static struct cipher_testvec camellia_ct
- },
- };
-
--static struct cipher_testvec camellia_lrw_enc_tv_template[] = {
-+static const struct cipher_testvec camellia_lrw_enc_tv_template[] = {
- /* Generated from AES-LRW test vectors */
- {
- .key = "\x45\x62\xac\x25\xf8\x28\x17\x6d"
-@@ -30376,7 +30426,7 @@ static struct cipher_testvec camellia_lr
- },
- };
-
--static struct cipher_testvec camellia_lrw_dec_tv_template[] = {
-+static const struct cipher_testvec camellia_lrw_dec_tv_template[] = {
- /* Generated from AES-LRW test vectors */
- /* same as enc vectors with input and result reversed */
- {
-@@ -30629,7 +30679,7 @@ static struct cipher_testvec camellia_lr
- },
- };
-
--static struct cipher_testvec camellia_xts_enc_tv_template[] = {
-+static const struct cipher_testvec camellia_xts_enc_tv_template[] = {
- /* Generated from AES-XTS test vectors */
- {
- .key = "\x00\x00\x00\x00\x00\x00\x00\x00"
-@@ -30971,7 +31021,7 @@ static struct cipher_testvec camellia_xt
- },
- };
-
--static struct cipher_testvec camellia_xts_dec_tv_template[] = {
-+static const struct cipher_testvec camellia_xts_dec_tv_template[] = {
- /* Generated from AES-XTS test vectors */
- /* same as enc vectors with input and result reversed */
- {
-@@ -31317,10 +31367,7 @@ static struct cipher_testvec camellia_xt
- /*
- * SEED test vectors
- */
--#define SEED_ENC_TEST_VECTORS 4
--#define SEED_DEC_TEST_VECTORS 4
--
--static struct cipher_testvec seed_enc_tv_template[] = {
-+static const struct cipher_testvec seed_enc_tv_template[] = {
- {
- .key = zeroed_string,
- .klen = 16,
-@@ -31362,7 +31409,7 @@ static struct cipher_testvec seed_enc_tv
- }
- };
-
--static struct cipher_testvec seed_dec_tv_template[] = {
-+static const struct cipher_testvec seed_dec_tv_template[] = {
- {
- .key = zeroed_string,
- .klen = 16,
-@@ -31404,8 +31451,7 @@ static struct cipher_testvec seed_dec_tv
- }
- };
-
--#define SALSA20_STREAM_ENC_TEST_VECTORS 5
--static struct cipher_testvec salsa20_stream_enc_tv_template[] = {
-+static const struct cipher_testvec salsa20_stream_enc_tv_template[] = {
- /*
- * Testvectors from verified.test-vectors submitted to ECRYPT.
- * They are truncated to size 39, 64, 111, 129 to test a variety
-@@ -32574,8 +32620,7 @@ static struct cipher_testvec salsa20_str
- },
- };
-
--#define CHACHA20_ENC_TEST_VECTORS 4
--static struct cipher_testvec chacha20_enc_tv_template[] = {
-+static const struct cipher_testvec chacha20_enc_tv_template[] = {
- { /* RFC7539 A.2. Test Vector #1 */
- .key = "\x00\x00\x00\x00\x00\x00\x00\x00"
- "\x00\x00\x00\x00\x00\x00\x00\x00"
-@@ -33086,9 +33131,7 @@ static struct cipher_testvec chacha20_en
- /*
- * CTS (Cipher Text Stealing) mode tests
- */
--#define CTS_MODE_ENC_TEST_VECTORS 6
--#define CTS_MODE_DEC_TEST_VECTORS 6
--static struct cipher_testvec cts_mode_enc_tv_template[] = {
-+static const struct cipher_testvec cts_mode_enc_tv_template[] = {
- { /* from rfc3962 */
- .klen = 16,
- .key = "\x63\x68\x69\x63\x6b\x65\x6e\x20"
-@@ -33190,7 +33233,7 @@ static struct cipher_testvec cts_mode_en
- }
- };
-
--static struct cipher_testvec cts_mode_dec_tv_template[] = {
-+static const struct cipher_testvec cts_mode_dec_tv_template[] = {
- { /* from rfc3962 */
- .klen = 16,
- .key = "\x63\x68\x69\x63\x6b\x65\x6e\x20"
-@@ -33308,10 +33351,7 @@ struct comp_testvec {
- * Params: winbits=-11, Z_DEFAULT_COMPRESSION, MAX_MEM_LEVEL.
- */
-
--#define DEFLATE_COMP_TEST_VECTORS 2
--#define DEFLATE_DECOMP_TEST_VECTORS 2
--
--static struct comp_testvec deflate_comp_tv_template[] = {
-+static const struct comp_testvec deflate_comp_tv_template[] = {
- {
- .inlen = 70,
- .outlen = 38,
-@@ -33347,7 +33387,7 @@ static struct comp_testvec deflate_comp_
- },
- };
-
--static struct comp_testvec deflate_decomp_tv_template[] = {
-+static const struct comp_testvec deflate_decomp_tv_template[] = {
- {
- .inlen = 122,
- .outlen = 191,
-@@ -33386,10 +33426,7 @@ static struct comp_testvec deflate_decom
- /*
- * LZO test vectors (null-terminated strings).
- */
--#define LZO_COMP_TEST_VECTORS 2
--#define LZO_DECOMP_TEST_VECTORS 2
--
--static struct comp_testvec lzo_comp_tv_template[] = {
-+static const struct comp_testvec lzo_comp_tv_template[] = {
- {
- .inlen = 70,
- .outlen = 57,
-@@ -33429,7 +33466,7 @@ static struct comp_testvec lzo_comp_tv_t
- },
- };
-
--static struct comp_testvec lzo_decomp_tv_template[] = {
-+static const struct comp_testvec lzo_decomp_tv_template[] = {
- {
- .inlen = 133,
- .outlen = 159,
-@@ -33472,7 +33509,7 @@ static struct comp_testvec lzo_decomp_tv
- */
- #define MICHAEL_MIC_TEST_VECTORS 6
-
--static struct hash_testvec michael_mic_tv_template[] = {
-+static const struct hash_testvec michael_mic_tv_template[] = {
- {
- .key = "\x00\x00\x00\x00\x00\x00\x00\x00",
- .ksize = 8,
-@@ -33520,9 +33557,7 @@ static struct hash_testvec michael_mic_t
- /*
- * CRC32 test vectors
- */
--#define CRC32_TEST_VECTORS 14
--
--static struct hash_testvec crc32_tv_template[] = {
-+static const struct hash_testvec crc32_tv_template[] = {
- {
- .key = "\x87\xa9\xcb\xed",
- .ksize = 4,
-@@ -33954,9 +33989,7 @@ static struct hash_testvec crc32_tv_temp
- /*
- * CRC32C test vectors
- */
--#define CRC32C_TEST_VECTORS 15
--
--static struct hash_testvec crc32c_tv_template[] = {
-+static const struct hash_testvec crc32c_tv_template[] = {
- {
- .psize = 0,
- .digest = "\x00\x00\x00\x00",
-@@ -34392,9 +34425,7 @@ static struct hash_testvec crc32c_tv_tem
- /*
- * Blakcifn CRC test vectors
- */
--#define BFIN_CRC_TEST_VECTORS 6
--
--static struct hash_testvec bfin_crc_tv_template[] = {
-+static const struct hash_testvec bfin_crc_tv_template[] = {
- {
- .psize = 0,
- .digest = "\x00\x00\x00\x00",
-@@ -34479,9 +34510,6 @@ static struct hash_testvec bfin_crc_tv_t
-
- };
-
--#define LZ4_COMP_TEST_VECTORS 1
--#define LZ4_DECOMP_TEST_VECTORS 1
--
- static struct comp_testvec lz4_comp_tv_template[] = {
- {
- .inlen = 70,
-@@ -34512,9 +34540,6 @@ static struct comp_testvec lz4_decomp_tv
- },
- };
-
--#define LZ4HC_COMP_TEST_VECTORS 1
--#define LZ4HC_DECOMP_TEST_VECTORS 1
--
- static struct comp_testvec lz4hc_comp_tv_template[] = {
- {
- .inlen = 70,
---- /dev/null
-+++ b/crypto/tls.c
-@@ -0,0 +1,607 @@
-+/*
-+ * Copyright 2013 Freescale Semiconductor, Inc.
-+ * Copyright 2017 NXP Semiconductor, Inc.
-+ *
-+ * This program is free software; you can redistribute it and/or modify it
-+ * under the terms of the GNU General Public License as published by the Free
-+ * Software Foundation; either version 2 of the License, or (at your option)
-+ * any later version.
-+ *
-+ */
-+
-+#include <crypto/internal/aead.h>
-+#include <crypto/internal/hash.h>
-+#include <crypto/internal/skcipher.h>
-+#include <crypto/authenc.h>
-+#include <crypto/null.h>
-+#include <crypto/scatterwalk.h>
-+#include <linux/err.h>
-+#include <linux/init.h>
-+#include <linux/module.h>
-+#include <linux/rtnetlink.h>
-+
-+struct tls_instance_ctx {
-+ struct crypto_ahash_spawn auth;
-+ struct crypto_skcipher_spawn enc;
-+};
-+
-+struct crypto_tls_ctx {
-+ unsigned int reqoff;
-+ struct crypto_ahash *auth;
-+ struct crypto_skcipher *enc;
-+ struct crypto_skcipher *null;
-+};
-+
-+struct tls_request_ctx {
-+ /*
-+ * cryptlen holds the payload length in the case of encryption or
-+ * payload_len + icv_len + padding_len in case of decryption
-+ */
-+ unsigned int cryptlen;
-+ /* working space for partial results */
-+ struct scatterlist tmp[2];
-+ struct scatterlist cipher[2];
-+ struct scatterlist dst[2];
-+ char tail[];
-+};
-+
-+struct async_op {
-+ struct completion completion;
-+ int err;
-+};
-+
-+static void tls_async_op_done(struct crypto_async_request *req, int err)
-+{
-+ struct async_op *areq = req->data;
-+
-+ if (err == -EINPROGRESS)
-+ return;
-+
-+ areq->err = err;
-+ complete(&areq->completion);
-+}
-+
-+static int crypto_tls_setkey(struct crypto_aead *tls, const u8 *key,
-+ unsigned int keylen)
-+{
-+ struct crypto_tls_ctx *ctx = crypto_aead_ctx(tls);
-+ struct crypto_ahash *auth = ctx->auth;
-+ struct crypto_skcipher *enc = ctx->enc;
-+ struct crypto_authenc_keys keys;
-+ int err = -EINVAL;
-+
-+ if (crypto_authenc_extractkeys(&keys, key, keylen) != 0)
-+ goto badkey;
-+
-+ crypto_ahash_clear_flags(auth, CRYPTO_TFM_REQ_MASK);
-+ crypto_ahash_set_flags(auth, crypto_aead_get_flags(tls) &
-+ CRYPTO_TFM_REQ_MASK);
-+ err = crypto_ahash_setkey(auth, keys.authkey, keys.authkeylen);
-+ crypto_aead_set_flags(tls, crypto_ahash_get_flags(auth) &
-+ CRYPTO_TFM_RES_MASK);
-+
-+ if (err)
-+ goto out;
-+
-+ crypto_skcipher_clear_flags(enc, CRYPTO_TFM_REQ_MASK);
-+ crypto_skcipher_set_flags(enc, crypto_aead_get_flags(tls) &
-+ CRYPTO_TFM_REQ_MASK);
-+ err = crypto_skcipher_setkey(enc, keys.enckey, keys.enckeylen);
-+ crypto_aead_set_flags(tls, crypto_skcipher_get_flags(enc) &
-+ CRYPTO_TFM_RES_MASK);
-+
-+out:
-+ return err;
-+
-+badkey:
-+ crypto_aead_set_flags(tls, CRYPTO_TFM_RES_BAD_KEY_LEN);
-+ goto out;
-+}
-+
-+/**
-+ * crypto_tls_genicv - Calculate hmac digest for a TLS record
-+ * @hash: (output) buffer to save the digest into
-+ * @src: (input) scatterlist with the assoc and payload data
-+ * @srclen: (input) size of the source buffer (assoclen + cryptlen)
-+ * @req: (input) aead request
-+ **/
-+static int crypto_tls_genicv(u8 *hash, struct scatterlist *src,
-+ unsigned int srclen, struct aead_request *req)
-+{
-+ struct crypto_aead *tls = crypto_aead_reqtfm(req);
-+ struct crypto_tls_ctx *ctx = crypto_aead_ctx(tls);
-+ struct tls_request_ctx *treq_ctx = aead_request_ctx(req);
-+ struct async_op ahash_op;
-+ struct ahash_request *ahreq = (void *)(treq_ctx->tail + ctx->reqoff);
-+ unsigned int flags = CRYPTO_TFM_REQ_MAY_SLEEP;
-+ int err = -EBADMSG;
-+
-+ /* Bail out if the request assoc len is 0 */
-+ if (!req->assoclen)
-+ return err;
-+
-+ init_completion(&ahash_op.completion);
-+
-+ /* the hash transform to be executed comes from the original request */
-+ ahash_request_set_tfm(ahreq, ctx->auth);
-+ /* prepare the hash request with input data and result pointer */
-+ ahash_request_set_crypt(ahreq, src, hash, srclen);
-+ /* set the notifier for when the async hash function returns */
-+ ahash_request_set_callback(ahreq, aead_request_flags(req) & flags,
-+ tls_async_op_done, &ahash_op);
-+
-+ /* Calculate the digest on the given data. The result is put in hash */
-+ err = crypto_ahash_digest(ahreq);
-+ if (err == -EINPROGRESS) {
-+ err = wait_for_completion_interruptible(&ahash_op.completion);
-+ if (!err)
-+ err = ahash_op.err;
-+ }
-+
-+ return err;
-+}
-+
-+/**
-+ * crypto_tls_gen_padicv - Calculate and pad hmac digest for a TLS record
-+ * @hash: (output) buffer to save the digest and padding into
-+ * @phashlen: (output) the size of digest + padding
-+ * @req: (input) aead request
-+ **/
-+static int crypto_tls_gen_padicv(u8 *hash, unsigned int *phashlen,
-+ struct aead_request *req)
-+{
-+ struct crypto_aead *tls = crypto_aead_reqtfm(req);
-+ unsigned int hash_size = crypto_aead_authsize(tls);
-+ unsigned int block_size = crypto_aead_blocksize(tls);
-+ unsigned int srclen = req->cryptlen + hash_size;
-+ unsigned int icvlen = req->cryptlen + req->assoclen;
-+ unsigned int padlen;
-+ int err;
-+
-+ err = crypto_tls_genicv(hash, req->src, icvlen, req);
-+ if (err)
-+ goto out;
-+
-+ /* add padding after digest */
-+ padlen = block_size - (srclen % block_size);
-+ memset(hash + hash_size, padlen - 1, padlen);
-+
-+ *phashlen = hash_size + padlen;
-+out:
-+ return err;
-+}
-+
-+static int crypto_tls_copy_data(struct aead_request *req,
-+ struct scatterlist *src,
-+ struct scatterlist *dst,
-+ unsigned int len)
-+{
-+ struct crypto_aead *tls = crypto_aead_reqtfm(req);
-+ struct crypto_tls_ctx *ctx = crypto_aead_ctx(tls);
-+ SKCIPHER_REQUEST_ON_STACK(skreq, ctx->null);
-+
-+ skcipher_request_set_tfm(skreq, ctx->null);
-+ skcipher_request_set_callback(skreq, aead_request_flags(req),
-+ NULL, NULL);
-+ skcipher_request_set_crypt(skreq, src, dst, len, NULL);
-+
-+ return crypto_skcipher_encrypt(skreq);
-+}
-+
-+static int crypto_tls_encrypt(struct aead_request *req)
-+{
-+ struct crypto_aead *tls = crypto_aead_reqtfm(req);
-+ struct crypto_tls_ctx *ctx = crypto_aead_ctx(tls);
-+ struct tls_request_ctx *treq_ctx = aead_request_ctx(req);
-+ struct skcipher_request *skreq;
-+ struct scatterlist *cipher = treq_ctx->cipher;
-+ struct scatterlist *tmp = treq_ctx->tmp;
-+ struct scatterlist *sg, *src, *dst;
-+ unsigned int cryptlen, phashlen;
-+ u8 *hash = treq_ctx->tail;
-+ int err;
-+
-+ /*
-+ * The hash result is saved at the beginning of the tls request ctx
-+ * and is aligned as required by the hash transform. Enough space was
-+ * allocated in crypto_tls_init_tfm to accommodate the difference. The
-+ * requests themselves start later at treq_ctx->tail + ctx->reqoff so
-+ * the result is not overwritten by the second (cipher) request.
-+ */
-+ hash = (u8 *)ALIGN((unsigned long)hash +
-+ crypto_ahash_alignmask(ctx->auth),
-+ crypto_ahash_alignmask(ctx->auth) + 1);
-+
-+ /*
-+ * STEP 1: create ICV together with necessary padding
-+ */
-+ err = crypto_tls_gen_padicv(hash, &phashlen, req);
-+ if (err)
-+ return err;
-+
-+ /*
-+ * STEP 2: Hash and padding are combined with the payload
-+ * depending on the form it arrives. Scatter tables must have at least
-+ * one page of data before chaining with another table and can't have
-+ * an empty data page. The following code addresses these requirements.
-+ *
-+ * If the payload is empty, only the hash is encrypted, otherwise the
-+ * payload scatterlist is merged with the hash. A special merging case
-+ * is when the payload has only one page of data. In that case the
-+ * payload page is moved to another scatterlist and prepared there for
-+ * encryption.
-+ */
-+ if (req->cryptlen) {
-+ src = scatterwalk_ffwd(tmp, req->src, req->assoclen);
-+
-+ sg_init_table(cipher, 2);
-+ sg_set_buf(cipher + 1, hash, phashlen);
-+
-+ if (sg_is_last(src)) {
-+ sg_set_page(cipher, sg_page(src), req->cryptlen,
-+ src->offset);
-+ src = cipher;
-+ } else {
-+ unsigned int rem_len = req->cryptlen;
-+
-+ for (sg = src; rem_len > sg->length; sg = sg_next(sg))
-+ rem_len -= min(rem_len, sg->length);
-+
-+ sg_set_page(cipher, sg_page(sg), rem_len, sg->offset);
-+ sg_chain(sg, 1, cipher);
-+ }
-+ } else {
-+ sg_init_one(cipher, hash, phashlen);
-+ src = cipher;
-+ }
-+
-+ /**
-+ * If src != dst copy the associated data from source to destination.
-+ * In both cases fast-forward passed the associated data in the dest.
-+ */
-+ if (req->src != req->dst) {
-+ err = crypto_tls_copy_data(req, req->src, req->dst,
-+ req->assoclen);
-+ if (err)
-+ return err;
-+ }
-+ dst = scatterwalk_ffwd(treq_ctx->dst, req->dst, req->assoclen);
-+
-+ /*
-+ * STEP 3: encrypt the frame and return the result
-+ */
-+ cryptlen = req->cryptlen + phashlen;
-+
-+ /*
-+ * The hash and the cipher are applied at different times and their
-+ * requests can use the same memory space without interference
-+ */
-+ skreq = (void *)(treq_ctx->tail + ctx->reqoff);
-+ skcipher_request_set_tfm(skreq, ctx->enc);
-+ skcipher_request_set_crypt(skreq, src, dst, cryptlen, req->iv);
-+ skcipher_request_set_callback(skreq, aead_request_flags(req),
-+ req->base.complete, req->base.data);
-+ /*
-+ * Apply the cipher transform. The result will be in req->dst when the
-+ * asynchronuous call terminates
-+ */
-+ err = crypto_skcipher_encrypt(skreq);
-+
-+ return err;
-+}
-+
-+static int crypto_tls_decrypt(struct aead_request *req)
-+{
-+ struct crypto_aead *tls = crypto_aead_reqtfm(req);
-+ struct crypto_tls_ctx *ctx = crypto_aead_ctx(tls);
-+ struct tls_request_ctx *treq_ctx = aead_request_ctx(req);
-+ unsigned int cryptlen = req->cryptlen;
-+ unsigned int hash_size = crypto_aead_authsize(tls);
-+ unsigned int block_size = crypto_aead_blocksize(tls);
-+ struct skcipher_request *skreq = (void *)(treq_ctx->tail + ctx->reqoff);
-+ struct scatterlist *tmp = treq_ctx->tmp;
-+ struct scatterlist *src, *dst;
-+
-+ u8 padding[255]; /* padding can be 0-255 bytes */
-+ u8 pad_size;
-+ u16 *len_field;
-+ u8 *ihash, *hash = treq_ctx->tail;
-+
-+ int paderr = 0;
-+ int err = -EINVAL;
-+ int i;
-+ struct async_op ciph_op;
-+
-+ /*
-+ * Rule out bad packets. The input packet length must be at least one
-+ * byte more than the hash_size
-+ */
-+ if (cryptlen <= hash_size || cryptlen % block_size)
-+ goto out;
-+
-+ /*
-+ * Step 1 - Decrypt the source. Fast-forward past the associated data
-+ * to the encrypted data. The result will be overwritten in place so
-+ * that the decrypted data will be adjacent to the associated data. The
-+ * last step (computing the hash) will have it's input data already
-+ * prepared and ready to be accessed at req->src.
-+ */
-+ src = scatterwalk_ffwd(tmp, req->src, req->assoclen);
-+ dst = src;
-+
-+ init_completion(&ciph_op.completion);
-+ skcipher_request_set_tfm(skreq, ctx->enc);
-+ skcipher_request_set_callback(skreq, aead_request_flags(req),
-+ tls_async_op_done, &ciph_op);
-+ skcipher_request_set_crypt(skreq, src, dst, cryptlen, req->iv);
-+ err = crypto_skcipher_decrypt(skreq);
-+ if (err == -EINPROGRESS) {
-+ err = wait_for_completion_interruptible(&ciph_op.completion);
-+ if (!err)
-+ err = ciph_op.err;
-+ }
-+ if (err)
-+ goto out;
-+
-+ /*
-+ * Step 2 - Verify padding
-+ * Retrieve the last byte of the payload; this is the padding size.
-+ */
-+ cryptlen -= 1;
-+ scatterwalk_map_and_copy(&pad_size, dst, cryptlen, 1, 0);
-+
-+ /* RFC recommendation for invalid padding size. */
-+ if (cryptlen < pad_size + hash_size) {
-+ pad_size = 0;
-+ paderr = -EBADMSG;
-+ }
-+ cryptlen -= pad_size;
-+ scatterwalk_map_and_copy(padding, dst, cryptlen, pad_size, 0);
-+
-+ /* Padding content must be equal with pad_size. We verify it all */
-+ for (i = 0; i < pad_size; i++)
-+ if (padding[i] != pad_size)
-+ paderr = -EBADMSG;
-+
-+ /*
-+ * Step 3 - Verify hash
-+ * Align the digest result as required by the hash transform. Enough
-+ * space was allocated in crypto_tls_init_tfm
-+ */
-+ hash = (u8 *)ALIGN((unsigned long)hash +
-+ crypto_ahash_alignmask(ctx->auth),
-+ crypto_ahash_alignmask(ctx->auth) + 1);
-+ /*
-+ * Two bytes at the end of the associated data make the length field.
-+ * It must be updated with the length of the cleartext message before
-+ * the hash is calculated.
-+ */
-+ len_field = sg_virt(req->src) + req->assoclen - 2;
-+ cryptlen -= hash_size;
-+ *len_field = htons(cryptlen);
-+
-+ /* This is the hash from the decrypted packet. Save it for later */
-+ ihash = hash + hash_size;
-+ scatterwalk_map_and_copy(ihash, dst, cryptlen, hash_size, 0);
-+
-+ /* Now compute and compare our ICV with the one from the packet */
-+ err = crypto_tls_genicv(hash, req->src, cryptlen + req->assoclen, req);
-+ if (!err)
-+ err = memcmp(hash, ihash, hash_size) ? -EBADMSG : 0;
-+
-+ if (req->src != req->dst) {
-+ err = crypto_tls_copy_data(req, req->src, req->dst, cryptlen +
-+ req->assoclen);
-+ if (err)
-+ goto out;
-+ }
-+
-+ /* return the first found error */
-+ if (paderr)
-+ err = paderr;
-+
-+out:
-+ aead_request_complete(req, err);
-+ return err;
-+}
-+
-+static int crypto_tls_init_tfm(struct crypto_aead *tfm)
-+{
-+ struct aead_instance *inst = aead_alg_instance(tfm);
-+ struct tls_instance_ctx *ictx = aead_instance_ctx(inst);
-+ struct crypto_tls_ctx *ctx = crypto_aead_ctx(tfm);
-+ struct crypto_ahash *auth;
-+ struct crypto_skcipher *enc;
-+ struct crypto_skcipher *null;
-+ int err;
-+
-+ auth = crypto_spawn_ahash(&ictx->auth);
-+ if (IS_ERR(auth))
-+ return PTR_ERR(auth);
-+
-+ enc = crypto_spawn_skcipher(&ictx->enc);
-+ err = PTR_ERR(enc);
-+ if (IS_ERR(enc))
-+ goto err_free_ahash;
-+
-+ null = crypto_get_default_null_skcipher2();
-+ err = PTR_ERR(null);
-+ if (IS_ERR(null))
-+ goto err_free_skcipher;
-+
-+ ctx->auth = auth;
-+ ctx->enc = enc;
-+ ctx->null = null;
-+
-+ /*
-+ * Allow enough space for two digests. The two digests will be compared
-+ * during the decryption phase. One will come from the decrypted packet
-+ * and the other will be calculated. For encryption, one digest is
-+ * padded (up to a cipher blocksize) and chained with the payload
-+ */
-+ ctx->reqoff = ALIGN(crypto_ahash_digestsize(auth) +
-+ crypto_ahash_alignmask(auth),
-+ crypto_ahash_alignmask(auth) + 1) +
-+ max(crypto_ahash_digestsize(auth),
-+ crypto_skcipher_blocksize(enc));
-+
-+ crypto_aead_set_reqsize(tfm,
-+ sizeof(struct tls_request_ctx) +
-+ ctx->reqoff +
-+ max_t(unsigned int,
-+ crypto_ahash_reqsize(auth) +
-+ sizeof(struct ahash_request),
-+ crypto_skcipher_reqsize(enc) +
-+ sizeof(struct skcipher_request)));
-+
-+ return 0;
-+
-+err_free_skcipher:
-+ crypto_free_skcipher(enc);
-+err_free_ahash:
-+ crypto_free_ahash(auth);
-+ return err;
-+}
-+
-+static void crypto_tls_exit_tfm(struct crypto_aead *tfm)
-+{
-+ struct crypto_tls_ctx *ctx = crypto_aead_ctx(tfm);
-+
-+ crypto_free_ahash(ctx->auth);
-+ crypto_free_skcipher(ctx->enc);
-+ crypto_put_default_null_skcipher2();
-+}
-+
-+static void crypto_tls_free(struct aead_instance *inst)
-+{
-+ struct tls_instance_ctx *ctx = aead_instance_ctx(inst);
-+
-+ crypto_drop_skcipher(&ctx->enc);
-+ crypto_drop_ahash(&ctx->auth);
-+ kfree(inst);
-+}
-+
-+static int crypto_tls_create(struct crypto_template *tmpl, struct rtattr **tb)
-+{
-+ struct crypto_attr_type *algt;
-+ struct aead_instance *inst;
-+ struct hash_alg_common *auth;
-+ struct crypto_alg *auth_base;
-+ struct skcipher_alg *enc;
-+ struct tls_instance_ctx *ctx;
-+ const char *enc_name;
-+ int err;
-+
-+ algt = crypto_get_attr_type(tb);
-+ if (IS_ERR(algt))
-+ return PTR_ERR(algt);
-+
-+ if ((algt->type ^ CRYPTO_ALG_TYPE_AEAD) & algt->mask)
-+ return -EINVAL;
-+
-+ auth = ahash_attr_alg(tb[1], CRYPTO_ALG_TYPE_HASH,
-+ CRYPTO_ALG_TYPE_AHASH_MASK |
-+ crypto_requires_sync(algt->type, algt->mask));
-+ if (IS_ERR(auth))
-+ return PTR_ERR(auth);
-+
-+ auth_base = &auth->base;
-+
-+ enc_name = crypto_attr_alg_name(tb[2]);
-+ err = PTR_ERR(enc_name);
-+ if (IS_ERR(enc_name))
-+ goto out_put_auth;
-+
-+ inst = kzalloc(sizeof(*inst) + sizeof(*ctx), GFP_KERNEL);
-+ err = -ENOMEM;
-+ if (!inst)
-+ goto out_put_auth;
-+
-+ ctx = aead_instance_ctx(inst);
-+
-+ err = crypto_init_ahash_spawn(&ctx->auth, auth,
-+ aead_crypto_instance(inst));
-+ if (err)
-+ goto err_free_inst;
-+
-+ crypto_set_skcipher_spawn(&ctx->enc, aead_crypto_instance(inst));
-+ err = crypto_grab_skcipher(&ctx->enc, enc_name, 0,
-+ crypto_requires_sync(algt->type,
-+ algt->mask));
-+ if (err)
-+ goto err_drop_auth;
-+
-+ enc = crypto_spawn_skcipher_alg(&ctx->enc);
-+
-+ err = -ENAMETOOLONG;
-+ if (snprintf(inst->alg.base.cra_name, CRYPTO_MAX_ALG_NAME,
-+ "tls10(%s,%s)", auth_base->cra_name,
-+ enc->base.cra_name) >= CRYPTO_MAX_ALG_NAME)
-+ goto err_drop_enc;
-+
-+ if (snprintf(inst->alg.base.cra_driver_name, CRYPTO_MAX_ALG_NAME,
-+ "tls10(%s,%s)", auth_base->cra_driver_name,
-+ enc->base.cra_driver_name) >= CRYPTO_MAX_ALG_NAME)
-+ goto err_drop_enc;
-+
-+ inst->alg.base.cra_flags = (auth_base->cra_flags |
-+ enc->base.cra_flags) & CRYPTO_ALG_ASYNC;
-+ inst->alg.base.cra_priority = enc->base.cra_priority * 10 +
-+ auth_base->cra_priority;
-+ inst->alg.base.cra_blocksize = enc->base.cra_blocksize;
-+ inst->alg.base.cra_alignmask = auth_base->cra_alignmask |
-+ enc->base.cra_alignmask;
-+ inst->alg.base.cra_ctxsize = sizeof(struct crypto_tls_ctx);
-+
-+ inst->alg.ivsize = crypto_skcipher_alg_ivsize(enc);
-+ inst->alg.chunksize = crypto_skcipher_alg_chunksize(enc);
-+ inst->alg.maxauthsize = auth->digestsize;
-+
-+ inst->alg.init = crypto_tls_init_tfm;
-+ inst->alg.exit = crypto_tls_exit_tfm;
-+
-+ inst->alg.setkey = crypto_tls_setkey;
-+ inst->alg.encrypt = crypto_tls_encrypt;
-+ inst->alg.decrypt = crypto_tls_decrypt;
-+
-+ inst->free = crypto_tls_free;
-+
-+ err = aead_register_instance(tmpl, inst);
-+ if (err)
-+ goto err_drop_enc;
-+
-+out:
-+ crypto_mod_put(auth_base);
-+ return err;
-+
-+err_drop_enc:
-+ crypto_drop_skcipher(&ctx->enc);
-+err_drop_auth:
-+ crypto_drop_ahash(&ctx->auth);
-+err_free_inst:
-+ kfree(inst);
-+out_put_auth:
-+ goto out;
-+}
-+
-+static struct crypto_template crypto_tls_tmpl = {
-+ .name = "tls10",
-+ .create = crypto_tls_create,
-+ .module = THIS_MODULE,
-+};
-+
-+static int __init crypto_tls_module_init(void)
-+{
-+ return crypto_register_template(&crypto_tls_tmpl);
-+}
-+
-+static void __exit crypto_tls_module_exit(void)
-+{
-+ crypto_unregister_template(&crypto_tls_tmpl);
-+}
-+
-+module_init(crypto_tls_module_init);
-+module_exit(crypto_tls_module_exit);
-+
-+MODULE_LICENSE("GPL");
-+MODULE_DESCRIPTION("TLS 1.0 record encryption");
---- a/drivers/crypto/caam/Kconfig
-+++ b/drivers/crypto/caam/Kconfig
-@@ -1,6 +1,11 @@
-+config CRYPTO_DEV_FSL_CAAM_COMMON
-+ tristate
-+
- config CRYPTO_DEV_FSL_CAAM
-- tristate "Freescale CAAM-Multicore driver backend"
-+ tristate "Freescale CAAM-Multicore platform driver backend"
- depends on FSL_SOC || ARCH_MXC || ARCH_LAYERSCAPE
-+ select CRYPTO_DEV_FSL_CAAM_COMMON
-+ select SOC_BUS
- help
- Enables the driver module for Freescale's Cryptographic Accelerator
- and Assurance Module (CAAM), also known as the SEC version 4 (SEC4).
-@@ -11,9 +16,16 @@ config CRYPTO_DEV_FSL_CAAM
- To compile this driver as a module, choose M here: the module
- will be called caam.
-
-+if CRYPTO_DEV_FSL_CAAM
-+
-+config CRYPTO_DEV_FSL_CAAM_DEBUG
-+ bool "Enable debug output in CAAM driver"
-+ help
-+ Selecting this will enable printing of various debug
-+ information in the CAAM driver.
-+
- config CRYPTO_DEV_FSL_CAAM_JR
- tristate "Freescale CAAM Job Ring driver backend"
-- depends on CRYPTO_DEV_FSL_CAAM
- default y
- help
- Enables the driver module for Job Rings which are part of
-@@ -24,9 +36,10 @@ config CRYPTO_DEV_FSL_CAAM_JR
- To compile this driver as a module, choose M here: the module
- will be called caam_jr.
-
-+if CRYPTO_DEV_FSL_CAAM_JR
-+
- config CRYPTO_DEV_FSL_CAAM_RINGSIZE
- int "Job Ring size"
-- depends on CRYPTO_DEV_FSL_CAAM_JR
- range 2 9
- default "9"
- help
-@@ -44,7 +57,6 @@ config CRYPTO_DEV_FSL_CAAM_RINGSIZE
-
- config CRYPTO_DEV_FSL_CAAM_INTC
- bool "Job Ring interrupt coalescing"
-- depends on CRYPTO_DEV_FSL_CAAM_JR
- help
- Enable the Job Ring's interrupt coalescing feature.
-
-@@ -74,7 +86,6 @@ config CRYPTO_DEV_FSL_CAAM_INTC_TIME_THL
-
- config CRYPTO_DEV_FSL_CAAM_CRYPTO_API
- tristate "Register algorithm implementations with the Crypto API"
-- depends on CRYPTO_DEV_FSL_CAAM && CRYPTO_DEV_FSL_CAAM_JR
- default y
- select CRYPTO_AEAD
- select CRYPTO_AUTHENC
-@@ -87,9 +98,25 @@ config CRYPTO_DEV_FSL_CAAM_CRYPTO_API
- To compile this as a module, choose M here: the module
- will be called caamalg.
-
-+config CRYPTO_DEV_FSL_CAAM_CRYPTO_API_QI
-+ tristate "Queue Interface as Crypto API backend"
-+ depends on FSL_SDK_DPA && NET
-+ default y
-+ select CRYPTO_AUTHENC
-+ select CRYPTO_BLKCIPHER
-+ help
-+ Selecting this will use CAAM Queue Interface (QI) for sending
-+ & receiving crypto jobs to/from CAAM. This gives better performance
-+ than job ring interface when the number of cores are more than the
-+ number of job rings assigned to the kernel. The number of portals
-+ assigned to the kernel should also be more than the number of
-+ job rings.
-+
-+ To compile this as a module, choose M here: the module
-+ will be called caamalg_qi.
-+
- config CRYPTO_DEV_FSL_CAAM_AHASH_API
- tristate "Register hash algorithm implementations with Crypto API"
-- depends on CRYPTO_DEV_FSL_CAAM && CRYPTO_DEV_FSL_CAAM_JR
- default y
- select CRYPTO_HASH
- help
-@@ -101,7 +128,6 @@ config CRYPTO_DEV_FSL_CAAM_AHASH_API
-
- config CRYPTO_DEV_FSL_CAAM_PKC_API
- tristate "Register public key cryptography implementations with Crypto API"
-- depends on CRYPTO_DEV_FSL_CAAM && CRYPTO_DEV_FSL_CAAM_JR
- default y
- select CRYPTO_RSA
- help
-@@ -113,7 +139,6 @@ config CRYPTO_DEV_FSL_CAAM_PKC_API
-
- config CRYPTO_DEV_FSL_CAAM_RNG_API
- tristate "Register caam device for hwrng API"
-- depends on CRYPTO_DEV_FSL_CAAM && CRYPTO_DEV_FSL_CAAM_JR
- default y
- select CRYPTO_RNG
- select HW_RANDOM
-@@ -124,13 +149,31 @@ config CRYPTO_DEV_FSL_CAAM_RNG_API
- To compile this as a module, choose M here: the module
- will be called caamrng.
-
--config CRYPTO_DEV_FSL_CAAM_IMX
-- def_bool SOC_IMX6 || SOC_IMX7D
-- depends on CRYPTO_DEV_FSL_CAAM
-+endif # CRYPTO_DEV_FSL_CAAM_JR
-
--config CRYPTO_DEV_FSL_CAAM_DEBUG
-- bool "Enable debug output in CAAM driver"
-- depends on CRYPTO_DEV_FSL_CAAM
-- help
-- Selecting this will enable printing of various debug
-- information in the CAAM driver.
-+endif # CRYPTO_DEV_FSL_CAAM
-+
-+config CRYPTO_DEV_FSL_DPAA2_CAAM
-+ tristate "QorIQ DPAA2 CAAM (DPSECI) driver"
-+ depends on FSL_MC_DPIO
-+ select CRYPTO_DEV_FSL_CAAM_COMMON
-+ select CRYPTO_BLKCIPHER
-+ select CRYPTO_AUTHENC
-+ select CRYPTO_AEAD
-+ select CRYPTO_HASH
-+ ---help---
-+ CAAM driver for QorIQ Data Path Acceleration Architecture 2.
-+ It handles DPSECI DPAA2 objects that sit on the Management Complex
-+ (MC) fsl-mc bus.
-+
-+ To compile this as a module, choose M here: the module
-+ will be called dpaa2_caam.
-+
-+config CRYPTO_DEV_FSL_CAAM_CRYPTO_API_DESC
-+ def_tristate (CRYPTO_DEV_FSL_CAAM_CRYPTO_API || \
-+ CRYPTO_DEV_FSL_CAAM_CRYPTO_API_QI || \
-+ CRYPTO_DEV_FSL_DPAA2_CAAM)
-+
-+config CRYPTO_DEV_FSL_CAAM_AHASH_API_DESC
-+ def_tristate (CRYPTO_DEV_FSL_CAAM_AHASH_API || \
-+ CRYPTO_DEV_FSL_DPAA2_CAAM)
---- a/drivers/crypto/caam/Makefile
-+++ b/drivers/crypto/caam/Makefile
-@@ -5,13 +5,27 @@ ifeq ($(CONFIG_CRYPTO_DEV_FSL_CAAM_DEBUG
- ccflags-y := -DDEBUG
- endif
-
-+ccflags-y += -DVERSION=\"\"
-+
-+obj-$(CONFIG_CRYPTO_DEV_FSL_CAAM_COMMON) += error.o
- obj-$(CONFIG_CRYPTO_DEV_FSL_CAAM) += caam.o
- obj-$(CONFIG_CRYPTO_DEV_FSL_CAAM_JR) += caam_jr.o
- obj-$(CONFIG_CRYPTO_DEV_FSL_CAAM_CRYPTO_API) += caamalg.o
-+obj-$(CONFIG_CRYPTO_DEV_FSL_CAAM_CRYPTO_API_QI) += caamalg_qi.o
-+obj-$(CONFIG_CRYPTO_DEV_FSL_CAAM_CRYPTO_API_DESC) += caamalg_desc.o
- obj-$(CONFIG_CRYPTO_DEV_FSL_CAAM_AHASH_API) += caamhash.o
-+obj-$(CONFIG_CRYPTO_DEV_FSL_CAAM_AHASH_API_DESC) += caamhash_desc.o
- obj-$(CONFIG_CRYPTO_DEV_FSL_CAAM_RNG_API) += caamrng.o
- obj-$(CONFIG_CRYPTO_DEV_FSL_CAAM_PKC_API) += caam_pkc.o
-
- caam-objs := ctrl.o
--caam_jr-objs := jr.o key_gen.o error.o
-+caam_jr-objs := jr.o key_gen.o
- caam_pkc-y := caampkc.o pkc_desc.o
-+ifneq ($(CONFIG_CRYPTO_DEV_FSL_CAAM_CRYPTO_API_QI),)
-+ ccflags-y += -DCONFIG_CAAM_QI
-+ caam-objs += qi.o
-+endif
-+
-+obj-$(CONFIG_CRYPTO_DEV_FSL_DPAA2_CAAM) += dpaa2_caam.o
-+
-+dpaa2_caam-y := caamalg_qi2.o dpseci.o
---- a/drivers/crypto/caam/caamalg.c
-+++ b/drivers/crypto/caam/caamalg.c
-@@ -2,6 +2,7 @@
- * caam - Freescale FSL CAAM support for crypto API
- *
- * Copyright 2008-2011 Freescale Semiconductor, Inc.
-+ * Copyright 2016 NXP
- *
- * Based on talitos crypto API driver.
- *
-@@ -53,6 +54,7 @@
- #include "error.h"
- #include "sg_sw_sec4.h"
- #include "key_gen.h"
-+#include "caamalg_desc.h"
-
- /*
- * crypto alg
-@@ -62,8 +64,6 @@
- #define CAAM_MAX_KEY_SIZE (AES_MAX_KEY_SIZE + \
- CTR_RFC3686_NONCE_SIZE + \
- SHA512_DIGEST_SIZE * 2)
--/* max IV is max of AES_BLOCK_SIZE, DES3_EDE_BLOCK_SIZE */
--#define CAAM_MAX_IV_LENGTH 16
-
- #define AEAD_DESC_JOB_IO_LEN (DESC_JOB_IO_LEN + CAAM_CMD_SZ * 2)
- #define GCM_DESC_JOB_IO_LEN (AEAD_DESC_JOB_IO_LEN + \
-@@ -71,37 +71,6 @@
- #define AUTHENC_DESC_JOB_IO_LEN (AEAD_DESC_JOB_IO_LEN + \
- CAAM_CMD_SZ * 5)
-
--/* length of descriptors text */
--#define DESC_AEAD_BASE (4 * CAAM_CMD_SZ)
--#define DESC_AEAD_ENC_LEN (DESC_AEAD_BASE + 11 * CAAM_CMD_SZ)
--#define DESC_AEAD_DEC_LEN (DESC_AEAD_BASE + 15 * CAAM_CMD_SZ)
--#define DESC_AEAD_GIVENC_LEN (DESC_AEAD_ENC_LEN + 9 * CAAM_CMD_SZ)
--
--/* Note: Nonce is counted in enckeylen */
--#define DESC_AEAD_CTR_RFC3686_LEN (4 * CAAM_CMD_SZ)
--
--#define DESC_AEAD_NULL_BASE (3 * CAAM_CMD_SZ)
--#define DESC_AEAD_NULL_ENC_LEN (DESC_AEAD_NULL_BASE + 11 * CAAM_CMD_SZ)
--#define DESC_AEAD_NULL_DEC_LEN (DESC_AEAD_NULL_BASE + 13 * CAAM_CMD_SZ)
--
--#define DESC_GCM_BASE (3 * CAAM_CMD_SZ)
--#define DESC_GCM_ENC_LEN (DESC_GCM_BASE + 16 * CAAM_CMD_SZ)
--#define DESC_GCM_DEC_LEN (DESC_GCM_BASE + 12 * CAAM_CMD_SZ)
--
--#define DESC_RFC4106_BASE (3 * CAAM_CMD_SZ)
--#define DESC_RFC4106_ENC_LEN (DESC_RFC4106_BASE + 13 * CAAM_CMD_SZ)
--#define DESC_RFC4106_DEC_LEN (DESC_RFC4106_BASE + 13 * CAAM_CMD_SZ)
--
--#define DESC_RFC4543_BASE (3 * CAAM_CMD_SZ)
--#define DESC_RFC4543_ENC_LEN (DESC_RFC4543_BASE + 11 * CAAM_CMD_SZ)
--#define DESC_RFC4543_DEC_LEN (DESC_RFC4543_BASE + 12 * CAAM_CMD_SZ)
--
--#define DESC_ABLKCIPHER_BASE (3 * CAAM_CMD_SZ)
--#define DESC_ABLKCIPHER_ENC_LEN (DESC_ABLKCIPHER_BASE + \
-- 20 * CAAM_CMD_SZ)
--#define DESC_ABLKCIPHER_DEC_LEN (DESC_ABLKCIPHER_BASE + \
-- 15 * CAAM_CMD_SZ)
--
- #define DESC_MAX_USED_BYTES (CAAM_DESC_BYTES_MAX - DESC_JOB_IO_LEN)
- #define DESC_MAX_USED_LEN (DESC_MAX_USED_BYTES / CAAM_CMD_SZ)
-
-@@ -112,47 +81,11 @@
- #define debug(format, arg...)
- #endif
-
--#ifdef DEBUG
--#include <linux/highmem.h>
--
--static void dbg_dump_sg(const char *level, const char *prefix_str,
-- int prefix_type, int rowsize, int groupsize,
-- struct scatterlist *sg, size_t tlen, bool ascii,
-- bool may_sleep)
--{
-- struct scatterlist *it;
-- void *it_page;
-- size_t len;
-- void *buf;
--
-- for (it = sg; it != NULL && tlen > 0 ; it = sg_next(sg)) {
-- /*
-- * make sure the scatterlist's page
-- * has a valid virtual memory mapping
-- */
-- it_page = kmap_atomic(sg_page(it));
-- if (unlikely(!it_page)) {
-- printk(KERN_ERR "dbg_dump_sg: kmap failed\n");
-- return;
-- }
--
-- buf = it_page + it->offset;
-- len = min_t(size_t, tlen, it->length);
-- print_hex_dump(level, prefix_str, prefix_type, rowsize,
-- groupsize, buf, len, ascii);
-- tlen -= len;
--
-- kunmap_atomic(it_page);
-- }
--}
--#endif
--
- static struct list_head alg_list;
-
- struct caam_alg_entry {
- int class1_alg_type;
- int class2_alg_type;
-- int alg_op;
- bool rfc3686;
- bool geniv;
- };
-@@ -163,302 +96,71 @@ struct caam_aead_alg {
- bool registered;
- };
-
--/* Set DK bit in class 1 operation if shared */
--static inline void append_dec_op1(u32 *desc, u32 type)
--{
-- u32 *jump_cmd, *uncond_jump_cmd;
--
-- /* DK bit is valid only for AES */
-- if ((type & OP_ALG_ALGSEL_MASK) != OP_ALG_ALGSEL_AES) {
-- append_operation(desc, type | OP_ALG_AS_INITFINAL |
-- OP_ALG_DECRYPT);
-- return;
-- }
--
-- jump_cmd = append_jump(desc, JUMP_TEST_ALL | JUMP_COND_SHRD);
-- append_operation(desc, type | OP_ALG_AS_INITFINAL |
-- OP_ALG_DECRYPT);
-- uncond_jump_cmd = append_jump(desc, JUMP_TEST_ALL);
-- set_jump_tgt_here(desc, jump_cmd);
-- append_operation(desc, type | OP_ALG_AS_INITFINAL |
-- OP_ALG_DECRYPT | OP_ALG_AAI_DK);
-- set_jump_tgt_here(desc, uncond_jump_cmd);
--}
--
--/*
-- * For aead functions, read payload and write payload,
-- * both of which are specified in req->src and req->dst
-- */
--static inline void aead_append_src_dst(u32 *desc, u32 msg_type)
--{
-- append_seq_fifo_store(desc, 0, FIFOST_TYPE_MESSAGE_DATA | KEY_VLF);
-- append_seq_fifo_load(desc, 0, FIFOLD_CLASS_BOTH |
-- KEY_VLF | msg_type | FIFOLD_TYPE_LASTBOTH);
--}
--
--/*
-- * For ablkcipher encrypt and decrypt, read from req->src and
-- * write to req->dst
-- */
--static inline void ablkcipher_append_src_dst(u32 *desc)
--{
-- append_math_add(desc, VARSEQOUTLEN, SEQINLEN, REG0, CAAM_CMD_SZ);
-- append_math_add(desc, VARSEQINLEN, SEQINLEN, REG0, CAAM_CMD_SZ);
-- append_seq_fifo_load(desc, 0, FIFOLD_CLASS_CLASS1 |
-- KEY_VLF | FIFOLD_TYPE_MSG | FIFOLD_TYPE_LAST1);
-- append_seq_fifo_store(desc, 0, FIFOST_TYPE_MESSAGE_DATA | KEY_VLF);
--}
--
- /*
- * per-session context
- */
- struct caam_ctx {
-- struct device *jrdev;
- u32 sh_desc_enc[DESC_MAX_USED_LEN];
- u32 sh_desc_dec[DESC_MAX_USED_LEN];
- u32 sh_desc_givenc[DESC_MAX_USED_LEN];
-+ u8 key[CAAM_MAX_KEY_SIZE];
- dma_addr_t sh_desc_enc_dma;
- dma_addr_t sh_desc_dec_dma;
- dma_addr_t sh_desc_givenc_dma;
-- u32 class1_alg_type;
-- u32 class2_alg_type;
-- u32 alg_op;
-- u8 key[CAAM_MAX_KEY_SIZE];
- dma_addr_t key_dma;
-- unsigned int enckeylen;
-- unsigned int split_key_len;
-- unsigned int split_key_pad_len;
-+ enum dma_data_direction dir;
-+ struct device *jrdev;
-+ struct alginfo adata;
-+ struct alginfo cdata;
- unsigned int authsize;
- };
-
--static void append_key_aead(u32 *desc, struct caam_ctx *ctx,
-- int keys_fit_inline, bool is_rfc3686)
--{
-- u32 *nonce;
-- unsigned int enckeylen = ctx->enckeylen;
--
-- /*
-- * RFC3686 specific:
-- * | ctx->key = {AUTH_KEY, ENC_KEY, NONCE}
-- * | enckeylen = encryption key size + nonce size
-- */
-- if (is_rfc3686)
-- enckeylen -= CTR_RFC3686_NONCE_SIZE;
--
-- if (keys_fit_inline) {
-- append_key_as_imm(desc, ctx->key, ctx->split_key_pad_len,
-- ctx->split_key_len, CLASS_2 |
-- KEY_DEST_MDHA_SPLIT | KEY_ENC);
-- append_key_as_imm(desc, (void *)ctx->key +
-- ctx->split_key_pad_len, enckeylen,
-- enckeylen, CLASS_1 | KEY_DEST_CLASS_REG);
-- } else {
-- append_key(desc, ctx->key_dma, ctx->split_key_len, CLASS_2 |
-- KEY_DEST_MDHA_SPLIT | KEY_ENC);
-- append_key(desc, ctx->key_dma + ctx->split_key_pad_len,
-- enckeylen, CLASS_1 | KEY_DEST_CLASS_REG);
-- }
--
-- /* Load Counter into CONTEXT1 reg */
-- if (is_rfc3686) {
-- nonce = (u32 *)((void *)ctx->key + ctx->split_key_pad_len +
-- enckeylen);
-- append_load_as_imm(desc, nonce, CTR_RFC3686_NONCE_SIZE,
-- LDST_CLASS_IND_CCB |
-- LDST_SRCDST_BYTE_OUTFIFO | LDST_IMM);
-- append_move(desc,
-- MOVE_SRC_OUTFIFO |
-- MOVE_DEST_CLASS1CTX |
-- (16 << MOVE_OFFSET_SHIFT) |
-- (CTR_RFC3686_NONCE_SIZE << MOVE_LEN_SHIFT));
-- }
--}
--
--static void init_sh_desc_key_aead(u32 *desc, struct caam_ctx *ctx,
-- int keys_fit_inline, bool is_rfc3686)
--{
-- u32 *key_jump_cmd;
--
-- /* Note: Context registers are saved. */
-- init_sh_desc(desc, HDR_SHARE_SERIAL | HDR_SAVECTX);
--
-- /* Skip if already shared */
-- key_jump_cmd = append_jump(desc, JUMP_JSL | JUMP_TEST_ALL |
-- JUMP_COND_SHRD);
--
-- append_key_aead(desc, ctx, keys_fit_inline, is_rfc3686);
--
-- set_jump_tgt_here(desc, key_jump_cmd);
--}
--
- static int aead_null_set_sh_desc(struct crypto_aead *aead)
- {
- struct caam_ctx *ctx = crypto_aead_ctx(aead);
- struct device *jrdev = ctx->jrdev;
-- bool keys_fit_inline = false;
-- u32 *key_jump_cmd, *jump_cmd, *read_move_cmd, *write_move_cmd;
-+ struct caam_drv_private *ctrlpriv = dev_get_drvdata(jrdev->parent);
- u32 *desc;
-+ int rem_bytes = CAAM_DESC_BYTES_MAX - AEAD_DESC_JOB_IO_LEN -
-+ ctx->adata.keylen_pad;
-
- /*
- * Job Descriptor and Shared Descriptors
- * must all fit into the 64-word Descriptor h/w Buffer
- */
-- if (DESC_AEAD_NULL_ENC_LEN + AEAD_DESC_JOB_IO_LEN +
-- ctx->split_key_pad_len <= CAAM_DESC_BYTES_MAX)
-- keys_fit_inline = true;
-+ if (rem_bytes >= DESC_AEAD_NULL_ENC_LEN) {
-+ ctx->adata.key_inline = true;
-+ ctx->adata.key_virt = ctx->key;
-+ } else {
-+ ctx->adata.key_inline = false;
-+ ctx->adata.key_dma = ctx->key_dma;
-+ }
-
- /* aead_encrypt shared descriptor */
- desc = ctx->sh_desc_enc;
--
-- init_sh_desc(desc, HDR_SHARE_SERIAL);
--
-- /* Skip if already shared */
-- key_jump_cmd = append_jump(desc, JUMP_JSL | JUMP_TEST_ALL |
-- JUMP_COND_SHRD);
-- if (keys_fit_inline)
-- append_key_as_imm(desc, ctx->key, ctx->split_key_pad_len,
-- ctx->split_key_len, CLASS_2 |
-- KEY_DEST_MDHA_SPLIT | KEY_ENC);
-- else
-- append_key(desc, ctx->key_dma, ctx->split_key_len, CLASS_2 |
-- KEY_DEST_MDHA_SPLIT | KEY_ENC);
-- set_jump_tgt_here(desc, key_jump_cmd);
--
-- /* assoclen + cryptlen = seqinlen */
-- append_math_sub(desc, REG3, SEQINLEN, REG0, CAAM_CMD_SZ);
--
-- /* Prepare to read and write cryptlen + assoclen bytes */
-- append_math_add(desc, VARSEQINLEN, ZERO, REG3, CAAM_CMD_SZ);
-- append_math_add(desc, VARSEQOUTLEN, ZERO, REG3, CAAM_CMD_SZ);
--
-- /*
-- * MOVE_LEN opcode is not available in all SEC HW revisions,
-- * thus need to do some magic, i.e. self-patch the descriptor
-- * buffer.
-- */
-- read_move_cmd = append_move(desc, MOVE_SRC_DESCBUF |
-- MOVE_DEST_MATH3 |
-- (0x6 << MOVE_LEN_SHIFT));
-- write_move_cmd = append_move(desc, MOVE_SRC_MATH3 |
-- MOVE_DEST_DESCBUF |
-- MOVE_WAITCOMP |
-- (0x8 << MOVE_LEN_SHIFT));
--
-- /* Class 2 operation */
-- append_operation(desc, ctx->class2_alg_type |
-- OP_ALG_AS_INITFINAL | OP_ALG_ENCRYPT);
--
-- /* Read and write cryptlen bytes */
-- aead_append_src_dst(desc, FIFOLD_TYPE_MSG | FIFOLD_TYPE_FLUSH1);
--
-- set_move_tgt_here(desc, read_move_cmd);
-- set_move_tgt_here(desc, write_move_cmd);
-- append_cmd(desc, CMD_LOAD | DISABLE_AUTO_INFO_FIFO);
-- append_move(desc, MOVE_SRC_INFIFO_CL | MOVE_DEST_OUTFIFO |
-- MOVE_AUX_LS);
--
-- /* Write ICV */
-- append_seq_store(desc, ctx->authsize, LDST_CLASS_2_CCB |
-- LDST_SRCDST_BYTE_CONTEXT);
--
-- ctx->sh_desc_enc_dma = dma_map_single(jrdev, desc,
-- desc_bytes(desc),
-- DMA_TO_DEVICE);
-- if (dma_mapping_error(jrdev, ctx->sh_desc_enc_dma)) {
-- dev_err(jrdev, "unable to map shared descriptor\n");
-- return -ENOMEM;
-- }
--#ifdef DEBUG
-- print_hex_dump(KERN_ERR,
-- "aead null enc shdesc@"__stringify(__LINE__)": ",
-- DUMP_PREFIX_ADDRESS, 16, 4, desc,
-- desc_bytes(desc), 1);
--#endif
-+ cnstr_shdsc_aead_null_encap(desc, &ctx->adata, ctx->authsize,
-+ ctrlpriv->era);
-+ dma_sync_single_for_device(jrdev, ctx->sh_desc_enc_dma,
-+ desc_bytes(desc), ctx->dir);
-
- /*
- * Job Descriptor and Shared Descriptors
- * must all fit into the 64-word Descriptor h/w Buffer
- */
-- keys_fit_inline = false;
-- if (DESC_AEAD_NULL_DEC_LEN + DESC_JOB_IO_LEN +
-- ctx->split_key_pad_len <= CAAM_DESC_BYTES_MAX)
-- keys_fit_inline = true;
--
-- desc = ctx->sh_desc_dec;
-+ if (rem_bytes >= DESC_AEAD_NULL_DEC_LEN) {
-+ ctx->adata.key_inline = true;
-+ ctx->adata.key_virt = ctx->key;
-+ } else {
-+ ctx->adata.key_inline = false;
-+ ctx->adata.key_dma = ctx->key_dma;
-+ }
-
- /* aead_decrypt shared descriptor */
-- init_sh_desc(desc, HDR_SHARE_SERIAL);
--
-- /* Skip if already shared */
-- key_jump_cmd = append_jump(desc, JUMP_JSL | JUMP_TEST_ALL |
-- JUMP_COND_SHRD);
-- if (keys_fit_inline)
-- append_key_as_imm(desc, ctx->key, ctx->split_key_pad_len,
-- ctx->split_key_len, CLASS_2 |
-- KEY_DEST_MDHA_SPLIT | KEY_ENC);
-- else
-- append_key(desc, ctx->key_dma, ctx->split_key_len, CLASS_2 |
-- KEY_DEST_MDHA_SPLIT | KEY_ENC);
-- set_jump_tgt_here(desc, key_jump_cmd);
--
-- /* Class 2 operation */
-- append_operation(desc, ctx->class2_alg_type |
-- OP_ALG_AS_INITFINAL | OP_ALG_DECRYPT | OP_ALG_ICV_ON);
--
-- /* assoclen + cryptlen = seqoutlen */
-- append_math_sub(desc, REG2, SEQOUTLEN, REG0, CAAM_CMD_SZ);
--
-- /* Prepare to read and write cryptlen + assoclen bytes */
-- append_math_add(desc, VARSEQINLEN, ZERO, REG2, CAAM_CMD_SZ);
-- append_math_add(desc, VARSEQOUTLEN, ZERO, REG2, CAAM_CMD_SZ);
--
-- /*
-- * MOVE_LEN opcode is not available in all SEC HW revisions,
-- * thus need to do some magic, i.e. self-patch the descriptor
-- * buffer.
-- */
-- read_move_cmd = append_move(desc, MOVE_SRC_DESCBUF |
-- MOVE_DEST_MATH2 |
-- (0x6 << MOVE_LEN_SHIFT));
-- write_move_cmd = append_move(desc, MOVE_SRC_MATH2 |
-- MOVE_DEST_DESCBUF |
-- MOVE_WAITCOMP |
-- (0x8 << MOVE_LEN_SHIFT));
--
-- /* Read and write cryptlen bytes */
-- aead_append_src_dst(desc, FIFOLD_TYPE_MSG | FIFOLD_TYPE_FLUSH1);
--
-- /*
-- * Insert a NOP here, since we need at least 4 instructions between
-- * code patching the descriptor buffer and the location being patched.
-- */
-- jump_cmd = append_jump(desc, JUMP_TEST_ALL);
-- set_jump_tgt_here(desc, jump_cmd);
--
-- set_move_tgt_here(desc, read_move_cmd);
-- set_move_tgt_here(desc, write_move_cmd);
-- append_cmd(desc, CMD_LOAD | DISABLE_AUTO_INFO_FIFO);
-- append_move(desc, MOVE_SRC_INFIFO_CL | MOVE_DEST_OUTFIFO |
-- MOVE_AUX_LS);
-- append_cmd(desc, CMD_LOAD | ENABLE_AUTO_INFO_FIFO);
--
-- /* Load ICV */
-- append_seq_fifo_load(desc, ctx->authsize, FIFOLD_CLASS_CLASS2 |
-- FIFOLD_TYPE_LAST2 | FIFOLD_TYPE_ICV);
--
-- ctx->sh_desc_dec_dma = dma_map_single(jrdev, desc,
-- desc_bytes(desc),
-- DMA_TO_DEVICE);
-- if (dma_mapping_error(jrdev, ctx->sh_desc_dec_dma)) {
-- dev_err(jrdev, "unable to map shared descriptor\n");
-- return -ENOMEM;
-- }
--#ifdef DEBUG
-- print_hex_dump(KERN_ERR,
-- "aead null dec shdesc@"__stringify(__LINE__)": ",
-- DUMP_PREFIX_ADDRESS, 16, 4, desc,
-- desc_bytes(desc), 1);
--#endif
-+ desc = ctx->sh_desc_dec;
-+ cnstr_shdsc_aead_null_decap(desc, &ctx->adata, ctx->authsize,
-+ ctrlpriv->era);
-+ dma_sync_single_for_device(jrdev, ctx->sh_desc_dec_dma,
-+ desc_bytes(desc), ctx->dir);
-
- return 0;
- }
-@@ -470,11 +172,12 @@ static int aead_set_sh_desc(struct crypt
- unsigned int ivsize = crypto_aead_ivsize(aead);
- struct caam_ctx *ctx = crypto_aead_ctx(aead);
- struct device *jrdev = ctx->jrdev;
-- bool keys_fit_inline;
-- u32 geniv, moveiv;
-+ struct caam_drv_private *ctrlpriv = dev_get_drvdata(jrdev->parent);
- u32 ctx1_iv_off = 0;
-- u32 *desc;
-- const bool ctr_mode = ((ctx->class1_alg_type & OP_ALG_AAI_MASK) ==
-+ u32 *desc, *nonce = NULL;
-+ u32 inl_mask;
-+ unsigned int data_len[2];
-+ const bool ctr_mode = ((ctx->cdata.algtype & OP_ALG_AAI_MASK) ==
- OP_ALG_AAI_CTR_MOD128);
- const bool is_rfc3686 = alg->caam.rfc3686;
-
-@@ -482,7 +185,7 @@ static int aead_set_sh_desc(struct crypt
- return 0;
-
- /* NULL encryption / decryption */
-- if (!ctx->enckeylen)
-+ if (!ctx->cdata.keylen)
- return aead_null_set_sh_desc(aead);
-
- /*
-@@ -497,8 +200,14 @@ static int aead_set_sh_desc(struct crypt
- * RFC3686 specific:
- * CONTEXT1[255:128] = {NONCE, IV, COUNTER}
- */
-- if (is_rfc3686)
-+ if (is_rfc3686) {
- ctx1_iv_off = 16 + CTR_RFC3686_NONCE_SIZE;
-+ nonce = (u32 *)((void *)ctx->key + ctx->adata.keylen_pad +
-+ ctx->cdata.keylen - CTR_RFC3686_NONCE_SIZE);
-+ }
-+
-+ data_len[0] = ctx->adata.keylen_pad;
-+ data_len[1] = ctx->cdata.keylen;
-
- if (alg->caam.geniv)
- goto skip_enc;
-@@ -507,146 +216,64 @@ static int aead_set_sh_desc(struct crypt
- * Job Descriptor and Shared Descriptors
- * must all fit into the 64-word Descriptor h/w Buffer
- */
-- keys_fit_inline = false;
-- if (DESC_AEAD_ENC_LEN + AUTHENC_DESC_JOB_IO_LEN +
-- ctx->split_key_pad_len + ctx->enckeylen +
-- (is_rfc3686 ? DESC_AEAD_CTR_RFC3686_LEN : 0) <=
-- CAAM_DESC_BYTES_MAX)
-- keys_fit_inline = true;
--
-- /* aead_encrypt shared descriptor */
-- desc = ctx->sh_desc_enc;
--
-- /* Note: Context registers are saved. */
-- init_sh_desc_key_aead(desc, ctx, keys_fit_inline, is_rfc3686);
--
-- /* Class 2 operation */
-- append_operation(desc, ctx->class2_alg_type |
-- OP_ALG_AS_INITFINAL | OP_ALG_ENCRYPT);
--
-- /* Read and write assoclen bytes */
-- append_math_add(desc, VARSEQINLEN, ZERO, REG3, CAAM_CMD_SZ);
-- append_math_add(desc, VARSEQOUTLEN, ZERO, REG3, CAAM_CMD_SZ);
-+ if (desc_inline_query(DESC_AEAD_ENC_LEN +
-+ (is_rfc3686 ? DESC_AEAD_CTR_RFC3686_LEN : 0),
-+ AUTHENC_DESC_JOB_IO_LEN, data_len, &inl_mask,
-+ ARRAY_SIZE(data_len)) < 0)
-+ return -EINVAL;
-
-- /* Skip assoc data */
-- append_seq_fifo_store(desc, 0, FIFOST_TYPE_SKIP | FIFOLDST_VLF);
-+ if (inl_mask & 1)
-+ ctx->adata.key_virt = ctx->key;
-+ else
-+ ctx->adata.key_dma = ctx->key_dma;
-
-- /* read assoc before reading payload */
-- append_seq_fifo_load(desc, 0, FIFOLD_CLASS_CLASS2 | FIFOLD_TYPE_MSG |
-- FIFOLDST_VLF);
-+ if (inl_mask & 2)
-+ ctx->cdata.key_virt = ctx->key + ctx->adata.keylen_pad;
-+ else
-+ ctx->cdata.key_dma = ctx->key_dma + ctx->adata.keylen_pad;
-
-- /* Load Counter into CONTEXT1 reg */
-- if (is_rfc3686)
-- append_load_imm_be32(desc, 1, LDST_IMM | LDST_CLASS_1_CCB |
-- LDST_SRCDST_BYTE_CONTEXT |
-- ((ctx1_iv_off + CTR_RFC3686_IV_SIZE) <<
-- LDST_OFFSET_SHIFT));
--
-- /* Class 1 operation */
-- append_operation(desc, ctx->class1_alg_type |
-- OP_ALG_AS_INITFINAL | OP_ALG_ENCRYPT);
--
-- /* Read and write cryptlen bytes */
-- append_math_add(desc, VARSEQINLEN, SEQINLEN, REG0, CAAM_CMD_SZ);
-- append_math_add(desc, VARSEQOUTLEN, SEQINLEN, REG0, CAAM_CMD_SZ);
-- aead_append_src_dst(desc, FIFOLD_TYPE_MSG1OUT2);
--
-- /* Write ICV */
-- append_seq_store(desc, ctx->authsize, LDST_CLASS_2_CCB |
-- LDST_SRCDST_BYTE_CONTEXT);
-+ ctx->adata.key_inline = !!(inl_mask & 1);
-+ ctx->cdata.key_inline = !!(inl_mask & 2);
-
-- ctx->sh_desc_enc_dma = dma_map_single(jrdev, desc,
-- desc_bytes(desc),
-- DMA_TO_DEVICE);
-- if (dma_mapping_error(jrdev, ctx->sh_desc_enc_dma)) {
-- dev_err(jrdev, "unable to map shared descriptor\n");
-- return -ENOMEM;
-- }
--#ifdef DEBUG
-- print_hex_dump(KERN_ERR, "aead enc shdesc@"__stringify(__LINE__)": ",
-- DUMP_PREFIX_ADDRESS, 16, 4, desc,
-- desc_bytes(desc), 1);
--#endif
-+ /* aead_encrypt shared descriptor */
-+ desc = ctx->sh_desc_enc;
-+ cnstr_shdsc_aead_encap(desc, &ctx->cdata, &ctx->adata, ivsize,
-+ ctx->authsize, is_rfc3686, nonce, ctx1_iv_off,
-+ false, ctrlpriv->era);
-+ dma_sync_single_for_device(jrdev, ctx->sh_desc_enc_dma,
-+ desc_bytes(desc), ctx->dir);
-
- skip_enc:
- /*
- * Job Descriptor and Shared Descriptors
- * must all fit into the 64-word Descriptor h/w Buffer
- */
-- keys_fit_inline = false;
-- if (DESC_AEAD_DEC_LEN + AUTHENC_DESC_JOB_IO_LEN +
-- ctx->split_key_pad_len + ctx->enckeylen +
-- (is_rfc3686 ? DESC_AEAD_CTR_RFC3686_LEN : 0) <=
-- CAAM_DESC_BYTES_MAX)
-- keys_fit_inline = true;
--
-- /* aead_decrypt shared descriptor */
-- desc = ctx->sh_desc_dec;
--
-- /* Note: Context registers are saved. */
-- init_sh_desc_key_aead(desc, ctx, keys_fit_inline, is_rfc3686);
--
-- /* Class 2 operation */
-- append_operation(desc, ctx->class2_alg_type |
-- OP_ALG_AS_INITFINAL | OP_ALG_DECRYPT | OP_ALG_ICV_ON);
-+ if (desc_inline_query(DESC_AEAD_DEC_LEN +
-+ (is_rfc3686 ? DESC_AEAD_CTR_RFC3686_LEN : 0),
-+ AUTHENC_DESC_JOB_IO_LEN, data_len, &inl_mask,
-+ ARRAY_SIZE(data_len)) < 0)
-+ return -EINVAL;
-
-- /* Read and write assoclen bytes */
-- append_math_add(desc, VARSEQINLEN, ZERO, REG3, CAAM_CMD_SZ);
-- if (alg->caam.geniv)
-- append_math_add_imm_u32(desc, VARSEQOUTLEN, REG3, IMM, ivsize);
-+ if (inl_mask & 1)
-+ ctx->adata.key_virt = ctx->key;
- else
-- append_math_add(desc, VARSEQOUTLEN, ZERO, REG3, CAAM_CMD_SZ);
--
-- /* Skip assoc data */
-- append_seq_fifo_store(desc, 0, FIFOST_TYPE_SKIP | FIFOLDST_VLF);
--
-- /* read assoc before reading payload */
-- append_seq_fifo_load(desc, 0, FIFOLD_CLASS_CLASS2 | FIFOLD_TYPE_MSG |
-- KEY_VLF);
--
-- if (alg->caam.geniv) {
-- append_seq_load(desc, ivsize, LDST_CLASS_1_CCB |
-- LDST_SRCDST_BYTE_CONTEXT |
-- (ctx1_iv_off << LDST_OFFSET_SHIFT));
-- append_move(desc, MOVE_SRC_CLASS1CTX | MOVE_DEST_CLASS2INFIFO |
-- (ctx1_iv_off << MOVE_OFFSET_SHIFT) | ivsize);
-- }
--
-- /* Load Counter into CONTEXT1 reg */
-- if (is_rfc3686)
-- append_load_imm_be32(desc, 1, LDST_IMM | LDST_CLASS_1_CCB |
-- LDST_SRCDST_BYTE_CONTEXT |
-- ((ctx1_iv_off + CTR_RFC3686_IV_SIZE) <<
-- LDST_OFFSET_SHIFT));
-+ ctx->adata.key_dma = ctx->key_dma;
-
-- /* Choose operation */
-- if (ctr_mode)
-- append_operation(desc, ctx->class1_alg_type |
-- OP_ALG_AS_INITFINAL | OP_ALG_DECRYPT);
-+ if (inl_mask & 2)
-+ ctx->cdata.key_virt = ctx->key + ctx->adata.keylen_pad;
- else
-- append_dec_op1(desc, ctx->class1_alg_type);
-+ ctx->cdata.key_dma = ctx->key_dma + ctx->adata.keylen_pad;
-
-- /* Read and write cryptlen bytes */
-- append_math_add(desc, VARSEQINLEN, SEQOUTLEN, REG0, CAAM_CMD_SZ);
-- append_math_add(desc, VARSEQOUTLEN, SEQOUTLEN, REG0, CAAM_CMD_SZ);
-- aead_append_src_dst(desc, FIFOLD_TYPE_MSG);
--
-- /* Load ICV */
-- append_seq_fifo_load(desc, ctx->authsize, FIFOLD_CLASS_CLASS2 |
-- FIFOLD_TYPE_LAST2 | FIFOLD_TYPE_ICV);
-+ ctx->adata.key_inline = !!(inl_mask & 1);
-+ ctx->cdata.key_inline = !!(inl_mask & 2);
-
-- ctx->sh_desc_dec_dma = dma_map_single(jrdev, desc,
-- desc_bytes(desc),
-- DMA_TO_DEVICE);
-- if (dma_mapping_error(jrdev, ctx->sh_desc_dec_dma)) {
-- dev_err(jrdev, "unable to map shared descriptor\n");
-- return -ENOMEM;
-- }
--#ifdef DEBUG
-- print_hex_dump(KERN_ERR, "aead dec shdesc@"__stringify(__LINE__)": ",
-- DUMP_PREFIX_ADDRESS, 16, 4, desc,
-- desc_bytes(desc), 1);
--#endif
-+ /* aead_decrypt shared descriptor */
-+ desc = ctx->sh_desc_dec;
-+ cnstr_shdsc_aead_decap(desc, &ctx->cdata, &ctx->adata, ivsize,
-+ ctx->authsize, alg->caam.geniv, is_rfc3686,
-+ nonce, ctx1_iv_off, false, ctrlpriv->era);
-+ dma_sync_single_for_device(jrdev, ctx->sh_desc_dec_dma,
-+ desc_bytes(desc), ctx->dir);
-
- if (!alg->caam.geniv)
- goto skip_givenc;
-@@ -655,107 +282,32 @@ skip_enc:
- * Job Descriptor and Shared Descriptors
- * must all fit into the 64-word Descriptor h/w Buffer
- */
-- keys_fit_inline = false;
-- if (DESC_AEAD_GIVENC_LEN + AUTHENC_DESC_JOB_IO_LEN +
-- ctx->split_key_pad_len + ctx->enckeylen +
-- (is_rfc3686 ? DESC_AEAD_CTR_RFC3686_LEN : 0) <=
-- CAAM_DESC_BYTES_MAX)
-- keys_fit_inline = true;
--
-- /* aead_givencrypt shared descriptor */
-- desc = ctx->sh_desc_enc;
--
-- /* Note: Context registers are saved. */
-- init_sh_desc_key_aead(desc, ctx, keys_fit_inline, is_rfc3686);
-+ if (desc_inline_query(DESC_AEAD_GIVENC_LEN +
-+ (is_rfc3686 ? DESC_AEAD_CTR_RFC3686_LEN : 0),
-+ AUTHENC_DESC_JOB_IO_LEN, data_len, &inl_mask,
-+ ARRAY_SIZE(data_len)) < 0)
-+ return -EINVAL;
-
-- if (is_rfc3686)
-- goto copy_iv;
-+ if (inl_mask & 1)
-+ ctx->adata.key_virt = ctx->key;
-+ else
-+ ctx->adata.key_dma = ctx->key_dma;
-
-- /* Generate IV */
-- geniv = NFIFOENTRY_STYPE_PAD | NFIFOENTRY_DEST_DECO |
-- NFIFOENTRY_DTYPE_MSG | NFIFOENTRY_LC1 |
-- NFIFOENTRY_PTYPE_RND | (ivsize << NFIFOENTRY_DLEN_SHIFT);
-- append_load_imm_u32(desc, geniv, LDST_CLASS_IND_CCB |
-- LDST_SRCDST_WORD_INFO_FIFO | LDST_IMM);
-- append_cmd(desc, CMD_LOAD | DISABLE_AUTO_INFO_FIFO);
-- append_move(desc, MOVE_WAITCOMP |
-- MOVE_SRC_INFIFO | MOVE_DEST_CLASS1CTX |
-- (ctx1_iv_off << MOVE_OFFSET_SHIFT) |
-- (ivsize << MOVE_LEN_SHIFT));
-- append_cmd(desc, CMD_LOAD | ENABLE_AUTO_INFO_FIFO);
--
--copy_iv:
-- /* Copy IV to class 1 context */
-- append_move(desc, MOVE_SRC_CLASS1CTX | MOVE_DEST_OUTFIFO |
-- (ctx1_iv_off << MOVE_OFFSET_SHIFT) |
-- (ivsize << MOVE_LEN_SHIFT));
--
-- /* Return to encryption */
-- append_operation(desc, ctx->class2_alg_type |
-- OP_ALG_AS_INITFINAL | OP_ALG_ENCRYPT);
--
-- /* Read and write assoclen bytes */
-- append_math_add(desc, VARSEQINLEN, ZERO, REG3, CAAM_CMD_SZ);
-- append_math_add(desc, VARSEQOUTLEN, ZERO, REG3, CAAM_CMD_SZ);
--
-- /* ivsize + cryptlen = seqoutlen - authsize */
-- append_math_sub_imm_u32(desc, REG3, SEQOUTLEN, IMM, ctx->authsize);
--
-- /* Skip assoc data */
-- append_seq_fifo_store(desc, 0, FIFOST_TYPE_SKIP | FIFOLDST_VLF);
--
-- /* read assoc before reading payload */
-- append_seq_fifo_load(desc, 0, FIFOLD_CLASS_CLASS2 | FIFOLD_TYPE_MSG |
-- KEY_VLF);
--
-- /* Copy iv from outfifo to class 2 fifo */
-- moveiv = NFIFOENTRY_STYPE_OFIFO | NFIFOENTRY_DEST_CLASS2 |
-- NFIFOENTRY_DTYPE_MSG | (ivsize << NFIFOENTRY_DLEN_SHIFT);
-- append_load_imm_u32(desc, moveiv, LDST_CLASS_IND_CCB |
-- LDST_SRCDST_WORD_INFO_FIFO | LDST_IMM);
-- append_load_imm_u32(desc, ivsize, LDST_CLASS_2_CCB |
-- LDST_SRCDST_WORD_DATASZ_REG | LDST_IMM);
-+ if (inl_mask & 2)
-+ ctx->cdata.key_virt = ctx->key + ctx->adata.keylen_pad;
-+ else
-+ ctx->cdata.key_dma = ctx->key_dma + ctx->adata.keylen_pad;
-
-- /* Load Counter into CONTEXT1 reg */
-- if (is_rfc3686)
-- append_load_imm_be32(desc, 1, LDST_IMM | LDST_CLASS_1_CCB |
-- LDST_SRCDST_BYTE_CONTEXT |
-- ((ctx1_iv_off + CTR_RFC3686_IV_SIZE) <<
-- LDST_OFFSET_SHIFT));
--
-- /* Class 1 operation */
-- append_operation(desc, ctx->class1_alg_type |
-- OP_ALG_AS_INITFINAL | OP_ALG_ENCRYPT);
--
-- /* Will write ivsize + cryptlen */
-- append_math_add(desc, VARSEQOUTLEN, SEQINLEN, REG0, CAAM_CMD_SZ);
--
-- /* Not need to reload iv */
-- append_seq_fifo_load(desc, ivsize,
-- FIFOLD_CLASS_SKIP);
--
-- /* Will read cryptlen */
-- append_math_add(desc, VARSEQINLEN, SEQINLEN, REG0, CAAM_CMD_SZ);
-- append_seq_fifo_load(desc, 0, FIFOLD_CLASS_BOTH | KEY_VLF |
-- FIFOLD_TYPE_MSG1OUT2 | FIFOLD_TYPE_LASTBOTH);
-- append_seq_fifo_store(desc, 0, FIFOST_TYPE_MESSAGE_DATA | KEY_VLF);
--
-- /* Write ICV */
-- append_seq_store(desc, ctx->authsize, LDST_CLASS_2_CCB |
-- LDST_SRCDST_BYTE_CONTEXT);
-+ ctx->adata.key_inline = !!(inl_mask & 1);
-+ ctx->cdata.key_inline = !!(inl_mask & 2);
-
-- ctx->sh_desc_enc_dma = dma_map_single(jrdev, desc,
-- desc_bytes(desc),
-- DMA_TO_DEVICE);
-- if (dma_mapping_error(jrdev, ctx->sh_desc_enc_dma)) {
-- dev_err(jrdev, "unable to map shared descriptor\n");
-- return -ENOMEM;
-- }
--#ifdef DEBUG
-- print_hex_dump(KERN_ERR, "aead givenc shdesc@"__stringify(__LINE__)": ",
-- DUMP_PREFIX_ADDRESS, 16, 4, desc,
-- desc_bytes(desc), 1);
--#endif
-+ /* aead_givencrypt shared descriptor */
-+ desc = ctx->sh_desc_enc;
-+ cnstr_shdsc_aead_givencap(desc, &ctx->cdata, &ctx->adata, ivsize,
-+ ctx->authsize, is_rfc3686, nonce,
-+ ctx1_iv_off, false, ctrlpriv->era);
-+ dma_sync_single_for_device(jrdev, ctx->sh_desc_enc_dma,
-+ desc_bytes(desc), ctx->dir);
-
- skip_givenc:
- return 0;
-@@ -776,12 +328,12 @@ static int gcm_set_sh_desc(struct crypto
- {
- struct caam_ctx *ctx = crypto_aead_ctx(aead);
- struct device *jrdev = ctx->jrdev;
-- bool keys_fit_inline = false;
-- u32 *key_jump_cmd, *zero_payload_jump_cmd,
-- *zero_assoc_jump_cmd1, *zero_assoc_jump_cmd2;
-+ unsigned int ivsize = crypto_aead_ivsize(aead);
- u32 *desc;
-+ int rem_bytes = CAAM_DESC_BYTES_MAX - GCM_DESC_JOB_IO_LEN -
-+ ctx->cdata.keylen;
-
-- if (!ctx->enckeylen || !ctx->authsize)
-+ if (!ctx->cdata.keylen || !ctx->authsize)
- return 0;
-
- /*
-@@ -789,175 +341,35 @@ static int gcm_set_sh_desc(struct crypto
- * Job Descriptor and Shared Descriptor
- * must fit into the 64-word Descriptor h/w Buffer
- */
-- if (DESC_GCM_ENC_LEN + GCM_DESC_JOB_IO_LEN +
-- ctx->enckeylen <= CAAM_DESC_BYTES_MAX)
-- keys_fit_inline = true;
-+ if (rem_bytes >= DESC_GCM_ENC_LEN) {
-+ ctx->cdata.key_inline = true;
-+ ctx->cdata.key_virt = ctx->key;
-+ } else {
-+ ctx->cdata.key_inline = false;
-+ ctx->cdata.key_dma = ctx->key_dma;
-+ }
-
- desc = ctx->sh_desc_enc;
--
-- init_sh_desc(desc, HDR_SHARE_SERIAL);
--
-- /* skip key loading if they are loaded due to sharing */
-- key_jump_cmd = append_jump(desc, JUMP_JSL | JUMP_TEST_ALL |
-- JUMP_COND_SHRD | JUMP_COND_SELF);
-- if (keys_fit_inline)
-- append_key_as_imm(desc, (void *)ctx->key, ctx->enckeylen,
-- ctx->enckeylen, CLASS_1 | KEY_DEST_CLASS_REG);
-- else
-- append_key(desc, ctx->key_dma, ctx->enckeylen,
-- CLASS_1 | KEY_DEST_CLASS_REG);
-- set_jump_tgt_here(desc, key_jump_cmd);
--
-- /* class 1 operation */
-- append_operation(desc, ctx->class1_alg_type |
-- OP_ALG_AS_INITFINAL | OP_ALG_ENCRYPT);
--
-- /* if assoclen + cryptlen is ZERO, skip to ICV write */
-- append_math_sub(desc, VARSEQOUTLEN, SEQINLEN, REG0, CAAM_CMD_SZ);
-- zero_assoc_jump_cmd2 = append_jump(desc, JUMP_TEST_ALL |
-- JUMP_COND_MATH_Z);
--
-- /* if assoclen is ZERO, skip reading the assoc data */
-- append_math_add(desc, VARSEQINLEN, ZERO, REG3, CAAM_CMD_SZ);
-- zero_assoc_jump_cmd1 = append_jump(desc, JUMP_TEST_ALL |
-- JUMP_COND_MATH_Z);
--
-- append_math_add(desc, VARSEQOUTLEN, ZERO, REG3, CAAM_CMD_SZ);
--
-- /* skip assoc data */
-- append_seq_fifo_store(desc, 0, FIFOST_TYPE_SKIP | FIFOLDST_VLF);
--
-- /* cryptlen = seqinlen - assoclen */
-- append_math_sub(desc, VARSEQOUTLEN, SEQINLEN, REG3, CAAM_CMD_SZ);
--
-- /* if cryptlen is ZERO jump to zero-payload commands */
-- zero_payload_jump_cmd = append_jump(desc, JUMP_TEST_ALL |
-- JUMP_COND_MATH_Z);
--
-- /* read assoc data */
-- append_seq_fifo_load(desc, 0, FIFOLD_CLASS_CLASS1 | FIFOLDST_VLF |
-- FIFOLD_TYPE_AAD | FIFOLD_TYPE_FLUSH1);
-- set_jump_tgt_here(desc, zero_assoc_jump_cmd1);
--
-- append_math_sub(desc, VARSEQINLEN, SEQINLEN, REG0, CAAM_CMD_SZ);
--
-- /* write encrypted data */
-- append_seq_fifo_store(desc, 0, FIFOST_TYPE_MESSAGE_DATA | FIFOLDST_VLF);
--
-- /* read payload data */
-- append_seq_fifo_load(desc, 0, FIFOLD_CLASS_CLASS1 | FIFOLDST_VLF |
-- FIFOLD_TYPE_MSG | FIFOLD_TYPE_LAST1);
--
-- /* jump the zero-payload commands */
-- append_jump(desc, JUMP_TEST_ALL | 2);
--
-- /* zero-payload commands */
-- set_jump_tgt_here(desc, zero_payload_jump_cmd);
--
-- /* read assoc data */
-- append_seq_fifo_load(desc, 0, FIFOLD_CLASS_CLASS1 | FIFOLDST_VLF |
-- FIFOLD_TYPE_AAD | FIFOLD_TYPE_LAST1);
--
-- /* There is no input data */
-- set_jump_tgt_here(desc, zero_assoc_jump_cmd2);
--
-- /* write ICV */
-- append_seq_store(desc, ctx->authsize, LDST_CLASS_1_CCB |
-- LDST_SRCDST_BYTE_CONTEXT);
--
-- ctx->sh_desc_enc_dma = dma_map_single(jrdev, desc,
-- desc_bytes(desc),
-- DMA_TO_DEVICE);
-- if (dma_mapping_error(jrdev, ctx->sh_desc_enc_dma)) {
-- dev_err(jrdev, "unable to map shared descriptor\n");
-- return -ENOMEM;
-- }
--#ifdef DEBUG
-- print_hex_dump(KERN_ERR, "gcm enc shdesc@"__stringify(__LINE__)": ",
-- DUMP_PREFIX_ADDRESS, 16, 4, desc,
-- desc_bytes(desc), 1);
--#endif
-+ cnstr_shdsc_gcm_encap(desc, &ctx->cdata, ivsize, ctx->authsize, false);
-+ dma_sync_single_for_device(jrdev, ctx->sh_desc_enc_dma,
-+ desc_bytes(desc), ctx->dir);
-
- /*
- * Job Descriptor and Shared Descriptors
- * must all fit into the 64-word Descriptor h/w Buffer
- */
-- keys_fit_inline = false;
-- if (DESC_GCM_DEC_LEN + GCM_DESC_JOB_IO_LEN +
-- ctx->enckeylen <= CAAM_DESC_BYTES_MAX)
-- keys_fit_inline = true;
-+ if (rem_bytes >= DESC_GCM_DEC_LEN) {
-+ ctx->cdata.key_inline = true;
-+ ctx->cdata.key_virt = ctx->key;
-+ } else {
-+ ctx->cdata.key_inline = false;
-+ ctx->cdata.key_dma = ctx->key_dma;
-+ }
-
- desc = ctx->sh_desc_dec;
--
-- init_sh_desc(desc, HDR_SHARE_SERIAL);
--
-- /* skip key loading if they are loaded due to sharing */
-- key_jump_cmd = append_jump(desc, JUMP_JSL |
-- JUMP_TEST_ALL | JUMP_COND_SHRD |
-- JUMP_COND_SELF);
-- if (keys_fit_inline)
-- append_key_as_imm(desc, (void *)ctx->key, ctx->enckeylen,
-- ctx->enckeylen, CLASS_1 | KEY_DEST_CLASS_REG);
-- else
-- append_key(desc, ctx->key_dma, ctx->enckeylen,
-- CLASS_1 | KEY_DEST_CLASS_REG);
-- set_jump_tgt_here(desc, key_jump_cmd);
--
-- /* class 1 operation */
-- append_operation(desc, ctx->class1_alg_type |
-- OP_ALG_AS_INITFINAL | OP_ALG_DECRYPT | OP_ALG_ICV_ON);
--
-- /* if assoclen is ZERO, skip reading the assoc data */
-- append_math_add(desc, VARSEQINLEN, ZERO, REG3, CAAM_CMD_SZ);
-- zero_assoc_jump_cmd1 = append_jump(desc, JUMP_TEST_ALL |
-- JUMP_COND_MATH_Z);
--
-- append_math_add(desc, VARSEQOUTLEN, ZERO, REG3, CAAM_CMD_SZ);
--
-- /* skip assoc data */
-- append_seq_fifo_store(desc, 0, FIFOST_TYPE_SKIP | FIFOLDST_VLF);
--
-- /* read assoc data */
-- append_seq_fifo_load(desc, 0, FIFOLD_CLASS_CLASS1 | FIFOLDST_VLF |
-- FIFOLD_TYPE_AAD | FIFOLD_TYPE_FLUSH1);
--
-- set_jump_tgt_here(desc, zero_assoc_jump_cmd1);
--
-- /* cryptlen = seqoutlen - assoclen */
-- append_math_sub(desc, VARSEQINLEN, SEQOUTLEN, REG0, CAAM_CMD_SZ);
--
-- /* jump to zero-payload command if cryptlen is zero */
-- zero_payload_jump_cmd = append_jump(desc, JUMP_TEST_ALL |
-- JUMP_COND_MATH_Z);
--
-- append_math_sub(desc, VARSEQOUTLEN, SEQOUTLEN, REG0, CAAM_CMD_SZ);
--
-- /* store encrypted data */
-- append_seq_fifo_store(desc, 0, FIFOST_TYPE_MESSAGE_DATA | FIFOLDST_VLF);
--
-- /* read payload data */
-- append_seq_fifo_load(desc, 0, FIFOLD_CLASS_CLASS1 | FIFOLDST_VLF |
-- FIFOLD_TYPE_MSG | FIFOLD_TYPE_FLUSH1);
--
-- /* zero-payload command */
-- set_jump_tgt_here(desc, zero_payload_jump_cmd);
--
-- /* read ICV */
-- append_seq_fifo_load(desc, ctx->authsize, FIFOLD_CLASS_CLASS1 |
-- FIFOLD_TYPE_ICV | FIFOLD_TYPE_LAST1);
--
-- ctx->sh_desc_dec_dma = dma_map_single(jrdev, desc,
-- desc_bytes(desc),
-- DMA_TO_DEVICE);
-- if (dma_mapping_error(jrdev, ctx->sh_desc_dec_dma)) {
-- dev_err(jrdev, "unable to map shared descriptor\n");
-- return -ENOMEM;
-- }
--#ifdef DEBUG
-- print_hex_dump(KERN_ERR, "gcm dec shdesc@"__stringify(__LINE__)": ",
-- DUMP_PREFIX_ADDRESS, 16, 4, desc,
-- desc_bytes(desc), 1);
--#endif
-+ cnstr_shdsc_gcm_decap(desc, &ctx->cdata, ivsize, ctx->authsize, false);
-+ dma_sync_single_for_device(jrdev, ctx->sh_desc_dec_dma,
-+ desc_bytes(desc), ctx->dir);
-
- return 0;
- }
-@@ -976,11 +388,12 @@ static int rfc4106_set_sh_desc(struct cr
- {
- struct caam_ctx *ctx = crypto_aead_ctx(aead);
- struct device *jrdev = ctx->jrdev;
-- bool keys_fit_inline = false;
-- u32 *key_jump_cmd;
-+ unsigned int ivsize = crypto_aead_ivsize(aead);
- u32 *desc;
-+ int rem_bytes = CAAM_DESC_BYTES_MAX - GCM_DESC_JOB_IO_LEN -
-+ ctx->cdata.keylen;
-
-- if (!ctx->enckeylen || !ctx->authsize)
-+ if (!ctx->cdata.keylen || !ctx->authsize)
- return 0;
-
- /*
-@@ -988,148 +401,37 @@ static int rfc4106_set_sh_desc(struct cr
- * Job Descriptor and Shared Descriptor
- * must fit into the 64-word Descriptor h/w Buffer
- */
-- if (DESC_RFC4106_ENC_LEN + GCM_DESC_JOB_IO_LEN +
-- ctx->enckeylen <= CAAM_DESC_BYTES_MAX)
-- keys_fit_inline = true;
-+ if (rem_bytes >= DESC_RFC4106_ENC_LEN) {
-+ ctx->cdata.key_inline = true;
-+ ctx->cdata.key_virt = ctx->key;
-+ } else {
-+ ctx->cdata.key_inline = false;
-+ ctx->cdata.key_dma = ctx->key_dma;
-+ }
-
- desc = ctx->sh_desc_enc;
--
-- init_sh_desc(desc, HDR_SHARE_SERIAL);
--
-- /* Skip key loading if it is loaded due to sharing */
-- key_jump_cmd = append_jump(desc, JUMP_JSL | JUMP_TEST_ALL |
-- JUMP_COND_SHRD);
-- if (keys_fit_inline)
-- append_key_as_imm(desc, (void *)ctx->key, ctx->enckeylen,
-- ctx->enckeylen, CLASS_1 | KEY_DEST_CLASS_REG);
-- else
-- append_key(desc, ctx->key_dma, ctx->enckeylen,
-- CLASS_1 | KEY_DEST_CLASS_REG);
-- set_jump_tgt_here(desc, key_jump_cmd);
--
-- /* Class 1 operation */
-- append_operation(desc, ctx->class1_alg_type |
-- OP_ALG_AS_INITFINAL | OP_ALG_ENCRYPT);
--
-- append_math_sub_imm_u32(desc, VARSEQINLEN, REG3, IMM, 8);
-- append_math_add(desc, VARSEQOUTLEN, ZERO, REG3, CAAM_CMD_SZ);
--
-- /* Read assoc data */
-- append_seq_fifo_load(desc, 0, FIFOLD_CLASS_CLASS1 | FIFOLDST_VLF |
-- FIFOLD_TYPE_AAD | FIFOLD_TYPE_FLUSH1);
--
-- /* Skip IV */
-- append_seq_fifo_load(desc, 8, FIFOLD_CLASS_SKIP);
--
-- /* Will read cryptlen bytes */
-- append_math_sub(desc, VARSEQINLEN, SEQINLEN, REG0, CAAM_CMD_SZ);
--
-- /* Workaround for erratum A-005473 (simultaneous SEQ FIFO skips) */
-- append_seq_fifo_load(desc, 0, FIFOLD_CLASS_CLASS1 | FIFOLD_TYPE_MSG);
--
-- /* Skip assoc data */
-- append_seq_fifo_store(desc, 0, FIFOST_TYPE_SKIP | FIFOLDST_VLF);
--
-- /* cryptlen = seqoutlen - assoclen */
-- append_math_sub(desc, VARSEQOUTLEN, VARSEQINLEN, REG0, CAAM_CMD_SZ);
--
-- /* Write encrypted data */
-- append_seq_fifo_store(desc, 0, FIFOST_TYPE_MESSAGE_DATA | FIFOLDST_VLF);
--
-- /* Read payload data */
-- append_seq_fifo_load(desc, 0, FIFOLD_CLASS_CLASS1 | FIFOLDST_VLF |
-- FIFOLD_TYPE_MSG | FIFOLD_TYPE_LAST1);
--
-- /* Write ICV */
-- append_seq_store(desc, ctx->authsize, LDST_CLASS_1_CCB |
-- LDST_SRCDST_BYTE_CONTEXT);
--
-- ctx->sh_desc_enc_dma = dma_map_single(jrdev, desc,
-- desc_bytes(desc),
-- DMA_TO_DEVICE);
-- if (dma_mapping_error(jrdev, ctx->sh_desc_enc_dma)) {
-- dev_err(jrdev, "unable to map shared descriptor\n");
-- return -ENOMEM;
-- }
--#ifdef DEBUG
-- print_hex_dump(KERN_ERR, "rfc4106 enc shdesc@"__stringify(__LINE__)": ",
-- DUMP_PREFIX_ADDRESS, 16, 4, desc,
-- desc_bytes(desc), 1);
--#endif
-+ cnstr_shdsc_rfc4106_encap(desc, &ctx->cdata, ivsize, ctx->authsize,
-+ false);
-+ dma_sync_single_for_device(jrdev, ctx->sh_desc_enc_dma,
-+ desc_bytes(desc), ctx->dir);
-
- /*
- * Job Descriptor and Shared Descriptors
- * must all fit into the 64-word Descriptor h/w Buffer
- */
-- keys_fit_inline = false;
-- if (DESC_RFC4106_DEC_LEN + DESC_JOB_IO_LEN +
-- ctx->enckeylen <= CAAM_DESC_BYTES_MAX)
-- keys_fit_inline = true;
-+ if (rem_bytes >= DESC_RFC4106_DEC_LEN) {
-+ ctx->cdata.key_inline = true;
-+ ctx->cdata.key_virt = ctx->key;
-+ } else {
-+ ctx->cdata.key_inline = false;
-+ ctx->cdata.key_dma = ctx->key_dma;
-+ }
-
- desc = ctx->sh_desc_dec;
--
-- init_sh_desc(desc, HDR_SHARE_SERIAL);
--
-- /* Skip key loading if it is loaded due to sharing */
-- key_jump_cmd = append_jump(desc, JUMP_JSL |
-- JUMP_TEST_ALL | JUMP_COND_SHRD);
-- if (keys_fit_inline)
-- append_key_as_imm(desc, (void *)ctx->key, ctx->enckeylen,
-- ctx->enckeylen, CLASS_1 | KEY_DEST_CLASS_REG);
-- else
-- append_key(desc, ctx->key_dma, ctx->enckeylen,
-- CLASS_1 | KEY_DEST_CLASS_REG);
-- set_jump_tgt_here(desc, key_jump_cmd);
--
-- /* Class 1 operation */
-- append_operation(desc, ctx->class1_alg_type |
-- OP_ALG_AS_INITFINAL | OP_ALG_DECRYPT | OP_ALG_ICV_ON);
--
-- append_math_sub_imm_u32(desc, VARSEQINLEN, REG3, IMM, 8);
-- append_math_add(desc, VARSEQOUTLEN, ZERO, REG3, CAAM_CMD_SZ);
--
-- /* Read assoc data */
-- append_seq_fifo_load(desc, 0, FIFOLD_CLASS_CLASS1 | FIFOLDST_VLF |
-- FIFOLD_TYPE_AAD | FIFOLD_TYPE_FLUSH1);
--
-- /* Skip IV */
-- append_seq_fifo_load(desc, 8, FIFOLD_CLASS_SKIP);
--
-- /* Will read cryptlen bytes */
-- append_math_sub(desc, VARSEQINLEN, SEQOUTLEN, REG3, CAAM_CMD_SZ);
--
-- /* Workaround for erratum A-005473 (simultaneous SEQ FIFO skips) */
-- append_seq_fifo_load(desc, 0, FIFOLD_CLASS_CLASS1 | FIFOLD_TYPE_MSG);
--
-- /* Skip assoc data */
-- append_seq_fifo_store(desc, 0, FIFOST_TYPE_SKIP | FIFOLDST_VLF);
--
-- /* Will write cryptlen bytes */
-- append_math_sub(desc, VARSEQOUTLEN, SEQOUTLEN, REG0, CAAM_CMD_SZ);
--
-- /* Store payload data */
-- append_seq_fifo_store(desc, 0, FIFOST_TYPE_MESSAGE_DATA | FIFOLDST_VLF);
--
-- /* Read encrypted data */
-- append_seq_fifo_load(desc, 0, FIFOLD_CLASS_CLASS1 | FIFOLDST_VLF |
-- FIFOLD_TYPE_MSG | FIFOLD_TYPE_FLUSH1);
--
-- /* Read ICV */
-- append_seq_fifo_load(desc, ctx->authsize, FIFOLD_CLASS_CLASS1 |
-- FIFOLD_TYPE_ICV | FIFOLD_TYPE_LAST1);
--
-- ctx->sh_desc_dec_dma = dma_map_single(jrdev, desc,
-- desc_bytes(desc),
-- DMA_TO_DEVICE);
-- if (dma_mapping_error(jrdev, ctx->sh_desc_dec_dma)) {
-- dev_err(jrdev, "unable to map shared descriptor\n");
-- return -ENOMEM;
-- }
--#ifdef DEBUG
-- print_hex_dump(KERN_ERR, "rfc4106 dec shdesc@"__stringify(__LINE__)": ",
-- DUMP_PREFIX_ADDRESS, 16, 4, desc,
-- desc_bytes(desc), 1);
--#endif
-+ cnstr_shdsc_rfc4106_decap(desc, &ctx->cdata, ivsize, ctx->authsize,
-+ false);
-+ dma_sync_single_for_device(jrdev, ctx->sh_desc_dec_dma,
-+ desc_bytes(desc), ctx->dir);
-
- return 0;
- }
-@@ -1149,12 +451,12 @@ static int rfc4543_set_sh_desc(struct cr
- {
- struct caam_ctx *ctx = crypto_aead_ctx(aead);
- struct device *jrdev = ctx->jrdev;
-- bool keys_fit_inline = false;
-- u32 *key_jump_cmd;
-- u32 *read_move_cmd, *write_move_cmd;
-+ unsigned int ivsize = crypto_aead_ivsize(aead);
- u32 *desc;
-+ int rem_bytes = CAAM_DESC_BYTES_MAX - GCM_DESC_JOB_IO_LEN -
-+ ctx->cdata.keylen;
-
-- if (!ctx->enckeylen || !ctx->authsize)
-+ if (!ctx->cdata.keylen || !ctx->authsize)
- return 0;
-
- /*
-@@ -1162,151 +464,37 @@ static int rfc4543_set_sh_desc(struct cr
- * Job Descriptor and Shared Descriptor
- * must fit into the 64-word Descriptor h/w Buffer
- */
-- if (DESC_RFC4543_ENC_LEN + GCM_DESC_JOB_IO_LEN +
-- ctx->enckeylen <= CAAM_DESC_BYTES_MAX)
-- keys_fit_inline = true;
-+ if (rem_bytes >= DESC_RFC4543_ENC_LEN) {
-+ ctx->cdata.key_inline = true;
-+ ctx->cdata.key_virt = ctx->key;
-+ } else {
-+ ctx->cdata.key_inline = false;
-+ ctx->cdata.key_dma = ctx->key_dma;
-+ }
-
- desc = ctx->sh_desc_enc;
--
-- init_sh_desc(desc, HDR_SHARE_SERIAL);
--
-- /* Skip key loading if it is loaded due to sharing */
-- key_jump_cmd = append_jump(desc, JUMP_JSL | JUMP_TEST_ALL |
-- JUMP_COND_SHRD);
-- if (keys_fit_inline)
-- append_key_as_imm(desc, (void *)ctx->key, ctx->enckeylen,
-- ctx->enckeylen, CLASS_1 | KEY_DEST_CLASS_REG);
-- else
-- append_key(desc, ctx->key_dma, ctx->enckeylen,
-- CLASS_1 | KEY_DEST_CLASS_REG);
-- set_jump_tgt_here(desc, key_jump_cmd);
--
-- /* Class 1 operation */
-- append_operation(desc, ctx->class1_alg_type |
-- OP_ALG_AS_INITFINAL | OP_ALG_ENCRYPT);
--
-- /* assoclen + cryptlen = seqinlen */
-- append_math_sub(desc, REG3, SEQINLEN, REG0, CAAM_CMD_SZ);
--
-- /*
-- * MOVE_LEN opcode is not available in all SEC HW revisions,
-- * thus need to do some magic, i.e. self-patch the descriptor
-- * buffer.
-- */
-- read_move_cmd = append_move(desc, MOVE_SRC_DESCBUF | MOVE_DEST_MATH3 |
-- (0x6 << MOVE_LEN_SHIFT));
-- write_move_cmd = append_move(desc, MOVE_SRC_MATH3 | MOVE_DEST_DESCBUF |
-- (0x8 << MOVE_LEN_SHIFT));
--
-- /* Will read assoclen + cryptlen bytes */
-- append_math_sub(desc, VARSEQINLEN, SEQINLEN, REG0, CAAM_CMD_SZ);
--
-- /* Will write assoclen + cryptlen bytes */
-- append_math_sub(desc, VARSEQOUTLEN, SEQINLEN, REG0, CAAM_CMD_SZ);
--
-- /* Read and write assoclen + cryptlen bytes */
-- aead_append_src_dst(desc, FIFOLD_TYPE_AAD);
--
-- set_move_tgt_here(desc, read_move_cmd);
-- set_move_tgt_here(desc, write_move_cmd);
-- append_cmd(desc, CMD_LOAD | DISABLE_AUTO_INFO_FIFO);
-- /* Move payload data to OFIFO */
-- append_move(desc, MOVE_SRC_INFIFO_CL | MOVE_DEST_OUTFIFO);
--
-- /* Write ICV */
-- append_seq_store(desc, ctx->authsize, LDST_CLASS_1_CCB |
-- LDST_SRCDST_BYTE_CONTEXT);
--
-- ctx->sh_desc_enc_dma = dma_map_single(jrdev, desc,
-- desc_bytes(desc),
-- DMA_TO_DEVICE);
-- if (dma_mapping_error(jrdev, ctx->sh_desc_enc_dma)) {
-- dev_err(jrdev, "unable to map shared descriptor\n");
-- return -ENOMEM;
-- }
--#ifdef DEBUG
-- print_hex_dump(KERN_ERR, "rfc4543 enc shdesc@"__stringify(__LINE__)": ",
-- DUMP_PREFIX_ADDRESS, 16, 4, desc,
-- desc_bytes(desc), 1);
--#endif
-+ cnstr_shdsc_rfc4543_encap(desc, &ctx->cdata, ivsize, ctx->authsize,
-+ false);
-+ dma_sync_single_for_device(jrdev, ctx->sh_desc_enc_dma,
-+ desc_bytes(desc), ctx->dir);
-
- /*
- * Job Descriptor and Shared Descriptors
- * must all fit into the 64-word Descriptor h/w Buffer
- */
-- keys_fit_inline = false;
-- if (DESC_RFC4543_DEC_LEN + GCM_DESC_JOB_IO_LEN +
-- ctx->enckeylen <= CAAM_DESC_BYTES_MAX)
-- keys_fit_inline = true;
-+ if (rem_bytes >= DESC_RFC4543_DEC_LEN) {
-+ ctx->cdata.key_inline = true;
-+ ctx->cdata.key_virt = ctx->key;
-+ } else {
-+ ctx->cdata.key_inline = false;
-+ ctx->cdata.key_dma = ctx->key_dma;
-+ }
-
- desc = ctx->sh_desc_dec;
--
-- init_sh_desc(desc, HDR_SHARE_SERIAL);
--
-- /* Skip key loading if it is loaded due to sharing */
-- key_jump_cmd = append_jump(desc, JUMP_JSL |
-- JUMP_TEST_ALL | JUMP_COND_SHRD);
-- if (keys_fit_inline)
-- append_key_as_imm(desc, (void *)ctx->key, ctx->enckeylen,
-- ctx->enckeylen, CLASS_1 | KEY_DEST_CLASS_REG);
-- else
-- append_key(desc, ctx->key_dma, ctx->enckeylen,
-- CLASS_1 | KEY_DEST_CLASS_REG);
-- set_jump_tgt_here(desc, key_jump_cmd);
--
-- /* Class 1 operation */
-- append_operation(desc, ctx->class1_alg_type |
-- OP_ALG_AS_INITFINAL | OP_ALG_DECRYPT | OP_ALG_ICV_ON);
--
-- /* assoclen + cryptlen = seqoutlen */
-- append_math_sub(desc, REG3, SEQOUTLEN, REG0, CAAM_CMD_SZ);
--
-- /*
-- * MOVE_LEN opcode is not available in all SEC HW revisions,
-- * thus need to do some magic, i.e. self-patch the descriptor
-- * buffer.
-- */
-- read_move_cmd = append_move(desc, MOVE_SRC_DESCBUF | MOVE_DEST_MATH3 |
-- (0x6 << MOVE_LEN_SHIFT));
-- write_move_cmd = append_move(desc, MOVE_SRC_MATH3 | MOVE_DEST_DESCBUF |
-- (0x8 << MOVE_LEN_SHIFT));
--
-- /* Will read assoclen + cryptlen bytes */
-- append_math_sub(desc, VARSEQINLEN, SEQOUTLEN, REG0, CAAM_CMD_SZ);
--
-- /* Will write assoclen + cryptlen bytes */
-- append_math_sub(desc, VARSEQOUTLEN, SEQOUTLEN, REG0, CAAM_CMD_SZ);
--
-- /* Store payload data */
-- append_seq_fifo_store(desc, 0, FIFOST_TYPE_MESSAGE_DATA | FIFOLDST_VLF);
--
-- /* In-snoop assoclen + cryptlen data */
-- append_seq_fifo_load(desc, 0, FIFOLD_CLASS_BOTH | FIFOLDST_VLF |
-- FIFOLD_TYPE_AAD | FIFOLD_TYPE_LAST2FLUSH1);
--
-- set_move_tgt_here(desc, read_move_cmd);
-- set_move_tgt_here(desc, write_move_cmd);
-- append_cmd(desc, CMD_LOAD | DISABLE_AUTO_INFO_FIFO);
-- /* Move payload data to OFIFO */
-- append_move(desc, MOVE_SRC_INFIFO_CL | MOVE_DEST_OUTFIFO);
-- append_cmd(desc, CMD_LOAD | ENABLE_AUTO_INFO_FIFO);
--
-- /* Read ICV */
-- append_seq_fifo_load(desc, ctx->authsize, FIFOLD_CLASS_CLASS1 |
-- FIFOLD_TYPE_ICV | FIFOLD_TYPE_LAST1);
--
-- ctx->sh_desc_dec_dma = dma_map_single(jrdev, desc,
-- desc_bytes(desc),
-- DMA_TO_DEVICE);
-- if (dma_mapping_error(jrdev, ctx->sh_desc_dec_dma)) {
-- dev_err(jrdev, "unable to map shared descriptor\n");
-- return -ENOMEM;
-- }
--#ifdef DEBUG
-- print_hex_dump(KERN_ERR, "rfc4543 dec shdesc@"__stringify(__LINE__)": ",
-- DUMP_PREFIX_ADDRESS, 16, 4, desc,
-- desc_bytes(desc), 1);
--#endif
-+ cnstr_shdsc_rfc4543_decap(desc, &ctx->cdata, ivsize, ctx->authsize,
-+ false);
-+ dma_sync_single_for_device(jrdev, ctx->sh_desc_dec_dma,
-+ desc_bytes(desc), ctx->dir);
-
- return 0;
- }
-@@ -1322,74 +510,67 @@ static int rfc4543_setauthsize(struct cr
- return 0;
- }
-
--static u32 gen_split_aead_key(struct caam_ctx *ctx, const u8 *key_in,
-- u32 authkeylen)
--{
-- return gen_split_key(ctx->jrdev, ctx->key, ctx->split_key_len,
-- ctx->split_key_pad_len, key_in, authkeylen,
-- ctx->alg_op);
--}
--
- static int aead_setkey(struct crypto_aead *aead,
- const u8 *key, unsigned int keylen)
- {
-- /* Sizes for MDHA pads (*not* keys): MD5, SHA1, 224, 256, 384, 512 */
-- static const u8 mdpadlen[] = { 16, 20, 32, 32, 64, 64 };
- struct caam_ctx *ctx = crypto_aead_ctx(aead);
- struct device *jrdev = ctx->jrdev;
-+ struct caam_drv_private *ctrlpriv = dev_get_drvdata(jrdev->parent);
- struct crypto_authenc_keys keys;
- int ret = 0;
-
- if (crypto_authenc_extractkeys(&keys, key, keylen) != 0)
- goto badkey;
-
-- /* Pick class 2 key length from algorithm submask */
-- ctx->split_key_len = mdpadlen[(ctx->alg_op & OP_ALG_ALGSEL_SUBMASK) >>
-- OP_ALG_ALGSEL_SHIFT] * 2;
-- ctx->split_key_pad_len = ALIGN(ctx->split_key_len, 16);
--
-- if (ctx->split_key_pad_len + keys.enckeylen > CAAM_MAX_KEY_SIZE)
-- goto badkey;
--
- #ifdef DEBUG
- printk(KERN_ERR "keylen %d enckeylen %d authkeylen %d\n",
- keys.authkeylen + keys.enckeylen, keys.enckeylen,
- keys.authkeylen);
-- printk(KERN_ERR "split_key_len %d split_key_pad_len %d\n",
-- ctx->split_key_len, ctx->split_key_pad_len);
- print_hex_dump(KERN_ERR, "key in @"__stringify(__LINE__)": ",
- DUMP_PREFIX_ADDRESS, 16, 4, key, keylen, 1);
- #endif
-
-- ret = gen_split_aead_key(ctx, keys.authkey, keys.authkeylen);
-+ /*
-+ * If DKP is supported, use it in the shared descriptor to generate
-+ * the split key.
-+ */
-+ if (ctrlpriv->era >= 6) {
-+ ctx->adata.keylen = keys.authkeylen;
-+ ctx->adata.keylen_pad = split_key_len(ctx->adata.algtype &
-+ OP_ALG_ALGSEL_MASK);
-+
-+ if (ctx->adata.keylen_pad + keys.enckeylen > CAAM_MAX_KEY_SIZE)
-+ goto badkey;
-+
-+ memcpy(ctx->key, keys.authkey, keys.authkeylen);
-+ memcpy(ctx->key + ctx->adata.keylen_pad, keys.enckey,
-+ keys.enckeylen);
-+ dma_sync_single_for_device(jrdev, ctx->key_dma,
-+ ctx->adata.keylen_pad +
-+ keys.enckeylen, ctx->dir);
-+ goto skip_split_key;
-+ }
-+
-+ ret = gen_split_key(ctx->jrdev, ctx->key, &ctx->adata, keys.authkey,
-+ keys.authkeylen, CAAM_MAX_KEY_SIZE -
-+ keys.enckeylen);
- if (ret) {
- goto badkey;
- }
-
- /* postpend encryption key to auth split key */
-- memcpy(ctx->key + ctx->split_key_pad_len, keys.enckey, keys.enckeylen);
--
-- ctx->key_dma = dma_map_single(jrdev, ctx->key, ctx->split_key_pad_len +
-- keys.enckeylen, DMA_TO_DEVICE);
-- if (dma_mapping_error(jrdev, ctx->key_dma)) {
-- dev_err(jrdev, "unable to map key i/o memory\n");
-- return -ENOMEM;
-- }
-+ memcpy(ctx->key + ctx->adata.keylen_pad, keys.enckey, keys.enckeylen);
-+ dma_sync_single_for_device(jrdev, ctx->key_dma, ctx->adata.keylen_pad +
-+ keys.enckeylen, ctx->dir);
- #ifdef DEBUG
- print_hex_dump(KERN_ERR, "ctx.key@"__stringify(__LINE__)": ",
- DUMP_PREFIX_ADDRESS, 16, 4, ctx->key,
-- ctx->split_key_pad_len + keys.enckeylen, 1);
-+ ctx->adata.keylen_pad + keys.enckeylen, 1);
- #endif
-
-- ctx->enckeylen = keys.enckeylen;
--
-- ret = aead_set_sh_desc(aead);
-- if (ret) {
-- dma_unmap_single(jrdev, ctx->key_dma, ctx->split_key_pad_len +
-- keys.enckeylen, DMA_TO_DEVICE);
-- }
--
-- return ret;
-+skip_split_key:
-+ ctx->cdata.keylen = keys.enckeylen;
-+ return aead_set_sh_desc(aead);
- badkey:
- crypto_aead_set_flags(aead, CRYPTO_TFM_RES_BAD_KEY_LEN);
- return -EINVAL;
-@@ -1400,7 +581,6 @@ static int gcm_setkey(struct crypto_aead
- {
- struct caam_ctx *ctx = crypto_aead_ctx(aead);
- struct device *jrdev = ctx->jrdev;
-- int ret = 0;
-
- #ifdef DEBUG
- print_hex_dump(KERN_ERR, "key in @"__stringify(__LINE__)": ",
-@@ -1408,21 +588,10 @@ static int gcm_setkey(struct crypto_aead
- #endif
-
- memcpy(ctx->key, key, keylen);
-- ctx->key_dma = dma_map_single(jrdev, ctx->key, keylen,
-- DMA_TO_DEVICE);
-- if (dma_mapping_error(jrdev, ctx->key_dma)) {
-- dev_err(jrdev, "unable to map key i/o memory\n");
-- return -ENOMEM;
-- }
-- ctx->enckeylen = keylen;
-+ dma_sync_single_for_device(jrdev, ctx->key_dma, keylen, ctx->dir);
-+ ctx->cdata.keylen = keylen;
-
-- ret = gcm_set_sh_desc(aead);
-- if (ret) {
-- dma_unmap_single(jrdev, ctx->key_dma, ctx->enckeylen,
-- DMA_TO_DEVICE);
-- }
--
-- return ret;
-+ return gcm_set_sh_desc(aead);
- }
-
- static int rfc4106_setkey(struct crypto_aead *aead,
-@@ -1430,7 +599,6 @@ static int rfc4106_setkey(struct crypto_
- {
- struct caam_ctx *ctx = crypto_aead_ctx(aead);
- struct device *jrdev = ctx->jrdev;
-- int ret = 0;
-
- if (keylen < 4)
- return -EINVAL;
-@@ -1446,22 +614,10 @@ static int rfc4106_setkey(struct crypto_
- * The last four bytes of the key material are used as the salt value
- * in the nonce. Update the AES key length.
- */
-- ctx->enckeylen = keylen - 4;
--
-- ctx->key_dma = dma_map_single(jrdev, ctx->key, ctx->enckeylen,
-- DMA_TO_DEVICE);
-- if (dma_mapping_error(jrdev, ctx->key_dma)) {
-- dev_err(jrdev, "unable to map key i/o memory\n");
-- return -ENOMEM;
-- }
--
-- ret = rfc4106_set_sh_desc(aead);
-- if (ret) {
-- dma_unmap_single(jrdev, ctx->key_dma, ctx->enckeylen,
-- DMA_TO_DEVICE);
-- }
--
-- return ret;
-+ ctx->cdata.keylen = keylen - 4;
-+ dma_sync_single_for_device(jrdev, ctx->key_dma, ctx->cdata.keylen,
-+ ctx->dir);
-+ return rfc4106_set_sh_desc(aead);
- }
-
- static int rfc4543_setkey(struct crypto_aead *aead,
-@@ -1469,7 +625,6 @@ static int rfc4543_setkey(struct crypto_
- {
- struct caam_ctx *ctx = crypto_aead_ctx(aead);
- struct device *jrdev = ctx->jrdev;
-- int ret = 0;
-
- if (keylen < 4)
- return -EINVAL;
-@@ -1485,43 +640,28 @@ static int rfc4543_setkey(struct crypto_
- * The last four bytes of the key material are used as the salt value
- * in the nonce. Update the AES key length.
- */
-- ctx->enckeylen = keylen - 4;
--
-- ctx->key_dma = dma_map_single(jrdev, ctx->key, ctx->enckeylen,
-- DMA_TO_DEVICE);
-- if (dma_mapping_error(jrdev, ctx->key_dma)) {
-- dev_err(jrdev, "unable to map key i/o memory\n");
-- return -ENOMEM;
-- }
--
-- ret = rfc4543_set_sh_desc(aead);
-- if (ret) {
-- dma_unmap_single(jrdev, ctx->key_dma, ctx->enckeylen,
-- DMA_TO_DEVICE);
-- }
--
-- return ret;
-+ ctx->cdata.keylen = keylen - 4;
-+ dma_sync_single_for_device(jrdev, ctx->key_dma, ctx->cdata.keylen,
-+ ctx->dir);
-+ return rfc4543_set_sh_desc(aead);
- }
-
- static int ablkcipher_setkey(struct crypto_ablkcipher *ablkcipher,
- const u8 *key, unsigned int keylen)
- {
- struct caam_ctx *ctx = crypto_ablkcipher_ctx(ablkcipher);
-- struct ablkcipher_tfm *crt = &ablkcipher->base.crt_ablkcipher;
- struct crypto_tfm *tfm = crypto_ablkcipher_tfm(ablkcipher);
- const char *alg_name = crypto_tfm_alg_name(tfm);
- struct device *jrdev = ctx->jrdev;
-- int ret = 0;
-- u32 *key_jump_cmd;
-+ unsigned int ivsize = crypto_ablkcipher_ivsize(ablkcipher);
- u32 *desc;
-- u8 *nonce;
-- u32 geniv;
- u32 ctx1_iv_off = 0;
-- const bool ctr_mode = ((ctx->class1_alg_type & OP_ALG_AAI_MASK) ==
-+ const bool ctr_mode = ((ctx->cdata.algtype & OP_ALG_AAI_MASK) ==
- OP_ALG_AAI_CTR_MOD128);
- const bool is_rfc3686 = (ctr_mode &&
- (strstr(alg_name, "rfc3686") != NULL));
-
-+ memcpy(ctx->key, key, keylen);
- #ifdef DEBUG
- print_hex_dump(KERN_ERR, "key in @"__stringify(__LINE__)": ",
- DUMP_PREFIX_ADDRESS, 16, 4, key, keylen, 1);
-@@ -1544,215 +684,33 @@ static int ablkcipher_setkey(struct cryp
- keylen -= CTR_RFC3686_NONCE_SIZE;
- }
-
-- memcpy(ctx->key, key, keylen);
-- ctx->key_dma = dma_map_single(jrdev, ctx->key, keylen,
-- DMA_TO_DEVICE);
-- if (dma_mapping_error(jrdev, ctx->key_dma)) {
-- dev_err(jrdev, "unable to map key i/o memory\n");
-- return -ENOMEM;
-- }
-- ctx->enckeylen = keylen;
-+ dma_sync_single_for_device(jrdev, ctx->key_dma, keylen, DMA_TO_DEVICE);
-+ ctx->cdata.keylen = keylen;
-+ ctx->cdata.key_virt = ctx->key;
-+ ctx->cdata.key_inline = true;
-
- /* ablkcipher_encrypt shared descriptor */
- desc = ctx->sh_desc_enc;
-- init_sh_desc(desc, HDR_SHARE_SERIAL | HDR_SAVECTX);
-- /* Skip if already shared */
-- key_jump_cmd = append_jump(desc, JUMP_JSL | JUMP_TEST_ALL |
-- JUMP_COND_SHRD);
--
-- /* Load class1 key only */
-- append_key_as_imm(desc, (void *)ctx->key, ctx->enckeylen,
-- ctx->enckeylen, CLASS_1 |
-- KEY_DEST_CLASS_REG);
--
-- /* Load nonce into CONTEXT1 reg */
-- if (is_rfc3686) {
-- nonce = (u8 *)key + keylen;
-- append_load_as_imm(desc, nonce, CTR_RFC3686_NONCE_SIZE,
-- LDST_CLASS_IND_CCB |
-- LDST_SRCDST_BYTE_OUTFIFO | LDST_IMM);
-- append_move(desc, MOVE_WAITCOMP |
-- MOVE_SRC_OUTFIFO |
-- MOVE_DEST_CLASS1CTX |
-- (16 << MOVE_OFFSET_SHIFT) |
-- (CTR_RFC3686_NONCE_SIZE << MOVE_LEN_SHIFT));
-- }
--
-- set_jump_tgt_here(desc, key_jump_cmd);
--
-- /* Load iv */
-- append_seq_load(desc, crt->ivsize, LDST_SRCDST_BYTE_CONTEXT |
-- LDST_CLASS_1_CCB | (ctx1_iv_off << LDST_OFFSET_SHIFT));
--
-- /* Load counter into CONTEXT1 reg */
-- if (is_rfc3686)
-- append_load_imm_be32(desc, 1, LDST_IMM | LDST_CLASS_1_CCB |
-- LDST_SRCDST_BYTE_CONTEXT |
-- ((ctx1_iv_off + CTR_RFC3686_IV_SIZE) <<
-- LDST_OFFSET_SHIFT));
--
-- /* Load operation */
-- append_operation(desc, ctx->class1_alg_type |
-- OP_ALG_AS_INITFINAL | OP_ALG_ENCRYPT);
--
-- /* Perform operation */
-- ablkcipher_append_src_dst(desc);
-+ cnstr_shdsc_ablkcipher_encap(desc, &ctx->cdata, ivsize, is_rfc3686,
-+ ctx1_iv_off);
-+ dma_sync_single_for_device(jrdev, ctx->sh_desc_enc_dma,
-+ desc_bytes(desc), ctx->dir);
-
-- ctx->sh_desc_enc_dma = dma_map_single(jrdev, desc,
-- desc_bytes(desc),
-- DMA_TO_DEVICE);
-- if (dma_mapping_error(jrdev, ctx->sh_desc_enc_dma)) {
-- dev_err(jrdev, "unable to map shared descriptor\n");
-- return -ENOMEM;
-- }
--#ifdef DEBUG
-- print_hex_dump(KERN_ERR,
-- "ablkcipher enc shdesc@"__stringify(__LINE__)": ",
-- DUMP_PREFIX_ADDRESS, 16, 4, desc,
-- desc_bytes(desc), 1);
--#endif
- /* ablkcipher_decrypt shared descriptor */
- desc = ctx->sh_desc_dec;
-+ cnstr_shdsc_ablkcipher_decap(desc, &ctx->cdata, ivsize, is_rfc3686,
-+ ctx1_iv_off);
-+ dma_sync_single_for_device(jrdev, ctx->sh_desc_dec_dma,
-+ desc_bytes(desc), ctx->dir);
-
-- init_sh_desc(desc, HDR_SHARE_SERIAL | HDR_SAVECTX);
-- /* Skip if already shared */
-- key_jump_cmd = append_jump(desc, JUMP_JSL | JUMP_TEST_ALL |
-- JUMP_COND_SHRD);
--
-- /* Load class1 key only */
-- append_key_as_imm(desc, (void *)ctx->key, ctx->enckeylen,
-- ctx->enckeylen, CLASS_1 |
-- KEY_DEST_CLASS_REG);
--
-- /* Load nonce into CONTEXT1 reg */
-- if (is_rfc3686) {
-- nonce = (u8 *)key + keylen;
-- append_load_as_imm(desc, nonce, CTR_RFC3686_NONCE_SIZE,
-- LDST_CLASS_IND_CCB |
-- LDST_SRCDST_BYTE_OUTFIFO | LDST_IMM);
-- append_move(desc, MOVE_WAITCOMP |
-- MOVE_SRC_OUTFIFO |
-- MOVE_DEST_CLASS1CTX |
-- (16 << MOVE_OFFSET_SHIFT) |
-- (CTR_RFC3686_NONCE_SIZE << MOVE_LEN_SHIFT));
-- }
--
-- set_jump_tgt_here(desc, key_jump_cmd);
--
-- /* load IV */
-- append_seq_load(desc, crt->ivsize, LDST_SRCDST_BYTE_CONTEXT |
-- LDST_CLASS_1_CCB | (ctx1_iv_off << LDST_OFFSET_SHIFT));
--
-- /* Load counter into CONTEXT1 reg */
-- if (is_rfc3686)
-- append_load_imm_be32(desc, 1, LDST_IMM | LDST_CLASS_1_CCB |
-- LDST_SRCDST_BYTE_CONTEXT |
-- ((ctx1_iv_off + CTR_RFC3686_IV_SIZE) <<
-- LDST_OFFSET_SHIFT));
--
-- /* Choose operation */
-- if (ctr_mode)
-- append_operation(desc, ctx->class1_alg_type |
-- OP_ALG_AS_INITFINAL | OP_ALG_DECRYPT);
-- else
-- append_dec_op1(desc, ctx->class1_alg_type);
--
-- /* Perform operation */
-- ablkcipher_append_src_dst(desc);
--
-- ctx->sh_desc_dec_dma = dma_map_single(jrdev, desc,
-- desc_bytes(desc),
-- DMA_TO_DEVICE);
-- if (dma_mapping_error(jrdev, ctx->sh_desc_dec_dma)) {
-- dev_err(jrdev, "unable to map shared descriptor\n");
-- return -ENOMEM;
-- }
--
--#ifdef DEBUG
-- print_hex_dump(KERN_ERR,
-- "ablkcipher dec shdesc@"__stringify(__LINE__)": ",
-- DUMP_PREFIX_ADDRESS, 16, 4, desc,
-- desc_bytes(desc), 1);
--#endif
- /* ablkcipher_givencrypt shared descriptor */
- desc = ctx->sh_desc_givenc;
-+ cnstr_shdsc_ablkcipher_givencap(desc, &ctx->cdata, ivsize, is_rfc3686,
-+ ctx1_iv_off);
-+ dma_sync_single_for_device(jrdev, ctx->sh_desc_givenc_dma,
-+ desc_bytes(desc), ctx->dir);
-
-- init_sh_desc(desc, HDR_SHARE_SERIAL | HDR_SAVECTX);
-- /* Skip if already shared */
-- key_jump_cmd = append_jump(desc, JUMP_JSL | JUMP_TEST_ALL |
-- JUMP_COND_SHRD);
--
-- /* Load class1 key only */
-- append_key_as_imm(desc, (void *)ctx->key, ctx->enckeylen,
-- ctx->enckeylen, CLASS_1 |
-- KEY_DEST_CLASS_REG);
--
-- /* Load Nonce into CONTEXT1 reg */
-- if (is_rfc3686) {
-- nonce = (u8 *)key + keylen;
-- append_load_as_imm(desc, nonce, CTR_RFC3686_NONCE_SIZE,
-- LDST_CLASS_IND_CCB |
-- LDST_SRCDST_BYTE_OUTFIFO | LDST_IMM);
-- append_move(desc, MOVE_WAITCOMP |
-- MOVE_SRC_OUTFIFO |
-- MOVE_DEST_CLASS1CTX |
-- (16 << MOVE_OFFSET_SHIFT) |
-- (CTR_RFC3686_NONCE_SIZE << MOVE_LEN_SHIFT));
-- }
-- set_jump_tgt_here(desc, key_jump_cmd);
--
-- /* Generate IV */
-- geniv = NFIFOENTRY_STYPE_PAD | NFIFOENTRY_DEST_DECO |
-- NFIFOENTRY_DTYPE_MSG | NFIFOENTRY_LC1 |
-- NFIFOENTRY_PTYPE_RND | (crt->ivsize << NFIFOENTRY_DLEN_SHIFT);
-- append_load_imm_u32(desc, geniv, LDST_CLASS_IND_CCB |
-- LDST_SRCDST_WORD_INFO_FIFO | LDST_IMM);
-- append_cmd(desc, CMD_LOAD | DISABLE_AUTO_INFO_FIFO);
-- append_move(desc, MOVE_WAITCOMP |
-- MOVE_SRC_INFIFO |
-- MOVE_DEST_CLASS1CTX |
-- (crt->ivsize << MOVE_LEN_SHIFT) |
-- (ctx1_iv_off << MOVE_OFFSET_SHIFT));
-- append_cmd(desc, CMD_LOAD | ENABLE_AUTO_INFO_FIFO);
--
-- /* Copy generated IV to memory */
-- append_seq_store(desc, crt->ivsize,
-- LDST_SRCDST_BYTE_CONTEXT | LDST_CLASS_1_CCB |
-- (ctx1_iv_off << LDST_OFFSET_SHIFT));
--
-- /* Load Counter into CONTEXT1 reg */
-- if (is_rfc3686)
-- append_load_imm_be32(desc, 1, LDST_IMM | LDST_CLASS_1_CCB |
-- LDST_SRCDST_BYTE_CONTEXT |
-- ((ctx1_iv_off + CTR_RFC3686_IV_SIZE) <<
-- LDST_OFFSET_SHIFT));
--
-- if (ctx1_iv_off)
-- append_jump(desc, JUMP_JSL | JUMP_TEST_ALL | JUMP_COND_NCP |
-- (1 << JUMP_OFFSET_SHIFT));
--
-- /* Load operation */
-- append_operation(desc, ctx->class1_alg_type |
-- OP_ALG_AS_INITFINAL | OP_ALG_ENCRYPT);
--
-- /* Perform operation */
-- ablkcipher_append_src_dst(desc);
--
-- ctx->sh_desc_givenc_dma = dma_map_single(jrdev, desc,
-- desc_bytes(desc),
-- DMA_TO_DEVICE);
-- if (dma_mapping_error(jrdev, ctx->sh_desc_givenc_dma)) {
-- dev_err(jrdev, "unable to map shared descriptor\n");
-- return -ENOMEM;
-- }
--#ifdef DEBUG
-- print_hex_dump(KERN_ERR,
-- "ablkcipher givenc shdesc@" __stringify(__LINE__) ": ",
-- DUMP_PREFIX_ADDRESS, 16, 4, desc,
-- desc_bytes(desc), 1);
--#endif
--
-- return ret;
-+ return 0;
- }
-
- static int xts_ablkcipher_setkey(struct crypto_ablkcipher *ablkcipher,
-@@ -1760,8 +718,7 @@ static int xts_ablkcipher_setkey(struct
- {
- struct caam_ctx *ctx = crypto_ablkcipher_ctx(ablkcipher);
- struct device *jrdev = ctx->jrdev;
-- u32 *key_jump_cmd, *desc;
-- __be64 sector_size = cpu_to_be64(512);
-+ u32 *desc;
-
- if (keylen != 2 * AES_MIN_KEY_SIZE && keylen != 2 * AES_MAX_KEY_SIZE) {
- crypto_ablkcipher_set_flags(ablkcipher,
-@@ -1771,126 +728,38 @@ static int xts_ablkcipher_setkey(struct
- }
-
- memcpy(ctx->key, key, keylen);
-- ctx->key_dma = dma_map_single(jrdev, ctx->key, keylen, DMA_TO_DEVICE);
-- if (dma_mapping_error(jrdev, ctx->key_dma)) {
-- dev_err(jrdev, "unable to map key i/o memory\n");
-- return -ENOMEM;
-- }
-- ctx->enckeylen = keylen;
-+ dma_sync_single_for_device(jrdev, ctx->key_dma, keylen, DMA_TO_DEVICE);
-+ ctx->cdata.keylen = keylen;
-+ ctx->cdata.key_virt = ctx->key;
-+ ctx->cdata.key_inline = true;
-
- /* xts_ablkcipher_encrypt shared descriptor */
- desc = ctx->sh_desc_enc;
-- init_sh_desc(desc, HDR_SHARE_SERIAL | HDR_SAVECTX);
-- /* Skip if already shared */
-- key_jump_cmd = append_jump(desc, JUMP_JSL | JUMP_TEST_ALL |
-- JUMP_COND_SHRD);
--
-- /* Load class1 keys only */
-- append_key_as_imm(desc, (void *)ctx->key, ctx->enckeylen,
-- ctx->enckeylen, CLASS_1 | KEY_DEST_CLASS_REG);
--
-- /* Load sector size with index 40 bytes (0x28) */
-- append_cmd(desc, CMD_LOAD | IMMEDIATE | LDST_SRCDST_BYTE_CONTEXT |
-- LDST_CLASS_1_CCB | (0x28 << LDST_OFFSET_SHIFT) | 8);
-- append_data(desc, (void *)&sector_size, 8);
--
-- set_jump_tgt_here(desc, key_jump_cmd);
--
-- /*
-- * create sequence for loading the sector index
-- * Upper 8B of IV - will be used as sector index
-- * Lower 8B of IV - will be discarded
-- */
-- append_cmd(desc, CMD_SEQ_LOAD | LDST_SRCDST_BYTE_CONTEXT |
-- LDST_CLASS_1_CCB | (0x20 << LDST_OFFSET_SHIFT) | 8);
-- append_seq_fifo_load(desc, 8, FIFOLD_CLASS_SKIP);
--
-- /* Load operation */
-- append_operation(desc, ctx->class1_alg_type | OP_ALG_AS_INITFINAL |
-- OP_ALG_ENCRYPT);
--
-- /* Perform operation */
-- ablkcipher_append_src_dst(desc);
--
-- ctx->sh_desc_enc_dma = dma_map_single(jrdev, desc, desc_bytes(desc),
-- DMA_TO_DEVICE);
-- if (dma_mapping_error(jrdev, ctx->sh_desc_enc_dma)) {
-- dev_err(jrdev, "unable to map shared descriptor\n");
-- return -ENOMEM;
-- }
--#ifdef DEBUG
-- print_hex_dump(KERN_ERR,
-- "xts ablkcipher enc shdesc@" __stringify(__LINE__) ": ",
-- DUMP_PREFIX_ADDRESS, 16, 4, desc, desc_bytes(desc), 1);
--#endif
-+ cnstr_shdsc_xts_ablkcipher_encap(desc, &ctx->cdata);
-+ dma_sync_single_for_device(jrdev, ctx->sh_desc_enc_dma,
-+ desc_bytes(desc), ctx->dir);
-
- /* xts_ablkcipher_decrypt shared descriptor */
- desc = ctx->sh_desc_dec;
--
-- init_sh_desc(desc, HDR_SHARE_SERIAL | HDR_SAVECTX);
-- /* Skip if already shared */
-- key_jump_cmd = append_jump(desc, JUMP_JSL | JUMP_TEST_ALL |
-- JUMP_COND_SHRD);
--
-- /* Load class1 key only */
-- append_key_as_imm(desc, (void *)ctx->key, ctx->enckeylen,
-- ctx->enckeylen, CLASS_1 | KEY_DEST_CLASS_REG);
--
-- /* Load sector size with index 40 bytes (0x28) */
-- append_cmd(desc, CMD_LOAD | IMMEDIATE | LDST_SRCDST_BYTE_CONTEXT |
-- LDST_CLASS_1_CCB | (0x28 << LDST_OFFSET_SHIFT) | 8);
-- append_data(desc, (void *)&sector_size, 8);
--
-- set_jump_tgt_here(desc, key_jump_cmd);
--
-- /*
-- * create sequence for loading the sector index
-- * Upper 8B of IV - will be used as sector index
-- * Lower 8B of IV - will be discarded
-- */
-- append_cmd(desc, CMD_SEQ_LOAD | LDST_SRCDST_BYTE_CONTEXT |
-- LDST_CLASS_1_CCB | (0x20 << LDST_OFFSET_SHIFT) | 8);
-- append_seq_fifo_load(desc, 8, FIFOLD_CLASS_SKIP);
--
-- /* Load operation */
-- append_dec_op1(desc, ctx->class1_alg_type);
--
-- /* Perform operation */
-- ablkcipher_append_src_dst(desc);
--
-- ctx->sh_desc_dec_dma = dma_map_single(jrdev, desc, desc_bytes(desc),
-- DMA_TO_DEVICE);
-- if (dma_mapping_error(jrdev, ctx->sh_desc_dec_dma)) {
-- dma_unmap_single(jrdev, ctx->sh_desc_enc_dma,
-- desc_bytes(ctx->sh_desc_enc), DMA_TO_DEVICE);
-- dev_err(jrdev, "unable to map shared descriptor\n");
-- return -ENOMEM;
-- }
--#ifdef DEBUG
-- print_hex_dump(KERN_ERR,
-- "xts ablkcipher dec shdesc@" __stringify(__LINE__) ": ",
-- DUMP_PREFIX_ADDRESS, 16, 4, desc, desc_bytes(desc), 1);
--#endif
-+ cnstr_shdsc_xts_ablkcipher_decap(desc, &ctx->cdata);
-+ dma_sync_single_for_device(jrdev, ctx->sh_desc_dec_dma,
-+ desc_bytes(desc), ctx->dir);
-
- return 0;
- }
-
- /*
- * aead_edesc - s/w-extended aead descriptor
-- * @assoc_nents: number of segments in associated data (SPI+Seq) scatterlist
-- * @src_nents: number of segments in input scatterlist
-- * @dst_nents: number of segments in output scatterlist
-- * @iv_dma: dma address of iv for checking continuity and link table
-- * @desc: h/w descriptor (variable length; must not exceed MAX_CAAM_DESCSIZE)
-+ * @src_nents: number of segments in input s/w scatterlist
-+ * @dst_nents: number of segments in output s/w scatterlist
- * @sec4_sg_bytes: length of dma mapped sec4_sg space
- * @sec4_sg_dma: bus physical mapped address of h/w link table
-+ * @sec4_sg: pointer to h/w link table
- * @hw_desc: the h/w job descriptor followed by any referenced link tables
- */
- struct aead_edesc {
-- int assoc_nents;
- int src_nents;
- int dst_nents;
-- dma_addr_t iv_dma;
- int sec4_sg_bytes;
- dma_addr_t sec4_sg_dma;
- struct sec4_sg_entry *sec4_sg;
-@@ -1899,12 +768,12 @@ struct aead_edesc {
-
- /*
- * ablkcipher_edesc - s/w-extended ablkcipher descriptor
-- * @src_nents: number of segments in input scatterlist
-- * @dst_nents: number of segments in output scatterlist
-+ * @src_nents: number of segments in input s/w scatterlist
-+ * @dst_nents: number of segments in output s/w scatterlist
- * @iv_dma: dma address of iv for checking continuity and link table
-- * @desc: h/w descriptor (variable length; must not exceed MAX_CAAM_DESCSIZE)
- * @sec4_sg_bytes: length of dma mapped sec4_sg space
- * @sec4_sg_dma: bus physical mapped address of h/w link table
-+ * @sec4_sg: pointer to h/w link table
- * @hw_desc: the h/w job descriptor followed by any referenced link tables
- */
- struct ablkcipher_edesc {
-@@ -1924,10 +793,11 @@ static void caam_unmap(struct device *de
- int sec4_sg_bytes)
- {
- if (dst != src) {
-- dma_unmap_sg(dev, src, src_nents ? : 1, DMA_TO_DEVICE);
-- dma_unmap_sg(dev, dst, dst_nents ? : 1, DMA_FROM_DEVICE);
-+ if (src_nents)
-+ dma_unmap_sg(dev, src, src_nents, DMA_TO_DEVICE);
-+ dma_unmap_sg(dev, dst, dst_nents, DMA_FROM_DEVICE);
- } else {
-- dma_unmap_sg(dev, src, src_nents ? : 1, DMA_BIDIRECTIONAL);
-+ dma_unmap_sg(dev, src, src_nents, DMA_BIDIRECTIONAL);
- }
-
- if (iv_dma)
-@@ -2021,8 +891,7 @@ static void ablkcipher_encrypt_done(stru
- dev_err(jrdev, "%s %d: err 0x%x\n", __func__, __LINE__, err);
- #endif
-
-- edesc = (struct ablkcipher_edesc *)((char *)desc -
-- offsetof(struct ablkcipher_edesc, hw_desc));
-+ edesc = container_of(desc, struct ablkcipher_edesc, hw_desc[0]);
-
- if (err)
- caam_jr_strstatus(jrdev, err);
-@@ -2031,10 +900,10 @@ static void ablkcipher_encrypt_done(stru
- print_hex_dump(KERN_ERR, "dstiv @"__stringify(__LINE__)": ",
- DUMP_PREFIX_ADDRESS, 16, 4, req->info,
- edesc->src_nents > 1 ? 100 : ivsize, 1);
-- dbg_dump_sg(KERN_ERR, "dst @"__stringify(__LINE__)": ",
-- DUMP_PREFIX_ADDRESS, 16, 4, req->dst,
-- edesc->dst_nents > 1 ? 100 : req->nbytes, 1, true);
- #endif
-+ caam_dump_sg(KERN_ERR, "dst @" __stringify(__LINE__)": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, req->dst,
-+ edesc->dst_nents > 1 ? 100 : req->nbytes, 1);
-
- ablkcipher_unmap(jrdev, edesc, req);
-
-@@ -2062,8 +931,7 @@ static void ablkcipher_decrypt_done(stru
- dev_err(jrdev, "%s %d: err 0x%x\n", __func__, __LINE__, err);
- #endif
-
-- edesc = (struct ablkcipher_edesc *)((char *)desc -
-- offsetof(struct ablkcipher_edesc, hw_desc));
-+ edesc = container_of(desc, struct ablkcipher_edesc, hw_desc[0]);
- if (err)
- caam_jr_strstatus(jrdev, err);
-
-@@ -2071,10 +939,10 @@ static void ablkcipher_decrypt_done(stru
- print_hex_dump(KERN_ERR, "dstiv @"__stringify(__LINE__)": ",
- DUMP_PREFIX_ADDRESS, 16, 4, req->info,
- ivsize, 1);
-- dbg_dump_sg(KERN_ERR, "dst @"__stringify(__LINE__)": ",
-- DUMP_PREFIX_ADDRESS, 16, 4, req->dst,
-- edesc->dst_nents > 1 ? 100 : req->nbytes, 1, true);
- #endif
-+ caam_dump_sg(KERN_ERR, "dst @" __stringify(__LINE__)": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, req->dst,
-+ edesc->dst_nents > 1 ? 100 : req->nbytes, 1);
-
- ablkcipher_unmap(jrdev, edesc, req);
-
-@@ -2114,7 +982,7 @@ static void init_aead_job(struct aead_re
- init_job_desc_shared(desc, ptr, len, HDR_SHARE_DEFER | HDR_REVERSE);
-
- if (all_contig) {
-- src_dma = sg_dma_address(req->src);
-+ src_dma = edesc->src_nents ? sg_dma_address(req->src) : 0;
- in_options = 0;
- } else {
- src_dma = edesc->sec4_sg_dma;
-@@ -2129,7 +997,7 @@ static void init_aead_job(struct aead_re
- out_options = in_options;
-
- if (unlikely(req->src != req->dst)) {
-- if (!edesc->dst_nents) {
-+ if (edesc->dst_nents == 1) {
- dst_dma = sg_dma_address(req->dst);
- } else {
- dst_dma = edesc->sec4_sg_dma +
-@@ -2147,9 +1015,6 @@ static void init_aead_job(struct aead_re
- append_seq_out_ptr(desc, dst_dma,
- req->assoclen + req->cryptlen - authsize,
- out_options);
--
-- /* REG3 = assoclen */
-- append_math_add_imm_u32(desc, REG3, ZERO, IMM, req->assoclen);
- }
-
- static void init_gcm_job(struct aead_request *req,
-@@ -2164,6 +1029,7 @@ static void init_gcm_job(struct aead_req
- unsigned int last;
-
- init_aead_job(req, edesc, all_contig, encrypt);
-+ append_math_add_imm_u32(desc, REG3, ZERO, IMM, req->assoclen);
-
- /* BUG This should not be specific to generic GCM. */
- last = 0;
-@@ -2175,7 +1041,7 @@ static void init_gcm_job(struct aead_req
- FIFOLD_TYPE_IV | FIFOLD_TYPE_FLUSH1 | 12 | last);
- /* Append Salt */
- if (!generic_gcm)
-- append_data(desc, ctx->key + ctx->enckeylen, 4);
-+ append_data(desc, ctx->key + ctx->cdata.keylen, 4);
- /* Append IV */
- append_data(desc, req->iv, ivsize);
- /* End of blank commands */
-@@ -2190,7 +1056,8 @@ static void init_authenc_job(struct aead
- struct caam_aead_alg, aead);
- unsigned int ivsize = crypto_aead_ivsize(aead);
- struct caam_ctx *ctx = crypto_aead_ctx(aead);
-- const bool ctr_mode = ((ctx->class1_alg_type & OP_ALG_AAI_MASK) ==
-+ struct caam_drv_private *ctrlpriv = dev_get_drvdata(ctx->jrdev->parent);
-+ const bool ctr_mode = ((ctx->cdata.algtype & OP_ALG_AAI_MASK) ==
- OP_ALG_AAI_CTR_MOD128);
- const bool is_rfc3686 = alg->caam.rfc3686;
- u32 *desc = edesc->hw_desc;
-@@ -2213,6 +1080,15 @@ static void init_authenc_job(struct aead
-
- init_aead_job(req, edesc, all_contig, encrypt);
-
-+ /*
-+ * {REG3, DPOVRD} = assoclen, depending on whether MATH command supports
-+ * having DPOVRD as destination.
-+ */
-+ if (ctrlpriv->era < 3)
-+ append_math_add_imm_u32(desc, REG3, ZERO, IMM, req->assoclen);
-+ else
-+ append_math_add_imm_u32(desc, DPOVRD, ZERO, IMM, req->assoclen);
-+
- if (ivsize && ((is_rfc3686 && encrypt) || !alg->caam.geniv))
- append_load_as_imm(desc, req->iv, ivsize,
- LDST_CLASS_1_CCB |
-@@ -2236,16 +1112,15 @@ static void init_ablkcipher_job(u32 *sh_
- int len, sec4_sg_index = 0;
-
- #ifdef DEBUG
-- bool may_sleep = ((req->base.flags & (CRYPTO_TFM_REQ_MAY_BACKLOG |
-- CRYPTO_TFM_REQ_MAY_SLEEP)) != 0);
- print_hex_dump(KERN_ERR, "presciv@"__stringify(__LINE__)": ",
- DUMP_PREFIX_ADDRESS, 16, 4, req->info,
- ivsize, 1);
-- printk(KERN_ERR "asked=%d, nbytes%d\n", (int)edesc->src_nents ? 100 : req->nbytes, req->nbytes);
-- dbg_dump_sg(KERN_ERR, "src @"__stringify(__LINE__)": ",
-- DUMP_PREFIX_ADDRESS, 16, 4, req->src,
-- edesc->src_nents ? 100 : req->nbytes, 1, may_sleep);
-+ pr_err("asked=%d, nbytes%d\n",
-+ (int)edesc->src_nents > 1 ? 100 : req->nbytes, req->nbytes);
- #endif
-+ caam_dump_sg(KERN_ERR, "src @" __stringify(__LINE__)": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, req->src,
-+ edesc->src_nents > 1 ? 100 : req->nbytes, 1);
-
- len = desc_len(sh_desc);
- init_job_desc_shared(desc, ptr, len, HDR_SHARE_DEFER | HDR_REVERSE);
-@@ -2261,7 +1136,7 @@ static void init_ablkcipher_job(u32 *sh_
- append_seq_in_ptr(desc, src_dma, req->nbytes + ivsize, in_options);
-
- if (likely(req->src == req->dst)) {
-- if (!edesc->src_nents && iv_contig) {
-+ if (edesc->src_nents == 1 && iv_contig) {
- dst_dma = sg_dma_address(req->src);
- } else {
- dst_dma = edesc->sec4_sg_dma +
-@@ -2269,7 +1144,7 @@ static void init_ablkcipher_job(u32 *sh_
- out_options = LDST_SGF;
- }
- } else {
-- if (!edesc->dst_nents) {
-+ if (edesc->dst_nents == 1) {
- dst_dma = sg_dma_address(req->dst);
- } else {
- dst_dma = edesc->sec4_sg_dma +
-@@ -2296,20 +1171,18 @@ static void init_ablkcipher_giv_job(u32
- int len, sec4_sg_index = 0;
-
- #ifdef DEBUG
-- bool may_sleep = ((req->base.flags & (CRYPTO_TFM_REQ_MAY_BACKLOG |
-- CRYPTO_TFM_REQ_MAY_SLEEP)) != 0);
- print_hex_dump(KERN_ERR, "presciv@" __stringify(__LINE__) ": ",
- DUMP_PREFIX_ADDRESS, 16, 4, req->info,
- ivsize, 1);
-- dbg_dump_sg(KERN_ERR, "src @" __stringify(__LINE__) ": ",
-- DUMP_PREFIX_ADDRESS, 16, 4, req->src,
-- edesc->src_nents ? 100 : req->nbytes, 1, may_sleep);
- #endif
-+ caam_dump_sg(KERN_ERR, "src @" __stringify(__LINE__) ": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, req->src,
-+ edesc->src_nents > 1 ? 100 : req->nbytes, 1);
-
- len = desc_len(sh_desc);
- init_job_desc_shared(desc, ptr, len, HDR_SHARE_DEFER | HDR_REVERSE);
-
-- if (!edesc->src_nents) {
-+ if (edesc->src_nents == 1) {
- src_dma = sg_dma_address(req->src);
- in_options = 0;
- } else {
-@@ -2340,87 +1213,100 @@ static struct aead_edesc *aead_edesc_all
- struct crypto_aead *aead = crypto_aead_reqtfm(req);
- struct caam_ctx *ctx = crypto_aead_ctx(aead);
- struct device *jrdev = ctx->jrdev;
-- gfp_t flags = (req->base.flags & (CRYPTO_TFM_REQ_MAY_BACKLOG |
-- CRYPTO_TFM_REQ_MAY_SLEEP)) ? GFP_KERNEL : GFP_ATOMIC;
-- int src_nents, dst_nents = 0;
-+ gfp_t flags = (req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP) ?
-+ GFP_KERNEL : GFP_ATOMIC;
-+ int src_nents, mapped_src_nents, dst_nents = 0, mapped_dst_nents = 0;
- struct aead_edesc *edesc;
-- int sgc;
-- bool all_contig = true;
-- int sec4_sg_index, sec4_sg_len = 0, sec4_sg_bytes;
-+ int sec4_sg_index, sec4_sg_len, sec4_sg_bytes;
- unsigned int authsize = ctx->authsize;
-
- if (unlikely(req->dst != req->src)) {
-- src_nents = sg_count(req->src, req->assoclen + req->cryptlen);
-- dst_nents = sg_count(req->dst,
-- req->assoclen + req->cryptlen +
-- (encrypt ? authsize : (-authsize)));
-- } else {
-- src_nents = sg_count(req->src,
-- req->assoclen + req->cryptlen +
-- (encrypt ? authsize : 0));
-- }
--
-- /* Check if data are contiguous. */
-- all_contig = !src_nents;
-- if (!all_contig) {
-- src_nents = src_nents ? : 1;
-- sec4_sg_len = src_nents;
-- }
--
-- sec4_sg_len += dst_nents;
--
-- sec4_sg_bytes = sec4_sg_len * sizeof(struct sec4_sg_entry);
-+ src_nents = sg_nents_for_len(req->src, req->assoclen +
-+ req->cryptlen);
-+ if (unlikely(src_nents < 0)) {
-+ dev_err(jrdev, "Insufficient bytes (%d) in src S/G\n",
-+ req->assoclen + req->cryptlen);
-+ return ERR_PTR(src_nents);
-+ }
-
-- /* allocate space for base edesc and hw desc commands, link tables */
-- edesc = kzalloc(sizeof(*edesc) + desc_bytes + sec4_sg_bytes,
-- GFP_DMA | flags);
-- if (!edesc) {
-- dev_err(jrdev, "could not allocate extended descriptor\n");
-- return ERR_PTR(-ENOMEM);
-+ dst_nents = sg_nents_for_len(req->dst, req->assoclen +
-+ req->cryptlen +
-+ (encrypt ? authsize :
-+ (-authsize)));
-+ if (unlikely(dst_nents < 0)) {
-+ dev_err(jrdev, "Insufficient bytes (%d) in dst S/G\n",
-+ req->assoclen + req->cryptlen +
-+ (encrypt ? authsize : (-authsize)));
-+ return ERR_PTR(dst_nents);
-+ }
-+ } else {
-+ src_nents = sg_nents_for_len(req->src, req->assoclen +
-+ req->cryptlen +
-+ (encrypt ? authsize : 0));
-+ if (unlikely(src_nents < 0)) {
-+ dev_err(jrdev, "Insufficient bytes (%d) in src S/G\n",
-+ req->assoclen + req->cryptlen +
-+ (encrypt ? authsize : 0));
-+ return ERR_PTR(src_nents);
-+ }
- }
-
- if (likely(req->src == req->dst)) {
-- sgc = dma_map_sg(jrdev, req->src, src_nents ? : 1,
-- DMA_BIDIRECTIONAL);
-- if (unlikely(!sgc)) {
-+ mapped_src_nents = dma_map_sg(jrdev, req->src, src_nents,
-+ DMA_BIDIRECTIONAL);
-+ if (unlikely(!mapped_src_nents)) {
- dev_err(jrdev, "unable to map source\n");
-- kfree(edesc);
- return ERR_PTR(-ENOMEM);
- }
- } else {
-- sgc = dma_map_sg(jrdev, req->src, src_nents ? : 1,
-- DMA_TO_DEVICE);
-- if (unlikely(!sgc)) {
-- dev_err(jrdev, "unable to map source\n");
-- kfree(edesc);
-- return ERR_PTR(-ENOMEM);
-+ /* Cover also the case of null (zero length) input data */
-+ if (src_nents) {
-+ mapped_src_nents = dma_map_sg(jrdev, req->src,
-+ src_nents, DMA_TO_DEVICE);
-+ if (unlikely(!mapped_src_nents)) {
-+ dev_err(jrdev, "unable to map source\n");
-+ return ERR_PTR(-ENOMEM);
-+ }
-+ } else {
-+ mapped_src_nents = 0;
- }
-
-- sgc = dma_map_sg(jrdev, req->dst, dst_nents ? : 1,
-- DMA_FROM_DEVICE);
-- if (unlikely(!sgc)) {
-+ mapped_dst_nents = dma_map_sg(jrdev, req->dst, dst_nents,
-+ DMA_FROM_DEVICE);
-+ if (unlikely(!mapped_dst_nents)) {
- dev_err(jrdev, "unable to map destination\n");
-- dma_unmap_sg(jrdev, req->src, src_nents ? : 1,
-- DMA_TO_DEVICE);
-- kfree(edesc);
-+ dma_unmap_sg(jrdev, req->src, src_nents, DMA_TO_DEVICE);
- return ERR_PTR(-ENOMEM);
- }
- }
-
-+ sec4_sg_len = mapped_src_nents > 1 ? mapped_src_nents : 0;
-+ sec4_sg_len += mapped_dst_nents > 1 ? mapped_dst_nents : 0;
-+ sec4_sg_bytes = sec4_sg_len * sizeof(struct sec4_sg_entry);
-+
-+ /* allocate space for base edesc and hw desc commands, link tables */
-+ edesc = kzalloc(sizeof(*edesc) + desc_bytes + sec4_sg_bytes,
-+ GFP_DMA | flags);
-+ if (!edesc) {
-+ caam_unmap(jrdev, req->src, req->dst, src_nents, dst_nents, 0,
-+ 0, 0, 0);
-+ return ERR_PTR(-ENOMEM);
-+ }
-+
- edesc->src_nents = src_nents;
- edesc->dst_nents = dst_nents;
- edesc->sec4_sg = (void *)edesc + sizeof(struct aead_edesc) +
- desc_bytes;
-- *all_contig_ptr = all_contig;
-+ *all_contig_ptr = !(mapped_src_nents > 1);
-
- sec4_sg_index = 0;
-- if (!all_contig) {
-- sg_to_sec4_sg_last(req->src, src_nents,
-- edesc->sec4_sg + sec4_sg_index, 0);
-- sec4_sg_index += src_nents;
-+ if (mapped_src_nents > 1) {
-+ sg_to_sec4_sg_last(req->src, mapped_src_nents,
-+ edesc->sec4_sg + sec4_sg_index, 0);
-+ sec4_sg_index += mapped_src_nents;
- }
-- if (dst_nents) {
-- sg_to_sec4_sg_last(req->dst, dst_nents,
-+ if (mapped_dst_nents > 1) {
-+ sg_to_sec4_sg_last(req->dst, mapped_dst_nents,
- edesc->sec4_sg + sec4_sg_index, 0);
- }
-
-@@ -2573,13 +1459,9 @@ static int aead_decrypt(struct aead_requ
- u32 *desc;
- int ret = 0;
-
--#ifdef DEBUG
-- bool may_sleep = ((req->base.flags & (CRYPTO_TFM_REQ_MAY_BACKLOG |
-- CRYPTO_TFM_REQ_MAY_SLEEP)) != 0);
-- dbg_dump_sg(KERN_ERR, "dec src@"__stringify(__LINE__)": ",
-- DUMP_PREFIX_ADDRESS, 16, 4, req->src,
-- req->assoclen + req->cryptlen, 1, may_sleep);
--#endif
-+ caam_dump_sg(KERN_ERR, "dec src@" __stringify(__LINE__)": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, req->src,
-+ req->assoclen + req->cryptlen, 1);
-
- /* allocate extended descriptor */
- edesc = aead_edesc_alloc(req, AUTHENC_DESC_JOB_IO_LEN,
-@@ -2619,51 +1501,80 @@ static struct ablkcipher_edesc *ablkciph
- struct device *jrdev = ctx->jrdev;
- gfp_t flags = (req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP) ?
- GFP_KERNEL : GFP_ATOMIC;
-- int src_nents, dst_nents = 0, sec4_sg_bytes;
-+ int src_nents, mapped_src_nents, dst_nents = 0, mapped_dst_nents = 0;
- struct ablkcipher_edesc *edesc;
- dma_addr_t iv_dma = 0;
-- bool iv_contig = false;
-- int sgc;
-+ bool in_contig;
- int ivsize = crypto_ablkcipher_ivsize(ablkcipher);
-- int sec4_sg_index;
-+ int dst_sg_idx, sec4_sg_ents, sec4_sg_bytes;
-
-- src_nents = sg_count(req->src, req->nbytes);
-+ src_nents = sg_nents_for_len(req->src, req->nbytes);
-+ if (unlikely(src_nents < 0)) {
-+ dev_err(jrdev, "Insufficient bytes (%d) in src S/G\n",
-+ req->nbytes);
-+ return ERR_PTR(src_nents);
-+ }
-
-- if (req->dst != req->src)
-- dst_nents = sg_count(req->dst, req->nbytes);
-+ if (req->dst != req->src) {
-+ dst_nents = sg_nents_for_len(req->dst, req->nbytes);
-+ if (unlikely(dst_nents < 0)) {
-+ dev_err(jrdev, "Insufficient bytes (%d) in dst S/G\n",
-+ req->nbytes);
-+ return ERR_PTR(dst_nents);
-+ }
-+ }
-
- if (likely(req->src == req->dst)) {
-- sgc = dma_map_sg(jrdev, req->src, src_nents ? : 1,
-- DMA_BIDIRECTIONAL);
-+ mapped_src_nents = dma_map_sg(jrdev, req->src, src_nents,
-+ DMA_BIDIRECTIONAL);
-+ if (unlikely(!mapped_src_nents)) {
-+ dev_err(jrdev, "unable to map source\n");
-+ return ERR_PTR(-ENOMEM);
-+ }
- } else {
-- sgc = dma_map_sg(jrdev, req->src, src_nents ? : 1,
-- DMA_TO_DEVICE);
-- sgc = dma_map_sg(jrdev, req->dst, dst_nents ? : 1,
-- DMA_FROM_DEVICE);
-+ mapped_src_nents = dma_map_sg(jrdev, req->src, src_nents,
-+ DMA_TO_DEVICE);
-+ if (unlikely(!mapped_src_nents)) {
-+ dev_err(jrdev, "unable to map source\n");
-+ return ERR_PTR(-ENOMEM);
-+ }
-+
-+ mapped_dst_nents = dma_map_sg(jrdev, req->dst, dst_nents,
-+ DMA_FROM_DEVICE);
-+ if (unlikely(!mapped_dst_nents)) {
-+ dev_err(jrdev, "unable to map destination\n");
-+ dma_unmap_sg(jrdev, req->src, src_nents, DMA_TO_DEVICE);
-+ return ERR_PTR(-ENOMEM);
-+ }
- }
-
- iv_dma = dma_map_single(jrdev, req->info, ivsize, DMA_TO_DEVICE);
- if (dma_mapping_error(jrdev, iv_dma)) {
- dev_err(jrdev, "unable to map IV\n");
-+ caam_unmap(jrdev, req->src, req->dst, src_nents, dst_nents, 0,
-+ 0, 0, 0);
- return ERR_PTR(-ENOMEM);
- }
-
-- /*
-- * Check if iv can be contiguous with source and destination.
-- * If so, include it. If not, create scatterlist.
-- */
-- if (!src_nents && iv_dma + ivsize == sg_dma_address(req->src))
-- iv_contig = true;
-- else
-- src_nents = src_nents ? : 1;
-- sec4_sg_bytes = ((iv_contig ? 0 : 1) + src_nents + dst_nents) *
-- sizeof(struct sec4_sg_entry);
-+ if (mapped_src_nents == 1 &&
-+ iv_dma + ivsize == sg_dma_address(req->src)) {
-+ in_contig = true;
-+ sec4_sg_ents = 0;
-+ } else {
-+ in_contig = false;
-+ sec4_sg_ents = 1 + mapped_src_nents;
-+ }
-+ dst_sg_idx = sec4_sg_ents;
-+ sec4_sg_ents += mapped_dst_nents > 1 ? mapped_dst_nents : 0;
-+ sec4_sg_bytes = sec4_sg_ents * sizeof(struct sec4_sg_entry);
-
- /* allocate space for base edesc and hw desc commands, link tables */
- edesc = kzalloc(sizeof(*edesc) + desc_bytes + sec4_sg_bytes,
- GFP_DMA | flags);
- if (!edesc) {
- dev_err(jrdev, "could not allocate extended descriptor\n");
-+ caam_unmap(jrdev, req->src, req->dst, src_nents, dst_nents,
-+ iv_dma, ivsize, 0, 0);
- return ERR_PTR(-ENOMEM);
- }
-
-@@ -2673,23 +1584,24 @@ static struct ablkcipher_edesc *ablkciph
- edesc->sec4_sg = (void *)edesc + sizeof(struct ablkcipher_edesc) +
- desc_bytes;
-
-- sec4_sg_index = 0;
-- if (!iv_contig) {
-+ if (!in_contig) {
- dma_to_sec4_sg_one(edesc->sec4_sg, iv_dma, ivsize, 0);
-- sg_to_sec4_sg_last(req->src, src_nents,
-+ sg_to_sec4_sg_last(req->src, mapped_src_nents,
- edesc->sec4_sg + 1, 0);
-- sec4_sg_index += 1 + src_nents;
- }
-
-- if (dst_nents) {
-- sg_to_sec4_sg_last(req->dst, dst_nents,
-- edesc->sec4_sg + sec4_sg_index, 0);
-+ if (mapped_dst_nents > 1) {
-+ sg_to_sec4_sg_last(req->dst, mapped_dst_nents,
-+ edesc->sec4_sg + dst_sg_idx, 0);
- }
-
- edesc->sec4_sg_dma = dma_map_single(jrdev, edesc->sec4_sg,
- sec4_sg_bytes, DMA_TO_DEVICE);
- if (dma_mapping_error(jrdev, edesc->sec4_sg_dma)) {
- dev_err(jrdev, "unable to map S/G table\n");
-+ caam_unmap(jrdev, req->src, req->dst, src_nents, dst_nents,
-+ iv_dma, ivsize, 0, 0);
-+ kfree(edesc);
- return ERR_PTR(-ENOMEM);
- }
-
-@@ -2701,7 +1613,7 @@ static struct ablkcipher_edesc *ablkciph
- sec4_sg_bytes, 1);
- #endif
-
-- *iv_contig_out = iv_contig;
-+ *iv_contig_out = in_contig;
- return edesc;
- }
-
-@@ -2792,30 +1704,54 @@ static struct ablkcipher_edesc *ablkciph
- struct crypto_ablkcipher *ablkcipher = crypto_ablkcipher_reqtfm(req);
- struct caam_ctx *ctx = crypto_ablkcipher_ctx(ablkcipher);
- struct device *jrdev = ctx->jrdev;
-- gfp_t flags = (req->base.flags & (CRYPTO_TFM_REQ_MAY_BACKLOG |
-- CRYPTO_TFM_REQ_MAY_SLEEP)) ?
-+ gfp_t flags = (req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP) ?
- GFP_KERNEL : GFP_ATOMIC;
-- int src_nents, dst_nents = 0, sec4_sg_bytes;
-+ int src_nents, mapped_src_nents, dst_nents, mapped_dst_nents;
- struct ablkcipher_edesc *edesc;
- dma_addr_t iv_dma = 0;
-- bool iv_contig = false;
-- int sgc;
-+ bool out_contig;
- int ivsize = crypto_ablkcipher_ivsize(ablkcipher);
-- int sec4_sg_index;
-+ int dst_sg_idx, sec4_sg_ents, sec4_sg_bytes;
-
-- src_nents = sg_count(req->src, req->nbytes);
--
-- if (unlikely(req->dst != req->src))
-- dst_nents = sg_count(req->dst, req->nbytes);
-+ src_nents = sg_nents_for_len(req->src, req->nbytes);
-+ if (unlikely(src_nents < 0)) {
-+ dev_err(jrdev, "Insufficient bytes (%d) in src S/G\n",
-+ req->nbytes);
-+ return ERR_PTR(src_nents);
-+ }
-
- if (likely(req->src == req->dst)) {
-- sgc = dma_map_sg(jrdev, req->src, src_nents ? : 1,
-- DMA_BIDIRECTIONAL);
-+ mapped_src_nents = dma_map_sg(jrdev, req->src, src_nents,
-+ DMA_BIDIRECTIONAL);
-+ if (unlikely(!mapped_src_nents)) {
-+ dev_err(jrdev, "unable to map source\n");
-+ return ERR_PTR(-ENOMEM);
-+ }
-+
-+ dst_nents = src_nents;
-+ mapped_dst_nents = src_nents;
- } else {
-- sgc = dma_map_sg(jrdev, req->src, src_nents ? : 1,
-- DMA_TO_DEVICE);
-- sgc = dma_map_sg(jrdev, req->dst, dst_nents ? : 1,
-- DMA_FROM_DEVICE);
-+ mapped_src_nents = dma_map_sg(jrdev, req->src, src_nents,
-+ DMA_TO_DEVICE);
-+ if (unlikely(!mapped_src_nents)) {
-+ dev_err(jrdev, "unable to map source\n");
-+ return ERR_PTR(-ENOMEM);
-+ }
-+
-+ dst_nents = sg_nents_for_len(req->dst, req->nbytes);
-+ if (unlikely(dst_nents < 0)) {
-+ dev_err(jrdev, "Insufficient bytes (%d) in dst S/G\n",
-+ req->nbytes);
-+ return ERR_PTR(dst_nents);
-+ }
-+
-+ mapped_dst_nents = dma_map_sg(jrdev, req->dst, dst_nents,
-+ DMA_FROM_DEVICE);
-+ if (unlikely(!mapped_dst_nents)) {
-+ dev_err(jrdev, "unable to map destination\n");
-+ dma_unmap_sg(jrdev, req->src, src_nents, DMA_TO_DEVICE);
-+ return ERR_PTR(-ENOMEM);
-+ }
- }
-
- /*
-@@ -2825,21 +1761,29 @@ static struct ablkcipher_edesc *ablkciph
- iv_dma = dma_map_single(jrdev, greq->giv, ivsize, DMA_TO_DEVICE);
- if (dma_mapping_error(jrdev, iv_dma)) {
- dev_err(jrdev, "unable to map IV\n");
-+ caam_unmap(jrdev, req->src, req->dst, src_nents, dst_nents, 0,
-+ 0, 0, 0);
- return ERR_PTR(-ENOMEM);
- }
-
-- if (!dst_nents && iv_dma + ivsize == sg_dma_address(req->dst))
-- iv_contig = true;
-- else
-- dst_nents = dst_nents ? : 1;
-- sec4_sg_bytes = ((iv_contig ? 0 : 1) + src_nents + dst_nents) *
-- sizeof(struct sec4_sg_entry);
-+ sec4_sg_ents = mapped_src_nents > 1 ? mapped_src_nents : 0;
-+ dst_sg_idx = sec4_sg_ents;
-+ if (mapped_dst_nents == 1 &&
-+ iv_dma + ivsize == sg_dma_address(req->dst)) {
-+ out_contig = true;
-+ } else {
-+ out_contig = false;
-+ sec4_sg_ents += 1 + mapped_dst_nents;
-+ }
-
- /* allocate space for base edesc and hw desc commands, link tables */
-+ sec4_sg_bytes = sec4_sg_ents * sizeof(struct sec4_sg_entry);
- edesc = kzalloc(sizeof(*edesc) + desc_bytes + sec4_sg_bytes,
- GFP_DMA | flags);
- if (!edesc) {
- dev_err(jrdev, "could not allocate extended descriptor\n");
-+ caam_unmap(jrdev, req->src, req->dst, src_nents, dst_nents,
-+ iv_dma, ivsize, 0, 0);
- return ERR_PTR(-ENOMEM);
- }
-
-@@ -2849,24 +1793,24 @@ static struct ablkcipher_edesc *ablkciph
- edesc->sec4_sg = (void *)edesc + sizeof(struct ablkcipher_edesc) +
- desc_bytes;
-
-- sec4_sg_index = 0;
-- if (src_nents) {
-- sg_to_sec4_sg_last(req->src, src_nents, edesc->sec4_sg, 0);
-- sec4_sg_index += src_nents;
-- }
-+ if (mapped_src_nents > 1)
-+ sg_to_sec4_sg_last(req->src, mapped_src_nents, edesc->sec4_sg,
-+ 0);
-
-- if (!iv_contig) {
-- dma_to_sec4_sg_one(edesc->sec4_sg + sec4_sg_index,
-+ if (!out_contig) {
-+ dma_to_sec4_sg_one(edesc->sec4_sg + dst_sg_idx,
- iv_dma, ivsize, 0);
-- sec4_sg_index += 1;
-- sg_to_sec4_sg_last(req->dst, dst_nents,
-- edesc->sec4_sg + sec4_sg_index, 0);
-+ sg_to_sec4_sg_last(req->dst, mapped_dst_nents,
-+ edesc->sec4_sg + dst_sg_idx + 1, 0);
- }
-
- edesc->sec4_sg_dma = dma_map_single(jrdev, edesc->sec4_sg,
- sec4_sg_bytes, DMA_TO_DEVICE);
- if (dma_mapping_error(jrdev, edesc->sec4_sg_dma)) {
- dev_err(jrdev, "unable to map S/G table\n");
-+ caam_unmap(jrdev, req->src, req->dst, src_nents, dst_nents,
-+ iv_dma, ivsize, 0, 0);
-+ kfree(edesc);
- return ERR_PTR(-ENOMEM);
- }
- edesc->iv_dma = iv_dma;
-@@ -2878,7 +1822,7 @@ static struct ablkcipher_edesc *ablkciph
- sec4_sg_bytes, 1);
- #endif
-
-- *iv_contig_out = iv_contig;
-+ *iv_contig_out = out_contig;
- return edesc;
- }
-
-@@ -2889,7 +1833,7 @@ static int ablkcipher_givencrypt(struct
- struct crypto_ablkcipher *ablkcipher = crypto_ablkcipher_reqtfm(req);
- struct caam_ctx *ctx = crypto_ablkcipher_ctx(ablkcipher);
- struct device *jrdev = ctx->jrdev;
-- bool iv_contig;
-+ bool iv_contig = false;
- u32 *desc;
- int ret = 0;
-
-@@ -2933,7 +1877,6 @@ struct caam_alg_template {
- } template_u;
- u32 class1_alg_type;
- u32 class2_alg_type;
-- u32 alg_op;
- };
-
- static struct caam_alg_template driver_algs[] = {
-@@ -3118,7 +2061,6 @@ static struct caam_aead_alg driver_aeads
- .caam = {
- .class2_alg_type = OP_ALG_ALGSEL_MD5 |
- OP_ALG_AAI_HMAC_PRECOMP,
-- .alg_op = OP_ALG_ALGSEL_MD5 | OP_ALG_AAI_HMAC,
- },
- },
- {
-@@ -3140,7 +2082,6 @@ static struct caam_aead_alg driver_aeads
- .caam = {
- .class2_alg_type = OP_ALG_ALGSEL_SHA1 |
- OP_ALG_AAI_HMAC_PRECOMP,
-- .alg_op = OP_ALG_ALGSEL_SHA1 | OP_ALG_AAI_HMAC,
- },
- },
- {
-@@ -3162,7 +2103,6 @@ static struct caam_aead_alg driver_aeads
- .caam = {
- .class2_alg_type = OP_ALG_ALGSEL_SHA224 |
- OP_ALG_AAI_HMAC_PRECOMP,
-- .alg_op = OP_ALG_ALGSEL_SHA224 | OP_ALG_AAI_HMAC,
- },
- },
- {
-@@ -3184,7 +2124,6 @@ static struct caam_aead_alg driver_aeads
- .caam = {
- .class2_alg_type = OP_ALG_ALGSEL_SHA256 |
- OP_ALG_AAI_HMAC_PRECOMP,
-- .alg_op = OP_ALG_ALGSEL_SHA256 | OP_ALG_AAI_HMAC,
- },
- },
- {
-@@ -3206,7 +2145,6 @@ static struct caam_aead_alg driver_aeads
- .caam = {
- .class2_alg_type = OP_ALG_ALGSEL_SHA384 |
- OP_ALG_AAI_HMAC_PRECOMP,
-- .alg_op = OP_ALG_ALGSEL_SHA384 | OP_ALG_AAI_HMAC,
- },
- },
- {
-@@ -3228,7 +2166,6 @@ static struct caam_aead_alg driver_aeads
- .caam = {
- .class2_alg_type = OP_ALG_ALGSEL_SHA512 |
- OP_ALG_AAI_HMAC_PRECOMP,
-- .alg_op = OP_ALG_ALGSEL_SHA512 | OP_ALG_AAI_HMAC,
- },
- },
- {
-@@ -3250,7 +2187,6 @@ static struct caam_aead_alg driver_aeads
- .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_CBC,
- .class2_alg_type = OP_ALG_ALGSEL_MD5 |
- OP_ALG_AAI_HMAC_PRECOMP,
-- .alg_op = OP_ALG_ALGSEL_MD5 | OP_ALG_AAI_HMAC,
- },
- },
- {
-@@ -3273,7 +2209,6 @@ static struct caam_aead_alg driver_aeads
- .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_CBC,
- .class2_alg_type = OP_ALG_ALGSEL_MD5 |
- OP_ALG_AAI_HMAC_PRECOMP,
-- .alg_op = OP_ALG_ALGSEL_MD5 | OP_ALG_AAI_HMAC,
- .geniv = true,
- },
- },
-@@ -3296,7 +2231,6 @@ static struct caam_aead_alg driver_aeads
- .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_CBC,
- .class2_alg_type = OP_ALG_ALGSEL_SHA1 |
- OP_ALG_AAI_HMAC_PRECOMP,
-- .alg_op = OP_ALG_ALGSEL_SHA1 | OP_ALG_AAI_HMAC,
- },
- },
- {
-@@ -3319,7 +2253,6 @@ static struct caam_aead_alg driver_aeads
- .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_CBC,
- .class2_alg_type = OP_ALG_ALGSEL_SHA1 |
- OP_ALG_AAI_HMAC_PRECOMP,
-- .alg_op = OP_ALG_ALGSEL_SHA1 | OP_ALG_AAI_HMAC,
- .geniv = true,
- },
- },
-@@ -3342,7 +2275,6 @@ static struct caam_aead_alg driver_aeads
- .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_CBC,
- .class2_alg_type = OP_ALG_ALGSEL_SHA224 |
- OP_ALG_AAI_HMAC_PRECOMP,
-- .alg_op = OP_ALG_ALGSEL_SHA224 | OP_ALG_AAI_HMAC,
- },
- },
- {
-@@ -3365,7 +2297,6 @@ static struct caam_aead_alg driver_aeads
- .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_CBC,
- .class2_alg_type = OP_ALG_ALGSEL_SHA224 |
- OP_ALG_AAI_HMAC_PRECOMP,
-- .alg_op = OP_ALG_ALGSEL_SHA224 | OP_ALG_AAI_HMAC,
- .geniv = true,
- },
- },
-@@ -3388,7 +2319,6 @@ static struct caam_aead_alg driver_aeads
- .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_CBC,
- .class2_alg_type = OP_ALG_ALGSEL_SHA256 |
- OP_ALG_AAI_HMAC_PRECOMP,
-- .alg_op = OP_ALG_ALGSEL_SHA256 | OP_ALG_AAI_HMAC,
- },
- },
- {
-@@ -3411,7 +2341,6 @@ static struct caam_aead_alg driver_aeads
- .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_CBC,
- .class2_alg_type = OP_ALG_ALGSEL_SHA256 |
- OP_ALG_AAI_HMAC_PRECOMP,
-- .alg_op = OP_ALG_ALGSEL_SHA256 | OP_ALG_AAI_HMAC,
- .geniv = true,
- },
- },
-@@ -3434,7 +2363,6 @@ static struct caam_aead_alg driver_aeads
- .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_CBC,
- .class2_alg_type = OP_ALG_ALGSEL_SHA384 |
- OP_ALG_AAI_HMAC_PRECOMP,
-- .alg_op = OP_ALG_ALGSEL_SHA384 | OP_ALG_AAI_HMAC,
- },
- },
- {
-@@ -3457,7 +2385,6 @@ static struct caam_aead_alg driver_aeads
- .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_CBC,
- .class2_alg_type = OP_ALG_ALGSEL_SHA384 |
- OP_ALG_AAI_HMAC_PRECOMP,
-- .alg_op = OP_ALG_ALGSEL_SHA384 | OP_ALG_AAI_HMAC,
- .geniv = true,
- },
- },
-@@ -3480,7 +2407,6 @@ static struct caam_aead_alg driver_aeads
- .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_CBC,
- .class2_alg_type = OP_ALG_ALGSEL_SHA512 |
- OP_ALG_AAI_HMAC_PRECOMP,
-- .alg_op = OP_ALG_ALGSEL_SHA512 | OP_ALG_AAI_HMAC,
- },
- },
- {
-@@ -3503,7 +2429,6 @@ static struct caam_aead_alg driver_aeads
- .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_CBC,
- .class2_alg_type = OP_ALG_ALGSEL_SHA512 |
- OP_ALG_AAI_HMAC_PRECOMP,
-- .alg_op = OP_ALG_ALGSEL_SHA512 | OP_ALG_AAI_HMAC,
- .geniv = true,
- },
- },
-@@ -3526,7 +2451,6 @@ static struct caam_aead_alg driver_aeads
- .class1_alg_type = OP_ALG_ALGSEL_3DES | OP_ALG_AAI_CBC,
- .class2_alg_type = OP_ALG_ALGSEL_MD5 |
- OP_ALG_AAI_HMAC_PRECOMP,
-- .alg_op = OP_ALG_ALGSEL_MD5 | OP_ALG_AAI_HMAC,
- }
- },
- {
-@@ -3549,7 +2473,6 @@ static struct caam_aead_alg driver_aeads
- .class1_alg_type = OP_ALG_ALGSEL_3DES | OP_ALG_AAI_CBC,
- .class2_alg_type = OP_ALG_ALGSEL_MD5 |
- OP_ALG_AAI_HMAC_PRECOMP,
-- .alg_op = OP_ALG_ALGSEL_MD5 | OP_ALG_AAI_HMAC,
- .geniv = true,
- }
- },
-@@ -3573,7 +2496,6 @@ static struct caam_aead_alg driver_aeads
- .class1_alg_type = OP_ALG_ALGSEL_3DES | OP_ALG_AAI_CBC,
- .class2_alg_type = OP_ALG_ALGSEL_SHA1 |
- OP_ALG_AAI_HMAC_PRECOMP,
-- .alg_op = OP_ALG_ALGSEL_SHA1 | OP_ALG_AAI_HMAC,
- },
- },
- {
-@@ -3597,7 +2519,6 @@ static struct caam_aead_alg driver_aeads
- .class1_alg_type = OP_ALG_ALGSEL_3DES | OP_ALG_AAI_CBC,
- .class2_alg_type = OP_ALG_ALGSEL_SHA1 |
- OP_ALG_AAI_HMAC_PRECOMP,
-- .alg_op = OP_ALG_ALGSEL_SHA1 | OP_ALG_AAI_HMAC,
- .geniv = true,
- },
- },
-@@ -3621,7 +2542,6 @@ static struct caam_aead_alg driver_aeads
- .class1_alg_type = OP_ALG_ALGSEL_3DES | OP_ALG_AAI_CBC,
- .class2_alg_type = OP_ALG_ALGSEL_SHA224 |
- OP_ALG_AAI_HMAC_PRECOMP,
-- .alg_op = OP_ALG_ALGSEL_SHA224 | OP_ALG_AAI_HMAC,
- },
- },
- {
-@@ -3645,7 +2565,6 @@ static struct caam_aead_alg driver_aeads
- .class1_alg_type = OP_ALG_ALGSEL_3DES | OP_ALG_AAI_CBC,
- .class2_alg_type = OP_ALG_ALGSEL_SHA224 |
- OP_ALG_AAI_HMAC_PRECOMP,
-- .alg_op = OP_ALG_ALGSEL_SHA224 | OP_ALG_AAI_HMAC,
- .geniv = true,
- },
- },
-@@ -3669,7 +2588,6 @@ static struct caam_aead_alg driver_aeads
- .class1_alg_type = OP_ALG_ALGSEL_3DES | OP_ALG_AAI_CBC,
- .class2_alg_type = OP_ALG_ALGSEL_SHA256 |
- OP_ALG_AAI_HMAC_PRECOMP,
-- .alg_op = OP_ALG_ALGSEL_SHA256 | OP_ALG_AAI_HMAC,
- },
- },
- {
-@@ -3693,7 +2611,6 @@ static struct caam_aead_alg driver_aeads
- .class1_alg_type = OP_ALG_ALGSEL_3DES | OP_ALG_AAI_CBC,
- .class2_alg_type = OP_ALG_ALGSEL_SHA256 |
- OP_ALG_AAI_HMAC_PRECOMP,
-- .alg_op = OP_ALG_ALGSEL_SHA256 | OP_ALG_AAI_HMAC,
- .geniv = true,
- },
- },
-@@ -3717,7 +2634,6 @@ static struct caam_aead_alg driver_aeads
- .class1_alg_type = OP_ALG_ALGSEL_3DES | OP_ALG_AAI_CBC,
- .class2_alg_type = OP_ALG_ALGSEL_SHA384 |
- OP_ALG_AAI_HMAC_PRECOMP,
-- .alg_op = OP_ALG_ALGSEL_SHA384 | OP_ALG_AAI_HMAC,
- },
- },
- {
-@@ -3741,7 +2657,6 @@ static struct caam_aead_alg driver_aeads
- .class1_alg_type = OP_ALG_ALGSEL_3DES | OP_ALG_AAI_CBC,
- .class2_alg_type = OP_ALG_ALGSEL_SHA384 |
- OP_ALG_AAI_HMAC_PRECOMP,
-- .alg_op = OP_ALG_ALGSEL_SHA384 | OP_ALG_AAI_HMAC,
- .geniv = true,
- },
- },
-@@ -3765,7 +2680,6 @@ static struct caam_aead_alg driver_aeads
- .class1_alg_type = OP_ALG_ALGSEL_3DES | OP_ALG_AAI_CBC,
- .class2_alg_type = OP_ALG_ALGSEL_SHA512 |
- OP_ALG_AAI_HMAC_PRECOMP,
-- .alg_op = OP_ALG_ALGSEL_SHA512 | OP_ALG_AAI_HMAC,
- },
- },
- {
-@@ -3789,7 +2703,6 @@ static struct caam_aead_alg driver_aeads
- .class1_alg_type = OP_ALG_ALGSEL_3DES | OP_ALG_AAI_CBC,
- .class2_alg_type = OP_ALG_ALGSEL_SHA512 |
- OP_ALG_AAI_HMAC_PRECOMP,
-- .alg_op = OP_ALG_ALGSEL_SHA512 | OP_ALG_AAI_HMAC,
- .geniv = true,
- },
- },
-@@ -3812,7 +2725,6 @@ static struct caam_aead_alg driver_aeads
- .class1_alg_type = OP_ALG_ALGSEL_DES | OP_ALG_AAI_CBC,
- .class2_alg_type = OP_ALG_ALGSEL_MD5 |
- OP_ALG_AAI_HMAC_PRECOMP,
-- .alg_op = OP_ALG_ALGSEL_MD5 | OP_ALG_AAI_HMAC,
- },
- },
- {
-@@ -3835,7 +2747,6 @@ static struct caam_aead_alg driver_aeads
- .class1_alg_type = OP_ALG_ALGSEL_DES | OP_ALG_AAI_CBC,
- .class2_alg_type = OP_ALG_ALGSEL_MD5 |
- OP_ALG_AAI_HMAC_PRECOMP,
-- .alg_op = OP_ALG_ALGSEL_MD5 | OP_ALG_AAI_HMAC,
- .geniv = true,
- },
- },
-@@ -3858,7 +2769,6 @@ static struct caam_aead_alg driver_aeads
- .class1_alg_type = OP_ALG_ALGSEL_DES | OP_ALG_AAI_CBC,
- .class2_alg_type = OP_ALG_ALGSEL_SHA1 |
- OP_ALG_AAI_HMAC_PRECOMP,
-- .alg_op = OP_ALG_ALGSEL_SHA1 | OP_ALG_AAI_HMAC,
- },
- },
- {
-@@ -3881,7 +2791,6 @@ static struct caam_aead_alg driver_aeads
- .class1_alg_type = OP_ALG_ALGSEL_DES | OP_ALG_AAI_CBC,
- .class2_alg_type = OP_ALG_ALGSEL_SHA1 |
- OP_ALG_AAI_HMAC_PRECOMP,
-- .alg_op = OP_ALG_ALGSEL_SHA1 | OP_ALG_AAI_HMAC,
- .geniv = true,
- },
- },
-@@ -3904,7 +2813,6 @@ static struct caam_aead_alg driver_aeads
- .class1_alg_type = OP_ALG_ALGSEL_DES | OP_ALG_AAI_CBC,
- .class2_alg_type = OP_ALG_ALGSEL_SHA224 |
- OP_ALG_AAI_HMAC_PRECOMP,
-- .alg_op = OP_ALG_ALGSEL_SHA224 | OP_ALG_AAI_HMAC,
- },
- },
- {
-@@ -3927,7 +2835,6 @@ static struct caam_aead_alg driver_aeads
- .class1_alg_type = OP_ALG_ALGSEL_DES | OP_ALG_AAI_CBC,
- .class2_alg_type = OP_ALG_ALGSEL_SHA224 |
- OP_ALG_AAI_HMAC_PRECOMP,
-- .alg_op = OP_ALG_ALGSEL_SHA224 | OP_ALG_AAI_HMAC,
- .geniv = true,
- },
- },
-@@ -3950,7 +2857,6 @@ static struct caam_aead_alg driver_aeads
- .class1_alg_type = OP_ALG_ALGSEL_DES | OP_ALG_AAI_CBC,
- .class2_alg_type = OP_ALG_ALGSEL_SHA256 |
- OP_ALG_AAI_HMAC_PRECOMP,
-- .alg_op = OP_ALG_ALGSEL_SHA256 | OP_ALG_AAI_HMAC,
- },
- },
- {
-@@ -3973,7 +2879,6 @@ static struct caam_aead_alg driver_aeads
- .class1_alg_type = OP_ALG_ALGSEL_DES | OP_ALG_AAI_CBC,
- .class2_alg_type = OP_ALG_ALGSEL_SHA256 |
- OP_ALG_AAI_HMAC_PRECOMP,
-- .alg_op = OP_ALG_ALGSEL_SHA256 | OP_ALG_AAI_HMAC,
- .geniv = true,
- },
- },
-@@ -3996,7 +2901,6 @@ static struct caam_aead_alg driver_aeads
- .class1_alg_type = OP_ALG_ALGSEL_DES | OP_ALG_AAI_CBC,
- .class2_alg_type = OP_ALG_ALGSEL_SHA384 |
- OP_ALG_AAI_HMAC_PRECOMP,
-- .alg_op = OP_ALG_ALGSEL_SHA384 | OP_ALG_AAI_HMAC,
- },
- },
- {
-@@ -4019,7 +2923,6 @@ static struct caam_aead_alg driver_aeads
- .class1_alg_type = OP_ALG_ALGSEL_DES | OP_ALG_AAI_CBC,
- .class2_alg_type = OP_ALG_ALGSEL_SHA384 |
- OP_ALG_AAI_HMAC_PRECOMP,
-- .alg_op = OP_ALG_ALGSEL_SHA384 | OP_ALG_AAI_HMAC,
- .geniv = true,
- },
- },
-@@ -4042,7 +2945,6 @@ static struct caam_aead_alg driver_aeads
- .class1_alg_type = OP_ALG_ALGSEL_DES | OP_ALG_AAI_CBC,
- .class2_alg_type = OP_ALG_ALGSEL_SHA512 |
- OP_ALG_AAI_HMAC_PRECOMP,
-- .alg_op = OP_ALG_ALGSEL_SHA512 | OP_ALG_AAI_HMAC,
- },
- },
- {
-@@ -4065,7 +2967,6 @@ static struct caam_aead_alg driver_aeads
- .class1_alg_type = OP_ALG_ALGSEL_DES | OP_ALG_AAI_CBC,
- .class2_alg_type = OP_ALG_ALGSEL_SHA512 |
- OP_ALG_AAI_HMAC_PRECOMP,
-- .alg_op = OP_ALG_ALGSEL_SHA512 | OP_ALG_AAI_HMAC,
- .geniv = true,
- },
- },
-@@ -4090,7 +2991,6 @@ static struct caam_aead_alg driver_aeads
- OP_ALG_AAI_CTR_MOD128,
- .class2_alg_type = OP_ALG_ALGSEL_MD5 |
- OP_ALG_AAI_HMAC_PRECOMP,
-- .alg_op = OP_ALG_ALGSEL_MD5 | OP_ALG_AAI_HMAC,
- .rfc3686 = true,
- },
- },
-@@ -4115,7 +3015,6 @@ static struct caam_aead_alg driver_aeads
- OP_ALG_AAI_CTR_MOD128,
- .class2_alg_type = OP_ALG_ALGSEL_MD5 |
- OP_ALG_AAI_HMAC_PRECOMP,
-- .alg_op = OP_ALG_ALGSEL_MD5 | OP_ALG_AAI_HMAC,
- .rfc3686 = true,
- .geniv = true,
- },
-@@ -4141,7 +3040,6 @@ static struct caam_aead_alg driver_aeads
- OP_ALG_AAI_CTR_MOD128,
- .class2_alg_type = OP_ALG_ALGSEL_SHA1 |
- OP_ALG_AAI_HMAC_PRECOMP,
-- .alg_op = OP_ALG_ALGSEL_SHA1 | OP_ALG_AAI_HMAC,
- .rfc3686 = true,
- },
- },
-@@ -4166,7 +3064,6 @@ static struct caam_aead_alg driver_aeads
- OP_ALG_AAI_CTR_MOD128,
- .class2_alg_type = OP_ALG_ALGSEL_SHA1 |
- OP_ALG_AAI_HMAC_PRECOMP,
-- .alg_op = OP_ALG_ALGSEL_SHA1 | OP_ALG_AAI_HMAC,
- .rfc3686 = true,
- .geniv = true,
- },
-@@ -4192,7 +3089,6 @@ static struct caam_aead_alg driver_aeads
- OP_ALG_AAI_CTR_MOD128,
- .class2_alg_type = OP_ALG_ALGSEL_SHA224 |
- OP_ALG_AAI_HMAC_PRECOMP,
-- .alg_op = OP_ALG_ALGSEL_SHA224 | OP_ALG_AAI_HMAC,
- .rfc3686 = true,
- },
- },
-@@ -4217,7 +3113,6 @@ static struct caam_aead_alg driver_aeads
- OP_ALG_AAI_CTR_MOD128,
- .class2_alg_type = OP_ALG_ALGSEL_SHA224 |
- OP_ALG_AAI_HMAC_PRECOMP,
-- .alg_op = OP_ALG_ALGSEL_SHA224 | OP_ALG_AAI_HMAC,
- .rfc3686 = true,
- .geniv = true,
- },
-@@ -4243,7 +3138,6 @@ static struct caam_aead_alg driver_aeads
- OP_ALG_AAI_CTR_MOD128,
- .class2_alg_type = OP_ALG_ALGSEL_SHA256 |
- OP_ALG_AAI_HMAC_PRECOMP,
-- .alg_op = OP_ALG_ALGSEL_SHA256 | OP_ALG_AAI_HMAC,
- .rfc3686 = true,
- },
- },
-@@ -4268,7 +3162,6 @@ static struct caam_aead_alg driver_aeads
- OP_ALG_AAI_CTR_MOD128,
- .class2_alg_type = OP_ALG_ALGSEL_SHA256 |
- OP_ALG_AAI_HMAC_PRECOMP,
-- .alg_op = OP_ALG_ALGSEL_SHA256 | OP_ALG_AAI_HMAC,
- .rfc3686 = true,
- .geniv = true,
- },
-@@ -4294,7 +3187,6 @@ static struct caam_aead_alg driver_aeads
- OP_ALG_AAI_CTR_MOD128,
- .class2_alg_type = OP_ALG_ALGSEL_SHA384 |
- OP_ALG_AAI_HMAC_PRECOMP,
-- .alg_op = OP_ALG_ALGSEL_SHA384 | OP_ALG_AAI_HMAC,
- .rfc3686 = true,
- },
- },
-@@ -4319,7 +3211,6 @@ static struct caam_aead_alg driver_aeads
- OP_ALG_AAI_CTR_MOD128,
- .class2_alg_type = OP_ALG_ALGSEL_SHA384 |
- OP_ALG_AAI_HMAC_PRECOMP,
-- .alg_op = OP_ALG_ALGSEL_SHA384 | OP_ALG_AAI_HMAC,
- .rfc3686 = true,
- .geniv = true,
- },
-@@ -4345,7 +3236,6 @@ static struct caam_aead_alg driver_aeads
- OP_ALG_AAI_CTR_MOD128,
- .class2_alg_type = OP_ALG_ALGSEL_SHA512 |
- OP_ALG_AAI_HMAC_PRECOMP,
-- .alg_op = OP_ALG_ALGSEL_SHA512 | OP_ALG_AAI_HMAC,
- .rfc3686 = true,
- },
- },
-@@ -4370,7 +3260,6 @@ static struct caam_aead_alg driver_aeads
- OP_ALG_AAI_CTR_MOD128,
- .class2_alg_type = OP_ALG_ALGSEL_SHA512 |
- OP_ALG_AAI_HMAC_PRECOMP,
-- .alg_op = OP_ALG_ALGSEL_SHA512 | OP_ALG_AAI_HMAC,
- .rfc3686 = true,
- .geniv = true,
- },
-@@ -4383,18 +3272,44 @@ struct caam_crypto_alg {
- struct caam_alg_entry caam;
- };
-
--static int caam_init_common(struct caam_ctx *ctx, struct caam_alg_entry *caam)
-+static int caam_init_common(struct caam_ctx *ctx, struct caam_alg_entry *caam,
-+ bool uses_dkp)
- {
-+ dma_addr_t dma_addr;
-+ struct caam_drv_private *priv;
-+
- ctx->jrdev = caam_jr_alloc();
- if (IS_ERR(ctx->jrdev)) {
- pr_err("Job Ring Device allocation for transform failed\n");
- return PTR_ERR(ctx->jrdev);
- }
-
-+ priv = dev_get_drvdata(ctx->jrdev->parent);
-+ if (priv->era >= 6 && uses_dkp)
-+ ctx->dir = DMA_BIDIRECTIONAL;
-+ else
-+ ctx->dir = DMA_TO_DEVICE;
-+
-+ dma_addr = dma_map_single_attrs(ctx->jrdev, ctx->sh_desc_enc,
-+ offsetof(struct caam_ctx,
-+ sh_desc_enc_dma),
-+ ctx->dir, DMA_ATTR_SKIP_CPU_SYNC);
-+ if (dma_mapping_error(ctx->jrdev, dma_addr)) {
-+ dev_err(ctx->jrdev, "unable to map key, shared descriptors\n");
-+ caam_jr_free(ctx->jrdev);
-+ return -ENOMEM;
-+ }
-+
-+ ctx->sh_desc_enc_dma = dma_addr;
-+ ctx->sh_desc_dec_dma = dma_addr + offsetof(struct caam_ctx,
-+ sh_desc_dec);
-+ ctx->sh_desc_givenc_dma = dma_addr + offsetof(struct caam_ctx,
-+ sh_desc_givenc);
-+ ctx->key_dma = dma_addr + offsetof(struct caam_ctx, key);
-+
- /* copy descriptor header template value */
-- ctx->class1_alg_type = OP_TYPE_CLASS1_ALG | caam->class1_alg_type;
-- ctx->class2_alg_type = OP_TYPE_CLASS2_ALG | caam->class2_alg_type;
-- ctx->alg_op = OP_TYPE_CLASS2_ALG | caam->alg_op;
-+ ctx->cdata.algtype = OP_TYPE_CLASS1_ALG | caam->class1_alg_type;
-+ ctx->adata.algtype = OP_TYPE_CLASS2_ALG | caam->class2_alg_type;
-
- return 0;
- }
-@@ -4406,7 +3321,7 @@ static int caam_cra_init(struct crypto_t
- container_of(alg, struct caam_crypto_alg, crypto_alg);
- struct caam_ctx *ctx = crypto_tfm_ctx(tfm);
-
-- return caam_init_common(ctx, &caam_alg->caam);
-+ return caam_init_common(ctx, &caam_alg->caam, false);
- }
-
- static int caam_aead_init(struct crypto_aead *tfm)
-@@ -4416,30 +3331,15 @@ static int caam_aead_init(struct crypto_
- container_of(alg, struct caam_aead_alg, aead);
- struct caam_ctx *ctx = crypto_aead_ctx(tfm);
-
-- return caam_init_common(ctx, &caam_alg->caam);
-+ return caam_init_common(ctx, &caam_alg->caam,
-+ alg->setkey == aead_setkey);
- }
-
- static void caam_exit_common(struct caam_ctx *ctx)
- {
-- if (ctx->sh_desc_enc_dma &&
-- !dma_mapping_error(ctx->jrdev, ctx->sh_desc_enc_dma))
-- dma_unmap_single(ctx->jrdev, ctx->sh_desc_enc_dma,
-- desc_bytes(ctx->sh_desc_enc), DMA_TO_DEVICE);
-- if (ctx->sh_desc_dec_dma &&
-- !dma_mapping_error(ctx->jrdev, ctx->sh_desc_dec_dma))
-- dma_unmap_single(ctx->jrdev, ctx->sh_desc_dec_dma,
-- desc_bytes(ctx->sh_desc_dec), DMA_TO_DEVICE);
-- if (ctx->sh_desc_givenc_dma &&
-- !dma_mapping_error(ctx->jrdev, ctx->sh_desc_givenc_dma))
-- dma_unmap_single(ctx->jrdev, ctx->sh_desc_givenc_dma,
-- desc_bytes(ctx->sh_desc_givenc),
-- DMA_TO_DEVICE);
-- if (ctx->key_dma &&
-- !dma_mapping_error(ctx->jrdev, ctx->key_dma))
-- dma_unmap_single(ctx->jrdev, ctx->key_dma,
-- ctx->enckeylen + ctx->split_key_pad_len,
-- DMA_TO_DEVICE);
--
-+ dma_unmap_single_attrs(ctx->jrdev, ctx->sh_desc_enc_dma,
-+ offsetof(struct caam_ctx, sh_desc_enc_dma),
-+ ctx->dir, DMA_ATTR_SKIP_CPU_SYNC);
- caam_jr_free(ctx->jrdev);
- }
-
-@@ -4515,7 +3415,6 @@ static struct caam_crypto_alg *caam_alg_
-
- t_alg->caam.class1_alg_type = template->class1_alg_type;
- t_alg->caam.class2_alg_type = template->class2_alg_type;
-- t_alg->caam.alg_op = template->alg_op;
-
- return t_alg;
- }
---- /dev/null
-+++ b/drivers/crypto/caam/caamalg_desc.c
-@@ -0,0 +1,1961 @@
-+/*
-+ * Shared descriptors for aead, ablkcipher algorithms
-+ *
-+ * Copyright 2016 NXP
-+ */
-+
-+#include "compat.h"
-+#include "desc_constr.h"
-+#include "caamalg_desc.h"
-+
-+/*
-+ * For aead functions, read payload and write payload,
-+ * both of which are specified in req->src and req->dst
-+ */
-+static inline void aead_append_src_dst(u32 *desc, u32 msg_type)
-+{
-+ append_seq_fifo_store(desc, 0, FIFOST_TYPE_MESSAGE_DATA | KEY_VLF);
-+ append_seq_fifo_load(desc, 0, FIFOLD_CLASS_BOTH |
-+ KEY_VLF | msg_type | FIFOLD_TYPE_LASTBOTH);
-+}
-+
-+/* Set DK bit in class 1 operation if shared */
-+static inline void append_dec_op1(u32 *desc, u32 type)
-+{
-+ u32 *jump_cmd, *uncond_jump_cmd;
-+
-+ /* DK bit is valid only for AES */
-+ if ((type & OP_ALG_ALGSEL_MASK) != OP_ALG_ALGSEL_AES) {
-+ append_operation(desc, type | OP_ALG_AS_INITFINAL |
-+ OP_ALG_DECRYPT);
-+ return;
-+ }
-+
-+ jump_cmd = append_jump(desc, JUMP_TEST_ALL | JUMP_COND_SHRD);
-+ append_operation(desc, type | OP_ALG_AS_INITFINAL |
-+ OP_ALG_DECRYPT);
-+ uncond_jump_cmd = append_jump(desc, JUMP_TEST_ALL);
-+ set_jump_tgt_here(desc, jump_cmd);
-+ append_operation(desc, type | OP_ALG_AS_INITFINAL |
-+ OP_ALG_DECRYPT | OP_ALG_AAI_DK);
-+ set_jump_tgt_here(desc, uncond_jump_cmd);
-+}
-+
-+/**
-+ * cnstr_shdsc_aead_null_encap - IPSec ESP encapsulation shared descriptor
-+ * (non-protocol) with no (null) encryption.
-+ * @desc: pointer to buffer used for descriptor construction
-+ * @adata: pointer to authentication transform definitions.
-+ * A split key is required for SEC Era < 6; the size of the split key
-+ * is specified in this case. Valid algorithm values - one of
-+ * OP_ALG_ALGSEL_{MD5, SHA1, SHA224, SHA256, SHA384, SHA512} ANDed
-+ * with OP_ALG_AAI_HMAC_PRECOMP.
-+ * @icvsize: integrity check value (ICV) size (truncated or full)
-+ * @era: SEC Era
-+ */
-+void cnstr_shdsc_aead_null_encap(u32 * const desc, struct alginfo *adata,
-+ unsigned int icvsize, int era)
-+{
-+ u32 *key_jump_cmd, *read_move_cmd, *write_move_cmd;
-+
-+ init_sh_desc(desc, HDR_SHARE_SERIAL);
-+
-+ /* Skip if already shared */
-+ key_jump_cmd = append_jump(desc, JUMP_JSL | JUMP_TEST_ALL |
-+ JUMP_COND_SHRD);
-+ if (era < 6) {
-+ if (adata->key_inline)
-+ append_key_as_imm(desc, adata->key_virt,
-+ adata->keylen_pad, adata->keylen,
-+ CLASS_2 | KEY_DEST_MDHA_SPLIT |
-+ KEY_ENC);
-+ else
-+ append_key(desc, adata->key_dma, adata->keylen,
-+ CLASS_2 | KEY_DEST_MDHA_SPLIT | KEY_ENC);
-+ } else {
-+ append_proto_dkp(desc, adata);
-+ }
-+ set_jump_tgt_here(desc, key_jump_cmd);
-+
-+ /* assoclen + cryptlen = seqinlen */
-+ append_math_sub(desc, REG3, SEQINLEN, REG0, CAAM_CMD_SZ);
-+
-+ /* Prepare to read and write cryptlen + assoclen bytes */
-+ append_math_add(desc, VARSEQINLEN, ZERO, REG3, CAAM_CMD_SZ);
-+ append_math_add(desc, VARSEQOUTLEN, ZERO, REG3, CAAM_CMD_SZ);
-+
-+ /*
-+ * MOVE_LEN opcode is not available in all SEC HW revisions,
-+ * thus need to do some magic, i.e. self-patch the descriptor
-+ * buffer.
-+ */
-+ read_move_cmd = append_move(desc, MOVE_SRC_DESCBUF |
-+ MOVE_DEST_MATH3 |
-+ (0x6 << MOVE_LEN_SHIFT));
-+ write_move_cmd = append_move(desc, MOVE_SRC_MATH3 |
-+ MOVE_DEST_DESCBUF |
-+ MOVE_WAITCOMP |
-+ (0x8 << MOVE_LEN_SHIFT));
-+
-+ /* Class 2 operation */
-+ append_operation(desc, adata->algtype | OP_ALG_AS_INITFINAL |
-+ OP_ALG_ENCRYPT);
-+
-+ /* Read and write cryptlen bytes */
-+ aead_append_src_dst(desc, FIFOLD_TYPE_MSG | FIFOLD_TYPE_FLUSH1);
-+
-+ set_move_tgt_here(desc, read_move_cmd);
-+ set_move_tgt_here(desc, write_move_cmd);
-+ append_cmd(desc, CMD_LOAD | DISABLE_AUTO_INFO_FIFO);
-+ append_move(desc, MOVE_SRC_INFIFO_CL | MOVE_DEST_OUTFIFO |
-+ MOVE_AUX_LS);
-+
-+ /* Write ICV */
-+ append_seq_store(desc, icvsize, LDST_CLASS_2_CCB |
-+ LDST_SRCDST_BYTE_CONTEXT);
-+
-+#ifdef DEBUG
-+ print_hex_dump(KERN_ERR,
-+ "aead null enc shdesc@" __stringify(__LINE__)": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, desc, desc_bytes(desc), 1);
-+#endif
-+}
-+EXPORT_SYMBOL(cnstr_shdsc_aead_null_encap);
-+
-+/**
-+ * cnstr_shdsc_aead_null_decap - IPSec ESP decapsulation shared descriptor
-+ * (non-protocol) with no (null) decryption.
-+ * @desc: pointer to buffer used for descriptor construction
-+ * @adata: pointer to authentication transform definitions.
-+ * A split key is required for SEC Era < 6; the size of the split key
-+ * is specified in this case. Valid algorithm values - one of
-+ * OP_ALG_ALGSEL_{MD5, SHA1, SHA224, SHA256, SHA384, SHA512} ANDed
-+ * with OP_ALG_AAI_HMAC_PRECOMP.
-+ * @icvsize: integrity check value (ICV) size (truncated or full)
-+ * @era: SEC Era
-+ */
-+void cnstr_shdsc_aead_null_decap(u32 * const desc, struct alginfo *adata,
-+ unsigned int icvsize, int era)
-+{
-+ u32 *key_jump_cmd, *read_move_cmd, *write_move_cmd, *jump_cmd;
-+
-+ init_sh_desc(desc, HDR_SHARE_SERIAL);
-+
-+ /* Skip if already shared */
-+ key_jump_cmd = append_jump(desc, JUMP_JSL | JUMP_TEST_ALL |
-+ JUMP_COND_SHRD);
-+ if (era < 6) {
-+ if (adata->key_inline)
-+ append_key_as_imm(desc, adata->key_virt,
-+ adata->keylen_pad, adata->keylen,
-+ CLASS_2 | KEY_DEST_MDHA_SPLIT |
-+ KEY_ENC);
-+ else
-+ append_key(desc, adata->key_dma, adata->keylen,
-+ CLASS_2 | KEY_DEST_MDHA_SPLIT | KEY_ENC);
-+ } else {
-+ append_proto_dkp(desc, adata);
-+ }
-+ set_jump_tgt_here(desc, key_jump_cmd);
-+
-+ /* Class 2 operation */
-+ append_operation(desc, adata->algtype | OP_ALG_AS_INITFINAL |
-+ OP_ALG_DECRYPT | OP_ALG_ICV_ON);
-+
-+ /* assoclen + cryptlen = seqoutlen */
-+ append_math_sub(desc, REG2, SEQOUTLEN, REG0, CAAM_CMD_SZ);
-+
-+ /* Prepare to read and write cryptlen + assoclen bytes */
-+ append_math_add(desc, VARSEQINLEN, ZERO, REG2, CAAM_CMD_SZ);
-+ append_math_add(desc, VARSEQOUTLEN, ZERO, REG2, CAAM_CMD_SZ);
-+
-+ /*
-+ * MOVE_LEN opcode is not available in all SEC HW revisions,
-+ * thus need to do some magic, i.e. self-patch the descriptor
-+ * buffer.
-+ */
-+ read_move_cmd = append_move(desc, MOVE_SRC_DESCBUF |
-+ MOVE_DEST_MATH2 |
-+ (0x6 << MOVE_LEN_SHIFT));
-+ write_move_cmd = append_move(desc, MOVE_SRC_MATH2 |
-+ MOVE_DEST_DESCBUF |
-+ MOVE_WAITCOMP |
-+ (0x8 << MOVE_LEN_SHIFT));
-+
-+ /* Read and write cryptlen bytes */
-+ aead_append_src_dst(desc, FIFOLD_TYPE_MSG | FIFOLD_TYPE_FLUSH1);
-+
-+ /*
-+ * Insert a NOP here, since we need at least 4 instructions between
-+ * code patching the descriptor buffer and the location being patched.
-+ */
-+ jump_cmd = append_jump(desc, JUMP_TEST_ALL);
-+ set_jump_tgt_here(desc, jump_cmd);
-+
-+ set_move_tgt_here(desc, read_move_cmd);
-+ set_move_tgt_here(desc, write_move_cmd);
-+ append_cmd(desc, CMD_LOAD | DISABLE_AUTO_INFO_FIFO);
-+ append_move(desc, MOVE_SRC_INFIFO_CL | MOVE_DEST_OUTFIFO |
-+ MOVE_AUX_LS);
-+ append_cmd(desc, CMD_LOAD | ENABLE_AUTO_INFO_FIFO);
-+
-+ /* Load ICV */
-+ append_seq_fifo_load(desc, icvsize, FIFOLD_CLASS_CLASS2 |
-+ FIFOLD_TYPE_LAST2 | FIFOLD_TYPE_ICV);
-+
-+#ifdef DEBUG
-+ print_hex_dump(KERN_ERR,
-+ "aead null dec shdesc@" __stringify(__LINE__)": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, desc, desc_bytes(desc), 1);
-+#endif
-+}
-+EXPORT_SYMBOL(cnstr_shdsc_aead_null_decap);
-+
-+static void init_sh_desc_key_aead(u32 * const desc,
-+ struct alginfo * const cdata,
-+ struct alginfo * const adata,
-+ const bool is_rfc3686, u32 *nonce, int era)
-+{
-+ u32 *key_jump_cmd;
-+ unsigned int enckeylen = cdata->keylen;
-+
-+ /* Note: Context registers are saved. */
-+ init_sh_desc(desc, HDR_SHARE_SERIAL | HDR_SAVECTX);
-+
-+ /* Skip if already shared */
-+ key_jump_cmd = append_jump(desc, JUMP_JSL | JUMP_TEST_ALL |
-+ JUMP_COND_SHRD);
-+
-+ /*
-+ * RFC3686 specific:
-+ * | key = {AUTH_KEY, ENC_KEY, NONCE}
-+ * | enckeylen = encryption key size + nonce size
-+ */
-+ if (is_rfc3686)
-+ enckeylen -= CTR_RFC3686_NONCE_SIZE;
-+
-+ if (era < 6) {
-+ if (adata->key_inline)
-+ append_key_as_imm(desc, adata->key_virt,
-+ adata->keylen_pad, adata->keylen,
-+ CLASS_2 | KEY_DEST_MDHA_SPLIT |
-+ KEY_ENC);
-+ else
-+ append_key(desc, adata->key_dma, adata->keylen,
-+ CLASS_2 | KEY_DEST_MDHA_SPLIT | KEY_ENC);
-+ } else {
-+ append_proto_dkp(desc, adata);
-+ }
-+
-+ if (cdata->key_inline)
-+ append_key_as_imm(desc, cdata->key_virt, enckeylen,
-+ enckeylen, CLASS_1 | KEY_DEST_CLASS_REG);
-+ else
-+ append_key(desc, cdata->key_dma, enckeylen, CLASS_1 |
-+ KEY_DEST_CLASS_REG);
-+
-+ /* Load Counter into CONTEXT1 reg */
-+ if (is_rfc3686) {
-+ append_load_as_imm(desc, nonce, CTR_RFC3686_NONCE_SIZE,
-+ LDST_CLASS_IND_CCB |
-+ LDST_SRCDST_BYTE_OUTFIFO | LDST_IMM);
-+ append_move(desc,
-+ MOVE_SRC_OUTFIFO |
-+ MOVE_DEST_CLASS1CTX |
-+ (16 << MOVE_OFFSET_SHIFT) |
-+ (CTR_RFC3686_NONCE_SIZE << MOVE_LEN_SHIFT));
-+ }
-+
-+ set_jump_tgt_here(desc, key_jump_cmd);
-+}
-+
-+/**
-+ * cnstr_shdsc_aead_encap - IPSec ESP encapsulation shared descriptor
-+ * (non-protocol).
-+ * @desc: pointer to buffer used for descriptor construction
-+ * @cdata: pointer to block cipher transform definitions
-+ * Valid algorithm values - one of OP_ALG_ALGSEL_{AES, DES, 3DES} ANDed
-+ * with OP_ALG_AAI_CBC or OP_ALG_AAI_CTR_MOD128.
-+ * @adata: pointer to authentication transform definitions.
-+ * A split key is required for SEC Era < 6; the size of the split key
-+ * is specified in this case. Valid algorithm values - one of
-+ * OP_ALG_ALGSEL_{MD5, SHA1, SHA224, SHA256, SHA384, SHA512} ANDed
-+ * with OP_ALG_AAI_HMAC_PRECOMP.
-+ * @ivsize: initialization vector size
-+ * @icvsize: integrity check value (ICV) size (truncated or full)
-+ * @is_rfc3686: true when ctr(aes) is wrapped by rfc3686 template
-+ * @nonce: pointer to rfc3686 nonce
-+ * @ctx1_iv_off: IV offset in CONTEXT1 register
-+ * @is_qi: true when called from caam/qi
-+ * @era: SEC Era
-+ */
-+void cnstr_shdsc_aead_encap(u32 * const desc, struct alginfo *cdata,
-+ struct alginfo *adata, unsigned int ivsize,
-+ unsigned int icvsize, const bool is_rfc3686,
-+ u32 *nonce, const u32 ctx1_iv_off, const bool is_qi,
-+ int era)
-+{
-+ /* Note: Context registers are saved. */
-+ init_sh_desc_key_aead(desc, cdata, adata, is_rfc3686, nonce, era);
-+
-+ /* Class 2 operation */
-+ append_operation(desc, adata->algtype | OP_ALG_AS_INITFINAL |
-+ OP_ALG_ENCRYPT);
-+
-+ if (is_qi) {
-+ u32 *wait_load_cmd;
-+
-+ /* REG3 = assoclen */
-+ append_seq_load(desc, 4, LDST_CLASS_DECO |
-+ LDST_SRCDST_WORD_DECO_MATH3 |
-+ (4 << LDST_OFFSET_SHIFT));
-+
-+ wait_load_cmd = append_jump(desc, JUMP_JSL | JUMP_TEST_ALL |
-+ JUMP_COND_CALM | JUMP_COND_NCP |
-+ JUMP_COND_NOP | JUMP_COND_NIP |
-+ JUMP_COND_NIFP);
-+ set_jump_tgt_here(desc, wait_load_cmd);
-+
-+ append_seq_load(desc, ivsize, LDST_CLASS_1_CCB |
-+ LDST_SRCDST_BYTE_CONTEXT |
-+ (ctx1_iv_off << LDST_OFFSET_SHIFT));
-+ }
-+
-+ /* Read and write assoclen bytes */
-+ if (is_qi || era < 3) {
-+ append_math_add(desc, VARSEQINLEN, ZERO, REG3, CAAM_CMD_SZ);
-+ append_math_add(desc, VARSEQOUTLEN, ZERO, REG3, CAAM_CMD_SZ);
-+ } else {
-+ append_math_add(desc, VARSEQINLEN, ZERO, DPOVRD, CAAM_CMD_SZ);
-+ append_math_add(desc, VARSEQOUTLEN, ZERO, DPOVRD, CAAM_CMD_SZ);
-+ }
-+
-+ /* Skip assoc data */
-+ append_seq_fifo_store(desc, 0, FIFOST_TYPE_SKIP | FIFOLDST_VLF);
-+
-+ /* read assoc before reading payload */
-+ append_seq_fifo_load(desc, 0, FIFOLD_CLASS_CLASS2 | FIFOLD_TYPE_MSG |
-+ FIFOLDST_VLF);
-+
-+ /* Load Counter into CONTEXT1 reg */
-+ if (is_rfc3686)
-+ append_load_imm_be32(desc, 1, LDST_IMM | LDST_CLASS_1_CCB |
-+ LDST_SRCDST_BYTE_CONTEXT |
-+ ((ctx1_iv_off + CTR_RFC3686_IV_SIZE) <<
-+ LDST_OFFSET_SHIFT));
-+
-+ /* Class 1 operation */
-+ append_operation(desc, cdata->algtype | OP_ALG_AS_INITFINAL |
-+ OP_ALG_ENCRYPT);
-+
-+ /* Read and write cryptlen bytes */
-+ append_math_add(desc, VARSEQINLEN, SEQINLEN, REG0, CAAM_CMD_SZ);
-+ append_math_add(desc, VARSEQOUTLEN, SEQINLEN, REG0, CAAM_CMD_SZ);
-+ aead_append_src_dst(desc, FIFOLD_TYPE_MSG1OUT2);
-+
-+ /* Write ICV */
-+ append_seq_store(desc, icvsize, LDST_CLASS_2_CCB |
-+ LDST_SRCDST_BYTE_CONTEXT);
-+
-+#ifdef DEBUG
-+ print_hex_dump(KERN_ERR, "aead enc shdesc@" __stringify(__LINE__)": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, desc, desc_bytes(desc), 1);
-+#endif
-+}
-+EXPORT_SYMBOL(cnstr_shdsc_aead_encap);
-+
-+/**
-+ * cnstr_shdsc_aead_decap - IPSec ESP decapsulation shared descriptor
-+ * (non-protocol).
-+ * @desc: pointer to buffer used for descriptor construction
-+ * @cdata: pointer to block cipher transform definitions
-+ * Valid algorithm values - one of OP_ALG_ALGSEL_{AES, DES, 3DES} ANDed
-+ * with OP_ALG_AAI_CBC or OP_ALG_AAI_CTR_MOD128.
-+ * @adata: pointer to authentication transform definitions.
-+ * A split key is required for SEC Era < 6; the size of the split key
-+ * is specified in this case. Valid algorithm values - one of
-+ * OP_ALG_ALGSEL_{MD5, SHA1, SHA224, SHA256, SHA384, SHA512} ANDed
-+ * with OP_ALG_AAI_HMAC_PRECOMP.
-+ * @ivsize: initialization vector size
-+ * @icvsize: integrity check value (ICV) size (truncated or full)
-+ * @is_rfc3686: true when ctr(aes) is wrapped by rfc3686 template
-+ * @nonce: pointer to rfc3686 nonce
-+ * @ctx1_iv_off: IV offset in CONTEXT1 register
-+ * @is_qi: true when called from caam/qi
-+ * @era: SEC Era
-+ */
-+void cnstr_shdsc_aead_decap(u32 * const desc, struct alginfo *cdata,
-+ struct alginfo *adata, unsigned int ivsize,
-+ unsigned int icvsize, const bool geniv,
-+ const bool is_rfc3686, u32 *nonce,
-+ const u32 ctx1_iv_off, const bool is_qi, int era)
-+{
-+ /* Note: Context registers are saved. */
-+ init_sh_desc_key_aead(desc, cdata, adata, is_rfc3686, nonce, era);
-+
-+ /* Class 2 operation */
-+ append_operation(desc, adata->algtype | OP_ALG_AS_INITFINAL |
-+ OP_ALG_DECRYPT | OP_ALG_ICV_ON);
-+
-+ if (is_qi) {
-+ u32 *wait_load_cmd;
-+
-+ /* REG3 = assoclen */
-+ append_seq_load(desc, 4, LDST_CLASS_DECO |
-+ LDST_SRCDST_WORD_DECO_MATH3 |
-+ (4 << LDST_OFFSET_SHIFT));
-+
-+ wait_load_cmd = append_jump(desc, JUMP_JSL | JUMP_TEST_ALL |
-+ JUMP_COND_CALM | JUMP_COND_NCP |
-+ JUMP_COND_NOP | JUMP_COND_NIP |
-+ JUMP_COND_NIFP);
-+ set_jump_tgt_here(desc, wait_load_cmd);
-+
-+ if (!geniv)
-+ append_seq_load(desc, ivsize, LDST_CLASS_1_CCB |
-+ LDST_SRCDST_BYTE_CONTEXT |
-+ (ctx1_iv_off << LDST_OFFSET_SHIFT));
-+ }
-+
-+ /* Read and write assoclen bytes */
-+ if (is_qi || era < 3) {
-+ append_math_add(desc, VARSEQINLEN, ZERO, REG3, CAAM_CMD_SZ);
-+ if (geniv)
-+ append_math_add_imm_u32(desc, VARSEQOUTLEN, REG3, IMM,
-+ ivsize);
-+ else
-+ append_math_add(desc, VARSEQOUTLEN, ZERO, REG3,
-+ CAAM_CMD_SZ);
-+ } else {
-+ append_math_add(desc, VARSEQINLEN, ZERO, DPOVRD, CAAM_CMD_SZ);
-+ if (geniv)
-+ append_math_add_imm_u32(desc, VARSEQOUTLEN, DPOVRD, IMM,
-+ ivsize);
-+ else
-+ append_math_add(desc, VARSEQOUTLEN, ZERO, DPOVRD,
-+ CAAM_CMD_SZ);
-+ }
-+
-+ /* Skip assoc data */
-+ append_seq_fifo_store(desc, 0, FIFOST_TYPE_SKIP | FIFOLDST_VLF);
-+
-+ /* read assoc before reading payload */
-+ append_seq_fifo_load(desc, 0, FIFOLD_CLASS_CLASS2 | FIFOLD_TYPE_MSG |
-+ KEY_VLF);
-+
-+ if (geniv) {
-+ append_seq_load(desc, ivsize, LDST_CLASS_1_CCB |
-+ LDST_SRCDST_BYTE_CONTEXT |
-+ (ctx1_iv_off << LDST_OFFSET_SHIFT));
-+ append_move(desc, MOVE_SRC_CLASS1CTX | MOVE_DEST_CLASS2INFIFO |
-+ (ctx1_iv_off << MOVE_OFFSET_SHIFT) | ivsize);
-+ }
-+
-+ /* Load Counter into CONTEXT1 reg */
-+ if (is_rfc3686)
-+ append_load_imm_be32(desc, 1, LDST_IMM | LDST_CLASS_1_CCB |
-+ LDST_SRCDST_BYTE_CONTEXT |
-+ ((ctx1_iv_off + CTR_RFC3686_IV_SIZE) <<
-+ LDST_OFFSET_SHIFT));
-+
-+ /* Choose operation */
-+ if (ctx1_iv_off)
-+ append_operation(desc, cdata->algtype | OP_ALG_AS_INITFINAL |
-+ OP_ALG_DECRYPT);
-+ else
-+ append_dec_op1(desc, cdata->algtype);
-+
-+ /* Read and write cryptlen bytes */
-+ append_math_add(desc, VARSEQINLEN, SEQOUTLEN, REG0, CAAM_CMD_SZ);
-+ append_math_add(desc, VARSEQOUTLEN, SEQOUTLEN, REG0, CAAM_CMD_SZ);
-+ aead_append_src_dst(desc, FIFOLD_TYPE_MSG);
-+
-+ /* Load ICV */
-+ append_seq_fifo_load(desc, icvsize, FIFOLD_CLASS_CLASS2 |
-+ FIFOLD_TYPE_LAST2 | FIFOLD_TYPE_ICV);
-+
-+#ifdef DEBUG
-+ print_hex_dump(KERN_ERR, "aead dec shdesc@" __stringify(__LINE__)": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, desc, desc_bytes(desc), 1);
-+#endif
-+}
-+EXPORT_SYMBOL(cnstr_shdsc_aead_decap);
-+
-+/**
-+ * cnstr_shdsc_aead_givencap - IPSec ESP encapsulation shared descriptor
-+ * (non-protocol) with HW-generated initialization
-+ * vector.
-+ * @desc: pointer to buffer used for descriptor construction
-+ * @cdata: pointer to block cipher transform definitions
-+ * Valid algorithm values - one of OP_ALG_ALGSEL_{AES, DES, 3DES} ANDed
-+ * with OP_ALG_AAI_CBC or OP_ALG_AAI_CTR_MOD128.
-+ * @adata: pointer to authentication transform definitions.
-+ * A split key is required for SEC Era < 6; the size of the split key
-+ * is specified in this case. Valid algorithm values - one of
-+ * OP_ALG_ALGSEL_{MD5, SHA1, SHA224, SHA256, SHA384, SHA512} ANDed
-+ * with OP_ALG_AAI_HMAC_PRECOMP.
-+ * @ivsize: initialization vector size
-+ * @icvsize: integrity check value (ICV) size (truncated or full)
-+ * @is_rfc3686: true when ctr(aes) is wrapped by rfc3686 template
-+ * @nonce: pointer to rfc3686 nonce
-+ * @ctx1_iv_off: IV offset in CONTEXT1 register
-+ * @is_qi: true when called from caam/qi
-+ * @era: SEC Era
-+ */
-+void cnstr_shdsc_aead_givencap(u32 * const desc, struct alginfo *cdata,
-+ struct alginfo *adata, unsigned int ivsize,
-+ unsigned int icvsize, const bool is_rfc3686,
-+ u32 *nonce, const u32 ctx1_iv_off,
-+ const bool is_qi, int era)
-+{
-+ u32 geniv, moveiv;
-+
-+ /* Note: Context registers are saved. */
-+ init_sh_desc_key_aead(desc, cdata, adata, is_rfc3686, nonce, era);
-+
-+ if (is_qi) {
-+ u32 *wait_load_cmd;
-+
-+ /* REG3 = assoclen */
-+ append_seq_load(desc, 4, LDST_CLASS_DECO |
-+ LDST_SRCDST_WORD_DECO_MATH3 |
-+ (4 << LDST_OFFSET_SHIFT));
-+
-+ wait_load_cmd = append_jump(desc, JUMP_JSL | JUMP_TEST_ALL |
-+ JUMP_COND_CALM | JUMP_COND_NCP |
-+ JUMP_COND_NOP | JUMP_COND_NIP |
-+ JUMP_COND_NIFP);
-+ set_jump_tgt_here(desc, wait_load_cmd);
-+ }
-+
-+ if (is_rfc3686) {
-+ if (is_qi)
-+ append_seq_load(desc, ivsize, LDST_CLASS_1_CCB |
-+ LDST_SRCDST_BYTE_CONTEXT |
-+ (ctx1_iv_off << LDST_OFFSET_SHIFT));
-+
-+ goto copy_iv;
-+ }
-+
-+ /* Generate IV */
-+ geniv = NFIFOENTRY_STYPE_PAD | NFIFOENTRY_DEST_DECO |
-+ NFIFOENTRY_DTYPE_MSG | NFIFOENTRY_LC1 |
-+ NFIFOENTRY_PTYPE_RND | (ivsize << NFIFOENTRY_DLEN_SHIFT);
-+ append_load_imm_u32(desc, geniv, LDST_CLASS_IND_CCB |
-+ LDST_SRCDST_WORD_INFO_FIFO | LDST_IMM);
-+ append_cmd(desc, CMD_LOAD | DISABLE_AUTO_INFO_FIFO);
-+ append_move(desc, MOVE_WAITCOMP |
-+ MOVE_SRC_INFIFO | MOVE_DEST_CLASS1CTX |
-+ (ctx1_iv_off << MOVE_OFFSET_SHIFT) |
-+ (ivsize << MOVE_LEN_SHIFT));
-+ append_cmd(desc, CMD_LOAD | ENABLE_AUTO_INFO_FIFO);
-+
-+copy_iv:
-+ /* Copy IV to class 1 context */
-+ append_move(desc, MOVE_SRC_CLASS1CTX | MOVE_DEST_OUTFIFO |
-+ (ctx1_iv_off << MOVE_OFFSET_SHIFT) |
-+ (ivsize << MOVE_LEN_SHIFT));
-+
-+ /* Return to encryption */
-+ append_operation(desc, adata->algtype | OP_ALG_AS_INITFINAL |
-+ OP_ALG_ENCRYPT);
-+
-+ /* Read and write assoclen bytes */
-+ if (is_qi || era < 3) {
-+ append_math_add(desc, VARSEQINLEN, ZERO, REG3, CAAM_CMD_SZ);
-+ append_math_add(desc, VARSEQOUTLEN, ZERO, REG3, CAAM_CMD_SZ);
-+ } else {
-+ append_math_add(desc, VARSEQINLEN, ZERO, DPOVRD, CAAM_CMD_SZ);
-+ append_math_add(desc, VARSEQOUTLEN, ZERO, DPOVRD, CAAM_CMD_SZ);
-+ }
-+
-+ /* Skip assoc data */
-+ append_seq_fifo_store(desc, 0, FIFOST_TYPE_SKIP | FIFOLDST_VLF);
-+
-+ /* read assoc before reading payload */
-+ append_seq_fifo_load(desc, 0, FIFOLD_CLASS_CLASS2 | FIFOLD_TYPE_MSG |
-+ KEY_VLF);
-+
-+ /* Copy iv from outfifo to class 2 fifo */
-+ moveiv = NFIFOENTRY_STYPE_OFIFO | NFIFOENTRY_DEST_CLASS2 |
-+ NFIFOENTRY_DTYPE_MSG | (ivsize << NFIFOENTRY_DLEN_SHIFT);
-+ append_load_imm_u32(desc, moveiv, LDST_CLASS_IND_CCB |
-+ LDST_SRCDST_WORD_INFO_FIFO | LDST_IMM);
-+ append_load_imm_u32(desc, ivsize, LDST_CLASS_2_CCB |
-+ LDST_SRCDST_WORD_DATASZ_REG | LDST_IMM);
-+
-+ /* Load Counter into CONTEXT1 reg */
-+ if (is_rfc3686)
-+ append_load_imm_be32(desc, 1, LDST_IMM | LDST_CLASS_1_CCB |
-+ LDST_SRCDST_BYTE_CONTEXT |
-+ ((ctx1_iv_off + CTR_RFC3686_IV_SIZE) <<
-+ LDST_OFFSET_SHIFT));
-+
-+ /* Class 1 operation */
-+ append_operation(desc, cdata->algtype | OP_ALG_AS_INITFINAL |
-+ OP_ALG_ENCRYPT);
-+
-+ /* Will write ivsize + cryptlen */
-+ append_math_add(desc, VARSEQOUTLEN, SEQINLEN, REG0, CAAM_CMD_SZ);
-+
-+ /* Not need to reload iv */
-+ append_seq_fifo_load(desc, ivsize,
-+ FIFOLD_CLASS_SKIP);
-+
-+ /* Will read cryptlen */
-+ append_math_add(desc, VARSEQINLEN, SEQINLEN, REG0, CAAM_CMD_SZ);
-+ append_seq_fifo_load(desc, 0, FIFOLD_CLASS_BOTH | KEY_VLF |
-+ FIFOLD_TYPE_MSG1OUT2 | FIFOLD_TYPE_LASTBOTH);
-+ append_seq_fifo_store(desc, 0, FIFOST_TYPE_MESSAGE_DATA | KEY_VLF);
-+
-+ /* Write ICV */
-+ append_seq_store(desc, icvsize, LDST_CLASS_2_CCB |
-+ LDST_SRCDST_BYTE_CONTEXT);
-+
-+#ifdef DEBUG
-+ print_hex_dump(KERN_ERR,
-+ "aead givenc shdesc@" __stringify(__LINE__)": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, desc, desc_bytes(desc), 1);
-+#endif
-+}
-+EXPORT_SYMBOL(cnstr_shdsc_aead_givencap);
-+
-+/**
-+ * cnstr_shdsc_tls_encap - tls encapsulation shared descriptor
-+ * @desc: pointer to buffer used for descriptor construction
-+ * @cdata: pointer to block cipher transform definitions
-+ * Valid algorithm values - one of OP_ALG_ALGSEL_AES ANDed
-+ * with OP_ALG_AAI_CBC
-+ * @adata: pointer to authentication transform definitions.
-+ * A split key is required for SEC Era < 6; the size of the split key
-+ * is specified in this case. Valid algorithm values OP_ALG_ALGSEL_SHA1
-+ * ANDed with OP_ALG_AAI_HMAC_PRECOMP.
-+ * @assoclen: associated data length
-+ * @ivsize: initialization vector size
-+ * @authsize: authentication data size
-+ * @blocksize: block cipher size
-+ * @era: SEC Era
-+ */
-+void cnstr_shdsc_tls_encap(u32 * const desc, struct alginfo *cdata,
-+ struct alginfo *adata, unsigned int assoclen,
-+ unsigned int ivsize, unsigned int authsize,
-+ unsigned int blocksize, int era)
-+{
-+ u32 *key_jump_cmd, *zero_payload_jump_cmd;
-+ u32 genpad, idx_ld_datasz, idx_ld_pad, stidx;
-+
-+ /*
-+ * Compute the index (in bytes) for the LOAD with destination of
-+ * Class 1 Data Size Register and for the LOAD that generates padding
-+ */
-+ if (adata->key_inline) {
-+ idx_ld_datasz = DESC_TLS10_ENC_LEN + adata->keylen_pad +
-+ cdata->keylen - 4 * CAAM_CMD_SZ;
-+ idx_ld_pad = DESC_TLS10_ENC_LEN + adata->keylen_pad +
-+ cdata->keylen - 2 * CAAM_CMD_SZ;
-+ } else {
-+ idx_ld_datasz = DESC_TLS10_ENC_LEN + 2 * CAAM_PTR_SZ -
-+ 4 * CAAM_CMD_SZ;
-+ idx_ld_pad = DESC_TLS10_ENC_LEN + 2 * CAAM_PTR_SZ -
-+ 2 * CAAM_CMD_SZ;
-+ }
-+
-+ stidx = 1 << HDR_START_IDX_SHIFT;
-+ init_sh_desc(desc, HDR_SHARE_SERIAL | stidx);
-+
-+ /* skip key loading if they are loaded due to sharing */
-+ key_jump_cmd = append_jump(desc, JUMP_JSL | JUMP_TEST_ALL |
-+ JUMP_COND_SHRD);
-+
-+ if (era < 6) {
-+ if (adata->key_inline)
-+ append_key_as_imm(desc, adata->key_virt,
-+ adata->keylen_pad, adata->keylen,
-+ CLASS_2 | KEY_DEST_MDHA_SPLIT |
-+ KEY_ENC);
-+ else
-+ append_key(desc, adata->key_dma, adata->keylen,
-+ CLASS_2 | KEY_DEST_MDHA_SPLIT | KEY_ENC);
-+ } else {
-+ append_proto_dkp(desc, adata);
-+ }
-+
-+ if (cdata->key_inline)
-+ append_key_as_imm(desc, cdata->key_virt, cdata->keylen,
-+ cdata->keylen, CLASS_1 | KEY_DEST_CLASS_REG);
-+ else
-+ append_key(desc, cdata->key_dma, cdata->keylen, CLASS_1 |
-+ KEY_DEST_CLASS_REG);
-+
-+ set_jump_tgt_here(desc, key_jump_cmd);
-+
-+ /* class 2 operation */
-+ append_operation(desc, adata->algtype | OP_ALG_AS_INITFINAL |
-+ OP_ALG_ENCRYPT);
-+ /* class 1 operation */
-+ append_operation(desc, cdata->algtype | OP_ALG_AS_INITFINAL |
-+ OP_ALG_ENCRYPT);
-+
-+ /* payloadlen = input data length - (assoclen + ivlen) */
-+ append_math_sub_imm_u32(desc, REG0, SEQINLEN, IMM, assoclen + ivsize);
-+
-+ /* math1 = payloadlen + icvlen */
-+ append_math_add_imm_u32(desc, REG1, REG0, IMM, authsize);
-+
-+ /* padlen = block_size - math1 % block_size */
-+ append_math_and_imm_u32(desc, REG3, REG1, IMM, blocksize - 1);
-+ append_math_sub_imm_u32(desc, REG2, IMM, REG3, blocksize);
-+
-+ /* cryptlen = payloadlen + icvlen + padlen */
-+ append_math_add(desc, VARSEQOUTLEN, REG1, REG2, 4);
-+
-+ /*
-+ * update immediate data with the padding length value
-+ * for the LOAD in the class 1 data size register.
-+ */
-+ append_move(desc, MOVE_SRC_DESCBUF | MOVE_DEST_MATH2 |
-+ (idx_ld_datasz << MOVE_OFFSET_SHIFT) | 7);
-+ append_move(desc, MOVE_WAITCOMP | MOVE_SRC_MATH2 | MOVE_DEST_DESCBUF |
-+ (idx_ld_datasz << MOVE_OFFSET_SHIFT) | 8);
-+
-+ /* overwrite PL field for the padding iNFO FIFO entry */
-+ append_move(desc, MOVE_SRC_DESCBUF | MOVE_DEST_MATH2 |
-+ (idx_ld_pad << MOVE_OFFSET_SHIFT) | 7);
-+ append_move(desc, MOVE_WAITCOMP | MOVE_SRC_MATH2 | MOVE_DEST_DESCBUF |
-+ (idx_ld_pad << MOVE_OFFSET_SHIFT) | 8);
-+
-+ /* store encrypted payload, icv and padding */
-+ append_seq_fifo_store(desc, 0, FIFOST_TYPE_MESSAGE_DATA | LDST_VLF);
-+
-+ /* if payload length is zero, jump to zero-payload commands */
-+ append_math_add(desc, VARSEQINLEN, ZERO, REG0, 4);
-+ zero_payload_jump_cmd = append_jump(desc, JUMP_TEST_ALL |
-+ JUMP_COND_MATH_Z);
-+
-+ /* load iv in context1 */
-+ append_cmd(desc, CMD_SEQ_LOAD | LDST_SRCDST_WORD_CLASS_CTX |
-+ LDST_CLASS_1_CCB | ivsize);
-+
-+ /* read assoc for authentication */
-+ append_seq_fifo_load(desc, assoclen, FIFOLD_CLASS_CLASS2 |
-+ FIFOLD_TYPE_MSG);
-+ /* insnoop payload */
-+ append_seq_fifo_load(desc, 0, FIFOLD_CLASS_BOTH | FIFOLD_TYPE_MSG |
-+ FIFOLD_TYPE_LAST2 | FIFOLDST_VLF);
-+
-+ /* jump the zero-payload commands */
-+ append_jump(desc, JUMP_TEST_ALL | 3);
-+
-+ /* zero-payload commands */
-+ set_jump_tgt_here(desc, zero_payload_jump_cmd);
-+
-+ /* load iv in context1 */
-+ append_cmd(desc, CMD_SEQ_LOAD | LDST_SRCDST_WORD_CLASS_CTX |
-+ LDST_CLASS_1_CCB | ivsize);
-+
-+ /* assoc data is the only data for authentication */
-+ append_seq_fifo_load(desc, assoclen, FIFOLD_CLASS_CLASS2 |
-+ FIFOLD_TYPE_MSG | FIFOLD_TYPE_LAST2);
-+
-+ /* send icv to encryption */
-+ append_move(desc, MOVE_SRC_CLASS2CTX | MOVE_DEST_CLASS1INFIFO |
-+ authsize);
-+
-+ /* update class 1 data size register with padding length */
-+ append_load_imm_u32(desc, 0, LDST_CLASS_1_CCB |
-+ LDST_SRCDST_WORD_DATASZ_REG | LDST_IMM);
-+
-+ /* generate padding and send it to encryption */
-+ genpad = NFIFOENTRY_DEST_CLASS1 | NFIFOENTRY_LC1 | NFIFOENTRY_FC1 |
-+ NFIFOENTRY_STYPE_PAD | NFIFOENTRY_DTYPE_MSG | NFIFOENTRY_PTYPE_N;
-+ append_load_imm_u32(desc, genpad, LDST_CLASS_IND_CCB |
-+ LDST_SRCDST_WORD_INFO_FIFO | LDST_IMM);
-+
-+#ifdef DEBUG
-+ print_hex_dump(KERN_ERR, "tls enc shdesc@" __stringify(__LINE__) ": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, desc,
-+ desc_bytes(desc), 1);
-+#endif
-+}
-+EXPORT_SYMBOL(cnstr_shdsc_tls_encap);
-+
-+/**
-+ * cnstr_shdsc_tls_decap - tls decapsulation shared descriptor
-+ * @desc: pointer to buffer used for descriptor construction
-+ * @cdata: pointer to block cipher transform definitions
-+ * Valid algorithm values - one of OP_ALG_ALGSEL_AES ANDed
-+ * with OP_ALG_AAI_CBC
-+ * @adata: pointer to authentication transform definitions.
-+ * A split key is required for SEC Era < 6; the size of the split key
-+ * is specified in this case. Valid algorithm values OP_ALG_ALGSEL_SHA1
-+ * ANDed with OP_ALG_AAI_HMAC_PRECOMP.
-+ * @assoclen: associated data length
-+ * @ivsize: initialization vector size
-+ * @authsize: authentication data size
-+ * @blocksize: block cipher size
-+ * @era: SEC Era
-+ */
-+void cnstr_shdsc_tls_decap(u32 * const desc, struct alginfo *cdata,
-+ struct alginfo *adata, unsigned int assoclen,
-+ unsigned int ivsize, unsigned int authsize,
-+ unsigned int blocksize, int era)
-+{
-+ u32 stidx, jumpback;
-+ u32 *key_jump_cmd, *zero_payload_jump_cmd, *skip_zero_jump_cmd;
-+ /*
-+ * Pointer Size bool determines the size of address pointers.
-+ * false - Pointers fit in one 32-bit word.
-+ * true - Pointers fit in two 32-bit words.
-+ */
-+ static const bool ps = (CAAM_PTR_SZ != CAAM_CMD_SZ);
-+
-+ stidx = 1 << HDR_START_IDX_SHIFT;
-+ init_sh_desc(desc, HDR_SHARE_SERIAL | stidx);
-+
-+ /* skip key loading if they are loaded due to sharing */
-+ key_jump_cmd = append_jump(desc, JUMP_JSL | JUMP_TEST_ALL |
-+ JUMP_COND_SHRD);
-+
-+ if (era < 6)
-+ append_key(desc, adata->key_dma, adata->keylen, CLASS_2 |
-+ KEY_DEST_MDHA_SPLIT | KEY_ENC);
-+ else
-+ append_proto_dkp(desc, adata);
-+
-+ append_key(desc, cdata->key_dma, cdata->keylen, CLASS_1 |
-+ KEY_DEST_CLASS_REG);
-+
-+ set_jump_tgt_here(desc, key_jump_cmd);
-+
-+ /* class 2 operation */
-+ append_operation(desc, adata->algtype | OP_ALG_AS_INITFINAL |
-+ OP_ALG_DECRYPT | OP_ALG_ICV_ON);
-+ /* class 1 operation */
-+ append_operation(desc, cdata->algtype | OP_ALG_AS_INITFINAL |
-+ OP_ALG_DECRYPT);
-+
-+ /* VSIL = input data length - 2 * block_size */
-+ append_math_sub_imm_u32(desc, VARSEQINLEN, SEQINLEN, IMM, 2 *
-+ blocksize);
-+
-+ /*
-+ * payloadlen + icvlen + padlen = input data length - (assoclen +
-+ * ivsize)
-+ */
-+ append_math_sub_imm_u32(desc, REG3, SEQINLEN, IMM, assoclen + ivsize);
-+
-+ /* skip data to the last but one cipher block */
-+ append_seq_fifo_load(desc, 0, FIFOLD_CLASS_SKIP | LDST_VLF);
-+
-+ /* load iv for the last cipher block */
-+ append_cmd(desc, CMD_SEQ_LOAD | LDST_SRCDST_WORD_CLASS_CTX |
-+ LDST_CLASS_1_CCB | ivsize);
-+
-+ /* read last cipher block */
-+ append_seq_fifo_load(desc, 0, FIFOLD_CLASS_CLASS1 | FIFOLD_TYPE_MSG |
-+ FIFOLD_TYPE_LAST1 | blocksize);
-+
-+ /* move decrypted block into math0 and math1 */
-+ append_move(desc, MOVE_WAITCOMP | MOVE_SRC_OUTFIFO | MOVE_DEST_MATH0 |
-+ blocksize);
-+
-+ /* reset AES CHA */
-+ append_load_imm_u32(desc, CCTRL_RESET_CHA_AESA, LDST_CLASS_IND_CCB |
-+ LDST_SRCDST_WORD_CHACTRL | LDST_IMM);
-+
-+ /* rewind input sequence */
-+ append_seq_in_ptr_intlen(desc, 0, 65535, SQIN_RTO);
-+
-+ /* key1 is in decryption form */
-+ append_operation(desc, cdata->algtype | OP_ALG_AAI_DK |
-+ OP_ALG_AS_INITFINAL | OP_ALG_DECRYPT);
-+
-+ /* load iv in context1 */
-+ append_cmd(desc, CMD_SEQ_LOAD | LDST_CLASS_1_CCB |
-+ LDST_SRCDST_WORD_CLASS_CTX | ivsize);
-+
-+ /* read sequence number */
-+ append_seq_fifo_load(desc, 8, FIFOLD_CLASS_CLASS2 | FIFOLD_TYPE_MSG);
-+ /* load Type, Version and Len fields in math0 */
-+ append_cmd(desc, CMD_SEQ_LOAD | LDST_CLASS_DECO |
-+ LDST_SRCDST_WORD_DECO_MATH0 | (3 << LDST_OFFSET_SHIFT) | 5);
-+
-+ /* compute (padlen - 1) */
-+ append_math_and_imm_u64(desc, REG1, REG1, IMM, 255);
-+
-+ /* math2 = icvlen + (padlen - 1) + 1 */
-+ append_math_add_imm_u32(desc, REG2, REG1, IMM, authsize + 1);
-+
-+ append_jump(desc, JUMP_TEST_ALL | JUMP_COND_CALM | 1);
-+
-+ /* VSOL = payloadlen + icvlen + padlen */
-+ append_math_add(desc, VARSEQOUTLEN, ZERO, REG3, 4);
-+
-+ if (caam_little_end)
-+ append_moveb(desc, MOVE_WAITCOMP |
-+ MOVE_SRC_MATH0 | MOVE_DEST_MATH0 | 8);
-+
-+ /* update Len field */
-+ append_math_sub(desc, REG0, REG0, REG2, 8);
-+
-+ /* store decrypted payload, icv and padding */
-+ append_seq_fifo_store(desc, 0, FIFOST_TYPE_MESSAGE_DATA | LDST_VLF);
-+
-+ /* VSIL = (payloadlen + icvlen + padlen) - (icvlen + padlen)*/
-+ append_math_sub(desc, VARSEQINLEN, REG3, REG2, 4);
-+
-+ zero_payload_jump_cmd = append_jump(desc, JUMP_TEST_ALL |
-+ JUMP_COND_MATH_Z);
-+
-+ /* send Type, Version and Len(pre ICV) fields to authentication */
-+ append_move(desc, MOVE_WAITCOMP |
-+ MOVE_SRC_MATH0 | MOVE_DEST_CLASS2INFIFO |
-+ (3 << MOVE_OFFSET_SHIFT) | 5);
-+
-+ /* outsnooping payload */
-+ append_seq_fifo_load(desc, 0, FIFOLD_CLASS_BOTH |
-+ FIFOLD_TYPE_MSG1OUT2 | FIFOLD_TYPE_LAST2 |
-+ FIFOLDST_VLF);
-+ skip_zero_jump_cmd = append_jump(desc, JUMP_TEST_ALL | 2);
-+
-+ set_jump_tgt_here(desc, zero_payload_jump_cmd);
-+ /* send Type, Version and Len(pre ICV) fields to authentication */
-+ append_move(desc, MOVE_WAITCOMP | MOVE_AUX_LS |
-+ MOVE_SRC_MATH0 | MOVE_DEST_CLASS2INFIFO |
-+ (3 << MOVE_OFFSET_SHIFT) | 5);
-+
-+ set_jump_tgt_here(desc, skip_zero_jump_cmd);
-+ append_math_add(desc, VARSEQINLEN, ZERO, REG2, 4);
-+
-+ /* load icvlen and padlen */
-+ append_seq_fifo_load(desc, 0, FIFOLD_CLASS_CLASS1 | FIFOLD_TYPE_MSG |
-+ FIFOLD_TYPE_LAST1 | FIFOLDST_VLF);
-+
-+ /* VSIL = (payloadlen + icvlen + padlen) - icvlen + padlen */
-+ append_math_sub(desc, VARSEQINLEN, REG3, REG2, 4);
-+
-+ /*
-+ * Start a new input sequence using the SEQ OUT PTR command options,
-+ * pointer and length used when the current output sequence was defined.
-+ */
-+ if (ps) {
-+ /*
-+ * Move the lower 32 bits of Shared Descriptor address, the
-+ * SEQ OUT PTR command, Output Pointer (2 words) and
-+ * Output Length into math registers.
-+ */
-+ if (caam_little_end)
-+ append_move(desc, MOVE_WAITCOMP | MOVE_SRC_DESCBUF |
-+ MOVE_DEST_MATH0 |
-+ (55 * 4 << MOVE_OFFSET_SHIFT) | 20);
-+ else
-+ append_move(desc, MOVE_WAITCOMP | MOVE_SRC_DESCBUF |
-+ MOVE_DEST_MATH0 |
-+ (54 * 4 << MOVE_OFFSET_SHIFT) | 20);
-+
-+ /* Transform SEQ OUT PTR command in SEQ IN PTR command */
-+ append_math_and_imm_u32(desc, REG0, REG0, IMM,
-+ ~(CMD_SEQ_IN_PTR ^ CMD_SEQ_OUT_PTR));
-+ /* Append a JUMP command after the copied fields */
-+ jumpback = CMD_JUMP | (char)-9;
-+ append_load_imm_u32(desc, jumpback, LDST_CLASS_DECO | LDST_IMM |
-+ LDST_SRCDST_WORD_DECO_MATH2 |
-+ (4 << LDST_OFFSET_SHIFT));
-+ append_jump(desc, JUMP_TEST_ALL | JUMP_COND_CALM | 1);
-+ /* Move the updated fields back to the Job Descriptor */
-+ if (caam_little_end)
-+ append_move(desc, MOVE_WAITCOMP | MOVE_SRC_MATH0 |
-+ MOVE_DEST_DESCBUF |
-+ (55 * 4 << MOVE_OFFSET_SHIFT) | 24);
-+ else
-+ append_move(desc, MOVE_WAITCOMP | MOVE_SRC_MATH0 |
-+ MOVE_DEST_DESCBUF |
-+ (54 * 4 << MOVE_OFFSET_SHIFT) | 24);
-+
-+ /*
-+ * Read the new SEQ IN PTR command, Input Pointer, Input Length
-+ * and then jump back to the next command from the
-+ * Shared Descriptor.
-+ */
-+ append_jump(desc, JUMP_TEST_ALL | JUMP_COND_CALM | 6);
-+ } else {
-+ /*
-+ * Move the SEQ OUT PTR command, Output Pointer (1 word) and
-+ * Output Length into math registers.
-+ */
-+ if (caam_little_end)
-+ append_move(desc, MOVE_WAITCOMP | MOVE_SRC_DESCBUF |
-+ MOVE_DEST_MATH0 |
-+ (54 * 4 << MOVE_OFFSET_SHIFT) | 12);
-+ else
-+ append_move(desc, MOVE_WAITCOMP | MOVE_SRC_DESCBUF |
-+ MOVE_DEST_MATH0 |
-+ (53 * 4 << MOVE_OFFSET_SHIFT) | 12);
-+
-+ /* Transform SEQ OUT PTR command in SEQ IN PTR command */
-+ append_math_and_imm_u64(desc, REG0, REG0, IMM,
-+ ~(((u64)(CMD_SEQ_IN_PTR ^
-+ CMD_SEQ_OUT_PTR)) << 32));
-+ /* Append a JUMP command after the copied fields */
-+ jumpback = CMD_JUMP | (char)-7;
-+ append_load_imm_u32(desc, jumpback, LDST_CLASS_DECO | LDST_IMM |
-+ LDST_SRCDST_WORD_DECO_MATH1 |
-+ (4 << LDST_OFFSET_SHIFT));
-+ append_jump(desc, JUMP_TEST_ALL | JUMP_COND_CALM | 1);
-+ /* Move the updated fields back to the Job Descriptor */
-+ if (caam_little_end)
-+ append_move(desc, MOVE_WAITCOMP | MOVE_SRC_MATH0 |
-+ MOVE_DEST_DESCBUF |
-+ (54 * 4 << MOVE_OFFSET_SHIFT) | 16);
-+ else
-+ append_move(desc, MOVE_WAITCOMP | MOVE_SRC_MATH0 |
-+ MOVE_DEST_DESCBUF |
-+ (53 * 4 << MOVE_OFFSET_SHIFT) | 16);
-+
-+ /*
-+ * Read the new SEQ IN PTR command, Input Pointer, Input Length
-+ * and then jump back to the next command from the
-+ * Shared Descriptor.
-+ */
-+ append_jump(desc, JUMP_TEST_ALL | JUMP_COND_CALM | 5);
-+ }
-+
-+ /* skip payload */
-+ append_seq_fifo_load(desc, 0, FIFOLD_CLASS_SKIP | FIFOLDST_VLF);
-+ /* check icv */
-+ append_seq_fifo_load(desc, 0, FIFOLD_CLASS_CLASS2 | FIFOLD_TYPE_ICV |
-+ FIFOLD_TYPE_LAST2 | authsize);
-+
-+#ifdef DEBUG
-+ print_hex_dump(KERN_ERR, "tls dec shdesc@" __stringify(__LINE__) ": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, desc,
-+ desc_bytes(desc), 1);
-+#endif
-+}
-+EXPORT_SYMBOL(cnstr_shdsc_tls_decap);
-+
-+/**
-+ * cnstr_shdsc_gcm_encap - gcm encapsulation shared descriptor
-+ * @desc: pointer to buffer used for descriptor construction
-+ * @cdata: pointer to block cipher transform definitions
-+ * Valid algorithm values - OP_ALG_ALGSEL_AES ANDed with OP_ALG_AAI_GCM.
-+ * @ivsize: initialization vector size
-+ * @icvsize: integrity check value (ICV) size (truncated or full)
-+ * @is_qi: true when called from caam/qi
-+ */
-+void cnstr_shdsc_gcm_encap(u32 * const desc, struct alginfo *cdata,
-+ unsigned int ivsize, unsigned int icvsize,
-+ const bool is_qi)
-+{
-+ u32 *key_jump_cmd, *zero_payload_jump_cmd, *zero_assoc_jump_cmd1,
-+ *zero_assoc_jump_cmd2;
-+
-+ init_sh_desc(desc, HDR_SHARE_SERIAL);
-+
-+ /* skip key loading if they are loaded due to sharing */
-+ key_jump_cmd = append_jump(desc, JUMP_JSL | JUMP_TEST_ALL |
-+ JUMP_COND_SHRD);
-+ if (cdata->key_inline)
-+ append_key_as_imm(desc, cdata->key_virt, cdata->keylen,
-+ cdata->keylen, CLASS_1 | KEY_DEST_CLASS_REG);
-+ else
-+ append_key(desc, cdata->key_dma, cdata->keylen, CLASS_1 |
-+ KEY_DEST_CLASS_REG);
-+ set_jump_tgt_here(desc, key_jump_cmd);
-+
-+ /* class 1 operation */
-+ append_operation(desc, cdata->algtype | OP_ALG_AS_INITFINAL |
-+ OP_ALG_ENCRYPT);
-+
-+ if (is_qi) {
-+ u32 *wait_load_cmd;
-+
-+ /* REG3 = assoclen */
-+ append_seq_load(desc, 4, LDST_CLASS_DECO |
-+ LDST_SRCDST_WORD_DECO_MATH3 |
-+ (4 << LDST_OFFSET_SHIFT));
-+
-+ wait_load_cmd = append_jump(desc, JUMP_JSL | JUMP_TEST_ALL |
-+ JUMP_COND_CALM | JUMP_COND_NCP |
-+ JUMP_COND_NOP | JUMP_COND_NIP |
-+ JUMP_COND_NIFP);
-+ set_jump_tgt_here(desc, wait_load_cmd);
-+
-+ append_math_sub_imm_u32(desc, VARSEQOUTLEN, SEQINLEN, IMM,
-+ ivsize);
-+ } else {
-+ append_math_sub(desc, VARSEQOUTLEN, SEQINLEN, REG0,
-+ CAAM_CMD_SZ);
-+ }
-+
-+ /* if assoclen + cryptlen is ZERO, skip to ICV write */
-+ zero_assoc_jump_cmd2 = append_jump(desc, JUMP_TEST_ALL |
-+ JUMP_COND_MATH_Z);
-+
-+ if (is_qi)
-+ append_seq_fifo_load(desc, ivsize, FIFOLD_CLASS_CLASS1 |
-+ FIFOLD_TYPE_IV | FIFOLD_TYPE_FLUSH1);
-+
-+ /* if assoclen is ZERO, skip reading the assoc data */
-+ append_math_add(desc, VARSEQINLEN, ZERO, REG3, CAAM_CMD_SZ);
-+ zero_assoc_jump_cmd1 = append_jump(desc, JUMP_TEST_ALL |
-+ JUMP_COND_MATH_Z);
-+
-+ append_math_add(desc, VARSEQOUTLEN, ZERO, REG3, CAAM_CMD_SZ);
-+
-+ /* skip assoc data */
-+ append_seq_fifo_store(desc, 0, FIFOST_TYPE_SKIP | FIFOLDST_VLF);
-+
-+ /* cryptlen = seqinlen - assoclen */
-+ append_math_sub(desc, VARSEQOUTLEN, SEQINLEN, REG3, CAAM_CMD_SZ);
-+
-+ /* if cryptlen is ZERO jump to zero-payload commands */
-+ zero_payload_jump_cmd = append_jump(desc, JUMP_TEST_ALL |
-+ JUMP_COND_MATH_Z);
-+
-+ /* read assoc data */
-+ append_seq_fifo_load(desc, 0, FIFOLD_CLASS_CLASS1 | FIFOLDST_VLF |
-+ FIFOLD_TYPE_AAD | FIFOLD_TYPE_FLUSH1);
-+ set_jump_tgt_here(desc, zero_assoc_jump_cmd1);
-+
-+ append_math_sub(desc, VARSEQINLEN, SEQINLEN, REG0, CAAM_CMD_SZ);
-+
-+ /* write encrypted data */
-+ append_seq_fifo_store(desc, 0, FIFOST_TYPE_MESSAGE_DATA | FIFOLDST_VLF);
-+
-+ /* read payload data */
-+ append_seq_fifo_load(desc, 0, FIFOLD_CLASS_CLASS1 | FIFOLDST_VLF |
-+ FIFOLD_TYPE_MSG | FIFOLD_TYPE_LAST1);
-+
-+ /* jump to ICV writing */
-+ if (is_qi)
-+ append_jump(desc, JUMP_TEST_ALL | 4);
-+ else
-+ append_jump(desc, JUMP_TEST_ALL | 2);
-+
-+ /* zero-payload commands */
-+ set_jump_tgt_here(desc, zero_payload_jump_cmd);
-+
-+ /* read assoc data */
-+ append_seq_fifo_load(desc, 0, FIFOLD_CLASS_CLASS1 | FIFOLDST_VLF |
-+ FIFOLD_TYPE_AAD | FIFOLD_TYPE_LAST1);
-+ if (is_qi)
-+ /* jump to ICV writing */
-+ append_jump(desc, JUMP_TEST_ALL | 2);
-+
-+ /* There is no input data */
-+ set_jump_tgt_here(desc, zero_assoc_jump_cmd2);
-+
-+ if (is_qi)
-+ append_seq_fifo_load(desc, ivsize, FIFOLD_CLASS_CLASS1 |
-+ FIFOLD_TYPE_IV | FIFOLD_TYPE_FLUSH1 |
-+ FIFOLD_TYPE_LAST1);
-+
-+ /* write ICV */
-+ append_seq_store(desc, icvsize, LDST_CLASS_1_CCB |
-+ LDST_SRCDST_BYTE_CONTEXT);
-+
-+#ifdef DEBUG
-+ print_hex_dump(KERN_ERR, "gcm enc shdesc@" __stringify(__LINE__)": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, desc, desc_bytes(desc), 1);
-+#endif
-+}
-+EXPORT_SYMBOL(cnstr_shdsc_gcm_encap);
-+
-+/**
-+ * cnstr_shdsc_gcm_decap - gcm decapsulation shared descriptor
-+ * @desc: pointer to buffer used for descriptor construction
-+ * @cdata: pointer to block cipher transform definitions
-+ * Valid algorithm values - OP_ALG_ALGSEL_AES ANDed with OP_ALG_AAI_GCM.
-+ * @ivsize: initialization vector size
-+ * @icvsize: integrity check value (ICV) size (truncated or full)
-+ * @is_qi: true when called from caam/qi
-+ */
-+void cnstr_shdsc_gcm_decap(u32 * const desc, struct alginfo *cdata,
-+ unsigned int ivsize, unsigned int icvsize,
-+ const bool is_qi)
-+{
-+ u32 *key_jump_cmd, *zero_payload_jump_cmd, *zero_assoc_jump_cmd1;
-+
-+ init_sh_desc(desc, HDR_SHARE_SERIAL);
-+
-+ /* skip key loading if they are loaded due to sharing */
-+ key_jump_cmd = append_jump(desc, JUMP_JSL |
-+ JUMP_TEST_ALL | JUMP_COND_SHRD);
-+ if (cdata->key_inline)
-+ append_key_as_imm(desc, cdata->key_virt, cdata->keylen,
-+ cdata->keylen, CLASS_1 | KEY_DEST_CLASS_REG);
-+ else
-+ append_key(desc, cdata->key_dma, cdata->keylen, CLASS_1 |
-+ KEY_DEST_CLASS_REG);
-+ set_jump_tgt_here(desc, key_jump_cmd);
-+
-+ /* class 1 operation */
-+ append_operation(desc, cdata->algtype | OP_ALG_AS_INITFINAL |
-+ OP_ALG_DECRYPT | OP_ALG_ICV_ON);
-+
-+ if (is_qi) {
-+ u32 *wait_load_cmd;
-+
-+ /* REG3 = assoclen */
-+ append_seq_load(desc, 4, LDST_CLASS_DECO |
-+ LDST_SRCDST_WORD_DECO_MATH3 |
-+ (4 << LDST_OFFSET_SHIFT));
-+
-+ wait_load_cmd = append_jump(desc, JUMP_JSL | JUMP_TEST_ALL |
-+ JUMP_COND_CALM | JUMP_COND_NCP |
-+ JUMP_COND_NOP | JUMP_COND_NIP |
-+ JUMP_COND_NIFP);
-+ set_jump_tgt_here(desc, wait_load_cmd);
-+
-+ append_seq_fifo_load(desc, ivsize, FIFOLD_CLASS_CLASS1 |
-+ FIFOLD_TYPE_IV | FIFOLD_TYPE_FLUSH1);
-+ }
-+
-+ /* if assoclen is ZERO, skip reading the assoc data */
-+ append_math_add(desc, VARSEQINLEN, ZERO, REG3, CAAM_CMD_SZ);
-+ zero_assoc_jump_cmd1 = append_jump(desc, JUMP_TEST_ALL |
-+ JUMP_COND_MATH_Z);
-+
-+ append_math_add(desc, VARSEQOUTLEN, ZERO, REG3, CAAM_CMD_SZ);
-+
-+ /* skip assoc data */
-+ append_seq_fifo_store(desc, 0, FIFOST_TYPE_SKIP | FIFOLDST_VLF);
-+
-+ /* read assoc data */
-+ append_seq_fifo_load(desc, 0, FIFOLD_CLASS_CLASS1 | FIFOLDST_VLF |
-+ FIFOLD_TYPE_AAD | FIFOLD_TYPE_FLUSH1);
-+
-+ set_jump_tgt_here(desc, zero_assoc_jump_cmd1);
-+
-+ /* cryptlen = seqoutlen - assoclen */
-+ append_math_sub(desc, VARSEQINLEN, SEQOUTLEN, REG0, CAAM_CMD_SZ);
-+
-+ /* jump to zero-payload command if cryptlen is zero */
-+ zero_payload_jump_cmd = append_jump(desc, JUMP_TEST_ALL |
-+ JUMP_COND_MATH_Z);
-+
-+ append_math_sub(desc, VARSEQOUTLEN, SEQOUTLEN, REG0, CAAM_CMD_SZ);
-+
-+ /* store encrypted data */
-+ append_seq_fifo_store(desc, 0, FIFOST_TYPE_MESSAGE_DATA | FIFOLDST_VLF);
-+
-+ /* read payload data */
-+ append_seq_fifo_load(desc, 0, FIFOLD_CLASS_CLASS1 | FIFOLDST_VLF |
-+ FIFOLD_TYPE_MSG | FIFOLD_TYPE_FLUSH1);
-+
-+ /* zero-payload command */
-+ set_jump_tgt_here(desc, zero_payload_jump_cmd);
-+
-+ /* read ICV */
-+ append_seq_fifo_load(desc, icvsize, FIFOLD_CLASS_CLASS1 |
-+ FIFOLD_TYPE_ICV | FIFOLD_TYPE_LAST1);
-+
-+#ifdef DEBUG
-+ print_hex_dump(KERN_ERR, "gcm dec shdesc@" __stringify(__LINE__)": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, desc, desc_bytes(desc), 1);
-+#endif
-+}
-+EXPORT_SYMBOL(cnstr_shdsc_gcm_decap);
-+
-+/**
-+ * cnstr_shdsc_rfc4106_encap - IPSec ESP gcm encapsulation shared descriptor
-+ * (non-protocol).
-+ * @desc: pointer to buffer used for descriptor construction
-+ * @cdata: pointer to block cipher transform definitions
-+ * Valid algorithm values - OP_ALG_ALGSEL_AES ANDed with OP_ALG_AAI_GCM.
-+ * @ivsize: initialization vector size
-+ * @icvsize: integrity check value (ICV) size (truncated or full)
-+ * @is_qi: true when called from caam/qi
-+ */
-+void cnstr_shdsc_rfc4106_encap(u32 * const desc, struct alginfo *cdata,
-+ unsigned int ivsize, unsigned int icvsize,
-+ const bool is_qi)
-+{
-+ u32 *key_jump_cmd;
-+
-+ init_sh_desc(desc, HDR_SHARE_SERIAL);
-+
-+ /* Skip key loading if it is loaded due to sharing */
-+ key_jump_cmd = append_jump(desc, JUMP_JSL | JUMP_TEST_ALL |
-+ JUMP_COND_SHRD);
-+ if (cdata->key_inline)
-+ append_key_as_imm(desc, cdata->key_virt, cdata->keylen,
-+ cdata->keylen, CLASS_1 | KEY_DEST_CLASS_REG);
-+ else
-+ append_key(desc, cdata->key_dma, cdata->keylen, CLASS_1 |
-+ KEY_DEST_CLASS_REG);
-+ set_jump_tgt_here(desc, key_jump_cmd);
-+
-+ /* Class 1 operation */
-+ append_operation(desc, cdata->algtype | OP_ALG_AS_INITFINAL |
-+ OP_ALG_ENCRYPT);
-+
-+ if (is_qi) {
-+ u32 *wait_load_cmd;
-+
-+ /* REG3 = assoclen */
-+ append_seq_load(desc, 4, LDST_CLASS_DECO |
-+ LDST_SRCDST_WORD_DECO_MATH3 |
-+ (4 << LDST_OFFSET_SHIFT));
-+
-+ wait_load_cmd = append_jump(desc, JUMP_JSL | JUMP_TEST_ALL |
-+ JUMP_COND_CALM | JUMP_COND_NCP |
-+ JUMP_COND_NOP | JUMP_COND_NIP |
-+ JUMP_COND_NIFP);
-+ set_jump_tgt_here(desc, wait_load_cmd);
-+
-+ /* Read salt and IV */
-+ append_fifo_load_as_imm(desc, (void *)(cdata->key_virt +
-+ cdata->keylen), 4, FIFOLD_CLASS_CLASS1 |
-+ FIFOLD_TYPE_IV);
-+ append_seq_fifo_load(desc, ivsize, FIFOLD_CLASS_CLASS1 |
-+ FIFOLD_TYPE_IV | FIFOLD_TYPE_FLUSH1);
-+ }
-+
-+ append_math_sub_imm_u32(desc, VARSEQINLEN, REG3, IMM, ivsize);
-+ append_math_add(desc, VARSEQOUTLEN, ZERO, REG3, CAAM_CMD_SZ);
-+
-+ /* Read assoc data */
-+ append_seq_fifo_load(desc, 0, FIFOLD_CLASS_CLASS1 | FIFOLDST_VLF |
-+ FIFOLD_TYPE_AAD | FIFOLD_TYPE_FLUSH1);
-+
-+ /* Skip IV */
-+ append_seq_fifo_load(desc, ivsize, FIFOLD_CLASS_SKIP);
-+
-+ /* Will read cryptlen bytes */
-+ append_math_sub(desc, VARSEQINLEN, SEQINLEN, REG0, CAAM_CMD_SZ);
-+
-+ /* Workaround for erratum A-005473 (simultaneous SEQ FIFO skips) */
-+ append_seq_fifo_load(desc, 0, FIFOLD_CLASS_CLASS1 | FIFOLD_TYPE_MSG);
-+
-+ /* Skip assoc data */
-+ append_seq_fifo_store(desc, 0, FIFOST_TYPE_SKIP | FIFOLDST_VLF);
-+
-+ /* cryptlen = seqoutlen - assoclen */
-+ append_math_sub(desc, VARSEQOUTLEN, VARSEQINLEN, REG0, CAAM_CMD_SZ);
-+
-+ /* Write encrypted data */
-+ append_seq_fifo_store(desc, 0, FIFOST_TYPE_MESSAGE_DATA | FIFOLDST_VLF);
-+
-+ /* Read payload data */
-+ append_seq_fifo_load(desc, 0, FIFOLD_CLASS_CLASS1 | FIFOLDST_VLF |
-+ FIFOLD_TYPE_MSG | FIFOLD_TYPE_LAST1);
-+
-+ /* Write ICV */
-+ append_seq_store(desc, icvsize, LDST_CLASS_1_CCB |
-+ LDST_SRCDST_BYTE_CONTEXT);
-+
-+#ifdef DEBUG
-+ print_hex_dump(KERN_ERR,
-+ "rfc4106 enc shdesc@" __stringify(__LINE__)": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, desc, desc_bytes(desc), 1);
-+#endif
-+}
-+EXPORT_SYMBOL(cnstr_shdsc_rfc4106_encap);
-+
-+/**
-+ * cnstr_shdsc_rfc4106_decap - IPSec ESP gcm decapsulation shared descriptor
-+ * (non-protocol).
-+ * @desc: pointer to buffer used for descriptor construction
-+ * @cdata: pointer to block cipher transform definitions
-+ * Valid algorithm values - OP_ALG_ALGSEL_AES ANDed with OP_ALG_AAI_GCM.
-+ * @ivsize: initialization vector size
-+ * @icvsize: integrity check value (ICV) size (truncated or full)
-+ * @is_qi: true when called from caam/qi
-+ */
-+void cnstr_shdsc_rfc4106_decap(u32 * const desc, struct alginfo *cdata,
-+ unsigned int ivsize, unsigned int icvsize,
-+ const bool is_qi)
-+{
-+ u32 *key_jump_cmd;
-+
-+ init_sh_desc(desc, HDR_SHARE_SERIAL);
-+
-+ /* Skip key loading if it is loaded due to sharing */
-+ key_jump_cmd = append_jump(desc, JUMP_JSL | JUMP_TEST_ALL |
-+ JUMP_COND_SHRD);
-+ if (cdata->key_inline)
-+ append_key_as_imm(desc, cdata->key_virt, cdata->keylen,
-+ cdata->keylen, CLASS_1 |
-+ KEY_DEST_CLASS_REG);
-+ else
-+ append_key(desc, cdata->key_dma, cdata->keylen, CLASS_1 |
-+ KEY_DEST_CLASS_REG);
-+ set_jump_tgt_here(desc, key_jump_cmd);
-+
-+ /* Class 1 operation */
-+ append_operation(desc, cdata->algtype | OP_ALG_AS_INITFINAL |
-+ OP_ALG_DECRYPT | OP_ALG_ICV_ON);
-+
-+ if (is_qi) {
-+ u32 *wait_load_cmd;
-+
-+ /* REG3 = assoclen */
-+ append_seq_load(desc, 4, LDST_CLASS_DECO |
-+ LDST_SRCDST_WORD_DECO_MATH3 |
-+ (4 << LDST_OFFSET_SHIFT));
-+
-+ wait_load_cmd = append_jump(desc, JUMP_JSL | JUMP_TEST_ALL |
-+ JUMP_COND_CALM | JUMP_COND_NCP |
-+ JUMP_COND_NOP | JUMP_COND_NIP |
-+ JUMP_COND_NIFP);
-+ set_jump_tgt_here(desc, wait_load_cmd);
-+
-+ /* Read salt and IV */
-+ append_fifo_load_as_imm(desc, (void *)(cdata->key_virt +
-+ cdata->keylen), 4, FIFOLD_CLASS_CLASS1 |
-+ FIFOLD_TYPE_IV);
-+ append_seq_fifo_load(desc, ivsize, FIFOLD_CLASS_CLASS1 |
-+ FIFOLD_TYPE_IV | FIFOLD_TYPE_FLUSH1);
-+ }
-+
-+ append_math_sub_imm_u32(desc, VARSEQINLEN, REG3, IMM, ivsize);
-+ append_math_add(desc, VARSEQOUTLEN, ZERO, REG3, CAAM_CMD_SZ);
-+
-+ /* Read assoc data */
-+ append_seq_fifo_load(desc, 0, FIFOLD_CLASS_CLASS1 | FIFOLDST_VLF |
-+ FIFOLD_TYPE_AAD | FIFOLD_TYPE_FLUSH1);
-+
-+ /* Skip IV */
-+ append_seq_fifo_load(desc, ivsize, FIFOLD_CLASS_SKIP);
-+
-+ /* Will read cryptlen bytes */
-+ append_math_sub(desc, VARSEQINLEN, SEQOUTLEN, REG3, CAAM_CMD_SZ);
-+
-+ /* Workaround for erratum A-005473 (simultaneous SEQ FIFO skips) */
-+ append_seq_fifo_load(desc, 0, FIFOLD_CLASS_CLASS1 | FIFOLD_TYPE_MSG);
-+
-+ /* Skip assoc data */
-+ append_seq_fifo_store(desc, 0, FIFOST_TYPE_SKIP | FIFOLDST_VLF);
-+
-+ /* Will write cryptlen bytes */
-+ append_math_sub(desc, VARSEQOUTLEN, SEQOUTLEN, REG0, CAAM_CMD_SZ);
-+
-+ /* Store payload data */
-+ append_seq_fifo_store(desc, 0, FIFOST_TYPE_MESSAGE_DATA | FIFOLDST_VLF);
-+
-+ /* Read encrypted data */
-+ append_seq_fifo_load(desc, 0, FIFOLD_CLASS_CLASS1 | FIFOLDST_VLF |
-+ FIFOLD_TYPE_MSG | FIFOLD_TYPE_FLUSH1);
-+
-+ /* Read ICV */
-+ append_seq_fifo_load(desc, icvsize, FIFOLD_CLASS_CLASS1 |
-+ FIFOLD_TYPE_ICV | FIFOLD_TYPE_LAST1);
-+
-+#ifdef DEBUG
-+ print_hex_dump(KERN_ERR,
-+ "rfc4106 dec shdesc@" __stringify(__LINE__)": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, desc, desc_bytes(desc), 1);
-+#endif
-+}
-+EXPORT_SYMBOL(cnstr_shdsc_rfc4106_decap);
-+
-+/**
-+ * cnstr_shdsc_rfc4543_encap - IPSec ESP gmac encapsulation shared descriptor
-+ * (non-protocol).
-+ * @desc: pointer to buffer used for descriptor construction
-+ * @cdata: pointer to block cipher transform definitions
-+ * Valid algorithm values - OP_ALG_ALGSEL_AES ANDed with OP_ALG_AAI_GCM.
-+ * @ivsize: initialization vector size
-+ * @icvsize: integrity check value (ICV) size (truncated or full)
-+ * @is_qi: true when called from caam/qi
-+ */
-+void cnstr_shdsc_rfc4543_encap(u32 * const desc, struct alginfo *cdata,
-+ unsigned int ivsize, unsigned int icvsize,
-+ const bool is_qi)
-+{
-+ u32 *key_jump_cmd, *read_move_cmd, *write_move_cmd;
-+
-+ init_sh_desc(desc, HDR_SHARE_SERIAL);
-+
-+ /* Skip key loading if it is loaded due to sharing */
-+ key_jump_cmd = append_jump(desc, JUMP_JSL | JUMP_TEST_ALL |
-+ JUMP_COND_SHRD);
-+ if (cdata->key_inline)
-+ append_key_as_imm(desc, cdata->key_virt, cdata->keylen,
-+ cdata->keylen, CLASS_1 | KEY_DEST_CLASS_REG);
-+ else
-+ append_key(desc, cdata->key_dma, cdata->keylen, CLASS_1 |
-+ KEY_DEST_CLASS_REG);
-+ set_jump_tgt_here(desc, key_jump_cmd);
-+
-+ /* Class 1 operation */
-+ append_operation(desc, cdata->algtype | OP_ALG_AS_INITFINAL |
-+ OP_ALG_ENCRYPT);
-+
-+ if (is_qi) {
-+ /* assoclen is not needed, skip it */
-+ append_seq_fifo_load(desc, 4, FIFOLD_CLASS_SKIP);
-+
-+ /* Read salt and IV */
-+ append_fifo_load_as_imm(desc, (void *)(cdata->key_virt +
-+ cdata->keylen), 4, FIFOLD_CLASS_CLASS1 |
-+ FIFOLD_TYPE_IV);
-+ append_seq_fifo_load(desc, ivsize, FIFOLD_CLASS_CLASS1 |
-+ FIFOLD_TYPE_IV | FIFOLD_TYPE_FLUSH1);
-+ }
-+
-+ /* assoclen + cryptlen = seqinlen */
-+ append_math_sub(desc, REG3, SEQINLEN, REG0, CAAM_CMD_SZ);
-+
-+ /*
-+ * MOVE_LEN opcode is not available in all SEC HW revisions,
-+ * thus need to do some magic, i.e. self-patch the descriptor
-+ * buffer.
-+ */
-+ read_move_cmd = append_move(desc, MOVE_SRC_DESCBUF | MOVE_DEST_MATH3 |
-+ (0x6 << MOVE_LEN_SHIFT));
-+ write_move_cmd = append_move(desc, MOVE_SRC_MATH3 | MOVE_DEST_DESCBUF |
-+ (0x8 << MOVE_LEN_SHIFT));
-+
-+ /* Will read assoclen + cryptlen bytes */
-+ append_math_sub(desc, VARSEQINLEN, SEQINLEN, REG0, CAAM_CMD_SZ);
-+
-+ /* Will write assoclen + cryptlen bytes */
-+ append_math_sub(desc, VARSEQOUTLEN, SEQINLEN, REG0, CAAM_CMD_SZ);
-+
-+ /* Read and write assoclen + cryptlen bytes */
-+ aead_append_src_dst(desc, FIFOLD_TYPE_AAD);
-+
-+ set_move_tgt_here(desc, read_move_cmd);
-+ set_move_tgt_here(desc, write_move_cmd);
-+ append_cmd(desc, CMD_LOAD | DISABLE_AUTO_INFO_FIFO);
-+ /* Move payload data to OFIFO */
-+ append_move(desc, MOVE_SRC_INFIFO_CL | MOVE_DEST_OUTFIFO);
-+
-+ /* Write ICV */
-+ append_seq_store(desc, icvsize, LDST_CLASS_1_CCB |
-+ LDST_SRCDST_BYTE_CONTEXT);
-+
-+#ifdef DEBUG
-+ print_hex_dump(KERN_ERR,
-+ "rfc4543 enc shdesc@" __stringify(__LINE__)": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, desc, desc_bytes(desc), 1);
-+#endif
-+}
-+EXPORT_SYMBOL(cnstr_shdsc_rfc4543_encap);
-+
-+/**
-+ * cnstr_shdsc_rfc4543_decap - IPSec ESP gmac decapsulation shared descriptor
-+ * (non-protocol).
-+ * @desc: pointer to buffer used for descriptor construction
-+ * @cdata: pointer to block cipher transform definitions
-+ * Valid algorithm values - OP_ALG_ALGSEL_AES ANDed with OP_ALG_AAI_GCM.
-+ * @ivsize: initialization vector size
-+ * @icvsize: integrity check value (ICV) size (truncated or full)
-+ * @is_qi: true when called from caam/qi
-+ */
-+void cnstr_shdsc_rfc4543_decap(u32 * const desc, struct alginfo *cdata,
-+ unsigned int ivsize, unsigned int icvsize,
-+ const bool is_qi)
-+{
-+ u32 *key_jump_cmd, *read_move_cmd, *write_move_cmd;
-+
-+ init_sh_desc(desc, HDR_SHARE_SERIAL);
-+
-+ /* Skip key loading if it is loaded due to sharing */
-+ key_jump_cmd = append_jump(desc, JUMP_JSL | JUMP_TEST_ALL |
-+ JUMP_COND_SHRD);
-+ if (cdata->key_inline)
-+ append_key_as_imm(desc, cdata->key_virt, cdata->keylen,
-+ cdata->keylen, CLASS_1 | KEY_DEST_CLASS_REG);
-+ else
-+ append_key(desc, cdata->key_dma, cdata->keylen, CLASS_1 |
-+ KEY_DEST_CLASS_REG);
-+ set_jump_tgt_here(desc, key_jump_cmd);
-+
-+ /* Class 1 operation */
-+ append_operation(desc, cdata->algtype | OP_ALG_AS_INITFINAL |
-+ OP_ALG_DECRYPT | OP_ALG_ICV_ON);
-+
-+ if (is_qi) {
-+ /* assoclen is not needed, skip it */
-+ append_seq_fifo_load(desc, 4, FIFOLD_CLASS_SKIP);
-+
-+ /* Read salt and IV */
-+ append_fifo_load_as_imm(desc, (void *)(cdata->key_virt +
-+ cdata->keylen), 4, FIFOLD_CLASS_CLASS1 |
-+ FIFOLD_TYPE_IV);
-+ append_seq_fifo_load(desc, ivsize, FIFOLD_CLASS_CLASS1 |
-+ FIFOLD_TYPE_IV | FIFOLD_TYPE_FLUSH1);
-+ }
-+
-+ /* assoclen + cryptlen = seqoutlen */
-+ append_math_sub(desc, REG3, SEQOUTLEN, REG0, CAAM_CMD_SZ);
-+
-+ /*
-+ * MOVE_LEN opcode is not available in all SEC HW revisions,
-+ * thus need to do some magic, i.e. self-patch the descriptor
-+ * buffer.
-+ */
-+ read_move_cmd = append_move(desc, MOVE_SRC_DESCBUF | MOVE_DEST_MATH3 |
-+ (0x6 << MOVE_LEN_SHIFT));
-+ write_move_cmd = append_move(desc, MOVE_SRC_MATH3 | MOVE_DEST_DESCBUF |
-+ (0x8 << MOVE_LEN_SHIFT));
-+
-+ /* Will read assoclen + cryptlen bytes */
-+ append_math_sub(desc, VARSEQINLEN, SEQOUTLEN, REG0, CAAM_CMD_SZ);
-+
-+ /* Will write assoclen + cryptlen bytes */
-+ append_math_sub(desc, VARSEQOUTLEN, SEQOUTLEN, REG0, CAAM_CMD_SZ);
-+
-+ /* Store payload data */
-+ append_seq_fifo_store(desc, 0, FIFOST_TYPE_MESSAGE_DATA | FIFOLDST_VLF);
-+
-+ /* In-snoop assoclen + cryptlen data */
-+ append_seq_fifo_load(desc, 0, FIFOLD_CLASS_BOTH | FIFOLDST_VLF |
-+ FIFOLD_TYPE_AAD | FIFOLD_TYPE_LAST2FLUSH1);
-+
-+ set_move_tgt_here(desc, read_move_cmd);
-+ set_move_tgt_here(desc, write_move_cmd);
-+ append_cmd(desc, CMD_LOAD | DISABLE_AUTO_INFO_FIFO);
-+ /* Move payload data to OFIFO */
-+ append_move(desc, MOVE_SRC_INFIFO_CL | MOVE_DEST_OUTFIFO);
-+ append_cmd(desc, CMD_LOAD | ENABLE_AUTO_INFO_FIFO);
-+
-+ /* Read ICV */
-+ append_seq_fifo_load(desc, icvsize, FIFOLD_CLASS_CLASS1 |
-+ FIFOLD_TYPE_ICV | FIFOLD_TYPE_LAST1);
-+
-+#ifdef DEBUG
-+ print_hex_dump(KERN_ERR,
-+ "rfc4543 dec shdesc@" __stringify(__LINE__)": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, desc, desc_bytes(desc), 1);
-+#endif
-+}
-+EXPORT_SYMBOL(cnstr_shdsc_rfc4543_decap);
-+
-+/*
-+ * For ablkcipher encrypt and decrypt, read from req->src and
-+ * write to req->dst
-+ */
-+static inline void ablkcipher_append_src_dst(u32 *desc)
-+{
-+ append_math_add(desc, VARSEQOUTLEN, SEQINLEN, REG0, CAAM_CMD_SZ);
-+ append_math_add(desc, VARSEQINLEN, SEQINLEN, REG0, CAAM_CMD_SZ);
-+ append_seq_fifo_load(desc, 0, FIFOLD_CLASS_CLASS1 |
-+ KEY_VLF | FIFOLD_TYPE_MSG | FIFOLD_TYPE_LAST1);
-+ append_seq_fifo_store(desc, 0, FIFOST_TYPE_MESSAGE_DATA | KEY_VLF);
-+}
-+
-+/**
-+ * cnstr_shdsc_ablkcipher_encap - ablkcipher encapsulation shared descriptor
-+ * @desc: pointer to buffer used for descriptor construction
-+ * @cdata: pointer to block cipher transform definitions
-+ * Valid algorithm values - one of OP_ALG_ALGSEL_{AES, DES, 3DES} ANDed
-+ * with OP_ALG_AAI_CBC or OP_ALG_AAI_CTR_MOD128.
-+ * @ivsize: initialization vector size
-+ * @is_rfc3686: true when ctr(aes) is wrapped by rfc3686 template
-+ * @ctx1_iv_off: IV offset in CONTEXT1 register
-+ */
-+void cnstr_shdsc_ablkcipher_encap(u32 * const desc, struct alginfo *cdata,
-+ unsigned int ivsize, const bool is_rfc3686,
-+ const u32 ctx1_iv_off)
-+{
-+ u32 *key_jump_cmd;
-+
-+ init_sh_desc(desc, HDR_SHARE_SERIAL | HDR_SAVECTX);
-+ /* Skip if already shared */
-+ key_jump_cmd = append_jump(desc, JUMP_JSL | JUMP_TEST_ALL |
-+ JUMP_COND_SHRD);
-+
-+ /* Load class1 key only */
-+ append_key_as_imm(desc, cdata->key_virt, cdata->keylen,
-+ cdata->keylen, CLASS_1 | KEY_DEST_CLASS_REG);
-+
-+ /* Load nonce into CONTEXT1 reg */
-+ if (is_rfc3686) {
-+ const u8 *nonce = cdata->key_virt + cdata->keylen;
-+
-+ append_load_as_imm(desc, nonce, CTR_RFC3686_NONCE_SIZE,
-+ LDST_CLASS_IND_CCB |
-+ LDST_SRCDST_BYTE_OUTFIFO | LDST_IMM);
-+ append_move(desc, MOVE_WAITCOMP | MOVE_SRC_OUTFIFO |
-+ MOVE_DEST_CLASS1CTX | (16 << MOVE_OFFSET_SHIFT) |
-+ (CTR_RFC3686_NONCE_SIZE << MOVE_LEN_SHIFT));
-+ }
-+
-+ set_jump_tgt_here(desc, key_jump_cmd);
-+
-+ /* Load iv */
-+ append_seq_load(desc, ivsize, LDST_SRCDST_BYTE_CONTEXT |
-+ LDST_CLASS_1_CCB | (ctx1_iv_off << LDST_OFFSET_SHIFT));
-+
-+ /* Load counter into CONTEXT1 reg */
-+ if (is_rfc3686)
-+ append_load_imm_be32(desc, 1, LDST_IMM | LDST_CLASS_1_CCB |
-+ LDST_SRCDST_BYTE_CONTEXT |
-+ ((ctx1_iv_off + CTR_RFC3686_IV_SIZE) <<
-+ LDST_OFFSET_SHIFT));
-+
-+ /* Load operation */
-+ append_operation(desc, cdata->algtype | OP_ALG_AS_INITFINAL |
-+ OP_ALG_ENCRYPT);
-+
-+ /* Perform operation */
-+ ablkcipher_append_src_dst(desc);
-+
-+#ifdef DEBUG
-+ print_hex_dump(KERN_ERR,
-+ "ablkcipher enc shdesc@" __stringify(__LINE__)": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, desc, desc_bytes(desc), 1);
-+#endif
-+}
-+EXPORT_SYMBOL(cnstr_shdsc_ablkcipher_encap);
-+
-+/**
-+ * cnstr_shdsc_ablkcipher_decap - ablkcipher decapsulation shared descriptor
-+ * @desc: pointer to buffer used for descriptor construction
-+ * @cdata: pointer to block cipher transform definitions
-+ * Valid algorithm values - one of OP_ALG_ALGSEL_{AES, DES, 3DES} ANDed
-+ * with OP_ALG_AAI_CBC or OP_ALG_AAI_CTR_MOD128.
-+ * @ivsize: initialization vector size
-+ * @is_rfc3686: true when ctr(aes) is wrapped by rfc3686 template
-+ * @ctx1_iv_off: IV offset in CONTEXT1 register
-+ */
-+void cnstr_shdsc_ablkcipher_decap(u32 * const desc, struct alginfo *cdata,
-+ unsigned int ivsize, const bool is_rfc3686,
-+ const u32 ctx1_iv_off)
-+{
-+ u32 *key_jump_cmd;
-+
-+ init_sh_desc(desc, HDR_SHARE_SERIAL | HDR_SAVECTX);
-+ /* Skip if already shared */
-+ key_jump_cmd = append_jump(desc, JUMP_JSL | JUMP_TEST_ALL |
-+ JUMP_COND_SHRD);
-+
-+ /* Load class1 key only */
-+ append_key_as_imm(desc, cdata->key_virt, cdata->keylen,
-+ cdata->keylen, CLASS_1 | KEY_DEST_CLASS_REG);
-+
-+ /* Load nonce into CONTEXT1 reg */
-+ if (is_rfc3686) {
-+ const u8 *nonce = cdata->key_virt + cdata->keylen;
-+
-+ append_load_as_imm(desc, nonce, CTR_RFC3686_NONCE_SIZE,
-+ LDST_CLASS_IND_CCB |
-+ LDST_SRCDST_BYTE_OUTFIFO | LDST_IMM);
-+ append_move(desc, MOVE_WAITCOMP | MOVE_SRC_OUTFIFO |
-+ MOVE_DEST_CLASS1CTX | (16 << MOVE_OFFSET_SHIFT) |
-+ (CTR_RFC3686_NONCE_SIZE << MOVE_LEN_SHIFT));
-+ }
-+
-+ set_jump_tgt_here(desc, key_jump_cmd);
-+
-+ /* load IV */
-+ append_seq_load(desc, ivsize, LDST_SRCDST_BYTE_CONTEXT |
-+ LDST_CLASS_1_CCB | (ctx1_iv_off << LDST_OFFSET_SHIFT));
-+
-+ /* Load counter into CONTEXT1 reg */
-+ if (is_rfc3686)
-+ append_load_imm_be32(desc, 1, LDST_IMM | LDST_CLASS_1_CCB |
-+ LDST_SRCDST_BYTE_CONTEXT |
-+ ((ctx1_iv_off + CTR_RFC3686_IV_SIZE) <<
-+ LDST_OFFSET_SHIFT));
-+
-+ /* Choose operation */
-+ if (ctx1_iv_off)
-+ append_operation(desc, cdata->algtype | OP_ALG_AS_INITFINAL |
-+ OP_ALG_DECRYPT);
-+ else
-+ append_dec_op1(desc, cdata->algtype);
-+
-+ /* Perform operation */
-+ ablkcipher_append_src_dst(desc);
-+
-+#ifdef DEBUG
-+ print_hex_dump(KERN_ERR,
-+ "ablkcipher dec shdesc@" __stringify(__LINE__)": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, desc, desc_bytes(desc), 1);
-+#endif
-+}
-+EXPORT_SYMBOL(cnstr_shdsc_ablkcipher_decap);
-+
-+/**
-+ * cnstr_shdsc_ablkcipher_givencap - ablkcipher encapsulation shared descriptor
-+ * with HW-generated initialization vector.
-+ * @desc: pointer to buffer used for descriptor construction
-+ * @cdata: pointer to block cipher transform definitions
-+ * Valid algorithm values - one of OP_ALG_ALGSEL_{AES, DES, 3DES} ANDed
-+ * with OP_ALG_AAI_CBC.
-+ * @ivsize: initialization vector size
-+ * @is_rfc3686: true when ctr(aes) is wrapped by rfc3686 template
-+ * @ctx1_iv_off: IV offset in CONTEXT1 register
-+ */
-+void cnstr_shdsc_ablkcipher_givencap(u32 * const desc, struct alginfo *cdata,
-+ unsigned int ivsize, const bool is_rfc3686,
-+ const u32 ctx1_iv_off)
-+{
-+ u32 *key_jump_cmd, geniv;
-+
-+ init_sh_desc(desc, HDR_SHARE_SERIAL | HDR_SAVECTX);
-+ /* Skip if already shared */
-+ key_jump_cmd = append_jump(desc, JUMP_JSL | JUMP_TEST_ALL |
-+ JUMP_COND_SHRD);
-+
-+ /* Load class1 key only */
-+ append_key_as_imm(desc, cdata->key_virt, cdata->keylen,
-+ cdata->keylen, CLASS_1 | KEY_DEST_CLASS_REG);
-+
-+ /* Load Nonce into CONTEXT1 reg */
-+ if (is_rfc3686) {
-+ const u8 *nonce = cdata->key_virt + cdata->keylen;
-+
-+ append_load_as_imm(desc, nonce, CTR_RFC3686_NONCE_SIZE,
-+ LDST_CLASS_IND_CCB |
-+ LDST_SRCDST_BYTE_OUTFIFO | LDST_IMM);
-+ append_move(desc, MOVE_WAITCOMP | MOVE_SRC_OUTFIFO |
-+ MOVE_DEST_CLASS1CTX | (16 << MOVE_OFFSET_SHIFT) |
-+ (CTR_RFC3686_NONCE_SIZE << MOVE_LEN_SHIFT));
-+ }
-+ set_jump_tgt_here(desc, key_jump_cmd);
-+
-+ /* Generate IV */
-+ geniv = NFIFOENTRY_STYPE_PAD | NFIFOENTRY_DEST_DECO |
-+ NFIFOENTRY_DTYPE_MSG | NFIFOENTRY_LC1 | NFIFOENTRY_PTYPE_RND |
-+ (ivsize << NFIFOENTRY_DLEN_SHIFT);
-+ append_load_imm_u32(desc, geniv, LDST_CLASS_IND_CCB |
-+ LDST_SRCDST_WORD_INFO_FIFO | LDST_IMM);
-+ append_cmd(desc, CMD_LOAD | DISABLE_AUTO_INFO_FIFO);
-+ append_move(desc, MOVE_WAITCOMP | MOVE_SRC_INFIFO |
-+ MOVE_DEST_CLASS1CTX | (ivsize << MOVE_LEN_SHIFT) |
-+ (ctx1_iv_off << MOVE_OFFSET_SHIFT));
-+ append_cmd(desc, CMD_LOAD | ENABLE_AUTO_INFO_FIFO);
-+
-+ /* Copy generated IV to memory */
-+ append_seq_store(desc, ivsize, LDST_SRCDST_BYTE_CONTEXT |
-+ LDST_CLASS_1_CCB | (ctx1_iv_off << LDST_OFFSET_SHIFT));
-+
-+ /* Load Counter into CONTEXT1 reg */
-+ if (is_rfc3686)
-+ append_load_imm_be32(desc, 1, LDST_IMM | LDST_CLASS_1_CCB |
-+ LDST_SRCDST_BYTE_CONTEXT |
-+ ((ctx1_iv_off + CTR_RFC3686_IV_SIZE) <<
-+ LDST_OFFSET_SHIFT));
-+
-+ if (ctx1_iv_off)
-+ append_jump(desc, JUMP_JSL | JUMP_TEST_ALL | JUMP_COND_NCP |
-+ (1 << JUMP_OFFSET_SHIFT));
-+
-+ /* Load operation */
-+ append_operation(desc, cdata->algtype | OP_ALG_AS_INITFINAL |
-+ OP_ALG_ENCRYPT);
-+
-+ /* Perform operation */
-+ ablkcipher_append_src_dst(desc);
-+
-+#ifdef DEBUG
-+ print_hex_dump(KERN_ERR,
-+ "ablkcipher givenc shdesc@" __stringify(__LINE__) ": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, desc, desc_bytes(desc), 1);
-+#endif
-+}
-+EXPORT_SYMBOL(cnstr_shdsc_ablkcipher_givencap);
-+
-+/**
-+ * cnstr_shdsc_xts_ablkcipher_encap - xts ablkcipher encapsulation shared
-+ * descriptor
-+ * @desc: pointer to buffer used for descriptor construction
-+ * @cdata: pointer to block cipher transform definitions
-+ * Valid algorithm values - OP_ALG_ALGSEL_AES ANDed with OP_ALG_AAI_XTS.
-+ */
-+void cnstr_shdsc_xts_ablkcipher_encap(u32 * const desc, struct alginfo *cdata)
-+{
-+ __be64 sector_size = cpu_to_be64(512);
-+ u32 *key_jump_cmd;
-+
-+ init_sh_desc(desc, HDR_SHARE_SERIAL | HDR_SAVECTX);
-+ /* Skip if already shared */
-+ key_jump_cmd = append_jump(desc, JUMP_JSL | JUMP_TEST_ALL |
-+ JUMP_COND_SHRD);
-+
-+ /* Load class1 keys only */
-+ append_key_as_imm(desc, cdata->key_virt, cdata->keylen,
-+ cdata->keylen, CLASS_1 | KEY_DEST_CLASS_REG);
-+
-+ /* Load sector size with index 40 bytes (0x28) */
-+ append_load_as_imm(desc, (void *)&sector_size, 8, LDST_CLASS_1_CCB |
-+ LDST_SRCDST_BYTE_CONTEXT |
-+ (0x28 << LDST_OFFSET_SHIFT));
-+
-+ set_jump_tgt_here(desc, key_jump_cmd);
-+
-+ /*
-+ * create sequence for loading the sector index
-+ * Upper 8B of IV - will be used as sector index
-+ * Lower 8B of IV - will be discarded
-+ */
-+ append_seq_load(desc, 8, LDST_SRCDST_BYTE_CONTEXT | LDST_CLASS_1_CCB |
-+ (0x20 << LDST_OFFSET_SHIFT));
-+ append_seq_fifo_load(desc, 8, FIFOLD_CLASS_SKIP);
-+
-+ /* Load operation */
-+ append_operation(desc, cdata->algtype | OP_ALG_AS_INITFINAL |
-+ OP_ALG_ENCRYPT);
-+
-+ /* Perform operation */
-+ ablkcipher_append_src_dst(desc);
-+
-+#ifdef DEBUG
-+ print_hex_dump(KERN_ERR,
-+ "xts ablkcipher enc shdesc@" __stringify(__LINE__) ": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, desc, desc_bytes(desc), 1);
-+#endif
-+}
-+EXPORT_SYMBOL(cnstr_shdsc_xts_ablkcipher_encap);
-+
-+/**
-+ * cnstr_shdsc_xts_ablkcipher_decap - xts ablkcipher decapsulation shared
-+ * descriptor
-+ * @desc: pointer to buffer used for descriptor construction
-+ * @cdata: pointer to block cipher transform definitions
-+ * Valid algorithm values - OP_ALG_ALGSEL_AES ANDed with OP_ALG_AAI_XTS.
-+ */
-+void cnstr_shdsc_xts_ablkcipher_decap(u32 * const desc, struct alginfo *cdata)
-+{
-+ __be64 sector_size = cpu_to_be64(512);
-+ u32 *key_jump_cmd;
-+
-+ init_sh_desc(desc, HDR_SHARE_SERIAL | HDR_SAVECTX);
-+ /* Skip if already shared */
-+ key_jump_cmd = append_jump(desc, JUMP_JSL | JUMP_TEST_ALL |
-+ JUMP_COND_SHRD);
-+
-+ /* Load class1 key only */
-+ append_key_as_imm(desc, cdata->key_virt, cdata->keylen,
-+ cdata->keylen, CLASS_1 | KEY_DEST_CLASS_REG);
-+
-+ /* Load sector size with index 40 bytes (0x28) */
-+ append_load_as_imm(desc, (void *)&sector_size, 8, LDST_CLASS_1_CCB |
-+ LDST_SRCDST_BYTE_CONTEXT |
-+ (0x28 << LDST_OFFSET_SHIFT));
-+
-+ set_jump_tgt_here(desc, key_jump_cmd);
-+
-+ /*
-+ * create sequence for loading the sector index
-+ * Upper 8B of IV - will be used as sector index
-+ * Lower 8B of IV - will be discarded
-+ */
-+ append_seq_load(desc, 8, LDST_SRCDST_BYTE_CONTEXT | LDST_CLASS_1_CCB |
-+ (0x20 << LDST_OFFSET_SHIFT));
-+ append_seq_fifo_load(desc, 8, FIFOLD_CLASS_SKIP);
-+
-+ /* Load operation */
-+ append_dec_op1(desc, cdata->algtype);
-+
-+ /* Perform operation */
-+ ablkcipher_append_src_dst(desc);
-+
-+#ifdef DEBUG
-+ print_hex_dump(KERN_ERR,
-+ "xts ablkcipher dec shdesc@" __stringify(__LINE__) ": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, desc, desc_bytes(desc), 1);
-+#endif
-+}
-+EXPORT_SYMBOL(cnstr_shdsc_xts_ablkcipher_decap);
-+
-+MODULE_LICENSE("GPL");
-+MODULE_DESCRIPTION("FSL CAAM descriptor support");
-+MODULE_AUTHOR("Freescale Semiconductor - NMG/STC");
---- /dev/null
-+++ b/drivers/crypto/caam/caamalg_desc.h
-@@ -0,0 +1,127 @@
-+/*
-+ * Shared descriptors for aead, ablkcipher algorithms
-+ *
-+ * Copyright 2016 NXP
-+ */
-+
-+#ifndef _CAAMALG_DESC_H_
-+#define _CAAMALG_DESC_H_
-+
-+/* length of descriptors text */
-+#define DESC_AEAD_BASE (4 * CAAM_CMD_SZ)
-+#define DESC_AEAD_ENC_LEN (DESC_AEAD_BASE + 11 * CAAM_CMD_SZ)
-+#define DESC_AEAD_DEC_LEN (DESC_AEAD_BASE + 15 * CAAM_CMD_SZ)
-+#define DESC_AEAD_GIVENC_LEN (DESC_AEAD_ENC_LEN + 7 * CAAM_CMD_SZ)
-+#define DESC_QI_AEAD_ENC_LEN (DESC_AEAD_ENC_LEN + 3 * CAAM_CMD_SZ)
-+#define DESC_QI_AEAD_DEC_LEN (DESC_AEAD_DEC_LEN + 3 * CAAM_CMD_SZ)
-+#define DESC_QI_AEAD_GIVENC_LEN (DESC_AEAD_GIVENC_LEN + 3 * CAAM_CMD_SZ)
-+
-+#define DESC_TLS_BASE (4 * CAAM_CMD_SZ)
-+#define DESC_TLS10_ENC_LEN (DESC_TLS_BASE + 29 * CAAM_CMD_SZ)
-+
-+/* Note: Nonce is counted in cdata.keylen */
-+#define DESC_AEAD_CTR_RFC3686_LEN (4 * CAAM_CMD_SZ)
-+
-+#define DESC_AEAD_NULL_BASE (3 * CAAM_CMD_SZ)
-+#define DESC_AEAD_NULL_ENC_LEN (DESC_AEAD_NULL_BASE + 11 * CAAM_CMD_SZ)
-+#define DESC_AEAD_NULL_DEC_LEN (DESC_AEAD_NULL_BASE + 13 * CAAM_CMD_SZ)
-+
-+#define DESC_GCM_BASE (3 * CAAM_CMD_SZ)
-+#define DESC_GCM_ENC_LEN (DESC_GCM_BASE + 16 * CAAM_CMD_SZ)
-+#define DESC_GCM_DEC_LEN (DESC_GCM_BASE + 12 * CAAM_CMD_SZ)
-+#define DESC_QI_GCM_ENC_LEN (DESC_GCM_ENC_LEN + 6 * CAAM_CMD_SZ)
-+#define DESC_QI_GCM_DEC_LEN (DESC_GCM_DEC_LEN + 3 * CAAM_CMD_SZ)
-+
-+#define DESC_RFC4106_BASE (3 * CAAM_CMD_SZ)
-+#define DESC_RFC4106_ENC_LEN (DESC_RFC4106_BASE + 13 * CAAM_CMD_SZ)
-+#define DESC_RFC4106_DEC_LEN (DESC_RFC4106_BASE + 13 * CAAM_CMD_SZ)
-+#define DESC_QI_RFC4106_ENC_LEN (DESC_RFC4106_ENC_LEN + 5 * CAAM_CMD_SZ)
-+#define DESC_QI_RFC4106_DEC_LEN (DESC_RFC4106_DEC_LEN + 5 * CAAM_CMD_SZ)
-+
-+#define DESC_RFC4543_BASE (3 * CAAM_CMD_SZ)
-+#define DESC_RFC4543_ENC_LEN (DESC_RFC4543_BASE + 11 * CAAM_CMD_SZ)
-+#define DESC_RFC4543_DEC_LEN (DESC_RFC4543_BASE + 12 * CAAM_CMD_SZ)
-+#define DESC_QI_RFC4543_ENC_LEN (DESC_RFC4543_ENC_LEN + 4 * CAAM_CMD_SZ)
-+#define DESC_QI_RFC4543_DEC_LEN (DESC_RFC4543_DEC_LEN + 4 * CAAM_CMD_SZ)
-+
-+#define DESC_ABLKCIPHER_BASE (3 * CAAM_CMD_SZ)
-+#define DESC_ABLKCIPHER_ENC_LEN (DESC_ABLKCIPHER_BASE + \
-+ 20 * CAAM_CMD_SZ)
-+#define DESC_ABLKCIPHER_DEC_LEN (DESC_ABLKCIPHER_BASE + \
-+ 15 * CAAM_CMD_SZ)
-+
-+void cnstr_shdsc_aead_null_encap(u32 * const desc, struct alginfo *adata,
-+ unsigned int icvsize, int era);
-+
-+void cnstr_shdsc_aead_null_decap(u32 * const desc, struct alginfo *adata,
-+ unsigned int icvsize, int era);
-+
-+void cnstr_shdsc_aead_encap(u32 * const desc, struct alginfo *cdata,
-+ struct alginfo *adata, unsigned int ivsize,
-+ unsigned int icvsize, const bool is_rfc3686,
-+ u32 *nonce, const u32 ctx1_iv_off,
-+ const bool is_qi, int era);
-+
-+void cnstr_shdsc_aead_decap(u32 * const desc, struct alginfo *cdata,
-+ struct alginfo *adata, unsigned int ivsize,
-+ unsigned int icvsize, const bool geniv,
-+ const bool is_rfc3686, u32 *nonce,
-+ const u32 ctx1_iv_off, const bool is_qi, int era);
-+
-+void cnstr_shdsc_aead_givencap(u32 * const desc, struct alginfo *cdata,
-+ struct alginfo *adata, unsigned int ivsize,
-+ unsigned int icvsize, const bool is_rfc3686,
-+ u32 *nonce, const u32 ctx1_iv_off,
-+ const bool is_qi, int era);
-+
-+void cnstr_shdsc_tls_encap(u32 *const desc, struct alginfo *cdata,
-+ struct alginfo *adata, unsigned int assoclen,
-+ unsigned int ivsize, unsigned int authsize,
-+ unsigned int blocksize, int era);
-+
-+void cnstr_shdsc_tls_decap(u32 *const desc, struct alginfo *cdata,
-+ struct alginfo *adata, unsigned int assoclen,
-+ unsigned int ivsize, unsigned int authsize,
-+ unsigned int blocksize, int era);
-+
-+void cnstr_shdsc_gcm_encap(u32 * const desc, struct alginfo *cdata,
-+ unsigned int ivsize, unsigned int icvsize,
-+ const bool is_qi);
-+
-+void cnstr_shdsc_gcm_decap(u32 * const desc, struct alginfo *cdata,
-+ unsigned int ivsize, unsigned int icvsize,
-+ const bool is_qi);
-+
-+void cnstr_shdsc_rfc4106_encap(u32 * const desc, struct alginfo *cdata,
-+ unsigned int ivsize, unsigned int icvsize,
-+ const bool is_qi);
-+
-+void cnstr_shdsc_rfc4106_decap(u32 * const desc, struct alginfo *cdata,
-+ unsigned int ivsize, unsigned int icvsize,
-+ const bool is_qi);
-+
-+void cnstr_shdsc_rfc4543_encap(u32 * const desc, struct alginfo *cdata,
-+ unsigned int ivsize, unsigned int icvsize,
-+ const bool is_qi);
-+
-+void cnstr_shdsc_rfc4543_decap(u32 * const desc, struct alginfo *cdata,
-+ unsigned int ivsize, unsigned int icvsize,
-+ const bool is_qi);
-+
-+void cnstr_shdsc_ablkcipher_encap(u32 * const desc, struct alginfo *cdata,
-+ unsigned int ivsize, const bool is_rfc3686,
-+ const u32 ctx1_iv_off);
-+
-+void cnstr_shdsc_ablkcipher_decap(u32 * const desc, struct alginfo *cdata,
-+ unsigned int ivsize, const bool is_rfc3686,
-+ const u32 ctx1_iv_off);
-+
-+void cnstr_shdsc_ablkcipher_givencap(u32 * const desc, struct alginfo *cdata,
-+ unsigned int ivsize, const bool is_rfc3686,
-+ const u32 ctx1_iv_off);
-+
-+void cnstr_shdsc_xts_ablkcipher_encap(u32 * const desc, struct alginfo *cdata);
-+
-+void cnstr_shdsc_xts_ablkcipher_decap(u32 * const desc, struct alginfo *cdata);
-+
-+#endif /* _CAAMALG_DESC_H_ */
---- /dev/null
-+++ b/drivers/crypto/caam/caamalg_qi.c
-@@ -0,0 +1,3321 @@
-+/*
-+ * Freescale FSL CAAM support for crypto API over QI backend.
-+ * Based on caamalg.c
-+ *
-+ * Copyright 2013-2016 Freescale Semiconductor, Inc.
-+ * Copyright 2016-2017 NXP
-+ */
-+
-+#include "compat.h"
-+#include "ctrl.h"
-+#include "regs.h"
-+#include "intern.h"
-+#include "desc_constr.h"
-+#include "error.h"
-+#include "sg_sw_qm.h"
-+#include "key_gen.h"
-+#include "qi.h"
-+#include "jr.h"
-+#include "caamalg_desc.h"
-+
-+/*
-+ * crypto alg
-+ */
-+#define CAAM_CRA_PRIORITY 2000
-+/* max key is sum of AES_MAX_KEY_SIZE, max split key size */
-+#define CAAM_MAX_KEY_SIZE (AES_MAX_KEY_SIZE + \
-+ SHA512_DIGEST_SIZE * 2)
-+
-+#define DESC_MAX_USED_BYTES (DESC_QI_AEAD_GIVENC_LEN + \
-+ CAAM_MAX_KEY_SIZE)
-+#define DESC_MAX_USED_LEN (DESC_MAX_USED_BYTES / CAAM_CMD_SZ)
-+
-+struct caam_alg_entry {
-+ int class1_alg_type;
-+ int class2_alg_type;
-+ bool rfc3686;
-+ bool geniv;
-+};
-+
-+struct caam_aead_alg {
-+ struct aead_alg aead;
-+ struct caam_alg_entry caam;
-+ bool registered;
-+};
-+
-+/*
-+ * per-session context
-+ */
-+struct caam_ctx {
-+ struct device *jrdev;
-+ u32 sh_desc_enc[DESC_MAX_USED_LEN];
-+ u32 sh_desc_dec[DESC_MAX_USED_LEN];
-+ u32 sh_desc_givenc[DESC_MAX_USED_LEN];
-+ u8 key[CAAM_MAX_KEY_SIZE];
-+ dma_addr_t key_dma;
-+ enum dma_data_direction dir;
-+ struct alginfo adata;
-+ struct alginfo cdata;
-+ unsigned int authsize;
-+ struct device *qidev;
-+ spinlock_t lock; /* Protects multiple init of driver context */
-+ struct caam_drv_ctx *drv_ctx[NUM_OP];
-+};
-+
-+static int aead_set_sh_desc(struct crypto_aead *aead)
-+{
-+ struct caam_aead_alg *alg = container_of(crypto_aead_alg(aead),
-+ typeof(*alg), aead);
-+ struct caam_ctx *ctx = crypto_aead_ctx(aead);
-+ unsigned int ivsize = crypto_aead_ivsize(aead);
-+ u32 ctx1_iv_off = 0;
-+ u32 *nonce = NULL;
-+ unsigned int data_len[2];
-+ u32 inl_mask;
-+ const bool ctr_mode = ((ctx->cdata.algtype & OP_ALG_AAI_MASK) ==
-+ OP_ALG_AAI_CTR_MOD128);
-+ const bool is_rfc3686 = alg->caam.rfc3686;
-+ struct caam_drv_private *ctrlpriv = dev_get_drvdata(ctx->jrdev->parent);
-+
-+ if (!ctx->cdata.keylen || !ctx->authsize)
-+ return 0;
-+
-+ /*
-+ * AES-CTR needs to load IV in CONTEXT1 reg
-+ * at an offset of 128bits (16bytes)
-+ * CONTEXT1[255:128] = IV
-+ */
-+ if (ctr_mode)
-+ ctx1_iv_off = 16;
-+
-+ /*
-+ * RFC3686 specific:
-+ * CONTEXT1[255:128] = {NONCE, IV, COUNTER}
-+ */
-+ if (is_rfc3686) {
-+ ctx1_iv_off = 16 + CTR_RFC3686_NONCE_SIZE;
-+ nonce = (u32 *)((void *)ctx->key + ctx->adata.keylen_pad +
-+ ctx->cdata.keylen - CTR_RFC3686_NONCE_SIZE);
-+ }
-+
-+ data_len[0] = ctx->adata.keylen_pad;
-+ data_len[1] = ctx->cdata.keylen;
-+
-+ if (alg->caam.geniv)
-+ goto skip_enc;
-+
-+ /* aead_encrypt shared descriptor */
-+ if (desc_inline_query(DESC_QI_AEAD_ENC_LEN +
-+ (is_rfc3686 ? DESC_AEAD_CTR_RFC3686_LEN : 0),
-+ DESC_JOB_IO_LEN, data_len, &inl_mask,
-+ ARRAY_SIZE(data_len)) < 0)
-+ return -EINVAL;
-+
-+ if (inl_mask & 1)
-+ ctx->adata.key_virt = ctx->key;
-+ else
-+ ctx->adata.key_dma = ctx->key_dma;
-+
-+ if (inl_mask & 2)
-+ ctx->cdata.key_virt = ctx->key + ctx->adata.keylen_pad;
-+ else
-+ ctx->cdata.key_dma = ctx->key_dma + ctx->adata.keylen_pad;
-+
-+ ctx->adata.key_inline = !!(inl_mask & 1);
-+ ctx->cdata.key_inline = !!(inl_mask & 2);
-+
-+ cnstr_shdsc_aead_encap(ctx->sh_desc_enc, &ctx->cdata, &ctx->adata,
-+ ivsize, ctx->authsize, is_rfc3686, nonce,
-+ ctx1_iv_off, true, ctrlpriv->era);
-+
-+skip_enc:
-+ /* aead_decrypt shared descriptor */
-+ if (desc_inline_query(DESC_QI_AEAD_DEC_LEN +
-+ (is_rfc3686 ? DESC_AEAD_CTR_RFC3686_LEN : 0),
-+ DESC_JOB_IO_LEN, data_len, &inl_mask,
-+ ARRAY_SIZE(data_len)) < 0)
-+ return -EINVAL;
-+
-+ if (inl_mask & 1)
-+ ctx->adata.key_virt = ctx->key;
-+ else
-+ ctx->adata.key_dma = ctx->key_dma;
-+
-+ if (inl_mask & 2)
-+ ctx->cdata.key_virt = ctx->key + ctx->adata.keylen_pad;
-+ else
-+ ctx->cdata.key_dma = ctx->key_dma + ctx->adata.keylen_pad;
-+
-+ ctx->adata.key_inline = !!(inl_mask & 1);
-+ ctx->cdata.key_inline = !!(inl_mask & 2);
-+
-+ cnstr_shdsc_aead_decap(ctx->sh_desc_dec, &ctx->cdata, &ctx->adata,
-+ ivsize, ctx->authsize, alg->caam.geniv,
-+ is_rfc3686, nonce, ctx1_iv_off, true,
-+ ctrlpriv->era);
-+
-+ if (!alg->caam.geniv)
-+ goto skip_givenc;
-+
-+ /* aead_givencrypt shared descriptor */
-+ if (desc_inline_query(DESC_QI_AEAD_GIVENC_LEN +
-+ (is_rfc3686 ? DESC_AEAD_CTR_RFC3686_LEN : 0),
-+ DESC_JOB_IO_LEN, data_len, &inl_mask,
-+ ARRAY_SIZE(data_len)) < 0)
-+ return -EINVAL;
-+
-+ if (inl_mask & 1)
-+ ctx->adata.key_virt = ctx->key;
-+ else
-+ ctx->adata.key_dma = ctx->key_dma;
-+
-+ if (inl_mask & 2)
-+ ctx->cdata.key_virt = ctx->key + ctx->adata.keylen_pad;
-+ else
-+ ctx->cdata.key_dma = ctx->key_dma + ctx->adata.keylen_pad;
-+
-+ ctx->adata.key_inline = !!(inl_mask & 1);
-+ ctx->cdata.key_inline = !!(inl_mask & 2);
-+
-+ cnstr_shdsc_aead_givencap(ctx->sh_desc_enc, &ctx->cdata, &ctx->adata,
-+ ivsize, ctx->authsize, is_rfc3686, nonce,
-+ ctx1_iv_off, true, ctrlpriv->era);
-+
-+skip_givenc:
-+ return 0;
-+}
-+
-+static int aead_setauthsize(struct crypto_aead *authenc, unsigned int authsize)
-+{
-+ struct caam_ctx *ctx = crypto_aead_ctx(authenc);
-+
-+ ctx->authsize = authsize;
-+ aead_set_sh_desc(authenc);
-+
-+ return 0;
-+}
-+
-+static int aead_setkey(struct crypto_aead *aead, const u8 *key,
-+ unsigned int keylen)
-+{
-+ struct caam_ctx *ctx = crypto_aead_ctx(aead);
-+ struct device *jrdev = ctx->jrdev;
-+ struct caam_drv_private *ctrlpriv = dev_get_drvdata(jrdev->parent);
-+ struct crypto_authenc_keys keys;
-+ int ret = 0;
-+
-+ if (crypto_authenc_extractkeys(&keys, key, keylen) != 0)
-+ goto badkey;
-+
-+#ifdef DEBUG
-+ dev_err(jrdev, "keylen %d enckeylen %d authkeylen %d\n",
-+ keys.authkeylen + keys.enckeylen, keys.enckeylen,
-+ keys.authkeylen);
-+ print_hex_dump(KERN_ERR, "key in @" __stringify(__LINE__)": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, key, keylen, 1);
-+#endif
-+
-+ /*
-+ * If DKP is supported, use it in the shared descriptor to generate
-+ * the split key.
-+ */
-+ if (ctrlpriv->era >= 6) {
-+ ctx->adata.keylen = keys.authkeylen;
-+ ctx->adata.keylen_pad = split_key_len(ctx->adata.algtype &
-+ OP_ALG_ALGSEL_MASK);
-+
-+ if (ctx->adata.keylen_pad + keys.enckeylen > CAAM_MAX_KEY_SIZE)
-+ goto badkey;
-+
-+ memcpy(ctx->key, keys.authkey, keys.authkeylen);
-+ memcpy(ctx->key + ctx->adata.keylen_pad, keys.enckey,
-+ keys.enckeylen);
-+ dma_sync_single_for_device(jrdev, ctx->key_dma,
-+ ctx->adata.keylen_pad +
-+ keys.enckeylen, ctx->dir);
-+ goto skip_split_key;
-+ }
-+
-+ ret = gen_split_key(jrdev, ctx->key, &ctx->adata, keys.authkey,
-+ keys.authkeylen, CAAM_MAX_KEY_SIZE -
-+ keys.enckeylen);
-+ if (ret)
-+ goto badkey;
-+
-+ /* postpend encryption key to auth split key */
-+ memcpy(ctx->key + ctx->adata.keylen_pad, keys.enckey, keys.enckeylen);
-+ dma_sync_single_for_device(jrdev, ctx->key_dma, ctx->adata.keylen_pad +
-+ keys.enckeylen, ctx->dir);
-+#ifdef DEBUG
-+ print_hex_dump(KERN_ERR, "ctx.key@" __stringify(__LINE__)": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, ctx->key,
-+ ctx->adata.keylen_pad + keys.enckeylen, 1);
-+#endif
-+
-+skip_split_key:
-+ ctx->cdata.keylen = keys.enckeylen;
-+
-+ ret = aead_set_sh_desc(aead);
-+ if (ret)
-+ goto badkey;
-+
-+ /* Now update the driver contexts with the new shared descriptor */
-+ if (ctx->drv_ctx[ENCRYPT]) {
-+ ret = caam_drv_ctx_update(ctx->drv_ctx[ENCRYPT],
-+ ctx->sh_desc_enc);
-+ if (ret) {
-+ dev_err(jrdev, "driver enc context update failed\n");
-+ goto badkey;
-+ }
-+ }
-+
-+ if (ctx->drv_ctx[DECRYPT]) {
-+ ret = caam_drv_ctx_update(ctx->drv_ctx[DECRYPT],
-+ ctx->sh_desc_dec);
-+ if (ret) {
-+ dev_err(jrdev, "driver dec context update failed\n");
-+ goto badkey;
-+ }
-+ }
-+
-+ return ret;
-+badkey:
-+ crypto_aead_set_flags(aead, CRYPTO_TFM_RES_BAD_KEY_LEN);
-+ return -EINVAL;
-+}
-+
-+static int tls_set_sh_desc(struct crypto_aead *tls)
-+{
-+ struct caam_ctx *ctx = crypto_aead_ctx(tls);
-+ unsigned int ivsize = crypto_aead_ivsize(tls);
-+ unsigned int blocksize = crypto_aead_blocksize(tls);
-+ unsigned int assoclen = 13; /* always 13 bytes for TLS */
-+ unsigned int data_len[2];
-+ u32 inl_mask;
-+ struct caam_drv_private *ctrlpriv = dev_get_drvdata(ctx->jrdev->parent);
-+
-+ if (!ctx->cdata.keylen || !ctx->authsize)
-+ return 0;
-+
-+ /*
-+ * TLS 1.0 encrypt shared descriptor
-+ * Job Descriptor and Shared Descriptor
-+ * must fit into the 64-word Descriptor h/w Buffer
-+ */
-+ data_len[0] = ctx->adata.keylen_pad;
-+ data_len[1] = ctx->cdata.keylen;
-+
-+ if (desc_inline_query(DESC_TLS10_ENC_LEN, DESC_JOB_IO_LEN, data_len,
-+ &inl_mask, ARRAY_SIZE(data_len)) < 0)
-+ return -EINVAL;
-+
-+ if (inl_mask & 1)
-+ ctx->adata.key_virt = ctx->key;
-+ else
-+ ctx->adata.key_dma = ctx->key_dma;
-+
-+ if (inl_mask & 2)
-+ ctx->cdata.key_virt = ctx->key + ctx->adata.keylen_pad;
-+ else
-+ ctx->cdata.key_dma = ctx->key_dma + ctx->adata.keylen_pad;
-+
-+ ctx->adata.key_inline = !!(inl_mask & 1);
-+ ctx->cdata.key_inline = !!(inl_mask & 2);
-+
-+ cnstr_shdsc_tls_encap(ctx->sh_desc_enc, &ctx->cdata, &ctx->adata,
-+ assoclen, ivsize, ctx->authsize, blocksize,
-+ ctrlpriv->era);
-+
-+ /*
-+ * TLS 1.0 decrypt shared descriptor
-+ * Keys do not fit inline, regardless of algorithms used
-+ */
-+ ctx->adata.key_inline = false;
-+ ctx->adata.key_dma = ctx->key_dma;
-+ ctx->cdata.key_dma = ctx->key_dma + ctx->adata.keylen_pad;
-+
-+ cnstr_shdsc_tls_decap(ctx->sh_desc_dec, &ctx->cdata, &ctx->adata,
-+ assoclen, ivsize, ctx->authsize, blocksize,
-+ ctrlpriv->era);
-+
-+ return 0;
-+}
-+
-+static int tls_setauthsize(struct crypto_aead *tls, unsigned int authsize)
-+{
-+ struct caam_ctx *ctx = crypto_aead_ctx(tls);
-+
-+ ctx->authsize = authsize;
-+ tls_set_sh_desc(tls);
-+
-+ return 0;
-+}
-+
-+static int tls_setkey(struct crypto_aead *tls, const u8 *key,
-+ unsigned int keylen)
-+{
-+ struct caam_ctx *ctx = crypto_aead_ctx(tls);
-+ struct device *jrdev = ctx->jrdev;
-+ struct caam_drv_private *ctrlpriv = dev_get_drvdata(jrdev->parent);
-+ struct crypto_authenc_keys keys;
-+ int ret = 0;
-+
-+ if (crypto_authenc_extractkeys(&keys, key, keylen) != 0)
-+ goto badkey;
-+
-+#ifdef DEBUG
-+ dev_err(jrdev, "keylen %d enckeylen %d authkeylen %d\n",
-+ keys.authkeylen + keys.enckeylen, keys.enckeylen,
-+ keys.authkeylen);
-+ print_hex_dump(KERN_ERR, "key in @" __stringify(__LINE__)": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, key, keylen, 1);
-+#endif
-+
-+ /*
-+ * If DKP is supported, use it in the shared descriptor to generate
-+ * the split key.
-+ */
-+ if (ctrlpriv->era >= 6) {
-+ ctx->adata.keylen = keys.authkeylen;
-+ ctx->adata.keylen_pad = split_key_len(ctx->adata.algtype &
-+ OP_ALG_ALGSEL_MASK);
-+
-+ if (ctx->adata.keylen_pad + keys.enckeylen > CAAM_MAX_KEY_SIZE)
-+ goto badkey;
-+
-+ memcpy(ctx->key, keys.authkey, keys.authkeylen);
-+ memcpy(ctx->key + ctx->adata.keylen_pad, keys.enckey,
-+ keys.enckeylen);
-+ dma_sync_single_for_device(jrdev, ctx->key_dma,
-+ ctx->adata.keylen_pad +
-+ keys.enckeylen, ctx->dir);
-+ goto skip_split_key;
-+ }
-+
-+ ret = gen_split_key(jrdev, ctx->key, &ctx->adata, keys.authkey,
-+ keys.authkeylen, CAAM_MAX_KEY_SIZE -
-+ keys.enckeylen);
-+ if (ret)
-+ goto badkey;
-+
-+ /* postpend encryption key to auth split key */
-+ memcpy(ctx->key + ctx->adata.keylen_pad, keys.enckey, keys.enckeylen);
-+ dma_sync_single_for_device(jrdev, ctx->key_dma, ctx->adata.keylen_pad +
-+ keys.enckeylen, ctx->dir);
-+
-+#ifdef DEBUG
-+ dev_err(jrdev, "split keylen %d split keylen padded %d\n",
-+ ctx->adata.keylen, ctx->adata.keylen_pad);
-+ print_hex_dump(KERN_ERR, "ctx.key@" __stringify(__LINE__)": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, ctx->key,
-+ ctx->adata.keylen_pad + keys.enckeylen, 1);
-+#endif
-+
-+skip_split_key:
-+ ctx->cdata.keylen = keys.enckeylen;
-+
-+ ret = tls_set_sh_desc(tls);
-+ if (ret)
-+ goto badkey;
-+
-+ /* Now update the driver contexts with the new shared descriptor */
-+ if (ctx->drv_ctx[ENCRYPT]) {
-+ ret = caam_drv_ctx_update(ctx->drv_ctx[ENCRYPT],
-+ ctx->sh_desc_enc);
-+ if (ret) {
-+ dev_err(jrdev, "driver enc context update failed\n");
-+ goto badkey;
-+ }
-+ }
-+
-+ if (ctx->drv_ctx[DECRYPT]) {
-+ ret = caam_drv_ctx_update(ctx->drv_ctx[DECRYPT],
-+ ctx->sh_desc_dec);
-+ if (ret) {
-+ dev_err(jrdev, "driver dec context update failed\n");
-+ goto badkey;
-+ }
-+ }
-+
-+ return ret;
-+badkey:
-+ crypto_aead_set_flags(tls, CRYPTO_TFM_RES_BAD_KEY_LEN);
-+ return -EINVAL;
-+}
-+
-+static int gcm_set_sh_desc(struct crypto_aead *aead)
-+{
-+ struct caam_ctx *ctx = crypto_aead_ctx(aead);
-+ unsigned int ivsize = crypto_aead_ivsize(aead);
-+ int rem_bytes = CAAM_DESC_BYTES_MAX - DESC_JOB_IO_LEN -
-+ ctx->cdata.keylen;
-+
-+ if (!ctx->cdata.keylen || !ctx->authsize)
-+ return 0;
-+
-+ /*
-+ * Job Descriptor and Shared Descriptor
-+ * must fit into the 64-word Descriptor h/w Buffer
-+ */
-+ if (rem_bytes >= DESC_QI_GCM_ENC_LEN) {
-+ ctx->cdata.key_inline = true;
-+ ctx->cdata.key_virt = ctx->key;
-+ } else {
-+ ctx->cdata.key_inline = false;
-+ ctx->cdata.key_dma = ctx->key_dma;
-+ }
-+
-+ cnstr_shdsc_gcm_encap(ctx->sh_desc_enc, &ctx->cdata, ivsize,
-+ ctx->authsize, true);
-+
-+ /*
-+ * Job Descriptor and Shared Descriptor
-+ * must fit into the 64-word Descriptor h/w Buffer
-+ */
-+ if (rem_bytes >= DESC_QI_GCM_DEC_LEN) {
-+ ctx->cdata.key_inline = true;
-+ ctx->cdata.key_virt = ctx->key;
-+ } else {
-+ ctx->cdata.key_inline = false;
-+ ctx->cdata.key_dma = ctx->key_dma;
-+ }
-+
-+ cnstr_shdsc_gcm_decap(ctx->sh_desc_dec, &ctx->cdata, ivsize,
-+ ctx->authsize, true);
-+
-+ return 0;
-+}
-+
-+static int gcm_setauthsize(struct crypto_aead *authenc, unsigned int authsize)
-+{
-+ struct caam_ctx *ctx = crypto_aead_ctx(authenc);
-+
-+ ctx->authsize = authsize;
-+ gcm_set_sh_desc(authenc);
-+
-+ return 0;
-+}
-+
-+static int gcm_setkey(struct crypto_aead *aead,
-+ const u8 *key, unsigned int keylen)
-+{
-+ struct caam_ctx *ctx = crypto_aead_ctx(aead);
-+ struct device *jrdev = ctx->jrdev;
-+ int ret;
-+
-+#ifdef DEBUG
-+ print_hex_dump(KERN_ERR, "key in @" __stringify(__LINE__)": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, key, keylen, 1);
-+#endif
-+
-+ memcpy(ctx->key, key, keylen);
-+ dma_sync_single_for_device(jrdev, ctx->key_dma, keylen, ctx->dir);
-+ ctx->cdata.keylen = keylen;
-+
-+ ret = gcm_set_sh_desc(aead);
-+ if (ret)
-+ return ret;
-+
-+ /* Now update the driver contexts with the new shared descriptor */
-+ if (ctx->drv_ctx[ENCRYPT]) {
-+ ret = caam_drv_ctx_update(ctx->drv_ctx[ENCRYPT],
-+ ctx->sh_desc_enc);
-+ if (ret) {
-+ dev_err(jrdev, "driver enc context update failed\n");
-+ return ret;
-+ }
-+ }
-+
-+ if (ctx->drv_ctx[DECRYPT]) {
-+ ret = caam_drv_ctx_update(ctx->drv_ctx[DECRYPT],
-+ ctx->sh_desc_dec);
-+ if (ret) {
-+ dev_err(jrdev, "driver dec context update failed\n");
-+ return ret;
-+ }
-+ }
-+
-+ return 0;
-+}
-+
-+static int rfc4106_set_sh_desc(struct crypto_aead *aead)
-+{
-+ struct caam_ctx *ctx = crypto_aead_ctx(aead);
-+ unsigned int ivsize = crypto_aead_ivsize(aead);
-+ int rem_bytes = CAAM_DESC_BYTES_MAX - DESC_JOB_IO_LEN -
-+ ctx->cdata.keylen;
-+
-+ if (!ctx->cdata.keylen || !ctx->authsize)
-+ return 0;
-+
-+ ctx->cdata.key_virt = ctx->key;
-+
-+ /*
-+ * Job Descriptor and Shared Descriptor
-+ * must fit into the 64-word Descriptor h/w Buffer
-+ */
-+ if (rem_bytes >= DESC_QI_RFC4106_ENC_LEN) {
-+ ctx->cdata.key_inline = true;
-+ } else {
-+ ctx->cdata.key_inline = false;
-+ ctx->cdata.key_dma = ctx->key_dma;
-+ }
-+
-+ cnstr_shdsc_rfc4106_encap(ctx->sh_desc_enc, &ctx->cdata, ivsize,
-+ ctx->authsize, true);
-+
-+ /*
-+ * Job Descriptor and Shared Descriptor
-+ * must fit into the 64-word Descriptor h/w Buffer
-+ */
-+ if (rem_bytes >= DESC_QI_RFC4106_DEC_LEN) {
-+ ctx->cdata.key_inline = true;
-+ } else {
-+ ctx->cdata.key_inline = false;
-+ ctx->cdata.key_dma = ctx->key_dma;
-+ }
-+
-+ cnstr_shdsc_rfc4106_decap(ctx->sh_desc_dec, &ctx->cdata, ivsize,
-+ ctx->authsize, true);
-+
-+ return 0;
-+}
-+
-+static int rfc4106_setauthsize(struct crypto_aead *authenc,
-+ unsigned int authsize)
-+{
-+ struct caam_ctx *ctx = crypto_aead_ctx(authenc);
-+
-+ ctx->authsize = authsize;
-+ rfc4106_set_sh_desc(authenc);
-+
-+ return 0;
-+}
-+
-+static int rfc4106_setkey(struct crypto_aead *aead,
-+ const u8 *key, unsigned int keylen)
-+{
-+ struct caam_ctx *ctx = crypto_aead_ctx(aead);
-+ struct device *jrdev = ctx->jrdev;
-+ int ret;
-+
-+ if (keylen < 4)
-+ return -EINVAL;
-+
-+#ifdef DEBUG
-+ print_hex_dump(KERN_ERR, "key in @" __stringify(__LINE__)": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, key, keylen, 1);
-+#endif
-+
-+ memcpy(ctx->key, key, keylen);
-+ /*
-+ * The last four bytes of the key material are used as the salt value
-+ * in the nonce. Update the AES key length.
-+ */
-+ ctx->cdata.keylen = keylen - 4;
-+ dma_sync_single_for_device(jrdev, ctx->key_dma, ctx->cdata.keylen,
-+ ctx->dir);
-+
-+ ret = rfc4106_set_sh_desc(aead);
-+ if (ret)
-+ return ret;
-+
-+ /* Now update the driver contexts with the new shared descriptor */
-+ if (ctx->drv_ctx[ENCRYPT]) {
-+ ret = caam_drv_ctx_update(ctx->drv_ctx[ENCRYPT],
-+ ctx->sh_desc_enc);
-+ if (ret) {
-+ dev_err(jrdev, "driver enc context update failed\n");
-+ return ret;
-+ }
-+ }
-+
-+ if (ctx->drv_ctx[DECRYPT]) {
-+ ret = caam_drv_ctx_update(ctx->drv_ctx[DECRYPT],
-+ ctx->sh_desc_dec);
-+ if (ret) {
-+ dev_err(jrdev, "driver dec context update failed\n");
-+ return ret;
-+ }
-+ }
-+
-+ return 0;
-+}
-+
-+static int rfc4543_set_sh_desc(struct crypto_aead *aead)
-+{
-+ struct caam_ctx *ctx = crypto_aead_ctx(aead);
-+ unsigned int ivsize = crypto_aead_ivsize(aead);
-+ int rem_bytes = CAAM_DESC_BYTES_MAX - DESC_JOB_IO_LEN -
-+ ctx->cdata.keylen;
-+
-+ if (!ctx->cdata.keylen || !ctx->authsize)
-+ return 0;
-+
-+ ctx->cdata.key_virt = ctx->key;
-+
-+ /*
-+ * Job Descriptor and Shared Descriptor
-+ * must fit into the 64-word Descriptor h/w Buffer
-+ */
-+ if (rem_bytes >= DESC_QI_RFC4543_ENC_LEN) {
-+ ctx->cdata.key_inline = true;
-+ } else {
-+ ctx->cdata.key_inline = false;
-+ ctx->cdata.key_dma = ctx->key_dma;
-+ }
-+
-+ cnstr_shdsc_rfc4543_encap(ctx->sh_desc_enc, &ctx->cdata, ivsize,
-+ ctx->authsize, true);
-+
-+ /*
-+ * Job Descriptor and Shared Descriptor
-+ * must fit into the 64-word Descriptor h/w Buffer
-+ */
-+ if (rem_bytes >= DESC_QI_RFC4543_DEC_LEN) {
-+ ctx->cdata.key_inline = true;
-+ } else {
-+ ctx->cdata.key_inline = false;
-+ ctx->cdata.key_dma = ctx->key_dma;
-+ }
-+
-+ cnstr_shdsc_rfc4543_decap(ctx->sh_desc_dec, &ctx->cdata, ivsize,
-+ ctx->authsize, true);
-+
-+ return 0;
-+}
-+
-+static int rfc4543_setauthsize(struct crypto_aead *authenc,
-+ unsigned int authsize)
-+{
-+ struct caam_ctx *ctx = crypto_aead_ctx(authenc);
-+
-+ ctx->authsize = authsize;
-+ rfc4543_set_sh_desc(authenc);
-+
-+ return 0;
-+}
-+
-+static int rfc4543_setkey(struct crypto_aead *aead,
-+ const u8 *key, unsigned int keylen)
-+{
-+ struct caam_ctx *ctx = crypto_aead_ctx(aead);
-+ struct device *jrdev = ctx->jrdev;
-+ int ret;
-+
-+ if (keylen < 4)
-+ return -EINVAL;
-+
-+#ifdef DEBUG
-+ print_hex_dump(KERN_ERR, "key in @" __stringify(__LINE__)": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, key, keylen, 1);
-+#endif
-+
-+ memcpy(ctx->key, key, keylen);
-+ /*
-+ * The last four bytes of the key material are used as the salt value
-+ * in the nonce. Update the AES key length.
-+ */
-+ ctx->cdata.keylen = keylen - 4;
-+ dma_sync_single_for_device(jrdev, ctx->key_dma, ctx->cdata.keylen,
-+ ctx->dir);
-+
-+ ret = rfc4543_set_sh_desc(aead);
-+ if (ret)
-+ return ret;
-+
-+ /* Now update the driver contexts with the new shared descriptor */
-+ if (ctx->drv_ctx[ENCRYPT]) {
-+ ret = caam_drv_ctx_update(ctx->drv_ctx[ENCRYPT],
-+ ctx->sh_desc_enc);
-+ if (ret) {
-+ dev_err(jrdev, "driver enc context update failed\n");
-+ return ret;
-+ }
-+ }
-+
-+ if (ctx->drv_ctx[DECRYPT]) {
-+ ret = caam_drv_ctx_update(ctx->drv_ctx[DECRYPT],
-+ ctx->sh_desc_dec);
-+ if (ret) {
-+ dev_err(jrdev, "driver dec context update failed\n");
-+ return ret;
-+ }
-+ }
-+
-+ return 0;
-+}
-+
-+static int ablkcipher_setkey(struct crypto_ablkcipher *ablkcipher,
-+ const u8 *key, unsigned int keylen)
-+{
-+ struct caam_ctx *ctx = crypto_ablkcipher_ctx(ablkcipher);
-+ struct crypto_tfm *tfm = crypto_ablkcipher_tfm(ablkcipher);
-+ const char *alg_name = crypto_tfm_alg_name(tfm);
-+ struct device *jrdev = ctx->jrdev;
-+ unsigned int ivsize = crypto_ablkcipher_ivsize(ablkcipher);
-+ u32 ctx1_iv_off = 0;
-+ const bool ctr_mode = ((ctx->cdata.algtype & OP_ALG_AAI_MASK) ==
-+ OP_ALG_AAI_CTR_MOD128);
-+ const bool is_rfc3686 = (ctr_mode && strstr(alg_name, "rfc3686"));
-+ int ret = 0;
-+
-+ memcpy(ctx->key, key, keylen);
-+#ifdef DEBUG
-+ print_hex_dump(KERN_ERR, "key in @" __stringify(__LINE__)": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, key, keylen, 1);
-+#endif
-+ /*
-+ * AES-CTR needs to load IV in CONTEXT1 reg
-+ * at an offset of 128bits (16bytes)
-+ * CONTEXT1[255:128] = IV
-+ */
-+ if (ctr_mode)
-+ ctx1_iv_off = 16;
-+
-+ /*
-+ * RFC3686 specific:
-+ * | CONTEXT1[255:128] = {NONCE, IV, COUNTER}
-+ * | *key = {KEY, NONCE}
-+ */
-+ if (is_rfc3686) {
-+ ctx1_iv_off = 16 + CTR_RFC3686_NONCE_SIZE;
-+ keylen -= CTR_RFC3686_NONCE_SIZE;
-+ }
-+
-+ dma_sync_single_for_device(jrdev, ctx->key_dma, keylen, DMA_TO_DEVICE);
-+ ctx->cdata.keylen = keylen;
-+ ctx->cdata.key_virt = ctx->key;
-+ ctx->cdata.key_inline = true;
-+
-+ /* ablkcipher encrypt, decrypt, givencrypt shared descriptors */
-+ cnstr_shdsc_ablkcipher_encap(ctx->sh_desc_enc, &ctx->cdata, ivsize,
-+ is_rfc3686, ctx1_iv_off);
-+ cnstr_shdsc_ablkcipher_decap(ctx->sh_desc_dec, &ctx->cdata, ivsize,
-+ is_rfc3686, ctx1_iv_off);
-+ cnstr_shdsc_ablkcipher_givencap(ctx->sh_desc_givenc, &ctx->cdata,
-+ ivsize, is_rfc3686, ctx1_iv_off);
-+
-+ /* Now update the driver contexts with the new shared descriptor */
-+ if (ctx->drv_ctx[ENCRYPT]) {
-+ ret = caam_drv_ctx_update(ctx->drv_ctx[ENCRYPT],
-+ ctx->sh_desc_enc);
-+ if (ret) {
-+ dev_err(jrdev, "driver enc context update failed\n");
-+ goto badkey;
-+ }
-+ }
-+
-+ if (ctx->drv_ctx[DECRYPT]) {
-+ ret = caam_drv_ctx_update(ctx->drv_ctx[DECRYPT],
-+ ctx->sh_desc_dec);
-+ if (ret) {
-+ dev_err(jrdev, "driver dec context update failed\n");
-+ goto badkey;
-+ }
-+ }
-+
-+ if (ctx->drv_ctx[GIVENCRYPT]) {
-+ ret = caam_drv_ctx_update(ctx->drv_ctx[GIVENCRYPT],
-+ ctx->sh_desc_givenc);
-+ if (ret) {
-+ dev_err(jrdev, "driver givenc context update failed\n");
-+ goto badkey;
-+ }
-+ }
-+
-+ return ret;
-+badkey:
-+ crypto_ablkcipher_set_flags(ablkcipher, CRYPTO_TFM_RES_BAD_KEY_LEN);
-+ return -EINVAL;
-+}
-+
-+static int xts_ablkcipher_setkey(struct crypto_ablkcipher *ablkcipher,
-+ const u8 *key, unsigned int keylen)
-+{
-+ struct caam_ctx *ctx = crypto_ablkcipher_ctx(ablkcipher);
-+ struct device *jrdev = ctx->jrdev;
-+ int ret = 0;
-+
-+ if (keylen != 2 * AES_MIN_KEY_SIZE && keylen != 2 * AES_MAX_KEY_SIZE) {
-+ crypto_ablkcipher_set_flags(ablkcipher,
-+ CRYPTO_TFM_RES_BAD_KEY_LEN);
-+ dev_err(jrdev, "key size mismatch\n");
-+ return -EINVAL;
-+ }
-+
-+ memcpy(ctx->key, key, keylen);
-+ dma_sync_single_for_device(jrdev, ctx->key_dma, keylen, DMA_TO_DEVICE);
-+ ctx->cdata.keylen = keylen;
-+ ctx->cdata.key_virt = ctx->key;
-+ ctx->cdata.key_inline = true;
-+
-+ /* xts ablkcipher encrypt, decrypt shared descriptors */
-+ cnstr_shdsc_xts_ablkcipher_encap(ctx->sh_desc_enc, &ctx->cdata);
-+ cnstr_shdsc_xts_ablkcipher_decap(ctx->sh_desc_dec, &ctx->cdata);
-+
-+ /* Now update the driver contexts with the new shared descriptor */
-+ if (ctx->drv_ctx[ENCRYPT]) {
-+ ret = caam_drv_ctx_update(ctx->drv_ctx[ENCRYPT],
-+ ctx->sh_desc_enc);
-+ if (ret) {
-+ dev_err(jrdev, "driver enc context update failed\n");
-+ goto badkey;
-+ }
-+ }
-+
-+ if (ctx->drv_ctx[DECRYPT]) {
-+ ret = caam_drv_ctx_update(ctx->drv_ctx[DECRYPT],
-+ ctx->sh_desc_dec);
-+ if (ret) {
-+ dev_err(jrdev, "driver dec context update failed\n");
-+ goto badkey;
-+ }
-+ }
-+
-+ return ret;
-+badkey:
-+ crypto_ablkcipher_set_flags(ablkcipher, CRYPTO_TFM_RES_BAD_KEY_LEN);
-+ return 0;
-+}
-+
-+/*
-+ * aead_edesc - s/w-extended aead descriptor
-+ * @src_nents: number of segments in input scatterlist
-+ * @dst_nents: number of segments in output scatterlist
-+ * @iv_dma: dma address of iv for checking continuity and link table
-+ * @qm_sg_bytes: length of dma mapped h/w link table
-+ * @qm_sg_dma: bus physical mapped address of h/w link table
-+ * @assoclen: associated data length, in CAAM endianness
-+ * @assoclen_dma: bus physical mapped address of req->assoclen
-+ * @drv_req: driver-specific request structure
-+ * @sgt: the h/w link table
-+ */
-+struct aead_edesc {
-+ int src_nents;
-+ int dst_nents;
-+ dma_addr_t iv_dma;
-+ int qm_sg_bytes;
-+ dma_addr_t qm_sg_dma;
-+ unsigned int assoclen;
-+ dma_addr_t assoclen_dma;
-+ struct caam_drv_req drv_req;
-+#define CAAM_QI_MAX_AEAD_SG \
-+ ((CAAM_QI_MEMCACHE_SIZE - offsetof(struct aead_edesc, sgt)) / \
-+ sizeof(struct qm_sg_entry))
-+ struct qm_sg_entry sgt[0];
-+};
-+
-+/*
-+ * tls_edesc - s/w-extended tls descriptor
-+ * @src_nents: number of segments in input scatterlist
-+ * @dst_nents: number of segments in output scatterlist
-+ * @iv_dma: dma address of iv for checking continuity and link table
-+ * @qm_sg_bytes: length of dma mapped h/w link table
-+ * @tmp: array of scatterlists used by 'scatterwalk_ffwd'
-+ * @qm_sg_dma: bus physical mapped address of h/w link table
-+ * @drv_req: driver-specific request structure
-+ * @sgt: the h/w link table
-+ */
-+struct tls_edesc {
-+ int src_nents;
-+ int dst_nents;
-+ dma_addr_t iv_dma;
-+ int qm_sg_bytes;
-+ dma_addr_t qm_sg_dma;
-+ struct scatterlist tmp[2];
-+ struct scatterlist *dst;
-+ struct caam_drv_req drv_req;
-+ struct qm_sg_entry sgt[0];
-+};
-+
-+/*
-+ * ablkcipher_edesc - s/w-extended ablkcipher descriptor
-+ * @src_nents: number of segments in input scatterlist
-+ * @dst_nents: number of segments in output scatterlist
-+ * @iv_dma: dma address of iv for checking continuity and link table
-+ * @qm_sg_bytes: length of dma mapped h/w link table
-+ * @qm_sg_dma: bus physical mapped address of h/w link table
-+ * @drv_req: driver-specific request structure
-+ * @sgt: the h/w link table
-+ */
-+struct ablkcipher_edesc {
-+ int src_nents;
-+ int dst_nents;
-+ dma_addr_t iv_dma;
-+ int qm_sg_bytes;
-+ dma_addr_t qm_sg_dma;
-+ struct caam_drv_req drv_req;
-+#define CAAM_QI_MAX_ABLKCIPHER_SG \
-+ ((CAAM_QI_MEMCACHE_SIZE - offsetof(struct ablkcipher_edesc, sgt)) / \
-+ sizeof(struct qm_sg_entry))
-+ struct qm_sg_entry sgt[0];
-+};
-+
-+static struct caam_drv_ctx *get_drv_ctx(struct caam_ctx *ctx,
-+ enum optype type)
-+{
-+ /*
-+ * This function is called on the fast path with values of 'type'
-+ * known at compile time. Invalid arguments are not expected and
-+ * thus no checks are made.
-+ */
-+ struct caam_drv_ctx *drv_ctx = ctx->drv_ctx[type];
-+ u32 *desc;
-+
-+ if (unlikely(!drv_ctx)) {
-+ spin_lock(&ctx->lock);
-+
-+ /* Read again to check if some other core init drv_ctx */
-+ drv_ctx = ctx->drv_ctx[type];
-+ if (!drv_ctx) {
-+ int cpu;
-+
-+ if (type == ENCRYPT)
-+ desc = ctx->sh_desc_enc;
-+ else if (type == DECRYPT)
-+ desc = ctx->sh_desc_dec;
-+ else /* (type == GIVENCRYPT) */
-+ desc = ctx->sh_desc_givenc;
-+
-+ cpu = smp_processor_id();
-+ drv_ctx = caam_drv_ctx_init(ctx->qidev, &cpu, desc);
-+ if (likely(!IS_ERR_OR_NULL(drv_ctx)))
-+ drv_ctx->op_type = type;
-+
-+ ctx->drv_ctx[type] = drv_ctx;
-+ }
-+
-+ spin_unlock(&ctx->lock);
-+ }
-+
-+ return drv_ctx;
-+}
-+
-+static void caam_unmap(struct device *dev, struct scatterlist *src,
-+ struct scatterlist *dst, int src_nents,
-+ int dst_nents, dma_addr_t iv_dma, int ivsize,
-+ enum optype op_type, dma_addr_t qm_sg_dma,
-+ int qm_sg_bytes)
-+{
-+ if (dst != src) {
-+ if (src_nents)
-+ dma_unmap_sg(dev, src, src_nents, DMA_TO_DEVICE);
-+ dma_unmap_sg(dev, dst, dst_nents, DMA_FROM_DEVICE);
-+ } else {
-+ dma_unmap_sg(dev, src, src_nents, DMA_BIDIRECTIONAL);
-+ }
-+
-+ if (iv_dma)
-+ dma_unmap_single(dev, iv_dma, ivsize,
-+ op_type == GIVENCRYPT ? DMA_FROM_DEVICE :
-+ DMA_TO_DEVICE);
-+ if (qm_sg_bytes)
-+ dma_unmap_single(dev, qm_sg_dma, qm_sg_bytes, DMA_TO_DEVICE);
-+}
-+
-+static void aead_unmap(struct device *dev,
-+ struct aead_edesc *edesc,
-+ struct aead_request *req)
-+{
-+ struct crypto_aead *aead = crypto_aead_reqtfm(req);
-+ int ivsize = crypto_aead_ivsize(aead);
-+
-+ caam_unmap(dev, req->src, req->dst, edesc->src_nents, edesc->dst_nents,
-+ edesc->iv_dma, ivsize, edesc->drv_req.drv_ctx->op_type,
-+ edesc->qm_sg_dma, edesc->qm_sg_bytes);
-+ dma_unmap_single(dev, edesc->assoclen_dma, 4, DMA_TO_DEVICE);
-+}
-+
-+static void tls_unmap(struct device *dev,
-+ struct tls_edesc *edesc,
-+ struct aead_request *req)
-+{
-+ struct crypto_aead *aead = crypto_aead_reqtfm(req);
-+ int ivsize = crypto_aead_ivsize(aead);
-+
-+ caam_unmap(dev, req->src, edesc->dst, edesc->src_nents,
-+ edesc->dst_nents, edesc->iv_dma, ivsize,
-+ edesc->drv_req.drv_ctx->op_type, edesc->qm_sg_dma,
-+ edesc->qm_sg_bytes);
-+}
-+
-+static void ablkcipher_unmap(struct device *dev,
-+ struct ablkcipher_edesc *edesc,
-+ struct ablkcipher_request *req)
-+{
-+ struct crypto_ablkcipher *ablkcipher = crypto_ablkcipher_reqtfm(req);
-+ int ivsize = crypto_ablkcipher_ivsize(ablkcipher);
-+
-+ caam_unmap(dev, req->src, req->dst, edesc->src_nents, edesc->dst_nents,
-+ edesc->iv_dma, ivsize, edesc->drv_req.drv_ctx->op_type,
-+ edesc->qm_sg_dma, edesc->qm_sg_bytes);
-+}
-+
-+static void aead_done(struct caam_drv_req *drv_req, u32 status)
-+{
-+ struct device *qidev;
-+ struct aead_edesc *edesc;
-+ struct aead_request *aead_req = drv_req->app_ctx;
-+ struct crypto_aead *aead = crypto_aead_reqtfm(aead_req);
-+ struct caam_ctx *caam_ctx = crypto_aead_ctx(aead);
-+ int ecode = 0;
-+
-+ qidev = caam_ctx->qidev;
-+
-+ if (unlikely(status)) {
-+ u32 ssrc = status & JRSTA_SSRC_MASK;
-+ u8 err_id = status & JRSTA_CCBERR_ERRID_MASK;
-+
-+ caam_jr_strstatus(qidev, status);
-+ /*
-+ * verify hw auth check passed else return -EBADMSG
-+ */
-+ if (ssrc == JRSTA_SSRC_CCB_ERROR &&
-+ err_id == JRSTA_CCBERR_ERRID_ICVCHK)
-+ ecode = -EBADMSG;
-+ else
-+ ecode = -EIO;
-+ }
-+
-+ edesc = container_of(drv_req, typeof(*edesc), drv_req);
-+ aead_unmap(qidev, edesc, aead_req);
-+
-+ aead_request_complete(aead_req, ecode);
-+ qi_cache_free(edesc);
-+}
-+
-+/*
-+ * allocate and map the aead extended descriptor
-+ */
-+static struct aead_edesc *aead_edesc_alloc(struct aead_request *req,
-+ bool encrypt)
-+{
-+ struct crypto_aead *aead = crypto_aead_reqtfm(req);
-+ struct caam_ctx *ctx = crypto_aead_ctx(aead);
-+ struct caam_aead_alg *alg = container_of(crypto_aead_alg(aead),
-+ typeof(*alg), aead);
-+ struct device *qidev = ctx->qidev;
-+ gfp_t flags = (req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP) ?
-+ GFP_KERNEL : GFP_ATOMIC;
-+ int src_nents, mapped_src_nents, dst_nents = 0, mapped_dst_nents = 0;
-+ struct aead_edesc *edesc;
-+ dma_addr_t qm_sg_dma, iv_dma = 0;
-+ int ivsize = 0;
-+ unsigned int authsize = ctx->authsize;
-+ int qm_sg_index = 0, qm_sg_ents = 0, qm_sg_bytes;
-+ int in_len, out_len;
-+ struct qm_sg_entry *sg_table, *fd_sgt;
-+ struct caam_drv_ctx *drv_ctx;
-+ enum optype op_type = encrypt ? ENCRYPT : DECRYPT;
-+
-+ drv_ctx = get_drv_ctx(ctx, op_type);
-+ if (unlikely(IS_ERR_OR_NULL(drv_ctx)))
-+ return (struct aead_edesc *)drv_ctx;
-+
-+ /* allocate space for base edesc and hw desc commands, link tables */
-+ edesc = qi_cache_alloc(GFP_DMA | flags);
-+ if (unlikely(!edesc)) {
-+ dev_err(qidev, "could not allocate extended descriptor\n");
-+ return ERR_PTR(-ENOMEM);
-+ }
-+
-+ if (likely(req->src == req->dst)) {
-+ src_nents = sg_nents_for_len(req->src, req->assoclen +
-+ req->cryptlen +
-+ (encrypt ? authsize : 0));
-+ if (unlikely(src_nents < 0)) {
-+ dev_err(qidev, "Insufficient bytes (%d) in src S/G\n",
-+ req->assoclen + req->cryptlen +
-+ (encrypt ? authsize : 0));
-+ qi_cache_free(edesc);
-+ return ERR_PTR(src_nents);
-+ }
-+
-+ mapped_src_nents = dma_map_sg(qidev, req->src, src_nents,
-+ DMA_BIDIRECTIONAL);
-+ if (unlikely(!mapped_src_nents)) {
-+ dev_err(qidev, "unable to map source\n");
-+ qi_cache_free(edesc);
-+ return ERR_PTR(-ENOMEM);
-+ }
-+ } else {
-+ src_nents = sg_nents_for_len(req->src, req->assoclen +
-+ req->cryptlen);
-+ if (unlikely(src_nents < 0)) {
-+ dev_err(qidev, "Insufficient bytes (%d) in src S/G\n",
-+ req->assoclen + req->cryptlen);
-+ qi_cache_free(edesc);
-+ return ERR_PTR(src_nents);
-+ }
-+
-+ dst_nents = sg_nents_for_len(req->dst, req->assoclen +
-+ req->cryptlen +
-+ (encrypt ? authsize :
-+ (-authsize)));
-+ if (unlikely(dst_nents < 0)) {
-+ dev_err(qidev, "Insufficient bytes (%d) in dst S/G\n",
-+ req->assoclen + req->cryptlen +
-+ (encrypt ? authsize : (-authsize)));
-+ qi_cache_free(edesc);
-+ return ERR_PTR(dst_nents);
-+ }
-+
-+ if (src_nents) {
-+ mapped_src_nents = dma_map_sg(qidev, req->src,
-+ src_nents, DMA_TO_DEVICE);
-+ if (unlikely(!mapped_src_nents)) {
-+ dev_err(qidev, "unable to map source\n");
-+ qi_cache_free(edesc);
-+ return ERR_PTR(-ENOMEM);
-+ }
-+ } else {
-+ mapped_src_nents = 0;
-+ }
-+
-+ mapped_dst_nents = dma_map_sg(qidev, req->dst, dst_nents,
-+ DMA_FROM_DEVICE);
-+ if (unlikely(!mapped_dst_nents)) {
-+ dev_err(qidev, "unable to map destination\n");
-+ dma_unmap_sg(qidev, req->src, src_nents, DMA_TO_DEVICE);
-+ qi_cache_free(edesc);
-+ return ERR_PTR(-ENOMEM);
-+ }
-+ }
-+
-+ if ((alg->caam.rfc3686 && encrypt) || !alg->caam.geniv) {
-+ ivsize = crypto_aead_ivsize(aead);
-+ iv_dma = dma_map_single(qidev, req->iv, ivsize, DMA_TO_DEVICE);
-+ if (dma_mapping_error(qidev, iv_dma)) {
-+ dev_err(qidev, "unable to map IV\n");
-+ caam_unmap(qidev, req->src, req->dst, src_nents,
-+ dst_nents, 0, 0, op_type, 0, 0);
-+ qi_cache_free(edesc);
-+ return ERR_PTR(-ENOMEM);
-+ }
-+ }
-+
-+ /*
-+ * Create S/G table: req->assoclen, [IV,] req->src [, req->dst].
-+ * Input is not contiguous.
-+ */
-+ qm_sg_ents = 1 + !!ivsize + mapped_src_nents +
-+ (mapped_dst_nents > 1 ? mapped_dst_nents : 0);
-+ if (unlikely(qm_sg_ents > CAAM_QI_MAX_AEAD_SG)) {
-+ dev_err(qidev, "Insufficient S/G entries: %d > %lu\n",
-+ qm_sg_ents, CAAM_QI_MAX_AEAD_SG);
-+ caam_unmap(qidev, req->src, req->dst, src_nents, dst_nents,
-+ iv_dma, ivsize, op_type, 0, 0);
-+ qi_cache_free(edesc);
-+ return ERR_PTR(-ENOMEM);
-+ }
-+ sg_table = &edesc->sgt[0];
-+ qm_sg_bytes = qm_sg_ents * sizeof(*sg_table);
-+
-+ edesc->src_nents = src_nents;
-+ edesc->dst_nents = dst_nents;
-+ edesc->iv_dma = iv_dma;
-+ edesc->drv_req.app_ctx = req;
-+ edesc->drv_req.cbk = aead_done;
-+ edesc->drv_req.drv_ctx = drv_ctx;
-+
-+ edesc->assoclen = cpu_to_caam32(req->assoclen);
-+ edesc->assoclen_dma = dma_map_single(qidev, &edesc->assoclen, 4,
-+ DMA_TO_DEVICE);
-+ if (dma_mapping_error(qidev, edesc->assoclen_dma)) {
-+ dev_err(qidev, "unable to map assoclen\n");
-+ caam_unmap(qidev, req->src, req->dst, src_nents, dst_nents,
-+ iv_dma, ivsize, op_type, 0, 0);
-+ qi_cache_free(edesc);
-+ return ERR_PTR(-ENOMEM);
-+ }
-+
-+ dma_to_qm_sg_one(sg_table, edesc->assoclen_dma, 4, 0);
-+ qm_sg_index++;
-+ if (ivsize) {
-+ dma_to_qm_sg_one(sg_table + qm_sg_index, iv_dma, ivsize, 0);
-+ qm_sg_index++;
-+ }
-+ sg_to_qm_sg_last(req->src, mapped_src_nents, sg_table + qm_sg_index, 0);
-+ qm_sg_index += mapped_src_nents;
-+
-+ if (mapped_dst_nents > 1)
-+ sg_to_qm_sg_last(req->dst, mapped_dst_nents, sg_table +
-+ qm_sg_index, 0);
-+
-+ qm_sg_dma = dma_map_single(qidev, sg_table, qm_sg_bytes, DMA_TO_DEVICE);
-+ if (dma_mapping_error(qidev, qm_sg_dma)) {
-+ dev_err(qidev, "unable to map S/G table\n");
-+ dma_unmap_single(qidev, edesc->assoclen_dma, 4, DMA_TO_DEVICE);
-+ caam_unmap(qidev, req->src, req->dst, src_nents, dst_nents,
-+ iv_dma, ivsize, op_type, 0, 0);
-+ qi_cache_free(edesc);
-+ return ERR_PTR(-ENOMEM);
-+ }
-+
-+ edesc->qm_sg_dma = qm_sg_dma;
-+ edesc->qm_sg_bytes = qm_sg_bytes;
-+
-+ out_len = req->assoclen + req->cryptlen +
-+ (encrypt ? ctx->authsize : (-ctx->authsize));
-+ in_len = 4 + ivsize + req->assoclen + req->cryptlen;
-+
-+ fd_sgt = &edesc->drv_req.fd_sgt[0];
-+ dma_to_qm_sg_one_last_ext(&fd_sgt[1], qm_sg_dma, in_len, 0);
-+
-+ if (req->dst == req->src) {
-+ if (mapped_src_nents == 1)
-+ dma_to_qm_sg_one(&fd_sgt[0], sg_dma_address(req->src),
-+ out_len, 0);
-+ else
-+ dma_to_qm_sg_one_ext(&fd_sgt[0], qm_sg_dma +
-+ (1 + !!ivsize) * sizeof(*sg_table),
-+ out_len, 0);
-+ } else if (mapped_dst_nents == 1) {
-+ dma_to_qm_sg_one(&fd_sgt[0], sg_dma_address(req->dst), out_len,
-+ 0);
-+ } else {
-+ dma_to_qm_sg_one_ext(&fd_sgt[0], qm_sg_dma + sizeof(*sg_table) *
-+ qm_sg_index, out_len, 0);
-+ }
-+
-+ return edesc;
-+}
-+
-+static inline int aead_crypt(struct aead_request *req, bool encrypt)
-+{
-+ struct aead_edesc *edesc;
-+ struct crypto_aead *aead = crypto_aead_reqtfm(req);
-+ struct caam_ctx *ctx = crypto_aead_ctx(aead);
-+ int ret;
-+
-+ if (unlikely(caam_congested))
-+ return -EAGAIN;
-+
-+ /* allocate extended descriptor */
-+ edesc = aead_edesc_alloc(req, encrypt);
-+ if (IS_ERR_OR_NULL(edesc))
-+ return PTR_ERR(edesc);
-+
-+ /* Create and submit job descriptor */
-+ ret = caam_qi_enqueue(ctx->qidev, &edesc->drv_req);
-+ if (!ret) {
-+ ret = -EINPROGRESS;
-+ } else {
-+ aead_unmap(ctx->qidev, edesc, req);
-+ qi_cache_free(edesc);
-+ }
-+
-+ return ret;
-+}
-+
-+static int aead_encrypt(struct aead_request *req)
-+{
-+ return aead_crypt(req, true);
-+}
-+
-+static int aead_decrypt(struct aead_request *req)
-+{
-+ return aead_crypt(req, false);
-+}
-+
-+static int ipsec_gcm_encrypt(struct aead_request *req)
-+{
-+ if (req->assoclen < 8)
-+ return -EINVAL;
-+
-+ return aead_crypt(req, true);
-+}
-+
-+static int ipsec_gcm_decrypt(struct aead_request *req)
-+{
-+ if (req->assoclen < 8)
-+ return -EINVAL;
-+
-+ return aead_crypt(req, false);
-+}
-+
-+static void tls_done(struct caam_drv_req *drv_req, u32 status)
-+{
-+ struct device *qidev;
-+ struct tls_edesc *edesc;
-+ struct aead_request *aead_req = drv_req->app_ctx;
-+ struct crypto_aead *aead = crypto_aead_reqtfm(aead_req);
-+ struct caam_ctx *caam_ctx = crypto_aead_ctx(aead);
-+ int ecode = 0;
-+
-+ qidev = caam_ctx->qidev;
-+
-+ if (unlikely(status)) {
-+ caam_jr_strstatus(qidev, status);
-+ ecode = -EIO;
-+ }
-+
-+ edesc = container_of(drv_req, typeof(*edesc), drv_req);
-+ tls_unmap(qidev, edesc, aead_req);
-+
-+ aead_request_complete(aead_req, ecode);
-+ qi_cache_free(edesc);
-+}
-+
-+/*
-+ * allocate and map the tls extended descriptor
-+ */
-+static struct tls_edesc *tls_edesc_alloc(struct aead_request *req, bool encrypt)
-+{
-+ struct crypto_aead *aead = crypto_aead_reqtfm(req);
-+ struct caam_ctx *ctx = crypto_aead_ctx(aead);
-+ unsigned int blocksize = crypto_aead_blocksize(aead);
-+ unsigned int padsize, authsize;
-+ struct caam_aead_alg *alg = container_of(crypto_aead_alg(aead),
-+ typeof(*alg), aead);
-+ struct device *qidev = ctx->qidev;
-+ gfp_t flags = (req->base.flags & (CRYPTO_TFM_REQ_MAY_BACKLOG |
-+ CRYPTO_TFM_REQ_MAY_SLEEP)) ? GFP_KERNEL : GFP_ATOMIC;
-+ int src_nents, mapped_src_nents, dst_nents = 0, mapped_dst_nents = 0;
-+ struct tls_edesc *edesc;
-+ dma_addr_t qm_sg_dma, iv_dma = 0;
-+ int ivsize = 0;
-+ int qm_sg_index, qm_sg_ents = 0, qm_sg_bytes;
-+ int in_len, out_len;
-+ struct qm_sg_entry *sg_table, *fd_sgt;
-+ struct caam_drv_ctx *drv_ctx;
-+ enum optype op_type = encrypt ? ENCRYPT : DECRYPT;
-+ struct scatterlist *dst;
-+
-+ if (encrypt) {
-+ padsize = blocksize - ((req->cryptlen + ctx->authsize) %
-+ blocksize);
-+ authsize = ctx->authsize + padsize;
-+ } else {
-+ authsize = ctx->authsize;
-+ }
-+
-+ drv_ctx = get_drv_ctx(ctx, op_type);
-+ if (unlikely(IS_ERR_OR_NULL(drv_ctx)))
-+ return (struct tls_edesc *)drv_ctx;
-+
-+ /* allocate space for base edesc and hw desc commands, link tables */
-+ edesc = qi_cache_alloc(GFP_DMA | flags);
-+ if (unlikely(!edesc)) {
-+ dev_err(qidev, "could not allocate extended descriptor\n");
-+ return ERR_PTR(-ENOMEM);
-+ }
-+
-+ if (likely(req->src == req->dst)) {
-+ src_nents = sg_nents_for_len(req->src, req->assoclen +
-+ req->cryptlen +
-+ (encrypt ? authsize : 0));
-+ if (unlikely(src_nents < 0)) {
-+ dev_err(qidev, "Insufficient bytes (%d) in src S/G\n",
-+ req->assoclen + req->cryptlen +
-+ (encrypt ? authsize : 0));
-+ qi_cache_free(edesc);
-+ return ERR_PTR(src_nents);
-+ }
-+
-+ mapped_src_nents = dma_map_sg(qidev, req->src, src_nents,
-+ DMA_BIDIRECTIONAL);
-+ if (unlikely(!mapped_src_nents)) {
-+ dev_err(qidev, "unable to map source\n");
-+ qi_cache_free(edesc);
-+ return ERR_PTR(-ENOMEM);
-+ }
-+ dst = req->dst;
-+ } else {
-+ src_nents = sg_nents_for_len(req->src, req->assoclen +
-+ req->cryptlen);
-+ if (unlikely(src_nents < 0)) {
-+ dev_err(qidev, "Insufficient bytes (%d) in src S/G\n",
-+ req->assoclen + req->cryptlen);
-+ qi_cache_free(edesc);
-+ return ERR_PTR(src_nents);
-+ }
-+
-+ dst = scatterwalk_ffwd(edesc->tmp, req->dst, req->assoclen);
-+ dst_nents = sg_nents_for_len(dst, req->cryptlen +
-+ (encrypt ? authsize : 0));
-+ if (unlikely(dst_nents < 0)) {
-+ dev_err(qidev, "Insufficient bytes (%d) in dst S/G\n",
-+ req->cryptlen +
-+ (encrypt ? authsize : 0));
-+ qi_cache_free(edesc);
-+ return ERR_PTR(dst_nents);
-+ }
-+
-+ if (src_nents) {
-+ mapped_src_nents = dma_map_sg(qidev, req->src,
-+ src_nents, DMA_TO_DEVICE);
-+ if (unlikely(!mapped_src_nents)) {
-+ dev_err(qidev, "unable to map source\n");
-+ qi_cache_free(edesc);
-+ return ERR_PTR(-ENOMEM);
-+ }
-+ } else {
-+ mapped_src_nents = 0;
-+ }
-+
-+ mapped_dst_nents = dma_map_sg(qidev, dst, dst_nents,
-+ DMA_FROM_DEVICE);
-+ if (unlikely(!mapped_dst_nents)) {
-+ dev_err(qidev, "unable to map destination\n");
-+ dma_unmap_sg(qidev, req->src, src_nents, DMA_TO_DEVICE);
-+ qi_cache_free(edesc);
-+ return ERR_PTR(-ENOMEM);
-+ }
-+ }
-+
-+ ivsize = crypto_aead_ivsize(aead);
-+ iv_dma = dma_map_single(qidev, req->iv, ivsize, DMA_TO_DEVICE);
-+ if (dma_mapping_error(qidev, iv_dma)) {
-+ dev_err(qidev, "unable to map IV\n");
-+ caam_unmap(qidev, req->src, dst, src_nents, dst_nents, 0, 0,
-+ op_type, 0, 0);
-+ qi_cache_free(edesc);
-+ return ERR_PTR(-ENOMEM);
-+ }
-+
-+ /*
-+ * Create S/G table: IV, src, dst.
-+ * Input is not contiguous.
-+ */
-+ qm_sg_ents = 1 + mapped_src_nents +
-+ (mapped_dst_nents > 1 ? mapped_dst_nents : 0);
-+ sg_table = &edesc->sgt[0];
-+ qm_sg_bytes = qm_sg_ents * sizeof(*sg_table);
-+
-+ edesc->src_nents = src_nents;
-+ edesc->dst_nents = dst_nents;
-+ edesc->dst = dst;
-+ edesc->iv_dma = iv_dma;
-+ edesc->drv_req.app_ctx = req;
-+ edesc->drv_req.cbk = tls_done;
-+ edesc->drv_req.drv_ctx = drv_ctx;
-+
-+ dma_to_qm_sg_one(sg_table, iv_dma, ivsize, 0);
-+ qm_sg_index = 1;
-+
-+ sg_to_qm_sg_last(req->src, mapped_src_nents, sg_table + qm_sg_index, 0);
-+ qm_sg_index += mapped_src_nents;
-+
-+ if (mapped_dst_nents > 1)
-+ sg_to_qm_sg_last(dst, mapped_dst_nents, sg_table +
-+ qm_sg_index, 0);
-+
-+ qm_sg_dma = dma_map_single(qidev, sg_table, qm_sg_bytes, DMA_TO_DEVICE);
-+ if (dma_mapping_error(qidev, qm_sg_dma)) {
-+ dev_err(qidev, "unable to map S/G table\n");
-+ caam_unmap(qidev, req->src, dst, src_nents, dst_nents, iv_dma,
-+ ivsize, op_type, 0, 0);
-+ qi_cache_free(edesc);
-+ return ERR_PTR(-ENOMEM);
-+ }
-+
-+ edesc->qm_sg_dma = qm_sg_dma;
-+ edesc->qm_sg_bytes = qm_sg_bytes;
-+
-+ out_len = req->cryptlen + (encrypt ? authsize : 0);
-+ in_len = ivsize + req->assoclen + req->cryptlen;
-+
-+ fd_sgt = &edesc->drv_req.fd_sgt[0];
-+
-+ dma_to_qm_sg_one_last_ext(&fd_sgt[1], qm_sg_dma, in_len, 0);
-+
-+ if (req->dst == req->src)
-+ dma_to_qm_sg_one_ext(&fd_sgt[0], qm_sg_dma +
-+ (sg_nents_for_len(req->src, req->assoclen) +
-+ 1) * sizeof(*sg_table), out_len, 0);
-+ else if (mapped_dst_nents == 1)
-+ dma_to_qm_sg_one(&fd_sgt[0], sg_dma_address(dst), out_len, 0);
-+ else
-+ dma_to_qm_sg_one_ext(&fd_sgt[0], qm_sg_dma + sizeof(*sg_table) *
-+ qm_sg_index, out_len, 0);
-+
-+ return edesc;
-+}
-+
-+static int tls_crypt(struct aead_request *req, bool encrypt)
-+{
-+ struct tls_edesc *edesc;
-+ struct crypto_aead *aead = crypto_aead_reqtfm(req);
-+ struct caam_ctx *ctx = crypto_aead_ctx(aead);
-+ int ret;
-+
-+ if (unlikely(caam_congested))
-+ return -EAGAIN;
-+
-+ edesc = tls_edesc_alloc(req, encrypt);
-+ if (IS_ERR_OR_NULL(edesc))
-+ return PTR_ERR(edesc);
-+
-+ ret = caam_qi_enqueue(ctx->qidev, &edesc->drv_req);
-+ if (!ret) {
-+ ret = -EINPROGRESS;
-+ } else {
-+ tls_unmap(ctx->qidev, edesc, req);
-+ qi_cache_free(edesc);
-+ }
-+
-+ return ret;
-+}
-+
-+static int tls_encrypt(struct aead_request *req)
-+{
-+ return tls_crypt(req, true);
-+}
-+
-+static int tls_decrypt(struct aead_request *req)
-+{
-+ return tls_crypt(req, false);
-+}
-+
-+static void ablkcipher_done(struct caam_drv_req *drv_req, u32 status)
-+{
-+ struct ablkcipher_edesc *edesc;
-+ struct ablkcipher_request *req = drv_req->app_ctx;
-+ struct crypto_ablkcipher *ablkcipher = crypto_ablkcipher_reqtfm(req);
-+ struct caam_ctx *caam_ctx = crypto_ablkcipher_ctx(ablkcipher);
-+ struct device *qidev = caam_ctx->qidev;
-+ int ivsize = crypto_ablkcipher_ivsize(ablkcipher);
-+
-+#ifdef DEBUG
-+ dev_err(qidev, "%s %d: status 0x%x\n", __func__, __LINE__, status);
-+#endif
-+
-+ edesc = container_of(drv_req, typeof(*edesc), drv_req);
-+
-+ if (status)
-+ caam_jr_strstatus(qidev, status);
-+
-+#ifdef DEBUG
-+ print_hex_dump(KERN_ERR, "dstiv @" __stringify(__LINE__)": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, req->info,
-+ edesc->src_nents > 1 ? 100 : ivsize, 1);
-+ caam_dump_sg(KERN_ERR, "dst @" __stringify(__LINE__)": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, req->dst,
-+ edesc->dst_nents > 1 ? 100 : req->nbytes, 1);
-+#endif
-+
-+ ablkcipher_unmap(qidev, edesc, req);
-+ qi_cache_free(edesc);
-+
-+ /*
-+ * The crypto API expects us to set the IV (req->info) to the last
-+ * ciphertext block. This is used e.g. by the CTS mode.
-+ */
-+ scatterwalk_map_and_copy(req->info, req->dst, req->nbytes - ivsize,
-+ ivsize, 0);
-+
-+ ablkcipher_request_complete(req, status);
-+}
-+
-+static struct ablkcipher_edesc *ablkcipher_edesc_alloc(struct ablkcipher_request
-+ *req, bool encrypt)
-+{
-+ struct crypto_ablkcipher *ablkcipher = crypto_ablkcipher_reqtfm(req);
-+ struct caam_ctx *ctx = crypto_ablkcipher_ctx(ablkcipher);
-+ struct device *qidev = ctx->qidev;
-+ gfp_t flags = (req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP) ?
-+ GFP_KERNEL : GFP_ATOMIC;
-+ int src_nents, mapped_src_nents, dst_nents = 0, mapped_dst_nents = 0;
-+ struct ablkcipher_edesc *edesc;
-+ dma_addr_t iv_dma;
-+ bool in_contig;
-+ int ivsize = crypto_ablkcipher_ivsize(ablkcipher);
-+ int dst_sg_idx, qm_sg_ents;
-+ struct qm_sg_entry *sg_table, *fd_sgt;
-+ struct caam_drv_ctx *drv_ctx;
-+ enum optype op_type = encrypt ? ENCRYPT : DECRYPT;
-+
-+ drv_ctx = get_drv_ctx(ctx, op_type);
-+ if (unlikely(IS_ERR_OR_NULL(drv_ctx)))
-+ return (struct ablkcipher_edesc *)drv_ctx;
-+
-+ src_nents = sg_nents_for_len(req->src, req->nbytes);
-+ if (unlikely(src_nents < 0)) {
-+ dev_err(qidev, "Insufficient bytes (%d) in src S/G\n",
-+ req->nbytes);
-+ return ERR_PTR(src_nents);
-+ }
-+
-+ if (unlikely(req->src != req->dst)) {
-+ dst_nents = sg_nents_for_len(req->dst, req->nbytes);
-+ if (unlikely(dst_nents < 0)) {
-+ dev_err(qidev, "Insufficient bytes (%d) in dst S/G\n",
-+ req->nbytes);
-+ return ERR_PTR(dst_nents);
-+ }
-+
-+ mapped_src_nents = dma_map_sg(qidev, req->src, src_nents,
-+ DMA_TO_DEVICE);
-+ if (unlikely(!mapped_src_nents)) {
-+ dev_err(qidev, "unable to map source\n");
-+ return ERR_PTR(-ENOMEM);
-+ }
-+
-+ mapped_dst_nents = dma_map_sg(qidev, req->dst, dst_nents,
-+ DMA_FROM_DEVICE);
-+ if (unlikely(!mapped_dst_nents)) {
-+ dev_err(qidev, "unable to map destination\n");
-+ dma_unmap_sg(qidev, req->src, src_nents, DMA_TO_DEVICE);
-+ return ERR_PTR(-ENOMEM);
-+ }
-+ } else {
-+ mapped_src_nents = dma_map_sg(qidev, req->src, src_nents,
-+ DMA_BIDIRECTIONAL);
-+ if (unlikely(!mapped_src_nents)) {
-+ dev_err(qidev, "unable to map source\n");
-+ return ERR_PTR(-ENOMEM);
-+ }
-+ }
-+
-+ iv_dma = dma_map_single(qidev, req->info, ivsize, DMA_TO_DEVICE);
-+ if (dma_mapping_error(qidev, iv_dma)) {
-+ dev_err(qidev, "unable to map IV\n");
-+ caam_unmap(qidev, req->src, req->dst, src_nents, dst_nents, 0,
-+ 0, 0, 0, 0);
-+ return ERR_PTR(-ENOMEM);
-+ }
-+
-+ if (mapped_src_nents == 1 &&
-+ iv_dma + ivsize == sg_dma_address(req->src)) {
-+ in_contig = true;
-+ qm_sg_ents = 0;
-+ } else {
-+ in_contig = false;
-+ qm_sg_ents = 1 + mapped_src_nents;
-+ }
-+ dst_sg_idx = qm_sg_ents;
-+
-+ qm_sg_ents += mapped_dst_nents > 1 ? mapped_dst_nents : 0;
-+ if (unlikely(qm_sg_ents > CAAM_QI_MAX_ABLKCIPHER_SG)) {
-+ dev_err(qidev, "Insufficient S/G entries: %d > %lu\n",
-+ qm_sg_ents, CAAM_QI_MAX_ABLKCIPHER_SG);
-+ caam_unmap(qidev, req->src, req->dst, src_nents, dst_nents,
-+ iv_dma, ivsize, op_type, 0, 0);
-+ return ERR_PTR(-ENOMEM);
-+ }
-+
-+ /* allocate space for base edesc and link tables */
-+ edesc = qi_cache_alloc(GFP_DMA | flags);
-+ if (unlikely(!edesc)) {
-+ dev_err(qidev, "could not allocate extended descriptor\n");
-+ caam_unmap(qidev, req->src, req->dst, src_nents, dst_nents,
-+ iv_dma, ivsize, op_type, 0, 0);
-+ return ERR_PTR(-ENOMEM);
-+ }
-+
-+ edesc->src_nents = src_nents;
-+ edesc->dst_nents = dst_nents;
-+ edesc->iv_dma = iv_dma;
-+ sg_table = &edesc->sgt[0];
-+ edesc->qm_sg_bytes = qm_sg_ents * sizeof(*sg_table);
-+ edesc->drv_req.app_ctx = req;
-+ edesc->drv_req.cbk = ablkcipher_done;
-+ edesc->drv_req.drv_ctx = drv_ctx;
-+
-+ if (!in_contig) {
-+ dma_to_qm_sg_one(sg_table, iv_dma, ivsize, 0);
-+ sg_to_qm_sg_last(req->src, mapped_src_nents, sg_table + 1, 0);
-+ }
-+
-+ if (mapped_dst_nents > 1)
-+ sg_to_qm_sg_last(req->dst, mapped_dst_nents, sg_table +
-+ dst_sg_idx, 0);
-+
-+ edesc->qm_sg_dma = dma_map_single(qidev, sg_table, edesc->qm_sg_bytes,
-+ DMA_TO_DEVICE);
-+ if (dma_mapping_error(qidev, edesc->qm_sg_dma)) {
-+ dev_err(qidev, "unable to map S/G table\n");
-+ caam_unmap(qidev, req->src, req->dst, src_nents, dst_nents,
-+ iv_dma, ivsize, op_type, 0, 0);
-+ qi_cache_free(edesc);
-+ return ERR_PTR(-ENOMEM);
-+ }
-+
-+ fd_sgt = &edesc->drv_req.fd_sgt[0];
-+
-+ if (!in_contig)
-+ dma_to_qm_sg_one_last_ext(&fd_sgt[1], edesc->qm_sg_dma,
-+ ivsize + req->nbytes, 0);
-+ else
-+ dma_to_qm_sg_one_last(&fd_sgt[1], iv_dma, ivsize + req->nbytes,
-+ 0);
-+
-+ if (req->src == req->dst) {
-+ if (!in_contig)
-+ dma_to_qm_sg_one_ext(&fd_sgt[0], edesc->qm_sg_dma +
-+ sizeof(*sg_table), req->nbytes, 0);
-+ else
-+ dma_to_qm_sg_one(&fd_sgt[0], sg_dma_address(req->src),
-+ req->nbytes, 0);
-+ } else if (mapped_dst_nents > 1) {
-+ dma_to_qm_sg_one_ext(&fd_sgt[0], edesc->qm_sg_dma + dst_sg_idx *
-+ sizeof(*sg_table), req->nbytes, 0);
-+ } else {
-+ dma_to_qm_sg_one(&fd_sgt[0], sg_dma_address(req->dst),
-+ req->nbytes, 0);
-+ }
-+
-+ return edesc;
-+}
-+
-+static struct ablkcipher_edesc *ablkcipher_giv_edesc_alloc(
-+ struct skcipher_givcrypt_request *creq)
-+{
-+ struct ablkcipher_request *req = &creq->creq;
-+ struct crypto_ablkcipher *ablkcipher = crypto_ablkcipher_reqtfm(req);
-+ struct caam_ctx *ctx = crypto_ablkcipher_ctx(ablkcipher);
-+ struct device *qidev = ctx->qidev;
-+ gfp_t flags = (req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP) ?
-+ GFP_KERNEL : GFP_ATOMIC;
-+ int src_nents, mapped_src_nents, dst_nents, mapped_dst_nents;
-+ struct ablkcipher_edesc *edesc;
-+ dma_addr_t iv_dma;
-+ bool out_contig;
-+ int ivsize = crypto_ablkcipher_ivsize(ablkcipher);
-+ struct qm_sg_entry *sg_table, *fd_sgt;
-+ int dst_sg_idx, qm_sg_ents;
-+ struct caam_drv_ctx *drv_ctx;
-+
-+ drv_ctx = get_drv_ctx(ctx, GIVENCRYPT);
-+ if (unlikely(IS_ERR_OR_NULL(drv_ctx)))
-+ return (struct ablkcipher_edesc *)drv_ctx;
-+
-+ src_nents = sg_nents_for_len(req->src, req->nbytes);
-+ if (unlikely(src_nents < 0)) {
-+ dev_err(qidev, "Insufficient bytes (%d) in src S/G\n",
-+ req->nbytes);
-+ return ERR_PTR(src_nents);
-+ }
-+
-+ if (unlikely(req->src != req->dst)) {
-+ dst_nents = sg_nents_for_len(req->dst, req->nbytes);
-+ if (unlikely(dst_nents < 0)) {
-+ dev_err(qidev, "Insufficient bytes (%d) in dst S/G\n",
-+ req->nbytes);
-+ return ERR_PTR(dst_nents);
-+ }
-+
-+ mapped_src_nents = dma_map_sg(qidev, req->src, src_nents,
-+ DMA_TO_DEVICE);
-+ if (unlikely(!mapped_src_nents)) {
-+ dev_err(qidev, "unable to map source\n");
-+ return ERR_PTR(-ENOMEM);
-+ }
-+
-+ mapped_dst_nents = dma_map_sg(qidev, req->dst, dst_nents,
-+ DMA_FROM_DEVICE);
-+ if (unlikely(!mapped_dst_nents)) {
-+ dev_err(qidev, "unable to map destination\n");
-+ dma_unmap_sg(qidev, req->src, src_nents, DMA_TO_DEVICE);
-+ return ERR_PTR(-ENOMEM);
-+ }
-+ } else {
-+ mapped_src_nents = dma_map_sg(qidev, req->src, src_nents,
-+ DMA_BIDIRECTIONAL);
-+ if (unlikely(!mapped_src_nents)) {
-+ dev_err(qidev, "unable to map source\n");
-+ return ERR_PTR(-ENOMEM);
-+ }
-+
-+ dst_nents = src_nents;
-+ mapped_dst_nents = src_nents;
-+ }
-+
-+ iv_dma = dma_map_single(qidev, creq->giv, ivsize, DMA_FROM_DEVICE);
-+ if (dma_mapping_error(qidev, iv_dma)) {
-+ dev_err(qidev, "unable to map IV\n");
-+ caam_unmap(qidev, req->src, req->dst, src_nents, dst_nents, 0,
-+ 0, 0, 0, 0);
-+ return ERR_PTR(-ENOMEM);
-+ }
-+
-+ qm_sg_ents = mapped_src_nents > 1 ? mapped_src_nents : 0;
-+ dst_sg_idx = qm_sg_ents;
-+ if (mapped_dst_nents == 1 &&
-+ iv_dma + ivsize == sg_dma_address(req->dst)) {
-+ out_contig = true;
-+ } else {
-+ out_contig = false;
-+ qm_sg_ents += 1 + mapped_dst_nents;
-+ }
-+
-+ if (unlikely(qm_sg_ents > CAAM_QI_MAX_ABLKCIPHER_SG)) {
-+ dev_err(qidev, "Insufficient S/G entries: %d > %lu\n",
-+ qm_sg_ents, CAAM_QI_MAX_ABLKCIPHER_SG);
-+ caam_unmap(qidev, req->src, req->dst, src_nents, dst_nents,
-+ iv_dma, ivsize, GIVENCRYPT, 0, 0);
-+ return ERR_PTR(-ENOMEM);
-+ }
-+
-+ /* allocate space for base edesc and link tables */
-+ edesc = qi_cache_alloc(GFP_DMA | flags);
-+ if (!edesc) {
-+ dev_err(qidev, "could not allocate extended descriptor\n");
-+ caam_unmap(qidev, req->src, req->dst, src_nents, dst_nents,
-+ iv_dma, ivsize, GIVENCRYPT, 0, 0);
-+ return ERR_PTR(-ENOMEM);
-+ }
-+
-+ edesc->src_nents = src_nents;
-+ edesc->dst_nents = dst_nents;
-+ edesc->iv_dma = iv_dma;
-+ sg_table = &edesc->sgt[0];
-+ edesc->qm_sg_bytes = qm_sg_ents * sizeof(*sg_table);
-+ edesc->drv_req.app_ctx = req;
-+ edesc->drv_req.cbk = ablkcipher_done;
-+ edesc->drv_req.drv_ctx = drv_ctx;
-+
-+ if (mapped_src_nents > 1)
-+ sg_to_qm_sg_last(req->src, mapped_src_nents, sg_table, 0);
-+
-+ if (!out_contig) {
-+ dma_to_qm_sg_one(sg_table + dst_sg_idx, iv_dma, ivsize, 0);
-+ sg_to_qm_sg_last(req->dst, mapped_dst_nents, sg_table +
-+ dst_sg_idx + 1, 0);
-+ }
-+
-+ edesc->qm_sg_dma = dma_map_single(qidev, sg_table, edesc->qm_sg_bytes,
-+ DMA_TO_DEVICE);
-+ if (dma_mapping_error(qidev, edesc->qm_sg_dma)) {
-+ dev_err(qidev, "unable to map S/G table\n");
-+ caam_unmap(qidev, req->src, req->dst, src_nents, dst_nents,
-+ iv_dma, ivsize, GIVENCRYPT, 0, 0);
-+ qi_cache_free(edesc);
-+ return ERR_PTR(-ENOMEM);
-+ }
-+
-+ fd_sgt = &edesc->drv_req.fd_sgt[0];
-+
-+ if (mapped_src_nents > 1)
-+ dma_to_qm_sg_one_ext(&fd_sgt[1], edesc->qm_sg_dma, req->nbytes,
-+ 0);
-+ else
-+ dma_to_qm_sg_one(&fd_sgt[1], sg_dma_address(req->src),
-+ req->nbytes, 0);
-+
-+ if (!out_contig)
-+ dma_to_qm_sg_one_ext(&fd_sgt[0], edesc->qm_sg_dma + dst_sg_idx *
-+ sizeof(*sg_table), ivsize + req->nbytes,
-+ 0);
-+ else
-+ dma_to_qm_sg_one(&fd_sgt[0], sg_dma_address(req->dst),
-+ ivsize + req->nbytes, 0);
-+
-+ return edesc;
-+}
-+
-+static inline int ablkcipher_crypt(struct ablkcipher_request *req, bool encrypt)
-+{
-+ struct ablkcipher_edesc *edesc;
-+ struct crypto_ablkcipher *ablkcipher = crypto_ablkcipher_reqtfm(req);
-+ struct caam_ctx *ctx = crypto_ablkcipher_ctx(ablkcipher);
-+ int ret;
-+
-+ if (unlikely(caam_congested))
-+ return -EAGAIN;
-+
-+ /* allocate extended descriptor */
-+ edesc = ablkcipher_edesc_alloc(req, encrypt);
-+ if (IS_ERR(edesc))
-+ return PTR_ERR(edesc);
-+
-+ ret = caam_qi_enqueue(ctx->qidev, &edesc->drv_req);
-+ if (!ret) {
-+ ret = -EINPROGRESS;
-+ } else {
-+ ablkcipher_unmap(ctx->qidev, edesc, req);
-+ qi_cache_free(edesc);
-+ }
-+
-+ return ret;
-+}
-+
-+static int ablkcipher_encrypt(struct ablkcipher_request *req)
-+{
-+ return ablkcipher_crypt(req, true);
-+}
-+
-+static int ablkcipher_decrypt(struct ablkcipher_request *req)
-+{
-+ return ablkcipher_crypt(req, false);
-+}
-+
-+static int ablkcipher_givencrypt(struct skcipher_givcrypt_request *creq)
-+{
-+ struct ablkcipher_request *req = &creq->creq;
-+ struct ablkcipher_edesc *edesc;
-+ struct crypto_ablkcipher *ablkcipher = crypto_ablkcipher_reqtfm(req);
-+ struct caam_ctx *ctx = crypto_ablkcipher_ctx(ablkcipher);
-+ int ret;
-+
-+ if (unlikely(caam_congested))
-+ return -EAGAIN;
-+
-+ /* allocate extended descriptor */
-+ edesc = ablkcipher_giv_edesc_alloc(creq);
-+ if (IS_ERR(edesc))
-+ return PTR_ERR(edesc);
-+
-+ ret = caam_qi_enqueue(ctx->qidev, &edesc->drv_req);
-+ if (!ret) {
-+ ret = -EINPROGRESS;
-+ } else {
-+ ablkcipher_unmap(ctx->qidev, edesc, req);
-+ qi_cache_free(edesc);
-+ }
-+
-+ return ret;
-+}
-+
-+#define template_ablkcipher template_u.ablkcipher
-+struct caam_alg_template {
-+ char name[CRYPTO_MAX_ALG_NAME];
-+ char driver_name[CRYPTO_MAX_ALG_NAME];
-+ unsigned int blocksize;
-+ u32 type;
-+ union {
-+ struct ablkcipher_alg ablkcipher;
-+ } template_u;
-+ u32 class1_alg_type;
-+ u32 class2_alg_type;
-+};
-+
-+static struct caam_alg_template driver_algs[] = {
-+ /* ablkcipher descriptor */
-+ {
-+ .name = "cbc(aes)",
-+ .driver_name = "cbc-aes-caam-qi",
-+ .blocksize = AES_BLOCK_SIZE,
-+ .type = CRYPTO_ALG_TYPE_GIVCIPHER,
-+ .template_ablkcipher = {
-+ .setkey = ablkcipher_setkey,
-+ .encrypt = ablkcipher_encrypt,
-+ .decrypt = ablkcipher_decrypt,
-+ .givencrypt = ablkcipher_givencrypt,
-+ .geniv = "<built-in>",
-+ .min_keysize = AES_MIN_KEY_SIZE,
-+ .max_keysize = AES_MAX_KEY_SIZE,
-+ .ivsize = AES_BLOCK_SIZE,
-+ },
-+ .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_CBC,
-+ },
-+ {
-+ .name = "cbc(des3_ede)",
-+ .driver_name = "cbc-3des-caam-qi",
-+ .blocksize = DES3_EDE_BLOCK_SIZE,
-+ .type = CRYPTO_ALG_TYPE_GIVCIPHER,
-+ .template_ablkcipher = {
-+ .setkey = ablkcipher_setkey,
-+ .encrypt = ablkcipher_encrypt,
-+ .decrypt = ablkcipher_decrypt,
-+ .givencrypt = ablkcipher_givencrypt,
-+ .geniv = "<built-in>",
-+ .min_keysize = DES3_EDE_KEY_SIZE,
-+ .max_keysize = DES3_EDE_KEY_SIZE,
-+ .ivsize = DES3_EDE_BLOCK_SIZE,
-+ },
-+ .class1_alg_type = OP_ALG_ALGSEL_3DES | OP_ALG_AAI_CBC,
-+ },
-+ {
-+ .name = "cbc(des)",
-+ .driver_name = "cbc-des-caam-qi",
-+ .blocksize = DES_BLOCK_SIZE,
-+ .type = CRYPTO_ALG_TYPE_GIVCIPHER,
-+ .template_ablkcipher = {
-+ .setkey = ablkcipher_setkey,
-+ .encrypt = ablkcipher_encrypt,
-+ .decrypt = ablkcipher_decrypt,
-+ .givencrypt = ablkcipher_givencrypt,
-+ .geniv = "<built-in>",
-+ .min_keysize = DES_KEY_SIZE,
-+ .max_keysize = DES_KEY_SIZE,
-+ .ivsize = DES_BLOCK_SIZE,
-+ },
-+ .class1_alg_type = OP_ALG_ALGSEL_DES | OP_ALG_AAI_CBC,
-+ },
-+ {
-+ .name = "ctr(aes)",
-+ .driver_name = "ctr-aes-caam-qi",
-+ .blocksize = 1,
-+ .type = CRYPTO_ALG_TYPE_ABLKCIPHER,
-+ .template_ablkcipher = {
-+ .setkey = ablkcipher_setkey,
-+ .encrypt = ablkcipher_encrypt,
-+ .decrypt = ablkcipher_decrypt,
-+ .geniv = "chainiv",
-+ .min_keysize = AES_MIN_KEY_SIZE,
-+ .max_keysize = AES_MAX_KEY_SIZE,
-+ .ivsize = AES_BLOCK_SIZE,
-+ },
-+ .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_CTR_MOD128,
-+ },
-+ {
-+ .name = "rfc3686(ctr(aes))",
-+ .driver_name = "rfc3686-ctr-aes-caam-qi",
-+ .blocksize = 1,
-+ .type = CRYPTO_ALG_TYPE_GIVCIPHER,
-+ .template_ablkcipher = {
-+ .setkey = ablkcipher_setkey,
-+ .encrypt = ablkcipher_encrypt,
-+ .decrypt = ablkcipher_decrypt,
-+ .givencrypt = ablkcipher_givencrypt,
-+ .geniv = "<built-in>",
-+ .min_keysize = AES_MIN_KEY_SIZE +
-+ CTR_RFC3686_NONCE_SIZE,
-+ .max_keysize = AES_MAX_KEY_SIZE +
-+ CTR_RFC3686_NONCE_SIZE,
-+ .ivsize = CTR_RFC3686_IV_SIZE,
-+ },
-+ .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_CTR_MOD128,
-+ },
-+ {
-+ .name = "xts(aes)",
-+ .driver_name = "xts-aes-caam-qi",
-+ .blocksize = AES_BLOCK_SIZE,
-+ .type = CRYPTO_ALG_TYPE_ABLKCIPHER,
-+ .template_ablkcipher = {
-+ .setkey = xts_ablkcipher_setkey,
-+ .encrypt = ablkcipher_encrypt,
-+ .decrypt = ablkcipher_decrypt,
-+ .geniv = "eseqiv",
-+ .min_keysize = 2 * AES_MIN_KEY_SIZE,
-+ .max_keysize = 2 * AES_MAX_KEY_SIZE,
-+ .ivsize = AES_BLOCK_SIZE,
-+ },
-+ .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_XTS,
-+ },
-+};
-+
-+static struct caam_aead_alg driver_aeads[] = {
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "rfc4106(gcm(aes))",
-+ .cra_driver_name = "rfc4106-gcm-aes-caam-qi",
-+ .cra_blocksize = 1,
-+ },
-+ .setkey = rfc4106_setkey,
-+ .setauthsize = rfc4106_setauthsize,
-+ .encrypt = ipsec_gcm_encrypt,
-+ .decrypt = ipsec_gcm_decrypt,
-+ .ivsize = 8,
-+ .maxauthsize = AES_BLOCK_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_GCM,
-+ },
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "rfc4543(gcm(aes))",
-+ .cra_driver_name = "rfc4543-gcm-aes-caam-qi",
-+ .cra_blocksize = 1,
-+ },
-+ .setkey = rfc4543_setkey,
-+ .setauthsize = rfc4543_setauthsize,
-+ .encrypt = ipsec_gcm_encrypt,
-+ .decrypt = ipsec_gcm_decrypt,
-+ .ivsize = 8,
-+ .maxauthsize = AES_BLOCK_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_GCM,
-+ },
-+ },
-+ /* Galois Counter Mode */
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "gcm(aes)",
-+ .cra_driver_name = "gcm-aes-caam-qi",
-+ .cra_blocksize = 1,
-+ },
-+ .setkey = gcm_setkey,
-+ .setauthsize = gcm_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = 12,
-+ .maxauthsize = AES_BLOCK_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_GCM,
-+ }
-+ },
-+ /* single-pass ipsec_esp descriptor */
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "authenc(hmac(md5),cbc(aes))",
-+ .cra_driver_name = "authenc-hmac-md5-"
-+ "cbc-aes-caam-qi",
-+ .cra_blocksize = AES_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = AES_BLOCK_SIZE,
-+ .maxauthsize = MD5_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_MD5 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ }
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "echainiv(authenc(hmac(md5),"
-+ "cbc(aes)))",
-+ .cra_driver_name = "echainiv-authenc-hmac-md5-"
-+ "cbc-aes-caam-qi",
-+ .cra_blocksize = AES_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = AES_BLOCK_SIZE,
-+ .maxauthsize = MD5_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_MD5 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ .geniv = true,
-+ }
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "authenc(hmac(sha1),cbc(aes))",
-+ .cra_driver_name = "authenc-hmac-sha1-"
-+ "cbc-aes-caam-qi",
-+ .cra_blocksize = AES_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = AES_BLOCK_SIZE,
-+ .maxauthsize = SHA1_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA1 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ }
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "echainiv(authenc(hmac(sha1),"
-+ "cbc(aes)))",
-+ .cra_driver_name = "echainiv-authenc-"
-+ "hmac-sha1-cbc-aes-caam-qi",
-+ .cra_blocksize = AES_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = AES_BLOCK_SIZE,
-+ .maxauthsize = SHA1_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA1 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ .geniv = true,
-+ },
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "authenc(hmac(sha224),cbc(aes))",
-+ .cra_driver_name = "authenc-hmac-sha224-"
-+ "cbc-aes-caam-qi",
-+ .cra_blocksize = AES_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = AES_BLOCK_SIZE,
-+ .maxauthsize = SHA224_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA224 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ }
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "echainiv(authenc(hmac(sha224),"
-+ "cbc(aes)))",
-+ .cra_driver_name = "echainiv-authenc-"
-+ "hmac-sha224-cbc-aes-caam-qi",
-+ .cra_blocksize = AES_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = AES_BLOCK_SIZE,
-+ .maxauthsize = SHA224_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA224 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ .geniv = true,
-+ }
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "authenc(hmac(sha256),cbc(aes))",
-+ .cra_driver_name = "authenc-hmac-sha256-"
-+ "cbc-aes-caam-qi",
-+ .cra_blocksize = AES_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = AES_BLOCK_SIZE,
-+ .maxauthsize = SHA256_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA256 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ }
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "echainiv(authenc(hmac(sha256),"
-+ "cbc(aes)))",
-+ .cra_driver_name = "echainiv-authenc-"
-+ "hmac-sha256-cbc-aes-"
-+ "caam-qi",
-+ .cra_blocksize = AES_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = AES_BLOCK_SIZE,
-+ .maxauthsize = SHA256_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA256 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ .geniv = true,
-+ }
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "authenc(hmac(sha384),cbc(aes))",
-+ .cra_driver_name = "authenc-hmac-sha384-"
-+ "cbc-aes-caam-qi",
-+ .cra_blocksize = AES_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = AES_BLOCK_SIZE,
-+ .maxauthsize = SHA384_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA384 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ }
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "echainiv(authenc(hmac(sha384),"
-+ "cbc(aes)))",
-+ .cra_driver_name = "echainiv-authenc-"
-+ "hmac-sha384-cbc-aes-"
-+ "caam-qi",
-+ .cra_blocksize = AES_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = AES_BLOCK_SIZE,
-+ .maxauthsize = SHA384_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA384 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ .geniv = true,
-+ }
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "authenc(hmac(sha512),cbc(aes))",
-+ .cra_driver_name = "authenc-hmac-sha512-"
-+ "cbc-aes-caam-qi",
-+ .cra_blocksize = AES_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = AES_BLOCK_SIZE,
-+ .maxauthsize = SHA512_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA512 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ }
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "echainiv(authenc(hmac(sha512),"
-+ "cbc(aes)))",
-+ .cra_driver_name = "echainiv-authenc-"
-+ "hmac-sha512-cbc-aes-"
-+ "caam-qi",
-+ .cra_blocksize = AES_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = AES_BLOCK_SIZE,
-+ .maxauthsize = SHA512_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA512 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ .geniv = true,
-+ }
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "authenc(hmac(md5),cbc(des3_ede))",
-+ .cra_driver_name = "authenc-hmac-md5-"
-+ "cbc-des3_ede-caam-qi",
-+ .cra_blocksize = DES3_EDE_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = DES3_EDE_BLOCK_SIZE,
-+ .maxauthsize = MD5_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_3DES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_MD5 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ }
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "echainiv(authenc(hmac(md5),"
-+ "cbc(des3_ede)))",
-+ .cra_driver_name = "echainiv-authenc-hmac-md5-"
-+ "cbc-des3_ede-caam-qi",
-+ .cra_blocksize = DES3_EDE_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = DES3_EDE_BLOCK_SIZE,
-+ .maxauthsize = MD5_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_3DES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_MD5 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ .geniv = true,
-+ }
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "authenc(hmac(sha1),"
-+ "cbc(des3_ede))",
-+ .cra_driver_name = "authenc-hmac-sha1-"
-+ "cbc-des3_ede-caam-qi",
-+ .cra_blocksize = DES3_EDE_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = DES3_EDE_BLOCK_SIZE,
-+ .maxauthsize = SHA1_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_3DES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA1 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ },
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "echainiv(authenc(hmac(sha1),"
-+ "cbc(des3_ede)))",
-+ .cra_driver_name = "echainiv-authenc-"
-+ "hmac-sha1-"
-+ "cbc-des3_ede-caam-qi",
-+ .cra_blocksize = DES3_EDE_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = DES3_EDE_BLOCK_SIZE,
-+ .maxauthsize = SHA1_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_3DES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA1 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ .geniv = true,
-+ }
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "authenc(hmac(sha224),"
-+ "cbc(des3_ede))",
-+ .cra_driver_name = "authenc-hmac-sha224-"
-+ "cbc-des3_ede-caam-qi",
-+ .cra_blocksize = DES3_EDE_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = DES3_EDE_BLOCK_SIZE,
-+ .maxauthsize = SHA224_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_3DES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA224 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ },
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "echainiv(authenc(hmac(sha224),"
-+ "cbc(des3_ede)))",
-+ .cra_driver_name = "echainiv-authenc-"
-+ "hmac-sha224-"
-+ "cbc-des3_ede-caam-qi",
-+ .cra_blocksize = DES3_EDE_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = DES3_EDE_BLOCK_SIZE,
-+ .maxauthsize = SHA224_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_3DES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA224 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ .geniv = true,
-+ }
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "authenc(hmac(sha256),"
-+ "cbc(des3_ede))",
-+ .cra_driver_name = "authenc-hmac-sha256-"
-+ "cbc-des3_ede-caam-qi",
-+ .cra_blocksize = DES3_EDE_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = DES3_EDE_BLOCK_SIZE,
-+ .maxauthsize = SHA256_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_3DES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA256 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ },
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "echainiv(authenc(hmac(sha256),"
-+ "cbc(des3_ede)))",
-+ .cra_driver_name = "echainiv-authenc-"
-+ "hmac-sha256-"
-+ "cbc-des3_ede-caam-qi",
-+ .cra_blocksize = DES3_EDE_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = DES3_EDE_BLOCK_SIZE,
-+ .maxauthsize = SHA256_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_3DES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA256 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ .geniv = true,
-+ }
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "authenc(hmac(sha384),"
-+ "cbc(des3_ede))",
-+ .cra_driver_name = "authenc-hmac-sha384-"
-+ "cbc-des3_ede-caam-qi",
-+ .cra_blocksize = DES3_EDE_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = DES3_EDE_BLOCK_SIZE,
-+ .maxauthsize = SHA384_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_3DES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA384 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ },
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "echainiv(authenc(hmac(sha384),"
-+ "cbc(des3_ede)))",
-+ .cra_driver_name = "echainiv-authenc-"
-+ "hmac-sha384-"
-+ "cbc-des3_ede-caam-qi",
-+ .cra_blocksize = DES3_EDE_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = DES3_EDE_BLOCK_SIZE,
-+ .maxauthsize = SHA384_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_3DES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA384 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ .geniv = true,
-+ }
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "authenc(hmac(sha512),"
-+ "cbc(des3_ede))",
-+ .cra_driver_name = "authenc-hmac-sha512-"
-+ "cbc-des3_ede-caam-qi",
-+ .cra_blocksize = DES3_EDE_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = DES3_EDE_BLOCK_SIZE,
-+ .maxauthsize = SHA512_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_3DES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA512 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ },
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "echainiv(authenc(hmac(sha512),"
-+ "cbc(des3_ede)))",
-+ .cra_driver_name = "echainiv-authenc-"
-+ "hmac-sha512-"
-+ "cbc-des3_ede-caam-qi",
-+ .cra_blocksize = DES3_EDE_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = DES3_EDE_BLOCK_SIZE,
-+ .maxauthsize = SHA512_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_3DES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA512 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ .geniv = true,
-+ }
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "authenc(hmac(md5),cbc(des))",
-+ .cra_driver_name = "authenc-hmac-md5-"
-+ "cbc-des-caam-qi",
-+ .cra_blocksize = DES_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = DES_BLOCK_SIZE,
-+ .maxauthsize = MD5_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_DES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_MD5 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ },
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "echainiv(authenc(hmac(md5),"
-+ "cbc(des)))",
-+ .cra_driver_name = "echainiv-authenc-hmac-md5-"
-+ "cbc-des-caam-qi",
-+ .cra_blocksize = DES_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = DES_BLOCK_SIZE,
-+ .maxauthsize = MD5_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_DES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_MD5 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ .geniv = true,
-+ }
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "authenc(hmac(sha1),cbc(des))",
-+ .cra_driver_name = "authenc-hmac-sha1-"
-+ "cbc-des-caam-qi",
-+ .cra_blocksize = DES_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = DES_BLOCK_SIZE,
-+ .maxauthsize = SHA1_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_DES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA1 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ },
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "echainiv(authenc(hmac(sha1),"
-+ "cbc(des)))",
-+ .cra_driver_name = "echainiv-authenc-"
-+ "hmac-sha1-cbc-des-caam-qi",
-+ .cra_blocksize = DES_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = DES_BLOCK_SIZE,
-+ .maxauthsize = SHA1_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_DES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA1 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ .geniv = true,
-+ }
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "authenc(hmac(sha224),cbc(des))",
-+ .cra_driver_name = "authenc-hmac-sha224-"
-+ "cbc-des-caam-qi",
-+ .cra_blocksize = DES_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = DES_BLOCK_SIZE,
-+ .maxauthsize = SHA224_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_DES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA224 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ },
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "echainiv(authenc(hmac(sha224),"
-+ "cbc(des)))",
-+ .cra_driver_name = "echainiv-authenc-"
-+ "hmac-sha224-cbc-des-"
-+ "caam-qi",
-+ .cra_blocksize = DES_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = DES_BLOCK_SIZE,
-+ .maxauthsize = SHA224_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_DES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA224 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ .geniv = true,
-+ }
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "authenc(hmac(sha256),cbc(des))",
-+ .cra_driver_name = "authenc-hmac-sha256-"
-+ "cbc-des-caam-qi",
-+ .cra_blocksize = DES_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = DES_BLOCK_SIZE,
-+ .maxauthsize = SHA256_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_DES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA256 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ },
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "echainiv(authenc(hmac(sha256),"
-+ "cbc(des)))",
-+ .cra_driver_name = "echainiv-authenc-"
-+ "hmac-sha256-cbc-des-"
-+ "caam-qi",
-+ .cra_blocksize = DES_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = DES_BLOCK_SIZE,
-+ .maxauthsize = SHA256_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_DES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA256 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ .geniv = true,
-+ },
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "authenc(hmac(sha384),cbc(des))",
-+ .cra_driver_name = "authenc-hmac-sha384-"
-+ "cbc-des-caam-qi",
-+ .cra_blocksize = DES_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = DES_BLOCK_SIZE,
-+ .maxauthsize = SHA384_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_DES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA384 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ },
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "echainiv(authenc(hmac(sha384),"
-+ "cbc(des)))",
-+ .cra_driver_name = "echainiv-authenc-"
-+ "hmac-sha384-cbc-des-"
-+ "caam-qi",
-+ .cra_blocksize = DES_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = DES_BLOCK_SIZE,
-+ .maxauthsize = SHA384_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_DES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA384 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ .geniv = true,
-+ }
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "authenc(hmac(sha512),cbc(des))",
-+ .cra_driver_name = "authenc-hmac-sha512-"
-+ "cbc-des-caam-qi",
-+ .cra_blocksize = DES_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = DES_BLOCK_SIZE,
-+ .maxauthsize = SHA512_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_DES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA512 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ }
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "echainiv(authenc(hmac(sha512),"
-+ "cbc(des)))",
-+ .cra_driver_name = "echainiv-authenc-"
-+ "hmac-sha512-cbc-des-"
-+ "caam-qi",
-+ .cra_blocksize = DES_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = DES_BLOCK_SIZE,
-+ .maxauthsize = SHA512_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_DES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA512 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ .geniv = true,
-+ }
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "tls10(hmac(sha1),cbc(aes))",
-+ .cra_driver_name = "tls10-hmac-sha1-cbc-aes-caam-qi",
-+ .cra_blocksize = AES_BLOCK_SIZE,
-+ },
-+ .setkey = tls_setkey,
-+ .setauthsize = tls_setauthsize,
-+ .encrypt = tls_encrypt,
-+ .decrypt = tls_decrypt,
-+ .ivsize = AES_BLOCK_SIZE,
-+ .maxauthsize = SHA1_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA1 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ }
-+ }
-+};
-+
-+struct caam_crypto_alg {
-+ struct list_head entry;
-+ struct crypto_alg crypto_alg;
-+ struct caam_alg_entry caam;
-+};
-+
-+static int caam_init_common(struct caam_ctx *ctx, struct caam_alg_entry *caam,
-+ bool uses_dkp)
-+{
-+ struct caam_drv_private *priv;
-+ /* Digest sizes for MD5, SHA1, SHA-224, SHA-256, SHA-384, SHA-512 */
-+ static const u8 digest_size[] = {
-+ MD5_DIGEST_SIZE,
-+ SHA1_DIGEST_SIZE,
-+ SHA224_DIGEST_SIZE,
-+ SHA256_DIGEST_SIZE,
-+ SHA384_DIGEST_SIZE,
-+ SHA512_DIGEST_SIZE
-+ };
-+ u8 op_id;
-+
-+ /*
-+ * distribute tfms across job rings to ensure in-order
-+ * crypto request processing per tfm
-+ */
-+ ctx->jrdev = caam_jr_alloc();
-+ if (IS_ERR(ctx->jrdev)) {
-+ pr_err("Job Ring Device allocation for transform failed\n");
-+ return PTR_ERR(ctx->jrdev);
-+ }
-+
-+ priv = dev_get_drvdata(ctx->jrdev->parent);
-+ if (priv->era >= 6 && uses_dkp)
-+ ctx->dir = DMA_BIDIRECTIONAL;
-+ else
-+ ctx->dir = DMA_TO_DEVICE;
-+
-+ ctx->key_dma = dma_map_single(ctx->jrdev, ctx->key, sizeof(ctx->key),
-+ ctx->dir);
-+ if (dma_mapping_error(ctx->jrdev, ctx->key_dma)) {
-+ dev_err(ctx->jrdev, "unable to map key\n");
-+ caam_jr_free(ctx->jrdev);
-+ return -ENOMEM;
-+ }
-+
-+ /* copy descriptor header template value */
-+ ctx->cdata.algtype = OP_TYPE_CLASS1_ALG | caam->class1_alg_type;
-+ ctx->adata.algtype = OP_TYPE_CLASS2_ALG | caam->class2_alg_type;
-+
-+ if (ctx->adata.algtype) {
-+ op_id = (ctx->adata.algtype & OP_ALG_ALGSEL_SUBMASK)
-+ >> OP_ALG_ALGSEL_SHIFT;
-+ if (op_id < ARRAY_SIZE(digest_size)) {
-+ ctx->authsize = digest_size[op_id];
-+ } else {
-+ dev_err(ctx->jrdev,
-+ "incorrect op_id %d; must be less than %zu\n",
-+ op_id, ARRAY_SIZE(digest_size));
-+ caam_jr_free(ctx->jrdev);
-+ return -EINVAL;
-+ }
-+ } else {
-+ ctx->authsize = 0;
-+ }
-+
-+ ctx->qidev = priv->qidev;
-+
-+ spin_lock_init(&ctx->lock);
-+ ctx->drv_ctx[ENCRYPT] = NULL;
-+ ctx->drv_ctx[DECRYPT] = NULL;
-+ ctx->drv_ctx[GIVENCRYPT] = NULL;
-+
-+ return 0;
-+}
-+
-+static int caam_cra_init(struct crypto_tfm *tfm)
-+{
-+ struct crypto_alg *alg = tfm->__crt_alg;
-+ struct caam_crypto_alg *caam_alg = container_of(alg, typeof(*caam_alg),
-+ crypto_alg);
-+ struct caam_ctx *ctx = crypto_tfm_ctx(tfm);
-+
-+ return caam_init_common(ctx, &caam_alg->caam, false);
-+}
-+
-+static int caam_aead_init(struct crypto_aead *tfm)
-+{
-+ struct aead_alg *alg = crypto_aead_alg(tfm);
-+ struct caam_aead_alg *caam_alg = container_of(alg, typeof(*caam_alg),
-+ aead);
-+ struct caam_ctx *ctx = crypto_aead_ctx(tfm);
-+
-+ return caam_init_common(ctx, &caam_alg->caam,
-+ (alg->setkey == aead_setkey) ||
-+ (alg->setkey == tls_setkey));
-+}
-+
-+static void caam_exit_common(struct caam_ctx *ctx)
-+{
-+ caam_drv_ctx_rel(ctx->drv_ctx[ENCRYPT]);
-+ caam_drv_ctx_rel(ctx->drv_ctx[DECRYPT]);
-+ caam_drv_ctx_rel(ctx->drv_ctx[GIVENCRYPT]);
-+
-+ dma_unmap_single(ctx->jrdev, ctx->key_dma, sizeof(ctx->key), ctx->dir);
-+
-+ caam_jr_free(ctx->jrdev);
-+}
-+
-+static void caam_cra_exit(struct crypto_tfm *tfm)
-+{
-+ caam_exit_common(crypto_tfm_ctx(tfm));
-+}
-+
-+static void caam_aead_exit(struct crypto_aead *tfm)
-+{
-+ caam_exit_common(crypto_aead_ctx(tfm));
-+}
-+
-+static struct list_head alg_list;
-+static void __exit caam_qi_algapi_exit(void)
-+{
-+ struct caam_crypto_alg *t_alg, *n;
-+ int i;
-+
-+ for (i = 0; i < ARRAY_SIZE(driver_aeads); i++) {
-+ struct caam_aead_alg *t_alg = driver_aeads + i;
-+
-+ if (t_alg->registered)
-+ crypto_unregister_aead(&t_alg->aead);
-+ }
-+
-+ if (!alg_list.next)
-+ return;
-+
-+ list_for_each_entry_safe(t_alg, n, &alg_list, entry) {
-+ crypto_unregister_alg(&t_alg->crypto_alg);
-+ list_del(&t_alg->entry);
-+ kfree(t_alg);
-+ }
-+}
-+
-+static struct caam_crypto_alg *caam_alg_alloc(struct caam_alg_template
-+ *template)
-+{
-+ struct caam_crypto_alg *t_alg;
-+ struct crypto_alg *alg;
-+
-+ t_alg = kzalloc(sizeof(*t_alg), GFP_KERNEL);
-+ if (!t_alg)
-+ return ERR_PTR(-ENOMEM);
-+
-+ alg = &t_alg->crypto_alg;
-+
-+ snprintf(alg->cra_name, CRYPTO_MAX_ALG_NAME, "%s", template->name);
-+ snprintf(alg->cra_driver_name, CRYPTO_MAX_ALG_NAME, "%s",
-+ template->driver_name);
-+ alg->cra_module = THIS_MODULE;
-+ alg->cra_init = caam_cra_init;
-+ alg->cra_exit = caam_cra_exit;
-+ alg->cra_priority = CAAM_CRA_PRIORITY;
-+ alg->cra_blocksize = template->blocksize;
-+ alg->cra_alignmask = 0;
-+ alg->cra_ctxsize = sizeof(struct caam_ctx);
-+ alg->cra_flags = CRYPTO_ALG_ASYNC | CRYPTO_ALG_KERN_DRIVER_ONLY |
-+ template->type;
-+ switch (template->type) {
-+ case CRYPTO_ALG_TYPE_GIVCIPHER:
-+ alg->cra_type = &crypto_givcipher_type;
-+ alg->cra_ablkcipher = template->template_ablkcipher;
-+ break;
-+ case CRYPTO_ALG_TYPE_ABLKCIPHER:
-+ alg->cra_type = &crypto_ablkcipher_type;
-+ alg->cra_ablkcipher = template->template_ablkcipher;
-+ break;
-+ }
-+
-+ t_alg->caam.class1_alg_type = template->class1_alg_type;
-+ t_alg->caam.class2_alg_type = template->class2_alg_type;
-+
-+ return t_alg;
-+}
-+
-+static void caam_aead_alg_init(struct caam_aead_alg *t_alg)
-+{
-+ struct aead_alg *alg = &t_alg->aead;
-+
-+ alg->base.cra_module = THIS_MODULE;
-+ alg->base.cra_priority = CAAM_CRA_PRIORITY;
-+ alg->base.cra_ctxsize = sizeof(struct caam_ctx);
-+ alg->base.cra_flags = CRYPTO_ALG_ASYNC | CRYPTO_ALG_KERN_DRIVER_ONLY;
-+
-+ alg->init = caam_aead_init;
-+ alg->exit = caam_aead_exit;
-+}
-+
-+static int __init caam_qi_algapi_init(void)
-+{
-+ struct device_node *dev_node;
-+ struct platform_device *pdev;
-+ struct device *ctrldev;
-+ struct caam_drv_private *priv;
-+ int i = 0, err = 0;
-+ u32 cha_vid, cha_inst, des_inst, aes_inst, md_inst;
-+ unsigned int md_limit = SHA512_DIGEST_SIZE;
-+ bool registered = false;
-+
-+ dev_node = of_find_compatible_node(NULL, NULL, "fsl,sec-v4.0");
-+ if (!dev_node) {
-+ dev_node = of_find_compatible_node(NULL, NULL, "fsl,sec4.0");
-+ if (!dev_node)
-+ return -ENODEV;
-+ }
-+
-+ pdev = of_find_device_by_node(dev_node);
-+ of_node_put(dev_node);
-+ if (!pdev)
-+ return -ENODEV;
-+
-+ ctrldev = &pdev->dev;
-+ priv = dev_get_drvdata(ctrldev);
-+
-+ /*
-+ * If priv is NULL, it's probably because the caam driver wasn't
-+ * properly initialized (e.g. RNG4 init failed). Thus, bail out here.
-+ */
-+ if (!priv || !priv->qi_present)
-+ return -ENODEV;
-+
-+ if (caam_dpaa2) {
-+ dev_info(ctrldev, "caam/qi frontend driver not suitable for DPAA 2.x, aborting...\n");
-+ return -ENODEV;
-+ }
-+
-+ INIT_LIST_HEAD(&alg_list);
-+
-+ /*
-+ * Register crypto algorithms the device supports.
-+ * First, detect presence and attributes of DES, AES, and MD blocks.
-+ */
-+ cha_vid = rd_reg32(&priv->ctrl->perfmon.cha_id_ls);
-+ cha_inst = rd_reg32(&priv->ctrl->perfmon.cha_num_ls);
-+ des_inst = (cha_inst & CHA_ID_LS_DES_MASK) >> CHA_ID_LS_DES_SHIFT;
-+ aes_inst = (cha_inst & CHA_ID_LS_AES_MASK) >> CHA_ID_LS_AES_SHIFT;
-+ md_inst = (cha_inst & CHA_ID_LS_MD_MASK) >> CHA_ID_LS_MD_SHIFT;
-+
-+ /* If MD is present, limit digest size based on LP256 */
-+ if (md_inst && ((cha_vid & CHA_ID_LS_MD_MASK) == CHA_ID_LS_MD_LP256))
-+ md_limit = SHA256_DIGEST_SIZE;
-+
-+ for (i = 0; i < ARRAY_SIZE(driver_algs); i++) {
-+ struct caam_crypto_alg *t_alg;
-+ struct caam_alg_template *alg = driver_algs + i;
-+ u32 alg_sel = alg->class1_alg_type & OP_ALG_ALGSEL_MASK;
-+
-+ /* Skip DES algorithms if not supported by device */
-+ if (!des_inst &&
-+ ((alg_sel == OP_ALG_ALGSEL_3DES) ||
-+ (alg_sel == OP_ALG_ALGSEL_DES)))
-+ continue;
-+
-+ /* Skip AES algorithms if not supported by device */
-+ if (!aes_inst && (alg_sel == OP_ALG_ALGSEL_AES))
-+ continue;
-+
-+ t_alg = caam_alg_alloc(alg);
-+ if (IS_ERR(t_alg)) {
-+ err = PTR_ERR(t_alg);
-+ dev_warn(priv->qidev, "%s alg allocation failed\n",
-+ alg->driver_name);
-+ continue;
-+ }
-+
-+ err = crypto_register_alg(&t_alg->crypto_alg);
-+ if (err) {
-+ dev_warn(priv->qidev, "%s alg registration failed\n",
-+ t_alg->crypto_alg.cra_driver_name);
-+ kfree(t_alg);
-+ continue;
-+ }
-+
-+ list_add_tail(&t_alg->entry, &alg_list);
-+ registered = true;
-+ }
-+
-+ for (i = 0; i < ARRAY_SIZE(driver_aeads); i++) {
-+ struct caam_aead_alg *t_alg = driver_aeads + i;
-+ u32 c1_alg_sel = t_alg->caam.class1_alg_type &
-+ OP_ALG_ALGSEL_MASK;
-+ u32 c2_alg_sel = t_alg->caam.class2_alg_type &
-+ OP_ALG_ALGSEL_MASK;
-+ u32 alg_aai = t_alg->caam.class1_alg_type & OP_ALG_AAI_MASK;
-+
-+ /* Skip DES algorithms if not supported by device */
-+ if (!des_inst &&
-+ ((c1_alg_sel == OP_ALG_ALGSEL_3DES) ||
-+ (c1_alg_sel == OP_ALG_ALGSEL_DES)))
-+ continue;
-+
-+ /* Skip AES algorithms if not supported by device */
-+ if (!aes_inst && (c1_alg_sel == OP_ALG_ALGSEL_AES))
-+ continue;
-+
-+ /*
-+ * Check support for AES algorithms not available
-+ * on LP devices.
-+ */
-+ if (((cha_vid & CHA_ID_LS_AES_MASK) == CHA_ID_LS_AES_LP) &&
-+ (alg_aai == OP_ALG_AAI_GCM))
-+ continue;
-+
-+ /*
-+ * Skip algorithms requiring message digests
-+ * if MD or MD size is not supported by device.
-+ */
-+ if (c2_alg_sel &&
-+ (!md_inst || (t_alg->aead.maxauthsize > md_limit)))
-+ continue;
-+
-+ caam_aead_alg_init(t_alg);
-+
-+ err = crypto_register_aead(&t_alg->aead);
-+ if (err) {
-+ pr_warn("%s alg registration failed\n",
-+ t_alg->aead.base.cra_driver_name);
-+ continue;
-+ }
-+
-+ t_alg->registered = true;
-+ registered = true;
-+ }
-+
-+ if (registered)
-+ dev_info(priv->qidev, "algorithms registered in /proc/crypto\n");
-+
-+ return err;
-+}
-+
-+module_init(caam_qi_algapi_init);
-+module_exit(caam_qi_algapi_exit);
-+
-+MODULE_LICENSE("GPL");
-+MODULE_DESCRIPTION("Support for crypto API using CAAM-QI backend");
-+MODULE_AUTHOR("Freescale Semiconductor");
---- /dev/null
-+++ b/drivers/crypto/caam/caamalg_qi2.c
-@@ -0,0 +1,5938 @@
-+/*
-+ * Copyright 2015-2016 Freescale Semiconductor Inc.
-+ * Copyright 2017 NXP
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the names of the above-listed copyright holders nor the
-+ * names of any contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
-+ * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
-+ * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
-+ * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE
-+ * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
-+ * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
-+ * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
-+ * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
-+ * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
-+ * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
-+ * POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#include <linux/fsl/mc.h>
-+#include "compat.h"
-+#include "regs.h"
-+#include "caamalg_qi2.h"
-+#include "dpseci_cmd.h"
-+#include "desc_constr.h"
-+#include "error.h"
-+#include "sg_sw_sec4.h"
-+#include "sg_sw_qm2.h"
-+#include "key_gen.h"
-+#include "caamalg_desc.h"
-+#include "caamhash_desc.h"
-+#include "../../../drivers/staging/fsl-mc/include/dpaa2-io.h"
-+#include "../../../drivers/staging/fsl-mc/include/dpaa2-fd.h"
-+
-+#define CAAM_CRA_PRIORITY 2000
-+
-+/* max key is sum of AES_MAX_KEY_SIZE, max split key size */
-+#define CAAM_MAX_KEY_SIZE (AES_MAX_KEY_SIZE + CTR_RFC3686_NONCE_SIZE + \
-+ SHA512_DIGEST_SIZE * 2)
-+
-+#ifndef CONFIG_CRYPTO_DEV_FSL_CAAM
-+bool caam_little_end;
-+EXPORT_SYMBOL(caam_little_end);
-+bool caam_imx;
-+EXPORT_SYMBOL(caam_imx);
-+#endif
-+
-+/*
-+ * This is a a cache of buffers, from which the users of CAAM QI driver
-+ * can allocate short buffers. It's speedier than doing kmalloc on the hotpath.
-+ * NOTE: A more elegant solution would be to have some headroom in the frames
-+ * being processed. This can be added by the dpaa2-eth driver. This would
-+ * pose a problem for userspace application processing which cannot
-+ * know of this limitation. So for now, this will work.
-+ * NOTE: The memcache is SMP-safe. No need to handle spinlocks in-here
-+ */
-+static struct kmem_cache *qi_cache;
-+
-+struct caam_alg_entry {
-+ struct device *dev;
-+ int class1_alg_type;
-+ int class2_alg_type;
-+ bool rfc3686;
-+ bool geniv;
-+};
-+
-+struct caam_aead_alg {
-+ struct aead_alg aead;
-+ struct caam_alg_entry caam;
-+ bool registered;
-+};
-+
-+/**
-+ * caam_ctx - per-session context
-+ * @flc: Flow Contexts array
-+ * @key: virtual address of the key(s): [authentication key], encryption key
-+ * @flc_dma: I/O virtual addresses of the Flow Contexts
-+ * @key_dma: I/O virtual address of the key
-+ * @dir: DMA direction for mapping key and Flow Contexts
-+ * @dev: dpseci device
-+ * @adata: authentication algorithm details
-+ * @cdata: encryption algorithm details
-+ * @authsize: authentication tag (a.k.a. ICV / MAC) size
-+ */
-+struct caam_ctx {
-+ struct caam_flc flc[NUM_OP];
-+ u8 key[CAAM_MAX_KEY_SIZE];
-+ dma_addr_t flc_dma[NUM_OP];
-+ dma_addr_t key_dma;
-+ enum dma_data_direction dir;
-+ struct device *dev;
-+ struct alginfo adata;
-+ struct alginfo cdata;
-+ unsigned int authsize;
-+};
-+
-+void *dpaa2_caam_iova_to_virt(struct dpaa2_caam_priv *priv,
-+ dma_addr_t iova_addr)
-+{
-+ phys_addr_t phys_addr;
-+
-+ phys_addr = priv->domain ? iommu_iova_to_phys(priv->domain, iova_addr) :
-+ iova_addr;
-+
-+ return phys_to_virt(phys_addr);
-+}
-+
-+/*
-+ * qi_cache_zalloc - Allocate buffers from CAAM-QI cache
-+ *
-+ * Allocate data on the hotpath. Instead of using kzalloc, one can use the
-+ * services of the CAAM QI memory cache (backed by kmem_cache). The buffers
-+ * will have a size of CAAM_QI_MEMCACHE_SIZE, which should be sufficient for
-+ * hosting 16 SG entries.
-+ *
-+ * @flags - flags that would be used for the equivalent kmalloc(..) call
-+ *
-+ * Returns a pointer to a retrieved buffer on success or NULL on failure.
-+ */
-+static inline void *qi_cache_zalloc(gfp_t flags)
-+{
-+ return kmem_cache_zalloc(qi_cache, flags);
-+}
-+
-+/*
-+ * qi_cache_free - Frees buffers allocated from CAAM-QI cache
-+ *
-+ * @obj - buffer previously allocated by qi_cache_zalloc
-+ *
-+ * No checking is being done, the call is a passthrough call to
-+ * kmem_cache_free(...)
-+ */
-+static inline void qi_cache_free(void *obj)
-+{
-+ kmem_cache_free(qi_cache, obj);
-+}
-+
-+static struct caam_request *to_caam_req(struct crypto_async_request *areq)
-+{
-+ switch (crypto_tfm_alg_type(areq->tfm)) {
-+ case CRYPTO_ALG_TYPE_ABLKCIPHER:
-+ case CRYPTO_ALG_TYPE_GIVCIPHER:
-+ return ablkcipher_request_ctx(ablkcipher_request_cast(areq));
-+ case CRYPTO_ALG_TYPE_AEAD:
-+ return aead_request_ctx(container_of(areq, struct aead_request,
-+ base));
-+ case CRYPTO_ALG_TYPE_AHASH:
-+ return ahash_request_ctx(ahash_request_cast(areq));
-+ default:
-+ return ERR_PTR(-EINVAL);
-+ }
-+}
-+
-+static void caam_unmap(struct device *dev, struct scatterlist *src,
-+ struct scatterlist *dst, int src_nents,
-+ int dst_nents, dma_addr_t iv_dma, int ivsize,
-+ enum optype op_type, dma_addr_t qm_sg_dma,
-+ int qm_sg_bytes)
-+{
-+ if (dst != src) {
-+ if (src_nents)
-+ dma_unmap_sg(dev, src, src_nents, DMA_TO_DEVICE);
-+ dma_unmap_sg(dev, dst, dst_nents, DMA_FROM_DEVICE);
-+ } else {
-+ dma_unmap_sg(dev, src, src_nents, DMA_BIDIRECTIONAL);
-+ }
-+
-+ if (iv_dma)
-+ dma_unmap_single(dev, iv_dma, ivsize,
-+ op_type == GIVENCRYPT ? DMA_FROM_DEVICE :
-+ DMA_TO_DEVICE);
-+
-+ if (qm_sg_bytes)
-+ dma_unmap_single(dev, qm_sg_dma, qm_sg_bytes, DMA_TO_DEVICE);
-+}
-+
-+static int aead_set_sh_desc(struct crypto_aead *aead)
-+{
-+ struct caam_aead_alg *alg = container_of(crypto_aead_alg(aead),
-+ typeof(*alg), aead);
-+ struct caam_ctx *ctx = crypto_aead_ctx(aead);
-+ unsigned int ivsize = crypto_aead_ivsize(aead);
-+ struct device *dev = ctx->dev;
-+ struct dpaa2_caam_priv *priv = dev_get_drvdata(dev);
-+ struct caam_flc *flc;
-+ u32 *desc;
-+ u32 ctx1_iv_off = 0;
-+ u32 *nonce = NULL;
-+ unsigned int data_len[2];
-+ u32 inl_mask;
-+ const bool ctr_mode = ((ctx->cdata.algtype & OP_ALG_AAI_MASK) ==
-+ OP_ALG_AAI_CTR_MOD128);
-+ const bool is_rfc3686 = alg->caam.rfc3686;
-+
-+ if (!ctx->cdata.keylen || !ctx->authsize)
-+ return 0;
-+
-+ /*
-+ * AES-CTR needs to load IV in CONTEXT1 reg
-+ * at an offset of 128bits (16bytes)
-+ * CONTEXT1[255:128] = IV
-+ */
-+ if (ctr_mode)
-+ ctx1_iv_off = 16;
-+
-+ /*
-+ * RFC3686 specific:
-+ * CONTEXT1[255:128] = {NONCE, IV, COUNTER}
-+ */
-+ if (is_rfc3686) {
-+ ctx1_iv_off = 16 + CTR_RFC3686_NONCE_SIZE;
-+ nonce = (u32 *)((void *)ctx->key + ctx->adata.keylen_pad +
-+ ctx->cdata.keylen - CTR_RFC3686_NONCE_SIZE);
-+ }
-+
-+ data_len[0] = ctx->adata.keylen_pad;
-+ data_len[1] = ctx->cdata.keylen;
-+
-+ /* aead_encrypt shared descriptor */
-+ if (desc_inline_query((alg->caam.geniv ? DESC_QI_AEAD_GIVENC_LEN :
-+ DESC_QI_AEAD_ENC_LEN) +
-+ (is_rfc3686 ? DESC_AEAD_CTR_RFC3686_LEN : 0),
-+ DESC_JOB_IO_LEN, data_len, &inl_mask,
-+ ARRAY_SIZE(data_len)) < 0)
-+ return -EINVAL;
-+
-+ if (inl_mask & 1)
-+ ctx->adata.key_virt = ctx->key;
-+ else
-+ ctx->adata.key_dma = ctx->key_dma;
-+
-+ if (inl_mask & 2)
-+ ctx->cdata.key_virt = ctx->key + ctx->adata.keylen_pad;
-+ else
-+ ctx->cdata.key_dma = ctx->key_dma + ctx->adata.keylen_pad;
-+
-+ ctx->adata.key_inline = !!(inl_mask & 1);
-+ ctx->cdata.key_inline = !!(inl_mask & 2);
-+
-+ flc = &ctx->flc[ENCRYPT];
-+ desc = flc->sh_desc;
-+
-+ if (alg->caam.geniv)
-+ cnstr_shdsc_aead_givencap(desc, &ctx->cdata, &ctx->adata,
-+ ivsize, ctx->authsize, is_rfc3686,
-+ nonce, ctx1_iv_off, true,
-+ priv->sec_attr.era);
-+ else
-+ cnstr_shdsc_aead_encap(desc, &ctx->cdata, &ctx->adata,
-+ ivsize, ctx->authsize, is_rfc3686, nonce,
-+ ctx1_iv_off, true, priv->sec_attr.era);
-+
-+ flc->flc[1] = cpu_to_caam32(desc_len(desc)); /* SDL */
-+ dma_sync_single_for_device(dev, ctx->flc_dma[ENCRYPT],
-+ sizeof(flc->flc) + desc_bytes(desc),
-+ ctx->dir);
-+
-+ /* aead_decrypt shared descriptor */
-+ if (desc_inline_query(DESC_QI_AEAD_DEC_LEN +
-+ (is_rfc3686 ? DESC_AEAD_CTR_RFC3686_LEN : 0),
-+ DESC_JOB_IO_LEN, data_len, &inl_mask,
-+ ARRAY_SIZE(data_len)) < 0)
-+ return -EINVAL;
-+
-+ if (inl_mask & 1)
-+ ctx->adata.key_virt = ctx->key;
-+ else
-+ ctx->adata.key_dma = ctx->key_dma;
-+
-+ if (inl_mask & 2)
-+ ctx->cdata.key_virt = ctx->key + ctx->adata.keylen_pad;
-+ else
-+ ctx->cdata.key_dma = ctx->key_dma + ctx->adata.keylen_pad;
-+
-+ ctx->adata.key_inline = !!(inl_mask & 1);
-+ ctx->cdata.key_inline = !!(inl_mask & 2);
-+
-+ flc = &ctx->flc[DECRYPT];
-+ desc = flc->sh_desc;
-+ cnstr_shdsc_aead_decap(desc, &ctx->cdata, &ctx->adata,
-+ ivsize, ctx->authsize, alg->caam.geniv,
-+ is_rfc3686, nonce, ctx1_iv_off, true,
-+ priv->sec_attr.era);
-+ flc->flc[1] = cpu_to_caam32(desc_len(desc)); /* SDL */
-+ dma_sync_single_for_device(dev, ctx->flc_dma[DECRYPT],
-+ sizeof(flc->flc) + desc_bytes(desc),
-+ ctx->dir);
-+
-+ return 0;
-+}
-+
-+static int aead_setauthsize(struct crypto_aead *authenc, unsigned int authsize)
-+{
-+ struct caam_ctx *ctx = crypto_aead_ctx(authenc);
-+
-+ ctx->authsize = authsize;
-+ aead_set_sh_desc(authenc);
-+
-+ return 0;
-+}
-+
-+struct split_key_sh_result {
-+ struct completion completion;
-+ int err;
-+ struct device *dev;
-+};
-+
-+static void split_key_sh_done(void *cbk_ctx, u32 err)
-+{
-+ struct split_key_sh_result *res = cbk_ctx;
-+
-+#ifdef DEBUG
-+ dev_err(res->dev, "%s %d: err 0x%x\n", __func__, __LINE__, err);
-+#endif
-+
-+ if (err)
-+ caam_qi2_strstatus(res->dev, err);
-+
-+ res->err = err;
-+ complete(&res->completion);
-+}
-+
-+static int aead_setkey(struct crypto_aead *aead, const u8 *key,
-+ unsigned int keylen)
-+{
-+ struct caam_ctx *ctx = crypto_aead_ctx(aead);
-+ struct device *dev = ctx->dev;
-+ struct crypto_authenc_keys keys;
-+
-+ if (crypto_authenc_extractkeys(&keys, key, keylen) != 0)
-+ goto badkey;
-+
-+#ifdef DEBUG
-+ dev_err(dev, "keylen %d enckeylen %d authkeylen %d\n",
-+ keys.authkeylen + keys.enckeylen, keys.enckeylen,
-+ keys.authkeylen);
-+ print_hex_dump(KERN_ERR, "key in @" __stringify(__LINE__)": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, key, keylen, 1);
-+#endif
-+
-+ ctx->adata.keylen = keys.authkeylen;
-+ ctx->adata.keylen_pad = split_key_len(ctx->adata.algtype &
-+ OP_ALG_ALGSEL_MASK);
-+
-+ if (ctx->adata.keylen_pad + keys.enckeylen > CAAM_MAX_KEY_SIZE)
-+ goto badkey;
-+
-+ memcpy(ctx->key, keys.authkey, keys.authkeylen);
-+ memcpy(ctx->key + ctx->adata.keylen_pad, keys.enckey, keys.enckeylen);
-+ dma_sync_single_for_device(dev, ctx->key_dma, ctx->adata.keylen_pad +
-+ keys.enckeylen, ctx->dir);
-+#ifdef DEBUG
-+ print_hex_dump(KERN_ERR, "ctx.key@" __stringify(__LINE__)": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, ctx->key,
-+ ctx->adata.keylen_pad + keys.enckeylen, 1);
-+#endif
-+
-+ ctx->cdata.keylen = keys.enckeylen;
-+
-+ return aead_set_sh_desc(aead);
-+badkey:
-+ crypto_aead_set_flags(aead, CRYPTO_TFM_RES_BAD_KEY_LEN);
-+ return -EINVAL;
-+}
-+
-+static struct aead_edesc *aead_edesc_alloc(struct aead_request *req,
-+ bool encrypt)
-+{
-+ struct crypto_aead *aead = crypto_aead_reqtfm(req);
-+ struct caam_request *req_ctx = aead_request_ctx(req);
-+ struct dpaa2_fl_entry *in_fle = &req_ctx->fd_flt[1];
-+ struct dpaa2_fl_entry *out_fle = &req_ctx->fd_flt[0];
-+ struct caam_ctx *ctx = crypto_aead_ctx(aead);
-+ struct caam_aead_alg *alg = container_of(crypto_aead_alg(aead),
-+ typeof(*alg), aead);
-+ struct device *dev = ctx->dev;
-+ gfp_t flags = (req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP) ?
-+ GFP_KERNEL : GFP_ATOMIC;
-+ int src_nents, mapped_src_nents, dst_nents = 0, mapped_dst_nents = 0;
-+ struct aead_edesc *edesc;
-+ dma_addr_t qm_sg_dma, iv_dma = 0;
-+ int ivsize = 0;
-+ unsigned int authsize = ctx->authsize;
-+ int qm_sg_index = 0, qm_sg_nents = 0, qm_sg_bytes;
-+ int in_len, out_len;
-+ struct dpaa2_sg_entry *sg_table;
-+ enum optype op_type = encrypt ? ENCRYPT : DECRYPT;
-+
-+ /* allocate space for base edesc and link tables */
-+ edesc = qi_cache_zalloc(GFP_DMA | flags);
-+ if (unlikely(!edesc)) {
-+ dev_err(dev, "could not allocate extended descriptor\n");
-+ return ERR_PTR(-ENOMEM);
-+ }
-+
-+ if (unlikely(req->dst != req->src)) {
-+ src_nents = sg_nents_for_len(req->src, req->assoclen +
-+ req->cryptlen);
-+ if (unlikely(src_nents < 0)) {
-+ dev_err(dev, "Insufficient bytes (%d) in src S/G\n",
-+ req->assoclen + req->cryptlen);
-+ qi_cache_free(edesc);
-+ return ERR_PTR(src_nents);
-+ }
-+
-+ dst_nents = sg_nents_for_len(req->dst, req->assoclen +
-+ req->cryptlen +
-+ (encrypt ? authsize :
-+ (-authsize)));
-+ if (unlikely(dst_nents < 0)) {
-+ dev_err(dev, "Insufficient bytes (%d) in dst S/G\n",
-+ req->assoclen + req->cryptlen +
-+ (encrypt ? authsize : (-authsize)));
-+ qi_cache_free(edesc);
-+ return ERR_PTR(dst_nents);
-+ }
-+
-+ if (src_nents) {
-+ mapped_src_nents = dma_map_sg(dev, req->src, src_nents,
-+ DMA_TO_DEVICE);
-+ if (unlikely(!mapped_src_nents)) {
-+ dev_err(dev, "unable to map source\n");
-+ qi_cache_free(edesc);
-+ return ERR_PTR(-ENOMEM);
-+ }
-+ } else {
-+ mapped_src_nents = 0;
-+ }
-+
-+ mapped_dst_nents = dma_map_sg(dev, req->dst, dst_nents,
-+ DMA_FROM_DEVICE);
-+ if (unlikely(!mapped_dst_nents)) {
-+ dev_err(dev, "unable to map destination\n");
-+ dma_unmap_sg(dev, req->src, src_nents, DMA_TO_DEVICE);
-+ qi_cache_free(edesc);
-+ return ERR_PTR(-ENOMEM);
-+ }
-+ } else {
-+ src_nents = sg_nents_for_len(req->src, req->assoclen +
-+ req->cryptlen +
-+ (encrypt ? authsize : 0));
-+ if (unlikely(src_nents < 0)) {
-+ dev_err(dev, "Insufficient bytes (%d) in src S/G\n",
-+ req->assoclen + req->cryptlen +
-+ (encrypt ? authsize : 0));
-+ qi_cache_free(edesc);
-+ return ERR_PTR(src_nents);
-+ }
-+
-+ mapped_src_nents = dma_map_sg(dev, req->src, src_nents,
-+ DMA_BIDIRECTIONAL);
-+ if (unlikely(!mapped_src_nents)) {
-+ dev_err(dev, "unable to map source\n");
-+ qi_cache_free(edesc);
-+ return ERR_PTR(-ENOMEM);
-+ }
-+ }
-+
-+ if ((alg->caam.rfc3686 && encrypt) || !alg->caam.geniv) {
-+ ivsize = crypto_aead_ivsize(aead);
-+ iv_dma = dma_map_single(dev, req->iv, ivsize, DMA_TO_DEVICE);
-+ if (dma_mapping_error(dev, iv_dma)) {
-+ dev_err(dev, "unable to map IV\n");
-+ caam_unmap(dev, req->src, req->dst, src_nents,
-+ dst_nents, 0, 0, op_type, 0, 0);
-+ qi_cache_free(edesc);
-+ return ERR_PTR(-ENOMEM);
-+ }
-+ }
-+
-+ /*
-+ * Create S/G table: req->assoclen, [IV,] req->src [, req->dst].
-+ * Input is not contiguous.
-+ */
-+ qm_sg_nents = 1 + !!ivsize + mapped_src_nents +
-+ (mapped_dst_nents > 1 ? mapped_dst_nents : 0);
-+ if (unlikely(qm_sg_nents > CAAM_QI_MAX_AEAD_SG)) {
-+ dev_err(dev, "Insufficient S/G entries: %d > %lu\n",
-+ qm_sg_nents, CAAM_QI_MAX_AEAD_SG);
-+ caam_unmap(dev, req->src, req->dst, src_nents, dst_nents,
-+ iv_dma, ivsize, op_type, 0, 0);
-+ qi_cache_free(edesc);
-+ return ERR_PTR(-ENOMEM);
-+ }
-+ sg_table = &edesc->sgt[0];
-+ qm_sg_bytes = qm_sg_nents * sizeof(*sg_table);
-+
-+ edesc->src_nents = src_nents;
-+ edesc->dst_nents = dst_nents;
-+ edesc->iv_dma = iv_dma;
-+
-+ edesc->assoclen = cpu_to_caam32(req->assoclen);
-+ edesc->assoclen_dma = dma_map_single(dev, &edesc->assoclen, 4,
-+ DMA_TO_DEVICE);
-+ if (dma_mapping_error(dev, edesc->assoclen_dma)) {
-+ dev_err(dev, "unable to map assoclen\n");
-+ caam_unmap(dev, req->src, req->dst, src_nents, dst_nents,
-+ iv_dma, ivsize, op_type, 0, 0);
-+ qi_cache_free(edesc);
-+ return ERR_PTR(-ENOMEM);
-+ }
-+
-+ dma_to_qm_sg_one(sg_table, edesc->assoclen_dma, 4, 0);
-+ qm_sg_index++;
-+ if (ivsize) {
-+ dma_to_qm_sg_one(sg_table + qm_sg_index, iv_dma, ivsize, 0);
-+ qm_sg_index++;
-+ }
-+ sg_to_qm_sg_last(req->src, mapped_src_nents, sg_table + qm_sg_index, 0);
-+ qm_sg_index += mapped_src_nents;
-+
-+ if (mapped_dst_nents > 1)
-+ sg_to_qm_sg_last(req->dst, mapped_dst_nents, sg_table +
-+ qm_sg_index, 0);
-+
-+ qm_sg_dma = dma_map_single(dev, sg_table, qm_sg_bytes, DMA_TO_DEVICE);
-+ if (dma_mapping_error(dev, qm_sg_dma)) {
-+ dev_err(dev, "unable to map S/G table\n");
-+ dma_unmap_single(dev, edesc->assoclen_dma, 4, DMA_TO_DEVICE);
-+ caam_unmap(dev, req->src, req->dst, src_nents, dst_nents,
-+ iv_dma, ivsize, op_type, 0, 0);
-+ qi_cache_free(edesc);
-+ return ERR_PTR(-ENOMEM);
-+ }
-+
-+ edesc->qm_sg_dma = qm_sg_dma;
-+ edesc->qm_sg_bytes = qm_sg_bytes;
-+
-+ out_len = req->assoclen + req->cryptlen +
-+ (encrypt ? ctx->authsize : (-ctx->authsize));
-+ in_len = 4 + ivsize + req->assoclen + req->cryptlen;
-+
-+ memset(&req_ctx->fd_flt, 0, sizeof(req_ctx->fd_flt));
-+ dpaa2_fl_set_final(in_fle, true);
-+ dpaa2_fl_set_format(in_fle, dpaa2_fl_sg);
-+ dpaa2_fl_set_addr(in_fle, qm_sg_dma);
-+ dpaa2_fl_set_len(in_fle, in_len);
-+
-+ if (req->dst == req->src) {
-+ if (mapped_src_nents == 1) {
-+ dpaa2_fl_set_format(out_fle, dpaa2_fl_single);
-+ dpaa2_fl_set_addr(out_fle, sg_dma_address(req->src));
-+ } else {
-+ dpaa2_fl_set_format(out_fle, dpaa2_fl_sg);
-+ dpaa2_fl_set_addr(out_fle, qm_sg_dma +
-+ (1 + !!ivsize) * sizeof(*sg_table));
-+ }
-+ } else if (mapped_dst_nents == 1) {
-+ dpaa2_fl_set_format(out_fle, dpaa2_fl_single);
-+ dpaa2_fl_set_addr(out_fle, sg_dma_address(req->dst));
-+ } else {
-+ dpaa2_fl_set_format(out_fle, dpaa2_fl_sg);
-+ dpaa2_fl_set_addr(out_fle, qm_sg_dma + qm_sg_index *
-+ sizeof(*sg_table));
-+ }
-+
-+ dpaa2_fl_set_len(out_fle, out_len);
-+
-+ return edesc;
-+}
-+
-+static struct tls_edesc *tls_edesc_alloc(struct aead_request *req,
-+ bool encrypt)
-+{
-+ struct crypto_aead *tls = crypto_aead_reqtfm(req);
-+ unsigned int blocksize = crypto_aead_blocksize(tls);
-+ unsigned int padsize, authsize;
-+ struct caam_request *req_ctx = aead_request_ctx(req);
-+ struct dpaa2_fl_entry *in_fle = &req_ctx->fd_flt[1];
-+ struct dpaa2_fl_entry *out_fle = &req_ctx->fd_flt[0];
-+ struct caam_ctx *ctx = crypto_aead_ctx(tls);
-+ struct caam_aead_alg *alg = container_of(crypto_aead_alg(tls),
-+ typeof(*alg), aead);
-+ struct device *dev = ctx->dev;
-+ gfp_t flags = (req->base.flags & (CRYPTO_TFM_REQ_MAY_BACKLOG |
-+ CRYPTO_TFM_REQ_MAY_SLEEP)) ? GFP_KERNEL : GFP_ATOMIC;
-+ int src_nents, mapped_src_nents, dst_nents = 0, mapped_dst_nents = 0;
-+ struct tls_edesc *edesc;
-+ dma_addr_t qm_sg_dma, iv_dma = 0;
-+ int ivsize = 0;
-+ int qm_sg_index, qm_sg_ents = 0, qm_sg_bytes;
-+ int in_len, out_len;
-+ struct dpaa2_sg_entry *sg_table;
-+ enum optype op_type = encrypt ? ENCRYPT : DECRYPT;
-+ struct scatterlist *dst;
-+
-+ if (encrypt) {
-+ padsize = blocksize - ((req->cryptlen + ctx->authsize) %
-+ blocksize);
-+ authsize = ctx->authsize + padsize;
-+ } else {
-+ authsize = ctx->authsize;
-+ }
-+
-+ /* allocate space for base edesc and link tables */
-+ edesc = qi_cache_zalloc(GFP_DMA | flags);
-+ if (unlikely(!edesc)) {
-+ dev_err(dev, "could not allocate extended descriptor\n");
-+ return ERR_PTR(-ENOMEM);
-+ }
-+
-+ if (likely(req->src == req->dst)) {
-+ src_nents = sg_nents_for_len(req->src, req->assoclen +
-+ req->cryptlen +
-+ (encrypt ? authsize : 0));
-+ if (unlikely(src_nents < 0)) {
-+ dev_err(dev, "Insufficient bytes (%d) in src S/G\n",
-+ req->assoclen + req->cryptlen +
-+ (encrypt ? authsize : 0));
-+ qi_cache_free(edesc);
-+ return ERR_PTR(src_nents);
-+ }
-+
-+ mapped_src_nents = dma_map_sg(dev, req->src, src_nents,
-+ DMA_BIDIRECTIONAL);
-+ if (unlikely(!mapped_src_nents)) {
-+ dev_err(dev, "unable to map source\n");
-+ qi_cache_free(edesc);
-+ return ERR_PTR(-ENOMEM);
-+ }
-+ dst = req->dst;
-+ } else {
-+ src_nents = sg_nents_for_len(req->src, req->assoclen +
-+ req->cryptlen);
-+ if (unlikely(src_nents < 0)) {
-+ dev_err(dev, "Insufficient bytes (%d) in src S/G\n",
-+ req->assoclen + req->cryptlen);
-+ qi_cache_free(edesc);
-+ return ERR_PTR(src_nents);
-+ }
-+
-+ dst = scatterwalk_ffwd(edesc->tmp, req->dst, req->assoclen);
-+ dst_nents = sg_nents_for_len(dst, req->cryptlen +
-+ (encrypt ? authsize : 0));
-+ if (unlikely(dst_nents < 0)) {
-+ dev_err(dev, "Insufficient bytes (%d) in dst S/G\n",
-+ req->cryptlen +
-+ (encrypt ? authsize : 0));
-+ qi_cache_free(edesc);
-+ return ERR_PTR(dst_nents);
-+ }
-+
-+ if (src_nents) {
-+ mapped_src_nents = dma_map_sg(dev, req->src,
-+ src_nents, DMA_TO_DEVICE);
-+ if (unlikely(!mapped_src_nents)) {
-+ dev_err(dev, "unable to map source\n");
-+ qi_cache_free(edesc);
-+ return ERR_PTR(-ENOMEM);
-+ }
-+ } else {
-+ mapped_src_nents = 0;
-+ }
-+
-+ mapped_dst_nents = dma_map_sg(dev, dst, dst_nents,
-+ DMA_FROM_DEVICE);
-+ if (unlikely(!mapped_dst_nents)) {
-+ dev_err(dev, "unable to map destination\n");
-+ dma_unmap_sg(dev, req->src, src_nents, DMA_TO_DEVICE);
-+ qi_cache_free(edesc);
-+ return ERR_PTR(-ENOMEM);
-+ }
-+ }
-+
-+ ivsize = crypto_aead_ivsize(tls);
-+ iv_dma = dma_map_single(dev, req->iv, ivsize, DMA_TO_DEVICE);
-+ if (dma_mapping_error(dev, iv_dma)) {
-+ dev_err(dev, "unable to map IV\n");
-+ caam_unmap(dev, req->src, dst, src_nents, dst_nents, 0, 0,
-+ op_type, 0, 0);
-+ qi_cache_free(edesc);
-+ return ERR_PTR(-ENOMEM);
-+ }
-+
-+ /*
-+ * Create S/G table: IV, src, dst.
-+ * Input is not contiguous.
-+ */
-+ qm_sg_ents = 1 + mapped_src_nents +
-+ (mapped_dst_nents > 1 ? mapped_dst_nents : 0);
-+ sg_table = &edesc->sgt[0];
-+ qm_sg_bytes = qm_sg_ents * sizeof(*sg_table);
-+
-+ edesc->src_nents = src_nents;
-+ edesc->dst_nents = dst_nents;
-+ edesc->dst = dst;
-+ edesc->iv_dma = iv_dma;
-+
-+ dma_to_qm_sg_one(sg_table, iv_dma, ivsize, 0);
-+ qm_sg_index = 1;
-+
-+ sg_to_qm_sg_last(req->src, mapped_src_nents, sg_table + qm_sg_index, 0);
-+ qm_sg_index += mapped_src_nents;
-+
-+ if (mapped_dst_nents > 1)
-+ sg_to_qm_sg_last(dst, mapped_dst_nents, sg_table +
-+ qm_sg_index, 0);
-+
-+ qm_sg_dma = dma_map_single(dev, sg_table, qm_sg_bytes, DMA_TO_DEVICE);
-+ if (dma_mapping_error(dev, qm_sg_dma)) {
-+ dev_err(dev, "unable to map S/G table\n");
-+ caam_unmap(dev, req->src, dst, src_nents, dst_nents, iv_dma,
-+ ivsize, op_type, 0, 0);
-+ qi_cache_free(edesc);
-+ return ERR_PTR(-ENOMEM);
-+ }
-+
-+ edesc->qm_sg_dma = qm_sg_dma;
-+ edesc->qm_sg_bytes = qm_sg_bytes;
-+
-+ out_len = req->cryptlen + (encrypt ? authsize : 0);
-+ in_len = ivsize + req->assoclen + req->cryptlen;
-+
-+ memset(&req_ctx->fd_flt, 0, sizeof(req_ctx->fd_flt));
-+ dpaa2_fl_set_final(in_fle, true);
-+ dpaa2_fl_set_format(in_fle, dpaa2_fl_sg);
-+ dpaa2_fl_set_addr(in_fle, qm_sg_dma);
-+ dpaa2_fl_set_len(in_fle, in_len);
-+
-+ if (req->dst == req->src) {
-+ dpaa2_fl_set_format(out_fle, dpaa2_fl_sg);
-+ dpaa2_fl_set_addr(out_fle, qm_sg_dma +
-+ (sg_nents_for_len(req->src, req->assoclen) +
-+ 1) * sizeof(*sg_table));
-+ } else if (mapped_dst_nents == 1) {
-+ dpaa2_fl_set_format(out_fle, dpaa2_fl_single);
-+ dpaa2_fl_set_addr(out_fle, sg_dma_address(dst));
-+ } else {
-+ dpaa2_fl_set_format(out_fle, dpaa2_fl_sg);
-+ dpaa2_fl_set_addr(out_fle, qm_sg_dma + qm_sg_index *
-+ sizeof(*sg_table));
-+ }
-+
-+ dpaa2_fl_set_len(out_fle, out_len);
-+
-+ return edesc;
-+}
-+
-+static int tls_set_sh_desc(struct crypto_aead *tls)
-+{
-+ struct caam_ctx *ctx = crypto_aead_ctx(tls);
-+ unsigned int ivsize = crypto_aead_ivsize(tls);
-+ unsigned int blocksize = crypto_aead_blocksize(tls);
-+ struct device *dev = ctx->dev;
-+ struct dpaa2_caam_priv *priv = dev_get_drvdata(dev);
-+ struct caam_flc *flc;
-+ u32 *desc;
-+ unsigned int assoclen = 13; /* always 13 bytes for TLS */
-+ unsigned int data_len[2];
-+ u32 inl_mask;
-+
-+ if (!ctx->cdata.keylen || !ctx->authsize)
-+ return 0;
-+
-+ /*
-+ * TLS 1.0 encrypt shared descriptor
-+ * Job Descriptor and Shared Descriptor
-+ * must fit into the 64-word Descriptor h/w Buffer
-+ */
-+ data_len[0] = ctx->adata.keylen_pad;
-+ data_len[1] = ctx->cdata.keylen;
-+
-+ if (desc_inline_query(DESC_TLS10_ENC_LEN, DESC_JOB_IO_LEN, data_len,
-+ &inl_mask, ARRAY_SIZE(data_len)) < 0)
-+ return -EINVAL;
-+
-+ if (inl_mask & 1)
-+ ctx->adata.key_virt = ctx->key;
-+ else
-+ ctx->adata.key_dma = ctx->key_dma;
-+
-+ if (inl_mask & 2)
-+ ctx->cdata.key_virt = ctx->key + ctx->adata.keylen_pad;
-+ else
-+ ctx->cdata.key_dma = ctx->key_dma + ctx->adata.keylen_pad;
-+
-+ ctx->adata.key_inline = !!(inl_mask & 1);
-+ ctx->cdata.key_inline = !!(inl_mask & 2);
-+
-+ flc = &ctx->flc[ENCRYPT];
-+ desc = flc->sh_desc;
-+ cnstr_shdsc_tls_encap(desc, &ctx->cdata, &ctx->adata,
-+ assoclen, ivsize, ctx->authsize, blocksize,
-+ priv->sec_attr.era);
-+ flc->flc[1] = cpu_to_caam32(desc_len(desc));
-+ dma_sync_single_for_device(dev, ctx->flc_dma[ENCRYPT],
-+ sizeof(flc->flc) + desc_bytes(desc),
-+ ctx->dir);
-+
-+ /*
-+ * TLS 1.0 decrypt shared descriptor
-+ * Keys do not fit inline, regardless of algorithms used
-+ */
-+ ctx->adata.key_inline = false;
-+ ctx->adata.key_dma = ctx->key_dma;
-+ ctx->cdata.key_dma = ctx->key_dma + ctx->adata.keylen_pad;
-+
-+ flc = &ctx->flc[DECRYPT];
-+ desc = flc->sh_desc;
-+ cnstr_shdsc_tls_decap(desc, &ctx->cdata, &ctx->adata, assoclen, ivsize,
-+ ctx->authsize, blocksize, priv->sec_attr.era);
-+ flc->flc[1] = cpu_to_caam32(desc_len(desc)); /* SDL */
-+ dma_sync_single_for_device(dev, ctx->flc_dma[DECRYPT],
-+ sizeof(flc->flc) + desc_bytes(desc),
-+ ctx->dir);
-+
-+ return 0;
-+}
-+
-+static int tls_setkey(struct crypto_aead *tls, const u8 *key,
-+ unsigned int keylen)
-+{
-+ struct caam_ctx *ctx = crypto_aead_ctx(tls);
-+ struct device *dev = ctx->dev;
-+ struct crypto_authenc_keys keys;
-+
-+ if (crypto_authenc_extractkeys(&keys, key, keylen) != 0)
-+ goto badkey;
-+
-+#ifdef DEBUG
-+ dev_err(dev, "keylen %d enckeylen %d authkeylen %d\n",
-+ keys.authkeylen + keys.enckeylen, keys.enckeylen,
-+ keys.authkeylen);
-+ print_hex_dump(KERN_ERR, "key in @" __stringify(__LINE__)": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, key, keylen, 1);
-+#endif
-+
-+ ctx->adata.keylen = keys.authkeylen;
-+ ctx->adata.keylen_pad = split_key_len(ctx->adata.algtype &
-+ OP_ALG_ALGSEL_MASK);
-+
-+ if (ctx->adata.keylen_pad + keys.enckeylen > CAAM_MAX_KEY_SIZE)
-+ goto badkey;
-+
-+ memcpy(ctx->key, keys.authkey, keys.authkeylen);
-+ memcpy(ctx->key + ctx->adata.keylen_pad, keys.enckey, keys.enckeylen);
-+ dma_sync_single_for_device(dev, ctx->key_dma, ctx->adata.keylen_pad +
-+ keys.enckeylen, ctx->dir);
-+#ifdef DEBUG
-+ print_hex_dump(KERN_ERR, "ctx.key@" __stringify(__LINE__)": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, ctx->key,
-+ ctx->adata.keylen_pad + keys.enckeylen, 1);
-+#endif
-+
-+ ctx->cdata.keylen = keys.enckeylen;
-+
-+ return tls_set_sh_desc(tls);
-+badkey:
-+ crypto_aead_set_flags(tls, CRYPTO_TFM_RES_BAD_KEY_LEN);
-+ return -EINVAL;
-+}
-+
-+static int tls_setauthsize(struct crypto_aead *tls, unsigned int authsize)
-+{
-+ struct caam_ctx *ctx = crypto_aead_ctx(tls);
-+
-+ ctx->authsize = authsize;
-+ tls_set_sh_desc(tls);
-+
-+ return 0;
-+}
-+
-+static int gcm_set_sh_desc(struct crypto_aead *aead)
-+{
-+ struct caam_ctx *ctx = crypto_aead_ctx(aead);
-+ struct device *dev = ctx->dev;
-+ unsigned int ivsize = crypto_aead_ivsize(aead);
-+ struct caam_flc *flc;
-+ u32 *desc;
-+ int rem_bytes = CAAM_DESC_BYTES_MAX - DESC_JOB_IO_LEN -
-+ ctx->cdata.keylen;
-+
-+ if (!ctx->cdata.keylen || !ctx->authsize)
-+ return 0;
-+
-+ /*
-+ * AES GCM encrypt shared descriptor
-+ * Job Descriptor and Shared Descriptor
-+ * must fit into the 64-word Descriptor h/w Buffer
-+ */
-+ if (rem_bytes >= DESC_QI_GCM_ENC_LEN) {
-+ ctx->cdata.key_inline = true;
-+ ctx->cdata.key_virt = ctx->key;
-+ } else {
-+ ctx->cdata.key_inline = false;
-+ ctx->cdata.key_dma = ctx->key_dma;
-+ }
-+
-+ flc = &ctx->flc[ENCRYPT];
-+ desc = flc->sh_desc;
-+ cnstr_shdsc_gcm_encap(desc, &ctx->cdata, ivsize, ctx->authsize, true);
-+ flc->flc[1] = cpu_to_caam32(desc_len(desc)); /* SDL */
-+ dma_sync_single_for_device(dev, ctx->flc_dma[ENCRYPT],
-+ sizeof(flc->flc) + desc_bytes(desc),
-+ ctx->dir);
-+
-+ /*
-+ * Job Descriptor and Shared Descriptors
-+ * must all fit into the 64-word Descriptor h/w Buffer
-+ */
-+ if (rem_bytes >= DESC_QI_GCM_DEC_LEN) {
-+ ctx->cdata.key_inline = true;
-+ ctx->cdata.key_virt = ctx->key;
-+ } else {
-+ ctx->cdata.key_inline = false;
-+ ctx->cdata.key_dma = ctx->key_dma;
-+ }
-+
-+ flc = &ctx->flc[DECRYPT];
-+ desc = flc->sh_desc;
-+ cnstr_shdsc_gcm_decap(desc, &ctx->cdata, ivsize, ctx->authsize, true);
-+ flc->flc[1] = cpu_to_caam32(desc_len(desc)); /* SDL */
-+ dma_sync_single_for_device(dev, ctx->flc_dma[DECRYPT],
-+ sizeof(flc->flc) + desc_bytes(desc),
-+ ctx->dir);
-+
-+ return 0;
-+}
-+
-+static int gcm_setauthsize(struct crypto_aead *authenc, unsigned int authsize)
-+{
-+ struct caam_ctx *ctx = crypto_aead_ctx(authenc);
-+
-+ ctx->authsize = authsize;
-+ gcm_set_sh_desc(authenc);
-+
-+ return 0;
-+}
-+
-+static int gcm_setkey(struct crypto_aead *aead,
-+ const u8 *key, unsigned int keylen)
-+{
-+ struct caam_ctx *ctx = crypto_aead_ctx(aead);
-+ struct device *dev = ctx->dev;
-+
-+#ifdef DEBUG
-+ print_hex_dump(KERN_ERR, "key in @" __stringify(__LINE__)": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, key, keylen, 1);
-+#endif
-+
-+ memcpy(ctx->key, key, keylen);
-+ dma_sync_single_for_device(dev, ctx->key_dma, keylen, ctx->dir);
-+ ctx->cdata.keylen = keylen;
-+
-+ return gcm_set_sh_desc(aead);
-+}
-+
-+static int rfc4106_set_sh_desc(struct crypto_aead *aead)
-+{
-+ struct caam_ctx *ctx = crypto_aead_ctx(aead);
-+ struct device *dev = ctx->dev;
-+ unsigned int ivsize = crypto_aead_ivsize(aead);
-+ struct caam_flc *flc;
-+ u32 *desc;
-+ int rem_bytes = CAAM_DESC_BYTES_MAX - DESC_JOB_IO_LEN -
-+ ctx->cdata.keylen;
-+
-+ if (!ctx->cdata.keylen || !ctx->authsize)
-+ return 0;
-+
-+ ctx->cdata.key_virt = ctx->key;
-+
-+ /*
-+ * RFC4106 encrypt shared descriptor
-+ * Job Descriptor and Shared Descriptor
-+ * must fit into the 64-word Descriptor h/w Buffer
-+ */
-+ if (rem_bytes >= DESC_QI_RFC4106_ENC_LEN) {
-+ ctx->cdata.key_inline = true;
-+ } else {
-+ ctx->cdata.key_inline = false;
-+ ctx->cdata.key_dma = ctx->key_dma;
-+ }
-+
-+ flc = &ctx->flc[ENCRYPT];
-+ desc = flc->sh_desc;
-+ cnstr_shdsc_rfc4106_encap(desc, &ctx->cdata, ivsize, ctx->authsize,
-+ true);
-+ flc->flc[1] = cpu_to_caam32(desc_len(desc)); /* SDL */
-+ dma_sync_single_for_device(dev, ctx->flc_dma[ENCRYPT],
-+ sizeof(flc->flc) + desc_bytes(desc),
-+ ctx->dir);
-+
-+ /*
-+ * Job Descriptor and Shared Descriptors
-+ * must all fit into the 64-word Descriptor h/w Buffer
-+ */
-+ if (rem_bytes >= DESC_QI_RFC4106_DEC_LEN) {
-+ ctx->cdata.key_inline = true;
-+ } else {
-+ ctx->cdata.key_inline = false;
-+ ctx->cdata.key_dma = ctx->key_dma;
-+ }
-+
-+ flc = &ctx->flc[DECRYPT];
-+ desc = flc->sh_desc;
-+ cnstr_shdsc_rfc4106_decap(desc, &ctx->cdata, ivsize, ctx->authsize,
-+ true);
-+ flc->flc[1] = cpu_to_caam32(desc_len(desc)); /* SDL */
-+ dma_sync_single_for_device(dev, ctx->flc_dma[DECRYPT],
-+ sizeof(flc->flc) + desc_bytes(desc),
-+ ctx->dir);
-+
-+ return 0;
-+}
-+
-+static int rfc4106_setauthsize(struct crypto_aead *authenc,
-+ unsigned int authsize)
-+{
-+ struct caam_ctx *ctx = crypto_aead_ctx(authenc);
-+
-+ ctx->authsize = authsize;
-+ rfc4106_set_sh_desc(authenc);
-+
-+ return 0;
-+}
-+
-+static int rfc4106_setkey(struct crypto_aead *aead,
-+ const u8 *key, unsigned int keylen)
-+{
-+ struct caam_ctx *ctx = crypto_aead_ctx(aead);
-+ struct device *dev = ctx->dev;
-+
-+ if (keylen < 4)
-+ return -EINVAL;
-+
-+#ifdef DEBUG
-+ print_hex_dump(KERN_ERR, "key in @" __stringify(__LINE__)": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, key, keylen, 1);
-+#endif
-+
-+ memcpy(ctx->key, key, keylen);
-+ /*
-+ * The last four bytes of the key material are used as the salt value
-+ * in the nonce. Update the AES key length.
-+ */
-+ ctx->cdata.keylen = keylen - 4;
-+ dma_sync_single_for_device(dev, ctx->key_dma, ctx->cdata.keylen,
-+ ctx->dir);
-+
-+ return rfc4106_set_sh_desc(aead);
-+}
-+
-+static int rfc4543_set_sh_desc(struct crypto_aead *aead)
-+{
-+ struct caam_ctx *ctx = crypto_aead_ctx(aead);
-+ struct device *dev = ctx->dev;
-+ unsigned int ivsize = crypto_aead_ivsize(aead);
-+ struct caam_flc *flc;
-+ u32 *desc;
-+ int rem_bytes = CAAM_DESC_BYTES_MAX - DESC_JOB_IO_LEN -
-+ ctx->cdata.keylen;
-+
-+ if (!ctx->cdata.keylen || !ctx->authsize)
-+ return 0;
-+
-+ ctx->cdata.key_virt = ctx->key;
-+
-+ /*
-+ * RFC4543 encrypt shared descriptor
-+ * Job Descriptor and Shared Descriptor
-+ * must fit into the 64-word Descriptor h/w Buffer
-+ */
-+ if (rem_bytes >= DESC_QI_RFC4543_ENC_LEN) {
-+ ctx->cdata.key_inline = true;
-+ } else {
-+ ctx->cdata.key_inline = false;
-+ ctx->cdata.key_dma = ctx->key_dma;
-+ }
-+
-+ flc = &ctx->flc[ENCRYPT];
-+ desc = flc->sh_desc;
-+ cnstr_shdsc_rfc4543_encap(desc, &ctx->cdata, ivsize, ctx->authsize,
-+ true);
-+ flc->flc[1] = cpu_to_caam32(desc_len(desc)); /* SDL */
-+ dma_sync_single_for_device(dev, ctx->flc_dma[ENCRYPT],
-+ sizeof(flc->flc) + desc_bytes(desc),
-+ ctx->dir);
-+
-+ /*
-+ * Job Descriptor and Shared Descriptors
-+ * must all fit into the 64-word Descriptor h/w Buffer
-+ */
-+ if (rem_bytes >= DESC_QI_RFC4543_DEC_LEN) {
-+ ctx->cdata.key_inline = true;
-+ } else {
-+ ctx->cdata.key_inline = false;
-+ ctx->cdata.key_dma = ctx->key_dma;
-+ }
-+
-+ flc = &ctx->flc[DECRYPT];
-+ desc = flc->sh_desc;
-+ cnstr_shdsc_rfc4543_decap(desc, &ctx->cdata, ivsize, ctx->authsize,
-+ true);
-+ flc->flc[1] = cpu_to_caam32(desc_len(desc)); /* SDL */
-+ dma_sync_single_for_device(dev, ctx->flc_dma[DECRYPT],
-+ sizeof(flc->flc) + desc_bytes(desc),
-+ ctx->dir);
-+
-+ return 0;
-+}
-+
-+static int rfc4543_setauthsize(struct crypto_aead *authenc,
-+ unsigned int authsize)
-+{
-+ struct caam_ctx *ctx = crypto_aead_ctx(authenc);
-+
-+ ctx->authsize = authsize;
-+ rfc4543_set_sh_desc(authenc);
-+
-+ return 0;
-+}
-+
-+static int rfc4543_setkey(struct crypto_aead *aead,
-+ const u8 *key, unsigned int keylen)
-+{
-+ struct caam_ctx *ctx = crypto_aead_ctx(aead);
-+ struct device *dev = ctx->dev;
-+
-+ if (keylen < 4)
-+ return -EINVAL;
-+
-+#ifdef DEBUG
-+ print_hex_dump(KERN_ERR, "key in @" __stringify(__LINE__)": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, key, keylen, 1);
-+#endif
-+
-+ memcpy(ctx->key, key, keylen);
-+ /*
-+ * The last four bytes of the key material are used as the salt value
-+ * in the nonce. Update the AES key length.
-+ */
-+ ctx->cdata.keylen = keylen - 4;
-+ dma_sync_single_for_device(dev, ctx->key_dma, ctx->cdata.keylen,
-+ ctx->dir);
-+
-+ return rfc4543_set_sh_desc(aead);
-+}
-+
-+static int ablkcipher_setkey(struct crypto_ablkcipher *ablkcipher,
-+ const u8 *key, unsigned int keylen)
-+{
-+ struct caam_ctx *ctx = crypto_ablkcipher_ctx(ablkcipher);
-+ struct crypto_tfm *tfm = crypto_ablkcipher_tfm(ablkcipher);
-+ const char *alg_name = crypto_tfm_alg_name(tfm);
-+ struct device *dev = ctx->dev;
-+ struct caam_flc *flc;
-+ unsigned int ivsize = crypto_ablkcipher_ivsize(ablkcipher);
-+ u32 *desc;
-+ u32 ctx1_iv_off = 0;
-+ const bool ctr_mode = ((ctx->cdata.algtype & OP_ALG_AAI_MASK) ==
-+ OP_ALG_AAI_CTR_MOD128);
-+ const bool is_rfc3686 = (ctr_mode && strstr(alg_name, "rfc3686"));
-+
-+#ifdef DEBUG
-+ print_hex_dump(KERN_ERR, "key in @" __stringify(__LINE__)": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, key, keylen, 1);
-+#endif
-+ /*
-+ * AES-CTR needs to load IV in CONTEXT1 reg
-+ * at an offset of 128bits (16bytes)
-+ * CONTEXT1[255:128] = IV
-+ */
-+ if (ctr_mode)
-+ ctx1_iv_off = 16;
-+
-+ /*
-+ * RFC3686 specific:
-+ * | CONTEXT1[255:128] = {NONCE, IV, COUNTER}
-+ * | *key = {KEY, NONCE}
-+ */
-+ if (is_rfc3686) {
-+ ctx1_iv_off = 16 + CTR_RFC3686_NONCE_SIZE;
-+ keylen -= CTR_RFC3686_NONCE_SIZE;
-+ }
-+
-+ ctx->cdata.keylen = keylen;
-+ ctx->cdata.key_virt = key;
-+ ctx->cdata.key_inline = true;
-+
-+ /* ablkcipher_encrypt shared descriptor */
-+ flc = &ctx->flc[ENCRYPT];
-+ desc = flc->sh_desc;
-+ cnstr_shdsc_ablkcipher_encap(desc, &ctx->cdata, ivsize,
-+ is_rfc3686, ctx1_iv_off);
-+ flc->flc[1] = cpu_to_caam32(desc_len(desc)); /* SDL */
-+ dma_sync_single_for_device(dev, ctx->flc_dma[ENCRYPT],
-+ sizeof(flc->flc) + desc_bytes(desc),
-+ ctx->dir);
-+
-+ /* ablkcipher_decrypt shared descriptor */
-+ flc = &ctx->flc[DECRYPT];
-+ desc = flc->sh_desc;
-+ cnstr_shdsc_ablkcipher_decap(desc, &ctx->cdata, ivsize,
-+ is_rfc3686, ctx1_iv_off);
-+ flc->flc[1] = cpu_to_caam32(desc_len(desc)); /* SDL */
-+ dma_sync_single_for_device(dev, ctx->flc_dma[DECRYPT],
-+ sizeof(flc->flc) + desc_bytes(desc),
-+ ctx->dir);
-+
-+ /* ablkcipher_givencrypt shared descriptor */
-+ flc = &ctx->flc[GIVENCRYPT];
-+ desc = flc->sh_desc;
-+ cnstr_shdsc_ablkcipher_givencap(desc, &ctx->cdata,
-+ ivsize, is_rfc3686, ctx1_iv_off);
-+ flc->flc[1] = cpu_to_caam32(desc_len(desc)); /* SDL */
-+ dma_sync_single_for_device(dev, ctx->flc_dma[GIVENCRYPT],
-+ sizeof(flc->flc) + desc_bytes(desc),
-+ ctx->dir);
-+
-+ return 0;
-+}
-+
-+static int xts_ablkcipher_setkey(struct crypto_ablkcipher *ablkcipher,
-+ const u8 *key, unsigned int keylen)
-+{
-+ struct caam_ctx *ctx = crypto_ablkcipher_ctx(ablkcipher);
-+ struct device *dev = ctx->dev;
-+ struct caam_flc *flc;
-+ u32 *desc;
-+
-+ if (keylen != 2 * AES_MIN_KEY_SIZE && keylen != 2 * AES_MAX_KEY_SIZE) {
-+ dev_err(dev, "key size mismatch\n");
-+ crypto_ablkcipher_set_flags(ablkcipher,
-+ CRYPTO_TFM_RES_BAD_KEY_LEN);
-+ return -EINVAL;
-+ }
-+
-+ ctx->cdata.keylen = keylen;
-+ ctx->cdata.key_virt = key;
-+ ctx->cdata.key_inline = true;
-+
-+ /* xts_ablkcipher_encrypt shared descriptor */
-+ flc = &ctx->flc[ENCRYPT];
-+ desc = flc->sh_desc;
-+ cnstr_shdsc_xts_ablkcipher_encap(desc, &ctx->cdata);
-+ flc->flc[1] = cpu_to_caam32(desc_len(desc)); /* SDL */
-+ dma_sync_single_for_device(dev, ctx->flc_dma[ENCRYPT],
-+ sizeof(flc->flc) + desc_bytes(desc),
-+ ctx->dir);
-+
-+ /* xts_ablkcipher_decrypt shared descriptor */
-+ flc = &ctx->flc[DECRYPT];
-+ desc = flc->sh_desc;
-+ cnstr_shdsc_xts_ablkcipher_decap(desc, &ctx->cdata);
-+ flc->flc[1] = cpu_to_caam32(desc_len(desc)); /* SDL */
-+ dma_sync_single_for_device(dev, ctx->flc_dma[DECRYPT],
-+ sizeof(flc->flc) + desc_bytes(desc),
-+ ctx->dir);
-+
-+ return 0;
-+}
-+
-+static struct ablkcipher_edesc *ablkcipher_edesc_alloc(struct ablkcipher_request
-+ *req, bool encrypt)
-+{
-+ struct crypto_ablkcipher *ablkcipher = crypto_ablkcipher_reqtfm(req);
-+ struct caam_request *req_ctx = ablkcipher_request_ctx(req);
-+ struct dpaa2_fl_entry *in_fle = &req_ctx->fd_flt[1];
-+ struct dpaa2_fl_entry *out_fle = &req_ctx->fd_flt[0];
-+ struct caam_ctx *ctx = crypto_ablkcipher_ctx(ablkcipher);
-+ struct device *dev = ctx->dev;
-+ gfp_t flags = (req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP) ?
-+ GFP_KERNEL : GFP_ATOMIC;
-+ int src_nents, mapped_src_nents, dst_nents = 0, mapped_dst_nents = 0;
-+ struct ablkcipher_edesc *edesc;
-+ dma_addr_t iv_dma;
-+ bool in_contig;
-+ int ivsize = crypto_ablkcipher_ivsize(ablkcipher);
-+ int dst_sg_idx, qm_sg_ents;
-+ struct dpaa2_sg_entry *sg_table;
-+ enum optype op_type = encrypt ? ENCRYPT : DECRYPT;
-+
-+ src_nents = sg_nents_for_len(req->src, req->nbytes);
-+ if (unlikely(src_nents < 0)) {
-+ dev_err(dev, "Insufficient bytes (%d) in src S/G\n",
-+ req->nbytes);
-+ return ERR_PTR(src_nents);
-+ }
-+
-+ if (unlikely(req->dst != req->src)) {
-+ dst_nents = sg_nents_for_len(req->dst, req->nbytes);
-+ if (unlikely(dst_nents < 0)) {
-+ dev_err(dev, "Insufficient bytes (%d) in dst S/G\n",
-+ req->nbytes);
-+ return ERR_PTR(dst_nents);
-+ }
-+
-+ mapped_src_nents = dma_map_sg(dev, req->src, src_nents,
-+ DMA_TO_DEVICE);
-+ if (unlikely(!mapped_src_nents)) {
-+ dev_err(dev, "unable to map source\n");
-+ return ERR_PTR(-ENOMEM);
-+ }
-+
-+ mapped_dst_nents = dma_map_sg(dev, req->dst, dst_nents,
-+ DMA_FROM_DEVICE);
-+ if (unlikely(!mapped_dst_nents)) {
-+ dev_err(dev, "unable to map destination\n");
-+ dma_unmap_sg(dev, req->src, src_nents, DMA_TO_DEVICE);
-+ return ERR_PTR(-ENOMEM);
-+ }
-+ } else {
-+ mapped_src_nents = dma_map_sg(dev, req->src, src_nents,
-+ DMA_BIDIRECTIONAL);
-+ if (unlikely(!mapped_src_nents)) {
-+ dev_err(dev, "unable to map source\n");
-+ return ERR_PTR(-ENOMEM);
-+ }
-+ }
-+
-+ iv_dma = dma_map_single(dev, req->info, ivsize, DMA_TO_DEVICE);
-+ if (dma_mapping_error(dev, iv_dma)) {
-+ dev_err(dev, "unable to map IV\n");
-+ caam_unmap(dev, req->src, req->dst, src_nents, dst_nents, 0,
-+ 0, 0, 0, 0);
-+ return ERR_PTR(-ENOMEM);
-+ }
-+
-+ if (mapped_src_nents == 1 &&
-+ iv_dma + ivsize == sg_dma_address(req->src)) {
-+ in_contig = true;
-+ qm_sg_ents = 0;
-+ } else {
-+ in_contig = false;
-+ qm_sg_ents = 1 + mapped_src_nents;
-+ }
-+ dst_sg_idx = qm_sg_ents;
-+
-+ qm_sg_ents += mapped_dst_nents > 1 ? mapped_dst_nents : 0;
-+ if (unlikely(qm_sg_ents > CAAM_QI_MAX_ABLKCIPHER_SG)) {
-+ dev_err(dev, "Insufficient S/G entries: %d > %lu\n",
-+ qm_sg_ents, CAAM_QI_MAX_ABLKCIPHER_SG);
-+ caam_unmap(dev, req->src, req->dst, src_nents, dst_nents,
-+ iv_dma, ivsize, op_type, 0, 0);
-+ return ERR_PTR(-ENOMEM);
-+ }
-+
-+ /* allocate space for base edesc and link tables */
-+ edesc = qi_cache_zalloc(GFP_DMA | flags);
-+ if (unlikely(!edesc)) {
-+ dev_err(dev, "could not allocate extended descriptor\n");
-+ caam_unmap(dev, req->src, req->dst, src_nents, dst_nents,
-+ iv_dma, ivsize, op_type, 0, 0);
-+ return ERR_PTR(-ENOMEM);
-+ }
-+
-+ edesc->src_nents = src_nents;
-+ edesc->dst_nents = dst_nents;
-+ edesc->iv_dma = iv_dma;
-+ sg_table = &edesc->sgt[0];
-+ edesc->qm_sg_bytes = qm_sg_ents * sizeof(*sg_table);
-+
-+ if (!in_contig) {
-+ dma_to_qm_sg_one(sg_table, iv_dma, ivsize, 0);
-+ sg_to_qm_sg_last(req->src, mapped_src_nents, sg_table + 1, 0);
-+ }
-+
-+ if (mapped_dst_nents > 1)
-+ sg_to_qm_sg_last(req->dst, mapped_dst_nents, sg_table +
-+ dst_sg_idx, 0);
-+
-+ edesc->qm_sg_dma = dma_map_single(dev, sg_table, edesc->qm_sg_bytes,
-+ DMA_TO_DEVICE);
-+ if (dma_mapping_error(dev, edesc->qm_sg_dma)) {
-+ dev_err(dev, "unable to map S/G table\n");
-+ caam_unmap(dev, req->src, req->dst, src_nents, dst_nents,
-+ iv_dma, ivsize, op_type, 0, 0);
-+ qi_cache_free(edesc);
-+ return ERR_PTR(-ENOMEM);
-+ }
-+
-+ memset(&req_ctx->fd_flt, 0, sizeof(req_ctx->fd_flt));
-+ dpaa2_fl_set_final(in_fle, true);
-+ dpaa2_fl_set_len(in_fle, req->nbytes + ivsize);
-+ dpaa2_fl_set_len(out_fle, req->nbytes);
-+
-+ if (!in_contig) {
-+ dpaa2_fl_set_format(in_fle, dpaa2_fl_sg);
-+ dpaa2_fl_set_addr(in_fle, edesc->qm_sg_dma);
-+ } else {
-+ dpaa2_fl_set_format(in_fle, dpaa2_fl_single);
-+ dpaa2_fl_set_addr(in_fle, iv_dma);
-+ }
-+
-+ if (req->src == req->dst) {
-+ if (!in_contig) {
-+ dpaa2_fl_set_format(out_fle, dpaa2_fl_sg);
-+ dpaa2_fl_set_addr(out_fle, edesc->qm_sg_dma +
-+ sizeof(*sg_table));
-+ } else {
-+ dpaa2_fl_set_format(out_fle, dpaa2_fl_single);
-+ dpaa2_fl_set_addr(out_fle, sg_dma_address(req->src));
-+ }
-+ } else if (mapped_dst_nents > 1) {
-+ dpaa2_fl_set_format(out_fle, dpaa2_fl_sg);
-+ dpaa2_fl_set_addr(out_fle, edesc->qm_sg_dma + dst_sg_idx *
-+ sizeof(*sg_table));
-+ } else {
-+ dpaa2_fl_set_format(out_fle, dpaa2_fl_single);
-+ dpaa2_fl_set_addr(out_fle, sg_dma_address(req->dst));
-+ }
-+
-+ return edesc;
-+}
-+
-+static struct ablkcipher_edesc *ablkcipher_giv_edesc_alloc(
-+ struct skcipher_givcrypt_request *greq)
-+{
-+ struct ablkcipher_request *req = &greq->creq;
-+ struct crypto_ablkcipher *ablkcipher = crypto_ablkcipher_reqtfm(req);
-+ struct caam_request *req_ctx = ablkcipher_request_ctx(req);
-+ struct dpaa2_fl_entry *in_fle = &req_ctx->fd_flt[1];
-+ struct dpaa2_fl_entry *out_fle = &req_ctx->fd_flt[0];
-+ struct caam_ctx *ctx = crypto_ablkcipher_ctx(ablkcipher);
-+ struct device *dev = ctx->dev;
-+ gfp_t flags = (req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP) ?
-+ GFP_KERNEL : GFP_ATOMIC;
-+ int src_nents, mapped_src_nents, dst_nents, mapped_dst_nents;
-+ struct ablkcipher_edesc *edesc;
-+ dma_addr_t iv_dma;
-+ bool out_contig;
-+ int ivsize = crypto_ablkcipher_ivsize(ablkcipher);
-+ struct dpaa2_sg_entry *sg_table;
-+ int dst_sg_idx, qm_sg_ents;
-+
-+ src_nents = sg_nents_for_len(req->src, req->nbytes);
-+ if (unlikely(src_nents < 0)) {
-+ dev_err(dev, "Insufficient bytes (%d) in src S/G\n",
-+ req->nbytes);
-+ return ERR_PTR(src_nents);
-+ }
-+
-+ if (unlikely(req->dst != req->src)) {
-+ dst_nents = sg_nents_for_len(req->dst, req->nbytes);
-+ if (unlikely(dst_nents < 0)) {
-+ dev_err(dev, "Insufficient bytes (%d) in dst S/G\n",
-+ req->nbytes);
-+ return ERR_PTR(dst_nents);
-+ }
-+
-+ mapped_src_nents = dma_map_sg(dev, req->src, src_nents,
-+ DMA_TO_DEVICE);
-+ if (unlikely(!mapped_src_nents)) {
-+ dev_err(dev, "unable to map source\n");
-+ return ERR_PTR(-ENOMEM);
-+ }
-+
-+ mapped_dst_nents = dma_map_sg(dev, req->dst, dst_nents,
-+ DMA_FROM_DEVICE);
-+ if (unlikely(!mapped_dst_nents)) {
-+ dev_err(dev, "unable to map destination\n");
-+ dma_unmap_sg(dev, req->src, src_nents, DMA_TO_DEVICE);
-+ return ERR_PTR(-ENOMEM);
-+ }
-+ } else {
-+ mapped_src_nents = dma_map_sg(dev, req->src, src_nents,
-+ DMA_BIDIRECTIONAL);
-+ if (unlikely(!mapped_src_nents)) {
-+ dev_err(dev, "unable to map source\n");
-+ return ERR_PTR(-ENOMEM);
-+ }
-+
-+ dst_nents = src_nents;
-+ mapped_dst_nents = src_nents;
-+ }
-+
-+ iv_dma = dma_map_single(dev, greq->giv, ivsize, DMA_FROM_DEVICE);
-+ if (dma_mapping_error(dev, iv_dma)) {
-+ dev_err(dev, "unable to map IV\n");
-+ caam_unmap(dev, req->src, req->dst, src_nents, dst_nents, 0,
-+ 0, 0, 0, 0);
-+ return ERR_PTR(-ENOMEM);
-+ }
-+
-+ qm_sg_ents = mapped_src_nents > 1 ? mapped_src_nents : 0;
-+ dst_sg_idx = qm_sg_ents;
-+ if (mapped_dst_nents == 1 &&
-+ iv_dma + ivsize == sg_dma_address(req->dst)) {
-+ out_contig = true;
-+ } else {
-+ out_contig = false;
-+ qm_sg_ents += 1 + mapped_dst_nents;
-+ }
-+
-+ if (unlikely(qm_sg_ents > CAAM_QI_MAX_ABLKCIPHER_SG)) {
-+ dev_err(dev, "Insufficient S/G entries: %d > %lu\n",
-+ qm_sg_ents, CAAM_QI_MAX_ABLKCIPHER_SG);
-+ caam_unmap(dev, req->src, req->dst, src_nents, dst_nents,
-+ iv_dma, ivsize, GIVENCRYPT, 0, 0);
-+ return ERR_PTR(-ENOMEM);
-+ }
-+
-+ /* allocate space for base edesc and link tables */
-+ edesc = qi_cache_zalloc(GFP_DMA | flags);
-+ if (!edesc) {
-+ dev_err(dev, "could not allocate extended descriptor\n");
-+ caam_unmap(dev, req->src, req->dst, src_nents, dst_nents,
-+ iv_dma, ivsize, GIVENCRYPT, 0, 0);
-+ return ERR_PTR(-ENOMEM);
-+ }
-+
-+ edesc->src_nents = src_nents;
-+ edesc->dst_nents = dst_nents;
-+ edesc->iv_dma = iv_dma;
-+ sg_table = &edesc->sgt[0];
-+ edesc->qm_sg_bytes = qm_sg_ents * sizeof(*sg_table);
-+
-+ if (mapped_src_nents > 1)
-+ sg_to_qm_sg_last(req->src, mapped_src_nents, sg_table, 0);
-+
-+ if (!out_contig) {
-+ dma_to_qm_sg_one(sg_table + dst_sg_idx, iv_dma, ivsize, 0);
-+ sg_to_qm_sg_last(req->dst, mapped_dst_nents, sg_table +
-+ dst_sg_idx + 1, 0);
-+ }
-+
-+ edesc->qm_sg_dma = dma_map_single(dev, sg_table, edesc->qm_sg_bytes,
-+ DMA_TO_DEVICE);
-+ if (dma_mapping_error(dev, edesc->qm_sg_dma)) {
-+ dev_err(dev, "unable to map S/G table\n");
-+ caam_unmap(dev, req->src, req->dst, src_nents, dst_nents,
-+ iv_dma, ivsize, GIVENCRYPT, 0, 0);
-+ qi_cache_free(edesc);
-+ return ERR_PTR(-ENOMEM);
-+ }
-+
-+ memset(&req_ctx->fd_flt, 0, sizeof(req_ctx->fd_flt));
-+ dpaa2_fl_set_final(in_fle, true);
-+ dpaa2_fl_set_len(in_fle, req->nbytes);
-+ dpaa2_fl_set_len(out_fle, ivsize + req->nbytes);
-+
-+ if (mapped_src_nents > 1) {
-+ dpaa2_fl_set_format(in_fle, dpaa2_fl_sg);
-+ dpaa2_fl_set_addr(in_fle, edesc->qm_sg_dma);
-+ } else {
-+ dpaa2_fl_set_format(in_fle, dpaa2_fl_single);
-+ dpaa2_fl_set_addr(in_fle, sg_dma_address(req->src));
-+ }
-+
-+ if (!out_contig) {
-+ dpaa2_fl_set_format(out_fle, dpaa2_fl_sg);
-+ dpaa2_fl_set_addr(out_fle, edesc->qm_sg_dma + dst_sg_idx *
-+ sizeof(*sg_table));
-+ } else {
-+ dpaa2_fl_set_format(out_fle, dpaa2_fl_single);
-+ dpaa2_fl_set_addr(out_fle, sg_dma_address(req->dst));
-+ }
-+
-+ return edesc;
-+}
-+
-+static void aead_unmap(struct device *dev, struct aead_edesc *edesc,
-+ struct aead_request *req)
-+{
-+ struct crypto_aead *aead = crypto_aead_reqtfm(req);
-+ int ivsize = crypto_aead_ivsize(aead);
-+ struct caam_request *caam_req = aead_request_ctx(req);
-+
-+ caam_unmap(dev, req->src, req->dst, edesc->src_nents, edesc->dst_nents,
-+ edesc->iv_dma, ivsize, caam_req->op_type,
-+ edesc->qm_sg_dma, edesc->qm_sg_bytes);
-+ dma_unmap_single(dev, edesc->assoclen_dma, 4, DMA_TO_DEVICE);
-+}
-+
-+static void tls_unmap(struct device *dev, struct tls_edesc *edesc,
-+ struct aead_request *req)
-+{
-+ struct crypto_aead *tls = crypto_aead_reqtfm(req);
-+ int ivsize = crypto_aead_ivsize(tls);
-+ struct caam_request *caam_req = aead_request_ctx(req);
-+
-+ caam_unmap(dev, req->src, edesc->dst, edesc->src_nents,
-+ edesc->dst_nents, edesc->iv_dma, ivsize, caam_req->op_type,
-+ edesc->qm_sg_dma, edesc->qm_sg_bytes);
-+}
-+
-+static void ablkcipher_unmap(struct device *dev,
-+ struct ablkcipher_edesc *edesc,
-+ struct ablkcipher_request *req)
-+{
-+ struct crypto_ablkcipher *ablkcipher = crypto_ablkcipher_reqtfm(req);
-+ int ivsize = crypto_ablkcipher_ivsize(ablkcipher);
-+ struct caam_request *caam_req = ablkcipher_request_ctx(req);
-+
-+ caam_unmap(dev, req->src, req->dst, edesc->src_nents, edesc->dst_nents,
-+ edesc->iv_dma, ivsize, caam_req->op_type,
-+ edesc->qm_sg_dma, edesc->qm_sg_bytes);
-+}
-+
-+static void aead_encrypt_done(void *cbk_ctx, u32 status)
-+{
-+ struct crypto_async_request *areq = cbk_ctx;
-+ struct aead_request *req = container_of(areq, struct aead_request,
-+ base);
-+ struct caam_request *req_ctx = to_caam_req(areq);
-+ struct aead_edesc *edesc = req_ctx->edesc;
-+ struct crypto_aead *aead = crypto_aead_reqtfm(req);
-+ struct caam_ctx *ctx = crypto_aead_ctx(aead);
-+ int ecode = 0;
-+
-+#ifdef DEBUG
-+ dev_err(ctx->dev, "%s %d: err 0x%x\n", __func__, __LINE__, status);
-+#endif
-+
-+ if (unlikely(status)) {
-+ caam_qi2_strstatus(ctx->dev, status);
-+ ecode = -EIO;
-+ }
-+
-+ aead_unmap(ctx->dev, edesc, req);
-+ qi_cache_free(edesc);
-+ aead_request_complete(req, ecode);
-+}
-+
-+static void aead_decrypt_done(void *cbk_ctx, u32 status)
-+{
-+ struct crypto_async_request *areq = cbk_ctx;
-+ struct aead_request *req = container_of(areq, struct aead_request,
-+ base);
-+ struct caam_request *req_ctx = to_caam_req(areq);
-+ struct aead_edesc *edesc = req_ctx->edesc;
-+ struct crypto_aead *aead = crypto_aead_reqtfm(req);
-+ struct caam_ctx *ctx = crypto_aead_ctx(aead);
-+ int ecode = 0;
-+
-+#ifdef DEBUG
-+ dev_err(ctx->dev, "%s %d: err 0x%x\n", __func__, __LINE__, status);
-+#endif
-+
-+ if (unlikely(status)) {
-+ caam_qi2_strstatus(ctx->dev, status);
-+ /*
-+ * verify hw auth check passed else return -EBADMSG
-+ */
-+ if ((status & JRSTA_CCBERR_ERRID_MASK) ==
-+ JRSTA_CCBERR_ERRID_ICVCHK)
-+ ecode = -EBADMSG;
-+ else
-+ ecode = -EIO;
-+ }
-+
-+ aead_unmap(ctx->dev, edesc, req);
-+ qi_cache_free(edesc);
-+ aead_request_complete(req, ecode);
-+}
-+
-+static int aead_encrypt(struct aead_request *req)
-+{
-+ struct aead_edesc *edesc;
-+ struct crypto_aead *aead = crypto_aead_reqtfm(req);
-+ struct caam_ctx *ctx = crypto_aead_ctx(aead);
-+ struct caam_request *caam_req = aead_request_ctx(req);
-+ int ret;
-+
-+ /* allocate extended descriptor */
-+ edesc = aead_edesc_alloc(req, true);
-+ if (IS_ERR(edesc))
-+ return PTR_ERR(edesc);
-+
-+ caam_req->flc = &ctx->flc[ENCRYPT];
-+ caam_req->flc_dma = ctx->flc_dma[ENCRYPT];
-+ caam_req->op_type = ENCRYPT;
-+ caam_req->cbk = aead_encrypt_done;
-+ caam_req->ctx = &req->base;
-+ caam_req->edesc = edesc;
-+ ret = dpaa2_caam_enqueue(ctx->dev, caam_req);
-+ if (ret != -EINPROGRESS &&
-+ !(ret == -EBUSY && req->base.flags & CRYPTO_TFM_REQ_MAY_BACKLOG)) {
-+ aead_unmap(ctx->dev, edesc, req);
-+ qi_cache_free(edesc);
-+ }
-+
-+ return ret;
-+}
-+
-+static int aead_decrypt(struct aead_request *req)
-+{
-+ struct aead_edesc *edesc;
-+ struct crypto_aead *aead = crypto_aead_reqtfm(req);
-+ struct caam_ctx *ctx = crypto_aead_ctx(aead);
-+ struct caam_request *caam_req = aead_request_ctx(req);
-+ int ret;
-+
-+ /* allocate extended descriptor */
-+ edesc = aead_edesc_alloc(req, false);
-+ if (IS_ERR(edesc))
-+ return PTR_ERR(edesc);
-+
-+ caam_req->flc = &ctx->flc[DECRYPT];
-+ caam_req->flc_dma = ctx->flc_dma[DECRYPT];
-+ caam_req->op_type = DECRYPT;
-+ caam_req->cbk = aead_decrypt_done;
-+ caam_req->ctx = &req->base;
-+ caam_req->edesc = edesc;
-+ ret = dpaa2_caam_enqueue(ctx->dev, caam_req);
-+ if (ret != -EINPROGRESS &&
-+ !(ret == -EBUSY && req->base.flags & CRYPTO_TFM_REQ_MAY_BACKLOG)) {
-+ aead_unmap(ctx->dev, edesc, req);
-+ qi_cache_free(edesc);
-+ }
-+
-+ return ret;
-+}
-+
-+static void tls_encrypt_done(void *cbk_ctx, u32 status)
-+{
-+ struct crypto_async_request *areq = cbk_ctx;
-+ struct aead_request *req = container_of(areq, struct aead_request,
-+ base);
-+ struct caam_request *req_ctx = to_caam_req(areq);
-+ struct tls_edesc *edesc = req_ctx->edesc;
-+ struct crypto_aead *tls = crypto_aead_reqtfm(req);
-+ struct caam_ctx *ctx = crypto_aead_ctx(tls);
-+ int ecode = 0;
-+
-+#ifdef DEBUG
-+ dev_err(ctx->dev, "%s %d: err 0x%x\n", __func__, __LINE__, status);
-+#endif
-+
-+ if (unlikely(status)) {
-+ caam_qi2_strstatus(ctx->dev, status);
-+ ecode = -EIO;
-+ }
-+
-+ tls_unmap(ctx->dev, edesc, req);
-+ qi_cache_free(edesc);
-+ aead_request_complete(req, ecode);
-+}
-+
-+static void tls_decrypt_done(void *cbk_ctx, u32 status)
-+{
-+ struct crypto_async_request *areq = cbk_ctx;
-+ struct aead_request *req = container_of(areq, struct aead_request,
-+ base);
-+ struct caam_request *req_ctx = to_caam_req(areq);
-+ struct tls_edesc *edesc = req_ctx->edesc;
-+ struct crypto_aead *tls = crypto_aead_reqtfm(req);
-+ struct caam_ctx *ctx = crypto_aead_ctx(tls);
-+ int ecode = 0;
-+
-+#ifdef DEBUG
-+ dev_err(ctx->dev, "%s %d: err 0x%x\n", __func__, __LINE__, status);
-+#endif
-+
-+ if (unlikely(status)) {
-+ caam_qi2_strstatus(ctx->dev, status);
-+ /*
-+ * verify hw auth check passed else return -EBADMSG
-+ */
-+ if ((status & JRSTA_CCBERR_ERRID_MASK) ==
-+ JRSTA_CCBERR_ERRID_ICVCHK)
-+ ecode = -EBADMSG;
-+ else
-+ ecode = -EIO;
-+ }
-+
-+ tls_unmap(ctx->dev, edesc, req);
-+ qi_cache_free(edesc);
-+ aead_request_complete(req, ecode);
-+}
-+
-+static int tls_encrypt(struct aead_request *req)
-+{
-+ struct tls_edesc *edesc;
-+ struct crypto_aead *tls = crypto_aead_reqtfm(req);
-+ struct caam_ctx *ctx = crypto_aead_ctx(tls);
-+ struct caam_request *caam_req = aead_request_ctx(req);
-+ int ret;
-+
-+ /* allocate extended descriptor */
-+ edesc = tls_edesc_alloc(req, true);
-+ if (IS_ERR(edesc))
-+ return PTR_ERR(edesc);
-+
-+ caam_req->flc = &ctx->flc[ENCRYPT];
-+ caam_req->flc_dma = ctx->flc_dma[ENCRYPT];
-+ caam_req->op_type = ENCRYPT;
-+ caam_req->cbk = tls_encrypt_done;
-+ caam_req->ctx = &req->base;
-+ caam_req->edesc = edesc;
-+ ret = dpaa2_caam_enqueue(ctx->dev, caam_req);
-+ if (ret != -EINPROGRESS &&
-+ !(ret == -EBUSY && req->base.flags & CRYPTO_TFM_REQ_MAY_BACKLOG)) {
-+ tls_unmap(ctx->dev, edesc, req);
-+ qi_cache_free(edesc);
-+ }
-+
-+ return ret;
-+}
-+
-+static int tls_decrypt(struct aead_request *req)
-+{
-+ struct tls_edesc *edesc;
-+ struct crypto_aead *tls = crypto_aead_reqtfm(req);
-+ struct caam_ctx *ctx = crypto_aead_ctx(tls);
-+ struct caam_request *caam_req = aead_request_ctx(req);
-+ int ret;
-+
-+ /* allocate extended descriptor */
-+ edesc = tls_edesc_alloc(req, false);
-+ if (IS_ERR(edesc))
-+ return PTR_ERR(edesc);
-+
-+ caam_req->flc = &ctx->flc[DECRYPT];
-+ caam_req->flc_dma = ctx->flc_dma[DECRYPT];
-+ caam_req->op_type = DECRYPT;
-+ caam_req->cbk = tls_decrypt_done;
-+ caam_req->ctx = &req->base;
-+ caam_req->edesc = edesc;
-+ ret = dpaa2_caam_enqueue(ctx->dev, caam_req);
-+ if (ret != -EINPROGRESS &&
-+ !(ret == -EBUSY && req->base.flags & CRYPTO_TFM_REQ_MAY_BACKLOG)) {
-+ tls_unmap(ctx->dev, edesc, req);
-+ qi_cache_free(edesc);
-+ }
-+
-+ return ret;
-+}
-+
-+static int ipsec_gcm_encrypt(struct aead_request *req)
-+{
-+ if (req->assoclen < 8)
-+ return -EINVAL;
-+
-+ return aead_encrypt(req);
-+}
-+
-+static int ipsec_gcm_decrypt(struct aead_request *req)
-+{
-+ if (req->assoclen < 8)
-+ return -EINVAL;
-+
-+ return aead_decrypt(req);
-+}
-+
-+static void ablkcipher_done(void *cbk_ctx, u32 status)
-+{
-+ struct crypto_async_request *areq = cbk_ctx;
-+ struct ablkcipher_request *req = ablkcipher_request_cast(areq);
-+ struct caam_request *req_ctx = to_caam_req(areq);
-+ struct crypto_ablkcipher *ablkcipher = crypto_ablkcipher_reqtfm(req);
-+ struct caam_ctx *ctx = crypto_ablkcipher_ctx(ablkcipher);
-+ struct ablkcipher_edesc *edesc = req_ctx->edesc;
-+ int ecode = 0;
-+ int ivsize = crypto_ablkcipher_ivsize(ablkcipher);
-+
-+#ifdef DEBUG
-+ dev_err(ctx->dev, "%s %d: err 0x%x\n", __func__, __LINE__, status);
-+#endif
-+
-+ if (unlikely(status)) {
-+ caam_qi2_strstatus(ctx->dev, status);
-+ ecode = -EIO;
-+ }
-+
-+#ifdef DEBUG
-+ print_hex_dump(KERN_ERR, "dstiv @" __stringify(__LINE__)": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, req->info,
-+ edesc->src_nents > 1 ? 100 : ivsize, 1);
-+ caam_dump_sg(KERN_ERR, "dst @" __stringify(__LINE__)": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, req->dst,
-+ edesc->dst_nents > 1 ? 100 : req->nbytes, 1);
-+#endif
-+
-+ ablkcipher_unmap(ctx->dev, edesc, req);
-+ qi_cache_free(edesc);
-+
-+ /*
-+ * The crypto API expects us to set the IV (req->info) to the last
-+ * ciphertext block. This is used e.g. by the CTS mode.
-+ */
-+ scatterwalk_map_and_copy(req->info, req->dst, req->nbytes - ivsize,
-+ ivsize, 0);
-+
-+ ablkcipher_request_complete(req, ecode);
-+}
-+
-+static int ablkcipher_encrypt(struct ablkcipher_request *req)
-+{
-+ struct ablkcipher_edesc *edesc;
-+ struct crypto_ablkcipher *ablkcipher = crypto_ablkcipher_reqtfm(req);
-+ struct caam_ctx *ctx = crypto_ablkcipher_ctx(ablkcipher);
-+ struct caam_request *caam_req = ablkcipher_request_ctx(req);
-+ int ret;
-+
-+ /* allocate extended descriptor */
-+ edesc = ablkcipher_edesc_alloc(req, true);
-+ if (IS_ERR(edesc))
-+ return PTR_ERR(edesc);
-+
-+ caam_req->flc = &ctx->flc[ENCRYPT];
-+ caam_req->flc_dma = ctx->flc_dma[ENCRYPT];
-+ caam_req->op_type = ENCRYPT;
-+ caam_req->cbk = ablkcipher_done;
-+ caam_req->ctx = &req->base;
-+ caam_req->edesc = edesc;
-+ ret = dpaa2_caam_enqueue(ctx->dev, caam_req);
-+ if (ret != -EINPROGRESS &&
-+ !(ret == -EBUSY && req->base.flags & CRYPTO_TFM_REQ_MAY_BACKLOG)) {
-+ ablkcipher_unmap(ctx->dev, edesc, req);
-+ qi_cache_free(edesc);
-+ }
-+
-+ return ret;
-+}
-+
-+static int ablkcipher_givencrypt(struct skcipher_givcrypt_request *greq)
-+{
-+ struct ablkcipher_request *req = &greq->creq;
-+ struct ablkcipher_edesc *edesc;
-+ struct crypto_ablkcipher *ablkcipher = crypto_ablkcipher_reqtfm(req);
-+ struct caam_ctx *ctx = crypto_ablkcipher_ctx(ablkcipher);
-+ struct caam_request *caam_req = ablkcipher_request_ctx(req);
-+ int ret;
-+
-+ /* allocate extended descriptor */
-+ edesc = ablkcipher_giv_edesc_alloc(greq);
-+ if (IS_ERR(edesc))
-+ return PTR_ERR(edesc);
-+
-+ caam_req->flc = &ctx->flc[GIVENCRYPT];
-+ caam_req->flc_dma = ctx->flc_dma[GIVENCRYPT];
-+ caam_req->op_type = GIVENCRYPT;
-+ caam_req->cbk = ablkcipher_done;
-+ caam_req->ctx = &req->base;
-+ caam_req->edesc = edesc;
-+ ret = dpaa2_caam_enqueue(ctx->dev, caam_req);
-+ if (ret != -EINPROGRESS &&
-+ !(ret == -EBUSY && req->base.flags & CRYPTO_TFM_REQ_MAY_BACKLOG)) {
-+ ablkcipher_unmap(ctx->dev, edesc, req);
-+ qi_cache_free(edesc);
-+ }
-+
-+ return ret;
-+}
-+
-+static int ablkcipher_decrypt(struct ablkcipher_request *req)
-+{
-+ struct ablkcipher_edesc *edesc;
-+ struct crypto_ablkcipher *ablkcipher = crypto_ablkcipher_reqtfm(req);
-+ struct caam_ctx *ctx = crypto_ablkcipher_ctx(ablkcipher);
-+ struct caam_request *caam_req = ablkcipher_request_ctx(req);
-+ int ret;
-+
-+ /* allocate extended descriptor */
-+ edesc = ablkcipher_edesc_alloc(req, false);
-+ if (IS_ERR(edesc))
-+ return PTR_ERR(edesc);
-+
-+ caam_req->flc = &ctx->flc[DECRYPT];
-+ caam_req->flc_dma = ctx->flc_dma[DECRYPT];
-+ caam_req->op_type = DECRYPT;
-+ caam_req->cbk = ablkcipher_done;
-+ caam_req->ctx = &req->base;
-+ caam_req->edesc = edesc;
-+ ret = dpaa2_caam_enqueue(ctx->dev, caam_req);
-+ if (ret != -EINPROGRESS &&
-+ !(ret == -EBUSY && req->base.flags & CRYPTO_TFM_REQ_MAY_BACKLOG)) {
-+ ablkcipher_unmap(ctx->dev, edesc, req);
-+ qi_cache_free(edesc);
-+ }
-+
-+ return ret;
-+}
-+
-+struct caam_crypto_alg {
-+ struct list_head entry;
-+ struct crypto_alg crypto_alg;
-+ struct caam_alg_entry caam;
-+};
-+
-+static int caam_cra_init(struct crypto_tfm *tfm, bool uses_dkp)
-+{
-+ struct crypto_alg *alg = tfm->__crt_alg;
-+ struct caam_crypto_alg *caam_alg = container_of(alg, typeof(*caam_alg),
-+ crypto_alg);
-+ struct caam_ctx *ctx = crypto_tfm_ctx(tfm);
-+ dma_addr_t dma_addr;
-+ int i;
-+
-+ /* copy descriptor header template value */
-+ ctx->cdata.algtype = OP_TYPE_CLASS1_ALG |
-+ caam_alg->caam.class1_alg_type;
-+ ctx->adata.algtype = OP_TYPE_CLASS2_ALG |
-+ caam_alg->caam.class2_alg_type;
-+
-+ ctx->dev = caam_alg->caam.dev;
-+ ctx->dir = uses_dkp ? DMA_BIDIRECTIONAL : DMA_TO_DEVICE;
-+
-+ dma_addr = dma_map_single_attrs(ctx->dev, ctx->flc,
-+ offsetof(struct caam_ctx, flc_dma),
-+ ctx->dir, DMA_ATTR_SKIP_CPU_SYNC);
-+ if (dma_mapping_error(ctx->dev, dma_addr)) {
-+ dev_err(ctx->dev, "unable to map key, shared descriptors\n");
-+ return -ENOMEM;
-+ }
-+
-+ for (i = 0; i < NUM_OP; i++)
-+ ctx->flc_dma[i] = dma_addr + i * sizeof(ctx->flc[i]);
-+ ctx->key_dma = dma_addr + NUM_OP * sizeof(ctx->flc[0]);
-+
-+ return 0;
-+}
-+
-+static int caam_cra_init_ablkcipher(struct crypto_tfm *tfm)
-+{
-+ struct ablkcipher_tfm *ablkcipher_tfm =
-+ crypto_ablkcipher_crt(__crypto_ablkcipher_cast(tfm));
-+
-+ ablkcipher_tfm->reqsize = sizeof(struct caam_request);
-+ return caam_cra_init(tfm, false);
-+}
-+
-+static int caam_cra_init_aead(struct crypto_aead *tfm)
-+{
-+ struct aead_alg *alg = crypto_aead_alg(tfm);
-+
-+ crypto_aead_set_reqsize(tfm, sizeof(struct caam_request));
-+ return caam_cra_init(crypto_aead_tfm(tfm),
-+ (alg->setkey == aead_setkey) ||
-+ (alg->setkey == tls_setkey));
-+}
-+
-+static void caam_exit_common(struct caam_ctx *ctx)
-+{
-+ dma_unmap_single_attrs(ctx->dev, ctx->flc_dma[0],
-+ offsetof(struct caam_ctx, flc_dma), ctx->dir,
-+ DMA_ATTR_SKIP_CPU_SYNC);
-+}
-+
-+static void caam_cra_exit(struct crypto_tfm *tfm)
-+{
-+ caam_exit_common(crypto_tfm_ctx(tfm));
-+}
-+
-+static void caam_cra_exit_aead(struct crypto_aead *tfm)
-+{
-+ caam_exit_common(crypto_aead_ctx(tfm));
-+}
-+
-+#define template_ablkcipher template_u.ablkcipher
-+struct caam_alg_template {
-+ char name[CRYPTO_MAX_ALG_NAME];
-+ char driver_name[CRYPTO_MAX_ALG_NAME];
-+ unsigned int blocksize;
-+ u32 type;
-+ union {
-+ struct ablkcipher_alg ablkcipher;
-+ } template_u;
-+ u32 class1_alg_type;
-+ u32 class2_alg_type;
-+};
-+
-+static struct caam_alg_template driver_algs[] = {
-+ /* ablkcipher descriptor */
-+ {
-+ .name = "cbc(aes)",
-+ .driver_name = "cbc-aes-caam-qi2",
-+ .blocksize = AES_BLOCK_SIZE,
-+ .type = CRYPTO_ALG_TYPE_GIVCIPHER,
-+ .template_ablkcipher = {
-+ .setkey = ablkcipher_setkey,
-+ .encrypt = ablkcipher_encrypt,
-+ .decrypt = ablkcipher_decrypt,
-+ .givencrypt = ablkcipher_givencrypt,
-+ .geniv = "<built-in>",
-+ .min_keysize = AES_MIN_KEY_SIZE,
-+ .max_keysize = AES_MAX_KEY_SIZE,
-+ .ivsize = AES_BLOCK_SIZE,
-+ },
-+ .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_CBC,
-+ },
-+ {
-+ .name = "cbc(des3_ede)",
-+ .driver_name = "cbc-3des-caam-qi2",
-+ .blocksize = DES3_EDE_BLOCK_SIZE,
-+ .type = CRYPTO_ALG_TYPE_GIVCIPHER,
-+ .template_ablkcipher = {
-+ .setkey = ablkcipher_setkey,
-+ .encrypt = ablkcipher_encrypt,
-+ .decrypt = ablkcipher_decrypt,
-+ .givencrypt = ablkcipher_givencrypt,
-+ .geniv = "<built-in>",
-+ .min_keysize = DES3_EDE_KEY_SIZE,
-+ .max_keysize = DES3_EDE_KEY_SIZE,
-+ .ivsize = DES3_EDE_BLOCK_SIZE,
-+ },
-+ .class1_alg_type = OP_ALG_ALGSEL_3DES | OP_ALG_AAI_CBC,
-+ },
-+ {
-+ .name = "cbc(des)",
-+ .driver_name = "cbc-des-caam-qi2",
-+ .blocksize = DES_BLOCK_SIZE,
-+ .type = CRYPTO_ALG_TYPE_GIVCIPHER,
-+ .template_ablkcipher = {
-+ .setkey = ablkcipher_setkey,
-+ .encrypt = ablkcipher_encrypt,
-+ .decrypt = ablkcipher_decrypt,
-+ .givencrypt = ablkcipher_givencrypt,
-+ .geniv = "<built-in>",
-+ .min_keysize = DES_KEY_SIZE,
-+ .max_keysize = DES_KEY_SIZE,
-+ .ivsize = DES_BLOCK_SIZE,
-+ },
-+ .class1_alg_type = OP_ALG_ALGSEL_DES | OP_ALG_AAI_CBC,
-+ },
-+ {
-+ .name = "ctr(aes)",
-+ .driver_name = "ctr-aes-caam-qi2",
-+ .blocksize = 1,
-+ .type = CRYPTO_ALG_TYPE_ABLKCIPHER,
-+ .template_ablkcipher = {
-+ .setkey = ablkcipher_setkey,
-+ .encrypt = ablkcipher_encrypt,
-+ .decrypt = ablkcipher_decrypt,
-+ .geniv = "chainiv",
-+ .min_keysize = AES_MIN_KEY_SIZE,
-+ .max_keysize = AES_MAX_KEY_SIZE,
-+ .ivsize = AES_BLOCK_SIZE,
-+ },
-+ .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_CTR_MOD128,
-+ },
-+ {
-+ .name = "rfc3686(ctr(aes))",
-+ .driver_name = "rfc3686-ctr-aes-caam-qi2",
-+ .blocksize = 1,
-+ .type = CRYPTO_ALG_TYPE_GIVCIPHER,
-+ .template_ablkcipher = {
-+ .setkey = ablkcipher_setkey,
-+ .encrypt = ablkcipher_encrypt,
-+ .decrypt = ablkcipher_decrypt,
-+ .givencrypt = ablkcipher_givencrypt,
-+ .geniv = "<built-in>",
-+ .min_keysize = AES_MIN_KEY_SIZE +
-+ CTR_RFC3686_NONCE_SIZE,
-+ .max_keysize = AES_MAX_KEY_SIZE +
-+ CTR_RFC3686_NONCE_SIZE,
-+ .ivsize = CTR_RFC3686_IV_SIZE,
-+ },
-+ .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_CTR_MOD128,
-+ },
-+ {
-+ .name = "xts(aes)",
-+ .driver_name = "xts-aes-caam-qi2",
-+ .blocksize = AES_BLOCK_SIZE,
-+ .type = CRYPTO_ALG_TYPE_ABLKCIPHER,
-+ .template_ablkcipher = {
-+ .setkey = xts_ablkcipher_setkey,
-+ .encrypt = ablkcipher_encrypt,
-+ .decrypt = ablkcipher_decrypt,
-+ .geniv = "eseqiv",
-+ .min_keysize = 2 * AES_MIN_KEY_SIZE,
-+ .max_keysize = 2 * AES_MAX_KEY_SIZE,
-+ .ivsize = AES_BLOCK_SIZE,
-+ },
-+ .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_XTS,
-+ }
-+};
-+
-+static struct caam_aead_alg driver_aeads[] = {
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "rfc4106(gcm(aes))",
-+ .cra_driver_name = "rfc4106-gcm-aes-caam-qi2",
-+ .cra_blocksize = 1,
-+ },
-+ .setkey = rfc4106_setkey,
-+ .setauthsize = rfc4106_setauthsize,
-+ .encrypt = ipsec_gcm_encrypt,
-+ .decrypt = ipsec_gcm_decrypt,
-+ .ivsize = 8,
-+ .maxauthsize = AES_BLOCK_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_GCM,
-+ },
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "rfc4543(gcm(aes))",
-+ .cra_driver_name = "rfc4543-gcm-aes-caam-qi2",
-+ .cra_blocksize = 1,
-+ },
-+ .setkey = rfc4543_setkey,
-+ .setauthsize = rfc4543_setauthsize,
-+ .encrypt = ipsec_gcm_encrypt,
-+ .decrypt = ipsec_gcm_decrypt,
-+ .ivsize = 8,
-+ .maxauthsize = AES_BLOCK_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_GCM,
-+ },
-+ },
-+ /* Galois Counter Mode */
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "gcm(aes)",
-+ .cra_driver_name = "gcm-aes-caam-qi2",
-+ .cra_blocksize = 1,
-+ },
-+ .setkey = gcm_setkey,
-+ .setauthsize = gcm_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = 12,
-+ .maxauthsize = AES_BLOCK_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_GCM,
-+ }
-+ },
-+ /* single-pass ipsec_esp descriptor */
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "authenc(hmac(md5),cbc(aes))",
-+ .cra_driver_name = "authenc-hmac-md5-"
-+ "cbc-aes-caam-qi2",
-+ .cra_blocksize = AES_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = AES_BLOCK_SIZE,
-+ .maxauthsize = MD5_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_MD5 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ }
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "echainiv(authenc(hmac(md5),"
-+ "cbc(aes)))",
-+ .cra_driver_name = "echainiv-authenc-hmac-md5-"
-+ "cbc-aes-caam-qi2",
-+ .cra_blocksize = AES_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = AES_BLOCK_SIZE,
-+ .maxauthsize = MD5_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_MD5 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ .geniv = true,
-+ }
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "authenc(hmac(sha1),cbc(aes))",
-+ .cra_driver_name = "authenc-hmac-sha1-"
-+ "cbc-aes-caam-qi2",
-+ .cra_blocksize = AES_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = AES_BLOCK_SIZE,
-+ .maxauthsize = SHA1_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA1 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ }
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "echainiv(authenc(hmac(sha1),"
-+ "cbc(aes)))",
-+ .cra_driver_name = "echainiv-authenc-"
-+ "hmac-sha1-cbc-aes-caam-qi2",
-+ .cra_blocksize = AES_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = AES_BLOCK_SIZE,
-+ .maxauthsize = SHA1_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA1 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ .geniv = true,
-+ },
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "authenc(hmac(sha224),cbc(aes))",
-+ .cra_driver_name = "authenc-hmac-sha224-"
-+ "cbc-aes-caam-qi2",
-+ .cra_blocksize = AES_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = AES_BLOCK_SIZE,
-+ .maxauthsize = SHA224_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA224 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ }
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "echainiv(authenc(hmac(sha224),"
-+ "cbc(aes)))",
-+ .cra_driver_name = "echainiv-authenc-"
-+ "hmac-sha224-cbc-aes-caam-qi2",
-+ .cra_blocksize = AES_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = AES_BLOCK_SIZE,
-+ .maxauthsize = SHA224_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA224 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ .geniv = true,
-+ }
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "authenc(hmac(sha256),cbc(aes))",
-+ .cra_driver_name = "authenc-hmac-sha256-"
-+ "cbc-aes-caam-qi2",
-+ .cra_blocksize = AES_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = AES_BLOCK_SIZE,
-+ .maxauthsize = SHA256_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA256 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ }
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "echainiv(authenc(hmac(sha256),"
-+ "cbc(aes)))",
-+ .cra_driver_name = "echainiv-authenc-"
-+ "hmac-sha256-cbc-aes-"
-+ "caam-qi2",
-+ .cra_blocksize = AES_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = AES_BLOCK_SIZE,
-+ .maxauthsize = SHA256_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA256 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ .geniv = true,
-+ }
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "authenc(hmac(sha384),cbc(aes))",
-+ .cra_driver_name = "authenc-hmac-sha384-"
-+ "cbc-aes-caam-qi2",
-+ .cra_blocksize = AES_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = AES_BLOCK_SIZE,
-+ .maxauthsize = SHA384_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA384 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ }
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "echainiv(authenc(hmac(sha384),"
-+ "cbc(aes)))",
-+ .cra_driver_name = "echainiv-authenc-"
-+ "hmac-sha384-cbc-aes-"
-+ "caam-qi2",
-+ .cra_blocksize = AES_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = AES_BLOCK_SIZE,
-+ .maxauthsize = SHA384_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA384 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ .geniv = true,
-+ }
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "authenc(hmac(sha512),cbc(aes))",
-+ .cra_driver_name = "authenc-hmac-sha512-"
-+ "cbc-aes-caam-qi2",
-+ .cra_blocksize = AES_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = AES_BLOCK_SIZE,
-+ .maxauthsize = SHA512_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA512 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ }
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "echainiv(authenc(hmac(sha512),"
-+ "cbc(aes)))",
-+ .cra_driver_name = "echainiv-authenc-"
-+ "hmac-sha512-cbc-aes-"
-+ "caam-qi2",
-+ .cra_blocksize = AES_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = AES_BLOCK_SIZE,
-+ .maxauthsize = SHA512_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA512 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ .geniv = true,
-+ }
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "authenc(hmac(md5),cbc(des3_ede))",
-+ .cra_driver_name = "authenc-hmac-md5-"
-+ "cbc-des3_ede-caam-qi2",
-+ .cra_blocksize = DES3_EDE_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = DES3_EDE_BLOCK_SIZE,
-+ .maxauthsize = MD5_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_3DES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_MD5 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ }
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "echainiv(authenc(hmac(md5),"
-+ "cbc(des3_ede)))",
-+ .cra_driver_name = "echainiv-authenc-hmac-md5-"
-+ "cbc-des3_ede-caam-qi2",
-+ .cra_blocksize = DES3_EDE_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = DES3_EDE_BLOCK_SIZE,
-+ .maxauthsize = MD5_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_3DES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_MD5 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ .geniv = true,
-+ }
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "authenc(hmac(sha1),"
-+ "cbc(des3_ede))",
-+ .cra_driver_name = "authenc-hmac-sha1-"
-+ "cbc-des3_ede-caam-qi2",
-+ .cra_blocksize = DES3_EDE_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = DES3_EDE_BLOCK_SIZE,
-+ .maxauthsize = SHA1_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_3DES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA1 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ },
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "echainiv(authenc(hmac(sha1),"
-+ "cbc(des3_ede)))",
-+ .cra_driver_name = "echainiv-authenc-"
-+ "hmac-sha1-"
-+ "cbc-des3_ede-caam-qi2",
-+ .cra_blocksize = DES3_EDE_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = DES3_EDE_BLOCK_SIZE,
-+ .maxauthsize = SHA1_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_3DES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA1 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ .geniv = true,
-+ }
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "authenc(hmac(sha224),"
-+ "cbc(des3_ede))",
-+ .cra_driver_name = "authenc-hmac-sha224-"
-+ "cbc-des3_ede-caam-qi2",
-+ .cra_blocksize = DES3_EDE_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = DES3_EDE_BLOCK_SIZE,
-+ .maxauthsize = SHA224_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_3DES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA224 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ },
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "echainiv(authenc(hmac(sha224),"
-+ "cbc(des3_ede)))",
-+ .cra_driver_name = "echainiv-authenc-"
-+ "hmac-sha224-"
-+ "cbc-des3_ede-caam-qi2",
-+ .cra_blocksize = DES3_EDE_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = DES3_EDE_BLOCK_SIZE,
-+ .maxauthsize = SHA224_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_3DES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA224 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ .geniv = true,
-+ }
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "authenc(hmac(sha256),"
-+ "cbc(des3_ede))",
-+ .cra_driver_name = "authenc-hmac-sha256-"
-+ "cbc-des3_ede-caam-qi2",
-+ .cra_blocksize = DES3_EDE_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = DES3_EDE_BLOCK_SIZE,
-+ .maxauthsize = SHA256_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_3DES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA256 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ },
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "echainiv(authenc(hmac(sha256),"
-+ "cbc(des3_ede)))",
-+ .cra_driver_name = "echainiv-authenc-"
-+ "hmac-sha256-"
-+ "cbc-des3_ede-caam-qi2",
-+ .cra_blocksize = DES3_EDE_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = DES3_EDE_BLOCK_SIZE,
-+ .maxauthsize = SHA256_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_3DES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA256 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ .geniv = true,
-+ }
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "authenc(hmac(sha384),"
-+ "cbc(des3_ede))",
-+ .cra_driver_name = "authenc-hmac-sha384-"
-+ "cbc-des3_ede-caam-qi2",
-+ .cra_blocksize = DES3_EDE_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = DES3_EDE_BLOCK_SIZE,
-+ .maxauthsize = SHA384_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_3DES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA384 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ },
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "echainiv(authenc(hmac(sha384),"
-+ "cbc(des3_ede)))",
-+ .cra_driver_name = "echainiv-authenc-"
-+ "hmac-sha384-"
-+ "cbc-des3_ede-caam-qi2",
-+ .cra_blocksize = DES3_EDE_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = DES3_EDE_BLOCK_SIZE,
-+ .maxauthsize = SHA384_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_3DES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA384 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ .geniv = true,
-+ }
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "authenc(hmac(sha512),"
-+ "cbc(des3_ede))",
-+ .cra_driver_name = "authenc-hmac-sha512-"
-+ "cbc-des3_ede-caam-qi2",
-+ .cra_blocksize = DES3_EDE_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = DES3_EDE_BLOCK_SIZE,
-+ .maxauthsize = SHA512_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_3DES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA512 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ },
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "echainiv(authenc(hmac(sha512),"
-+ "cbc(des3_ede)))",
-+ .cra_driver_name = "echainiv-authenc-"
-+ "hmac-sha512-"
-+ "cbc-des3_ede-caam-qi2",
-+ .cra_blocksize = DES3_EDE_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = DES3_EDE_BLOCK_SIZE,
-+ .maxauthsize = SHA512_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_3DES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA512 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ .geniv = true,
-+ }
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "authenc(hmac(md5),cbc(des))",
-+ .cra_driver_name = "authenc-hmac-md5-"
-+ "cbc-des-caam-qi2",
-+ .cra_blocksize = DES_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = DES_BLOCK_SIZE,
-+ .maxauthsize = MD5_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_DES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_MD5 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ },
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "echainiv(authenc(hmac(md5),"
-+ "cbc(des)))",
-+ .cra_driver_name = "echainiv-authenc-hmac-md5-"
-+ "cbc-des-caam-qi2",
-+ .cra_blocksize = DES_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = DES_BLOCK_SIZE,
-+ .maxauthsize = MD5_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_DES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_MD5 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ .geniv = true,
-+ }
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "authenc(hmac(sha1),cbc(des))",
-+ .cra_driver_name = "authenc-hmac-sha1-"
-+ "cbc-des-caam-qi2",
-+ .cra_blocksize = DES_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = DES_BLOCK_SIZE,
-+ .maxauthsize = SHA1_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_DES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA1 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ },
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "echainiv(authenc(hmac(sha1),"
-+ "cbc(des)))",
-+ .cra_driver_name = "echainiv-authenc-"
-+ "hmac-sha1-cbc-des-caam-qi2",
-+ .cra_blocksize = DES_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = DES_BLOCK_SIZE,
-+ .maxauthsize = SHA1_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_DES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA1 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ .geniv = true,
-+ }
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "authenc(hmac(sha224),cbc(des))",
-+ .cra_driver_name = "authenc-hmac-sha224-"
-+ "cbc-des-caam-qi2",
-+ .cra_blocksize = DES_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = DES_BLOCK_SIZE,
-+ .maxauthsize = SHA224_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_DES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA224 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ },
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "echainiv(authenc(hmac(sha224),"
-+ "cbc(des)))",
-+ .cra_driver_name = "echainiv-authenc-"
-+ "hmac-sha224-cbc-des-"
-+ "caam-qi2",
-+ .cra_blocksize = DES_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = DES_BLOCK_SIZE,
-+ .maxauthsize = SHA224_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_DES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA224 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ .geniv = true,
-+ }
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "authenc(hmac(sha256),cbc(des))",
-+ .cra_driver_name = "authenc-hmac-sha256-"
-+ "cbc-des-caam-qi2",
-+ .cra_blocksize = DES_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = DES_BLOCK_SIZE,
-+ .maxauthsize = SHA256_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_DES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA256 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ },
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "echainiv(authenc(hmac(sha256),"
-+ "cbc(des)))",
-+ .cra_driver_name = "echainiv-authenc-"
-+ "hmac-sha256-cbc-desi-"
-+ "caam-qi2",
-+ .cra_blocksize = DES_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = DES_BLOCK_SIZE,
-+ .maxauthsize = SHA256_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_DES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA256 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ .geniv = true,
-+ },
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "authenc(hmac(sha384),cbc(des))",
-+ .cra_driver_name = "authenc-hmac-sha384-"
-+ "cbc-des-caam-qi2",
-+ .cra_blocksize = DES_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = DES_BLOCK_SIZE,
-+ .maxauthsize = SHA384_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_DES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA384 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ },
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "echainiv(authenc(hmac(sha384),"
-+ "cbc(des)))",
-+ .cra_driver_name = "echainiv-authenc-"
-+ "hmac-sha384-cbc-des-"
-+ "caam-qi2",
-+ .cra_blocksize = DES_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = DES_BLOCK_SIZE,
-+ .maxauthsize = SHA384_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_DES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA384 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ .geniv = true,
-+ }
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "authenc(hmac(sha512),cbc(des))",
-+ .cra_driver_name = "authenc-hmac-sha512-"
-+ "cbc-des-caam-qi2",
-+ .cra_blocksize = DES_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = DES_BLOCK_SIZE,
-+ .maxauthsize = SHA512_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_DES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA512 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ }
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "echainiv(authenc(hmac(sha512),"
-+ "cbc(des)))",
-+ .cra_driver_name = "echainiv-authenc-"
-+ "hmac-sha512-cbc-des-"
-+ "caam-qi2",
-+ .cra_blocksize = DES_BLOCK_SIZE,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = DES_BLOCK_SIZE,
-+ .maxauthsize = SHA512_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_DES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA512 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ .geniv = true,
-+ }
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "authenc(hmac(md5),"
-+ "rfc3686(ctr(aes)))",
-+ .cra_driver_name = "authenc-hmac-md5-"
-+ "rfc3686-ctr-aes-caam-qi2",
-+ .cra_blocksize = 1,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = CTR_RFC3686_IV_SIZE,
-+ .maxauthsize = MD5_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_AES |
-+ OP_ALG_AAI_CTR_MOD128,
-+ .class2_alg_type = OP_ALG_ALGSEL_MD5 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ .rfc3686 = true,
-+ },
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "seqiv(authenc("
-+ "hmac(md5),rfc3686(ctr(aes))))",
-+ .cra_driver_name = "seqiv-authenc-hmac-md5-"
-+ "rfc3686-ctr-aes-caam-qi2",
-+ .cra_blocksize = 1,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = CTR_RFC3686_IV_SIZE,
-+ .maxauthsize = MD5_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_AES |
-+ OP_ALG_AAI_CTR_MOD128,
-+ .class2_alg_type = OP_ALG_ALGSEL_MD5 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ .rfc3686 = true,
-+ .geniv = true,
-+ },
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "authenc(hmac(sha1),"
-+ "rfc3686(ctr(aes)))",
-+ .cra_driver_name = "authenc-hmac-sha1-"
-+ "rfc3686-ctr-aes-caam-qi2",
-+ .cra_blocksize = 1,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = CTR_RFC3686_IV_SIZE,
-+ .maxauthsize = SHA1_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_AES |
-+ OP_ALG_AAI_CTR_MOD128,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA1 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ .rfc3686 = true,
-+ },
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "seqiv(authenc("
-+ "hmac(sha1),rfc3686(ctr(aes))))",
-+ .cra_driver_name = "seqiv-authenc-hmac-sha1-"
-+ "rfc3686-ctr-aes-caam-qi2",
-+ .cra_blocksize = 1,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = CTR_RFC3686_IV_SIZE,
-+ .maxauthsize = SHA1_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_AES |
-+ OP_ALG_AAI_CTR_MOD128,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA1 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ .rfc3686 = true,
-+ .geniv = true,
-+ },
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "authenc(hmac(sha224),"
-+ "rfc3686(ctr(aes)))",
-+ .cra_driver_name = "authenc-hmac-sha224-"
-+ "rfc3686-ctr-aes-caam-qi2",
-+ .cra_blocksize = 1,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = CTR_RFC3686_IV_SIZE,
-+ .maxauthsize = SHA224_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_AES |
-+ OP_ALG_AAI_CTR_MOD128,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA224 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ .rfc3686 = true,
-+ },
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "seqiv(authenc("
-+ "hmac(sha224),rfc3686(ctr(aes))))",
-+ .cra_driver_name = "seqiv-authenc-hmac-sha224-"
-+ "rfc3686-ctr-aes-caam-qi2",
-+ .cra_blocksize = 1,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = CTR_RFC3686_IV_SIZE,
-+ .maxauthsize = SHA224_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_AES |
-+ OP_ALG_AAI_CTR_MOD128,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA224 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ .rfc3686 = true,
-+ .geniv = true,
-+ },
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "authenc(hmac(sha256),"
-+ "rfc3686(ctr(aes)))",
-+ .cra_driver_name = "authenc-hmac-sha256-"
-+ "rfc3686-ctr-aes-caam-qi2",
-+ .cra_blocksize = 1,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = CTR_RFC3686_IV_SIZE,
-+ .maxauthsize = SHA256_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_AES |
-+ OP_ALG_AAI_CTR_MOD128,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA256 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ .rfc3686 = true,
-+ },
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "seqiv(authenc(hmac(sha256),"
-+ "rfc3686(ctr(aes))))",
-+ .cra_driver_name = "seqiv-authenc-hmac-sha256-"
-+ "rfc3686-ctr-aes-caam-qi2",
-+ .cra_blocksize = 1,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = CTR_RFC3686_IV_SIZE,
-+ .maxauthsize = SHA256_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_AES |
-+ OP_ALG_AAI_CTR_MOD128,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA256 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ .rfc3686 = true,
-+ .geniv = true,
-+ },
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "authenc(hmac(sha384),"
-+ "rfc3686(ctr(aes)))",
-+ .cra_driver_name = "authenc-hmac-sha384-"
-+ "rfc3686-ctr-aes-caam-qi2",
-+ .cra_blocksize = 1,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = CTR_RFC3686_IV_SIZE,
-+ .maxauthsize = SHA384_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_AES |
-+ OP_ALG_AAI_CTR_MOD128,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA384 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ .rfc3686 = true,
-+ },
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "seqiv(authenc(hmac(sha384),"
-+ "rfc3686(ctr(aes))))",
-+ .cra_driver_name = "seqiv-authenc-hmac-sha384-"
-+ "rfc3686-ctr-aes-caam-qi2",
-+ .cra_blocksize = 1,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = CTR_RFC3686_IV_SIZE,
-+ .maxauthsize = SHA384_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_AES |
-+ OP_ALG_AAI_CTR_MOD128,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA384 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ .rfc3686 = true,
-+ .geniv = true,
-+ },
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "authenc(hmac(sha512),"
-+ "rfc3686(ctr(aes)))",
-+ .cra_driver_name = "authenc-hmac-sha512-"
-+ "rfc3686-ctr-aes-caam-qi2",
-+ .cra_blocksize = 1,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = CTR_RFC3686_IV_SIZE,
-+ .maxauthsize = SHA512_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_AES |
-+ OP_ALG_AAI_CTR_MOD128,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA512 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ .rfc3686 = true,
-+ },
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "seqiv(authenc(hmac(sha512),"
-+ "rfc3686(ctr(aes))))",
-+ .cra_driver_name = "seqiv-authenc-hmac-sha512-"
-+ "rfc3686-ctr-aes-caam-qi2",
-+ .cra_blocksize = 1,
-+ },
-+ .setkey = aead_setkey,
-+ .setauthsize = aead_setauthsize,
-+ .encrypt = aead_encrypt,
-+ .decrypt = aead_decrypt,
-+ .ivsize = CTR_RFC3686_IV_SIZE,
-+ .maxauthsize = SHA512_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_AES |
-+ OP_ALG_AAI_CTR_MOD128,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA512 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ .rfc3686 = true,
-+ .geniv = true,
-+ },
-+ },
-+ {
-+ .aead = {
-+ .base = {
-+ .cra_name = "tls10(hmac(sha1),cbc(aes))",
-+ .cra_driver_name = "tls10-hmac-sha1-cbc-aes-caam-qi2",
-+ .cra_blocksize = AES_BLOCK_SIZE,
-+ },
-+ .setkey = tls_setkey,
-+ .setauthsize = tls_setauthsize,
-+ .encrypt = tls_encrypt,
-+ .decrypt = tls_decrypt,
-+ .ivsize = AES_BLOCK_SIZE,
-+ .maxauthsize = SHA1_DIGEST_SIZE,
-+ },
-+ .caam = {
-+ .class1_alg_type = OP_ALG_ALGSEL_AES | OP_ALG_AAI_CBC,
-+ .class2_alg_type = OP_ALG_ALGSEL_SHA1 |
-+ OP_ALG_AAI_HMAC_PRECOMP,
-+ },
-+ },
-+};
-+
-+static struct caam_crypto_alg *caam_alg_alloc(struct caam_alg_template
-+ *template)
-+{
-+ struct caam_crypto_alg *t_alg;
-+ struct crypto_alg *alg;
-+
-+ t_alg = kzalloc(sizeof(*t_alg), GFP_KERNEL);
-+ if (!t_alg)
-+ return ERR_PTR(-ENOMEM);
-+
-+ alg = &t_alg->crypto_alg;
-+
-+ snprintf(alg->cra_name, CRYPTO_MAX_ALG_NAME, "%s", template->name);
-+ snprintf(alg->cra_driver_name, CRYPTO_MAX_ALG_NAME, "%s",
-+ template->driver_name);
-+ alg->cra_module = THIS_MODULE;
-+ alg->cra_exit = caam_cra_exit;
-+ alg->cra_priority = CAAM_CRA_PRIORITY;
-+ alg->cra_blocksize = template->blocksize;
-+ alg->cra_alignmask = 0;
-+ alg->cra_ctxsize = sizeof(struct caam_ctx);
-+ alg->cra_flags = CRYPTO_ALG_ASYNC | CRYPTO_ALG_KERN_DRIVER_ONLY |
-+ template->type;
-+ switch (template->type) {
-+ case CRYPTO_ALG_TYPE_GIVCIPHER:
-+ alg->cra_init = caam_cra_init_ablkcipher;
-+ alg->cra_type = &crypto_givcipher_type;
-+ alg->cra_ablkcipher = template->template_ablkcipher;
-+ break;
-+ case CRYPTO_ALG_TYPE_ABLKCIPHER:
-+ alg->cra_init = caam_cra_init_ablkcipher;
-+ alg->cra_type = &crypto_ablkcipher_type;
-+ alg->cra_ablkcipher = template->template_ablkcipher;
-+ break;
-+ }
-+
-+ t_alg->caam.class1_alg_type = template->class1_alg_type;
-+ t_alg->caam.class2_alg_type = template->class2_alg_type;
-+
-+ return t_alg;
-+}
-+
-+static void caam_aead_alg_init(struct caam_aead_alg *t_alg)
-+{
-+ struct aead_alg *alg = &t_alg->aead;
-+
-+ alg->base.cra_module = THIS_MODULE;
-+ alg->base.cra_priority = CAAM_CRA_PRIORITY;
-+ alg->base.cra_ctxsize = sizeof(struct caam_ctx);
-+ alg->base.cra_flags = CRYPTO_ALG_ASYNC | CRYPTO_ALG_KERN_DRIVER_ONLY;
-+
-+ alg->init = caam_cra_init_aead;
-+ alg->exit = caam_cra_exit_aead;
-+}
-+
-+/* max hash key is max split key size */
-+#define CAAM_MAX_HASH_KEY_SIZE (SHA512_DIGEST_SIZE * 2)
-+
-+#define CAAM_MAX_HASH_BLOCK_SIZE SHA512_BLOCK_SIZE
-+#define CAAM_MAX_HASH_DIGEST_SIZE SHA512_DIGEST_SIZE
-+
-+#define DESC_HASH_MAX_USED_BYTES (DESC_AHASH_FINAL_LEN + \
-+ CAAM_MAX_HASH_KEY_SIZE)
-+#define DESC_HASH_MAX_USED_LEN (DESC_HASH_MAX_USED_BYTES / CAAM_CMD_SZ)
-+
-+/* caam context sizes for hashes: running digest + 8 */
-+#define HASH_MSG_LEN 8
-+#define MAX_CTX_LEN (HASH_MSG_LEN + SHA512_DIGEST_SIZE)
-+
-+enum hash_optype {
-+ UPDATE = 0,
-+ UPDATE_FIRST,
-+ FINALIZE,
-+ DIGEST,
-+ HASH_NUM_OP
-+};
-+
-+/**
-+ * caam_hash_ctx - ahash per-session context
-+ * @flc: Flow Contexts array
-+ * @flc_dma: I/O virtual addresses of the Flow Contexts
-+ * @key: virtual address of the authentication key
-+ * @dev: dpseci device
-+ * @ctx_len: size of Context Register
-+ * @adata: hashing algorithm details
-+ */
-+struct caam_hash_ctx {
-+ struct caam_flc flc[HASH_NUM_OP];
-+ dma_addr_t flc_dma[HASH_NUM_OP];
-+ u8 key[CAAM_MAX_HASH_KEY_SIZE];
-+ struct device *dev;
-+ int ctx_len;
-+ struct alginfo adata;
-+};
-+
-+/* ahash state */
-+struct caam_hash_state {
-+ struct caam_request caam_req;
-+ dma_addr_t buf_dma;
-+ dma_addr_t ctx_dma;
-+ u8 buf_0[CAAM_MAX_HASH_BLOCK_SIZE] ____cacheline_aligned;
-+ int buflen_0;
-+ u8 buf_1[CAAM_MAX_HASH_BLOCK_SIZE] ____cacheline_aligned;
-+ int buflen_1;
-+ u8 caam_ctx[MAX_CTX_LEN] ____cacheline_aligned;
-+ int (*update)(struct ahash_request *req);
-+ int (*final)(struct ahash_request *req);
-+ int (*finup)(struct ahash_request *req);
-+ int current_buf;
-+};
-+
-+struct caam_export_state {
-+ u8 buf[CAAM_MAX_HASH_BLOCK_SIZE];
-+ u8 caam_ctx[MAX_CTX_LEN];
-+ int buflen;
-+ int (*update)(struct ahash_request *req);
-+ int (*final)(struct ahash_request *req);
-+ int (*finup)(struct ahash_request *req);
-+};
-+
-+static inline void switch_buf(struct caam_hash_state *state)
-+{
-+ state->current_buf ^= 1;
-+}
-+
-+static inline u8 *current_buf(struct caam_hash_state *state)
-+{
-+ return state->current_buf ? state->buf_1 : state->buf_0;
-+}
-+
-+static inline u8 *alt_buf(struct caam_hash_state *state)
-+{
-+ return state->current_buf ? state->buf_0 : state->buf_1;
-+}
-+
-+static inline int *current_buflen(struct caam_hash_state *state)
-+{
-+ return state->current_buf ? &state->buflen_1 : &state->buflen_0;
-+}
-+
-+static inline int *alt_buflen(struct caam_hash_state *state)
-+{
-+ return state->current_buf ? &state->buflen_0 : &state->buflen_1;
-+}
-+
-+/* Map current buffer in state (if length > 0) and put it in link table */
-+static inline int buf_map_to_qm_sg(struct device *dev,
-+ struct dpaa2_sg_entry *qm_sg,
-+ struct caam_hash_state *state)
-+{
-+ int buflen = *current_buflen(state);
-+
-+ if (!buflen)
-+ return 0;
-+
-+ state->buf_dma = dma_map_single(dev, current_buf(state), buflen,
-+ DMA_TO_DEVICE);
-+ if (dma_mapping_error(dev, state->buf_dma)) {
-+ dev_err(dev, "unable to map buf\n");
-+ state->buf_dma = 0;
-+ return -ENOMEM;
-+ }
-+
-+ dma_to_qm_sg_one(qm_sg, state->buf_dma, buflen, 0);
-+
-+ return 0;
-+}
-+
-+/* Map state->caam_ctx, and add it to link table */
-+static inline int ctx_map_to_qm_sg(struct device *dev,
-+ struct caam_hash_state *state, int ctx_len,
-+ struct dpaa2_sg_entry *qm_sg, u32 flag)
-+{
-+ state->ctx_dma = dma_map_single(dev, state->caam_ctx, ctx_len, flag);
-+ if (dma_mapping_error(dev, state->ctx_dma)) {
-+ dev_err(dev, "unable to map ctx\n");
-+ state->ctx_dma = 0;
-+ return -ENOMEM;
-+ }
-+
-+ dma_to_qm_sg_one(qm_sg, state->ctx_dma, ctx_len, 0);
-+
-+ return 0;
-+}
-+
-+static int ahash_set_sh_desc(struct crypto_ahash *ahash)
-+{
-+ struct caam_hash_ctx *ctx = crypto_ahash_ctx(ahash);
-+ int digestsize = crypto_ahash_digestsize(ahash);
-+ struct dpaa2_caam_priv *priv = dev_get_drvdata(ctx->dev);
-+ struct caam_flc *flc;
-+ u32 *desc;
-+
-+ ctx->adata.key_virt = ctx->key;
-+ ctx->adata.key_inline = true;
-+
-+ /* ahash_update shared descriptor */
-+ flc = &ctx->flc[UPDATE];
-+ desc = flc->sh_desc;
-+ cnstr_shdsc_ahash(desc, &ctx->adata, OP_ALG_AS_UPDATE, ctx->ctx_len,
-+ ctx->ctx_len, true, priv->sec_attr.era);
-+ flc->flc[1] = cpu_to_caam32(desc_len(desc)); /* SDL */
-+ dma_sync_single_for_device(ctx->dev, ctx->flc_dma[UPDATE],
-+ desc_bytes(desc), DMA_BIDIRECTIONAL);
-+#ifdef DEBUG
-+ print_hex_dump(KERN_ERR,
-+ "ahash update shdesc@" __stringify(__LINE__)": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, desc, desc_bytes(desc), 1);
-+#endif
-+
-+ /* ahash_update_first shared descriptor */
-+ flc = &ctx->flc[UPDATE_FIRST];
-+ desc = flc->sh_desc;
-+ cnstr_shdsc_ahash(desc, &ctx->adata, OP_ALG_AS_INIT, ctx->ctx_len,
-+ ctx->ctx_len, false, priv->sec_attr.era);
-+ flc->flc[1] = cpu_to_caam32(desc_len(desc)); /* SDL */
-+ dma_sync_single_for_device(ctx->dev, ctx->flc_dma[UPDATE_FIRST],
-+ desc_bytes(desc), DMA_BIDIRECTIONAL);
-+#ifdef DEBUG
-+ print_hex_dump(KERN_ERR,
-+ "ahash update first shdesc@" __stringify(__LINE__)": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, desc, desc_bytes(desc), 1);
-+#endif
-+
-+ /* ahash_final shared descriptor */
-+ flc = &ctx->flc[FINALIZE];
-+ desc = flc->sh_desc;
-+ cnstr_shdsc_ahash(desc, &ctx->adata, OP_ALG_AS_FINALIZE, digestsize,
-+ ctx->ctx_len, true, priv->sec_attr.era);
-+ flc->flc[1] = cpu_to_caam32(desc_len(desc)); /* SDL */
-+ dma_sync_single_for_device(ctx->dev, ctx->flc_dma[FINALIZE],
-+ desc_bytes(desc), DMA_BIDIRECTIONAL);
-+#ifdef DEBUG
-+ print_hex_dump(KERN_ERR,
-+ "ahash final shdesc@" __stringify(__LINE__)": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, desc, desc_bytes(desc), 1);
-+#endif
-+
-+ /* ahash_digest shared descriptor */
-+ flc = &ctx->flc[DIGEST];
-+ desc = flc->sh_desc;
-+ cnstr_shdsc_ahash(desc, &ctx->adata, OP_ALG_AS_INITFINAL, digestsize,
-+ ctx->ctx_len, false, priv->sec_attr.era);
-+ flc->flc[1] = cpu_to_caam32(desc_len(desc)); /* SDL */
-+ dma_sync_single_for_device(ctx->dev, ctx->flc_dma[DIGEST],
-+ desc_bytes(desc), DMA_BIDIRECTIONAL);
-+#ifdef DEBUG
-+ print_hex_dump(KERN_ERR,
-+ "ahash digest shdesc@" __stringify(__LINE__)": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, desc, desc_bytes(desc), 1);
-+#endif
-+
-+ return 0;
-+}
-+
-+/* Digest hash size if it is too large */
-+static int hash_digest_key(struct caam_hash_ctx *ctx, const u8 *key_in,
-+ u32 *keylen, u8 *key_out, u32 digestsize)
-+{
-+ struct caam_request *req_ctx;
-+ u32 *desc;
-+ struct split_key_sh_result result;
-+ dma_addr_t src_dma, dst_dma;
-+ struct caam_flc *flc;
-+ dma_addr_t flc_dma;
-+ int ret = -ENOMEM;
-+ struct dpaa2_fl_entry *in_fle, *out_fle;
-+
-+ req_ctx = kzalloc(sizeof(*req_ctx), GFP_KERNEL | GFP_DMA);
-+ if (!req_ctx)
-+ return -ENOMEM;
-+
-+ in_fle = &req_ctx->fd_flt[1];
-+ out_fle = &req_ctx->fd_flt[0];
-+
-+ flc = kzalloc(sizeof(*flc), GFP_KERNEL | GFP_DMA);
-+ if (!flc)
-+ goto err_flc;
-+
-+ src_dma = dma_map_single(ctx->dev, (void *)key_in, *keylen,
-+ DMA_TO_DEVICE);
-+ if (dma_mapping_error(ctx->dev, src_dma)) {
-+ dev_err(ctx->dev, "unable to map key input memory\n");
-+ goto err_src_dma;
-+ }
-+ dst_dma = dma_map_single(ctx->dev, (void *)key_out, digestsize,
-+ DMA_FROM_DEVICE);
-+ if (dma_mapping_error(ctx->dev, dst_dma)) {
-+ dev_err(ctx->dev, "unable to map key output memory\n");
-+ goto err_dst_dma;
-+ }
-+
-+ desc = flc->sh_desc;
-+
-+ init_sh_desc(desc, 0);
-+
-+ /* descriptor to perform unkeyed hash on key_in */
-+ append_operation(desc, ctx->adata.algtype | OP_ALG_ENCRYPT |
-+ OP_ALG_AS_INITFINAL);
-+ append_seq_fifo_load(desc, *keylen, FIFOLD_CLASS_CLASS2 |
-+ FIFOLD_TYPE_LAST2 | FIFOLD_TYPE_MSG);
-+ append_seq_store(desc, digestsize, LDST_CLASS_2_CCB |
-+ LDST_SRCDST_BYTE_CONTEXT);
-+
-+ flc->flc[1] = cpu_to_caam32(desc_len(desc)); /* SDL */
-+ flc_dma = dma_map_single(ctx->dev, flc, sizeof(flc->flc) +
-+ desc_bytes(desc), DMA_TO_DEVICE);
-+ if (dma_mapping_error(ctx->dev, flc_dma)) {
-+ dev_err(ctx->dev, "unable to map shared descriptor\n");
-+ goto err_flc_dma;
-+ }
-+
-+ dpaa2_fl_set_final(in_fle, true);
-+ dpaa2_fl_set_format(in_fle, dpaa2_fl_single);
-+ dpaa2_fl_set_addr(in_fle, src_dma);
-+ dpaa2_fl_set_len(in_fle, *keylen);
-+ dpaa2_fl_set_format(out_fle, dpaa2_fl_single);
-+ dpaa2_fl_set_addr(out_fle, dst_dma);
-+ dpaa2_fl_set_len(out_fle, digestsize);
-+
-+#ifdef DEBUG
-+ print_hex_dump(KERN_ERR, "key_in@" __stringify(__LINE__)": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, key_in, *keylen, 1);
-+ print_hex_dump(KERN_ERR, "shdesc@" __stringify(__LINE__)": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, desc, desc_bytes(desc), 1);
-+#endif
-+
-+ result.err = 0;
-+ init_completion(&result.completion);
-+ result.dev = ctx->dev;
-+
-+ req_ctx->flc = flc;
-+ req_ctx->flc_dma = flc_dma;
-+ req_ctx->cbk = split_key_sh_done;
-+ req_ctx->ctx = &result;
-+
-+ ret = dpaa2_caam_enqueue(ctx->dev, req_ctx);
-+ if (ret == -EINPROGRESS) {
-+ /* in progress */
-+ wait_for_completion(&result.completion);
-+ ret = result.err;
-+#ifdef DEBUG
-+ print_hex_dump(KERN_ERR,
-+ "digested key@" __stringify(__LINE__)": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, key_in, digestsize,
-+ 1);
-+#endif
-+ }
-+
-+ dma_unmap_single(ctx->dev, flc_dma, sizeof(flc->flc) + desc_bytes(desc),
-+ DMA_TO_DEVICE);
-+err_flc_dma:
-+ dma_unmap_single(ctx->dev, dst_dma, digestsize, DMA_FROM_DEVICE);
-+err_dst_dma:
-+ dma_unmap_single(ctx->dev, src_dma, *keylen, DMA_TO_DEVICE);
-+err_src_dma:
-+ kfree(flc);
-+err_flc:
-+ kfree(req_ctx);
-+
-+ *keylen = digestsize;
-+
-+ return ret;
-+}
-+
-+static int ahash_setkey(struct crypto_ahash *ahash, const u8 *key,
-+ unsigned int keylen)
-+{
-+ struct caam_hash_ctx *ctx = crypto_ahash_ctx(ahash);
-+ unsigned int blocksize = crypto_tfm_alg_blocksize(&ahash->base);
-+ unsigned int digestsize = crypto_ahash_digestsize(ahash);
-+ int ret;
-+ u8 *hashed_key = NULL;
-+
-+#ifdef DEBUG
-+ dev_err(ctx->dev, "keylen %d blocksize %d\n", keylen, blocksize);
-+#endif
-+
-+ if (keylen > blocksize) {
-+ hashed_key = kmalloc_array(digestsize, sizeof(*hashed_key),
-+ GFP_KERNEL | GFP_DMA);
-+ if (!hashed_key)
-+ return -ENOMEM;
-+ ret = hash_digest_key(ctx, key, &keylen, hashed_key,
-+ digestsize);
-+ if (ret)
-+ goto bad_free_key;
-+ key = hashed_key;
-+ }
-+
-+ ctx->adata.keylen = keylen;
-+ ctx->adata.keylen_pad = split_key_len(ctx->adata.algtype &
-+ OP_ALG_ALGSEL_MASK);
-+ if (ctx->adata.keylen_pad > CAAM_MAX_HASH_KEY_SIZE)
-+ goto bad_free_key;
-+
-+ memcpy(ctx->key, key, keylen);
-+
-+ kfree(hashed_key);
-+ return ahash_set_sh_desc(ahash);
-+bad_free_key:
-+ kfree(hashed_key);
-+ crypto_ahash_set_flags(ahash, CRYPTO_TFM_RES_BAD_KEY_LEN);
-+ return -EINVAL;
-+}
-+
-+static inline void ahash_unmap(struct device *dev, struct ahash_edesc *edesc,
-+ struct ahash_request *req, int dst_len)
-+{
-+ struct caam_hash_state *state = ahash_request_ctx(req);
-+
-+ if (edesc->src_nents)
-+ dma_unmap_sg(dev, req->src, edesc->src_nents, DMA_TO_DEVICE);
-+ if (edesc->dst_dma)
-+ dma_unmap_single(dev, edesc->dst_dma, dst_len, DMA_FROM_DEVICE);
-+
-+ if (edesc->qm_sg_bytes)
-+ dma_unmap_single(dev, edesc->qm_sg_dma, edesc->qm_sg_bytes,
-+ DMA_TO_DEVICE);
-+
-+ if (state->buf_dma) {
-+ dma_unmap_single(dev, state->buf_dma, *current_buflen(state),
-+ DMA_TO_DEVICE);
-+ state->buf_dma = 0;
-+ }
-+}
-+
-+static inline void ahash_unmap_ctx(struct device *dev,
-+ struct ahash_edesc *edesc,
-+ struct ahash_request *req, int dst_len,
-+ u32 flag)
-+{
-+ struct crypto_ahash *ahash = crypto_ahash_reqtfm(req);
-+ struct caam_hash_ctx *ctx = crypto_ahash_ctx(ahash);
-+ struct caam_hash_state *state = ahash_request_ctx(req);
-+
-+ if (state->ctx_dma) {
-+ dma_unmap_single(dev, state->ctx_dma, ctx->ctx_len, flag);
-+ state->ctx_dma = 0;
-+ }
-+ ahash_unmap(dev, edesc, req, dst_len);
-+}
-+
-+static void ahash_done(void *cbk_ctx, u32 status)
-+{
-+ struct crypto_async_request *areq = cbk_ctx;
-+ struct ahash_request *req = ahash_request_cast(areq);
-+ struct crypto_ahash *ahash = crypto_ahash_reqtfm(req);
-+ struct caam_hash_state *state = ahash_request_ctx(req);
-+ struct ahash_edesc *edesc = state->caam_req.edesc;
-+ struct caam_hash_ctx *ctx = crypto_ahash_ctx(ahash);
-+ int digestsize = crypto_ahash_digestsize(ahash);
-+ int ecode = 0;
-+
-+#ifdef DEBUG
-+ dev_err(ctx->dev, "%s %d: err 0x%x\n", __func__, __LINE__, status);
-+#endif
-+
-+ if (unlikely(status)) {
-+ caam_qi2_strstatus(ctx->dev, status);
-+ ecode = -EIO;
-+ }
-+
-+ ahash_unmap(ctx->dev, edesc, req, digestsize);
-+ qi_cache_free(edesc);
-+
-+#ifdef DEBUG
-+ print_hex_dump(KERN_ERR, "ctx@" __stringify(__LINE__)": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, state->caam_ctx,
-+ ctx->ctx_len, 1);
-+ if (req->result)
-+ print_hex_dump(KERN_ERR, "result@" __stringify(__LINE__)": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, req->result,
-+ digestsize, 1);
-+#endif
-+
-+ req->base.complete(&req->base, ecode);
-+}
-+
-+static void ahash_done_bi(void *cbk_ctx, u32 status)
-+{
-+ struct crypto_async_request *areq = cbk_ctx;
-+ struct ahash_request *req = ahash_request_cast(areq);
-+ struct crypto_ahash *ahash = crypto_ahash_reqtfm(req);
-+ struct caam_hash_state *state = ahash_request_ctx(req);
-+ struct ahash_edesc *edesc = state->caam_req.edesc;
-+ struct caam_hash_ctx *ctx = crypto_ahash_ctx(ahash);
-+ int ecode = 0;
-+#ifdef DEBUG
-+ int digestsize = crypto_ahash_digestsize(ahash);
-+
-+ dev_err(ctx->dev, "%s %d: err 0x%x\n", __func__, __LINE__, status);
-+#endif
-+
-+ if (unlikely(status)) {
-+ caam_qi2_strstatus(ctx->dev, status);
-+ ecode = -EIO;
-+ }
-+
-+ ahash_unmap_ctx(ctx->dev, edesc, req, ctx->ctx_len, DMA_BIDIRECTIONAL);
-+ switch_buf(state);
-+ qi_cache_free(edesc);
-+
-+#ifdef DEBUG
-+ print_hex_dump(KERN_ERR, "ctx@" __stringify(__LINE__)": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, state->caam_ctx,
-+ ctx->ctx_len, 1);
-+ if (req->result)
-+ print_hex_dump(KERN_ERR, "result@" __stringify(__LINE__)": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, req->result,
-+ digestsize, 1);
-+#endif
-+
-+ req->base.complete(&req->base, ecode);
-+}
-+
-+static void ahash_done_ctx_src(void *cbk_ctx, u32 status)
-+{
-+ struct crypto_async_request *areq = cbk_ctx;
-+ struct ahash_request *req = ahash_request_cast(areq);
-+ struct crypto_ahash *ahash = crypto_ahash_reqtfm(req);
-+ struct caam_hash_state *state = ahash_request_ctx(req);
-+ struct ahash_edesc *edesc = state->caam_req.edesc;
-+ struct caam_hash_ctx *ctx = crypto_ahash_ctx(ahash);
-+ int digestsize = crypto_ahash_digestsize(ahash);
-+ int ecode = 0;
-+
-+#ifdef DEBUG
-+ dev_err(ctx->dev, "%s %d: err 0x%x\n", __func__, __LINE__, status);
-+#endif
-+
-+ if (unlikely(status)) {
-+ caam_qi2_strstatus(ctx->dev, status);
-+ ecode = -EIO;
-+ }
-+
-+ ahash_unmap_ctx(ctx->dev, edesc, req, digestsize, DMA_TO_DEVICE);
-+ qi_cache_free(edesc);
-+
-+#ifdef DEBUG
-+ print_hex_dump(KERN_ERR, "ctx@" __stringify(__LINE__)": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, state->caam_ctx,
-+ ctx->ctx_len, 1);
-+ if (req->result)
-+ print_hex_dump(KERN_ERR, "result@" __stringify(__LINE__)": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, req->result,
-+ digestsize, 1);
-+#endif
-+
-+ req->base.complete(&req->base, ecode);
-+}
-+
-+static void ahash_done_ctx_dst(void *cbk_ctx, u32 status)
-+{
-+ struct crypto_async_request *areq = cbk_ctx;
-+ struct ahash_request *req = ahash_request_cast(areq);
-+ struct crypto_ahash *ahash = crypto_ahash_reqtfm(req);
-+ struct caam_hash_state *state = ahash_request_ctx(req);
-+ struct ahash_edesc *edesc = state->caam_req.edesc;
-+ struct caam_hash_ctx *ctx = crypto_ahash_ctx(ahash);
-+ int ecode = 0;
-+#ifdef DEBUG
-+ int digestsize = crypto_ahash_digestsize(ahash);
-+
-+ dev_err(ctx->dev, "%s %d: err 0x%x\n", __func__, __LINE__, status);
-+#endif
-+
-+ if (unlikely(status)) {
-+ caam_qi2_strstatus(ctx->dev, status);
-+ ecode = -EIO;
-+ }
-+
-+ ahash_unmap_ctx(ctx->dev, edesc, req, ctx->ctx_len, DMA_FROM_DEVICE);
-+ switch_buf(state);
-+ qi_cache_free(edesc);
-+
-+#ifdef DEBUG
-+ print_hex_dump(KERN_ERR, "ctx@" __stringify(__LINE__)": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, state->caam_ctx,
-+ ctx->ctx_len, 1);
-+ if (req->result)
-+ print_hex_dump(KERN_ERR, "result@" __stringify(__LINE__)": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, req->result,
-+ digestsize, 1);
-+#endif
-+
-+ req->base.complete(&req->base, ecode);
-+}
-+
-+static int ahash_update_ctx(struct ahash_request *req)
-+{
-+ struct crypto_ahash *ahash = crypto_ahash_reqtfm(req);
-+ struct caam_hash_ctx *ctx = crypto_ahash_ctx(ahash);
-+ struct caam_hash_state *state = ahash_request_ctx(req);
-+ struct caam_request *req_ctx = &state->caam_req;
-+ struct dpaa2_fl_entry *in_fle = &req_ctx->fd_flt[1];
-+ struct dpaa2_fl_entry *out_fle = &req_ctx->fd_flt[0];
-+ gfp_t flags = (req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP) ?
-+ GFP_KERNEL : GFP_ATOMIC;
-+ u8 *buf = current_buf(state);
-+ int *buflen = current_buflen(state);
-+ u8 *next_buf = alt_buf(state);
-+ int *next_buflen = alt_buflen(state), last_buflen;
-+ int in_len = *buflen + req->nbytes, to_hash;
-+ int src_nents, mapped_nents, qm_sg_bytes, qm_sg_src_index;
-+ struct ahash_edesc *edesc;
-+ int ret = 0;
-+
-+ last_buflen = *next_buflen;
-+ *next_buflen = in_len & (crypto_tfm_alg_blocksize(&ahash->base) - 1);
-+ to_hash = in_len - *next_buflen;
-+
-+ if (to_hash) {
-+ struct dpaa2_sg_entry *sg_table;
-+
-+ src_nents = sg_nents_for_len(req->src,
-+ req->nbytes - (*next_buflen));
-+ if (src_nents < 0) {
-+ dev_err(ctx->dev, "Invalid number of src SG.\n");
-+ return src_nents;
-+ }
-+
-+ if (src_nents) {
-+ mapped_nents = dma_map_sg(ctx->dev, req->src, src_nents,
-+ DMA_TO_DEVICE);
-+ if (!mapped_nents) {
-+ dev_err(ctx->dev, "unable to DMA map source\n");
-+ return -ENOMEM;
-+ }
-+ } else {
-+ mapped_nents = 0;
-+ }
-+
-+ /* allocate space for base edesc and link tables */
-+ edesc = qi_cache_zalloc(GFP_DMA | flags);
-+ if (!edesc) {
-+ dma_unmap_sg(ctx->dev, req->src, src_nents,
-+ DMA_TO_DEVICE);
-+ return -ENOMEM;
-+ }
-+
-+ edesc->src_nents = src_nents;
-+ qm_sg_src_index = 1 + (*buflen ? 1 : 0);
-+ qm_sg_bytes = (qm_sg_src_index + mapped_nents) *
-+ sizeof(*sg_table);
-+ sg_table = &edesc->sgt[0];
-+
-+ ret = ctx_map_to_qm_sg(ctx->dev, state, ctx->ctx_len, sg_table,
-+ DMA_BIDIRECTIONAL);
-+ if (ret)
-+ goto unmap_ctx;
-+
-+ ret = buf_map_to_qm_sg(ctx->dev, sg_table + 1, state);
-+ if (ret)
-+ goto unmap_ctx;
-+
-+ if (mapped_nents) {
-+ sg_to_qm_sg_last(req->src, mapped_nents,
-+ sg_table + qm_sg_src_index, 0);
-+ if (*next_buflen)
-+ scatterwalk_map_and_copy(next_buf, req->src,
-+ to_hash - *buflen,
-+ *next_buflen, 0);
-+ } else {
-+ dpaa2_sg_set_final(sg_table + qm_sg_src_index - 1,
-+ true);
-+ }
-+
-+ edesc->qm_sg_dma = dma_map_single(ctx->dev, sg_table,
-+ qm_sg_bytes, DMA_TO_DEVICE);
-+ if (dma_mapping_error(ctx->dev, edesc->qm_sg_dma)) {
-+ dev_err(ctx->dev, "unable to map S/G table\n");
-+ ret = -ENOMEM;
-+ goto unmap_ctx;
-+ }
-+ edesc->qm_sg_bytes = qm_sg_bytes;
-+
-+ memset(&req_ctx->fd_flt, 0, sizeof(req_ctx->fd_flt));
-+ dpaa2_fl_set_final(in_fle, true);
-+ dpaa2_fl_set_format(in_fle, dpaa2_fl_sg);
-+ dpaa2_fl_set_addr(in_fle, edesc->qm_sg_dma);
-+ dpaa2_fl_set_len(in_fle, ctx->ctx_len + to_hash);
-+ dpaa2_fl_set_format(out_fle, dpaa2_fl_single);
-+ dpaa2_fl_set_addr(out_fle, state->ctx_dma);
-+ dpaa2_fl_set_len(out_fle, ctx->ctx_len);
-+
-+ req_ctx->flc = &ctx->flc[UPDATE];
-+ req_ctx->flc_dma = ctx->flc_dma[UPDATE];
-+ req_ctx->cbk = ahash_done_bi;
-+ req_ctx->ctx = &req->base;
-+ req_ctx->edesc = edesc;
-+
-+ ret = dpaa2_caam_enqueue(ctx->dev, req_ctx);
-+ if (ret != -EINPROGRESS &&
-+ !(ret == -EBUSY &&
-+ req->base.flags & CRYPTO_TFM_REQ_MAY_BACKLOG))
-+ goto unmap_ctx;
-+ } else if (*next_buflen) {
-+ scatterwalk_map_and_copy(buf + *buflen, req->src, 0,
-+ req->nbytes, 0);
-+ *buflen = *next_buflen;
-+ *next_buflen = last_buflen;
-+ }
-+#ifdef DEBUG
-+ print_hex_dump(KERN_ERR, "buf@" __stringify(__LINE__)": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, buf, *buflen, 1);
-+ print_hex_dump(KERN_ERR, "next buf@" __stringify(__LINE__)": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, next_buf,
-+ *next_buflen, 1);
-+#endif
-+
-+ return ret;
-+unmap_ctx:
-+ ahash_unmap_ctx(ctx->dev, edesc, req, ctx->ctx_len, DMA_BIDIRECTIONAL);
-+ qi_cache_free(edesc);
-+ return ret;
-+}
-+
-+static int ahash_final_ctx(struct ahash_request *req)
-+{
-+ struct crypto_ahash *ahash = crypto_ahash_reqtfm(req);
-+ struct caam_hash_ctx *ctx = crypto_ahash_ctx(ahash);
-+ struct caam_hash_state *state = ahash_request_ctx(req);
-+ struct caam_request *req_ctx = &state->caam_req;
-+ struct dpaa2_fl_entry *in_fle = &req_ctx->fd_flt[1];
-+ struct dpaa2_fl_entry *out_fle = &req_ctx->fd_flt[0];
-+ gfp_t flags = (req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP) ?
-+ GFP_KERNEL : GFP_ATOMIC;
-+ int buflen = *current_buflen(state);
-+ int qm_sg_bytes, qm_sg_src_index;
-+ int digestsize = crypto_ahash_digestsize(ahash);
-+ struct ahash_edesc *edesc;
-+ struct dpaa2_sg_entry *sg_table;
-+ int ret;
-+
-+ /* allocate space for base edesc and link tables */
-+ edesc = qi_cache_zalloc(GFP_DMA | flags);
-+ if (!edesc)
-+ return -ENOMEM;
-+
-+ qm_sg_src_index = 1 + (buflen ? 1 : 0);
-+ qm_sg_bytes = qm_sg_src_index * sizeof(*sg_table);
-+ sg_table = &edesc->sgt[0];
-+
-+ ret = ctx_map_to_qm_sg(ctx->dev, state, ctx->ctx_len, sg_table,
-+ DMA_TO_DEVICE);
-+ if (ret)
-+ goto unmap_ctx;
-+
-+ ret = buf_map_to_qm_sg(ctx->dev, sg_table + 1, state);
-+ if (ret)
-+ goto unmap_ctx;
-+
-+ dpaa2_sg_set_final(sg_table + qm_sg_src_index - 1, true);
-+
-+ edesc->qm_sg_dma = dma_map_single(ctx->dev, sg_table, qm_sg_bytes,
-+ DMA_TO_DEVICE);
-+ if (dma_mapping_error(ctx->dev, edesc->qm_sg_dma)) {
-+ dev_err(ctx->dev, "unable to map S/G table\n");
-+ ret = -ENOMEM;
-+ goto unmap_ctx;
-+ }
-+ edesc->qm_sg_bytes = qm_sg_bytes;
-+
-+ edesc->dst_dma = dma_map_single(ctx->dev, req->result, digestsize,
-+ DMA_FROM_DEVICE);
-+ if (dma_mapping_error(ctx->dev, edesc->dst_dma)) {
-+ dev_err(ctx->dev, "unable to map dst\n");
-+ edesc->dst_dma = 0;
-+ ret = -ENOMEM;
-+ goto unmap_ctx;
-+ }
-+
-+ memset(&req_ctx->fd_flt, 0, sizeof(req_ctx->fd_flt));
-+ dpaa2_fl_set_final(in_fle, true);
-+ dpaa2_fl_set_format(in_fle, dpaa2_fl_sg);
-+ dpaa2_fl_set_addr(in_fle, edesc->qm_sg_dma);
-+ dpaa2_fl_set_len(in_fle, ctx->ctx_len + buflen);
-+ dpaa2_fl_set_format(out_fle, dpaa2_fl_single);
-+ dpaa2_fl_set_addr(out_fle, edesc->dst_dma);
-+ dpaa2_fl_set_len(out_fle, digestsize);
-+
-+ req_ctx->flc = &ctx->flc[FINALIZE];
-+ req_ctx->flc_dma = ctx->flc_dma[FINALIZE];
-+ req_ctx->cbk = ahash_done_ctx_src;
-+ req_ctx->ctx = &req->base;
-+ req_ctx->edesc = edesc;
-+
-+ ret = dpaa2_caam_enqueue(ctx->dev, req_ctx);
-+ if (ret == -EINPROGRESS ||
-+ (ret == -EBUSY && req->base.flags & CRYPTO_TFM_REQ_MAY_BACKLOG))
-+ return ret;
-+
-+unmap_ctx:
-+ ahash_unmap_ctx(ctx->dev, edesc, req, digestsize, DMA_FROM_DEVICE);
-+ qi_cache_free(edesc);
-+ return ret;
-+}
-+
-+static int ahash_finup_ctx(struct ahash_request *req)
-+{
-+ struct crypto_ahash *ahash = crypto_ahash_reqtfm(req);
-+ struct caam_hash_ctx *ctx = crypto_ahash_ctx(ahash);
-+ struct caam_hash_state *state = ahash_request_ctx(req);
-+ struct caam_request *req_ctx = &state->caam_req;
-+ struct dpaa2_fl_entry *in_fle = &req_ctx->fd_flt[1];
-+ struct dpaa2_fl_entry *out_fle = &req_ctx->fd_flt[0];
-+ gfp_t flags = (req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP) ?
-+ GFP_KERNEL : GFP_ATOMIC;
-+ int buflen = *current_buflen(state);
-+ int qm_sg_bytes, qm_sg_src_index;
-+ int src_nents, mapped_nents;
-+ int digestsize = crypto_ahash_digestsize(ahash);
-+ struct ahash_edesc *edesc;
-+ struct dpaa2_sg_entry *sg_table;
-+ int ret;
-+
-+ src_nents = sg_nents_for_len(req->src, req->nbytes);
-+ if (src_nents < 0) {
-+ dev_err(ctx->dev, "Invalid number of src SG.\n");
-+ return src_nents;
-+ }
-+
-+ if (src_nents) {
-+ mapped_nents = dma_map_sg(ctx->dev, req->src, src_nents,
-+ DMA_TO_DEVICE);
-+ if (!mapped_nents) {
-+ dev_err(ctx->dev, "unable to DMA map source\n");
-+ return -ENOMEM;
-+ }
-+ } else {
-+ mapped_nents = 0;
-+ }
-+
-+ /* allocate space for base edesc and link tables */
-+ edesc = qi_cache_zalloc(GFP_DMA | flags);
-+ if (!edesc) {
-+ dma_unmap_sg(ctx->dev, req->src, src_nents, DMA_TO_DEVICE);
-+ return -ENOMEM;
-+ }
-+
-+ edesc->src_nents = src_nents;
-+ qm_sg_src_index = 1 + (buflen ? 1 : 0);
-+ qm_sg_bytes = (qm_sg_src_index + mapped_nents) * sizeof(*sg_table);
-+ sg_table = &edesc->sgt[0];
-+
-+ ret = ctx_map_to_qm_sg(ctx->dev, state, ctx->ctx_len, sg_table,
-+ DMA_TO_DEVICE);
-+ if (ret)
-+ goto unmap_ctx;
-+
-+ ret = buf_map_to_qm_sg(ctx->dev, sg_table + 1, state);
-+ if (ret)
-+ goto unmap_ctx;
-+
-+ sg_to_qm_sg_last(req->src, mapped_nents, sg_table + qm_sg_src_index, 0);
-+
-+ edesc->qm_sg_dma = dma_map_single(ctx->dev, sg_table, qm_sg_bytes,
-+ DMA_TO_DEVICE);
-+ if (dma_mapping_error(ctx->dev, edesc->qm_sg_dma)) {
-+ dev_err(ctx->dev, "unable to map S/G table\n");
-+ ret = -ENOMEM;
-+ goto unmap_ctx;
-+ }
-+ edesc->qm_sg_bytes = qm_sg_bytes;
-+
-+ edesc->dst_dma = dma_map_single(ctx->dev, req->result, digestsize,
-+ DMA_FROM_DEVICE);
-+ if (dma_mapping_error(ctx->dev, edesc->dst_dma)) {
-+ dev_err(ctx->dev, "unable to map dst\n");
-+ edesc->dst_dma = 0;
-+ ret = -ENOMEM;
-+ goto unmap_ctx;
-+ }
-+
-+ memset(&req_ctx->fd_flt, 0, sizeof(req_ctx->fd_flt));
-+ dpaa2_fl_set_final(in_fle, true);
-+ dpaa2_fl_set_format(in_fle, dpaa2_fl_sg);
-+ dpaa2_fl_set_addr(in_fle, edesc->qm_sg_dma);
-+ dpaa2_fl_set_len(in_fle, ctx->ctx_len + buflen + req->nbytes);
-+ dpaa2_fl_set_format(out_fle, dpaa2_fl_single);
-+ dpaa2_fl_set_addr(out_fle, edesc->dst_dma);
-+ dpaa2_fl_set_len(out_fle, digestsize);
-+
-+ req_ctx->flc = &ctx->flc[FINALIZE];
-+ req_ctx->flc_dma = ctx->flc_dma[FINALIZE];
-+ req_ctx->cbk = ahash_done_ctx_src;
-+ req_ctx->ctx = &req->base;
-+ req_ctx->edesc = edesc;
-+
-+ ret = dpaa2_caam_enqueue(ctx->dev, req_ctx);
-+ if (ret == -EINPROGRESS ||
-+ (ret == -EBUSY && req->base.flags & CRYPTO_TFM_REQ_MAY_BACKLOG))
-+ return ret;
-+
-+unmap_ctx:
-+ ahash_unmap_ctx(ctx->dev, edesc, req, digestsize, DMA_FROM_DEVICE);
-+ qi_cache_free(edesc);
-+ return ret;
-+}
-+
-+static int ahash_digest(struct ahash_request *req)
-+{
-+ struct crypto_ahash *ahash = crypto_ahash_reqtfm(req);
-+ struct caam_hash_ctx *ctx = crypto_ahash_ctx(ahash);
-+ struct caam_hash_state *state = ahash_request_ctx(req);
-+ struct caam_request *req_ctx = &state->caam_req;
-+ struct dpaa2_fl_entry *in_fle = &req_ctx->fd_flt[1];
-+ struct dpaa2_fl_entry *out_fle = &req_ctx->fd_flt[0];
-+ gfp_t flags = (req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP) ?
-+ GFP_KERNEL : GFP_ATOMIC;
-+ int digestsize = crypto_ahash_digestsize(ahash);
-+ int src_nents, mapped_nents;
-+ struct ahash_edesc *edesc;
-+ int ret = -ENOMEM;
-+
-+ state->buf_dma = 0;
-+
-+ src_nents = sg_nents_for_len(req->src, req->nbytes);
-+ if (src_nents < 0) {
-+ dev_err(ctx->dev, "Invalid number of src SG.\n");
-+ return src_nents;
-+ }
-+
-+ if (src_nents) {
-+ mapped_nents = dma_map_sg(ctx->dev, req->src, src_nents,
-+ DMA_TO_DEVICE);
-+ if (!mapped_nents) {
-+ dev_err(ctx->dev, "unable to map source for DMA\n");
-+ return ret;
-+ }
-+ } else {
-+ mapped_nents = 0;
-+ }
-+
-+ /* allocate space for base edesc and link tables */
-+ edesc = qi_cache_zalloc(GFP_DMA | flags);
-+ if (!edesc) {
-+ dma_unmap_sg(ctx->dev, req->src, src_nents, DMA_TO_DEVICE);
-+ return ret;
-+ }
-+
-+ edesc->src_nents = src_nents;
-+ memset(&req_ctx->fd_flt, 0, sizeof(req_ctx->fd_flt));
-+
-+ if (mapped_nents > 1) {
-+ int qm_sg_bytes;
-+ struct dpaa2_sg_entry *sg_table = &edesc->sgt[0];
-+
-+ qm_sg_bytes = mapped_nents * sizeof(*sg_table);
-+ sg_to_qm_sg_last(req->src, mapped_nents, sg_table, 0);
-+ edesc->qm_sg_dma = dma_map_single(ctx->dev, sg_table,
-+ qm_sg_bytes, DMA_TO_DEVICE);
-+ if (dma_mapping_error(ctx->dev, edesc->qm_sg_dma)) {
-+ dev_err(ctx->dev, "unable to map S/G table\n");
-+ goto unmap;
-+ }
-+ edesc->qm_sg_bytes = qm_sg_bytes;
-+ dpaa2_fl_set_format(in_fle, dpaa2_fl_sg);
-+ dpaa2_fl_set_addr(in_fle, edesc->qm_sg_dma);
-+ } else {
-+ dpaa2_fl_set_format(in_fle, dpaa2_fl_single);
-+ dpaa2_fl_set_addr(in_fle, sg_dma_address(req->src));
-+ }
-+
-+ edesc->dst_dma = dma_map_single(ctx->dev, req->result, digestsize,
-+ DMA_FROM_DEVICE);
-+ if (dma_mapping_error(ctx->dev, edesc->dst_dma)) {
-+ dev_err(ctx->dev, "unable to map dst\n");
-+ edesc->dst_dma = 0;
-+ goto unmap;
-+ }
-+
-+ dpaa2_fl_set_final(in_fle, true);
-+ dpaa2_fl_set_len(in_fle, req->nbytes);
-+ dpaa2_fl_set_format(out_fle, dpaa2_fl_single);
-+ dpaa2_fl_set_addr(out_fle, edesc->dst_dma);
-+ dpaa2_fl_set_len(out_fle, digestsize);
-+
-+ req_ctx->flc = &ctx->flc[DIGEST];
-+ req_ctx->flc_dma = ctx->flc_dma[DIGEST];
-+ req_ctx->cbk = ahash_done;
-+ req_ctx->ctx = &req->base;
-+ req_ctx->edesc = edesc;
-+ ret = dpaa2_caam_enqueue(ctx->dev, req_ctx);
-+ if (ret == -EINPROGRESS ||
-+ (ret == -EBUSY && req->base.flags & CRYPTO_TFM_REQ_MAY_BACKLOG))
-+ return ret;
-+
-+unmap:
-+ ahash_unmap(ctx->dev, edesc, req, digestsize);
-+ qi_cache_free(edesc);
-+ return ret;
-+}
-+
-+static int ahash_final_no_ctx(struct ahash_request *req)
-+{
-+ struct crypto_ahash *ahash = crypto_ahash_reqtfm(req);
-+ struct caam_hash_ctx *ctx = crypto_ahash_ctx(ahash);
-+ struct caam_hash_state *state = ahash_request_ctx(req);
-+ struct caam_request *req_ctx = &state->caam_req;
-+ struct dpaa2_fl_entry *in_fle = &req_ctx->fd_flt[1];
-+ struct dpaa2_fl_entry *out_fle = &req_ctx->fd_flt[0];
-+ gfp_t flags = (req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP) ?
-+ GFP_KERNEL : GFP_ATOMIC;
-+ u8 *buf = current_buf(state);
-+ int buflen = *current_buflen(state);
-+ int digestsize = crypto_ahash_digestsize(ahash);
-+ struct ahash_edesc *edesc;
-+ int ret = -ENOMEM;
-+
-+ /* allocate space for base edesc and link tables */
-+ edesc = qi_cache_zalloc(GFP_DMA | flags);
-+ if (!edesc)
-+ return ret;
-+
-+ state->buf_dma = dma_map_single(ctx->dev, buf, buflen, DMA_TO_DEVICE);
-+ if (dma_mapping_error(ctx->dev, state->buf_dma)) {
-+ dev_err(ctx->dev, "unable to map src\n");
-+ goto unmap;
-+ }
-+
-+ edesc->dst_dma = dma_map_single(ctx->dev, req->result, digestsize,
-+ DMA_FROM_DEVICE);
-+ if (dma_mapping_error(ctx->dev, edesc->dst_dma)) {
-+ dev_err(ctx->dev, "unable to map dst\n");
-+ edesc->dst_dma = 0;
-+ goto unmap;
-+ }
-+
-+ memset(&req_ctx->fd_flt, 0, sizeof(req_ctx->fd_flt));
-+ dpaa2_fl_set_final(in_fle, true);
-+ dpaa2_fl_set_format(in_fle, dpaa2_fl_single);
-+ dpaa2_fl_set_addr(in_fle, state->buf_dma);
-+ dpaa2_fl_set_len(in_fle, buflen);
-+ dpaa2_fl_set_format(out_fle, dpaa2_fl_single);
-+ dpaa2_fl_set_addr(out_fle, edesc->dst_dma);
-+ dpaa2_fl_set_len(out_fle, digestsize);
-+
-+ req_ctx->flc = &ctx->flc[DIGEST];
-+ req_ctx->flc_dma = ctx->flc_dma[DIGEST];
-+ req_ctx->cbk = ahash_done;
-+ req_ctx->ctx = &req->base;
-+ req_ctx->edesc = edesc;
-+
-+ ret = dpaa2_caam_enqueue(ctx->dev, req_ctx);
-+ if (ret == -EINPROGRESS ||
-+ (ret == -EBUSY && req->base.flags & CRYPTO_TFM_REQ_MAY_BACKLOG))
-+ return ret;
-+
-+unmap:
-+ ahash_unmap(ctx->dev, edesc, req, digestsize);
-+ qi_cache_free(edesc);
-+ return ret;
-+}
-+
-+static int ahash_update_no_ctx(struct ahash_request *req)
-+{
-+ struct crypto_ahash *ahash = crypto_ahash_reqtfm(req);
-+ struct caam_hash_ctx *ctx = crypto_ahash_ctx(ahash);
-+ struct caam_hash_state *state = ahash_request_ctx(req);
-+ struct caam_request *req_ctx = &state->caam_req;
-+ struct dpaa2_fl_entry *in_fle = &req_ctx->fd_flt[1];
-+ struct dpaa2_fl_entry *out_fle = &req_ctx->fd_flt[0];
-+ gfp_t flags = (req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP) ?
-+ GFP_KERNEL : GFP_ATOMIC;
-+ u8 *buf = current_buf(state);
-+ int *buflen = current_buflen(state);
-+ u8 *next_buf = alt_buf(state);
-+ int *next_buflen = alt_buflen(state);
-+ int in_len = *buflen + req->nbytes, to_hash;
-+ int qm_sg_bytes, src_nents, mapped_nents;
-+ struct ahash_edesc *edesc;
-+ int ret = 0;
-+
-+ *next_buflen = in_len & (crypto_tfm_alg_blocksize(&ahash->base) - 1);
-+ to_hash = in_len - *next_buflen;
-+
-+ if (to_hash) {
-+ struct dpaa2_sg_entry *sg_table;
-+
-+ src_nents = sg_nents_for_len(req->src,
-+ req->nbytes - *next_buflen);
-+ if (src_nents < 0) {
-+ dev_err(ctx->dev, "Invalid number of src SG.\n");
-+ return src_nents;
-+ }
-+
-+ if (src_nents) {
-+ mapped_nents = dma_map_sg(ctx->dev, req->src, src_nents,
-+ DMA_TO_DEVICE);
-+ if (!mapped_nents) {
-+ dev_err(ctx->dev, "unable to DMA map source\n");
-+ return -ENOMEM;
-+ }
-+ } else {
-+ mapped_nents = 0;
-+ }
-+
-+ /* allocate space for base edesc and link tables */
-+ edesc = qi_cache_zalloc(GFP_DMA | flags);
-+ if (!edesc) {
-+ dma_unmap_sg(ctx->dev, req->src, src_nents,
-+ DMA_TO_DEVICE);
-+ return -ENOMEM;
-+ }
-+
-+ edesc->src_nents = src_nents;
-+ qm_sg_bytes = (1 + mapped_nents) * sizeof(*sg_table);
-+ sg_table = &edesc->sgt[0];
-+
-+ ret = buf_map_to_qm_sg(ctx->dev, sg_table, state);
-+ if (ret)
-+ goto unmap_ctx;
-+
-+ sg_to_qm_sg_last(req->src, mapped_nents, sg_table + 1, 0);
-+
-+ if (*next_buflen)
-+ scatterwalk_map_and_copy(next_buf, req->src,
-+ to_hash - *buflen,
-+ *next_buflen, 0);
-+
-+ edesc->qm_sg_dma = dma_map_single(ctx->dev, sg_table,
-+ qm_sg_bytes, DMA_TO_DEVICE);
-+ if (dma_mapping_error(ctx->dev, edesc->qm_sg_dma)) {
-+ dev_err(ctx->dev, "unable to map S/G table\n");
-+ ret = -ENOMEM;
-+ goto unmap_ctx;
-+ }
-+ edesc->qm_sg_bytes = qm_sg_bytes;
-+
-+ state->ctx_dma = dma_map_single(ctx->dev, state->caam_ctx,
-+ ctx->ctx_len, DMA_FROM_DEVICE);
-+ if (dma_mapping_error(ctx->dev, state->ctx_dma)) {
-+ dev_err(ctx->dev, "unable to map ctx\n");
-+ state->ctx_dma = 0;
-+ ret = -ENOMEM;
-+ goto unmap_ctx;
-+ }
-+
-+ memset(&req_ctx->fd_flt, 0, sizeof(req_ctx->fd_flt));
-+ dpaa2_fl_set_final(in_fle, true);
-+ dpaa2_fl_set_format(in_fle, dpaa2_fl_sg);
-+ dpaa2_fl_set_addr(in_fle, edesc->qm_sg_dma);
-+ dpaa2_fl_set_len(in_fle, to_hash);
-+ dpaa2_fl_set_format(out_fle, dpaa2_fl_single);
-+ dpaa2_fl_set_addr(out_fle, state->ctx_dma);
-+ dpaa2_fl_set_len(out_fle, ctx->ctx_len);
-+
-+ req_ctx->flc = &ctx->flc[UPDATE_FIRST];
-+ req_ctx->flc_dma = ctx->flc_dma[UPDATE_FIRST];
-+ req_ctx->cbk = ahash_done_ctx_dst;
-+ req_ctx->ctx = &req->base;
-+ req_ctx->edesc = edesc;
-+
-+ ret = dpaa2_caam_enqueue(ctx->dev, req_ctx);
-+ if (ret != -EINPROGRESS &&
-+ !(ret == -EBUSY &&
-+ req->base.flags & CRYPTO_TFM_REQ_MAY_BACKLOG))
-+ goto unmap_ctx;
-+
-+ state->update = ahash_update_ctx;
-+ state->finup = ahash_finup_ctx;
-+ state->final = ahash_final_ctx;
-+ } else if (*next_buflen) {
-+ scatterwalk_map_and_copy(buf + *buflen, req->src, 0,
-+ req->nbytes, 0);
-+ *buflen = *next_buflen;
-+ *next_buflen = 0;
-+ }
-+#ifdef DEBUG
-+ print_hex_dump(KERN_ERR, "buf@" __stringify(__LINE__)": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, buf, *buflen, 1);
-+ print_hex_dump(KERN_ERR, "next buf@" __stringify(__LINE__)": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, next_buf,
-+ *next_buflen, 1);
-+#endif
-+
-+ return ret;
-+unmap_ctx:
-+ ahash_unmap_ctx(ctx->dev, edesc, req, ctx->ctx_len, DMA_TO_DEVICE);
-+ qi_cache_free(edesc);
-+ return ret;
-+}
-+
-+static int ahash_finup_no_ctx(struct ahash_request *req)
-+{
-+ struct crypto_ahash *ahash = crypto_ahash_reqtfm(req);
-+ struct caam_hash_ctx *ctx = crypto_ahash_ctx(ahash);
-+ struct caam_hash_state *state = ahash_request_ctx(req);
-+ struct caam_request *req_ctx = &state->caam_req;
-+ struct dpaa2_fl_entry *in_fle = &req_ctx->fd_flt[1];
-+ struct dpaa2_fl_entry *out_fle = &req_ctx->fd_flt[0];
-+ gfp_t flags = (req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP) ?
-+ GFP_KERNEL : GFP_ATOMIC;
-+ int buflen = *current_buflen(state);
-+ int qm_sg_bytes, src_nents, mapped_nents;
-+ int digestsize = crypto_ahash_digestsize(ahash);
-+ struct ahash_edesc *edesc;
-+ struct dpaa2_sg_entry *sg_table;
-+ int ret;
-+
-+ src_nents = sg_nents_for_len(req->src, req->nbytes);
-+ if (src_nents < 0) {
-+ dev_err(ctx->dev, "Invalid number of src SG.\n");
-+ return src_nents;
-+ }
-+
-+ if (src_nents) {
-+ mapped_nents = dma_map_sg(ctx->dev, req->src, src_nents,
-+ DMA_TO_DEVICE);
-+ if (!mapped_nents) {
-+ dev_err(ctx->dev, "unable to DMA map source\n");
-+ return -ENOMEM;
-+ }
-+ } else {
-+ mapped_nents = 0;
-+ }
-+
-+ /* allocate space for base edesc and link tables */
-+ edesc = qi_cache_zalloc(GFP_DMA | flags);
-+ if (!edesc) {
-+ dma_unmap_sg(ctx->dev, req->src, src_nents, DMA_TO_DEVICE);
-+ return -ENOMEM;
-+ }
-+
-+ edesc->src_nents = src_nents;
-+ qm_sg_bytes = (2 + mapped_nents) * sizeof(*sg_table);
-+ sg_table = &edesc->sgt[0];
-+
-+ ret = buf_map_to_qm_sg(ctx->dev, sg_table, state);
-+ if (ret)
-+ goto unmap;
-+
-+ sg_to_qm_sg_last(req->src, mapped_nents, sg_table + 1, 0);
-+
-+ edesc->qm_sg_dma = dma_map_single(ctx->dev, sg_table, qm_sg_bytes,
-+ DMA_TO_DEVICE);
-+ if (dma_mapping_error(ctx->dev, edesc->qm_sg_dma)) {
-+ dev_err(ctx->dev, "unable to map S/G table\n");
-+ ret = -ENOMEM;
-+ goto unmap;
-+ }
-+ edesc->qm_sg_bytes = qm_sg_bytes;
-+
-+ edesc->dst_dma = dma_map_single(ctx->dev, req->result, digestsize,
-+ DMA_FROM_DEVICE);
-+ if (dma_mapping_error(ctx->dev, edesc->dst_dma)) {
-+ dev_err(ctx->dev, "unable to map dst\n");
-+ edesc->dst_dma = 0;
-+ ret = -ENOMEM;
-+ goto unmap;
-+ }
-+
-+ memset(&req_ctx->fd_flt, 0, sizeof(req_ctx->fd_flt));
-+ dpaa2_fl_set_final(in_fle, true);
-+ dpaa2_fl_set_format(in_fle, dpaa2_fl_sg);
-+ dpaa2_fl_set_addr(in_fle, edesc->qm_sg_dma);
-+ dpaa2_fl_set_len(in_fle, buflen + req->nbytes);
-+ dpaa2_fl_set_format(out_fle, dpaa2_fl_single);
-+ dpaa2_fl_set_addr(out_fle, edesc->dst_dma);
-+ dpaa2_fl_set_len(out_fle, digestsize);
-+
-+ req_ctx->flc = &ctx->flc[DIGEST];
-+ req_ctx->flc_dma = ctx->flc_dma[DIGEST];
-+ req_ctx->cbk = ahash_done;
-+ req_ctx->ctx = &req->base;
-+ req_ctx->edesc = edesc;
-+ ret = dpaa2_caam_enqueue(ctx->dev, req_ctx);
-+ if (ret != -EINPROGRESS &&
-+ !(ret == -EBUSY && req->base.flags & CRYPTO_TFM_REQ_MAY_BACKLOG))
-+ goto unmap;
-+
-+ return ret;
-+unmap:
-+ ahash_unmap(ctx->dev, edesc, req, digestsize);
-+ qi_cache_free(edesc);
-+ return -ENOMEM;
-+}
-+
-+static int ahash_update_first(struct ahash_request *req)
-+{
-+ struct crypto_ahash *ahash = crypto_ahash_reqtfm(req);
-+ struct caam_hash_ctx *ctx = crypto_ahash_ctx(ahash);
-+ struct caam_hash_state *state = ahash_request_ctx(req);
-+ struct caam_request *req_ctx = &state->caam_req;
-+ struct dpaa2_fl_entry *in_fle = &req_ctx->fd_flt[1];
-+ struct dpaa2_fl_entry *out_fle = &req_ctx->fd_flt[0];
-+ gfp_t flags = (req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP) ?
-+ GFP_KERNEL : GFP_ATOMIC;
-+ u8 *next_buf = alt_buf(state);
-+ int *next_buflen = alt_buflen(state);
-+ int to_hash;
-+ int src_nents, mapped_nents;
-+ struct ahash_edesc *edesc;
-+ int ret = 0;
-+
-+ *next_buflen = req->nbytes & (crypto_tfm_alg_blocksize(&ahash->base) -
-+ 1);
-+ to_hash = req->nbytes - *next_buflen;
-+
-+ if (to_hash) {
-+ struct dpaa2_sg_entry *sg_table;
-+
-+ src_nents = sg_nents_for_len(req->src,
-+ req->nbytes - (*next_buflen));
-+ if (src_nents < 0) {
-+ dev_err(ctx->dev, "Invalid number of src SG.\n");
-+ return src_nents;
-+ }
-+
-+ if (src_nents) {
-+ mapped_nents = dma_map_sg(ctx->dev, req->src, src_nents,
-+ DMA_TO_DEVICE);
-+ if (!mapped_nents) {
-+ dev_err(ctx->dev, "unable to map source for DMA\n");
-+ return -ENOMEM;
-+ }
-+ } else {
-+ mapped_nents = 0;
-+ }
-+
-+ /* allocate space for base edesc and link tables */
-+ edesc = qi_cache_zalloc(GFP_DMA | flags);
-+ if (!edesc) {
-+ dma_unmap_sg(ctx->dev, req->src, src_nents,
-+ DMA_TO_DEVICE);
-+ return -ENOMEM;
-+ }
-+
-+ edesc->src_nents = src_nents;
-+ sg_table = &edesc->sgt[0];
-+
-+ memset(&req_ctx->fd_flt, 0, sizeof(req_ctx->fd_flt));
-+ dpaa2_fl_set_final(in_fle, true);
-+ dpaa2_fl_set_len(in_fle, to_hash);
-+
-+ if (mapped_nents > 1) {
-+ int qm_sg_bytes;
-+
-+ sg_to_qm_sg_last(req->src, mapped_nents, sg_table, 0);
-+ qm_sg_bytes = mapped_nents * sizeof(*sg_table);
-+ edesc->qm_sg_dma = dma_map_single(ctx->dev, sg_table,
-+ qm_sg_bytes,
-+ DMA_TO_DEVICE);
-+ if (dma_mapping_error(ctx->dev, edesc->qm_sg_dma)) {
-+ dev_err(ctx->dev, "unable to map S/G table\n");
-+ ret = -ENOMEM;
-+ goto unmap_ctx;
-+ }
-+ edesc->qm_sg_bytes = qm_sg_bytes;
-+ dpaa2_fl_set_format(in_fle, dpaa2_fl_sg);
-+ dpaa2_fl_set_addr(in_fle, edesc->qm_sg_dma);
-+ } else {
-+ dpaa2_fl_set_format(in_fle, dpaa2_fl_single);
-+ dpaa2_fl_set_addr(in_fle, sg_dma_address(req->src));
-+ }
-+
-+ if (*next_buflen)
-+ scatterwalk_map_and_copy(next_buf, req->src, to_hash,
-+ *next_buflen, 0);
-+
-+ state->ctx_dma = dma_map_single(ctx->dev, state->caam_ctx,
-+ ctx->ctx_len, DMA_FROM_DEVICE);
-+ if (dma_mapping_error(ctx->dev, state->ctx_dma)) {
-+ dev_err(ctx->dev, "unable to map ctx\n");
-+ state->ctx_dma = 0;
-+ ret = -ENOMEM;
-+ goto unmap_ctx;
-+ }
-+
-+ dpaa2_fl_set_format(out_fle, dpaa2_fl_single);
-+ dpaa2_fl_set_addr(out_fle, state->ctx_dma);
-+ dpaa2_fl_set_len(out_fle, ctx->ctx_len);
-+
-+ req_ctx->flc = &ctx->flc[UPDATE_FIRST];
-+ req_ctx->flc_dma = ctx->flc_dma[UPDATE_FIRST];
-+ req_ctx->cbk = ahash_done_ctx_dst;
-+ req_ctx->ctx = &req->base;
-+ req_ctx->edesc = edesc;
-+
-+ ret = dpaa2_caam_enqueue(ctx->dev, req_ctx);
-+ if (ret != -EINPROGRESS &&
-+ !(ret == -EBUSY && req->base.flags &
-+ CRYPTO_TFM_REQ_MAY_BACKLOG))
-+ goto unmap_ctx;
-+
-+ state->update = ahash_update_ctx;
-+ state->finup = ahash_finup_ctx;
-+ state->final = ahash_final_ctx;
-+ } else if (*next_buflen) {
-+ state->update = ahash_update_no_ctx;
-+ state->finup = ahash_finup_no_ctx;
-+ state->final = ahash_final_no_ctx;
-+ scatterwalk_map_and_copy(next_buf, req->src, 0,
-+ req->nbytes, 0);
-+ switch_buf(state);
-+ }
-+#ifdef DEBUG
-+ print_hex_dump(KERN_ERR, "next buf@" __stringify(__LINE__)": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, next_buf, *next_buflen, 1);
-+#endif
-+
-+ return ret;
-+unmap_ctx:
-+ ahash_unmap_ctx(ctx->dev, edesc, req, ctx->ctx_len, DMA_TO_DEVICE);
-+ qi_cache_free(edesc);
-+ return ret;
-+}
-+
-+static int ahash_finup_first(struct ahash_request *req)
-+{
-+ return ahash_digest(req);
-+}
-+
-+static int ahash_init(struct ahash_request *req)
-+{
-+ struct caam_hash_state *state = ahash_request_ctx(req);
-+
-+ state->update = ahash_update_first;
-+ state->finup = ahash_finup_first;
-+ state->final = ahash_final_no_ctx;
-+
-+ state->ctx_dma = 0;
-+ state->current_buf = 0;
-+ state->buf_dma = 0;
-+ state->buflen_0 = 0;
-+ state->buflen_1 = 0;
-+
-+ return 0;
-+}
-+
-+static int ahash_update(struct ahash_request *req)
-+{
-+ struct caam_hash_state *state = ahash_request_ctx(req);
-+
-+ return state->update(req);
-+}
-+
-+static int ahash_finup(struct ahash_request *req)
-+{
-+ struct caam_hash_state *state = ahash_request_ctx(req);
-+
-+ return state->finup(req);
-+}
-+
-+static int ahash_final(struct ahash_request *req)
-+{
-+ struct caam_hash_state *state = ahash_request_ctx(req);
-+
-+ return state->final(req);
-+}
-+
-+static int ahash_export(struct ahash_request *req, void *out)
-+{
-+ struct caam_hash_state *state = ahash_request_ctx(req);
-+ struct caam_export_state *export = out;
-+ int len;
-+ u8 *buf;
-+
-+ if (state->current_buf) {
-+ buf = state->buf_1;
-+ len = state->buflen_1;
-+ } else {
-+ buf = state->buf_0;
-+ len = state->buflen_0;
-+ }
-+
-+ memcpy(export->buf, buf, len);
-+ memcpy(export->caam_ctx, state->caam_ctx, sizeof(export->caam_ctx));
-+ export->buflen = len;
-+ export->update = state->update;
-+ export->final = state->final;
-+ export->finup = state->finup;
-+
-+ return 0;
-+}
-+
-+static int ahash_import(struct ahash_request *req, const void *in)
-+{
-+ struct caam_hash_state *state = ahash_request_ctx(req);
-+ const struct caam_export_state *export = in;
-+
-+ memset(state, 0, sizeof(*state));
-+ memcpy(state->buf_0, export->buf, export->buflen);
-+ memcpy(state->caam_ctx, export->caam_ctx, sizeof(state->caam_ctx));
-+ state->buflen_0 = export->buflen;
-+ state->update = export->update;
-+ state->final = export->final;
-+ state->finup = export->finup;
-+
-+ return 0;
-+}
-+
-+struct caam_hash_template {
-+ char name[CRYPTO_MAX_ALG_NAME];
-+ char driver_name[CRYPTO_MAX_ALG_NAME];
-+ char hmac_name[CRYPTO_MAX_ALG_NAME];
-+ char hmac_driver_name[CRYPTO_MAX_ALG_NAME];
-+ unsigned int blocksize;
-+ struct ahash_alg template_ahash;
-+ u32 alg_type;
-+};
-+
-+/* ahash descriptors */
-+static struct caam_hash_template driver_hash[] = {
-+ {
-+ .name = "sha1",
-+ .driver_name = "sha1-caam-qi2",
-+ .hmac_name = "hmac(sha1)",
-+ .hmac_driver_name = "hmac-sha1-caam-qi2",
-+ .blocksize = SHA1_BLOCK_SIZE,
-+ .template_ahash = {
-+ .init = ahash_init,
-+ .update = ahash_update,
-+ .final = ahash_final,
-+ .finup = ahash_finup,
-+ .digest = ahash_digest,
-+ .export = ahash_export,
-+ .import = ahash_import,
-+ .setkey = ahash_setkey,
-+ .halg = {
-+ .digestsize = SHA1_DIGEST_SIZE,
-+ .statesize = sizeof(struct caam_export_state),
-+ },
-+ },
-+ .alg_type = OP_ALG_ALGSEL_SHA1,
-+ }, {
-+ .name = "sha224",
-+ .driver_name = "sha224-caam-qi2",
-+ .hmac_name = "hmac(sha224)",
-+ .hmac_driver_name = "hmac-sha224-caam-qi2",
-+ .blocksize = SHA224_BLOCK_SIZE,
-+ .template_ahash = {
-+ .init = ahash_init,
-+ .update = ahash_update,
-+ .final = ahash_final,
-+ .finup = ahash_finup,
-+ .digest = ahash_digest,
-+ .export = ahash_export,
-+ .import = ahash_import,
-+ .setkey = ahash_setkey,
-+ .halg = {
-+ .digestsize = SHA224_DIGEST_SIZE,
-+ .statesize = sizeof(struct caam_export_state),
-+ },
-+ },
-+ .alg_type = OP_ALG_ALGSEL_SHA224,
-+ }, {
-+ .name = "sha256",
-+ .driver_name = "sha256-caam-qi2",
-+ .hmac_name = "hmac(sha256)",
-+ .hmac_driver_name = "hmac-sha256-caam-qi2",
-+ .blocksize = SHA256_BLOCK_SIZE,
-+ .template_ahash = {
-+ .init = ahash_init,
-+ .update = ahash_update,
-+ .final = ahash_final,
-+ .finup = ahash_finup,
-+ .digest = ahash_digest,
-+ .export = ahash_export,
-+ .import = ahash_import,
-+ .setkey = ahash_setkey,
-+ .halg = {
-+ .digestsize = SHA256_DIGEST_SIZE,
-+ .statesize = sizeof(struct caam_export_state),
-+ },
-+ },
-+ .alg_type = OP_ALG_ALGSEL_SHA256,
-+ }, {
-+ .name = "sha384",
-+ .driver_name = "sha384-caam-qi2",
-+ .hmac_name = "hmac(sha384)",
-+ .hmac_driver_name = "hmac-sha384-caam-qi2",
-+ .blocksize = SHA384_BLOCK_SIZE,
-+ .template_ahash = {
-+ .init = ahash_init,
-+ .update = ahash_update,
-+ .final = ahash_final,
-+ .finup = ahash_finup,
-+ .digest = ahash_digest,
-+ .export = ahash_export,
-+ .import = ahash_import,
-+ .setkey = ahash_setkey,
-+ .halg = {
-+ .digestsize = SHA384_DIGEST_SIZE,
-+ .statesize = sizeof(struct caam_export_state),
-+ },
-+ },
-+ .alg_type = OP_ALG_ALGSEL_SHA384,
-+ }, {
-+ .name = "sha512",
-+ .driver_name = "sha512-caam-qi2",
-+ .hmac_name = "hmac(sha512)",
-+ .hmac_driver_name = "hmac-sha512-caam-qi2",
-+ .blocksize = SHA512_BLOCK_SIZE,
-+ .template_ahash = {
-+ .init = ahash_init,
-+ .update = ahash_update,
-+ .final = ahash_final,
-+ .finup = ahash_finup,
-+ .digest = ahash_digest,
-+ .export = ahash_export,
-+ .import = ahash_import,
-+ .setkey = ahash_setkey,
-+ .halg = {
-+ .digestsize = SHA512_DIGEST_SIZE,
-+ .statesize = sizeof(struct caam_export_state),
-+ },
-+ },
-+ .alg_type = OP_ALG_ALGSEL_SHA512,
-+ }, {
-+ .name = "md5",
-+ .driver_name = "md5-caam-qi2",
-+ .hmac_name = "hmac(md5)",
-+ .hmac_driver_name = "hmac-md5-caam-qi2",
-+ .blocksize = MD5_BLOCK_WORDS * 4,
-+ .template_ahash = {
-+ .init = ahash_init,
-+ .update = ahash_update,
-+ .final = ahash_final,
-+ .finup = ahash_finup,
-+ .digest = ahash_digest,
-+ .export = ahash_export,
-+ .import = ahash_import,
-+ .setkey = ahash_setkey,
-+ .halg = {
-+ .digestsize = MD5_DIGEST_SIZE,
-+ .statesize = sizeof(struct caam_export_state),
-+ },
-+ },
-+ .alg_type = OP_ALG_ALGSEL_MD5,
-+ }
-+};
-+
-+struct caam_hash_alg {
-+ struct list_head entry;
-+ struct device *dev;
-+ int alg_type;
-+ struct ahash_alg ahash_alg;
-+};
-+
-+static int caam_hash_cra_init(struct crypto_tfm *tfm)
-+{
-+ struct crypto_ahash *ahash = __crypto_ahash_cast(tfm);
-+ struct crypto_alg *base = tfm->__crt_alg;
-+ struct hash_alg_common *halg =
-+ container_of(base, struct hash_alg_common, base);
-+ struct ahash_alg *alg =
-+ container_of(halg, struct ahash_alg, halg);
-+ struct caam_hash_alg *caam_hash =
-+ container_of(alg, struct caam_hash_alg, ahash_alg);
-+ struct caam_hash_ctx *ctx = crypto_tfm_ctx(tfm);
-+ /* Sizes for MDHA running digests: MD5, SHA1, 224, 256, 384, 512 */
-+ static const u8 runninglen[] = { HASH_MSG_LEN + MD5_DIGEST_SIZE,
-+ HASH_MSG_LEN + SHA1_DIGEST_SIZE,
-+ HASH_MSG_LEN + 32,
-+ HASH_MSG_LEN + SHA256_DIGEST_SIZE,
-+ HASH_MSG_LEN + 64,
-+ HASH_MSG_LEN + SHA512_DIGEST_SIZE };
-+ dma_addr_t dma_addr;
-+ int i;
-+
-+ ctx->dev = caam_hash->dev;
-+
-+ dma_addr = dma_map_single_attrs(ctx->dev, ctx->flc, sizeof(ctx->flc),
-+ DMA_BIDIRECTIONAL,
-+ DMA_ATTR_SKIP_CPU_SYNC);
-+ if (dma_mapping_error(ctx->dev, dma_addr)) {
-+ dev_err(ctx->dev, "unable to map shared descriptors\n");
-+ return -ENOMEM;
-+ }
-+
-+ for (i = 0; i < HASH_NUM_OP; i++)
-+ ctx->flc_dma[i] = dma_addr + i * sizeof(ctx->flc[i]);
-+
-+ /* copy descriptor header template value */
-+ ctx->adata.algtype = OP_TYPE_CLASS2_ALG | caam_hash->alg_type;
-+
-+ ctx->ctx_len = runninglen[(ctx->adata.algtype &
-+ OP_ALG_ALGSEL_SUBMASK) >>
-+ OP_ALG_ALGSEL_SHIFT];
-+
-+ crypto_ahash_set_reqsize(__crypto_ahash_cast(tfm),
-+ sizeof(struct caam_hash_state));
-+
-+ return ahash_set_sh_desc(ahash);
-+}
-+
-+static void caam_hash_cra_exit(struct crypto_tfm *tfm)
-+{
-+ struct caam_hash_ctx *ctx = crypto_tfm_ctx(tfm);
-+
-+ dma_unmap_single_attrs(ctx->dev, ctx->flc_dma[0], sizeof(ctx->flc),
-+ DMA_BIDIRECTIONAL, DMA_ATTR_SKIP_CPU_SYNC);
-+}
-+
-+static struct caam_hash_alg *caam_hash_alloc(struct device *dev,
-+ struct caam_hash_template *template, bool keyed)
-+{
-+ struct caam_hash_alg *t_alg;
-+ struct ahash_alg *halg;
-+ struct crypto_alg *alg;
-+
-+ t_alg = kzalloc(sizeof(*t_alg), GFP_KERNEL);
-+ if (!t_alg)
-+ return ERR_PTR(-ENOMEM);
-+
-+ t_alg->ahash_alg = template->template_ahash;
-+ halg = &t_alg->ahash_alg;
-+ alg = &halg->halg.base;
-+
-+ if (keyed) {
-+ snprintf(alg->cra_name, CRYPTO_MAX_ALG_NAME, "%s",
-+ template->hmac_name);
-+ snprintf(alg->cra_driver_name, CRYPTO_MAX_ALG_NAME, "%s",
-+ template->hmac_driver_name);
-+ } else {
-+ snprintf(alg->cra_name, CRYPTO_MAX_ALG_NAME, "%s",
-+ template->name);
-+ snprintf(alg->cra_driver_name, CRYPTO_MAX_ALG_NAME, "%s",
-+ template->driver_name);
-+ t_alg->ahash_alg.setkey = NULL;
-+ }
-+ alg->cra_module = THIS_MODULE;
-+ alg->cra_init = caam_hash_cra_init;
-+ alg->cra_exit = caam_hash_cra_exit;
-+ alg->cra_ctxsize = sizeof(struct caam_hash_ctx);
-+ alg->cra_priority = CAAM_CRA_PRIORITY;
-+ alg->cra_blocksize = template->blocksize;
-+ alg->cra_alignmask = 0;
-+ alg->cra_flags = CRYPTO_ALG_ASYNC | CRYPTO_ALG_TYPE_AHASH;
-+ alg->cra_type = &crypto_ahash_type;
-+
-+ t_alg->alg_type = template->alg_type;
-+ t_alg->dev = dev;
-+
-+ return t_alg;
-+}
-+
-+static void dpaa2_caam_fqdan_cb(struct dpaa2_io_notification_ctx *nctx)
-+{
-+ struct dpaa2_caam_priv_per_cpu *ppriv;
-+
-+ ppriv = container_of(nctx, struct dpaa2_caam_priv_per_cpu, nctx);
-+ napi_schedule_irqoff(&ppriv->napi);
-+}
-+
-+static int __cold dpaa2_dpseci_dpio_setup(struct dpaa2_caam_priv *priv)
-+{
-+ struct device *dev = priv->dev;
-+ struct dpaa2_io_notification_ctx *nctx;
-+ struct dpaa2_caam_priv_per_cpu *ppriv;
-+ int err, i = 0, cpu;
-+
-+ for_each_online_cpu(cpu) {
-+ ppriv = per_cpu_ptr(priv->ppriv, cpu);
-+ ppriv->priv = priv;
-+ nctx = &ppriv->nctx;
-+ nctx->is_cdan = 0;
-+ nctx->id = ppriv->rsp_fqid;
-+ nctx->desired_cpu = cpu;
-+ nctx->cb = dpaa2_caam_fqdan_cb;
-+
-+ /* Register notification callbacks */
-+ err = dpaa2_io_service_register(NULL, nctx);
-+ if (unlikely(err)) {
-+ dev_dbg(dev, "No affine DPIO for cpu %d\n", cpu);
-+ nctx->cb = NULL;
-+ /*
-+ * If no affine DPIO for this core, there's probably
-+ * none available for next cores either. Signal we want
-+ * to retry later, in case the DPIO devices weren't
-+ * probed yet.
-+ */
-+ err = -EPROBE_DEFER;
-+ goto err;
-+ }
-+
-+ ppriv->store = dpaa2_io_store_create(DPAA2_CAAM_STORE_SIZE,
-+ dev);
-+ if (unlikely(!ppriv->store)) {
-+ dev_err(dev, "dpaa2_io_store_create() failed\n");
-+ goto err;
-+ }
-+
-+ if (++i == priv->num_pairs)
-+ break;
-+ }
-+
-+ return 0;
-+
-+err:
-+ for_each_online_cpu(cpu) {
-+ ppriv = per_cpu_ptr(priv->ppriv, cpu);
-+ if (!ppriv->nctx.cb)
-+ break;
-+ dpaa2_io_service_deregister(NULL, &ppriv->nctx);
-+ }
-+
-+ for_each_online_cpu(cpu) {
-+ ppriv = per_cpu_ptr(priv->ppriv, cpu);
-+ if (!ppriv->store)
-+ break;
-+ dpaa2_io_store_destroy(ppriv->store);
-+ }
-+
-+ return err;
-+}
-+
-+static void __cold dpaa2_dpseci_dpio_free(struct dpaa2_caam_priv *priv)
-+{
-+ struct dpaa2_caam_priv_per_cpu *ppriv;
-+ int i = 0, cpu;
-+
-+ for_each_online_cpu(cpu) {
-+ ppriv = per_cpu_ptr(priv->ppriv, cpu);
-+ dpaa2_io_service_deregister(NULL, &ppriv->nctx);
-+ dpaa2_io_store_destroy(ppriv->store);
-+
-+ if (++i == priv->num_pairs)
-+ return;
-+ }
-+}
-+
-+static int dpaa2_dpseci_bind(struct dpaa2_caam_priv *priv)
-+{
-+ struct dpseci_rx_queue_cfg rx_queue_cfg;
-+ struct device *dev = priv->dev;
-+ struct fsl_mc_device *ls_dev = to_fsl_mc_device(dev);
-+ struct dpaa2_caam_priv_per_cpu *ppriv;
-+ int err = 0, i = 0, cpu;
-+
-+ /* Configure Rx queues */
-+ for_each_online_cpu(cpu) {
-+ ppriv = per_cpu_ptr(priv->ppriv, cpu);
-+
-+ rx_queue_cfg.options = DPSECI_QUEUE_OPT_DEST |
-+ DPSECI_QUEUE_OPT_USER_CTX;
-+ rx_queue_cfg.order_preservation_en = 0;
-+ rx_queue_cfg.dest_cfg.dest_type = DPSECI_DEST_DPIO;
-+ rx_queue_cfg.dest_cfg.dest_id = ppriv->nctx.dpio_id;
-+ /*
-+ * Rx priority (WQ) doesn't really matter, since we use
-+ * pull mode, i.e. volatile dequeues from specific FQs
-+ */
-+ rx_queue_cfg.dest_cfg.priority = 0;
-+ rx_queue_cfg.user_ctx = ppriv->nctx.qman64;
-+
-+ err = dpseci_set_rx_queue(priv->mc_io, 0, ls_dev->mc_handle, i,
-+ &rx_queue_cfg);
-+ if (err) {
-+ dev_err(dev, "dpseci_set_rx_queue() failed with err %d\n",
-+ err);
-+ return err;
-+ }
-+
-+ if (++i == priv->num_pairs)
-+ break;
-+ }
-+
-+ return err;
-+}
-+
-+static void dpaa2_dpseci_congestion_free(struct dpaa2_caam_priv *priv)
-+{
-+ struct device *dev = priv->dev;
-+
-+ if (!priv->cscn_mem)
-+ return;
-+
-+ dma_unmap_single(dev, priv->cscn_dma, DPAA2_CSCN_SIZE, DMA_FROM_DEVICE);
-+ kfree(priv->cscn_mem);
-+}
-+
-+static void dpaa2_dpseci_free(struct dpaa2_caam_priv *priv)
-+{
-+ struct device *dev = priv->dev;
-+ struct fsl_mc_device *ls_dev = to_fsl_mc_device(dev);
-+
-+ dpaa2_dpseci_congestion_free(priv);
-+ dpseci_close(priv->mc_io, 0, ls_dev->mc_handle);
-+}
-+
-+static void dpaa2_caam_process_fd(struct dpaa2_caam_priv *priv,
-+ const struct dpaa2_fd *fd)
-+{
-+ struct caam_request *req;
-+ u32 fd_err;
-+
-+ if (dpaa2_fd_get_format(fd) != dpaa2_fd_list) {
-+ dev_err(priv->dev, "Only Frame List FD format is supported!\n");
-+ return;
-+ }
-+
-+ fd_err = dpaa2_fd_get_ctrl(fd) & FD_CTRL_ERR_MASK;
-+ if (unlikely(fd_err))
-+ dev_err(priv->dev, "FD error: %08x\n", fd_err);
-+
-+ /*
-+ * FD[ADDR] is guaranteed to be valid, irrespective of errors reported
-+ * in FD[ERR] or FD[FRC].
-+ */
-+ req = dpaa2_caam_iova_to_virt(priv, dpaa2_fd_get_addr(fd));
-+ dma_unmap_single(priv->dev, req->fd_flt_dma, sizeof(req->fd_flt),
-+ DMA_BIDIRECTIONAL);
-+ req->cbk(req->ctx, dpaa2_fd_get_frc(fd));
-+}
-+
-+static int dpaa2_caam_pull_fq(struct dpaa2_caam_priv_per_cpu *ppriv)
-+{
-+ int err;
-+
-+ /* Retry while portal is busy */
-+ do {
-+ err = dpaa2_io_service_pull_fq(NULL, ppriv->rsp_fqid,
-+ ppriv->store);
-+ } while (err == -EBUSY);
-+
-+ if (unlikely(err))
-+ dev_err(ppriv->priv->dev, "dpaa2_io_service_pull err %d", err);
-+
-+ return err;
-+}
-+
-+static int dpaa2_caam_store_consume(struct dpaa2_caam_priv_per_cpu *ppriv)
-+{
-+ struct dpaa2_dq *dq;
-+ int cleaned = 0, is_last;
-+
-+ do {
-+ dq = dpaa2_io_store_next(ppriv->store, &is_last);
-+ if (unlikely(!dq)) {
-+ if (unlikely(!is_last)) {
-+ dev_dbg(ppriv->priv->dev,
-+ "FQ %d returned no valid frames\n",
-+ ppriv->rsp_fqid);
-+ /*
-+ * MUST retry until we get some sort of
-+ * valid response token (be it "empty dequeue"
-+ * or a valid frame).
-+ */
-+ continue;
-+ }
-+ break;
-+ }
-+
-+ /* Process FD */
-+ dpaa2_caam_process_fd(ppriv->priv, dpaa2_dq_fd(dq));
-+ cleaned++;
-+ } while (!is_last);
-+
-+ return cleaned;
-+}
-+
-+static int dpaa2_dpseci_poll(struct napi_struct *napi, int budget)
-+{
-+ struct dpaa2_caam_priv_per_cpu *ppriv;
-+ struct dpaa2_caam_priv *priv;
-+ int err, cleaned = 0, store_cleaned;
-+
-+ ppriv = container_of(napi, struct dpaa2_caam_priv_per_cpu, napi);
-+ priv = ppriv->priv;
-+
-+ if (unlikely(dpaa2_caam_pull_fq(ppriv)))
-+ return 0;
-+
-+ do {
-+ store_cleaned = dpaa2_caam_store_consume(ppriv);
-+ cleaned += store_cleaned;
-+
-+ if (store_cleaned == 0 ||
-+ cleaned > budget - DPAA2_CAAM_STORE_SIZE)
-+ break;
-+
-+ /* Try to dequeue some more */
-+ err = dpaa2_caam_pull_fq(ppriv);
-+ if (unlikely(err))
-+ break;
-+ } while (1);
-+
-+ if (cleaned < budget) {
-+ napi_complete_done(napi, cleaned);
-+ err = dpaa2_io_service_rearm(NULL, &ppriv->nctx);
-+ if (unlikely(err))
-+ dev_err(priv->dev, "Notification rearm failed: %d\n",
-+ err);
-+ }
-+
-+ return cleaned;
-+}
-+
-+static int dpaa2_dpseci_congestion_setup(struct dpaa2_caam_priv *priv,
-+ u16 token)
-+{
-+ struct dpseci_congestion_notification_cfg cong_notif_cfg = { 0 };
-+ struct device *dev = priv->dev;
-+ int err;
-+
-+ /*
-+ * Congestion group feature supported starting with DPSECI API v5.1
-+ * and only when object has been created with this capability.
-+ */
-+ if ((DPSECI_VER(priv->major_ver, priv->minor_ver) < DPSECI_VER(5, 1)) ||
-+ !(priv->dpseci_attr.options & DPSECI_OPT_HAS_CG))
-+ return 0;
-+
-+ priv->cscn_mem = kzalloc(DPAA2_CSCN_SIZE + DPAA2_CSCN_ALIGN,
-+ GFP_KERNEL | GFP_DMA);
-+ if (!priv->cscn_mem)
-+ return -ENOMEM;
-+
-+ priv->cscn_mem_aligned = PTR_ALIGN(priv->cscn_mem, DPAA2_CSCN_ALIGN);
-+ priv->cscn_dma = dma_map_single(dev, priv->cscn_mem_aligned,
-+ DPAA2_CSCN_SIZE, DMA_FROM_DEVICE);
-+ if (dma_mapping_error(dev, priv->cscn_dma)) {
-+ dev_err(dev, "Error mapping CSCN memory area\n");
-+ err = -ENOMEM;
-+ goto err_dma_map;
-+ }
-+
-+ cong_notif_cfg.units = DPSECI_CONGESTION_UNIT_BYTES;
-+ cong_notif_cfg.threshold_entry = DPAA2_SEC_CONG_ENTRY_THRESH;
-+ cong_notif_cfg.threshold_exit = DPAA2_SEC_CONG_EXIT_THRESH;
-+ cong_notif_cfg.message_ctx = (u64)priv;
-+ cong_notif_cfg.message_iova = priv->cscn_dma;
-+ cong_notif_cfg.notification_mode = DPSECI_CGN_MODE_WRITE_MEM_ON_ENTER |
-+ DPSECI_CGN_MODE_WRITE_MEM_ON_EXIT |
-+ DPSECI_CGN_MODE_COHERENT_WRITE;
-+
-+ err = dpseci_set_congestion_notification(priv->mc_io, 0, token,
-+ &cong_notif_cfg);
-+ if (err) {
-+ dev_err(dev, "dpseci_set_congestion_notification failed\n");
-+ goto err_set_cong;
-+ }
-+
-+ return 0;
-+
-+err_set_cong:
-+ dma_unmap_single(dev, priv->cscn_dma, DPAA2_CSCN_SIZE, DMA_FROM_DEVICE);
-+err_dma_map:
-+ kfree(priv->cscn_mem);
-+
-+ return err;
-+}
-+
-+static int __cold dpaa2_dpseci_setup(struct fsl_mc_device *ls_dev)
-+{
-+ struct device *dev = &ls_dev->dev;
-+ struct dpaa2_caam_priv *priv;
-+ struct dpaa2_caam_priv_per_cpu *ppriv;
-+ int err, cpu;
-+ u8 i;
-+
-+ priv = dev_get_drvdata(dev);
-+
-+ priv->dev = dev;
-+ priv->dpsec_id = ls_dev->obj_desc.id;
-+
-+ /* Get a handle for the DPSECI this interface is associate with */
-+ err = dpseci_open(priv->mc_io, 0, priv->dpsec_id, &ls_dev->mc_handle);
-+ if (err) {
-+ dev_err(dev, "dpsec_open() failed: %d\n", err);
-+ goto err_open;
-+ }
-+
-+ dev_info(dev, "Opened dpseci object successfully\n");
-+
-+ err = dpseci_get_api_version(priv->mc_io, 0, &priv->major_ver,
-+ &priv->minor_ver);
-+ if (err) {
-+ dev_err(dev, "dpseci_get_api_version() failed\n");
-+ goto err_get_vers;
-+ }
-+
-+ err = dpseci_get_attributes(priv->mc_io, 0, ls_dev->mc_handle,
-+ &priv->dpseci_attr);
-+ if (err) {
-+ dev_err(dev, "dpseci_get_attributes() failed\n");
-+ goto err_get_vers;
-+ }
-+
-+ err = dpseci_get_sec_attr(priv->mc_io, 0, ls_dev->mc_handle,
-+ &priv->sec_attr);
-+ if (err) {
-+ dev_err(dev, "dpseci_get_sec_attr() failed\n");
-+ goto err_get_vers;
-+ }
-+
-+ err = dpaa2_dpseci_congestion_setup(priv, ls_dev->mc_handle);
-+ if (err) {
-+ dev_err(dev, "setup_congestion() failed\n");
-+ goto err_get_vers;
-+ }
-+
-+ priv->num_pairs = min(priv->dpseci_attr.num_rx_queues,
-+ priv->dpseci_attr.num_tx_queues);
-+ if (priv->num_pairs > num_online_cpus()) {
-+ dev_warn(dev, "%d queues won't be used\n",
-+ priv->num_pairs - num_online_cpus());
-+ priv->num_pairs = num_online_cpus();
-+ }
-+
-+ for (i = 0; i < priv->dpseci_attr.num_rx_queues; i++) {
-+ err = dpseci_get_rx_queue(priv->mc_io, 0, ls_dev->mc_handle, i,
-+ &priv->rx_queue_attr[i]);
-+ if (err) {
-+ dev_err(dev, "dpseci_get_rx_queue() failed\n");
-+ goto err_get_rx_queue;
-+ }
-+ }
-+
-+ for (i = 0; i < priv->dpseci_attr.num_tx_queues; i++) {
-+ err = dpseci_get_tx_queue(priv->mc_io, 0, ls_dev->mc_handle, i,
-+ &priv->tx_queue_attr[i]);
-+ if (err) {
-+ dev_err(dev, "dpseci_get_tx_queue() failed\n");
-+ goto err_get_rx_queue;
-+ }
-+ }
-+
-+ i = 0;
-+ for_each_online_cpu(cpu) {
-+ dev_info(dev, "prio %d: rx queue %d, tx queue %d\n", i,
-+ priv->rx_queue_attr[i].fqid,
-+ priv->tx_queue_attr[i].fqid);
-+
-+ ppriv = per_cpu_ptr(priv->ppriv, cpu);
-+ ppriv->req_fqid = priv->tx_queue_attr[i].fqid;
-+ ppriv->rsp_fqid = priv->rx_queue_attr[i].fqid;
-+ ppriv->prio = i;
-+
-+ ppriv->net_dev.dev = *dev;
-+ INIT_LIST_HEAD(&ppriv->net_dev.napi_list);
-+ netif_napi_add(&ppriv->net_dev, &ppriv->napi, dpaa2_dpseci_poll,
-+ DPAA2_CAAM_NAPI_WEIGHT);
-+ if (++i == priv->num_pairs)
-+ break;
-+ }
-+
-+ return 0;
-+
-+err_get_rx_queue:
-+ dpaa2_dpseci_congestion_free(priv);
-+err_get_vers:
-+ dpseci_close(priv->mc_io, 0, ls_dev->mc_handle);
-+err_open:
-+ return err;
-+}
-+
-+static int dpaa2_dpseci_enable(struct dpaa2_caam_priv *priv)
-+{
-+ struct device *dev = priv->dev;
-+ struct fsl_mc_device *ls_dev = to_fsl_mc_device(dev);
-+ struct dpaa2_caam_priv_per_cpu *ppriv;
-+ int err, i;
-+
-+ for (i = 0; i < priv->num_pairs; i++) {
-+ ppriv = per_cpu_ptr(priv->ppriv, i);
-+ napi_enable(&ppriv->napi);
-+ }
-+
-+ err = dpseci_enable(priv->mc_io, 0, ls_dev->mc_handle);
-+ if (err) {
-+ dev_err(dev, "dpseci_enable() failed\n");
-+ return err;
-+ }
-+
-+ dev_info(dev, "DPSECI version %d.%d\n",
-+ priv->major_ver,
-+ priv->minor_ver);
-+
-+ return 0;
-+}
-+
-+static int __cold dpaa2_dpseci_disable(struct dpaa2_caam_priv *priv)
-+{
-+ struct device *dev = priv->dev;
-+ struct dpaa2_caam_priv_per_cpu *ppriv;
-+ struct fsl_mc_device *ls_dev = to_fsl_mc_device(dev);
-+ int i, err = 0, enabled;
-+
-+ err = dpseci_disable(priv->mc_io, 0, ls_dev->mc_handle);
-+ if (err) {
-+ dev_err(dev, "dpseci_disable() failed\n");
-+ return err;
-+ }
-+
-+ err = dpseci_is_enabled(priv->mc_io, 0, ls_dev->mc_handle, &enabled);
-+ if (err) {
-+ dev_err(dev, "dpseci_is_enabled() failed\n");
-+ return err;
-+ }
-+
-+ dev_dbg(dev, "disable: %s\n", enabled ? "false" : "true");
-+
-+ for (i = 0; i < priv->num_pairs; i++) {
-+ ppriv = per_cpu_ptr(priv->ppriv, i);
-+ napi_disable(&ppriv->napi);
-+ netif_napi_del(&ppriv->napi);
-+ }
-+
-+ return 0;
-+}
-+
-+static struct list_head alg_list;
-+static struct list_head hash_list;
-+
-+static int dpaa2_caam_probe(struct fsl_mc_device *dpseci_dev)
-+{
-+ struct device *dev;
-+ struct dpaa2_caam_priv *priv;
-+ int i, err = 0;
-+ bool registered = false;
-+
-+ /*
-+ * There is no way to get CAAM endianness - there is no direct register
-+ * space access and MC f/w does not provide this attribute.
-+ * All DPAA2-based SoCs have little endian CAAM, thus hard-code this
-+ * property.
-+ */
-+ caam_little_end = true;
-+
-+ caam_imx = false;
-+
-+ dev = &dpseci_dev->dev;
-+
-+ priv = kzalloc(sizeof(*priv), GFP_KERNEL);
-+ if (!priv)
-+ return -ENOMEM;
-+
-+ dev_set_drvdata(dev, priv);
-+
-+ priv->domain = iommu_get_domain_for_dev(dev);
-+
-+ qi_cache = kmem_cache_create("dpaa2_caamqicache", CAAM_QI_MEMCACHE_SIZE,
-+ 0, SLAB_CACHE_DMA, NULL);
-+ if (!qi_cache) {
-+ dev_err(dev, "Can't allocate SEC cache\n");
-+ err = -ENOMEM;
-+ goto err_qicache;
-+ }
-+
-+ err = dma_set_mask_and_coherent(dev, DMA_BIT_MASK(49));
-+ if (err) {
-+ dev_err(dev, "dma_set_mask_and_coherent() failed\n");
-+ goto err_dma_mask;
-+ }
-+
-+ /* Obtain a MC portal */
-+ err = fsl_mc_portal_allocate(dpseci_dev, 0, &priv->mc_io);
-+ if (err) {
-+ if (err == -ENXIO)
-+ err = -EPROBE_DEFER;
-+ else
-+ dev_err(dev, "MC portal allocation failed\n");
-+
-+ goto err_dma_mask;
-+ }
-+
-+ priv->ppriv = alloc_percpu(*priv->ppriv);
-+ if (!priv->ppriv) {
-+ dev_err(dev, "alloc_percpu() failed\n");
-+ goto err_alloc_ppriv;
-+ }
-+
-+ /* DPSECI initialization */
-+ err = dpaa2_dpseci_setup(dpseci_dev);
-+ if (err < 0) {
-+ dev_err(dev, "dpaa2_dpseci_setup() failed\n");
-+ goto err_dpseci_setup;
-+ }
-+
-+ /* DPIO */
-+ err = dpaa2_dpseci_dpio_setup(priv);
-+ if (err) {
-+ dev_err(dev, "dpaa2_dpseci_dpio_setup() failed\n");
-+ goto err_dpio_setup;
-+ }
-+
-+ /* DPSECI binding to DPIO */
-+ err = dpaa2_dpseci_bind(priv);
-+ if (err) {
-+ dev_err(dev, "dpaa2_dpseci_bind() failed\n");
-+ goto err_bind;
-+ }
-+
-+ /* DPSECI enable */
-+ err = dpaa2_dpseci_enable(priv);
-+ if (err) {
-+ dev_err(dev, "dpaa2_dpseci_enable() failed");
-+ goto err_bind;
-+ }
-+
-+ /* register crypto algorithms the device supports */
-+ INIT_LIST_HEAD(&alg_list);
-+ for (i = 0; i < ARRAY_SIZE(driver_algs); i++) {
-+ struct caam_crypto_alg *t_alg;
-+ struct caam_alg_template *alg = driver_algs + i;
-+ u32 alg_sel = alg->class1_alg_type & OP_ALG_ALGSEL_MASK;
-+
-+ /* Skip DES algorithms if not supported by device */
-+ if (!priv->sec_attr.des_acc_num &&
-+ ((alg_sel == OP_ALG_ALGSEL_3DES) ||
-+ (alg_sel == OP_ALG_ALGSEL_DES)))
-+ continue;
-+
-+ /* Skip AES algorithms if not supported by device */
-+ if (!priv->sec_attr.aes_acc_num &&
-+ (alg_sel == OP_ALG_ALGSEL_AES))
-+ continue;
-+
-+ t_alg = caam_alg_alloc(alg);
-+ if (IS_ERR(t_alg)) {
-+ err = PTR_ERR(t_alg);
-+ dev_warn(dev, "%s alg allocation failed: %d\n",
-+ alg->driver_name, err);
-+ continue;
-+ }
-+ t_alg->caam.dev = dev;
-+
-+ err = crypto_register_alg(&t_alg->crypto_alg);
-+ if (err) {
-+ dev_warn(dev, "%s alg registration failed: %d\n",
-+ t_alg->crypto_alg.cra_driver_name, err);
-+ kfree(t_alg);
-+ continue;
-+ }
-+
-+ list_add_tail(&t_alg->entry, &alg_list);
-+ registered = true;
-+ }
-+
-+ for (i = 0; i < ARRAY_SIZE(driver_aeads); i++) {
-+ struct caam_aead_alg *t_alg = driver_aeads + i;
-+ u32 c1_alg_sel = t_alg->caam.class1_alg_type &
-+ OP_ALG_ALGSEL_MASK;
-+ u32 c2_alg_sel = t_alg->caam.class2_alg_type &
-+ OP_ALG_ALGSEL_MASK;
-+
-+ /* Skip DES algorithms if not supported by device */
-+ if (!priv->sec_attr.des_acc_num &&
-+ ((c1_alg_sel == OP_ALG_ALGSEL_3DES) ||
-+ (c1_alg_sel == OP_ALG_ALGSEL_DES)))
-+ continue;
-+
-+ /* Skip AES algorithms if not supported by device */
-+ if (!priv->sec_attr.aes_acc_num &&
-+ (c1_alg_sel == OP_ALG_ALGSEL_AES))
-+ continue;
-+
-+ /*
-+ * Skip algorithms requiring message digests
-+ * if MD not supported by device.
-+ */
-+ if (!priv->sec_attr.md_acc_num && c2_alg_sel)
-+ continue;
-+
-+ t_alg->caam.dev = dev;
-+ caam_aead_alg_init(t_alg);
-+
-+ err = crypto_register_aead(&t_alg->aead);
-+ if (err) {
-+ dev_warn(dev, "%s alg registration failed: %d\n",
-+ t_alg->aead.base.cra_driver_name, err);
-+ continue;
-+ }
-+
-+ t_alg->registered = true;
-+ registered = true;
-+ }
-+ if (registered)
-+ dev_info(dev, "algorithms registered in /proc/crypto\n");
-+
-+ /* register hash algorithms the device supports */
-+ INIT_LIST_HEAD(&hash_list);
-+
-+ /*
-+ * Skip registration of any hashing algorithms if MD block
-+ * is not present.
-+ */
-+ if (!priv->sec_attr.md_acc_num)
-+ return 0;
-+
-+ for (i = 0; i < ARRAY_SIZE(driver_hash); i++) {
-+ struct caam_hash_alg *t_alg;
-+ struct caam_hash_template *alg = driver_hash + i;
-+
-+ /* register hmac version */
-+ t_alg = caam_hash_alloc(dev, alg, true);
-+ if (IS_ERR(t_alg)) {
-+ err = PTR_ERR(t_alg);
-+ dev_warn(dev, "%s hash alg allocation failed: %d\n",
-+ alg->driver_name, err);
-+ continue;
-+ }
-+
-+ err = crypto_register_ahash(&t_alg->ahash_alg);
-+ if (err) {
-+ dev_warn(dev, "%s alg registration failed: %d\n",
-+ t_alg->ahash_alg.halg.base.cra_driver_name,
-+ err);
-+ kfree(t_alg);
-+ } else {
-+ list_add_tail(&t_alg->entry, &hash_list);
-+ }
-+
-+ /* register unkeyed version */
-+ t_alg = caam_hash_alloc(dev, alg, false);
-+ if (IS_ERR(t_alg)) {
-+ err = PTR_ERR(t_alg);
-+ dev_warn(dev, "%s alg allocation failed: %d\n",
-+ alg->driver_name, err);
-+ continue;
-+ }
-+
-+ err = crypto_register_ahash(&t_alg->ahash_alg);
-+ if (err) {
-+ dev_warn(dev, "%s alg registration failed: %d\n",
-+ t_alg->ahash_alg.halg.base.cra_driver_name,
-+ err);
-+ kfree(t_alg);
-+ } else {
-+ list_add_tail(&t_alg->entry, &hash_list);
-+ }
-+ }
-+ if (!list_empty(&hash_list))
-+ dev_info(dev, "hash algorithms registered in /proc/crypto\n");
-+
-+ return err;
-+
-+err_bind:
-+ dpaa2_dpseci_dpio_free(priv);
-+err_dpio_setup:
-+ dpaa2_dpseci_free(priv);
-+err_dpseci_setup:
-+ free_percpu(priv->ppriv);
-+err_alloc_ppriv:
-+ fsl_mc_portal_free(priv->mc_io);
-+err_dma_mask:
-+ kmem_cache_destroy(qi_cache);
-+err_qicache:
-+ dev_set_drvdata(dev, NULL);
-+
-+ return err;
-+}
-+
-+static int __cold dpaa2_caam_remove(struct fsl_mc_device *ls_dev)
-+{
-+ struct device *dev;
-+ struct dpaa2_caam_priv *priv;
-+ int i;
-+
-+ dev = &ls_dev->dev;
-+ priv = dev_get_drvdata(dev);
-+
-+ for (i = 0; i < ARRAY_SIZE(driver_aeads); i++) {
-+ struct caam_aead_alg *t_alg = driver_aeads + i;
-+
-+ if (t_alg->registered)
-+ crypto_unregister_aead(&t_alg->aead);
-+ }
-+
-+ if (alg_list.next) {
-+ struct caam_crypto_alg *t_alg, *n;
-+
-+ list_for_each_entry_safe(t_alg, n, &alg_list, entry) {
-+ crypto_unregister_alg(&t_alg->crypto_alg);
-+ list_del(&t_alg->entry);
-+ kfree(t_alg);
-+ }
-+ }
-+
-+ if (hash_list.next) {
-+ struct caam_hash_alg *t_hash_alg, *p;
-+
-+ list_for_each_entry_safe(t_hash_alg, p, &hash_list, entry) {
-+ crypto_unregister_ahash(&t_hash_alg->ahash_alg);
-+ list_del(&t_hash_alg->entry);
-+ kfree(t_hash_alg);
-+ }
-+ }
-+
-+ dpaa2_dpseci_disable(priv);
-+ dpaa2_dpseci_dpio_free(priv);
-+ dpaa2_dpseci_free(priv);
-+ free_percpu(priv->ppriv);
-+ fsl_mc_portal_free(priv->mc_io);
-+ dev_set_drvdata(dev, NULL);
-+ kmem_cache_destroy(qi_cache);
-+
-+ return 0;
-+}
-+
-+int dpaa2_caam_enqueue(struct device *dev, struct caam_request *req)
-+{
-+ struct dpaa2_fd fd;
-+ struct dpaa2_caam_priv *priv = dev_get_drvdata(dev);
-+ int err = 0, i, id;
-+
-+ if (IS_ERR(req))
-+ return PTR_ERR(req);
-+
-+ if (priv->cscn_mem) {
-+ dma_sync_single_for_cpu(priv->dev, priv->cscn_dma,
-+ DPAA2_CSCN_SIZE,
-+ DMA_FROM_DEVICE);
-+ if (unlikely(dpaa2_cscn_state_congested(priv->cscn_mem_aligned))) {
-+ dev_dbg_ratelimited(dev, "Dropping request\n");
-+ return -EBUSY;
-+ }
-+ }
-+
-+ dpaa2_fl_set_flc(&req->fd_flt[1], req->flc_dma);
-+
-+ req->fd_flt_dma = dma_map_single(dev, req->fd_flt, sizeof(req->fd_flt),
-+ DMA_BIDIRECTIONAL);
-+ if (dma_mapping_error(dev, req->fd_flt_dma)) {
-+ dev_err(dev, "DMA mapping error for QI enqueue request\n");
-+ goto err_out;
-+ }
-+
-+ memset(&fd, 0, sizeof(fd));
-+ dpaa2_fd_set_format(&fd, dpaa2_fd_list);
-+ dpaa2_fd_set_addr(&fd, req->fd_flt_dma);
-+ dpaa2_fd_set_len(&fd, dpaa2_fl_get_len(&req->fd_flt[1]));
-+ dpaa2_fd_set_flc(&fd, req->flc_dma);
-+
-+ /*
-+ * There is no guarantee that preemption is disabled here,
-+ * thus take action.
-+ */
-+ preempt_disable();
-+ id = smp_processor_id() % priv->dpseci_attr.num_tx_queues;
-+ for (i = 0; i < (priv->dpseci_attr.num_tx_queues << 1); i++) {
-+ err = dpaa2_io_service_enqueue_fq(NULL,
-+ priv->tx_queue_attr[id].fqid,
-+ &fd);
-+ if (err != -EBUSY)
-+ break;
-+ }
-+ preempt_enable();
-+
-+ if (unlikely(err < 0)) {
-+ dev_err(dev, "Error enqueuing frame: %d\n", err);
-+ goto err_out;
-+ }
-+
-+ return -EINPROGRESS;
-+
-+err_out:
-+ dma_unmap_single(dev, req->fd_flt_dma, sizeof(req->fd_flt),
-+ DMA_BIDIRECTIONAL);
-+ return -EIO;
-+}
-+EXPORT_SYMBOL(dpaa2_caam_enqueue);
-+
-+const struct fsl_mc_device_id dpaa2_caam_match_id_table[] = {
-+ {
-+ .vendor = FSL_MC_VENDOR_FREESCALE,
-+ .obj_type = "dpseci",
-+ },
-+ { .vendor = 0x0 }
-+};
-+
-+static struct fsl_mc_driver dpaa2_caam_driver = {
-+ .driver = {
-+ .name = KBUILD_MODNAME,
-+ .owner = THIS_MODULE,
-+ },
-+ .probe = dpaa2_caam_probe,
-+ .remove = dpaa2_caam_remove,
-+ .match_id_table = dpaa2_caam_match_id_table
-+};
-+
-+MODULE_LICENSE("Dual BSD/GPL");
-+MODULE_AUTHOR("Freescale Semiconductor, Inc");
-+MODULE_DESCRIPTION("Freescale DPAA2 CAAM Driver");
-+
-+module_fsl_mc_driver(dpaa2_caam_driver);
---- /dev/null
-+++ b/drivers/crypto/caam/caamalg_qi2.h
-@@ -0,0 +1,283 @@
-+/*
-+ * Copyright 2015-2016 Freescale Semiconductor Inc.
-+ * Copyright 2017 NXP
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the names of the above-listed copyright holders nor the
-+ * names of any contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
-+ * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
-+ * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
-+ * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE
-+ * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
-+ * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
-+ * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
-+ * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
-+ * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
-+ * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
-+ * POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#ifndef _CAAMALG_QI2_H_
-+#define _CAAMALG_QI2_H_
-+
-+#include "../../../drivers/staging/fsl-mc/include/dpaa2-io.h"
-+#include "../../../drivers/staging/fsl-mc/include/dpaa2-fd.h"
-+#include <linux/threads.h>
-+#include "dpseci.h"
-+#include "desc_constr.h"
-+
-+#define DPAA2_CAAM_STORE_SIZE 16
-+/* NAPI weight *must* be a multiple of the store size. */
-+#define DPAA2_CAAM_NAPI_WEIGHT 64
-+
-+/* The congestion entrance threshold was chosen so that on LS2088
-+ * we support the maximum throughput for the available memory
-+ */
-+#define DPAA2_SEC_CONG_ENTRY_THRESH (128 * 1024 * 1024)
-+#define DPAA2_SEC_CONG_EXIT_THRESH (DPAA2_SEC_CONG_ENTRY_THRESH * 9 / 10)
-+
-+/**
-+ * dpaa2_caam_priv - driver private data
-+ * @dpseci_id: DPSECI object unique ID
-+ * @major_ver: DPSECI major version
-+ * @minor_ver: DPSECI minor version
-+ * @dpseci_attr: DPSECI attributes
-+ * @sec_attr: SEC engine attributes
-+ * @rx_queue_attr: array of Rx queue attributes
-+ * @tx_queue_attr: array of Tx queue attributes
-+ * @cscn_mem: pointer to memory region containing the
-+ * dpaa2_cscn struct; it's size is larger than
-+ * sizeof(struct dpaa2_cscn) to accommodate alignment
-+ * @cscn_mem_aligned: pointer to struct dpaa2_cscn; it is computed
-+ * as PTR_ALIGN(cscn_mem, DPAA2_CSCN_ALIGN)
-+ * @cscn_dma: dma address used by the QMAN to write CSCN messages
-+ * @dev: device associated with the DPSECI object
-+ * @mc_io: pointer to MC portal's I/O object
-+ * @domain: IOMMU domain
-+ * @ppriv: per CPU pointers to privata data
-+ */
-+struct dpaa2_caam_priv {
-+ int dpsec_id;
-+
-+ u16 major_ver;
-+ u16 minor_ver;
-+
-+ struct dpseci_attr dpseci_attr;
-+ struct dpseci_sec_attr sec_attr;
-+ struct dpseci_rx_queue_attr rx_queue_attr[DPSECI_PRIO_NUM];
-+ struct dpseci_tx_queue_attr tx_queue_attr[DPSECI_PRIO_NUM];
-+ int num_pairs;
-+
-+ /* congestion */
-+ void *cscn_mem;
-+ void *cscn_mem_aligned;
-+ dma_addr_t cscn_dma;
-+
-+ struct device *dev;
-+ struct fsl_mc_io *mc_io;
-+ struct iommu_domain *domain;
-+
-+ struct dpaa2_caam_priv_per_cpu __percpu *ppriv;
-+};
-+
-+/**
-+ * dpaa2_caam_priv_per_cpu - per CPU private data
-+ * @napi: napi structure
-+ * @net_dev: netdev used by napi
-+ * @req_fqid: (virtual) request (Tx / enqueue) FQID
-+ * @rsp_fqid: (virtual) response (Rx / dequeue) FQID
-+ * @prio: internal queue number - index for dpaa2_caam_priv.*_queue_attr
-+ * @nctx: notification context of response FQ
-+ * @store: where dequeued frames are stored
-+ * @priv: backpointer to dpaa2_caam_priv
-+ */
-+struct dpaa2_caam_priv_per_cpu {
-+ struct napi_struct napi;
-+ struct net_device net_dev;
-+ int req_fqid;
-+ int rsp_fqid;
-+ int prio;
-+ struct dpaa2_io_notification_ctx nctx;
-+ struct dpaa2_io_store *store;
-+ struct dpaa2_caam_priv *priv;
-+};
-+
-+/*
-+ * The CAAM QI hardware constructs a job descriptor which points
-+ * to shared descriptor (as pointed by context_a of FQ to CAAM).
-+ * When the job descriptor is executed by deco, the whole job
-+ * descriptor together with shared descriptor gets loaded in
-+ * deco buffer which is 64 words long (each 32-bit).
-+ *
-+ * The job descriptor constructed by QI hardware has layout:
-+ *
-+ * HEADER (1 word)
-+ * Shdesc ptr (1 or 2 words)
-+ * SEQ_OUT_PTR (1 word)
-+ * Out ptr (1 or 2 words)
-+ * Out length (1 word)
-+ * SEQ_IN_PTR (1 word)
-+ * In ptr (1 or 2 words)
-+ * In length (1 word)
-+ *
-+ * The shdesc ptr is used to fetch shared descriptor contents
-+ * into deco buffer.
-+ *
-+ * Apart from shdesc contents, the total number of words that
-+ * get loaded in deco buffer are '8' or '11'. The remaining words
-+ * in deco buffer can be used for storing shared descriptor.
-+ */
-+#define MAX_SDLEN ((CAAM_DESC_BYTES_MAX - DESC_JOB_IO_LEN) / CAAM_CMD_SZ)
-+
-+/* Length of a single buffer in the QI driver memory cache */
-+#define CAAM_QI_MEMCACHE_SIZE 512
-+
-+/*
-+ * aead_edesc - s/w-extended aead descriptor
-+ * @src_nents: number of segments in input scatterlist
-+ * @dst_nents: number of segments in output scatterlist
-+ * @iv_dma: dma address of iv for checking continuity and link table
-+ * @qm_sg_bytes: length of dma mapped h/w link table
-+ * @qm_sg_dma: bus physical mapped address of h/w link table
-+ * @assoclen: associated data length, in CAAM endianness
-+ * @assoclen_dma: bus physical mapped address of req->assoclen
-+ * @sgt: the h/w link table
-+ */
-+struct aead_edesc {
-+ int src_nents;
-+ int dst_nents;
-+ dma_addr_t iv_dma;
-+ int qm_sg_bytes;
-+ dma_addr_t qm_sg_dma;
-+ unsigned int assoclen;
-+ dma_addr_t assoclen_dma;
-+#define CAAM_QI_MAX_AEAD_SG \
-+ ((CAAM_QI_MEMCACHE_SIZE - offsetof(struct aead_edesc, sgt)) / \
-+ sizeof(struct dpaa2_sg_entry))
-+ struct dpaa2_sg_entry sgt[0];
-+};
-+
-+/*
-+ * tls_edesc - s/w-extended tls descriptor
-+ * @src_nents: number of segments in input scatterlist
-+ * @dst_nents: number of segments in output scatterlist
-+ * @iv_dma: dma address of iv for checking continuity and link table
-+ * @qm_sg_bytes: length of dma mapped h/w link table
-+ * @qm_sg_dma: bus physical mapped address of h/w link table
-+ * @tmp: array of scatterlists used by 'scatterwalk_ffwd'
-+ * @dst: pointer to output scatterlist, usefull for unmapping
-+ * @sgt: the h/w link table
-+ */
-+struct tls_edesc {
-+ int src_nents;
-+ int dst_nents;
-+ dma_addr_t iv_dma;
-+ int qm_sg_bytes;
-+ dma_addr_t qm_sg_dma;
-+ struct scatterlist tmp[2];
-+ struct scatterlist *dst;
-+ struct dpaa2_sg_entry sgt[0];
-+};
-+
-+/*
-+ * ablkcipher_edesc - s/w-extended ablkcipher descriptor
-+ * @src_nents: number of segments in input scatterlist
-+ * @dst_nents: number of segments in output scatterlist
-+ * @iv_dma: dma address of iv for checking continuity and link table
-+ * @qm_sg_bytes: length of dma mapped qm_sg space
-+ * @qm_sg_dma: I/O virtual address of h/w link table
-+ * @sgt: the h/w link table
-+ */
-+struct ablkcipher_edesc {
-+ int src_nents;
-+ int dst_nents;
-+ dma_addr_t iv_dma;
-+ int qm_sg_bytes;
-+ dma_addr_t qm_sg_dma;
-+#define CAAM_QI_MAX_ABLKCIPHER_SG \
-+ ((CAAM_QI_MEMCACHE_SIZE - offsetof(struct ablkcipher_edesc, sgt)) / \
-+ sizeof(struct dpaa2_sg_entry))
-+ struct dpaa2_sg_entry sgt[0];
-+};
-+
-+/*
-+ * ahash_edesc - s/w-extended ahash descriptor
-+ * @dst_dma: I/O virtual address of req->result
-+ * @qm_sg_dma: I/O virtual address of h/w link table
-+ * @src_nents: number of segments in input scatterlist
-+ * @qm_sg_bytes: length of dma mapped qm_sg space
-+ * @sgt: pointer to h/w link table
-+ */
-+struct ahash_edesc {
-+ dma_addr_t dst_dma;
-+ dma_addr_t qm_sg_dma;
-+ int src_nents;
-+ int qm_sg_bytes;
-+ struct dpaa2_sg_entry sgt[0];
-+};
-+
-+/**
-+ * caam_flc - Flow Context (FLC)
-+ * @flc: Flow Context options
-+ * @sh_desc: Shared Descriptor
-+ */
-+struct caam_flc {
-+ u32 flc[16];
-+ u32 sh_desc[MAX_SDLEN];
-+} ____cacheline_aligned;
-+
-+enum optype {
-+ ENCRYPT = 0,
-+ DECRYPT,
-+ GIVENCRYPT,
-+ NUM_OP
-+};
-+
-+/**
-+ * caam_request - the request structure the driver application should fill while
-+ * submitting a job to driver.
-+ * @fd_flt: Frame list table defining input and output
-+ * fd_flt[0] - FLE pointing to output buffer
-+ * fd_flt[1] - FLE pointing to input buffer
-+ * @fd_flt_dma: DMA address for the frame list table
-+ * @flc: Flow Context
-+ * @flc_dma: I/O virtual address of Flow Context
-+ * @op_type: operation type
-+ * @cbk: Callback function to invoke when job is completed
-+ * @ctx: arbit context attached with request by the application
-+ * @edesc: extended descriptor; points to one of {ablkcipher,aead}_edesc
-+ */
-+struct caam_request {
-+ struct dpaa2_fl_entry fd_flt[2];
-+ dma_addr_t fd_flt_dma;
-+ struct caam_flc *flc;
-+ dma_addr_t flc_dma;
-+ enum optype op_type;
-+ void (*cbk)(void *ctx, u32 err);
-+ void *ctx;
-+ void *edesc;
-+};
-+
-+/**
-+ * dpaa2_caam_enqueue() - enqueue a crypto request
-+ * @dev: device associated with the DPSECI object
-+ * @req: pointer to caam_request
-+ */
-+int dpaa2_caam_enqueue(struct device *dev, struct caam_request *req);
-+
-+#endif /* _CAAMALG_QI2_H_ */
---- a/drivers/crypto/caam/caamhash.c
-+++ b/drivers/crypto/caam/caamhash.c
-@@ -62,6 +62,7 @@
- #include "error.h"
- #include "sg_sw_sec4.h"
- #include "key_gen.h"
-+#include "caamhash_desc.h"
-
- #define CAAM_CRA_PRIORITY 3000
-
-@@ -71,14 +72,6 @@
- #define CAAM_MAX_HASH_BLOCK_SIZE SHA512_BLOCK_SIZE
- #define CAAM_MAX_HASH_DIGEST_SIZE SHA512_DIGEST_SIZE
-
--/* length of descriptors text */
--#define DESC_AHASH_BASE (4 * CAAM_CMD_SZ)
--#define DESC_AHASH_UPDATE_LEN (6 * CAAM_CMD_SZ)
--#define DESC_AHASH_UPDATE_FIRST_LEN (DESC_AHASH_BASE + 4 * CAAM_CMD_SZ)
--#define DESC_AHASH_FINAL_LEN (DESC_AHASH_BASE + 5 * CAAM_CMD_SZ)
--#define DESC_AHASH_FINUP_LEN (DESC_AHASH_BASE + 5 * CAAM_CMD_SZ)
--#define DESC_AHASH_DIGEST_LEN (DESC_AHASH_BASE + 4 * CAAM_CMD_SZ)
--
- #define DESC_HASH_MAX_USED_BYTES (DESC_AHASH_FINAL_LEN + \
- CAAM_MAX_HASH_KEY_SIZE)
- #define DESC_HASH_MAX_USED_LEN (DESC_HASH_MAX_USED_BYTES / CAAM_CMD_SZ)
-@@ -103,20 +96,15 @@ struct caam_hash_ctx {
- u32 sh_desc_update_first[DESC_HASH_MAX_USED_LEN] ____cacheline_aligned;
- u32 sh_desc_fin[DESC_HASH_MAX_USED_LEN] ____cacheline_aligned;
- u32 sh_desc_digest[DESC_HASH_MAX_USED_LEN] ____cacheline_aligned;
-- u32 sh_desc_finup[DESC_HASH_MAX_USED_LEN] ____cacheline_aligned;
- dma_addr_t sh_desc_update_dma ____cacheline_aligned;
- dma_addr_t sh_desc_update_first_dma;
- dma_addr_t sh_desc_fin_dma;
- dma_addr_t sh_desc_digest_dma;
-- dma_addr_t sh_desc_finup_dma;
-+ enum dma_data_direction dir;
- struct device *jrdev;
-- u32 alg_type;
-- u32 alg_op;
- u8 key[CAAM_MAX_HASH_KEY_SIZE];
-- dma_addr_t key_dma;
- int ctx_len;
-- unsigned int split_key_len;
-- unsigned int split_key_pad_len;
-+ struct alginfo adata;
- };
-
- /* ahash state */
-@@ -143,6 +131,31 @@ struct caam_export_state {
- int (*finup)(struct ahash_request *req);
- };
-
-+static inline void switch_buf(struct caam_hash_state *state)
-+{
-+ state->current_buf ^= 1;
-+}
-+
-+static inline u8 *current_buf(struct caam_hash_state *state)
-+{
-+ return state->current_buf ? state->buf_1 : state->buf_0;
-+}
-+
-+static inline u8 *alt_buf(struct caam_hash_state *state)
-+{
-+ return state->current_buf ? state->buf_0 : state->buf_1;
-+}
-+
-+static inline int *current_buflen(struct caam_hash_state *state)
-+{
-+ return state->current_buf ? &state->buflen_1 : &state->buflen_0;
-+}
-+
-+static inline int *alt_buflen(struct caam_hash_state *state)
-+{
-+ return state->current_buf ? &state->buflen_0 : &state->buflen_1;
-+}
-+
- /* Common job descriptor seq in/out ptr routines */
-
- /* Map state->caam_ctx, and append seq_out_ptr command that points to it */
-@@ -175,40 +188,31 @@ static inline dma_addr_t map_seq_out_ptr
- return dst_dma;
- }
-
--/* Map current buffer in state and put it in link table */
--static inline dma_addr_t buf_map_to_sec4_sg(struct device *jrdev,
-- struct sec4_sg_entry *sec4_sg,
-- u8 *buf, int buflen)
-+/* Map current buffer in state (if length > 0) and put it in link table */
-+static inline int buf_map_to_sec4_sg(struct device *jrdev,
-+ struct sec4_sg_entry *sec4_sg,
-+ struct caam_hash_state *state)
- {
-- dma_addr_t buf_dma;
-+ int buflen = *current_buflen(state);
-
-- buf_dma = dma_map_single(jrdev, buf, buflen, DMA_TO_DEVICE);
-- dma_to_sec4_sg_one(sec4_sg, buf_dma, buflen, 0);
-+ if (!buflen)
-+ return 0;
-
-- return buf_dma;
--}
-+ state->buf_dma = dma_map_single(jrdev, current_buf(state), buflen,
-+ DMA_TO_DEVICE);
-+ if (dma_mapping_error(jrdev, state->buf_dma)) {
-+ dev_err(jrdev, "unable to map buf\n");
-+ state->buf_dma = 0;
-+ return -ENOMEM;
-+ }
-
--/*
-- * Only put buffer in link table if it contains data, which is possible,
-- * since a buffer has previously been used, and needs to be unmapped,
-- */
--static inline dma_addr_t
--try_buf_map_to_sec4_sg(struct device *jrdev, struct sec4_sg_entry *sec4_sg,
-- u8 *buf, dma_addr_t buf_dma, int buflen,
-- int last_buflen)
--{
-- if (buf_dma && !dma_mapping_error(jrdev, buf_dma))
-- dma_unmap_single(jrdev, buf_dma, last_buflen, DMA_TO_DEVICE);
-- if (buflen)
-- buf_dma = buf_map_to_sec4_sg(jrdev, sec4_sg, buf, buflen);
-- else
-- buf_dma = 0;
-+ dma_to_sec4_sg_one(sec4_sg, state->buf_dma, buflen, 0);
-
-- return buf_dma;
-+ return 0;
- }
-
- /* Map state->caam_ctx, and add it to link table */
--static inline int ctx_map_to_sec4_sg(u32 *desc, struct device *jrdev,
-+static inline int ctx_map_to_sec4_sg(struct device *jrdev,
- struct caam_hash_state *state, int ctx_len,
- struct sec4_sg_entry *sec4_sg, u32 flag)
- {
-@@ -224,124 +228,22 @@ static inline int ctx_map_to_sec4_sg(u32
- return 0;
- }
-
--/* Common shared descriptor commands */
--static inline void append_key_ahash(u32 *desc, struct caam_hash_ctx *ctx)
--{
-- append_key_as_imm(desc, ctx->key, ctx->split_key_pad_len,
-- ctx->split_key_len, CLASS_2 |
-- KEY_DEST_MDHA_SPLIT | KEY_ENC);
--}
--
--/* Append key if it has been set */
--static inline void init_sh_desc_key_ahash(u32 *desc, struct caam_hash_ctx *ctx)
--{
-- u32 *key_jump_cmd;
--
-- init_sh_desc(desc, HDR_SHARE_SERIAL);
--
-- if (ctx->split_key_len) {
-- /* Skip if already shared */
-- key_jump_cmd = append_jump(desc, JUMP_JSL | JUMP_TEST_ALL |
-- JUMP_COND_SHRD);
--
-- append_key_ahash(desc, ctx);
--
-- set_jump_tgt_here(desc, key_jump_cmd);
-- }
--
-- /* Propagate errors from shared to job descriptor */
-- append_cmd(desc, SET_OK_NO_PROP_ERRORS | CMD_LOAD);
--}
--
--/*
-- * For ahash read data from seqin following state->caam_ctx,
-- * and write resulting class2 context to seqout, which may be state->caam_ctx
-- * or req->result
-- */
--static inline void ahash_append_load_str(u32 *desc, int digestsize)
--{
-- /* Calculate remaining bytes to read */
-- append_math_add(desc, VARSEQINLEN, SEQINLEN, REG0, CAAM_CMD_SZ);
--
-- /* Read remaining bytes */
-- append_seq_fifo_load(desc, 0, FIFOLD_CLASS_CLASS2 | FIFOLD_TYPE_LAST2 |
-- FIFOLD_TYPE_MSG | KEY_VLF);
--
-- /* Store class2 context bytes */
-- append_seq_store(desc, digestsize, LDST_CLASS_2_CCB |
-- LDST_SRCDST_BYTE_CONTEXT);
--}
--
--/*
-- * For ahash update, final and finup, import context, read and write to seqout
-- */
--static inline void ahash_ctx_data_to_out(u32 *desc, u32 op, u32 state,
-- int digestsize,
-- struct caam_hash_ctx *ctx)
--{
-- init_sh_desc_key_ahash(desc, ctx);
--
-- /* Import context from software */
-- append_cmd(desc, CMD_SEQ_LOAD | LDST_SRCDST_BYTE_CONTEXT |
-- LDST_CLASS_2_CCB | ctx->ctx_len);
--
-- /* Class 2 operation */
-- append_operation(desc, op | state | OP_ALG_ENCRYPT);
--
-- /*
-- * Load from buf and/or src and write to req->result or state->context
-- */
-- ahash_append_load_str(desc, digestsize);
--}
--
--/* For ahash firsts and digest, read and write to seqout */
--static inline void ahash_data_to_out(u32 *desc, u32 op, u32 state,
-- int digestsize, struct caam_hash_ctx *ctx)
--{
-- init_sh_desc_key_ahash(desc, ctx);
--
-- /* Class 2 operation */
-- append_operation(desc, op | state | OP_ALG_ENCRYPT);
--
-- /*
-- * Load from buf and/or src and write to req->result or state->context
-- */
-- ahash_append_load_str(desc, digestsize);
--}
--
- static int ahash_set_sh_desc(struct crypto_ahash *ahash)
- {
- struct caam_hash_ctx *ctx = crypto_ahash_ctx(ahash);
- int digestsize = crypto_ahash_digestsize(ahash);
- struct device *jrdev = ctx->jrdev;
-- u32 have_key = 0;
-+ struct caam_drv_private *ctrlpriv = dev_get_drvdata(jrdev->parent);
- u32 *desc;
-
-- if (ctx->split_key_len)
-- have_key = OP_ALG_AAI_HMAC_PRECOMP;
-+ ctx->adata.key_virt = ctx->key;
-
- /* ahash_update shared descriptor */
- desc = ctx->sh_desc_update;
--
-- init_sh_desc(desc, HDR_SHARE_SERIAL);
--
-- /* Import context from software */
-- append_cmd(desc, CMD_SEQ_LOAD | LDST_SRCDST_BYTE_CONTEXT |
-- LDST_CLASS_2_CCB | ctx->ctx_len);
--
-- /* Class 2 operation */
-- append_operation(desc, ctx->alg_type | OP_ALG_AS_UPDATE |
-- OP_ALG_ENCRYPT);
--
-- /* Load data and write to result or context */
-- ahash_append_load_str(desc, ctx->ctx_len);
--
-- ctx->sh_desc_update_dma = dma_map_single(jrdev, desc, desc_bytes(desc),
-- DMA_TO_DEVICE);
-- if (dma_mapping_error(jrdev, ctx->sh_desc_update_dma)) {
-- dev_err(jrdev, "unable to map shared descriptor\n");
-- return -ENOMEM;
-- }
-+ cnstr_shdsc_ahash(desc, &ctx->adata, OP_ALG_AS_UPDATE, ctx->ctx_len,
-+ ctx->ctx_len, true, ctrlpriv->era);
-+ dma_sync_single_for_device(jrdev, ctx->sh_desc_update_dma,
-+ desc_bytes(desc), ctx->dir);
- #ifdef DEBUG
- print_hex_dump(KERN_ERR,
- "ahash update shdesc@"__stringify(__LINE__)": ",
-@@ -350,17 +252,10 @@ static int ahash_set_sh_desc(struct cryp
-
- /* ahash_update_first shared descriptor */
- desc = ctx->sh_desc_update_first;
--
-- ahash_data_to_out(desc, have_key | ctx->alg_type, OP_ALG_AS_INIT,
-- ctx->ctx_len, ctx);
--
-- ctx->sh_desc_update_first_dma = dma_map_single(jrdev, desc,
-- desc_bytes(desc),
-- DMA_TO_DEVICE);
-- if (dma_mapping_error(jrdev, ctx->sh_desc_update_first_dma)) {
-- dev_err(jrdev, "unable to map shared descriptor\n");
-- return -ENOMEM;
-- }
-+ cnstr_shdsc_ahash(desc, &ctx->adata, OP_ALG_AS_INIT, ctx->ctx_len,
-+ ctx->ctx_len, false, ctrlpriv->era);
-+ dma_sync_single_for_device(jrdev, ctx->sh_desc_update_first_dma,
-+ desc_bytes(desc), ctx->dir);
- #ifdef DEBUG
- print_hex_dump(KERN_ERR,
- "ahash update first shdesc@"__stringify(__LINE__)": ",
-@@ -369,53 +264,22 @@ static int ahash_set_sh_desc(struct cryp
-
- /* ahash_final shared descriptor */
- desc = ctx->sh_desc_fin;
--
-- ahash_ctx_data_to_out(desc, have_key | ctx->alg_type,
-- OP_ALG_AS_FINALIZE, digestsize, ctx);
--
-- ctx->sh_desc_fin_dma = dma_map_single(jrdev, desc, desc_bytes(desc),
-- DMA_TO_DEVICE);
-- if (dma_mapping_error(jrdev, ctx->sh_desc_fin_dma)) {
-- dev_err(jrdev, "unable to map shared descriptor\n");
-- return -ENOMEM;
-- }
-+ cnstr_shdsc_ahash(desc, &ctx->adata, OP_ALG_AS_FINALIZE, digestsize,
-+ ctx->ctx_len, true, ctrlpriv->era);
-+ dma_sync_single_for_device(jrdev, ctx->sh_desc_fin_dma,
-+ desc_bytes(desc), ctx->dir);
- #ifdef DEBUG
- print_hex_dump(KERN_ERR, "ahash final shdesc@"__stringify(__LINE__)": ",
- DUMP_PREFIX_ADDRESS, 16, 4, desc,
- desc_bytes(desc), 1);
- #endif
-
-- /* ahash_finup shared descriptor */
-- desc = ctx->sh_desc_finup;
--
-- ahash_ctx_data_to_out(desc, have_key | ctx->alg_type,
-- OP_ALG_AS_FINALIZE, digestsize, ctx);
--
-- ctx->sh_desc_finup_dma = dma_map_single(jrdev, desc, desc_bytes(desc),
-- DMA_TO_DEVICE);
-- if (dma_mapping_error(jrdev, ctx->sh_desc_finup_dma)) {
-- dev_err(jrdev, "unable to map shared descriptor\n");
-- return -ENOMEM;
-- }
--#ifdef DEBUG
-- print_hex_dump(KERN_ERR, "ahash finup shdesc@"__stringify(__LINE__)": ",
-- DUMP_PREFIX_ADDRESS, 16, 4, desc,
-- desc_bytes(desc), 1);
--#endif
--
- /* ahash_digest shared descriptor */
- desc = ctx->sh_desc_digest;
--
-- ahash_data_to_out(desc, have_key | ctx->alg_type, OP_ALG_AS_INITFINAL,
-- digestsize, ctx);
--
-- ctx->sh_desc_digest_dma = dma_map_single(jrdev, desc,
-- desc_bytes(desc),
-- DMA_TO_DEVICE);
-- if (dma_mapping_error(jrdev, ctx->sh_desc_digest_dma)) {
-- dev_err(jrdev, "unable to map shared descriptor\n");
-- return -ENOMEM;
-- }
-+ cnstr_shdsc_ahash(desc, &ctx->adata, OP_ALG_AS_INITFINAL, digestsize,
-+ ctx->ctx_len, false, ctrlpriv->era);
-+ dma_sync_single_for_device(jrdev, ctx->sh_desc_digest_dma,
-+ desc_bytes(desc), ctx->dir);
- #ifdef DEBUG
- print_hex_dump(KERN_ERR,
- "ahash digest shdesc@"__stringify(__LINE__)": ",
-@@ -426,14 +290,6 @@ static int ahash_set_sh_desc(struct cryp
- return 0;
- }
-
--static int gen_split_hash_key(struct caam_hash_ctx *ctx, const u8 *key_in,
-- u32 keylen)
--{
-- return gen_split_key(ctx->jrdev, ctx->key, ctx->split_key_len,
-- ctx->split_key_pad_len, key_in, keylen,
-- ctx->alg_op);
--}
--
- /* Digest hash size if it is too large */
- static int hash_digest_key(struct caam_hash_ctx *ctx, const u8 *key_in,
- u32 *keylen, u8 *key_out, u32 digestsize)
-@@ -469,7 +325,7 @@ static int hash_digest_key(struct caam_h
- }
-
- /* Job descriptor to perform unkeyed hash on key_in */
-- append_operation(desc, ctx->alg_type | OP_ALG_ENCRYPT |
-+ append_operation(desc, ctx->adata.algtype | OP_ALG_ENCRYPT |
- OP_ALG_AS_INITFINAL);
- append_seq_in_ptr(desc, src_dma, *keylen, 0);
- append_seq_fifo_load(desc, *keylen, FIFOLD_CLASS_CLASS2 |
-@@ -513,12 +369,10 @@ static int hash_digest_key(struct caam_h
- static int ahash_setkey(struct crypto_ahash *ahash,
- const u8 *key, unsigned int keylen)
- {
-- /* Sizes for MDHA pads (*not* keys): MD5, SHA1, 224, 256, 384, 512 */
-- static const u8 mdpadlen[] = { 16, 20, 32, 32, 64, 64 };
- struct caam_hash_ctx *ctx = crypto_ahash_ctx(ahash);
-- struct device *jrdev = ctx->jrdev;
- int blocksize = crypto_tfm_alg_blocksize(&ahash->base);
- int digestsize = crypto_ahash_digestsize(ahash);
-+ struct caam_drv_private *ctrlpriv = dev_get_drvdata(ctx->jrdev->parent);
- int ret;
- u8 *hashed_key = NULL;
-
-@@ -539,43 +393,29 @@ static int ahash_setkey(struct crypto_ah
- key = hashed_key;
- }
-
-- /* Pick class 2 key length from algorithm submask */
-- ctx->split_key_len = mdpadlen[(ctx->alg_op & OP_ALG_ALGSEL_SUBMASK) >>
-- OP_ALG_ALGSEL_SHIFT] * 2;
-- ctx->split_key_pad_len = ALIGN(ctx->split_key_len, 16);
--
--#ifdef DEBUG
-- printk(KERN_ERR "split_key_len %d split_key_pad_len %d\n",
-- ctx->split_key_len, ctx->split_key_pad_len);
-- print_hex_dump(KERN_ERR, "key in @"__stringify(__LINE__)": ",
-- DUMP_PREFIX_ADDRESS, 16, 4, key, keylen, 1);
--#endif
-+ /*
-+ * If DKP is supported, use it in the shared descriptor to generate
-+ * the split key.
-+ */
-+ if (ctrlpriv->era >= 6) {
-+ ctx->adata.key_inline = true;
-+ ctx->adata.keylen = keylen;
-+ ctx->adata.keylen_pad = split_key_len(ctx->adata.algtype &
-+ OP_ALG_ALGSEL_MASK);
-
-- ret = gen_split_hash_key(ctx, key, keylen);
-- if (ret)
-- goto bad_free_key;
-+ if (ctx->adata.keylen_pad > CAAM_MAX_HASH_KEY_SIZE)
-+ goto bad_free_key;
-
-- ctx->key_dma = dma_map_single(jrdev, ctx->key, ctx->split_key_pad_len,
-- DMA_TO_DEVICE);
-- if (dma_mapping_error(jrdev, ctx->key_dma)) {
-- dev_err(jrdev, "unable to map key i/o memory\n");
-- ret = -ENOMEM;
-- goto error_free_key;
-+ memcpy(ctx->key, key, keylen);
-+ } else {
-+ ret = gen_split_key(ctx->jrdev, ctx->key, &ctx->adata, key,
-+ keylen, CAAM_MAX_HASH_KEY_SIZE);
-+ if (ret)
-+ goto bad_free_key;
- }
--#ifdef DEBUG
-- print_hex_dump(KERN_ERR, "ctx.key@"__stringify(__LINE__)": ",
-- DUMP_PREFIX_ADDRESS, 16, 4, ctx->key,
-- ctx->split_key_pad_len, 1);
--#endif
-
-- ret = ahash_set_sh_desc(ahash);
-- if (ret) {
-- dma_unmap_single(jrdev, ctx->key_dma, ctx->split_key_pad_len,
-- DMA_TO_DEVICE);
-- }
-- error_free_key:
- kfree(hashed_key);
-- return ret;
-+ return ahash_set_sh_desc(ahash);
- bad_free_key:
- kfree(hashed_key);
- crypto_ahash_set_flags(ahash, CRYPTO_TFM_RES_BAD_KEY_LEN);
-@@ -604,6 +444,8 @@ static inline void ahash_unmap(struct de
- struct ahash_edesc *edesc,
- struct ahash_request *req, int dst_len)
- {
-+ struct caam_hash_state *state = ahash_request_ctx(req);
-+
- if (edesc->src_nents)
- dma_unmap_sg(dev, req->src, edesc->src_nents, DMA_TO_DEVICE);
- if (edesc->dst_dma)
-@@ -612,6 +454,12 @@ static inline void ahash_unmap(struct de
- if (edesc->sec4_sg_bytes)
- dma_unmap_single(dev, edesc->sec4_sg_dma,
- edesc->sec4_sg_bytes, DMA_TO_DEVICE);
-+
-+ if (state->buf_dma) {
-+ dma_unmap_single(dev, state->buf_dma, *current_buflen(state),
-+ DMA_TO_DEVICE);
-+ state->buf_dma = 0;
-+ }
- }
-
- static inline void ahash_unmap_ctx(struct device *dev,
-@@ -643,8 +491,7 @@ static void ahash_done(struct device *jr
- dev_err(jrdev, "%s %d: err 0x%x\n", __func__, __LINE__, err);
- #endif
-
-- edesc = (struct ahash_edesc *)((char *)desc -
-- offsetof(struct ahash_edesc, hw_desc));
-+ edesc = container_of(desc, struct ahash_edesc, hw_desc[0]);
- if (err)
- caam_jr_strstatus(jrdev, err);
-
-@@ -671,19 +518,19 @@ static void ahash_done_bi(struct device
- struct ahash_edesc *edesc;
- struct crypto_ahash *ahash = crypto_ahash_reqtfm(req);
- struct caam_hash_ctx *ctx = crypto_ahash_ctx(ahash);
--#ifdef DEBUG
- struct caam_hash_state *state = ahash_request_ctx(req);
-+#ifdef DEBUG
- int digestsize = crypto_ahash_digestsize(ahash);
-
- dev_err(jrdev, "%s %d: err 0x%x\n", __func__, __LINE__, err);
- #endif
-
-- edesc = (struct ahash_edesc *)((char *)desc -
-- offsetof(struct ahash_edesc, hw_desc));
-+ edesc = container_of(desc, struct ahash_edesc, hw_desc[0]);
- if (err)
- caam_jr_strstatus(jrdev, err);
-
- ahash_unmap_ctx(jrdev, edesc, req, ctx->ctx_len, DMA_BIDIRECTIONAL);
-+ switch_buf(state);
- kfree(edesc);
-
- #ifdef DEBUG
-@@ -713,8 +560,7 @@ static void ahash_done_ctx_src(struct de
- dev_err(jrdev, "%s %d: err 0x%x\n", __func__, __LINE__, err);
- #endif
-
-- edesc = (struct ahash_edesc *)((char *)desc -
-- offsetof(struct ahash_edesc, hw_desc));
-+ edesc = container_of(desc, struct ahash_edesc, hw_desc[0]);
- if (err)
- caam_jr_strstatus(jrdev, err);
-
-@@ -741,19 +587,19 @@ static void ahash_done_ctx_dst(struct de
- struct ahash_edesc *edesc;
- struct crypto_ahash *ahash = crypto_ahash_reqtfm(req);
- struct caam_hash_ctx *ctx = crypto_ahash_ctx(ahash);
--#ifdef DEBUG
- struct caam_hash_state *state = ahash_request_ctx(req);
-+#ifdef DEBUG
- int digestsize = crypto_ahash_digestsize(ahash);
-
- dev_err(jrdev, "%s %d: err 0x%x\n", __func__, __LINE__, err);
- #endif
-
-- edesc = (struct ahash_edesc *)((char *)desc -
-- offsetof(struct ahash_edesc, hw_desc));
-+ edesc = container_of(desc, struct ahash_edesc, hw_desc[0]);
- if (err)
- caam_jr_strstatus(jrdev, err);
-
- ahash_unmap_ctx(jrdev, edesc, req, ctx->ctx_len, DMA_FROM_DEVICE);
-+ switch_buf(state);
- kfree(edesc);
-
- #ifdef DEBUG
-@@ -835,13 +681,12 @@ static int ahash_update_ctx(struct ahash
- struct caam_hash_ctx *ctx = crypto_ahash_ctx(ahash);
- struct caam_hash_state *state = ahash_request_ctx(req);
- struct device *jrdev = ctx->jrdev;
-- gfp_t flags = (req->base.flags & (CRYPTO_TFM_REQ_MAY_BACKLOG |
-- CRYPTO_TFM_REQ_MAY_SLEEP)) ? GFP_KERNEL : GFP_ATOMIC;
-- u8 *buf = state->current_buf ? state->buf_1 : state->buf_0;
-- int *buflen = state->current_buf ? &state->buflen_1 : &state->buflen_0;
-- u8 *next_buf = state->current_buf ? state->buf_0 : state->buf_1;
-- int *next_buflen = state->current_buf ? &state->buflen_0 :
-- &state->buflen_1, last_buflen;
-+ gfp_t flags = (req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP) ?
-+ GFP_KERNEL : GFP_ATOMIC;
-+ u8 *buf = current_buf(state);
-+ int *buflen = current_buflen(state);
-+ u8 *next_buf = alt_buf(state);
-+ int *next_buflen = alt_buflen(state), last_buflen;
- int in_len = *buflen + req->nbytes, to_hash;
- u32 *desc;
- int src_nents, mapped_nents, sec4_sg_bytes, sec4_sg_src_index;
-@@ -890,15 +735,14 @@ static int ahash_update_ctx(struct ahash
- edesc->src_nents = src_nents;
- edesc->sec4_sg_bytes = sec4_sg_bytes;
-
-- ret = ctx_map_to_sec4_sg(desc, jrdev, state, ctx->ctx_len,
-+ ret = ctx_map_to_sec4_sg(jrdev, state, ctx->ctx_len,
- edesc->sec4_sg, DMA_BIDIRECTIONAL);
- if (ret)
- goto unmap_ctx;
-
-- state->buf_dma = try_buf_map_to_sec4_sg(jrdev,
-- edesc->sec4_sg + 1,
-- buf, state->buf_dma,
-- *buflen, last_buflen);
-+ ret = buf_map_to_sec4_sg(jrdev, edesc->sec4_sg + 1, state);
-+ if (ret)
-+ goto unmap_ctx;
-
- if (mapped_nents) {
- sg_to_sec4_sg_last(req->src, mapped_nents,
-@@ -909,12 +753,10 @@ static int ahash_update_ctx(struct ahash
- to_hash - *buflen,
- *next_buflen, 0);
- } else {
-- (edesc->sec4_sg + sec4_sg_src_index - 1)->len |=
-- cpu_to_caam32(SEC4_SG_LEN_FIN);
-+ sg_to_sec4_set_last(edesc->sec4_sg + sec4_sg_src_index -
-+ 1);
- }
-
-- state->current_buf = !state->current_buf;
--
- desc = edesc->hw_desc;
-
- edesc->sec4_sg_dma = dma_map_single(jrdev, edesc->sec4_sg,
-@@ -969,12 +811,9 @@ static int ahash_final_ctx(struct ahash_
- struct caam_hash_ctx *ctx = crypto_ahash_ctx(ahash);
- struct caam_hash_state *state = ahash_request_ctx(req);
- struct device *jrdev = ctx->jrdev;
-- gfp_t flags = (req->base.flags & (CRYPTO_TFM_REQ_MAY_BACKLOG |
-- CRYPTO_TFM_REQ_MAY_SLEEP)) ? GFP_KERNEL : GFP_ATOMIC;
-- u8 *buf = state->current_buf ? state->buf_1 : state->buf_0;
-- int buflen = state->current_buf ? state->buflen_1 : state->buflen_0;
-- int last_buflen = state->current_buf ? state->buflen_0 :
-- state->buflen_1;
-+ gfp_t flags = (req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP) ?
-+ GFP_KERNEL : GFP_ATOMIC;
-+ int buflen = *current_buflen(state);
- u32 *desc;
- int sec4_sg_bytes, sec4_sg_src_index;
- int digestsize = crypto_ahash_digestsize(ahash);
-@@ -994,18 +833,17 @@ static int ahash_final_ctx(struct ahash_
- desc = edesc->hw_desc;
-
- edesc->sec4_sg_bytes = sec4_sg_bytes;
-- edesc->src_nents = 0;
-
-- ret = ctx_map_to_sec4_sg(desc, jrdev, state, ctx->ctx_len,
-+ ret = ctx_map_to_sec4_sg(jrdev, state, ctx->ctx_len,
- edesc->sec4_sg, DMA_TO_DEVICE);
- if (ret)
- goto unmap_ctx;
-
-- state->buf_dma = try_buf_map_to_sec4_sg(jrdev, edesc->sec4_sg + 1,
-- buf, state->buf_dma, buflen,
-- last_buflen);
-- (edesc->sec4_sg + sec4_sg_src_index - 1)->len |=
-- cpu_to_caam32(SEC4_SG_LEN_FIN);
-+ ret = buf_map_to_sec4_sg(jrdev, edesc->sec4_sg + 1, state);
-+ if (ret)
-+ goto unmap_ctx;
-+
-+ sg_to_sec4_set_last(edesc->sec4_sg + sec4_sg_src_index - 1);
-
- edesc->sec4_sg_dma = dma_map_single(jrdev, edesc->sec4_sg,
- sec4_sg_bytes, DMA_TO_DEVICE);
-@@ -1048,12 +886,9 @@ static int ahash_finup_ctx(struct ahash_
- struct caam_hash_ctx *ctx = crypto_ahash_ctx(ahash);
- struct caam_hash_state *state = ahash_request_ctx(req);
- struct device *jrdev = ctx->jrdev;
-- gfp_t flags = (req->base.flags & (CRYPTO_TFM_REQ_MAY_BACKLOG |
-- CRYPTO_TFM_REQ_MAY_SLEEP)) ? GFP_KERNEL : GFP_ATOMIC;
-- u8 *buf = state->current_buf ? state->buf_1 : state->buf_0;
-- int buflen = state->current_buf ? state->buflen_1 : state->buflen_0;
-- int last_buflen = state->current_buf ? state->buflen_0 :
-- state->buflen_1;
-+ gfp_t flags = (req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP) ?
-+ GFP_KERNEL : GFP_ATOMIC;
-+ int buflen = *current_buflen(state);
- u32 *desc;
- int sec4_sg_src_index;
- int src_nents, mapped_nents;
-@@ -1082,7 +917,7 @@ static int ahash_finup_ctx(struct ahash_
-
- /* allocate space for base edesc and hw desc commands, link tables */
- edesc = ahash_edesc_alloc(ctx, sec4_sg_src_index + mapped_nents,
-- ctx->sh_desc_finup, ctx->sh_desc_finup_dma,
-+ ctx->sh_desc_fin, ctx->sh_desc_fin_dma,
- flags);
- if (!edesc) {
- dma_unmap_sg(jrdev, req->src, src_nents, DMA_TO_DEVICE);
-@@ -1093,14 +928,14 @@ static int ahash_finup_ctx(struct ahash_
-
- edesc->src_nents = src_nents;
-
-- ret = ctx_map_to_sec4_sg(desc, jrdev, state, ctx->ctx_len,
-+ ret = ctx_map_to_sec4_sg(jrdev, state, ctx->ctx_len,
- edesc->sec4_sg, DMA_TO_DEVICE);
- if (ret)
- goto unmap_ctx;
-
-- state->buf_dma = try_buf_map_to_sec4_sg(jrdev, edesc->sec4_sg + 1,
-- buf, state->buf_dma, buflen,
-- last_buflen);
-+ ret = buf_map_to_sec4_sg(jrdev, edesc->sec4_sg + 1, state);
-+ if (ret)
-+ goto unmap_ctx;
-
- ret = ahash_edesc_add_src(ctx, edesc, req, mapped_nents,
- sec4_sg_src_index, ctx->ctx_len + buflen,
-@@ -1136,15 +971,18 @@ static int ahash_digest(struct ahash_req
- {
- struct crypto_ahash *ahash = crypto_ahash_reqtfm(req);
- struct caam_hash_ctx *ctx = crypto_ahash_ctx(ahash);
-+ struct caam_hash_state *state = ahash_request_ctx(req);
- struct device *jrdev = ctx->jrdev;
-- gfp_t flags = (req->base.flags & (CRYPTO_TFM_REQ_MAY_BACKLOG |
-- CRYPTO_TFM_REQ_MAY_SLEEP)) ? GFP_KERNEL : GFP_ATOMIC;
-+ gfp_t flags = (req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP) ?
-+ GFP_KERNEL : GFP_ATOMIC;
- u32 *desc;
- int digestsize = crypto_ahash_digestsize(ahash);
- int src_nents, mapped_nents;
- struct ahash_edesc *edesc;
- int ret;
-
-+ state->buf_dma = 0;
-+
- src_nents = sg_nents_for_len(req->src, req->nbytes);
- if (src_nents < 0) {
- dev_err(jrdev, "Invalid number of src SG.\n");
-@@ -1215,10 +1053,10 @@ static int ahash_final_no_ctx(struct aha
- struct caam_hash_ctx *ctx = crypto_ahash_ctx(ahash);
- struct caam_hash_state *state = ahash_request_ctx(req);
- struct device *jrdev = ctx->jrdev;
-- gfp_t flags = (req->base.flags & (CRYPTO_TFM_REQ_MAY_BACKLOG |
-- CRYPTO_TFM_REQ_MAY_SLEEP)) ? GFP_KERNEL : GFP_ATOMIC;
-- u8 *buf = state->current_buf ? state->buf_1 : state->buf_0;
-- int buflen = state->current_buf ? state->buflen_1 : state->buflen_0;
-+ gfp_t flags = (req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP) ?
-+ GFP_KERNEL : GFP_ATOMIC;
-+ u8 *buf = current_buf(state);
-+ int buflen = *current_buflen(state);
- u32 *desc;
- int digestsize = crypto_ahash_digestsize(ahash);
- struct ahash_edesc *edesc;
-@@ -1246,7 +1084,6 @@ static int ahash_final_no_ctx(struct aha
- dev_err(jrdev, "unable to map dst\n");
- goto unmap;
- }
-- edesc->src_nents = 0;
-
- #ifdef DEBUG
- print_hex_dump(KERN_ERR, "jobdesc@"__stringify(__LINE__)": ",
-@@ -1276,13 +1113,12 @@ static int ahash_update_no_ctx(struct ah
- struct caam_hash_ctx *ctx = crypto_ahash_ctx(ahash);
- struct caam_hash_state *state = ahash_request_ctx(req);
- struct device *jrdev = ctx->jrdev;
-- gfp_t flags = (req->base.flags & (CRYPTO_TFM_REQ_MAY_BACKLOG |
-- CRYPTO_TFM_REQ_MAY_SLEEP)) ? GFP_KERNEL : GFP_ATOMIC;
-- u8 *buf = state->current_buf ? state->buf_1 : state->buf_0;
-- int *buflen = state->current_buf ? &state->buflen_1 : &state->buflen_0;
-- u8 *next_buf = state->current_buf ? state->buf_0 : state->buf_1;
-- int *next_buflen = state->current_buf ? &state->buflen_0 :
-- &state->buflen_1;
-+ gfp_t flags = (req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP) ?
-+ GFP_KERNEL : GFP_ATOMIC;
-+ u8 *buf = current_buf(state);
-+ int *buflen = current_buflen(state);
-+ u8 *next_buf = alt_buf(state);
-+ int *next_buflen = alt_buflen(state);
- int in_len = *buflen + req->nbytes, to_hash;
- int sec4_sg_bytes, src_nents, mapped_nents;
- struct ahash_edesc *edesc;
-@@ -1329,10 +1165,11 @@ static int ahash_update_no_ctx(struct ah
-
- edesc->src_nents = src_nents;
- edesc->sec4_sg_bytes = sec4_sg_bytes;
-- edesc->dst_dma = 0;
-
-- state->buf_dma = buf_map_to_sec4_sg(jrdev, edesc->sec4_sg,
-- buf, *buflen);
-+ ret = buf_map_to_sec4_sg(jrdev, edesc->sec4_sg, state);
-+ if (ret)
-+ goto unmap_ctx;
-+
- sg_to_sec4_sg_last(req->src, mapped_nents,
- edesc->sec4_sg + 1, 0);
-
-@@ -1342,8 +1179,6 @@ static int ahash_update_no_ctx(struct ah
- *next_buflen, 0);
- }
-
-- state->current_buf = !state->current_buf;
--
- desc = edesc->hw_desc;
-
- edesc->sec4_sg_dma = dma_map_single(jrdev, edesc->sec4_sg,
-@@ -1403,12 +1238,9 @@ static int ahash_finup_no_ctx(struct aha
- struct caam_hash_ctx *ctx = crypto_ahash_ctx(ahash);
- struct caam_hash_state *state = ahash_request_ctx(req);
- struct device *jrdev = ctx->jrdev;
-- gfp_t flags = (req->base.flags & (CRYPTO_TFM_REQ_MAY_BACKLOG |
-- CRYPTO_TFM_REQ_MAY_SLEEP)) ? GFP_KERNEL : GFP_ATOMIC;
-- u8 *buf = state->current_buf ? state->buf_1 : state->buf_0;
-- int buflen = state->current_buf ? state->buflen_1 : state->buflen_0;
-- int last_buflen = state->current_buf ? state->buflen_0 :
-- state->buflen_1;
-+ gfp_t flags = (req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP) ?
-+ GFP_KERNEL : GFP_ATOMIC;
-+ int buflen = *current_buflen(state);
- u32 *desc;
- int sec4_sg_bytes, sec4_sg_src_index, src_nents, mapped_nents;
- int digestsize = crypto_ahash_digestsize(ahash);
-@@ -1450,9 +1282,9 @@ static int ahash_finup_no_ctx(struct aha
- edesc->src_nents = src_nents;
- edesc->sec4_sg_bytes = sec4_sg_bytes;
-
-- state->buf_dma = try_buf_map_to_sec4_sg(jrdev, edesc->sec4_sg, buf,
-- state->buf_dma, buflen,
-- last_buflen);
-+ ret = buf_map_to_sec4_sg(jrdev, edesc->sec4_sg, state);
-+ if (ret)
-+ goto unmap;
-
- ret = ahash_edesc_add_src(ctx, edesc, req, mapped_nents, 1, buflen,
- req->nbytes);
-@@ -1496,11 +1328,10 @@ static int ahash_update_first(struct aha
- struct caam_hash_ctx *ctx = crypto_ahash_ctx(ahash);
- struct caam_hash_state *state = ahash_request_ctx(req);
- struct device *jrdev = ctx->jrdev;
-- gfp_t flags = (req->base.flags & (CRYPTO_TFM_REQ_MAY_BACKLOG |
-- CRYPTO_TFM_REQ_MAY_SLEEP)) ? GFP_KERNEL : GFP_ATOMIC;
-- u8 *next_buf = state->current_buf ? state->buf_1 : state->buf_0;
-- int *next_buflen = state->current_buf ?
-- &state->buflen_1 : &state->buflen_0;
-+ gfp_t flags = (req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP) ?
-+ GFP_KERNEL : GFP_ATOMIC;
-+ u8 *next_buf = alt_buf(state);
-+ int *next_buflen = alt_buflen(state);
- int to_hash;
- u32 *desc;
- int src_nents, mapped_nents;
-@@ -1545,7 +1376,6 @@ static int ahash_update_first(struct aha
- }
-
- edesc->src_nents = src_nents;
-- edesc->dst_dma = 0;
-
- ret = ahash_edesc_add_src(ctx, edesc, req, mapped_nents, 0, 0,
- to_hash);
-@@ -1582,6 +1412,7 @@ static int ahash_update_first(struct aha
- state->final = ahash_final_no_ctx;
- scatterwalk_map_and_copy(next_buf, req->src, 0,
- req->nbytes, 0);
-+ switch_buf(state);
- }
- #ifdef DEBUG
- print_hex_dump(KERN_ERR, "next buf@"__stringify(__LINE__)": ",
-@@ -1688,7 +1519,6 @@ struct caam_hash_template {
- unsigned int blocksize;
- struct ahash_alg template_ahash;
- u32 alg_type;
-- u32 alg_op;
- };
-
- /* ahash descriptors */
-@@ -1714,7 +1544,6 @@ static struct caam_hash_template driver_
- },
- },
- .alg_type = OP_ALG_ALGSEL_SHA1,
-- .alg_op = OP_ALG_ALGSEL_SHA1 | OP_ALG_AAI_HMAC,
- }, {
- .name = "sha224",
- .driver_name = "sha224-caam",
-@@ -1736,7 +1565,6 @@ static struct caam_hash_template driver_
- },
- },
- .alg_type = OP_ALG_ALGSEL_SHA224,
-- .alg_op = OP_ALG_ALGSEL_SHA224 | OP_ALG_AAI_HMAC,
- }, {
- .name = "sha256",
- .driver_name = "sha256-caam",
-@@ -1758,7 +1586,6 @@ static struct caam_hash_template driver_
- },
- },
- .alg_type = OP_ALG_ALGSEL_SHA256,
-- .alg_op = OP_ALG_ALGSEL_SHA256 | OP_ALG_AAI_HMAC,
- }, {
- .name = "sha384",
- .driver_name = "sha384-caam",
-@@ -1780,7 +1607,6 @@ static struct caam_hash_template driver_
- },
- },
- .alg_type = OP_ALG_ALGSEL_SHA384,
-- .alg_op = OP_ALG_ALGSEL_SHA384 | OP_ALG_AAI_HMAC,
- }, {
- .name = "sha512",
- .driver_name = "sha512-caam",
-@@ -1802,7 +1628,6 @@ static struct caam_hash_template driver_
- },
- },
- .alg_type = OP_ALG_ALGSEL_SHA512,
-- .alg_op = OP_ALG_ALGSEL_SHA512 | OP_ALG_AAI_HMAC,
- }, {
- .name = "md5",
- .driver_name = "md5-caam",
-@@ -1824,14 +1649,12 @@ static struct caam_hash_template driver_
- },
- },
- .alg_type = OP_ALG_ALGSEL_MD5,
-- .alg_op = OP_ALG_ALGSEL_MD5 | OP_ALG_AAI_HMAC,
- },
- };
-
- struct caam_hash_alg {
- struct list_head entry;
- int alg_type;
-- int alg_op;
- struct ahash_alg ahash_alg;
- };
-
-@@ -1853,6 +1676,8 @@ static int caam_hash_cra_init(struct cry
- HASH_MSG_LEN + SHA256_DIGEST_SIZE,
- HASH_MSG_LEN + 64,
- HASH_MSG_LEN + SHA512_DIGEST_SIZE };
-+ dma_addr_t dma_addr;
-+ struct caam_drv_private *priv;
-
- /*
- * Get a Job ring from Job Ring driver to ensure in-order
-@@ -1863,11 +1688,34 @@ static int caam_hash_cra_init(struct cry
- pr_err("Job Ring Device allocation for transform failed\n");
- return PTR_ERR(ctx->jrdev);
- }
-+
-+ priv = dev_get_drvdata(ctx->jrdev->parent);
-+ ctx->dir = priv->era >= 6 ? DMA_BIDIRECTIONAL : DMA_TO_DEVICE;
-+
-+ dma_addr = dma_map_single_attrs(ctx->jrdev, ctx->sh_desc_update,
-+ offsetof(struct caam_hash_ctx,
-+ sh_desc_update_dma),
-+ ctx->dir, DMA_ATTR_SKIP_CPU_SYNC);
-+ if (dma_mapping_error(ctx->jrdev, dma_addr)) {
-+ dev_err(ctx->jrdev, "unable to map shared descriptors\n");
-+ caam_jr_free(ctx->jrdev);
-+ return -ENOMEM;
-+ }
-+
-+ ctx->sh_desc_update_dma = dma_addr;
-+ ctx->sh_desc_update_first_dma = dma_addr +
-+ offsetof(struct caam_hash_ctx,
-+ sh_desc_update_first);
-+ ctx->sh_desc_fin_dma = dma_addr + offsetof(struct caam_hash_ctx,
-+ sh_desc_fin);
-+ ctx->sh_desc_digest_dma = dma_addr + offsetof(struct caam_hash_ctx,
-+ sh_desc_digest);
-+
- /* copy descriptor header template value */
-- ctx->alg_type = OP_TYPE_CLASS2_ALG | caam_hash->alg_type;
-- ctx->alg_op = OP_TYPE_CLASS2_ALG | caam_hash->alg_op;
-+ ctx->adata.algtype = OP_TYPE_CLASS2_ALG | caam_hash->alg_type;
-
-- ctx->ctx_len = runninglen[(ctx->alg_op & OP_ALG_ALGSEL_SUBMASK) >>
-+ ctx->ctx_len = runninglen[(ctx->adata.algtype &
-+ OP_ALG_ALGSEL_SUBMASK) >>
- OP_ALG_ALGSEL_SHIFT];
-
- crypto_ahash_set_reqsize(__crypto_ahash_cast(tfm),
-@@ -1879,30 +1727,10 @@ static void caam_hash_cra_exit(struct cr
- {
- struct caam_hash_ctx *ctx = crypto_tfm_ctx(tfm);
-
-- if (ctx->sh_desc_update_dma &&
-- !dma_mapping_error(ctx->jrdev, ctx->sh_desc_update_dma))
-- dma_unmap_single(ctx->jrdev, ctx->sh_desc_update_dma,
-- desc_bytes(ctx->sh_desc_update),
-- DMA_TO_DEVICE);
-- if (ctx->sh_desc_update_first_dma &&
-- !dma_mapping_error(ctx->jrdev, ctx->sh_desc_update_first_dma))
-- dma_unmap_single(ctx->jrdev, ctx->sh_desc_update_first_dma,
-- desc_bytes(ctx->sh_desc_update_first),
-- DMA_TO_DEVICE);
-- if (ctx->sh_desc_fin_dma &&
-- !dma_mapping_error(ctx->jrdev, ctx->sh_desc_fin_dma))
-- dma_unmap_single(ctx->jrdev, ctx->sh_desc_fin_dma,
-- desc_bytes(ctx->sh_desc_fin), DMA_TO_DEVICE);
-- if (ctx->sh_desc_digest_dma &&
-- !dma_mapping_error(ctx->jrdev, ctx->sh_desc_digest_dma))
-- dma_unmap_single(ctx->jrdev, ctx->sh_desc_digest_dma,
-- desc_bytes(ctx->sh_desc_digest),
-- DMA_TO_DEVICE);
-- if (ctx->sh_desc_finup_dma &&
-- !dma_mapping_error(ctx->jrdev, ctx->sh_desc_finup_dma))
-- dma_unmap_single(ctx->jrdev, ctx->sh_desc_finup_dma,
-- desc_bytes(ctx->sh_desc_finup), DMA_TO_DEVICE);
--
-+ dma_unmap_single_attrs(ctx->jrdev, ctx->sh_desc_update_dma,
-+ offsetof(struct caam_hash_ctx,
-+ sh_desc_update_dma),
-+ ctx->dir, DMA_ATTR_SKIP_CPU_SYNC);
- caam_jr_free(ctx->jrdev);
- }
-
-@@ -1961,7 +1789,6 @@ caam_hash_alloc(struct caam_hash_templat
- alg->cra_type = &crypto_ahash_type;
-
- t_alg->alg_type = template->alg_type;
-- t_alg->alg_op = template->alg_op;
-
- return t_alg;
- }
---- /dev/null
-+++ b/drivers/crypto/caam/caamhash_desc.c
-@@ -0,0 +1,108 @@
-+/*
-+ * Shared descriptors for ahash algorithms
-+ *
-+ * Copyright 2017 NXP
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the names of the above-listed copyright holders nor the
-+ * names of any contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
-+ * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
-+ * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
-+ * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE
-+ * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
-+ * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
-+ * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
-+ * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
-+ * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
-+ * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
-+ * POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#include "compat.h"
-+#include "desc_constr.h"
-+#include "caamhash_desc.h"
-+
-+/**
-+ * cnstr_shdsc_ahash - ahash shared descriptor
-+ * @desc: pointer to buffer used for descriptor construction
-+ * @adata: pointer to authentication transform definitions.
-+ * A split key is required for SEC Era < 6; the size of the split key
-+ * is specified in this case.
-+ * Valid algorithm values - one of OP_ALG_ALGSEL_{MD5, SHA1, SHA224,
-+ * SHA256, SHA384, SHA512}.
-+ * @state: algorithm state OP_ALG_AS_{INIT, FINALIZE, INITFINALIZE, UPDATE}
-+ * @digestsize: algorithm's digest size
-+ * @ctx_len: size of Context Register
-+ * @import_ctx: true if previous Context Register needs to be restored
-+ * must be true for ahash update and final
-+ * must be false for for ahash first and digest
-+ * @era: SEC Era
-+ */
-+void cnstr_shdsc_ahash(u32 * const desc, struct alginfo *adata, u32 state,
-+ int digestsize, int ctx_len, bool import_ctx, int era)
-+{
-+ u32 op = adata->algtype;
-+
-+ init_sh_desc(desc, HDR_SHARE_SERIAL);
-+
-+ /* Append key if it has been set; ahash update excluded */
-+ if (state != OP_ALG_AS_UPDATE && adata->keylen) {
-+ u32 *skip_key_load;
-+
-+ /* Skip key loading if already shared */
-+ skip_key_load = append_jump(desc, JUMP_JSL | JUMP_TEST_ALL |
-+ JUMP_COND_SHRD);
-+
-+ if (era < 6)
-+ append_key_as_imm(desc, adata->key_virt,
-+ adata->keylen_pad,
-+ adata->keylen, CLASS_2 |
-+ KEY_DEST_MDHA_SPLIT | KEY_ENC);
-+ else
-+ append_proto_dkp(desc, adata);
-+
-+ set_jump_tgt_here(desc, skip_key_load);
-+
-+ op |= OP_ALG_AAI_HMAC_PRECOMP;
-+ }
-+
-+ /* If needed, import context from software */
-+ if (import_ctx)
-+ append_seq_load(desc, ctx_len, LDST_CLASS_2_CCB |
-+ LDST_SRCDST_BYTE_CONTEXT);
-+
-+ /* Class 2 operation */
-+ append_operation(desc, op | state | OP_ALG_ENCRYPT);
-+
-+ /*
-+ * Load from buf and/or src and write to req->result or state->context
-+ * Calculate remaining bytes to read
-+ */
-+ append_math_add(desc, VARSEQINLEN, SEQINLEN, REG0, CAAM_CMD_SZ);
-+ /* Read remaining bytes */
-+ append_seq_fifo_load(desc, 0, FIFOLD_CLASS_CLASS2 | FIFOLD_TYPE_LAST2 |
-+ FIFOLD_TYPE_MSG | KEY_VLF);
-+ /* Store class2 context bytes */
-+ append_seq_store(desc, digestsize, LDST_CLASS_2_CCB |
-+ LDST_SRCDST_BYTE_CONTEXT);
-+}
-+EXPORT_SYMBOL(cnstr_shdsc_ahash);
-+
-+MODULE_LICENSE("Dual BSD/GPL");
-+MODULE_DESCRIPTION("FSL CAAM ahash descriptors support");
-+MODULE_AUTHOR("NXP Semiconductors");
---- /dev/null
-+++ b/drivers/crypto/caam/caamhash_desc.h
-@@ -0,0 +1,49 @@
-+/*
-+ * Shared descriptors for ahash algorithms
-+ *
-+ * Copyright 2017 NXP
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the names of the above-listed copyright holders nor the
-+ * names of any contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
-+ * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
-+ * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
-+ * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE
-+ * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
-+ * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
-+ * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
-+ * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
-+ * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
-+ * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
-+ * POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#ifndef _CAAMHASH_DESC_H_
-+#define _CAAMHASH_DESC_H_
-+
-+/* length of descriptors text */
-+#define DESC_AHASH_BASE (3 * CAAM_CMD_SZ)
-+#define DESC_AHASH_UPDATE_LEN (6 * CAAM_CMD_SZ)
-+#define DESC_AHASH_UPDATE_FIRST_LEN (DESC_AHASH_BASE + 4 * CAAM_CMD_SZ)
-+#define DESC_AHASH_FINAL_LEN (DESC_AHASH_BASE + 5 * CAAM_CMD_SZ)
-+#define DESC_AHASH_DIGEST_LEN (DESC_AHASH_BASE + 4 * CAAM_CMD_SZ)
-+
-+void cnstr_shdsc_ahash(u32 * const desc, struct alginfo *adata, u32 state,
-+ int digestsize, int ctx_len, bool import_ctx, int era);
-+
-+#endif /* _CAAMHASH_DESC_H_ */
---- a/drivers/crypto/caam/caampkc.c
-+++ b/drivers/crypto/caam/caampkc.c
-@@ -18,6 +18,10 @@
- #define DESC_RSA_PUB_LEN (2 * CAAM_CMD_SZ + sizeof(struct rsa_pub_pdb))
- #define DESC_RSA_PRIV_F1_LEN (2 * CAAM_CMD_SZ + \
- sizeof(struct rsa_priv_f1_pdb))
-+#define DESC_RSA_PRIV_F2_LEN (2 * CAAM_CMD_SZ + \
-+ sizeof(struct rsa_priv_f2_pdb))
-+#define DESC_RSA_PRIV_F3_LEN (2 * CAAM_CMD_SZ + \
-+ sizeof(struct rsa_priv_f3_pdb))
-
- static void rsa_io_unmap(struct device *dev, struct rsa_edesc *edesc,
- struct akcipher_request *req)
-@@ -54,6 +58,42 @@ static void rsa_priv_f1_unmap(struct dev
- dma_unmap_single(dev, pdb->d_dma, key->d_sz, DMA_TO_DEVICE);
- }
-
-+static void rsa_priv_f2_unmap(struct device *dev, struct rsa_edesc *edesc,
-+ struct akcipher_request *req)
-+{
-+ struct crypto_akcipher *tfm = crypto_akcipher_reqtfm(req);
-+ struct caam_rsa_ctx *ctx = akcipher_tfm_ctx(tfm);
-+ struct caam_rsa_key *key = &ctx->key;
-+ struct rsa_priv_f2_pdb *pdb = &edesc->pdb.priv_f2;
-+ size_t p_sz = key->p_sz;
-+ size_t q_sz = key->p_sz;
-+
-+ dma_unmap_single(dev, pdb->d_dma, key->d_sz, DMA_TO_DEVICE);
-+ dma_unmap_single(dev, pdb->p_dma, p_sz, DMA_TO_DEVICE);
-+ dma_unmap_single(dev, pdb->q_dma, q_sz, DMA_TO_DEVICE);
-+ dma_unmap_single(dev, pdb->tmp1_dma, p_sz, DMA_TO_DEVICE);
-+ dma_unmap_single(dev, pdb->tmp2_dma, q_sz, DMA_TO_DEVICE);
-+}
-+
-+static void rsa_priv_f3_unmap(struct device *dev, struct rsa_edesc *edesc,
-+ struct akcipher_request *req)
-+{
-+ struct crypto_akcipher *tfm = crypto_akcipher_reqtfm(req);
-+ struct caam_rsa_ctx *ctx = akcipher_tfm_ctx(tfm);
-+ struct caam_rsa_key *key = &ctx->key;
-+ struct rsa_priv_f3_pdb *pdb = &edesc->pdb.priv_f3;
-+ size_t p_sz = key->p_sz;
-+ size_t q_sz = key->p_sz;
-+
-+ dma_unmap_single(dev, pdb->p_dma, p_sz, DMA_TO_DEVICE);
-+ dma_unmap_single(dev, pdb->q_dma, q_sz, DMA_TO_DEVICE);
-+ dma_unmap_single(dev, pdb->dp_dma, p_sz, DMA_TO_DEVICE);
-+ dma_unmap_single(dev, pdb->dq_dma, q_sz, DMA_TO_DEVICE);
-+ dma_unmap_single(dev, pdb->c_dma, p_sz, DMA_TO_DEVICE);
-+ dma_unmap_single(dev, pdb->tmp1_dma, p_sz, DMA_TO_DEVICE);
-+ dma_unmap_single(dev, pdb->tmp2_dma, q_sz, DMA_TO_DEVICE);
-+}
-+
- /* RSA Job Completion handler */
- static void rsa_pub_done(struct device *dev, u32 *desc, u32 err, void *context)
- {
-@@ -90,6 +130,42 @@ static void rsa_priv_f1_done(struct devi
- akcipher_request_complete(req, err);
- }
-
-+static void rsa_priv_f2_done(struct device *dev, u32 *desc, u32 err,
-+ void *context)
-+{
-+ struct akcipher_request *req = context;
-+ struct rsa_edesc *edesc;
-+
-+ if (err)
-+ caam_jr_strstatus(dev, err);
-+
-+ edesc = container_of(desc, struct rsa_edesc, hw_desc[0]);
-+
-+ rsa_priv_f2_unmap(dev, edesc, req);
-+ rsa_io_unmap(dev, edesc, req);
-+ kfree(edesc);
-+
-+ akcipher_request_complete(req, err);
-+}
-+
-+static void rsa_priv_f3_done(struct device *dev, u32 *desc, u32 err,
-+ void *context)
-+{
-+ struct akcipher_request *req = context;
-+ struct rsa_edesc *edesc;
-+
-+ if (err)
-+ caam_jr_strstatus(dev, err);
-+
-+ edesc = container_of(desc, struct rsa_edesc, hw_desc[0]);
-+
-+ rsa_priv_f3_unmap(dev, edesc, req);
-+ rsa_io_unmap(dev, edesc, req);
-+ kfree(edesc);
-+
-+ akcipher_request_complete(req, err);
-+}
-+
- static struct rsa_edesc *rsa_edesc_alloc(struct akcipher_request *req,
- size_t desclen)
- {
-@@ -97,8 +173,8 @@ static struct rsa_edesc *rsa_edesc_alloc
- struct caam_rsa_ctx *ctx = akcipher_tfm_ctx(tfm);
- struct device *dev = ctx->dev;
- struct rsa_edesc *edesc;
-- gfp_t flags = (req->base.flags & (CRYPTO_TFM_REQ_MAY_BACKLOG |
-- CRYPTO_TFM_REQ_MAY_SLEEP)) ? GFP_KERNEL : GFP_ATOMIC;
-+ gfp_t flags = (req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP) ?
-+ GFP_KERNEL : GFP_ATOMIC;
- int sgc;
- int sec4_sg_index, sec4_sg_len = 0, sec4_sg_bytes;
- int src_nents, dst_nents;
-@@ -258,6 +334,172 @@ static int set_rsa_priv_f1_pdb(struct ak
- return 0;
- }
-
-+static int set_rsa_priv_f2_pdb(struct akcipher_request *req,
-+ struct rsa_edesc *edesc)
-+{
-+ struct crypto_akcipher *tfm = crypto_akcipher_reqtfm(req);
-+ struct caam_rsa_ctx *ctx = akcipher_tfm_ctx(tfm);
-+ struct caam_rsa_key *key = &ctx->key;
-+ struct device *dev = ctx->dev;
-+ struct rsa_priv_f2_pdb *pdb = &edesc->pdb.priv_f2;
-+ int sec4_sg_index = 0;
-+ size_t p_sz = key->p_sz;
-+ size_t q_sz = key->p_sz;
-+
-+ pdb->d_dma = dma_map_single(dev, key->d, key->d_sz, DMA_TO_DEVICE);
-+ if (dma_mapping_error(dev, pdb->d_dma)) {
-+ dev_err(dev, "Unable to map RSA private exponent memory\n");
-+ return -ENOMEM;
-+ }
-+
-+ pdb->p_dma = dma_map_single(dev, key->p, p_sz, DMA_TO_DEVICE);
-+ if (dma_mapping_error(dev, pdb->p_dma)) {
-+ dev_err(dev, "Unable to map RSA prime factor p memory\n");
-+ goto unmap_d;
-+ }
-+
-+ pdb->q_dma = dma_map_single(dev, key->q, q_sz, DMA_TO_DEVICE);
-+ if (dma_mapping_error(dev, pdb->q_dma)) {
-+ dev_err(dev, "Unable to map RSA prime factor q memory\n");
-+ goto unmap_p;
-+ }
-+
-+ pdb->tmp1_dma = dma_map_single(dev, key->tmp1, p_sz, DMA_TO_DEVICE);
-+ if (dma_mapping_error(dev, pdb->tmp1_dma)) {
-+ dev_err(dev, "Unable to map RSA tmp1 memory\n");
-+ goto unmap_q;
-+ }
-+
-+ pdb->tmp2_dma = dma_map_single(dev, key->tmp2, q_sz, DMA_TO_DEVICE);
-+ if (dma_mapping_error(dev, pdb->tmp2_dma)) {
-+ dev_err(dev, "Unable to map RSA tmp2 memory\n");
-+ goto unmap_tmp1;
-+ }
-+
-+ if (edesc->src_nents > 1) {
-+ pdb->sgf |= RSA_PRIV_PDB_SGF_G;
-+ pdb->g_dma = edesc->sec4_sg_dma;
-+ sec4_sg_index += edesc->src_nents;
-+ } else {
-+ pdb->g_dma = sg_dma_address(req->src);
-+ }
-+
-+ if (edesc->dst_nents > 1) {
-+ pdb->sgf |= RSA_PRIV_PDB_SGF_F;
-+ pdb->f_dma = edesc->sec4_sg_dma +
-+ sec4_sg_index * sizeof(struct sec4_sg_entry);
-+ } else {
-+ pdb->f_dma = sg_dma_address(req->dst);
-+ }
-+
-+ pdb->sgf |= (key->d_sz << RSA_PDB_D_SHIFT) | key->n_sz;
-+ pdb->p_q_len = (q_sz << RSA_PDB_Q_SHIFT) | p_sz;
-+
-+ return 0;
-+
-+unmap_tmp1:
-+ dma_unmap_single(dev, pdb->tmp1_dma, p_sz, DMA_TO_DEVICE);
-+unmap_q:
-+ dma_unmap_single(dev, pdb->q_dma, q_sz, DMA_TO_DEVICE);
-+unmap_p:
-+ dma_unmap_single(dev, pdb->p_dma, p_sz, DMA_TO_DEVICE);
-+unmap_d:
-+ dma_unmap_single(dev, pdb->d_dma, key->d_sz, DMA_TO_DEVICE);
-+
-+ return -ENOMEM;
-+}
-+
-+static int set_rsa_priv_f3_pdb(struct akcipher_request *req,
-+ struct rsa_edesc *edesc)
-+{
-+ struct crypto_akcipher *tfm = crypto_akcipher_reqtfm(req);
-+ struct caam_rsa_ctx *ctx = akcipher_tfm_ctx(tfm);
-+ struct caam_rsa_key *key = &ctx->key;
-+ struct device *dev = ctx->dev;
-+ struct rsa_priv_f3_pdb *pdb = &edesc->pdb.priv_f3;
-+ int sec4_sg_index = 0;
-+ size_t p_sz = key->p_sz;
-+ size_t q_sz = key->p_sz;
-+
-+ pdb->p_dma = dma_map_single(dev, key->p, p_sz, DMA_TO_DEVICE);
-+ if (dma_mapping_error(dev, pdb->p_dma)) {
-+ dev_err(dev, "Unable to map RSA prime factor p memory\n");
-+ return -ENOMEM;
-+ }
-+
-+ pdb->q_dma = dma_map_single(dev, key->q, q_sz, DMA_TO_DEVICE);
-+ if (dma_mapping_error(dev, pdb->q_dma)) {
-+ dev_err(dev, "Unable to map RSA prime factor q memory\n");
-+ goto unmap_p;
-+ }
-+
-+ pdb->dp_dma = dma_map_single(dev, key->dp, p_sz, DMA_TO_DEVICE);
-+ if (dma_mapping_error(dev, pdb->dp_dma)) {
-+ dev_err(dev, "Unable to map RSA exponent dp memory\n");
-+ goto unmap_q;
-+ }
-+
-+ pdb->dq_dma = dma_map_single(dev, key->dq, q_sz, DMA_TO_DEVICE);
-+ if (dma_mapping_error(dev, pdb->dq_dma)) {
-+ dev_err(dev, "Unable to map RSA exponent dq memory\n");
-+ goto unmap_dp;
-+ }
-+
-+ pdb->c_dma = dma_map_single(dev, key->qinv, p_sz, DMA_TO_DEVICE);
-+ if (dma_mapping_error(dev, pdb->c_dma)) {
-+ dev_err(dev, "Unable to map RSA CRT coefficient qinv memory\n");
-+ goto unmap_dq;
-+ }
-+
-+ pdb->tmp1_dma = dma_map_single(dev, key->tmp1, p_sz, DMA_TO_DEVICE);
-+ if (dma_mapping_error(dev, pdb->tmp1_dma)) {
-+ dev_err(dev, "Unable to map RSA tmp1 memory\n");
-+ goto unmap_qinv;
-+ }
-+
-+ pdb->tmp2_dma = dma_map_single(dev, key->tmp2, q_sz, DMA_TO_DEVICE);
-+ if (dma_mapping_error(dev, pdb->tmp2_dma)) {
-+ dev_err(dev, "Unable to map RSA tmp2 memory\n");
-+ goto unmap_tmp1;
-+ }
-+
-+ if (edesc->src_nents > 1) {
-+ pdb->sgf |= RSA_PRIV_PDB_SGF_G;
-+ pdb->g_dma = edesc->sec4_sg_dma;
-+ sec4_sg_index += edesc->src_nents;
-+ } else {
-+ pdb->g_dma = sg_dma_address(req->src);
-+ }
-+
-+ if (edesc->dst_nents > 1) {
-+ pdb->sgf |= RSA_PRIV_PDB_SGF_F;
-+ pdb->f_dma = edesc->sec4_sg_dma +
-+ sec4_sg_index * sizeof(struct sec4_sg_entry);
-+ } else {
-+ pdb->f_dma = sg_dma_address(req->dst);
-+ }
-+
-+ pdb->sgf |= key->n_sz;
-+ pdb->p_q_len = (q_sz << RSA_PDB_Q_SHIFT) | p_sz;
-+
-+ return 0;
-+
-+unmap_tmp1:
-+ dma_unmap_single(dev, pdb->tmp1_dma, p_sz, DMA_TO_DEVICE);
-+unmap_qinv:
-+ dma_unmap_single(dev, pdb->c_dma, p_sz, DMA_TO_DEVICE);
-+unmap_dq:
-+ dma_unmap_single(dev, pdb->dq_dma, q_sz, DMA_TO_DEVICE);
-+unmap_dp:
-+ dma_unmap_single(dev, pdb->dp_dma, p_sz, DMA_TO_DEVICE);
-+unmap_q:
-+ dma_unmap_single(dev, pdb->q_dma, q_sz, DMA_TO_DEVICE);
-+unmap_p:
-+ dma_unmap_single(dev, pdb->p_dma, p_sz, DMA_TO_DEVICE);
-+
-+ return -ENOMEM;
-+}
-+
- static int caam_rsa_enc(struct akcipher_request *req)
- {
- struct crypto_akcipher *tfm = crypto_akcipher_reqtfm(req);
-@@ -301,24 +543,14 @@ init_fail:
- return ret;
- }
-
--static int caam_rsa_dec(struct akcipher_request *req)
-+static int caam_rsa_dec_priv_f1(struct akcipher_request *req)
- {
- struct crypto_akcipher *tfm = crypto_akcipher_reqtfm(req);
- struct caam_rsa_ctx *ctx = akcipher_tfm_ctx(tfm);
-- struct caam_rsa_key *key = &ctx->key;
- struct device *jrdev = ctx->dev;
- struct rsa_edesc *edesc;
- int ret;
-
-- if (unlikely(!key->n || !key->d))
-- return -EINVAL;
--
-- if (req->dst_len < key->n_sz) {
-- req->dst_len = key->n_sz;
-- dev_err(jrdev, "Output buffer length less than parameter n\n");
-- return -EOVERFLOW;
-- }
--
- /* Allocate extended descriptor */
- edesc = rsa_edesc_alloc(req, DESC_RSA_PRIV_F1_LEN);
- if (IS_ERR(edesc))
-@@ -344,17 +576,147 @@ init_fail:
- return ret;
- }
-
-+static int caam_rsa_dec_priv_f2(struct akcipher_request *req)
-+{
-+ struct crypto_akcipher *tfm = crypto_akcipher_reqtfm(req);
-+ struct caam_rsa_ctx *ctx = akcipher_tfm_ctx(tfm);
-+ struct device *jrdev = ctx->dev;
-+ struct rsa_edesc *edesc;
-+ int ret;
-+
-+ /* Allocate extended descriptor */
-+ edesc = rsa_edesc_alloc(req, DESC_RSA_PRIV_F2_LEN);
-+ if (IS_ERR(edesc))
-+ return PTR_ERR(edesc);
-+
-+ /* Set RSA Decrypt Protocol Data Block - Private Key Form #2 */
-+ ret = set_rsa_priv_f2_pdb(req, edesc);
-+ if (ret)
-+ goto init_fail;
-+
-+ /* Initialize Job Descriptor */
-+ init_rsa_priv_f2_desc(edesc->hw_desc, &edesc->pdb.priv_f2);
-+
-+ ret = caam_jr_enqueue(jrdev, edesc->hw_desc, rsa_priv_f2_done, req);
-+ if (!ret)
-+ return -EINPROGRESS;
-+
-+ rsa_priv_f2_unmap(jrdev, edesc, req);
-+
-+init_fail:
-+ rsa_io_unmap(jrdev, edesc, req);
-+ kfree(edesc);
-+ return ret;
-+}
-+
-+static int caam_rsa_dec_priv_f3(struct akcipher_request *req)
-+{
-+ struct crypto_akcipher *tfm = crypto_akcipher_reqtfm(req);
-+ struct caam_rsa_ctx *ctx = akcipher_tfm_ctx(tfm);
-+ struct device *jrdev = ctx->dev;
-+ struct rsa_edesc *edesc;
-+ int ret;
-+
-+ /* Allocate extended descriptor */
-+ edesc = rsa_edesc_alloc(req, DESC_RSA_PRIV_F3_LEN);
-+ if (IS_ERR(edesc))
-+ return PTR_ERR(edesc);
-+
-+ /* Set RSA Decrypt Protocol Data Block - Private Key Form #3 */
-+ ret = set_rsa_priv_f3_pdb(req, edesc);
-+ if (ret)
-+ goto init_fail;
-+
-+ /* Initialize Job Descriptor */
-+ init_rsa_priv_f3_desc(edesc->hw_desc, &edesc->pdb.priv_f3);
-+
-+ ret = caam_jr_enqueue(jrdev, edesc->hw_desc, rsa_priv_f3_done, req);
-+ if (!ret)
-+ return -EINPROGRESS;
-+
-+ rsa_priv_f3_unmap(jrdev, edesc, req);
-+
-+init_fail:
-+ rsa_io_unmap(jrdev, edesc, req);
-+ kfree(edesc);
-+ return ret;
-+}
-+
-+static int caam_rsa_dec(struct akcipher_request *req)
-+{
-+ struct crypto_akcipher *tfm = crypto_akcipher_reqtfm(req);
-+ struct caam_rsa_ctx *ctx = akcipher_tfm_ctx(tfm);
-+ struct caam_rsa_key *key = &ctx->key;
-+ int ret;
-+
-+ if (unlikely(!key->n || !key->d))
-+ return -EINVAL;
-+
-+ if (req->dst_len < key->n_sz) {
-+ req->dst_len = key->n_sz;
-+ dev_err(ctx->dev, "Output buffer length less than parameter n\n");
-+ return -EOVERFLOW;
-+ }
-+
-+ if (key->priv_form == FORM3)
-+ ret = caam_rsa_dec_priv_f3(req);
-+ else if (key->priv_form == FORM2)
-+ ret = caam_rsa_dec_priv_f2(req);
-+ else
-+ ret = caam_rsa_dec_priv_f1(req);
-+
-+ return ret;
-+}
-+
- static void caam_rsa_free_key(struct caam_rsa_key *key)
- {
- kzfree(key->d);
-+ kzfree(key->p);
-+ kzfree(key->q);
-+ kzfree(key->dp);
-+ kzfree(key->dq);
-+ kzfree(key->qinv);
-+ kzfree(key->tmp1);
-+ kzfree(key->tmp2);
- kfree(key->e);
- kfree(key->n);
-- key->d = NULL;
-- key->e = NULL;
-- key->n = NULL;
-- key->d_sz = 0;
-- key->e_sz = 0;
-- key->n_sz = 0;
-+ memset(key, 0, sizeof(*key));
-+}
-+
-+static void caam_rsa_drop_leading_zeros(const u8 **ptr, size_t *nbytes)
-+{
-+ while (!**ptr && *nbytes) {
-+ (*ptr)++;
-+ (*nbytes)--;
-+ }
-+}
-+
-+/**
-+ * caam_read_rsa_crt - Used for reading dP, dQ, qInv CRT members.
-+ * dP, dQ and qInv could decode to less than corresponding p, q length, as the
-+ * BER-encoding requires that the minimum number of bytes be used to encode the
-+ * integer. dP, dQ, qInv decoded values have to be zero-padded to appropriate
-+ * length.
-+ *
-+ * @ptr : pointer to {dP, dQ, qInv} CRT member
-+ * @nbytes: length in bytes of {dP, dQ, qInv} CRT member
-+ * @dstlen: length in bytes of corresponding p or q prime factor
-+ */
-+static u8 *caam_read_rsa_crt(const u8 *ptr, size_t nbytes, size_t dstlen)
-+{
-+ u8 *dst;
-+
-+ caam_rsa_drop_leading_zeros(&ptr, &nbytes);
-+ if (!nbytes)
-+ return NULL;
-+
-+ dst = kzalloc(dstlen, GFP_DMA | GFP_KERNEL);
-+ if (!dst)
-+ return NULL;
-+
-+ memcpy(dst + (dstlen - nbytes), ptr, nbytes);
-+
-+ return dst;
- }
-
- /**
-@@ -370,10 +732,9 @@ static inline u8 *caam_read_raw_data(con
- {
- u8 *val;
-
-- while (!*buf && *nbytes) {
-- buf++;
-- (*nbytes)--;
-- }
-+ caam_rsa_drop_leading_zeros(&buf, nbytes);
-+ if (!*nbytes)
-+ return NULL;
-
- val = kzalloc(*nbytes, GFP_DMA | GFP_KERNEL);
- if (!val)
-@@ -395,7 +756,7 @@ static int caam_rsa_set_pub_key(struct c
- unsigned int keylen)
- {
- struct caam_rsa_ctx *ctx = akcipher_tfm_ctx(tfm);
-- struct rsa_key raw_key = {0};
-+ struct rsa_key raw_key = {NULL};
- struct caam_rsa_key *rsa_key = &ctx->key;
- int ret;
-
-@@ -437,11 +798,69 @@ err:
- return -ENOMEM;
- }
-
-+static void caam_rsa_set_priv_key_form(struct caam_rsa_ctx *ctx,
-+ struct rsa_key *raw_key)
-+{
-+ struct caam_rsa_key *rsa_key = &ctx->key;
-+ size_t p_sz = raw_key->p_sz;
-+ size_t q_sz = raw_key->q_sz;
-+
-+ rsa_key->p = caam_read_raw_data(raw_key->p, &p_sz);
-+ if (!rsa_key->p)
-+ return;
-+ rsa_key->p_sz = p_sz;
-+
-+ rsa_key->q = caam_read_raw_data(raw_key->q, &q_sz);
-+ if (!rsa_key->q)
-+ goto free_p;
-+ rsa_key->q_sz = q_sz;
-+
-+ rsa_key->tmp1 = kzalloc(raw_key->p_sz, GFP_DMA | GFP_KERNEL);
-+ if (!rsa_key->tmp1)
-+ goto free_q;
-+
-+ rsa_key->tmp2 = kzalloc(raw_key->q_sz, GFP_DMA | GFP_KERNEL);
-+ if (!rsa_key->tmp2)
-+ goto free_tmp1;
-+
-+ rsa_key->priv_form = FORM2;
-+
-+ rsa_key->dp = caam_read_rsa_crt(raw_key->dp, raw_key->dp_sz, p_sz);
-+ if (!rsa_key->dp)
-+ goto free_tmp2;
-+
-+ rsa_key->dq = caam_read_rsa_crt(raw_key->dq, raw_key->dq_sz, q_sz);
-+ if (!rsa_key->dq)
-+ goto free_dp;
-+
-+ rsa_key->qinv = caam_read_rsa_crt(raw_key->qinv, raw_key->qinv_sz,
-+ q_sz);
-+ if (!rsa_key->qinv)
-+ goto free_dq;
-+
-+ rsa_key->priv_form = FORM3;
-+
-+ return;
-+
-+free_dq:
-+ kzfree(rsa_key->dq);
-+free_dp:
-+ kzfree(rsa_key->dp);
-+free_tmp2:
-+ kzfree(rsa_key->tmp2);
-+free_tmp1:
-+ kzfree(rsa_key->tmp1);
-+free_q:
-+ kzfree(rsa_key->q);
-+free_p:
-+ kzfree(rsa_key->p);
-+}
-+
- static int caam_rsa_set_priv_key(struct crypto_akcipher *tfm, const void *key,
- unsigned int keylen)
- {
- struct caam_rsa_ctx *ctx = akcipher_tfm_ctx(tfm);
-- struct rsa_key raw_key = {0};
-+ struct rsa_key raw_key = {NULL};
- struct caam_rsa_key *rsa_key = &ctx->key;
- int ret;
-
-@@ -483,6 +902,8 @@ static int caam_rsa_set_priv_key(struct
- memcpy(rsa_key->d, raw_key.d, raw_key.d_sz);
- memcpy(rsa_key->e, raw_key.e, raw_key.e_sz);
-
-+ caam_rsa_set_priv_key_form(ctx, &raw_key);
-+
- return 0;
-
- err:
---- a/drivers/crypto/caam/caampkc.h
-+++ b/drivers/crypto/caam/caampkc.h
-@@ -13,21 +13,75 @@
- #include "pdb.h"
-
- /**
-+ * caam_priv_key_form - CAAM RSA private key representation
-+ * CAAM RSA private key may have either of three forms.
-+ *
-+ * 1. The first representation consists of the pair (n, d), where the
-+ * components have the following meanings:
-+ * n the RSA modulus
-+ * d the RSA private exponent
-+ *
-+ * 2. The second representation consists of the triplet (p, q, d), where the
-+ * components have the following meanings:
-+ * p the first prime factor of the RSA modulus n
-+ * q the second prime factor of the RSA modulus n
-+ * d the RSA private exponent
-+ *
-+ * 3. The third representation consists of the quintuple (p, q, dP, dQ, qInv),
-+ * where the components have the following meanings:
-+ * p the first prime factor of the RSA modulus n
-+ * q the second prime factor of the RSA modulus n
-+ * dP the first factors's CRT exponent
-+ * dQ the second factors's CRT exponent
-+ * qInv the (first) CRT coefficient
-+ *
-+ * The benefit of using the third or the second key form is lower computational
-+ * cost for the decryption and signature operations.
-+ */
-+enum caam_priv_key_form {
-+ FORM1,
-+ FORM2,
-+ FORM3
-+};
-+
-+/**
- * caam_rsa_key - CAAM RSA key structure. Keys are allocated in DMA zone.
- * @n : RSA modulus raw byte stream
- * @e : RSA public exponent raw byte stream
- * @d : RSA private exponent raw byte stream
-+ * @p : RSA prime factor p of RSA modulus n
-+ * @q : RSA prime factor q of RSA modulus n
-+ * @dp : RSA CRT exponent of p
-+ * @dp : RSA CRT exponent of q
-+ * @qinv : RSA CRT coefficient
-+ * @tmp1 : CAAM uses this temporary buffer as internal state buffer.
-+ * It is assumed to be as long as p.
-+ * @tmp2 : CAAM uses this temporary buffer as internal state buffer.
-+ * It is assumed to be as long as q.
- * @n_sz : length in bytes of RSA modulus n
- * @e_sz : length in bytes of RSA public exponent
- * @d_sz : length in bytes of RSA private exponent
-+ * @p_sz : length in bytes of RSA prime factor p of RSA modulus n
-+ * @q_sz : length in bytes of RSA prime factor q of RSA modulus n
-+ * @priv_form : CAAM RSA private key representation
- */
- struct caam_rsa_key {
- u8 *n;
- u8 *e;
- u8 *d;
-+ u8 *p;
-+ u8 *q;
-+ u8 *dp;
-+ u8 *dq;
-+ u8 *qinv;
-+ u8 *tmp1;
-+ u8 *tmp2;
- size_t n_sz;
- size_t e_sz;
- size_t d_sz;
-+ size_t p_sz;
-+ size_t q_sz;
-+ enum caam_priv_key_form priv_form;
- };
-
- /**
-@@ -59,6 +113,8 @@ struct rsa_edesc {
- union {
- struct rsa_pub_pdb pub;
- struct rsa_priv_f1_pdb priv_f1;
-+ struct rsa_priv_f2_pdb priv_f2;
-+ struct rsa_priv_f3_pdb priv_f3;
- } pdb;
- u32 hw_desc[];
- };
-@@ -66,5 +122,7 @@ struct rsa_edesc {
- /* Descriptor construction primitives. */
- void init_rsa_pub_desc(u32 *desc, struct rsa_pub_pdb *pdb);
- void init_rsa_priv_f1_desc(u32 *desc, struct rsa_priv_f1_pdb *pdb);
-+void init_rsa_priv_f2_desc(u32 *desc, struct rsa_priv_f2_pdb *pdb);
-+void init_rsa_priv_f3_desc(u32 *desc, struct rsa_priv_f3_pdb *pdb);
-
- #endif
---- a/drivers/crypto/caam/caamrng.c
-+++ b/drivers/crypto/caam/caamrng.c
-@@ -52,7 +52,7 @@
-
- /* length of descriptors */
- #define DESC_JOB_O_LEN (CAAM_CMD_SZ * 2 + CAAM_PTR_SZ * 2)
--#define DESC_RNG_LEN (4 * CAAM_CMD_SZ)
-+#define DESC_RNG_LEN (3 * CAAM_CMD_SZ)
-
- /* Buffer, its dma address and lock */
- struct buf_data {
-@@ -100,8 +100,7 @@ static void rng_done(struct device *jrde
- {
- struct buf_data *bd;
-
-- bd = (struct buf_data *)((char *)desc -
-- offsetof(struct buf_data, hw_desc));
-+ bd = container_of(desc, struct buf_data, hw_desc[0]);
-
- if (err)
- caam_jr_strstatus(jrdev, err);
-@@ -196,9 +195,6 @@ static inline int rng_create_sh_desc(str
-
- init_sh_desc(desc, HDR_SHARE_SERIAL);
-
-- /* Propagate errors from shared to job descriptor */
-- append_cmd(desc, SET_OK_NO_PROP_ERRORS | CMD_LOAD);
--
- /* Generate random bytes */
- append_operation(desc, OP_ALG_ALGSEL_RNG | OP_TYPE_CLASS1_ALG);
-
-@@ -289,11 +285,7 @@ static int caam_init_rng(struct caam_rng
- if (err)
- return err;
-
-- err = caam_init_buf(ctx, 1);
-- if (err)
-- return err;
--
-- return 0;
-+ return caam_init_buf(ctx, 1);
- }
-
- static struct hwrng caam_rng = {
-@@ -351,7 +343,7 @@ static int __init caam_rng_init(void)
- pr_err("Job Ring Device allocation for transform failed\n");
- return PTR_ERR(dev);
- }
-- rng_ctx = kmalloc(sizeof(*rng_ctx), GFP_DMA);
-+ rng_ctx = kmalloc(sizeof(*rng_ctx), GFP_DMA | GFP_KERNEL);
- if (!rng_ctx) {
- err = -ENOMEM;
- goto free_caam_alloc;
---- a/drivers/crypto/caam/compat.h
-+++ b/drivers/crypto/caam/compat.h
-@@ -16,6 +16,7 @@
- #include <linux/of_platform.h>
- #include <linux/dma-mapping.h>
- #include <linux/io.h>
-+#include <linux/iommu.h>
- #include <linux/spinlock.h>
- #include <linux/rtnetlink.h>
- #include <linux/in.h>
---- a/drivers/crypto/caam/ctrl.c
-+++ b/drivers/crypto/caam/ctrl.c
-@@ -2,40 +2,41 @@
- * Controller-level driver, kernel property detection, initialization
- *
- * Copyright 2008-2012 Freescale Semiconductor, Inc.
-+ * Copyright 2017 NXP
- */
-
- #include <linux/device.h>
- #include <linux/of_address.h>
- #include <linux/of_irq.h>
-+#include <linux/sys_soc.h>
-
- #include "compat.h"
- #include "regs.h"
- #include "intern.h"
- #include "jr.h"
- #include "desc_constr.h"
--#include "error.h"
- #include "ctrl.h"
-
- bool caam_little_end;
- EXPORT_SYMBOL(caam_little_end);
-+bool caam_imx;
-+EXPORT_SYMBOL(caam_imx);
-+bool caam_dpaa2;
-+EXPORT_SYMBOL(caam_dpaa2);
-+
-+#ifdef CONFIG_CAAM_QI
-+#include "qi.h"
-+#endif
-
- /*
- * i.MX targets tend to have clock control subsystems that can
- * enable/disable clocking to our device.
- */
--#ifdef CONFIG_CRYPTO_DEV_FSL_CAAM_IMX
--static inline struct clk *caam_drv_identify_clk(struct device *dev,
-- char *clk_name)
--{
-- return devm_clk_get(dev, clk_name);
--}
--#else
- static inline struct clk *caam_drv_identify_clk(struct device *dev,
- char *clk_name)
- {
-- return NULL;
-+ return caam_imx ? devm_clk_get(dev, clk_name) : NULL;
- }
--#endif
-
- /*
- * Descriptor to instantiate RNG State Handle 0 in normal mode and
-@@ -274,7 +275,7 @@ static int deinstantiate_rng(struct devi
- /*
- * If the corresponding bit is set, then it means the state
- * handle was initialized by us, and thus it needs to be
-- * deintialized as well
-+ * deinitialized as well
- */
- if ((1 << sh_idx) & state_handle_mask) {
- /*
-@@ -307,20 +308,24 @@ static int caam_remove(struct platform_d
- struct device *ctrldev;
- struct caam_drv_private *ctrlpriv;
- struct caam_ctrl __iomem *ctrl;
-- int ring;
-
- ctrldev = &pdev->dev;
- ctrlpriv = dev_get_drvdata(ctrldev);
- ctrl = (struct caam_ctrl __iomem *)ctrlpriv->ctrl;
-
-- /* Remove platform devices for JobRs */
-- for (ring = 0; ring < ctrlpriv->total_jobrs; ring++) {
-- if (ctrlpriv->jrpdev[ring])
-- of_device_unregister(ctrlpriv->jrpdev[ring]);
-- }
-+ /* Remove platform devices under the crypto node */
-+ of_platform_depopulate(ctrldev);
-+
-+#ifdef CONFIG_CAAM_QI
-+ if (ctrlpriv->qidev)
-+ caam_qi_shutdown(ctrlpriv->qidev);
-+#endif
-
-- /* De-initialize RNG state handles initialized by this driver. */
-- if (ctrlpriv->rng4_sh_init)
-+ /*
-+ * De-initialize RNG state handles initialized by this driver.
-+ * In case of DPAA 2.x, RNG is managed by MC firmware.
-+ */
-+ if (!caam_dpaa2 && ctrlpriv->rng4_sh_init)
- deinstantiate_rng(ctrldev, ctrlpriv->rng4_sh_init);
-
- /* Shut down debug views */
-@@ -335,8 +340,8 @@ static int caam_remove(struct platform_d
- clk_disable_unprepare(ctrlpriv->caam_ipg);
- clk_disable_unprepare(ctrlpriv->caam_mem);
- clk_disable_unprepare(ctrlpriv->caam_aclk);
-- clk_disable_unprepare(ctrlpriv->caam_emi_slow);
--
-+ if (ctrlpriv->caam_emi_slow)
-+ clk_disable_unprepare(ctrlpriv->caam_emi_slow);
- return 0;
- }
-
-@@ -370,11 +375,8 @@ static void kick_trng(struct platform_de
- */
- val = (rd_reg32(&r4tst->rtsdctl) & RTSDCTL_ENT_DLY_MASK)
- >> RTSDCTL_ENT_DLY_SHIFT;
-- if (ent_delay <= val) {
-- /* put RNG4 into run mode */
-- clrsetbits_32(&r4tst->rtmctl, RTMCTL_PRGM, 0);
-- return;
-- }
-+ if (ent_delay <= val)
-+ goto start_rng;
-
- val = rd_reg32(&r4tst->rtsdctl);
- val = (val & ~RTSDCTL_ENT_DLY_MASK) |
-@@ -386,15 +388,12 @@ static void kick_trng(struct platform_de
- wr_reg32(&r4tst->rtfrqmax, RTFRQMAX_DISABLE);
- /* read the control register */
- val = rd_reg32(&r4tst->rtmctl);
-+start_rng:
- /*
- * select raw sampling in both entropy shifter
-- * and statistical checker
-+ * and statistical checker; ; put RNG4 into run mode
- */
-- clrsetbits_32(&val, 0, RTMCTL_SAMP_MODE_RAW_ES_SC);
-- /* put RNG4 into run mode */
-- clrsetbits_32(&val, RTMCTL_PRGM, 0);
-- /* write back the control register */
-- wr_reg32(&r4tst->rtmctl, val);
-+ clrsetbits_32(&r4tst->rtmctl, RTMCTL_PRGM, RTMCTL_SAMP_MODE_RAW_ES_SC);
- }
-
- /**
-@@ -415,28 +414,26 @@ int caam_get_era(void)
- }
- EXPORT_SYMBOL(caam_get_era);
-
--#ifdef CONFIG_DEBUG_FS
--static int caam_debugfs_u64_get(void *data, u64 *val)
--{
-- *val = caam64_to_cpu(*(u64 *)data);
-- return 0;
--}
--
--static int caam_debugfs_u32_get(void *data, u64 *val)
--{
-- *val = caam32_to_cpu(*(u32 *)data);
-- return 0;
--}
--
--DEFINE_SIMPLE_ATTRIBUTE(caam_fops_u32_ro, caam_debugfs_u32_get, NULL, "%llu\n");
--DEFINE_SIMPLE_ATTRIBUTE(caam_fops_u64_ro, caam_debugfs_u64_get, NULL, "%llu\n");
--#endif
-+static const struct of_device_id caam_match[] = {
-+ {
-+ .compatible = "fsl,sec-v4.0",
-+ },
-+ {
-+ .compatible = "fsl,sec4.0",
-+ },
-+ {},
-+};
-+MODULE_DEVICE_TABLE(of, caam_match);
-
- /* Probe routine for CAAM top (controller) level */
- static int caam_probe(struct platform_device *pdev)
- {
-- int ret, ring, rspec, gen_sk, ent_delay = RTSDCTL_ENT_DLY_MIN;
-+ int ret, ring, gen_sk, ent_delay = RTSDCTL_ENT_DLY_MIN;
- u64 caam_id;
-+ static const struct soc_device_attribute imx_soc[] = {
-+ {.family = "Freescale i.MX"},
-+ {},
-+ };
- struct device *dev;
- struct device_node *nprop, *np;
- struct caam_ctrl __iomem *ctrl;
-@@ -456,9 +453,10 @@ static int caam_probe(struct platform_de
-
- dev = &pdev->dev;
- dev_set_drvdata(dev, ctrlpriv);
-- ctrlpriv->pdev = pdev;
- nprop = pdev->dev.of_node;
-
-+ caam_imx = (bool)soc_device_match(imx_soc);
-+
- /* Enable clocking */
- clk = caam_drv_identify_clk(&pdev->dev, "ipg");
- if (IS_ERR(clk)) {
-@@ -487,14 +485,16 @@ static int caam_probe(struct platform_de
- }
- ctrlpriv->caam_aclk = clk;
-
-- clk = caam_drv_identify_clk(&pdev->dev, "emi_slow");
-- if (IS_ERR(clk)) {
-- ret = PTR_ERR(clk);
-- dev_err(&pdev->dev,
-- "can't identify CAAM emi_slow clk: %d\n", ret);
-- return ret;
-+ if (!of_machine_is_compatible("fsl,imx6ul")) {
-+ clk = caam_drv_identify_clk(&pdev->dev, "emi_slow");
-+ if (IS_ERR(clk)) {
-+ ret = PTR_ERR(clk);
-+ dev_err(&pdev->dev,
-+ "can't identify CAAM emi_slow clk: %d\n", ret);
-+ return ret;
-+ }
-+ ctrlpriv->caam_emi_slow = clk;
- }
-- ctrlpriv->caam_emi_slow = clk;
-
- ret = clk_prepare_enable(ctrlpriv->caam_ipg);
- if (ret < 0) {
-@@ -515,11 +515,13 @@ static int caam_probe(struct platform_de
- goto disable_caam_mem;
- }
-
-- ret = clk_prepare_enable(ctrlpriv->caam_emi_slow);
-- if (ret < 0) {
-- dev_err(&pdev->dev, "can't enable CAAM emi slow clock: %d\n",
-- ret);
-- goto disable_caam_aclk;
-+ if (ctrlpriv->caam_emi_slow) {
-+ ret = clk_prepare_enable(ctrlpriv->caam_emi_slow);
-+ if (ret < 0) {
-+ dev_err(&pdev->dev, "can't enable CAAM emi slow clock: %d\n",
-+ ret);
-+ goto disable_caam_aclk;
-+ }
- }
-
- /* Get configuration properties from device tree */
-@@ -546,13 +548,13 @@ static int caam_probe(struct platform_de
- else
- BLOCK_OFFSET = PG_SIZE_64K;
-
-- ctrlpriv->ctrl = (struct caam_ctrl __force *)ctrl;
-- ctrlpriv->assure = (struct caam_assurance __force *)
-- ((uint8_t *)ctrl +
-+ ctrlpriv->ctrl = (struct caam_ctrl __iomem __force *)ctrl;
-+ ctrlpriv->assure = (struct caam_assurance __iomem __force *)
-+ ((__force uint8_t *)ctrl +
- BLOCK_OFFSET * ASSURE_BLOCK_NUMBER
- );
-- ctrlpriv->deco = (struct caam_deco __force *)
-- ((uint8_t *)ctrl +
-+ ctrlpriv->deco = (struct caam_deco __iomem __force *)
-+ ((__force uint8_t *)ctrl +
- BLOCK_OFFSET * DECO_BLOCK_NUMBER
- );
-
-@@ -561,12 +563,17 @@ static int caam_probe(struct platform_de
-
- /*
- * Enable DECO watchdogs and, if this is a PHYS_ADDR_T_64BIT kernel,
-- * long pointers in master configuration register
-+ * long pointers in master configuration register.
-+ * In case of DPAA 2.x, Management Complex firmware performs
-+ * the configuration.
- */
-- clrsetbits_32(&ctrl->mcr, MCFGR_AWCACHE_MASK | MCFGR_LONG_PTR,
-- MCFGR_AWCACHE_CACH | MCFGR_AWCACHE_BUFF |
-- MCFGR_WDENABLE | MCFGR_LARGE_BURST |
-- (sizeof(dma_addr_t) == sizeof(u64) ? MCFGR_LONG_PTR : 0));
-+ caam_dpaa2 = !!(comp_params & CTPR_MS_DPAA2);
-+ if (!caam_dpaa2)
-+ clrsetbits_32(&ctrl->mcr, MCFGR_AWCACHE_MASK | MCFGR_LONG_PTR,
-+ MCFGR_AWCACHE_CACH | MCFGR_AWCACHE_BUFF |
-+ MCFGR_WDENABLE | MCFGR_LARGE_BURST |
-+ (sizeof(dma_addr_t) == sizeof(u64) ?
-+ MCFGR_LONG_PTR : 0));
-
- /*
- * Read the Compile Time paramters and SCFGR to determine
-@@ -594,64 +601,69 @@ static int caam_probe(struct platform_de
- JRSTART_JR1_START | JRSTART_JR2_START |
- JRSTART_JR3_START);
-
-- if (sizeof(dma_addr_t) == sizeof(u64))
-- if (of_device_is_compatible(nprop, "fsl,sec-v5.0"))
-- dma_set_mask_and_coherent(dev, DMA_BIT_MASK(40));
-+ if (sizeof(dma_addr_t) == sizeof(u64)) {
-+ if (caam_dpaa2)
-+ ret = dma_set_mask_and_coherent(dev, DMA_BIT_MASK(49));
-+ else if (of_device_is_compatible(nprop, "fsl,sec-v5.0"))
-+ ret = dma_set_mask_and_coherent(dev, DMA_BIT_MASK(40));
- else
-- dma_set_mask_and_coherent(dev, DMA_BIT_MASK(36));
-- else
-- dma_set_mask_and_coherent(dev, DMA_BIT_MASK(32));
-+ ret = dma_set_mask_and_coherent(dev, DMA_BIT_MASK(36));
-+ } else {
-+ ret = dma_set_mask_and_coherent(dev, DMA_BIT_MASK(32));
-+ }
-+ if (ret) {
-+ dev_err(dev, "dma_set_mask_and_coherent failed (%d)\n", ret);
-+ goto iounmap_ctrl;
-+ }
-
-- /*
-- * Detect and enable JobRs
-- * First, find out how many ring spec'ed, allocate references
-- * for all, then go probe each one.
-- */
-- rspec = 0;
-- for_each_available_child_of_node(nprop, np)
-- if (of_device_is_compatible(np, "fsl,sec-v4.0-job-ring") ||
-- of_device_is_compatible(np, "fsl,sec4.0-job-ring"))
-- rspec++;
-+ ctrlpriv->era = caam_get_era();
-
-- ctrlpriv->jrpdev = devm_kcalloc(&pdev->dev, rspec,
-- sizeof(*ctrlpriv->jrpdev), GFP_KERNEL);
-- if (ctrlpriv->jrpdev == NULL) {
-- ret = -ENOMEM;
-+ ret = of_platform_populate(nprop, caam_match, NULL, dev);
-+ if (ret) {
-+ dev_err(dev, "JR platform devices creation error\n");
- goto iounmap_ctrl;
- }
-
-+#ifdef CONFIG_DEBUG_FS
-+ /*
-+ * FIXME: needs better naming distinction, as some amalgamation of
-+ * "caam" and nprop->full_name. The OF name isn't distinctive,
-+ * but does separate instances
-+ */
-+ perfmon = (struct caam_perfmon __force *)&ctrl->perfmon;
-+
-+ ctrlpriv->dfs_root = debugfs_create_dir(dev_name(dev), NULL);
-+ ctrlpriv->ctl = debugfs_create_dir("ctl", ctrlpriv->dfs_root);
-+#endif
- ring = 0;
-- ctrlpriv->total_jobrs = 0;
- for_each_available_child_of_node(nprop, np)
- if (of_device_is_compatible(np, "fsl,sec-v4.0-job-ring") ||
- of_device_is_compatible(np, "fsl,sec4.0-job-ring")) {
-- ctrlpriv->jrpdev[ring] =
-- of_platform_device_create(np, NULL, dev);
-- if (!ctrlpriv->jrpdev[ring]) {
-- pr_warn("JR%d Platform device creation error\n",
-- ring);
-- continue;
-- }
-- ctrlpriv->jr[ring] = (struct caam_job_ring __force *)
-- ((uint8_t *)ctrl +
-+ ctrlpriv->jr[ring] = (struct caam_job_ring __iomem __force *)
-+ ((__force uint8_t *)ctrl +
- (ring + JR_BLOCK_NUMBER) *
- BLOCK_OFFSET
- );
- ctrlpriv->total_jobrs++;
- ring++;
-- }
-+ }
-
-- /* Check to see if QI present. If so, enable */
-- ctrlpriv->qi_present =
-- !!(rd_reg32(&ctrl->perfmon.comp_parms_ms) &
-- CTPR_MS_QI_MASK);
-- if (ctrlpriv->qi_present) {
-- ctrlpriv->qi = (struct caam_queue_if __force *)
-- ((uint8_t *)ctrl +
-+ /* Check to see if (DPAA 1.x) QI present. If so, enable */
-+ ctrlpriv->qi_present = !!(comp_params & CTPR_MS_QI_MASK);
-+ if (ctrlpriv->qi_present && !caam_dpaa2) {
-+ ctrlpriv->qi = (struct caam_queue_if __iomem __force *)
-+ ((__force uint8_t *)ctrl +
- BLOCK_OFFSET * QI_BLOCK_NUMBER
- );
- /* This is all that's required to physically enable QI */
- wr_reg32(&ctrlpriv->qi->qi_control_lo, QICTL_DQEN);
-+
-+ /* If QMAN driver is present, init CAAM-QI backend */
-+#ifdef CONFIG_CAAM_QI
-+ ret = caam_qi_init(pdev);
-+ if (ret)
-+ dev_err(dev, "caam qi i/f init failed: %d\n", ret);
-+#endif
- }
-
- /* If no QI and no rings specified, quit and go home */
-@@ -666,8 +678,10 @@ static int caam_probe(struct platform_de
- /*
- * If SEC has RNG version >= 4 and RNG state handle has not been
- * already instantiated, do RNG instantiation
-+ * In case of DPAA 2.x, RNG is managed by MC firmware.
- */
-- if ((cha_vid_ls & CHA_ID_LS_RNG_MASK) >> CHA_ID_LS_RNG_SHIFT >= 4) {
-+ if (!caam_dpaa2 &&
-+ (cha_vid_ls & CHA_ID_LS_RNG_MASK) >> CHA_ID_LS_RNG_SHIFT >= 4) {
- ctrlpriv->rng4_sh_init =
- rd_reg32(&ctrl->r4tst[0].rdsta);
- /*
-@@ -734,78 +748,47 @@ static int caam_probe(struct platform_de
-
- /* Report "alive" for developer to see */
- dev_info(dev, "device ID = 0x%016llx (Era %d)\n", caam_id,
-- caam_get_era());
-- dev_info(dev, "job rings = %d, qi = %d\n",
-- ctrlpriv->total_jobrs, ctrlpriv->qi_present);
-+ ctrlpriv->era);
-+ dev_info(dev, "job rings = %d, qi = %d, dpaa2 = %s\n",
-+ ctrlpriv->total_jobrs, ctrlpriv->qi_present,
-+ caam_dpaa2 ? "yes" : "no");
-
- #ifdef CONFIG_DEBUG_FS
-- /*
-- * FIXME: needs better naming distinction, as some amalgamation of
-- * "caam" and nprop->full_name. The OF name isn't distinctive,
-- * but does separate instances
-- */
-- perfmon = (struct caam_perfmon __force *)&ctrl->perfmon;
--
-- ctrlpriv->dfs_root = debugfs_create_dir(dev_name(dev), NULL);
-- ctrlpriv->ctl = debugfs_create_dir("ctl", ctrlpriv->dfs_root);
--
-- /* Controller-level - performance monitor counters */
--
-- ctrlpriv->ctl_rq_dequeued =
-- debugfs_create_file("rq_dequeued",
-- S_IRUSR | S_IRGRP | S_IROTH,
-- ctrlpriv->ctl, &perfmon->req_dequeued,
-- &caam_fops_u64_ro);
-- ctrlpriv->ctl_ob_enc_req =
-- debugfs_create_file("ob_rq_encrypted",
-- S_IRUSR | S_IRGRP | S_IROTH,
-- ctrlpriv->ctl, &perfmon->ob_enc_req,
-- &caam_fops_u64_ro);
-- ctrlpriv->ctl_ib_dec_req =
-- debugfs_create_file("ib_rq_decrypted",
-- S_IRUSR | S_IRGRP | S_IROTH,
-- ctrlpriv->ctl, &perfmon->ib_dec_req,
-- &caam_fops_u64_ro);
-- ctrlpriv->ctl_ob_enc_bytes =
-- debugfs_create_file("ob_bytes_encrypted",
-- S_IRUSR | S_IRGRP | S_IROTH,
-- ctrlpriv->ctl, &perfmon->ob_enc_bytes,
-- &caam_fops_u64_ro);
-- ctrlpriv->ctl_ob_prot_bytes =
-- debugfs_create_file("ob_bytes_protected",
-- S_IRUSR | S_IRGRP | S_IROTH,
-- ctrlpriv->ctl, &perfmon->ob_prot_bytes,
-- &caam_fops_u64_ro);
-- ctrlpriv->ctl_ib_dec_bytes =
-- debugfs_create_file("ib_bytes_decrypted",
-- S_IRUSR | S_IRGRP | S_IROTH,
-- ctrlpriv->ctl, &perfmon->ib_dec_bytes,
-- &caam_fops_u64_ro);
-- ctrlpriv->ctl_ib_valid_bytes =
-- debugfs_create_file("ib_bytes_validated",
-- S_IRUSR | S_IRGRP | S_IROTH,
-- ctrlpriv->ctl, &perfmon->ib_valid_bytes,
-- &caam_fops_u64_ro);
-+ debugfs_create_file("rq_dequeued", S_IRUSR | S_IRGRP | S_IROTH,
-+ ctrlpriv->ctl, &perfmon->req_dequeued,
-+ &caam_fops_u64_ro);
-+ debugfs_create_file("ob_rq_encrypted", S_IRUSR | S_IRGRP | S_IROTH,
-+ ctrlpriv->ctl, &perfmon->ob_enc_req,
-+ &caam_fops_u64_ro);
-+ debugfs_create_file("ib_rq_decrypted", S_IRUSR | S_IRGRP | S_IROTH,
-+ ctrlpriv->ctl, &perfmon->ib_dec_req,
-+ &caam_fops_u64_ro);
-+ debugfs_create_file("ob_bytes_encrypted", S_IRUSR | S_IRGRP | S_IROTH,
-+ ctrlpriv->ctl, &perfmon->ob_enc_bytes,
-+ &caam_fops_u64_ro);
-+ debugfs_create_file("ob_bytes_protected", S_IRUSR | S_IRGRP | S_IROTH,
-+ ctrlpriv->ctl, &perfmon->ob_prot_bytes,
-+ &caam_fops_u64_ro);
-+ debugfs_create_file("ib_bytes_decrypted", S_IRUSR | S_IRGRP | S_IROTH,
-+ ctrlpriv->ctl, &perfmon->ib_dec_bytes,
-+ &caam_fops_u64_ro);
-+ debugfs_create_file("ib_bytes_validated", S_IRUSR | S_IRGRP | S_IROTH,
-+ ctrlpriv->ctl, &perfmon->ib_valid_bytes,
-+ &caam_fops_u64_ro);
-
- /* Controller level - global status values */
-- ctrlpriv->ctl_faultaddr =
-- debugfs_create_file("fault_addr",
-- S_IRUSR | S_IRGRP | S_IROTH,
-- ctrlpriv->ctl, &perfmon->faultaddr,
-- &caam_fops_u32_ro);
-- ctrlpriv->ctl_faultdetail =
-- debugfs_create_file("fault_detail",
-- S_IRUSR | S_IRGRP | S_IROTH,
-- ctrlpriv->ctl, &perfmon->faultdetail,
-- &caam_fops_u32_ro);
-- ctrlpriv->ctl_faultstatus =
-- debugfs_create_file("fault_status",
-- S_IRUSR | S_IRGRP | S_IROTH,
-- ctrlpriv->ctl, &perfmon->status,
-- &caam_fops_u32_ro);
-+ debugfs_create_file("fault_addr", S_IRUSR | S_IRGRP | S_IROTH,
-+ ctrlpriv->ctl, &perfmon->faultaddr,
-+ &caam_fops_u32_ro);
-+ debugfs_create_file("fault_detail", S_IRUSR | S_IRGRP | S_IROTH,
-+ ctrlpriv->ctl, &perfmon->faultdetail,
-+ &caam_fops_u32_ro);
-+ debugfs_create_file("fault_status", S_IRUSR | S_IRGRP | S_IROTH,
-+ ctrlpriv->ctl, &perfmon->status,
-+ &caam_fops_u32_ro);
-
- /* Internal covering keys (useful in non-secure mode only) */
-- ctrlpriv->ctl_kek_wrap.data = &ctrlpriv->ctrl->kek[0];
-+ ctrlpriv->ctl_kek_wrap.data = (__force void *)&ctrlpriv->ctrl->kek[0];
- ctrlpriv->ctl_kek_wrap.size = KEK_KEY_SIZE * sizeof(u32);
- ctrlpriv->ctl_kek = debugfs_create_blob("kek",
- S_IRUSR |
-@@ -813,7 +796,7 @@ static int caam_probe(struct platform_de
- ctrlpriv->ctl,
- &ctrlpriv->ctl_kek_wrap);
-
-- ctrlpriv->ctl_tkek_wrap.data = &ctrlpriv->ctrl->tkek[0];
-+ ctrlpriv->ctl_tkek_wrap.data = (__force void *)&ctrlpriv->ctrl->tkek[0];
- ctrlpriv->ctl_tkek_wrap.size = KEK_KEY_SIZE * sizeof(u32);
- ctrlpriv->ctl_tkek = debugfs_create_blob("tkek",
- S_IRUSR |
-@@ -821,7 +804,7 @@ static int caam_probe(struct platform_de
- ctrlpriv->ctl,
- &ctrlpriv->ctl_tkek_wrap);
-
-- ctrlpriv->ctl_tdsk_wrap.data = &ctrlpriv->ctrl->tdsk[0];
-+ ctrlpriv->ctl_tdsk_wrap.data = (__force void *)&ctrlpriv->ctrl->tdsk[0];
- ctrlpriv->ctl_tdsk_wrap.size = KEK_KEY_SIZE * sizeof(u32);
- ctrlpriv->ctl_tdsk = debugfs_create_blob("tdsk",
- S_IRUSR |
-@@ -832,13 +815,17 @@ static int caam_probe(struct platform_de
- return 0;
-
- caam_remove:
-+#ifdef CONFIG_DEBUG_FS
-+ debugfs_remove_recursive(ctrlpriv->dfs_root);
-+#endif
- caam_remove(pdev);
- return ret;
-
- iounmap_ctrl:
- iounmap(ctrl);
- disable_caam_emi_slow:
-- clk_disable_unprepare(ctrlpriv->caam_emi_slow);
-+ if (ctrlpriv->caam_emi_slow)
-+ clk_disable_unprepare(ctrlpriv->caam_emi_slow);
- disable_caam_aclk:
- clk_disable_unprepare(ctrlpriv->caam_aclk);
- disable_caam_mem:
-@@ -848,17 +835,6 @@ disable_caam_ipg:
- return ret;
- }
-
--static struct of_device_id caam_match[] = {
-- {
-- .compatible = "fsl,sec-v4.0",
-- },
-- {
-- .compatible = "fsl,sec4.0",
-- },
-- {},
--};
--MODULE_DEVICE_TABLE(of, caam_match);
--
- static struct platform_driver caam_driver = {
- .driver = {
- .name = "caam",
---- a/drivers/crypto/caam/ctrl.h
-+++ b/drivers/crypto/caam/ctrl.h
-@@ -10,4 +10,6 @@
- /* Prototypes for backend-level services exposed to APIs */
- int caam_get_era(void);
-
-+extern bool caam_dpaa2;
-+
- #endif /* CTRL_H */
---- a/drivers/crypto/caam/desc.h
-+++ b/drivers/crypto/caam/desc.h
-@@ -22,12 +22,6 @@
- #define SEC4_SG_LEN_MASK 0x3fffffff /* Excludes EXT and FINAL */
- #define SEC4_SG_OFFSET_MASK 0x00001fff
-
--struct sec4_sg_entry {
-- u64 ptr;
-- u32 len;
-- u32 bpid_offset;
--};
--
- /* Max size of any CAAM descriptor in 32-bit words, inclusive of header */
- #define MAX_CAAM_DESCSIZE 64
-
-@@ -47,6 +41,7 @@ struct sec4_sg_entry {
- #define CMD_SEQ_LOAD (0x03 << CMD_SHIFT)
- #define CMD_FIFO_LOAD (0x04 << CMD_SHIFT)
- #define CMD_SEQ_FIFO_LOAD (0x05 << CMD_SHIFT)
-+#define CMD_MOVEB (0x07 << CMD_SHIFT)
- #define CMD_STORE (0x0a << CMD_SHIFT)
- #define CMD_SEQ_STORE (0x0b << CMD_SHIFT)
- #define CMD_FIFO_STORE (0x0c << CMD_SHIFT)
-@@ -90,8 +85,8 @@ struct sec4_sg_entry {
- #define HDR_ZRO 0x00008000
-
- /* Start Index or SharedDesc Length */
--#define HDR_START_IDX_MASK 0x3f
- #define HDR_START_IDX_SHIFT 16
-+#define HDR_START_IDX_MASK (0x3f << HDR_START_IDX_SHIFT)
-
- /* If shared descriptor header, 6-bit length */
- #define HDR_DESCLEN_SHR_MASK 0x3f
-@@ -121,10 +116,10 @@ struct sec4_sg_entry {
- #define HDR_PROP_DNR 0x00000800
-
- /* JobDesc/SharedDesc share property */
--#define HDR_SD_SHARE_MASK 0x03
- #define HDR_SD_SHARE_SHIFT 8
--#define HDR_JD_SHARE_MASK 0x07
-+#define HDR_SD_SHARE_MASK (0x03 << HDR_SD_SHARE_SHIFT)
- #define HDR_JD_SHARE_SHIFT 8
-+#define HDR_JD_SHARE_MASK (0x07 << HDR_JD_SHARE_SHIFT)
-
- #define HDR_SHARE_NEVER (0x00 << HDR_SD_SHARE_SHIFT)
- #define HDR_SHARE_WAIT (0x01 << HDR_SD_SHARE_SHIFT)
-@@ -235,7 +230,7 @@ struct sec4_sg_entry {
- #define LDST_SRCDST_WORD_DECO_MATH2 (0x0a << LDST_SRCDST_SHIFT)
- #define LDST_SRCDST_WORD_DECO_AAD_SZ (0x0b << LDST_SRCDST_SHIFT)
- #define LDST_SRCDST_WORD_DECO_MATH3 (0x0b << LDST_SRCDST_SHIFT)
--#define LDST_SRCDST_WORD_CLASS1_ICV_SZ (0x0c << LDST_SRCDST_SHIFT)
-+#define LDST_SRCDST_WORD_CLASS1_IV_SZ (0x0c << LDST_SRCDST_SHIFT)
- #define LDST_SRCDST_WORD_ALTDS_CLASS1 (0x0f << LDST_SRCDST_SHIFT)
- #define LDST_SRCDST_WORD_PKHA_A_SZ (0x10 << LDST_SRCDST_SHIFT)
- #define LDST_SRCDST_WORD_PKHA_B_SZ (0x11 << LDST_SRCDST_SHIFT)
-@@ -360,6 +355,7 @@ struct sec4_sg_entry {
- #define FIFOLD_TYPE_PK_N (0x08 << FIFOLD_TYPE_SHIFT)
- #define FIFOLD_TYPE_PK_A (0x0c << FIFOLD_TYPE_SHIFT)
- #define FIFOLD_TYPE_PK_B (0x0d << FIFOLD_TYPE_SHIFT)
-+#define FIFOLD_TYPE_IFIFO (0x0f << FIFOLD_TYPE_SHIFT)
-
- /* Other types. Need to OR in last/flush bits as desired */
- #define FIFOLD_TYPE_MSG_MASK (0x38 << FIFOLD_TYPE_SHIFT)
-@@ -400,7 +396,7 @@ struct sec4_sg_entry {
- #define FIFOST_TYPE_PKHA_N (0x08 << FIFOST_TYPE_SHIFT)
- #define FIFOST_TYPE_PKHA_A (0x0c << FIFOST_TYPE_SHIFT)
- #define FIFOST_TYPE_PKHA_B (0x0d << FIFOST_TYPE_SHIFT)
--#define FIFOST_TYPE_AF_SBOX_JKEK (0x10 << FIFOST_TYPE_SHIFT)
-+#define FIFOST_TYPE_AF_SBOX_JKEK (0x20 << FIFOST_TYPE_SHIFT)
- #define FIFOST_TYPE_AF_SBOX_TKEK (0x21 << FIFOST_TYPE_SHIFT)
- #define FIFOST_TYPE_PKHA_E_JKEK (0x22 << FIFOST_TYPE_SHIFT)
- #define FIFOST_TYPE_PKHA_E_TKEK (0x23 << FIFOST_TYPE_SHIFT)
-@@ -413,6 +409,7 @@ struct sec4_sg_entry {
- #define FIFOST_TYPE_MESSAGE_DATA (0x30 << FIFOST_TYPE_SHIFT)
- #define FIFOST_TYPE_RNGSTORE (0x34 << FIFOST_TYPE_SHIFT)
- #define FIFOST_TYPE_RNGFIFO (0x35 << FIFOST_TYPE_SHIFT)
-+#define FIFOST_TYPE_METADATA (0x3e << FIFOST_TYPE_SHIFT)
- #define FIFOST_TYPE_SKIP (0x3f << FIFOST_TYPE_SHIFT)
-
- /*
-@@ -449,6 +446,18 @@ struct sec4_sg_entry {
- #define OP_PCLID_DSAVERIFY (0x16 << OP_PCLID_SHIFT)
- #define OP_PCLID_RSAENC_PUBKEY (0x18 << OP_PCLID_SHIFT)
- #define OP_PCLID_RSADEC_PRVKEY (0x19 << OP_PCLID_SHIFT)
-+#define OP_PCLID_DKP_MD5 (0x20 << OP_PCLID_SHIFT)
-+#define OP_PCLID_DKP_SHA1 (0x21 << OP_PCLID_SHIFT)
-+#define OP_PCLID_DKP_SHA224 (0x22 << OP_PCLID_SHIFT)
-+#define OP_PCLID_DKP_SHA256 (0x23 << OP_PCLID_SHIFT)
-+#define OP_PCLID_DKP_SHA384 (0x24 << OP_PCLID_SHIFT)
-+#define OP_PCLID_DKP_SHA512 (0x25 << OP_PCLID_SHIFT)
-+#define OP_PCLID_DKP_RIF_MD5 (0x60 << OP_PCLID_SHIFT)
-+#define OP_PCLID_DKP_RIF_SHA1 (0x61 << OP_PCLID_SHIFT)
-+#define OP_PCLID_DKP_RIF_SHA224 (0x62 << OP_PCLID_SHIFT)
-+#define OP_PCLID_DKP_RIF_SHA256 (0x63 << OP_PCLID_SHIFT)
-+#define OP_PCLID_DKP_RIF_SHA384 (0x64 << OP_PCLID_SHIFT)
-+#define OP_PCLID_DKP_RIF_SHA512 (0x65 << OP_PCLID_SHIFT)
-
- /* Assuming OP_TYPE = OP_TYPE_DECAP_PROTOCOL/ENCAP_PROTOCOL */
- #define OP_PCLID_IPSEC (0x01 << OP_PCLID_SHIFT)
-@@ -1098,6 +1107,22 @@ struct sec4_sg_entry {
- /* MacSec protinfos */
- #define OP_PCL_MACSEC 0x0001
-
-+/* Derived Key Protocol (DKP) Protinfo */
-+#define OP_PCL_DKP_SRC_SHIFT 14
-+#define OP_PCL_DKP_SRC_MASK (3 << OP_PCL_DKP_SRC_SHIFT)
-+#define OP_PCL_DKP_SRC_IMM (0 << OP_PCL_DKP_SRC_SHIFT)
-+#define OP_PCL_DKP_SRC_SEQ (1 << OP_PCL_DKP_SRC_SHIFT)
-+#define OP_PCL_DKP_SRC_PTR (2 << OP_PCL_DKP_SRC_SHIFT)
-+#define OP_PCL_DKP_SRC_SGF (3 << OP_PCL_DKP_SRC_SHIFT)
-+#define OP_PCL_DKP_DST_SHIFT 12
-+#define OP_PCL_DKP_DST_MASK (3 << OP_PCL_DKP_DST_SHIFT)
-+#define OP_PCL_DKP_DST_IMM (0 << OP_PCL_DKP_DST_SHIFT)
-+#define OP_PCL_DKP_DST_SEQ (1 << OP_PCL_DKP_DST_SHIFT)
-+#define OP_PCL_DKP_DST_PTR (2 << OP_PCL_DKP_DST_SHIFT)
-+#define OP_PCL_DKP_DST_SGF (3 << OP_PCL_DKP_DST_SHIFT)
-+#define OP_PCL_DKP_KEY_SHIFT 0
-+#define OP_PCL_DKP_KEY_MASK (0xfff << OP_PCL_DKP_KEY_SHIFT)
-+
- /* PKI unidirectional protocol protinfo bits */
- #define OP_PCL_PKPROT_TEST 0x0008
- #define OP_PCL_PKPROT_DECRYPT 0x0004
-@@ -1107,8 +1132,8 @@ struct sec4_sg_entry {
- /* For non-protocol/alg-only op commands */
- #define OP_ALG_TYPE_SHIFT 24
- #define OP_ALG_TYPE_MASK (0x7 << OP_ALG_TYPE_SHIFT)
--#define OP_ALG_TYPE_CLASS1 2
--#define OP_ALG_TYPE_CLASS2 4
-+#define OP_ALG_TYPE_CLASS1 (2 << OP_ALG_TYPE_SHIFT)
-+#define OP_ALG_TYPE_CLASS2 (4 << OP_ALG_TYPE_SHIFT)
-
- #define OP_ALG_ALGSEL_SHIFT 16
- #define OP_ALG_ALGSEL_MASK (0xff << OP_ALG_ALGSEL_SHIFT)
-@@ -1249,7 +1274,7 @@ struct sec4_sg_entry {
- #define OP_ALG_PKMODE_MOD_PRIMALITY 0x00f
-
- /* PKHA mode copy-memory functions */
--#define OP_ALG_PKMODE_SRC_REG_SHIFT 13
-+#define OP_ALG_PKMODE_SRC_REG_SHIFT 17
- #define OP_ALG_PKMODE_SRC_REG_MASK (7 << OP_ALG_PKMODE_SRC_REG_SHIFT)
- #define OP_ALG_PKMODE_DST_REG_SHIFT 10
- #define OP_ALG_PKMODE_DST_REG_MASK (7 << OP_ALG_PKMODE_DST_REG_SHIFT)
-@@ -1445,10 +1470,11 @@ struct sec4_sg_entry {
- #define MATH_SRC1_REG2 (0x02 << MATH_SRC1_SHIFT)
- #define MATH_SRC1_REG3 (0x03 << MATH_SRC1_SHIFT)
- #define MATH_SRC1_IMM (0x04 << MATH_SRC1_SHIFT)
--#define MATH_SRC1_DPOVRD (0x07 << MATH_SRC0_SHIFT)
-+#define MATH_SRC1_DPOVRD (0x07 << MATH_SRC1_SHIFT)
- #define MATH_SRC1_INFIFO (0x0a << MATH_SRC1_SHIFT)
- #define MATH_SRC1_OUTFIFO (0x0b << MATH_SRC1_SHIFT)
- #define MATH_SRC1_ONE (0x0c << MATH_SRC1_SHIFT)
-+#define MATH_SRC1_ZERO (0x0f << MATH_SRC1_SHIFT)
-
- /* Destination selectors */
- #define MATH_DEST_SHIFT 8
-@@ -1457,6 +1483,7 @@ struct sec4_sg_entry {
- #define MATH_DEST_REG1 (0x01 << MATH_DEST_SHIFT)
- #define MATH_DEST_REG2 (0x02 << MATH_DEST_SHIFT)
- #define MATH_DEST_REG3 (0x03 << MATH_DEST_SHIFT)
-+#define MATH_DEST_DPOVRD (0x07 << MATH_DEST_SHIFT)
- #define MATH_DEST_SEQINLEN (0x08 << MATH_DEST_SHIFT)
- #define MATH_DEST_SEQOUTLEN (0x09 << MATH_DEST_SHIFT)
- #define MATH_DEST_VARSEQINLEN (0x0a << MATH_DEST_SHIFT)
-@@ -1629,4 +1656,31 @@ struct sec4_sg_entry {
- /* Frame Descriptor Command for Replacement Job Descriptor */
- #define FD_CMD_REPLACE_JOB_DESC 0x20000000
-
-+/* CHA Control Register bits */
-+#define CCTRL_RESET_CHA_ALL 0x1
-+#define CCTRL_RESET_CHA_AESA 0x2
-+#define CCTRL_RESET_CHA_DESA 0x4
-+#define CCTRL_RESET_CHA_AFHA 0x8
-+#define CCTRL_RESET_CHA_KFHA 0x10
-+#define CCTRL_RESET_CHA_SF8A 0x20
-+#define CCTRL_RESET_CHA_PKHA 0x40
-+#define CCTRL_RESET_CHA_MDHA 0x80
-+#define CCTRL_RESET_CHA_CRCA 0x100
-+#define CCTRL_RESET_CHA_RNG 0x200
-+#define CCTRL_RESET_CHA_SF9A 0x400
-+#define CCTRL_RESET_CHA_ZUCE 0x800
-+#define CCTRL_RESET_CHA_ZUCA 0x1000
-+#define CCTRL_UNLOAD_PK_A0 0x10000
-+#define CCTRL_UNLOAD_PK_A1 0x20000
-+#define CCTRL_UNLOAD_PK_A2 0x40000
-+#define CCTRL_UNLOAD_PK_A3 0x80000
-+#define CCTRL_UNLOAD_PK_B0 0x100000
-+#define CCTRL_UNLOAD_PK_B1 0x200000
-+#define CCTRL_UNLOAD_PK_B2 0x400000
-+#define CCTRL_UNLOAD_PK_B3 0x800000
-+#define CCTRL_UNLOAD_PK_N 0x1000000
-+#define CCTRL_UNLOAD_PK_A 0x4000000
-+#define CCTRL_UNLOAD_PK_B 0x8000000
-+#define CCTRL_UNLOAD_SBOX 0x10000000
-+
- #endif /* DESC_H */
---- a/drivers/crypto/caam/desc_constr.h
-+++ b/drivers/crypto/caam/desc_constr.h
-@@ -4,6 +4,9 @@
- * Copyright 2008-2012 Freescale Semiconductor, Inc.
- */
-
-+#ifndef DESC_CONSTR_H
-+#define DESC_CONSTR_H
-+
- #include "desc.h"
- #include "regs.h"
-
-@@ -33,38 +36,39 @@
-
- extern bool caam_little_end;
-
--static inline int desc_len(u32 *desc)
-+static inline int desc_len(u32 * const desc)
- {
- return caam32_to_cpu(*desc) & HDR_DESCLEN_MASK;
- }
-
--static inline int desc_bytes(void *desc)
-+static inline int desc_bytes(void * const desc)
- {
- return desc_len(desc) * CAAM_CMD_SZ;
- }
-
--static inline u32 *desc_end(u32 *desc)
-+static inline u32 *desc_end(u32 * const desc)
- {
- return desc + desc_len(desc);
- }
-
--static inline void *sh_desc_pdb(u32 *desc)
-+static inline void *sh_desc_pdb(u32 * const desc)
- {
- return desc + 1;
- }
-
--static inline void init_desc(u32 *desc, u32 options)
-+static inline void init_desc(u32 * const desc, u32 options)
- {
- *desc = cpu_to_caam32((options | HDR_ONE) + 1);
- }
-
--static inline void init_sh_desc(u32 *desc, u32 options)
-+static inline void init_sh_desc(u32 * const desc, u32 options)
- {
- PRINT_POS;
- init_desc(desc, CMD_SHARED_DESC_HDR | options);
- }
-
--static inline void init_sh_desc_pdb(u32 *desc, u32 options, size_t pdb_bytes)
-+static inline void init_sh_desc_pdb(u32 * const desc, u32 options,
-+ size_t pdb_bytes)
- {
- u32 pdb_len = (pdb_bytes + CAAM_CMD_SZ - 1) / CAAM_CMD_SZ;
-
-@@ -72,19 +76,20 @@ static inline void init_sh_desc_pdb(u32
- options);
- }
-
--static inline void init_job_desc(u32 *desc, u32 options)
-+static inline void init_job_desc(u32 * const desc, u32 options)
- {
- init_desc(desc, CMD_DESC_HDR | options);
- }
-
--static inline void init_job_desc_pdb(u32 *desc, u32 options, size_t pdb_bytes)
-+static inline void init_job_desc_pdb(u32 * const desc, u32 options,
-+ size_t pdb_bytes)
- {
- u32 pdb_len = (pdb_bytes + CAAM_CMD_SZ - 1) / CAAM_CMD_SZ;
-
- init_job_desc(desc, (((pdb_len + 1) << HDR_START_IDX_SHIFT)) | options);
- }
-
--static inline void append_ptr(u32 *desc, dma_addr_t ptr)
-+static inline void append_ptr(u32 * const desc, dma_addr_t ptr)
- {
- dma_addr_t *offset = (dma_addr_t *)desc_end(desc);
-
-@@ -94,8 +99,8 @@ static inline void append_ptr(u32 *desc,
- CAAM_PTR_SZ / CAAM_CMD_SZ);
- }
-
--static inline void init_job_desc_shared(u32 *desc, dma_addr_t ptr, int len,
-- u32 options)
-+static inline void init_job_desc_shared(u32 * const desc, dma_addr_t ptr,
-+ int len, u32 options)
- {
- PRINT_POS;
- init_job_desc(desc, HDR_SHARED | options |
-@@ -103,7 +108,7 @@ static inline void init_job_desc_shared(
- append_ptr(desc, ptr);
- }
-
--static inline void append_data(u32 *desc, void *data, int len)
-+static inline void append_data(u32 * const desc, const void *data, int len)
- {
- u32 *offset = desc_end(desc);
-
-@@ -114,7 +119,7 @@ static inline void append_data(u32 *desc
- (len + CAAM_CMD_SZ - 1) / CAAM_CMD_SZ);
- }
-
--static inline void append_cmd(u32 *desc, u32 command)
-+static inline void append_cmd(u32 * const desc, u32 command)
- {
- u32 *cmd = desc_end(desc);
-
-@@ -125,7 +130,7 @@ static inline void append_cmd(u32 *desc,
-
- #define append_u32 append_cmd
-
--static inline void append_u64(u32 *desc, u64 data)
-+static inline void append_u64(u32 * const desc, u64 data)
- {
- u32 *offset = desc_end(desc);
-
-@@ -142,14 +147,14 @@ static inline void append_u64(u32 *desc,
- }
-
- /* Write command without affecting header, and return pointer to next word */
--static inline u32 *write_cmd(u32 *desc, u32 command)
-+static inline u32 *write_cmd(u32 * const desc, u32 command)
- {
- *desc = cpu_to_caam32(command);
-
- return desc + 1;
- }
-
--static inline void append_cmd_ptr(u32 *desc, dma_addr_t ptr, int len,
-+static inline void append_cmd_ptr(u32 * const desc, dma_addr_t ptr, int len,
- u32 command)
- {
- append_cmd(desc, command | len);
-@@ -157,7 +162,7 @@ static inline void append_cmd_ptr(u32 *d
- }
-
- /* Write length after pointer, rather than inside command */
--static inline void append_cmd_ptr_extlen(u32 *desc, dma_addr_t ptr,
-+static inline void append_cmd_ptr_extlen(u32 * const desc, dma_addr_t ptr,
- unsigned int len, u32 command)
- {
- append_cmd(desc, command);
-@@ -166,7 +171,7 @@ static inline void append_cmd_ptr_extlen
- append_cmd(desc, len);
- }
-
--static inline void append_cmd_data(u32 *desc, void *data, int len,
-+static inline void append_cmd_data(u32 * const desc, const void *data, int len,
- u32 command)
- {
- append_cmd(desc, command | IMMEDIATE | len);
-@@ -174,7 +179,7 @@ static inline void append_cmd_data(u32 *
- }
-
- #define APPEND_CMD_RET(cmd, op) \
--static inline u32 *append_##cmd(u32 *desc, u32 options) \
-+static inline u32 *append_##cmd(u32 * const desc, u32 options) \
- { \
- u32 *cmd = desc_end(desc); \
- PRINT_POS; \
-@@ -183,14 +188,15 @@ static inline u32 *append_##cmd(u32 *des
- }
- APPEND_CMD_RET(jump, JUMP)
- APPEND_CMD_RET(move, MOVE)
-+APPEND_CMD_RET(moveb, MOVEB)
-
--static inline void set_jump_tgt_here(u32 *desc, u32 *jump_cmd)
-+static inline void set_jump_tgt_here(u32 * const desc, u32 *jump_cmd)
- {
- *jump_cmd = cpu_to_caam32(caam32_to_cpu(*jump_cmd) |
- (desc_len(desc) - (jump_cmd - desc)));
- }
-
--static inline void set_move_tgt_here(u32 *desc, u32 *move_cmd)
-+static inline void set_move_tgt_here(u32 * const desc, u32 *move_cmd)
- {
- u32 val = caam32_to_cpu(*move_cmd);
-
-@@ -200,7 +206,7 @@ static inline void set_move_tgt_here(u32
- }
-
- #define APPEND_CMD(cmd, op) \
--static inline void append_##cmd(u32 *desc, u32 options) \
-+static inline void append_##cmd(u32 * const desc, u32 options) \
- { \
- PRINT_POS; \
- append_cmd(desc, CMD_##op | options); \
-@@ -208,7 +214,8 @@ static inline void append_##cmd(u32 *des
- APPEND_CMD(operation, OPERATION)
-
- #define APPEND_CMD_LEN(cmd, op) \
--static inline void append_##cmd(u32 *desc, unsigned int len, u32 options) \
-+static inline void append_##cmd(u32 * const desc, unsigned int len, \
-+ u32 options) \
- { \
- PRINT_POS; \
- append_cmd(desc, CMD_##op | len | options); \
-@@ -220,8 +227,8 @@ APPEND_CMD_LEN(seq_fifo_load, SEQ_FIFO_L
- APPEND_CMD_LEN(seq_fifo_store, SEQ_FIFO_STORE)
-
- #define APPEND_CMD_PTR(cmd, op) \
--static inline void append_##cmd(u32 *desc, dma_addr_t ptr, unsigned int len, \
-- u32 options) \
-+static inline void append_##cmd(u32 * const desc, dma_addr_t ptr, \
-+ unsigned int len, u32 options) \
- { \
- PRINT_POS; \
- append_cmd_ptr(desc, ptr, len, CMD_##op | options); \
-@@ -231,8 +238,8 @@ APPEND_CMD_PTR(load, LOAD)
- APPEND_CMD_PTR(fifo_load, FIFO_LOAD)
- APPEND_CMD_PTR(fifo_store, FIFO_STORE)
-
--static inline void append_store(u32 *desc, dma_addr_t ptr, unsigned int len,
-- u32 options)
-+static inline void append_store(u32 * const desc, dma_addr_t ptr,
-+ unsigned int len, u32 options)
- {
- u32 cmd_src;
-
-@@ -249,7 +256,8 @@ static inline void append_store(u32 *des
- }
-
- #define APPEND_SEQ_PTR_INTLEN(cmd, op) \
--static inline void append_seq_##cmd##_ptr_intlen(u32 *desc, dma_addr_t ptr, \
-+static inline void append_seq_##cmd##_ptr_intlen(u32 * const desc, \
-+ dma_addr_t ptr, \
- unsigned int len, \
- u32 options) \
- { \
-@@ -263,7 +271,7 @@ APPEND_SEQ_PTR_INTLEN(in, IN)
- APPEND_SEQ_PTR_INTLEN(out, OUT)
-
- #define APPEND_CMD_PTR_TO_IMM(cmd, op) \
--static inline void append_##cmd##_as_imm(u32 *desc, void *data, \
-+static inline void append_##cmd##_as_imm(u32 * const desc, const void *data, \
- unsigned int len, u32 options) \
- { \
- PRINT_POS; \
-@@ -273,7 +281,7 @@ APPEND_CMD_PTR_TO_IMM(load, LOAD);
- APPEND_CMD_PTR_TO_IMM(fifo_load, FIFO_LOAD);
-
- #define APPEND_CMD_PTR_EXTLEN(cmd, op) \
--static inline void append_##cmd##_extlen(u32 *desc, dma_addr_t ptr, \
-+static inline void append_##cmd##_extlen(u32 * const desc, dma_addr_t ptr, \
- unsigned int len, u32 options) \
- { \
- PRINT_POS; \
-@@ -287,7 +295,7 @@ APPEND_CMD_PTR_EXTLEN(seq_out_ptr, SEQ_O
- * the size of its type
- */
- #define APPEND_CMD_PTR_LEN(cmd, op, type) \
--static inline void append_##cmd(u32 *desc, dma_addr_t ptr, \
-+static inline void append_##cmd(u32 * const desc, dma_addr_t ptr, \
- type len, u32 options) \
- { \
- PRINT_POS; \
-@@ -304,7 +312,7 @@ APPEND_CMD_PTR_LEN(seq_out_ptr, SEQ_OUT_
- * from length of immediate data provided, e.g., split keys
- */
- #define APPEND_CMD_PTR_TO_IMM2(cmd, op) \
--static inline void append_##cmd##_as_imm(u32 *desc, void *data, \
-+static inline void append_##cmd##_as_imm(u32 * const desc, const void *data, \
- unsigned int data_len, \
- unsigned int len, u32 options) \
- { \
-@@ -315,7 +323,7 @@ static inline void append_##cmd##_as_imm
- APPEND_CMD_PTR_TO_IMM2(key, KEY);
-
- #define APPEND_CMD_RAW_IMM(cmd, op, type) \
--static inline void append_##cmd##_imm_##type(u32 *desc, type immediate, \
-+static inline void append_##cmd##_imm_##type(u32 * const desc, type immediate, \
- u32 options) \
- { \
- PRINT_POS; \
-@@ -426,3 +434,107 @@ do { \
- APPEND_MATH_IMM_u64(LSHIFT, desc, dest, src0, src1, data)
- #define append_math_rshift_imm_u64(desc, dest, src0, src1, data) \
- APPEND_MATH_IMM_u64(RSHIFT, desc, dest, src0, src1, data)
-+
-+/**
-+ * struct alginfo - Container for algorithm details
-+ * @algtype: algorithm selector; for valid values, see documentation of the
-+ * functions where it is used.
-+ * @keylen: length of the provided algorithm key, in bytes
-+ * @keylen_pad: padded length of the provided algorithm key, in bytes
-+ * @key: address where algorithm key resides; virtual address if key_inline
-+ * is true, dma (bus) address if key_inline is false.
-+ * @key_inline: true - key can be inlined in the descriptor; false - key is
-+ * referenced by the descriptor
-+ */
-+struct alginfo {
-+ u32 algtype;
-+ unsigned int keylen;
-+ unsigned int keylen_pad;
-+ union {
-+ dma_addr_t key_dma;
-+ const void *key_virt;
-+ };
-+ bool key_inline;
-+};
-+
-+/**
-+ * desc_inline_query() - Provide indications on which data items can be inlined
-+ * and which shall be referenced in a shared descriptor.
-+ * @sd_base_len: Shared descriptor base length - bytes consumed by the commands,
-+ * excluding the data items to be inlined (or corresponding
-+ * pointer if an item is not inlined). Each cnstr_* function that
-+ * generates descriptors should have a define mentioning
-+ * corresponding length.
-+ * @jd_len: Maximum length of the job descriptor(s) that will be used
-+ * together with the shared descriptor.
-+ * @data_len: Array of lengths of the data items trying to be inlined
-+ * @inl_mask: 32bit mask with bit x = 1 if data item x can be inlined, 0
-+ * otherwise.
-+ * @count: Number of data items (size of @data_len array); must be <= 32
-+ *
-+ * Return: 0 if data can be inlined / referenced, negative value if not. If 0,
-+ * check @inl_mask for details.
-+ */
-+static inline int desc_inline_query(unsigned int sd_base_len,
-+ unsigned int jd_len, unsigned int *data_len,
-+ u32 *inl_mask, unsigned int count)
-+{
-+ int rem_bytes = (int)(CAAM_DESC_BYTES_MAX - sd_base_len - jd_len);
-+ unsigned int i;
-+
-+ *inl_mask = 0;
-+ for (i = 0; (i < count) && (rem_bytes > 0); i++) {
-+ if (rem_bytes - (int)(data_len[i] +
-+ (count - i - 1) * CAAM_PTR_SZ) >= 0) {
-+ rem_bytes -= data_len[i];
-+ *inl_mask |= (1 << i);
-+ } else {
-+ rem_bytes -= CAAM_PTR_SZ;
-+ }
-+ }
-+
-+ return (rem_bytes >= 0) ? 0 : -1;
-+}
-+
-+/**
-+ * append_proto_dkp - Derived Key Protocol (DKP): key -> split key
-+ * @desc: pointer to buffer used for descriptor construction
-+ * @adata: pointer to authentication transform definitions.
-+ * keylen should be the length of initial key, while keylen_pad
-+ * the length of the derived (split) key.
-+ * Valid algorithm values - one of OP_ALG_ALGSEL_{MD5, SHA1, SHA224,
-+ * SHA256, SHA384, SHA512}.
-+ */
-+static inline void append_proto_dkp(u32 * const desc, struct alginfo *adata)
-+{
-+ u32 protid;
-+
-+ /*
-+ * Quick & dirty translation from OP_ALG_ALGSEL_{MD5, SHA*}
-+ * to OP_PCLID_DKP_{MD5, SHA*}
-+ */
-+ protid = (adata->algtype & OP_ALG_ALGSEL_SUBMASK) |
-+ (0x20 << OP_ALG_ALGSEL_SHIFT);
-+
-+ if (adata->key_inline) {
-+ int words;
-+
-+ append_operation(desc, OP_TYPE_UNI_PROTOCOL | protid |
-+ OP_PCL_DKP_SRC_IMM | OP_PCL_DKP_DST_IMM |
-+ adata->keylen);
-+ append_data(desc, adata->key_virt, adata->keylen);
-+
-+ /* Reserve space in descriptor buffer for the derived key */
-+ words = (ALIGN(adata->keylen_pad, CAAM_CMD_SZ) -
-+ ALIGN(adata->keylen, CAAM_CMD_SZ)) / CAAM_CMD_SZ;
-+ if (words)
-+ (*desc) = cpu_to_caam32(caam32_to_cpu(*desc) + words);
-+ } else {
-+ append_operation(desc, OP_TYPE_UNI_PROTOCOL | protid |
-+ OP_PCL_DKP_SRC_PTR | OP_PCL_DKP_DST_PTR |
-+ adata->keylen);
-+ append_ptr(desc, adata->key_dma);
-+ }
-+}
-+
-+#endif /* DESC_CONSTR_H */
---- /dev/null
-+++ b/drivers/crypto/caam/dpseci.c
-@@ -0,0 +1,858 @@
-+/*
-+ * Copyright 2013-2016 Freescale Semiconductor Inc.
-+ * Copyright 2017 NXP
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the names of the above-listed copyright holders nor the
-+ * names of any contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
-+ * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
-+ * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
-+ * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE
-+ * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
-+ * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
-+ * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
-+ * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
-+ * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
-+ * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
-+ * POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#include <linux/fsl/mc.h>
-+#include "../../../drivers/staging/fsl-mc/include/dpopr.h"
-+#include "dpseci.h"
-+#include "dpseci_cmd.h"
-+
-+/**
-+ * dpseci_open() - Open a control session for the specified object
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @dpseci_id: DPSECI unique ID
-+ * @token: Returned token; use in subsequent API calls
-+ *
-+ * This function can be used to open a control session for an already created
-+ * object; an object may have been declared in the DPL or by calling the
-+ * dpseci_create() function.
-+ * This function returns a unique authentication token, associated with the
-+ * specific object ID and the specific MC portal; this token must be used in all
-+ * subsequent commands for this specific object.
-+ *
-+ * Return: '0' on success, error code otherwise
-+ */
-+int dpseci_open(struct fsl_mc_io *mc_io, u32 cmd_flags, int dpseci_id,
-+ u16 *token)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpseci_cmd_open *cmd_params;
-+ int err;
-+
-+ cmd.header = mc_encode_cmd_header(DPSECI_CMDID_OPEN,
-+ cmd_flags,
-+ 0);
-+ cmd_params = (struct dpseci_cmd_open *)cmd.params;
-+ cmd_params->dpseci_id = cpu_to_le32(dpseci_id);
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ *token = mc_cmd_hdr_read_token(&cmd);
-+
-+ return 0;
-+}
-+
-+/**
-+ * dpseci_close() - Close the control session of the object
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPSECI object
-+ *
-+ * After this function is called, no further operations are allowed on the
-+ * object without opening a new control session.
-+ *
-+ * Return: '0' on success, error code otherwise
-+ */
-+int dpseci_close(struct fsl_mc_io *mc_io, u32 cmd_flags, u16 token)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ cmd.header = mc_encode_cmd_header(DPSECI_CMDID_CLOSE,
-+ cmd_flags,
-+ token);
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpseci_create() - Create the DPSECI object
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @dprc_token: Parent container token; '0' for default container
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @cfg: Configuration structure
-+ * @obj_id: returned object id
-+ *
-+ * Create the DPSECI object, allocate required resources and perform required
-+ * initialization.
-+ *
-+ * The object can be created either by declaring it in the DPL file, or by
-+ * calling this function.
-+ *
-+ * The function accepts an authentication token of a parent container that this
-+ * object should be assigned to. The token can be '0' so the object will be
-+ * assigned to the default container.
-+ * The newly created object can be opened with the returned object id and using
-+ * the container's associated tokens and MC portals.
-+ *
-+ * Return: '0' on success, error code otherwise
-+ */
-+int dpseci_create(struct fsl_mc_io *mc_io, u16 dprc_token, u32 cmd_flags,
-+ const struct dpseci_cfg *cfg, u32 *obj_id)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpseci_cmd_create *cmd_params;
-+ int i, err;
-+
-+ cmd.header = mc_encode_cmd_header(DPSECI_CMDID_CREATE,
-+ cmd_flags,
-+ dprc_token);
-+ cmd_params = (struct dpseci_cmd_create *)cmd.params;
-+ for (i = 0; i < 8; i++)
-+ cmd_params->priorities[i] = cfg->priorities[i];
-+ cmd_params->num_tx_queues = cfg->num_tx_queues;
-+ cmd_params->num_rx_queues = cfg->num_rx_queues;
-+ cmd_params->options = cpu_to_le32(cfg->options);
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ *obj_id = mc_cmd_read_object_id(&cmd);
-+
-+ return 0;
-+}
-+
-+/**
-+ * dpseci_destroy() - Destroy the DPSECI object and release all its resources
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @dprc_token: Parent container token; '0' for default container
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @object_id: The object id; it must be a valid id within the container that
-+ * created this object
-+ *
-+ * The function accepts the authentication token of the parent container that
-+ * created the object (not the one that currently owns the object). The object
-+ * is searched within parent using the provided 'object_id'.
-+ * All tokens to the object must be closed before calling destroy.
-+ *
-+ * Return: '0' on success, error code otherwise
-+ */
-+int dpseci_destroy(struct fsl_mc_io *mc_io, u16 dprc_token, u32 cmd_flags,
-+ u32 object_id)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpseci_cmd_destroy *cmd_params;
-+
-+ cmd.header = mc_encode_cmd_header(DPSECI_CMDID_DESTROY,
-+ cmd_flags,
-+ dprc_token);
-+ cmd_params = (struct dpseci_cmd_destroy *)cmd.params;
-+ cmd_params->object_id = cpu_to_le32(object_id);
-+
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpseci_enable() - Enable the DPSECI, allow sending and receiving frames
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPSECI object
-+ *
-+ * Return: '0' on success, error code otherwise
-+ */
-+int dpseci_enable(struct fsl_mc_io *mc_io, u32 cmd_flags, u16 token)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ cmd.header = mc_encode_cmd_header(DPSECI_CMDID_ENABLE,
-+ cmd_flags,
-+ token);
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpseci_disable() - Disable the DPSECI, stop sending and receiving frames
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPSECI object
-+ *
-+ * Return: '0' on success, error code otherwise
-+ */
-+int dpseci_disable(struct fsl_mc_io *mc_io, u32 cmd_flags, u16 token)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ cmd.header = mc_encode_cmd_header(DPSECI_CMDID_DISABLE,
-+ cmd_flags,
-+ token);
-+
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpseci_is_enabled() - Check if the DPSECI is enabled.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPSECI object
-+ * @en: Returns '1' if object is enabled; '0' otherwise
-+ *
-+ * Return: '0' on success, error code otherwise
-+ */
-+int dpseci_is_enabled(struct fsl_mc_io *mc_io, u32 cmd_flags, u16 token,
-+ int *en)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpseci_rsp_is_enabled *rsp_params;
-+ int err;
-+
-+ cmd.header = mc_encode_cmd_header(DPSECI_CMDID_IS_ENABLED,
-+ cmd_flags,
-+ token);
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ rsp_params = (struct dpseci_rsp_is_enabled *)cmd.params;
-+ *en = le32_to_cpu(rsp_params->is_enabled);
-+
-+ return 0;
-+}
-+
-+/**
-+ * dpseci_reset() - Reset the DPSECI, returns the object to initial state.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPSECI object
-+ *
-+ * Return: '0' on success, error code otherwise
-+ */
-+int dpseci_reset(struct fsl_mc_io *mc_io, u32 cmd_flags, u16 token)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ cmd.header = mc_encode_cmd_header(DPSECI_CMDID_RESET,
-+ cmd_flags,
-+ token);
-+
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpseci_get_irq_enable() - Get overall interrupt state
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPSECI object
-+ * @irq_index: The interrupt index to configure
-+ * @en: Returned Interrupt state - enable = 1, disable = 0
-+ *
-+ * Return: '0' on success, error code otherwise
-+ */
-+int dpseci_get_irq_enable(struct fsl_mc_io *mc_io, u32 cmd_flags, u16 token,
-+ u8 irq_index, u8 *en)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpseci_cmd_irq_enable *cmd_params;
-+ struct dpseci_rsp_get_irq_enable *rsp_params;
-+ int err;
-+
-+ cmd.header = mc_encode_cmd_header(DPSECI_CMDID_GET_IRQ_ENABLE,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpseci_cmd_irq_enable *)cmd.params;
-+ cmd_params->irq_index = irq_index;
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ rsp_params = (struct dpseci_rsp_get_irq_enable *)cmd.params;
-+ *en = rsp_params->enable_state;
-+
-+ return 0;
-+}
-+
-+/**
-+ * dpseci_set_irq_enable() - Set overall interrupt state.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPSECI object
-+ * @irq_index: The interrupt index to configure
-+ * @en: Interrupt state - enable = 1, disable = 0
-+ *
-+ * Allows GPP software to control when interrupts are generated.
-+ * Each interrupt can have up to 32 causes. The enable/disable control's the
-+ * overall interrupt state. If the interrupt is disabled no causes will cause
-+ * an interrupt.
-+ *
-+ * Return: '0' on success, error code otherwise
-+ */
-+int dpseci_set_irq_enable(struct fsl_mc_io *mc_io, u32 cmd_flags, u16 token,
-+ u8 irq_index, u8 en)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpseci_cmd_irq_enable *cmd_params;
-+
-+ cmd.header = mc_encode_cmd_header(DPSECI_CMDID_SET_IRQ_ENABLE,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpseci_cmd_irq_enable *)cmd.params;
-+ cmd_params->irq_index = irq_index;
-+ cmd_params->enable_state = en;
-+
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpseci_get_irq_mask() - Get interrupt mask.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPSECI object
-+ * @irq_index: The interrupt index to configure
-+ * @mask: Returned event mask to trigger interrupt
-+ *
-+ * Every interrupt can have up to 32 causes and the interrupt model supports
-+ * masking/unmasking each cause independently.
-+ *
-+ * Return: '0' on success, error code otherwise
-+ */
-+int dpseci_get_irq_mask(struct fsl_mc_io *mc_io, u32 cmd_flags, u16 token,
-+ u8 irq_index, u32 *mask)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpseci_cmd_irq_mask *cmd_params;
-+ int err;
-+
-+ cmd.header = mc_encode_cmd_header(DPSECI_CMDID_GET_IRQ_MASK,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpseci_cmd_irq_mask *)cmd.params;
-+ cmd_params->irq_index = irq_index;
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ *mask = le32_to_cpu(cmd_params->mask);
-+
-+ return 0;
-+}
-+
-+/**
-+ * dpseci_set_irq_mask() - Set interrupt mask.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPSECI object
-+ * @irq_index: The interrupt index to configure
-+ * @mask: event mask to trigger interrupt;
-+ * each bit:
-+ * 0 = ignore event
-+ * 1 = consider event for asserting IRQ
-+ *
-+ * Every interrupt can have up to 32 causes and the interrupt model supports
-+ * masking/unmasking each cause independently
-+ *
-+ * Return: '0' on success, error code otherwise
-+ */
-+int dpseci_set_irq_mask(struct fsl_mc_io *mc_io, u32 cmd_flags, u16 token,
-+ u8 irq_index, u32 mask)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpseci_cmd_irq_mask *cmd_params;
-+
-+ cmd.header = mc_encode_cmd_header(DPSECI_CMDID_SET_IRQ_MASK,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpseci_cmd_irq_mask *)cmd.params;
-+ cmd_params->mask = cpu_to_le32(mask);
-+ cmd_params->irq_index = irq_index;
-+
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpseci_get_irq_status() - Get the current status of any pending interrupts
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPSECI object
-+ * @irq_index: The interrupt index to configure
-+ * @status: Returned interrupts status - one bit per cause:
-+ * 0 = no interrupt pending
-+ * 1 = interrupt pending
-+ *
-+ * Return: '0' on success, error code otherwise
-+ */
-+int dpseci_get_irq_status(struct fsl_mc_io *mc_io, u32 cmd_flags, u16 token,
-+ u8 irq_index, u32 *status)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpseci_cmd_irq_status *cmd_params;
-+ int err;
-+
-+ cmd.header = mc_encode_cmd_header(DPSECI_CMDID_GET_IRQ_STATUS,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpseci_cmd_irq_status *)cmd.params;
-+ cmd_params->status = cpu_to_le32(*status);
-+ cmd_params->irq_index = irq_index;
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ *status = le32_to_cpu(cmd_params->status);
-+
-+ return 0;
-+}
-+
-+/**
-+ * dpseci_clear_irq_status() - Clear a pending interrupt's status
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPSECI object
-+ * @irq_index: The interrupt index to configure
-+ * @status: bits to clear (W1C) - one bit per cause:
-+ * 0 = don't change
-+ * 1 = clear status bit
-+ *
-+ * Return: '0' on success, error code otherwise
-+ */
-+int dpseci_clear_irq_status(struct fsl_mc_io *mc_io, u32 cmd_flags, u16 token,
-+ u8 irq_index, u32 status)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpseci_cmd_irq_status *cmd_params;
-+
-+ cmd.header = mc_encode_cmd_header(DPSECI_CMDID_CLEAR_IRQ_STATUS,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpseci_cmd_irq_status *)cmd.params;
-+ cmd_params->status = cpu_to_le32(status);
-+ cmd_params->irq_index = irq_index;
-+
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpseci_get_attributes() - Retrieve DPSECI attributes
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPSECI object
-+ * @attr: Returned object's attributes
-+ *
-+ * Return: '0' on success, error code otherwise
-+ */
-+int dpseci_get_attributes(struct fsl_mc_io *mc_io, u32 cmd_flags, u16 token,
-+ struct dpseci_attr *attr)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpseci_rsp_get_attributes *rsp_params;
-+ int err;
-+
-+ cmd.header = mc_encode_cmd_header(DPSECI_CMDID_GET_ATTR,
-+ cmd_flags,
-+ token);
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ rsp_params = (struct dpseci_rsp_get_attributes *)cmd.params;
-+ attr->id = le32_to_cpu(rsp_params->id);
-+ attr->num_tx_queues = rsp_params->num_tx_queues;
-+ attr->num_rx_queues = rsp_params->num_rx_queues;
-+ attr->options = le32_to_cpu(rsp_params->options);
-+
-+ return 0;
-+}
-+
-+/**
-+ * dpseci_set_rx_queue() - Set Rx queue configuration
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPSECI object
-+ * @queue: Select the queue relative to number of priorities configured at
-+ * DPSECI creation; use DPSECI_ALL_QUEUES to configure all
-+ * Rx queues identically.
-+ * @cfg: Rx queue configuration
-+ *
-+ * Return: '0' on success, error code otherwise
-+ */
-+int dpseci_set_rx_queue(struct fsl_mc_io *mc_io, u32 cmd_flags, u16 token,
-+ u8 queue, const struct dpseci_rx_queue_cfg *cfg)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpseci_cmd_queue *cmd_params;
-+
-+ cmd.header = mc_encode_cmd_header(DPSECI_CMDID_SET_RX_QUEUE,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpseci_cmd_queue *)cmd.params;
-+ cmd_params->dest_id = cpu_to_le32(cfg->dest_cfg.dest_id);
-+ cmd_params->priority = cfg->dest_cfg.priority;
-+ cmd_params->queue = queue;
-+ cmd_params->dest_type = cfg->dest_cfg.dest_type;
-+ cmd_params->user_ctx = cpu_to_le64(cfg->user_ctx);
-+ cmd_params->options = cpu_to_le32(cfg->options);
-+ cmd_params->order_preservation_en =
-+ cpu_to_le32(cfg->order_preservation_en);
-+
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpseci_get_rx_queue() - Retrieve Rx queue attributes
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPSECI object
-+ * @queue: Select the queue relative to number of priorities configured at
-+ * DPSECI creation
-+ * @attr: Returned Rx queue attributes
-+ *
-+ * Return: '0' on success, error code otherwise
-+ */
-+int dpseci_get_rx_queue(struct fsl_mc_io *mc_io, u32 cmd_flags, u16 token,
-+ u8 queue, struct dpseci_rx_queue_attr *attr)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpseci_cmd_queue *cmd_params;
-+ int err;
-+
-+ cmd.header = mc_encode_cmd_header(DPSECI_CMDID_GET_RX_QUEUE,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpseci_cmd_queue *)cmd.params;
-+ cmd_params->queue = queue;
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ attr->dest_cfg.dest_id = le32_to_cpu(cmd_params->dest_id);
-+ attr->dest_cfg.priority = cmd_params->priority;
-+ attr->dest_cfg.dest_type = cmd_params->dest_type;
-+ attr->user_ctx = le64_to_cpu(cmd_params->user_ctx);
-+ attr->fqid = le32_to_cpu(cmd_params->fqid);
-+ attr->order_preservation_en =
-+ le32_to_cpu(cmd_params->order_preservation_en);
-+
-+ return 0;
-+}
-+
-+/**
-+ * dpseci_get_tx_queue() - Retrieve Tx queue attributes
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPSECI object
-+ * @queue: Select the queue relative to number of priorities configured at
-+ * DPSECI creation
-+ * @attr: Returned Tx queue attributes
-+ *
-+ * Return: '0' on success, error code otherwise
-+ */
-+int dpseci_get_tx_queue(struct fsl_mc_io *mc_io, u32 cmd_flags, u16 token,
-+ u8 queue, struct dpseci_tx_queue_attr *attr)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpseci_cmd_queue *cmd_params;
-+ struct dpseci_rsp_get_tx_queue *rsp_params;
-+ int err;
-+
-+ cmd.header = mc_encode_cmd_header(DPSECI_CMDID_GET_TX_QUEUE,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpseci_cmd_queue *)cmd.params;
-+ cmd_params->queue = queue;
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ rsp_params = (struct dpseci_rsp_get_tx_queue *)cmd.params;
-+ attr->fqid = le32_to_cpu(rsp_params->fqid);
-+ attr->priority = rsp_params->priority;
-+
-+ return 0;
-+}
-+
-+/**
-+ * dpseci_get_sec_attr() - Retrieve SEC accelerator attributes
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPSECI object
-+ * @attr: Returned SEC attributes
-+ *
-+ * Return: '0' on success, error code otherwise
-+ */
-+int dpseci_get_sec_attr(struct fsl_mc_io *mc_io, u32 cmd_flags, u16 token,
-+ struct dpseci_sec_attr *attr)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpseci_rsp_get_sec_attr *rsp_params;
-+ int err;
-+
-+ cmd.header = mc_encode_cmd_header(DPSECI_CMDID_GET_SEC_ATTR,
-+ cmd_flags,
-+ token);
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ rsp_params = (struct dpseci_rsp_get_sec_attr *)cmd.params;
-+ attr->ip_id = le16_to_cpu(rsp_params->ip_id);
-+ attr->major_rev = rsp_params->major_rev;
-+ attr->minor_rev = rsp_params->minor_rev;
-+ attr->era = rsp_params->era;
-+ attr->deco_num = rsp_params->deco_num;
-+ attr->zuc_auth_acc_num = rsp_params->zuc_auth_acc_num;
-+ attr->zuc_enc_acc_num = rsp_params->zuc_enc_acc_num;
-+ attr->snow_f8_acc_num = rsp_params->snow_f8_acc_num;
-+ attr->snow_f9_acc_num = rsp_params->snow_f9_acc_num;
-+ attr->crc_acc_num = rsp_params->crc_acc_num;
-+ attr->pk_acc_num = rsp_params->pk_acc_num;
-+ attr->kasumi_acc_num = rsp_params->kasumi_acc_num;
-+ attr->rng_acc_num = rsp_params->rng_acc_num;
-+ attr->md_acc_num = rsp_params->md_acc_num;
-+ attr->arc4_acc_num = rsp_params->arc4_acc_num;
-+ attr->des_acc_num = rsp_params->des_acc_num;
-+ attr->aes_acc_num = rsp_params->aes_acc_num;
-+
-+ return 0;
-+}
-+
-+/**
-+ * dpseci_get_sec_counters() - Retrieve SEC accelerator counters
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPSECI object
-+ * @counters: Returned SEC counters
-+ *
-+ * Return: '0' on success, error code otherwise
-+ */
-+int dpseci_get_sec_counters(struct fsl_mc_io *mc_io, u32 cmd_flags, u16 token,
-+ struct dpseci_sec_counters *counters)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpseci_rsp_get_sec_counters *rsp_params;
-+ int err;
-+
-+ cmd.header = mc_encode_cmd_header(DPSECI_CMDID_GET_SEC_COUNTERS,
-+ cmd_flags,
-+ token);
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ rsp_params = (struct dpseci_rsp_get_sec_counters *)cmd.params;
-+ counters->dequeued_requests =
-+ le64_to_cpu(rsp_params->dequeued_requests);
-+ counters->ob_enc_requests = le64_to_cpu(rsp_params->ob_enc_requests);
-+ counters->ib_dec_requests = le64_to_cpu(rsp_params->ib_dec_requests);
-+ counters->ob_enc_bytes = le64_to_cpu(rsp_params->ob_enc_bytes);
-+ counters->ob_prot_bytes = le64_to_cpu(rsp_params->ob_prot_bytes);
-+ counters->ib_dec_bytes = le64_to_cpu(rsp_params->ib_dec_bytes);
-+ counters->ib_valid_bytes = le64_to_cpu(rsp_params->ib_valid_bytes);
-+
-+ return 0;
-+}
-+
-+/**
-+ * dpseci_get_api_version() - Get Data Path SEC Interface API version
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @major_ver: Major version of data path sec API
-+ * @minor_ver: Minor version of data path sec API
-+ *
-+ * Return: '0' on success, error code otherwise
-+ */
-+int dpseci_get_api_version(struct fsl_mc_io *mc_io, u32 cmd_flags,
-+ u16 *major_ver, u16 *minor_ver)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpseci_rsp_get_api_version *rsp_params;
-+ int err;
-+
-+ cmd.header = mc_encode_cmd_header(DPSECI_CMDID_GET_API_VERSION,
-+ cmd_flags, 0);
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ rsp_params = (struct dpseci_rsp_get_api_version *)cmd.params;
-+ *major_ver = le16_to_cpu(rsp_params->major);
-+ *minor_ver = le16_to_cpu(rsp_params->minor);
-+
-+ return 0;
-+}
-+
-+/**
-+ * dpseci_set_opr() - Set Order Restoration configuration
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPSECI object
-+ * @index: The queue index
-+ * @options: Configuration mode options; can be OPR_OPT_CREATE or
-+ * OPR_OPT_RETIRE
-+ * @cfg: Configuration options for the OPR
-+ *
-+ * Return: '0' on success, error code otherwise
-+ */
-+int dpseci_set_opr(struct fsl_mc_io *mc_io, u32 cmd_flags, u16 token, u8 index,
-+ u8 options, struct opr_cfg *cfg)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpseci_cmd_opr *cmd_params;
-+
-+ cmd.header = mc_encode_cmd_header(
-+ DPSECI_CMDID_SET_OPR,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpseci_cmd_opr *)cmd.params;
-+ cmd_params->index = index;
-+ cmd_params->options = options;
-+ cmd_params->oloe = cfg->oloe;
-+ cmd_params->oeane = cfg->oeane;
-+ cmd_params->olws = cfg->olws;
-+ cmd_params->oa = cfg->oa;
-+ cmd_params->oprrws = cfg->oprrws;
-+
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpseci_get_opr() - Retrieve Order Restoration config and query
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPSECI object
-+ * @index: The queue index
-+ * @cfg: Returned OPR configuration
-+ * @qry: Returned OPR query
-+ *
-+ * Return: '0' on success, error code otherwise
-+ */
-+int dpseci_get_opr(struct fsl_mc_io *mc_io, u32 cmd_flags, u16 token, u8 index,
-+ struct opr_cfg *cfg, struct opr_qry *qry)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpseci_cmd_opr *cmd_params;
-+ struct dpseci_rsp_get_opr *rsp_params;
-+ int err;
-+
-+ cmd.header = mc_encode_cmd_header(DPSECI_CMDID_GET_OPR,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpseci_cmd_opr *)cmd.params;
-+ cmd_params->index = index;
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ rsp_params = (struct dpseci_rsp_get_opr *)cmd.params;
-+ qry->rip = dpseci_get_field(rsp_params->rip_enable, OPR_RIP);
-+ qry->enable = dpseci_get_field(rsp_params->rip_enable, OPR_ENABLE);
-+ cfg->oloe = rsp_params->oloe;
-+ cfg->oeane = rsp_params->oeane;
-+ cfg->olws = rsp_params->olws;
-+ cfg->oa = rsp_params->oa;
-+ cfg->oprrws = rsp_params->oprrws;
-+ qry->nesn = le16_to_cpu(rsp_params->nesn);
-+ qry->ndsn = le16_to_cpu(rsp_params->ndsn);
-+ qry->ea_tseq = le16_to_cpu(rsp_params->ea_tseq);
-+ qry->tseq_nlis = dpseci_get_field(rsp_params->tseq_nlis, OPR_TSEQ_NLIS);
-+ qry->ea_hseq = le16_to_cpu(rsp_params->ea_hseq);
-+ qry->hseq_nlis = dpseci_get_field(rsp_params->tseq_nlis, OPR_HSEQ_NLIS);
-+ qry->ea_hptr = le16_to_cpu(rsp_params->ea_hptr);
-+ qry->ea_tptr = le16_to_cpu(rsp_params->ea_tptr);
-+ qry->opr_vid = le16_to_cpu(rsp_params->opr_vid);
-+ qry->opr_id = le16_to_cpu(rsp_params->opr_id);
-+
-+ return 0;
-+}
-+
-+/**
-+ * dpseci_set_congestion_notification() - Set congestion group
-+ * notification configuration
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPSECI object
-+ * @cfg: congestion notification configuration
-+ *
-+ * Return: '0' on success, error code otherwise
-+ */
-+int dpseci_set_congestion_notification(struct fsl_mc_io *mc_io, u32 cmd_flags,
-+ u16 token, const struct dpseci_congestion_notification_cfg *cfg)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpseci_cmd_congestion_notification *cmd_params;
-+
-+ cmd.header = mc_encode_cmd_header(
-+ DPSECI_CMDID_SET_CONGESTION_NOTIFICATION,
-+ cmd_flags,
-+ token);
-+ cmd_params = (struct dpseci_cmd_congestion_notification *)cmd.params;
-+ cmd_params->dest_id = cpu_to_le32(cfg->dest_cfg.dest_id);
-+ cmd_params->notification_mode = cpu_to_le16(cfg->notification_mode);
-+ cmd_params->priority = cfg->dest_cfg.priority;
-+ dpseci_set_field(cmd_params->options, CGN_DEST_TYPE,
-+ cfg->dest_cfg.dest_type);
-+ dpseci_set_field(cmd_params->options, CGN_UNITS, cfg->units);
-+ cmd_params->message_iova = cpu_to_le64(cfg->message_iova);
-+ cmd_params->message_ctx = cpu_to_le64(cfg->message_ctx);
-+ cmd_params->threshold_entry = cpu_to_le32(cfg->threshold_entry);
-+ cmd_params->threshold_exit = cpu_to_le32(cfg->threshold_exit);
-+
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+/**
-+ * dpseci_get_congestion_notification() - Get congestion group notification
-+ * configuration
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPSECI object
-+ * @cfg: congestion notification configuration
-+ *
-+ * Return: '0' on success, error code otherwise
-+ */
-+int dpseci_get_congestion_notification(struct fsl_mc_io *mc_io, u32 cmd_flags,
-+ u16 token, struct dpseci_congestion_notification_cfg *cfg)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpseci_cmd_congestion_notification *rsp_params;
-+ int err;
-+
-+ cmd.header = mc_encode_cmd_header(
-+ DPSECI_CMDID_GET_CONGESTION_NOTIFICATION,
-+ cmd_flags,
-+ token);
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ rsp_params = (struct dpseci_cmd_congestion_notification *)cmd.params;
-+ cfg->dest_cfg.dest_id = le32_to_cpu(rsp_params->dest_id);
-+ cfg->notification_mode = le16_to_cpu(rsp_params->notification_mode);
-+ cfg->dest_cfg.priority = rsp_params->priority;
-+ cfg->dest_cfg.dest_type = dpseci_get_field(rsp_params->options,
-+ CGN_DEST_TYPE);
-+ cfg->units = dpseci_get_field(rsp_params->options, CGN_UNITS);
-+ cfg->message_iova = le64_to_cpu(rsp_params->message_iova);
-+ cfg->message_ctx = le64_to_cpu(rsp_params->message_ctx);
-+ cfg->threshold_entry = le32_to_cpu(rsp_params->threshold_entry);
-+ cfg->threshold_exit = le32_to_cpu(rsp_params->threshold_exit);
-+
-+ return 0;
-+}
---- /dev/null
-+++ b/drivers/crypto/caam/dpseci.h
-@@ -0,0 +1,395 @@
-+/*
-+ * Copyright 2013-2016 Freescale Semiconductor Inc.
-+ * Copyright 2017 NXP
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the names of the above-listed copyright holders nor the
-+ * names of any contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
-+ * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
-+ * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
-+ * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE
-+ * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
-+ * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
-+ * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
-+ * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
-+ * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
-+ * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
-+ * POSSIBILITY OF SUCH DAMAGE.
-+ */
-+#ifndef _DPSECI_H_
-+#define _DPSECI_H_
-+
-+/*
-+ * Data Path SEC Interface API
-+ * Contains initialization APIs and runtime control APIs for DPSECI
-+ */
-+
-+struct fsl_mc_io;
-+struct opr_cfg;
-+struct opr_qry;
-+
-+/**
-+ * General DPSECI macros
-+ */
-+
-+/**
-+ * Maximum number of Tx/Rx priorities per DPSECI object
-+ */
-+#define DPSECI_PRIO_NUM 8
-+
-+/**
-+ * All queues considered; see dpseci_set_rx_queue()
-+ */
-+#define DPSECI_ALL_QUEUES (u8)(-1)
-+
-+int dpseci_open(struct fsl_mc_io *mc_io, u32 cmd_flags, int dpseci_id,
-+ u16 *token);
-+
-+int dpseci_close(struct fsl_mc_io *mc_io, u32 cmd_flags, u16 token);
-+
-+/**
-+ * Enable the Congestion Group support
-+ */
-+#define DPSECI_OPT_HAS_CG 0x000020
-+
-+/**
-+ * Enable the Order Restoration support
-+ */
-+#define DPSECI_OPT_HAS_OPR 0x000040
-+
-+/**
-+ * Order Point Records are shared for the entire DPSECI
-+ */
-+#define DPSECI_OPT_OPR_SHARED 0x000080
-+
-+/**
-+ * struct dpseci_cfg - Structure representing DPSECI configuration
-+ * @options: Any combination of the following options:
-+ * DPSECI_OPT_HAS_CG
-+ * DPSECI_OPT_HAS_OPR
-+ * DPSECI_OPT_OPR_SHARED
-+ * @num_tx_queues: num of queues towards the SEC
-+ * @num_rx_queues: num of queues back from the SEC
-+ * @priorities: Priorities for the SEC hardware processing;
-+ * each place in the array is the priority of the tx queue
-+ * towards the SEC;
-+ * valid priorities are configured with values 1-8;
-+ */
-+struct dpseci_cfg {
-+ u32 options;
-+ u8 num_tx_queues;
-+ u8 num_rx_queues;
-+ u8 priorities[DPSECI_PRIO_NUM];
-+};
-+
-+int dpseci_create(struct fsl_mc_io *mc_io, u16 dprc_token, u32 cmd_flags,
-+ const struct dpseci_cfg *cfg, u32 *obj_id);
-+
-+int dpseci_destroy(struct fsl_mc_io *mc_io, u16 dprc_token, u32 cmd_flags,
-+ u32 object_id);
-+
-+int dpseci_enable(struct fsl_mc_io *mc_io, u32 cmd_flags, u16 token);
-+
-+int dpseci_disable(struct fsl_mc_io *mc_io, u32 cmd_flags, u16 token);
-+
-+int dpseci_is_enabled(struct fsl_mc_io *mc_io, u32 cmd_flags, u16 token,
-+ int *en);
-+
-+int dpseci_reset(struct fsl_mc_io *mc_io, u32 cmd_flags, u16 token);
-+
-+int dpseci_get_irq_enable(struct fsl_mc_io *mc_io, u32 cmd_flags, u16 token,
-+ u8 irq_index, u8 *en);
-+
-+int dpseci_set_irq_enable(struct fsl_mc_io *mc_io, u32 cmd_flags, u16 token,
-+ u8 irq_index, u8 en);
-+
-+int dpseci_get_irq_mask(struct fsl_mc_io *mc_io, u32 cmd_flags, u16 token,
-+ u8 irq_index, u32 *mask);
-+
-+int dpseci_set_irq_mask(struct fsl_mc_io *mc_io, u32 cmd_flags, u16 token,
-+ u8 irq_index, u32 mask);
-+
-+int dpseci_get_irq_status(struct fsl_mc_io *mc_io, u32 cmd_flags, u16 token,
-+ u8 irq_index, u32 *status);
-+
-+int dpseci_clear_irq_status(struct fsl_mc_io *mc_io, u32 cmd_flags, u16 token,
-+ u8 irq_index, u32 status);
-+
-+/**
-+ * struct dpseci_attr - Structure representing DPSECI attributes
-+ * @id: DPSECI object ID
-+ * @num_tx_queues: number of queues towards the SEC
-+ * @num_rx_queues: number of queues back from the SEC
-+ * @options: any combination of the following options:
-+ * DPSECI_OPT_HAS_CG
-+ * DPSECI_OPT_HAS_OPR
-+ * DPSECI_OPT_OPR_SHARED
-+ */
-+struct dpseci_attr {
-+ int id;
-+ u8 num_tx_queues;
-+ u8 num_rx_queues;
-+ u32 options;
-+};
-+
-+int dpseci_get_attributes(struct fsl_mc_io *mc_io, u32 cmd_flags, u16 token,
-+ struct dpseci_attr *attr);
-+
-+/**
-+ * enum dpseci_dest - DPSECI destination types
-+ * @DPSECI_DEST_NONE: Unassigned destination; The queue is set in parked mode
-+ * and does not generate FQDAN notifications; user is expected to dequeue
-+ * from the queue based on polling or other user-defined method
-+ * @DPSECI_DEST_DPIO: The queue is set in schedule mode and generates FQDAN
-+ * notifications to the specified DPIO; user is expected to dequeue from
-+ * the queue only after notification is received
-+ * @DPSECI_DEST_DPCON: The queue is set in schedule mode and does not generate
-+ * FQDAN notifications, but is connected to the specified DPCON object;
-+ * user is expected to dequeue from the DPCON channel
-+ */
-+enum dpseci_dest {
-+ DPSECI_DEST_NONE = 0,
-+ DPSECI_DEST_DPIO,
-+ DPSECI_DEST_DPCON
-+};
-+
-+/**
-+ * struct dpseci_dest_cfg - Structure representing DPSECI destination parameters
-+ * @dest_type: Destination type
-+ * @dest_id: Either DPIO ID or DPCON ID, depending on the destination type
-+ * @priority: Priority selection within the DPIO or DPCON channel; valid values
-+ * are 0-1 or 0-7, depending on the number of priorities in that channel;
-+ * not relevant for 'DPSECI_DEST_NONE' option
-+ */
-+struct dpseci_dest_cfg {
-+ enum dpseci_dest dest_type;
-+ int dest_id;
-+ u8 priority;
-+};
-+
-+/**
-+ * DPSECI queue modification options
-+ */
-+
-+/**
-+ * Select to modify the user's context associated with the queue
-+ */
-+#define DPSECI_QUEUE_OPT_USER_CTX 0x00000001
-+
-+/**
-+ * Select to modify the queue's destination
-+ */
-+#define DPSECI_QUEUE_OPT_DEST 0x00000002
-+
-+/**
-+ * Select to modify the queue's order preservation
-+ */
-+#define DPSECI_QUEUE_OPT_ORDER_PRESERVATION 0x00000004
-+
-+/**
-+ * struct dpseci_rx_queue_cfg - DPSECI RX queue configuration
-+ * @options: Flags representing the suggested modifications to the queue;
-+ * Use any combination of 'DPSECI_QUEUE_OPT_<X>' flags
-+ * @order_preservation_en: order preservation configuration for the rx queue
-+ * valid only if 'DPSECI_QUEUE_OPT_ORDER_PRESERVATION' is contained in 'options'
-+ * @user_ctx: User context value provided in the frame descriptor of each
-+ * dequeued frame; valid only if 'DPSECI_QUEUE_OPT_USER_CTX' is contained
-+ * in 'options'
-+ * @dest_cfg: Queue destination parameters; valid only if
-+ * 'DPSECI_QUEUE_OPT_DEST' is contained in 'options'
-+ */
-+struct dpseci_rx_queue_cfg {
-+ u32 options;
-+ int order_preservation_en;
-+ u64 user_ctx;
-+ struct dpseci_dest_cfg dest_cfg;
-+};
-+
-+int dpseci_set_rx_queue(struct fsl_mc_io *mc_io, u32 cmd_flags, u16 token,
-+ u8 queue, const struct dpseci_rx_queue_cfg *cfg);
-+
-+/**
-+ * struct dpseci_rx_queue_attr - Structure representing attributes of Rx queues
-+ * @user_ctx: User context value provided in the frame descriptor of each
-+ * dequeued frame
-+ * @order_preservation_en: Status of the order preservation configuration on the
-+ * queue
-+ * @dest_cfg: Queue destination configuration
-+ * @fqid: Virtual FQID value to be used for dequeue operations
-+ */
-+struct dpseci_rx_queue_attr {
-+ u64 user_ctx;
-+ int order_preservation_en;
-+ struct dpseci_dest_cfg dest_cfg;
-+ u32 fqid;
-+};
-+
-+int dpseci_get_rx_queue(struct fsl_mc_io *mc_io, u32 cmd_flags, u16 token,
-+ u8 queue, struct dpseci_rx_queue_attr *attr);
-+
-+/**
-+ * struct dpseci_tx_queue_attr - Structure representing attributes of Tx queues
-+ * @fqid: Virtual FQID to be used for sending frames to SEC hardware
-+ * @priority: SEC hardware processing priority for the queue
-+ */
-+struct dpseci_tx_queue_attr {
-+ u32 fqid;
-+ u8 priority;
-+};
-+
-+int dpseci_get_tx_queue(struct fsl_mc_io *mc_io, u32 cmd_flags, u16 token,
-+ u8 queue, struct dpseci_tx_queue_attr *attr);
-+
-+/**
-+ * struct dpseci_sec_attr - Structure representing attributes of the SEC
-+ * hardware accelerator
-+ * @ip_id: ID for SEC
-+ * @major_rev: Major revision number for SEC
-+ * @minor_rev: Minor revision number for SEC
-+ * @era: SEC Era
-+ * @deco_num: The number of copies of the DECO that are implemented in this
-+ * version of SEC
-+ * @zuc_auth_acc_num: The number of copies of ZUCA that are implemented in this
-+ * version of SEC
-+ * @zuc_enc_acc_num: The number of copies of ZUCE that are implemented in this
-+ * version of SEC
-+ * @snow_f8_acc_num: The number of copies of the SNOW-f8 module that are
-+ * implemented in this version of SEC
-+ * @snow_f9_acc_num: The number of copies of the SNOW-f9 module that are
-+ * implemented in this version of SEC
-+ * @crc_acc_num: The number of copies of the CRC module that are implemented in
-+ * this version of SEC
-+ * @pk_acc_num: The number of copies of the Public Key module that are
-+ * implemented in this version of SEC
-+ * @kasumi_acc_num: The number of copies of the Kasumi module that are
-+ * implemented in this version of SEC
-+ * @rng_acc_num: The number of copies of the Random Number Generator that are
-+ * implemented in this version of SEC
-+ * @md_acc_num: The number of copies of the MDHA (Hashing module) that are
-+ * implemented in this version of SEC
-+ * @arc4_acc_num: The number of copies of the ARC4 module that are implemented
-+ * in this version of SEC
-+ * @des_acc_num: The number of copies of the DES module that are implemented in
-+ * this version of SEC
-+ * @aes_acc_num: The number of copies of the AES module that are implemented in
-+ * this version of SEC
-+ **/
-+struct dpseci_sec_attr {
-+ u16 ip_id;
-+ u8 major_rev;
-+ u8 minor_rev;
-+ u8 era;
-+ u8 deco_num;
-+ u8 zuc_auth_acc_num;
-+ u8 zuc_enc_acc_num;
-+ u8 snow_f8_acc_num;
-+ u8 snow_f9_acc_num;
-+ u8 crc_acc_num;
-+ u8 pk_acc_num;
-+ u8 kasumi_acc_num;
-+ u8 rng_acc_num;
-+ u8 md_acc_num;
-+ u8 arc4_acc_num;
-+ u8 des_acc_num;
-+ u8 aes_acc_num;
-+};
-+
-+int dpseci_get_sec_attr(struct fsl_mc_io *mc_io, u32 cmd_flags, u16 token,
-+ struct dpseci_sec_attr *attr);
-+
-+/**
-+ * struct dpseci_sec_counters - Structure representing global SEC counters and
-+ * not per dpseci counters
-+ * @dequeued_requests: Number of Requests Dequeued
-+ * @ob_enc_requests: Number of Outbound Encrypt Requests
-+ * @ib_dec_requests: Number of Inbound Decrypt Requests
-+ * @ob_enc_bytes: Number of Outbound Bytes Encrypted
-+ * @ob_prot_bytes: Number of Outbound Bytes Protected
-+ * @ib_dec_bytes: Number of Inbound Bytes Decrypted
-+ * @ib_valid_bytes: Number of Inbound Bytes Validated
-+ */
-+struct dpseci_sec_counters {
-+ u64 dequeued_requests;
-+ u64 ob_enc_requests;
-+ u64 ib_dec_requests;
-+ u64 ob_enc_bytes;
-+ u64 ob_prot_bytes;
-+ u64 ib_dec_bytes;
-+ u64 ib_valid_bytes;
-+};
-+
-+int dpseci_get_sec_counters(struct fsl_mc_io *mc_io, u32 cmd_flags, u16 token,
-+ struct dpseci_sec_counters *counters);
-+
-+int dpseci_get_api_version(struct fsl_mc_io *mc_io, u32 cmd_flags,
-+ u16 *major_ver, u16 *minor_ver);
-+
-+int dpseci_set_opr(struct fsl_mc_io *mc_io, u32 cmd_flags, u16 token, u8 index,
-+ u8 options, struct opr_cfg *cfg);
-+
-+int dpseci_get_opr(struct fsl_mc_io *mc_io, u32 cmd_flags, u16 token, u8 index,
-+ struct opr_cfg *cfg, struct opr_qry *qry);
-+
-+/**
-+ * enum dpseci_congestion_unit - DPSECI congestion units
-+ * @DPSECI_CONGESTION_UNIT_BYTES: bytes units
-+ * @DPSECI_CONGESTION_UNIT_FRAMES: frames units
-+ */
-+enum dpseci_congestion_unit {
-+ DPSECI_CONGESTION_UNIT_BYTES = 0,
-+ DPSECI_CONGESTION_UNIT_FRAMES
-+};
-+
-+#define DPSECI_CGN_MODE_WRITE_MEM_ON_ENTER 0x00000001
-+#define DPSECI_CGN_MODE_WRITE_MEM_ON_EXIT 0x00000002
-+#define DPSECI_CGN_MODE_COHERENT_WRITE 0x00000004
-+#define DPSECI_CGN_MODE_NOTIFY_DEST_ON_ENTER 0x00000008
-+#define DPSECI_CGN_MODE_NOTIFY_DEST_ON_EXIT 0x00000010
-+#define DPSECI_CGN_MODE_INTR_COALESCING_DISABLED 0x00000020
-+
-+/**
-+ * struct dpseci_congestion_notification_cfg - congestion notification
-+ * configuration
-+ * @units: units type
-+ * @threshold_entry: above this threshold we enter a congestion state.
-+ * set it to '0' to disable it
-+ * @threshold_exit: below this threshold we exit the congestion state.
-+ * @message_ctx: The context that will be part of the CSCN message
-+ * @message_iova: I/O virtual address (must be in DMA-able memory),
-+ * must be 16B aligned;
-+ * @dest_cfg: CSCN can be send to either DPIO or DPCON WQ channel
-+ * @notification_mode: Mask of available options; use 'DPSECI_CGN_MODE_<X>'
-+ * values
-+ */
-+struct dpseci_congestion_notification_cfg {
-+ enum dpseci_congestion_unit units;
-+ u32 threshold_entry;
-+ u32 threshold_exit;
-+ u64 message_ctx;
-+ u64 message_iova;
-+ struct dpseci_dest_cfg dest_cfg;
-+ u16 notification_mode;
-+};
-+
-+int dpseci_set_congestion_notification(struct fsl_mc_io *mc_io, u32 cmd_flags,
-+ u16 token, const struct dpseci_congestion_notification_cfg *cfg);
-+
-+int dpseci_get_congestion_notification(struct fsl_mc_io *mc_io, u32 cmd_flags,
-+ u16 token, struct dpseci_congestion_notification_cfg *cfg);
-+
-+#endif /* _DPSECI_H_ */
---- /dev/null
-+++ b/drivers/crypto/caam/dpseci_cmd.h
-@@ -0,0 +1,261 @@
-+/*
-+ * Copyright 2013-2016 Freescale Semiconductor Inc.
-+ * Copyright 2017 NXP
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the names of the above-listed copyright holders nor the
-+ * names of any contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
-+ * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
-+ * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
-+ * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE
-+ * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
-+ * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
-+ * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
-+ * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
-+ * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
-+ * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
-+ * POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#ifndef _DPSECI_CMD_H_
-+#define _DPSECI_CMD_H_
-+
-+/* DPSECI Version */
-+#define DPSECI_VER_MAJOR 5
-+#define DPSECI_VER_MINOR 1
-+
-+#define DPSECI_VER(maj, min) (((maj) << 16) | (min))
-+#define DPSECI_VERSION DPSECI_VER(DPSECI_VER_MAJOR, DPSECI_VER_MINOR)
-+
-+/* Command IDs */
-+
-+#define DPSECI_CMDID_CLOSE 0x8001
-+#define DPSECI_CMDID_OPEN 0x8091
-+#define DPSECI_CMDID_CREATE 0x9092
-+#define DPSECI_CMDID_DESTROY 0x9891
-+#define DPSECI_CMDID_GET_API_VERSION 0xa091
-+
-+#define DPSECI_CMDID_ENABLE 0x0021
-+#define DPSECI_CMDID_DISABLE 0x0031
-+#define DPSECI_CMDID_GET_ATTR 0x0041
-+#define DPSECI_CMDID_RESET 0x0051
-+#define DPSECI_CMDID_IS_ENABLED 0x0061
-+
-+#define DPSECI_CMDID_SET_IRQ_ENABLE 0x0121
-+#define DPSECI_CMDID_GET_IRQ_ENABLE 0x0131
-+#define DPSECI_CMDID_SET_IRQ_MASK 0x0141
-+#define DPSECI_CMDID_GET_IRQ_MASK 0x0151
-+#define DPSECI_CMDID_GET_IRQ_STATUS 0x0161
-+#define DPSECI_CMDID_CLEAR_IRQ_STATUS 0x0171
-+
-+#define DPSECI_CMDID_SET_RX_QUEUE 0x1941
-+#define DPSECI_CMDID_GET_RX_QUEUE 0x1961
-+#define DPSECI_CMDID_GET_TX_QUEUE 0x1971
-+#define DPSECI_CMDID_GET_SEC_ATTR 0x1981
-+#define DPSECI_CMDID_GET_SEC_COUNTERS 0x1991
-+#define DPSECI_CMDID_SET_OPR 0x19A1
-+#define DPSECI_CMDID_GET_OPR 0x19B1
-+
-+#define DPSECI_CMDID_SET_CONGESTION_NOTIFICATION 0x1701
-+#define DPSECI_CMDID_GET_CONGESTION_NOTIFICATION 0x1711
-+
-+/* Macros for accessing command fields smaller than 1 byte */
-+#define DPSECI_MASK(field) \
-+ GENMASK(DPSECI_##field##_SHIFT + DPSECI_##field##_SIZE - 1, \
-+ DPSECI_##field##_SHIFT)
-+
-+#define dpseci_set_field(var, field, val) \
-+ ((var) |= (((val) << DPSECI_##field##_SHIFT) & DPSECI_MASK(field)))
-+
-+#define dpseci_get_field(var, field) \
-+ (((var) & DPSECI_MASK(field)) >> DPSECI_##field##_SHIFT)
-+
-+struct dpseci_cmd_open {
-+ __le32 dpseci_id;
-+};
-+
-+struct dpseci_cmd_create {
-+ u8 priorities[8];
-+ u8 num_tx_queues;
-+ u8 num_rx_queues;
-+ __le16 pad;
-+ __le32 options;
-+};
-+
-+struct dpseci_cmd_destroy {
-+ __le32 object_id;
-+};
-+
-+struct dpseci_rsp_is_enabled {
-+ __le32 is_enabled;
-+};
-+
-+struct dpseci_cmd_irq_enable {
-+ u8 enable_state;
-+ u8 pad[3];
-+ u8 irq_index;
-+};
-+
-+struct dpseci_rsp_get_irq_enable {
-+ u8 enable_state;
-+};
-+
-+struct dpseci_cmd_irq_mask {
-+ __le32 mask;
-+ u8 irq_index;
-+};
-+
-+struct dpseci_cmd_irq_status {
-+ __le32 status;
-+ u8 irq_index;
-+};
-+
-+struct dpseci_rsp_get_attributes {
-+ __le32 id;
-+ __le32 pad0;
-+ u8 num_tx_queues;
-+ u8 num_rx_queues;
-+ u8 pad1[6];
-+ __le32 options;
-+};
-+
-+struct dpseci_cmd_queue {
-+ __le32 dest_id;
-+ u8 priority;
-+ u8 queue;
-+ u8 dest_type;
-+ u8 pad;
-+ __le64 user_ctx;
-+ union {
-+ __le32 options;
-+ __le32 fqid;
-+ };
-+ __le32 order_preservation_en;
-+};
-+
-+struct dpseci_rsp_get_tx_queue {
-+ __le32 pad;
-+ __le32 fqid;
-+ u8 priority;
-+};
-+
-+struct dpseci_rsp_get_sec_attr {
-+ __le16 ip_id;
-+ u8 major_rev;
-+ u8 minor_rev;
-+ u8 era;
-+ u8 pad0[3];
-+ u8 deco_num;
-+ u8 zuc_auth_acc_num;
-+ u8 zuc_enc_acc_num;
-+ u8 pad1;
-+ u8 snow_f8_acc_num;
-+ u8 snow_f9_acc_num;
-+ u8 crc_acc_num;
-+ u8 pad2;
-+ u8 pk_acc_num;
-+ u8 kasumi_acc_num;
-+ u8 rng_acc_num;
-+ u8 pad3;
-+ u8 md_acc_num;
-+ u8 arc4_acc_num;
-+ u8 des_acc_num;
-+ u8 aes_acc_num;
-+};
-+
-+struct dpseci_rsp_get_sec_counters {
-+ __le64 dequeued_requests;
-+ __le64 ob_enc_requests;
-+ __le64 ib_dec_requests;
-+ __le64 ob_enc_bytes;
-+ __le64 ob_prot_bytes;
-+ __le64 ib_dec_bytes;
-+ __le64 ib_valid_bytes;
-+};
-+
-+struct dpseci_rsp_get_api_version {
-+ __le16 major;
-+ __le16 minor;
-+};
-+
-+struct dpseci_cmd_opr {
-+ __le16 pad;
-+ u8 index;
-+ u8 options;
-+ u8 pad1[7];
-+ u8 oloe;
-+ u8 oeane;
-+ u8 olws;
-+ u8 oa;
-+ u8 oprrws;
-+};
-+
-+#define DPSECI_OPR_RIP_SHIFT 0
-+#define DPSECI_OPR_RIP_SIZE 1
-+#define DPSECI_OPR_ENABLE_SHIFT 1
-+#define DPSECI_OPR_ENABLE_SIZE 1
-+#define DPSECI_OPR_TSEQ_NLIS_SHIFT 1
-+#define DPSECI_OPR_TSEQ_NLIS_SIZE 1
-+#define DPSECI_OPR_HSEQ_NLIS_SHIFT 1
-+#define DPSECI_OPR_HSEQ_NLIS_SIZE 1
-+
-+struct dpseci_rsp_get_opr {
-+ __le64 pad;
-+ u8 rip_enable;
-+ u8 pad0[2];
-+ u8 oloe;
-+ u8 oeane;
-+ u8 olws;
-+ u8 oa;
-+ u8 oprrws;
-+ __le16 nesn;
-+ __le16 pad1;
-+ __le16 ndsn;
-+ __le16 pad2;
-+ __le16 ea_tseq;
-+ u8 tseq_nlis;
-+ u8 pad3;
-+ __le16 ea_hseq;
-+ u8 hseq_nlis;
-+ u8 pad4;
-+ __le16 ea_hptr;
-+ __le16 pad5;
-+ __le16 ea_tptr;
-+ __le16 pad6;
-+ __le16 opr_vid;
-+ __le16 pad7;
-+ __le16 opr_id;
-+};
-+
-+#define DPSECI_CGN_DEST_TYPE_SHIFT 0
-+#define DPSECI_CGN_DEST_TYPE_SIZE 4
-+#define DPSECI_CGN_UNITS_SHIFT 4
-+#define DPSECI_CGN_UNITS_SIZE 2
-+
-+struct dpseci_cmd_congestion_notification {
-+ __le32 dest_id;
-+ __le16 notification_mode;
-+ u8 priority;
-+ u8 options;
-+ __le64 message_iova;
-+ __le64 message_ctx;
-+ __le32 threshold_entry;
-+ __le32 threshold_exit;
-+};
-+
-+#endif /* _DPSECI_CMD_H_ */
---- a/drivers/crypto/caam/error.c
-+++ b/drivers/crypto/caam/error.c
-@@ -6,11 +6,54 @@
-
- #include "compat.h"
- #include "regs.h"
--#include "intern.h"
- #include "desc.h"
--#include "jr.h"
- #include "error.h"
-
-+#ifdef DEBUG
-+
-+#include <linux/highmem.h>
-+
-+void caam_dump_sg(const char *level, const char *prefix_str, int prefix_type,
-+ int rowsize, int groupsize, struct scatterlist *sg,
-+ size_t tlen, bool ascii)
-+{
-+ struct scatterlist *it;
-+ void *it_page;
-+ size_t len;
-+ void *buf;
-+
-+ for (it = sg; it && tlen > 0 ; it = sg_next(sg)) {
-+ /*
-+ * make sure the scatterlist's page
-+ * has a valid virtual memory mapping
-+ */
-+ it_page = kmap_atomic(sg_page(it));
-+ if (unlikely(!it_page)) {
-+ pr_err("caam_dump_sg: kmap failed\n");
-+ return;
-+ }
-+
-+ buf = it_page + it->offset;
-+ len = min_t(size_t, tlen, it->length);
-+ print_hex_dump(level, prefix_str, prefix_type, rowsize,
-+ groupsize, buf, len, ascii);
-+ tlen -= len;
-+
-+ kunmap_atomic(it_page);
-+ }
-+}
-+
-+#else
-+
-+void caam_dump_sg(const char *level, const char *prefix_str, int prefix_type,
-+ int rowsize, int groupsize, struct scatterlist *sg,
-+ size_t tlen, bool ascii)
-+{}
-+
-+#endif
-+
-+EXPORT_SYMBOL(caam_dump_sg);
-+
- static const struct {
- u8 value;
- const char *error_text;
-@@ -69,6 +112,54 @@ static const struct {
- { 0xF1, "3GPP HFN matches or exceeds the Threshold" },
- };
-
-+static const struct {
-+ u8 value;
-+ const char *error_text;
-+} qi_error_list[] = {
-+ { 0x1F, "Job terminated by FQ or ICID flush" },
-+ { 0x20, "FD format error"},
-+ { 0x21, "FD command format error"},
-+ { 0x23, "FL format error"},
-+ { 0x25, "CRJD specified in FD, but not enabled in FLC"},
-+ { 0x30, "Max. buffer size too small"},
-+ { 0x31, "DHR exceeds max. buffer size (allocate mode, S/G format)"},
-+ { 0x32, "SGT exceeds max. buffer size (allocate mode, S/G format"},
-+ { 0x33, "Size over/underflow (allocate mode)"},
-+ { 0x34, "Size over/underflow (reuse mode)"},
-+ { 0x35, "Length exceeds max. short length (allocate mode, S/G/ format)"},
-+ { 0x36, "Memory footprint exceeds max. value (allocate mode, S/G/ format)"},
-+ { 0x41, "SBC frame format not supported (allocate mode)"},
-+ { 0x42, "Pool 0 invalid / pool 1 size < pool 0 size (allocate mode)"},
-+ { 0x43, "Annotation output enabled but ASAR = 0 (allocate mode)"},
-+ { 0x44, "Unsupported or reserved frame format or SGHR = 1 (reuse mode)"},
-+ { 0x45, "DHR correction underflow (reuse mode, single buffer format)"},
-+ { 0x46, "Annotation length exceeds offset (reuse mode)"},
-+ { 0x48, "Annotation output enabled but ASA limited by ASAR (reuse mode)"},
-+ { 0x49, "Data offset correction exceeds input frame data length (reuse mode)"},
-+ { 0x4B, "Annotation output enabled but ASA cannote be expanded (frame list)"},
-+ { 0x51, "Unsupported IF reuse mode"},
-+ { 0x52, "Unsupported FL use mode"},
-+ { 0x53, "Unsupported RJD use mode"},
-+ { 0x54, "Unsupported inline descriptor use mode"},
-+ { 0xC0, "Table buffer pool 0 depletion"},
-+ { 0xC1, "Table buffer pool 1 depletion"},
-+ { 0xC2, "Data buffer pool 0 depletion, no OF allocated"},
-+ { 0xC3, "Data buffer pool 1 depletion, no OF allocated"},
-+ { 0xC4, "Data buffer pool 0 depletion, partial OF allocated"},
-+ { 0xC5, "Data buffer pool 1 depletion, partial OF allocated"},
-+ { 0xD0, "FLC read error"},
-+ { 0xD1, "FL read error"},
-+ { 0xD2, "FL write error"},
-+ { 0xD3, "OF SGT write error"},
-+ { 0xD4, "PTA read error"},
-+ { 0xD5, "PTA write error"},
-+ { 0xD6, "OF SGT F-bit write error"},
-+ { 0xD7, "ASA write error"},
-+ { 0xE1, "FLC[ICR]=0 ICID error"},
-+ { 0xE2, "FLC[ICR]=1 ICID error"},
-+ { 0xE4, "source of ICID flush not trusted (BDI = 0)"},
-+};
-+
- static const char * const cha_id_list[] = {
- "",
- "AES",
-@@ -146,10 +237,9 @@ static void report_ccb_status(struct dev
- strlen(rng_err_id_list[err_id])) {
- /* RNG-only error */
- err_str = rng_err_id_list[err_id];
-- } else if (err_id < ARRAY_SIZE(err_id_list))
-+ } else {
- err_str = err_id_list[err_id];
-- else
-- snprintf(err_err_code, sizeof(err_err_code), "%02x", err_id);
-+ }
-
- /*
- * CCB ICV check failures are part of normal operation life;
-@@ -198,6 +288,27 @@ static void report_deco_status(struct de
- status, error, idx_str, idx, err_str, err_err_code);
- }
-
-+static void report_qi_status(struct device *qidev, const u32 status,
-+ const char *error)
-+{
-+ u8 err_id = status & JRSTA_QIERR_ERROR_MASK;
-+ const char *err_str = "unidentified error value 0x";
-+ char err_err_code[3] = { 0 };
-+ int i;
-+
-+ for (i = 0; i < ARRAY_SIZE(qi_error_list); i++)
-+ if (qi_error_list[i].value == err_id)
-+ break;
-+
-+ if (i != ARRAY_SIZE(qi_error_list) && qi_error_list[i].error_text)
-+ err_str = qi_error_list[i].error_text;
-+ else
-+ snprintf(err_err_code, sizeof(err_err_code), "%02x", err_id);
-+
-+ dev_err(qidev, "%08x: %s: %s%s\n",
-+ status, error, err_str, err_err_code);
-+}
-+
- static void report_jr_status(struct device *jrdev, const u32 status,
- const char *error)
- {
-@@ -212,7 +323,7 @@ static void report_cond_code_status(stru
- status, error, __func__);
- }
-
--void caam_jr_strstatus(struct device *jrdev, u32 status)
-+void caam_strstatus(struct device *jrdev, u32 status, bool qi_v2)
- {
- static const struct stat_src {
- void (*report_ssed)(struct device *jrdev, const u32 status,
-@@ -224,7 +335,7 @@ void caam_jr_strstatus(struct device *jr
- { report_ccb_status, "CCB" },
- { report_jump_status, "Jump" },
- { report_deco_status, "DECO" },
-- { NULL, "Queue Manager Interface" },
-+ { report_qi_status, "Queue Manager Interface" },
- { report_jr_status, "Job Ring" },
- { report_cond_code_status, "Condition Code" },
- { NULL, NULL },
-@@ -250,4 +361,4 @@ void caam_jr_strstatus(struct device *jr
- else
- dev_err(jrdev, "%d: unknown error source\n", ssrc);
- }
--EXPORT_SYMBOL(caam_jr_strstatus);
-+EXPORT_SYMBOL(caam_strstatus);
---- a/drivers/crypto/caam/error.h
-+++ b/drivers/crypto/caam/error.h
-@@ -7,5 +7,13 @@
- #ifndef CAAM_ERROR_H
- #define CAAM_ERROR_H
- #define CAAM_ERROR_STR_MAX 302
--void caam_jr_strstatus(struct device *jrdev, u32 status);
-+
-+void caam_strstatus(struct device *dev, u32 status, bool qi_v2);
-+
-+#define caam_jr_strstatus(jrdev, status) caam_strstatus(jrdev, status, false)
-+#define caam_qi2_strstatus(qidev, status) caam_strstatus(qidev, status, true)
-+
-+void caam_dump_sg(const char *level, const char *prefix_str, int prefix_type,
-+ int rowsize, int groupsize, struct scatterlist *sg,
-+ size_t tlen, bool ascii);
- #endif /* CAAM_ERROR_H */
---- a/drivers/crypto/caam/intern.h
-+++ b/drivers/crypto/caam/intern.h
-@@ -64,10 +64,9 @@ struct caam_drv_private_jr {
- * Driver-private storage for a single CAAM block instance
- */
- struct caam_drv_private {
--
-- struct device *dev;
-- struct platform_device **jrpdev; /* Alloc'ed array per sub-device */
-- struct platform_device *pdev;
-+#ifdef CONFIG_CAAM_QI
-+ struct device *qidev;
-+#endif
-
- /* Physical-presence section */
- struct caam_ctrl __iomem *ctrl; /* controller region */
-@@ -84,6 +83,7 @@ struct caam_drv_private {
- u8 qi_present; /* Nonzero if QI present in device */
- int secvio_irq; /* Security violation interrupt number */
- int virt_en; /* Virtualization enabled in CAAM */
-+ int era; /* CAAM Era (internal HW revision) */
-
- #define RNG4_MAX_HANDLES 2
- /* RNG4 block */
-@@ -103,11 +103,6 @@ struct caam_drv_private {
- #ifdef CONFIG_DEBUG_FS
- struct dentry *dfs_root;
- struct dentry *ctl; /* controller dir */
-- struct dentry *ctl_rq_dequeued, *ctl_ob_enc_req, *ctl_ib_dec_req;
-- struct dentry *ctl_ob_enc_bytes, *ctl_ob_prot_bytes;
-- struct dentry *ctl_ib_dec_bytes, *ctl_ib_valid_bytes;
-- struct dentry *ctl_faultaddr, *ctl_faultdetail, *ctl_faultstatus;
--
- struct debugfs_blob_wrapper ctl_kek_wrap, ctl_tkek_wrap, ctl_tdsk_wrap;
- struct dentry *ctl_kek, *ctl_tkek, *ctl_tdsk;
- #endif
-@@ -115,4 +110,22 @@ struct caam_drv_private {
-
- void caam_jr_algapi_init(struct device *dev);
- void caam_jr_algapi_remove(struct device *dev);
-+
-+#ifdef CONFIG_DEBUG_FS
-+static int caam_debugfs_u64_get(void *data, u64 *val)
-+{
-+ *val = caam64_to_cpu(*(u64 *)data);
-+ return 0;
-+}
-+
-+static int caam_debugfs_u32_get(void *data, u64 *val)
-+{
-+ *val = caam32_to_cpu(*(u32 *)data);
-+ return 0;
-+}
-+
-+DEFINE_SIMPLE_ATTRIBUTE(caam_fops_u32_ro, caam_debugfs_u32_get, NULL, "%llu\n");
-+DEFINE_SIMPLE_ATTRIBUTE(caam_fops_u64_ro, caam_debugfs_u64_get, NULL, "%llu\n");
-+#endif
-+
- #endif /* INTERN_H */
---- a/drivers/crypto/caam/jr.c
-+++ b/drivers/crypto/caam/jr.c
-@@ -9,6 +9,7 @@
- #include <linux/of_address.h>
-
- #include "compat.h"
-+#include "ctrl.h"
- #include "regs.h"
- #include "jr.h"
- #include "desc.h"
-@@ -22,6 +23,14 @@ struct jr_driver_data {
-
- static struct jr_driver_data driver_data;
-
-+static int jr_driver_probed;
-+
-+int caam_jr_driver_probed(void)
-+{
-+ return jr_driver_probed;
-+}
-+EXPORT_SYMBOL(caam_jr_driver_probed);
-+
- static int caam_reset_hw_jr(struct device *dev)
- {
- struct caam_drv_private_jr *jrp = dev_get_drvdata(dev);
-@@ -118,6 +127,8 @@ static int caam_jr_remove(struct platfor
- dev_err(jrdev, "Failed to shut down job ring\n");
- irq_dispose_mapping(jrpriv->irq);
-
-+ jr_driver_probed--;
-+
- return ret;
- }
-
-@@ -281,6 +292,36 @@ struct device *caam_jr_alloc(void)
- EXPORT_SYMBOL(caam_jr_alloc);
-
- /**
-+ * caam_jridx_alloc() - Alloc a specific job ring based on its index.
-+ *
-+ * returns : pointer to the newly allocated physical
-+ * JobR dev can be written to if successful.
-+ **/
-+struct device *caam_jridx_alloc(int idx)
-+{
-+ struct caam_drv_private_jr *jrpriv;
-+ struct device *dev = ERR_PTR(-ENODEV);
-+
-+ spin_lock(&driver_data.jr_alloc_lock);
-+
-+ if (list_empty(&driver_data.jr_list))
-+ goto end;
-+
-+ list_for_each_entry(jrpriv, &driver_data.jr_list, list_node) {
-+ if (jrpriv->ridx == idx) {
-+ atomic_inc(&jrpriv->tfm_count);
-+ dev = jrpriv->dev;
-+ break;
-+ }
-+ }
-+
-+end:
-+ spin_unlock(&driver_data.jr_alloc_lock);
-+ return dev;
-+}
-+EXPORT_SYMBOL(caam_jridx_alloc);
-+
-+/**
- * caam_jr_free() - Free the Job Ring
- * @rdev - points to the dev that identifies the Job ring to
- * be released.
-@@ -497,15 +538,28 @@ static int caam_jr_probe(struct platform
- return -ENOMEM;
- }
-
-- jrpriv->rregs = (struct caam_job_ring __force *)ctrl;
-+ jrpriv->rregs = (struct caam_job_ring __iomem __force *)ctrl;
-
-- if (sizeof(dma_addr_t) == sizeof(u64))
-- if (of_device_is_compatible(nprop, "fsl,sec-v5.0-job-ring"))
-- dma_set_mask_and_coherent(jrdev, DMA_BIT_MASK(40));
-+ if (sizeof(dma_addr_t) == sizeof(u64)) {
-+ if (caam_dpaa2)
-+ error = dma_set_mask_and_coherent(jrdev,
-+ DMA_BIT_MASK(49));
-+ else if (of_device_is_compatible(nprop,
-+ "fsl,sec-v5.0-job-ring"))
-+ error = dma_set_mask_and_coherent(jrdev,
-+ DMA_BIT_MASK(40));
- else
-- dma_set_mask_and_coherent(jrdev, DMA_BIT_MASK(36));
-- else
-- dma_set_mask_and_coherent(jrdev, DMA_BIT_MASK(32));
-+ error = dma_set_mask_and_coherent(jrdev,
-+ DMA_BIT_MASK(36));
-+ } else {
-+ error = dma_set_mask_and_coherent(jrdev, DMA_BIT_MASK(32));
-+ }
-+ if (error) {
-+ dev_err(jrdev, "dma_set_mask_and_coherent failed (%d)\n",
-+ error);
-+ iounmap(ctrl);
-+ return error;
-+ }
-
- /* Identify the interrupt */
- jrpriv->irq = irq_of_parse_and_map(nprop, 0);
-@@ -525,10 +579,12 @@ static int caam_jr_probe(struct platform
-
- atomic_set(&jrpriv->tfm_count, 0);
-
-+ jr_driver_probed++;
-+
- return 0;
- }
-
--static struct of_device_id caam_jr_match[] = {
-+static const struct of_device_id caam_jr_match[] = {
- {
- .compatible = "fsl,sec-v4.0-job-ring",
- },
---- a/drivers/crypto/caam/jr.h
-+++ b/drivers/crypto/caam/jr.h
-@@ -8,7 +8,9 @@
- #define JR_H
-
- /* Prototypes for backend-level services exposed to APIs */
-+int caam_jr_driver_probed(void);
- struct device *caam_jr_alloc(void);
-+struct device *caam_jridx_alloc(int idx);
- void caam_jr_free(struct device *rdev);
- int caam_jr_enqueue(struct device *dev, u32 *desc,
- void (*cbk)(struct device *dev, u32 *desc, u32 status,
---- a/drivers/crypto/caam/key_gen.c
-+++ b/drivers/crypto/caam/key_gen.c
-@@ -41,15 +41,29 @@ Split key generation--------------------
- [06] 0x64260028 fifostr: class2 mdsplit-jdk len=40
- @0xffe04000
- */
--int gen_split_key(struct device *jrdev, u8 *key_out, int split_key_len,
-- int split_key_pad_len, const u8 *key_in, u32 keylen,
-- u32 alg_op)
-+int gen_split_key(struct device *jrdev, u8 *key_out,
-+ struct alginfo * const adata, const u8 *key_in, u32 keylen,
-+ int max_keylen)
- {
- u32 *desc;
- struct split_key_result result;
- dma_addr_t dma_addr_in, dma_addr_out;
- int ret = -ENOMEM;
-
-+ adata->keylen = split_key_len(adata->algtype & OP_ALG_ALGSEL_MASK);
-+ adata->keylen_pad = split_key_pad_len(adata->algtype &
-+ OP_ALG_ALGSEL_MASK);
-+
-+#ifdef DEBUG
-+ dev_err(jrdev, "split keylen %d split keylen padded %d\n",
-+ adata->keylen, adata->keylen_pad);
-+ print_hex_dump(KERN_ERR, "ctx.key@" __stringify(__LINE__)": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, key_in, keylen, 1);
-+#endif
-+
-+ if (adata->keylen_pad > max_keylen)
-+ return -EINVAL;
-+
- desc = kmalloc(CAAM_CMD_SZ * 6 + CAAM_PTR_SZ * 2, GFP_KERNEL | GFP_DMA);
- if (!desc) {
- dev_err(jrdev, "unable to allocate key input memory\n");
-@@ -63,7 +77,7 @@ int gen_split_key(struct device *jrdev,
- goto out_free;
- }
-
-- dma_addr_out = dma_map_single(jrdev, key_out, split_key_pad_len,
-+ dma_addr_out = dma_map_single(jrdev, key_out, adata->keylen_pad,
- DMA_FROM_DEVICE);
- if (dma_mapping_error(jrdev, dma_addr_out)) {
- dev_err(jrdev, "unable to map key output memory\n");
-@@ -74,7 +88,9 @@ int gen_split_key(struct device *jrdev,
- append_key(desc, dma_addr_in, keylen, CLASS_2 | KEY_DEST_CLASS_REG);
-
- /* Sets MDHA up into an HMAC-INIT */
-- append_operation(desc, alg_op | OP_ALG_DECRYPT | OP_ALG_AS_INIT);
-+ append_operation(desc, (adata->algtype & OP_ALG_ALGSEL_MASK) |
-+ OP_ALG_AAI_HMAC | OP_TYPE_CLASS2_ALG | OP_ALG_DECRYPT |
-+ OP_ALG_AS_INIT);
-
- /*
- * do a FIFO_LOAD of zero, this will trigger the internal key expansion
-@@ -87,7 +103,7 @@ int gen_split_key(struct device *jrdev,
- * FIFO_STORE with the explicit split-key content store
- * (0x26 output type)
- */
-- append_fifo_store(desc, dma_addr_out, split_key_len,
-+ append_fifo_store(desc, dma_addr_out, adata->keylen,
- LDST_CLASS_2_CCB | FIFOST_TYPE_SPLIT_KEK);
-
- #ifdef DEBUG
-@@ -108,11 +124,11 @@ int gen_split_key(struct device *jrdev,
- #ifdef DEBUG
- print_hex_dump(KERN_ERR, "ctx.key@"__stringify(__LINE__)": ",
- DUMP_PREFIX_ADDRESS, 16, 4, key_out,
-- split_key_pad_len, 1);
-+ adata->keylen_pad, 1);
- #endif
- }
-
-- dma_unmap_single(jrdev, dma_addr_out, split_key_pad_len,
-+ dma_unmap_single(jrdev, dma_addr_out, adata->keylen_pad,
- DMA_FROM_DEVICE);
- out_unmap_in:
- dma_unmap_single(jrdev, dma_addr_in, keylen, DMA_TO_DEVICE);
---- a/drivers/crypto/caam/key_gen.h
-+++ b/drivers/crypto/caam/key_gen.h
-@@ -5,6 +5,36 @@
- *
- */
-
-+/**
-+ * split_key_len - Compute MDHA split key length for a given algorithm
-+ * @hash: Hashing algorithm selection, one of OP_ALG_ALGSEL_* - MD5, SHA1,
-+ * SHA224, SHA384, SHA512.
-+ *
-+ * Return: MDHA split key length
-+ */
-+static inline u32 split_key_len(u32 hash)
-+{
-+ /* Sizes for MDHA pads (*not* keys): MD5, SHA1, 224, 256, 384, 512 */
-+ static const u8 mdpadlen[] = { 16, 20, 32, 32, 64, 64 };
-+ u32 idx;
-+
-+ idx = (hash & OP_ALG_ALGSEL_SUBMASK) >> OP_ALG_ALGSEL_SHIFT;
-+
-+ return (u32)(mdpadlen[idx] * 2);
-+}
-+
-+/**
-+ * split_key_pad_len - Compute MDHA split key pad length for a given algorithm
-+ * @hash: Hashing algorithm selection, one of OP_ALG_ALGSEL_* - MD5, SHA1,
-+ * SHA224, SHA384, SHA512.
-+ *
-+ * Return: MDHA split key pad length
-+ */
-+static inline u32 split_key_pad_len(u32 hash)
-+{
-+ return ALIGN(split_key_len(hash), 16);
-+}
-+
- struct split_key_result {
- struct completion completion;
- int err;
-@@ -12,6 +42,6 @@ struct split_key_result {
-
- void split_key_done(struct device *dev, u32 *desc, u32 err, void *context);
-
--int gen_split_key(struct device *jrdev, u8 *key_out, int split_key_len,
-- int split_key_pad_len, const u8 *key_in, u32 keylen,
-- u32 alg_op);
-+int gen_split_key(struct device *jrdev, u8 *key_out,
-+ struct alginfo * const adata, const u8 *key_in, u32 keylen,
-+ int max_keylen);
---- a/drivers/crypto/caam/pdb.h
-+++ b/drivers/crypto/caam/pdb.h
-@@ -483,6 +483,8 @@ struct dsa_verify_pdb {
- #define RSA_PDB_E_MASK (0xFFF << RSA_PDB_E_SHIFT)
- #define RSA_PDB_D_SHIFT 12
- #define RSA_PDB_D_MASK (0xFFF << RSA_PDB_D_SHIFT)
-+#define RSA_PDB_Q_SHIFT 12
-+#define RSA_PDB_Q_MASK (0xFFF << RSA_PDB_Q_SHIFT)
-
- #define RSA_PDB_SGF_F (0x8 << RSA_PDB_SGF_SHIFT)
- #define RSA_PDB_SGF_G (0x4 << RSA_PDB_SGF_SHIFT)
-@@ -490,6 +492,8 @@ struct dsa_verify_pdb {
- #define RSA_PRIV_PDB_SGF_G (0x8 << RSA_PDB_SGF_SHIFT)
-
- #define RSA_PRIV_KEY_FRM_1 0
-+#define RSA_PRIV_KEY_FRM_2 1
-+#define RSA_PRIV_KEY_FRM_3 2
-
- /**
- * RSA Encrypt Protocol Data Block
-@@ -525,4 +529,62 @@ struct rsa_priv_f1_pdb {
- dma_addr_t d_dma;
- } __packed;
-
-+/**
-+ * RSA Decrypt PDB - Private Key Form #2
-+ * @sgf : scatter-gather field
-+ * @g_dma : dma address of encrypted input data
-+ * @f_dma : dma address of output data
-+ * @d_dma : dma address of RSA private exponent
-+ * @p_dma : dma address of RSA prime factor p of RSA modulus n
-+ * @q_dma : dma address of RSA prime factor q of RSA modulus n
-+ * @tmp1_dma: dma address of temporary buffer. CAAM uses this temporary buffer
-+ * as internal state buffer. It is assumed to be as long as p.
-+ * @tmp2_dma: dma address of temporary buffer. CAAM uses this temporary buffer
-+ * as internal state buffer. It is assumed to be as long as q.
-+ * @p_q_len : length in bytes of first two prime factors of the RSA modulus n
-+ */
-+struct rsa_priv_f2_pdb {
-+ u32 sgf;
-+ dma_addr_t g_dma;
-+ dma_addr_t f_dma;
-+ dma_addr_t d_dma;
-+ dma_addr_t p_dma;
-+ dma_addr_t q_dma;
-+ dma_addr_t tmp1_dma;
-+ dma_addr_t tmp2_dma;
-+ u32 p_q_len;
-+} __packed;
-+
-+/**
-+ * RSA Decrypt PDB - Private Key Form #3
-+ * This is the RSA Chinese Reminder Theorem (CRT) form for two prime factors of
-+ * the RSA modulus.
-+ * @sgf : scatter-gather field
-+ * @g_dma : dma address of encrypted input data
-+ * @f_dma : dma address of output data
-+ * @c_dma : dma address of RSA CRT coefficient
-+ * @p_dma : dma address of RSA prime factor p of RSA modulus n
-+ * @q_dma : dma address of RSA prime factor q of RSA modulus n
-+ * @dp_dma : dma address of RSA CRT exponent of RSA prime factor p
-+ * @dp_dma : dma address of RSA CRT exponent of RSA prime factor q
-+ * @tmp1_dma: dma address of temporary buffer. CAAM uses this temporary buffer
-+ * as internal state buffer. It is assumed to be as long as p.
-+ * @tmp2_dma: dma address of temporary buffer. CAAM uses this temporary buffer
-+ * as internal state buffer. It is assumed to be as long as q.
-+ * @p_q_len : length in bytes of first two prime factors of the RSA modulus n
-+ */
-+struct rsa_priv_f3_pdb {
-+ u32 sgf;
-+ dma_addr_t g_dma;
-+ dma_addr_t f_dma;
-+ dma_addr_t c_dma;
-+ dma_addr_t p_dma;
-+ dma_addr_t q_dma;
-+ dma_addr_t dp_dma;
-+ dma_addr_t dq_dma;
-+ dma_addr_t tmp1_dma;
-+ dma_addr_t tmp2_dma;
-+ u32 p_q_len;
-+} __packed;
-+
- #endif
---- a/drivers/crypto/caam/pkc_desc.c
-+++ b/drivers/crypto/caam/pkc_desc.c
-@@ -34,3 +34,39 @@ void init_rsa_priv_f1_desc(u32 *desc, st
- append_operation(desc, OP_TYPE_UNI_PROTOCOL | OP_PCLID_RSADEC_PRVKEY |
- RSA_PRIV_KEY_FRM_1);
- }
-+
-+/* Descriptor for RSA Private operation - Private Key Form #2 */
-+void init_rsa_priv_f2_desc(u32 *desc, struct rsa_priv_f2_pdb *pdb)
-+{
-+ init_job_desc_pdb(desc, 0, sizeof(*pdb));
-+ append_cmd(desc, pdb->sgf);
-+ append_ptr(desc, pdb->g_dma);
-+ append_ptr(desc, pdb->f_dma);
-+ append_ptr(desc, pdb->d_dma);
-+ append_ptr(desc, pdb->p_dma);
-+ append_ptr(desc, pdb->q_dma);
-+ append_ptr(desc, pdb->tmp1_dma);
-+ append_ptr(desc, pdb->tmp2_dma);
-+ append_cmd(desc, pdb->p_q_len);
-+ append_operation(desc, OP_TYPE_UNI_PROTOCOL | OP_PCLID_RSADEC_PRVKEY |
-+ RSA_PRIV_KEY_FRM_2);
-+}
-+
-+/* Descriptor for RSA Private operation - Private Key Form #3 */
-+void init_rsa_priv_f3_desc(u32 *desc, struct rsa_priv_f3_pdb *pdb)
-+{
-+ init_job_desc_pdb(desc, 0, sizeof(*pdb));
-+ append_cmd(desc, pdb->sgf);
-+ append_ptr(desc, pdb->g_dma);
-+ append_ptr(desc, pdb->f_dma);
-+ append_ptr(desc, pdb->c_dma);
-+ append_ptr(desc, pdb->p_dma);
-+ append_ptr(desc, pdb->q_dma);
-+ append_ptr(desc, pdb->dp_dma);
-+ append_ptr(desc, pdb->dq_dma);
-+ append_ptr(desc, pdb->tmp1_dma);
-+ append_ptr(desc, pdb->tmp2_dma);
-+ append_cmd(desc, pdb->p_q_len);
-+ append_operation(desc, OP_TYPE_UNI_PROTOCOL | OP_PCLID_RSADEC_PRVKEY |
-+ RSA_PRIV_KEY_FRM_3);
-+}
---- /dev/null
-+++ b/drivers/crypto/caam/qi.c
-@@ -0,0 +1,804 @@
-+/*
-+ * CAAM/SEC 4.x QI transport/backend driver
-+ * Queue Interface backend functionality
-+ *
-+ * Copyright 2013-2016 Freescale Semiconductor, Inc.
-+ * Copyright 2016-2017 NXP
-+ */
-+
-+#include <linux/cpumask.h>
-+#include <linux/kthread.h>
-+#include <linux/fsl_qman.h>
-+
-+#include "regs.h"
-+#include "qi.h"
-+#include "desc.h"
-+#include "intern.h"
-+#include "desc_constr.h"
-+
-+#define PREHDR_RSLS_SHIFT 31
-+
-+/*
-+ * Use a reasonable backlog of frames (per CPU) as congestion threshold,
-+ * so that resources used by the in-flight buffers do not become a memory hog.
-+ */
-+#define MAX_RSP_FQ_BACKLOG_PER_CPU 256
-+
-+#define CAAM_QI_ENQUEUE_RETRIES 10000
-+
-+#define CAAM_NAPI_WEIGHT 63
-+
-+/*
-+ * caam_napi - struct holding CAAM NAPI-related params
-+ * @irqtask: IRQ task for QI backend
-+ * @p: QMan portal
-+ */
-+struct caam_napi {
-+ struct napi_struct irqtask;
-+ struct qman_portal *p;
-+};
-+
-+/*
-+ * caam_qi_pcpu_priv - percpu private data structure to main list of pending
-+ * responses expected on each cpu.
-+ * @caam_napi: CAAM NAPI params
-+ * @net_dev: netdev used by NAPI
-+ * @rsp_fq: response FQ from CAAM
-+ */
-+struct caam_qi_pcpu_priv {
-+ struct caam_napi caam_napi;
-+ struct net_device net_dev;
-+ struct qman_fq *rsp_fq;
-+} ____cacheline_aligned;
-+
-+static DEFINE_PER_CPU(struct caam_qi_pcpu_priv, pcpu_qipriv);
-+static DEFINE_PER_CPU(int, last_cpu);
-+
-+/*
-+ * caam_qi_priv - CAAM QI backend private params
-+ * @cgr: QMan congestion group
-+ * @qi_pdev: platform device for QI backend
-+ */
-+struct caam_qi_priv {
-+ struct qman_cgr cgr;
-+ struct platform_device *qi_pdev;
-+};
-+
-+static struct caam_qi_priv qipriv ____cacheline_aligned;
-+
-+/*
-+ * This is written by only one core - the one that initialized the CGR - and
-+ * read by multiple cores (all the others).
-+ */
-+bool caam_congested __read_mostly;
-+EXPORT_SYMBOL(caam_congested);
-+
-+#ifdef CONFIG_DEBUG_FS
-+/*
-+ * This is a counter for the number of times the congestion group (where all
-+ * the request and response queueus are) reached congestion. Incremented
-+ * each time the congestion callback is called with congested == true.
-+ */
-+static u64 times_congested;
-+#endif
-+
-+/*
-+ * CPU from where the module initialised. This is required because QMan driver
-+ * requires CGRs to be removed from same CPU from where they were originally
-+ * allocated.
-+ */
-+static int mod_init_cpu;
-+
-+/*
-+ * This is a a cache of buffers, from which the users of CAAM QI driver
-+ * can allocate short (CAAM_QI_MEMCACHE_SIZE) buffers. It's faster than
-+ * doing malloc on the hotpath.
-+ * NOTE: A more elegant solution would be to have some headroom in the frames
-+ * being processed. This could be added by the dpaa-ethernet driver.
-+ * This would pose a problem for userspace application processing which
-+ * cannot know of this limitation. So for now, this will work.
-+ * NOTE: The memcache is SMP-safe. No need to handle spinlocks in-here
-+ */
-+static struct kmem_cache *qi_cache;
-+
-+int caam_qi_enqueue(struct device *qidev, struct caam_drv_req *req)
-+{
-+ struct qm_fd fd;
-+ int ret;
-+ int num_retries = 0;
-+
-+ fd.cmd = 0;
-+ fd.format = qm_fd_compound;
-+ fd.cong_weight = caam32_to_cpu(req->fd_sgt[1].length);
-+ fd.addr = dma_map_single(qidev, req->fd_sgt, sizeof(req->fd_sgt),
-+ DMA_BIDIRECTIONAL);
-+ if (dma_mapping_error(qidev, fd.addr)) {
-+ dev_err(qidev, "DMA mapping error for QI enqueue request\n");
-+ return -EIO;
-+ }
-+
-+ do {
-+ ret = qman_enqueue(req->drv_ctx->req_fq, &fd, 0);
-+ if (likely(!ret))
-+ return 0;
-+
-+ if (ret != -EBUSY)
-+ break;
-+ num_retries++;
-+ } while (num_retries < CAAM_QI_ENQUEUE_RETRIES);
-+
-+ dev_err(qidev, "qman_enqueue failed: %d\n", ret);
-+
-+ return ret;
-+}
-+EXPORT_SYMBOL(caam_qi_enqueue);
-+
-+static void caam_fq_ern_cb(struct qman_portal *qm, struct qman_fq *fq,
-+ const struct qm_mr_entry *msg)
-+{
-+ const struct qm_fd *fd;
-+ struct caam_drv_req *drv_req;
-+ struct device *qidev = &(raw_cpu_ptr(&pcpu_qipriv)->net_dev.dev);
-+
-+ fd = &msg->ern.fd;
-+
-+ if (fd->format != qm_fd_compound) {
-+ dev_err(qidev, "Non-compound FD from CAAM\n");
-+ return;
-+ }
-+
-+ drv_req = (struct caam_drv_req *)phys_to_virt(qm_fd_addr_get64(fd));
-+ if (!drv_req) {
-+ dev_err(qidev,
-+ "Can't find original request for CAAM response\n");
-+ return;
-+ }
-+
-+ dma_unmap_single(drv_req->drv_ctx->qidev, qm_fd_addr(fd),
-+ sizeof(drv_req->fd_sgt), DMA_BIDIRECTIONAL);
-+
-+ drv_req->cbk(drv_req, -EIO);
-+}
-+
-+static struct qman_fq *create_caam_req_fq(struct device *qidev,
-+ struct qman_fq *rsp_fq,
-+ dma_addr_t hwdesc,
-+ int fq_sched_flag)
-+{
-+ int ret;
-+ struct qman_fq *req_fq;
-+ struct qm_mcc_initfq opts;
-+
-+ req_fq = kzalloc(sizeof(*req_fq), GFP_ATOMIC);
-+ if (!req_fq)
-+ return ERR_PTR(-ENOMEM);
-+
-+ req_fq->cb.ern = caam_fq_ern_cb;
-+ req_fq->cb.fqs = NULL;
-+
-+ ret = qman_create_fq(0, QMAN_FQ_FLAG_DYNAMIC_FQID |
-+ QMAN_FQ_FLAG_TO_DCPORTAL | QMAN_FQ_FLAG_LOCKED,
-+ req_fq);
-+ if (ret) {
-+ dev_err(qidev, "Failed to create session req FQ\n");
-+ goto create_req_fq_fail;
-+ }
-+
-+ opts.we_mask = QM_INITFQ_WE_FQCTRL | QM_INITFQ_WE_DESTWQ |
-+ QM_INITFQ_WE_CONTEXTB | QM_INITFQ_WE_CONTEXTA |
-+ QM_INITFQ_WE_CGID;
-+ opts.fqd.fq_ctrl = QM_FQCTRL_CPCSTASH | QM_FQCTRL_CGE;
-+ opts.fqd.dest.channel = qm_channel_caam;
-+ opts.fqd.dest.wq = 2;
-+ opts.fqd.context_b = qman_fq_fqid(rsp_fq);
-+ opts.fqd.context_a.hi = upper_32_bits(hwdesc);
-+ opts.fqd.context_a.lo = lower_32_bits(hwdesc);
-+ opts.fqd.cgid = qipriv.cgr.cgrid;
-+
-+ ret = qman_init_fq(req_fq, fq_sched_flag, &opts);
-+ if (ret) {
-+ dev_err(qidev, "Failed to init session req FQ\n");
-+ goto init_req_fq_fail;
-+ }
-+
-+ dev_dbg(qidev, "Allocated request FQ %u for CPU %u\n", req_fq->fqid,
-+ smp_processor_id());
-+ return req_fq;
-+
-+init_req_fq_fail:
-+ qman_destroy_fq(req_fq, 0);
-+create_req_fq_fail:
-+ kfree(req_fq);
-+ return ERR_PTR(ret);
-+}
-+
-+static int empty_retired_fq(struct device *qidev, struct qman_fq *fq)
-+{
-+ int ret;
-+
-+ ret = qman_volatile_dequeue(fq, QMAN_VOLATILE_FLAG_WAIT_INT |
-+ QMAN_VOLATILE_FLAG_FINISH,
-+ QM_VDQCR_PRECEDENCE_VDQCR |
-+ QM_VDQCR_NUMFRAMES_TILLEMPTY);
-+ if (ret) {
-+ dev_err(qidev, "Volatile dequeue fail for FQ: %u\n", fq->fqid);
-+ return ret;
-+ }
-+
-+ do {
-+ struct qman_portal *p;
-+
-+ p = qman_get_affine_portal(smp_processor_id());
-+ qman_p_poll_dqrr(p, 16);
-+ } while (fq->flags & QMAN_FQ_STATE_NE);
-+
-+ return 0;
-+}
-+
-+static int kill_fq(struct device *qidev, struct qman_fq *fq)
-+{
-+ u32 flags;
-+ int ret;
-+
-+ ret = qman_retire_fq(fq, &flags);
-+ if (ret < 0) {
-+ dev_err(qidev, "qman_retire_fq failed: %d\n", ret);
-+ return ret;
-+ }
-+
-+ if (!ret)
-+ goto empty_fq;
-+
-+ /* Async FQ retirement condition */
-+ if (ret == 1) {
-+ /* Retry till FQ gets in retired state */
-+ do {
-+ msleep(20);
-+ } while (fq->state != qman_fq_state_retired);
-+
-+ WARN_ON(fq->flags & QMAN_FQ_STATE_BLOCKOOS);
-+ WARN_ON(fq->flags & QMAN_FQ_STATE_ORL);
-+ }
-+
-+empty_fq:
-+ if (fq->flags & QMAN_FQ_STATE_NE) {
-+ ret = empty_retired_fq(qidev, fq);
-+ if (ret) {
-+ dev_err(qidev, "empty_retired_fq fail for FQ: %u\n",
-+ fq->fqid);
-+ return ret;
-+ }
-+ }
-+
-+ ret = qman_oos_fq(fq);
-+ if (ret)
-+ dev_err(qidev, "OOS of FQID: %u failed\n", fq->fqid);
-+
-+ qman_destroy_fq(fq, 0);
-+ kfree(fq);
-+
-+ return ret;
-+}
-+
-+static int empty_caam_fq(struct qman_fq *fq)
-+{
-+ int ret;
-+ struct qm_mcr_queryfq_np np;
-+
-+ /* Wait till the older CAAM FQ get empty */
-+ do {
-+ ret = qman_query_fq_np(fq, &np);
-+ if (ret)
-+ return ret;
-+
-+ if (!np.frm_cnt)
-+ break;
-+
-+ msleep(20);
-+ } while (1);
-+
-+ /*
-+ * Give extra time for pending jobs from this FQ in holding tanks
-+ * to get processed
-+ */
-+ msleep(20);
-+ return 0;
-+}
-+
-+int caam_drv_ctx_update(struct caam_drv_ctx *drv_ctx, u32 *sh_desc)
-+{
-+ int ret;
-+ u32 num_words;
-+ struct qman_fq *new_fq, *old_fq;
-+ struct device *qidev = drv_ctx->qidev;
-+
-+ num_words = desc_len(sh_desc);
-+ if (num_words > MAX_SDLEN) {
-+ dev_err(qidev, "Invalid descriptor len: %d words\n", num_words);
-+ return -EINVAL;
-+ }
-+
-+ /* Note down older req FQ */
-+ old_fq = drv_ctx->req_fq;
-+
-+ /* Create a new req FQ in parked state */
-+ new_fq = create_caam_req_fq(drv_ctx->qidev, drv_ctx->rsp_fq,
-+ drv_ctx->context_a, 0);
-+ if (unlikely(IS_ERR_OR_NULL(new_fq))) {
-+ dev_err(qidev, "FQ allocation for shdesc update failed\n");
-+ return PTR_ERR(new_fq);
-+ }
-+
-+ /* Hook up new FQ to context so that new requests keep queuing */
-+ drv_ctx->req_fq = new_fq;
-+
-+ /* Empty and remove the older FQ */
-+ ret = empty_caam_fq(old_fq);
-+ if (ret) {
-+ dev_err(qidev, "Old CAAM FQ empty failed: %d\n", ret);
-+
-+ /* We can revert to older FQ */
-+ drv_ctx->req_fq = old_fq;
-+
-+ if (kill_fq(qidev, new_fq))
-+ dev_warn(qidev, "New CAAM FQ kill failed\n");
-+
-+ return ret;
-+ }
-+
-+ /*
-+ * Re-initialise pre-header. Set RSLS and SDLEN.
-+ * Update the shared descriptor for driver context.
-+ */
-+ drv_ctx->prehdr[0] = cpu_to_caam32((1 << PREHDR_RSLS_SHIFT) |
-+ num_words);
-+ memcpy(drv_ctx->sh_desc, sh_desc, desc_bytes(sh_desc));
-+ dma_sync_single_for_device(qidev, drv_ctx->context_a,
-+ sizeof(drv_ctx->sh_desc) +
-+ sizeof(drv_ctx->prehdr),
-+ DMA_BIDIRECTIONAL);
-+
-+ /* Put the new FQ in scheduled state */
-+ ret = qman_schedule_fq(new_fq);
-+ if (ret) {
-+ dev_err(qidev, "Fail to sched new CAAM FQ, ecode = %d\n", ret);
-+
-+ /*
-+ * We can kill new FQ and revert to old FQ.
-+ * Since the desc is already modified, it is success case
-+ */
-+
-+ drv_ctx->req_fq = old_fq;
-+
-+ if (kill_fq(qidev, new_fq))
-+ dev_warn(qidev, "New CAAM FQ kill failed\n");
-+ } else if (kill_fq(qidev, old_fq)) {
-+ dev_warn(qidev, "Old CAAM FQ kill failed\n");
-+ }
-+
-+ return 0;
-+}
-+EXPORT_SYMBOL(caam_drv_ctx_update);
-+
-+struct caam_drv_ctx *caam_drv_ctx_init(struct device *qidev,
-+ int *cpu,
-+ u32 *sh_desc)
-+{
-+ size_t size;
-+ u32 num_words;
-+ dma_addr_t hwdesc;
-+ struct caam_drv_ctx *drv_ctx;
-+ const cpumask_t *cpus = qman_affine_cpus();
-+
-+ num_words = desc_len(sh_desc);
-+ if (num_words > MAX_SDLEN) {
-+ dev_err(qidev, "Invalid descriptor len: %d words\n",
-+ num_words);
-+ return ERR_PTR(-EINVAL);
-+ }
-+
-+ drv_ctx = kzalloc(sizeof(*drv_ctx), GFP_ATOMIC);
-+ if (!drv_ctx)
-+ return ERR_PTR(-ENOMEM);
-+
-+ /*
-+ * Initialise pre-header - set RSLS and SDLEN - and shared descriptor
-+ * and dma-map them.
-+ */
-+ drv_ctx->prehdr[0] = cpu_to_caam32((1 << PREHDR_RSLS_SHIFT) |
-+ num_words);
-+ memcpy(drv_ctx->sh_desc, sh_desc, desc_bytes(sh_desc));
-+ size = sizeof(drv_ctx->prehdr) + sizeof(drv_ctx->sh_desc);
-+ hwdesc = dma_map_single(qidev, drv_ctx->prehdr, size,
-+ DMA_BIDIRECTIONAL);
-+ if (dma_mapping_error(qidev, hwdesc)) {
-+ dev_err(qidev, "DMA map error for preheader + shdesc\n");
-+ kfree(drv_ctx);
-+ return ERR_PTR(-ENOMEM);
-+ }
-+ drv_ctx->context_a = hwdesc;
-+
-+ /* If given CPU does not own the portal, choose another one that does */
-+ if (!cpumask_test_cpu(*cpu, cpus)) {
-+ int *pcpu = &get_cpu_var(last_cpu);
-+
-+ *pcpu = cpumask_next(*pcpu, cpus);
-+ if (*pcpu >= nr_cpu_ids)
-+ *pcpu = cpumask_first(cpus);
-+ *cpu = *pcpu;
-+
-+ put_cpu_var(last_cpu);
-+ }
-+ drv_ctx->cpu = *cpu;
-+
-+ /* Find response FQ hooked with this CPU */
-+ drv_ctx->rsp_fq = per_cpu(pcpu_qipriv.rsp_fq, drv_ctx->cpu);
-+
-+ /* Attach request FQ */
-+ drv_ctx->req_fq = create_caam_req_fq(qidev, drv_ctx->rsp_fq, hwdesc,
-+ QMAN_INITFQ_FLAG_SCHED);
-+ if (unlikely(IS_ERR_OR_NULL(drv_ctx->req_fq))) {
-+ dev_err(qidev, "create_caam_req_fq failed\n");
-+ dma_unmap_single(qidev, hwdesc, size, DMA_BIDIRECTIONAL);
-+ kfree(drv_ctx);
-+ return ERR_PTR(-ENOMEM);
-+ }
-+
-+ drv_ctx->qidev = qidev;
-+ return drv_ctx;
-+}
-+EXPORT_SYMBOL(caam_drv_ctx_init);
-+
-+void *qi_cache_alloc(gfp_t flags)
-+{
-+ return kmem_cache_alloc(qi_cache, flags);
-+}
-+EXPORT_SYMBOL(qi_cache_alloc);
-+
-+void qi_cache_free(void *obj)
-+{
-+ kmem_cache_free(qi_cache, obj);
-+}
-+EXPORT_SYMBOL(qi_cache_free);
-+
-+static int caam_qi_poll(struct napi_struct *napi, int budget)
-+{
-+ struct caam_napi *np = container_of(napi, struct caam_napi, irqtask);
-+
-+ int cleaned = qman_p_poll_dqrr(np->p, budget);
-+
-+ if (cleaned < budget) {
-+ napi_complete(napi);
-+ qman_p_irqsource_add(np->p, QM_PIRQ_DQRI);
-+ }
-+
-+ return cleaned;
-+}
-+
-+void caam_drv_ctx_rel(struct caam_drv_ctx *drv_ctx)
-+{
-+ if (IS_ERR_OR_NULL(drv_ctx))
-+ return;
-+
-+ /* Remove request FQ */
-+ if (kill_fq(drv_ctx->qidev, drv_ctx->req_fq))
-+ dev_err(drv_ctx->qidev, "Crypto session req FQ kill failed\n");
-+
-+ dma_unmap_single(drv_ctx->qidev, drv_ctx->context_a,
-+ sizeof(drv_ctx->sh_desc) + sizeof(drv_ctx->prehdr),
-+ DMA_BIDIRECTIONAL);
-+ kfree(drv_ctx);
-+}
-+EXPORT_SYMBOL(caam_drv_ctx_rel);
-+
-+int caam_qi_shutdown(struct device *qidev)
-+{
-+ int i, ret;
-+ struct caam_qi_priv *priv = dev_get_drvdata(qidev);
-+ const cpumask_t *cpus = qman_affine_cpus();
-+ struct cpumask old_cpumask = current->cpus_allowed;
-+
-+ for_each_cpu(i, cpus) {
-+ struct napi_struct *irqtask;
-+
-+ irqtask = &per_cpu_ptr(&pcpu_qipriv.caam_napi, i)->irqtask;
-+ napi_disable(irqtask);
-+ netif_napi_del(irqtask);
-+
-+ if (kill_fq(qidev, per_cpu(pcpu_qipriv.rsp_fq, i)))
-+ dev_err(qidev, "Rsp FQ kill failed, cpu: %d\n", i);
-+ }
-+
-+ /*
-+ * QMan driver requires CGRs to be deleted from same CPU from where they
-+ * were instantiated. Hence we get the module removal execute from the
-+ * same CPU from where it was originally inserted.
-+ */
-+ set_cpus_allowed_ptr(current, get_cpu_mask(mod_init_cpu));
-+
-+ ret = qman_delete_cgr(&priv->cgr);
-+ if (ret)
-+ dev_err(qidev, "Deletion of CGR failed: %d\n", ret);
-+ else
-+ qman_release_cgrid(priv->cgr.cgrid);
-+
-+ kmem_cache_destroy(qi_cache);
-+
-+ /* Now that we're done with the CGRs, restore the cpus allowed mask */
-+ set_cpus_allowed_ptr(current, &old_cpumask);
-+
-+ platform_device_unregister(priv->qi_pdev);
-+ return ret;
-+}
-+
-+static void cgr_cb(struct qman_portal *qm, struct qman_cgr *cgr, int congested)
-+{
-+ caam_congested = congested;
-+
-+ if (congested) {
-+#ifdef CONFIG_DEBUG_FS
-+ times_congested++;
-+#endif
-+ pr_debug_ratelimited("CAAM entered congestion\n");
-+
-+ } else {
-+ pr_debug_ratelimited("CAAM exited congestion\n");
-+ }
-+}
-+
-+static int caam_qi_napi_schedule(struct qman_portal *p, struct caam_napi *np)
-+{
-+ /*
-+ * In case of threaded ISR, for RT kernels in_irq() does not return
-+ * appropriate value, so use in_serving_softirq to distinguish between
-+ * softirq and irq contexts.
-+ */
-+ if (unlikely(in_irq() || !in_serving_softirq())) {
-+ /* Disable QMan IRQ source and invoke NAPI */
-+ qman_p_irqsource_remove(p, QM_PIRQ_DQRI);
-+ np->p = p;
-+ napi_schedule(&np->irqtask);
-+ return 1;
-+ }
-+ return 0;
-+}
-+
-+static enum qman_cb_dqrr_result caam_rsp_fq_dqrr_cb(struct qman_portal *p,
-+ struct qman_fq *rsp_fq,
-+ const struct qm_dqrr_entry *dqrr)
-+{
-+ struct caam_napi *caam_napi = raw_cpu_ptr(&pcpu_qipriv.caam_napi);
-+ struct caam_drv_req *drv_req;
-+ const struct qm_fd *fd;
-+ struct device *qidev = &(raw_cpu_ptr(&pcpu_qipriv)->net_dev.dev);
-+
-+ if (caam_qi_napi_schedule(p, caam_napi))
-+ return qman_cb_dqrr_stop;
-+
-+ fd = &dqrr->fd;
-+ if (unlikely(fd->status)) {
-+ u32 ssrc = fd->status & JRSTA_SSRC_MASK;
-+ u8 err_id = fd->status & JRSTA_CCBERR_ERRID_MASK;
-+
-+ if (ssrc != JRSTA_SSRC_CCB_ERROR ||
-+ err_id != JRSTA_CCBERR_ERRID_ICVCHK)
-+ dev_err(qidev, "Error: %#x in CAAM response FD\n",
-+ fd->status);
-+ }
-+
-+ if (unlikely(fd->format != fd->format)) {
-+ dev_err(qidev, "Non-compound FD from CAAM\n");
-+ return qman_cb_dqrr_consume;
-+ }
-+
-+ drv_req = (struct caam_drv_req *)phys_to_virt(fd->addr);
-+ if (unlikely(!drv_req)) {
-+ dev_err(qidev,
-+ "Can't find original request for caam response\n");
-+ return qman_cb_dqrr_consume;
-+ }
-+
-+ dma_unmap_single(drv_req->drv_ctx->qidev, qm_fd_addr(fd),
-+ sizeof(drv_req->fd_sgt), DMA_BIDIRECTIONAL);
-+
-+ drv_req->cbk(drv_req, fd->status);
-+ return qman_cb_dqrr_consume;
-+}
-+
-+static int alloc_rsp_fq_cpu(struct device *qidev, unsigned int cpu)
-+{
-+ struct qm_mcc_initfq opts;
-+ struct qman_fq *fq;
-+ int ret;
-+
-+ fq = kzalloc(sizeof(*fq), GFP_KERNEL | GFP_DMA);
-+ if (!fq)
-+ return -ENOMEM;
-+
-+ fq->cb.dqrr = caam_rsp_fq_dqrr_cb;
-+
-+ ret = qman_create_fq(0, QMAN_FQ_FLAG_NO_ENQUEUE |
-+ QMAN_FQ_FLAG_DYNAMIC_FQID, fq);
-+ if (ret) {
-+ dev_err(qidev, "Rsp FQ create failed\n");
-+ kfree(fq);
-+ return -ENODEV;
-+ }
-+
-+ opts.we_mask = QM_INITFQ_WE_FQCTRL | QM_INITFQ_WE_DESTWQ |
-+ QM_INITFQ_WE_CONTEXTB | QM_INITFQ_WE_CONTEXTA |
-+ QM_INITFQ_WE_CGID;
-+ opts.fqd.fq_ctrl = QM_FQCTRL_CTXASTASHING | QM_FQCTRL_CPCSTASH |
-+ QM_FQCTRL_CGE;
-+ opts.fqd.dest.channel = qman_affine_channel(cpu);
-+ opts.fqd.dest.wq = 3;
-+ opts.fqd.cgid = qipriv.cgr.cgrid;
-+ opts.fqd.context_a.stashing.exclusive = QM_STASHING_EXCL_CTX |
-+ QM_STASHING_EXCL_DATA;
-+ opts.fqd.context_a.stashing.data_cl = 1;
-+ opts.fqd.context_a.stashing.context_cl = 1;
-+
-+ ret = qman_init_fq(fq, QMAN_INITFQ_FLAG_SCHED, &opts);
-+ if (ret) {
-+ dev_err(qidev, "Rsp FQ init failed\n");
-+ kfree(fq);
-+ return -ENODEV;
-+ }
-+
-+ per_cpu(pcpu_qipriv.rsp_fq, cpu) = fq;
-+
-+ dev_dbg(qidev, "Allocated response FQ %u for CPU %u", fq->fqid, cpu);
-+ return 0;
-+}
-+
-+static int init_cgr(struct device *qidev)
-+{
-+ int ret;
-+ struct qm_mcc_initcgr opts;
-+ const u64 cpus = *(u64 *)qman_affine_cpus();
-+ const int num_cpus = hweight64(cpus);
-+ const u64 val = num_cpus * MAX_RSP_FQ_BACKLOG_PER_CPU;
-+
-+ ret = qman_alloc_cgrid(&qipriv.cgr.cgrid);
-+ if (ret) {
-+ dev_err(qidev, "CGR alloc failed for rsp FQs: %d\n", ret);
-+ return ret;
-+ }
-+
-+ qipriv.cgr.cb = cgr_cb;
-+ memset(&opts, 0, sizeof(opts));
-+ opts.we_mask = QM_CGR_WE_CSCN_EN | QM_CGR_WE_CS_THRES | QM_CGR_WE_MODE;
-+ opts.cgr.cscn_en = QM_CGR_EN;
-+ opts.cgr.mode = QMAN_CGR_MODE_FRAME;
-+ qm_cgr_cs_thres_set64(&opts.cgr.cs_thres, val, 1);
-+
-+ ret = qman_create_cgr(&qipriv.cgr, QMAN_CGR_FLAG_USE_INIT, &opts);
-+ if (ret) {
-+ dev_err(qidev, "Error %d creating CAAM CGRID: %u\n", ret,
-+ qipriv.cgr.cgrid);
-+ return ret;
-+ }
-+
-+ dev_dbg(qidev, "Congestion threshold set to %llu\n", val);
-+ return 0;
-+}
-+
-+static int alloc_rsp_fqs(struct device *qidev)
-+{
-+ int ret, i;
-+ const cpumask_t *cpus = qman_affine_cpus();
-+
-+ /*Now create response FQs*/
-+ for_each_cpu(i, cpus) {
-+ ret = alloc_rsp_fq_cpu(qidev, i);
-+ if (ret) {
-+ dev_err(qidev, "CAAM rsp FQ alloc failed, cpu: %u", i);
-+ return ret;
-+ }
-+ }
-+
-+ return 0;
-+}
-+
-+static void free_rsp_fqs(void)
-+{
-+ int i;
-+ const cpumask_t *cpus = qman_affine_cpus();
-+
-+ for_each_cpu(i, cpus)
-+ kfree(per_cpu(pcpu_qipriv.rsp_fq, i));
-+}
-+
-+int caam_qi_init(struct platform_device *caam_pdev)
-+{
-+ int err, i;
-+ struct platform_device *qi_pdev;
-+ struct device *ctrldev = &caam_pdev->dev, *qidev;
-+ struct caam_drv_private *ctrlpriv;
-+ const cpumask_t *cpus = qman_affine_cpus();
-+ struct cpumask old_cpumask = current->cpus_allowed;
-+ static struct platform_device_info qi_pdev_info = {
-+ .name = "caam_qi",
-+ .id = PLATFORM_DEVID_NONE
-+ };
-+
-+ /*
-+ * QMAN requires CGRs to be removed from same CPU+portal from where it
-+ * was originally allocated. Hence we need to note down the
-+ * initialisation CPU and use the same CPU for module exit.
-+ * We select the first CPU to from the list of portal owning CPUs.
-+ * Then we pin module init to this CPU.
-+ */
-+ mod_init_cpu = cpumask_first(cpus);
-+ set_cpus_allowed_ptr(current, get_cpu_mask(mod_init_cpu));
-+
-+ qi_pdev_info.parent = ctrldev;
-+ qi_pdev_info.dma_mask = dma_get_mask(ctrldev);
-+ qi_pdev = platform_device_register_full(&qi_pdev_info);
-+ if (IS_ERR(qi_pdev))
-+ return PTR_ERR(qi_pdev);
-+ arch_setup_dma_ops(&qi_pdev->dev, 0, 0, NULL, true);
-+
-+ ctrlpriv = dev_get_drvdata(ctrldev);
-+ qidev = &qi_pdev->dev;
-+
-+ qipriv.qi_pdev = qi_pdev;
-+ dev_set_drvdata(qidev, &qipriv);
-+
-+ /* Initialize the congestion detection */
-+ err = init_cgr(qidev);
-+ if (err) {
-+ dev_err(qidev, "CGR initialization failed: %d\n", err);
-+ platform_device_unregister(qi_pdev);
-+ return err;
-+ }
-+
-+ /* Initialise response FQs */
-+ err = alloc_rsp_fqs(qidev);
-+ if (err) {
-+ dev_err(qidev, "Can't allocate CAAM response FQs: %d\n", err);
-+ free_rsp_fqs();
-+ platform_device_unregister(qi_pdev);
-+ return err;
-+ }
-+
-+ /*
-+ * Enable the NAPI contexts on each of the core which has an affine
-+ * portal.
-+ */
-+ for_each_cpu(i, cpus) {
-+ struct caam_qi_pcpu_priv *priv = per_cpu_ptr(&pcpu_qipriv, i);
-+ struct caam_napi *caam_napi = &priv->caam_napi;
-+ struct napi_struct *irqtask = &caam_napi->irqtask;
-+ struct net_device *net_dev = &priv->net_dev;
-+
-+ net_dev->dev = *qidev;
-+ INIT_LIST_HEAD(&net_dev->napi_list);
-+
-+ netif_napi_add(net_dev, irqtask, caam_qi_poll,
-+ CAAM_NAPI_WEIGHT);
-+
-+ napi_enable(irqtask);
-+ }
-+
-+ /* Hook up QI device to parent controlling caam device */
-+ ctrlpriv->qidev = qidev;
-+
-+ qi_cache = kmem_cache_create("caamqicache", CAAM_QI_MEMCACHE_SIZE, 0,
-+ SLAB_CACHE_DMA, NULL);
-+ if (!qi_cache) {
-+ dev_err(qidev, "Can't allocate CAAM cache\n");
-+ free_rsp_fqs();
-+ platform_device_unregister(qi_pdev);
-+ return -ENOMEM;
-+ }
-+
-+ /* Done with the CGRs; restore the cpus allowed mask */
-+ set_cpus_allowed_ptr(current, &old_cpumask);
-+#ifdef CONFIG_DEBUG_FS
-+ debugfs_create_file("qi_congested", 0444, ctrlpriv->ctl,
-+ &times_congested, &caam_fops_u64_ro);
-+#endif
-+ dev_info(qidev, "Linux CAAM Queue I/F driver initialised\n");
-+ return 0;
-+}
---- /dev/null
-+++ b/drivers/crypto/caam/qi.h
-@@ -0,0 +1,204 @@
-+/*
-+ * Public definitions for the CAAM/QI (Queue Interface) backend.
-+ *
-+ * Copyright 2013-2016 Freescale Semiconductor, Inc.
-+ * Copyright 2016-2017 NXP
-+ */
-+
-+#ifndef __QI_H__
-+#define __QI_H__
-+
-+#include <linux/fsl_qman.h>
-+#include "compat.h"
-+#include "desc.h"
-+#include "desc_constr.h"
-+
-+/*
-+ * CAAM hardware constructs a job descriptor which points to a shared descriptor
-+ * (as pointed by context_a of to-CAAM FQ).
-+ * When the job descriptor is executed by DECO, the whole job descriptor
-+ * together with shared descriptor gets loaded in DECO buffer, which is
-+ * 64 words (each 32-bit) long.
-+ *
-+ * The job descriptor constructed by CAAM hardware has the following layout:
-+ *
-+ * HEADER (1 word)
-+ * Shdesc ptr (1 or 2 words)
-+ * SEQ_OUT_PTR (1 word)
-+ * Out ptr (1 or 2 words)
-+ * Out length (1 word)
-+ * SEQ_IN_PTR (1 word)
-+ * In ptr (1 or 2 words)
-+ * In length (1 word)
-+ *
-+ * The shdesc ptr is used to fetch shared descriptor contents into DECO buffer.
-+ *
-+ * Apart from shdesc contents, the total number of words that get loaded in DECO
-+ * buffer are '8' or '11'. The remaining words in DECO buffer can be used for
-+ * storing shared descriptor.
-+ */
-+#define MAX_SDLEN ((CAAM_DESC_BYTES_MAX - DESC_JOB_IO_LEN) / CAAM_CMD_SZ)
-+
-+/* Length of a single buffer in the QI driver memory cache */
-+#define CAAM_QI_MEMCACHE_SIZE 768
-+
-+extern bool caam_congested __read_mostly;
-+
-+/*
-+ * This is the request structure the driver application should fill while
-+ * submitting a job to driver.
-+ */
-+struct caam_drv_req;
-+
-+/*
-+ * caam_qi_cbk - application's callback function invoked by the driver when the
-+ * request has been successfully processed.
-+ * @drv_req: original request that was submitted
-+ * @status: completion status of request (0 - success, non-zero - error code)
-+ */
-+typedef void (*caam_qi_cbk)(struct caam_drv_req *drv_req, u32 status);
-+
-+enum optype {
-+ ENCRYPT,
-+ DECRYPT,
-+ GIVENCRYPT,
-+ NUM_OP
-+};
-+
-+/**
-+ * caam_drv_ctx - CAAM/QI backend driver context
-+ *
-+ * The jobs are processed by the driver against a driver context.
-+ * With every cryptographic context, a driver context is attached.
-+ * The driver context contains data for private use by driver.
-+ * For the applications, this is an opaque structure.
-+ *
-+ * @prehdr: preheader placed before shrd desc
-+ * @sh_desc: shared descriptor
-+ * @context_a: shared descriptor dma address
-+ * @req_fq: to-CAAM request frame queue
-+ * @rsp_fq: from-CAAM response frame queue
-+ * @cpu: cpu on which to receive CAAM response
-+ * @op_type: operation type
-+ * @qidev: device pointer for CAAM/QI backend
-+ */
-+struct caam_drv_ctx {
-+ u32 prehdr[2];
-+ u32 sh_desc[MAX_SDLEN];
-+ dma_addr_t context_a;
-+ struct qman_fq *req_fq;
-+ struct qman_fq *rsp_fq;
-+ int cpu;
-+ enum optype op_type;
-+ struct device *qidev;
-+} ____cacheline_aligned;
-+
-+/**
-+ * caam_drv_req - The request structure the driver application should fill while
-+ * submitting a job to driver.
-+ * @fd_sgt: QMan S/G pointing to output (fd_sgt[0]) and input (fd_sgt[1])
-+ * buffers.
-+ * @cbk: callback function to invoke when job is completed
-+ * @app_ctx: arbitrary context attached with request by the application
-+ *
-+ * The fields mentioned below should not be used by application.
-+ * These are for private use by driver.
-+ *
-+ * @hdr__: linked list header to maintain list of outstanding requests to CAAM
-+ * @hwaddr: DMA address for the S/G table.
-+ */
-+struct caam_drv_req {
-+ struct qm_sg_entry fd_sgt[2];
-+ struct caam_drv_ctx *drv_ctx;
-+ caam_qi_cbk cbk;
-+ void *app_ctx;
-+} ____cacheline_aligned;
-+
-+/**
-+ * caam_drv_ctx_init - Initialise a CAAM/QI driver context
-+ *
-+ * A CAAM/QI driver context must be attached with each cryptographic context.
-+ * This function allocates memory for CAAM/QI context and returns a handle to
-+ * the application. This handle must be submitted along with each enqueue
-+ * request to the driver by the application.
-+ *
-+ * @cpu: CPU where the application prefers to the driver to receive CAAM
-+ * responses. The request completion callback would be issued from this
-+ * CPU.
-+ * @sh_desc: shared descriptor pointer to be attached with CAAM/QI driver
-+ * context.
-+ *
-+ * Returns a driver context on success or negative error code on failure.
-+ */
-+struct caam_drv_ctx *caam_drv_ctx_init(struct device *qidev, int *cpu,
-+ u32 *sh_desc);
-+
-+/**
-+ * caam_qi_enqueue - Submit a request to QI backend driver.
-+ *
-+ * The request structure must be properly filled as described above.
-+ *
-+ * @qidev: device pointer for QI backend
-+ * @req: CAAM QI request structure
-+ *
-+ * Returns 0 on success or negative error code on failure.
-+ */
-+int caam_qi_enqueue(struct device *qidev, struct caam_drv_req *req);
-+
-+/**
-+ * caam_drv_ctx_busy - Check if there are too many jobs pending with CAAM
-+ * or too many CAAM responses are pending to be processed.
-+ * @drv_ctx: driver context for which job is to be submitted
-+ *
-+ * Returns caam congestion status 'true/false'
-+ */
-+bool caam_drv_ctx_busy(struct caam_drv_ctx *drv_ctx);
-+
-+/**
-+ * caam_drv_ctx_update - Update QI driver context
-+ *
-+ * Invoked when shared descriptor is required to be change in driver context.
-+ *
-+ * @drv_ctx: driver context to be updated
-+ * @sh_desc: new shared descriptor pointer to be updated in QI driver context
-+ *
-+ * Returns 0 on success or negative error code on failure.
-+ */
-+int caam_drv_ctx_update(struct caam_drv_ctx *drv_ctx, u32 *sh_desc);
-+
-+/**
-+ * caam_drv_ctx_rel - Release a QI driver context
-+ * @drv_ctx: context to be released
-+ */
-+void caam_drv_ctx_rel(struct caam_drv_ctx *drv_ctx);
-+
-+int caam_qi_init(struct platform_device *pdev);
-+int caam_qi_shutdown(struct device *dev);
-+
-+/**
-+ * qi_cache_alloc - Allocate buffers from CAAM-QI cache
-+ *
-+ * Invoked when a user of the CAAM-QI (i.e. caamalg-qi) needs data which has
-+ * to be allocated on the hotpath. Instead of using malloc, one can use the
-+ * services of the CAAM QI memory cache (backed by kmem_cache). The buffers
-+ * will have a size of 256B, which is sufficient for hosting 16 SG entries.
-+ *
-+ * @flags: flags that would be used for the equivalent malloc(..) call
-+ *
-+ * Returns a pointer to a retrieved buffer on success or NULL on failure.
-+ */
-+void *qi_cache_alloc(gfp_t flags);
-+
-+/**
-+ * qi_cache_free - Frees buffers allocated from CAAM-QI cache
-+ *
-+ * Invoked when a user of the CAAM-QI (i.e. caamalg-qi) no longer needs
-+ * the buffer previously allocated by a qi_cache_alloc call.
-+ * No checking is being done, the call is a passthrough call to
-+ * kmem_cache_free(...)
-+ *
-+ * @obj: object previously allocated using qi_cache_alloc()
-+ */
-+void qi_cache_free(void *obj);
-+
-+#endif /* __QI_H__ */
---- a/drivers/crypto/caam/regs.h
-+++ b/drivers/crypto/caam/regs.h
-@@ -2,6 +2,7 @@
- * CAAM hardware register-level view
- *
- * Copyright 2008-2011 Freescale Semiconductor, Inc.
-+ * Copyright 2017 NXP
- */
-
- #ifndef REGS_H
-@@ -67,6 +68,7 @@
- */
-
- extern bool caam_little_end;
-+extern bool caam_imx;
-
- #define caam_to_cpu(len) \
- static inline u##len caam##len ## _to_cpu(u##len val) \
-@@ -154,13 +156,10 @@ static inline u64 rd_reg64(void __iomem
- #else /* CONFIG_64BIT */
- static inline void wr_reg64(void __iomem *reg, u64 data)
- {
--#ifndef CONFIG_CRYPTO_DEV_FSL_CAAM_IMX
-- if (caam_little_end) {
-+ if (!caam_imx && caam_little_end) {
- wr_reg32((u32 __iomem *)(reg) + 1, data >> 32);
- wr_reg32((u32 __iomem *)(reg), data);
-- } else
--#endif
-- {
-+ } else {
- wr_reg32((u32 __iomem *)(reg), data >> 32);
- wr_reg32((u32 __iomem *)(reg) + 1, data);
- }
-@@ -168,41 +167,40 @@ static inline void wr_reg64(void __iomem
-
- static inline u64 rd_reg64(void __iomem *reg)
- {
--#ifndef CONFIG_CRYPTO_DEV_FSL_CAAM_IMX
-- if (caam_little_end)
-+ if (!caam_imx && caam_little_end)
- return ((u64)rd_reg32((u32 __iomem *)(reg) + 1) << 32 |
- (u64)rd_reg32((u32 __iomem *)(reg)));
-- else
--#endif
-- return ((u64)rd_reg32((u32 __iomem *)(reg)) << 32 |
-- (u64)rd_reg32((u32 __iomem *)(reg) + 1));
-+
-+ return ((u64)rd_reg32((u32 __iomem *)(reg)) << 32 |
-+ (u64)rd_reg32((u32 __iomem *)(reg) + 1));
- }
- #endif /* CONFIG_64BIT */
-
-+static inline u64 cpu_to_caam_dma64(dma_addr_t value)
-+{
-+ if (caam_imx)
-+ return (((u64)cpu_to_caam32(lower_32_bits(value)) << 32) |
-+ (u64)cpu_to_caam32(upper_32_bits(value)));
-+
-+ return cpu_to_caam64(value);
-+}
-+
-+static inline u64 caam_dma64_to_cpu(u64 value)
-+{
-+ if (caam_imx)
-+ return (((u64)caam32_to_cpu(lower_32_bits(value)) << 32) |
-+ (u64)caam32_to_cpu(upper_32_bits(value)));
-+
-+ return caam64_to_cpu(value);
-+}
-+
- #ifdef CONFIG_ARCH_DMA_ADDR_T_64BIT
--#ifdef CONFIG_SOC_IMX7D
--#define cpu_to_caam_dma(value) \
-- (((u64)cpu_to_caam32(lower_32_bits(value)) << 32) | \
-- (u64)cpu_to_caam32(upper_32_bits(value)))
--#define caam_dma_to_cpu(value) \
-- (((u64)caam32_to_cpu(lower_32_bits(value)) << 32) | \
-- (u64)caam32_to_cpu(upper_32_bits(value)))
--#else
--#define cpu_to_caam_dma(value) cpu_to_caam64(value)
--#define caam_dma_to_cpu(value) caam64_to_cpu(value)
--#endif /* CONFIG_SOC_IMX7D */
-+#define cpu_to_caam_dma(value) cpu_to_caam_dma64(value)
-+#define caam_dma_to_cpu(value) caam_dma64_to_cpu(value)
- #else
- #define cpu_to_caam_dma(value) cpu_to_caam32(value)
- #define caam_dma_to_cpu(value) caam32_to_cpu(value)
--#endif /* CONFIG_ARCH_DMA_ADDR_T_64BIT */
--
--#ifdef CONFIG_CRYPTO_DEV_FSL_CAAM_IMX
--#define cpu_to_caam_dma64(value) \
-- (((u64)cpu_to_caam32(lower_32_bits(value)) << 32) | \
-- (u64)cpu_to_caam32(upper_32_bits(value)))
--#else
--#define cpu_to_caam_dma64(value) cpu_to_caam64(value)
--#endif
-+#endif /* CONFIG_ARCH_DMA_ADDR_T_64BIT */
-
- /*
- * jr_outentry
-@@ -293,6 +291,7 @@ struct caam_perfmon {
- u32 cha_rev_ls; /* CRNR - CHA Rev No. Least significant half*/
- #define CTPR_MS_QI_SHIFT 25
- #define CTPR_MS_QI_MASK (0x1ull << CTPR_MS_QI_SHIFT)
-+#define CTPR_MS_DPAA2 BIT(13)
- #define CTPR_MS_VIRT_EN_INCL 0x00000001
- #define CTPR_MS_VIRT_EN_POR 0x00000002
- #define CTPR_MS_PG_SZ_MASK 0x10
-@@ -628,6 +627,8 @@ struct caam_job_ring {
- #define JRSTA_DECOERR_INVSIGN 0x86
- #define JRSTA_DECOERR_DSASIGN 0x87
-
-+#define JRSTA_QIERR_ERROR_MASK 0x00ff
-+
- #define JRSTA_CCBERR_JUMP 0x08000000
- #define JRSTA_CCBERR_INDEX_MASK 0xff00
- #define JRSTA_CCBERR_INDEX_SHIFT 8
---- /dev/null
-+++ b/drivers/crypto/caam/sg_sw_qm.h
-@@ -0,0 +1,126 @@
-+/*
-+ * Copyright 2013-2016 Freescale Semiconductor, Inc.
-+ * Copyright 2016-2017 NXP
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of Freescale Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#ifndef __SG_SW_QM_H
-+#define __SG_SW_QM_H
-+
-+#include <linux/fsl_qman.h>
-+#include "regs.h"
-+
-+static inline void cpu_to_hw_sg(struct qm_sg_entry *qm_sg_ptr)
-+{
-+ dma_addr_t addr = qm_sg_ptr->opaque;
-+
-+ qm_sg_ptr->opaque = cpu_to_caam64(addr);
-+ qm_sg_ptr->sgt_efl = cpu_to_caam32(qm_sg_ptr->sgt_efl);
-+}
-+
-+static inline void __dma_to_qm_sg(struct qm_sg_entry *qm_sg_ptr, dma_addr_t dma,
-+ u32 len, u16 offset)
-+{
-+ qm_sg_ptr->addr = dma;
-+ qm_sg_ptr->length = len;
-+ qm_sg_ptr->__reserved2 = 0;
-+ qm_sg_ptr->bpid = 0;
-+ qm_sg_ptr->__reserved3 = 0;
-+ qm_sg_ptr->offset = offset & QM_SG_OFFSET_MASK;
-+
-+ cpu_to_hw_sg(qm_sg_ptr);
-+}
-+
-+static inline void dma_to_qm_sg_one(struct qm_sg_entry *qm_sg_ptr,
-+ dma_addr_t dma, u32 len, u16 offset)
-+{
-+ qm_sg_ptr->extension = 0;
-+ qm_sg_ptr->final = 0;
-+ __dma_to_qm_sg(qm_sg_ptr, dma, len, offset);
-+}
-+
-+static inline void dma_to_qm_sg_one_last(struct qm_sg_entry *qm_sg_ptr,
-+ dma_addr_t dma, u32 len, u16 offset)
-+{
-+ qm_sg_ptr->extension = 0;
-+ qm_sg_ptr->final = 1;
-+ __dma_to_qm_sg(qm_sg_ptr, dma, len, offset);
-+}
-+
-+static inline void dma_to_qm_sg_one_ext(struct qm_sg_entry *qm_sg_ptr,
-+ dma_addr_t dma, u32 len, u16 offset)
-+{
-+ qm_sg_ptr->extension = 1;
-+ qm_sg_ptr->final = 0;
-+ __dma_to_qm_sg(qm_sg_ptr, dma, len, offset);
-+}
-+
-+static inline void dma_to_qm_sg_one_last_ext(struct qm_sg_entry *qm_sg_ptr,
-+ dma_addr_t dma, u32 len,
-+ u16 offset)
-+{
-+ qm_sg_ptr->extension = 1;
-+ qm_sg_ptr->final = 1;
-+ __dma_to_qm_sg(qm_sg_ptr, dma, len, offset);
-+}
-+
-+/*
-+ * convert scatterlist to h/w link table format
-+ * but does not have final bit; instead, returns last entry
-+ */
-+static inline struct qm_sg_entry *
-+sg_to_qm_sg(struct scatterlist *sg, int sg_count,
-+ struct qm_sg_entry *qm_sg_ptr, u16 offset)
-+{
-+ while (sg_count && sg) {
-+ dma_to_qm_sg_one(qm_sg_ptr, sg_dma_address(sg),
-+ sg_dma_len(sg), offset);
-+ qm_sg_ptr++;
-+ sg = sg_next(sg);
-+ sg_count--;
-+ }
-+ return qm_sg_ptr - 1;
-+}
-+
-+/*
-+ * convert scatterlist to h/w link table format
-+ * scatterlist must have been previously dma mapped
-+ */
-+static inline void sg_to_qm_sg_last(struct scatterlist *sg, int sg_count,
-+ struct qm_sg_entry *qm_sg_ptr, u16 offset)
-+{
-+ qm_sg_ptr = sg_to_qm_sg(sg, sg_count, qm_sg_ptr, offset);
-+
-+ qm_sg_ptr->sgt_efl = caam32_to_cpu(qm_sg_ptr->sgt_efl);
-+ qm_sg_ptr->final = 1;
-+ qm_sg_ptr->sgt_efl = cpu_to_caam32(qm_sg_ptr->sgt_efl);
-+}
-+
-+#endif /* __SG_SW_QM_H */
---- /dev/null
-+++ b/drivers/crypto/caam/sg_sw_qm2.h
-@@ -0,0 +1,81 @@
-+/*
-+ * Copyright 2015-2016 Freescale Semiconductor, Inc.
-+ * Copyright 2017 NXP
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the names of the above-listed copyright holders nor the
-+ * names of any contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
-+ * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
-+ * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
-+ * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE
-+ * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
-+ * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
-+ * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
-+ * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
-+ * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
-+ * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
-+ * POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#ifndef _SG_SW_QM2_H_
-+#define _SG_SW_QM2_H_
-+
-+#include "../../../drivers/staging/fsl-mc/include/dpaa2-fd.h"
-+
-+static inline void dma_to_qm_sg_one(struct dpaa2_sg_entry *qm_sg_ptr,
-+ dma_addr_t dma, u32 len, u16 offset)
-+{
-+ dpaa2_sg_set_addr(qm_sg_ptr, dma);
-+ dpaa2_sg_set_format(qm_sg_ptr, dpaa2_sg_single);
-+ dpaa2_sg_set_final(qm_sg_ptr, false);
-+ dpaa2_sg_set_len(qm_sg_ptr, len);
-+ dpaa2_sg_set_bpid(qm_sg_ptr, 0);
-+ dpaa2_sg_set_offset(qm_sg_ptr, offset);
-+}
-+
-+/*
-+ * convert scatterlist to h/w link table format
-+ * but does not have final bit; instead, returns last entry
-+ */
-+static inline struct dpaa2_sg_entry *
-+sg_to_qm_sg(struct scatterlist *sg, int sg_count,
-+ struct dpaa2_sg_entry *qm_sg_ptr, u16 offset)
-+{
-+ while (sg_count && sg) {
-+ dma_to_qm_sg_one(qm_sg_ptr, sg_dma_address(sg),
-+ sg_dma_len(sg), offset);
-+ qm_sg_ptr++;
-+ sg = sg_next(sg);
-+ sg_count--;
-+ }
-+ return qm_sg_ptr - 1;
-+}
-+
-+/*
-+ * convert scatterlist to h/w link table format
-+ * scatterlist must have been previously dma mapped
-+ */
-+static inline void sg_to_qm_sg_last(struct scatterlist *sg, int sg_count,
-+ struct dpaa2_sg_entry *qm_sg_ptr,
-+ u16 offset)
-+{
-+ qm_sg_ptr = sg_to_qm_sg(sg, sg_count, qm_sg_ptr, offset);
-+ dpaa2_sg_set_final(qm_sg_ptr, true);
-+}
-+
-+#endif /* _SG_SW_QM2_H_ */
---- a/drivers/crypto/caam/sg_sw_sec4.h
-+++ b/drivers/crypto/caam/sg_sw_sec4.h
-@@ -5,9 +5,19 @@
- *
- */
-
-+#ifndef _SG_SW_SEC4_H_
-+#define _SG_SW_SEC4_H_
-+
-+#include "ctrl.h"
- #include "regs.h"
-+#include "sg_sw_qm2.h"
-+#include "../../../drivers/staging/fsl-mc/include/dpaa2-fd.h"
-
--struct sec4_sg_entry;
-+struct sec4_sg_entry {
-+ u64 ptr;
-+ u32 len;
-+ u32 bpid_offset;
-+};
-
- /*
- * convert single dma address to h/w link table format
-@@ -15,9 +25,15 @@ struct sec4_sg_entry;
- static inline void dma_to_sec4_sg_one(struct sec4_sg_entry *sec4_sg_ptr,
- dma_addr_t dma, u32 len, u16 offset)
- {
-- sec4_sg_ptr->ptr = cpu_to_caam_dma64(dma);
-- sec4_sg_ptr->len = cpu_to_caam32(len);
-- sec4_sg_ptr->bpid_offset = cpu_to_caam32(offset & SEC4_SG_OFFSET_MASK);
-+ if (caam_dpaa2) {
-+ dma_to_qm_sg_one((struct dpaa2_sg_entry *)sec4_sg_ptr, dma, len,
-+ offset);
-+ } else {
-+ sec4_sg_ptr->ptr = cpu_to_caam_dma64(dma);
-+ sec4_sg_ptr->len = cpu_to_caam32(len);
-+ sec4_sg_ptr->bpid_offset = cpu_to_caam32(offset &
-+ SEC4_SG_OFFSET_MASK);
-+ }
- #ifdef DEBUG
- print_hex_dump(KERN_ERR, "sec4_sg_ptr@: ",
- DUMP_PREFIX_ADDRESS, 16, 4, sec4_sg_ptr,
-@@ -43,6 +59,14 @@ sg_to_sec4_sg(struct scatterlist *sg, in
- return sec4_sg_ptr - 1;
- }
-
-+static inline void sg_to_sec4_set_last(struct sec4_sg_entry *sec4_sg_ptr)
-+{
-+ if (caam_dpaa2)
-+ dpaa2_sg_set_final((struct dpaa2_sg_entry *)sec4_sg_ptr, true);
-+ else
-+ sec4_sg_ptr->len |= cpu_to_caam32(SEC4_SG_LEN_FIN);
-+}
-+
- /*
- * convert scatterlist to h/w link table format
- * scatterlist must have been previously dma mapped
-@@ -52,31 +76,7 @@ static inline void sg_to_sec4_sg_last(st
- u16 offset)
- {
- sec4_sg_ptr = sg_to_sec4_sg(sg, sg_count, sec4_sg_ptr, offset);
-- sec4_sg_ptr->len |= cpu_to_caam32(SEC4_SG_LEN_FIN);
--}
--
--static inline struct sec4_sg_entry *sg_to_sec4_sg_len(
-- struct scatterlist *sg, unsigned int total,
-- struct sec4_sg_entry *sec4_sg_ptr)
--{
-- do {
-- unsigned int len = min(sg_dma_len(sg), total);
--
-- dma_to_sec4_sg_one(sec4_sg_ptr, sg_dma_address(sg), len, 0);
-- sec4_sg_ptr++;
-- sg = sg_next(sg);
-- total -= len;
-- } while (total);
-- return sec4_sg_ptr - 1;
-+ sg_to_sec4_set_last(sec4_sg_ptr);
- }
-
--/* derive number of elements in scatterlist, but return 0 for 1 */
--static inline int sg_count(struct scatterlist *sg_list, int nbytes)
--{
-- int sg_nents = sg_nents_for_len(sg_list, nbytes);
--
-- if (likely(sg_nents == 1))
-- return 0;
--
-- return sg_nents;
--}
-+#endif /* _SG_SW_SEC4_H_ */
---- a/drivers/crypto/talitos.c
-+++ b/drivers/crypto/talitos.c
-@@ -1241,6 +1241,14 @@ static int ipsec_esp(struct talitos_edes
- ret = talitos_sg_map_ext(dev, areq->src, cryptlen, edesc, &desc->ptr[4],
- sg_count, areq->assoclen, tbl_off, elen);
-
-+ /*
-+ * In case of SEC 2.x+, cipher in len must include only the ciphertext,
-+ * while extent is used for ICV len.
-+ */
-+ if ((edesc->desc.hdr & DESC_HDR_TYPE_IPSEC_ESP) &&
-+ (desc->hdr & DESC_HDR_MODE1_MDEU_CICV))
-+ desc->ptr[4].len = cpu_to_be16(cryptlen);
-+
- if (ret > 1) {
- tbl_off += ret;
- sync_needed = true;
---- a/drivers/net/wireless/rsi/rsi_91x_usb.c
-+++ b/drivers/net/wireless/rsi/rsi_91x_usb.c
-@@ -516,7 +516,7 @@ err:
-
- /**
- * rsi_disconnect() - This function performs the reverse of the probe function,
-- * it deintialize the driver structure.
-+ * it deinitialize the driver structure.
- * @pfunction: Pointer to the USB interface structure.
- *
- * Return: None.
---- a/drivers/staging/wilc1000/linux_wlan.c
-+++ b/drivers/staging/wilc1000/linux_wlan.c
-@@ -211,7 +211,7 @@ static void deinit_irq(struct net_device
- vif = netdev_priv(dev);
- wilc = vif->wilc;
-
-- /* Deintialize IRQ */
-+ /* Deinitialize IRQ */
- if (wilc->dev_irq_num) {
- free_irq(wilc->dev_irq_num, wilc);
- gpio_free(wilc->gpio);
---- a/drivers/staging/wilc1000/wilc_wfi_cfgoperations.c
-+++ b/drivers/staging/wilc1000/wilc_wfi_cfgoperations.c
-@@ -2359,7 +2359,7 @@ int wilc_deinit_host_int(struct net_devi
- del_timer_sync(&wilc_during_ip_timer);
-
- if (s32Error)
-- netdev_err(net, "Error while deintializing host interface\n");
-+ netdev_err(net, "Error while deinitializing host interface\n");
-
- return s32Error;
- }
---- /dev/null
-+++ b/include/crypto/acompress.h
-@@ -0,0 +1,269 @@
-+/*
-+ * Asynchronous Compression operations
-+ *
-+ * Copyright (c) 2016, Intel Corporation
-+ * Authors: Weigang Li <weigang.li@intel.com>
-+ * Giovanni Cabiddu <giovanni.cabiddu@intel.com>
-+ *
-+ * This program is free software; you can redistribute it and/or modify it
-+ * under the terms of the GNU General Public License as published by the Free
-+ * Software Foundation; either version 2 of the License, or (at your option)
-+ * any later version.
-+ *
-+ */
-+#ifndef _CRYPTO_ACOMP_H
-+#define _CRYPTO_ACOMP_H
-+#include <linux/crypto.h>
-+
-+#define CRYPTO_ACOMP_ALLOC_OUTPUT 0x00000001
-+
-+/**
-+ * struct acomp_req - asynchronous (de)compression request
-+ *
-+ * @base: Common attributes for asynchronous crypto requests
-+ * @src: Source Data
-+ * @dst: Destination data
-+ * @slen: Size of the input buffer
-+ * @dlen: Size of the output buffer and number of bytes produced
-+ * @flags: Internal flags
-+ * @__ctx: Start of private context data
-+ */
-+struct acomp_req {
-+ struct crypto_async_request base;
-+ struct scatterlist *src;
-+ struct scatterlist *dst;
-+ unsigned int slen;
-+ unsigned int dlen;
-+ u32 flags;
-+ void *__ctx[] CRYPTO_MINALIGN_ATTR;
-+};
-+
-+/**
-+ * struct crypto_acomp - user-instantiated objects which encapsulate
-+ * algorithms and core processing logic
-+ *
-+ * @compress: Function performs a compress operation
-+ * @decompress: Function performs a de-compress operation
-+ * @dst_free: Frees destination buffer if allocated inside the
-+ * algorithm
-+ * @reqsize: Context size for (de)compression requests
-+ * @base: Common crypto API algorithm data structure
-+ */
-+struct crypto_acomp {
-+ int (*compress)(struct acomp_req *req);
-+ int (*decompress)(struct acomp_req *req);
-+ void (*dst_free)(struct scatterlist *dst);
-+ unsigned int reqsize;
-+ struct crypto_tfm base;
-+};
-+
-+/**
-+ * struct acomp_alg - asynchronous compression algorithm
-+ *
-+ * @compress: Function performs a compress operation
-+ * @decompress: Function performs a de-compress operation
-+ * @dst_free: Frees destination buffer if allocated inside the algorithm
-+ * @init: Initialize the cryptographic transformation object.
-+ * This function is used to initialize the cryptographic
-+ * transformation object. This function is called only once at
-+ * the instantiation time, right after the transformation context
-+ * was allocated. In case the cryptographic hardware has some
-+ * special requirements which need to be handled by software, this
-+ * function shall check for the precise requirement of the
-+ * transformation and put any software fallbacks in place.
-+ * @exit: Deinitialize the cryptographic transformation object. This is a
-+ * counterpart to @init, used to remove various changes set in
-+ * @init.
-+ *
-+ * @reqsize: Context size for (de)compression requests
-+ * @base: Common crypto API algorithm data structure
-+ */
-+struct acomp_alg {
-+ int (*compress)(struct acomp_req *req);
-+ int (*decompress)(struct acomp_req *req);
-+ void (*dst_free)(struct scatterlist *dst);
-+ int (*init)(struct crypto_acomp *tfm);
-+ void (*exit)(struct crypto_acomp *tfm);
-+ unsigned int reqsize;
-+ struct crypto_alg base;
-+};
-+
-+/**
-+ * DOC: Asynchronous Compression API
-+ *
-+ * The Asynchronous Compression API is used with the algorithms of type
-+ * CRYPTO_ALG_TYPE_ACOMPRESS (listed as type "acomp" in /proc/crypto)
-+ */
-+
-+/**
-+ * crypto_alloc_acomp() -- allocate ACOMPRESS tfm handle
-+ * @alg_name: is the cra_name / name or cra_driver_name / driver name of the
-+ * compression algorithm e.g. "deflate"
-+ * @type: specifies the type of the algorithm
-+ * @mask: specifies the mask for the algorithm
-+ *
-+ * Allocate a handle for a compression algorithm. The returned struct
-+ * crypto_acomp is the handle that is required for any subsequent
-+ * API invocation for the compression operations.
-+ *
-+ * Return: allocated handle in case of success; IS_ERR() is true in case
-+ * of an error, PTR_ERR() returns the error code.
-+ */
-+struct crypto_acomp *crypto_alloc_acomp(const char *alg_name, u32 type,
-+ u32 mask);
-+
-+static inline struct crypto_tfm *crypto_acomp_tfm(struct crypto_acomp *tfm)
-+{
-+ return &tfm->base;
-+}
-+
-+static inline struct acomp_alg *__crypto_acomp_alg(struct crypto_alg *alg)
-+{
-+ return container_of(alg, struct acomp_alg, base);
-+}
-+
-+static inline struct crypto_acomp *__crypto_acomp_tfm(struct crypto_tfm *tfm)
-+{
-+ return container_of(tfm, struct crypto_acomp, base);
-+}
-+
-+static inline struct acomp_alg *crypto_acomp_alg(struct crypto_acomp *tfm)
-+{
-+ return __crypto_acomp_alg(crypto_acomp_tfm(tfm)->__crt_alg);
-+}
-+
-+static inline unsigned int crypto_acomp_reqsize(struct crypto_acomp *tfm)
-+{
-+ return tfm->reqsize;
-+}
-+
-+static inline void acomp_request_set_tfm(struct acomp_req *req,
-+ struct crypto_acomp *tfm)
-+{
-+ req->base.tfm = crypto_acomp_tfm(tfm);
-+}
-+
-+static inline struct crypto_acomp *crypto_acomp_reqtfm(struct acomp_req *req)
-+{
-+ return __crypto_acomp_tfm(req->base.tfm);
-+}
-+
-+/**
-+ * crypto_free_acomp() -- free ACOMPRESS tfm handle
-+ *
-+ * @tfm: ACOMPRESS tfm handle allocated with crypto_alloc_acomp()
-+ */
-+static inline void crypto_free_acomp(struct crypto_acomp *tfm)
-+{
-+ crypto_destroy_tfm(tfm, crypto_acomp_tfm(tfm));
-+}
-+
-+static inline int crypto_has_acomp(const char *alg_name, u32 type, u32 mask)
-+{
-+ type &= ~CRYPTO_ALG_TYPE_MASK;
-+ type |= CRYPTO_ALG_TYPE_ACOMPRESS;
-+ mask |= CRYPTO_ALG_TYPE_MASK;
-+
-+ return crypto_has_alg(alg_name, type, mask);
-+}
-+
-+/**
-+ * acomp_request_alloc() -- allocates asynchronous (de)compression request
-+ *
-+ * @tfm: ACOMPRESS tfm handle allocated with crypto_alloc_acomp()
-+ *
-+ * Return: allocated handle in case of success or NULL in case of an error
-+ */
-+struct acomp_req *acomp_request_alloc(struct crypto_acomp *tfm);
-+
-+/**
-+ * acomp_request_free() -- zeroize and free asynchronous (de)compression
-+ * request as well as the output buffer if allocated
-+ * inside the algorithm
-+ *
-+ * @req: request to free
-+ */
-+void acomp_request_free(struct acomp_req *req);
-+
-+/**
-+ * acomp_request_set_callback() -- Sets an asynchronous callback
-+ *
-+ * Callback will be called when an asynchronous operation on a given
-+ * request is finished.
-+ *
-+ * @req: request that the callback will be set for
-+ * @flgs: specify for instance if the operation may backlog
-+ * @cmlp: callback which will be called
-+ * @data: private data used by the caller
-+ */
-+static inline void acomp_request_set_callback(struct acomp_req *req,
-+ u32 flgs,
-+ crypto_completion_t cmpl,
-+ void *data)
-+{
-+ req->base.complete = cmpl;
-+ req->base.data = data;
-+ req->base.flags = flgs;
-+}
-+
-+/**
-+ * acomp_request_set_params() -- Sets request parameters
-+ *
-+ * Sets parameters required by an acomp operation
-+ *
-+ * @req: asynchronous compress request
-+ * @src: pointer to input buffer scatterlist
-+ * @dst: pointer to output buffer scatterlist. If this is NULL, the
-+ * acomp layer will allocate the output memory
-+ * @slen: size of the input buffer
-+ * @dlen: size of the output buffer. If dst is NULL, this can be used by
-+ * the user to specify the maximum amount of memory to allocate
-+ */
-+static inline void acomp_request_set_params(struct acomp_req *req,
-+ struct scatterlist *src,
-+ struct scatterlist *dst,
-+ unsigned int slen,
-+ unsigned int dlen)
-+{
-+ req->src = src;
-+ req->dst = dst;
-+ req->slen = slen;
-+ req->dlen = dlen;
-+
-+ if (!req->dst)
-+ req->flags |= CRYPTO_ACOMP_ALLOC_OUTPUT;
-+}
-+
-+/**
-+ * crypto_acomp_compress() -- Invoke asynchronous compress operation
-+ *
-+ * Function invokes the asynchronous compress operation
-+ *
-+ * @req: asynchronous compress request
-+ *
-+ * Return: zero on success; error code in case of error
-+ */
-+static inline int crypto_acomp_compress(struct acomp_req *req)
-+{
-+ struct crypto_acomp *tfm = crypto_acomp_reqtfm(req);
-+
-+ return tfm->compress(req);
-+}
-+
-+/**
-+ * crypto_acomp_decompress() -- Invoke asynchronous decompress operation
-+ *
-+ * Function invokes the asynchronous decompress operation
-+ *
-+ * @req: asynchronous compress request
-+ *
-+ * Return: zero on success; error code in case of error
-+ */
-+static inline int crypto_acomp_decompress(struct acomp_req *req)
-+{
-+ struct crypto_acomp *tfm = crypto_acomp_reqtfm(req);
-+
-+ return tfm->decompress(req);
-+}
-+
-+#endif
---- /dev/null
-+++ b/include/crypto/internal/acompress.h
-@@ -0,0 +1,81 @@
-+/*
-+ * Asynchronous Compression operations
-+ *
-+ * Copyright (c) 2016, Intel Corporation
-+ * Authors: Weigang Li <weigang.li@intel.com>
-+ * Giovanni Cabiddu <giovanni.cabiddu@intel.com>
-+ *
-+ * This program is free software; you can redistribute it and/or modify it
-+ * under the terms of the GNU General Public License as published by the Free
-+ * Software Foundation; either version 2 of the License, or (at your option)
-+ * any later version.
-+ *
-+ */
-+#ifndef _CRYPTO_ACOMP_INT_H
-+#define _CRYPTO_ACOMP_INT_H
-+#include <crypto/acompress.h>
-+
-+/*
-+ * Transform internal helpers.
-+ */
-+static inline void *acomp_request_ctx(struct acomp_req *req)
-+{
-+ return req->__ctx;
-+}
-+
-+static inline void *acomp_tfm_ctx(struct crypto_acomp *tfm)
-+{
-+ return tfm->base.__crt_ctx;
-+}
-+
-+static inline void acomp_request_complete(struct acomp_req *req,
-+ int err)
-+{
-+ req->base.complete(&req->base, err);
-+}
-+
-+static inline const char *acomp_alg_name(struct crypto_acomp *tfm)
-+{
-+ return crypto_acomp_tfm(tfm)->__crt_alg->cra_name;
-+}
-+
-+static inline struct acomp_req *__acomp_request_alloc(struct crypto_acomp *tfm)
-+{
-+ struct acomp_req *req;
-+
-+ req = kzalloc(sizeof(*req) + crypto_acomp_reqsize(tfm), GFP_KERNEL);
-+ if (likely(req))
-+ acomp_request_set_tfm(req, tfm);
-+ return req;
-+}
-+
-+static inline void __acomp_request_free(struct acomp_req *req)
-+{
-+ kzfree(req);
-+}
-+
-+/**
-+ * crypto_register_acomp() -- Register asynchronous compression algorithm
-+ *
-+ * Function registers an implementation of an asynchronous
-+ * compression algorithm
-+ *
-+ * @alg: algorithm definition
-+ *
-+ * Return: zero on success; error code in case of error
-+ */
-+int crypto_register_acomp(struct acomp_alg *alg);
-+
-+/**
-+ * crypto_unregister_acomp() -- Unregister asynchronous compression algorithm
-+ *
-+ * Function unregisters an implementation of an asynchronous
-+ * compression algorithm
-+ *
-+ * @alg: algorithm definition
-+ *
-+ * Return: zero on success; error code in case of error
-+ */
-+int crypto_unregister_acomp(struct acomp_alg *alg);
-+
-+#endif
---- /dev/null
-+++ b/include/crypto/internal/scompress.h
-@@ -0,0 +1,136 @@
-+/*
-+ * Synchronous Compression operations
-+ *
-+ * Copyright 2015 LG Electronics Inc.
-+ * Copyright (c) 2016, Intel Corporation
-+ * Author: Giovanni Cabiddu <giovanni.cabiddu@intel.com>
-+ *
-+ * This program is free software; you can redistribute it and/or modify it
-+ * under the terms of the GNU General Public License as published by the Free
-+ * Software Foundation; either version 2 of the License, or (at your option)
-+ * any later version.
-+ *
-+ */
-+#ifndef _CRYPTO_SCOMP_INT_H
-+#define _CRYPTO_SCOMP_INT_H
-+#include <linux/crypto.h>
-+
-+#define SCOMP_SCRATCH_SIZE 131072
-+
-+struct crypto_scomp {
-+ struct crypto_tfm base;
-+};
-+
-+/**
-+ * struct scomp_alg - synchronous compression algorithm
-+ *
-+ * @alloc_ctx: Function allocates algorithm specific context
-+ * @free_ctx: Function frees context allocated with alloc_ctx
-+ * @compress: Function performs a compress operation
-+ * @decompress: Function performs a de-compress operation
-+ * @init: Initialize the cryptographic transformation object.
-+ * This function is used to initialize the cryptographic
-+ * transformation object. This function is called only once at
-+ * the instantiation time, right after the transformation context
-+ * was allocated. In case the cryptographic hardware has some
-+ * special requirements which need to be handled by software, this
-+ * function shall check for the precise requirement of the
-+ * transformation and put any software fallbacks in place.
-+ * @exit: Deinitialize the cryptographic transformation object. This is a
-+ * counterpart to @init, used to remove various changes set in
-+ * @init.
-+ * @base: Common crypto API algorithm data structure
-+ */
-+struct scomp_alg {
-+ void *(*alloc_ctx)(struct crypto_scomp *tfm);
-+ void (*free_ctx)(struct crypto_scomp *tfm, void *ctx);
-+ int (*compress)(struct crypto_scomp *tfm, const u8 *src,
-+ unsigned int slen, u8 *dst, unsigned int *dlen,
-+ void *ctx);
-+ int (*decompress)(struct crypto_scomp *tfm, const u8 *src,
-+ unsigned int slen, u8 *dst, unsigned int *dlen,
-+ void *ctx);
-+ struct crypto_alg base;
-+};
-+
-+static inline struct scomp_alg *__crypto_scomp_alg(struct crypto_alg *alg)
-+{
-+ return container_of(alg, struct scomp_alg, base);
-+}
-+
-+static inline struct crypto_scomp *__crypto_scomp_tfm(struct crypto_tfm *tfm)
-+{
-+ return container_of(tfm, struct crypto_scomp, base);
-+}
-+
-+static inline struct crypto_tfm *crypto_scomp_tfm(struct crypto_scomp *tfm)
-+{
-+ return &tfm->base;
-+}
-+
-+static inline void crypto_free_scomp(struct crypto_scomp *tfm)
-+{
-+ crypto_destroy_tfm(tfm, crypto_scomp_tfm(tfm));
-+}
-+
-+static inline struct scomp_alg *crypto_scomp_alg(struct crypto_scomp *tfm)
-+{
-+ return __crypto_scomp_alg(crypto_scomp_tfm(tfm)->__crt_alg);
-+}
-+
-+static inline void *crypto_scomp_alloc_ctx(struct crypto_scomp *tfm)
-+{
-+ return crypto_scomp_alg(tfm)->alloc_ctx(tfm);
-+}
-+
-+static inline void crypto_scomp_free_ctx(struct crypto_scomp *tfm,
-+ void *ctx)
-+{
-+ return crypto_scomp_alg(tfm)->free_ctx(tfm, ctx);
-+}
-+
-+static inline int crypto_scomp_compress(struct crypto_scomp *tfm,
-+ const u8 *src, unsigned int slen,
-+ u8 *dst, unsigned int *dlen, void *ctx)
-+{
-+ return crypto_scomp_alg(tfm)->compress(tfm, src, slen, dst, dlen, ctx);
-+}
-+
-+static inline int crypto_scomp_decompress(struct crypto_scomp *tfm,
-+ const u8 *src, unsigned int slen,
-+ u8 *dst, unsigned int *dlen,
-+ void *ctx)
-+{
-+ return crypto_scomp_alg(tfm)->decompress(tfm, src, slen, dst, dlen,
-+ ctx);
-+}
-+
-+int crypto_init_scomp_ops_async(struct crypto_tfm *tfm);
-+struct acomp_req *crypto_acomp_scomp_alloc_ctx(struct acomp_req *req);
-+void crypto_acomp_scomp_free_ctx(struct acomp_req *req);
-+
-+/**
-+ * crypto_register_scomp() -- Register synchronous compression algorithm
-+ *
-+ * Function registers an implementation of a synchronous
-+ * compression algorithm
-+ *
-+ * @alg: algorithm definition
-+ *
-+ * Return: zero on success; error code in case of error
-+ */
-+int crypto_register_scomp(struct scomp_alg *alg);
-+
-+/**
-+ * crypto_unregister_scomp() -- Unregister synchronous compression algorithm
-+ *
-+ * Function unregisters an implementation of a synchronous
-+ * compression algorithm
-+ *
-+ * @alg: algorithm definition
-+ *
-+ * Return: zero on success; error code in case of error
-+ */
-+int crypto_unregister_scomp(struct scomp_alg *alg);
-+
-+#endif
---- a/include/linux/crypto.h
-+++ b/include/linux/crypto.h
-@@ -50,6 +50,8 @@
- #define CRYPTO_ALG_TYPE_SKCIPHER 0x00000005
- #define CRYPTO_ALG_TYPE_GIVCIPHER 0x00000006
- #define CRYPTO_ALG_TYPE_KPP 0x00000008
-+#define CRYPTO_ALG_TYPE_ACOMPRESS 0x0000000a
-+#define CRYPTO_ALG_TYPE_SCOMPRESS 0x0000000b
- #define CRYPTO_ALG_TYPE_RNG 0x0000000c
- #define CRYPTO_ALG_TYPE_AKCIPHER 0x0000000d
- #define CRYPTO_ALG_TYPE_DIGEST 0x0000000e
-@@ -60,6 +62,7 @@
- #define CRYPTO_ALG_TYPE_HASH_MASK 0x0000000e
- #define CRYPTO_ALG_TYPE_AHASH_MASK 0x0000000e
- #define CRYPTO_ALG_TYPE_BLKCIPHER_MASK 0x0000000c
-+#define CRYPTO_ALG_TYPE_ACOMPRESS_MASK 0x0000000e
-
- #define CRYPTO_ALG_LARVAL 0x00000010
- #define CRYPTO_ALG_DEAD 0x00000020
---- a/include/uapi/linux/cryptouser.h
-+++ b/include/uapi/linux/cryptouser.h
-@@ -46,6 +46,7 @@ enum crypto_attr_type_t {
- CRYPTOCFGA_REPORT_CIPHER, /* struct crypto_report_cipher */
- CRYPTOCFGA_REPORT_AKCIPHER, /* struct crypto_report_akcipher */
- CRYPTOCFGA_REPORT_KPP, /* struct crypto_report_kpp */
-+ CRYPTOCFGA_REPORT_ACOMP, /* struct crypto_report_acomp */
- __CRYPTOCFGA_MAX
-
- #define CRYPTOCFGA_MAX (__CRYPTOCFGA_MAX - 1)
-@@ -112,5 +113,9 @@ struct crypto_report_kpp {
- char type[CRYPTO_MAX_NAME];
- };
-
-+struct crypto_report_acomp {
-+ char type[CRYPTO_MAX_NAME];
-+};
-+
- #define CRYPTO_REPORT_MAXSIZE (sizeof(struct crypto_user_alg) + \
- sizeof(struct crypto_report_blkcipher))
---- a/scripts/spelling.txt
-+++ b/scripts/spelling.txt
-@@ -305,6 +305,9 @@ defintion||definition
- defintions||definitions
- defualt||default
- defult||default
-+deintializing||deinitializing
-+deintialize||deinitialize
-+deintialized||deinitialized
- deivce||device
- delared||declared
- delare||declare
---- a/sound/soc/amd/acp-pcm-dma.c
-+++ b/sound/soc/amd/acp-pcm-dma.c
-@@ -506,7 +506,7 @@ static int acp_init(void __iomem *acp_mm
- return 0;
- }
-
--/* Deintialize ACP */
-+/* Deinitialize ACP */
- static int acp_deinit(void __iomem *acp_mmio)
- {
- u32 val;
diff --git a/target/linux/layerscape/patches-4.9/805-dma-support-layerscape.patch b/target/linux/layerscape/patches-4.9/805-dma-support-layerscape.patch
deleted file mode 100644
index fb42fdaa5f..0000000000
--- a/target/linux/layerscape/patches-4.9/805-dma-support-layerscape.patch
+++ /dev/null
@@ -1,4372 +0,0 @@
-From d3d537ebe9884e7d945ab74bb02312d0c2c9b08d Mon Sep 17 00:00:00 2001
-From: Yangbo Lu <yangbo.lu@nxp.com>
-Date: Thu, 5 Jul 2018 17:32:53 +0800
-Subject: [PATCH 17/32] dma: support layerscape
-
-This is an integrated patch for layerscape dma support.
-
-Signed-off-by: jiaheng.fan <jiaheng.fan@nxp.com>
-Signed-off-by: Yangbo Lu <yangbo.lu@nxp.com>
----
- drivers/dma/Kconfig | 31 +
- drivers/dma/Makefile | 3 +
- drivers/dma/caam_dma.c | 563 ++++++++++
- drivers/dma/dpaa2-qdma/Kconfig | 8 +
- drivers/dma/dpaa2-qdma/Makefile | 8 +
- drivers/dma/dpaa2-qdma/dpaa2-qdma.c | 940 +++++++++++++++++
- drivers/dma/dpaa2-qdma/dpaa2-qdma.h | 227 +++++
- drivers/dma/dpaa2-qdma/dpdmai.c | 515 ++++++++++
- drivers/dma/dpaa2-qdma/fsl_dpdmai.h | 521 ++++++++++
- drivers/dma/dpaa2-qdma/fsl_dpdmai_cmd.h | 222 ++++
- drivers/dma/fsl-qdma.c | 1243 +++++++++++++++++++++++
- 11 files changed, 4281 insertions(+)
- create mode 100644 drivers/dma/caam_dma.c
- create mode 100644 drivers/dma/dpaa2-qdma/Kconfig
- create mode 100644 drivers/dma/dpaa2-qdma/Makefile
- create mode 100644 drivers/dma/dpaa2-qdma/dpaa2-qdma.c
- create mode 100644 drivers/dma/dpaa2-qdma/dpaa2-qdma.h
- create mode 100644 drivers/dma/dpaa2-qdma/dpdmai.c
- create mode 100644 drivers/dma/dpaa2-qdma/fsl_dpdmai.h
- create mode 100644 drivers/dma/dpaa2-qdma/fsl_dpdmai_cmd.h
- create mode 100644 drivers/dma/fsl-qdma.c
-
---- a/drivers/dma/Kconfig
-+++ b/drivers/dma/Kconfig
-@@ -192,6 +192,20 @@ config FSL_EDMA
- multiplexing capability for DMA request sources(slot).
- This module can be found on Freescale Vybrid and LS-1 SoCs.
-
-+config FSL_QDMA
-+ tristate "Freescale qDMA engine support"
-+ select DMA_ENGINE
-+ select DMA_VIRTUAL_CHANNELS
-+ select DMA_ENGINE_RAID
-+ select ASYNC_TX_ENABLE_CHANNEL_SWITCH
-+ help
-+ Support the Freescale qDMA engine with command queue and legacy mode.
-+ Channel virtualization is supported through enqueuing of DMA jobs to,
-+ or dequeuing DMA jobs from, different work queues.
-+ This module can be found on Freescale LS SoCs.
-+
-+source drivers/dma/dpaa2-qdma/Kconfig
-+
- config FSL_RAID
- tristate "Freescale RAID engine Support"
- depends on FSL_SOC && !ASYNC_TX_ENABLE_CHANNEL_SWITCH
-@@ -564,6 +578,23 @@ config ZX_DMA
- help
- Support the DMA engine for ZTE ZX296702 platform devices.
-
-+config CRYPTO_DEV_FSL_CAAM_DMA
-+ tristate "CAAM DMA engine support"
-+ depends on CRYPTO_DEV_FSL_CAAM_JR
-+ default y
-+ select DMA_ENGINE
-+ select ASYNC_CORE
-+ select ASYNC_TX_ENABLE_CHANNEL_SWITCH
-+ help
-+ Selecting this will offload the DMA operations for users of
-+ the scatter gather memcopy API to the CAAM via job rings. The
-+ CAAM is a hardware module that provides hardware acceleration to
-+ cryptographic operations. It has a built-in DMA controller that can
-+ be programmed to read/write cryptographic data. This module defines
-+ a DMA driver that uses the DMA capabilities of the CAAM.
-+
-+ To compile this as a module, choose M here: the module
-+ will be called caam_dma.
-
- # driver files
- source "drivers/dma/bestcomm/Kconfig"
---- a/drivers/dma/Makefile
-+++ b/drivers/dma/Makefile
-@@ -29,6 +29,8 @@ obj-$(CONFIG_DW_DMAC_CORE) += dw/
- obj-$(CONFIG_EP93XX_DMA) += ep93xx_dma.o
- obj-$(CONFIG_FSL_DMA) += fsldma.o
- obj-$(CONFIG_FSL_EDMA) += fsl-edma.o
-+obj-$(CONFIG_FSL_QDMA) += fsl-qdma.o
-+obj-$(CONFIG_FSL_DPAA2_QDMA) += dpaa2-qdma/
- obj-$(CONFIG_FSL_RAID) += fsl_raid.o
- obj-$(CONFIG_HSU_DMA) += hsu/
- obj-$(CONFIG_IMG_MDC_DMA) += img-mdc-dma.o
-@@ -67,6 +69,7 @@ obj-$(CONFIG_TI_DMA_CROSSBAR) += ti-dma-
- obj-$(CONFIG_TI_EDMA) += edma.o
- obj-$(CONFIG_XGENE_DMA) += xgene-dma.o
- obj-$(CONFIG_ZX_DMA) += zx296702_dma.o
-+obj-$(CONFIG_CRYPTO_DEV_FSL_CAAM_DMA) += caam_dma.o
-
- obj-y += qcom/
- obj-y += xilinx/
---- /dev/null
-+++ b/drivers/dma/caam_dma.c
-@@ -0,0 +1,563 @@
-+/*
-+ * caam support for SG DMA
-+ *
-+ * Copyright 2016 Freescale Semiconductor, Inc
-+ * Copyright 2017 NXP
-+ */
-+
-+#include <linux/module.h>
-+#include <linux/platform_device.h>
-+#include <linux/dma-mapping.h>
-+#include <linux/interrupt.h>
-+#include <linux/slab.h>
-+#include <linux/debugfs.h>
-+
-+#include <linux/dmaengine.h>
-+#include "dmaengine.h"
-+
-+#include "../crypto/caam/regs.h"
-+#include "../crypto/caam/jr.h"
-+#include "../crypto/caam/error.h"
-+#include "../crypto/caam/intern.h"
-+#include "../crypto/caam/desc_constr.h"
-+#include "../crypto/caam/sg_sw_sec4.h"
-+
-+#define DESC_DMA_MEMCPY_LEN ((CAAM_DESC_BYTES_MAX - DESC_JOB_IO_LEN) / \
-+ CAAM_CMD_SZ)
-+
-+/* This is max chunk size of a DMA transfer. If a buffer is larger than this
-+ * value it is internally broken into chunks of max CAAM_DMA_CHUNK_SIZE bytes
-+ * and for each chunk a DMA transfer request is issued.
-+ * This value is the largest number on 16 bits that is a multiple of 256 bytes
-+ * (the largest configurable CAAM DMA burst size).
-+ */
-+#define CAAM_DMA_CHUNK_SIZE 65280
-+
-+struct caam_dma_sh_desc {
-+ u32 desc[DESC_DMA_MEMCPY_LEN] ____cacheline_aligned;
-+ dma_addr_t desc_dma;
-+};
-+
-+/* caam dma extended descriptor */
-+struct caam_dma_edesc {
-+ struct dma_async_tx_descriptor async_tx;
-+ struct list_head node;
-+ struct caam_dma_ctx *ctx;
-+ dma_addr_t src_dma;
-+ dma_addr_t dst_dma;
-+ unsigned int src_len;
-+ unsigned int dst_len;
-+ struct sec4_sg_entry *sec4_sg;
-+ u32 jd[] ____cacheline_aligned;
-+};
-+
-+/*
-+ * caam_dma_ctx - per jr/channel context
-+ * @chan: dma channel used by async_tx API
-+ * @node: list_head used to attach to the global dma_ctx_list
-+ * @jrdev: Job Ring device
-+ * @submit_q: queue of pending (submitted, but not enqueued) jobs
-+ * @done_not_acked: jobs that have been completed by jr, but maybe not acked
-+ * @edesc_lock: protects extended descriptor
-+ */
-+struct caam_dma_ctx {
-+ struct dma_chan chan;
-+ struct list_head node;
-+ struct device *jrdev;
-+ struct list_head submit_q;
-+ struct list_head done_not_acked;
-+ spinlock_t edesc_lock;
-+};
-+
-+static struct dma_device *dma_dev;
-+static struct caam_dma_sh_desc *dma_sh_desc;
-+static LIST_HEAD(dma_ctx_list);
-+
-+static dma_cookie_t caam_dma_tx_submit(struct dma_async_tx_descriptor *tx)
-+{
-+ struct caam_dma_edesc *edesc = NULL;
-+ struct caam_dma_ctx *ctx = NULL;
-+ dma_cookie_t cookie;
-+
-+ edesc = container_of(tx, struct caam_dma_edesc, async_tx);
-+ ctx = container_of(tx->chan, struct caam_dma_ctx, chan);
-+
-+ spin_lock_bh(&ctx->edesc_lock);
-+
-+ cookie = dma_cookie_assign(tx);
-+ list_add_tail(&edesc->node, &ctx->submit_q);
-+
-+ spin_unlock_bh(&ctx->edesc_lock);
-+
-+ return cookie;
-+}
-+
-+static unsigned int caam_dma_sg_dma_len(struct scatterlist *sg,
-+ unsigned int nents)
-+{
-+ unsigned int len;
-+
-+ for (len = 0; sg && nents; sg = sg_next(sg), nents--)
-+ len += sg_dma_len(sg);
-+
-+ return len;
-+}
-+
-+static struct caam_dma_edesc *
-+caam_dma_sg_edesc_alloc(struct dma_chan *chan,
-+ struct scatterlist *dst_sg, unsigned int dst_nents,
-+ struct scatterlist *src_sg, unsigned int src_nents,
-+ unsigned long flags)
-+{
-+ struct caam_dma_ctx *ctx = container_of(chan, struct caam_dma_ctx,
-+ chan);
-+ struct device *jrdev = ctx->jrdev;
-+ struct caam_dma_edesc *edesc;
-+ struct sec4_sg_entry *sec4_sg;
-+ dma_addr_t sec4_sg_dma_src;
-+ unsigned int sec4_sg_bytes;
-+
-+ if (!dst_sg || !src_sg || !dst_nents || !src_nents)
-+ return NULL;
-+
-+ sec4_sg_bytes = (src_nents + dst_nents) * sizeof(*sec4_sg);
-+
-+ edesc = kzalloc(sizeof(*edesc) + DESC_JOB_IO_LEN + sec4_sg_bytes,
-+ GFP_DMA | GFP_NOWAIT);
-+ if (!edesc)
-+ return ERR_PTR(-ENOMEM);
-+
-+ edesc->src_len = caam_dma_sg_dma_len(src_sg, src_nents);
-+ edesc->dst_len = caam_dma_sg_dma_len(dst_sg, dst_nents);
-+ if (edesc->src_len != edesc->dst_len) {
-+ dev_err(jrdev, "%s: src(%u) and dst(%u) len mismatch.\n",
-+ __func__, edesc->src_len, edesc->dst_len);
-+ kfree(edesc);
-+ return ERR_PTR(-EINVAL);
-+ }
-+
-+ dma_async_tx_descriptor_init(&edesc->async_tx, chan);
-+ edesc->async_tx.tx_submit = caam_dma_tx_submit;
-+ edesc->async_tx.flags = flags;
-+ edesc->async_tx.cookie = -EBUSY;
-+
-+ /* Prepare SEC SGs */
-+ edesc->sec4_sg = (void *)edesc + offsetof(struct caam_dma_edesc, jd) +
-+ DESC_JOB_IO_LEN;
-+
-+ sec4_sg = edesc->sec4_sg;
-+ sg_to_sec4_sg_last(src_sg, src_nents, sec4_sg, 0);
-+
-+ sec4_sg += src_nents;
-+ sg_to_sec4_sg_last(dst_sg, dst_nents, sec4_sg, 0);
-+
-+ sec4_sg_dma_src = dma_map_single(jrdev, edesc->sec4_sg, sec4_sg_bytes,
-+ DMA_TO_DEVICE);
-+ if (dma_mapping_error(jrdev, sec4_sg_dma_src)) {
-+ dev_err(jrdev, "error mapping segments to device\n");
-+ kfree(edesc);
-+ return ERR_PTR(-ENOMEM);
-+ }
-+
-+ edesc->src_dma = sec4_sg_dma_src;
-+ edesc->dst_dma = sec4_sg_dma_src + src_nents * sizeof(*sec4_sg);
-+ edesc->ctx = ctx;
-+
-+ return edesc;
-+}
-+
-+static void caam_jr_chan_free_edesc(struct caam_dma_edesc *edesc)
-+{
-+ struct caam_dma_ctx *ctx = edesc->ctx;
-+ struct caam_dma_edesc *_edesc = NULL;
-+
-+ spin_lock_bh(&ctx->edesc_lock);
-+
-+ list_add_tail(&edesc->node, &ctx->done_not_acked);
-+ list_for_each_entry_safe(edesc, _edesc, &ctx->done_not_acked, node) {
-+ if (async_tx_test_ack(&edesc->async_tx)) {
-+ list_del(&edesc->node);
-+ kfree(edesc);
-+ }
-+ }
-+
-+ spin_unlock_bh(&ctx->edesc_lock);
-+}
-+
-+static void caam_dma_done(struct device *dev, u32 *hwdesc, u32 err,
-+ void *context)
-+{
-+ struct caam_dma_edesc *edesc = context;
-+ struct caam_dma_ctx *ctx = edesc->ctx;
-+ dma_async_tx_callback callback;
-+ void *callback_param;
-+
-+ if (err)
-+ caam_jr_strstatus(ctx->jrdev, err);
-+
-+ dma_run_dependencies(&edesc->async_tx);
-+
-+ spin_lock_bh(&ctx->edesc_lock);
-+ dma_cookie_complete(&edesc->async_tx);
-+ spin_unlock_bh(&ctx->edesc_lock);
-+
-+ callback = edesc->async_tx.callback;
-+ callback_param = edesc->async_tx.callback_param;
-+
-+ dma_descriptor_unmap(&edesc->async_tx);
-+
-+ caam_jr_chan_free_edesc(edesc);
-+
-+ if (callback)
-+ callback(callback_param);
-+}
-+
-+static void caam_dma_sg_init_job_desc(struct caam_dma_edesc *edesc)
-+{
-+ u32 *jd = edesc->jd;
-+ u32 *sh_desc = dma_sh_desc->desc;
-+ dma_addr_t desc_dma = dma_sh_desc->desc_dma;
-+
-+ /* init the job descriptor */
-+ init_job_desc_shared(jd, desc_dma, desc_len(sh_desc), HDR_REVERSE);
-+
-+ /* set SEQIN PTR */
-+ append_seq_in_ptr(jd, edesc->src_dma, edesc->src_len, LDST_SGF);
-+
-+ /* set SEQOUT PTR */
-+ append_seq_out_ptr(jd, edesc->dst_dma, edesc->dst_len, LDST_SGF);
-+
-+#ifdef DEBUG
-+ print_hex_dump(KERN_ERR, "caam dma desc@" __stringify(__LINE__) ": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, jd, desc_bytes(jd), 1);
-+#endif
-+}
-+
-+/* This function can be called from an interrupt context */
-+static struct dma_async_tx_descriptor *
-+caam_dma_prep_sg(struct dma_chan *chan, struct scatterlist *dst_sg,
-+ unsigned int dst_nents, struct scatterlist *src_sg,
-+ unsigned int src_nents, unsigned long flags)
-+{
-+ struct caam_dma_edesc *edesc;
-+
-+ /* allocate extended descriptor */
-+ edesc = caam_dma_sg_edesc_alloc(chan, dst_sg, dst_nents, src_sg,
-+ src_nents, flags);
-+ if (IS_ERR_OR_NULL(edesc))
-+ return ERR_CAST(edesc);
-+
-+ /* Initialize job descriptor */
-+ caam_dma_sg_init_job_desc(edesc);
-+
-+ return &edesc->async_tx;
-+}
-+
-+static void caam_dma_memcpy_init_job_desc(struct caam_dma_edesc *edesc)
-+{
-+ u32 *jd = edesc->jd;
-+ u32 *sh_desc = dma_sh_desc->desc;
-+ dma_addr_t desc_dma = dma_sh_desc->desc_dma;
-+
-+ /* init the job descriptor */
-+ init_job_desc_shared(jd, desc_dma, desc_len(sh_desc), HDR_REVERSE);
-+
-+ /* set SEQIN PTR */
-+ append_seq_in_ptr(jd, edesc->src_dma, edesc->src_len, 0);
-+
-+ /* set SEQOUT PTR */
-+ append_seq_out_ptr(jd, edesc->dst_dma, edesc->dst_len, 0);
-+
-+#ifdef DEBUG
-+ print_hex_dump(KERN_ERR, "caam dma desc@" __stringify(__LINE__) ": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, jd, desc_bytes(jd), 1);
-+#endif
-+}
-+
-+static struct dma_async_tx_descriptor *
-+caam_dma_prep_memcpy(struct dma_chan *chan, dma_addr_t dst, dma_addr_t src,
-+ size_t len, unsigned long flags)
-+{
-+ struct caam_dma_edesc *edesc;
-+ struct caam_dma_ctx *ctx = container_of(chan, struct caam_dma_ctx,
-+ chan);
-+
-+ edesc = kzalloc(sizeof(*edesc) + DESC_JOB_IO_LEN, GFP_DMA | GFP_NOWAIT);
-+ if (!edesc)
-+ return ERR_PTR(-ENOMEM);
-+
-+ dma_async_tx_descriptor_init(&edesc->async_tx, chan);
-+ edesc->async_tx.tx_submit = caam_dma_tx_submit;
-+ edesc->async_tx.flags = flags;
-+ edesc->async_tx.cookie = -EBUSY;
-+
-+ edesc->src_dma = src;
-+ edesc->src_len = len;
-+ edesc->dst_dma = dst;
-+ edesc->dst_len = len;
-+ edesc->ctx = ctx;
-+
-+ caam_dma_memcpy_init_job_desc(edesc);
-+
-+ return &edesc->async_tx;
-+}
-+
-+/* This function can be called in an interrupt context */
-+static void caam_dma_issue_pending(struct dma_chan *chan)
-+{
-+ struct caam_dma_ctx *ctx = container_of(chan, struct caam_dma_ctx,
-+ chan);
-+ struct caam_dma_edesc *edesc, *_edesc;
-+
-+ spin_lock_bh(&ctx->edesc_lock);
-+ list_for_each_entry_safe(edesc, _edesc, &ctx->submit_q, node) {
-+ if (caam_jr_enqueue(ctx->jrdev, edesc->jd,
-+ caam_dma_done, edesc) < 0)
-+ break;
-+ list_del(&edesc->node);
-+ }
-+ spin_unlock_bh(&ctx->edesc_lock);
-+}
-+
-+static void caam_dma_free_chan_resources(struct dma_chan *chan)
-+{
-+ struct caam_dma_ctx *ctx = container_of(chan, struct caam_dma_ctx,
-+ chan);
-+ struct caam_dma_edesc *edesc, *_edesc;
-+
-+ spin_lock_bh(&ctx->edesc_lock);
-+ list_for_each_entry_safe(edesc, _edesc, &ctx->submit_q, node) {
-+ list_del(&edesc->node);
-+ kfree(edesc);
-+ }
-+ list_for_each_entry_safe(edesc, _edesc, &ctx->done_not_acked, node) {
-+ list_del(&edesc->node);
-+ kfree(edesc);
-+ }
-+ spin_unlock_bh(&ctx->edesc_lock);
-+}
-+
-+static int caam_dma_jr_chan_bind(void)
-+{
-+ struct device *jrdev;
-+ struct caam_dma_ctx *ctx;
-+ int bonds = 0;
-+ int i;
-+
-+ for (i = 0; i < caam_jr_driver_probed(); i++) {
-+ jrdev = caam_jridx_alloc(i);
-+ if (IS_ERR(jrdev)) {
-+ pr_err("job ring device %d allocation failed\n", i);
-+ continue;
-+ }
-+
-+ ctx = kzalloc(sizeof(*ctx), GFP_KERNEL);
-+ if (!ctx) {
-+ caam_jr_free(jrdev);
-+ continue;
-+ }
-+
-+ ctx->chan.device = dma_dev;
-+ ctx->chan.private = ctx;
-+
-+ ctx->jrdev = jrdev;
-+
-+ INIT_LIST_HEAD(&ctx->submit_q);
-+ INIT_LIST_HEAD(&ctx->done_not_acked);
-+ INIT_LIST_HEAD(&ctx->node);
-+ spin_lock_init(&ctx->edesc_lock);
-+
-+ dma_cookie_init(&ctx->chan);
-+
-+ /* add the context of this channel to the context list */
-+ list_add_tail(&ctx->node, &dma_ctx_list);
-+
-+ /* add this channel to the device chan list */
-+ list_add_tail(&ctx->chan.device_node, &dma_dev->channels);
-+
-+ bonds++;
-+ }
-+
-+ return bonds;
-+}
-+
-+static inline void caam_jr_dma_free(struct dma_chan *chan)
-+{
-+ struct caam_dma_ctx *ctx = container_of(chan, struct caam_dma_ctx,
-+ chan);
-+
-+ list_del(&ctx->node);
-+ list_del(&chan->device_node);
-+ caam_jr_free(ctx->jrdev);
-+ kfree(ctx);
-+}
-+
-+static void set_caam_dma_desc(u32 *desc)
-+{
-+ u32 *jmp_cmd;
-+
-+ /* dma shared descriptor */
-+ init_sh_desc(desc, HDR_SHARE_NEVER | (1 << HDR_START_IDX_SHIFT));
-+
-+ /* REG1 = CAAM_DMA_CHUNK_SIZE */
-+ append_math_add_imm_u32(desc, REG1, ZERO, IMM, CAAM_DMA_CHUNK_SIZE);
-+
-+ /* REG0 = SEQINLEN - CAAM_DMA_CHUNK_SIZE */
-+ append_math_sub_imm_u32(desc, REG0, SEQINLEN, IMM, CAAM_DMA_CHUNK_SIZE);
-+
-+ /* if (REG0 > 0)
-+ * jmp to LABEL1
-+ */
-+ jmp_cmd = append_jump(desc, JUMP_TEST_INVALL | JUMP_COND_MATH_N |
-+ JUMP_COND_MATH_Z);
-+
-+ /* REG1 = SEQINLEN */
-+ append_math_sub(desc, REG1, SEQINLEN, ZERO, CAAM_CMD_SZ);
-+
-+ /* LABEL1 */
-+ set_jump_tgt_here(desc, jmp_cmd);
-+
-+ /* VARSEQINLEN = REG1 */
-+ append_math_add(desc, VARSEQINLEN, REG1, ZERO, CAAM_CMD_SZ);
-+
-+ /* VARSEQOUTLEN = REG1 */
-+ append_math_add(desc, VARSEQOUTLEN, REG1, ZERO, CAAM_CMD_SZ);
-+
-+ /* do FIFO STORE */
-+ append_seq_fifo_store(desc, 0, FIFOST_TYPE_METADATA | LDST_VLF);
-+
-+ /* do FIFO LOAD */
-+ append_seq_fifo_load(desc, 0, FIFOLD_CLASS_CLASS1 |
-+ FIFOLD_TYPE_IFIFO | LDST_VLF);
-+
-+ /* if (REG0 > 0)
-+ * jmp 0xF8 (after shared desc header)
-+ */
-+ append_jump(desc, JUMP_TEST_INVALL | JUMP_COND_MATH_N |
-+ JUMP_COND_MATH_Z | 0xF8);
-+
-+#ifdef DEBUG
-+ print_hex_dump(KERN_ERR, "caam dma shdesc@" __stringify(__LINE__) ": ",
-+ DUMP_PREFIX_ADDRESS, 16, 4, desc, desc_bytes(desc), 1);
-+#endif
-+}
-+
-+static int __init caam_dma_probe(struct platform_device *pdev)
-+{
-+ struct device *dev = &pdev->dev;
-+ struct device *ctrldev = dev->parent;
-+ struct dma_chan *chan, *_chan;
-+ u32 *sh_desc;
-+ int err = -ENOMEM;
-+ int bonds;
-+
-+ if (!caam_jr_driver_probed()) {
-+ dev_info(dev, "Defer probing after JR driver probing\n");
-+ return -EPROBE_DEFER;
-+ }
-+
-+ dma_dev = kzalloc(sizeof(*dma_dev), GFP_KERNEL);
-+ if (!dma_dev)
-+ return -ENOMEM;
-+
-+ dma_sh_desc = kzalloc(sizeof(*dma_sh_desc), GFP_KERNEL | GFP_DMA);
-+ if (!dma_sh_desc)
-+ goto desc_err;
-+
-+ sh_desc = dma_sh_desc->desc;
-+ set_caam_dma_desc(sh_desc);
-+ dma_sh_desc->desc_dma = dma_map_single(ctrldev, sh_desc,
-+ desc_bytes(sh_desc),
-+ DMA_TO_DEVICE);
-+ if (dma_mapping_error(ctrldev, dma_sh_desc->desc_dma)) {
-+ dev_err(dev, "unable to map dma descriptor\n");
-+ goto map_err;
-+ }
-+
-+ INIT_LIST_HEAD(&dma_dev->channels);
-+
-+ bonds = caam_dma_jr_chan_bind();
-+ if (!bonds) {
-+ err = -ENODEV;
-+ goto jr_bind_err;
-+ }
-+
-+ dma_dev->dev = dev;
-+ dma_dev->residue_granularity = DMA_RESIDUE_GRANULARITY_DESCRIPTOR;
-+ dma_cap_set(DMA_SG, dma_dev->cap_mask);
-+ dma_cap_set(DMA_MEMCPY, dma_dev->cap_mask);
-+ dma_cap_set(DMA_PRIVATE, dma_dev->cap_mask);
-+ dma_dev->device_tx_status = dma_cookie_status;
-+ dma_dev->device_issue_pending = caam_dma_issue_pending;
-+ dma_dev->device_prep_dma_sg = caam_dma_prep_sg;
-+ dma_dev->device_prep_dma_memcpy = caam_dma_prep_memcpy;
-+ dma_dev->device_free_chan_resources = caam_dma_free_chan_resources;
-+
-+ err = dma_async_device_register(dma_dev);
-+ if (err) {
-+ dev_err(dev, "Failed to register CAAM DMA engine\n");
-+ goto jr_bind_err;
-+ }
-+
-+ dev_info(dev, "caam dma support with %d job rings\n", bonds);
-+
-+ return err;
-+
-+jr_bind_err:
-+ list_for_each_entry_safe(chan, _chan, &dma_dev->channels, device_node)
-+ caam_jr_dma_free(chan);
-+
-+ dma_unmap_single(ctrldev, dma_sh_desc->desc_dma, desc_bytes(sh_desc),
-+ DMA_TO_DEVICE);
-+map_err:
-+ kfree(dma_sh_desc);
-+desc_err:
-+ kfree(dma_dev);
-+ return err;
-+}
-+
-+static int caam_dma_remove(struct platform_device *pdev)
-+{
-+ struct device *dev = &pdev->dev;
-+ struct device *ctrldev = dev->parent;
-+ struct caam_dma_ctx *ctx, *_ctx;
-+
-+ dma_async_device_unregister(dma_dev);
-+
-+ list_for_each_entry_safe(ctx, _ctx, &dma_ctx_list, node) {
-+ list_del(&ctx->node);
-+ caam_jr_free(ctx->jrdev);
-+ kfree(ctx);
-+ }
-+
-+ dma_unmap_single(ctrldev, dma_sh_desc->desc_dma,
-+ desc_bytes(dma_sh_desc->desc), DMA_TO_DEVICE);
-+
-+ kfree(dma_sh_desc);
-+ kfree(dma_dev);
-+
-+ dev_info(dev, "caam dma support disabled\n");
-+ return 0;
-+}
-+
-+static const struct of_device_id caam_dma_match[] = {
-+ { .compatible = "fsl,sec-v5.4-dma", },
-+ { .compatible = "fsl,sec-v5.0-dma", },
-+ { .compatible = "fsl,sec-v4.0-dma", },
-+ {},
-+};
-+MODULE_DEVICE_TABLE(of, caam_dma_match);
-+
-+static struct platform_driver caam_dma_driver = {
-+ .driver = {
-+ .name = "caam-dma",
-+ .of_match_table = caam_dma_match,
-+ },
-+ .probe = caam_dma_probe,
-+ .remove = caam_dma_remove,
-+};
-+module_platform_driver(caam_dma_driver);
-+
-+MODULE_LICENSE("Dual BSD/GPL");
-+MODULE_DESCRIPTION("NXP CAAM support for SG DMA");
-+MODULE_AUTHOR("NXP Semiconductors");
---- /dev/null
-+++ b/drivers/dma/dpaa2-qdma/Kconfig
-@@ -0,0 +1,8 @@
-+menuconfig FSL_DPAA2_QDMA
-+ tristate "NXP DPAA2 QDMA"
-+ depends on FSL_MC_BUS && FSL_MC_DPIO
-+ select DMA_ENGINE
-+ select DMA_VIRTUAL_CHANNELS
-+ ---help---
-+ NXP Data Path Acceleration Architecture 2 QDMA driver,
-+ using the NXP MC bus driver.
---- /dev/null
-+++ b/drivers/dma/dpaa2-qdma/Makefile
-@@ -0,0 +1,8 @@
-+#
-+# Makefile for the NXP DPAA2 CAAM controllers
-+#
-+ccflags-y += -DVERSION=\"\"
-+
-+obj-$(CONFIG_FSL_DPAA2_QDMA) += fsl-dpaa2-qdma.o
-+
-+fsl-dpaa2-qdma-objs := dpaa2-qdma.o dpdmai.o
---- /dev/null
-+++ b/drivers/dma/dpaa2-qdma/dpaa2-qdma.c
-@@ -0,0 +1,940 @@
-+/*
-+ * drivers/dma/dpaa2-qdma/dpaa2-qdma.c
-+ *
-+ * Copyright 2015-2017 NXP Semiconductor, Inc.
-+ * Author: Changming Huang <jerry.huang@nxp.com>
-+ *
-+ * Driver for the NXP QDMA engine with QMan mode.
-+ * Channel virtualization is supported through enqueuing of DMA jobs to,
-+ * or dequeuing DMA jobs from different work queues with QMan portal.
-+ * This module can be found on NXP LS2 SoCs.
-+ *
-+ * This program is free software; you can redistribute it and/or modify it
-+ * under the terms of the GNU General Public License as published by the
-+ * Free Software Foundation; either version 2 of the License, or (at your
-+ * option) any later version.
-+ */
-+
-+#include <linux/init.h>
-+#include <linux/module.h>
-+#include <linux/interrupt.h>
-+#include <linux/clk.h>
-+#include <linux/dma-mapping.h>
-+#include <linux/dmapool.h>
-+#include <linux/slab.h>
-+#include <linux/spinlock.h>
-+#include <linux/of.h>
-+#include <linux/of_device.h>
-+#include <linux/of_address.h>
-+#include <linux/of_irq.h>
-+#include <linux/of_dma.h>
-+#include <linux/types.h>
-+#include <linux/delay.h>
-+#include <linux/iommu.h>
-+
-+#include "../virt-dma.h"
-+
-+#include <linux/fsl/mc.h>
-+#include "../../../drivers/staging/fsl-mc/include/dpaa2-io.h"
-+#include "../../../drivers/staging/fsl-mc/include/dpaa2-fd.h"
-+#include "fsl_dpdmai_cmd.h"
-+#include "fsl_dpdmai.h"
-+#include "dpaa2-qdma.h"
-+
-+static bool smmu_disable = true;
-+
-+static struct dpaa2_qdma_chan *to_dpaa2_qdma_chan(struct dma_chan *chan)
-+{
-+ return container_of(chan, struct dpaa2_qdma_chan, vchan.chan);
-+}
-+
-+static struct dpaa2_qdma_comp *to_fsl_qdma_comp(struct virt_dma_desc *vd)
-+{
-+ return container_of(vd, struct dpaa2_qdma_comp, vdesc);
-+}
-+
-+static int dpaa2_qdma_alloc_chan_resources(struct dma_chan *chan)
-+{
-+ return 0;
-+}
-+
-+static void dpaa2_qdma_free_chan_resources(struct dma_chan *chan)
-+{
-+ struct dpaa2_qdma_chan *dpaa2_chan = to_dpaa2_qdma_chan(chan);
-+ unsigned long flags;
-+ LIST_HEAD(head);
-+
-+ spin_lock_irqsave(&dpaa2_chan->vchan.lock, flags);
-+ vchan_get_all_descriptors(&dpaa2_chan->vchan, &head);
-+ spin_unlock_irqrestore(&dpaa2_chan->vchan.lock, flags);
-+
-+ vchan_dma_desc_free_list(&dpaa2_chan->vchan, &head);
-+}
-+
-+/*
-+ * Request a command descriptor for enqueue.
-+ */
-+static struct dpaa2_qdma_comp *
-+dpaa2_qdma_request_desc(struct dpaa2_qdma_chan *dpaa2_chan)
-+{
-+ struct dpaa2_qdma_comp *comp_temp = NULL;
-+ unsigned long flags;
-+
-+ spin_lock_irqsave(&dpaa2_chan->queue_lock, flags);
-+ if (list_empty(&dpaa2_chan->comp_free)) {
-+ spin_unlock_irqrestore(&dpaa2_chan->queue_lock, flags);
-+ comp_temp = kzalloc(sizeof(*comp_temp), GFP_KERNEL);
-+ if (!comp_temp)
-+ goto err;
-+ comp_temp->fd_virt_addr = dma_pool_alloc(dpaa2_chan->fd_pool,
-+ GFP_NOWAIT, &comp_temp->fd_bus_addr);
-+ if (!comp_temp->fd_virt_addr)
-+ goto err;
-+
-+ comp_temp->fl_virt_addr =
-+ (void *)((struct dpaa2_fd *)
-+ comp_temp->fd_virt_addr + 1);
-+ comp_temp->fl_bus_addr = comp_temp->fd_bus_addr +
-+ sizeof(struct dpaa2_fd);
-+ comp_temp->desc_virt_addr =
-+ (void *)((struct dpaa2_fl_entry *)
-+ comp_temp->fl_virt_addr + 3);
-+ comp_temp->desc_bus_addr = comp_temp->fl_bus_addr +
-+ sizeof(struct dpaa2_fl_entry) * 3;
-+
-+ comp_temp->qchan = dpaa2_chan;
-+ comp_temp->sg_blk_num = 0;
-+ INIT_LIST_HEAD(&comp_temp->sg_src_head);
-+ INIT_LIST_HEAD(&comp_temp->sg_dst_head);
-+ return comp_temp;
-+ }
-+ comp_temp = list_first_entry(&dpaa2_chan->comp_free,
-+ struct dpaa2_qdma_comp, list);
-+ list_del(&comp_temp->list);
-+ spin_unlock_irqrestore(&dpaa2_chan->queue_lock, flags);
-+
-+ comp_temp->qchan = dpaa2_chan;
-+err:
-+ return comp_temp;
-+}
-+
-+static void dpaa2_qdma_populate_fd(uint32_t format,
-+ struct dpaa2_qdma_comp *dpaa2_comp)
-+{
-+ struct dpaa2_fd *fd;
-+
-+ fd = (struct dpaa2_fd *)dpaa2_comp->fd_virt_addr;
-+ memset(fd, 0, sizeof(struct dpaa2_fd));
-+
-+ /* fd populated */
-+ dpaa2_fd_set_addr(fd, dpaa2_comp->fl_bus_addr);
-+ /* Bypass memory translation, Frame list format, short length disable */
-+ /* we need to disable BMT if fsl-mc use iova addr */
-+ if (smmu_disable)
-+ dpaa2_fd_set_bpid(fd, QMAN_FD_BMT_ENABLE);
-+ dpaa2_fd_set_format(fd, QMAN_FD_FMT_ENABLE | QMAN_FD_SL_DISABLE);
-+
-+ dpaa2_fd_set_frc(fd, format | QDMA_SER_CTX);
-+}
-+
-+/* first frame list for descriptor buffer */
-+static void dpaa2_qdma_populate_first_framel(
-+ struct dpaa2_fl_entry *f_list,
-+ struct dpaa2_qdma_comp *dpaa2_comp)
-+{
-+ struct dpaa2_qdma_sd_d *sdd;
-+
-+ sdd = (struct dpaa2_qdma_sd_d *)dpaa2_comp->desc_virt_addr;
-+ memset(sdd, 0, 2 * (sizeof(*sdd)));
-+ /* source and destination descriptor */
-+ sdd->cmd = cpu_to_le32(QDMA_SD_CMD_RDTTYPE_COHERENT); /* source descriptor CMD */
-+ sdd++;
-+ sdd->cmd = cpu_to_le32(QDMA_DD_CMD_WRTTYPE_COHERENT); /* dest descriptor CMD */
-+
-+ memset(f_list, 0, sizeof(struct dpaa2_fl_entry));
-+ /* first frame list to source descriptor */
-+
-+ dpaa2_fl_set_addr(f_list, dpaa2_comp->desc_bus_addr);
-+ dpaa2_fl_set_len(f_list, 0x20);
-+ dpaa2_fl_set_format(f_list, QDMA_FL_FMT_SBF | QDMA_FL_SL_LONG);
-+
-+ if (smmu_disable)
-+ f_list->bpid = cpu_to_le16(QDMA_FL_BMT_ENABLE); /* bypass memory translation */
-+}
-+
-+/* source and destination frame list */
-+static void dpaa2_qdma_populate_frames(struct dpaa2_fl_entry *f_list,
-+ dma_addr_t dst, dma_addr_t src, size_t len, uint8_t fmt)
-+{
-+ /* source frame list to source buffer */
-+ memset(f_list, 0, sizeof(struct dpaa2_fl_entry));
-+
-+
-+ dpaa2_fl_set_addr(f_list, src);
-+ dpaa2_fl_set_len(f_list, len);
-+ dpaa2_fl_set_format(f_list, (fmt | QDMA_FL_SL_LONG)); /* single buffer frame or scatter gather frame */
-+ if (smmu_disable)
-+ f_list->bpid = cpu_to_le16(QDMA_FL_BMT_ENABLE); /* bypass memory translation */
-+
-+ f_list++;
-+ /* destination frame list to destination buffer */
-+ memset(f_list, 0, sizeof(struct dpaa2_fl_entry));
-+
-+ dpaa2_fl_set_addr(f_list, dst);
-+ dpaa2_fl_set_len(f_list, len);
-+ dpaa2_fl_set_format(f_list, (fmt | QDMA_FL_SL_LONG));
-+ dpaa2_fl_set_final(f_list, QDMA_FL_F); /* single buffer frame or scatter gather frame */
-+ if (smmu_disable)
-+ f_list->bpid = cpu_to_le16(QDMA_FL_BMT_ENABLE); /* bypass memory translation */
-+}
-+
-+static struct dma_async_tx_descriptor *dpaa2_qdma_prep_memcpy(
-+ struct dma_chan *chan, dma_addr_t dst,
-+ dma_addr_t src, size_t len, unsigned long flags)
-+{
-+ struct dpaa2_qdma_chan *dpaa2_chan = to_dpaa2_qdma_chan(chan);
-+ struct dpaa2_qdma_comp *dpaa2_comp;
-+ struct dpaa2_fl_entry *f_list;
-+ uint32_t format;
-+
-+ dpaa2_comp = dpaa2_qdma_request_desc(dpaa2_chan);
-+
-+#ifdef LONG_FORMAT
-+ format = QDMA_FD_LONG_FORMAT;
-+#else
-+ format = QDMA_FD_SHORT_FORMAT;
-+#endif
-+ /* populate Frame descriptor */
-+ dpaa2_qdma_populate_fd(format, dpaa2_comp);
-+
-+ f_list = (struct dpaa2_fl_entry *)dpaa2_comp->fl_virt_addr;
-+
-+#ifdef LONG_FORMAT
-+ /* first frame list for descriptor buffer (logn format) */
-+ dpaa2_qdma_populate_first_framel(f_list, dpaa2_comp);
-+
-+ f_list++;
-+#endif
-+
-+ dpaa2_qdma_populate_frames(f_list, dst, src, len, QDMA_FL_FMT_SBF);
-+
-+ return vchan_tx_prep(&dpaa2_chan->vchan, &dpaa2_comp->vdesc, flags);
-+}
-+
-+static struct qdma_sg_blk *dpaa2_qdma_get_sg_blk(
-+ struct dpaa2_qdma_comp *dpaa2_comp,
-+ struct dpaa2_qdma_chan *dpaa2_chan)
-+{
-+ struct qdma_sg_blk *sg_blk = NULL;
-+ dma_addr_t phy_sgb;
-+ unsigned long flags;
-+
-+ spin_lock_irqsave(&dpaa2_chan->queue_lock, flags);
-+ if (list_empty(&dpaa2_chan->sgb_free)) {
-+ sg_blk = (struct qdma_sg_blk *)dma_pool_alloc(
-+ dpaa2_chan->sg_blk_pool,
-+ GFP_NOWAIT, &phy_sgb);
-+ if (!sg_blk) {
-+ spin_unlock_irqrestore(&dpaa2_chan->queue_lock, flags);
-+ return sg_blk;
-+ }
-+ sg_blk->blk_virt_addr = (void *)(sg_blk + 1);
-+ sg_blk->blk_bus_addr = phy_sgb + sizeof(*sg_blk);
-+ } else {
-+ sg_blk = list_first_entry(&dpaa2_chan->sgb_free,
-+ struct qdma_sg_blk, list);
-+ list_del(&sg_blk->list);
-+ }
-+ spin_unlock_irqrestore(&dpaa2_chan->queue_lock, flags);
-+
-+ return sg_blk;
-+}
-+
-+static uint32_t dpaa2_qdma_populate_sg(struct device *dev,
-+ struct dpaa2_qdma_chan *dpaa2_chan,
-+ struct dpaa2_qdma_comp *dpaa2_comp,
-+ struct scatterlist *dst_sg, u32 dst_nents,
-+ struct scatterlist *src_sg, u32 src_nents)
-+{
-+ struct dpaa2_qdma_sg *src_sge;
-+ struct dpaa2_qdma_sg *dst_sge;
-+ struct qdma_sg_blk *sg_blk;
-+ struct qdma_sg_blk *sg_blk_dst;
-+ dma_addr_t src;
-+ dma_addr_t dst;
-+ uint32_t num;
-+ uint32_t blocks;
-+ uint32_t len = 0;
-+ uint32_t total_len = 0;
-+ int i, j = 0;
-+
-+ num = min(dst_nents, src_nents);
-+ blocks = num / (NUM_SG_PER_BLK - 1);
-+ if (num % (NUM_SG_PER_BLK - 1))
-+ blocks += 1;
-+ if (dpaa2_comp->sg_blk_num < blocks) {
-+ len = blocks - dpaa2_comp->sg_blk_num;
-+ for (i = 0; i < len; i++) {
-+ /* source sg blocks */
-+ sg_blk = dpaa2_qdma_get_sg_blk(dpaa2_comp, dpaa2_chan);
-+ if (!sg_blk)
-+ return 0;
-+ list_add_tail(&sg_blk->list, &dpaa2_comp->sg_src_head);
-+ /* destination sg blocks */
-+ sg_blk = dpaa2_qdma_get_sg_blk(dpaa2_comp, dpaa2_chan);
-+ if (!sg_blk)
-+ return 0;
-+ list_add_tail(&sg_blk->list, &dpaa2_comp->sg_dst_head);
-+ }
-+ } else {
-+ len = dpaa2_comp->sg_blk_num - blocks;
-+ for (i = 0; i < len; i++) {
-+ spin_lock(&dpaa2_chan->queue_lock);
-+ /* handle source sg blocks */
-+ sg_blk = list_first_entry(&dpaa2_comp->sg_src_head,
-+ struct qdma_sg_blk, list);
-+ list_del(&sg_blk->list);
-+ list_add_tail(&sg_blk->list, &dpaa2_chan->sgb_free);
-+ /* handle destination sg blocks */
-+ sg_blk = list_first_entry(&dpaa2_comp->sg_dst_head,
-+ struct qdma_sg_blk, list);
-+ list_del(&sg_blk->list);
-+ list_add_tail(&sg_blk->list, &dpaa2_chan->sgb_free);
-+ spin_unlock(&dpaa2_chan->queue_lock);
-+ }
-+ }
-+ dpaa2_comp->sg_blk_num = blocks;
-+
-+ /* get the first source sg phy address */
-+ sg_blk = list_first_entry(&dpaa2_comp->sg_src_head,
-+ struct qdma_sg_blk, list);
-+ dpaa2_comp->sge_src_bus_addr = sg_blk->blk_bus_addr;
-+ /* get the first destinaiton sg phy address */
-+ sg_blk_dst = list_first_entry(&dpaa2_comp->sg_dst_head,
-+ struct qdma_sg_blk, list);
-+ dpaa2_comp->sge_dst_bus_addr = sg_blk_dst->blk_bus_addr;
-+
-+ for (i = 0; i < blocks; i++) {
-+ src_sge = (struct dpaa2_qdma_sg *)sg_blk->blk_virt_addr;
-+ dst_sge = (struct dpaa2_qdma_sg *)sg_blk_dst->blk_virt_addr;
-+
-+ for (j = 0; j < (NUM_SG_PER_BLK - 1); j++) {
-+ len = min(sg_dma_len(dst_sg), sg_dma_len(src_sg));
-+ if (0 == len)
-+ goto fetch;
-+ total_len += len;
-+ src = sg_dma_address(src_sg);
-+ dst = sg_dma_address(dst_sg);
-+
-+ /* source SG */
-+ src_sge->addr_lo = src;
-+ src_sge->addr_hi = (src >> 32);
-+ src_sge->data_len.data_len_sl0 = len;
-+ src_sge->ctrl.sl = QDMA_SG_SL_LONG;
-+ src_sge->ctrl.fmt = QDMA_SG_FMT_SDB;
-+ /* destination SG */
-+ dst_sge->addr_lo = dst;
-+ dst_sge->addr_hi = (dst >> 32);
-+ dst_sge->data_len.data_len_sl0 = len;
-+ dst_sge->ctrl.sl = QDMA_SG_SL_LONG;
-+ dst_sge->ctrl.fmt = QDMA_SG_FMT_SDB;
-+fetch:
-+ num--;
-+ if (0 == num) {
-+ src_sge->ctrl.f = QDMA_SG_F;
-+ dst_sge->ctrl.f = QDMA_SG_F;
-+ goto end;
-+ }
-+ dst_sg = sg_next(dst_sg);
-+ src_sg = sg_next(src_sg);
-+ src_sge++;
-+ dst_sge++;
-+ if (j == (NUM_SG_PER_BLK - 2)) {
-+ /* for next blocks, extension */
-+ sg_blk = list_next_entry(sg_blk, list);
-+ sg_blk_dst = list_next_entry(sg_blk_dst, list);
-+ src_sge->addr_lo = sg_blk->blk_bus_addr;
-+ src_sge->addr_hi = sg_blk->blk_bus_addr >> 32;
-+ src_sge->ctrl.sl = QDMA_SG_SL_LONG;
-+ src_sge->ctrl.fmt = QDMA_SG_FMT_SGTE;
-+ dst_sge->addr_lo = sg_blk_dst->blk_bus_addr;
-+ dst_sge->addr_hi =
-+ sg_blk_dst->blk_bus_addr >> 32;
-+ dst_sge->ctrl.sl = QDMA_SG_SL_LONG;
-+ dst_sge->ctrl.fmt = QDMA_SG_FMT_SGTE;
-+ }
-+ }
-+ }
-+
-+end:
-+ return total_len;
-+}
-+
-+static enum dma_status dpaa2_qdma_tx_status(struct dma_chan *chan,
-+ dma_cookie_t cookie, struct dma_tx_state *txstate)
-+{
-+ return dma_cookie_status(chan, cookie, txstate);
-+}
-+
-+static void dpaa2_qdma_free_desc(struct virt_dma_desc *vdesc)
-+{
-+}
-+
-+static void dpaa2_qdma_issue_pending(struct dma_chan *chan)
-+{
-+ struct dpaa2_qdma_comp *dpaa2_comp;
-+ struct dpaa2_qdma_chan *dpaa2_chan = to_dpaa2_qdma_chan(chan);
-+ struct dpaa2_qdma_engine *dpaa2_qdma = dpaa2_chan->qdma;
-+ struct dpaa2_qdma_priv *priv = dpaa2_qdma->priv;
-+ struct virt_dma_desc *vdesc;
-+ struct dpaa2_fd *fd;
-+ int err;
-+ unsigned long flags;
-+
-+ spin_lock_irqsave(&dpaa2_chan->queue_lock, flags);
-+ spin_lock(&dpaa2_chan->vchan.lock);
-+ if (vchan_issue_pending(&dpaa2_chan->vchan)) {
-+ vdesc = vchan_next_desc(&dpaa2_chan->vchan);
-+ if (!vdesc)
-+ goto err_enqueue;
-+ dpaa2_comp = to_fsl_qdma_comp(vdesc);
-+
-+ fd = (struct dpaa2_fd *)dpaa2_comp->fd_virt_addr;
-+
-+ list_del(&vdesc->node);
-+ list_add_tail(&dpaa2_comp->list, &dpaa2_chan->comp_used);
-+
-+ /* TOBO: priority hard-coded to zero */
-+ err = dpaa2_io_service_enqueue_fq(NULL,
-+ priv->tx_queue_attr[0].fqid, fd);
-+ if (err) {
-+ list_del(&dpaa2_comp->list);
-+ list_add_tail(&dpaa2_comp->list,
-+ &dpaa2_chan->comp_free);
-+ }
-+
-+ }
-+err_enqueue:
-+ spin_unlock(&dpaa2_chan->vchan.lock);
-+ spin_unlock_irqrestore(&dpaa2_chan->queue_lock, flags);
-+}
-+
-+static int __cold dpaa2_qdma_setup(struct fsl_mc_device *ls_dev)
-+{
-+ struct device *dev = &ls_dev->dev;
-+ struct dpaa2_qdma_priv *priv;
-+ struct dpaa2_qdma_priv_per_prio *ppriv;
-+ uint8_t prio_def = DPDMAI_PRIO_NUM;
-+ int err;
-+ int i;
-+
-+ priv = dev_get_drvdata(dev);
-+
-+ priv->dev = dev;
-+ priv->dpqdma_id = ls_dev->obj_desc.id;
-+
-+ /*Get the handle for the DPDMAI this interface is associate with */
-+ err = dpdmai_open(priv->mc_io, 0, priv->dpqdma_id, &ls_dev->mc_handle);
-+ if (err) {
-+ dev_err(dev, "dpdmai_open() failed\n");
-+ return err;
-+ }
-+ dev_info(dev, "Opened dpdmai object successfully\n");
-+
-+ err = dpdmai_get_attributes(priv->mc_io, 0, ls_dev->mc_handle,
-+ &priv->dpdmai_attr);
-+ if (err) {
-+ dev_err(dev, "dpdmai_get_attributes() failed\n");
-+ return err;
-+ }
-+
-+ if (priv->dpdmai_attr.version.major > DPDMAI_VER_MAJOR) {
-+ dev_err(dev, "DPDMAI major version mismatch\n"
-+ "Found %u.%u, supported version is %u.%u\n",
-+ priv->dpdmai_attr.version.major,
-+ priv->dpdmai_attr.version.minor,
-+ DPDMAI_VER_MAJOR, DPDMAI_VER_MINOR);
-+ }
-+
-+ if (priv->dpdmai_attr.version.minor > DPDMAI_VER_MINOR) {
-+ dev_err(dev, "DPDMAI minor version mismatch\n"
-+ "Found %u.%u, supported version is %u.%u\n",
-+ priv->dpdmai_attr.version.major,
-+ priv->dpdmai_attr.version.minor,
-+ DPDMAI_VER_MAJOR, DPDMAI_VER_MINOR);
-+ }
-+
-+ priv->num_pairs = min(priv->dpdmai_attr.num_of_priorities, prio_def);
-+ ppriv = kcalloc(priv->num_pairs, sizeof(*ppriv), GFP_KERNEL);
-+ if (!ppriv) {
-+ dev_err(dev, "kzalloc for ppriv failed\n");
-+ return -1;
-+ }
-+ priv->ppriv = ppriv;
-+
-+ for (i = 0; i < priv->num_pairs; i++) {
-+ err = dpdmai_get_rx_queue(priv->mc_io, 0, ls_dev->mc_handle,
-+ i, &priv->rx_queue_attr[i]);
-+ if (err) {
-+ dev_err(dev, "dpdmai_get_rx_queue() failed\n");
-+ return err;
-+ }
-+ ppriv->rsp_fqid = priv->rx_queue_attr[i].fqid;
-+
-+ err = dpdmai_get_tx_queue(priv->mc_io, 0, ls_dev->mc_handle,
-+ i, &priv->tx_queue_attr[i]);
-+ if (err) {
-+ dev_err(dev, "dpdmai_get_tx_queue() failed\n");
-+ return err;
-+ }
-+ ppriv->req_fqid = priv->tx_queue_attr[i].fqid;
-+ ppriv->prio = i;
-+ ppriv->priv = priv;
-+ ppriv++;
-+ }
-+
-+ return 0;
-+}
-+
-+static void dpaa2_qdma_fqdan_cb(struct dpaa2_io_notification_ctx *ctx)
-+{
-+ struct dpaa2_qdma_priv_per_prio *ppriv = container_of(ctx,
-+ struct dpaa2_qdma_priv_per_prio, nctx);
-+ struct dpaa2_qdma_priv *priv = ppriv->priv;
-+ struct dpaa2_qdma_comp *dpaa2_comp, *_comp_tmp;
-+ struct dpaa2_qdma_chan *qchan;
-+ const struct dpaa2_fd *fd;
-+ const struct dpaa2_fd *fd_eq;
-+ struct dpaa2_dq *dq;
-+ int err;
-+ int is_last = 0;
-+ uint8_t status;
-+ int i;
-+ int found;
-+ uint32_t n_chans = priv->dpaa2_qdma->n_chans;
-+
-+ do {
-+ err = dpaa2_io_service_pull_fq(NULL, ppriv->rsp_fqid,
-+ ppriv->store);
-+ } while (err);
-+
-+ while (!is_last) {
-+ do {
-+ dq = dpaa2_io_store_next(ppriv->store, &is_last);
-+ } while (!is_last && !dq);
-+ if (!dq) {
-+ dev_err(priv->dev, "FQID returned no valid frames!\n");
-+ continue;
-+ }
-+
-+ /* obtain FD and process the error */
-+ fd = dpaa2_dq_fd(dq);
-+
-+ status = dpaa2_fd_get_ctrl(fd) & 0xff;
-+ if (status)
-+ dev_err(priv->dev, "FD error occurred\n");
-+ found = 0;
-+ for (i = 0; i < n_chans; i++) {
-+ qchan = &priv->dpaa2_qdma->chans[i];
-+ spin_lock(&qchan->queue_lock);
-+ if (list_empty(&qchan->comp_used)) {
-+ spin_unlock(&qchan->queue_lock);
-+ continue;
-+ }
-+ list_for_each_entry_safe(dpaa2_comp, _comp_tmp,
-+ &qchan->comp_used, list) {
-+ fd_eq = (struct dpaa2_fd *)
-+ dpaa2_comp->fd_virt_addr;
-+
-+ if (le64_to_cpu(fd_eq->simple.addr) ==
-+ le64_to_cpu(fd->simple.addr)) {
-+
-+ list_del(&dpaa2_comp->list);
-+ list_add_tail(&dpaa2_comp->list,
-+ &qchan->comp_free);
-+
-+ spin_lock(&qchan->vchan.lock);
-+ vchan_cookie_complete(
-+ &dpaa2_comp->vdesc);
-+ spin_unlock(&qchan->vchan.lock);
-+ found = 1;
-+ break;
-+ }
-+ }
-+ spin_unlock(&qchan->queue_lock);
-+ if (found)
-+ break;
-+ }
-+ }
-+
-+ dpaa2_io_service_rearm(NULL, ctx);
-+}
-+
-+static int __cold dpaa2_qdma_dpio_setup(struct dpaa2_qdma_priv *priv)
-+{
-+ int err, i, num;
-+ struct device *dev = priv->dev;
-+ struct dpaa2_qdma_priv_per_prio *ppriv;
-+
-+ num = priv->num_pairs;
-+ ppriv = priv->ppriv;
-+ for (i = 0; i < num; i++) {
-+ ppriv->nctx.is_cdan = 0;
-+ ppriv->nctx.desired_cpu = 1;
-+ ppriv->nctx.id = ppriv->rsp_fqid;
-+ ppriv->nctx.cb = dpaa2_qdma_fqdan_cb;
-+ err = dpaa2_io_service_register(NULL, &ppriv->nctx);
-+ if (err) {
-+ dev_err(dev, "Notification register failed\n");
-+ goto err_service;
-+ }
-+
-+ ppriv->store = dpaa2_io_store_create(DPAA2_QDMA_STORE_SIZE,
-+ dev);
-+ if (!ppriv->store) {
-+ dev_err(dev, "dpaa2_io_store_create() failed\n");
-+ goto err_store;
-+ }
-+
-+ ppriv++;
-+ }
-+ return 0;
-+
-+err_store:
-+ dpaa2_io_service_deregister(NULL, &ppriv->nctx);
-+err_service:
-+ ppriv--;
-+ while (ppriv >= priv->ppriv) {
-+ dpaa2_io_service_deregister(NULL, &ppriv->nctx);
-+ dpaa2_io_store_destroy(ppriv->store);
-+ ppriv--;
-+ }
-+ return -1;
-+}
-+
-+static void __cold dpaa2_dpmai_store_free(struct dpaa2_qdma_priv *priv)
-+{
-+ struct dpaa2_qdma_priv_per_prio *ppriv = priv->ppriv;
-+ int i;
-+
-+ for (i = 0; i < priv->num_pairs; i++) {
-+ dpaa2_io_store_destroy(ppriv->store);
-+ ppriv++;
-+ }
-+}
-+
-+static void __cold dpaa2_dpdmai_dpio_free(struct dpaa2_qdma_priv *priv)
-+{
-+ struct dpaa2_qdma_priv_per_prio *ppriv = priv->ppriv;
-+ int i;
-+
-+ for (i = 0; i < priv->num_pairs; i++) {
-+ dpaa2_io_service_deregister(NULL, &ppriv->nctx);
-+ ppriv++;
-+ }
-+}
-+
-+static int __cold dpaa2_dpdmai_bind(struct dpaa2_qdma_priv *priv)
-+{
-+ int err;
-+ struct dpdmai_rx_queue_cfg rx_queue_cfg;
-+ struct device *dev = priv->dev;
-+ struct dpaa2_qdma_priv_per_prio *ppriv;
-+ struct fsl_mc_device *ls_dev = to_fsl_mc_device(dev);
-+ int i, num;
-+
-+ num = priv->num_pairs;
-+ ppriv = priv->ppriv;
-+ for (i = 0; i < num; i++) {
-+ rx_queue_cfg.options = DPDMAI_QUEUE_OPT_USER_CTX |
-+ DPDMAI_QUEUE_OPT_DEST;
-+ rx_queue_cfg.user_ctx = ppriv->nctx.qman64;
-+ rx_queue_cfg.dest_cfg.dest_type = DPDMAI_DEST_DPIO;
-+ rx_queue_cfg.dest_cfg.dest_id = ppriv->nctx.dpio_id;
-+ rx_queue_cfg.dest_cfg.priority = ppriv->prio;
-+ err = dpdmai_set_rx_queue(priv->mc_io, 0, ls_dev->mc_handle,
-+ rx_queue_cfg.dest_cfg.priority, &rx_queue_cfg);
-+ if (err) {
-+ dev_err(dev, "dpdmai_set_rx_queue() failed\n");
-+ return err;
-+ }
-+
-+ ppriv++;
-+ }
-+
-+ return 0;
-+}
-+
-+static int __cold dpaa2_dpdmai_dpio_unbind(struct dpaa2_qdma_priv *priv)
-+{
-+ int err = 0;
-+ struct device *dev = priv->dev;
-+ struct fsl_mc_device *ls_dev = to_fsl_mc_device(dev);
-+ struct dpaa2_qdma_priv_per_prio *ppriv = priv->ppriv;
-+ int i;
-+
-+ for (i = 0; i < priv->num_pairs; i++) {
-+ ppriv->nctx.qman64 = 0;
-+ ppriv->nctx.dpio_id = 0;
-+ ppriv++;
-+ }
-+
-+ err = dpdmai_reset(priv->mc_io, 0, ls_dev->mc_handle);
-+ if (err)
-+ dev_err(dev, "dpdmai_reset() failed\n");
-+
-+ return err;
-+}
-+
-+static void __cold dpaa2_dpdmai_free_pool(struct dpaa2_qdma_chan *qchan,
-+ struct list_head *head)
-+{
-+ struct qdma_sg_blk *sgb_tmp, *_sgb_tmp;
-+ /* free the QDMA SG pool block */
-+ list_for_each_entry_safe(sgb_tmp, _sgb_tmp, head, list) {
-+ sgb_tmp->blk_virt_addr = (void *)((struct qdma_sg_blk *)
-+ sgb_tmp->blk_virt_addr - 1);
-+ sgb_tmp->blk_bus_addr = sgb_tmp->blk_bus_addr
-+ - sizeof(*sgb_tmp);
-+ dma_pool_free(qchan->sg_blk_pool, sgb_tmp->blk_virt_addr,
-+ sgb_tmp->blk_bus_addr);
-+ }
-+
-+}
-+
-+static void __cold dpaa2_dpdmai_free_comp(struct dpaa2_qdma_chan *qchan,
-+ struct list_head *head)
-+{
-+ struct dpaa2_qdma_comp *comp_tmp, *_comp_tmp;
-+ /* free the QDMA comp resource */
-+ list_for_each_entry_safe(comp_tmp, _comp_tmp,
-+ head, list) {
-+ dma_pool_free(qchan->fd_pool,
-+ comp_tmp->fd_virt_addr,
-+ comp_tmp->fd_bus_addr);
-+ /* free the SG source block on comp */
-+ dpaa2_dpdmai_free_pool(qchan, &comp_tmp->sg_src_head);
-+ /* free the SG destination block on comp */
-+ dpaa2_dpdmai_free_pool(qchan, &comp_tmp->sg_dst_head);
-+ list_del(&comp_tmp->list);
-+ kfree(comp_tmp);
-+ }
-+
-+}
-+
-+static void __cold dpaa2_dpdmai_free_channels(
-+ struct dpaa2_qdma_engine *dpaa2_qdma)
-+{
-+ struct dpaa2_qdma_chan *qchan;
-+ int num, i;
-+
-+ num = dpaa2_qdma->n_chans;
-+ for (i = 0; i < num; i++) {
-+ qchan = &dpaa2_qdma->chans[i];
-+ dpaa2_dpdmai_free_comp(qchan, &qchan->comp_used);
-+ dpaa2_dpdmai_free_comp(qchan, &qchan->comp_free);
-+ dpaa2_dpdmai_free_pool(qchan, &qchan->sgb_free);
-+ dma_pool_destroy(qchan->fd_pool);
-+ dma_pool_destroy(qchan->sg_blk_pool);
-+ }
-+}
-+
-+static int dpaa2_dpdmai_alloc_channels(struct dpaa2_qdma_engine *dpaa2_qdma)
-+{
-+ struct dpaa2_qdma_chan *dpaa2_chan;
-+ struct device *dev = &dpaa2_qdma->priv->dpdmai_dev->dev;
-+ int i;
-+
-+ INIT_LIST_HEAD(&dpaa2_qdma->dma_dev.channels);
-+ for (i = 0; i < dpaa2_qdma->n_chans; i++) {
-+ dpaa2_chan = &dpaa2_qdma->chans[i];
-+ dpaa2_chan->qdma = dpaa2_qdma;
-+ dpaa2_chan->vchan.desc_free = dpaa2_qdma_free_desc;
-+ vchan_init(&dpaa2_chan->vchan, &dpaa2_qdma->dma_dev);
-+
-+ dpaa2_chan->fd_pool = dma_pool_create("fd_pool",
-+ dev, FD_POOL_SIZE, 32, 0);
-+ if (!dpaa2_chan->fd_pool)
-+ return -1;
-+ dpaa2_chan->sg_blk_pool = dma_pool_create("sg_blk_pool",
-+ dev, SG_POOL_SIZE, 32, 0);
-+ if (!dpaa2_chan->sg_blk_pool)
-+ return -1;
-+
-+ spin_lock_init(&dpaa2_chan->queue_lock);
-+ INIT_LIST_HEAD(&dpaa2_chan->comp_used);
-+ INIT_LIST_HEAD(&dpaa2_chan->comp_free);
-+ INIT_LIST_HEAD(&dpaa2_chan->sgb_free);
-+ }
-+ return 0;
-+}
-+
-+static int dpaa2_qdma_probe(struct fsl_mc_device *dpdmai_dev)
-+{
-+ struct dpaa2_qdma_priv *priv;
-+ struct device *dev = &dpdmai_dev->dev;
-+ struct dpaa2_qdma_engine *dpaa2_qdma;
-+ int err;
-+
-+ priv = kzalloc(sizeof(*priv), GFP_KERNEL);
-+ if (!priv)
-+ return -ENOMEM;
-+ dev_set_drvdata(dev, priv);
-+ priv->dpdmai_dev = dpdmai_dev;
-+
-+ priv->iommu_domain = iommu_get_domain_for_dev(dev);
-+ if (priv->iommu_domain)
-+ smmu_disable = false;
-+
-+ /* obtain a MC portal */
-+ err = fsl_mc_portal_allocate(dpdmai_dev, 0, &priv->mc_io);
-+ if (err) {
-+ dev_err(dev, "MC portal allocation failed\n");
-+ goto err_mcportal;
-+ }
-+
-+ /* DPDMAI initialization */
-+ err = dpaa2_qdma_setup(dpdmai_dev);
-+ if (err) {
-+ dev_err(dev, "dpaa2_dpdmai_setup() failed\n");
-+ goto err_dpdmai_setup;
-+ }
-+
-+ /* DPIO */
-+ err = dpaa2_qdma_dpio_setup(priv);
-+ if (err) {
-+ dev_err(dev, "dpaa2_dpdmai_dpio_setup() failed\n");
-+ goto err_dpio_setup;
-+ }
-+
-+ /* DPDMAI binding to DPIO */
-+ err = dpaa2_dpdmai_bind(priv);
-+ if (err) {
-+ dev_err(dev, "dpaa2_dpdmai_bind() failed\n");
-+ goto err_bind;
-+ }
-+
-+ /* DPDMAI enable */
-+ err = dpdmai_enable(priv->mc_io, 0, dpdmai_dev->mc_handle);
-+ if (err) {
-+ dev_err(dev, "dpdmai_enable() faile\n");
-+ goto err_enable;
-+ }
-+
-+ dpaa2_qdma = kzalloc(sizeof(*dpaa2_qdma), GFP_KERNEL);
-+ if (!dpaa2_qdma) {
-+ err = -ENOMEM;
-+ goto err_eng;
-+ }
-+
-+ priv->dpaa2_qdma = dpaa2_qdma;
-+ dpaa2_qdma->priv = priv;
-+
-+ dpaa2_qdma->n_chans = NUM_CH;
-+
-+ err = dpaa2_dpdmai_alloc_channels(dpaa2_qdma);
-+ if (err) {
-+ dev_err(dev, "QDMA alloc channels faile\n");
-+ goto err_reg;
-+ }
-+
-+ dma_cap_set(DMA_PRIVATE, dpaa2_qdma->dma_dev.cap_mask);
-+ dma_cap_set(DMA_SLAVE, dpaa2_qdma->dma_dev.cap_mask);
-+ dma_cap_set(DMA_MEMCPY, dpaa2_qdma->dma_dev.cap_mask);
-+
-+ dpaa2_qdma->dma_dev.dev = dev;
-+ dpaa2_qdma->dma_dev.device_alloc_chan_resources
-+ = dpaa2_qdma_alloc_chan_resources;
-+ dpaa2_qdma->dma_dev.device_free_chan_resources
-+ = dpaa2_qdma_free_chan_resources;
-+ dpaa2_qdma->dma_dev.device_tx_status = dpaa2_qdma_tx_status;
-+ dpaa2_qdma->dma_dev.device_prep_dma_memcpy = dpaa2_qdma_prep_memcpy;
-+ dpaa2_qdma->dma_dev.device_issue_pending = dpaa2_qdma_issue_pending;
-+
-+ err = dma_async_device_register(&dpaa2_qdma->dma_dev);
-+ if (err) {
-+ dev_err(dev, "Can't register NXP QDMA engine.\n");
-+ goto err_reg;
-+ }
-+
-+ return 0;
-+
-+err_reg:
-+ dpaa2_dpdmai_free_channels(dpaa2_qdma);
-+ kfree(dpaa2_qdma);
-+err_eng:
-+ dpdmai_disable(priv->mc_io, 0, dpdmai_dev->mc_handle);
-+err_enable:
-+ dpaa2_dpdmai_dpio_unbind(priv);
-+err_bind:
-+ dpaa2_dpmai_store_free(priv);
-+ dpaa2_dpdmai_dpio_free(priv);
-+err_dpio_setup:
-+ dpdmai_close(priv->mc_io, 0, dpdmai_dev->mc_handle);
-+err_dpdmai_setup:
-+ fsl_mc_portal_free(priv->mc_io);
-+err_mcportal:
-+ kfree(priv->ppriv);
-+ kfree(priv);
-+ dev_set_drvdata(dev, NULL);
-+ return err;
-+}
-+
-+static int dpaa2_qdma_remove(struct fsl_mc_device *ls_dev)
-+{
-+ struct device *dev;
-+ struct dpaa2_qdma_priv *priv;
-+ struct dpaa2_qdma_engine *dpaa2_qdma;
-+
-+ dev = &ls_dev->dev;
-+ priv = dev_get_drvdata(dev);
-+ dpaa2_qdma = priv->dpaa2_qdma;
-+
-+ dpdmai_disable(priv->mc_io, 0, ls_dev->mc_handle);
-+ dpaa2_dpdmai_dpio_unbind(priv);
-+ dpaa2_dpmai_store_free(priv);
-+ dpaa2_dpdmai_dpio_free(priv);
-+ dpdmai_close(priv->mc_io, 0, ls_dev->mc_handle);
-+ fsl_mc_portal_free(priv->mc_io);
-+ dev_set_drvdata(dev, NULL);
-+ dpaa2_dpdmai_free_channels(dpaa2_qdma);
-+
-+ dma_async_device_unregister(&dpaa2_qdma->dma_dev);
-+ kfree(priv);
-+ kfree(dpaa2_qdma);
-+
-+ return 0;
-+}
-+
-+static const struct fsl_mc_device_id dpaa2_qdma_id_table[] = {
-+ {
-+ .vendor = FSL_MC_VENDOR_FREESCALE,
-+ .obj_type = "dpdmai",
-+ },
-+ { .vendor = 0x0 }
-+};
-+
-+static struct fsl_mc_driver dpaa2_qdma_driver = {
-+ .driver = {
-+ .name = "dpaa2-qdma",
-+ .owner = THIS_MODULE,
-+ },
-+ .probe = dpaa2_qdma_probe,
-+ .remove = dpaa2_qdma_remove,
-+ .match_id_table = dpaa2_qdma_id_table
-+};
-+
-+static int __init dpaa2_qdma_driver_init(void)
-+{
-+ return fsl_mc_driver_register(&(dpaa2_qdma_driver));
-+}
-+late_initcall(dpaa2_qdma_driver_init);
-+
-+static void __exit fsl_qdma_exit(void)
-+{
-+ fsl_mc_driver_unregister(&(dpaa2_qdma_driver));
-+}
-+module_exit(fsl_qdma_exit);
-+
-+MODULE_DESCRIPTION("NXP DPAA2 qDMA driver");
-+MODULE_LICENSE("Dual BSD/GPL");
---- /dev/null
-+++ b/drivers/dma/dpaa2-qdma/dpaa2-qdma.h
-@@ -0,0 +1,227 @@
-+/* Copyright 2015 NXP Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of NXP Semiconductor nor the
-+ * names of its contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY NXP Semiconductor ``AS IS'' AND ANY
-+ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-+ * DISCLAIMED. IN NO EVENT SHALL NXP Semiconductor BE LIABLE FOR ANY
-+ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-+ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
-+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-+ */
-+
-+#ifndef __DPAA2_QDMA_H
-+#define __DPAA2_QDMA_H
-+
-+#define LONG_FORMAT 1
-+
-+#define DPAA2_QDMA_STORE_SIZE 16
-+#define NUM_CH 8
-+#define NUM_SG_PER_BLK 16
-+
-+#define QDMA_DMR_OFFSET 0x0
-+#define QDMA_DQ_EN (0 << 30)
-+#define QDMA_DQ_DIS (1 << 30)
-+
-+#define QDMA_DSR_M_OFFSET 0x10004
-+
-+struct dpaa2_qdma_sd_d {
-+ uint32_t rsv:32;
-+ union {
-+ struct {
-+ uint32_t ssd:12; /* souce stride distance */
-+ uint32_t sss:12; /* souce stride size */
-+ uint32_t rsv1:8;
-+ } sdf;
-+ struct {
-+ uint32_t dsd:12; /* Destination stride distance */
-+ uint32_t dss:12; /* Destination stride size */
-+ uint32_t rsv2:8;
-+ } ddf;
-+ } df;
-+ uint32_t rbpcmd; /* Route-by-port command */
-+ uint32_t cmd;
-+} __attribute__((__packed__));
-+/* Source descriptor command read transaction type for RBP=0:
-+ coherent copy of cacheable memory */
-+#define QDMA_SD_CMD_RDTTYPE_COHERENT (0xb << 28)
-+/* Destination descriptor command write transaction type for RBP=0:
-+ coherent copy of cacheable memory */
-+#define QDMA_DD_CMD_WRTTYPE_COHERENT (0x6 << 28)
-+
-+#define QDMA_SG_FMT_SDB 0x0 /* single data buffer */
-+#define QDMA_SG_FMT_FDS 0x1 /* frame data section */
-+#define QDMA_SG_FMT_SGTE 0x2 /* SGT extension */
-+#define QDMA_SG_SL_SHORT 0x1 /* short length */
-+#define QDMA_SG_SL_LONG 0x0 /* short length */
-+#define QDMA_SG_F 0x1 /* last sg entry */
-+struct dpaa2_qdma_sg {
-+ uint32_t addr_lo; /* address 0:31 */
-+ uint32_t addr_hi:17; /* address 32:48 */
-+ uint32_t rsv:15;
-+ union {
-+ uint32_t data_len_sl0; /* SL=0, the long format */
-+ struct {
-+ uint32_t len:17; /* SL=1, the short format */
-+ uint32_t reserve:3;
-+ uint32_t sf:1;
-+ uint32_t sr:1;
-+ uint32_t size:10; /* buff size */
-+ } data_len_sl1;
-+ } data_len; /* AVAIL_LENGTH */
-+ struct {
-+ uint32_t bpid:14;
-+ uint32_t ivp:1;
-+ uint32_t mbt:1;
-+ uint32_t offset:12;
-+ uint32_t fmt:2;
-+ uint32_t sl:1;
-+ uint32_t f:1;
-+ } ctrl;
-+} __attribute__((__packed__));
-+
-+#define QMAN_FD_FMT_ENABLE (1) /* frame list table enable */
-+#define QMAN_FD_BMT_ENABLE (1 << 15) /* bypass memory translation */
-+#define QMAN_FD_BMT_DISABLE (0 << 15) /* bypass memory translation */
-+#define QMAN_FD_SL_DISABLE (0 << 14) /* short lengthe disabled */
-+#define QMAN_FD_SL_ENABLE (1 << 14) /* short lengthe enabled */
-+
-+#define QDMA_SB_FRAME (0 << 28) /* single frame */
-+#define QDMA_SG_FRAME (2 << 28) /* scatter gather frames */
-+#define QDMA_FINAL_BIT_DISABLE (0 << 31) /* final bit disable */
-+#define QDMA_FINAL_BIT_ENABLE (1 << 31) /* final bit enable */
-+
-+#define QDMA_FD_SHORT_FORMAT (1 << 11) /* short format */
-+#define QDMA_FD_LONG_FORMAT (0 << 11) /* long format */
-+#define QDMA_SER_DISABLE (0 << 8) /* no notification */
-+#define QDMA_SER_CTX (1 << 8) /* notification by FQD_CTX[fqid] */
-+#define QDMA_SER_DEST (2 << 8) /* notification by destination desc */
-+#define QDMA_SER_BOTH (3 << 8) /* soruce and dest notification */
-+#define QDMA_FD_SPF_ENALBE (1 << 30) /* source prefetch enable */
-+
-+#define QMAN_FD_VA_ENABLE (1 << 14) /* Address used is virtual address */
-+#define QMAN_FD_VA_DISABLE (0 << 14)/* Address used is a real address */
-+#define QMAN_FD_CBMT_ENABLE (1 << 15) /* Flow Context: 49bit physical address */
-+#define QMAN_FD_CBMT_DISABLE (0 << 15) /* Flow Context: 64bit virtual address */
-+#define QMAN_FD_SC_DISABLE (0 << 27) /* stashing control */
-+
-+#define QDMA_FL_FMT_SBF (0x0) /* Single buffer frame */
-+#define QDMA_FL_FMT_SGE 0x2 /* Scatter gather frame */
-+#define QDMA_FL_BMT_ENABLE (0x1 << 15)/* enable bypass memory translation */
-+#define QDMA_FL_BMT_DISABLE 0x0 /* enable bypass memory translation */
-+#define QDMA_FL_SL_LONG (0x0 << 2)/* long length */
-+#define QDMA_FL_SL_SHORT 0x1 /* short length */
-+#define QDMA_FL_F (0x1)/* last frame list bit */
-+/*Description of Frame list table structure*/
-+
-+struct dpaa2_qdma_chan {
-+ struct virt_dma_chan vchan;
-+ struct virt_dma_desc vdesc;
-+ enum dma_status status;
-+ struct dpaa2_qdma_engine *qdma;
-+
-+ struct mutex dpaa2_queue_mutex;
-+ spinlock_t queue_lock;
-+ struct dma_pool *fd_pool;
-+ struct dma_pool *sg_blk_pool;
-+
-+ struct list_head comp_used;
-+ struct list_head comp_free;
-+
-+ struct list_head sgb_free;
-+};
-+
-+struct qdma_sg_blk {
-+ dma_addr_t blk_bus_addr;
-+ void *blk_virt_addr;
-+ struct list_head list;
-+};
-+
-+struct dpaa2_qdma_comp {
-+ dma_addr_t fd_bus_addr;
-+ dma_addr_t fl_bus_addr;
-+ dma_addr_t desc_bus_addr;
-+ dma_addr_t sge_src_bus_addr;
-+ dma_addr_t sge_dst_bus_addr;
-+ void *fd_virt_addr;
-+ void *fl_virt_addr;
-+ void *desc_virt_addr;
-+ void *sg_src_virt_addr;
-+ void *sg_dst_virt_addr;
-+ struct qdma_sg_blk *sg_blk;
-+ uint32_t sg_blk_num;
-+ struct list_head sg_src_head;
-+ struct list_head sg_dst_head;
-+ struct dpaa2_qdma_chan *qchan;
-+ struct virt_dma_desc vdesc;
-+ struct list_head list;
-+};
-+
-+struct dpaa2_qdma_engine {
-+ struct dma_device dma_dev;
-+ u32 n_chans;
-+ struct dpaa2_qdma_chan chans[NUM_CH];
-+
-+ struct dpaa2_qdma_priv *priv;
-+};
-+
-+/*
-+ * dpaa2_qdma_priv - driver private data
-+ */
-+struct dpaa2_qdma_priv {
-+ int dpqdma_id;
-+
-+ struct iommu_domain *iommu_domain;
-+ struct dpdmai_attr dpdmai_attr;
-+ struct device *dev;
-+ struct fsl_mc_io *mc_io;
-+ struct fsl_mc_device *dpdmai_dev;
-+
-+ struct dpdmai_rx_queue_attr rx_queue_attr[DPDMAI_PRIO_NUM];
-+ struct dpdmai_tx_queue_attr tx_queue_attr[DPDMAI_PRIO_NUM];
-+
-+ uint8_t num_pairs;
-+
-+ struct dpaa2_qdma_engine *dpaa2_qdma;
-+ struct dpaa2_qdma_priv_per_prio *ppriv;
-+};
-+
-+struct dpaa2_qdma_priv_per_prio {
-+ int req_fqid;
-+ int rsp_fqid;
-+ int prio;
-+
-+ struct dpaa2_io_store *store;
-+ struct dpaa2_io_notification_ctx nctx;
-+
-+ struct dpaa2_qdma_priv *priv;
-+};
-+
-+/* FD pool size: one FD + 3 Frame list + 2 source/destination descriptor */
-+#define FD_POOL_SIZE (sizeof(struct dpaa2_fd) + \
-+ sizeof(struct dpaa2_fl_entry) * 3 + \
-+ sizeof(struct dpaa2_qdma_sd_d) * 2)
-+
-+/* qdma_sg_blk + 16 SGs */
-+#define SG_POOL_SIZE (sizeof(struct qdma_sg_blk) +\
-+ sizeof(struct dpaa2_qdma_sg) * NUM_SG_PER_BLK)
-+#endif /* __DPAA2_QDMA_H */
---- /dev/null
-+++ b/drivers/dma/dpaa2-qdma/dpdmai.c
-@@ -0,0 +1,515 @@
-+/* Copyright 2013-2015 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of the above-listed copyright holders nor the
-+ * names of any contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
-+ * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
-+ * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
-+ * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE
-+ * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
-+ * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
-+ * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
-+ * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
-+ * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
-+ * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
-+ * POSSIBILITY OF SUCH DAMAGE.
-+ */
-+#include <linux/types.h>
-+#include <linux/io.h>
-+#include "fsl_dpdmai.h"
-+#include "fsl_dpdmai_cmd.h"
-+#include <linux/fsl/mc.h>
-+
-+struct dpdmai_cmd_open {
-+ __le32 dpdmai_id;
-+};
-+
-+struct dpdmai_rsp_get_attributes {
-+ __le32 id;
-+ u8 num_of_priorities;
-+ u8 pad0[3];
-+ __le16 major;
-+ __le16 minor;
-+};
-+
-+
-+struct dpdmai_cmd_queue {
-+ __le32 dest_id;
-+ u8 priority;
-+ u8 queue;
-+ u8 dest_type;
-+ u8 pad;
-+ __le64 user_ctx;
-+ union {
-+ __le32 options;
-+ __le32 fqid;
-+ };
-+};
-+
-+struct dpdmai_rsp_get_tx_queue {
-+ __le64 pad;
-+ __le32 fqid;
-+};
-+
-+
-+int dpdmai_open(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ int dpdmai_id,
-+ uint16_t *token)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpdmai_cmd_open *cmd_params;
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPDMAI_CMDID_OPEN,
-+ cmd_flags,
-+ 0);
-+
-+ cmd_params = (struct dpdmai_cmd_open *)cmd.params;
-+ cmd_params->dpdmai_id = cpu_to_le32(dpdmai_id);
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ *token = mc_cmd_hdr_read_token(&cmd);
-+ return 0;
-+}
-+
-+int dpdmai_close(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPDMAI_CMDID_CLOSE,
-+ cmd_flags, token);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+int dpdmai_create(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ const struct dpdmai_cfg *cfg,
-+ uint16_t *token)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPDMAI_CMDID_CREATE,
-+ cmd_flags,
-+ 0);
-+ DPDMAI_CMD_CREATE(cmd, cfg);
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ *token = MC_CMD_HDR_READ_TOKEN(cmd.header);
-+
-+ return 0;
-+}
-+
-+int dpdmai_destroy(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPDMAI_CMDID_DESTROY,
-+ cmd_flags,
-+ token);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+int dpdmai_enable(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPDMAI_CMDID_ENABLE,
-+ cmd_flags,
-+ token);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+int dpdmai_disable(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPDMAI_CMDID_DISABLE,
-+ cmd_flags,
-+ token);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+int dpdmai_is_enabled(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token,
-+ int *en)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ int err;
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPDMAI_CMDID_IS_ENABLED,
-+ cmd_flags,
-+ token);
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ DPDMAI_RSP_IS_ENABLED(cmd, *en);
-+
-+ return 0;
-+}
-+
-+int dpdmai_reset(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPDMAI_CMDID_RESET,
-+ cmd_flags,
-+ token);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+int dpdmai_get_irq(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token,
-+ uint8_t irq_index,
-+ int *type,
-+ struct dpdmai_irq_cfg *irq_cfg)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPDMAI_CMDID_GET_IRQ,
-+ cmd_flags,
-+ token);
-+ DPDMAI_CMD_GET_IRQ(cmd, irq_index);
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ DPDMAI_RSP_GET_IRQ(cmd, *type, irq_cfg);
-+
-+ return 0;
-+}
-+
-+int dpdmai_set_irq(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token,
-+ uint8_t irq_index,
-+ struct dpdmai_irq_cfg *irq_cfg)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPDMAI_CMDID_SET_IRQ,
-+ cmd_flags,
-+ token);
-+ DPDMAI_CMD_SET_IRQ(cmd, irq_index, irq_cfg);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+int dpdmai_get_irq_enable(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token,
-+ uint8_t irq_index,
-+ uint8_t *en)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPDMAI_CMDID_GET_IRQ_ENABLE,
-+ cmd_flags,
-+ token);
-+ DPDMAI_CMD_GET_IRQ_ENABLE(cmd, irq_index);
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ DPDMAI_RSP_GET_IRQ_ENABLE(cmd, *en);
-+
-+ return 0;
-+}
-+
-+int dpdmai_set_irq_enable(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token,
-+ uint8_t irq_index,
-+ uint8_t en)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPDMAI_CMDID_SET_IRQ_ENABLE,
-+ cmd_flags,
-+ token);
-+ DPDMAI_CMD_SET_IRQ_ENABLE(cmd, irq_index, en);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+int dpdmai_get_irq_mask(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token,
-+ uint8_t irq_index,
-+ uint32_t *mask)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPDMAI_CMDID_GET_IRQ_MASK,
-+ cmd_flags,
-+ token);
-+ DPDMAI_CMD_GET_IRQ_MASK(cmd, irq_index);
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ DPDMAI_RSP_GET_IRQ_MASK(cmd, *mask);
-+
-+ return 0;
-+}
-+
-+int dpdmai_set_irq_mask(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token,
-+ uint8_t irq_index,
-+ uint32_t mask)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPDMAI_CMDID_SET_IRQ_MASK,
-+ cmd_flags,
-+ token);
-+ DPDMAI_CMD_SET_IRQ_MASK(cmd, irq_index, mask);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+int dpdmai_get_irq_status(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token,
-+ uint8_t irq_index,
-+ uint32_t *status)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPDMAI_CMDID_GET_IRQ_STATUS,
-+ cmd_flags,
-+ token);
-+ DPDMAI_CMD_GET_IRQ_STATUS(cmd, irq_index, *status);
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ DPDMAI_RSP_GET_IRQ_STATUS(cmd, *status);
-+
-+ return 0;
-+}
-+
-+int dpdmai_clear_irq_status(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token,
-+ uint8_t irq_index,
-+ uint32_t status)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPDMAI_CMDID_CLEAR_IRQ_STATUS,
-+ cmd_flags,
-+ token);
-+ DPDMAI_CMD_CLEAR_IRQ_STATUS(cmd, irq_index, status);
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+int dpdmai_get_attributes(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token,
-+ struct dpdmai_attr *attr)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ int err;
-+ struct dpdmai_rsp_get_attributes *rsp_params;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPDMAI_CMDID_GET_ATTR,
-+ cmd_flags,
-+ token);
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ rsp_params = (struct dpdmai_rsp_get_attributes *)cmd.params;
-+ attr->id = le32_to_cpu(rsp_params->id);
-+ attr->version.major = le16_to_cpu(rsp_params->major);
-+ attr->version.minor = le16_to_cpu(rsp_params->minor);
-+ attr->num_of_priorities = rsp_params->num_of_priorities;
-+
-+
-+ return 0;
-+}
-+
-+int dpdmai_set_rx_queue(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token,
-+ uint8_t priority,
-+ const struct dpdmai_rx_queue_cfg *cfg)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpdmai_cmd_queue *cmd_params;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPDMAI_CMDID_SET_RX_QUEUE,
-+ cmd_flags,
-+ token);
-+
-+ cmd_params = (struct dpdmai_cmd_queue *)cmd.params;
-+ cmd_params->dest_id = cpu_to_le32(cfg->dest_cfg.dest_id);
-+ cmd_params->priority = cfg->dest_cfg.priority;
-+ cmd_params->queue = priority;
-+ cmd_params->dest_type = cfg->dest_cfg.dest_type;
-+ cmd_params->user_ctx = cpu_to_le64(cfg->user_ctx);
-+ cmd_params->options = cpu_to_le32(cfg->options);
-+
-+
-+ /* send command to mc*/
-+ return mc_send_command(mc_io, &cmd);
-+}
-+
-+int dpdmai_get_rx_queue(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token,
-+ uint8_t priority, struct dpdmai_rx_queue_attr *attr)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpdmai_cmd_queue *cmd_params;
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPDMAI_CMDID_GET_RX_QUEUE,
-+ cmd_flags,
-+ token);
-+
-+ cmd_params = (struct dpdmai_cmd_queue *)cmd.params;
-+ cmd_params->queue = priority;
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+ attr->dest_cfg.dest_id = le32_to_cpu(cmd_params->dest_id);
-+ attr->dest_cfg.priority = cmd_params->priority;
-+ attr->dest_cfg.dest_type = cmd_params->dest_type;
-+ attr->user_ctx = le64_to_cpu(cmd_params->user_ctx);
-+ attr->fqid = le32_to_cpu(cmd_params->fqid);
-+
-+ return 0;
-+}
-+
-+int dpdmai_get_tx_queue(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token,
-+ uint8_t priority,
-+ struct dpdmai_tx_queue_attr *attr)
-+{
-+ struct fsl_mc_command cmd = { 0 };
-+ struct dpdmai_cmd_queue *cmd_params;
-+ struct dpdmai_rsp_get_tx_queue *rsp_params;
-+ int err;
-+
-+ /* prepare command */
-+ cmd.header = mc_encode_cmd_header(DPDMAI_CMDID_GET_TX_QUEUE,
-+ cmd_flags,
-+ token);
-+
-+ cmd_params = (struct dpdmai_cmd_queue *)cmd.params;
-+ cmd_params->queue = priority;
-+
-+ /* send command to mc*/
-+ err = mc_send_command(mc_io, &cmd);
-+ if (err)
-+ return err;
-+
-+ /* retrieve response parameters */
-+
-+ rsp_params = (struct dpdmai_rsp_get_tx_queue *)cmd.params;
-+ attr->fqid = le32_to_cpu(rsp_params->fqid);
-+
-+ return 0;
-+}
---- /dev/null
-+++ b/drivers/dma/dpaa2-qdma/fsl_dpdmai.h
-@@ -0,0 +1,521 @@
-+/* Copyright 2013-2015 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of the above-listed copyright holders nor the
-+ * names of any contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
-+ * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
-+ * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
-+ * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE
-+ * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
-+ * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
-+ * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
-+ * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
-+ * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
-+ * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
-+ * POSSIBILITY OF SUCH DAMAGE.
-+ */
-+#ifndef __FSL_DPDMAI_H
-+#define __FSL_DPDMAI_H
-+
-+struct fsl_mc_io;
-+
-+/* Data Path DMA Interface API
-+ * Contains initialization APIs and runtime control APIs for DPDMAI
-+ */
-+
-+/* General DPDMAI macros */
-+
-+/**
-+ * Maximum number of Tx/Rx priorities per DPDMAI object
-+ */
-+#define DPDMAI_PRIO_NUM 2
-+
-+/**
-+ * All queues considered; see dpdmai_set_rx_queue()
-+ */
-+#define DPDMAI_ALL_QUEUES (uint8_t)(-1)
-+
-+/**
-+ * dpdmai_open() - Open a control session for the specified object
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @dpdmai_id: DPDMAI unique ID
-+ * @token: Returned token; use in subsequent API calls
-+ *
-+ * This function can be used to open a control session for an
-+ * already created object; an object may have been declared in
-+ * the DPL or by calling the dpdmai_create() function.
-+ * This function returns a unique authentication token,
-+ * associated with the specific object ID and the specific MC
-+ * portal; this token must be used in all subsequent commands for
-+ * this specific object.
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpdmai_open(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ int dpdmai_id,
-+ uint16_t *token);
-+
-+/**
-+ * dpdmai_close() - Close the control session of the object
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPDMAI object
-+ *
-+ * After this function is called, no further operations are
-+ * allowed on the object without opening a new control session.
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpdmai_close(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token);
-+
-+/**
-+ * struct dpdmai_cfg - Structure representing DPDMAI configuration
-+ * @priorities: Priorities for the DMA hardware processing; valid priorities are
-+ * configured with values 1-8; the entry following last valid entry
-+ * should be configured with 0
-+ */
-+struct dpdmai_cfg {
-+ uint8_t priorities[DPDMAI_PRIO_NUM];
-+};
-+
-+/**
-+ * dpdmai_create() - Create the DPDMAI object
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @cfg: Configuration structure
-+ * @token: Returned token; use in subsequent API calls
-+ *
-+ * Create the DPDMAI object, allocate required resources and
-+ * perform required initialization.
-+ *
-+ * The object can be created either by declaring it in the
-+ * DPL file, or by calling this function.
-+ *
-+ * This function returns a unique authentication token,
-+ * associated with the specific object ID and the specific MC
-+ * portal; this token must be used in all subsequent calls to
-+ * this specific object. For objects that are created using the
-+ * DPL file, call dpdmai_open() function to get an authentication
-+ * token first.
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpdmai_create(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ const struct dpdmai_cfg *cfg,
-+ uint16_t *token);
-+
-+/**
-+ * dpdmai_destroy() - Destroy the DPDMAI object and release all its resources.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPDMAI object
-+ *
-+ * Return: '0' on Success; error code otherwise.
-+ */
-+int dpdmai_destroy(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token);
-+
-+/**
-+ * dpdmai_enable() - Enable the DPDMAI, allow sending and receiving frames.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPDMAI object
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpdmai_enable(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token);
-+
-+/**
-+ * dpdmai_disable() - Disable the DPDMAI, stop sending and receiving frames.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPDMAI object
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpdmai_disable(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token);
-+
-+/**
-+ * dpdmai_is_enabled() - Check if the DPDMAI is enabled.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPDMAI object
-+ * @en: Returns '1' if object is enabled; '0' otherwise
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpdmai_is_enabled(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token,
-+ int *en);
-+
-+/**
-+ * dpdmai_reset() - Reset the DPDMAI, returns the object to initial state.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPDMAI object
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpdmai_reset(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token);
-+
-+/**
-+ * struct dpdmai_irq_cfg - IRQ configuration
-+ * @addr: Address that must be written to signal a message-based interrupt
-+ * @val: Value to write into irq_addr address
-+ * @irq_num: A user defined number associated with this IRQ
-+ */
-+struct dpdmai_irq_cfg {
-+ uint64_t addr;
-+ uint32_t val;
-+ int irq_num;
-+};
-+
-+/**
-+ * dpdmai_set_irq() - Set IRQ information for the DPDMAI to trigger an interrupt.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPDMAI object
-+ * @irq_index: Identifies the interrupt index to configure
-+ * @irq_cfg: IRQ configuration
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpdmai_set_irq(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token,
-+ uint8_t irq_index,
-+ struct dpdmai_irq_cfg *irq_cfg);
-+
-+/**
-+ * dpdmai_get_irq() - Get IRQ information from the DPDMAI
-+ *
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPDMAI object
-+ * @irq_index: The interrupt index to configure
-+ * @type: Interrupt type: 0 represents message interrupt
-+ * type (both irq_addr and irq_val are valid)
-+ * @irq_cfg: IRQ attributes
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpdmai_get_irq(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token,
-+ uint8_t irq_index,
-+ int *type,
-+ struct dpdmai_irq_cfg *irq_cfg);
-+
-+/**
-+ * dpdmai_set_irq_enable() - Set overall interrupt state.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPDMAI object
-+ * @irq_index: The interrupt index to configure
-+ * @en: Interrupt state - enable = 1, disable = 0
-+ *
-+ * Allows GPP software to control when interrupts are generated.
-+ * Each interrupt can have up to 32 causes. The enable/disable control's the
-+ * overall interrupt state. if the interrupt is disabled no causes will cause
-+ * an interrupt
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpdmai_set_irq_enable(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token,
-+ uint8_t irq_index,
-+ uint8_t en);
-+
-+/**
-+ * dpdmai_get_irq_enable() - Get overall interrupt state
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPDMAI object
-+ * @irq_index: The interrupt index to configure
-+ * @en: Returned Interrupt state - enable = 1, disable = 0
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpdmai_get_irq_enable(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token,
-+ uint8_t irq_index,
-+ uint8_t *en);
-+
-+/**
-+ * dpdmai_set_irq_mask() - Set interrupt mask.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPDMAI object
-+ * @irq_index: The interrupt index to configure
-+ * @mask: event mask to trigger interrupt;
-+ * each bit:
-+ * 0 = ignore event
-+ * 1 = consider event for asserting IRQ
-+ *
-+ * Every interrupt can have up to 32 causes and the interrupt model supports
-+ * masking/unmasking each cause independently
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpdmai_set_irq_mask(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token,
-+ uint8_t irq_index,
-+ uint32_t mask);
-+
-+/**
-+ * dpdmai_get_irq_mask() - Get interrupt mask.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPDMAI object
-+ * @irq_index: The interrupt index to configure
-+ * @mask: Returned event mask to trigger interrupt
-+ *
-+ * Every interrupt can have up to 32 causes and the interrupt model supports
-+ * masking/unmasking each cause independently
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpdmai_get_irq_mask(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token,
-+ uint8_t irq_index,
-+ uint32_t *mask);
-+
-+/**
-+ * dpdmai_get_irq_status() - Get the current status of any pending interrupts
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPDMAI object
-+ * @irq_index: The interrupt index to configure
-+ * @status: Returned interrupts status - one bit per cause:
-+ * 0 = no interrupt pending
-+ * 1 = interrupt pending
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpdmai_get_irq_status(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token,
-+ uint8_t irq_index,
-+ uint32_t *status);
-+
-+/**
-+ * dpdmai_clear_irq_status() - Clear a pending interrupt's status
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPDMAI object
-+ * @irq_index: The interrupt index to configure
-+ * @status: bits to clear (W1C) - one bit per cause:
-+ * 0 = don't change
-+ * 1 = clear status bit
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpdmai_clear_irq_status(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token,
-+ uint8_t irq_index,
-+ uint32_t status);
-+
-+/**
-+ * struct dpdmai_attr - Structure representing DPDMAI attributes
-+ * @id: DPDMAI object ID
-+ * @version: DPDMAI version
-+ * @num_of_priorities: number of priorities
-+ */
-+struct dpdmai_attr {
-+ int id;
-+ /**
-+ * struct version - DPDMAI version
-+ * @major: DPDMAI major version
-+ * @minor: DPDMAI minor version
-+ */
-+ struct {
-+ uint16_t major;
-+ uint16_t minor;
-+ } version;
-+ uint8_t num_of_priorities;
-+};
-+
-+/**
-+ * dpdmai_get_attributes() - Retrieve DPDMAI attributes.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPDMAI object
-+ * @attr: Returned object's attributes
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpdmai_get_attributes(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token,
-+ struct dpdmai_attr *attr);
-+
-+/**
-+ * enum dpdmai_dest - DPDMAI destination types
-+ * @DPDMAI_DEST_NONE: Unassigned destination; The queue is set in parked mode
-+ * and does not generate FQDAN notifications; user is expected to dequeue
-+ * from the queue based on polling or other user-defined method
-+ * @DPDMAI_DEST_DPIO: The queue is set in schedule mode and generates FQDAN
-+ * notifications to the specified DPIO; user is expected to dequeue
-+ * from the queue only after notification is received
-+ * @DPDMAI_DEST_DPCON: The queue is set in schedule mode and does not generate
-+ * FQDAN notifications, but is connected to the specified DPCON object;
-+ * user is expected to dequeue from the DPCON channel
-+ */
-+enum dpdmai_dest {
-+ DPDMAI_DEST_NONE = 0,
-+ DPDMAI_DEST_DPIO = 1,
-+ DPDMAI_DEST_DPCON = 2
-+};
-+
-+/**
-+ * struct dpdmai_dest_cfg - Structure representing DPDMAI destination parameters
-+ * @dest_type: Destination type
-+ * @dest_id: Either DPIO ID or DPCON ID, depending on the destination type
-+ * @priority: Priority selection within the DPIO or DPCON channel; valid values
-+ * are 0-1 or 0-7, depending on the number of priorities in that
-+ * channel; not relevant for 'DPDMAI_DEST_NONE' option
-+ */
-+struct dpdmai_dest_cfg {
-+ enum dpdmai_dest dest_type;
-+ int dest_id;
-+ uint8_t priority;
-+};
-+
-+/* DPDMAI queue modification options */
-+
-+/**
-+ * Select to modify the user's context associated with the queue
-+ */
-+#define DPDMAI_QUEUE_OPT_USER_CTX 0x00000001
-+
-+/**
-+ * Select to modify the queue's destination
-+ */
-+#define DPDMAI_QUEUE_OPT_DEST 0x00000002
-+
-+/**
-+ * struct dpdmai_rx_queue_cfg - DPDMAI RX queue configuration
-+ * @options: Flags representing the suggested modifications to the queue;
-+ * Use any combination of 'DPDMAI_QUEUE_OPT_<X>' flags
-+ * @user_ctx: User context value provided in the frame descriptor of each
-+ * dequeued frame;
-+ * valid only if 'DPDMAI_QUEUE_OPT_USER_CTX' is contained in 'options'
-+ * @dest_cfg: Queue destination parameters;
-+ * valid only if 'DPDMAI_QUEUE_OPT_DEST' is contained in 'options'
-+ */
-+struct dpdmai_rx_queue_cfg {
-+ uint32_t options;
-+ uint64_t user_ctx;
-+ struct dpdmai_dest_cfg dest_cfg;
-+
-+};
-+
-+/**
-+ * dpdmai_set_rx_queue() - Set Rx queue configuration
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPDMAI object
-+ * @priority: Select the queue relative to number of
-+ * priorities configured at DPDMAI creation; use
-+ * DPDMAI_ALL_QUEUES to configure all Rx queues
-+ * identically.
-+ * @cfg: Rx queue configuration
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpdmai_set_rx_queue(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token,
-+ uint8_t priority,
-+ const struct dpdmai_rx_queue_cfg *cfg);
-+
-+/**
-+ * struct dpdmai_rx_queue_attr - Structure representing attributes of Rx queues
-+ * @user_ctx: User context value provided in the frame descriptor of each
-+ * dequeued frame
-+ * @dest_cfg: Queue destination configuration
-+ * @fqid: Virtual FQID value to be used for dequeue operations
-+ */
-+struct dpdmai_rx_queue_attr {
-+ uint64_t user_ctx;
-+ struct dpdmai_dest_cfg dest_cfg;
-+ uint32_t fqid;
-+};
-+
-+/**
-+ * dpdmai_get_rx_queue() - Retrieve Rx queue attributes.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPDMAI object
-+ * @priority: Select the queue relative to number of
-+ * priorities configured at DPDMAI creation
-+ * @attr: Returned Rx queue attributes
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpdmai_get_rx_queue(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token,
-+ uint8_t priority,
-+ struct dpdmai_rx_queue_attr *attr);
-+
-+/**
-+ * struct dpdmai_tx_queue_attr - Structure representing attributes of Tx queues
-+ * @fqid: Virtual FQID to be used for sending frames to DMA hardware
-+ */
-+
-+struct dpdmai_tx_queue_attr {
-+ uint32_t fqid;
-+};
-+
-+/**
-+ * dpdmai_get_tx_queue() - Retrieve Tx queue attributes.
-+ * @mc_io: Pointer to MC portal's I/O object
-+ * @cmd_flags: Command flags; one or more of 'MC_CMD_FLAG_'
-+ * @token: Token of DPDMAI object
-+ * @priority: Select the queue relative to number of
-+ * priorities configured at DPDMAI creation
-+ * @attr: Returned Tx queue attributes
-+ *
-+ * Return: '0' on Success; Error code otherwise.
-+ */
-+int dpdmai_get_tx_queue(struct fsl_mc_io *mc_io,
-+ uint32_t cmd_flags,
-+ uint16_t token,
-+ uint8_t priority,
-+ struct dpdmai_tx_queue_attr *attr);
-+
-+#endif /* __FSL_DPDMAI_H */
---- /dev/null
-+++ b/drivers/dma/dpaa2-qdma/fsl_dpdmai_cmd.h
-@@ -0,0 +1,222 @@
-+/* Copyright 2013-2016 Freescale Semiconductor Inc.
-+ *
-+ * Redistribution and use in source and binary forms, with or without
-+ * modification, are permitted provided that the following conditions are met:
-+ * * Redistributions of source code must retain the above copyright
-+ * notice, this list of conditions and the following disclaimer.
-+ * * Redistributions in binary form must reproduce the above copyright
-+ * notice, this list of conditions and the following disclaimer in the
-+ * documentation and/or other materials provided with the distribution.
-+ * * Neither the name of the above-listed copyright holders nor the
-+ * names of any contributors may be used to endorse or promote products
-+ * derived from this software without specific prior written permission.
-+ *
-+ *
-+ * ALTERNATIVELY, this software may be distributed under the terms of the
-+ * GNU General Public License ("GPL") as published by the Free Software
-+ * Foundation, either version 2 of that License or (at your option) any
-+ * later version.
-+ *
-+ * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
-+ * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
-+ * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
-+ * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE
-+ * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
-+ * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
-+ * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
-+ * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
-+ * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
-+ * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
-+ * POSSIBILITY OF SUCH DAMAGE.
-+ */
-+#ifndef _FSL_DPDMAI_CMD_H
-+#define _FSL_DPDMAI_CMD_H
-+
-+/* DPDMAI Version */
-+#define DPDMAI_VER_MAJOR 2
-+#define DPDMAI_VER_MINOR 2
-+
-+#define DPDMAI_CMD_BASE_VERSION 0
-+#define DPDMAI_CMD_ID_OFFSET 4
-+
-+/* Command IDs */
-+#define DPDMAI_CMDID_CLOSE ((0x800 << DPDMAI_CMD_ID_OFFSET) | DPDMAI_CMD_BASE_VERSION)
-+#define DPDMAI_CMDID_OPEN ((0x80E << DPDMAI_CMD_ID_OFFSET) | DPDMAI_CMD_BASE_VERSION)
-+#define DPDMAI_CMDID_CREATE ((0x90E << DPDMAI_CMD_ID_OFFSET) | DPDMAI_CMD_BASE_VERSION)
-+#define DPDMAI_CMDID_DESTROY ((0x900 << DPDMAI_CMD_ID_OFFSET) | DPDMAI_CMD_BASE_VERSION)
-+
-+#define DPDMAI_CMDID_ENABLE ((0x002 << DPDMAI_CMD_ID_OFFSET) | DPDMAI_CMD_BASE_VERSION)
-+#define DPDMAI_CMDID_DISABLE ((0x003 << DPDMAI_CMD_ID_OFFSET) | DPDMAI_CMD_BASE_VERSION)
-+#define DPDMAI_CMDID_GET_ATTR ((0x004 << DPDMAI_CMD_ID_OFFSET) | DPDMAI_CMD_BASE_VERSION)
-+#define DPDMAI_CMDID_RESET ((0x005 << DPDMAI_CMD_ID_OFFSET) | DPDMAI_CMD_BASE_VERSION)
-+#define DPDMAI_CMDID_IS_ENABLED ((0x006 << DPDMAI_CMD_ID_OFFSET) | DPDMAI_CMD_BASE_VERSION)
-+
-+#define DPDMAI_CMDID_SET_IRQ ((0x010 << DPDMAI_CMD_ID_OFFSET) | DPDMAI_CMD_BASE_VERSION)
-+#define DPDMAI_CMDID_GET_IRQ ((0x011 << DPDMAI_CMD_ID_OFFSET) | DPDMAI_CMD_BASE_VERSION)
-+#define DPDMAI_CMDID_SET_IRQ_ENABLE ((0x012 << DPDMAI_CMD_ID_OFFSET) | DPDMAI_CMD_BASE_VERSION)
-+#define DPDMAI_CMDID_GET_IRQ_ENABLE ((0x013 << DPDMAI_CMD_ID_OFFSET) | DPDMAI_CMD_BASE_VERSION)
-+#define DPDMAI_CMDID_SET_IRQ_MASK ((0x014 << DPDMAI_CMD_ID_OFFSET) | DPDMAI_CMD_BASE_VERSION)
-+#define DPDMAI_CMDID_GET_IRQ_MASK ((0x015 << DPDMAI_CMD_ID_OFFSET) | DPDMAI_CMD_BASE_VERSION)
-+#define DPDMAI_CMDID_GET_IRQ_STATUS ((0x016 << DPDMAI_CMD_ID_OFFSET) | DPDMAI_CMD_BASE_VERSION)
-+#define DPDMAI_CMDID_CLEAR_IRQ_STATUS ((0x017 << DPDMAI_CMD_ID_OFFSET) | DPDMAI_CMD_BASE_VERSION)
-+
-+#define DPDMAI_CMDID_SET_RX_QUEUE ((0x1A0 << DPDMAI_CMD_ID_OFFSET) | DPDMAI_CMD_BASE_VERSION)
-+#define DPDMAI_CMDID_GET_RX_QUEUE ((0x1A1 << DPDMAI_CMD_ID_OFFSET) | DPDMAI_CMD_BASE_VERSION)
-+#define DPDMAI_CMDID_GET_TX_QUEUE ((0x1A2 << DPDMAI_CMD_ID_OFFSET) | DPDMAI_CMD_BASE_VERSION)
-+
-+
-+#define MC_CMD_HDR_TOKEN_O 32 /* Token field offset */
-+#define MC_CMD_HDR_TOKEN_S 16 /* Token field size */
-+
-+
-+#define MAKE_UMASK64(_width) \
-+ ((uint64_t)((_width) < 64 ? ((uint64_t)1 << (_width)) - 1 : \
-+ (uint64_t)-1))
-+
-+static inline uint64_t mc_enc(int lsoffset, int width, uint64_t val)
-+{
-+ return (uint64_t)(((uint64_t)val & MAKE_UMASK64(width)) << lsoffset);
-+}
-+
-+static inline uint64_t mc_dec(uint64_t val, int lsoffset, int width)
-+{
-+ return (uint64_t)((val >> lsoffset) & MAKE_UMASK64(width));
-+}
-+
-+#define MC_CMD_OP(_cmd, _param, _offset, _width, _type, _arg) \
-+ ((_cmd).params[_param] |= mc_enc((_offset), (_width), _arg))
-+
-+#define MC_RSP_OP(_cmd, _param, _offset, _width, _type, _arg) \
-+ (_arg = (_type)mc_dec(_cmd.params[_param], (_offset), (_width)))
-+
-+#define MC_CMD_HDR_READ_TOKEN(_hdr) \
-+ ((uint16_t)mc_dec((_hdr), MC_CMD_HDR_TOKEN_O, MC_CMD_HDR_TOKEN_S))
-+
-+/* cmd, param, offset, width, type, arg_name */
-+#define DPDMAI_CMD_OPEN(cmd, dpdmai_id) \
-+ MC_CMD_OP(cmd, 0, 0, 32, int, dpdmai_id)
-+
-+/* cmd, param, offset, width, type, arg_name */
-+#define DPDMAI_CMD_CREATE(cmd, cfg) \
-+do { \
-+ MC_CMD_OP(cmd, 0, 8, 8, uint8_t, cfg->priorities[0]);\
-+ MC_CMD_OP(cmd, 0, 16, 8, uint8_t, cfg->priorities[1]);\
-+} while (0)
-+
-+/* cmd, param, offset, width, type, arg_name */
-+#define DPDMAI_RSP_IS_ENABLED(cmd, en) \
-+ MC_RSP_OP(cmd, 0, 0, 1, int, en)
-+
-+/* cmd, param, offset, width, type, arg_name */
-+#define DPDMAI_CMD_SET_IRQ(cmd, irq_index, irq_cfg) \
-+do { \
-+ MC_CMD_OP(cmd, 0, 0, 8, uint8_t, irq_index);\
-+ MC_CMD_OP(cmd, 0, 32, 32, uint32_t, irq_cfg->val);\
-+ MC_CMD_OP(cmd, 1, 0, 64, uint64_t, irq_cfg->addr);\
-+ MC_CMD_OP(cmd, 2, 0, 32, int, irq_cfg->irq_num); \
-+} while (0)
-+
-+/* cmd, param, offset, width, type, arg_name */
-+#define DPDMAI_CMD_GET_IRQ(cmd, irq_index) \
-+ MC_CMD_OP(cmd, 0, 32, 8, uint8_t, irq_index)
-+
-+/* cmd, param, offset, width, type, arg_name */
-+#define DPDMAI_RSP_GET_IRQ(cmd, type, irq_cfg) \
-+do { \
-+ MC_RSP_OP(cmd, 0, 0, 32, uint32_t, irq_cfg->val); \
-+ MC_RSP_OP(cmd, 1, 0, 64, uint64_t, irq_cfg->addr);\
-+ MC_RSP_OP(cmd, 2, 0, 32, int, irq_cfg->irq_num); \
-+ MC_RSP_OP(cmd, 2, 32, 32, int, type); \
-+} while (0)
-+
-+/* cmd, param, offset, width, type, arg_name */
-+#define DPDMAI_CMD_SET_IRQ_ENABLE(cmd, irq_index, enable_state) \
-+do { \
-+ MC_CMD_OP(cmd, 0, 0, 8, uint8_t, enable_state); \
-+ MC_CMD_OP(cmd, 0, 32, 8, uint8_t, irq_index); \
-+} while (0)
-+
-+/* cmd, param, offset, width, type, arg_name */
-+#define DPDMAI_CMD_GET_IRQ_ENABLE(cmd, irq_index) \
-+ MC_CMD_OP(cmd, 0, 32, 8, uint8_t, irq_index)
-+
-+/* cmd, param, offset, width, type, arg_name */
-+#define DPDMAI_RSP_GET_IRQ_ENABLE(cmd, enable_state) \
-+ MC_RSP_OP(cmd, 0, 0, 8, uint8_t, enable_state)
-+
-+/* cmd, param, offset, width, type, arg_name */
-+#define DPDMAI_CMD_SET_IRQ_MASK(cmd, irq_index, mask) \
-+do { \
-+ MC_CMD_OP(cmd, 0, 0, 32, uint32_t, mask); \
-+ MC_CMD_OP(cmd, 0, 32, 8, uint8_t, irq_index); \
-+} while (0)
-+
-+/* cmd, param, offset, width, type, arg_name */
-+#define DPDMAI_CMD_GET_IRQ_MASK(cmd, irq_index) \
-+ MC_CMD_OP(cmd, 0, 32, 8, uint8_t, irq_index)
-+
-+/* cmd, param, offset, width, type, arg_name */
-+#define DPDMAI_RSP_GET_IRQ_MASK(cmd, mask) \
-+ MC_RSP_OP(cmd, 0, 0, 32, uint32_t, mask)
-+
-+/* cmd, param, offset, width, type, arg_name */
-+#define DPDMAI_CMD_GET_IRQ_STATUS(cmd, irq_index, status) \
-+do { \
-+ MC_CMD_OP(cmd, 0, 0, 32, uint32_t, status);\
-+ MC_CMD_OP(cmd, 0, 32, 8, uint8_t, irq_index);\
-+} while (0)
-+
-+/* cmd, param, offset, width, type, arg_name */
-+#define DPDMAI_RSP_GET_IRQ_STATUS(cmd, status) \
-+ MC_RSP_OP(cmd, 0, 0, 32, uint32_t, status)
-+
-+/* cmd, param, offset, width, type, arg_name */
-+#define DPDMAI_CMD_CLEAR_IRQ_STATUS(cmd, irq_index, status) \
-+do { \
-+ MC_CMD_OP(cmd, 0, 0, 32, uint32_t, status); \
-+ MC_CMD_OP(cmd, 0, 32, 8, uint8_t, irq_index); \
-+} while (0)
-+
-+/* cmd, param, offset, width, type, arg_name */
-+#define DPDMAI_RSP_GET_ATTR(cmd, attr) \
-+do { \
-+ MC_RSP_OP(cmd, 0, 0, 32, int, attr->id); \
-+ MC_RSP_OP(cmd, 0, 32, 8, uint8_t, attr->num_of_priorities); \
-+ MC_RSP_OP(cmd, 1, 0, 16, uint16_t, attr->version.major);\
-+ MC_RSP_OP(cmd, 1, 16, 16, uint16_t, attr->version.minor);\
-+} while (0)
-+
-+/* cmd, param, offset, width, type, arg_name */
-+#define DPDMAI_CMD_SET_RX_QUEUE(cmd, priority, cfg) \
-+do { \
-+ MC_CMD_OP(cmd, 0, 0, 32, int, cfg->dest_cfg.dest_id); \
-+ MC_CMD_OP(cmd, 0, 32, 8, uint8_t, cfg->dest_cfg.priority); \
-+ MC_CMD_OP(cmd, 0, 40, 8, uint8_t, priority); \
-+ MC_CMD_OP(cmd, 0, 48, 4, enum dpdmai_dest, cfg->dest_cfg.dest_type); \
-+ MC_CMD_OP(cmd, 1, 0, 64, uint64_t, cfg->user_ctx); \
-+ MC_CMD_OP(cmd, 2, 0, 32, uint32_t, cfg->options);\
-+} while (0)
-+
-+/* cmd, param, offset, width, type, arg_name */
-+#define DPDMAI_CMD_GET_RX_QUEUE(cmd, priority) \
-+ MC_CMD_OP(cmd, 0, 40, 8, uint8_t, priority)
-+
-+/* cmd, param, offset, width, type, arg_name */
-+#define DPDMAI_RSP_GET_RX_QUEUE(cmd, attr) \
-+do { \
-+ MC_RSP_OP(cmd, 0, 0, 32, int, attr->dest_cfg.dest_id);\
-+ MC_RSP_OP(cmd, 0, 32, 8, uint8_t, attr->dest_cfg.priority);\
-+ MC_RSP_OP(cmd, 0, 48, 4, enum dpdmai_dest, attr->dest_cfg.dest_type);\
-+ MC_RSP_OP(cmd, 1, 0, 64, uint64_t, attr->user_ctx);\
-+ MC_RSP_OP(cmd, 2, 0, 32, uint32_t, attr->fqid);\
-+} while (0)
-+
-+/* cmd, param, offset, width, type, arg_name */
-+#define DPDMAI_CMD_GET_TX_QUEUE(cmd, priority) \
-+ MC_CMD_OP(cmd, 0, 40, 8, uint8_t, priority)
-+
-+/* cmd, param, offset, width, type, arg_name */
-+#define DPDMAI_RSP_GET_TX_QUEUE(cmd, attr) \
-+ MC_RSP_OP(cmd, 1, 0, 32, uint32_t, attr->fqid)
-+
-+#endif /* _FSL_DPDMAI_CMD_H */
---- /dev/null
-+++ b/drivers/dma/fsl-qdma.c
-@@ -0,0 +1,1243 @@
-+/*
-+ * drivers/dma/fsl-qdma.c
-+ *
-+ * Copyright 2014-2015 Freescale Semiconductor, Inc.
-+ *
-+ * Driver for the Freescale qDMA engine with software command queue mode.
-+ * Channel virtualization is supported through enqueuing of DMA jobs to,
-+ * or dequeuing DMA jobs from, different work queues.
-+ * This module can be found on Freescale LS SoCs.
-+ *
-+ * This program is free software; you can redistribute it and/or modify it
-+ * under the terms of the GNU General Public License as published by the
-+ * Free Software Foundation; either version 2 of the License, or (at your
-+ * option) any later version.
-+ */
-+
-+#include <asm/cacheflush.h>
-+#include <linux/clk.h>
-+#include <linux/delay.h>
-+#include <linux/dma-mapping.h>
-+#include <linux/dmapool.h>
-+#include <linux/init.h>
-+#include <linux/interrupt.h>
-+#include <linux/module.h>
-+#include <linux/of.h>
-+#include <linux/of_address.h>
-+#include <linux/of_device.h>
-+#include <linux/of_dma.h>
-+#include <linux/of_irq.h>
-+#include <linux/slab.h>
-+#include <linux/spinlock.h>
-+
-+#include "virt-dma.h"
-+
-+#define FSL_QDMA_DMR 0x0
-+#define FSL_QDMA_DSR 0x4
-+#define FSL_QDMA_DEIER 0xe00
-+#define FSL_QDMA_DEDR 0xe04
-+#define FSL_QDMA_DECFDW0R 0xe10
-+#define FSL_QDMA_DECFDW1R 0xe14
-+#define FSL_QDMA_DECFDW2R 0xe18
-+#define FSL_QDMA_DECFDW3R 0xe1c
-+#define FSL_QDMA_DECFQIDR 0xe30
-+#define FSL_QDMA_DECBR 0xe34
-+
-+#define FSL_QDMA_BCQMR(x) (0xc0 + 0x100 * (x))
-+#define FSL_QDMA_BCQSR(x) (0xc4 + 0x100 * (x))
-+#define FSL_QDMA_BCQEDPA_SADDR(x) (0xc8 + 0x100 * (x))
-+#define FSL_QDMA_BCQDPA_SADDR(x) (0xcc + 0x100 * (x))
-+#define FSL_QDMA_BCQEEPA_SADDR(x) (0xd0 + 0x100 * (x))
-+#define FSL_QDMA_BCQEPA_SADDR(x) (0xd4 + 0x100 * (x))
-+#define FSL_QDMA_BCQIER(x) (0xe0 + 0x100 * (x))
-+#define FSL_QDMA_BCQIDR(x) (0xe4 + 0x100 * (x))
-+
-+#define FSL_QDMA_SQDPAR 0x80c
-+#define FSL_QDMA_SQEPAR 0x814
-+#define FSL_QDMA_BSQMR 0x800
-+#define FSL_QDMA_BSQSR 0x804
-+#define FSL_QDMA_BSQICR 0x828
-+#define FSL_QDMA_CQMR 0xa00
-+#define FSL_QDMA_CQDSCR1 0xa08
-+#define FSL_QDMA_CQDSCR2 0xa0c
-+#define FSL_QDMA_CQIER 0xa10
-+#define FSL_QDMA_CQEDR 0xa14
-+#define FSL_QDMA_SQCCMR 0xa20
-+
-+#define FSL_QDMA_SQICR_ICEN
-+
-+#define FSL_QDMA_CQIDR_CQT 0xff000000
-+#define FSL_QDMA_CQIDR_SQPE 0x800000
-+#define FSL_QDMA_CQIDR_SQT 0x8000
-+
-+#define FSL_QDMA_BCQIER_CQTIE 0x8000
-+#define FSL_QDMA_BCQIER_CQPEIE 0x800000
-+#define FSL_QDMA_BSQICR_ICEN 0x80000000
-+#define FSL_QDMA_BSQICR_ICST(x) ((x) << 16)
-+#define FSL_QDMA_CQIER_MEIE 0x80000000
-+#define FSL_QDMA_CQIER_TEIE 0x1
-+#define FSL_QDMA_SQCCMR_ENTER_WM 0x200000
-+
-+#define FSL_QDMA_QUEUE_MAX 8
-+
-+#define FSL_QDMA_BCQMR_EN 0x80000000
-+#define FSL_QDMA_BCQMR_EI 0x40000000
-+#define FSL_QDMA_BCQMR_CD_THLD(x) ((x) << 20)
-+#define FSL_QDMA_BCQMR_CQ_SIZE(x) ((x) << 16)
-+
-+#define FSL_QDMA_BCQSR_QF 0x10000
-+#define FSL_QDMA_BCQSR_XOFF 0x1
-+
-+#define FSL_QDMA_BSQMR_EN 0x80000000
-+#define FSL_QDMA_BSQMR_DI 0x40000000
-+#define FSL_QDMA_BSQMR_CQ_SIZE(x) ((x) << 16)
-+
-+#define FSL_QDMA_BSQSR_QE 0x20000
-+
-+#define FSL_QDMA_DMR_DQD 0x40000000
-+#define FSL_QDMA_DSR_DB 0x80000000
-+
-+#define FSL_QDMA_BASE_BUFFER_SIZE 96
-+#define FSL_QDMA_EXPECT_SG_ENTRY_NUM 16
-+#define FSL_QDMA_CIRCULAR_DESC_SIZE_MIN 64
-+#define FSL_QDMA_CIRCULAR_DESC_SIZE_MAX 16384
-+#define FSL_QDMA_QUEUE_NUM_MAX 8
-+
-+#define FSL_QDMA_CMD_RWTTYPE 0x4
-+#define FSL_QDMA_CMD_LWC 0x2
-+
-+#define FSL_QDMA_CMD_RWTTYPE_OFFSET 28
-+#define FSL_QDMA_CMD_NS_OFFSET 27
-+#define FSL_QDMA_CMD_DQOS_OFFSET 24
-+#define FSL_QDMA_CMD_WTHROTL_OFFSET 20
-+#define FSL_QDMA_CMD_DSEN_OFFSET 19
-+#define FSL_QDMA_CMD_LWC_OFFSET 16
-+
-+#define FSL_QDMA_E_SG_TABLE 1
-+#define FSL_QDMA_E_DATA_BUFFER 0
-+#define FSL_QDMA_F_LAST_ENTRY 1
-+
-+u64 pre_addr, pre_queue;
-+
-+/* qDMA Command Descriptor Fotmats */
-+
-+/* Compound Command Descriptor Fotmat */
-+struct fsl_qdma_ccdf {
-+ __le32 status; /* ser, status */
-+ __le32 cfg; /* format, offset */
-+ union {
-+ struct {
-+ __le32 addr_lo; /* low 32-bits of 40-bit address */
-+ u8 addr_hi; /* high 8-bits of 40-bit address */
-+ u8 __reserved1[2];
-+ u8 cfg8b_w1; /* dd, queue*/
-+ } __packed;
-+ __le64 data;
-+ };
-+} __packed;
-+
-+#define QDMA_CCDF_STATUS 20
-+#define QDMA_CCDF_OFFSET 20
-+#define QDMA_CCDF_MASK GENMASK(28, 20)
-+#define QDMA_CCDF_FOTMAT BIT(29)
-+#define QDMA_CCDF_SER BIT(30)
-+
-+static inline u64 qdma_ccdf_addr_get64(const struct fsl_qdma_ccdf *ccdf)
-+{
-+ return le64_to_cpu(ccdf->data) & 0xffffffffffLLU;
-+}
-+static inline u64 qdma_ccdf_get_queue(const struct fsl_qdma_ccdf *ccdf)
-+{
-+ return ccdf->cfg8b_w1 & 0xff;
-+}
-+static inline void qdma_ccdf_addr_set64(struct fsl_qdma_ccdf *ccdf, u64 addr)
-+{
-+ ccdf->addr_hi = upper_32_bits(addr);
-+ ccdf->addr_lo = cpu_to_le32(lower_32_bits(addr));
-+}
-+static inline int qdma_ccdf_get_offset(const struct fsl_qdma_ccdf *ccdf)
-+{
-+ return (le32_to_cpu(ccdf->cfg) & QDMA_CCDF_MASK) >> QDMA_CCDF_OFFSET;
-+}
-+static inline void qdma_ccdf_set_format(struct fsl_qdma_ccdf *ccdf, int offset)
-+{
-+ ccdf->cfg = cpu_to_le32(QDMA_CCDF_FOTMAT | offset);
-+}
-+static inline int qdma_ccdf_get_status(const struct fsl_qdma_ccdf *ccdf)
-+{
-+ return (le32_to_cpu(ccdf->status) & QDMA_CCDF_MASK) >> QDMA_CCDF_STATUS;
-+}
-+static inline void qdma_ccdf_set_ser(struct fsl_qdma_ccdf *ccdf, int status)
-+{
-+ ccdf->status = cpu_to_le32(QDMA_CCDF_SER | status);
-+}
-+/* qDMA Compound S/G Format */
-+struct fsl_qdma_csgf {
-+ __le32 offset; /* offset */
-+ __le32 cfg; /* E bit, F bit, length */
-+ union {
-+ struct {
-+ __le32 addr_lo; /* low 32-bits of 40-bit address */
-+ u8 addr_hi; /* high 8-bits of 40-bit address */
-+ u8 __reserved1[3];
-+ };
-+ __le64 data;
-+ };
-+} __packed;
-+
-+#define QDMA_SG_FIN BIT(30)
-+#define QDMA_SG_EXT BIT(31)
-+#define QDMA_SG_LEN_MASK GENMASK(29, 0)
-+static inline u64 qdma_csgf_addr_get64(const struct fsl_qdma_csgf *sg)
-+{
-+ return be64_to_cpu(sg->data) & 0xffffffffffLLU;
-+}
-+static inline void qdma_csgf_addr_set64(struct fsl_qdma_csgf *sg, u64 addr)
-+{
-+ sg->addr_hi = upper_32_bits(addr);
-+ sg->addr_lo = cpu_to_le32(lower_32_bits(addr));
-+}
-+static inline void qdma_csgf_set_len(struct fsl_qdma_csgf *csgf, int len)
-+{
-+ csgf->cfg = cpu_to_le32(len & QDMA_SG_LEN_MASK);
-+}
-+static inline void qdma_csgf_set_f(struct fsl_qdma_csgf *csgf, int len)
-+{
-+ csgf->cfg = cpu_to_le32(QDMA_SG_FIN | (len & QDMA_SG_LEN_MASK));
-+}
-+static inline void qdma_csgf_set_e(struct fsl_qdma_csgf *csgf, int len)
-+{
-+ csgf->cfg = cpu_to_le32(QDMA_SG_EXT | (len & QDMA_SG_LEN_MASK));
-+}
-+
-+/* qDMA Source Descriptor Format */
-+struct fsl_qdma_sdf {
-+ __le32 rev3;
-+ __le32 cfg; /* rev4, bit[0-11] - ssd, bit[12-23] sss */
-+ __le32 rev5;
-+ __le32 cmd;
-+} __packed;
-+
-+/*qDMA Destination Descriptor Format*/
-+struct fsl_qdma_ddf {
-+ __le32 rev1;
-+ __le32 cfg; /* rev2, bit[0-11] - dsd, bit[12-23] - dss */
-+ __le32 rev3;
-+ __le32 cmd;
-+} __packed;
-+
-+struct fsl_qdma_chan {
-+ struct virt_dma_chan vchan;
-+ struct virt_dma_desc vdesc;
-+ enum dma_status status;
-+ u32 slave_id;
-+ struct fsl_qdma_engine *qdma;
-+ struct fsl_qdma_queue *queue;
-+ struct list_head qcomp;
-+};
-+
-+struct fsl_qdma_queue {
-+ struct fsl_qdma_ccdf *virt_head;
-+ struct fsl_qdma_ccdf *virt_tail;
-+ struct list_head comp_used;
-+ struct list_head comp_free;
-+ struct dma_pool *comp_pool;
-+ struct dma_pool *sg_pool;
-+ spinlock_t queue_lock;
-+ dma_addr_t bus_addr;
-+ u32 n_cq;
-+ u32 id;
-+ struct fsl_qdma_ccdf *cq;
-+};
-+
-+struct fsl_qdma_sg {
-+ dma_addr_t bus_addr;
-+ void *virt_addr;
-+};
-+
-+struct fsl_qdma_comp {
-+ dma_addr_t bus_addr;
-+ void *virt_addr;
-+ struct fsl_qdma_chan *qchan;
-+ struct fsl_qdma_sg *sg_block;
-+ struct virt_dma_desc vdesc;
-+ struct list_head list;
-+ u32 sg_block_src;
-+ u32 sg_block_dst;
-+};
-+
-+struct fsl_qdma_engine {
-+ struct dma_device dma_dev;
-+ void __iomem *ctrl_base;
-+ void __iomem *status_base;
-+ void __iomem *block_base;
-+ u32 n_chans;
-+ u32 n_queues;
-+ struct mutex fsl_qdma_mutex;
-+ int error_irq;
-+ int queue_irq;
-+ bool big_endian;
-+ struct fsl_qdma_queue *queue;
-+ struct fsl_qdma_queue *status;
-+ struct fsl_qdma_chan chans[];
-+
-+};
-+
-+static u32 qdma_readl(struct fsl_qdma_engine *qdma, void __iomem *addr)
-+{
-+ if (qdma->big_endian)
-+ return ioread32be(addr);
-+ else
-+ return ioread32(addr);
-+}
-+
-+static void qdma_writel(struct fsl_qdma_engine *qdma, u32 val,
-+ void __iomem *addr)
-+{
-+ if (qdma->big_endian)
-+ iowrite32be(val, addr);
-+ else
-+ iowrite32(val, addr);
-+}
-+
-+static struct fsl_qdma_chan *to_fsl_qdma_chan(struct dma_chan *chan)
-+{
-+ return container_of(chan, struct fsl_qdma_chan, vchan.chan);
-+}
-+
-+static struct fsl_qdma_comp *to_fsl_qdma_comp(struct virt_dma_desc *vd)
-+{
-+ return container_of(vd, struct fsl_qdma_comp, vdesc);
-+}
-+
-+static int fsl_qdma_alloc_chan_resources(struct dma_chan *chan)
-+{
-+ /*
-+ * In QDMA mode, We don't need to do anything.
-+ */
-+ return 0;
-+}
-+
-+static void fsl_qdma_free_chan_resources(struct dma_chan *chan)
-+{
-+ struct fsl_qdma_chan *fsl_chan = to_fsl_qdma_chan(chan);
-+ unsigned long flags;
-+ LIST_HEAD(head);
-+
-+ spin_lock_irqsave(&fsl_chan->vchan.lock, flags);
-+ vchan_get_all_descriptors(&fsl_chan->vchan, &head);
-+ spin_unlock_irqrestore(&fsl_chan->vchan.lock, flags);
-+
-+ vchan_dma_desc_free_list(&fsl_chan->vchan, &head);
-+}
-+
-+static void fsl_qdma_comp_fill_memcpy(struct fsl_qdma_comp *fsl_comp,
-+ dma_addr_t dst, dma_addr_t src, u32 len)
-+{
-+ struct fsl_qdma_ccdf *ccdf;
-+ struct fsl_qdma_csgf *csgf_desc, *csgf_src, *csgf_dest;
-+ struct fsl_qdma_sdf *sdf;
-+ struct fsl_qdma_ddf *ddf;
-+
-+ ccdf = (struct fsl_qdma_ccdf *)fsl_comp->virt_addr;
-+ csgf_desc = (struct fsl_qdma_csgf *)fsl_comp->virt_addr + 1;
-+ csgf_src = (struct fsl_qdma_csgf *)fsl_comp->virt_addr + 2;
-+ csgf_dest = (struct fsl_qdma_csgf *)fsl_comp->virt_addr + 3;
-+ sdf = (struct fsl_qdma_sdf *)fsl_comp->virt_addr + 4;
-+ ddf = (struct fsl_qdma_ddf *)fsl_comp->virt_addr + 5;
-+
-+ memset(fsl_comp->virt_addr, 0, FSL_QDMA_BASE_BUFFER_SIZE);
-+ /* Head Command Descriptor(Frame Descriptor) */
-+ qdma_ccdf_addr_set64(ccdf, fsl_comp->bus_addr + 16);
-+ qdma_ccdf_set_format(ccdf, qdma_ccdf_get_offset(ccdf));
-+ qdma_ccdf_set_ser(ccdf, qdma_ccdf_get_status(ccdf));
-+ /* Status notification is enqueued to status queue. */
-+ /* Compound Command Descriptor(Frame List Table) */
-+ qdma_csgf_addr_set64(csgf_desc, fsl_comp->bus_addr + 64);
-+ /* It must be 32 as Compound S/G Descriptor */
-+ qdma_csgf_set_len(csgf_desc, 32);
-+ qdma_csgf_addr_set64(csgf_src, src);
-+ qdma_csgf_set_len(csgf_src, len);
-+ qdma_csgf_addr_set64(csgf_dest, dst);
-+ qdma_csgf_set_len(csgf_dest, len);
-+ /* This entry is the last entry. */
-+ qdma_csgf_set_f(csgf_dest, len);
-+ /* Descriptor Buffer */
-+ sdf->cmd = cpu_to_le32(
-+ FSL_QDMA_CMD_RWTTYPE << FSL_QDMA_CMD_RWTTYPE_OFFSET);
-+ ddf->cmd = cpu_to_le32(
-+ FSL_QDMA_CMD_RWTTYPE << FSL_QDMA_CMD_RWTTYPE_OFFSET);
-+ ddf->cmd |= cpu_to_le32(
-+ FSL_QDMA_CMD_LWC << FSL_QDMA_CMD_LWC_OFFSET);
-+}
-+
-+static void fsl_qdma_comp_fill_sg(
-+ struct fsl_qdma_comp *fsl_comp,
-+ struct scatterlist *dst_sg, unsigned int dst_nents,
-+ struct scatterlist *src_sg, unsigned int src_nents)
-+{
-+ struct fsl_qdma_ccdf *ccdf;
-+ struct fsl_qdma_csgf *csgf_desc, *csgf_src, *csgf_dest, *csgf_sg;
-+ struct fsl_qdma_sdf *sdf;
-+ struct fsl_qdma_ddf *ddf;
-+ struct fsl_qdma_sg *sg_block, *temp;
-+ struct scatterlist *sg;
-+ u64 total_src_len = 0;
-+ u64 total_dst_len = 0;
-+ u32 i;
-+
-+ ccdf = (struct fsl_qdma_ccdf *)fsl_comp->virt_addr;
-+ csgf_desc = (struct fsl_qdma_csgf *)fsl_comp->virt_addr + 1;
-+ csgf_src = (struct fsl_qdma_csgf *)fsl_comp->virt_addr + 2;
-+ csgf_dest = (struct fsl_qdma_csgf *)fsl_comp->virt_addr + 3;
-+ sdf = (struct fsl_qdma_sdf *)fsl_comp->virt_addr + 4;
-+ ddf = (struct fsl_qdma_ddf *)fsl_comp->virt_addr + 5;
-+ memset(fsl_comp->virt_addr, 0, FSL_QDMA_BASE_BUFFER_SIZE);
-+ /* Head Command Descriptor(Frame Descriptor) */
-+ qdma_ccdf_addr_set64(ccdf, fsl_comp->bus_addr + 16);
-+ qdma_ccdf_set_format(ccdf, qdma_ccdf_get_offset(ccdf));
-+ /* Status notification is enqueued to status queue. */
-+ qdma_ccdf_set_ser(ccdf, qdma_ccdf_get_status(ccdf));
-+
-+ /* Compound Command Descriptor(Frame List Table) */
-+ qdma_csgf_addr_set64(csgf_desc, fsl_comp->bus_addr + 64);
-+ /* It must be 32 as Compound S/G Descriptor */
-+ qdma_csgf_set_len(csgf_desc, 32);
-+
-+ sg_block = fsl_comp->sg_block;
-+ qdma_csgf_addr_set64(csgf_src, sg_block->bus_addr);
-+ /* This entry link to the s/g entry. */
-+ qdma_csgf_set_e(csgf_src, 32);
-+
-+ temp = sg_block + fsl_comp->sg_block_src;
-+ qdma_csgf_addr_set64(csgf_dest, temp->bus_addr);
-+ /* This entry is the last entry. */
-+ qdma_csgf_set_f(csgf_dest, 32);
-+ /* This entry link to the s/g entry. */
-+ qdma_csgf_set_e(csgf_dest, 32);
-+
-+ for_each_sg(src_sg, sg, src_nents, i) {
-+ temp = sg_block + i / (FSL_QDMA_EXPECT_SG_ENTRY_NUM - 1);
-+ csgf_sg = (struct fsl_qdma_csgf *)temp->virt_addr +
-+ i % (FSL_QDMA_EXPECT_SG_ENTRY_NUM - 1);
-+ qdma_csgf_addr_set64(csgf_sg, sg_dma_address(sg));
-+ qdma_csgf_set_len(csgf_sg, sg_dma_len(sg));
-+ total_src_len += sg_dma_len(sg);
-+
-+ if (i == src_nents - 1)
-+ qdma_csgf_set_f(csgf_sg, sg_dma_len(sg));
-+ if (i % (FSL_QDMA_EXPECT_SG_ENTRY_NUM - 1) ==
-+ FSL_QDMA_EXPECT_SG_ENTRY_NUM - 2) {
-+ csgf_sg = (struct fsl_qdma_csgf *)temp->virt_addr +
-+ FSL_QDMA_EXPECT_SG_ENTRY_NUM - 1;
-+ temp = sg_block +
-+ i / (FSL_QDMA_EXPECT_SG_ENTRY_NUM - 1) + 1;
-+ qdma_csgf_addr_set64(csgf_sg, temp->bus_addr);
-+ qdma_csgf_set_e(csgf_sg, sg_dma_len(sg));
-+ }
-+ }
-+
-+ sg_block += fsl_comp->sg_block_src;
-+ for_each_sg(dst_sg, sg, dst_nents, i) {
-+ temp = sg_block + i / (FSL_QDMA_EXPECT_SG_ENTRY_NUM - 1);
-+ csgf_sg = (struct fsl_qdma_csgf *)temp->virt_addr +
-+ i % (FSL_QDMA_EXPECT_SG_ENTRY_NUM - 1);
-+ qdma_csgf_addr_set64(csgf_sg, sg_dma_address(sg));
-+ qdma_csgf_set_len(csgf_sg, sg_dma_len(sg));
-+ total_dst_len += sg_dma_len(sg);
-+
-+ if (i == dst_nents - 1)
-+ qdma_csgf_set_f(csgf_sg, sg_dma_len(sg));
-+ if (i % (FSL_QDMA_EXPECT_SG_ENTRY_NUM - 1) ==
-+ FSL_QDMA_EXPECT_SG_ENTRY_NUM - 2) {
-+ csgf_sg = (struct fsl_qdma_csgf *)temp->virt_addr +
-+ FSL_QDMA_EXPECT_SG_ENTRY_NUM - 1;
-+ temp = sg_block +
-+ i / (FSL_QDMA_EXPECT_SG_ENTRY_NUM - 1) + 1;
-+ qdma_csgf_addr_set64(csgf_sg, temp->bus_addr);
-+ qdma_csgf_set_e(csgf_sg, sg_dma_len(sg));
-+ }
-+ }
-+
-+ if (total_src_len != total_dst_len)
-+ dev_err(&fsl_comp->qchan->vchan.chan.dev->device,
-+ "The data length for src and dst isn't match.\n");
-+
-+ qdma_csgf_set_len(csgf_src, total_src_len);
-+ qdma_csgf_set_len(csgf_dest, total_dst_len);
-+
-+ /* Descriptor Buffer */
-+}
-+
-+/*
-+ * Prei-request full command descriptor for enqueue.
-+ */
-+static int fsl_qdma_pre_request_enqueue_desc(struct fsl_qdma_queue *queue)
-+{
-+ struct fsl_qdma_comp *comp_temp;
-+ int i;
-+
-+ for (i = 0; i < queue->n_cq; i++) {
-+ comp_temp = kzalloc(sizeof(*comp_temp), GFP_KERNEL);
-+ if (!comp_temp)
-+ return -1;
-+ comp_temp->virt_addr = dma_pool_alloc(queue->comp_pool,
-+ GFP_NOWAIT,
-+ &comp_temp->bus_addr);
-+ if (!comp_temp->virt_addr)
-+ return -1;
-+ list_add_tail(&comp_temp->list, &queue->comp_free);
-+ }
-+ return 0;
-+}
-+
-+/*
-+ * Request a command descriptor for enqueue.
-+ */
-+static struct fsl_qdma_comp *fsl_qdma_request_enqueue_desc(
-+ struct fsl_qdma_chan *fsl_chan,
-+ unsigned int dst_nents,
-+ unsigned int src_nents)
-+{
-+ struct fsl_qdma_comp *comp_temp;
-+ struct fsl_qdma_sg *sg_block;
-+ struct fsl_qdma_queue *queue = fsl_chan->queue;
-+ unsigned long flags;
-+ unsigned int dst_sg_entry_block, src_sg_entry_block, sg_entry_total, i;
-+
-+ spin_lock_irqsave(&queue->queue_lock, flags);
-+ if (list_empty(&queue->comp_free)) {
-+ spin_unlock_irqrestore(&queue->queue_lock, flags);
-+ comp_temp = kzalloc(sizeof(*comp_temp), GFP_KERNEL);
-+ if (!comp_temp)
-+ return NULL;
-+ comp_temp->virt_addr = dma_pool_alloc(queue->comp_pool,
-+ GFP_NOWAIT,
-+ &comp_temp->bus_addr);
-+ if (!comp_temp->virt_addr)
-+ return NULL;
-+ } else {
-+ comp_temp = list_first_entry(&queue->comp_free,
-+ struct fsl_qdma_comp,
-+ list);
-+ list_del(&comp_temp->list);
-+ spin_unlock_irqrestore(&queue->queue_lock, flags);
-+ }
-+
-+ if (dst_nents != 0)
-+ dst_sg_entry_block = dst_nents /
-+ (FSL_QDMA_EXPECT_SG_ENTRY_NUM - 1) + 1;
-+ else
-+ dst_sg_entry_block = 0;
-+
-+ if (src_nents != 0)
-+ src_sg_entry_block = src_nents /
-+ (FSL_QDMA_EXPECT_SG_ENTRY_NUM - 1) + 1;
-+ else
-+ src_sg_entry_block = 0;
-+
-+ sg_entry_total = dst_sg_entry_block + src_sg_entry_block;
-+ if (sg_entry_total) {
-+ sg_block = kzalloc(sizeof(*sg_block) *
-+ sg_entry_total,
-+ GFP_KERNEL);
-+ if (!sg_block)
-+ return NULL;
-+ comp_temp->sg_block = sg_block;
-+ for (i = 0; i < sg_entry_total; i++) {
-+ sg_block->virt_addr = dma_pool_alloc(queue->sg_pool,
-+ GFP_NOWAIT,
-+ &sg_block->bus_addr);
-+ memset(sg_block->virt_addr, 0,
-+ FSL_QDMA_EXPECT_SG_ENTRY_NUM * 16);
-+ sg_block++;
-+ }
-+ }
-+
-+ comp_temp->sg_block_src = src_sg_entry_block;
-+ comp_temp->sg_block_dst = dst_sg_entry_block;
-+ comp_temp->qchan = fsl_chan;
-+
-+ return comp_temp;
-+}
-+
-+static struct fsl_qdma_queue *fsl_qdma_alloc_queue_resources(
-+ struct platform_device *pdev,
-+ unsigned int queue_num)
-+{
-+ struct device_node *np = pdev->dev.of_node;
-+ struct fsl_qdma_queue *queue_head, *queue_temp;
-+ int ret, len, i;
-+ unsigned int queue_size[FSL_QDMA_QUEUE_MAX];
-+
-+ if (queue_num > FSL_QDMA_QUEUE_MAX)
-+ queue_num = FSL_QDMA_QUEUE_MAX;
-+ len = sizeof(*queue_head) * queue_num;
-+ queue_head = devm_kzalloc(&pdev->dev, len, GFP_KERNEL);
-+ if (!queue_head)
-+ return NULL;
-+
-+ ret = of_property_read_u32_array(np, "queue-sizes", queue_size,
-+ queue_num);
-+ if (ret) {
-+ dev_err(&pdev->dev, "Can't get queue-sizes.\n");
-+ return NULL;
-+ }
-+
-+ for (i = 0; i < queue_num; i++) {
-+ if (queue_size[i] > FSL_QDMA_CIRCULAR_DESC_SIZE_MAX
-+ || queue_size[i] < FSL_QDMA_CIRCULAR_DESC_SIZE_MIN) {
-+ dev_err(&pdev->dev, "Get wrong queue-sizes.\n");
-+ return NULL;
-+ }
-+ queue_temp = queue_head + i;
-+ queue_temp->cq = dma_alloc_coherent(&pdev->dev,
-+ sizeof(struct fsl_qdma_ccdf) *
-+ queue_size[i],
-+ &queue_temp->bus_addr,
-+ GFP_KERNEL);
-+ if (!queue_temp->cq)
-+ return NULL;
-+ queue_temp->n_cq = queue_size[i];
-+ queue_temp->id = i;
-+ queue_temp->virt_head = queue_temp->cq;
-+ queue_temp->virt_tail = queue_temp->cq;
-+ /*
-+ * The dma pool for queue command buffer
-+ */
-+ queue_temp->comp_pool = dma_pool_create("comp_pool",
-+ &pdev->dev,
-+ FSL_QDMA_BASE_BUFFER_SIZE,
-+ 16, 0);
-+ if (!queue_temp->comp_pool) {
-+ dma_free_coherent(&pdev->dev,
-+ sizeof(struct fsl_qdma_ccdf) *
-+ queue_size[i],
-+ queue_temp->cq,
-+ queue_temp->bus_addr);
-+ return NULL;
-+ }
-+ /*
-+ * The dma pool for queue command buffer
-+ */
-+ queue_temp->sg_pool = dma_pool_create("sg_pool",
-+ &pdev->dev,
-+ FSL_QDMA_EXPECT_SG_ENTRY_NUM * 16,
-+ 64, 0);
-+ if (!queue_temp->sg_pool) {
-+ dma_free_coherent(&pdev->dev,
-+ sizeof(struct fsl_qdma_ccdf) *
-+ queue_size[i],
-+ queue_temp->cq,
-+ queue_temp->bus_addr);
-+ dma_pool_destroy(queue_temp->comp_pool);
-+ return NULL;
-+ }
-+ /*
-+ * List for queue command buffer
-+ */
-+ INIT_LIST_HEAD(&queue_temp->comp_used);
-+ INIT_LIST_HEAD(&queue_temp->comp_free);
-+ spin_lock_init(&queue_temp->queue_lock);
-+ }
-+
-+ return queue_head;
-+}
-+
-+static struct fsl_qdma_queue *fsl_qdma_prep_status_queue(
-+ struct platform_device *pdev)
-+{
-+ struct device_node *np = pdev->dev.of_node;
-+ struct fsl_qdma_queue *status_head;
-+ unsigned int status_size;
-+ int ret;
-+
-+ ret = of_property_read_u32(np, "status-sizes", &status_size);
-+ if (ret) {
-+ dev_err(&pdev->dev, "Can't get status-sizes.\n");
-+ return NULL;
-+ }
-+ if (status_size > FSL_QDMA_CIRCULAR_DESC_SIZE_MAX
-+ || status_size < FSL_QDMA_CIRCULAR_DESC_SIZE_MIN) {
-+ dev_err(&pdev->dev, "Get wrong status_size.\n");
-+ return NULL;
-+ }
-+ status_head = devm_kzalloc(&pdev->dev, sizeof(*status_head),
-+ GFP_KERNEL);
-+ if (!status_head)
-+ return NULL;
-+
-+ /*
-+ * Buffer for queue command
-+ */
-+ status_head->cq = dma_alloc_coherent(&pdev->dev,
-+ sizeof(struct fsl_qdma_ccdf) *
-+ status_size,
-+ &status_head->bus_addr,
-+ GFP_KERNEL);
-+ if (!status_head->cq)
-+ return NULL;
-+ status_head->n_cq = status_size;
-+ status_head->virt_head = status_head->cq;
-+ status_head->virt_tail = status_head->cq;
-+ status_head->comp_pool = NULL;
-+
-+ return status_head;
-+}
-+
-+static int fsl_qdma_halt(struct fsl_qdma_engine *fsl_qdma)
-+{
-+ void __iomem *ctrl = fsl_qdma->ctrl_base;
-+ void __iomem *block = fsl_qdma->block_base;
-+ int i, count = 5;
-+ u32 reg;
-+
-+ /* Disable the command queue and wait for idle state. */
-+ reg = qdma_readl(fsl_qdma, ctrl + FSL_QDMA_DMR);
-+ reg |= FSL_QDMA_DMR_DQD;
-+ qdma_writel(fsl_qdma, reg, ctrl + FSL_QDMA_DMR);
-+ for (i = 0; i < FSL_QDMA_QUEUE_NUM_MAX; i++)
-+ qdma_writel(fsl_qdma, 0, block + FSL_QDMA_BCQMR(i));
-+
-+ while (1) {
-+ reg = qdma_readl(fsl_qdma, ctrl + FSL_QDMA_DSR);
-+ if (!(reg & FSL_QDMA_DSR_DB))
-+ break;
-+ if (count-- < 0)
-+ return -EBUSY;
-+ udelay(100);
-+ }
-+
-+ /* Disable status queue. */
-+ qdma_writel(fsl_qdma, 0, block + FSL_QDMA_BSQMR);
-+
-+ /*
-+ * Clear the command queue interrupt detect register for all queues.
-+ */
-+ qdma_writel(fsl_qdma, 0xffffffff, block + FSL_QDMA_BCQIDR(0));
-+
-+ return 0;
-+}
-+
-+static int fsl_qdma_queue_transfer_complete(struct fsl_qdma_engine *fsl_qdma)
-+{
-+ struct fsl_qdma_queue *fsl_queue = fsl_qdma->queue;
-+ struct fsl_qdma_queue *fsl_status = fsl_qdma->status;
-+ struct fsl_qdma_queue *temp_queue;
-+ struct fsl_qdma_comp *fsl_comp;
-+ struct fsl_qdma_ccdf *status_addr;
-+ struct fsl_qdma_csgf *csgf_src;
-+ void __iomem *block = fsl_qdma->block_base;
-+ u32 reg, i;
-+ bool duplicate, duplicate_handle;
-+
-+ while (1) {
-+ duplicate = 0;
-+ duplicate_handle = 0;
-+ reg = qdma_readl(fsl_qdma, block + FSL_QDMA_BSQSR);
-+ if (reg & FSL_QDMA_BSQSR_QE)
-+ return 0;
-+ status_addr = fsl_status->virt_head;
-+ if (qdma_ccdf_get_queue(status_addr) == pre_queue &&
-+ qdma_ccdf_addr_get64(status_addr) == pre_addr)
-+ duplicate = 1;
-+ i = qdma_ccdf_get_queue(status_addr);
-+ pre_queue = qdma_ccdf_get_queue(status_addr);
-+ pre_addr = qdma_ccdf_addr_get64(status_addr);
-+ temp_queue = fsl_queue + i;
-+ spin_lock(&temp_queue->queue_lock);
-+ if (list_empty(&temp_queue->comp_used)) {
-+ if (duplicate)
-+ duplicate_handle = 1;
-+ else {
-+ spin_unlock(&temp_queue->queue_lock);
-+ return -1;
-+ }
-+ } else {
-+ fsl_comp = list_first_entry(&temp_queue->comp_used,
-+ struct fsl_qdma_comp,
-+ list);
-+ csgf_src = (struct fsl_qdma_csgf *)fsl_comp->virt_addr
-+ + 2;
-+ if (fsl_comp->bus_addr + 16 != pre_addr) {
-+ if (duplicate)
-+ duplicate_handle = 1;
-+ else {
-+ spin_unlock(&temp_queue->queue_lock);
-+ return -1;
-+ }
-+ }
-+ }
-+
-+ if (duplicate_handle) {
-+ reg = qdma_readl(fsl_qdma, block + FSL_QDMA_BSQMR);
-+ reg |= FSL_QDMA_BSQMR_DI;
-+ qdma_ccdf_addr_set64(status_addr, 0x0);
-+ fsl_status->virt_head++;
-+ if (fsl_status->virt_head == fsl_status->cq
-+ + fsl_status->n_cq)
-+ fsl_status->virt_head = fsl_status->cq;
-+ qdma_writel(fsl_qdma, reg, block + FSL_QDMA_BSQMR);
-+ spin_unlock(&temp_queue->queue_lock);
-+ continue;
-+ }
-+ list_del(&fsl_comp->list);
-+
-+ reg = qdma_readl(fsl_qdma, block + FSL_QDMA_BSQMR);
-+ reg |= FSL_QDMA_BSQMR_DI;
-+ qdma_ccdf_addr_set64(status_addr, 0x0);
-+ fsl_status->virt_head++;
-+ if (fsl_status->virt_head == fsl_status->cq + fsl_status->n_cq)
-+ fsl_status->virt_head = fsl_status->cq;
-+ qdma_writel(fsl_qdma, reg, block + FSL_QDMA_BSQMR);
-+ spin_unlock(&temp_queue->queue_lock);
-+
-+ spin_lock(&fsl_comp->qchan->vchan.lock);
-+ vchan_cookie_complete(&fsl_comp->vdesc);
-+ fsl_comp->qchan->status = DMA_COMPLETE;
-+ spin_unlock(&fsl_comp->qchan->vchan.lock);
-+ }
-+ return 0;
-+}
-+
-+static irqreturn_t fsl_qdma_error_handler(int irq, void *dev_id)
-+{
-+ struct fsl_qdma_engine *fsl_qdma = dev_id;
-+ unsigned int intr;
-+ void __iomem *status = fsl_qdma->status_base;
-+
-+ intr = qdma_readl(fsl_qdma, status + FSL_QDMA_DEDR);
-+
-+ if (intr)
-+ dev_err(fsl_qdma->dma_dev.dev, "DMA transaction error!\n");
-+
-+ qdma_writel(fsl_qdma, 0xffffffff, status + FSL_QDMA_DEDR);
-+ return IRQ_HANDLED;
-+}
-+
-+static irqreturn_t fsl_qdma_queue_handler(int irq, void *dev_id)
-+{
-+ struct fsl_qdma_engine *fsl_qdma = dev_id;
-+ unsigned int intr, reg;
-+ void __iomem *block = fsl_qdma->block_base;
-+ void __iomem *ctrl = fsl_qdma->ctrl_base;
-+
-+ intr = qdma_readl(fsl_qdma, block + FSL_QDMA_BCQIDR(0));
-+
-+ if ((intr & FSL_QDMA_CQIDR_SQT) != 0)
-+ intr = fsl_qdma_queue_transfer_complete(fsl_qdma);
-+
-+ if (intr != 0) {
-+ reg = qdma_readl(fsl_qdma, ctrl + FSL_QDMA_DMR);
-+ reg |= FSL_QDMA_DMR_DQD;
-+ qdma_writel(fsl_qdma, reg, ctrl + FSL_QDMA_DMR);
-+ qdma_writel(fsl_qdma, 0, block + FSL_QDMA_BCQIER(0));
-+ dev_err(fsl_qdma->dma_dev.dev, "QDMA: status err!\n");
-+ }
-+
-+ qdma_writel(fsl_qdma, 0xffffffff, block + FSL_QDMA_BCQIDR(0));
-+
-+ return IRQ_HANDLED;
-+}
-+
-+static int
-+fsl_qdma_irq_init(struct platform_device *pdev,
-+ struct fsl_qdma_engine *fsl_qdma)
-+{
-+ int ret;
-+
-+ fsl_qdma->error_irq = platform_get_irq_byname(pdev,
-+ "qdma-error");
-+ if (fsl_qdma->error_irq < 0) {
-+ dev_err(&pdev->dev, "Can't get qdma controller irq.\n");
-+ return fsl_qdma->error_irq;
-+ }
-+
-+ fsl_qdma->queue_irq = platform_get_irq_byname(pdev, "qdma-queue");
-+ if (fsl_qdma->queue_irq < 0) {
-+ dev_err(&pdev->dev, "Can't get qdma queue irq.\n");
-+ return fsl_qdma->queue_irq;
-+ }
-+
-+ ret = devm_request_irq(&pdev->dev, fsl_qdma->error_irq,
-+ fsl_qdma_error_handler, 0, "qDMA error", fsl_qdma);
-+ if (ret) {
-+ dev_err(&pdev->dev, "Can't register qDMA controller IRQ.\n");
-+ return ret;
-+ }
-+ ret = devm_request_irq(&pdev->dev, fsl_qdma->queue_irq,
-+ fsl_qdma_queue_handler, 0, "qDMA queue", fsl_qdma);
-+ if (ret) {
-+ dev_err(&pdev->dev, "Can't register qDMA queue IRQ.\n");
-+ return ret;
-+ }
-+
-+ return 0;
-+}
-+
-+static int fsl_qdma_reg_init(struct fsl_qdma_engine *fsl_qdma)
-+{
-+ struct fsl_qdma_queue *fsl_queue = fsl_qdma->queue;
-+ struct fsl_qdma_queue *temp;
-+ void __iomem *ctrl = fsl_qdma->ctrl_base;
-+ void __iomem *status = fsl_qdma->status_base;
-+ void __iomem *block = fsl_qdma->block_base;
-+ int i, ret;
-+ u32 reg;
-+
-+ /* Try to halt the qDMA engine first. */
-+ ret = fsl_qdma_halt(fsl_qdma);
-+ if (ret) {
-+ dev_err(fsl_qdma->dma_dev.dev, "DMA halt failed!");
-+ return ret;
-+ }
-+
-+ /*
-+ * Clear the command queue interrupt detect register for all queues.
-+ */
-+ qdma_writel(fsl_qdma, 0xffffffff, block + FSL_QDMA_BCQIDR(0));
-+
-+ for (i = 0; i < fsl_qdma->n_queues; i++) {
-+ temp = fsl_queue + i;
-+ /*
-+ * Initialize Command Queue registers to point to the first
-+ * command descriptor in memory.
-+ * Dequeue Pointer Address Registers
-+ * Enqueue Pointer Address Registers
-+ */
-+ qdma_writel(fsl_qdma, temp->bus_addr,
-+ block + FSL_QDMA_BCQDPA_SADDR(i));
-+ qdma_writel(fsl_qdma, temp->bus_addr,
-+ block + FSL_QDMA_BCQEPA_SADDR(i));
-+
-+ /* Initialize the queue mode. */
-+ reg = FSL_QDMA_BCQMR_EN;
-+ reg |= FSL_QDMA_BCQMR_CD_THLD(ilog2(temp->n_cq)-4);
-+ reg |= FSL_QDMA_BCQMR_CQ_SIZE(ilog2(temp->n_cq)-6);
-+ qdma_writel(fsl_qdma, reg, block + FSL_QDMA_BCQMR(i));
-+ }
-+
-+ /*
-+ * Workaround for erratum: ERR010812.
-+ * We must enable XOFF to avoid the enqueue rejection occurs.
-+ * Setting SQCCMR ENTER_WM to 0x20.
-+ */
-+ qdma_writel(fsl_qdma, FSL_QDMA_SQCCMR_ENTER_WM,
-+ block + FSL_QDMA_SQCCMR);
-+ /*
-+ * Initialize status queue registers to point to the first
-+ * command descriptor in memory.
-+ * Dequeue Pointer Address Registers
-+ * Enqueue Pointer Address Registers
-+ */
-+ qdma_writel(fsl_qdma, fsl_qdma->status->bus_addr,
-+ block + FSL_QDMA_SQEPAR);
-+ qdma_writel(fsl_qdma, fsl_qdma->status->bus_addr,
-+ block + FSL_QDMA_SQDPAR);
-+ /* Initialize status queue interrupt. */
-+ qdma_writel(fsl_qdma, FSL_QDMA_BCQIER_CQTIE,
-+ block + FSL_QDMA_BCQIER(0));
-+ qdma_writel(fsl_qdma, FSL_QDMA_BSQICR_ICEN | FSL_QDMA_BSQICR_ICST(5)
-+ | 0x8000,
-+ block + FSL_QDMA_BSQICR);
-+ qdma_writel(fsl_qdma, FSL_QDMA_CQIER_MEIE | FSL_QDMA_CQIER_TEIE,
-+ block + FSL_QDMA_CQIER);
-+ /* Initialize controller interrupt register. */
-+ qdma_writel(fsl_qdma, 0xffffffff, status + FSL_QDMA_DEDR);
-+ qdma_writel(fsl_qdma, 0xffffffff, status + FSL_QDMA_DEIER);
-+
-+ /* Initialize the status queue mode. */
-+ reg = FSL_QDMA_BSQMR_EN;
-+ reg |= FSL_QDMA_BSQMR_CQ_SIZE(ilog2(fsl_qdma->status->n_cq)-6);
-+ qdma_writel(fsl_qdma, reg, block + FSL_QDMA_BSQMR);
-+
-+ reg = qdma_readl(fsl_qdma, ctrl + FSL_QDMA_DMR);
-+ reg &= ~FSL_QDMA_DMR_DQD;
-+ qdma_writel(fsl_qdma, reg, ctrl + FSL_QDMA_DMR);
-+
-+ return 0;
-+}
-+
-+static struct dma_async_tx_descriptor *fsl_qdma_prep_dma_sg(
-+ struct dma_chan *chan,
-+ struct scatterlist *dst_sg, unsigned int dst_nents,
-+ struct scatterlist *src_sg, unsigned int src_nents,
-+ unsigned long flags)
-+{
-+ struct fsl_qdma_chan *fsl_chan = to_fsl_qdma_chan(chan);
-+ struct fsl_qdma_comp *fsl_comp;
-+
-+ fsl_comp = fsl_qdma_request_enqueue_desc(fsl_chan,
-+ dst_nents,
-+ src_nents);
-+ fsl_qdma_comp_fill_sg(fsl_comp, dst_sg, dst_nents, src_sg, src_nents);
-+
-+ return vchan_tx_prep(&fsl_chan->vchan, &fsl_comp->vdesc, flags);
-+}
-+
-+static struct dma_async_tx_descriptor *
-+fsl_qdma_prep_memcpy(struct dma_chan *chan, dma_addr_t dst,
-+ dma_addr_t src, size_t len, unsigned long flags)
-+{
-+ struct fsl_qdma_chan *fsl_chan = to_fsl_qdma_chan(chan);
-+ struct fsl_qdma_comp *fsl_comp;
-+
-+ fsl_comp = fsl_qdma_request_enqueue_desc(fsl_chan, 0, 0);
-+ fsl_qdma_comp_fill_memcpy(fsl_comp, dst, src, len);
-+
-+ return vchan_tx_prep(&fsl_chan->vchan, &fsl_comp->vdesc, flags);
-+}
-+
-+static void fsl_qdma_enqueue_desc(struct fsl_qdma_chan *fsl_chan)
-+{
-+ void __iomem *block = fsl_chan->qdma->block_base;
-+ struct fsl_qdma_queue *fsl_queue = fsl_chan->queue;
-+ struct fsl_qdma_comp *fsl_comp;
-+ struct virt_dma_desc *vdesc;
-+ u32 reg;
-+
-+ reg = qdma_readl(fsl_chan->qdma, block + FSL_QDMA_BCQSR(fsl_queue->id));
-+ if (reg & (FSL_QDMA_BCQSR_QF | FSL_QDMA_BCQSR_XOFF))
-+ return;
-+ vdesc = vchan_next_desc(&fsl_chan->vchan);
-+ if (!vdesc)
-+ return;
-+ list_del(&vdesc->node);
-+ fsl_comp = to_fsl_qdma_comp(vdesc);
-+
-+ memcpy(fsl_queue->virt_head++, fsl_comp->virt_addr, 16);
-+ if (fsl_queue->virt_head == fsl_queue->cq + fsl_queue->n_cq)
-+ fsl_queue->virt_head = fsl_queue->cq;
-+
-+ list_add_tail(&fsl_comp->list, &fsl_queue->comp_used);
-+ barrier();
-+ reg = qdma_readl(fsl_chan->qdma, block + FSL_QDMA_BCQMR(fsl_queue->id));
-+ reg |= FSL_QDMA_BCQMR_EI;
-+ qdma_writel(fsl_chan->qdma, reg, block + FSL_QDMA_BCQMR(fsl_queue->id));
-+ fsl_chan->status = DMA_IN_PROGRESS;
-+}
-+
-+static enum dma_status fsl_qdma_tx_status(struct dma_chan *chan,
-+ dma_cookie_t cookie, struct dma_tx_state *txstate)
-+{
-+ return dma_cookie_status(chan, cookie, txstate);
-+}
-+
-+static void fsl_qdma_free_desc(struct virt_dma_desc *vdesc)
-+{
-+ struct fsl_qdma_comp *fsl_comp;
-+ struct fsl_qdma_queue *fsl_queue;
-+ struct fsl_qdma_sg *sg_block;
-+ unsigned long flags;
-+ unsigned int i;
-+
-+ fsl_comp = to_fsl_qdma_comp(vdesc);
-+ fsl_queue = fsl_comp->qchan->queue;
-+
-+ if (fsl_comp->sg_block) {
-+ for (i = 0; i < fsl_comp->sg_block_src +
-+ fsl_comp->sg_block_dst; i++) {
-+ sg_block = fsl_comp->sg_block + i;
-+ dma_pool_free(fsl_queue->sg_pool,
-+ sg_block->virt_addr,
-+ sg_block->bus_addr);
-+ }
-+ kfree(fsl_comp->sg_block);
-+ }
-+
-+ spin_lock_irqsave(&fsl_queue->queue_lock, flags);
-+ list_add_tail(&fsl_comp->list, &fsl_queue->comp_free);
-+ spin_unlock_irqrestore(&fsl_queue->queue_lock, flags);
-+}
-+
-+static void fsl_qdma_issue_pending(struct dma_chan *chan)
-+{
-+ struct fsl_qdma_chan *fsl_chan = to_fsl_qdma_chan(chan);
-+ struct fsl_qdma_queue *fsl_queue = fsl_chan->queue;
-+ unsigned long flags;
-+
-+ spin_lock_irqsave(&fsl_queue->queue_lock, flags);
-+ spin_lock(&fsl_chan->vchan.lock);
-+ if (vchan_issue_pending(&fsl_chan->vchan))
-+ fsl_qdma_enqueue_desc(fsl_chan);
-+ spin_unlock(&fsl_chan->vchan.lock);
-+ spin_unlock_irqrestore(&fsl_queue->queue_lock, flags);
-+}
-+
-+static int fsl_qdma_probe(struct platform_device *pdev)
-+{
-+ struct device_node *np = pdev->dev.of_node;
-+ struct fsl_qdma_engine *fsl_qdma;
-+ struct fsl_qdma_chan *fsl_chan;
-+ struct resource *res;
-+ unsigned int len, chans, queues;
-+ int ret, i;
-+
-+ ret = of_property_read_u32(np, "channels", &chans);
-+ if (ret) {
-+ dev_err(&pdev->dev, "Can't get channels.\n");
-+ return ret;
-+ }
-+
-+ len = sizeof(*fsl_qdma) + sizeof(*fsl_chan) * chans;
-+ fsl_qdma = devm_kzalloc(&pdev->dev, len, GFP_KERNEL);
-+ if (!fsl_qdma)
-+ return -ENOMEM;
-+
-+ ret = of_property_read_u32(np, "queues", &queues);
-+ if (ret) {
-+ dev_err(&pdev->dev, "Can't get queues.\n");
-+ return ret;
-+ }
-+
-+ fsl_qdma->queue = fsl_qdma_alloc_queue_resources(pdev, queues);
-+ if (!fsl_qdma->queue)
-+ return -ENOMEM;
-+
-+ fsl_qdma->status = fsl_qdma_prep_status_queue(pdev);
-+ if (!fsl_qdma->status)
-+ return -ENOMEM;
-+
-+ fsl_qdma->n_chans = chans;
-+ fsl_qdma->n_queues = queues;
-+ mutex_init(&fsl_qdma->fsl_qdma_mutex);
-+
-+ res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
-+ fsl_qdma->ctrl_base = devm_ioremap_resource(&pdev->dev, res);
-+ if (IS_ERR(fsl_qdma->ctrl_base))
-+ return PTR_ERR(fsl_qdma->ctrl_base);
-+
-+ res = platform_get_resource(pdev, IORESOURCE_MEM, 1);
-+ fsl_qdma->status_base = devm_ioremap_resource(&pdev->dev, res);
-+ if (IS_ERR(fsl_qdma->status_base))
-+ return PTR_ERR(fsl_qdma->status_base);
-+
-+ res = platform_get_resource(pdev, IORESOURCE_MEM, 2);
-+ fsl_qdma->block_base = devm_ioremap_resource(&pdev->dev, res);
-+ if (IS_ERR(fsl_qdma->block_base))
-+ return PTR_ERR(fsl_qdma->block_base);
-+
-+ ret = fsl_qdma_irq_init(pdev, fsl_qdma);
-+ if (ret)
-+ return ret;
-+
-+ fsl_qdma->big_endian = of_property_read_bool(np, "big-endian");
-+ INIT_LIST_HEAD(&fsl_qdma->dma_dev.channels);
-+ for (i = 0; i < fsl_qdma->n_chans; i++) {
-+ struct fsl_qdma_chan *fsl_chan = &fsl_qdma->chans[i];
-+
-+ fsl_chan->qdma = fsl_qdma;
-+ fsl_chan->queue = fsl_qdma->queue + i % fsl_qdma->n_queues;
-+ fsl_chan->vchan.desc_free = fsl_qdma_free_desc;
-+ INIT_LIST_HEAD(&fsl_chan->qcomp);
-+ vchan_init(&fsl_chan->vchan, &fsl_qdma->dma_dev);
-+ }
-+ for (i = 0; i < fsl_qdma->n_queues; i++)
-+ fsl_qdma_pre_request_enqueue_desc(fsl_qdma->queue + i);
-+
-+ dma_cap_set(DMA_MEMCPY, fsl_qdma->dma_dev.cap_mask);
-+ dma_cap_set(DMA_SG, fsl_qdma->dma_dev.cap_mask);
-+
-+ fsl_qdma->dma_dev.dev = &pdev->dev;
-+ fsl_qdma->dma_dev.device_alloc_chan_resources
-+ = fsl_qdma_alloc_chan_resources;
-+ fsl_qdma->dma_dev.device_free_chan_resources
-+ = fsl_qdma_free_chan_resources;
-+ fsl_qdma->dma_dev.device_tx_status = fsl_qdma_tx_status;
-+ fsl_qdma->dma_dev.device_prep_dma_memcpy = fsl_qdma_prep_memcpy;
-+ fsl_qdma->dma_dev.device_prep_dma_sg = fsl_qdma_prep_dma_sg;
-+ fsl_qdma->dma_dev.device_issue_pending = fsl_qdma_issue_pending;
-+
-+ dma_set_mask(&pdev->dev, DMA_BIT_MASK(40));
-+
-+ platform_set_drvdata(pdev, fsl_qdma);
-+
-+ ret = dma_async_device_register(&fsl_qdma->dma_dev);
-+ if (ret) {
-+ dev_err(&pdev->dev, "Can't register Freescale qDMA engine.\n");
-+ return ret;
-+ }
-+
-+ ret = fsl_qdma_reg_init(fsl_qdma);
-+ if (ret) {
-+ dev_err(&pdev->dev, "Can't Initialize the qDMA engine.\n");
-+ return ret;
-+ }
-+
-+
-+ return 0;
-+}
-+
-+static int fsl_qdma_remove(struct platform_device *pdev)
-+{
-+ struct device_node *np = pdev->dev.of_node;
-+ struct fsl_qdma_engine *fsl_qdma = platform_get_drvdata(pdev);
-+ struct fsl_qdma_queue *queue_temp;
-+ struct fsl_qdma_queue *status = fsl_qdma->status;
-+ struct fsl_qdma_comp *comp_temp, *_comp_temp;
-+ int i;
-+
-+ of_dma_controller_free(np);
-+ dma_async_device_unregister(&fsl_qdma->dma_dev);
-+
-+ /* Free descriptor areas */
-+ for (i = 0; i < fsl_qdma->n_queues; i++) {
-+ queue_temp = fsl_qdma->queue + i;
-+ list_for_each_entry_safe(comp_temp, _comp_temp,
-+ &queue_temp->comp_used, list) {
-+ dma_pool_free(queue_temp->comp_pool,
-+ comp_temp->virt_addr,
-+ comp_temp->bus_addr);
-+ list_del(&comp_temp->list);
-+ kfree(comp_temp);
-+ }
-+ list_for_each_entry_safe(comp_temp, _comp_temp,
-+ &queue_temp->comp_free, list) {
-+ dma_pool_free(queue_temp->comp_pool,
-+ comp_temp->virt_addr,
-+ comp_temp->bus_addr);
-+ list_del(&comp_temp->list);
-+ kfree(comp_temp);
-+ }
-+ dma_free_coherent(&pdev->dev, sizeof(struct fsl_qdma_ccdf) *
-+ queue_temp->n_cq, queue_temp->cq,
-+ queue_temp->bus_addr);
-+ dma_pool_destroy(queue_temp->comp_pool);
-+ }
-+
-+ dma_free_coherent(&pdev->dev, sizeof(struct fsl_qdma_ccdf) *
-+ status->n_cq, status->cq, status->bus_addr);
-+ return 0;
-+}
-+
-+static const struct of_device_id fsl_qdma_dt_ids[] = {
-+ { .compatible = "fsl,ls1021a-qdma", },
-+ { /* sentinel */ }
-+};
-+MODULE_DEVICE_TABLE(of, fsl_qdma_dt_ids);
-+
-+static struct platform_driver fsl_qdma_driver = {
-+ .driver = {
-+ .name = "fsl-qdma",
-+ .owner = THIS_MODULE,
-+ .of_match_table = fsl_qdma_dt_ids,
-+ },
-+ .probe = fsl_qdma_probe,
-+ .remove = fsl_qdma_remove,
-+};
-+
-+static int __init fsl_qdma_init(void)
-+{
-+ return platform_driver_register(&fsl_qdma_driver);
-+}
-+subsys_initcall(fsl_qdma_init);
-+
-+static void __exit fsl_qdma_exit(void)
-+{
-+ platform_driver_unregister(&fsl_qdma_driver);
-+}
-+module_exit(fsl_qdma_exit);
-+
-+MODULE_ALIAS("platform:fsl-qdma");
-+MODULE_DESCRIPTION("Freescale qDMA engine driver");
-+MODULE_LICENSE("GPL v2");
diff --git a/target/linux/layerscape/patches-4.9/806-flextimer-support-layerscape.patch b/target/linux/layerscape/patches-4.9/806-flextimer-support-layerscape.patch
deleted file mode 100644
index 9826d822e9..0000000000
--- a/target/linux/layerscape/patches-4.9/806-flextimer-support-layerscape.patch
+++ /dev/null
@@ -1,404 +0,0 @@
-From b92e223750a07b28f175eae97d5ce3978df41be8 Mon Sep 17 00:00:00 2001
-From: Yangbo Lu <yangbo.lu@nxp.com>
-Date: Wed, 17 Jan 2018 15:32:05 +0800
-Subject: [PATCH 18/30] flextimer: support layerscape
-
-This is an integrated patch for layerscape flextimer support.
-
-Signed-off-by: Wang Dongsheng <dongsheng.wang@nxp.com>
-Signed-off-by: Meng Yi <meng.yi@nxp.com>
-Signed-off-by: Yangbo Lu <yangbo.lu@nxp.com>
----
- drivers/clocksource/fsl_ftm_timer.c | 8 +-
- drivers/soc/fsl/layerscape/ftm_alarm.c | 367 +++++++++++++++++++++++++++++++++
- 2 files changed, 371 insertions(+), 4 deletions(-)
- create mode 100644 drivers/soc/fsl/layerscape/ftm_alarm.c
-
---- a/drivers/clocksource/fsl_ftm_timer.c
-+++ b/drivers/clocksource/fsl_ftm_timer.c
-@@ -83,11 +83,11 @@ static inline void ftm_counter_disable(v
-
- static inline void ftm_irq_acknowledge(void __iomem *base)
- {
-- u32 val;
-+ unsigned int timeout = 100;
-
-- val = ftm_readl(base + FTM_SC);
-- val &= ~FTM_SC_TOF;
-- ftm_writel(val, base + FTM_SC);
-+ while ((FTM_SC_TOF & ftm_readl(base + FTM_SC)) && timeout--)
-+ ftm_writel(ftm_readl(base + FTM_SC) & (~FTM_SC_TOF),
-+ base + FTM_SC);
- }
-
- static inline void ftm_irq_enable(void __iomem *base)
---- /dev/null
-+++ b/drivers/soc/fsl/layerscape/ftm_alarm.c
-@@ -0,0 +1,367 @@
-+/*
-+ * Freescale FlexTimer Module (FTM) Alarm driver.
-+ *
-+ * Copyright 2014 Freescale Semiconductor, Inc.
-+ *
-+ * This program is free software; you can redistribute it and/or
-+ * modify it under the terms of the GNU General Public License
-+ * as published by the Free Software Foundation; either version 2
-+ * of the License, or (at your option) any later version.
-+ */
-+
-+#include <linux/device.h>
-+#include <linux/err.h>
-+#include <linux/interrupt.h>
-+#include <linux/io.h>
-+#include <linux/of_address.h>
-+#include <linux/of_irq.h>
-+#include <linux/platform_device.h>
-+#include <linux/of.h>
-+#include <linux/of_device.h>
-+#include <linux/libata.h>
-+#include <linux/module.h>
-+
-+#define FTM_SC 0x00
-+#define FTM_SC_CLK_SHIFT 3
-+#define FTM_SC_CLK_MASK (0x3 << FTM_SC_CLK_SHIFT)
-+#define FTM_SC_CLK(c) ((c) << FTM_SC_CLK_SHIFT)
-+#define FTM_SC_PS_MASK 0x7
-+#define FTM_SC_TOIE BIT(6)
-+#define FTM_SC_TOF BIT(7)
-+
-+#define FTM_SC_CLKS_FIXED_FREQ 0x02
-+
-+#define FTM_CNT 0x04
-+#define FTM_MOD 0x08
-+#define FTM_CNTIN 0x4C
-+
-+#define FIXED_FREQ_CLK 32000
-+#define MAX_FREQ_DIV (1 << FTM_SC_PS_MASK)
-+#define MAX_COUNT_VAL 0xffff
-+
-+static void __iomem *ftm1_base;
-+static void __iomem *rcpm_ftm_addr;
-+static u32 alarm_freq;
-+static bool big_endian;
-+
-+enum pmu_endian_type {
-+ BIG_ENDIAN,
-+ LITTLE_ENDIAN,
-+};
-+
-+struct rcpm_cfg {
-+ enum pmu_endian_type big_endian; /* Big/Little endian of PMU module */
-+ u32 flextimer_set_bit; /* FlexTimer1 is not powerdown during device LPM20 */
-+};
-+
-+static struct rcpm_cfg ls1012a_rcpm_cfg = {
-+ .big_endian = BIG_ENDIAN,
-+ .flextimer_set_bit = 0x20000,
-+};
-+
-+static struct rcpm_cfg ls1021a_rcpm_cfg = {
-+ .big_endian = BIG_ENDIAN,
-+ .flextimer_set_bit = 0x20000,
-+};
-+
-+static struct rcpm_cfg ls1043a_rcpm_cfg = {
-+ .big_endian = BIG_ENDIAN,
-+ .flextimer_set_bit = 0x20000,
-+};
-+
-+static struct rcpm_cfg ls1046a_rcpm_cfg = {
-+ .big_endian = BIG_ENDIAN,
-+ .flextimer_set_bit = 0x20000,
-+};
-+
-+static struct rcpm_cfg ls1088a_rcpm_cfg = {
-+ .big_endian = LITTLE_ENDIAN,
-+ .flextimer_set_bit = 0x4000,
-+};
-+
-+static struct rcpm_cfg ls208xa_rcpm_cfg = {
-+ .big_endian = LITTLE_ENDIAN,
-+ .flextimer_set_bit = 0x4000,
-+};
-+
-+static const struct of_device_id ippdexpcr_of_match[] = {
-+ { .compatible = "fsl,ls1012a-ftm", .data = &ls1012a_rcpm_cfg},
-+ { .compatible = "fsl,ls1021a-ftm", .data = &ls1021a_rcpm_cfg},
-+ { .compatible = "fsl,ls1043a-ftm", .data = &ls1043a_rcpm_cfg},
-+ { .compatible = "fsl,ls1046a-ftm", .data = &ls1046a_rcpm_cfg},
-+ { .compatible = "fsl,ls1088a-ftm", .data = &ls1088a_rcpm_cfg},
-+ { .compatible = "fsl,ls208xa-ftm", .data = &ls208xa_rcpm_cfg},
-+ {},
-+};
-+MODULE_DEVICE_TABLE(of, ippdexpcr_of_match);
-+
-+static inline u32 ftm_readl(void __iomem *addr)
-+{
-+ if (big_endian)
-+ return ioread32be(addr);
-+
-+ return ioread32(addr);
-+}
-+
-+static inline void ftm_writel(u32 val, void __iomem *addr)
-+{
-+ if (big_endian)
-+ iowrite32be(val, addr);
-+ else
-+ iowrite32(val, addr);
-+}
-+
-+static inline void ftm_counter_enable(void __iomem *base)
-+{
-+ u32 val;
-+
-+ /* select and enable counter clock source */
-+ val = ftm_readl(base + FTM_SC);
-+ val &= ~(FTM_SC_PS_MASK | FTM_SC_CLK_MASK);
-+ val |= (FTM_SC_PS_MASK | FTM_SC_CLK(FTM_SC_CLKS_FIXED_FREQ));
-+ ftm_writel(val, base + FTM_SC);
-+}
-+
-+static inline void ftm_counter_disable(void __iomem *base)
-+{
-+ u32 val;
-+
-+ /* disable counter clock source */
-+ val = ftm_readl(base + FTM_SC);
-+ val &= ~(FTM_SC_PS_MASK | FTM_SC_CLK_MASK);
-+ ftm_writel(val, base + FTM_SC);
-+}
-+
-+static inline void ftm_irq_acknowledge(void __iomem *base)
-+{
-+ unsigned int timeout = 100;
-+
-+ while ((FTM_SC_TOF & ftm_readl(base + FTM_SC)) && timeout--)
-+ ftm_writel(ftm_readl(base + FTM_SC) & (~FTM_SC_TOF),
-+ base + FTM_SC);
-+}
-+
-+static inline void ftm_irq_enable(void __iomem *base)
-+{
-+ u32 val;
-+
-+ val = ftm_readl(base + FTM_SC);
-+ val |= FTM_SC_TOIE;
-+ ftm_writel(val, base + FTM_SC);
-+}
-+
-+static inline void ftm_irq_disable(void __iomem *base)
-+{
-+ u32 val;
-+
-+ val = ftm_readl(base + FTM_SC);
-+ val &= ~FTM_SC_TOIE;
-+ ftm_writel(val, base + FTM_SC);
-+}
-+
-+static inline void ftm_reset_counter(void __iomem *base)
-+{
-+ /*
-+ * The CNT register contains the FTM counter value.
-+ * Reset clears the CNT register. Writing any value to COUNT
-+ * updates the counter with its initial value, CNTIN.
-+ */
-+ ftm_writel(0x00, base + FTM_CNT);
-+}
-+
-+static u32 time_to_cycle(unsigned long time)
-+{
-+ u32 cycle;
-+
-+ cycle = time * alarm_freq;
-+ if (cycle > MAX_COUNT_VAL) {
-+ pr_err("Out of alarm range.\n");
-+ cycle = 0;
-+ }
-+
-+ return cycle;
-+}
-+
-+static u32 cycle_to_time(u32 cycle)
-+{
-+ return cycle / alarm_freq + 1;
-+}
-+
-+static void ftm_clean_alarm(void)
-+{
-+ ftm_counter_disable(ftm1_base);
-+
-+ ftm_writel(0x00, ftm1_base + FTM_CNTIN);
-+ ftm_writel(~0U, ftm1_base + FTM_MOD);
-+
-+ ftm_reset_counter(ftm1_base);
-+}
-+
-+static int ftm_set_alarm(u64 cycle)
-+{
-+ ftm_irq_disable(ftm1_base);
-+
-+ /*
-+ * The counter increments until the value of MOD is reached,
-+ * at which point the counter is reloaded with the value of CNTIN.
-+ * The TOF (the overflow flag) bit is set when the FTM counter
-+ * changes from MOD to CNTIN. So we should using the cycle - 1.
-+ */
-+ ftm_writel(cycle - 1, ftm1_base + FTM_MOD);
-+
-+ ftm_counter_enable(ftm1_base);
-+
-+ ftm_irq_enable(ftm1_base);
-+
-+ return 0;
-+}
-+
-+static irqreturn_t ftm_alarm_interrupt(int irq, void *dev_id)
-+{
-+ ftm_irq_acknowledge(ftm1_base);
-+ ftm_irq_disable(ftm1_base);
-+ ftm_clean_alarm();
-+
-+ return IRQ_HANDLED;
-+}
-+
-+static ssize_t ftm_alarm_show(struct device *dev,
-+ struct device_attribute *attr,
-+ char *buf)
-+{
-+ u32 count, val;
-+
-+ count = ftm_readl(ftm1_base + FTM_MOD);
-+ val = ftm_readl(ftm1_base + FTM_CNT);
-+ val = (count & MAX_COUNT_VAL) - val;
-+ val = cycle_to_time(val);
-+
-+ return sprintf(buf, "%u\n", val);
-+}
-+
-+static ssize_t ftm_alarm_store(struct device *dev,
-+ struct device_attribute *attr,
-+ const char *buf, size_t count)
-+{
-+ u32 cycle;
-+ unsigned long time;
-+
-+ if (kstrtoul(buf, 0, &time))
-+ return -EINVAL;
-+
-+ ftm_clean_alarm();
-+
-+ cycle = time_to_cycle(time);
-+ if (!cycle)
-+ return -EINVAL;
-+
-+ ftm_set_alarm(cycle);
-+
-+ return count;
-+}
-+
-+static struct device_attribute ftm_alarm_attributes = __ATTR(ftm_alarm, 0644,
-+ ftm_alarm_show, ftm_alarm_store);
-+
-+static int ftm_alarm_probe(struct platform_device *pdev)
-+{
-+ struct device_node *np = pdev->dev.of_node;
-+ struct resource *r;
-+ int irq;
-+ int ret;
-+ struct rcpm_cfg *rcpm_cfg;
-+ u32 ippdexpcr, flextimer;
-+ const struct of_device_id *of_id;
-+ enum pmu_endian_type endian;
-+
-+ r = platform_get_resource(pdev, IORESOURCE_MEM, 0);
-+ if (!r)
-+ return -ENODEV;
-+
-+ ftm1_base = devm_ioremap_resource(&pdev->dev, r);
-+ if (IS_ERR(ftm1_base))
-+ return PTR_ERR(ftm1_base);
-+
-+ of_id = of_match_node(ippdexpcr_of_match, np);
-+ if (!of_id)
-+ return -ENODEV;
-+
-+ rcpm_cfg = devm_kzalloc(&pdev->dev, sizeof(*rcpm_cfg), GFP_KERNEL);
-+ if (!rcpm_cfg)
-+ return -ENOMEM;
-+
-+ rcpm_cfg = (struct rcpm_cfg*)of_id->data;
-+ endian = rcpm_cfg->big_endian;
-+ flextimer = rcpm_cfg->flextimer_set_bit;
-+
-+ r = platform_get_resource_byname(pdev, IORESOURCE_MEM, "FlexTimer1");
-+ if (r) {
-+ rcpm_ftm_addr = devm_ioremap_resource(&pdev->dev, r);
-+ if (IS_ERR(rcpm_ftm_addr))
-+ return PTR_ERR(rcpm_ftm_addr);
-+ if (endian == BIG_ENDIAN)
-+ ippdexpcr = ioread32be(rcpm_ftm_addr);
-+ else
-+ ippdexpcr = ioread32(rcpm_ftm_addr);
-+ ippdexpcr |= flextimer;
-+ if (endian == BIG_ENDIAN)
-+ iowrite32be(ippdexpcr, rcpm_ftm_addr);
-+ else
-+ iowrite32(ippdexpcr, rcpm_ftm_addr);
-+ }
-+
-+ irq = irq_of_parse_and_map(np, 0);
-+ if (irq <= 0) {
-+ pr_err("ftm: unable to get IRQ from DT, %d\n", irq);
-+ return -EINVAL;
-+ }
-+
-+ big_endian = of_property_read_bool(np, "big-endian");
-+
-+ ret = devm_request_irq(&pdev->dev, irq, ftm_alarm_interrupt,
-+ IRQF_NO_SUSPEND, dev_name(&pdev->dev), NULL);
-+ if (ret < 0) {
-+ dev_err(&pdev->dev, "failed to request irq\n");
-+ return ret;
-+ }
-+
-+ ret = device_create_file(&pdev->dev, &ftm_alarm_attributes);
-+ if (ret) {
-+ dev_err(&pdev->dev, "create sysfs fail.\n");
-+ return ret;
-+ }
-+
-+ alarm_freq = (u32)FIXED_FREQ_CLK / (u32)MAX_FREQ_DIV;
-+
-+ ftm_clean_alarm();
-+
-+ device_init_wakeup(&pdev->dev, true);
-+
-+ return ret;
-+}
-+
-+static const struct of_device_id ftm_alarm_match[] = {
-+ { .compatible = "fsl,ls1012a-ftm", },
-+ { .compatible = "fsl,ls1021a-ftm", },
-+ { .compatible = "fsl,ls1043a-ftm", },
-+ { .compatible = "fsl,ls1046a-ftm", },
-+ { .compatible = "fsl,ls1088a-ftm", },
-+ { .compatible = "fsl,ls208xa-ftm", },
-+ { .compatible = "fsl,ftm-timer", },
-+ { },
-+};
-+
-+static struct platform_driver ftm_alarm_driver = {
-+ .probe = ftm_alarm_probe,
-+ .driver = {
-+ .name = "ftm-alarm",
-+ .owner = THIS_MODULE,
-+ .of_match_table = ftm_alarm_match,
-+ },
-+};
-+
-+static int __init ftm_alarm_init(void)
-+{
-+ return platform_driver_register(&ftm_alarm_driver);
-+}
-+device_initcall(ftm_alarm_init);
diff --git a/target/linux/layerscape/patches-4.9/807-gpu-support-layerscape.patch b/target/linux/layerscape/patches-4.9/807-gpu-support-layerscape.patch
deleted file mode 100644
index 5aea9ea738..0000000000
--- a/target/linux/layerscape/patches-4.9/807-gpu-support-layerscape.patch
+++ /dev/null
@@ -1,58 +0,0 @@
-From 177f92a14d8177124f37db0fafc11182e2dcdd62 Mon Sep 17 00:00:00 2001
-From: Yangbo Lu <yangbo.lu@nxp.com>
-Date: Wed, 17 Jan 2018 15:33:05 +0800
-Subject: [PATCH 19/30] gpu: support layerscape
-
-This is an integrated patch for layerscape dcu support.
-
-Signed-off-by: Alison Wang <b18965@freescale.com>
-Signed-off-by: Yangbo Lu <yangbo.lu@nxp.com>
----
- drivers/gpu/drm/fsl-dcu/fsl_dcu_drm_drv.c | 15 ++++++++++++++-
- 1 file changed, 14 insertions(+), 1 deletion(-)
-
---- a/drivers/gpu/drm/fsl-dcu/fsl_dcu_drm_drv.c
-+++ b/drivers/gpu/drm/fsl-dcu/fsl_dcu_drm_drv.c
-@@ -225,7 +225,6 @@ static int fsl_dcu_drm_pm_suspend(struct
- if (!fsl_dev)
- return 0;
-
-- disable_irq(fsl_dev->irq);
- drm_kms_helper_poll_disable(fsl_dev->drm);
-
- console_lock();
-@@ -243,6 +242,8 @@ static int fsl_dcu_drm_pm_suspend(struct
- return PTR_ERR(fsl_dev->state);
- }
-
-+ disable_irq(fsl_dev->irq);
-+
- clk_disable_unprepare(fsl_dev->clk);
-
- return 0;
-@@ -262,6 +263,12 @@ static int fsl_dcu_drm_pm_resume(struct
- return ret;
- }
-
-+ ret = clk_prepare_enable(fsl_dev->pix_clk);
-+ if (ret < 0) {
-+ dev_err(dev, "failed to enable dcu pix clk\n");
-+ return ret;
-+ }
-+
- if (fsl_dev->tcon)
- fsl_tcon_bypass_enable(fsl_dev->tcon);
- fsl_dcu_drm_init_planes(fsl_dev->drm);
-@@ -388,6 +395,12 @@ static int fsl_dcu_drm_probe(struct plat
- goto disable_clk;
- }
-
-+ ret = clk_prepare_enable(fsl_dev->pix_clk);
-+ if (ret < 0) {
-+ dev_err(dev, "failed to enable dcu pix clk\n");
-+ return ret;
-+ }
-+
- fsl_dev->tcon = fsl_tcon_init(dev);
-
- drm = drm_dev_alloc(driver, dev);
diff --git a/target/linux/layerscape/patches-4.9/808-guts-support-layerscape.patch b/target/linux/layerscape/patches-4.9/808-guts-support-layerscape.patch
deleted file mode 100644
index 26412c5ebe..0000000000
--- a/target/linux/layerscape/patches-4.9/808-guts-support-layerscape.patch
+++ /dev/null
@@ -1,421 +0,0 @@
-From 45b0e1589b25ea3106a8c8d18bf653fde95baa9f Mon Sep 17 00:00:00 2001
-From: Yangbo Lu <yangbo.lu@nxp.com>
-Date: Wed, 17 Jan 2018 15:34:22 +0800
-Subject: [PATCH 20/30] guts: support layerscape
-
-This is an integrated patch for layerscape guts support.
-
-Signed-off-by: Roy Pledge <roy.pledge@nxp.com>
-Signed-off-by: Geert Uytterhoeven <geert+renesas@glider.be>
-Signed-off-by: Amrita Kumari <amrita.kumari@nxp.com>
-Signed-off-by: Yangbo Lu <yangbo.lu@nxp.com>
----
- drivers/soc/fsl/guts.c | 238 +++++++++++++++++++++++++++++++++++++++++++++++
- include/linux/fsl/guts.h | 125 +++++++++++++++----------
- 2 files changed, 315 insertions(+), 48 deletions(-)
- create mode 100644 drivers/soc/fsl/guts.c
-
---- /dev/null
-+++ b/drivers/soc/fsl/guts.c
-@@ -0,0 +1,238 @@
-+/*
-+ * Freescale QorIQ Platforms GUTS Driver
-+ *
-+ * Copyright (C) 2016 Freescale Semiconductor, Inc.
-+ *
-+ * This program is free software; you can redistribute it and/or modify
-+ * it under the terms of the GNU General Public License as published by
-+ * the Free Software Foundation; either version 2 of the License, or
-+ * (at your option) any later version.
-+ */
-+
-+#include <linux/io.h>
-+#include <linux/slab.h>
-+#include <linux/module.h>
-+#include <linux/of_fdt.h>
-+#include <linux/sys_soc.h>
-+#include <linux/of_address.h>
-+#include <linux/platform_device.h>
-+#include <linux/fsl/guts.h>
-+
-+struct guts {
-+ struct ccsr_guts __iomem *regs;
-+ bool little_endian;
-+};
-+
-+struct fsl_soc_die_attr {
-+ char *die;
-+ u32 svr;
-+ u32 mask;
-+};
-+
-+static struct guts *guts;
-+static struct soc_device_attribute soc_dev_attr;
-+static struct soc_device *soc_dev;
-+
-+
-+/* SoC die attribute definition for QorIQ platform */
-+static const struct fsl_soc_die_attr fsl_soc_die[] = {
-+ /*
-+ * Power Architecture-based SoCs T Series
-+ */
-+
-+ /* Die: T4240, SoC: T4240/T4160/T4080 */
-+ { .die = "T4240",
-+ .svr = 0x82400000,
-+ .mask = 0xfff00000,
-+ },
-+ /* Die: T1040, SoC: T1040/T1020/T1042/T1022 */
-+ { .die = "T1040",
-+ .svr = 0x85200000,
-+ .mask = 0xfff00000,
-+ },
-+ /* Die: T2080, SoC: T2080/T2081 */
-+ { .die = "T2080",
-+ .svr = 0x85300000,
-+ .mask = 0xfff00000,
-+ },
-+ /* Die: T1024, SoC: T1024/T1014/T1023/T1013 */
-+ { .die = "T1024",
-+ .svr = 0x85400000,
-+ .mask = 0xfff00000,
-+ },
-+
-+ /*
-+ * ARM-based SoCs LS Series
-+ */
-+
-+ /* Die: LS1043A, SoC: LS1043A/LS1023A */
-+ { .die = "LS1043A",
-+ .svr = 0x87920000,
-+ .mask = 0xffff0000,
-+ },
-+ /* Die: LS2080A, SoC: LS2080A/LS2040A/LS2085A */
-+ { .die = "LS2080A",
-+ .svr = 0x87010000,
-+ .mask = 0xff3f0000,
-+ },
-+ /* Die: LS1088A, SoC: LS1088A/LS1048A/LS1084A/LS1044A */
-+ { .die = "LS1088A",
-+ .svr = 0x87030000,
-+ .mask = 0xff3f0000,
-+ },
-+ /* Die: LS1012A, SoC: LS1012A */
-+ { .die = "LS1012A",
-+ .svr = 0x87040000,
-+ .mask = 0xffff0000,
-+ },
-+ /* Die: LS1046A, SoC: LS1046A/LS1026A */
-+ { .die = "LS1046A",
-+ .svr = 0x87070000,
-+ .mask = 0xffff0000,
-+ },
-+ /* Die: LS2088A, SoC: LS2088A/LS2048A/LS2084A/LS2044A */
-+ { .die = "LS2088A",
-+ .svr = 0x87090000,
-+ .mask = 0xff3f0000,
-+ },
-+ /* Die: LS1021A, SoC: LS1021A/LS1020A/LS1022A */
-+ { .die = "LS1021A",
-+ .svr = 0x87000000,
-+ .mask = 0xfff70000,
-+ },
-+ { },
-+};
-+
-+static const struct fsl_soc_die_attr *fsl_soc_die_match(
-+ u32 svr, const struct fsl_soc_die_attr *matches)
-+{
-+ while (matches->svr) {
-+ if (matches->svr == (svr & matches->mask))
-+ return matches;
-+ matches++;
-+ };
-+ return NULL;
-+}
-+
-+u32 fsl_guts_get_svr(void)
-+{
-+ u32 svr = 0;
-+
-+ if (!guts || !guts->regs)
-+ return svr;
-+
-+ if (guts->little_endian)
-+ svr = ioread32(&guts->regs->svr);
-+ else
-+ svr = ioread32be(&guts->regs->svr);
-+
-+ return svr;
-+}
-+EXPORT_SYMBOL(fsl_guts_get_svr);
-+
-+static int fsl_guts_probe(struct platform_device *pdev)
-+{
-+ struct device_node *np = pdev->dev.of_node;
-+ struct device *dev = &pdev->dev;
-+ struct resource *res;
-+ const struct fsl_soc_die_attr *soc_die;
-+ const char *machine;
-+ u32 svr;
-+
-+ /* Initialize guts */
-+ guts = devm_kzalloc(dev, sizeof(*guts), GFP_KERNEL);
-+ if (!guts)
-+ return -ENOMEM;
-+
-+ guts->little_endian = of_property_read_bool(np, "little-endian");
-+
-+ res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
-+ guts->regs = devm_ioremap_resource(dev, res);
-+ if (IS_ERR(guts->regs))
-+ return PTR_ERR(guts->regs);
-+
-+ /* Register soc device */
-+ machine = of_flat_dt_get_machine_name();
-+ if (machine)
-+ soc_dev_attr.machine = devm_kstrdup(dev, machine, GFP_KERNEL);
-+
-+ svr = fsl_guts_get_svr();
-+ soc_die = fsl_soc_die_match(svr, fsl_soc_die);
-+ if (soc_die) {
-+ soc_dev_attr.family = devm_kasprintf(dev, GFP_KERNEL,
-+ "QorIQ %s", soc_die->die);
-+ } else {
-+ soc_dev_attr.family = devm_kasprintf(dev, GFP_KERNEL, "QorIQ");
-+ }
-+ soc_dev_attr.soc_id = devm_kasprintf(dev, GFP_KERNEL,
-+ "svr:0x%08x", svr);
-+ soc_dev_attr.revision = devm_kasprintf(dev, GFP_KERNEL, "%d.%d",
-+ (svr >> 4) & 0xf, svr & 0xf);
-+
-+ soc_dev = soc_device_register(&soc_dev_attr);
-+ if (IS_ERR(soc_dev))
-+ return PTR_ERR(soc_dev);
-+
-+ pr_info("Machine: %s\n", soc_dev_attr.machine);
-+ pr_info("SoC family: %s\n", soc_dev_attr.family);
-+ pr_info("SoC ID: %s, Revision: %s\n",
-+ soc_dev_attr.soc_id, soc_dev_attr.revision);
-+ return 0;
-+}
-+
-+static int fsl_guts_remove(struct platform_device *dev)
-+{
-+ soc_device_unregister(soc_dev);
-+ return 0;
-+}
-+
-+/*
-+ * Table for matching compatible strings, for device tree
-+ * guts node, for Freescale QorIQ SOCs.
-+ */
-+static const struct of_device_id fsl_guts_of_match[] = {
-+ { .compatible = "fsl,qoriq-device-config-1.0", },
-+ { .compatible = "fsl,qoriq-device-config-2.0", },
-+ { .compatible = "fsl,p1010-guts", },
-+ { .compatible = "fsl,p1020-guts", },
-+ { .compatible = "fsl,p1021-guts", },
-+ { .compatible = "fsl,p1022-guts", },
-+ { .compatible = "fsl,p1023-guts", },
-+ { .compatible = "fsl,p2020-guts", },
-+ { .compatible = "fsl,bsc9131-guts", },
-+ { .compatible = "fsl,bsc9132-guts", },
-+ { .compatible = "fsl,mpc8536-guts", },
-+ { .compatible = "fsl,mpc8544-guts", },
-+ { .compatible = "fsl,mpc8548-guts", },
-+ { .compatible = "fsl,mpc8568-guts", },
-+ { .compatible = "fsl,mpc8569-guts", },
-+ { .compatible = "fsl,mpc8572-guts", },
-+ { .compatible = "fsl,ls1021a-dcfg", },
-+ { .compatible = "fsl,ls1043a-dcfg", },
-+ { .compatible = "fsl,ls1046a-dcfg", },
-+ { .compatible = "fsl,ls2080a-dcfg", },
-+ { .compatible = "fsl,ls1088a-dcfg", },
-+ {}
-+};
-+MODULE_DEVICE_TABLE(of, fsl_guts_of_match);
-+
-+static struct platform_driver fsl_guts_driver = {
-+ .driver = {
-+ .name = "fsl-guts",
-+ .of_match_table = fsl_guts_of_match,
-+ },
-+ .probe = fsl_guts_probe,
-+ .remove = fsl_guts_remove,
-+};
-+
-+static int __init fsl_guts_init(void)
-+{
-+ return platform_driver_register(&fsl_guts_driver);
-+}
-+core_initcall(fsl_guts_init);
-+
-+static void __exit fsl_guts_exit(void)
-+{
-+ platform_driver_unregister(&fsl_guts_driver);
-+}
-+module_exit(fsl_guts_exit);
---- a/include/linux/fsl/guts.h
-+++ b/include/linux/fsl/guts.h
-@@ -30,83 +30,112 @@
- * #ifdefs.
- */
- struct ccsr_guts {
-- __be32 porpllsr; /* 0x.0000 - POR PLL Ratio Status Register */
-- __be32 porbmsr; /* 0x.0004 - POR Boot Mode Status Register */
-- __be32 porimpscr; /* 0x.0008 - POR I/O Impedance Status and Control Register */
-- __be32 pordevsr; /* 0x.000c - POR I/O Device Status Register */
-- __be32 pordbgmsr; /* 0x.0010 - POR Debug Mode Status Register */
-- __be32 pordevsr2; /* 0x.0014 - POR device status register 2 */
-+ u32 porpllsr; /* 0x.0000 - POR PLL Ratio Status Register */
-+ u32 porbmsr; /* 0x.0004 - POR Boot Mode Status Register */
-+ u32 porimpscr; /* 0x.0008 - POR I/O Impedance Status and
-+ * Control Register
-+ */
-+ u32 pordevsr; /* 0x.000c - POR I/O Device Status Register */
-+ u32 pordbgmsr; /* 0x.0010 - POR Debug Mode Status Register */
-+ u32 pordevsr2; /* 0x.0014 - POR device status register 2 */
- u8 res018[0x20 - 0x18];
-- __be32 porcir; /* 0x.0020 - POR Configuration Information Register */
-+ u32 porcir; /* 0x.0020 - POR Configuration Information
-+ * Register
-+ */
- u8 res024[0x30 - 0x24];
-- __be32 gpiocr; /* 0x.0030 - GPIO Control Register */
-+ u32 gpiocr; /* 0x.0030 - GPIO Control Register */
- u8 res034[0x40 - 0x34];
-- __be32 gpoutdr; /* 0x.0040 - General-Purpose Output Data Register */
-+ u32 gpoutdr; /* 0x.0040 - General-Purpose Output Data
-+ * Register
-+ */
- u8 res044[0x50 - 0x44];
-- __be32 gpindr; /* 0x.0050 - General-Purpose Input Data Register */
-+ u32 gpindr; /* 0x.0050 - General-Purpose Input Data
-+ * Register
-+ */
- u8 res054[0x60 - 0x54];
-- __be32 pmuxcr; /* 0x.0060 - Alternate Function Signal Multiplex Control */
-- __be32 pmuxcr2; /* 0x.0064 - Alternate function signal multiplex control 2 */
-- __be32 dmuxcr; /* 0x.0068 - DMA Mux Control Register */
-+ u32 pmuxcr; /* 0x.0060 - Alternate Function Signal
-+ * Multiplex Control
-+ */
-+ u32 pmuxcr2; /* 0x.0064 - Alternate function signal
-+ * multiplex control 2
-+ */
-+ u32 dmuxcr; /* 0x.0068 - DMA Mux Control Register */
- u8 res06c[0x70 - 0x6c];
-- __be32 devdisr; /* 0x.0070 - Device Disable Control */
-+ u32 devdisr; /* 0x.0070 - Device Disable Control */
- #define CCSR_GUTS_DEVDISR_TB1 0x00001000
- #define CCSR_GUTS_DEVDISR_TB0 0x00004000
-- __be32 devdisr2; /* 0x.0074 - Device Disable Control 2 */
-+ u32 devdisr2; /* 0x.0074 - Device Disable Control 2 */
- u8 res078[0x7c - 0x78];
-- __be32 pmjcr; /* 0x.007c - 4 Power Management Jog Control Register */
-- __be32 powmgtcsr; /* 0x.0080 - Power Management Status and Control Register */
-- __be32 pmrccr; /* 0x.0084 - Power Management Reset Counter Configuration Register */
-- __be32 pmpdccr; /* 0x.0088 - Power Management Power Down Counter Configuration Register */
-- __be32 pmcdr; /* 0x.008c - 4Power management clock disable register */
-- __be32 mcpsumr; /* 0x.0090 - Machine Check Summary Register */
-- __be32 rstrscr; /* 0x.0094 - Reset Request Status and Control Register */
-- __be32 ectrstcr; /* 0x.0098 - Exception reset control register */
-- __be32 autorstsr; /* 0x.009c - Automatic reset status register */
-- __be32 pvr; /* 0x.00a0 - Processor Version Register */
-- __be32 svr; /* 0x.00a4 - System Version Register */
-+ u32 pmjcr; /* 0x.007c - 4 Power Management Jog Control
-+ * Register
-+ */
-+ u32 powmgtcsr; /* 0x.0080 - Power Management Status and
-+ * Control Register
-+ */
-+ u32 pmrccr; /* 0x.0084 - Power Management Reset Counter
-+ * Configuration Register
-+ */
-+ u32 pmpdccr; /* 0x.0088 - Power Management Power Down Counter
-+ * Configuration Register
-+ */
-+ u32 pmcdr; /* 0x.008c - 4Power management clock disable
-+ * register
-+ */
-+ u32 mcpsumr; /* 0x.0090 - Machine Check Summary Register */
-+ u32 rstrscr; /* 0x.0094 - Reset Request Status and
-+ * Control Register
-+ */
-+ u32 ectrstcr; /* 0x.0098 - Exception reset control register */
-+ u32 autorstsr; /* 0x.009c - Automatic reset status register */
-+ u32 pvr; /* 0x.00a0 - Processor Version Register */
-+ u32 svr; /* 0x.00a4 - System Version Register */
- u8 res0a8[0xb0 - 0xa8];
-- __be32 rstcr; /* 0x.00b0 - Reset Control Register */
-+ u32 rstcr; /* 0x.00b0 - Reset Control Register */
- u8 res0b4[0xc0 - 0xb4];
-- __be32 iovselsr; /* 0x.00c0 - I/O voltage select status register
-+ u32 iovselsr; /* 0x.00c0 - I/O voltage select status register
- Called 'elbcvselcr' on 86xx SOCs */
- u8 res0c4[0x100 - 0xc4];
-- __be32 rcwsr[16]; /* 0x.0100 - Reset Control Word Status registers
-+ u32 rcwsr[16]; /* 0x.0100 - Reset Control Word Status registers
- There are 16 registers */
- u8 res140[0x224 - 0x140];
-- __be32 iodelay1; /* 0x.0224 - IO delay control register 1 */
-- __be32 iodelay2; /* 0x.0228 - IO delay control register 2 */
-+ u32 iodelay1; /* 0x.0224 - IO delay control register 1 */
-+ u32 iodelay2; /* 0x.0228 - IO delay control register 2 */
- u8 res22c[0x604 - 0x22c];
-- __be32 pamubypenr; /* 0x.604 - PAMU bypass enable register */
-+ u32 pamubypenr; /* 0x.604 - PAMU bypass enable register */
- u8 res608[0x800 - 0x608];
-- __be32 clkdvdr; /* 0x.0800 - Clock Divide Register */
-+ u32 clkdvdr; /* 0x.0800 - Clock Divide Register */
- u8 res804[0x900 - 0x804];
-- __be32 ircr; /* 0x.0900 - Infrared Control Register */
-+ u32 ircr; /* 0x.0900 - Infrared Control Register */
- u8 res904[0x908 - 0x904];
-- __be32 dmacr; /* 0x.0908 - DMA Control Register */
-+ u32 dmacr; /* 0x.0908 - DMA Control Register */
- u8 res90c[0x914 - 0x90c];
-- __be32 elbccr; /* 0x.0914 - eLBC Control Register */
-+ u32 elbccr; /* 0x.0914 - eLBC Control Register */
- u8 res918[0xb20 - 0x918];
-- __be32 ddr1clkdr; /* 0x.0b20 - DDR1 Clock Disable Register */
-- __be32 ddr2clkdr; /* 0x.0b24 - DDR2 Clock Disable Register */
-- __be32 ddrclkdr; /* 0x.0b28 - DDR Clock Disable Register */
-+ u32 ddr1clkdr; /* 0x.0b20 - DDR1 Clock Disable Register */
-+ u32 ddr2clkdr; /* 0x.0b24 - DDR2 Clock Disable Register */
-+ u32 ddrclkdr; /* 0x.0b28 - DDR Clock Disable Register */
- u8 resb2c[0xe00 - 0xb2c];
-- __be32 clkocr; /* 0x.0e00 - Clock Out Select Register */
-+ u32 clkocr; /* 0x.0e00 - Clock Out Select Register */
- u8 rese04[0xe10 - 0xe04];
-- __be32 ddrdllcr; /* 0x.0e10 - DDR DLL Control Register */
-+ u32 ddrdllcr; /* 0x.0e10 - DDR DLL Control Register */
- u8 rese14[0xe20 - 0xe14];
-- __be32 lbcdllcr; /* 0x.0e20 - LBC DLL Control Register */
-- __be32 cpfor; /* 0x.0e24 - L2 charge pump fuse override register */
-+ u32 lbcdllcr; /* 0x.0e20 - LBC DLL Control Register */
-+ u32 cpfor; /* 0x.0e24 - L2 charge pump fuse override
-+ * register
-+ */
- u8 rese28[0xf04 - 0xe28];
-- __be32 srds1cr0; /* 0x.0f04 - SerDes1 Control Register 0 */
-- __be32 srds1cr1; /* 0x.0f08 - SerDes1 Control Register 0 */
-+ u32 srds1cr0; /* 0x.0f04 - SerDes1 Control Register 0 */
-+ u32 srds1cr1; /* 0x.0f08 - SerDes1 Control Register 0 */
- u8 resf0c[0xf2c - 0xf0c];
-- __be32 itcr; /* 0x.0f2c - Internal transaction control register */
-+ u32 itcr; /* 0x.0f2c - Internal transaction control
-+ * register
-+ */
- u8 resf30[0xf40 - 0xf30];
-- __be32 srds2cr0; /* 0x.0f40 - SerDes2 Control Register 0 */
-- __be32 srds2cr1; /* 0x.0f44 - SerDes2 Control Register 0 */
-+ u32 srds2cr0; /* 0x.0f40 - SerDes2 Control Register 0 */
-+ u32 srds2cr1; /* 0x.0f44 - SerDes2 Control Register 0 */
- } __attribute__ ((packed));
-
-+u32 fsl_guts_get_svr(void);
-
- /* Alternate function signal multiplex control */
- #define MPC85xx_PMUXCR_QE(x) (0x8000 >> (x))
diff --git a/target/linux/layerscape/patches-4.9/809-i2c-support-layerscape.patch b/target/linux/layerscape/patches-4.9/809-i2c-support-layerscape.patch
deleted file mode 100644
index ffb95cc342..0000000000
--- a/target/linux/layerscape/patches-4.9/809-i2c-support-layerscape.patch
+++ /dev/null
@@ -1,360 +0,0 @@
-From 659aa30c59fb188b533a7edcb9bd38ac007a2739 Mon Sep 17 00:00:00 2001
-From: Yangbo Lu <yangbo.lu@nxp.com>
-Date: Wed, 17 Jan 2018 15:35:11 +0800
-Subject: [PATCH 21/30] i2c: support layerscape
-
-This is an integrated patch for layerscape i2c support.
-
-Signed-off-by: Zhang Ying-22455 <ying.zhang22455@nxp.com>
-Signed-off-by: Priyanka Jain <Priyanka.Jain@freescale.com>
-Signed-off-by: Yangbo Lu <yangbo.lu@nxp.com>
----
- drivers/i2c/busses/i2c-imx.c | 195 +++++++++++++++++++++++++++++++++++-
- drivers/i2c/muxes/i2c-mux-pca954x.c | 43 ++++++++
- 2 files changed, 237 insertions(+), 1 deletion(-)
-
---- a/drivers/i2c/busses/i2c-imx.c
-+++ b/drivers/i2c/busses/i2c-imx.c
-@@ -53,6 +53,11 @@
- #include <linux/pm_runtime.h>
- #include <linux/sched.h>
- #include <linux/slab.h>
-+#include <linux/gpio.h>
-+#include <linux/of_address.h>
-+#include <linux/of.h>
-+#include <linux/of_device.h>
-+#include <linux/libata.h>
-
- /* This will be the driver name the kernel reports */
- #define DRIVER_NAME "imx-i2c"
-@@ -117,6 +122,54 @@
-
- #define I2C_PM_TIMEOUT 10 /* ms */
-
-+enum pinmux_endian_type {
-+ BIG_ENDIAN,
-+ LITTLE_ENDIAN,
-+};
-+
-+struct pinmux_cfg {
-+ enum pinmux_endian_type endian; /* endian of RCWPMUXCR0 */
-+ u32 pmuxcr_offset;
-+ u32 pmuxcr_set_bit; /* pin mux of RCWPMUXCR0 */
-+};
-+
-+static struct pinmux_cfg ls1012a_pinmux_cfg = {
-+ .endian = BIG_ENDIAN,
-+ .pmuxcr_offset = 0x430,
-+ .pmuxcr_set_bit = 0x10,
-+};
-+
-+static struct pinmux_cfg ls1043a_pinmux_cfg = {
-+ .endian = BIG_ENDIAN,
-+ .pmuxcr_offset = 0x40C,
-+ .pmuxcr_set_bit = 0x10,
-+};
-+
-+static struct pinmux_cfg ls1046a_pinmux_cfg = {
-+ .endian = BIG_ENDIAN,
-+ .pmuxcr_offset = 0x40C,
-+ .pmuxcr_set_bit = 0x80000000,
-+};
-+
-+static const struct of_device_id pinmux_of_match[] = {
-+ { .compatible = "fsl,ls1012a-vf610-i2c", .data = &ls1012a_pinmux_cfg},
-+ { .compatible = "fsl,ls1043a-vf610-i2c", .data = &ls1043a_pinmux_cfg},
-+ { .compatible = "fsl,ls1046a-vf610-i2c", .data = &ls1046a_pinmux_cfg},
-+ {},
-+};
-+MODULE_DEVICE_TABLE(of, pinmux_of_match);
-+
-+/* The SCFG, Supplemental Configuration Unit, provides SoC specific
-+ * configuration and status registers for the device. There is a
-+ * SDHC IO VSEL control register on SCFG for some platforms. It's
-+ * used to support SDHC IO voltage switching.
-+ */
-+static const struct of_device_id scfg_device_ids[] = {
-+ { .compatible = "fsl,ls1012a-scfg", },
-+ { .compatible = "fsl,ls1043a-scfg", },
-+ { .compatible = "fsl,ls1046a-scfg", },
-+ {}
-+};
- /*
- * sorted list of clock divider, register value pairs
- * taken from table 26-5, p.26-9, Freescale i.MX
-@@ -210,6 +263,12 @@ struct imx_i2c_struct {
- struct pinctrl_state *pinctrl_pins_gpio;
-
- struct imx_i2c_dma *dma;
-+ int layerscape_bus_recover;
-+ int gpio;
-+ int need_set_pmuxcr;
-+ int pmuxcr_set;
-+ int pmuxcr_endian;
-+ void __iomem *pmuxcr_addr;
- };
-
- static const struct imx_i2c_hwdata imx1_i2c_hwdata = {
-@@ -878,6 +937,78 @@ static int i2c_imx_read(struct imx_i2c_s
- return 0;
- }
-
-+/*
-+ * Based on the I2C specification, if the data line (SDA) is
-+ * stuck low, the master should send nine * clock pulses.
-+ * The I2C slave device that held the bus low should release it
-+ * sometime within * those nine clocks. Due to this erratum,
-+ * the I2C controller cannot generate nine clock pulses.
-+ */
-+static int i2c_imx_recovery_for_layerscape(struct imx_i2c_struct *i2c_imx)
-+{
-+ u32 pmuxcr = 0;
-+ int ret;
-+ unsigned int i, temp;
-+
-+ /* configure IICx_SCL/GPIO pin as a GPIO */
-+ if (i2c_imx->need_set_pmuxcr == 1) {
-+ pmuxcr = ioread32be(i2c_imx->pmuxcr_addr);
-+ if (i2c_imx->pmuxcr_endian == BIG_ENDIAN)
-+ iowrite32be(i2c_imx->pmuxcr_set|pmuxcr,
-+ i2c_imx->pmuxcr_addr);
-+ else
-+ iowrite32(i2c_imx->pmuxcr_set|pmuxcr,
-+ i2c_imx->pmuxcr_addr);
-+ }
-+
-+ ret = gpio_request(i2c_imx->gpio, i2c_imx->adapter.name);
-+ if (ret) {
-+ dev_err(&i2c_imx->adapter.dev,
-+ "can't get gpio: %d\n", ret);
-+ return ret;
-+ }
-+
-+ /* Configure GPIO pin as an output and open drain. */
-+ gpio_direction_output(i2c_imx->gpio, 1);
-+ udelay(10);
-+
-+ /* Write data to generate 9 pulses */
-+ for (i = 0; i < 9; i++) {
-+ gpio_set_value(i2c_imx->gpio, 1);
-+ udelay(10);
-+ gpio_set_value(i2c_imx->gpio, 0);
-+ udelay(10);
-+ }
-+ /* ensure that the last level sent is always high */
-+ gpio_set_value(i2c_imx->gpio, 1);
-+
-+ /*
-+ * Set I2Cx_IBCR = 0h00 to generate a STOP and then
-+ * set I2Cx_IBCR = 0h80 to reset
-+ */
-+ temp = imx_i2c_read_reg(i2c_imx, IMX_I2C_I2CR);
-+ temp &= ~(I2CR_MSTA | I2CR_MTX);
-+ imx_i2c_write_reg(temp, i2c_imx, IMX_I2C_I2CR);
-+
-+ /* Restore the saved value of the register SCFG_RCWPMUXCR0 */
-+ if (i2c_imx->need_set_pmuxcr == 1) {
-+ if (i2c_imx->pmuxcr_endian == BIG_ENDIAN)
-+ iowrite32be(pmuxcr, i2c_imx->pmuxcr_addr);
-+ else
-+ iowrite32(pmuxcr, i2c_imx->pmuxcr_addr);
-+ }
-+ /*
-+ * Set I2C_IBSR[IBAL] to clear the IBAL bit if-
-+ * I2C_IBSR[IBAL] = 1
-+ */
-+ temp = imx_i2c_read_reg(i2c_imx, IMX_I2C_I2SR);
-+ if (temp & I2SR_IAL) {
-+ temp &= ~I2SR_IAL;
-+ imx_i2c_write_reg(temp, i2c_imx, IMX_I2C_I2SR);
-+ }
-+ return 0;
-+}
-+
- static int i2c_imx_xfer(struct i2c_adapter *adapter,
- struct i2c_msg *msgs, int num)
- {
-@@ -888,6 +1019,19 @@ static int i2c_imx_xfer(struct i2c_adapt
-
- dev_dbg(&i2c_imx->adapter.dev, "<%s>\n", __func__);
-
-+ /*
-+ * workround for ERR010027: ensure that the I2C BUS is idle
-+ * before switching to master mode and attempting a Start cycle
-+ */
-+ result = i2c_imx_bus_busy(i2c_imx, 0);
-+ if (result) {
-+ /* timeout */
-+ if ((result == -ETIMEDOUT) && (i2c_imx->layerscape_bus_recover == 1))
-+ i2c_imx_recovery_for_layerscape(i2c_imx);
-+ else
-+ goto out;
-+ }
-+
- result = pm_runtime_get_sync(i2c_imx->adapter.dev.parent);
- if (result < 0)
- goto out;
-@@ -1030,6 +1174,50 @@ static int i2c_imx_init_recovery_info(st
- return 0;
- }
-
-+/*
-+ * switch SCL and SDA to their GPIO function and do some bitbanging
-+ * for bus recovery.
-+ * There are platforms such as Layerscape that don't support pinctrl, so add
-+ * workaround for layerscape, it has no effect for other platforms.
-+ */
-+static int i2c_imx_init_recovery_for_layerscape(
-+ struct imx_i2c_struct *i2c_imx,
-+ struct platform_device *pdev)
-+{
-+ const struct of_device_id *of_id;
-+ struct device_node *np = pdev->dev.of_node;
-+ struct pinmux_cfg *pinmux_cfg;
-+ struct device_node *scfg_node;
-+ void __iomem *scfg_base = NULL;
-+
-+ i2c_imx->gpio = of_get_named_gpio(np, "fsl-scl-gpio", 0);
-+ if (!gpio_is_valid(i2c_imx->gpio)) {
-+ dev_info(&pdev->dev, "fsl-scl-gpio not found\n");
-+ return 0;
-+ }
-+ pinmux_cfg = devm_kzalloc(&pdev->dev, sizeof(*pinmux_cfg), GFP_KERNEL);
-+ if (!pinmux_cfg)
-+ return -ENOMEM;
-+
-+ i2c_imx->need_set_pmuxcr = 0;
-+ of_id = of_match_node(pinmux_of_match, np);
-+ if (of_id) {
-+ pinmux_cfg = (struct pinmux_cfg *)of_id->data;
-+ i2c_imx->pmuxcr_endian = pinmux_cfg->endian;
-+ i2c_imx->pmuxcr_set = pinmux_cfg->pmuxcr_set_bit;
-+ scfg_node = of_find_matching_node(NULL, scfg_device_ids);
-+ if (scfg_node) {
-+ scfg_base = of_iomap(scfg_node, 0);
-+ if (scfg_base) {
-+ i2c_imx->pmuxcr_addr = scfg_base + pinmux_cfg->pmuxcr_offset;
-+ i2c_imx->need_set_pmuxcr = 1;
-+ }
-+ }
-+ }
-+ i2c_imx->layerscape_bus_recover = 1;
-+ return 0;
-+}
-+
- static u32 i2c_imx_func(struct i2c_adapter *adapter)
- {
- return I2C_FUNC_I2C | I2C_FUNC_SMBUS_EMUL
-@@ -1085,6 +1273,11 @@ static int i2c_imx_probe(struct platform
- i2c_imx->adapter.dev.of_node = pdev->dev.of_node;
- i2c_imx->base = base;
-
-+ /* Init optional bus recovery for layerscape */
-+ ret = i2c_imx_init_recovery_for_layerscape(i2c_imx, pdev);
-+ if (ret)
-+ return ret;
-+
- /* Get I2C clock */
- i2c_imx->clk = devm_clk_get(&pdev->dev, NULL);
- if (IS_ERR(i2c_imx->clk)) {
-@@ -1099,7 +1292,7 @@ static int i2c_imx_probe(struct platform
- }
-
- /* Request IRQ */
-- ret = devm_request_irq(&pdev->dev, irq, i2c_imx_isr, 0,
-+ ret = devm_request_irq(&pdev->dev, irq, i2c_imx_isr, IRQF_SHARED,
- pdev->name, i2c_imx);
- if (ret) {
- dev_err(&pdev->dev, "can't claim irq %d\n", irq);
---- a/drivers/i2c/muxes/i2c-mux-pca954x.c
-+++ b/drivers/i2c/muxes/i2c-mux-pca954x.c
-@@ -74,6 +74,7 @@ struct pca954x {
- u8 last_chan; /* last register value */
- u8 deselect;
- struct i2c_client *client;
-+ u8 disable_mux; /* do not disable mux if val not 0 */
- };
-
- /* Provide specs for the PCA954x types we know about */
-@@ -196,6 +197,13 @@ static int pca954x_deselect_mux(struct i
- if (!(data->deselect & (1 << chan)))
- return 0;
-
-+#ifdef CONFIG_ARCH_LAYERSCAPE
-+ if (data->disable_mux != 0)
-+ data->last_chan = data->chip->nchans;
-+ else
-+ data->last_chan = 0;
-+ return pca954x_reg_write(muxc->parent, client, data->disable_mux);
-+#endif
- /* Deselect active channel */
- data->last_chan = 0;
- return pca954x_reg_write(muxc->parent, client, data->last_chan);
-@@ -228,6 +236,28 @@ static int pca954x_probe(struct i2c_clie
- return -ENOMEM;
- data = i2c_mux_priv(muxc);
-
-+#ifdef CONFIG_ARCH_LAYERSCAPE
-+ /* The point here is that you must not disable a mux if there
-+ * are no pullups on the input or you mess up the I2C. This
-+ * needs to be put into the DTS really as the kernel cannot
-+ * know this otherwise.
-+ */
-+ match = of_match_device(of_match_ptr(pca954x_of_match), &client->dev);
-+ if (match)
-+ data->chip = of_device_get_match_data(&client->dev);
-+ else
-+ data->chip = &chips[id->driver_data];
-+
-+ data->disable_mux = of_node &&
-+ of_property_read_bool(of_node, "i2c-mux-never-disable") &&
-+ data->chip->muxtype == pca954x_ismux ?
-+ data->chip->enable : 0;
-+ /* force the first selection */
-+ if (data->disable_mux != 0)
-+ data->last_chan = data->chip->nchans;
-+ else
-+ data->last_chan = 0;
-+#endif
- i2c_set_clientdata(client, muxc);
- data->client = client;
-
-@@ -240,11 +270,16 @@ static int pca954x_probe(struct i2c_clie
- * that the mux is in fact present. This also
- * initializes the mux to disconnected state.
- */
-+#ifdef CONFIG_ARCH_LAYERSCAPE
-+ if (i2c_smbus_write_byte(client, data->disable_mux) < 0) {
-+#else
- if (i2c_smbus_write_byte(client, 0) < 0) {
-+#endif
- dev_warn(&client->dev, "probe failed\n");
- return -ENODEV;
- }
-
-+#ifndef CONFIG_ARCH_LAYERSCAPE
- match = of_match_device(of_match_ptr(pca954x_of_match), &client->dev);
- if (match)
- data->chip = of_device_get_match_data(&client->dev);
-@@ -252,6 +287,7 @@ static int pca954x_probe(struct i2c_clie
- data->chip = &chips[id->driver_data];
-
- data->last_chan = 0; /* force the first selection */
-+#endif
-
- idle_disconnect_dt = of_node &&
- of_property_read_bool(of_node, "i2c-mux-idle-disconnect");
-@@ -312,6 +348,13 @@ static int pca954x_resume(struct device
- struct i2c_mux_core *muxc = i2c_get_clientdata(client);
- struct pca954x *data = i2c_mux_priv(muxc);
-
-+#ifdef CONFIG_ARCH_LAYERSCAPE
-+ if (data->disable_mux != 0)
-+ data->last_chan = data->chip->nchans;
-+ else
-+ data->last_chan = 0;
-+ return i2c_smbus_write_byte(client, data->disable_mux);
-+#endif
- data->last_chan = 0;
- return i2c_smbus_write_byte(client, 0);
- }
diff --git a/target/linux/layerscape/patches-4.9/810-iommu-support-layerscape.patch b/target/linux/layerscape/patches-4.9/810-iommu-support-layerscape.patch
deleted file mode 100644
index 5dbeb0c45d..0000000000
--- a/target/linux/layerscape/patches-4.9/810-iommu-support-layerscape.patch
+++ /dev/null
@@ -1,1671 +0,0 @@
-From 89a1f0d7826df69d8e02268b97bc3da02e07203f Mon Sep 17 00:00:00 2001
-From: Yangbo Lu <yangbo.lu@nxp.com>
-Date: Thu, 5 Jul 2018 17:35:15 +0800
-Subject: [PATCH 22/32] iommu: support layerscape
-
-This is an integrated patch for layerscape smmu support.
-
-Signed-off-by: Eric Auger <eric.auger@redhat.com>
-Signed-off-by: Robin Murphy <robin.murphy@arm.com>
-Signed-off-by: Nipun Gupta <nipun.gupta@nxp.com>
-Signed-off-by: Sunil Goutham <sgoutham@cavium.com>
-Signed-off-by: Yangbo Lu <yangbo.lu@nxp.com>
----
- drivers/iommu/amd_iommu.c | 56 +++++---
- drivers/iommu/arm-smmu-v3.c | 111 +++++++++++-----
- drivers/iommu/arm-smmu.c | 100 ++++++++++++---
- drivers/iommu/dma-iommu.c | 242 +++++++++++++++++++++++++++++------
- drivers/iommu/intel-iommu.c | 92 ++++++++++---
- drivers/iommu/iommu.c | 240 ++++++++++++++++++++++++++++++++--
- drivers/iommu/mtk_iommu.c | 2 +
- drivers/iommu/mtk_iommu_v1.c | 2 +
- include/linux/dma-iommu.h | 11 ++
- include/linux/iommu.h | 57 +++++++--
- 10 files changed, 762 insertions(+), 151 deletions(-)
-
---- a/drivers/iommu/amd_iommu.c
-+++ b/drivers/iommu/amd_iommu.c
-@@ -379,6 +379,8 @@ static struct iommu_group *acpihid_devic
-
- if (!entry->group)
- entry->group = generic_device_group(dev);
-+ else
-+ iommu_group_ref_get(entry->group);
-
- return entry->group;
- }
-@@ -3166,9 +3168,10 @@ static bool amd_iommu_capable(enum iommu
- return false;
- }
-
--static void amd_iommu_get_dm_regions(struct device *dev,
-- struct list_head *head)
-+static void amd_iommu_get_resv_regions(struct device *dev,
-+ struct list_head *head)
- {
-+ struct iommu_resv_region *region;
- struct unity_map_entry *entry;
- int devid;
-
-@@ -3177,41 +3180,56 @@ static void amd_iommu_get_dm_regions(str
- return;
-
- list_for_each_entry(entry, &amd_iommu_unity_map, list) {
-- struct iommu_dm_region *region;
-+ size_t length;
-+ int prot = 0;
-
- if (devid < entry->devid_start || devid > entry->devid_end)
- continue;
-
-- region = kzalloc(sizeof(*region), GFP_KERNEL);
-+ length = entry->address_end - entry->address_start;
-+ if (entry->prot & IOMMU_PROT_IR)
-+ prot |= IOMMU_READ;
-+ if (entry->prot & IOMMU_PROT_IW)
-+ prot |= IOMMU_WRITE;
-+
-+ region = iommu_alloc_resv_region(entry->address_start,
-+ length, prot,
-+ IOMMU_RESV_DIRECT);
- if (!region) {
- pr_err("Out of memory allocating dm-regions for %s\n",
- dev_name(dev));
- return;
- }
--
-- region->start = entry->address_start;
-- region->length = entry->address_end - entry->address_start;
-- if (entry->prot & IOMMU_PROT_IR)
-- region->prot |= IOMMU_READ;
-- if (entry->prot & IOMMU_PROT_IW)
-- region->prot |= IOMMU_WRITE;
--
- list_add_tail(&region->list, head);
- }
-+
-+ region = iommu_alloc_resv_region(MSI_RANGE_START,
-+ MSI_RANGE_END - MSI_RANGE_START + 1,
-+ 0, IOMMU_RESV_MSI);
-+ if (!region)
-+ return;
-+ list_add_tail(&region->list, head);
-+
-+ region = iommu_alloc_resv_region(HT_RANGE_START,
-+ HT_RANGE_END - HT_RANGE_START + 1,
-+ 0, IOMMU_RESV_RESERVED);
-+ if (!region)
-+ return;
-+ list_add_tail(&region->list, head);
- }
-
--static void amd_iommu_put_dm_regions(struct device *dev,
-+static void amd_iommu_put_resv_regions(struct device *dev,
- struct list_head *head)
- {
-- struct iommu_dm_region *entry, *next;
-+ struct iommu_resv_region *entry, *next;
-
- list_for_each_entry_safe(entry, next, head, list)
- kfree(entry);
- }
-
--static void amd_iommu_apply_dm_region(struct device *dev,
-+static void amd_iommu_apply_resv_region(struct device *dev,
- struct iommu_domain *domain,
-- struct iommu_dm_region *region)
-+ struct iommu_resv_region *region)
- {
- struct dma_ops_domain *dma_dom = to_dma_ops_domain(to_pdomain(domain));
- unsigned long start, end;
-@@ -3235,9 +3253,9 @@ static const struct iommu_ops amd_iommu_
- .add_device = amd_iommu_add_device,
- .remove_device = amd_iommu_remove_device,
- .device_group = amd_iommu_device_group,
-- .get_dm_regions = amd_iommu_get_dm_regions,
-- .put_dm_regions = amd_iommu_put_dm_regions,
-- .apply_dm_region = amd_iommu_apply_dm_region,
-+ .get_resv_regions = amd_iommu_get_resv_regions,
-+ .put_resv_regions = amd_iommu_put_resv_regions,
-+ .apply_resv_region = amd_iommu_apply_resv_region,
- .pgsize_bitmap = AMD_IOMMU_PGSIZES,
- };
-
---- a/drivers/iommu/arm-smmu-v3.c
-+++ b/drivers/iommu/arm-smmu-v3.c
-@@ -410,6 +410,9 @@
- /* High-level queue structures */
- #define ARM_SMMU_POLL_TIMEOUT_US 100
-
-+#define MSI_IOVA_BASE 0x8000000
-+#define MSI_IOVA_LENGTH 0x100000
-+
- static bool disable_bypass;
- module_param_named(disable_bypass, disable_bypass, bool, S_IRUGO);
- MODULE_PARM_DESC(disable_bypass,
-@@ -552,9 +555,14 @@ struct arm_smmu_s2_cfg {
- };
-
- struct arm_smmu_strtab_ent {
-- bool valid;
--
-- bool bypass; /* Overrides s1/s2 config */
-+ /*
-+ * An STE is "assigned" if the master emitting the corresponding SID
-+ * is attached to a domain. The behaviour of an unassigned STE is
-+ * determined by the disable_bypass parameter, whereas an assigned
-+ * STE behaves according to s1_cfg/s2_cfg, which themselves are
-+ * configured according to the domain type.
-+ */
-+ bool assigned;
- struct arm_smmu_s1_cfg *s1_cfg;
- struct arm_smmu_s2_cfg *s2_cfg;
- };
-@@ -627,6 +635,7 @@ enum arm_smmu_domain_stage {
- ARM_SMMU_DOMAIN_S1 = 0,
- ARM_SMMU_DOMAIN_S2,
- ARM_SMMU_DOMAIN_NESTED,
-+ ARM_SMMU_DOMAIN_BYPASS,
- };
-
- struct arm_smmu_domain {
-@@ -1000,9 +1009,9 @@ static void arm_smmu_write_strtab_ent(st
- * This is hideously complicated, but we only really care about
- * three cases at the moment:
- *
-- * 1. Invalid (all zero) -> bypass (init)
-- * 2. Bypass -> translation (attach)
-- * 3. Translation -> bypass (detach)
-+ * 1. Invalid (all zero) -> bypass/fault (init)
-+ * 2. Bypass/fault -> translation/bypass (attach)
-+ * 3. Translation/bypass -> bypass/fault (detach)
- *
- * Given that we can't update the STE atomically and the SMMU
- * doesn't read the thing in a defined order, that leaves us
-@@ -1041,11 +1050,15 @@ static void arm_smmu_write_strtab_ent(st
- }
-
- /* Nuke the existing STE_0 value, as we're going to rewrite it */
-- val = ste->valid ? STRTAB_STE_0_V : 0;
-+ val = STRTAB_STE_0_V;
-+
-+ /* Bypass/fault */
-+ if (!ste->assigned || !(ste->s1_cfg || ste->s2_cfg)) {
-+ if (!ste->assigned && disable_bypass)
-+ val |= STRTAB_STE_0_CFG_ABORT;
-+ else
-+ val |= STRTAB_STE_0_CFG_BYPASS;
-
-- if (ste->bypass) {
-- val |= disable_bypass ? STRTAB_STE_0_CFG_ABORT
-- : STRTAB_STE_0_CFG_BYPASS;
- dst[0] = cpu_to_le64(val);
- dst[1] = cpu_to_le64(STRTAB_STE_1_SHCFG_INCOMING
- << STRTAB_STE_1_SHCFG_SHIFT);
-@@ -1108,10 +1121,7 @@ static void arm_smmu_write_strtab_ent(st
- static void arm_smmu_init_bypass_stes(u64 *strtab, unsigned int nent)
- {
- unsigned int i;
-- struct arm_smmu_strtab_ent ste = {
-- .valid = true,
-- .bypass = true,
-- };
-+ struct arm_smmu_strtab_ent ste = { .assigned = false };
-
- for (i = 0; i < nent; ++i) {
- arm_smmu_write_strtab_ent(NULL, -1, strtab, &ste);
-@@ -1365,8 +1375,6 @@ static bool arm_smmu_capable(enum iommu_
- switch (cap) {
- case IOMMU_CAP_CACHE_COHERENCY:
- return true;
-- case IOMMU_CAP_INTR_REMAP:
-- return true; /* MSIs are just memory writes */
- case IOMMU_CAP_NOEXEC:
- return true;
- default:
-@@ -1378,7 +1386,9 @@ static struct iommu_domain *arm_smmu_dom
- {
- struct arm_smmu_domain *smmu_domain;
-
-- if (type != IOMMU_DOMAIN_UNMANAGED && type != IOMMU_DOMAIN_DMA)
-+ if (type != IOMMU_DOMAIN_UNMANAGED &&
-+ type != IOMMU_DOMAIN_DMA &&
-+ type != IOMMU_DOMAIN_IDENTITY)
- return NULL;
-
- /*
-@@ -1509,6 +1519,11 @@ static int arm_smmu_domain_finalise(stru
- struct arm_smmu_domain *smmu_domain = to_smmu_domain(domain);
- struct arm_smmu_device *smmu = smmu_domain->smmu;
-
-+ if (domain->type == IOMMU_DOMAIN_IDENTITY) {
-+ smmu_domain->stage = ARM_SMMU_DOMAIN_BYPASS;
-+ return 0;
-+ }
-+
- /* Restrict the stage to what we can actually support */
- if (!(smmu->features & ARM_SMMU_FEAT_TRANS_S1))
- smmu_domain->stage = ARM_SMMU_DOMAIN_S2;
-@@ -1581,7 +1596,7 @@ static __le64 *arm_smmu_get_step_for_sid
- return step;
- }
-
--static int arm_smmu_install_ste_for_dev(struct iommu_fwspec *fwspec)
-+static void arm_smmu_install_ste_for_dev(struct iommu_fwspec *fwspec)
- {
- int i, j;
- struct arm_smmu_master_data *master = fwspec->iommu_priv;
-@@ -1600,17 +1615,14 @@ static int arm_smmu_install_ste_for_dev(
-
- arm_smmu_write_strtab_ent(smmu, sid, step, &master->ste);
- }
--
-- return 0;
- }
-
- static void arm_smmu_detach_dev(struct device *dev)
- {
- struct arm_smmu_master_data *master = dev->iommu_fwspec->iommu_priv;
-
-- master->ste.bypass = true;
-- if (arm_smmu_install_ste_for_dev(dev->iommu_fwspec) < 0)
-- dev_warn(dev, "failed to install bypass STE\n");
-+ master->ste.assigned = false;
-+ arm_smmu_install_ste_for_dev(dev->iommu_fwspec);
- }
-
- static int arm_smmu_attach_dev(struct iommu_domain *domain, struct device *dev)
-@@ -1629,7 +1641,7 @@ static int arm_smmu_attach_dev(struct io
- ste = &master->ste;
-
- /* Already attached to a different domain? */
-- if (!ste->bypass)
-+ if (ste->assigned)
- arm_smmu_detach_dev(dev);
-
- mutex_lock(&smmu_domain->init_mutex);
-@@ -1650,10 +1662,12 @@ static int arm_smmu_attach_dev(struct io
- goto out_unlock;
- }
-
-- ste->bypass = false;
-- ste->valid = true;
-+ ste->assigned = true;
-
-- if (smmu_domain->stage == ARM_SMMU_DOMAIN_S1) {
-+ if (smmu_domain->stage == ARM_SMMU_DOMAIN_BYPASS) {
-+ ste->s1_cfg = NULL;
-+ ste->s2_cfg = NULL;
-+ } else if (smmu_domain->stage == ARM_SMMU_DOMAIN_S1) {
- ste->s1_cfg = &smmu_domain->s1_cfg;
- ste->s2_cfg = NULL;
- arm_smmu_write_ctx_desc(smmu, ste->s1_cfg);
-@@ -1662,10 +1676,7 @@ static int arm_smmu_attach_dev(struct io
- ste->s2_cfg = &smmu_domain->s2_cfg;
- }
-
-- ret = arm_smmu_install_ste_for_dev(dev->iommu_fwspec);
-- if (ret < 0)
-- ste->valid = false;
--
-+ arm_smmu_install_ste_for_dev(dev->iommu_fwspec);
- out_unlock:
- mutex_unlock(&smmu_domain->init_mutex);
- return ret;
-@@ -1696,6 +1707,9 @@ arm_smmu_unmap(struct iommu_domain *doma
- struct arm_smmu_domain *smmu_domain = to_smmu_domain(domain);
- struct io_pgtable_ops *ops = smmu_domain->pgtbl_ops;
-
-+ if (domain->type == IOMMU_DOMAIN_IDENTITY)
-+ return iova;
-+
- if (!ops)
- return 0;
-
-@@ -1811,7 +1825,7 @@ static void arm_smmu_remove_device(struc
- return;
-
- master = fwspec->iommu_priv;
-- if (master && master->ste.valid)
-+ if (master && master->ste.assigned)
- arm_smmu_detach_dev(dev);
- iommu_group_remove_device(dev);
- kfree(master);
-@@ -1840,6 +1854,9 @@ static int arm_smmu_domain_get_attr(stru
- {
- struct arm_smmu_domain *smmu_domain = to_smmu_domain(domain);
-
-+ if (domain->type != IOMMU_DOMAIN_UNMANAGED)
-+ return -EINVAL;
-+
- switch (attr) {
- case DOMAIN_ATTR_NESTING:
- *(int *)data = (smmu_domain->stage == ARM_SMMU_DOMAIN_NESTED);
-@@ -1855,6 +1872,9 @@ static int arm_smmu_domain_set_attr(stru
- int ret = 0;
- struct arm_smmu_domain *smmu_domain = to_smmu_domain(domain);
-
-+ if (domain->type != IOMMU_DOMAIN_UNMANAGED)
-+ return -EINVAL;
-+
- mutex_lock(&smmu_domain->init_mutex);
-
- switch (attr) {
-@@ -1884,6 +1904,31 @@ static int arm_smmu_of_xlate(struct devi
- return iommu_fwspec_add_ids(dev, args->args, 1);
- }
-
-+static void arm_smmu_get_resv_regions(struct device *dev,
-+ struct list_head *head)
-+{
-+ struct iommu_resv_region *region;
-+ int prot = IOMMU_WRITE | IOMMU_NOEXEC | IOMMU_MMIO;
-+
-+ region = iommu_alloc_resv_region(MSI_IOVA_BASE, MSI_IOVA_LENGTH,
-+ prot, IOMMU_RESV_SW_MSI);
-+ if (!region)
-+ return;
-+
-+ list_add_tail(&region->list, head);
-+
-+ iommu_dma_get_resv_regions(dev, head);
-+}
-+
-+static void arm_smmu_put_resv_regions(struct device *dev,
-+ struct list_head *head)
-+{
-+ struct iommu_resv_region *entry, *next;
-+
-+ list_for_each_entry_safe(entry, next, head, list)
-+ kfree(entry);
-+}
-+
- static struct iommu_ops arm_smmu_ops = {
- .capable = arm_smmu_capable,
- .domain_alloc = arm_smmu_domain_alloc,
-@@ -1899,6 +1944,8 @@ static struct iommu_ops arm_smmu_ops = {
- .domain_get_attr = arm_smmu_domain_get_attr,
- .domain_set_attr = arm_smmu_domain_set_attr,
- .of_xlate = arm_smmu_of_xlate,
-+ .get_resv_regions = arm_smmu_get_resv_regions,
-+ .put_resv_regions = arm_smmu_put_resv_regions,
- .pgsize_bitmap = -1UL, /* Restricted during device attach */
- };
-
---- a/drivers/iommu/arm-smmu.c
-+++ b/drivers/iommu/arm-smmu.c
-@@ -49,6 +49,7 @@
- #include <linux/spinlock.h>
-
- #include <linux/amba/bus.h>
-+#include <linux/fsl/mc.h>
-
- #include "io-pgtable.h"
-
-@@ -247,6 +248,7 @@ enum arm_smmu_s2cr_privcfg {
- #define ARM_MMU500_ACTLR_CPRE (1 << 1)
-
- #define ARM_MMU500_ACR_CACHE_LOCK (1 << 26)
-+#define ARM_MMU500_ACR_SMTNMB_TLBEN (1 << 8)
-
- #define CB_PAR_F (1 << 0)
-
-@@ -278,6 +280,9 @@ enum arm_smmu_s2cr_privcfg {
-
- #define FSYNR0_WNR (1 << 4)
-
-+#define MSI_IOVA_BASE 0x8000000
-+#define MSI_IOVA_LENGTH 0x100000
-+
- static int force_stage;
- module_param(force_stage, int, S_IRUGO);
- MODULE_PARM_DESC(force_stage,
-@@ -401,6 +406,7 @@ enum arm_smmu_domain_stage {
- ARM_SMMU_DOMAIN_S1 = 0,
- ARM_SMMU_DOMAIN_S2,
- ARM_SMMU_DOMAIN_NESTED,
-+ ARM_SMMU_DOMAIN_BYPASS,
- };
-
- struct arm_smmu_domain {
-@@ -821,6 +827,12 @@ static int arm_smmu_init_domain_context(
- if (smmu_domain->smmu)
- goto out_unlock;
-
-+ if (domain->type == IOMMU_DOMAIN_IDENTITY) {
-+ smmu_domain->stage = ARM_SMMU_DOMAIN_BYPASS;
-+ smmu_domain->smmu = smmu;
-+ goto out_unlock;
-+ }
-+
- /*
- * Mapping the requested stage onto what we support is surprisingly
- * complicated, mainly because the spec allows S1+S2 SMMUs without
-@@ -981,7 +993,7 @@ static void arm_smmu_destroy_domain_cont
- void __iomem *cb_base;
- int irq;
-
-- if (!smmu)
-+ if (!smmu || domain->type == IOMMU_DOMAIN_IDENTITY)
- return;
-
- /*
-@@ -1004,7 +1016,9 @@ static struct iommu_domain *arm_smmu_dom
- {
- struct arm_smmu_domain *smmu_domain;
-
-- if (type != IOMMU_DOMAIN_UNMANAGED && type != IOMMU_DOMAIN_DMA)
-+ if (type != IOMMU_DOMAIN_UNMANAGED &&
-+ type != IOMMU_DOMAIN_DMA &&
-+ type != IOMMU_DOMAIN_IDENTITY)
- return NULL;
- /*
- * Allocate the domain and initialise some of its data structures.
-@@ -1202,10 +1216,15 @@ static int arm_smmu_domain_add_master(st
- {
- struct arm_smmu_device *smmu = smmu_domain->smmu;
- struct arm_smmu_s2cr *s2cr = smmu->s2crs;
-- enum arm_smmu_s2cr_type type = S2CR_TYPE_TRANS;
- u8 cbndx = smmu_domain->cfg.cbndx;
-+ enum arm_smmu_s2cr_type type;
- int i, idx;
-
-+ if (smmu_domain->stage == ARM_SMMU_DOMAIN_BYPASS)
-+ type = S2CR_TYPE_BYPASS;
-+ else
-+ type = S2CR_TYPE_TRANS;
-+
- for_each_cfg_sme(fwspec, i, idx) {
- if (type == s2cr[idx].type && cbndx == s2cr[idx].cbndx)
- continue;
-@@ -1343,6 +1362,9 @@ static phys_addr_t arm_smmu_iova_to_phys
- struct arm_smmu_domain *smmu_domain = to_smmu_domain(domain);
- struct io_pgtable_ops *ops= smmu_domain->pgtbl_ops;
-
-+ if (domain->type == IOMMU_DOMAIN_IDENTITY)
-+ return iova;
-+
- if (!ops)
- return 0;
-
-@@ -1368,8 +1390,6 @@ static bool arm_smmu_capable(enum iommu_
- * requests.
- */
- return true;
-- case IOMMU_CAP_INTR_REMAP:
-- return true; /* MSIs are just memory writes */
- case IOMMU_CAP_NOEXEC:
- return true;
- default:
-@@ -1478,10 +1498,12 @@ static struct iommu_group *arm_smmu_devi
- }
-
- if (group)
-- return group;
-+ return iommu_group_ref_get(group);
-
- if (dev_is_pci(dev))
- group = pci_device_group(dev);
-+ else if (dev_is_fsl_mc(dev))
-+ group = fsl_mc_device_group(dev);
- else
- group = generic_device_group(dev);
-
-@@ -1493,6 +1515,9 @@ static int arm_smmu_domain_get_attr(stru
- {
- struct arm_smmu_domain *smmu_domain = to_smmu_domain(domain);
-
-+ if (domain->type != IOMMU_DOMAIN_UNMANAGED)
-+ return -EINVAL;
-+
- switch (attr) {
- case DOMAIN_ATTR_NESTING:
- *(int *)data = (smmu_domain->stage == ARM_SMMU_DOMAIN_NESTED);
-@@ -1508,6 +1533,9 @@ static int arm_smmu_domain_set_attr(stru
- int ret = 0;
- struct arm_smmu_domain *smmu_domain = to_smmu_domain(domain);
-
-+ if (domain->type != IOMMU_DOMAIN_UNMANAGED)
-+ return -EINVAL;
-+
- mutex_lock(&smmu_domain->init_mutex);
-
- switch (attr) {
-@@ -1534,17 +1562,44 @@ out_unlock:
-
- static int arm_smmu_of_xlate(struct device *dev, struct of_phandle_args *args)
- {
-- u32 fwid = 0;
-+ u32 mask, fwid = 0;
-
- if (args->args_count > 0)
- fwid |= (u16)args->args[0];
-
- if (args->args_count > 1)
- fwid |= (u16)args->args[1] << SMR_MASK_SHIFT;
-+ else if (!of_property_read_u32(args->np, "stream-match-mask", &mask))
-+ fwid |= (u16)mask << SMR_MASK_SHIFT;
-
- return iommu_fwspec_add_ids(dev, &fwid, 1);
- }
-
-+static void arm_smmu_get_resv_regions(struct device *dev,
-+ struct list_head *head)
-+{
-+ struct iommu_resv_region *region;
-+ int prot = IOMMU_WRITE | IOMMU_NOEXEC | IOMMU_MMIO;
-+
-+ region = iommu_alloc_resv_region(MSI_IOVA_BASE, MSI_IOVA_LENGTH,
-+ prot, IOMMU_RESV_SW_MSI);
-+ if (!region)
-+ return;
-+
-+ list_add_tail(&region->list, head);
-+
-+ iommu_dma_get_resv_regions(dev, head);
-+}
-+
-+static void arm_smmu_put_resv_regions(struct device *dev,
-+ struct list_head *head)
-+{
-+ struct iommu_resv_region *entry, *next;
-+
-+ list_for_each_entry_safe(entry, next, head, list)
-+ kfree(entry);
-+}
-+
- static struct iommu_ops arm_smmu_ops = {
- .capable = arm_smmu_capable,
- .domain_alloc = arm_smmu_domain_alloc,
-@@ -1560,6 +1615,8 @@ static struct iommu_ops arm_smmu_ops = {
- .domain_get_attr = arm_smmu_domain_get_attr,
- .domain_set_attr = arm_smmu_domain_set_attr,
- .of_xlate = arm_smmu_of_xlate,
-+ .get_resv_regions = arm_smmu_get_resv_regions,
-+ .put_resv_regions = arm_smmu_put_resv_regions,
- .pgsize_bitmap = -1UL, /* Restricted during device attach */
- };
-
-@@ -1581,16 +1638,22 @@ static void arm_smmu_device_reset(struct
- for (i = 0; i < smmu->num_mapping_groups; ++i)
- arm_smmu_write_sme(smmu, i);
-
-- /*
-- * Before clearing ARM_MMU500_ACTLR_CPRE, need to
-- * clear CACHE_LOCK bit of ACR first. And, CACHE_LOCK
-- * bit is only present in MMU-500r2 onwards.
-- */
-- reg = readl_relaxed(gr0_base + ARM_SMMU_GR0_ID7);
-- major = (reg >> ID7_MAJOR_SHIFT) & ID7_MAJOR_MASK;
-- if ((smmu->model == ARM_MMU500) && (major >= 2)) {
-+ if (smmu->model == ARM_MMU500) {
-+ /*
-+ * Before clearing ARM_MMU500_ACTLR_CPRE, need to
-+ * clear CACHE_LOCK bit of ACR first. And, CACHE_LOCK
-+ * bit is only present in MMU-500r2 onwards.
-+ */
-+ reg = readl_relaxed(gr0_base + ARM_SMMU_GR0_ID7);
-+ major = (reg >> ID7_MAJOR_SHIFT) & ID7_MAJOR_MASK;
- reg = readl_relaxed(gr0_base + ARM_SMMU_GR0_sACR);
-- reg &= ~ARM_MMU500_ACR_CACHE_LOCK;
-+ if (major >= 2)
-+ reg &= ~ARM_MMU500_ACR_CACHE_LOCK;
-+ /*
-+ * Allow unmatched Stream IDs to allocate bypass
-+ * TLB entries for reduced latency.
-+ */
-+ reg |= ARM_MMU500_ACR_SMTNMB_TLBEN;
- writel_relaxed(reg, gr0_base + ARM_SMMU_GR0_sACR);
- }
-
-@@ -2024,6 +2087,11 @@ static int arm_smmu_device_dt_probe(stru
- bus_set_iommu(&pci_bus_type, &arm_smmu_ops);
- }
- #endif
-+#ifdef CONFIG_FSL_MC_BUS
-+ if (!iommu_present(&fsl_mc_bus_type))
-+ bus_set_iommu(&fsl_mc_bus_type, &arm_smmu_ops);
-+#endif
-+
- return 0;
- }
-
---- a/drivers/iommu/dma-iommu.c
-+++ b/drivers/iommu/dma-iommu.c
-@@ -37,15 +37,50 @@ struct iommu_dma_msi_page {
- phys_addr_t phys;
- };
-
-+enum iommu_dma_cookie_type {
-+ IOMMU_DMA_IOVA_COOKIE,
-+ IOMMU_DMA_MSI_COOKIE,
-+};
-+
- struct iommu_dma_cookie {
-- struct iova_domain iovad;
-- struct list_head msi_page_list;
-- spinlock_t msi_lock;
-+ enum iommu_dma_cookie_type type;
-+ union {
-+ /* Full allocator for IOMMU_DMA_IOVA_COOKIE */
-+ struct iova_domain iovad;
-+ /* Trivial linear page allocator for IOMMU_DMA_MSI_COOKIE */
-+ dma_addr_t msi_iova;
-+ };
-+ struct list_head msi_page_list;
-+ spinlock_t msi_lock;
- };
-
-+static inline size_t cookie_msi_granule(struct iommu_dma_cookie *cookie)
-+{
-+ if (cookie->type == IOMMU_DMA_IOVA_COOKIE)
-+ return cookie->iovad.granule;
-+ return PAGE_SIZE;
-+}
-+
- static inline struct iova_domain *cookie_iovad(struct iommu_domain *domain)
- {
-- return &((struct iommu_dma_cookie *)domain->iova_cookie)->iovad;
-+ struct iommu_dma_cookie *cookie = domain->iova_cookie;
-+
-+ if (cookie->type == IOMMU_DMA_IOVA_COOKIE)
-+ return &cookie->iovad;
-+ return NULL;
-+}
-+
-+static struct iommu_dma_cookie *cookie_alloc(enum iommu_dma_cookie_type type)
-+{
-+ struct iommu_dma_cookie *cookie;
-+
-+ cookie = kzalloc(sizeof(*cookie), GFP_KERNEL);
-+ if (cookie) {
-+ spin_lock_init(&cookie->msi_lock);
-+ INIT_LIST_HEAD(&cookie->msi_page_list);
-+ cookie->type = type;
-+ }
-+ return cookie;
- }
-
- int iommu_dma_init(void)
-@@ -62,25 +97,53 @@ int iommu_dma_init(void)
- */
- int iommu_get_dma_cookie(struct iommu_domain *domain)
- {
-+ if (domain->iova_cookie)
-+ return -EEXIST;
-+
-+ domain->iova_cookie = cookie_alloc(IOMMU_DMA_IOVA_COOKIE);
-+ if (!domain->iova_cookie)
-+ return -ENOMEM;
-+
-+ return 0;
-+}
-+EXPORT_SYMBOL(iommu_get_dma_cookie);
-+
-+/**
-+ * iommu_get_msi_cookie - Acquire just MSI remapping resources
-+ * @domain: IOMMU domain to prepare
-+ * @base: Start address of IOVA region for MSI mappings
-+ *
-+ * Users who manage their own IOVA allocation and do not want DMA API support,
-+ * but would still like to take advantage of automatic MSI remapping, can use
-+ * this to initialise their own domain appropriately. Users should reserve a
-+ * contiguous IOVA region, starting at @base, large enough to accommodate the
-+ * number of PAGE_SIZE mappings necessary to cover every MSI doorbell address
-+ * used by the devices attached to @domain.
-+ */
-+int iommu_get_msi_cookie(struct iommu_domain *domain, dma_addr_t base)
-+{
- struct iommu_dma_cookie *cookie;
-
-+ if (domain->type != IOMMU_DOMAIN_UNMANAGED)
-+ return -EINVAL;
-+
- if (domain->iova_cookie)
- return -EEXIST;
-
-- cookie = kzalloc(sizeof(*cookie), GFP_KERNEL);
-+ cookie = cookie_alloc(IOMMU_DMA_MSI_COOKIE);
- if (!cookie)
- return -ENOMEM;
-
-- spin_lock_init(&cookie->msi_lock);
-- INIT_LIST_HEAD(&cookie->msi_page_list);
-+ cookie->msi_iova = base;
- domain->iova_cookie = cookie;
- return 0;
- }
--EXPORT_SYMBOL(iommu_get_dma_cookie);
-+EXPORT_SYMBOL(iommu_get_msi_cookie);
-
- /**
- * iommu_put_dma_cookie - Release a domain's DMA mapping resources
-- * @domain: IOMMU domain previously prepared by iommu_get_dma_cookie()
-+ * @domain: IOMMU domain previously prepared by iommu_get_dma_cookie() or
-+ * iommu_get_msi_cookie()
- *
- * IOMMU drivers should normally call this from their domain_free callback.
- */
-@@ -92,7 +155,7 @@ void iommu_put_dma_cookie(struct iommu_d
- if (!cookie)
- return;
-
-- if (cookie->iovad.granule)
-+ if (cookie->type == IOMMU_DMA_IOVA_COOKIE && cookie->iovad.granule)
- put_iova_domain(&cookie->iovad);
-
- list_for_each_entry_safe(msi, tmp, &cookie->msi_page_list, list) {
-@@ -104,21 +167,99 @@ void iommu_put_dma_cookie(struct iommu_d
- }
- EXPORT_SYMBOL(iommu_put_dma_cookie);
-
--static void iova_reserve_pci_windows(struct pci_dev *dev,
-- struct iova_domain *iovad)
-+/**
-+ * iommu_dma_get_resv_regions - Reserved region driver helper
-+ * @dev: Device from iommu_get_resv_regions()
-+ * @list: Reserved region list from iommu_get_resv_regions()
-+ *
-+ * IOMMU drivers can use this to implement their .get_resv_regions callback
-+ * for general non-IOMMU-specific reservations. Currently, this covers host
-+ * bridge windows for PCI devices.
-+ */
-+void iommu_dma_get_resv_regions(struct device *dev, struct list_head *list)
- {
-- struct pci_host_bridge *bridge = pci_find_host_bridge(dev->bus);
-+ struct pci_host_bridge *bridge;
- struct resource_entry *window;
-- unsigned long lo, hi;
-
-+ if (!dev_is_pci(dev))
-+ return;
-+
-+ bridge = pci_find_host_bridge(to_pci_dev(dev)->bus);
- resource_list_for_each_entry(window, &bridge->windows) {
-+ struct iommu_resv_region *region;
-+ phys_addr_t start;
-+ size_t length;
-+
- if (resource_type(window->res) != IORESOURCE_MEM)
- continue;
-
-- lo = iova_pfn(iovad, window->res->start - window->offset);
-- hi = iova_pfn(iovad, window->res->end - window->offset);
-+ start = window->res->start - window->offset;
-+ length = window->res->end - window->res->start + 1;
-+ region = iommu_alloc_resv_region(start, length, 0,
-+ IOMMU_RESV_RESERVED);
-+ if (!region)
-+ return;
-+
-+ list_add_tail(&region->list, list);
-+ }
-+}
-+EXPORT_SYMBOL(iommu_dma_get_resv_regions);
-+
-+static int cookie_init_hw_msi_region(struct iommu_dma_cookie *cookie,
-+ phys_addr_t start, phys_addr_t end)
-+{
-+ struct iova_domain *iovad = &cookie->iovad;
-+ struct iommu_dma_msi_page *msi_page;
-+ int i, num_pages;
-+
-+ start -= iova_offset(iovad, start);
-+ num_pages = iova_align(iovad, end - start) >> iova_shift(iovad);
-+
-+ msi_page = kcalloc(num_pages, sizeof(*msi_page), GFP_KERNEL);
-+ if (!msi_page)
-+ return -ENOMEM;
-+
-+ for (i = 0; i < num_pages; i++) {
-+ msi_page[i].phys = start;
-+ msi_page[i].iova = start;
-+ INIT_LIST_HEAD(&msi_page[i].list);
-+ list_add(&msi_page[i].list, &cookie->msi_page_list);
-+ start += iovad->granule;
-+ }
-+
-+ return 0;
-+}
-+
-+static int iova_reserve_iommu_regions(struct device *dev,
-+ struct iommu_domain *domain)
-+{
-+ struct iommu_dma_cookie *cookie = domain->iova_cookie;
-+ struct iova_domain *iovad = &cookie->iovad;
-+ struct iommu_resv_region *region;
-+ LIST_HEAD(resv_regions);
-+ int ret = 0;
-+
-+ iommu_get_resv_regions(dev, &resv_regions);
-+ list_for_each_entry(region, &resv_regions, list) {
-+ unsigned long lo, hi;
-+
-+ /* We ARE the software that manages these! */
-+ if (region->type == IOMMU_RESV_SW_MSI)
-+ continue;
-+
-+ lo = iova_pfn(iovad, region->start);
-+ hi = iova_pfn(iovad, region->start + region->length - 1);
- reserve_iova(iovad, lo, hi);
-+
-+ if (region->type == IOMMU_RESV_MSI)
-+ ret = cookie_init_hw_msi_region(cookie, region->start,
-+ region->start + region->length);
-+ if (ret)
-+ break;
- }
-+ iommu_put_resv_regions(dev, &resv_regions);
-+
-+ return ret;
- }
-
- /**
-@@ -136,11 +277,12 @@ static void iova_reserve_pci_windows(str
- int iommu_dma_init_domain(struct iommu_domain *domain, dma_addr_t base,
- u64 size, struct device *dev)
- {
-- struct iova_domain *iovad = cookie_iovad(domain);
-+ struct iommu_dma_cookie *cookie = domain->iova_cookie;
-+ struct iova_domain *iovad = &cookie->iovad;
- unsigned long order, base_pfn, end_pfn;
-
-- if (!iovad)
-- return -ENODEV;
-+ if (!cookie || cookie->type != IOMMU_DMA_IOVA_COOKIE)
-+ return -EINVAL;
-
- /* Use the smallest supported page size for IOVA granularity */
- order = __ffs(domain->pgsize_bitmap);
-@@ -160,22 +302,37 @@ int iommu_dma_init_domain(struct iommu_d
- end_pfn = min_t(unsigned long, end_pfn,
- domain->geometry.aperture_end >> order);
- }
-+ /*
-+ * PCI devices may have larger DMA masks, but still prefer allocating
-+ * within a 32-bit mask to avoid DAC addressing. Such limitations don't
-+ * apply to the typical platform device, so for those we may as well
-+ * leave the cache limit at the top of their range to save an rb_last()
-+ * traversal on every allocation.
-+ */
-+ if (dev && dev_is_pci(dev))
-+ end_pfn &= DMA_BIT_MASK(32) >> order;
-
-- /* All we can safely do with an existing domain is enlarge it */
-+ /* start_pfn is always nonzero for an already-initialised domain */
- if (iovad->start_pfn) {
- if (1UL << order != iovad->granule ||
-- base_pfn != iovad->start_pfn ||
-- end_pfn < iovad->dma_32bit_pfn) {
-+ base_pfn != iovad->start_pfn) {
- pr_warn("Incompatible range for DMA domain\n");
- return -EFAULT;
- }
-- iovad->dma_32bit_pfn = end_pfn;
-- } else {
-- init_iova_domain(iovad, 1UL << order, base_pfn, end_pfn);
-- if (dev && dev_is_pci(dev))
-- iova_reserve_pci_windows(to_pci_dev(dev), iovad);
-+ /*
-+ * If we have devices with different DMA masks, move the free
-+ * area cache limit down for the benefit of the smaller one.
-+ */
-+ iovad->dma_32bit_pfn = min(end_pfn, iovad->dma_32bit_pfn);
-+
-+ return 0;
- }
-- return 0;
-+
-+ init_iova_domain(iovad, 1UL << order, base_pfn, end_pfn);
-+ if (!dev)
-+ return 0;
-+
-+ return iova_reserve_iommu_regions(dev, domain);
- }
- EXPORT_SYMBOL(iommu_dma_init_domain);
-
-@@ -643,11 +800,12 @@ static struct iommu_dma_msi_page *iommu_
- {
- struct iommu_dma_cookie *cookie = domain->iova_cookie;
- struct iommu_dma_msi_page *msi_page;
-- struct iova_domain *iovad = &cookie->iovad;
-+ struct iova_domain *iovad = cookie_iovad(domain);
- struct iova *iova;
- int prot = IOMMU_WRITE | IOMMU_NOEXEC | IOMMU_MMIO;
-+ size_t size = cookie_msi_granule(cookie);
-
-- msi_addr &= ~(phys_addr_t)iova_mask(iovad);
-+ msi_addr &= ~(phys_addr_t)(size - 1);
- list_for_each_entry(msi_page, &cookie->msi_page_list, list)
- if (msi_page->phys == msi_addr)
- return msi_page;
-@@ -656,13 +814,18 @@ static struct iommu_dma_msi_page *iommu_
- if (!msi_page)
- return NULL;
-
-- iova = __alloc_iova(domain, iovad->granule, dma_get_mask(dev));
-- if (!iova)
-- goto out_free_page;
--
- msi_page->phys = msi_addr;
-- msi_page->iova = iova_dma_addr(iovad, iova);
-- if (iommu_map(domain, msi_page->iova, msi_addr, iovad->granule, prot))
-+ if (iovad) {
-+ iova = __alloc_iova(domain, size, dma_get_mask(dev));
-+ if (!iova)
-+ goto out_free_page;
-+ msi_page->iova = iova_dma_addr(iovad, iova);
-+ } else {
-+ msi_page->iova = cookie->msi_iova;
-+ cookie->msi_iova += size;
-+ }
-+
-+ if (iommu_map(domain, msi_page->iova, msi_addr, size, prot))
- goto out_free_iova;
-
- INIT_LIST_HEAD(&msi_page->list);
-@@ -670,7 +833,10 @@ static struct iommu_dma_msi_page *iommu_
- return msi_page;
-
- out_free_iova:
-- __free_iova(iovad, iova);
-+ if (iovad)
-+ __free_iova(iovad, iova);
-+ else
-+ cookie->msi_iova -= size;
- out_free_page:
- kfree(msi_page);
- return NULL;
-@@ -711,7 +877,7 @@ void iommu_dma_map_msi_msg(int irq, stru
- msg->data = ~0U;
- } else {
- msg->address_hi = upper_32_bits(msi_page->iova);
-- msg->address_lo &= iova_mask(&cookie->iovad);
-+ msg->address_lo &= cookie_msi_granule(cookie) - 1;
- msg->address_lo += lower_32_bits(msi_page->iova);
- }
- }
---- a/drivers/iommu/intel-iommu.c
-+++ b/drivers/iommu/intel-iommu.c
-@@ -441,6 +441,7 @@ struct dmar_rmrr_unit {
- u64 end_address; /* reserved end address */
- struct dmar_dev_scope *devices; /* target devices */
- int devices_cnt; /* target device count */
-+ struct iommu_resv_region *resv; /* reserved region handle */
- };
-
- struct dmar_atsr_unit {
-@@ -4267,27 +4268,40 @@ static inline void init_iommu_pm_ops(voi
- int __init dmar_parse_one_rmrr(struct acpi_dmar_header *header, void *arg)
- {
- struct acpi_dmar_reserved_memory *rmrr;
-+ int prot = DMA_PTE_READ|DMA_PTE_WRITE;
- struct dmar_rmrr_unit *rmrru;
-+ size_t length;
-
- rmrru = kzalloc(sizeof(*rmrru), GFP_KERNEL);
- if (!rmrru)
-- return -ENOMEM;
-+ goto out;
-
- rmrru->hdr = header;
- rmrr = (struct acpi_dmar_reserved_memory *)header;
- rmrru->base_address = rmrr->base_address;
- rmrru->end_address = rmrr->end_address;
-+
-+ length = rmrr->end_address - rmrr->base_address + 1;
-+ rmrru->resv = iommu_alloc_resv_region(rmrr->base_address, length, prot,
-+ IOMMU_RESV_DIRECT);
-+ if (!rmrru->resv)
-+ goto free_rmrru;
-+
- rmrru->devices = dmar_alloc_dev_scope((void *)(rmrr + 1),
- ((void *)rmrr) + rmrr->header.length,
- &rmrru->devices_cnt);
-- if (rmrru->devices_cnt && rmrru->devices == NULL) {
-- kfree(rmrru);
-- return -ENOMEM;
-- }
-+ if (rmrru->devices_cnt && rmrru->devices == NULL)
-+ goto free_all;
-
- list_add(&rmrru->list, &dmar_rmrr_units);
-
- return 0;
-+free_all:
-+ kfree(rmrru->resv);
-+free_rmrru:
-+ kfree(rmrru);
-+out:
-+ return -ENOMEM;
- }
-
- static struct dmar_atsr_unit *dmar_find_atsr(struct acpi_dmar_atsr *atsr)
-@@ -4501,6 +4515,7 @@ static void intel_iommu_free_dmars(void)
- list_for_each_entry_safe(rmrru, rmrr_n, &dmar_rmrr_units, list) {
- list_del(&rmrru->list);
- dmar_free_dev_scope(&rmrru->devices, &rmrru->devices_cnt);
-+ kfree(rmrru->resv);
- kfree(rmrru);
- }
-
-@@ -5236,6 +5251,45 @@ static void intel_iommu_remove_device(st
- iommu_device_unlink(iommu->iommu_dev, dev);
- }
-
-+static void intel_iommu_get_resv_regions(struct device *device,
-+ struct list_head *head)
-+{
-+ struct iommu_resv_region *reg;
-+ struct dmar_rmrr_unit *rmrr;
-+ struct device *i_dev;
-+ int i;
-+
-+ rcu_read_lock();
-+ for_each_rmrr_units(rmrr) {
-+ for_each_active_dev_scope(rmrr->devices, rmrr->devices_cnt,
-+ i, i_dev) {
-+ if (i_dev != device)
-+ continue;
-+
-+ list_add_tail(&rmrr->resv->list, head);
-+ }
-+ }
-+ rcu_read_unlock();
-+
-+ reg = iommu_alloc_resv_region(IOAPIC_RANGE_START,
-+ IOAPIC_RANGE_END - IOAPIC_RANGE_START + 1,
-+ 0, IOMMU_RESV_MSI);
-+ if (!reg)
-+ return;
-+ list_add_tail(&reg->list, head);
-+}
-+
-+static void intel_iommu_put_resv_regions(struct device *dev,
-+ struct list_head *head)
-+{
-+ struct iommu_resv_region *entry, *next;
-+
-+ list_for_each_entry_safe(entry, next, head, list) {
-+ if (entry->type == IOMMU_RESV_RESERVED)
-+ kfree(entry);
-+ }
-+}
-+
- #ifdef CONFIG_INTEL_IOMMU_SVM
- #define MAX_NR_PASID_BITS (20)
- static inline unsigned long intel_iommu_get_pts(struct intel_iommu *iommu)
-@@ -5366,19 +5420,21 @@ struct intel_iommu *intel_svm_device_to_
- #endif /* CONFIG_INTEL_IOMMU_SVM */
-
- static const struct iommu_ops intel_iommu_ops = {
-- .capable = intel_iommu_capable,
-- .domain_alloc = intel_iommu_domain_alloc,
-- .domain_free = intel_iommu_domain_free,
-- .attach_dev = intel_iommu_attach_device,
-- .detach_dev = intel_iommu_detach_device,
-- .map = intel_iommu_map,
-- .unmap = intel_iommu_unmap,
-- .map_sg = default_iommu_map_sg,
-- .iova_to_phys = intel_iommu_iova_to_phys,
-- .add_device = intel_iommu_add_device,
-- .remove_device = intel_iommu_remove_device,
-- .device_group = pci_device_group,
-- .pgsize_bitmap = INTEL_IOMMU_PGSIZES,
-+ .capable = intel_iommu_capable,
-+ .domain_alloc = intel_iommu_domain_alloc,
-+ .domain_free = intel_iommu_domain_free,
-+ .attach_dev = intel_iommu_attach_device,
-+ .detach_dev = intel_iommu_detach_device,
-+ .map = intel_iommu_map,
-+ .unmap = intel_iommu_unmap,
-+ .map_sg = default_iommu_map_sg,
-+ .iova_to_phys = intel_iommu_iova_to_phys,
-+ .add_device = intel_iommu_add_device,
-+ .remove_device = intel_iommu_remove_device,
-+ .get_resv_regions = intel_iommu_get_resv_regions,
-+ .put_resv_regions = intel_iommu_put_resv_regions,
-+ .device_group = pci_device_group,
-+ .pgsize_bitmap = INTEL_IOMMU_PGSIZES,
- };
-
- static void quirk_iommu_g4x_gfx(struct pci_dev *dev)
---- a/drivers/iommu/iommu.c
-+++ b/drivers/iommu/iommu.c
-@@ -33,9 +33,11 @@
- #include <linux/bitops.h>
- #include <linux/property.h>
- #include <trace/events/iommu.h>
-+#include <linux/fsl/mc.h>
-
- static struct kset *iommu_group_kset;
- static DEFINE_IDA(iommu_group_ida);
-+static unsigned int iommu_def_domain_type = IOMMU_DOMAIN_DMA;
-
- struct iommu_callback_data {
- const struct iommu_ops *ops;
-@@ -68,6 +70,13 @@ struct iommu_group_attribute {
- const char *buf, size_t count);
- };
-
-+static const char * const iommu_group_resv_type_string[] = {
-+ [IOMMU_RESV_DIRECT] = "direct",
-+ [IOMMU_RESV_RESERVED] = "reserved",
-+ [IOMMU_RESV_MSI] = "msi",
-+ [IOMMU_RESV_SW_MSI] = "msi",
-+};
-+
- #define IOMMU_GROUP_ATTR(_name, _mode, _show, _store) \
- struct iommu_group_attribute iommu_group_attr_##_name = \
- __ATTR(_name, _mode, _show, _store)
-@@ -86,6 +95,18 @@ static int __iommu_attach_group(struct i
- static void __iommu_detach_group(struct iommu_domain *domain,
- struct iommu_group *group);
-
-+static int __init iommu_set_def_domain_type(char *str)
-+{
-+ bool pt;
-+
-+ if (!str || strtobool(str, &pt))
-+ return -EINVAL;
-+
-+ iommu_def_domain_type = pt ? IOMMU_DOMAIN_IDENTITY : IOMMU_DOMAIN_DMA;
-+ return 0;
-+}
-+early_param("iommu.passthrough", iommu_set_def_domain_type);
-+
- static ssize_t iommu_group_attr_show(struct kobject *kobj,
- struct attribute *__attr, char *buf)
- {
-@@ -133,8 +154,131 @@ static ssize_t iommu_group_show_name(str
- return sprintf(buf, "%s\n", group->name);
- }
-
-+/**
-+ * iommu_insert_resv_region - Insert a new region in the
-+ * list of reserved regions.
-+ * @new: new region to insert
-+ * @regions: list of regions
-+ *
-+ * The new element is sorted by address with respect to the other
-+ * regions of the same type. In case it overlaps with another
-+ * region of the same type, regions are merged. In case it
-+ * overlaps with another region of different type, regions are
-+ * not merged.
-+ */
-+static int iommu_insert_resv_region(struct iommu_resv_region *new,
-+ struct list_head *regions)
-+{
-+ struct iommu_resv_region *region;
-+ phys_addr_t start = new->start;
-+ phys_addr_t end = new->start + new->length - 1;
-+ struct list_head *pos = regions->next;
-+
-+ while (pos != regions) {
-+ struct iommu_resv_region *entry =
-+ list_entry(pos, struct iommu_resv_region, list);
-+ phys_addr_t a = entry->start;
-+ phys_addr_t b = entry->start + entry->length - 1;
-+ int type = entry->type;
-+
-+ if (end < a) {
-+ goto insert;
-+ } else if (start > b) {
-+ pos = pos->next;
-+ } else if ((start >= a) && (end <= b)) {
-+ if (new->type == type)
-+ goto done;
-+ else
-+ pos = pos->next;
-+ } else {
-+ if (new->type == type) {
-+ phys_addr_t new_start = min(a, start);
-+ phys_addr_t new_end = max(b, end);
-+
-+ list_del(&entry->list);
-+ entry->start = new_start;
-+ entry->length = new_end - new_start + 1;
-+ iommu_insert_resv_region(entry, regions);
-+ } else {
-+ pos = pos->next;
-+ }
-+ }
-+ }
-+insert:
-+ region = iommu_alloc_resv_region(new->start, new->length,
-+ new->prot, new->type);
-+ if (!region)
-+ return -ENOMEM;
-+
-+ list_add_tail(&region->list, pos);
-+done:
-+ return 0;
-+}
-+
-+static int
-+iommu_insert_device_resv_regions(struct list_head *dev_resv_regions,
-+ struct list_head *group_resv_regions)
-+{
-+ struct iommu_resv_region *entry;
-+ int ret;
-+
-+ list_for_each_entry(entry, dev_resv_regions, list) {
-+ ret = iommu_insert_resv_region(entry, group_resv_regions);
-+ if (ret)
-+ break;
-+ }
-+ return ret;
-+}
-+
-+int iommu_get_group_resv_regions(struct iommu_group *group,
-+ struct list_head *head)
-+{
-+ struct iommu_device *device;
-+ int ret = 0;
-+
-+ mutex_lock(&group->mutex);
-+ list_for_each_entry(device, &group->devices, list) {
-+ struct list_head dev_resv_regions;
-+
-+ INIT_LIST_HEAD(&dev_resv_regions);
-+ iommu_get_resv_regions(device->dev, &dev_resv_regions);
-+ ret = iommu_insert_device_resv_regions(&dev_resv_regions, head);
-+ iommu_put_resv_regions(device->dev, &dev_resv_regions);
-+ if (ret)
-+ break;
-+ }
-+ mutex_unlock(&group->mutex);
-+ return ret;
-+}
-+EXPORT_SYMBOL_GPL(iommu_get_group_resv_regions);
-+
-+static ssize_t iommu_group_show_resv_regions(struct iommu_group *group,
-+ char *buf)
-+{
-+ struct iommu_resv_region *region, *next;
-+ struct list_head group_resv_regions;
-+ char *str = buf;
-+
-+ INIT_LIST_HEAD(&group_resv_regions);
-+ iommu_get_group_resv_regions(group, &group_resv_regions);
-+
-+ list_for_each_entry_safe(region, next, &group_resv_regions, list) {
-+ str += sprintf(str, "0x%016llx 0x%016llx %s\n",
-+ (long long int)region->start,
-+ (long long int)(region->start +
-+ region->length - 1),
-+ iommu_group_resv_type_string[region->type]);
-+ kfree(region);
-+ }
-+
-+ return (str - buf);
-+}
-+
- static IOMMU_GROUP_ATTR(name, S_IRUGO, iommu_group_show_name, NULL);
-
-+static IOMMU_GROUP_ATTR(reserved_regions, 0444,
-+ iommu_group_show_resv_regions, NULL);
-+
- static void iommu_group_release(struct kobject *kobj)
- {
- struct iommu_group *group = to_iommu_group(kobj);
-@@ -212,6 +356,11 @@ struct iommu_group *iommu_group_alloc(vo
- */
- kobject_put(&group->kobj);
-
-+ ret = iommu_group_create_file(group,
-+ &iommu_group_attr_reserved_regions);
-+ if (ret)
-+ return ERR_PTR(ret);
-+
- pr_debug("Allocated group %d\n", group->id);
-
- return group;
-@@ -318,7 +467,7 @@ static int iommu_group_create_direct_map
- struct device *dev)
- {
- struct iommu_domain *domain = group->default_domain;
-- struct iommu_dm_region *entry;
-+ struct iommu_resv_region *entry;
- struct list_head mappings;
- unsigned long pg_size;
- int ret = 0;
-@@ -331,18 +480,21 @@ static int iommu_group_create_direct_map
- pg_size = 1UL << __ffs(domain->pgsize_bitmap);
- INIT_LIST_HEAD(&mappings);
-
-- iommu_get_dm_regions(dev, &mappings);
-+ iommu_get_resv_regions(dev, &mappings);
-
- /* We need to consider overlapping regions for different devices */
- list_for_each_entry(entry, &mappings, list) {
- dma_addr_t start, end, addr;
-
-- if (domain->ops->apply_dm_region)
-- domain->ops->apply_dm_region(dev, domain, entry);
-+ if (domain->ops->apply_resv_region)
-+ domain->ops->apply_resv_region(dev, domain, entry);
-
- start = ALIGN(entry->start, pg_size);
- end = ALIGN(entry->start + entry->length, pg_size);
-
-+ if (entry->type != IOMMU_RESV_DIRECT)
-+ continue;
-+
- for (addr = start; addr < end; addr += pg_size) {
- phys_addr_t phys_addr;
-
-@@ -358,7 +510,7 @@ static int iommu_group_create_direct_map
- }
-
- out:
-- iommu_put_dm_regions(dev, &mappings);
-+ iommu_put_resv_regions(dev, &mappings);
-
- return ret;
- }
-@@ -563,6 +715,19 @@ struct iommu_group *iommu_group_get(stru
- EXPORT_SYMBOL_GPL(iommu_group_get);
-
- /**
-+ * iommu_group_ref_get - Increment reference on a group
-+ * @group: the group to use, must not be NULL
-+ *
-+ * This function is called by iommu drivers to take additional references on an
-+ * existing group. Returns the given group for convenience.
-+ */
-+struct iommu_group *iommu_group_ref_get(struct iommu_group *group)
-+{
-+ kobject_get(group->devices_kobj);
-+ return group;
-+}
-+
-+/**
- * iommu_group_put - Decrement group reference
- * @group: the group to use
- *
-@@ -812,6 +977,26 @@ struct iommu_group *pci_device_group(str
- return group;
- }
-
-+/* Get the IOMMU group for device on fsl-mc bus */
-+struct iommu_group *fsl_mc_device_group(struct device *dev)
-+{
-+ struct device *cont_dev = fsl_mc_cont_dev(dev);
-+ struct iommu_group *group;
-+
-+ /* Container device is responsible for creating the iommu group */
-+ if (fsl_mc_is_cont_dev(dev)) {
-+ group = iommu_group_alloc();
-+ if (IS_ERR(group))
-+ return NULL;
-+ } else {
-+ get_device(cont_dev);
-+ group = iommu_group_get(cont_dev);
-+ put_device(cont_dev);
-+ }
-+
-+ return group;
-+}
-+
- /**
- * iommu_group_get_for_dev - Find or create the IOMMU group for a device
- * @dev: target device
-@@ -845,10 +1030,19 @@ struct iommu_group *iommu_group_get_for_
- * IOMMU driver.
- */
- if (!group->default_domain) {
-- group->default_domain = __iommu_domain_alloc(dev->bus,
-- IOMMU_DOMAIN_DMA);
-+ struct iommu_domain *dom;
-+
-+ dom = __iommu_domain_alloc(dev->bus, iommu_def_domain_type);
-+ if (!dom && iommu_def_domain_type != IOMMU_DOMAIN_DMA) {
-+ dev_warn(dev,
-+ "failed to allocate default IOMMU domain of type %u; falling back to IOMMU_DOMAIN_DMA",
-+ iommu_def_domain_type);
-+ dom = __iommu_domain_alloc(dev->bus, IOMMU_DOMAIN_DMA);
-+ }
-+
-+ group->default_domain = dom;
- if (!group->domain)
-- group->domain = group->default_domain;
-+ group->domain = dom;
- }
-
- ret = iommu_group_add_device(group, dev);
-@@ -1557,20 +1751,38 @@ int iommu_domain_set_attr(struct iommu_d
- }
- EXPORT_SYMBOL_GPL(iommu_domain_set_attr);
-
--void iommu_get_dm_regions(struct device *dev, struct list_head *list)
-+void iommu_get_resv_regions(struct device *dev, struct list_head *list)
- {
- const struct iommu_ops *ops = dev->bus->iommu_ops;
-
-- if (ops && ops->get_dm_regions)
-- ops->get_dm_regions(dev, list);
-+ if (ops && ops->get_resv_regions)
-+ ops->get_resv_regions(dev, list);
- }
-
--void iommu_put_dm_regions(struct device *dev, struct list_head *list)
-+void iommu_put_resv_regions(struct device *dev, struct list_head *list)
- {
- const struct iommu_ops *ops = dev->bus->iommu_ops;
-
-- if (ops && ops->put_dm_regions)
-- ops->put_dm_regions(dev, list);
-+ if (ops && ops->put_resv_regions)
-+ ops->put_resv_regions(dev, list);
-+}
-+
-+struct iommu_resv_region *iommu_alloc_resv_region(phys_addr_t start,
-+ size_t length, int prot,
-+ enum iommu_resv_type type)
-+{
-+ struct iommu_resv_region *region;
-+
-+ region = kzalloc(sizeof(*region), GFP_KERNEL);
-+ if (!region)
-+ return NULL;
-+
-+ INIT_LIST_HEAD(&region->list);
-+ region->start = start;
-+ region->length = length;
-+ region->prot = prot;
-+ region->type = type;
-+ return region;
- }
-
- /* Request that a device is direct mapped by the IOMMU */
---- a/drivers/iommu/mtk_iommu.c
-+++ b/drivers/iommu/mtk_iommu.c
-@@ -410,6 +410,8 @@ static struct iommu_group *mtk_iommu_dev
- data->m4u_group = iommu_group_alloc();
- if (IS_ERR(data->m4u_group))
- dev_err(dev, "Failed to allocate M4U IOMMU group\n");
-+ } else {
-+ iommu_group_ref_get(data->m4u_group);
- }
- return data->m4u_group;
- }
---- a/drivers/iommu/mtk_iommu_v1.c
-+++ b/drivers/iommu/mtk_iommu_v1.c
-@@ -502,6 +502,8 @@ static struct iommu_group *mtk_iommu_dev
- data->m4u_group = iommu_group_alloc();
- if (IS_ERR(data->m4u_group))
- dev_err(dev, "Failed to allocate M4U IOMMU group\n");
-+ } else {
-+ iommu_group_ref_get(data->m4u_group);
- }
- return data->m4u_group;
- }
---- a/include/linux/dma-iommu.h
-+++ b/include/linux/dma-iommu.h
-@@ -28,6 +28,7 @@ int iommu_dma_init(void);
-
- /* Domain management interface for IOMMU drivers */
- int iommu_get_dma_cookie(struct iommu_domain *domain);
-+int iommu_get_msi_cookie(struct iommu_domain *domain, dma_addr_t base);
- void iommu_put_dma_cookie(struct iommu_domain *domain);
-
- /* Setup call for arch DMA mapping code */
-@@ -67,6 +68,7 @@ int iommu_dma_mapping_error(struct devic
-
- /* The DMA API isn't _quite_ the whole story, though... */
- void iommu_dma_map_msi_msg(int irq, struct msi_msg *msg);
-+void iommu_dma_get_resv_regions(struct device *dev, struct list_head *list);
-
- #else
-
-@@ -83,6 +85,11 @@ static inline int iommu_get_dma_cookie(s
- return -ENODEV;
- }
-
-+static inline int iommu_get_msi_cookie(struct iommu_domain *domain, dma_addr_t base)
-+{
-+ return -ENODEV;
-+}
-+
- static inline void iommu_put_dma_cookie(struct iommu_domain *domain)
- {
- }
-@@ -91,6 +98,10 @@ static inline void iommu_dma_map_msi_msg
- {
- }
-
-+static inline void iommu_dma_get_resv_regions(struct device *dev, struct list_head *list)
-+{
-+}
-+
- #endif /* CONFIG_IOMMU_DMA */
- #endif /* __KERNEL__ */
- #endif /* __DMA_IOMMU_H */
---- a/include/linux/iommu.h
-+++ b/include/linux/iommu.h
-@@ -117,18 +117,32 @@ enum iommu_attr {
- DOMAIN_ATTR_MAX,
- };
-
-+/* These are the possible reserved region types */
-+enum iommu_resv_type {
-+ /* Memory regions which must be mapped 1:1 at all times */
-+ IOMMU_RESV_DIRECT,
-+ /* Arbitrary "never map this or give it to a device" address ranges */
-+ IOMMU_RESV_RESERVED,
-+ /* Hardware MSI region (untranslated) */
-+ IOMMU_RESV_MSI,
-+ /* Software-managed MSI translation window */
-+ IOMMU_RESV_SW_MSI,
-+};
-+
- /**
-- * struct iommu_dm_region - descriptor for a direct mapped memory region
-+ * struct iommu_resv_region - descriptor for a reserved memory region
- * @list: Linked list pointers
- * @start: System physical start address of the region
- * @length: Length of the region in bytes
- * @prot: IOMMU Protection flags (READ/WRITE/...)
-+ * @type: Type of the reserved region
- */
--struct iommu_dm_region {
-+struct iommu_resv_region {
- struct list_head list;
- phys_addr_t start;
- size_t length;
- int prot;
-+ enum iommu_resv_type type;
- };
-
- #ifdef CONFIG_IOMMU_API
-@@ -150,9 +164,9 @@ struct iommu_dm_region {
- * @device_group: find iommu group for a particular device
- * @domain_get_attr: Query domain attributes
- * @domain_set_attr: Change domain attributes
-- * @get_dm_regions: Request list of direct mapping requirements for a device
-- * @put_dm_regions: Free list of direct mapping requirements for a device
-- * @apply_dm_region: Temporary helper call-back for iova reserved ranges
-+ * @get_resv_regions: Request list of reserved regions for a device
-+ * @put_resv_regions: Free list of reserved regions for a device
-+ * @apply_resv_region: Temporary helper call-back for iova reserved ranges
- * @domain_window_enable: Configure and enable a particular window for a domain
- * @domain_window_disable: Disable a particular window for a domain
- * @domain_set_windows: Set the number of windows for a domain
-@@ -184,11 +198,12 @@ struct iommu_ops {
- int (*domain_set_attr)(struct iommu_domain *domain,
- enum iommu_attr attr, void *data);
-
-- /* Request/Free a list of direct mapping requirements for a device */
-- void (*get_dm_regions)(struct device *dev, struct list_head *list);
-- void (*put_dm_regions)(struct device *dev, struct list_head *list);
-- void (*apply_dm_region)(struct device *dev, struct iommu_domain *domain,
-- struct iommu_dm_region *region);
-+ /* Request/Free a list of reserved regions for a device */
-+ void (*get_resv_regions)(struct device *dev, struct list_head *list);
-+ void (*put_resv_regions)(struct device *dev, struct list_head *list);
-+ void (*apply_resv_region)(struct device *dev,
-+ struct iommu_domain *domain,
-+ struct iommu_resv_region *region);
-
- /* Window handling functions */
- int (*domain_window_enable)(struct iommu_domain *domain, u32 wnd_nr,
-@@ -233,9 +248,14 @@ extern phys_addr_t iommu_iova_to_phys(st
- extern void iommu_set_fault_handler(struct iommu_domain *domain,
- iommu_fault_handler_t handler, void *token);
-
--extern void iommu_get_dm_regions(struct device *dev, struct list_head *list);
--extern void iommu_put_dm_regions(struct device *dev, struct list_head *list);
-+extern void iommu_get_resv_regions(struct device *dev, struct list_head *list);
-+extern void iommu_put_resv_regions(struct device *dev, struct list_head *list);
- extern int iommu_request_dm_for_dev(struct device *dev);
-+extern struct iommu_resv_region *
-+iommu_alloc_resv_region(phys_addr_t start, size_t length, int prot,
-+ enum iommu_resv_type type);
-+extern int iommu_get_group_resv_regions(struct iommu_group *group,
-+ struct list_head *head);
-
- extern int iommu_attach_group(struct iommu_domain *domain,
- struct iommu_group *group);
-@@ -253,6 +273,7 @@ extern void iommu_group_remove_device(st
- extern int iommu_group_for_each_dev(struct iommu_group *group, void *data,
- int (*fn)(struct device *, void *));
- extern struct iommu_group *iommu_group_get(struct device *dev);
-+extern struct iommu_group *iommu_group_ref_get(struct iommu_group *group);
- extern void iommu_group_put(struct iommu_group *group);
- extern int iommu_group_register_notifier(struct iommu_group *group,
- struct notifier_block *nb);
-@@ -330,6 +351,8 @@ static inline size_t iommu_map_sg(struct
- extern struct iommu_group *pci_device_group(struct device *dev);
- /* Generic device grouping function */
- extern struct iommu_group *generic_device_group(struct device *dev);
-+/* FSL-MC device grouping function */
-+struct iommu_group *fsl_mc_device_group(struct device *dev);
-
- /**
- * struct iommu_fwspec - per-device IOMMU instance data
-@@ -439,16 +462,22 @@ static inline void iommu_set_fault_handl
- {
- }
-
--static inline void iommu_get_dm_regions(struct device *dev,
-+static inline void iommu_get_resv_regions(struct device *dev,
- struct list_head *list)
- {
- }
-
--static inline void iommu_put_dm_regions(struct device *dev,
-+static inline void iommu_put_resv_regions(struct device *dev,
- struct list_head *list)
- {
- }
-
-+static inline int iommu_get_group_resv_regions(struct iommu_group *group,
-+ struct list_head *head)
-+{
-+ return -ENODEV;
-+}
-+
- static inline int iommu_request_dm_for_dev(struct device *dev)
- {
- return -ENODEV;
diff --git a/target/linux/layerscape/patches-4.9/811-irqchip-support-layerscape.patch b/target/linux/layerscape/patches-4.9/811-irqchip-support-layerscape.patch
deleted file mode 100644
index 3aca252675..0000000000
--- a/target/linux/layerscape/patches-4.9/811-irqchip-support-layerscape.patch
+++ /dev/null
@@ -1,182 +0,0 @@
-From dab02a7cc54494740e849cd51b554d100eb5541d Mon Sep 17 00:00:00 2001
-From: Yangbo Lu <yangbo.lu@nxp.com>
-Date: Thu, 5 Jul 2018 17:36:09 +0800
-Subject: [PATCH 23/32] irqchip: support layerscape
-
-This is an integrated patch for layerscape gic support.
-
-Signed-off-by: Eric Auger <eric.auger@redhat.com>
-Signed-off-by: Zhao Qiang <qiang.zhao@nxp.com>
-Signed-off-by: Yangbo Lu <yangbo.lu@nxp.com>
----
- drivers/irqchip/Makefile | 1 +
- drivers/irqchip/irq-gic-v3-its.c | 1 +
- include/linux/irqchip/arm-gic-v3.h | 3 +++
- include/linux/irqdomain.h | 36 +++++++++++++++++++++++++++
- kernel/irq/irqdomain.c | 39 ++++++++++++++++++++++++++++++
- kernel/irq/msi.c | 4 +--
- 6 files changed, 82 insertions(+), 2 deletions(-)
-
---- a/drivers/irqchip/Makefile
-+++ b/drivers/irqchip/Makefile
-@@ -75,3 +75,4 @@ obj-$(CONFIG_LS_SCFG_MSI) += irq-ls-scf
- obj-$(CONFIG_EZNPS_GIC) += irq-eznps.o
- obj-$(CONFIG_ARCH_ASPEED) += irq-aspeed-vic.o
- obj-$(CONFIG_STM32_EXTI) += irq-stm32-exti.o
-+obj-$(CONFIG_QUICC_ENGINE) += irq-qeic.o
---- a/drivers/irqchip/irq-gic-v3-its.c
-+++ b/drivers/irqchip/irq-gic-v3-its.c
-@@ -1658,6 +1658,7 @@ static int its_init_domain(struct fwnode
-
- inner_domain->parent = its_parent;
- inner_domain->bus_token = DOMAIN_BUS_NEXUS;
-+ inner_domain->flags |= IRQ_DOMAIN_FLAG_MSI_REMAP;
- info->ops = &its_msi_domain_ops;
- info->data = its;
- inner_domain->host_data = info;
---- a/include/linux/irqchip/arm-gic-v3.h
-+++ b/include/linux/irqchip/arm-gic-v3.h
-@@ -133,6 +133,9 @@
- #define GIC_BASER_SHAREABILITY(reg, type) \
- (GIC_BASER_##type << reg##_SHAREABILITY_SHIFT)
-
-+/* encode a size field of width @w containing @n - 1 units */
-+#define GIC_ENCODE_SZ(n, w) (((unsigned long)(n) - 1) & GENMASK_ULL(((w) - 1), 0))
-+
- #define GICR_PROPBASER_SHAREABILITY_SHIFT (10)
- #define GICR_PROPBASER_INNER_CACHEABILITY_SHIFT (7)
- #define GICR_PROPBASER_OUTER_CACHEABILITY_SHIFT (56)
---- a/include/linux/irqdomain.h
-+++ b/include/linux/irqdomain.h
-@@ -187,6 +187,12 @@ enum {
- /* Irq domain is an IPI domain with single virq */
- IRQ_DOMAIN_FLAG_IPI_SINGLE = (1 << 3),
-
-+ /* Irq domain implements MSIs */
-+ IRQ_DOMAIN_FLAG_MSI = (1 << 4),
-+
-+ /* Irq domain implements MSI remapping */
-+ IRQ_DOMAIN_FLAG_MSI_REMAP = (1 << 5),
-+
- /*
- * Flags starting from IRQ_DOMAIN_FLAG_NONCORE are reserved
- * for implementation specific purposes and ignored by the
-@@ -220,6 +226,7 @@ struct irq_domain *irq_domain_add_legacy
- void *host_data);
- extern struct irq_domain *irq_find_matching_fwspec(struct irq_fwspec *fwspec,
- enum irq_domain_bus_token bus_token);
-+extern bool irq_domain_check_msi_remap(void);
- extern void irq_set_default_host(struct irq_domain *host);
- extern int irq_domain_alloc_descs(int virq, unsigned int nr_irqs,
- irq_hw_number_t hwirq, int node,
-@@ -453,6 +460,19 @@ static inline bool irq_domain_is_ipi_sin
- {
- return domain->flags & IRQ_DOMAIN_FLAG_IPI_SINGLE;
- }
-+
-+static inline bool irq_domain_is_msi(struct irq_domain *domain)
-+{
-+ return domain->flags & IRQ_DOMAIN_FLAG_MSI;
-+}
-+
-+static inline bool irq_domain_is_msi_remap(struct irq_domain *domain)
-+{
-+ return domain->flags & IRQ_DOMAIN_FLAG_MSI_REMAP;
-+}
-+
-+extern bool irq_domain_hierarchical_is_msi_remap(struct irq_domain *domain);
-+
- #else /* CONFIG_IRQ_DOMAIN_HIERARCHY */
- static inline void irq_domain_activate_irq(struct irq_data *data) { }
- static inline void irq_domain_deactivate_irq(struct irq_data *data) { }
-@@ -484,6 +504,22 @@ static inline bool irq_domain_is_ipi_sin
- {
- return false;
- }
-+
-+static inline bool irq_domain_is_msi(struct irq_domain *domain)
-+{
-+ return false;
-+}
-+
-+static inline bool irq_domain_is_msi_remap(struct irq_domain *domain)
-+{
-+ return false;
-+}
-+
-+static inline bool
-+irq_domain_hierarchical_is_msi_remap(struct irq_domain *domain)
-+{
-+ return false;
-+}
- #endif /* CONFIG_IRQ_DOMAIN_HIERARCHY */
-
- #else /* CONFIG_IRQ_DOMAIN */
---- a/kernel/irq/irqdomain.c
-+++ b/kernel/irq/irqdomain.c
-@@ -319,6 +319,31 @@ struct irq_domain *irq_find_matching_fws
- EXPORT_SYMBOL_GPL(irq_find_matching_fwspec);
-
- /**
-+ * irq_domain_check_msi_remap - Check whether all MSI irq domains implement
-+ * IRQ remapping
-+ *
-+ * Return: false if any MSI irq domain does not support IRQ remapping,
-+ * true otherwise (including if there is no MSI irq domain)
-+ */
-+bool irq_domain_check_msi_remap(void)
-+{
-+ struct irq_domain *h;
-+ bool ret = true;
-+
-+ mutex_lock(&irq_domain_mutex);
-+ list_for_each_entry(h, &irq_domain_list, link) {
-+ if (irq_domain_is_msi(h) &&
-+ !irq_domain_hierarchical_is_msi_remap(h)) {
-+ ret = false;
-+ break;
-+ }
-+ }
-+ mutex_unlock(&irq_domain_mutex);
-+ return ret;
-+}
-+EXPORT_SYMBOL_GPL(irq_domain_check_msi_remap);
-+
-+/**
- * irq_set_default_host() - Set a "default" irq domain
- * @domain: default domain pointer
- *
-@@ -1420,6 +1445,20 @@ static void irq_domain_check_hierarchy(s
- if (domain->ops->alloc)
- domain->flags |= IRQ_DOMAIN_FLAG_HIERARCHY;
- }
-+
-+/**
-+ * irq_domain_hierarchical_is_msi_remap - Check if the domain or any
-+ * parent has MSI remapping support
-+ * @domain: domain pointer
-+ */
-+bool irq_domain_hierarchical_is_msi_remap(struct irq_domain *domain)
-+{
-+ for (; domain; domain = domain->parent) {
-+ if (irq_domain_is_msi_remap(domain))
-+ return true;
-+ }
-+ return false;
-+}
- #else /* CONFIG_IRQ_DOMAIN_HIERARCHY */
- /**
- * irq_domain_get_irq_data - Get irq_data associated with @virq and @domain
---- a/kernel/irq/msi.c
-+++ b/kernel/irq/msi.c
-@@ -272,8 +272,8 @@ struct irq_domain *msi_create_irq_domain
- if (info->flags & MSI_FLAG_USE_DEF_CHIP_OPS)
- msi_domain_update_chip_ops(info);
-
-- return irq_domain_create_hierarchy(parent, 0, 0, fwnode,
-- &msi_domain_ops, info);
-+ return irq_domain_create_hierarchy(parent, IRQ_DOMAIN_FLAG_MSI, 0,
-+ fwnode, &msi_domain_ops, info);
- }
-
- int msi_domain_prepare_irqs(struct irq_domain *domain, struct device *dev,
diff --git a/target/linux/layerscape/patches-4.9/812-mmc-layerscape-support.patch b/target/linux/layerscape/patches-4.9/812-mmc-layerscape-support.patch
deleted file mode 100644
index 4b32417379..0000000000
--- a/target/linux/layerscape/patches-4.9/812-mmc-layerscape-support.patch
+++ /dev/null
@@ -1,594 +0,0 @@
-From 4215d5757595e7ec7ca146c2b901beb177f415d8 Mon Sep 17 00:00:00 2001
-From: Yangbo Lu <yangbo.lu@nxp.com>
-Date: Wed, 17 Jan 2018 15:37:13 +0800
-Subject: [PATCH 24/30] mmc: layerscape support
-
-This is an integrated patch for layerscape mmc support.
-
-Adrian Hunter <adrian.hunter@intel.com>
-Jaehoon Chung <jh80.chung@samsung.com>
-Masahiro Yamada <yamada.masahiro@socionext.com>
-Signed-off-by: Yangbo Lu <yangbo.lu@nxp.com>
----
- drivers/mmc/host/Kconfig | 1 +
- drivers/mmc/host/sdhci-esdhc.h | 52 +++++---
- drivers/mmc/host/sdhci-of-esdhc.c | 265 ++++++++++++++++++++++++++++++++++++--
- drivers/mmc/host/sdhci.c | 45 ++++---
- drivers/mmc/host/sdhci.h | 3 +
- 5 files changed, 320 insertions(+), 46 deletions(-)
-
---- a/drivers/mmc/host/Kconfig
-+++ b/drivers/mmc/host/Kconfig
-@@ -144,6 +144,7 @@ config MMC_SDHCI_OF_ESDHC
- depends on MMC_SDHCI_PLTFM
- depends on PPC || ARCH_MXC || ARCH_LAYERSCAPE
- select MMC_SDHCI_IO_ACCESSORS
-+ select FSL_GUTS
- help
- This selects the Freescale eSDHC controller support.
-
---- a/drivers/mmc/host/sdhci-esdhc.h
-+++ b/drivers/mmc/host/sdhci-esdhc.h
-@@ -24,30 +24,46 @@
- SDHCI_QUIRK_PIO_NEEDS_DELAY | \
- SDHCI_QUIRK_NO_HISPD_BIT)
-
--#define ESDHC_PROCTL 0x28
--
--#define ESDHC_SYSTEM_CONTROL 0x2c
--#define ESDHC_CLOCK_MASK 0x0000fff0
--#define ESDHC_PREDIV_SHIFT 8
--#define ESDHC_DIVIDER_SHIFT 4
--#define ESDHC_CLOCK_PEREN 0x00000004
--#define ESDHC_CLOCK_HCKEN 0x00000002
--#define ESDHC_CLOCK_IPGEN 0x00000001
--
- /* pltfm-specific */
- #define ESDHC_HOST_CONTROL_LE 0x20
-
- /*
-- * P2020 interpretation of the SDHCI_HOST_CONTROL register
-+ * eSDHC register definition
- */
--#define ESDHC_CTRL_4BITBUS (0x1 << 1)
--#define ESDHC_CTRL_8BITBUS (0x2 << 1)
--#define ESDHC_CTRL_BUSWIDTH_MASK (0x3 << 1)
--
--/* OF-specific */
--#define ESDHC_DMA_SYSCTL 0x40c
--#define ESDHC_DMA_SNOOP 0x00000040
-
--#define ESDHC_HOST_CONTROL_RES 0x01
-+/* Present State Register */
-+#define ESDHC_PRSSTAT 0x24
-+#define ESDHC_CLOCK_STABLE 0x00000008
-+
-+/* Protocol Control Register */
-+#define ESDHC_PROCTL 0x28
-+#define ESDHC_VOLT_SEL 0x00000400
-+#define ESDHC_CTRL_4BITBUS (0x1 << 1)
-+#define ESDHC_CTRL_8BITBUS (0x2 << 1)
-+#define ESDHC_CTRL_BUSWIDTH_MASK (0x3 << 1)
-+#define ESDHC_HOST_CONTROL_RES 0x01
-+
-+/* System Control Register */
-+#define ESDHC_SYSTEM_CONTROL 0x2c
-+#define ESDHC_CLOCK_MASK 0x0000fff0
-+#define ESDHC_PREDIV_SHIFT 8
-+#define ESDHC_DIVIDER_SHIFT 4
-+#define ESDHC_CLOCK_SDCLKEN 0x00000008
-+#define ESDHC_CLOCK_PEREN 0x00000004
-+#define ESDHC_CLOCK_HCKEN 0x00000002
-+#define ESDHC_CLOCK_IPGEN 0x00000001
-+
-+/* Host Controller Capabilities Register 2 */
-+#define ESDHC_CAPABILITIES_1 0x114
-+
-+/* Tuning Block Control Register */
-+#define ESDHC_TBCTL 0x120
-+#define ESDHC_TB_EN 0x00000004
-+
-+/* Control Register for DMA transfer */
-+#define ESDHC_DMA_SYSCTL 0x40c
-+#define ESDHC_PERIPHERAL_CLK_SEL 0x00080000
-+#define ESDHC_FLUSH_ASYNC_FIFO 0x00040000
-+#define ESDHC_DMA_SNOOP 0x00000040
-
- #endif /* _DRIVERS_MMC_SDHCI_ESDHC_H */
---- a/drivers/mmc/host/sdhci-of-esdhc.c
-+++ b/drivers/mmc/host/sdhci-of-esdhc.c
-@@ -16,8 +16,12 @@
- #include <linux/err.h>
- #include <linux/io.h>
- #include <linux/of.h>
-+#include <linux/of_address.h>
- #include <linux/delay.h>
- #include <linux/module.h>
-+#include <linux/sys_soc.h>
-+#include <linux/clk.h>
-+#include <linux/ktime.h>
- #include <linux/mmc/host.h>
- #include "sdhci-pltfm.h"
- #include "sdhci-esdhc.h"
-@@ -28,8 +32,12 @@
- struct sdhci_esdhc {
- u8 vendor_ver;
- u8 spec_ver;
-+ bool quirk_incorrect_hostver;
-+ unsigned int peripheral_clock;
- };
-
-+static void esdhc_clock_enable(struct sdhci_host *host, bool enable);
-+
- /**
- * esdhc_read*_fixup - Fixup the value read from incompatible eSDHC register
- * to make it compatible with SD spec.
-@@ -80,6 +88,17 @@ static u32 esdhc_readl_fixup(struct sdhc
- return ret;
- }
-
-+ /*
-+ * DTS properties of mmc host are used to enable each speed mode
-+ * according to soc and board capability. So clean up
-+ * SDR50/SDR104/DDR50 support bits here.
-+ */
-+ if (spec_reg == SDHCI_CAPABILITIES_1) {
-+ ret = value & (~(SDHCI_SUPPORT_SDR50 | SDHCI_SUPPORT_SDR104 |
-+ SDHCI_SUPPORT_DDR50));
-+ return ret;
-+ }
-+
- ret = value;
- return ret;
- }
-@@ -87,6 +106,8 @@ static u32 esdhc_readl_fixup(struct sdhc
- static u16 esdhc_readw_fixup(struct sdhci_host *host,
- int spec_reg, u32 value)
- {
-+ struct sdhci_pltfm_host *pltfm_host = sdhci_priv(host);
-+ struct sdhci_esdhc *esdhc = sdhci_pltfm_priv(pltfm_host);
- u16 ret;
- int shift = (spec_reg & 0x2) * 8;
-
-@@ -94,6 +115,12 @@ static u16 esdhc_readw_fixup(struct sdhc
- ret = value & 0xffff;
- else
- ret = (value >> shift) & 0xffff;
-+ /* Workaround for T4240-R1.0-R2.0 eSDHC which has incorrect
-+ * vendor version and spec version information.
-+ */
-+ if ((spec_reg == SDHCI_HOST_VERSION) &&
-+ (esdhc->quirk_incorrect_hostver))
-+ ret = (VENDOR_V_23 << SDHCI_VENDOR_VER_SHIFT) | SDHCI_SPEC_200;
- return ret;
- }
-
-@@ -235,7 +262,11 @@ static u32 esdhc_be_readl(struct sdhci_h
- u32 ret;
- u32 value;
-
-- value = ioread32be(host->ioaddr + reg);
-+ if (reg == SDHCI_CAPABILITIES_1)
-+ value = ioread32be(host->ioaddr + ESDHC_CAPABILITIES_1);
-+ else
-+ value = ioread32be(host->ioaddr + reg);
-+
- ret = esdhc_readl_fixup(host, reg, value);
-
- return ret;
-@@ -246,7 +277,11 @@ static u32 esdhc_le_readl(struct sdhci_h
- u32 ret;
- u32 value;
-
-- value = ioread32(host->ioaddr + reg);
-+ if (reg == SDHCI_CAPABILITIES_1)
-+ value = ioread32(host->ioaddr + ESDHC_CAPABILITIES_1);
-+ else
-+ value = ioread32(host->ioaddr + reg);
-+
- ret = esdhc_readl_fixup(host, reg, value);
-
- return ret;
-@@ -404,15 +439,25 @@ static int esdhc_of_enable_dma(struct sd
- static unsigned int esdhc_of_get_max_clock(struct sdhci_host *host)
- {
- struct sdhci_pltfm_host *pltfm_host = sdhci_priv(host);
-+ struct sdhci_esdhc *esdhc = sdhci_pltfm_priv(pltfm_host);
-
-- return pltfm_host->clock;
-+ if (esdhc->peripheral_clock)
-+ return esdhc->peripheral_clock;
-+ else
-+ return pltfm_host->clock;
- }
-
- static unsigned int esdhc_of_get_min_clock(struct sdhci_host *host)
- {
- struct sdhci_pltfm_host *pltfm_host = sdhci_priv(host);
-+ struct sdhci_esdhc *esdhc = sdhci_pltfm_priv(pltfm_host);
-+ unsigned int clock;
-
-- return pltfm_host->clock / 256 / 16;
-+ if (esdhc->peripheral_clock)
-+ clock = esdhc->peripheral_clock;
-+ else
-+ clock = pltfm_host->clock;
-+ return clock / 256 / 16;
- }
-
- static void esdhc_of_set_clock(struct sdhci_host *host, unsigned int clock)
-@@ -421,12 +466,15 @@ static void esdhc_of_set_clock(struct sd
- struct sdhci_esdhc *esdhc = sdhci_pltfm_priv(pltfm_host);
- int pre_div = 1;
- int div = 1;
-+ ktime_t timeout;
- u32 temp;
-
- host->mmc->actual_clock = 0;
-
-- if (clock == 0)
-+ if (clock == 0) {
-+ esdhc_clock_enable(host, false);
- return;
-+ }
-
- /* Workaround to start pre_div at 2 for VNN < VENDOR_V_23 */
- if (esdhc->vendor_ver < VENDOR_V_23)
-@@ -454,9 +502,15 @@ static void esdhc_of_set_clock(struct sd
- clock -= 5000000;
- }
-
-+ /* Workaround to reduce the clock frequency for ls1021a esdhc */
-+ if (of_find_compatible_node(NULL, NULL, "fsl,ls1021a-esdhc")) {
-+ if (clock == 50000000)
-+ clock = 46500000;
-+ }
-+
- temp = sdhci_readl(host, ESDHC_SYSTEM_CONTROL);
-- temp &= ~(ESDHC_CLOCK_IPGEN | ESDHC_CLOCK_HCKEN | ESDHC_CLOCK_PEREN
-- | ESDHC_CLOCK_MASK);
-+ temp &= ~(ESDHC_CLOCK_SDCLKEN | ESDHC_CLOCK_IPGEN | ESDHC_CLOCK_HCKEN |
-+ ESDHC_CLOCK_PEREN | ESDHC_CLOCK_MASK);
- sdhci_writel(host, temp, ESDHC_SYSTEM_CONTROL);
-
- while (host->max_clk / pre_div / 16 > clock && pre_div < 256)
-@@ -476,7 +530,20 @@ static void esdhc_of_set_clock(struct sd
- | (div << ESDHC_DIVIDER_SHIFT)
- | (pre_div << ESDHC_PREDIV_SHIFT));
- sdhci_writel(host, temp, ESDHC_SYSTEM_CONTROL);
-- mdelay(1);
-+
-+ /* Wait max 20 ms */
-+ timeout = ktime_add_ms(ktime_get(), 20);
-+ while (!(sdhci_readl(host, ESDHC_PRSSTAT) & ESDHC_CLOCK_STABLE)) {
-+ if (ktime_after(ktime_get(), timeout)) {
-+ pr_err("%s: Internal clock never stabilised.\n",
-+ mmc_hostname(host->mmc));
-+ return;
-+ }
-+ udelay(10);
-+ }
-+
-+ temp |= ESDHC_CLOCK_SDCLKEN;
-+ sdhci_writel(host, temp, ESDHC_SYSTEM_CONTROL);
- }
-
- static void esdhc_pltfm_set_bus_width(struct sdhci_host *host, int width)
-@@ -501,12 +568,136 @@ static void esdhc_pltfm_set_bus_width(st
- sdhci_writel(host, ctrl, ESDHC_PROCTL);
- }
-
-+static void esdhc_clock_enable(struct sdhci_host *host, bool enable)
-+{
-+ u32 val;
-+ ktime_t timeout;
-+
-+ val = sdhci_readl(host, ESDHC_SYSTEM_CONTROL);
-+
-+ if (enable)
-+ val |= ESDHC_CLOCK_SDCLKEN;
-+ else
-+ val &= ~ESDHC_CLOCK_SDCLKEN;
-+
-+ sdhci_writel(host, val, ESDHC_SYSTEM_CONTROL);
-+
-+ /* Wait max 20 ms */
-+ timeout = ktime_add_ms(ktime_get(), 20);
-+ val = ESDHC_CLOCK_STABLE;
-+ while (!(sdhci_readl(host, ESDHC_PRSSTAT) & val)) {
-+ if (ktime_after(ktime_get(), timeout)) {
-+ pr_err("%s: Internal clock never stabilised.\n",
-+ mmc_hostname(host->mmc));
-+ break;
-+ }
-+ udelay(10);
-+ }
-+}
-+
- static void esdhc_reset(struct sdhci_host *host, u8 mask)
- {
-+ u32 val;
-+
- sdhci_reset(host, mask);
-
- sdhci_writel(host, host->ier, SDHCI_INT_ENABLE);
- sdhci_writel(host, host->ier, SDHCI_SIGNAL_ENABLE);
-+
-+ if (mask & SDHCI_RESET_ALL) {
-+ val = sdhci_readl(host, ESDHC_TBCTL);
-+ val &= ~ESDHC_TB_EN;
-+ sdhci_writel(host, val, ESDHC_TBCTL);
-+ }
-+}
-+
-+/* The SCFG, Supplemental Configuration Unit, provides SoC specific
-+ * configuration and status registers for the device. There is a
-+ * SDHC IO VSEL control register on SCFG for some platforms. It's
-+ * used to support SDHC IO voltage switching.
-+ */
-+static const struct of_device_id scfg_device_ids[] = {
-+ { .compatible = "fsl,t1040-scfg", },
-+ { .compatible = "fsl,ls1012a-scfg", },
-+ { .compatible = "fsl,ls1046a-scfg", },
-+ {}
-+};
-+
-+/* SDHC IO VSEL control register definition */
-+#define SCFG_SDHCIOVSELCR 0x408
-+#define SDHCIOVSELCR_TGLEN 0x80000000
-+#define SDHCIOVSELCR_VSELVAL 0x60000000
-+#define SDHCIOVSELCR_SDHC_VS 0x00000001
-+
-+static int esdhc_signal_voltage_switch(struct mmc_host *mmc,
-+ struct mmc_ios *ios)
-+{
-+ struct sdhci_host *host = mmc_priv(mmc);
-+ struct device_node *scfg_node;
-+ void __iomem *scfg_base = NULL;
-+ u32 sdhciovselcr;
-+ u32 val;
-+
-+ /*
-+ * Signal Voltage Switching is only applicable for Host Controllers
-+ * v3.00 and above.
-+ */
-+ if (host->version < SDHCI_SPEC_300)
-+ return 0;
-+
-+ val = sdhci_readl(host, ESDHC_PROCTL);
-+
-+ switch (ios->signal_voltage) {
-+ case MMC_SIGNAL_VOLTAGE_330:
-+ val &= ~ESDHC_VOLT_SEL;
-+ sdhci_writel(host, val, ESDHC_PROCTL);
-+ return 0;
-+ case MMC_SIGNAL_VOLTAGE_180:
-+ scfg_node = of_find_matching_node(NULL, scfg_device_ids);
-+ if (scfg_node)
-+ scfg_base = of_iomap(scfg_node, 0);
-+ if (scfg_base) {
-+ sdhciovselcr = SDHCIOVSELCR_TGLEN |
-+ SDHCIOVSELCR_VSELVAL;
-+ iowrite32be(sdhciovselcr,
-+ scfg_base + SCFG_SDHCIOVSELCR);
-+
-+ val |= ESDHC_VOLT_SEL;
-+ sdhci_writel(host, val, ESDHC_PROCTL);
-+ mdelay(5);
-+
-+ sdhciovselcr = SDHCIOVSELCR_TGLEN |
-+ SDHCIOVSELCR_SDHC_VS;
-+ iowrite32be(sdhciovselcr,
-+ scfg_base + SCFG_SDHCIOVSELCR);
-+ iounmap(scfg_base);
-+ } else {
-+ val |= ESDHC_VOLT_SEL;
-+ sdhci_writel(host, val, ESDHC_PROCTL);
-+ }
-+ return 0;
-+ default:
-+ return 0;
-+ }
-+}
-+
-+static int esdhc_execute_tuning(struct mmc_host *mmc, u32 opcode)
-+{
-+ struct sdhci_host *host = mmc_priv(mmc);
-+ u32 val;
-+
-+ /* Use tuning block for tuning procedure */
-+ esdhc_clock_enable(host, false);
-+ val = sdhci_readl(host, ESDHC_DMA_SYSCTL);
-+ val |= ESDHC_FLUSH_ASYNC_FIFO;
-+ sdhci_writel(host, val, ESDHC_DMA_SYSCTL);
-+
-+ val = sdhci_readl(host, ESDHC_TBCTL);
-+ val |= ESDHC_TB_EN;
-+ sdhci_writel(host, val, ESDHC_TBCTL);
-+ esdhc_clock_enable(host, true);
-+
-+ return sdhci_execute_tuning(mmc, opcode);
- }
-
- #ifdef CONFIG_PM_SLEEP
-@@ -589,10 +780,19 @@ static const struct sdhci_pltfm_data sdh
- .ops = &sdhci_esdhc_le_ops,
- };
-
-+static struct soc_device_attribute soc_incorrect_hostver[] = {
-+ { .family = "QorIQ T4240", .revision = "1.0", },
-+ { .family = "QorIQ T4240", .revision = "2.0", },
-+ { },
-+};
-+
- static void esdhc_init(struct platform_device *pdev, struct sdhci_host *host)
- {
- struct sdhci_pltfm_host *pltfm_host;
- struct sdhci_esdhc *esdhc;
-+ struct device_node *np;
-+ struct clk *clk;
-+ u32 val;
- u16 host_ver;
-
- pltfm_host = sdhci_priv(host);
-@@ -602,6 +802,36 @@ static void esdhc_init(struct platform_d
- esdhc->vendor_ver = (host_ver & SDHCI_VENDOR_VER_MASK) >>
- SDHCI_VENDOR_VER_SHIFT;
- esdhc->spec_ver = host_ver & SDHCI_SPEC_VER_MASK;
-+ if (soc_device_match(soc_incorrect_hostver))
-+ esdhc->quirk_incorrect_hostver = true;
-+ else
-+ esdhc->quirk_incorrect_hostver = false;
-+
-+ np = pdev->dev.of_node;
-+ clk = of_clk_get(np, 0);
-+ if (!IS_ERR(clk)) {
-+ /*
-+ * esdhc->peripheral_clock would be assigned with a value
-+ * which is eSDHC base clock when use periperal clock.
-+ * For ls1046a, the clock value got by common clk API is
-+ * peripheral clock while the eSDHC base clock is 1/2
-+ * peripheral clock.
-+ */
-+ if (of_device_is_compatible(np, "fsl,ls1046a-esdhc"))
-+ esdhc->peripheral_clock = clk_get_rate(clk) / 2;
-+ else
-+ esdhc->peripheral_clock = clk_get_rate(clk);
-+
-+ clk_put(clk);
-+ }
-+
-+ if (esdhc->peripheral_clock) {
-+ esdhc_clock_enable(host, false);
-+ val = sdhci_readl(host, ESDHC_DMA_SYSCTL);
-+ val |= ESDHC_PERIPHERAL_CLK_SEL;
-+ sdhci_writel(host, val, ESDHC_DMA_SYSCTL);
-+ esdhc_clock_enable(host, true);
-+ }
- }
-
- static int sdhci_esdhc_probe(struct platform_device *pdev)
-@@ -624,6 +854,11 @@ static int sdhci_esdhc_probe(struct plat
- if (IS_ERR(host))
- return PTR_ERR(host);
-
-+ host->mmc_host_ops.start_signal_voltage_switch =
-+ esdhc_signal_voltage_switch;
-+ host->mmc_host_ops.execute_tuning = esdhc_execute_tuning;
-+ host->tuning_delay = 1;
-+
- esdhc_init(pdev, host);
-
- sdhci_get_of_property(pdev);
---- a/drivers/mmc/host/sdhci.c
-+++ b/drivers/mmc/host/sdhci.c
-@@ -1631,26 +1631,24 @@ static void sdhci_set_ios(struct mmc_hos
-
- ctrl = sdhci_readb(host, SDHCI_HOST_CONTROL);
-
-- if ((ios->timing == MMC_TIMING_SD_HS ||
-- ios->timing == MMC_TIMING_MMC_HS)
-- && !(host->quirks & SDHCI_QUIRK_NO_HISPD_BIT))
-- ctrl |= SDHCI_CTRL_HISPD;
-- else
-- ctrl &= ~SDHCI_CTRL_HISPD;
-+ if (!(host->quirks & SDHCI_QUIRK_NO_HISPD_BIT)) {
-+ if ((ios->timing == MMC_TIMING_SD_HS ||
-+ ios->timing == MMC_TIMING_MMC_HS ||
-+ ios->timing == MMC_TIMING_MMC_HS400 ||
-+ ios->timing == MMC_TIMING_MMC_HS200 ||
-+ ios->timing == MMC_TIMING_MMC_DDR52 ||
-+ ios->timing == MMC_TIMING_UHS_SDR50 ||
-+ ios->timing == MMC_TIMING_UHS_SDR104 ||
-+ ios->timing == MMC_TIMING_UHS_DDR50 ||
-+ ios->timing == MMC_TIMING_UHS_SDR25))
-+ ctrl |= SDHCI_CTRL_HISPD;
-+ else
-+ ctrl &= ~SDHCI_CTRL_HISPD;
-+ }
-
- if (host->version >= SDHCI_SPEC_300) {
- u16 clk, ctrl_2;
-
-- /* In case of UHS-I modes, set High Speed Enable */
-- if ((ios->timing == MMC_TIMING_MMC_HS400) ||
-- (ios->timing == MMC_TIMING_MMC_HS200) ||
-- (ios->timing == MMC_TIMING_MMC_DDR52) ||
-- (ios->timing == MMC_TIMING_UHS_SDR50) ||
-- (ios->timing == MMC_TIMING_UHS_SDR104) ||
-- (ios->timing == MMC_TIMING_UHS_DDR50) ||
-- (ios->timing == MMC_TIMING_UHS_SDR25))
-- ctrl |= SDHCI_CTRL_HISPD;
--
- if (!host->preset_enabled) {
- sdhci_writeb(host, ctrl, SDHCI_HOST_CONTROL);
- /*
-@@ -1963,7 +1961,7 @@ static int sdhci_prepare_hs400_tuning(st
- return 0;
- }
-
--static int sdhci_execute_tuning(struct mmc_host *mmc, u32 opcode)
-+int sdhci_execute_tuning(struct mmc_host *mmc, u32 opcode)
- {
- struct sdhci_host *host = mmc_priv(mmc);
- u16 ctrl;
-@@ -2022,6 +2020,9 @@ static int sdhci_execute_tuning(struct m
- return err;
- }
-
-+ if (host->tuning_delay < 0)
-+ host->tuning_delay = opcode == MMC_SEND_TUNING_BLOCK;
-+
- ctrl = sdhci_readw(host, SDHCI_HOST_CONTROL2);
- ctrl |= SDHCI_CTRL_EXEC_TUNING;
- if (host->quirks2 & SDHCI_QUIRK2_TUNING_WORK_AROUND)
-@@ -2134,9 +2135,10 @@ static int sdhci_execute_tuning(struct m
-
- ctrl = sdhci_readw(host, SDHCI_HOST_CONTROL2);
-
-- /* eMMC spec does not require a delay between tuning cycles */
-- if (opcode == MMC_SEND_TUNING_BLOCK)
-- mdelay(1);
-+ /* Spec does not require a delay between tuning cycles */
-+ if (host->tuning_delay > 0)
-+ mdelay(host->tuning_delay);
-+
- } while (ctrl & SDHCI_CTRL_EXEC_TUNING);
-
- /*
-@@ -2172,6 +2174,7 @@ out_unlock:
- spin_unlock_irqrestore(&host->lock, flags);
- return err;
- }
-+EXPORT_SYMBOL_GPL(sdhci_execute_tuning);
-
- static int sdhci_select_drive_strength(struct mmc_card *card,
- unsigned int max_dtr, int host_drv,
-@@ -3004,6 +3007,8 @@ struct sdhci_host *sdhci_alloc_host(stru
-
- host->flags = SDHCI_SIGNALING_330;
-
-+ host->tuning_delay = -1;
-+
- return host;
- }
-
---- a/drivers/mmc/host/sdhci.h
-+++ b/drivers/mmc/host/sdhci.h
-@@ -524,6 +524,8 @@ struct sdhci_host {
- #define SDHCI_TUNING_MODE_1 0
- #define SDHCI_TUNING_MODE_2 1
- #define SDHCI_TUNING_MODE_3 2
-+ /* Delay (ms) between tuning commands */
-+ int tuning_delay;
-
- unsigned long private[0] ____cacheline_aligned;
- };
-@@ -689,6 +691,7 @@ void sdhci_set_power_noreg(struct sdhci_
- void sdhci_set_bus_width(struct sdhci_host *host, int width);
- void sdhci_reset(struct sdhci_host *host, u8 mask);
- void sdhci_set_uhs_signaling(struct sdhci_host *host, unsigned timing);
-+int sdhci_execute_tuning(struct mmc_host *mmc, u32 opcode);
-
- #ifdef CONFIG_PM
- extern int sdhci_suspend_host(struct sdhci_host *host);
diff --git a/target/linux/layerscape/patches-4.9/813-qe-support-layerscape.patch b/target/linux/layerscape/patches-4.9/813-qe-support-layerscape.patch
deleted file mode 100644
index 1dfa46e839..0000000000
--- a/target/linux/layerscape/patches-4.9/813-qe-support-layerscape.patch
+++ /dev/null
@@ -1,1976 +0,0 @@
-From 2ab544f7e943c63c300933d34815e78451cc0c26 Mon Sep 17 00:00:00 2001
-From: Yangbo Lu <yangbo.lu@nxp.com>
-Date: Wed, 17 Jan 2018 15:37:56 +0800
-Subject: [PATCH 25/30] qe: support layerscape
-
-This is an integrated patch for layerscape qe support.
-
-Signed-off-by: Zhao Qiang <qiang.zhao@nxp.com>
-Signed-off-by: Yangbo Lu <yangbo.lu@nxp.com>
----
- drivers/{soc/fsl/qe/qe_ic.c => irqchip/irq-qeic.c} | 389 +++++++++++++--------
- drivers/net/wan/fsl_ucc_hdlc.c | 4 +-
- drivers/soc/fsl/qe/Kconfig | 2 +-
- drivers/soc/fsl/qe/Makefile | 2 +-
- drivers/soc/fsl/qe/qe.c | 80 +++--
- drivers/soc/fsl/qe/qe_ic.h | 103 ------
- drivers/soc/fsl/qe/qe_io.c | 42 +--
- drivers/soc/fsl/qe/qe_tdm.c | 8 +-
- drivers/soc/fsl/qe/ucc.c | 10 +-
- drivers/soc/fsl/qe/ucc_fast.c | 74 ++--
- drivers/tty/serial/ucc_uart.c | 1 +
- include/soc/fsl/qe/qe.h | 1 -
- include/soc/fsl/qe/qe_ic.h | 139 --------
- 13 files changed, 359 insertions(+), 496 deletions(-)
- rename drivers/{soc/fsl/qe/qe_ic.c => irqchip/irq-qeic.c} (54%)
- delete mode 100644 drivers/soc/fsl/qe/qe_ic.h
- delete mode 100644 include/soc/fsl/qe/qe_ic.h
-
---- a/drivers/soc/fsl/qe/qe_ic.c
-+++ /dev/null
-@@ -1,512 +0,0 @@
--/*
-- * arch/powerpc/sysdev/qe_lib/qe_ic.c
-- *
-- * Copyright (C) 2006 Freescale Semiconductor, Inc. All rights reserved.
-- *
-- * Author: Li Yang <leoli@freescale.com>
-- * Based on code from Shlomi Gridish <gridish@freescale.com>
-- *
-- * QUICC ENGINE Interrupt Controller
-- *
-- * This program is free software; you can redistribute it and/or modify it
-- * under the terms of the GNU General Public License as published by the
-- * Free Software Foundation; either version 2 of the License, or (at your
-- * option) any later version.
-- */
--
--#include <linux/of_irq.h>
--#include <linux/of_address.h>
--#include <linux/kernel.h>
--#include <linux/init.h>
--#include <linux/errno.h>
--#include <linux/reboot.h>
--#include <linux/slab.h>
--#include <linux/stddef.h>
--#include <linux/sched.h>
--#include <linux/signal.h>
--#include <linux/device.h>
--#include <linux/spinlock.h>
--#include <asm/irq.h>
--#include <asm/io.h>
--#include <soc/fsl/qe/qe_ic.h>
--
--#include "qe_ic.h"
--
--static DEFINE_RAW_SPINLOCK(qe_ic_lock);
--
--static struct qe_ic_info qe_ic_info[] = {
-- [1] = {
-- .mask = 0x00008000,
-- .mask_reg = QEIC_CIMR,
-- .pri_code = 0,
-- .pri_reg = QEIC_CIPWCC,
-- },
-- [2] = {
-- .mask = 0x00004000,
-- .mask_reg = QEIC_CIMR,
-- .pri_code = 1,
-- .pri_reg = QEIC_CIPWCC,
-- },
-- [3] = {
-- .mask = 0x00002000,
-- .mask_reg = QEIC_CIMR,
-- .pri_code = 2,
-- .pri_reg = QEIC_CIPWCC,
-- },
-- [10] = {
-- .mask = 0x00000040,
-- .mask_reg = QEIC_CIMR,
-- .pri_code = 1,
-- .pri_reg = QEIC_CIPZCC,
-- },
-- [11] = {
-- .mask = 0x00000020,
-- .mask_reg = QEIC_CIMR,
-- .pri_code = 2,
-- .pri_reg = QEIC_CIPZCC,
-- },
-- [12] = {
-- .mask = 0x00000010,
-- .mask_reg = QEIC_CIMR,
-- .pri_code = 3,
-- .pri_reg = QEIC_CIPZCC,
-- },
-- [13] = {
-- .mask = 0x00000008,
-- .mask_reg = QEIC_CIMR,
-- .pri_code = 4,
-- .pri_reg = QEIC_CIPZCC,
-- },
-- [14] = {
-- .mask = 0x00000004,
-- .mask_reg = QEIC_CIMR,
-- .pri_code = 5,
-- .pri_reg = QEIC_CIPZCC,
-- },
-- [15] = {
-- .mask = 0x00000002,
-- .mask_reg = QEIC_CIMR,
-- .pri_code = 6,
-- .pri_reg = QEIC_CIPZCC,
-- },
-- [20] = {
-- .mask = 0x10000000,
-- .mask_reg = QEIC_CRIMR,
-- .pri_code = 3,
-- .pri_reg = QEIC_CIPRTA,
-- },
-- [25] = {
-- .mask = 0x00800000,
-- .mask_reg = QEIC_CRIMR,
-- .pri_code = 0,
-- .pri_reg = QEIC_CIPRTB,
-- },
-- [26] = {
-- .mask = 0x00400000,
-- .mask_reg = QEIC_CRIMR,
-- .pri_code = 1,
-- .pri_reg = QEIC_CIPRTB,
-- },
-- [27] = {
-- .mask = 0x00200000,
-- .mask_reg = QEIC_CRIMR,
-- .pri_code = 2,
-- .pri_reg = QEIC_CIPRTB,
-- },
-- [28] = {
-- .mask = 0x00100000,
-- .mask_reg = QEIC_CRIMR,
-- .pri_code = 3,
-- .pri_reg = QEIC_CIPRTB,
-- },
-- [32] = {
-- .mask = 0x80000000,
-- .mask_reg = QEIC_CIMR,
-- .pri_code = 0,
-- .pri_reg = QEIC_CIPXCC,
-- },
-- [33] = {
-- .mask = 0x40000000,
-- .mask_reg = QEIC_CIMR,
-- .pri_code = 1,
-- .pri_reg = QEIC_CIPXCC,
-- },
-- [34] = {
-- .mask = 0x20000000,
-- .mask_reg = QEIC_CIMR,
-- .pri_code = 2,
-- .pri_reg = QEIC_CIPXCC,
-- },
-- [35] = {
-- .mask = 0x10000000,
-- .mask_reg = QEIC_CIMR,
-- .pri_code = 3,
-- .pri_reg = QEIC_CIPXCC,
-- },
-- [36] = {
-- .mask = 0x08000000,
-- .mask_reg = QEIC_CIMR,
-- .pri_code = 4,
-- .pri_reg = QEIC_CIPXCC,
-- },
-- [40] = {
-- .mask = 0x00800000,
-- .mask_reg = QEIC_CIMR,
-- .pri_code = 0,
-- .pri_reg = QEIC_CIPYCC,
-- },
-- [41] = {
-- .mask = 0x00400000,
-- .mask_reg = QEIC_CIMR,
-- .pri_code = 1,
-- .pri_reg = QEIC_CIPYCC,
-- },
-- [42] = {
-- .mask = 0x00200000,
-- .mask_reg = QEIC_CIMR,
-- .pri_code = 2,
-- .pri_reg = QEIC_CIPYCC,
-- },
-- [43] = {
-- .mask = 0x00100000,
-- .mask_reg = QEIC_CIMR,
-- .pri_code = 3,
-- .pri_reg = QEIC_CIPYCC,
-- },
--};
--
--static inline u32 qe_ic_read(volatile __be32 __iomem * base, unsigned int reg)
--{
-- return in_be32(base + (reg >> 2));
--}
--
--static inline void qe_ic_write(volatile __be32 __iomem * base, unsigned int reg,
-- u32 value)
--{
-- out_be32(base + (reg >> 2), value);
--}
--
--static inline struct qe_ic *qe_ic_from_irq(unsigned int virq)
--{
-- return irq_get_chip_data(virq);
--}
--
--static inline struct qe_ic *qe_ic_from_irq_data(struct irq_data *d)
--{
-- return irq_data_get_irq_chip_data(d);
--}
--
--static void qe_ic_unmask_irq(struct irq_data *d)
--{
-- struct qe_ic *qe_ic = qe_ic_from_irq_data(d);
-- unsigned int src = irqd_to_hwirq(d);
-- unsigned long flags;
-- u32 temp;
--
-- raw_spin_lock_irqsave(&qe_ic_lock, flags);
--
-- temp = qe_ic_read(qe_ic->regs, qe_ic_info[src].mask_reg);
-- qe_ic_write(qe_ic->regs, qe_ic_info[src].mask_reg,
-- temp | qe_ic_info[src].mask);
--
-- raw_spin_unlock_irqrestore(&qe_ic_lock, flags);
--}
--
--static void qe_ic_mask_irq(struct irq_data *d)
--{
-- struct qe_ic *qe_ic = qe_ic_from_irq_data(d);
-- unsigned int src = irqd_to_hwirq(d);
-- unsigned long flags;
-- u32 temp;
--
-- raw_spin_lock_irqsave(&qe_ic_lock, flags);
--
-- temp = qe_ic_read(qe_ic->regs, qe_ic_info[src].mask_reg);
-- qe_ic_write(qe_ic->regs, qe_ic_info[src].mask_reg,
-- temp & ~qe_ic_info[src].mask);
--
-- /* Flush the above write before enabling interrupts; otherwise,
-- * spurious interrupts will sometimes happen. To be 100% sure
-- * that the write has reached the device before interrupts are
-- * enabled, the mask register would have to be read back; however,
-- * this is not required for correctness, only to avoid wasting
-- * time on a large number of spurious interrupts. In testing,
-- * a sync reduced the observed spurious interrupts to zero.
-- */
-- mb();
--
-- raw_spin_unlock_irqrestore(&qe_ic_lock, flags);
--}
--
--static struct irq_chip qe_ic_irq_chip = {
-- .name = "QEIC",
-- .irq_unmask = qe_ic_unmask_irq,
-- .irq_mask = qe_ic_mask_irq,
-- .irq_mask_ack = qe_ic_mask_irq,
--};
--
--static int qe_ic_host_match(struct irq_domain *h, struct device_node *node,
-- enum irq_domain_bus_token bus_token)
--{
-- /* Exact match, unless qe_ic node is NULL */
-- struct device_node *of_node = irq_domain_get_of_node(h);
-- return of_node == NULL || of_node == node;
--}
--
--static int qe_ic_host_map(struct irq_domain *h, unsigned int virq,
-- irq_hw_number_t hw)
--{
-- struct qe_ic *qe_ic = h->host_data;
-- struct irq_chip *chip;
--
-- if (hw >= ARRAY_SIZE(qe_ic_info)) {
-- pr_err("%s: Invalid hw irq number for QEIC\n", __func__);
-- return -EINVAL;
-- }
--
-- if (qe_ic_info[hw].mask == 0) {
-- printk(KERN_ERR "Can't map reserved IRQ\n");
-- return -EINVAL;
-- }
-- /* Default chip */
-- chip = &qe_ic->hc_irq;
--
-- irq_set_chip_data(virq, qe_ic);
-- irq_set_status_flags(virq, IRQ_LEVEL);
--
-- irq_set_chip_and_handler(virq, chip, handle_level_irq);
--
-- return 0;
--}
--
--static const struct irq_domain_ops qe_ic_host_ops = {
-- .match = qe_ic_host_match,
-- .map = qe_ic_host_map,
-- .xlate = irq_domain_xlate_onetwocell,
--};
--
--/* Return an interrupt vector or NO_IRQ if no interrupt is pending. */
--unsigned int qe_ic_get_low_irq(struct qe_ic *qe_ic)
--{
-- int irq;
--
-- BUG_ON(qe_ic == NULL);
--
-- /* get the interrupt source vector. */
-- irq = qe_ic_read(qe_ic->regs, QEIC_CIVEC) >> 26;
--
-- if (irq == 0)
-- return NO_IRQ;
--
-- return irq_linear_revmap(qe_ic->irqhost, irq);
--}
--
--/* Return an interrupt vector or NO_IRQ if no interrupt is pending. */
--unsigned int qe_ic_get_high_irq(struct qe_ic *qe_ic)
--{
-- int irq;
--
-- BUG_ON(qe_ic == NULL);
--
-- /* get the interrupt source vector. */
-- irq = qe_ic_read(qe_ic->regs, QEIC_CHIVEC) >> 26;
--
-- if (irq == 0)
-- return NO_IRQ;
--
-- return irq_linear_revmap(qe_ic->irqhost, irq);
--}
--
--void __init qe_ic_init(struct device_node *node, unsigned int flags,
-- void (*low_handler)(struct irq_desc *desc),
-- void (*high_handler)(struct irq_desc *desc))
--{
-- struct qe_ic *qe_ic;
-- struct resource res;
-- u32 temp = 0, ret, high_active = 0;
--
-- ret = of_address_to_resource(node, 0, &res);
-- if (ret)
-- return;
--
-- qe_ic = kzalloc(sizeof(*qe_ic), GFP_KERNEL);
-- if (qe_ic == NULL)
-- return;
--
-- qe_ic->irqhost = irq_domain_add_linear(node, NR_QE_IC_INTS,
-- &qe_ic_host_ops, qe_ic);
-- if (qe_ic->irqhost == NULL) {
-- kfree(qe_ic);
-- return;
-- }
--
-- qe_ic->regs = ioremap(res.start, resource_size(&res));
--
-- qe_ic->hc_irq = qe_ic_irq_chip;
--
-- qe_ic->virq_high = irq_of_parse_and_map(node, 0);
-- qe_ic->virq_low = irq_of_parse_and_map(node, 1);
--
-- if (qe_ic->virq_low == NO_IRQ) {
-- printk(KERN_ERR "Failed to map QE_IC low IRQ\n");
-- kfree(qe_ic);
-- return;
-- }
--
-- /* default priority scheme is grouped. If spread mode is */
-- /* required, configure cicr accordingly. */
-- if (flags & QE_IC_SPREADMODE_GRP_W)
-- temp |= CICR_GWCC;
-- if (flags & QE_IC_SPREADMODE_GRP_X)
-- temp |= CICR_GXCC;
-- if (flags & QE_IC_SPREADMODE_GRP_Y)
-- temp |= CICR_GYCC;
-- if (flags & QE_IC_SPREADMODE_GRP_Z)
-- temp |= CICR_GZCC;
-- if (flags & QE_IC_SPREADMODE_GRP_RISCA)
-- temp |= CICR_GRTA;
-- if (flags & QE_IC_SPREADMODE_GRP_RISCB)
-- temp |= CICR_GRTB;
--
-- /* choose destination signal for highest priority interrupt */
-- if (flags & QE_IC_HIGH_SIGNAL) {
-- temp |= (SIGNAL_HIGH << CICR_HPIT_SHIFT);
-- high_active = 1;
-- }
--
-- qe_ic_write(qe_ic->regs, QEIC_CICR, temp);
--
-- irq_set_handler_data(qe_ic->virq_low, qe_ic);
-- irq_set_chained_handler(qe_ic->virq_low, low_handler);
--
-- if (qe_ic->virq_high != NO_IRQ &&
-- qe_ic->virq_high != qe_ic->virq_low) {
-- irq_set_handler_data(qe_ic->virq_high, qe_ic);
-- irq_set_chained_handler(qe_ic->virq_high, high_handler);
-- }
--}
--
--void qe_ic_set_highest_priority(unsigned int virq, int high)
--{
-- struct qe_ic *qe_ic = qe_ic_from_irq(virq);
-- unsigned int src = virq_to_hw(virq);
-- u32 temp = 0;
--
-- temp = qe_ic_read(qe_ic->regs, QEIC_CICR);
--
-- temp &= ~CICR_HP_MASK;
-- temp |= src << CICR_HP_SHIFT;
--
-- temp &= ~CICR_HPIT_MASK;
-- temp |= (high ? SIGNAL_HIGH : SIGNAL_LOW) << CICR_HPIT_SHIFT;
--
-- qe_ic_write(qe_ic->regs, QEIC_CICR, temp);
--}
--
--/* Set Priority level within its group, from 1 to 8 */
--int qe_ic_set_priority(unsigned int virq, unsigned int priority)
--{
-- struct qe_ic *qe_ic = qe_ic_from_irq(virq);
-- unsigned int src = virq_to_hw(virq);
-- u32 temp;
--
-- if (priority > 8 || priority == 0)
-- return -EINVAL;
-- if (WARN_ONCE(src >= ARRAY_SIZE(qe_ic_info),
-- "%s: Invalid hw irq number for QEIC\n", __func__))
-- return -EINVAL;
-- if (qe_ic_info[src].pri_reg == 0)
-- return -EINVAL;
--
-- temp = qe_ic_read(qe_ic->regs, qe_ic_info[src].pri_reg);
--
-- if (priority < 4) {
-- temp &= ~(0x7 << (32 - priority * 3));
-- temp |= qe_ic_info[src].pri_code << (32 - priority * 3);
-- } else {
-- temp &= ~(0x7 << (24 - priority * 3));
-- temp |= qe_ic_info[src].pri_code << (24 - priority * 3);
-- }
--
-- qe_ic_write(qe_ic->regs, qe_ic_info[src].pri_reg, temp);
--
-- return 0;
--}
--
--/* Set a QE priority to use high irq, only priority 1~2 can use high irq */
--int qe_ic_set_high_priority(unsigned int virq, unsigned int priority, int high)
--{
-- struct qe_ic *qe_ic = qe_ic_from_irq(virq);
-- unsigned int src = virq_to_hw(virq);
-- u32 temp, control_reg = QEIC_CICNR, shift = 0;
--
-- if (priority > 2 || priority == 0)
-- return -EINVAL;
-- if (WARN_ONCE(src >= ARRAY_SIZE(qe_ic_info),
-- "%s: Invalid hw irq number for QEIC\n", __func__))
-- return -EINVAL;
--
-- switch (qe_ic_info[src].pri_reg) {
-- case QEIC_CIPZCC:
-- shift = CICNR_ZCC1T_SHIFT;
-- break;
-- case QEIC_CIPWCC:
-- shift = CICNR_WCC1T_SHIFT;
-- break;
-- case QEIC_CIPYCC:
-- shift = CICNR_YCC1T_SHIFT;
-- break;
-- case QEIC_CIPXCC:
-- shift = CICNR_XCC1T_SHIFT;
-- break;
-- case QEIC_CIPRTA:
-- shift = CRICR_RTA1T_SHIFT;
-- control_reg = QEIC_CRICR;
-- break;
-- case QEIC_CIPRTB:
-- shift = CRICR_RTB1T_SHIFT;
-- control_reg = QEIC_CRICR;
-- break;
-- default:
-- return -EINVAL;
-- }
--
-- shift += (2 - priority) * 2;
-- temp = qe_ic_read(qe_ic->regs, control_reg);
-- temp &= ~(SIGNAL_MASK << shift);
-- temp |= (high ? SIGNAL_HIGH : SIGNAL_LOW) << shift;
-- qe_ic_write(qe_ic->regs, control_reg, temp);
--
-- return 0;
--}
--
--static struct bus_type qe_ic_subsys = {
-- .name = "qe_ic",
-- .dev_name = "qe_ic",
--};
--
--static struct device device_qe_ic = {
-- .id = 0,
-- .bus = &qe_ic_subsys,
--};
--
--static int __init init_qe_ic_sysfs(void)
--{
-- int rc;
--
-- printk(KERN_DEBUG "Registering qe_ic with sysfs...\n");
--
-- rc = subsys_system_register(&qe_ic_subsys, NULL);
-- if (rc) {
-- printk(KERN_ERR "Failed registering qe_ic sys class\n");
-- return -ENODEV;
-- }
-- rc = device_register(&device_qe_ic);
-- if (rc) {
-- printk(KERN_ERR "Failed registering qe_ic sys device\n");
-- return -ENODEV;
-- }
-- return 0;
--}
--
--subsys_initcall(init_qe_ic_sysfs);
---- /dev/null
-+++ b/drivers/irqchip/irq-qeic.c
-@@ -0,0 +1,605 @@
-+/*
-+ * drivers/irqchip/irq-qeic.c
-+ *
-+ * Copyright (C) 2016 Freescale Semiconductor, Inc. All rights reserved.
-+ *
-+ * Author: Li Yang <leoli@freescale.com>
-+ * Based on code from Shlomi Gridish <gridish@freescale.com>
-+ *
-+ * QUICC ENGINE Interrupt Controller
-+ *
-+ * This program is free software; you can redistribute it and/or modify it
-+ * under the terms of the GNU General Public License as published by the
-+ * Free Software Foundation; either version 2 of the License, or (at your
-+ * option) any later version.
-+ */
-+
-+#include <linux/of_irq.h>
-+#include <linux/of_address.h>
-+#include <linux/kernel.h>
-+#include <linux/init.h>
-+#include <linux/irqdomain.h>
-+#include <linux/irqchip.h>
-+#include <linux/errno.h>
-+#include <linux/of_address.h>
-+#include <linux/of_irq.h>
-+#include <linux/reboot.h>
-+#include <linux/slab.h>
-+#include <linux/stddef.h>
-+#include <linux/sched.h>
-+#include <linux/signal.h>
-+#include <linux/device.h>
-+#include <linux/spinlock.h>
-+#include <linux/irq.h>
-+#include <asm/io.h>
-+
-+#define NR_QE_IC_INTS 64
-+
-+/* QE IC registers offset */
-+#define QEIC_CICR 0x00
-+#define QEIC_CIVEC 0x04
-+#define QEIC_CRIPNR 0x08
-+#define QEIC_CIPNR 0x0c
-+#define QEIC_CIPXCC 0x10
-+#define QEIC_CIPYCC 0x14
-+#define QEIC_CIPWCC 0x18
-+#define QEIC_CIPZCC 0x1c
-+#define QEIC_CIMR 0x20
-+#define QEIC_CRIMR 0x24
-+#define QEIC_CICNR 0x28
-+#define QEIC_CIPRTA 0x30
-+#define QEIC_CIPRTB 0x34
-+#define QEIC_CRICR 0x3c
-+#define QEIC_CHIVEC 0x60
-+
-+/* Interrupt priority registers */
-+#define CIPCC_SHIFT_PRI0 29
-+#define CIPCC_SHIFT_PRI1 26
-+#define CIPCC_SHIFT_PRI2 23
-+#define CIPCC_SHIFT_PRI3 20
-+#define CIPCC_SHIFT_PRI4 13
-+#define CIPCC_SHIFT_PRI5 10
-+#define CIPCC_SHIFT_PRI6 7
-+#define CIPCC_SHIFT_PRI7 4
-+
-+/* CICR priority modes */
-+#define CICR_GWCC 0x00040000
-+#define CICR_GXCC 0x00020000
-+#define CICR_GYCC 0x00010000
-+#define CICR_GZCC 0x00080000
-+#define CICR_GRTA 0x00200000
-+#define CICR_GRTB 0x00400000
-+#define CICR_HPIT_SHIFT 8
-+#define CICR_HPIT_MASK 0x00000300
-+#define CICR_HP_SHIFT 24
-+#define CICR_HP_MASK 0x3f000000
-+
-+/* CICNR */
-+#define CICNR_WCC1T_SHIFT 20
-+#define CICNR_ZCC1T_SHIFT 28
-+#define CICNR_YCC1T_SHIFT 12
-+#define CICNR_XCC1T_SHIFT 4
-+
-+/* CRICR */
-+#define CRICR_RTA1T_SHIFT 20
-+#define CRICR_RTB1T_SHIFT 28
-+
-+/* Signal indicator */
-+#define SIGNAL_MASK 3
-+#define SIGNAL_HIGH 2
-+#define SIGNAL_LOW 0
-+
-+#define NUM_OF_QE_IC_GROUPS 6
-+
-+/* Flags when we init the QE IC */
-+#define QE_IC_SPREADMODE_GRP_W 0x00000001
-+#define QE_IC_SPREADMODE_GRP_X 0x00000002
-+#define QE_IC_SPREADMODE_GRP_Y 0x00000004
-+#define QE_IC_SPREADMODE_GRP_Z 0x00000008
-+#define QE_IC_SPREADMODE_GRP_RISCA 0x00000010
-+#define QE_IC_SPREADMODE_GRP_RISCB 0x00000020
-+
-+#define QE_IC_LOW_SIGNAL 0x00000100
-+#define QE_IC_HIGH_SIGNAL 0x00000200
-+
-+#define QE_IC_GRP_W_PRI0_DEST_SIGNAL_HIGH 0x00001000
-+#define QE_IC_GRP_W_PRI1_DEST_SIGNAL_HIGH 0x00002000
-+#define QE_IC_GRP_X_PRI0_DEST_SIGNAL_HIGH 0x00004000
-+#define QE_IC_GRP_X_PRI1_DEST_SIGNAL_HIGH 0x00008000
-+#define QE_IC_GRP_Y_PRI0_DEST_SIGNAL_HIGH 0x00010000
-+#define QE_IC_GRP_Y_PRI1_DEST_SIGNAL_HIGH 0x00020000
-+#define QE_IC_GRP_Z_PRI0_DEST_SIGNAL_HIGH 0x00040000
-+#define QE_IC_GRP_Z_PRI1_DEST_SIGNAL_HIGH 0x00080000
-+#define QE_IC_GRP_RISCA_PRI0_DEST_SIGNAL_HIGH 0x00100000
-+#define QE_IC_GRP_RISCA_PRI1_DEST_SIGNAL_HIGH 0x00200000
-+#define QE_IC_GRP_RISCB_PRI0_DEST_SIGNAL_HIGH 0x00400000
-+#define QE_IC_GRP_RISCB_PRI1_DEST_SIGNAL_HIGH 0x00800000
-+#define QE_IC_GRP_W_DEST_SIGNAL_SHIFT (12)
-+
-+/* QE interrupt sources groups */
-+enum qe_ic_grp_id {
-+ QE_IC_GRP_W = 0, /* QE interrupt controller group W */
-+ QE_IC_GRP_X, /* QE interrupt controller group X */
-+ QE_IC_GRP_Y, /* QE interrupt controller group Y */
-+ QE_IC_GRP_Z, /* QE interrupt controller group Z */
-+ QE_IC_GRP_RISCA, /* QE interrupt controller RISC group A */
-+ QE_IC_GRP_RISCB /* QE interrupt controller RISC group B */
-+};
-+
-+struct qe_ic {
-+ /* Control registers offset */
-+ u32 __iomem *regs;
-+
-+ /* The remapper for this QEIC */
-+ struct irq_domain *irqhost;
-+
-+ /* The "linux" controller struct */
-+ struct irq_chip hc_irq;
-+
-+ /* VIRQ numbers of QE high/low irqs */
-+ unsigned int virq_high;
-+ unsigned int virq_low;
-+};
-+
-+/*
-+ * QE interrupt controller internal structure
-+ */
-+struct qe_ic_info {
-+ /* location of this source at the QIMR register. */
-+ u32 mask;
-+
-+ /* Mask register offset */
-+ u32 mask_reg;
-+
-+ /*
-+ * for grouped interrupts sources - the interrupt
-+ * code as appears at the group priority register
-+ */
-+ u8 pri_code;
-+
-+ /* Group priority register offset */
-+ u32 pri_reg;
-+};
-+
-+static DEFINE_RAW_SPINLOCK(qe_ic_lock);
-+
-+static struct qe_ic_info qe_ic_info[] = {
-+ [1] = {
-+ .mask = 0x00008000,
-+ .mask_reg = QEIC_CIMR,
-+ .pri_code = 0,
-+ .pri_reg = QEIC_CIPWCC,
-+ },
-+ [2] = {
-+ .mask = 0x00004000,
-+ .mask_reg = QEIC_CIMR,
-+ .pri_code = 1,
-+ .pri_reg = QEIC_CIPWCC,
-+ },
-+ [3] = {
-+ .mask = 0x00002000,
-+ .mask_reg = QEIC_CIMR,
-+ .pri_code = 2,
-+ .pri_reg = QEIC_CIPWCC,
-+ },
-+ [10] = {
-+ .mask = 0x00000040,
-+ .mask_reg = QEIC_CIMR,
-+ .pri_code = 1,
-+ .pri_reg = QEIC_CIPZCC,
-+ },
-+ [11] = {
-+ .mask = 0x00000020,
-+ .mask_reg = QEIC_CIMR,
-+ .pri_code = 2,
-+ .pri_reg = QEIC_CIPZCC,
-+ },
-+ [12] = {
-+ .mask = 0x00000010,
-+ .mask_reg = QEIC_CIMR,
-+ .pri_code = 3,
-+ .pri_reg = QEIC_CIPZCC,
-+ },
-+ [13] = {
-+ .mask = 0x00000008,
-+ .mask_reg = QEIC_CIMR,
-+ .pri_code = 4,
-+ .pri_reg = QEIC_CIPZCC,
-+ },
-+ [14] = {
-+ .mask = 0x00000004,
-+ .mask_reg = QEIC_CIMR,
-+ .pri_code = 5,
-+ .pri_reg = QEIC_CIPZCC,
-+ },
-+ [15] = {
-+ .mask = 0x00000002,
-+ .mask_reg = QEIC_CIMR,
-+ .pri_code = 6,
-+ .pri_reg = QEIC_CIPZCC,
-+ },
-+ [20] = {
-+ .mask = 0x10000000,
-+ .mask_reg = QEIC_CRIMR,
-+ .pri_code = 3,
-+ .pri_reg = QEIC_CIPRTA,
-+ },
-+ [25] = {
-+ .mask = 0x00800000,
-+ .mask_reg = QEIC_CRIMR,
-+ .pri_code = 0,
-+ .pri_reg = QEIC_CIPRTB,
-+ },
-+ [26] = {
-+ .mask = 0x00400000,
-+ .mask_reg = QEIC_CRIMR,
-+ .pri_code = 1,
-+ .pri_reg = QEIC_CIPRTB,
-+ },
-+ [27] = {
-+ .mask = 0x00200000,
-+ .mask_reg = QEIC_CRIMR,
-+ .pri_code = 2,
-+ .pri_reg = QEIC_CIPRTB,
-+ },
-+ [28] = {
-+ .mask = 0x00100000,
-+ .mask_reg = QEIC_CRIMR,
-+ .pri_code = 3,
-+ .pri_reg = QEIC_CIPRTB,
-+ },
-+ [32] = {
-+ .mask = 0x80000000,
-+ .mask_reg = QEIC_CIMR,
-+ .pri_code = 0,
-+ .pri_reg = QEIC_CIPXCC,
-+ },
-+ [33] = {
-+ .mask = 0x40000000,
-+ .mask_reg = QEIC_CIMR,
-+ .pri_code = 1,
-+ .pri_reg = QEIC_CIPXCC,
-+ },
-+ [34] = {
-+ .mask = 0x20000000,
-+ .mask_reg = QEIC_CIMR,
-+ .pri_code = 2,
-+ .pri_reg = QEIC_CIPXCC,
-+ },
-+ [35] = {
-+ .mask = 0x10000000,
-+ .mask_reg = QEIC_CIMR,
-+ .pri_code = 3,
-+ .pri_reg = QEIC_CIPXCC,
-+ },
-+ [36] = {
-+ .mask = 0x08000000,
-+ .mask_reg = QEIC_CIMR,
-+ .pri_code = 4,
-+ .pri_reg = QEIC_CIPXCC,
-+ },
-+ [40] = {
-+ .mask = 0x00800000,
-+ .mask_reg = QEIC_CIMR,
-+ .pri_code = 0,
-+ .pri_reg = QEIC_CIPYCC,
-+ },
-+ [41] = {
-+ .mask = 0x00400000,
-+ .mask_reg = QEIC_CIMR,
-+ .pri_code = 1,
-+ .pri_reg = QEIC_CIPYCC,
-+ },
-+ [42] = {
-+ .mask = 0x00200000,
-+ .mask_reg = QEIC_CIMR,
-+ .pri_code = 2,
-+ .pri_reg = QEIC_CIPYCC,
-+ },
-+ [43] = {
-+ .mask = 0x00100000,
-+ .mask_reg = QEIC_CIMR,
-+ .pri_code = 3,
-+ .pri_reg = QEIC_CIPYCC,
-+ },
-+};
-+
-+static inline u32 qe_ic_read(__be32 __iomem *base, unsigned int reg)
-+{
-+ return ioread32be(base + (reg >> 2));
-+}
-+
-+static inline void qe_ic_write(__be32 __iomem *base, unsigned int reg,
-+ u32 value)
-+{
-+ iowrite32be(value, base + (reg >> 2));
-+}
-+
-+static inline struct qe_ic *qe_ic_from_irq(unsigned int virq)
-+{
-+ return irq_get_chip_data(virq);
-+}
-+
-+static inline struct qe_ic *qe_ic_from_irq_data(struct irq_data *d)
-+{
-+ return irq_data_get_irq_chip_data(d);
-+}
-+
-+static void qe_ic_unmask_irq(struct irq_data *d)
-+{
-+ struct qe_ic *qe_ic = qe_ic_from_irq_data(d);
-+ unsigned int src = irqd_to_hwirq(d);
-+ unsigned long flags;
-+ u32 temp;
-+
-+ raw_spin_lock_irqsave(&qe_ic_lock, flags);
-+
-+ temp = qe_ic_read(qe_ic->regs, qe_ic_info[src].mask_reg);
-+ qe_ic_write(qe_ic->regs, qe_ic_info[src].mask_reg,
-+ temp | qe_ic_info[src].mask);
-+
-+ raw_spin_unlock_irqrestore(&qe_ic_lock, flags);
-+}
-+
-+static void qe_ic_mask_irq(struct irq_data *d)
-+{
-+ struct qe_ic *qe_ic = qe_ic_from_irq_data(d);
-+ unsigned int src = irqd_to_hwirq(d);
-+ unsigned long flags;
-+ u32 temp;
-+
-+ raw_spin_lock_irqsave(&qe_ic_lock, flags);
-+
-+ temp = qe_ic_read(qe_ic->regs, qe_ic_info[src].mask_reg);
-+ qe_ic_write(qe_ic->regs, qe_ic_info[src].mask_reg,
-+ temp & ~qe_ic_info[src].mask);
-+
-+ /* Flush the above write before enabling interrupts; otherwise,
-+ * spurious interrupts will sometimes happen. To be 100% sure
-+ * that the write has reached the device before interrupts are
-+ * enabled, the mask register would have to be read back; however,
-+ * this is not required for correctness, only to avoid wasting
-+ * time on a large number of spurious interrupts. In testing,
-+ * a sync reduced the observed spurious interrupts to zero.
-+ */
-+ mb();
-+
-+ raw_spin_unlock_irqrestore(&qe_ic_lock, flags);
-+}
-+
-+static struct irq_chip qe_ic_irq_chip = {
-+ .name = "QEIC",
-+ .irq_unmask = qe_ic_unmask_irq,
-+ .irq_mask = qe_ic_mask_irq,
-+ .irq_mask_ack = qe_ic_mask_irq,
-+};
-+
-+static int qe_ic_host_match(struct irq_domain *h, struct device_node *node,
-+ enum irq_domain_bus_token bus_token)
-+{
-+ /* Exact match, unless qe_ic node is NULL */
-+ struct device_node *of_node = irq_domain_get_of_node(h);
-+ return of_node == NULL || of_node == node;
-+}
-+
-+static int qe_ic_host_map(struct irq_domain *h, unsigned int virq,
-+ irq_hw_number_t hw)
-+{
-+ struct qe_ic *qe_ic = h->host_data;
-+ struct irq_chip *chip;
-+
-+ if (hw >= ARRAY_SIZE(qe_ic_info)) {
-+ pr_err("%s: Invalid hw irq number for QEIC\n", __func__);
-+ return -EINVAL;
-+ }
-+
-+ if (qe_ic_info[hw].mask == 0) {
-+ printk(KERN_ERR "Can't map reserved IRQ\n");
-+ return -EINVAL;
-+ }
-+ /* Default chip */
-+ chip = &qe_ic->hc_irq;
-+
-+ irq_set_chip_data(virq, qe_ic);
-+ irq_set_status_flags(virq, IRQ_LEVEL);
-+
-+ irq_set_chip_and_handler(virq, chip, handle_level_irq);
-+
-+ return 0;
-+}
-+
-+static const struct irq_domain_ops qe_ic_host_ops = {
-+ .match = qe_ic_host_match,
-+ .map = qe_ic_host_map,
-+ .xlate = irq_domain_xlate_onetwocell,
-+};
-+
-+/* Return an interrupt vector or 0 if no interrupt is pending. */
-+static unsigned int qe_ic_get_low_irq(struct qe_ic *qe_ic)
-+{
-+ int irq;
-+
-+ BUG_ON(qe_ic == NULL);
-+
-+ /* get the interrupt source vector. */
-+ irq = qe_ic_read(qe_ic->regs, QEIC_CIVEC) >> 26;
-+
-+ if (irq == 0)
-+ return 0;
-+
-+ return irq_linear_revmap(qe_ic->irqhost, irq);
-+}
-+
-+/* Return an interrupt vector or 0 if no interrupt is pending. */
-+static unsigned int qe_ic_get_high_irq(struct qe_ic *qe_ic)
-+{
-+ int irq;
-+
-+ BUG_ON(qe_ic == NULL);
-+
-+ /* get the interrupt source vector. */
-+ irq = qe_ic_read(qe_ic->regs, QEIC_CHIVEC) >> 26;
-+
-+ if (irq == 0)
-+ return 0;
-+
-+ return irq_linear_revmap(qe_ic->irqhost, irq);
-+}
-+
-+static inline void qe_ic_cascade_low_ipic(struct irq_desc *desc)
-+{
-+ struct qe_ic *qe_ic = irq_desc_get_handler_data(desc);
-+ unsigned int cascade_irq = qe_ic_get_low_irq(qe_ic);
-+
-+ if (cascade_irq != 0)
-+ generic_handle_irq(cascade_irq);
-+}
-+
-+static inline void qe_ic_cascade_high_ipic(struct irq_desc *desc)
-+{
-+ struct qe_ic *qe_ic = irq_desc_get_handler_data(desc);
-+ unsigned int cascade_irq = qe_ic_get_high_irq(qe_ic);
-+
-+ if (cascade_irq != 0)
-+ generic_handle_irq(cascade_irq);
-+}
-+
-+static inline void qe_ic_cascade_low_mpic(struct irq_desc *desc)
-+{
-+ struct qe_ic *qe_ic = irq_desc_get_handler_data(desc);
-+ unsigned int cascade_irq = qe_ic_get_low_irq(qe_ic);
-+ struct irq_chip *chip = irq_desc_get_chip(desc);
-+
-+ if (cascade_irq != 0)
-+ generic_handle_irq(cascade_irq);
-+
-+ chip->irq_eoi(&desc->irq_data);
-+}
-+
-+static inline void qe_ic_cascade_high_mpic(struct irq_desc *desc)
-+{
-+ struct qe_ic *qe_ic = irq_desc_get_handler_data(desc);
-+ unsigned int cascade_irq = qe_ic_get_high_irq(qe_ic);
-+ struct irq_chip *chip = irq_desc_get_chip(desc);
-+
-+ if (cascade_irq != 0)
-+ generic_handle_irq(cascade_irq);
-+
-+ chip->irq_eoi(&desc->irq_data);
-+}
-+
-+static inline void qe_ic_cascade_muxed_mpic(struct irq_desc *desc)
-+{
-+ struct qe_ic *qe_ic = irq_desc_get_handler_data(desc);
-+ unsigned int cascade_irq;
-+ struct irq_chip *chip = irq_desc_get_chip(desc);
-+
-+ cascade_irq = qe_ic_get_high_irq(qe_ic);
-+ if (cascade_irq == 0)
-+ cascade_irq = qe_ic_get_low_irq(qe_ic);
-+
-+ if (cascade_irq != 0)
-+ generic_handle_irq(cascade_irq);
-+
-+ chip->irq_eoi(&desc->irq_data);
-+}
-+
-+static int __init qe_ic_init(struct device_node *node, unsigned int flags)
-+{
-+ struct qe_ic *qe_ic;
-+ struct resource res;
-+ u32 temp = 0, high_active = 0;
-+ int ret = 0;
-+
-+ if (!node)
-+ return -ENODEV;
-+
-+ ret = of_address_to_resource(node, 0, &res);
-+ if (ret) {
-+ ret = -ENODEV;
-+ goto err_put_node;
-+ }
-+
-+ qe_ic = kzalloc(sizeof(*qe_ic), GFP_KERNEL);
-+ if (qe_ic == NULL) {
-+ ret = -ENOMEM;
-+ goto err_put_node;
-+ }
-+
-+ qe_ic->irqhost = irq_domain_add_linear(node, NR_QE_IC_INTS,
-+ &qe_ic_host_ops, qe_ic);
-+ if (qe_ic->irqhost == NULL) {
-+ ret = -ENOMEM;
-+ goto err_free_qe_ic;
-+ }
-+
-+ qe_ic->regs = ioremap(res.start, resource_size(&res));
-+
-+ qe_ic->hc_irq = qe_ic_irq_chip;
-+
-+ qe_ic->virq_high = irq_of_parse_and_map(node, 0);
-+ qe_ic->virq_low = irq_of_parse_and_map(node, 1);
-+
-+ if (qe_ic->virq_low == 0) {
-+ pr_err("Failed to map QE_IC low IRQ\n");
-+ ret = -ENOMEM;
-+ goto err_domain_remove;
-+ }
-+
-+ /* default priority scheme is grouped. If spread mode is */
-+ /* required, configure cicr accordingly. */
-+ if (flags & QE_IC_SPREADMODE_GRP_W)
-+ temp |= CICR_GWCC;
-+ if (flags & QE_IC_SPREADMODE_GRP_X)
-+ temp |= CICR_GXCC;
-+ if (flags & QE_IC_SPREADMODE_GRP_Y)
-+ temp |= CICR_GYCC;
-+ if (flags & QE_IC_SPREADMODE_GRP_Z)
-+ temp |= CICR_GZCC;
-+ if (flags & QE_IC_SPREADMODE_GRP_RISCA)
-+ temp |= CICR_GRTA;
-+ if (flags & QE_IC_SPREADMODE_GRP_RISCB)
-+ temp |= CICR_GRTB;
-+
-+ /* choose destination signal for highest priority interrupt */
-+ if (flags & QE_IC_HIGH_SIGNAL) {
-+ temp |= (SIGNAL_HIGH << CICR_HPIT_SHIFT);
-+ high_active = 1;
-+ }
-+
-+ qe_ic_write(qe_ic->regs, QEIC_CICR, temp);
-+
-+ irq_set_handler_data(qe_ic->virq_low, qe_ic);
-+ irq_set_chained_handler(qe_ic->virq_low, qe_ic_cascade_low_mpic);
-+
-+ if (qe_ic->virq_high != 0 &&
-+ qe_ic->virq_high != qe_ic->virq_low) {
-+ irq_set_handler_data(qe_ic->virq_high, qe_ic);
-+ irq_set_chained_handler(qe_ic->virq_high,
-+ qe_ic_cascade_high_mpic);
-+ }
-+ of_node_put(node);
-+ return 0;
-+
-+err_domain_remove:
-+ irq_domain_remove(qe_ic->irqhost);
-+err_free_qe_ic:
-+ kfree(qe_ic);
-+err_put_node:
-+ of_node_put(node);
-+ return ret;
-+}
-+
-+static int __init init_qe_ic(struct device_node *node,
-+ struct device_node *parent)
-+{
-+ int ret;
-+
-+ ret = qe_ic_init(node, 0);
-+ if (ret)
-+ return ret;
-+
-+ return 0;
-+}
-+
-+IRQCHIP_DECLARE(qeic, "fsl,qe-ic", init_qe_ic);
---- a/drivers/net/wan/fsl_ucc_hdlc.c
-+++ b/drivers/net/wan/fsl_ucc_hdlc.c
-@@ -381,8 +381,8 @@ static netdev_tx_t ucc_hdlc_tx(struct sk
- /* set bd status and length */
- bd_status = (bd_status & T_W_S) | T_R_S | T_I_S | T_L_S | T_TC_S;
-
-- iowrite16be(bd_status, &bd->status);
- iowrite16be(skb->len, &bd->length);
-+ iowrite16be(bd_status, &bd->status);
-
- /* Move to next BD in the ring */
- if (!(bd_status & T_W_S))
-@@ -457,7 +457,7 @@ static int hdlc_rx_done(struct ucc_hdlc_
- struct sk_buff *skb = NULL;
- hdlc_device *hdlc = dev_to_hdlc(dev);
- struct qe_bd *bd;
-- u32 bd_status;
-+ u16 bd_status;
- u16 length, howmany = 0;
- u8 *bdbuffer;
- int i;
---- a/drivers/soc/fsl/qe/Kconfig
-+++ b/drivers/soc/fsl/qe/Kconfig
-@@ -4,7 +4,7 @@
-
- config QUICC_ENGINE
- bool "Freescale QUICC Engine (QE) Support"
-- depends on FSL_SOC && PPC32
-+ depends on OF && HAS_IOMEM
- select GENERIC_ALLOCATOR
- select CRC32
- help
---- a/drivers/soc/fsl/qe/Makefile
-+++ b/drivers/soc/fsl/qe/Makefile
-@@ -1,7 +1,7 @@
- #
- # Makefile for the linux ppc-specific parts of QE
- #
--obj-$(CONFIG_QUICC_ENGINE)+= qe.o qe_common.o qe_ic.o qe_io.o
-+obj-$(CONFIG_QUICC_ENGINE)+= qe.o qe_common.o qe_io.o
- obj-$(CONFIG_CPM) += qe_common.o
- obj-$(CONFIG_UCC) += ucc.o
- obj-$(CONFIG_UCC_SLOW) += ucc_slow.o
---- a/drivers/soc/fsl/qe/qe.c
-+++ b/drivers/soc/fsl/qe/qe.c
-@@ -33,8 +33,6 @@
- #include <asm/pgtable.h>
- #include <soc/fsl/qe/immap_qe.h>
- #include <soc/fsl/qe/qe.h>
--#include <asm/prom.h>
--#include <asm/rheap.h>
-
- static void qe_snums_init(void);
- static int qe_sdma_init(void);
-@@ -109,15 +107,27 @@ void qe_reset(void)
- panic("sdma init failed!");
- }
-
-+/* issue commands to QE, return 0 on success while -EIO on error
-+ *
-+ * @cmd: the command code, should be QE_INIT_TX_RX, QE_STOP_TX and so on
-+ * @device: which sub-block will run the command, QE_CR_SUBBLOCK_UCCFAST1 - 8
-+ * , QE_CR_SUBBLOCK_UCCSLOW1 - 8, QE_CR_SUBBLOCK_MCC1 - 3,
-+ * QE_CR_SUBBLOCK_IDMA1 - 4 and such on.
-+ * @mcn_protocol: specifies mode for the command for non-MCC, should be
-+ * QE_CR_PROTOCOL_HDLC_TRANSPARENT, QE_CR_PROTOCOL_QMC, QE_CR_PROTOCOL_UART
-+ * and such on.
-+ * @cmd_input: command related data.
-+ */
- int qe_issue_cmd(u32 cmd, u32 device, u8 mcn_protocol, u32 cmd_input)
- {
- unsigned long flags;
- u8 mcn_shift = 0, dev_shift = 0;
-- u32 ret;
-+ int ret;
-+ int i;
-
- spin_lock_irqsave(&qe_lock, flags);
- if (cmd == QE_RESET) {
-- out_be32(&qe_immr->cp.cecr, (u32) (cmd | QE_CR_FLG));
-+ iowrite32be((cmd | QE_CR_FLG), &qe_immr->cp.cecr);
- } else {
- if (cmd == QE_ASSIGN_PAGE) {
- /* Here device is the SNUM, not sub-block */
-@@ -134,20 +144,26 @@ int qe_issue_cmd(u32 cmd, u32 device, u8
- mcn_shift = QE_CR_MCN_NORMAL_SHIFT;
- }
-
-- out_be32(&qe_immr->cp.cecdr, cmd_input);
-- out_be32(&qe_immr->cp.cecr,
-- (cmd | QE_CR_FLG | ((u32) device << dev_shift) | (u32)
-- mcn_protocol << mcn_shift));
-+ iowrite32be(cmd_input, &qe_immr->cp.cecdr);
-+ iowrite32be((cmd | QE_CR_FLG | ((u32)device << dev_shift) |
-+ (u32)mcn_protocol << mcn_shift), &qe_immr->cp.cecr);
- }
-
- /* wait for the QE_CR_FLG to clear */
-- ret = spin_event_timeout((in_be32(&qe_immr->cp.cecr) & QE_CR_FLG) == 0,
-- 100, 0);
-+ ret = -EIO;
-+ for (i = 0; i < 100; i++) {
-+ if ((ioread32be(&qe_immr->cp.cecr) & QE_CR_FLG) == 0) {
-+ ret = 0;
-+ break;
-+ }
-+ udelay(1);
-+ }
-+
- /* On timeout (e.g. failure), the expression will be false (ret == 0),
- otherwise it will be true (ret == 1). */
- spin_unlock_irqrestore(&qe_lock, flags);
-
-- return ret == 1;
-+ return ret;
- }
- EXPORT_SYMBOL(qe_issue_cmd);
-
-@@ -169,8 +185,8 @@ static unsigned int brg_clk = 0;
- unsigned int qe_get_brg_clk(void)
- {
- struct device_node *qe;
-- int size;
-- const u32 *prop;
-+ u32 val;
-+ int ret;
- unsigned int mod;
-
- if (brg_clk)
-@@ -183,9 +199,9 @@ unsigned int qe_get_brg_clk(void)
- return brg_clk;
- }
-
-- prop = of_get_property(qe, "brg-frequency", &size);
-- if (prop && size == sizeof(*prop))
-- brg_clk = *prop;
-+ ret = of_property_read_u32(qe, "brg-frequency", &val);
-+ if (!ret)
-+ brg_clk = val;
-
- of_node_put(qe);
-
-@@ -234,7 +250,7 @@ int qe_setbrg(enum qe_clock brg, unsigne
- tempval = ((divisor - 1) << QE_BRGC_DIVISOR_SHIFT) |
- QE_BRGC_ENABLE | div16;
-
-- out_be32(&qe_immr->brg.brgc[brg - QE_BRG1], tempval);
-+ iowrite32be(tempval, &qe_immr->brg.brgc[brg - QE_BRG1]);
-
- return 0;
- }
-@@ -368,9 +384,9 @@ static int qe_sdma_init(void)
- return -ENOMEM;
- }
-
-- out_be32(&sdma->sdebcr, (u32) sdma_buf_offset & QE_SDEBCR_BA_MASK);
-- out_be32(&sdma->sdmr, (QE_SDMR_GLB_1_MSK |
-- (0x1 << QE_SDMR_CEN_SHIFT)));
-+ iowrite32be((u32)sdma_buf_offset & QE_SDEBCR_BA_MASK, &sdma->sdebcr);
-+ iowrite32be((QE_SDMR_GLB_1_MSK | (0x1 << QE_SDMR_CEN_SHIFT)),
-+ &sdma->sdmr);
-
- return 0;
- }
-@@ -408,14 +424,14 @@ static void qe_upload_microcode(const vo
- "uploading microcode '%s'\n", ucode->id);
-
- /* Use auto-increment */
-- out_be32(&qe_immr->iram.iadd, be32_to_cpu(ucode->iram_offset) |
-- QE_IRAM_IADD_AIE | QE_IRAM_IADD_BADDR);
-+ iowrite32be(be32_to_cpu(ucode->iram_offset) | QE_IRAM_IADD_AIE |
-+ QE_IRAM_IADD_BADDR, &qe_immr->iram.iadd);
-
- for (i = 0; i < be32_to_cpu(ucode->count); i++)
-- out_be32(&qe_immr->iram.idata, be32_to_cpu(code[i]));
-+ iowrite32be(be32_to_cpu(code[i]), &qe_immr->iram.idata);
-
- /* Set I-RAM Ready Register */
-- out_be32(&qe_immr->iram.iready, be32_to_cpu(QE_IRAM_READY));
-+ iowrite32be(be32_to_cpu(QE_IRAM_READY), &qe_immr->iram.iready);
- }
-
- /*
-@@ -500,7 +516,7 @@ int qe_upload_firmware(const struct qe_f
- * If the microcode calls for it, split the I-RAM.
- */
- if (!firmware->split)
-- setbits16(&qe_immr->cp.cercr, QE_CP_CERCR_CIR);
-+ qe_setbits16(&qe_immr->cp.cercr, QE_CP_CERCR_CIR);
-
- if (firmware->soc.model)
- printk(KERN_INFO
-@@ -534,11 +550,11 @@ int qe_upload_firmware(const struct qe_f
- u32 trap = be32_to_cpu(ucode->traps[j]);
-
- if (trap)
-- out_be32(&qe_immr->rsp[i].tibcr[j], trap);
-+ iowrite32be(trap, &qe_immr->rsp[i].tibcr[j]);
- }
-
- /* Enable traps */
-- out_be32(&qe_immr->rsp[i].eccr, be32_to_cpu(ucode->eccr));
-+ iowrite32be(be32_to_cpu(ucode->eccr), &qe_immr->rsp[i].eccr);
- }
-
- qe_firmware_uploaded = 1;
-@@ -657,9 +673,9 @@ EXPORT_SYMBOL(qe_get_num_of_risc);
- unsigned int qe_get_num_of_snums(void)
- {
- struct device_node *qe;
-- int size;
- unsigned int num_of_snums;
-- const u32 *prop;
-+ u32 val;
-+ int ret;
-
- num_of_snums = 28; /* The default number of snum for threads is 28 */
- qe = of_find_compatible_node(NULL, NULL, "fsl,qe");
-@@ -673,9 +689,9 @@ unsigned int qe_get_num_of_snums(void)
- return num_of_snums;
- }
-
-- prop = of_get_property(qe, "fsl,qe-num-snums", &size);
-- if (prop && size == sizeof(*prop)) {
-- num_of_snums = *prop;
-+ ret = of_property_read_u32(qe, "fsl,qe-num-snums", &val);
-+ if (!ret) {
-+ num_of_snums = val;
- if ((num_of_snums < 28) || (num_of_snums > QE_NUM_OF_SNUM)) {
- /* No QE ever has fewer than 28 SNUMs */
- pr_err("QE: number of snum is invalid\n");
---- a/drivers/soc/fsl/qe/qe_ic.h
-+++ /dev/null
-@@ -1,103 +0,0 @@
--/*
-- * drivers/soc/fsl/qe/qe_ic.h
-- *
-- * QUICC ENGINE Interrupt Controller Header
-- *
-- * Copyright (C) 2006 Freescale Semiconductor, Inc. All rights reserved.
-- *
-- * Author: Li Yang <leoli@freescale.com>
-- * Based on code from Shlomi Gridish <gridish@freescale.com>
-- *
-- * This program is free software; you can redistribute it and/or modify it
-- * under the terms of the GNU General Public License as published by the
-- * Free Software Foundation; either version 2 of the License, or (at your
-- * option) any later version.
-- */
--#ifndef _POWERPC_SYSDEV_QE_IC_H
--#define _POWERPC_SYSDEV_QE_IC_H
--
--#include <soc/fsl/qe/qe_ic.h>
--
--#define NR_QE_IC_INTS 64
--
--/* QE IC registers offset */
--#define QEIC_CICR 0x00
--#define QEIC_CIVEC 0x04
--#define QEIC_CRIPNR 0x08
--#define QEIC_CIPNR 0x0c
--#define QEIC_CIPXCC 0x10
--#define QEIC_CIPYCC 0x14
--#define QEIC_CIPWCC 0x18
--#define QEIC_CIPZCC 0x1c
--#define QEIC_CIMR 0x20
--#define QEIC_CRIMR 0x24
--#define QEIC_CICNR 0x28
--#define QEIC_CIPRTA 0x30
--#define QEIC_CIPRTB 0x34
--#define QEIC_CRICR 0x3c
--#define QEIC_CHIVEC 0x60
--
--/* Interrupt priority registers */
--#define CIPCC_SHIFT_PRI0 29
--#define CIPCC_SHIFT_PRI1 26
--#define CIPCC_SHIFT_PRI2 23
--#define CIPCC_SHIFT_PRI3 20
--#define CIPCC_SHIFT_PRI4 13
--#define CIPCC_SHIFT_PRI5 10
--#define CIPCC_SHIFT_PRI6 7
--#define CIPCC_SHIFT_PRI7 4
--
--/* CICR priority modes */
--#define CICR_GWCC 0x00040000
--#define CICR_GXCC 0x00020000
--#define CICR_GYCC 0x00010000
--#define CICR_GZCC 0x00080000
--#define CICR_GRTA 0x00200000
--#define CICR_GRTB 0x00400000
--#define CICR_HPIT_SHIFT 8
--#define CICR_HPIT_MASK 0x00000300
--#define CICR_HP_SHIFT 24
--#define CICR_HP_MASK 0x3f000000
--
--/* CICNR */
--#define CICNR_WCC1T_SHIFT 20
--#define CICNR_ZCC1T_SHIFT 28
--#define CICNR_YCC1T_SHIFT 12
--#define CICNR_XCC1T_SHIFT 4
--
--/* CRICR */
--#define CRICR_RTA1T_SHIFT 20
--#define CRICR_RTB1T_SHIFT 28
--
--/* Signal indicator */
--#define SIGNAL_MASK 3
--#define SIGNAL_HIGH 2
--#define SIGNAL_LOW 0
--
--struct qe_ic {
-- /* Control registers offset */
-- volatile u32 __iomem *regs;
--
-- /* The remapper for this QEIC */
-- struct irq_domain *irqhost;
--
-- /* The "linux" controller struct */
-- struct irq_chip hc_irq;
--
-- /* VIRQ numbers of QE high/low irqs */
-- unsigned int virq_high;
-- unsigned int virq_low;
--};
--
--/*
-- * QE interrupt controller internal structure
-- */
--struct qe_ic_info {
-- u32 mask; /* location of this source at the QIMR register. */
-- u32 mask_reg; /* Mask register offset */
-- u8 pri_code; /* for grouped interrupts sources - the interrupt
-- code as appears at the group priority register */
-- u32 pri_reg; /* Group priority register offset */
--};
--
--#endif /* _POWERPC_SYSDEV_QE_IC_H */
---- a/drivers/soc/fsl/qe/qe_io.c
-+++ b/drivers/soc/fsl/qe/qe_io.c
-@@ -22,8 +22,6 @@
-
- #include <asm/io.h>
- #include <soc/fsl/qe/qe.h>
--#include <asm/prom.h>
--#include <sysdev/fsl_soc.h>
-
- #undef DEBUG
-
-@@ -61,16 +59,16 @@ void __par_io_config_pin(struct qe_pio_r
- pin_mask1bit = (u32) (1 << (QE_PIO_PINS - (pin + 1)));
-
- /* Set open drain, if required */
-- tmp_val = in_be32(&par_io->cpodr);
-+ tmp_val = ioread32be(&par_io->cpodr);
- if (open_drain)
-- out_be32(&par_io->cpodr, pin_mask1bit | tmp_val);
-+ iowrite32be(pin_mask1bit | tmp_val, &par_io->cpodr);
- else
-- out_be32(&par_io->cpodr, ~pin_mask1bit & tmp_val);
-+ iowrite32be(~pin_mask1bit & tmp_val, &par_io->cpodr);
-
- /* define direction */
- tmp_val = (pin > (QE_PIO_PINS / 2) - 1) ?
-- in_be32(&par_io->cpdir2) :
-- in_be32(&par_io->cpdir1);
-+ ioread32be(&par_io->cpdir2) :
-+ ioread32be(&par_io->cpdir1);
-
- /* get all bits mask for 2 bit per port */
- pin_mask2bits = (u32) (0x3 << (QE_PIO_PINS -
-@@ -82,34 +80,30 @@ void __par_io_config_pin(struct qe_pio_r
-
- /* clear and set 2 bits mask */
- if (pin > (QE_PIO_PINS / 2) - 1) {
-- out_be32(&par_io->cpdir2,
-- ~pin_mask2bits & tmp_val);
-+ iowrite32be(~pin_mask2bits & tmp_val, &par_io->cpdir2);
- tmp_val &= ~pin_mask2bits;
-- out_be32(&par_io->cpdir2, new_mask2bits | tmp_val);
-+ iowrite32be(new_mask2bits | tmp_val, &par_io->cpdir2);
- } else {
-- out_be32(&par_io->cpdir1,
-- ~pin_mask2bits & tmp_val);
-+ iowrite32be(~pin_mask2bits & tmp_val, &par_io->cpdir1);
- tmp_val &= ~pin_mask2bits;
-- out_be32(&par_io->cpdir1, new_mask2bits | tmp_val);
-+ iowrite32be(new_mask2bits | tmp_val, &par_io->cpdir1);
- }
- /* define pin assignment */
- tmp_val = (pin > (QE_PIO_PINS / 2) - 1) ?
-- in_be32(&par_io->cppar2) :
-- in_be32(&par_io->cppar1);
-+ ioread32be(&par_io->cppar2) :
-+ ioread32be(&par_io->cppar1);
-
- new_mask2bits = (u32) (assignment << (QE_PIO_PINS -
- (pin % (QE_PIO_PINS / 2) + 1) * 2));
- /* clear and set 2 bits mask */
- if (pin > (QE_PIO_PINS / 2) - 1) {
-- out_be32(&par_io->cppar2,
-- ~pin_mask2bits & tmp_val);
-+ iowrite32be(~pin_mask2bits & tmp_val, &par_io->cppar2);
- tmp_val &= ~pin_mask2bits;
-- out_be32(&par_io->cppar2, new_mask2bits | tmp_val);
-+ iowrite32be(new_mask2bits | tmp_val, &par_io->cppar2);
- } else {
-- out_be32(&par_io->cppar1,
-- ~pin_mask2bits & tmp_val);
-+ iowrite32be(~pin_mask2bits & tmp_val, &par_io->cppar1);
- tmp_val &= ~pin_mask2bits;
-- out_be32(&par_io->cppar1, new_mask2bits | tmp_val);
-+ iowrite32be(new_mask2bits | tmp_val, &par_io->cppar1);
- }
- }
- EXPORT_SYMBOL(__par_io_config_pin);
-@@ -137,12 +131,12 @@ int par_io_data_set(u8 port, u8 pin, u8
- /* calculate pin location */
- pin_mask = (u32) (1 << (QE_PIO_PINS - 1 - pin));
-
-- tmp_val = in_be32(&par_io[port].cpdata);
-+ tmp_val = ioread32be(&par_io[port].cpdata);
-
- if (val == 0) /* clear */
-- out_be32(&par_io[port].cpdata, ~pin_mask & tmp_val);
-+ iowrite32be(~pin_mask & tmp_val, &par_io[port].cpdata);
- else /* set */
-- out_be32(&par_io[port].cpdata, pin_mask | tmp_val);
-+ iowrite32be(pin_mask | tmp_val, &par_io[port].cpdata);
-
- return 0;
- }
---- a/drivers/soc/fsl/qe/qe_tdm.c
-+++ b/drivers/soc/fsl/qe/qe_tdm.c
-@@ -227,10 +227,10 @@ void ucc_tdm_init(struct ucc_tdm *utdm,
- &siram[siram_entry_id * 32 + 0x200 + i]);
- }
-
-- setbits16(&siram[(siram_entry_id * 32) + (utdm->num_of_ts - 1)],
-- SIR_LAST);
-- setbits16(&siram[(siram_entry_id * 32) + 0x200 + (utdm->num_of_ts - 1)],
-- SIR_LAST);
-+ qe_setbits16(&siram[(siram_entry_id * 32) + (utdm->num_of_ts - 1)],
-+ SIR_LAST);
-+ qe_setbits16(&siram[(siram_entry_id * 32) + 0x200 +
-+ (utdm->num_of_ts - 1)], SIR_LAST);
-
- /* Set SIxMR register */
- sixmr = SIMR_SAD(siram_entry_id);
---- a/drivers/soc/fsl/qe/ucc.c
-+++ b/drivers/soc/fsl/qe/ucc.c
-@@ -39,7 +39,7 @@ int ucc_set_qe_mux_mii_mng(unsigned int
- return -EINVAL;
-
- spin_lock_irqsave(&cmxgcr_lock, flags);
-- clrsetbits_be32(&qe_immr->qmx.cmxgcr, QE_CMXGCR_MII_ENET_MNG,
-+ qe_clrsetbits32(&qe_immr->qmx.cmxgcr, QE_CMXGCR_MII_ENET_MNG,
- ucc_num << QE_CMXGCR_MII_ENET_MNG_SHIFT);
- spin_unlock_irqrestore(&cmxgcr_lock, flags);
-
-@@ -84,7 +84,7 @@ int ucc_set_type(unsigned int ucc_num, e
- return -EINVAL;
- }
-
-- clrsetbits_8(guemr, UCC_GUEMR_MODE_MASK,
-+ qe_clrsetbits8(guemr, UCC_GUEMR_MODE_MASK,
- UCC_GUEMR_SET_RESERVED3 | speed);
-
- return 0;
-@@ -113,9 +113,9 @@ int ucc_mux_set_grant_tsa_bkpt(unsigned
- get_cmxucr_reg(ucc_num, &cmxucr, &reg_num, &shift);
-
- if (set)
-- setbits32(cmxucr, mask << shift);
-+ qe_setbits32(cmxucr, mask << shift);
- else
-- clrbits32(cmxucr, mask << shift);
-+ qe_clrbits32(cmxucr, mask << shift);
-
- return 0;
- }
-@@ -211,7 +211,7 @@ int ucc_set_qe_mux_rxtx(unsigned int ucc
- if (mode == COMM_DIR_RX)
- shift += 4;
-
-- clrsetbits_be32(cmxucr, QE_CMXUCR_TX_CLK_SRC_MASK << shift,
-+ qe_clrsetbits32(cmxucr, QE_CMXUCR_TX_CLK_SRC_MASK << shift,
- clock_bits << shift);
-
- return 0;
---- a/drivers/soc/fsl/qe/ucc_fast.c
-+++ b/drivers/soc/fsl/qe/ucc_fast.c
-@@ -33,41 +33,41 @@ void ucc_fast_dump_regs(struct ucc_fast_
- printk(KERN_INFO "Base address: 0x%p\n", uccf->uf_regs);
-
- printk(KERN_INFO "gumr : addr=0x%p, val=0x%08x\n",
-- &uccf->uf_regs->gumr, in_be32(&uccf->uf_regs->gumr));
-+ &uccf->uf_regs->gumr, ioread32be(&uccf->uf_regs->gumr));
- printk(KERN_INFO "upsmr : addr=0x%p, val=0x%08x\n",
-- &uccf->uf_regs->upsmr, in_be32(&uccf->uf_regs->upsmr));
-+ &uccf->uf_regs->upsmr, ioread32be(&uccf->uf_regs->upsmr));
- printk(KERN_INFO "utodr : addr=0x%p, val=0x%04x\n",
-- &uccf->uf_regs->utodr, in_be16(&uccf->uf_regs->utodr));
-+ &uccf->uf_regs->utodr, ioread16be(&uccf->uf_regs->utodr));
- printk(KERN_INFO "udsr : addr=0x%p, val=0x%04x\n",
-- &uccf->uf_regs->udsr, in_be16(&uccf->uf_regs->udsr));
-+ &uccf->uf_regs->udsr, ioread16be(&uccf->uf_regs->udsr));
- printk(KERN_INFO "ucce : addr=0x%p, val=0x%08x\n",
-- &uccf->uf_regs->ucce, in_be32(&uccf->uf_regs->ucce));
-+ &uccf->uf_regs->ucce, ioread32be(&uccf->uf_regs->ucce));
- printk(KERN_INFO "uccm : addr=0x%p, val=0x%08x\n",
-- &uccf->uf_regs->uccm, in_be32(&uccf->uf_regs->uccm));
-+ &uccf->uf_regs->uccm, ioread32be(&uccf->uf_regs->uccm));
- printk(KERN_INFO "uccs : addr=0x%p, val=0x%02x\n",
-- &uccf->uf_regs->uccs, in_8(&uccf->uf_regs->uccs));
-+ &uccf->uf_regs->uccs, ioread8(&uccf->uf_regs->uccs));
- printk(KERN_INFO "urfb : addr=0x%p, val=0x%08x\n",
-- &uccf->uf_regs->urfb, in_be32(&uccf->uf_regs->urfb));
-+ &uccf->uf_regs->urfb, ioread32be(&uccf->uf_regs->urfb));
- printk(KERN_INFO "urfs : addr=0x%p, val=0x%04x\n",
-- &uccf->uf_regs->urfs, in_be16(&uccf->uf_regs->urfs));
-+ &uccf->uf_regs->urfs, ioread16be(&uccf->uf_regs->urfs));
- printk(KERN_INFO "urfet : addr=0x%p, val=0x%04x\n",
-- &uccf->uf_regs->urfet, in_be16(&uccf->uf_regs->urfet));
-+ &uccf->uf_regs->urfet, ioread16be(&uccf->uf_regs->urfet));
- printk(KERN_INFO "urfset: addr=0x%p, val=0x%04x\n",
-- &uccf->uf_regs->urfset, in_be16(&uccf->uf_regs->urfset));
-+ &uccf->uf_regs->urfset, ioread16be(&uccf->uf_regs->urfset));
- printk(KERN_INFO "utfb : addr=0x%p, val=0x%08x\n",
-- &uccf->uf_regs->utfb, in_be32(&uccf->uf_regs->utfb));
-+ &uccf->uf_regs->utfb, ioread32be(&uccf->uf_regs->utfb));
- printk(KERN_INFO "utfs : addr=0x%p, val=0x%04x\n",
-- &uccf->uf_regs->utfs, in_be16(&uccf->uf_regs->utfs));
-+ &uccf->uf_regs->utfs, ioread16be(&uccf->uf_regs->utfs));
- printk(KERN_INFO "utfet : addr=0x%p, val=0x%04x\n",
-- &uccf->uf_regs->utfet, in_be16(&uccf->uf_regs->utfet));
-+ &uccf->uf_regs->utfet, ioread16be(&uccf->uf_regs->utfet));
- printk(KERN_INFO "utftt : addr=0x%p, val=0x%04x\n",
-- &uccf->uf_regs->utftt, in_be16(&uccf->uf_regs->utftt));
-+ &uccf->uf_regs->utftt, ioread16be(&uccf->uf_regs->utftt));
- printk(KERN_INFO "utpt : addr=0x%p, val=0x%04x\n",
-- &uccf->uf_regs->utpt, in_be16(&uccf->uf_regs->utpt));
-+ &uccf->uf_regs->utpt, ioread16be(&uccf->uf_regs->utpt));
- printk(KERN_INFO "urtry : addr=0x%p, val=0x%08x\n",
-- &uccf->uf_regs->urtry, in_be32(&uccf->uf_regs->urtry));
-+ &uccf->uf_regs->urtry, ioread32be(&uccf->uf_regs->urtry));
- printk(KERN_INFO "guemr : addr=0x%p, val=0x%02x\n",
-- &uccf->uf_regs->guemr, in_8(&uccf->uf_regs->guemr));
-+ &uccf->uf_regs->guemr, ioread8(&uccf->uf_regs->guemr));
- }
- EXPORT_SYMBOL(ucc_fast_dump_regs);
-
-@@ -89,7 +89,7 @@ EXPORT_SYMBOL(ucc_fast_get_qe_cr_subbloc
-
- void ucc_fast_transmit_on_demand(struct ucc_fast_private * uccf)
- {
-- out_be16(&uccf->uf_regs->utodr, UCC_FAST_TOD);
-+ iowrite16be(UCC_FAST_TOD, &uccf->uf_regs->utodr);
- }
- EXPORT_SYMBOL(ucc_fast_transmit_on_demand);
-
-@@ -101,7 +101,7 @@ void ucc_fast_enable(struct ucc_fast_pri
- uf_regs = uccf->uf_regs;
-
- /* Enable reception and/or transmission on this UCC. */
-- gumr = in_be32(&uf_regs->gumr);
-+ gumr = ioread32be(&uf_regs->gumr);
- if (mode & COMM_DIR_TX) {
- gumr |= UCC_FAST_GUMR_ENT;
- uccf->enabled_tx = 1;
-@@ -110,7 +110,7 @@ void ucc_fast_enable(struct ucc_fast_pri
- gumr |= UCC_FAST_GUMR_ENR;
- uccf->enabled_rx = 1;
- }
-- out_be32(&uf_regs->gumr, gumr);
-+ iowrite32be(gumr, &uf_regs->gumr);
- }
- EXPORT_SYMBOL(ucc_fast_enable);
-
-@@ -122,7 +122,7 @@ void ucc_fast_disable(struct ucc_fast_pr
- uf_regs = uccf->uf_regs;
-
- /* Disable reception and/or transmission on this UCC. */
-- gumr = in_be32(&uf_regs->gumr);
-+ gumr = ioread32be(&uf_regs->gumr);
- if (mode & COMM_DIR_TX) {
- gumr &= ~UCC_FAST_GUMR_ENT;
- uccf->enabled_tx = 0;
-@@ -131,7 +131,7 @@ void ucc_fast_disable(struct ucc_fast_pr
- gumr &= ~UCC_FAST_GUMR_ENR;
- uccf->enabled_rx = 0;
- }
-- out_be32(&uf_regs->gumr, gumr);
-+ iowrite32be(gumr, &uf_regs->gumr);
- }
- EXPORT_SYMBOL(ucc_fast_disable);
-
-@@ -263,12 +263,13 @@ int ucc_fast_init(struct ucc_fast_info *
- gumr |= uf_info->tenc;
- gumr |= uf_info->tcrc;
- gumr |= uf_info->mode;
-- out_be32(&uf_regs->gumr, gumr);
-+ iowrite32be(gumr, &uf_regs->gumr);
-
- /* Allocate memory for Tx Virtual Fifo */
- uccf->ucc_fast_tx_virtual_fifo_base_offset =
- qe_muram_alloc(uf_info->utfs, UCC_FAST_VIRT_FIFO_REGS_ALIGNMENT);
-- if (IS_ERR_VALUE(uccf->ucc_fast_tx_virtual_fifo_base_offset)) {
-+ if (IS_ERR_VALUE((unsigned long)uccf->
-+ ucc_fast_tx_virtual_fifo_base_offset)) {
- printk(KERN_ERR "%s: cannot allocate MURAM for TX FIFO\n",
- __func__);
- uccf->ucc_fast_tx_virtual_fifo_base_offset = 0;
-@@ -281,7 +282,8 @@ int ucc_fast_init(struct ucc_fast_info *
- qe_muram_alloc(uf_info->urfs +
- UCC_FAST_RECEIVE_VIRTUAL_FIFO_SIZE_FUDGE_FACTOR,
- UCC_FAST_VIRT_FIFO_REGS_ALIGNMENT);
-- if (IS_ERR_VALUE(uccf->ucc_fast_rx_virtual_fifo_base_offset)) {
-+ if (IS_ERR_VALUE((unsigned long)uccf->
-+ ucc_fast_rx_virtual_fifo_base_offset)) {
- printk(KERN_ERR "%s: cannot allocate MURAM for RX FIFO\n",
- __func__);
- uccf->ucc_fast_rx_virtual_fifo_base_offset = 0;
-@@ -290,15 +292,15 @@ int ucc_fast_init(struct ucc_fast_info *
- }
-
- /* Set Virtual Fifo registers */
-- out_be16(&uf_regs->urfs, uf_info->urfs);
-- out_be16(&uf_regs->urfet, uf_info->urfet);
-- out_be16(&uf_regs->urfset, uf_info->urfset);
-- out_be16(&uf_regs->utfs, uf_info->utfs);
-- out_be16(&uf_regs->utfet, uf_info->utfet);
-- out_be16(&uf_regs->utftt, uf_info->utftt);
-+ iowrite16be(uf_info->urfs, &uf_regs->urfs);
-+ iowrite16be(uf_info->urfet, &uf_regs->urfet);
-+ iowrite16be(uf_info->urfset, &uf_regs->urfset);
-+ iowrite16be(uf_info->utfs, &uf_regs->utfs);
-+ iowrite16be(uf_info->utfet, &uf_regs->utfet);
-+ iowrite16be(uf_info->utftt, &uf_regs->utftt);
- /* utfb, urfb are offsets from MURAM base */
-- out_be32(&uf_regs->utfb, uccf->ucc_fast_tx_virtual_fifo_base_offset);
-- out_be32(&uf_regs->urfb, uccf->ucc_fast_rx_virtual_fifo_base_offset);
-+ iowrite32be(uccf->ucc_fast_tx_virtual_fifo_base_offset, &uf_regs->utfb);
-+ iowrite32be(uccf->ucc_fast_rx_virtual_fifo_base_offset, &uf_regs->urfb);
-
- /* Mux clocking */
- /* Grant Support */
-@@ -366,14 +368,14 @@ int ucc_fast_init(struct ucc_fast_info *
- }
-
- /* Set interrupt mask register at UCC level. */
-- out_be32(&uf_regs->uccm, uf_info->uccm_mask);
-+ iowrite32be(uf_info->uccm_mask, &uf_regs->uccm);
-
- /* First, clear anything pending at UCC level,
- * otherwise, old garbage may come through
- * as soon as the dam is opened. */
-
- /* Writing '1' clears */
-- out_be32(&uf_regs->ucce, 0xffffffff);
-+ iowrite32be(0xffffffff, &uf_regs->ucce);
-
- *uccf_ret = uccf;
- return 0;
---- a/drivers/tty/serial/ucc_uart.c
-+++ b/drivers/tty/serial/ucc_uart.c
-@@ -34,6 +34,7 @@
- #include <soc/fsl/qe/ucc_slow.h>
-
- #include <linux/firmware.h>
-+#include <asm/cpm.h>
- #include <asm/reg.h>
-
- /*
---- a/include/soc/fsl/qe/qe.h
-+++ b/include/soc/fsl/qe/qe.h
-@@ -21,7 +21,6 @@
- #include <linux/spinlock.h>
- #include <linux/errno.h>
- #include <linux/err.h>
--#include <asm/cpm.h>
- #include <soc/fsl/qe/immap_qe.h>
- #include <linux/of.h>
- #include <linux/of_address.h>
---- a/include/soc/fsl/qe/qe_ic.h
-+++ /dev/null
-@@ -1,139 +0,0 @@
--/*
-- * Copyright (C) 2006 Freescale Semiconductor, Inc. All rights reserved.
-- *
-- * Authors: Shlomi Gridish <gridish@freescale.com>
-- * Li Yang <leoli@freescale.com>
-- *
-- * Description:
-- * QE IC external definitions and structure.
-- *
-- * This program is free software; you can redistribute it and/or modify it
-- * under the terms of the GNU General Public License as published by the
-- * Free Software Foundation; either version 2 of the License, or (at your
-- * option) any later version.
-- */
--#ifndef _ASM_POWERPC_QE_IC_H
--#define _ASM_POWERPC_QE_IC_H
--
--#include <linux/irq.h>
--
--struct device_node;
--struct qe_ic;
--
--#define NUM_OF_QE_IC_GROUPS 6
--
--/* Flags when we init the QE IC */
--#define QE_IC_SPREADMODE_GRP_W 0x00000001
--#define QE_IC_SPREADMODE_GRP_X 0x00000002
--#define QE_IC_SPREADMODE_GRP_Y 0x00000004
--#define QE_IC_SPREADMODE_GRP_Z 0x00000008
--#define QE_IC_SPREADMODE_GRP_RISCA 0x00000010
--#define QE_IC_SPREADMODE_GRP_RISCB 0x00000020
--
--#define QE_IC_LOW_SIGNAL 0x00000100
--#define QE_IC_HIGH_SIGNAL 0x00000200
--
--#define QE_IC_GRP_W_PRI0_DEST_SIGNAL_HIGH 0x00001000
--#define QE_IC_GRP_W_PRI1_DEST_SIGNAL_HIGH 0x00002000
--#define QE_IC_GRP_X_PRI0_DEST_SIGNAL_HIGH 0x00004000
--#define QE_IC_GRP_X_PRI1_DEST_SIGNAL_HIGH 0x00008000
--#define QE_IC_GRP_Y_PRI0_DEST_SIGNAL_HIGH 0x00010000
--#define QE_IC_GRP_Y_PRI1_DEST_SIGNAL_HIGH 0x00020000
--#define QE_IC_GRP_Z_PRI0_DEST_SIGNAL_HIGH 0x00040000
--#define QE_IC_GRP_Z_PRI1_DEST_SIGNAL_HIGH 0x00080000
--#define QE_IC_GRP_RISCA_PRI0_DEST_SIGNAL_HIGH 0x00100000
--#define QE_IC_GRP_RISCA_PRI1_DEST_SIGNAL_HIGH 0x00200000
--#define QE_IC_GRP_RISCB_PRI0_DEST_SIGNAL_HIGH 0x00400000
--#define QE_IC_GRP_RISCB_PRI1_DEST_SIGNAL_HIGH 0x00800000
--#define QE_IC_GRP_W_DEST_SIGNAL_SHIFT (12)
--
--/* QE interrupt sources groups */
--enum qe_ic_grp_id {
-- QE_IC_GRP_W = 0, /* QE interrupt controller group W */
-- QE_IC_GRP_X, /* QE interrupt controller group X */
-- QE_IC_GRP_Y, /* QE interrupt controller group Y */
-- QE_IC_GRP_Z, /* QE interrupt controller group Z */
-- QE_IC_GRP_RISCA, /* QE interrupt controller RISC group A */
-- QE_IC_GRP_RISCB /* QE interrupt controller RISC group B */
--};
--
--#ifdef CONFIG_QUICC_ENGINE
--void qe_ic_init(struct device_node *node, unsigned int flags,
-- void (*low_handler)(struct irq_desc *desc),
-- void (*high_handler)(struct irq_desc *desc));
--unsigned int qe_ic_get_low_irq(struct qe_ic *qe_ic);
--unsigned int qe_ic_get_high_irq(struct qe_ic *qe_ic);
--#else
--static inline void qe_ic_init(struct device_node *node, unsigned int flags,
-- void (*low_handler)(struct irq_desc *desc),
-- void (*high_handler)(struct irq_desc *desc))
--{}
--static inline unsigned int qe_ic_get_low_irq(struct qe_ic *qe_ic)
--{ return 0; }
--static inline unsigned int qe_ic_get_high_irq(struct qe_ic *qe_ic)
--{ return 0; }
--#endif /* CONFIG_QUICC_ENGINE */
--
--void qe_ic_set_highest_priority(unsigned int virq, int high);
--int qe_ic_set_priority(unsigned int virq, unsigned int priority);
--int qe_ic_set_high_priority(unsigned int virq, unsigned int priority, int high);
--
--static inline void qe_ic_cascade_low_ipic(struct irq_desc *desc)
--{
-- struct qe_ic *qe_ic = irq_desc_get_handler_data(desc);
-- unsigned int cascade_irq = qe_ic_get_low_irq(qe_ic);
--
-- if (cascade_irq != NO_IRQ)
-- generic_handle_irq(cascade_irq);
--}
--
--static inline void qe_ic_cascade_high_ipic(struct irq_desc *desc)
--{
-- struct qe_ic *qe_ic = irq_desc_get_handler_data(desc);
-- unsigned int cascade_irq = qe_ic_get_high_irq(qe_ic);
--
-- if (cascade_irq != NO_IRQ)
-- generic_handle_irq(cascade_irq);
--}
--
--static inline void qe_ic_cascade_low_mpic(struct irq_desc *desc)
--{
-- struct qe_ic *qe_ic = irq_desc_get_handler_data(desc);
-- unsigned int cascade_irq = qe_ic_get_low_irq(qe_ic);
-- struct irq_chip *chip = irq_desc_get_chip(desc);
--
-- if (cascade_irq != NO_IRQ)
-- generic_handle_irq(cascade_irq);
--
-- chip->irq_eoi(&desc->irq_data);
--}
--
--static inline void qe_ic_cascade_high_mpic(struct irq_desc *desc)
--{
-- struct qe_ic *qe_ic = irq_desc_get_handler_data(desc);
-- unsigned int cascade_irq = qe_ic_get_high_irq(qe_ic);
-- struct irq_chip *chip = irq_desc_get_chip(desc);
--
-- if (cascade_irq != NO_IRQ)
-- generic_handle_irq(cascade_irq);
--
-- chip->irq_eoi(&desc->irq_data);
--}
--
--static inline void qe_ic_cascade_muxed_mpic(struct irq_desc *desc)
--{
-- struct qe_ic *qe_ic = irq_desc_get_handler_data(desc);
-- unsigned int cascade_irq;
-- struct irq_chip *chip = irq_desc_get_chip(desc);
--
-- cascade_irq = qe_ic_get_high_irq(qe_ic);
-- if (cascade_irq == NO_IRQ)
-- cascade_irq = qe_ic_get_low_irq(qe_ic);
--
-- if (cascade_irq != NO_IRQ)
-- generic_handle_irq(cascade_irq);
--
-- chip->irq_eoi(&desc->irq_data);
--}
--
--#endif /* _ASM_POWERPC_QE_IC_H */
diff --git a/target/linux/layerscape/patches-4.9/814-rtc-support-layerscape.patch b/target/linux/layerscape/patches-4.9/814-rtc-support-layerscape.patch
deleted file mode 100644
index 9510a524d8..0000000000
--- a/target/linux/layerscape/patches-4.9/814-rtc-support-layerscape.patch
+++ /dev/null
@@ -1,682 +0,0 @@
-From bda12381598c3df43f4e60362a8cd4af58b7f5b0 Mon Sep 17 00:00:00 2001
-From: Yangbo Lu <yangbo.lu@nxp.com>
-Date: Wed, 17 Jan 2018 15:38:54 +0800
-Subject: [PATCH 26/30] rtc: support layerscape
-
-This is an integrated patch for layerscape rtc support.
-
-Signed-off-by: Zhang Ying-22455 <ying.zhang22455@nxp.com>
-Signed-off-by: Yangbo Lu <yangbo.lu@nxp.com>
----
- drivers/rtc/rtc-pcf85263.c | 665 +++++++++++++++++++++++++++++++++++++++++++++
- 1 file changed, 665 insertions(+)
- create mode 100644 drivers/rtc/rtc-pcf85263.c
-
---- /dev/null
-+++ b/drivers/rtc/rtc-pcf85263.c
-@@ -0,0 +1,665 @@
-+/*
-+ * rtc-pcf85263 Driver for the NXP PCF85263 RTC
-+ * Copyright 2016 Parkeon
-+ *
-+ * This program is free software; you can redistribute it and/or modify
-+ * it under the terms of the GNU General Public License version 2 as
-+ * published by the Free Software Foundation.
-+ */
-+
-+#include <linux/module.h>
-+#include <linux/mutex.h>
-+#include <linux/rtc.h>
-+#include <linux/i2c.h>
-+#include <linux/bcd.h>
-+#include <linux/of.h>
-+#include <linux/of_device.h>
-+#include <linux/regmap.h>
-+
-+
-+#define DRV_NAME "rtc-pcf85263"
-+
-+/* Quartz capacitance */
-+#define PCF85263_QUARTZCAP_7pF 0
-+#define PCF85263_QUARTZCAP_6pF 1
-+#define PCF85263_QUARTZCAP_12p5pF 2
-+
-+/* Quartz drive strength */
-+#define PCF85263_QUARTZDRIVE_NORMAL 0
-+#define PCF85263_QUARTZDRIVE_LOW 1
-+#define PCF85263_QUARTZDRIVE_HIGH 2
-+
-+
-+#define PCF85263_REG_RTC_SC 0x01 /* Seconds */
-+#define PCF85263_REG_RTC_SC_OS BIT(7) /* Oscilator stopped flag */
-+
-+#define PCF85263_REG_RTC_MN 0x02 /* Minutes */
-+#define PCF85263_REG_RTC_HR 0x03 /* Hours */
-+#define PCF85263_REG_RTC_DT 0x04 /* Day of month 1-31 */
-+#define PCF85263_REG_RTC_DW 0x05 /* Day of week 0-6 */
-+#define PCF85263_REG_RTC_MO 0x06 /* Month 1-12 */
-+#define PCF85263_REG_RTC_YR 0x07 /* Year 0-99 */
-+
-+#define PCF85263_REG_ALM1_SC 0x08 /* Seconds */
-+#define PCF85263_REG_ALM1_MN 0x09 /* Minutes */
-+#define PCF85263_REG_ALM1_HR 0x0a /* Hours */
-+#define PCF85263_REG_ALM1_DT 0x0b /* Day of month 1-31 */
-+#define PCF85263_REG_ALM1_MO 0x0c /* Month 1-12 */
-+
-+#define PCF85263_REG_ALM_CTL 0x10
-+#define PCF85263_REG_ALM_CTL_ALL_A1E 0x1f /* sec,min,hr,day,mon alarm 1 */
-+
-+#define PCF85263_REG_OSC 0x25
-+#define PCF85263_REG_OSC_CL_MASK (BIT(0) | BIT(1))
-+#define PCF85263_REG_OSC_CL_SHIFT 0
-+#define PCF85263_REG_OSC_OSCD_MASK (BIT(2) | BIT(3))
-+#define PCF85263_REG_OSC_OSCD_SHIFT 2
-+#define PCF85263_REG_OSC_LOWJ BIT(4)
-+#define PCF85263_REG_OSC_12H BIT(5)
-+
-+#define PCF85263_REG_PINIO 0x27
-+#define PCF85263_REG_PINIO_INTAPM_MASK (BIT(0) | BIT(1))
-+#define PCF85263_REG_PINIO_INTAPM_SHIFT 0
-+#define PCF85263_INTAPM_INTA (0x2 << PCF85263_REG_PINIO_INTAPM_SHIFT)
-+#define PCF85263_INTAPM_HIGHZ (0x3 << PCF85263_REG_PINIO_INTAPM_SHIFT)
-+#define PCF85263_REG_PINIO_TSPM_MASK (BIT(2) | BIT(3))
-+#define PCF85263_REG_PINIO_TSPM_SHIFT 2
-+#define PCF85263_TSPM_DISABLED (0x0 << PCF85263_REG_PINIO_TSPM_SHIFT)
-+#define PCF85263_TSPM_INTB (0x1 << PCF85263_REG_PINIO_TSPM_SHIFT)
-+#define PCF85263_REG_PINIO_CLKDISABLE BIT(7)
-+
-+#define PCF85263_REG_FUNCTION 0x28
-+#define PCF85263_REG_FUNCTION_COF_MASK 0x7
-+#define PCF85263_REG_FUNCTION_COF_OFF 0x7 /* No clock output */
-+
-+#define PCF85263_REG_INTA_CTL 0x29
-+#define PCF85263_REG_INTB_CTL 0x2A
-+#define PCF85263_REG_INTx_CTL_A1E BIT(4) /* Alarm 1 */
-+#define PCF85263_REG_INTx_CTL_ILP BIT(7) /* 0=pulse, 1=level */
-+
-+#define PCF85263_REG_FLAGS 0x2B
-+#define PCF85263_REG_FLAGS_A1F BIT(5)
-+
-+#define PCF85263_REG_RAM_BYTE 0x2c
-+
-+#define PCF85263_REG_STOPENABLE 0x2e
-+#define PCF85263_REG_STOPENABLE_STOP BIT(0)
-+
-+#define PCF85263_REG_RESET 0x2f /* Reset command */
-+#define PCF85263_REG_RESET_CMD_CPR 0xa4 /* Clear prescaler */
-+
-+#define PCF85263_MAX_REG 0x2f
-+
-+#define PCF85263_HR_PM BIT(5)
-+
-+enum pcf85263_irqpin {
-+ PCF85263_IRQPIN_NONE,
-+ PCF85263_IRQPIN_INTA,
-+ PCF85263_IRQPIN_INTB
-+};
-+
-+static const char *const pcf85263_irqpin_names[] = {
-+ [PCF85263_IRQPIN_NONE] = "None",
-+ [PCF85263_IRQPIN_INTA] = "INTA",
-+ [PCF85263_IRQPIN_INTB] = "INTB"
-+};
-+
-+struct pcf85263 {
-+ struct device *dev;
-+ struct rtc_device *rtc;
-+ struct regmap *regmap;
-+ enum pcf85263_irqpin irq_pin;
-+ int irq;
-+ bool mode_12h;
-+};
-+
-+/*
-+ * Helpers to convert 12h to 24h and vice versa.
-+ * Values in register are stored in BCD with a PM flag in bit 5
-+ *
-+ * 23:00 <=> 11PM <=> 0x31
-+ * 00:00 <=> 12AM <=> 0x12
-+ * 01:00 <=> 1AM <=> 0x01
-+ * 12:00 <=> 12PM <=> 0x32
-+ * 13:00 <=> 1PM <=> 0x21
-+ */
-+static int pcf85263_bcd12h_to_bin24h(int regval)
-+{
-+ int hr = bcd2bin(regval & 0x1f);
-+ bool pm = regval & PCF85263_HR_PM;
-+
-+ if (hr == 12)
-+ return pm ? 12 : 0;
-+
-+ return pm ? hr + 12 : hr;
-+}
-+
-+static int pcf85263_bin24h_to_bcd12h(int hr24)
-+{
-+ bool pm = hr24 >= 12;
-+ int hr12 = hr24 % 12;
-+
-+ if (!hr12)
-+ hr12++;
-+
-+ return bin2bcd(hr12) | pm ? 0 : PCF85263_HR_PM;
-+}
-+
-+static int pcf85263_read_time(struct device *dev, struct rtc_time *tm)
-+{
-+ struct pcf85263 *pcf85263 = dev_get_drvdata(dev);
-+ const int first = PCF85263_REG_RTC_SC;
-+ const int last = PCF85263_REG_RTC_YR;
-+ const int len = last - first + 1;
-+ u8 regs[len];
-+ u8 hr_reg;
-+ int ret;
-+
-+ ret = regmap_bulk_read(pcf85263->regmap, first, regs, len);
-+ if (ret)
-+ return ret;
-+
-+ if (regs[PCF85263_REG_RTC_SC - first] & PCF85263_REG_RTC_SC_OS) {
-+ dev_warn(dev, "Oscillator stop detected, date/time is not reliable.\n");
-+ return -EINVAL;
-+ }
-+
-+ tm->tm_sec = bcd2bin(regs[PCF85263_REG_RTC_SC - first] & 0x7f);
-+ tm->tm_min = bcd2bin(regs[PCF85263_REG_RTC_MN - first] & 0x7f);
-+
-+ hr_reg = regs[PCF85263_REG_RTC_HR - first];
-+ if (pcf85263->mode_12h)
-+ tm->tm_hour = pcf85263_bcd12h_to_bin24h(hr_reg);
-+ else
-+ tm->tm_hour = bcd2bin(hr_reg & 0x3f);
-+
-+ tm->tm_mday = bcd2bin(regs[PCF85263_REG_RTC_DT - first]);
-+ tm->tm_wday = bcd2bin(regs[PCF85263_REG_RTC_DW - first]);
-+ tm->tm_mon = bcd2bin(regs[PCF85263_REG_RTC_MO - first]) - 1;
-+ tm->tm_year = bcd2bin(regs[PCF85263_REG_RTC_YR - first]);
-+
-+ tm->tm_year += 100; /* Assume 21st century */
-+
-+ return 0;
-+}
-+
-+static int pcf85263_set_time(struct device *dev, struct rtc_time *tm)
-+{
-+ struct pcf85263 *pcf85263 = dev_get_drvdata(dev);
-+
-+ /*
-+ * Before setting time need to stop RTC and disable prescaler
-+ * Do this all in a single I2C transaction exploiting wraparound
-+ * as described in data sheet.
-+ * This means that the array below must be in register order
-+ */
-+ u8 regs[] = {
-+ PCF85263_REG_STOPENABLE_STOP, /* STOP */
-+ PCF85263_REG_RESET_CMD_CPR, /* Disable prescaler */
-+ /* Wrap around to register 0 (1/100s) */
-+ 0, /* 1/100s always zero. */
-+ bin2bcd(tm->tm_sec),
-+ bin2bcd(tm->tm_min),
-+ bin2bcd(tm->tm_hour), /* 24-hour */
-+ bin2bcd(tm->tm_mday),
-+ bin2bcd(tm->tm_wday + 1),
-+ bin2bcd(tm->tm_mon + 1),
-+ bin2bcd(tm->tm_year % 100)
-+ };
-+ int ret;
-+
-+ ret = regmap_bulk_write(pcf85263->regmap, PCF85263_REG_STOPENABLE,
-+ regs, sizeof(regs));
-+ if (ret)
-+ return ret;
-+
-+ /* As we have set the time in 24H update the hardware for that */
-+ if (pcf85263->mode_12h) {
-+ pcf85263->mode_12h = false;
-+ ret = regmap_update_bits(pcf85263->regmap, PCF85263_REG_OSC,
-+ PCF85263_REG_OSC_12H, 0);
-+ if (ret)
-+ return ret;
-+ }
-+
-+ /* Start it again */
-+ return regmap_write(pcf85263->regmap, PCF85263_REG_STOPENABLE, 0);
-+}
-+
-+static int pcf85263_enable_alarm(struct pcf85263 *pcf85263, bool enable)
-+{
-+ int reg;
-+ int ret;
-+
-+ ret = regmap_update_bits(pcf85263->regmap, PCF85263_REG_ALM_CTL,
-+ PCF85263_REG_ALM_CTL_ALL_A1E,
-+ enable ? PCF85263_REG_ALM_CTL_ALL_A1E : 0);
-+ if (ret)
-+ return ret;
-+
-+ switch (pcf85263->irq_pin) {
-+ case PCF85263_IRQPIN_NONE:
-+ return 0;
-+
-+ case PCF85263_IRQPIN_INTA:
-+ reg = PCF85263_REG_INTA_CTL;
-+ break;
-+
-+ case PCF85263_IRQPIN_INTB:
-+ reg = PCF85263_REG_INTB_CTL;
-+ break;
-+
-+ default:
-+ return -EINVAL;
-+ }
-+
-+ return regmap_update_bits(pcf85263->regmap, reg,
-+ PCF85263_REG_INTx_CTL_A1E,
-+ enable ? PCF85263_REG_INTx_CTL_A1E : 0);
-+}
-+
-+static int pcf85263_read_alarm(struct device *dev, struct rtc_wkalrm *alarm)
-+{
-+ struct pcf85263 *pcf85263 = dev_get_drvdata(dev);
-+ struct rtc_time *tm = &alarm->time;
-+ const int first = PCF85263_REG_ALM1_SC;
-+ const int last = PCF85263_REG_ALM1_MO;
-+ const int len = last - first + 1;
-+ u8 regs[len];
-+ u8 hr_reg;
-+ unsigned int regval;
-+ int ret;
-+
-+ ret = regmap_bulk_read(pcf85263->regmap, first, regs, len);
-+ if (ret)
-+ return ret;
-+
-+ tm->tm_sec = bcd2bin(regs[PCF85263_REG_ALM1_SC - first] & 0x7f);
-+ tm->tm_min = bcd2bin(regs[PCF85263_REG_ALM1_MN - first] & 0x7f);
-+
-+ hr_reg = regs[PCF85263_REG_ALM1_HR - first];
-+ if (pcf85263->mode_12h)
-+ tm->tm_hour = pcf85263_bcd12h_to_bin24h(hr_reg);
-+ else
-+ tm->tm_hour = bcd2bin(hr_reg & 0x3f);
-+
-+ tm->tm_mday = bcd2bin(regs[PCF85263_REG_ALM1_DT - first]);
-+ tm->tm_mon = bcd2bin(regs[PCF85263_REG_ALM1_MO - first]) - 1;
-+ tm->tm_year = -1;
-+ tm->tm_wday = -1;
-+
-+ ret = regmap_read(pcf85263->regmap, PCF85263_REG_ALM_CTL, &regval);
-+ if (ret)
-+ return ret;
-+ alarm->enabled = !!(regval & PCF85263_REG_ALM_CTL_ALL_A1E);
-+
-+ ret = regmap_read(pcf85263->regmap, PCF85263_REG_FLAGS, &regval);
-+ if (ret)
-+ return ret;
-+ alarm->pending = !!(regval & PCF85263_REG_FLAGS_A1F);
-+
-+ return 0;
-+}
-+
-+static int pcf85263_set_alarm(struct device *dev, struct rtc_wkalrm *alarm)
-+{
-+ struct pcf85263 *pcf85263 = dev_get_drvdata(dev);
-+ struct rtc_time *tm = &alarm->time;
-+ const int first = PCF85263_REG_ALM1_SC;
-+ const int last = PCF85263_REG_ALM1_MO;
-+ const int len = last - first + 1;
-+ u8 regs[len];
-+ int ret;
-+
-+ /* Disable alarm comparison during update */
-+ ret = pcf85263_enable_alarm(pcf85263, false);
-+ if (ret)
-+ return ret;
-+
-+ /* Clear any pending alarm (write 0=>clr, 1=>no change) */
-+ ret = regmap_write(pcf85263->regmap, PCF85263_REG_FLAGS,
-+ (unsigned int)(~PCF85263_REG_FLAGS_A1F));
-+ if (ret)
-+ return ret;
-+
-+ /* Set the alarm time registers */
-+ regs[PCF85263_REG_ALM1_SC - first] = bin2bcd(tm->tm_sec);
-+ regs[PCF85263_REG_ALM1_MN - first] = bin2bcd(tm->tm_min);
-+ regs[PCF85263_REG_ALM1_HR - first] = pcf85263->mode_12h ?
-+ pcf85263_bin24h_to_bcd12h(tm->tm_hour) :
-+ bin2bcd(tm->tm_hour);
-+ regs[PCF85263_REG_ALM1_DT - first] = bin2bcd(tm->tm_mday);
-+ regs[PCF85263_REG_ALM1_MO - first] = bin2bcd(tm->tm_mon + 1);
-+
-+ ret = regmap_bulk_write(pcf85263->regmap, first, regs, sizeof(regs));
-+ if (ret)
-+ return ret;
-+
-+ if (alarm->enabled)
-+ ret = pcf85263_enable_alarm(pcf85263, true);
-+
-+ return ret;
-+}
-+
-+static int pcf85263_alarm_irq_enable(struct device *dev, unsigned int enable)
-+{
-+ struct pcf85263 *pcf85263 = dev_get_drvdata(dev);
-+
-+ return pcf85263_enable_alarm(pcf85263, !!enable);
-+}
-+
-+static irqreturn_t pcf85263_irq(int irq, void *data)
-+{
-+ struct pcf85263 *pcf85263 = data;
-+ unsigned int regval;
-+ int ret;
-+
-+ ret = regmap_read(pcf85263->regmap, PCF85263_REG_FLAGS, &regval);
-+ if (ret)
-+ return IRQ_NONE;
-+
-+ if (regval & PCF85263_REG_FLAGS_A1F) {
-+ regmap_write(pcf85263->regmap, PCF85263_REG_FLAGS,
-+ (unsigned int)(~PCF85263_REG_FLAGS_A1F));
-+
-+ rtc_update_irq(pcf85263->rtc, 1, RTC_IRQF | RTC_AF);
-+
-+ return IRQ_HANDLED;
-+ }
-+
-+ return IRQ_NONE;
-+}
-+
-+static int pcf85263_check_osc_stopped(struct pcf85263 *pcf85263)
-+{
-+ unsigned int regval;
-+ int ret;
-+
-+ ret = regmap_read(pcf85263->regmap, PCF85263_REG_RTC_SC, &regval);
-+ if (ret)
-+ return ret;
-+
-+ ret = regval & PCF85263_REG_RTC_SC_OS ? 1 : 0;
-+ if (ret)
-+ dev_warn(pcf85263->dev, "Oscillator stop detected, date/time is not reliable.\n");
-+
-+ return ret;
-+}
-+
-+#ifdef CONFIG_RTC_INTF_DEV
-+static int pcf85263_ioctl(struct device *dev,
-+ unsigned int cmd, unsigned long arg)
-+{
-+ struct pcf85263 *pcf85263 = dev_get_drvdata(dev);
-+ int ret;
-+
-+ switch (cmd) {
-+ case RTC_VL_READ:
-+ ret = pcf85263_check_osc_stopped(pcf85263);
-+ if (ret < 0)
-+ return ret;
-+
-+ if (copy_to_user((void __user *)arg, &ret, sizeof(int)))
-+ return -EFAULT;
-+ return 0;
-+
-+ case RTC_VL_CLR:
-+ return regmap_update_bits(pcf85263->regmap,
-+ PCF85263_REG_RTC_SC,
-+ PCF85263_REG_RTC_SC_OS, 0);
-+ default:
-+ return -ENOIOCTLCMD;
-+ }
-+}
-+#else
-+#define pcf85263_ioctl NULL
-+#endif
-+
-+static int pcf85263_init_hw(struct pcf85263 *pcf85263)
-+{
-+ struct device_node *np = pcf85263->dev->of_node;
-+ unsigned int regval;
-+ u32 propval;
-+ int ret;
-+
-+ /* Determine if oscilator has been stopped (probably low power) */
-+ ret = pcf85263_check_osc_stopped(pcf85263);
-+ if (ret < 0) {
-+ /* Log here since this is the first hw access on probe */
-+ dev_err(pcf85263->dev, "Unable to read register\n");
-+
-+ return ret;
-+ }
-+
-+ /* Determine 12/24H mode */
-+ ret = regmap_read(pcf85263->regmap, PCF85263_REG_OSC, &regval);
-+ if (ret)
-+ return ret;
-+ pcf85263->mode_12h = !!(regval & PCF85263_REG_OSC_12H);
-+
-+ /* Set oscilator register */
-+ regval &= ~PCF85263_REG_OSC_12H; /* keep current 12/24 h setting */
-+
-+ propval = PCF85263_QUARTZCAP_12p5pF;
-+ of_property_read_u32(np, "quartz-load-capacitance", &propval);
-+ regval |= ((propval << PCF85263_REG_OSC_CL_SHIFT)
-+ & PCF85263_REG_OSC_CL_MASK);
-+
-+ propval = PCF85263_QUARTZDRIVE_NORMAL;
-+ of_property_read_u32(np, "quartz-drive-strength", &propval);
-+ regval |= ((propval << PCF85263_REG_OSC_OSCD_SHIFT)
-+ & PCF85263_REG_OSC_OSCD_MASK);
-+
-+ if (of_property_read_bool(np, "quartz-low-jitter"))
-+ regval |= PCF85263_REG_OSC_LOWJ;
-+
-+ ret = regmap_write(pcf85263->regmap, PCF85263_REG_OSC, regval);
-+ if (ret)
-+ return ret;
-+
-+ /* Set function register (RTC mode, 1s tick, clock output static) */
-+ ret = regmap_write(pcf85263->regmap, PCF85263_REG_FUNCTION,
-+ PCF85263_REG_FUNCTION_COF_OFF);
-+ if (ret)
-+ return ret;
-+
-+ /* Set all interrupts to disabled, level mode */
-+ ret = regmap_write(pcf85263->regmap, PCF85263_REG_INTA_CTL,
-+ PCF85263_REG_INTx_CTL_ILP);
-+ if (ret)
-+ return ret;
-+ ret = regmap_write(pcf85263->regmap, PCF85263_REG_INTB_CTL,
-+ PCF85263_REG_INTx_CTL_ILP);
-+ if (ret)
-+ return ret;
-+
-+ /* Setup IO pin config register */
-+ regval = PCF85263_REG_PINIO_CLKDISABLE;
-+ switch (pcf85263->irq_pin) {
-+ case PCF85263_IRQPIN_INTA:
-+ regval |= (PCF85263_INTAPM_INTA | PCF85263_TSPM_DISABLED);
-+ break;
-+ case PCF85263_IRQPIN_INTB:
-+ regval |= (PCF85263_INTAPM_HIGHZ | PCF85263_TSPM_INTB);
-+ break;
-+ case PCF85263_IRQPIN_NONE:
-+ regval |= (PCF85263_INTAPM_HIGHZ | PCF85263_TSPM_DISABLED);
-+ break;
-+ }
-+ ret = regmap_write(pcf85263->regmap, PCF85263_REG_PINIO, regval);
-+
-+ return ret;
-+}
-+
-+static const struct rtc_class_ops rtc_ops = {
-+ .ioctl = pcf85263_ioctl,
-+ .read_time = pcf85263_read_time,
-+ .set_time = pcf85263_set_time,
-+ .read_alarm = pcf85263_read_alarm,
-+ .set_alarm = pcf85263_set_alarm,
-+ .alarm_irq_enable = pcf85263_alarm_irq_enable,
-+};
-+
-+static const struct regmap_config pcf85263_regmap_cfg = {
-+ .reg_bits = 8,
-+ .val_bits = 8,
-+ .max_register = PCF85263_MAX_REG,
-+};
-+
-+/*
-+ * On some boards the interrupt line may not be wired to the CPU but only to
-+ * a power supply circuit.
-+ * In that case no interrupt will be specified in the device tree but the
-+ * wakeup-source DT property may be used to enable wakeup programming in
-+ * sysfs
-+ */
-+static bool pcf85263_can_wakeup_machine(struct pcf85263 *pcf85263)
-+{
-+ return pcf85263->irq ||
-+ of_property_read_bool(pcf85263->dev->of_node, "wakeup-source");
-+}
-+
-+static int pcf85263_probe(struct i2c_client *client,
-+ const struct i2c_device_id *id)
-+{
-+ struct device *dev = &client->dev;
-+ struct pcf85263 *pcf85263;
-+ int ret;
-+
-+ if (!i2c_check_functionality(client->adapter, I2C_FUNC_I2C |
-+ I2C_FUNC_SMBUS_BYTE_DATA |
-+ I2C_FUNC_SMBUS_I2C_BLOCK))
-+ return -ENODEV;
-+
-+ pcf85263 = devm_kzalloc(dev, sizeof(*pcf85263), GFP_KERNEL);
-+ if (!pcf85263)
-+ return -ENOMEM;
-+
-+ pcf85263->dev = dev;
-+ pcf85263->irq = client->irq;
-+ dev_set_drvdata(dev, pcf85263);
-+
-+ pcf85263->regmap = devm_regmap_init_i2c(client, &pcf85263_regmap_cfg);
-+ if (IS_ERR(pcf85263->regmap)) {
-+ ret = PTR_ERR(pcf85263->regmap);
-+ dev_err(dev, "regmap allocation failed (%d)\n", ret);
-+
-+ return ret;
-+ }
-+
-+ /* Determine which interrupt pin the board uses */
-+ if (pcf85263_can_wakeup_machine(pcf85263)) {
-+ if (of_property_match_string(dev->of_node,
-+ "interrupt-names", "INTB") >= 0)
-+ pcf85263->irq_pin = PCF85263_IRQPIN_INTB;
-+ else
-+ pcf85263->irq_pin = PCF85263_IRQPIN_INTA;
-+ } else {
-+ pcf85263->irq_pin = PCF85263_IRQPIN_NONE;
-+ }
-+
-+ ret = pcf85263_init_hw(pcf85263);
-+ if (ret)
-+ return ret;
-+
-+ if (pcf85263->irq) {
-+ ret = devm_request_threaded_irq(dev, pcf85263->irq, NULL,
-+ pcf85263_irq,
-+ IRQF_ONESHOT,
-+ dev->driver->name, pcf85263);
-+ if (ret) {
-+ dev_err(dev, "irq %d unavailable (%d)\n",
-+ pcf85263->irq, ret);
-+ pcf85263->irq = 0;
-+ }
-+ }
-+
-+ if (pcf85263_can_wakeup_machine(pcf85263))
-+ device_init_wakeup(dev, true);
-+
-+ pcf85263->rtc = devm_rtc_device_register(dev, dev->driver->name,
-+ &rtc_ops, THIS_MODULE);
-+ ret = PTR_ERR_OR_ZERO(pcf85263->rtc);
-+ if (ret)
-+ return ret;
-+
-+ /* We cannot support UIE mode if we do not have an IRQ line */
-+ if (!pcf85263->irq)
-+ pcf85263->rtc->uie_unsupported = 1;
-+
-+ dev_info(pcf85263->dev,
-+ "PCF85263 RTC (irqpin=%s irq=%d)\n",
-+ pcf85263_irqpin_names[pcf85263->irq_pin],
-+ pcf85263->irq);
-+
-+ return 0;
-+}
-+
-+static int pcf85263_remove(struct i2c_client *client)
-+{
-+ struct pcf85263 *pcf85263 = i2c_get_clientdata(client);
-+
-+ if (pcf85263_can_wakeup_machine(pcf85263))
-+ device_init_wakeup(pcf85263->dev, false);
-+
-+ return 0;
-+}
-+
-+#ifdef CONFIG_PM_SLEEP
-+static int pcf85263_suspend(struct device *dev)
-+{
-+ struct pcf85263 *pcf85263 = dev_get_drvdata(dev);
-+ int ret = 0;
-+
-+ if (device_may_wakeup(dev))
-+ ret = enable_irq_wake(pcf85263->irq);
-+
-+ return ret;
-+}
-+
-+static int pcf85263_resume(struct device *dev)
-+{
-+ struct pcf85263 *pcf85263 = dev_get_drvdata(dev);
-+ int ret = 0;
-+
-+ if (device_may_wakeup(dev))
-+ ret = disable_irq_wake(pcf85263->irq);
-+
-+ return ret;
-+}
-+
-+#endif
-+
-+static const struct i2c_device_id pcf85263_id[] = {
-+ { "pcf85263", 0 },
-+ { }
-+};
-+MODULE_DEVICE_TABLE(i2c, pcf85263_id);
-+
-+#ifdef CONFIG_OF
-+static const struct of_device_id pcf85263_of_match[] = {
-+ { .compatible = "nxp,pcf85263" },
-+ {}
-+};
-+MODULE_DEVICE_TABLE(of, pcf85263_of_match);
-+#endif
-+
-+static SIMPLE_DEV_PM_OPS(pcf85263_pm_ops, pcf85263_suspend, pcf85263_resume);
-+
-+static struct i2c_driver pcf85263_driver = {
-+ .driver = {
-+ .name = "rtc-pcf85263",
-+ .of_match_table = of_match_ptr(pcf85263_of_match),
-+ .pm = &pcf85263_pm_ops,
-+ },
-+ .probe = pcf85263_probe,
-+ .remove = pcf85263_remove,
-+ .id_table = pcf85263_id,
-+};
-+
-+module_i2c_driver(pcf85263_driver);
-+
-+MODULE_AUTHOR("Martin Fuzzey <mfuzzey@parkeon.com>");
-+MODULE_DESCRIPTION("PCF85263 RTC Driver");
-+MODULE_LICENSE("GPL");
-+
diff --git a/target/linux/layerscape/patches-4.9/815-spi-support-layerscape.patch b/target/linux/layerscape/patches-4.9/815-spi-support-layerscape.patch
deleted file mode 100644
index b798a366b0..0000000000
--- a/target/linux/layerscape/patches-4.9/815-spi-support-layerscape.patch
+++ /dev/null
@@ -1,427 +0,0 @@
-From 027b679f248f15dea36c6cd6782d6643e2151057 Mon Sep 17 00:00:00 2001
-From: Yangbo Lu <yangbo.lu@nxp.com>
-Date: Wed, 17 Jan 2018 15:39:43 +0800
-Subject: [PATCH 27/30] spi: support layerscape
-
-This is an integrated patch for layerscape dspi support.
-
-Signed-off-by: Christophe JAILLET <christophe.jaillet@wanadoo.fr>
-Signed-off-by: Sanchayan Maity <maitysanchayan@gmail.com>
-Signed-off-by: Geert Uytterhoeven <geert@linux-m68k.org>
-Signed-off-by: Sanchayan Maity <maitysanchayan@gmail.com>
-Signed-off-by: Yangbo Lu <yangbo.lu@nxp.com>
----
- drivers/spi/spi-fsl-dspi.c | 309 ++++++++++++++++++++++++++++++++++++++++++++-
- 1 file changed, 304 insertions(+), 5 deletions(-)
-
---- a/drivers/spi/spi-fsl-dspi.c
-+++ b/drivers/spi/spi-fsl-dspi.c
-@@ -15,6 +15,8 @@
-
- #include <linux/clk.h>
- #include <linux/delay.h>
-+#include <linux/dmaengine.h>
-+#include <linux/dma-mapping.h>
- #include <linux/err.h>
- #include <linux/errno.h>
- #include <linux/interrupt.h>
-@@ -40,6 +42,7 @@
- #define TRAN_STATE_WORD_ODD_NUM 0x04
-
- #define DSPI_FIFO_SIZE 4
-+#define DSPI_DMA_BUFSIZE (DSPI_FIFO_SIZE * 1024)
-
- #define SPI_MCR 0x00
- #define SPI_MCR_MASTER (1 << 31)
-@@ -72,6 +75,11 @@
- #define SPI_SR_TCFQF 0x80000000
- #define SPI_SR_CLEAR 0xdaad0000
-
-+#define SPI_RSER_TFFFE BIT(25)
-+#define SPI_RSER_TFFFD BIT(24)
-+#define SPI_RSER_RFDFE BIT(17)
-+#define SPI_RSER_RFDFD BIT(16)
-+
- #define SPI_RSER 0x30
- #define SPI_RSER_EOQFE 0x10000000
- #define SPI_RSER_TCFQE 0x80000000
-@@ -109,6 +117,8 @@
-
- #define SPI_TCR_TCNT_MAX 0x10000
-
-+#define DMA_COMPLETION_TIMEOUT msecs_to_jiffies(3000)
-+
- struct chip_data {
- u32 mcr_val;
- u32 ctar_val;
-@@ -118,6 +128,7 @@ struct chip_data {
- enum dspi_trans_mode {
- DSPI_EOQ_MODE = 0,
- DSPI_TCFQ_MODE,
-+ DSPI_DMA_MODE,
- };
-
- struct fsl_dspi_devtype_data {
-@@ -126,7 +137,7 @@ struct fsl_dspi_devtype_data {
- };
-
- static const struct fsl_dspi_devtype_data vf610_data = {
-- .trans_mode = DSPI_EOQ_MODE,
-+ .trans_mode = DSPI_DMA_MODE,
- .max_clock_factor = 2,
- };
-
-@@ -140,6 +151,23 @@ static const struct fsl_dspi_devtype_dat
- .max_clock_factor = 8,
- };
-
-+struct fsl_dspi_dma {
-+ /* Length of transfer in words of DSPI_FIFO_SIZE */
-+ u32 curr_xfer_len;
-+
-+ u32 *tx_dma_buf;
-+ struct dma_chan *chan_tx;
-+ dma_addr_t tx_dma_phys;
-+ struct completion cmd_tx_complete;
-+ struct dma_async_tx_descriptor *tx_desc;
-+
-+ u32 *rx_dma_buf;
-+ struct dma_chan *chan_rx;
-+ dma_addr_t rx_dma_phys;
-+ struct completion cmd_rx_complete;
-+ struct dma_async_tx_descriptor *rx_desc;
-+};
-+
- struct fsl_dspi {
- struct spi_master *master;
- struct platform_device *pdev;
-@@ -166,8 +194,11 @@ struct fsl_dspi {
- u32 waitflags;
-
- u32 spi_tcnt;
-+ struct fsl_dspi_dma *dma;
- };
-
-+static u32 dspi_data_to_pushr(struct fsl_dspi *dspi, int tx_word);
-+
- static inline int is_double_byte_mode(struct fsl_dspi *dspi)
- {
- unsigned int val;
-@@ -177,6 +208,255 @@ static inline int is_double_byte_mode(st
- return ((val & SPI_FRAME_BITS_MASK) == SPI_FRAME_BITS(8)) ? 0 : 1;
- }
-
-+static void dspi_tx_dma_callback(void *arg)
-+{
-+ struct fsl_dspi *dspi = arg;
-+ struct fsl_dspi_dma *dma = dspi->dma;
-+
-+ complete(&dma->cmd_tx_complete);
-+}
-+
-+static void dspi_rx_dma_callback(void *arg)
-+{
-+ struct fsl_dspi *dspi = arg;
-+ struct fsl_dspi_dma *dma = dspi->dma;
-+ int rx_word;
-+ int i;
-+ u16 d;
-+
-+ rx_word = is_double_byte_mode(dspi);
-+
-+ if (!(dspi->dataflags & TRAN_STATE_RX_VOID)) {
-+ for (i = 0; i < dma->curr_xfer_len; i++) {
-+ d = dspi->dma->rx_dma_buf[i];
-+ rx_word ? (*(u16 *)dspi->rx = d) :
-+ (*(u8 *)dspi->rx = d);
-+ dspi->rx += rx_word + 1;
-+ }
-+ }
-+
-+ complete(&dma->cmd_rx_complete);
-+}
-+
-+static int dspi_next_xfer_dma_submit(struct fsl_dspi *dspi)
-+{
-+ struct fsl_dspi_dma *dma = dspi->dma;
-+ struct device *dev = &dspi->pdev->dev;
-+ int time_left;
-+ int tx_word;
-+ int i;
-+
-+ tx_word = is_double_byte_mode(dspi);
-+
-+ for (i = 0; i < dma->curr_xfer_len; i++) {
-+ dspi->dma->tx_dma_buf[i] = dspi_data_to_pushr(dspi, tx_word);
-+ if ((dspi->cs_change) && (!dspi->len))
-+ dspi->dma->tx_dma_buf[i] &= ~SPI_PUSHR_CONT;
-+ }
-+
-+ dma->tx_desc = dmaengine_prep_slave_single(dma->chan_tx,
-+ dma->tx_dma_phys,
-+ dma->curr_xfer_len *
-+ DMA_SLAVE_BUSWIDTH_4_BYTES,
-+ DMA_MEM_TO_DEV,
-+ DMA_PREP_INTERRUPT | DMA_CTRL_ACK);
-+ if (!dma->tx_desc) {
-+ dev_err(dev, "Not able to get desc for DMA xfer\n");
-+ return -EIO;
-+ }
-+
-+ dma->tx_desc->callback = dspi_tx_dma_callback;
-+ dma->tx_desc->callback_param = dspi;
-+ if (dma_submit_error(dmaengine_submit(dma->tx_desc))) {
-+ dev_err(dev, "DMA submit failed\n");
-+ return -EINVAL;
-+ }
-+
-+ dma->rx_desc = dmaengine_prep_slave_single(dma->chan_rx,
-+ dma->rx_dma_phys,
-+ dma->curr_xfer_len *
-+ DMA_SLAVE_BUSWIDTH_4_BYTES,
-+ DMA_DEV_TO_MEM,
-+ DMA_PREP_INTERRUPT | DMA_CTRL_ACK);
-+ if (!dma->rx_desc) {
-+ dev_err(dev, "Not able to get desc for DMA xfer\n");
-+ return -EIO;
-+ }
-+
-+ dma->rx_desc->callback = dspi_rx_dma_callback;
-+ dma->rx_desc->callback_param = dspi;
-+ if (dma_submit_error(dmaengine_submit(dma->rx_desc))) {
-+ dev_err(dev, "DMA submit failed\n");
-+ return -EINVAL;
-+ }
-+
-+ reinit_completion(&dspi->dma->cmd_rx_complete);
-+ reinit_completion(&dspi->dma->cmd_tx_complete);
-+
-+ dma_async_issue_pending(dma->chan_rx);
-+ dma_async_issue_pending(dma->chan_tx);
-+
-+ time_left = wait_for_completion_timeout(&dspi->dma->cmd_tx_complete,
-+ DMA_COMPLETION_TIMEOUT);
-+ if (time_left == 0) {
-+ dev_err(dev, "DMA tx timeout\n");
-+ dmaengine_terminate_all(dma->chan_tx);
-+ dmaengine_terminate_all(dma->chan_rx);
-+ return -ETIMEDOUT;
-+ }
-+
-+ time_left = wait_for_completion_timeout(&dspi->dma->cmd_rx_complete,
-+ DMA_COMPLETION_TIMEOUT);
-+ if (time_left == 0) {
-+ dev_err(dev, "DMA rx timeout\n");
-+ dmaengine_terminate_all(dma->chan_tx);
-+ dmaengine_terminate_all(dma->chan_rx);
-+ return -ETIMEDOUT;
-+ }
-+
-+ return 0;
-+}
-+
-+static int dspi_dma_xfer(struct fsl_dspi *dspi)
-+{
-+ struct fsl_dspi_dma *dma = dspi->dma;
-+ struct device *dev = &dspi->pdev->dev;
-+ int curr_remaining_bytes;
-+ int bytes_per_buffer;
-+ int word = 1;
-+ int ret = 0;
-+
-+ if (is_double_byte_mode(dspi))
-+ word = 2;
-+ curr_remaining_bytes = dspi->len;
-+ bytes_per_buffer = DSPI_DMA_BUFSIZE / DSPI_FIFO_SIZE;
-+ while (curr_remaining_bytes) {
-+ /* Check if current transfer fits the DMA buffer */
-+ dma->curr_xfer_len = curr_remaining_bytes / word;
-+ if (dma->curr_xfer_len > bytes_per_buffer)
-+ dma->curr_xfer_len = bytes_per_buffer;
-+
-+ ret = dspi_next_xfer_dma_submit(dspi);
-+ if (ret) {
-+ dev_err(dev, "DMA transfer failed\n");
-+ goto exit;
-+
-+ } else {
-+ curr_remaining_bytes -= dma->curr_xfer_len * word;
-+ if (curr_remaining_bytes < 0)
-+ curr_remaining_bytes = 0;
-+ }
-+ }
-+
-+exit:
-+ return ret;
-+}
-+
-+static int dspi_request_dma(struct fsl_dspi *dspi, phys_addr_t phy_addr)
-+{
-+ struct fsl_dspi_dma *dma;
-+ struct dma_slave_config cfg;
-+ struct device *dev = &dspi->pdev->dev;
-+ int ret;
-+
-+ dma = devm_kzalloc(dev, sizeof(*dma), GFP_KERNEL);
-+ if (!dma)
-+ return -ENOMEM;
-+
-+ dma->chan_rx = dma_request_slave_channel(dev, "rx");
-+ if (!dma->chan_rx) {
-+ dev_err(dev, "rx dma channel not available\n");
-+ ret = -ENODEV;
-+ return ret;
-+ }
-+
-+ dma->chan_tx = dma_request_slave_channel(dev, "tx");
-+ if (!dma->chan_tx) {
-+ dev_err(dev, "tx dma channel not available\n");
-+ ret = -ENODEV;
-+ goto err_tx_channel;
-+ }
-+
-+ dma->tx_dma_buf = dma_alloc_coherent(dev, DSPI_DMA_BUFSIZE,
-+ &dma->tx_dma_phys, GFP_KERNEL);
-+ if (!dma->tx_dma_buf) {
-+ ret = -ENOMEM;
-+ goto err_tx_dma_buf;
-+ }
-+
-+ dma->rx_dma_buf = dma_alloc_coherent(dev, DSPI_DMA_BUFSIZE,
-+ &dma->rx_dma_phys, GFP_KERNEL);
-+ if (!dma->rx_dma_buf) {
-+ ret = -ENOMEM;
-+ goto err_rx_dma_buf;
-+ }
-+
-+ cfg.src_addr = phy_addr + SPI_POPR;
-+ cfg.dst_addr = phy_addr + SPI_PUSHR;
-+ cfg.src_addr_width = DMA_SLAVE_BUSWIDTH_4_BYTES;
-+ cfg.dst_addr_width = DMA_SLAVE_BUSWIDTH_4_BYTES;
-+ cfg.src_maxburst = 1;
-+ cfg.dst_maxburst = 1;
-+
-+ cfg.direction = DMA_DEV_TO_MEM;
-+ ret = dmaengine_slave_config(dma->chan_rx, &cfg);
-+ if (ret) {
-+ dev_err(dev, "can't configure rx dma channel\n");
-+ ret = -EINVAL;
-+ goto err_slave_config;
-+ }
-+
-+ cfg.direction = DMA_MEM_TO_DEV;
-+ ret = dmaengine_slave_config(dma->chan_tx, &cfg);
-+ if (ret) {
-+ dev_err(dev, "can't configure tx dma channel\n");
-+ ret = -EINVAL;
-+ goto err_slave_config;
-+ }
-+
-+ dspi->dma = dma;
-+ init_completion(&dma->cmd_tx_complete);
-+ init_completion(&dma->cmd_rx_complete);
-+
-+ return 0;
-+
-+err_slave_config:
-+ dma_free_coherent(dev, DSPI_DMA_BUFSIZE,
-+ dma->rx_dma_buf, dma->rx_dma_phys);
-+err_rx_dma_buf:
-+ dma_free_coherent(dev, DSPI_DMA_BUFSIZE,
-+ dma->tx_dma_buf, dma->tx_dma_phys);
-+err_tx_dma_buf:
-+ dma_release_channel(dma->chan_tx);
-+err_tx_channel:
-+ dma_release_channel(dma->chan_rx);
-+
-+ devm_kfree(dev, dma);
-+ dspi->dma = NULL;
-+
-+ return ret;
-+}
-+
-+static void dspi_release_dma(struct fsl_dspi *dspi)
-+{
-+ struct fsl_dspi_dma *dma = dspi->dma;
-+ struct device *dev = &dspi->pdev->dev;
-+
-+ if (dma) {
-+ if (dma->chan_tx) {
-+ dma_unmap_single(dev, dma->tx_dma_phys,
-+ DSPI_DMA_BUFSIZE, DMA_TO_DEVICE);
-+ dma_release_channel(dma->chan_tx);
-+ }
-+
-+ if (dma->chan_rx) {
-+ dma_unmap_single(dev, dma->rx_dma_phys,
-+ DSPI_DMA_BUFSIZE, DMA_FROM_DEVICE);
-+ dma_release_channel(dma->chan_rx);
-+ }
-+ }
-+}
-+
- static void hz_to_spi_baud(char *pbr, char *br, int speed_hz,
- unsigned long clkrate)
- {
-@@ -425,6 +705,12 @@ static int dspi_transfer_one_message(str
- regmap_write(dspi->regmap, SPI_RSER, SPI_RSER_TCFQE);
- dspi_tcfq_write(dspi);
- break;
-+ case DSPI_DMA_MODE:
-+ regmap_write(dspi->regmap, SPI_RSER,
-+ SPI_RSER_TFFFE | SPI_RSER_TFFFD |
-+ SPI_RSER_RFDFE | SPI_RSER_RFDFD);
-+ status = dspi_dma_xfer(dspi);
-+ break;
- default:
- dev_err(&dspi->pdev->dev, "unsupported trans_mode %u\n",
- trans_mode);
-@@ -432,9 +718,13 @@ static int dspi_transfer_one_message(str
- goto out;
- }
-
-- if (wait_event_interruptible(dspi->waitq, dspi->waitflags))
-- dev_err(&dspi->pdev->dev, "wait transfer complete fail!\n");
-- dspi->waitflags = 0;
-+ if (trans_mode != DSPI_DMA_MODE) {
-+ if (wait_event_interruptible(dspi->waitq,
-+ dspi->waitflags))
-+ dev_err(&dspi->pdev->dev,
-+ "wait transfer complete fail!\n");
-+ dspi->waitflags = 0;
-+ }
-
- if (transfer->delay_usecs)
- udelay(transfer->delay_usecs);
-@@ -712,7 +1002,8 @@ static int dspi_probe(struct platform_de
- if (IS_ERR(dspi->regmap)) {
- dev_err(&pdev->dev, "failed to init regmap: %ld\n",
- PTR_ERR(dspi->regmap));
-- return PTR_ERR(dspi->regmap);
-+ ret = PTR_ERR(dspi->regmap);
-+ goto out_master_put;
- }
-
- dspi->clk = devm_clk_get(&pdev->dev, "dspi");
-@@ -740,6 +1031,13 @@ static int dspi_probe(struct platform_de
- goto out_clk_put;
- }
-
-+ if (dspi->devtype_data->trans_mode == DSPI_DMA_MODE) {
-+ if (dspi_request_dma(dspi, res->start)) {
-+ dev_err(&pdev->dev, "can't get dma channels\n");
-+ goto out_clk_put;
-+ }
-+ }
-+
- master->max_speed_hz =
- clk_get_rate(dspi->clk) / dspi->devtype_data->max_clock_factor;
-
-@@ -768,6 +1066,7 @@ static int dspi_remove(struct platform_d
- struct fsl_dspi *dspi = spi_master_get_devdata(master);
-
- /* Disconnect from the SPI framework */
-+ dspi_release_dma(dspi);
- clk_disable_unprepare(dspi->clk);
- spi_unregister_master(dspi->master);
-
diff --git a/target/linux/layerscape/patches-4.9/816-tty-serial-support-layerscape.patch b/target/linux/layerscape/patches-4.9/816-tty-serial-support-layerscape.patch
deleted file mode 100644
index 3271563f7f..0000000000
--- a/target/linux/layerscape/patches-4.9/816-tty-serial-support-layerscape.patch
+++ /dev/null
@@ -1,162 +0,0 @@
-From c35aec61e5bb0faafb2847a0d750ebd7345a4b0f Mon Sep 17 00:00:00 2001
-From: Yangbo Lu <yangbo.lu@nxp.com>
-Date: Wed, 17 Jan 2018 15:40:24 +0800
-Subject: [PATCH 28/30] tty: serial: support layerscape
-
-This is an integrated patch for layerscape uart support.
-
-Signed-off-by: Nikita Yushchenko <nikita.yoush@cogentembedded.com>
-Signed-off-by: Yuan Yao <yao.yuan@nxp.com>
-Signed-off-by: Stefan Agner <stefan@agner.ch>
-Signed-off-by: Yangbo Lu <yangbo.lu@nxp.com>
----
- drivers/tty/serial/fsl_lpuart.c | 66 ++++++++++++++++++++++++++++-------------
- 1 file changed, 46 insertions(+), 20 deletions(-)
-
---- a/drivers/tty/serial/fsl_lpuart.c
-+++ b/drivers/tty/serial/fsl_lpuart.c
-@@ -231,6 +231,8 @@
- #define DEV_NAME "ttyLP"
- #define UART_NR 6
-
-+static DECLARE_BITMAP(linemap, UART_NR);
-+
- struct lpuart_port {
- struct uart_port port;
- struct clk *clk;
-@@ -1349,6 +1351,18 @@ lpuart_set_termios(struct uart_port *por
- /* ask the core to calculate the divisor */
- baud = uart_get_baud_rate(port, termios, old, 50, port->uartclk / 16);
-
-+ /*
-+ * Need to update the Ring buffer length according to the selected
-+ * baud rate and restart Rx DMA path.
-+ *
-+ * Since timer function acqures sport->port.lock, need to stop before
-+ * acquring same lock because otherwise del_timer_sync() can deadlock.
-+ */
-+ if (old && sport->lpuart_dma_rx_use) {
-+ del_timer_sync(&sport->lpuart_timer);
-+ lpuart_dma_rx_free(&sport->port);
-+ }
-+
- spin_lock_irqsave(&sport->port.lock, flags);
-
- sport->port.read_status_mask = 0;
-@@ -1398,22 +1412,11 @@ lpuart_set_termios(struct uart_port *por
- /* restore control register */
- writeb(old_cr2, sport->port.membase + UARTCR2);
-
-- /*
-- * If new baud rate is set, we will also need to update the Ring buffer
-- * length according to the selected baud rate and restart Rx DMA path.
-- */
-- if (old) {
-- if (sport->lpuart_dma_rx_use) {
-- del_timer_sync(&sport->lpuart_timer);
-- lpuart_dma_rx_free(&sport->port);
-- }
--
-- if (sport->dma_rx_chan && !lpuart_start_rx_dma(sport)) {
-- sport->lpuart_dma_rx_use = true;
-+ if (old && sport->lpuart_dma_rx_use) {
-+ if (!lpuart_start_rx_dma(sport))
- rx_dma_timer_init(sport);
-- } else {
-+ else
- sport->lpuart_dma_rx_use = false;
-- }
- }
-
- spin_unlock_irqrestore(&sport->port.lock, flags);
-@@ -1641,6 +1644,13 @@ lpuart_console_write(struct console *co,
- {
- struct lpuart_port *sport = lpuart_ports[co->index];
- unsigned char old_cr2, cr2;
-+ unsigned long flags;
-+ int locked = 1;
-+
-+ if (sport->port.sysrq || oops_in_progress)
-+ locked = spin_trylock_irqsave(&sport->port.lock, flags);
-+ else
-+ spin_lock_irqsave(&sport->port.lock, flags);
-
- /* first save CR2 and then disable interrupts */
- cr2 = old_cr2 = readb(sport->port.membase + UARTCR2);
-@@ -1655,6 +1665,9 @@ lpuart_console_write(struct console *co,
- barrier();
-
- writeb(old_cr2, sport->port.membase + UARTCR2);
-+
-+ if (locked)
-+ spin_unlock_irqrestore(&sport->port.lock, flags);
- }
-
- static void
-@@ -1662,6 +1675,13 @@ lpuart32_console_write(struct console *c
- {
- struct lpuart_port *sport = lpuart_ports[co->index];
- unsigned long old_cr, cr;
-+ unsigned long flags;
-+ int locked = 1;
-+
-+ if (sport->port.sysrq || oops_in_progress)
-+ locked = spin_trylock_irqsave(&sport->port.lock, flags);
-+ else
-+ spin_lock_irqsave(&sport->port.lock, flags);
-
- /* first save CR2 and then disable interrupts */
- cr = old_cr = lpuart32_read(sport->port.membase + UARTCTRL);
-@@ -1676,6 +1696,9 @@ lpuart32_console_write(struct console *c
- barrier();
-
- lpuart32_write(old_cr, sport->port.membase + UARTCTRL);
-+
-+ if (locked)
-+ spin_unlock_irqrestore(&sport->port.lock, flags);
- }
-
- /*
-@@ -1900,13 +1923,13 @@ static int lpuart_probe(struct platform_
-
- ret = of_alias_get_id(np, "serial");
- if (ret < 0) {
-- dev_err(&pdev->dev, "failed to get alias id, errno %d\n", ret);
-- return ret;
-- }
-- if (ret >= ARRAY_SIZE(lpuart_ports)) {
-- dev_err(&pdev->dev, "serial%d out of range\n", ret);
-- return -EINVAL;
-+ ret = find_first_zero_bit(linemap, UART_NR);
-+ if (ret >= UART_NR) {
-+ dev_err(&pdev->dev, "port line is full, add device failed\n");
-+ return ret;
-+ }
- }
-+ set_bit(ret, linemap);
- sport->port.line = ret;
- sport->lpuart32 = of_device_is_compatible(np, "fsl,ls1021a-lpuart");
-
-@@ -1988,6 +2011,7 @@ static int lpuart_remove(struct platform
- struct lpuart_port *sport = platform_get_drvdata(pdev);
-
- uart_remove_one_port(&lpuart_reg, &sport->port);
-+ clear_bit(sport->port.line, linemap);
-
- clk_disable_unprepare(sport->clk);
-
-@@ -2072,12 +2096,10 @@ static int lpuart_resume(struct device *
-
- if (sport->lpuart_dma_rx_use) {
- if (sport->port.irq_wake) {
-- if (!lpuart_start_rx_dma(sport)) {
-- sport->lpuart_dma_rx_use = true;
-+ if (!lpuart_start_rx_dma(sport))
- rx_dma_timer_init(sport);
-- } else {
-+ else
- sport->lpuart_dma_rx_use = false;
-- }
- }
- }
-
diff --git a/target/linux/layerscape/patches-4.9/817-usb-support-layerscape.patch b/target/linux/layerscape/patches-4.9/817-usb-support-layerscape.patch
deleted file mode 100644
index a6a1594a9a..0000000000
--- a/target/linux/layerscape/patches-4.9/817-usb-support-layerscape.patch
+++ /dev/null
@@ -1,1790 +0,0 @@
-From 1d35e363dd6e8bb1733bca0dfc186e3f70e692fe Mon Sep 17 00:00:00 2001
-From: Yangbo Lu <yangbo.lu@nxp.com>
-Date: Thu, 5 Jul 2018 17:38:52 +0800
-Subject: [PATCH 29/32] usb: support layerscape
-
-This is an integrated patch for layerscape usb support.
-
-Signed-off-by: yinbo.zhu <yinbo.zhu@nxp.com>
-Signed-off-by: Ramneek Mehresh <ramneek.mehresh@freescale.com>
-Signed-off-by: Nikhil Badola <nikhil.badola@freescale.com>
-Signed-off-by: Changming Huang <jerry.huang@nxp.com>
-Signed-off-by: Catalin Marinas <catalin.marinas@arm.com>
-Signed-off-by: Rajesh Bhagat <rajesh.bhagat@freescale.com>
-Signed-off-by: Suresh Gupta <suresh.gupta@freescale.com>
-Signed-off-by: Zhao Chenhui <chenhui.zhao@freescale.com>
-Signed-off-by: Yangbo Lu <yangbo.lu@nxp.com>
----
- drivers/net/usb/cdc_ether.c | 8 +
- drivers/net/usb/r8152.c | 6 +
- drivers/usb/common/common.c | 50 +++++
- drivers/usb/core/hub.c | 8 +
- drivers/usb/dwc3/core.c | 243 +++++++++++++++++++++-
- drivers/usb/dwc3/core.h | 51 ++++-
- drivers/usb/dwc3/ep0.c | 4 +-
- drivers/usb/dwc3/gadget.c | 7 +
- drivers/usb/dwc3/host.c | 24 ++-
- drivers/usb/gadget/udc/fsl_udc_core.c | 46 +++--
- drivers/usb/gadget/udc/fsl_usb2_udc.h | 16 +-
- drivers/usb/host/Kconfig | 4 +-
- drivers/usb/host/ehci-fsl.c | 279 ++++++++++++++++++++++++--
- drivers/usb/host/ehci-fsl.h | 3 +
- drivers/usb/host/ehci-hub.c | 4 +
- drivers/usb/host/ehci.h | 9 +
- drivers/usb/host/fsl-mph-dr-of.c | 16 +-
- drivers/usb/host/xhci-hub.c | 22 ++
- drivers/usb/host/xhci-plat.c | 16 +-
- drivers/usb/host/xhci-ring.c | 29 ++-
- drivers/usb/host/xhci.c | 38 +++-
- drivers/usb/host/xhci.h | 6 +-
- drivers/usb/phy/phy-fsl-usb.c | 59 ++++--
- drivers/usb/phy/phy-fsl-usb.h | 8 +
- include/linux/usb.h | 1 +
- include/linux/usb/of.h | 2 +
- 26 files changed, 867 insertions(+), 92 deletions(-)
-
---- a/drivers/net/usb/cdc_ether.c
-+++ b/drivers/net/usb/cdc_ether.c
-@@ -533,6 +533,7 @@ static const struct driver_info wwan_inf
- #define LINKSYS_VENDOR_ID 0x13b1
- #define NVIDIA_VENDOR_ID 0x0955
- #define HP_VENDOR_ID 0x03f0
-+#define TPLINK_VENDOR_ID 0x2357
-
- static const struct usb_device_id products[] = {
- /* BLACKLIST !!
-@@ -742,6 +743,13 @@ static const struct usb_device_id produc
- USB_CDC_SUBCLASS_ETHERNET, USB_CDC_PROTO_NONE),
- .driver_info = 0,
- },
-+
-+ /* TP-LINK UE300 USB 3.0 Ethernet Adapters (based on Realtek RTL8153) */
-+{
-+ USB_DEVICE_AND_INTERFACE_INFO(TPLINK_VENDOR_ID, 0x0601, USB_CLASS_COMM,
-+ USB_CDC_SUBCLASS_ETHERNET, USB_CDC_PROTO_NONE),
-+ .driver_info = 0,
-+},
-
- /* WHITELIST!!!
- *
---- a/drivers/net/usb/r8152.c
-+++ b/drivers/net/usb/r8152.c
-@@ -521,6 +521,7 @@ enum rtl8152_flags {
- #define VENDOR_ID_LENOVO 0x17ef
- #define VENDOR_ID_LINKSYS 0x13b1
- #define VENDOR_ID_NVIDIA 0x0955
-+#define VENDOR_ID_TPLINK 0x2357
-
- #define MCU_TYPE_PLA 0x0100
- #define MCU_TYPE_USB 0x0000
-@@ -1817,6 +1818,10 @@ static int rx_bottom(struct r8152 *tp, i
- unsigned int pkt_len;
- struct sk_buff *skb;
-
-+ /* limite the skb numbers for rx_queue */
-+ if (unlikely(skb_queue_len(&tp->rx_queue) >= 1000))
-+ break;
-+
- pkt_len = le32_to_cpu(rx_desc->opts1) & RX_LEN_MASK;
- if (pkt_len < ETH_ZLEN)
- break;
-@@ -4513,6 +4518,7 @@ static struct usb_device_id rtl8152_tabl
- {REALTEK_USB_DEVICE(VENDOR_ID_LENOVO, 0x304f)},
- {REALTEK_USB_DEVICE(VENDOR_ID_LINKSYS, 0x0041)},
- {REALTEK_USB_DEVICE(VENDOR_ID_NVIDIA, 0x09ff)},
-+ {REALTEK_USB_DEVICE(VENDOR_ID_TPLINK, 0x0601)},
- {}
- };
-
---- a/drivers/usb/common/common.c
-+++ b/drivers/usb/common/common.c
-@@ -105,6 +105,56 @@ static const char *const usb_dr_modes[]
- [USB_DR_MODE_OTG] = "otg",
- };
-
-+/**
-+ * of_usb_get_dr_mode - Get dual role mode for given device_node
-+ * @np: Pointer to the given device_node
-+ *
-+ * The function gets phy interface string from property 'dr_mode',
-+ * and returns the correspondig enum usb_dr_mode
-+ */
-+enum usb_dr_mode of_usb_get_dr_mode(struct device_node *np)
-+{
-+ const char *dr_mode;
-+ int err, i;
-+
-+ err = of_property_read_string(np, "dr_mode", &dr_mode);
-+ if (err < 0)
-+ return USB_DR_MODE_UNKNOWN;
-+
-+ for (i = 0; i < ARRAY_SIZE(usb_dr_modes); i++)
-+ if (!strcmp(dr_mode, usb_dr_modes[i]))
-+ return i;
-+
-+ return USB_DR_MODE_UNKNOWN;
-+}
-+EXPORT_SYMBOL_GPL(of_usb_get_dr_mode);
-+
-+/**
-+ * of_usb_get_maximum_speed - Get maximum requested speed for a given USB
-+ * controller.
-+ * @np: Pointer to the given device_node
-+ *
-+ * The function gets the maximum speed string from property "maximum-speed",
-+ * and returns the corresponding enum usb_device_speed.
-+ */
-+enum usb_device_speed of_usb_get_maximum_speed(struct device_node *np)
-+{
-+ const char *maximum_speed;
-+ int err;
-+ int i;
-+
-+ err = of_property_read_string(np, "maximum-speed", &maximum_speed);
-+ if (err < 0)
-+ return USB_SPEED_UNKNOWN;
-+
-+ for (i = 0; i < ARRAY_SIZE(speed_names); i++)
-+ if (strcmp(maximum_speed, speed_names[i]) == 0)
-+ return i;
-+
-+ return USB_SPEED_UNKNOWN;
-+}
-+EXPORT_SYMBOL_GPL(of_usb_get_maximum_speed);
-+
- static enum usb_dr_mode usb_get_dr_mode_from_string(const char *str)
- {
- int ret;
---- a/drivers/usb/core/hub.c
-+++ b/drivers/usb/core/hub.c
-@@ -4433,6 +4433,14 @@ hub_port_init(struct usb_hub *hub, struc
- else
- speed = usb_speed_string(udev->speed);
-
-+#if !defined(CONFIG_FSL_USB2_OTG) && !defined(CONFIG_FSL_USB2_OTG_MODULE)
-+if (udev->speed != USB_SPEED_SUPER)
-+ dev_info(&udev->dev,
-+ "%s %s USB device number %d using %s\n",
-+ (udev->config) ? "reset" : "new", speed,
-+ devnum, udev->bus->controller->driver->name);
-+#endif
-+
- if (udev->speed < USB_SPEED_SUPER)
- dev_info(&udev->dev,
- "%s %s USB device number %d using %s\n",
---- a/drivers/usb/dwc3/core.c
-+++ b/drivers/usb/dwc3/core.c
-@@ -58,6 +58,7 @@ static int dwc3_get_dr_mode(struct dwc3
- enum usb_dr_mode mode;
- struct device *dev = dwc->dev;
- unsigned int hw_mode;
-+ struct device_node *node = dev->of_node;
-
- if (dwc->dr_mode == USB_DR_MODE_UNKNOWN)
- dwc->dr_mode = USB_DR_MODE_OTG;
-@@ -83,6 +84,24 @@ static int dwc3_get_dr_mode(struct dwc3
- mode = USB_DR_MODE_HOST;
- break;
- default:
-+ /* Adjust Frame Length */
-+ if (dwc->configure_gfladj)
-+ dwc3_writel(dwc->regs, DWC3_GFLADJ, GFLADJ_30MHZ_REG_SEL |
-+ GFLADJ_30MHZ(GFLADJ_30MHZ_DEFAULT));
-+
-+ /* Change burst beat and outstanding pipelined transfers requests */
-+ dwc3_writel(dwc->regs, DWC3_GSBUSCFG0,
-+ (dwc3_readl(dwc->regs, DWC3_GSBUSCFG0) & ~0xff) | 0xf);
-+ dwc3_writel(dwc->regs, DWC3_GSBUSCFG1,
-+ dwc3_readl(dwc->regs, DWC3_GSBUSCFG1) | 0xf00);
-+
-+ /* Enable Snooping */
-+ if (node && of_dma_is_coherent(node)) {
-+ dwc3_writel(dwc->regs, DWC3_GSBUSCFG0,
-+ dwc3_readl(dwc->regs, DWC3_GSBUSCFG0) | 0x22220000);
-+ dev_dbg(dev, "enabled snooping for usb\n");
-+ }
-+
- if (IS_ENABLED(CONFIG_USB_DWC3_HOST))
- mode = USB_DR_MODE_HOST;
- else if (IS_ENABLED(CONFIG_USB_DWC3_GADGET))
-@@ -227,8 +246,9 @@ static void dwc3_frame_length_adjustment
-
- reg = dwc3_readl(dwc->regs, DWC3_GFLADJ);
- dft = reg & DWC3_GFLADJ_30MHZ_MASK;
-- if (!dev_WARN_ONCE(dwc->dev, dft == dwc->fladj,
-- "request value same as default, ignoring\n")) {
-+ if (dft == dwc->fladj) {
-+ dev_warn(dwc->dev, "request value same as default, ignoring\n");
-+ } else {
- reg &= ~DWC3_GFLADJ_30MHZ_MASK;
- reg |= DWC3_GFLADJ_30MHZ_SDBND_SEL | dwc->fladj;
- dwc3_writel(dwc->regs, DWC3_GFLADJ, reg);
-@@ -599,6 +619,99 @@ static int dwc3_phy_setup(struct dwc3 *d
- return 0;
- }
-
-+/* set global soc bus configuration registers */
-+static void dwc3_set_soc_bus_cfg(struct dwc3 *dwc)
-+{
-+ struct device *dev = dwc->dev;
-+ u32 *vals;
-+ u32 cfg;
-+ int ntype;
-+ int ret;
-+ int i;
-+
-+ cfg = dwc3_readl(dwc->regs, DWC3_GSBUSCFG0);
-+
-+ /*
-+ * Handle property "snps,incr-burst-type-adjustment".
-+ * Get the number of value from this property:
-+ * result <= 0, means this property is not supported.
-+ * result = 1, means INCRx burst mode supported.
-+ * result > 1, means undefined length burst mode supported.
-+ */
-+ ntype = device_property_read_u32_array(dev,
-+ "snps,incr-burst-type-adjustment", NULL, 0);
-+ if (ntype > 0) {
-+ vals = kcalloc(ntype, sizeof(u32), GFP_KERNEL);
-+ if (!vals) {
-+ dev_err(dev, "Error to get memory\n");
-+ return;
-+ }
-+ /* Get INCR burst type, and parse it */
-+ ret = device_property_read_u32_array(dev,
-+ "snps,incr-burst-type-adjustment", vals, ntype);
-+ if (ret) {
-+ dev_err(dev, "Error to get property\n");
-+ return;
-+ }
-+ *(dwc->incrx_type + 1) = vals[0];
-+ if (ntype > 1) {
-+ *dwc->incrx_type = 1;
-+ for (i = 1; i < ntype; i++) {
-+ if (vals[i] > *(dwc->incrx_type + 1))
-+ *(dwc->incrx_type + 1) = vals[i];
-+ }
-+ } else
-+ *dwc->incrx_type = 0;
-+
-+ /* Enable Undefined Length INCR Burst and Enable INCRx Burst */
-+ cfg &= ~DWC3_GSBUSCFG0_INCRBRST_MASK;
-+ if (*dwc->incrx_type)
-+ cfg |= DWC3_GSBUSCFG0_INCRBRSTENA;
-+ switch (*(dwc->incrx_type + 1)) {
-+ case 256:
-+ cfg |= DWC3_GSBUSCFG0_INCR256BRSTENA;
-+ break;
-+ case 128:
-+ cfg |= DWC3_GSBUSCFG0_INCR128BRSTENA;
-+ break;
-+ case 64:
-+ cfg |= DWC3_GSBUSCFG0_INCR64BRSTENA;
-+ break;
-+ case 32:
-+ cfg |= DWC3_GSBUSCFG0_INCR32BRSTENA;
-+ break;
-+ case 16:
-+ cfg |= DWC3_GSBUSCFG0_INCR16BRSTENA;
-+ break;
-+ case 8:
-+ cfg |= DWC3_GSBUSCFG0_INCR8BRSTENA;
-+ break;
-+ case 4:
-+ cfg |= DWC3_GSBUSCFG0_INCR4BRSTENA;
-+ break;
-+ case 1:
-+ break;
-+ default:
-+ dev_err(dev, "Invalid property\n");
-+ break;
-+ }
-+ }
-+
-+ /* Handle usb snooping */
-+ if (dwc->dma_snooping_quirk) {
-+ cfg &= ~DWC3_GSBUSCFG0_SNP_MASK;
-+ cfg |= (AXI3_CACHE_TYPE_SNP << DWC3_GSBUSCFG0_DATARD_SHIFT) |
-+ (AXI3_CACHE_TYPE_SNP << DWC3_GSBUSCFG0_DESCRD_SHIFT) |
-+ (AXI3_CACHE_TYPE_SNP << DWC3_GSBUSCFG0_DATAWR_SHIFT) |
-+ (AXI3_CACHE_TYPE_SNP << DWC3_GSBUSCFG0_DESCWR_SHIFT);
-+ }
-+
-+ dwc3_writel(dwc->regs, DWC3_GSBUSCFG0, cfg);
-+
-+}
-+
-+
-+
- static void dwc3_core_exit(struct dwc3 *dwc)
- {
- dwc3_event_buffers_cleanup(dwc);
-@@ -741,6 +854,8 @@ static int dwc3_core_init(struct dwc3 *d
- if (ret)
- goto err1;
-
-+ dwc3_set_soc_bus_cfg(dwc);
-+
- /* Adjust Frame Length */
- dwc3_frame_length_adjustment(dwc);
-
-@@ -939,11 +1054,117 @@ static void dwc3_core_exit_mode(struct d
- }
- }
-
-+static void dwc3_get_properties(struct dwc3 *dwc)
-+{
-+ struct device *dev = dwc->dev;
-+ struct device_node *node = dev->of_node;
-+ u8 lpm_nyet_threshold;
-+ u8 tx_de_emphasis;
-+ u8 hird_threshold;
-+
-+ /* default to highest possible threshold */
-+ lpm_nyet_threshold = 0xff;
-+
-+ /* default to -3.5dB de-emphasis */
-+ tx_de_emphasis = 1;
-+
-+ /*
-+ * default to assert utmi_sleep_n and use maximum allowed HIRD
-+ * threshold value of 0b1100
-+ */
-+ hird_threshold = 12;
-+
-+ dwc->maximum_speed = usb_get_maximum_speed(dev);
-+ dwc->dr_mode = usb_get_dr_mode(dev);
-+ dwc->hsphy_mode = of_usb_get_phy_mode(dev->of_node);
-+
-+ dwc->sysdev_is_parent = device_property_read_bool(dev,
-+ "linux,sysdev_is_parent");
-+ if (dwc->sysdev_is_parent)
-+ dwc->sysdev = dwc->dev->parent;
-+ else
-+ dwc->sysdev = dwc->dev;
-+
-+ dwc->has_lpm_erratum = device_property_read_bool(dev,
-+ "snps,has-lpm-erratum");
-+ device_property_read_u8(dev, "snps,lpm-nyet-threshold",
-+ &lpm_nyet_threshold);
-+ dwc->is_utmi_l1_suspend = device_property_read_bool(dev,
-+ "snps,is-utmi-l1-suspend");
-+ device_property_read_u8(dev, "snps,hird-threshold",
-+ &hird_threshold);
-+ dwc->usb3_lpm_capable = device_property_read_bool(dev,
-+ "snps,usb3_lpm_capable");
-+ dwc->quirk_reverse_in_out = device_property_read_bool(dev,
-+ "snps,quirk_reverse_in_out");
-+ dwc->quirk_stop_transfer_in_block = device_property_read_bool(dev,
-+ "snps,quirk_stop_transfer_in_block");
-+ dwc->quirk_stop_ep_in_u1 = device_property_read_bool(dev,
-+ "snps,quirk_stop_ep_in_u1");
-+
-+ dwc->needs_fifo_resize = of_property_read_bool(node, "tx-fifo-resize");
-+
-+ dwc->configure_gfladj =
-+ of_property_read_bool(node, "configure-gfladj");
-+ dwc->dr_mode = usb_get_dr_mode(dev);
-+
-+ dwc->disable_scramble_quirk = device_property_read_bool(dev,
-+ "snps,disable_scramble_quirk");
-+ dwc->u2exit_lfps_quirk = device_property_read_bool(dev,
-+ "snps,u2exit_lfps_quirk");
-+ dwc->u2ss_inp3_quirk = device_property_read_bool(dev,
-+ "snps,u2ss_inp3_quirk");
-+ dwc->req_p1p2p3_quirk = device_property_read_bool(dev,
-+ "snps,req_p1p2p3_quirk");
-+ dwc->del_p1p2p3_quirk = device_property_read_bool(dev,
-+ "snps,del_p1p2p3_quirk");
-+ dwc->del_phy_power_chg_quirk = device_property_read_bool(dev,
-+ "snps,del_phy_power_chg_quirk");
-+ dwc->lfps_filter_quirk = device_property_read_bool(dev,
-+ "snps,lfps_filter_quirk");
-+ dwc->rx_detect_poll_quirk = device_property_read_bool(dev,
-+ "snps,rx_detect_poll_quirk");
-+ dwc->dis_u3_susphy_quirk = device_property_read_bool(dev,
-+ "snps,dis_u3_susphy_quirk");
-+ dwc->dis_u2_susphy_quirk = device_property_read_bool(dev,
-+ "snps,dis_u2_susphy_quirk");
-+ dwc->dis_enblslpm_quirk = device_property_read_bool(dev,
-+ "snps,dis_enblslpm_quirk");
-+ dwc->dis_rxdet_inp3_quirk = device_property_read_bool(dev,
-+ "snps,dis_rxdet_inp3_quirk");
-+ dwc->dis_u2_freeclk_exists_quirk = device_property_read_bool(dev,
-+ "snps,dis-u2-freeclk-exists-quirk");
-+ dwc->dis_del_phy_power_chg_quirk = device_property_read_bool(dev,
-+ "snps,dis-del-phy-power-chg-quirk");
-+ dwc->dma_snooping_quirk = device_property_read_bool(dev,
-+ "snps,dma-snooping");
-+
-+ dwc->tx_de_emphasis_quirk = device_property_read_bool(dev,
-+ "snps,tx_de_emphasis_quirk");
-+ dwc->disable_devinit_u1u2_quirk = device_property_read_bool(dev,
-+ "snps,disable_devinit_u1u2");
-+ device_property_read_u8(dev, "snps,tx_de_emphasis",
-+ &tx_de_emphasis);
-+ device_property_read_string(dev, "snps,hsphy_interface",
-+ &dwc->hsphy_interface);
-+ device_property_read_u32(dev, "snps,quirk-frame-length-adjustment",
-+ &dwc->fladj);
-+
-+ dwc->lpm_nyet_threshold = lpm_nyet_threshold;
-+ dwc->tx_de_emphasis = tx_de_emphasis;
-+
-+ dwc->hird_threshold = hird_threshold
-+ | (dwc->is_utmi_l1_suspend << 4);
-+
-+ dwc->imod_interval = 0;
-+}
-+
- #define DWC3_ALIGN_MASK (16 - 1)
-
- static int dwc3_probe(struct platform_device *pdev)
- {
- struct device *dev = &pdev->dev;
-+ struct device_node *node = dev->of_node;
- struct resource *res;
- struct dwc3 *dwc;
- u8 lpm_nyet_threshold;
-@@ -975,6 +1196,11 @@ static int dwc3_probe(struct platform_de
- dwc->xhci_resources[0].flags = res->flags;
- dwc->xhci_resources[0].name = res->name;
-
-+ if (node) {
-+ dwc->configure_gfladj =
-+ of_property_read_bool(node, "configure-gfladj");
-+ }
-+
- res->start += DWC3_GLOBALS_REGS_START;
-
- /*
-@@ -1017,6 +1243,12 @@ static int dwc3_probe(struct platform_de
- dwc->usb3_lpm_capable = device_property_read_bool(dev,
- "snps,usb3_lpm_capable");
-
-+ dwc->needs_fifo_resize = of_property_read_bool(node, "tx-fifo-resize");
-+
-+ dwc->configure_gfladj =
-+ of_property_read_bool(node, "configure-gfladj");
-+ dwc->dr_mode = of_usb_get_dr_mode(node);
-+
- dwc->disable_scramble_quirk = device_property_read_bool(dev,
- "snps,disable_scramble_quirk");
- dwc->u2exit_lfps_quirk = device_property_read_bool(dev,
-@@ -1061,6 +1293,8 @@ static int dwc3_probe(struct platform_de
- dwc->hird_threshold = hird_threshold
- | (dwc->is_utmi_l1_suspend << 4);
-
-+ dwc3_get_properties(dwc);
-+
- platform_set_drvdata(pdev, dwc);
- dwc3_cache_hwparams(dwc);
-
-@@ -1084,6 +1318,11 @@ static int dwc3_probe(struct platform_de
- if (ret < 0)
- goto err1;
-
-+ /* Adjust Frame Length */
-+ if (dwc->configure_gfladj)
-+ dwc3_writel(dwc->regs, DWC3_GFLADJ, GFLADJ_30MHZ_REG_SEL |
-+ GFLADJ_30MHZ(GFLADJ_30MHZ_DEFAULT));
-+
- pm_runtime_forbid(dev);
-
- ret = dwc3_alloc_event_buffers(dwc, DWC3_EVENT_BUFFERS_SIZE);
---- a/drivers/usb/dwc3/core.h
-+++ b/drivers/usb/dwc3/core.h
-@@ -26,6 +26,7 @@
- #include <linux/dma-mapping.h>
- #include <linux/mm.h>
- #include <linux/debugfs.h>
-+#include <linux/of_address.h>
-
- #include <linux/usb/ch9.h>
- #include <linux/usb/gadget.h>
-@@ -154,6 +155,32 @@
-
- /* Bit fields */
-
-+/* Global SoC Bus Configuration Register 0 */
-+#define AXI3_CACHE_TYPE_AW 0x8 /* write allocate */
-+#define AXI3_CACHE_TYPE_AR 0x4 /* read allocate */
-+#define AXI3_CACHE_TYPE_SNP 0x2 /* cacheable */
-+#define AXI3_CACHE_TYPE_BUF 0x1 /* bufferable */
-+#define DWC3_GSBUSCFG0_DATARD_SHIFT 28
-+#define DWC3_GSBUSCFG0_DESCRD_SHIFT 24
-+#define DWC3_GSBUSCFG0_DATAWR_SHIFT 20
-+#define DWC3_GSBUSCFG0_DESCWR_SHIFT 16
-+#define DWC3_GSBUSCFG0_SNP_MASK 0xffff0000
-+#define DWC3_GSBUSCFG0_DATABIGEND (1 << 11)
-+#define DWC3_GSBUSCFG0_DESCBIGEND (1 << 10)
-+#define DWC3_GSBUSCFG0_INCR256BRSTENA (1 << 7) /* INCR256 burst */
-+#define DWC3_GSBUSCFG0_INCR128BRSTENA (1 << 6) /* INCR128 burst */
-+#define DWC3_GSBUSCFG0_INCR64BRSTENA (1 << 5) /* INCR64 burst */
-+#define DWC3_GSBUSCFG0_INCR32BRSTENA (1 << 4) /* INCR32 burst */
-+#define DWC3_GSBUSCFG0_INCR16BRSTENA (1 << 3) /* INCR16 burst */
-+#define DWC3_GSBUSCFG0_INCR8BRSTENA (1 << 2) /* INCR8 burst */
-+#define DWC3_GSBUSCFG0_INCR4BRSTENA (1 << 1) /* INCR4 burst */
-+#define DWC3_GSBUSCFG0_INCRBRSTENA (1 << 0) /* undefined length enable */
-+#define DWC3_GSBUSCFG0_INCRBRST_MASK 0xff
-+
-+/* Global SoC Bus Configuration Register 1 */
-+#define DWC3_GSBUSCFG1_1KPAGEENA (1 << 12) /* 1K page boundary enable */
-+#define DWC3_GSBUSCFG1_PTRANSLIMIT_MASK 0xf00
-+
- /* Global Debug Queue/FIFO Space Available Register */
- #define DWC3_GDBGFIFOSPACE_NUM(n) ((n) & 0x1f)
- #define DWC3_GDBGFIFOSPACE_TYPE(n) (((n) << 5) & 0x1e0)
-@@ -182,7 +209,6 @@
- #define DWC3_GCTL_CLK_PIPE (1)
- #define DWC3_GCTL_CLK_PIPEHALF (2)
- #define DWC3_GCTL_CLK_MASK (3)
--
- #define DWC3_GCTL_PRTCAP(n) (((n) & (3 << 12)) >> 12)
- #define DWC3_GCTL_PRTCAPDIR(n) ((n) << 12)
- #define DWC3_GCTL_PRTCAP_HOST 1
-@@ -294,6 +320,10 @@
- /* Global Frame Length Adjustment Register */
- #define DWC3_GFLADJ_30MHZ_SDBND_SEL (1 << 7)
- #define DWC3_GFLADJ_30MHZ_MASK 0x3f
-+#define GFLADJ_30MHZ_REG_SEL (1 << 7)
-+#define GFLADJ_30MHZ(n) ((n) & 0x3f)
-+#define GFLADJ_30MHZ_DEFAULT 0x20
-+
-
- /* Global User Control Register 2 */
- #define DWC3_GUCTL2_RST_ACTBITLATER (1 << 14)
-@@ -758,6 +788,7 @@ struct dwc3_scratchpad_array {
- * @regs: base address for our registers
- * @regs_size: address space size
- * @fladj: frame length adjustment
-+ * @incrx_type: INCR burst type adjustment
- * @irq_gadget: peripheral controller's IRQ number
- * @nr_scratch: number of scratch buffers
- * @u1u2: only used on revisions <1.83a for workaround
-@@ -834,6 +865,7 @@ struct dwc3_scratchpad_array {
- * 1 - -3.5dB de-emphasis
- * 2 - No de-emphasis
- * 3 - Reserved
-+ * @disable_devinit_u1u2_quirk: disable device-initiated U1/U2 request.
- */
- struct dwc3 {
- struct usb_ctrlrequest *ctrl_req;
-@@ -852,6 +884,7 @@ struct dwc3 {
- spinlock_t lock;
-
- struct device *dev;
-+ struct device *sysdev;
-
- struct platform_device *xhci;
- struct resource xhci_resources[DWC3_XHCI_RESOURCES_NUM];
-@@ -877,6 +910,12 @@ struct dwc3 {
- enum usb_phy_interface hsphy_mode;
-
- u32 fladj;
-+ /*
-+ * For INCR burst type.
-+ * First field: for undefined length INCR burst type enable.
-+ * Second field: for INCRx burst type enable
-+ */
-+ u32 incrx_type[2];
- u32 irq_gadget;
- u32 nr_scratch;
- u32 u1u2;
-@@ -953,9 +992,12 @@ struct dwc3 {
- unsigned ep0_bounced:1;
- unsigned ep0_expect_in:1;
- unsigned has_hibernation:1;
-+ unsigned sysdev_is_parent:1;
- unsigned has_lpm_erratum:1;
- unsigned is_utmi_l1_suspend:1;
- unsigned is_fpga:1;
-+ unsigned needs_fifo_resize:1;
-+ unsigned configure_gfladj:1;
- unsigned pending_events:1;
- unsigned pullups_connected:1;
- unsigned setup_packet_pending:1;
-@@ -976,9 +1018,16 @@ struct dwc3 {
- unsigned dis_rxdet_inp3_quirk:1;
- unsigned dis_u2_freeclk_exists_quirk:1;
- unsigned dis_del_phy_power_chg_quirk:1;
-+ unsigned dma_snooping_quirk:1;
-
- unsigned tx_de_emphasis_quirk:1;
- unsigned tx_de_emphasis:2;
-+ unsigned disable_devinit_u1u2_quirk:1;
-+ unsigned quirk_reverse_in_out:1;
-+ unsigned quirk_stop_transfer_in_block:1;
-+ unsigned quirk_stop_ep_in_u1:1;
-+
-+ u16 imod_interval;
- };
-
- /* -------------------------------------------------------------------------- */
---- a/drivers/usb/dwc3/ep0.c
-+++ b/drivers/usb/dwc3/ep0.c
-@@ -360,9 +360,9 @@ static int dwc3_ep0_handle_status(struct
- if ((dwc->speed == DWC3_DSTS_SUPERSPEED) ||
- (dwc->speed == DWC3_DSTS_SUPERSPEED_PLUS)) {
- reg = dwc3_readl(dwc->regs, DWC3_DCTL);
-- if (reg & DWC3_DCTL_INITU1ENA)
-+ if ((reg & DWC3_DCTL_INITU1ENA) && !dwc->disable_devinit_u1u2_quirk)
- usb_status |= 1 << USB_DEV_STAT_U1_ENABLED;
-- if (reg & DWC3_DCTL_INITU2ENA)
-+ if ((reg & DWC3_DCTL_INITU2ENA) && !dwc->disable_devinit_u1u2_quirk)
- usb_status |= 1 << USB_DEV_STAT_U2_ENABLED;
- }
-
---- a/drivers/usb/dwc3/gadget.c
-+++ b/drivers/usb/dwc3/gadget.c
-@@ -2927,6 +2927,7 @@ static irqreturn_t dwc3_interrupt(int ir
- int dwc3_gadget_init(struct dwc3 *dwc)
- {
- int ret, irq;
-+ u32 reg;
- struct platform_device *dwc3_pdev = to_platform_device(dwc->dev);
-
- irq = platform_get_irq_byname(dwc3_pdev, "peripheral");
-@@ -3041,6 +3042,12 @@ int dwc3_gadget_init(struct dwc3 *dwc)
- goto err5;
- }
-
-+ if (dwc->disable_devinit_u1u2_quirk) {
-+ reg = dwc3_readl(dwc->regs, DWC3_DCTL);
-+ reg &= ~(DWC3_DCTL_INITU1ENA | DWC3_DCTL_INITU2ENA);
-+ dwc3_writel(dwc->regs, DWC3_DCTL, reg);
-+ }
-+
- return 0;
-
- err5:
---- a/drivers/usb/dwc3/host.c
-+++ b/drivers/usb/dwc3/host.c
-@@ -17,6 +17,8 @@
-
- #include <linux/platform_device.h>
-
-+#include <linux/of_device.h>
-+
- #include "core.h"
-
- int dwc3_host_init(struct dwc3 *dwc)
-@@ -73,12 +75,21 @@ int dwc3_host_init(struct dwc3 *dwc)
- return -ENOMEM;
- }
-
-- dma_set_coherent_mask(&xhci->dev, dwc->dev->coherent_dma_mask);
-+ if (IS_ENABLED(CONFIG_OF) && dwc->dev->of_node)
-+ of_dma_configure(&xhci->dev, dwc->dev->of_node);
-+ else
-+ dma_set_coherent_mask(&xhci->dev, dwc->dev->coherent_dma_mask);
-
-- xhci->dev.parent = dwc->dev;
-+ xhci->dev.parent = dwc->dev;
- xhci->dev.dma_mask = dwc->dev->dma_mask;
- xhci->dev.dma_parms = dwc->dev->dma_parms;
-
-+ /* set DMA operations */
-+ if (dwc->dev->of_node && of_dma_is_coherent(dwc->dev->of_node)) {
-+ xhci->dev.archdata.dma_ops = dwc->dev->archdata.dma_ops;
-+ dev_dbg(dwc->dev, "set dma_ops for usb\n");
-+ }
-+
- dwc->xhci = xhci;
-
- ret = platform_device_add_resources(xhci, dwc->xhci_resources,
-@@ -90,6 +101,15 @@ int dwc3_host_init(struct dwc3 *dwc)
-
- memset(props, 0, sizeof(struct property_entry) * ARRAY_SIZE(props));
-
-+ if (dwc->quirk_reverse_in_out)
-+ props[prop_idx++].name = "quirk-reverse-in-out";
-+
-+ if (dwc->quirk_stop_transfer_in_block)
-+ props[prop_idx++].name = "quirk-stop-transfer-in-block";
-+
-+ if (dwc->quirk_stop_ep_in_u1)
-+ props[prop_idx++].name = "quirk-stop-ep-in-u1";
-+
- if (dwc->usb3_lpm_capable)
- props[prop_idx++].name = "usb3-lpm-capable";
-
---- a/drivers/usb/gadget/udc/fsl_udc_core.c
-+++ b/drivers/usb/gadget/udc/fsl_udc_core.c
-@@ -198,7 +198,11 @@ __acquires(ep->udc->lock)
-
- spin_unlock(&ep->udc->lock);
-
-- usb_gadget_giveback_request(&ep->ep, &req->req);
-+ /* this complete() should a func implemented by gadget layer,
-+ * eg fsg->bulk_in_complete()
-+ */
-+ if (req->req.complete)
-+ usb_gadget_giveback_request(&ep->ep, &req->req);
-
- spin_lock(&ep->udc->lock);
- ep->stopped = stopped;
-@@ -245,10 +249,10 @@ static int dr_controller_setup(struct fs
- if (udc->pdata->have_sysif_regs) {
- if (udc->pdata->controller_ver) {
- /* controller version 1.6 or above */
-- ctrl = __raw_readl(&usb_sys_regs->control);
-+ ctrl = ioread32be(&usb_sys_regs->control);
- ctrl &= ~USB_CTRL_UTMI_PHY_EN;
- ctrl |= USB_CTRL_USB_EN;
-- __raw_writel(ctrl, &usb_sys_regs->control);
-+ iowrite32be(ctrl, &usb_sys_regs->control);
- }
- }
- portctrl |= PORTSCX_PTS_ULPI;
-@@ -257,13 +261,14 @@ static int dr_controller_setup(struct fs
- portctrl |= PORTSCX_PTW_16BIT;
- /* fall through */
- case FSL_USB2_PHY_UTMI:
-+ case FSL_USB2_PHY_UTMI_DUAL:
- if (udc->pdata->have_sysif_regs) {
- if (udc->pdata->controller_ver) {
- /* controller version 1.6 or above */
-- ctrl = __raw_readl(&usb_sys_regs->control);
-+ ctrl = ioread32be(&usb_sys_regs->control);
- ctrl |= (USB_CTRL_UTMI_PHY_EN |
- USB_CTRL_USB_EN);
-- __raw_writel(ctrl, &usb_sys_regs->control);
-+ iowrite32be(ctrl, &usb_sys_regs->control);
- mdelay(FSL_UTMI_PHY_DLY); /* Delay for UTMI
- PHY CLK to become stable - 10ms*/
- }
-@@ -329,22 +334,22 @@ static int dr_controller_setup(struct fs
- /* Config control enable i/o output, cpu endian register */
- #ifndef CONFIG_ARCH_MXC
- if (udc->pdata->have_sysif_regs) {
-- ctrl = __raw_readl(&usb_sys_regs->control);
-+ ctrl = ioread32be(&usb_sys_regs->control);
- ctrl |= USB_CTRL_IOENB;
-- __raw_writel(ctrl, &usb_sys_regs->control);
-+ iowrite32be(ctrl, &usb_sys_regs->control);
- }
- #endif
-
--#if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
-+#if !defined(CONFIG_NOT_COHERENT_CACHE)
- /* Turn on cache snooping hardware, since some PowerPC platforms
- * wholly rely on hardware to deal with cache coherent. */
-
- if (udc->pdata->have_sysif_regs) {
- /* Setup Snooping for all the 4GB space */
- tmp = SNOOP_SIZE_2GB; /* starts from 0x0, size 2G */
-- __raw_writel(tmp, &usb_sys_regs->snoop1);
-+ iowrite32be(tmp, &usb_sys_regs->snoop1);
- tmp |= 0x80000000; /* starts from 0x8000000, size 2G */
-- __raw_writel(tmp, &usb_sys_regs->snoop2);
-+ iowrite32be(tmp, &usb_sys_regs->snoop2);
- }
- #endif
-
-@@ -1057,7 +1062,7 @@ static int fsl_ep_fifo_status(struct usb
- struct ep_queue_head *qh;
-
- ep = container_of(_ep, struct fsl_ep, ep);
-- if (!_ep || (!ep->ep.desc && ep_index(ep) != 0))
-+ if (!_ep || !ep->ep.desc || (ep_index(ep) == 0))
- return -ENODEV;
-
- udc = (struct fsl_udc *)ep->udc;
-@@ -1599,14 +1604,13 @@ static int process_ep_req(struct fsl_udc
- struct fsl_req *curr_req)
- {
- struct ep_td_struct *curr_td;
-- int td_complete, actual, remaining_length, j, tmp;
-+ int actual, remaining_length, j, tmp;
- int status = 0;
- int errors = 0;
- struct ep_queue_head *curr_qh = &udc->ep_qh[pipe];
- int direction = pipe % 2;
-
- curr_td = curr_req->head;
-- td_complete = 0;
- actual = curr_req->req.length;
-
- for (j = 0; j < curr_req->dtd_count; j++) {
-@@ -1651,11 +1655,9 @@ static int process_ep_req(struct fsl_udc
- status = -EPROTO;
- break;
- } else {
-- td_complete++;
- break;
- }
- } else {
-- td_complete++;
- VDBG("dTD transmitted successful");
- }
-
-@@ -1698,7 +1700,7 @@ static void dtd_complete_irq(struct fsl_
- curr_ep = get_ep_by_pipe(udc, i);
-
- /* If the ep is configured */
-- if (!curr_ep->ep.name) {
-+ if (strncmp(curr_ep->name, "ep", 2)) {
- WARNING("Invalid EP?");
- continue;
- }
-@@ -2420,10 +2422,12 @@ static int fsl_udc_probe(struct platform
- usb_sys_regs = (void *)dr_regs + USB_DR_SYS_OFFSET;
- #endif
-
-+#ifdef CONFIG_ARCH_MXC
- /* Initialize USB clocks */
- ret = fsl_udc_clk_init(pdev);
- if (ret < 0)
- goto err_iounmap_noclk;
-+#endif
-
- /* Read Device Controller Capability Parameters register */
- dccparams = fsl_readl(&dr_regs->dccparams);
-@@ -2463,9 +2467,11 @@ static int fsl_udc_probe(struct platform
- dr_controller_setup(udc_controller);
- }
-
-+#ifdef CONFIG_ARCH_MXC
- ret = fsl_udc_clk_finalize(pdev);
- if (ret)
- goto err_free_irq;
-+#endif
-
- /* Setup gadget structure */
- udc_controller->gadget.ops = &fsl_gadget_ops;
-@@ -2478,6 +2484,7 @@ static int fsl_udc_probe(struct platform
- /* Setup gadget.dev and register with kernel */
- dev_set_name(&udc_controller->gadget.dev, "gadget");
- udc_controller->gadget.dev.of_node = pdev->dev.of_node;
-+ set_dma_ops(&udc_controller->gadget.dev, pdev->dev.archdata.dma_ops);
-
- if (!IS_ERR_OR_NULL(udc_controller->transceiver))
- udc_controller->gadget.is_otg = 1;
-@@ -2529,7 +2536,9 @@ err_free_irq:
- err_iounmap:
- if (pdata->exit)
- pdata->exit(pdev);
-+#ifdef CONFIG_ARCH_MXC
- fsl_udc_clk_release();
-+#endif
- err_iounmap_noclk:
- iounmap(dr_regs);
- err_release_mem_region:
-@@ -2557,8 +2566,9 @@ static int fsl_udc_remove(struct platfor
- udc_controller->done = &done;
- usb_del_gadget_udc(&udc_controller->gadget);
-
-+#ifdef CONFIG_ARCH_MXC
- fsl_udc_clk_release();
--
-+#endif
- /* DR has been stopped in usb_gadget_unregister_driver() */
- remove_proc_file();
-
-@@ -2570,7 +2580,7 @@ static int fsl_udc_remove(struct platfor
- dma_pool_destroy(udc_controller->td_pool);
- free_irq(udc_controller->irq, udc_controller);
- iounmap(dr_regs);
-- if (pdata->operating_mode == FSL_USB2_DR_DEVICE)
-+ if (res && (pdata->operating_mode == FSL_USB2_DR_DEVICE))
- release_mem_region(res->start, resource_size(res));
-
- /* free udc --wait for the release() finished */
---- a/drivers/usb/gadget/udc/fsl_usb2_udc.h
-+++ b/drivers/usb/gadget/udc/fsl_usb2_udc.h
-@@ -20,6 +20,10 @@
- #define USB_MAX_CTRL_PAYLOAD 64
- #define USB_DR_SYS_OFFSET 0x400
-
-+#ifdef CONFIG_SOC_LS1021A
-+#undef CONFIG_ARCH_MXC
-+#endif
-+
- /* USB DR device mode registers (Little Endian) */
- struct usb_dr_device {
- /* Capability register */
-@@ -597,18 +601,6 @@ struct platform_device;
- int fsl_udc_clk_init(struct platform_device *pdev);
- int fsl_udc_clk_finalize(struct platform_device *pdev);
- void fsl_udc_clk_release(void);
--#else
--static inline int fsl_udc_clk_init(struct platform_device *pdev)
--{
-- return 0;
--}
--static inline int fsl_udc_clk_finalize(struct platform_device *pdev)
--{
-- return 0;
--}
--static inline void fsl_udc_clk_release(void)
--{
--}
- #endif
-
- #endif
---- a/drivers/usb/host/Kconfig
-+++ b/drivers/usb/host/Kconfig
-@@ -164,8 +164,8 @@ config XPS_USB_HCD_XILINX
- devices only.
-
- config USB_EHCI_FSL
-- tristate "Support for Freescale PPC on-chip EHCI USB controller"
-- depends on FSL_SOC
-+ tristate "Support for Freescale QorIQ(ARM/PPC) on-chip EHCI USB controller"
-+ depends on USB_EHCI_HCD
- select USB_EHCI_ROOT_HUB_TT
- ---help---
- Variation of ARC USB block used in some Freescale chips.
---- a/drivers/usb/host/ehci-fsl.c
-+++ b/drivers/usb/host/ehci-fsl.c
-@@ -36,15 +36,127 @@
- #include <linux/platform_device.h>
- #include <linux/fsl_devices.h>
- #include <linux/of_platform.h>
-+#include <linux/io.h>
-+
-+#ifdef CONFIG_PPC
-+#include <asm/fsl_pm.h>
-+#include <linux/suspend.h>
-+#endif
-
- #include "ehci.h"
- #include "ehci-fsl.h"
-
-+#define FSL_USB_PHY_ADDR 0xffe214000
-+
-+struct ccsr_usb_port_ctrl {
-+ u32 ctrl;
-+ u32 drvvbuscfg;
-+ u32 pwrfltcfg;
-+ u32 sts;
-+ u8 res_14[0xc];
-+ u32 bistcfg;
-+ u32 biststs;
-+ u32 abistcfg;
-+ u32 abiststs;
-+ u8 res_30[0x10];
-+ u32 xcvrprg;
-+ u32 anaprg;
-+ u32 anadrv;
-+ u32 anasts;
-+};
-+
-+struct ccsr_usb_phy {
-+ u32 id;
-+ struct ccsr_usb_port_ctrl port1;
-+ u8 res_50[0xc];
-+ u32 tvr;
-+ u32 pllprg[4];
-+ u8 res_70[0x4];
-+ u32 anaccfg;
-+ u32 dbg;
-+ u8 res_7c[0x4];
-+ struct ccsr_usb_port_ctrl port2;
-+ u8 res_dc[0x334];
-+};
-+
- #define DRIVER_DESC "Freescale EHCI Host controller driver"
- #define DRV_NAME "ehci-fsl"
-
- static struct hc_driver __read_mostly fsl_ehci_hc_driver;
-
-+struct ehci_fsl {
-+ struct ehci_hcd ehci;
-+
-+#ifdef CONFIG_PM
-+struct ehci_regs saved_regs;
-+struct ccsr_usb_phy saved_phy_regs;
-+/* Saved USB PHY settings, need to restore after deep sleep. */
-+u32 usb_ctrl;
-+#endif
-+ /*
-+ * store current hcd state for otg;
-+ * have_hcd is true when host drv al already part of otg framework,
-+ * otherwise false;
-+ * hcd_add is true when otg framework wants to add host
-+ * drv as part of otg;flase when it wants to remove it
-+ */
-+unsigned have_hcd:1;
-+unsigned hcd_add:1;
-+};
-+
-+static struct ehci_fsl *hcd_to_ehci_fsl(struct usb_hcd *hcd)
-+{
-+struct ehci_hcd *ehci = hcd_to_ehci(hcd);
-+
-+return container_of(ehci, struct ehci_fsl, ehci);
-+}
-+
-+#if defined(CONFIG_FSL_USB2_OTG) || defined(CONFIG_FSL_USB2_OTG_MODULE)
-+static void do_change_hcd(struct work_struct *work)
-+{
-+struct ehci_hcd *ehci = container_of(work, struct ehci_hcd,
-+ change_hcd_work);
-+struct usb_hcd *hcd = ehci_to_hcd(ehci);
-+struct ehci_fsl *ehci_fsl = hcd_to_ehci_fsl(hcd);
-+void __iomem *non_ehci = hcd->regs;
-+int retval;
-+
-+ if (ehci_fsl->hcd_add && !ehci_fsl->have_hcd) {
-+ writel(USBMODE_CM_HOST, non_ehci + FSL_SOC_USB_USBMODE);
-+ /* host, gadget and otg share same int line */
-+ retval = usb_add_hcd(hcd, hcd->irq, IRQF_SHARED);
-+ if (retval == 0)
-+ ehci_fsl->have_hcd = 1;
-+ } else if (!ehci_fsl->hcd_add && ehci_fsl->have_hcd) {
-+ usb_remove_hcd(hcd);
-+ ehci_fsl->have_hcd = 0;
-+ }
-+}
-+#endif
-+
-+#if defined(CONFIG_FSL_USB2_OTG) || defined(CONFIG_FSL_USB2_OTG_MODULE)
-+static void do_change_hcd(struct work_struct *work)
-+{
-+ struct ehci_hcd *ehci = container_of(work, struct ehci_hcd,
-+ change_hcd_work);
-+ struct usb_hcd *hcd = ehci_to_hcd(ehci);
-+ struct ehci_fsl *ehci_fsl = hcd_to_ehci_fsl(hcd);
-+ void __iomem *non_ehci = hcd->regs;
-+ int retval;
-+
-+ if (ehci_fsl->hcd_add && !ehci_fsl->have_hcd) {
-+ writel(USBMODE_CM_HOST, non_ehci + FSL_SOC_USB_USBMODE);
-+ /* host, gadget and otg share same int line */
-+ retval = usb_add_hcd(hcd, hcd->irq, IRQF_SHARED);
-+ if (retval == 0)
-+ ehci_fsl->have_hcd = 1;
-+ } else if (!ehci_fsl->hcd_add && ehci_fsl->have_hcd) {
-+ usb_remove_hcd(hcd);
-+ ehci_fsl->have_hcd = 0;
-+ }
-+}
-+#endif
-+
- /* configure so an HC device and id are always provided */
- /* always called with process context; sleeping is OK */
-
-@@ -131,6 +243,12 @@ static int fsl_ehci_drv_probe(struct pla
- clrsetbits_be32(hcd->regs + FSL_SOC_USB_CTRL,
- CONTROL_REGISTER_W1C_MASK, 0x4);
-
-+ /* Set USB_EN bit to select ULPI phy for USB controller version 2.5 */
-+ if (pdata->controller_ver == FSL_USB_VER_2_5 &&
-+ pdata->phy_mode == FSL_USB2_PHY_ULPI)
-+ iowrite32be(USB_CTRL_USB_EN, hcd->regs + FSL_SOC_USB_CTRL);
-+
-+
- /*
- * Enable UTMI phy and program PTS field in UTMI mode before asserting
- * controller reset for USB Controller version 2.5
-@@ -143,16 +261,20 @@ static int fsl_ehci_drv_probe(struct pla
-
- /* Don't need to set host mode here. It will be done by tdi_reset() */
-
-- retval = usb_add_hcd(hcd, irq, IRQF_SHARED);
-+ retval = usb_add_hcd(hcd, irq, IRQF_SHARED | IRQF_NO_SUSPEND);
- if (retval != 0)
- goto err2;
- device_wakeup_enable(hcd->self.controller);
-
--#ifdef CONFIG_USB_OTG
-+#if defined(CONFIG_FSL_USB2_OTG) || defined(CONFIG_FSL_USB2_OTG_MODULE)
- if (pdata->operating_mode == FSL_USB2_DR_OTG) {
- struct ehci_hcd *ehci = hcd_to_ehci(hcd);
-+ struct ehci_fsl *ehci_fsl = hcd_to_ehci_fsl(hcd);
-
- hcd->usb_phy = usb_get_phy(USB_PHY_TYPE_USB2);
-+
-+ INIT_WORK(&ehci->change_hcd_work, do_change_hcd);
-+
- dev_dbg(&pdev->dev, "hcd=0x%p ehci=0x%p, phy=0x%p\n",
- hcd, ehci, hcd->usb_phy);
-
-@@ -168,6 +290,11 @@ static int fsl_ehci_drv_probe(struct pla
- retval = -ENODEV;
- goto err2;
- }
-+
-+ ehci_fsl->have_hcd = 1;
-+ } else {
-+ dev_err(&pdev->dev, "wrong operating mode\n");
-+ return -ENODEV;
- }
- #endif
- return retval;
-@@ -181,6 +308,17 @@ static int fsl_ehci_drv_probe(struct pla
- return retval;
- }
-
-+static bool usb_phy_clk_valid(struct usb_hcd *hcd)
-+{
-+ void __iomem *non_ehci = hcd->regs;
-+ bool ret = true;
-+
-+ if (!(ioread32be(non_ehci + FSL_SOC_USB_CTRL) & PHY_CLK_VALID))
-+ ret = false;
-+
-+ return ret;
-+}
-+
- static int ehci_fsl_setup_phy(struct usb_hcd *hcd,
- enum fsl_usb2_phy_modes phy_mode,
- unsigned int port_offset)
-@@ -219,6 +357,21 @@ static int ehci_fsl_setup_phy(struct usb
- /* fall through */
- case FSL_USB2_PHY_UTMI:
- case FSL_USB2_PHY_UTMI_DUAL:
-+ if (pdata->has_fsl_erratum_a006918) {
-+ pr_warn("fsl-ehci: USB PHY clock invalid\n");
-+ return -EINVAL;
-+ }
-+
-+ /* PHY_CLK_VALID bit is de-featured from all controller
-+ * versions below 2.4 and is to be checked only for
-+ * internal UTMI phy
-+ */
-+ if (pdata->controller_ver > FSL_USB_VER_2_4 &&
-+ pdata->have_sysif_regs && !usb_phy_clk_valid(hcd)) {
-+ pr_err("fsl-ehci: USB PHY clock invalid\n");
-+ return -EINVAL;
-+ }
-+
- if (pdata->have_sysif_regs && pdata->controller_ver) {
- /* controller version 1.6 or above */
- clrsetbits_be32(non_ehci + FSL_SOC_USB_CTRL,
-@@ -286,20 +439,18 @@ static int ehci_fsl_usb_setup(struct ehc
- if (pdata->has_fsl_erratum_a005275 == 1)
- ehci->has_fsl_hs_errata = 1;
-
-+ if (pdata->has_fsl_erratum_a005697 == 1)
-+ ehci->has_fsl_susp_errata = 1;
-+
- if ((pdata->operating_mode == FSL_USB2_DR_HOST) ||
- (pdata->operating_mode == FSL_USB2_DR_OTG))
- if (ehci_fsl_setup_phy(hcd, pdata->phy_mode, 0))
- return -EINVAL;
-
- if (pdata->operating_mode == FSL_USB2_MPH_HOST) {
-- unsigned int chip, rev, svr;
--
-- svr = mfspr(SPRN_SVR);
-- chip = svr >> 16;
-- rev = (svr >> 4) & 0xf;
-
- /* Deal with USB Erratum #14 on MPC834x Rev 1.0 & 1.1 chips */
-- if ((rev == 1) && (chip >= 0x8050) && (chip <= 0x8055))
-+ if (pdata->has_fsl_erratum_14 == 1)
- ehci->has_fsl_port_bug = 1;
-
- if (pdata->port_enables & FSL_USB2_PORT0_ENABLED)
-@@ -379,16 +530,57 @@ static int ehci_fsl_setup(struct usb_hcd
- return retval;
- }
-
--struct ehci_fsl {
-- struct ehci_hcd ehci;
-
- #ifdef CONFIG_PM
-- /* Saved USB PHY settings, need to restore after deep sleep. */
-- u32 usb_ctrl;
--#endif
--};
-+void __iomem *phy_reg;
-
--#ifdef CONFIG_PM
-+#ifdef CONFIG_PPC
-+/* save usb registers */
-+static int ehci_fsl_save_context(struct usb_hcd *hcd)
-+{
-+ struct ehci_fsl *ehci_fsl = hcd_to_ehci_fsl(hcd);
-+ struct ehci_hcd *ehci = hcd_to_ehci(hcd);
-+ void __iomem *non_ehci = hcd->regs;
-+ struct device *dev = hcd->self.controller;
-+ struct fsl_usb2_platform_data *pdata = dev_get_platdata(dev);
-+
-+ if (pdata->phy_mode == FSL_USB2_PHY_UTMI_DUAL) {
-+ phy_reg = ioremap(FSL_USB_PHY_ADDR,
-+ sizeof(struct ccsr_usb_phy));
-+ _memcpy_fromio((void *)&ehci_fsl->saved_phy_regs, phy_reg,
-+ sizeof(struct ccsr_usb_phy));
-+ }
-+
-+ _memcpy_fromio((void *)&ehci_fsl->saved_regs, ehci->regs,
-+ sizeof(struct ehci_regs));
-+ ehci_fsl->usb_ctrl = ioread32be(non_ehci + FSL_SOC_USB_CTRL);
-+
-+ return 0;
-+}
-+
-+/*Restore usb registers */
-+static int ehci_fsl_restore_context(struct usb_hcd *hcd)
-+{
-+ struct ehci_fsl *ehci_fsl = hcd_to_ehci_fsl(hcd);
-+ struct ehci_hcd *ehci = hcd_to_ehci(hcd);
-+ void __iomem *non_ehci = hcd->regs;
-+ struct device *dev = hcd->self.controller;
-+ struct fsl_usb2_platform_data *pdata = dev_get_platdata(dev);
-+
-+ if (pdata->phy_mode == FSL_USB2_PHY_UTMI_DUAL) {
-+ if (phy_reg)
-+ _memcpy_toio(phy_reg,
-+ (void *)&ehci_fsl->saved_phy_regs,
-+ sizeof(struct ccsr_usb_phy));
-+ }
-+
-+ _memcpy_toio(ehci->regs, (void *)&ehci_fsl->saved_regs,
-+ sizeof(struct ehci_regs));
-+ iowrite32be(ehci_fsl->usb_ctrl, non_ehci + FSL_SOC_USB_CTRL);
-+
-+ return 0;
-+}
-+#endif
-
- #ifdef CONFIG_PPC_MPC512x
- static int ehci_fsl_mpc512x_drv_suspend(struct device *dev)
-@@ -535,26 +727,45 @@ static inline int ehci_fsl_mpc512x_drv_r
- }
- #endif /* CONFIG_PPC_MPC512x */
-
--static struct ehci_fsl *hcd_to_ehci_fsl(struct usb_hcd *hcd)
--{
-- struct ehci_hcd *ehci = hcd_to_ehci(hcd);
--
-- return container_of(ehci, struct ehci_fsl, ehci);
--}
--
- static int ehci_fsl_drv_suspend(struct device *dev)
- {
- struct usb_hcd *hcd = dev_get_drvdata(dev);
- struct ehci_fsl *ehci_fsl = hcd_to_ehci_fsl(hcd);
- void __iomem *non_ehci = hcd->regs;
-+#if defined(CONFIG_FSL_USB2_OTG) || defined(CONFIG_FSL_USB2_OTG_MODULE)
-+ struct usb_bus host = hcd->self;
-+#endif
-+
-+#ifdef CONFIG_PPC
-+suspend_state_t pm_state;
-+/* FIXME:Need to port fsl_pm.h before enable below code. */
-+/*pm_state = pm_suspend_state();*/
-+pm_state = PM_SUSPEND_MEM;
-+
-+if (pm_state == PM_SUSPEND_MEM)
-+ ehci_fsl_save_context(hcd);
-+#endif
-
- if (of_device_is_compatible(dev->parent->of_node,
- "fsl,mpc5121-usb2-dr")) {
- return ehci_fsl_mpc512x_drv_suspend(dev);
- }
-
-+#if defined(CONFIG_FSL_USB2_OTG) || defined(CONFIG_FSL_USB2_OTG_MODULE)
-+ if (host.is_otg) {
-+ struct ehci_hcd *ehci = hcd_to_ehci(hcd);
-+
-+ /* remove hcd */
-+ ehci_fsl->hcd_add = 0;
-+ schedule_work(&ehci->change_hcd_work);
-+ host.is_otg = 0;
-+ return 0;
-+ }
-+#endif
-+
- ehci_prepare_ports_for_controller_suspend(hcd_to_ehci(hcd),
- device_may_wakeup(dev));
-+
- if (!fsl_deep_sleep())
- return 0;
-
-@@ -568,12 +779,36 @@ static int ehci_fsl_drv_resume(struct de
- struct ehci_fsl *ehci_fsl = hcd_to_ehci_fsl(hcd);
- struct ehci_hcd *ehci = hcd_to_ehci(hcd);
- void __iomem *non_ehci = hcd->regs;
-+#if defined(CONFIG_FSL_USB2_OTG) || defined(CONFIG_FSL_USB2_OTG_MODULE)
-+ struct usb_bus host = hcd->self;
-+#endif
-+
-+#ifdef CONFIG_PPC
-+suspend_state_t pm_state;
-+/* FIXME:Need to port fsl_pm.h before enable below code.*/
-+/* pm_state = pm_suspend_state(); */
-+pm_state = PM_SUSPEND_MEM;
-+
-+if (pm_state == PM_SUSPEND_MEM)
-+ ehci_fsl_restore_context(hcd);
-+#endif
-
- if (of_device_is_compatible(dev->parent->of_node,
- "fsl,mpc5121-usb2-dr")) {
- return ehci_fsl_mpc512x_drv_resume(dev);
- }
-
-+#if defined(CONFIG_FSL_USB2_OTG) || defined(CONFIG_FSL_USB2_OTG_MODULE)
-+ if (host.is_otg) {
-+ /* add hcd */
-+ ehci_fsl->hcd_add = 1;
-+ schedule_work(&ehci->change_hcd_work);
-+ usb_hcd_resume_root_hub(hcd);
-+ host.is_otg = 0;
-+ return 0;
-+ }
-+#endif
-+
- ehci_prepare_ports_for_controller_resume(ehci);
- if (!fsl_deep_sleep())
- return 0;
---- a/drivers/usb/host/ehci-fsl.h
-+++ b/drivers/usb/host/ehci-fsl.h
-@@ -63,4 +63,7 @@
- #define UTMI_PHY_EN (1<<9)
- #define ULPI_PHY_CLK_SEL (1<<10)
- #define PHY_CLK_VALID (1<<17)
-+
-+/* Retry count for checking UTMI PHY CLK validity */
-+#define UTMI_PHY_CLK_VALID_CHK_RETRY 5
- #endif /* _EHCI_FSL_H */
---- a/drivers/usb/host/ehci-hub.c
-+++ b/drivers/usb/host/ehci-hub.c
-@@ -278,6 +278,8 @@ static int ehci_bus_suspend (struct usb_
- else if ((t1 & PORT_PE) && !(t1 & PORT_SUSPEND)) {
- t2 |= PORT_SUSPEND;
- set_bit(port, &ehci->bus_suspended);
-+ if (ehci_has_fsl_susp_errata(ehci))
-+ usleep_range(10000, 20000);
- }
-
- /* enable remote wakeup on all ports, if told to do so */
-@@ -305,6 +307,8 @@ static int ehci_bus_suspend (struct usb_
- USB_PORT_STAT_HIGH_SPEED)
- fs_idle_delay = true;
- ehci_writel(ehci, t2, reg);
-+ if (ehci_has_fsl_susp_errata(ehci))
-+ usleep_range(10000, 20000);
- changed = 1;
- }
- }
---- a/drivers/usb/host/ehci.h
-+++ b/drivers/usb/host/ehci.h
-@@ -180,6 +180,9 @@ struct ehci_hcd { /* one per controlle
- unsigned periodic_count; /* periodic activity count */
- unsigned uframe_periodic_max; /* max periodic time per uframe */
-
-+#if defined(CONFIG_FSL_USB2_OTG) || defined(CONFIG_FSL_USB2_OTG_MODULE)
-+ struct work_struct change_hcd_work;
-+#endif
-
- /* list of itds & sitds completed while now_frame was still active */
- struct list_head cached_itd_list;
-@@ -219,6 +222,7 @@ struct ehci_hcd { /* one per controlle
- unsigned no_selective_suspend:1;
- unsigned has_fsl_port_bug:1; /* FreeScale */
- unsigned has_fsl_hs_errata:1; /* Freescale HS quirk */
-+ unsigned has_fsl_susp_errata:1; /*Freescale SUSP quirk*/
- unsigned big_endian_mmio:1;
- unsigned big_endian_desc:1;
- unsigned big_endian_capbase:1;
-@@ -704,10 +708,15 @@ ehci_port_speed(struct ehci_hcd *ehci, u
- #if defined(CONFIG_PPC_85xx)
- /* Some Freescale processors have an erratum (USB A-005275) in which
- * incoming packets get corrupted in HS mode
-+ * Some Freescale processors have an erratum (USB A-005697) in which
-+ * we need to wait for 10ms for bus to fo into suspend mode after
-+ * setting SUSP bit
- */
- #define ehci_has_fsl_hs_errata(e) ((e)->has_fsl_hs_errata)
-+#define ehci_has_fsl_susp_errata(e) ((e)->has_fsl_susp_errata)
- #else
- #define ehci_has_fsl_hs_errata(e) (0)
-+#define ehci_has_fsl_susp_errata(e) (0)
- #endif
-
- /*
---- a/drivers/usb/host/fsl-mph-dr-of.c
-+++ b/drivers/usb/host/fsl-mph-dr-of.c
-@@ -98,8 +98,8 @@ static struct platform_device *fsl_usb2_
-
- pdev->dev.coherent_dma_mask = ofdev->dev.coherent_dma_mask;
-
-- if (!pdev->dev.dma_mask)
-- pdev->dev.dma_mask = &ofdev->dev.coherent_dma_mask;
-+ if (!pdev->dev.dma_mask && ofdev->dev.of_node)
-+ of_dma_configure(&pdev->dev, ofdev->dev.of_node);
- else
- dma_set_mask(&pdev->dev, DMA_BIT_MASK(32));
-
-@@ -226,6 +226,18 @@ static int fsl_usb2_mph_dr_of_probe(stru
- of_property_read_bool(np, "fsl,usb-erratum-a007792");
- pdata->has_fsl_erratum_a005275 =
- of_property_read_bool(np, "fsl,usb-erratum-a005275");
-+ pdata->has_fsl_erratum_a005697 =
-+ of_property_read_bool(np, "fsl,usb_erratum-a005697");
-+ if (of_get_property(np, "fsl,erratum_a006918", NULL))
-+ pdata->has_fsl_erratum_a006918 = 1;
-+ else
-+ pdata->has_fsl_erratum_a006918 = 0;
-+
-+ if (of_get_property(np, "fsl,usb_erratum_14", NULL))
-+ pdata->has_fsl_erratum_14 = 1;
-+ else
-+ pdata->has_fsl_erratum_14 = 0;
-+
-
- /*
- * Determine whether phy_clk_valid needs to be checked
---- a/drivers/usb/host/xhci-hub.c
-+++ b/drivers/usb/host/xhci-hub.c
-@@ -562,12 +562,34 @@ void xhci_set_link_state(struct xhci_hcd
- int port_id, u32 link_state)
- {
- u32 temp;
-+ u32 portpmsc_u2_backup = 0;
-+
-+ /* Backup U2 timeout info before initiating U3 entry erratum A-010131 */
-+ if (xhci->shared_hcd->speed >= HCD_USB3 &&
-+ link_state == USB_SS_PORT_LS_U3 &&
-+ (xhci->quirks & XHCI_DIS_U1U2_WHEN_U3)) {
-+ portpmsc_u2_backup = readl(port_array[port_id] + PORTPMSC);
-+ portpmsc_u2_backup &= PORT_U2_TIMEOUT_MASK;
-+ temp = readl(port_array[port_id] + PORTPMSC);
-+ temp |= PORT_U2_TIMEOUT_MASK;
-+ writel(temp, port_array[port_id] + PORTPMSC);
-+ }
-
- temp = readl(port_array[port_id]);
- temp = xhci_port_state_to_neutral(temp);
- temp &= ~PORT_PLS_MASK;
- temp |= PORT_LINK_STROBE | link_state;
- writel(temp, port_array[port_id]);
-+
-+ /* Restore U2 timeout info after U3 entry complete */
-+ if (xhci->shared_hcd->speed >= HCD_USB3 &&
-+ link_state == USB_SS_PORT_LS_U3 &&
-+ (xhci->quirks & XHCI_DIS_U1U2_WHEN_U3)) {
-+ temp = readl(port_array[port_id] + PORTPMSC);
-+ temp &= ~PORT_U2_TIMEOUT_MASK;
-+ temp |= portpmsc_u2_backup;
-+ writel(temp, port_array[port_id] + PORTPMSC);
-+ }
- }
-
- static void xhci_set_remote_wake_mask(struct xhci_hcd *xhci,
---- a/drivers/usb/host/xhci-plat.c
-+++ b/drivers/usb/host/xhci-plat.c
-@@ -220,8 +220,22 @@ static int xhci_plat_probe(struct platfo
- goto disable_clk;
- }
-
-- if (device_property_read_bool(&pdev->dev, "usb3-lpm-capable"))
-+ if (device_property_read_bool(&pdev->dev, "usb3-lpm-capable")) {
- xhci->quirks |= XHCI_LPM_SUPPORT;
-+ if (device_property_read_bool(&pdev->dev,
-+ "snps,dis-u1u2-when-u3-quirk"))
-+ xhci->quirks |= XHCI_DIS_U1U2_WHEN_U3;
-+ }
-+
-+ if (device_property_read_bool(&pdev->dev, "quirk-reverse-in-out"))
-+ xhci->quirks |= XHCI_REVERSE_IN_OUT;
-+
-+ if (device_property_read_bool(&pdev->dev,
-+ "quirk-stop-transfer-in-block"))
-+ xhci->quirks |= XHCI_STOP_TRANSFER_IN_BLOCK;
-+
-+ if (device_property_read_bool(&pdev->dev, "quirk-stop-ep-in-u1"))
-+ xhci->quirks |= XHCI_STOP_EP_IN_U1;
-
- if (device_property_read_bool(&pdev->dev, "quirk-broken-port-ped"))
- xhci->quirks |= XHCI_BROKEN_PORT_PED;
---- a/drivers/usb/host/xhci-ring.c
-+++ b/drivers/usb/host/xhci-ring.c
-@@ -1852,14 +1852,17 @@ static int finish_td(struct xhci_hcd *xh
- union xhci_trb *event_trb, struct xhci_transfer_event *event,
- struct xhci_virt_ep *ep, int *status, bool skip)
- {
-+ struct xhci_dequeue_state deq_state;
- struct xhci_virt_device *xdev;
- struct xhci_ring *ep_ring;
-+ unsigned int stream_id;
- unsigned int slot_id;
- int ep_index;
- struct urb *urb = NULL;
- struct xhci_ep_ctx *ep_ctx;
- int ret = 0;
- struct urb_priv *urb_priv;
-+ u32 remaining;
- u32 trb_comp_code;
-
- slot_id = TRB_TO_SLOT_ID(le32_to_cpu(event->flags));
-@@ -1885,13 +1888,29 @@ static int finish_td(struct xhci_hcd *xh
- if (trb_comp_code == COMP_STALL ||
- xhci_requires_manual_halt_cleanup(xhci, ep_ctx,
- trb_comp_code)) {
-- /* Issue a reset endpoint command to clear the host side
-- * halt, followed by a set dequeue command to move the
-- * dequeue pointer past the TD.
-- * The class driver clears the device side halt later.
-+ /*
-+ * A-007463: After transaction error, controller switches
-+ * control transfer data stage from IN to OUT direction.
- */
-- xhci_cleanup_halted_endpoint(xhci, slot_id, ep_index,
-+ remaining = EVENT_TRB_LEN(le32_to_cpu(event->transfer_len));
-+ if (remaining && xhci_requires_manual_halt_cleanup(xhci, ep_ctx,
-+ trb_comp_code) &&
-+ (xhci->quirks & XHCI_REVERSE_IN_OUT)) {
-+ memset(&deq_state, 0, sizeof(deq_state));
-+ xhci_find_new_dequeue_state(xhci, slot_id,
-+ ep_index, td->urb->stream_id, td, &deq_state);
-+ xhci_queue_new_dequeue_state(xhci, slot_id, ep_index,
-+ stream_id, &deq_state);
-+ xhci_ring_cmd_db(xhci);
-+ } else {
-+ /* Issue a reset endpoint command to clear the host side
-+ * halt, followed by a set dequeue command to move the
-+ * dequeue pointer past the TD.
-+ * The class driver clears the device side halt later.
-+ */
-+ xhci_cleanup_halted_endpoint(xhci, slot_id, ep_index,
- ep_ring->stream_id, td, event_trb);
-+ }
- } else {
- /* Update ring dequeue pointer */
- while (ep_ring->dequeue != td->last_trb)
---- a/drivers/usb/host/xhci.c
-+++ b/drivers/usb/host/xhci.c
-@@ -1609,14 +1609,38 @@ int xhci_urb_dequeue(struct usb_hcd *hcd
- ret = -ENOMEM;
- goto done;
- }
-- ep->ep_state |= EP_HALT_PENDING;
-- ep->stop_cmds_pending++;
-- ep->stop_cmd_timer.expires = jiffies +
-+ /*
-+ *A-009611: Issuing an End Transfer command on an IN endpoint.
-+ *when a transfer is in progress on USB blocks the transmission
-+ *Workaround: Software must wait for all existing TRBs to
-+ *complete before issuing End transfer command.
-+ */
-+ if ((ep_ring->enqueue == ep_ring->dequeue &&
-+ (xhci->quirks & XHCI_STOP_TRANSFER_IN_BLOCK)) ||
-+ !(xhci->quirks & XHCI_STOP_TRANSFER_IN_BLOCK)) {
-+ ep->ep_state |= EP_HALT_PENDING;
-+ ep->stop_cmds_pending++;
-+ ep->stop_cmd_timer.expires = jiffies +
- XHCI_STOP_EP_CMD_TIMEOUT * HZ;
-- add_timer(&ep->stop_cmd_timer);
-- xhci_queue_stop_endpoint(xhci, command, urb->dev->slot_id,
-- ep_index, 0);
-- xhci_ring_cmd_db(xhci);
-+ add_timer(&ep->stop_cmd_timer);
-+ xhci_queue_stop_endpoint(xhci, command,
-+ urb->dev->slot_id,
-+ ep_index, 0);
-+ xhci_ring_cmd_db(xhci);
-+ }
-+
-+ /*
-+ *A-009668: Stop Endpoint Command does not complete.
-+ *Workaround: Instead of issuing a Stop Endpoint Command,
-+ *issue a Disable Slot Command with the corresponding slot ID.
-+ *Alternately, you can issue an Address Device Command with
-+ *BSR=1
-+ */
-+ if ((urb->dev->speed <= USB_SPEED_HIGH) &&
-+ (xhci->quirks & XHCI_STOP_EP_IN_U1)) {
-+ xhci_queue_slot_control(xhci, command, TRB_DISABLE_SLOT,
-+ urb->dev->slot_id);
-+ }
- }
- done:
- spin_unlock_irqrestore(&xhci->lock, flags);
---- a/drivers/usb/host/xhci.h
-+++ b/drivers/usb/host/xhci.h
-@@ -1625,7 +1625,7 @@ struct xhci_hcd {
- #define XHCI_STATE_REMOVING (1 << 2)
- /* Statistics */
- int error_bitmask;
-- unsigned int quirks;
-+ u64 quirks;
- #define XHCI_LINK_TRB_QUIRK (1 << 0)
- #define XHCI_RESET_EP_QUIRK (1 << 1)
- #define XHCI_NEC_HOST (1 << 2)
-@@ -1661,6 +1661,10 @@ struct xhci_hcd {
- #define XHCI_SSIC_PORT_UNUSED (1 << 22)
- #define XHCI_NO_64BIT_SUPPORT (1 << 23)
- #define XHCI_MISSING_CAS (1 << 24)
-+#define XHCI_REVERSE_IN_OUT (1 << 29)
-+#define XHCI_STOP_TRANSFER_IN_BLOCK (1 << 30)
-+#define XHCI_STOP_EP_IN_U1 (1 << 31)
-+#define XHCI_DIS_U1U2_WHEN_U3 (1 << 32)
- /* For controller with a broken Port Disable implementation */
- #define XHCI_BROKEN_PORT_PED (1 << 25)
- #define XHCI_LIMIT_ENDPOINT_INTERVAL_7 (1 << 26)
---- a/drivers/usb/phy/phy-fsl-usb.c
-+++ b/drivers/usb/phy/phy-fsl-usb.c
-@@ -1,5 +1,5 @@
- /*
-- * Copyright (C) 2007,2008 Freescale semiconductor, Inc.
-+ * Copyright 2007,2008 Freescale Semiconductor, Inc.
- *
- * Author: Li Yang <LeoLi@freescale.com>
- * Jerry Huang <Chang-Ming.Huang@freescale.com>
-@@ -463,6 +463,7 @@ void otg_reset_controller(void)
- int fsl_otg_start_host(struct otg_fsm *fsm, int on)
- {
- struct usb_otg *otg = fsm->otg;
-+ struct usb_bus *host = otg->host;
- struct device *dev;
- struct fsl_otg *otg_dev =
- container_of(otg->usb_phy, struct fsl_otg, phy);
-@@ -486,6 +487,7 @@ int fsl_otg_start_host(struct otg_fsm *f
- otg_reset_controller();
- VDBG("host on......\n");
- if (dev->driver->pm && dev->driver->pm->resume) {
-+ host->is_otg = 1;
- retval = dev->driver->pm->resume(dev);
- if (fsm->id) {
- /* default-b */
-@@ -510,8 +512,11 @@ int fsl_otg_start_host(struct otg_fsm *f
- else {
- VDBG("host off......\n");
- if (dev && dev->driver) {
-- if (dev->driver->pm && dev->driver->pm->suspend)
-+ if (dev->driver->pm &&
-+ dev->driver->pm->suspend) {
-+ host->is_otg = 1;
- retval = dev->driver->pm->suspend(dev);
-+ }
- if (fsm->id)
- /* default-b */
- fsl_otg_drv_vbus(fsm, 0);
-@@ -539,8 +544,17 @@ int fsl_otg_start_gadget(struct otg_fsm
- dev = otg->gadget->dev.parent;
-
- if (on) {
-- if (dev->driver->resume)
-+ /* Delay gadget resume to synchronize between host and gadget
-+ * drivers. Upon role-reversal host drv is shutdown by kernel
-+ * worker thread. By the time host drv shuts down, controller
-+ * gets programmed for gadget role. Shutting host drv after
-+ * this results in controller getting reset, and it stops
-+ * responding to otg events
-+ */
-+ if (dev->driver->resume) {
-+ msleep(1000);
- dev->driver->resume(dev);
-+ }
- } else {
- if (dev->driver->suspend)
- dev->driver->suspend(dev, otg_suspend_state);
-@@ -672,6 +686,10 @@ static void fsl_otg_event(struct work_st
- fsl_otg_start_host(fsm, 0);
- otg_drv_vbus(fsm, 0);
- fsl_otg_start_gadget(fsm, 1);
-+ } else {
-+ fsl_otg_start_gadget(fsm, 0);
-+ otg_drv_vbus(fsm, 1);
-+ fsl_otg_start_host(fsm, 1);
- }
- }
-
-@@ -724,6 +742,7 @@ irqreturn_t fsl_otg_isr(int irq, void *d
- {
- struct otg_fsm *fsm = &((struct fsl_otg *)dev_id)->fsm;
- struct usb_otg *otg = ((struct fsl_otg *)dev_id)->phy.otg;
-+ struct fsl_otg *otg_dev = dev_id;
- u32 otg_int_src, otg_sc;
-
- otg_sc = fsl_readl(&usb_dr_regs->otgsc);
-@@ -753,18 +772,8 @@ irqreturn_t fsl_otg_isr(int irq, void *d
- otg->gadget->is_a_peripheral = !fsm->id;
- VDBG("ID int (ID is %d)\n", fsm->id);
-
-- if (fsm->id) { /* switch to gadget */
-- schedule_delayed_work(
-- &((struct fsl_otg *)dev_id)->otg_event,
-- 100);
-- } else { /* switch to host */
-- cancel_delayed_work(&
-- ((struct fsl_otg *)dev_id)->
-- otg_event);
-- fsl_otg_start_gadget(fsm, 0);
-- otg_drv_vbus(fsm, 1);
-- fsl_otg_start_host(fsm, 1);
-- }
-+ schedule_delayed_work(&otg_dev->otg_event, 100);
-+
- return IRQ_HANDLED;
- }
- }
-@@ -925,12 +934,32 @@ int usb_otg_start(struct platform_device
- temp &= ~(PORTSC_PHY_TYPE_SEL | PORTSC_PTW);
- switch (pdata->phy_mode) {
- case FSL_USB2_PHY_ULPI:
-+ if (pdata->controller_ver) {
-+ /* controller version 1.6 or above */
-+ setbits32(&p_otg->dr_mem_map->control,
-+ USB_CTRL_ULPI_PHY_CLK_SEL);
-+ /*
-+ * Due to controller issue of PHY_CLK_VALID in ULPI
-+ * mode, we set USB_CTRL_USB_EN before checking
-+ * PHY_CLK_VALID, otherwise PHY_CLK_VALID doesn't work.
-+ */
-+ clrsetbits_be32(&p_otg->dr_mem_map->control,
-+ USB_CTRL_UTMI_PHY_EN, USB_CTRL_IOENB);
-+ }
- temp |= PORTSC_PTS_ULPI;
- break;
- case FSL_USB2_PHY_UTMI_WIDE:
- temp |= PORTSC_PTW_16BIT;
- /* fall through */
- case FSL_USB2_PHY_UTMI:
-+ if (pdata->controller_ver) {
-+ /* controller version 1.6 or above */
-+ setbits32(&p_otg->dr_mem_map->control,
-+ USB_CTRL_UTMI_PHY_EN);
-+ /* Delay for UTMI PHY CLK to become stable - 10ms */
-+ mdelay(FSL_UTMI_PHY_DLY);
-+ }
-+ setbits32(&p_otg->dr_mem_map->control, USB_CTRL_UTMI_PHY_EN);
- temp |= PORTSC_PTS_UTMI;
- /* fall through */
- default:
---- a/drivers/usb/phy/phy-fsl-usb.h
-+++ b/drivers/usb/phy/phy-fsl-usb.h
-@@ -199,6 +199,14 @@
- /* control Register Bit Masks */
- #define USB_CTRL_IOENB (0x1<<2)
- #define USB_CTRL_ULPI_INT0EN (0x1<<0)
-+#define USB_CTRL_WU_INT_EN (0x1<<1)
-+#define USB_CTRL_LINE_STATE_FILTER__EN (0x1<<3)
-+#define USB_CTRL_KEEP_OTG_ON (0x1<<4)
-+#define USB_CTRL_OTG_PORT (0x1<<5)
-+#define USB_CTRL_PLL_RESET (0x1<<8)
-+#define USB_CTRL_UTMI_PHY_EN (0x1<<9)
-+#define USB_CTRL_ULPI_PHY_CLK_SEL (0x1<<10)
-+#define USB_CTRL_PHY_CLK_VALID (0x1<<17)
-
- /* BCSR5 */
- #define BCSR5_INT_USB (0x02)
---- a/include/linux/usb.h
-+++ b/include/linux/usb.h
-@@ -362,6 +362,7 @@ struct usb_bus {
- * for control transfers?
- */
- u8 otg_port; /* 0, or number of OTG/HNP port */
-+ unsigned is_otg:1; /* true when host is also otg */
- unsigned is_b_host:1; /* true during some HNP roleswitches */
- unsigned b_hnp_enable:1; /* OTG: did A-Host enable HNP? */
- unsigned no_stop_on_short:1; /*
---- a/include/linux/usb/of.h
-+++ b/include/linux/usb/of.h
-@@ -11,6 +11,8 @@
- #include <linux/usb/otg.h>
- #include <linux/usb/phy.h>
-
-+enum usb_dr_mode of_usb_get_dr_mode(struct device_node *np);
-+
- #if IS_ENABLED(CONFIG_OF)
- enum usb_dr_mode of_usb_get_dr_mode_by_phy(struct device_node *np, int arg0);
- bool of_usb_host_tpl_support(struct device_node *np);
diff --git a/target/linux/layerscape/patches-4.9/818-vfio-support-layerscape.patch b/target/linux/layerscape/patches-4.9/818-vfio-support-layerscape.patch
deleted file mode 100644
index f2a65ffba5..0000000000
--- a/target/linux/layerscape/patches-4.9/818-vfio-support-layerscape.patch
+++ /dev/null
@@ -1,1165 +0,0 @@
-From e6af99cc1d56322fc960d072af1a7e0e9285b90c Mon Sep 17 00:00:00 2001
-From: Yangbo Lu <yangbo.lu@nxp.com>
-Date: Thu, 5 Jul 2018 17:39:43 +0800
-Subject: [PATCH 30/32] vfio: support layerscape
-
-This is an integrated patch for layerscape vfio support.
-
-Signed-off-by: Bharat Bhushan <Bharat.Bhushan@nxp.com>
-Signed-off-by: Eric Auger <eric.auger@redhat.com>
-Signed-off-by: Robin Murphy <robin.murphy@arm.com>
-Signed-off-by: Wei Yongjun <weiyongjun1@huawei.com>
-Signed-off-by: Yangbo Lu <yangbo.lu@nxp.com>
----
- drivers/vfio/Kconfig | 1 +
- drivers/vfio/Makefile | 1 +
- drivers/vfio/fsl-mc/Kconfig | 9 +
- drivers/vfio/fsl-mc/Makefile | 2 +
- drivers/vfio/fsl-mc/vfio_fsl_mc.c | 752 ++++++++++++++++++++++
- drivers/vfio/fsl-mc/vfio_fsl_mc_intr.c | 199 ++++++
- drivers/vfio/fsl-mc/vfio_fsl_mc_private.h | 55 ++
- drivers/vfio/vfio_iommu_type1.c | 39 +-
- include/uapi/linux/vfio.h | 1 +
- 9 files changed, 1057 insertions(+), 2 deletions(-)
- create mode 100644 drivers/vfio/fsl-mc/Kconfig
- create mode 100644 drivers/vfio/fsl-mc/Makefile
- create mode 100644 drivers/vfio/fsl-mc/vfio_fsl_mc.c
- create mode 100644 drivers/vfio/fsl-mc/vfio_fsl_mc_intr.c
- create mode 100644 drivers/vfio/fsl-mc/vfio_fsl_mc_private.h
-
---- a/drivers/vfio/Kconfig
-+++ b/drivers/vfio/Kconfig
-@@ -48,4 +48,5 @@ menuconfig VFIO_NOIOMMU
-
- source "drivers/vfio/pci/Kconfig"
- source "drivers/vfio/platform/Kconfig"
-+source "drivers/vfio/fsl-mc/Kconfig"
- source "virt/lib/Kconfig"
---- a/drivers/vfio/Makefile
-+++ b/drivers/vfio/Makefile
-@@ -7,3 +7,4 @@ obj-$(CONFIG_VFIO_IOMMU_SPAPR_TCE) += vf
- obj-$(CONFIG_VFIO_SPAPR_EEH) += vfio_spapr_eeh.o
- obj-$(CONFIG_VFIO_PCI) += pci/
- obj-$(CONFIG_VFIO_PLATFORM) += platform/
-+obj-$(CONFIG_VFIO_FSL_MC) += fsl-mc/
---- /dev/null
-+++ b/drivers/vfio/fsl-mc/Kconfig
-@@ -0,0 +1,9 @@
-+config VFIO_FSL_MC
-+ tristate "VFIO support for QorIQ DPAA2 fsl-mc bus devices"
-+ depends on VFIO && FSL_MC_BUS && EVENTFD
-+ help
-+ Driver to enable support for the VFIO QorIQ DPAA2 fsl-mc
-+ (Management Complex) devices. This is required to passthrough
-+ fsl-mc bus devices using the VFIO framework.
-+
-+ If you don't know what to do here, say N.
---- /dev/null
-+++ b/drivers/vfio/fsl-mc/Makefile
-@@ -0,0 +1,2 @@
-+vfio-fsl_mc-y := vfio_fsl_mc.o
-+obj-$(CONFIG_VFIO_FSL_MC) += vfio_fsl_mc.o vfio_fsl_mc_intr.o
---- /dev/null
-+++ b/drivers/vfio/fsl-mc/vfio_fsl_mc.c
-@@ -0,0 +1,752 @@
-+/*
-+ * Freescale Management Complex (MC) device passthrough using VFIO
-+ *
-+ * Copyright (C) 2013-2016 Freescale Semiconductor, Inc.
-+ * Copyright 2016-2017 NXP
-+ * Author: Bharat Bhushan <bharat.bhushan@nxp.com>
-+ *
-+ * This file is licensed under the terms of the GNU General Public
-+ * License version 2. This program is licensed "as is" without any
-+ * warranty of any kind, whether express or implied.
-+ */
-+
-+#include <linux/device.h>
-+#include <linux/iommu.h>
-+#include <linux/module.h>
-+#include <linux/mutex.h>
-+#include <linux/slab.h>
-+#include <linux/types.h>
-+#include <linux/vfio.h>
-+#include <linux/delay.h>
-+
-+#include <linux/fsl/mc.h>
-+
-+#include "vfio_fsl_mc_private.h"
-+
-+#define DRIVER_VERSION "0.10"
-+#define DRIVER_AUTHOR "Bharat Bhushan <bharat.bhushan@nxp.com>"
-+#define DRIVER_DESC "VFIO for FSL-MC devices - User Level meta-driver"
-+
-+static DEFINE_MUTEX(driver_lock);
-+
-+/* FSl-MC device regions (address and size) are aligned to 64K.
-+ * While MC firmware reports size less than 64K for some objects (it actually
-+ * reports size which does not include reserved space beyond valid bytes).
-+ * Align the size to PAGE_SIZE for userspace to mmap.
-+ */
-+static size_t aligned_region_size(struct fsl_mc_device *mc_dev, int index)
-+{
-+ size_t size;
-+
-+ size = resource_size(&mc_dev->regions[index]);
-+ return PAGE_ALIGN(size);
-+}
-+
-+static int vfio_fsl_mc_regions_init(struct vfio_fsl_mc_device *vdev)
-+{
-+ struct fsl_mc_device *mc_dev = vdev->mc_dev;
-+ int count = mc_dev->obj_desc.region_count;
-+ int i;
-+
-+ vdev->regions = kcalloc(count, sizeof(struct vfio_fsl_mc_region),
-+ GFP_KERNEL);
-+ if (!vdev->regions)
-+ return -ENOMEM;
-+
-+ for (i = 0; i < mc_dev->obj_desc.region_count; i++) {
-+ vdev->regions[i].addr = mc_dev->regions[i].start;
-+ vdev->regions[i].size = aligned_region_size(mc_dev, i);
-+ vdev->regions[i].type = VFIO_FSL_MC_REGION_TYPE_MMIO;
-+ if (mc_dev->regions[i].flags & IORESOURCE_CACHEABLE)
-+ vdev->regions[i].type |=
-+ VFIO_FSL_MC_REGION_TYPE_CACHEABLE;
-+ vdev->regions[i].flags = VFIO_REGION_INFO_FLAG_MMAP;
-+ vdev->regions[i].flags |= VFIO_REGION_INFO_FLAG_READ;
-+ if (!(mc_dev->regions[i].flags & IORESOURCE_READONLY))
-+ vdev->regions[i].flags |= VFIO_REGION_INFO_FLAG_WRITE;
-+ }
-+
-+ vdev->num_regions = mc_dev->obj_desc.region_count;
-+ return 0;
-+}
-+
-+static void vfio_fsl_mc_regions_cleanup(struct vfio_fsl_mc_device *vdev)
-+{
-+ int i;
-+
-+ for (i = 0; i < vdev->num_regions; i++)
-+ iounmap(vdev->regions[i].ioaddr);
-+
-+ vdev->num_regions = 0;
-+ kfree(vdev->regions);
-+}
-+
-+static int vfio_fsl_mc_open(void *device_data)
-+{
-+ struct vfio_fsl_mc_device *vdev = device_data;
-+ int ret;
-+
-+ if (!try_module_get(THIS_MODULE))
-+ return -ENODEV;
-+
-+ mutex_lock(&driver_lock);
-+ if (!vdev->refcnt) {
-+ ret = vfio_fsl_mc_regions_init(vdev);
-+ if (ret)
-+ goto error_region_init;
-+
-+ ret = vfio_fsl_mc_irqs_init(vdev);
-+ if (ret)
-+ goto error_irq_init;
-+ }
-+
-+ vdev->refcnt++;
-+ mutex_unlock(&driver_lock);
-+ return 0;
-+
-+error_irq_init:
-+ vfio_fsl_mc_regions_cleanup(vdev);
-+error_region_init:
-+ mutex_unlock(&driver_lock);
-+ if (ret)
-+ module_put(THIS_MODULE);
-+
-+ return ret;
-+}
-+
-+static void vfio_fsl_mc_release(void *device_data)
-+{
-+ struct vfio_fsl_mc_device *vdev = device_data;
-+ struct fsl_mc_device *mc_dev = vdev->mc_dev;
-+
-+ mutex_lock(&driver_lock);
-+
-+ if (!(--vdev->refcnt)) {
-+ vfio_fsl_mc_regions_cleanup(vdev);
-+ vfio_fsl_mc_irqs_cleanup(vdev);
-+ }
-+
-+ if (strcmp(mc_dev->obj_desc.type, "dprc") == 0)
-+ dprc_reset_container(mc_dev->mc_io, 0, mc_dev->mc_handle,
-+ mc_dev->obj_desc.id);
-+
-+ mutex_unlock(&driver_lock);
-+
-+ module_put(THIS_MODULE);
-+}
-+
-+static long vfio_fsl_mc_ioctl(void *device_data, unsigned int cmd,
-+ unsigned long arg)
-+{
-+ struct vfio_fsl_mc_device *vdev = device_data;
-+ struct fsl_mc_device *mc_dev = vdev->mc_dev;
-+ unsigned long minsz;
-+
-+ if (WARN_ON(!mc_dev))
-+ return -ENODEV;
-+
-+ switch (cmd) {
-+ case VFIO_DEVICE_GET_INFO:
-+ {
-+ struct vfio_device_info info;
-+
-+ minsz = offsetofend(struct vfio_device_info, num_irqs);
-+
-+ if (copy_from_user(&info, (void __user *)arg, minsz))
-+ return -EFAULT;
-+
-+ if (info.argsz < minsz)
-+ return -EINVAL;
-+
-+ info.flags = VFIO_DEVICE_FLAGS_FSL_MC;
-+ info.num_regions = mc_dev->obj_desc.region_count;
-+ info.num_irqs = mc_dev->obj_desc.irq_count;
-+
-+ return copy_to_user((void __user *)arg, &info, minsz);
-+ }
-+ case VFIO_DEVICE_GET_REGION_INFO:
-+ {
-+ struct vfio_region_info info;
-+
-+ minsz = offsetofend(struct vfio_region_info, offset);
-+
-+ if (copy_from_user(&info, (void __user *)arg, minsz))
-+ return -EFAULT;
-+
-+ if (info.argsz < minsz)
-+ return -EINVAL;
-+
-+ if (info.index >= vdev->num_regions)
-+ return -EINVAL;
-+
-+ /* map offset to the physical address */
-+ info.offset = VFIO_FSL_MC_INDEX_TO_OFFSET(info.index);
-+ info.size = vdev->regions[info.index].size;
-+ info.flags = vdev->regions[info.index].flags;
-+
-+ return copy_to_user((void __user *)arg, &info, minsz);
-+ }
-+ case VFIO_DEVICE_GET_IRQ_INFO:
-+ {
-+ struct vfio_irq_info info;
-+
-+ minsz = offsetofend(struct vfio_irq_info, count);
-+ if (copy_from_user(&info, (void __user *)arg, minsz))
-+ return -EFAULT;
-+
-+ if (info.argsz < minsz)
-+ return -EINVAL;
-+
-+ if (info.index >= mc_dev->obj_desc.irq_count)
-+ return -EINVAL;
-+
-+ if (vdev->mc_irqs != NULL) {
-+ info.flags = vdev->mc_irqs[info.index].flags;
-+ info.count = vdev->mc_irqs[info.index].count;
-+ } else {
-+ /*
-+ * If IRQs are not initialized then these can not
-+ * be configuted and used by user-space/
-+ */
-+ info.flags = 0;
-+ info.count = 0;
-+ }
-+
-+ return copy_to_user((void __user *)arg, &info, minsz);
-+ }
-+ case VFIO_DEVICE_SET_IRQS:
-+ {
-+ struct vfio_irq_set hdr;
-+ u8 *data = NULL;
-+ int ret = 0;
-+
-+ minsz = offsetofend(struct vfio_irq_set, count);
-+
-+ if (copy_from_user(&hdr, (void __user *)arg, minsz))
-+ return -EFAULT;
-+
-+ if (hdr.argsz < minsz)
-+ return -EINVAL;
-+
-+ if (hdr.index >= mc_dev->obj_desc.irq_count)
-+ return -EINVAL;
-+
-+ if (hdr.start != 0 || hdr.count > 1)
-+ return -EINVAL;
-+
-+ if (hdr.count == 0 &&
-+ (!(hdr.flags & VFIO_IRQ_SET_DATA_NONE) ||
-+ !(hdr.flags & VFIO_IRQ_SET_ACTION_TRIGGER)))
-+ return -EINVAL;
-+
-+ if (hdr.flags & ~(VFIO_IRQ_SET_DATA_TYPE_MASK |
-+ VFIO_IRQ_SET_ACTION_TYPE_MASK))
-+ return -EINVAL;
-+
-+ if (!(hdr.flags & VFIO_IRQ_SET_DATA_NONE)) {
-+ size_t size;
-+
-+ if (hdr.flags & VFIO_IRQ_SET_DATA_BOOL)
-+ size = sizeof(uint8_t);
-+ else if (hdr.flags & VFIO_IRQ_SET_DATA_EVENTFD)
-+ size = sizeof(int32_t);
-+ else
-+ return -EINVAL;
-+
-+ if (hdr.argsz - minsz < hdr.count * size)
-+ return -EINVAL;
-+
-+ data = memdup_user((void __user *)(arg + minsz),
-+ hdr.count * size);
-+ if (IS_ERR(data))
-+ return PTR_ERR(data);
-+ }
-+
-+ ret = vfio_fsl_mc_set_irqs_ioctl(vdev, hdr.flags,
-+ hdr.index, hdr.start,
-+ hdr.count, data);
-+ return ret;
-+ }
-+ case VFIO_DEVICE_RESET:
-+ {
-+ return -EINVAL;
-+ }
-+ default:
-+ return -EINVAL;
-+ }
-+}
-+
-+static ssize_t vfio_fsl_mc_read(void *device_data, char __user *buf,
-+ size_t count, loff_t *ppos)
-+{
-+ struct vfio_fsl_mc_device *vdev = device_data;
-+ unsigned int index = VFIO_FSL_MC_OFFSET_TO_INDEX(*ppos);
-+ loff_t off = *ppos & VFIO_FSL_MC_OFFSET_MASK;
-+ struct vfio_fsl_mc_region *region;
-+ uint64_t data[8];
-+ int i;
-+
-+ /* Read ioctl supported only for DPRC and DPMCP device */
-+ if (strcmp(vdev->mc_dev->obj_desc.type, "dprc") &&
-+ strcmp(vdev->mc_dev->obj_desc.type, "dpmcp"))
-+ return -EINVAL;
-+
-+ if (index >= vdev->num_regions)
-+ return -EINVAL;
-+
-+ region = &vdev->regions[index];
-+
-+ if (!(region->flags & VFIO_REGION_INFO_FLAG_READ))
-+ return -EINVAL;
-+
-+ if (!region->type & VFIO_FSL_MC_REGION_TYPE_MMIO)
-+ return -EINVAL;
-+
-+ if (!region->ioaddr) {
-+ region->ioaddr = ioremap_nocache(region->addr, region->size);
-+ if (!region->ioaddr)
-+ return -ENOMEM;
-+ }
-+
-+ if (count != 64 || off != 0)
-+ return -EINVAL;
-+
-+ for (i = 7; i >= 0; i--)
-+ data[i] = readq(region->ioaddr + i * sizeof(uint64_t));
-+
-+ if (copy_to_user(buf, data, 64))
-+ return -EFAULT;
-+
-+ return count;
-+}
-+
-+#define MC_CMD_COMPLETION_TIMEOUT_MS 5000
-+#define MC_CMD_COMPLETION_POLLING_MAX_SLEEP_USECS 500
-+
-+static int vfio_fsl_mc_dprc_wait_for_response(void __iomem *ioaddr)
-+{
-+ enum mc_cmd_status status;
-+ unsigned long timeout_usecs = MC_CMD_COMPLETION_TIMEOUT_MS * 1000;
-+
-+ for (;;) {
-+ u64 header;
-+ struct mc_cmd_header *resp_hdr;
-+
-+ __iormb();
-+ header = readq(ioaddr);
-+ __iormb();
-+
-+ resp_hdr = (struct mc_cmd_header *)&header;
-+ status = (enum mc_cmd_status)resp_hdr->status;
-+ if (status != MC_CMD_STATUS_READY)
-+ break;
-+
-+ udelay(MC_CMD_COMPLETION_POLLING_MAX_SLEEP_USECS);
-+ timeout_usecs -= MC_CMD_COMPLETION_POLLING_MAX_SLEEP_USECS;
-+ if (timeout_usecs == 0)
-+ return -ETIMEDOUT;
-+ }
-+
-+ return 0;
-+}
-+
-+static int vfio_fsl_mc_send_command(void __iomem *ioaddr, uint64_t *cmd_data)
-+{
-+ int i;
-+
-+ /* Write at command header in the end */
-+ for (i = 7; i >= 0; i--)
-+ writeq(cmd_data[i], ioaddr + i * sizeof(uint64_t));
-+
-+ /* Wait for response before returning to user-space
-+ * This can be optimized in future to even prepare response
-+ * before returning to user-space and avoid read ioctl.
-+ */
-+ return vfio_fsl_mc_dprc_wait_for_response(ioaddr);
-+}
-+
-+static int vfio_handle_dprc_commands(void __iomem *ioaddr, uint64_t *cmd_data)
-+{
-+ uint64_t cmd_hdr = cmd_data[0];
-+ int cmd = (cmd_hdr >> 52) & 0xfff;
-+
-+ switch (cmd) {
-+ case DPRC_CMDID_OPEN:
-+ default:
-+ return vfio_fsl_mc_send_command(ioaddr, cmd_data);
-+ }
-+
-+ return 0;
-+}
-+
-+static ssize_t vfio_fsl_mc_write(void *device_data, const char __user *buf,
-+ size_t count, loff_t *ppos)
-+{
-+ struct vfio_fsl_mc_device *vdev = device_data;
-+ unsigned int index = VFIO_FSL_MC_OFFSET_TO_INDEX(*ppos);
-+ loff_t off = *ppos & VFIO_FSL_MC_OFFSET_MASK;
-+ struct vfio_fsl_mc_region *region;
-+ uint64_t data[8];
-+ int ret;
-+
-+ /* Write ioctl supported only for DPRC and DPMCP device */
-+ if (strcmp(vdev->mc_dev->obj_desc.type, "dprc") &&
-+ strcmp(vdev->mc_dev->obj_desc.type, "dpmcp"))
-+ return -EINVAL;
-+
-+ if (index >= vdev->num_regions)
-+ return -EINVAL;
-+
-+ region = &vdev->regions[index];
-+
-+ if (!(region->flags & VFIO_REGION_INFO_FLAG_WRITE))
-+ return -EINVAL;
-+
-+ if (!region->type & VFIO_FSL_MC_REGION_TYPE_MMIO)
-+ return -EINVAL;
-+
-+ if (!region->ioaddr) {
-+ region->ioaddr = ioremap_nocache(region->addr, region->size);
-+ if (!region->ioaddr)
-+ return -ENOMEM;
-+ }
-+
-+ if (count != 64 || off != 0)
-+ return -EINVAL;
-+
-+ if (copy_from_user(&data, buf, 64))
-+ return -EFAULT;
-+
-+ ret = vfio_handle_dprc_commands(region->ioaddr, data);
-+ if (ret)
-+ return ret;
-+
-+ return count;
-+}
-+
-+static int vfio_fsl_mc_mmap_mmio(struct vfio_fsl_mc_region region,
-+ struct vm_area_struct *vma)
-+{
-+ u64 size = vma->vm_end - vma->vm_start;
-+ u64 pgoff, base;
-+
-+ pgoff = vma->vm_pgoff &
-+ ((1U << (VFIO_FSL_MC_OFFSET_SHIFT - PAGE_SHIFT)) - 1);
-+ base = pgoff << PAGE_SHIFT;
-+
-+ if (region.size < PAGE_SIZE || base + size > region.size)
-+ return -EINVAL;
-+ /*
-+ * Set the REGION_TYPE_CACHEABLE (QBman CENA regs) to be the
-+ * cache inhibited area of the portal to avoid coherency issues
-+ * if a user migrates to another core.
-+ */
-+ if (region.type & VFIO_FSL_MC_REGION_TYPE_CACHEABLE)
-+ vma->vm_page_prot = pgprot_cached_ns(vma->vm_page_prot);
-+ else
-+ vma->vm_page_prot = pgprot_noncached(vma->vm_page_prot);
-+
-+ vma->vm_pgoff = (region.addr >> PAGE_SHIFT) + pgoff;
-+
-+ return remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff,
-+ size, vma->vm_page_prot);
-+}
-+
-+/* Allows mmaping fsl_mc device regions in assigned DPRC */
-+static int vfio_fsl_mc_mmap(void *device_data, struct vm_area_struct *vma)
-+{
-+ struct vfio_fsl_mc_device *vdev = device_data;
-+ struct fsl_mc_device *mc_dev = vdev->mc_dev;
-+ unsigned long size, addr;
-+ int index;
-+
-+ index = vma->vm_pgoff >> (VFIO_FSL_MC_OFFSET_SHIFT - PAGE_SHIFT);
-+
-+ if (vma->vm_end < vma->vm_start)
-+ return -EINVAL;
-+ if (vma->vm_start & ~PAGE_MASK)
-+ return -EINVAL;
-+ if (vma->vm_end & ~PAGE_MASK)
-+ return -EINVAL;
-+ if (!(vma->vm_flags & VM_SHARED))
-+ return -EINVAL;
-+ if (index >= vdev->num_regions)
-+ return -EINVAL;
-+
-+ if (!(vdev->regions[index].flags & VFIO_REGION_INFO_FLAG_MMAP))
-+ return -EINVAL;
-+
-+ if (!(vdev->regions[index].flags & VFIO_REGION_INFO_FLAG_READ)
-+ && (vma->vm_flags & VM_READ))
-+ return -EINVAL;
-+
-+ if (!(vdev->regions[index].flags & VFIO_REGION_INFO_FLAG_WRITE)
-+ && (vma->vm_flags & VM_WRITE))
-+ return -EINVAL;
-+
-+ addr = vdev->regions[index].addr;
-+ size = vdev->regions[index].size;
-+
-+ vma->vm_private_data = mc_dev;
-+
-+ if (vdev->regions[index].type & VFIO_FSL_MC_REGION_TYPE_MMIO)
-+ return vfio_fsl_mc_mmap_mmio(vdev->regions[index], vma);
-+
-+ return -EFAULT;
-+}
-+
-+static const struct vfio_device_ops vfio_fsl_mc_ops = {
-+ .name = "vfio-fsl-mc",
-+ .open = vfio_fsl_mc_open,
-+ .release = vfio_fsl_mc_release,
-+ .ioctl = vfio_fsl_mc_ioctl,
-+ .read = vfio_fsl_mc_read,
-+ .write = vfio_fsl_mc_write,
-+ .mmap = vfio_fsl_mc_mmap,
-+};
-+
-+static int vfio_fsl_mc_initialize_dprc(struct vfio_fsl_mc_device *vdev)
-+{
-+ struct device *root_dprc_dev;
-+ struct fsl_mc_device *mc_dev = vdev->mc_dev;
-+ struct device *dev = &mc_dev->dev;
-+ struct fsl_mc_bus *mc_bus;
-+ struct irq_domain *mc_msi_domain;
-+ unsigned int irq_count;
-+ int ret;
-+
-+ /* device must be DPRC */
-+ if (strcmp(mc_dev->obj_desc.type, "dprc"))
-+ return -EINVAL;
-+
-+ /* mc_io must be un-initialized */
-+ WARN_ON(mc_dev->mc_io);
-+
-+ /* allocate a portal from the root DPRC for vfio use */
-+ fsl_mc_get_root_dprc(dev, &root_dprc_dev);
-+ if (WARN_ON(!root_dprc_dev))
-+ return -EINVAL;
-+
-+ ret = fsl_mc_portal_allocate(to_fsl_mc_device(root_dprc_dev),
-+ FSL_MC_IO_ATOMIC_CONTEXT_PORTAL,
-+ &mc_dev->mc_io);
-+ if (ret < 0)
-+ goto clean_msi_domain;
-+
-+ /* Reset MCP before move on */
-+ ret = fsl_mc_portal_reset(mc_dev->mc_io);
-+ if (ret < 0) {
-+ dev_err(dev, "dprc portal reset failed: error = %d\n", ret);
-+ goto free_mc_portal;
-+ }
-+
-+ /* MSI domain set up */
-+ ret = fsl_mc_find_msi_domain(root_dprc_dev->parent, &mc_msi_domain);
-+ if (ret < 0)
-+ goto free_mc_portal;
-+
-+ dev_set_msi_domain(&mc_dev->dev, mc_msi_domain);
-+
-+ ret = dprc_open(mc_dev->mc_io, 0, mc_dev->obj_desc.id,
-+ &mc_dev->mc_handle);
-+ if (ret) {
-+ dev_err(dev, "dprc_open() failed: error = %d\n", ret);
-+ goto free_mc_portal;
-+ }
-+
-+ /* Initialize resource pool */
-+ fsl_mc_init_all_resource_pools(mc_dev);
-+
-+ mc_bus = to_fsl_mc_bus(mc_dev);
-+
-+ if (!mc_bus->irq_resources) {
-+ irq_count = FSL_MC_IRQ_POOL_MAX_TOTAL_IRQS;
-+ ret = fsl_mc_populate_irq_pool(mc_bus, irq_count);
-+ if (ret < 0) {
-+ dev_err(dev, "%s: Failed to init irq-pool\n", __func__);
-+ goto clean_resource_pool;
-+ }
-+ }
-+
-+ mutex_init(&mc_bus->scan_mutex);
-+
-+ mutex_lock(&mc_bus->scan_mutex);
-+ ret = dprc_scan_objects(mc_dev, mc_dev->driver_override,
-+ &irq_count);
-+ mutex_unlock(&mc_bus->scan_mutex);
-+ if (ret) {
-+ dev_err(dev, "dprc_scan_objects() fails (%d)\n", ret);
-+ goto clean_irq_pool;
-+ }
-+
-+ if (irq_count > FSL_MC_IRQ_POOL_MAX_TOTAL_IRQS) {
-+ dev_warn(&mc_dev->dev,
-+ "IRQs needed (%u) exceed IRQs preallocated (%u)\n",
-+ irq_count, FSL_MC_IRQ_POOL_MAX_TOTAL_IRQS);
-+ }
-+
-+ return 0;
-+
-+clean_irq_pool:
-+ fsl_mc_cleanup_irq_pool(mc_bus);
-+
-+clean_resource_pool:
-+ fsl_mc_cleanup_all_resource_pools(mc_dev);
-+ dprc_close(mc_dev->mc_io, 0, mc_dev->mc_handle);
-+
-+free_mc_portal:
-+ fsl_mc_portal_free(mc_dev->mc_io);
-+
-+clean_msi_domain:
-+ dev_set_msi_domain(&mc_dev->dev, NULL);
-+
-+ return ret;
-+}
-+
-+static int vfio_fsl_mc_device_remove(struct device *dev, void *data)
-+{
-+ struct fsl_mc_device *mc_dev;
-+
-+ WARN_ON(dev == NULL);
-+
-+ mc_dev = to_fsl_mc_device(dev);
-+ if (WARN_ON(mc_dev == NULL))
-+ return -ENODEV;
-+
-+ fsl_mc_device_remove(mc_dev);
-+ return 0;
-+}
-+
-+static void vfio_fsl_mc_cleanup_dprc(struct vfio_fsl_mc_device *vdev)
-+{
-+ struct fsl_mc_device *mc_dev = vdev->mc_dev;
-+ struct fsl_mc_bus *mc_bus;
-+
-+ /* device must be DPRC */
-+ if (strcmp(mc_dev->obj_desc.type, "dprc"))
-+ return;
-+
-+ device_for_each_child(&mc_dev->dev, NULL, vfio_fsl_mc_device_remove);
-+
-+ mc_bus = to_fsl_mc_bus(mc_dev);
-+ if (dev_get_msi_domain(&mc_dev->dev))
-+ fsl_mc_cleanup_irq_pool(mc_bus);
-+
-+ dev_set_msi_domain(&mc_dev->dev, NULL);
-+
-+ fsl_mc_cleanup_all_resource_pools(mc_dev);
-+ dprc_close(mc_dev->mc_io, 0, mc_dev->mc_handle);
-+ fsl_mc_portal_free(mc_dev->mc_io);
-+}
-+
-+static int vfio_fsl_mc_probe(struct fsl_mc_device *mc_dev)
-+{
-+ struct iommu_group *group;
-+ struct vfio_fsl_mc_device *vdev;
-+ struct device *dev = &mc_dev->dev;
-+ int ret;
-+
-+ group = vfio_iommu_group_get(dev);
-+ if (!group) {
-+ dev_err(dev, "%s: VFIO: No IOMMU group\n", __func__);
-+ return -EINVAL;
-+ }
-+
-+ vdev = kzalloc(sizeof(*vdev), GFP_KERNEL);
-+ if (!vdev) {
-+ vfio_iommu_group_put(group, dev);
-+ return -ENOMEM;
-+ }
-+
-+ vdev->mc_dev = mc_dev;
-+
-+ ret = vfio_add_group_dev(dev, &vfio_fsl_mc_ops, vdev);
-+ if (ret) {
-+ dev_err(dev, "%s: Failed to add to vfio group\n", __func__);
-+ goto free_vfio_device;
-+ }
-+
-+ /* DPRC container scanned and it's chilren bound with vfio driver */
-+ if (strcmp(mc_dev->obj_desc.type, "dprc") == 0) {
-+ ret = vfio_fsl_mc_initialize_dprc(vdev);
-+ if (ret) {
-+ vfio_del_group_dev(dev);
-+ goto free_vfio_device;
-+ }
-+ } else {
-+ struct fsl_mc_device *mc_bus_dev;
-+
-+ /* Non-dprc devices share mc_io from the parent dprc */
-+ mc_bus_dev = to_fsl_mc_device(mc_dev->dev.parent);
-+ if (mc_bus_dev == NULL) {
-+ vfio_del_group_dev(dev);
-+ goto free_vfio_device;
-+ }
-+
-+ mc_dev->mc_io = mc_bus_dev->mc_io;
-+
-+ /* Inherit parent MSI domain */
-+ dev_set_msi_domain(&mc_dev->dev,
-+ dev_get_msi_domain(mc_dev->dev.parent));
-+ }
-+ return 0;
-+
-+free_vfio_device:
-+ kfree(vdev);
-+ vfio_iommu_group_put(group, dev);
-+ return ret;
-+}
-+
-+static int vfio_fsl_mc_remove(struct fsl_mc_device *mc_dev)
-+{
-+ struct vfio_fsl_mc_device *vdev;
-+ struct device *dev = &mc_dev->dev;
-+
-+ vdev = vfio_del_group_dev(dev);
-+ if (!vdev)
-+ return -EINVAL;
-+
-+ if (strcmp(mc_dev->obj_desc.type, "dprc") == 0)
-+ vfio_fsl_mc_cleanup_dprc(vdev);
-+ else
-+ dev_set_msi_domain(&mc_dev->dev, NULL);
-+
-+ mc_dev->mc_io = NULL;
-+
-+ vfio_iommu_group_put(mc_dev->dev.iommu_group, dev);
-+ kfree(vdev);
-+
-+ return 0;
-+}
-+
-+/*
-+ * vfio-fsl_mc is a meta-driver, so use driver_override interface to
-+ * bind a fsl_mc container with this driver and match_id_table is NULL.
-+ */
-+static struct fsl_mc_driver vfio_fsl_mc_driver = {
-+ .probe = vfio_fsl_mc_probe,
-+ .remove = vfio_fsl_mc_remove,
-+ .match_id_table = NULL,
-+ .driver = {
-+ .name = "vfio-fsl-mc",
-+ .owner = THIS_MODULE,
-+ },
-+};
-+
-+static int __init vfio_fsl_mc_driver_init(void)
-+{
-+ return fsl_mc_driver_register(&vfio_fsl_mc_driver);
-+}
-+
-+static void __exit vfio_fsl_mc_driver_exit(void)
-+{
-+ fsl_mc_driver_unregister(&vfio_fsl_mc_driver);
-+}
-+
-+module_init(vfio_fsl_mc_driver_init);
-+module_exit(vfio_fsl_mc_driver_exit);
-+
-+MODULE_VERSION(DRIVER_VERSION);
-+MODULE_LICENSE("GPL v2");
-+MODULE_AUTHOR(DRIVER_AUTHOR);
-+MODULE_DESCRIPTION(DRIVER_DESC);
---- /dev/null
-+++ b/drivers/vfio/fsl-mc/vfio_fsl_mc_intr.c
-@@ -0,0 +1,199 @@
-+/*
-+ * Freescale Management Complex (MC) device passthrough using VFIO
-+ *
-+ * Copyright (C) 2013-2016 Freescale Semiconductor, Inc.
-+ * Author: Bharat Bhushan <bharat.bhushan@nxp.com>
-+ *
-+ * This file is licensed under the terms of the GNU General Public
-+ * License version 2. This program is licensed "as is" without any
-+ * warranty of any kind, whether express or implied.
-+ */
-+
-+#include <linux/vfio.h>
-+#include <linux/slab.h>
-+#include <linux/types.h>
-+#include <linux/eventfd.h>
-+#include <linux/msi.h>
-+
-+#include "linux/fsl/mc.h"
-+#include "vfio_fsl_mc_private.h"
-+
-+static irqreturn_t vfio_fsl_mc_irq_handler(int irq_num, void *arg)
-+{
-+ struct vfio_fsl_mc_irq *mc_irq = (struct vfio_fsl_mc_irq *)arg;
-+
-+ eventfd_signal(mc_irq->trigger, 1);
-+ return IRQ_HANDLED;
-+}
-+
-+static int vfio_fsl_mc_irq_mask(struct vfio_fsl_mc_device *vdev,
-+ unsigned int index, unsigned int start,
-+ unsigned int count, uint32_t flags,
-+ void *data)
-+{
-+ return -EINVAL;
-+}
-+
-+static int vfio_fsl_mc_irq_unmask(struct vfio_fsl_mc_device *vdev,
-+ unsigned int index, unsigned int start,
-+ unsigned int count, uint32_t flags,
-+ void *data)
-+{
-+ return -EINVAL;
-+}
-+
-+static int vfio_set_trigger(struct vfio_fsl_mc_device *vdev,
-+ int index, int fd)
-+{
-+ struct vfio_fsl_mc_irq *irq = &vdev->mc_irqs[index];
-+ struct eventfd_ctx *trigger;
-+ int hwirq;
-+ int ret;
-+
-+ hwirq = vdev->mc_dev->irqs[index]->msi_desc->irq;
-+ if (irq->trigger) {
-+ free_irq(hwirq, irq);
-+ kfree(irq->name);
-+ eventfd_ctx_put(irq->trigger);
-+ irq->trigger = NULL;
-+ }
-+
-+ if (fd < 0) /* Disable only */
-+ return 0;
-+
-+ irq->name = kasprintf(GFP_KERNEL, "vfio-irq[%d](%s)",
-+ hwirq, dev_name(&vdev->mc_dev->dev));
-+ if (!irq->name)
-+ return -ENOMEM;
-+
-+ trigger = eventfd_ctx_fdget(fd);
-+ if (IS_ERR(trigger)) {
-+ kfree(irq->name);
-+ return PTR_ERR(trigger);
-+ }
-+
-+ irq->trigger = trigger;
-+
-+ ret = request_irq(hwirq, vfio_fsl_mc_irq_handler, 0,
-+ irq->name, irq);
-+ if (ret) {
-+ kfree(irq->name);
-+ eventfd_ctx_put(trigger);
-+ irq->trigger = NULL;
-+ return ret;
-+ }
-+
-+ return 0;
-+}
-+
-+int vfio_fsl_mc_irqs_init(struct vfio_fsl_mc_device *vdev)
-+{
-+ struct fsl_mc_device *mc_dev = vdev->mc_dev;
-+ struct vfio_fsl_mc_irq *mc_irq;
-+ int irq_count;
-+ int ret, i;
-+
-+ /* Device does not support any interrupt */
-+ if (mc_dev->obj_desc.irq_count == 0)
-+ return 0;
-+
-+ irq_count = mc_dev->obj_desc.irq_count;
-+
-+ mc_irq = kcalloc(irq_count, sizeof(*mc_irq), GFP_KERNEL);
-+ if (mc_irq == NULL)
-+ return -ENOMEM;
-+
-+ /* Allocate IRQs */
-+ ret = fsl_mc_allocate_irqs(mc_dev);
-+ if (ret) {
-+ kfree(mc_irq);
-+ return ret;
-+ }
-+
-+ for (i = 0; i < irq_count; i++) {
-+ mc_irq[i].count = 1;
-+ mc_irq[i].flags = VFIO_IRQ_INFO_EVENTFD;
-+ }
-+
-+ vdev->mc_irqs = mc_irq;
-+
-+ return 0;
-+}
-+
-+/* Free All IRQs for the given MC object */
-+void vfio_fsl_mc_irqs_cleanup(struct vfio_fsl_mc_device *vdev)
-+{
-+ struct fsl_mc_device *mc_dev = vdev->mc_dev;
-+ int irq_count = mc_dev->obj_desc.irq_count;
-+ int i;
-+
-+ /* Device does not support any interrupt */
-+ if (mc_dev->obj_desc.irq_count == 0)
-+ return;
-+
-+ for (i = 0; i < irq_count; i++)
-+ vfio_set_trigger(vdev, i, -1);
-+
-+ fsl_mc_free_irqs(mc_dev);
-+ kfree(vdev->mc_irqs);
-+}
-+
-+static int vfio_fsl_mc_set_irq_trigger(struct vfio_fsl_mc_device *vdev,
-+ unsigned int index, unsigned int start,
-+ unsigned int count, uint32_t flags,
-+ void *data)
-+{
-+ struct vfio_fsl_mc_irq *irq = &vdev->mc_irqs[index];
-+ int hwirq;
-+
-+ if (!count && (flags & VFIO_IRQ_SET_DATA_NONE))
-+ return vfio_set_trigger(vdev, index, -1);
-+
-+ if (start != 0 || count != 1)
-+ return -EINVAL;
-+
-+ if (flags & VFIO_IRQ_SET_DATA_EVENTFD) {
-+ int32_t fd = *(int32_t *)data;
-+
-+ return vfio_set_trigger(vdev, index, fd);
-+ }
-+
-+ hwirq = vdev->mc_dev->irqs[index]->msi_desc->irq;
-+
-+ if (flags & VFIO_IRQ_SET_DATA_NONE) {
-+ vfio_fsl_mc_irq_handler(hwirq, irq);
-+
-+ } else if (flags & VFIO_IRQ_SET_DATA_BOOL) {
-+ uint8_t trigger = *(uint8_t *)data;
-+
-+ if (trigger)
-+ vfio_fsl_mc_irq_handler(hwirq, irq);
-+ }
-+
-+ return 0;
-+}
-+
-+int vfio_fsl_mc_set_irqs_ioctl(struct vfio_fsl_mc_device *vdev,
-+ uint32_t flags, unsigned int index,
-+ unsigned int start, unsigned int count,
-+ void *data)
-+{
-+ int ret = -ENOTTY;
-+
-+ switch (flags & VFIO_IRQ_SET_ACTION_TYPE_MASK) {
-+ case VFIO_IRQ_SET_ACTION_MASK:
-+ ret = vfio_fsl_mc_irq_mask(vdev, index, start, count,
-+ flags, data);
-+ break;
-+ case VFIO_IRQ_SET_ACTION_UNMASK:
-+ ret = vfio_fsl_mc_irq_unmask(vdev, index, start, count,
-+ flags, data);
-+ break;
-+ case VFIO_IRQ_SET_ACTION_TRIGGER:
-+ ret = vfio_fsl_mc_set_irq_trigger(vdev, index, start,
-+ count, flags, data);
-+ break;
-+ }
-+
-+ return ret;
-+}
---- /dev/null
-+++ b/drivers/vfio/fsl-mc/vfio_fsl_mc_private.h
-@@ -0,0 +1,55 @@
-+/*
-+ * Freescale Management Complex VFIO private declarations
-+ *
-+ * Copyright (C) 2013-2016 Freescale Semiconductor, Inc.
-+ * Copyright 2016 NXP
-+ * Author: Bharat Bhushan <bharat.bhushan@nxp.com>
-+ *
-+ * This file is licensed under the terms of the GNU General Public
-+ * License version 2. This program is licensed "as is" without any
-+ * warranty of any kind, whether express or implied.
-+ */
-+
-+#ifndef VFIO_FSL_MC_PRIVATE_H
-+#define VFIO_FSL_MC_PRIVATE_H
-+
-+#define VFIO_FSL_MC_OFFSET_SHIFT 40
-+#define VFIO_FSL_MC_OFFSET_MASK (((u64)(1) << VFIO_FSL_MC_OFFSET_SHIFT) - 1)
-+
-+#define VFIO_FSL_MC_OFFSET_TO_INDEX(off) (off >> VFIO_FSL_MC_OFFSET_SHIFT)
-+
-+#define VFIO_FSL_MC_INDEX_TO_OFFSET(index) \
-+ ((u64)(index) << VFIO_FSL_MC_OFFSET_SHIFT)
-+
-+struct vfio_fsl_mc_irq {
-+ u32 flags;
-+ u32 count;
-+ struct eventfd_ctx *trigger;
-+ char *name;
-+};
-+
-+struct vfio_fsl_mc_region {
-+ u32 flags;
-+#define VFIO_FSL_MC_REGION_TYPE_MMIO 1
-+#define VFIO_FSL_MC_REGION_TYPE_CACHEABLE 2
-+ u32 type;
-+ u64 addr;
-+ resource_size_t size;
-+ void __iomem *ioaddr;
-+};
-+
-+struct vfio_fsl_mc_device {
-+ struct fsl_mc_device *mc_dev;
-+ int refcnt;
-+ u32 num_regions;
-+ struct vfio_fsl_mc_region *regions;
-+ struct vfio_fsl_mc_irq *mc_irqs;
-+};
-+
-+int vfio_fsl_mc_irqs_init(struct vfio_fsl_mc_device *vdev);
-+void vfio_fsl_mc_irqs_cleanup(struct vfio_fsl_mc_device *vdev);
-+int vfio_fsl_mc_set_irqs_ioctl(struct vfio_fsl_mc_device *vdev,
-+ uint32_t flags, unsigned int index,
-+ unsigned int start, unsigned int count,
-+ void *data);
-+#endif /* VFIO_PCI_PRIVATE_H */
---- a/drivers/vfio/vfio_iommu_type1.c
-+++ b/drivers/vfio/vfio_iommu_type1.c
-@@ -36,6 +36,8 @@
- #include <linux/uaccess.h>
- #include <linux/vfio.h>
- #include <linux/workqueue.h>
-+#include <linux/dma-iommu.h>
-+#include <linux/irqdomain.h>
-
- #define DRIVER_VERSION "0.2"
- #define DRIVER_AUTHOR "Alex Williamson <alex.williamson@redhat.com>"
-@@ -720,6 +722,27 @@ static void vfio_test_domain_fgsp(struct
- __free_pages(pages, order);
- }
-
-+static bool vfio_iommu_has_sw_msi(struct iommu_group *group, phys_addr_t *base)
-+{
-+ struct list_head group_resv_regions;
-+ struct iommu_resv_region *region, *next;
-+ bool ret = false;
-+
-+ INIT_LIST_HEAD(&group_resv_regions);
-+ iommu_get_group_resv_regions(group, &group_resv_regions);
-+ list_for_each_entry(region, &group_resv_regions, list) {
-+ if (region->type == IOMMU_RESV_SW_MSI) {
-+ *base = region->start;
-+ ret = true;
-+ goto out;
-+ }
-+ }
-+out:
-+ list_for_each_entry_safe(region, next, &group_resv_regions, list)
-+ kfree(region);
-+ return ret;
-+}
-+
- static int vfio_iommu_type1_attach_group(void *iommu_data,
- struct iommu_group *iommu_group)
- {
-@@ -728,6 +751,8 @@ static int vfio_iommu_type1_attach_group
- struct vfio_domain *domain, *d;
- struct bus_type *bus = NULL;
- int ret;
-+ bool resv_msi, msi_remap;
-+ phys_addr_t resv_msi_base;
-
- mutex_lock(&iommu->lock);
-
-@@ -774,11 +799,15 @@ static int vfio_iommu_type1_attach_group
- if (ret)
- goto out_domain;
-
-+ resv_msi = vfio_iommu_has_sw_msi(iommu_group, &resv_msi_base);
-+
- INIT_LIST_HEAD(&domain->group_list);
- list_add(&group->next, &domain->group_list);
-
-- if (!allow_unsafe_interrupts &&
-- !iommu_capable(bus, IOMMU_CAP_INTR_REMAP)) {
-+ msi_remap = resv_msi ? irq_domain_check_msi_remap() :
-+ iommu_capable(bus, IOMMU_CAP_INTR_REMAP);
-+
-+ if (!allow_unsafe_interrupts && !msi_remap) {
- pr_warn("%s: No interrupt remapping support. Use the module param \"allow_unsafe_interrupts\" to enable VFIO IOMMU support on this platform\n",
- __func__);
- ret = -EPERM;
-@@ -820,6 +849,12 @@ static int vfio_iommu_type1_attach_group
- if (ret)
- goto out_detach;
-
-+ if (resv_msi) {
-+ ret = iommu_get_msi_cookie(domain->domain, resv_msi_base);
-+ if (ret)
-+ goto out_detach;
-+ }
-+
- list_add(&domain->next, &iommu->domain_list);
-
- mutex_unlock(&iommu->lock);
---- a/include/uapi/linux/vfio.h
-+++ b/include/uapi/linux/vfio.h
-@@ -198,6 +198,7 @@ struct vfio_device_info {
- #define VFIO_DEVICE_FLAGS_PCI (1 << 1) /* vfio-pci device */
- #define VFIO_DEVICE_FLAGS_PLATFORM (1 << 2) /* vfio-platform device */
- #define VFIO_DEVICE_FLAGS_AMBA (1 << 3) /* vfio-amba device */
-+#define VFIO_DEVICE_FLAGS_FSL_MC (1 << 5) /* vfio-fsl-mc device */
- __u32 num_regions; /* Max region index + 1 */
- __u32 num_irqs; /* Max IRQ index + 1 */
- };
diff --git a/target/linux/layerscape/patches-4.9/819-flexcan-support-layerscape.patch b/target/linux/layerscape/patches-4.9/819-flexcan-support-layerscape.patch
deleted file mode 100644
index 5b0b3fd88e..0000000000
--- a/target/linux/layerscape/patches-4.9/819-flexcan-support-layerscape.patch
+++ /dev/null
@@ -1,542 +0,0 @@
-From 2887442bd13bc8be687afc7172cb01c2b7f0dd3b Mon Sep 17 00:00:00 2001
-From: Yangbo Lu <yangbo.lu@nxp.com>
-Date: Thu, 5 Jul 2018 17:41:14 +0800
-Subject: [PATCH 31/32] flexcan: support layerscape
-
-This is an integrated patch for layerscape flexcan support.
-
-Signed-off-by: Pankaj Bansal <pankaj.bansal@nxp.com>
-Signed-off-by: Bhupesh Sharma <bhupesh.sharma@freescale.com>
-Signed-off-by: Sakar Arora <Sakar.Arora@freescale.com>
-Signed-off-by: Yangbo Lu <yangbo.lu@nxp.com>
----
- drivers/net/can/flexcan.c | 212 ++++++++++++++++++++++----------------
- 1 file changed, 123 insertions(+), 89 deletions(-)
-
---- a/drivers/net/can/flexcan.c
-+++ b/drivers/net/can/flexcan.c
-@@ -184,6 +184,7 @@
- * MX53 FlexCAN2 03.00.00.00 yes no no no
- * MX6s FlexCAN3 10.00.12.00 yes yes no yes
- * VF610 FlexCAN3 ? no yes yes yes?
-+ * LS1021A FlexCAN2 03.00.04.00 no yes no yes
- *
- * Some SOCs do not have the RX_WARN & TX_WARN interrupt line connected.
- */
-@@ -260,6 +261,10 @@ struct flexcan_priv {
- struct flexcan_platform_data *pdata;
- const struct flexcan_devtype_data *devtype_data;
- struct regulator *reg_xceiver;
-+
-+ /* Read and Write APIs */
-+ u32 (*read)(void __iomem *addr);
-+ void (*write)(u32 val, void __iomem *addr);
- };
-
- static struct flexcan_devtype_data fsl_p1010_devtype_data = {
-@@ -276,6 +281,10 @@ static struct flexcan_devtype_data fsl_v
- .quirks = FLEXCAN_QUIRK_DISABLE_RXFG | FLEXCAN_QUIRK_DISABLE_MECR,
- };
-
-+static const struct flexcan_devtype_data fsl_ls1021a_r2_devtype_data = {
-+ .quirks = FLEXCAN_QUIRK_DISABLE_RXFG | FLEXCAN_QUIRK_DISABLE_MECR,
-+};
-+
- static const struct can_bittiming_const flexcan_bittiming_const = {
- .name = DRV_NAME,
- .tseg1_min = 4,
-@@ -288,32 +297,38 @@ static const struct can_bittiming_const
- .brp_inc = 1,
- };
-
--/* Abstract off the read/write for arm versus ppc. This
-- * assumes that PPC uses big-endian registers and everything
-- * else uses little-endian registers, independent of CPU
-- * endianness.
-+/* FlexCAN module is essentially modelled as a little-endian IP in most
-+ * SoCs, i.e the registers as well as the message buffer areas are
-+ * implemented in a little-endian fashion.
-+ *
-+ * However there are some SoCs (e.g. LS1021A) which implement the FlexCAN
-+ * module in a big-endian fashion (i.e the registers as well as the
-+ * message buffer areas are implemented in a big-endian way).
-+ *
-+ * In addition, the FlexCAN module can be found on SoCs having ARM or
-+ * PPC cores. So, we need to abstract off the register read/write
-+ * functions, ensuring that these cater to all the combinations of module
-+ * endianness and underlying CPU endianness.
- */
--#if defined(CONFIG_PPC)
--static inline u32 flexcan_read(void __iomem *addr)
-+static inline u32 flexcan_read_be(void __iomem *addr)
- {
-- return in_be32(addr);
-+ return ioread32be(addr);
- }
-
--static inline void flexcan_write(u32 val, void __iomem *addr)
-+static inline void flexcan_write_be(u32 val, void __iomem *addr)
- {
-- out_be32(addr, val);
-+ iowrite32be(val, addr);
- }
--#else
--static inline u32 flexcan_read(void __iomem *addr)
-+
-+static inline u32 flexcan_read_le(void __iomem *addr)
- {
-- return readl(addr);
-+ return ioread32(addr);
- }
-
--static inline void flexcan_write(u32 val, void __iomem *addr)
-+static inline void flexcan_write_le(u32 val, void __iomem *addr)
- {
-- writel(val, addr);
-+ iowrite32(val, addr);
- }
--#endif
-
- static inline int flexcan_transceiver_enable(const struct flexcan_priv *priv)
- {
-@@ -344,14 +359,14 @@ static int flexcan_chip_enable(struct fl
- unsigned int timeout = FLEXCAN_TIMEOUT_US / 10;
- u32 reg;
-
-- reg = flexcan_read(&regs->mcr);
-+ reg = priv->read(&regs->mcr);
- reg &= ~FLEXCAN_MCR_MDIS;
-- flexcan_write(reg, &regs->mcr);
-+ priv->write(reg, &regs->mcr);
-
-- while (timeout-- && (flexcan_read(&regs->mcr) & FLEXCAN_MCR_LPM_ACK))
-+ while (timeout-- && (priv->read(&regs->mcr) & FLEXCAN_MCR_LPM_ACK))
- udelay(10);
-
-- if (flexcan_read(&regs->mcr) & FLEXCAN_MCR_LPM_ACK)
-+ if (priv->read(&regs->mcr) & FLEXCAN_MCR_LPM_ACK)
- return -ETIMEDOUT;
-
- return 0;
-@@ -363,14 +378,14 @@ static int flexcan_chip_disable(struct f
- unsigned int timeout = FLEXCAN_TIMEOUT_US / 10;
- u32 reg;
-
-- reg = flexcan_read(&regs->mcr);
-+ reg = priv->read(&regs->mcr);
- reg |= FLEXCAN_MCR_MDIS;
-- flexcan_write(reg, &regs->mcr);
-+ priv->write(reg, &regs->mcr);
-
-- while (timeout-- && !(flexcan_read(&regs->mcr) & FLEXCAN_MCR_LPM_ACK))
-+ while (timeout-- && !(priv->read(&regs->mcr) & FLEXCAN_MCR_LPM_ACK))
- udelay(10);
-
-- if (!(flexcan_read(&regs->mcr) & FLEXCAN_MCR_LPM_ACK))
-+ if (!(priv->read(&regs->mcr) & FLEXCAN_MCR_LPM_ACK))
- return -ETIMEDOUT;
-
- return 0;
-@@ -382,14 +397,14 @@ static int flexcan_chip_freeze(struct fl
- unsigned int timeout = 1000 * 1000 * 10 / priv->can.bittiming.bitrate;
- u32 reg;
-
-- reg = flexcan_read(&regs->mcr);
-+ reg = priv->read(&regs->mcr);
- reg |= FLEXCAN_MCR_HALT;
-- flexcan_write(reg, &regs->mcr);
-+ priv->write(reg, &regs->mcr);
-
-- while (timeout-- && !(flexcan_read(&regs->mcr) & FLEXCAN_MCR_FRZ_ACK))
-+ while (timeout-- && !(priv->read(&regs->mcr) & FLEXCAN_MCR_FRZ_ACK))
- udelay(100);
-
-- if (!(flexcan_read(&regs->mcr) & FLEXCAN_MCR_FRZ_ACK))
-+ if (!(priv->read(&regs->mcr) & FLEXCAN_MCR_FRZ_ACK))
- return -ETIMEDOUT;
-
- return 0;
-@@ -401,14 +416,14 @@ static int flexcan_chip_unfreeze(struct
- unsigned int timeout = FLEXCAN_TIMEOUT_US / 10;
- u32 reg;
-
-- reg = flexcan_read(&regs->mcr);
-+ reg = priv->read(&regs->mcr);
- reg &= ~FLEXCAN_MCR_HALT;
-- flexcan_write(reg, &regs->mcr);
-+ priv->write(reg, &regs->mcr);
-
-- while (timeout-- && (flexcan_read(&regs->mcr) & FLEXCAN_MCR_FRZ_ACK))
-+ while (timeout-- && (priv->read(&regs->mcr) & FLEXCAN_MCR_FRZ_ACK))
- udelay(10);
-
-- if (flexcan_read(&regs->mcr) & FLEXCAN_MCR_FRZ_ACK)
-+ if (priv->read(&regs->mcr) & FLEXCAN_MCR_FRZ_ACK)
- return -ETIMEDOUT;
-
- return 0;
-@@ -419,11 +434,11 @@ static int flexcan_chip_softreset(struct
- struct flexcan_regs __iomem *regs = priv->regs;
- unsigned int timeout = FLEXCAN_TIMEOUT_US / 10;
-
-- flexcan_write(FLEXCAN_MCR_SOFTRST, &regs->mcr);
-- while (timeout-- && (flexcan_read(&regs->mcr) & FLEXCAN_MCR_SOFTRST))
-+ priv->write(FLEXCAN_MCR_SOFTRST, &regs->mcr);
-+ while (timeout-- && (priv->read(&regs->mcr) & FLEXCAN_MCR_SOFTRST))
- udelay(10);
-
-- if (flexcan_read(&regs->mcr) & FLEXCAN_MCR_SOFTRST)
-+ if (priv->read(&regs->mcr) & FLEXCAN_MCR_SOFTRST)
- return -ETIMEDOUT;
-
- return 0;
-@@ -434,7 +449,7 @@ static int __flexcan_get_berr_counter(co
- {
- const struct flexcan_priv *priv = netdev_priv(dev);
- struct flexcan_regs __iomem *regs = priv->regs;
-- u32 reg = flexcan_read(&regs->ecr);
-+ u32 reg = priv->read(&regs->ecr);
-
- bec->txerr = (reg >> 0) & 0xff;
- bec->rxerr = (reg >> 8) & 0xff;
-@@ -491,24 +506,24 @@ static int flexcan_start_xmit(struct sk_
-
- if (cf->can_dlc > 0) {
- data = be32_to_cpup((__be32 *)&cf->data[0]);
-- flexcan_write(data, &regs->mb[FLEXCAN_TX_BUF_ID].data[0]);
-+ priv->write(data, &regs->mb[FLEXCAN_TX_BUF_ID].data[0]);
- }
- if (cf->can_dlc > 4) {
- data = be32_to_cpup((__be32 *)&cf->data[4]);
-- flexcan_write(data, &regs->mb[FLEXCAN_TX_BUF_ID].data[1]);
-+ priv->write(data, &regs->mb[FLEXCAN_TX_BUF_ID].data[1]);
- }
-
- can_put_echo_skb(skb, dev, 0);
-
-- flexcan_write(can_id, &regs->mb[FLEXCAN_TX_BUF_ID].can_id);
-- flexcan_write(ctrl, &regs->mb[FLEXCAN_TX_BUF_ID].can_ctrl);
-+ priv->write(can_id, &regs->mb[FLEXCAN_TX_BUF_ID].can_id);
-+ priv->write(ctrl, &regs->mb[FLEXCAN_TX_BUF_ID].can_ctrl);
-
- /* Errata ERR005829 step8:
- * Write twice INACTIVE(0x8) code to first MB.
- */
-- flexcan_write(FLEXCAN_MB_CODE_TX_INACTIVE,
-+ priv->write(FLEXCAN_MB_CODE_TX_INACTIVE,
- &regs->mb[FLEXCAN_TX_BUF_RESERVED].can_ctrl);
-- flexcan_write(FLEXCAN_MB_CODE_TX_INACTIVE,
-+ priv->write(FLEXCAN_MB_CODE_TX_INACTIVE,
- &regs->mb[FLEXCAN_TX_BUF_RESERVED].can_ctrl);
-
- return NETDEV_TX_OK;
-@@ -632,8 +647,8 @@ static void flexcan_read_fifo(const stru
- struct flexcan_mb __iomem *mb = &regs->mb[0];
- u32 reg_ctrl, reg_id;
-
-- reg_ctrl = flexcan_read(&mb->can_ctrl);
-- reg_id = flexcan_read(&mb->can_id);
-+ reg_ctrl = priv->read(&mb->can_ctrl);
-+ reg_id = priv->read(&mb->can_id);
- if (reg_ctrl & FLEXCAN_MB_CNT_IDE)
- cf->can_id = ((reg_id >> 0) & CAN_EFF_MASK) | CAN_EFF_FLAG;
- else
-@@ -643,12 +658,12 @@ static void flexcan_read_fifo(const stru
- cf->can_id |= CAN_RTR_FLAG;
- cf->can_dlc = get_can_dlc((reg_ctrl >> 16) & 0xf);
-
-- *(__be32 *)(cf->data + 0) = cpu_to_be32(flexcan_read(&mb->data[0]));
-- *(__be32 *)(cf->data + 4) = cpu_to_be32(flexcan_read(&mb->data[1]));
-+ *(__be32 *)(cf->data + 0) = cpu_to_be32(priv->read(&mb->data[0]));
-+ *(__be32 *)(cf->data + 4) = cpu_to_be32(priv->read(&mb->data[1]));
-
- /* mark as read */
-- flexcan_write(FLEXCAN_IFLAG_RX_FIFO_AVAILABLE, &regs->iflag1);
-- flexcan_read(&regs->timer);
-+ priv->write(FLEXCAN_IFLAG_RX_FIFO_AVAILABLE, &regs->iflag1);
-+ priv->read(&regs->timer);
- }
-
- static int flexcan_read_frame(struct net_device *dev)
-@@ -685,17 +700,17 @@ static int flexcan_poll(struct napi_stru
- /* The error bits are cleared on read,
- * use saved value from irq handler.
- */
-- reg_esr = flexcan_read(&regs->esr) | priv->reg_esr;
-+ reg_esr = priv->read(&regs->esr) | priv->reg_esr;
-
- /* handle state changes */
- work_done += flexcan_poll_state(dev, reg_esr);
-
- /* handle RX-FIFO */
-- reg_iflag1 = flexcan_read(&regs->iflag1);
-+ reg_iflag1 = priv->read(&regs->iflag1);
- while (reg_iflag1 & FLEXCAN_IFLAG_RX_FIFO_AVAILABLE &&
- work_done < quota) {
- work_done += flexcan_read_frame(dev);
-- reg_iflag1 = flexcan_read(&regs->iflag1);
-+ reg_iflag1 = priv->read(&regs->iflag1);
- }
-
- /* report bus errors */
-@@ -705,8 +720,8 @@ static int flexcan_poll(struct napi_stru
- if (work_done < quota) {
- napi_complete_done(napi, work_done);
- /* enable IRQs */
-- flexcan_write(FLEXCAN_IFLAG_DEFAULT, &regs->imask1);
-- flexcan_write(priv->reg_ctrl_default, &regs->ctrl);
-+ priv->write(FLEXCAN_IFLAG_DEFAULT, &regs->imask1);
-+ priv->write(priv->reg_ctrl_default, &regs->ctrl);
- }
-
- return work_done;
-@@ -720,12 +735,12 @@ static irqreturn_t flexcan_irq(int irq,
- struct flexcan_regs __iomem *regs = priv->regs;
- u32 reg_iflag1, reg_esr;
-
-- reg_iflag1 = flexcan_read(&regs->iflag1);
-- reg_esr = flexcan_read(&regs->esr);
-+ reg_iflag1 = priv->read(&regs->iflag1);
-+ reg_esr = priv->read(&regs->esr);
-
- /* ACK all bus error and state change IRQ sources */
- if (reg_esr & FLEXCAN_ESR_ALL_INT)
-- flexcan_write(reg_esr & FLEXCAN_ESR_ALL_INT, &regs->esr);
-+ priv->write(reg_esr & FLEXCAN_ESR_ALL_INT, &regs->esr);
-
- /* schedule NAPI in case of:
- * - rx IRQ
-@@ -739,16 +754,16 @@ static irqreturn_t flexcan_irq(int irq,
- * save them for later use.
- */
- priv->reg_esr = reg_esr & FLEXCAN_ESR_ERR_BUS;
-- flexcan_write(FLEXCAN_IFLAG_DEFAULT &
-+ priv->write(FLEXCAN_IFLAG_DEFAULT &
- ~FLEXCAN_IFLAG_RX_FIFO_AVAILABLE, &regs->imask1);
-- flexcan_write(priv->reg_ctrl_default & ~FLEXCAN_CTRL_ERR_ALL,
-+ priv->write(priv->reg_ctrl_default & ~FLEXCAN_CTRL_ERR_ALL,
- &regs->ctrl);
- napi_schedule(&priv->napi);
- }
-
- /* FIFO overflow */
- if (reg_iflag1 & FLEXCAN_IFLAG_RX_FIFO_OVERFLOW) {
-- flexcan_write(FLEXCAN_IFLAG_RX_FIFO_OVERFLOW, &regs->iflag1);
-+ priv->write(FLEXCAN_IFLAG_RX_FIFO_OVERFLOW, &regs->iflag1);
- dev->stats.rx_over_errors++;
- dev->stats.rx_errors++;
- }
-@@ -760,9 +775,9 @@ static irqreturn_t flexcan_irq(int irq,
- can_led_event(dev, CAN_LED_EVENT_TX);
-
- /* after sending a RTR frame MB is in RX mode */
-- flexcan_write(FLEXCAN_MB_CODE_TX_INACTIVE,
-+ priv->write(FLEXCAN_MB_CODE_TX_INACTIVE,
- &regs->mb[FLEXCAN_TX_BUF_ID].can_ctrl);
-- flexcan_write((1 << FLEXCAN_TX_BUF_ID), &regs->iflag1);
-+ priv->write((1 << FLEXCAN_TX_BUF_ID), &regs->iflag1);
- netif_wake_queue(dev);
- }
-
-@@ -776,7 +791,7 @@ static void flexcan_set_bittiming(struct
- struct flexcan_regs __iomem *regs = priv->regs;
- u32 reg;
-
-- reg = flexcan_read(&regs->ctrl);
-+ reg = priv->read(&regs->ctrl);
- reg &= ~(FLEXCAN_CTRL_PRESDIV(0xff) |
- FLEXCAN_CTRL_RJW(0x3) |
- FLEXCAN_CTRL_PSEG1(0x7) |
-@@ -800,11 +815,11 @@ static void flexcan_set_bittiming(struct
- reg |= FLEXCAN_CTRL_SMP;
-
- netdev_dbg(dev, "writing ctrl=0x%08x\n", reg);
-- flexcan_write(reg, &regs->ctrl);
-+ priv->write(reg, &regs->ctrl);
-
- /* print chip status */
- netdev_dbg(dev, "%s: mcr=0x%08x ctrl=0x%08x\n", __func__,
-- flexcan_read(&regs->mcr), flexcan_read(&regs->ctrl));
-+ priv->read(&regs->mcr), priv->read(&regs->ctrl));
- }
-
- /* flexcan_chip_start
-@@ -842,13 +857,13 @@ static int flexcan_chip_start(struct net
- * choose format C
- * set max mailbox number
- */
-- reg_mcr = flexcan_read(&regs->mcr);
-+ reg_mcr = priv->read(&regs->mcr);
- reg_mcr &= ~FLEXCAN_MCR_MAXMB(0xff);
- reg_mcr |= FLEXCAN_MCR_FRZ | FLEXCAN_MCR_FEN | FLEXCAN_MCR_HALT |
- FLEXCAN_MCR_SUPV | FLEXCAN_MCR_WRN_EN | FLEXCAN_MCR_SRX_DIS |
- FLEXCAN_MCR_IDAM_C | FLEXCAN_MCR_MAXMB(FLEXCAN_TX_BUF_ID);
- netdev_dbg(dev, "%s: writing mcr=0x%08x", __func__, reg_mcr);
-- flexcan_write(reg_mcr, &regs->mcr);
-+ priv->write(reg_mcr, &regs->mcr);
-
- /* CTRL
- *
-@@ -861,7 +876,7 @@ static int flexcan_chip_start(struct net
- * enable bus off interrupt
- * (== FLEXCAN_CTRL_ERR_STATE)
- */
-- reg_ctrl = flexcan_read(&regs->ctrl);
-+ reg_ctrl = priv->read(&regs->ctrl);
- reg_ctrl &= ~FLEXCAN_CTRL_TSYN;
- reg_ctrl |= FLEXCAN_CTRL_BOFF_REC | FLEXCAN_CTRL_LBUF |
- FLEXCAN_CTRL_ERR_STATE;
-@@ -881,29 +896,29 @@ static int flexcan_chip_start(struct net
- /* leave interrupts disabled for now */
- reg_ctrl &= ~FLEXCAN_CTRL_ERR_ALL;
- netdev_dbg(dev, "%s: writing ctrl=0x%08x", __func__, reg_ctrl);
-- flexcan_write(reg_ctrl, &regs->ctrl);
-+ priv->write(reg_ctrl, &regs->ctrl);
-
- /* clear and invalidate all mailboxes first */
- for (i = FLEXCAN_TX_BUF_ID; i < ARRAY_SIZE(regs->mb); i++) {
-- flexcan_write(FLEXCAN_MB_CODE_RX_INACTIVE,
-+ priv->write(FLEXCAN_MB_CODE_RX_INACTIVE,
- &regs->mb[i].can_ctrl);
- }
-
- /* Errata ERR005829: mark first TX mailbox as INACTIVE */
-- flexcan_write(FLEXCAN_MB_CODE_TX_INACTIVE,
-+ priv->write(FLEXCAN_MB_CODE_TX_INACTIVE,
- &regs->mb[FLEXCAN_TX_BUF_RESERVED].can_ctrl);
-
- /* mark TX mailbox as INACTIVE */
-- flexcan_write(FLEXCAN_MB_CODE_TX_INACTIVE,
-+ priv->write(FLEXCAN_MB_CODE_TX_INACTIVE,
- &regs->mb[FLEXCAN_TX_BUF_ID].can_ctrl);
-
- /* acceptance mask/acceptance code (accept everything) */
-- flexcan_write(0x0, &regs->rxgmask);
-- flexcan_write(0x0, &regs->rx14mask);
-- flexcan_write(0x0, &regs->rx15mask);
-+ priv->write(0x0, &regs->rxgmask);
-+ priv->write(0x0, &regs->rx14mask);
-+ priv->write(0x0, &regs->rx15mask);
-
- if (priv->devtype_data->quirks & FLEXCAN_QUIRK_DISABLE_RXFG)
-- flexcan_write(0x0, &regs->rxfgmask);
-+ priv->write(0x0, &regs->rxfgmask);
-
- /* On Vybrid, disable memory error detection interrupts
- * and freeze mode.
-@@ -916,16 +931,16 @@ static int flexcan_chip_start(struct net
- * and Correction of Memory Errors" to write to
- * MECR register
- */
-- reg_ctrl2 = flexcan_read(&regs->ctrl2);
-+ reg_ctrl2 = priv->read(&regs->ctrl2);
- reg_ctrl2 |= FLEXCAN_CTRL2_ECRWRE;
-- flexcan_write(reg_ctrl2, &regs->ctrl2);
-+ priv->write(reg_ctrl2, &regs->ctrl2);
-
-- reg_mecr = flexcan_read(&regs->mecr);
-+ reg_mecr = priv->read(&regs->mecr);
- reg_mecr &= ~FLEXCAN_MECR_ECRWRDIS;
-- flexcan_write(reg_mecr, &regs->mecr);
-+ priv->write(reg_mecr, &regs->mecr);
- reg_mecr &= ~(FLEXCAN_MECR_NCEFAFRZ | FLEXCAN_MECR_HANCEI_MSK |
- FLEXCAN_MECR_FANCEI_MSK);
-- flexcan_write(reg_mecr, &regs->mecr);
-+ priv->write(reg_mecr, &regs->mecr);
- }
-
- err = flexcan_transceiver_enable(priv);
-@@ -941,13 +956,13 @@ static int flexcan_chip_start(struct net
-
- /* enable interrupts atomically */
- disable_irq(dev->irq);
-- flexcan_write(priv->reg_ctrl_default, &regs->ctrl);
-- flexcan_write(FLEXCAN_IFLAG_DEFAULT, &regs->imask1);
-+ priv->write(priv->reg_ctrl_default, &regs->ctrl);
-+ priv->write(FLEXCAN_IFLAG_DEFAULT, &regs->imask1);
- enable_irq(dev->irq);
-
- /* print chip status */
- netdev_dbg(dev, "%s: reading mcr=0x%08x ctrl=0x%08x\n", __func__,
-- flexcan_read(&regs->mcr), flexcan_read(&regs->ctrl));
-+ priv->read(&regs->mcr), priv->read(&regs->ctrl));
-
- return 0;
-
-@@ -972,8 +987,8 @@ static void flexcan_chip_stop(struct net
- flexcan_chip_disable(priv);
-
- /* Disable all interrupts */
-- flexcan_write(0, &regs->imask1);
-- flexcan_write(priv->reg_ctrl_default & ~FLEXCAN_CTRL_ERR_ALL,
-+ priv->write(0, &regs->imask1);
-+ priv->write(priv->reg_ctrl_default & ~FLEXCAN_CTRL_ERR_ALL,
- &regs->ctrl);
-
- flexcan_transceiver_disable(priv);
-@@ -1089,25 +1104,25 @@ static int register_flexcandev(struct ne
- err = flexcan_chip_disable(priv);
- if (err)
- goto out_disable_per;
-- reg = flexcan_read(&regs->ctrl);
-+ reg = priv->read(&regs->ctrl);
- reg |= FLEXCAN_CTRL_CLK_SRC;
-- flexcan_write(reg, &regs->ctrl);
-+ priv->write(reg, &regs->ctrl);
-
- err = flexcan_chip_enable(priv);
- if (err)
- goto out_chip_disable;
-
- /* set freeze, halt and activate FIFO, restrict register access */
-- reg = flexcan_read(&regs->mcr);
-+ reg = priv->read(&regs->mcr);
- reg |= FLEXCAN_MCR_FRZ | FLEXCAN_MCR_HALT |
- FLEXCAN_MCR_FEN | FLEXCAN_MCR_SUPV;
-- flexcan_write(reg, &regs->mcr);
-+ priv->write(reg, &regs->mcr);
-
- /* Currently we only support newer versions of this core
- * featuring a RX FIFO. Older cores found on some Coldfire
- * derivates are not yet supported.
- */
-- reg = flexcan_read(&regs->mcr);
-+ reg = priv->read(&regs->mcr);
- if (!(reg & FLEXCAN_MCR_FEN)) {
- netdev_err(dev, "Could not enable RX FIFO, unsupported core\n");
- err = -ENODEV;
-@@ -1135,8 +1150,12 @@ static void unregister_flexcandev(struct
- static const struct of_device_id flexcan_of_match[] = {
- { .compatible = "fsl,imx6q-flexcan", .data = &fsl_imx6q_devtype_data, },
- { .compatible = "fsl,imx28-flexcan", .data = &fsl_imx28_devtype_data, },
-+ { .compatible = "fsl,imx53-flexcan", .data = &fsl_p1010_devtype_data, },
-+ { .compatible = "fsl,imx35-flexcan", .data = &fsl_p1010_devtype_data, },
-+ { .compatible = "fsl,imx25-flexcan", .data = &fsl_p1010_devtype_data, },
- { .compatible = "fsl,p1010-flexcan", .data = &fsl_p1010_devtype_data, },
- { .compatible = "fsl,vf610-flexcan", .data = &fsl_vf610_devtype_data, },
-+ { .compatible = "fsl,ls1021ar2-flexcan", .data = &fsl_ls1021a_r2_devtype_data, },
- { /* sentinel */ },
- };
- MODULE_DEVICE_TABLE(of, flexcan_of_match);
-@@ -1213,6 +1232,21 @@ static int flexcan_probe(struct platform
- dev->flags |= IFF_ECHO;
-
- priv = netdev_priv(dev);
-+
-+ if (of_property_read_bool(pdev->dev.of_node, "big-endian")) {
-+ priv->read = flexcan_read_be;
-+ priv->write = flexcan_write_be;
-+ } else {
-+ if (of_device_is_compatible(pdev->dev.of_node,
-+ "fsl,p1010-flexcan")) {
-+ priv->read = flexcan_read_be;
-+ priv->write = flexcan_write_be;
-+ } else {
-+ priv->read = flexcan_read_le;
-+ priv->write = flexcan_write_le;
-+ }
-+ }
-+
- priv->can.clock.freq = clock_freq;
- priv->can.bittiming_const = &flexcan_bittiming_const;
- priv->can.do_set_mode = flexcan_set_mode;
diff --git a/target/linux/layerscape/patches-4.9/820-kvm-support-layerscape.patch b/target/linux/layerscape/patches-4.9/820-kvm-support-layerscape.patch
deleted file mode 100644
index a652660f9a..0000000000
--- a/target/linux/layerscape/patches-4.9/820-kvm-support-layerscape.patch
+++ /dev/null
@@ -1,239 +0,0 @@
-From fe22151c95c02c6bb145ea6c3685941e8fb09d60 Mon Sep 17 00:00:00 2001
-From: Yangbo Lu <yangbo.lu@nxp.com>
-Date: Thu, 5 Jul 2018 17:43:16 +0800
-Subject: [PATCH 32/32] kvm: support layerscape
-
-This is an integrated patch for layerscape kvm support.
-
-Signed-off-by: Laurentiu Tudor <laurentiu.tudor@nxp.com>
-Signed-off-by: Bharat Bhushan <Bharat.Bhushan@nxp.com>
-Signed-off-by: Ioana Ciornei <ioana.ciornei@nxp.com>
-Signed-off-by: Yangbo Lu <yangbo.lu@nxp.com>
----
- arch/arm/include/asm/kvm_mmu.h | 3 +-
- arch/arm/kvm/mmu.c | 56 ++++++++++++++++++++++++++++++--
- arch/arm64/include/asm/kvm_mmu.h | 14 ++++++--
- virt/kvm/arm/vgic/vgic-its.c | 24 +++++++++++---
- virt/kvm/arm/vgic/vgic-v2.c | 3 +-
- 5 files changed, 88 insertions(+), 12 deletions(-)
-
---- a/arch/arm/include/asm/kvm_mmu.h
-+++ b/arch/arm/include/asm/kvm_mmu.h
-@@ -55,7 +55,8 @@ void stage2_unmap_vm(struct kvm *kvm);
- int kvm_alloc_stage2_pgd(struct kvm *kvm);
- void kvm_free_stage2_pgd(struct kvm *kvm);
- int kvm_phys_addr_ioremap(struct kvm *kvm, phys_addr_t guest_ipa,
-- phys_addr_t pa, unsigned long size, bool writable);
-+ phys_addr_t pa, unsigned long size, bool writable,
-+ pgprot_t prot);
-
- int kvm_handle_guest_abort(struct kvm_vcpu *vcpu, struct kvm_run *run);
-
---- a/arch/arm/kvm/mmu.c
-+++ b/arch/arm/kvm/mmu.c
-@@ -1020,9 +1020,11 @@ static int stage2_pmdp_test_and_clear_yo
- * @guest_ipa: The IPA at which to insert the mapping
- * @pa: The physical address of the device
- * @size: The size of the mapping
-+ * @prot: S2 page translation bits
- */
- int kvm_phys_addr_ioremap(struct kvm *kvm, phys_addr_t guest_ipa,
-- phys_addr_t pa, unsigned long size, bool writable)
-+ phys_addr_t pa, unsigned long size, bool writable,
-+ pgprot_t prot)
- {
- phys_addr_t addr, end;
- int ret = 0;
-@@ -1033,7 +1035,7 @@ int kvm_phys_addr_ioremap(struct kvm *kv
- pfn = __phys_to_pfn(pa);
-
- for (addr = guest_ipa; addr < end; addr += PAGE_SIZE) {
-- pte_t pte = pfn_pte(pfn, PAGE_S2_DEVICE);
-+ pte_t pte = pfn_pte(pfn, prot);
-
- if (writable)
- pte = kvm_s2pte_mkwrite(pte);
-@@ -1057,6 +1059,30 @@ out:
- return ret;
- }
-
-+#ifdef CONFIG_ARM64
-+static pgprot_t stage1_to_stage2_pgprot(pgprot_t prot)
-+{
-+ switch (pgprot_val(prot) & PTE_ATTRINDX_MASK) {
-+ case PTE_ATTRINDX(MT_DEVICE_nGnRE):
-+ case PTE_ATTRINDX(MT_DEVICE_nGnRnE):
-+ case PTE_ATTRINDX(MT_DEVICE_GRE):
-+ return PAGE_S2_DEVICE;
-+ case PTE_ATTRINDX(MT_NORMAL_NC):
-+ case PTE_ATTRINDX(MT_NORMAL):
-+ return (pgprot_val(prot) & PTE_SHARED)
-+ ? PAGE_S2
-+ : PAGE_S2_NS;
-+ }
-+
-+ return PAGE_S2_DEVICE;
-+}
-+#else
-+static pgprot_t stage1_to_stage2_pgprot(pgprot_t prot)
-+{
-+ return PAGE_S2_DEVICE;
-+}
-+#endif
-+
- static bool transparent_hugepage_adjust(kvm_pfn_t *pfnp, phys_addr_t *ipap)
- {
- kvm_pfn_t pfn = *pfnp;
-@@ -1308,6 +1334,19 @@ static int user_mem_abort(struct kvm_vcp
- hugetlb = true;
- gfn = (fault_ipa & PMD_MASK) >> PAGE_SHIFT;
- } else {
-+ if (!is_vm_hugetlb_page(vma)) {
-+ pte_t *pte;
-+ spinlock_t *ptl;
-+ pgprot_t prot;
-+
-+ pte = get_locked_pte(current->mm, memslot->userspace_addr, &ptl);
-+ prot = stage1_to_stage2_pgprot(__pgprot(pte_val(*pte)));
-+ pte_unmap_unlock(pte, ptl);
-+#ifdef CONFIG_ARM64
-+ if (pgprot_val(prot) == pgprot_val(PAGE_S2_NS))
-+ mem_type = PAGE_S2_NS;
-+#endif
-+ }
- /*
- * Pages belonging to memslots that don't have the same
- * alignment for userspace and IPA cannot be mapped using
-@@ -1345,6 +1384,11 @@ static int user_mem_abort(struct kvm_vcp
- if (is_error_noslot_pfn(pfn))
- return -EFAULT;
-
-+#ifdef CONFIG_ARM64
-+ if (pgprot_val(mem_type) == pgprot_val(PAGE_S2_NS)) {
-+ flags |= KVM_S2PTE_FLAG_IS_IOMAP;
-+ } else
-+#endif
- if (kvm_is_device_pfn(pfn)) {
- mem_type = PAGE_S2_DEVICE;
- flags |= KVM_S2PTE_FLAG_IS_IOMAP;
-@@ -1882,6 +1926,9 @@ int kvm_arch_prepare_memory_region(struc
- gpa_t gpa = mem->guest_phys_addr +
- (vm_start - mem->userspace_addr);
- phys_addr_t pa;
-+ pgprot_t prot;
-+ pte_t *pte;
-+ spinlock_t *ptl;
-
- pa = (phys_addr_t)vma->vm_pgoff << PAGE_SHIFT;
- pa += vm_start - vma->vm_start;
-@@ -1891,10 +1938,13 @@ int kvm_arch_prepare_memory_region(struc
- ret = -EINVAL;
- goto out;
- }
-+ pte = get_locked_pte(current->mm, mem->userspace_addr, &ptl);
-+ prot = stage1_to_stage2_pgprot(__pgprot(pte_val(*pte)));
-+ pte_unmap_unlock(pte, ptl);
-
- ret = kvm_phys_addr_ioremap(kvm, gpa, pa,
- vm_end - vm_start,
-- writable);
-+ writable, prot);
- if (ret)
- break;
- }
---- a/arch/arm64/include/asm/kvm_mmu.h
-+++ b/arch/arm64/include/asm/kvm_mmu.h
-@@ -167,7 +167,8 @@ void stage2_unmap_vm(struct kvm *kvm);
- int kvm_alloc_stage2_pgd(struct kvm *kvm);
- void kvm_free_stage2_pgd(struct kvm *kvm);
- int kvm_phys_addr_ioremap(struct kvm *kvm, phys_addr_t guest_ipa,
-- phys_addr_t pa, unsigned long size, bool writable);
-+ phys_addr_t pa, unsigned long size, bool writable,
-+ pgprot_t prot);
-
- int kvm_handle_guest_abort(struct kvm_vcpu *vcpu, struct kvm_run *run);
-
-@@ -274,8 +275,15 @@ static inline void __coherent_cache_gues
-
- static inline void __kvm_flush_dcache_pte(pte_t pte)
- {
-- struct page *page = pte_page(pte);
-- kvm_flush_dcache_to_poc(page_address(page), PAGE_SIZE);
-+ if (pfn_valid(pte_pfn(pte))) {
-+ struct page *page = pte_page(pte);
-+ kvm_flush_dcache_to_poc(page_address(page), PAGE_SIZE);
-+ } else {
-+ void __iomem *va = ioremap_cache_ns(pte_pfn(pte) << PAGE_SHIFT, PAGE_SIZE);
-+
-+ kvm_flush_dcache_to_poc(va, PAGE_SIZE);
-+ iounmap(va);
-+ }
- }
-
- static inline void __kvm_flush_dcache_pmd(pmd_t pmd)
---- a/virt/kvm/arm/vgic/vgic-its.c
-+++ b/virt/kvm/arm/vgic/vgic-its.c
-@@ -176,6 +176,8 @@ static struct its_itte *find_itte(struct
-
- #define GIC_LPI_OFFSET 8192
-
-+#define VITS_TYPER_DEVBITS 17
-+
- /*
- * Finds and returns a collection in the ITS collection table.
- * Must be called with the its_lock mutex held.
-@@ -375,7 +377,7 @@ static unsigned long vgic_mmio_read_its_
- * To avoid memory waste in the guest, we keep the number of IDBits and
- * DevBits low - as least for the time being.
- */
-- reg |= 0x0f << GITS_TYPER_DEVBITS_SHIFT;
-+ reg |= GIC_ENCODE_SZ(VITS_TYPER_DEVBITS, 5) << GITS_TYPER_DEVBITS_SHIFT;
- reg |= 0x0f << GITS_TYPER_IDBITS_SHIFT;
-
- return extract_bytes(reg, addr & 7, len);
-@@ -601,16 +603,30 @@ static int vgic_its_cmd_handle_movi(stru
- * Check whether an ID can be stored into the corresponding guest table.
- * For a direct table this is pretty easy, but gets a bit nasty for
- * indirect tables. We check whether the resulting guest physical address
-- * is actually valid (covered by a memslot and guest accessbible).
-+ * is actually valid (covered by a memslot and guest accessible).
- * For this we have to read the respective first level entry.
- */
--static bool vgic_its_check_id(struct vgic_its *its, u64 baser, int id)
-+static bool vgic_its_check_id(struct vgic_its *its, u64 baser, u32 id)
- {
- int l1_tbl_size = GITS_BASER_NR_PAGES(baser) * SZ_64K;
-+ u64 indirect_ptr, type = GITS_BASER_TYPE(baser);
- int index;
-- u64 indirect_ptr;
- gfn_t gfn;
-
-+ switch (type) {
-+ case GITS_BASER_TYPE_DEVICE:
-+ if (id >= BIT_ULL(VITS_TYPER_DEVBITS))
-+ return false;
-+ break;
-+ case GITS_BASER_TYPE_COLLECTION:
-+ /* as GITS_TYPER.CIL == 0, ITS supports 16-bit collection ID */
-+ if (id >= BIT_ULL(16))
-+ return false;
-+ break;
-+ default:
-+ return false;
-+ }
-+
- if (!(baser & GITS_BASER_INDIRECT)) {
- phys_addr_t addr;
-
---- a/virt/kvm/arm/vgic/vgic-v2.c
-+++ b/virt/kvm/arm/vgic/vgic-v2.c
-@@ -290,7 +290,8 @@ int vgic_v2_map_resources(struct kvm *kv
- if (!static_branch_unlikely(&vgic_v2_cpuif_trap)) {
- ret = kvm_phys_addr_ioremap(kvm, dist->vgic_cpu_base,
- kvm_vgic_global_state.vcpu_base,
-- KVM_VGIC_V2_CPU_SIZE, true);
-+ KVM_VGIC_V2_CPU_SIZE, true,
-+ PAGE_S2_DEVICE);
- if (ret) {
- kvm_err("Unable to remap VGIC CPU to VCPU\n");
- goto out;
diff --git a/target/linux/layerscape/patches-4.9/821-add-esdhc-vsel-to-ls1043.patch b/target/linux/layerscape/patches-4.9/821-add-esdhc-vsel-to-ls1043.patch
deleted file mode 100644
index c45c6b4a22..0000000000
--- a/target/linux/layerscape/patches-4.9/821-add-esdhc-vsel-to-ls1043.patch
+++ /dev/null
@@ -1,10 +0,0 @@
---- a/drivers/mmc/host/sdhci-of-esdhc.c
-+++ b/drivers/mmc/host/sdhci-of-esdhc.c
-@@ -619,6 +619,7 @@ static void esdhc_reset(struct sdhci_hos
- static const struct of_device_id scfg_device_ids[] = {
- { .compatible = "fsl,t1040-scfg", },
- { .compatible = "fsl,ls1012a-scfg", },
-+ { .compatible = "fsl,ls1043a-scfg", },
- { .compatible = "fsl,ls1046a-scfg", },
- {}
- };
diff --git a/target/linux/layerscape/patches-4.9/822-rgmii-fixed-link.patch b/target/linux/layerscape/patches-4.9/822-rgmii-fixed-link.patch
deleted file mode 100644
index 593bd3e79a..0000000000
--- a/target/linux/layerscape/patches-4.9/822-rgmii-fixed-link.patch
+++ /dev/null
@@ -1,62 +0,0 @@
-From f27ef8941ca29b2d10428754be51e8ee06bb1263 Mon Sep 17 00:00:00 2001
-From: Mathew McBride <matt@traverse.com.au>
-Date: Mon, 7 Aug 2017 10:19:48 +1000
-Subject: [PATCH] Recognize when an RGMII Link is set as fixed (in the device
- tree) and set up the MAC accordingly
-
----
- drivers/net/ethernet/freescale/sdk_dpaa/mac.c | 1
- drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MAC/memac.c | 13 ++++++++++
- 2 files changed, 14 insertions(+)
-
---- a/drivers/net/ethernet/freescale/sdk_dpaa/mac.c
-+++ b/drivers/net/ethernet/freescale/sdk_dpaa/mac.c
-@@ -386,6 +386,7 @@ static int __cold mac_probe(struct platf
- mac_dev->fixed_link->duplex = phy->duplex;
- mac_dev->fixed_link->pause = phy->pause;
- mac_dev->fixed_link->asym_pause = phy->asym_pause;
-+ printk(KERN_INFO "Setting up fixed link, speed %d duplex %d\n", mac_dev->fixed_link->speed, mac_dev->fixed_link->duplex);
- }
-
- _errno = mac_dev->init(mac_dev);
---- a/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MAC/memac.c
-+++ b/drivers/net/ethernet/freescale/sdk_fman/Peripherals/FM/MAC/memac.c
-@@ -36,6 +36,8 @@
-
- @Description FM mEMAC driver
- *//***************************************************************************/
-+#include <../../../../sdk_dpaa/mac.h>
-+#include <linux/phy_fixed.h>
-
- #include "std_ext.h"
- #include "string_ext.h"
-@@ -48,6 +50,8 @@
- #include "memac.h"
-
-
-+static t_Error MemacAdjustLink(t_Handle h_Memac, e_EnetSpeed speed, bool fullDuplex);
-+
- /*****************************************************************************/
- /* Internal routines */
- /*****************************************************************************/
-@@ -276,11 +280,20 @@ static t_Error MemacEnable(t_Handle h_Me
- {
- t_Memac *p_Memac = (t_Memac *)h_Memac;
-
-+ struct mac_device *mac_dev = (struct mac_device *)p_Memac->h_App;
-+
- SANITY_CHECK_RETURN_ERROR(p_Memac, E_INVALID_HANDLE);
- SANITY_CHECK_RETURN_ERROR(!p_Memac->p_MemacDriverParam, E_INVALID_STATE);
-
- fman_memac_enable(p_Memac->p_MemMap, (mode & e_COMM_MODE_RX), (mode & e_COMM_MODE_TX));
-
-+ if (ENET_INTERFACE_FROM_MODE(p_Memac->enetMode) == e_ENET_IF_RGMII) {
-+ if (mac_dev->fixed_link) {
-+ printk(KERN_INFO "This is a fixed-link, forcing speed %d duplex %d\n",mac_dev->fixed_link->speed,mac_dev->fixed_link->duplex);
-+ MemacAdjustLink(h_Memac,mac_dev->fixed_link->speed,mac_dev->fixed_link->duplex);
-+ }
-+ }
-+
- return E_OK;
- }
-